diff --git a/Gopkg.lock b/Gopkg.lock index 1b8fccadd..a0e23656d 100644 --- a/Gopkg.lock +++ b/Gopkg.lock @@ -3,19 +3,15 @@ [[projects]] name = "github.com/DataDog/dd-trace-go" - packages = [ - "opentracing", - "tracer", - "tracer/ext" - ] + packages = ["opentracing","tracer","tracer/ext"] revision = "27617015d45e6cd550b9a7ac7715c37cc2f7d020" version = "v0.6.1" [[projects]] name = "github.com/Shopify/sarama" packages = ["."] - revision = "f7be6aa2bc7b2e38edf816b08b582782194a1c02" - version = "v1.16.0" + revision = "35324cf48e33d8260e1c7c18854465a904ade249" + version = "v1.17.0" [[projects]] name = "github.com/apache/thrift" @@ -24,61 +20,15 @@ [[projects]] name = "github.com/aws/aws-sdk-go" - packages = [ - "aws", - "aws/awserr", - "aws/awsutil", - "aws/client", - "aws/client/metadata", - "aws/corehandlers", - "aws/credentials", - "aws/credentials/ec2rolecreds", - "aws/credentials/endpointcreds", - "aws/credentials/stscreds", - "aws/defaults", - "aws/ec2metadata", - "aws/endpoints", - "aws/request", - "aws/session", - "aws/signer/v4", - "internal/sdkio", - "internal/sdkrand", - "internal/shareddefaults", - "private/protocol", - "private/protocol/query", - "private/protocol/query/queryutil", - "private/protocol/rest", - "private/protocol/restxml", - "private/protocol/xml/xmlutil", - "service/route53", - "service/route53/route53iface", - "service/sts" - ] - revision = "ee7b4b1162937cba700de23bd90acb742982e626" - version = "v1.13.50" + packages = ["aws","aws/awserr","aws/awsutil","aws/client","aws/client/metadata","aws/corehandlers","aws/credentials","aws/credentials/ec2rolecreds","aws/credentials/endpointcreds","aws/credentials/stscreds","aws/csm","aws/defaults","aws/ec2metadata","aws/endpoints","aws/request","aws/session","aws/signer/v4","internal/sdkio","internal/sdkrand","internal/shareddefaults","private/protocol","private/protocol/query","private/protocol/query/queryutil","private/protocol/rest","private/protocol/restxml","private/protocol/xml/xmlutil","service/route53","service/route53/route53iface","service/sts"] + revision = "852052a10992d92f68b9a60862a3312292524903" + version = "v1.14.17" [[projects]] name = "github.com/coreos/etcd" - packages = [ - "auth/authpb", - "client", - "clientv3", - "etcdserver/api/v3rpc/rpctypes", - "etcdserver/etcdserverpb", - "mvcc/mvccpb", - "pkg/pathutil", - "pkg/srv", - "pkg/types", - "version" - ] - revision = "70c8726202dd91e482fb4029fd14af1d4ed1d5af" - version = "v3.3.5" - -[[projects]] - name = "github.com/coreos/go-semver" - packages = ["semver"] - revision = "8ab6407b697782a06568d4b7f1db25550ec2e4c6" - version = "v0.2.0" + packages = ["auth/authpb","clientv3","etcdserver/api/v3rpc/rpctypes","etcdserver/etcdserverpb","mvcc/mvccpb","pkg/types"] + revision = "33245c6b5b49130ca99280408fadfab01aac0e48" + version = "v3.3.8" [[projects]] name = "github.com/davecgh/go-spew" @@ -125,8 +75,8 @@ [[projects]] name = "github.com/go-ini/ini" packages = ["."] - revision = "6529cf7c58879c08d927016dde4477f18a0634cb" - version = "v1.36.0" + revision = "06f5f3d67269ccec1fe5fe4134ba6e982984f7f5" + version = "v1.37.0" [[projects]] name = "github.com/go-logfmt/logfmt" @@ -136,12 +86,7 @@ [[projects]] name = "github.com/gogo/protobuf" - packages = [ - "gogoproto", - "proto", - "protoc-gen-gogo/descriptor", - "sortkeys" - ] + packages = ["gogoproto","proto","protoc-gen-gogo/descriptor","sortkeys"] revision = "1adfc126b41513cc696b209667c8656ea7aac67c" version = "v1.0.0" @@ -153,13 +98,7 @@ [[projects]] name = "github.com/golang/protobuf" - packages = [ - "proto", - "ptypes", - "ptypes/any", - "ptypes/duration", - "ptypes/timestamp" - ] + packages = ["proto","ptypes","ptypes/any","ptypes/duration","ptypes/timestamp"] revision = "b4deda0973fb4c70b50d226b1af49f3da59f5265" version = "v1.1.0" @@ -167,7 +106,7 @@ branch = "master" name = "github.com/golang/snappy" packages = ["."] - revision = "553a641470496b2327abcac10b36396bd98e45c9" + revision = "2e65f85255dbc3072edf28d6b5b8efc472979f5a" [[projects]] branch = "master" @@ -177,13 +116,9 @@ [[projects]] name = "github.com/googleapis/gnostic" - packages = [ - "OpenAPIv2", - "compiler", - "extensions" - ] - revision = "ee43cbb60db7bd22502942cccbc39059117352ab" - version = "v0.1.0" + packages = ["OpenAPIv2","compiler","extensions"] + revision = "7c663266750e7d82587642f65e60bc4083f1f84e" + version = "v0.2.0" [[projects]] branch = "master" @@ -194,10 +129,7 @@ [[projects]] branch = "master" name = "github.com/hashicorp/golang-lru" - packages = [ - ".", - "simplelru" - ] + packages = [".","simplelru"] revision = "0fb14efe8c47ae851c0034ed7a448854d3d34cf3" [[projects]] @@ -209,8 +141,8 @@ [[projects]] name = "github.com/imdario/mergo" packages = ["."] - revision = "9d5f1277e9a8ed20c3684bda8fde67c05628518c" - version = "v0.3.4" + revision = "9316a62528ac99aaecb4e47eadd6dc8aa6533d58" + version = "v0.3.5" [[projects]] name = "github.com/jmespath/go-jmespath" @@ -232,8 +164,8 @@ [[projects]] name = "github.com/matttproud/golang_protobuf_extensions" packages = ["pbutil"] - revision = "3247c84500bff8d9fb6d579d800f20b3e091582c" - version = "v1.0.0" + revision = "c12348ce28de40eed0136aa2b644d0ee0650e56c" + version = "v1.0.1" [[projects]] name = "github.com/modern-go/concurrent" @@ -255,38 +187,21 @@ [[projects]] name = "github.com/opentracing/opentracing-go" - packages = [ - ".", - "ext", - "log" - ] + packages = [".","ext","log"] revision = "1949ddbfd147afd4d964a9f00b24eb291e0e7c38" version = "v1.0.2" [[projects]] name = "github.com/openzipkin/zipkin-go-opentracing" - packages = [ - ".", - "flag", - "thrift/gen-go/scribe", - "thrift/gen-go/zipkincore", - "types", - "wire" - ] + packages = [".","flag","thrift/gen-go/scribe","thrift/gen-go/zipkincore","types","wire"] revision = "26cf9707480e6b90e5eff22cf0bbf05319154232" version = "v0.3.4" [[projects]] name = "github.com/pierrec/lz4" - packages = ["."] - revision = "2fcda4cb7018ce05a25959d2fe08c83e3329f169" - version = "v1.1" - -[[projects]] - name = "github.com/pierrec/xxHash" - packages = ["xxHash32"] - revision = "f051bb7f1d1aaf1b5a665d74fb6b0217712c69f7" - version = "v0.1.1" + packages = [".","internal/xxh32"] + revision = "1958fd8fff7f115e79725b1288e0b878b3e06b00" + version = "v2.0.3" [[projects]] branch = "master" @@ -297,12 +212,8 @@ [[projects]] branch = "master" name = "github.com/prometheus/common" - packages = [ - "expfmt", - "internal/bitbucket.org/ww/goautoneg", - "model" - ] - revision = "d811d2e9bf898806ecfb6ef6296774b13ffc314c" + packages = ["expfmt","internal/bitbucket.org/ww/goautoneg","model"] + revision = "7600349dcfe1abd18d72d3a1770870d9800a7801" [[projects]] branch = "master" @@ -325,49 +236,23 @@ branch = "master" name = "golang.org/x/crypto" packages = ["ssh/terminal"] - revision = "1a580b3eff7814fc9b40602fd35256c63b50f491" + revision = "a49355c7e3f8fe157a85be2f77e6e269a0f89602" [[projects]] branch = "master" name = "golang.org/x/net" - packages = [ - "context", - "http/httpguts", - "http2", - "http2/hpack", - "idna", - "internal/timeseries", - "trace" - ] - revision = "2491c5de3490fced2f6cff376127c667efeed857" + packages = ["context","http/httpguts","http2","http2/hpack","idna","internal/timeseries","trace"] + revision = "4cb1c02c05b0e749b0365f61ae859a8e0cfceed9" [[projects]] branch = "master" name = "golang.org/x/sys" - packages = [ - "unix", - "windows" - ] - revision = "7c87d13f8e835d2fb3a70a2912c811ed0c1d241b" + packages = ["unix","windows"] + revision = "7138fd3d9dc8335c567ca206f4333fb75eb05d56" [[projects]] name = "golang.org/x/text" - packages = [ - "collate", - "collate/build", - "internal/colltab", - "internal/gen", - "internal/tag", - "internal/triegen", - "internal/ucd", - "language", - "secure/bidirule", - "transform", - "unicode/bidi", - "unicode/cldr", - "unicode/norm", - "unicode/rangetable" - ] + packages = ["collate","collate/build","internal/colltab","internal/gen","internal/tag","internal/triegen","internal/ucd","language","secure/bidirule","transform","unicode/bidi","unicode/cldr","unicode/norm","unicode/rangetable"] revision = "f21a4dfb5e38f5895301dc265a8def02365cc3d0" version = "v0.3.0" @@ -381,39 +266,13 @@ branch = "master" name = "google.golang.org/genproto" packages = ["googleapis/rpc/status"] - revision = "7bb2a897381c9c5ab2aeb8614f758d7766af68ff" + revision = "ff3583edef7de132f219f0efc00e097cabcc0ec0" [[projects]] name = "google.golang.org/grpc" - packages = [ - ".", - "balancer", - "balancer/base", - "balancer/roundrobin", - "channelz", - "codes", - "connectivity", - "credentials", - "encoding", - "encoding/proto", - "grpclb/grpc_lb_v1/messages", - "grpclog", - "health/grpc_health_v1", - "internal", - "keepalive", - "metadata", - "naming", - "peer", - "resolver", - "resolver/dns", - "resolver/passthrough", - "stats", - "status", - "tap", - "transport" - ] - revision = "41344da2231b913fa3d983840a57a6b1b7b631a1" - version = "v1.12.0" + packages = [".","balancer","balancer/base","balancer/roundrobin","codes","connectivity","credentials","encoding","encoding/proto","grpclog","health/grpc_health_v1","internal","internal/backoff","internal/channelz","internal/grpcrand","keepalive","metadata","naming","peer","resolver","resolver/dns","resolver/passthrough","stats","status","tap","transport"] + revision = "168a6198bcb0ef175f7dacec0b8691fc141dc9b8" + version = "v1.13.0" [[projects]] name = "gopkg.in/inf.v0" @@ -429,145 +288,23 @@ [[projects]] name = "k8s.io/api" - packages = [ - "admissionregistration/v1alpha1", - "admissionregistration/v1beta1", - "apps/v1", - "apps/v1beta1", - "apps/v1beta2", - "authentication/v1", - "authentication/v1beta1", - "authorization/v1", - "authorization/v1beta1", - "autoscaling/v1", - "autoscaling/v2beta1", - "batch/v1", - "batch/v1beta1", - "batch/v2alpha1", - "certificates/v1beta1", - "core/v1", - "events/v1beta1", - "extensions/v1beta1", - "networking/v1", - "policy/v1beta1", - "rbac/v1", - "rbac/v1alpha1", - "rbac/v1beta1", - "scheduling/v1alpha1", - "settings/v1alpha1", - "storage/v1", - "storage/v1alpha1", - "storage/v1beta1" - ] + packages = ["admissionregistration/v1alpha1","admissionregistration/v1beta1","apps/v1","apps/v1beta1","apps/v1beta2","authentication/v1","authentication/v1beta1","authorization/v1","authorization/v1beta1","autoscaling/v1","autoscaling/v2beta1","batch/v1","batch/v1beta1","batch/v2alpha1","certificates/v1beta1","core/v1","events/v1beta1","extensions/v1beta1","networking/v1","policy/v1beta1","rbac/v1","rbac/v1alpha1","rbac/v1beta1","scheduling/v1alpha1","settings/v1alpha1","storage/v1","storage/v1alpha1","storage/v1beta1"] revision = "73d903622b7391f3312dcbac6483fed484e185f8" [[projects]] name = "k8s.io/apimachinery" - packages = [ - "pkg/api/errors", - "pkg/api/meta", - "pkg/api/resource", - "pkg/apis/meta/internalversion", - "pkg/apis/meta/v1", - "pkg/apis/meta/v1/unstructured", - "pkg/apis/meta/v1beta1", - "pkg/conversion", - "pkg/conversion/queryparams", - "pkg/fields", - "pkg/labels", - "pkg/runtime", - "pkg/runtime/schema", - "pkg/runtime/serializer", - "pkg/runtime/serializer/json", - "pkg/runtime/serializer/protobuf", - "pkg/runtime/serializer/recognizer", - "pkg/runtime/serializer/streaming", - "pkg/runtime/serializer/versioning", - "pkg/selection", - "pkg/types", - "pkg/util/cache", - "pkg/util/clock", - "pkg/util/diff", - "pkg/util/errors", - "pkg/util/framer", - "pkg/util/intstr", - "pkg/util/json", - "pkg/util/net", - "pkg/util/runtime", - "pkg/util/sets", - "pkg/util/validation", - "pkg/util/validation/field", - "pkg/util/wait", - "pkg/util/yaml", - "pkg/version", - "pkg/watch", - "third_party/forked/golang/reflect" - ] + packages = ["pkg/api/errors","pkg/api/meta","pkg/api/resource","pkg/apis/meta/internalversion","pkg/apis/meta/v1","pkg/apis/meta/v1/unstructured","pkg/apis/meta/v1beta1","pkg/conversion","pkg/conversion/queryparams","pkg/fields","pkg/labels","pkg/runtime","pkg/runtime/schema","pkg/runtime/serializer","pkg/runtime/serializer/json","pkg/runtime/serializer/protobuf","pkg/runtime/serializer/recognizer","pkg/runtime/serializer/streaming","pkg/runtime/serializer/versioning","pkg/selection","pkg/types","pkg/util/cache","pkg/util/clock","pkg/util/diff","pkg/util/errors","pkg/util/framer","pkg/util/intstr","pkg/util/json","pkg/util/net","pkg/util/runtime","pkg/util/sets","pkg/util/validation","pkg/util/validation/field","pkg/util/wait","pkg/util/yaml","pkg/version","pkg/watch","third_party/forked/golang/reflect"] revision = "302974c03f7e50f16561ba237db776ab93594ef6" [[projects]] name = "k8s.io/client-go" - packages = [ - "discovery", - "kubernetes", - "kubernetes/scheme", - "kubernetes/typed/admissionregistration/v1alpha1", - "kubernetes/typed/admissionregistration/v1beta1", - "kubernetes/typed/apps/v1", - "kubernetes/typed/apps/v1beta1", - "kubernetes/typed/apps/v1beta2", - "kubernetes/typed/authentication/v1", - "kubernetes/typed/authentication/v1beta1", - "kubernetes/typed/authorization/v1", - "kubernetes/typed/authorization/v1beta1", - "kubernetes/typed/autoscaling/v1", - "kubernetes/typed/autoscaling/v2beta1", - "kubernetes/typed/batch/v1", - "kubernetes/typed/batch/v1beta1", - "kubernetes/typed/batch/v2alpha1", - "kubernetes/typed/certificates/v1beta1", - "kubernetes/typed/core/v1", - "kubernetes/typed/events/v1beta1", - "kubernetes/typed/extensions/v1beta1", - "kubernetes/typed/networking/v1", - "kubernetes/typed/policy/v1beta1", - "kubernetes/typed/rbac/v1", - "kubernetes/typed/rbac/v1alpha1", - "kubernetes/typed/rbac/v1beta1", - "kubernetes/typed/scheduling/v1alpha1", - "kubernetes/typed/settings/v1alpha1", - "kubernetes/typed/storage/v1", - "kubernetes/typed/storage/v1alpha1", - "kubernetes/typed/storage/v1beta1", - "pkg/apis/clientauthentication", - "pkg/apis/clientauthentication/v1alpha1", - "pkg/version", - "plugin/pkg/client/auth/exec", - "rest", - "rest/watch", - "tools/auth", - "tools/cache", - "tools/clientcmd", - "tools/clientcmd/api", - "tools/clientcmd/api/latest", - "tools/clientcmd/api/v1", - "tools/metrics", - "tools/pager", - "tools/reference", - "transport", - "util/buffer", - "util/cert", - "util/flowcontrol", - "util/homedir", - "util/integer", - "util/retry" - ] + packages = ["discovery","kubernetes","kubernetes/scheme","kubernetes/typed/admissionregistration/v1alpha1","kubernetes/typed/admissionregistration/v1beta1","kubernetes/typed/apps/v1","kubernetes/typed/apps/v1beta1","kubernetes/typed/apps/v1beta2","kubernetes/typed/authentication/v1","kubernetes/typed/authentication/v1beta1","kubernetes/typed/authorization/v1","kubernetes/typed/authorization/v1beta1","kubernetes/typed/autoscaling/v1","kubernetes/typed/autoscaling/v2beta1","kubernetes/typed/batch/v1","kubernetes/typed/batch/v1beta1","kubernetes/typed/batch/v2alpha1","kubernetes/typed/certificates/v1beta1","kubernetes/typed/core/v1","kubernetes/typed/events/v1beta1","kubernetes/typed/extensions/v1beta1","kubernetes/typed/networking/v1","kubernetes/typed/policy/v1beta1","kubernetes/typed/rbac/v1","kubernetes/typed/rbac/v1alpha1","kubernetes/typed/rbac/v1beta1","kubernetes/typed/scheduling/v1alpha1","kubernetes/typed/settings/v1alpha1","kubernetes/typed/storage/v1","kubernetes/typed/storage/v1alpha1","kubernetes/typed/storage/v1beta1","pkg/apis/clientauthentication","pkg/apis/clientauthentication/v1alpha1","pkg/version","plugin/pkg/client/auth/exec","rest","rest/watch","tools/auth","tools/cache","tools/clientcmd","tools/clientcmd/api","tools/clientcmd/api/latest","tools/clientcmd/api/v1","tools/metrics","tools/pager","tools/reference","transport","util/buffer","util/cert","util/flowcontrol","util/homedir","util/integer","util/retry"] revision = "23781f4d6632d88e869066eaebb743857aa1ef9b" version = "v7.0.0" [solve-meta] analyzer-name = "dep" analyzer-version = 1 - inputs-digest = "435926fcc83a4f1a93fd257248f2b1256eaa5a212159b07743408b8cafdbffff" + inputs-digest = "cb1fb57b8b1053263afdf42b13581b0fe510bdc842616466ea7824563bfcf3ed" solver-name = "gps-cdcl" solver-version = 1 diff --git a/vendor/github.com/Shopify/sarama/.travis.yml b/vendor/github.com/Shopify/sarama/.travis.yml index cc38769fa..ea295ec5f 100644 --- a/vendor/github.com/Shopify/sarama/.travis.yml +++ b/vendor/github.com/Shopify/sarama/.travis.yml @@ -2,6 +2,7 @@ language: go go: - 1.8.x - 1.9.x +- 1.10.x env: global: @@ -11,9 +12,9 @@ env: - KAFKA_HOSTNAME=localhost - DEBUG=true matrix: - - KAFKA_VERSION=0.10.2.1 - KAFKA_VERSION=0.11.0.2 - KAFKA_VERSION=1.0.0 + - KAFKA_VERSION=1.1.0 before_install: - export REPOSITORY_ROOT=${TRAVIS_BUILD_DIR} diff --git a/vendor/github.com/Shopify/sarama/CHANGELOG.md b/vendor/github.com/Shopify/sarama/CHANGELOG.md index 836841650..16d5829c9 100644 --- a/vendor/github.com/Shopify/sarama/CHANGELOG.md +++ b/vendor/github.com/Shopify/sarama/CHANGELOG.md @@ -1,5 +1,43 @@ # Changelog +#### Version 1.17.0 (2018-05-30) + +New Features: + - Add support for gzip compression levels + ([#1044](https://github.com/Shopify/sarama/pull/1044)). + - Add support for Metadata request/response pairs versions v1 to v5 + ([#1047](https://github.com/Shopify/sarama/pull/1047), + [#1069](https://github.com/Shopify/sarama/pull/1069)). + - Add versioning to JoinGroup request/response pairs + ([#1098](https://github.com/Shopify/sarama/pull/1098)) + - Add support for CreatePartitions, DeleteGroups, DeleteRecords request/response pairs + ([#1065](https://github.com/Shopify/sarama/pull/1065), + [#1096](https://github.com/Shopify/sarama/pull/1096), + [#1027](https://github.com/Shopify/sarama/pull/1027)). + - Add `Controller()` method to Client interface + ([#1063](https://github.com/Shopify/sarama/pull/1063)). + +Improvements: + - ConsumerMetadataReq/Resp has been migrated to FindCoordinatorReq/Resp + ([#1010](https://github.com/Shopify/sarama/pull/1010)). + - Expose missing protocol parts: `msgSet` and `recordBatch` + ([#1049](https://github.com/Shopify/sarama/pull/1049)). + - Add support for v1 DeleteTopics Request + ([#1052](https://github.com/Shopify/sarama/pull/1052)). + - Add support for Go 1.10 + ([#1064](https://github.com/Shopify/sarama/pull/1064)). + - Claim support for Kafka 1.1.0 + ([#1073](https://github.com/Shopify/sarama/pull/1073)). + +Bug Fixes: + - Fix FindCoordinatorResponse.encode to allow nil Coordinator + ([#1050](https://github.com/Shopify/sarama/pull/1050), + [#1051](https://github.com/Shopify/sarama/pull/1051)). + - Clear all metadata when we have the latest topic info + ([#1033](https://github.com/Shopify/sarama/pull/1033)). + - Make `PartitionConsumer.Close` idempotent + ([#1092](https://github.com/Shopify/sarama/pull/1092)). + #### Version 1.16.0 (2018-02-12) New Features: diff --git a/vendor/github.com/Shopify/sarama/LICENSE b/vendor/github.com/Shopify/sarama/LICENSE index 8121b63b1..d2bf4352f 100644 --- a/vendor/github.com/Shopify/sarama/LICENSE +++ b/vendor/github.com/Shopify/sarama/LICENSE @@ -1,4 +1,4 @@ -Copyright (c) 2013 Evan Huus +Copyright (c) 2013 Shopify Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the diff --git a/vendor/github.com/Shopify/sarama/Makefile b/vendor/github.com/Shopify/sarama/Makefile index 58a39e4f3..b9a453dd2 100644 --- a/vendor/github.com/Shopify/sarama/Makefile +++ b/vendor/github.com/Shopify/sarama/Makefile @@ -4,7 +4,7 @@ default: fmt vet errcheck test test: echo "" > coverage.txt for d in `go list ./... | grep -v vendor`; do \ - go test -v -timeout 60s -race -coverprofile=profile.out -covermode=atomic $$d; \ + go test -p 1 -v -timeout 90s -race -coverprofile=profile.out -covermode=atomic $$d || exit 1; \ if [ -f profile.out ]; then \ cat profile.out >> coverage.txt; \ rm profile.out; \ @@ -14,8 +14,9 @@ test: vet: go vet ./... +# See https://github.com/kisielk/errcheck/pull/141 for details on ignorepkg errcheck: - errcheck github.com/Shopify/sarama/... + errcheck -ignorepkg fmt github.com/Shopify/sarama/... fmt: @if [ -n "$$(go fmt ./...)" ]; then echo 'Please run go fmt on your code.' && exit 1; fi diff --git a/vendor/github.com/Shopify/sarama/README.md b/vendor/github.com/Shopify/sarama/README.md index 28431f13e..4fc0cc600 100644 --- a/vendor/github.com/Shopify/sarama/README.md +++ b/vendor/github.com/Shopify/sarama/README.md @@ -21,7 +21,7 @@ You might also want to look at the [Frequently Asked Questions](https://github.c Sarama provides a "2 releases + 2 months" compatibility guarantee: we support the two latest stable releases of Kafka and Go, and we provide a two month grace period for older releases. This means we currently officially support -Go 1.9 and 1.8, and Kafka 1.0 through 0.10, although older releases are +Go 1.8 through 1.10, and Kafka 0.11 through 1.1, although older releases are still likely to work. Sarama follows semantic versioning and provides API stability via the gopkg.in service. diff --git a/vendor/github.com/Shopify/sarama/alter_configs_request_test.go b/vendor/github.com/Shopify/sarama/alter_configs_request_test.go index bfeff4491..b9407ca7e 100644 --- a/vendor/github.com/Shopify/sarama/alter_configs_request_test.go +++ b/vendor/github.com/Shopify/sarama/alter_configs_request_test.go @@ -31,11 +31,7 @@ var ( '1', '0', '0', '0', 2, // a topic 0, 3, 'b', 'a', 'r', // topic name: foo - 0, 0, 0, 2, //2 config - 0, 10, // 10 chars - 's', 'e', 'g', 'm', 'e', 'n', 't', '.', 'm', 's', - 0, 4, - '1', '0', '0', '0', + 0, 0, 0, 1, //2 config 0, 12, // 12 chars 'r', 'e', 't', 'e', 'n', 't', 'i', 'o', 'n', '.', 'm', 's', 0, 4, @@ -80,7 +76,6 @@ func TestAlterConfigsRequest(t *testing.T) { Type: TopicResource, Name: "bar", ConfigEntries: map[string]*string{ - "segment.ms": &configValue, "retention.ms": &configValue, }, }, diff --git a/vendor/github.com/Shopify/sarama/async_producer_test.go b/vendor/github.com/Shopify/sarama/async_producer_test.go index 716b654ea..478dca4cf 100644 --- a/vendor/github.com/Shopify/sarama/async_producer_test.go +++ b/vendor/github.com/Shopify/sarama/async_producer_test.go @@ -131,9 +131,12 @@ func TestAsyncProducer(t *testing.T) { if msg.Metadata.(int) != i { t.Error("Message metadata did not match") } + case <-time.After(time.Second): + t.Errorf("Timeout waiting for msg #%d", i) + goto done } } - +done: closeProducer(t, producer) leader.Close() seedBroker.Close() diff --git a/vendor/github.com/Shopify/sarama/broker.go b/vendor/github.com/Shopify/sarama/broker.go index b759f8f78..d836bee6d 100644 --- a/vendor/github.com/Shopify/sarama/broker.go +++ b/vendor/github.com/Shopify/sarama/broker.go @@ -18,6 +18,7 @@ import ( type Broker struct { id int32 addr string + rack *string conf *Config correlationID int32 @@ -230,6 +231,18 @@ func (b *Broker) GetConsumerMetadata(request *ConsumerMetadataRequest) (*Consume return response, nil } +func (b *Broker) FindCoordinator(request *FindCoordinatorRequest) (*FindCoordinatorResponse, error) { + response := new(FindCoordinatorResponse) + + err := b.sendAndReceive(request, response) + + if err != nil { + return nil, err + } + + return response, nil +} + func (b *Broker) GetAvailableOffsets(request *OffsetRequest) (*OffsetResponse, error) { response := new(OffsetResponse) @@ -373,6 +386,17 @@ func (b *Broker) ApiVersions(request *ApiVersionsRequest) (*ApiVersionsResponse, return response, nil } +func (b *Broker) CreatePartitions(request *CreatePartitionsRequest) (*CreatePartitionsResponse, error) { + response := new(CreatePartitionsResponse) + + err := b.sendAndReceive(request, response) + if err != nil { + return nil, err + } + + return response, nil +} + func (b *Broker) CreateTopics(request *CreateTopicsRequest) (*CreateTopicsResponse, error) { response := new(CreateTopicsResponse) @@ -395,6 +419,17 @@ func (b *Broker) DeleteTopics(request *DeleteTopicsRequest) (*DeleteTopicsRespon return response, nil } +func (b *Broker) DeleteRecords(request *DeleteRecordsRequest) (*DeleteRecordsResponse, error) { + response := new(DeleteRecordsResponse) + + err := b.sendAndReceive(request, response) + if err != nil { + return nil, err + } + + return response, nil +} + func (b *Broker) DescribeAcls(request *DescribeAclsRequest) (*DescribeAclsResponse, error) { response := new(DescribeAclsResponse) @@ -504,6 +539,17 @@ func (b *Broker) AlterConfigs(request *AlterConfigsRequest) (*AlterConfigsRespon return response, nil } + +func (b *Broker) DeleteGroups(request *DeleteGroupsRequest) (*DeleteGroupsResponse, error) { + response := new(DeleteGroupsResponse) + + if err := b.sendAndReceive(request, response); err != nil { + return nil, err + } + + return response, nil +} + func (b *Broker) send(rb protocolBody, promiseResponse bool) (*responsePromise, error) { b.lock.Lock() defer b.lock.Unlock() @@ -569,7 +615,7 @@ func (b *Broker) sendAndReceive(req protocolBody, res versionedDecoder) error { } } -func (b *Broker) decode(pd packetDecoder) (err error) { +func (b *Broker) decode(pd packetDecoder, version int16) (err error) { b.id, err = pd.getInt32() if err != nil { return err @@ -585,6 +631,13 @@ func (b *Broker) decode(pd packetDecoder) (err error) { return err } + if version >= 1 { + b.rack, err = pd.getNullableString() + if err != nil { + return err + } + } + b.addr = net.JoinHostPort(host, fmt.Sprint(port)) if _, _, err := net.SplitHostPort(b.addr); err != nil { return err @@ -593,7 +646,7 @@ func (b *Broker) decode(pd packetDecoder) (err error) { return nil } -func (b *Broker) encode(pe packetEncoder) (err error) { +func (b *Broker) encode(pe packetEncoder, version int16) (err error) { host, portstr, err := net.SplitHostPort(b.addr) if err != nil { @@ -613,6 +666,13 @@ func (b *Broker) encode(pe packetEncoder) (err error) { pe.putInt32(int32(port)) + if version >= 1 { + err = pe.putNullableString(b.rack) + if err != nil { + return err + } + } + return nil } diff --git a/vendor/github.com/Shopify/sarama/broker_test.go b/vendor/github.com/Shopify/sarama/broker_test.go index cc73b4440..9263cef8b 100644 --- a/vendor/github.com/Shopify/sarama/broker_test.go +++ b/vendor/github.com/Shopify/sarama/broker_test.go @@ -71,7 +71,7 @@ func TestSimpleBrokerCommunication(t *testing.T) { // Set the broker id in order to validate local broker metrics broker.id = 0 conf := NewConfig() - conf.Version = V0_10_0_0 + conf.Version = tt.version err := broker.Open(conf) if err != nil { t.Fatal(err) @@ -97,11 +97,13 @@ func TestSimpleBrokerCommunication(t *testing.T) { // We're not testing encoding/decoding here, so most of the requests/responses will be empty for simplicity's sake var brokerTestTable = []struct { + version KafkaVersion name string response []byte runner func(*testing.T, *Broker) }{ - {"MetadataRequest", + {V0_10_0_0, + "MetadataRequest", []byte{0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := MetadataRequest{} @@ -114,7 +116,8 @@ var brokerTestTable = []struct { } }}, - {"ConsumerMetadataRequest", + {V0_10_0_0, + "ConsumerMetadataRequest", []byte{0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 't', 0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := ConsumerMetadataRequest{} @@ -127,7 +130,8 @@ var brokerTestTable = []struct { } }}, - {"ProduceRequest (NoResponse)", + {V0_10_0_0, + "ProduceRequest (NoResponse)", []byte{}, func(t *testing.T, broker *Broker) { request := ProduceRequest{} @@ -141,7 +145,8 @@ var brokerTestTable = []struct { } }}, - {"ProduceRequest (WaitForLocal)", + {V0_10_0_0, + "ProduceRequest (WaitForLocal)", []byte{0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := ProduceRequest{} @@ -155,7 +160,8 @@ var brokerTestTable = []struct { } }}, - {"FetchRequest", + {V0_10_0_0, + "FetchRequest", []byte{0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := FetchRequest{} @@ -168,7 +174,8 @@ var brokerTestTable = []struct { } }}, - {"OffsetFetchRequest", + {V0_10_0_0, + "OffsetFetchRequest", []byte{0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := OffsetFetchRequest{} @@ -181,7 +188,8 @@ var brokerTestTable = []struct { } }}, - {"OffsetCommitRequest", + {V0_10_0_0, + "OffsetCommitRequest", []byte{0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := OffsetCommitRequest{} @@ -194,7 +202,8 @@ var brokerTestTable = []struct { } }}, - {"OffsetRequest", + {V0_10_0_0, + "OffsetRequest", []byte{0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := OffsetRequest{} @@ -207,7 +216,8 @@ var brokerTestTable = []struct { } }}, - {"JoinGroupRequest", + {V0_10_0_0, + "JoinGroupRequest", []byte{0x00, 0x17, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := JoinGroupRequest{} @@ -220,7 +230,8 @@ var brokerTestTable = []struct { } }}, - {"SyncGroupRequest", + {V0_10_0_0, + "SyncGroupRequest", []byte{0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := SyncGroupRequest{} @@ -233,7 +244,8 @@ var brokerTestTable = []struct { } }}, - {"LeaveGroupRequest", + {V0_10_0_0, + "LeaveGroupRequest", []byte{0x00, 0x00}, func(t *testing.T, broker *Broker) { request := LeaveGroupRequest{} @@ -246,7 +258,8 @@ var brokerTestTable = []struct { } }}, - {"HeartbeatRequest", + {V0_10_0_0, + "HeartbeatRequest", []byte{0x00, 0x00}, func(t *testing.T, broker *Broker) { request := HeartbeatRequest{} @@ -259,7 +272,8 @@ var brokerTestTable = []struct { } }}, - {"ListGroupsRequest", + {V0_10_0_0, + "ListGroupsRequest", []byte{0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := ListGroupsRequest{} @@ -272,7 +286,8 @@ var brokerTestTable = []struct { } }}, - {"DescribeGroupsRequest", + {V0_10_0_0, + "DescribeGroupsRequest", []byte{0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := DescribeGroupsRequest{} @@ -285,7 +300,8 @@ var brokerTestTable = []struct { } }}, - {"ApiVersionsRequest", + {V0_10_0_0, + "ApiVersionsRequest", []byte{0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, func(t *testing.T, broker *Broker) { request := ApiVersionsRequest{} @@ -297,6 +313,20 @@ var brokerTestTable = []struct { t.Error("ApiVersions request got no response!") } }}, + + {V1_1_0_0, + "DeleteGroupsRequest", + []byte{0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00}, + func(t *testing.T, broker *Broker) { + request := DeleteGroupsRequest{} + response, err := broker.DeleteGroups(&request) + if err != nil { + t.Error(err) + } + if response == nil { + t.Error("DeleteGroups request got no response!") + } + }}, } func validateBrokerMetrics(t *testing.T, broker *Broker, mockBrokerMetrics brokerMetrics) { diff --git a/vendor/github.com/Shopify/sarama/client.go b/vendor/github.com/Shopify/sarama/client.go index 3dbfc4b06..019cb4373 100644 --- a/vendor/github.com/Shopify/sarama/client.go +++ b/vendor/github.com/Shopify/sarama/client.go @@ -17,6 +17,9 @@ type Client interface { // altered after it has been created. Config() *Config + // Controller returns the cluster controller broker. + Controller() (*Broker, error) + // Brokers returns the current set of active brokers as retrieved from cluster metadata. Brokers() []*Broker @@ -97,6 +100,7 @@ type client struct { seedBrokers []*Broker deadSeeds []*Broker + controllerID int32 // cluster controller broker id brokers map[int32]*Broker // maps broker ids to brokers metadata map[string]map[int32]*PartitionMetadata // maps topics to partition ids to metadata coordinators map[string]int32 // Maps consumer group names to coordinating broker IDs @@ -379,6 +383,27 @@ func (client *client) GetOffset(topic string, partitionID int32, time int64) (in return offset, err } +func (client *client) Controller() (*Broker, error) { + if client.Closed() { + return nil, ErrClosedClient + } + + controller := client.cachedController() + if controller == nil { + if err := client.refreshMetadata(); err != nil { + return nil, err + } + controller = client.cachedController() + } + + if controller == nil { + return nil, ErrControllerNotAvailable + } + + _ = controller.Open(client.conf) + return controller, nil +} + func (client *client) Coordinator(consumerGroup string) (*Broker, error) { if client.Closed() { return nil, ErrClosedClient @@ -607,20 +632,7 @@ func (client *client) backgroundMetadataUpdater() { for { select { case <-ticker.C: - topics := []string{} - if !client.conf.Metadata.Full { - if specificTopics, err := client.Topics(); err != nil { - Logger.Println("Client background metadata topic load:", err) - break - } else if len(specificTopics) == 0 { - Logger.Println("Client background metadata update: no specific topics to update") - break - } else { - topics = specificTopics - } - } - - if err := client.RefreshMetadata(topics...); err != nil { + if err := client.refreshMetadata(); err != nil { Logger.Println("Client background metadata update:", err) } case <-client.closer: @@ -629,6 +641,26 @@ func (client *client) backgroundMetadataUpdater() { } } +func (client *client) refreshMetadata() error { + topics := []string{} + + if !client.conf.Metadata.Full { + if specificTopics, err := client.Topics(); err != nil { + return err + } else if len(specificTopics) == 0 { + return ErrNoTopicsToUpdateMetadata + } else { + topics = specificTopics + } + } + + if err := client.RefreshMetadata(topics...); err != nil { + return err + } + + return nil +} + func (client *client) tryRefreshMetadata(topics []string, attemptsRemaining int) error { retry := func(err error) error { if attemptsRemaining > 0 { @@ -645,12 +677,18 @@ func (client *client) tryRefreshMetadata(topics []string, attemptsRemaining int) } else { Logger.Printf("client/metadata fetching metadata for all topics from broker %s\n", broker.addr) } - response, err := broker.GetMetadata(&MetadataRequest{Topics: topics}) + + req := &MetadataRequest{Topics: topics} + if client.conf.Version.IsAtLeast(V0_10_0_0) { + req.Version = 1 + } + response, err := broker.GetMetadata(req) switch err.(type) { case nil: + allKnownMetaData := len(topics) == 0 // valid response, use it - shouldRetry, err := client.updateMetadata(response) + shouldRetry, err := client.updateMetadata(response, allKnownMetaData) if shouldRetry { Logger.Println("client/metadata found some partitions to be leaderless") return retry(err) // note: err can be nil @@ -674,7 +712,7 @@ func (client *client) tryRefreshMetadata(topics []string, attemptsRemaining int) } // if no fatal error, returns a list of topics that need retrying due to ErrLeaderNotAvailable -func (client *client) updateMetadata(data *MetadataResponse) (retry bool, err error) { +func (client *client) updateMetadata(data *MetadataResponse, allKnownMetaData bool) (retry bool, err error) { client.lock.Lock() defer client.lock.Unlock() @@ -686,6 +724,12 @@ func (client *client) updateMetadata(data *MetadataResponse) (retry bool, err er client.registerBroker(broker) } + client.controllerID = data.ControllerID + + if allKnownMetaData { + client.metadata = make(map[string]map[int32]*PartitionMetadata) + client.cachedPartitionsResults = make(map[string][maxPartitionIndex][]int32) + } for _, topic := range data.Topics { delete(client.metadata, topic.Name) delete(client.cachedPartitionsResults, topic.Name) @@ -735,8 +779,15 @@ func (client *client) cachedCoordinator(consumerGroup string) *Broker { return nil } -func (client *client) getConsumerMetadata(consumerGroup string, attemptsRemaining int) (*ConsumerMetadataResponse, error) { - retry := func(err error) (*ConsumerMetadataResponse, error) { +func (client *client) cachedController() *Broker { + client.lock.RLock() + defer client.lock.RUnlock() + + return client.brokers[client.controllerID] +} + +func (client *client) getConsumerMetadata(consumerGroup string, attemptsRemaining int) (*FindCoordinatorResponse, error) { + retry := func(err error) (*FindCoordinatorResponse, error) { if attemptsRemaining > 0 { Logger.Printf("client/coordinator retrying after %dms... (%d attempts remaining)\n", client.conf.Metadata.Retry.Backoff/time.Millisecond, attemptsRemaining) time.Sleep(client.conf.Metadata.Retry.Backoff) @@ -748,10 +799,11 @@ func (client *client) getConsumerMetadata(consumerGroup string, attemptsRemainin for broker := client.any(); broker != nil; broker = client.any() { Logger.Printf("client/coordinator requesting coordinator for consumergroup %s from %s\n", consumerGroup, broker.Addr()) - request := new(ConsumerMetadataRequest) - request.ConsumerGroup = consumerGroup + request := new(FindCoordinatorRequest) + request.CoordinatorKey = consumerGroup + request.CoordinatorType = CoordinatorGroup - response, err := broker.GetConsumerMetadata(request) + response, err := broker.FindCoordinator(request) if err != nil { Logger.Printf("client/coordinator request to broker %s failed: %s\n", broker.Addr(), err) diff --git a/vendor/github.com/Shopify/sarama/client_test.go b/vendor/github.com/Shopify/sarama/client_test.go index 2e1198d27..fc255a730 100644 --- a/vendor/github.com/Shopify/sarama/client_test.go +++ b/vendor/github.com/Shopify/sarama/client_test.go @@ -444,6 +444,47 @@ func TestClientResurrectDeadSeeds(t *testing.T) { safeClose(t, c) } +func TestClientController(t *testing.T) { + seedBroker := NewMockBroker(t, 1) + defer seedBroker.Close() + controllerBroker := NewMockBroker(t, 2) + defer controllerBroker.Close() + + seedBroker.SetHandlerByMap(map[string]MockResponse{ + "MetadataRequest": NewMockMetadataResponse(t). + SetController(controllerBroker.BrokerID()). + SetBroker(seedBroker.Addr(), seedBroker.BrokerID()). + SetBroker(controllerBroker.Addr(), controllerBroker.BrokerID()), + }) + + cfg := NewConfig() + + // test kafka version greater than 0.10.0.0 + cfg.Version = V0_10_0_0 + client1, err := NewClient([]string{seedBroker.Addr()}, cfg) + if err != nil { + t.Fatal(err) + } + defer safeClose(t, client1) + broker, err := client1.Controller() + if err != nil { + t.Fatal(err) + } + if broker.Addr() != controllerBroker.Addr() { + t.Errorf("Expected controller to have address %s, found %s", controllerBroker.Addr(), broker.Addr()) + } + + // test kafka version earlier than 0.10.0.0 + cfg.Version = V0_9_0_1 + client2, err := NewClient([]string{seedBroker.Addr()}, cfg) + if err != nil { + t.Fatal(err) + } + defer safeClose(t, client2) + if _, err = client2.Controller(); err != ErrControllerNotAvailable { + t.Errorf("Expected Contoller() to return %s, found %s", ErrControllerNotAvailable, err) + } +} func TestClientCoordinatorWithConsumerOffsetsTopic(t *testing.T) { seedBroker := NewMockBroker(t, 1) staleCoordinator := NewMockBroker(t, 2) diff --git a/vendor/github.com/Shopify/sarama/config.go b/vendor/github.com/Shopify/sarama/config.go index 29ea5c2b3..a564b5c23 100644 --- a/vendor/github.com/Shopify/sarama/config.go +++ b/vendor/github.com/Shopify/sarama/config.go @@ -1,7 +1,10 @@ package sarama import ( + "compress/gzip" "crypto/tls" + "fmt" + "io/ioutil" "regexp" "time" @@ -99,6 +102,10 @@ type Config struct { // The type of compression to use on messages (defaults to no compression). // Similar to `compression.codec` setting of the JVM producer. Compression CompressionCodec + // The level of compression to use on messages. The meaning depends + // on the actual compression type used and defaults to default compression + // level for the codec. + CompressionLevel int // Generates partitioners for choosing the partition to send messages to // (defaults to hashing the message key). Similar to the `partitioner.class` // setting for the JVM producer. @@ -290,6 +297,7 @@ func NewConfig() *Config { c.Producer.Retry.Max = 3 c.Producer.Retry.Backoff = 100 * time.Millisecond c.Producer.Return.Errors = true + c.Producer.CompressionLevel = CompressionLevelDefault c.Consumer.Fetch.Min = 1 c.Consumer.Fetch.Default = 1024 * 1024 @@ -302,7 +310,7 @@ func NewConfig() *Config { c.ClientID = defaultClientID c.ChannelBufferSize = 256 - c.Version = minVersion + c.Version = MinVersion c.MetricRegistry = metrics.NewRegistry() return c @@ -409,6 +417,14 @@ func (c *Config) Validate() error { return ConfigurationError("lz4 compression requires Version >= V0_10_0_0") } + if c.Producer.Compression == CompressionGZIP { + if c.Producer.CompressionLevel != CompressionLevelDefault { + if _, err := gzip.NewWriterLevel(ioutil.Discard, c.Producer.CompressionLevel); err != nil { + return ConfigurationError(fmt.Sprintf("gzip compression does not work with level %d: %v", c.Producer.CompressionLevel, err)) + } + } + } + // validate the Consumer values switch { case c.Consumer.Fetch.Min <= 0: diff --git a/vendor/github.com/Shopify/sarama/consumer.go b/vendor/github.com/Shopify/sarama/consumer.go index 48d231cf9..96226ac5b 100644 --- a/vendor/github.com/Shopify/sarama/consumer.go +++ b/vendor/github.com/Shopify/sarama/consumer.go @@ -310,6 +310,7 @@ type partitionConsumer struct { trigger, dying chan none responseResult error + closeOnce sync.Once fetchSize int32 offset int64 @@ -412,7 +413,9 @@ func (child *partitionConsumer) AsyncClose() { // the dispatcher to exit its loop, which removes it from the consumer then closes its 'messages' and // 'errors' channel (alternatively, if the child is already at the dispatcher for some reason, that will // also just close itself) - close(child.dying) + child.closeOnce.Do(func() { + close(child.dying) + }) } func (child *partitionConsumer) Close() error { @@ -461,7 +464,6 @@ feederLoop: child.messages <- msg } child.broker.input <- child - expiryTicker.Stop() continue feederLoop } else { // current message has not been sent, return to select @@ -482,9 +484,6 @@ feederLoop: func (child *partitionConsumer) parseMessages(msgSet *MessageSet) ([]*ConsumerMessage, error) { var messages []*ConsumerMessage - var incomplete bool - prelude := true - for _, msgBlock := range msgSet.Messages { for _, msg := range msgBlock.Messages() { offset := msg.Offset @@ -492,29 +491,22 @@ func (child *partitionConsumer) parseMessages(msgSet *MessageSet) ([]*ConsumerMe baseOffset := msgBlock.Offset - msgBlock.Messages()[len(msgBlock.Messages())-1].Offset offset += baseOffset } - if prelude && offset < child.offset { + if offset < child.offset { continue } - prelude = false - - if offset >= child.offset { - messages = append(messages, &ConsumerMessage{ - Topic: child.topic, - Partition: child.partition, - Key: msg.Msg.Key, - Value: msg.Msg.Value, - Offset: offset, - Timestamp: msg.Msg.Timestamp, - BlockTimestamp: msgBlock.Msg.Timestamp, - }) - child.offset = offset + 1 - } else { - incomplete = true - } + messages = append(messages, &ConsumerMessage{ + Topic: child.topic, + Partition: child.partition, + Key: msg.Msg.Key, + Value: msg.Msg.Value, + Offset: offset, + Timestamp: msg.Msg.Timestamp, + BlockTimestamp: msgBlock.Msg.Timestamp, + }) + child.offset = offset + 1 } } - - if incomplete || len(messages) == 0 { + if len(messages) == 0 { return nil, ErrIncompleteResponse } return messages, nil @@ -522,42 +514,25 @@ func (child *partitionConsumer) parseMessages(msgSet *MessageSet) ([]*ConsumerMe func (child *partitionConsumer) parseRecords(batch *RecordBatch) ([]*ConsumerMessage, error) { var messages []*ConsumerMessage - var incomplete bool - prelude := true - originalOffset := child.offset - for _, rec := range batch.Records { offset := batch.FirstOffset + rec.OffsetDelta - if prelude && offset < child.offset { + if offset < child.offset { continue } - prelude = false - - if offset >= child.offset { - messages = append(messages, &ConsumerMessage{ - Topic: child.topic, - Partition: child.partition, - Key: rec.Key, - Value: rec.Value, - Offset: offset, - Timestamp: batch.FirstTimestamp.Add(rec.TimestampDelta), - Headers: rec.Headers, - }) - child.offset = offset + 1 - } else { - incomplete = true - } + messages = append(messages, &ConsumerMessage{ + Topic: child.topic, + Partition: child.partition, + Key: rec.Key, + Value: rec.Value, + Offset: offset, + Timestamp: batch.FirstTimestamp.Add(rec.TimestampDelta), + Headers: rec.Headers, + }) + child.offset = offset + 1 } - - if incomplete { + if len(messages) == 0 { return nil, ErrIncompleteResponse } - - child.offset = batch.FirstOffset + int64(batch.LastOffsetDelta) + 1 - if child.offset <= originalOffset { - return nil, ErrConsumerOffsetNotAdvanced - } - return messages, nil } @@ -610,14 +585,14 @@ func (child *partitionConsumer) parseResponse(response *FetchResponse) ([]*Consu switch records.recordsType { case legacyRecords: - messageSetMessages, err := child.parseMessages(records.msgSet) + messageSetMessages, err := child.parseMessages(records.MsgSet) if err != nil { return nil, err } messages = append(messages, messageSetMessages...) case defaultRecords: - recordBatchMessages, err := child.parseRecords(records.recordBatch) + recordBatchMessages, err := child.parseRecords(records.RecordBatch) if err != nil { return nil, err } diff --git a/vendor/github.com/Shopify/sarama/consumer_metadata_request.go b/vendor/github.com/Shopify/sarama/consumer_metadata_request.go index 483be3354..4de45e7bf 100644 --- a/vendor/github.com/Shopify/sarama/consumer_metadata_request.go +++ b/vendor/github.com/Shopify/sarama/consumer_metadata_request.go @@ -5,12 +5,19 @@ type ConsumerMetadataRequest struct { } func (r *ConsumerMetadataRequest) encode(pe packetEncoder) error { - return pe.putString(r.ConsumerGroup) + tmp := new(FindCoordinatorRequest) + tmp.CoordinatorKey = r.ConsumerGroup + tmp.CoordinatorType = CoordinatorGroup + return tmp.encode(pe) } func (r *ConsumerMetadataRequest) decode(pd packetDecoder, version int16) (err error) { - r.ConsumerGroup, err = pd.getString() - return err + tmp := new(FindCoordinatorRequest) + if err := tmp.decode(pd, version); err != nil { + return err + } + r.ConsumerGroup = tmp.CoordinatorKey + return nil } func (r *ConsumerMetadataRequest) key() int16 { diff --git a/vendor/github.com/Shopify/sarama/consumer_metadata_request_test.go b/vendor/github.com/Shopify/sarama/consumer_metadata_request_test.go index 4509631a0..24e5f0a43 100644 --- a/vendor/github.com/Shopify/sarama/consumer_metadata_request_test.go +++ b/vendor/github.com/Shopify/sarama/consumer_metadata_request_test.go @@ -1,6 +1,8 @@ package sarama -import "testing" +import ( + "testing" +) var ( consumerMetadataRequestEmpty = []byte{ @@ -12,8 +14,10 @@ var ( func TestConsumerMetadataRequest(t *testing.T) { request := new(ConsumerMetadataRequest) - testRequest(t, "empty string", request, consumerMetadataRequestEmpty) + testEncodable(t, "empty string", request, consumerMetadataRequestEmpty) + testVersionDecodable(t, "empty string", request, consumerMetadataRequestEmpty, 0) request.ConsumerGroup = "foobar" - testRequest(t, "with string", request, consumerMetadataRequestString) + testEncodable(t, "with string", request, consumerMetadataRequestString) + testVersionDecodable(t, "with string", request, consumerMetadataRequestString, 0) } diff --git a/vendor/github.com/Shopify/sarama/consumer_metadata_response.go b/vendor/github.com/Shopify/sarama/consumer_metadata_response.go index 6b9632bba..442cbde7a 100644 --- a/vendor/github.com/Shopify/sarama/consumer_metadata_response.go +++ b/vendor/github.com/Shopify/sarama/consumer_metadata_response.go @@ -14,20 +14,18 @@ type ConsumerMetadataResponse struct { } func (r *ConsumerMetadataResponse) decode(pd packetDecoder, version int16) (err error) { - tmp, err := pd.getInt16() - if err != nil { - return err - } - r.Err = KError(tmp) + tmp := new(FindCoordinatorResponse) - coordinator := new(Broker) - if err := coordinator.decode(pd); err != nil { + if err := tmp.decode(pd, version); err != nil { return err } - if coordinator.addr == ":0" { + + r.Err = tmp.Err + + r.Coordinator = tmp.Coordinator + if tmp.Coordinator == nil { return nil } - r.Coordinator = coordinator // this can all go away in 2.0, but we have to fill in deprecated fields to maintain // backwards compatibility @@ -47,28 +45,22 @@ func (r *ConsumerMetadataResponse) decode(pd packetDecoder, version int16) (err } func (r *ConsumerMetadataResponse) encode(pe packetEncoder) error { - pe.putInt16(int16(r.Err)) - if r.Coordinator != nil { - host, portstr, err := net.SplitHostPort(r.Coordinator.Addr()) - if err != nil { - return err - } - port, err := strconv.ParseInt(portstr, 10, 32) - if err != nil { - return err - } - pe.putInt32(r.Coordinator.ID()) - if err := pe.putString(host); err != nil { - return err - } - pe.putInt32(int32(port)) - return nil + if r.Coordinator == nil { + r.Coordinator = new(Broker) + r.Coordinator.id = r.CoordinatorID + r.Coordinator.addr = net.JoinHostPort(r.CoordinatorHost, strconv.Itoa(int(r.CoordinatorPort))) } - pe.putInt32(r.CoordinatorID) - if err := pe.putString(r.CoordinatorHost); err != nil { + + tmp := &FindCoordinatorResponse{ + Version: 0, + Err: r.Err, + Coordinator: r.Coordinator, + } + + if err := tmp.encode(pe); err != nil { return err } - pe.putInt32(r.CoordinatorPort) + return nil } diff --git a/vendor/github.com/Shopify/sarama/consumer_metadata_response_test.go b/vendor/github.com/Shopify/sarama/consumer_metadata_response_test.go index b748784d7..8482f6ff1 100644 --- a/vendor/github.com/Shopify/sarama/consumer_metadata_response_test.go +++ b/vendor/github.com/Shopify/sarama/consumer_metadata_response_test.go @@ -17,8 +17,17 @@ var ( ) func TestConsumerMetadataResponseError(t *testing.T) { - response := ConsumerMetadataResponse{Err: ErrOffsetsLoadInProgress} - testResponse(t, "error", &response, consumerMetadataResponseError) + response := &ConsumerMetadataResponse{Err: ErrOffsetsLoadInProgress} + testEncodable(t, "", response, consumerMetadataResponseError) + + decodedResp := &ConsumerMetadataResponse{} + if err := versionedDecode(consumerMetadataResponseError, decodedResp, 0); err != nil { + t.Error("could not decode: ", err) + } + + if decodedResp.Err != ErrOffsetsLoadInProgress { + t.Errorf("got %s, want %s", decodedResp.Err, ErrOffsetsLoadInProgress) + } } func TestConsumerMetadataResponseSuccess(t *testing.T) { diff --git a/vendor/github.com/Shopify/sarama/delete_groups_request.go b/vendor/github.com/Shopify/sarama/delete_groups_request.go new file mode 100644 index 000000000..305a324ac --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_groups_request.go @@ -0,0 +1,30 @@ +package sarama + +type DeleteGroupsRequest struct { + Groups []string +} + +func (r *DeleteGroupsRequest) encode(pe packetEncoder) error { + return pe.putStringArray(r.Groups) +} + +func (r *DeleteGroupsRequest) decode(pd packetDecoder, version int16) (err error) { + r.Groups, err = pd.getStringArray() + return +} + +func (r *DeleteGroupsRequest) key() int16 { + return 42 +} + +func (r *DeleteGroupsRequest) version() int16 { + return 0 +} + +func (r *DeleteGroupsRequest) requiredVersion() KafkaVersion { + return V1_1_0_0 +} + +func (r *DeleteGroupsRequest) AddGroup(group string) { + r.Groups = append(r.Groups, group) +} diff --git a/vendor/github.com/Shopify/sarama/delete_groups_request_test.go b/vendor/github.com/Shopify/sarama/delete_groups_request_test.go new file mode 100644 index 000000000..908172498 --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_groups_request_test.go @@ -0,0 +1,34 @@ +package sarama + +import "testing" + +var ( + emptyDeleteGroupsRequest = []byte{0, 0, 0, 0} + + singleDeleteGroupsRequest = []byte{ + 0, 0, 0, 1, // 1 group + 0, 3, 'f', 'o', 'o', // group name: foo + } + + doubleDeleteGroupsRequest = []byte{ + 0, 0, 0, 2, // 2 groups + 0, 3, 'f', 'o', 'o', // group name: foo + 0, 3, 'b', 'a', 'r', // group name: foo + } +) + +func TestDeleteGroupsRequest(t *testing.T) { + var request *DeleteGroupsRequest + + request = new(DeleteGroupsRequest) + testRequest(t, "no groups", request, emptyDeleteGroupsRequest) + + request = new(DeleteGroupsRequest) + request.AddGroup("foo") + testRequest(t, "one group", request, singleDeleteGroupsRequest) + + request = new(DeleteGroupsRequest) + request.AddGroup("foo") + request.AddGroup("bar") + testRequest(t, "two groups", request, doubleDeleteGroupsRequest) +} diff --git a/vendor/github.com/Shopify/sarama/delete_groups_response.go b/vendor/github.com/Shopify/sarama/delete_groups_response.go new file mode 100644 index 000000000..c067ebb42 --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_groups_response.go @@ -0,0 +1,70 @@ +package sarama + +import ( + "time" +) + +type DeleteGroupsResponse struct { + ThrottleTime time.Duration + GroupErrorCodes map[string]KError +} + +func (r *DeleteGroupsResponse) encode(pe packetEncoder) error { + pe.putInt32(int32(r.ThrottleTime / time.Millisecond)) + + if err := pe.putArrayLength(len(r.GroupErrorCodes)); err != nil { + return err + } + for groupID, errorCode := range r.GroupErrorCodes { + if err := pe.putString(groupID); err != nil { + return err + } + pe.putInt16(int16(errorCode)) + } + + return nil +} + +func (r *DeleteGroupsResponse) decode(pd packetDecoder, version int16) error { + throttleTime, err := pd.getInt32() + if err != nil { + return err + } + r.ThrottleTime = time.Duration(throttleTime) * time.Millisecond + + n, err := pd.getArrayLength() + if err != nil { + return err + } + if n == 0 { + return nil + } + + r.GroupErrorCodes = make(map[string]KError, n) + for i := 0; i < n; i++ { + groupID, err := pd.getString() + if err != nil { + return err + } + errorCode, err := pd.getInt16() + if err != nil { + return err + } + + r.GroupErrorCodes[groupID] = KError(errorCode) + } + + return nil +} + +func (r *DeleteGroupsResponse) key() int16 { + return 42 +} + +func (r *DeleteGroupsResponse) version() int16 { + return 0 +} + +func (r *DeleteGroupsResponse) requiredVersion() KafkaVersion { + return V1_1_0_0 +} diff --git a/vendor/github.com/Shopify/sarama/delete_groups_response_test.go b/vendor/github.com/Shopify/sarama/delete_groups_response_test.go new file mode 100644 index 000000000..6f622b5f0 --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_groups_response_test.go @@ -0,0 +1,57 @@ +package sarama + +import ( + "testing" +) + +var ( + emptyDeleteGroupsResponse = []byte{ + 0, 0, 0, 0, // does not violate any quota + 0, 0, 0, 0, // no groups + } + + errorDeleteGroupsResponse = []byte{ + 0, 0, 0, 0, // does not violate any quota + 0, 0, 0, 1, // 1 group + 0, 3, 'f', 'o', 'o', // group name + 0, 31, // error ErrClusterAuthorizationFailed + } + + noErrorDeleteGroupsResponse = []byte{ + 0, 0, 0, 0, // does not violate any quota + 0, 0, 0, 1, // 1 group + 0, 3, 'f', 'o', 'o', // group name + 0, 0, // no error + } +) + +func TestDeleteGroupsResponse(t *testing.T) { + var response *DeleteGroupsResponse + + response = new(DeleteGroupsResponse) + testVersionDecodable(t, "empty", response, emptyDeleteGroupsResponse, 0) + if response.ThrottleTime != 0 { + t.Error("Expected no violation") + } + if len(response.GroupErrorCodes) != 0 { + t.Error("Expected no groups") + } + + response = new(DeleteGroupsResponse) + testVersionDecodable(t, "error", response, errorDeleteGroupsResponse, 0) + if response.ThrottleTime != 0 { + t.Error("Expected no violation") + } + if response.GroupErrorCodes["foo"] != ErrClusterAuthorizationFailed { + t.Error("Expected error ErrClusterAuthorizationFailed, found:", response.GroupErrorCodes["foo"]) + } + + response = new(DeleteGroupsResponse) + testVersionDecodable(t, "no error", response, noErrorDeleteGroupsResponse, 0) + if response.ThrottleTime != 0 { + t.Error("Expected no violation") + } + if response.GroupErrorCodes["foo"] != ErrNoError { + t.Error("Expected error ErrClusterAuthorizationFailed, found:", response.GroupErrorCodes["foo"]) + } +} diff --git a/vendor/github.com/Shopify/sarama/delete_records_request.go b/vendor/github.com/Shopify/sarama/delete_records_request.go new file mode 100644 index 000000000..93efafd4d --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_records_request.go @@ -0,0 +1,126 @@ +package sarama + +import ( + "sort" + "time" +) + +// request message format is: +// [topic] timeout(int32) +// where topic is: +// name(string) [partition] +// where partition is: +// id(int32) offset(int64) + +type DeleteRecordsRequest struct { + Topics map[string]*DeleteRecordsRequestTopic + Timeout time.Duration +} + +func (d *DeleteRecordsRequest) encode(pe packetEncoder) error { + if err := pe.putArrayLength(len(d.Topics)); err != nil { + return err + } + keys := make([]string, 0, len(d.Topics)) + for topic := range d.Topics { + keys = append(keys, topic) + } + sort.Strings(keys) + for _, topic := range keys { + if err := pe.putString(topic); err != nil { + return err + } + if err := d.Topics[topic].encode(pe); err != nil { + return err + } + } + pe.putInt32(int32(d.Timeout / time.Millisecond)) + + return nil +} + +func (d *DeleteRecordsRequest) decode(pd packetDecoder, version int16) error { + n, err := pd.getArrayLength() + if err != nil { + return err + } + + if n > 0 { + d.Topics = make(map[string]*DeleteRecordsRequestTopic, n) + for i := 0; i < n; i++ { + topic, err := pd.getString() + if err != nil { + return err + } + details := new(DeleteRecordsRequestTopic) + if err = details.decode(pd, version); err != nil { + return err + } + d.Topics[topic] = details + } + } + + timeout, err := pd.getInt32() + if err != nil { + return err + } + d.Timeout = time.Duration(timeout) * time.Millisecond + + return nil +} + +func (d *DeleteRecordsRequest) key() int16 { + return 21 +} + +func (d *DeleteRecordsRequest) version() int16 { + return 0 +} + +func (d *DeleteRecordsRequest) requiredVersion() KafkaVersion { + return V0_11_0_0 +} + +type DeleteRecordsRequestTopic struct { + PartitionOffsets map[int32]int64 // partition => offset +} + +func (t *DeleteRecordsRequestTopic) encode(pe packetEncoder) error { + if err := pe.putArrayLength(len(t.PartitionOffsets)); err != nil { + return err + } + keys := make([]int32, 0, len(t.PartitionOffsets)) + for partition := range t.PartitionOffsets { + keys = append(keys, partition) + } + sort.Slice(keys, func(i, j int) bool { return keys[i] < keys[j] }) + for _, partition := range keys { + pe.putInt32(partition) + pe.putInt64(t.PartitionOffsets[partition]) + } + return nil +} + +func (t *DeleteRecordsRequestTopic) decode(pd packetDecoder, version int16) error { + n, err := pd.getArrayLength() + if err != nil { + return err + } + + if n > 0 { + t.PartitionOffsets = make(map[int32]int64, n) + for i := 0; i < n; i++ { + partition, err := pd.getInt32() + if err != nil { + return err + } + offset, err := pd.getInt64() + if err != nil { + return err + } + t.PartitionOffsets[partition] = offset + } + } + + return nil +} diff --git a/vendor/github.com/Shopify/sarama/delete_records_request_test.go b/vendor/github.com/Shopify/sarama/delete_records_request_test.go new file mode 100644 index 000000000..c72960cfb --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_records_request_test.go @@ -0,0 +1,36 @@ +package sarama + +import ( + "testing" + "time" +) + +var deleteRecordsRequest = []byte{ + 0, 0, 0, 2, + 0, 5, 'o', 't', 'h', 'e', 'r', + 0, 0, 0, 0, + 0, 5, 't', 'o', 'p', 'i', 'c', + 0, 0, 0, 2, + 0, 0, 0, 19, + 0, 0, 0, 0, 0, 0, 0, 200, + 0, 0, 0, 20, + 0, 0, 0, 0, 0, 0, 0, 190, + 0, 0, 0, 100, +} + +func TestDeleteRecordsRequest(t *testing.T) { + req := &DeleteRecordsRequest{ + Topics: map[string]*DeleteRecordsRequestTopic{ + "topic": { + PartitionOffsets: map[int32]int64{ + 19: 200, + 20: 190, + }, + }, + "other": {}, + }, + Timeout: 100 * time.Millisecond, + } + + testRequest(t, "", req, deleteRecordsRequest) +} diff --git a/vendor/github.com/Shopify/sarama/delete_records_response.go b/vendor/github.com/Shopify/sarama/delete_records_response.go new file mode 100644 index 000000000..733a58b6b --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_records_response.go @@ -0,0 +1,158 @@ +package sarama + +import ( + "sort" + "time" +) + +// response message format is: +// throttleMs(int32) [topic] +// where topic is: +// name(string) [partition] +// where partition is: +// id(int32) low_watermark(int64) error_code(int16) + +type DeleteRecordsResponse struct { + Version int16 + ThrottleTime time.Duration + Topics map[string]*DeleteRecordsResponseTopic +} + +func (d *DeleteRecordsResponse) encode(pe packetEncoder) error { + pe.putInt32(int32(d.ThrottleTime / time.Millisecond)) + + if err := pe.putArrayLength(len(d.Topics)); err != nil { + return err + } + keys := make([]string, 0, len(d.Topics)) + for topic := range d.Topics { + keys = append(keys, topic) + } + sort.Strings(keys) + for _, topic := range keys { + if err := pe.putString(topic); err != nil { + return err + } + if err := d.Topics[topic].encode(pe); err != nil { + return err + } + } + return nil +} + +func (d *DeleteRecordsResponse) decode(pd packetDecoder, version int16) error { + d.Version = version + + throttleTime, err := pd.getInt32() + if err != nil { + return err + } + d.ThrottleTime = time.Duration(throttleTime) * time.Millisecond + + n, err := pd.getArrayLength() + if err != nil { + return err + } + + if n > 0 { + d.Topics = make(map[string]*DeleteRecordsResponseTopic, n) + for i := 0; i < n; i++ { + topic, err := pd.getString() + if err != nil { + return err + } + details := new(DeleteRecordsResponseTopic) + if err = details.decode(pd, version); err != nil { + return err + } + d.Topics[topic] = details + } + } + + return nil +} + +func (d *DeleteRecordsResponse) key() int16 { + return 21 +} + +func (d *DeleteRecordsResponse) version() int16 { + return 0 +} + +func (d *DeleteRecordsResponse) requiredVersion() KafkaVersion { + return V0_11_0_0 +} + +type DeleteRecordsResponseTopic struct { + Partitions map[int32]*DeleteRecordsResponsePartition +} + +func (t *DeleteRecordsResponseTopic) encode(pe packetEncoder) error { + if err := pe.putArrayLength(len(t.Partitions)); err != nil { + return err + } + keys := make([]int32, 0, len(t.Partitions)) + for partition := range t.Partitions { + keys = append(keys, partition) + } + sort.Slice(keys, func(i, j int) bool { return keys[i] < keys[j] }) + for _, partition := range keys { + pe.putInt32(partition) + if err := t.Partitions[partition].encode(pe); err != nil { + return err + } + } + return nil +} + +func (t *DeleteRecordsResponseTopic) decode(pd packetDecoder, version int16) error { + n, err := pd.getArrayLength() + if err != nil { + return err + } + + if n > 0 { + t.Partitions = make(map[int32]*DeleteRecordsResponsePartition, n) + for i := 0; i < n; i++ { + partition, err := pd.getInt32() + if err != nil { + return err + } + details := new(DeleteRecordsResponsePartition) + if err = details.decode(pd, version); err != nil { + return err + } + t.Partitions[partition] = details + } + } + + return nil +} + +type DeleteRecordsResponsePartition struct { + LowWatermark int64 + Err KError +} + +func (t *DeleteRecordsResponsePartition) encode(pe packetEncoder) error { + pe.putInt64(t.LowWatermark) + pe.putInt16(int16(t.Err)) + return nil +} + +func (t *DeleteRecordsResponsePartition) decode(pd packetDecoder, version int16) error { + lowWatermark, err := pd.getInt64() + if err != nil { + return err + } + t.LowWatermark = lowWatermark + + kErr, err := pd.getInt16() + if err != nil { + return err + } + t.Err = KError(kErr) + + return nil +} diff --git a/vendor/github.com/Shopify/sarama/delete_records_response_test.go b/vendor/github.com/Shopify/sarama/delete_records_response_test.go new file mode 100644 index 000000000..3653cdc41 --- /dev/null +++ b/vendor/github.com/Shopify/sarama/delete_records_response_test.go @@ -0,0 +1,39 @@ +package sarama + +import ( + "testing" + "time" +) + +var deleteRecordsResponse = []byte{ + 0, 0, 0, 100, + 0, 0, 0, 2, + 0, 5, 'o', 't', 'h', 'e', 'r', + 0, 0, 0, 0, + 0, 5, 't', 'o', 'p', 'i', 'c', + 0, 0, 0, 2, + 0, 0, 0, 19, + 0, 0, 0, 0, 0, 0, 0, 200, + 0, 0, + 0, 0, 0, 20, + 255, 255, 255, 255, 255, 255, 255, 255, + 0, 3, +} + +func TestDeleteRecordsResponse(t *testing.T) { + resp := &DeleteRecordsResponse{ + Version: 0, + ThrottleTime: 100 * time.Millisecond, + Topics: map[string]*DeleteRecordsResponseTopic{ + "topic": { + Partitions: map[int32]*DeleteRecordsResponsePartition{ + 19: {LowWatermark: 200, Err: 0}, + 20: {LowWatermark: -1, Err: 3}, + }, + }, + "other": {}, + }, + } + + testResponse(t, "", resp, deleteRecordsResponse) +} diff --git a/vendor/github.com/Shopify/sarama/delete_topics_request.go b/vendor/github.com/Shopify/sarama/delete_topics_request.go index ed9089ea4..911f67d31 100644 --- a/vendor/github.com/Shopify/sarama/delete_topics_request.go +++ b/vendor/github.com/Shopify/sarama/delete_topics_request.go @@ -3,6 +3,7 @@ package sarama import "time" type DeleteTopicsRequest struct { + Version int16 Topics []string Timeout time.Duration } @@ -25,6 +26,7 @@ func (d *DeleteTopicsRequest) decode(pd packetDecoder, version int16) (err error return err } d.Timeout = time.Duration(timeout) * time.Millisecond + d.Version = version return nil } @@ -33,9 +35,14 @@ func (d *DeleteTopicsRequest) key() int16 { } func (d *DeleteTopicsRequest) version() int16 { - return 0 + return d.Version } func (d *DeleteTopicsRequest) requiredVersion() KafkaVersion { - return V0_10_1_0 + switch d.Version { + case 1: + return V0_11_0_0 + default: + return V0_10_1_0 + } } diff --git a/vendor/github.com/Shopify/sarama/delete_topics_request_test.go b/vendor/github.com/Shopify/sarama/delete_topics_request_test.go index ce940823c..c313a2f3b 100644 --- a/vendor/github.com/Shopify/sarama/delete_topics_request_test.go +++ b/vendor/github.com/Shopify/sarama/delete_topics_request_test.go @@ -12,8 +12,19 @@ var deleteTopicsRequest = []byte{ 0, 0, 0, 100, } -func TestDeleteTopicsRequest(t *testing.T) { +func TestDeleteTopicsRequestV0(t *testing.T) { req := &DeleteTopicsRequest{ + Version: 0, + Topics: []string{"topic", "other"}, + Timeout: 100 * time.Millisecond, + } + + testRequest(t, "", req, deleteTopicsRequest) +} + +func TestDeleteTopicsRequestV1(t *testing.T) { + req := &DeleteTopicsRequest{ + Version: 1, Topics: []string{"topic", "other"}, Timeout: 100 * time.Millisecond, } diff --git a/vendor/github.com/Shopify/sarama/errors.go b/vendor/github.com/Shopify/sarama/errors.go index 54f431a4a..c578ef5fb 100644 --- a/vendor/github.com/Shopify/sarama/errors.go +++ b/vendor/github.com/Shopify/sarama/errors.go @@ -41,6 +41,14 @@ var ErrMessageTooLarge = errors.New("kafka: message is larger than Consumer.Fetc // a RecordBatch. var ErrConsumerOffsetNotAdvanced = errors.New("kafka: consumer offset was not advanced after a RecordBatch") +// ErrControllerNotAvailable is returned when server didn't give correct controller id. May be kafka server's version +// is lower than 0.10.0.0. +var ErrControllerNotAvailable = errors.New("kafka: controller is not available") + +// ErrNoTopicsToUpdateMetadata is returned when Meta.Full is set to false but no specific topics were found to update +// the metadata. +var ErrNoTopicsToUpdateMetadata = errors.New("kafka: no specific topics to update metadata") + // PacketEncodingError is returned from a failure while encoding a Kafka packet. This can happen, for example, // if you try to encode a string over 2^15 characters in length, since Kafka's encoding rules do not permit that. type PacketEncodingError struct { diff --git a/vendor/github.com/Shopify/sarama/fetch_request.go b/vendor/github.com/Shopify/sarama/fetch_request.go index 8c8e3a5af..462ab8afb 100644 --- a/vendor/github.com/Shopify/sarama/fetch_request.go +++ b/vendor/github.com/Shopify/sarama/fetch_request.go @@ -149,7 +149,7 @@ func (r *FetchRequest) requiredVersion() KafkaVersion { case 4: return V0_11_0_0 default: - return minVersion + return MinVersion } } diff --git a/vendor/github.com/Shopify/sarama/fetch_response.go b/vendor/github.com/Shopify/sarama/fetch_response.go index 0e81ad89f..ae91bb9eb 100644 --- a/vendor/github.com/Shopify/sarama/fetch_response.go +++ b/vendor/github.com/Shopify/sarama/fetch_response.go @@ -280,7 +280,7 @@ func (r *FetchResponse) requiredVersion() KafkaVersion { case 4: return V0_11_0_0 default: - return minVersion + return MinVersion } } @@ -353,7 +353,7 @@ func (r *FetchResponse) AddMessage(topic string, partition int32, key, value Enc records := newLegacyRecords(&MessageSet{}) frb.RecordsSet = []*Records{&records} } - set := frb.RecordsSet[0].msgSet + set := frb.RecordsSet[0].MsgSet set.Messages = append(set.Messages, msgBlock) } @@ -365,7 +365,7 @@ func (r *FetchResponse) AddRecord(topic string, partition int32, key, value Enco records := newDefaultRecords(&RecordBatch{Version: 2}) frb.RecordsSet = []*Records{&records} } - batch := frb.RecordsSet[0].recordBatch + batch := frb.RecordsSet[0].RecordBatch batch.addRecord(rec) } @@ -375,7 +375,7 @@ func (r *FetchResponse) SetLastOffsetDelta(topic string, partition int32, offset records := newDefaultRecords(&RecordBatch{Version: 2}) frb.RecordsSet = []*Records{&records} } - batch := frb.RecordsSet[0].recordBatch + batch := frb.RecordsSet[0].RecordBatch batch.LastOffsetDelta = offset } diff --git a/vendor/github.com/Shopify/sarama/fetch_response_test.go b/vendor/github.com/Shopify/sarama/fetch_response_test.go index b82a69f3e..c6b6b46e4 100644 --- a/vendor/github.com/Shopify/sarama/fetch_response_test.go +++ b/vendor/github.com/Shopify/sarama/fetch_response_test.go @@ -132,7 +132,7 @@ func TestOneMessageFetchResponse(t *testing.T) { if n != 1 { t.Fatal("Decoding produced incorrect number of messages.") } - msgBlock := block.RecordsSet[0].msgSet.Messages[0] + msgBlock := block.RecordsSet[0].MsgSet.Messages[0] if msgBlock.Offset != 0x550000 { t.Error("Decoding produced incorrect message offset.") } @@ -185,7 +185,7 @@ func TestOneRecordFetchResponse(t *testing.T) { if n != 1 { t.Fatal("Decoding produced incorrect number of records.") } - rec := block.RecordsSet[0].recordBatch.Records[0] + rec := block.RecordsSet[0].RecordBatch.Records[0] if !bytes.Equal(rec.Key, []byte{0x01, 0x02, 0x03, 0x04}) { t.Error("Decoding produced incorrect record key.") } @@ -231,7 +231,7 @@ func TestOneMessageFetchResponseV4(t *testing.T) { if n != 1 { t.Fatal("Decoding produced incorrect number of records.") } - msgBlock := block.RecordsSet[0].msgSet.Messages[0] + msgBlock := block.RecordsSet[0].MsgSet.Messages[0] if msgBlock.Offset != 0x550000 { t.Error("Decoding produced incorrect message offset.") } diff --git a/vendor/github.com/Shopify/sarama/find_coordinator_request.go b/vendor/github.com/Shopify/sarama/find_coordinator_request.go new file mode 100644 index 000000000..0ab5cb5ff --- /dev/null +++ b/vendor/github.com/Shopify/sarama/find_coordinator_request.go @@ -0,0 +1,61 @@ +package sarama + +type CoordinatorType int8 + +const ( + CoordinatorGroup CoordinatorType = 0 + CoordinatorTransaction CoordinatorType = 1 +) + +type FindCoordinatorRequest struct { + Version int16 + CoordinatorKey string + CoordinatorType CoordinatorType +} + +func (f *FindCoordinatorRequest) encode(pe packetEncoder) error { + if err := pe.putString(f.CoordinatorKey); err != nil { + return err + } + + if f.Version >= 1 { + pe.putInt8(int8(f.CoordinatorType)) + } + + return nil +} + +func (f *FindCoordinatorRequest) decode(pd packetDecoder, version int16) (err error) { + if f.CoordinatorKey, err = pd.getString(); err != nil { + return err + } + + if version >= 1 { + f.Version = version + coordinatorType, err := pd.getInt8() + if err != nil { + return err + } + + f.CoordinatorType = CoordinatorType(coordinatorType) + } + + return nil +} + +func (f *FindCoordinatorRequest) key() int16 { + return 10 +} + +func (f *FindCoordinatorRequest) version() int16 { + return f.Version +} + +func (f *FindCoordinatorRequest) requiredVersion() KafkaVersion { + switch f.Version { + case 1: + return V0_11_0_0 + default: + return V0_8_2_0 + } +} diff --git a/vendor/github.com/Shopify/sarama/find_coordinator_request_test.go b/vendor/github.com/Shopify/sarama/find_coordinator_request_test.go new file mode 100644 index 000000000..7e889b074 --- /dev/null +++ b/vendor/github.com/Shopify/sarama/find_coordinator_request_test.go @@ -0,0 +1,33 @@ +package sarama + +import "testing" + +var ( + findCoordinatorRequestConsumerGroup = []byte{ + 0, 5, 'g', 'r', 'o', 'u', 'p', + 0, + } + + findCoordinatorRequestTransaction = []byte{ + 0, 13, 't', 'r', 'a', 'n', 's', 'a', 'c', 't', 'i', 'o', 'n', 'i', 'd', + 1, + } +) + +func TestFindCoordinatorRequest(t *testing.T) { + req := &FindCoordinatorRequest{ + Version: 1, + CoordinatorKey: "group", + CoordinatorType: CoordinatorGroup, + } + + testRequest(t, "version 1 - group", req, findCoordinatorRequestConsumerGroup) + + req = &FindCoordinatorRequest{ + Version: 1, + CoordinatorKey: "transactionid", + CoordinatorType: CoordinatorTransaction, + } + + testRequest(t, "version 1 - transaction", req, findCoordinatorRequestTransaction) +} diff --git a/vendor/github.com/Shopify/sarama/find_coordinator_response.go b/vendor/github.com/Shopify/sarama/find_coordinator_response.go new file mode 100644 index 000000000..9c900e8b7 --- /dev/null +++ b/vendor/github.com/Shopify/sarama/find_coordinator_response.go @@ -0,0 +1,92 @@ +package sarama + +import ( + "time" +) + +var NoNode = &Broker{id: -1, addr: ":-1"} + +type FindCoordinatorResponse struct { + Version int16 + ThrottleTime time.Duration + Err KError + ErrMsg *string + Coordinator *Broker +} + +func (f *FindCoordinatorResponse) decode(pd packetDecoder, version int16) (err error) { + if version >= 1 { + f.Version = version + + throttleTime, err := pd.getInt32() + if err != nil { + return err + } + f.ThrottleTime = time.Duration(throttleTime) * time.Millisecond + } + + tmp, err := pd.getInt16() + if err != nil { + return err + } + f.Err = KError(tmp) + + if version >= 1 { + if f.ErrMsg, err = pd.getNullableString(); err != nil { + return err + } + } + + coordinator := new(Broker) + // The version is hardcoded to 0, as version 1 of the Broker-decode + // contains the rack-field which is not present in the FindCoordinatorResponse. + if err := coordinator.decode(pd, 0); err != nil { + return err + } + if coordinator.addr == ":0" { + return nil + } + f.Coordinator = coordinator + + return nil +} + +func (f *FindCoordinatorResponse) encode(pe packetEncoder) error { + if f.Version >= 1 { + pe.putInt32(int32(f.ThrottleTime / time.Millisecond)) + } + + pe.putInt16(int16(f.Err)) + + if f.Version >= 1 { + if err := pe.putNullableString(f.ErrMsg); err != nil { + return err + } + } + + coordinator := f.Coordinator + if coordinator == nil { + coordinator = NoNode + } + if err := coordinator.encode(pe, 0); err != nil { + return err + } + return nil +} + +func (f *FindCoordinatorResponse) key() int16 { + return 10 +} + +func (f *FindCoordinatorResponse) version() int16 { + return f.Version +} + +func (f *FindCoordinatorResponse) requiredVersion() KafkaVersion { + switch f.Version { + case 1: + return V0_11_0_0 + default: + return V0_8_2_0 + } +} diff --git a/vendor/github.com/Shopify/sarama/find_coordinator_response_test.go b/vendor/github.com/Shopify/sarama/find_coordinator_response_test.go new file mode 100644 index 000000000..417a76c6e --- /dev/null +++ b/vendor/github.com/Shopify/sarama/find_coordinator_response_test.go @@ -0,0 +1,83 @@ +package sarama + +import ( + "testing" + "time" +) + +func TestFindCoordinatorResponse(t *testing.T) { + errMsg := "kaboom" + + for _, tc := range []struct { + desc string + response *FindCoordinatorResponse + encoded []byte + }{{ + desc: "version 0 - no error", + response: &FindCoordinatorResponse{ + Version: 0, + Err: ErrNoError, + Coordinator: &Broker{ + id: 7, + addr: "host:9092", + }, + }, + encoded: []byte{ + 0, 0, // Err + 0, 0, 0, 7, // Coordinator.ID + 0, 4, 'h', 'o', 's', 't', // Coordinator.Host + 0, 0, 35, 132, // Coordinator.Port + }, + }, { + desc: "version 1 - no error", + response: &FindCoordinatorResponse{ + Version: 1, + ThrottleTime: 100 * time.Millisecond, + Err: ErrNoError, + Coordinator: &Broker{ + id: 7, + addr: "host:9092", + }, + }, + encoded: []byte{ + 0, 0, 0, 100, // ThrottleTime + 0, 0, // Err + 255, 255, // ErrMsg: empty + 0, 0, 0, 7, // Coordinator.ID + 0, 4, 'h', 'o', 's', 't', // Coordinator.Host + 0, 0, 35, 132, // Coordinator.Port + }, + }, { + desc: "version 0 - error", + response: &FindCoordinatorResponse{ + Version: 0, + Err: ErrConsumerCoordinatorNotAvailable, + Coordinator: NoNode, + }, + encoded: []byte{ + 0, 15, // Err + 255, 255, 255, 255, // Coordinator.ID: -1 + 0, 0, // Coordinator.Host: "" + 255, 255, 255, 255, // Coordinator.Port: -1 + }, + }, { + desc: "version 1 - error", + response: &FindCoordinatorResponse{ + Version: 1, + ThrottleTime: 100 * time.Millisecond, + Err: ErrConsumerCoordinatorNotAvailable, + ErrMsg: &errMsg, + Coordinator: NoNode, + }, + encoded: []byte{ + 0, 0, 0, 100, // ThrottleTime + 0, 15, // Err + 0, 6, 'k', 'a', 'b', 'o', 'o', 'm', // ErrMsg + 255, 255, 255, 255, // Coordinator.ID: -1 + 0, 0, // Coordinator.Host: "" + 255, 255, 255, 255, // Coordinator.Port: -1 + }, + }} { + testResponse(t, tc.desc, tc.response, tc.encoded) + } +} diff --git a/vendor/github.com/Shopify/sarama/functional_consumer_test.go b/vendor/github.com/Shopify/sarama/functional_consumer_test.go index ab8433109..83bec0331 100644 --- a/vendor/github.com/Shopify/sarama/functional_consumer_test.go +++ b/vendor/github.com/Shopify/sarama/functional_consumer_test.go @@ -1,8 +1,13 @@ package sarama import ( + "fmt" "math" + "os" + "sort" + "sync" "testing" + "time" ) func TestFuncConsumerOffsetOutOfRange(t *testing.T) { @@ -46,7 +51,7 @@ func TestConsumerHighWaterMarkOffset(t *testing.T) { } defer safeClose(t, c) - pc, err := c.ConsumePartition("test.1", 0, OffsetOldest) + pc, err := c.ConsumePartition("test.1", 0, offset) if err != nil { t.Fatal(err) } @@ -59,3 +64,163 @@ func TestConsumerHighWaterMarkOffset(t *testing.T) { safeClose(t, pc) } + +// Makes sure that messages produced by all supported client versions/ +// compression codecs (except LZ4) combinations can be consumed by all +// supported consumer versions. It relies on the KAFKA_VERSION environment +// variable to provide the version of the test Kafka cluster. +// +// Note that LZ4 codec was introduced in v0.10.0.0 and therefore is excluded +// from this test case. It has a similar version matrix test case below that +// only checks versions from v0.10.0.0 until KAFKA_VERSION. +func TestVersionMatrix(t *testing.T) { + setupFunctionalTest(t) + defer teardownFunctionalTest(t) + + // Produce lot's of message with all possible combinations of supported + // protocol versions and compressions for the except of LZ4. + testVersions := versionRange(V0_8_2_0) + allCodecsButLZ4 := []CompressionCodec{CompressionNone, CompressionGZIP, CompressionSnappy} + producedMessages := produceMsgs(t, testVersions, allCodecsButLZ4, 17, 100) + + // When/Then + consumeMsgs(t, testVersions, producedMessages) +} + +// Support for LZ4 codec was introduced in v0.10.0.0 so a version matrix to +// test LZ4 should start with v0.10.0.0. +func TestVersionMatrixLZ4(t *testing.T) { + setupFunctionalTest(t) + defer teardownFunctionalTest(t) + + // Produce lot's of message with all possible combinations of supported + // protocol versions starting with v0.10 (first where LZ4 was supported) + // and all possible compressions. + testVersions := versionRange(V0_10_0_0) + allCodecs := []CompressionCodec{CompressionNone, CompressionGZIP, CompressionSnappy, CompressionLZ4} + producedMessages := produceMsgs(t, testVersions, allCodecs, 17, 100) + + // When/Then + consumeMsgs(t, testVersions, producedMessages) +} + +func prodMsg2Str(prodMsg *ProducerMessage) string { + return fmt.Sprintf("{offset: %d, value: %s}", prodMsg.Offset, string(prodMsg.Value.(StringEncoder))) +} + +func consMsg2Str(consMsg *ConsumerMessage) string { + return fmt.Sprintf("{offset: %d, value: %s}", consMsg.Offset, string(consMsg.Value)) +} + +func versionRange(lower KafkaVersion) []KafkaVersion { + // Get the test cluster version from the environment. If there is nothing + // there then assume the highest. + upper, err := ParseKafkaVersion(os.Getenv("KAFKA_VERSION")) + if err != nil { + upper = MaxVersion + } + + versions := make([]KafkaVersion, 0, len(SupportedVersions)) + for _, v := range SupportedVersions { + if !v.IsAtLeast(lower) { + continue + } + if !upper.IsAtLeast(v) { + return versions + } + versions = append(versions, v) + } + return versions +} + +func produceMsgs(t *testing.T, clientVersions []KafkaVersion, codecs []CompressionCodec, flush int, countPerVerCodec int) []*ProducerMessage { + var wg sync.WaitGroup + var producedMessagesMu sync.Mutex + var producedMessages []*ProducerMessage + for _, prodVer := range clientVersions { + for _, codec := range codecs { + prodCfg := NewConfig() + prodCfg.Version = prodVer + prodCfg.Producer.Return.Successes = true + prodCfg.Producer.Return.Errors = true + prodCfg.Producer.Flush.MaxMessages = flush + prodCfg.Producer.Compression = codec + + p, err := NewSyncProducer(kafkaBrokers, prodCfg) + if err != nil { + t.Errorf("Failed to create producer: version=%s, compression=%s, err=%v", prodVer, codec, err) + continue + } + defer safeClose(t, p) + for i := 0; i < countPerVerCodec; i++ { + msg := &ProducerMessage{ + Topic: "test.1", + Value: StringEncoder(fmt.Sprintf("msg:%s:%s:%d", prodVer, codec, i)), + } + wg.Add(1) + go func() { + defer wg.Done() + _, _, err := p.SendMessage(msg) + if err != nil { + t.Errorf("Failed to produce message: %s, err=%v", msg.Value, err) + } + producedMessagesMu.Lock() + producedMessages = append(producedMessages, msg) + producedMessagesMu.Unlock() + }() + } + } + } + wg.Wait() + + // Sort produced message in ascending offset order. + sort.Slice(producedMessages, func(i, j int) bool { + return producedMessages[i].Offset < producedMessages[j].Offset + }) + t.Logf("*** Total produced %d, firstOffset=%d, lastOffset=%d\n", + len(producedMessages), producedMessages[0].Offset, producedMessages[len(producedMessages)-1].Offset) + return producedMessages +} + +func consumeMsgs(t *testing.T, clientVersions []KafkaVersion, producedMessages []*ProducerMessage) { + // Consume all produced messages with all client versions supported by the + // cluster. +consumerVersionLoop: + for _, consVer := range clientVersions { + t.Logf("*** Consuming with client version %s\n", consVer) + // Create a partition consumer that should start from the first produced + // message. + consCfg := NewConfig() + consCfg.Version = consVer + c, err := NewConsumer(kafkaBrokers, consCfg) + if err != nil { + t.Fatal(err) + } + defer safeClose(t, c) + pc, err := c.ConsumePartition("test.1", 0, producedMessages[0].Offset) + if err != nil { + t.Fatal(err) + } + defer safeClose(t, pc) + + // Consume as many messages as there have been produced and make sure that + // order is preserved. + for i, prodMsg := range producedMessages { + select { + case consMsg := <-pc.Messages(): + if consMsg.Offset != prodMsg.Offset { + t.Errorf("Consumed unexpected offset: version=%s, index=%d, want=%s, got=%s", + consVer, i, prodMsg2Str(prodMsg), consMsg2Str(consMsg)) + continue consumerVersionLoop + } + if string(consMsg.Value) != string(prodMsg.Value.(StringEncoder)) { + t.Errorf("Consumed unexpected msg: version=%s, index=%d, want=%s, got=%s", + consVer, i, prodMsg2Str(prodMsg), consMsg2Str(consMsg)) + continue consumerVersionLoop + } + case <-time.After(3 * time.Second): + t.Fatalf("Timeout waiting for: index=%d, offset=%d, msg=%s", i, prodMsg.Offset, prodMsg.Value) + } + } + } +} diff --git a/vendor/github.com/Shopify/sarama/join_group_request.go b/vendor/github.com/Shopify/sarama/join_group_request.go index 3a7ba1712..97e9299ea 100644 --- a/vendor/github.com/Shopify/sarama/join_group_request.go +++ b/vendor/github.com/Shopify/sarama/join_group_request.go @@ -25,8 +25,10 @@ func (p *GroupProtocol) encode(pe packetEncoder) (err error) { } type JoinGroupRequest struct { + Version int16 GroupId string SessionTimeout int32 + RebalanceTimeout int32 MemberId string ProtocolType string GroupProtocols map[string][]byte // deprecated; use OrderedGroupProtocols @@ -38,6 +40,9 @@ func (r *JoinGroupRequest) encode(pe packetEncoder) error { return err } pe.putInt32(r.SessionTimeout) + if r.Version >= 1 { + pe.putInt32(r.RebalanceTimeout) + } if err := pe.putString(r.MemberId); err != nil { return err } @@ -76,6 +81,8 @@ func (r *JoinGroupRequest) encode(pe packetEncoder) error { } func (r *JoinGroupRequest) decode(pd packetDecoder, version int16) (err error) { + r.Version = version + if r.GroupId, err = pd.getString(); err != nil { return } @@ -84,6 +91,12 @@ func (r *JoinGroupRequest) decode(pd packetDecoder, version int16) (err error) { return } + if version >= 1 { + if r.RebalanceTimeout, err = pd.getInt32(); err != nil { + return err + } + } + if r.MemberId, err = pd.getString(); err != nil { return } @@ -118,11 +131,18 @@ func (r *JoinGroupRequest) key() int16 { } func (r *JoinGroupRequest) version() int16 { - return 0 + return r.Version } func (r *JoinGroupRequest) requiredVersion() KafkaVersion { - return V0_9_0_0 + switch r.Version { + case 2: + return V0_11_0_0 + case 1: + return V0_10_1_0 + default: + return V0_9_0_0 + } } func (r *JoinGroupRequest) AddGroupProtocol(name string, metadata []byte) { diff --git a/vendor/github.com/Shopify/sarama/join_group_request_test.go b/vendor/github.com/Shopify/sarama/join_group_request_test.go index 1ba3308bb..a2e17f980 100644 --- a/vendor/github.com/Shopify/sarama/join_group_request_test.go +++ b/vendor/github.com/Shopify/sarama/join_group_request_test.go @@ -3,7 +3,7 @@ package sarama import "testing" var ( - joinGroupRequestNoProtocols = []byte{ + joinGroupRequestV0_NoProtocols = []byte{ 0, 9, 'T', 'e', 's', 't', 'G', 'r', 'o', 'u', 'p', // Group ID 0, 0, 0, 100, // Session timeout 0, 0, // Member ID @@ -11,7 +11,7 @@ var ( 0, 0, 0, 0, // 0 protocol groups } - joinGroupRequestOneProtocol = []byte{ + joinGroupRequestV0_OneProtocol = []byte{ 0, 9, 'T', 'e', 's', 't', 'G', 'r', 'o', 'u', 'p', // Group ID 0, 0, 0, 100, // Session timeout 0, 11, 'O', 'n', 'e', 'P', 'r', 'o', 't', 'o', 'c', 'o', 'l', // Member ID @@ -20,6 +20,17 @@ var ( 0, 3, 'o', 'n', 'e', // Protocol name 0, 0, 0, 3, 0x01, 0x02, 0x03, // protocol metadata } + + joinGroupRequestV1 = []byte{ + 0, 9, 'T', 'e', 's', 't', 'G', 'r', 'o', 'u', 'p', // Group ID + 0, 0, 0, 100, // Session timeout + 0, 0, 0, 200, // Rebalance timeout + 0, 11, 'O', 'n', 'e', 'P', 'r', 'o', 't', 'o', 'c', 'o', 'l', // Member ID + 0, 8, 'c', 'o', 'n', 's', 'u', 'm', 'e', 'r', // Protocol Type + 0, 0, 0, 1, // 1 group protocol + 0, 3, 'o', 'n', 'e', // Protocol name + 0, 0, 0, 3, 0x01, 0x02, 0x03, // protocol metadata + } ) func TestJoinGroupRequest(t *testing.T) { @@ -27,20 +38,20 @@ func TestJoinGroupRequest(t *testing.T) { request.GroupId = "TestGroup" request.SessionTimeout = 100 request.ProtocolType = "consumer" - testRequest(t, "no protocols", request, joinGroupRequestNoProtocols) + testRequest(t, "V0: no protocols", request, joinGroupRequestV0_NoProtocols) } -func TestJoinGroupRequestOneProtocol(t *testing.T) { +func TestJoinGroupRequestV0_OneProtocol(t *testing.T) { request := new(JoinGroupRequest) request.GroupId = "TestGroup" request.SessionTimeout = 100 request.MemberId = "OneProtocol" request.ProtocolType = "consumer" request.AddGroupProtocol("one", []byte{0x01, 0x02, 0x03}) - packet := testRequestEncode(t, "one protocol", request, joinGroupRequestOneProtocol) + packet := testRequestEncode(t, "V0: one protocol", request, joinGroupRequestV0_OneProtocol) request.GroupProtocols = make(map[string][]byte) request.GroupProtocols["one"] = []byte{0x01, 0x02, 0x03} - testRequestDecode(t, "one protocol", request, packet) + testRequestDecode(t, "V0: one protocol", request, packet) } func TestJoinGroupRequestDeprecatedEncode(t *testing.T) { @@ -51,7 +62,22 @@ func TestJoinGroupRequestDeprecatedEncode(t *testing.T) { request.ProtocolType = "consumer" request.GroupProtocols = make(map[string][]byte) request.GroupProtocols["one"] = []byte{0x01, 0x02, 0x03} - packet := testRequestEncode(t, "one protocol", request, joinGroupRequestOneProtocol) + packet := testRequestEncode(t, "V0: one protocol", request, joinGroupRequestV0_OneProtocol) request.AddGroupProtocol("one", []byte{0x01, 0x02, 0x03}) - testRequestDecode(t, "one protocol", request, packet) + testRequestDecode(t, "V0: one protocol", request, packet) +} + +func TestJoinGroupRequestV1(t *testing.T) { + request := new(JoinGroupRequest) + request.Version = 1 + request.GroupId = "TestGroup" + request.SessionTimeout = 100 + request.RebalanceTimeout = 200 + request.MemberId = "OneProtocol" + request.ProtocolType = "consumer" + request.AddGroupProtocol("one", []byte{0x01, 0x02, 0x03}) + packet := testRequestEncode(t, "V1", request, joinGroupRequestV1) + request.GroupProtocols = make(map[string][]byte) + request.GroupProtocols["one"] = []byte{0x01, 0x02, 0x03} + testRequestDecode(t, "V1", request, packet) } diff --git a/vendor/github.com/Shopify/sarama/join_group_response.go b/vendor/github.com/Shopify/sarama/join_group_response.go index 6d35fe364..5752acc8a 100644 --- a/vendor/github.com/Shopify/sarama/join_group_response.go +++ b/vendor/github.com/Shopify/sarama/join_group_response.go @@ -1,6 +1,8 @@ package sarama type JoinGroupResponse struct { + Version int16 + ThrottleTime int32 Err KError GenerationId int32 GroupProtocol string @@ -22,6 +24,9 @@ func (r *JoinGroupResponse) GetMembers() (map[string]ConsumerGroupMemberMetadata } func (r *JoinGroupResponse) encode(pe packetEncoder) error { + if r.Version >= 2 { + pe.putInt32(r.ThrottleTime) + } pe.putInt16(int16(r.Err)) pe.putInt32(r.GenerationId) @@ -53,6 +58,14 @@ func (r *JoinGroupResponse) encode(pe packetEncoder) error { } func (r *JoinGroupResponse) decode(pd packetDecoder, version int16) (err error) { + r.Version = version + + if version >= 2 { + if r.ThrottleTime, err = pd.getInt32(); err != nil { + return + } + } + kerr, err := pd.getInt16() if err != nil { return err @@ -107,9 +120,16 @@ func (r *JoinGroupResponse) key() int16 { } func (r *JoinGroupResponse) version() int16 { - return 0 + return r.Version } func (r *JoinGroupResponse) requiredVersion() KafkaVersion { - return V0_9_0_0 + switch r.Version { + case 2: + return V0_11_0_0 + case 1: + return V0_10_1_0 + default: + return V0_9_0_0 + } } diff --git a/vendor/github.com/Shopify/sarama/join_group_response_test.go b/vendor/github.com/Shopify/sarama/join_group_response_test.go index ba7f71f20..a43b37a95 100644 --- a/vendor/github.com/Shopify/sarama/join_group_response_test.go +++ b/vendor/github.com/Shopify/sarama/join_group_response_test.go @@ -6,7 +6,7 @@ import ( ) var ( - joinGroupResponseNoError = []byte{ + joinGroupResponseV0_NoError = []byte{ 0x00, 0x00, // No error 0x00, 0x01, 0x02, 0x03, // Generation ID 0, 8, 'p', 'r', 'o', 't', 'o', 'c', 'o', 'l', // Protocol name chosen @@ -15,7 +15,7 @@ var ( 0, 0, 0, 0, // No member info } - joinGroupResponseWithError = []byte{ + joinGroupResponseV0_WithError = []byte{ 0, 23, // Error: inconsistent group protocol 0x00, 0x00, 0x00, 0x00, // Generation ID 0, 0, // Protocol name chosen @@ -24,7 +24,7 @@ var ( 0, 0, 0, 0, // No member info } - joinGroupResponseLeader = []byte{ + joinGroupResponseV0_Leader = []byte{ 0x00, 0x00, // No error 0x00, 0x01, 0x02, 0x03, // Generation ID 0, 8, 'p', 'r', 'o', 't', 'o', 'c', 'o', 'l', // Protocol name chosen @@ -34,13 +34,32 @@ var ( 0, 3, 'f', 'o', 'o', // Member ID 0, 0, 0, 3, 0x01, 0x02, 0x03, // Member metadata } + + joinGroupResponseV1 = []byte{ + 0x00, 0x00, // No error + 0x00, 0x01, 0x02, 0x03, // Generation ID + 0, 8, 'p', 'r', 'o', 't', 'o', 'c', 'o', 'l', // Protocol name chosen + 0, 3, 'f', 'o', 'o', // Leader ID + 0, 3, 'b', 'a', 'r', // Member ID + 0, 0, 0, 0, // No member info + } + + joinGroupResponseV2 = []byte{ + 0, 0, 0, 100, + 0x00, 0x00, // No error + 0x00, 0x01, 0x02, 0x03, // Generation ID + 0, 8, 'p', 'r', 'o', 't', 'o', 'c', 'o', 'l', // Protocol name chosen + 0, 3, 'f', 'o', 'o', // Leader ID + 0, 3, 'b', 'a', 'r', // Member ID + 0, 0, 0, 0, // No member info + } ) -func TestJoinGroupResponse(t *testing.T) { +func TestJoinGroupResponseV0(t *testing.T) { var response *JoinGroupResponse response = new(JoinGroupResponse) - testVersionDecodable(t, "no error", response, joinGroupResponseNoError, 0) + testVersionDecodable(t, "no error", response, joinGroupResponseV0_NoError, 0) if response.Err != ErrNoError { t.Error("Decoding Err failed: no error expected but found", response.Err) } @@ -58,7 +77,7 @@ func TestJoinGroupResponse(t *testing.T) { } response = new(JoinGroupResponse) - testVersionDecodable(t, "with error", response, joinGroupResponseWithError, 0) + testVersionDecodable(t, "with error", response, joinGroupResponseV0_WithError, 0) if response.Err != ErrInconsistentGroupProtocol { t.Error("Decoding Err failed: ErrInconsistentGroupProtocol expected but found", response.Err) } @@ -76,7 +95,7 @@ func TestJoinGroupResponse(t *testing.T) { } response = new(JoinGroupResponse) - testVersionDecodable(t, "with error", response, joinGroupResponseLeader, 0) + testVersionDecodable(t, "with error", response, joinGroupResponseV0_Leader, 0) if response.Err != ErrNoError { t.Error("Decoding Err failed: ErrNoError expected but found", response.Err) } @@ -96,3 +115,58 @@ func TestJoinGroupResponse(t *testing.T) { t.Error("Decoding foo member failed, found:", response.Members["foo"]) } } + +func TestJoinGroupResponseV1(t *testing.T) { + response := new(JoinGroupResponse) + testVersionDecodable(t, "no error", response, joinGroupResponseV1, 1) + if response.Err != ErrNoError { + t.Error("Decoding Err failed: no error expected but found", response.Err) + } + if response.GenerationId != 66051 { + t.Error("Decoding GenerationId failed, found:", response.GenerationId) + } + if response.GroupProtocol != "protocol" { + t.Error("Decoding GroupProtocol failed, found:", response.GroupProtocol) + } + if response.LeaderId != "foo" { + t.Error("Decoding LeaderId failed, found:", response.LeaderId) + } + if response.MemberId != "bar" { + t.Error("Decoding MemberId failed, found:", response.MemberId) + } + if response.Version != 1 { + t.Error("Decoding Version failed, found:", response.Version) + } + if len(response.Members) != 0 { + t.Error("Decoding Members failed, found:", response.Members) + } +} + +func TestJoinGroupResponseV2(t *testing.T) { + response := new(JoinGroupResponse) + testVersionDecodable(t, "no error", response, joinGroupResponseV2, 2) + if response.ThrottleTime != 100 { + t.Error("Decoding ThrottleTime failed, found:", response.ThrottleTime) + } + if response.Err != ErrNoError { + t.Error("Decoding Err failed: no error expected but found", response.Err) + } + if response.GenerationId != 66051 { + t.Error("Decoding GenerationId failed, found:", response.GenerationId) + } + if response.GroupProtocol != "protocol" { + t.Error("Decoding GroupProtocol failed, found:", response.GroupProtocol) + } + if response.LeaderId != "foo" { + t.Error("Decoding LeaderId failed, found:", response.LeaderId) + } + if response.MemberId != "bar" { + t.Error("Decoding MemberId failed, found:", response.MemberId) + } + if response.Version != 2 { + t.Error("Decoding Version failed, found:", response.Version) + } + if len(response.Members) != 0 { + t.Error("Decoding Members failed, found:", response.Members) + } +} diff --git a/vendor/github.com/Shopify/sarama/message.go b/vendor/github.com/Shopify/sarama/message.go index bd5650bbc..fecdbfdef 100644 --- a/vendor/github.com/Shopify/sarama/message.go +++ b/vendor/github.com/Shopify/sarama/message.go @@ -24,13 +24,28 @@ const ( CompressionLZ4 CompressionCodec = 3 ) +func (cc CompressionCodec) String() string { + return []string{ + "none", + "gzip", + "snappy", + "lz4", + }[int(cc)] +} + +// CompressionLevelDefault is the constant to use in CompressionLevel +// to have the default compression level for any codec. The value is picked +// that we don't use any existing compression levels. +const CompressionLevelDefault = -1000 + type Message struct { - Codec CompressionCodec // codec used to compress the message contents - Key []byte // the message key, may be nil - Value []byte // the message contents - Set *MessageSet // the message set a message might wrap - Version int8 // v1 requires Kafka 0.10 - Timestamp time.Time // the timestamp of the message (version 1+ only) + Codec CompressionCodec // codec used to compress the message contents + CompressionLevel int // compression level + Key []byte // the message key, may be nil + Value []byte // the message contents + Set *MessageSet // the message set a message might wrap + Version int8 // v1 requires Kafka 0.10 + Timestamp time.Time // the timestamp of the message (version 1+ only) compressedCache []byte compressedSize int // used for computing the compression ratio metrics @@ -66,7 +81,15 @@ func (m *Message) encode(pe packetEncoder) error { payload = m.Value case CompressionGZIP: var buf bytes.Buffer - writer := gzip.NewWriter(&buf) + var writer *gzip.Writer + if m.CompressionLevel != CompressionLevelDefault { + writer, err = gzip.NewWriterLevel(&buf, m.CompressionLevel) + if err != nil { + return err + } + } else { + writer = gzip.NewWriter(&buf) + } if _, err = writer.Write(m.Value); err != nil { return err } diff --git a/vendor/github.com/Shopify/sarama/message_test.go b/vendor/github.com/Shopify/sarama/message_test.go index 630ad6e2a..0eb02f263 100644 --- a/vendor/github.com/Shopify/sarama/message_test.go +++ b/vendor/github.com/Shopify/sarama/message_test.go @@ -1,8 +1,6 @@ package sarama import ( - "runtime" - "strings" "testing" "time" ) @@ -31,17 +29,6 @@ var ( 0xFF, 0xFF, 0xFF, 0xFF} // value emptyGzipMessage = []byte{ - 97, 79, 149, 90, //CRC - 0x00, // magic version byte - 0x01, // attribute flags - 0xFF, 0xFF, 0xFF, 0xFF, // key - // value - 0x00, 0x00, 0x00, 0x17, - 0x1f, 0x8b, - 0x08, - 0, 0, 9, 110, 136, 0, 255, 1, 0, 0, 255, 255, 0, 0, 0, 0, 0, 0, 0, 0} - - emptyGzipMessage18 = []byte{ 132, 99, 80, 148, //CRC 0x00, // magic version byte 0x01, // attribute flags @@ -107,11 +94,7 @@ func TestMessageEncoding(t *testing.T) { message.Value = []byte{} message.Codec = CompressionGZIP - if strings.HasPrefix(runtime.Version(), "go1.8") || strings.HasPrefix(runtime.Version(), "go1.9") { - testEncodable(t, "empty gzip", &message, emptyGzipMessage18) - } else { - testEncodable(t, "empty gzip", &message, emptyGzipMessage) - } + testEncodable(t, "empty gzip", &message, emptyGzipMessage) message.Value = []byte{} message.Codec = CompressionLZ4 diff --git a/vendor/github.com/Shopify/sarama/metadata_request.go b/vendor/github.com/Shopify/sarama/metadata_request.go index 9a26b55fd..48adfa28c 100644 --- a/vendor/github.com/Shopify/sarama/metadata_request.go +++ b/vendor/github.com/Shopify/sarama/metadata_request.go @@ -1,40 +1,65 @@ package sarama type MetadataRequest struct { - Topics []string + Version int16 + Topics []string + AllowAutoTopicCreation bool } func (r *MetadataRequest) encode(pe packetEncoder) error { - err := pe.putArrayLength(len(r.Topics)) - if err != nil { - return err + if r.Version < 0 || r.Version > 5 { + return PacketEncodingError{"invalid or unsupported MetadataRequest version field"} } - - for i := range r.Topics { - err = pe.putString(r.Topics[i]) + if r.Version == 0 || r.Topics != nil || len(r.Topics) > 0 { + err := pe.putArrayLength(len(r.Topics)) if err != nil { return err } + + for i := range r.Topics { + err = pe.putString(r.Topics[i]) + if err != nil { + return err + } + } + } else { + pe.putInt32(-1) + } + if r.Version > 3 { + pe.putBool(r.AllowAutoTopicCreation) } return nil } func (r *MetadataRequest) decode(pd packetDecoder, version int16) error { - topicCount, err := pd.getArrayLength() + r.Version = version + size, err := pd.getInt32() if err != nil { return err } - if topicCount == 0 { + if size < 0 { return nil - } + } else { + topicCount := size + if topicCount == 0 { + return nil + } - r.Topics = make([]string, topicCount) - for i := range r.Topics { - topic, err := pd.getString() + r.Topics = make([]string, topicCount) + for i := range r.Topics { + topic, err := pd.getString() + if err != nil { + return err + } + r.Topics[i] = topic + } + } + if r.Version > 3 { + autoCreation, err := pd.getBool() if err != nil { return err } - r.Topics[i] = topic + r.AllowAutoTopicCreation = autoCreation } return nil } @@ -44,9 +69,20 @@ func (r *MetadataRequest) key() int16 { } func (r *MetadataRequest) version() int16 { - return 0 + return r.Version } func (r *MetadataRequest) requiredVersion() KafkaVersion { - return minVersion + switch r.Version { + case 1: + return V0_10_0_0 + case 2: + return V0_10_1_0 + case 3, 4: + return V0_11_0_0 + case 5: + return V1_0_0_0 + default: + return MinVersion + } } diff --git a/vendor/github.com/Shopify/sarama/metadata_request_test.go b/vendor/github.com/Shopify/sarama/metadata_request_test.go index 44f3146e4..727e48a2c 100644 --- a/vendor/github.com/Shopify/sarama/metadata_request_test.go +++ b/vendor/github.com/Shopify/sarama/metadata_request_test.go @@ -3,27 +3,74 @@ package sarama import "testing" var ( - metadataRequestNoTopics = []byte{ + metadataRequestNoTopicsV0 = []byte{ 0x00, 0x00, 0x00, 0x00} - metadataRequestOneTopic = []byte{ + metadataRequestOneTopicV0 = []byte{ 0x00, 0x00, 0x00, 0x01, 0x00, 0x06, 't', 'o', 'p', 'i', 'c', '1'} - metadataRequestThreeTopics = []byte{ + metadataRequestThreeTopicsV0 = []byte{ 0x00, 0x00, 0x00, 0x03, 0x00, 0x03, 'f', 'o', 'o', 0x00, 0x03, 'b', 'a', 'r', 0x00, 0x03, 'b', 'a', 'z'} + + metadataRequestNoTopicsV1 = []byte{ + 0xff, 0xff, 0xff, 0xff} + + metadataRequestAutoCreateV4 = append(metadataRequestOneTopicV0, byte(1)) + metadataRequestNoAutoCreateV4 = append(metadataRequestOneTopicV0, byte(0)) ) -func TestMetadataRequest(t *testing.T) { +func TestMetadataRequestV0(t *testing.T) { request := new(MetadataRequest) - testRequest(t, "no topics", request, metadataRequestNoTopics) + testRequest(t, "no topics", request, metadataRequestNoTopicsV0) request.Topics = []string{"topic1"} - testRequest(t, "one topic", request, metadataRequestOneTopic) + testRequest(t, "one topic", request, metadataRequestOneTopicV0) request.Topics = []string{"foo", "bar", "baz"} - testRequest(t, "three topics", request, metadataRequestThreeTopics) + testRequest(t, "three topics", request, metadataRequestThreeTopicsV0) +} + +func TestMetadataRequestV1(t *testing.T) { + request := new(MetadataRequest) + request.Version = 1 + testRequest(t, "no topics", request, metadataRequestNoTopicsV1) + + request.Topics = []string{"topic1"} + testRequest(t, "one topic", request, metadataRequestOneTopicV0) + + request.Topics = []string{"foo", "bar", "baz"} + testRequest(t, "three topics", request, metadataRequestThreeTopicsV0) +} + +func TestMetadataRequestV2(t *testing.T) { + request := new(MetadataRequest) + request.Version = 2 + testRequest(t, "no topics", request, metadataRequestNoTopicsV1) + + request.Topics = []string{"topic1"} + testRequest(t, "one topic", request, metadataRequestOneTopicV0) +} + +func TestMetadataRequestV3(t *testing.T) { + request := new(MetadataRequest) + request.Version = 3 + testRequest(t, "no topics", request, metadataRequestNoTopicsV1) + + request.Topics = []string{"topic1"} + testRequest(t, "one topic", request, metadataRequestOneTopicV0) +} + +func TestMetadataRequestV4(t *testing.T) { + request := new(MetadataRequest) + request.Version = 4 + request.Topics = []string{"topic1"} + request.AllowAutoTopicCreation = true + testRequest(t, "one topic", request, metadataRequestAutoCreateV4) + + request.AllowAutoTopicCreation = false + testRequest(t, "one topic", request, metadataRequestNoAutoCreateV4) } diff --git a/vendor/github.com/Shopify/sarama/metadata_response.go b/vendor/github.com/Shopify/sarama/metadata_response.go index f9d6a4271..bf8a67bbc 100644 --- a/vendor/github.com/Shopify/sarama/metadata_response.go +++ b/vendor/github.com/Shopify/sarama/metadata_response.go @@ -1,14 +1,15 @@ package sarama type PartitionMetadata struct { - Err KError - ID int32 - Leader int32 - Replicas []int32 - Isr []int32 + Err KError + ID int32 + Leader int32 + Replicas []int32 + Isr []int32 + OfflineReplicas []int32 } -func (pm *PartitionMetadata) decode(pd packetDecoder) (err error) { +func (pm *PartitionMetadata) decode(pd packetDecoder, version int16) (err error) { tmp, err := pd.getInt16() if err != nil { return err @@ -35,10 +36,17 @@ func (pm *PartitionMetadata) decode(pd packetDecoder) (err error) { return err } + if version >= 5 { + pm.OfflineReplicas, err = pd.getInt32Array() + if err != nil { + return err + } + } + return nil } -func (pm *PartitionMetadata) encode(pe packetEncoder) (err error) { +func (pm *PartitionMetadata) encode(pe packetEncoder, version int16) (err error) { pe.putInt16(int16(pm.Err)) pe.putInt32(pm.ID) pe.putInt32(pm.Leader) @@ -53,16 +61,24 @@ func (pm *PartitionMetadata) encode(pe packetEncoder) (err error) { return err } + if version >= 5 { + err = pe.putInt32Array(pm.OfflineReplicas) + if err != nil { + return err + } + } + return nil } type TopicMetadata struct { Err KError Name string + IsInternal bool // Only valid for Version >= 1 Partitions []*PartitionMetadata } -func (tm *TopicMetadata) decode(pd packetDecoder) (err error) { +func (tm *TopicMetadata) decode(pd packetDecoder, version int16) (err error) { tmp, err := pd.getInt16() if err != nil { return err @@ -74,6 +90,13 @@ func (tm *TopicMetadata) decode(pd packetDecoder) (err error) { return err } + if version >= 1 { + tm.IsInternal, err = pd.getBool() + if err != nil { + return err + } + } + n, err := pd.getArrayLength() if err != nil { return err @@ -81,7 +104,7 @@ func (tm *TopicMetadata) decode(pd packetDecoder) (err error) { tm.Partitions = make([]*PartitionMetadata, n) for i := 0; i < n; i++ { tm.Partitions[i] = new(PartitionMetadata) - err = tm.Partitions[i].decode(pd) + err = tm.Partitions[i].decode(pd, version) if err != nil { return err } @@ -90,7 +113,7 @@ func (tm *TopicMetadata) decode(pd packetDecoder) (err error) { return nil } -func (tm *TopicMetadata) encode(pe packetEncoder) (err error) { +func (tm *TopicMetadata) encode(pe packetEncoder, version int16) (err error) { pe.putInt16(int16(tm.Err)) err = pe.putString(tm.Name) @@ -98,13 +121,17 @@ func (tm *TopicMetadata) encode(pe packetEncoder) (err error) { return err } + if version >= 1 { + pe.putBool(tm.IsInternal) + } + err = pe.putArrayLength(len(tm.Partitions)) if err != nil { return err } for _, pm := range tm.Partitions { - err = pm.encode(pe) + err = pm.encode(pe, version) if err != nil { return err } @@ -114,11 +141,24 @@ func (tm *TopicMetadata) encode(pe packetEncoder) (err error) { } type MetadataResponse struct { - Brokers []*Broker - Topics []*TopicMetadata + Version int16 + ThrottleTimeMs int32 + Brokers []*Broker + ClusterID *string + ControllerID int32 + Topics []*TopicMetadata } func (r *MetadataResponse) decode(pd packetDecoder, version int16) (err error) { + r.Version = version + + if version >= 3 { + r.ThrottleTimeMs, err = pd.getInt32() + if err != nil { + return err + } + } + n, err := pd.getArrayLength() if err != nil { return err @@ -127,12 +167,28 @@ func (r *MetadataResponse) decode(pd packetDecoder, version int16) (err error) { r.Brokers = make([]*Broker, n) for i := 0; i < n; i++ { r.Brokers[i] = new(Broker) - err = r.Brokers[i].decode(pd) + err = r.Brokers[i].decode(pd, version) if err != nil { return err } } + if version >= 2 { + r.ClusterID, err = pd.getNullableString() + if err != nil { + return err + } + } + + if version >= 1 { + r.ControllerID, err = pd.getInt32() + if err != nil { + return err + } + } else { + r.ControllerID = -1 + } + n, err = pd.getArrayLength() if err != nil { return err @@ -141,7 +197,7 @@ func (r *MetadataResponse) decode(pd packetDecoder, version int16) (err error) { r.Topics = make([]*TopicMetadata, n) for i := 0; i < n; i++ { r.Topics[i] = new(TopicMetadata) - err = r.Topics[i].decode(pd) + err = r.Topics[i].decode(pd, version) if err != nil { return err } @@ -156,18 +212,22 @@ func (r *MetadataResponse) encode(pe packetEncoder) error { return err } for _, broker := range r.Brokers { - err = broker.encode(pe) + err = broker.encode(pe, r.Version) if err != nil { return err } } + if r.Version >= 1 { + pe.putInt32(r.ControllerID) + } + err = pe.putArrayLength(len(r.Topics)) if err != nil { return err } for _, tm := range r.Topics { - err = tm.encode(pe) + err = tm.encode(pe, r.Version) if err != nil { return err } @@ -181,11 +241,22 @@ func (r *MetadataResponse) key() int16 { } func (r *MetadataResponse) version() int16 { - return 0 + return r.Version } func (r *MetadataResponse) requiredVersion() KafkaVersion { - return minVersion + switch r.Version { + case 1: + return V0_10_0_0 + case 2: + return V0_10_1_0 + case 3, 4: + return V0_11_0_0 + case 5: + return V1_0_0_0 + default: + return MinVersion + } } // testing API diff --git a/vendor/github.com/Shopify/sarama/metadata_response_test.go b/vendor/github.com/Shopify/sarama/metadata_response_test.go index ea62a4f1b..04a4ce7fc 100644 --- a/vendor/github.com/Shopify/sarama/metadata_response_test.go +++ b/vendor/github.com/Shopify/sarama/metadata_response_test.go @@ -3,11 +3,11 @@ package sarama import "testing" var ( - emptyMetadataResponse = []byte{ + emptyMetadataResponseV0 = []byte{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00} - brokersNoTopicsMetadataResponse = []byte{ + brokersNoTopicsMetadataResponseV0 = []byte{ 0x00, 0x00, 0x00, 0x02, 0x00, 0x00, 0xab, 0xff, @@ -20,7 +20,7 @@ var ( 0x00, 0x00, 0x00, 0x00} - topicsNoBrokersMetadataResponse = []byte{ + topicsNoBrokersMetadataResponseV0 = []byte{ 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, @@ -36,12 +36,76 @@ var ( 0x00, 0x00, 0x00, 0x03, 'b', 'a', 'r', 0x00, 0x00, 0x00, 0x00} + + brokersNoTopicsMetadataResponseV1 = []byte{ + 0x00, 0x00, 0x00, 0x02, + + 0x00, 0x00, 0xab, 0xff, + 0x00, 0x09, 'l', 'o', 'c', 'a', 'l', 'h', 'o', 's', 't', + 0x00, 0x00, 0x00, 0x33, + 0x00, 0x05, 'r', 'a', 'c', 'k', '0', + + 0x00, 0x01, 0x02, 0x03, + 0x00, 0x0a, 'g', 'o', 'o', 'g', 'l', 'e', '.', 'c', 'o', 'm', + 0x00, 0x00, 0x01, 0x11, + 0x00, 0x05, 'r', 'a', 'c', 'k', '1', + + 0x00, 0x00, 0x00, 0x01, + + 0x00, 0x00, 0x00, 0x00} + + topicsNoBrokersMetadataResponseV1 = []byte{ + 0x00, 0x00, 0x00, 0x00, + + 0x00, 0x00, 0x00, 0x04, + + 0x00, 0x00, 0x00, 0x02, + + 0x00, 0x00, + 0x00, 0x03, 'f', 'o', 'o', + 0x00, + 0x00, 0x00, 0x00, 0x01, + 0x00, 0x04, + 0x00, 0x00, 0x00, 0x01, + 0x00, 0x00, 0x00, 0x07, + 0x00, 0x00, 0x00, 0x03, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x02, 0x00, 0x00, 0x00, 0x03, + 0x00, 0x00, 0x00, 0x00, + + 0x00, 0x00, + 0x00, 0x03, 'b', 'a', 'r', + 0x01, + 0x00, 0x00, 0x00, 0x00} + + noBrokersNoTopicsWithThrottleTimeAndClusterIDV3 = []byte{ + 0x00, 0x00, 0x00, 0x10, + 0x00, 0x00, 0x00, 0x00, + 0x00, 0x09, 'c', 'l', 'u', 's', 't', 'e', 'r', 'I', 'd', + 0x00, 0x00, 0x00, 0x01, + 0x00, 0x00, 0x00, 0x00} + + noBrokersOneTopicWithOfflineReplicasV5 = []byte{ + 0x00, 0x00, 0x00, 0x05, + 0x00, 0x00, 0x00, 0x00, + 0x00, 0x09, 'c', 'l', 'u', 's', 't', 'e', 'r', 'I', 'd', + 0x00, 0x00, 0x00, 0x02, + 0x00, 0x00, 0x00, 0x01, + 0x00, 0x00, + 0x00, 0x03, 'f', 'o', 'o', + 0x00, + 0x00, 0x00, 0x00, 0x01, + 0x00, 0x04, + 0x00, 0x00, 0x00, 0x01, + 0x00, 0x00, 0x00, 0x07, + 0x00, 0x00, 0x00, 0x03, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x02, 0x00, 0x00, 0x00, 0x03, + 0x00, 0x00, 0x00, 0x02, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x02, + 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x03, + } ) -func TestEmptyMetadataResponse(t *testing.T) { +func TestEmptyMetadataResponseV0(t *testing.T) { response := MetadataResponse{} - testVersionDecodable(t, "empty", &response, emptyMetadataResponse, 0) + testVersionDecodable(t, "empty, V0", &response, emptyMetadataResponseV0, 0) if len(response.Brokers) != 0 { t.Error("Decoding produced", len(response.Brokers), "brokers where there were none!") } @@ -50,10 +114,10 @@ func TestEmptyMetadataResponse(t *testing.T) { } } -func TestMetadataResponseWithBrokers(t *testing.T) { +func TestMetadataResponseWithBrokersV0(t *testing.T) { response := MetadataResponse{} - testVersionDecodable(t, "brokers, no topics", &response, brokersNoTopicsMetadataResponse, 0) + testVersionDecodable(t, "brokers, no topics, V0", &response, brokersNoTopicsMetadataResponseV0, 0) if len(response.Brokers) != 2 { t.Fatal("Decoding produced", len(response.Brokers), "brokers where there were two!") } @@ -76,10 +140,10 @@ func TestMetadataResponseWithBrokers(t *testing.T) { } } -func TestMetadataResponseWithTopics(t *testing.T) { +func TestMetadataResponseWithTopicsV0(t *testing.T) { response := MetadataResponse{} - testVersionDecodable(t, "topics, no brokers", &response, topicsNoBrokersMetadataResponse, 0) + testVersionDecodable(t, "topics, no brokers, V0", &response, topicsNoBrokersMetadataResponseV0, 0) if len(response.Brokers) != 0 { t.Error("Decoding produced", len(response.Brokers), "brokers where there were none!") } @@ -137,3 +201,90 @@ func TestMetadataResponseWithTopics(t *testing.T) { t.Error("Decoding produced invalid partition count for topic 1.") } } + +func TestMetadataResponseWithBrokersV1(t *testing.T) { + response := MetadataResponse{} + + testVersionDecodable(t, "topics, V1", &response, brokersNoTopicsMetadataResponseV1, 1) + if len(response.Brokers) != 2 { + t.Error("Decoding produced", len(response.Brokers), "brokers where there were 2!") + } + if response.Brokers[0].rack == nil || *response.Brokers[0].rack != "rack0" { + t.Error("Decoding produced invalid broker 0 rack.") + } + if response.Brokers[1].rack == nil || *response.Brokers[1].rack != "rack1" { + t.Error("Decoding produced invalid broker 1 rack.") + } + if response.ControllerID != 1 { + t.Error("Decoding produced", response.ControllerID, "should have been 1!") + } + if len(response.Topics) != 0 { + t.Error("Decoding produced", len(response.Brokers), "brokers where there were none!") + } +} + +func TestMetadataResponseWithTopicsV1(t *testing.T) { + response := MetadataResponse{} + + testVersionDecodable(t, "topics, V1", &response, topicsNoBrokersMetadataResponseV1, 1) + if len(response.Brokers) != 0 { + t.Error("Decoding produced", len(response.Brokers), "brokers where there were none!") + } + if response.ControllerID != 4 { + t.Error("Decoding produced", response.ControllerID, "should have been 4!") + } + if len(response.Topics) != 2 { + t.Error("Decoding produced", len(response.Topics), "topics where there were 2!") + } + if response.Topics[0].IsInternal { + t.Error("Decoding produced", response.Topics[0], "topic0 should have been false!") + } + if !response.Topics[1].IsInternal { + t.Error("Decoding produced", response.Topics[1], "topic1 should have been true!") + } +} + +func TestMetadataResponseWithThrottleTime(t *testing.T) { + response := MetadataResponse{} + + testVersionDecodable(t, "no topics, no brokers, throttle time and cluster Id V3", &response, noBrokersNoTopicsWithThrottleTimeAndClusterIDV3, 3) + if response.ThrottleTimeMs != int32(16) { + t.Error("Decoding produced", response.ThrottleTimeMs, "should have been 16!") + } + if len(response.Brokers) != 0 { + t.Error("Decoding produced", response.Brokers, "should have been 0!") + } + if response.ControllerID != int32(1) { + t.Error("Decoding produced", response.ControllerID, "should have been 1!") + } + if *response.ClusterID != "clusterId" { + t.Error("Decoding produced", response.ClusterID, "should have been clusterId!") + } + if len(response.Topics) != 0 { + t.Error("Decoding produced", len(response.Topics), "should have been 0!") + } +} + +func TestMetadataResponseWithOfflineReplicasV5(t *testing.T) { + response := MetadataResponse{} + + testVersionDecodable(t, "no brokers, 1 topic with offline replica V5", &response, noBrokersOneTopicWithOfflineReplicasV5, 5) + if response.ThrottleTimeMs != int32(5) { + t.Error("Decoding produced", response.ThrottleTimeMs, "should have been 5!") + } + if len(response.Brokers) != 0 { + t.Error("Decoding produced", response.Brokers, "should have been 0!") + } + if response.ControllerID != int32(2) { + t.Error("Decoding produced", response.ControllerID, "should have been 21!") + } + if *response.ClusterID != "clusterId" { + t.Error("Decoding produced", response.ClusterID, "should have been clusterId!") + } + if len(response.Topics) != 1 { + t.Error("Decoding produced", len(response.Topics), "should have been 1!") + } + if len(response.Topics[0].Partitions[0].OfflineReplicas) != 1 { + t.Error("Decoding produced", len(response.Topics[0].Partitions[0].OfflineReplicas), "should have been 1!") + } +} diff --git a/vendor/github.com/Shopify/sarama/mockresponses.go b/vendor/github.com/Shopify/sarama/mockresponses.go index f79a9d5e9..5541d32ec 100644 --- a/vendor/github.com/Shopify/sarama/mockresponses.go +++ b/vendor/github.com/Shopify/sarama/mockresponses.go @@ -68,9 +68,10 @@ func (mc *MockSequence) For(reqBody versionedDecoder) (res encoder) { // MockMetadataResponse is a `MetadataResponse` builder. type MockMetadataResponse struct { - leaders map[string]map[int32]int32 - brokers map[string]int32 - t TestReporter + controllerID int32 + leaders map[string]map[int32]int32 + brokers map[string]int32 + t TestReporter } func NewMockMetadataResponse(t TestReporter) *MockMetadataResponse { @@ -96,9 +97,17 @@ func (mmr *MockMetadataResponse) SetBroker(addr string, brokerID int32) *MockMet return mmr } +func (mmr *MockMetadataResponse) SetController(brokerID int32) *MockMetadataResponse { + mmr.controllerID = brokerID + return mmr +} + func (mmr *MockMetadataResponse) For(reqBody versionedDecoder) encoder { metadataRequest := reqBody.(*MetadataRequest) - metadataResponse := &MetadataResponse{} + metadataResponse := &MetadataResponse{ + Version: metadataRequest.version(), + ControllerID: mmr.controllerID, + } for addr, brokerID := range mmr.brokers { metadataResponse.AddBroker(addr, brokerID) } @@ -326,6 +335,60 @@ func (mr *MockConsumerMetadataResponse) For(reqBody versionedDecoder) encoder { return res } +// MockFindCoordinatorResponse is a `FindCoordinatorResponse` builder. +type MockFindCoordinatorResponse struct { + groupCoordinators map[string]interface{} + transCoordinators map[string]interface{} + t TestReporter +} + +func NewMockFindCoordinatorResponse(t TestReporter) *MockFindCoordinatorResponse { + return &MockFindCoordinatorResponse{ + groupCoordinators: make(map[string]interface{}), + transCoordinators: make(map[string]interface{}), + t: t, + } +} + +func (mr *MockFindCoordinatorResponse) SetCoordinator(coordinatorType CoordinatorType, group string, broker *MockBroker) *MockFindCoordinatorResponse { + switch coordinatorType { + case CoordinatorGroup: + mr.groupCoordinators[group] = broker + case CoordinatorTransaction: + mr.transCoordinators[group] = broker + } + return mr +} + +func (mr *MockFindCoordinatorResponse) SetError(coordinatorType CoordinatorType, group string, kerror KError) *MockFindCoordinatorResponse { + switch coordinatorType { + case CoordinatorGroup: + mr.groupCoordinators[group] = kerror + case CoordinatorTransaction: + mr.transCoordinators[group] = kerror + } + return mr +} + +func (mr *MockFindCoordinatorResponse) For(reqBody versionedDecoder) encoder { + req := reqBody.(*FindCoordinatorRequest) + res := &FindCoordinatorResponse{} + var v interface{} + switch req.CoordinatorType { + case CoordinatorGroup: + v = mr.groupCoordinators[req.CoordinatorKey] + case CoordinatorTransaction: + v = mr.transCoordinators[req.CoordinatorKey] + } + switch v := v.(type) { + case *MockBroker: + res.Coordinator = &Broker{id: v.BrokerID(), addr: v.Addr()} + case KError: + res.Err = v + } + return res +} + // MockOffsetCommitResponse is a `OffsetCommitResponse` builder. type MockOffsetCommitResponse struct { errors map[string]map[string]map[int32]KError diff --git a/vendor/github.com/Shopify/sarama/offset_commit_request.go b/vendor/github.com/Shopify/sarama/offset_commit_request.go index b21ea634b..37e99fbf5 100644 --- a/vendor/github.com/Shopify/sarama/offset_commit_request.go +++ b/vendor/github.com/Shopify/sarama/offset_commit_request.go @@ -1,5 +1,7 @@ package sarama +import "errors" + // ReceiveTime is a special value for the timestamp field of Offset Commit Requests which // tells the broker to set the timestamp to the time at which the request was received. // The timestamp is only used if message version 1 is used, which requires kafka 0.8.2. @@ -173,7 +175,7 @@ func (r *OffsetCommitRequest) requiredVersion() KafkaVersion { case 2: return V0_9_0_0 default: - return minVersion + return MinVersion } } @@ -188,3 +190,15 @@ func (r *OffsetCommitRequest) AddBlock(topic string, partitionID int32, offset i r.blocks[topic][partitionID] = &offsetCommitRequestBlock{offset, timestamp, metadata} } + +func (r *OffsetCommitRequest) Offset(topic string, partitionID int32) (int64, string, error) { + partitions := r.blocks[topic] + if partitions == nil { + return 0, "", errors.New("No such offset") + } + block := partitions[partitionID] + if block == nil { + return 0, "", errors.New("No such offset") + } + return block.offset, block.metadata, nil +} diff --git a/vendor/github.com/Shopify/sarama/offset_commit_response.go b/vendor/github.com/Shopify/sarama/offset_commit_response.go index 7f277e775..a4b18acdf 100644 --- a/vendor/github.com/Shopify/sarama/offset_commit_response.go +++ b/vendor/github.com/Shopify/sarama/offset_commit_response.go @@ -81,5 +81,5 @@ func (r *OffsetCommitResponse) version() int16 { } func (r *OffsetCommitResponse) requiredVersion() KafkaVersion { - return minVersion + return MinVersion } diff --git a/vendor/github.com/Shopify/sarama/offset_fetch_request.go b/vendor/github.com/Shopify/sarama/offset_fetch_request.go index b19fe79ba..5a05014b4 100644 --- a/vendor/github.com/Shopify/sarama/offset_fetch_request.go +++ b/vendor/github.com/Shopify/sarama/offset_fetch_request.go @@ -68,7 +68,7 @@ func (r *OffsetFetchRequest) requiredVersion() KafkaVersion { case 1: return V0_8_2_0 default: - return minVersion + return MinVersion } } diff --git a/vendor/github.com/Shopify/sarama/offset_fetch_response.go b/vendor/github.com/Shopify/sarama/offset_fetch_response.go index 323220eac..11e4b1f3f 100644 --- a/vendor/github.com/Shopify/sarama/offset_fetch_response.go +++ b/vendor/github.com/Shopify/sarama/offset_fetch_response.go @@ -115,7 +115,7 @@ func (r *OffsetFetchResponse) version() int16 { } func (r *OffsetFetchResponse) requiredVersion() KafkaVersion { - return minVersion + return MinVersion } func (r *OffsetFetchResponse) GetBlock(topic string, partition int32) *OffsetFetchResponseBlock { diff --git a/vendor/github.com/Shopify/sarama/offset_request.go b/vendor/github.com/Shopify/sarama/offset_request.go index 6c2696016..4c5df75df 100644 --- a/vendor/github.com/Shopify/sarama/offset_request.go +++ b/vendor/github.com/Shopify/sarama/offset_request.go @@ -109,7 +109,7 @@ func (r *OffsetRequest) requiredVersion() KafkaVersion { case 1: return V0_10_1_0 default: - return minVersion + return MinVersion } } diff --git a/vendor/github.com/Shopify/sarama/offset_response.go b/vendor/github.com/Shopify/sarama/offset_response.go index 9a9cfe96f..8b2193f9a 100644 --- a/vendor/github.com/Shopify/sarama/offset_response.go +++ b/vendor/github.com/Shopify/sarama/offset_response.go @@ -155,7 +155,7 @@ func (r *OffsetResponse) requiredVersion() KafkaVersion { case 1: return V0_10_1_0 default: - return minVersion + return MinVersion } } diff --git a/vendor/github.com/Shopify/sarama/produce_request.go b/vendor/github.com/Shopify/sarama/produce_request.go index 0ec4d8d53..0c755d02b 100644 --- a/vendor/github.com/Shopify/sarama/produce_request.go +++ b/vendor/github.com/Shopify/sarama/produce_request.go @@ -113,9 +113,9 @@ func (r *ProduceRequest) encode(pe packetEncoder) error { } if metricRegistry != nil { if r.Version >= 3 { - topicRecordCount += updateBatchMetrics(records.recordBatch, compressionRatioMetric, topicCompressionRatioMetric) + topicRecordCount += updateBatchMetrics(records.RecordBatch, compressionRatioMetric, topicCompressionRatioMetric) } else { - topicRecordCount += updateMsgSetMetrics(records.msgSet, compressionRatioMetric, topicCompressionRatioMetric) + topicRecordCount += updateMsgSetMetrics(records.MsgSet, compressionRatioMetric, topicCompressionRatioMetric) } batchSize := int64(pe.offset() - startOffset) batchSizeMetric.Update(batchSize) @@ -215,7 +215,7 @@ func (r *ProduceRequest) requiredVersion() KafkaVersion { case 3: return V0_11_0_0 default: - return minVersion + return MinVersion } } @@ -231,7 +231,7 @@ func (r *ProduceRequest) ensureRecords(topic string, partition int32) { func (r *ProduceRequest) AddMessage(topic string, partition int32, msg *Message) { r.ensureRecords(topic, partition) - set := r.records[topic][partition].msgSet + set := r.records[topic][partition].MsgSet if set == nil { set = new(MessageSet) diff --git a/vendor/github.com/Shopify/sarama/produce_request_test.go b/vendor/github.com/Shopify/sarama/produce_request_test.go index f6e84bc9b..b9896eb61 100644 --- a/vendor/github.com/Shopify/sarama/produce_request_test.go +++ b/vendor/github.com/Shopify/sarama/produce_request_test.go @@ -99,6 +99,8 @@ func TestProduceRequest(t *testing.T) { } request.AddBatch("topic", 0xAD, batch) packet := testRequestEncode(t, "one record", request, produceRequestOneRecord) - batch.Records[0].length.startOffset = 0 + // compressRecords field is not populated on decoding because consumers + // are only interested in decoded records. + batch.compressedRecords = nil testRequestDecode(t, "one record", request, packet) } diff --git a/vendor/github.com/Shopify/sarama/produce_response.go b/vendor/github.com/Shopify/sarama/produce_response.go index 043c40f87..667e34c66 100644 --- a/vendor/github.com/Shopify/sarama/produce_response.go +++ b/vendor/github.com/Shopify/sarama/produce_response.go @@ -152,7 +152,7 @@ func (r *ProduceResponse) requiredVersion() KafkaVersion { case 3: return V0_11_0_0 default: - return minVersion + return MinVersion } } diff --git a/vendor/github.com/Shopify/sarama/produce_set.go b/vendor/github.com/Shopify/sarama/produce_set.go index 3cbaeb5f9..13be2b3c9 100644 --- a/vendor/github.com/Shopify/sarama/produce_set.go +++ b/vendor/github.com/Shopify/sarama/produce_set.go @@ -59,10 +59,11 @@ func (ps *produceSet) add(msg *ProducerMessage) error { if set == nil { if ps.parent.conf.Version.IsAtLeast(V0_11_0_0) { batch := &RecordBatch{ - FirstTimestamp: timestamp, - Version: 2, - ProducerID: -1, /* No producer id */ - Codec: ps.parent.conf.Producer.Compression, + FirstTimestamp: timestamp, + Version: 2, + ProducerID: -1, /* No producer id */ + Codec: ps.parent.conf.Producer.Compression, + CompressionLevel: ps.parent.conf.Producer.CompressionLevel, } set = &partitionSet{recordsToSend: newDefaultRecords(batch)} size = recordBatchOverhead @@ -79,7 +80,7 @@ func (ps *produceSet) add(msg *ProducerMessage) error { rec := &Record{ Key: key, Value: val, - TimestampDelta: timestamp.Sub(set.recordsToSend.recordBatch.FirstTimestamp), + TimestampDelta: timestamp.Sub(set.recordsToSend.RecordBatch.FirstTimestamp), } size += len(key) + len(val) if len(msg.Headers) > 0 { @@ -89,14 +90,14 @@ func (ps *produceSet) add(msg *ProducerMessage) error { size += len(rec.Headers[i].Key) + len(rec.Headers[i].Value) + 2*binary.MaxVarintLen32 } } - set.recordsToSend.recordBatch.addRecord(rec) + set.recordsToSend.RecordBatch.addRecord(rec) } else { msgToSend := &Message{Codec: CompressionNone, Key: key, Value: val} if ps.parent.conf.Version.IsAtLeast(V0_10_0_0) { msgToSend.Timestamp = timestamp msgToSend.Version = 1 } - set.recordsToSend.msgSet.addMessage(msgToSend) + set.recordsToSend.MsgSet.addMessage(msgToSend) size = producerMessageOverhead + len(key) + len(val) } @@ -122,7 +123,14 @@ func (ps *produceSet) buildRequest() *ProduceRequest { for topic, partitionSet := range ps.msgs { for partition, set := range partitionSet { if req.Version >= 3 { - rb := set.recordsToSend.recordBatch + // If the API version we're hitting is 3 or greater, we need to calculate + // offsets for each record in the batch relative to FirstOffset. + // Additionally, we must set LastOffsetDelta to the value of the last offset + // in the batch. Since the OffsetDelta of the first record is 0, we know that the + // final record of any batch will have an offset of (# of records in batch) - 1. + // (See https://cwiki.apache.org/confluence/display/KAFKA/A+Guide+To+The+Kafka+Protocol#AGuideToTheKafkaProtocol-Messagesets + // under the RecordBatch section for details.) + rb := set.recordsToSend.RecordBatch if len(rb.Records) > 0 { rb.LastOffsetDelta = int32(len(rb.Records) - 1) for i, record := range rb.Records { @@ -134,7 +142,7 @@ func (ps *produceSet) buildRequest() *ProduceRequest { continue } if ps.parent.conf.Producer.Compression == CompressionNone { - req.AddSet(topic, partition, set.recordsToSend.msgSet) + req.AddSet(topic, partition, set.recordsToSend.MsgSet) } else { // When compression is enabled, the entire set for each partition is compressed // and sent as the payload of a single fake "message" with the appropriate codec @@ -147,24 +155,25 @@ func (ps *produceSet) buildRequest() *ProduceRequest { // recompressing the message set. // (See https://cwiki.apache.org/confluence/display/KAFKA/KIP-31+-+Move+to+relative+offsets+in+compressed+message+sets // for details on relative offsets.) - for i, msg := range set.recordsToSend.msgSet.Messages { + for i, msg := range set.recordsToSend.MsgSet.Messages { msg.Offset = int64(i) } } - payload, err := encode(set.recordsToSend.msgSet, ps.parent.conf.MetricRegistry) + payload, err := encode(set.recordsToSend.MsgSet, ps.parent.conf.MetricRegistry) if err != nil { Logger.Println(err) // if this happens, it's basically our fault. panic(err) } compMsg := &Message{ - Codec: ps.parent.conf.Producer.Compression, - Key: nil, - Value: payload, - Set: set.recordsToSend.msgSet, // Provide the underlying message set for accurate metrics + Codec: ps.parent.conf.Producer.Compression, + CompressionLevel: ps.parent.conf.Producer.CompressionLevel, + Key: nil, + Value: payload, + Set: set.recordsToSend.MsgSet, // Provide the underlying message set for accurate metrics } if ps.parent.conf.Version.IsAtLeast(V0_10_0_0) { compMsg.Version = 1 - compMsg.Timestamp = set.recordsToSend.msgSet.Messages[0].Msg.Timestamp + compMsg.Timestamp = set.recordsToSend.MsgSet.Messages[0].Msg.Timestamp } req.AddMessage(topic, partition, compMsg) } diff --git a/vendor/github.com/Shopify/sarama/produce_set_test.go b/vendor/github.com/Shopify/sarama/produce_set_test.go index fee6e3fa1..6663f36f7 100644 --- a/vendor/github.com/Shopify/sarama/produce_set_test.go +++ b/vendor/github.com/Shopify/sarama/produce_set_test.go @@ -167,7 +167,7 @@ func TestProduceSetCompressedRequestBuilding(t *testing.T) { t.Error("Wrong request version") } - for _, msgBlock := range req.records["t1"][0].msgSet.Messages { + for _, msgBlock := range req.records["t1"][0].MsgSet.Messages { msg := msgBlock.Msg err := msg.decodeSet() if err != nil { @@ -227,7 +227,7 @@ func TestProduceSetV3RequestBuilding(t *testing.T) { t.Error("Wrong request version") } - batch := req.records["t1"][0].recordBatch + batch := req.records["t1"][0].RecordBatch if batch.FirstTimestamp != now { t.Errorf("Wrong first timestamp: %v", batch.FirstTimestamp) } diff --git a/vendor/github.com/Shopify/sarama/record_batch.go b/vendor/github.com/Shopify/sarama/record_batch.go index 321de485b..845318aa3 100644 --- a/vendor/github.com/Shopify/sarama/record_batch.go +++ b/vendor/github.com/Shopify/sarama/record_batch.go @@ -40,6 +40,7 @@ type RecordBatch struct { PartitionLeaderEpoch int32 Version int8 Codec CompressionCodec + CompressionLevel int Control bool LastOffsetDelta int32 FirstTimestamp time.Time @@ -219,7 +220,15 @@ func (b *RecordBatch) encodeRecords(pe packetEncoder) error { b.compressedRecords = raw case CompressionGZIP: var buf bytes.Buffer - writer := gzip.NewWriter(&buf) + var writer *gzip.Writer + if b.CompressionLevel != CompressionLevelDefault { + writer, err = gzip.NewWriterLevel(&buf, b.CompressionLevel) + if err != nil { + return err + } + } else { + writer = gzip.NewWriter(&buf) + } if _, err := writer.Write(raw); err != nil { return err } diff --git a/vendor/github.com/Shopify/sarama/record_test.go b/vendor/github.com/Shopify/sarama/record_test.go index a3fb8f501..2756c5b25 100644 --- a/vendor/github.com/Shopify/sarama/record_test.go +++ b/vendor/github.com/Shopify/sarama/record_test.go @@ -116,11 +116,12 @@ var recordBatchTestCases = []struct { { name: "gzipped record", batch: RecordBatch{ - Version: 2, - Codec: CompressionGZIP, - FirstTimestamp: time.Unix(1479847795, 0), - MaxTimestamp: time.Unix(0, 0), - LastOffsetDelta: 0, + Version: 2, + Codec: CompressionGZIP, + CompressionLevel: CompressionLevelDefault, + FirstTimestamp: time.Unix(1479847795, 0), + MaxTimestamp: time.Unix(0, 0), + LastOffsetDelta: 0, Records: []*Record{{ TimestampDelta: 5 * time.Millisecond, Key: []byte{1, 2, 3, 4}, @@ -281,6 +282,9 @@ func TestRecordBatchDecoding(t *testing.T) { for _, r := range tc.batch.Records { r.length = varintLengthField{} } + // The compression level is not restored on decoding. It is not needed + // anyway. We only set it here to ensure that comparision succeeds. + batch.CompressionLevel = tc.batch.CompressionLevel if !reflect.DeepEqual(batch, tc.batch) { t.Errorf(spew.Sprintf("invalid decode of %s\ngot %+v\nwanted %+v", tc.name, batch, tc.batch)) } diff --git a/vendor/github.com/Shopify/sarama/records.go b/vendor/github.com/Shopify/sarama/records.go index 258dcbac8..301055bb0 100644 --- a/vendor/github.com/Shopify/sarama/records.go +++ b/vendor/github.com/Shopify/sarama/records.go @@ -14,30 +14,30 @@ const ( // Records implements a union type containing either a RecordBatch or a legacy MessageSet. type Records struct { recordsType int - msgSet *MessageSet - recordBatch *RecordBatch + MsgSet *MessageSet + RecordBatch *RecordBatch } func newLegacyRecords(msgSet *MessageSet) Records { - return Records{recordsType: legacyRecords, msgSet: msgSet} + return Records{recordsType: legacyRecords, MsgSet: msgSet} } func newDefaultRecords(batch *RecordBatch) Records { - return Records{recordsType: defaultRecords, recordBatch: batch} + return Records{recordsType: defaultRecords, RecordBatch: batch} } -// setTypeFromFields sets type of Records depending on which of msgSet or recordBatch is not nil. +// setTypeFromFields sets type of Records depending on which of MsgSet or RecordBatch is not nil. // The first return value indicates whether both fields are nil (and the type is not set). // If both fields are not nil, it returns an error. func (r *Records) setTypeFromFields() (bool, error) { - if r.msgSet == nil && r.recordBatch == nil { + if r.MsgSet == nil && r.RecordBatch == nil { return true, nil } - if r.msgSet != nil && r.recordBatch != nil { - return false, fmt.Errorf("both msgSet and recordBatch are set, but record type is unknown") + if r.MsgSet != nil && r.RecordBatch != nil { + return false, fmt.Errorf("both MsgSet and RecordBatch are set, but record type is unknown") } r.recordsType = defaultRecords - if r.msgSet != nil { + if r.MsgSet != nil { r.recordsType = legacyRecords } return false, nil @@ -52,15 +52,15 @@ func (r *Records) encode(pe packetEncoder) error { switch r.recordsType { case legacyRecords: - if r.msgSet == nil { + if r.MsgSet == nil { return nil } - return r.msgSet.encode(pe) + return r.MsgSet.encode(pe) case defaultRecords: - if r.recordBatch == nil { + if r.RecordBatch == nil { return nil } - return r.recordBatch.encode(pe) + return r.RecordBatch.encode(pe) } return fmt.Errorf("unknown records type: %v", r.recordsType) @@ -89,11 +89,11 @@ func (r *Records) decode(pd packetDecoder) error { switch r.recordsType { case legacyRecords: - r.msgSet = &MessageSet{} - return r.msgSet.decode(pd) + r.MsgSet = &MessageSet{} + return r.MsgSet.decode(pd) case defaultRecords: - r.recordBatch = &RecordBatch{} - return r.recordBatch.decode(pd) + r.RecordBatch = &RecordBatch{} + return r.RecordBatch.decode(pd) } return fmt.Errorf("unknown records type: %v", r.recordsType) } @@ -107,15 +107,15 @@ func (r *Records) numRecords() (int, error) { switch r.recordsType { case legacyRecords: - if r.msgSet == nil { + if r.MsgSet == nil { return 0, nil } - return len(r.msgSet.Messages), nil + return len(r.MsgSet.Messages), nil case defaultRecords: - if r.recordBatch == nil { + if r.RecordBatch == nil { return 0, nil } - return len(r.recordBatch.Records), nil + return len(r.RecordBatch.Records), nil } return 0, fmt.Errorf("unknown records type: %v", r.recordsType) } @@ -131,15 +131,15 @@ func (r *Records) isPartial() (bool, error) { case unknownRecords: return false, nil case legacyRecords: - if r.msgSet == nil { + if r.MsgSet == nil { return false, nil } - return r.msgSet.PartialTrailingMessage, nil + return r.MsgSet.PartialTrailingMessage, nil case defaultRecords: - if r.recordBatch == nil { + if r.RecordBatch == nil { return false, nil } - return r.recordBatch.PartialTrailingRecord, nil + return r.RecordBatch.PartialTrailingRecord, nil } return false, fmt.Errorf("unknown records type: %v", r.recordsType) } @@ -155,10 +155,10 @@ func (r *Records) isControl() (bool, error) { case legacyRecords: return false, nil case defaultRecords: - if r.recordBatch == nil { + if r.RecordBatch == nil { return false, nil } - return r.recordBatch.Control, nil + return r.RecordBatch.Control, nil } return false, fmt.Errorf("unknown records type: %v", r.recordsType) } diff --git a/vendor/github.com/Shopify/sarama/records_test.go b/vendor/github.com/Shopify/sarama/records_test.go index e8bcb6c35..cd6aa2784 100644 --- a/vendor/github.com/Shopify/sarama/records_test.go +++ b/vendor/github.com/Shopify/sarama/records_test.go @@ -45,8 +45,8 @@ func TestLegacyRecords(t *testing.T) { if r.recordsType != legacyRecords { t.Fatalf("Wrong records type %v, expected %v", r.recordsType, legacyRecords) } - if !reflect.DeepEqual(set, r.msgSet) { - t.Errorf("Wrong decoding for legacy records, wanted %#+v, got %#+v", set, r.msgSet) + if !reflect.DeepEqual(set, r.MsgSet) { + t.Errorf("Wrong decoding for legacy records, wanted %#+v, got %#+v", set, r.MsgSet) } n, err := r.numRecords() @@ -113,8 +113,8 @@ func TestDefaultRecords(t *testing.T) { if r.recordsType != defaultRecords { t.Fatalf("Wrong records type %v, expected %v", r.recordsType, defaultRecords) } - if !reflect.DeepEqual(batch, r.recordBatch) { - t.Errorf("Wrong decoding for default records, wanted %#+v, got %#+v", batch, r.recordBatch) + if !reflect.DeepEqual(batch, r.RecordBatch) { + t.Errorf("Wrong decoding for default records, wanted %#+v, got %#+v", batch, r.RecordBatch) } n, err := r.numRecords() diff --git a/vendor/github.com/Shopify/sarama/request.go b/vendor/github.com/Shopify/sarama/request.go index 5f7cb76e9..4d211a14f 100644 --- a/vendor/github.com/Shopify/sarama/request.go +++ b/vendor/github.com/Shopify/sarama/request.go @@ -97,7 +97,7 @@ func allocateBody(key, version int16) protocolBody { case 9: return &OffsetFetchRequest{} case 10: - return &ConsumerMetadataRequest{} + return &FindCoordinatorRequest{} case 11: return &JoinGroupRequest{} case 12: @@ -118,6 +118,8 @@ func allocateBody(key, version int16) protocolBody { return &CreateTopicsRequest{} case 20: return &DeleteTopicsRequest{} + case 21: + return &DeleteRecordsRequest{} case 22: return &InitProducerIDRequest{} case 24: @@ -140,6 +142,8 @@ func allocateBody(key, version int16) protocolBody { return &AlterConfigsRequest{} case 37: return &CreatePartitionsRequest{} + case 42: + return &DeleteGroupsRequest{} } return nil } diff --git a/vendor/github.com/Shopify/sarama/request_test.go b/vendor/github.com/Shopify/sarama/request_test.go index ba830d613..ffc2bb7a8 100644 --- a/vendor/github.com/Shopify/sarama/request_test.go +++ b/vendor/github.com/Shopify/sarama/request_test.go @@ -50,6 +50,9 @@ func testVersionDecodable(t *testing.T, name string, out versionedDecoder, in [] } func testRequest(t *testing.T, name string, rb protocolBody, expected []byte) { + if !rb.requiredVersion().IsAtLeast(MinVersion) { + t.Errorf("Request %s has invalid required version", name) + } packet := testRequestEncode(t, name, rb, expected) testRequestDecode(t, name, rb, packet) } @@ -76,6 +79,8 @@ func testRequestDecode(t *testing.T, name string, rb protocolBody, packet []byte t.Error(spew.Sprintf("Decoded request %q does not match the encoded one\nencoded: %+v\ndecoded: %+v", name, rb, decoded.body)) } else if n != len(packet) { t.Errorf("Decoded request %q bytes: %d does not match the encoded one: %d\n", name, n, len(packet)) + } else if rb.version() != decoded.body.version() { + t.Errorf("Decoded request %q version: %d does not match the encoded one: %d\n", name, decoded.body.version(), rb.version()) } } diff --git a/vendor/github.com/Shopify/sarama/utils.go b/vendor/github.com/Shopify/sarama/utils.go index 9d7b60f16..702e22627 100644 --- a/vendor/github.com/Shopify/sarama/utils.go +++ b/vendor/github.com/Shopify/sarama/utils.go @@ -139,21 +139,49 @@ func (v KafkaVersion) IsAtLeast(other KafkaVersion) bool { // Effective constants defining the supported kafka versions. var ( - V0_8_2_0 = newKafkaVersion(0, 8, 2, 0) - V0_8_2_1 = newKafkaVersion(0, 8, 2, 1) - V0_8_2_2 = newKafkaVersion(0, 8, 2, 2) - V0_9_0_0 = newKafkaVersion(0, 9, 0, 0) - V0_9_0_1 = newKafkaVersion(0, 9, 0, 1) - V0_10_0_0 = newKafkaVersion(0, 10, 0, 0) - V0_10_0_1 = newKafkaVersion(0, 10, 0, 1) - V0_10_1_0 = newKafkaVersion(0, 10, 1, 0) - V0_10_2_0 = newKafkaVersion(0, 10, 2, 0) - V0_11_0_0 = newKafkaVersion(0, 11, 0, 0) - V1_0_0_0 = newKafkaVersion(1, 0, 0, 0) - minVersion = V0_8_2_0 + V0_8_2_0 = newKafkaVersion(0, 8, 2, 0) + V0_8_2_1 = newKafkaVersion(0, 8, 2, 1) + V0_8_2_2 = newKafkaVersion(0, 8, 2, 2) + V0_9_0_0 = newKafkaVersion(0, 9, 0, 0) + V0_9_0_1 = newKafkaVersion(0, 9, 0, 1) + V0_10_0_0 = newKafkaVersion(0, 10, 0, 0) + V0_10_0_1 = newKafkaVersion(0, 10, 0, 1) + V0_10_1_0 = newKafkaVersion(0, 10, 1, 0) + V0_10_1_1 = newKafkaVersion(0, 10, 1, 1) + V0_10_2_0 = newKafkaVersion(0, 10, 2, 0) + V0_10_2_1 = newKafkaVersion(0, 10, 2, 1) + V0_11_0_0 = newKafkaVersion(0, 11, 0, 0) + V0_11_0_1 = newKafkaVersion(0, 11, 0, 1) + V0_11_0_2 = newKafkaVersion(0, 11, 0, 2) + V1_0_0_0 = newKafkaVersion(1, 0, 0, 0) + V1_1_0_0 = newKafkaVersion(1, 1, 0, 0) + + SupportedVersions = []KafkaVersion{ + V0_8_2_0, + V0_8_2_1, + V0_8_2_2, + V0_9_0_0, + V0_9_0_1, + V0_10_0_0, + V0_10_0_1, + V0_10_1_0, + V0_10_1_1, + V0_10_2_0, + V0_10_2_1, + V0_11_0_0, + V0_11_0_1, + V0_11_0_2, + V1_0_0_0, + V1_1_0_0, + } + MinVersion = V0_8_2_0 + MaxVersion = V1_1_0_0 ) func ParseKafkaVersion(s string) (KafkaVersion, error) { + if len(s) < 5 { + return MinVersion, fmt.Errorf("invalid version `%s`", s) + } var major, minor, veryMinor, patch uint var err error if s[0] == '0' { @@ -162,7 +190,7 @@ func ParseKafkaVersion(s string) (KafkaVersion, error) { err = scanKafkaVersion(s, `^\d+\.\d+\.\d+$`, "%d.%d.%d", [3]*uint{&major, &minor, &veryMinor}) } if err != nil { - return minVersion, err + return MinVersion, err } return newKafkaVersion(major, minor, veryMinor, patch), nil } diff --git a/vendor/github.com/aws/aws-sdk-go/.github/ISSUE_TEMPLATE.md b/vendor/github.com/aws/aws-sdk-go/.github/ISSUE_TEMPLATE.md index 1ecb636bf..245949d00 100644 --- a/vendor/github.com/aws/aws-sdk-go/.github/ISSUE_TEMPLATE.md +++ b/vendor/github.com/aws/aws-sdk-go/.github/ISSUE_TEMPLATE.md @@ -10,5 +10,5 @@ Please fill out the sections below to help us address your issue. ### Steps to reproduce -If you have have an runnable example, please include it. +If you have an runnable example, please include it. diff --git a/vendor/github.com/aws/aws-sdk-go/.godoc_config b/vendor/github.com/aws/aws-sdk-go/.godoc_config index 395878d6c..0c6400bc7 100644 --- a/vendor/github.com/aws/aws-sdk-go/.godoc_config +++ b/vendor/github.com/aws/aws-sdk-go/.godoc_config @@ -3,7 +3,7 @@ "Pattern": "/sdk-for-go/api/", "StripPrefix": "/sdk-for-go/api", "Include": ["/src/github.com/aws/aws-sdk-go/aws", "/src/github.com/aws/aws-sdk-go/service"], - "Exclude": ["/src/cmd", "/src/github.com/aws/aws-sdk-go/awstesting", "/src/github.com/aws/aws-sdk-go/awsmigrate"], + "Exclude": ["/src/cmd", "/src/github.com/aws/aws-sdk-go/awstesting", "/src/github.com/aws/aws-sdk-go/awsmigrate", "/src/github.com/aws/aws-sdk-go/private"], "IgnoredSuffixes": ["iface"] }, "Github": { diff --git a/vendor/github.com/aws/aws-sdk-go/.travis.yml b/vendor/github.com/aws/aws-sdk-go/.travis.yml index 7659cd7d5..0b34640df 100644 --- a/vendor/github.com/aws/aws-sdk-go/.travis.yml +++ b/vendor/github.com/aws/aws-sdk-go/.travis.yml @@ -40,11 +40,8 @@ matrix: - os: osx go: tip -install: - - make get-deps - script: - - make unit-with-race-cover + - make ci-test branches: only: diff --git a/vendor/github.com/aws/aws-sdk-go/CHANGELOG.md b/vendor/github.com/aws/aws-sdk-go/CHANGELOG.md index 55b9b6beb..4b63e1053 100644 --- a/vendor/github.com/aws/aws-sdk-go/CHANGELOG.md +++ b/vendor/github.com/aws/aws-sdk-go/CHANGELOG.md @@ -1,3 +1,306 @@ +Release v1.14.17 (2018-06-29) +=== + +### Service Client Updates +* `service/secretsmanager`: Updates service examples + * New SDK code snippet examples for the new APIs released for the Resource-based Policy support in Secrets Manager + +Release v1.14.16 (2018-06-28) +=== + +### Service Client Updates +* `service/elasticbeanstalk`: Updates service API, documentation, and examples + * Elastic Beanstalk adds "Suspended" health status to the EnvironmentHealthStatus enum type and updates document. +* `service/lambda`: Updates service API and documentation + * Support for SQS as an event source. +* `service/storagegateway`: Updates service API, documentation, and examples + * AWS Storage Gateway now enables you to use Server Message Block (SMB) protocol to store and access objects in Amazon Simple Storage Service (S3). + +Release v1.14.15 (2018-06-27) +=== + +### Service Client Updates +* `service/cloudfront`: Updates service API and documentation + * Unpublish delete-service-linked-role API. +* `service/codepipeline`: Updates service API + * UpdatePipeline may now throw a LimitExceededException when adding or updating Source Actions that use periodic checks for change detection +* `service/comprehend`: Updates service API, documentation, and paginators +* `service/secretsmanager`: Updates service documentation, paginators, and examples + * Documentation updates for secretsmanager + +### SDK Bugs +* `aws/csm`: Final API Call Attempt events were not being called [#2008](https://github.com/aws/aws-sdk-go/pull/2008) +Release v1.14.14 (2018-06-26) +=== + +### Service Client Updates +* `service/inspector`: Updates service API, documentation, and paginators + * Introduce four new APIs to view and preview Exclusions. Exclusions show which intended security checks are excluded from an assessment, along with reasons and recommendations to fix. The APIs are CreateExclusionsPreview, GetExclusionsPreview, ListExclusions, and DescribeExclusions. +* `service/s3`: Updates service API and documentation + * Add AllowQuotedRecordDelimiter to Amazon S3 Select API. Please refer to https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html for usage details. +* `service/secretsmanager`: Updates service API, documentation, paginators, and examples + * This release adds support for resource-based policies that attach directly to your secrets. These policies provide an additional way to control who can access your secrets and what they can do with them. For more information, see https://docs.aws.amazon.com/secretsmanager/latest/userguide/auth-and-access_resource-based-policies.html in the Secrets Manager User Guide. + +Release v1.14.13 (2018-06-22) +=== + +### Service Client Updates +* `service/alexaforbusiness`: Updates service API and documentation +* `service/appstream`: Updates service API, documentation, paginators, and examples + * This API update enables customers to find their VPC private IP address and ENI ID associated with AppStream streaming sessions. +* `aws/endpoints`: Updated Regions and Endpoints metadata. + +Release v1.14.12 (2018-06-21) +=== + +### Service Client Updates +* `service/clouddirectory`: Adds new service + * SDK release to support Flexible Schema initiative being carried out by Amazon Cloud Directory. This feature lets customers using new capabilities like: variant typed attributes, dynamic facets and AWS managed Cloud Directory schemas. + +Release v1.14.11 (2018-06-21) +=== + +### Service Client Updates +* `service/macie`: Adds new service + * Amazon Macie is a security service that uses machine learning to automatically discover, classify, and protect sensitive data in AWS. With this release, we are launching the following Macie HTTPS API operations: AssociateMemberAccount, AssociateS3Resources, DisassociateMemberAccount, DisassociateS3Resources, ListMemberAccounts, ListS3Resources, and UpdateS3Resources. With these API operations you can issue HTTPS requests directly to the service. +* `service/neptune`: Updates service API, documentation, and examples + * Deprecates the PubliclyAccessible parameter that is not supported by Amazon Neptune. +* `service/ssm`: Updates service API, documentation, and examples + * Adds Amazon Linux 2 support to Patch Manager + +Release v1.14.10 (2018-06-20) +=== + +### Service Client Updates +* `service/acm-pca`: Updates service API, documentation, paginators, and examples +* `service/medialive`: Updates service API, documentation, and paginators + * AWS Elemental MediaLive now makes Reserved Outputs and Inputs available through the AWS Management Console and API. You can reserve outputs and inputs with a 12 month commitment in exchange for discounted hourly rates. Pricing is available at https://aws.amazon.com/medialive/pricing/ +* `service/rds`: Updates service API, documentation, and examples + * This release adds a new parameter to specify the retention period for Performance Insights data for RDS instances. You can either choose 7 days (default) or 731 days. For more information, see Amazon RDS Documentation. + +### SDK Enhancements +* `service/s3`: Update SelectObjectContent doc example to be on the API not nested type. ([#1991](https://github.com/aws/aws-sdk-go/pull/1991)) + +### SDK Bugs +* `aws/client`: Fix HTTP debug log EventStream payloads ([#2000](https://github.com/aws/aws-sdk-go/pull/2000)) + * Fixes the SDK's HTTP client debug logging to not log the HTTP response body for EventStreams. This prevents the SDK from buffering a very large amount of data to be logged at once. The aws.LogDebugWithEventStreamBody should be used to log the event stream events. + * Fixes a bug in the SDK's response logger which will buffer the response body's content if LogDebug is enabled but LogDebugWithHTTPBody is not. +Release v1.14.9 (2018-06-19) +=== + +### Service Client Updates +* `aws/endpoints`: Updated Regions and Endpoints metadata. +* `service/rekognition`: Updates service documentation and examples + * Documentation updates for rekognition + +### SDK Bugs +* `private/model/api`: Update client ServiceName to be based on name of service for new services. ([#1997](https://github.com/aws/aws-sdk-go/pull/1997)) + * Fixes the SDK's `ServiceName` AWS service client package value to be unique based on the service name for new AWS services. Does not change exiting client packages. +Release v1.14.8 (2018-06-15) +=== + +### Service Client Updates +* `service/mediaconvert`: Updates service API and documentation + * This release adds language code support according to the ISO-639-3 standard. Custom 3-character language codes are now supported on input and output for both audio and captions. + +Release v1.14.7 (2018-06-14) +=== + +### Service Client Updates +* `service/apigateway`: Updates service API and documentation + * Support for PRIVATE endpoint configuration type +* `service/dynamodb`: Updates service API and documentation + * Added two new fields SSEType and KMSMasterKeyArn to SSEDescription block in describe-table output. +* `service/iotanalytics`: Updates service API and documentation + +Release v1.14.6 (2018-06-13) +=== + +### Service Client Updates +* `service/servicecatalog`: Updates service API + * Introduced new length limitations for few of the product fields. +* `service/ssm`: Updates service API and documentation + * Added support for new parameter, CloudWatchOutputConfig, for SendCommand API. Users can now have RunCommand output sent to CloudWatchLogs. + +Release v1.14.5 (2018-06-12) +=== + +### Service Client Updates +* `service/devicefarm`: Updates service API and documentation + * Adding VPCEndpoint support for Remote access. Allows customers to be able to access their private endpoints/services running in their VPC during remote access. +* `service/ecs`: Updates service API and documentation + * Introduces daemon scheduling capability to deploy one task per instance on selected instances in a cluster. Adds a "force" flag to the DeleteService API to delete a service without requiring to scale down the number of tasks to zero. + +### SDK Enhancements +* `service/rds/rdsutils`: Clean up the rdsutils package and adds a new builder to construct connection strings ([#1985](https://github.com/aws/aws-sdk-go/pull/1985)) + * Rewords documentation to be more useful and provides links to prior setup needed to support authentication tokens. Introduces a builder that allows for building connection strings + +### SDK Bugs +* `aws/signer/v4`: Fix X-Amz-Content-Sha256 being in to query for presign ([#1976](https://github.com/aws/aws-sdk-go/pull/1976)) + * Fixes the bug which would allow the X-Amz-Content-Sha256 header to be promoted to the query string when presigning a S3 request. This bug also was preventing users from setting their own sha256 value for a presigned URL. Presigned requests generated with the custom sha256 would of always failed with invalid signature. + * Fixes [#1974](https://github.com/aws/aws-sdk-go/pull/1974) +Release v1.14.4 (2018-06-11) +=== + +### Service Client Updates +* `service/clouddirectory`: Updates service API and documentation + * Amazon Cloud Directory now supports optional attributes on Typed Links, giving users the ability to associate and manage data on Typed Links. +* `service/rds`: Updates service documentation + * Changed lists of valid EngineVersion values to links to the RDS User Guide. +* `service/storagegateway`: Updates service API and documentation + * AWS Storage Gateway now enables you to create cached volumes and tapes with AWS KMS support. + +Release v1.14.3 (2018-06-08) +=== + +### Service Client Updates +* `service/mediatailor`: Updates service API + +Release v1.14.2 (2018-06-07) +=== + +### Service Client Updates +* `service/medialive`: Updates service API, documentation, and paginators + * AWS Elemental MediaLive now makes channel log information available through Amazon CloudWatch Logs. You can set up each MediaLive channel with a logging level; when the channel is run, logs will automatically be published to your account on Amazon CloudWatch Logs + +Release v1.14.1 (2018-06-05) +=== + +### Service Client Updates +* `service/ce`: Updates service API and documentation +* `service/polly`: Updates service API and documentation + * Amazon Polly adds new French voice - "Lea" +* `service/rds`: Updates service API and documentation + * This release adds customizable processor features for RDS instances. +* `service/secretsmanager`: Updates service documentation + * Documentation updates for secretsmanager +* `service/shield`: Updates service API and documentation + * DDoS Response Team access management for AWS Shield + +Release v1.14.0 (2018-06-04) +=== + +### Service Client Updates +* `service/AWSMigrationHub`: Updates service documentation +* `service/appstream`: Updates service API and documentation + * Amazon AppStream 2.0 adds support for Google Drive for G Suite. With this feature, customers will be able to connect their G Suite accounts with AppStream 2.0 and enable Google Drive access for an AppStream 2.0 stack. Users of the stack can then link their Google Drive using their G Suite login credentials and use their existing files stored in Drive with their AppStream 2.0 applications. File changes will be synced automatically to Google cloud. +* `service/ec2`: Updates service API and documentation + * You are now able to use instance storage (up to 3600 GB of NVMe based SSD) on M5 instances, the next generation of EC2's General Purpose instances in us-east-1, us-west-2, us-east-2, eu-west-1 and ca-central-1. M5 instances offer up to 96 vCPUs, 384 GiB of DDR4 instance memory, 25 Gbps in Network bandwidth and improved EBS and Networking bandwidth on smaller instance sizes and provide a balance of compute, memory and network resources for many applications. +* `service/eks`: Adds new service +* `service/mediaconvert`: Updates service API and documentation + * This release adds the support for Common Media Application Format (CMAF) fragmented outputs, RF64 WAV audio output format, and HEV1 or HEVC1 MP4 packaging types when using HEVC in DASH or CMAF outputs. +* `service/sagemaker`: Updates service API, documentation, and paginators + * Amazon SageMaker has added the ability to run hyperparameter tuning jobs. A hyperparameter tuning job will create and evaluate multiple training jobs while tuning algorithm hyperparameters, to optimize a customer specified objective metric. + +### SDK Features +* Add support for EventStream based APIs (S3 SelectObjectContent) ([#1941](https://github.com/aws/aws-sdk-go/pull/1941)) + * Adds support for EventStream asynchronous APIs such as S3 SelectObjectContents API. This API allows your application to receiving multiple events asynchronously from the API response. Your application recieves these events from a channel on the API response. + * See PR [#1941](https://github.com/aws/aws-sdk-go/pull/1941) for example. + * Fixes [#1895](https://github.com/aws/aws-sdk-go/issues/1895) + +Release v1.13.60 (2018-06-01) +=== + +### Service Client Updates +* `service/ds`: Updates service API and documentation + * Added ResetUserPassword API. Customers can now reset their users' passwords without providing the old passwords in Simple AD and Microsoft AD. +* `aws/endpoints`: Updated Regions and Endpoints metadata. +* `service/iot`: Updates service API and documentation + * We are releasing force CancelJob and CancelJobExecution functionalities to customers. +* `service/mediatailor`: Adds new service +* `service/redshift`: Updates service documentation + * Documentation updates for redshift +* `service/sns`: Updates service API, documentation, and paginators + * The SNS Subscribe API has been updated with two new optional parameters: Attributes and ReturnSubscriptionArn. Attributes is a map of subscription attributes which can be one or more of: FilterPolicy, DeliveryPolicy, and RawMessageDelivery. ReturnSubscriptionArn is a boolean parameter that overrides the default behavior of returning "pending confirmation" for subscriptions that require confirmation instead of returning the subscription ARN. + +### SDK Bugs +* `private/mode/api`: Fix error code constants being generated incorrectly.([#1958](https://github.com/aws/aws-sdk-go/issues/1958)) + * Fixes the SDK's code generation to not modify the error code text value when generating error code constants. This prevents generating error code values which are invalid and will never be sent by the service. This change does not change the error code constant variable name generated by the SDK, only the value of the error code. + * Fixes [#1856](https://github.com/aws/aws-sdk-go/issues/1856) +Release v1.13.59 (2018-05-31) +=== + +* `aws/endpoints`: Updated Regions and Endpoints metadata. +Release v1.13.58 (2018-05-30) +=== + +### Service Client Updates +* `service/elasticloadbalancingv2`: Updates service API and documentation +* `aws/endpoints`: Updated Regions and Endpoints metadata. +* `service/neptune`: Adds new service + * Amazon Neptune is a fast, reliable graph database service that makes it easy to build and run applications that work with highly connected datasets. Neptune supports popular graph models Property Graph and W3C's Resource Description Frame (RDF), and their respective query languages Apache TinkerPop Gremlin 3.3.2 and SPARQL 1.1. + +Release v1.13.57 (2018-05-29) +=== + +### Service Client Updates +* `aws/endpoints`: Updated Regions and Endpoints metadata. +* `service/pi`: Adds new service + +Release v1.13.56 (2018-05-25) +=== + +### Service Client Updates +* `service/appstream`: Updates service API and documentation + * This API update enables customers to control whether users can transfer data between their local devices and their streaming applications through file uploads and downloads, clipboard operations, or printing to local devices +* `service/config`: Updates service API and documentation +* `service/glue`: Updates service API and documentation + * AWS Glue now sends a delay notification to Amazon CloudWatch Events when an ETL job runs longer than the specified delay notification threshold. +* `service/iot`: Updates service API + * We are exposing DELETION_IN_PROGRESS as a new job status in regards to the release of DeleteJob API. + +Release v1.13.55 (2018-05-24) +=== + +### Service Client Updates +* `service/codebuild`: Updates service API + * AWS CodeBuild Adds Support for Windows Builds. +* `service/elasticloadbalancingv2`: Updates service documentation +* `service/rds`: Updates service API and documentation + * This release adds CloudWatch Logs integration capabilities to RDS Aurora MySQL clusters +* `service/secretsmanager`: Updates service documentation + * Documentation updates for secretsmanager + +### SDK Bugs +* `service/cloudwatchlogs`: Fix pagination with cloudwatchlogs ([#1945](https://github.com/aws/aws-sdk-go/pull/1945)) + * Fixes the SDK's behavior with CloudWatchLogs APIs which return duplicate `NextToken` values to signal end of pagination. + * Fixes [#1908](https://github.com/aws/aws-sdk-go/pull/1908) + +Release v1.13.54 (2018-05-22) +=== + +### Service Client Updates +* `service/ecs`: Updates service API and documentation + * Amazon Elastic Container Service (ECS) adds service discovery for services that use host or bridged network mode. ECS can now also register instance IPs for active tasks using bridged and host networking with Route 53, making them available via DNS. +* `service/inspector`: Updates service API + * We are launching the ability to target all EC2 instances. With this launch, resourceGroupArn is now optional for CreateAssessmentTarget and UpdateAssessmentTarget. If resourceGroupArn is not specified, all EC2 instances in the account in the AWS region are included in the assessment target. + +Release v1.13.53 (2018-05-21) +=== + +### Service Client Updates +* `service/cloudformation`: Updates service API and documentation + * 1) Filtered Update for StackSet based on Accounts and Regions: This feature will allow flexibility for the customers to roll out updates on a StackSet based on specific Accounts and Regions. 2) Support for customized ExecutionRoleName: This feature will allow customers to attach ExecutionRoleName to the StackSet thus ensuring more security and controlling the behavior of any AWS resources in the target accounts. + +Release v1.13.52 (2018-05-18) +=== + +### Service Client Updates +* `service/email`: Updates service documentation + * Fixed a broken link in the documentation for S3Action. +* `service/iot`: Updates service API and documentation + * We are releasing DeleteJob and DeleteJobExecution APIs to allow customer to delete resources created using AWS IoT Jobs. + +Release v1.13.51 (2018-05-17) +=== + +### Service Client Updates +* `service/codedeploy`: Updates service documentation + * Documentation updates for codedeploy +* `service/cognito-idp`: Updates service API and documentation +* `service/ec2`: Updates service API and documentation + * You are now able to use instance storage (up to 1800 GB of NVMe based SSD) on C5 instances, the next generation of EC2's compute optimized instances in us-east-1, us-west-2, us-east-2, eu-west-1 and ca-central-1. C5 instances offer up to 72 vCPUs, 144 GiB of DDR4 instance memory, 25 Gbps in Network bandwidth and improved EBS and Networking bandwidth on smaller instance sizes to deliver improved performance for compute-intensive workloads.You can now run bare metal workloads on EC2 with i3.metal instances. As a new instance size belonging to the I3 instance family, i3.metal instances have the same characteristics as other instances in the family, including NVMe SSD-backed instance storage optimized for low latency, very high random I/O performance, and high sequential read throughput. I3.metal instances are powered by 2.3 GHz Intel Xeon processors, offering 36 hyper-threaded cores (72 logical processors), 512 GiB of memory, and 15.2 TB of NVMe SSD-backed instance storage. These instances deliver high networking throughput and lower latency with up to 25 Gbps of aggregate network bandwidth using Elastic Network Adapter (ENA)-based Enhanced Networking. + Release v1.13.50 (2018-05-16) === @@ -341,7 +644,7 @@ Release v1.13.19 (2018-03-22) ### SDK Bugs * `aws/endpoints`: Use service metadata for fallback signing name ([#1854](https://github.com/aws/aws-sdk-go/pull/1854)) - * Updates the SDK's endpoint resolution to fallback deriving the service's signing name from the service's modeled metadata in addition the the endpoints modeled data. + * Updates the SDK's endpoint resolution to fallback deriving the service's signing name from the service's modeled metadata in addition the endpoints modeled data. * Fixes [#1850](https://github.com/aws/aws-sdk-go/issues/1850) Release v1.13.18 (2018-03-21) === @@ -1078,7 +1381,7 @@ Release v1.12.31 (2017-11-20) * DescribeGroups API and miscellaneous enhancements ### SDK Bugs -* `aws/client`: Retry delays for throttled exception were not limited to 5 mintues [#1654](https://github.com/aws/aws-sdk-go/pull/1654) +* `aws/client`: Retry delays for throttled exception were not limited to 5 minutes [#1654](https://github.com/aws/aws-sdk-go/pull/1654) * Fixes [#1653](https://github.com/aws/aws-sdk-go/issues/1653) Release v1.12.30 (2017-11-17) === @@ -1210,7 +1513,7 @@ Release v1.12.21 (2017-11-02) * `aws/endpoints`: Updated Regions and Endpoints metadata. ### SDK Bugs -* `aws/request`: Fix bug in request presign creating invalide URL ([#1624](https://github.com/aws/aws-sdk-go/pull/1624)) +* `aws/request`: Fix bug in request presign creating invalid URL ([#1624](https://github.com/aws/aws-sdk-go/pull/1624)) * Fixes a bug the Request Presign and PresignRequest methods that would allow a invalid expire duration as input. A expire time of 0 would be interpreted by the SDK to generate a normal request signature, not a presigned URL. This caused the returned URL unusable. * Fixes [#1617](https://github.com/aws/aws-sdk-go/issues/1617) Release v1.12.20 (2017-11-01) diff --git a/vendor/github.com/aws/aws-sdk-go/CONTRIBUTING.md b/vendor/github.com/aws/aws-sdk-go/CONTRIBUTING.md index 6f422a95e..9f4a0c14c 100644 --- a/vendor/github.com/aws/aws-sdk-go/CONTRIBUTING.md +++ b/vendor/github.com/aws/aws-sdk-go/CONTRIBUTING.md @@ -89,7 +89,7 @@ go test -tags codegen ./private/... See the `Makefile` for additional testing tags that can be used in testing. To test on multiple platform the SDK includes several DockerFiles under the -`awstesting/sandbox` folder, and associated make recipes to to execute +`awstesting/sandbox` folder, and associated make recipes to execute unit testing within environments configured for specific Go versions. ``` diff --git a/vendor/github.com/aws/aws-sdk-go/Makefile b/vendor/github.com/aws/aws-sdk-go/Makefile index 83ccc1e62..89f4b7a82 100644 --- a/vendor/github.com/aws/aws-sdk-go/Makefile +++ b/vendor/github.com/aws/aws-sdk-go/Makefile @@ -13,6 +13,7 @@ SDK_ONLY_PKGS=$(shell go list ./... | grep -v "/vendor/") SDK_UNIT_TEST_ONLY_PKGS=$(shell go list -tags ${UNIT_TEST_TAGS} ./... | grep -v "/vendor/") SDK_GO_1_4=$(shell go version | grep "go1.4") SDK_GO_1_5=$(shell go version | grep "go1.5") +SDK_GO_1_6=$(shell go version | grep "go1.6") SDK_GO_VERSION=$(shell go version | awk '''{print $$3}''' | tr -d '''\n''') all: get-deps generate unit @@ -35,7 +36,7 @@ help: @echo " get-deps-tests to get the SDK's test dependencies" @echo " get-deps-verify to get the SDK's verification dependencies" -generate: gen-test gen-endpoints gen-services +generate: cleanup-models gen-test gen-endpoints gen-services gen-test: gen-protocol-test @@ -48,6 +49,10 @@ gen-protocol-test: gen-endpoints: go generate ./models/endpoints/ +cleanup-models: + @echo "Cleaning up stale model versions" + @./cleanup_models.sh + build: @echo "go build SDK and vendor packages" @go build ${SDK_ONLY_PKGS} @@ -60,6 +65,18 @@ unit-with-race-cover: get-deps-tests build verify @echo "go test SDK and vendor packages" @go test -tags ${UNIT_TEST_TAGS} -race -cpu=1,2,4 $(SDK_UNIT_TEST_ONLY_PKGS) +ci-test: ci-test-generate unit-with-race-cover ci-test-generate-validate + +ci-test-generate: get-deps + @echo "CI test generated code" + @if [ \( -z "${SDK_GO_1_6}" \) -a \( -z "${SDK_GO_1_5}" \) ]; then make generate; else echo "skipping generate"; fi + +ci-test-generate-validate: + @echo "CI test validate no generated code changes" + @gitstatus=`if [ \( -z "${SDK_GO_1_6}" \) -a \( -z "${SDK_GO_1_5}" \) ]; then git status --porcelain; else echo "skipping validation"; fi`; \ + echo "$$gitstatus"; \ + if [ "$$gitstatus" != "" ] && [ "$$gitstatus" != "skipping validation" ]; then git diff; exit 1; fi + integration: get-deps-tests integ-custom smoke-tests performance integ-custom: @@ -118,6 +135,13 @@ sandbox-go19: sandbox-build-go19 sandbox-test-go19: sandbox-build-go19 docker run -t aws-sdk-go-1.9 +sandbox-build-go110: + docker build -f ./awstesting/sandbox/Dockerfile.test.go1.8 -t "aws-sdk-go-1.10" . +sandbox-go110: sandbox-build-go110 + docker run -i -t aws-sdk-go-1.10 bash +sandbox-test-go110: sandbox-build-go110 + docker run -t aws-sdk-go-1.10 + sandbox-build-gotip: @echo "Run make update-aws-golang-tip, if this test fails because missing aws-golang:tip container" docker build -f ./awstesting/sandbox/Dockerfile.test.gotip -t "aws-sdk-go-tip" . diff --git a/vendor/github.com/aws/aws-sdk-go/README.md b/vendor/github.com/aws/aws-sdk-go/README.md index 9f10e76f3..3200765ea 100644 --- a/vendor/github.com/aws/aws-sdk-go/README.md +++ b/vendor/github.com/aws/aws-sdk-go/README.md @@ -334,7 +334,7 @@ take a callback function that will be called for each page of the API's response ``` Waiter helper methods provide the functionality to wait for an AWS resource -state. These methods abstract the logic needed to to check the state of an +state. These methods abstract the logic needed to check the state of an AWS resource, and wait until that resource is in a desired state. The waiter will block until the resource is in the state that is desired, an error occurs, or the waiter times out. If a resource times out the error code returned will diff --git a/vendor/github.com/aws/aws-sdk-go/aws/client/client.go b/vendor/github.com/aws/aws-sdk-go/aws/client/client.go index 3271a18e8..212fe25e7 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/client/client.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/client/client.go @@ -91,6 +91,6 @@ func (c *Client) AddDebugHandlers() { return } - c.Handlers.Send.PushFrontNamed(request.NamedHandler{Name: "awssdk.client.LogRequest", Fn: logRequest}) - c.Handlers.Send.PushBackNamed(request.NamedHandler{Name: "awssdk.client.LogResponse", Fn: logResponse}) + c.Handlers.Send.PushFrontNamed(LogHTTPRequestHandler) + c.Handlers.Send.PushBackNamed(LogHTTPResponseHandler) } diff --git a/vendor/github.com/aws/aws-sdk-go/aws/client/logger.go b/vendor/github.com/aws/aws-sdk-go/aws/client/logger.go index e223c54cc..ce9fb896d 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/client/logger.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/client/logger.go @@ -44,12 +44,22 @@ func (reader *teeReaderCloser) Close() error { return reader.Source.Close() } +// LogHTTPRequestHandler is a SDK request handler to log the HTTP request sent +// to a service. Will include the HTTP request body if the LogLevel of the +// request matches LogDebugWithHTTPBody. +var LogHTTPRequestHandler = request.NamedHandler{ + Name: "awssdk.client.LogRequest", + Fn: logRequest, +} + func logRequest(r *request.Request) { logBody := r.Config.LogLevel.Matches(aws.LogDebugWithHTTPBody) bodySeekable := aws.IsReaderSeekable(r.Body) - dumpedBody, err := httputil.DumpRequestOut(r.HTTPRequest, logBody) + + b, err := httputil.DumpRequestOut(r.HTTPRequest, logBody) if err != nil { - r.Config.Logger.Log(fmt.Sprintf(logReqErrMsg, r.ClientInfo.ServiceName, r.Operation.Name, err)) + r.Config.Logger.Log(fmt.Sprintf(logReqErrMsg, + r.ClientInfo.ServiceName, r.Operation.Name, err)) return } @@ -63,7 +73,28 @@ func logRequest(r *request.Request) { r.ResetBody() } - r.Config.Logger.Log(fmt.Sprintf(logReqMsg, r.ClientInfo.ServiceName, r.Operation.Name, string(dumpedBody))) + r.Config.Logger.Log(fmt.Sprintf(logReqMsg, + r.ClientInfo.ServiceName, r.Operation.Name, string(b))) +} + +// LogHTTPRequestHeaderHandler is a SDK request handler to log the HTTP request sent +// to a service. Will only log the HTTP request's headers. The request payload +// will not be read. +var LogHTTPRequestHeaderHandler = request.NamedHandler{ + Name: "awssdk.client.LogRequestHeader", + Fn: logRequestHeader, +} + +func logRequestHeader(r *request.Request) { + b, err := httputil.DumpRequestOut(r.HTTPRequest, false) + if err != nil { + r.Config.Logger.Log(fmt.Sprintf(logReqErrMsg, + r.ClientInfo.ServiceName, r.Operation.Name, err)) + return + } + + r.Config.Logger.Log(fmt.Sprintf(logReqMsg, + r.ClientInfo.ServiceName, r.Operation.Name, string(b))) } const logRespMsg = `DEBUG: Response %s/%s Details: @@ -76,27 +107,44 @@ const logRespErrMsg = `DEBUG ERROR: Response %s/%s: %s -----------------------------------------------------` +// LogHTTPResponseHandler is a SDK request handler to log the HTTP response +// received from a service. Will include the HTTP response body if the LogLevel +// of the request matches LogDebugWithHTTPBody. +var LogHTTPResponseHandler = request.NamedHandler{ + Name: "awssdk.client.LogResponse", + Fn: logResponse, +} + func logResponse(r *request.Request) { lw := &logWriter{r.Config.Logger, bytes.NewBuffer(nil)} - r.HTTPResponse.Body = &teeReaderCloser{ - Reader: io.TeeReader(r.HTTPResponse.Body, lw), - Source: r.HTTPResponse.Body, + + logBody := r.Config.LogLevel.Matches(aws.LogDebugWithHTTPBody) + if logBody { + r.HTTPResponse.Body = &teeReaderCloser{ + Reader: io.TeeReader(r.HTTPResponse.Body, lw), + Source: r.HTTPResponse.Body, + } } handlerFn := func(req *request.Request) { - body, err := httputil.DumpResponse(req.HTTPResponse, false) + b, err := httputil.DumpResponse(req.HTTPResponse, false) if err != nil { - lw.Logger.Log(fmt.Sprintf(logRespErrMsg, req.ClientInfo.ServiceName, req.Operation.Name, err)) + lw.Logger.Log(fmt.Sprintf(logRespErrMsg, + req.ClientInfo.ServiceName, req.Operation.Name, err)) return } - b, err := ioutil.ReadAll(lw.buf) - if err != nil { - lw.Logger.Log(fmt.Sprintf(logRespErrMsg, req.ClientInfo.ServiceName, req.Operation.Name, err)) - return - } - lw.Logger.Log(fmt.Sprintf(logRespMsg, req.ClientInfo.ServiceName, req.Operation.Name, string(body))) - if req.Config.LogLevel.Matches(aws.LogDebugWithHTTPBody) { + lw.Logger.Log(fmt.Sprintf(logRespMsg, + req.ClientInfo.ServiceName, req.Operation.Name, string(b))) + + if logBody { + b, err := ioutil.ReadAll(lw.buf) + if err != nil { + lw.Logger.Log(fmt.Sprintf(logRespErrMsg, + req.ClientInfo.ServiceName, req.Operation.Name, err)) + return + } + lw.Logger.Log(string(b)) } } @@ -110,3 +158,27 @@ func logResponse(r *request.Request) { Name: handlerName, Fn: handlerFn, }) } + +// LogHTTPResponseHeaderHandler is a SDK request handler to log the HTTP +// response received from a service. Will only log the HTTP response's headers. +// The response payload will not be read. +var LogHTTPResponseHeaderHandler = request.NamedHandler{ + Name: "awssdk.client.LogResponseHeader", + Fn: logResponseHeader, +} + +func logResponseHeader(r *request.Request) { + if r.Config.Logger == nil { + return + } + + b, err := httputil.DumpResponse(r.HTTPResponse, false) + if err != nil { + r.Config.Logger.Log(fmt.Sprintf(logRespErrMsg, + r.ClientInfo.ServiceName, r.Operation.Name, err)) + return + } + + r.Config.Logger.Log(fmt.Sprintf(logRespMsg, + r.ClientInfo.ServiceName, r.Operation.Name, string(b))) +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/client/logger_test.go b/vendor/github.com/aws/aws-sdk-go/aws/client/logger_test.go index f0e06b805..ee3630e6c 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/client/logger_test.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/client/logger_test.go @@ -5,6 +5,7 @@ import ( "fmt" "io" "io/ioutil" + "net/http" "reflect" "testing" @@ -127,6 +128,83 @@ func TestLogRequest(t *testing.T) { } } +func TestLogResponse(t *testing.T) { + cases := []struct { + Body *bytes.Buffer + ExpectBody []byte + ReadBody bool + LogLevel aws.LogLevelType + }{ + { + Body: bytes.NewBuffer([]byte("body content")), + ExpectBody: []byte("body content"), + }, + { + Body: bytes.NewBuffer([]byte("body content")), + LogLevel: aws.LogDebug, + ExpectBody: []byte("body content"), + }, + { + Body: bytes.NewBuffer([]byte("body content")), + LogLevel: aws.LogDebugWithHTTPBody, + ReadBody: true, + ExpectBody: []byte("body content"), + }, + } + + for i, c := range cases { + var logW bytes.Buffer + req := request.New( + aws.Config{ + Credentials: credentials.AnonymousCredentials, + Logger: &bufLogger{w: &logW}, + LogLevel: aws.LogLevel(c.LogLevel), + }, + metadata.ClientInfo{ + Endpoint: "https://mock-service.mock-region.amazonaws.com", + }, + testHandlers(), + nil, + &request.Operation{ + Name: "APIName", + HTTPMethod: "POST", + HTTPPath: "/", + }, + struct{}{}, nil, + ) + req.HTTPResponse = &http.Response{ + StatusCode: 200, + Status: "OK", + Header: http.Header{ + "ABC": []string{"123"}, + }, + Body: ioutil.NopCloser(c.Body), + } + + logResponse(req) + req.Handlers.Unmarshal.Run(req) + + if c.ReadBody { + if e, a := len(c.ExpectBody), c.Body.Len(); e != a { + t.Errorf("%d, expect orginal body not to of been read", i) + } + } + + if logW.Len() == 0 { + t.Errorf("%d, expect HTTP Response headers to be logged", i) + } + + b, err := ioutil.ReadAll(req.HTTPResponse.Body) + if err != nil { + t.Fatalf("%d, expect to read SDK request Body", i) + } + + if e, a := c.ExpectBody, b; !bytes.Equal(e, a) { + t.Errorf("%d, expect %v body, got %v", i, e, a) + } + } +} + type bufLogger struct { w *bytes.Buffer } diff --git a/vendor/github.com/aws/aws-sdk-go/aws/client/metadata/client_info.go b/vendor/github.com/aws/aws-sdk-go/aws/client/metadata/client_info.go index 4778056dd..920e9fddf 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/client/metadata/client_info.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/client/metadata/client_info.go @@ -3,6 +3,7 @@ package metadata // ClientInfo wraps immutable data from the client.Client structure. type ClientInfo struct { ServiceName string + ServiceID string APIVersion string Endpoint string SigningName string diff --git a/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials.go b/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials.go index 42416fc2f..ed086992f 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials.go @@ -178,7 +178,8 @@ func (e *Expiry) IsExpired() bool { type Credentials struct { creds Value forceRefresh bool - m sync.Mutex + + m sync.RWMutex provider Provider } @@ -201,6 +202,17 @@ func NewCredentials(provider Provider) *Credentials { // If Credentials.Expire() was called the credentials Value will be force // expired, and the next call to Get() will cause them to be refreshed. func (c *Credentials) Get() (Value, error) { + // Check the cached credentials first with just the read lock. + c.m.RLock() + if !c.isExpired() { + creds := c.creds + c.m.RUnlock() + return creds, nil + } + c.m.RUnlock() + + // Credentials are expired need to retrieve the credentials taking the full + // lock. c.m.Lock() defer c.m.Unlock() @@ -234,8 +246,8 @@ func (c *Credentials) Expire() { // If the Credentials were forced to be expired with Expire() this will // reflect that override. func (c *Credentials) IsExpired() bool { - c.m.Lock() - defer c.m.Unlock() + c.m.RLock() + defer c.m.RUnlock() return c.isExpired() } diff --git a/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials_bench_test.go b/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials_bench_test.go new file mode 100644 index 000000000..01a5d633b --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/credentials/credentials_bench_test.go @@ -0,0 +1,90 @@ +// +build go1.9 + +package credentials + +import ( + "fmt" + "strconv" + "sync" + "testing" + "time" +) + +func BenchmarkCredentials_Get(b *testing.B) { + stub := &stubProvider{} + + cases := []int{1, 10, 100, 500, 1000, 10000} + + for _, c := range cases { + b.Run(strconv.Itoa(c), func(b *testing.B) { + creds := NewCredentials(stub) + var wg sync.WaitGroup + wg.Add(c) + for i := 0; i < c; i++ { + go func() { + for j := 0; j < b.N; j++ { + v, err := creds.Get() + if err != nil { + b.Fatalf("expect no error %v, %v", v, err) + } + } + wg.Done() + }() + } + b.ResetTimer() + + wg.Wait() + }) + } +} + +func BenchmarkCredentials_Get_Expire(b *testing.B) { + p := &blockProvider{} + + expRates := []int{10000, 1000, 100} + cases := []int{1, 10, 100, 500, 1000, 10000} + + for _, expRate := range expRates { + for _, c := range cases { + b.Run(fmt.Sprintf("%d-%d", expRate, c), func(b *testing.B) { + creds := NewCredentials(p) + var wg sync.WaitGroup + wg.Add(c) + for i := 0; i < c; i++ { + go func(id int) { + for j := 0; j < b.N; j++ { + v, err := creds.Get() + if err != nil { + b.Fatalf("expect no error %v, %v", v, err) + } + // periodically expire creds to cause rwlock + if id == 0 && j%expRate == 0 { + creds.Expire() + } + } + wg.Done() + }(i) + } + b.ResetTimer() + + wg.Wait() + }) + } + } +} + +type blockProvider struct { + creds Value + expired bool + err error +} + +func (s *blockProvider) Retrieve() (Value, error) { + s.expired = false + s.creds.ProviderName = "blockProvider" + time.Sleep(time.Millisecond) + return s.creds, s.err +} +func (s *blockProvider) IsExpired() bool { + return s.expired +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/doc.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/doc.go new file mode 100644 index 000000000..152d785b3 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/doc.go @@ -0,0 +1,46 @@ +// Package csm provides Client Side Monitoring (CSM) which enables sending metrics +// via UDP connection. Using the Start function will enable the reporting of +// metrics on a given port. If Start is called, with different parameters, again, +// a panic will occur. +// +// Pause can be called to pause any metrics publishing on a given port. Sessions +// that have had their handlers modified via InjectHandlers may still be used. +// However, the handlers will act as a no-op meaning no metrics will be published. +// +// Example: +// r, err := csm.Start("clientID", ":31000") +// if err != nil { +// panic(fmt.Errorf("failed starting CSM: %v", err)) +// } +// +// sess, err := session.NewSession(&aws.Config{}) +// if err != nil { +// panic(fmt.Errorf("failed loading session: %v", err)) +// } +// +// r.InjectHandlers(&sess.Handlers) +// +// client := s3.New(sess) +// resp, err := client.GetObject(&s3.GetObjectInput{ +// Bucket: aws.String("bucket"), +// Key: aws.String("key"), +// }) +// +// // Will pause monitoring +// r.Pause() +// resp, err = client.GetObject(&s3.GetObjectInput{ +// Bucket: aws.String("bucket"), +// Key: aws.String("key"), +// }) +// +// // Resume monitoring +// r.Continue() +// +// Start returns a Reporter that is used to enable or disable monitoring. If +// access to the Reporter is required later, calling Get will return the Reporter +// singleton. +// +// Example: +// r := csm.Get() +// r.Continue() +package csm diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/enable.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/enable.go new file mode 100644 index 000000000..2f0c6eac9 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/enable.go @@ -0,0 +1,67 @@ +package csm + +import ( + "fmt" + "sync" +) + +var ( + lock sync.Mutex +) + +// Client side metric handler names +const ( + APICallMetricHandlerName = "awscsm.SendAPICallMetric" + APICallAttemptMetricHandlerName = "awscsm.SendAPICallAttemptMetric" +) + +// Start will start the a long running go routine to capture +// client side metrics. Calling start multiple time will only +// start the metric listener once and will panic if a different +// client ID or port is passed in. +// +// Example: +// r, err := csm.Start("clientID", "127.0.0.1:8094") +// if err != nil { +// panic(fmt.Errorf("expected no error, but received %v", err)) +// } +// sess := session.NewSession() +// r.InjectHandlers(sess.Handlers) +// +// svc := s3.New(sess) +// out, err := svc.GetObject(&s3.GetObjectInput{ +// Bucket: aws.String("bucket"), +// Key: aws.String("key"), +// }) +func Start(clientID string, url string) (*Reporter, error) { + lock.Lock() + defer lock.Unlock() + + if sender == nil { + sender = newReporter(clientID, url) + } else { + if sender.clientID != clientID { + panic(fmt.Errorf("inconsistent client IDs. %q was expected, but received %q", sender.clientID, clientID)) + } + + if sender.url != url { + panic(fmt.Errorf("inconsistent URLs. %q was expected, but received %q", sender.url, url)) + } + } + + if err := connect(url); err != nil { + sender = nil + return nil, err + } + + return sender, nil +} + +// Get will return a reporter if one exists, if one does not exist, nil will +// be returned. +func Get() *Reporter { + lock.Lock() + defer lock.Unlock() + + return sender +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/enable_test.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/enable_test.go new file mode 100644 index 000000000..a7504d3f9 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/enable_test.go @@ -0,0 +1,74 @@ +package csm + +import ( + "encoding/json" + "fmt" + "net" + "testing" +) + +func startUDPServer(done chan struct{}, fn func([]byte)) (string, error) { + addr, err := net.ResolveUDPAddr("udp", "127.0.0.1:0") + if err != nil { + return "", err + } + + conn, err := net.ListenUDP("udp", addr) + if err != nil { + return "", err + } + + buf := make([]byte, 1024) + go func() { + defer conn.Close() + + for { + select { + case <-done: + return + default: + } + + n, _, err := conn.ReadFromUDP(buf) + fn(buf[:n]) + + if err != nil { + panic(err) + } + } + }() + + return conn.LocalAddr().String(), nil +} + +func TestDifferentParams(t *testing.T) { + defer func() { + if r := recover(); r == nil { + t.Errorf("expected panic with different parameters") + } + }() + Start("clientID2", ":0") +} + +var MetricsCh = make(chan map[string]interface{}, 1) +var Done = make(chan struct{}) + +func init() { + url, err := startUDPServer(Done, func(b []byte) { + m := map[string]interface{}{} + if err := json.Unmarshal(b, &m); err != nil { + panic(fmt.Sprintf("expected no error, but received %v", err)) + } + + MetricsCh <- m + }) + + if err != nil { + panic(err) + } + + _, err = Start("clientID", url) + if err != nil { + panic(err) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/example_test.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/example_test.go new file mode 100644 index 000000000..3da43dccd --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/example_test.go @@ -0,0 +1,40 @@ +package csm_test + +import ( + "fmt" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/csm" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/service/s3" +) + +func ExampleStart() { + r, err := csm.Start("clientID", ":31000") + if err != nil { + panic(fmt.Errorf("failed starting CSM: %v", err)) + } + + sess, err := session.NewSession(&aws.Config{}) + if err != nil { + panic(fmt.Errorf("failed loading session: %v", err)) + } + + r.InjectHandlers(&sess.Handlers) + + client := s3.New(sess) + client.GetObject(&s3.GetObjectInput{ + Bucket: aws.String("bucket"), + Key: aws.String("key"), + }) + + // Pauses monitoring + r.Pause() + client.GetObject(&s3.GetObjectInput{ + Bucket: aws.String("bucket"), + Key: aws.String("key"), + }) + + // Resume monitoring + r.Continue() +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/metric.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/metric.go new file mode 100644 index 000000000..4b0d630e4 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/metric.go @@ -0,0 +1,51 @@ +package csm + +import ( + "strconv" + "time" +) + +type metricTime time.Time + +func (t metricTime) MarshalJSON() ([]byte, error) { + ns := time.Duration(time.Time(t).UnixNano()) + return []byte(strconv.FormatInt(int64(ns/time.Millisecond), 10)), nil +} + +type metric struct { + ClientID *string `json:"ClientId,omitempty"` + API *string `json:"Api,omitempty"` + Service *string `json:"Service,omitempty"` + Timestamp *metricTime `json:"Timestamp,omitempty"` + Type *string `json:"Type,omitempty"` + Version *int `json:"Version,omitempty"` + + AttemptCount *int `json:"AttemptCount,omitempty"` + Latency *int `json:"Latency,omitempty"` + + Fqdn *string `json:"Fqdn,omitempty"` + UserAgent *string `json:"UserAgent,omitempty"` + AttemptLatency *int `json:"AttemptLatency,omitempty"` + + SessionToken *string `json:"SessionToken,omitempty"` + Region *string `json:"Region,omitempty"` + AccessKey *string `json:"AccessKey,omitempty"` + HTTPStatusCode *int `json:"HttpStatusCode,omitempty"` + XAmzID2 *string `json:"XAmzId2,omitempty"` + XAmzRequestID *string `json:"XAmznRequestId,omitempty"` + + AWSException *string `json:"AwsException,omitempty"` + AWSExceptionMessage *string `json:"AwsExceptionMessage,omitempty"` + SDKException *string `json:"SdkException,omitempty"` + SDKExceptionMessage *string `json:"SdkExceptionMessage,omitempty"` + + DestinationIP *string `json:"DestinationIp,omitempty"` + ConnectionReused *int `json:"ConnectionReused,omitempty"` + + AcquireConnectionLatency *int `json:"AcquireConnectionLatency,omitempty"` + ConnectLatency *int `json:"ConnectLatency,omitempty"` + RequestLatency *int `json:"RequestLatency,omitempty"` + DNSLatency *int `json:"DnsLatency,omitempty"` + TCPLatency *int `json:"TcpLatency,omitempty"` + SSLLatency *int `json:"SslLatency,omitempty"` +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/metric_chan.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/metric_chan.go new file mode 100644 index 000000000..514fc3739 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/metric_chan.go @@ -0,0 +1,54 @@ +package csm + +import ( + "sync/atomic" +) + +const ( + runningEnum = iota + pausedEnum +) + +var ( + // MetricsChannelSize of metrics to hold in the channel + MetricsChannelSize = 100 +) + +type metricChan struct { + ch chan metric + paused int64 +} + +func newMetricChan(size int) metricChan { + return metricChan{ + ch: make(chan metric, size), + } +} + +func (ch *metricChan) Pause() { + atomic.StoreInt64(&ch.paused, pausedEnum) +} + +func (ch *metricChan) Continue() { + atomic.StoreInt64(&ch.paused, runningEnum) +} + +func (ch *metricChan) IsPaused() bool { + v := atomic.LoadInt64(&ch.paused) + return v == pausedEnum +} + +// Push will push metrics to the metric channel if the channel +// is not paused +func (ch *metricChan) Push(m metric) bool { + if ch.IsPaused() { + return false + } + + select { + case ch.ch <- m: + return true + default: + return false + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/metric_chan_test.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/metric_chan_test.go new file mode 100644 index 000000000..81bdb25c8 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/metric_chan_test.go @@ -0,0 +1,72 @@ +package csm + +import ( + "testing" +) + +func TestMetricChanPush(t *testing.T) { + ch := newMetricChan(5) + defer close(ch.ch) + + pushed := ch.Push(metric{}) + if !pushed { + t.Errorf("expected metrics to be pushed") + } + + if e, a := 1, len(ch.ch); e != a { + t.Errorf("expected %d, but received %d", e, a) + } +} + +func TestMetricChanPauseContinue(t *testing.T) { + ch := newMetricChan(5) + defer close(ch.ch) + ch.Pause() + + if !ch.IsPaused() { + t.Errorf("expected to be paused, but did not pause properly") + } + + ch.Continue() + if ch.IsPaused() { + t.Errorf("expected to be not paused, but did not continue properly") + } + + pushed := ch.Push(metric{}) + if !pushed { + t.Errorf("expected metrics to be pushed") + } + + if e, a := 1, len(ch.ch); e != a { + t.Errorf("expected %d, but received %d", e, a) + } +} + +func TestMetricChanPushWhenPaused(t *testing.T) { + ch := newMetricChan(5) + defer close(ch.ch) + ch.Pause() + + pushed := ch.Push(metric{}) + if pushed { + t.Errorf("expected metrics to not be pushed") + } + + if e, a := 0, len(ch.ch); e != a { + t.Errorf("expected %d, but received %d", e, a) + } +} + +func TestMetricChanNonBlocking(t *testing.T) { + ch := newMetricChan(0) + defer close(ch.ch) + + pushed := ch.Push(metric{}) + if pushed { + t.Errorf("expected metrics to be not pushed") + } + + if e, a := 0, len(ch.ch); e != a { + t.Errorf("expected %d, but received %d", e, a) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/reporter.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/reporter.go new file mode 100644 index 000000000..d2481c18d --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/reporter.go @@ -0,0 +1,232 @@ +package csm + +import ( + "encoding/json" + "net" + "time" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awserr" + "github.com/aws/aws-sdk-go/aws/request" +) + +const ( + // DefaultPort is used when no port is specified + DefaultPort = "31000" +) + +// Reporter will gather metrics of API requests made and +// send those metrics to the CSM endpoint. +type Reporter struct { + clientID string + url string + conn net.Conn + metricsCh metricChan + done chan struct{} +} + +var ( + sender *Reporter +) + +func connect(url string) error { + const network = "udp" + if err := sender.connect(network, url); err != nil { + return err + } + + if sender.done == nil { + sender.done = make(chan struct{}) + go sender.start() + } + + return nil +} + +func newReporter(clientID, url string) *Reporter { + return &Reporter{ + clientID: clientID, + url: url, + metricsCh: newMetricChan(MetricsChannelSize), + } +} + +func (rep *Reporter) sendAPICallAttemptMetric(r *request.Request) { + if rep == nil { + return + } + + now := time.Now() + creds, _ := r.Config.Credentials.Get() + + m := metric{ + ClientID: aws.String(rep.clientID), + API: aws.String(r.Operation.Name), + Service: aws.String(r.ClientInfo.ServiceID), + Timestamp: (*metricTime)(&now), + UserAgent: aws.String(r.HTTPRequest.Header.Get("User-Agent")), + Region: r.Config.Region, + Type: aws.String("ApiCallAttempt"), + Version: aws.Int(1), + + XAmzRequestID: aws.String(r.RequestID), + + AttemptCount: aws.Int(r.RetryCount + 1), + AttemptLatency: aws.Int(int(now.Sub(r.AttemptTime).Nanoseconds() / int64(time.Millisecond))), + AccessKey: aws.String(creds.AccessKeyID), + } + + if r.HTTPResponse != nil { + m.HTTPStatusCode = aws.Int(r.HTTPResponse.StatusCode) + } + + if r.Error != nil { + if awserr, ok := r.Error.(awserr.Error); ok { + setError(&m, awserr) + } + } + + rep.metricsCh.Push(m) +} + +func setError(m *metric, err awserr.Error) { + msg := err.Message() + code := err.Code() + + switch code { + case "RequestError", + "SerializationError", + request.CanceledErrorCode: + + m.SDKException = &code + m.SDKExceptionMessage = &msg + default: + m.AWSException = &code + m.AWSExceptionMessage = &msg + } +} + +func (rep *Reporter) sendAPICallMetric(r *request.Request) { + if rep == nil { + return + } + + now := time.Now() + m := metric{ + ClientID: aws.String(rep.clientID), + API: aws.String(r.Operation.Name), + Service: aws.String(r.ClientInfo.ServiceID), + Timestamp: (*metricTime)(&now), + Type: aws.String("ApiCall"), + AttemptCount: aws.Int(r.RetryCount + 1), + Latency: aws.Int(int(time.Now().Sub(r.Time) / time.Millisecond)), + XAmzRequestID: aws.String(r.RequestID), + } + + // TODO: Probably want to figure something out for logging dropped + // metrics + rep.metricsCh.Push(m) +} + +func (rep *Reporter) connect(network, url string) error { + if rep.conn != nil { + rep.conn.Close() + } + + conn, err := net.Dial(network, url) + if err != nil { + return awserr.New("UDPError", "Could not connect", err) + } + + rep.conn = conn + + return nil +} + +func (rep *Reporter) close() { + if rep.done != nil { + close(rep.done) + } + + rep.metricsCh.Pause() +} + +func (rep *Reporter) start() { + defer func() { + rep.metricsCh.Pause() + }() + + for { + select { + case <-rep.done: + rep.done = nil + return + case m := <-rep.metricsCh.ch: + // TODO: What to do with this error? Probably should just log + b, err := json.Marshal(m) + if err != nil { + continue + } + + rep.conn.Write(b) + } + } +} + +// Pause will pause the metric channel preventing any new metrics from +// being added. +func (rep *Reporter) Pause() { + lock.Lock() + defer lock.Unlock() + + if rep == nil { + return + } + + rep.close() +} + +// Continue will reopen the metric channel and allow for monitoring +// to be resumed. +func (rep *Reporter) Continue() { + lock.Lock() + defer lock.Unlock() + if rep == nil { + return + } + + if !rep.metricsCh.IsPaused() { + return + } + + rep.metricsCh.Continue() +} + +// InjectHandlers will will enable client side metrics and inject the proper +// handlers to handle how metrics are sent. +// +// Example: +// // Start must be called in order to inject the correct handlers +// r, err := csm.Start("clientID", "127.0.0.1:8094") +// if err != nil { +// panic(fmt.Errorf("expected no error, but received %v", err)) +// } +// +// sess := session.NewSession() +// r.InjectHandlers(&sess.Handlers) +// +// // create a new service client with our client side metric session +// svc := s3.New(sess) +func (rep *Reporter) InjectHandlers(handlers *request.Handlers) { + if rep == nil { + return + } + + apiCallHandler := request.NamedHandler{Name: APICallMetricHandlerName, Fn: rep.sendAPICallMetric} + apiCallAttemptHandler := request.NamedHandler{Name: APICallAttemptMetricHandlerName, Fn: rep.sendAPICallAttemptMetric} + + handlers.Complete.PushFrontNamed(apiCallHandler) + handlers.Complete.PushFrontNamed(apiCallAttemptHandler) + + handlers.AfterRetry.PushFrontNamed(apiCallAttemptHandler) +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/csm/reporter_test.go b/vendor/github.com/aws/aws-sdk-go/aws/csm/reporter_test.go new file mode 100644 index 000000000..4a1001983 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/aws/csm/reporter_test.go @@ -0,0 +1,249 @@ +package csm_test + +import ( + "fmt" + "net" + "net/http" + "net/http/httptest" + "strings" + "testing" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/csm" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/aws/signer/v4" + "github.com/aws/aws-sdk-go/private/protocol/jsonrpc" +) + +func startUDPServer(done chan struct{}, fn func([]byte)) (string, error) { + addr, err := net.ResolveUDPAddr("udp", "127.0.0.1:0") + if err != nil { + return "", err + } + + conn, err := net.ListenUDP("udp", addr) + if err != nil { + return "", err + } + + buf := make([]byte, 1024) + i := 0 + go func() { + defer conn.Close() + for { + i++ + select { + case <-done: + return + default: + } + + n, _, err := conn.ReadFromUDP(buf) + fn(buf[:n]) + + if err != nil { + panic(err) + } + } + }() + + return conn.LocalAddr().String(), nil +} + +func TestReportingMetrics(t *testing.T) { + reporter := csm.Get() + if reporter == nil { + t.Errorf("expected non-nil reporter") + } + + sess := session.New() + sess.Handlers.Clear() + reporter.InjectHandlers(&sess.Handlers) + + md := metadata.ClientInfo{} + op := &request.Operation{} + r := request.New(*sess.Config, md, sess.Handlers, client.DefaultRetryer{NumMaxRetries: 0}, op, nil, nil) + sess.Handlers.Complete.Run(r) + + foundAttempt := false + foundCall := false + + expectedMetrics := 2 + + for i := 0; i < expectedMetrics; i++ { + m := <-csm.MetricsCh + for k, v := range m { + switch k { + case "Type": + a := v.(string) + foundCall = foundCall || a == "ApiCall" + foundAttempt = foundAttempt || a == "ApiCallAttempt" + + if prefix := "ApiCall"; !strings.HasPrefix(a, prefix) { + t.Errorf("expected 'APICall' prefix, but received %q", a) + } + } + } + } + + if !foundAttempt { + t.Errorf("expected attempt event to have occurred") + } + + if !foundCall { + t.Errorf("expected call event to have occurred") + } +} + +type mockService struct { + *client.Client +} + +type input struct{} +type output struct{} + +func (s *mockService) Request(i input) *request.Request { + op := &request.Operation{ + Name: "foo", + HTTPMethod: "POST", + HTTPPath: "/", + } + + o := output{} + req := s.NewRequest(op, &i, &o) + return req +} + +func BenchmarkWithCSM(b *testing.B) { + server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.Write([]byte(fmt.Sprintf("{}"))) + })) + + cfg := aws.Config{ + Endpoint: aws.String(server.URL), + } + + sess := session.New(&cfg) + r := csm.Get() + + r.InjectHandlers(&sess.Handlers) + + c := sess.ClientConfig("id", &cfg) + + svc := mockService{ + client.New( + *c.Config, + metadata.ClientInfo{ + ServiceName: "service", + ServiceID: "id", + SigningName: "signing", + SigningRegion: "region", + Endpoint: server.URL, + APIVersion: "0", + JSONVersion: "1.1", + TargetPrefix: "prefix", + }, + c.Handlers, + ), + } + + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(jsonrpc.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(jsonrpc.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(jsonrpc.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(jsonrpc.UnmarshalErrorHandler) + + for i := 0; i < b.N; i++ { + req := svc.Request(input{}) + req.Send() + } +} + +func BenchmarkWithCSMNoUDPConnection(b *testing.B) { + server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.Write([]byte(fmt.Sprintf("{}"))) + })) + + cfg := aws.Config{ + Endpoint: aws.String(server.URL), + } + + sess := session.New(&cfg) + r := csm.Get() + r.Pause() + r.InjectHandlers(&sess.Handlers) + defer r.Pause() + + c := sess.ClientConfig("id", &cfg) + + svc := mockService{ + client.New( + *c.Config, + metadata.ClientInfo{ + ServiceName: "service", + ServiceID: "id", + SigningName: "signing", + SigningRegion: "region", + Endpoint: server.URL, + APIVersion: "0", + JSONVersion: "1.1", + TargetPrefix: "prefix", + }, + c.Handlers, + ), + } + + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(jsonrpc.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(jsonrpc.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(jsonrpc.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(jsonrpc.UnmarshalErrorHandler) + + for i := 0; i < b.N; i++ { + req := svc.Request(input{}) + req.Send() + } +} + +func BenchmarkWithoutCSM(b *testing.B) { + server := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.Write([]byte(fmt.Sprintf("{}"))) + })) + + cfg := aws.Config{ + Endpoint: aws.String(server.URL), + } + sess := session.New(&cfg) + c := sess.ClientConfig("id", &cfg) + + svc := mockService{ + client.New( + *c.Config, + metadata.ClientInfo{ + ServiceName: "service", + ServiceID: "id", + SigningName: "signing", + SigningRegion: "region", + Endpoint: server.URL, + APIVersion: "0", + JSONVersion: "1.1", + TargetPrefix: "prefix", + }, + c.Handlers, + ), + } + + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(jsonrpc.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(jsonrpc.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(jsonrpc.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(jsonrpc.UnmarshalErrorHandler) + + for i := 0; i < b.N; i++ { + req := svc.Request(input{}) + req.Send() + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/endpoints/defaults.go b/vendor/github.com/aws/aws-sdk-go/aws/endpoints/defaults.go index 857f677dd..75c351d0f 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/endpoints/defaults.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/endpoints/defaults.go @@ -48,6 +48,7 @@ const ( A4bServiceID = "a4b" // A4b. AcmServiceID = "acm" // Acm. AcmPcaServiceID = "acm-pca" // AcmPca. + ApiMediatailorServiceID = "api.mediatailor" // ApiMediatailor. ApiPricingServiceID = "api.pricing" // ApiPricing. ApigatewayServiceID = "apigateway" // Apigateway. ApplicationAutoscalingServiceID = "application-autoscaling" // ApplicationAutoscaling. @@ -130,6 +131,7 @@ const ( ModelsLexServiceID = "models.lex" // ModelsLex. MonitoringServiceID = "monitoring" // Monitoring. MturkRequesterServiceID = "mturk-requester" // MturkRequester. + NeptuneServiceID = "neptune" // Neptune. OpsworksServiceID = "opsworks" // Opsworks. OpsworksCmServiceID = "opsworks-cm" // OpsworksCm. OrganizationsServiceID = "organizations" // Organizations. @@ -307,6 +309,16 @@ var awsPartition = partition{ "us-west-2": endpoint{}, }, }, + "api.mediatailor": service{ + + Endpoints: endpoints{ + "ap-northeast-1": endpoint{}, + "ap-southeast-1": endpoint{}, + "ap-southeast-2": endpoint{}, + "eu-west-1": endpoint{}, + "us-east-1": endpoint{}, + }, + }, "api.pricing": service{ Defaults: endpoint{ CredentialScope: credentialScope{ @@ -434,6 +446,7 @@ var awsPartition = partition{ Endpoints: endpoints{ "ap-northeast-1": endpoint{}, "ap-northeast-2": endpoint{}, + "ap-south-1": endpoint{}, "ap-southeast-1": endpoint{}, "ap-southeast-2": endpoint{}, "ca-central-1": endpoint{}, @@ -558,6 +571,7 @@ var awsPartition = partition{ "ca-central-1": endpoint{}, "eu-central-1": endpoint{}, "eu-west-1": endpoint{}, + "eu-west-2": endpoint{}, "us-east-1": endpoint{}, "us-east-2": endpoint{}, "us-west-1": endpoint{}, @@ -1046,6 +1060,7 @@ var awsPartition = partition{ "elasticfilesystem": service{ Endpoints: endpoints{ + "ap-northeast-2": endpoint{}, "ap-southeast-2": endpoint{}, "eu-central-1": endpoint{}, "eu-west-1": endpoint{}, @@ -1179,8 +1194,10 @@ var awsPartition = partition{ Endpoints: endpoints{ "ap-northeast-1": endpoint{}, + "ap-northeast-2": endpoint{}, "ap-southeast-1": endpoint{}, "ap-southeast-2": endpoint{}, + "ca-central-1": endpoint{}, "eu-central-1": endpoint{}, "eu-west-1": endpoint{}, "us-east-1": endpoint{}, @@ -1194,6 +1211,7 @@ var awsPartition = partition{ Protocols: []string{"https"}, }, Endpoints: endpoints{ + "eu-west-1": endpoint{}, "us-east-1": endpoint{}, "us-west-2": endpoint{}, }, @@ -1242,11 +1260,13 @@ var awsPartition = partition{ Endpoints: endpoints{ "ap-northeast-1": endpoint{}, + "ap-northeast-2": endpoint{}, "ap-south-1": endpoint{}, "ap-southeast-1": endpoint{}, "ap-southeast-2": endpoint{}, "eu-central-1": endpoint{}, "eu-west-1": endpoint{}, + "eu-west-2": endpoint{}, "us-east-1": endpoint{}, "us-east-2": endpoint{}, "us-west-2": endpoint{}, @@ -1509,8 +1529,10 @@ var awsPartition = partition{ Endpoints: endpoints{ "ap-northeast-1": endpoint{}, + "ap-northeast-2": endpoint{}, "ap-southeast-1": endpoint{}, "ap-southeast-2": endpoint{}, + "eu-central-1": endpoint{}, "eu-west-1": endpoint{}, "us-east-1": endpoint{}, "us-west-2": endpoint{}, @@ -1622,6 +1644,35 @@ var awsPartition = partition{ "us-east-1": endpoint{}, }, }, + "neptune": service{ + + Endpoints: endpoints{ + "eu-west-1": endpoint{ + Hostname: "rds.eu-west-1.amazonaws.com", + CredentialScope: credentialScope{ + Region: "eu-west-1", + }, + }, + "us-east-1": endpoint{ + Hostname: "rds.us-east-1.amazonaws.com", + CredentialScope: credentialScope{ + Region: "us-east-1", + }, + }, + "us-east-2": endpoint{ + Hostname: "rds.us-east-2.amazonaws.com", + CredentialScope: credentialScope{ + Region: "us-east-2", + }, + }, + "us-west-2": endpoint{ + Hostname: "rds.us-west-2.amazonaws.com", + CredentialScope: credentialScope{ + Region: "us-west-2", + }, + }, + }, + }, "opsworks": service{ Endpoints: endpoints{ @@ -1805,10 +1856,11 @@ var awsPartition = partition{ "runtime.sagemaker": service{ Endpoints: endpoints{ - "eu-west-1": endpoint{}, - "us-east-1": endpoint{}, - "us-east-2": endpoint{}, - "us-west-2": endpoint{}, + "ap-northeast-1": endpoint{}, + "eu-west-1": endpoint{}, + "us-east-1": endpoint{}, + "us-east-2": endpoint{}, + "us-west-2": endpoint{}, }, }, "s3": service{ @@ -1873,10 +1925,11 @@ var awsPartition = partition{ "sagemaker": service{ Endpoints: endpoints{ - "eu-west-1": endpoint{}, - "us-east-1": endpoint{}, - "us-east-2": endpoint{}, - "us-west-2": endpoint{}, + "ap-northeast-1": endpoint{}, + "eu-west-1": endpoint{}, + "us-east-1": endpoint{}, + "us-east-2": endpoint{}, + "us-west-2": endpoint{}, }, }, "sdb": service{ @@ -1991,6 +2044,7 @@ var awsPartition = partition{ "eu-west-1": endpoint{}, "us-east-1": endpoint{}, "us-east-2": endpoint{}, + "us-west-1": endpoint{}, "us-west-2": endpoint{}, }, }, @@ -2081,6 +2135,10 @@ var awsPartition = partition{ "eu-west-1": endpoint{}, "eu-west-2": endpoint{}, "eu-west-3": endpoint{}, + "fips-us-east-1": endpoint{}, + "fips-us-east-2": endpoint{}, + "fips-us-west-1": endpoint{}, + "fips-us-west-2": endpoint{}, "sa-east-1": endpoint{}, "us-east-1": endpoint{ SSLCommonName: "queue.{dnsSuffix}", @@ -2507,13 +2565,15 @@ var awscnPartition = partition{ "ecr": service{ Endpoints: endpoints{ - "cn-north-1": endpoint{}, + "cn-north-1": endpoint{}, + "cn-northwest-1": endpoint{}, }, }, "ecs": service{ Endpoints: endpoints{ - "cn-north-1": endpoint{}, + "cn-north-1": endpoint{}, + "cn-northwest-1": endpoint{}, }, }, "elasticache": service{ @@ -2927,6 +2987,12 @@ var awsusgovPartition = partition{ }, }, }, + "inspector": service{ + + Endpoints: endpoints{ + "us-gov-west-1": endpoint{}, + }, + }, "kinesis": service{ Endpoints: endpoints{ diff --git a/vendor/github.com/aws/aws-sdk-go/aws/logger.go b/vendor/github.com/aws/aws-sdk-go/aws/logger.go index 3babb5abd..6ed15b2ec 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/logger.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/logger.go @@ -71,6 +71,12 @@ const ( // LogDebugWithRequestErrors states the SDK should log when service requests fail // to build, send, validate, or unmarshal. LogDebugWithRequestErrors + + // LogDebugWithEventStreamBody states the SDK should log EventStream + // request and response bodys. This should be used to log the EventStream + // wire unmarshaled message content of requests and responses made while + // using the SDK Will also enable LogDebug. + LogDebugWithEventStreamBody ) // A Logger is a minimalistic interface for the SDK to log messages to. Should diff --git a/vendor/github.com/aws/aws-sdk-go/aws/request/handlers.go b/vendor/github.com/aws/aws-sdk-go/aws/request/handlers.go index 802ac88ad..605a72d3c 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/request/handlers.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/request/handlers.go @@ -14,6 +14,7 @@ type Handlers struct { Send HandlerList ValidateResponse HandlerList Unmarshal HandlerList + UnmarshalStream HandlerList UnmarshalMeta HandlerList UnmarshalError HandlerList Retry HandlerList @@ -30,6 +31,7 @@ func (h *Handlers) Copy() Handlers { Send: h.Send.copy(), ValidateResponse: h.ValidateResponse.copy(), Unmarshal: h.Unmarshal.copy(), + UnmarshalStream: h.UnmarshalStream.copy(), UnmarshalError: h.UnmarshalError.copy(), UnmarshalMeta: h.UnmarshalMeta.copy(), Retry: h.Retry.copy(), @@ -45,6 +47,7 @@ func (h *Handlers) Clear() { h.Send.Clear() h.Sign.Clear() h.Unmarshal.Clear() + h.UnmarshalStream.Clear() h.UnmarshalMeta.Clear() h.UnmarshalError.Clear() h.ValidateResponse.Clear() @@ -172,6 +175,21 @@ func (l *HandlerList) SwapNamed(n NamedHandler) (swapped bool) { return swapped } +// Swap will swap out all handlers matching the name passed in. The matched +// handlers will be swapped in. True is returned if the handlers were swapped. +func (l *HandlerList) Swap(name string, replace NamedHandler) bool { + var swapped bool + + for i := 0; i < len(l.list); i++ { + if l.list[i].Name == name { + l.list[i] = replace + swapped = true + } + } + + return swapped +} + // SetBackNamed will replace the named handler if it exists in the handler list. // If the handler does not exist the handler will be added to the end of the list. func (l *HandlerList) SetBackNamed(n NamedHandler) { diff --git a/vendor/github.com/aws/aws-sdk-go/aws/request/request.go b/vendor/github.com/aws/aws-sdk-go/aws/request/request.go index 69b7a01ad..75f0fe077 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/request/request.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/request/request.go @@ -46,6 +46,7 @@ type Request struct { Handlers Handlers Retryer + AttemptTime time.Time Time time.Time Operation *Operation HTTPRequest *http.Request @@ -121,6 +122,7 @@ func New(cfg aws.Config, clientInfo metadata.ClientInfo, handlers Handlers, Handlers: handlers.Copy(), Retryer: retryer, + AttemptTime: time.Now(), Time: time.Now(), ExpireTime: 0, Operation: operation, @@ -368,9 +370,9 @@ func (r *Request) Build() error { return r.Error } -// Sign will sign the request returning error if errors are encountered. +// Sign will sign the request, returning error if errors are encountered. // -// Send will build the request prior to signing. All Sign Handlers will +// Sign will build the request prior to signing. All Sign Handlers will // be executed in the order they were set. func (r *Request) Sign() error { r.Build() @@ -440,7 +442,7 @@ func (r *Request) GetBody() io.ReadSeeker { return r.safeBody } -// Send will send the request returning error if errors are encountered. +// Send will send the request, returning error if errors are encountered. // // Send will sign the request prior to sending. All Send Handlers will // be executed in the order they were set. @@ -461,6 +463,7 @@ func (r *Request) Send() error { }() for { + r.AttemptTime = time.Now() if aws.BoolValue(r.Retryable) { if r.Config.LogLevel.Matches(aws.LogDebugWithRequestRetries) { r.Config.Logger.Log(fmt.Sprintf("DEBUG: Retrying Request %s/%s, attempt %d", diff --git a/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_7.go b/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_7.go index 869b97a1a..e36e468b7 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_7.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_7.go @@ -21,7 +21,7 @@ func (noBody) WriteTo(io.Writer) (int64, error) { return 0, nil } var NoBody = noBody{} // ResetBody rewinds the request body back to its starting position, and -// set's the HTTP Request body reference. When the body is read prior +// sets the HTTP Request body reference. When the body is read prior // to being sent in the HTTP request it will need to be rewound. // // ResetBody will automatically be called by the SDK's build handler, but if diff --git a/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_8.go b/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_8.go index c32fc69bc..7c6a8000f 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_8.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/request/request_1_8.go @@ -11,7 +11,7 @@ import ( var NoBody = http.NoBody // ResetBody rewinds the request body back to its starting position, and -// set's the HTTP Request body reference. When the body is read prior +// sets the HTTP Request body reference. When the body is read prior // to being sent in the HTTP request it will need to be rewound. // // ResetBody will automatically be called by the SDK's build handler, but if diff --git a/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination.go b/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination.go index 159518a75..a633ed5ac 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination.go @@ -35,8 +35,12 @@ type Pagination struct { // NewRequest should always be built from the same API operations. It is // undefined if different API operations are returned on subsequent calls. NewRequest func() (*Request, error) + // EndPageOnSameToken, when enabled, will allow the paginator to stop on + // token that are the same as its previous tokens. + EndPageOnSameToken bool started bool + prevTokens []interface{} nextTokens []interface{} err error @@ -49,7 +53,15 @@ type Pagination struct { // // Will always return true if Next has not been called yet. func (p *Pagination) HasNextPage() bool { - return !(p.started && len(p.nextTokens) == 0) + if !p.started { + return true + } + + hasNextPage := len(p.nextTokens) != 0 + if p.EndPageOnSameToken { + return hasNextPage && !awsutil.DeepEqual(p.nextTokens, p.prevTokens) + } + return hasNextPage } // Err returns the error Pagination encountered when retrieving the next page. @@ -96,6 +108,7 @@ func (p *Pagination) Next() bool { return false } + p.prevTokens = p.nextTokens p.nextTokens = req.nextPageTokens() p.curPage = req.Data diff --git a/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination_test.go b/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination_test.go index b942e014f..03aed301a 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination_test.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/request/request_pagination_test.go @@ -466,20 +466,48 @@ func TestPagination_Standalone(t *testing.T) { Value, PrevToken, NextToken *string } - cases := [][]testCase{ + type testCaseList struct { + StopOnSameToken bool + Cases []testCase + } + + cases := []testCaseList{ { - testCase{aws.String("FirstValue"), aws.String("InitalToken"), aws.String("FirstToken")}, - testCase{aws.String("SecondValue"), aws.String("FirstToken"), aws.String("SecondToken")}, - testCase{aws.String("ThirdValue"), aws.String("SecondToken"), nil}, + Cases: []testCase{ + testCase{aws.String("FirstValue"), aws.String("InitalToken"), aws.String("FirstToken")}, + testCase{aws.String("SecondValue"), aws.String("FirstToken"), aws.String("SecondToken")}, + testCase{aws.String("ThirdValue"), aws.String("SecondToken"), nil}, + }, + StopOnSameToken: false, }, { - testCase{aws.String("FirstValue"), aws.String("InitalToken"), aws.String("FirstToken")}, - testCase{aws.String("SecondValue"), aws.String("FirstToken"), aws.String("SecondToken")}, - testCase{aws.String("ThirdValue"), aws.String("SecondToken"), aws.String("")}, + Cases: []testCase{ + testCase{aws.String("FirstValue"), aws.String("InitalToken"), aws.String("FirstToken")}, + testCase{aws.String("SecondValue"), aws.String("FirstToken"), aws.String("SecondToken")}, + testCase{aws.String("ThirdValue"), aws.String("SecondToken"), aws.String("")}, + }, + StopOnSameToken: false, + }, + { + Cases: []testCase{ + testCase{aws.String("FirstValue"), aws.String("InitalToken"), aws.String("FirstToken")}, + testCase{aws.String("SecondValue"), aws.String("FirstToken"), aws.String("SecondToken")}, + testCase{nil, aws.String("SecondToken"), aws.String("SecondToken")}, + }, + StopOnSameToken: true, + }, + { + Cases: []testCase{ + testCase{aws.String("FirstValue"), aws.String("InitalToken"), aws.String("FirstToken")}, + testCase{aws.String("SecondValue"), aws.String("FirstToken"), aws.String("SecondToken")}, + testCase{aws.String("SecondValue"), aws.String("SecondToken"), aws.String("SecondToken")}, + }, + StopOnSameToken: true, }, } - for _, c := range cases { + for _, testcase := range cases { + c := testcase.Cases input := testPageInput{ NextToken: c[0].PrevToken, } @@ -487,6 +515,7 @@ func TestPagination_Standalone(t *testing.T) { svc := awstesting.NewClient() i := 0 p := request.Pagination{ + EndPageOnSameToken: testcase.StopOnSameToken, NewRequest: func() (*request.Request, error) { r := svc.NewRequest( &request.Operation{ diff --git a/vendor/github.com/aws/aws-sdk-go/aws/session/env_config.go b/vendor/github.com/aws/aws-sdk-go/aws/session/env_config.go index 12b452177..82e04d76c 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/session/env_config.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/session/env_config.go @@ -96,9 +96,23 @@ type envConfig struct { // // AWS_CA_BUNDLE=$HOME/my_custom_ca_bundle CustomCABundle string + + csmEnabled string + CSMEnabled bool + CSMPort string + CSMClientID string } var ( + csmEnabledEnvKey = []string{ + "AWS_CSM_ENABLED", + } + csmPortEnvKey = []string{ + "AWS_CSM_PORT", + } + csmClientIDEnvKey = []string{ + "AWS_CSM_CLIENT_ID", + } credAccessEnvKey = []string{ "AWS_ACCESS_KEY_ID", "AWS_ACCESS_KEY", @@ -157,6 +171,12 @@ func envConfigLoad(enableSharedConfig bool) envConfig { setFromEnvVal(&cfg.Creds.SecretAccessKey, credSecretEnvKey) setFromEnvVal(&cfg.Creds.SessionToken, credSessionEnvKey) + // CSM environment variables + setFromEnvVal(&cfg.csmEnabled, csmEnabledEnvKey) + setFromEnvVal(&cfg.CSMPort, csmPortEnvKey) + setFromEnvVal(&cfg.CSMClientID, csmClientIDEnvKey) + cfg.CSMEnabled = len(cfg.csmEnabled) > 0 + // Require logical grouping of credentials if len(cfg.Creds.AccessKeyID) == 0 || len(cfg.Creds.SecretAccessKey) == 0 { cfg.Creds = credentials.Value{} diff --git a/vendor/github.com/aws/aws-sdk-go/aws/session/session.go b/vendor/github.com/aws/aws-sdk-go/aws/session/session.go index 259b5c0fe..51f305563 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/session/session.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/session/session.go @@ -15,6 +15,7 @@ import ( "github.com/aws/aws-sdk-go/aws/corehandlers" "github.com/aws/aws-sdk-go/aws/credentials" "github.com/aws/aws-sdk-go/aws/credentials/stscreds" + "github.com/aws/aws-sdk-go/aws/csm" "github.com/aws/aws-sdk-go/aws/defaults" "github.com/aws/aws-sdk-go/aws/endpoints" "github.com/aws/aws-sdk-go/aws/request" @@ -81,10 +82,16 @@ func New(cfgs ...*aws.Config) *Session { r.Error = err }) } + return s } - return deprecatedNewSession(cfgs...) + s := deprecatedNewSession(cfgs...) + if envCfg.CSMEnabled { + enableCSM(&s.Handlers, envCfg.CSMClientID, envCfg.CSMPort, s.Config.Logger) + } + + return s } // NewSession returns a new Session created from SDK defaults, config files, @@ -300,10 +307,22 @@ func deprecatedNewSession(cfgs ...*aws.Config) *Session { } initHandlers(s) - return s } +func enableCSM(handlers *request.Handlers, clientID string, port string, logger aws.Logger) { + logger.Log("Enabling CSM") + if len(port) == 0 { + port = csm.DefaultPort + } + + r, err := csm.Start(clientID, "127.0.0.1:"+port) + if err != nil { + return + } + r.InjectHandlers(handlers) +} + func newSession(opts Options, envCfg envConfig, cfgs ...*aws.Config) (*Session, error) { cfg := defaults.Config() handlers := defaults.Handlers() @@ -343,6 +362,9 @@ func newSession(opts Options, envCfg envConfig, cfgs ...*aws.Config) (*Session, } initHandlers(s) + if envCfg.CSMEnabled { + enableCSM(&s.Handlers, envCfg.CSMClientID, envCfg.CSMPort, s.Config.Logger) + } // Setup HTTP client with custom cert bundle if enabled if opts.CustomCABundle != nil { diff --git a/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/functional_test.go b/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/functional_test.go index 93d664b23..817494522 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/functional_test.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/functional_test.go @@ -48,7 +48,7 @@ func TestPresignHandler(t *testing.T) { expectedHost := "bucket.s3.mock-region.amazonaws.com" expectedDate := "19700101T000000Z" expectedHeaders := "content-disposition;host;x-amz-acl" - expectedSig := "a46583256431b09eb45ba4af2e6286d96a9835ed13721023dc8076dfdcb90fcb" + expectedSig := "2d76a414208c0eac2a23ef9c834db9635ecd5a0fbb447a00ad191f82d854f55b" expectedCred := "AKID/19700101/mock-region/s3/aws4_request" u, _ := url.Parse(urlstr) @@ -71,8 +71,8 @@ func TestPresignHandler(t *testing.T) { if e, a := "300", urlQ.Get("X-Amz-Expires"); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := "UNSIGNED-PAYLOAD", urlQ.Get("X-Amz-Content-Sha256"); e != a { - t.Errorf("expect %v, got %v", e, a) + if a := urlQ.Get("X-Amz-Content-Sha256"); len(a) != 0 { + t.Errorf("expect no content sha256 got %v", a) } if e, a := "+", urlstr; strings.Contains(a, e) { // + encoded as %20 @@ -98,7 +98,7 @@ func TestPresignRequest(t *testing.T) { expectedHost := "bucket.s3.mock-region.amazonaws.com" expectedDate := "19700101T000000Z" expectedHeaders := "content-disposition;host;x-amz-acl" - expectedSig := "a46583256431b09eb45ba4af2e6286d96a9835ed13721023dc8076dfdcb90fcb" + expectedSig := "2d76a414208c0eac2a23ef9c834db9635ecd5a0fbb447a00ad191f82d854f55b" expectedCred := "AKID/19700101/mock-region/s3/aws4_request" expectedHeaderMap := http.Header{ "x-amz-acl": []string{"public-read"}, @@ -128,8 +128,8 @@ func TestPresignRequest(t *testing.T) { if e, a := "300", urlQ.Get("X-Amz-Expires"); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := "UNSIGNED-PAYLOAD", urlQ.Get("X-Amz-Content-Sha256"); e != a { - t.Errorf("expect %v, got %v", e, a) + if a := urlQ.Get("X-Amz-Content-Sha256"); len(a) != 0 { + t.Errorf("expect no content sha256 got %v", a) } if e, a := "+", urlstr; strings.Contains(a, e) { // + encoded as %20 @@ -169,7 +169,7 @@ func TestStandaloneSign_WithPort(t *testing.T) { cases := []struct { description string - url string + url string expectedSig string }{ { @@ -213,7 +213,7 @@ func TestStandalonePresign_WithPort(t *testing.T) { cases := []struct { description string - url string + url string expectedSig string }{ { @@ -241,7 +241,7 @@ func TestStandalonePresign_WithPort(t *testing.T) { for _, c := range cases { signer := v4.NewSigner(unit.Session.Config.Credentials) req, _ := http.NewRequest("GET", c.url, nil) - _, err := signer.Presign(req, nil, "es", "us-east-1", 5 * time.Minute, time.Unix(0, 0)) + _, err := signer.Presign(req, nil, "es", "us-east-1", 5*time.Minute, time.Unix(0, 0)) if err != nil { t.Fatalf("expect no error, got %v", err) } diff --git a/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4.go b/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4.go index 6e4637612..f35861315 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4.go @@ -135,6 +135,7 @@ var requiredSignedHeaders = rules{ "X-Amz-Server-Side-Encryption-Customer-Key-Md5": struct{}{}, "X-Amz-Storage-Class": struct{}{}, "X-Amz-Website-Redirect-Location": struct{}{}, + "X-Amz-Content-Sha256": struct{}{}, }, }, patterns{"X-Amz-Meta-"}, @@ -671,8 +672,15 @@ func (ctx *signingCtx) buildSignature() { func (ctx *signingCtx) buildBodyDigest() error { hash := ctx.Request.Header.Get("X-Amz-Content-Sha256") if hash == "" { - if ctx.unsignedPayload || (ctx.isPresign && ctx.ServiceName == "s3") { + includeSHA256Header := ctx.unsignedPayload || + ctx.ServiceName == "s3" || + ctx.ServiceName == "glacier" + + s3Presign := ctx.isPresign && ctx.ServiceName == "s3" + + if ctx.unsignedPayload || s3Presign { hash = "UNSIGNED-PAYLOAD" + includeSHA256Header = !s3Presign } else if ctx.Body == nil { hash = emptyStringSHA256 } else { @@ -681,7 +689,8 @@ func (ctx *signingCtx) buildBodyDigest() error { } hash = hex.EncodeToString(makeSha256Reader(ctx.Body)) } - if ctx.unsignedPayload || ctx.ServiceName == "s3" || ctx.ServiceName == "glacier" { + + if includeSHA256Header { ctx.Request.Header.Set("X-Amz-Content-Sha256", hash) } } diff --git a/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4_test.go b/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4_test.go index 16bc7378a..09c054d06 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4_test.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/signer/v4/v4_test.go @@ -223,16 +223,36 @@ func TestSignBodyGlacier(t *testing.T) { } } -func TestPresignEmptyBodyS3(t *testing.T) { - req, body := buildRequest("s3", "us-east-1", "hello") +func TestPresign_SignedPayload(t *testing.T) { + req, body := buildRequest("glacier", "us-east-1", "hello") signer := buildSigner() - signer.Presign(req, body, "s3", "us-east-1", 5*time.Minute, time.Now()) + signer.Presign(req, body, "glacier", "us-east-1", 5*time.Minute, time.Now()) + hash := req.Header.Get("X-Amz-Content-Sha256") + if e, a := "2cf24dba5fb0a30e26e83b2ac5b9e29e1b161e5c1fa7425e73043362938b9824", hash; e != a { + t.Errorf("expect %v, got %v", e, a) + } +} + +func TestPresign_UnsignedPayload(t *testing.T) { + req, body := buildRequest("service-name", "us-east-1", "hello") + signer := buildSigner() + signer.UnsignedPayload = true + signer.Presign(req, body, "service-name", "us-east-1", 5*time.Minute, time.Now()) hash := req.Header.Get("X-Amz-Content-Sha256") if e, a := "UNSIGNED-PAYLOAD", hash; e != a { t.Errorf("expect %v, got %v", e, a) } } +func TestPresign_UnsignedPayload_S3(t *testing.T) { + req, body := buildRequest("s3", "us-east-1", "hello") + signer := buildSigner() + signer.Presign(req, body, "s3", "us-east-1", 5*time.Minute, time.Now()) + if a := req.Header.Get("X-Amz-Content-Sha256"); len(a) != 0 { + t.Errorf("expect no content sha256 got %v", a) + } +} + func TestSignUnseekableBody(t *testing.T) { req, body := buildRequestWithBodyReader("mock-service", "mock-region", bytes.NewBuffer([]byte("hello"))) signer := buildSigner() diff --git a/vendor/github.com/aws/aws-sdk-go/aws/version.go b/vendor/github.com/aws/aws-sdk-go/aws/version.go index 82d77b75e..560f7912e 100644 --- a/vendor/github.com/aws/aws-sdk-go/aws/version.go +++ b/vendor/github.com/aws/aws-sdk-go/aws/version.go @@ -5,4 +5,4 @@ package aws const SDKName = "aws-sdk-go" // SDKVersion is the version of this SDK -const SDKVersion = "1.13.50" +const SDKVersion = "1.14.17" diff --git a/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/eventstream_test.go b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/eventstream_test.go new file mode 100644 index 000000000..0e1c67a0d --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/eventstream_test.go @@ -0,0 +1,188 @@ +// +build integration + +package s3 + +import ( + "bytes" + "encoding/csv" + "fmt" + "io" + "os" + "path/filepath" + "strings" + "testing" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awserr" + "github.com/aws/aws-sdk-go/service/s3" +) + +func TestSelectObjectContent(t *testing.T) { + keyName := "selectObject.csv" + putTestFile(t, filepath.Join("testdata", "positive_select.csv"), keyName) + + resp, err := svc.SelectObjectContent(&s3.SelectObjectContentInput{ + Bucket: bucketName, + Key: &keyName, + Expression: aws.String("Select * from S3Object"), + ExpressionType: aws.String(s3.ExpressionTypeSql), + InputSerialization: &s3.InputSerialization{ + CSV: &s3.CSVInput{ + FieldDelimiter: aws.String(","), + FileHeaderInfo: aws.String(s3.FileHeaderInfoIgnore), + }, + }, + OutputSerialization: &s3.OutputSerialization{ + CSV: &s3.CSVOutput{ + FieldDelimiter: aws.String(","), + }, + }, + }) + if err != nil { + t.Fatalf("expect no error, %v", err) + } + defer resp.EventStream.Close() + + var sum int64 + var processed int64 + for event := range resp.EventStream.Events() { + switch tv := event.(type) { + case *s3.RecordsEvent: + sum += int64(len(tv.Payload)) + case *s3.StatsEvent: + processed = *tv.Details.BytesProcessed + } + } + + if sum == 0 { + t.Errorf("expect selected content, got none") + } + + if processed == 0 { + t.Errorf("expect selected status bytes processed, got none") + } + + if err := resp.EventStream.Err(); err != nil { + t.Fatalf("exect no error, %v", err) + } +} + +func TestSelectObjectContent_Error(t *testing.T) { + keyName := "negativeSelect.csv" + + buf := make([]byte, 0, 1024*1024*6) + buf = append(buf, []byte("name,number\n")...) + line := []byte("jj,0\n") + for i := 0; i < (cap(buf)/len(line))-2; i++ { + buf = append(buf, line...) + } + buf = append(buf, []byte("gg,NaN\n")...) + + putTestContent(t, bytes.NewReader(buf), keyName) + + resp, err := svc.SelectObjectContent(&s3.SelectObjectContentInput{ + Bucket: bucketName, + Key: &keyName, + Expression: aws.String("SELECT name FROM S3Object WHERE cast(number as int) < 1"), + ExpressionType: aws.String(s3.ExpressionTypeSql), + InputSerialization: &s3.InputSerialization{ + CSV: &s3.CSVInput{ + FileHeaderInfo: aws.String(s3.FileHeaderInfoUse), + }, + }, + OutputSerialization: &s3.OutputSerialization{ + CSV: &s3.CSVOutput{ + FieldDelimiter: aws.String(","), + }, + }, + }) + if err != nil { + t.Fatalf("expect no error, %v", err) + } + defer resp.EventStream.Close() + + var sum int64 + for event := range resp.EventStream.Events() { + switch tv := event.(type) { + case *s3.RecordsEvent: + sum += int64(len(tv.Payload)) + } + } + + if sum == 0 { + t.Errorf("expect selected content") + } + + err = resp.EventStream.Err() + if err == nil { + t.Fatalf("exepct error") + } + + aerr := err.(awserr.Error) + if a := aerr.Code(); len(a) == 0 { + t.Errorf("expect, error code") + } + if a := aerr.Message(); len(a) == 0 { + t.Errorf("expect, error message") + } +} + +func TestSelectObjectContent_Stream(t *testing.T) { + keyName := "selectGopher.csv" + + buf := `name,number +gopher,0 +ᵷodɥǝɹ,1 +` + // Put a mock CSV file to the S3 bucket so that its contents can be + // selected. + putTestContent(t, strings.NewReader(buf), keyName) + + // Make the Select Object Content API request using the object uploaded. + resp, err := svc.SelectObjectContent(&s3.SelectObjectContentInput{ + Bucket: bucketName, + Key: &keyName, + Expression: aws.String("SELECT name FROM S3Object WHERE cast(number as int) < 1"), + ExpressionType: aws.String(s3.ExpressionTypeSql), + InputSerialization: &s3.InputSerialization{ + CSV: &s3.CSVInput{ + FileHeaderInfo: aws.String(s3.FileHeaderInfoUse), + }, + }, + OutputSerialization: &s3.OutputSerialization{ + CSV: &s3.CSVOutput{}, + }, + }) + if err != nil { + fmt.Fprintf(os.Stderr, "failed making API request, %v\n", err) + return + } + defer resp.EventStream.Close() + + results, resultWriter := io.Pipe() + go func() { + defer resultWriter.Close() + for event := range resp.EventStream.Events() { + switch e := event.(type) { + case *s3.RecordsEvent: + resultWriter.Write(e.Payload) + case *s3.StatsEvent: + fmt.Printf("Processed %d bytes\n", *e.Details.BytesProcessed) + } + } + }() + + // Printout the results + resReader := csv.NewReader(results) + for { + record, err := resReader.Read() + if err == io.EOF { + break + } + fmt.Println(record) + } + + if err := resp.EventStream.Err(); err != nil { + fmt.Fprintf(os.Stderr, "reading from event stream failed, %v\n", err) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/shared_test.go b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/shared_test.go index 5e00b46cf..58ef0042b 100644 --- a/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/shared_test.go +++ b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/shared_test.go @@ -4,6 +4,7 @@ package s3 import ( "fmt" + "io" "os" "testing" @@ -65,3 +66,24 @@ func teardown() { svc.DeleteBucket(&s3.DeleteBucketInput{Bucket: bucketName}) } + +func putTestFile(t *testing.T, filename, key string) { + f, err := os.Open(filename) + if err != nil { + t.Fatalf("failed to open testfile, %v", err) + } + defer f.Close() + + putTestContent(t, f, key) +} + +func putTestContent(t *testing.T, reader io.ReadSeeker, key string) { + _, err := svc.PutObject(&s3.PutObjectInput{ + Bucket: bucketName, + Key: aws.String(key), + Body: reader, + }) + if err != nil { + t.Errorf("expect no error, got %v", err) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/testdata/positive_select.csv b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/testdata/positive_select.csv new file mode 100644 index 000000000..f14f901e3 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/customizations/s3/testdata/positive_select.csv @@ -0,0 +1,556 @@ +A,B,C,D,E,F,G,H,I,J +0,0,0.5,217.371,217.658,218.002,269.445,487.447,2.106,489.554 +0,0,490.077,0.657,0.84,1.588,121.092,122.681,2.185,124.867 +0,9,490.077,1.602,1.676,1.977,184.155,186.132,1.198,187.331 +0,6,490.384,1.787,1.962,2.451,270.592,273.044,4.158,277.202 +0,5,491.125,0.693,0.877,1.9,295.589,297.49,19.456,316.946 +0,8,491.348,1.833,1.924,66.432,250.035,316.467,2.379,318.847 +0,0,614.955,0.455,0.507,8.554,229.261,237.815,15.761,253.577 +0,9,677.418,0.493,0.546,14.288,176.628,190.917,6.522,197.44 +0,3,491.864,1.034,1.109,250.552,132.254,382.806,2.485,385.291 +0,2,491.351,1.763,1.807,276.239,122.283,398.523,26.436,424.96 +0,4,490.154,1.867,1.935,341.544,86.243,427.787,15.659,443.447 +0,8,810.203,0.381,0.453,21.4,101.836,123.236,9.01,132.246 +0,6,767.595,0.54,0.592,68.928,105.705,174.633,0.93,175.564 +0,0,868.54,0.384,0.483,8.721,82.83,91.552,37.154,128.706 +0,9,874.866,0.472,0.574,41.728,80.617,122.345,11.14,133.486 +0,1,490.757,2.284,2.336,426.145,91.084,517.229,0.507,517.736 +0,7,490.755,1.652,1.816,426.521,90.913,517.435,2.244,519.679 +0,6,943.165,0.324,0.391,2.676,116.421,119.098,0.452,119.551 +0,2,916.32,0.609,0.884,81.097,83.55,164.647,1.342,165.989 +0,8,942.462,0.374,0.464,55.682,84.915,140.597,46.844,187.442 +0,4,933.61,0.377,0.451,64.784,131.147,195.932,9.997,205.929 +0,5,808.083,0.565,0.641,190.27,140.808,331.078,0.465,331.543 +0,3,877.165,0.47,0.546,121.153,141.078,262.231,2.558,264.789 +0,7,1010.442,0.372,0.441,2.201,129.116,131.318,7.148,138.467 +0,6,1062.725,0.483,0.581,6.402,80.848,87.251,45.416,132.668 +0,1,1008.502,1.579,1.706,65.348,121.324,186.672,1.251,187.924 +0,0,997.256,0.402,0.832,76.472,121.367,197.839,18.513,216.353 +0,9,1008.364,1.428,1.582,70.936,133.941,204.878,2.255,207.133 +0,8,1129.916,0.458,0.551,19.07,113.216,132.287,0.38,132.667 +0,4,1139.547,0.388,0.457,9.775,113.036,122.812,1.531,124.343 +0,2,1082.318,0.352,0.43,67.775,112.768,180.543,22.578,203.122 +0,5,1139.629,0.416,0.466,10.429,135.269,145.699,2.425,148.124 +0,3,1141.962,0.394,0.472,53.375,92.295,145.67,0.212,145.883 +0,0,1213.618,0.408,0.481,2.655,112.525,115.18,0.547,115.728 +0,9,1215.504,0.427,0.672,2.233,111.382,113.615,1.145,114.761 +0,6,1195.403,0.437,0.511,25.098,109.564,134.662,22.852,157.515 +0,7,1148.919,0.486,1.021,138.981,91.884,230.865,16.504,247.369 +0,4,1263.894,0.711,0.788,33.761,98.396,132.158,0.671,132.83 +0,1,1196.433,0.301,0.364,100.757,99.252,200.01,1.116,201.126 +0,8,1262.594,0.964,1.248,66.394,96.747,163.141,1.173,164.315 +0,2,1285.445,0.209,0.254,43.808,97.338,141.146,0.548,141.694 +0,9,1330.271,0.288,0.355,35.329,84.511,119.84,16.98,136.821 +0,6,1352.928,0.474,0.579,12.904,100.981,113.886,0.517,114.403 +0,0,1329.353,0.294,0.366,36.781,100.857,137.638,0.532,138.171 +0,2,1427.143,0.532,0.58,9.336,79.756,89.093,1.239,90.332 +0,4,1396.736,0.503,0.592,44.925,80.622,125.548,10.972,136.52 +0,8,1426.921,0.631,0.71,15.906,90.068,105.975,1.38,107.356 +0,5,1287.759,0.278,0.316,181.454,80.529,261.984,10.774,272.759 +0,1,1397.566,0.268,0.324,99.635,82.55,182.186,8.445,190.632 +0,0,1467.53,1.354,1.536,33.298,87.09,120.388,0.517,120.905 +0,7,1396.3,0.593,0.684,104.729,87.108,191.838,1.133,192.971 +0,9,1467.103,1.422,1.546,42.719,82.771,125.491,9.398,134.89 +0,3,1287.85,0.245,0.272,248.437,79.849,328.286,9.318,337.605 +0,6,1467.339,0.933,1.132,69.074,88.797,157.871,0.891,158.762 +0,4,1533.267,0.449,0.548,42.003,81.379,123.383,9.108,132.491 +0,2,1517.483,0.32,0.399,56.455,91.585,148.04,0.44,148.48 +0,8,1534.284,0.425,0.48,39.738,91.427,131.166,1.247,132.413 +0,5,1560.528,0.396,0.499,19.318,86.69,106.008,1.54,107.548 +0,6,1626.109,0.34,0.43,6.88,82.575,89.455,0.568,90.023 +0,1,1588.231,0.451,0.541,44.902,84.635,129.537,24.285,153.823 +0,7,1589.279,0.288,0.352,57.597,94.892,152.489,0.555,153.044 +0,9,1602.002,0.424,0.68,45.095,94.862,139.957,0.797,140.755 +0,5,1668.084,0.29,0.351,29.143,84.196,113.34,26.201,139.542 +0,4,1665.766,0.373,0.457,34.457,107.029,141.486,0.715,142.201 +0,2,1665.971,0.303,0.375,47.175,94.572,141.748,0.861,142.609 +0,0,1588.445,0.517,0.625,154.081,80.415,234.496,7.112,241.608 +0,3,1625.463,0.647,0.751,143.666,83.075,226.742,21.951,248.694 +0,7,1742.325,0.53,0.624,25.438,84.393,109.831,22.061,131.893 +0,1,1742.079,0.555,0.659,34.769,97.152,131.921,0.364,132.286 +0,8,1666.705,0.883,1.046,113.237,94.162,207.4,1.118,208.519 +0,2,1808.586,0.3,0.362,21.568,78.537,100.106,1.979,102.085 +0,0,1830.064,0.405,0.481,7.114,81.532,88.647,2.392,91.04 +0,6,1716.143,0.49,0.557,121.371,87.089,208.46,14.395,222.856 +0,5,1807.634,0.442,0.527,36.463,94.602,131.065,1.369,132.434 +0,8,1875.228,0.164,0.198,3.582,79.597,83.179,1.504,84.684 +0,1,1874.367,0.562,0.645,28.304,79.054,107.358,0.456,107.815 +0,4,1807.975,0.4,0.482,99.4,81.109,180.509,1.133,181.643 +0,9,1742.773,0.339,0.39,168.948,80.077,249.025,2.367,251.393 +0,3,1874.166,0.495,0.593,38.988,84.602,123.59,7.982,131.572 +0,8,1959.921,0.415,0.494,9.861,82.855,92.717,2.765,95.483 +0,2,1910.682,0.995,1.109,69.161,85.346,154.507,1.625,156.133 +0,0,1921.114,0.394,0.473,92.984,80.599,173.584,1.153,174.738 +0,3,2005.748,0.476,0.556,7.441,83.841,91.283,1.422,92.705 +0,5,1940.076,0.325,0.401,73.91,84.342,158.253,1.718,159.972 +0,7,1874.22,0.557,0.642,140.027,91.031,231.059,1.754,232.814 +0,6,1939.008,0.377,0.455,95.473,81.569,177.042,1.072,178.115 +0,4,1989.626,0.314,0.388,51.23,81.724,132.955,11.694,144.65 +0,9,1994.183,0.321,0.394,57.158,82.743,139.902,0.758,140.66 +0,2,2066.819,0.204,0.259,13.808,84.448,98.256,1.366,99.623 +0,1,1982.189,0.295,0.368,99.448,85.395,184.843,1.56,186.403 +0,8,2055.408,0.203,0.253,51.567,82.101,133.668,1.073,134.741 +0,9,2134.846,0.25,0.306,16.106,85.649,101.755,0.176,101.931 +0,1,2168.598,0.229,0.302,6.826,81.481,88.307,1.143,89.45 +0,4,2134.304,0.373,0.492,45.962,81.137,127.099,0.549,127.648 +0,7,2107.039,0.483,0.555,73.813,81.641,155.455,1.562,157.017 +0,6,2117.128,0.208,0.255,63.776,83.028,146.805,1.656,148.461 +0,8,2190.154,0.225,0.285,29.098,80.996,110.094,1.344,111.439 +0,9,2236.784,0.256,0.32,4.01,82.88,86.89,2.12,89.011 +0,2,2166.449,0.317,0.395,80.763,84.208,164.971,2.32,167.291 +0,5,2100.052,0.294,0.365,146.743,86.973,233.716,0.672,234.389 +0,3,2098.458,0.241,0.3,150.5,84.733,235.234,1.292,236.526 +0,0,2095.857,0.215,0.271,153.005,85.917,238.923,0.534,239.458 +0,6,2265.593,0.182,0.218,20.159,80.738,100.897,1.449,102.347 +0,4,2261.957,0.207,0.256,42.386,82.309,124.696,1.433,126.13 +0,7,2264.061,0.243,0.288,51.339,80.631,131.97,0.973,132.943 +0,8,2301.604,0.391,0.474,24.05,81.886,105.937,1.805,107.743 +0,1,2258.053,0.206,0.26,93.644,81.876,175.52,1.331,176.852 +0,0,2335.321,0.204,0.245,21.603,81.849,103.452,0.941,104.394 +0,6,2367.949,0.434,0.515,6.274,83.161,89.435,4.495,93.931 +0,3,2334.991,0.332,0.403,58.507,88.463,146.971,1.116,148.088 +0,8,2409.356,0.385,0.463,13.78,83.24,97.02,0.344,97.364 +0,5,2334.448,0.364,0.451,106.034,82.488,188.523,1.39,189.914 +0,9,2325.809,0.429,0.506,114.736,84.279,199.015,1.209,200.225 +0,2,2333.745,0.423,0.517,106.853,85.698,192.551,1.745,194.296 +0,4,2388.097,0.399,0.498,67.532,84.096,151.628,0.599,152.228 +0,3,2483.086,0.35,0.427,19.21,81.612,100.822,3.51,104.333 +0,1,2434.913,0.435,0.577,86.727,83.002,169.729,1.902,171.632 +0,7,2397.012,0.331,0.416,142.874,80.866,223.74,1.672,225.413 +0,6,2461.891,0.36,0.441,78.194,82.238,160.433,0.613,161.046 +0,9,2526.038,0.665,0.74,32.614,86.809,119.423,1.275,120.699 +0,4,2540.332,0.326,0.387,42.093,80.618,122.711,2.268,124.979 +0,8,2506.727,0.378,0.456,99.838,79.225,179.064,0.294,179.358 +0,6,2622.939,0.33,0.385,1.186,81.73,82.917,2.248,85.165 +0,3,2587.429,0.61,0.72,59.939,82.437,142.376,0.97,143.346 +0,1,2606.549,0.391,0.459,40.636,83.436,124.072,2.096,126.169 +0,7,2622.432,0.383,0.463,30.735,80.765,111.501,0.733,112.234 +0,2,2528.046,0.199,0.244,128.905,85.696,214.602,0.334,214.936 +0,4,2665.318,0.312,0.399,26.866,81.414,108.281,0.222,108.504 +0,5,2524.369,0.329,0.413,167.907,84.934,252.841,1.305,254.147 +0,8,2686.096,0.401,0.494,7.747,85.181,92.928,2.125,95.053 +0,0,2439.722,0.357,0.696,254.259,89.099,343.358,2.809,346.167 +0,9,2646.75,0.681,0.799,73.064,84.639,157.704,3.532,161.236 +0,6,2708.115,0.4,0.481,14.501,86.758,101.259,0.934,102.194 +0,3,2730.783,0.303,0.377,35.013,88.845,123.858,1.666,125.524 +0,1,2732.726,0.318,0.414,53.138,78.873,132.011,0.237,132.249 +0,0,2785.893,0.375,0.447,25.451,83.295,108.746,4.165,112.911 +0,9,2807.993,0.31,0.384,35.981,91.657,127.639,0.466,128.106 +0,2,2742.992,0.403,0.56,101.119,91.707,192.827,5.458,198.285 +0,8,2781.157,0.365,0.446,70.781,90.886,161.667,0.817,162.484 +0,1,2864.982,0.311,0.402,19.474,86.691,106.165,3.435,109.601 +0,3,2856.319,0.429,0.493,54.672,82.88,137.553,0.33,137.884 +0,5,2778.523,0.309,0.392,132.818,84.58,217.399,1.527,218.927 +0,0,2898.815,0.362,0.463,12.416,86.002,98.418,1.107,99.525 +0,7,2734.674,0.744,0.873,195.477,83.728,279.205,7.848,287.053 +0,4,2773.831,0.339,0.428,156.128,91.457,247.585,1.311,248.897 +0,6,2810.317,0.339,0.432,125.657,102.335,227.993,2.034,230.027 +0,2,2941.285,0.294,0.367,38.02,79.84,117.86,1.696,119.556 +0,8,2943.648,0.293,0.373,38.288,79.728,118.016,2.042,120.058 +0,9,2936.108,0.466,0.563,63.933,82.084,146.017,1.602,147.619 +0,4,3022.735,0.269,0.339,3.697,87.616,91.313,0.516,91.83 +0,3,2994.213,0.418,0.495,42.946,81.806,124.752,0.29,125.043 +0,1,2974.591,0.641,0.762,72.809,81.187,153.997,1.512,155.51 +0,9,3083.737,0.352,0.425,15.144,84.807,99.951,1.383,101.335 +0,6,3040.353,0.399,0.48,61.605,83.294,144.899,9.906,154.806 +0,2,3060.852,0.407,0.487,40.928,92.521,133.449,0.893,134.342 +0,0,2998.348,0.336,0.417,115.561,82.329,197.89,2.808,200.698 +0,8,3063.714,0.314,0.391,50.53,84.619,135.15,28.56,163.71 +0,1,3130.111,0.381,0.484,36.604,82.182,118.787,1.306,120.094 +0,5,2997.458,0.349,0.427,169.477,83.501,252.978,2.447,255.425 +0,7,3021.738,0.425,0.518,148.774,83.974,232.748,0.411,233.16 +0,3,3119.263,0.315,0.392,50.462,85,135.463,4.92,140.383 +0,4,3114.576,0.397,0.465,66.492,81.543,148.035,1.216,149.251 +0,9,3185.086,0.49,0.563,0.843,79.106,79.95,28.271,108.222 +0,6,3195.164,0.659,0.878,41.861,81.999,123.86,0.305,124.166 +0,8,3227.436,0.588,0.685,13.471,80.559,94.03,0.675,94.705 +0,0,3199.056,0.344,0.417,55.856,81.147,137.003,2.313,139.317 +0,2,3195.197,0.89,0.993,59.866,83.95,143.817,2.518,146.336 +0,1,3250.212,0.555,0.641,53.457,80.43,133.887,1.541,135.428 +0,5,3252.89,0.347,0.424,55.768,81.876,137.644,2.326,139.971 +0,9,3293.317,0.516,0.622,39.115,78.826,117.941,1.674,119.615 +0,2,3341.541,0.379,0.456,26.056,81.181,107.238,1.453,108.691 +0,4,3263.836,0.304,0.385,109.176,79.223,188.399,1.336,189.736 +0,6,3319.341,0.424,0.509,52.086,83.572,135.658,1.93,137.589 +0,3,3259.654,0.318,0.4,115.781,84.483,200.264,2.851,203.116 +0,9,3412.942,0.36,0.432,19.904,83.186,103.091,0.294,103.386 +0,5,3392.869,0.364,0.438,46.674,81.708,128.382,2.336,130.718 +0,7,3254.902,0.434,0.504,184.8,83.725,268.526,1.536,270.063 +0,0,3338.38,0.334,0.412,104.769,84.635,189.405,0.579,189.984 +0,8,3322.15,0.363,0.429,120.337,85.709,206.047,1.064,207.111 +0,3,3462.777,0.285,0.363,32.857,78.802,111.659,3.064,114.724 +0,2,3450.24,0.329,0.416,53.507,82.338,135.845,0.291,136.137 +0,1,3385.654,0.404,0.479,125.574,82.017,207.591,1.116,208.708 +0,6,3456.937,0.306,0.374,58.496,80.921,139.418,1.87,141.288 +0,4,3453.579,0.31,0.387,61.685,82.969,144.655,1.418,146.073 +0,8,3529.268,0.324,0.408,31.325,78.86,110.186,1.213,111.4 +0,5,3523.596,0.334,0.417,39.494,83.382,122.877,0.347,123.225 +0,7,3524.971,0.36,0.472,47.432,80.801,128.234,0.953,129.187 +0,4,3599.659,0.319,0.398,27.195,80.69,107.885,1.895,109.781 +0,3,3577.512,0.571,0.652,51.889,82.948,134.837,1.141,135.979 +0,1,3594.371,0.341,0.422,42.685,81.099,123.785,1.473,125.259 +0,7,3654.167,0.306,0.383,15.528,81.986,97.515,2.405,99.92 +0,9,3516.338,0.397,0.472,178.897,79.745,258.642,2.238,260.881 +0,2,3586.389,1.185,1.333,109.11,81.551,190.661,2.03,192.692 +0,5,3646.833,0.424,0.488,56.484,81.305,137.789,1.209,138.999 +0,0,3528.372,0.397,0.487,176.378,80.819,257.198,0.746,257.944 +0,6,3598.234,0.336,0.428,102.676,85.142,187.818,1.845,189.664 +0,8,3640.677,0.476,0.58,83.915,81.6,165.515,12.681,178.196 +0,4,3709.449,0.415,0.495,25.988,83.141,109.13,1.996,111.126 +0,3,3713.499,0.322,0.402,55.534,81.807,137.341,0.906,138.248 +0,0,3786.324,0.919,1.147,3.983,80.348,84.331,1.885,86.217 +0,7,3754.097,0.438,0.543,36.421,81.782,118.204,1.217,119.421 +0,9,3777.227,0.339,0.419,18.041,81.599,99.641,2.512,102.154 +0,1,3719.638,0.353,0.419,112.793,82.398,195.191,1.433,196.624 +0,4,3820.583,0.299,0.38,14.112,83.485,97.598,1.551,99.149 +0,6,3787.905,0.358,0.44,49.391,82.265,131.656,1.218,132.874 +0,2,3779.087,0.323,0.402,81.512,79.373,160.885,3.793,164.679 +0,5,3785.843,1.116,1.253,82.986,77.901,160.888,1.176,162.064 +0,8,3818.882,0.383,0.46,80.581,80.539,161.121,2.24,163.361 +0,9,3879.424,0.314,0.394,21.002,81.687,102.689,1.579,104.268 +0,6,3920.787,0.287,0.38,3.32,80.808,84.129,2.54,86.669 +0,7,3873.527,0.371,0.436,60.962,79.343,140.305,1.693,141.998 +0,5,3947.917,0.324,0.401,18.55,86.417,104.968,1.132,106.101 +0,3,3851.755,0.345,0.42,114.969,87.007,201.977,0.317,202.294 +0,9,3983.709,0.467,0.534,6.466,81.73,88.196,0.443,88.64 +0,8,3982.255,0.767,0.998,14.279,81.449,95.729,1.705,97.435 +0,4,3919.74,0.346,0.424,85.31,79.932,165.243,0.644,165.887 +0,7,4015.534,0.333,0.409,17.541,80.366,97.907,1.668,99.575 +0,1,3916.272,0.432,0.512,128.903,84.8,213.703,2.03,215.733 +0,0,3872.552,0.463,0.605,190.345,81.085,271.43,2.323,273.754 +0,2,3943.776,0.456,0.565,124.062,79.417,203.479,2.947,206.427 +0,5,4054.025,0.473,0.519,16.707,81.618,98.325,2.546,100.872 +0,4,4085.637,0.444,0.528,14.533,83.168,97.701,1.309,99.01 +0,7,4115.136,0.466,0.563,10.979,80.789,91.768,1.994,93.762 +0,9,4072.356,0.332,0.411,61.405,81.35,142.756,1.96,144.716 +0,6,4007.465,0.323,0.404,173.194,81.587,254.782,1.562,256.344 +0,0,4146.315,0.415,0.495,47.446,82.791,130.237,1.332,131.569 +0,3,4054.052,0.334,0.407,140.693,83.369,224.063,5.103,229.166 +0,8,4079.697,0.352,0.431,114.177,84.118,198.295,7.426,205.722 +0,4,4184.657,0.346,0.42,13.748,86.813,100.561,0.308,100.869 +0,2,4150.211,0.297,0.391,50.058,85.067,135.125,2.111,137.237 +0,9,4217.079,0.289,0.372,15.913,78.546,94.459,1.492,95.952 +0,7,4208.903,0.592,0.799,46.416,79.377,125.794,1.363,127.157 +0,5,4154.904,0.378,0.458,105.65,86.733,192.384,0.306,192.69 +0,1,4132.012,0.351,0.423,128.658,87.255,215.914,1.377,217.291 +0,6,4263.817,0.316,0.392,7.054,80.022,87.076,2.867,89.944 +0,2,4287.456,0.418,0.519,13.658,77.869,91.528,1.837,93.365 +0,8,4285.427,0.371,0.448,46.607,81.282,127.89,1.193,129.083 +0,6,4353.769,0.428,0.512,12.728,83.385,96.114,1.372,97.486 +0,9,4313.041,0.452,0.544,65.025,81.466,146.492,1.454,147.947 +0,7,4336.069,0.547,0.631,62.669,80.678,143.347,1.741,145.089 +0,4,4285.532,0.421,0.489,126.035,80.128,206.164,1.865,208.029 +0,1,4349.311,0.344,0.419,83.199,81.257,164.457,2.457,166.915 +0,5,4347.602,0.336,0.415,84.785,84.577,169.362,0.205,169.568 +0,3,4283.225,0.311,0.39,165.412,81.631,247.043,2.736,249.779 +0,6,4451.266,0.349,0.435,16.483,81.492,97.976,1.693,99.669 +0,2,4380.832,0.957,1.096,87.309,82.649,169.959,1.588,171.547 +0,8,4414.518,0.362,0.479,53.482,84.438,137.92,1.534,139.454 +0,0,4277.9,0.615,0.698,190.489,85.361,275.85,1.139,276.99 +0,4,4493.572,0.353,0.433,5.668,79.869,85.538,1.985,87.523 +0,9,4460.995,0.297,0.379,72.698,82.185,154.884,1.312,156.196 +0,7,4481.166,0.353,0.43,52.934,82.767,135.702,0.9,136.602 +0,1,4516.236,0.426,0.513,21.016,82.575,103.591,1.242,104.834 +0,0,4554.897,0.284,0.36,14.035,80.027,94.063,0.644,94.708 +0,2,4552.387,0.34,0.416,49.053,82.256,131.309,1.498,132.807 +0,6,4550.944,0.374,0.452,58.083,82.241,140.324,0.226,140.55 +0,5,4517.178,0.287,0.348,92.038,83.638,175.677,2.136,177.813 +0,3,4533.015,0.387,0.482,81.677,80.321,161.999,0.881,162.88 +0,8,4553.982,0.403,0.5,92.788,79.698,172.487,2.855,175.343 +0,0,4649.615,0.455,0.528,19.45,84.334,103.785,1.69,105.475 +0,4,4581.108,0.727,0.888,88.144,85.538,173.683,0.515,174.198 +0,7,4617.775,0.309,0.38,57.266,80.933,138.2,1.523,139.723 +0,9,4617.201,0.408,0.513,79.382,81.334,160.716,0.872,161.589 +0,1,4621.077,0.313,0.394,79.38,80.484,159.864,1.538,161.403 +0,5,4695,0.323,0.398,26.916,80.04,106.957,1.254,108.212 +0,8,4729.336,0.417,0.504,8.58,81.443,90.024,1.481,91.506 +0,6,4691.503,0.315,0.393,52.131,81.54,133.672,1.764,135.436 +0,7,4757.506,0.336,0.402,8.604,82.634,91.239,2.208,93.447 +0,3,4695.901,0.364,0.612,110.355,79.703,190.059,2.086,192.145 +0,4,4755.316,0.387,0.444,71.444,80.424,151.868,1.02,152.889 +0,1,4782.487,0.804,0.913,71.209,80.168,151.377,1.373,152.751 +0,2,4685.202,0.318,0.395,168.695,81.247,249.943,1.572,251.515 +0,0,4755.102,0.475,0.548,109.227,80.705,189.933,0.478,190.411 +0,7,4850.962,0.47,0.583,25.01,82.997,108.007,1.157,109.164 +0,5,4803.219,0.334,0.445,72.976,85.095,158.071,1.135,159.207 +0,9,4778.799,0.486,0.589,120.634,80.375,201.009,1.331,202.341 +0,4,4908.214,0.415,0.488,15.698,82.28,97.979,1.139,99.118 +0,6,4826.949,0.393,0.479,97.059,83.146,180.206,1.672,181.879 +0,8,4820.854,0.484,0.577,109.039,81.594,190.634,0.318,190.953 +0,2,4936.725,0.354,0.431,6.303,83.521,89.824,1.867,91.692 +0,3,4888.057,0.498,0.577,71.038,81.397,152.435,0.342,152.777 +0,5,4962.434,0.279,0.331,14.723,82.679,97.402,0.279,97.682 +0,7,4960.133,0.327,0.397,44.946,83.613,128.56,0.245,128.805 +0,3,5040.845,0.439,0.546,14.402,85.033,99.435,2.166,101.602 +0,8,5011.814,0.411,0.481,59.326,83.914,143.24,1.398,144.639 +0,2,5028.425,0.366,0.446,42.929,84.889,127.819,0.305,128.124 +0,0,4945.521,0.346,0.418,150.986,80.727,231.713,2.023,233.737 +0,9,4981.151,0.45,0.557,115.5,82.904,198.405,2.823,201.228 +0,5,5060.124,0.398,0.501,64.278,83.236,147.514,2.355,149.869 +0,6,5008.836,0.28,0.353,115.728,85.538,201.267,0.603,201.871 +0,1,4935.247,0.379,0.458,189.72,85.303,275.023,0.465,275.488 +0,4,5007.34,0.542,0.741,117.463,85.645,203.108,2.146,205.255 +0,7,5088.946,0.313,0.394,49.444,81.597,131.042,1.646,132.688 +0,3,5142.457,0.399,0.478,5.147,81.287,86.435,0.881,87.317 +0,2,5156.557,0.291,0.362,5.203,81.382,86.585,1.92,88.505 +0,8,5156.462,0.521,0.69,19.219,80.308,99.528,1.821,101.349 +0,9,5182.387,0.32,0.399,23.899,80.953,104.852,1.121,105.973 +0,7,5221.642,0.307,0.391,5.745,81.815,87.561,1.817,89.378 +0,0,5179.266,0.363,0.443,49.741,84.892,134.633,2.054,136.688 +0,5,5210.018,0.568,0.643,61.334,81.577,142.911,1.418,144.33 +0,3,5229.782,0.339,0.44,49.632,85.029,134.661,1.34,136.002 +0,1,5210.738,0.792,1.471,86.688,81.747,168.435,0.87,169.305 +0,9,5288.37,0.428,0.512,5.724,86.497,92.221,2.111,94.333 +0,2,5245.07,0.351,0.431,61.536,81.418,142.954,1.728,144.682 +0,0,5315.961,0.382,0.447,20.382,79.869,100.252,1.409,101.661 +0,8,5257.815,0.423,0.49,92.566,80.892,173.459,1.496,174.955 +0,7,5311.029,0.435,0.516,52.528,81.313,133.841,1.413,135.255 +0,6,5210.712,1.532,1.63,151.833,85.155,236.989,1.118,238.107 +0,3,5365.791,0.299,0.377,50.784,81.585,132.369,1.452,133.822 +0,2,5389.757,0.474,0.554,26.667,82.959,109.627,1.814,111.441 +0,0,5417.631,0.784,0.907,8.897,81.026,89.924,1.412,91.336 +0,4,5212.602,0.386,0.447,217.003,82.217,299.221,1.166,300.387 +0,9,5382.712,0.311,0.389,75.127,83.211,158.339,1.983,160.323 +0,5,5354.356,0.305,0.39,112.1,80.95,193.05,1.376,194.427 +0,7,5446.296,0.356,0.432,36.105,80.157,116.262,0.92,117.183 +0,1,5380.053,0.405,0.494,111.048,85.12,196.168,1.345,197.514 +0,6,5448.827,0.315,0.388,42.469,86.109,128.578,0.571,129.149 +0,8,5432.778,0.329,0.409,64.109,83.938,148.048,0.344,148.393 +0,4,5512.997,0.298,0.374,12.336,81.353,93.69,2.014,95.704 +0,3,5499.624,0.486,0.595,59.894,82.285,142.179,1.619,143.799 +0,5,5548.792,0.428,0.535,10.625,83.853,94.479,1.992,96.472 +0,2,5501.208,0.575,0.684,90.412,78.346,168.759,0.685,169.444 +0,4,5608.705,0.558,0.815,8.557,78.683,87.24,1.117,88.358 +0,1,5577.576,0.426,0.513,55.869,79.618,135.487,1.607,137.095 +0,0,5508.975,0.328,0.404,126.867,81.307,208.175,1.943,210.118 +0,9,5543.047,0.36,0.436,90.905,85.163,176.069,2.65,178.719 +0,2,5670.662,0.41,0.496,11.867,80.168,92.035,2.243,94.278 +0,7,5563.487,0.303,0.384,127.359,84.02,211.379,1.073,212.452 +0,6,5577.982,0.339,0.414,117.634,83.729,201.363,0.301,201.665 +0,5,5645.292,0.373,0.455,50.568,84.436,135.005,1.355,136.36 +0,4,5697.074,0.417,0.514,14.751,79.581,94.332,2.424,96.757 +0,3,5643.43,0.367,0.462,102.833,81.493,184.327,2.916,187.244 +0,0,5719.101,0.348,0.427,42.935,78.555,121.491,1.886,123.377 +0,8,5581.178,0.307,0.381,181.001,83.057,264.059,1.295,265.354 +0,9,5721.773,0.324,0.406,46.37,81.578,127.949,1.29,129.239 +0,2,5764.95,0.399,0.477,28.761,79.071,107.832,2.316,110.149 +0,1,5714.679,0.318,0.404,98.197,82.491,180.689,1.885,182.574 +0,4,5793.839,0.314,0.382,30.597,81.38,111.978,1.234,113.212 +0,7,5775.948,0.332,0.413,51.577,84.088,135.665,0.377,136.043 +0,3,5830.686,0.955,1.072,11.735,81.606,93.341,1.701,95.043 +0,9,5851.02,0.333,0.414,18.739,81.701,100.441,1.33,101.771 +0,2,5875.109,0.396,0.478,15.164,81.893,97.058,1.217,98.275 +0,5,5781.659,0.318,0.386,119.27,83.15,202.42,0.376,202.797 +0,7,5911.998,0.339,0.418,11.726,82.093,93.82,1.495,95.315 +0,3,5925.738,0.447,0.516,21.925,79.87,101.795,0.344,102.14 +0,0,5842.488,0.338,0.409,124.073,81.734,205.807,2.047,207.855 +0,6,5779.654,0.326,0.406,188.882,82.945,271.828,0.958,272.786 +0,1,5897.261,0.301,0.377,74.047,81.4,155.448,0.967,156.415 +0,8,5846.539,0.337,0.429,140.2,79.976,220.176,2.328,222.505 +0,9,5952.801,0.344,0.421,54.536,79.755,134.291,1.41,135.702 +0,7,6007.318,0.522,0.618,14.816,79.956,94.772,16.419,111.192 +0,2,5973.394,0.518,0.592,60.307,84.753,145.06,0.75,145.81 +0,5,5984.463,0.449,0.524,50.467,84.478,134.945,1.93,136.876 +0,4,5907.056,0.402,0.484,129.229,84.9,214.129,1.722,215.851 +0,3,6027.888,0.42,0.494,39.048,80.015,119.064,1.978,121.042 +0,0,6050.355,1.316,1.485,29.107,82.111,111.218,1.742,112.96 +0,1,6053.684,0.311,0.379,47.288,81.844,129.132,1.182,130.315 +0,6,6052.444,0.503,0.615,67.027,81.05,148.077,1.152,149.229 +0,5,6121.346,0.337,0.446,23.684,78.068,101.752,2.019,103.772 +0,8,6069.053,0.397,0.478,79.682,79.537,159.219,1.582,160.802 +0,9,6088.518,0.344,0.419,65.681,81.3,146.982,13.495,160.477 +0,7,6118.543,0.547,0.608,50.691,81.862,132.554,0.377,132.931 +0,3,6148.938,0.351,0.416,38.998,82.443,121.442,10.62,132.062 +0,1,6184.009,0.625,0.847,30.097,79.57,109.667,2.015,111.682 +0,4,6122.915,0.478,0.554,133.957,92.839,226.796,1.7,228.497 +0,2,6119.207,0.42,0.468,136.749,95.595,232.345,3.165,235.51 +0,5,6225.127,0.307,0.389,30.625,98.714,129.339,0.471,129.811 +0,8,6229.86,0.669,0.746,51.831,97.397,149.228,1.67,150.898 +0,6,6201.683,0.393,0.475,85.686,96.242,181.928,5.483,187.412 +0,9,6249.004,0.394,0.467,68.533,82.42,150.954,0.531,151.485 +0,0,6163.324,0.47,0.577,154.586,83.082,237.668,12.538,250.207 +0,7,6251.483,0.329,0.409,71.341,90.405,161.747,0.44,162.188 +0,3,6281.008,0.413,0.572,45.589,86.798,132.388,1.088,133.476 +0,1,6295.702,0.418,0.499,51.953,79.591,131.545,1.322,132.867 +0,4,6351.453,0.526,0.622,1.002,79.648,80.651,2.773,83.424 +0,2,6354.728,0.554,0.647,34.466,80.677,115.143,0.939,116.083 +0,5,6354.95,0.507,0.659,65.208,80.513,145.721,4.54,150.262 +0,6,6389.103,0.371,0.453,31.17,84.714,115.884,0.492,116.376 +0,0,6413.542,0.446,0.573,10.238,84.236,94.475,0.223,94.699 +0,1,6428.577,0.312,0.401,27.804,81.855,109.659,1.676,111.336 +0,7,6413.677,0.485,0.6,45.942,80.918,126.86,2.002,128.863 +0,2,6470.818,0.322,0.4,18.925,81.521,100.446,0.952,101.399 +0,4,6434.885,0.367,0.43,63.441,84.026,147.468,2.273,149.741 +0,0,6508.249,0.291,0.367,14.539,80.234,94.774,2.012,96.787 +0,9,6400.496,0.321,0.422,124.092,86.419,210.512,2.121,212.633 +0,6,6505.483,0.332,0.395,50.891,82.623,133.515,2.367,135.883 +0,5,6505.215,0.47,0.546,66.876,80.667,147.543,1.659,149.202 +0,2,6572.224,0.316,0.376,17.54,84.437,101.978,0.329,102.308 +0,4,6584.634,0.383,0.482,15.541,81.204,96.746,1.192,97.939 +0,9,6613.139,0.441,0.554,10.492,79.826,90.318,1.596,91.915 +0,7,6542.547,0.337,0.417,113.17,80.125,193.295,1.239,194.535 +0,0,6605.043,0.334,0.415,57.477,82.496,139.973,1.355,141.329 +0,3,6414.491,0.775,0.919,247.767,84.12,331.888,2.017,333.905 +0,8,6380.768,0.437,0.526,281.598,85.832,367.431,0.649,368.081 +0,5,6654.422,0.74,0.895,7.715,86.517,94.233,1.298,95.531 +0,1,6539.923,0.391,0.483,149.274,83.239,232.513,2.996,235.509 +0,4,6682.579,0.336,0.416,6.481,86.19,92.672,10.253,102.925 +0,6,6641.381,0.317,0.393,59.657,84.226,143.884,0.521,144.405 +0,9,6705.064,0.411,0.478,19.864,80.4,100.264,1.428,101.693 +0,2,6674.542,0.419,0.494,53.331,81.556,134.888,1.243,136.131 +0,7,6737.093,0.433,0.512,18.865,80.421,99.287,0.362,99.649 +0,8,6748.856,0.377,0.434,17.681,84.385,102.067,0.776,102.843 +0,6,6785.795,0.516,0.607,9.629,79.503,89.133,2.355,91.488 +0,5,6749.962,0.412,0.48,73.922,79.8,153.722,1.568,155.291 +0,0,6746.38,0.315,0.39,110.351,80.636,190.987,1.112,192.1 +0,7,6836.754,0.55,0.613,24.537,80.93,105.467,2.694,108.162 +0,8,6851.707,0.33,0.393,20.472,81.812,102.285,1.242,103.527 +0,6,6877.293,0.396,0.464,12.523,81.162,93.686,1.659,95.345 +0,4,6785.508,0.589,0.692,115.734,86.656,202.391,0.733,203.124 +0,1,6775.441,0.33,0.414,130.532,83.319,213.851,1.687,215.538 +0,9,6806.765,0.34,0.423,99.015,85.013,184.029,3.078,187.108 +0,3,6748.406,0.486,0.618,174.309,81.219,255.529,4.272,259.801 +0,5,6905.264,0.568,0.647,23.753,80.364,104.118,0.631,104.749 +0,8,6955.243,0.326,0.406,2.524,80.645,83.169,0.755,83.925 +0,0,6938.49,0.461,0.541,27.562,81.761,109.323,1.228,110.552 +0,2,6810.68,0.93,1.1,197.267,80.838,278.105,1.271,279.376 +0,7,6944.924,0.37,0.455,63.266,82.346,145.613,1.637,147.251 +0,4,6988.643,0.466,0.606,34.89,83.461,118.352,3.998,122.35 +0,3,7008.212,0.533,0.621,66.77,79.833,146.603,1.377,147.98 +0,0,7049.053,0.377,0.459,25.731,82.163,107.895,1.657,109.553 +0,6,6972.667,0.354,0.428,119.457,80.472,199.93,0.336,200.266 +0,7,7092.184,0.337,0.406,4.213,82.36,86.573,1.179,87.752 +0,9,6993.88,0.439,0.511,110.675,82.611,193.287,1.277,194.564 +0,1,6990.987,0.309,0.389,131.279,80.628,211.907,1.199,213.107 +0,8,7039.181,1.011,1.109,91.879,83.571,175.451,1.855,177.306 +0,2,7090.066,0.356,0.455,51.305,82.594,133.899,1.56,135.459 +0,4,7111.004,0.415,0.497,30.584,84.877,115.461,1.764,117.225 +0,5,7010.022,0.464,0.541,141.61,82.79,224.4,2.016,226.417 +0,0,7158.615,0.364,0.442,4.359,80.082,84.441,1.28,85.722 +0,6,7172.942,0.521,0.598,26.196,79.661,105.858,1.066,106.924 +0,3,7156.204,0.42,0.502,53.726,79.747,133.473,1.584,135.058 +0,1,7204.101,0.331,0.416,26.093,81.632,107.725,2,109.726 +0,0,7244.345,0.32,0.398,21.466,80.972,102.438,1.166,103.605 +0,5,7236.446,0.493,0.55,41.039,79.52,120.56,2.102,122.662 +0,9,7188.454,0.601,0.747,110.165,80.817,190.983,0.996,191.979 +0,8,7216.494,0.341,0.427,81.98,80.888,162.869,2.197,165.067 +0,7,7179.945,0.403,0.485,119.842,81.613,201.455,0.476,201.931 +0,2,7225.536,0.484,0.582,74.136,81.65,155.786,1.136,156.923 +0,1,7313.837,0.452,0.567,18.595,83.189,101.784,2.503,104.288 +0,6,7279.875,0.328,0.411,61.184,81.339,142.524,2.233,144.757 +0,5,7359.116,0.325,0.402,6.764,82.524,89.288,1.53,90.819 +0,4,7228.237,0.716,0.876,138.937,84.177,223.115,0.301,223.416 +0,7,7381.883,0.288,0.383,22.346,80.881,103.228,1.238,104.466 +0,3,7291.272,0.443,0.52,113.137,81.756,194.893,1.745,196.639 +0,8,7381.568,0.403,0.472,54.796,80.499,135.296,0.451,135.748 +0,0,7347.958,0.324,0.405,118.366,81.119,199.485,1.507,200.992 +0,9,7380.444,0.554,0.68,93.055,79.533,172.588,1.514,174.103 +0,4,7451.66,0.316,0.384,22.329,80.797,103.127,2.351,105.478 +0,3,7487.919,0.325,0.403,15.077,82.38,97.457,17.729,115.187 +0,1,7418.133,0.432,0.498,97.07,87.573,184.644,0.619,185.263 +0,8,7517.326,0.441,0.512,15.017,86.247,101.265,0.225,101.49 +0,6,7424.648,0.311,0.394,113.912,82.157,196.07,0.29,196.361 +0,5,7449.946,0.966,1.145,91.384,80.285,171.67,1.546,173.216 +0,2,7382.465,0.311,0.368,155.965,84.581,240.547,6.505,247.052 +0,7,7486.358,0.32,0.403,62.472,82.007,144.479,5.539,150.018 +0,4,7557.146,0.366,0.452,15.698,80.87,96.568,13.833,110.402 +0,0,7548.957,0.315,0.408,53.927,80.333,134.26,1.998,136.259 +0,9,7554.554,0.424,0.512,48.698,81.789,130.488,2.084,132.572 +0,3,7603.119,0.564,0.632,66.287,80.673,146.961,1.472,148.433 +0,6,7621.016,0.287,0.361,50.611,79.755,130.366,2.06,132.427 +0,1,7603.404,0.35,0.44,71.549,81.794,153.343,1.394,154.737 +0,7,7636.385,0.328,0.393,38.377,83.192,121.569,7.821,129.391 +0,8,7618.824,0.315,0.391,58.371,88.394,146.766,0.268,147.034 +0,4,7667.559,0.458,0.544,41.885,80.925,122.811,2.257,125.068 +0,0,7685.224,0.379,0.462,51.5,85.236,136.737,1.835,138.572 +0,5,7623.17,0.345,0.429,113.716,86.773,200.489,2.099,202.589 +0,9,7687.135,0.365,0.44,55.114,83.348,138.463,1.93,140.394 +0,2,7629.525,0.75,0.881,112.85,84.984,197.834,1.481,199.316 +0,6,7753.451,0.348,0.443,19.425,81.92,101.345,0.235,101.581 +0,8,7765.861,0.295,0.382,42.989,83.301,126.291,0.309,126.6 +0,3,7751.56,0.371,0.447,57.895,83.908,141.804,1.772,143.577 +0,1,7758.148,0.305,0.375,61.756,83.904,145.661,0.303,145.964 +0,4,7792.635,0.385,0.467,48.907,82.184,131.091,1.903,132.995 +0,7,7765.787,0.508,0.598,82.233,80.202,162.436,0.937,163.373 +0,5,7825.768,0.352,0.433,48.817,83.915,132.733,1.856,134.59 +0,0,7823.804,0.347,0.42,81.14,78.737,159.878,2.87,162.748 +0,3,7895.144,0.324,0.404,19.162,79.062,98.225,1.353,99.579 +0,1,7904.124,0.531,0.75,44.278,80.547,124.825,2.869,127.694 +0,7,7929.169,0.305,0.4,22.345,80.727,103.073,0.274,103.347 +0,9,7827.535,0.315,0.398,125.914,79.985,205.899,1.407,207.306 +0,4,7925.637,0.295,0.37,55.306,79.662,134.969,1.413,136.382 +0,6,7855.04,0.348,0.424,129.077,82.84,211.917,1.083,213 +0,2,7828.844,0.332,0.377,154.911,84.522,239.433,0.778,240.211 +0,8,7892.473,0.449,0.525,91.578,85.603,177.182,9.727,186.909 +0,5,7960.366,0.389,0.468,55.094,82.706,137.801,1.851,139.652 +0,3,7994.732,0.337,0.41,22.79,83.043,105.834,1.32,107.154 +0,7,8032.525,0.351,0.425,16.842,80.964,97.806,3.531,101.338 +0,0,7986.561,0.295,0.37,65.496,81.656,147.152,1.49,148.643 +0,9,8034.848,0.327,0.409,49.351,81.008,130.359,9.179,139.539 +0,4,8062.031,0.837,1.09,46.288,82.287,128.575,3.955,132.531 +0,5,8100.03,0.389,0.486,18.157,82.12,100.278,2.442,102.72 +0,1,8031.827,0.478,0.608,96.665,82.449,179.114,1.973,181.087 +0,2,8069.062,0.342,0.407,85.309,81.496,166.805,0.329,167.135 +0,9,8174.399,0.474,0.569,11.749,83.111,94.861,1.975,96.836 +0,3,8101.893,0.317,0.394,85.135,84.047,169.183,2.215,171.398 +0,8,8079.389,0.343,0.409,114.95,81.574,196.524,0.902,197.427 +0,1,8212.925,0.393,0.494,5.508,81.466,86.974,2.467,89.442 +0,6,8068.049,0.4,0.476,153.286,83.743,237.03,1.172,238.202 +0,4,8194.572,0.388,0.455,46.45,82.449,128.899,3.487,132.387 +0,7,8133.87,0.316,0.402,127.895,81.295,209.19,0.378,209.569 +0,0,8135.211,0.37,0.49,149.827,95.401,245.228,0.525,245.753 +0,5,8202.758,0.299,0.378,83.997,93.829,177.826,0.433,178.259 +0,2,8236.206,0.419,0.527,47.407,97.117,144.525,0.519,145.044 +0,6,8306.259,0.336,0.413,3.622,81.754,85.377,3.954,89.331 +0,9,8271.247,0.468,1.138,48.801,82.285,131.086,5.775,136.861 +0,7,8343.45,0.445,0.55,7.533,83.535,91.069,12.373,103.442 +0,1,8302.377,0.352,0.429,54.306,89.95,144.257,1.044,145.301 +0,8,8276.825,0.397,0.494,104.45,87.657,192.107,1.003,193.111 +0,3,8273.295,0.582,0.657,107.827,88.457,196.284,0.432,196.717 +0,4,8326.967,0.382,0.454,55.066,88.862,143.928,3.529,147.457 +0,9,8408.118,0.387,0.47,10.387,81.323,91.711,2.096,93.808 +0,5,8381.02,0.737,0.809,72.154,82.173,154.327,2.08,156.408 +0,6,8395.599,0.31,0.381,60.746,88.193,148.94,0.223,149.163 +0,1,8447.686,0.287,0.348,15.042,82.895,97.937,3.174,101.111 +0,0,8380.969,0.55,0.685,84.005,88.633,172.638,1.193,173.831 +0,4,8474.432,0.295,0.375,10.992,87.808,98.8,15.063,113.864 +0,8,8469.946,0.996,1.127,55.763,87.871,143.635,0.35,143.985 +0,0,8554.808,0.306,0.389,4.062,97.274,101.336,1.629,102.966 +0,9,8501.936,0.364,0.441,57.994,97.65,155.644,0.333,155.977 +0,5,8537.438,0.409,0.487,24.686,95.581,120.268,0.765,121.034 +0,7,8446.903,0.418,0.538,166.93,82.688,249.618,12.816,262.434 +0,3,8470.014,1.107,1.204,144.07,95.186,239.257,0.458,239.715 +0,1,8548.805,0.291,0.366,65.599,95.008,160.607,1.759,162.367 +0,2,8381.252,0.64,0.703,232.706,95.11,327.816,2.124,329.941 +0,4,8588.308,0.451,0.53,47.676,81.686,129.363,19.95,149.313 +0,6,8544.77,0.425,0.514,94.256,98.361,192.617,1.175,193.793 +0,8,8613.935,0.482,0.648,66.74,79.822,146.562,13.637,160.199 +0,5,8658.478,0.364,0.461,23.727,91.555,115.283,0.678,115.961 +0,0,8657.783,0.359,0.437,29.469,86.909,116.378,1.615,117.994 +0,9,8657.92,0.434,0.494,52.377,80.121,132.499,2.162,134.661 +0,7,8709.341,1.143,1.286,2.176,84.412,86.589,18.531,105.121 +0,0,8775.784,0.34,0.447,2.582,677.582,680.165,1.972,682.138 +0,4,8737.632,0.477,0.553,40.936,679.191,720.127,46.877,767.005 +0,1,8711.177,0.458,0.532,46.203,746.914,793.118,0.676,793.795 +0,2,8711.197,0.549,0.614,36.637,756.865,793.503,0.62,794.123 +0,3,8709.732,0.968,1.401,729.574,148.167,877.741,1.855,879.597 +0,9,8792.592,0.382,0.461,647.063,212.999,860.063,2.327,862.391 +0,5,8774.449,0.442,0.497,665.322,278.888,944.21,1.71,945.921 +0,8,8774.139,0.542,0.675,666.023,348.013,1014.037,3.225,1017.262 +0,7,8814.472,0.42,0.499,625.747,350.894,976.641,1.077,977.719 +0,3,9589.338,0.74,0.84,52.156,212.561,264.717,1.988,266.706 +0,1,9504.98,0.395,0.478,199.746,152.4,352.146,1.436,353.583 +0,2,9505.326,0.369,0.408,200.027,212.787,412.815,0.346,413.162 +0,5,9720.381,0.408,0.468,50.974,150.027,201.002,0.889,201.892 +0,9,9654.994,0.436,0.496,119.325,206.595,325.92,1.9,327.82 +0,7,9792.196,0.259,0.853,45.479,147.983,193.463,1.292,194.756 +0,8,9791.412,0.416,0.474,47.556,209.663,257.22,1.399,258.62 +0,1,9858.571,0.332,0.388,47.169,146.521,193.691,1.358,195.049 +0,6,8738.566,0.293,0.349,1167.354,156.58,1323.935,0.934,1324.869 +0,5,9922.281,0.382,0.438,49.221,145.247,194.469,1.392,195.862 +0,7,9986.962,0.364,0.42,47.021,150.528,197.549,1.334,198.884 +0,0,9457.939,0.316,0.395,578.457,150.278,728.735,1.254,729.99 +0,9,9982.825,0.43,0.5,64.168,199.563,263.731,0.345,264.077 +0,3,9856.054,0.416,0.472,247.039,147.388,394.428,1.72,396.148 +0,4,9504.641,0.548,0.619,664.554,141.019,805.573,1.417,806.991 +0,2,9918.499,0.404,0.459,250.86,142.241,393.101,1.155,394.257 diff --git a/vendor/github.com/aws/aws-sdk-go/awstesting/integration/smoke/secretsmanager/client.go b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/smoke/secretsmanager/client.go new file mode 100644 index 000000000..014fd62cb --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/smoke/secretsmanager/client.go @@ -0,0 +1,16 @@ +// +build integration + +//Package secretsmanager provides gucumber integration tests support. +package secretsmanager + +import ( + "github.com/aws/aws-sdk-go/awstesting/integration/smoke" + "github.com/aws/aws-sdk-go/service/secretsmanager" + "github.com/gucumber/gucumber" +) + +func init() { + gucumber.Before("@secretsmanager", func() { + gucumber.World["client"] = secretsmanager.New(smoke.Session) + }) +} diff --git a/vendor/github.com/aws/aws-sdk-go/awstesting/integration/smoke/secretsmanager/secretsmanager.feature b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/smoke/secretsmanager/secretsmanager.feature new file mode 100644 index 000000000..31824c86a --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/awstesting/integration/smoke/secretsmanager/secretsmanager.feature @@ -0,0 +1,13 @@ +# language: en +@smoke @secretsmanager +Feature: Amazon SecretsManager + + Scenario: Making a request + When I call the "ListSecrets" API + Then the request should be successful + + Scenario: Handling errors + When I attempt to call the "DescribeSecret" API with: + | SecretId | fake-secret-id | + Then I expect the response error code to be "ResourceNotFoundException" + diff --git a/vendor/github.com/aws/aws-sdk-go/awstesting/sandbox/Dockerfile.test.go1.10 b/vendor/github.com/aws/aws-sdk-go/awstesting/sandbox/Dockerfile.test.go1.10 new file mode 100644 index 000000000..7cf47942d --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/awstesting/sandbox/Dockerfile.test.go1.10 @@ -0,0 +1,11 @@ +FROM ubuntu:12.04 +FROM golang:1.10 + +ADD . /go/src/github.com/aws/aws-sdk-go + +RUN apt-get update && apt-get install -y --no-install-recommends \ + vim \ + && rm -rf /var/list/apt/lists/* + +WORKDIR /go/src/github.com/aws/aws-sdk-go +CMD ["make", "unit"] diff --git a/vendor/github.com/aws/aws-sdk-go/buildspec.yml b/vendor/github.com/aws/aws-sdk-go/buildspec.yml index 427208edf..2defb7f6c 100644 --- a/vendor/github.com/aws/aws-sdk-go/buildspec.yml +++ b/vendor/github.com/aws/aws-sdk-go/buildspec.yml @@ -10,7 +10,7 @@ phases: - mkdir -p /go/src/github.com/aws - ln -s $SDK_CB_ROOT $SDK_GO_ROOT - cd $SDK_GO_ROOT - - make unit + - make ci-test - cd $SDK_CB_ROOT - #echo Compiling the Go code... post_build: diff --git a/vendor/github.com/aws/aws-sdk-go/cleanup_models.sh b/vendor/github.com/aws/aws-sdk-go/cleanup_models.sh new file mode 100755 index 000000000..184ec40a9 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/cleanup_models.sh @@ -0,0 +1,9 @@ +#/usr/bin/env bash + +# Deletes all but the most current model version. +for v in $(ls ./models/apis | grep -v '.go' ); do + for vm in $(ls -r models/apis/$v/ | sed -n '1!p' ); do + echo "rm -r models/apis/$v/$vm" + rm -r models/apis/$v/$vm + done +done diff --git a/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/godoc.html b/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/godoc.html index fbaaa8111..979569685 100644 --- a/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/godoc.html +++ b/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/godoc.html @@ -158,7 +158,7 @@ window.onload = function() { - + diff --git a/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/style.css b/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/style.css index ce9d6cca1..2318e4a8d 100644 --- a/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/style.css +++ b/vendor/github.com/aws/aws-sdk-go/doc-src/aws-godoc/templates/style.css @@ -186,7 +186,11 @@ div#mobile_only { div#fixed { position: fixed; width: 100%; + max-width: 100%; height: 100%; + overflow-y: scroll; + overflow-x: hidden; + -webkit-overflow-scrolling: touch; } div .top_link { diff --git a/vendor/github.com/aws/aws-sdk-go/example/service/ec2/filterInstances/README.md b/vendor/github.com/aws/aws-sdk-go/example/service/ec2/filterInstances/README.md index 74716262e..91bf16dd8 100644 --- a/vendor/github.com/aws/aws-sdk-go/example/service/ec2/filterInstances/README.md +++ b/vendor/github.com/aws/aws-sdk-go/example/service/ec2/filterInstances/README.md @@ -5,10 +5,10 @@ This is an example using the AWS SDK for Go to list ec2 instances that match pro # Usage -The example uses the bucket name provided, and lists all object keys in a bucket. +The example uses the tag name provided, and lists all matching ec2 instances. ```sh -go run -tags example filter_ec2_by_tag.go +go run filter_ec2_by_tag.go ``` Output: diff --git a/vendor/github.com/aws/aws-sdk-go/example/service/s3/putObjectWithProcess/README.md b/vendor/github.com/aws/aws-sdk-go/example/service/s3/putObjectWithProcess/README.md new file mode 100644 index 000000000..c4665d9a9 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/example/service/s3/putObjectWithProcess/README.md @@ -0,0 +1,14 @@ +# Example + +This is an example using the AWS SDK for Go to upload object with progress. +We use CustomReader to implement it + + +# Usage + +The example uses the bucket name provided, one key for object, and output the progress to stdout. +The Object size should larger than 5M or your will not see the progress + +```sh +AWS_REGION= go run putObjWithProcess.go +``` diff --git a/vendor/github.com/aws/aws-sdk-go/example/service/s3/putObjectWithProcess/putObjWithProcess.go b/vendor/github.com/aws/aws-sdk-go/example/service/s3/putObjectWithProcess/putObjWithProcess.go new file mode 100644 index 000000000..4cdff6bbd --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/example/service/s3/putObjectWithProcess/putObjWithProcess.go @@ -0,0 +1,102 @@ +// +build example + +package main + +import ( + "log" + "os" + "sync/atomic" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/credentials" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/service/s3/s3manager" +) + +type CustomReader struct { + fp *os.File + size int64 + read int64 +} + +func (r *CustomReader) Read(p []byte) (int, error) { + return r.fp.Read(p) +} + +func (r *CustomReader) ReadAt(p []byte, off int64) (int, error) { + n, err := r.fp.ReadAt(p, off) + if err != nil { + return n, err + } + + // Got the length have read( or means has uploaded), and you can construct your message + atomic.AddInt64(&r.read, int64(n)) + + // I have no idea why the read length need to be div 2, + // maybe the request read once when Sign and actually send call ReadAt again + // It works for me + log.Printf("total read:%d progress:%d%%\n", r.read/2, int(float32(r.read*100/2)/float32(r.size))) + + return n, err +} + +func (r *CustomReader) Seek(offset int64, whence int) (int64, error) { + return r.fp.Seek(offset, whence) +} + +func main() { + if len(os.Args) < 4 { + log.Println("USAGE ERROR: AWS_REGION=us-east-1 go run putObjWithProcess.go ") + return + } + + credential := os.Args[1] + bucket := os.Args[2] + key := os.Args[3] + fileName := os.Args[4] + + creds := credentials.NewSharedCredentials(credential, "default") + if _, err := creds.Get(); err != nil { + log.Println("ERROR:", err) + return + } + + sess := session.New(&aws.Config{ + Credentials: creds, + }) + + file, err := os.Open(fileName) + if err != nil { + log.Println("ERROR:", err) + return + } + + fileInfo, err := file.Stat() + if err != nil { + log.Println("ERROR:", err) + return + } + + reader := &CustomReader{ + fp: file, + size: fileInfo.Size(), + } + + uploader := s3manager.NewUploader(sess, func(u *s3manager.Uploader) { + u.PartSize = 5 * 1024 * 1024 + u.LeavePartsOnError = true + }) + + output, err := uploader.Upload(&s3manager.UploadInput{ + Bucket: aws.String(bucket), + Key: aws.String(key), + Body: reader, + }) + + if err != nil { + log.Println("ERROR:", err) + return + } + + log.Println(output.Location) +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/AWSMigrationHub/2017-05-31/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/AWSMigrationHub/2017-05-31/docs-2.json index 2b2541bb3..5cb5869d5 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/AWSMigrationHub/2017-05-31/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/AWSMigrationHub/2017-05-31/docs-2.json @@ -17,7 +17,7 @@ "ListProgressUpdateStreams": "

Lists progress update streams associated with the user account making this call.

", "NotifyApplicationState": "

Sets the migration state of an application. For a given application identified by the value passed to ApplicationId, its status is set or updated by passing one of three values to Status: NOT_STARTED | IN_PROGRESS | COMPLETED.

", "NotifyMigrationTaskState": "

Notifies Migration Hub of the current status, progress, or other detail regarding a migration task. This API has the following traits:

  • Migration tools will call the NotifyMigrationTaskState API to share the latest progress and status.

  • MigrationTaskName is used for addressing updates to the correct target.

  • ProgressUpdateStream is used for access control and to provide a namespace for each migration tool.

", - "PutResourceAttributes": "

Provides identifying details of the resource being migrated so that it can be associated in the Application Discovery Service (ADS)'s repository. This association occurs asynchronously after PutResourceAttributes returns.

  • Keep in mind that subsequent calls to PutResourceAttributes will override previously stored attributes. For example, if it is first called with a MAC address, but later, it is desired to add an IP address, it will then be required to call it with both the IP and MAC addresses to prevent overiding the MAC address.

  • Note the instructions regarding the special use case of the ResourceAttributeList parameter when specifying any \"VM\" related value.

Because this is an asynchronous call, it will always return 200, whether an association occurs or not. To confirm if an association was found based on the provided details, call ListDiscoveredResources.

" + "PutResourceAttributes": "

Provides identifying details of the resource being migrated so that it can be associated in the Application Discovery Service (ADS)'s repository. This association occurs asynchronously after PutResourceAttributes returns.

  • Keep in mind that subsequent calls to PutResourceAttributes will override previously stored attributes. For example, if it is first called with a MAC address, but later, it is desired to add an IP address, it will then be required to call it with both the IP and MAC addresses to prevent overiding the MAC address.

  • Note the instructions regarding the special use case of the ResourceAttributeList parameter when specifying any \"VM\" related value.

Because this is an asynchronous call, it will always return 200, whether an association occurs or not. To confirm if an association was found based on the provided details, call ListDiscoveredResources.

" }, "shapes": { "AccessDeniedException": { @@ -414,7 +414,7 @@ "ResourceAttributeList": { "base": null, "refs": { - "PutResourceAttributesRequest$ResourceAttributeList": "

Information about the resource that is being migrated. This data will be used to map the task to a resource in the Application Discovery Service (ADS)'s repository.

In the ResourceAttribute object array, the Type field is reserved for the following values: IPV4_ADDRESS | IPV6_ADDRESS | MAC_ADDRESS | FQDN | VM_MANAGER_ID | VM_MANAGED_OBJECT_REFERENCE | VM_NAME | VM_PATH | BIOS_ID | MOTHERBOARD_SERIAL_NUMBER, and the identifying value can be a string up to 256 characters.

If any \"VM\" related value is used for a ResourceAttribute object, it is required that VM_MANAGER_ID, as a minimum, is always used. If it is not used, the server will not be associated in the Application Discovery Service (ADS)'s repository using any of the other \"VM\" related values, and you will experience data loss. See the Example section below for a use case of specifying \"VM\" related values.

" + "PutResourceAttributesRequest$ResourceAttributeList": "

Information about the resource that is being migrated. This data will be used to map the task to a resource in the Application Discovery Service (ADS)'s repository.

Takes the object array of ResourceAttribute where the Type field is reserved for the following values: IPV4_ADDRESS | IPV6_ADDRESS | MAC_ADDRESS | FQDN | VM_MANAGER_ID | VM_MANAGED_OBJECT_REFERENCE | VM_NAME | VM_PATH | BIOS_ID | MOTHERBOARD_SERIAL_NUMBER where the identifying value can be a string up to 256 characters.

  • If any \"VM\" related value is set for a ResourceAttribute object, it is required that VM_MANAGER_ID, as a minimum, is always set. If VM_MANAGER_ID is not set, then all \"VM\" fields will be discarded and \"VM\" fields will not be used for matching the migration task to a server in Application Discovery Service (ADS)'s repository. See the Example section below for a use case of specifying \"VM\" related values.

  • If a server you are trying to match has multiple IP or MAC addresses, you should provide as many as you know in separate type/value pairs passed to the ResourceAttributeList parameter to maximize the chances of matching.

" } }, "ResourceAttributeType": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/api-2.json index ecebfa0e5..d1b89f1d9 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/api-2.json @@ -83,6 +83,7 @@ "output":{"shape":"DescribeCertificateAuthorityAuditReportResponse"}, "errors":[ {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidArnException"}, {"shape":"InvalidArgsException"} ] }, @@ -128,7 +129,8 @@ {"shape":"RequestInProgressException"}, {"shape":"RequestFailedException"}, {"shape":"ResourceNotFoundException"}, - {"shape":"InvalidArnException"} + {"shape":"InvalidArnException"}, + {"shape":"InvalidStateException"} ] }, "ImportCertificateAuthorityCertificate":{ @@ -144,6 +146,7 @@ {"shape":"RequestFailedException"}, {"shape":"ResourceNotFoundException"}, {"shape":"InvalidArnException"}, + {"shape":"InvalidStateException"}, {"shape":"MalformedCertificateException"}, {"shape":"CertificateMismatchException"} ] @@ -191,6 +194,19 @@ {"shape":"InvalidArnException"} ] }, + "RestoreCertificateAuthority":{ + "name":"RestoreCertificateAuthority", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"RestoreCertificateAuthorityRequest"}, + "errors":[ + {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidStateException"}, + {"shape":"InvalidArnException"} + ] + }, "RevokeCertificate":{ "name":"RevokeCertificate", "http":{ @@ -218,6 +234,7 @@ "errors":[ {"shape":"ResourceNotFoundException"}, {"shape":"InvalidArnException"}, + {"shape":"InvalidStateException"}, {"shape":"InvalidTagException"}, {"shape":"TooManyTagsException"} ] @@ -232,6 +249,7 @@ "errors":[ {"shape":"ResourceNotFoundException"}, {"shape":"InvalidArnException"}, + {"shape":"InvalidStateException"}, {"shape":"InvalidTagException"} ] }, @@ -317,7 +335,8 @@ "NotAfter":{"shape":"TStamp"}, "FailureReason":{"shape":"FailureReason"}, "CertificateAuthorityConfiguration":{"shape":"CertificateAuthorityConfiguration"}, - "RevocationConfiguration":{"shape":"RevocationConfiguration"} + "RevocationConfiguration":{"shape":"RevocationConfiguration"}, + "RestorableUntil":{"shape":"TStamp"} } }, "CertificateAuthorityConfiguration":{ @@ -339,6 +358,7 @@ "CREATING", "PENDING_CERTIFICATE", "ACTIVE", + "DELETED", "DISABLED", "EXPIRED", "FAILED" @@ -443,7 +463,8 @@ "type":"structure", "required":["CertificateAuthorityArn"], "members":{ - "CertificateAuthorityArn":{"shape":"Arn"} + "CertificateAuthorityArn":{"shape":"Arn"}, + "PermanentDeletionTimeInDays":{"shape":"PermanentDeletionTimeInDays"} } }, "DescribeCertificateAuthorityAuditReportRequest":{ @@ -696,6 +717,11 @@ "max":500, "min":1 }, + "PermanentDeletionTimeInDays":{ + "type":"integer", + "max":30, + "min":7 + }, "PositiveLong":{ "type":"long", "min":1 @@ -728,6 +754,13 @@ }, "exception":true }, + "RestoreCertificateAuthorityRequest":{ + "type":"structure", + "required":["CertificateAuthorityArn"], + "members":{ + "CertificateAuthorityArn":{"shape":"Arn"} + } + }, "RevocationConfiguration":{ "type":"structure", "members":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/docs-2.json index 3ee8ca8fa..ed5836460 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/docs-2.json @@ -1,574 +1,588 @@ { - "version": "2.0", - "service": "

You can use the ACM PCA API to create a private certificate authority (CA). You must first call the CreateCertificateAuthority function. If successful, the function returns an Amazon Resource Name (ARN) for your private CA. Use this ARN as input to the GetCertificateAuthorityCsr function to retrieve the certificate signing request (CSR) for your private CA certificate. Sign the CSR using the root or an intermediate CA in your on-premises PKI hierarchy, and call the ImportCertificateAuthorityCertificate to import your signed private CA certificate into ACM PCA.

Use your private CA to issue and revoke certificates. These are private certificates that identify and secure client computers, servers, applications, services, devices, and users over SSLS/TLS connections within your organization. Call the IssueCertificate function to issue a certificate. Call the RevokeCertificate function to revoke a certificate.

Certificates issued by your private CA can be trusted only within your organization, not publicly.

Your private CA can optionally create a certificate revocation list (CRL) to track the certificates you revoke. To create a CRL, you must specify a RevocationConfiguration object when you call the CreateCertificateAuthority function. ACM PCA writes the CRL to an S3 bucket that you specify. You must specify a bucket policy that grants ACM PCA write permission.

You can also call the CreateCertificateAuthorityAuditReport to create an optional audit report that lists every time the CA private key is used. The private key is used for signing when the IssueCertificate or RevokeCertificate function is called.

", - "operations": { - "CreateCertificateAuthority": "

Creates a private subordinate certificate authority (CA). You must specify the CA configuration, the revocation configuration, the CA type, and an optional idempotency token. The CA configuration specifies the name of the algorithm and key size to be used to create the CA private key, the type of signing algorithm that the CA uses to sign, and X.500 subject information. The CRL (certificate revocation list) configuration specifies the CRL expiration period in days (the validity period of the CRL), the Amazon S3 bucket that will contain the CRL, and a CNAME alias for the S3 bucket that is included in certificates issued by the CA. If successful, this function returns the Amazon Resource Name (ARN) of the CA.

", - "CreateCertificateAuthorityAuditReport": "

Creates an audit report that lists every time that the your CA private key is used. The report is saved in the Amazon S3 bucket that you specify on input. The IssueCertificate and RevokeCertificate functions use the private key. You can generate a new report every 30 minutes.

", - "DeleteCertificateAuthority": "

Deletes the private certificate authority (CA) that you created or started to create by calling the CreateCertificateAuthority function. This action requires that you enter an ARN (Amazon Resource Name) for the private CA that you want to delete. You can find the ARN by calling the ListCertificateAuthorities function. You can delete the CA if you are waiting for it to be created (the Status field of the CertificateAuthority is CREATING) or if the CA has been created but you haven't yet imported the signed certificate (the Status is PENDING_CERTIFICATE) into ACM PCA. If you've already imported the certificate, you cannot delete the CA unless it has been disabled for more than 30 days. To disable a CA, call the UpdateCertificateAuthority function and set the CertificateAuthorityStatus argument to DISABLED.

", - "DescribeCertificateAuthority": "

Lists information about your private certificate authority (CA). You specify the private CA on input by its ARN (Amazon Resource Name). The output contains the status of your CA. This can be any of the following:

  • CREATING: ACM PCA is creating your private certificate authority.

  • PENDING_CERTIFICATE: The certificate is pending. You must use your on-premises root or subordinate CA to sign your private CA CSR and then import it into PCA.

  • ACTIVE: Your private CA is active.

  • DISABLED: Your private CA has been disabled.

  • EXPIRED: Your private CA certificate has expired.

  • FAILED: Your private CA has failed. Your CA can fail for problems such a network outage or backend AWS failure or other errors. A failed CA can never return to the pending state. You must create a new CA.

", - "DescribeCertificateAuthorityAuditReport": "

Lists information about a specific audit report created by calling the CreateCertificateAuthorityAuditReport function. Audit information is created every time the certificate authority (CA) private key is used. The private key is used when you call the IssueCertificate function or the RevokeCertificate function.

", - "GetCertificate": "

Retrieves a certificate from your private CA. The ARN of the certificate is returned when you call the IssueCertificate function. You must specify both the ARN of your private CA and the ARN of the issued certificate when calling the GetCertificate function. You can retrieve the certificate if it is in the ISSUED state. You can call the CreateCertificateAuthorityAuditReport function to create a report that contains information about all of the certificates issued and revoked by your private CA.

", - "GetCertificateAuthorityCertificate": "

Retrieves the certificate and certificate chain for your private certificate authority (CA). Both the certificate and the chain are base64 PEM-encoded. The chain does not include the CA certificate. Each certificate in the chain signs the one before it.

", - "GetCertificateAuthorityCsr": "

Retrieves the certificate signing request (CSR) for your private certificate authority (CA). The CSR is created when you call the CreateCertificateAuthority function. Take the CSR to your on-premises X.509 infrastructure and sign it by using your root or a subordinate CA. Then import the signed certificate back into ACM PCA by calling the ImportCertificateAuthorityCertificate function. The CSR is returned as a base64 PEM-encoded string.

", - "ImportCertificateAuthorityCertificate": "

Imports your signed private CA certificate into ACM PCA. Before you can call this function, you must create the private certificate authority by calling the CreateCertificateAuthority function. You must then generate a certificate signing request (CSR) by calling the GetCertificateAuthorityCsr function. Take the CSR to your on-premises CA and use the root certificate or a subordinate certificate to sign it. Create a certificate chain and copy the signed certificate and the certificate chain to your working directory.

Your certificate chain must not include the private CA certificate that you are importing.

Your on-premises CA certificate must be the last certificate in your chain. The subordinate certificate, if any, that your root CA signed must be next to last. The subordinate certificate signed by the preceding subordinate CA must come next, and so on until your chain is built.

The chain must be PEM-encoded.

", - "IssueCertificate": "

Uses your private certificate authority (CA) to issue a client certificate. This function returns the Amazon Resource Name (ARN) of the certificate. You can retrieve the certificate by calling the GetCertificate function and specifying the ARN.

You cannot use the ACM ListCertificateAuthorities function to retrieve the ARNs of the certificates that you issue by using ACM PCA.

", - "ListCertificateAuthorities": "

Lists the private certificate authorities that you created by using the CreateCertificateAuthority function.

", - "ListTags": "

Lists the tags, if any, that are associated with your private CA. Tags are labels that you can use to identify and organize your CAs. Each tag consists of a key and an optional value. Call the TagCertificateAuthority function to add one or more tags to your CA. Call the UntagCertificateAuthority function to remove tags.

", - "RevokeCertificate": "

Revokes a certificate that you issued by calling the IssueCertificate function. If you enable a certificate revocation list (CRL) when you create or update your private CA, information about the revoked certificates will be included in the CRL. ACM PCA writes the CRL to an S3 bucket that you specify. For more information about revocation, see the CrlConfiguration structure. ACM PCA also writes revocation information to the audit report. For more information, see CreateCertificateAuthorityAuditReport.

", - "TagCertificateAuthority": "

Adds one or more tags to your private CA. Tags are labels that you can use to identify and organize your AWS resources. Each tag consists of a key and an optional value. You specify the private CA on input by its Amazon Resource Name (ARN). You specify the tag by using a key-value pair. You can apply a tag to just one private CA if you want to identify a specific characteristic of that CA, or you can apply the same tag to multiple private CAs if you want to filter for a common relationship among those CAs. To remove one or more tags, use the UntagCertificateAuthority function. Call the ListTags function to see what tags are associated with your CA.

", - "UntagCertificateAuthority": "

Remove one or more tags from your private CA. A tag consists of a key-value pair. If you do not specify the value portion of the tag when calling this function, the tag will be removed regardless of value. If you specify a value, the tag is removed only if it is associated with the specified value. To add tags to a private CA, use the TagCertificateAuthority. Call the ListTags function to see what tags are associated with your CA.

", - "UpdateCertificateAuthority": "

Updates the status or configuration of a private certificate authority (CA). Your private CA must be in the ACTIVE or DISABLED state before you can update it. You can disable a private CA that is in the ACTIVE state or make a CA that is in the DISABLED state active again.

" + "version":"2.0", + "service":"

You can use the ACM PCA API to create a private certificate authority (CA). You must first call the CreateCertificateAuthority operation. If successful, the operation returns an Amazon Resource Name (ARN) for your private CA. Use this ARN as input to the GetCertificateAuthorityCsr operation to retrieve the certificate signing request (CSR) for your private CA certificate. Sign the CSR using the root or an intermediate CA in your on-premises PKI hierarchy, and call the ImportCertificateAuthorityCertificate to import your signed private CA certificate into ACM PCA.

Use your private CA to issue and revoke certificates. These are private certificates that identify and secure client computers, servers, applications, services, devices, and users over SSLS/TLS connections within your organization. Call the IssueCertificate operation to issue a certificate. Call the RevokeCertificate operation to revoke a certificate.

Certificates issued by your private CA can be trusted only within your organization, not publicly.

Your private CA can optionally create a certificate revocation list (CRL) to track the certificates you revoke. To create a CRL, you must specify a RevocationConfiguration object when you call the CreateCertificateAuthority operation. ACM PCA writes the CRL to an S3 bucket that you specify. You must specify a bucket policy that grants ACM PCA write permission.

You can also call the CreateCertificateAuthorityAuditReport to create an optional audit report that lists every time the CA private key is used. The private key is used for signing when the IssueCertificate or RevokeCertificate operation is called.

", + "operations":{ + "CreateCertificateAuthority":"

Creates a private subordinate certificate authority (CA). You must specify the CA configuration, the revocation configuration, the CA type, and an optional idempotency token. The CA configuration specifies the name of the algorithm and key size to be used to create the CA private key, the type of signing algorithm that the CA uses to sign, and X.500 subject information. The CRL (certificate revocation list) configuration specifies the CRL expiration period in days (the validity period of the CRL), the Amazon S3 bucket that will contain the CRL, and a CNAME alias for the S3 bucket that is included in certificates issued by the CA. If successful, this operation returns the Amazon Resource Name (ARN) of the CA.

", + "CreateCertificateAuthorityAuditReport":"

Creates an audit report that lists every time that the your CA private key is used. The report is saved in the Amazon S3 bucket that you specify on input. The IssueCertificate and RevokeCertificate operations use the private key. You can generate a new report every 30 minutes.

", + "DeleteCertificateAuthority":"

Deletes a private certificate authority (CA). You must provide the ARN (Amazon Resource Name) of the private CA that you want to delete. You can find the ARN by calling the ListCertificateAuthorities operation. Before you can delete a CA, you must disable it. Call the UpdateCertificateAuthority operation and set the CertificateAuthorityStatus parameter to DISABLED.

Additionally, you can delete a CA if you are waiting for it to be created (the Status field of the CertificateAuthority is CREATING). You can also delete it if the CA has been created but you haven't yet imported the signed certificate (the Status is PENDING_CERTIFICATE) into ACM PCA.

If the CA is in one of the aforementioned states and you call DeleteCertificateAuthority, the CA's status changes to DELETED. However, the CA won't be permentantly deleted until the restoration period has passed. By default, if you do not set the PermanentDeletionTimeInDays parameter, the CA remains restorable for 30 days. You can set the parameter from 7 to 30 days. The DescribeCertificateAuthority operation returns the time remaining in the restoration window of a Private CA in the DELETED state. To restore an eligable CA, call the RestoreCertificateAuthority operation.

", + "DescribeCertificateAuthority":"

Lists information about your private certificate authority (CA). You specify the private CA on input by its ARN (Amazon Resource Name). The output contains the status of your CA. This can be any of the following:

  • CREATING - ACM PCA is creating your private certificate authority.

  • PENDING_CERTIFICATE - The certificate is pending. You must use your on-premises root or subordinate CA to sign your private CA CSR and then import it into PCA.

  • ACTIVE - Your private CA is active.

  • DISABLED - Your private CA has been disabled.

  • EXPIRED - Your private CA certificate has expired.

  • FAILED - Your private CA has failed. Your CA can fail because of problems such a network outage or backend AWS failure or other errors. A failed CA can never return to the pending state. You must create a new CA.

  • DELETED - Your private CA is within the restoration period, after which it will be permanently deleted. The length of time remaining in the CA's restoration period will also be included in this operation's output.

", + "DescribeCertificateAuthorityAuditReport":"

Lists information about a specific audit report created by calling the CreateCertificateAuthorityAuditReport operation. Audit information is created every time the certificate authority (CA) private key is used. The private key is used when you call the IssueCertificate operation or the RevokeCertificate operation.

", + "GetCertificate":"

Retrieves a certificate from your private CA. The ARN of the certificate is returned when you call the IssueCertificate operation. You must specify both the ARN of your private CA and the ARN of the issued certificate when calling the GetCertificate operation. You can retrieve the certificate if it is in the ISSUED state. You can call the CreateCertificateAuthorityAuditReport operation to create a report that contains information about all of the certificates issued and revoked by your private CA.

", + "GetCertificateAuthorityCertificate":"

Retrieves the certificate and certificate chain for your private certificate authority (CA). Both the certificate and the chain are base64 PEM-encoded. The chain does not include the CA certificate. Each certificate in the chain signs the one before it.

", + "GetCertificateAuthorityCsr":"

Retrieves the certificate signing request (CSR) for your private certificate authority (CA). The CSR is created when you call the CreateCertificateAuthority operation. Take the CSR to your on-premises X.509 infrastructure and sign it by using your root or a subordinate CA. Then import the signed certificate back into ACM PCA by calling the ImportCertificateAuthorityCertificate operation. The CSR is returned as a base64 PEM-encoded string.

", + "ImportCertificateAuthorityCertificate":"

Imports your signed private CA certificate into ACM PCA. Before you can call this operation, you must create the private certificate authority by calling the CreateCertificateAuthority operation. You must then generate a certificate signing request (CSR) by calling the GetCertificateAuthorityCsr operation. Take the CSR to your on-premises CA and use the root certificate or a subordinate certificate to sign it. Create a certificate chain and copy the signed certificate and the certificate chain to your working directory.

Your certificate chain must not include the private CA certificate that you are importing.

Your on-premises CA certificate must be the last certificate in your chain. The subordinate certificate, if any, that your root CA signed must be next to last. The subordinate certificate signed by the preceding subordinate CA must come next, and so on until your chain is built.

The chain must be PEM-encoded.

", + "IssueCertificate":"

Uses your private certificate authority (CA) to issue a client certificate. This operation returns the Amazon Resource Name (ARN) of the certificate. You can retrieve the certificate by calling the GetCertificate operation and specifying the ARN.

You cannot use the ACM ListCertificateAuthorities operation to retrieve the ARNs of the certificates that you issue by using ACM PCA.

", + "ListCertificateAuthorities":"

Lists the private certificate authorities that you created by using the CreateCertificateAuthority operation.

", + "ListTags":"

Lists the tags, if any, that are associated with your private CA. Tags are labels that you can use to identify and organize your CAs. Each tag consists of a key and an optional value. Call the TagCertificateAuthority operation to add one or more tags to your CA. Call the UntagCertificateAuthority operation to remove tags.

", + "RestoreCertificateAuthority":"

Restores a certificate authority (CA) that is in the DELETED state. You can restore a CA during the period that you defined in the PermanentDeletionTimeInDays parameter of the DeleteCertificateAuthority operation. Currently, you can specify 7 to 30 days. If you did not specify a PermanentDeletionTimeInDays value, by default you can restore the CA at any time in a 30 day period. You can check the time remaining in the restoration period of a private CA in the DELETED state by calling the DescribeCertificateAuthority or ListCertificateAuthorities operations. The status of a restored CA is set to its pre-deletion status when the RestoreCertificateAuthority operation returns. To change its status to ACTIVE, call the UpdateCertificateAuthority operation. If the private CA was in the PENDING_CERTIFICATE state at deletion, you must use the ImportCertificateAuthorityCertificate operation to import a certificate authority into the private CA before it can be activated. You cannot restore a CA after the restoration period has ended.

", + "RevokeCertificate":"

Revokes a certificate that you issued by calling the IssueCertificate operation. If you enable a certificate revocation list (CRL) when you create or update your private CA, information about the revoked certificates will be included in the CRL. ACM PCA writes the CRL to an S3 bucket that you specify. For more information about revocation, see the CrlConfiguration structure. ACM PCA also writes revocation information to the audit report. For more information, see CreateCertificateAuthorityAuditReport.

", + "TagCertificateAuthority":"

Adds one or more tags to your private CA. Tags are labels that you can use to identify and organize your AWS resources. Each tag consists of a key and an optional value. You specify the private CA on input by its Amazon Resource Name (ARN). You specify the tag by using a key-value pair. You can apply a tag to just one private CA if you want to identify a specific characteristic of that CA, or you can apply the same tag to multiple private CAs if you want to filter for a common relationship among those CAs. To remove one or more tags, use the UntagCertificateAuthority operation. Call the ListTags operation to see what tags are associated with your CA.

", + "UntagCertificateAuthority":"

Remove one or more tags from your private CA. A tag consists of a key-value pair. If you do not specify the value portion of the tag when calling this operation, the tag will be removed regardless of value. If you specify a value, the tag is removed only if it is associated with the specified value. To add tags to a private CA, use the TagCertificateAuthority. Call the ListTags operation to see what tags are associated with your CA.

", + "UpdateCertificateAuthority":"

Updates the status or configuration of a private certificate authority (CA). Your private CA must be in the ACTIVE or DISABLED state before you can update it. You can disable a private CA that is in the ACTIVE state or make a CA that is in the DISABLED state active again.

" }, - "shapes": { - "ASN1Subject": { - "base": "

Contains information about the certificate subject. The certificate can be one issued by your private certificate authority (CA) or it can be your private CA certificate. The Subject field in the certificate identifies the entity that owns or controls the public key in the certificate. The entity can be a user, computer, device, or service. The Subject must contain an X.500 distinguished name (DN). A DN is a sequence of relative distinguished names (RDNs). The RDNs are separated by commas in the certificate. The DN must be unique for each for each entity, but your private CA can issue more than one certificate with the same DN to the same entity.

", - "refs": { - "CertificateAuthorityConfiguration$Subject": "

Structure that contains X.500 distinguished name information for your private CA.

" + "shapes":{ + "ASN1Subject":{ + "base":"

Contains information about the certificate subject. The certificate can be one issued by your private certificate authority (CA) or it can be your private CA certificate. The Subject field in the certificate identifies the entity that owns or controls the public key in the certificate. The entity can be a user, computer, device, or service. The Subject must contain an X.500 distinguished name (DN). A DN is a sequence of relative distinguished names (RDNs). The RDNs are separated by commas in the certificate. The DN must be unique for each entity, but your private CA can issue more than one certificate with the same DN to the same entity.

", + "refs":{ + "CertificateAuthorityConfiguration$Subject":"

Structure that contains X.500 distinguished name information for your private CA.

" } }, - "Arn": { - "base": null, - "refs": { - "CertificateAuthority$Arn": "

Amazon Resource Name (ARN) for your private certificate authority (CA). The format is 12345678-1234-1234-1234-123456789012 .

", - "CreateCertificateAuthorityAuditReportRequest$CertificateAuthorityArn": "

Amazon Resource Name (ARN) of the CA to be audited. This is of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "CreateCertificateAuthorityResponse$CertificateAuthorityArn": "

If successful, the Amazon Resource Name (ARN) of the certificate authority (CA). This is of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "DeleteCertificateAuthorityRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "DescribeCertificateAuthorityAuditReportRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) of the private CA. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "DescribeCertificateAuthorityRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "GetCertificateAuthorityCertificateRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) of your private CA. This is of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "GetCertificateAuthorityCsrRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority function. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "GetCertificateRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", - "GetCertificateRequest$CertificateArn": "

The ARN of the issued certificate. The ARN contains the certificate serial number and must be in the following form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245

", - "ImportCertificateAuthorityCertificateRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "IssueCertificateRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "IssueCertificateResponse$CertificateArn": "

The Amazon Resource Name (ARN) of the issued certificate and the certificate serial number. This is of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245

", - "ListTagsRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority function. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "RevokeCertificateRequest$CertificateAuthorityArn": "

Amazon Resource Name (ARN) of the private CA that issued the certificate to be revoked. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "TagCertificateAuthorityRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "UntagCertificateAuthorityRequest$CertificateAuthorityArn": "

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", - "UpdateCertificateAuthorityRequest$CertificateAuthorityArn": "

Amazon Resource Name (ARN) of the private CA that issued the certificate to be revoked. This must be of the form:

arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

" + "Arn":{ + "base":null, + "refs":{ + "CertificateAuthority$Arn":"

Amazon Resource Name (ARN) for your private certificate authority (CA). The format is 12345678-1234-1234-1234-123456789012 .

", + "CreateCertificateAuthorityAuditReportRequest$CertificateAuthorityArn":"

Amazon Resource Name (ARN) of the CA to be audited. This is of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "CreateCertificateAuthorityResponse$CertificateAuthorityArn":"

If successful, the Amazon Resource Name (ARN) of the certificate authority (CA). This is of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "DeleteCertificateAuthorityRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must have the following form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "DescribeCertificateAuthorityAuditReportRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) of the private CA. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "DescribeCertificateAuthorityRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "GetCertificateAuthorityCertificateRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) of your private CA. This is of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "GetCertificateAuthorityCsrRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority operation. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "GetCertificateRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 .

", + "GetCertificateRequest$CertificateArn":"

The ARN of the issued certificate. The ARN contains the certificate serial number and must be in the following form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245

", + "ImportCertificateAuthorityCertificateRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "IssueCertificateRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "IssueCertificateResponse$CertificateArn":"

The Amazon Resource Name (ARN) of the issued certificate and the certificate serial number. This is of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245

", + "ListTagsRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority operation. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "RestoreCertificateAuthorityRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority operation. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "RevokeCertificateRequest$CertificateAuthorityArn":"

Amazon Resource Name (ARN) of the private CA that issued the certificate to be revoked. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "TagCertificateAuthorityRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "UntagCertificateAuthorityRequest$CertificateAuthorityArn":"

The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

", + "UpdateCertificateAuthorityRequest$CertificateAuthorityArn":"

Amazon Resource Name (ARN) of the private CA that issued the certificate to be revoked. This must be of the form:

arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012

" } }, - "AuditReportId": { - "base": null, - "refs": { - "CreateCertificateAuthorityAuditReportResponse$AuditReportId": "

An alphanumeric string that contains a report identifier.

", - "DescribeCertificateAuthorityAuditReportRequest$AuditReportId": "

The report ID returned by calling the CreateCertificateAuthorityAuditReport function.

" + "AuditReportId":{ + "base":null, + "refs":{ + "CreateCertificateAuthorityAuditReportResponse$AuditReportId":"

An alphanumeric string that contains a report identifier.

", + "DescribeCertificateAuthorityAuditReportRequest$AuditReportId":"

The report ID returned by calling the CreateCertificateAuthorityAuditReport operation.

" } }, - "AuditReportResponseFormat": { - "base": null, - "refs": { - "CreateCertificateAuthorityAuditReportRequest$AuditReportResponseFormat": "

Format in which to create the report. This can be either JSON or CSV.

" + "AuditReportResponseFormat":{ + "base":null, + "refs":{ + "CreateCertificateAuthorityAuditReportRequest$AuditReportResponseFormat":"

Format in which to create the report. This can be either JSON or CSV.

" } }, - "AuditReportStatus": { - "base": null, - "refs": { - "DescribeCertificateAuthorityAuditReportResponse$AuditReportStatus": "

Specifies whether report creation is in progress, has succeeded, or has failed.

" + "AuditReportStatus":{ + "base":null, + "refs":{ + "DescribeCertificateAuthorityAuditReportResponse$AuditReportStatus":"

Specifies whether report creation is in progress, has succeeded, or has failed.

" } }, - "Boolean": { - "base": null, - "refs": { - "CrlConfiguration$Enabled": "

Boolean value that specifies whether certificate revocation lists (CRLs) are enabled. You can use this value to enable certificate revocation for a new CA when you call the CreateCertificateAuthority function or for an existing CA when you call the UpdateCertificateAuthority function.

" + "Boolean":{ + "base":null, + "refs":{ + "CrlConfiguration$Enabled":"

Boolean value that specifies whether certificate revocation lists (CRLs) are enabled. You can use this value to enable certificate revocation for a new CA when you call the CreateCertificateAuthority operation or for an existing CA when you call the UpdateCertificateAuthority operation.

" } }, - "CertificateAuthorities": { - "base": null, - "refs": { - "ListCertificateAuthoritiesResponse$CertificateAuthorities": "

Summary information about each certificate authority you have created.

" + "CertificateAuthorities":{ + "base":null, + "refs":{ + "ListCertificateAuthoritiesResponse$CertificateAuthorities":"

Summary information about each certificate authority you have created.

" } }, - "CertificateAuthority": { - "base": "

Contains information about your private certificate authority (CA). Your private CA can issue and revoke X.509 digital certificates. Digital certificates verify that the entity named in the certificate Subject field owns or controls the public key contained in the Subject Public Key Info field. Call the CreateCertificateAuthority function to create your private CA. You must then call the GetCertificateAuthorityCertificate function to retrieve a private CA certificate signing request (CSR). Take the CSR to your on-premises CA and sign it with the root CA certificate or a subordinate certificate. Call the ImportCertificateAuthorityCertificate function to import the signed certificate into AWS Certificate Manager (ACM).

", - "refs": { - "CertificateAuthorities$member": null, - "DescribeCertificateAuthorityResponse$CertificateAuthority": "

A CertificateAuthority structure that contains information about your private CA.

" + "CertificateAuthority":{ + "base":"

Contains information about your private certificate authority (CA). Your private CA can issue and revoke X.509 digital certificates. Digital certificates verify that the entity named in the certificate Subject field owns or controls the public key contained in the Subject Public Key Info field. Call the CreateCertificateAuthority operation to create your private CA. You must then call the GetCertificateAuthorityCertificate operation to retrieve a private CA certificate signing request (CSR). Take the CSR to your on-premises CA and sign it with the root CA certificate or a subordinate certificate. Call the ImportCertificateAuthorityCertificate operation to import the signed certificate into AWS Certificate Manager (ACM).

", + "refs":{ + "CertificateAuthorities$member":null, + "DescribeCertificateAuthorityResponse$CertificateAuthority":"

A CertificateAuthority structure that contains information about your private CA.

" } }, - "CertificateAuthorityConfiguration": { - "base": "

Contains configuration information for your private certificate authority (CA). This includes information about the class of public key algorithm and the key pair that your private CA creates when it issues a certificate, the signature algorithm it uses used when issuing certificates, and its X.500 distinguished name. You must specify this information when you call the CreateCertificateAuthority function.

", - "refs": { - "CertificateAuthority$CertificateAuthorityConfiguration": "

Your private CA configuration.

", - "CreateCertificateAuthorityRequest$CertificateAuthorityConfiguration": "

Name and bit size of the private key algorithm, the name of the signing algorithm, and X.500 certificate subject information.

" + "CertificateAuthorityConfiguration":{ + "base":"

Contains configuration information for your private certificate authority (CA). This includes information about the class of public key algorithm and the key pair that your private CA creates when it issues a certificate, the signature algorithm it uses used when issuing certificates, and its X.500 distinguished name. You must specify this information when you call the CreateCertificateAuthority operation.

", + "refs":{ + "CertificateAuthority$CertificateAuthorityConfiguration":"

Your private CA configuration.

", + "CreateCertificateAuthorityRequest$CertificateAuthorityConfiguration":"

Name and bit size of the private key algorithm, the name of the signing algorithm, and X.500 certificate subject information.

" } }, - "CertificateAuthorityStatus": { - "base": null, - "refs": { - "CertificateAuthority$Status": "

Status of your private CA.

", - "UpdateCertificateAuthorityRequest$Status": "

Status of your private CA.

" + "CertificateAuthorityStatus":{ + "base":null, + "refs":{ + "CertificateAuthority$Status":"

Status of your private CA.

", + "UpdateCertificateAuthorityRequest$Status":"

Status of your private CA.

" } }, - "CertificateAuthorityType": { - "base": null, - "refs": { - "CertificateAuthority$Type": "

Type of your private CA.

", - "CreateCertificateAuthorityRequest$CertificateAuthorityType": "

The type of the certificate authority. Currently, this must be SUBORDINATE.

" + "CertificateAuthorityType":{ + "base":null, + "refs":{ + "CertificateAuthority$Type":"

Type of your private CA.

", + "CreateCertificateAuthorityRequest$CertificateAuthorityType":"

The type of the certificate authority. Currently, this must be SUBORDINATE.

" } }, - "CertificateBody": { - "base": null, - "refs": { - "GetCertificateAuthorityCertificateResponse$Certificate": "

Base64-encoded certificate authority (CA) certificate.

", - "GetCertificateResponse$Certificate": "

The base64 PEM-encoded certificate specified by the CertificateArn parameter.

" + "CertificateBody":{ + "base":null, + "refs":{ + "GetCertificateAuthorityCertificateResponse$Certificate":"

Base64-encoded certificate authority (CA) certificate.

", + "GetCertificateResponse$Certificate":"

The base64 PEM-encoded certificate specified by the CertificateArn parameter.

" } }, - "CertificateBodyBlob": { - "base": null, - "refs": { - "ImportCertificateAuthorityCertificateRequest$Certificate": "

The PEM-encoded certificate for your private CA. This must be signed by using your on-premises CA.

" + "CertificateBodyBlob":{ + "base":null, + "refs":{ + "ImportCertificateAuthorityCertificateRequest$Certificate":"

The PEM-encoded certificate for your private CA. This must be signed by using your on-premises CA.

" } }, - "CertificateChain": { - "base": null, - "refs": { - "GetCertificateAuthorityCertificateResponse$CertificateChain": "

Base64-encoded certificate chain that includes any intermediate certificates and chains up to root on-premises certificate that you used to sign your private CA certificate. The chain does not include your private CA certificate.

", - "GetCertificateResponse$CertificateChain": "

The base64 PEM-encoded certificate chain that chains up to the on-premises root CA certificate that you used to sign your private CA certificate.

" + "CertificateChain":{ + "base":null, + "refs":{ + "GetCertificateAuthorityCertificateResponse$CertificateChain":"

Base64-encoded certificate chain that includes any intermediate certificates and chains up to root on-premises certificate that you used to sign your private CA certificate. The chain does not include your private CA certificate.

", + "GetCertificateResponse$CertificateChain":"

The base64 PEM-encoded certificate chain that chains up to the on-premises root CA certificate that you used to sign your private CA certificate.

" } }, - "CertificateChainBlob": { - "base": null, - "refs": { - "ImportCertificateAuthorityCertificateRequest$CertificateChain": "

A PEM-encoded file that contains all of your certificates, other than the certificate you're importing, chaining up to your root CA. Your on-premises root certificate is the last in the chain, and each certificate in the chain signs the one preceding.

" + "CertificateChainBlob":{ + "base":null, + "refs":{ + "ImportCertificateAuthorityCertificateRequest$CertificateChain":"

A PEM-encoded file that contains all of your certificates, other than the certificate you're importing, chaining up to your root CA. Your on-premises root certificate is the last in the chain, and each certificate in the chain signs the one preceding.

" } }, - "CertificateMismatchException": { - "base": "

The certificate authority certificate you are importing does not comply with conditions specified in the certificate that signed it.

", - "refs": { + "CertificateMismatchException":{ + "base":"

The certificate authority certificate you are importing does not comply with conditions specified in the certificate that signed it.

", + "refs":{ } }, - "ConcurrentModificationException": { - "base": "

A previous update to your private CA is still ongoing.

", - "refs": { + "ConcurrentModificationException":{ + "base":"

A previous update to your private CA is still ongoing.

", + "refs":{ } }, - "CountryCodeString": { - "base": null, - "refs": { - "ASN1Subject$Country": "

Two digit code that specifies the country in which the certificate subject located.

" + "CountryCodeString":{ + "base":null, + "refs":{ + "ASN1Subject$Country":"

Two-digit code that specifies the country in which the certificate subject located.

" } }, - "CreateCertificateAuthorityAuditReportRequest": { - "base": null, - "refs": { + "CreateCertificateAuthorityAuditReportRequest":{ + "base":null, + "refs":{ } }, - "CreateCertificateAuthorityAuditReportResponse": { - "base": null, - "refs": { + "CreateCertificateAuthorityAuditReportResponse":{ + "base":null, + "refs":{ } }, - "CreateCertificateAuthorityRequest": { - "base": null, - "refs": { + "CreateCertificateAuthorityRequest":{ + "base":null, + "refs":{ } }, - "CreateCertificateAuthorityResponse": { - "base": null, - "refs": { + "CreateCertificateAuthorityResponse":{ + "base":null, + "refs":{ } }, - "CrlConfiguration": { - "base": "

Contains configuration information for a certificate revocation list (CRL). Your private certificate authority (CA) creates base CRLs. Delta CRLs are not supported. You can enable CRLs for your new or an existing private CA by setting the Enabled parameter to true. Your private CA writes CRLs to an S3 bucket that you specify in the S3BucketName parameter. You can hide the name of your bucket by specifying a value for the CustomCname parameter. Your private CA copies the CNAME or the S3 bucket name to the CRL Distribution Points extension of each certificate it issues. Your S3 bucket policy must give write permission to ACM PCA.

Your private CA uses the value in the ExpirationInDays parameter to calculate the nextUpdate field in the CRL. The CRL is refreshed at 1/2 the age of next update or when a certificate is revoked. When a certificate is revoked, it is recorded in the next CRL that is generated and in the next audit report. Only time valid certificates are listed in the CRL. Expired certificates are not included.

CRLs contain the following fields:

  • Version: The current version number defined in RFC 5280 is V2. The integer value is 0x1.

  • Signature Algorithm: The name of the algorithm used to sign the CRL.

  • Issuer: The X.500 distinguished name of your private CA that issued the CRL.

  • Last Update: The issue date and time of this CRL.

  • Next Update: The day and time by which the next CRL will be issued.

  • Revoked Certificates: List of revoked certificates. Each list item contains the following information.

    • Serial Number: The serial number, in hexadecimal format, of the revoked certificate.

    • Revocation Date: Date and time the certificate was revoked.

    • CRL Entry Extensions: Optional extensions for the CRL entry.

      • X509v3 CRL Reason Code: Reason the certificate was revoked.

  • CRL Extensions: Optional extensions for the CRL.

    • X509v3 Authority Key Identifier: Identifies the public key associated with the private key used to sign the certificate.

    • X509v3 CRL Number:: Decimal sequence number for the CRL.

  • Signature Algorithm: Algorithm used by your private CA to sign the CRL.

  • Signature Value: Signature computed over the CRL.

Certificate revocation lists created by ACM PCA are DER-encoded. You can use the following OpenSSL command to list a CRL.

openssl crl -inform DER -text -in crl_path -noout

", - "refs": { - "RevocationConfiguration$CrlConfiguration": "

Configuration of the certificate revocation list (CRL), if any, maintained by your private CA.

" + "CrlConfiguration":{ + "base":"

Contains configuration information for a certificate revocation list (CRL). Your private certificate authority (CA) creates base CRLs. Delta CRLs are not supported. You can enable CRLs for your new or an existing private CA by setting the Enabled parameter to true. Your private CA writes CRLs to an S3 bucket that you specify in the S3BucketName parameter. You can hide the name of your bucket by specifying a value for the CustomCname parameter. Your private CA copies the CNAME or the S3 bucket name to the CRL Distribution Points extension of each certificate it issues. Your S3 bucket policy must give write permission to ACM PCA.

Your private CA uses the value in the ExpirationInDays parameter to calculate the nextUpdate field in the CRL. The CRL is refreshed at 1/2 the age of next update or when a certificate is revoked. When a certificate is revoked, it is recorded in the next CRL that is generated and in the next audit report. Only time valid certificates are listed in the CRL. Expired certificates are not included.

CRLs contain the following fields:

  • Version: The current version number defined in RFC 5280 is V2. The integer value is 0x1.

  • Signature Algorithm: The name of the algorithm used to sign the CRL.

  • Issuer: The X.500 distinguished name of your private CA that issued the CRL.

  • Last Update: The issue date and time of this CRL.

  • Next Update: The day and time by which the next CRL will be issued.

  • Revoked Certificates: List of revoked certificates. Each list item contains the following information.

    • Serial Number: The serial number, in hexadecimal format, of the revoked certificate.

    • Revocation Date: Date and time the certificate was revoked.

    • CRL Entry Extensions: Optional extensions for the CRL entry.

      • X509v3 CRL Reason Code: Reason the certificate was revoked.

  • CRL Extensions: Optional extensions for the CRL.

    • X509v3 Authority Key Identifier: Identifies the public key associated with the private key used to sign the certificate.

    • X509v3 CRL Number:: Decimal sequence number for the CRL.

  • Signature Algorithm: Algorithm used by your private CA to sign the CRL.

  • Signature Value: Signature computed over the CRL.

Certificate revocation lists created by ACM PCA are DER-encoded. You can use the following OpenSSL command to list a CRL.

openssl crl -inform DER -text -in crl_path -noout

", + "refs":{ + "RevocationConfiguration$CrlConfiguration":"

Configuration of the certificate revocation list (CRL), if any, maintained by your private CA.

" } }, - "CsrBlob": { - "base": null, - "refs": { - "IssueCertificateRequest$Csr": "

The certificate signing request (CSR) for the certificate you want to issue. You can use the following OpenSSL command to create the CSR and a 2048 bit RSA private key.

openssl req -new -newkey rsa:2048 -days 365 -keyout private/test_cert_priv_key.pem -out csr/test_cert_.csr

If you have a configuration file, you can use the following OpenSSL command. The usr_cert block in the configuration file contains your X509 version 3 extensions.

openssl req -new -config openssl_rsa.cnf -extensions usr_cert -newkey rsa:2048 -days -365 -keyout private/test_cert_priv_key.pem -out csr/test_cert_.csr

" + "CsrBlob":{ + "base":null, + "refs":{ + "IssueCertificateRequest$Csr":"

The certificate signing request (CSR) for the certificate you want to issue. You can use the following OpenSSL command to create the CSR and a 2048 bit RSA private key.

openssl req -new -newkey rsa:2048 -days 365 -keyout private/test_cert_priv_key.pem -out csr/test_cert_.csr

If you have a configuration file, you can use the following OpenSSL command. The usr_cert block in the configuration file contains your X509 version 3 extensions.

openssl req -new -config openssl_rsa.cnf -extensions usr_cert -newkey rsa:2048 -days -365 -keyout private/test_cert_priv_key.pem -out csr/test_cert_.csr

" } }, - "CsrBody": { - "base": null, - "refs": { - "GetCertificateAuthorityCsrResponse$Csr": "

The base64 PEM-encoded certificate signing request (CSR) for your private CA certificate.

" + "CsrBody":{ + "base":null, + "refs":{ + "GetCertificateAuthorityCsrResponse$Csr":"

The base64 PEM-encoded certificate signing request (CSR) for your private CA certificate.

" } }, - "DeleteCertificateAuthorityRequest": { - "base": null, - "refs": { + "DeleteCertificateAuthorityRequest":{ + "base":null, + "refs":{ } }, - "DescribeCertificateAuthorityAuditReportRequest": { - "base": null, - "refs": { + "DescribeCertificateAuthorityAuditReportRequest":{ + "base":null, + "refs":{ } }, - "DescribeCertificateAuthorityAuditReportResponse": { - "base": null, - "refs": { + "DescribeCertificateAuthorityAuditReportResponse":{ + "base":null, + "refs":{ } }, - "DescribeCertificateAuthorityRequest": { - "base": null, - "refs": { + "DescribeCertificateAuthorityRequest":{ + "base":null, + "refs":{ } }, - "DescribeCertificateAuthorityResponse": { - "base": null, - "refs": { + "DescribeCertificateAuthorityResponse":{ + "base":null, + "refs":{ } }, - "DistinguishedNameQualifierString": { - "base": null, - "refs": { - "ASN1Subject$DistinguishedNameQualifier": "

Disambiguating information for the certificate subject.

" + "DistinguishedNameQualifierString":{ + "base":null, + "refs":{ + "ASN1Subject$DistinguishedNameQualifier":"

Disambiguating information for the certificate subject.

" } }, - "FailureReason": { - "base": null, - "refs": { - "CertificateAuthority$FailureReason": "

Reason the request to create your private CA failed.

" + "FailureReason":{ + "base":null, + "refs":{ + "CertificateAuthority$FailureReason":"

Reason the request to create your private CA failed.

" } }, - "GetCertificateAuthorityCertificateRequest": { - "base": null, - "refs": { + "GetCertificateAuthorityCertificateRequest":{ + "base":null, + "refs":{ } }, - "GetCertificateAuthorityCertificateResponse": { - "base": null, - "refs": { + "GetCertificateAuthorityCertificateResponse":{ + "base":null, + "refs":{ } }, - "GetCertificateAuthorityCsrRequest": { - "base": null, - "refs": { + "GetCertificateAuthorityCsrRequest":{ + "base":null, + "refs":{ } }, - "GetCertificateAuthorityCsrResponse": { - "base": null, - "refs": { + "GetCertificateAuthorityCsrResponse":{ + "base":null, + "refs":{ } }, - "GetCertificateRequest": { - "base": null, - "refs": { + "GetCertificateRequest":{ + "base":null, + "refs":{ } }, - "GetCertificateResponse": { - "base": null, - "refs": { + "GetCertificateResponse":{ + "base":null, + "refs":{ } }, - "IdempotencyToken": { - "base": null, - "refs": { - "CreateCertificateAuthorityRequest$IdempotencyToken": "

Alphanumeric string that can be used to distinguish between calls to CreateCertificateAuthority. Idempotency tokens time out after five minutes. Therefore, if you call CreateCertificateAuthority multiple times with the same idempotency token within a five minute period, ACM PCA recognizes that you are requesting only one certificate and will issue only one. If you change the idempotency token for each call, however, ACM PCA recognizes that you are requesting multiple certificates.

", - "IssueCertificateRequest$IdempotencyToken": "

Custom string that can be used to distinguish between calls to the IssueCertificate function. Idempotency tokens time out after one hour. Therefore, if you call IssueCertificate multiple times with the same idempotency token within 5 minutes, ACM PCA recognizes that you are requesting only one certificate and will issue only one. If you change the idempotency token for each call, PCA recognizes that you are requesting multiple certificates.

" + "IdempotencyToken":{ + "base":null, + "refs":{ + "CreateCertificateAuthorityRequest$IdempotencyToken":"

Alphanumeric string that can be used to distinguish between calls to CreateCertificateAuthority. Idempotency tokens time out after five minutes. Therefore, if you call CreateCertificateAuthority multiple times with the same idempotency token within a five minute period, ACM PCA recognizes that you are requesting only one certificate. As a result, ACM PCA issues only one. If you change the idempotency token for each call, however, ACM PCA recognizes that you are requesting multiple certificates.

", + "IssueCertificateRequest$IdempotencyToken":"

Custom string that can be used to distinguish between calls to the IssueCertificate operation. Idempotency tokens time out after one hour. Therefore, if you call IssueCertificate multiple times with the same idempotency token within 5 minutes, ACM PCA recognizes that you are requesting only one certificate and will issue only one. If you change the idempotency token for each call, PCA recognizes that you are requesting multiple certificates.

" } }, - "ImportCertificateAuthorityCertificateRequest": { - "base": null, - "refs": { + "ImportCertificateAuthorityCertificateRequest":{ + "base":null, + "refs":{ } }, - "Integer1To5000": { - "base": null, - "refs": { - "CrlConfiguration$ExpirationInDays": "

Number of days until a certificate expires.

" + "Integer1To5000":{ + "base":null, + "refs":{ + "CrlConfiguration$ExpirationInDays":"

Number of days until a certificate expires.

" } }, - "InvalidArgsException": { - "base": "

One or more of the specified arguments was not valid.

", - "refs": { + "InvalidArgsException":{ + "base":"

One or more of the specified arguments was not valid.

", + "refs":{ } }, - "InvalidArnException": { - "base": "

The requested Amazon Resource Name (ARN) does not refer to an existing resource.

", - "refs": { + "InvalidArnException":{ + "base":"

The requested Amazon Resource Name (ARN) does not refer to an existing resource.

", + "refs":{ } }, - "InvalidNextTokenException": { - "base": "

The token specified in the NextToken argument is not valid. Use the token returned from your previous call to ListCertificateAuthorities.

", - "refs": { + "InvalidNextTokenException":{ + "base":"

The token specified in the NextToken argument is not valid. Use the token returned from your previous call to ListCertificateAuthorities.

", + "refs":{ } }, - "InvalidPolicyException": { - "base": "

The S3 bucket policy is not valid. The policy must give ACM PCA rights to read from and write to the bucket and find the bucket location.

", - "refs": { + "InvalidPolicyException":{ + "base":"

The S3 bucket policy is not valid. The policy must give ACM PCA rights to read from and write to the bucket and find the bucket location.

", + "refs":{ } }, - "InvalidStateException": { - "base": "

The private CA is in a state during which a report cannot be generated.

", - "refs": { + "InvalidStateException":{ + "base":"

The private CA is in a state during which a report cannot be generated.

", + "refs":{ } }, - "InvalidTagException": { - "base": "

The tag associated with the CA is not valid. The invalid argument is contained in the message field.

", - "refs": { + "InvalidTagException":{ + "base":"

The tag associated with the CA is not valid. The invalid argument is contained in the message field.

", + "refs":{ } }, - "IssueCertificateRequest": { - "base": null, - "refs": { + "IssueCertificateRequest":{ + "base":null, + "refs":{ } }, - "IssueCertificateResponse": { - "base": null, - "refs": { + "IssueCertificateResponse":{ + "base":null, + "refs":{ } }, - "KeyAlgorithm": { - "base": null, - "refs": { - "CertificateAuthorityConfiguration$KeyAlgorithm": "

Type of the public key algorithm and size, in bits, of the key pair that your key pair creates when it issues a certificate.

" + "KeyAlgorithm":{ + "base":null, + "refs":{ + "CertificateAuthorityConfiguration$KeyAlgorithm":"

Type of the public key algorithm and size, in bits, of the key pair that your key pair creates when it issues a certificate.

" } }, - "LimitExceededException": { - "base": "

An ACM PCA limit has been exceeded. See the exception message returned to determine the limit that was exceeded.

", - "refs": { + "LimitExceededException":{ + "base":"

An ACM PCA limit has been exceeded. See the exception message returned to determine the limit that was exceeded.

", + "refs":{ } }, - "ListCertificateAuthoritiesRequest": { - "base": null, - "refs": { + "ListCertificateAuthoritiesRequest":{ + "base":null, + "refs":{ } }, - "ListCertificateAuthoritiesResponse": { - "base": null, - "refs": { + "ListCertificateAuthoritiesResponse":{ + "base":null, + "refs":{ } }, - "ListTagsRequest": { - "base": null, - "refs": { + "ListTagsRequest":{ + "base":null, + "refs":{ } }, - "ListTagsResponse": { - "base": null, - "refs": { + "ListTagsResponse":{ + "base":null, + "refs":{ } }, - "MalformedCSRException": { - "base": "

The certificate signing request is invalid.

", - "refs": { + "MalformedCSRException":{ + "base":"

The certificate signing request is invalid.

", + "refs":{ } }, - "MalformedCertificateException": { - "base": "

One or more fields in the certificate are invalid.

", - "refs": { + "MalformedCertificateException":{ + "base":"

One or more fields in the certificate are invalid.

", + "refs":{ } }, - "MaxResults": { - "base": null, - "refs": { - "ListCertificateAuthoritiesRequest$MaxResults": "

Use this parameter when paginating results to specify the maximum number of items to return in the response on each page. If additional items exist beyond the number you specify, the NextToken element is sent in the response. Use this NextToken value in a subsequent request to retrieve additional items.

", - "ListTagsRequest$MaxResults": "

Use this parameter when paginating results to specify the maximum number of items to return in the response. If additional items exist beyond the number you specify, the NextToken element is sent in the response. Use this NextToken value in a subsequent request to retrieve additional items.

" + "MaxResults":{ + "base":null, + "refs":{ + "ListCertificateAuthoritiesRequest$MaxResults":"

Use this parameter when paginating results to specify the maximum number of items to return in the response on each page. If additional items exist beyond the number you specify, the NextToken element is sent in the response. Use this NextToken value in a subsequent request to retrieve additional items.

", + "ListTagsRequest$MaxResults":"

Use this parameter when paginating results to specify the maximum number of items to return in the response. If additional items exist beyond the number you specify, the NextToken element is sent in the response. Use this NextToken value in a subsequent request to retrieve additional items.

" } }, - "NextToken": { - "base": null, - "refs": { - "ListCertificateAuthoritiesRequest$NextToken": "

Use this parameter when paginating results in a subsequent request after you receive a response with truncated results. Set it to the value of the NextToken parameter from the response you just received.

", - "ListCertificateAuthoritiesResponse$NextToken": "

When the list is truncated, this value is present and should be used for the NextToken parameter in a subsequent pagination request.

", - "ListTagsRequest$NextToken": "

Use this parameter when paginating results in a subsequent request after you receive a response with truncated results. Set it to the value of NextToken from the response you just received.

", - "ListTagsResponse$NextToken": "

When the list is truncated, this value is present and should be used for the NextToken parameter in a subsequent pagination request.

" + "NextToken":{ + "base":null, + "refs":{ + "ListCertificateAuthoritiesRequest$NextToken":"

Use this parameter when paginating results in a subsequent request after you receive a response with truncated results. Set it to the value of the NextToken parameter from the response you just received.

", + "ListCertificateAuthoritiesResponse$NextToken":"

When the list is truncated, this value is present and should be used for the NextToken parameter in a subsequent pagination request.

", + "ListTagsRequest$NextToken":"

Use this parameter when paginating results in a subsequent request after you receive a response with truncated results. Set it to the value of NextToken from the response you just received.

", + "ListTagsResponse$NextToken":"

When the list is truncated, this value is present and should be used for the NextToken parameter in a subsequent pagination request.

" } }, - "PositiveLong": { - "base": null, - "refs": { - "Validity$Value": "

Time period.

" + "PermanentDeletionTimeInDays":{ + "base":null, + "refs":{ + "DeleteCertificateAuthorityRequest$PermanentDeletionTimeInDays":"

The number of days to make a CA restorable after it has been deleted. This can be anywhere from 7 to 30 days, with 30 being the default.

" } }, - "RequestAlreadyProcessedException": { - "base": "

Your request has already been completed.

", - "refs": { + "PositiveLong":{ + "base":null, + "refs":{ + "Validity$Value":"

Time period.

" } }, - "RequestFailedException": { - "base": "

The request has failed for an unspecified reason.

", - "refs": { + "RequestAlreadyProcessedException":{ + "base":"

Your request has already been completed.

", + "refs":{ } }, - "RequestInProgressException": { - "base": "

Your request is already in progress.

", - "refs": { + "RequestFailedException":{ + "base":"

The request has failed for an unspecified reason.

", + "refs":{ } }, - "ResourceNotFoundException": { - "base": "

A resource such as a private CA, S3 bucket, certificate, or audit report cannot be found.

", - "refs": { + "RequestInProgressException":{ + "base":"

Your request is already in progress.

", + "refs":{ } }, - "RevocationConfiguration": { - "base": "

Certificate revocation information used by the CreateCertificateAuthority and UpdateCertificateAuthority functions. Your private certificate authority (CA) can create and maintain a certificate revocation list (CRL). A CRL contains information about certificates revoked by your CA. For more information, see RevokeCertificate.

", - "refs": { - "CertificateAuthority$RevocationConfiguration": "

Information about the certificate revocation list (CRL) created and maintained by your private CA.

", - "CreateCertificateAuthorityRequest$RevocationConfiguration": "

Contains a Boolean value that you can use to enable a certification revocation list (CRL) for the CA, the name of the S3 bucket to which ACM PCA will write the CRL, and an optional CNAME alias that you can use to hide the name of your bucket in the CRL Distribution Points extension of your CA certificate. For more information, see the CrlConfiguration structure.

", - "UpdateCertificateAuthorityRequest$RevocationConfiguration": "

Revocation information for your private CA.

" + "ResourceNotFoundException":{ + "base":"

A resource such as a private CA, S3 bucket, certificate, or audit report cannot be found.

", + "refs":{ } }, - "RevocationReason": { - "base": null, - "refs": { - "RevokeCertificateRequest$RevocationReason": "

Specifies why you revoked the certificate.

" + "RestoreCertificateAuthorityRequest":{ + "base":null, + "refs":{ } }, - "RevokeCertificateRequest": { - "base": null, - "refs": { + "RevocationConfiguration":{ + "base":"

Certificate revocation information used by the CreateCertificateAuthority and UpdateCertificateAuthority operations. Your private certificate authority (CA) can create and maintain a certificate revocation list (CRL). A CRL contains information about certificates revoked by your CA. For more information, see RevokeCertificate.

", + "refs":{ + "CertificateAuthority$RevocationConfiguration":"

Information about the certificate revocation list (CRL) created and maintained by your private CA.

", + "CreateCertificateAuthorityRequest$RevocationConfiguration":"

Contains a Boolean value that you can use to enable a certification revocation list (CRL) for the CA, the name of the S3 bucket to which ACM PCA will write the CRL, and an optional CNAME alias that you can use to hide the name of your bucket in the CRL Distribution Points extension of your CA certificate. For more information, see the CrlConfiguration structure.

", + "UpdateCertificateAuthorityRequest$RevocationConfiguration":"

Revocation information for your private CA.

" } }, - "SigningAlgorithm": { - "base": null, - "refs": { - "CertificateAuthorityConfiguration$SigningAlgorithm": "

Name of the algorithm your private CA uses to sign certificate requests.

", - "IssueCertificateRequest$SigningAlgorithm": "

The name of the algorithm that will be used to sign the certificate to be issued.

" + "RevocationReason":{ + "base":null, + "refs":{ + "RevokeCertificateRequest$RevocationReason":"

Specifies why you revoked the certificate.

" } }, - "String": { - "base": null, - "refs": { - "CertificateAuthority$Serial": "

Serial number of your private CA.

", - "CertificateMismatchException$message": null, - "ConcurrentModificationException$message": null, - "CreateCertificateAuthorityAuditReportRequest$S3BucketName": "

Name of the S3 bucket that will contain the audit report.

", - "CreateCertificateAuthorityAuditReportResponse$S3Key": "

The key that uniquely identifies the report file in your S3 bucket.

", - "DescribeCertificateAuthorityAuditReportResponse$S3BucketName": "

Name of the S3 bucket that contains the report.

", - "DescribeCertificateAuthorityAuditReportResponse$S3Key": "

S3 key that uniquely identifies the report file in your S3 bucket.

", - "InvalidArgsException$message": null, - "InvalidArnException$message": null, - "InvalidNextTokenException$message": null, - "InvalidPolicyException$message": null, - "InvalidStateException$message": null, - "InvalidTagException$message": null, - "LimitExceededException$message": null, - "MalformedCSRException$message": null, - "MalformedCertificateException$message": null, - "RequestAlreadyProcessedException$message": null, - "RequestFailedException$message": null, - "RequestInProgressException$message": null, - "ResourceNotFoundException$message": null, - "TooManyTagsException$message": null + "RevokeCertificateRequest":{ + "base":null, + "refs":{ } }, - "String128": { - "base": null, - "refs": { - "ASN1Subject$State": "

State in which the subject of the certificate is located.

", - "ASN1Subject$Locality": "

The locality (such as a city or town) in which the certificate subject is located.

", - "ASN1Subject$Pseudonym": "

Typically a shortened version of a longer GivenName. For example, Jonathan is often shortened to John. Elizabeth is often shortened to Beth, Liz, or Eliza.

", - "RevokeCertificateRequest$CertificateSerial": "

Serial number of the certificate to be revoked. This must be in hexadecimal format. You can retrieve the serial number by calling GetCertificate with the Amazon Resource Name (ARN) of the certificate you want and the ARN of your private CA. The GetCertificate function retrieves the certificate in the PEM format. You can use the following OpenSSL command to list the certificate in text format and copy the hexadecimal serial number.

openssl x509 -in file_path -text -noout

You can also copy the serial number from the console or use the DescribeCertificate function in the AWS Certificate Manager API Reference.

" + "SigningAlgorithm":{ + "base":null, + "refs":{ + "CertificateAuthorityConfiguration$SigningAlgorithm":"

Name of the algorithm your private CA uses to sign certificate requests.

", + "IssueCertificateRequest$SigningAlgorithm":"

The name of the algorithm that will be used to sign the certificate to be issued.

" } }, - "String16": { - "base": null, - "refs": { - "ASN1Subject$GivenName": "

First name.

" + "String":{ + "base":null, + "refs":{ + "CertificateAuthority$Serial":"

Serial number of your private CA.

", + "CertificateMismatchException$message":null, + "ConcurrentModificationException$message":null, + "CreateCertificateAuthorityAuditReportRequest$S3BucketName":"

Name of the S3 bucket that will contain the audit report.

", + "CreateCertificateAuthorityAuditReportResponse$S3Key":"

The key that uniquely identifies the report file in your S3 bucket.

", + "DescribeCertificateAuthorityAuditReportResponse$S3BucketName":"

Name of the S3 bucket that contains the report.

", + "DescribeCertificateAuthorityAuditReportResponse$S3Key":"

S3 key that uniquely identifies the report file in your S3 bucket.

", + "InvalidArgsException$message":null, + "InvalidArnException$message":null, + "InvalidNextTokenException$message":null, + "InvalidPolicyException$message":null, + "InvalidStateException$message":null, + "InvalidTagException$message":null, + "LimitExceededException$message":null, + "MalformedCSRException$message":null, + "MalformedCertificateException$message":null, + "RequestAlreadyProcessedException$message":null, + "RequestFailedException$message":null, + "RequestInProgressException$message":null, + "ResourceNotFoundException$message":null, + "TooManyTagsException$message":null } }, - "String253": { - "base": null, - "refs": { - "CrlConfiguration$CustomCname": "

Name inserted into the certificate CRL Distribution Points extension that enables the use of an alias for the CRL distribution point. Use this value if you don't want the name of your S3 bucket to be public.

" + "String128":{ + "base":null, + "refs":{ + "ASN1Subject$State":"

State in which the subject of the certificate is located.

", + "ASN1Subject$Locality":"

The locality (such as a city or town) in which the certificate subject is located.

", + "ASN1Subject$Pseudonym":"

Typically a shortened version of a longer GivenName. For example, Jonathan is often shortened to John. Elizabeth is often shortened to Beth, Liz, or Eliza.

", + "RevokeCertificateRequest$CertificateSerial":"

Serial number of the certificate to be revoked. This must be in hexadecimal format. You can retrieve the serial number by calling GetCertificate with the Amazon Resource Name (ARN) of the certificate you want and the ARN of your private CA. The GetCertificate operation retrieves the certificate in the PEM format. You can use the following OpenSSL command to list the certificate in text format and copy the hexadecimal serial number.

openssl x509 -in file_path -text -noout

You can also copy the serial number from the console or use the DescribeCertificate operation in the AWS Certificate Manager API Reference.

" } }, - "String3": { - "base": null, - "refs": { - "ASN1Subject$GenerationQualifier": "

Typically a qualifier appended to the name of an individual. Examples include Jr. for junior, Sr. for senior, and III for third.

" + "String16":{ + "base":null, + "refs":{ + "ASN1Subject$GivenName":"

First name.

" } }, - "String3To255": { - "base": null, - "refs": { - "CrlConfiguration$S3BucketName": "

Name of the S3 bucket that contains the CRL. If you do not provide a value for the CustomCname argument, the name of your S3 bucket is placed into the CRL Distribution Points extension of the issued certificate. You can change the name of your bucket by calling the UpdateCertificateAuthority function. You must specify a bucket policy that allows ACM PCA to write the CRL to your bucket.

" + "String253":{ + "base":null, + "refs":{ + "CrlConfiguration$CustomCname":"

Name inserted into the certificate CRL Distribution Points extension that enables the use of an alias for the CRL distribution point. Use this value if you don't want the name of your S3 bucket to be public.

" } }, - "String40": { - "base": null, - "refs": { - "ASN1Subject$Surname": "

Family name. In the US and the UK for example, the surname of an individual is ordered last. In Asian cultures the surname is typically ordered first.

" + "String3":{ + "base":null, + "refs":{ + "ASN1Subject$GenerationQualifier":"

Typically a qualifier appended to the name of an individual. Examples include Jr. for junior, Sr. for senior, and III for third.

" } }, - "String5": { - "base": null, - "refs": { - "ASN1Subject$Initials": "

Concatenation that typically contains the first letter of the GivenName, the first letter of the middle name if one exists, and the first letter of the SurName.

" + "String3To255":{ + "base":null, + "refs":{ + "CrlConfiguration$S3BucketName":"

Name of the S3 bucket that contains the CRL. If you do not provide a value for the CustomCname argument, the name of your S3 bucket is placed into the CRL Distribution Points extension of the issued certificate. You can change the name of your bucket by calling the UpdateCertificateAuthority operation. You must specify a bucket policy that allows ACM PCA to write the CRL to your bucket.

" } }, - "String64": { - "base": null, - "refs": { - "ASN1Subject$Organization": "

Legal name of the organization with which the certificate subject is affiliated.

", - "ASN1Subject$OrganizationalUnit": "

A subdivision or unit of the organization (such as sales or finance) with which the certificate subject is affiliated.

", - "ASN1Subject$CommonName": "

Fully qualified domain name (FQDN) associated with the certificate subject.

", - "ASN1Subject$SerialNumber": "

The certificate serial number.

", - "ASN1Subject$Title": "

A title such as Mr. or Ms. which is pre-pended to the name to refer formally to the certificate subject.

" + "String40":{ + "base":null, + "refs":{ + "ASN1Subject$Surname":"

Family name. In the US and the UK, for example, the surname of an individual is ordered last. In Asian cultures the surname is typically ordered first.

" } }, - "TStamp": { - "base": null, - "refs": { - "CertificateAuthority$CreatedAt": "

Date and time at which your private CA was created.

", - "CertificateAuthority$LastStateChangeAt": "

Date and time at which your private CA was last updated.

", - "CertificateAuthority$NotBefore": "

Date and time before which your private CA certificate is not valid.

", - "CertificateAuthority$NotAfter": "

Date and time after which your private CA certificate is not valid.

", - "DescribeCertificateAuthorityAuditReportResponse$CreatedAt": "

The date and time at which the report was created.

" + "String5":{ + "base":null, + "refs":{ + "ASN1Subject$Initials":"

Concatenation that typically contains the first letter of the GivenName, the first letter of the middle name if one exists, and the first letter of the SurName.

" } }, - "Tag": { - "base": "

Tags are labels that you can use to identify and organize your private CAs. Each tag consists of a key and an optional value. You can associate up to 50 tags with a private CA. To add one or more tags to a private CA, call the TagCertificateAuthority function. To remove a tag, call the UntagCertificateAuthority function.

", - "refs": { - "TagList$member": null + "String64":{ + "base":null, + "refs":{ + "ASN1Subject$Organization":"

Legal name of the organization with which the certificate subject is affiliated.

", + "ASN1Subject$OrganizationalUnit":"

A subdivision or unit of the organization (such as sales or finance) with which the certificate subject is affiliated.

", + "ASN1Subject$CommonName":"

Fully qualified domain name (FQDN) associated with the certificate subject.

", + "ASN1Subject$SerialNumber":"

The certificate serial number.

", + "ASN1Subject$Title":"

A title such as Mr. or Ms., which is pre-pended to the name to refer formally to the certificate subject.

" } }, - "TagCertificateAuthorityRequest": { - "base": null, - "refs": { + "TStamp":{ + "base":null, + "refs":{ + "CertificateAuthority$CreatedAt":"

Date and time at which your private CA was created.

", + "CertificateAuthority$LastStateChangeAt":"

Date and time at which your private CA was last updated.

", + "CertificateAuthority$NotBefore":"

Date and time before which your private CA certificate is not valid.

", + "CertificateAuthority$NotAfter":"

Date and time after which your private CA certificate is not valid.

", + "CertificateAuthority$RestorableUntil":"

The period during which a deleted CA can be restored. For more information, see the PermanentDeletionTimeInDays parameter of the DeleteCertificateAuthorityRequest operation.

", + "DescribeCertificateAuthorityAuditReportResponse$CreatedAt":"

The date and time at which the report was created.

" } }, - "TagKey": { - "base": null, - "refs": { - "Tag$Key": "

Key (name) of the tag.

" + "Tag":{ + "base":"

Tags are labels that you can use to identify and organize your private CAs. Each tag consists of a key and an optional value. You can associate up to 50 tags with a private CA. To add one or more tags to a private CA, call the TagCertificateAuthority operation. To remove a tag, call the UntagCertificateAuthority operation.

", + "refs":{ + "TagList$member":null } }, - "TagList": { - "base": null, - "refs": { - "ListTagsResponse$Tags": "

The tags associated with your private CA.

", - "TagCertificateAuthorityRequest$Tags": "

List of tags to be associated with the CA.

", - "UntagCertificateAuthorityRequest$Tags": "

List of tags to be removed from the CA.

" + "TagCertificateAuthorityRequest":{ + "base":null, + "refs":{ } }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

Value of the tag.

" + "TagKey":{ + "base":null, + "refs":{ + "Tag$Key":"

Key (name) of the tag.

" } }, - "TooManyTagsException": { - "base": "

You can associate up to 50 tags with a private CA. Exception information is contained in the exception message field.

", - "refs": { + "TagList":{ + "base":null, + "refs":{ + "ListTagsResponse$Tags":"

The tags associated with your private CA.

", + "TagCertificateAuthorityRequest$Tags":"

List of tags to be associated with the CA.

", + "UntagCertificateAuthorityRequest$Tags":"

List of tags to be removed from the CA.

" } }, - "UntagCertificateAuthorityRequest": { - "base": null, - "refs": { + "TagValue":{ + "base":null, + "refs":{ + "Tag$Value":"

Value of the tag.

" } }, - "UpdateCertificateAuthorityRequest": { - "base": null, - "refs": { + "TooManyTagsException":{ + "base":"

You can associate up to 50 tags with a private CA. Exception information is contained in the exception message field.

", + "refs":{ } }, - "Validity": { - "base": "

Length of time for which the certificate issued by your private certificate authority (CA), or by the private CA itself, is valid in days, months, or years. You can issue a certificate by calling the IssueCertificate function.

", - "refs": { - "IssueCertificateRequest$Validity": "

The type of the validity period.

" + "UntagCertificateAuthorityRequest":{ + "base":null, + "refs":{ } }, - "ValidityPeriodType": { - "base": null, - "refs": { - "Validity$Type": "

Specifies whether the Value parameter represents days, months, or years.

" + "UpdateCertificateAuthorityRequest":{ + "base":null, + "refs":{ + } + }, + "Validity":{ + "base":"

Length of time for which the certificate issued by your private certificate authority (CA), or by the private CA itself, is valid in days, months, or years. You can issue a certificate by calling the IssueCertificate operation.

", + "refs":{ + "IssueCertificateRequest$Validity":"

The type of the validity period.

" + } + }, + "ValidityPeriodType":{ + "base":null, + "refs":{ + "Validity$Type":"

Specifies whether the Value parameter represents days, months, or years.

" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/examples-1.json index 0ea7e3b0b..faff76894 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/examples-1.json @@ -1,5 +1,5 @@ { - "version": "1.0", - "examples": { + "version":"1.0", + "examples":{ } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/paginators-1.json index 5677bd8e4..f85268a04 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/acm-pca/2017-08-22/paginators-1.json @@ -1,4 +1,4 @@ { - "pagination": { + "pagination":{ } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/api-2.json index e178b4c03..a677eb6fe 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/api-2.json @@ -32,7 +32,8 @@ "input":{"shape":"AssociateDeviceWithRoomRequest"}, "output":{"shape":"AssociateDeviceWithRoomResponse"}, "errors":[ - {"shape":"LimitExceededException"} + {"shape":"LimitExceededException"}, + {"shape":"DeviceNotRegisteredException"} ] }, "AssociateSkillGroupWithRoom":{ @@ -219,7 +220,10 @@ "requestUri":"/" }, "input":{"shape":"DisassociateDeviceFromRoomRequest"}, - "output":{"shape":"DisassociateDeviceFromRoomResponse"} + "output":{"shape":"DisassociateDeviceFromRoomResponse"}, + "errors":[ + {"shape":"DeviceNotRegisteredException"} + ] }, "DisassociateSkillGroupFromRoom":{ "name":"DisassociateSkillGroupFromRoom", @@ -463,7 +467,10 @@ "requestUri":"/" }, "input":{"shape":"StartDeviceSyncRequest"}, - "output":{"shape":"StartDeviceSyncResponse"} + "output":{"shape":"StartDeviceSyncResponse"}, + "errors":[ + {"shape":"DeviceNotRegisteredException"} + ] }, "TagResource":{ "name":"TagResource", @@ -523,7 +530,8 @@ "input":{"shape":"UpdateDeviceRequest"}, "output":{"shape":"UpdateDeviceResponse"}, "errors":[ - {"shape":"NotFoundException"} + {"shape":"NotFoundException"}, + {"shape":"DeviceNotRegisteredException"} ] }, "UpdateProfile":{ @@ -978,6 +986,13 @@ "min":2, "pattern":"[\\u0009\\u000A\\u000D\\u0020-\\u007E\\u0085\\u00A0-\\uD7FF\\uE000-\\uFFFD\\u10000-\\u10FFFF]*" }, + "DeviceNotRegisteredException":{ + "type":"structure", + "members":{ + "Message":{"shape":"ErrorMessage"} + }, + "exception":true + }, "DeviceSerialNumber":{ "type":"string", "pattern":"[a-zA-Z0-9]{1,200}" diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/docs-2.json index 00ce2b01b..ff3faee2c 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/alexaforbusiness/2017-11-09/docs-2.json @@ -450,6 +450,11 @@ "UpdateDeviceRequest$DeviceName": "

The updated device name. Required.

" } }, + "DeviceNotRegisteredException": { + "base": null, + "refs": { + } + }, "DeviceSerialNumber": { "base": null, "refs": { @@ -569,6 +574,7 @@ "base": null, "refs": { "AlreadyExistsException$Message": null, + "DeviceNotRegisteredException$Message": null, "InvalidUserStatusException$Message": null, "LimitExceededException$Message": null, "NameInUseException$Message": null, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/api-2.json index f592e3cb6..d913a6066 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/api-2.json @@ -2858,7 +2858,8 @@ "type":"string", "enum":[ "REGIONAL", - "EDGE" + "EDGE", + "PRIVATE" ] }, "ExportResponse":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/docs-2.json index 5652bf1a5..3297315e9 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/apigateway/2015-07-09/docs-2.json @@ -572,7 +572,7 @@ } }, "EndpointType": { - "base": "

The endpoint type. The valid value is EDGE for edge-optimized API setup, most suitable for mobile applications, REGIONAL for regional API endpoint setup, most suitable for calling from AWS Region

", + "base": "

The endpoint type. The valid values are EDGE for edge-optimized API setup, most suitable for mobile applications; REGIONAL for regional API endpoint setup, most suitable for calling from AWS Region; and PRIVATE for private APIs.

", "refs": { "ListOfEndpointType$member": null } @@ -965,7 +965,7 @@ "ListOfEndpointType": { "base": null, "refs": { - "EndpointConfiguration$types": "

A list of endpoint types of an API (RestApi) or its custom domain name (DomainName). For an edge-optimized API and its custom domain name, the endpoint type is \"EDGE\". For a regional API and its custom domain name, the endpoint type is REGIONAL.

" + "EndpointConfiguration$types": "

A list of endpoint types of an API (RestApi) or its custom domain name (DomainName). For an edge-optimized API and its custom domain name, the endpoint type is \"EDGE\". For a regional API and its custom domain name, the endpoint type is REGIONAL. For a private API, the endpoint type is PRIVATE.

" } }, "ListOfGatewayResponse": { @@ -1188,9 +1188,9 @@ "DeploymentCanarySettings$stageVariableOverrides": "

A stage variable overrides used for the canary release deployment. They can override existing stage variables or add new stage variables for the canary release deployment. These stage variables are represented as a string-to-string map between stage variable names and their values.

", "GatewayResponse$responseParameters": "

Response parameters (paths, query strings and headers) of the GatewayResponse as a string-to-string map of key-value pairs.

", "GatewayResponse$responseTemplates": "

Response templates of the GatewayResponse as a string-to-string map of key-value pairs.

", - "GetExportRequest$parameters": "

A key-value map of query string parameters that specify properties of the export, depending on the requested exportType. For exportType swagger, any combination of the following parameters are supported: integrations will export the API with x-amazon-apigateway-integration extensions. authorizers will export the API with x-amazon-apigateway-authorizer extensions. postman will export the API with Postman extensions, allowing for import to the Postman tool

", + "GetExportRequest$parameters": "

A key-value map of query string parameters that specify properties of the export, depending on the requested exportType. For exportType swagger, any combination of the following parameters are supported: extensions='integrations' or extensions='apigateway' will export the API with x-amazon-apigateway-integration extensions. extensions='authorizers' will export the API with x-amazon-apigateway-authorizer extensions. postman will export the API with Postman extensions, allowing for import to the Postman tool

", "GetSdkRequest$parameters": "

A string-to-string key-value map of query parameters sdkType-dependent properties of the SDK. For sdkType of objectivec or swift, a parameter named classPrefix is required. For sdkType of android, parameters named groupId, artifactId, artifactVersion, and invokerPackage are required. For sdkType of java, parameters named serviceName and javaPackageName are required.

", - "ImportRestApiRequest$parameters": "

A key-value map of context-specific query string parameters specifying the behavior of different API importing operations. The following shows operation-specific parameters and their supported values.

To exclude DocumentationParts from the import, set parameters as ignore=documentation.

To configure the endpoint type, set parameters as endpointConfigurationTypes=EDGE orendpointConfigurationTypes=REGIONAL. The default endpoint type is EDGE.

To handle imported basePath, set parameters as basePath=ignore, basePath=prepend or basePath=split.

For example, the AWS CLI command to exclude documentation from the imported API is:

aws apigateway import-rest-api --parameters ignore=documentation --body 'file:///path/to/imported-api-body.json

The AWS CLI command to set the regional endpoint on the imported API is:

aws apigateway import-rest-api --parameters endpointConfigurationTypes=REGIONAL --body 'file:///path/to/imported-api-body.json
", + "ImportRestApiRequest$parameters": "

A key-value map of context-specific query string parameters specifying the behavior of different API importing operations. The following shows operation-specific parameters and their supported values.

To exclude DocumentationParts from the import, set parameters as ignore=documentation.

To configure the endpoint type, set parameters as endpointConfigurationTypes=EDGE, endpointConfigurationTypes=REGIONAL, or endpointConfigurationTypes=PRIVATE. The default endpoint type is EDGE.

To handle imported basePath, set parameters as basePath=ignore, basePath=prepend or basePath=split.

For example, the AWS CLI command to exclude documentation from the imported API is:

aws apigateway import-rest-api --parameters ignore=documentation --body 'file:///path/to/imported-api-body.json'

The AWS CLI command to set the regional endpoint on the imported API is:

aws apigateway import-rest-api --parameters endpointConfigurationTypes=REGIONAL --body 'file:///path/to/imported-api-body.json'
", "Integration$requestParameters": "

A key-value map specifying request parameters that are passed from the method request to the back end. The key is an integration request parameter name and the associated value is a method request parameter value or static value that must be enclosed within single quotes and pre-encoded as required by the back end. The method request parameter value must match the pattern of method.request.{location}.{name}, where location is querystring, path, or header and name must be a valid and unique method request parameter name.

", "Integration$requestTemplates": "

Represents a map of Velocity templates that are applied on the request payload based on the value of the Content-Type header sent by the client. The content type value is the key in this map, and the template (as a String) is the value.

", "IntegrationResponse$responseParameters": "

A key-value map specifying response parameters that are passed to the method response from the back end. The key is a method response header parameter name and the mapped value is an integration response header value, a static value enclosed within a pair of single quotes, or a JSON expression from the integration response body. The mapping key must match the pattern of method.response.header.{name}, where name is a valid and unique header name. The mapped non-static value must match the pattern of integration.response.header.{name} or integration.response.body.{JSON-expression}, where name is a valid and unique response header name and JSON-expression is a valid JSON expression without the $ prefix.

", @@ -1205,7 +1205,7 @@ "PutIntegrationResponseRequest$responseTemplates": "

Specifies a put integration response's templates.

", "PutMethodRequest$requestModels": "

Specifies the Model resources used for the request's content type. Request models are represented as a key/value map, with a content type as the key and a Model name as the value.

", "PutMethodResponseRequest$responseModels": "

Specifies the Model resources used for the response's content type. Response models are represented as a key/value map, with a content type as the key and a Model name as the value.

", - "PutRestApiRequest$parameters": "

Custom header parameters as part of the request. For example, to exclude DocumentationParts from an imported API, set ignore=documentation as a parameters value, as in the AWS CLI command of aws apigateway import-rest-api --parameters ignore=documentation --body 'file:///path/to/imported-api-body.json.

", + "PutRestApiRequest$parameters": "

Custom header parameters as part of the request. For example, to exclude DocumentationParts from an imported API, set ignore=documentation as a parameters value, as in the AWS CLI command of aws apigateway import-rest-api --parameters ignore=documentation --body 'file:///path/to/imported-api-body.json'.

", "Stage$variables": "

A map that defines the stage variables for a Stage resource. Variable names can have alphanumeric and underscore characters, and the values must match [A-Za-z0-9-._~:/?#&=,]+.

", "Stage$tags": "

The collection of tags. Each tag element is associated with a given resource.

", "TagResourceRequest$tags": "

[Required] The key-value map of strings. The valid character set is [a-zA-Z+-=._:/]. The tag key can be up to 128 characters and must not start with aws:. The tag value can be up to 256 characters.

", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/api-2.json index c8184969e..885188ddc 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/api-2.json @@ -477,7 +477,8 @@ {"shape":"InvalidParameterCombinationException"}, {"shape":"LimitExceededException"}, {"shape":"InvalidAccountStatusException"}, - {"shape":"IncompatibleImageException"} + {"shape":"IncompatibleImageException"}, + {"shape":"OperationNotPermittedException"} ] } }, @@ -493,6 +494,16 @@ "min":1, "sensitive":true }, + "Action":{ + "type":"string", + "enum":[ + "CLIPBOARD_COPY_FROM_LOCAL_DEVICE", + "CLIPBOARD_COPY_TO_LOCAL_DEVICE", + "FILE_UPLOAD", + "FILE_DOWNLOAD", + "PRINTING_TO_LOCAL_DEVICE" + ] + }, "Application":{ "type":"structure", "members":{ @@ -680,12 +691,13 @@ "type":"structure", "required":["Name"], "members":{ - "Name":{"shape":"String"}, + "Name":{"shape":"Name"}, "Description":{"shape":"Description"}, "DisplayName":{"shape":"DisplayName"}, "StorageConnectors":{"shape":"StorageConnectorList"}, "RedirectURL":{"shape":"RedirectURL"}, - "FeedbackURL":{"shape":"FeedbackURL"} + "FeedbackURL":{"shape":"FeedbackURL"}, + "UserSettings":{"shape":"UserSettingList"} } }, "CreateStackResult":{ @@ -914,6 +926,10 @@ "type":"string", "max":100 }, + "Domain":{ + "type":"string", + "max":64 + }, "DomainJoinInfo":{ "type":"structure", "members":{ @@ -921,6 +937,11 @@ "OrganizationalUnitDistinguishedName":{"shape":"OrganizationalUnitDistinguishedName"} } }, + "DomainList":{ + "type":"list", + "member":{"shape":"Domain"}, + "max":10 + }, "ErrorMessage":{"type":"string"}, "ExpireSessionRequest":{ "type":"structure", @@ -1232,6 +1253,13 @@ "type":"string", "pattern":"^[a-zA-Z0-9][a-zA-Z0-9_.-]{0,100}$" }, + "NetworkAccessConfiguration":{ + "type":"structure", + "members":{ + "EniPrivateIpAddress":{"shape":"String"}, + "EniId":{"shape":"String"} + } + }, "OperationNotPermittedException":{ "type":"structure", "members":{ @@ -1247,6 +1275,13 @@ "type":"list", "member":{"shape":"OrganizationalUnitDistinguishedName"} }, + "Permission":{ + "type":"string", + "enum":[ + "ENABLED", + "DISABLED" + ] + }, "PlatformType":{ "type":"string", "enum":["WINDOWS"] @@ -1335,7 +1370,8 @@ "StackName":{"shape":"String"}, "FleetName":{"shape":"String"}, "State":{"shape":"SessionState"}, - "AuthenticationType":{"shape":"AuthenticationType"} + "AuthenticationType":{"shape":"AuthenticationType"}, + "NetworkAccessConfiguration":{"shape":"NetworkAccessConfiguration"} } }, "SessionList":{ @@ -1362,16 +1398,20 @@ "StorageConnectors":{"shape":"StorageConnectorList"}, "RedirectURL":{"shape":"RedirectURL"}, "FeedbackURL":{"shape":"FeedbackURL"}, - "StackErrors":{"shape":"StackErrors"} + "StackErrors":{"shape":"StackErrors"}, + "UserSettings":{"shape":"UserSettingList"} } }, "StackAttribute":{ "type":"string", "enum":[ "STORAGE_CONNECTORS", + "STORAGE_CONNECTOR_HOMEFOLDERS", + "STORAGE_CONNECTOR_GOOGLE_DRIVE", "REDIRECT_URL", "FEEDBACK_URL", - "THEME_NAME" + "THEME_NAME", + "USER_SETTINGS" ] }, "StackAttributes":{ @@ -1456,7 +1496,8 @@ "required":["ConnectorType"], "members":{ "ConnectorType":{"shape":"StorageConnectorType"}, - "ResourceIdentifier":{"shape":"ResourceIdentifier"} + "ResourceIdentifier":{"shape":"ResourceIdentifier"}, + "Domains":{"shape":"DomainList"} } }, "StorageConnectorList":{ @@ -1465,7 +1506,10 @@ }, "StorageConnectorType":{ "type":"string", - "enum":["HOMEFOLDERS"] + "enum":[ + "HOMEFOLDERS", + "GOOGLE_DRIVE" + ] }, "StreamingUrlUserId":{ "type":"string", @@ -1600,7 +1644,8 @@ }, "RedirectURL":{"shape":"RedirectURL"}, "FeedbackURL":{"shape":"FeedbackURL"}, - "AttributesToDelete":{"shape":"StackAttributes"} + "AttributesToDelete":{"shape":"StackAttributes"}, + "UserSettings":{"shape":"UserSettingList"} } }, "UpdateStackResult":{ @@ -1614,6 +1659,22 @@ "max":32, "min":2 }, + "UserSetting":{ + "type":"structure", + "required":[ + "Action", + "Permission" + ], + "members":{ + "Action":{"shape":"Action"}, + "Permission":{"shape":"Permission"} + } + }, + "UserSettingList":{ + "type":"list", + "member":{"shape":"UserSetting"}, + "min":1 + }, "VisibilityType":{ "type":"string", "enum":[ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/docs-2.json index e4911fdee..a99fa8e0f 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/docs-2.json @@ -1,1083 +1,1129 @@ { - "version": "2.0", - "service": "Amazon AppStream 2.0

You can use Amazon AppStream 2.0 to stream desktop applications to any device running a web browser, without rewriting them.

", - "operations": { - "AssociateFleet": "

Associates the specified fleet with the specified stack.

", - "CopyImage": "

Copies the image within the same region or to a new region within the same AWS account. Note that any tags you added to the image will not be copied.

", - "CreateDirectoryConfig": "

Creates a directory configuration.

", - "CreateFleet": "

Creates a fleet.

", - "CreateImageBuilder": "

Creates an image builder.

The initial state of the builder is PENDING. When it is ready, the state is RUNNING.

", - "CreateImageBuilderStreamingURL": "

Creates a URL to start an image builder streaming session.

", - "CreateStack": "

Creates a stack.

", - "CreateStreamingURL": "

Creates a URL to start a streaming session for the specified user.

", - "DeleteDirectoryConfig": "

Deletes the specified directory configuration.

", - "DeleteFleet": "

Deletes the specified fleet.

", - "DeleteImage": "

Deletes the specified image. You cannot delete an image that is currently in use. After you delete an image, you cannot provision new capacity using the image.

", - "DeleteImageBuilder": "

Deletes the specified image builder and releases the capacity.

", - "DeleteStack": "

Deletes the specified stack. After this operation completes, the environment can no longer be activated and any reservations made for the stack are released.

", - "DescribeDirectoryConfigs": "

Describes the specified directory configurations. Note that although the response syntax in this topic includes the account password, this password is not returned in the actual response.

", - "DescribeFleets": "

Describes the specified fleets or all fleets in the account.

", - "DescribeImageBuilders": "

Describes the specified image builders or all image builders in the account.

", - "DescribeImages": "

Describes the specified images or all images in the account.

", - "DescribeSessions": "

Describes the streaming sessions for the specified stack and fleet. If a user ID is provided, only the streaming sessions for only that user are returned. If an authentication type is not provided, the default is to authenticate users using a streaming URL.

", - "DescribeStacks": "

Describes the specified stacks or all stacks in the account.

", - "DisassociateFleet": "

Disassociates the specified fleet from the specified stack.

", - "ExpireSession": "

Stops the specified streaming session.

", - "ListAssociatedFleets": "

Lists the fleets associated with the specified stack.

", - "ListAssociatedStacks": "

Lists the stacks associated with the specified fleet.

", - "ListTagsForResource": "

Lists the tags for the specified AppStream 2.0 resource. You can tag AppStream 2.0 image builders, images, fleets, and stacks.

For more information about tags, see Tagging Your Resources in the Amazon AppStream 2.0 Developer Guide.

", - "StartFleet": "

Starts the specified fleet.

", - "StartImageBuilder": "

Starts the specified image builder.

", - "StopFleet": "

Stops the specified fleet.

", - "StopImageBuilder": "

Stops the specified image builder.

", - "TagResource": "

Adds or overwrites one or more tags for the specified AppStream 2.0 resource. You can tag AppStream 2.0 image builders, images, fleets, and stacks.

Each tag consists of a key and an optional value. If a resource already has a tag with the same key, this operation updates its value.

To list the current tags for your resources, use ListTagsForResource. To disassociate tags from your resources, use UntagResource.

For more information about tags, see Tagging Your Resources in the Amazon AppStream 2.0 Developer Guide.

", - "UntagResource": "

Disassociates the specified tags from the specified AppStream 2.0 resource.

To list the current tags for your resources, use ListTagsForResource.

For more information about tags, see Tagging Your Resources in the Amazon AppStream 2.0 Developer Guide.

", - "UpdateDirectoryConfig": "

Updates the specified directory configuration.

", - "UpdateFleet": "

Updates the specified fleet.

If the fleet is in the STOPPED state, you can update any attribute except the fleet name. If the fleet is in the RUNNING state, you can update the DisplayName and ComputeCapacity attributes. If the fleet is in the STARTING or STOPPING state, you can't update it.

", - "UpdateStack": "

Updates the specified stack.

" + "version":"2.0", + "service":"Amazon AppStream 2.0

You can use Amazon AppStream 2.0 to stream desktop applications to any device running a web browser, without rewriting them.

", + "operations":{ + "AssociateFleet":"

Associates the specified fleet with the specified stack.

", + "CopyImage":"

Copies the image within the same region or to a new region within the same AWS account. Note that any tags you added to the image will not be copied.

", + "CreateDirectoryConfig":"

Creates a Directory Config object in AppStream 2.0. This object includes the information required to join streaming instances to an Active Directory domain.

", + "CreateFleet":"

Creates a fleet. A fleet consists of streaming instances that run a specified image.

", + "CreateImageBuilder":"

Creates an image builder. An image builder is a virtual machine that is used to create an image.

The initial state of the builder is PENDING. When it is ready, the state is RUNNING.

", + "CreateImageBuilderStreamingURL":"

Creates a URL to start an image builder streaming session.

", + "CreateStack":"

Creates a stack to start streaming applications to users. A stack consists of an associated fleet, user access policies, and storage configurations.

", + "CreateStreamingURL":"

Creates a temporary URL to start an AppStream 2.0 streaming session for the specified user. A streaming URL enables application streaming to be tested without user setup.

", + "DeleteDirectoryConfig":"

Deletes the specified Directory Config object from AppStream 2.0. This object includes the information required to join streaming instances to an Active Directory domain.

", + "DeleteFleet":"

Deletes the specified fleet.

", + "DeleteImage":"

Deletes the specified image. You cannot delete an image when it is in use. After you delete an image, you cannot provision new capacity using the image.

", + "DeleteImageBuilder":"

Deletes the specified image builder and releases the capacity.

", + "DeleteStack":"

Deletes the specified stack. After the stack is deleted, the application streaming environment provided by the stack is no longer available to users. Also, any reservations made for application streaming sessions for the stack are released.

", + "DescribeDirectoryConfigs":"

Retrieves a list that describes one or more specified Directory Config objects for AppStream 2.0, if the names for these objects are provided. Otherwise, all Directory Config objects in the account are described. These objects include the information required to join streaming instances to an Active Directory domain.

Although the response syntax in this topic includes the account password, this password is not returned in the actual response.

", + "DescribeFleets":"

Retrieves a list that describes one or more specified fleets, if the fleet names are provided. Otherwise, all fleets in the account are described.

", + "DescribeImageBuilders":"

Retrieves a list that describes one or more specified image builders, if the image builder names are provided. Otherwise, all image builders in the account are described.

", + "DescribeImages":"

Retrieves a list that describes one or more specified images, if the image names are provided. Otherwise, all images in the account are described.

", + "DescribeSessions":"

Retrieves a list that describes the streaming sessions for a specified stack and fleet. If a user ID is provided for the stack and fleet, only streaming sessions for that user are described. If an authentication type is not provided, the default is to authenticate users using a streaming URL.

", + "DescribeStacks":"

Retrieves a list that describes one or more specified stacks, if the stack names are provided. Otherwise, all stacks in the account are described.

", + "DisassociateFleet":"

Disassociates the specified fleet from the specified stack.

", + "ExpireSession":"

Immediately stops the specified streaming session.

", + "ListAssociatedFleets":"

Retrieves the name of the fleet that is associated with the specified stack.

", + "ListAssociatedStacks":"

Retrieves the name of the stack with which the specified fleet is associated.

", + "ListTagsForResource":"

Retrieves a list of all tags for the specified AppStream 2.0 resource. You can tag AppStream 2.0 image builders, images, fleets, and stacks.

For more information about tags, see Tagging Your Resources in the Amazon AppStream 2.0 Developer Guide.

", + "StartFleet":"

Starts the specified fleet.

", + "StartImageBuilder":"

Starts the specified image builder.

", + "StopFleet":"

Stops the specified fleet.

", + "StopImageBuilder":"

Stops the specified image builder.

", + "TagResource":"

Adds or overwrites one or more tags for the specified AppStream 2.0 resource. You can tag AppStream 2.0 image builders, images, fleets, and stacks.

Each tag consists of a key and an optional value. If a resource already has a tag with the same key, this operation updates its value.

To list the current tags for your resources, use ListTagsForResource. To disassociate tags from your resources, use UntagResource.

For more information about tags, see Tagging Your Resources in the Amazon AppStream 2.0 Developer Guide.

", + "UntagResource":"

Disassociates one or more specified tags from the specified AppStream 2.0 resource.

To list the current tags for your resources, use ListTagsForResource.

For more information about tags, see Tagging Your Resources in the Amazon AppStream 2.0 Developer Guide.

", + "UpdateDirectoryConfig":"

Updates the specified Directory Config object in AppStream 2.0. This object includes the information required to join streaming instances to an Active Directory domain.

", + "UpdateFleet":"

Updates the specified fleet.

If the fleet is in the STOPPED state, you can update any attribute except the fleet name. If the fleet is in the RUNNING state, you can update the DisplayName and ComputeCapacity attributes. If the fleet is in the STARTING or STOPPING state, you can't update it.

", + "UpdateStack":"

Updates the specified fields for the specified stack.

" }, - "shapes": { - "AccountName": { - "base": null, - "refs": { - "ServiceAccountCredentials$AccountName": "

The user name of the account. This account must have the following privileges: create computer objects, join computers to the domain, and change/reset the password on descendant computer objects for the organizational units specified.

" + "shapes":{ + "AccountName":{ + "base":null, + "refs":{ + "ServiceAccountCredentials$AccountName":"

The user name of the account. This account must have the following privileges: create computer objects, join computers to the domain, and change/reset the password on descendant computer objects for the organizational units specified.

" } }, - "AccountPassword": { - "base": null, - "refs": { - "ServiceAccountCredentials$AccountPassword": "

The password for the account.

" + "AccountPassword":{ + "base":null, + "refs":{ + "ServiceAccountCredentials$AccountPassword":"

The password for the account.

" } }, - "Application": { - "base": "

Describes an application in the application catalog.

", - "refs": { - "Applications$member": null + "Action":{ + "base":null, + "refs":{ + "UserSetting$Action":"

The action that is enabled or disabled.

" } }, - "Applications": { - "base": null, - "refs": { - "Image$Applications": "

The applications associated with the image.

" + "Application":{ + "base":"

Describes an application in the application catalog.

", + "refs":{ + "Applications$member":null } }, - "AppstreamAgentVersion": { - "base": null, - "refs": { - "CreateImageBuilderRequest$AppstreamAgentVersion": "

The version of the AppStream 2.0 agent to use for this image builder. To use the latest version of the AppStream 2.0 agent, specify [LATEST].

", - "Image$AppstreamAgentVersion": "

The version of the AppStream 2.0 agent to use for instances that are launched from this image.

", - "ImageBuilder$AppstreamAgentVersion": "

The version of the AppStream 2.0 agent that is currently being used by this image builder.

", - "StartImageBuilderRequest$AppstreamAgentVersion": "

The version of the AppStream 2.0 agent to use for this image builder. To use the latest version of the AppStream 2.0 agent, specify [LATEST].

" + "Applications":{ + "base":null, + "refs":{ + "Image$Applications":"

The applications associated with the image.

" } }, - "Arn": { - "base": null, - "refs": { - "Fleet$Arn": "

The ARN for the fleet.

", - "Image$Arn": "

The ARN of the image.

", - "Image$BaseImageArn": "

The ARN of the image from which this image was created.

", - "ImageBuilder$Arn": "

The ARN for the image builder.

", - "ImageBuilder$ImageArn": "

The ARN of the image from which this builder was created.

", - "ListTagsForResourceRequest$ResourceArn": "

The Amazon Resource Name (ARN) of the resource.

", - "Stack$Arn": "

The ARN of the stack.

", - "TagResourceRequest$ResourceArn": "

The Amazon Resource Name (ARN) of the resource.

", - "UntagResourceRequest$ResourceArn": "

The Amazon Resource Name (ARN) of the resource.

" + "AppstreamAgentVersion":{ + "base":null, + "refs":{ + "CreateImageBuilderRequest$AppstreamAgentVersion":"

The version of the AppStream 2.0 agent to use for this image builder. To use the latest version of the AppStream 2.0 agent, specify [LATEST].

", + "Image$AppstreamAgentVersion":"

The version of the AppStream 2.0 agent to use for instances that are launched from this image.

", + "ImageBuilder$AppstreamAgentVersion":"

The version of the AppStream 2.0 agent that is currently being used by this image builder.

", + "StartImageBuilderRequest$AppstreamAgentVersion":"

The version of the AppStream 2.0 agent to use for this image builder. To use the latest version of the AppStream 2.0 agent, specify [LATEST].

" } }, - "AssociateFleetRequest": { - "base": null, - "refs": { + "Arn":{ + "base":null, + "refs":{ + "Fleet$Arn":"

The ARN for the fleet.

", + "Image$Arn":"

The ARN of the image.

", + "Image$BaseImageArn":"

The ARN of the image from which this image was created.

", + "ImageBuilder$Arn":"

The ARN for the image builder.

", + "ImageBuilder$ImageArn":"

The ARN of the image from which this builder was created.

", + "ListTagsForResourceRequest$ResourceArn":"

The Amazon Resource Name (ARN) of the resource.

", + "Stack$Arn":"

The ARN of the stack.

", + "TagResourceRequest$ResourceArn":"

The Amazon Resource Name (ARN) of the resource.

", + "UntagResourceRequest$ResourceArn":"

The Amazon Resource Name (ARN) of the resource.

" } }, - "AssociateFleetResult": { - "base": null, - "refs": { + "AssociateFleetRequest":{ + "base":null, + "refs":{ } }, - "AuthenticationType": { - "base": null, - "refs": { - "DescribeSessionsRequest$AuthenticationType": "

The authentication method. Specify API for a user authenticated using a streaming URL or SAML for a SAML federated user. The default is to authenticate users using a streaming URL.

", - "Session$AuthenticationType": "

The authentication method. The user is authenticated using a streaming URL (API) or SAML federation (SAML).

" + "AssociateFleetResult":{ + "base":null, + "refs":{ } }, - "Boolean": { - "base": null, - "refs": { - "Application$Enabled": "

If there is a problem, the application can be disabled after image creation.

", - "Image$ImageBuilderSupported": "

Indicates whether an image builder can be launched from this image.

", - "UpdateFleetRequest$DeleteVpcConfig": "

Deletes the VPC association for the specified fleet.

", - "UpdateStackRequest$DeleteStorageConnectors": "

Deletes the storage connectors currently enabled for the stack.

" + "AuthenticationType":{ + "base":null, + "refs":{ + "DescribeSessionsRequest$AuthenticationType":"

The authentication method. Specify API for a user authenticated using a streaming URL or SAML for a SAML federated user. The default is to authenticate users using a streaming URL.

", + "Session$AuthenticationType":"

The authentication method. The user is authenticated using a streaming URL (API) or SAML federation (SAML).

" } }, - "BooleanObject": { - "base": null, - "refs": { - "CreateFleetRequest$EnableDefaultInternetAccess": "

Enables or disables default internet access for the fleet.

", - "CreateImageBuilderRequest$EnableDefaultInternetAccess": "

Enables or disables default internet access for the image builder.

", - "Fleet$EnableDefaultInternetAccess": "

Indicates whether default internet access is enabled for the fleet.

", - "ImageBuilder$EnableDefaultInternetAccess": "

Enables or disables default internet access for the image builder.

", - "UpdateFleetRequest$EnableDefaultInternetAccess": "

Enables or disables default internet access for the fleet.

" + "Boolean":{ + "base":null, + "refs":{ + "Application$Enabled":"

If there is a problem, the application can be disabled after image creation.

", + "Image$ImageBuilderSupported":"

Indicates whether an image builder can be launched from this image.

", + "UpdateFleetRequest$DeleteVpcConfig":"

Deletes the VPC association for the specified fleet.

", + "UpdateStackRequest$DeleteStorageConnectors":"

Deletes the storage connectors currently enabled for the stack.

" } }, - "ComputeCapacity": { - "base": "

Describes the capacity for a fleet.

", - "refs": { - "CreateFleetRequest$ComputeCapacity": "

The desired capacity for the fleet.

", - "UpdateFleetRequest$ComputeCapacity": "

The desired capacity for the fleet.

" + "BooleanObject":{ + "base":null, + "refs":{ + "CreateFleetRequest$EnableDefaultInternetAccess":"

Enables or disables default internet access for the fleet.

", + "CreateImageBuilderRequest$EnableDefaultInternetAccess":"

Enables or disables default internet access for the image builder.

", + "Fleet$EnableDefaultInternetAccess":"

Indicates whether default internet access is enabled for the fleet.

", + "ImageBuilder$EnableDefaultInternetAccess":"

Enables or disables default internet access for the image builder.

", + "UpdateFleetRequest$EnableDefaultInternetAccess":"

Enables or disables default internet access for the fleet.

" } }, - "ComputeCapacityStatus": { - "base": "

Describes the capacity status for a fleet.

", - "refs": { - "Fleet$ComputeCapacityStatus": "

The capacity status for the fleet.

" + "ComputeCapacity":{ + "base":"

Describes the capacity for a fleet.

", + "refs":{ + "CreateFleetRequest$ComputeCapacity":"

The desired capacity for the fleet.

", + "UpdateFleetRequest$ComputeCapacity":"

The desired capacity for the fleet.

" } }, - "ConcurrentModificationException": { - "base": "

An API error occurred. Wait a few minutes and try again.

", - "refs": { + "ComputeCapacityStatus":{ + "base":"

Describes the capacity status for a fleet.

", + "refs":{ + "Fleet$ComputeCapacityStatus":"

The capacity status for the fleet.

" } }, - "CopyImageRequest": { - "base": null, - "refs": { + "ConcurrentModificationException":{ + "base":"

An API error occurred. Wait a few minutes and try again.

", + "refs":{ } }, - "CopyImageResponse": { - "base": null, - "refs": { + "CopyImageRequest":{ + "base":null, + "refs":{ } }, - "CreateDirectoryConfigRequest": { - "base": null, - "refs": { + "CopyImageResponse":{ + "base":null, + "refs":{ } }, - "CreateDirectoryConfigResult": { - "base": null, - "refs": { + "CreateDirectoryConfigRequest":{ + "base":null, + "refs":{ } }, - "CreateFleetRequest": { - "base": null, - "refs": { + "CreateDirectoryConfigResult":{ + "base":null, + "refs":{ } }, - "CreateFleetResult": { - "base": null, - "refs": { + "CreateFleetRequest":{ + "base":null, + "refs":{ } }, - "CreateImageBuilderRequest": { - "base": null, - "refs": { + "CreateFleetResult":{ + "base":null, + "refs":{ } }, - "CreateImageBuilderResult": { - "base": null, - "refs": { + "CreateImageBuilderRequest":{ + "base":null, + "refs":{ } }, - "CreateImageBuilderStreamingURLRequest": { - "base": null, - "refs": { + "CreateImageBuilderResult":{ + "base":null, + "refs":{ } }, - "CreateImageBuilderStreamingURLResult": { - "base": null, - "refs": { + "CreateImageBuilderStreamingURLRequest":{ + "base":null, + "refs":{ } }, - "CreateStackRequest": { - "base": null, - "refs": { + "CreateImageBuilderStreamingURLResult":{ + "base":null, + "refs":{ } }, - "CreateStackResult": { - "base": null, - "refs": { + "CreateStackRequest":{ + "base":null, + "refs":{ } }, - "CreateStreamingURLRequest": { - "base": null, - "refs": { + "CreateStackResult":{ + "base":null, + "refs":{ } }, - "CreateStreamingURLResult": { - "base": null, - "refs": { + "CreateStreamingURLRequest":{ + "base":null, + "refs":{ } }, - "DeleteDirectoryConfigRequest": { - "base": null, - "refs": { + "CreateStreamingURLResult":{ + "base":null, + "refs":{ } }, - "DeleteDirectoryConfigResult": { - "base": null, - "refs": { + "DeleteDirectoryConfigRequest":{ + "base":null, + "refs":{ } }, - "DeleteFleetRequest": { - "base": null, - "refs": { + "DeleteDirectoryConfigResult":{ + "base":null, + "refs":{ } }, - "DeleteFleetResult": { - "base": null, - "refs": { + "DeleteFleetRequest":{ + "base":null, + "refs":{ } }, - "DeleteImageBuilderRequest": { - "base": null, - "refs": { + "DeleteFleetResult":{ + "base":null, + "refs":{ } }, - "DeleteImageBuilderResult": { - "base": null, - "refs": { + "DeleteImageBuilderRequest":{ + "base":null, + "refs":{ } }, - "DeleteImageRequest": { - "base": null, - "refs": { + "DeleteImageBuilderResult":{ + "base":null, + "refs":{ } }, - "DeleteImageResult": { - "base": null, - "refs": { + "DeleteImageRequest":{ + "base":null, + "refs":{ } }, - "DeleteStackRequest": { - "base": null, - "refs": { + "DeleteImageResult":{ + "base":null, + "refs":{ } }, - "DeleteStackResult": { - "base": null, - "refs": { + "DeleteStackRequest":{ + "base":null, + "refs":{ } }, - "DescribeDirectoryConfigsRequest": { - "base": null, - "refs": { + "DeleteStackResult":{ + "base":null, + "refs":{ } }, - "DescribeDirectoryConfigsResult": { - "base": null, - "refs": { + "DescribeDirectoryConfigsRequest":{ + "base":null, + "refs":{ } }, - "DescribeFleetsRequest": { - "base": null, - "refs": { + "DescribeDirectoryConfigsResult":{ + "base":null, + "refs":{ } }, - "DescribeFleetsResult": { - "base": null, - "refs": { + "DescribeFleetsRequest":{ + "base":null, + "refs":{ } }, - "DescribeImageBuildersRequest": { - "base": null, - "refs": { + "DescribeFleetsResult":{ + "base":null, + "refs":{ } }, - "DescribeImageBuildersResult": { - "base": null, - "refs": { + "DescribeImageBuildersRequest":{ + "base":null, + "refs":{ } }, - "DescribeImagesRequest": { - "base": null, - "refs": { + "DescribeImageBuildersResult":{ + "base":null, + "refs":{ } }, - "DescribeImagesResult": { - "base": null, - "refs": { + "DescribeImagesRequest":{ + "base":null, + "refs":{ } }, - "DescribeSessionsRequest": { - "base": null, - "refs": { + "DescribeImagesResult":{ + "base":null, + "refs":{ } }, - "DescribeSessionsResult": { - "base": null, - "refs": { + "DescribeSessionsRequest":{ + "base":null, + "refs":{ } }, - "DescribeStacksRequest": { - "base": null, - "refs": { + "DescribeSessionsResult":{ + "base":null, + "refs":{ } }, - "DescribeStacksResult": { - "base": null, - "refs": { + "DescribeStacksRequest":{ + "base":null, + "refs":{ } }, - "Description": { - "base": null, - "refs": { - "CopyImageRequest$DestinationImageDescription": "

The description that the image will have when it is copied to the destination.

", - "CreateFleetRequest$Description": "

The description for display.

", - "CreateImageBuilderRequest$Description": "

The description for display.

", - "CreateStackRequest$Description": "

The description for display.

", - "UpdateFleetRequest$Description": "

The description for display.

", - "UpdateStackRequest$Description": "

The description for display.

" + "DescribeStacksResult":{ + "base":null, + "refs":{ } }, - "DirectoryConfig": { - "base": "

Configuration information for the directory used to join domains.

", - "refs": { - "CreateDirectoryConfigResult$DirectoryConfig": "

Information about the directory configuration.

", - "DirectoryConfigList$member": null, - "UpdateDirectoryConfigResult$DirectoryConfig": "

Information about the directory configuration.

" + "Description":{ + "base":null, + "refs":{ + "CopyImageRequest$DestinationImageDescription":"

The description that the image will have when it is copied to the destination.

", + "CreateFleetRequest$Description":"

The description for display.

", + "CreateImageBuilderRequest$Description":"

The description for display.

", + "CreateStackRequest$Description":"

The description for display.

", + "UpdateFleetRequest$Description":"

The description for display.

", + "UpdateStackRequest$Description":"

The description for display.

" } }, - "DirectoryConfigList": { - "base": null, - "refs": { - "DescribeDirectoryConfigsResult$DirectoryConfigs": "

Information about the directory configurations. Note that although the response syntax in this topic includes the account password, this password is not returned in the actual response.

" + "DirectoryConfig":{ + "base":"

Configuration information for the directory used to join domains.

", + "refs":{ + "CreateDirectoryConfigResult$DirectoryConfig":"

Information about the directory configuration.

", + "DirectoryConfigList$member":null, + "UpdateDirectoryConfigResult$DirectoryConfig":"

Information about the Directory Config object.

" } }, - "DirectoryName": { - "base": null, - "refs": { - "CreateDirectoryConfigRequest$DirectoryName": "

The fully qualified name of the directory (for example, corp.example.com).

", - "DeleteDirectoryConfigRequest$DirectoryName": "

The name of the directory configuration.

", - "DirectoryConfig$DirectoryName": "

The fully qualified name of the directory (for example, corp.example.com).

", - "DirectoryNameList$member": null, - "DomainJoinInfo$DirectoryName": "

The fully qualified name of the directory (for example, corp.example.com).

", - "UpdateDirectoryConfigRequest$DirectoryName": "

The name of the directory configuration.

" + "DirectoryConfigList":{ + "base":null, + "refs":{ + "DescribeDirectoryConfigsResult$DirectoryConfigs":"

Information about the directory configurations. Note that although the response syntax in this topic includes the account password, this password is not returned in the actual response.

" } }, - "DirectoryNameList": { - "base": null, - "refs": { - "DescribeDirectoryConfigsRequest$DirectoryNames": "

The directory names.

" + "DirectoryName":{ + "base":null, + "refs":{ + "CreateDirectoryConfigRequest$DirectoryName":"

The fully qualified name of the directory (for example, corp.example.com).

", + "DeleteDirectoryConfigRequest$DirectoryName":"

The name of the directory configuration.

", + "DirectoryConfig$DirectoryName":"

The fully qualified name of the directory (for example, corp.example.com).

", + "DirectoryNameList$member":null, + "DomainJoinInfo$DirectoryName":"

The fully qualified name of the directory (for example, corp.example.com).

", + "UpdateDirectoryConfigRequest$DirectoryName":"

The name of the Directory Config object.

" } }, - "DisassociateFleetRequest": { - "base": null, - "refs": { + "DirectoryNameList":{ + "base":null, + "refs":{ + "DescribeDirectoryConfigsRequest$DirectoryNames":"

The directory names.

" } }, - "DisassociateFleetResult": { - "base": null, - "refs": { + "DisassociateFleetRequest":{ + "base":null, + "refs":{ } }, - "DisplayName": { - "base": null, - "refs": { - "CreateFleetRequest$DisplayName": "

The fleet name for display.

", - "CreateImageBuilderRequest$DisplayName": "

The image builder name for display.

", - "CreateStackRequest$DisplayName": "

The stack name for display.

", - "UpdateFleetRequest$DisplayName": "

The fleet name for display.

", - "UpdateStackRequest$DisplayName": "

The stack name for display.

" + "DisassociateFleetResult":{ + "base":null, + "refs":{ } }, - "DomainJoinInfo": { - "base": "

Contains the information needed to join a Microsoft Active Directory domain.

", - "refs": { - "CreateFleetRequest$DomainJoinInfo": "

The information needed to join a Microsoft Active Directory domain.

", - "CreateImageBuilderRequest$DomainJoinInfo": "

The information needed to join a Microsoft Active Directory domain.

", - "Fleet$DomainJoinInfo": "

The information needed to join a Microsoft Active Directory domain.

", - "ImageBuilder$DomainJoinInfo": "

The information needed to join a Microsoft Active Directory domain.

", - "UpdateFleetRequest$DomainJoinInfo": "

The information needed to join a Microsoft Active Directory domain.

" + "DisplayName":{ + "base":null, + "refs":{ + "CreateFleetRequest$DisplayName":"

The fleet name for display.

", + "CreateImageBuilderRequest$DisplayName":"

The image builder name for display.

", + "CreateStackRequest$DisplayName":"

The stack name for display.

", + "UpdateFleetRequest$DisplayName":"

The fleet name for display.

", + "UpdateStackRequest$DisplayName":"

The stack name for display.

" } }, - "ErrorMessage": { - "base": "

The error message in the exception.

", - "refs": { - "ConcurrentModificationException$Message": null, - "IncompatibleImageException$Message": null, - "InvalidAccountStatusException$Message": null, - "InvalidParameterCombinationException$Message": null, - "InvalidRoleException$Message": null, - "LimitExceededException$Message": null, - "OperationNotPermittedException$Message": null, - "ResourceAlreadyExistsException$Message": null, - "ResourceInUseException$Message": null, - "ResourceNotAvailableException$Message": null, - "ResourceNotFoundException$Message": null + "Domain":{ + "base":"GSuite domain for GDrive integration.", + "refs":{ + "DomainList$member":null } }, - "ExpireSessionRequest": { - "base": null, - "refs": { + "DomainJoinInfo":{ + "base":"

Contains the information needed to join a Microsoft Active Directory domain.

", + "refs":{ + "CreateFleetRequest$DomainJoinInfo":"

The information needed to join a Microsoft Active Directory domain.

", + "CreateImageBuilderRequest$DomainJoinInfo":"

The information needed to join a Microsoft Active Directory domain.

", + "Fleet$DomainJoinInfo":"

The information needed to join a Microsoft Active Directory domain.

", + "ImageBuilder$DomainJoinInfo":"

The information needed to join a Microsoft Active Directory domain.

", + "UpdateFleetRequest$DomainJoinInfo":"

The information needed to join a Microsoft Active Directory domain.

" } }, - "ExpireSessionResult": { - "base": null, - "refs": { + "DomainList":{ + "base":null, + "refs":{ + "StorageConnector$Domains":"

The names of the domains for the G Suite account.

" } }, - "FeedbackURL": { - "base": null, - "refs": { - "CreateStackRequest$FeedbackURL": "

The URL that users are redirected to after they click the Send Feedback link. If no URL is specified, no Send Feedback link is displayed.

", - "Stack$FeedbackURL": "

The URL that users are redirected to after they click the Send Feedback link. If no URL is specified, no Send Feedback link is displayed.

", - "UpdateStackRequest$FeedbackURL": "

The URL that users are redirected to after they click the Send Feedback link. If no URL is specified, no Send Feedback link is displayed.

" + "ErrorMessage":{ + "base":"

The error message in the exception.

", + "refs":{ + "ConcurrentModificationException$Message":null, + "IncompatibleImageException$Message":null, + "InvalidAccountStatusException$Message":null, + "InvalidParameterCombinationException$Message":null, + "InvalidRoleException$Message":null, + "LimitExceededException$Message":null, + "OperationNotPermittedException$Message":null, + "ResourceAlreadyExistsException$Message":null, + "ResourceInUseException$Message":null, + "ResourceNotAvailableException$Message":null, + "ResourceNotFoundException$Message":null } }, - "Fleet": { - "base": "

Contains the parameters for a fleet.

", - "refs": { - "CreateFleetResult$Fleet": "

Information about the fleet.

", - "FleetList$member": null, - "UpdateFleetResult$Fleet": "

Information about the fleet.

" + "ExpireSessionRequest":{ + "base":null, + "refs":{ } }, - "FleetAttribute": { - "base": "

The fleet attribute.

", - "refs": { - "FleetAttributes$member": null + "ExpireSessionResult":{ + "base":null, + "refs":{ } }, - "FleetAttributes": { - "base": "

The fleet attributes.

", - "refs": { - "UpdateFleetRequest$AttributesToDelete": "

The fleet attributes to delete.

" + "FeedbackURL":{ + "base":null, + "refs":{ + "CreateStackRequest$FeedbackURL":"

The URL that users are redirected to after they click the Send Feedback link. If no URL is specified, no Send Feedback link is displayed.

", + "Stack$FeedbackURL":"

The URL that users are redirected to after they click the Send Feedback link. If no URL is specified, no Send Feedback link is displayed.

", + "UpdateStackRequest$FeedbackURL":"

The URL that users are redirected to after they click the Send Feedback link. If no URL is specified, no Send Feedback link is displayed.

" } }, - "FleetError": { - "base": "

Describes a fleet error.

", - "refs": { - "FleetErrors$member": null + "Fleet":{ + "base":"

Contains the parameters for a fleet.

", + "refs":{ + "CreateFleetResult$Fleet":"

Information about the fleet.

", + "FleetList$member":null, + "UpdateFleetResult$Fleet":"

Information about the fleet.

" } }, - "FleetErrorCode": { - "base": null, - "refs": { - "FleetError$ErrorCode": "

The error code.

", - "ResourceError$ErrorCode": "

The error code.

" + "FleetAttribute":{ + "base":"

The fleet attribute.

", + "refs":{ + "FleetAttributes$member":null } }, - "FleetErrors": { - "base": null, - "refs": { - "Fleet$FleetErrors": "

The fleet errors.

" + "FleetAttributes":{ + "base":"

The fleet attributes.

", + "refs":{ + "UpdateFleetRequest$AttributesToDelete":"

The fleet attributes to delete.

" } }, - "FleetList": { - "base": "

The fleets.

", - "refs": { - "DescribeFleetsResult$Fleets": "

Information about the fleets.

" + "FleetError":{ + "base":"

Describes a fleet error.

", + "refs":{ + "FleetErrors$member":null } }, - "FleetState": { - "base": null, - "refs": { - "Fleet$State": "

The current state for the fleet.

" + "FleetErrorCode":{ + "base":null, + "refs":{ + "FleetError$ErrorCode":"

The error code.

", + "ResourceError$ErrorCode":"

The error code.

" } }, - "FleetType": { - "base": null, - "refs": { - "CreateFleetRequest$FleetType": "

The fleet type.

ALWAYS_ON

Provides users with instant-on access to their apps. You are charged for all running instances in your fleet, even if no users are streaming apps.

ON_DEMAND

Provide users with access to applications after they connect, which takes one to two minutes. You are charged for instance streaming when users are connected and a small hourly fee for instances that are not streaming apps.

", - "Fleet$FleetType": "

The fleet type.

ALWAYS_ON

Provides users with instant-on access to their apps. You are charged for all running instances in your fleet, even if no users are streaming apps.

ON_DEMAND

Provide users with access to applications after they connect, which takes one to two minutes. You are charged for instance streaming when users are connected and a small hourly fee for instances that are not streaming apps.

" + "FleetErrors":{ + "base":null, + "refs":{ + "Fleet$FleetErrors":"

The fleet errors.

" } }, - "Image": { - "base": "

Describes an image.

", - "refs": { - "DeleteImageResult$Image": "

Information about the image.

", - "ImageList$member": null + "FleetList":{ + "base":"

The fleets.

", + "refs":{ + "DescribeFleetsResult$Fleets":"

Information about the fleets.

" } }, - "ImageBuilder": { - "base": "

Describes a streaming instance used for editing an image. New images are created from a snapshot through an image builder.

", - "refs": { - "CreateImageBuilderResult$ImageBuilder": "

Information about the image builder.

", - "DeleteImageBuilderResult$ImageBuilder": "

Information about the image builder.

", - "ImageBuilderList$member": null, - "StartImageBuilderResult$ImageBuilder": "

Information about the image builder.

", - "StopImageBuilderResult$ImageBuilder": "

Information about the image builder.

" + "FleetState":{ + "base":null, + "refs":{ + "Fleet$State":"

The current state for the fleet.

" } }, - "ImageBuilderList": { - "base": null, - "refs": { - "DescribeImageBuildersResult$ImageBuilders": "

Information about the image builders.

" + "FleetType":{ + "base":null, + "refs":{ + "CreateFleetRequest$FleetType":"

The fleet type.

ALWAYS_ON

Provides users with instant-on access to their apps. You are charged for all running instances in your fleet, even if no users are streaming apps.

ON_DEMAND

Provide users with access to applications after they connect, which takes one to two minutes. You are charged for instance streaming when users are connected and a small hourly fee for instances that are not streaming apps.

", + "Fleet$FleetType":"

The fleet type.

ALWAYS_ON

Provides users with instant-on access to their apps. You are charged for all running instances in your fleet, even if no users are streaming apps.

ON_DEMAND

Provide users with access to applications after they connect, which takes one to two minutes. You are charged for instance streaming when users are connected and a small hourly fee for instances that are not streaming apps.

" } }, - "ImageBuilderState": { - "base": null, - "refs": { - "ImageBuilder$State": "

The state of the image builder.

" + "Image":{ + "base":"

Describes an image.

", + "refs":{ + "DeleteImageResult$Image":"

Information about the image.

", + "ImageList$member":null } }, - "ImageBuilderStateChangeReason": { - "base": "

Describes the reason why the last image builder state change occurred.

", - "refs": { - "ImageBuilder$StateChangeReason": "

The reason why the last state change occurred.

" + "ImageBuilder":{ + "base":"

Describes a streaming instance used for editing an image. New images are created from a snapshot through an image builder.

", + "refs":{ + "CreateImageBuilderResult$ImageBuilder":"

Information about the image builder.

", + "DeleteImageBuilderResult$ImageBuilder":"

Information about the image builder.

", + "ImageBuilderList$member":null, + "StartImageBuilderResult$ImageBuilder":"

Information about the image builder.

", + "StopImageBuilderResult$ImageBuilder":"

Information about the image builder.

" } }, - "ImageBuilderStateChangeReasonCode": { - "base": null, - "refs": { - "ImageBuilderStateChangeReason$Code": "

The state change reason code.

" + "ImageBuilderList":{ + "base":null, + "refs":{ + "DescribeImageBuildersResult$ImageBuilders":"

Information about the image builders.

" } }, - "ImageList": { - "base": null, - "refs": { - "DescribeImagesResult$Images": "

Information about the images.

" + "ImageBuilderState":{ + "base":null, + "refs":{ + "ImageBuilder$State":"

The state of the image builder.

" } }, - "ImageState": { - "base": null, - "refs": { - "Image$State": "

The image starts in the PENDING state. If image creation succeeds, the state is AVAILABLE. If image creation fails, the state is FAILED.

" + "ImageBuilderStateChangeReason":{ + "base":"

Describes the reason why the last image builder state change occurred.

", + "refs":{ + "ImageBuilder$StateChangeReason":"

The reason why the last state change occurred.

" } }, - "ImageStateChangeReason": { - "base": "

Describes the reason why the last image state change occurred.

", - "refs": { - "Image$StateChangeReason": "

The reason why the last state change occurred.

" + "ImageBuilderStateChangeReasonCode":{ + "base":null, + "refs":{ + "ImageBuilderStateChangeReason$Code":"

The state change reason code.

" } }, - "ImageStateChangeReasonCode": { - "base": null, - "refs": { - "ImageStateChangeReason$Code": "

The state change reason code.

" + "ImageList":{ + "base":null, + "refs":{ + "DescribeImagesResult$Images":"

Information about the images.

" } }, - "IncompatibleImageException": { - "base": "

The image does not support storage connectors.

", - "refs": { + "ImageState":{ + "base":null, + "refs":{ + "Image$State":"

The image starts in the PENDING state. If image creation succeeds, the state is AVAILABLE. If image creation fails, the state is FAILED.

" } }, - "Integer": { - "base": null, - "refs": { - "ComputeCapacity$DesiredInstances": "

The desired number of streaming instances.

", - "ComputeCapacityStatus$Desired": "

The desired number of streaming instances.

", - "ComputeCapacityStatus$Running": "

The total number of simultaneous streaming instances that are running.

", - "ComputeCapacityStatus$InUse": "

The number of instances in use for streaming.

", - "ComputeCapacityStatus$Available": "

The number of currently available instances that can be used to stream sessions.

", - "CreateFleetRequest$MaxUserDurationInSeconds": "

The maximum time that a streaming session can run, in seconds. Specify a value between 600 and 57600.

", - "CreateFleetRequest$DisconnectTimeoutInSeconds": "

The time after disconnection when a session is considered to have ended, in seconds. If a user who was disconnected reconnects within this time interval, the user is connected to their previous session. Specify a value between 60 and 57600.

", - "DescribeDirectoryConfigsRequest$MaxResults": "

The maximum size of each page of results.

", - "DescribeImageBuildersRequest$MaxResults": "

The maximum size of each page of results.

", - "DescribeSessionsRequest$Limit": "

The size of each page of results. The default value is 20 and the maximum value is 50.

", - "Fleet$MaxUserDurationInSeconds": "

The maximum time that a streaming session can run, in seconds. Specify a value between 600 and 57600.

", - "Fleet$DisconnectTimeoutInSeconds": "

The time after disconnection when a session is considered to have ended, in seconds. If a user who was disconnected reconnects within this time interval, the user is connected to their previous session. Specify a value between 60 and 57600.

", - "UpdateFleetRequest$MaxUserDurationInSeconds": "

The maximum time that a streaming session can run, in seconds. Specify a value between 600 and 57600.

", - "UpdateFleetRequest$DisconnectTimeoutInSeconds": "

The time after disconnection when a session is considered to have ended, in seconds. If a user who was disconnected reconnects within this time interval, the user is connected to their previous session. Specify a value between 60 and 57600.

" + "ImageStateChangeReason":{ + "base":"

Describes the reason why the last image state change occurred.

", + "refs":{ + "Image$StateChangeReason":"

The reason why the last state change occurred.

" } }, - "InvalidAccountStatusException": { - "base": "

The resource cannot be created because your AWS account is suspended. For assistance, contact AWS Support.

", - "refs": { + "ImageStateChangeReasonCode":{ + "base":null, + "refs":{ + "ImageStateChangeReason$Code":"

The state change reason code.

" } }, - "InvalidParameterCombinationException": { - "base": "

Indicates an incorrect combination of parameters, or a missing parameter.

", - "refs": { + "IncompatibleImageException":{ + "base":"

The image does not support storage connectors.

", + "refs":{ } }, - "InvalidRoleException": { - "base": "

The specified role is invalid.

", - "refs": { + "Integer":{ + "base":null, + "refs":{ + "ComputeCapacity$DesiredInstances":"

The desired number of streaming instances.

", + "ComputeCapacityStatus$Desired":"

The desired number of streaming instances.

", + "ComputeCapacityStatus$Running":"

The total number of simultaneous streaming instances that are running.

", + "ComputeCapacityStatus$InUse":"

The number of instances in use for streaming.

", + "ComputeCapacityStatus$Available":"

The number of currently available instances that can be used to stream sessions.

", + "CreateFleetRequest$MaxUserDurationInSeconds":"

The maximum time that a streaming session can run, in seconds. Specify a value between 600 and 57600.

", + "CreateFleetRequest$DisconnectTimeoutInSeconds":"

The time after disconnection when a session is considered to have ended, in seconds. If a user who was disconnected reconnects within this time interval, the user is connected to their previous session. Specify a value between 60 and 57600.

", + "DescribeDirectoryConfigsRequest$MaxResults":"

The maximum size of each page of results.

", + "DescribeImageBuildersRequest$MaxResults":"

The maximum size of each page of results.

", + "DescribeSessionsRequest$Limit":"

The size of each page of results. The default value is 20 and the maximum value is 50.

", + "Fleet$MaxUserDurationInSeconds":"

The maximum time that a streaming session can run, in seconds. Specify a value between 600 and 57600.

", + "Fleet$DisconnectTimeoutInSeconds":"

The time after disconnection when a session is considered to have ended, in seconds. If a user who was disconnected reconnects within this time interval, the user is connected to their previous session. Specify a value between 60 and 57600.

", + "UpdateFleetRequest$MaxUserDurationInSeconds":"

The maximum time that a streaming session can run, in seconds. Specify a value between 600 and 57600.

", + "UpdateFleetRequest$DisconnectTimeoutInSeconds":"

The time after disconnection when a session is considered to have ended, in seconds. If a user who was disconnected reconnects within this time interval, the user is connected to their previous session. Specify a value between 60 and 57600.

" } }, - "LimitExceededException": { - "base": "

The requested limit exceeds the permitted limit for an account.

", - "refs": { + "InvalidAccountStatusException":{ + "base":"

The resource cannot be created because your AWS account is suspended. For assistance, contact AWS Support.

", + "refs":{ } }, - "ListAssociatedFleetsRequest": { - "base": null, - "refs": { + "InvalidParameterCombinationException":{ + "base":"

Indicates an incorrect combination of parameters, or a missing parameter.

", + "refs":{ } }, - "ListAssociatedFleetsResult": { - "base": null, - "refs": { + "InvalidRoleException":{ + "base":"

The specified role is invalid.

", + "refs":{ } }, - "ListAssociatedStacksRequest": { - "base": null, - "refs": { + "LimitExceededException":{ + "base":"

The requested limit exceeds the permitted limit for an account.

", + "refs":{ } }, - "ListAssociatedStacksResult": { - "base": null, - "refs": { + "ListAssociatedFleetsRequest":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceRequest": { - "base": null, - "refs": { + "ListAssociatedFleetsResult":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceResponse": { - "base": null, - "refs": { + "ListAssociatedStacksRequest":{ + "base":null, + "refs":{ } }, - "Long": { - "base": null, - "refs": { - "CreateImageBuilderStreamingURLRequest$Validity": "

The time that the streaming URL will be valid, in seconds. Specify a value between 1 and 604800 seconds. The default is 3600 seconds.

", - "CreateStreamingURLRequest$Validity": "

The time that the streaming URL will be valid, in seconds. Specify a value between 1 and 604800 seconds. The default is 60 seconds.

" + "ListAssociatedStacksResult":{ + "base":null, + "refs":{ } }, - "Metadata": { - "base": null, - "refs": { - "Application$Metadata": "

Additional attributes that describe the application.

" + "ListTagsForResourceRequest":{ + "base":null, + "refs":{ } }, - "Name": { - "base": null, - "refs": { - "CopyImageRequest$SourceImageName": "

The name of the image to copy.

", - "CopyImageRequest$DestinationImageName": "

The name that the image will have when it is copied to the destination.

", - "CopyImageResponse$DestinationImageName": "

The name of the destination image.

", - "CreateFleetRequest$Name": "

A unique name for the fleet.

", - "CreateImageBuilderRequest$Name": "

A unique name for the image builder.

", - "DeleteImageBuilderRequest$Name": "

The name of the image builder.

", - "DeleteImageRequest$Name": "

The name of the image.

" + "ListTagsForResourceResponse":{ + "base":null, + "refs":{ } }, - "OperationNotPermittedException": { - "base": "

The attempted operation is not permitted.

", - "refs": { + "Long":{ + "base":null, + "refs":{ + "CreateImageBuilderStreamingURLRequest$Validity":"

The time that the streaming URL will be valid, in seconds. Specify a value between 1 and 604800 seconds. The default is 3600 seconds.

", + "CreateStreamingURLRequest$Validity":"

The time that the streaming URL will be valid, in seconds. Specify a value between 1 and 604800 seconds. The default is 60 seconds.

" } }, - "OrganizationalUnitDistinguishedName": { - "base": null, - "refs": { - "DomainJoinInfo$OrganizationalUnitDistinguishedName": "

The distinguished name of the organizational unit for computer accounts.

", - "OrganizationalUnitDistinguishedNamesList$member": null + "Metadata":{ + "base":null, + "refs":{ + "Application$Metadata":"

Additional attributes that describe the application.

" } }, - "OrganizationalUnitDistinguishedNamesList": { - "base": null, - "refs": { - "CreateDirectoryConfigRequest$OrganizationalUnitDistinguishedNames": "

The distinguished names of the organizational units for computer accounts.

", - "DirectoryConfig$OrganizationalUnitDistinguishedNames": "

The distinguished names of the organizational units for computer accounts.

", - "UpdateDirectoryConfigRequest$OrganizationalUnitDistinguishedNames": "

The distinguished names of the organizational units for computer accounts.

" + "Name":{ + "base":null, + "refs":{ + "CopyImageRequest$SourceImageName":"

The name of the image to copy.

", + "CopyImageRequest$DestinationImageName":"

The name that the image will have when it is copied to the destination.

", + "CopyImageResponse$DestinationImageName":"

The name of the destination image.

", + "CreateFleetRequest$Name":"

A unique name for the fleet.

", + "CreateImageBuilderRequest$Name":"

A unique name for the image builder.

", + "CreateStackRequest$Name":"

The name of the stack.

", + "DeleteImageBuilderRequest$Name":"

The name of the image builder.

", + "DeleteImageRequest$Name":"

The name of the image.

" } }, - "PlatformType": { - "base": null, - "refs": { - "Image$Platform": "

The operating system platform of the image.

", - "ImageBuilder$Platform": "

The operating system platform of the image builder.

" + "NetworkAccessConfiguration":{ + "base":"

The network details of the fleet instance for the streaming session.

", + "refs":{ + "Session$NetworkAccessConfiguration":"

The network details for the streaming session.

" } }, - "RedirectURL": { - "base": null, - "refs": { - "CreateStackRequest$RedirectURL": "

The URL that users are redirected to after their streaming session ends.

", - "Stack$RedirectURL": "

The URL that users are redirected to after their streaming session ends.

", - "UpdateStackRequest$RedirectURL": "

The URL that users are redirected to after their streaming session ends.

" + "OperationNotPermittedException":{ + "base":"

The attempted operation is not permitted.

", + "refs":{ } }, - "RegionName": { - "base": null, - "refs": { - "CopyImageRequest$DestinationRegion": "

The destination region to which the image will be copied. This parameter is required, even if you are copying an image within the same region.

" + "OrganizationalUnitDistinguishedName":{ + "base":null, + "refs":{ + "DomainJoinInfo$OrganizationalUnitDistinguishedName":"

The distinguished name of the organizational unit for computer accounts.

", + "OrganizationalUnitDistinguishedNamesList$member":null } }, - "ResourceAlreadyExistsException": { - "base": "

The specified resource already exists.

", - "refs": { + "OrganizationalUnitDistinguishedNamesList":{ + "base":null, + "refs":{ + "CreateDirectoryConfigRequest$OrganizationalUnitDistinguishedNames":"

The distinguished names of the organizational units for computer accounts.

", + "DirectoryConfig$OrganizationalUnitDistinguishedNames":"

The distinguished names of the organizational units for computer accounts.

", + "UpdateDirectoryConfigRequest$OrganizationalUnitDistinguishedNames":"

The distinguished names of the organizational units for computer accounts.

" } }, - "ResourceError": { - "base": "

Describes a resource error.

", - "refs": { - "ResourceErrors$member": null + "Permission":{ + "base":null, + "refs":{ + "UserSetting$Permission":"

Indicates whether the action is enabled or disabled.

" } }, - "ResourceErrors": { - "base": null, - "refs": { - "ImageBuilder$ImageBuilderErrors": "

The image builder errors.

" + "PlatformType":{ + "base":null, + "refs":{ + "Image$Platform":"

The operating system platform of the image.

", + "ImageBuilder$Platform":"

The operating system platform of the image builder.

" } }, - "ResourceIdentifier": { - "base": "

The ARN of the resource.

", - "refs": { - "StorageConnector$ResourceIdentifier": "

The ARN of the storage connector.

" + "RedirectURL":{ + "base":null, + "refs":{ + "CreateStackRequest$RedirectURL":"

The URL that users are redirected to after their streaming session ends.

", + "Stack$RedirectURL":"

The URL that users are redirected to after their streaming session ends.

", + "UpdateStackRequest$RedirectURL":"

The URL that users are redirected to after their streaming session ends.

" } }, - "ResourceInUseException": { - "base": "

The specified resource is in use.

", - "refs": { + "RegionName":{ + "base":null, + "refs":{ + "CopyImageRequest$DestinationRegion":"

The destination region to which the image will be copied. This parameter is required, even if you are copying an image within the same region.

" } }, - "ResourceNotAvailableException": { - "base": "

The specified resource exists and is not in use, but isn't available.

", - "refs": { + "ResourceAlreadyExistsException":{ + "base":"

The specified resource already exists.

", + "refs":{ } }, - "ResourceNotFoundException": { - "base": "

The specified resource was not found.

", - "refs": { + "ResourceError":{ + "base":"

Describes a resource error.

", + "refs":{ + "ResourceErrors$member":null } }, - "SecurityGroupIdList": { - "base": "

The security group IDs.

", - "refs": { - "VpcConfig$SecurityGroupIds": "

The security groups for the fleet.

" + "ResourceErrors":{ + "base":null, + "refs":{ + "ImageBuilder$ImageBuilderErrors":"

The image builder errors.

" } }, - "ServiceAccountCredentials": { - "base": "

Describes the credentials for the service account used by the streaming instance to connect to the directory.

", - "refs": { - "CreateDirectoryConfigRequest$ServiceAccountCredentials": "

The credentials for the service account used by the streaming instance to connect to the directory.

", - "DirectoryConfig$ServiceAccountCredentials": "

The credentials for the service account used by the streaming instance to connect to the directory.

", - "UpdateDirectoryConfigRequest$ServiceAccountCredentials": "

The credentials for the service account used by the streaming instance to connect to the directory.

" + "ResourceIdentifier":{ + "base":"

The ARN of the resource.

", + "refs":{ + "StorageConnector$ResourceIdentifier":"

The ARN of the storage connector.

" } }, - "Session": { - "base": "

Describes a streaming session.

", - "refs": { - "SessionList$member": null + "ResourceInUseException":{ + "base":"

The specified resource is in use.

", + "refs":{ } }, - "SessionList": { - "base": "

List of sessions.

", - "refs": { - "DescribeSessionsResult$Sessions": "

Information about the streaming sessions.

" + "ResourceNotAvailableException":{ + "base":"

The specified resource exists and is not in use, but isn't available.

", + "refs":{ } }, - "SessionState": { - "base": "

Possible values for the state of a streaming session.

", - "refs": { - "Session$State": "

The current state of the streaming session.

" + "ResourceNotFoundException":{ + "base":"

The specified resource was not found.

", + "refs":{ } }, - "Stack": { - "base": "

Describes a stack.

", - "refs": { - "CreateStackResult$Stack": "

Information about the stack.

", - "StackList$member": null, - "UpdateStackResult$Stack": "

Information about the stack.

" + "SecurityGroupIdList":{ + "base":"

The security group IDs.

", + "refs":{ + "VpcConfig$SecurityGroupIds":"

The security groups for the fleet.

" } }, - "StackAttribute": { - "base": null, - "refs": { - "StackAttributes$member": null + "ServiceAccountCredentials":{ + "base":"

Describes the credentials for the service account used by the streaming instance to connect to the directory.

", + "refs":{ + "CreateDirectoryConfigRequest$ServiceAccountCredentials":"

The credentials for the service account used by the streaming instance to connect to the directory.

", + "DirectoryConfig$ServiceAccountCredentials":"

The credentials for the service account used by the streaming instance to connect to the directory.

", + "UpdateDirectoryConfigRequest$ServiceAccountCredentials":"

The credentials for the service account used by the streaming instance to connect to the directory.

" } }, - "StackAttributes": { - "base": null, - "refs": { - "UpdateStackRequest$AttributesToDelete": "

The stack attributes to delete.

" + "Session":{ + "base":"

Describes a streaming session.

", + "refs":{ + "SessionList$member":null } }, - "StackError": { - "base": "

Describes a stack error.

", - "refs": { - "StackErrors$member": null + "SessionList":{ + "base":"

List of sessions.

", + "refs":{ + "DescribeSessionsResult$Sessions":"

Information about the streaming sessions.

" } }, - "StackErrorCode": { - "base": null, - "refs": { - "StackError$ErrorCode": "

The error code.

" + "SessionState":{ + "base":"

Possible values for the state of a streaming session.

", + "refs":{ + "Session$State":"

The current state of the streaming session.

" } }, - "StackErrors": { - "base": "

The stack errors.

", - "refs": { - "Stack$StackErrors": "

The errors for the stack.

" + "Stack":{ + "base":"

Describes a stack.

", + "refs":{ + "CreateStackResult$Stack":"

Information about the stack.

", + "StackList$member":null, + "UpdateStackResult$Stack":"

Information about the stack.

" } }, - "StackList": { - "base": "

The stacks.

", - "refs": { - "DescribeStacksResult$Stacks": "

Information about the stacks.

" + "StackAttribute":{ + "base":null, + "refs":{ + "StackAttributes$member":null } }, - "StartFleetRequest": { - "base": null, - "refs": { + "StackAttributes":{ + "base":null, + "refs":{ + "UpdateStackRequest$AttributesToDelete":"

The stack attributes to delete.

" } }, - "StartFleetResult": { - "base": null, - "refs": { + "StackError":{ + "base":"

Describes a stack error.

", + "refs":{ + "StackErrors$member":null } }, - "StartImageBuilderRequest": { - "base": null, - "refs": { + "StackErrorCode":{ + "base":null, + "refs":{ + "StackError$ErrorCode":"

The error code.

" } }, - "StartImageBuilderResult": { - "base": null, - "refs": { + "StackErrors":{ + "base":"

The stack errors.

", + "refs":{ + "Stack$StackErrors":"

The errors for the stack.

" } }, - "StopFleetRequest": { - "base": null, - "refs": { - } - }, - "StopFleetResult": { - "base": null, - "refs": { - } - }, - "StopImageBuilderRequest": { - "base": null, - "refs": { - } - }, - "StopImageBuilderResult": { - "base": null, - "refs": { - } - }, - "StorageConnector": { - "base": "

Describes a storage connector.

", - "refs": { - "StorageConnectorList$member": null - } - }, - "StorageConnectorList": { - "base": "

The storage connectors.

", - "refs": { - "CreateStackRequest$StorageConnectors": "

The storage connectors to enable.

", - "Stack$StorageConnectors": "

The storage connectors to enable.

", - "UpdateStackRequest$StorageConnectors": "

The storage connectors to enable.

" - } - }, - "StorageConnectorType": { - "base": "

The type of storage connector.

", - "refs": { - "StorageConnector$ConnectorType": "

The type of storage connector.

" - } - }, - "StreamingUrlUserId": { - "base": null, - "refs": { - "CreateStreamingURLRequest$UserId": "

The ID of the user.

" - } - }, - "String": { - "base": null, - "refs": { - "Application$Name": "

The name of the application.

", - "Application$DisplayName": "

The application name for display.

", - "Application$IconURL": "

The URL for the application icon. This URL might be time-limited.

", - "Application$LaunchPath": "

The path to the application executable in the instance.

", - "Application$LaunchParameters": "

The arguments that are passed to the application at launch.

", - "AssociateFleetRequest$FleetName": "

The name of the fleet.

", - "AssociateFleetRequest$StackName": "

The name of the stack.

", - "CreateFleetRequest$ImageName": "

The name of the image used to create the fleet.

", - "CreateFleetRequest$InstanceType": "

The instance type to use when launching fleet instances. The following instance types are available:

  • stream.standard.medium

  • stream.standard.large

  • stream.compute.large

  • stream.compute.xlarge

  • stream.compute.2xlarge

  • stream.compute.4xlarge

  • stream.compute.8xlarge

  • stream.memory.large

  • stream.memory.xlarge

  • stream.memory.2xlarge

  • stream.memory.4xlarge

  • stream.memory.8xlarge

  • stream.graphics-design.large

  • stream.graphics-design.xlarge

  • stream.graphics-design.2xlarge

  • stream.graphics-design.4xlarge

  • stream.graphics-desktop.2xlarge

  • stream.graphics-pro.4xlarge

  • stream.graphics-pro.8xlarge

  • stream.graphics-pro.16xlarge

", - "CreateImageBuilderRequest$ImageName": "

The name of the image used to create the builder.

", - "CreateImageBuilderRequest$InstanceType": "

The instance type to use when launching the image builder.

", - "CreateImageBuilderStreamingURLRequest$Name": "

The name of the image builder.

", - "CreateImageBuilderStreamingURLResult$StreamingURL": "

The URL to start the AppStream 2.0 streaming session.

", - "CreateStackRequest$Name": "

The name of the stack.

", - "CreateStreamingURLRequest$StackName": "

The name of the stack.

", - "CreateStreamingURLRequest$FleetName": "

The name of the fleet.

", - "CreateStreamingURLRequest$ApplicationId": "

The name of the application to launch after the session starts. This is the name that you specified as Name in the Image Assistant.

", - "CreateStreamingURLRequest$SessionContext": "

The session context. For more information, see Session Context in the Amazon AppStream 2.0 Developer Guide.

", - "CreateStreamingURLResult$StreamingURL": "

The URL to start the AppStream 2.0 streaming session.

", - "DeleteFleetRequest$Name": "

The name of the fleet.

", - "DeleteStackRequest$Name": "

The name of the stack.

", - "DescribeDirectoryConfigsRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "DescribeDirectoryConfigsResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "DescribeFleetsRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "DescribeFleetsResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "DescribeImageBuildersRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "DescribeImageBuildersResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "DescribeSessionsRequest$StackName": "

The name of the stack. This value is case-sensitive.

", - "DescribeSessionsRequest$FleetName": "

The name of the fleet. This value is case-sensitive.

", - "DescribeSessionsRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "DescribeSessionsResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "DescribeStacksRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "DescribeStacksResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "DisassociateFleetRequest$FleetName": "

The name of the fleet.

", - "DisassociateFleetRequest$StackName": "

The name of the stack.

", - "ExpireSessionRequest$SessionId": "

The ID of the streaming session.

", - "Fleet$Name": "

The name of the fleet.

", - "Fleet$DisplayName": "

The fleet name for display.

", - "Fleet$Description": "

The description for display.

", - "Fleet$ImageName": "

The name of the image used to create the fleet.

", - "Fleet$InstanceType": "

The instance type to use when launching fleet instances.

", - "FleetError$ErrorMessage": "

The error message.

", - "Image$Name": "

The name of the image.

", - "Image$DisplayName": "

The image name for display.

", - "Image$Description": "

The description for display.

", - "ImageBuilder$Name": "

The name of the image builder.

", - "ImageBuilder$Description": "

The description for display.

", - "ImageBuilder$DisplayName": "

The image builder name for display.

", - "ImageBuilder$InstanceType": "

The instance type for the image builder.

", - "ImageBuilderStateChangeReason$Message": "

The state change reason message.

", - "ImageStateChangeReason$Message": "

The state change reason message.

", - "ListAssociatedFleetsRequest$StackName": "

The name of the stack.

", - "ListAssociatedFleetsRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "ListAssociatedFleetsResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "ListAssociatedStacksRequest$FleetName": "

The name of the fleet.

", - "ListAssociatedStacksRequest$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", - "ListAssociatedStacksResult$NextToken": "

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", - "Metadata$key": null, - "Metadata$value": null, - "ResourceError$ErrorMessage": "

The error message.

", - "SecurityGroupIdList$member": null, - "Session$Id": "

The ID of the streaming session.

", - "Session$StackName": "

The name of the stack for the streaming session.

", - "Session$FleetName": "

The name of the fleet for the streaming session.

", - "Stack$Name": "

The name of the stack.

", - "Stack$Description": "

The description for display.

", - "Stack$DisplayName": "

The stack name for display.

", - "StackError$ErrorMessage": "

The error message.

", - "StartFleetRequest$Name": "

The name of the fleet.

", - "StartImageBuilderRequest$Name": "

The name of the image builder.

", - "StopFleetRequest$Name": "

The name of the fleet.

", - "StopImageBuilderRequest$Name": "

The name of the image builder.

", - "StringList$member": null, - "SubnetIdList$member": null, - "UpdateFleetRequest$ImageName": "

The name of the image used to create the fleet.

", - "UpdateFleetRequest$Name": "

A unique name for the fleet.

", - "UpdateFleetRequest$InstanceType": "

The instance type to use when launching fleet instances. The following instance types are available:

  • stream.standard.medium

  • stream.standard.large

  • stream.compute.large

  • stream.compute.xlarge

  • stream.compute.2xlarge

  • stream.compute.4xlarge

  • stream.compute.8xlarge

  • stream.memory.large

  • stream.memory.xlarge

  • stream.memory.2xlarge

  • stream.memory.4xlarge

  • stream.memory.8xlarge

  • stream.graphics-design.large

  • stream.graphics-design.xlarge

  • stream.graphics-design.2xlarge

  • stream.graphics-design.4xlarge

  • stream.graphics-desktop.2xlarge

  • stream.graphics-pro.4xlarge

  • stream.graphics-pro.8xlarge

  • stream.graphics-pro.16xlarge

", - "UpdateStackRequest$Name": "

The name of the stack.

" - } - }, - "StringList": { - "base": null, - "refs": { - "DescribeFleetsRequest$Names": "

The names of the fleets to describe.

", - "DescribeImageBuildersRequest$Names": "

The names of the image builders to describe.

", - "DescribeImagesRequest$Names": "

The names of the images to describe.

", - "DescribeStacksRequest$Names": "

The names of the stacks to describe.

", - "ListAssociatedFleetsResult$Names": "

The names of the fleets.

", - "ListAssociatedStacksResult$Names": "

The names of the stacks.

" - } - }, - "SubnetIdList": { - "base": "

The subnet IDs.

", - "refs": { - "VpcConfig$SubnetIds": "

The subnets to which a network interface is established from the fleet instance.

" - } - }, - "TagKey": { - "base": null, - "refs": { - "TagKeyList$member": null, - "Tags$key": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "UntagResourceRequest$TagKeys": "

The tag keys for the tags to disassociate.

" - } - }, - "TagResourceRequest": { - "base": null, - "refs": { - } - }, - "TagResourceResponse": { - "base": null, - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tags$value": null - } - }, - "Tags": { - "base": null, - "refs": { - "ListTagsForResourceResponse$Tags": "

The information about the tags.

", - "TagResourceRequest$Tags": "

The tags to associate. A tag is a key-value pair (the value is optional). For example, Environment=Test, or, if you do not specify a value, Environment=.

If you do not specify a value, we set the value to an empty string.

" - } - }, - "Timestamp": { - "base": null, - "refs": { - "CreateImageBuilderStreamingURLResult$Expires": "

The elapsed time, in seconds after the Unix epoch, when this URL expires.

", - "CreateStreamingURLResult$Expires": "

The elapsed time, in seconds after the Unix epoch, when this URL expires.

", - "DirectoryConfig$CreatedTime": "

The time the directory configuration was created.

", - "Fleet$CreatedTime": "

The time the fleet was created.

", - "Image$CreatedTime": "

The time the image was created.

", - "Image$PublicBaseImageReleasedDate": "

The release date of the public base image. For private images, this date is the release date of the base image from which the image was created.

", - "ImageBuilder$CreatedTime": "

The time stamp when the image builder was created.

", - "ResourceError$ErrorTimestamp": "

The time the error occurred.

", - "Stack$CreatedTime": "

The time the stack was created.

" + "StackList":{ + "base":"

The stacks.

", + "refs":{ + "DescribeStacksResult$Stacks":"

Information about the stacks.

" } }, - "UntagResourceRequest": { - "base": null, - "refs": { + "StartFleetRequest":{ + "base":null, + "refs":{ } }, - "UntagResourceResponse": { - "base": null, - "refs": { + "StartFleetResult":{ + "base":null, + "refs":{ } }, - "UpdateDirectoryConfigRequest": { - "base": null, - "refs": { + "StartImageBuilderRequest":{ + "base":null, + "refs":{ } }, - "UpdateDirectoryConfigResult": { - "base": null, - "refs": { + "StartImageBuilderResult":{ + "base":null, + "refs":{ } }, - "UpdateFleetRequest": { - "base": null, - "refs": { + "StopFleetRequest":{ + "base":null, + "refs":{ + } + }, + "StopFleetResult":{ + "base":null, + "refs":{ + } + }, + "StopImageBuilderRequest":{ + "base":null, + "refs":{ + } + }, + "StopImageBuilderResult":{ + "base":null, + "refs":{ + } + }, + "StorageConnector":{ + "base":"

Describes a connector to enable persistent storage for users.

", + "refs":{ + "StorageConnectorList$member":null + } + }, + "StorageConnectorList":{ + "base":"

The storage connectors.

", + "refs":{ + "CreateStackRequest$StorageConnectors":"

The storage connectors to enable.

", + "Stack$StorageConnectors":"

The storage connectors to enable.

", + "UpdateStackRequest$StorageConnectors":"

The storage connectors to enable.

" + } + }, + "StorageConnectorType":{ + "base":"

The type of storage connector.

", + "refs":{ + "StorageConnector$ConnectorType":"

The type of storage connector.

" + } + }, + "StreamingUrlUserId":{ + "base":null, + "refs":{ + "CreateStreamingURLRequest$UserId":"

The ID of the user.

" + } + }, + "String":{ + "base":null, + "refs":{ + "Application$Name":"

The name of the application.

", + "Application$DisplayName":"

The application name for display.

", + "Application$IconURL":"

The URL for the application icon. This URL might be time-limited.

", + "Application$LaunchPath":"

The path to the application executable in the instance.

", + "Application$LaunchParameters":"

The arguments that are passed to the application at launch.

", + "AssociateFleetRequest$FleetName":"

The name of the fleet.

", + "AssociateFleetRequest$StackName":"

The name of the stack.

", + "CreateFleetRequest$ImageName":"

The name of the image used to create the fleet.

", + "CreateFleetRequest$InstanceType":"

The instance type to use when launching fleet instances. The following instance types are available:

  • stream.standard.medium

  • stream.standard.large

  • stream.compute.large

  • stream.compute.xlarge

  • stream.compute.2xlarge

  • stream.compute.4xlarge

  • stream.compute.8xlarge

  • stream.memory.large

  • stream.memory.xlarge

  • stream.memory.2xlarge

  • stream.memory.4xlarge

  • stream.memory.8xlarge

  • stream.graphics-design.large

  • stream.graphics-design.xlarge

  • stream.graphics-design.2xlarge

  • stream.graphics-design.4xlarge

  • stream.graphics-desktop.2xlarge

  • stream.graphics-pro.4xlarge

  • stream.graphics-pro.8xlarge

  • stream.graphics-pro.16xlarge

", + "CreateImageBuilderRequest$ImageName":"

The name of the image used to create the builder.

", + "CreateImageBuilderRequest$InstanceType":"

The instance type to use when launching the image builder.

", + "CreateImageBuilderStreamingURLRequest$Name":"

The name of the image builder.

", + "CreateImageBuilderStreamingURLResult$StreamingURL":"

The URL to start the AppStream 2.0 streaming session.

", + "CreateStreamingURLRequest$StackName":"

The name of the stack.

", + "CreateStreamingURLRequest$FleetName":"

The name of the fleet.

", + "CreateStreamingURLRequest$ApplicationId":"

The name of the application to launch after the session starts. This is the name that you specified as Name in the Image Assistant.

", + "CreateStreamingURLRequest$SessionContext":"

The session context. For more information, see Session Context in the Amazon AppStream 2.0 Developer Guide.

", + "CreateStreamingURLResult$StreamingURL":"

The URL to start the AppStream 2.0 streaming session.

", + "DeleteFleetRequest$Name":"

The name of the fleet.

", + "DeleteStackRequest$Name":"

The name of the stack.

", + "DescribeDirectoryConfigsRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "DescribeDirectoryConfigsResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "DescribeFleetsRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "DescribeFleetsResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "DescribeImageBuildersRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "DescribeImageBuildersResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "DescribeSessionsRequest$StackName":"

The name of the stack. This value is case-sensitive.

", + "DescribeSessionsRequest$FleetName":"

The name of the fleet. This value is case-sensitive.

", + "DescribeSessionsRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "DescribeSessionsResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "DescribeStacksRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "DescribeStacksResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "DisassociateFleetRequest$FleetName":"

The name of the fleet.

", + "DisassociateFleetRequest$StackName":"

The name of the stack.

", + "ExpireSessionRequest$SessionId":"

The ID of the streaming session.

", + "Fleet$Name":"

The name of the fleet.

", + "Fleet$DisplayName":"

The fleet name for display.

", + "Fleet$Description":"

The description for display.

", + "Fleet$ImageName":"

The name of the image used to create the fleet.

", + "Fleet$InstanceType":"

The instance type to use when launching fleet instances.

", + "FleetError$ErrorMessage":"

The error message.

", + "Image$Name":"

The name of the image.

", + "Image$DisplayName":"

The image name for display.

", + "Image$Description":"

The description for display.

", + "ImageBuilder$Name":"

The name of the image builder.

", + "ImageBuilder$Description":"

The description for display.

", + "ImageBuilder$DisplayName":"

The image builder name for display.

", + "ImageBuilder$InstanceType":"

The instance type for the image builder.

", + "ImageBuilderStateChangeReason$Message":"

The state change reason message.

", + "ImageStateChangeReason$Message":"

The state change reason message.

", + "ListAssociatedFleetsRequest$StackName":"

The name of the stack.

", + "ListAssociatedFleetsRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "ListAssociatedFleetsResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "ListAssociatedStacksRequest$FleetName":"

The name of the fleet.

", + "ListAssociatedStacksRequest$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If this value is null, it retrieves the first page.

", + "ListAssociatedStacksResult$NextToken":"

The pagination token to use to retrieve the next page of results for this operation. If there are no more pages, this value is null.

", + "Metadata$key":null, + "Metadata$value":null, + "NetworkAccessConfiguration$EniPrivateIpAddress":"

The private IP address of the elastic network interface that is attached to instances in your VPC.

", + "NetworkAccessConfiguration$EniId":"

The resource identifier of the elastic network interface that is attached to instances in your VPC. All network interfaces have the eni-xxxxxxxx resource identifier.

", + "ResourceError$ErrorMessage":"

The error message.

", + "SecurityGroupIdList$member":null, + "Session$Id":"

The ID of the streaming session.

", + "Session$StackName":"

The name of the stack for the streaming session.

", + "Session$FleetName":"

The name of the fleet for the streaming session.

", + "Stack$Name":"

The name of the stack.

", + "Stack$Description":"

The description for display.

", + "Stack$DisplayName":"

The stack name for display.

", + "StackError$ErrorMessage":"

The error message.

", + "StartFleetRequest$Name":"

The name of the fleet.

", + "StartImageBuilderRequest$Name":"

The name of the image builder.

", + "StopFleetRequest$Name":"

The name of the fleet.

", + "StopImageBuilderRequest$Name":"

The name of the image builder.

", + "StringList$member":null, + "SubnetIdList$member":null, + "UpdateFleetRequest$ImageName":"

The name of the image used to create the fleet.

", + "UpdateFleetRequest$Name":"

A unique name for the fleet.

", + "UpdateFleetRequest$InstanceType":"

The instance type to use when launching fleet instances. The following instance types are available:

  • stream.standard.medium

  • stream.standard.large

  • stream.compute.large

  • stream.compute.xlarge

  • stream.compute.2xlarge

  • stream.compute.4xlarge

  • stream.compute.8xlarge

  • stream.memory.large

  • stream.memory.xlarge

  • stream.memory.2xlarge

  • stream.memory.4xlarge

  • stream.memory.8xlarge

  • stream.graphics-design.large

  • stream.graphics-design.xlarge

  • stream.graphics-design.2xlarge

  • stream.graphics-design.4xlarge

  • stream.graphics-desktop.2xlarge

  • stream.graphics-pro.4xlarge

  • stream.graphics-pro.8xlarge

  • stream.graphics-pro.16xlarge

", + "UpdateStackRequest$Name":"

The name of the stack.

" + } + }, + "StringList":{ + "base":null, + "refs":{ + "DescribeFleetsRequest$Names":"

The names of the fleets to describe.

", + "DescribeImageBuildersRequest$Names":"

The names of the image builders to describe.

", + "DescribeImagesRequest$Names":"

The names of the images to describe.

", + "DescribeStacksRequest$Names":"

The names of the stacks to describe.

", + "ListAssociatedFleetsResult$Names":"

The name of the fleet.

", + "ListAssociatedStacksResult$Names":"

The name of the stack.

" + } + }, + "SubnetIdList":{ + "base":"

The subnet IDs.

", + "refs":{ + "VpcConfig$SubnetIds":"

The subnets to which a network interface is established from the fleet instance.

" + } + }, + "TagKey":{ + "base":null, + "refs":{ + "TagKeyList$member":null, + "Tags$key":null + } + }, + "TagKeyList":{ + "base":null, + "refs":{ + "UntagResourceRequest$TagKeys":"

The tag keys for the tags to disassociate.

" + } + }, + "TagResourceRequest":{ + "base":null, + "refs":{ + } + }, + "TagResourceResponse":{ + "base":null, + "refs":{ + } + }, + "TagValue":{ + "base":null, + "refs":{ + "Tags$value":null + } + }, + "Tags":{ + "base":null, + "refs":{ + "ListTagsForResourceResponse$Tags":"

The information about the tags.

", + "TagResourceRequest$Tags":"

The tags to associate. A tag is a key-value pair (the value is optional). For example, Environment=Test, or, if you do not specify a value, Environment=.

If you do not specify a value, we set the value to an empty string.

" + } + }, + "Timestamp":{ + "base":null, + "refs":{ + "CreateImageBuilderStreamingURLResult$Expires":"

The elapsed time, in seconds after the Unix epoch, when this URL expires.

", + "CreateStreamingURLResult$Expires":"

The elapsed time, in seconds after the Unix epoch, when this URL expires.

", + "DirectoryConfig$CreatedTime":"

The time the directory configuration was created.

", + "Fleet$CreatedTime":"

The time the fleet was created.

", + "Image$CreatedTime":"

The time the image was created.

", + "Image$PublicBaseImageReleasedDate":"

The release date of the public base image. For private images, this date is the release date of the base image from which the image was created.

", + "ImageBuilder$CreatedTime":"

The time stamp when the image builder was created.

", + "ResourceError$ErrorTimestamp":"

The time the error occurred.

", + "Stack$CreatedTime":"

The time the stack was created.

" } }, - "UpdateFleetResult": { - "base": null, - "refs": { + "UntagResourceRequest":{ + "base":null, + "refs":{ } }, - "UpdateStackRequest": { - "base": null, - "refs": { + "UntagResourceResponse":{ + "base":null, + "refs":{ } }, - "UpdateStackResult": { - "base": null, - "refs": { + "UpdateDirectoryConfigRequest":{ + "base":null, + "refs":{ } }, - "UserId": { - "base": null, - "refs": { - "DescribeSessionsRequest$UserId": "

The user ID.

", - "Session$UserId": "

The identifier of the user for whom the session was created.

" + "UpdateDirectoryConfigResult":{ + "base":null, + "refs":{ } }, - "VisibilityType": { - "base": null, - "refs": { - "Image$Visibility": "

Indicates whether the image is public or private.

" + "UpdateFleetRequest":{ + "base":null, + "refs":{ } }, - "VpcConfig": { - "base": "

Describes VPC configuration information.

", - "refs": { - "CreateFleetRequest$VpcConfig": "

The VPC configuration for the fleet.

", - "CreateImageBuilderRequest$VpcConfig": "

The VPC configuration for the image builder. You can specify only one subnet.

", - "Fleet$VpcConfig": "

The VPC configuration for the fleet.

", - "ImageBuilder$VpcConfig": "

The VPC configuration of the image builder.

", - "UpdateFleetRequest$VpcConfig": "

The VPC configuration for the fleet.

" + "UpdateFleetResult":{ + "base":null, + "refs":{ + } + }, + "UpdateStackRequest":{ + "base":null, + "refs":{ + } + }, + "UpdateStackResult":{ + "base":null, + "refs":{ + } + }, + "UserId":{ + "base":null, + "refs":{ + "DescribeSessionsRequest$UserId":"

The user ID.

", + "Session$UserId":"

The identifier of the user for whom the session was created.

" + } + }, + "UserSetting":{ + "base":"

Describes an action and whether the action is enabled or disabled for users during their streaming sessions.

", + "refs":{ + "UserSettingList$member":null + } + }, + "UserSettingList":{ + "base":null, + "refs":{ + "CreateStackRequest$UserSettings":"

The actions that are enabled or disabled for users during their streaming sessions. By default, these actions are enabled.

", + "Stack$UserSettings":"

The actions that are enabled or disabled for users during their streaming sessions. By default these actions are enabled.

", + "UpdateStackRequest$UserSettings":"

The actions that are enabled or disabled for users during their streaming sessions. By default, these actions are enabled.

" + } + }, + "VisibilityType":{ + "base":null, + "refs":{ + "Image$Visibility":"

Indicates whether the image is public or private.

" + } + }, + "VpcConfig":{ + "base":"

Describes VPC configuration information.

", + "refs":{ + "CreateFleetRequest$VpcConfig":"

The VPC configuration for the fleet.

", + "CreateImageBuilderRequest$VpcConfig":"

The VPC configuration for the image builder. You can specify only one subnet.

", + "Fleet$VpcConfig":"

The VPC configuration for the fleet.

", + "ImageBuilder$VpcConfig":"

The VPC configuration of the image builder.

", + "UpdateFleetRequest$VpcConfig":"

The VPC configuration for the fleet.

" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/examples-1.json index 0ea7e3b0b..faff76894 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/examples-1.json @@ -1,5 +1,5 @@ { - "version": "1.0", - "examples": { + "version":"1.0", + "examples":{ } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/paginators-1.json index 5677bd8e4..f85268a04 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/paginators-1.json @@ -1,4 +1,4 @@ { - "pagination": { + "pagination":{ } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/smoke.json b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/smoke.json new file mode 100644 index 000000000..49dca4295 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/appstream/2016-12-01/smoke.json @@ -0,0 +1,11 @@ +{ + "version": 1, + "defaultRegion": "us-west-2", + "testCases": [ + { + "operationName": "DescribeStacks", + "input": {}, + "errorExpectedFromService": false + } + ] +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/api-2.json index d80187c88..32a9b5eda 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/api-2.json @@ -110,6 +110,8 @@ "type":"string", "enum":["PAYER"] }, + "AmortizedRecurringFee":{"type":"string"}, + "AmortizedUpfrontFee":{"type":"string"}, "AttributeType":{"type":"string"}, "AttributeValue":{"type":"string"}, "Attributes":{ @@ -482,6 +484,7 @@ "key":{"shape":"MetricName"}, "value":{"shape":"MetricValue"} }, + "NetRISavings":{"type":"string"}, "NextPageToken":{"type":"string"}, "NonNegativeInteger":{ "type":"integer", @@ -494,6 +497,7 @@ "CONVERTIBLE" ] }, + "OnDemandCostOfRIHoursUsed":{"type":"string"}, "OnDemandHours":{"type":"string"}, "PageSize":{"type":"integer"}, "PaymentOption":{ @@ -531,7 +535,13 @@ "UtilizationPercentage":{"shape":"UtilizationPercentage"}, "PurchasedHours":{"shape":"PurchasedHours"}, "TotalActualHours":{"shape":"TotalActualHours"}, - "UnusedHours":{"shape":"UnusedHours"} + "UnusedHours":{"shape":"UnusedHours"}, + "OnDemandCostOfRIHoursUsed":{"shape":"OnDemandCostOfRIHoursUsed"}, + "NetRISavings":{"shape":"NetRISavings"}, + "TotalPotentialRISavings":{"shape":"TotalPotentialRISavings"}, + "AmortizedUpfrontFee":{"shape":"AmortizedUpfrontFee"}, + "AmortizedRecurringFee":{"shape":"AmortizedRecurringFee"}, + "TotalAmortizedFee":{"shape":"TotalAmortizedFee"} } }, "ReservationCoverageGroup":{ @@ -659,6 +669,8 @@ ] }, "TotalActualHours":{"type":"string"}, + "TotalAmortizedFee":{"type":"string"}, + "TotalPotentialRISavings":{"type":"string"}, "TotalRunningHours":{"type":"string"}, "UnusedHours":{"type":"string"}, "UtilizationByTime":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/docs-2.json index c90748c75..6ac75a417 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ce/2017-10-25/docs-2.json @@ -17,6 +17,18 @@ "ReservationPurchaseRecommendation$AccountScope": "

The account scope that AWS recommends that you purchase this instance for. For example, you can purchase this reservation for an entire organization in AWS Organizations.

" } }, + "AmortizedRecurringFee": { + "base": null, + "refs": { + "ReservationAggregates$AmortizedRecurringFee": "

The monthly cost of your RI, amortized over the RI period.

" + } + }, + "AmortizedUpfrontFee": { + "base": null, + "refs": { + "ReservationAggregates$AmortizedUpfrontFee": "

The upfront cost of your RI, amortized over the RI period.

" + } + }, "AttributeType": { "base": null, "refs": { @@ -158,7 +170,7 @@ } }, "Expression": { - "base": "

Use Expression to filter by cost or by usage. There are two patterns:

  • Simple dimension values - You can set the dimension name and values for the filters that you plan to use. For example, you can filter for INSTANCE_TYPE==m4.xlarge OR INSTANCE_TYPE==c4.large. The Expression for that looks like this:

    { \"Dimensions\": { \"Key\": \"INSTANCE_TYPE\", \"Values\": [ \"m4.xlarge\", “c4.large” ] } }

    The list of dimension values are OR'd together to retrieve cost or usage data. You can create Expression and DimensionValues objects using either with* methods or set* methods in multiple lines.

  • Compound dimension values with logical operations - You can use multiple Expression types and the logical operators AND/OR/NOT to create a list of one or more Expression objects. This allows you to filter on more advanced options. For example, you can filter on ((INSTANCE_TYPE == m4.large OR INSTANCE_TYPE == m3.large) OR (TAG.Type == Type1)) AND (USAGE_TYPE != DataTransfer). The Expression for that looks like this:

    { \"And\": [ {\"Or\": [ {\"Dimensions\": { \"Key\": \"INSTANCE_TYPE\", \"Values\": [ \"m4.x.large\", \"c4.large\" ] }}, {\"Tag\": { \"Key\": \"TagName\", \"Values\": [\"Value1\"] } } ]}, {\"Not\": {\"dimensions\": { \"Key\": \"USAGE_TYPE\", \"Values\": [\"DataTransfer\"] }}} ] }

    Because each Expression can have only one operator, the service returns an error if more than one is specified. The following example shows an Expression object that creates an error.

    { \"And\": [ ... ], \"DimensionValues\": { \"Dimension\": \"USAGE_TYPE\", \"Values\": [ \"DataTransfer\" ] } }

", + "base": "

Use Expression to filter by cost or by usage. There are two patterns:

  • Simple dimension values - You can set the dimension name and values for the filters that you plan to use. For example, you can filter for INSTANCE_TYPE==m4.xlarge OR INSTANCE_TYPE==c4.large. The Expression for that looks like this:

    { \"Dimensions\": { \"Key\": \"INSTANCE_TYPE\", \"Values\": [ \"m4.xlarge\", “c4.large” ] } }

    The list of dimension values are OR'd together to retrieve cost or usage data. You can create Expression and DimensionValues objects using either with* methods or set* methods in multiple lines.

  • Compound dimension values with logical operations - You can use multiple Expression types and the logical operators AND/OR/NOT to create a list of one or more Expression objects. This allows you to filter on more advanced options. For example, you can filter on ((INSTANCE_TYPE == m4.large OR INSTANCE_TYPE == m3.large) OR (TAG.Type == Type1)) AND (USAGE_TYPE != DataTransfer). The Expression for that looks like this:

    { \"And\": [ {\"Or\": [ {\"Dimensions\": { \"Key\": \"INSTANCE_TYPE\", \"Values\": [ \"m4.x.large\", \"c4.large\" ] }}, {\"Tags\": { \"Key\": \"TagName\", \"Values\": [\"Value1\"] } } ]}, {\"Not\": {\"Dimensions\": { \"Key\": \"USAGE_TYPE\", \"Values\": [\"DataTransfer\"] }}} ] }

    Because each Expression can have only one operator, the service returns an error if more than one is specified. The following example shows an Expression object that creates an error.

    { \"And\": [ ... ], \"DimensionValues\": { \"Dimension\": \"USAGE_TYPE\", \"Values\": [ \"DataTransfer\" ] } }

", "refs": { "Expression$Not": "

Return results that don't match a Dimension object.

", "Expressions$member": null, @@ -404,6 +416,12 @@ "ResultByTime$Total": "

The total amount of cost or usage accrued during the time period.

" } }, + "NetRISavings": { + "base": null, + "refs": { + "ReservationAggregates$NetRISavings": "

How much you saved due to purchasing and utilizing RIs. This is calculated by subtracting the TotalAmortizedFee from the OnDemandCostOfRIHoursUsed.

" + } + }, "NextPageToken": { "base": null, "refs": { @@ -433,6 +451,12 @@ "EC2Specification$OfferingClass": "

Whether you want a recommendation for standard or convertible reservations.

" } }, + "OnDemandCostOfRIHoursUsed": { + "base": null, + "refs": { + "ReservationAggregates$OnDemandCostOfRIHoursUsed": "

How much your RIs would cost if charged On-Demand rates.

" + } + }, "OnDemandHours": { "base": null, "refs": { @@ -616,6 +640,18 @@ "ReservationAggregates$TotalActualHours": "

The total number of RI hours that you used.

" } }, + "TotalAmortizedFee": { + "base": null, + "refs": { + "ReservationAggregates$TotalAmortizedFee": "

The total cost of your RI, amortized over the RI period.

" + } + }, + "TotalPotentialRISavings": { + "base": null, + "refs": { + "ReservationAggregates$TotalPotentialRISavings": "

How much you could save if you use your entire reservation.

" + } + }, "TotalRunningHours": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/api-2.json similarity index 95% rename from vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/api-2.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/api-2.json index e166d0373..dea3930f3 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/api-2.json @@ -1,13 +1,13 @@ { "version":"2.0", "metadata":{ - "apiVersion":"2016-05-10", + "apiVersion":"2017-01-11", "endpointPrefix":"clouddirectory", "protocol":"rest-json", "serviceFullName":"Amazon CloudDirectory", "signatureVersion":"v4", "signingName":"clouddirectory", - "uid":"clouddirectory-2016-05-10" + "uid":"clouddirectory-2017-01-11" }, "operations":{ "AddFacetToObject":{ @@ -47,6 +47,7 @@ {"shape":"ValidationException"}, {"shape":"LimitExceededException"}, {"shape":"AccessDeniedException"}, + {"shape":"SchemaAlreadyExistsException"}, {"shape":"ResourceNotFoundException"}, {"shape":"InvalidAttachmentException"} ] @@ -596,6 +597,27 @@ {"shape":"FacetNotFoundException"} ] }, + "GetLinkAttributes":{ + "name":"GetLinkAttributes", + "http":{ + "method":"POST", + "requestUri":"/amazonclouddirectory/2017-01-11/typedlink/attributes/get", + "responseCode":200 + }, + "input":{"shape":"GetLinkAttributesRequest"}, + "output":{"shape":"GetLinkAttributesResponse"}, + "errors":[ + {"shape":"InternalServiceException"}, + {"shape":"InvalidArnException"}, + {"shape":"RetryableConflictException"}, + {"shape":"ValidationException"}, + {"shape":"LimitExceededException"}, + {"shape":"AccessDeniedException"}, + {"shape":"DirectoryNotEnabledException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"FacetValidationException"} + ] + }, "GetObjectAttributes":{ "name":"GetObjectAttributes", "http":{ @@ -843,6 +865,24 @@ {"shape":"NotIndexException"} ] }, + "ListManagedSchemaArns":{ + "name":"ListManagedSchemaArns", + "http":{ + "method":"POST", + "requestUri":"/amazonclouddirectory/2017-01-11/schema/managed", + "responseCode":200 + }, + "input":{"shape":"ListManagedSchemaArnsRequest"}, + "output":{"shape":"ListManagedSchemaArnsResponse"}, + "errors":[ + {"shape":"InternalServiceException"}, + {"shape":"InvalidArnException"}, + {"shape":"ValidationException"}, + {"shape":"AccessDeniedException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidNextTokenException"} + ] + }, "ListObjectAttributes":{ "name":"ListObjectAttributes", "http":{ @@ -1215,11 +1255,33 @@ {"shape":"LimitExceededException"}, {"shape":"AccessDeniedException"}, {"shape":"InvalidFacetUpdateException"}, + {"shape":"FacetValidationException"}, {"shape":"ResourceNotFoundException"}, {"shape":"FacetNotFoundException"}, {"shape":"InvalidRuleException"} ] }, + "UpdateLinkAttributes":{ + "name":"UpdateLinkAttributes", + "http":{ + "method":"POST", + "requestUri":"/amazonclouddirectory/2017-01-11/typedlink/attributes/update", + "responseCode":200 + }, + "input":{"shape":"UpdateLinkAttributesRequest"}, + "output":{"shape":"UpdateLinkAttributesResponse"}, + "errors":[ + {"shape":"InternalServiceException"}, + {"shape":"InvalidArnException"}, + {"shape":"RetryableConflictException"}, + {"shape":"ValidationException"}, + {"shape":"LimitExceededException"}, + {"shape":"AccessDeniedException"}, + {"shape":"DirectoryNotEnabledException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"FacetValidationException"} + ] + }, "UpdateObjectAttributes":{ "name":"UpdateObjectAttributes", "http":{ @@ -1238,6 +1300,7 @@ {"shape":"AccessDeniedException"}, {"shape":"DirectoryNotEnabledException"}, {"shape":"ResourceNotFoundException"}, + {"shape":"LinkNameAlreadyInUseException"}, {"shape":"FacetValidationException"} ] }, @@ -1300,7 +1363,8 @@ {"shape":"IncompatibleSchemaException"}, {"shape":"AccessDeniedException"}, {"shape":"ResourceNotFoundException"}, - {"shape":"InvalidAttachmentException"} + {"shape":"InvalidAttachmentException"}, + {"shape":"SchemaAlreadyExistsException"} ] }, "UpgradePublishedSchema":{ @@ -1744,6 +1808,23 @@ "members":{ } }, + "BatchGetLinkAttributes":{ + "type":"structure", + "required":[ + "TypedLinkSpecifier", + "AttributeNames" + ], + "members":{ + "TypedLinkSpecifier":{"shape":"TypedLinkSpecifier"}, + "AttributeNames":{"shape":"AttributeNameList"} + } + }, + "BatchGetLinkAttributesResponse":{ + "type":"structure", + "members":{ + "Attributes":{"shape":"AttributeKeyAndValueList"} + } + }, "BatchGetObjectAttributes":{ "type":"structure", "required":[ @@ -1983,7 +2064,8 @@ "LookupPolicy":{"shape":"BatchLookupPolicy"}, "ListIndex":{"shape":"BatchListIndex"}, "ListOutgoingTypedLinks":{"shape":"BatchListOutgoingTypedLinks"}, - "ListIncomingTypedLinks":{"shape":"BatchListIncomingTypedLinks"} + "ListIncomingTypedLinks":{"shape":"BatchListIncomingTypedLinks"}, + "GetLinkAttributes":{"shape":"BatchGetLinkAttributes"} } }, "BatchReadOperationList":{ @@ -2041,7 +2123,8 @@ "LookupPolicy":{"shape":"BatchLookupPolicyResponse"}, "ListIndex":{"shape":"BatchListIndexResponse"}, "ListOutgoingTypedLinks":{"shape":"BatchListOutgoingTypedLinksResponse"}, - "ListIncomingTypedLinks":{"shape":"BatchListIncomingTypedLinksResponse"} + "ListIncomingTypedLinks":{"shape":"BatchListIncomingTypedLinksResponse"}, + "GetLinkAttributes":{"shape":"BatchGetLinkAttributesResponse"} } }, "BatchReferenceName":{"type":"string"}, @@ -2061,6 +2144,22 @@ "members":{ } }, + "BatchUpdateLinkAttributes":{ + "type":"structure", + "required":[ + "TypedLinkSpecifier", + "AttributeUpdates" + ], + "members":{ + "TypedLinkSpecifier":{"shape":"TypedLinkSpecifier"}, + "AttributeUpdates":{"shape":"LinkAttributeUpdateList"} + } + }, + "BatchUpdateLinkAttributesResponse":{ + "type":"structure", + "members":{ + } + }, "BatchUpdateObjectAttributes":{ "type":"structure", "required":[ @@ -2126,7 +2225,8 @@ "AttachToIndex":{"shape":"BatchAttachToIndex"}, "DetachFromIndex":{"shape":"BatchDetachFromIndex"}, "AttachTypedLink":{"shape":"BatchAttachTypedLink"}, - "DetachTypedLink":{"shape":"BatchDetachTypedLink"} + "DetachTypedLink":{"shape":"BatchDetachTypedLink"}, + "UpdateLinkAttributes":{"shape":"BatchUpdateLinkAttributes"} } }, "BatchWriteOperationList":{ @@ -2149,7 +2249,8 @@ "AttachToIndex":{"shape":"BatchAttachToIndexResponse"}, "DetachFromIndex":{"shape":"BatchDetachFromIndexResponse"}, "AttachTypedLink":{"shape":"BatchAttachTypedLinkResponse"}, - "DetachTypedLink":{"shape":"BatchDetachTypedLinkResponse"} + "DetachTypedLink":{"shape":"BatchDetachTypedLinkResponse"}, + "UpdateLinkAttributes":{"shape":"BatchUpdateLinkAttributesResponse"} } }, "BatchWriteOperationResponseList":{ @@ -2229,8 +2330,7 @@ "type":"structure", "required":[ "SchemaArn", - "Name", - "ObjectType" + "Name" ], "members":{ "SchemaArn":{ @@ -2240,7 +2340,8 @@ }, "Name":{"shape":"FacetName"}, "Attributes":{"shape":"FacetAttributeList"}, - "ObjectType":{"shape":"ObjectType"} + "ObjectType":{"shape":"ObjectType"}, + "FacetStyle":{"shape":"FacetStyle"} } }, "CreateFacetResponse":{ @@ -2611,7 +2712,8 @@ "type":"structure", "members":{ "Name":{"shape":"FacetName"}, - "ObjectType":{"shape":"ObjectType"} + "ObjectType":{"shape":"ObjectType"}, + "FacetStyle":{"shape":"FacetStyle"} } }, "FacetAlreadyExistsException":{ @@ -2664,7 +2766,8 @@ "BINARY", "BOOLEAN", "NUMBER", - "DATETIME" + "DATETIME", + "VARIANT" ] }, "FacetAttributeUpdate":{ @@ -2704,6 +2807,13 @@ "error":{"httpStatusCode":400}, "exception":true }, + "FacetStyle":{ + "type":"string", + "enum":[ + "STATIC", + "DYNAMIC" + ] + }, "FacetValidationException":{ "type":"structure", "members":{ @@ -2764,6 +2874,30 @@ "Facet":{"shape":"Facet"} } }, + "GetLinkAttributesRequest":{ + "type":"structure", + "required":[ + "DirectoryArn", + "TypedLinkSpecifier", + "AttributeNames" + ], + "members":{ + "DirectoryArn":{ + "shape":"Arn", + "location":"header", + "locationName":"x-amz-data-partition" + }, + "TypedLinkSpecifier":{"shape":"TypedLinkSpecifier"}, + "AttributeNames":{"shape":"AttributeNameList"}, + "ConsistencyLevel":{"shape":"ConsistencyLevel"} + } + }, + "GetLinkAttributesResponse":{ + "type":"structure", + "members":{ + "Attributes":{"shape":"AttributeKeyAndValueList"} + } + }, "GetObjectAttributesRequest":{ "type":"structure", "required":[ @@ -2959,6 +3093,24 @@ "error":{"httpStatusCode":400}, "exception":true }, + "LinkAttributeAction":{ + "type":"structure", + "members":{ + "AttributeActionType":{"shape":"UpdateActionType"}, + "AttributeUpdateValue":{"shape":"TypedAttributeValue"} + } + }, + "LinkAttributeUpdate":{ + "type":"structure", + "members":{ + "AttributeKey":{"shape":"AttributeKey"}, + "AttributeAction":{"shape":"LinkAttributeAction"} + } + }, + "LinkAttributeUpdateList":{ + "type":"list", + "member":{"shape":"LinkAttributeUpdate"} + }, "LinkName":{ "type":"string", "max":64, @@ -3155,6 +3307,21 @@ "NextToken":{"shape":"NextToken"} } }, + "ListManagedSchemaArnsRequest":{ + "type":"structure", + "members":{ + "SchemaArn":{"shape":"Arn"}, + "NextToken":{"shape":"NextToken"}, + "MaxResults":{"shape":"NumberResults"} + } + }, + "ListManagedSchemaArnsResponse":{ + "type":"structure", + "members":{ + "SchemaArns":{"shape":"Arns"}, + "NextToken":{"shape":"NextToken"} + } + }, "ListObjectAttributesRequest":{ "type":"structure", "required":[ @@ -3975,6 +4142,28 @@ "members":{ } }, + "UpdateLinkAttributesRequest":{ + "type":"structure", + "required":[ + "DirectoryArn", + "TypedLinkSpecifier", + "AttributeUpdates" + ], + "members":{ + "DirectoryArn":{ + "shape":"Arn", + "location":"header", + "locationName":"x-amz-data-partition" + }, + "TypedLinkSpecifier":{"shape":"TypedLinkSpecifier"}, + "AttributeUpdates":{"shape":"LinkAttributeUpdateList"} + } + }, + "UpdateLinkAttributesResponse":{ + "type":"structure", + "members":{ + } + }, "UpdateObjectAttributesRequest":{ "type":"structure", "required":[ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/docs-2.json similarity index 95% rename from vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/docs-2.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/docs-2.json index a9280cb40..3c60778f7 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/docs-2.json @@ -30,6 +30,7 @@ "GetAppliedSchemaVersion": "

Returns current applied schema version ARN, including the minor version in use.

", "GetDirectory": "

Retrieves metadata about a directory.

", "GetFacet": "

Gets details of the Facet, such as facet name, attributes, Rules, or ObjectType. You can call this on all kinds of schema facets -- published, development, or applied.

", + "GetLinkAttributes": "

Retrieves attributes that are associated with a typed link.

", "GetObjectAttributes": "

Retrieves attributes within a facet that are associated with an object.

", "GetObjectInformation": "

Retrieves metadata about an object.

", "GetSchemaAsJson": "

Retrieves a JSON representation of the schema. See JSON Schema Format for more information.

", @@ -42,6 +43,7 @@ "ListFacetNames": "

Retrieves the names of facets that exist in a schema.

", "ListIncomingTypedLinks": "

Returns a paginated list of all the incoming TypedLinkSpecifier information for an object. It also supports filtering by typed link facet and identity attributes. For more information, see Typed link.

", "ListIndex": "

Lists objects attached to the specified index.

", + "ListManagedSchemaArns": "

Lists the major version families of each managed schema. If a major version ARN is provided as SchemaArn, the minor version revisions in that family are listed instead.

", "ListObjectAttributes": "

Lists all attributes that are associated with an object.

", "ListObjectChildren": "

Returns a paginated list of child objects that are associated with a given object.

", "ListObjectParentPaths": "

Retrieves all available parent paths for any object type such as node, leaf node, policy node, and index node objects. For more information about objects, see Directory Structure.

Use this API to evaluate all parents for an object. The call returns all objects from the root of the directory up to the requested object. The API returns the number of paths based on user-defined MaxResults, in case there are multiple paths to the parent. The order of the paths and nodes returned is consistent among multiple API calls unless the objects are deleted or moved. Paths not leading to the directory root are ignored from the target object.

", @@ -60,6 +62,7 @@ "TagResource": "

An API operation for adding tags to a resource.

", "UntagResource": "

An API operation for removing tags from a resource.

", "UpdateFacet": "

Does the following:

  1. Adds new Attributes, Rules, or ObjectTypes.

  2. Updates existing Attributes, Rules, or ObjectTypes.

  3. Deletes existing Attributes, Rules, or ObjectTypes.

", + "UpdateLinkAttributes": "

Updates a given typed link’s attributes. Attributes to be updated must not contribute to the typed link’s identity, as defined by its IdentityAttributeOrder.

", "UpdateObjectAttributes": "

Updates a given object's attributes.

", "UpdateSchema": "

Updates the schema name with a new name. Only development schema names can be updated.

", "UpdateTypedLinkFacet": "

Updates a TypedLinkFacet. For more information, see Typed link.

", @@ -133,6 +136,7 @@ "GetAppliedSchemaVersionRequest$SchemaArn": "

The ARN of the applied schema.

", "GetAppliedSchemaVersionResponse$AppliedSchemaArn": "

Current applied schema ARN, including the minor version in use if one was provided.

", "GetFacetRequest$SchemaArn": "

The Amazon Resource Name (ARN) that is associated with the Facet. For more information, see arns.

", + "GetLinkAttributesRequest$DirectoryArn": "

The Amazon Resource Name (ARN) that is associated with the Directory where the typed link resides. For more information, see arns or Typed link.

", "GetObjectAttributesRequest$DirectoryArn": "

The Amazon Resource Name (ARN) that is associated with the Directory where the object resides.

", "GetObjectInformationRequest$DirectoryArn": "

The ARN of the directory being retrieved.

", "GetSchemaAsJsonRequest$SchemaArn": "

The ARN of the schema to retrieve.

", @@ -144,6 +148,7 @@ "ListFacetNamesRequest$SchemaArn": "

The Amazon Resource Name (ARN) to retrieve facet names from.

", "ListIncomingTypedLinksRequest$DirectoryArn": "

The Amazon Resource Name (ARN) of the directory where you want to list the typed links.

", "ListIndexRequest$DirectoryArn": "

The ARN of the directory that the index exists in.

", + "ListManagedSchemaArnsRequest$SchemaArn": "

The response for ListManagedSchemaArns. When this parameter is used, all minor version ARNs for a major version are listed.

", "ListObjectAttributesRequest$DirectoryArn": "

The Amazon Resource Name (ARN) that is associated with the Directory where the object resides. For more information, see arns.

", "ListObjectChildrenRequest$DirectoryArn": "

The Amazon Resource Name (ARN) that is associated with the Directory where the object resides. For more information, see arns.

", "ListObjectParentPathsRequest$DirectoryArn": "

The ARN of the directory to which the parent path applies.

", @@ -166,6 +171,7 @@ "TypedLinkSchemaAndFacetName$SchemaArn": "

The Amazon Resource Name (ARN) that is associated with the schema. For more information, see arns.

", "UntagResourceRequest$ResourceArn": "

The Amazon Resource Name (ARN) of the resource. Tagging is only supported for directories.

", "UpdateFacetRequest$SchemaArn": "

The Amazon Resource Name (ARN) that is associated with the Facet. For more information, see arns.

", + "UpdateLinkAttributesRequest$DirectoryArn": "

The Amazon Resource Name (ARN) that is associated with the Directory where the updated typed link resides. For more information, see arns or Typed link.

", "UpdateObjectAttributesRequest$DirectoryArn": "

The Amazon Resource Name (ARN) that is associated with the Directory where the object resides. For more information, see arns.

", "UpdateSchemaRequest$SchemaArn": "

The Amazon Resource Name (ARN) of the development schema. For more information, see arns.

", "UpdateSchemaResponse$SchemaArn": "

The ARN that is associated with the updated schema. For more information, see arns.

", @@ -184,6 +190,7 @@ "refs": { "ListAppliedSchemaArnsResponse$SchemaArns": "

The ARNs of schemas that are applied to the directory.

", "ListDevelopmentSchemaArnsResponse$SchemaArns": "

The ARNs of retrieved development schemas.

", + "ListManagedSchemaArnsResponse$SchemaArns": "

The ARNs for all AWS managed schemas.

", "ListPublishedSchemaArnsResponse$SchemaArns": "

The ARNs of published schemas.

" } }, @@ -232,6 +239,7 @@ "refs": { "AttributeKeyAndValue$Key": "

The key of the attribute.

", "AttributeKeyList$member": null, + "LinkAttributeUpdate$AttributeKey": "

The key of the attribute being updated.

", "ObjectAttributeRange$AttributeKey": "

The key of the attribute that the attribute range covers.

", "ObjectAttributeUpdate$ObjectAttributeKey": "

The key of the attribute being updated.

" } @@ -248,9 +256,11 @@ "AddFacetToObjectRequest$ObjectAttributeList": "

Attributes on the facet that you are adding to the object.

", "BatchAddFacetToObject$ObjectAttributeList": "

The attributes to set on the object.

", "BatchCreateObject$ObjectAttributeList": "

An attribute map, which contains an attribute ARN as the key and attribute value as the map value.

", + "BatchGetLinkAttributesResponse$Attributes": "

The attributes that are associated with the typed link.

", "BatchGetObjectAttributesResponse$Attributes": "

The attribute values that are associated with an object.

", "BatchListObjectAttributesResponse$Attributes": "

The attributes map that is associated with the object. AttributeArn is the key; attribute value is the value.

", "CreateObjectRequest$ObjectAttributeList": "

The attribute map whose attribute ARN contains the key and attribute value as the map value.

", + "GetLinkAttributesResponse$Attributes": "

The attributes that are associated with the typed link.

", "GetObjectAttributesResponse$Attributes": "

The attributes that are associated with the object.

", "IndexAttachment$IndexedAttributes": "

The indexed attribute values.

", "ListObjectAttributesResponse$Attributes": "

Attributes map that is associated with the object. AttributeArn is the key, and attribute value is the value.

" @@ -292,7 +302,9 @@ "AttributeNameList": { "base": null, "refs": { + "BatchGetLinkAttributes$AttributeNames": "

A list of attribute names whose values will be retrieved.

", "BatchGetObjectAttributes$AttributeNames": "

List of attribute names whose values will be retrieved.

", + "GetLinkAttributesRequest$AttributeNames": "

A list of attribute names whose values will be retrieved.

", "GetObjectAttributesRequest$AttributeNames": "

List of attribute names whose values will be retrieved.

", "GetTypedLinkFacetInformationResponse$IdentityAttributeOrder": "

The order of identity attributes for the facet, from most significant to least significant. The ability to filter typed links considers the order that the attributes are defined on the typed link facet. When providing ranges to typed link selection, any inexact ranges must be specified at the end. Any attributes that do not have a range specified are presumed to match the entire range. Filters are interpreted in the order of the attributes on the typed link facet, not the order in which they are supplied to any API calls. For more information about identity attributes, see Typed link.

", "TypedLinkFacet$IdentityAttributeOrder": "

The set of attributes that distinguish links made from this facet from each other, in the order of significance. Listing typed links can filter on the values of these attributes. See ListOutgoingTypedLinks and ListIncomingTypedLinks for details.

", @@ -443,6 +455,18 @@ "BatchWriteOperationResponse$DetachTypedLink": "

Detaches a typed link from a specified source and target object. For more information, see Typed link.

" } }, + "BatchGetLinkAttributes": { + "base": "

Retrieves attributes that are associated with a typed link inside a BatchRead operation. For more information, see GetLinkAttributes and BatchReadRequest$Operations.

", + "refs": { + "BatchReadOperation$GetLinkAttributes": "

Retrieves attributes that are associated with a typed link.

" + } + }, + "BatchGetLinkAttributesResponse": { + "base": "

Represents the output of a GetLinkAttributes response operation.

", + "refs": { + "BatchReadSuccessfulResponse$GetLinkAttributes": "

The list of attributes to retrieve from the typed link.

" + } + }, "BatchGetObjectAttributes": { "base": "

Retrieves attributes within a facet that are associated with an object inside an BatchRead operation. For more information, see GetObjectAttributes and BatchReadRequest$Operations.

", "refs": { @@ -665,6 +689,18 @@ "BatchWriteOperationResponse$RemoveFacetFromObject": "

The result of a batch remove facet from object operation.

" } }, + "BatchUpdateLinkAttributes": { + "base": "

Updates a given typed link’s attributes inside a BatchRead operation. Attributes to be updated must not contribute to the typed link’s identity, as defined by its IdentityAttributeOrder. For more information, see UpdateLinkAttributes and BatchReadRequest$Operations.

", + "refs": { + "BatchWriteOperation$UpdateLinkAttributes": "

Updates a given object's attributes.

" + } + }, + "BatchUpdateLinkAttributesResponse": { + "base": "

Represents the output of a UpdateLinkAttributes response operation.

", + "refs": { + "BatchWriteOperationResponse$UpdateLinkAttributes": "

Represents the output of a BatchWrite response operation.

" + } + }, "BatchUpdateObjectAttributes": { "base": "

Represents the output of a BatchUpdate operation.

", "refs": { @@ -754,6 +790,7 @@ "base": null, "refs": { "BatchReadRequest$ConsistencyLevel": "

Represents the manner and timing in which the successful write or update of an object is reflected in a subsequent read operation of that same object.

", + "GetLinkAttributesRequest$ConsistencyLevel": "

The consistency level at which to retrieve the attributes on a typed link.

", "GetObjectAttributesRequest$ConsistencyLevel": "

The consistency level at which to retrieve the attributes on an object.

", "GetObjectInformationRequest$ConsistencyLevel": "

The consistency level at which to retrieve the object information.

", "ListAttachedIndicesRequest$ConsistencyLevel": "

The consistency level to use for this operation.

", @@ -1128,6 +1165,13 @@ "refs": { } }, + "FacetStyle": { + "base": null, + "refs": { + "CreateFacetRequest$FacetStyle": "

There are two different styles that you can define on any given facet, Static and Dynamic. For static facets, all attributes must be defined in the schema. For dynamic facets, attributes can be defined during data plane operations.

", + "Facet$FacetStyle": "

There are two different styles that you can define on any given facet, Static and Dynamic. For static facets, all attributes must be defined in the schema. For dynamic facets, attributes can be defined during data plane operations.

" + } + }, "FacetValidationException": { "base": "

The Facet that you provided was not well formed or could not be validated with the schema.

", "refs": { @@ -1163,6 +1207,16 @@ "refs": { } }, + "GetLinkAttributesRequest": { + "base": null, + "refs": { + } + }, + "GetLinkAttributesResponse": { + "base": null, + "refs": { + } + }, "GetObjectAttributesRequest": { "base": null, "refs": { @@ -1273,6 +1327,25 @@ "refs": { } }, + "LinkAttributeAction": { + "base": "

The action to take on a typed link attribute value. Updates are only supported for attributes which don’t contribute to link identity.

", + "refs": { + "LinkAttributeUpdate$AttributeAction": "

The action to perform as part of the attribute update.

" + } + }, + "LinkAttributeUpdate": { + "base": "

Structure that contains attribute update information.

", + "refs": { + "LinkAttributeUpdateList$member": null + } + }, + "LinkAttributeUpdateList": { + "base": null, + "refs": { + "BatchUpdateLinkAttributes$AttributeUpdates": "

The attributes update structure.

", + "UpdateLinkAttributesRequest$AttributeUpdates": "

The attributes update structure.

" + } + }, "LinkName": { "base": null, "refs": { @@ -1380,6 +1453,16 @@ "refs": { } }, + "ListManagedSchemaArnsRequest": { + "base": null, + "refs": { + } + }, + "ListManagedSchemaArnsResponse": { + "base": null, + "refs": { + } + }, "ListObjectAttributesRequest": { "base": null, "refs": { @@ -1539,6 +1622,8 @@ "ListIncomingTypedLinksResponse$NextToken": "

The pagination token.

", "ListIndexRequest$NextToken": "

The pagination token.

", "ListIndexResponse$NextToken": "

The pagination token.

", + "ListManagedSchemaArnsRequest$NextToken": "

The pagination token.

", + "ListManagedSchemaArnsResponse$NextToken": "

The pagination token.

", "ListObjectAttributesRequest$NextToken": "

The pagination token.

", "ListObjectAttributesResponse$NextToken": "

The pagination token.

", "ListObjectChildrenRequest$NextToken": "

The pagination token.

", @@ -1607,6 +1692,7 @@ "ListFacetNamesRequest$MaxResults": "

The maximum number of results to retrieve.

", "ListIncomingTypedLinksRequest$MaxResults": "

The maximum number of results to retrieve.

", "ListIndexRequest$MaxResults": "

The maximum number of objects in a single page to retrieve from the index during a request. For more information, see AWS Directory Service Limits.

", + "ListManagedSchemaArnsRequest$MaxResults": "

The maximum number of results to retrieve.

", "ListObjectAttributesRequest$MaxResults": "

The maximum number of items to be retrieved in a single call. This is an approximate number.

", "ListObjectChildrenRequest$MaxResults": "

The maximum number of items to be retrieved in a single call. This is an approximate number.

", "ListObjectParentPathsRequest$MaxResults": "

The maximum number of items to be retrieved in a single call. This is an approximate number.

", @@ -2052,6 +2138,7 @@ "AttributeKeyAndValue$Value": "

The value of the attribute.

", "AttributeNameAndValue$Value": "

The value for the typed link.

", "FacetAttributeDefinition$DefaultValue": "

The default value of the attribute (if configured).

", + "LinkAttributeAction$AttributeUpdateValue": "

The value that you want to update to.

", "ObjectAttributeAction$ObjectAttributeUpdateValue": "

The value that you want to update to.

", "TypedAttributeValueRange$StartValue": "

The value to start the range at.

", "TypedAttributeValueRange$EndValue": "

The attribute value to terminate the range at.

", @@ -2148,8 +2235,12 @@ "AttachTypedLinkResponse$TypedLinkSpecifier": "

Returns a typed link specifier as output.

", "BatchAttachTypedLinkResponse$TypedLinkSpecifier": "

Returns a typed link specifier as output.

", "BatchDetachTypedLink$TypedLinkSpecifier": "

Used to accept a typed link specifier as input.

", + "BatchGetLinkAttributes$TypedLinkSpecifier": "

Allows a typed link specifier to be accepted as input.

", + "BatchUpdateLinkAttributes$TypedLinkSpecifier": "

Allows a typed link specifier to be accepted as input.

", "DetachTypedLinkRequest$TypedLinkSpecifier": "

Used to accept a typed link specifier as input.

", - "TypedLinkSpecifierList$member": null + "GetLinkAttributesRequest$TypedLinkSpecifier": "

Allows a typed link specifier to be accepted as input.

", + "TypedLinkSpecifierList$member": null, + "UpdateLinkAttributesRequest$TypedLinkSpecifier": "

Allows a typed link specifier to be accepted as input.

" } }, "TypedLinkSpecifierList": { @@ -2180,6 +2271,7 @@ "base": null, "refs": { "FacetAttributeUpdate$Action": "

The action to perform when updating the attribute.

", + "LinkAttributeAction$AttributeActionType": "

A type that can be either UPDATE_OR_CREATE or DELETE.

", "ObjectAttributeAction$ObjectAttributeActionType": "

A type that can be either Update or Delete.

", "TypedLinkFacetAttributeUpdate$Action": "

The action to perform when updating the attribute.

" } @@ -2194,6 +2286,16 @@ "refs": { } }, + "UpdateLinkAttributesRequest": { + "base": null, + "refs": { + } + }, + "UpdateLinkAttributesResponse": { + "base": null, + "refs": { + } + }, "UpdateObjectAttributesRequest": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/examples-1.json similarity index 100% rename from vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/examples-1.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/examples-1.json diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/paginators-1.json similarity index 94% rename from vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/paginators-1.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/paginators-1.json index d4164db11..2f01e0bf6 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2016-05-10/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/clouddirectory/2017-01-11/paginators-1.json @@ -35,6 +35,11 @@ "output_token": "NextToken", "limit_key": "MaxResults" }, + "ListManagedSchemaArns": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" + }, "ListObjectAttributes": { "input_token": "NextToken", "output_token": "NextToken", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/api-2.json index 095a76d6b..b4e7efef5 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/api-2.json @@ -567,7 +567,8 @@ {"shape":"OperationInProgressException"}, {"shape":"OperationIdAlreadyExistsException"}, {"shape":"StaleRequestException"}, - {"shape":"InvalidOperationException"} + {"shape":"InvalidOperationException"}, + {"shape":"StackInstanceNotFoundException"} ] }, "UpdateTerminationProtection":{ @@ -889,6 +890,7 @@ "Capabilities":{"shape":"Capabilities"}, "Tags":{"shape":"Tags"}, "AdministrationRoleARN":{"shape":"RoleARN"}, + "ExecutionRoleName":{"shape":"ExecutionRoleName"}, "ClientRequestToken":{ "shape":"ClientRequestToken", "idempotencyToken":true @@ -1169,6 +1171,12 @@ "members":{ } }, + "ExecutionRoleName":{ + "type":"string", + "max":64, + "min":1, + "pattern":"[a-zA-Z_0-9+=,.@-]+" + }, "ExecutionStatus":{ "type":"string", "enum":[ @@ -1966,7 +1974,8 @@ "Capabilities":{"shape":"Capabilities"}, "Tags":{"shape":"Tags"}, "StackSetARN":{"shape":"StackSetARN"}, - "AdministrationRoleARN":{"shape":"RoleARN"} + "AdministrationRoleARN":{"shape":"RoleARN"}, + "ExecutionRoleName":{"shape":"ExecutionRoleName"} } }, "StackSetARN":{"type":"string"}, @@ -2008,6 +2017,7 @@ "OperationPreferences":{"shape":"StackSetOperationPreferences"}, "RetainStacks":{"shape":"RetainStacksNullable"}, "AdministrationRoleARN":{"shape":"RoleARN"}, + "ExecutionRoleName":{"shape":"ExecutionRoleName"}, "CreationTimestamp":{"shape":"Timestamp"}, "EndTimestamp":{"shape":"Timestamp"} } @@ -2292,7 +2302,7 @@ "Regions" ], "members":{ - "StackSetName":{"shape":"StackSetName"}, + "StackSetName":{"shape":"StackSetNameOrId"}, "Accounts":{"shape":"AccountList"}, "Regions":{"shape":"RegionList"}, "ParameterOverrides":{"shape":"Parameters"}, @@ -2329,10 +2339,13 @@ "Tags":{"shape":"Tags"}, "OperationPreferences":{"shape":"StackSetOperationPreferences"}, "AdministrationRoleARN":{"shape":"RoleARN"}, + "ExecutionRoleName":{"shape":"ExecutionRoleName"}, "OperationId":{ "shape":"ClientRequestToken", "idempotencyToken":true - } + }, + "Accounts":{"shape":"AccountList"}, + "Regions":{"shape":"RegionList"} } }, "UpdateStackSetOutput":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/docs-2.json index 8f0567bc3..276dcdb1a 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudformation/2010-05-15/docs-2.json @@ -40,7 +40,7 @@ "StopStackSetOperation": "

Stops an in-progress operation on a stack set and its associated stack instances.

", "UpdateStack": "

Updates a stack as specified in the template. After the call completes successfully, the stack update starts. You can check the status of the stack via the DescribeStacks action.

To get a copy of the template for an existing stack, you can use the GetTemplate action.

For more information about creating an update template, updating a stack, and monitoring the progress of the update, see Updating a Stack.

", "UpdateStackInstances": "

Updates the parameter values for stack instances for the specified accounts, within the specified regions. A stack instance refers to a stack in a specific account and region.

You can only update stack instances in regions and accounts where they already exist; to create additional stack instances, use CreateStackInstances.

During stack set updates, any parameters overridden for a stack instance are not updated, but retain their overridden value.

You can only update the parameter values that are specified in the stack set; to add or delete a parameter itself, use UpdateStackSet to update the stack set template. If you add a parameter to a template, before you can override the parameter value specified in the stack set you must first use UpdateStackSet to update all stack instances with the updated template and parameter value specified in the stack set. Once a stack instance has been updated with the new parameter, you can then override the parameter value using UpdateStackInstances.

", - "UpdateStackSet": "

Updates the stack set and all associated stack instances.

Even if the stack set operation created by updating the stack set fails (completely or partially, below or above a specified failure tolerance), the stack set is updated with your changes. Subsequent CreateStackInstances calls on the specified stack set use the updated stack set.

", + "UpdateStackSet": "

Updates the stack set, and associated stack instances in the specified accounts and regions.

Even if the stack set operation created by updating the stack set fails (completely or partially, below or above a specified failure tolerance), the stack set is updated with your changes. Subsequent CreateStackInstances calls on the specified stack set use the updated stack set.

", "UpdateTerminationProtection": "

Updates termination protection for the specified stack. If a user attempts to delete a stack with termination protection enabled, the operation fails and the stack remains unchanged. For more information, see Protecting a Stack From Being Deleted in the AWS CloudFormation User Guide.

For nested stacks, termination protection is set on the root stack and cannot be changed directly on the nested stack.

", "ValidateTemplate": "

Validates a specified template. AWS CloudFormation first checks if the template is valid JSON. If it isn't, AWS CloudFormation checks if the template is valid YAML. If both these checks fail, AWS CloudFormation returns a template validation error.

" }, @@ -91,7 +91,8 @@ "refs": { "CreateStackInstancesInput$Accounts": "

The names of one or more AWS accounts that you want to create stack instances in the specified region(s) for.

", "DeleteStackInstancesInput$Accounts": "

The names of the AWS accounts that you want to delete stack instances for.

", - "UpdateStackInstancesInput$Accounts": "

The names of one or more AWS accounts for which you want to update parameter values for stack instances. The overridden parameter values will be applied to all stack instances in the specified accounts and regions.

" + "UpdateStackInstancesInput$Accounts": "

The names of one or more AWS accounts for which you want to update parameter values for stack instances. The overridden parameter values will be applied to all stack instances in the specified accounts and regions.

", + "UpdateStackSetInput$Accounts": "

The accounts in which to update associated stack instances. If you specify accounts, you must also specify the regions in which to update stack set instances.

To update all the stack instances associated with this stack set, do not specify the Accounts or Regions properties.

If the stack set update includes changes to the template (that is, if the TemplateBody or TemplateURL properties are specified), or the Parameters property, AWS CloudFormation marks all stack instances with a status of OUTDATED prior to updating the stack instances in the specified accounts and regions. If the stack set update does not include changes to the template or parameters, AWS CloudFormation updates the stack instances in the specified accounts and regions, while leaving all other stack instances with their existing stack instance status.

" } }, "AllowedValue": { @@ -544,6 +545,15 @@ "refs": { } }, + "ExecutionRoleName": { + "base": null, + "refs": { + "CreateStackSetInput$ExecutionRoleName": "

The name of the IAM execution role to use to create the stack set. If you do not specify an execution role, AWS CloudFormation uses the AWSCloudFormationStackSetExecutionRole role for the stack set operation.

Specify an IAM role only if you are using customized execution roles to control which stack resources users and groups can include in their stack sets.

", + "StackSet$ExecutionRoleName": "

The name of the IAM execution role used to create or update the stack set.

Use customized execution roles to control which stack resources users and groups can include in their stack sets.

", + "StackSetOperation$ExecutionRoleName": "

The name of the IAM execution role used to create or update the stack set.

Use customized execution roles to control which stack resources users and groups can include in their stack sets.

", + "UpdateStackSetInput$ExecutionRoleName": "

The name of the IAM execution role to use to update the stack set. If you do not specify an execution role, AWS CloudFormation uses the AWSCloudFormationStackSetExecutionRole role for the stack set operation.

Specify an IAM role only if you are using customized execution roles to control which stack resources users and groups can include in their stack sets.

If you specify a customized execution role, AWS CloudFormation uses that role to update the stack. If you do not specify a customized execution role, AWS CloudFormation performs the update using the role previously associated with the stack set, so long as you have permissions to perform operations on the stack set.

" + } + }, "ExecutionStatus": { "base": null, "refs": { @@ -1012,7 +1022,8 @@ "CreateStackInstancesInput$Regions": "

The names of one or more regions where you want to create stack instances using the specified AWS account(s).

", "DeleteStackInstancesInput$Regions": "

The regions where you want to delete stack set instances.

", "StackSetOperationPreferences$RegionOrder": "

The order of the regions in where you want to perform the stack operation.

", - "UpdateStackInstancesInput$Regions": "

The names of one or more regions in which you want to update parameter values for stack instances. The overridden parameter values will be applied to all stack instances in the specified accounts and regions.

" + "UpdateStackInstancesInput$Regions": "

The names of one or more regions in which you want to update parameter values for stack instances. The overridden parameter values will be applied to all stack instances in the specified accounts and regions.

", + "UpdateStackSetInput$Regions": "

The regions in which to update associated stack instances. If you specify regions, you must also specify accounts in which to update stack set instances.

To update all the stack instances associated with this stack set, do not specify the Accounts or Regions properties.

If the stack set update includes changes to the template (that is, if the TemplateBody or TemplateURL properties are specified), or the Parameters property, AWS CloudFormation marks all stack instances with a status of OUTDATED prior to updating the stack instances in the specified accounts and regions. If the stack set update does not include changes to the template or parameters, AWS CloudFormation updates the stack instances in the specified accounts and regions, while leaving all other stack instances with their existing stack instance status.

" } }, "Replacement": { @@ -1150,10 +1161,10 @@ "ContinueUpdateRollbackInput$RoleARN": "

The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that AWS CloudFormation assumes to roll back the stack. AWS CloudFormation uses the role's credentials to make calls on your behalf. AWS CloudFormation always uses this role for all future operations on the stack. As long as users have permission to operate on the stack, AWS CloudFormation uses this role even if the users don't have permission to pass it. Ensure that the role grants least privilege.

If you don't specify a value, AWS CloudFormation uses the role that was previously associated with the stack. If no role is available, AWS CloudFormation uses a temporary session that is generated from your user credentials.

", "CreateChangeSetInput$RoleARN": "

The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that AWS CloudFormation assumes when executing the change set. AWS CloudFormation uses the role's credentials to make calls on your behalf. AWS CloudFormation uses this role for all future operations on the stack. As long as users have permission to operate on the stack, AWS CloudFormation uses this role even if the users don't have permission to pass it. Ensure that the role grants least privilege.

If you don't specify a value, AWS CloudFormation uses the role that was previously associated with the stack. If no role is available, AWS CloudFormation uses a temporary session that is generated from your user credentials.

", "CreateStackInput$RoleARN": "

The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that AWS CloudFormation assumes to create the stack. AWS CloudFormation uses the role's credentials to make calls on your behalf. AWS CloudFormation always uses this role for all future operations on the stack. As long as users have permission to operate on the stack, AWS CloudFormation uses this role even if the users don't have permission to pass it. Ensure that the role grants least privilege.

If you don't specify a value, AWS CloudFormation uses the role that was previously associated with the stack. If no role is available, AWS CloudFormation uses a temporary session that is generated from your user credentials.

", - "CreateStackSetInput$AdministrationRoleARN": "

The Amazon Resource Number (ARN) of the IAM role to use to create this stack set.

Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Define Permissions for Multiple Administrators in the AWS CloudFormation User Guide.

", + "CreateStackSetInput$AdministrationRoleARN": "

The Amazon Resource Number (ARN) of the IAM role to use to create this stack set.

Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Prerequisites: Granting Permissions for Stack Set Operations in the AWS CloudFormation User Guide.

", "DeleteStackInput$RoleARN": "

The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that AWS CloudFormation assumes to delete the stack. AWS CloudFormation uses the role's credentials to make calls on your behalf.

If you don't specify a value, AWS CloudFormation uses the role that was previously associated with the stack. If no role is available, AWS CloudFormation uses a temporary session that is generated from your user credentials.

", "Stack$RoleARN": "

The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that is associated with the stack. During a stack operation, AWS CloudFormation uses this role's credentials to make calls on your behalf.

", - "StackSet$AdministrationRoleARN": "

The Amazon Resource Number (ARN) of the IAM role used to create or update the stack set.

Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Define Permissions for Multiple Administrators in the AWS CloudFormation User Guide.

", + "StackSet$AdministrationRoleARN": "

The Amazon Resource Number (ARN) of the IAM role used to create or update the stack set.

Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Prerequisites: Granting Permissions for Stack Set Operations in the AWS CloudFormation User Guide.

", "StackSetOperation$AdministrationRoleARN": "

The Amazon Resource Number (ARN) of the IAM role used to perform this stack set operation.

Use customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Define Permissions for Multiple Administrators in the AWS CloudFormation User Guide.

", "UpdateStackInput$RoleARN": "

The Amazon Resource Name (ARN) of an AWS Identity and Access Management (IAM) role that AWS CloudFormation assumes to update the stack. AWS CloudFormation uses the role's credentials to make calls on your behalf. AWS CloudFormation always uses this role for all future operations on the stack. As long as users have permission to operate on the stack, AWS CloudFormation uses this role even if the users don't have permission to pass it. Ensure that the role grants least privilege.

If you don't specify a value, AWS CloudFormation uses the role that was previously associated with the stack. If no role is available, AWS CloudFormation uses a temporary session that is generated from your user credentials.

", "UpdateStackSetInput$AdministrationRoleARN": "

The Amazon Resource Number (ARN) of the IAM role to use to update this stack set.

Specify an IAM role only if you are using customized administrator roles to control which users or groups can manage specific stack sets within the same administrator account. For more information, see Define Permissions for Multiple Administrators in the AWS CloudFormation User Guide.

If you specify a customized administrator role, AWS CloudFormation uses that role to update the stack. If you do not specify a customized administrator role, AWS CloudFormation performs the update using the role previously associated with the stack set, so long as you have permissions to perform operations on the stack set.

" @@ -1405,14 +1416,14 @@ "StackSet$StackSetName": "

The name that's associated with the stack set.

", "StackSetSummary$StackSetName": "

The name of the stack set.

", "StopStackSetOperationInput$StackSetName": "

The name or unique ID of the stack set that you want to stop the operation for.

", - "UpdateStackInstancesInput$StackSetName": "

The name or unique ID of the stack set associated with the stack instances.

", "UpdateStackSetInput$StackSetName": "

The name or unique ID of the stack set that you want to update.

" } }, "StackSetNameOrId": { "base": null, "refs": { - "GetTemplateSummaryInput$StackSetName": "

The name or unique ID of the stack set from which the stack was created.

Conditional: You must specify only one of the following parameters: StackName, StackSetName, TemplateBody, or TemplateURL.

" + "GetTemplateSummaryInput$StackSetName": "

The name or unique ID of the stack set from which the stack was created.

Conditional: You must specify only one of the following parameters: StackName, StackSetName, TemplateBody, or TemplateURL.

", + "UpdateStackInstancesInput$StackSetName": "

The name or unique ID of the stack set associated with the stack instances.

" } }, "StackSetNotEmptyException": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/api-2.json deleted file mode 100644 index 421d0d98b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/api-2.json +++ /dev/null @@ -1,2651 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2015-04-17", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4", - "protocol":"rest-xml" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2015_04_17", - "http":{ - "method":"POST", - "requestUri":"/2015-04-17/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - { - "shape":"CloudFrontOriginAccessIdentityAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCloudFrontOriginAccessIdentities", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2015_04_17", - "http":{ - "method":"POST", - "requestUri":"/2015-04-17/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"DistributionAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidOrigin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidViewerCertificate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidMinimumProtocolVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyDistributions", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidDefaultRootObject", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRelativePath", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidErrorCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidResponseCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRequiredProtocol", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchOrigin", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"TooManyOrigins", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCacheBehaviors", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCookieNamesInWhiteList", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidForwardCookies", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyHeadersInForwardedValues", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidHeadersForS3Origin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCertificates", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidLocationCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidGeoRestrictionParameter", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidProtocolSettings", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidTTLOrder", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2015_04_17", - "http":{ - "method":"POST", - "requestUri":"/2015-04-17/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"BatchTooLarge", - "error":{"httpStatusCode":413}, - "exception":true - }, - { - "shape":"TooManyInvalidationsInProgress", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2015_04_17", - "http":{ - "method":"POST", - "requestUri":"/2015-04-17/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"StreamingDistributionAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidOrigin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyStreamingDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyStreamingDistributions", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2015_04_17", - "http":{ - "method":"DELETE", - "requestUri":"/2015-04-17/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"CloudFrontOriginAccessIdentityInUse", - "error":{"httpStatusCode":409}, - "exception":true - } - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2015_04_17", - "http":{ - "method":"DELETE", - "requestUri":"/2015-04-17/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"DistributionNotDisabled", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - } - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2015_04_17", - "http":{ - "method":"DELETE", - "requestUri":"/2015-04-17/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"StreamingDistributionNotDisabled", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - } - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetDistribution":{ - "name":"GetDistribution2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - { - "shape":"NoSuchInvalidation", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListDistributions":{ - "name":"ListDistributions2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2015_04_17", - "http":{ - "method":"GET", - "requestUri":"/2015-04-17/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2015_04_17", - "http":{ - "method":"PUT", - "requestUri":"/2015-04-17/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"IllegalUpdate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2015_04_17", - "http":{ - "method":"PUT", - "requestUri":"/2015-04-17/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"IllegalUpdate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"TooManyDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidDefaultRootObject", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRelativePath", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidErrorCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidResponseCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidViewerCertificate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidMinimumProtocolVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRequiredProtocol", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchOrigin", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"TooManyOrigins", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCacheBehaviors", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCookieNamesInWhiteList", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidForwardCookies", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyHeadersInForwardedValues", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidHeadersForS3Origin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCertificates", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidLocationCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidGeoRestrictionParameter", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidTTLOrder", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2015_04_17", - "http":{ - "method":"PUT", - "requestUri":"/2015-04-17/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"IllegalUpdate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"TooManyStreamingDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"CloudFrontOriginAccessIdentityConfig" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"DistributionConfig" - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"InvalidationBatch" - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"StreamingDistributionConfig" - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "required":["Id"] - }, - "DeleteDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "required":["Id"] - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "required":["Id"] - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer" - ] - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"CloudFrontOriginAccessIdentityConfig" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"DistributionConfig" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-04-17/"}, - "locationName":"StreamingDistributionConfig" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "IAMCertificateId":{"shape":"string"}, - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"} - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/docs-2.json deleted file mode 100644 index bd2eb39f1..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/docs-2.json +++ /dev/null @@ -1,1141 +0,0 @@ -{ - "version": "2.0", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "service": null, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidViewerCertificate$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions you want in the response body.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOrigins$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$IAMCertificateId": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value or CloudFrontDefaultCertificate." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/waiters-2.json deleted file mode 100644 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-04-17/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/api-2.json deleted file mode 100644 index 5da9d56e9..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/api-2.json +++ /dev/null @@ -1,2721 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2015-07-27", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4", - "protocol":"rest-xml" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2015_07_27", - "http":{ - "method":"POST", - "requestUri":"/2015-07-27/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - { - "shape":"CloudFrontOriginAccessIdentityAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCloudFrontOriginAccessIdentities", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2015_07_27", - "http":{ - "method":"POST", - "requestUri":"/2015-07-27/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"DistributionAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidOrigin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidViewerCertificate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidMinimumProtocolVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyDistributions", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidDefaultRootObject", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRelativePath", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidErrorCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidResponseCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRequiredProtocol", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchOrigin", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"TooManyOrigins", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCacheBehaviors", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCookieNamesInWhiteList", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidForwardCookies", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyHeadersInForwardedValues", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidHeadersForS3Origin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCertificates", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidLocationCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidGeoRestrictionParameter", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidProtocolSettings", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidTTLOrder", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidWebACLId", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2015_07_27", - "http":{ - "method":"POST", - "requestUri":"/2015-07-27/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"BatchTooLarge", - "error":{"httpStatusCode":413}, - "exception":true - }, - { - "shape":"TooManyInvalidationsInProgress", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2015_07_27", - "http":{ - "method":"POST", - "requestUri":"/2015-07-27/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"StreamingDistributionAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidOrigin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyStreamingDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyStreamingDistributions", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2015_07_27", - "http":{ - "method":"DELETE", - "requestUri":"/2015-07-27/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"CloudFrontOriginAccessIdentityInUse", - "error":{"httpStatusCode":409}, - "exception":true - } - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2015_07_27", - "http":{ - "method":"DELETE", - "requestUri":"/2015-07-27/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"DistributionNotDisabled", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - } - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2015_07_27", - "http":{ - "method":"DELETE", - "requestUri":"/2015-07-27/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"StreamingDistributionNotDisabled", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - } - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetDistribution":{ - "name":"GetDistribution2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - { - "shape":"NoSuchInvalidation", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListDistributions":{ - "name":"ListDistributions2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidWebACLId", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - } - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2015_07_27", - "http":{ - "method":"GET", - "requestUri":"/2015-07-27/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2015_07_27", - "http":{ - "method":"PUT", - "requestUri":"/2015-07-27/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"IllegalUpdate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchCloudFrontOriginAccessIdentity", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2015_07_27", - "http":{ - "method":"PUT", - "requestUri":"/2015-07-27/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"IllegalUpdate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"TooManyDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidDefaultRootObject", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRelativePath", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidErrorCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidResponseCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidViewerCertificate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidMinimumProtocolVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidRequiredProtocol", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchOrigin", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"TooManyOrigins", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCacheBehaviors", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCookieNamesInWhiteList", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidForwardCookies", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyHeadersInForwardedValues", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidHeadersForS3Origin", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyCertificates", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidLocationCode", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidGeoRestrictionParameter", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidTTLOrder", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidWebACLId", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2015_07_27", - "http":{ - "method":"PUT", - "requestUri":"/2015-07-27/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - { - "shape":"AccessDenied", - "error":{"httpStatusCode":403}, - "exception":true - }, - { - "shape":"CNAMEAlreadyExists", - "error":{"httpStatusCode":409}, - "exception":true - }, - { - "shape":"IllegalUpdate", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidIfMatchVersion", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"MissingBody", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"NoSuchStreamingDistribution", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"PreconditionFailed", - "error":{"httpStatusCode":412}, - "exception":true - }, - { - "shape":"TooManyStreamingDistributionCNAMEs", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidArgument", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InvalidOriginAccessIdentity", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TooManyTrustedSigners", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"TrustedSignerDoesNotExist", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"InconsistentQuantities", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"CloudFrontOriginAccessIdentityConfig" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"DistributionConfig" - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"InvalidationBatch" - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"StreamingDistributionConfig" - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "required":["Id"] - }, - "DeleteDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "required":["Id"] - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "required":["Id"] - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - }, - "required":["Id"] - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer" - ] - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"CloudFrontOriginAccessIdentityConfig" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"DistributionConfig" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-07-27/"}, - "locationName":"StreamingDistributionConfig" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "IAMCertificateId":{"shape":"string"}, - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"} - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/docs-2.json deleted file mode 100644 index 07747194c..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/docs-2.json +++ /dev/null @@ -1,1164 +0,0 @@ -{ - "version": "2.0", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "service": null, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOrigins$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$IAMCertificateId": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value or CloudFrontDefaultCertificate." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/waiters-2.json deleted file mode 100644 index f6d3ba7bc..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-07-27/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/api-2.json deleted file mode 100644 index 374b7e5e9..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/api-2.json +++ /dev/null @@ -1,2150 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2015-09-17", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2015_09_17", - "http":{ - "method":"POST", - "requestUri":"/2015-09-17/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2015_09_17", - "http":{ - "method":"POST", - "requestUri":"/2015-09-17/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2015_09_17", - "http":{ - "method":"POST", - "requestUri":"/2015-09-17/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2015_09_17", - "http":{ - "method":"POST", - "requestUri":"/2015-09-17/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2015_09_17", - "http":{ - "method":"DELETE", - "requestUri":"/2015-09-17/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2015_09_17", - "http":{ - "method":"DELETE", - "requestUri":"/2015-09-17/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2015_09_17", - "http":{ - "method":"DELETE", - "requestUri":"/2015-09-17/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2015_09_17", - "http":{ - "method":"GET", - "requestUri":"/2015-09-17/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2015_09_17", - "http":{ - "method":"PUT", - "requestUri":"/2015-09-17/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2015_09_17", - "http":{ - "method":"PUT", - "requestUri":"/2015-09-17/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2015_09_17", - "http":{ - "method":"PUT", - "requestUri":"/2015-09-17/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer" - ] - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2015-09-17/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "Certificate":{"shape":"string"}, - "CertificateSource":{"shape":"CertificateSource"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "IAMCertificateId":{ - "shape":"string", - "deprecated":true - }, - "CloudFrontDefaultCertificate":{ - "shape":"boolean", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/docs-2.json deleted file mode 100644 index 3e23d960f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/docs-2.json +++ /dev/null @@ -1,1173 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to \"cloudfront\". If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), set to \"iam\", and update the Certificate field with the IAM certificate identifier of the custom viewer certificate for this distribution." - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DefaultCacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "Note: this field is deprecated. Please use \"cloudfront\" as CertificateSource and omit specifying a Certificate. If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOrigins$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$Certificate": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), set to the IAM certificate identifier of the custom viewer certificate for this distribution.", - "ViewerCertificate$IAMCertificateId": "Note: this field is deprecated. Please use \"iam\" as CertificateSource and specify the IAM certificate Id as the Certificate. If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value or CloudFrontDefaultCertificate." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/waiters-2.json deleted file mode 100644 index f6d3ba7bc..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/api-2.json deleted file mode 100644 index 8fbe7298b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/api-2.json +++ /dev/null @@ -1,2216 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2016-01-13", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_01_13", - "http":{ - "method":"POST", - "requestUri":"/2016-01-13/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_01_13", - "http":{ - "method":"POST", - "requestUri":"/2016-01-13/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_01_13", - "http":{ - "method":"POST", - "requestUri":"/2016-01-13/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_01_13", - "http":{ - "method":"POST", - "requestUri":"/2016-01-13/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_01_13", - "http":{ - "method":"DELETE", - "requestUri":"/2016-01-13/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_01_13", - "http":{ - "method":"DELETE", - "requestUri":"/2016-01-13/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_01_13", - "http":{ - "method":"DELETE", - "requestUri":"/2016-01-13/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_01_13", - "http":{ - "method":"GET", - "requestUri":"/2016-01-13/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_01_13", - "http":{ - "method":"PUT", - "requestUri":"/2016-01-13/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_01_13", - "http":{ - "method":"PUT", - "requestUri":"/2016-01-13/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_01_13", - "http":{ - "method":"PUT", - "requestUri":"/2016-01-13/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-13/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "Certificate":{"shape":"string"}, - "CertificateSource":{"shape":"CertificateSource"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "IAMCertificateId":{ - "shape":"string", - "deprecated":true - }, - "CloudFrontDefaultCertificate":{ - "shape":"boolean", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/docs-2.json deleted file mode 100644 index a9e293427..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/docs-2.json +++ /dev/null @@ -1,1219 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to \"cloudfront\". If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), you can use your own IAM or ACM certificate. To use an ACM certificate, set to \"acm\" and update the Certificate to the ACM certificate ARN. To use an IAM certificate, set to \"iam\" and update the Certificate to the IAM certificate identifier." - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomHeaders": { - "base": "A complex type that contains the list of Custom Headers for each origin.", - "refs": { - "Origin$CustomHeaders": "A complex type that contains information about the custom headers associated with this Origin." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "A complex type that contains information related to a Header", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "A complex type that contains the custom headers for this Origin." - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "OriginSslProtocols": { - "base": "A complex type that contains the list of SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "A complex type that contains one SslProtocol element for each SSL/TLS protocol that you want to allow CloudFront to use when establishing an HTTPS connection with this origin." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DefaultCacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "Note: this field is deprecated. Please use \"cloudfront\" as CertificateSource and omit specifying a Certificate. If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomHeaders$Quantity": "The number of custom headers for this origin.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "OriginSslProtocols$Quantity": "The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "OriginCustomHeader$HeaderName": "The header's name.", - "OriginCustomHeader$HeaderValue": "The header's value.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$Certificate": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), you can use your own IAM or ACM certificate. For ACM, set to the ACM certificate ARN. For IAM, set to the IAM certificate identifier.", - "ViewerCertificate$IAMCertificateId": "Note: this field is deprecated. Please use \"iam\" as CertificateSource and specify the IAM certificate Id as the Certificate. If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value or CloudFrontDefaultCertificate." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/waiters-2.json deleted file mode 100644 index f6d3ba7bc..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-13/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/api-2.json deleted file mode 100644 index 549e4efd8..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/api-2.json +++ /dev/null @@ -1,2219 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"cloudfront-2016-01-28", - "apiVersion":"2016-01-28", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_01_28", - "http":{ - "method":"POST", - "requestUri":"/2016-01-28/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_01_28", - "http":{ - "method":"POST", - "requestUri":"/2016-01-28/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_01_28", - "http":{ - "method":"POST", - "requestUri":"/2016-01-28/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_01_28", - "http":{ - "method":"POST", - "requestUri":"/2016-01-28/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_01_28", - "http":{ - "method":"DELETE", - "requestUri":"/2016-01-28/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_01_28", - "http":{ - "method":"DELETE", - "requestUri":"/2016-01-28/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_01_28", - "http":{ - "method":"DELETE", - "requestUri":"/2016-01-28/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_01_28", - "http":{ - "method":"GET", - "requestUri":"/2016-01-28/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_01_28", - "http":{ - "method":"PUT", - "requestUri":"/2016-01-28/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_01_28", - "http":{ - "method":"PUT", - "requestUri":"/2016-01-28/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_01_28", - "http":{ - "method":"PUT", - "requestUri":"/2016-01-28/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-01-28/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/docs-2.json deleted file mode 100644 index 8174625b4..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/docs-2.json +++ /dev/null @@ -1,1220 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomHeaders": { - "base": "A complex type that contains the list of Custom Headers for each origin.", - "refs": { - "Origin$CustomHeaders": "A complex type that contains information about the custom headers associated with this Origin." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "A complex type that contains information related to a Header", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "A complex type that contains the custom headers for this Origin." - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "OriginSslProtocols": { - "base": "A complex type that contains the list of SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "A complex type that contains one SslProtocol element for each SSL/TLS protocol that you want to allow CloudFront to use when establishing an HTTPS connection with this origin." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DefaultCacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an ACMCertificateArn or IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomHeaders$Quantity": "The number of custom headers for this origin.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "OriginSslProtocols$Quantity": "The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "OriginCustomHeader$HeaderName": "The header's name.", - "OriginCustomHeader$HeaderValue": "The header's value.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$IAMCertificateId": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value, ACMCertificateArn, or CloudFrontDefaultCertificate.", - "ViewerCertificate$ACMCertificateArn": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the ACM certificate ARN of the custom viewer certificate for this distribution. Specify either this value, IAMCertificateId, or CloudFrontDefaultCertificate.", - "ViewerCertificate$Certificate": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/waiters-2.json deleted file mode 100644 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-01-28/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/api-2.json deleted file mode 100644 index d9d838f55..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/api-2.json +++ /dev/null @@ -1,2548 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"cloudfront-2016-08-01", - "apiVersion":"2016-08-01", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"} - ] - }, - "CreateDistributionWithTags":{ - "name":"CreateDistributionWithTags2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionWithTagsRequest"}, - "output":{"shape":"CreateDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"InvalidTagging"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistributionWithTags":{ - "name":"CreateStreamingDistributionWithTags2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/streaming-distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionWithTagsRequest"}, - "output":{"shape":"CreateStreamingDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"}, - {"shape":"InvalidTagging"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_08_01", - "http":{ - "method":"DELETE", - "requestUri":"/2016-08-01/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_08_01", - "http":{ - "method":"DELETE", - "requestUri":"/2016-08-01/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_08_01", - "http":{ - "method":"DELETE", - "requestUri":"/2016-08-01/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource2016_08_01", - "http":{ - "method":"GET", - "requestUri":"/2016-08-01/tagging" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "TagResource":{ - "name":"TagResource2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/tagging?Operation=Tag", - "responseCode":204 - }, - "input":{"shape":"TagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UntagResource":{ - "name":"UntagResource2016_08_01", - "http":{ - "method":"POST", - "requestUri":"/2016-08-01/tagging?Operation=Untag", - "responseCode":204 - }, - "input":{"shape":"UntagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_08_01", - "http":{ - "method":"PUT", - "requestUri":"/2016-08-01/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_08_01", - "http":{ - "method":"PUT", - "requestUri":"/2016-08-01/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_08_01", - "http":{ - "method":"PUT", - "requestUri":"/2016-08-01/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateDistributionWithTagsRequest":{ - "type":"structure", - "required":["DistributionConfigWithTags"], - "members":{ - "DistributionConfigWithTags":{ - "shape":"DistributionConfigWithTags", - "locationName":"DistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"DistributionConfigWithTags" - }, - "CreateDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CreateStreamingDistributionWithTagsRequest":{ - "type":"structure", - "required":["StreamingDistributionConfigWithTags"], - "members":{ - "StreamingDistributionConfigWithTags":{ - "shape":"StreamingDistributionConfigWithTags", - "locationName":"StreamingDistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"StreamingDistributionConfigWithTags" - }, - "CreateStreamingDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionConfigWithTags":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Tags" - ], - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTagging":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["Resource"], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - } - } - }, - "ListTagsForResourceResult":{ - "type":"structure", - "required":["Tags"], - "members":{ - "Tags":{"shape":"Tags"} - }, - "payload":"Tags" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchResource":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "ResourceARN":{ - "type":"string", - "pattern":"arn:aws:cloudfront::[0-9]+:.*" - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionConfigWithTags":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Tags" - ], - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Tag":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{ - "type":"string", - "max":128, - "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "TagKeyList":{ - "type":"list", - "member":{ - "shape":"TagKey", - "locationName":"Key" - } - }, - "TagKeys":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagKeyList"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "Tags" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "Tags":{ - "shape":"Tags", - "locationName":"Tags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"Tags" - }, - "TagValue":{ - "type":"string", - "max":256, - "min":0, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "Tags":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagList"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UntagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "TagKeys" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "TagKeys":{ - "shape":"TagKeys", - "locationName":"TagKeys", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - } - }, - "payload":"TagKeys" - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-01/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/docs-2.json deleted file mode 100644 index 68fcff76a..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/docs-2.json +++ /dev/null @@ -1,1355 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon CloudFront

Amazon CloudFront is a global content delivery network (CDN) service that accelerates delivery of your websites, APIs, video content or other web assets. It integrates with other Amazon Web Services products to give developers and businesses an easy way to accelerate content to end users with no minimum usage commitments.

", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateDistributionWithTags": "Create a new distribution with tags.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "CreateStreamingDistributionWithTags": "Create a new streaming distribution with tags.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "ListTagsForResource": "List tags for a CloudFront resource.", - "TagResource": "Add tags to a CloudFront resource.", - "UntagResource": "Remove tags from a CloudFront resource.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionWithTagsRequest": { - "base": "The request to create a new distribution with tags", - "refs": { - } - }, - "CreateDistributionWithTagsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsRequest": { - "base": "The request to create a new streaming distribution with tags.", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomHeaders": { - "base": "A complex type that contains the list of Custom Headers for each origin.", - "refs": { - "Origin$CustomHeaders": "A complex type that contains information about the custom headers associated with this Origin." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "CreateDistributionWithTagsResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "DistributionConfigWithTags$DistributionConfig": "A distribution Configuration.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionConfigWithTags": { - "base": "A distribution Configuration and a list of tags to be associated with the distribution.", - "refs": { - "CreateDistributionWithTagsRequest$DistributionConfigWithTags": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidTagging": { - "base": "The specified tagging for a CloudFront resource is invalid. For more information, see the error text.", - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": "The request to list tags for a CloudFront resource.", - "refs": { - } - }, - "ListTagsForResourceResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchResource": { - "base": "The specified CloudFront resource does not exist.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "A complex type that contains information related to a Header", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "A complex type that contains the custom headers for this Origin." - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "OriginSslProtocols": { - "base": "A complex type that contains the list of SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "ResourceARN": { - "base": null, - "refs": { - "ListTagsForResourceRequest$Resource": "An ARN of a CloudFront resource.", - "TagResourceRequest$Resource": "An ARN of a CloudFront resource.", - "UntagResourceRequest$Resource": "An ARN of a CloudFront resource." - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "A complex type that contains one SslProtocol element for each SSL/TLS protocol that you want to allow CloudFront to use when establishing an HTTPS connection with this origin." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "CreateStreamingDistributionWithTagsResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "StreamingDistributionConfigWithTags$StreamingDistributionConfig": "A streaming distribution Configuration.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionConfigWithTags": { - "base": "A streaming distribution Configuration and a list of tags to be associated with the streaming distribution.", - "refs": { - "CreateStreamingDistributionWithTagsRequest$StreamingDistributionConfigWithTags": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "Tag": { - "base": "A complex type that contains Tag key and Tag value.", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": "A string that contains Tag key. The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.", - "refs": { - "Tag$Key": "A string that contains Tag key. The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.", - "TagKeyList$member": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "TagKeys$Items": "A complex type that contains Tag key elements" - } - }, - "TagKeys": { - "base": "A complex type that contains zero or more Tag elements.", - "refs": { - "UntagResourceRequest$TagKeys": "A complex type that contains zero or more Tag key elements." - } - }, - "TagList": { - "base": null, - "refs": { - "Tags$Items": "A complex type that contains Tag elements" - } - }, - "TagResourceRequest": { - "base": "The request to add tags to a CloudFront resource.", - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "A string that contains an optional Tag value. The string length should be between 0 and 256 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @." - } - }, - "Tags": { - "base": "A complex type that contains zero or more Tag elements.", - "refs": { - "DistributionConfigWithTags$Tags": "A complex type that contains zero or more Tag elements.", - "ListTagsForResourceResult$Tags": "A complex type that contains zero or more Tag elements.", - "StreamingDistributionConfigWithTags$Tags": "A complex type that contains zero or more Tag elements.", - "TagResourceRequest$Tags": "A complex type that contains zero or more Tag elements." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UntagResourceRequest": { - "base": "The request to remove tags from a CloudFront resource.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DefaultCacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an ACMCertificateArn or IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomHeaders$Quantity": "The number of custom headers for this origin.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "OriginSslProtocols$Quantity": "The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateDistributionWithTagsResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionWithTagsResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CreateStreamingDistributionWithTagsResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionWithTagsResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$ARN": "The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$ARN": "The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidTagging$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchResource$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "OriginCustomHeader$HeaderName": "The header's name.", - "OriginCustomHeader$HeaderValue": "The header's value.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$ARN": "The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$ARN": "The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$IAMCertificateId": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value, ACMCertificateArn, or CloudFrontDefaultCertificate.", - "ViewerCertificate$ACMCertificateArn": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the ACM certificate ARN of the custom viewer certificate for this distribution. Specify either this value, IAMCertificateId, or CloudFrontDefaultCertificate.", - "ViewerCertificate$Certificate": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/waiters-2.json deleted file mode 100644 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-01/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/api-2.json deleted file mode 100755 index a8f3caf4c..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/api-2.json +++ /dev/null @@ -1,2586 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"cloudfront-2016-08-20", - "apiVersion":"2016-08-20", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "CreateDistributionWithTags":{ - "name":"CreateDistributionWithTags2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionWithTagsRequest"}, - "output":{"shape":"CreateDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"InvalidTagging"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistributionWithTags":{ - "name":"CreateStreamingDistributionWithTags2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/streaming-distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionWithTagsRequest"}, - "output":{"shape":"CreateStreamingDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"}, - {"shape":"InvalidTagging"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_08_20", - "http":{ - "method":"DELETE", - "requestUri":"/2016-08-20/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_08_20", - "http":{ - "method":"DELETE", - "requestUri":"/2016-08-20/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_08_20", - "http":{ - "method":"DELETE", - "requestUri":"/2016-08-20/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource2016_08_20", - "http":{ - "method":"GET", - "requestUri":"/2016-08-20/tagging" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "TagResource":{ - "name":"TagResource2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/tagging?Operation=Tag", - "responseCode":204 - }, - "input":{"shape":"TagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UntagResource":{ - "name":"UntagResource2016_08_20", - "http":{ - "method":"POST", - "requestUri":"/2016-08-20/tagging?Operation=Untag", - "responseCode":204 - }, - "input":{"shape":"UntagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_08_20", - "http":{ - "method":"PUT", - "requestUri":"/2016-08-20/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_08_20", - "http":{ - "method":"PUT", - "requestUri":"/2016-08-20/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_08_20", - "http":{ - "method":"PUT", - "requestUri":"/2016-08-20/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateDistributionWithTagsRequest":{ - "type":"structure", - "required":["DistributionConfigWithTags"], - "members":{ - "DistributionConfigWithTags":{ - "shape":"DistributionConfigWithTags", - "locationName":"DistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"DistributionConfigWithTags" - }, - "CreateDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CreateStreamingDistributionWithTagsRequest":{ - "type":"structure", - "required":["StreamingDistributionConfigWithTags"], - "members":{ - "StreamingDistributionConfigWithTags":{ - "shape":"StreamingDistributionConfigWithTags", - "locationName":"StreamingDistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"StreamingDistributionConfigWithTags" - }, - "CreateStreamingDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionConfigWithTags":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Tags" - ], - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"}, - "QueryStringCacheKeys":{"shape":"QueryStringCacheKeys"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTagging":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["Resource"], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - } - } - }, - "ListTagsForResourceResult":{ - "type":"structure", - "required":["Tags"], - "members":{ - "Tags":{"shape":"Tags"} - }, - "payload":"Tags" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchResource":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "QueryStringCacheKeys":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"QueryStringCacheKeysList"} - } - }, - "QueryStringCacheKeysList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "ResourceARN":{ - "type":"string", - "pattern":"arn:aws:cloudfront::[0-9]+:.*" - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionConfigWithTags":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Tags" - ], - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Tag":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{ - "type":"string", - "max":128, - "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "TagKeyList":{ - "type":"list", - "member":{ - "shape":"TagKey", - "locationName":"Key" - } - }, - "TagKeys":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagKeyList"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "Tags" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "Tags":{ - "shape":"Tags", - "locationName":"Tags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"Tags" - }, - "TagValue":{ - "type":"string", - "max":256, - "min":0, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "Tags":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagList"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UntagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "TagKeys" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "TagKeys":{ - "shape":"TagKeys", - "locationName":"TagKeys", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - } - }, - "payload":"TagKeys" - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-08-20/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/docs-2.json deleted file mode 100755 index 0497df084..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/docs-2.json +++ /dev/null @@ -1,1381 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon CloudFront

Amazon CloudFront is a global content delivery network (CDN) service that accelerates delivery of your websites, APIs, video content or other web assets. It integrates with other Amazon Web Services products to give developers and businesses an easy way to accelerate content to end users with no minimum usage commitments.

", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateDistributionWithTags": "Create a new distribution with tags.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "CreateStreamingDistributionWithTags": "Create a new streaming distribution with tags.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "ListTagsForResource": "List tags for a CloudFront resource.", - "TagResource": "Add tags to a CloudFront resource.", - "UntagResource": "Remove tags from a CloudFront resource.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionWithTagsRequest": { - "base": "The request to create a new distribution with tags", - "refs": { - } - }, - "CreateDistributionWithTagsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsRequest": { - "base": "The request to create a new streaming distribution with tags.", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomHeaders": { - "base": "A complex type that contains the list of Custom Headers for each origin.", - "refs": { - "Origin$CustomHeaders": "A complex type that contains information about the custom headers associated with this Origin." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "CreateDistributionWithTagsResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "DistributionConfigWithTags$DistributionConfig": "A distribution Configuration.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionConfigWithTags": { - "base": "A distribution Configuration and a list of tags to be associated with the distribution.", - "refs": { - "CreateDistributionWithTagsRequest$DistributionConfigWithTags": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidQueryStringParameters": { - "base": null, - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidTagging": { - "base": "The specified tagging for a CloudFront resource is invalid. For more information, see the error text.", - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": "The request to list tags for a CloudFront resource.", - "refs": { - } - }, - "ListTagsForResourceResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchResource": { - "base": "The specified CloudFront resource does not exist.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "A complex type that contains information related to a Header", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "A complex type that contains the custom headers for this Origin." - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "OriginSslProtocols": { - "base": "A complex type that contains the list of SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "QueryStringCacheKeys": { - "base": null, - "refs": { - "ForwardedValues$QueryStringCacheKeys": "

A complex type that contains information about the query string parameters that you want CloudFront to use for caching for this cache behavior.

" - } - }, - "QueryStringCacheKeysList": { - "base": null, - "refs": { - "QueryStringCacheKeys$Items": "Optional: A list that contains the query string parameters that you want CloudFront to use as a basis for caching for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "ResourceARN": { - "base": null, - "refs": { - "ListTagsForResourceRequest$Resource": "An ARN of a CloudFront resource.", - "TagResourceRequest$Resource": "An ARN of a CloudFront resource.", - "UntagResourceRequest$Resource": "An ARN of a CloudFront resource." - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "A complex type that contains one SslProtocol element for each SSL/TLS protocol that you want to allow CloudFront to use when establishing an HTTPS connection with this origin." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "CreateStreamingDistributionWithTagsResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "StreamingDistributionConfigWithTags$StreamingDistributionConfig": "A streaming distribution Configuration.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionConfigWithTags": { - "base": "A streaming distribution Configuration and a list of tags to be associated with the streaming distribution.", - "refs": { - "CreateStreamingDistributionWithTagsRequest$StreamingDistributionConfigWithTags": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "Tag": { - "base": "A complex type that contains Tag key and Tag value.", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": "A string that contains Tag key. The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.", - "refs": { - "Tag$Key": "A string that contains Tag key. The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.", - "TagKeyList$member": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "TagKeys$Items": "A complex type that contains Tag key elements" - } - }, - "TagKeys": { - "base": "A complex type that contains zero or more Tag elements.", - "refs": { - "UntagResourceRequest$TagKeys": "A complex type that contains zero or more Tag key elements." - } - }, - "TagList": { - "base": null, - "refs": { - "Tags$Items": "A complex type that contains Tag elements" - } - }, - "TagResourceRequest": { - "base": "The request to add tags to a CloudFront resource.", - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "A string that contains an optional Tag value. The string length should be between 0 and 256 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @." - } - }, - "Tags": { - "base": "A complex type that contains zero or more Tag elements.", - "refs": { - "DistributionConfigWithTags$Tags": "A complex type that contains zero or more Tag elements.", - "ListTagsForResourceResult$Tags": "A complex type that contains zero or more Tag elements.", - "StreamingDistributionConfigWithTags$Tags": "A complex type that contains zero or more Tag elements.", - "TagResourceRequest$Tags": "A complex type that contains zero or more Tag elements." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyQueryStringParameters": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UntagResourceRequest": { - "base": "The request to remove tags from a CloudFront resource.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DefaultCacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "

Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior and cache based on the query string parameters. CloudFront behavior depends on the value of QueryString and on the values that you specify for QueryStringCacheKeys, if any:

  • If you specify true for QueryString and you don't specify any values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin and caches based on all query string parameters. Depending on how many query string parameters and values you have, this can adversely affect performance because CloudFront must forward more requests to the origin.
  • If you specify true for QueryString and you specify one or more values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin, but it only caches based on the query string parameters that you specify.
  • If you specify false for QueryString, CloudFront doesn't forward any query string parameters to the origin, and doesn't cache based on query string parameters.
", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an ACMCertificateArn or IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomHeaders$Quantity": "The number of custom headers for this origin.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "OriginSslProtocols$Quantity": "The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "QueryStringCacheKeys$Quantity": "The number of whitelisted query string parameters for this cache behavior.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateDistributionWithTagsResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionWithTagsResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CreateStreamingDistributionWithTagsResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionWithTagsResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$ARN": "The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$ARN": "The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidQueryStringParameters$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidTagging$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchResource$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "OriginCustomHeader$HeaderName": "The header's name.", - "OriginCustomHeader$HeaderValue": "The header's value.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "QueryStringCacheKeysList$member": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$ARN": "The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$ARN": "The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyQueryStringParameters$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$IAMCertificateId": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value, ACMCertificateArn, or CloudFrontDefaultCertificate.", - "ViewerCertificate$ACMCertificateArn": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the ACM certificate ARN of the custom viewer certificate for this distribution. Specify either this value, IAMCertificateId, or CloudFrontDefaultCertificate.", - "ViewerCertificate$Certificate": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/paginators-1.json deleted file mode 100755 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/waiters-2.json deleted file mode 100755 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-08-20/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/api-2.json deleted file mode 100755 index ec9b60af4..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/api-2.json +++ /dev/null @@ -1,2596 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"cloudfront-2016-09-07", - "apiVersion":"2016-09-07", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "CreateDistributionWithTags":{ - "name":"CreateDistributionWithTags2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionWithTagsRequest"}, - "output":{"shape":"CreateDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"InvalidTagging"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistributionWithTags":{ - "name":"CreateStreamingDistributionWithTags2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/streaming-distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionWithTagsRequest"}, - "output":{"shape":"CreateStreamingDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"}, - {"shape":"InvalidTagging"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_09_07", - "http":{ - "method":"DELETE", - "requestUri":"/2016-09-07/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_09_07", - "http":{ - "method":"DELETE", - "requestUri":"/2016-09-07/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_09_07", - "http":{ - "method":"DELETE", - "requestUri":"/2016-09-07/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource2016_09_07", - "http":{ - "method":"GET", - "requestUri":"/2016-09-07/tagging" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "TagResource":{ - "name":"TagResource2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/tagging?Operation=Tag", - "responseCode":204 - }, - "input":{"shape":"TagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UntagResource":{ - "name":"UntagResource2016_09_07", - "http":{ - "method":"POST", - "requestUri":"/2016-09-07/tagging?Operation=Untag", - "responseCode":204 - }, - "input":{"shape":"UntagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_09_07", - "http":{ - "method":"PUT", - "requestUri":"/2016-09-07/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_09_07", - "http":{ - "method":"PUT", - "requestUri":"/2016-09-07/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_09_07", - "http":{ - "method":"PUT", - "requestUri":"/2016-09-07/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateDistributionWithTagsRequest":{ - "type":"structure", - "required":["DistributionConfigWithTags"], - "members":{ - "DistributionConfigWithTags":{ - "shape":"DistributionConfigWithTags", - "locationName":"DistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"DistributionConfigWithTags" - }, - "CreateDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CreateStreamingDistributionWithTagsRequest":{ - "type":"structure", - "required":["StreamingDistributionConfigWithTags"], - "members":{ - "StreamingDistributionConfigWithTags":{ - "shape":"StreamingDistributionConfigWithTags", - "locationName":"StreamingDistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"StreamingDistributionConfigWithTags" - }, - "CreateStreamingDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"} - } - }, - "DistributionConfigWithTags":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Tags" - ], - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId", - "HttpVersion" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"}, - "QueryStringCacheKeys":{"shape":"QueryStringCacheKeys"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "HttpVersion":{ - "type":"string", - "enum":[ - "http1.1", - "http2" - ] - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTagging":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["Resource"], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - } - } - }, - "ListTagsForResourceResult":{ - "type":"structure", - "required":["Tags"], - "members":{ - "Tags":{"shape":"Tags"} - }, - "payload":"Tags" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchResource":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "QueryStringCacheKeys":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"QueryStringCacheKeysList"} - } - }, - "QueryStringCacheKeysList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "ResourceARN":{ - "type":"string", - "pattern":"arn:aws:cloudfront::[0-9]+:.*" - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionConfigWithTags":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Tags" - ], - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Tag":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{ - "type":"string", - "max":128, - "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "TagKeyList":{ - "type":"list", - "member":{ - "shape":"TagKey", - "locationName":"Key" - } - }, - "TagKeys":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagKeyList"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "Tags" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "Tags":{ - "shape":"Tags", - "locationName":"Tags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"Tags" - }, - "TagValue":{ - "type":"string", - "max":256, - "min":0, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "Tags":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagList"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UntagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "TagKeys" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "TagKeys":{ - "shape":"TagKeys", - "locationName":"TagKeys", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - } - }, - "payload":"TagKeys" - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-07/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/docs-2.json deleted file mode 100755 index 45b1c473d..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/docs-2.json +++ /dev/null @@ -1,1388 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon CloudFront

Amazon CloudFront is a global content delivery network (CDN) service that accelerates delivery of your websites, APIs, video content or other web assets. It integrates with other Amazon Web Services products to give developers and businesses an easy way to accelerate content to end users with no minimum usage commitments.

", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "Create a new origin access identity.", - "CreateDistribution": "Create a new distribution.", - "CreateDistributionWithTags": "Create a new distribution with tags.", - "CreateInvalidation": "Create a new invalidation.", - "CreateStreamingDistribution": "Create a new streaming distribution.", - "CreateStreamingDistributionWithTags": "Create a new streaming distribution with tags.", - "DeleteCloudFrontOriginAccessIdentity": "Delete an origin access identity.", - "DeleteDistribution": "Delete a distribution.", - "DeleteStreamingDistribution": "Delete a streaming distribution.", - "GetCloudFrontOriginAccessIdentity": "Get the information about an origin access identity.", - "GetCloudFrontOriginAccessIdentityConfig": "Get the configuration information about an origin access identity.", - "GetDistribution": "Get the information about a distribution.", - "GetDistributionConfig": "Get the configuration information about a distribution.", - "GetInvalidation": "Get the information about an invalidation.", - "GetStreamingDistribution": "Get the information about a streaming distribution.", - "GetStreamingDistributionConfig": "Get the configuration information about a streaming distribution.", - "ListCloudFrontOriginAccessIdentities": "List origin access identities.", - "ListDistributions": "List distributions.", - "ListDistributionsByWebACLId": "List the distributions that are associated with a specified AWS WAF web ACL.", - "ListInvalidations": "List invalidation batches.", - "ListStreamingDistributions": "List streaming distributions.", - "ListTagsForResource": "List tags for a CloudFront resource.", - "TagResource": "Add tags to a CloudFront resource.", - "UntagResource": "Remove tags from a CloudFront resource.", - "UpdateCloudFrontOriginAccessIdentity": "Update an origin access identity.", - "UpdateDistribution": "Update a distribution.", - "UpdateStreamingDistribution": "Update a streaming distribution." - }, - "shapes": { - "AccessDenied": { - "base": "Access denied.", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for the default cache behavior or for any of the other cache behaviors for this distribution. These are accounts that you want to allow to create signed URLs for private content.", - "refs": { - "Distribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.", - "StreamingDistribution$ActiveTrustedSigners": "CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs." - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "Optional: A complex type that contains CNAME elements, if any, for this distribution. If Quantity is 0, you can omit Items." - } - }, - "Aliases": { - "base": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "refs": { - "DistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "DistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.", - "StreamingDistributionConfig$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.", - "StreamingDistributionSummary$Aliases": "A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution." - } - }, - "AllowedMethods": { - "base": "A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices: - CloudFront forwards only GET and HEAD requests. - CloudFront forwards only GET, HEAD and OPTIONS requests. - CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests. If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you may not want users to have permission to delete objects from your origin.", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "A complex type that describes how CloudFront processes requests. You can create up to 10 cache behaviors.You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin will never be used. If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error. To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element. To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CacheBehaviors": { - "base": "A complex type that contains zero or more CacheBehavior elements.", - "refs": { - "DistributionConfig$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements.", - "DistributionSummary$CacheBehaviors": "A complex type that contains zero or more CacheBehavior elements." - } - }, - "CachedMethods": { - "base": "A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices: - CloudFront caches responses to GET and HEAD requests. - CloudFront caches responses to GET, HEAD, and OPTIONS requests. If you pick the second choice for your S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers and Origin headers for the responses to be cached correctly.", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "CloudFront origin access identity.", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information.", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "The origin access identity's information." - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "Origin access identity configuration.", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "The current configuration information for the identity.", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "The origin access identity's configuration information.", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "The identity's configuration information." - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "The CloudFrontOriginAccessIdentityList type.", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "The CloudFrontOriginAccessIdentityList type." - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "Summary of the information about a CloudFront origin access identity.", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account." - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "Optional: A complex type that contains whitelisted cookies for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "CookieNames": { - "base": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior.", - "refs": { - "CookiePreference$WhitelistedNames": "A complex type that specifies the whitelisted cookies, if any, that you want CloudFront to forward to your origin that is associated with this cache behavior." - } - }, - "CookiePreference": { - "base": "A complex type that specifies the cookie preferences associated with this cache behavior.", - "refs": { - "ForwardedValues$Cookies": "A complex type that specifies how CloudFront handles cookies." - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to create a new origin access identity.", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "The request to create a new distribution.", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateDistributionWithTagsRequest": { - "base": "The request to create a new distribution with tags", - "refs": { - } - }, - "CreateDistributionWithTagsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "The request to create an invalidation.", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "The request to create a new streaming distribution.", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsRequest": { - "base": "The request to create a new streaming distribution with tags.", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "A complex type that describes how you'd prefer CloudFront to respond to requests that result in either a 4xx or 5xx response. You can control whether a custom error page should be displayed, what the desired response code should be for this error page and how long should the error response be cached by CloudFront. If you don't want to specify any custom error responses, include only an empty CustomErrorResponses element. To delete all custom error responses in an existing distribution, update the distribution configuration and include only an empty CustomErrorResponses element. To add, change, or remove one or more custom error responses, update the distribution configuration and specify all of the custom error responses that you want to include in the updated distribution.", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "Optional: A complex type that contains custom error responses for this distribution. If Quantity is 0, you can omit Items." - } - }, - "CustomErrorResponses": { - "base": "A complex type that contains zero or more CustomErrorResponse elements.", - "refs": { - "DistributionConfig$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponse elements.", - "DistributionSummary$CustomErrorResponses": "A complex type that contains zero or more CustomErrorResponses elements." - } - }, - "CustomHeaders": { - "base": "A complex type that contains the list of Custom Headers for each origin.", - "refs": { - "Origin$CustomHeaders": "A complex type that contains information about the custom headers associated with this Origin." - } - }, - "CustomOriginConfig": { - "base": "A customer origin.", - "refs": { - "Origin$CustomOriginConfig": "A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead." - } - }, - "DefaultCacheBehavior": { - "base": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior.", - "DistributionSummary$DefaultCacheBehavior": "A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements.You must create exactly one default cache behavior." - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "The request to delete a origin access identity.", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "The request to delete a distribution.", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "The request to delete a streaming distribution.", - "refs": { - } - }, - "Distribution": { - "base": "A distribution.", - "refs": { - "CreateDistributionResult$Distribution": "The distribution's information.", - "CreateDistributionWithTagsResult$Distribution": "The distribution's information.", - "GetDistributionResult$Distribution": "The distribution's information.", - "UpdateDistributionResult$Distribution": "The distribution's information." - } - }, - "DistributionAlreadyExists": { - "base": "The caller reference you attempted to create the distribution with is associated with another distribution.", - "refs": { - } - }, - "DistributionConfig": { - "base": "A distribution Configuration.", - "refs": { - "CreateDistributionRequest$DistributionConfig": "The distribution's configuration information.", - "Distribution$DistributionConfig": "The current configuration information for the distribution.", - "DistributionConfigWithTags$DistributionConfig": "A distribution Configuration.", - "GetDistributionConfigResult$DistributionConfig": "The distribution's configuration information.", - "UpdateDistributionRequest$DistributionConfig": "The distribution's configuration information." - } - }, - "DistributionConfigWithTags": { - "base": "A distribution Configuration and a list of tags to be associated with the distribution.", - "refs": { - "CreateDistributionWithTagsRequest$DistributionConfigWithTags": "The distribution's configuration information." - } - }, - "DistributionList": { - "base": "A distribution list.", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "The DistributionList type.", - "ListDistributionsResult$DistributionList": "The DistributionList type." - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront distribution.", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "ForwardedValues": { - "base": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "refs": { - "CacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers.", - "DefaultCacheBehavior$ForwardedValues": "A complex type that specifies how CloudFront handles query strings, cookies and headers." - } - }, - "GeoRestriction": { - "base": "A complex type that controls the countries in which your content is distributed. For more information about geo restriction, go to Customizing Error Responses in the Amazon CloudFront Developer Guide. CloudFront determines the location of your users using MaxMind GeoIP databases. For information about the accuracy of these databases, see How accurate are your GeoIP databases? on the MaxMind website.", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "The method that you want to use to restrict distribution of your content by country: - none: No geo restriction is enabled, meaning access to content is not restricted by client geo location. - blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content. - whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content." - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "The request to get an origin access identity's configuration.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "The request to get an origin access identity's information.", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "The request to get a distribution configuration.", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "The request to get a distribution's information.", - "refs": { - } - }, - "GetDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "The request to get an invalidation's information.", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "To request to get a streaming distribution configuration.", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "The request to get a streaming distribution's information.", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "Optional: A complex type that contains a Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items." - } - }, - "Headers": { - "base": "A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior. For the headers that you specify, CloudFront also caches separate versions of a given object based on the header values in viewer requests; this is known as varying on headers. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to vary on the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value.", - "refs": { - "ForwardedValues$Headers": "A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior." - } - }, - "HttpVersion": { - "base": null, - "refs": { - "DistributionConfig$HttpVersion": "(Optional) Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 will automatically use an earlier version.", - "DistributionSummary$HttpVersion": "Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 will automatically use an earlier version." - } - }, - "IllegalUpdate": { - "base": "Origin and CallerReference cannot be updated.", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "The value of Quantity and the size of Items do not match.", - "refs": { - } - }, - "InvalidArgument": { - "base": "The argument is invalid.", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "The default root object file name is too big or contains an invalid character.", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "The If-Match version is missing or not valid for the distribution.", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "The origin access identity is not valid or doesn't exist.", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that Support Server Name Indication (SNI).", - "refs": { - } - }, - "InvalidQueryStringParameters": { - "base": null, - "refs": { - } - }, - "InvalidRelativePath": { - "base": "The relative path is too big, is not URL-encoded, or does not begin with a slash (/).", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidTagging": { - "base": "The specified tagging for a CloudFront resource is invalid. For more information, see the error text.", - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "An invalidation.", - "refs": { - "CreateInvalidationResult$Invalidation": "The invalidation's information.", - "GetInvalidationResult$Invalidation": "The invalidation's information." - } - }, - "InvalidationBatch": { - "base": "An invalidation batch.", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "The batch information for the invalidation.", - "Invalidation$InvalidationBatch": "The current invalidation information for the batch request." - } - }, - "InvalidationList": { - "base": "An invalidation list.", - "refs": { - "ListInvalidationsResult$InvalidationList": "Information about invalidation batches." - } - }, - "InvalidationSummary": { - "base": "Summary of an invalidation request.", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "A complex type that contains one InvalidationSummary element for each invalidation batch that was created by the current AWS account." - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "Use this element to specify whether you want CloudFront to forward cookies to the origin that is associated with this cache behavior. You can specify all, none or whitelist. If you choose All, CloudFront forwards all cookies regardless of how many your application uses." - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "KeyPairIds": { - "base": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.", - "refs": { - "Signer$KeyPairIds": "A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber." - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "The request to list origin access identities.", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "The request to list distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "The request to list your distributions.", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "The request to list invalidations.", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "The request to list your streaming distributions.", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": "The request to list tags for a CloudFront resource.", - "refs": { - } - }, - "ListTagsForResourceResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist). The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country. CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes." - } - }, - "LoggingConfig": { - "base": "A complex type that controls whether access logs are written for the distribution.", - "refs": { - "DistributionConfig$Logging": "A complex type that controls whether access logs are written for the distribution." - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.", - "CachedMethods$Items": "A complex type that contains the HTTP methods that you want CloudFront to cache responses to." - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "Specify the minimum version of the SSL protocol that you want CloudFront to use, SSLv3 or TLSv1, for HTTPS connections. CloudFront will serve your objects only to browsers or devices that support at least the SSL version that you specify. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using dedicated IP (if you specify vip for SSLSupportMethod), you can choose SSLv3 or TLSv1 as the MinimumProtocolVersion. If you're using a custom certificate (if you specify a value for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion." - } - }, - "MissingBody": { - "base": "This operation requires a body. Ensure that the body is present and the Content-Type header is set.", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "The specified origin access identity does not exist.", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "The specified distribution does not exist.", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "The specified invalidation does not exist.", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "No origin exists with the specified Origin Id.", - "refs": { - } - }, - "NoSuchResource": { - "base": "The specified CloudFront resource does not exist.", - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "The specified streaming distribution does not exist.", - "refs": { - } - }, - "Origin": { - "base": "A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files.You must create at least one origin.", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "A complex type that contains information related to a Header", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "A complex type that contains the custom headers for this Origin." - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "A complex type that contains origins for this distribution." - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "The origin protocol policy to apply to your origin." - } - }, - "OriginSslProtocols": { - "base": "A complex type that contains the list of SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS." - } - }, - "Origins": { - "base": "A complex type that contains information about origins for this distribution.", - "refs": { - "DistributionConfig$Origins": "A complex type that contains information about origins for this distribution.", - "DistributionSummary$Origins": "A complex type that contains information about origins for this distribution." - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "A complex type that contains a list of the objects that you want to invalidate." - } - }, - "Paths": { - "base": "A complex type that contains information about the objects that you want to invalidate.", - "refs": { - "InvalidationBatch$Paths": "The path of the object to invalidate. The path is relative to the distribution and must begin with a slash (/). You must enclose each invalidation object with the Path element tags. If the path includes non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not invalidate the old version of the updated object." - } - }, - "PreconditionFailed": { - "base": "The precondition given in one or more of the request-header fields evaluated to false.", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "A complex type that contains information about price class for this distribution.", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "A complex type that contains information about price class for this streaming distribution.", - "StreamingDistributionSummary$PriceClass": null - } - }, - "QueryStringCacheKeys": { - "base": null, - "refs": { - "ForwardedValues$QueryStringCacheKeys": "

A complex type that contains information about the query string parameters that you want CloudFront to use for caching for this cache behavior.

" - } - }, - "QueryStringCacheKeysList": { - "base": null, - "refs": { - "QueryStringCacheKeys$Items": "Optional: A list that contains the query string parameters that you want CloudFront to use as a basis for caching for this cache behavior. If Quantity is 0, you can omit Items." - } - }, - "ResourceARN": { - "base": null, - "refs": { - "ListTagsForResourceRequest$Resource": "An ARN of a CloudFront resource.", - "TagResourceRequest$Resource": "An ARN of a CloudFront resource.", - "UntagResourceRequest$Resource": "An ARN of a CloudFront resource." - } - }, - "Restrictions": { - "base": "A complex type that identifies ways in which you want to restrict distribution of your content.", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "refs": { - "StreamingDistributionConfig$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.", - "StreamingDistributionSummary$S3Origin": "A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution." - } - }, - "S3OriginConfig": { - "base": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.", - "refs": { - "Origin$S3OriginConfig": "A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead." - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "If you specify a value for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests. Valid values are vip and sni-only. If you specify vip, CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges. If you specify sni-only, CloudFront can only respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. Do not specify a value for SSLSupportMethod if you specified true for CloudFrontDefaultCertificate." - } - }, - "Signer": { - "base": "A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "A complex type that contains one Signer complex type for each unique trusted signer that is specified in the TrustedSigners complex type, including trusted signers in the default cache behavior and in all of the other cache behaviors." - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "A complex type that contains one SslProtocol element for each SSL/TLS protocol that you want to allow CloudFront to use when establishing an HTTPS connection with this origin." - } - }, - "StreamingDistribution": { - "base": "A streaming distribution.", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "CreateStreamingDistributionWithTagsResult$StreamingDistribution": "The streaming distribution's information.", - "GetStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information.", - "UpdateStreamingDistributionResult$StreamingDistribution": "The streaming distribution's information." - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "The configuration for the streaming distribution.", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "The streaming distribution's configuration information.", - "StreamingDistribution$StreamingDistributionConfig": "The current configuration information for the streaming distribution.", - "StreamingDistributionConfigWithTags$StreamingDistributionConfig": "A streaming distribution Configuration.", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionConfigWithTags": { - "base": "A streaming distribution Configuration and a list of tags to be associated with the streaming distribution.", - "refs": { - "CreateStreamingDistributionWithTagsRequest$StreamingDistributionConfigWithTags": "The streaming distribution's configuration information." - } - }, - "StreamingDistributionList": { - "base": "A streaming distribution list.", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "The StreamingDistributionList type." - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "A summary of the information for an Amazon CloudFront streaming distribution.", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account." - } - }, - "StreamingLoggingConfig": { - "base": "A complex type that controls whether access logs are written for this streaming distribution.", - "refs": { - "StreamingDistributionConfig$Logging": "A complex type that controls whether access logs are written for the streaming distribution." - } - }, - "Tag": { - "base": "A complex type that contains Tag key and Tag value.", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": "A string that contains Tag key. The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.", - "refs": { - "Tag$Key": "A string that contains Tag key. The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.", - "TagKeyList$member": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "TagKeys$Items": "A complex type that contains Tag key elements" - } - }, - "TagKeys": { - "base": "A complex type that contains zero or more Tag elements.", - "refs": { - "UntagResourceRequest$TagKeys": "A complex type that contains zero or more Tag key elements." - } - }, - "TagList": { - "base": null, - "refs": { - "Tags$Items": "A complex type that contains Tag elements" - } - }, - "TagResourceRequest": { - "base": "The request to add tags to a CloudFront resource.", - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "A string that contains an optional Tag value. The string length should be between 0 and 256 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @." - } - }, - "Tags": { - "base": "A complex type that contains zero or more Tag elements.", - "refs": { - "DistributionConfigWithTags$Tags": "A complex type that contains zero or more Tag elements.", - "ListTagsForResourceResult$Tags": "A complex type that contains zero or more Tag elements.", - "StreamingDistributionConfigWithTags$Tags": "A complex type that contains zero or more Tag elements.", - "TagResourceRequest$Tags": "A complex type that contains zero or more Tag elements." - } - }, - "TooManyCacheBehaviors": { - "base": "You cannot create anymore cache behaviors for the distribution.", - "refs": { - } - }, - "TooManyCertificates": { - "base": "You cannot create anymore custom ssl certificates.", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "Processing your request would cause you to exceed the maximum number of origin access identities allowed.", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "Your request contains more cookie names in the whitelist than are allowed per cache behavior.", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "Your request contains more CNAMEs than are allowed per distribution.", - "refs": { - } - }, - "TooManyDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of distributions allowed.", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "You cannot create anymore origins for the distribution.", - "refs": { - } - }, - "TooManyQueryStringParameters": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "Processing your request would cause you to exceed the maximum number of streaming distributions allowed.", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "Your request contains more trusted signers than are allowed per distribution.", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "One or more of your trusted signers do not exist.", - "refs": { - } - }, - "TrustedSigners": { - "base": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "refs": { - "CacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "DefaultCacheBehavior$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionConfig$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.", - "StreamingDistributionSummary$TrustedSigners": "A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, go to Using a Signed URL to Serve Private Content in the Amazon CloudFront Developer Guide. If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution." - } - }, - "UntagResourceRequest": { - "base": "The request to remove tags from a CloudFront resource.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "The request to update an origin access identity.", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "The request to update a distribution.", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "The request to update a streaming distribution.", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "The returned result of the corresponding request.", - "refs": { - } - }, - "ViewerCertificate": { - "base": "A complex type that contains information about viewer certificates for this distribution.", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL.", - "DefaultCacheBehavior$ViewerProtocolPolicy": "Use this element to specify the protocol that users can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. If you want CloudFront to allow end users to use any available protocol, specify allow-all. If you want CloudFront to require HTTPS, specify https. If you want CloudFront to respond to an HTTP request with an HTTP status code of 301 (Moved Permanently) and the HTTPS URL, specify redirect-to-https. The viewer then resubmits the request using the HTTPS URL." - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "Each active trusted signer.", - "CacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "CacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "CloudFrontOriginAccessIdentityList$IsTruncated": "A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.", - "DefaultCacheBehavior$SmoothStreaming": "Indicates whether you want to distribute media files in Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false.", - "DefaultCacheBehavior$Compress": "Whether you want CloudFront to automatically compress content for web requests that include Accept-Encoding: gzip in the request header. If so, specify true; if not, specify false. CloudFront compresses files larger than 1000 bytes and less than 1 megabyte for both Amazon S3 and custom origins. When a CloudFront edge location is unusually busy, some files might not be compressed. The value of the Content-Type header must be on the list of file types that CloudFront will compress. For the current list, see Serving Compressed Content in the Amazon CloudFront Developer Guide. If you configure CloudFront to compress content, CloudFront removes the ETag response header from the objects that it compresses. The ETag header indicates that the version in a CloudFront edge cache is identical to the version on the origin server, but after compression the two versions are no longer identical. As a result, for compressed objects, CloudFront can't use the ETag header to determine whether an expired object in the CloudFront edge cache is still the latest version.", - "DistributionConfig$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "DistributionList$IsTruncated": "A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "DistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "ForwardedValues$QueryString": "

Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior and cache based on the query string parameters. CloudFront behavior depends on the value of QueryString and on the values that you specify for QueryStringCacheKeys, if any:

  • If you specify true for QueryString and you don't specify any values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin and caches based on all query string parameters. Depending on how many query string parameters and values you have, this can adversely affect performance because CloudFront must forward more requests to the origin.
  • If you specify true for QueryString and you specify one or more values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin, but it only caches based on the query string parameters that you specify.
  • If you specify false for QueryString, CloudFront doesn't forward any query string parameters to the origin, and doesn't cache based on query string parameters.
", - "InvalidationList$IsTruncated": "A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.", - "LoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix and IncludeCookies, the values are automatically deleted.", - "LoggingConfig$IncludeCookies": "Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.", - "StreamingDistributionConfig$Enabled": "Whether the streaming distribution is enabled to accept end user requests for content.", - "StreamingDistributionList$IsTruncated": "A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.", - "StreamingDistributionSummary$Enabled": "Whether the distribution is enabled to accept end user requests for content.", - "StreamingLoggingConfig$Enabled": "Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.", - "TrustedSigners$Enabled": "Specifies whether you want to require end users to use signed URLs to access the files specified by PathPattern and TargetOriginId.", - "ViewerCertificate$CloudFrontDefaultCertificate": "If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an ACMCertificateArn or IAMCertificateId." - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "The number of unique trusted signers included in all cache behaviors. For example, if three cache behaviors all list the same three AWS accounts, the value of Quantity for ActiveTrustedSigners will be 3.", - "Aliases$Quantity": "The number of CNAMEs, if any, for this distribution.", - "AllowedMethods$Quantity": "The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).", - "CacheBehaviors$Quantity": "The number of cache behaviors for this distribution.", - "CachedMethods$Quantity": "The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).", - "CloudFrontOriginAccessIdentityList$MaxItems": "The value you provided for the MaxItems request parameter.", - "CloudFrontOriginAccessIdentityList$Quantity": "The number of CloudFront origin access identities that were created by the current AWS account.", - "CookieNames$Quantity": "The number of whitelisted cookies for this cache behavior.", - "CustomErrorResponse$ErrorCode": "The 4xx or 5xx HTTP status code that you want to customize. For a list of HTTP status codes that you can customize, see CloudFront documentation.", - "CustomErrorResponses$Quantity": "The number of custom error responses for this distribution.", - "CustomHeaders$Quantity": "The number of custom headers for this origin.", - "CustomOriginConfig$HTTPPort": "The HTTP port the custom origin listens on.", - "CustomOriginConfig$HTTPSPort": "The HTTPS port the custom origin listens on.", - "Distribution$InProgressInvalidationBatches": "The number of invalidation batches currently in progress.", - "DistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "DistributionList$Quantity": "The number of distributions that were created by the current AWS account.", - "GeoRestriction$Quantity": "When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.", - "Headers$Quantity": "The number of different headers that you want CloudFront to forward to the origin and to vary on for this cache behavior. The maximum number of headers that you can specify by name is 10. If you want CloudFront to forward all headers to the origin and vary on all of them, specify 1 for Quantity and * for Name. If you don't want CloudFront to forward any additional headers to the origin or to vary on any headers, specify 0 for Quantity and omit Items.", - "InvalidationList$MaxItems": "The value you provided for the MaxItems request parameter.", - "InvalidationList$Quantity": "The number of invalidation batches that were created by the current AWS account.", - "KeyPairIds$Quantity": "The number of active CloudFront key pairs for AwsAccountNumber.", - "OriginSslProtocols$Quantity": "The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.", - "Origins$Quantity": "The number of origins for this distribution.", - "Paths$Quantity": "The number of objects that you want to invalidate.", - "QueryStringCacheKeys$Quantity": "The number of whitelisted query string parameters for this cache behavior.", - "StreamingDistributionList$MaxItems": "The value you provided for the MaxItems request parameter.", - "StreamingDistributionList$Quantity": "The number of streaming distributions that were created by the current AWS account.", - "TrustedSigners$Quantity": "The number of trusted signers for this cache behavior." - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "CustomErrorResponse$ErrorCachingMinTTL": "The minimum amount of time you want HTTP error codes to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated. You can specify a value from 0 to 31,536,000.", - "DefaultCacheBehavior$MinTTL": "The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront queries your origin to see whether the object has been updated.You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$DefaultTTL": "If you don't configure your origin to add a Cache-Control max-age directive or an Expires header, DefaultTTL is the default amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years).", - "DefaultCacheBehavior$MaxTTL": "The maximum amount of time (in seconds) that an object is in a CloudFront cache before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. You can specify a value from 0 to 3,153,600,000 seconds (100 years)." - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "The pattern (for example, images/*.jpg) that specifies which requests you want this cache behavior to apply to. When CloudFront receives an end-user request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution. The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.", - "CacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "CloudFrontOriginAccessIdentity$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created. If the CallerReference is a value you already sent in a previous request to create an identity, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.", - "CloudFrontOriginAccessIdentityConfig$Comment": "Any comments you want to include about the origin access identity.", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "The value you provided for the Marker request parameter.", - "CloudFrontOriginAccessIdentityList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.", - "CloudFrontOriginAccessIdentitySummary$Id": "The ID for the origin access identity. For example: E74FTE3AJFJ256A.", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.", - "CloudFrontOriginAccessIdentitySummary$Comment": "The comment for this origin access identity, as originally specified when created.", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity created.", - "CreateDistributionResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionResult$ETag": "The current version of the distribution created.", - "CreateDistributionWithTagsResult$Location": "The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.", - "CreateDistributionWithTagsResult$ETag": "The current version of the distribution created.", - "CreateInvalidationRequest$DistributionId": "The distribution's id.", - "CreateInvalidationResult$Location": "The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.", - "CreateStreamingDistributionResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionResult$ETag": "The current version of the streaming distribution created.", - "CreateStreamingDistributionWithTagsResult$Location": "The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.", - "CreateStreamingDistributionWithTagsResult$ETag": "The current version of the streaming distribution created.", - "CustomErrorResponse$ResponsePagePath": "The path of the custom error page (for example, /custom_404.html). The path is relative to the distribution and must begin with a slash (/). If the path includes any non-ASCII characters or unsafe characters as defined in RFC 1783 (http://www.ietf.org/rfc/rfc1738.txt), URL encode those characters. Do not URL encode any other characters in the path, or CloudFront will not return the custom error page to the viewer.", - "CustomErrorResponse$ResponseCode": "The HTTP status code that you want CloudFront to return with the custom error page to the viewer. For a list of HTTP status codes that you can replace, see CloudFront Documentation.", - "DefaultCacheBehavior$TargetOriginId": "The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "The origin access identity's id.", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.", - "DeleteDistributionRequest$Id": "The distribution id.", - "DeleteDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.", - "DeleteStreamingDistributionRequest$Id": "The distribution id.", - "DeleteStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.", - "Distribution$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "Distribution$ARN": "The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "Distribution$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "Distribution$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the DistributionConfig object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create a distribution, and the content of the DistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "DistributionConfig$DefaultRootObject": "The object that you want CloudFront to return (for example, index.html) when an end user requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/index.html). Specifying a default root object avoids exposing the contents of your distribution. If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element. To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element. To replace the default root object, update the distribution configuration and specify the new object.", - "DistributionConfig$Comment": "Any comments you want to include about the distribution.", - "DistributionConfig$WebACLId": "(Optional) If you're using AWS WAF to filter CloudFront requests, the Id of the AWS WAF web ACL that is associated with the distribution.", - "DistributionList$Marker": "The value you provided for the Marker request parameter.", - "DistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "DistributionSummary$ARN": "The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "DistributionSummary$Status": "This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "DistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "DistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "DistributionSummary$WebACLId": "The Web ACL Id (if any) associated with the distribution.", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "GetCloudFrontOriginAccessIdentityResult$ETag": "The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.", - "GetDistributionConfigRequest$Id": "The distribution's id.", - "GetDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetDistributionRequest$Id": "The distribution's id.", - "GetDistributionResult$ETag": "The current version of the distribution's information. For example: E2QWRUHAPOMQZL.", - "GetInvalidationRequest$DistributionId": "The distribution's id.", - "GetInvalidationRequest$Id": "The invalidation's id.", - "GetStreamingDistributionConfigRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionConfigResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "GetStreamingDistributionRequest$Id": "The streaming distribution's id.", - "GetStreamingDistributionResult$ETag": "The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidQueryStringParameters$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidTagging$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "The identifier for the invalidation request. For example: IDFDVBD632BHDS5.", - "Invalidation$Status": "The status of the invalidation request. When the invalidation batch is finished, the status is Completed.", - "InvalidationBatch$CallerReference": "A unique name that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the Path object), a new distribution is created. If the CallerReference is a value you already sent in a previous request to create an invalidation batch, and the content of each Path element is identical to the original request, the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a distribution but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.", - "InvalidationList$Marker": "The value you provided for the Marker request parameter.", - "InvalidationList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your invalidation batches where they left off.", - "InvalidationSummary$Id": "The unique ID for an invalidation request.", - "InvalidationSummary$Status": "The status of an invalidation request.", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "The maximum number of origin access identities you want in the response body.", - "ListDistributionsByWebACLIdRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsByWebACLIdRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListDistributionsByWebACLIdRequest$WebACLId": "The Id of the AWS WAF web ACL for which you want to list the associated distributions. If you specify \"null\" for the Id, the request returns a list of the distributions that aren't associated with a web ACL.", - "ListDistributionsRequest$Marker": "Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)", - "ListDistributionsRequest$MaxItems": "The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.", - "ListInvalidationsRequest$DistributionId": "The distribution's id.", - "ListInvalidationsRequest$Marker": "Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.", - "ListInvalidationsRequest$MaxItems": "The maximum number of invalidation batches you want in the response body.", - "ListStreamingDistributionsRequest$Marker": "Use this when paginating results to indicate where to begin in your list of streaming distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).", - "ListStreamingDistributionsRequest$MaxItems": "The maximum number of streaming distributions you want in the response body.", - "LocationList$member": null, - "LoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "LoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchResource$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "A unique identifier for the origin. The value of Id must be unique within the distribution. You use the value of Id when you create a cache behavior. The Id identifies the origin that CloudFront routes a request to when the request matches the path pattern for that cache behavior.", - "Origin$DomainName": "Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com. Custom origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.", - "Origin$OriginPath": "An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName.", - "OriginCustomHeader$HeaderName": "The header's name.", - "OriginCustomHeader$HeaderValue": "The header's value.", - "PathList$member": null, - "PreconditionFailed$Message": null, - "QueryStringCacheKeysList$member": null, - "S3Origin$DomainName": "The DNS name of the S3 origin.", - "S3Origin$OriginAccessIdentity": "Your S3 origin's origin access identity.", - "S3OriginConfig$OriginAccessIdentity": "The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that end users can only access objects in an Amazon S3 bucket through CloudFront. If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element. To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element. To replace the origin access identity, update the distribution configuration and specify the new origin access identity. Use the format origin-access-identity/cloudfront/Id where Id is the value that CloudFront returned in the Id element when you created the origin access identity.", - "Signer$AwsAccountNumber": "Specifies an AWS account that can create signed URLs. Values: self, which indicates that the AWS account that was used to create the distribution can created signed URLs, or an AWS account number. Omit the dashes in the account number.", - "StreamingDistribution$Id": "The identifier for the streaming distribution. For example: EGTXBD79H29TRA8.", - "StreamingDistribution$ARN": "The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "StreamingDistribution$Status": "The current status of the streaming distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistribution$DomainName": "The domain name corresponding to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "A unique number that ensures the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.", - "StreamingDistributionConfig$Comment": "Any comments you want to include about the streaming distribution.", - "StreamingDistributionList$Marker": "The value you provided for the Marker request parameter.", - "StreamingDistributionList$NextMarker": "If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your streaming distributions where they left off.", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "The identifier for the distribution. For example: EDFDVBD632BHDS5.", - "StreamingDistributionSummary$ARN": "The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account Id.", - "StreamingDistributionSummary$Status": "Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.", - "StreamingDistributionSummary$DomainName": "The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.", - "StreamingDistributionSummary$Comment": "The comment originally specified when this distribution was created.", - "StreamingLoggingConfig$Bucket": "The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.", - "StreamingLoggingConfig$Prefix": "An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyQueryStringParameters$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "The identity's id.", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "The value of the ETag header you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionRequest$Id": "The distribution's id.", - "UpdateDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionRequest$Id": "The streaming distribution's id.", - "UpdateStreamingDistributionRequest$IfMatch": "The value of the ETag header you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.", - "UpdateStreamingDistributionResult$ETag": "The current version of the configuration. For example: E2QWRUHAPOMQZL.", - "ViewerCertificate$IAMCertificateId": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value, ACMCertificateArn, or CloudFrontDefaultCertificate.", - "ViewerCertificate$ACMCertificateArn": "If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the ACM certificate ARN of the custom viewer certificate for this distribution. Specify either this value, IAMCertificateId, or CloudFrontDefaultCertificate.", - "ViewerCertificate$Certificate": "Note: this field is deprecated. Please use one of [ACMCertificateArn, IAMCertificateId, CloudFrontDefaultCertificate]." - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "The date and time the distribution was last modified.", - "DistributionSummary$LastModifiedTime": "The date and time the distribution was last modified.", - "Invalidation$CreateTime": "The date and time the invalidation request was first made.", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "The date and time the distribution was last modified.", - "StreamingDistributionSummary$LastModifiedTime": "The date and time the distribution was last modified." - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/examples-1.json deleted file mode 100755 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/paginators-1.json deleted file mode 100755 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/waiters-2.json deleted file mode 100755 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-07/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/api-2.json deleted file mode 100644 index b905c836f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/api-2.json +++ /dev/null @@ -1,2599 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2016-09-29", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4", - "uid":"cloudfront-2016-09-07" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "CreateDistributionWithTags":{ - "name":"CreateDistributionWithTags2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionWithTagsRequest"}, - "output":{"shape":"CreateDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"InvalidTagging"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistributionWithTags":{ - "name":"CreateStreamingDistributionWithTags2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/streaming-distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionWithTagsRequest"}, - "output":{"shape":"CreateStreamingDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"}, - {"shape":"InvalidTagging"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_09_29", - "http":{ - "method":"DELETE", - "requestUri":"/2016-09-29/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_09_29", - "http":{ - "method":"DELETE", - "requestUri":"/2016-09-29/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_09_29", - "http":{ - "method":"DELETE", - "requestUri":"/2016-09-29/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource2016_09_29", - "http":{ - "method":"GET", - "requestUri":"/2016-09-29/tagging" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "TagResource":{ - "name":"TagResource2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/tagging?Operation=Tag", - "responseCode":204 - }, - "input":{"shape":"TagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UntagResource":{ - "name":"UntagResource2016_09_29", - "http":{ - "method":"POST", - "requestUri":"/2016-09-29/tagging?Operation=Untag", - "responseCode":204 - }, - "input":{"shape":"UntagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_09_29", - "http":{ - "method":"PUT", - "requestUri":"/2016-09-29/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_09_29", - "http":{ - "method":"PUT", - "requestUri":"/2016-09-29/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_09_29", - "http":{ - "method":"PUT", - "requestUri":"/2016-09-29/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateDistributionWithTagsRequest":{ - "type":"structure", - "required":["DistributionConfigWithTags"], - "members":{ - "DistributionConfigWithTags":{ - "shape":"DistributionConfigWithTags", - "locationName":"DistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"DistributionConfigWithTags" - }, - "CreateDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CreateStreamingDistributionWithTagsRequest":{ - "type":"structure", - "required":["StreamingDistributionConfigWithTags"], - "members":{ - "StreamingDistributionConfigWithTags":{ - "shape":"StreamingDistributionConfigWithTags", - "locationName":"StreamingDistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"StreamingDistributionConfigWithTags" - }, - "CreateStreamingDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"}, - "IsIPV6Enabled":{"shape":"boolean"} - } - }, - "DistributionConfigWithTags":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Tags" - ], - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId", - "HttpVersion", - "IsIPV6Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"}, - "IsIPV6Enabled":{"shape":"boolean"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"}, - "QueryStringCacheKeys":{"shape":"QueryStringCacheKeys"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "HttpVersion":{ - "type":"string", - "enum":[ - "http1.1", - "http2" - ] - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTagging":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["Resource"], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - } - } - }, - "ListTagsForResourceResult":{ - "type":"structure", - "required":["Tags"], - "members":{ - "Tags":{"shape":"Tags"} - }, - "payload":"Tags" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchResource":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "QueryStringCacheKeys":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"QueryStringCacheKeysList"} - } - }, - "QueryStringCacheKeysList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "ResourceARN":{ - "type":"string", - "pattern":"arn:aws:cloudfront::[0-9]+:.*" - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionConfigWithTags":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Tags" - ], - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Tag":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{ - "type":"string", - "max":128, - "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "TagKeyList":{ - "type":"list", - "member":{ - "shape":"TagKey", - "locationName":"Key" - } - }, - "TagKeys":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagKeyList"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "Tags" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "Tags":{ - "shape":"Tags", - "locationName":"Tags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"Tags" - }, - "TagValue":{ - "type":"string", - "max":256, - "min":0, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "Tags":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagList"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UntagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "TagKeys" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "TagKeys":{ - "shape":"TagKeys", - "locationName":"TagKeys", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - } - }, - "payload":"TagKeys" - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-09-29/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/docs-2.json deleted file mode 100644 index 78e85b1a6..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/docs-2.json +++ /dev/null @@ -1,1390 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon CloudFront

This is the Amazon CloudFront API Reference. This guide is for developers who need detailed information about the CloudFront API actions, data types, and errors. For detailed information about CloudFront features and their associated API calls, see the Amazon CloudFront Developer Guide.

", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "

Creates a new origin access identity. If you're using Amazon S3 for your origin, you can use an origin access identity to require users to access your content using a CloudFront URL instead of the Amazon S3 URL. For more information about how to use origin access identities, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "CreateDistribution": "

Creates a new web distribution. Send a GET request to the /CloudFront API version/distribution/distribution ID resource.

", - "CreateDistributionWithTags": "

Create a new distribution with tags.

", - "CreateInvalidation": "

Create a new invalidation.

", - "CreateStreamingDistribution": "

Creates a new RMTP distribution. An RTMP distribution is similar to a web distribution, but an RTMP distribution streams media files using the Adobe Real-Time Messaging Protocol (RTMP) instead of serving files using HTTP.

To create a new web distribution, submit a POST request to the CloudFront API version/distribution resource. The request body must include a document with a StreamingDistributionConfig element. The response echoes the StreamingDistributionConfig element and returns other information about the RTMP distribution.

To get the status of your request, use the GET StreamingDistribution API action. When the value of Enabled is true and the value of Status is Deployed, your distribution is ready. A distribution usually deploys in less than 15 minutes.

For more information about web distributions, see Working with RTMP Distributions in the Amazon CloudFront Developer Guide.

Beginning with the 2012-05-05 version of the CloudFront API, we made substantial changes to the format of the XML document that you include in the request body when you create or update a web distribution or an RTMP distribution, and when you invalidate objects. With previous versions of the API, we discovered that it was too easy to accidentally delete one or more values for an element that accepts multiple values, for example, CNAMEs and trusted signers. Our changes for the 2012-05-05 release are intended to prevent these accidental deletions and to notify you when there's a mismatch between the number of values you say you're specifying in the Quantity element and the number of values specified.

", - "CreateStreamingDistributionWithTags": "

Create a new streaming distribution with tags.

", - "DeleteCloudFrontOriginAccessIdentity": "

Delete an origin access identity.

", - "DeleteDistribution": "

Delete a distribution.

", - "DeleteStreamingDistribution": "

Delete a streaming distribution. To delete an RTMP distribution using the CloudFront API, perform the following steps.

To delete an RTMP distribution using the CloudFront API:

  1. Disable the RTMP distribution.

  2. Submit a GET Streaming Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Streaming Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Streaming Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Then set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  5. Review the response to the PUT Streaming Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Streaming Distribution Config request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Streaming Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  8. Review the response to your DELETE Streaming Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", - "GetCloudFrontOriginAccessIdentity": "

Get the information about an origin access identity.

", - "GetCloudFrontOriginAccessIdentityConfig": "

Get the configuration information about an origin access identity.

", - "GetDistribution": "

Get the information about a distribution.

", - "GetDistributionConfig": "

Get the configuration information about a distribution.

", - "GetInvalidation": "

Get the information about an invalidation.

", - "GetStreamingDistribution": "

Gets information about a specified RTMP distribution, including the distribution configuration.

", - "GetStreamingDistributionConfig": "

Get the configuration information about a streaming distribution.

", - "ListCloudFrontOriginAccessIdentities": "

Lists origin access identities.

", - "ListDistributions": "

List distributions.

", - "ListDistributionsByWebACLId": "

List the distributions that are associated with a specified AWS WAF web ACL.

", - "ListInvalidations": "

Lists invalidation batches.

", - "ListStreamingDistributions": "

List streaming distributions.

", - "ListTagsForResource": "

List tags for a CloudFront resource.

", - "TagResource": "

Add tags to a CloudFront resource.

", - "UntagResource": "

Remove tags from a CloudFront resource.

", - "UpdateCloudFrontOriginAccessIdentity": "

Update an origin access identity.

", - "UpdateDistribution": "

Update a distribution.

", - "UpdateStreamingDistribution": "

Update a streaming distribution.

" - }, - "shapes": { - "AccessDenied": { - "base": "

Access denied.

", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "

A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for this distribution. These are the accounts that you want to allow to create signed URLs for private content.

The Signer complex type lists the AWS account number of the trusted signer or self if the signer is the AWS account that created the distribution. The Signer element also includes the IDs of any active CloudFront key pairs that are associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create signed URLs.

For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "refs": { - "Distribution$ActiveTrustedSigners": "

CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.

", - "StreamingDistribution$ActiveTrustedSigners": "

A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for this distribution. These are the accounts that you want to allow to create signed URLs for private content.

The Signer complex type lists the AWS account number of the trusted signer or self if the signer is the AWS account that created the distribution. The Signer element also includes the IDs of any active CloudFront key pairs that are associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create signed URLs.

For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

" - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "

A complex type that contains the CNAME aliases, if any, that you want to associate with this distribution.

" - } - }, - "Aliases": { - "base": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "refs": { - "DistributionConfig$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "DistributionSummary$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "StreamingDistributionConfig$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.

", - "StreamingDistributionSummary$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.

" - } - }, - "AllowedMethods": { - "base": "

A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices:

  • CloudFront forwards only GET and HEAD requests.

  • CloudFront forwards only GET, HEAD, and OPTIONS requests.

  • CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests.

If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you might not want users to have permissions to delete objects from your origin.

", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "

Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "

A complex type that describes how CloudFront processes requests.

You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin is never used.

For the current limit on the number of cache behaviors that you can add to a distribution, see Amazon CloudFront Limits in the AWS General Reference.

If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error.

To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element.

To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.

For more information about cache behaviors, see Cache Behaviors in the Amazon CloudFront Developer Guide.

", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "

Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items.

" - } - }, - "CacheBehaviors": { - "base": "

A complex type that contains zero or more CacheBehavior elements.

", - "refs": { - "DistributionConfig$CacheBehaviors": "

A complex type that contains zero or more CacheBehavior elements.

", - "DistributionSummary$CacheBehaviors": "

A complex type that contains zero or more CacheBehavior elements.

" - } - }, - "CachedMethods": { - "base": "

A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices:

  • CloudFront caches responses to GET and HEAD requests.

  • CloudFront caches responses to GET, HEAD, and OPTIONS requests.

If you pick the second choice for your Amazon S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers, and Origin headers for the responses to be cached correctly.

", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "

This field is deprecated. You can use one of the following: [ACMCertificateArn, IAMCertificateId, or CloudFrontDefaultCertificate].

" - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "

CloudFront origin access identity.

", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

" - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "

If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.

", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "

Origin access identity configuration. Send a GET request to the /CloudFront API version/CloudFront/identity ID/config resource.

", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "

The current configuration information for the identity.

", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "

The current configuration information for the identity.

", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "

The origin access identity's configuration information.

", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "

The identity's configuration information.

" - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "

Lists the origin access identities for CloudFront.Send a GET request to the /CloudFront API version/origin-access-identity/cloudfront resource. The response includes a CloudFrontOriginAccessIdentityList element with zero or more CloudFrontOriginAccessIdentitySummary child elements. By default, your entire list of origin access identities is returned in one single page. If the list is long, you can paginate it using the MaxItems and Marker parameters.

", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "

The CloudFrontOriginAccessIdentityList type.

" - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "

Summary of the information about a CloudFront origin access identity.

", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "

A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account.

" - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "

A complex type that contains one Name element for each cookie that you want CloudFront to forward to the origin for this cache behavior.

" - } - }, - "CookieNames": { - "base": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

", - "refs": { - "CookiePreference$WhitelistedNames": "

Required if you specify whitelist for the value of Forward:. A complex type that specifies how many different cookies you want CloudFront to forward to the origin for this cache behavior and, if you want to forward selected cookies, the names of those cookies.

If you specify all or none for the value of Forward, omit WhitelistedNames. If you change the value of Forward from whitelist to all or none and you don't delete the WhitelistedNames element and its child elements, CloudFront deletes them automatically.

For the current limit on the number of cookie names that you can whitelist for each cache behavior, see Amazon CloudFront Limits in the AWS General Reference.

" - } - }, - "CookiePreference": { - "base": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

", - "refs": { - "ForwardedValues$Cookies": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

" - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to create a new origin access identity.

", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "

The request to create a new distribution.

", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateDistributionWithTagsRequest": { - "base": "

The request to create a new distribution with tags.

", - "refs": { - } - }, - "CreateDistributionWithTagsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "

The request to create an invalidation.

", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "

The request to create a new streaming distribution.

", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsRequest": { - "base": "

The request to create a new streaming distribution with tags.

", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "

A complex type that controls:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "

A complex type that contains a CustomErrorResponse element for each HTTP status code for which you want to specify a custom error page and/or a caching duration.

" - } - }, - "CustomErrorResponses": { - "base": "

A complex type that controls:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "refs": { - "DistributionConfig$CustomErrorResponses": "

A complex type that controls the following:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "DistributionSummary$CustomErrorResponses": "

A complex type that contains zero or more CustomErrorResponses elements.

" - } - }, - "CustomHeaders": { - "base": "

A complex type that contains the list of Custom Headers for each origin.

", - "refs": { - "Origin$CustomHeaders": "

A complex type that contains names and values for the custom headers that you want.

" - } - }, - "CustomOriginConfig": { - "base": "

A customer origin.

", - "refs": { - "Origin$CustomOriginConfig": "

A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead.

" - } - }, - "DefaultCacheBehavior": { - "base": "

A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "

A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

", - "DistributionSummary$DefaultCacheBehavior": "

A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

" - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "

Deletes a origin access identity.

", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "

This action deletes a web distribution. To delete a web distribution using the CloudFront API, perform the following steps.

To delete a web distribution using the CloudFront API:

  1. Disable the web distribution

  2. Submit a GET Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Distribution Config request in Step 2.

  5. Review the response to the PUT Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Distribution request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Distribution Config request in Step 6.

  8. Review the response to your DELETE Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "

The request to delete a streaming distribution.

", - "refs": { - } - }, - "Distribution": { - "base": "

The distribution's information.

", - "refs": { - "CreateDistributionResult$Distribution": "

The distribution's information.

", - "CreateDistributionWithTagsResult$Distribution": "

The distribution's information.

", - "GetDistributionResult$Distribution": "

The distribution's information.

", - "UpdateDistributionResult$Distribution": "

The distribution's information.

" - } - }, - "DistributionAlreadyExists": { - "base": "

The caller reference you attempted to create the distribution with is associated with another distribution.

", - "refs": { - } - }, - "DistributionConfig": { - "base": "

A distribution configuration.

", - "refs": { - "CreateDistributionRequest$DistributionConfig": "

The distribution's configuration information.

", - "Distribution$DistributionConfig": "

The current configuration information for the distribution. Send a GET request to the /CloudFront API version/distribution ID/config resource.

", - "DistributionConfigWithTags$DistributionConfig": "

A distribution configuration.

", - "GetDistributionConfigResult$DistributionConfig": "

The distribution's configuration information.

", - "UpdateDistributionRequest$DistributionConfig": "

The distribution's configuration information.

" - } - }, - "DistributionConfigWithTags": { - "base": "

A distribution Configuration and a list of tags to be associated with the distribution.

", - "refs": { - "CreateDistributionWithTagsRequest$DistributionConfigWithTags": "

The distribution's configuration information.

" - } - }, - "DistributionList": { - "base": "

A distribution list.

", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "

The DistributionList type.

", - "ListDistributionsResult$DistributionList": "

The DistributionList type.

" - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "

A summary of the information about a CloudFront distribution.

", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "

A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account.

" - } - }, - "ForwardedValues": { - "base": "

A complex type that specifies how CloudFront handles query strings and cookies.

", - "refs": { - "CacheBehavior$ForwardedValues": "

A complex type that specifies how CloudFront handles query strings and cookies.

", - "DefaultCacheBehavior$ForwardedValues": "

A complex type that specifies how CloudFront handles query strings and cookies.

" - } - }, - "GeoRestriction": { - "base": "

A complex type that controls the countries in which your content is distributed. CloudFront determines the location of your users using MaxMind GeoIP databases.

", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "

The method that you want to use to restrict distribution of your content by country:

  • none: No geo restriction is enabled, meaning access to content is not restricted by client geo location.

  • blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content.

  • whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content.

" - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "

The origin access identity's configuration information. For more information, see CloudFrontOriginAccessIdentityConfigComplexType.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to get an origin access identity's information.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "

The request to get a distribution configuration.

", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "

The request to get a distribution's information.

", - "refs": { - } - }, - "GetDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "

The request to get an invalidation's information.

", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "

To request to get a streaming distribution configuration.

", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "

The request to get a streaming distribution's information.

", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "

A complex type that contains one Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items.

" - } - }, - "Headers": { - "base": "

A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior.

For the headers that you specify, CloudFront also caches separate versions of a specified object based on the header values in viewer requests. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to cache your content based on values in the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value. For more information about caching based on header values, see How CloudFront Forwards and Caches Headers in the Amazon CloudFront Developer Guide.

", - "refs": { - "ForwardedValues$Headers": "

A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior.

" - } - }, - "HttpVersion": { - "base": null, - "refs": { - "DistributionConfig$HttpVersion": "

(Optional) Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 automatically use an earlier HTTP version.

For viewers and CloudFront to use HTTP/2, viewers must support TLS 1.2 or later, and must support Server Name Identification (SNI).

In general, configuring CloudFront to communicate with viewers using HTTP/2 reduces latency. You can improve performance by optimizing for HTTP/2. For more information, do an Internet search for \"http/2 optimization.\"

", - "DistributionSummary$HttpVersion": "

Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 will automatically use an earlier version.

" - } - }, - "IllegalUpdate": { - "base": "

Origin and CallerReference cannot be updated.

", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "

The value of Quantity and the size of Items do not match.

", - "refs": { - } - }, - "InvalidArgument": { - "base": "

The argument is invalid.

", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "

The default root object file name is too big or contains an invalid character.

", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "

Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.

", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "

The If-Match version is missing or not valid for the distribution.

", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "

The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.

", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "

The origin access identity is not valid or doesn't exist.

", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "

You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that support Server Name Indication (SNI).

", - "refs": { - } - }, - "InvalidQueryStringParameters": { - "base": null, - "refs": { - } - }, - "InvalidRelativePath": { - "base": "

The relative path is too big, is not URL-encoded, or does not begin with a slash (/).

", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "

This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.

", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidTagging": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "

An invalidation.

", - "refs": { - "CreateInvalidationResult$Invalidation": "

The invalidation's information.

", - "GetInvalidationResult$Invalidation": "

The invalidation's information. For more information, see Invalidation Complex Type.

" - } - }, - "InvalidationBatch": { - "base": "

An invalidation batch.

", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "

The batch information for the invalidation.

", - "Invalidation$InvalidationBatch": "

The current invalidation information for the batch request.

" - } - }, - "InvalidationList": { - "base": "

The InvalidationList complex type describes the list of invalidation objects. For more information about invalidation, see Invalidating Objects (Web Distributions Only) in the Amazon CloudFront Developer Guide.

", - "refs": { - "ListInvalidationsResult$InvalidationList": "

Information about invalidation batches.

" - } - }, - "InvalidationSummary": { - "base": "

A summary of an invalidation request.

", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "

A complex type that contains one InvalidationSummary element for each invalidation batch created by the current AWS account.

" - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "

Specifies which cookies to forward to the origin for this cache behavior: all, none, or the list of cookies specified in the WhitelistedNames complex type.

Amazon S3 doesn't process cookies. When the cache behavior is forwarding requests to an Amazon S3 origin, specify none for the Forward element.

" - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

For more information, see ActiveTrustedSigners.

" - } - }, - "KeyPairIds": { - "base": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

For more information, see ActiveTrustedSigners.

", - "refs": { - "Signer$KeyPairIds": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

" - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "

The request to list origin access identities.

", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "

The request to list distributions that are associated with a specified AWS WAF web ACL.

", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "

The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.

", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "

The request to list your distributions.

", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "

The request to list invalidations.

", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "

The request to list your streaming distributions.

", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": "

The request to list tags for a CloudFront resource.

", - "refs": { - } - }, - "ListTagsForResourceResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "

A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist).

The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country.

CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes.

" - } - }, - "LoggingConfig": { - "base": "

A complex type that controls whether access logs are written for the distribution.

", - "refs": { - "DistributionConfig$Logging": "

A complex type that controls whether access logs are written for the distribution.

For more information about logging, see Access Logs in the Amazon CloudFront Developer Guide.

" - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "

A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.

", - "CachedMethods$Items": "

A complex type that contains the HTTP methods that you want CloudFront to cache responses to.

" - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "

Specify the minimum version of the SSL/TLS protocol that you want CloudFront to use for HTTPS connections between viewers and CloudFront: SSLv3 or TLSv1. CloudFront serves your objects only to viewers that support SSL/TLS version that you specify and later versions. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. Note the following:

  • If you specify <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>, the minimum SSL protocol version is TLSv1 and can't be changed.

  • If you're using a custom certificate (if you specify a value for ACMCertificateArn or for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion.

" - } - }, - "MissingBody": { - "base": "

This operation requires a body. Ensure that the body is present and the Content-Type header is set.

", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "

The specified origin access identity does not exist.

", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "

The specified distribution does not exist.

", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "

The specified invalidation does not exist.

", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "

No origin exists with the specified Origin Id.

", - "refs": { - } - }, - "NoSuchResource": { - "base": null, - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "

The specified streaming distribution does not exist.

", - "refs": { - } - }, - "Origin": { - "base": "

A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files. You must create at least one origin.

For the current limit on the number of origins that you can create for a distribution, see Amazon CloudFront Limits in the AWS General Reference.

", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "

A complex type that contains HeaderName and HeaderValue elements, if any, for this distribution.

", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "

Optional: A list that contains one OriginCustomHeader element for each custom header that you want CloudFront to forward to the origin. If Quantity is 0, omit Items.

" - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "

A complex type that contains origins for this distribution.

" - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "

The origin protocol policy to apply to your origin.

" - } - }, - "OriginSslProtocols": { - "base": "

A complex type that contains information about the SSL/TLS protocols that CloudFront can use when establishing an HTTPS connection with your origin.

", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "

The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.

" - } - }, - "Origins": { - "base": "

A complex type that contains information about origins for this distribution.

", - "refs": { - "DistributionConfig$Origins": "

A complex type that contains information about origins for this distribution.

", - "DistributionSummary$Origins": "

A complex type that contains information about origins for this distribution.

" - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "

A complex type that contains a list of the paths that you want to invalidate.

" - } - }, - "Paths": { - "base": "

A complex type that contains information about the objects that you want to invalidate. For more information, see Specifying the Objects to Invalidate in the Amazon CloudFront Developer Guide.

", - "refs": { - "InvalidationBatch$Paths": "

A complex type that contains information about the objects that you want to invalidate. For more information, see Specifying the Objects to Invalidate in the Amazon CloudFront Developer Guide.

" - } - }, - "PreconditionFailed": { - "base": "

The precondition given in one or more of the request-header fields evaluated to false.

", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "

The price class that corresponds with the maximum price that you want to pay for CloudFront service. If you specify PriceClass_All, CloudFront responds to requests for your objects from all CloudFront edge locations.

If you specify a price class other than PriceClass_All, CloudFront serves your objects from the CloudFront edge location that has the lowest latency among the edge locations in your price class. Viewers who are in or near regions that are excluded from your specified price class may encounter slower performance.

For more information about price classes, see Choosing the Price Class for a CloudFront Distribution in the Amazon CloudFront Developer Guide. For information about CloudFront pricing, including how price classes map to CloudFront regions, see Amazon CloudFront Pricing.

", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "

A complex type that contains information about price class for this streaming distribution.

", - "StreamingDistributionSummary$PriceClass": null - } - }, - "QueryStringCacheKeys": { - "base": null, - "refs": { - "ForwardedValues$QueryStringCacheKeys": "

A complex type that contains information about the query string parameters that you want CloudFront to use for caching for this cache behavior.

" - } - }, - "QueryStringCacheKeysList": { - "base": null, - "refs": { - "QueryStringCacheKeys$Items": "

(Optional) A list that contains the query string parameters that you want CloudFront to use as a basis for caching for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "ResourceARN": { - "base": null, - "refs": { - "ListTagsForResourceRequest$Resource": "

An ARN of a CloudFront resource.

", - "TagResourceRequest$Resource": "

An ARN of a CloudFront resource.

", - "UntagResourceRequest$Resource": "

An ARN of a CloudFront resource.

" - } - }, - "Restrictions": { - "base": "

A complex type that identifies ways in which you want to restrict distribution of your content.

", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

", - "refs": { - "StreamingDistributionConfig$S3Origin": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

", - "StreamingDistributionSummary$S3Origin": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

" - } - }, - "S3OriginConfig": { - "base": "

A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.

", - "refs": { - "Origin$S3OriginConfig": "

A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.

" - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "

If you specify a value for ACMCertificateArn or for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests: using a method that works for all clients or one that works for most clients:

  • vip: CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges.

  • sni-only: CloudFront can respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. If some of your users' browsers don't support SNI, we recommend that you do one of the following:

    • Use the vip option (dedicated IP addresses) instead of sni-only.

    • Use the CloudFront SSL/TLS certificate instead of a custom certificate. This requires that you use the CloudFront domain name of your distribution in the URLs for your objects, for example, https://d111111abcdef8.cloudfront.net/logo.png.

    • If you can control which browser your users use, upgrade the browser to one that supports SNI.

    • Use HTTP instead of HTTPS.

Do not specify a value for SSLSupportMethod if you specified <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>.

For more information, see Using Alternate Domain Names and HTTPS in the Amazon CloudFront Developer Guide.

" - } - }, - "Signer": { - "base": "

A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.

", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "

A complex type that contains one Signer complex type for each trusted signer that is specified in the TrustedSigners complex type.

For more information, see ActiveTrustedSigners.

" - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "

A list that contains allowed SSL/TLS protocols for this distribution.

" - } - }, - "StreamingDistribution": { - "base": "

A streaming distribution.

", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

", - "CreateStreamingDistributionWithTagsResult$StreamingDistribution": "

The streaming distribution's information.

", - "GetStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

", - "UpdateStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

" - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "

The RTMP distribution's configuration information.

", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "

The streaming distribution's configuration information.

", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "

The streaming distribution's configuration information.

", - "StreamingDistribution$StreamingDistributionConfig": "

The current configuration information for the RTMP distribution.

", - "StreamingDistributionConfigWithTags$StreamingDistributionConfig": "

A streaming distribution Configuration.

", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "

The streaming distribution's configuration information.

" - } - }, - "StreamingDistributionConfigWithTags": { - "base": "

A streaming distribution Configuration and a list of tags to be associated with the streaming distribution.

", - "refs": { - "CreateStreamingDistributionWithTagsRequest$StreamingDistributionConfigWithTags": "

The streaming distribution's configuration information.

" - } - }, - "StreamingDistributionList": { - "base": "

A streaming distribution list.

", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "

The StreamingDistributionList type.

" - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "

A summary of the information for an Amazon CloudFront streaming distribution.

", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "

A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account.

" - } - }, - "StreamingLoggingConfig": { - "base": "

A complex type that controls whether access logs are written for this streaming distribution.

", - "refs": { - "StreamingDistributionConfig$Logging": "

A complex type that controls whether access logs are written for the streaming distribution.

" - } - }, - "Tag": { - "base": "

A complex type that contains Tag key and Tag value.

", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": "

A string that contains Tag key.

The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

", - "refs": { - "Tag$Key": "

A string that contains Tag key.

The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

", - "TagKeyList$member": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "TagKeys$Items": "

A complex type that contains Tag key elements.

" - } - }, - "TagKeys": { - "base": "

A complex type that contains zero or more Tag elements.

", - "refs": { - "UntagResourceRequest$TagKeys": "

A complex type that contains zero or more Tag key elements.

" - } - }, - "TagList": { - "base": null, - "refs": { - "Tags$Items": "

A complex type that contains Tag elements.

" - } - }, - "TagResourceRequest": { - "base": "

The request to add tags to a CloudFront resource.

", - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

A string that contains an optional Tag value.

The string length should be between 0 and 256 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

" - } - }, - "Tags": { - "base": "

A complex type that contains zero or more Tag elements.

", - "refs": { - "DistributionConfigWithTags$Tags": "

A complex type that contains zero or more Tag elements.

", - "ListTagsForResourceResult$Tags": "

A complex type that contains zero or more Tag elements.

", - "StreamingDistributionConfigWithTags$Tags": "

A complex type that contains zero or more Tag elements.

", - "TagResourceRequest$Tags": "

A complex type that contains zero or more Tag elements.

" - } - }, - "TooManyCacheBehaviors": { - "base": "

You cannot create more cache behaviors for the distribution.

", - "refs": { - } - }, - "TooManyCertificates": { - "base": "

You cannot create anymore custom SSL/TLS certificates.

", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "

Processing your request would cause you to exceed the maximum number of origin access identities allowed.

", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "

Your request contains more cookie names in the whitelist than are allowed per cache behavior.

", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "

Your request contains more CNAMEs than are allowed per distribution.

", - "refs": { - } - }, - "TooManyDistributions": { - "base": "

Processing your request would cause you to exceed the maximum number of distributions allowed.

", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "

You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.

", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "

You cannot create more origins for the distribution.

", - "refs": { - } - }, - "TooManyQueryStringParameters": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "

Processing your request would cause you to exceed the maximum number of streaming distributions allowed.

", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "

Your request contains more trusted signers than are allowed per distribution.

", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "

One or more of your trusted signers do not exist.

", - "refs": { - } - }, - "TrustedSigners": { - "base": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

For more information about updating the distribution configuration, see DistributionConfig .

", - "refs": { - "CacheBehavior$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

", - "DefaultCacheBehavior$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

", - "StreamingDistributionConfig$TrustedSigners": "

A complex type that specifies any AWS accounts that you want to permit to create signed URLs for private content. If you want the distribution to use signed URLs, include this element; if you want the distribution to use public URLs, remove this element. For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "StreamingDistributionSummary$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items.If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

" - } - }, - "UntagResourceRequest": { - "base": "

The request to remove tags from a CloudFront resource.

", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to update an origin access identity.

", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "

The request to update a distribution.

", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "

The request to update a streaming distribution.

", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ViewerCertificate": { - "base": "

A complex type that specifies the following:

  • Which SSL/TLS certificate to use when viewers request objects using HTTPS

  • Whether you want CloudFront to use dedicated IP addresses or SNI when you're using alternate domain names in your object names

  • The minimum protocol version that you want CloudFront to use when communicating with viewers

For more information, see Using an HTTPS Connection to Access Your Objects in the Amazon Amazon CloudFront Developer Guide.

", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "

The protocol that viewers can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. You can specify the following options:

  • allow-all: Viewers can use HTTP or HTTPS.

  • redirect-to-https: If a viewer submits an HTTP request, CloudFront returns an HTTP status code of 301 (Moved Permanently) to the viewer along with the HTTPS URL. The viewer then resubmits the request using the new URL.

  • https-only: If a viewer sends an HTTP request, CloudFront returns an HTTP status code of 403 (Forbidden).

For more information about requiring the HTTPS protocol, see Using an HTTPS Connection to Access Your Objects in the Amazon CloudFront Developer Guide.

The only way to guarantee that viewers retrieve an object that was fetched from the origin using HTTPS is never to use any other protocol to fetch the object. If you have recently changed from HTTP to HTTPS, we recommend that you clear your objects' cache because cached objects are protocol agnostic. That means that an edge location will return an object from the cache regardless of whether the current request protocol matches the protocol used previously. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$ViewerProtocolPolicy": "

The protocol that viewers can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. You can specify the following options:

  • allow-all: Viewers can use HTTP or HTTPS.

  • redirect-to-https: If a viewer submits an HTTP request, CloudFront returns an HTTP status code of 301 (Moved Permanently) to the viewer along with the HTTPS URL. The viewer then resubmits the request using the new URL.

  • https-only: If a viewer sends an HTTP request, CloudFront returns an HTTP status code of 403 (Forbidden).

For more information about requiring the HTTPS protocol, see Using an HTTPS Connection to Access Your Objects in the Amazon CloudFront Developer Guide.

The only way to guarantee that viewers retrieve an object that was fetched from the origin using HTTPS is never to use any other protocol to fetch the object. If you have recently changed from HTTP to HTTPS, we recommend that you clear your objects' cache because cached objects are protocol agnostic. That means that an edge location will return an object from the cache regardless of whether the current request protocol matches the protocol used previously. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

" - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "

Enabled is true if any of the AWS accounts listed in the TrustedSigners complex type for this RTMP distribution have active CloudFront key pairs. If not, Enabled is false.

For more information, see ActiveTrustedSigners.

", - "CacheBehavior$SmoothStreaming": "

Indicates whether you want to distribute media files in the Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false. If you specify true for SmoothStreaming, you can still distribute other content using this cache behavior if the content matches the value of PathPattern.

", - "CacheBehavior$Compress": "

Whether you want CloudFront to automatically compress certain files for this cache behavior. If so, specify true; if not, specify false. For more information, see Serving Compressed Files in the Amazon CloudFront Developer Guide.

", - "CloudFrontOriginAccessIdentityList$IsTruncated": "

A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.

", - "DefaultCacheBehavior$SmoothStreaming": "

Indicates whether you want to distribute media files in the Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false. If you specify true for SmoothStreaming, you can still distribute other content using this cache behavior if the content matches the value of PathPattern.

", - "DefaultCacheBehavior$Compress": "

Whether you want CloudFront to automatically compress certain files for this cache behavior. If so, specify true; if not, specify false. For more information, see Serving Compressed Files in the Amazon CloudFront Developer Guide.

", - "DistributionConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket.

If you do not want to enable logging when you create a distribution, or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements.

If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.

", - "DistributionConfig$IsIPV6Enabled": "

If you want CloudFront to respond to IPv6 DNS requests with an IPv6 address for your distribution, specify true. If you specify false, CloudFront responds to IPv6 DNS requests with the DNS response code NOERROR and with no IP addresses. This allows viewers to submit a second request, for an IPv4 address for your distribution.

In general, you should enable IPv6 if you have users on IPv6 networks who want to access your content. However, if you're using signed URLs or signed cookies to restrict access to your content, and if you're using a custom policy that includes the IpAddress parameter to restrict the IP addresses that can access your content, do not enable IPv6. If you want to restrict access to some content by IP address and not restrict access to other content (or restrict access but not by IP address), you can create two distributions. For more information, see Creating a Signed URL Using a Custom Policy in the Amazon CloudFront Developer Guide.

If you're using an Amazon Route 53 alias resource record set to route traffic to your CloudFront distribution, you need to create a second alias resource record set when both of the following are true:

  • You enable IPv6 for the distribution

  • You're using alternate domain names in the URLs for your objects

For more information, see Routing Traffic to an Amazon CloudFront Web Distribution by Using Your Domain Name in the Amazon Route 53 Developer Guide.

If you created a CNAME resource record set, either with Amazon Route 53 or with another DNS service, you don't need to make any changes. A CNAME record will route traffic to your distribution regardless of the IP address format of the viewer request.

", - "DistributionList$IsTruncated": "

A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.

", - "DistributionSummary$Enabled": "

Whether the distribution is enabled to accept user requests for content.

", - "DistributionSummary$IsIPV6Enabled": "

Whether CloudFront responds to IPv6 DNS requests with an IPv6 address for your distribution.

", - "ForwardedValues$QueryString": "

Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior and cache based on the query string parameters. CloudFront behavior depends on the value of QueryString and on the values that you specify for QueryStringCacheKeys, if any:

If you specify true for QueryString and you don't specify any values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin and caches based on all query string parameters. Depending on how many query string parameters and values you have, this can adversely affect performance because CloudFront must forward more requests to the origin.

If you specify true for QueryString and you specify one or more values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin, but it only caches based on the query string parameters that you specify.

If you specify false for QueryString, CloudFront doesn't forward any query string parameters to the origin, and doesn't cache based on query string parameters.

For more information, see Configuring CloudFront to Cache Based on Query String Parameters in the Amazon CloudFront Developer Guide.

", - "InvalidationList$IsTruncated": "

A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.

", - "LoggingConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix, and IncludeCookies, the values are automatically deleted.

", - "LoggingConfig$IncludeCookies": "

Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.

", - "StreamingDistributionConfig$Enabled": "

Whether the streaming distribution is enabled to accept user requests for content.

", - "StreamingDistributionList$IsTruncated": "

A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.

", - "StreamingDistributionSummary$Enabled": "

Whether the distribution is enabled to accept end user requests for content.

", - "StreamingLoggingConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.

", - "TrustedSigners$Enabled": "

Specifies whether you want to require viewers to use signed URLs to access the files specified by PathPattern and TargetOriginId.

", - "ViewerCertificate$CloudFrontDefaultCertificate": "

If you want viewers to use HTTPS to request your objects and you're using the CloudFront domain name of your distribution in your object URLs (for example, https://d111111abcdef8.cloudfront.net/logo.jpg), set to true. Omit this value if you are setting an ACMCertificateArn or IAMCertificateId.

" - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "

A complex type that contains one Signer complex type for each trusted signer specified in the TrustedSigners complex type.

For more information, see ActiveTrustedSigners.

", - "Aliases$Quantity": "

The number of CNAME aliases, if any, that you want to associate with this distribution.

", - "AllowedMethods$Quantity": "

The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD, and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).

", - "CacheBehaviors$Quantity": "

The number of cache behaviors for this distribution.

", - "CachedMethods$Quantity": "

The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).

", - "CloudFrontOriginAccessIdentityList$MaxItems": "

The maximum number of origin access identities you want in the response body.

", - "CloudFrontOriginAccessIdentityList$Quantity": "

The number of CloudFront origin access identities that were created by the current AWS account.

", - "CookieNames$Quantity": "

The number of different cookies that you want CloudFront to forward to the origin for this cache behavior.

", - "CustomErrorResponse$ErrorCode": "

The HTTP status code for which you want to specify a custom error page and/or a caching duration.

", - "CustomErrorResponses$Quantity": "

The number of HTTP status codes for which you want to specify a custom error page and/or a caching duration. If Quantity is 0, you can omit Items.

", - "CustomHeaders$Quantity": "

The number of custom headers, if any, for this distribution.

", - "CustomOriginConfig$HTTPPort": "

The HTTP port the custom origin listens on.

", - "CustomOriginConfig$HTTPSPort": "

The HTTPS port the custom origin listens on.

", - "Distribution$InProgressInvalidationBatches": "

The number of invalidation batches currently in progress.

", - "DistributionList$MaxItems": "

The value you provided for the MaxItems request parameter.

", - "DistributionList$Quantity": "

The number of distributions that were created by the current AWS account.

", - "GeoRestriction$Quantity": "

When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.

", - "Headers$Quantity": "

The number of different headers that you want CloudFront to forward to the origin for this cache behavior. You can configure each cache behavior in a web distribution to do one of the following:

  • Forward all headers to your origin: Specify 1 for Quantity and * for Name.

    If you configure CloudFront to forward all headers to your origin, CloudFront doesn't cache the objects associated with this cache behavior. Instead, it sends every request to the origin.

  • Forward a whitelist of headers you specify: Specify the number of headers that you want to forward, and specify the header names in Name elements. CloudFront caches your objects based on the values in all of the specified headers. CloudFront also forwards the headers that it forwards by default, but it caches your objects based only on the headers that you specify.

  • Forward only the default headers: Specify 0 for Quantity and omit Items. In this configuration, CloudFront doesn't cache based on the values in the request headers.

", - "InvalidationList$MaxItems": "

The value that you provided for the MaxItems request parameter.

", - "InvalidationList$Quantity": "

The number of invalidation batches that were created by the current AWS account.

", - "KeyPairIds$Quantity": "

The number of active CloudFront key pairs for AwsAccountNumber.

For more information, see ActiveTrustedSigners.

", - "OriginSslProtocols$Quantity": "

The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.

", - "Origins$Quantity": "

The number of origins for this distribution.

", - "Paths$Quantity": "

The number of objects that you want to invalidate.

", - "QueryStringCacheKeys$Quantity": "

The number of whitelisted query string parameters for this cache behavior.

", - "StreamingDistributionList$MaxItems": "

The value you provided for the MaxItems request parameter.

", - "StreamingDistributionList$Quantity": "

The number of streaming distributions that were created by the current AWS account.

", - "TrustedSigners$Quantity": "

The number of trusted signers for this cache behavior.

" - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "

The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon Amazon CloudFront Developer Guide.

You must specify 0 for MinTTL if you configure CloudFront to forward all headers to your origin (under Headers, if you specify 1 for Quantity and * for Name).

", - "CacheBehavior$DefaultTTL": "

The default amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "CacheBehavior$MaxTTL": "

The maximum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "CustomErrorResponse$ErrorCachingMinTTL": "

The minimum amount of time, in seconds, that you want CloudFront to cache the HTTP status code specified in ErrorCode. When this time period has elapsed, CloudFront queries your origin to see whether the problem that caused the error has been resolved and the requested object is now available.

If you don't want to specify a value, include an empty element, <ErrorCachingMinTTL>, in the XML document.

For more information, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$MinTTL": "

The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon Amazon CloudFront Developer Guide.

You must specify 0 for MinTTL if you configure CloudFront to forward all headers to your origin (under Headers, if you specify 1 for Quantity and * for Name).

", - "DefaultCacheBehavior$DefaultTTL": "

The default amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$MaxTTL": null - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "

The pattern (for example, images/*.jpg) that specifies which requests to apply the behavior to. When CloudFront receives a viewer request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution.

You can optionally include a slash (/) at the beginning of the path pattern. For example, /images/*.jpg. CloudFront behavior is the same with or without the leading /.

The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.

For more information, see Path Pattern in the Amazon CloudFront Developer Guide.

", - "CacheBehavior$TargetOriginId": "

The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.

", - "CloudFrontOriginAccessIdentity$Id": "

The ID for the origin access identity. For example: E74FTE3AJFJ256A.

", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "

The Amazon S3 canonical user ID for the origin access identity, used when giving the origin access identity read permission to an object in Amazon S3.

", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "

A unique number that ensures the request can't be replayed.

If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created.

If the CallerReference is a value already sent in a previous identity request, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request.

If the CallerReference is a value you already sent in a previous request to create an identity, but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.

", - "CloudFrontOriginAccessIdentityConfig$Comment": "

Any comments you want to include about the origin access identity.

", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "

Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).

", - "CloudFrontOriginAccessIdentityList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.

", - "CloudFrontOriginAccessIdentitySummary$Id": "

The ID for the origin access identity. For example: E74FTE3AJFJ256A.

", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "

The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.

", - "CloudFrontOriginAccessIdentitySummary$Comment": "

The comment for this origin access identity, as originally specified when created.

", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "

The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.

", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the origin access identity created.

", - "CreateDistributionResult$Location": "

The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.

", - "CreateDistributionResult$ETag": "

The current version of the distribution created.

", - "CreateDistributionWithTagsResult$Location": "

The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.

", - "CreateDistributionWithTagsResult$ETag": "

The current version of the distribution created.

", - "CreateInvalidationRequest$DistributionId": "

The distribution's id.

", - "CreateInvalidationResult$Location": "

The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.

", - "CreateStreamingDistributionResult$Location": "

The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.

", - "CreateStreamingDistributionResult$ETag": "

The current version of the streaming distribution created.

", - "CreateStreamingDistributionWithTagsResult$Location": "

The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.

", - "CreateStreamingDistributionWithTagsResult$ETag": null, - "CustomErrorResponse$ResponsePagePath": "

The path to the custom error page that you want CloudFront to return to a viewer when your origin returns the HTTP status code specified by ErrorCode, for example, /4xx-errors/403-forbidden.html. If you want to store your objects and your custom error pages in different locations, your distribution must include a cache behavior for which the following is true:

  • The value of PathPattern matches the path to your custom error messages. For example, suppose you saved custom error pages for 4xx errors in an Amazon S3 bucket in a directory named /4xx-errors. Your distribution must include a cache behavior for which the path pattern routes requests for your custom error pages to that location, for example, /4xx-errors/*.

  • The value of TargetOriginId specifies the value of the ID element for the origin that contains your custom error pages.

If you specify a value for ResponsePagePath, you must also specify a value for ResponseCode. If you don't want to specify a value, include an empty element, <ResponsePagePath>, in the XML document.

We recommend that you store custom error pages in an Amazon S3 bucket. If you store custom error pages on an HTTP server and the server starts to return 5xx errors, CloudFront can't get the files that you want to return to viewers because the origin server is unavailable.

", - "CustomErrorResponse$ResponseCode": "

The HTTP status code that you want CloudFront to return to the viewer along with the custom error page. There are a variety of reasons that you might want CloudFront to return a status code different from the status code that your origin returned to CloudFront, for example:

  • Some Internet devices (some firewalls and corporate proxies, for example) intercept HTTP 4xx and 5xx and prevent the response from being returned to the viewer. If you substitute 200, the response typically won't be intercepted.

  • If you don't care about distinguishing among different client errors or server errors, you can specify 400 or 500 as the ResponseCode for all 4xx or 5xx errors.

  • You might want to return a 200 status code (OK) and static website so your customers don't know that your website is down.

If you specify a value for ResponseCode, you must also specify a value for ResponsePagePath. If you don't want to specify a value, include an empty element, <ResponseCode>, in the XML document.

", - "DefaultCacheBehavior$TargetOriginId": "

The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.

", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "

The origin access identity's ID.

", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "

The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.

", - "DeleteDistributionRequest$Id": "

The distribution ID.

", - "DeleteDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.

", - "DeleteStreamingDistributionRequest$Id": "

The distribution ID.

", - "DeleteStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.

", - "Distribution$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "Distribution$ARN": "

The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "Distribution$Status": "

This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated to all CloudFront edge locations.

", - "Distribution$DomainName": "

The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.

", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "

A unique value (for example, a date-time stamp) that ensures that the request can't be replayed.

If the value of CallerReference is new (regardless of the content of the DistributionConfig object), CloudFront creates a new distribution.

If CallerReference is a value you already sent in a previous request to create a distribution, and if the content of the DistributionConfig is identical to the original request (ignoring white space), CloudFront returns the same the response that it returned to the original request.

If CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.

", - "DistributionConfig$DefaultRootObject": "

The object that you want CloudFront to request from your origin (for example, index.html) when a viewer requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/product-description.html). Specifying a default root object avoids exposing the contents of your distribution.

Specify only the object name, for example, index.html. Do not add a / before the object name.

If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element.

To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element.

To replace the default root object, update the distribution configuration and specify the new object.

For more information about the default root object, see Creating a Default Root Object in the Amazon CloudFront Developer Guide.

", - "DistributionConfig$Comment": "

Any comments you want to include about the distribution.

If you don't want to specify a comment, include an empty Comment element.

To delete an existing comment, update the distribution configuration and include an empty Comment element.

To add or change a comment, update the distribution configuration and specify the new comment.

", - "DistributionConfig$WebACLId": "

A unique identifier that specifies the AWS WAF web ACL, if any, to associate with this distribution.

AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to CloudFront, and lets you control access to your content. Based on conditions that you specify, such as the IP addresses that requests originate from or the values of query strings, CloudFront responds to requests either with the requested content or with an HTTP 403 status code (Forbidden). You can also configure CloudFront to return a custom error page when a request is blocked. For more information about AWS WAF, see the AWS WAF Developer Guide.

", - "DistributionList$Marker": "

The value you provided for the Marker request parameter.

", - "DistributionList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.

", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "DistributionSummary$ARN": "

The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "DistributionSummary$Status": "

The current status of the distribution. When the status is Deployed, the distribution's information is propagated to all CloudFront edge locations.

", - "DistributionSummary$DomainName": "

The domain name that corresponds to the distribution. For example: d604721fxaaqy9.cloudfront.net.

", - "DistributionSummary$Comment": "

The comment originally specified when this distribution was created.

", - "DistributionSummary$WebACLId": "

The Web ACL Id (if any) associated with the distribution.

", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "

The identity's ID.

", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetCloudFrontOriginAccessIdentityRequest$Id": "

The identity's ID.

", - "GetCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.

", - "GetDistributionConfigRequest$Id": "

The distribution's ID.

", - "GetDistributionConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetDistributionRequest$Id": "

The distribution's ID.

", - "GetDistributionResult$ETag": "

The current version of the distribution's information. For example: E2QWRUHAPOMQZL.

", - "GetInvalidationRequest$DistributionId": "

The distribution's ID.

", - "GetInvalidationRequest$Id": "

The identifier for the invalidation request, for example, IDFDVBD632BHDS5.

", - "GetStreamingDistributionConfigRequest$Id": "

The streaming distribution's ID.

", - "GetStreamingDistributionConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetStreamingDistributionRequest$Id": "

The streaming distribution's ID.

", - "GetStreamingDistributionResult$ETag": "

The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.

", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidQueryStringParameters$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidTagging$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "

The identifier for the invalidation request. For example: IDFDVBD632BHDS5.

", - "Invalidation$Status": "

The status of the invalidation request. When the invalidation batch is finished, the status is Completed.

", - "InvalidationBatch$CallerReference": "

A value that you specify to uniquely identify an invalidation request. CloudFront uses the value to prevent you from accidentally resubmitting an identical request. Whenever you create a new invalidation request, you must specify a new value for CallerReference and change other values in the request as applicable. One way to ensure that the value of CallerReference is unique is to use a timestamp, for example, 20120301090000.

If you make a second invalidation request with the same value for CallerReference, and if the rest of the request is the same, CloudFront doesn't create a new invalidation request. Instead, CloudFront returns information about the invalidation request that you previously created with the same CallerReference.

If CallerReference is a value you already sent in a previous invalidation batch request but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.

", - "InvalidationList$Marker": "

The value that you provided for the Marker request parameter.

", - "InvalidationList$NextMarker": "

If IsTruncated is true, this element is present and contains the value that you can use for the Marker request parameter to continue listing your invalidation batches where they left off.

", - "InvalidationSummary$Id": "

The unique ID for an invalidation request.

", - "InvalidationSummary$Status": "

The status of an invalidation request.

", - "KeyPairIdList$member": null, - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "

Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).

", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "

The maximum number of origin access identities you want in the response body.

", - "ListDistributionsByWebACLIdRequest$Marker": "

Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)

", - "ListDistributionsByWebACLIdRequest$MaxItems": "

The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.

", - "ListDistributionsByWebACLIdRequest$WebACLId": "

The ID of the AWS WAF web ACL that you want to list the associated distributions. If you specify \"null\" for the ID, the request returns a list of the distributions that aren't associated with a web ACL.

", - "ListDistributionsRequest$Marker": "

Use this when paginating results to indicate where to begin in your list of distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).

", - "ListDistributionsRequest$MaxItems": "

The maximum number of distributions you want in the response body.

", - "ListInvalidationsRequest$DistributionId": "

The distribution's ID.

", - "ListInvalidationsRequest$Marker": "

Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.

", - "ListInvalidationsRequest$MaxItems": "

The maximum number of invalidation batches that you want in the response body.

", - "ListStreamingDistributionsRequest$Marker": "

The value that you provided for the Marker request parameter.

", - "ListStreamingDistributionsRequest$MaxItems": "

The value that you provided for the MaxItems request parameter.

", - "LocationList$member": null, - "LoggingConfig$Bucket": "

The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.

", - "LoggingConfig$Prefix": "

An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.

", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchResource$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "

A unique identifier for the origin. The value of Id must be unique within the distribution.

When you specify the value of TargetOriginId for the default cache behavior or for another cache behavior, you indicate the origin to which you want the cache behavior to route requests by specifying the value of the Id element for that origin. When a request matches the path pattern for that cache behavior, CloudFront routes the request to the specified origin. For more information, see Cache Behavior Settings in the Amazon CloudFront Developer Guide.

", - "Origin$DomainName": "

Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com.

Constraints for Amazon S3 origins:

  • If you configured Amazon S3 Transfer Acceleration for your bucket, do not specify the s3-accelerate endpoint for DomainName.

  • The bucket name must be between 3 and 63 characters long (inclusive).

  • The bucket name must contain only lowercase characters, numbers, periods, underscores, and dashes.

  • The bucket name must not contain adjacent periods.

Custom Origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.

Constraints for custom origins:

  • DomainName must be a valid DNS name that contains only a-z, A-Z, 0-9, dot (.), hyphen (-), or underscore (_) characters.

  • The name cannot exceed 128 characters.

", - "Origin$OriginPath": "

An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName, for example, example.com/production. Do not include a / at the end of the directory name.

For example, suppose you've specified the following values for your distribution:

  • DomainName: An Amazon S3 bucket named myawsbucket.

  • OriginPath: /production

  • CNAME: example.com

When a user enters example.com/index.html in a browser, CloudFront sends a request to Amazon S3 for myawsbucket/production/index.html.

When a user enters example.com/acme/index.html in a browser, CloudFront sends a request to Amazon S3 for myawsbucket/production/acme/index.html.

", - "OriginCustomHeader$HeaderName": "

The name of a header that you want CloudFront to forward to your origin. For more information, see Forwarding Custom Headers to Your Origin (Web Distributions Only) in the Amazon Amazon CloudFront Developer Guide.

", - "OriginCustomHeader$HeaderValue": "

The value for the header that you specified in the HeaderName field.

", - "PathList$member": null, - "PreconditionFailed$Message": null, - "QueryStringCacheKeysList$member": null, - "S3Origin$DomainName": "

The DNS name of the Amazon S3 origin.

", - "S3Origin$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the RTMP distribution. Use an origin access identity to configure the distribution so that end users can only access objects in an Amazon S3 bucket through CloudFront.

If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information, see Using an Origin Access Identity to Restrict Access to Your Amazon S3 Content in the Amazon Amazon CloudFront Developer Guide.

", - "S3OriginConfig$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that viewers can only access objects in an Amazon S3 bucket through CloudFront. The format of the value is:

origin-access-identity/CloudFront/ID-of-origin-access-identity

where ID-of-origin-access-identity is the value that CloudFront returned in the ID element when you created the origin access identity.

If you want viewers to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information about the origin access identity, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "Signer$AwsAccountNumber": "

An AWS account that is included in the TrustedSigners complex type for this RTMP distribution. Valid values include:

  • self, which is the AWS account used to create the distribution.

  • An AWS account number.

", - "StreamingDistribution$Id": "

The identifier for the RTMP distribution. For example: EGTXBD79EXAMPLE.

", - "StreamingDistribution$ARN": null, - "StreamingDistribution$Status": "

The current status of the RTMP distribution. When the status is Deployed, the distribution's information is propagated to all CloudFront edge locations.

", - "StreamingDistribution$DomainName": "

The domain name that corresponds to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.

", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "

A unique number that ensures that the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value that you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value that you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.

", - "StreamingDistributionConfig$Comment": "

Any comments you want to include about the streaming distribution.

", - "StreamingDistributionList$Marker": "

The value you provided for the Marker request parameter.

", - "StreamingDistributionList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your RTMP distributions where they left off.

", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "StreamingDistributionSummary$ARN": "

The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "StreamingDistributionSummary$Status": "

Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.

", - "StreamingDistributionSummary$DomainName": "

The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.

", - "StreamingDistributionSummary$Comment": "

The comment originally specified when this distribution was created.

", - "StreamingLoggingConfig$Bucket": "

The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.

", - "StreamingLoggingConfig$Prefix": "

An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.

", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyQueryStringParameters$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "

The identity's id.

", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "

The value of the ETag header that you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateDistributionRequest$Id": "

The distribution's id.

", - "UpdateDistributionRequest$IfMatch": "

The value of the ETag header that you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateDistributionResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateStreamingDistributionRequest$Id": "

The streaming distribution's id.

", - "UpdateStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateStreamingDistributionResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "ViewerCertificate$IAMCertificateId": "

If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the IAM certificate identifier of the custom viewer certificate for this distribution. Specify either this value, ACMCertificateArn, or CloudFrontDefaultCertificate.

", - "ViewerCertificate$ACMCertificateArn": "

If you want viewers to use HTTPS to request your objects and you're using an alternate domain name in your object URLs (for example, https://example.com/logo.jpg), specify the ACM certificate ARN of the custom viewer certificate for this distribution. Specify either this value, IAMCertificateId, or CloudFrontDefaultCertificate.

", - "ViewerCertificate$Certificate": "

Include one of these values to specify the following:

  • Whether you want viewers to use HTTP or HTTPS to request your objects.

  • If you want viewers to use HTTPS, whether you're using an alternate domain name such as example.com or the CloudFront domain name for your distribution, such as d111111abcdef8.cloudfront.net.

  • If you're using an alternate domain name, whether AWS Certificate Manager (ACM) provided the certificate, or you purchased a certificate from a third-party certificate authority and imported it into ACM or uploaded it to the IAM certificate store.

You must specify one (and only one) of the three values. Do not specify false for CloudFrontDefaultCertificate.

If you want viewers to use HTTP to request your objects: Specify the following value:

<CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>

In addition, specify allow-all for ViewerProtocolPolicy for all of your cache behaviors.

If you want viewers to use HTTPS to request your objects: Choose the type of certificate that you want to use based on whether you're using an alternate domain name for your objects or the CloudFront domain name:

  • If you're using an alternate domain name, such as example.com: Specify one of the following values, depending on whether ACM provided your certificate or you purchased your certificate from third-party certificate authority:

    • <ACMCertificateArn>ARN for ACM SSL/TLS certificate<ACMCertificateArn> where ARN for ACM SSL/TLS certificate is the ARN for the ACM SSL/TLS certificate that you want to use for this distribution.

    • <IAMCertificateId>IAM certificate ID<IAMCertificateId> where IAM certificate ID is the ID that IAM returned when you added the certificate to the IAM certificate store.

    If you specify ACMCertificateArn or IAMCertificateId, you must also specify a value for SSLSupportMethod.

    If you choose to use an ACM certificate or a certificate in the IAM certificate store, we recommend that you use only an alternate domain name in your object URLs (https://example.com/logo.jpg). If you use the domain name that is associated with your CloudFront distribution (https://d111111abcdef8.cloudfront.net/logo.jpg) and the viewer supports SNI, then CloudFront behaves normally. However, if the browser does not support SNI, the user's experience depends on the value that you choose for SSLSupportMethod:

    • vip: The viewer displays a warning because there is a mismatch between the CloudFront domain name and the domain name in your SSL/TLS certificate.

    • sni-only: CloudFront drops the connection with the browser without returning the object.

  • If you're using the CloudFront domain name for your distribution, such as d111111abcdef8.cloudfront.net : Specify the following value:

    <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>

    If you want viewers to use HTTPS, you must also specify one of the following values in your cache behaviors:

    • <ViewerProtocolPolicy>https-only<ViewerProtocolPolicy>

    • <ViewerProtocolPolicy>redirect-to-https<ViewerProtocolPolicy>

    You can also optionally require that CloudFront use HTTPS to communicate with your origin by specifying one of the following values for the applicable origins:

    • <OriginProtocolPolicy>https-only<OriginProtocolPolicy>

    • <OriginProtocolPolicy>match-viewer<OriginProtocolPolicy>

    For more information, see Using Alternate Domain Names and HTTPS in the Amazon CloudFront Developer Guide.

" - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "

The date and time the distribution was last modified.

", - "DistributionSummary$LastModifiedTime": "

The date and time the distribution was last modified.

", - "Invalidation$CreateTime": "

The date and time the invalidation request was first made.

", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "

The date and time that the distribution was last modified.

", - "StreamingDistributionSummary$LastModifiedTime": "

The date and time the distribution was last modified.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/waiters-2.json deleted file mode 100644 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-09-29/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/api-2.json deleted file mode 100644 index 3c644ead1..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/api-2.json +++ /dev/null @@ -1,2665 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2016-11-25", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4", - "uid":"cloudfront-2016-11-25" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"}, - {"shape":"TooManyDistributionsWithLambdaAssociations"}, - {"shape":"TooManyLambdaFunctionAssociations"}, - {"shape":"InvalidLambdaFunctionAssociation"} - ] - }, - "CreateDistributionWithTags":{ - "name":"CreateDistributionWithTags2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionWithTagsRequest"}, - "output":{"shape":"CreateDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"InvalidTagging"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"}, - {"shape":"TooManyDistributionsWithLambdaAssociations"}, - {"shape":"TooManyLambdaFunctionAssociations"}, - {"shape":"InvalidLambdaFunctionAssociation"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistributionWithTags":{ - "name":"CreateStreamingDistributionWithTags2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/streaming-distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionWithTagsRequest"}, - "output":{"shape":"CreateStreamingDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"}, - {"shape":"InvalidTagging"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2016_11_25", - "http":{ - "method":"DELETE", - "requestUri":"/2016-11-25/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2016_11_25", - "http":{ - "method":"DELETE", - "requestUri":"/2016-11-25/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2016_11_25", - "http":{ - "method":"DELETE", - "requestUri":"/2016-11-25/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource2016_11_25", - "http":{ - "method":"GET", - "requestUri":"/2016-11-25/tagging" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "TagResource":{ - "name":"TagResource2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/tagging?Operation=Tag", - "responseCode":204 - }, - "input":{"shape":"TagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UntagResource":{ - "name":"UntagResource2016_11_25", - "http":{ - "method":"POST", - "requestUri":"/2016-11-25/tagging?Operation=Untag", - "responseCode":204 - }, - "input":{"shape":"UntagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2016_11_25", - "http":{ - "method":"PUT", - "requestUri":"/2016-11-25/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2016_11_25", - "http":{ - "method":"PUT", - "requestUri":"/2016-11-25/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"}, - {"shape":"TooManyDistributionsWithLambdaAssociations"}, - {"shape":"TooManyLambdaFunctionAssociations"}, - {"shape":"InvalidLambdaFunctionAssociation"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2016_11_25", - "http":{ - "method":"PUT", - "requestUri":"/2016-11-25/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"}, - "LambdaFunctionAssociations":{"shape":"LambdaFunctionAssociations"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateDistributionWithTagsRequest":{ - "type":"structure", - "required":["DistributionConfigWithTags"], - "members":{ - "DistributionConfigWithTags":{ - "shape":"DistributionConfigWithTags", - "locationName":"DistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"DistributionConfigWithTags" - }, - "CreateDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CreateStreamingDistributionWithTagsRequest":{ - "type":"structure", - "required":["StreamingDistributionConfigWithTags"], - "members":{ - "StreamingDistributionConfigWithTags":{ - "shape":"StreamingDistributionConfigWithTags", - "locationName":"StreamingDistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"StreamingDistributionConfigWithTags" - }, - "CreateStreamingDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"}, - "LambdaFunctionAssociations":{"shape":"LambdaFunctionAssociations"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"}, - "IsIPV6Enabled":{"shape":"boolean"} - } - }, - "DistributionConfigWithTags":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Tags" - ], - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId", - "HttpVersion", - "IsIPV6Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"}, - "IsIPV6Enabled":{"shape":"boolean"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "EventType":{ - "type":"string", - "enum":[ - "viewer-request", - "viewer-response", - "origin-request", - "origin-response" - ] - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"}, - "QueryStringCacheKeys":{"shape":"QueryStringCacheKeys"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "HttpVersion":{ - "type":"string", - "enum":[ - "http1.1", - "http2" - ] - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLambdaFunctionAssociation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTagging":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "LambdaFunctionAssociation":{ - "type":"structure", - "members":{ - "LambdaFunctionARN":{"shape":"string"}, - "EventType":{"shape":"EventType"} - } - }, - "LambdaFunctionAssociationList":{ - "type":"list", - "member":{ - "shape":"LambdaFunctionAssociation", - "locationName":"LambdaFunctionAssociation" - } - }, - "LambdaFunctionAssociations":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LambdaFunctionAssociationList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["Resource"], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - } - } - }, - "ListTagsForResourceResult":{ - "type":"structure", - "required":["Tags"], - "members":{ - "Tags":{"shape":"Tags"} - }, - "payload":"Tags" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchResource":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "QueryStringCacheKeys":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"QueryStringCacheKeysList"} - } - }, - "QueryStringCacheKeysList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "ResourceARN":{ - "type":"string", - "pattern":"arn:aws:cloudfront::[0-9]+:.*" - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionConfigWithTags":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Tags" - ], - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Tag":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{ - "type":"string", - "max":128, - "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "TagKeyList":{ - "type":"list", - "member":{ - "shape":"TagKey", - "locationName":"Key" - } - }, - "TagKeys":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagKeyList"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "Tags" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "Tags":{ - "shape":"Tags", - "locationName":"Tags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"Tags" - }, - "TagValue":{ - "type":"string", - "max":256, - "min":0, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "Tags":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagList"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionsWithLambdaAssociations":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyLambdaFunctionAssociations":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UntagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "TagKeys" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "TagKeys":{ - "shape":"TagKeys", - "locationName":"TagKeys", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - } - }, - "payload":"TagKeys" - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2016-11-25/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/docs-2.json deleted file mode 100644 index de255b727..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/docs-2.json +++ /dev/null @@ -1,1435 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon CloudFront

This is the Amazon CloudFront API Reference. This guide is for developers who need detailed information about the CloudFront API actions, data types, and errors. For detailed information about CloudFront features and their associated API calls, see the Amazon CloudFront Developer Guide.

", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "

Creates a new origin access identity. If you're using Amazon S3 for your origin, you can use an origin access identity to require users to access your content using a CloudFront URL instead of the Amazon S3 URL. For more information about how to use origin access identities, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "CreateDistribution": "

Creates a new web distribution. Send a GET request to the /CloudFront API version/distribution/distribution ID resource.

", - "CreateDistributionWithTags": "

Create a new distribution with tags.

", - "CreateInvalidation": "

Create a new invalidation.

", - "CreateStreamingDistribution": "

Creates a new RMTP distribution. An RTMP distribution is similar to a web distribution, but an RTMP distribution streams media files using the Adobe Real-Time Messaging Protocol (RTMP) instead of serving files using HTTP.

To create a new web distribution, submit a POST request to the CloudFront API version/distribution resource. The request body must include a document with a StreamingDistributionConfig element. The response echoes the StreamingDistributionConfig element and returns other information about the RTMP distribution.

To get the status of your request, use the GET StreamingDistribution API action. When the value of Enabled is true and the value of Status is Deployed, your distribution is ready. A distribution usually deploys in less than 15 minutes.

For more information about web distributions, see Working with RTMP Distributions in the Amazon CloudFront Developer Guide.

Beginning with the 2012-05-05 version of the CloudFront API, we made substantial changes to the format of the XML document that you include in the request body when you create or update a web distribution or an RTMP distribution, and when you invalidate objects. With previous versions of the API, we discovered that it was too easy to accidentally delete one or more values for an element that accepts multiple values, for example, CNAMEs and trusted signers. Our changes for the 2012-05-05 release are intended to prevent these accidental deletions and to notify you when there's a mismatch between the number of values you say you're specifying in the Quantity element and the number of values specified.

", - "CreateStreamingDistributionWithTags": "

Create a new streaming distribution with tags.

", - "DeleteCloudFrontOriginAccessIdentity": "

Delete an origin access identity.

", - "DeleteDistribution": "

Delete a distribution.

", - "DeleteStreamingDistribution": "

Delete a streaming distribution. To delete an RTMP distribution using the CloudFront API, perform the following steps.

To delete an RTMP distribution using the CloudFront API:

  1. Disable the RTMP distribution.

  2. Submit a GET Streaming Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Streaming Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Streaming Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Then set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  5. Review the response to the PUT Streaming Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Streaming Distribution Config request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Streaming Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  8. Review the response to your DELETE Streaming Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", - "GetCloudFrontOriginAccessIdentity": "

Get the information about an origin access identity.

", - "GetCloudFrontOriginAccessIdentityConfig": "

Get the configuration information about an origin access identity.

", - "GetDistribution": "

Get the information about a distribution.

", - "GetDistributionConfig": "

Get the configuration information about a distribution.

", - "GetInvalidation": "

Get the information about an invalidation.

", - "GetStreamingDistribution": "

Gets information about a specified RTMP distribution, including the distribution configuration.

", - "GetStreamingDistributionConfig": "

Get the configuration information about a streaming distribution.

", - "ListCloudFrontOriginAccessIdentities": "

Lists origin access identities.

", - "ListDistributions": "

List distributions.

", - "ListDistributionsByWebACLId": "

List the distributions that are associated with a specified AWS WAF web ACL.

", - "ListInvalidations": "

Lists invalidation batches.

", - "ListStreamingDistributions": "

List streaming distributions.

", - "ListTagsForResource": "

List tags for a CloudFront resource.

", - "TagResource": "

Add tags to a CloudFront resource.

", - "UntagResource": "

Remove tags from a CloudFront resource.

", - "UpdateCloudFrontOriginAccessIdentity": "

Update an origin access identity.

", - "UpdateDistribution": "

Update a distribution.

", - "UpdateStreamingDistribution": "

Update a streaming distribution.

" - }, - "shapes": { - "AccessDenied": { - "base": "

Access denied.

", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "

A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for this distribution. These are the accounts that you want to allow to create signed URLs for private content.

The Signer complex type lists the AWS account number of the trusted signer or self if the signer is the AWS account that created the distribution. The Signer element also includes the IDs of any active CloudFront key pairs that are associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create signed URLs.

For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "refs": { - "Distribution$ActiveTrustedSigners": "

CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.

", - "StreamingDistribution$ActiveTrustedSigners": "

A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for this distribution. These are the accounts that you want to allow to create signed URLs for private content.

The Signer complex type lists the AWS account number of the trusted signer or self if the signer is the AWS account that created the distribution. The Signer element also includes the IDs of any active CloudFront key pairs that are associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create signed URLs.

For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

" - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "

A complex type that contains the CNAME aliases, if any, that you want to associate with this distribution.

" - } - }, - "Aliases": { - "base": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "refs": { - "DistributionConfig$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "DistributionSummary$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "StreamingDistributionConfig$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.

", - "StreamingDistributionSummary$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.

" - } - }, - "AllowedMethods": { - "base": "

A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices:

  • CloudFront forwards only GET and HEAD requests.

  • CloudFront forwards only GET, HEAD, and OPTIONS requests.

  • CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests.

If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you might not want users to have permissions to delete objects from your origin.

", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "

Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "

A complex type that describes how CloudFront processes requests.

You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin is never used.

For the current limit on the number of cache behaviors that you can add to a distribution, see Amazon CloudFront Limits in the AWS General Reference.

If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error.

To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element.

To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.

For more information about cache behaviors, see Cache Behaviors in the Amazon CloudFront Developer Guide.

", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "

Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items.

" - } - }, - "CacheBehaviors": { - "base": "

A complex type that contains zero or more CacheBehavior elements.

", - "refs": { - "DistributionConfig$CacheBehaviors": "

A complex type that contains zero or more CacheBehavior elements.

", - "DistributionSummary$CacheBehaviors": "

A complex type that contains zero or more CacheBehavior elements.

" - } - }, - "CachedMethods": { - "base": "

A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices:

  • CloudFront caches responses to GET and HEAD requests.

  • CloudFront caches responses to GET, HEAD, and OPTIONS requests.

If you pick the second choice for your Amazon S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers, and Origin headers for the responses to be cached correctly.

", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "

This field is deprecated. You can use one of the following: [ACMCertificateArn, IAMCertificateId, or CloudFrontDefaultCertificate].

" - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "

CloudFront origin access identity.

", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

" - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "

If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.

", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "

Origin access identity configuration. Send a GET request to the /CloudFront API version/CloudFront/identity ID/config resource.

", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "

The current configuration information for the identity.

", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "

The current configuration information for the identity.

", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "

The origin access identity's configuration information.

", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "

The identity's configuration information.

" - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "

Lists the origin access identities for CloudFront.Send a GET request to the /CloudFront API version/origin-access-identity/cloudfront resource. The response includes a CloudFrontOriginAccessIdentityList element with zero or more CloudFrontOriginAccessIdentitySummary child elements. By default, your entire list of origin access identities is returned in one single page. If the list is long, you can paginate it using the MaxItems and Marker parameters.

", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "

The CloudFrontOriginAccessIdentityList type.

" - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "

Summary of the information about a CloudFront origin access identity.

", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "

A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account.

" - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "

A complex type that contains one Name element for each cookie that you want CloudFront to forward to the origin for this cache behavior.

" - } - }, - "CookieNames": { - "base": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

", - "refs": { - "CookiePreference$WhitelistedNames": "

Required if you specify whitelist for the value of Forward:. A complex type that specifies how many different cookies you want CloudFront to forward to the origin for this cache behavior and, if you want to forward selected cookies, the names of those cookies.

If you specify all or none for the value of Forward, omit WhitelistedNames. If you change the value of Forward from whitelist to all or none and you don't delete the WhitelistedNames element and its child elements, CloudFront deletes them automatically.

For the current limit on the number of cookie names that you can whitelist for each cache behavior, see Amazon CloudFront Limits in the AWS General Reference.

" - } - }, - "CookiePreference": { - "base": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

", - "refs": { - "ForwardedValues$Cookies": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

" - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to create a new origin access identity.

", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "

The request to create a new distribution.

", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateDistributionWithTagsRequest": { - "base": "

The request to create a new distribution with tags.

", - "refs": { - } - }, - "CreateDistributionWithTagsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "

The request to create an invalidation.

", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "

The request to create a new streaming distribution.

", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsRequest": { - "base": "

The request to create a new streaming distribution with tags.

", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "

A complex type that controls:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "

A complex type that contains a CustomErrorResponse element for each HTTP status code for which you want to specify a custom error page and/or a caching duration.

" - } - }, - "CustomErrorResponses": { - "base": "

A complex type that controls:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "refs": { - "DistributionConfig$CustomErrorResponses": "

A complex type that controls the following:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "DistributionSummary$CustomErrorResponses": "

A complex type that contains zero or more CustomErrorResponses elements.

" - } - }, - "CustomHeaders": { - "base": "

A complex type that contains the list of Custom Headers for each origin.

", - "refs": { - "Origin$CustomHeaders": "

A complex type that contains names and values for the custom headers that you want.

" - } - }, - "CustomOriginConfig": { - "base": "

A customer origin.

", - "refs": { - "Origin$CustomOriginConfig": "

A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead.

" - } - }, - "DefaultCacheBehavior": { - "base": "

A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "

A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

", - "DistributionSummary$DefaultCacheBehavior": "

A complex type that describes the default cache behavior if you do not specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

" - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "

Deletes a origin access identity.

", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "

This action deletes a web distribution. To delete a web distribution using the CloudFront API, perform the following steps.

To delete a web distribution using the CloudFront API:

  1. Disable the web distribution

  2. Submit a GET Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Distribution Config request in Step 2.

  5. Review the response to the PUT Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Distribution request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Distribution Config request in Step 6.

  8. Review the response to your DELETE Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "

The request to delete a streaming distribution.

", - "refs": { - } - }, - "Distribution": { - "base": "

The distribution's information.

", - "refs": { - "CreateDistributionResult$Distribution": "

The distribution's information.

", - "CreateDistributionWithTagsResult$Distribution": "

The distribution's information.

", - "GetDistributionResult$Distribution": "

The distribution's information.

", - "UpdateDistributionResult$Distribution": "

The distribution's information.

" - } - }, - "DistributionAlreadyExists": { - "base": "

The caller reference you attempted to create the distribution with is associated with another distribution.

", - "refs": { - } - }, - "DistributionConfig": { - "base": "

A distribution configuration.

", - "refs": { - "CreateDistributionRequest$DistributionConfig": "

The distribution's configuration information.

", - "Distribution$DistributionConfig": "

The current configuration information for the distribution. Send a GET request to the /CloudFront API version/distribution ID/config resource.

", - "DistributionConfigWithTags$DistributionConfig": "

A distribution configuration.

", - "GetDistributionConfigResult$DistributionConfig": "

The distribution's configuration information.

", - "UpdateDistributionRequest$DistributionConfig": "

The distribution's configuration information.

" - } - }, - "DistributionConfigWithTags": { - "base": "

A distribution Configuration and a list of tags to be associated with the distribution.

", - "refs": { - "CreateDistributionWithTagsRequest$DistributionConfigWithTags": "

The distribution's configuration information.

" - } - }, - "DistributionList": { - "base": "

A distribution list.

", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "

The DistributionList type.

", - "ListDistributionsResult$DistributionList": "

The DistributionList type.

" - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "

A summary of the information about a CloudFront distribution.

", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "

A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account.

" - } - }, - "EventType": { - "base": null, - "refs": { - "LambdaFunctionAssociation$EventType": "

Specifies the event type that triggers a Lambda function invocation. Valid values are:

  • viewer-request

  • origin-request

  • viewer-response

  • origin-response

" - } - }, - "ForwardedValues": { - "base": "

A complex type that specifies how CloudFront handles query strings and cookies.

", - "refs": { - "CacheBehavior$ForwardedValues": "

A complex type that specifies how CloudFront handles query strings and cookies.

", - "DefaultCacheBehavior$ForwardedValues": "

A complex type that specifies how CloudFront handles query strings and cookies.

" - } - }, - "GeoRestriction": { - "base": "

A complex type that controls the countries in which your content is distributed. CloudFront determines the location of your users using MaxMind GeoIP databases.

", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "

The method that you want to use to restrict distribution of your content by country:

  • none: No geo restriction is enabled, meaning access to content is not restricted by client geo location.

  • blacklist: The Location elements specify the countries in which you do not want CloudFront to distribute your content.

  • whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content.

" - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "

The origin access identity's configuration information. For more information, see CloudFrontOriginAccessIdentityConfigComplexType.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to get an origin access identity's information.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "

The request to get a distribution configuration.

", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "

The request to get a distribution's information.

", - "refs": { - } - }, - "GetDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "

The request to get an invalidation's information.

", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "

To request to get a streaming distribution configuration.

", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "

The request to get a streaming distribution's information.

", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "

A complex type that contains one Name element for each header that you want CloudFront to forward to the origin and to vary on for this cache behavior. If Quantity is 0, omit Items.

" - } - }, - "Headers": { - "base": "

A complex type that specifies the headers that you want CloudFront to forward to the origin for this cache behavior.

For the headers that you specify, CloudFront also caches separate versions of a specified object based on the header values in viewer requests. For example, suppose viewer requests for logo.jpg contain a custom Product header that has a value of either Acme or Apex, and you configure CloudFront to cache your content based on values in the Product header. CloudFront forwards the Product header to the origin and caches the response from the origin once for each header value. For more information about caching based on header values, see How CloudFront Forwards and Caches Headers in the Amazon CloudFront Developer Guide.

", - "refs": { - "ForwardedValues$Headers": "

A complex type that specifies the Headers, if any, that you want CloudFront to vary upon for this cache behavior.

" - } - }, - "HttpVersion": { - "base": null, - "refs": { - "DistributionConfig$HttpVersion": "

(Optional) Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 automatically use an earlier HTTP version.

For viewers and CloudFront to use HTTP/2, viewers must support TLS 1.2 or later, and must support Server Name Identification (SNI).

In general, configuring CloudFront to communicate with viewers using HTTP/2 reduces latency. You can improve performance by optimizing for HTTP/2. For more information, do an Internet search for \"http/2 optimization.\"

", - "DistributionSummary$HttpVersion": "

Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 will automatically use an earlier version.

" - } - }, - "IllegalUpdate": { - "base": "

Origin and CallerReference cannot be updated.

", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "

The value of Quantity and the size of Items do not match.

", - "refs": { - } - }, - "InvalidArgument": { - "base": "

The argument is invalid.

", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "

The default root object file name is too big or contains an invalid character.

", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "

Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.

", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "

The If-Match version is missing or not valid for the distribution.

", - "refs": { - } - }, - "InvalidLambdaFunctionAssociation": { - "base": "

The specified Lambda function association is invalid.

", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "

The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.

", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "

The origin access identity is not valid or doesn't exist.

", - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "

You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that support Server Name Indication (SNI).

", - "refs": { - } - }, - "InvalidQueryStringParameters": { - "base": null, - "refs": { - } - }, - "InvalidRelativePath": { - "base": "

The relative path is too big, is not URL-encoded, or does not begin with a slash (/).

", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "

This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.

", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidTagging": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "

An invalidation.

", - "refs": { - "CreateInvalidationResult$Invalidation": "

The invalidation's information.

", - "GetInvalidationResult$Invalidation": "

The invalidation's information. For more information, see Invalidation Complex Type.

" - } - }, - "InvalidationBatch": { - "base": "

An invalidation batch.

", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "

The batch information for the invalidation.

", - "Invalidation$InvalidationBatch": "

The current invalidation information for the batch request.

" - } - }, - "InvalidationList": { - "base": "

The InvalidationList complex type describes the list of invalidation objects. For more information about invalidation, see Invalidating Objects (Web Distributions Only) in the Amazon CloudFront Developer Guide.

", - "refs": { - "ListInvalidationsResult$InvalidationList": "

Information about invalidation batches.

" - } - }, - "InvalidationSummary": { - "base": "

A summary of an invalidation request.

", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "

A complex type that contains one InvalidationSummary element for each invalidation batch created by the current AWS account.

" - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "

Specifies which cookies to forward to the origin for this cache behavior: all, none, or the list of cookies specified in the WhitelistedNames complex type.

Amazon S3 doesn't process cookies. When the cache behavior is forwarding requests to an Amazon S3 origin, specify none for the Forward element.

" - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

For more information, see ActiveTrustedSigners.

" - } - }, - "KeyPairIds": { - "base": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

For more information, see ActiveTrustedSigners.

", - "refs": { - "Signer$KeyPairIds": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

" - } - }, - "LambdaFunctionAssociation": { - "base": "

A complex type that contains a Lambda function association.

", - "refs": { - "LambdaFunctionAssociationList$member": null - } - }, - "LambdaFunctionAssociationList": { - "base": null, - "refs": { - "LambdaFunctionAssociations$Items": "

Optional: A complex type that contains LambdaFunctionAssociation items for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "LambdaFunctionAssociations": { - "base": "

A complex type that specifies a list of Lambda functions associations for a cache behavior.

If you want to invoke one or more Lambda functions triggered by requests that match the PathPattern of the cache behavior, specify the applicable values for Quantity and Items. Note that there can be up to 4 LambdaFunctionAssociation items in this list (one for each possible value of EventType) and each EventType can be associated with the Lambda function only once.

If you don't want to invoke any Lambda functions for the requests that match PathPattern, specify 0 for Quantity and omit Items.

", - "refs": { - "CacheBehavior$LambdaFunctionAssociations": "

A complex type that contains zero or more Lambda function associations for a cache behavior.

", - "DefaultCacheBehavior$LambdaFunctionAssociations": "

A complex type that contains zero or more Lambda function associations for a cache behavior.

" - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "

The request to list origin access identities.

", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "

The request to list distributions that are associated with a specified AWS WAF web ACL.

", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "

The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.

", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "

The request to list your distributions.

", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "

The request to list invalidations.

", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "

The request to list your streaming distributions.

", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": "

The request to list tags for a CloudFront resource.

", - "refs": { - } - }, - "ListTagsForResourceResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "

A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist).

The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country.

CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list in the CloudFront console, which includes both country names and codes.

" - } - }, - "LoggingConfig": { - "base": "

A complex type that controls whether access logs are written for the distribution.

", - "refs": { - "DistributionConfig$Logging": "

A complex type that controls whether access logs are written for the distribution.

For more information about logging, see Access Logs in the Amazon CloudFront Developer Guide.

" - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "

A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.

", - "CachedMethods$Items": "

A complex type that contains the HTTP methods that you want CloudFront to cache responses to.

" - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "

Specify the minimum version of the SSL/TLS protocol that you want CloudFront to use for HTTPS connections between viewers and CloudFront: SSLv3 or TLSv1. CloudFront serves your objects only to viewers that support SSL/TLS version that you specify and later versions. The TLSv1 protocol is more secure, so we recommend that you specify SSLv3 only if your users are using browsers or devices that don't support TLSv1. Note the following:

  • If you specify <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>, the minimum SSL protocol version is TLSv1 and can't be changed.

  • If you're using a custom certificate (if you specify a value for ACMCertificateArn or for IAMCertificateId) and if you're using SNI (if you specify sni-only for SSLSupportMethod), you must specify TLSv1 for MinimumProtocolVersion.

" - } - }, - "MissingBody": { - "base": "

This operation requires a body. Ensure that the body is present and the Content-Type header is set.

", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "

The specified origin access identity does not exist.

", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "

The specified distribution does not exist.

", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "

The specified invalidation does not exist.

", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "

No origin exists with the specified Origin Id.

", - "refs": { - } - }, - "NoSuchResource": { - "base": null, - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "

The specified streaming distribution does not exist.

", - "refs": { - } - }, - "Origin": { - "base": "

A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files. You must create at least one origin.

For the current limit on the number of origins that you can create for a distribution, see Amazon CloudFront Limits in the AWS General Reference.

", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "

A complex type that contains HeaderName and HeaderValue elements, if any, for this distribution.

", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "

Optional: A list that contains one OriginCustomHeader element for each custom header that you want CloudFront to forward to the origin. If Quantity is 0, omit Items.

" - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "

A complex type that contains origins for this distribution.

" - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "

The origin protocol policy to apply to your origin.

" - } - }, - "OriginSslProtocols": { - "base": "

A complex type that contains information about the SSL/TLS protocols that CloudFront can use when establishing an HTTPS connection with your origin.

", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "

The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.

" - } - }, - "Origins": { - "base": "

A complex type that contains information about origins for this distribution.

", - "refs": { - "DistributionConfig$Origins": "

A complex type that contains information about origins for this distribution.

", - "DistributionSummary$Origins": "

A complex type that contains information about origins for this distribution.

" - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "

A complex type that contains a list of the paths that you want to invalidate.

" - } - }, - "Paths": { - "base": "

A complex type that contains information about the objects that you want to invalidate. For more information, see Specifying the Objects to Invalidate in the Amazon CloudFront Developer Guide.

", - "refs": { - "InvalidationBatch$Paths": "

A complex type that contains information about the objects that you want to invalidate. For more information, see Specifying the Objects to Invalidate in the Amazon CloudFront Developer Guide.

" - } - }, - "PreconditionFailed": { - "base": "

The precondition given in one or more of the request-header fields evaluated to false.

", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "

The price class that corresponds with the maximum price that you want to pay for CloudFront service. If you specify PriceClass_All, CloudFront responds to requests for your objects from all CloudFront edge locations.

If you specify a price class other than PriceClass_All, CloudFront serves your objects from the CloudFront edge location that has the lowest latency among the edge locations in your price class. Viewers who are in or near regions that are excluded from your specified price class may encounter slower performance.

For more information about price classes, see Choosing the Price Class for a CloudFront Distribution in the Amazon CloudFront Developer Guide. For information about CloudFront pricing, including how price classes map to CloudFront regions, see Amazon CloudFront Pricing.

", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "

A complex type that contains information about price class for this streaming distribution.

", - "StreamingDistributionSummary$PriceClass": null - } - }, - "QueryStringCacheKeys": { - "base": null, - "refs": { - "ForwardedValues$QueryStringCacheKeys": "

A complex type that contains information about the query string parameters that you want CloudFront to use for caching for this cache behavior.

" - } - }, - "QueryStringCacheKeysList": { - "base": null, - "refs": { - "QueryStringCacheKeys$Items": "

(Optional) A list that contains the query string parameters that you want CloudFront to use as a basis for caching for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "ResourceARN": { - "base": null, - "refs": { - "ListTagsForResourceRequest$Resource": "

An ARN of a CloudFront resource.

", - "TagResourceRequest$Resource": "

An ARN of a CloudFront resource.

", - "UntagResourceRequest$Resource": "

An ARN of a CloudFront resource.

" - } - }, - "Restrictions": { - "base": "

A complex type that identifies ways in which you want to restrict distribution of your content.

", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

", - "refs": { - "StreamingDistributionConfig$S3Origin": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

", - "StreamingDistributionSummary$S3Origin": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

" - } - }, - "S3OriginConfig": { - "base": "

A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.

", - "refs": { - "Origin$S3OriginConfig": "

A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.

" - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "

If you specify a value for ACMCertificateArn or for IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests: using a method that works for all clients or one that works for most clients:

  • vip: CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you must request permission to use this feature, and you incur additional monthly charges.

  • sni-only: CloudFront can respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. If some of your users' browsers don't support SNI, we recommend that you do one of the following:

    • Use the vip option (dedicated IP addresses) instead of sni-only.

    • Use the CloudFront SSL/TLS certificate instead of a custom certificate. This requires that you use the CloudFront domain name of your distribution in the URLs for your objects, for example, https://d111111abcdef8.cloudfront.net/logo.png.

    • If you can control which browser your users use, upgrade the browser to one that supports SNI.

    • Use HTTP instead of HTTPS.

Do not specify a value for SSLSupportMethod if you specified <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>.

For more information, see Using Alternate Domain Names and HTTPS in the Amazon CloudFront Developer Guide.

" - } - }, - "Signer": { - "base": "

A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.

", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "

A complex type that contains one Signer complex type for each trusted signer that is specified in the TrustedSigners complex type.

For more information, see ActiveTrustedSigners.

" - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "

A list that contains allowed SSL/TLS protocols for this distribution.

" - } - }, - "StreamingDistribution": { - "base": "

A streaming distribution.

", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

", - "CreateStreamingDistributionWithTagsResult$StreamingDistribution": "

The streaming distribution's information.

", - "GetStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

", - "UpdateStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

" - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "

The RTMP distribution's configuration information.

", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "

The streaming distribution's configuration information.

", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "

The streaming distribution's configuration information.

", - "StreamingDistribution$StreamingDistributionConfig": "

The current configuration information for the RTMP distribution.

", - "StreamingDistributionConfigWithTags$StreamingDistributionConfig": "

A streaming distribution Configuration.

", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "

The streaming distribution's configuration information.

" - } - }, - "StreamingDistributionConfigWithTags": { - "base": "

A streaming distribution Configuration and a list of tags to be associated with the streaming distribution.

", - "refs": { - "CreateStreamingDistributionWithTagsRequest$StreamingDistributionConfigWithTags": "

The streaming distribution's configuration information.

" - } - }, - "StreamingDistributionList": { - "base": "

A streaming distribution list.

", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "

The StreamingDistributionList type.

" - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "

A summary of the information for an Amazon CloudFront streaming distribution.

", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "

A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account.

" - } - }, - "StreamingLoggingConfig": { - "base": "

A complex type that controls whether access logs are written for this streaming distribution.

", - "refs": { - "StreamingDistributionConfig$Logging": "

A complex type that controls whether access logs are written for the streaming distribution.

" - } - }, - "Tag": { - "base": "

A complex type that contains Tag key and Tag value.

", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": "

A string that contains Tag key.

The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

", - "refs": { - "Tag$Key": "

A string that contains Tag key.

The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

", - "TagKeyList$member": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "TagKeys$Items": "

A complex type that contains Tag key elements.

" - } - }, - "TagKeys": { - "base": "

A complex type that contains zero or more Tag elements.

", - "refs": { - "UntagResourceRequest$TagKeys": "

A complex type that contains zero or more Tag key elements.

" - } - }, - "TagList": { - "base": null, - "refs": { - "Tags$Items": "

A complex type that contains Tag elements.

" - } - }, - "TagResourceRequest": { - "base": "

The request to add tags to a CloudFront resource.

", - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

A string that contains an optional Tag value.

The string length should be between 0 and 256 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

" - } - }, - "Tags": { - "base": "

A complex type that contains zero or more Tag elements.

", - "refs": { - "DistributionConfigWithTags$Tags": "

A complex type that contains zero or more Tag elements.

", - "ListTagsForResourceResult$Tags": "

A complex type that contains zero or more Tag elements.

", - "StreamingDistributionConfigWithTags$Tags": "

A complex type that contains zero or more Tag elements.

", - "TagResourceRequest$Tags": "

A complex type that contains zero or more Tag elements.

" - } - }, - "TooManyCacheBehaviors": { - "base": "

You cannot create more cache behaviors for the distribution.

", - "refs": { - } - }, - "TooManyCertificates": { - "base": "

You cannot create anymore custom SSL/TLS certificates.

", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "

Processing your request would cause you to exceed the maximum number of origin access identities allowed.

", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "

Your request contains more cookie names in the whitelist than are allowed per cache behavior.

", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "

Your request contains more CNAMEs than are allowed per distribution.

", - "refs": { - } - }, - "TooManyDistributions": { - "base": "

Processing your request would cause you to exceed the maximum number of distributions allowed.

", - "refs": { - } - }, - "TooManyDistributionsWithLambdaAssociations": { - "base": "

Processing your request would cause the maximum number of distributions with Lambda function associations per owner to be exceeded.

", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "

You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.

", - "refs": { - } - }, - "TooManyLambdaFunctionAssociations": { - "base": "

Your request contains more Lambda function associations than are allowed per distribution.

", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "

You cannot create more origins for the distribution.

", - "refs": { - } - }, - "TooManyQueryStringParameters": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "

Processing your request would cause you to exceed the maximum number of streaming distributions allowed.

", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "

Your request contains more trusted signers than are allowed per distribution.

", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "

One or more of your trusted signers do not exist.

", - "refs": { - } - }, - "TrustedSigners": { - "base": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

For more information about updating the distribution configuration, see DistributionConfig .

", - "refs": { - "CacheBehavior$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

", - "DefaultCacheBehavior$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

", - "StreamingDistributionConfig$TrustedSigners": "

A complex type that specifies any AWS accounts that you want to permit to create signed URLs for private content. If you want the distribution to use signed URLs, include this element; if you want the distribution to use public URLs, remove this element. For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "StreamingDistributionSummary$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items.If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

" - } - }, - "UntagResourceRequest": { - "base": "

The request to remove tags from a CloudFront resource.

", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to update an origin access identity.

", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "

The request to update a distribution.

", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "

The request to update a streaming distribution.

", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ViewerCertificate": { - "base": "

A complex type that specifies the following:

  • Which SSL/TLS certificate to use when viewers request objects using HTTPS

  • Whether you want CloudFront to use dedicated IP addresses or SNI when you're using alternate domain names in your object names

  • The minimum protocol version that you want CloudFront to use when communicating with viewers

For more information, see Using an HTTPS Connection to Access Your Objects in the Amazon Amazon CloudFront Developer Guide.

", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "

The protocol that viewers can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. You can specify the following options:

  • allow-all: Viewers can use HTTP or HTTPS.

  • redirect-to-https: If a viewer submits an HTTP request, CloudFront returns an HTTP status code of 301 (Moved Permanently) to the viewer along with the HTTPS URL. The viewer then resubmits the request using the new URL.

  • https-only: If a viewer sends an HTTP request, CloudFront returns an HTTP status code of 403 (Forbidden).

For more information about requiring the HTTPS protocol, see Using an HTTPS Connection to Access Your Objects in the Amazon CloudFront Developer Guide.

The only way to guarantee that viewers retrieve an object that was fetched from the origin using HTTPS is never to use any other protocol to fetch the object. If you have recently changed from HTTP to HTTPS, we recommend that you clear your objects' cache because cached objects are protocol agnostic. That means that an edge location will return an object from the cache regardless of whether the current request protocol matches the protocol used previously. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$ViewerProtocolPolicy": "

The protocol that viewers can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. You can specify the following options:

  • allow-all: Viewers can use HTTP or HTTPS.

  • redirect-to-https: If a viewer submits an HTTP request, CloudFront returns an HTTP status code of 301 (Moved Permanently) to the viewer along with the HTTPS URL. The viewer then resubmits the request using the new URL.

  • https-only: If a viewer sends an HTTP request, CloudFront returns an HTTP status code of 403 (Forbidden).

For more information about requiring the HTTPS protocol, see Using an HTTPS Connection to Access Your Objects in the Amazon CloudFront Developer Guide.

The only way to guarantee that viewers retrieve an object that was fetched from the origin using HTTPS is never to use any other protocol to fetch the object. If you have recently changed from HTTP to HTTPS, we recommend that you clear your objects' cache because cached objects are protocol agnostic. That means that an edge location will return an object from the cache regardless of whether the current request protocol matches the protocol used previously. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

" - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "

Enabled is true if any of the AWS accounts listed in the TrustedSigners complex type for this RTMP distribution have active CloudFront key pairs. If not, Enabled is false.

For more information, see ActiveTrustedSigners.

", - "CacheBehavior$SmoothStreaming": "

Indicates whether you want to distribute media files in the Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false. If you specify true for SmoothStreaming, you can still distribute other content using this cache behavior if the content matches the value of PathPattern.

", - "CacheBehavior$Compress": "

Whether you want CloudFront to automatically compress certain files for this cache behavior. If so, specify true; if not, specify false. For more information, see Serving Compressed Files in the Amazon CloudFront Developer Guide.

", - "CloudFrontOriginAccessIdentityList$IsTruncated": "

A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.

", - "DefaultCacheBehavior$SmoothStreaming": "

Indicates whether you want to distribute media files in the Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false. If you specify true for SmoothStreaming, you can still distribute other content using this cache behavior if the content matches the value of PathPattern.

", - "DefaultCacheBehavior$Compress": "

Whether you want CloudFront to automatically compress certain files for this cache behavior. If so, specify true; if not, specify false. For more information, see Serving Compressed Files in the Amazon CloudFront Developer Guide.

", - "DistributionConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket.

If you do not want to enable logging when you create a distribution, or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements.

If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.

", - "DistributionConfig$IsIPV6Enabled": "

If you want CloudFront to respond to IPv6 DNS requests with an IPv6 address for your distribution, specify true. If you specify false, CloudFront responds to IPv6 DNS requests with the DNS response code NOERROR and with no IP addresses. This allows viewers to submit a second request, for an IPv4 address for your distribution.

In general, you should enable IPv6 if you have users on IPv6 networks who want to access your content. However, if you're using signed URLs or signed cookies to restrict access to your content, and if you're using a custom policy that includes the IpAddress parameter to restrict the IP addresses that can access your content, do not enable IPv6. If you want to restrict access to some content by IP address and not restrict access to other content (or restrict access but not by IP address), you can create two distributions. For more information, see Creating a Signed URL Using a Custom Policy in the Amazon CloudFront Developer Guide.

If you're using an Amazon Route 53 alias resource record set to route traffic to your CloudFront distribution, you need to create a second alias resource record set when both of the following are true:

  • You enable IPv6 for the distribution

  • You're using alternate domain names in the URLs for your objects

For more information, see Routing Traffic to an Amazon CloudFront Web Distribution by Using Your Domain Name in the Amazon Route 53 Developer Guide.

If you created a CNAME resource record set, either with Amazon Route 53 or with another DNS service, you don't need to make any changes. A CNAME record will route traffic to your distribution regardless of the IP address format of the viewer request.

", - "DistributionList$IsTruncated": "

A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.

", - "DistributionSummary$Enabled": "

Whether the distribution is enabled to accept user requests for content.

", - "DistributionSummary$IsIPV6Enabled": "

Whether CloudFront responds to IPv6 DNS requests with an IPv6 address for your distribution.

", - "ForwardedValues$QueryString": "

Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior and cache based on the query string parameters. CloudFront behavior depends on the value of QueryString and on the values that you specify for QueryStringCacheKeys, if any:

If you specify true for QueryString and you don't specify any values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin and caches based on all query string parameters. Depending on how many query string parameters and values you have, this can adversely affect performance because CloudFront must forward more requests to the origin.

If you specify true for QueryString and you specify one or more values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin, but it only caches based on the query string parameters that you specify.

If you specify false for QueryString, CloudFront doesn't forward any query string parameters to the origin, and doesn't cache based on query string parameters.

For more information, see Configuring CloudFront to Cache Based on Query String Parameters in the Amazon CloudFront Developer Guide.

", - "InvalidationList$IsTruncated": "

A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.

", - "LoggingConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix, and IncludeCookies, the values are automatically deleted.

", - "LoggingConfig$IncludeCookies": "

Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you do not want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.

", - "StreamingDistributionConfig$Enabled": "

Whether the streaming distribution is enabled to accept user requests for content.

", - "StreamingDistributionList$IsTruncated": "

A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.

", - "StreamingDistributionSummary$Enabled": "

Whether the distribution is enabled to accept end user requests for content.

", - "StreamingLoggingConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you do not want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.

", - "TrustedSigners$Enabled": "

Specifies whether you want to require viewers to use signed URLs to access the files specified by PathPattern and TargetOriginId.

", - "ViewerCertificate$CloudFrontDefaultCertificate": null - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "

A complex type that contains one Signer complex type for each trusted signer specified in the TrustedSigners complex type.

For more information, see ActiveTrustedSigners.

", - "Aliases$Quantity": "

The number of CNAME aliases, if any, that you want to associate with this distribution.

", - "AllowedMethods$Quantity": "

The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD, and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).

", - "CacheBehaviors$Quantity": "

The number of cache behaviors for this distribution.

", - "CachedMethods$Quantity": "

The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).

", - "CloudFrontOriginAccessIdentityList$MaxItems": "

The maximum number of origin access identities you want in the response body.

", - "CloudFrontOriginAccessIdentityList$Quantity": "

The number of CloudFront origin access identities that were created by the current AWS account.

", - "CookieNames$Quantity": "

The number of different cookies that you want CloudFront to forward to the origin for this cache behavior.

", - "CustomErrorResponse$ErrorCode": "

The HTTP status code for which you want to specify a custom error page and/or a caching duration.

", - "CustomErrorResponses$Quantity": "

The number of HTTP status codes for which you want to specify a custom error page and/or a caching duration. If Quantity is 0, you can omit Items.

", - "CustomHeaders$Quantity": "

The number of custom headers, if any, for this distribution.

", - "CustomOriginConfig$HTTPPort": "

The HTTP port the custom origin listens on.

", - "CustomOriginConfig$HTTPSPort": "

The HTTPS port the custom origin listens on.

", - "Distribution$InProgressInvalidationBatches": "

The number of invalidation batches currently in progress.

", - "DistributionList$MaxItems": "

The value you provided for the MaxItems request parameter.

", - "DistributionList$Quantity": "

The number of distributions that were created by the current AWS account.

", - "GeoRestriction$Quantity": "

When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.

", - "Headers$Quantity": "

The number of different headers that you want CloudFront to forward to the origin for this cache behavior. You can configure each cache behavior in a web distribution to do one of the following:

  • Forward all headers to your origin: Specify 1 for Quantity and * for Name.

    If you configure CloudFront to forward all headers to your origin, CloudFront doesn't cache the objects associated with this cache behavior. Instead, it sends every request to the origin.

  • Forward a whitelist of headers you specify: Specify the number of headers that you want to forward, and specify the header names in Name elements. CloudFront caches your objects based on the values in all of the specified headers. CloudFront also forwards the headers that it forwards by default, but it caches your objects based only on the headers that you specify.

  • Forward only the default headers: Specify 0 for Quantity and omit Items. In this configuration, CloudFront doesn't cache based on the values in the request headers.

", - "InvalidationList$MaxItems": "

The value that you provided for the MaxItems request parameter.

", - "InvalidationList$Quantity": "

The number of invalidation batches that were created by the current AWS account.

", - "KeyPairIds$Quantity": "

The number of active CloudFront key pairs for AwsAccountNumber.

For more information, see ActiveTrustedSigners.

", - "LambdaFunctionAssociations$Quantity": "

The number of Lambda function associations for this cache behavior.

", - "OriginSslProtocols$Quantity": "

The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.

", - "Origins$Quantity": "

The number of origins for this distribution.

", - "Paths$Quantity": "

The number of objects that you want to invalidate.

", - "QueryStringCacheKeys$Quantity": "

The number of whitelisted query string parameters for this cache behavior.

", - "StreamingDistributionList$MaxItems": "

The value you provided for the MaxItems request parameter.

", - "StreamingDistributionList$Quantity": "

The number of streaming distributions that were created by the current AWS account.

", - "TrustedSigners$Quantity": "

The number of trusted signers for this cache behavior.

" - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "

The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon Amazon CloudFront Developer Guide.

You must specify 0 for MinTTL if you configure CloudFront to forward all headers to your origin (under Headers, if you specify 1 for Quantity and * for Name).

", - "CacheBehavior$DefaultTTL": "

The default amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "CacheBehavior$MaxTTL": "

The maximum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "CustomErrorResponse$ErrorCachingMinTTL": "

The minimum amount of time, in seconds, that you want CloudFront to cache the HTTP status code specified in ErrorCode. When this time period has elapsed, CloudFront queries your origin to see whether the problem that caused the error has been resolved and the requested object is now available.

If you don't want to specify a value, include an empty element, <ErrorCachingMinTTL>, in the XML document.

For more information, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$MinTTL": "

The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon Amazon CloudFront Developer Guide.

You must specify 0 for MinTTL if you configure CloudFront to forward all headers to your origin (under Headers, if you specify 1 for Quantity and * for Name).

", - "DefaultCacheBehavior$DefaultTTL": "

The default amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$MaxTTL": null - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "

The pattern (for example, images/*.jpg) that specifies which requests to apply the behavior to. When CloudFront receives a viewer request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution.

You can optionally include a slash (/) at the beginning of the path pattern. For example, /images/*.jpg. CloudFront behavior is the same with or without the leading /.

The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.

For more information, see Path Pattern in the Amazon CloudFront Developer Guide.

", - "CacheBehavior$TargetOriginId": "

The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.

", - "CloudFrontOriginAccessIdentity$Id": "

The ID for the origin access identity. For example: E74FTE3AJFJ256A.

", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "

The Amazon S3 canonical user ID for the origin access identity, used when giving the origin access identity read permission to an object in Amazon S3.

", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "

A unique number that ensures the request can't be replayed.

If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created.

If the CallerReference is a value already sent in a previous identity request, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request.

If the CallerReference is a value you already sent in a previous request to create an identity, but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.

", - "CloudFrontOriginAccessIdentityConfig$Comment": "

Any comments you want to include about the origin access identity.

", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "

Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).

", - "CloudFrontOriginAccessIdentityList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.

", - "CloudFrontOriginAccessIdentitySummary$Id": "

The ID for the origin access identity. For example: E74FTE3AJFJ256A.

", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "

The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.

", - "CloudFrontOriginAccessIdentitySummary$Comment": "

The comment for this origin access identity, as originally specified when created.

", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "

The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.

", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the origin access identity created.

", - "CreateDistributionResult$Location": "

The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.

", - "CreateDistributionResult$ETag": "

The current version of the distribution created.

", - "CreateDistributionWithTagsResult$Location": "

The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.

", - "CreateDistributionWithTagsResult$ETag": "

The current version of the distribution created.

", - "CreateInvalidationRequest$DistributionId": "

The distribution's id.

", - "CreateInvalidationResult$Location": "

The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.

", - "CreateStreamingDistributionResult$Location": "

The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.

", - "CreateStreamingDistributionResult$ETag": "

The current version of the streaming distribution created.

", - "CreateStreamingDistributionWithTagsResult$Location": "

The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.

", - "CreateStreamingDistributionWithTagsResult$ETag": null, - "CustomErrorResponse$ResponsePagePath": "

The path to the custom error page that you want CloudFront to return to a viewer when your origin returns the HTTP status code specified by ErrorCode, for example, /4xx-errors/403-forbidden.html. If you want to store your objects and your custom error pages in different locations, your distribution must include a cache behavior for which the following is true:

  • The value of PathPattern matches the path to your custom error messages. For example, suppose you saved custom error pages for 4xx errors in an Amazon S3 bucket in a directory named /4xx-errors. Your distribution must include a cache behavior for which the path pattern routes requests for your custom error pages to that location, for example, /4xx-errors/*.

  • The value of TargetOriginId specifies the value of the ID element for the origin that contains your custom error pages.

If you specify a value for ResponsePagePath, you must also specify a value for ResponseCode. If you don't want to specify a value, include an empty element, <ResponsePagePath>, in the XML document.

We recommend that you store custom error pages in an Amazon S3 bucket. If you store custom error pages on an HTTP server and the server starts to return 5xx errors, CloudFront can't get the files that you want to return to viewers because the origin server is unavailable.

", - "CustomErrorResponse$ResponseCode": "

The HTTP status code that you want CloudFront to return to the viewer along with the custom error page. There are a variety of reasons that you might want CloudFront to return a status code different from the status code that your origin returned to CloudFront, for example:

  • Some Internet devices (some firewalls and corporate proxies, for example) intercept HTTP 4xx and 5xx and prevent the response from being returned to the viewer. If you substitute 200, the response typically won't be intercepted.

  • If you don't care about distinguishing among different client errors or server errors, you can specify 400 or 500 as the ResponseCode for all 4xx or 5xx errors.

  • You might want to return a 200 status code (OK) and static website so your customers don't know that your website is down.

If you specify a value for ResponseCode, you must also specify a value for ResponsePagePath. If you don't want to specify a value, include an empty element, <ResponseCode>, in the XML document.

", - "DefaultCacheBehavior$TargetOriginId": "

The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.

", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "

The origin access identity's ID.

", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "

The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.

", - "DeleteDistributionRequest$Id": "

The distribution ID.

", - "DeleteDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.

", - "DeleteStreamingDistributionRequest$Id": "

The distribution ID.

", - "DeleteStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.

", - "Distribution$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "Distribution$ARN": "

The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "Distribution$Status": "

This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated to all CloudFront edge locations.

", - "Distribution$DomainName": "

The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.

", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "

A unique value (for example, a date-time stamp) that ensures that the request can't be replayed.

If the value of CallerReference is new (regardless of the content of the DistributionConfig object), CloudFront creates a new distribution.

If CallerReference is a value you already sent in a previous request to create a distribution, and if the content of the DistributionConfig is identical to the original request (ignoring white space), CloudFront returns the same the response that it returned to the original request.

If CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.

", - "DistributionConfig$DefaultRootObject": "

The object that you want CloudFront to request from your origin (for example, index.html) when a viewer requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/product-description.html). Specifying a default root object avoids exposing the contents of your distribution.

Specify only the object name, for example, index.html. Do not add a / before the object name.

If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element.

To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element.

To replace the default root object, update the distribution configuration and specify the new object.

For more information about the default root object, see Creating a Default Root Object in the Amazon CloudFront Developer Guide.

", - "DistributionConfig$Comment": "

Any comments you want to include about the distribution.

If you don't want to specify a comment, include an empty Comment element.

To delete an existing comment, update the distribution configuration and include an empty Comment element.

To add or change a comment, update the distribution configuration and specify the new comment.

", - "DistributionConfig$WebACLId": "

A unique identifier that specifies the AWS WAF web ACL, if any, to associate with this distribution.

AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to CloudFront, and lets you control access to your content. Based on conditions that you specify, such as the IP addresses that requests originate from or the values of query strings, CloudFront responds to requests either with the requested content or with an HTTP 403 status code (Forbidden). You can also configure CloudFront to return a custom error page when a request is blocked. For more information about AWS WAF, see the AWS WAF Developer Guide.

", - "DistributionList$Marker": "

The value you provided for the Marker request parameter.

", - "DistributionList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.

", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "DistributionSummary$ARN": "

The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "DistributionSummary$Status": "

The current status of the distribution. When the status is Deployed, the distribution's information is propagated to all CloudFront edge locations.

", - "DistributionSummary$DomainName": "

The domain name that corresponds to the distribution. For example: d604721fxaaqy9.cloudfront.net.

", - "DistributionSummary$Comment": "

The comment originally specified when this distribution was created.

", - "DistributionSummary$WebACLId": "

The Web ACL Id (if any) associated with the distribution.

", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "

The identity's ID.

", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetCloudFrontOriginAccessIdentityRequest$Id": "

The identity's ID.

", - "GetCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.

", - "GetDistributionConfigRequest$Id": "

The distribution's ID.

", - "GetDistributionConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetDistributionRequest$Id": "

The distribution's ID.

", - "GetDistributionResult$ETag": "

The current version of the distribution's information. For example: E2QWRUHAPOMQZL.

", - "GetInvalidationRequest$DistributionId": "

The distribution's ID.

", - "GetInvalidationRequest$Id": "

The identifier for the invalidation request, for example, IDFDVBD632BHDS5.

", - "GetStreamingDistributionConfigRequest$Id": "

The streaming distribution's ID.

", - "GetStreamingDistributionConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetStreamingDistributionRequest$Id": "

The streaming distribution's ID.

", - "GetStreamingDistributionResult$ETag": "

The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.

", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLambdaFunctionAssociation$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidQueryStringParameters$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidTagging$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "

The identifier for the invalidation request. For example: IDFDVBD632BHDS5.

", - "Invalidation$Status": "

The status of the invalidation request. When the invalidation batch is finished, the status is Completed.

", - "InvalidationBatch$CallerReference": "

A value that you specify to uniquely identify an invalidation request. CloudFront uses the value to prevent you from accidentally resubmitting an identical request. Whenever you create a new invalidation request, you must specify a new value for CallerReference and change other values in the request as applicable. One way to ensure that the value of CallerReference is unique is to use a timestamp, for example, 20120301090000.

If you make a second invalidation request with the same value for CallerReference, and if the rest of the request is the same, CloudFront doesn't create a new invalidation request. Instead, CloudFront returns information about the invalidation request that you previously created with the same CallerReference.

If CallerReference is a value you already sent in a previous invalidation batch request but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.

", - "InvalidationList$Marker": "

The value that you provided for the Marker request parameter.

", - "InvalidationList$NextMarker": "

If IsTruncated is true, this element is present and contains the value that you can use for the Marker request parameter to continue listing your invalidation batches where they left off.

", - "InvalidationSummary$Id": "

The unique ID for an invalidation request.

", - "InvalidationSummary$Status": "

The status of an invalidation request.

", - "KeyPairIdList$member": null, - "LambdaFunctionAssociation$LambdaFunctionARN": "

The ARN of the Lambda function.

", - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "

Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).

", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "

The maximum number of origin access identities you want in the response body.

", - "ListDistributionsByWebACLIdRequest$Marker": "

Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)

", - "ListDistributionsByWebACLIdRequest$MaxItems": "

The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.

", - "ListDistributionsByWebACLIdRequest$WebACLId": "

The ID of the AWS WAF web ACL that you want to list the associated distributions. If you specify \"null\" for the ID, the request returns a list of the distributions that aren't associated with a web ACL.

", - "ListDistributionsRequest$Marker": "

Use this when paginating results to indicate where to begin in your list of distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).

", - "ListDistributionsRequest$MaxItems": "

The maximum number of distributions you want in the response body.

", - "ListInvalidationsRequest$DistributionId": "

The distribution's ID.

", - "ListInvalidationsRequest$Marker": "

Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.

", - "ListInvalidationsRequest$MaxItems": "

The maximum number of invalidation batches that you want in the response body.

", - "ListStreamingDistributionsRequest$Marker": "

The value that you provided for the Marker request parameter.

", - "ListStreamingDistributionsRequest$MaxItems": "

The value that you provided for the MaxItems request parameter.

", - "LocationList$member": null, - "LoggingConfig$Bucket": "

The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.

", - "LoggingConfig$Prefix": "

An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.

", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchResource$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "

A unique identifier for the origin. The value of Id must be unique within the distribution.

When you specify the value of TargetOriginId for the default cache behavior or for another cache behavior, you indicate the origin to which you want the cache behavior to route requests by specifying the value of the Id element for that origin. When a request matches the path pattern for that cache behavior, CloudFront routes the request to the specified origin. For more information, see Cache Behavior Settings in the Amazon CloudFront Developer Guide.

", - "Origin$DomainName": "

Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com.

Constraints for Amazon S3 origins:

  • If you configured Amazon S3 Transfer Acceleration for your bucket, do not specify the s3-accelerate endpoint for DomainName.

  • The bucket name must be between 3 and 63 characters long (inclusive).

  • The bucket name must contain only lowercase characters, numbers, periods, underscores, and dashes.

  • The bucket name must not contain adjacent periods.

Custom Origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.

Constraints for custom origins:

  • DomainName must be a valid DNS name that contains only a-z, A-Z, 0-9, dot (.), hyphen (-), or underscore (_) characters.

  • The name cannot exceed 128 characters.

", - "Origin$OriginPath": "

An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName, for example, example.com/production. Do not include a / at the end of the directory name.

For example, suppose you've specified the following values for your distribution:

  • DomainName: An Amazon S3 bucket named myawsbucket.

  • OriginPath: /production

  • CNAME: example.com

When a user enters example.com/index.html in a browser, CloudFront sends a request to Amazon S3 for myawsbucket/production/index.html.

When a user enters example.com/acme/index.html in a browser, CloudFront sends a request to Amazon S3 for myawsbucket/production/acme/index.html.

", - "OriginCustomHeader$HeaderName": "

The name of a header that you want CloudFront to forward to your origin. For more information, see Forwarding Custom Headers to Your Origin (Web Distributions Only) in the Amazon Amazon CloudFront Developer Guide.

", - "OriginCustomHeader$HeaderValue": "

The value for the header that you specified in the HeaderName field.

", - "PathList$member": null, - "PreconditionFailed$Message": null, - "QueryStringCacheKeysList$member": null, - "S3Origin$DomainName": "

The DNS name of the Amazon S3 origin.

", - "S3Origin$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the RTMP distribution. Use an origin access identity to configure the distribution so that end users can only access objects in an Amazon S3 bucket through CloudFront.

If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information, see Using an Origin Access Identity to Restrict Access to Your Amazon S3 Content in the Amazon Amazon CloudFront Developer Guide.

", - "S3OriginConfig$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that viewers can only access objects in an Amazon S3 bucket through CloudFront. The format of the value is:

origin-access-identity/CloudFront/ID-of-origin-access-identity

where ID-of-origin-access-identity is the value that CloudFront returned in the ID element when you created the origin access identity.

If you want viewers to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information about the origin access identity, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "Signer$AwsAccountNumber": "

An AWS account that is included in the TrustedSigners complex type for this RTMP distribution. Valid values include:

  • self, which is the AWS account used to create the distribution.

  • An AWS account number.

", - "StreamingDistribution$Id": "

The identifier for the RTMP distribution. For example: EGTXBD79EXAMPLE.

", - "StreamingDistribution$ARN": null, - "StreamingDistribution$Status": "

The current status of the RTMP distribution. When the status is Deployed, the distribution's information is propagated to all CloudFront edge locations.

", - "StreamingDistribution$DomainName": "

The domain name that corresponds to the streaming distribution. For example: s5c39gqb8ow64r.cloudfront.net.

", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "

A unique number that ensures that the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value that you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value that you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.

", - "StreamingDistributionConfig$Comment": "

Any comments you want to include about the streaming distribution.

", - "StreamingDistributionList$Marker": "

The value you provided for the Marker request parameter.

", - "StreamingDistributionList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your RTMP distributions where they left off.

", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "StreamingDistributionSummary$ARN": "

The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "StreamingDistributionSummary$Status": "

Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.

", - "StreamingDistributionSummary$DomainName": "

The domain name corresponding to the distribution. For example: d604721fxaaqy9.cloudfront.net.

", - "StreamingDistributionSummary$Comment": "

The comment originally specified when this distribution was created.

", - "StreamingLoggingConfig$Bucket": "

The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.

", - "StreamingLoggingConfig$Prefix": "

An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you do not want to specify a prefix, you still must include an empty Prefix element in the Logging element.

", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyDistributionsWithLambdaAssociations$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyLambdaFunctionAssociations$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyQueryStringParameters$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "

The identity's id.

", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "

The value of the ETag header that you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateDistributionRequest$Id": "

The distribution's id.

", - "UpdateDistributionRequest$IfMatch": "

The value of the ETag header that you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateDistributionResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateStreamingDistributionRequest$Id": "

The streaming distribution's id.

", - "UpdateStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateStreamingDistributionResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "ViewerCertificate$IAMCertificateId": null, - "ViewerCertificate$ACMCertificateArn": null, - "ViewerCertificate$Certificate": "

Include one of these values to specify the following:

  • Whether you want viewers to use HTTP or HTTPS to request your objects.

  • If you want viewers to use HTTPS, whether you're using an alternate domain name such as example.com or the CloudFront domain name for your distribution, such as d111111abcdef8.cloudfront.net.

  • If you're using an alternate domain name, whether AWS Certificate Manager (ACM) provided the certificate, or you purchased a certificate from a third-party certificate authority and imported it into ACM or uploaded it to the IAM certificate store.

You must specify one (and only one) of the three values. Do not specify false for CloudFrontDefaultCertificate.

If you want viewers to use HTTP to request your objects: Specify the following value:

<CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>

In addition, specify allow-all for ViewerProtocolPolicy for all of your cache behaviors.

If you want viewers to use HTTPS to request your objects: Choose the type of certificate that you want to use based on whether you're using an alternate domain name for your objects or the CloudFront domain name:

  • If you're using an alternate domain name, such as example.com: Specify one of the following values, depending on whether ACM provided your certificate or you purchased your certificate from third-party certificate authority:

    • <ACMCertificateArn>ARN for ACM SSL/TLS certificate<ACMCertificateArn> where ARN for ACM SSL/TLS certificate is the ARN for the ACM SSL/TLS certificate that you want to use for this distribution.

    • <IAMCertificateId>IAM certificate ID<IAMCertificateId> where IAM certificate ID is the ID that IAM returned when you added the certificate to the IAM certificate store.

    If you specify ACMCertificateArn or IAMCertificateId, you must also specify a value for SSLSupportMethod.

    If you choose to use an ACM certificate or a certificate in the IAM certificate store, we recommend that you use only an alternate domain name in your object URLs (https://example.com/logo.jpg). If you use the domain name that is associated with your CloudFront distribution (https://d111111abcdef8.cloudfront.net/logo.jpg) and the viewer supports SNI, then CloudFront behaves normally. However, if the browser does not support SNI, the user's experience depends on the value that you choose for SSLSupportMethod:

    • vip: The viewer displays a warning because there is a mismatch between the CloudFront domain name and the domain name in your SSL/TLS certificate.

    • sni-only: CloudFront drops the connection with the browser without returning the object.

  • If you're using the CloudFront domain name for your distribution, such as d111111abcdef8.cloudfront.net : Specify the following value:

    <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>

    If you want viewers to use HTTPS, you must also specify one of the following values in your cache behaviors:

    • <ViewerProtocolPolicy>https-only<ViewerProtocolPolicy>

    • <ViewerProtocolPolicy>redirect-to-https<ViewerProtocolPolicy>

    You can also optionally require that CloudFront use HTTPS to communicate with your origin by specifying one of the following values for the applicable origins:

    • <OriginProtocolPolicy>https-only<OriginProtocolPolicy>

    • <OriginProtocolPolicy>match-viewer<OriginProtocolPolicy>

    For more information, see Using Alternate Domain Names and HTTPS in the Amazon CloudFront Developer Guide.

" - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "

The date and time the distribution was last modified.

", - "DistributionSummary$LastModifiedTime": "

The date and time the distribution was last modified.

", - "Invalidation$CreateTime": "

The date and time the invalidation request was first made.

", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "

The date and time that the distribution was last modified.

", - "StreamingDistributionSummary$LastModifiedTime": "

The date and time the distribution was last modified.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/paginators-1.json deleted file mode 100644 index 51fbb907f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "output_token": "DistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "output_token": "InvalidationList.NextMarker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "output_token": "StreamingDistributionList.NextMarker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "result_key": "StreamingDistributionList.Items" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/waiters-2.json deleted file mode 100644 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2016-11-25/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/api-2.json deleted file mode 100644 index 3ff77f4f8..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/api-2.json +++ /dev/null @@ -1,2726 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2017-03-25", - "endpointPrefix":"cloudfront", - "globalEndpoint":"cloudfront.amazonaws.com", - "protocol":"rest-xml", - "serviceAbbreviation":"CloudFront", - "serviceFullName":"Amazon CloudFront", - "signatureVersion":"v4", - "uid":"cloudfront-2017-03-25" - }, - "operations":{ - "CreateCloudFrontOriginAccessIdentity":{ - "name":"CreateCloudFrontOriginAccessIdentity2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/origin-access-identity/cloudfront", - "responseCode":201 - }, - "input":{"shape":"CreateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"CreateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"CloudFrontOriginAccessIdentityAlreadyExists"}, - {"shape":"MissingBody"}, - {"shape":"TooManyCloudFrontOriginAccessIdentities"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateDistribution":{ - "name":"CreateDistribution2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/distribution", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionRequest"}, - "output":{"shape":"CreateDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"}, - {"shape":"TooManyDistributionsWithLambdaAssociations"}, - {"shape":"TooManyLambdaFunctionAssociations"}, - {"shape":"InvalidLambdaFunctionAssociation"}, - {"shape":"InvalidOriginReadTimeout"}, - {"shape":"InvalidOriginKeepaliveTimeout"} - ] - }, - "CreateDistributionWithTags":{ - "name":"CreateDistributionWithTags2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateDistributionWithTagsRequest"}, - "output":{"shape":"CreateDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"DistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"MissingBody"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"TooManyDistributions"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidProtocolSettings"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"InvalidTagging"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"}, - {"shape":"TooManyDistributionsWithLambdaAssociations"}, - {"shape":"TooManyLambdaFunctionAssociations"}, - {"shape":"InvalidLambdaFunctionAssociation"}, - {"shape":"InvalidOriginReadTimeout"}, - {"shape":"InvalidOriginKeepaliveTimeout"} - ] - }, - "CreateInvalidation":{ - "name":"CreateInvalidation2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/distribution/{DistributionId}/invalidation", - "responseCode":201 - }, - "input":{"shape":"CreateInvalidationRequest"}, - "output":{"shape":"CreateInvalidationResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"MissingBody"}, - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"BatchTooLarge"}, - {"shape":"TooManyInvalidationsInProgress"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistribution":{ - "name":"CreateStreamingDistribution2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/streaming-distribution", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionRequest"}, - "output":{"shape":"CreateStreamingDistributionResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "CreateStreamingDistributionWithTags":{ - "name":"CreateStreamingDistributionWithTags2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/streaming-distribution?WithTags", - "responseCode":201 - }, - "input":{"shape":"CreateStreamingDistributionWithTagsRequest"}, - "output":{"shape":"CreateStreamingDistributionWithTagsResult"}, - "errors":[ - {"shape":"CNAMEAlreadyExists"}, - {"shape":"StreamingDistributionAlreadyExists"}, - {"shape":"InvalidOrigin"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"AccessDenied"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"MissingBody"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"TooManyStreamingDistributions"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"}, - {"shape":"InvalidTagging"} - ] - }, - "DeleteCloudFrontOriginAccessIdentity":{ - "name":"DeleteCloudFrontOriginAccessIdentity2017_03_25", - "http":{ - "method":"DELETE", - "requestUri":"/2017-03-25/origin-access-identity/cloudfront/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteCloudFrontOriginAccessIdentityRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"CloudFrontOriginAccessIdentityInUse"} - ] - }, - "DeleteDistribution":{ - "name":"DeleteDistribution2017_03_25", - "http":{ - "method":"DELETE", - "requestUri":"/2017-03-25/distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"DistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "DeleteServiceLinkedRole":{ - "name":"DeleteServiceLinkedRole2017_03_25", - "http":{ - "method":"DELETE", - "requestUri":"/2017-03-25/service-linked-role/{RoleName}", - "responseCode":204 - }, - "input":{"shape":"DeleteServiceLinkedRoleRequest"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"AccessDenied"}, - {"shape":"ResourceInUse"}, - {"shape":"NoSuchResource"} - ] - }, - "DeleteStreamingDistribution":{ - "name":"DeleteStreamingDistribution2017_03_25", - "http":{ - "method":"DELETE", - "requestUri":"/2017-03-25/streaming-distribution/{Id}", - "responseCode":204 - }, - "input":{"shape":"DeleteStreamingDistributionRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"StreamingDistributionNotDisabled"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"} - ] - }, - "GetCloudFrontOriginAccessIdentity":{ - "name":"GetCloudFrontOriginAccessIdentity2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/origin-access-identity/cloudfront/{Id}" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetCloudFrontOriginAccessIdentityConfig":{ - "name":"GetCloudFrontOriginAccessIdentityConfig2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"GetCloudFrontOriginAccessIdentityConfigRequest"}, - "output":{"shape":"GetCloudFrontOriginAccessIdentityConfigResult"}, - "errors":[ - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistribution":{ - "name":"GetDistribution2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/distribution/{Id}" - }, - "input":{"shape":"GetDistributionRequest"}, - "output":{"shape":"GetDistributionResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetDistributionConfig":{ - "name":"GetDistributionConfig2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/distribution/{Id}/config" - }, - "input":{"shape":"GetDistributionConfigRequest"}, - "output":{"shape":"GetDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetInvalidation":{ - "name":"GetInvalidation2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/distribution/{DistributionId}/invalidation/{Id}" - }, - "input":{"shape":"GetInvalidationRequest"}, - "output":{"shape":"GetInvalidationResult"}, - "errors":[ - {"shape":"NoSuchInvalidation"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistribution":{ - "name":"GetStreamingDistribution2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/streaming-distribution/{Id}" - }, - "input":{"shape":"GetStreamingDistributionRequest"}, - "output":{"shape":"GetStreamingDistributionResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "GetStreamingDistributionConfig":{ - "name":"GetStreamingDistributionConfig2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/streaming-distribution/{Id}/config" - }, - "input":{"shape":"GetStreamingDistributionConfigRequest"}, - "output":{"shape":"GetStreamingDistributionConfigResult"}, - "errors":[ - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListCloudFrontOriginAccessIdentities":{ - "name":"ListCloudFrontOriginAccessIdentities2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/origin-access-identity/cloudfront" - }, - "input":{"shape":"ListCloudFrontOriginAccessIdentitiesRequest"}, - "output":{"shape":"ListCloudFrontOriginAccessIdentitiesResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributions":{ - "name":"ListDistributions2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/distribution" - }, - "input":{"shape":"ListDistributionsRequest"}, - "output":{"shape":"ListDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListDistributionsByWebACLId":{ - "name":"ListDistributionsByWebACLId2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/distributionsByWebACLId/{WebACLId}" - }, - "input":{"shape":"ListDistributionsByWebACLIdRequest"}, - "output":{"shape":"ListDistributionsByWebACLIdResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"InvalidWebACLId"} - ] - }, - "ListInvalidations":{ - "name":"ListInvalidations2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/distribution/{DistributionId}/invalidation" - }, - "input":{"shape":"ListInvalidationsRequest"}, - "output":{"shape":"ListInvalidationsResult"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"NoSuchDistribution"}, - {"shape":"AccessDenied"} - ] - }, - "ListStreamingDistributions":{ - "name":"ListStreamingDistributions2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/streaming-distribution" - }, - "input":{"shape":"ListStreamingDistributionsRequest"}, - "output":{"shape":"ListStreamingDistributionsResult"}, - "errors":[ - {"shape":"InvalidArgument"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource2017_03_25", - "http":{ - "method":"GET", - "requestUri":"/2017-03-25/tagging" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "TagResource":{ - "name":"TagResource2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/tagging?Operation=Tag", - "responseCode":204 - }, - "input":{"shape":"TagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UntagResource":{ - "name":"UntagResource2017_03_25", - "http":{ - "method":"POST", - "requestUri":"/2017-03-25/tagging?Operation=Untag", - "responseCode":204 - }, - "input":{"shape":"UntagResourceRequest"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidTagging"}, - {"shape":"NoSuchResource"} - ] - }, - "UpdateCloudFrontOriginAccessIdentity":{ - "name":"UpdateCloudFrontOriginAccessIdentity2017_03_25", - "http":{ - "method":"PUT", - "requestUri":"/2017-03-25/origin-access-identity/cloudfront/{Id}/config" - }, - "input":{"shape":"UpdateCloudFrontOriginAccessIdentityRequest"}, - "output":{"shape":"UpdateCloudFrontOriginAccessIdentityResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchCloudFrontOriginAccessIdentity"}, - {"shape":"PreconditionFailed"}, - {"shape":"InvalidArgument"}, - {"shape":"InconsistentQuantities"} - ] - }, - "UpdateDistribution":{ - "name":"UpdateDistribution2017_03_25", - "http":{ - "method":"PUT", - "requestUri":"/2017-03-25/distribution/{Id}/config" - }, - "input":{"shape":"UpdateDistributionRequest"}, - "output":{"shape":"UpdateDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyDistributionCNAMEs"}, - {"shape":"InvalidDefaultRootObject"}, - {"shape":"InvalidRelativePath"}, - {"shape":"InvalidErrorCode"}, - {"shape":"InvalidResponseCode"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InvalidViewerCertificate"}, - {"shape":"InvalidMinimumProtocolVersion"}, - {"shape":"InvalidRequiredProtocol"}, - {"shape":"NoSuchOrigin"}, - {"shape":"TooManyOrigins"}, - {"shape":"TooManyCacheBehaviors"}, - {"shape":"TooManyCookieNamesInWhiteList"}, - {"shape":"InvalidForwardCookies"}, - {"shape":"TooManyHeadersInForwardedValues"}, - {"shape":"InvalidHeadersForS3Origin"}, - {"shape":"InconsistentQuantities"}, - {"shape":"TooManyCertificates"}, - {"shape":"InvalidLocationCode"}, - {"shape":"InvalidGeoRestrictionParameter"}, - {"shape":"InvalidTTLOrder"}, - {"shape":"InvalidWebACLId"}, - {"shape":"TooManyOriginCustomHeaders"}, - {"shape":"TooManyQueryStringParameters"}, - {"shape":"InvalidQueryStringParameters"}, - {"shape":"TooManyDistributionsWithLambdaAssociations"}, - {"shape":"TooManyLambdaFunctionAssociations"}, - {"shape":"InvalidLambdaFunctionAssociation"}, - {"shape":"InvalidOriginReadTimeout"}, - {"shape":"InvalidOriginKeepaliveTimeout"} - ] - }, - "UpdateStreamingDistribution":{ - "name":"UpdateStreamingDistribution2017_03_25", - "http":{ - "method":"PUT", - "requestUri":"/2017-03-25/streaming-distribution/{Id}/config" - }, - "input":{"shape":"UpdateStreamingDistributionRequest"}, - "output":{"shape":"UpdateStreamingDistributionResult"}, - "errors":[ - {"shape":"AccessDenied"}, - {"shape":"CNAMEAlreadyExists"}, - {"shape":"IllegalUpdate"}, - {"shape":"InvalidIfMatchVersion"}, - {"shape":"MissingBody"}, - {"shape":"NoSuchStreamingDistribution"}, - {"shape":"PreconditionFailed"}, - {"shape":"TooManyStreamingDistributionCNAMEs"}, - {"shape":"InvalidArgument"}, - {"shape":"InvalidOriginAccessIdentity"}, - {"shape":"TooManyTrustedSigners"}, - {"shape":"TrustedSignerDoesNotExist"}, - {"shape":"InconsistentQuantities"} - ] - } - }, - "shapes":{ - "AccessDenied":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":403}, - "exception":true - }, - "ActiveTrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SignerList"} - } - }, - "AliasList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"CNAME" - } - }, - "Aliases":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AliasList"} - } - }, - "AllowedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"}, - "CachedMethods":{"shape":"CachedMethods"} - } - }, - "AwsAccountNumberList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"AwsAccountNumber" - } - }, - "BatchTooLarge":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":413}, - "exception":true - }, - "CNAMEAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CacheBehavior":{ - "type":"structure", - "required":[ - "PathPattern", - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "PathPattern":{"shape":"string"}, - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"}, - "LambdaFunctionAssociations":{"shape":"LambdaFunctionAssociations"} - } - }, - "CacheBehaviorList":{ - "type":"list", - "member":{ - "shape":"CacheBehavior", - "locationName":"CacheBehavior" - } - }, - "CacheBehaviors":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CacheBehaviorList"} - } - }, - "CachedMethods":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"MethodsList"} - } - }, - "CertificateSource":{ - "type":"string", - "enum":[ - "cloudfront", - "iam", - "acm" - ] - }, - "CloudFrontOriginAccessIdentity":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"} - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Comment" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentityInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "CloudFrontOriginAccessIdentityList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CloudFrontOriginAccessIdentitySummaryList"} - } - }, - "CloudFrontOriginAccessIdentitySummary":{ - "type":"structure", - "required":[ - "Id", - "S3CanonicalUserId", - "Comment" - ], - "members":{ - "Id":{"shape":"string"}, - "S3CanonicalUserId":{"shape":"string"}, - "Comment":{"shape":"string"} - } - }, - "CloudFrontOriginAccessIdentitySummaryList":{ - "type":"list", - "member":{ - "shape":"CloudFrontOriginAccessIdentitySummary", - "locationName":"CloudFrontOriginAccessIdentitySummary" - } - }, - "CookieNameList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "CookieNames":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CookieNameList"} - } - }, - "CookiePreference":{ - "type":"structure", - "required":["Forward"], - "members":{ - "Forward":{"shape":"ItemSelection"}, - "WhitelistedNames":{"shape":"CookieNames"} - } - }, - "CreateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["CloudFrontOriginAccessIdentityConfig"], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "CreateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "CreateDistributionRequest":{ - "type":"structure", - "required":["DistributionConfig"], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"DistributionConfig" - }, - "CreateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateDistributionWithTagsRequest":{ - "type":"structure", - "required":["DistributionConfigWithTags"], - "members":{ - "DistributionConfigWithTags":{ - "shape":"DistributionConfigWithTags", - "locationName":"DistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"DistributionConfigWithTags" - }, - "CreateDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "CreateInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "InvalidationBatch" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "InvalidationBatch":{ - "shape":"InvalidationBatch", - "locationName":"InvalidationBatch", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"InvalidationBatch" - }, - "CreateInvalidationResult":{ - "type":"structure", - "members":{ - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "CreateStreamingDistributionRequest":{ - "type":"structure", - "required":["StreamingDistributionConfig"], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"StreamingDistributionConfig" - }, - "CreateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CreateStreamingDistributionWithTagsRequest":{ - "type":"structure", - "required":["StreamingDistributionConfigWithTags"], - "members":{ - "StreamingDistributionConfigWithTags":{ - "shape":"StreamingDistributionConfigWithTags", - "locationName":"StreamingDistributionConfigWithTags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"StreamingDistributionConfigWithTags" - }, - "CreateStreamingDistributionWithTagsResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "Location":{ - "shape":"string", - "location":"header", - "locationName":"Location" - }, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "CustomErrorResponse":{ - "type":"structure", - "required":["ErrorCode"], - "members":{ - "ErrorCode":{"shape":"integer"}, - "ResponsePagePath":{"shape":"string"}, - "ResponseCode":{"shape":"string"}, - "ErrorCachingMinTTL":{"shape":"long"} - } - }, - "CustomErrorResponseList":{ - "type":"list", - "member":{ - "shape":"CustomErrorResponse", - "locationName":"CustomErrorResponse" - } - }, - "CustomErrorResponses":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"CustomErrorResponseList"} - } - }, - "CustomHeaders":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginCustomHeadersList"} - } - }, - "CustomOriginConfig":{ - "type":"structure", - "required":[ - "HTTPPort", - "HTTPSPort", - "OriginProtocolPolicy" - ], - "members":{ - "HTTPPort":{"shape":"integer"}, - "HTTPSPort":{"shape":"integer"}, - "OriginProtocolPolicy":{"shape":"OriginProtocolPolicy"}, - "OriginSslProtocols":{"shape":"OriginSslProtocols"}, - "OriginReadTimeout":{"shape":"integer"}, - "OriginKeepaliveTimeout":{"shape":"integer"} - } - }, - "DefaultCacheBehavior":{ - "type":"structure", - "required":[ - "TargetOriginId", - "ForwardedValues", - "TrustedSigners", - "ViewerProtocolPolicy", - "MinTTL" - ], - "members":{ - "TargetOriginId":{"shape":"string"}, - "ForwardedValues":{"shape":"ForwardedValues"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "ViewerProtocolPolicy":{"shape":"ViewerProtocolPolicy"}, - "MinTTL":{"shape":"long"}, - "AllowedMethods":{"shape":"AllowedMethods"}, - "SmoothStreaming":{"shape":"boolean"}, - "DefaultTTL":{"shape":"long"}, - "MaxTTL":{"shape":"long"}, - "Compress":{"shape":"boolean"}, - "LambdaFunctionAssociations":{"shape":"LambdaFunctionAssociations"} - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "DeleteServiceLinkedRoleRequest":{ - "type":"structure", - "required":["RoleName"], - "members":{ - "RoleName":{ - "shape":"string", - "location":"uri", - "locationName":"RoleName" - } - } - }, - "DeleteStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - } - }, - "Distribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "InProgressInvalidationBatches", - "DomainName", - "ActiveTrustedSigners", - "DistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "InProgressInvalidationBatches":{"shape":"integer"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "DistributionConfig":{"shape":"DistributionConfig"} - } - }, - "DistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "Origins", - "DefaultCacheBehavior", - "Comment", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "DefaultRootObject":{"shape":"string"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"LoggingConfig"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"}, - "IsIPV6Enabled":{"shape":"boolean"} - } - }, - "DistributionConfigWithTags":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Tags" - ], - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "DistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"DistributionSummaryList"} - } - }, - "DistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "DistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "Aliases", - "Origins", - "DefaultCacheBehavior", - "CacheBehaviors", - "CustomErrorResponses", - "Comment", - "PriceClass", - "Enabled", - "ViewerCertificate", - "Restrictions", - "WebACLId", - "HttpVersion", - "IsIPV6Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "Aliases":{"shape":"Aliases"}, - "Origins":{"shape":"Origins"}, - "DefaultCacheBehavior":{"shape":"DefaultCacheBehavior"}, - "CacheBehaviors":{"shape":"CacheBehaviors"}, - "CustomErrorResponses":{"shape":"CustomErrorResponses"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"}, - "ViewerCertificate":{"shape":"ViewerCertificate"}, - "Restrictions":{"shape":"Restrictions"}, - "WebACLId":{"shape":"string"}, - "HttpVersion":{"shape":"HttpVersion"}, - "IsIPV6Enabled":{"shape":"boolean"} - } - }, - "DistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"DistributionSummary", - "locationName":"DistributionSummary" - } - }, - "EventType":{ - "type":"string", - "enum":[ - "viewer-request", - "viewer-response", - "origin-request", - "origin-response" - ] - }, - "ForwardedValues":{ - "type":"structure", - "required":[ - "QueryString", - "Cookies" - ], - "members":{ - "QueryString":{"shape":"boolean"}, - "Cookies":{"shape":"CookiePreference"}, - "Headers":{"shape":"Headers"}, - "QueryStringCacheKeys":{"shape":"QueryStringCacheKeys"} - } - }, - "GeoRestriction":{ - "type":"structure", - "required":[ - "RestrictionType", - "Quantity" - ], - "members":{ - "RestrictionType":{"shape":"GeoRestrictionType"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LocationList"} - } - }, - "GeoRestrictionType":{ - "type":"string", - "enum":[ - "blacklist", - "whitelist", - "none" - ] - }, - "GetCloudFrontOriginAccessIdentityConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityConfig":{"shape":"CloudFrontOriginAccessIdentityConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "GetCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "GetDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionConfigResult":{ - "type":"structure", - "members":{ - "DistributionConfig":{"shape":"DistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"DistributionConfig" - }, - "GetDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "GetInvalidationRequest":{ - "type":"structure", - "required":[ - "DistributionId", - "Id" - ], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetInvalidationResult":{ - "type":"structure", - "members":{ - "Invalidation":{"shape":"Invalidation"} - }, - "payload":"Invalidation" - }, - "GetStreamingDistributionConfigRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionConfigResult":{ - "type":"structure", - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistributionConfig" - }, - "GetStreamingDistributionRequest":{ - "type":"structure", - "required":["Id"], - "members":{ - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - } - } - }, - "GetStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "HeaderList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "Headers":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"HeaderList"} - } - }, - "HttpVersion":{ - "type":"string", - "enum":[ - "http1.1", - "http2" - ] - }, - "IllegalUpdate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InconsistentQuantities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidArgument":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidDefaultRootObject":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidErrorCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidForwardCookies":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidGeoRestrictionParameter":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidHeadersForS3Origin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidIfMatchVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLambdaFunctionAssociation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidLocationCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidMinimumProtocolVersion":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginKeepaliveTimeout":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidOriginReadTimeout":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidProtocolSettings":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRelativePath":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequiredProtocol":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidResponseCode":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTTLOrder":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidTagging":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidViewerCertificate":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidWebACLId":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "Invalidation":{ - "type":"structure", - "required":[ - "Id", - "Status", - "CreateTime", - "InvalidationBatch" - ], - "members":{ - "Id":{"shape":"string"}, - "Status":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "InvalidationBatch":{"shape":"InvalidationBatch"} - } - }, - "InvalidationBatch":{ - "type":"structure", - "required":[ - "Paths", - "CallerReference" - ], - "members":{ - "Paths":{"shape":"Paths"}, - "CallerReference":{"shape":"string"} - } - }, - "InvalidationList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"InvalidationSummaryList"} - } - }, - "InvalidationSummary":{ - "type":"structure", - "required":[ - "Id", - "CreateTime", - "Status" - ], - "members":{ - "Id":{"shape":"string"}, - "CreateTime":{"shape":"timestamp"}, - "Status":{"shape":"string"} - } - }, - "InvalidationSummaryList":{ - "type":"list", - "member":{ - "shape":"InvalidationSummary", - "locationName":"InvalidationSummary" - } - }, - "ItemSelection":{ - "type":"string", - "enum":[ - "none", - "whitelist", - "all" - ] - }, - "KeyPairIdList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"KeyPairId" - } - }, - "KeyPairIds":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"KeyPairIdList"} - } - }, - "LambdaFunctionAssociation":{ - "type":"structure", - "members":{ - "LambdaFunctionARN":{"shape":"string"}, - "EventType":{"shape":"EventType"} - } - }, - "LambdaFunctionAssociationList":{ - "type":"list", - "member":{ - "shape":"LambdaFunctionAssociation", - "locationName":"LambdaFunctionAssociation" - } - }, - "LambdaFunctionAssociations":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"LambdaFunctionAssociationList"} - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListCloudFrontOriginAccessIdentitiesResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentityList":{"shape":"CloudFrontOriginAccessIdentityList"} - }, - "payload":"CloudFrontOriginAccessIdentityList" - }, - "ListDistributionsByWebACLIdRequest":{ - "type":"structure", - "required":["WebACLId"], - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - }, - "WebACLId":{ - "shape":"string", - "location":"uri", - "locationName":"WebACLId" - } - } - }, - "ListDistributionsByWebACLIdResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListDistributionsResult":{ - "type":"structure", - "members":{ - "DistributionList":{"shape":"DistributionList"} - }, - "payload":"DistributionList" - }, - "ListInvalidationsRequest":{ - "type":"structure", - "required":["DistributionId"], - "members":{ - "DistributionId":{ - "shape":"string", - "location":"uri", - "locationName":"DistributionId" - }, - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListInvalidationsResult":{ - "type":"structure", - "members":{ - "InvalidationList":{"shape":"InvalidationList"} - }, - "payload":"InvalidationList" - }, - "ListStreamingDistributionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"string", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"string", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListStreamingDistributionsResult":{ - "type":"structure", - "members":{ - "StreamingDistributionList":{"shape":"StreamingDistributionList"} - }, - "payload":"StreamingDistributionList" - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["Resource"], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - } - } - }, - "ListTagsForResourceResult":{ - "type":"structure", - "required":["Tags"], - "members":{ - "Tags":{"shape":"Tags"} - }, - "payload":"Tags" - }, - "LocationList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Location" - } - }, - "LoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "IncludeCookies", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "IncludeCookies":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Method":{ - "type":"string", - "enum":[ - "GET", - "HEAD", - "POST", - "PUT", - "PATCH", - "OPTIONS", - "DELETE" - ] - }, - "MethodsList":{ - "type":"list", - "member":{ - "shape":"Method", - "locationName":"Method" - } - }, - "MinimumProtocolVersion":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1_2016", - "TLSv1.1_2016", - "TLSv1.2_2018" - ] - }, - "MissingBody":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "NoSuchCloudFrontOriginAccessIdentity":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchInvalidation":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchOrigin":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchResource":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "NoSuchStreamingDistribution":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "Origin":{ - "type":"structure", - "required":[ - "Id", - "DomainName" - ], - "members":{ - "Id":{"shape":"string"}, - "DomainName":{"shape":"string"}, - "OriginPath":{"shape":"string"}, - "CustomHeaders":{"shape":"CustomHeaders"}, - "S3OriginConfig":{"shape":"S3OriginConfig"}, - "CustomOriginConfig":{"shape":"CustomOriginConfig"} - } - }, - "OriginCustomHeader":{ - "type":"structure", - "required":[ - "HeaderName", - "HeaderValue" - ], - "members":{ - "HeaderName":{"shape":"string"}, - "HeaderValue":{"shape":"string"} - } - }, - "OriginCustomHeadersList":{ - "type":"list", - "member":{ - "shape":"OriginCustomHeader", - "locationName":"OriginCustomHeader" - } - }, - "OriginList":{ - "type":"list", - "member":{ - "shape":"Origin", - "locationName":"Origin" - }, - "min":1 - }, - "OriginProtocolPolicy":{ - "type":"string", - "enum":[ - "http-only", - "match-viewer", - "https-only" - ] - }, - "OriginSslProtocols":{ - "type":"structure", - "required":[ - "Quantity", - "Items" - ], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"SslProtocolsList"} - } - }, - "Origins":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"OriginList"} - } - }, - "PathList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Path" - } - }, - "Paths":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"PathList"} - } - }, - "PreconditionFailed":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":412}, - "exception":true - }, - "PriceClass":{ - "type":"string", - "enum":[ - "PriceClass_100", - "PriceClass_200", - "PriceClass_All" - ] - }, - "QueryStringCacheKeys":{ - "type":"structure", - "required":["Quantity"], - "members":{ - "Quantity":{"shape":"integer"}, - "Items":{"shape":"QueryStringCacheKeysList"} - } - }, - "QueryStringCacheKeysList":{ - "type":"list", - "member":{ - "shape":"string", - "locationName":"Name" - } - }, - "ResourceARN":{ - "type":"string", - "pattern":"arn:aws:cloudfront::[0-9]+:.*" - }, - "ResourceInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "Restrictions":{ - "type":"structure", - "required":["GeoRestriction"], - "members":{ - "GeoRestriction":{"shape":"GeoRestriction"} - } - }, - "S3Origin":{ - "type":"structure", - "required":[ - "DomainName", - "OriginAccessIdentity" - ], - "members":{ - "DomainName":{"shape":"string"}, - "OriginAccessIdentity":{"shape":"string"} - } - }, - "S3OriginConfig":{ - "type":"structure", - "required":["OriginAccessIdentity"], - "members":{ - "OriginAccessIdentity":{"shape":"string"} - } - }, - "SSLSupportMethod":{ - "type":"string", - "enum":[ - "sni-only", - "vip" - ] - }, - "Signer":{ - "type":"structure", - "members":{ - "AwsAccountNumber":{"shape":"string"}, - "KeyPairIds":{"shape":"KeyPairIds"} - } - }, - "SignerList":{ - "type":"list", - "member":{ - "shape":"Signer", - "locationName":"Signer" - } - }, - "SslProtocol":{ - "type":"string", - "enum":[ - "SSLv3", - "TLSv1", - "TLSv1.1", - "TLSv1.2" - ] - }, - "SslProtocolsList":{ - "type":"list", - "member":{ - "shape":"SslProtocol", - "locationName":"SslProtocol" - } - }, - "StreamingDistribution":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "DomainName", - "ActiveTrustedSigners", - "StreamingDistributionConfig" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "ActiveTrustedSigners":{"shape":"ActiveTrustedSigners"}, - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"} - } - }, - "StreamingDistributionAlreadyExists":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionConfig":{ - "type":"structure", - "required":[ - "CallerReference", - "S3Origin", - "Comment", - "TrustedSigners", - "Enabled" - ], - "members":{ - "CallerReference":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "Comment":{"shape":"string"}, - "Logging":{"shape":"StreamingLoggingConfig"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionConfigWithTags":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Tags" - ], - "members":{ - "StreamingDistributionConfig":{"shape":"StreamingDistributionConfig"}, - "Tags":{"shape":"Tags"} - } - }, - "StreamingDistributionList":{ - "type":"structure", - "required":[ - "Marker", - "MaxItems", - "IsTruncated", - "Quantity" - ], - "members":{ - "Marker":{"shape":"string"}, - "NextMarker":{"shape":"string"}, - "MaxItems":{"shape":"integer"}, - "IsTruncated":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"StreamingDistributionSummaryList"} - } - }, - "StreamingDistributionNotDisabled":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, - "StreamingDistributionSummary":{ - "type":"structure", - "required":[ - "Id", - "ARN", - "Status", - "LastModifiedTime", - "DomainName", - "S3Origin", - "Aliases", - "TrustedSigners", - "Comment", - "PriceClass", - "Enabled" - ], - "members":{ - "Id":{"shape":"string"}, - "ARN":{"shape":"string"}, - "Status":{"shape":"string"}, - "LastModifiedTime":{"shape":"timestamp"}, - "DomainName":{"shape":"string"}, - "S3Origin":{"shape":"S3Origin"}, - "Aliases":{"shape":"Aliases"}, - "TrustedSigners":{"shape":"TrustedSigners"}, - "Comment":{"shape":"string"}, - "PriceClass":{"shape":"PriceClass"}, - "Enabled":{"shape":"boolean"} - } - }, - "StreamingDistributionSummaryList":{ - "type":"list", - "member":{ - "shape":"StreamingDistributionSummary", - "locationName":"StreamingDistributionSummary" - } - }, - "StreamingLoggingConfig":{ - "type":"structure", - "required":[ - "Enabled", - "Bucket", - "Prefix" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Bucket":{"shape":"string"}, - "Prefix":{"shape":"string"} - } - }, - "Tag":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{ - "type":"string", - "max":128, - "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "TagKeyList":{ - "type":"list", - "member":{ - "shape":"TagKey", - "locationName":"Key" - } - }, - "TagKeys":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagKeyList"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "Tags" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "Tags":{ - "shape":"Tags", - "locationName":"Tags", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"Tags" - }, - "TagValue":{ - "type":"string", - "max":256, - "min":0, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" - }, - "Tags":{ - "type":"structure", - "members":{ - "Items":{"shape":"TagList"} - } - }, - "TooManyCacheBehaviors":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCertificates":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCloudFrontOriginAccessIdentities":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyCookieNamesInWhiteList":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyDistributionsWithLambdaAssociations":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyHeadersInForwardedValues":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyInvalidationsInProgress":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyLambdaFunctionAssociations":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOriginCustomHeaders":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyOrigins":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyQueryStringParameters":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributionCNAMEs":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyStreamingDistributions":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TooManyTrustedSigners":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSignerDoesNotExist":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "TrustedSigners":{ - "type":"structure", - "required":[ - "Enabled", - "Quantity" - ], - "members":{ - "Enabled":{"shape":"boolean"}, - "Quantity":{"shape":"integer"}, - "Items":{"shape":"AwsAccountNumberList"} - } - }, - "UntagResourceRequest":{ - "type":"structure", - "required":[ - "Resource", - "TagKeys" - ], - "members":{ - "Resource":{ - "shape":"ResourceARN", - "location":"querystring", - "locationName":"Resource" - }, - "TagKeys":{ - "shape":"TagKeys", - "locationName":"TagKeys", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - } - }, - "payload":"TagKeys" - }, - "UpdateCloudFrontOriginAccessIdentityRequest":{ - "type":"structure", - "required":[ - "CloudFrontOriginAccessIdentityConfig", - "Id" - ], - "members":{ - "CloudFrontOriginAccessIdentityConfig":{ - "shape":"CloudFrontOriginAccessIdentityConfig", - "locationName":"CloudFrontOriginAccessIdentityConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"CloudFrontOriginAccessIdentityConfig" - }, - "UpdateCloudFrontOriginAccessIdentityResult":{ - "type":"structure", - "members":{ - "CloudFrontOriginAccessIdentity":{"shape":"CloudFrontOriginAccessIdentity"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"CloudFrontOriginAccessIdentity" - }, - "UpdateDistributionRequest":{ - "type":"structure", - "required":[ - "DistributionConfig", - "Id" - ], - "members":{ - "DistributionConfig":{ - "shape":"DistributionConfig", - "locationName":"DistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"DistributionConfig" - }, - "UpdateDistributionResult":{ - "type":"structure", - "members":{ - "Distribution":{"shape":"Distribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"Distribution" - }, - "UpdateStreamingDistributionRequest":{ - "type":"structure", - "required":[ - "StreamingDistributionConfig", - "Id" - ], - "members":{ - "StreamingDistributionConfig":{ - "shape":"StreamingDistributionConfig", - "locationName":"StreamingDistributionConfig", - "xmlNamespace":{"uri":"http://cloudfront.amazonaws.com/doc/2017-03-25/"} - }, - "Id":{ - "shape":"string", - "location":"uri", - "locationName":"Id" - }, - "IfMatch":{ - "shape":"string", - "location":"header", - "locationName":"If-Match" - } - }, - "payload":"StreamingDistributionConfig" - }, - "UpdateStreamingDistributionResult":{ - "type":"structure", - "members":{ - "StreamingDistribution":{"shape":"StreamingDistribution"}, - "ETag":{ - "shape":"string", - "location":"header", - "locationName":"ETag" - } - }, - "payload":"StreamingDistribution" - }, - "ViewerCertificate":{ - "type":"structure", - "members":{ - "CloudFrontDefaultCertificate":{"shape":"boolean"}, - "IAMCertificateId":{"shape":"string"}, - "ACMCertificateArn":{"shape":"string"}, - "SSLSupportMethod":{"shape":"SSLSupportMethod"}, - "MinimumProtocolVersion":{"shape":"MinimumProtocolVersion"}, - "Certificate":{ - "shape":"string", - "deprecated":true - }, - "CertificateSource":{ - "shape":"CertificateSource", - "deprecated":true - } - } - }, - "ViewerProtocolPolicy":{ - "type":"string", - "enum":[ - "allow-all", - "https-only", - "redirect-to-https" - ] - }, - "boolean":{"type":"boolean"}, - "integer":{"type":"integer"}, - "long":{"type":"long"}, - "string":{"type":"string"}, - "timestamp":{"type":"timestamp"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/docs-2.json deleted file mode 100644 index c0e86083c..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/docs-2.json +++ /dev/null @@ -1,1462 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon CloudFront

This is the Amazon CloudFront API Reference. This guide is for developers who need detailed information about CloudFront API actions, data types, and errors. For detailed information about CloudFront features, see the Amazon CloudFront Developer Guide.

", - "operations": { - "CreateCloudFrontOriginAccessIdentity": "

Creates a new origin access identity. If you're using Amazon S3 for your origin, you can use an origin access identity to require users to access your content using a CloudFront URL instead of the Amazon S3 URL. For more information about how to use origin access identities, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "CreateDistribution": "

Creates a new web distribution. Send a POST request to the /CloudFront API version/distribution/distribution ID resource.

", - "CreateDistributionWithTags": "

Create a new distribution with tags.

", - "CreateInvalidation": "

Create a new invalidation.

", - "CreateStreamingDistribution": "

Creates a new RMTP distribution. An RTMP distribution is similar to a web distribution, but an RTMP distribution streams media files using the Adobe Real-Time Messaging Protocol (RTMP) instead of serving files using HTTP.

To create a new web distribution, submit a POST request to the CloudFront API version/distribution resource. The request body must include a document with a StreamingDistributionConfig element. The response echoes the StreamingDistributionConfig element and returns other information about the RTMP distribution.

To get the status of your request, use the GET StreamingDistribution API action. When the value of Enabled is true and the value of Status is Deployed, your distribution is ready. A distribution usually deploys in less than 15 minutes.

For more information about web distributions, see Working with RTMP Distributions in the Amazon CloudFront Developer Guide.

Beginning with the 2012-05-05 version of the CloudFront API, we made substantial changes to the format of the XML document that you include in the request body when you create or update a web distribution or an RTMP distribution, and when you invalidate objects. With previous versions of the API, we discovered that it was too easy to accidentally delete one or more values for an element that accepts multiple values, for example, CNAMEs and trusted signers. Our changes for the 2012-05-05 release are intended to prevent these accidental deletions and to notify you when there's a mismatch between the number of values you say you're specifying in the Quantity element and the number of values specified.

", - "CreateStreamingDistributionWithTags": "

Create a new streaming distribution with tags.

", - "DeleteCloudFrontOriginAccessIdentity": "

Delete an origin access identity.

", - "DeleteDistribution": "

Delete a distribution.

", - "DeleteServiceLinkedRole": null, - "DeleteStreamingDistribution": "

Delete a streaming distribution. To delete an RTMP distribution using the CloudFront API, perform the following steps.

To delete an RTMP distribution using the CloudFront API:

  1. Disable the RTMP distribution.

  2. Submit a GET Streaming Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Streaming Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Streaming Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Then set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  5. Review the response to the PUT Streaming Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Streaming Distribution Config request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Streaming Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  8. Review the response to your DELETE Streaming Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", - "GetCloudFrontOriginAccessIdentity": "

Get the information about an origin access identity.

", - "GetCloudFrontOriginAccessIdentityConfig": "

Get the configuration information about an origin access identity.

", - "GetDistribution": "

Get the information about a distribution.

", - "GetDistributionConfig": "

Get the configuration information about a distribution.

", - "GetInvalidation": "

Get the information about an invalidation.

", - "GetStreamingDistribution": "

Gets information about a specified RTMP distribution, including the distribution configuration.

", - "GetStreamingDistributionConfig": "

Get the configuration information about a streaming distribution.

", - "ListCloudFrontOriginAccessIdentities": "

Lists origin access identities.

", - "ListDistributions": "

List distributions.

", - "ListDistributionsByWebACLId": "

List the distributions that are associated with a specified AWS WAF web ACL.

", - "ListInvalidations": "

Lists invalidation batches.

", - "ListStreamingDistributions": "

List streaming distributions.

", - "ListTagsForResource": "

List tags for a CloudFront resource.

", - "TagResource": "

Add tags to a CloudFront resource.

", - "UntagResource": "

Remove tags from a CloudFront resource.

", - "UpdateCloudFrontOriginAccessIdentity": "

Update an origin access identity.

", - "UpdateDistribution": "

Updates the configuration for a web distribution. Perform the following steps.

For information about updating a distribution using the CloudFront console, see Creating or Updating a Web Distribution Using the CloudFront Console in the Amazon CloudFront Developer Guide.

To update a web distribution using the CloudFront API

  1. Submit a GetDistributionConfig request to get the current configuration and an Etag header for the distribution.

    If you update the distribution again, you need to get a new Etag header.

  2. Update the XML document that was returned in the response to your GetDistributionConfig request to include the desired changes. You can't change the value of CallerReference. If you try to change this value, CloudFront returns an IllegalUpdate error.

    The new configuration replaces the existing configuration; the values that you specify in an UpdateDistribution request are not merged into the existing configuration. When you add, delete, or replace values in an element that allows multiple values (for example, CNAME), you must specify all of the values that you want to appear in the updated distribution. In addition, you must update the corresponding Quantity element.

  3. Submit an UpdateDistribution request to update the configuration for your distribution:

    • In the request body, include the XML document that you updated in Step 2. The request body must include an XML document with a DistributionConfig element.

    • Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GetDistributionConfig request in Step 1.

  4. Review the response to the UpdateDistribution request to confirm that the configuration was successfully updated.

  5. Optional: Submit a GetDistribution request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

    Beginning with the 2012-05-05 version of the CloudFront API, we made substantial changes to the format of the XML document that you include in the request body when you create or update a distribution. With previous versions of the API, we discovered that it was too easy to accidentally delete one or more values for an element that accepts multiple values, for example, CNAMEs and trusted signers. Our changes for the 2012-05-05 release are intended to prevent these accidental deletions and to notify you when there's a mismatch between the number of values you say you're specifying in the Quantity element and the number of values you're actually specifying.

", - "UpdateStreamingDistribution": "

Update a streaming distribution.

" - }, - "shapes": { - "AccessDenied": { - "base": "

Access denied.

", - "refs": { - } - }, - "ActiveTrustedSigners": { - "base": "

A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for this distribution. These are the accounts that you want to allow to create signed URLs for private content.

The Signer complex type lists the AWS account number of the trusted signer or self if the signer is the AWS account that created the distribution. The Signer element also includes the IDs of any active CloudFront key pairs that are associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create signed URLs.

For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "refs": { - "Distribution$ActiveTrustedSigners": "

CloudFront automatically adds this element to the response only if you've set up the distribution to serve private content with signed URLs. The element lists the key pair IDs that CloudFront is aware of for each trusted signer. The Signer child element lists the AWS account number of the trusted signer (or an empty Self element if the signer is you). The Signer element also includes the IDs of any active key pairs associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create working signed URLs.

", - "StreamingDistribution$ActiveTrustedSigners": "

A complex type that lists the AWS accounts, if any, that you included in the TrustedSigners complex type for this distribution. These are the accounts that you want to allow to create signed URLs for private content.

The Signer complex type lists the AWS account number of the trusted signer or self if the signer is the AWS account that created the distribution. The Signer element also includes the IDs of any active CloudFront key pairs that are associated with the trusted signer's AWS account. If no KeyPairId element appears for a Signer, that signer can't create signed URLs.

For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

" - } - }, - "AliasList": { - "base": null, - "refs": { - "Aliases$Items": "

A complex type that contains the CNAME aliases, if any, that you want to associate with this distribution.

" - } - }, - "Aliases": { - "base": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "refs": { - "DistributionConfig$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "DistributionSummary$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this distribution.

", - "StreamingDistributionConfig$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.

", - "StreamingDistributionSummary$Aliases": "

A complex type that contains information about CNAMEs (alternate domain names), if any, for this streaming distribution.

" - } - }, - "AllowedMethods": { - "base": "

A complex type that controls which HTTP methods CloudFront processes and forwards to your Amazon S3 bucket or your custom origin. There are three choices:

  • CloudFront forwards only GET and HEAD requests.

  • CloudFront forwards only GET, HEAD, and OPTIONS requests.

  • CloudFront forwards GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests.

If you pick the third choice, you may need to restrict access to your Amazon S3 bucket or to your custom origin so users can't perform operations that you don't want them to. For example, you might not want users to have permissions to delete objects from your origin.

", - "refs": { - "CacheBehavior$AllowedMethods": null, - "DefaultCacheBehavior$AllowedMethods": null - } - }, - "AwsAccountNumberList": { - "base": null, - "refs": { - "TrustedSigners$Items": "

Optional: A complex type that contains trusted signers for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "BatchTooLarge": { - "base": null, - "refs": { - } - }, - "CNAMEAlreadyExists": { - "base": null, - "refs": { - } - }, - "CacheBehavior": { - "base": "

A complex type that describes how CloudFront processes requests.

You must create at least as many cache behaviors (including the default cache behavior) as you have origins if you want CloudFront to distribute objects from all of the origins. Each cache behavior specifies the one origin from which you want CloudFront to get objects. If you have two origins and only the default cache behavior, the default cache behavior will cause CloudFront to get objects from one of the origins, but the other origin is never used.

For the current limit on the number of cache behaviors that you can add to a distribution, see Amazon CloudFront Limits in the AWS General Reference.

If you don't want to specify any cache behaviors, include only an empty CacheBehaviors element. Don't include an empty CacheBehavior element, or CloudFront returns a MalformedXML error.

To delete all cache behaviors in an existing distribution, update the distribution configuration and include only an empty CacheBehaviors element.

To add, change, or remove one or more cache behaviors, update the distribution configuration and specify all of the cache behaviors that you want to include in the updated distribution.

For more information about cache behaviors, see Cache Behaviors in the Amazon CloudFront Developer Guide.

", - "refs": { - "CacheBehaviorList$member": null - } - }, - "CacheBehaviorList": { - "base": null, - "refs": { - "CacheBehaviors$Items": "

Optional: A complex type that contains cache behaviors for this distribution. If Quantity is 0, you can omit Items.

" - } - }, - "CacheBehaviors": { - "base": "

A complex type that contains zero or more CacheBehavior elements.

", - "refs": { - "DistributionConfig$CacheBehaviors": "

A complex type that contains zero or more CacheBehavior elements.

", - "DistributionSummary$CacheBehaviors": "

A complex type that contains zero or more CacheBehavior elements.

" - } - }, - "CachedMethods": { - "base": "

A complex type that controls whether CloudFront caches the response to requests using the specified HTTP methods. There are two choices:

  • CloudFront caches responses to GET and HEAD requests.

  • CloudFront caches responses to GET, HEAD, and OPTIONS requests.

If you pick the second choice for your Amazon S3 Origin, you may need to forward Access-Control-Request-Method, Access-Control-Request-Headers, and Origin headers for the responses to be cached correctly.

", - "refs": { - "AllowedMethods$CachedMethods": null - } - }, - "CertificateSource": { - "base": null, - "refs": { - "ViewerCertificate$CertificateSource": "

This field has been deprecated. Use one of the following fields instead:

" - } - }, - "CloudFrontOriginAccessIdentity": { - "base": "

CloudFront origin access identity.

", - "refs": { - "CreateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

", - "GetCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

", - "UpdateCloudFrontOriginAccessIdentityResult$CloudFrontOriginAccessIdentity": "

The origin access identity's information.

" - } - }, - "CloudFrontOriginAccessIdentityAlreadyExists": { - "base": "

If the CallerReference is a value you already sent in a previous request to create an identity but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.

", - "refs": { - } - }, - "CloudFrontOriginAccessIdentityConfig": { - "base": "

Origin access identity configuration. Send a GET request to the /CloudFront API version/CloudFront/identity ID/config resource.

", - "refs": { - "CloudFrontOriginAccessIdentity$CloudFrontOriginAccessIdentityConfig": "

The current configuration information for the identity.

", - "CreateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "

The current configuration information for the identity.

", - "GetCloudFrontOriginAccessIdentityConfigResult$CloudFrontOriginAccessIdentityConfig": "

The origin access identity's configuration information.

", - "UpdateCloudFrontOriginAccessIdentityRequest$CloudFrontOriginAccessIdentityConfig": "

The identity's configuration information.

" - } - }, - "CloudFrontOriginAccessIdentityInUse": { - "base": null, - "refs": { - } - }, - "CloudFrontOriginAccessIdentityList": { - "base": "

Lists the origin access identities for CloudFront.Send a GET request to the /CloudFront API version/origin-access-identity/cloudfront resource. The response includes a CloudFrontOriginAccessIdentityList element with zero or more CloudFrontOriginAccessIdentitySummary child elements. By default, your entire list of origin access identities is returned in one single page. If the list is long, you can paginate it using the MaxItems and Marker parameters.

", - "refs": { - "ListCloudFrontOriginAccessIdentitiesResult$CloudFrontOriginAccessIdentityList": "

The CloudFrontOriginAccessIdentityList type.

" - } - }, - "CloudFrontOriginAccessIdentitySummary": { - "base": "

Summary of the information about a CloudFront origin access identity.

", - "refs": { - "CloudFrontOriginAccessIdentitySummaryList$member": null - } - }, - "CloudFrontOriginAccessIdentitySummaryList": { - "base": null, - "refs": { - "CloudFrontOriginAccessIdentityList$Items": "

A complex type that contains one CloudFrontOriginAccessIdentitySummary element for each origin access identity that was created by the current AWS account.

" - } - }, - "CookieNameList": { - "base": null, - "refs": { - "CookieNames$Items": "

A complex type that contains one Name element for each cookie that you want CloudFront to forward to the origin for this cache behavior.

" - } - }, - "CookieNames": { - "base": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

", - "refs": { - "CookiePreference$WhitelistedNames": "

Required if you specify whitelist for the value of Forward:. A complex type that specifies how many different cookies you want CloudFront to forward to the origin for this cache behavior and, if you want to forward selected cookies, the names of those cookies.

If you specify all or none for the value of Forward, omit WhitelistedNames. If you change the value of Forward from whitelist to all or none and you don't delete the WhitelistedNames element and its child elements, CloudFront deletes them automatically.

For the current limit on the number of cookie names that you can whitelist for each cache behavior, see Amazon CloudFront Limits in the AWS General Reference.

" - } - }, - "CookiePreference": { - "base": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

", - "refs": { - "ForwardedValues$Cookies": "

A complex type that specifies whether you want CloudFront to forward cookies to the origin and, if so, which ones. For more information about forwarding cookies to the origin, see How CloudFront Forwards, Caches, and Logs Cookies in the Amazon CloudFront Developer Guide.

" - } - }, - "CreateCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to create a new origin access identity.

", - "refs": { - } - }, - "CreateCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateDistributionRequest": { - "base": "

The request to create a new distribution.

", - "refs": { - } - }, - "CreateDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateDistributionWithTagsRequest": { - "base": "

The request to create a new distribution with tags.

", - "refs": { - } - }, - "CreateDistributionWithTagsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateInvalidationRequest": { - "base": "

The request to create an invalidation.

", - "refs": { - } - }, - "CreateInvalidationResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateStreamingDistributionRequest": { - "base": "

The request to create a new streaming distribution.

", - "refs": { - } - }, - "CreateStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsRequest": { - "base": "

The request to create a new streaming distribution with tags.

", - "refs": { - } - }, - "CreateStreamingDistributionWithTagsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "CustomErrorResponse": { - "base": "

A complex type that controls:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "refs": { - "CustomErrorResponseList$member": null - } - }, - "CustomErrorResponseList": { - "base": null, - "refs": { - "CustomErrorResponses$Items": "

A complex type that contains a CustomErrorResponse element for each HTTP status code for which you want to specify a custom error page and/or a caching duration.

" - } - }, - "CustomErrorResponses": { - "base": "

A complex type that controls:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "refs": { - "DistributionConfig$CustomErrorResponses": "

A complex type that controls the following:

  • Whether CloudFront replaces HTTP status codes in the 4xx and 5xx range with custom error messages before returning the response to the viewer.

  • How long CloudFront caches HTTP status codes in the 4xx and 5xx range.

For more information about custom error pages, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "DistributionSummary$CustomErrorResponses": "

A complex type that contains zero or more CustomErrorResponses elements.

" - } - }, - "CustomHeaders": { - "base": "

A complex type that contains the list of Custom Headers for each origin.

", - "refs": { - "Origin$CustomHeaders": "

A complex type that contains names and values for the custom headers that you want.

" - } - }, - "CustomOriginConfig": { - "base": "

A customer origin.

", - "refs": { - "Origin$CustomOriginConfig": "

A complex type that contains information about a custom origin. If the origin is an Amazon S3 bucket, use the S3OriginConfig element instead.

" - } - }, - "DefaultCacheBehavior": { - "base": "

A complex type that describes the default cache behavior if you don't specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

", - "refs": { - "DistributionConfig$DefaultCacheBehavior": "

A complex type that describes the default cache behavior if you don't specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

", - "DistributionSummary$DefaultCacheBehavior": "

A complex type that describes the default cache behavior if you don't specify a CacheBehavior element or if files don't match any of the values of PathPattern in CacheBehavior elements. You must create exactly one default cache behavior.

" - } - }, - "DeleteCloudFrontOriginAccessIdentityRequest": { - "base": "

Deletes a origin access identity.

", - "refs": { - } - }, - "DeleteDistributionRequest": { - "base": "

This action deletes a web distribution. To delete a web distribution using the CloudFront API, perform the following steps.

To delete a web distribution using the CloudFront API:

  1. Disable the web distribution

  2. Submit a GET Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Distribution Config request in Step 2.

  5. Review the response to the PUT Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Distribution request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Distribution Config request in Step 6.

  8. Review the response to your DELETE Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", - "refs": { - } - }, - "DeleteServiceLinkedRoleRequest": { - "base": null, - "refs": { - } - }, - "DeleteStreamingDistributionRequest": { - "base": "

The request to delete a streaming distribution.

", - "refs": { - } - }, - "Distribution": { - "base": "

The distribution's information.

", - "refs": { - "CreateDistributionResult$Distribution": "

The distribution's information.

", - "CreateDistributionWithTagsResult$Distribution": "

The distribution's information.

", - "GetDistributionResult$Distribution": "

The distribution's information.

", - "UpdateDistributionResult$Distribution": "

The distribution's information.

" - } - }, - "DistributionAlreadyExists": { - "base": "

The caller reference you attempted to create the distribution with is associated with another distribution.

", - "refs": { - } - }, - "DistributionConfig": { - "base": "

A distribution configuration.

", - "refs": { - "CreateDistributionRequest$DistributionConfig": "

The distribution's configuration information.

", - "Distribution$DistributionConfig": "

The current configuration information for the distribution. Send a GET request to the /CloudFront API version/distribution ID/config resource.

", - "DistributionConfigWithTags$DistributionConfig": "

A distribution configuration.

", - "GetDistributionConfigResult$DistributionConfig": "

The distribution's configuration information.

", - "UpdateDistributionRequest$DistributionConfig": "

The distribution's configuration information.

" - } - }, - "DistributionConfigWithTags": { - "base": "

A distribution Configuration and a list of tags to be associated with the distribution.

", - "refs": { - "CreateDistributionWithTagsRequest$DistributionConfigWithTags": "

The distribution's configuration information.

" - } - }, - "DistributionList": { - "base": "

A distribution list.

", - "refs": { - "ListDistributionsByWebACLIdResult$DistributionList": "

The DistributionList type.

", - "ListDistributionsResult$DistributionList": "

The DistributionList type.

" - } - }, - "DistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "DistributionSummary": { - "base": "

A summary of the information about a CloudFront distribution.

", - "refs": { - "DistributionSummaryList$member": null - } - }, - "DistributionSummaryList": { - "base": null, - "refs": { - "DistributionList$Items": "

A complex type that contains one DistributionSummary element for each distribution that was created by the current AWS account.

" - } - }, - "EventType": { - "base": null, - "refs": { - "LambdaFunctionAssociation$EventType": "

Specifies the event type that triggers a Lambda function invocation. You can specify the following values:

  • viewer-request: The function executes when CloudFront receives a request from a viewer and before it checks to see whether the requested object is in the edge cache.

  • origin-request: The function executes only when CloudFront forwards a request to your origin. When the requested object is in the edge cache, the function doesn't execute.

  • origin-response: The function executes after CloudFront receives a response from the origin and before it caches the object in the response. When the requested object is in the edge cache, the function doesn't execute.

    If the origin returns an HTTP status code other than HTTP 200 (OK), the function doesn't execute.

  • viewer-response: The function executes before CloudFront returns the requested object to the viewer. The function executes regardless of whether the object was already in the edge cache.

    If the origin returns an HTTP status code other than HTTP 200 (OK), the function doesn't execute.

" - } - }, - "ForwardedValues": { - "base": "

A complex type that specifies how CloudFront handles query strings and cookies.

", - "refs": { - "CacheBehavior$ForwardedValues": "

A complex type that specifies how CloudFront handles query strings and cookies.

", - "DefaultCacheBehavior$ForwardedValues": "

A complex type that specifies how CloudFront handles query strings and cookies.

" - } - }, - "GeoRestriction": { - "base": "

A complex type that controls the countries in which your content is distributed. CloudFront determines the location of your users using MaxMind GeoIP databases.

", - "refs": { - "Restrictions$GeoRestriction": null - } - }, - "GeoRestrictionType": { - "base": null, - "refs": { - "GeoRestriction$RestrictionType": "

The method that you want to use to restrict distribution of your content by country:

  • none: No geo restriction is enabled, meaning access to content is not restricted by client geo location.

  • blacklist: The Location elements specify the countries in which you don't want CloudFront to distribute your content.

  • whitelist: The Location elements specify the countries in which you want CloudFront to distribute your content.

" - } - }, - "GetCloudFrontOriginAccessIdentityConfigRequest": { - "base": "

The origin access identity's configuration information. For more information, see CloudFrontOriginAccessIdentityConfigComplexType.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to get an origin access identity's information.

", - "refs": { - } - }, - "GetCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetDistributionConfigRequest": { - "base": "

The request to get a distribution configuration.

", - "refs": { - } - }, - "GetDistributionConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetDistributionRequest": { - "base": "

The request to get a distribution's information.

", - "refs": { - } - }, - "GetDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetInvalidationRequest": { - "base": "

The request to get an invalidation's information.

", - "refs": { - } - }, - "GetInvalidationResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetStreamingDistributionConfigRequest": { - "base": "

To request to get a streaming distribution configuration.

", - "refs": { - } - }, - "GetStreamingDistributionConfigResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "GetStreamingDistributionRequest": { - "base": "

The request to get a streaming distribution's information.

", - "refs": { - } - }, - "GetStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "HeaderList": { - "base": null, - "refs": { - "Headers$Items": "

A list that contains one Name element for each header that you want CloudFront to use for caching in this cache behavior. If Quantity is 0, omit Items.

" - } - }, - "Headers": { - "base": "

A complex type that specifies the request headers, if any, that you want CloudFront to base caching on for this cache behavior.

For the headers that you specify, CloudFront caches separate versions of a specified object based on the header values in viewer requests. For example, suppose viewer requests for logo.jpg contain a custom product header that has a value of either acme or apex, and you configure CloudFront to cache your content based on values in the product header. CloudFront forwards the product header to the origin and caches the response from the origin once for each header value. For more information about caching based on header values, see How CloudFront Forwards and Caches Headers in the Amazon CloudFront Developer Guide.

", - "refs": { - "ForwardedValues$Headers": "

A complex type that specifies the Headers, if any, that you want CloudFront to base caching on for this cache behavior.

" - } - }, - "HttpVersion": { - "base": null, - "refs": { - "DistributionConfig$HttpVersion": "

(Optional) Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 automatically use an earlier HTTP version.

For viewers and CloudFront to use HTTP/2, viewers must support TLS 1.2 or later, and must support Server Name Identification (SNI).

In general, configuring CloudFront to communicate with viewers using HTTP/2 reduces latency. You can improve performance by optimizing for HTTP/2. For more information, do an Internet search for \"http/2 optimization.\"

", - "DistributionSummary$HttpVersion": "

Specify the maximum HTTP version that you want viewers to use to communicate with CloudFront. The default value for new web distributions is http2. Viewers that don't support HTTP/2 will automatically use an earlier version.

" - } - }, - "IllegalUpdate": { - "base": "

Origin and CallerReference cannot be updated.

", - "refs": { - } - }, - "InconsistentQuantities": { - "base": "

The value of Quantity and the size of Items don't match.

", - "refs": { - } - }, - "InvalidArgument": { - "base": "

The argument is invalid.

", - "refs": { - } - }, - "InvalidDefaultRootObject": { - "base": "

The default root object file name is too big or contains an invalid character.

", - "refs": { - } - }, - "InvalidErrorCode": { - "base": null, - "refs": { - } - }, - "InvalidForwardCookies": { - "base": "

Your request contains forward cookies option which doesn't match with the expectation for the whitelisted list of cookie names. Either list of cookie names has been specified when not allowed or list of cookie names is missing when expected.

", - "refs": { - } - }, - "InvalidGeoRestrictionParameter": { - "base": null, - "refs": { - } - }, - "InvalidHeadersForS3Origin": { - "base": null, - "refs": { - } - }, - "InvalidIfMatchVersion": { - "base": "

The If-Match version is missing or not valid for the distribution.

", - "refs": { - } - }, - "InvalidLambdaFunctionAssociation": { - "base": "

The specified Lambda function association is invalid.

", - "refs": { - } - }, - "InvalidLocationCode": { - "base": null, - "refs": { - } - }, - "InvalidMinimumProtocolVersion": { - "base": null, - "refs": { - } - }, - "InvalidOrigin": { - "base": "

The Amazon S3 origin server specified does not refer to a valid Amazon S3 bucket.

", - "refs": { - } - }, - "InvalidOriginAccessIdentity": { - "base": "

The origin access identity is not valid or doesn't exist.

", - "refs": { - } - }, - "InvalidOriginKeepaliveTimeout": { - "base": null, - "refs": { - } - }, - "InvalidOriginReadTimeout": { - "base": null, - "refs": { - } - }, - "InvalidProtocolSettings": { - "base": "

You cannot specify SSLv3 as the minimum protocol version if you only want to support only clients that support Server Name Indication (SNI).

", - "refs": { - } - }, - "InvalidQueryStringParameters": { - "base": null, - "refs": { - } - }, - "InvalidRelativePath": { - "base": "

The relative path is too big, is not URL-encoded, or does not begin with a slash (/).

", - "refs": { - } - }, - "InvalidRequiredProtocol": { - "base": "

This operation requires the HTTPS protocol. Ensure that you specify the HTTPS protocol in your request, or omit the RequiredProtocols element from your distribution configuration.

", - "refs": { - } - }, - "InvalidResponseCode": { - "base": null, - "refs": { - } - }, - "InvalidTTLOrder": { - "base": null, - "refs": { - } - }, - "InvalidTagging": { - "base": null, - "refs": { - } - }, - "InvalidViewerCertificate": { - "base": null, - "refs": { - } - }, - "InvalidWebACLId": { - "base": null, - "refs": { - } - }, - "Invalidation": { - "base": "

An invalidation.

", - "refs": { - "CreateInvalidationResult$Invalidation": "

The invalidation's information.

", - "GetInvalidationResult$Invalidation": "

The invalidation's information. For more information, see Invalidation Complex Type.

" - } - }, - "InvalidationBatch": { - "base": "

An invalidation batch.

", - "refs": { - "CreateInvalidationRequest$InvalidationBatch": "

The batch information for the invalidation.

", - "Invalidation$InvalidationBatch": "

The current invalidation information for the batch request.

" - } - }, - "InvalidationList": { - "base": "

The InvalidationList complex type describes the list of invalidation objects. For more information about invalidation, see Invalidating Objects (Web Distributions Only) in the Amazon CloudFront Developer Guide.

", - "refs": { - "ListInvalidationsResult$InvalidationList": "

Information about invalidation batches.

" - } - }, - "InvalidationSummary": { - "base": "

A summary of an invalidation request.

", - "refs": { - "InvalidationSummaryList$member": null - } - }, - "InvalidationSummaryList": { - "base": null, - "refs": { - "InvalidationList$Items": "

A complex type that contains one InvalidationSummary element for each invalidation batch created by the current AWS account.

" - } - }, - "ItemSelection": { - "base": null, - "refs": { - "CookiePreference$Forward": "

Specifies which cookies to forward to the origin for this cache behavior: all, none, or the list of cookies specified in the WhitelistedNames complex type.

Amazon S3 doesn't process cookies. When the cache behavior is forwarding requests to an Amazon S3 origin, specify none for the Forward element.

" - } - }, - "KeyPairIdList": { - "base": null, - "refs": { - "KeyPairIds$Items": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

For more information, see ActiveTrustedSigners.

" - } - }, - "KeyPairIds": { - "base": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

For more information, see ActiveTrustedSigners.

", - "refs": { - "Signer$KeyPairIds": "

A complex type that lists the active CloudFront key pairs, if any, that are associated with AwsAccountNumber.

" - } - }, - "LambdaFunctionAssociation": { - "base": "

A complex type that contains a Lambda function association.

", - "refs": { - "LambdaFunctionAssociationList$member": null - } - }, - "LambdaFunctionAssociationList": { - "base": null, - "refs": { - "LambdaFunctionAssociations$Items": "

Optional: A complex type that contains LambdaFunctionAssociation items for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "LambdaFunctionAssociations": { - "base": "

A complex type that specifies a list of Lambda functions associations for a cache behavior.

If you want to invoke one or more Lambda functions triggered by requests that match the PathPattern of the cache behavior, specify the applicable values for Quantity and Items. Note that there can be up to 4 LambdaFunctionAssociation items in this list (one for each possible value of EventType) and each EventType can be associated with the Lambda function only once.

If you don't want to invoke any Lambda functions for the requests that match PathPattern, specify 0 for Quantity and omit Items.

", - "refs": { - "CacheBehavior$LambdaFunctionAssociations": "

A complex type that contains zero or more Lambda function associations for a cache behavior.

", - "DefaultCacheBehavior$LambdaFunctionAssociations": "

A complex type that contains zero or more Lambda function associations for a cache behavior.

" - } - }, - "ListCloudFrontOriginAccessIdentitiesRequest": { - "base": "

The request to list origin access identities.

", - "refs": { - } - }, - "ListCloudFrontOriginAccessIdentitiesResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListDistributionsByWebACLIdRequest": { - "base": "

The request to list distributions that are associated with a specified AWS WAF web ACL.

", - "refs": { - } - }, - "ListDistributionsByWebACLIdResult": { - "base": "

The response to a request to list the distributions that are associated with a specified AWS WAF web ACL.

", - "refs": { - } - }, - "ListDistributionsRequest": { - "base": "

The request to list your distributions.

", - "refs": { - } - }, - "ListDistributionsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListInvalidationsRequest": { - "base": "

The request to list invalidations.

", - "refs": { - } - }, - "ListInvalidationsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListStreamingDistributionsRequest": { - "base": "

The request to list your streaming distributions.

", - "refs": { - } - }, - "ListStreamingDistributionsResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": "

The request to list tags for a CloudFront resource.

", - "refs": { - } - }, - "ListTagsForResourceResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "LocationList": { - "base": null, - "refs": { - "GeoRestriction$Items": "

A complex type that contains a Location element for each country in which you want CloudFront either to distribute your content (whitelist) or not distribute your content (blacklist).

The Location element is a two-letter, uppercase country code for a country that you want to include in your blacklist or whitelist. Include one Location element for each country.

CloudFront and MaxMind both use ISO 3166 country codes. For the current list of countries and the corresponding codes, see ISO 3166-1-alpha-2 code on the International Organization for Standardization website. You can also refer to the country list on the CloudFront console, which includes both country names and codes.

" - } - }, - "LoggingConfig": { - "base": "

A complex type that controls whether access logs are written for the distribution.

", - "refs": { - "DistributionConfig$Logging": "

A complex type that controls whether access logs are written for the distribution.

For more information about logging, see Access Logs in the Amazon CloudFront Developer Guide.

" - } - }, - "Method": { - "base": null, - "refs": { - "MethodsList$member": null - } - }, - "MethodsList": { - "base": null, - "refs": { - "AllowedMethods$Items": "

A complex type that contains the HTTP methods that you want CloudFront to process and forward to your origin.

", - "CachedMethods$Items": "

A complex type that contains the HTTP methods that you want CloudFront to cache responses to.

" - } - }, - "MinimumProtocolVersion": { - "base": null, - "refs": { - "ViewerCertificate$MinimumProtocolVersion": "

Specify the security policy that you want CloudFront to use for HTTPS connections. A security policy determines two settings:

  • The minimum SSL/TLS protocol that CloudFront uses to communicate with viewers

  • The cipher that CloudFront uses to encrypt the content that it returns to viewers

On the CloudFront console, this setting is called Security policy.

We recommend that you specify TLSv1.1_2016 unless your users are using browsers or devices that do not support TLSv1.1 or later.

When both of the following are true, you must specify TLSv1 or later for the security policy:

  • You're using a custom certificate: you specified a value for ACMCertificateArn or for IAMCertificateId

  • You're using SNI: you specified sni-only for SSLSupportMethod

If you specify true for CloudFrontDefaultCertificate, CloudFront automatically sets the security policy to TLSv1 regardless of the value that you specify for MinimumProtocolVersion.

For information about the relationship between the security policy that you choose and the protocols and ciphers that CloudFront uses to communicate with viewers, see Supported SSL/TLS Protocols and Ciphers for Communication Between Viewers and CloudFront in the Amazon CloudFront Developer Guide.

" - } - }, - "MissingBody": { - "base": "

This operation requires a body. Ensure that the body is present and the Content-Type header is set.

", - "refs": { - } - }, - "NoSuchCloudFrontOriginAccessIdentity": { - "base": "

The specified origin access identity does not exist.

", - "refs": { - } - }, - "NoSuchDistribution": { - "base": "

The specified distribution does not exist.

", - "refs": { - } - }, - "NoSuchInvalidation": { - "base": "

The specified invalidation does not exist.

", - "refs": { - } - }, - "NoSuchOrigin": { - "base": "

No origin exists with the specified Origin Id.

", - "refs": { - } - }, - "NoSuchResource": { - "base": null, - "refs": { - } - }, - "NoSuchStreamingDistribution": { - "base": "

The specified streaming distribution does not exist.

", - "refs": { - } - }, - "Origin": { - "base": "

A complex type that describes the Amazon S3 bucket or the HTTP server (for example, a web server) from which CloudFront gets your files. You must create at least one origin.

For the current limit on the number of origins that you can create for a distribution, see Amazon CloudFront Limits in the AWS General Reference.

", - "refs": { - "OriginList$member": null - } - }, - "OriginCustomHeader": { - "base": "

A complex type that contains HeaderName and HeaderValue elements, if any, for this distribution.

", - "refs": { - "OriginCustomHeadersList$member": null - } - }, - "OriginCustomHeadersList": { - "base": null, - "refs": { - "CustomHeaders$Items": "

Optional: A list that contains one OriginCustomHeader element for each custom header that you want CloudFront to forward to the origin. If Quantity is 0, omit Items.

" - } - }, - "OriginList": { - "base": null, - "refs": { - "Origins$Items": "

A complex type that contains origins for this distribution.

" - } - }, - "OriginProtocolPolicy": { - "base": null, - "refs": { - "CustomOriginConfig$OriginProtocolPolicy": "

The origin protocol policy to apply to your origin.

" - } - }, - "OriginSslProtocols": { - "base": "

A complex type that contains information about the SSL/TLS protocols that CloudFront can use when establishing an HTTPS connection with your origin.

", - "refs": { - "CustomOriginConfig$OriginSslProtocols": "

The SSL/TLS protocols that you want CloudFront to use when communicating with your origin over HTTPS.

" - } - }, - "Origins": { - "base": "

A complex type that contains information about origins for this distribution.

", - "refs": { - "DistributionConfig$Origins": "

A complex type that contains information about origins for this distribution.

", - "DistributionSummary$Origins": "

A complex type that contains information about origins for this distribution.

" - } - }, - "PathList": { - "base": null, - "refs": { - "Paths$Items": "

A complex type that contains a list of the paths that you want to invalidate.

" - } - }, - "Paths": { - "base": "

A complex type that contains information about the objects that you want to invalidate. For more information, see Specifying the Objects to Invalidate in the Amazon CloudFront Developer Guide.

", - "refs": { - "InvalidationBatch$Paths": "

A complex type that contains information about the objects that you want to invalidate. For more information, see Specifying the Objects to Invalidate in the Amazon CloudFront Developer Guide.

" - } - }, - "PreconditionFailed": { - "base": "

The precondition given in one or more of the request-header fields evaluated to false.

", - "refs": { - } - }, - "PriceClass": { - "base": null, - "refs": { - "DistributionConfig$PriceClass": "

The price class that corresponds with the maximum price that you want to pay for CloudFront service. If you specify PriceClass_All, CloudFront responds to requests for your objects from all CloudFront edge locations.

If you specify a price class other than PriceClass_All, CloudFront serves your objects from the CloudFront edge location that has the lowest latency among the edge locations in your price class. Viewers who are in or near regions that are excluded from your specified price class may encounter slower performance.

For more information about price classes, see Choosing the Price Class for a CloudFront Distribution in the Amazon CloudFront Developer Guide. For information about CloudFront pricing, including how price classes map to CloudFront regions, see Amazon CloudFront Pricing.

", - "DistributionSummary$PriceClass": null, - "StreamingDistributionConfig$PriceClass": "

A complex type that contains information about price class for this streaming distribution.

", - "StreamingDistributionSummary$PriceClass": null - } - }, - "QueryStringCacheKeys": { - "base": null, - "refs": { - "ForwardedValues$QueryStringCacheKeys": "

A complex type that contains information about the query string parameters that you want CloudFront to use for caching for this cache behavior.

" - } - }, - "QueryStringCacheKeysList": { - "base": null, - "refs": { - "QueryStringCacheKeys$Items": "

(Optional) A list that contains the query string parameters that you want CloudFront to use as a basis for caching for this cache behavior. If Quantity is 0, you can omit Items.

" - } - }, - "ResourceARN": { - "base": null, - "refs": { - "ListTagsForResourceRequest$Resource": "

An ARN of a CloudFront resource.

", - "TagResourceRequest$Resource": "

An ARN of a CloudFront resource.

", - "UntagResourceRequest$Resource": "

An ARN of a CloudFront resource.

" - } - }, - "ResourceInUse": { - "base": null, - "refs": { - } - }, - "Restrictions": { - "base": "

A complex type that identifies ways in which you want to restrict distribution of your content.

", - "refs": { - "DistributionConfig$Restrictions": null, - "DistributionSummary$Restrictions": null - } - }, - "S3Origin": { - "base": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

", - "refs": { - "StreamingDistributionConfig$S3Origin": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

", - "StreamingDistributionSummary$S3Origin": "

A complex type that contains information about the Amazon S3 bucket from which you want CloudFront to get your media files for distribution.

" - } - }, - "S3OriginConfig": { - "base": "

A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.

", - "refs": { - "Origin$S3OriginConfig": "

A complex type that contains information about the Amazon S3 origin. If the origin is a custom origin, use the CustomOriginConfig element instead.

" - } - }, - "SSLSupportMethod": { - "base": null, - "refs": { - "ViewerCertificate$SSLSupportMethod": "

If you specify a value for ViewerCertificate$ACMCertificateArn or for ViewerCertificate$IAMCertificateId, you must also specify how you want CloudFront to serve HTTPS requests: using a method that works for all clients or one that works for most clients:

  • vip: CloudFront uses dedicated IP addresses for your content and can respond to HTTPS requests from any viewer. However, you will incur additional monthly charges.

  • sni-only: CloudFront can respond to HTTPS requests from viewers that support Server Name Indication (SNI). All modern browsers support SNI, but some browsers still in use don't support SNI. If some of your users' browsers don't support SNI, we recommend that you do one of the following:

    • Use the vip option (dedicated IP addresses) instead of sni-only.

    • Use the CloudFront SSL/TLS certificate instead of a custom certificate. This requires that you use the CloudFront domain name of your distribution in the URLs for your objects, for example, https://d111111abcdef8.cloudfront.net/logo.png.

    • If you can control which browser your users use, upgrade the browser to one that supports SNI.

    • Use HTTP instead of HTTPS.

Don't specify a value for SSLSupportMethod if you specified <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>.

For more information, see Using Alternate Domain Names and HTTPS in the Amazon CloudFront Developer Guide.

" - } - }, - "Signer": { - "base": "

A complex type that lists the AWS accounts that were included in the TrustedSigners complex type, as well as their active CloudFront key pair IDs, if any.

", - "refs": { - "SignerList$member": null - } - }, - "SignerList": { - "base": null, - "refs": { - "ActiveTrustedSigners$Items": "

A complex type that contains one Signer complex type for each trusted signer that is specified in the TrustedSigners complex type.

For more information, see ActiveTrustedSigners.

" - } - }, - "SslProtocol": { - "base": null, - "refs": { - "SslProtocolsList$member": null - } - }, - "SslProtocolsList": { - "base": null, - "refs": { - "OriginSslProtocols$Items": "

A list that contains allowed SSL/TLS protocols for this distribution.

" - } - }, - "StreamingDistribution": { - "base": "

A streaming distribution.

", - "refs": { - "CreateStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

", - "CreateStreamingDistributionWithTagsResult$StreamingDistribution": "

The streaming distribution's information.

", - "GetStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

", - "UpdateStreamingDistributionResult$StreamingDistribution": "

The streaming distribution's information.

" - } - }, - "StreamingDistributionAlreadyExists": { - "base": null, - "refs": { - } - }, - "StreamingDistributionConfig": { - "base": "

The RTMP distribution's configuration information.

", - "refs": { - "CreateStreamingDistributionRequest$StreamingDistributionConfig": "

The streaming distribution's configuration information.

", - "GetStreamingDistributionConfigResult$StreamingDistributionConfig": "

The streaming distribution's configuration information.

", - "StreamingDistribution$StreamingDistributionConfig": "

The current configuration information for the RTMP distribution.

", - "StreamingDistributionConfigWithTags$StreamingDistributionConfig": "

A streaming distribution Configuration.

", - "UpdateStreamingDistributionRequest$StreamingDistributionConfig": "

The streaming distribution's configuration information.

" - } - }, - "StreamingDistributionConfigWithTags": { - "base": "

A streaming distribution Configuration and a list of tags to be associated with the streaming distribution.

", - "refs": { - "CreateStreamingDistributionWithTagsRequest$StreamingDistributionConfigWithTags": "

The streaming distribution's configuration information.

" - } - }, - "StreamingDistributionList": { - "base": "

A streaming distribution list.

", - "refs": { - "ListStreamingDistributionsResult$StreamingDistributionList": "

The StreamingDistributionList type.

" - } - }, - "StreamingDistributionNotDisabled": { - "base": null, - "refs": { - } - }, - "StreamingDistributionSummary": { - "base": "

A summary of the information for an Amazon CloudFront streaming distribution.

", - "refs": { - "StreamingDistributionSummaryList$member": null - } - }, - "StreamingDistributionSummaryList": { - "base": null, - "refs": { - "StreamingDistributionList$Items": "

A complex type that contains one StreamingDistributionSummary element for each distribution that was created by the current AWS account.

" - } - }, - "StreamingLoggingConfig": { - "base": "

A complex type that controls whether access logs are written for this streaming distribution.

", - "refs": { - "StreamingDistributionConfig$Logging": "

A complex type that controls whether access logs are written for the streaming distribution.

" - } - }, - "Tag": { - "base": "

A complex type that contains Tag key and Tag value.

", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": "

A string that contains Tag key.

The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

", - "refs": { - "Tag$Key": "

A string that contains Tag key.

The string length should be between 1 and 128 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

", - "TagKeyList$member": null - } - }, - "TagKeyList": { - "base": null, - "refs": { - "TagKeys$Items": "

A complex type that contains Tag key elements.

" - } - }, - "TagKeys": { - "base": "

A complex type that contains zero or more Tag elements.

", - "refs": { - "UntagResourceRequest$TagKeys": "

A complex type that contains zero or more Tag key elements.

" - } - }, - "TagList": { - "base": null, - "refs": { - "Tags$Items": "

A complex type that contains Tag elements.

" - } - }, - "TagResourceRequest": { - "base": "

The request to add tags to a CloudFront resource.

", - "refs": { - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

A string that contains an optional Tag value.

The string length should be between 0 and 256 characters. Valid characters include a-z, A-Z, 0-9, space, and the special characters _ - . : / = + @.

" - } - }, - "Tags": { - "base": "

A complex type that contains zero or more Tag elements.

", - "refs": { - "DistributionConfigWithTags$Tags": "

A complex type that contains zero or more Tag elements.

", - "ListTagsForResourceResult$Tags": "

A complex type that contains zero or more Tag elements.

", - "StreamingDistributionConfigWithTags$Tags": "

A complex type that contains zero or more Tag elements.

", - "TagResourceRequest$Tags": "

A complex type that contains zero or more Tag elements.

" - } - }, - "TooManyCacheBehaviors": { - "base": "

You cannot create more cache behaviors for the distribution.

", - "refs": { - } - }, - "TooManyCertificates": { - "base": "

You cannot create anymore custom SSL/TLS certificates.

", - "refs": { - } - }, - "TooManyCloudFrontOriginAccessIdentities": { - "base": "

Processing your request would cause you to exceed the maximum number of origin access identities allowed.

", - "refs": { - } - }, - "TooManyCookieNamesInWhiteList": { - "base": "

Your request contains more cookie names in the whitelist than are allowed per cache behavior.

", - "refs": { - } - }, - "TooManyDistributionCNAMEs": { - "base": "

Your request contains more CNAMEs than are allowed per distribution.

", - "refs": { - } - }, - "TooManyDistributions": { - "base": "

Processing your request would cause you to exceed the maximum number of distributions allowed.

", - "refs": { - } - }, - "TooManyDistributionsWithLambdaAssociations": { - "base": "

Processing your request would cause the maximum number of distributions with Lambda function associations per owner to be exceeded.

", - "refs": { - } - }, - "TooManyHeadersInForwardedValues": { - "base": null, - "refs": { - } - }, - "TooManyInvalidationsInProgress": { - "base": "

You have exceeded the maximum number of allowable InProgress invalidation batch requests, or invalidation objects.

", - "refs": { - } - }, - "TooManyLambdaFunctionAssociations": { - "base": "

Your request contains more Lambda function associations than are allowed per distribution.

", - "refs": { - } - }, - "TooManyOriginCustomHeaders": { - "base": null, - "refs": { - } - }, - "TooManyOrigins": { - "base": "

You cannot create more origins for the distribution.

", - "refs": { - } - }, - "TooManyQueryStringParameters": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributionCNAMEs": { - "base": null, - "refs": { - } - }, - "TooManyStreamingDistributions": { - "base": "

Processing your request would cause you to exceed the maximum number of streaming distributions allowed.

", - "refs": { - } - }, - "TooManyTrustedSigners": { - "base": "

Your request contains more trusted signers than are allowed per distribution.

", - "refs": { - } - }, - "TrustedSignerDoesNotExist": { - "base": "

One or more of your trusted signers don't exist.

", - "refs": { - } - }, - "TrustedSigners": { - "base": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

For more information about updating the distribution configuration, see DistributionConfig .

", - "refs": { - "CacheBehavior$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

", - "DefaultCacheBehavior$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content.

If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items. For more information, see Serving Private Content through CloudFront in the Amazon Amazon CloudFront Developer Guide.

If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items.

To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

", - "StreamingDistributionConfig$TrustedSigners": "

A complex type that specifies any AWS accounts that you want to permit to create signed URLs for private content. If you want the distribution to use signed URLs, include this element; if you want the distribution to use public URLs, remove this element. For more information, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "StreamingDistributionSummary$TrustedSigners": "

A complex type that specifies the AWS accounts, if any, that you want to allow to create signed URLs for private content. If you want to require signed URLs in requests for objects in the target origin that match the PathPattern for this cache behavior, specify true for Enabled, and specify the applicable values for Quantity and Items.If you don't want to require signed URLs in requests for objects that match PathPattern, specify false for Enabled and 0 for Quantity. Omit Items. To add, change, or remove one or more trusted signers, change Enabled to true (if it's currently false), change Quantity as applicable, and specify all of the trusted signers that you want to include in the updated distribution.

" - } - }, - "UntagResourceRequest": { - "base": "

The request to remove tags from a CloudFront resource.

", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityRequest": { - "base": "

The request to update an origin access identity.

", - "refs": { - } - }, - "UpdateCloudFrontOriginAccessIdentityResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "UpdateDistributionRequest": { - "base": "

The request to update a distribution.

", - "refs": { - } - }, - "UpdateDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "UpdateStreamingDistributionRequest": { - "base": "

The request to update a streaming distribution.

", - "refs": { - } - }, - "UpdateStreamingDistributionResult": { - "base": "

The returned result of the corresponding request.

", - "refs": { - } - }, - "ViewerCertificate": { - "base": "

A complex type that specifies the following:

  • Whether you want viewers to use HTTP or HTTPS to request your objects.

  • If you want viewers to use HTTPS, whether you're using an alternate domain name such as example.com or the CloudFront domain name for your distribution, such as d111111abcdef8.cloudfront.net.

  • If you're using an alternate domain name, whether AWS Certificate Manager (ACM) provided the certificate, or you purchased a certificate from a third-party certificate authority and imported it into ACM or uploaded it to the IAM certificate store.

You must specify only one of the following values:

Don't specify false for CloudFrontDefaultCertificate.

If you want viewers to use HTTP instead of HTTPS to request your objects: Specify the following value:

<CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>

In addition, specify allow-all for ViewerProtocolPolicy for all of your cache behaviors.

If you want viewers to use HTTPS to request your objects: Choose the type of certificate that you want to use based on whether you're using an alternate domain name for your objects or the CloudFront domain name:

  • If you're using an alternate domain name, such as example.com: Specify one of the following values, depending on whether ACM provided your certificate or you purchased your certificate from third-party certificate authority:

    • <ACMCertificateArn>ARN for ACM SSL/TLS certificate<ACMCertificateArn> where ARN for ACM SSL/TLS certificate is the ARN for the ACM SSL/TLS certificate that you want to use for this distribution.

    • <IAMCertificateId>IAM certificate ID<IAMCertificateId> where IAM certificate ID is the ID that IAM returned when you added the certificate to the IAM certificate store.

    If you specify ACMCertificateArn or IAMCertificateId, you must also specify a value for SSLSupportMethod.

    If you choose to use an ACM certificate or a certificate in the IAM certificate store, we recommend that you use only an alternate domain name in your object URLs (https://example.com/logo.jpg). If you use the domain name that is associated with your CloudFront distribution (such as https://d111111abcdef8.cloudfront.net/logo.jpg) and the viewer supports SNI, then CloudFront behaves normally. However, if the browser does not support SNI, the user's experience depends on the value that you choose for SSLSupportMethod:

    • vip: The viewer displays a warning because there is a mismatch between the CloudFront domain name and the domain name in your SSL/TLS certificate.

    • sni-only: CloudFront drops the connection with the browser without returning the object.

  • If you're using the CloudFront domain name for your distribution, such as d111111abcdef8.cloudfront.net : Specify the following value:

    <CloudFrontDefaultCertificate>true<CloudFrontDefaultCertificate>

If you want viewers to use HTTPS, you must also specify one of the following values in your cache behaviors:

  • <ViewerProtocolPolicy>https-only<ViewerProtocolPolicy>

  • <ViewerProtocolPolicy>redirect-to-https<ViewerProtocolPolicy>

You can also optionally require that CloudFront use HTTPS to communicate with your origin by specifying one of the following values for the applicable origins:

  • <OriginProtocolPolicy>https-only<OriginProtocolPolicy>

  • <OriginProtocolPolicy>match-viewer<OriginProtocolPolicy>

For more information, see Using Alternate Domain Names and HTTPS in the Amazon CloudFront Developer Guide.

", - "refs": { - "DistributionConfig$ViewerCertificate": null, - "DistributionSummary$ViewerCertificate": null - } - }, - "ViewerProtocolPolicy": { - "base": null, - "refs": { - "CacheBehavior$ViewerProtocolPolicy": "

The protocol that viewers can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. You can specify the following options:

  • allow-all: Viewers can use HTTP or HTTPS.

  • redirect-to-https: If a viewer submits an HTTP request, CloudFront returns an HTTP status code of 301 (Moved Permanently) to the viewer along with the HTTPS URL. The viewer then resubmits the request using the new URL.

  • https-only: If a viewer sends an HTTP request, CloudFront returns an HTTP status code of 403 (Forbidden).

For more information about requiring the HTTPS protocol, see Using an HTTPS Connection to Access Your Objects in the Amazon CloudFront Developer Guide.

The only way to guarantee that viewers retrieve an object that was fetched from the origin using HTTPS is never to use any other protocol to fetch the object. If you have recently changed from HTTP to HTTPS, we recommend that you clear your objects' cache because cached objects are protocol agnostic. That means that an edge location will return an object from the cache regardless of whether the current request protocol matches the protocol used previously. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$ViewerProtocolPolicy": "

The protocol that viewers can use to access the files in the origin specified by TargetOriginId when a request matches the path pattern in PathPattern. You can specify the following options:

  • allow-all: Viewers can use HTTP or HTTPS.

  • redirect-to-https: If a viewer submits an HTTP request, CloudFront returns an HTTP status code of 301 (Moved Permanently) to the viewer along with the HTTPS URL. The viewer then resubmits the request using the new URL.

  • https-only: If a viewer sends an HTTP request, CloudFront returns an HTTP status code of 403 (Forbidden).

For more information about requiring the HTTPS protocol, see Using an HTTPS Connection to Access Your Objects in the Amazon CloudFront Developer Guide.

The only way to guarantee that viewers retrieve an object that was fetched from the origin using HTTPS is never to use any other protocol to fetch the object. If you have recently changed from HTTP to HTTPS, we recommend that you clear your objects' cache because cached objects are protocol agnostic. That means that an edge location will return an object from the cache regardless of whether the current request protocol matches the protocol used previously. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

" - } - }, - "boolean": { - "base": null, - "refs": { - "ActiveTrustedSigners$Enabled": "

Enabled is true if any of the AWS accounts listed in the TrustedSigners complex type for this RTMP distribution have active CloudFront key pairs. If not, Enabled is false.

For more information, see ActiveTrustedSigners.

", - "CacheBehavior$SmoothStreaming": "

Indicates whether you want to distribute media files in the Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false. If you specify true for SmoothStreaming, you can still distribute other content using this cache behavior if the content matches the value of PathPattern.

", - "CacheBehavior$Compress": "

Whether you want CloudFront to automatically compress certain files for this cache behavior. If so, specify true; if not, specify false. For more information, see Serving Compressed Files in the Amazon CloudFront Developer Guide.

", - "CloudFrontOriginAccessIdentityList$IsTruncated": "

A flag that indicates whether more origin access identities remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more items in the list.

", - "DefaultCacheBehavior$SmoothStreaming": "

Indicates whether you want to distribute media files in the Microsoft Smooth Streaming format using the origin that is associated with this cache behavior. If so, specify true; if not, specify false. If you specify true for SmoothStreaming, you can still distribute other content using this cache behavior if the content matches the value of PathPattern.

", - "DefaultCacheBehavior$Compress": "

Whether you want CloudFront to automatically compress certain files for this cache behavior. If so, specify true; if not, specify false. For more information, see Serving Compressed Files in the Amazon CloudFront Developer Guide.

", - "DistributionConfig$Enabled": "

From this field, you can enable or disable the selected distribution.

If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.

", - "DistributionConfig$IsIPV6Enabled": "

If you want CloudFront to respond to IPv6 DNS requests with an IPv6 address for your distribution, specify true. If you specify false, CloudFront responds to IPv6 DNS requests with the DNS response code NOERROR and with no IP addresses. This allows viewers to submit a second request, for an IPv4 address for your distribution.

In general, you should enable IPv6 if you have users on IPv6 networks who want to access your content. However, if you're using signed URLs or signed cookies to restrict access to your content, and if you're using a custom policy that includes the IpAddress parameter to restrict the IP addresses that can access your content, don't enable IPv6. If you want to restrict access to some content by IP address and not restrict access to other content (or restrict access but not by IP address), you can create two distributions. For more information, see Creating a Signed URL Using a Custom Policy in the Amazon CloudFront Developer Guide.

If you're using an Amazon Route 53 alias resource record set to route traffic to your CloudFront distribution, you need to create a second alias resource record set when both of the following are true:

  • You enable IPv6 for the distribution

  • You're using alternate domain names in the URLs for your objects

For more information, see Routing Traffic to an Amazon CloudFront Web Distribution by Using Your Domain Name in the Amazon Route 53 Developer Guide.

If you created a CNAME resource record set, either with Amazon Route 53 or with another DNS service, you don't need to make any changes. A CNAME record will route traffic to your distribution regardless of the IP address format of the viewer request.

", - "DistributionList$IsTruncated": "

A flag that indicates whether more distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.

", - "DistributionSummary$Enabled": "

Whether the distribution is enabled to accept user requests for content.

", - "DistributionSummary$IsIPV6Enabled": "

Whether CloudFront responds to IPv6 DNS requests with an IPv6 address for your distribution.

", - "ForwardedValues$QueryString": "

Indicates whether you want CloudFront to forward query strings to the origin that is associated with this cache behavior and cache based on the query string parameters. CloudFront behavior depends on the value of QueryString and on the values that you specify for QueryStringCacheKeys, if any:

If you specify true for QueryString and you don't specify any values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin and caches based on all query string parameters. Depending on how many query string parameters and values you have, this can adversely affect performance because CloudFront must forward more requests to the origin.

If you specify true for QueryString and you specify one or more values for QueryStringCacheKeys, CloudFront forwards all query string parameters to the origin, but it only caches based on the query string parameters that you specify.

If you specify false for QueryString, CloudFront doesn't forward any query string parameters to the origin, and doesn't cache based on query string parameters.

For more information, see Configuring CloudFront to Cache Based on Query String Parameters in the Amazon CloudFront Developer Guide.

", - "InvalidationList$IsTruncated": "

A flag that indicates whether more invalidation batch requests remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more invalidation batches in the list.

", - "LoggingConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you don't want to enable logging when you create a distribution or if you want to disable logging for an existing distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket, prefix, and IncludeCookies, the values are automatically deleted.

", - "LoggingConfig$IncludeCookies": "

Specifies whether you want CloudFront to include cookies in access logs, specify true for IncludeCookies. If you choose to include cookies in logs, CloudFront logs all cookies regardless of how you configure the cache behaviors for this distribution. If you don't want to include cookies when you create a distribution or if you want to disable include cookies for an existing distribution, specify false for IncludeCookies.

", - "StreamingDistributionConfig$Enabled": "

Whether the streaming distribution is enabled to accept user requests for content.

", - "StreamingDistributionList$IsTruncated": "

A flag that indicates whether more streaming distributions remain to be listed. If your results were truncated, you can make a follow-up pagination request using the Marker request parameter to retrieve more distributions in the list.

", - "StreamingDistributionSummary$Enabled": "

Whether the distribution is enabled to accept end user requests for content.

", - "StreamingLoggingConfig$Enabled": "

Specifies whether you want CloudFront to save access logs to an Amazon S3 bucket. If you don't want to enable logging when you create a streaming distribution or if you want to disable logging for an existing streaming distribution, specify false for Enabled, and specify empty Bucket and Prefix elements. If you specify false for Enabled but you specify values for Bucket and Prefix, the values are automatically deleted.

", - "TrustedSigners$Enabled": "

Specifies whether you want to require viewers to use signed URLs to access the files specified by PathPattern and TargetOriginId.

", - "ViewerCertificate$CloudFrontDefaultCertificate": "

For information about how and when to use CloudFrontDefaultCertificate, see ViewerCertificate.

" - } - }, - "integer": { - "base": null, - "refs": { - "ActiveTrustedSigners$Quantity": "

A complex type that contains one Signer complex type for each trusted signer specified in the TrustedSigners complex type.

For more information, see ActiveTrustedSigners.

", - "Aliases$Quantity": "

The number of CNAME aliases, if any, that you want to associate with this distribution.

", - "AllowedMethods$Quantity": "

The number of HTTP methods that you want CloudFront to forward to your origin. Valid values are 2 (for GET and HEAD requests), 3 (for GET, HEAD, and OPTIONS requests) and 7 (for GET, HEAD, OPTIONS, PUT, PATCH, POST, and DELETE requests).

", - "CacheBehaviors$Quantity": "

The number of cache behaviors for this distribution.

", - "CachedMethods$Quantity": "

The number of HTTP methods for which you want CloudFront to cache responses. Valid values are 2 (for caching responses to GET and HEAD requests) and 3 (for caching responses to GET, HEAD, and OPTIONS requests).

", - "CloudFrontOriginAccessIdentityList$MaxItems": "

The maximum number of origin access identities you want in the response body.

", - "CloudFrontOriginAccessIdentityList$Quantity": "

The number of CloudFront origin access identities that were created by the current AWS account.

", - "CookieNames$Quantity": "

The number of different cookies that you want CloudFront to forward to the origin for this cache behavior.

", - "CustomErrorResponse$ErrorCode": "

The HTTP status code for which you want to specify a custom error page and/or a caching duration.

", - "CustomErrorResponses$Quantity": "

The number of HTTP status codes for which you want to specify a custom error page and/or a caching duration. If Quantity is 0, you can omit Items.

", - "CustomHeaders$Quantity": "

The number of custom headers, if any, for this distribution.

", - "CustomOriginConfig$HTTPPort": "

The HTTP port the custom origin listens on.

", - "CustomOriginConfig$HTTPSPort": "

The HTTPS port the custom origin listens on.

", - "CustomOriginConfig$OriginReadTimeout": "

You can create a custom origin read timeout. All timeout units are in seconds. The default origin read timeout is 30 seconds, but you can configure custom timeout lengths using the CloudFront API. The minimum timeout length is 4 seconds; the maximum is 60 seconds.

If you need to increase the maximum time limit, contact the AWS Support Center.

", - "CustomOriginConfig$OriginKeepaliveTimeout": "

You can create a custom keep-alive timeout. All timeout units are in seconds. The default keep-alive timeout is 5 seconds, but you can configure custom timeout lengths using the CloudFront API. The minimum timeout length is 1 second; the maximum is 60 seconds.

If you need to increase the maximum time limit, contact the AWS Support Center.

", - "Distribution$InProgressInvalidationBatches": "

The number of invalidation batches currently in progress.

", - "DistributionList$MaxItems": "

The value you provided for the MaxItems request parameter.

", - "DistributionList$Quantity": "

The number of distributions that were created by the current AWS account.

", - "GeoRestriction$Quantity": "

When geo restriction is enabled, this is the number of countries in your whitelist or blacklist. Otherwise, when it is not enabled, Quantity is 0, and you can omit Items.

", - "Headers$Quantity": "

The number of different headers that you want CloudFront to base caching on for this cache behavior. You can configure each cache behavior in a web distribution to do one of the following:

  • Forward all headers to your origin: Specify 1 for Quantity and * for Name.

    CloudFront doesn't cache the objects that are associated with this cache behavior. Instead, CloudFront sends every request to the origin.

  • Forward a whitelist of headers you specify: Specify the number of headers that you want CloudFront to base caching on. Then specify the header names in Name elements. CloudFront caches your objects based on the values in the specified headers.

  • Forward only the default headers: Specify 0 for Quantity and omit Items. In this configuration, CloudFront doesn't cache based on the values in the request headers.

Regardless of which option you choose, CloudFront forwards headers to your origin based on whether the origin is an S3 bucket or a custom origin. See the following documentation:

", - "InvalidationList$MaxItems": "

The value that you provided for the MaxItems request parameter.

", - "InvalidationList$Quantity": "

The number of invalidation batches that were created by the current AWS account.

", - "KeyPairIds$Quantity": "

The number of active CloudFront key pairs for AwsAccountNumber.

For more information, see ActiveTrustedSigners.

", - "LambdaFunctionAssociations$Quantity": "

The number of Lambda function associations for this cache behavior.

", - "OriginSslProtocols$Quantity": "

The number of SSL/TLS protocols that you want to allow CloudFront to use when establishing an HTTPS connection with this origin.

", - "Origins$Quantity": "

The number of origins for this distribution.

", - "Paths$Quantity": "

The number of objects that you want to invalidate.

", - "QueryStringCacheKeys$Quantity": "

The number of whitelisted query string parameters for this cache behavior.

", - "StreamingDistributionList$MaxItems": "

The value you provided for the MaxItems request parameter.

", - "StreamingDistributionList$Quantity": "

The number of streaming distributions that were created by the current AWS account.

", - "TrustedSigners$Quantity": "

The number of trusted signers for this cache behavior.

" - } - }, - "long": { - "base": null, - "refs": { - "CacheBehavior$MinTTL": "

The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon Amazon CloudFront Developer Guide.

You must specify 0 for MinTTL if you configure CloudFront to forward all headers to your origin (under Headers, if you specify 1 for Quantity and * for Name).

", - "CacheBehavior$DefaultTTL": "

The default amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "CacheBehavior$MaxTTL": "

The maximum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin adds HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "CustomErrorResponse$ErrorCachingMinTTL": "

The minimum amount of time, in seconds, that you want CloudFront to cache the HTTP status code specified in ErrorCode. When this time period has elapsed, CloudFront queries your origin to see whether the problem that caused the error has been resolved and the requested object is now available.

If you don't want to specify a value, include an empty element, <ErrorCachingMinTTL>, in the XML document.

For more information, see Customizing Error Responses in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$MinTTL": "

The minimum amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon Amazon CloudFront Developer Guide.

You must specify 0 for MinTTL if you configure CloudFront to forward all headers to your origin (under Headers, if you specify 1 for Quantity and * for Name).

", - "DefaultCacheBehavior$DefaultTTL": "

The default amount of time that you want objects to stay in CloudFront caches before CloudFront forwards another request to your origin to determine whether the object has been updated. The value that you specify applies only when your origin does not add HTTP headers such as Cache-Control max-age, Cache-Control s-maxage, and Expires to objects. For more information, see Specifying How Long Objects and Errors Stay in a CloudFront Edge Cache (Expiration) in the Amazon CloudFront Developer Guide.

", - "DefaultCacheBehavior$MaxTTL": null - } - }, - "string": { - "base": null, - "refs": { - "AccessDenied$Message": null, - "AliasList$member": null, - "AwsAccountNumberList$member": null, - "BatchTooLarge$Message": null, - "CNAMEAlreadyExists$Message": null, - "CacheBehavior$PathPattern": "

The pattern (for example, images/*.jpg) that specifies which requests to apply the behavior to. When CloudFront receives a viewer request, the requested path is compared with path patterns in the order in which cache behaviors are listed in the distribution.

You can optionally include a slash (/) at the beginning of the path pattern. For example, /images/*.jpg. CloudFront behavior is the same with or without the leading /.

The path pattern for the default cache behavior is * and cannot be changed. If the request for an object does not match the path pattern for any cache behaviors, CloudFront applies the behavior in the default cache behavior.

For more information, see Path Pattern in the Amazon CloudFront Developer Guide.

", - "CacheBehavior$TargetOriginId": "

The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.

", - "CloudFrontOriginAccessIdentity$Id": "

The ID for the origin access identity, for example, E74FTE3AJFJ256A.

", - "CloudFrontOriginAccessIdentity$S3CanonicalUserId": "

The Amazon S3 canonical user ID for the origin access identity, used when giving the origin access identity read permission to an object in Amazon S3.

", - "CloudFrontOriginAccessIdentityAlreadyExists$Message": null, - "CloudFrontOriginAccessIdentityConfig$CallerReference": "

A unique number that ensures the request can't be replayed.

If the CallerReference is new (no matter the content of the CloudFrontOriginAccessIdentityConfig object), a new origin access identity is created.

If the CallerReference is a value already sent in a previous identity request, and the content of the CloudFrontOriginAccessIdentityConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request.

If the CallerReference is a value you already sent in a previous request to create an identity, but the content of the CloudFrontOriginAccessIdentityConfig is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists error.

", - "CloudFrontOriginAccessIdentityConfig$Comment": "

Any comments you want to include about the origin access identity.

", - "CloudFrontOriginAccessIdentityInUse$Message": null, - "CloudFrontOriginAccessIdentityList$Marker": "

Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).

", - "CloudFrontOriginAccessIdentityList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your origin access identities where they left off.

", - "CloudFrontOriginAccessIdentitySummary$Id": "

The ID for the origin access identity. For example: E74FTE3AJFJ256A.

", - "CloudFrontOriginAccessIdentitySummary$S3CanonicalUserId": "

The Amazon S3 canonical user ID for the origin access identity, which you use when giving the origin access identity read permission to an object in Amazon S3.

", - "CloudFrontOriginAccessIdentitySummary$Comment": "

The comment for this origin access identity, as originally specified when created.

", - "CookieNameList$member": null, - "CreateCloudFrontOriginAccessIdentityResult$Location": "

The fully qualified URI of the new origin access identity just created. For example: https://cloudfront.amazonaws.com/2010-11-01/origin-access-identity/cloudfront/E74FTE3AJFJ256A.

", - "CreateCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the origin access identity created.

", - "CreateDistributionResult$Location": "

The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.

", - "CreateDistributionResult$ETag": "

The current version of the distribution created.

", - "CreateDistributionWithTagsResult$Location": "

The fully qualified URI of the new distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/distribution/EDFDVBD632BHDS5.

", - "CreateDistributionWithTagsResult$ETag": "

The current version of the distribution created.

", - "CreateInvalidationRequest$DistributionId": "

The distribution's id.

", - "CreateInvalidationResult$Location": "

The fully qualified URI of the distribution and invalidation batch request, including the Invalidation ID.

", - "CreateStreamingDistributionResult$Location": "

The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.

", - "CreateStreamingDistributionResult$ETag": "

The current version of the streaming distribution created.

", - "CreateStreamingDistributionWithTagsResult$Location": "

The fully qualified URI of the new streaming distribution resource just created. For example: https://cloudfront.amazonaws.com/2010-11-01/streaming-distribution/EGTXBD79H29TRA8.

", - "CreateStreamingDistributionWithTagsResult$ETag": null, - "CustomErrorResponse$ResponsePagePath": "

The path to the custom error page that you want CloudFront to return to a viewer when your origin returns the HTTP status code specified by ErrorCode, for example, /4xx-errors/403-forbidden.html. If you want to store your objects and your custom error pages in different locations, your distribution must include a cache behavior for which the following is true:

  • The value of PathPattern matches the path to your custom error messages. For example, suppose you saved custom error pages for 4xx errors in an Amazon S3 bucket in a directory named /4xx-errors. Your distribution must include a cache behavior for which the path pattern routes requests for your custom error pages to that location, for example, /4xx-errors/*.

  • The value of TargetOriginId specifies the value of the ID element for the origin that contains your custom error pages.

If you specify a value for ResponsePagePath, you must also specify a value for ResponseCode. If you don't want to specify a value, include an empty element, <ResponsePagePath>, in the XML document.

We recommend that you store custom error pages in an Amazon S3 bucket. If you store custom error pages on an HTTP server and the server starts to return 5xx errors, CloudFront can't get the files that you want to return to viewers because the origin server is unavailable.

", - "CustomErrorResponse$ResponseCode": "

The HTTP status code that you want CloudFront to return to the viewer along with the custom error page. There are a variety of reasons that you might want CloudFront to return a status code different from the status code that your origin returned to CloudFront, for example:

  • Some Internet devices (some firewalls and corporate proxies, for example) intercept HTTP 4xx and 5xx and prevent the response from being returned to the viewer. If you substitute 200, the response typically won't be intercepted.

  • If you don't care about distinguishing among different client errors or server errors, you can specify 400 or 500 as the ResponseCode for all 4xx or 5xx errors.

  • You might want to return a 200 status code (OK) and static website so your customers don't know that your website is down.

If you specify a value for ResponseCode, you must also specify a value for ResponsePagePath. If you don't want to specify a value, include an empty element, <ResponseCode>, in the XML document.

", - "DefaultCacheBehavior$TargetOriginId": "

The value of ID for the origin that you want CloudFront to route requests to when a request matches the path pattern either for a cache behavior or for the default cache behavior.

", - "DeleteCloudFrontOriginAccessIdentityRequest$Id": "

The origin access identity's ID.

", - "DeleteCloudFrontOriginAccessIdentityRequest$IfMatch": "

The value of the ETag header you received from a previous GET or PUT request. For example: E2QWRUHAPOMQZL.

", - "DeleteDistributionRequest$Id": "

The distribution ID.

", - "DeleteDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the distribution. For example: E2QWRUHAPOMQZL.

", - "DeleteServiceLinkedRoleRequest$RoleName": null, - "DeleteStreamingDistributionRequest$Id": "

The distribution ID.

", - "DeleteStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.

", - "Distribution$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "Distribution$ARN": "

The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "Distribution$Status": "

This response element indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated to all CloudFront edge locations.

", - "Distribution$DomainName": "

The domain name corresponding to the distribution, for example, d111111abcdef8.cloudfront.net.

", - "DistributionAlreadyExists$Message": null, - "DistributionConfig$CallerReference": "

A unique value (for example, a date-time stamp) that ensures that the request can't be replayed.

If the value of CallerReference is new (regardless of the content of the DistributionConfig object), CloudFront creates a new distribution.

If CallerReference is a value you already sent in a previous request to create a distribution, and if the content of the DistributionConfig is identical to the original request (ignoring white space), CloudFront returns the same the response that it returned to the original request.

If CallerReference is a value you already sent in a previous request to create a distribution but the content of the DistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.

", - "DistributionConfig$DefaultRootObject": "

The object that you want CloudFront to request from your origin (for example, index.html) when a viewer requests the root URL for your distribution (http://www.example.com) instead of an object in your distribution (http://www.example.com/product-description.html). Specifying a default root object avoids exposing the contents of your distribution.

Specify only the object name, for example, index.html. Don't add a / before the object name.

If you don't want to specify a default root object when you create a distribution, include an empty DefaultRootObject element.

To delete the default root object from an existing distribution, update the distribution configuration and include an empty DefaultRootObject element.

To replace the default root object, update the distribution configuration and specify the new object.

For more information about the default root object, see Creating a Default Root Object in the Amazon CloudFront Developer Guide.

", - "DistributionConfig$Comment": "

Any comments you want to include about the distribution.

If you don't want to specify a comment, include an empty Comment element.

To delete an existing comment, update the distribution configuration and include an empty Comment element.

To add or change a comment, update the distribution configuration and specify the new comment.

", - "DistributionConfig$WebACLId": "

A unique identifier that specifies the AWS WAF web ACL, if any, to associate with this distribution.

AWS WAF is a web application firewall that lets you monitor the HTTP and HTTPS requests that are forwarded to CloudFront, and lets you control access to your content. Based on conditions that you specify, such as the IP addresses that requests originate from or the values of query strings, CloudFront responds to requests either with the requested content or with an HTTP 403 status code (Forbidden). You can also configure CloudFront to return a custom error page when a request is blocked. For more information about AWS WAF, see the AWS WAF Developer Guide.

", - "DistributionList$Marker": "

The value you provided for the Marker request parameter.

", - "DistributionList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your distributions where they left off.

", - "DistributionNotDisabled$Message": null, - "DistributionSummary$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", - "DistributionSummary$ARN": "

The ARN (Amazon Resource Name) for the distribution. For example: arn:aws:cloudfront::123456789012:distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "DistributionSummary$Status": "

The current status of the distribution. When the status is Deployed, the distribution's information is propagated to all CloudFront edge locations.

", - "DistributionSummary$DomainName": "

The domain name that corresponds to the distribution, for example, d111111abcdef8.cloudfront.net.

", - "DistributionSummary$Comment": "

The comment originally specified when this distribution was created.

", - "DistributionSummary$WebACLId": "

The Web ACL Id (if any) associated with the distribution.

", - "GetCloudFrontOriginAccessIdentityConfigRequest$Id": "

The identity's ID.

", - "GetCloudFrontOriginAccessIdentityConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetCloudFrontOriginAccessIdentityRequest$Id": "

The identity's ID.

", - "GetCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the origin access identity's information. For example: E2QWRUHAPOMQZL.

", - "GetDistributionConfigRequest$Id": "

The distribution's ID.

", - "GetDistributionConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetDistributionRequest$Id": "

The distribution's ID.

", - "GetDistributionResult$ETag": "

The current version of the distribution's information. For example: E2QWRUHAPOMQZL.

", - "GetInvalidationRequest$DistributionId": "

The distribution's ID.

", - "GetInvalidationRequest$Id": "

The identifier for the invalidation request, for example, IDFDVBD632BHDS5.

", - "GetStreamingDistributionConfigRequest$Id": "

The streaming distribution's ID.

", - "GetStreamingDistributionConfigResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "GetStreamingDistributionRequest$Id": "

The streaming distribution's ID.

", - "GetStreamingDistributionResult$ETag": "

The current version of the streaming distribution's information. For example: E2QWRUHAPOMQZL.

", - "HeaderList$member": null, - "IllegalUpdate$Message": null, - "InconsistentQuantities$Message": null, - "InvalidArgument$Message": null, - "InvalidDefaultRootObject$Message": null, - "InvalidErrorCode$Message": null, - "InvalidForwardCookies$Message": null, - "InvalidGeoRestrictionParameter$Message": null, - "InvalidHeadersForS3Origin$Message": null, - "InvalidIfMatchVersion$Message": null, - "InvalidLambdaFunctionAssociation$Message": null, - "InvalidLocationCode$Message": null, - "InvalidMinimumProtocolVersion$Message": null, - "InvalidOrigin$Message": null, - "InvalidOriginAccessIdentity$Message": null, - "InvalidOriginKeepaliveTimeout$Message": null, - "InvalidOriginReadTimeout$Message": null, - "InvalidProtocolSettings$Message": null, - "InvalidQueryStringParameters$Message": null, - "InvalidRelativePath$Message": null, - "InvalidRequiredProtocol$Message": null, - "InvalidResponseCode$Message": null, - "InvalidTTLOrder$Message": null, - "InvalidTagging$Message": null, - "InvalidViewerCertificate$Message": null, - "InvalidWebACLId$Message": null, - "Invalidation$Id": "

The identifier for the invalidation request. For example: IDFDVBD632BHDS5.

", - "Invalidation$Status": "

The status of the invalidation request. When the invalidation batch is finished, the status is Completed.

", - "InvalidationBatch$CallerReference": "

A value that you specify to uniquely identify an invalidation request. CloudFront uses the value to prevent you from accidentally resubmitting an identical request. Whenever you create a new invalidation request, you must specify a new value for CallerReference and change other values in the request as applicable. One way to ensure that the value of CallerReference is unique is to use a timestamp, for example, 20120301090000.

If you make a second invalidation request with the same value for CallerReference, and if the rest of the request is the same, CloudFront doesn't create a new invalidation request. Instead, CloudFront returns information about the invalidation request that you previously created with the same CallerReference.

If CallerReference is a value you already sent in a previous invalidation batch request but the content of any Path is different from the original request, CloudFront returns an InvalidationBatchAlreadyExists error.

", - "InvalidationList$Marker": "

The value that you provided for the Marker request parameter.

", - "InvalidationList$NextMarker": "

If IsTruncated is true, this element is present and contains the value that you can use for the Marker request parameter to continue listing your invalidation batches where they left off.

", - "InvalidationSummary$Id": "

The unique ID for an invalidation request.

", - "InvalidationSummary$Status": "

The status of an invalidation request.

", - "KeyPairIdList$member": null, - "LambdaFunctionAssociation$LambdaFunctionARN": "

The ARN of the Lambda function. You must specify the ARN of a function version; you can't specify a Lambda alias or $LATEST.

", - "ListCloudFrontOriginAccessIdentitiesRequest$Marker": "

Use this when paginating results to indicate where to begin in your list of origin access identities. The results include identities in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last identity on that page).

", - "ListCloudFrontOriginAccessIdentitiesRequest$MaxItems": "

The maximum number of origin access identities you want in the response body.

", - "ListDistributionsByWebACLIdRequest$Marker": "

Use Marker and MaxItems to control pagination of results. If you have more than MaxItems distributions that satisfy the request, the response includes a NextMarker element. To get the next page of results, submit another request. For the value of Marker, specify the value of NextMarker from the last response. (For the first request, omit Marker.)

", - "ListDistributionsByWebACLIdRequest$MaxItems": "

The maximum number of distributions that you want CloudFront to return in the response body. The maximum and default values are both 100.

", - "ListDistributionsByWebACLIdRequest$WebACLId": "

The ID of the AWS WAF web ACL that you want to list the associated distributions. If you specify \"null\" for the ID, the request returns a list of the distributions that aren't associated with a web ACL.

", - "ListDistributionsRequest$Marker": "

Use this when paginating results to indicate where to begin in your list of distributions. The results include distributions in the list that occur after the marker. To get the next page of results, set the Marker to the value of the NextMarker from the current page's response (which is also the ID of the last distribution on that page).

", - "ListDistributionsRequest$MaxItems": "

The maximum number of distributions you want in the response body.

", - "ListInvalidationsRequest$DistributionId": "

The distribution's ID.

", - "ListInvalidationsRequest$Marker": "

Use this parameter when paginating results to indicate where to begin in your list of invalidation batches. Because the results are returned in decreasing order from most recent to oldest, the most recent results are on the first page, the second page will contain earlier results, and so on. To get the next page of results, set Marker to the value of the NextMarker from the current page's response. This value is the same as the ID of the last invalidation batch on that page.

", - "ListInvalidationsRequest$MaxItems": "

The maximum number of invalidation batches that you want in the response body.

", - "ListStreamingDistributionsRequest$Marker": "

The value that you provided for the Marker request parameter.

", - "ListStreamingDistributionsRequest$MaxItems": "

The value that you provided for the MaxItems request parameter.

", - "LocationList$member": null, - "LoggingConfig$Bucket": "

The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.

", - "LoggingConfig$Prefix": "

An optional string that you want CloudFront to prefix to the access log filenames for this distribution, for example, myprefix/. If you want to enable logging, but you don't want to specify a prefix, you still must include an empty Prefix element in the Logging element.

", - "MissingBody$Message": null, - "NoSuchCloudFrontOriginAccessIdentity$Message": null, - "NoSuchDistribution$Message": null, - "NoSuchInvalidation$Message": null, - "NoSuchOrigin$Message": null, - "NoSuchResource$Message": null, - "NoSuchStreamingDistribution$Message": null, - "Origin$Id": "

A unique identifier for the origin. The value of Id must be unique within the distribution.

When you specify the value of TargetOriginId for the default cache behavior or for another cache behavior, you indicate the origin to which you want the cache behavior to route requests by specifying the value of the Id element for that origin. When a request matches the path pattern for that cache behavior, CloudFront routes the request to the specified origin. For more information, see Cache Behavior Settings in the Amazon CloudFront Developer Guide.

", - "Origin$DomainName": "

Amazon S3 origins: The DNS name of the Amazon S3 bucket from which you want CloudFront to get objects for this origin, for example, myawsbucket.s3.amazonaws.com.

Constraints for Amazon S3 origins:

  • If you configured Amazon S3 Transfer Acceleration for your bucket, don't specify the s3-accelerate endpoint for DomainName.

  • The bucket name must be between 3 and 63 characters long (inclusive).

  • The bucket name must contain only lowercase characters, numbers, periods, underscores, and dashes.

  • The bucket name must not contain adjacent periods.

Custom Origins: The DNS domain name for the HTTP server from which you want CloudFront to get objects for this origin, for example, www.example.com.

Constraints for custom origins:

  • DomainName must be a valid DNS name that contains only a-z, A-Z, 0-9, dot (.), hyphen (-), or underscore (_) characters.

  • The name cannot exceed 128 characters.

", - "Origin$OriginPath": "

An optional element that causes CloudFront to request your content from a directory in your Amazon S3 bucket or your custom origin. When you include the OriginPath element, specify the directory name, beginning with a /. CloudFront appends the directory name to the value of DomainName, for example, example.com/production. Do not include a / at the end of the directory name.

For example, suppose you've specified the following values for your distribution:

  • DomainName: An Amazon S3 bucket named myawsbucket.

  • OriginPath: /production

  • CNAME: example.com

When a user enters example.com/index.html in a browser, CloudFront sends a request to Amazon S3 for myawsbucket/production/index.html.

When a user enters example.com/acme/index.html in a browser, CloudFront sends a request to Amazon S3 for myawsbucket/production/acme/index.html.

", - "OriginCustomHeader$HeaderName": "

The name of a header that you want CloudFront to forward to your origin. For more information, see Forwarding Custom Headers to Your Origin (Web Distributions Only) in the Amazon Amazon CloudFront Developer Guide.

", - "OriginCustomHeader$HeaderValue": "

The value for the header that you specified in the HeaderName field.

", - "PathList$member": null, - "PreconditionFailed$Message": null, - "QueryStringCacheKeysList$member": null, - "ResourceInUse$Message": null, - "S3Origin$DomainName": "

The DNS name of the Amazon S3 origin.

", - "S3Origin$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the RTMP distribution. Use an origin access identity to configure the distribution so that end users can only access objects in an Amazon S3 bucket through CloudFront.

If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information, see Using an Origin Access Identity to Restrict Access to Your Amazon S3 Content in the Amazon Amazon CloudFront Developer Guide.

", - "S3OriginConfig$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that viewers can only access objects in an Amazon S3 bucket through CloudFront. The format of the value is:

origin-access-identity/cloudfront/ID-of-origin-access-identity

where ID-of-origin-access-identity is the value that CloudFront returned in the ID element when you created the origin access identity.

If you want viewers to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information about the origin access identity, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", - "Signer$AwsAccountNumber": "

An AWS account that is included in the TrustedSigners complex type for this RTMP distribution. Valid values include:

  • self, which is the AWS account used to create the distribution.

  • An AWS account number.

", - "StreamingDistribution$Id": "

The identifier for the RTMP distribution. For example: EGTXBD79EXAMPLE.

", - "StreamingDistribution$ARN": null, - "StreamingDistribution$Status": "

The current status of the RTMP distribution. When the status is Deployed, the distribution's information is propagated to all CloudFront edge locations.

", - "StreamingDistribution$DomainName": "

The domain name that corresponds to the streaming distribution, for example, s5c39gqb8ow64r.cloudfront.net.

", - "StreamingDistributionAlreadyExists$Message": null, - "StreamingDistributionConfig$CallerReference": "

A unique number that ensures that the request can't be replayed. If the CallerReference is new (no matter the content of the StreamingDistributionConfig object), a new streaming distribution is created. If the CallerReference is a value that you already sent in a previous request to create a streaming distribution, and the content of the StreamingDistributionConfig is identical to the original request (ignoring white space), the response includes the same information returned to the original request. If the CallerReference is a value that you already sent in a previous request to create a streaming distribution but the content of the StreamingDistributionConfig is different from the original request, CloudFront returns a DistributionAlreadyExists error.

", - "StreamingDistributionConfig$Comment": "

Any comments you want to include about the streaming distribution.

", - "StreamingDistributionList$Marker": "

The value you provided for the Marker request parameter.

", - "StreamingDistributionList$NextMarker": "

If IsTruncated is true, this element is present and contains the value you can use for the Marker request parameter to continue listing your RTMP distributions where they left off.

", - "StreamingDistributionNotDisabled$Message": null, - "StreamingDistributionSummary$Id": "

The identifier for the distribution, for example, EDFDVBD632BHDS5.

", - "StreamingDistributionSummary$ARN": "

The ARN (Amazon Resource Name) for the streaming distribution. For example: arn:aws:cloudfront::123456789012:streaming-distribution/EDFDVBD632BHDS5, where 123456789012 is your AWS account ID.

", - "StreamingDistributionSummary$Status": "

Indicates the current status of the distribution. When the status is Deployed, the distribution's information is fully propagated throughout the Amazon CloudFront system.

", - "StreamingDistributionSummary$DomainName": "

The domain name corresponding to the distribution, for example, d111111abcdef8.cloudfront.net.

", - "StreamingDistributionSummary$Comment": "

The comment originally specified when this distribution was created.

", - "StreamingLoggingConfig$Bucket": "

The Amazon S3 bucket to store the access logs in, for example, myawslogbucket.s3.amazonaws.com.

", - "StreamingLoggingConfig$Prefix": "

An optional string that you want CloudFront to prefix to the access log filenames for this streaming distribution, for example, myprefix/. If you want to enable logging, but you don't want to specify a prefix, you still must include an empty Prefix element in the Logging element.

", - "TooManyCacheBehaviors$Message": null, - "TooManyCertificates$Message": null, - "TooManyCloudFrontOriginAccessIdentities$Message": null, - "TooManyCookieNamesInWhiteList$Message": null, - "TooManyDistributionCNAMEs$Message": null, - "TooManyDistributions$Message": null, - "TooManyDistributionsWithLambdaAssociations$Message": null, - "TooManyHeadersInForwardedValues$Message": null, - "TooManyInvalidationsInProgress$Message": null, - "TooManyLambdaFunctionAssociations$Message": null, - "TooManyOriginCustomHeaders$Message": null, - "TooManyOrigins$Message": null, - "TooManyQueryStringParameters$Message": null, - "TooManyStreamingDistributionCNAMEs$Message": null, - "TooManyStreamingDistributions$Message": null, - "TooManyTrustedSigners$Message": null, - "TrustedSignerDoesNotExist$Message": null, - "UpdateCloudFrontOriginAccessIdentityRequest$Id": "

The identity's id.

", - "UpdateCloudFrontOriginAccessIdentityRequest$IfMatch": "

The value of the ETag header that you received when retrieving the identity's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateCloudFrontOriginAccessIdentityResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateDistributionRequest$Id": "

The distribution's id.

", - "UpdateDistributionRequest$IfMatch": "

The value of the ETag header that you received when retrieving the distribution's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateDistributionResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateStreamingDistributionRequest$Id": "

The streaming distribution's id.

", - "UpdateStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when retrieving the streaming distribution's configuration. For example: E2QWRUHAPOMQZL.

", - "UpdateStreamingDistributionResult$ETag": "

The current version of the configuration. For example: E2QWRUHAPOMQZL.

", - "ViewerCertificate$IAMCertificateId": "

For information about how and when to use IAMCertificateId, see ViewerCertificate.

", - "ViewerCertificate$ACMCertificateArn": "

For information about how and when to use ACMCertificateArn, see ViewerCertificate.

", - "ViewerCertificate$Certificate": "

This field has been deprecated. Use one of the following fields instead:

" - } - }, - "timestamp": { - "base": null, - "refs": { - "Distribution$LastModifiedTime": "

The date and time the distribution was last modified.

", - "DistributionSummary$LastModifiedTime": "

The date and time the distribution was last modified.

", - "Invalidation$CreateTime": "

The date and time the invalidation request was first made.

", - "InvalidationSummary$CreateTime": null, - "StreamingDistribution$LastModifiedTime": "

The date and time that the distribution was last modified.

", - "StreamingDistributionSummary$LastModifiedTime": "

The date and time the distribution was last modified.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/examples-1.json deleted file mode 100644 index 69a60c20d..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/examples-1.json +++ /dev/null @@ -1,83 +0,0 @@ -{ - "version": "1.0", - "examples": { - "CreateCloudFrontOriginAccessIdentity": [ - null - ], - "CreateDistribution": [ - null - ], - "CreateDistributionWithTags": [ - null - ], - "CreateInvalidation": [ - null - ], - "CreateStreamingDistribution": [ - null - ], - "DeleteCloudFrontOriginAccessIdentity": [ - null - ], - "DeleteDistribution": [ - null - ], - "DeleteStreamingDistribution": [ - null - ], - "GetCloudFrontOriginAccessIdentity": [ - null - ], - "GetCloudFrontOriginAccessIdentityConfig": [ - null - ], - "GetDistribution": [ - null - ], - "GetDistributionConfig": [ - null - ], - "GetInvalidation": [ - null - ], - "GetStreamingDistribution": [ - null - ], - "GetStreamingDistributionConfig": [ - null - ], - "ListCloudFrontOriginAccessIdentities": [ - null - ], - "ListDistributions": [ - null - ], - "ListDistributionsByWebACLId": [ - null - ], - "ListInvalidations": [ - null - ], - "ListStreamingDistributions": [ - null - ], - "ListTagsForResource": [ - null - ], - "TagResource": [ - null - ], - "UntagResource": [ - null - ], - "UpdateCloudFrontOriginAccessIdentity": [ - null - ], - "UpdateDistribution": [ - null - ], - "UpdateStreamingDistribution": [ - null - ] - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/paginators-1.json deleted file mode 100644 index 8edbda230..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/paginators-1.json +++ /dev/null @@ -1,32 +0,0 @@ -{ - "pagination": { - "ListCloudFrontOriginAccessIdentities": { - "input_token": "Marker", - "limit_key": "MaxItems", - "more_results": "CloudFrontOriginAccessIdentityList.IsTruncated", - "output_token": "CloudFrontOriginAccessIdentityList.NextMarker", - "result_key": "CloudFrontOriginAccessIdentityList.Items" - }, - "ListDistributions": { - "input_token": "Marker", - "limit_key": "MaxItems", - "more_results": "DistributionList.IsTruncated", - "output_token": "DistributionList.NextMarker", - "result_key": "DistributionList.Items" - }, - "ListInvalidations": { - "input_token": "Marker", - "limit_key": "MaxItems", - "more_results": "InvalidationList.IsTruncated", - "output_token": "InvalidationList.NextMarker", - "result_key": "InvalidationList.Items" - }, - "ListStreamingDistributions": { - "input_token": "Marker", - "limit_key": "MaxItems", - "more_results": "StreamingDistributionList.IsTruncated", - "output_token": "StreamingDistributionList.NextMarker", - "result_key": "StreamingDistributionList.Items" - } - } -} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/waiters-2.json deleted file mode 100644 index edd74b2a3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-03-25/waiters-2.json +++ /dev/null @@ -1,47 +0,0 @@ -{ - "version": 2, - "waiters": { - "DistributionDeployed": { - "delay": 60, - "operation": "GetDistribution", - "maxAttempts": 25, - "description": "Wait until a distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "Distribution.Status" - } - ] - }, - "InvalidationCompleted": { - "delay": 20, - "operation": "GetInvalidation", - "maxAttempts": 30, - "description": "Wait until an invalidation has completed.", - "acceptors": [ - { - "expected": "Completed", - "matcher": "path", - "state": "success", - "argument": "Invalidation.Status" - } - ] - }, - "StreamingDistributionDeployed": { - "delay": 60, - "operation": "GetStreamingDistribution", - "maxAttempts": 25, - "description": "Wait until a streaming distribution is deployed.", - "acceptors": [ - { - "expected": "Deployed", - "matcher": "path", - "state": "success", - "argument": "StreamingDistribution.Status" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/api-2.json index 0ebda9457..24626f98b 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/api-2.json @@ -343,21 +343,6 @@ {"shape":"PreconditionFailed"} ] }, - "DeleteServiceLinkedRole":{ - "name":"DeleteServiceLinkedRole2017_10_30", - "http":{ - "method":"DELETE", - "requestUri":"/2017-10-30/service-linked-role/{RoleName}", - "responseCode":204 - }, - "input":{"shape":"DeleteServiceLinkedRoleRequest"}, - "errors":[ - {"shape":"InvalidArgument"}, - {"shape":"AccessDenied"}, - {"shape":"ResourceInUse"}, - {"shape":"NoSuchResource"} - ] - }, "DeleteStreamingDistribution":{ "name":"DeleteStreamingDistribution2017_10_30", "http":{ @@ -1539,17 +1524,6 @@ } } }, - "DeleteServiceLinkedRoleRequest":{ - "type":"structure", - "required":["RoleName"], - "members":{ - "RoleName":{ - "shape":"string", - "location":"uri", - "locationName":"RoleName" - } - } - }, "DeleteStreamingDistributionRequest":{ "type":"structure", "required":["Id"], @@ -3182,14 +3156,6 @@ "type":"string", "pattern":"arn:aws:cloudfront::[0-9]+:.*" }, - "ResourceInUse":{ - "type":"structure", - "members":{ - "Message":{"shape":"string"} - }, - "error":{"httpStatusCode":409}, - "exception":true - }, "Restrictions":{ "type":"structure", "required":["GeoRestriction"], diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/docs-2.json index a842f3490..a294a991a 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2017-10-30/docs-2.json @@ -16,7 +16,6 @@ "DeleteFieldLevelEncryptionConfig": "

Remove a field-level encryption configuration.

", "DeleteFieldLevelEncryptionProfile": "

Remove a field-level encryption profile.

", "DeletePublicKey": "

Remove a public key you previously added to CloudFront.

", - "DeleteServiceLinkedRole": null, "DeleteStreamingDistribution": "

Delete a streaming distribution. To delete an RTMP distribution using the CloudFront API, perform the following steps.

To delete an RTMP distribution using the CloudFront API:

  1. Disable the RTMP distribution.

  2. Submit a GET Streaming Distribution Config request to get the current configuration and the Etag header for the distribution.

  3. Update the XML document that was returned in the response to your GET Streaming Distribution Config request to change the value of Enabled to false.

  4. Submit a PUT Streaming Distribution Config request to update the configuration for your distribution. In the request body, include the XML document that you updated in Step 3. Then set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  5. Review the response to the PUT Streaming Distribution Config request to confirm that the distribution was successfully disabled.

  6. Submit a GET Streaming Distribution Config request to confirm that your changes have propagated. When propagation is complete, the value of Status is Deployed.

  7. Submit a DELETE Streaming Distribution request. Set the value of the HTTP If-Match header to the value of the ETag header that CloudFront returned when you submitted the GET Streaming Distribution Config request in Step 2.

  8. Review the response to your DELETE Streaming Distribution request to confirm that the distribution was successfully deleted.

For information about deleting a distribution using the CloudFront console, see Deleting a Distribution in the Amazon CloudFront Developer Guide.

", "GetCloudFrontOriginAccessIdentity": "

Get the information about an origin access identity.

", "GetCloudFrontOriginAccessIdentityConfig": "

Get the configuration information about an origin access identity.

", @@ -377,11 +376,6 @@ "refs": { } }, - "DeleteServiceLinkedRoleRequest": { - "base": null, - "refs": { - } - }, "DeleteStreamingDistributionRequest": { "base": "

The request to delete a streaming distribution.

", "refs": { @@ -1287,11 +1281,6 @@ "UntagResourceRequest$Resource": "

An ARN of a CloudFront resource.

" } }, - "ResourceInUse": { - "base": null, - "refs": { - } - }, "Restrictions": { "base": "

A complex type that identifies ways in which you want to restrict distribution of your content.

", "refs": { @@ -1809,7 +1798,6 @@ "DeleteFieldLevelEncryptionProfileRequest$IfMatch": "

The value of the ETag header that you received when retrieving the profile to delete. For example: E2QWRUHAPOMQZL.

", "DeletePublicKeyRequest$Id": "

The ID of the public key you want to remove from CloudFront.

", "DeletePublicKeyRequest$IfMatch": "

The value of the ETag header that you received when retrieving the public key identity to delete. For example: E2QWRUHAPOMQZL.

", - "DeleteServiceLinkedRoleRequest$RoleName": null, "DeleteStreamingDistributionRequest$Id": "

The distribution ID.

", "DeleteStreamingDistributionRequest$IfMatch": "

The value of the ETag header that you received when you disabled the streaming distribution. For example: E2QWRUHAPOMQZL.

", "Distribution$Id": "

The identifier for the distribution. For example: EDFDVBD632BHDS5.

", @@ -1967,7 +1955,6 @@ "QueryArgProfile$ProfileId": "

ID of profile to use for field-level encryption query argument-profile mapping

", "QueryArgProfileEmpty$Message": null, "QueryStringCacheKeysList$member": null, - "ResourceInUse$Message": null, "S3Origin$DomainName": "

The DNS name of the Amazon S3 origin.

", "S3Origin$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the RTMP distribution. Use an origin access identity to configure the distribution so that end users can only access objects in an Amazon S3 bucket through CloudFront.

If you want end users to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information, see Using an Origin Access Identity to Restrict Access to Your Amazon S3 Content in the Amazon Amazon CloudFront Developer Guide.

", "S3OriginConfig$OriginAccessIdentity": "

The CloudFront origin access identity to associate with the origin. Use an origin access identity to configure the origin so that viewers can only access objects in an Amazon S3 bucket through CloudFront. The format of the value is:

origin-access-identity/cloudfront/ID-of-origin-access-identity

where ID-of-origin-access-identity is the value that CloudFront returned in the ID element when you created the origin access identity.

If you want viewers to be able to access objects using either the CloudFront URL or the Amazon S3 URL, specify an empty OriginAccessIdentity element.

To delete the origin access identity from an existing distribution, update the distribution configuration and include an empty OriginAccessIdentity element.

To replace the origin access identity, update the distribution configuration and specify the new origin access identity.

For more information about the origin access identity, see Serving Private Content through CloudFront in the Amazon CloudFront Developer Guide.

", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/codebuild/2016-10-06/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/codebuild/2016-10-06/api-2.json index 3ebb4f52b..04f2e88ab 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/codebuild/2016-10-06/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/codebuild/2016-10-06/api-2.json @@ -490,7 +490,10 @@ }, "EnvironmentType":{ "type":"string", - "enum":["LINUX_CONTAINER"] + "enum":[ + "WINDOWS_CONTAINER", + "LINUX_CONTAINER" + ] }, "EnvironmentVariable":{ "type":"structure", @@ -658,7 +661,8 @@ "enum":[ "DEBIAN", "AMAZON_LINUX", - "UBUNTU" + "UBUNTU", + "WINDOWS_SERVER" ] }, "Project":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/codedeploy/2014-10-06/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/codedeploy/2014-10-06/docs-2.json index 4d58bd422..66de1a274 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/codedeploy/2014-10-06/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/codedeploy/2014-10-06/docs-2.json @@ -708,7 +708,7 @@ "Duration": { "base": null, "refs": { - "BlueInstanceTerminationOption$terminationWaitTimeInMinutes": "

The number of minutes to wait after a successful blue/green deployment before terminating instances from the original environment.

", + "BlueInstanceTerminationOption$terminationWaitTimeInMinutes": "

The number of minutes to wait after a successful blue/green deployment before terminating instances from the original environment. The maximum setting is 2880 minutes (2 days).

", "DeploymentReadyOption$waitTimeInMinutes": "

The number of minutes to wait before the status of a blue/green deployment changed to Stopped if rerouting is not started manually. Applies only to the STOP_DEPLOYMENT option for actionOnTimeout

" } }, @@ -758,7 +758,7 @@ "ELBInfoList": { "base": null, "refs": { - "LoadBalancerInfo$elbInfoList": "

An array containing information about the load balancer to use for load balancing in a deployment. In Elastic Load Balancing, load balancers are used with Classic Load Balancers.

" + "LoadBalancerInfo$elbInfoList": "

An array containing information about the load balancer to use for load balancing in a deployment. In Elastic Load Balancing, load balancers are used with Classic Load Balancers.

Adding more than one load balancer to the array is not supported.

" } }, "ELBName": { @@ -1824,7 +1824,7 @@ "TargetGroupInfoList": { "base": null, "refs": { - "LoadBalancerInfo$targetGroupInfoList": "

An array containing information about the target group to use for load balancing in a deployment. In Elastic Load Balancing, target groups are used with Application Load Balancers.

" + "LoadBalancerInfo$targetGroupInfoList": "

An array containing information about the target group to use for load balancing in a deployment. In Elastic Load Balancing, target groups are used with Application Load Balancers.

Adding more than one target group to the array is not supported.

" } }, "TargetGroupName": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/codepipeline/2015-07-09/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/codepipeline/2015-07-09/api-2.json index a8a6a44e8..8586604ab 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/codepipeline/2015-07-09/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/codepipeline/2015-07-09/api-2.json @@ -450,7 +450,8 @@ {"shape":"InvalidStageDeclarationException"}, {"shape":"InvalidActionDeclarationException"}, {"shape":"InvalidBlockerDeclarationException"}, - {"shape":"InvalidStructureException"} + {"shape":"InvalidStructureException"}, + {"shape":"LimitExceededException"} ] } }, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/api-2.json index 6c8f49a9a..99fd11b9f 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/api-2.json @@ -3614,7 +3614,8 @@ "SAML", "Facebook", "Google", - "LoginWithAmazon" + "LoginWithAmazon", + "OIDC" ] }, "IdpIdentifierType":{ @@ -5020,7 +5021,8 @@ "EmailConfigurationFailure":{"shape":"StringType"}, "Domain":{"shape":"DomainType"}, "AdminCreateUserConfig":{"shape":"AdminCreateUserConfigType"}, - "UserPoolAddOns":{"shape":"UserPoolAddOnsType"} + "UserPoolAddOns":{"shape":"UserPoolAddOnsType"}, + "Arn":{"shape":"ArnType"} } }, "UserStatusType":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/docs-2.json index af6da4b72..63d8d4f9d 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/cognito-idp/2016-04-18/docs-2.json @@ -95,7 +95,7 @@ "UpdateUserAttributes": "

Allows a user to update a specific attribute (one at a time).

", "UpdateUserPool": "

Updates the specified user pool with the specified attributes.

", "UpdateUserPoolClient": "

Allows the developer to update the specified user pool client and password policy.

", - "VerifySoftwareToken": "

Use this API to register a user's entered TOTP code and mark the user's software token MFA status as \"verified\" if successful,

", + "VerifySoftwareToken": "

Use this API to register a user's entered TOTP code and mark the user's software token MFA status as \"verified\" if successful. The request takes an access token or a session string, but not both.

", "VerifyUserAttribute": "

Verifies the specified user attributes in the user pool.

" }, "shapes": { @@ -460,7 +460,8 @@ "NotifyConfigurationType$SourceArn": "

The Amazon Resource Name (ARN) of the identity that is associated with the sending authorization policy. It permits Amazon Cognito to send for the email address specified in the From parameter.

", "SmsConfigurationType$SnsCallerArn": "

The Amazon Resource Name (ARN) of the Amazon Simple Notification Service (SNS) caller.

", "UpdateGroupRequest$RoleArn": "

The new role ARN for the group. This is used for setting the cognito:roles and cognito:preferred_role claims in the token.

", - "UserImportJobType$CloudWatchLogsRoleArn": "

The role ARN for the Amazon CloudWatch Logging role for the user import job. For more information, see \"Creating the CloudWatch Logs IAM Role\" in the Amazon Cognito Developer Guide.

" + "UserImportJobType$CloudWatchLogsRoleArn": "

The role ARN for the Amazon CloudWatch Logging role for the user import job. For more information, see \"Creating the CloudWatch Logs IAM Role\" in the Amazon Cognito Developer Guide.

", + "UserPoolType$Arn": "

The Amazon Resource Name (ARN) for the user pool.

" } }, "AssociateSoftwareTokenRequest": { @@ -597,7 +598,7 @@ "SMSMfaSettingsType$Enabled": "

Specifies whether SMS text message MFA is enabled.

", "SMSMfaSettingsType$PreferredMfa": "

The preferred MFA method.

", "SchemaAttributeType$DeveloperOnlyAttribute": "

Specifies whether the attribute type is developer only.

", - "SchemaAttributeType$Mutable": "

Specifies whether the attribute can be changed once it has been created.

", + "SchemaAttributeType$Mutable": "

Specifies whether the value of the attribute can be changed.

", "SchemaAttributeType$Required": "

Specifies whether a user pool attribute is required. If the attribute is required and the user does not provide a value, registration or sign-in will fail.

", "SignUpResponse$UserConfirmed": "

A response from the server indicating that a user registration has been confirmed.

", "SoftwareTokenMfaConfigType$Enabled": "

Specifies whether software token MFA is enabled.

", @@ -624,9 +625,9 @@ "CallbackURLsListType": { "base": null, "refs": { - "CreateUserPoolClientRequest$CallbackURLs": "

A list of allowed callback URLs for the identity providers.

", - "UpdateUserPoolClientRequest$CallbackURLs": "

A list of allowed callback URLs for the identity providers.

", - "UserPoolClientType$CallbackURLs": "

A list of allowed callback URLs for the identity providers.

" + "CreateUserPoolClientRequest$CallbackURLs": "

A list of allowed redirect (callback) URLs for the identity providers.

A redirect URI must:

  • Be an absolute URI.

  • Be registered with the authorization server.

  • Not use HTTP without TLS (i.e. use HTTPS instead of HTTP).

  • Not include a fragment component.

See OAuth 2.0 - Redirection Endpoint.

", + "UpdateUserPoolClientRequest$CallbackURLs": "

A list of allowed redirect (callback) URLs for the identity providers.

A redirect URI must:

  • Be an absolute URI.

  • Be registered with the authorization server.

  • Not use HTTP without TLS (i.e. use HTTPS instead of HTTP).

  • Not include a fragment component.

See OAuth 2.0 - Redirection Endpoint.

", + "UserPoolClientType$CallbackURLs": "

A list of allowed redirect (callback) URLs for the identity providers.

A redirect URI must:

  • Be an absolute URI.

  • Be registered with the authorization server.

  • Not use HTTP without TLS (i.e. use HTTPS instead of HTTP).

  • Not include a fragment component.

See OAuth 2.0 - Redirection Endpoint.

" } }, "ChallengeName": { @@ -638,7 +639,7 @@ "ChallengeNameType": { "base": null, "refs": { - "AdminInitiateAuthResponse$ChallengeName": "

The name of the challenge which you are responding to with this call. This is returned to you in the AdminInitiateAuth response if you need to pass another challenge.

  • SMS_MFA: Next challenge is to supply an SMS_MFA_CODE, delivered via SMS.

  • PASSWORD_VERIFIER: Next challenge is to supply PASSWORD_CLAIM_SIGNATURE, PASSWORD_CLAIM_SECRET_BLOCK, and TIMESTAMP after the client-side SRP calculations.

  • CUSTOM_CHALLENGE: This is returned if your custom authentication flow determines that the user should pass another challenge before tokens are issued.

  • DEVICE_SRP_AUTH: If device tracking was enabled on your user pool and the previous challenges were passed, this challenge is returned so that Amazon Cognito can start tracking this device.

  • DEVICE_PASSWORD_VERIFIER: Similar to PASSWORD_VERIFIER, but for devices only.

  • ADMIN_NO_SRP_AUTH: This is returned if you need to authenticate with USERNAME and PASSWORD directly. An app client must be enabled to use this flow.

  • NEW_PASSWORD_REQUIRED: For users which are required to change their passwords after successful first login. This challenge should be passed with NEW_PASSWORD and any other required attributes.

", + "AdminInitiateAuthResponse$ChallengeName": "

The name of the challenge which you are responding to with this call. This is returned to you in the AdminInitiateAuth response if you need to pass another challenge.

  • MFA_SETUP: If MFA is required, users who do not have at least one of the MFA methods set up are presented with an MFA_SETUP challenge. The user must set up at least one MFA type to continue to authenticate.

  • SELECT_MFA_TYPE: Selects the MFA type. Valid MFA options are SMS_MFA for text SMS MFA, and SOFTWARE_TOKEN_MFA for TOTP software token MFA.

  • SMS_MFA: Next challenge is to supply an SMS_MFA_CODE, delivered via SMS.

  • PASSWORD_VERIFIER: Next challenge is to supply PASSWORD_CLAIM_SIGNATURE, PASSWORD_CLAIM_SECRET_BLOCK, and TIMESTAMP after the client-side SRP calculations.

  • CUSTOM_CHALLENGE: This is returned if your custom authentication flow determines that the user should pass another challenge before tokens are issued.

  • DEVICE_SRP_AUTH: If device tracking was enabled on your user pool and the previous challenges were passed, this challenge is returned so that Amazon Cognito can start tracking this device.

  • DEVICE_PASSWORD_VERIFIER: Similar to PASSWORD_VERIFIER, but for devices only.

  • ADMIN_NO_SRP_AUTH: This is returned if you need to authenticate with USERNAME and PASSWORD directly. An app client must be enabled to use this flow.

  • NEW_PASSWORD_REQUIRED: For users which are required to change their passwords after successful first login. This challenge should be passed with NEW_PASSWORD and any other required attributes.

", "AdminRespondToAuthChallengeRequest$ChallengeName": "

The challenge name. For more information, see .

", "AdminRespondToAuthChallengeResponse$ChallengeName": "

The name of the challenge. For more information, see .

", "InitiateAuthResponse$ChallengeName": "

The name of the challenge which you are responding to with this call. This is returned to you in the AdminInitiateAuth response if you need to pass another challenge.

Valid values include the following. Note that all of these challenges require USERNAME and SECRET_HASH (if applicable) in the parameters.

  • SMS_MFA: Next challenge is to supply an SMS_MFA_CODE, delivered via SMS.

  • PASSWORD_VERIFIER: Next challenge is to supply PASSWORD_CLAIM_SIGNATURE, PASSWORD_CLAIM_SECRET_BLOCK, and TIMESTAMP after the client-side SRP calculations.

  • CUSTOM_CHALLENGE: This is returned if your custom authentication flow determines that the user should pass another challenge before tokens are issued.

  • DEVICE_SRP_AUTH: If device tracking was enabled on your user pool and the previous challenges were passed, this challenge is returned so that Amazon Cognito can start tracking this device.

  • DEVICE_PASSWORD_VERIFIER: Similar to PASSWORD_VERIFIER, but for devices only.

  • NEW_PASSWORD_REQUIRED: For users which are required to change their passwords after successful first login. This challenge should be passed with NEW_PASSWORD and any other required attributes.

", @@ -1583,7 +1584,7 @@ "IntegerType": { "base": null, "refs": { - "AuthenticationResultType$ExpiresIn": "

The expiration period of the authentication result.

", + "AuthenticationResultType$ExpiresIn": "

The expiration period of the authentication result in seconds.

", "UserPoolType$EstimatedNumberOfUsers": "

A number estimating the size of the user pool.

" } }, @@ -1638,7 +1639,7 @@ "CreateUserPoolRequest$LambdaConfig": "

The Lambda trigger configuration information for the new user pool.

In a push model, event sources (such as Amazon S3 and custom applications) need permission to invoke a function. So you will need to make an extra call to add permission for these event sources to invoke your Lambda function.

For more information on using the Lambda API to add permission, see AddPermission .

For adding permission using the AWS CLI, see add-permission .

", "UpdateUserPoolRequest$LambdaConfig": "

The AWS Lambda configuration information from the request to update the user pool.

", "UserPoolDescriptionType$LambdaConfig": "

The AWS Lambda configuration information in a user pool description.

", - "UserPoolType$LambdaConfig": "

The AWS Lambda triggers associated with tue user pool.

" + "UserPoolType$LambdaConfig": "

The AWS Lambda triggers associated with the user pool.

" } }, "LimitExceededException": { @@ -2039,10 +2040,10 @@ "base": null, "refs": { "CallbackURLsListType$member": null, - "CreateUserPoolClientRequest$DefaultRedirectURI": "

The default redirect URI. Must be in the CallbackURLs list.

", + "CreateUserPoolClientRequest$DefaultRedirectURI": "

The default redirect URI. Must be in the CallbackURLs list.

A redirect URI must:

  • Be an absolute URI.

  • Be registered with the authorization server.

  • Not use HTTP without TLS (i.e. use HTTPS instead of HTTP).

  • Not include a fragment component.

See OAuth 2.0 - Redirection Endpoint.

", "LogoutURLsListType$member": null, - "UpdateUserPoolClientRequest$DefaultRedirectURI": "

The default redirect URI. Must be in the CallbackURLs list.

", - "UserPoolClientType$DefaultRedirectURI": "

The default redirect URI. Must be in the CallbackURLs list.

" + "UpdateUserPoolClientRequest$DefaultRedirectURI": "

The default redirect URI. Must be in the CallbackURLs list.

A redirect URI must:

  • Be an absolute URI.

  • Be registered with the authorization server.

  • Not use HTTP without TLS (i.e. use HTTPS instead of HTTP).

  • Not include a fragment component.

See OAuth 2.0 - Redirection Endpoint.

", + "UserPoolClientType$DefaultRedirectURI": "

The default redirect URI. Must be in the CallbackURLs list.

A redirect URI must:

  • Be an absolute URI.

  • Be registered with the authorization server.

  • Not use HTTP without TLS (i.e. use HTTPS instead of HTTP).

  • Not include a fragment component.

See OAuth 2.0 - Redirection Endpoint.

" } }, "RefreshTokenValidityType": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/api-2.json index 2d06ef2c7..dd4627a90 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/api-2.json @@ -76,6 +76,66 @@ {"shape":"InternalServerException"} ] }, + "DescribeDominantLanguageDetectionJob":{ + "name":"DescribeDominantLanguageDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDominantLanguageDetectionJobRequest"}, + "output":{"shape":"DescribeDominantLanguageDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, + "DescribeEntitiesDetectionJob":{ + "name":"DescribeEntitiesDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeEntitiesDetectionJobRequest"}, + "output":{"shape":"DescribeEntitiesDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, + "DescribeKeyPhrasesDetectionJob":{ + "name":"DescribeKeyPhrasesDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeKeyPhrasesDetectionJobRequest"}, + "output":{"shape":"DescribeKeyPhrasesDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, + "DescribeSentimentDetectionJob":{ + "name":"DescribeSentimentDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeSentimentDetectionJobRequest"}, + "output":{"shape":"DescribeSentimentDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, "DescribeTopicsDetectionJob":{ "name":"DescribeTopicsDetectionJob", "http":{ @@ -150,6 +210,66 @@ {"shape":"InternalServerException"} ] }, + "ListDominantLanguageDetectionJobs":{ + "name":"ListDominantLanguageDetectionJobs", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListDominantLanguageDetectionJobsRequest"}, + "output":{"shape":"ListDominantLanguageDetectionJobsResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InvalidFilterException"}, + {"shape":"InternalServerException"} + ] + }, + "ListEntitiesDetectionJobs":{ + "name":"ListEntitiesDetectionJobs", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListEntitiesDetectionJobsRequest"}, + "output":{"shape":"ListEntitiesDetectionJobsResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InvalidFilterException"}, + {"shape":"InternalServerException"} + ] + }, + "ListKeyPhrasesDetectionJobs":{ + "name":"ListKeyPhrasesDetectionJobs", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListKeyPhrasesDetectionJobsRequest"}, + "output":{"shape":"ListKeyPhrasesDetectionJobsResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InvalidFilterException"}, + {"shape":"InternalServerException"} + ] + }, + "ListSentimentDetectionJobs":{ + "name":"ListSentimentDetectionJobs", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListSentimentDetectionJobsRequest"}, + "output":{"shape":"ListSentimentDetectionJobsResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InvalidFilterException"}, + {"shape":"InternalServerException"} + ] + }, "ListTopicsDetectionJobs":{ "name":"ListTopicsDetectionJobs", "http":{ @@ -165,6 +285,62 @@ {"shape":"InternalServerException"} ] }, + "StartDominantLanguageDetectionJob":{ + "name":"StartDominantLanguageDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StartDominantLanguageDetectionJobRequest"}, + "output":{"shape":"StartDominantLanguageDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, + "StartEntitiesDetectionJob":{ + "name":"StartEntitiesDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StartEntitiesDetectionJobRequest"}, + "output":{"shape":"StartEntitiesDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, + "StartKeyPhrasesDetectionJob":{ + "name":"StartKeyPhrasesDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StartKeyPhrasesDetectionJobRequest"}, + "output":{"shape":"StartKeyPhrasesDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, + "StartSentimentDetectionJob":{ + "name":"StartSentimentDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StartSentimentDetectionJobRequest"}, + "output":{"shape":"StartSentimentDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"TooManyRequestsException"}, + {"shape":"InternalServerException"} + ] + }, "StartTopicsDetectionJob":{ "name":"StartTopicsDetectionJob", "http":{ @@ -178,6 +354,62 @@ {"shape":"TooManyRequestsException"}, {"shape":"InternalServerException"} ] + }, + "StopDominantLanguageDetectionJob":{ + "name":"StopDominantLanguageDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StopDominantLanguageDetectionJobRequest"}, + "output":{"shape":"StopDominantLanguageDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"InternalServerException"} + ] + }, + "StopEntitiesDetectionJob":{ + "name":"StopEntitiesDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StopEntitiesDetectionJobRequest"}, + "output":{"shape":"StopEntitiesDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"InternalServerException"} + ] + }, + "StopKeyPhrasesDetectionJob":{ + "name":"StopKeyPhrasesDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StopKeyPhrasesDetectionJobRequest"}, + "output":{"shape":"StopKeyPhrasesDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"InternalServerException"} + ] + }, + "StopSentimentDetectionJob":{ + "name":"StopSentimentDetectionJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StopSentimentDetectionJobRequest"}, + "output":{"shape":"StopSentimentDetectionJobResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"JobNotFoundException"}, + {"shape":"InternalServerException"} + ] } }, "shapes":{ @@ -222,7 +454,7 @@ ], "members":{ "TextList":{"shape":"StringList"}, - "LanguageCode":{"shape":"String"} + "LanguageCode":{"shape":"LanguageCode"} } }, "BatchDetectEntitiesResponse":{ @@ -251,7 +483,7 @@ ], "members":{ "TextList":{"shape":"StringList"}, - "LanguageCode":{"shape":"String"} + "LanguageCode":{"shape":"LanguageCode"} } }, "BatchDetectKeyPhrasesResponse":{ @@ -281,7 +513,7 @@ ], "members":{ "TextList":{"shape":"StringList"}, - "LanguageCode":{"shape":"String"} + "LanguageCode":{"shape":"LanguageCode"} } }, "BatchDetectSentimentResponse":{ @@ -320,6 +552,58 @@ "min":1, "pattern":"^[a-zA-Z0-9-]+$" }, + "DescribeDominantLanguageDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "DescribeDominantLanguageDetectionJobResponse":{ + "type":"structure", + "members":{ + "DominantLanguageDetectionJobProperties":{"shape":"DominantLanguageDetectionJobProperties"} + } + }, + "DescribeEntitiesDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "DescribeEntitiesDetectionJobResponse":{ + "type":"structure", + "members":{ + "EntitiesDetectionJobProperties":{"shape":"EntitiesDetectionJobProperties"} + } + }, + "DescribeKeyPhrasesDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "DescribeKeyPhrasesDetectionJobResponse":{ + "type":"structure", + "members":{ + "KeyPhrasesDetectionJobProperties":{"shape":"KeyPhrasesDetectionJobProperties"} + } + }, + "DescribeSentimentDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "DescribeSentimentDetectionJobResponse":{ + "type":"structure", + "members":{ + "SentimentDetectionJobProperties":{"shape":"SentimentDetectionJobProperties"} + } + }, "DescribeTopicsDetectionJobRequest":{ "type":"structure", "required":["JobId"], @@ -405,6 +689,59 @@ "Score":{"shape":"Float"} } }, + "DominantLanguageDetectionJobFilter":{ + "type":"structure", + "members":{ + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "SubmitTimeBefore":{"shape":"Timestamp"}, + "SubmitTimeAfter":{"shape":"Timestamp"} + } + }, + "DominantLanguageDetectionJobProperties":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "Message":{"shape":"AnyLengthString"}, + "SubmitTime":{"shape":"Timestamp"}, + "EndTime":{"shape":"Timestamp"}, + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"} + } + }, + "DominantLanguageDetectionJobPropertiesList":{ + "type":"list", + "member":{"shape":"DominantLanguageDetectionJobProperties"} + }, + "EntitiesDetectionJobFilter":{ + "type":"structure", + "members":{ + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "SubmitTimeBefore":{"shape":"Timestamp"}, + "SubmitTimeAfter":{"shape":"Timestamp"} + } + }, + "EntitiesDetectionJobProperties":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "Message":{"shape":"AnyLengthString"}, + "SubmitTime":{"shape":"Timestamp"}, + "EndTime":{"shape":"Timestamp"}, + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "LanguageCode":{"shape":"LanguageCode"} + } + }, + "EntitiesDetectionJobPropertiesList":{ + "type":"list", + "member":{"shape":"EntitiesDetectionJobProperties"} + }, "Entity":{ "type":"structure", "members":{ @@ -496,7 +833,9 @@ "SUBMITTED", "IN_PROGRESS", "COMPLETED", - "FAILED" + "FAILED", + "STOP_REQUESTED", + "STOPPED" ] }, "KeyPhrase":{ @@ -508,6 +847,33 @@ "EndOffset":{"shape":"Integer"} } }, + "KeyPhrasesDetectionJobFilter":{ + "type":"structure", + "members":{ + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "SubmitTimeBefore":{"shape":"Timestamp"}, + "SubmitTimeAfter":{"shape":"Timestamp"} + } + }, + "KeyPhrasesDetectionJobProperties":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "Message":{"shape":"AnyLengthString"}, + "SubmitTime":{"shape":"Timestamp"}, + "EndTime":{"shape":"Timestamp"}, + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "LanguageCode":{"shape":"LanguageCode"} + } + }, + "KeyPhrasesDetectionJobPropertiesList":{ + "type":"list", + "member":{"shape":"KeyPhrasesDetectionJobProperties"} + }, "LanguageCode":{ "type":"string", "enum":[ @@ -515,6 +881,51 @@ "es" ] }, + "ListDominantLanguageDetectionJobsRequest":{ + "type":"structure", + "members":{ + "Filter":{"shape":"DominantLanguageDetectionJobFilter"}, + "NextToken":{"shape":"String"}, + "MaxResults":{"shape":"MaxResultsInteger"} + } + }, + "ListDominantLanguageDetectionJobsResponse":{ + "type":"structure", + "members":{ + "DominantLanguageDetectionJobPropertiesList":{"shape":"DominantLanguageDetectionJobPropertiesList"}, + "NextToken":{"shape":"String"} + } + }, + "ListEntitiesDetectionJobsRequest":{ + "type":"structure", + "members":{ + "Filter":{"shape":"EntitiesDetectionJobFilter"}, + "NextToken":{"shape":"String"}, + "MaxResults":{"shape":"MaxResultsInteger"} + } + }, + "ListEntitiesDetectionJobsResponse":{ + "type":"structure", + "members":{ + "EntitiesDetectionJobPropertiesList":{"shape":"EntitiesDetectionJobPropertiesList"}, + "NextToken":{"shape":"String"} + } + }, + "ListKeyPhrasesDetectionJobsRequest":{ + "type":"structure", + "members":{ + "Filter":{"shape":"KeyPhrasesDetectionJobFilter"}, + "NextToken":{"shape":"String"}, + "MaxResults":{"shape":"MaxResultsInteger"} + } + }, + "ListKeyPhrasesDetectionJobsResponse":{ + "type":"structure", + "members":{ + "KeyPhrasesDetectionJobPropertiesList":{"shape":"KeyPhrasesDetectionJobPropertiesList"}, + "NextToken":{"shape":"String"} + } + }, "ListOfDetectDominantLanguageResult":{ "type":"list", "member":{"shape":"BatchDetectDominantLanguageItemResult"} @@ -543,6 +954,21 @@ "type":"list", "member":{"shape":"KeyPhrase"} }, + "ListSentimentDetectionJobsRequest":{ + "type":"structure", + "members":{ + "Filter":{"shape":"SentimentDetectionJobFilter"}, + "NextToken":{"shape":"String"}, + "MaxResults":{"shape":"MaxResultsInteger"} + } + }, + "ListSentimentDetectionJobsResponse":{ + "type":"structure", + "members":{ + "SentimentDetectionJobPropertiesList":{"shape":"SentimentDetectionJobPropertiesList"}, + "NextToken":{"shape":"String"} + } + }, "ListTopicsDetectionJobsRequest":{ "type":"structure", "members":{ @@ -578,7 +1004,34 @@ "S3Uri":{ "type":"string", "max":1024, - "pattern":"s3://([^/]+)(/.*)?" + "pattern":"s3://[a-z0-9][\\.\\-a-z0-9]{1,61}[a-z0-9](/.*)?" + }, + "SentimentDetectionJobFilter":{ + "type":"structure", + "members":{ + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "SubmitTimeBefore":{"shape":"Timestamp"}, + "SubmitTimeAfter":{"shape":"Timestamp"} + } + }, + "SentimentDetectionJobProperties":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobName":{"shape":"JobName"}, + "JobStatus":{"shape":"JobStatus"}, + "Message":{"shape":"AnyLengthString"}, + "SubmitTime":{"shape":"Timestamp"}, + "EndTime":{"shape":"Timestamp"}, + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "LanguageCode":{"shape":"LanguageCode"} + } + }, + "SentimentDetectionJobPropertiesList":{ + "type":"list", + "member":{"shape":"SentimentDetectionJobProperties"} }, "SentimentScore":{ "type":"structure", @@ -598,6 +1051,112 @@ "MIXED" ] }, + "StartDominantLanguageDetectionJobRequest":{ + "type":"structure", + "required":[ + "InputDataConfig", + "OutputDataConfig", + "DataAccessRoleArn" + ], + "members":{ + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "DataAccessRoleArn":{"shape":"IamRoleArn"}, + "JobName":{"shape":"JobName"}, + "ClientRequestToken":{ + "shape":"ClientRequestTokenString", + "idempotencyToken":true + } + } + }, + "StartDominantLanguageDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, + "StartEntitiesDetectionJobRequest":{ + "type":"structure", + "required":[ + "InputDataConfig", + "OutputDataConfig", + "DataAccessRoleArn", + "LanguageCode" + ], + "members":{ + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "DataAccessRoleArn":{"shape":"IamRoleArn"}, + "JobName":{"shape":"JobName"}, + "LanguageCode":{"shape":"LanguageCode"}, + "ClientRequestToken":{ + "shape":"ClientRequestTokenString", + "idempotencyToken":true + } + } + }, + "StartEntitiesDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, + "StartKeyPhrasesDetectionJobRequest":{ + "type":"structure", + "required":[ + "InputDataConfig", + "OutputDataConfig", + "DataAccessRoleArn", + "LanguageCode" + ], + "members":{ + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "DataAccessRoleArn":{"shape":"IamRoleArn"}, + "JobName":{"shape":"JobName"}, + "LanguageCode":{"shape":"LanguageCode"}, + "ClientRequestToken":{ + "shape":"ClientRequestTokenString", + "idempotencyToken":true + } + } + }, + "StartKeyPhrasesDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, + "StartSentimentDetectionJobRequest":{ + "type":"structure", + "required":[ + "InputDataConfig", + "OutputDataConfig", + "DataAccessRoleArn", + "LanguageCode" + ], + "members":{ + "InputDataConfig":{"shape":"InputDataConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "DataAccessRoleArn":{"shape":"IamRoleArn"}, + "JobName":{"shape":"JobName"}, + "LanguageCode":{"shape":"LanguageCode"}, + "ClientRequestToken":{ + "shape":"ClientRequestTokenString", + "idempotencyToken":true + } + } + }, + "StartSentimentDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, "StartTopicsDetectionJobRequest":{ "type":"structure", "required":[ @@ -624,6 +1183,62 @@ "JobStatus":{"shape":"JobStatus"} } }, + "StopDominantLanguageDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "StopDominantLanguageDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, + "StopEntitiesDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "StopEntitiesDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, + "StopKeyPhrasesDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "StopKeyPhrasesDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, + "StopSentimentDetectionJobRequest":{ + "type":"structure", + "required":["JobId"], + "members":{ + "JobId":{"shape":"JobId"} + } + }, + "StopSentimentDetectionJobResponse":{ + "type":"structure", + "members":{ + "JobId":{"shape":"JobId"}, + "JobStatus":{"shape":"JobStatus"} + } + }, "String":{ "type":"string", "min":1 diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/docs-2.json index 429c6eebd..7a2e83d60 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/docs-2.json @@ -3,21 +3,41 @@ "service": "

Amazon Comprehend is an AWS service for gaining insight into the content of documents. Use these actions to determine the topics contained in your documents, the topics they discuss, the predominant sentiment expressed in them, the predominant language used, and more.

", "operations": { "BatchDetectDominantLanguage": "

Determines the dominant language of the input text for a batch of documents. For a list of languages that Amazon Comprehend can detect, see Amazon Comprehend Supported Languages.

", - "BatchDetectEntities": "

Inspects the text of a batch of documents and returns information about them. For more information about entities, see how-entities

", + "BatchDetectEntities": "

Inspects the text of a batch of documents for named entities and returns information about them. For more information about named entities, see how-entities

", "BatchDetectKeyPhrases": "

Detects the key noun phrases found in a batch of documents.

", "BatchDetectSentiment": "

Inspects a batch of documents and returns an inference of the prevailing sentiment, POSITIVE, NEUTRAL, MIXED, or NEGATIVE, in each one.

", + "DescribeDominantLanguageDetectionJob": "

Gets the properties associated with a dominant language detection job. Use this operation to get the status of a detection job.

", + "DescribeEntitiesDetectionJob": "

Gets the properties associated with an entities detection job. Use this operation to get the status of a detection job.

", + "DescribeKeyPhrasesDetectionJob": "

Gets the properties associated with a key phrases detection job. Use this operation to get the status of a detection job.

", + "DescribeSentimentDetectionJob": "

Gets the properties associated with a sentiment detection job. Use this operation to get the status of a detection job.

", "DescribeTopicsDetectionJob": "

Gets the properties associated with a topic detection job. Use this operation to get the status of a detection job.

", "DetectDominantLanguage": "

Determines the dominant language of the input text. For a list of languages that Amazon Comprehend can detect, see Amazon Comprehend Supported Languages.

", - "DetectEntities": "

Inspects text for entities, and returns information about them. For more information, about entities, see how-entities.

", + "DetectEntities": "

Inspects text for named entities, and returns information about them. For more information, about named entities, see how-entities.

", "DetectKeyPhrases": "

Detects the key noun phrases found in the text.

", "DetectSentiment": "

Inspects text and returns an inference of the prevailing sentiment (POSITIVE, NEUTRAL, MIXED, or NEGATIVE).

", + "ListDominantLanguageDetectionJobs": "

Gets a list of the dominant language detection jobs that you have submitted.

", + "ListEntitiesDetectionJobs": "

Gets a list of the entity detection jobs that you have submitted.

", + "ListKeyPhrasesDetectionJobs": "

Get a list of key phrase detection jobs that you have submitted.

", + "ListSentimentDetectionJobs": "

Gets a list of sentiment detection jobs that you have submitted.

", "ListTopicsDetectionJobs": "

Gets a list of the topic detection jobs that you have submitted.

", - "StartTopicsDetectionJob": "

Starts an asynchronous topic detection job. Use the DescribeTopicDetectionJob operation to track the status of a job.

" + "StartDominantLanguageDetectionJob": "

Starts an asynchronous dominant language detection job for a collection of documents. Use the operation to track the status of a job.

", + "StartEntitiesDetectionJob": "

Starts an asynchronous entity detection job for a collection of documents. Use the operation to track the status of a job.

", + "StartKeyPhrasesDetectionJob": "

Starts an asynchronous key phrase detection job for a collection of documents. Use the operation to track the status of a job.

", + "StartSentimentDetectionJob": "

Starts an asynchronous sentiment detection job for a collection of documents. use the operation to track the status of a job.

", + "StartTopicsDetectionJob": "

Starts an asynchronous topic detection job. Use the DescribeTopicDetectionJob operation to track the status of a job.

", + "StopDominantLanguageDetectionJob": "

Stops a dominant language detection job in progress.

If the job state is IN_PROGRESS the job will be marked for termination and put into the STOPPING state.

If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob operation, the operation will return a 400 Internal Request Exception.

When a job is stopped, any document that has already been processed will be written to the output location.

", + "StopEntitiesDetectionJob": "

Stops an entities detection job in progress.

If the job state is IN_PROGRESS the job will be marked for termination and put into the STOPPING state.

If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob operation, the operation will return a 400 Internal Request Exception.

When a job is stopped, any document that has already been processed will be written to the output location.

", + "StopKeyPhrasesDetectionJob": "

Stops a key phrases detection job in progress.

If the job state is IN_PROGRESS the job will be marked for termination and put into the STOPPING state.

If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob operation, the operation will return a 400 Internal Request Exception.

When a job is stopped, any document that has already been processed will be written to the output location.

", + "StopSentimentDetectionJob": "

Stops a sentiment detection job in progress.

If the job state is IN_PROGRESS the job will be marked for termination and put into the STOPPING state.

If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob operation, the operation will return a 400 Internal Request Exception.

When a job is stopped, any document that has already been processed will be written to the output location.

" }, "shapes": { "AnyLengthString": { "base": null, "refs": { + "DominantLanguageDetectionJobProperties$Message": "

A description for the status of a job.

", + "EntitiesDetectionJobProperties$Message": "

A description of the status of a job.

", + "KeyPhrasesDetectionJobProperties$Message": "

A description of the status of a job.

", + "SentimentDetectionJobProperties$Message": "

A description of the status of a job.

", "TopicsDetectionJobProperties$Message": "

A description for the status of a job.

" } }, @@ -108,9 +128,53 @@ "ClientRequestTokenString": { "base": null, "refs": { + "StartDominantLanguageDetectionJobRequest$ClientRequestToken": "

A unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.

", + "StartEntitiesDetectionJobRequest$ClientRequestToken": "

A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.

", + "StartKeyPhrasesDetectionJobRequest$ClientRequestToken": "

A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.

", + "StartSentimentDetectionJobRequest$ClientRequestToken": "

A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.

", "StartTopicsDetectionJobRequest$ClientRequestToken": "

A unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.

" } }, + "DescribeDominantLanguageDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "DescribeDominantLanguageDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "DescribeEntitiesDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "DescribeEntitiesDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "DescribeKeyPhrasesDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "DescribeKeyPhrasesDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "DescribeSentimentDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "DescribeSentimentDetectionJobResponse": { + "base": null, + "refs": { + } + }, "DescribeTopicsDetectionJobRequest": { "base": null, "refs": { @@ -167,6 +231,44 @@ "ListOfDominantLanguages$member": null } }, + "DominantLanguageDetectionJobFilter": { + "base": "

Provides information for filtering a list of dominant language detection jobs. For more information, see the operation.

", + "refs": { + "ListDominantLanguageDetectionJobsRequest$Filter": "

Filters that jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.

" + } + }, + "DominantLanguageDetectionJobProperties": { + "base": "

Provides information about a dominant language detection job.

", + "refs": { + "DescribeDominantLanguageDetectionJobResponse$DominantLanguageDetectionJobProperties": "

An object that contains the properties associated with a dominant language detection job.

", + "DominantLanguageDetectionJobPropertiesList$member": null + } + }, + "DominantLanguageDetectionJobPropertiesList": { + "base": null, + "refs": { + "ListDominantLanguageDetectionJobsResponse$DominantLanguageDetectionJobPropertiesList": "

A list containing the properties of each job that is returned.

" + } + }, + "EntitiesDetectionJobFilter": { + "base": "

Provides information for filtering a list of dominant language detection jobs. For more information, see the operation.

", + "refs": { + "ListEntitiesDetectionJobsRequest$Filter": "

Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.

" + } + }, + "EntitiesDetectionJobProperties": { + "base": "

Provides information about an entities detection job.

", + "refs": { + "DescribeEntitiesDetectionJobResponse$EntitiesDetectionJobProperties": "

An object that contains the properties associated with an entities detection job.

", + "EntitiesDetectionJobPropertiesList$member": null + } + }, + "EntitiesDetectionJobPropertiesList": { + "base": null, + "refs": { + "ListEntitiesDetectionJobsResponse$EntitiesDetectionJobPropertiesList": "

A list containing the properties of each job that is returned.

" + } + }, "Entity": { "base": "

Provides information about an entity.

", "refs": { @@ -194,12 +296,24 @@ "IamRoleArn": { "base": null, "refs": { + "StartDominantLanguageDetectionJobRequest$DataAccessRoleArn": "

The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.

", + "StartEntitiesDetectionJobRequest$DataAccessRoleArn": "

The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.

", + "StartKeyPhrasesDetectionJobRequest$DataAccessRoleArn": "

The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.

", + "StartSentimentDetectionJobRequest$DataAccessRoleArn": "

The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.

", "StartTopicsDetectionJobRequest$DataAccessRoleArn": "

The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.

" } }, "InputDataConfig": { "base": "

The input properties for a topic detection job.

", "refs": { + "DominantLanguageDetectionJobProperties$InputDataConfig": "

The input data configuration that you supplied when you created the dominant language detection job.

", + "EntitiesDetectionJobProperties$InputDataConfig": "

The input data configuration that you supplied when you created the entities detection job.

", + "KeyPhrasesDetectionJobProperties$InputDataConfig": "

The input data configuration that you supplied when you created the key phrases detection job.

", + "SentimentDetectionJobProperties$InputDataConfig": "

The input data configuration that you supplied when you created the sentiment detection job.

", + "StartDominantLanguageDetectionJobRequest$InputDataConfig": "

Specifies the format and location of the input data for the job.

", + "StartEntitiesDetectionJobRequest$InputDataConfig": "

Specifies the format and location of the input data for the job.

", + "StartKeyPhrasesDetectionJobRequest$InputDataConfig": "

Specifies the format and location of the input data for the job.

", + "StartSentimentDetectionJobRequest$InputDataConfig": "

Specifies the format and location of the input data for the job.

", "StartTopicsDetectionJobRequest$InputDataConfig": "

Specifies the format and location of the input data for the job.

", "TopicsDetectionJobProperties$InputDataConfig": "

The input data configuration supplied when you created the topic detection job.

" } @@ -243,14 +357,46 @@ "JobId": { "base": null, "refs": { + "DescribeDominantLanguageDetectionJobRequest$JobId": "

The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.

", + "DescribeEntitiesDetectionJobRequest$JobId": "

The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.

", + "DescribeKeyPhrasesDetectionJobRequest$JobId": "

The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.

", + "DescribeSentimentDetectionJobRequest$JobId": "

The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.

", "DescribeTopicsDetectionJobRequest$JobId": "

The identifier assigned by the user to the detection job.

", + "DominantLanguageDetectionJobProperties$JobId": "

The identifier assigned to the dominant language detection job.

", + "EntitiesDetectionJobProperties$JobId": "

The identifier assigned to the entities detection job.

", + "KeyPhrasesDetectionJobProperties$JobId": "

The identifier assigned to the key phrases detection job.

", + "SentimentDetectionJobProperties$JobId": "

The identifier assigned to the sentiment detection job.

", + "StartDominantLanguageDetectionJobResponse$JobId": "

The identifier generated for the job. To get the status of a job, use this identifier with the operation.

", + "StartEntitiesDetectionJobResponse$JobId": "

The identifier generated for the job. To get the status of job, use this identifier with the operation.

", + "StartKeyPhrasesDetectionJobResponse$JobId": "

The identifier generated for the job. To get the status of a job, use this identifier with the operation.

", + "StartSentimentDetectionJobResponse$JobId": "

The identifier generated for the job. To get the status of a job, use this identifier with the operation.

", "StartTopicsDetectionJobResponse$JobId": "

The identifier generated for the job. To get the status of the job, use this identifier with the DescribeTopicDetectionJob operation.

", + "StopDominantLanguageDetectionJobRequest$JobId": "

The identifier of the dominant language detection job to stop.

", + "StopDominantLanguageDetectionJobResponse$JobId": "

The identifier of the dominant language detection job to stop.

", + "StopEntitiesDetectionJobRequest$JobId": "

The identifier of the entities detection job to stop.

", + "StopEntitiesDetectionJobResponse$JobId": "

The identifier of the entities detection job to stop.

", + "StopKeyPhrasesDetectionJobRequest$JobId": "

The identifier of the key phrases detection job to stop.

", + "StopKeyPhrasesDetectionJobResponse$JobId": "

The identifier of the key phrases detection job to stop.

", + "StopSentimentDetectionJobRequest$JobId": "

The identifier of the sentiment detection job to stop.

", + "StopSentimentDetectionJobResponse$JobId": "

The identifier of the sentiment detection job to stop.

", "TopicsDetectionJobProperties$JobId": "

The identifier assigned to the topic detection job.

" } }, "JobName": { "base": null, "refs": { + "DominantLanguageDetectionJobFilter$JobName": "

Filters on the name of the job.

", + "DominantLanguageDetectionJobProperties$JobName": "

The name that you assigned to the dominant language detection job.

", + "EntitiesDetectionJobFilter$JobName": "

Filters on the name of the job.

", + "EntitiesDetectionJobProperties$JobName": "

The name that you assigned the entities detection job.

", + "KeyPhrasesDetectionJobFilter$JobName": "

Filters on the name of the job.

", + "KeyPhrasesDetectionJobProperties$JobName": "

The name that you assigned the key phrases detection job.

", + "SentimentDetectionJobFilter$JobName": "

Filters on the name of the job.

", + "SentimentDetectionJobProperties$JobName": "

The name that you assigned to the sentiment detection job

", + "StartDominantLanguageDetectionJobRequest$JobName": "

An identifier for the job.

", + "StartEntitiesDetectionJobRequest$JobName": "

The identifier of the job.

", + "StartKeyPhrasesDetectionJobRequest$JobName": "

The identifier of the job.

", + "StartSentimentDetectionJobRequest$JobName": "

The identifier of the job.

", "StartTopicsDetectionJobRequest$JobName": "

The identifier of the job.

", "TopicsDetectionJobFilter$JobName": "

", "TopicsDetectionJobProperties$JobName": "

The name of the topic detection job.

" @@ -264,7 +410,23 @@ "JobStatus": { "base": null, "refs": { + "DominantLanguageDetectionJobFilter$JobStatus": "

Filters the list of jobs based on job status. Returns only jobs with the specified status.

", + "DominantLanguageDetectionJobProperties$JobStatus": "

The current status of the dominant language detection job. If the status is FAILED, the Message field shows the reason for the failure.

", + "EntitiesDetectionJobFilter$JobStatus": "

Filters the list of jobs based on job status. Returns only jobs with the specified status.

", + "EntitiesDetectionJobProperties$JobStatus": "

The current status of the entities detection job. If the status is FAILED, the Message field shows the reason for the failure.

", + "KeyPhrasesDetectionJobFilter$JobStatus": "

Filters the list of jobs based on job status. Returns only jobs with the specified status.

", + "KeyPhrasesDetectionJobProperties$JobStatus": "

The current status of the key phrases detection job. If the status is FAILED, the Message field shows the reason for the failure.

", + "SentimentDetectionJobFilter$JobStatus": "

Filters the list of jobs based on job status. Returns only jobs with the specified status.

", + "SentimentDetectionJobProperties$JobStatus": "

The current status of the sentiment detection job. If the status is FAILED, the Messages field shows the reason for the failure.

", + "StartDominantLanguageDetectionJobResponse$JobStatus": "

The status of the job.

  • SUBMITTED - The job has been received and is queued for processing.

  • IN_PROGRESS - Amazon Comprehend is processing the job.

  • COMPLETED - The job was successfully completed and the output is available.

  • FAILED - The job did not complete. To get details, use the operation.

", + "StartEntitiesDetectionJobResponse$JobStatus": "

The status of the job.

  • SUBMITTED - The job has been received and is queued for processing.

  • IN_PROGRESS - Amazon Comprehend is processing the job.

  • COMPLETED - The job was successfully completed and the output is available.

  • FAILED - The job did not complete. To get details, use the operation.

", + "StartKeyPhrasesDetectionJobResponse$JobStatus": "

The status of the job.

  • SUBMITTED - The job has been received and is queued for processing.

  • IN_PROGRESS - Amazon Comprehend is processing the job.

  • COMPLETED - The job was successfully completed and the output is available.

  • FAILED - The job did not complete. To get details, use the operation.

", + "StartSentimentDetectionJobResponse$JobStatus": "

The status of the job.

  • SUBMITTED - The job has been received and is queued for processing.

  • IN_PROGRESS - Amazon Comprehend is processing the job.

  • COMPLETED - The job was successfully completed and the output is available.

  • FAILED - The job did not complete. To get details, use the operation.

", "StartTopicsDetectionJobResponse$JobStatus": "

The status of the job:

  • SUBMITTED - The job has been received and is queued for processing.

  • IN_PROGRESS - Amazon Comprehend is processing the job.

  • COMPLETED - The job was successfully completed and the output is available.

  • FAILED - The job did not complete. To get details, use the DescribeTopicDetectionJob operation.

", + "StopDominantLanguageDetectionJobResponse$JobStatus": "

Either STOPPING if the job is currently running, or STOPPED if the job was previously stopped with the StopDominantLanguageDetectionJob operation.

", + "StopEntitiesDetectionJobResponse$JobStatus": "

Either STOPPING if the job is currently running, or STOPPED if the job was previously stopped with the StopEntitiesDetectionJob operation.

", + "StopKeyPhrasesDetectionJobResponse$JobStatus": "

Either STOPPING if the job is currently running, or STOPPED if the job was previously stopped with the StopKeyPhrasesDetectionJob operation.

", + "StopSentimentDetectionJobResponse$JobStatus": "

Either STOPPING if the job is currently running, or STOPPED if the job was previously stopped with the StopSentimentDetectionJob operation.

", "TopicsDetectionJobFilter$JobStatus": "

Filters the list of topic detection jobs based on job status. Returns only jobs with the specified status.

", "TopicsDetectionJobProperties$JobStatus": "

The current status of the topic detection job. If the status is Failed, the reason for the failure is shown in the Message field.

" } @@ -275,12 +437,70 @@ "ListOfKeyPhrases$member": null } }, + "KeyPhrasesDetectionJobFilter": { + "base": "

Provides information for filtering a list of dominant language detection jobs. For more information, see the operation.

", + "refs": { + "ListKeyPhrasesDetectionJobsRequest$Filter": "

Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.

" + } + }, + "KeyPhrasesDetectionJobProperties": { + "base": "

Provides information about a key phrases detection job.

", + "refs": { + "DescribeKeyPhrasesDetectionJobResponse$KeyPhrasesDetectionJobProperties": "

An object that contains the properties associated with a key phrases detection job.

", + "KeyPhrasesDetectionJobPropertiesList$member": null + } + }, + "KeyPhrasesDetectionJobPropertiesList": { + "base": null, + "refs": { + "ListKeyPhrasesDetectionJobsResponse$KeyPhrasesDetectionJobPropertiesList": "

A list containing the properties of each job that is returned.

" + } + }, "LanguageCode": { "base": null, "refs": { - "DetectEntitiesRequest$LanguageCode": "

The RFC 5646 language code of the input text. If the request does not specify the language code, the service detects the dominant language. If you specify a language code that the service does not support, it returns UnsupportedLanguageException exception. For more information about RFC 5646, see Tags for Identifying Languages on the IETF Tools web site.

", - "DetectKeyPhrasesRequest$LanguageCode": "

The RFC 5646 language code for the input text. If you don't specify a language code, Amazon Comprehend detects the dominant language. If you specify the code for a language that Amazon Comprehend does not support, it returns and UnsupportedLanguageException. For more information about RFC 5646, see Tags for Identifying Languages on the IETF Tools web site.

", - "DetectSentimentRequest$LanguageCode": "

The RFC 5646 language code for the input text. If you don't specify a language code, Amazon Comprehend detects the dominant language. If you specify the code for a language that Amazon Comprehend does not support, it returns and UnsupportedLanguageException. For more information about RFC 5646, see Tags for Identifying Languages on the IETF Tools web site.

" + "BatchDetectEntitiesRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "BatchDetectKeyPhrasesRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "BatchDetectSentimentRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "DetectEntitiesRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "DetectKeyPhrasesRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "DetectSentimentRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "EntitiesDetectionJobProperties$LanguageCode": "

The language code of the input documents.

", + "KeyPhrasesDetectionJobProperties$LanguageCode": "

The language code of the input documents.

", + "SentimentDetectionJobProperties$LanguageCode": "

The language code of the input documents.

", + "StartEntitiesDetectionJobRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "StartKeyPhrasesDetectionJobRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

", + "StartSentimentDetectionJobRequest$LanguageCode": "

The language of the input documents. You can specify English (\"en\") or Spanish (\"es\"). All documents must be in the same language.

" + } + }, + "ListDominantLanguageDetectionJobsRequest": { + "base": null, + "refs": { + } + }, + "ListDominantLanguageDetectionJobsResponse": { + "base": null, + "refs": { + } + }, + "ListEntitiesDetectionJobsRequest": { + "base": null, + "refs": { + } + }, + "ListEntitiesDetectionJobsResponse": { + "base": null, + "refs": { + } + }, + "ListKeyPhrasesDetectionJobsRequest": { + "base": null, + "refs": { + } + }, + "ListKeyPhrasesDetectionJobsResponse": { + "base": null, + "refs": { } }, "ListOfDetectDominantLanguageResult": { @@ -328,6 +548,16 @@ "DetectKeyPhrasesResponse$KeyPhrases": "

A collection of key phrases that Amazon Comprehend identified in the input text. For each key phrase, the response provides the text of the key phrase, where the key phrase begins and ends, and the level of confidence that Amazon Comprehend has in the accuracy of the detection.

" } }, + "ListSentimentDetectionJobsRequest": { + "base": null, + "refs": { + } + }, + "ListSentimentDetectionJobsResponse": { + "base": null, + "refs": { + } + }, "ListTopicsDetectionJobsRequest": { "base": null, "refs": { @@ -341,7 +571,11 @@ "MaxResultsInteger": { "base": null, "refs": { - "ListTopicsDetectionJobsRequest$MaxResults": "

The maximum number of results to return in each page.

" + "ListDominantLanguageDetectionJobsRequest$MaxResults": "

The maximum number of results to return in each page. The default is 100.

", + "ListEntitiesDetectionJobsRequest$MaxResults": "

The maximum number of results to return in each page. The default is 100.

", + "ListKeyPhrasesDetectionJobsRequest$MaxResults": "

The maximum number of results to return in each page. The default is 100.

", + "ListSentimentDetectionJobsRequest$MaxResults": "

The maximum number of results to return in each page. The default is 100.

", + "ListTopicsDetectionJobsRequest$MaxResults": "

The maximum number of results to return in each page. The default is 100.

" } }, "NumberOfTopicsInteger": { @@ -353,7 +587,15 @@ "OutputDataConfig": { "base": "

Provides configuration parameters for the output of topic detection jobs.

", "refs": { - "StartTopicsDetectionJobRequest$OutputDataConfig": "

Specifies where to send the output files.

", + "DominantLanguageDetectionJobProperties$OutputDataConfig": "

The output data configuration that you supplied when you created the dominant language detection job.

", + "EntitiesDetectionJobProperties$OutputDataConfig": "

The output data configuration that you supplied when you created the entities detection job.

", + "KeyPhrasesDetectionJobProperties$OutputDataConfig": "

The output data configuration that you supplied when you created the key phrases detection job.

", + "SentimentDetectionJobProperties$OutputDataConfig": "

The output data configuration that you supplied when you created the sentiment detection job.

", + "StartDominantLanguageDetectionJobRequest$OutputDataConfig": "

Specifies where to send the output files.

", + "StartEntitiesDetectionJobRequest$OutputDataConfig": "

Specifies where to send the output files.

", + "StartKeyPhrasesDetectionJobRequest$OutputDataConfig": "

Specifies where to send the output files.

", + "StartSentimentDetectionJobRequest$OutputDataConfig": "

Specifies where to send the output files.

", + "StartTopicsDetectionJobRequest$OutputDataConfig": "

Specifies where to send the output files. The output is a compressed archive with two files, topic-terms.csv that lists the terms associated with each topic, and doc-topics.csv that lists the documents associated with each topic

", "TopicsDetectionJobProperties$OutputDataConfig": "

The output data configuration supplied when you created the topic detection job.

" } }, @@ -361,7 +603,26 @@ "base": null, "refs": { "InputDataConfig$S3Uri": "

The Amazon S3 URI for the input data. The URI must be in same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of data files.

For example, if you use the URI S3://bucketName/prefix, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.

", - "OutputDataConfig$S3Uri": "

The Amazon S3 URI where you want to write the output data. The URI must be in the same region as the API endpoint that you are calling.

The service creates an output file called output.tar.gz. It is a compressed archive that contains two files, topic-terms.csv that lists the terms associated with each topic, and doc-topics.csv that lists the documents associated with each topic. For more information, see topic-modeling.

" + "OutputDataConfig$S3Uri": "

When you use the OutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of the output file.

When the topic detection job is finished, the service creates an output file in a directory specific to the job. The S3Uri field contains the location of the output file, called output.tar.gz. It is a compressed archive that contains the ouput of the operation.

" + } + }, + "SentimentDetectionJobFilter": { + "base": "

Provides information for filtering a list of dominant language detection jobs. For more information, see the operation.

", + "refs": { + "ListSentimentDetectionJobsRequest$Filter": "

Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.

" + } + }, + "SentimentDetectionJobProperties": { + "base": "

Provides information about a sentiment detection job.

", + "refs": { + "DescribeSentimentDetectionJobResponse$SentimentDetectionJobProperties": "

An object that contains the properties associated with a sentiment detection job.

", + "SentimentDetectionJobPropertiesList$member": null + } + }, + "SentimentDetectionJobPropertiesList": { + "base": null, + "refs": { + "ListSentimentDetectionJobsResponse$SentimentDetectionJobPropertiesList": "

A list containing the properties of each job that is returned.

" } }, "SentimentScore": { @@ -378,6 +639,46 @@ "DetectSentimentResponse$Sentiment": "

The inferred sentiment that Amazon Comprehend has the highest level of confidence in.

" } }, + "StartDominantLanguageDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StartDominantLanguageDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "StartEntitiesDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StartEntitiesDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "StartKeyPhrasesDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StartKeyPhrasesDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "StartSentimentDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StartSentimentDetectionJobResponse": { + "base": null, + "refs": { + } + }, "StartTopicsDetectionJobRequest": { "base": null, "refs": { @@ -388,12 +689,49 @@ "refs": { } }, + "StopDominantLanguageDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StopDominantLanguageDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "StopEntitiesDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StopEntitiesDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "StopKeyPhrasesDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StopKeyPhrasesDetectionJobResponse": { + "base": null, + "refs": { + } + }, + "StopSentimentDetectionJobRequest": { + "base": null, + "refs": { + } + }, + "StopSentimentDetectionJobResponse": { + "base": null, + "refs": { + } + }, "String": { "base": null, "refs": { - "BatchDetectEntitiesRequest$LanguageCode": "

The language of the input documents. All documents must be in the same language.

", - "BatchDetectKeyPhrasesRequest$LanguageCode": "

The language of the input documents. All documents must be in the same language.

", - "BatchDetectSentimentRequest$LanguageCode": "

The language of the input documents. All documents must be in the same language.

", "BatchItemError$ErrorCode": "

The numeric error code of the error.

", "BatchItemError$ErrorMessage": "

A text description of the error.

", "BatchSizeLimitExceededException$Message": null, @@ -401,13 +739,21 @@ "DetectEntitiesRequest$Text": "

A UTF-8 text string. Each string must contain fewer that 5,000 bytes of UTF-8 encoded characters.

", "DetectKeyPhrasesRequest$Text": "

A UTF-8 text string. Each string must contain fewer that 5,000 bytes of UTF-8 encoded characters.

", "DetectSentimentRequest$Text": "

A UTF-8 text string. Each string must contain fewer that 5,000 bytes of UTF-8 encoded characters.

", - "DominantLanguage$LanguageCode": "

The RFC 5646 language code for the dominant language.

", + "DominantLanguage$LanguageCode": "

The RFC 5646 language code for the dominant language. For more information about RFC 5646, see Tags for Identifying Languages on the IETF Tools web site.

", "Entity$Text": "

The text of the entity.

", "InternalServerException$Message": null, "InvalidFilterException$Message": null, "InvalidRequestException$Message": null, "JobNotFoundException$Message": null, "KeyPhrase$Text": "

The text of a key noun phrase.

", + "ListDominantLanguageDetectionJobsRequest$NextToken": "

Identifies the next page of results to return.

", + "ListDominantLanguageDetectionJobsResponse$NextToken": "

Identifies the next page of results to return.

", + "ListEntitiesDetectionJobsRequest$NextToken": "

Identifies the next page of results to return.

", + "ListEntitiesDetectionJobsResponse$NextToken": "

Identifies the next page of results to return.

", + "ListKeyPhrasesDetectionJobsRequest$NextToken": "

Identifies the next page of results to return.

", + "ListKeyPhrasesDetectionJobsResponse$NextToken": "

Identifies the next page of results to return.

", + "ListSentimentDetectionJobsRequest$NextToken": "

Identifies the next page of results to return.

", + "ListSentimentDetectionJobsResponse$NextToken": "

Identifies the next page of results to return.

", "ListTopicsDetectionJobsRequest$NextToken": "

Identifies the next page of results to return.

", "ListTopicsDetectionJobsResponse$NextToken": "

Identifies the next page of results to return.

", "StringList$member": null, @@ -433,6 +779,22 @@ "Timestamp": { "base": null, "refs": { + "DominantLanguageDetectionJobFilter$SubmitTimeBefore": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest.

", + "DominantLanguageDetectionJobFilter$SubmitTimeAfter": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest.

", + "DominantLanguageDetectionJobProperties$SubmitTime": "

The time that the dominant language detection job was submitted for processing.

", + "DominantLanguageDetectionJobProperties$EndTime": "

The time that the dominant language detection job completed.

", + "EntitiesDetectionJobFilter$SubmitTimeBefore": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest.

", + "EntitiesDetectionJobFilter$SubmitTimeAfter": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest.

", + "EntitiesDetectionJobProperties$SubmitTime": "

The time that the entities detection job was submitted for processing.

", + "EntitiesDetectionJobProperties$EndTime": "

The time that the entities detection job completed

", + "KeyPhrasesDetectionJobFilter$SubmitTimeBefore": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest.

", + "KeyPhrasesDetectionJobFilter$SubmitTimeAfter": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest.

", + "KeyPhrasesDetectionJobProperties$SubmitTime": "

The time that the key phrases detection job was submitted for processing.

", + "KeyPhrasesDetectionJobProperties$EndTime": "

The time that the key phrases detection job completed.

", + "SentimentDetectionJobFilter$SubmitTimeBefore": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest.

", + "SentimentDetectionJobFilter$SubmitTimeAfter": "

Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest.

", + "SentimentDetectionJobProperties$SubmitTime": "

The time that the sentiment detection job was submitted for processing.

", + "SentimentDetectionJobProperties$EndTime": "

The time that the sentiment detection job ended.

", "TopicsDetectionJobFilter$SubmitTimeBefore": "

Filters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.

", "TopicsDetectionJobFilter$SubmitTimeAfter": "

Filters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.

", "TopicsDetectionJobProperties$SubmitTime": "

The time that the topic detection job was submitted for processing.

", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/paginators-1.json index a117c0e24..04005c5e5 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/comprehend/2017-11-27/paginators-1.json @@ -1,5 +1,25 @@ { "pagination": { + "ListDominantLanguageDetectionJobs": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" + }, + "ListEntitiesDetectionJobs": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" + }, + "ListKeyPhrasesDetectionJobs": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" + }, + "ListSentimentDetectionJobs": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" + }, "ListTopicsDetectionJobs": { "input_token": "NextToken", "output_token": "NextToken", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/api-2.json index 6dd713dde..a3d12e758 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/api-2.json @@ -107,6 +107,18 @@ {"shape":"InvalidParameterValueException"} ] }, + "DeleteRetentionConfiguration":{ + "name":"DeleteRetentionConfiguration", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DeleteRetentionConfigurationRequest"}, + "errors":[ + {"shape":"InvalidParameterValueException"}, + {"shape":"NoSuchRetentionConfigurationException"} + ] + }, "DeliverConfigSnapshot":{ "name":"DeliverConfigSnapshot", "http":{ @@ -296,6 +308,20 @@ {"shape":"InvalidLimitException"} ] }, + "DescribeRetentionConfigurations":{ + "name":"DescribeRetentionConfigurations", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeRetentionConfigurationsRequest"}, + "output":{"shape":"DescribeRetentionConfigurationsResponse"}, + "errors":[ + {"shape":"InvalidParameterValueException"}, + {"shape":"NoSuchRetentionConfigurationException"}, + {"shape":"InvalidNextTokenException"} + ] + }, "GetAggregateComplianceDetailsByConfigRule":{ "name":"GetAggregateComplianceDetailsByConfigRule", "http":{ @@ -507,6 +533,19 @@ {"shape":"NoSuchConfigRuleException"} ] }, + "PutRetentionConfiguration":{ + "name":"PutRetentionConfiguration", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"PutRetentionConfigurationRequest"}, + "output":{"shape":"PutRetentionConfigurationResponse"}, + "errors":[ + {"shape":"InvalidParameterValueException"}, + {"shape":"MaxNumberOfRetentionConfigurationsExceededException"} + ] + }, "StartConfigRulesEvaluation":{ "name":"StartConfigRulesEvaluation", "http":{ @@ -1081,6 +1120,13 @@ "RequesterAwsRegion":{"shape":"AwsRegion"} } }, + "DeleteRetentionConfigurationRequest":{ + "type":"structure", + "required":["RetentionConfigurationName"], + "members":{ + "RetentionConfigurationName":{"shape":"RetentionConfigurationName"} + } + }, "DeliverConfigSnapshotRequest":{ "type":"structure", "required":["deliveryChannelName"], @@ -1324,6 +1370,20 @@ "NextToken":{"shape":"String"} } }, + "DescribeRetentionConfigurationsRequest":{ + "type":"structure", + "members":{ + "RetentionConfigurationNames":{"shape":"RetentionConfigurationNameList"}, + "NextToken":{"shape":"NextToken"} + } + }, + "DescribeRetentionConfigurationsResponse":{ + "type":"structure", + "members":{ + "RetentionConfigurations":{"shape":"RetentionConfigurationList"}, + "NextToken":{"shape":"NextToken"} + } + }, "EarlierTime":{"type":"timestamp"}, "EmptiableStringWithCharLimit256":{ "type":"string", @@ -1665,6 +1725,12 @@ }, "exception":true }, + "MaxNumberOfRetentionConfigurationsExceededException":{ + "type":"structure", + "members":{ + }, + "exception":true + }, "MaximumExecutionFrequency":{ "type":"string", "enum":[ @@ -1740,6 +1806,12 @@ }, "exception":true }, + "NoSuchRetentionConfigurationException":{ + "type":"structure", + "members":{ + }, + "exception":true + }, "OrderingTimestamp":{"type":"timestamp"}, "OrganizationAccessDeniedException":{ "type":"structure", @@ -1848,6 +1920,19 @@ "FailedEvaluations":{"shape":"Evaluations"} } }, + "PutRetentionConfigurationRequest":{ + "type":"structure", + "required":["RetentionPeriodInDays"], + "members":{ + "RetentionPeriodInDays":{"shape":"RetentionPeriodInDays"} + } + }, + "PutRetentionConfigurationResponse":{ + "type":"structure", + "members":{ + "RetentionConfiguration":{"shape":"RetentionConfiguration"} + } + }, "RecorderName":{ "type":"string", "max":256, @@ -2028,6 +2113,38 @@ "max":20, "min":0 }, + "RetentionConfiguration":{ + "type":"structure", + "required":[ + "Name", + "RetentionPeriodInDays" + ], + "members":{ + "Name":{"shape":"RetentionConfigurationName"}, + "RetentionPeriodInDays":{"shape":"RetentionPeriodInDays"} + } + }, + "RetentionConfigurationList":{ + "type":"list", + "member":{"shape":"RetentionConfiguration"} + }, + "RetentionConfigurationName":{ + "type":"string", + "max":256, + "min":1, + "pattern":"[\\w\\-]+" + }, + "RetentionConfigurationNameList":{ + "type":"list", + "member":{"shape":"RetentionConfigurationName"}, + "max":1, + "min":0 + }, + "RetentionPeriodInDays":{ + "type":"integer", + "max":2557, + "min":30 + }, "RuleLimit":{ "type":"integer", "max":50, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/docs-2.json index c7eb92e02..265a307dc 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/config/2014-11-12/docs-2.json @@ -10,6 +10,7 @@ "DeleteDeliveryChannel": "

Deletes the delivery channel.

Before you can delete the delivery channel, you must stop the configuration recorder by using the StopConfigurationRecorder action.

", "DeleteEvaluationResults": "

Deletes the evaluation results for the specified AWS Config rule. You can specify one AWS Config rule per request. After you delete the evaluation results, you can call the StartConfigRulesEvaluation API to start evaluating your AWS resources against the rule.

", "DeletePendingAggregationRequest": "

Deletes pending authorization requests for a specified aggregator account in a specified region.

", + "DeleteRetentionConfiguration": "

Deletes the retention configuration.

", "DeliverConfigSnapshot": "

Schedules delivery of a configuration snapshot to the Amazon S3 bucket in the specified delivery channel. After the delivery has started, AWS Config sends the following notifications using an Amazon SNS topic that you have specified.

  • Notification of the start of the delivery.

  • Notification of the completion of the delivery, if the delivery was successfully completed.

  • Notification of delivery failure, if the delivery failed.

", "DescribeAggregateComplianceByConfigRules": "

Returns a list of compliant and noncompliant rules with the number of resources for compliant and noncompliant rules.

The results can return an empty result page, but if you have a nextToken, the results are displayed on the next page.

", "DescribeAggregationAuthorizations": "

Returns a list of authorizations granted to various aggregator accounts and regions.

", @@ -24,6 +25,7 @@ "DescribeDeliveryChannelStatus": "

Returns the current status of the specified delivery channel. If a delivery channel is not specified, this action returns the current status of all delivery channels associated with the account.

Currently, you can specify only one delivery channel per region in your account.

", "DescribeDeliveryChannels": "

Returns details about the specified delivery channel. If a delivery channel is not specified, this action returns the details of all delivery channels associated with the account.

Currently, you can specify only one delivery channel per region in your account.

", "DescribePendingAggregationRequests": "

Returns a list of all pending aggregation requests.

", + "DescribeRetentionConfigurations": "

Returns the details of one or more retention configurations. If the retention configuration name is not specified, this action returns the details for all the retention configurations for that account.

Currently, AWS Config supports only one retention configuration per region in your account.

", "GetAggregateComplianceDetailsByConfigRule": "

Returns the evaluation results for the specified AWS Config rule for a specific resource in a rule. The results indicate which AWS resources were evaluated by the rule, when each resource was last evaluated, and whether each resource complies with the rule.

The results can return an empty result page. But if you have a nextToken, the results are displayed on the next page.

", "GetAggregateConfigRuleComplianceSummary": "

Returns the number of compliant and noncompliant rules for one or more accounts and regions in an aggregator.

The results can return an empty result page, but if you have a nextToken, the results are displayed on the next page.

", "GetComplianceDetailsByConfigRule": "

Returns the evaluation results for the specified AWS Config rule. The results indicate which AWS resources were evaluated by the rule, when each resource was last evaluated, and whether each resource complies with the rule.

", @@ -31,7 +33,7 @@ "GetComplianceSummaryByConfigRule": "

Returns the number of AWS Config rules that are compliant and noncompliant, up to a maximum of 25 for each.

", "GetComplianceSummaryByResourceType": "

Returns the number of resources that are compliant and the number that are noncompliant. You can specify one or more resource types to get these numbers for each resource type. The maximum number returned is 100.

", "GetDiscoveredResourceCounts": "

Returns the resource types, the number of each resource type, and the total number of resources that AWS Config is recording in this region for your AWS account.

Example

  1. AWS Config is recording three resource types in the US East (Ohio) Region for your account: 25 EC2 instances, 20 IAM users, and 15 S3 buckets.

  2. You make a call to the GetDiscoveredResourceCounts action and specify that you want all resource types.

  3. AWS Config returns the following:

    • The resource types (EC2 instances, IAM users, and S3 buckets).

    • The number of each resource type (25, 20, and 15).

    • The total number of all resources (60).

The response is paginated. By default, AWS Config lists 100 ResourceCount objects on each page. You can customize this number with the limit parameter. The response includes a nextToken string. To get the next page of results, run the request again and specify the string for the nextToken parameter.

If you make a call to the GetDiscoveredResourceCounts action, you might not immediately receive resource counts in the following situations:

  • You are a new AWS Config customer.

  • You just enabled resource recording.

It might take a few minutes for AWS Config to record and count your resources. Wait a few minutes and then retry the GetDiscoveredResourceCounts action.

", - "GetResourceConfigHistory": "

Returns a list of configuration items for the specified resource. The list contains details about each state of the resource during the specified time interval.

The response is paginated. By default, AWS Config returns a limit of 10 configuration items per page. You can customize this number with the limit parameter. The response includes a nextToken string. To get the next page of results, run the request again and specify the string for the nextToken parameter.

Each call to the API is limited to span a duration of seven days. It is likely that the number of records returned is smaller than the specified limit. In such cases, you can make another call, using the nextToken.

", + "GetResourceConfigHistory": "

Returns a list of configuration items for the specified resource. The list contains details about each state of the resource during the specified time interval. If you specified a retention period to retain your ConfigurationItems between a minimum of 30 days and a maximum of 7 years (2557 days), AWS Config returns the ConfigurationItems for the specified retention period.

The response is paginated. By default, AWS Config returns a limit of 10 configuration items per page. You can customize this number with the limit parameter. The response includes a nextToken string. To get the next page of results, run the request again and specify the string for the nextToken parameter.

Each call to the API is limited to span a duration of seven days. It is likely that the number of records returned is smaller than the specified limit. In such cases, you can make another call, using the nextToken.

", "ListDiscoveredResources": "

Accepts a resource type and returns a list of resource identifiers for the resources of that type. A resource identifier includes the resource type, ID, and (if available) the custom resource name. The results consist of resources that AWS Config has discovered, including those that AWS Config is not currently recording. You can narrow the results to include only resources that have specific resource IDs or a resource name.

You can specify either resource IDs or a resource name, but not both, in the same request.

The response is paginated. By default, AWS Config lists 100 resource identifiers on each page. You can customize this number with the limit parameter. The response includes a nextToken string. To get the next page of results, run the request again and specify the string for the nextToken parameter.

", "PutAggregationAuthorization": "

Authorizes the aggregator account and region to collect data from the source account and region.

", "PutConfigRule": "

Adds or updates an AWS Config rule for evaluating whether your AWS resources comply with your desired configurations.

You can use this action for custom AWS Config rules and AWS managed Config rules. A custom AWS Config rule is a rule that you develop and maintain. An AWS managed Config rule is a customizable, predefined rule that AWS Config provides.

If you are adding a new custom AWS Config rule, you must first create the AWS Lambda function that the rule invokes to evaluate your resources. When you use the PutConfigRule action to add the rule to AWS Config, you must specify the Amazon Resource Name (ARN) that AWS Lambda assigns to the function. Specify the ARN for the SourceIdentifier key. This key is part of the Source object, which is part of the ConfigRule object.

If you are adding an AWS managed Config rule, specify the rule's identifier for the SourceIdentifier key. To reference AWS managed Config rule identifiers, see About AWS Managed Config Rules.

For any new rule that you add, specify the ConfigRuleName in the ConfigRule object. Do not specify the ConfigRuleArn or the ConfigRuleId. These values are generated by AWS Config for new rules.

If you are updating a rule that you added previously, you can specify the rule by ConfigRuleName, ConfigRuleId, or ConfigRuleArn in the ConfigRule data type that you use in this request.

The maximum number of rules that AWS Config supports is 50.

For information about requesting a rule limit increase, see AWS Config Limits in the AWS General Reference Guide.

For more information about developing and using AWS Config rules, see Evaluating AWS Resource Configurations with AWS Config in the AWS Config Developer Guide.

", @@ -39,6 +41,7 @@ "PutConfigurationRecorder": "

Creates a new configuration recorder to record the selected resource configurations.

You can use this action to change the role roleARN or the recordingGroup of an existing recorder. To change the role, call the action on the existing configuration recorder and specify a role.

Currently, you can specify only one configuration recorder per region in your account.

If ConfigurationRecorder does not have the recordingGroup parameter specified, the default is to record all supported resource types.

", "PutDeliveryChannel": "

Creates a delivery channel object to deliver configuration information to an Amazon S3 bucket and Amazon SNS topic.

Before you can create a delivery channel, you must create a configuration recorder.

You can use this action to change the Amazon S3 bucket or an Amazon SNS topic of the existing delivery channel. To change the Amazon S3 bucket or an Amazon SNS topic, call this action and specify the changed values for the S3 bucket and the SNS topic. If you specify a different value for either the S3 bucket or the SNS topic, this action will keep the existing value for the parameter that is not changed.

You can have only one delivery channel per region in your account.

", "PutEvaluations": "

Used by an AWS Lambda function to deliver evaluation results to AWS Config. This action is required in every AWS Lambda function that is invoked by an AWS Config rule.

", + "PutRetentionConfiguration": "

Creates and updates the retention configuration with details about retention period (number of days) that AWS Config stores your historical information. The API creates the RetentionConfiguration object and names the object as default. When you have a RetentionConfiguration object named default, calling the API modifies the default object.

Currently, AWS Config supports only one retention configuration per region in your account.

", "StartConfigRulesEvaluation": "

Runs an on-demand evaluation for the specified AWS Config rules against the last known configuration state of the resources. Use StartConfigRulesEvaluation when you want to test that a rule you updated is working as expected. StartConfigRulesEvaluation does not re-record the latest configuration state for your resources. It re-runs an evaluation against the last known state of your resources.

You can specify up to 25 AWS Config rules per request.

An existing StartConfigRulesEvaluation call for the specified rules must complete before you can call the API again. If you chose to have AWS Config stream to an Amazon SNS topic, you will receive a ConfigRuleEvaluationStarted notification when the evaluation starts.

You don't need to call the StartConfigRulesEvaluation API to run an evaluation for a new rule. When you create a rule, AWS Config evaluates your resources against the rule automatically.

The StartConfigRulesEvaluation API is useful if you want to run on-demand evaluations, such as the following example:

  1. You have a custom rule that evaluates your IAM resources every 24 hours.

  2. You update your Lambda function to add additional conditions to your rule.

  3. Instead of waiting for the next periodic evaluation, you call the StartConfigRulesEvaluation API.

  4. AWS Config invokes your Lambda function and evaluates your IAM resources.

  5. Your custom rule will still run periodic evaluations every 24 hours.

", "StartConfigurationRecorder": "

Starts recording configurations of the AWS resources you have selected to record in your AWS account.

You must have created at least one delivery channel to successfully start the configuration recorder.

", "StopConfigurationRecorder": "

Stops recording configurations of the AWS resources you have selected to record in your AWS account.

" @@ -133,7 +136,7 @@ "AggregatedSourceStatusList": { "base": null, "refs": { - "DescribeConfigurationAggregatorSourcesStatusResponse$AggregatedSourceStatusList": "

Retuns an AggregatedSourceStatus object.

" + "DescribeConfigurationAggregatorSourcesStatusResponse$AggregatedSourceStatusList": "

Returns an AggregatedSourceStatus object.

" } }, "AggregatedSourceStatusType": { @@ -242,12 +245,12 @@ "Boolean": { "base": null, "refs": { - "AccountAggregationSource$AllAwsRegions": "

If true, aggreagate existing AWS Config regions and future regions.

", + "AccountAggregationSource$AllAwsRegions": "

If true, aggregate existing AWS Config regions and future regions.

", "ComplianceContributorCount$CapExceeded": "

Indicates whether the maximum count is reached.

", "ConfigRuleEvaluationStatus$FirstEvaluationStarted": "

Indicates whether AWS Config has evaluated your resources against the rule at least once.

  • true - AWS Config has evaluated your AWS resources against the rule at least once.

  • false - AWS Config has not once finished evaluating your AWS resources against the rule.

", "ConfigurationRecorderStatus$recording": "

Specifies whether or not the recorder is currently recording.

", "ListDiscoveredResourcesRequest$includeDeletedResources": "

Specifies whether AWS Config includes deleted resources in the results. By default, deleted resources are not included.

", - "OrganizationAggregationSource$AllAwsRegions": "

If true, aggreagate existing AWS Config regions and future regions.

", + "OrganizationAggregationSource$AllAwsRegions": "

If true, aggregate existing AWS Config regions and future regions.

", "PutEvaluationsRequest$TestMode": "

Use this parameter to specify a test run for PutEvaluations. You can verify whether your AWS Lambda function will deliver evaluation results to AWS Config. No updates occur to your existing evaluations, and evaluation results are not sent to AWS Config.

When TestMode is true, PutEvaluations doesn't require a valid value for the ResultToken parameter, but the value cannot be null.

" } }, @@ -620,6 +623,11 @@ "refs": { } }, + "DeleteRetentionConfigurationRequest": { + "base": null, + "refs": { + } + }, "DeliverConfigSnapshotRequest": { "base": "

The input for the DeliverConfigSnapshot action.

", "refs": { @@ -805,6 +813,16 @@ "refs": { } }, + "DescribeRetentionConfigurationsRequest": { + "base": null, + "refs": { + } + }, + "DescribeRetentionConfigurationsResponse": { + "base": null, + "refs": { + } + }, "EarlierTime": { "base": null, "refs": { @@ -1083,6 +1101,11 @@ "refs": { } }, + "MaxNumberOfRetentionConfigurationsExceededException": { + "base": "

Failed to add the retention configuration because a retention configuration with that name already exists.

", + "refs": { + } + }, "MaximumExecutionFrequency": { "base": null, "refs": { @@ -1110,6 +1133,8 @@ "DescribeAggregateComplianceByConfigRulesResponse$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", "DescribeComplianceByResourceRequest$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", "DescribeComplianceByResourceResponse$NextToken": "

The string that you use in a subsequent request to get the next page of results in a paginated response.

", + "DescribeRetentionConfigurationsRequest$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", + "DescribeRetentionConfigurationsResponse$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", "GetAggregateComplianceDetailsByConfigRuleRequest$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", "GetAggregateComplianceDetailsByConfigRuleResponse$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", "GetAggregateConfigRuleComplianceSummaryRequest$NextToken": "

The nextToken string returned on a previous page that you use to get the next page of results in a paginated response.

", @@ -1169,6 +1194,11 @@ "refs": { } }, + "NoSuchRetentionConfigurationException": { + "base": "

You have specified a retention configuration that does not exist.

", + "refs": { + } + }, "OrderingTimestamp": { "base": null, "refs": { @@ -1255,6 +1285,16 @@ "refs": { } }, + "PutRetentionConfigurationRequest": { + "base": null, + "refs": { + } + }, + "PutRetentionConfigurationResponse": { + "base": null, + "refs": { + } + }, "RecorderName": { "base": null, "refs": { @@ -1428,6 +1468,40 @@ "GetDiscoveredResourceCountsRequest$resourceTypes": "

The comma-separated list that specifies the resource types that you want AWS Config to return (for example, \"AWS::EC2::Instance\", \"AWS::IAM::User\").

If a value for resourceTypes is not specified, AWS Config returns all resource types that AWS Config is recording in the region for your account.

If the configuration recorder is turned off, AWS Config returns an empty list of ResourceCount objects. If the configuration recorder is not recording a specific resource type (for example, S3 buckets), that resource type is not returned in the list of ResourceCount objects.

" } }, + "RetentionConfiguration": { + "base": "

An object with the name of the retention configuration and the retention period in days. The object stores the configuration for data retention in AWS Config.

", + "refs": { + "PutRetentionConfigurationResponse$RetentionConfiguration": "

Returns a retention configuration object.

", + "RetentionConfigurationList$member": null + } + }, + "RetentionConfigurationList": { + "base": null, + "refs": { + "DescribeRetentionConfigurationsResponse$RetentionConfigurations": "

Returns a retention configuration object.

" + } + }, + "RetentionConfigurationName": { + "base": null, + "refs": { + "DeleteRetentionConfigurationRequest$RetentionConfigurationName": "

The name of the retention configuration to delete.

", + "RetentionConfiguration$Name": "

The name of the retention configuration object.

", + "RetentionConfigurationNameList$member": null + } + }, + "RetentionConfigurationNameList": { + "base": null, + "refs": { + "DescribeRetentionConfigurationsRequest$RetentionConfigurationNames": "

A list of names of retention configurations for which you want details. If you do not specify a name, AWS Config returns details for all the retention configurations for that account.

Currently, AWS Config supports only one retention configuration per region in your account.

" + } + }, + "RetentionPeriodInDays": { + "base": null, + "refs": { + "PutRetentionConfigurationRequest$RetentionPeriodInDays": "

Number of days AWS Config stores your historical information.

Currently, only applicable to the configuration item history.

", + "RetentionConfiguration$RetentionPeriodInDays": "

Number of days AWS Config stores your historical information.

Currently, only applicable to the configuration item history.

" + } + }, "RuleLimit": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/api-2.json index b46f752c9..b2d750aee 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/api-2.json @@ -1159,7 +1159,8 @@ "CreateRemoteAccessSessionConfiguration":{ "type":"structure", "members":{ - "billingMethod":{"shape":"BillingMethod"} + "billingMethod":{"shape":"BillingMethod"}, + "vpceConfigurationArns":{"shape":"AmazonResourceNames"} } }, "CreateRemoteAccessSessionRequest":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/docs-2.json index a346fa94e..4447dba47 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/devicefarm/2015-06-23/docs-2.json @@ -175,6 +175,7 @@ "AmazonResourceNames": { "base": null, "refs": { + "CreateRemoteAccessSessionConfiguration$vpceConfigurationArns": "

An array of Amazon Resource Names (ARNs) included in the VPC endpoint configuration.

", "ScheduleRunConfiguration$vpceConfigurationArns": "

An array of Amazon Resource Names (ARNs) for your VPC endpoint configurations.

", "ScheduleRunConfiguration$auxiliaryApps": "

A list of auxiliary apps for the run.

" } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/api-2.json index 701d0f61c..7d1f06b33 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/api-2.json @@ -551,6 +551,24 @@ {"shape":"ServiceException"} ] }, + "ResetUserPassword":{ + "name":"ResetUserPassword", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ResetUserPasswordRequest"}, + "output":{"shape":"ResetUserPasswordResult"}, + "errors":[ + {"shape":"DirectoryUnavailableException"}, + {"shape":"UserDoesNotExistException"}, + {"shape":"InvalidPasswordException"}, + {"shape":"UnsupportedOperationException"}, + {"shape":"EntityDoesNotExistException"}, + {"shape":"ClientException"}, + {"shape":"ServiceException"} + ] + }, "RestoreFromSnapshot":{ "name":"RestoreFromSnapshot", "http":{ @@ -960,6 +978,12 @@ } }, "CreatedDateTime":{"type":"timestamp"}, + "CustomerUserName":{ + "type":"string", + "max":64, + "min":1, + "pattern":"^(?!.*\\\\|.*\"|.*\\/|.*\\[|.*\\]|.*:|.*;|.*\\||.*=|.*,|.*\\+|.*\\*|.*\\?|.*<|.*>|.*@).*$" + }, "DeleteAssociatedConditionalForwarder":{"type":"boolean"}, "DeleteConditionalForwarderRequest":{ "type":"structure", @@ -1481,6 +1505,14 @@ }, "exception":true }, + "InvalidPasswordException":{ + "type":"structure", + "members":{ + "Message":{"shape":"ExceptionMessage"}, + "RequestId":{"shape":"RequestId"} + }, + "exception":true + }, "IpAddr":{ "type":"string", "pattern":"^(?:(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.){3}(?:25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)$" @@ -1726,6 +1758,24 @@ "type":"string", "pattern":"^([A-Fa-f0-9]{8}-[A-Fa-f0-9]{4}-[A-Fa-f0-9]{4}-[A-Fa-f0-9]{4}-[A-Fa-f0-9]{12})$" }, + "ResetUserPasswordRequest":{ + "type":"structure", + "required":[ + "DirectoryId", + "UserName", + "NewPassword" + ], + "members":{ + "DirectoryId":{"shape":"DirectoryId"}, + "UserName":{"shape":"CustomerUserName"}, + "NewPassword":{"shape":"UserPassword"} + } + }, + "ResetUserPasswordResult":{ + "type":"structure", + "members":{ + } + }, "ResourceId":{ "type":"string", "pattern":"^[d]-[0-9a-f]{10}$" @@ -2077,11 +2127,25 @@ }, "UpdateSecurityGroupForDirectoryControllers":{"type":"boolean"}, "UseSameUsername":{"type":"boolean"}, + "UserDoesNotExistException":{ + "type":"structure", + "members":{ + "Message":{"shape":"ExceptionMessage"}, + "RequestId":{"shape":"RequestId"} + }, + "exception":true + }, "UserName":{ "type":"string", "min":1, "pattern":"[a-zA-Z0-9._-]+" }, + "UserPassword":{ + "type":"string", + "max":127, + "min":1, + "sensitive":true + }, "VerifyTrustRequest":{ "type":"structure", "required":["TrustId"], diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/docs-2.json index 0ed07bfd0..1c0a5c386 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ds/2015-04-16/docs-2.json @@ -36,6 +36,7 @@ "RegisterEventTopic": "

Associates a directory with an SNS topic. This establishes the directory as a publisher to the specified SNS topic. You can then receive email or text (SMS) messages when the status of your directory changes. You get notified if your directory goes from an Active status to an Impaired or Inoperable status. You also receive a notification when the directory returns to an Active status.

", "RemoveIpRoutes": "

Removes IP address blocks from a directory.

", "RemoveTagsFromResource": "

Removes tags from a directory.

", + "ResetUserPassword": "

Resets the password for any user in your AWS Managed Microsoft AD or Simple AD directory.

", "RestoreFromSnapshot": "

Restores a directory using an existing directory snapshot.

When you restore a directory from a snapshot, any changes made to the directory after the snapshot date are overwritten.

This action returns as soon as the restore operation is initiated. You can monitor the progress of the restore operation by calling the DescribeDirectories operation with the directory identifier. When the DirectoryDescription.Stage value changes to Active, the restore operation is complete.

", "StartSchemaExtension": "

Applies a schema extension to a Microsoft AD directory.

", "UpdateConditionalForwarder": "

Updates a conditional forwarder that has been set up for your AWS directory.

", @@ -302,6 +303,12 @@ "Trust$CreatedDateTime": "

The date and time that the trust relationship was created.

" } }, + "CustomerUserName": { + "base": null, + "refs": { + "ResetUserPasswordRequest$UserName": "

The username of the user whose password will be reset.

" + } + }, "DeleteAssociatedConditionalForwarder": { "base": null, "refs": { @@ -506,6 +513,7 @@ "ListSchemaExtensionsRequest$DirectoryId": "

The identifier of the directory from which to retrieve the schema extension information.

", "RegisterEventTopicRequest$DirectoryId": "

The Directory ID that will publish status messages to the SNS topic.

", "RemoveIpRoutesRequest$DirectoryId": "

Identifier (ID) of the directory from which you want to remove the IP addresses.

", + "ResetUserPasswordRequest$DirectoryId": "

Identifier of the AWS Managed Microsoft AD or Simple AD directory in which the user resides.

", "SchemaExtensionInfo$DirectoryId": "

The identifier of the directory to which the schema extension is applied.

", "Snapshot$DirectoryId": "

The directory identifier.

", "StartSchemaExtensionRequest$DirectoryId": "

The identifier of the directory for which the schema extension will be applied to.

", @@ -722,11 +730,13 @@ "InsufficientPermissionsException$Message": null, "InvalidNextTokenException$Message": null, "InvalidParameterException$Message": null, + "InvalidPasswordException$Message": null, "IpRouteLimitExceededException$Message": null, "ServiceException$Message": null, "SnapshotLimitExceededException$Message": null, "TagLimitExceededException$Message": null, - "UnsupportedOperationException$Message": null + "UnsupportedOperationException$Message": null, + "UserDoesNotExistException$Message": null } }, "GetDirectoryLimitsRequest": { @@ -764,6 +774,11 @@ "refs": { } }, + "InvalidPasswordException": { + "base": "

The new password provided by the user does not meet the password complexity requirements defined in your directory.

", + "refs": { + } + }, "IpAddr": { "base": null, "refs": { @@ -1045,11 +1060,23 @@ "InsufficientPermissionsException$RequestId": null, "InvalidNextTokenException$RequestId": null, "InvalidParameterException$RequestId": null, + "InvalidPasswordException$RequestId": null, "IpRouteLimitExceededException$RequestId": null, "ServiceException$RequestId": null, "SnapshotLimitExceededException$RequestId": null, "TagLimitExceededException$RequestId": null, - "UnsupportedOperationException$RequestId": null + "UnsupportedOperationException$RequestId": null, + "UserDoesNotExistException$RequestId": null + } + }, + "ResetUserPasswordRequest": { + "base": null, + "refs": { + } + }, + "ResetUserPasswordResult": { + "base": null, + "refs": { } }, "ResourceId": { @@ -1420,6 +1447,11 @@ "RadiusSettings$UseSameUsername": "

Not currently used.

" } }, + "UserDoesNotExistException": { + "base": "

The user provided a username that does not exist in your directory.

", + "refs": { + } + }, "UserName": { "base": null, "refs": { @@ -1429,6 +1461,12 @@ "EnableSsoRequest$UserName": "

The username of an alternate account to use to enable single-sign on. This is only used for AD Connector directories. This account must have privileges to add a service principal name.

If the AD Connector service account does not have privileges to add a service principal name, you can specify an alternate account with the UserName and Password parameters. These credentials are only used to enable single sign-on and are not stored by the service. The AD Connector service account is not changed.

" } }, + "UserPassword": { + "base": null, + "refs": { + "ResetUserPasswordRequest$NewPassword": "

The new password that will be reset.

" + } + }, "VerifyTrustRequest": { "base": "

Initiates the verification of an existing trust relationship between a Microsoft AD in the AWS cloud and an external domain.

", "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/api-2.json deleted file mode 100644 index e2f86cf4b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/api-2.json +++ /dev/null @@ -1,803 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2011-12-05", - "endpointPrefix":"dynamodb", - "jsonVersion":"1.0", - "protocol":"json", - "serviceAbbreviation":"DynamoDB", - "serviceFullName":"Amazon DynamoDB", - "serviceId":"DynamoDB", - "signatureVersion":"v4", - "targetPrefix":"DynamoDB_20111205", - "uid":"dynamodb-2011-12-05" - }, - "operations":{ - "BatchGetItem":{ - "name":"BatchGetItem", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"BatchGetItemInput"}, - "output":{"shape":"BatchGetItemOutput"}, - "errors":[ - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"InternalServerError"} - ] - }, - "BatchWriteItem":{ - "name":"BatchWriteItem", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"BatchWriteItemInput"}, - "output":{"shape":"BatchWriteItemOutput"}, - "errors":[ - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - }, - "CreateTable":{ - "name":"CreateTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateTableInput"}, - "output":{"shape":"CreateTableOutput"}, - "errors":[ - {"shape":"ResourceInUseException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - }, - "DeleteItem":{ - "name":"DeleteItem", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteItemInput"}, - "output":{"shape":"DeleteItemOutput"}, - "errors":[ - {"shape":"ConditionalCheckFailedException"}, - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - }, - "DeleteTable":{ - "name":"DeleteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteTableInput"}, - "output":{"shape":"DeleteTableOutput"}, - "errors":[ - {"shape":"ResourceInUseException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - }, - "DescribeTable":{ - "name":"DescribeTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeTableInput"}, - "output":{"shape":"DescribeTableOutput"}, - "errors":[ - {"shape":"ResourceNotFoundException"}, - {"shape":"InternalServerError"} - ] - }, - "GetItem":{ - "name":"GetItem", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetItemInput"}, - "output":{"shape":"GetItemOutput"}, - "errors":[ - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"InternalServerError"} - ] - }, - "ListTables":{ - "name":"ListTables", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListTablesInput"}, - "output":{"shape":"ListTablesOutput"}, - "errors":[ - {"shape":"InternalServerError"} - ] - }, - "PutItem":{ - "name":"PutItem", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PutItemInput"}, - "output":{"shape":"PutItemOutput"}, - "errors":[ - {"shape":"ConditionalCheckFailedException"}, - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - }, - "Query":{ - "name":"Query", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"QueryInput"}, - "output":{"shape":"QueryOutput"}, - "errors":[ - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"InternalServerError"} - ] - }, - "Scan":{ - "name":"Scan", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ScanInput"}, - "output":{"shape":"ScanOutput"}, - "errors":[ - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"InternalServerError"} - ] - }, - "UpdateItem":{ - "name":"UpdateItem", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UpdateItemInput"}, - "output":{"shape":"UpdateItemOutput"}, - "errors":[ - {"shape":"ConditionalCheckFailedException"}, - {"shape":"ProvisionedThroughputExceededException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - }, - "UpdateTable":{ - "name":"UpdateTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UpdateTableInput"}, - "output":{"shape":"UpdateTableOutput"}, - "errors":[ - {"shape":"ResourceInUseException"}, - {"shape":"ResourceNotFoundException"}, - {"shape":"LimitExceededException"}, - {"shape":"InternalServerError"} - ] - } - }, - "shapes":{ - "AttributeAction":{ - "type":"string", - "enum":[ - "ADD", - "PUT", - "DELETE" - ] - }, - "AttributeMap":{ - "type":"map", - "key":{"shape":"AttributeName"}, - "value":{"shape":"AttributeValue"} - }, - "AttributeName":{ - "type":"string", - "max":65535 - }, - "AttributeNameList":{ - "type":"list", - "member":{"shape":"AttributeName"}, - "min":1 - }, - "AttributeUpdates":{ - "type":"map", - "key":{"shape":"AttributeName"}, - "value":{"shape":"AttributeValueUpdate"} - }, - "AttributeValue":{ - "type":"structure", - "members":{ - "S":{"shape":"StringAttributeValue"}, - "N":{"shape":"NumberAttributeValue"}, - "B":{"shape":"BinaryAttributeValue"}, - "SS":{"shape":"StringSetAttributeValue"}, - "NS":{"shape":"NumberSetAttributeValue"}, - "BS":{"shape":"BinarySetAttributeValue"} - } - }, - "AttributeValueList":{ - "type":"list", - "member":{"shape":"AttributeValue"} - }, - "AttributeValueUpdate":{ - "type":"structure", - "members":{ - "Value":{"shape":"AttributeValue"}, - "Action":{"shape":"AttributeAction"} - } - }, - "BatchGetItemInput":{ - "type":"structure", - "required":["RequestItems"], - "members":{ - "RequestItems":{"shape":"BatchGetRequestMap"} - } - }, - "BatchGetItemOutput":{ - "type":"structure", - "members":{ - "Responses":{"shape":"BatchGetResponseMap"}, - "UnprocessedKeys":{"shape":"BatchGetRequestMap"} - } - }, - "BatchGetRequestMap":{ - "type":"map", - "key":{"shape":"TableName"}, - "value":{"shape":"KeysAndAttributes"}, - "max":100, - "min":1 - }, - "BatchGetResponseMap":{ - "type":"map", - "key":{"shape":"TableName"}, - "value":{"shape":"BatchResponse"} - }, - "BatchResponse":{ - "type":"structure", - "members":{ - "Items":{"shape":"ItemList"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "BatchWriteItemInput":{ - "type":"structure", - "required":["RequestItems"], - "members":{ - "RequestItems":{"shape":"BatchWriteItemRequestMap"} - } - }, - "BatchWriteItemOutput":{ - "type":"structure", - "members":{ - "Responses":{"shape":"BatchWriteResponseMap"}, - "UnprocessedItems":{"shape":"BatchWriteItemRequestMap"} - } - }, - "BatchWriteItemRequestMap":{ - "type":"map", - "key":{"shape":"TableName"}, - "value":{"shape":"WriteRequests"}, - "max":25, - "min":1 - }, - "BatchWriteResponse":{ - "type":"structure", - "members":{ - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "BatchWriteResponseMap":{ - "type":"map", - "key":{"shape":"TableName"}, - "value":{"shape":"BatchWriteResponse"} - }, - "BinaryAttributeValue":{"type":"blob"}, - "BinarySetAttributeValue":{ - "type":"list", - "member":{"shape":"BinaryAttributeValue"} - }, - "BooleanObject":{"type":"boolean"}, - "ComparisonOperator":{ - "type":"string", - "enum":[ - "EQ", - "NE", - "IN", - "LE", - "LT", - "GE", - "GT", - "BETWEEN", - "NOT_NULL", - "NULL", - "CONTAINS", - "NOT_CONTAINS", - "BEGINS_WITH" - ] - }, - "Condition":{ - "type":"structure", - "required":["ComparisonOperator"], - "members":{ - "AttributeValueList":{"shape":"AttributeValueList"}, - "ComparisonOperator":{"shape":"ComparisonOperator"} - } - }, - "ConditionalCheckFailedException":{ - "type":"structure", - "members":{ - "message":{"shape":"ErrorMessage"} - }, - "exception":true - }, - "ConsistentRead":{"type":"boolean"}, - "ConsumedCapacityUnits":{"type":"double"}, - "CreateTableInput":{ - "type":"structure", - "required":[ - "TableName", - "KeySchema", - "ProvisionedThroughput" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "KeySchema":{"shape":"KeySchema"}, - "ProvisionedThroughput":{"shape":"ProvisionedThroughput"} - } - }, - "CreateTableOutput":{ - "type":"structure", - "members":{ - "TableDescription":{"shape":"TableDescription"} - } - }, - "Date":{"type":"timestamp"}, - "DeleteItemInput":{ - "type":"structure", - "required":[ - "TableName", - "Key" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "Key":{"shape":"Key"}, - "Expected":{"shape":"ExpectedAttributeMap"}, - "ReturnValues":{"shape":"ReturnValue"} - } - }, - "DeleteItemOutput":{ - "type":"structure", - "members":{ - "Attributes":{"shape":"AttributeMap"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "DeleteRequest":{ - "type":"structure", - "required":["Key"], - "members":{ - "Key":{"shape":"Key"} - } - }, - "DeleteTableInput":{ - "type":"structure", - "required":["TableName"], - "members":{ - "TableName":{"shape":"TableName"} - } - }, - "DeleteTableOutput":{ - "type":"structure", - "members":{ - "TableDescription":{"shape":"TableDescription"} - } - }, - "DescribeTableInput":{ - "type":"structure", - "required":["TableName"], - "members":{ - "TableName":{"shape":"TableName"} - } - }, - "DescribeTableOutput":{ - "type":"structure", - "members":{ - "Table":{"shape":"TableDescription"} - } - }, - "ErrorMessage":{"type":"string"}, - "ExpectedAttributeMap":{ - "type":"map", - "key":{"shape":"AttributeName"}, - "value":{"shape":"ExpectedAttributeValue"} - }, - "ExpectedAttributeValue":{ - "type":"structure", - "members":{ - "Value":{"shape":"AttributeValue"}, - "Exists":{"shape":"BooleanObject"} - } - }, - "FilterConditionMap":{ - "type":"map", - "key":{"shape":"String"}, - "value":{"shape":"Condition"} - }, - "GetItemInput":{ - "type":"structure", - "required":[ - "TableName", - "Key" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "Key":{"shape":"Key"}, - "AttributesToGet":{"shape":"AttributeNameList"}, - "ConsistentRead":{"shape":"ConsistentRead"} - } - }, - "GetItemOutput":{ - "type":"structure", - "members":{ - "Item":{"shape":"AttributeMap"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "Integer":{"type":"integer"}, - "InternalServerError":{ - "type":"structure", - "members":{ - "message":{"shape":"ErrorMessage"} - }, - "exception":true, - "fault":true - }, - "ItemList":{ - "type":"list", - "member":{"shape":"AttributeMap"} - }, - "Key":{ - "type":"structure", - "required":["HashKeyElement"], - "members":{ - "HashKeyElement":{"shape":"AttributeValue"}, - "RangeKeyElement":{"shape":"AttributeValue"} - } - }, - "KeyList":{ - "type":"list", - "member":{"shape":"Key"}, - "max":100, - "min":1 - }, - "KeySchema":{ - "type":"structure", - "required":["HashKeyElement"], - "members":{ - "HashKeyElement":{"shape":"KeySchemaElement"}, - "RangeKeyElement":{"shape":"KeySchemaElement"} - } - }, - "KeySchemaAttributeName":{ - "type":"string", - "max":255, - "min":1 - }, - "KeySchemaElement":{ - "type":"structure", - "required":[ - "AttributeName", - "AttributeType" - ], - "members":{ - "AttributeName":{"shape":"KeySchemaAttributeName"}, - "AttributeType":{"shape":"ScalarAttributeType"} - } - }, - "KeysAndAttributes":{ - "type":"structure", - "required":["Keys"], - "members":{ - "Keys":{"shape":"KeyList"}, - "AttributesToGet":{"shape":"AttributeNameList"}, - "ConsistentRead":{"shape":"ConsistentRead"} - } - }, - "LimitExceededException":{ - "type":"structure", - "members":{ - "message":{"shape":"ErrorMessage"} - }, - "exception":true - }, - "ListTablesInput":{ - "type":"structure", - "members":{ - "ExclusiveStartTableName":{"shape":"TableName"}, - "Limit":{"shape":"ListTablesInputLimit"} - } - }, - "ListTablesInputLimit":{ - "type":"integer", - "max":100, - "min":1 - }, - "ListTablesOutput":{ - "type":"structure", - "members":{ - "TableNames":{"shape":"TableNameList"}, - "LastEvaluatedTableName":{"shape":"TableName"} - } - }, - "Long":{"type":"long"}, - "NumberAttributeValue":{"type":"string"}, - "NumberSetAttributeValue":{ - "type":"list", - "member":{"shape":"NumberAttributeValue"} - }, - "PositiveIntegerObject":{ - "type":"integer", - "min":1 - }, - "PositiveLongObject":{ - "type":"long", - "min":1 - }, - "ProvisionedThroughput":{ - "type":"structure", - "required":[ - "ReadCapacityUnits", - "WriteCapacityUnits" - ], - "members":{ - "ReadCapacityUnits":{"shape":"PositiveLongObject"}, - "WriteCapacityUnits":{"shape":"PositiveLongObject"} - } - }, - "ProvisionedThroughputDescription":{ - "type":"structure", - "members":{ - "LastIncreaseDateTime":{"shape":"Date"}, - "LastDecreaseDateTime":{"shape":"Date"}, - "NumberOfDecreasesToday":{"shape":"PositiveLongObject"}, - "ReadCapacityUnits":{"shape":"PositiveLongObject"}, - "WriteCapacityUnits":{"shape":"PositiveLongObject"} - } - }, - "ProvisionedThroughputExceededException":{ - "type":"structure", - "members":{ - "message":{"shape":"ErrorMessage"} - }, - "exception":true - }, - "PutItemInput":{ - "type":"structure", - "required":[ - "TableName", - "Item" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "Item":{"shape":"PutItemInputAttributeMap"}, - "Expected":{"shape":"ExpectedAttributeMap"}, - "ReturnValues":{"shape":"ReturnValue"} - } - }, - "PutItemInputAttributeMap":{ - "type":"map", - "key":{"shape":"AttributeName"}, - "value":{"shape":"AttributeValue"} - }, - "PutItemOutput":{ - "type":"structure", - "members":{ - "Attributes":{"shape":"AttributeMap"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "PutRequest":{ - "type":"structure", - "required":["Item"], - "members":{ - "Item":{"shape":"PutItemInputAttributeMap"} - } - }, - "QueryInput":{ - "type":"structure", - "required":[ - "TableName", - "HashKeyValue" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "AttributesToGet":{"shape":"AttributeNameList"}, - "Limit":{"shape":"PositiveIntegerObject"}, - "ConsistentRead":{"shape":"ConsistentRead"}, - "Count":{"shape":"BooleanObject"}, - "HashKeyValue":{"shape":"AttributeValue"}, - "RangeKeyCondition":{"shape":"Condition"}, - "ScanIndexForward":{"shape":"BooleanObject"}, - "ExclusiveStartKey":{"shape":"Key"} - } - }, - "QueryOutput":{ - "type":"structure", - "members":{ - "Items":{"shape":"ItemList"}, - "Count":{"shape":"Integer"}, - "LastEvaluatedKey":{"shape":"Key"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "ResourceInUseException":{ - "type":"structure", - "members":{ - "message":{"shape":"ErrorMessage"} - }, - "exception":true - }, - "ResourceNotFoundException":{ - "type":"structure", - "members":{ - "message":{"shape":"ErrorMessage"} - }, - "exception":true - }, - "ReturnValue":{ - "type":"string", - "enum":[ - "NONE", - "ALL_OLD", - "UPDATED_OLD", - "ALL_NEW", - "UPDATED_NEW" - ] - }, - "ScalarAttributeType":{ - "type":"string", - "enum":[ - "S", - "N", - "B" - ] - }, - "ScanInput":{ - "type":"structure", - "required":["TableName"], - "members":{ - "TableName":{"shape":"TableName"}, - "AttributesToGet":{"shape":"AttributeNameList"}, - "Limit":{"shape":"PositiveIntegerObject"}, - "Count":{"shape":"BooleanObject"}, - "ScanFilter":{"shape":"FilterConditionMap"}, - "ExclusiveStartKey":{"shape":"Key"} - } - }, - "ScanOutput":{ - "type":"structure", - "members":{ - "Items":{"shape":"ItemList"}, - "Count":{"shape":"Integer"}, - "ScannedCount":{"shape":"Integer"}, - "LastEvaluatedKey":{"shape":"Key"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "String":{"type":"string"}, - "StringAttributeValue":{"type":"string"}, - "StringSetAttributeValue":{ - "type":"list", - "member":{"shape":"StringAttributeValue"} - }, - "TableDescription":{ - "type":"structure", - "members":{ - "TableName":{"shape":"TableName"}, - "KeySchema":{"shape":"KeySchema"}, - "TableStatus":{"shape":"TableStatus"}, - "CreationDateTime":{"shape":"Date"}, - "ProvisionedThroughput":{"shape":"ProvisionedThroughputDescription"}, - "TableSizeBytes":{"shape":"Long"}, - "ItemCount":{"shape":"Long"} - } - }, - "TableName":{ - "type":"string", - "max":255, - "min":3, - "pattern":"[a-zA-Z0-9_.-]+" - }, - "TableNameList":{ - "type":"list", - "member":{"shape":"TableName"} - }, - "TableStatus":{ - "type":"string", - "enum":[ - "CREATING", - "UPDATING", - "DELETING", - "ACTIVE" - ] - }, - "UpdateItemInput":{ - "type":"structure", - "required":[ - "TableName", - "Key", - "AttributeUpdates" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "Key":{"shape":"Key"}, - "AttributeUpdates":{"shape":"AttributeUpdates"}, - "Expected":{"shape":"ExpectedAttributeMap"}, - "ReturnValues":{"shape":"ReturnValue"} - } - }, - "UpdateItemOutput":{ - "type":"structure", - "members":{ - "Attributes":{"shape":"AttributeMap"}, - "ConsumedCapacityUnits":{"shape":"ConsumedCapacityUnits"} - } - }, - "UpdateTableInput":{ - "type":"structure", - "required":[ - "TableName", - "ProvisionedThroughput" - ], - "members":{ - "TableName":{"shape":"TableName"}, - "ProvisionedThroughput":{"shape":"ProvisionedThroughput"} - } - }, - "UpdateTableOutput":{ - "type":"structure", - "members":{ - "TableDescription":{"shape":"TableDescription"} - } - }, - "WriteRequest":{ - "type":"structure", - "members":{ - "PutRequest":{"shape":"PutRequest"}, - "DeleteRequest":{"shape":"DeleteRequest"} - } - }, - "WriteRequests":{ - "type":"list", - "member":{"shape":"WriteRequest"}, - "max":25, - "min":1 - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/docs-2.json deleted file mode 100644 index 80242d9ee..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/docs-2.json +++ /dev/null @@ -1,606 +0,0 @@ -{ - "version": "2.0", - "service": "

Amazon DynamoDB is a fast, highly scalable, highly available, cost-effective non-relational database service.

Amazon DynamoDB removes traditional scalability limitations on data storage while maintaining low latency and predictable performance.

", - "operations": { - "BatchGetItem": "

Retrieves the attributes for multiple items from multiple tables using their primary keys.

The maximum number of item attributes that can be retrieved for a single operation is 100. Also, the number of items retrieved is constrained by a 1 MB the size limit. If the response size limit is exceeded or a partial result is returned due to an internal processing failure, Amazon DynamoDB returns an UnprocessedKeys value so you can retry the operation starting with the next item to get.

Amazon DynamoDB automatically adjusts the number of items returned per page to enforce this limit. For example, even if you ask to retrieve 100 items, but each individual item is 50k in size, the system returns 20 items and an appropriate UnprocessedKeys value so you can get the next page of results. If necessary, your application needs its own logic to assemble the pages of results into one set.

", - "BatchWriteItem": "

Allows to execute a batch of Put and/or Delete Requests for many tables in a single call. A total of 25 requests are allowed.

There are no transaction guarantees provided by this API. It does not allow conditional puts nor does it support return values.

", - "CreateTable": "

Adds a new table to your account.

The table name must be unique among those associated with the AWS Account issuing the request, and the AWS Region that receives the request (e.g. us-east-1).

The CreateTable operation triggers an asynchronous workflow to begin creating the table. Amazon DynamoDB immediately returns the state of the table (CREATING) until the table is in the ACTIVE state. Once the table is in the ACTIVE state, you can perform data plane operations.

", - "DeleteItem": "

Deletes a single item in a table by primary key.

You can perform a conditional delete operation that deletes the item if it exists, or if it has an expected attribute value.

", - "DeleteTable": "

Deletes a table and all of its items.

If the table is in the ACTIVE state, you can delete it. If a table is in CREATING or UPDATING states then Amazon DynamoDB returns a ResourceInUseException. If the specified table does not exist, Amazon DynamoDB returns a ResourceNotFoundException.

", - "DescribeTable": "

Retrieves information about the table, including the current status of the table, the primary key schema and when the table was created.

If the table does not exist, Amazon DynamoDB returns a ResourceNotFoundException.

", - "GetItem": "

Retrieves a set of Attributes for an item that matches the primary key.

The GetItem operation provides an eventually-consistent read by default. If eventually-consistent reads are not acceptable for your application, use ConsistentRead. Although this operation might take longer than a standard read, it always returns the last updated value.

", - "ListTables": "

Retrieves a paginated list of table names created by the AWS Account of the caller in the AWS Region (e.g. us-east-1).

", - "PutItem": "

Creates a new item, or replaces an old item with a new item (including all the attributes).

If an item already exists in the specified table with the same primary key, the new item completely replaces the existing item. You can perform a conditional put (insert a new item if one with the specified primary key doesn't exist), or replace an existing item if it has certain attribute values.

", - "Query": "

Gets the values of one or more items and its attributes by primary key (composite primary key, only).

Narrow the scope of the query using comparison operators on the RangeKeyValue of the composite key. Use the ScanIndexForward parameter to get results in forward or reverse order by range key.

", - "Scan": "

Retrieves one or more items and its attributes by performing a full scan of a table.

Provide a ScanFilter to get more specific results.

", - "UpdateItem": "

Edits an existing item's attributes.

You can perform a conditional update (insert a new attribute name-value pair if it doesn't exist, or replace an existing name-value pair if it has certain expected attribute values).

", - "UpdateTable": "

Updates the provisioned throughput for the given table.

Setting the throughput for a table helps you manage performance and is part of the Provisioned Throughput feature of Amazon DynamoDB.

" - }, - "shapes": { - "AttributeAction": { - "base": "

The type of action for an item update operation. Only use the add action for numbers or sets; the specified value is added to the existing value. If a set of values is specified, the values are added to the existing set. Adds the specified attribute. If the attribute exists, it is replaced by the new value. If no value is specified, this removes the attribute and its value. If a set of values is specified, then the values in the specified set are removed from the old set.

", - "refs": { - "AttributeValueUpdate$Action": null - } - }, - "AttributeMap": { - "base": null, - "refs": { - "DeleteItemOutput$Attributes": "

If the ReturnValues parameter is provided as ALL_OLD in the request, Amazon DynamoDB returns an array of attribute name-value pairs (essentially, the deleted item). Otherwise, the response contains an empty set.

", - "GetItemOutput$Item": "

Contains the requested attributes.

", - "ItemList$member": null, - "PutItemOutput$Attributes": "

Attribute values before the put operation, but only if the ReturnValues parameter is specified as ALL_OLD in the request.

", - "UpdateItemOutput$Attributes": "

A map of attribute name-value pairs, but only if the ReturnValues parameter is specified as something other than NONE in the request.

" - } - }, - "AttributeName": { - "base": null, - "refs": { - "AttributeMap$key": null, - "AttributeNameList$member": null, - "AttributeUpdates$key": null, - "ExpectedAttributeMap$key": null, - "PutItemInputAttributeMap$key": null - } - }, - "AttributeNameList": { - "base": "

List of Attribute names. If attribute names are not specified then all attributes will be returned. If some attributes are not found, they will not appear in the result.

", - "refs": { - "GetItemInput$AttributesToGet": null, - "KeysAndAttributes$AttributesToGet": null, - "QueryInput$AttributesToGet": null, - "ScanInput$AttributesToGet": null - } - }, - "AttributeUpdates": { - "base": "

Map of attribute name to the new value and action for the update. The attribute names specify the attributes to modify, and cannot contain any primary key attributes.

", - "refs": { - "UpdateItemInput$AttributeUpdates": null - } - }, - "AttributeValue": { - "base": "

AttributeValue can be String, Number, Binary, StringSet, NumberSet, BinarySet.

", - "refs": { - "AttributeMap$value": null, - "AttributeValueList$member": null, - "AttributeValueUpdate$Value": null, - "ExpectedAttributeValue$Value": "

Specify whether or not a value already exists and has a specific content for the attribute name-value pair.

", - "Key$HashKeyElement": "

A hash key element is treated as the primary key, and can be a string or a number. Single attribute primary keys have one index value. The value can be String, Number, StringSet, NumberSet.

", - "Key$RangeKeyElement": "

A range key element is treated as a secondary key (used in conjunction with the primary key), and can be a string or a number, and is only used for hash-and-range primary keys. The value can be String, Number, StringSet, NumberSet.

", - "PutItemInputAttributeMap$value": null, - "QueryInput$HashKeyValue": "

Attribute value of the hash component of the composite primary key.

" - } - }, - "AttributeValueList": { - "base": "

A list of attribute values to be used with a comparison operator for a scan or query operation. For comparisons that require more than one value, such as a BETWEEN comparison, the AttributeValueList contains two attribute values and the comparison operator.

", - "refs": { - "Condition$AttributeValueList": null - } - }, - "AttributeValueUpdate": { - "base": "

Specifies the attribute to update and how to perform the update. Possible values: PUT (default), ADD or DELETE.

", - "refs": { - "AttributeUpdates$value": null - } - }, - "BatchGetItemInput": { - "base": null, - "refs": { - } - }, - "BatchGetItemOutput": { - "base": null, - "refs": { - } - }, - "BatchGetRequestMap": { - "base": "

A map of the table name and corresponding items to get by primary key. While requesting items, each table name can be invoked only once per operation.

", - "refs": { - "BatchGetItemInput$RequestItems": null, - "BatchGetItemOutput$UnprocessedKeys": "

Contains a map of tables and their respective keys that were not processed with the current response, possibly due to reaching a limit on the response size. The UnprocessedKeys value is in the same form as a RequestItems parameter (so the value can be provided directly to a subsequent BatchGetItem operation). For more information, see the above RequestItems parameter.

" - } - }, - "BatchGetResponseMap": { - "base": "

Table names and the respective item attributes from the tables.

", - "refs": { - "BatchGetItemOutput$Responses": null - } - }, - "BatchResponse": { - "base": "

The item attributes from a response in a specific table, along with the read resources consumed on the table during the request.

", - "refs": { - "BatchGetResponseMap$value": null - } - }, - "BatchWriteItemInput": { - "base": null, - "refs": { - } - }, - "BatchWriteItemOutput": { - "base": "

A container for BatchWriteItem response

", - "refs": { - } - }, - "BatchWriteItemRequestMap": { - "base": "

A map of table name to list-of-write-requests.

Key: The table name corresponding to the list of requests

Value: Essentially a list of request items. Each request item could contain either a PutRequest or DeleteRequest. Never both.

", - "refs": { - "BatchWriteItemInput$RequestItems": "

A map of table name to list-of-write-requests. Used as input to the BatchWriteItem API call

", - "BatchWriteItemOutput$UnprocessedItems": "

The Items which we could not successfully process in a BatchWriteItem call is returned as UnprocessedItems

" - } - }, - "BatchWriteResponse": { - "base": null, - "refs": { - "BatchWriteResponseMap$value": null - } - }, - "BatchWriteResponseMap": { - "base": null, - "refs": { - "BatchWriteItemOutput$Responses": "

The response object as a result of BatchWriteItem call. This is essentially a map of table name to ConsumedCapacityUnits.

" - } - }, - "BinaryAttributeValue": { - "base": null, - "refs": { - "AttributeValue$B": "

Binary attributes are sequences of unsigned bytes.

", - "BinarySetAttributeValue$member": null - } - }, - "BinarySetAttributeValue": { - "base": null, - "refs": { - "AttributeValue$BS": "

A set of binary attributes.

" - } - }, - "BooleanObject": { - "base": null, - "refs": { - "ExpectedAttributeValue$Exists": "

Specify whether or not a value already exists for the attribute name-value pair.

", - "QueryInput$Count": "

If set to true, Amazon DynamoDB returns a total number of items that match the query parameters, instead of a list of the matching items and their attributes. Do not set Count to true while providing a list of AttributesToGet, otherwise Amazon DynamoDB returns a validation error.

", - "QueryInput$ScanIndexForward": "

Specifies forward or backward traversal of the index. Amazon DynamoDB returns results reflecting the requested order, determined by the range key. The default value is true (forward).

", - "ScanInput$Count": "

If set to true, Amazon DynamoDB returns a total number of items for the Scan operation, even if the operation has no matching items for the assigned filter. Do not set Count to true while providing a list of AttributesToGet, otherwise Amazon DynamoDB returns a validation error.

" - } - }, - "ComparisonOperator": { - "base": "

A comparison operator is an enumeration of several operations:

  • EQ for equal.
  • NE for not equal.
  • IN checks for exact matches.
  • LE for less than or equal to.
  • LT for less than.
  • GE for greater than or equal to.
  • GT for greater than.
  • BETWEEN for between.
  • NOT_NULL for exists.
  • NULL for not exists.
  • CONTAINS for substring or value in a set.
  • NOT_CONTAINS for absence of a substring or absence of a value in a set.
  • BEGINS_WITH for a substring prefix.

Scan operations support all available comparison operators.

Query operations support a subset of the available comparison operators: EQ, LE, LT, GE, GT, BETWEEN, and BEGINS_WITH.

", - "refs": { - "Condition$ComparisonOperator": null - } - }, - "Condition": { - "base": null, - "refs": { - "FilterConditionMap$value": null, - "QueryInput$RangeKeyCondition": "

A container for the attribute values and comparison operators to use for the query.

" - } - }, - "ConditionalCheckFailedException": { - "base": "

This exception is thrown when an expected value does not match what was found in the system.

", - "refs": { - } - }, - "ConsistentRead": { - "base": "

If set to true, then a consistent read is issued. Otherwise eventually-consistent is used.

", - "refs": { - "GetItemInput$ConsistentRead": null, - "KeysAndAttributes$ConsistentRead": null, - "QueryInput$ConsistentRead": null - } - }, - "ConsumedCapacityUnits": { - "base": "

The number of Capacity Units of the provisioned throughput of the table consumed during the operation. GetItem, BatchGetItem, BatchWriteItem, Query, and Scan operations consume ReadCapacityUnits, while PutItem, UpdateItem, and DeleteItem operations consume WriteCapacityUnits.

", - "refs": { - "BatchResponse$ConsumedCapacityUnits": null, - "BatchWriteResponse$ConsumedCapacityUnits": null, - "DeleteItemOutput$ConsumedCapacityUnits": null, - "GetItemOutput$ConsumedCapacityUnits": null, - "PutItemOutput$ConsumedCapacityUnits": null, - "QueryOutput$ConsumedCapacityUnits": null, - "ScanOutput$ConsumedCapacityUnits": null, - "UpdateItemOutput$ConsumedCapacityUnits": null - } - }, - "CreateTableInput": { - "base": null, - "refs": { - } - }, - "CreateTableOutput": { - "base": null, - "refs": { - } - }, - "Date": { - "base": null, - "refs": { - "ProvisionedThroughputDescription$LastIncreaseDateTime": null, - "ProvisionedThroughputDescription$LastDecreaseDateTime": null, - "TableDescription$CreationDateTime": null - } - }, - "DeleteItemInput": { - "base": null, - "refs": { - } - }, - "DeleteItemOutput": { - "base": null, - "refs": { - } - }, - "DeleteRequest": { - "base": "

A container for a Delete BatchWrite request

", - "refs": { - "WriteRequest$DeleteRequest": null - } - }, - "DeleteTableInput": { - "base": null, - "refs": { - } - }, - "DeleteTableOutput": { - "base": null, - "refs": { - } - }, - "DescribeTableInput": { - "base": null, - "refs": { - } - }, - "DescribeTableOutput": { - "base": null, - "refs": { - } - }, - "ErrorMessage": { - "base": null, - "refs": { - "ConditionalCheckFailedException$message": null, - "InternalServerError$message": null, - "LimitExceededException$message": null, - "ProvisionedThroughputExceededException$message": null, - "ResourceInUseException$message": null, - "ResourceNotFoundException$message": null - } - }, - "ExpectedAttributeMap": { - "base": "

Designates an attribute for a conditional modification. The Expected parameter allows you to provide an attribute name, and whether or not Amazon DynamoDB should check to see if the attribute has a particular value before modifying it.

", - "refs": { - "DeleteItemInput$Expected": null, - "PutItemInput$Expected": null, - "UpdateItemInput$Expected": null - } - }, - "ExpectedAttributeValue": { - "base": "

Allows you to provide an attribute name, and whether or not Amazon DynamoDB should check to see if the attribute value already exists; or if the attribute value exists and has a particular value before changing it.

", - "refs": { - "ExpectedAttributeMap$value": null - } - }, - "FilterConditionMap": { - "base": null, - "refs": { - "ScanInput$ScanFilter": "

Evaluates the scan results and returns only the desired values.

" - } - }, - "GetItemInput": { - "base": null, - "refs": { - } - }, - "GetItemOutput": { - "base": null, - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "QueryOutput$Count": "

Number of items in the response.

", - "ScanOutput$Count": "

Number of items in the response.

", - "ScanOutput$ScannedCount": "

Number of items in the complete scan before any filters are applied. A high ScannedCount value with few, or no, Count results indicates an inefficient Scan operation.

" - } - }, - "InternalServerError": { - "base": "

This exception is thrown when the service has a problem when trying to process the request.

", - "refs": { - } - }, - "ItemList": { - "base": null, - "refs": { - "BatchResponse$Items": null, - "QueryOutput$Items": null, - "ScanOutput$Items": null - } - }, - "Key": { - "base": "

The primary key that uniquely identifies each item in a table. A primary key can be a one attribute (hash) primary key or a two attribute (hash-and-range) primary key.

", - "refs": { - "DeleteItemInput$Key": null, - "DeleteRequest$Key": "

The item's key to be delete

", - "GetItemInput$Key": null, - "KeyList$member": null, - "QueryInput$ExclusiveStartKey": "

Primary key of the item from which to continue an earlier query. An earlier query might provide this value as the LastEvaluatedKey if that query operation was interrupted before completing the query; either because of the result set size or the Limit parameter. The LastEvaluatedKey can be passed back in a new query request to continue the operation from that point.

", - "QueryOutput$LastEvaluatedKey": "

Primary key of the item where the query operation stopped, inclusive of the previous result set. Use this value to start a new operation excluding this value in the new request. The LastEvaluatedKey is null when the entire query result set is complete (i.e. the operation processed the \"last page\").

", - "ScanInput$ExclusiveStartKey": "

Primary key of the item from which to continue an earlier scan. An earlier scan might provide this value if that scan operation was interrupted before scanning the entire table; either because of the result set size or the Limit parameter. The LastEvaluatedKey can be passed back in a new scan request to continue the operation from that point.

", - "ScanOutput$LastEvaluatedKey": "

Primary key of the item where the scan operation stopped. Provide this value in a subsequent scan operation to continue the operation from that point. The LastEvaluatedKey is null when the entire scan result set is complete (i.e. the operation processed the \"last page\").

", - "UpdateItemInput$Key": null - } - }, - "KeyList": { - "base": null, - "refs": { - "KeysAndAttributes$Keys": null - } - }, - "KeySchema": { - "base": "

The KeySchema identifies the primary key as a one attribute primary key (hash) or a composite two attribute (hash-and-range) primary key. Single attribute primary keys have one index value: a HashKeyElement. A composite hash-and-range primary key contains two attribute values: a HashKeyElement and a RangeKeyElement.

", - "refs": { - "CreateTableInput$KeySchema": null, - "TableDescription$KeySchema": null - } - }, - "KeySchemaAttributeName": { - "base": null, - "refs": { - "KeySchemaElement$AttributeName": "

The AttributeName of the KeySchemaElement.

" - } - }, - "KeySchemaElement": { - "base": "

KeySchemaElement is the primary key (hash or hash-and-range) structure for the table.

", - "refs": { - "KeySchema$HashKeyElement": "

A hash key element is treated as the primary key, and can be a string or a number. Single attribute primary keys have one index value. The value can be String, Number, StringSet, NumberSet.

", - "KeySchema$RangeKeyElement": "

A range key element is treated as a secondary key (used in conjunction with the primary key), and can be a string or a number, and is only used for hash-and-range primary keys. The value can be String, Number, StringSet, NumberSet.

" - } - }, - "KeysAndAttributes": { - "base": null, - "refs": { - "BatchGetRequestMap$value": null - } - }, - "LimitExceededException": { - "base": "

This exception is thrown when the subscriber exceeded the limits on the number of objects or operations.

", - "refs": { - } - }, - "ListTablesInput": { - "base": null, - "refs": { - } - }, - "ListTablesInputLimit": { - "base": "

A number of maximum table names to return.

", - "refs": { - "ListTablesInput$Limit": null - } - }, - "ListTablesOutput": { - "base": null, - "refs": { - } - }, - "Long": { - "base": null, - "refs": { - "TableDescription$TableSizeBytes": null, - "TableDescription$ItemCount": null - } - }, - "NumberAttributeValue": { - "base": null, - "refs": { - "AttributeValue$N": "

Numbers are positive or negative exact-value decimals and integers. A number can have up to 38 digits precision and can be between 10^-128 to 10^+126.

", - "NumberSetAttributeValue$member": null - } - }, - "NumberSetAttributeValue": { - "base": null, - "refs": { - "AttributeValue$NS": "

A set of numbers.

" - } - }, - "PositiveIntegerObject": { - "base": null, - "refs": { - "QueryInput$Limit": "

The maximum number of items to return. If Amazon DynamoDB hits this limit while querying the table, it stops the query and returns the matching values up to the limit, and a LastEvaluatedKey to apply in a subsequent operation to continue the query. Also, if the result set size exceeds 1MB before Amazon DynamoDB hits this limit, it stops the query and returns the matching values, and a LastEvaluatedKey to apply in a subsequent operation to continue the query.

", - "ScanInput$Limit": "

The maximum number of items to return. If Amazon DynamoDB hits this limit while scanning the table, it stops the scan and returns the matching values up to the limit, and a LastEvaluatedKey to apply in a subsequent operation to continue the scan. Also, if the scanned data set size exceeds 1 MB before Amazon DynamoDB hits this limit, it stops the scan and returns the matching values up to the limit, and a LastEvaluatedKey to apply in a subsequent operation to continue the scan.

" - } - }, - "PositiveLongObject": { - "base": null, - "refs": { - "ProvisionedThroughput$ReadCapacityUnits": "

ReadCapacityUnits are in terms of strictly consistent reads, assuming items of 1k. 2k items require twice the ReadCapacityUnits. Eventually-consistent reads only require half the ReadCapacityUnits of stirctly consistent reads.

", - "ProvisionedThroughput$WriteCapacityUnits": "

WriteCapacityUnits are in terms of strictly consistent reads, assuming items of 1k. 2k items require twice the WriteCapacityUnits.

", - "ProvisionedThroughputDescription$NumberOfDecreasesToday": null, - "ProvisionedThroughputDescription$ReadCapacityUnits": null, - "ProvisionedThroughputDescription$WriteCapacityUnits": null - } - }, - "ProvisionedThroughput": { - "base": "

Provisioned throughput reserves the required read and write resources for your table in terms of ReadCapacityUnits and WriteCapacityUnits. Values for provisioned throughput depend upon your expected read/write rates, item size, and consistency. Provide the expected number of read and write operations, assuming an item size of 1k and strictly consistent reads. For 2k item size, double the value. For 3k, triple the value, etc. Eventually-consistent reads consume half the resources of strictly consistent reads.

", - "refs": { - "CreateTableInput$ProvisionedThroughput": null, - "UpdateTableInput$ProvisionedThroughput": null - } - }, - "ProvisionedThroughputDescription": { - "base": null, - "refs": { - "TableDescription$ProvisionedThroughput": null - } - }, - "ProvisionedThroughputExceededException": { - "base": "

This exception is thrown when the level of provisioned throughput defined for the table is exceeded.

", - "refs": { - } - }, - "PutItemInput": { - "base": null, - "refs": { - } - }, - "PutItemInputAttributeMap": { - "base": "

A map of the attributes for the item, and must include the primary key values that define the item. Other attribute name-value pairs can be provided for the item.

", - "refs": { - "PutItemInput$Item": null, - "PutRequest$Item": "

The item to put

" - } - }, - "PutItemOutput": { - "base": null, - "refs": { - } - }, - "PutRequest": { - "base": "

A container for a Put BatchWrite request

", - "refs": { - "WriteRequest$PutRequest": null - } - }, - "QueryInput": { - "base": null, - "refs": { - } - }, - "QueryOutput": { - "base": null, - "refs": { - } - }, - "ResourceInUseException": { - "base": "

This exception is thrown when the resource which is being attempted to be changed is in use.

", - "refs": { - } - }, - "ResourceNotFoundException": { - "base": "

This exception is thrown when the resource which is being attempted to be changed is in use.

", - "refs": { - } - }, - "ReturnValue": { - "base": "

Use this parameter if you want to get the attribute name-value pairs before or after they are modified. For PUT operations, the possible parameter values are NONE (default) or ALL_OLD. For update operations, the possible parameter values are NONE (default) or ALL_OLD, UPDATED_OLD, ALL_NEW or UPDATED_NEW.

  • NONE: Nothing is returned.
  • ALL_OLD: Returns the attributes of the item as they were before the operation.
  • UPDATED_OLD: Returns the values of the updated attributes, only, as they were before the operation.
  • ALL_NEW: Returns all the attributes and their new values after the operation.
  • UPDATED_NEW: Returns the values of the updated attributes, only, as they are after the operation.
", - "refs": { - "DeleteItemInput$ReturnValues": null, - "PutItemInput$ReturnValues": null, - "UpdateItemInput$ReturnValues": null - } - }, - "ScalarAttributeType": { - "base": null, - "refs": { - "KeySchemaElement$AttributeType": "

The AttributeType of the KeySchemaElement which can be a String or a Number.

" - } - }, - "ScanInput": { - "base": null, - "refs": { - } - }, - "ScanOutput": { - "base": null, - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "FilterConditionMap$key": null - } - }, - "StringAttributeValue": { - "base": null, - "refs": { - "AttributeValue$S": "

Strings are Unicode with UTF-8 binary encoding. The maximum size is limited by the size of the primary key (1024 bytes as a range part of a key or 2048 bytes as a single part hash key) or the item size (64k).

", - "StringSetAttributeValue$member": null - } - }, - "StringSetAttributeValue": { - "base": null, - "refs": { - "AttributeValue$SS": "

A set of strings.

" - } - }, - "TableDescription": { - "base": null, - "refs": { - "CreateTableOutput$TableDescription": null, - "DeleteTableOutput$TableDescription": null, - "DescribeTableOutput$Table": null, - "UpdateTableOutput$TableDescription": null - } - }, - "TableName": { - "base": null, - "refs": { - "BatchGetRequestMap$key": null, - "BatchGetResponseMap$key": null, - "BatchWriteItemRequestMap$key": null, - "BatchWriteResponseMap$key": null, - "CreateTableInput$TableName": "

The name of the table you want to create. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "DeleteItemInput$TableName": "

The name of the table in which you want to delete an item. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "DeleteTableInput$TableName": "

The name of the table you want to delete. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "DescribeTableInput$TableName": "

The name of the table you want to describe. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "GetItemInput$TableName": "

The name of the table in which you want to get an item. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "ListTablesInput$ExclusiveStartTableName": "

The name of the table that starts the list. If you already ran a ListTables operation and received a LastEvaluatedTableName value in the response, use that value here to continue the list.

", - "ListTablesOutput$LastEvaluatedTableName": "

The name of the last table in the current list. Use this value as the ExclusiveStartTableName in a new request to continue the list until all the table names are returned. If this value is null, all table names have been returned.

", - "PutItemInput$TableName": "

The name of the table in which you want to put an item. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "QueryInput$TableName": "

The name of the table in which you want to query. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "ScanInput$TableName": "

The name of the table in which you want to scan. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "TableDescription$TableName": "

The name of the table being described.

", - "TableNameList$member": null, - "UpdateItemInput$TableName": "

The name of the table in which you want to update an item. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

", - "UpdateTableInput$TableName": "

The name of the table you want to update. Allowed characters are a-z, A-Z, 0-9, _ (underscore), - (hyphen) and . (period).

" - } - }, - "TableNameList": { - "base": null, - "refs": { - "ListTablesOutput$TableNames": null - } - }, - "TableStatus": { - "base": null, - "refs": { - "TableDescription$TableStatus": null - } - }, - "UpdateItemInput": { - "base": null, - "refs": { - } - }, - "UpdateItemOutput": { - "base": null, - "refs": { - } - }, - "UpdateTableInput": { - "base": null, - "refs": { - } - }, - "UpdateTableOutput": { - "base": null, - "refs": { - } - }, - "WriteRequest": { - "base": "

This structure is a Union of PutRequest and DeleteRequest. It can contain exactly one of PutRequest or DeleteRequest. Never Both. This is enforced in the code.

", - "refs": { - "WriteRequests$member": null - } - }, - "WriteRequests": { - "base": null, - "refs": { - "BatchWriteItemRequestMap$value": null - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/paginators-1.json deleted file mode 100644 index 3037d662a..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/paginators-1.json +++ /dev/null @@ -1,26 +0,0 @@ -{ - "pagination": { - "BatchGetItem": { - "input_token": "RequestItems", - "output_token": "UnprocessedKeys" - }, - "ListTables": { - "input_token": "ExclusiveStartTableName", - "limit_key": "Limit", - "output_token": "LastEvaluatedTableName", - "result_key": "TableNames" - }, - "Query": { - "input_token": "ExclusiveStartKey", - "limit_key": "Limit", - "output_token": "LastEvaluatedKey", - "result_key": "Items" - }, - "Scan": { - "input_token": "ExclusiveStartKey", - "limit_key": "Limit", - "output_token": "LastEvaluatedKey", - "result_key": "Items" - } - } -} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/waiters-2.json deleted file mode 100644 index 43a55ca7b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2011-12-05/waiters-2.json +++ /dev/null @@ -1,35 +0,0 @@ -{ - "version": 2, - "waiters": { - "TableExists": { - "delay": 20, - "operation": "DescribeTable", - "maxAttempts": 25, - "acceptors": [ - { - "expected": "ACTIVE", - "matcher": "path", - "state": "success", - "argument": "Table.TableStatus" - }, - { - "expected": "ResourceNotFoundException", - "matcher": "error", - "state": "retry" - } - ] - }, - "TableNotExists": { - "delay": 20, - "operation": "DescribeTable", - "maxAttempts": 25, - "acceptors": [ - { - "expected": "ResourceNotFoundException", - "matcher": "error", - "state": "success" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/api-2.json index 23de507f3..762389fd5 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/api-2.json @@ -1290,6 +1290,7 @@ "type":"list", "member":{"shape":"AttributeMap"} }, + "KMSMasterKeyArn":{"type":"string"}, "Key":{ "type":"map", "key":{"shape":"AttributeName"}, @@ -1855,7 +1856,9 @@ "SSEDescription":{ "type":"structure", "members":{ - "Status":{"shape":"SSEStatus"} + "Status":{"shape":"SSEStatus"}, + "SSEType":{"shape":"SSEType"}, + "KMSMasterKeyArn":{"shape":"KMSMasterKeyArn"} } }, "SSEEnabled":{"type":"boolean"}, @@ -1875,6 +1878,13 @@ "DISABLED" ] }, + "SSEType":{ + "type":"string", + "enum":[ + "AES256", + "KMS" + ] + }, "ScalarAttributeType":{ "type":"string", "enum":[ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/docs-2.json index 131c337ef..fea99c735 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/dynamodb/2012-08-10/docs-2.json @@ -5,7 +5,7 @@ "BatchGetItem": "

The BatchGetItem operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key.

A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. BatchGetItem will return a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for UnprocessedKeys. You can use this value to retry the operation starting with the next item to get.

If you request more than 100 items BatchGetItem will return a ValidationException with the message \"Too many items requested for the BatchGetItem call\".

For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate UnprocessedKeys value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one data set.

If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchGetItem will return a ProvisionedThroughputExceededException. If at least one of the items is successfully processed, then BatchGetItem completes successfully, while returning the keys of the unread items in UnprocessedKeys.

If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.

For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide.

By default, BatchGetItem performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set ConsistentRead to true for any or all tables.

In order to minimize response latency, BatchGetItem retrieves items in parallel.

When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the ProjectionExpression parameter.

If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see Capacity Units Calculations in the Amazon DynamoDB Developer Guide.

", "BatchWriteItem": "

The BatchWriteItem operation puts or deletes multiple items in one or more tables. A single call to BatchWriteItem can write up to 16 MB of data, which can comprise as many as 25 put or delete requests. Individual items to be written can be as large as 400 KB.

BatchWriteItem cannot update items. To update items, use the UpdateItem action.

The individual PutItem and DeleteItem operations specified in BatchWriteItem are atomic; however BatchWriteItem as a whole is not. If any requested operations fail because the table's provisioned throughput is exceeded or an internal processing failure occurs, the failed operations are returned in the UnprocessedItems response parameter. You can investigate and optionally resend the requests. Typically, you would call BatchWriteItem in a loop. Each iteration would check for unprocessed items and submit a new BatchWriteItem request with those unprocessed items until all items have been processed.

Note that if none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchWriteItem will return a ProvisionedThroughputExceededException.

If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.

For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide.

With BatchWriteItem, you can efficiently write or delete large amounts of data, such as from Amazon Elastic MapReduce (EMR), or copy data from another database into DynamoDB. In order to improve performance with these large-scale operations, BatchWriteItem does not behave in the same way as individual PutItem and DeleteItem calls would. For example, you cannot specify conditions on individual put and delete requests, and BatchWriteItem does not return deleted items in the response.

If you use a programming language that supports concurrency, you can use threads to write items in parallel. Your application must include the necessary logic to manage the threads. With languages that don't support threading, you must update or delete the specified items one at a time. In both situations, BatchWriteItem performs the specified put and delete operations in parallel, giving you the power of the thread pool approach without having to introduce complexity into your application.

Parallel processing reduces latency, but each specified put and delete request consumes the same number of write capacity units whether it is processed in parallel or not. Delete operations on nonexistent items consume one write capacity unit.

If one or more of the following is true, DynamoDB rejects the entire batch write operation:

  • One or more tables specified in the BatchWriteItem request does not exist.

  • Primary key attributes specified on an item in the request do not match those in the corresponding table's primary key schema.

  • You try to perform multiple operations on the same item in the same BatchWriteItem request. For example, you cannot put and delete the same item in the same BatchWriteItem request.

  • Your request contains at least two items with identical hash and range keys (which essentially is two put operations).

  • There are more than 25 requests in the batch.

  • Any individual item in a batch exceeds 400 KB.

  • The total request size exceeds 16 MB.

", "CreateBackup": "

Creates a backup for an existing table.

Each time you create an On-Demand Backup, the entire table data is backed up. There is no limit to the number of on-demand backups that can be taken.

When you create an On-Demand Backup, a time marker of the request is cataloged, and the backup is created asynchronously, by applying all changes until the time of the request to the last full table snapshot. Backup requests are processed instantaneously and become available for restore within minutes.

You can call CreateBackup at a maximum rate of 50 times per second.

All backups in DynamoDB work without consuming any provisioned throughput on the table.

If you submit a backup request on 2018-12-14 at 14:25:00, the backup is guaranteed to contain all data committed to the table up to 14:24:00, and data committed after 14:26:00 will not be. The backup may or may not contain data modifications made between 14:24:00 and 14:26:00. On-Demand Backup does not support causal consistency.

Along with data, the following are also included on the backups:

  • Global secondary indexes (GSIs)

  • Local secondary indexes (LSIs)

  • Streams

  • Provisioned read and write capacity

", - "CreateGlobalTable": "

Creates a global table from an existing table. A global table creates a replication relationship between two or more DynamoDB tables with the same table name in the provided regions.

Tables can only be added as the replicas of a global table group under the following conditions:

  • The tables must have the same name.

  • The tables must contain no items.

  • The tables must have the same hash key and sort key (if present).

  • The tables must have DynamoDB Streams enabled (NEW_AND_OLD_IMAGES).

  • The tables must have same provisioned and maximum write capacity units.

If global secondary indexes are specified, then the following conditions must also be met:

  • The global secondary indexes must have the same name.

  • The global secondary indexes must have the same hash key and sort key (if present).

  • The global secondary indexes must have the same provisioned and maximum write capacity units.

", + "CreateGlobalTable": "

Creates a global table from an existing table. A global table creates a replication relationship between two or more DynamoDB tables with the same table name in the provided regions.

If you want to add a new replica table to a global table, each of the following conditions must be true:

  • The table must have the same primary key as all of the other replicas.

  • The table must have the same name as all of the other replicas.

  • The table must have DynamoDB Streams enabled, with the stream containing both the new and the old images of the item.

  • None of the replica tables in the global table can contain any data.

If global secondary indexes are specified, then the following conditions must also be met:

  • The global secondary indexes must have the same name.

  • The global secondary indexes must have the same hash key and sort key (if present).

Write capacity settings should be set consistently across your replica tables and secondary indexes. DynamoDB strongly recommends enabling auto scaling to manage the write capacity settings for all of your global tables replicas and indexes.

If you prefer to manage write capacity settings manually, you should provision equal replicated write capacity units to your replica tables. You should also provision equal replicated write capacity units to matching secondary indexes across your global table.

", "CreateTable": "

The CreateTable operation adds a new table to your account. In an AWS account, table names must be unique within each region. That is, you can have two tables with same name if you create the tables in different regions.

CreateTable is an asynchronous operation. Upon receiving a CreateTable request, DynamoDB immediately returns a response with a TableStatus of CREATING. After the table is created, DynamoDB sets the TableStatus to ACTIVE. You can perform read and write operations only on an ACTIVE table.

You can optionally define secondary indexes on the new table, as part of the CreateTable operation. If you want to create multiple tables with secondary indexes on them, you must create the tables sequentially. Only one table with secondary indexes can be in the CREATING state at any given time.

You can use the DescribeTable action to check the table status.

", "DeleteBackup": "

Deletes an existing backup of a table.

You can call DeleteBackup at a maximum rate of 10 times per second.

", "DeleteItem": "

Deletes a single item in a table by primary key. You can perform a conditional delete operation that deletes the item if it exists, or if it has an expected attribute value.

In addition to deleting an item, you can also return the item's attribute values in the same operation, using the ReturnValues parameter.

Unless you specify conditions, the DeleteItem is an idempotent operation; running it multiple times on the same item or attribute does not result in an error response.

Conditional deletes are useful for deleting items only if specific conditions are met. If those conditions are met, DynamoDB performs the delete. Otherwise, the item is not deleted.

", @@ -139,8 +139,8 @@ "BackupSummary$BackupArn": "

ARN associated with the backup.

", "DeleteBackupInput$BackupArn": "

The ARN associated with the backup.

", "DescribeBackupInput$BackupArn": "

The ARN associated with the backup.

", - "ListBackupsInput$ExclusiveStartBackupArn": "

LastEvaluatedBackupARN returned by the previous ListBackups call.

", - "ListBackupsOutput$LastEvaluatedBackupArn": "

Last evaluated BackupARN.

", + "ListBackupsInput$ExclusiveStartBackupArn": "

LastEvaluatedBackupArn is the ARN of the backup last evaluated when the current page of results was returned, inclusive of the current page of results. This value may be specified as the ExclusiveStartBackupArn of a new ListBackups operation in order to fetch the next page of results.

", + "ListBackupsOutput$LastEvaluatedBackupArn": "

The ARN of the backup last evaluated when the current page of results was returned, inclusive of the current page of results. This value may be specified as the ExclusiveStartBackupArn of a new ListBackups operation in order to fetch the next page of results.

If LastEvaluatedBackupArn is empty, then the last page of results has been processed and there are no more results to be retrieved.

If LastEvaluatedBackupArn is not empty, this may or may not indicate there is more data to be returned. All results are guaranteed to have been returned if and only if no value for LastEvaluatedBackupArn is returned.

", "RestoreSummary$SourceBackupArn": "

ARN of the backup from which the table was restored.

", "RestoreTableFromBackupInput$BackupArn": "

The ARN associated with the backup.

" } @@ -368,7 +368,7 @@ "ContinuousBackupsDescription": { "base": "

Represents the continuous backups and point in time recovery settings on the table.

", "refs": { - "DescribeContinuousBackupsOutput$ContinuousBackupsDescription": "

ContinuousBackupsDescription can be one of the following : ENABLED, DISABLED.

", + "DescribeContinuousBackupsOutput$ContinuousBackupsDescription": "

Represents the continuous backups and point in time recovery settings on the table.

", "UpdateContinuousBackupsOutput$ContinuousBackupsDescription": "

Represents the continuous backups and point in time recovery settings on the table.

" } }, @@ -857,6 +857,12 @@ "ScanOutput$Items": "

An array of item attributes that match the scan criteria. Each element in this array consists of an attribute name and the value for that attribute.

" } }, + "KMSMasterKeyArn": { + "base": null, + "refs": { + "SSEDescription$KMSMasterKeyArn": "

The KMS master key ARN used for the KMS encryption.

" + } + }, "Key": { "base": null, "refs": { @@ -930,7 +936,7 @@ } }, "LimitExceededException": { - "base": "

Up to 50 CreateBackup operations are allowed per second, per account. There is no limit to the number of daily on-demand backups that can be taken.

Up to 10 simultaneous table operations are allowed per account. These operations include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, and RestoreTableToPointInTime.

For tables with secondary indexes, only one of those tables can be in the CREATING state at any point in time. Do not attempt to create more than one such table simultaneously.

The total limit of tables in the ACTIVE state is 250.

", + "base": "

There is no limit to the number of daily on-demand backups that can be taken.

Up to 10 simultaneous table operations are allowed per account. These operations include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, and RestoreTableToPointInTime.

For tables with secondary indexes, only one of those tables can be in the CREATING state at any point in time. Do not attempt to create more than one such table simultaneously.

The total limit of tables in the ACTIVE state is 250.

", "refs": { } }, @@ -1434,6 +1440,12 @@ "SSEDescription$Status": "

The current state of server-side encryption:

  • ENABLING - Server-side encryption is being enabled.

  • ENABLED - Server-side encryption is enabled.

  • DISABLING - Server-side encryption is being disabled.

  • DISABLED - Server-side encryption is disabled.

" } }, + "SSEType": { + "base": null, + "refs": { + "SSEDescription$SSEType": "

Server-side encryption type:

  • AES256 - Server-side encryption which uses the AES256 algorithm.

  • KMS - Server-side encryption which uses AWS Key Management Service.

" + } + }, "ScalarAttributeType": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/api-2.json deleted file mode 100644 index 3281d2ea0..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/api-2.json +++ /dev/null @@ -1,12049 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2015-04-15", - "endpointPrefix":"ec2", - "serviceAbbreviation":"Amazon EC2", - "serviceFullName":"Amazon Elastic Compute Cloud", - "signatureVersion":"v4", - "xmlNamespace":"http://ec2.amazonaws.com/doc/2015-04-15", - "protocol":"ec2" - }, - "operations":{ - "AcceptVpcPeeringConnection":{ - "name":"AcceptVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AcceptVpcPeeringConnectionRequest"}, - "output":{"shape":"AcceptVpcPeeringConnectionResult"} - }, - "AllocateAddress":{ - "name":"AllocateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateAddressRequest"}, - "output":{"shape":"AllocateAddressResult"} - }, - "AssignPrivateIpAddresses":{ - "name":"AssignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssignPrivateIpAddressesRequest"} - }, - "AssociateAddress":{ - "name":"AssociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateAddressRequest"}, - "output":{"shape":"AssociateAddressResult"} - }, - "AssociateDhcpOptions":{ - "name":"AssociateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateDhcpOptionsRequest"} - }, - "AssociateRouteTable":{ - "name":"AssociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateRouteTableRequest"}, - "output":{"shape":"AssociateRouteTableResult"} - }, - "AttachClassicLinkVpc":{ - "name":"AttachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachClassicLinkVpcRequest"}, - "output":{"shape":"AttachClassicLinkVpcResult"} - }, - "AttachInternetGateway":{ - "name":"AttachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachInternetGatewayRequest"} - }, - "AttachNetworkInterface":{ - "name":"AttachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachNetworkInterfaceRequest"}, - "output":{"shape":"AttachNetworkInterfaceResult"} - }, - "AttachVolume":{ - "name":"AttachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVolumeRequest"}, - "output":{ - "shape":"VolumeAttachment", - "locationName":"attachment" - } - }, - "AttachVpnGateway":{ - "name":"AttachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVpnGatewayRequest"}, - "output":{"shape":"AttachVpnGatewayResult"} - }, - "AuthorizeSecurityGroupEgress":{ - "name":"AuthorizeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupEgressRequest"} - }, - "AuthorizeSecurityGroupIngress":{ - "name":"AuthorizeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupIngressRequest"} - }, - "BundleInstance":{ - "name":"BundleInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"BundleInstanceRequest"}, - "output":{"shape":"BundleInstanceResult"} - }, - "CancelBundleTask":{ - "name":"CancelBundleTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelBundleTaskRequest"}, - "output":{"shape":"CancelBundleTaskResult"} - }, - "CancelConversionTask":{ - "name":"CancelConversionTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelConversionRequest"} - }, - "CancelExportTask":{ - "name":"CancelExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelExportTaskRequest"} - }, - "CancelImportTask":{ - "name":"CancelImportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelImportTaskRequest"}, - "output":{"shape":"CancelImportTaskResult"} - }, - "CancelReservedInstancesListing":{ - "name":"CancelReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelReservedInstancesListingRequest"}, - "output":{"shape":"CancelReservedInstancesListingResult"} - }, - "CancelSpotFleetRequests":{ - "name":"CancelSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotFleetRequestsRequest"}, - "output":{"shape":"CancelSpotFleetRequestsResponse"} - }, - "CancelSpotInstanceRequests":{ - "name":"CancelSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotInstanceRequestsRequest"}, - "output":{"shape":"CancelSpotInstanceRequestsResult"} - }, - "ConfirmProductInstance":{ - "name":"ConfirmProductInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ConfirmProductInstanceRequest"}, - "output":{"shape":"ConfirmProductInstanceResult"} - }, - "CopyImage":{ - "name":"CopyImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyImageRequest"}, - "output":{"shape":"CopyImageResult"} - }, - "CopySnapshot":{ - "name":"CopySnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopySnapshotRequest"}, - "output":{"shape":"CopySnapshotResult"} - }, - "CreateCustomerGateway":{ - "name":"CreateCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateCustomerGatewayRequest"}, - "output":{"shape":"CreateCustomerGatewayResult"} - }, - "CreateDhcpOptions":{ - "name":"CreateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDhcpOptionsRequest"}, - "output":{"shape":"CreateDhcpOptionsResult"} - }, - "CreateFlowLogs":{ - "name":"CreateFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateFlowLogsRequest"}, - "output":{"shape":"CreateFlowLogsResult"} - }, - "CreateImage":{ - "name":"CreateImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateImageRequest"}, - "output":{"shape":"CreateImageResult"} - }, - "CreateInstanceExportTask":{ - "name":"CreateInstanceExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInstanceExportTaskRequest"}, - "output":{"shape":"CreateInstanceExportTaskResult"} - }, - "CreateInternetGateway":{ - "name":"CreateInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInternetGatewayRequest"}, - "output":{"shape":"CreateInternetGatewayResult"} - }, - "CreateKeyPair":{ - "name":"CreateKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateKeyPairRequest"}, - "output":{ - "shape":"KeyPair", - "locationName":"keyPair" - } - }, - "CreateNetworkAcl":{ - "name":"CreateNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclRequest"}, - "output":{"shape":"CreateNetworkAclResult"} - }, - "CreateNetworkAclEntry":{ - "name":"CreateNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclEntryRequest"} - }, - "CreateNetworkInterface":{ - "name":"CreateNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkInterfaceRequest"}, - "output":{"shape":"CreateNetworkInterfaceResult"} - }, - "CreatePlacementGroup":{ - "name":"CreatePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreatePlacementGroupRequest"} - }, - "CreateReservedInstancesListing":{ - "name":"CreateReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateReservedInstancesListingRequest"}, - "output":{"shape":"CreateReservedInstancesListingResult"} - }, - "CreateRoute":{ - "name":"CreateRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteRequest"}, - "output":{"shape":"CreateRouteResult"} - }, - "CreateRouteTable":{ - "name":"CreateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteTableRequest"}, - "output":{"shape":"CreateRouteTableResult"} - }, - "CreateSecurityGroup":{ - "name":"CreateSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSecurityGroupRequest"}, - "output":{"shape":"CreateSecurityGroupResult"} - }, - "CreateSnapshot":{ - "name":"CreateSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSnapshotRequest"}, - "output":{ - "shape":"Snapshot", - "locationName":"snapshot" - } - }, - "CreateSpotDatafeedSubscription":{ - "name":"CreateSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"CreateSpotDatafeedSubscriptionResult"} - }, - "CreateSubnet":{ - "name":"CreateSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSubnetRequest"}, - "output":{"shape":"CreateSubnetResult"} - }, - "CreateTags":{ - "name":"CreateTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateTagsRequest"} - }, - "CreateVolume":{ - "name":"CreateVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVolumeRequest"}, - "output":{ - "shape":"Volume", - "locationName":"volume" - } - }, - "CreateVpc":{ - "name":"CreateVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcRequest"}, - "output":{"shape":"CreateVpcResult"} - }, - "CreateVpcEndpoint":{ - "name":"CreateVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcEndpointRequest"}, - "output":{"shape":"CreateVpcEndpointResult"} - }, - "CreateVpcPeeringConnection":{ - "name":"CreateVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcPeeringConnectionRequest"}, - "output":{"shape":"CreateVpcPeeringConnectionResult"} - }, - "CreateVpnConnection":{ - "name":"CreateVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRequest"}, - "output":{"shape":"CreateVpnConnectionResult"} - }, - "CreateVpnConnectionRoute":{ - "name":"CreateVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRouteRequest"} - }, - "CreateVpnGateway":{ - "name":"CreateVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnGatewayRequest"}, - "output":{"shape":"CreateVpnGatewayResult"} - }, - "DeleteCustomerGateway":{ - "name":"DeleteCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteCustomerGatewayRequest"} - }, - "DeleteDhcpOptions":{ - "name":"DeleteDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDhcpOptionsRequest"} - }, - "DeleteFlowLogs":{ - "name":"DeleteFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteFlowLogsRequest"}, - "output":{"shape":"DeleteFlowLogsResult"} - }, - "DeleteInternetGateway":{ - "name":"DeleteInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteInternetGatewayRequest"} - }, - "DeleteKeyPair":{ - "name":"DeleteKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteKeyPairRequest"} - }, - "DeleteNetworkAcl":{ - "name":"DeleteNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclRequest"} - }, - "DeleteNetworkAclEntry":{ - "name":"DeleteNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclEntryRequest"} - }, - "DeleteNetworkInterface":{ - "name":"DeleteNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkInterfaceRequest"} - }, - "DeletePlacementGroup":{ - "name":"DeletePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeletePlacementGroupRequest"} - }, - "DeleteRoute":{ - "name":"DeleteRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteRequest"} - }, - "DeleteRouteTable":{ - "name":"DeleteRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteTableRequest"} - }, - "DeleteSecurityGroup":{ - "name":"DeleteSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSecurityGroupRequest"} - }, - "DeleteSnapshot":{ - "name":"DeleteSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSnapshotRequest"} - }, - "DeleteSpotDatafeedSubscription":{ - "name":"DeleteSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSpotDatafeedSubscriptionRequest"} - }, - "DeleteSubnet":{ - "name":"DeleteSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSubnetRequest"} - }, - "DeleteTags":{ - "name":"DeleteTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteTagsRequest"} - }, - "DeleteVolume":{ - "name":"DeleteVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVolumeRequest"} - }, - "DeleteVpc":{ - "name":"DeleteVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcRequest"} - }, - "DeleteVpcEndpoints":{ - "name":"DeleteVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcEndpointsRequest"}, - "output":{"shape":"DeleteVpcEndpointsResult"} - }, - "DeleteVpcPeeringConnection":{ - "name":"DeleteVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcPeeringConnectionRequest"}, - "output":{"shape":"DeleteVpcPeeringConnectionResult"} - }, - "DeleteVpnConnection":{ - "name":"DeleteVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRequest"} - }, - "DeleteVpnConnectionRoute":{ - "name":"DeleteVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRouteRequest"} - }, - "DeleteVpnGateway":{ - "name":"DeleteVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnGatewayRequest"} - }, - "DeregisterImage":{ - "name":"DeregisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeregisterImageRequest"} - }, - "DescribeAccountAttributes":{ - "name":"DescribeAccountAttributes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAccountAttributesRequest"}, - "output":{"shape":"DescribeAccountAttributesResult"} - }, - "DescribeAddresses":{ - "name":"DescribeAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAddressesRequest"}, - "output":{"shape":"DescribeAddressesResult"} - }, - "DescribeAvailabilityZones":{ - "name":"DescribeAvailabilityZones", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAvailabilityZonesRequest"}, - "output":{"shape":"DescribeAvailabilityZonesResult"} - }, - "DescribeBundleTasks":{ - "name":"DescribeBundleTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeBundleTasksRequest"}, - "output":{"shape":"DescribeBundleTasksResult"} - }, - "DescribeClassicLinkInstances":{ - "name":"DescribeClassicLinkInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeClassicLinkInstancesRequest"}, - "output":{"shape":"DescribeClassicLinkInstancesResult"} - }, - "DescribeConversionTasks":{ - "name":"DescribeConversionTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeConversionTasksRequest"}, - "output":{"shape":"DescribeConversionTasksResult"} - }, - "DescribeCustomerGateways":{ - "name":"DescribeCustomerGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeCustomerGatewaysRequest"}, - "output":{"shape":"DescribeCustomerGatewaysResult"} - }, - "DescribeDhcpOptions":{ - "name":"DescribeDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDhcpOptionsRequest"}, - "output":{"shape":"DescribeDhcpOptionsResult"} - }, - "DescribeExportTasks":{ - "name":"DescribeExportTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeExportTasksRequest"}, - "output":{"shape":"DescribeExportTasksResult"} - }, - "DescribeFlowLogs":{ - "name":"DescribeFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeFlowLogsRequest"}, - "output":{"shape":"DescribeFlowLogsResult"} - }, - "DescribeImageAttribute":{ - "name":"DescribeImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImageAttributeRequest"}, - "output":{ - "shape":"ImageAttribute", - "locationName":"imageAttribute" - } - }, - "DescribeImages":{ - "name":"DescribeImages", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImagesRequest"}, - "output":{"shape":"DescribeImagesResult"} - }, - "DescribeImportImageTasks":{ - "name":"DescribeImportImageTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportImageTasksRequest"}, - "output":{"shape":"DescribeImportImageTasksResult"} - }, - "DescribeImportSnapshotTasks":{ - "name":"DescribeImportSnapshotTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportSnapshotTasksRequest"}, - "output":{"shape":"DescribeImportSnapshotTasksResult"} - }, - "DescribeInstanceAttribute":{ - "name":"DescribeInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceAttributeRequest"}, - "output":{"shape":"InstanceAttribute"} - }, - "DescribeInstanceStatus":{ - "name":"DescribeInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceStatusRequest"}, - "output":{"shape":"DescribeInstanceStatusResult"} - }, - "DescribeInstances":{ - "name":"DescribeInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstancesRequest"}, - "output":{"shape":"DescribeInstancesResult"} - }, - "DescribeInternetGateways":{ - "name":"DescribeInternetGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInternetGatewaysRequest"}, - "output":{"shape":"DescribeInternetGatewaysResult"} - }, - "DescribeKeyPairs":{ - "name":"DescribeKeyPairs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeKeyPairsRequest"}, - "output":{"shape":"DescribeKeyPairsResult"} - }, - "DescribeMovingAddresses":{ - "name":"DescribeMovingAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeMovingAddressesRequest"}, - "output":{"shape":"DescribeMovingAddressesResult"} - }, - "DescribeNetworkAcls":{ - "name":"DescribeNetworkAcls", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkAclsRequest"}, - "output":{"shape":"DescribeNetworkAclsResult"} - }, - "DescribeNetworkInterfaceAttribute":{ - "name":"DescribeNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfaceAttributeRequest"}, - "output":{"shape":"DescribeNetworkInterfaceAttributeResult"} - }, - "DescribeNetworkInterfaces":{ - "name":"DescribeNetworkInterfaces", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfacesRequest"}, - "output":{"shape":"DescribeNetworkInterfacesResult"} - }, - "DescribePlacementGroups":{ - "name":"DescribePlacementGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePlacementGroupsRequest"}, - "output":{"shape":"DescribePlacementGroupsResult"} - }, - "DescribePrefixLists":{ - "name":"DescribePrefixLists", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePrefixListsRequest"}, - "output":{"shape":"DescribePrefixListsResult"} - }, - "DescribeRegions":{ - "name":"DescribeRegions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRegionsRequest"}, - "output":{"shape":"DescribeRegionsResult"} - }, - "DescribeReservedInstances":{ - "name":"DescribeReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesRequest"}, - "output":{"shape":"DescribeReservedInstancesResult"} - }, - "DescribeReservedInstancesListings":{ - "name":"DescribeReservedInstancesListings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesListingsRequest"}, - "output":{"shape":"DescribeReservedInstancesListingsResult"} - }, - "DescribeReservedInstancesModifications":{ - "name":"DescribeReservedInstancesModifications", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesModificationsRequest"}, - "output":{"shape":"DescribeReservedInstancesModificationsResult"} - }, - "DescribeReservedInstancesOfferings":{ - "name":"DescribeReservedInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesOfferingsRequest"}, - "output":{"shape":"DescribeReservedInstancesOfferingsResult"} - }, - "DescribeRouteTables":{ - "name":"DescribeRouteTables", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRouteTablesRequest"}, - "output":{"shape":"DescribeRouteTablesResult"} - }, - "DescribeSecurityGroups":{ - "name":"DescribeSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupsRequest"}, - "output":{"shape":"DescribeSecurityGroupsResult"} - }, - "DescribeSnapshotAttribute":{ - "name":"DescribeSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotAttributeRequest"}, - "output":{"shape":"DescribeSnapshotAttributeResult"} - }, - "DescribeSnapshots":{ - "name":"DescribeSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotsRequest"}, - "output":{"shape":"DescribeSnapshotsResult"} - }, - "DescribeSpotDatafeedSubscription":{ - "name":"DescribeSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"DescribeSpotDatafeedSubscriptionResult"} - }, - "DescribeSpotFleetInstances":{ - "name":"DescribeSpotFleetInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetInstancesRequest"}, - "output":{"shape":"DescribeSpotFleetInstancesResponse"} - }, - "DescribeSpotFleetRequestHistory":{ - "name":"DescribeSpotFleetRequestHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestHistoryRequest"}, - "output":{"shape":"DescribeSpotFleetRequestHistoryResponse"} - }, - "DescribeSpotFleetRequests":{ - "name":"DescribeSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestsRequest"}, - "output":{"shape":"DescribeSpotFleetRequestsResponse"} - }, - "DescribeSpotInstanceRequests":{ - "name":"DescribeSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotInstanceRequestsRequest"}, - "output":{"shape":"DescribeSpotInstanceRequestsResult"} - }, - "DescribeSpotPriceHistory":{ - "name":"DescribeSpotPriceHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotPriceHistoryRequest"}, - "output":{"shape":"DescribeSpotPriceHistoryResult"} - }, - "DescribeSubnets":{ - "name":"DescribeSubnets", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSubnetsRequest"}, - "output":{"shape":"DescribeSubnetsResult"} - }, - "DescribeTags":{ - "name":"DescribeTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeTagsRequest"}, - "output":{"shape":"DescribeTagsResult"} - }, - "DescribeVolumeAttribute":{ - "name":"DescribeVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeAttributeRequest"}, - "output":{"shape":"DescribeVolumeAttributeResult"} - }, - "DescribeVolumeStatus":{ - "name":"DescribeVolumeStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeStatusRequest"}, - "output":{"shape":"DescribeVolumeStatusResult"} - }, - "DescribeVolumes":{ - "name":"DescribeVolumes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumesRequest"}, - "output":{"shape":"DescribeVolumesResult"} - }, - "DescribeVpcAttribute":{ - "name":"DescribeVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcAttributeRequest"}, - "output":{"shape":"DescribeVpcAttributeResult"} - }, - "DescribeVpcClassicLink":{ - "name":"DescribeVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkRequest"}, - "output":{"shape":"DescribeVpcClassicLinkResult"} - }, - "DescribeVpcEndpointServices":{ - "name":"DescribeVpcEndpointServices", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointServicesRequest"}, - "output":{"shape":"DescribeVpcEndpointServicesResult"} - }, - "DescribeVpcEndpoints":{ - "name":"DescribeVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointsRequest"}, - "output":{"shape":"DescribeVpcEndpointsResult"} - }, - "DescribeVpcPeeringConnections":{ - "name":"DescribeVpcPeeringConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcPeeringConnectionsRequest"}, - "output":{"shape":"DescribeVpcPeeringConnectionsResult"} - }, - "DescribeVpcs":{ - "name":"DescribeVpcs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcsRequest"}, - "output":{"shape":"DescribeVpcsResult"} - }, - "DescribeVpnConnections":{ - "name":"DescribeVpnConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnConnectionsRequest"}, - "output":{"shape":"DescribeVpnConnectionsResult"} - }, - "DescribeVpnGateways":{ - "name":"DescribeVpnGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnGatewaysRequest"}, - "output":{"shape":"DescribeVpnGatewaysResult"} - }, - "DetachClassicLinkVpc":{ - "name":"DetachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachClassicLinkVpcRequest"}, - "output":{"shape":"DetachClassicLinkVpcResult"} - }, - "DetachInternetGateway":{ - "name":"DetachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachInternetGatewayRequest"} - }, - "DetachNetworkInterface":{ - "name":"DetachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachNetworkInterfaceRequest"} - }, - "DetachVolume":{ - "name":"DetachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVolumeRequest"}, - "output":{ - "shape":"VolumeAttachment", - "locationName":"attachment" - } - }, - "DetachVpnGateway":{ - "name":"DetachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVpnGatewayRequest"} - }, - "DisableVgwRoutePropagation":{ - "name":"DisableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVgwRoutePropagationRequest"} - }, - "DisableVpcClassicLink":{ - "name":"DisableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkRequest"}, - "output":{"shape":"DisableVpcClassicLinkResult"} - }, - "DisassociateAddress":{ - "name":"DisassociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateAddressRequest"} - }, - "DisassociateRouteTable":{ - "name":"DisassociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateRouteTableRequest"} - }, - "EnableVgwRoutePropagation":{ - "name":"EnableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVgwRoutePropagationRequest"} - }, - "EnableVolumeIO":{ - "name":"EnableVolumeIO", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVolumeIORequest"} - }, - "EnableVpcClassicLink":{ - "name":"EnableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkRequest"}, - "output":{"shape":"EnableVpcClassicLinkResult"} - }, - "GetConsoleOutput":{ - "name":"GetConsoleOutput", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleOutputRequest"}, - "output":{"shape":"GetConsoleOutputResult"} - }, - "GetPasswordData":{ - "name":"GetPasswordData", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetPasswordDataRequest"}, - "output":{"shape":"GetPasswordDataResult"} - }, - "ImportImage":{ - "name":"ImportImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportImageRequest"}, - "output":{"shape":"ImportImageResult"} - }, - "ImportInstance":{ - "name":"ImportInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportInstanceRequest"}, - "output":{"shape":"ImportInstanceResult"} - }, - "ImportKeyPair":{ - "name":"ImportKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportKeyPairRequest"}, - "output":{"shape":"ImportKeyPairResult"} - }, - "ImportSnapshot":{ - "name":"ImportSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportSnapshotRequest"}, - "output":{"shape":"ImportSnapshotResult"} - }, - "ImportVolume":{ - "name":"ImportVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportVolumeRequest"}, - "output":{"shape":"ImportVolumeResult"} - }, - "ModifyImageAttribute":{ - "name":"ModifyImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyImageAttributeRequest"} - }, - "ModifyInstanceAttribute":{ - "name":"ModifyInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstanceAttributeRequest"} - }, - "ModifyNetworkInterfaceAttribute":{ - "name":"ModifyNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyNetworkInterfaceAttributeRequest"} - }, - "ModifyReservedInstances":{ - "name":"ModifyReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyReservedInstancesRequest"}, - "output":{"shape":"ModifyReservedInstancesResult"} - }, - "ModifySnapshotAttribute":{ - "name":"ModifySnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySnapshotAttributeRequest"} - }, - "ModifySubnetAttribute":{ - "name":"ModifySubnetAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySubnetAttributeRequest"} - }, - "ModifyVolumeAttribute":{ - "name":"ModifyVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVolumeAttributeRequest"} - }, - "ModifyVpcAttribute":{ - "name":"ModifyVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcAttributeRequest"} - }, - "ModifyVpcEndpoint":{ - "name":"ModifyVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcEndpointRequest"}, - "output":{"shape":"ModifyVpcEndpointResult"} - }, - "MonitorInstances":{ - "name":"MonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MonitorInstancesRequest"}, - "output":{"shape":"MonitorInstancesResult"} - }, - "MoveAddressToVpc":{ - "name":"MoveAddressToVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MoveAddressToVpcRequest"}, - "output":{"shape":"MoveAddressToVpcResult"} - }, - "PurchaseReservedInstancesOffering":{ - "name":"PurchaseReservedInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedInstancesOfferingRequest"}, - "output":{"shape":"PurchaseReservedInstancesOfferingResult"} - }, - "RebootInstances":{ - "name":"RebootInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootInstancesRequest"} - }, - "RegisterImage":{ - "name":"RegisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RegisterImageRequest"}, - "output":{"shape":"RegisterImageResult"} - }, - "RejectVpcPeeringConnection":{ - "name":"RejectVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RejectVpcPeeringConnectionRequest"}, - "output":{"shape":"RejectVpcPeeringConnectionResult"} - }, - "ReleaseAddress":{ - "name":"ReleaseAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseAddressRequest"} - }, - "ReplaceNetworkAclAssociation":{ - "name":"ReplaceNetworkAclAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclAssociationRequest"}, - "output":{"shape":"ReplaceNetworkAclAssociationResult"} - }, - "ReplaceNetworkAclEntry":{ - "name":"ReplaceNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclEntryRequest"} - }, - "ReplaceRoute":{ - "name":"ReplaceRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteRequest"} - }, - "ReplaceRouteTableAssociation":{ - "name":"ReplaceRouteTableAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteTableAssociationRequest"}, - "output":{"shape":"ReplaceRouteTableAssociationResult"} - }, - "ReportInstanceStatus":{ - "name":"ReportInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReportInstanceStatusRequest"} - }, - "RequestSpotFleet":{ - "name":"RequestSpotFleet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotFleetRequest"}, - "output":{"shape":"RequestSpotFleetResponse"} - }, - "RequestSpotInstances":{ - "name":"RequestSpotInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotInstancesRequest"}, - "output":{"shape":"RequestSpotInstancesResult"} - }, - "ResetImageAttribute":{ - "name":"ResetImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetImageAttributeRequest"} - }, - "ResetInstanceAttribute":{ - "name":"ResetInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetInstanceAttributeRequest"} - }, - "ResetNetworkInterfaceAttribute":{ - "name":"ResetNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetNetworkInterfaceAttributeRequest"} - }, - "ResetSnapshotAttribute":{ - "name":"ResetSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetSnapshotAttributeRequest"} - }, - "RestoreAddressToClassic":{ - "name":"RestoreAddressToClassic", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreAddressToClassicRequest"}, - "output":{"shape":"RestoreAddressToClassicResult"} - }, - "RevokeSecurityGroupEgress":{ - "name":"RevokeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupEgressRequest"} - }, - "RevokeSecurityGroupIngress":{ - "name":"RevokeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupIngressRequest"} - }, - "RunInstances":{ - "name":"RunInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunInstancesRequest"}, - "output":{ - "shape":"Reservation", - "locationName":"reservation" - } - }, - "StartInstances":{ - "name":"StartInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StartInstancesRequest"}, - "output":{"shape":"StartInstancesResult"} - }, - "StopInstances":{ - "name":"StopInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StopInstancesRequest"}, - "output":{"shape":"StopInstancesResult"} - }, - "TerminateInstances":{ - "name":"TerminateInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"TerminateInstancesRequest"}, - "output":{"shape":"TerminateInstancesResult"} - }, - "UnassignPrivateIpAddresses":{ - "name":"UnassignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnassignPrivateIpAddressesRequest"} - }, - "UnmonitorInstances":{ - "name":"UnmonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnmonitorInstancesRequest"}, - "output":{"shape":"UnmonitorInstancesResult"} - } - }, - "shapes":{ - "AcceptVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "AcceptVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "AccountAttribute":{ - "type":"structure", - "members":{ - "AttributeName":{ - "shape":"String", - "locationName":"attributeName" - }, - "AttributeValues":{ - "shape":"AccountAttributeValueList", - "locationName":"attributeValueSet" - } - } - }, - "AccountAttributeList":{ - "type":"list", - "member":{ - "shape":"AccountAttribute", - "locationName":"item" - } - }, - "AccountAttributeName":{ - "type":"string", - "enum":[ - "supported-platforms", - "default-vpc" - ] - }, - "AccountAttributeNameStringList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeName", - "locationName":"attributeName" - } - }, - "AccountAttributeValue":{ - "type":"structure", - "members":{ - "AttributeValue":{ - "shape":"String", - "locationName":"attributeValue" - } - } - }, - "AccountAttributeValueList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeValue", - "locationName":"item" - } - }, - "ActiveInstance":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - } - } - }, - "ActiveInstanceSet":{ - "type":"list", - "member":{ - "shape":"ActiveInstance", - "locationName":"item" - } - }, - "Address":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "NetworkInterfaceOwnerId":{ - "shape":"String", - "locationName":"networkInterfaceOwnerId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "AddressList":{ - "type":"list", - "member":{ - "shape":"Address", - "locationName":"item" - } - }, - "AllocateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Domain":{"shape":"DomainType"} - } - }, - "AllocateAddressResult":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - } - } - }, - "AllocationIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"AllocationId" - } - }, - "AllocationStrategy":{ - "type":"string", - "enum":[ - "lowestPrice", - "diversified" - ] - }, - "ArchitectureValues":{ - "type":"string", - "enum":[ - "i386", - "x86_64" - ] - }, - "AssignPrivateIpAddressesRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AllowReassignment":{ - "shape":"Boolean", - "locationName":"allowReassignment" - } - } - }, - "AssociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "AllowReassociation":{ - "shape":"Boolean", - "locationName":"allowReassociation" - } - } - }, - "AssociateAddressResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AssociateDhcpOptionsRequest":{ - "type":"structure", - "required":[ - "DhcpOptionsId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AssociateRouteTableRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "AssociateRouteTableResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AttachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId", - "Groups" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"SecurityGroupId" - } - } - }, - "AttachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "AttachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "AttachNetworkInterfaceRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "InstanceId", - "DeviceIndex" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - } - } - }, - "AttachNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - } - } - }, - "AttachVolumeRequest":{ - "type":"structure", - "required":[ - "VolumeId", - "InstanceId", - "Device" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"} - } - }, - "AttachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AttachVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpcAttachment":{ - "shape":"VpcAttachment", - "locationName":"attachment" - } - } - }, - "AttachmentStatus":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "AttributeBooleanValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Boolean", - "locationName":"value" - } - } - }, - "AttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "AuthorizeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "AuthorizeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "ZoneName":{ - "shape":"String", - "locationName":"zoneName" - }, - "State":{ - "shape":"AvailabilityZoneState", - "locationName":"zoneState" - }, - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Messages":{ - "shape":"AvailabilityZoneMessageList", - "locationName":"messageSet" - } - } - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"item" - } - }, - "AvailabilityZoneMessage":{ - "type":"structure", - "members":{ - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "AvailabilityZoneMessageList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZoneMessage", - "locationName":"item" - } - }, - "AvailabilityZoneState":{ - "type":"string", - "enum":["available"] - }, - "BatchState":{ - "type":"string", - "enum":[ - "submitted", - "active", - "cancelled", - "failed", - "cancelled_running", - "cancelled_terminating" - ] - }, - "BlockDeviceMapping":{ - "type":"structure", - "members":{ - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsBlockDevice", - "locationName":"ebs" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "BlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"item" - } - }, - "BlockDeviceMappingRequestList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "Boolean":{"type":"boolean"}, - "BundleIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"BundleId" - } - }, - "BundleInstanceRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Storage" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "Storage":{"shape":"Storage"} - } - }, - "BundleInstanceResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "BundleTask":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "BundleId":{ - "shape":"String", - "locationName":"bundleId" - }, - "State":{ - "shape":"BundleTaskState", - "locationName":"state" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Storage":{ - "shape":"Storage", - "locationName":"storage" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "BundleTaskError":{ - "shape":"BundleTaskError", - "locationName":"error" - } - } - }, - "BundleTaskError":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "BundleTaskList":{ - "type":"list", - "member":{ - "shape":"BundleTask", - "locationName":"item" - } - }, - "BundleTaskState":{ - "type":"string", - "enum":[ - "pending", - "waiting-for-shutdown", - "bundling", - "storing", - "cancelling", - "complete", - "failed" - ] - }, - "CancelBatchErrorCode":{ - "type":"string", - "enum":[ - "fleetRequestIdDoesNotExist", - "fleetRequestIdMalformed", - "fleetRequestNotInCancellableState", - "unexpectedError" - ] - }, - "CancelBundleTaskRequest":{ - "type":"structure", - "required":["BundleId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleId":{"shape":"String"} - } - }, - "CancelBundleTaskResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "CancelConversionRequest":{ - "type":"structure", - "required":["ConversionTaskId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ReasonMessage":{ - "shape":"String", - "locationName":"reasonMessage" - } - } - }, - "CancelExportTaskRequest":{ - "type":"structure", - "required":["ExportTaskId"], - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - } - } - }, - "CancelImportTaskRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskId":{"shape":"String"}, - "CancelReason":{"shape":"String"} - } - }, - "CancelImportTaskResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "PreviousState":{ - "shape":"String", - "locationName":"previousState" - } - } - }, - "CancelReservedInstancesListingRequest":{ - "type":"structure", - "required":["ReservedInstancesListingId"], - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - } - } - }, - "CancelReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CancelSpotFleetRequestsError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"CancelBatchErrorCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "CancelSpotFleetRequestsErrorItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "Error" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "Error":{ - "shape":"CancelSpotFleetRequestsError", - "locationName":"error" - } - } - }, - "CancelSpotFleetRequestsErrorSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsErrorItem", - "locationName":"item" - } - }, - "CancelSpotFleetRequestsRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestIds", - "TerminateInstances" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "TerminateInstances":{ - "shape":"Boolean", - "locationName":"terminateInstances" - } - } - }, - "CancelSpotFleetRequestsResponse":{ - "type":"structure", - "members":{ - "UnsuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsErrorSet", - "locationName":"unsuccessfulFleetRequestSet" - }, - "SuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsSuccessSet", - "locationName":"successfulFleetRequestSet" - } - } - }, - "CancelSpotFleetRequestsSuccessItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "CurrentSpotFleetRequestState", - "PreviousSpotFleetRequestState" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "CurrentSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"currentSpotFleetRequestState" - }, - "PreviousSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"previousSpotFleetRequestState" - } - } - }, - "CancelSpotFleetRequestsSuccessSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsSuccessItem", - "locationName":"item" - } - }, - "CancelSpotInstanceRequestState":{ - "type":"string", - "enum":[ - "active", - "open", - "closed", - "cancelled", - "completed" - ] - }, - "CancelSpotInstanceRequestsRequest":{ - "type":"structure", - "required":["SpotInstanceRequestIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - } - } - }, - "CancelSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "CancelledSpotInstanceRequests":{ - "shape":"CancelledSpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "CancelledSpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "State":{ - "shape":"CancelSpotInstanceRequestState", - "locationName":"state" - } - } - }, - "CancelledSpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"CancelledSpotInstanceRequest", - "locationName":"item" - } - }, - "ClassicLinkInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ClassicLinkInstanceList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkInstance", - "locationName":"item" - } - }, - "ClientData":{ - "type":"structure", - "members":{ - "UploadStart":{"shape":"DateTime"}, - "UploadEnd":{"shape":"DateTime"}, - "UploadSize":{"shape":"Double"}, - "Comment":{"shape":"String"} - } - }, - "ConfirmProductInstanceRequest":{ - "type":"structure", - "required":[ - "ProductCode", - "InstanceId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ProductCode":{"shape":"String"}, - "InstanceId":{"shape":"String"} - } - }, - "ConfirmProductInstanceResult":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ContainerFormat":{ - "type":"string", - "enum":["ova"] - }, - "ConversionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "ConversionTask":{ - "type":"structure", - "required":[ - "ConversionTaskId", - "State" - ], - "members":{ - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ExpirationTime":{ - "shape":"String", - "locationName":"expirationTime" - }, - "ImportInstance":{ - "shape":"ImportInstanceTaskDetails", - "locationName":"importInstance" - }, - "ImportVolume":{ - "shape":"ImportVolumeTaskDetails", - "locationName":"importVolume" - }, - "State":{ - "shape":"ConversionTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ConversionTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "CopyImageRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceImageId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceImageId":{"shape":"String"}, - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CopyImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CopySnapshotRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceSnapshotId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceSnapshotId":{"shape":"String"}, - "Description":{"shape":"String"}, - "DestinationRegion":{ - "shape":"String", - "locationName":"destinationRegion" - }, - "PresignedUrl":{ - "shape":"String", - "locationName":"presignedUrl" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopySnapshotResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - } - } - }, - "CreateCustomerGatewayRequest":{ - "type":"structure", - "required":[ - "Type", - "PublicIp", - "BgpAsn" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "PublicIp":{ - "shape":"String", - "locationName":"IpAddress" - }, - "BgpAsn":{"shape":"Integer"} - } - }, - "CreateCustomerGatewayResult":{ - "type":"structure", - "members":{ - "CustomerGateway":{ - "shape":"CustomerGateway", - "locationName":"customerGateway" - } - } - }, - "CreateDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpConfigurations"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpConfigurations":{ - "shape":"NewDhcpConfigurationList", - "locationName":"dhcpConfiguration" - } - } - }, - "CreateDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptions", - "locationName":"dhcpOptions" - } - } - }, - "CreateFlowLogsRequest":{ - "type":"structure", - "required":[ - "ResourceIds", - "ResourceType", - "TrafficType", - "LogGroupName", - "DeliverLogsPermissionArn" - ], - "members":{ - "ResourceIds":{ - "shape":"ValueStringList", - "locationName":"ResourceId" - }, - "ResourceType":{"shape":"FlowLogsResourceType"}, - "TrafficType":{"shape":"TrafficType"}, - "LogGroupName":{"shape":"String"}, - "DeliverLogsPermissionArn":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"flowLogIdSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "CreateImageRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NoReboot":{ - "shape":"Boolean", - "locationName":"noReboot" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"blockDeviceMapping" - } - } - }, - "CreateImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CreateInstanceExportTaskRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "Description":{ - "shape":"String", - "locationName":"description" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - }, - "ExportToS3Task":{ - "shape":"ExportToS3TaskSpecification", - "locationName":"exportToS3" - } - } - }, - "CreateInstanceExportTaskResult":{ - "type":"structure", - "members":{ - "ExportTask":{ - "shape":"ExportTask", - "locationName":"exportTask" - } - } - }, - "CreateInternetGatewayRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateInternetGatewayResult":{ - "type":"structure", - "members":{ - "InternetGateway":{ - "shape":"InternetGateway", - "locationName":"internetGateway" - } - } - }, - "CreateKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "CreateNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "CreateNetworkAclRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateNetworkAclResult":{ - "type":"structure", - "members":{ - "NetworkAcl":{ - "shape":"NetworkAcl", - "locationName":"networkAcl" - } - } - }, - "CreateNetworkInterfaceRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "NetworkInterface":{ - "shape":"NetworkInterface", - "locationName":"networkInterface" - } - } - }, - "CreatePlacementGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Strategy" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - } - } - }, - "CreateReservedInstancesListingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesId", - "InstanceCount", - "PriceSchedules", - "ClientToken" - ], - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "PriceSchedules":{ - "shape":"PriceScheduleSpecificationList", - "locationName":"priceSchedules" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CreateRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "CreateRouteResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "CreateRouteTableRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateRouteTableResult":{ - "type":"structure", - "members":{ - "RouteTable":{ - "shape":"RouteTable", - "locationName":"routeTable" - } - } - }, - "CreateSecurityGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Description" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "Description":{ - "shape":"String", - "locationName":"GroupDescription" - }, - "VpcId":{"shape":"String"} - } - }, - "CreateSecurityGroupResult":{ - "type":"structure", - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "CreateSnapshotRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Description":{"shape":"String"} - } - }, - "CreateSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "required":["Bucket"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - } - } - }, - "CreateSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "CreateSubnetRequest":{ - "type":"structure", - "required":[ - "VpcId", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "CidrBlock":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateSubnetResult":{ - "type":"structure", - "members":{ - "Subnet":{ - "shape":"Subnet", - "locationName":"subnet" - } - } - }, - "CreateTagsRequest":{ - "type":"structure", - "required":[ - "Resources", - "Tags" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"ResourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"Tag" - } - } - }, - "CreateVolumePermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "CreateVolumePermissionList":{ - "type":"list", - "member":{ - "shape":"CreateVolumePermission", - "locationName":"item" - } - }, - "CreateVolumePermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"CreateVolumePermissionList"}, - "Remove":{"shape":"CreateVolumePermissionList"} - } - }, - "CreateVolumeRequest":{ - "type":"structure", - "required":["AvailabilityZone"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Size":{"shape":"Integer"}, - "SnapshotId":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "VolumeType":{"shape":"VolumeType"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{"shape":"String"} - } - }, - "CreateVpcEndpointRequest":{ - "type":"structure", - "required":[ - "VpcId", - "ServiceName" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "ServiceName":{"shape":"String"}, - "PolicyDocument":{"shape":"String"}, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "ClientToken":{"shape":"String"} - } - }, - "CreateVpcEndpointResult":{ - "type":"structure", - "members":{ - "VpcEndpoint":{ - "shape":"VpcEndpoint", - "locationName":"vpcEndpoint" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeerVpcId":{ - "shape":"String", - "locationName":"peerVpcId" - }, - "PeerOwnerId":{ - "shape":"String", - "locationName":"peerOwnerId" - } - } - }, - "CreateVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "CreateVpcRequest":{ - "type":"structure", - "required":["CidrBlock"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CidrBlock":{"shape":"String"}, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - } - } - }, - "CreateVpcResult":{ - "type":"structure", - "members":{ - "Vpc":{ - "shape":"Vpc", - "locationName":"vpc" - } - } - }, - "CreateVpnConnectionRequest":{ - "type":"structure", - "required":[ - "Type", - "CustomerGatewayId", - "VpnGatewayId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"String"}, - "CustomerGatewayId":{"shape":"String"}, - "VpnGatewayId":{"shape":"String"}, - "Options":{ - "shape":"VpnConnectionOptionsSpecification", - "locationName":"options" - } - } - }, - "CreateVpnConnectionResult":{ - "type":"structure", - "members":{ - "VpnConnection":{ - "shape":"VpnConnection", - "locationName":"vpnConnection" - } - } - }, - "CreateVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "CreateVpnGatewayRequest":{ - "type":"structure", - "required":["Type"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpnGateway":{ - "shape":"VpnGateway", - "locationName":"vpnGateway" - } - } - }, - "CurrencyCodeValues":{ - "type":"string", - "enum":["USD"] - }, - "CustomerGateway":{ - "type":"structure", - "members":{ - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "Type":{ - "shape":"String", - "locationName":"type" - }, - "IpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "BgpAsn":{ - "shape":"String", - "locationName":"bgpAsn" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "CustomerGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"CustomerGatewayId" - } - }, - "CustomerGatewayList":{ - "type":"list", - "member":{ - "shape":"CustomerGateway", - "locationName":"item" - } - }, - "DatafeedSubscriptionState":{ - "type":"string", - "enum":[ - "Active", - "Inactive" - ] - }, - "DateTime":{"type":"timestamp"}, - "DeleteCustomerGatewayRequest":{ - "type":"structure", - "required":["CustomerGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayId":{"shape":"String"} - } - }, - "DeleteDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpOptionsId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"} - } - }, - "DeleteFlowLogsRequest":{ - "type":"structure", - "required":["FlowLogIds"], - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - } - } - }, - "DeleteFlowLogsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteInternetGatewayRequest":{ - "type":"structure", - "required":["InternetGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - } - } - }, - "DeleteKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "DeleteNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Egress" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - } - } - }, - "DeleteNetworkAclRequest":{ - "type":"structure", - "required":["NetworkAclId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "DeleteNetworkInterfaceRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "DeletePlacementGroupRequest":{ - "type":"structure", - "required":["GroupName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "DeleteRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - } - } - }, - "DeleteRouteTableRequest":{ - "type":"structure", - "required":["RouteTableId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "DeleteSecurityGroupRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"} - } - }, - "DeleteSnapshotRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"} - } - }, - "DeleteSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DeleteSubnetRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{"shape":"String"} - } - }, - "DeleteTagsRequest":{ - "type":"structure", - "required":["Resources"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"resourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tag" - } - } - }, - "DeleteVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"} - } - }, - "DeleteVpcEndpointsRequest":{ - "type":"structure", - "required":["VpcEndpointIds"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - } - } - }, - "DeleteVpcEndpointsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "DeleteVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DeleteVpcRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRequest":{ - "type":"structure", - "required":["VpnConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "DeleteVpnGatewayRequest":{ - "type":"structure", - "required":["VpnGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"} - } - }, - "DeregisterImageRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"} - } - }, - "DescribeAccountAttributesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttributeNames":{ - "shape":"AccountAttributeNameStringList", - "locationName":"attributeName" - } - } - }, - "DescribeAccountAttributesResult":{ - "type":"structure", - "members":{ - "AccountAttributes":{ - "shape":"AccountAttributeList", - "locationName":"accountAttributeSet" - } - } - }, - "DescribeAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"PublicIpStringList", - "locationName":"PublicIp" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AllocationIds":{ - "shape":"AllocationIdList", - "locationName":"AllocationId" - } - } - }, - "DescribeAddressesResult":{ - "type":"structure", - "members":{ - "Addresses":{ - "shape":"AddressList", - "locationName":"addressesSet" - } - } - }, - "DescribeAvailabilityZonesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ZoneNames":{ - "shape":"ZoneNameStringList", - "locationName":"ZoneName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeAvailabilityZonesResult":{ - "type":"structure", - "members":{ - "AvailabilityZones":{ - "shape":"AvailabilityZoneList", - "locationName":"availabilityZoneInfo" - } - } - }, - "DescribeBundleTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleIds":{ - "shape":"BundleIdStringList", - "locationName":"BundleId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeBundleTasksResult":{ - "type":"structure", - "members":{ - "BundleTasks":{ - "shape":"BundleTaskList", - "locationName":"bundleInstanceTasksSet" - } - } - }, - "DescribeClassicLinkInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeClassicLinkInstancesResult":{ - "type":"structure", - "members":{ - "Instances":{ - "shape":"ClassicLinkInstanceList", - "locationName":"instancesSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeConversionTaskList":{ - "type":"list", - "member":{ - "shape":"ConversionTask", - "locationName":"item" - } - }, - "DescribeConversionTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "ConversionTaskIds":{ - "shape":"ConversionIdStringList", - "locationName":"conversionTaskId" - } - } - }, - "DescribeConversionTasksResult":{ - "type":"structure", - "members":{ - "ConversionTasks":{ - "shape":"DescribeConversionTaskList", - "locationName":"conversionTasks" - } - } - }, - "DescribeCustomerGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayIds":{ - "shape":"CustomerGatewayIdStringList", - "locationName":"CustomerGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeCustomerGatewaysResult":{ - "type":"structure", - "members":{ - "CustomerGateways":{ - "shape":"CustomerGatewayList", - "locationName":"customerGatewaySet" - } - } - }, - "DescribeDhcpOptionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsIds":{ - "shape":"DhcpOptionsIdStringList", - "locationName":"DhcpOptionsId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptionsList", - "locationName":"dhcpOptionsSet" - } - } - }, - "DescribeExportTasksRequest":{ - "type":"structure", - "members":{ - "ExportTaskIds":{ - "shape":"ExportTaskIdStringList", - "locationName":"exportTaskId" - } - } - }, - "DescribeExportTasksResult":{ - "type":"structure", - "members":{ - "ExportTasks":{ - "shape":"ExportTaskList", - "locationName":"exportTaskSet" - } - } - }, - "DescribeFlowLogsRequest":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - }, - "Filter":{"shape":"FilterList"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogs":{ - "shape":"FlowLogSet", - "locationName":"flowLogSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ImageAttributeName"} - } - }, - "DescribeImagesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageIds":{ - "shape":"ImageIdStringList", - "locationName":"ImageId" - }, - "Owners":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "ExecutableUsers":{ - "shape":"ExecutableByStringList", - "locationName":"ExecutableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeImagesResult":{ - "type":"structure", - "members":{ - "Images":{ - "shape":"ImageList", - "locationName":"imagesSet" - } - } - }, - "DescribeImportImageTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportImageTasksResult":{ - "type":"structure", - "members":{ - "ImportImageTasks":{ - "shape":"ImportImageTaskList", - "locationName":"importImageTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeImportSnapshotTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportSnapshotTasksResult":{ - "type":"structure", - "members":{ - "ImportSnapshotTasks":{ - "shape":"ImportSnapshotTaskList", - "locationName":"importSnapshotTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "DescribeInstanceStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "IncludeAllInstances":{ - "shape":"Boolean", - "locationName":"includeAllInstances" - } - } - }, - "DescribeInstanceStatusResult":{ - "type":"structure", - "members":{ - "InstanceStatuses":{ - "shape":"InstanceStatusList", - "locationName":"instanceStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeInstancesResult":{ - "type":"structure", - "members":{ - "Reservations":{ - "shape":"ReservationList", - "locationName":"reservationSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInternetGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayIds":{ - "shape":"ValueStringList", - "locationName":"internetGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeInternetGatewaysResult":{ - "type":"structure", - "members":{ - "InternetGateways":{ - "shape":"InternetGatewayList", - "locationName":"internetGatewaySet" - } - } - }, - "DescribeKeyPairsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyNames":{ - "shape":"KeyNameStringList", - "locationName":"KeyName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeKeyPairsResult":{ - "type":"structure", - "members":{ - "KeyPairs":{ - "shape":"KeyPairList", - "locationName":"keySet" - } - } - }, - "DescribeMovingAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"ValueStringList", - "locationName":"publicIp" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeMovingAddressesResult":{ - "type":"structure", - "members":{ - "MovingAddressStatuses":{ - "shape":"MovingAddressStatusSet", - "locationName":"movingAddressStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNetworkAclsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclIds":{ - "shape":"ValueStringList", - "locationName":"NetworkAclId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeNetworkAclsResult":{ - "type":"structure", - "members":{ - "NetworkAcls":{ - "shape":"NetworkAclList", - "locationName":"networkAclSet" - } - } - }, - "DescribeNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Attribute":{ - "shape":"NetworkInterfaceAttribute", - "locationName":"attribute" - } - } - }, - "DescribeNetworkInterfaceAttributeResult":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - } - } - }, - "DescribeNetworkInterfacesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceIds":{ - "shape":"NetworkInterfaceIdList", - "locationName":"NetworkInterfaceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeNetworkInterfacesResult":{ - "type":"structure", - "members":{ - "NetworkInterfaces":{ - "shape":"NetworkInterfaceList", - "locationName":"networkInterfaceSet" - } - } - }, - "DescribePlacementGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"PlacementGroupStringList", - "locationName":"groupName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribePlacementGroupsResult":{ - "type":"structure", - "members":{ - "PlacementGroups":{ - "shape":"PlacementGroupList", - "locationName":"placementGroupSet" - } - } - }, - "DescribePrefixListsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "PrefixListIds":{ - "shape":"ValueStringList", - "locationName":"PrefixListId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribePrefixListsResult":{ - "type":"structure", - "members":{ - "PrefixLists":{ - "shape":"PrefixListSet", - "locationName":"prefixListSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeRegionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RegionNames":{ - "shape":"RegionNameStringList", - "locationName":"RegionName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRegionsResult":{ - "type":"structure", - "members":{ - "Regions":{ - "shape":"RegionList", - "locationName":"regionInfo" - } - } - }, - "DescribeReservedInstancesListingsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filters" - } - } - }, - "DescribeReservedInstancesListingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "DescribeReservedInstancesModificationsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationIds":{ - "shape":"ReservedInstancesModificationIdStringList", - "locationName":"ReservedInstancesModificationId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeReservedInstancesModificationsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModifications":{ - "shape":"ReservedInstancesModificationList", - "locationName":"reservedInstancesModificationsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesOfferingsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingIds":{ - "shape":"ReservedInstancesOfferingIdStringList", - "locationName":"ReservedInstancesOfferingId" - }, - "InstanceType":{"shape":"InstanceType"}, - "AvailabilityZone":{"shape":"String"}, - "ProductDescription":{"shape":"RIProductDescription"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "IncludeMarketplace":{"shape":"Boolean"}, - "MinDuration":{"shape":"Long"}, - "MaxDuration":{"shape":"Long"}, - "MaxInstanceCount":{"shape":"Integer"} - } - }, - "DescribeReservedInstancesOfferingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferings":{ - "shape":"ReservedInstancesOfferingList", - "locationName":"reservedInstancesOfferingsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - } - } - }, - "DescribeReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstances":{ - "shape":"ReservedInstancesList", - "locationName":"reservedInstancesSet" - } - } - }, - "DescribeRouteTablesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRouteTablesResult":{ - "type":"structure", - "members":{ - "RouteTables":{ - "shape":"RouteTableList", - "locationName":"routeTableSet" - } - } - }, - "DescribeSecurityGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSecurityGroupsResult":{ - "type":"structure", - "members":{ - "SecurityGroups":{ - "shape":"SecurityGroupList", - "locationName":"securityGroupInfo" - } - } - }, - "DescribeSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "DescribeSnapshotAttributeResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "CreateVolumePermissions":{ - "shape":"CreateVolumePermissionList", - "locationName":"createVolumePermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeSnapshotsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotIds":{ - "shape":"SnapshotIdStringList", - "locationName":"SnapshotId" - }, - "OwnerIds":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "RestorableByUserIds":{ - "shape":"RestorableByStringList", - "locationName":"RestorableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeSnapshotsResult":{ - "type":"structure", - "members":{ - "Snapshots":{ - "shape":"SnapshotList", - "locationName":"snapshotSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DescribeSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "DescribeSpotFleetInstancesRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetInstancesResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "ActiveInstances" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "ActiveInstances":{ - "shape":"ActiveInstanceSet", - "locationName":"activeInstanceSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestHistoryRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestHistoryResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime", - "LastEvaluatedTime", - "HistoryRecords" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "LastEvaluatedTime":{ - "shape":"DateTime", - "locationName":"lastEvaluatedTime" - }, - "HistoryRecords":{ - "shape":"HistoryRecords", - "locationName":"historyRecordSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestsResponse":{ - "type":"structure", - "required":["SpotFleetRequestConfigs"], - "members":{ - "SpotFleetRequestConfigs":{ - "shape":"SpotFleetRequestConfigSet", - "locationName":"spotFleetRequestConfigSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotInstanceRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "DescribeSpotPriceHistoryRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "InstanceTypes":{ - "shape":"InstanceTypeList", - "locationName":"InstanceType" - }, - "ProductDescriptions":{ - "shape":"ProductDescriptionList", - "locationName":"ProductDescription" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotPriceHistoryResult":{ - "type":"structure", - "members":{ - "SpotPriceHistory":{ - "shape":"SpotPriceHistoryList", - "locationName":"spotPriceHistorySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSubnetsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetIds":{ - "shape":"SubnetIdStringList", - "locationName":"SubnetId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSubnetsResult":{ - "type":"structure", - "members":{ - "Subnets":{ - "shape":"SubnetList", - "locationName":"subnetSet" - } - } - }, - "DescribeTagsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeTagsResult":{ - "type":"structure", - "members":{ - "Tags":{ - "shape":"TagDescriptionList", - "locationName":"tagSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Attribute":{"shape":"VolumeAttributeName"} - } - }, - "DescribeVolumeAttributeResult":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AutoEnableIO":{ - "shape":"AttributeBooleanValue", - "locationName":"autoEnableIO" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeVolumeStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeVolumeStatusResult":{ - "type":"structure", - "members":{ - "VolumeStatuses":{ - "shape":"VolumeStatusList", - "locationName":"volumeStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeVolumesResult":{ - "type":"structure", - "members":{ - "Volumes":{ - "shape":"VolumeList", - "locationName":"volumeSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcAttributeRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "Attribute":{"shape":"VpcAttributeName"} - } - }, - "DescribeVpcAttributeResult":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsSupport" - }, - "EnableDnsHostnames":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsHostnames" - } - } - }, - "DescribeVpcClassicLinkRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcClassicLinkIdList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcClassicLinkList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpcEndpointServicesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointServicesResult":{ - "type":"structure", - "members":{ - "ServiceNames":{ - "shape":"ValueStringList", - "locationName":"serviceNameSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcEndpointsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointsResult":{ - "type":"structure", - "members":{ - "VpcEndpoints":{ - "shape":"VpcEndpointSet", - "locationName":"vpcEndpointSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcPeeringConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionIds":{ - "shape":"ValueStringList", - "locationName":"VpcPeeringConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcPeeringConnectionsResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnections":{ - "shape":"VpcPeeringConnectionList", - "locationName":"vpcPeeringConnectionSet" - } - } - }, - "DescribeVpcsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcIdStringList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcsResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpnConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionIds":{ - "shape":"VpnConnectionIdStringList", - "locationName":"VpnConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnConnectionsResult":{ - "type":"structure", - "members":{ - "VpnConnections":{ - "shape":"VpnConnectionList", - "locationName":"vpnConnectionSet" - } - } - }, - "DescribeVpnGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayIds":{ - "shape":"VpnGatewayIdStringList", - "locationName":"VpnGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnGatewaysResult":{ - "type":"structure", - "members":{ - "VpnGateways":{ - "shape":"VpnGatewayList", - "locationName":"vpnGatewaySet" - } - } - }, - "DetachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DetachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachNetworkInterfaceRequest":{ - "type":"structure", - "required":["AttachmentId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "DetachVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"}, - "Force":{"shape":"Boolean"} - } - }, - "DetachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "DeviceType":{ - "type":"string", - "enum":[ - "ebs", - "instance-store" - ] - }, - "DhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"DhcpConfigurationValueList", - "locationName":"valueSet" - } - } - }, - "DhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"DhcpConfiguration", - "locationName":"item" - } - }, - "DhcpOptions":{ - "type":"structure", - "members":{ - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "DhcpConfigurations":{ - "shape":"DhcpConfigurationList", - "locationName":"dhcpConfigurationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "DhcpOptionsIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DhcpOptionsId" - } - }, - "DhcpOptionsList":{ - "type":"list", - "member":{ - "shape":"DhcpOptions", - "locationName":"item" - } - }, - "DisableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DisableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisassociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AssociationId":{"shape":"String"} - } - }, - "DisassociateRouteTableRequest":{ - "type":"structure", - "required":["AssociationId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "DiskImage":{ - "type":"structure", - "members":{ - "Image":{"shape":"DiskImageDetail"}, - "Description":{"shape":"String"}, - "Volume":{"shape":"VolumeDetail"} - } - }, - "DiskImageDescription":{ - "type":"structure", - "required":[ - "Format", - "Size", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - }, - "Checksum":{ - "shape":"String", - "locationName":"checksum" - } - } - }, - "DiskImageDetail":{ - "type":"structure", - "required":[ - "Format", - "Bytes", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Bytes":{ - "shape":"Long", - "locationName":"bytes" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - } - } - }, - "DiskImageFormat":{ - "type":"string", - "enum":[ - "VMDK", - "RAW", - "VHD" - ] - }, - "DiskImageList":{ - "type":"list", - "member":{"shape":"DiskImage"} - }, - "DiskImageVolumeDescription":{ - "type":"structure", - "required":["Id"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "DomainType":{ - "type":"string", - "enum":[ - "vpc", - "standard" - ] - }, - "Double":{"type":"double"}, - "EbsBlockDevice":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - } - } - }, - "EbsInstanceBlockDevice":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EbsInstanceBlockDeviceSpecification":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EnableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "EnableVolumeIORequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - } - } - }, - "EnableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "EnableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EventCode":{ - "type":"string", - "enum":[ - "instance-reboot", - "system-reboot", - "system-maintenance", - "instance-retirement", - "instance-stop" - ] - }, - "EventInformation":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "EventSubType":{ - "shape":"String", - "locationName":"eventSubType" - }, - "EventDescription":{ - "shape":"String", - "locationName":"eventDescription" - } - } - }, - "EventType":{ - "type":"string", - "enum":[ - "instanceChange", - "fleetRequestChange", - "error" - ] - }, - "ExecutableByStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExecutableBy" - } - }, - "ExportEnvironment":{ - "type":"string", - "enum":[ - "citrix", - "vmware", - "microsoft" - ] - }, - "ExportTask":{ - "type":"structure", - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "State":{ - "shape":"ExportTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceExportDetails":{ - "shape":"InstanceExportDetails", - "locationName":"instanceExport" - }, - "ExportToS3Task":{ - "shape":"ExportToS3Task", - "locationName":"exportToS3" - } - } - }, - "ExportTaskIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExportTaskId" - } - }, - "ExportTaskList":{ - "type":"list", - "member":{ - "shape":"ExportTask", - "locationName":"item" - } - }, - "ExportTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "ExportToS3Task":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "ExportToS3TaskSpecification":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Prefix":{ - "shape":"String", - "locationName":"s3Prefix" - } - } - }, - "Filter":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "FilterList":{ - "type":"list", - "member":{ - "shape":"Filter", - "locationName":"Filter" - } - }, - "Float":{"type":"float"}, - "FlowLog":{ - "type":"structure", - "members":{ - "CreationTime":{ - "shape":"DateTime", - "locationName":"creationTime" - }, - "FlowLogId":{ - "shape":"String", - "locationName":"flowLogId" - }, - "FlowLogStatus":{ - "shape":"String", - "locationName":"flowLogStatus" - }, - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "TrafficType":{ - "shape":"TrafficType", - "locationName":"trafficType" - }, - "LogGroupName":{ - "shape":"String", - "locationName":"logGroupName" - }, - "DeliverLogsStatus":{ - "shape":"String", - "locationName":"deliverLogsStatus" - }, - "DeliverLogsErrorMessage":{ - "shape":"String", - "locationName":"deliverLogsErrorMessage" - }, - "DeliverLogsPermissionArn":{ - "shape":"String", - "locationName":"deliverLogsPermissionArn" - } - } - }, - "FlowLogSet":{ - "type":"list", - "member":{ - "shape":"FlowLog", - "locationName":"item" - } - }, - "FlowLogsResourceType":{ - "type":"string", - "enum":[ - "VPC", - "Subnet", - "NetworkInterface" - ] - }, - "GatewayType":{ - "type":"string", - "enum":["ipsec.1"] - }, - "GetConsoleOutputRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetConsoleOutputResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "Output":{ - "shape":"String", - "locationName":"output" - } - } - }, - "GetPasswordDataRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetPasswordDataResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "PasswordData":{ - "shape":"String", - "locationName":"passwordData" - } - } - }, - "GroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"groupId" - } - }, - "GroupIdentifier":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "GroupIdentifierList":{ - "type":"list", - "member":{ - "shape":"GroupIdentifier", - "locationName":"item" - } - }, - "GroupNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"GroupName" - } - }, - "HistoryRecord":{ - "type":"structure", - "required":[ - "Timestamp", - "EventType", - "EventInformation" - ], - "members":{ - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "EventInformation":{ - "shape":"EventInformation", - "locationName":"eventInformation" - } - } - }, - "HistoryRecords":{ - "type":"list", - "member":{ - "shape":"HistoryRecord", - "locationName":"item" - } - }, - "HypervisorType":{ - "type":"string", - "enum":[ - "ovm", - "xen" - ] - }, - "IamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "IamInstanceProfileSpecification":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Name":{ - "shape":"String", - "locationName":"name" - } - } - }, - "IcmpTypeCode":{ - "type":"structure", - "members":{ - "Type":{ - "shape":"Integer", - "locationName":"type" - }, - "Code":{ - "shape":"Integer", - "locationName":"code" - } - } - }, - "Image":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "ImageLocation":{ - "shape":"String", - "locationName":"imageLocation" - }, - "State":{ - "shape":"ImageState", - "locationName":"imageState" - }, - "OwnerId":{ - "shape":"String", - "locationName":"imageOwnerId" - }, - "CreationDate":{ - "shape":"String", - "locationName":"creationDate" - }, - "Public":{ - "shape":"Boolean", - "locationName":"isPublic" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "ImageType":{ - "shape":"ImageTypeValues", - "locationName":"imageType" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "ImageOwnerAlias":{ - "shape":"String", - "locationName":"imageOwnerAlias" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - } - } - }, - "ImageAttribute":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "LaunchPermissions":{ - "shape":"LaunchPermissionList", - "locationName":"launchPermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - } - } - }, - "ImageAttributeName":{ - "type":"string", - "enum":[ - "description", - "kernel", - "ramdisk", - "launchPermission", - "productCodes", - "blockDeviceMapping", - "sriovNetSupport" - ] - }, - "ImageDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"}, - "DeviceName":{"shape":"String"}, - "SnapshotId":{"shape":"String"} - } - }, - "ImageDiskContainerList":{ - "type":"list", - "member":{ - "shape":"ImageDiskContainer", - "locationName":"item" - } - }, - "ImageIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImageId" - } - }, - "ImageList":{ - "type":"list", - "member":{ - "shape":"Image", - "locationName":"item" - } - }, - "ImageState":{ - "type":"string", - "enum":[ - "pending", - "available", - "invalid", - "deregistered", - "transient", - "failed", - "error" - ] - }, - "ImageTypeValues":{ - "type":"string", - "enum":[ - "machine", - "kernel", - "ramdisk" - ] - }, - "ImportImageRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainers":{ - "shape":"ImageDiskContainerList", - "locationName":"DiskContainer" - }, - "LicenseType":{"shape":"String"}, - "Hypervisor":{"shape":"String"}, - "Architecture":{"shape":"String"}, - "Platform":{"shape":"String"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportImageResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTaskList":{ - "type":"list", - "member":{ - "shape":"ImportImageTask", - "locationName":"item" - } - }, - "ImportInstanceLaunchSpecification":{ - "type":"structure", - "members":{ - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "GroupNames":{ - "shape":"SecurityGroupStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"GroupId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "UserData":{ - "shape":"UserData", - "locationName":"userData" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "Monitoring":{ - "shape":"Boolean", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "ImportInstanceRequest":{ - "type":"structure", - "required":["Platform"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "LaunchSpecification":{ - "shape":"ImportInstanceLaunchSpecification", - "locationName":"launchSpecification" - }, - "DiskImages":{ - "shape":"DiskImageList", - "locationName":"diskImage" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - } - } - }, - "ImportInstanceResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportInstanceTaskDetails":{ - "type":"structure", - "required":["Volumes"], - "members":{ - "Volumes":{ - "shape":"ImportInstanceVolumeDetailSet", - "locationName":"volumes" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailItem":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume", - "Status" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailSet":{ - "type":"list", - "member":{ - "shape":"ImportInstanceVolumeDetailItem", - "locationName":"item" - } - }, - "ImportKeyPairRequest":{ - "type":"structure", - "required":[ - "KeyName", - "PublicKeyMaterial" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "PublicKeyMaterial":{ - "shape":"Blob", - "locationName":"publicKeyMaterial" - } - } - }, - "ImportKeyPairResult":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "ImportSnapshotRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainer":{"shape":"SnapshotDiskContainer"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportSnapshotResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTaskList":{ - "type":"list", - "member":{ - "shape":"ImportSnapshotTask", - "locationName":"item" - } - }, - "ImportTaskIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImportTaskId" - } - }, - "ImportVolumeRequest":{ - "type":"structure", - "required":[ - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDetail", - "locationName":"image" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Volume":{ - "shape":"VolumeDetail", - "locationName":"volume" - } - } - }, - "ImportVolumeResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportVolumeTaskDetails":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - } - } - }, - "Instance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "State":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"dnsName" - }, - "StateTransitionReason":{ - "shape":"String", - "locationName":"reason" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "AmiLaunchIndex":{ - "shape":"Integer", - "locationName":"amiLaunchIndex" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "LaunchTime":{ - "shape":"DateTime", - "locationName":"launchTime" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PublicIpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "InstanceLifecycle":{ - "shape":"InstanceLifecycleType", - "locationName":"instanceLifecycle" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfile", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - } - } - }, - "InstanceAttribute":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"AttributeValue", - "locationName":"userData" - }, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "RootDeviceName":{ - "shape":"AttributeValue", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - } - } - }, - "InstanceAttributeName":{ - "type":"string", - "enum":[ - "instanceType", - "kernel", - "ramdisk", - "userData", - "disableApiTermination", - "instanceInitiatedShutdownBehavior", - "rootDeviceName", - "blockDeviceMapping", - "productCodes", - "sourceDestCheck", - "groupSet", - "ebsOptimized", - "sriovNetSupport" - ] - }, - "InstanceBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDevice", - "locationName":"ebs" - } - } - }, - "InstanceBlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMapping", - "locationName":"item" - } - }, - "InstanceBlockDeviceMappingSpecification":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDeviceSpecification", - "locationName":"ebs" - }, - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "InstanceBlockDeviceMappingSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMappingSpecification", - "locationName":"item" - } - }, - "InstanceCount":{ - "type":"structure", - "members":{ - "State":{ - "shape":"ListingState", - "locationName":"state" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - } - } - }, - "InstanceCountList":{ - "type":"list", - "member":{ - "shape":"InstanceCount", - "locationName":"item" - } - }, - "InstanceExportDetails":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - } - } - }, - "InstanceIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"InstanceId" - } - }, - "InstanceLifecycleType":{ - "type":"string", - "enum":["spot"] - }, - "InstanceList":{ - "type":"list", - "member":{ - "shape":"Instance", - "locationName":"item" - } - }, - "InstanceMonitoring":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - } - } - }, - "InstanceMonitoringList":{ - "type":"list", - "member":{ - "shape":"InstanceMonitoring", - "locationName":"item" - } - }, - "InstanceNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"InstanceNetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - }, - "PrivateIpAddresses":{ - "shape":"InstancePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - } - } - }, - "InstanceNetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - } - } - }, - "InstanceNetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "InstanceNetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterface", - "locationName":"item" - } - }, - "InstanceNetworkInterfaceSpecification":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddressesSet", - "queryName":"PrivateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AssociatePublicIpAddress":{ - "shape":"Boolean", - "locationName":"associatePublicIpAddress" - } - } - }, - "InstanceNetworkInterfaceSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterfaceSpecification", - "locationName":"item" - } - }, - "InstancePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "InstancePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"InstancePrivateIpAddress", - "locationName":"item" - } - }, - "InstanceState":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"Integer", - "locationName":"code" - }, - "Name":{ - "shape":"InstanceStateName", - "locationName":"name" - } - } - }, - "InstanceStateChange":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CurrentState":{ - "shape":"InstanceState", - "locationName":"currentState" - }, - "PreviousState":{ - "shape":"InstanceState", - "locationName":"previousState" - } - } - }, - "InstanceStateChangeList":{ - "type":"list", - "member":{ - "shape":"InstanceStateChange", - "locationName":"item" - } - }, - "InstanceStateName":{ - "type":"string", - "enum":[ - "pending", - "running", - "shutting-down", - "terminated", - "stopping", - "stopped" - ] - }, - "InstanceStatus":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Events":{ - "shape":"InstanceStatusEventList", - "locationName":"eventsSet" - }, - "InstanceState":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "SystemStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"systemStatus" - }, - "InstanceStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"instanceStatus" - } - } - }, - "InstanceStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"StatusName", - "locationName":"name" - }, - "Status":{ - "shape":"StatusType", - "locationName":"status" - }, - "ImpairedSince":{ - "shape":"DateTime", - "locationName":"impairedSince" - } - } - }, - "InstanceStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusDetails", - "locationName":"item" - } - }, - "InstanceStatusEvent":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"EventCode", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - } - } - }, - "InstanceStatusEventList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusEvent", - "locationName":"item" - } - }, - "InstanceStatusList":{ - "type":"list", - "member":{ - "shape":"InstanceStatus", - "locationName":"item" - } - }, - "InstanceStatusSummary":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"SummaryStatus", - "locationName":"status" - }, - "Details":{ - "shape":"InstanceStatusDetailsList", - "locationName":"details" - } - } - }, - "InstanceType":{ - "type":"string", - "enum":[ - "t1.micro", - "m1.small", - "m1.medium", - "m1.large", - "m1.xlarge", - "m3.medium", - "m3.large", - "m3.xlarge", - "m3.2xlarge", - "m4.large", - "m4.xlarge", - "m4.2xlarge", - "m4.4xlarge", - "m4.10xlarge", - "t2.micro", - "t2.small", - "t2.medium", - "t2.large", - "m2.xlarge", - "m2.2xlarge", - "m2.4xlarge", - "cr1.8xlarge", - "i2.xlarge", - "i2.2xlarge", - "i2.4xlarge", - "i2.8xlarge", - "hi1.4xlarge", - "hs1.8xlarge", - "c1.medium", - "c1.xlarge", - "c3.large", - "c3.xlarge", - "c3.2xlarge", - "c3.4xlarge", - "c3.8xlarge", - "c4.large", - "c4.xlarge", - "c4.2xlarge", - "c4.4xlarge", - "c4.8xlarge", - "cc1.4xlarge", - "cc2.8xlarge", - "g2.2xlarge", - "cg1.4xlarge", - "r3.large", - "r3.xlarge", - "r3.2xlarge", - "r3.4xlarge", - "r3.8xlarge", - "d2.xlarge", - "d2.2xlarge", - "d2.4xlarge", - "d2.8xlarge" - ] - }, - "InstanceTypeList":{ - "type":"list", - "member":{"shape":"InstanceType"} - }, - "Integer":{"type":"integer"}, - "InternetGateway":{ - "type":"structure", - "members":{ - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "Attachments":{ - "shape":"InternetGatewayAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "InternetGatewayAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "InternetGatewayAttachmentList":{ - "type":"list", - "member":{ - "shape":"InternetGatewayAttachment", - "locationName":"item" - } - }, - "InternetGatewayList":{ - "type":"list", - "member":{ - "shape":"InternetGateway", - "locationName":"item" - } - }, - "IpPermission":{ - "type":"structure", - "members":{ - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairList", - "locationName":"groups" - }, - "IpRanges":{ - "shape":"IpRangeList", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdList", - "locationName":"prefixListIds" - } - } - }, - "IpPermissionList":{ - "type":"list", - "member":{ - "shape":"IpPermission", - "locationName":"item" - } - }, - "IpRange":{ - "type":"structure", - "members":{ - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - } - } - }, - "IpRangeList":{ - "type":"list", - "member":{ - "shape":"IpRange", - "locationName":"item" - } - }, - "KeyNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"KeyName" - } - }, - "KeyPair":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - }, - "KeyMaterial":{ - "shape":"String", - "locationName":"keyMaterial" - } - } - }, - "KeyPairInfo":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "KeyPairList":{ - "type":"list", - "member":{ - "shape":"KeyPairInfo", - "locationName":"item" - } - }, - "LaunchPermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "LaunchPermissionList":{ - "type":"list", - "member":{ - "shape":"LaunchPermission", - "locationName":"item" - } - }, - "LaunchPermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"LaunchPermissionList"}, - "Remove":{"shape":"LaunchPermissionList"} - } - }, - "LaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - } - } - }, - "LaunchSpecsList":{ - "type":"list", - "member":{ - "shape":"SpotFleetLaunchSpecification", - "locationName":"item" - }, - "min":1 - }, - "ListingState":{ - "type":"string", - "enum":[ - "available", - "sold", - "cancelled", - "pending" - ] - }, - "ListingStatus":{ - "type":"string", - "enum":[ - "active", - "pending", - "cancelled", - "closed" - ] - }, - "Long":{"type":"long"}, - "ModifyImageAttributeRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"String"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "UserGroups":{ - "shape":"UserGroupStringList", - "locationName":"UserGroup" - }, - "ProductCodes":{ - "shape":"ProductCodeStringList", - "locationName":"ProductCode" - }, - "Value":{"shape":"String"}, - "LaunchPermission":{"shape":"LaunchPermissionModifications"}, - "Description":{"shape":"AttributeValue"} - } - }, - "ModifyInstanceAttributeRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - }, - "Value":{ - "shape":"String", - "locationName":"value" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingSpecificationList", - "locationName":"blockDeviceMapping" - }, - "SourceDestCheck":{"shape":"AttributeBooleanValue"}, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "Kernel":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "Ramdisk":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"BlobAttributeValue", - "locationName":"userData" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - } - } - }, - "ModifyNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachmentChanges", - "locationName":"attachment" - } - } - }, - "ModifyReservedInstancesRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesIds", - "TargetConfigurations" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "TargetConfigurations":{ - "shape":"ReservedInstancesConfigurationList", - "locationName":"ReservedInstancesConfigurationSetItemType" - } - } - }, - "ModifyReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - } - } - }, - "ModifySnapshotAttributeRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"UserGroup" - }, - "CreateVolumePermission":{"shape":"CreateVolumePermissionModifications"} - } - }, - "ModifySubnetAttributeRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "MapPublicIpOnLaunch":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "AutoEnableIO":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcAttributeRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{"shape":"AttributeBooleanValue"}, - "EnableDnsHostnames":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcEndpointRequest":{ - "type":"structure", - "required":["VpcEndpointId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointId":{"shape":"String"}, - "ResetPolicy":{"shape":"Boolean"}, - "PolicyDocument":{"shape":"String"}, - "AddRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"AddRouteTableId" - }, - "RemoveRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RemoveRouteTableId" - } - } - }, - "ModifyVpcEndpointResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "MonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "MonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "Monitoring":{ - "type":"structure", - "members":{ - "State":{ - "shape":"MonitoringState", - "locationName":"state" - } - } - }, - "MonitoringState":{ - "type":"string", - "enum":[ - "disabled", - "disabling", - "enabled", - "pending" - ] - }, - "MoveAddressToVpcRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "MoveAddressToVpcResult":{ - "type":"structure", - "members":{ - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "Status":{ - "shape":"Status", - "locationName":"status" - } - } - }, - "MoveStatus":{ - "type":"string", - "enum":[ - "movingToVpc", - "restoringToClassic" - ] - }, - "MovingAddressStatus":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "MoveStatus":{ - "shape":"MoveStatus", - "locationName":"moveStatus" - } - } - }, - "MovingAddressStatusSet":{ - "type":"list", - "member":{ - "shape":"MovingAddressStatus", - "locationName":"item" - } - }, - "NetworkAcl":{ - "type":"structure", - "members":{ - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"default" - }, - "Entries":{ - "shape":"NetworkAclEntryList", - "locationName":"entrySet" - }, - "Associations":{ - "shape":"NetworkAclAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "NetworkAclAssociation":{ - "type":"structure", - "members":{ - "NetworkAclAssociationId":{ - "shape":"String", - "locationName":"networkAclAssociationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - } - } - }, - "NetworkAclAssociationList":{ - "type":"list", - "member":{ - "shape":"NetworkAclAssociation", - "locationName":"item" - } - }, - "NetworkAclEntry":{ - "type":"structure", - "members":{ - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"icmpTypeCode" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "NetworkAclEntryList":{ - "type":"list", - "member":{ - "shape":"NetworkAclEntry", - "locationName":"item" - } - }, - "NetworkAclList":{ - "type":"list", - "member":{ - "shape":"NetworkAcl", - "locationName":"item" - } - }, - "NetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "RequesterManaged":{ - "shape":"Boolean", - "locationName":"requesterManaged" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - }, - "TagSet":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PrivateIpAddresses":{ - "shape":"NetworkInterfacePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - } - } - }, - "NetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "NetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttachmentChanges":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttribute":{ - "type":"string", - "enum":[ - "description", - "groupSet", - "sourceDestCheck", - "attachment" - ] - }, - "NetworkInterfaceIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "NetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"NetworkInterface", - "locationName":"item" - } - }, - "NetworkInterfacePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "NetworkInterfacePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"NetworkInterfacePrivateIpAddress", - "locationName":"item" - } - }, - "NetworkInterfaceStatus":{ - "type":"string", - "enum":[ - "available", - "attaching", - "in-use", - "detaching" - ] - }, - "OfferingTypeValues":{ - "type":"string", - "enum":[ - "Heavy Utilization", - "Medium Utilization", - "Light Utilization", - "No Upfront", - "Partial Upfront", - "All Upfront" - ] - }, - "OperationType":{ - "type":"string", - "enum":[ - "add", - "remove" - ] - }, - "OwnerStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"Owner" - } - }, - "PermissionGroup":{ - "type":"string", - "enum":["all"] - }, - "Placement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Tenancy":{ - "shape":"Tenancy", - "locationName":"tenancy" - } - } - }, - "PlacementGroup":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - }, - "State":{ - "shape":"PlacementGroupState", - "locationName":"state" - } - } - }, - "PlacementGroupList":{ - "type":"list", - "member":{ - "shape":"PlacementGroup", - "locationName":"item" - } - }, - "PlacementGroupState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "PlacementGroupStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PlacementStrategy":{ - "type":"string", - "enum":["cluster"] - }, - "PlatformValues":{ - "type":"string", - "enum":["Windows"] - }, - "PortRange":{ - "type":"structure", - "members":{ - "From":{ - "shape":"Integer", - "locationName":"from" - }, - "To":{ - "shape":"Integer", - "locationName":"to" - } - } - }, - "PrefixList":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - }, - "PrefixListName":{ - "shape":"String", - "locationName":"prefixListName" - }, - "Cidrs":{ - "shape":"ValueStringList", - "locationName":"cidrSet" - } - } - }, - "PrefixListId":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - } - } - }, - "PrefixListIdList":{ - "type":"list", - "member":{ - "shape":"PrefixListId", - "locationName":"item" - } - }, - "PrefixListSet":{ - "type":"list", - "member":{ - "shape":"PrefixList", - "locationName":"item" - } - }, - "PriceSchedule":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Active":{ - "shape":"Boolean", - "locationName":"active" - } - } - }, - "PriceScheduleList":{ - "type":"list", - "member":{ - "shape":"PriceSchedule", - "locationName":"item" - } - }, - "PriceScheduleSpecification":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "PriceScheduleSpecificationList":{ - "type":"list", - "member":{ - "shape":"PriceScheduleSpecification", - "locationName":"item" - } - }, - "PricingDetail":{ - "type":"structure", - "members":{ - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "Count":{ - "shape":"Integer", - "locationName":"count" - } - } - }, - "PricingDetailsList":{ - "type":"list", - "member":{ - "shape":"PricingDetail", - "locationName":"item" - } - }, - "PrivateIpAddressSpecification":{ - "type":"structure", - "required":["PrivateIpAddress"], - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - } - } - }, - "PrivateIpAddressSpecificationList":{ - "type":"list", - "member":{ - "shape":"PrivateIpAddressSpecification", - "locationName":"item" - } - }, - "PrivateIpAddressStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PrivateIpAddress" - } - }, - "ProductCode":{ - "type":"structure", - "members":{ - "ProductCodeId":{ - "shape":"String", - "locationName":"productCode" - }, - "ProductCodeType":{ - "shape":"ProductCodeValues", - "locationName":"type" - } - } - }, - "ProductCodeList":{ - "type":"list", - "member":{ - "shape":"ProductCode", - "locationName":"item" - } - }, - "ProductCodeStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ProductCode" - } - }, - "ProductCodeValues":{ - "type":"string", - "enum":[ - "devpay", - "marketplace" - ] - }, - "ProductDescriptionList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PropagatingVgw":{ - "type":"structure", - "members":{ - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - } - } - }, - "PropagatingVgwList":{ - "type":"list", - "member":{ - "shape":"PropagatingVgw", - "locationName":"item" - } - }, - "PublicIpStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PublicIp" - } - }, - "PurchaseReservedInstancesOfferingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesOfferingId", - "InstanceCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingId":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"}, - "LimitPrice":{ - "shape":"ReservedInstanceLimitPrice", - "locationName":"limitPrice" - } - } - }, - "PurchaseReservedInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "RIProductDescription":{ - "type":"string", - "enum":[ - "Linux/UNIX", - "Linux/UNIX (Amazon VPC)", - "Windows", - "Windows (Amazon VPC)" - ] - }, - "ReasonCodesList":{ - "type":"list", - "member":{ - "shape":"ReportInstanceReasonCodes", - "locationName":"item" - } - }, - "RebootInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"RecurringChargeFrequency", - "locationName":"frequency" - }, - "Amount":{ - "shape":"Double", - "locationName":"amount" - } - } - }, - "RecurringChargeFrequency":{ - "type":"string", - "enum":["Hourly"] - }, - "RecurringChargesList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"item" - } - }, - "Region":{ - "type":"structure", - "members":{ - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Endpoint":{ - "shape":"String", - "locationName":"regionEndpoint" - } - } - }, - "RegionList":{ - "type":"list", - "member":{ - "shape":"Region", - "locationName":"item" - } - }, - "RegionNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"RegionName" - } - }, - "RegisterImageRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageLocation":{"shape":"String"}, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"String", - "locationName":"virtualizationType" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - } - } - }, - "RegisterImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "RejectVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "RejectVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ReleaseAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"} - } - }, - "ReplaceNetworkAclAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "NetworkAclId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "ReplaceNetworkAclAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReplaceNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "ReplaceRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "ReplaceRouteTableAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "ReplaceRouteTableAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReportInstanceReasonCodes":{ - "type":"string", - "enum":[ - "instance-stuck-in-state", - "unresponsive", - "not-accepting-credentials", - "password-not-available", - "performance-network", - "performance-instance-store", - "performance-ebs-volume", - "performance-other", - "other" - ] - }, - "ReportInstanceStatusRequest":{ - "type":"structure", - "required":[ - "Instances", - "Status", - "ReasonCodes" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Instances":{ - "shape":"InstanceIdStringList", - "locationName":"instanceId" - }, - "Status":{ - "shape":"ReportStatusType", - "locationName":"status" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "ReasonCodes":{ - "shape":"ReasonCodesList", - "locationName":"reasonCode" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ReportStatusType":{ - "type":"string", - "enum":[ - "ok", - "impaired" - ] - }, - "RequestSpotFleetRequest":{ - "type":"structure", - "required":["SpotFleetRequestConfig"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - } - } - }, - "RequestSpotFleetResponse":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - } - } - }, - "RequestSpotInstancesRequest":{ - "type":"structure", - "required":["SpotPrice"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "LaunchSpecification":{"shape":"RequestSpotLaunchSpecification"} - } - }, - "RequestSpotInstancesResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "Reservation":{ - "type":"structure", - "members":{ - "ReservationId":{ - "shape":"String", - "locationName":"reservationId" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Instances":{ - "shape":"InstanceList", - "locationName":"instancesSet" - } - } - }, - "ReservationList":{ - "type":"list", - "member":{ - "shape":"Reservation", - "locationName":"item" - } - }, - "ReservedInstanceLimitPrice":{ - "type":"structure", - "members":{ - "Amount":{ - "shape":"Double", - "locationName":"amount" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "ReservedInstanceState":{ - "type":"string", - "enum":[ - "payment-pending", - "active", - "payment-failed", - "retired" - ] - }, - "ReservedInstances":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Start":{ - "shape":"DateTime", - "locationName":"start" - }, - "End":{ - "shape":"DateTime", - "locationName":"end" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "State":{ - "shape":"ReservedInstanceState", - "locationName":"state" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - } - } - }, - "ReservedInstancesConfiguration":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - } - } - }, - "ReservedInstancesConfigurationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"item" - } - }, - "ReservedInstancesId":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "ReservedInstancesIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesId" - } - }, - "ReservedInstancesList":{ - "type":"list", - "member":{ - "shape":"ReservedInstances", - "locationName":"item" - } - }, - "ReservedInstancesListing":{ - "type":"structure", - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "Status":{ - "shape":"ListingStatus", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceCounts":{ - "shape":"InstanceCountList", - "locationName":"instanceCounts" - }, - "PriceSchedules":{ - "shape":"PriceScheduleList", - "locationName":"priceSchedules" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesListingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesListing", - "locationName":"item" - } - }, - "ReservedInstancesModification":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - }, - "ReservedInstancesIds":{ - "shape":"ReservedIntancesIds", - "locationName":"reservedInstancesSet" - }, - "ModificationResults":{ - "shape":"ReservedInstancesModificationResultList", - "locationName":"modificationResultSet" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "EffectiveDate":{ - "shape":"DateTime", - "locationName":"effectiveDate" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesModificationIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesModificationId" - } - }, - "ReservedInstancesModificationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModification", - "locationName":"item" - } - }, - "ReservedInstancesModificationResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "TargetConfiguration":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"targetConfiguration" - } - } - }, - "ReservedInstancesModificationResultList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModificationResult", - "locationName":"item" - } - }, - "ReservedInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferingId":{ - "shape":"String", - "locationName":"reservedInstancesOfferingId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - }, - "Marketplace":{ - "shape":"Boolean", - "locationName":"marketplace" - }, - "PricingDetails":{ - "shape":"PricingDetailsList", - "locationName":"pricingDetailsSet" - } - } - }, - "ReservedInstancesOfferingIdStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ReservedInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesOffering", - "locationName":"item" - } - }, - "ReservedIntancesIds":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesId", - "locationName":"item" - } - }, - "ResetImageAttributeName":{ - "type":"string", - "enum":["launchPermission"] - }, - "ResetImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ResetImageAttributeName"} - } - }, - "ResetInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "ResetNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SourceDestCheck":{ - "shape":"String", - "locationName":"sourceDestCheck" - } - } - }, - "ResetSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "ResourceIdList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ResourceType":{ - "type":"string", - "enum":[ - "customer-gateway", - "dhcp-options", - "image", - "instance", - "internet-gateway", - "network-acl", - "network-interface", - "reserved-instances", - "route-table", - "snapshot", - "spot-instances-request", - "subnet", - "security-group", - "volume", - "vpc", - "vpn-connection", - "vpn-gateway" - ] - }, - "RestorableByStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "RestoreAddressToClassicRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RestoreAddressToClassicResult":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"Status", - "locationName":"status" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RevokeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "RevokeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "Route":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "DestinationPrefixListId":{ - "shape":"String", - "locationName":"destinationPrefixListId" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "State":{ - "shape":"RouteState", - "locationName":"state" - }, - "Origin":{ - "shape":"RouteOrigin", - "locationName":"origin" - } - } - }, - "RouteList":{ - "type":"list", - "member":{ - "shape":"Route", - "locationName":"item" - } - }, - "RouteOrigin":{ - "type":"string", - "enum":[ - "CreateRouteTable", - "CreateRoute", - "EnableVgwRoutePropagation" - ] - }, - "RouteState":{ - "type":"string", - "enum":[ - "active", - "blackhole" - ] - }, - "RouteTable":{ - "type":"structure", - "members":{ - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Routes":{ - "shape":"RouteList", - "locationName":"routeSet" - }, - "Associations":{ - "shape":"RouteTableAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PropagatingVgws":{ - "shape":"PropagatingVgwList", - "locationName":"propagatingVgwSet" - } - } - }, - "RouteTableAssociation":{ - "type":"structure", - "members":{ - "RouteTableAssociationId":{ - "shape":"String", - "locationName":"routeTableAssociationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Main":{ - "shape":"Boolean", - "locationName":"main" - } - } - }, - "RouteTableAssociationList":{ - "type":"list", - "member":{ - "shape":"RouteTableAssociation", - "locationName":"item" - } - }, - "RouteTableList":{ - "type":"list", - "member":{ - "shape":"RouteTable", - "locationName":"item" - } - }, - "RuleAction":{ - "type":"string", - "enum":[ - "allow", - "deny" - ] - }, - "RunInstancesMonitoringEnabled":{ - "type":"structure", - "required":["Enabled"], - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "RunInstancesRequest":{ - "type":"structure", - "required":[ - "ImageId", - "MinCount", - "MaxCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "MinCount":{"shape":"Integer"}, - "MaxCount":{"shape":"Integer"}, - "KeyName":{"shape":"String"}, - "SecurityGroups":{ - "shape":"SecurityGroupStringList", - "locationName":"SecurityGroup" - }, - "SecurityGroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "InstanceType":{"shape":"InstanceType"}, - "Placement":{"shape":"Placement"}, - "KernelId":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"RunInstancesMonitoringEnabled"}, - "SubnetId":{"shape":"String"}, - "DisableApiTermination":{ - "shape":"Boolean", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - } - } - }, - "S3Storage":{ - "type":"structure", - "members":{ - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "AWSAccessKeyId":{"shape":"String"}, - "UploadPolicy":{ - "shape":"Blob", - "locationName":"uploadPolicy" - }, - "UploadPolicySignature":{ - "shape":"String", - "locationName":"uploadPolicySignature" - } - } - }, - "SecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "Description":{ - "shape":"String", - "locationName":"groupDescription" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - }, - "IpPermissionsEgress":{ - "shape":"IpPermissionList", - "locationName":"ipPermissionsEgress" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SecurityGroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"SecurityGroup", - "locationName":"item" - } - }, - "SecurityGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroup" - } - }, - "ShutdownBehavior":{ - "type":"string", - "enum":[ - "stop", - "terminate" - ] - }, - "Snapshot":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "State":{ - "shape":"SnapshotState", - "locationName":"status" - }, - "StateMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "OwnerAlias":{ - "shape":"String", - "locationName":"ownerAlias" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - }, - "DataEncryptionKeyId":{ - "shape":"String", - "locationName":"dataEncryptionKeyId" - } - } - }, - "SnapshotAttributeName":{ - "type":"string", - "enum":[ - "productCodes", - "createVolumePermission" - ] - }, - "SnapshotDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SnapshotDetailList":{ - "type":"list", - "member":{ - "shape":"SnapshotDetail", - "locationName":"item" - } - }, - "SnapshotDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"} - } - }, - "SnapshotIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SnapshotId" - } - }, - "SnapshotList":{ - "type":"list", - "member":{ - "shape":"Snapshot", - "locationName":"item" - } - }, - "SnapshotState":{ - "type":"string", - "enum":[ - "pending", - "completed", - "error" - ] - }, - "SnapshotTaskDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SpotDatafeedSubscription":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "State":{ - "shape":"DatafeedSubscriptionState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - } - } - }, - "SpotFleetLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "Monitoring":{ - "shape":"SpotFleetMonitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "WeightedCapacity":{ - "shape":"Double", - "locationName":"weightedCapacity" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - } - } - }, - "SpotFleetMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "SpotFleetRequestConfig":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "SpotFleetRequestState", - "SpotFleetRequestConfig" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "SpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"spotFleetRequestState" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - } - } - }, - "SpotFleetRequestConfigData":{ - "type":"structure", - "required":[ - "SpotPrice", - "TargetCapacity", - "IamFleetRole", - "LaunchSpecifications" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "TerminateInstancesWithExpiration":{ - "shape":"Boolean", - "locationName":"terminateInstancesWithExpiration" - }, - "IamFleetRole":{ - "shape":"String", - "locationName":"iamFleetRole" - }, - "LaunchSpecifications":{ - "shape":"LaunchSpecsList", - "locationName":"launchSpecifications" - }, - "AllocationStrategy":{ - "shape":"AllocationStrategy", - "locationName":"allocationStrategy" - } - } - }, - "SpotFleetRequestConfigSet":{ - "type":"list", - "member":{ - "shape":"SpotFleetRequestConfig", - "locationName":"item" - } - }, - "SpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "State":{ - "shape":"SpotInstanceState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - }, - "Status":{ - "shape":"SpotInstanceStatus", - "locationName":"status" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "LaunchSpecification":{ - "shape":"LaunchSpecification", - "locationName":"launchSpecification" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "LaunchedAvailabilityZone":{ - "shape":"String", - "locationName":"launchedAvailabilityZone" - } - } - }, - "SpotInstanceRequestIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SpotInstanceRequestId" - } - }, - "SpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"SpotInstanceRequest", - "locationName":"item" - } - }, - "SpotInstanceState":{ - "type":"string", - "enum":[ - "open", - "active", - "closed", - "cancelled", - "failed" - ] - }, - "SpotInstanceStateFault":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceStatus":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceType":{ - "type":"string", - "enum":[ - "one-time", - "persistent" - ] - }, - "SpotPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "SpotPrice":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "SpotPriceHistoryList":{ - "type":"list", - "member":{ - "shape":"SpotPrice", - "locationName":"item" - } - }, - "StartInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "StartInstancesResult":{ - "type":"structure", - "members":{ - "StartingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "State":{ - "type":"string", - "enum":[ - "Pending", - "Available", - "Deleting", - "Deleted" - ] - }, - "StateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "Status":{ - "type":"string", - "enum":[ - "MoveInProgress", - "InVpc", - "InClassic" - ] - }, - "StatusName":{ - "type":"string", - "enum":["reachability"] - }, - "StatusType":{ - "type":"string", - "enum":[ - "passed", - "failed", - "insufficient-data", - "initializing" - ] - }, - "StopInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "StopInstancesResult":{ - "type":"structure", - "members":{ - "StoppingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "Storage":{ - "type":"structure", - "members":{ - "S3":{"shape":"S3Storage"} - } - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "State":{ - "shape":"SubnetState", - "locationName":"state" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "AvailableIpAddressCount":{ - "shape":"Integer", - "locationName":"availableIpAddressCount" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "DefaultForAz":{ - "shape":"Boolean", - "locationName":"defaultForAz" - }, - "MapPublicIpOnLaunch":{ - "shape":"Boolean", - "locationName":"mapPublicIpOnLaunch" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SubnetIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetId" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"item" - } - }, - "SubnetState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "SummaryStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data", - "not-applicable", - "initializing" - ] - }, - "Tag":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescription":{ - "type":"structure", - "members":{ - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "ResourceType":{ - "shape":"ResourceType", - "locationName":"resourceType" - }, - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescriptionList":{ - "type":"list", - "member":{ - "shape":"TagDescription", - "locationName":"item" - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"item" - } - }, - "TelemetryStatus":{ - "type":"string", - "enum":[ - "UP", - "DOWN" - ] - }, - "Tenancy":{ - "type":"string", - "enum":[ - "default", - "dedicated" - ] - }, - "TerminateInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "TerminateInstancesResult":{ - "type":"structure", - "members":{ - "TerminatingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "TrafficType":{ - "type":"string", - "enum":[ - "ACCEPT", - "REJECT", - "ALL" - ] - }, - "UnassignPrivateIpAddressesRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "PrivateIpAddresses" - ], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - } - } - }, - "UnmonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "UnmonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "UnsuccessfulItem":{ - "type":"structure", - "required":["Error"], - "members":{ - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "Error":{ - "shape":"UnsuccessfulItemError", - "locationName":"error" - } - } - }, - "UnsuccessfulItemError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "UnsuccessfulItemSet":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UserBucket":{ - "type":"structure", - "members":{ - "S3Bucket":{"shape":"String"}, - "S3Key":{"shape":"String"} - } - }, - "UserBucketDetails":{ - "type":"structure", - "members":{ - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "UserData":{ - "type":"structure", - "members":{ - "Data":{ - "shape":"String", - "locationName":"data" - } - } - }, - "UserGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserGroup" - } - }, - "UserIdGroupPair":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "UserIdGroupPairList":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserId" - } - }, - "ValueStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "VgwTelemetry":{ - "type":"structure", - "members":{ - "OutsideIpAddress":{ - "shape":"String", - "locationName":"outsideIpAddress" - }, - "Status":{ - "shape":"TelemetryStatus", - "locationName":"status" - }, - "LastStatusChange":{ - "shape":"DateTime", - "locationName":"lastStatusChange" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "AcceptedRouteCount":{ - "shape":"Integer", - "locationName":"acceptedRouteCount" - } - } - }, - "VgwTelemetryList":{ - "type":"list", - "member":{ - "shape":"VgwTelemetry", - "locationName":"item" - } - }, - "VirtualizationType":{ - "type":"string", - "enum":[ - "hvm", - "paravirtual" - ] - }, - "Volume":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Size":{ - "shape":"Integer", - "locationName":"size" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "State":{ - "shape":"VolumeState", - "locationName":"status" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "Attachments":{ - "shape":"VolumeAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "VolumeAttachment":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Device":{ - "shape":"String", - "locationName":"device" - }, - "State":{ - "shape":"VolumeAttachmentState", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "VolumeAttachmentList":{ - "type":"list", - "member":{ - "shape":"VolumeAttachment", - "locationName":"item" - } - }, - "VolumeAttachmentState":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "VolumeAttributeName":{ - "type":"string", - "enum":[ - "autoEnableIO", - "productCodes" - ] - }, - "VolumeDetail":{ - "type":"structure", - "required":["Size"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - } - } - }, - "VolumeIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VolumeId" - } - }, - "VolumeList":{ - "type":"list", - "member":{ - "shape":"Volume", - "locationName":"item" - } - }, - "VolumeState":{ - "type":"string", - "enum":[ - "creating", - "available", - "in-use", - "deleting", - "deleted", - "error" - ] - }, - "VolumeStatusAction":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusActionsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusAction", - "locationName":"item" - } - }, - "VolumeStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"VolumeStatusName", - "locationName":"name" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "VolumeStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusDetails", - "locationName":"item" - } - }, - "VolumeStatusEvent":{ - "type":"structure", - "members":{ - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusEventsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusEvent", - "locationName":"item" - } - }, - "VolumeStatusInfo":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"VolumeStatusInfoStatus", - "locationName":"status" - }, - "Details":{ - "shape":"VolumeStatusDetailsList", - "locationName":"details" - } - } - }, - "VolumeStatusInfoStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data" - ] - }, - "VolumeStatusItem":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VolumeStatus":{ - "shape":"VolumeStatusInfo", - "locationName":"volumeStatus" - }, - "Events":{ - "shape":"VolumeStatusEventsList", - "locationName":"eventsSet" - }, - "Actions":{ - "shape":"VolumeStatusActionsList", - "locationName":"actionsSet" - } - } - }, - "VolumeStatusList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusItem", - "locationName":"item" - } - }, - "VolumeStatusName":{ - "type":"string", - "enum":[ - "io-enabled", - "io-performance" - ] - }, - "VolumeType":{ - "type":"string", - "enum":[ - "standard", - "io1", - "gp2" - ] - }, - "Vpc":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"VpcState", - "locationName":"state" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"isDefault" - } - } - }, - "VpcAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "VpcAttachmentList":{ - "type":"list", - "member":{ - "shape":"VpcAttachment", - "locationName":"item" - } - }, - "VpcAttributeName":{ - "type":"string", - "enum":[ - "enableDnsSupport", - "enableDnsHostnames" - ] - }, - "VpcClassicLink":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkEnabled":{ - "shape":"Boolean", - "locationName":"classicLinkEnabled" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpcClassicLinkIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcClassicLinkList":{ - "type":"list", - "member":{ - "shape":"VpcClassicLink", - "locationName":"item" - } - }, - "VpcEndpoint":{ - "type":"structure", - "members":{ - "VpcEndpointId":{ - "shape":"String", - "locationName":"vpcEndpointId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ServiceName":{ - "shape":"String", - "locationName":"serviceName" - }, - "State":{ - "shape":"State", - "locationName":"state" - }, - "PolicyDocument":{ - "shape":"String", - "locationName":"policyDocument" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"routeTableIdSet" - }, - "CreationTimestamp":{ - "shape":"DateTime", - "locationName":"creationTimestamp" - } - } - }, - "VpcEndpointSet":{ - "type":"list", - "member":{ - "shape":"VpcEndpoint", - "locationName":"item" - } - }, - "VpcIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcList":{ - "type":"list", - "member":{ - "shape":"Vpc", - "locationName":"item" - } - }, - "VpcPeeringConnection":{ - "type":"structure", - "members":{ - "AccepterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"accepterVpcInfo" - }, - "ExpirationTime":{ - "shape":"DateTime", - "locationName":"expirationTime" - }, - "RequesterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"requesterVpcInfo" - }, - "Status":{ - "shape":"VpcPeeringConnectionStateReason", - "locationName":"status" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "VpcPeeringConnectionList":{ - "type":"list", - "member":{ - "shape":"VpcPeeringConnection", - "locationName":"item" - } - }, - "VpcPeeringConnectionStateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"VpcPeeringConnectionStateReasonCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "VpcPeeringConnectionStateReasonCode":{ - "type":"string", - "enum":[ - "initiating-request", - "pending-acceptance", - "active", - "deleted", - "rejected", - "failed", - "expired", - "provisioning", - "deleting" - ] - }, - "VpcPeeringConnectionVpcInfo":{ - "type":"structure", - "members":{ - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "VpcState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "VpnConnection":{ - "type":"structure", - "members":{ - "VpnConnectionId":{ - "shape":"String", - "locationName":"vpnConnectionId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "CustomerGatewayConfiguration":{ - "shape":"String", - "locationName":"customerGatewayConfiguration" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VgwTelemetry":{ - "shape":"VgwTelemetryList", - "locationName":"vgwTelemetry" - }, - "Options":{ - "shape":"VpnConnectionOptions", - "locationName":"options" - }, - "Routes":{ - "shape":"VpnStaticRouteList", - "locationName":"routes" - } - } - }, - "VpnConnectionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnConnectionId" - } - }, - "VpnConnectionList":{ - "type":"list", - "member":{ - "shape":"VpnConnection", - "locationName":"item" - } - }, - "VpnConnectionOptions":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnConnectionOptionsSpecification":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnGateway":{ - "type":"structure", - "members":{ - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VpcAttachments":{ - "shape":"VpcAttachmentList", - "locationName":"attachments" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpnGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnGatewayId" - } - }, - "VpnGatewayList":{ - "type":"list", - "member":{ - "shape":"VpnGateway", - "locationName":"item" - } - }, - "VpnState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "VpnStaticRoute":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "Source":{ - "shape":"VpnStaticRouteSource", - "locationName":"source" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - } - } - }, - "VpnStaticRouteList":{ - "type":"list", - "member":{ - "shape":"VpnStaticRoute", - "locationName":"item" - } - }, - "VpnStaticRouteSource":{ - "type":"string", - "enum":["Static"] - }, - "ZoneNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ZoneName" - } - }, - "NewDhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"NewDhcpConfiguration", - "locationName":"item" - } - }, - "NewDhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "DhcpConfigurationValueList":{ - "type":"list", - "member":{ - "shape":"AttributeValue", - "locationName":"item" - } - }, - "Blob":{"type":"blob"}, - "BlobAttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Blob", - "locationName":"value" - } - } - }, - "RequestSpotLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"ValueStringList", - "locationName":"SecurityGroup" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - }, - "SecurityGroupIds":{ - "shape":"ValueStringList", - "locationName":"SecurityGroupId" - } - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/docs-2.json deleted file mode 100644 index a970264b7..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/docs-2.json +++ /dev/null @@ -1,5495 +0,0 @@ -{ - "version": "2.0", - "operations": { - "AcceptVpcPeeringConnection": "

Accept a VPC peering connection request. To accept a request, the VPC peering connection must be in the pending-acceptance state, and you must be the owner of the peer VPC. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests.

", - "AllocateAddress": "

Acquires an Elastic IP address.

An Elastic IP address is for use either in the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "AssignPrivateIpAddresses": "

Assigns one or more secondary private IP addresses to the specified network interface. You can specify one or more specific secondary IP addresses, or you can specify the number of secondary IP addresses to be automatically assigned within the subnet's CIDR block range. The number of secondary IP addresses that you can assign to an instance varies by instance type. For information about instance types, see Instance Types in the Amazon Elastic Compute Cloud User Guide. For more information about Elastic IP addresses, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

AssignPrivateIpAddresses is available only in EC2-VPC.

", - "AssociateAddress": "

Associates an Elastic IP address with an instance or a network interface.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

[EC2-Classic, VPC in an EC2-VPC-only account] If the Elastic IP address is already associated with a different instance, it is disassociated from that instance and associated with the specified instance.

[VPC in an EC2-Classic account] If you don't specify a private IP address, the Elastic IP address is associated with the primary IP address. If the Elastic IP address is already associated with a different instance or a network interface, you get an error unless you allow reassociation.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

", - "AssociateDhcpOptions": "

Associates a set of DHCP options (that you've previously created) with the specified VPC, or associates no DHCP options with the VPC.

After you associate the options with the VPC, any existing instances and all new instances that you launch in that VPC use the options. You don't need to restart or relaunch the instances. They automatically pick up the changes within a few hours, depending on how frequently the instance renews its DHCP lease. You can explicitly renew the lease using the operating system on the instance.

For more information, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "AssociateRouteTable": "

Associates a subnet with a route table. The subnet and route table must be in the same VPC. This association causes traffic originating from the subnet to be routed according to the routes in the route table. The action returns an association ID, which you need in order to disassociate the route table from the subnet later. A route table can be associated with multiple subnets.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "AttachClassicLinkVpc": "

Links an EC2-Classic instance to a ClassicLink-enabled VPC through one or more of the VPC's security groups. You cannot link an EC2-Classic instance to more than one VPC at a time. You can only link an instance that's in the running state. An instance is automatically unlinked from a VPC when it's stopped - you can link it to the VPC again when you restart it.

After you've linked an instance, you cannot change the VPC security groups that are associated with it. To change the security groups, you must first unlink the instance, and then link it again.

Linking your instance to a VPC is sometimes referred to as attaching your instance.

", - "AttachInternetGateway": "

Attaches an Internet gateway to a VPC, enabling connectivity between the Internet and the VPC. For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "AttachNetworkInterface": "

Attaches a network interface to an instance.

", - "AttachVolume": "

Attaches an EBS volume to a running or stopped instance and exposes it to the instance with the specified device name.

Encrypted EBS volumes may only be attached to instances that support Amazon EBS encryption. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For a list of supported device names, see Attaching an EBS Volume to an Instance. Any device names that aren't reserved for instance store volumes can be used for EBS volumes. For more information, see Amazon EC2 Instance Store in the Amazon Elastic Compute Cloud User Guide.

If a volume has an AWS Marketplace product code:

  • The volume can be attached only to a stopped instance.
  • AWS Marketplace product codes are copied from the volume to the instance.
  • You must be subscribed to the product.
  • The instance type and operating system of the instance must support the product. For example, you can't detach a volume from a Windows instance and attach it to a Linux instance.

For an overview of the AWS Marketplace, see Introducing AWS Marketplace.

For more information about EBS volumes, see Attaching Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "AttachVpnGateway": "

Attaches a virtual private gateway to a VPC. For more information, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "AuthorizeSecurityGroupEgress": "

Adds one or more egress rules to a security group for use with a VPC. Specifically, this action permits instances to send traffic to one or more destination CIDR IP address ranges, or to one or more destination security groups for the same VPC.

You can have up to 50 rules per security group (covering both ingress and egress rules).

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. This action doesn't apply to security groups for use in EC2-Classic. For more information, see Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

Each rule consists of the protocol (for example, TCP), plus either a CIDR range or a source group. For the TCP and UDP protocols, you must also specify the destination port or port range. For the ICMP protocol, you must also specify the ICMP type and code. You can use -1 for the type or code to mean all types or all codes.

Rule changes are propagated to affected instances as quickly as possible. However, a small delay might occur.

", - "AuthorizeSecurityGroupIngress": "

Adds one or more ingress rules to a security group.

EC2-Classic: You can have up to 100 rules per group.

EC2-VPC: You can have up to 50 rules per group (covering both ingress and egress rules).

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

[EC2-Classic] This action gives one or more CIDR IP address ranges permission to access a security group in your account, or gives one or more security groups (called the source groups) permission to access a security group for your account. A source group can be for your own AWS account, or another.

[EC2-VPC] This action gives one or more CIDR IP address ranges permission to access a security group in your VPC, or gives one or more other security groups (called the source groups) permission to access a security group for your VPC. The security groups must all be for the same VPC.

", - "BundleInstance": "

Bundles an Amazon instance store-backed Windows instance.

During bundling, only the root device volume (C:\\) is bundled. Data on other instance store volumes is not preserved.

This action is not applicable for Linux/Unix instances or Windows instances that are backed by Amazon EBS.

For more information, see Creating an Instance Store-Backed Windows AMI.

", - "CancelBundleTask": "

Cancels a bundling operation for an instance store-backed Windows instance.

", - "CancelConversionTask": "

Cancels an active conversion task. The task can be the import of an instance or volume. The action removes all artifacts of the conversion, including a partially uploaded volume or instance. If the conversion is complete or is in the process of transferring the final disk image, the command fails and returns an exception.

For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

", - "CancelExportTask": "

Cancels an active export task. The request removes all artifacts of the export, including any partially-created Amazon S3 objects. If the export task is complete or is in the process of transferring the final disk image, the command fails and returns an error.

", - "CancelImportTask": "

Cancels an in-process import virtual machine or import snapshot task.

", - "CancelReservedInstancesListing": "

Cancels the specified Reserved Instance listing in the Reserved Instance Marketplace.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CancelSpotFleetRequests": "

Cancels the specified Spot fleet requests.

", - "CancelSpotInstanceRequests": "

Cancels one or more Spot instance requests. Spot instances are instances that Amazon EC2 starts on your behalf when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

Canceling a Spot instance request does not terminate running Spot instances associated with the request.

", - "ConfirmProductInstance": "

Determines whether a product code is associated with an instance. This action can only be used by the owner of the product code. It is useful when a product code owner needs to verify whether another user's instance is eligible for support.

", - "CopyImage": "

Initiates the copy of an AMI from the specified source region to the current region. You specify the destination region by using its endpoint when making the request. AMIs that use encrypted EBS snapshots cannot be copied with this method.

For more information, see Copying AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshot": "

Copies a point-in-time snapshot of an EBS volume and stores it in Amazon S3. You can copy the snapshot within the same region or from one region to another. You can use the snapshot to create EBS volumes or Amazon Machine Images (AMIs). The snapshot is copied to the regional endpoint that you send the HTTP request to.

Copies of encrypted EBS snapshots remain encrypted. Copies of unencrypted snapshots remain unencrypted, unless the Encrypted flag is specified during the snapshot copy operation. By default, encrypted snapshot copies use the default AWS Key Management Service (AWS KMS) customer master key (CMK); however, you can specify a non-default CMK with the KmsKeyId parameter.

For more information, see Copying an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGateway": "

Provides information to AWS about your VPN customer gateway device. The customer gateway is the appliance at your end of the VPN connection. (The device on the AWS side of the VPN connection is the virtual private gateway.) You must provide the Internet-routable IP address of the customer gateway's external interface. The IP address must be static and can't be behind a device performing network address translation (NAT).

For devices that use Border Gateway Protocol (BGP), you can also provide the device's BGP Autonomous System Number (ASN). You can use an existing ASN assigned to your network. If you don't have an ASN already, you can use a private ASN (in the 64512 - 65534 range).

Amazon EC2 supports all 2-byte ASN numbers in the range of 1 - 65534, with the exception of 7224, which is reserved in the us-east-1 region, and 9059, which is reserved in the eu-west-1 region.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

You cannot create more than one customer gateway with the same VPN type, IP address, and BGP ASN parameter values. If you run an identical request more than one time, the first request creates the customer gateway, and subsequent requests return information about the existing customer gateway. The subsequent requests do not create new customer gateway resources.

", - "CreateDhcpOptions": "

Creates a set of DHCP options for your VPC. After creating the set, you must associate it with the VPC, causing all existing and new instances that you launch in the VPC to use this set of DHCP options. The following are the individual DHCP options you can specify. For more information about the options, see RFC 2132.

  • domain-name-servers - The IP addresses of up to four domain name servers, or AmazonProvidedDNS. The default DHCP option set specifies AmazonProvidedDNS. If specifying more than one domain name server, specify the IP addresses in a single parameter, separated by commas.
  • domain-name - If you're using AmazonProvidedDNS in us-east-1, specify ec2.internal. If you're using AmazonProvidedDNS in another region, specify region.compute.internal (for example, ap-northeast-1.compute.internal). Otherwise, specify a domain name (for example, MyCompany.com). Important: Some Linux operating systems accept multiple domain names separated by spaces. However, Windows and other Linux operating systems treat the value as a single domain, which results in unexpected behavior. If your DHCP options set is associated with a VPC that has instances with multiple operating systems, specify only one domain name.
  • ntp-servers - The IP addresses of up to four Network Time Protocol (NTP) servers.
  • netbios-name-servers - The IP addresses of up to four NetBIOS name servers.
  • netbios-node-type - The NetBIOS node type (1, 2, 4, or 8). We recommend that you specify 2 (broadcast and multicast are not currently supported). For more information about these node types, see RFC 2132.

Your VPC automatically starts out with a set of DHCP options that includes only a DNS server that we provide (AmazonProvidedDNS). If you create a set of options, and if your VPC has an Internet gateway, make sure to set the domain-name-servers option either to AmazonProvidedDNS or to a domain name server of your choice. For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "CreateFlowLogs": "

Creates one or more flow logs to capture IP traffic for a specific network interface, subnet, or VPC. Flow logs are delivered to a specified log group in Amazon CloudWatch Logs. If you specify a VPC or subnet in the request, a log stream is created in CloudWatch Logs for each network interface in the subnet or VPC. Log streams can include information about accepted and rejected traffic to a network interface. You can view the data in your log streams using Amazon CloudWatch Logs.

In your request, you must also specify an IAM role that has permission to publish logs to CloudWatch Logs.

", - "CreateImage": "

Creates an Amazon EBS-backed AMI from an Amazon EBS-backed instance that is either running or stopped.

If you customized your instance with instance store volumes or EBS volumes in addition to the root device volume, the new AMI contains block device mapping information for those volumes. When you launch an instance from this new AMI, the instance automatically launches with those additional volumes.

For more information, see Creating Amazon EBS-Backed Linux AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateInstanceExportTask": "

Exports a running or stopped instance to an S3 bucket.

For information about the supported operating systems, image formats, and known limitations for the types of instances you can export, see Exporting EC2 Instances in the Amazon Elastic Compute Cloud User Guide.

", - "CreateInternetGateway": "

Creates an Internet gateway for use with a VPC. After creating the Internet gateway, you attach it to a VPC using AttachInternetGateway.

For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "CreateKeyPair": "

Creates a 2048-bit RSA key pair with the specified name. Amazon EC2 stores the public key and displays the private key for you to save to a file. The private key is returned as an unencrypted PEM encoded PKCS#8 private key. If a key with the specified name already exists, Amazon EC2 returns an error.

You can have up to five thousand key pairs per region.

The key pair returned to you is available only in the region in which you create it. To create a key pair that is available in all regions, use ImportKeyPair.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateNetworkAcl": "

Creates a network ACL in a VPC. Network ACLs provide an optional layer of security (in addition to security groups) for the instances in your VPC.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAclEntry": "

Creates an entry (a rule) in a network ACL with the specified rule number. Each network ACL has a set of numbered ingress rules and a separate set of numbered egress rules. When determining whether a packet should be allowed in or out of a subnet associated with the ACL, we process the entries in the ACL according to the rule numbers, in ascending order. Each network ACL has a set of ingress rules and a separate set of egress rules.

We recommend that you leave room between the rule numbers (for example, 100, 110, 120, ...), and not number them one right after the other (for example, 101, 102, 103, ...). This makes it easier to add a rule between existing ones without having to renumber the rules.

After you add an entry, you can't modify it; you must either replace it, or create an entry and delete the old one.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkInterface": "

Creates a network interface in the specified subnet.

For more information about network interfaces, see Elastic Network Interfaces in the Amazon Elastic Compute Cloud User Guide.

", - "CreatePlacementGroup": "

Creates a placement group that you launch cluster instances into. You must give the group a name that's unique within the scope of your account.

For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListing": "

Creates a listing for Amazon EC2 Reserved Instances to be sold in the Reserved Instance Marketplace. You can submit one Reserved Instance listing at a time. To get a list of your Reserved Instances, you can use the DescribeReservedInstances operation.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

To sell your Reserved Instances, you must first register as a seller in the Reserved Instance Marketplace. After completing the registration process, you can create a Reserved Instance Marketplace listing of some or all of your Reserved Instances, and specify the upfront price to receive for them. Your Reserved Instance listings then become available for purchase. To view the details of your Reserved Instance listing, you can use the DescribeReservedInstancesListings operation.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CreateRoute": "

Creates a route in a route table within a VPC.

You must specify one of the following targets: Internet gateway or virtual private gateway, NAT instance, VPC peering connection, or network interface.

When determining how to route traffic, we use the route with the most specific match. For example, let's say the traffic is destined for 192.0.2.3, and the route table includes the following two routes:

  • 192.0.2.0/24 (goes to some target A)

  • 192.0.2.0/28 (goes to some target B)

Both routes apply to the traffic destined for 192.0.2.3. However, the second route in the list covers a smaller number of IP addresses and is therefore more specific, so we use that route to determine where to target the traffic.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateRouteTable": "

Creates a route table for the specified VPC. After you create a route table, you can add routes and associate the table with a subnet.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateSecurityGroup": "

Creates a security group.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

EC2-Classic: You can have up to 500 security groups.

EC2-VPC: You can create up to 100 security groups per VPC.

When you create a security group, you specify a friendly name of your choice. You can have a security group for use in EC2-Classic with the same name as a security group for use in a VPC. However, you can't have two security groups for use in EC2-Classic with the same name or two security groups for use in a VPC with the same name.

You have a default security group for use in EC2-Classic and a default security group for use in your VPC. If you don't specify a security group when you launch an instance, the instance is launched into the appropriate default security group. A default security group includes a default rule that grants instances unrestricted network access to each other.

You can add or remove rules from your security groups using AuthorizeSecurityGroupIngress, AuthorizeSecurityGroupEgress, RevokeSecurityGroupIngress, and RevokeSecurityGroupEgress.

", - "CreateSnapshot": "

Creates a snapshot of an EBS volume and stores it in Amazon S3. You can use snapshots for backups, to make copies of EBS volumes, and to save data before shutting down an instance.

When a snapshot is created, any AWS Marketplace product codes that are associated with the source volume are propagated to the snapshot.

You can take a snapshot of an attached volume that is in use. However, snapshots only capture data that has been written to your EBS volume at the time the snapshot command is issued; this may exclude any data that has been cached by any applications or the operating system. If you can pause any file systems on the volume long enough to take a snapshot, your snapshot should be complete. However, if you cannot pause all file writes to the volume, you should unmount the volume from within the instance, issue the snapshot command, and then remount the volume to ensure a consistent and complete snapshot. You may remount and use your volume while the snapshot status is pending.

To create a snapshot for EBS volumes that serve as root devices, you should stop the instance before taking the snapshot.

Snapshots that are taken from encrypted volumes are automatically encrypted. Volumes that are created from encrypted snapshots are also automatically encrypted. Your encrypted volumes and any associated snapshots always remain protected.

For more information, see Amazon Elastic Block Store and Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSpotDatafeedSubscription": "

Creates a data feed for Spot instances, enabling you to view Spot instance usage logs. You can create one data feed per AWS account. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSubnet": "

Creates a subnet in an existing VPC.

When you create each subnet, you provide the VPC ID and the CIDR block you want for the subnet. After you create a subnet, you can't change its CIDR block. The subnet's CIDR block can be the same as the VPC's CIDR block (assuming you want only a single subnet in the VPC), or a subset of the VPC's CIDR block. If you create more than one subnet in a VPC, the subnets' CIDR blocks must not overlap. The smallest subnet (and VPC) you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses).

AWS reserves both the first four and the last IP address in each subnet's CIDR block. They're not available for use.

If you add more than one subnet to a VPC, they're set up in a star topology with a logical router in the middle.

If you launch an instance in a VPC using an Amazon EBS-backed AMI, the IP address doesn't change if you stop and restart the instance (unlike a similar instance launched outside a VPC, which gets a new IP address when restarted). It's therefore possible to have a subnet with no running instances (they're all stopped), but no remaining IP addresses available.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "CreateTags": "

Adds or overwrites one or more tags for the specified Amazon EC2 resource or resources. Each resource can have a maximum of 10 tags. Each tag consists of a key and optional value. Tag keys must be unique per resource.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVolume": "

Creates an EBS volume that can be attached to an instance in the same Availability Zone. The volume is created in the regional endpoint that you send the HTTP request to. For more information see Regions and Endpoints.

You can create a new empty volume or restore a volume from an EBS snapshot. Any AWS Marketplace product codes from the snapshot are propagated to the volume.

You can create encrypted volumes with the Encrypted parameter. Encrypted volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are also automatically encrypted. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For more information, see Creating or Restoring an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpc": "

Creates a VPC with the specified CIDR block.

The smallest VPC you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses). To help you decide how big to make your VPC, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

By default, each instance you launch in the VPC has the default DHCP options, which includes only a default DNS server that we provide (AmazonProvidedDNS). For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpcEndpoint": "

Creates a VPC endpoint for a specified AWS service. An endpoint enables you to create a private connection between your VPC and another AWS service in your account. You can specify an endpoint policy to attach to the endpoint that will control access to the service from your VPC. You can also specify the VPC route tables that use the endpoint.

Currently, only endpoints to Amazon S3 are supported.

", - "CreateVpcPeeringConnection": "

Requests a VPC peering connection between two VPCs: a requester VPC that you own and a peer VPC with which to create the connection. The peer VPC can belong to another AWS account. The requester VPC and peer VPC cannot have overlapping CIDR blocks.

The owner of the peer VPC must accept the peering request to activate the peering connection. The VPC peering connection request expires after 7 days, after which it cannot be accepted or rejected.

A CreateVpcPeeringConnection request between VPCs with overlapping CIDR blocks results in the VPC peering connection having a status of failed.

", - "CreateVpnConnection": "

Creates a VPN connection between an existing virtual private gateway and a VPN customer gateway. The only supported connection type is ipsec.1.

The response includes information that you need to give to your network administrator to configure your customer gateway.

We strongly recommend that you use HTTPS when calling this operation because the response contains sensitive cryptographic information for configuring your customer gateway.

If you decide to shut down your VPN connection for any reason and later create a new VPN connection, you must reconfigure your customer gateway with the new information returned from this call.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnConnectionRoute": "

Creates a static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnGateway": "

Creates a virtual private gateway. A virtual private gateway is the endpoint on the VPC side of your VPN connection. You can create a virtual private gateway before creating the VPC itself.

For more information about virtual private gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DeleteCustomerGateway": "

Deletes the specified customer gateway. You must delete the VPN connection before you can delete the customer gateway.

", - "DeleteDhcpOptions": "

Deletes the specified set of DHCP options. You must disassociate the set of DHCP options before you can delete it. You can disassociate the set of DHCP options by associating either a new set of options or the default set of options with the VPC.

", - "DeleteFlowLogs": "

Deletes one or more flow logs.

", - "DeleteInternetGateway": "

Deletes the specified Internet gateway. You must detach the Internet gateway from the VPC before you can delete it.

", - "DeleteKeyPair": "

Deletes the specified key pair, by removing the public key from Amazon EC2.

", - "DeleteNetworkAcl": "

Deletes the specified network ACL. You can't delete the ACL if it's associated with any subnets. You can't delete the default network ACL.

", - "DeleteNetworkAclEntry": "

Deletes the specified ingress or egress entry (rule) from the specified network ACL.

", - "DeleteNetworkInterface": "

Deletes the specified network interface. You must detach the network interface before you can delete it.

", - "DeletePlacementGroup": "

Deletes the specified placement group. You must terminate all instances in the placement group before you can delete the placement group. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteRoute": "

Deletes the specified route from the specified route table.

", - "DeleteRouteTable": "

Deletes the specified route table. You must disassociate the route table from any subnets before you can delete it. You can't delete the main route table.

", - "DeleteSecurityGroup": "

Deletes a security group.

If you attempt to delete a security group that is associated with an instance, or is referenced by another security group, the operation fails with InvalidGroup.InUse in EC2-Classic or DependencyViolation in EC2-VPC.

", - "DeleteSnapshot": "

Deletes the specified snapshot.

When you make periodic snapshots of a volume, the snapshots are incremental, and only the blocks on the device that have changed since your last snapshot are saved in the new snapshot. When you delete a snapshot, only the data not needed for any other snapshot is removed. So regardless of which prior snapshots have been deleted, all active snapshots will have access to all the information needed to restore the volume.

You cannot delete a snapshot of the root device of an EBS volume used by a registered AMI. You must first de-register the AMI before you can delete the snapshot.

For more information, see Deleting an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteSpotDatafeedSubscription": "

Deletes the data feed for Spot instances.

", - "DeleteSubnet": "

Deletes the specified subnet. You must terminate all running instances in the subnet before you can delete the subnet.

", - "DeleteTags": "

Deletes the specified set of tags from the specified set of resources. This call is designed to follow a DescribeTags request.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVolume": "

Deletes the specified EBS volume. The volume must be in the available state (not attached to an instance).

The volume may remain in the deleting state for several minutes.

For more information, see Deleting an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVpc": "

Deletes the specified VPC. You must detach or delete all gateways and resources that are associated with the VPC before you can delete it. For example, you must terminate all instances running in the VPC, delete all security groups associated with the VPC (except the default one), delete all route tables associated with the VPC (except the default one), and so on.

", - "DeleteVpcEndpoints": "

Deletes one or more specified VPC endpoints. Deleting the endpoint also deletes the endpoint routes in the route tables that were associated with the endpoint.

", - "DeleteVpcPeeringConnection": "

Deletes a VPC peering connection. Either the owner of the requester VPC or the owner of the peer VPC can delete the VPC peering connection if it's in the active state. The owner of the requester VPC can delete a VPC peering connection in the pending-acceptance state.

", - "DeleteVpnConnection": "

Deletes the specified VPN connection.

If you're deleting the VPC and its associated components, we recommend that you detach the virtual private gateway from the VPC and delete the VPC before deleting the VPN connection. If you believe that the tunnel credentials for your VPN connection have been compromised, you can delete the VPN connection and create a new one that has new keys, without needing to delete the VPC or virtual private gateway. If you create a new VPN connection, you must reconfigure the customer gateway using the new configuration information returned with the new VPN connection ID.

", - "DeleteVpnConnectionRoute": "

Deletes the specified static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

", - "DeleteVpnGateway": "

Deletes the specified virtual private gateway. We recommend that before you delete a virtual private gateway, you detach it from the VPC and delete the VPN connection. Note that you don't need to delete the virtual private gateway if you plan to delete and recreate the VPN connection between your VPC and your network.

", - "DeregisterImage": "

Deregisters the specified AMI. After you deregister an AMI, it can't be used to launch new instances.

This command does not delete the AMI.

", - "DescribeAccountAttributes": "

Describes attributes of your AWS account. The following are the supported account attributes:

  • supported-platforms: Indicates whether your account can launch instances into EC2-Classic and EC2-VPC, or only into EC2-VPC.

  • default-vpc: The ID of the default VPC for your account, or none.

  • max-instances: The maximum number of On-Demand instances that you can run.

  • vpc-max-security-groups-per-interface: The maximum number of security groups that you can assign to a network interface.

  • max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-Classic.

  • vpc-max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-VPC.

", - "DescribeAddresses": "

Describes one or more of your Elastic IP addresses.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeAvailabilityZones": "

Describes one or more of the Availability Zones that are available to you. The results include zones only for the region you're currently using. If there is an event impacting an Availability Zone, you can use this request to view the state and any provided message for that Availability Zone.

For more information, see Regions and Availability Zones in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeBundleTasks": "

Describes one or more of your bundling tasks.

Completed bundle tasks are listed for only a limited time. If your bundle task is no longer in the list, you can still register an AMI from it. Just use RegisterImage with the Amazon S3 bucket name and image manifest name you provided to the bundle task.

", - "DescribeClassicLinkInstances": "

Describes one or more of your linked EC2-Classic instances. This request only returns information about EC2-Classic instances linked to a VPC through ClassicLink; you cannot use this request to return information about other instances.

", - "DescribeConversionTasks": "

Describes one or more of your conversion tasks. For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeCustomerGateways": "

Describes one or more of your VPN customer gateways.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeDhcpOptions": "

Describes one or more of your DHCP options sets.

For more information about DHCP options sets, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeExportTasks": "

Describes one or more of your export tasks.

", - "DescribeFlowLogs": "

Describes one or more flow logs. To view the information in your flow logs (the log streams for the network interfaces), you must use the CloudWatch Logs console or the CloudWatch Logs API.

", - "DescribeImageAttribute": "

Describes the specified attribute of the specified AMI. You can specify only one attribute at a time.

", - "DescribeImages": "

Describes one or more of the images (AMIs, AKIs, and ARIs) available to you. Images available to you include public images, private images that you own, and private images owned by other AWS accounts but for which you have explicit launch permissions.

Deregistered images are included in the returned results for an unspecified interval after deregistration.

", - "DescribeImportImageTasks": "

Displays details about an import virtual machine or import snapshot tasks that are already created.

", - "DescribeImportSnapshotTasks": "

Describes your import snapshot tasks.

", - "DescribeInstanceAttribute": "

Describes the specified attribute of the specified instance. You can specify only one attribute at a time. Valid attribute values are: instanceType | kernel | ramdisk | userData | disableApiTermination | instanceInitiatedShutdownBehavior | rootDeviceName | blockDeviceMapping | productCodes | sourceDestCheck | groupSet | ebsOptimized | sriovNetSupport

", - "DescribeInstanceStatus": "

Describes the status of one or more instances.

Instance status includes the following components:

  • Status checks - Amazon EC2 performs status checks on running EC2 instances to identify hardware and software issues. For more information, see Status Checks for Your Instances and Troubleshooting Instances with Failed Status Checks in the Amazon Elastic Compute Cloud User Guide.

  • Scheduled events - Amazon EC2 can schedule events (such as reboot, stop, or terminate) for your instances related to hardware issues, software updates, or system maintenance. For more information, see Scheduled Events for Your Instances in the Amazon Elastic Compute Cloud User Guide.

  • Instance state - You can manage your instances from the moment you launch them through their termination. For more information, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeInstances": "

Describes one or more of your instances.

If you specify one or more instance IDs, Amazon EC2 returns information for those instances. If you do not specify instance IDs, Amazon EC2 returns information for all relevant instances. If you specify an instance ID that is not valid, an error is returned. If you specify an instance that you do not own, it is not included in the returned results.

Recently terminated instances might appear in the returned results. This interval is usually less than one hour.

", - "DescribeInternetGateways": "

Describes one or more of your Internet gateways.

", - "DescribeKeyPairs": "

Describes one or more of your key pairs.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeMovingAddresses": "

Describes your Elastic IP addresses that are being moved to the EC2-VPC platform, or that are being restored to the EC2-Classic platform. This request does not return information about any other Elastic IP addresses in your account.

", - "DescribeNetworkAcls": "

Describes one or more of your network ACLs.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "DescribeNetworkInterfaceAttribute": "

Describes a network interface attribute. You can specify only one attribute at a time.

", - "DescribeNetworkInterfaces": "

Describes one or more of your network interfaces.

", - "DescribePlacementGroups": "

Describes one or more of your placement groups. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribePrefixLists": "

Describes available AWS services in a prefix list format, which includes the prefix list name and prefix list ID of the service and the IP address range for the service. A prefix list ID is required for creating an outbound security group rule that allows traffic from a VPC to access an AWS service through a VPC endpoint.

", - "DescribeRegions": "

Describes one or more regions that are currently available to you.

For a list of the regions supported by Amazon EC2, see Regions and Endpoints.

", - "DescribeReservedInstances": "

Describes one or more of the Reserved Instances that you purchased.

For more information about Reserved Instances, see Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesListings": "

Describes your account's Reserved Instance listings in the Reserved Instance Marketplace.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

As a seller, you choose to list some or all of your Reserved Instances, and you specify the upfront price to receive for them. Your Reserved Instances are then listed in the Reserved Instance Marketplace and are available for purchase.

As a buyer, you specify the configuration of the Reserved Instance to purchase, and the Marketplace matches what you're searching for with what's available. The Marketplace first sells the lowest priced Reserved Instances to you, and continues to sell available Reserved Instance listings to you until your demand is met. You are charged based on the total price of all of the listings that you purchase.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesModifications": "

Describes the modifications made to your Reserved Instances. If no parameter is specified, information about all your Reserved Instances modification requests is returned. If a modification ID is specified, only information about the specific modification is returned.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesOfferings": "

Describes Reserved Instance offerings that are available for purchase. With Reserved Instances, you purchase the right to launch instances for a period of time. During that time period, you do not receive insufficient capacity errors, and you pay a lower usage rate than the rate charged for On-Demand instances for the actual time used.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeRouteTables": "

Describes one or more of your route tables.

Each subnet in your VPC must be associated with a route table. If a subnet is not explicitly associated with any route table, it is implicitly associated with the main route table. This command does not return the subnet ID for implicit associations.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "DescribeSecurityGroups": "

Describes one or more of your security groups.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeSnapshotAttribute": "

Describes the specified attribute of the specified snapshot. You can specify only one attribute at a time.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSnapshots": "

Describes one or more of the EBS snapshots available to you. Available snapshots include public snapshots available for any AWS account to launch, private snapshots that you own, and private snapshots owned by another AWS account but for which you've been given explicit create volume permissions.

The create volume permissions fall into the following categories:

  • public: The owner of the snapshot granted create volume permissions for the snapshot to the all group. All AWS accounts have create volume permissions for these snapshots.
  • explicit: The owner of the snapshot granted create volume permissions to a specific AWS account.
  • implicit: An AWS account has implicit create volume permissions for all snapshots it owns.

The list of snapshots returned can be modified by specifying snapshot IDs, snapshot owners, or AWS accounts with create volume permissions. If no options are specified, Amazon EC2 returns all snapshots for which you have create volume permissions.

If you specify one or more snapshot IDs, only snapshots that have the specified IDs are returned. If you specify an invalid snapshot ID, an error is returned. If you specify a snapshot ID for which you do not have access, it is not included in the returned results.

If you specify one or more snapshot owners, only snapshots from the specified owners and for which you have access are returned. The results can include the AWS account IDs of the specified owners, amazon for snapshots owned by Amazon, or self for snapshots that you own.

If you specify a list of restorable users, only snapshots with create snapshot permissions for those users are returned. You can specify AWS account IDs (if you own the snapshots), self for snapshots for which you own or have explicit permissions, or all for public snapshots.

If you are describing a long list of snapshots, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeSnapshots request to retrieve the remaining results.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotDatafeedSubscription": "

Describes the data feed for Spot instances. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotFleetInstances": "

Describes the running instances for the specified Spot fleet.

", - "DescribeSpotFleetRequestHistory": "

Describes the events for the specified Spot fleet request during the specified time.

Spot fleet events are delayed by up to 30 seconds before they can be described. This ensures that you can query by the last evaluated time and not miss a recorded event.

", - "DescribeSpotFleetRequests": "

Describes your Spot fleet requests.

", - "DescribeSpotInstanceRequests": "

Describes the Spot instance requests that belong to your account. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

You can use DescribeSpotInstanceRequests to find a running Spot instance by examining the response. If the status of the Spot instance is fulfilled, the instance ID appears in the response and contains the identifier of the instance. Alternatively, you can use DescribeInstances with a filter to look for instances where the instance lifecycle is spot.

", - "DescribeSpotPriceHistory": "

Describes the Spot price history. The prices returned are listed in chronological order, from the oldest to the most recent, for up to the past 90 days. For more information, see Spot Instance Pricing History in the Amazon Elastic Compute Cloud User Guide.

When you specify a start and end time, this operation returns the prices of the instance types within the time range that you specified and the time when the price changed. The price is valid within the time period that you specified; the response merely indicates the last time that the price changed.

", - "DescribeSubnets": "

Describes one or more of your subnets.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeTags": "

Describes one or more of the tags for your EC2 resources.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeAttribute": "

Describes the specified attribute of the specified volume. You can specify only one attribute at a time.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeStatus": "

Describes the status of the specified volumes. Volume status provides the result of the checks performed on your volumes to determine events that can impair the performance of your volumes. The performance of a volume can be affected if an issue occurs on the volume's underlying host. If the volume's underlying host experiences a power outage or system issue, after the system is restored, there could be data inconsistencies on the volume. Volume events notify you if this occurs. Volume actions notify you if any action needs to be taken in response to the event.

The DescribeVolumeStatus operation provides the following information about the specified volumes:

Status: Reflects the current status of the volume. The possible values are ok, impaired , warning, or insufficient-data. If all checks pass, the overall status of the volume is ok. If the check fails, the overall status is impaired. If the status is insufficient-data, then the checks may still be taking place on your volume at the time. We recommend that you retry the request. For more information on volume status, see Monitoring the Status of Your Volumes.

Events: Reflect the cause of a volume status and may require you to take action. For example, if your volume returns an impaired status, then the volume event might be potential-data-inconsistency. This means that your volume has been affected by an issue with the underlying host, has all I/O operations disabled, and may have inconsistent data.

Actions: Reflect the actions you may have to take in response to an event. For example, if the status of the volume is impaired and the volume event shows potential-data-inconsistency, then the action shows enable-volume-io. This means that you may want to enable the I/O operations for the volume by calling the EnableVolumeIO action and then check the volume for data consistency.

Volume status is based on the volume status checks, and does not reflect the volume state. Therefore, volume status does not indicate volumes in the error state (for example, when a volume is incapable of accepting I/O.)

", - "DescribeVolumes": "

Describes the specified EBS volumes.

If you are describing a long list of volumes, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeVolumes request to retrieve the remaining results.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcAttribute": "

Describes the specified attribute of the specified VPC. You can specify only one attribute at a time.

", - "DescribeVpcClassicLink": "

Describes the ClassicLink status of one or more VPCs.

", - "DescribeVpcEndpointServices": "

Describes all supported AWS services that can be specified when creating a VPC endpoint.

", - "DescribeVpcEndpoints": "

Describes one or more of your VPC endpoints.

", - "DescribeVpcPeeringConnections": "

Describes one or more of your VPC peering connections.

", - "DescribeVpcs": "

Describes one or more of your VPCs.

", - "DescribeVpnConnections": "

Describes one or more of your VPN connections.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeVpnGateways": "

Describes one or more of your virtual private gateways.

For more information about virtual private gateways, see Adding an IPsec Hardware VPN to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DetachClassicLinkVpc": "

Unlinks (detaches) a linked EC2-Classic instance from a VPC. After the instance has been unlinked, the VPC security groups are no longer associated with it. An instance is automatically unlinked from a VPC when it's stopped.

", - "DetachInternetGateway": "

Detaches an Internet gateway from a VPC, disabling connectivity between the Internet and the VPC. The VPC must not contain any running instances with Elastic IP addresses.

", - "DetachNetworkInterface": "

Detaches a network interface from an instance.

", - "DetachVolume": "

Detaches an EBS volume from an instance. Make sure to unmount any file systems on the device within your operating system before detaching the volume. Failure to do so results in the volume being stuck in a busy state while detaching.

If an Amazon EBS volume is the root device of an instance, it can't be detached while the instance is running. To detach the root volume, stop the instance first.

When a volume with an AWS Marketplace product code is detached from an instance, the product code is no longer associated with the instance.

For more information, see Detaching an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DetachVpnGateway": "

Detaches a virtual private gateway from a VPC. You do this if you're planning to turn off the VPC and not use it anymore. You can confirm a virtual private gateway has been completely detached from a VPC by describing the virtual private gateway (any attachments to the virtual private gateway are also described).

You must wait for the attachment's state to switch to detached before you can delete the VPC or attach a different VPC to the virtual private gateway.

", - "DisableVgwRoutePropagation": "

Disables a virtual private gateway (VGW) from propagating routes to a specified route table of a VPC.

", - "DisableVpcClassicLink": "

Disables ClassicLink for a VPC. You cannot disable ClassicLink for a VPC that has EC2-Classic instances linked to it.

", - "DisassociateAddress": "

Disassociates an Elastic IP address from the instance or network interface it's associated with.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

", - "DisassociateRouteTable": "

Disassociates a subnet from a route table.

After you perform this action, the subnet no longer uses the routes in the route table. Instead, it uses the routes in the VPC's main route table. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "EnableVgwRoutePropagation": "

Enables a virtual private gateway (VGW) to propagate routes to the specified route table of a VPC.

", - "EnableVolumeIO": "

Enables I/O operations for a volume that had I/O operations disabled because the data on the volume was potentially inconsistent.

", - "EnableVpcClassicLink": "

Enables a VPC for ClassicLink. You can then link EC2-Classic instances to your ClassicLink-enabled VPC to allow communication over private IP addresses. You cannot enable your VPC for ClassicLink if any of your VPC's route tables have existing routes for address ranges within the 10.0.0.0/8 IP address range, excluding local routes for VPCs in the 10.0.0.0/16 and 10.1.0.0/16 IP address ranges. For more information, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "GetConsoleOutput": "

Gets the console output for the specified instance.

Instances do not have a physical monitor through which you can view their console output. They also lack physical controls that allow you to power up, reboot, or shut them down. To allow these actions, we provide them through the Amazon EC2 API and command line interface.

Instance console output is buffered and posted shortly after instance boot, reboot, and termination. Amazon EC2 preserves the most recent 64 KB output which is available for at least one hour after the most recent post.

For Linux instances, the instance console output displays the exact console output that would normally be displayed on a physical monitor attached to a computer. This output is buffered because the instance produces it and then posts it to a store where the instance's owner can retrieve it.

For Windows instances, the instance console output includes output from the EC2Config service.

", - "GetPasswordData": "

Retrieves the encrypted administrator password for an instance running Windows.

The Windows password is generated at boot if the EC2Config service plugin, Ec2SetPassword, is enabled. This usually only happens the first time an AMI is launched, and then Ec2SetPassword is automatically disabled. The password is not generated for rebundled AMIs unless Ec2SetPassword is enabled before bundling.

The password is encrypted using the key pair that you specified when you launched the instance. You must provide the corresponding key pair file.

Password generation and encryption takes a few moments. We recommend that you wait up to 15 minutes after launching an instance before trying to retrieve the generated password.

", - "ImportImage": "

Import single or multi-volume disk images or EBS snapshots into an Amazon Machine Image (AMI).

", - "ImportInstance": "

Creates an import instance task using metadata from the specified disk image. ImportInstance only supports single-volume VMs. To import multi-volume VMs, use ImportImage. After importing the image, you then upload it using the ec2-import-volume command in the EC2 command line tools. For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

", - "ImportKeyPair": "

Imports the public key from an RSA key pair that you created with a third-party tool. Compare this with CreateKeyPair, in which AWS creates the key pair and gives the keys to you (AWS keeps a copy of the public key). With ImportKeyPair, you create the key pair and give AWS just the public key. The private key is never transferred between you and AWS.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "ImportSnapshot": "

Imports a disk into an EBS snapshot.

", - "ImportVolume": "

Creates an import volume task using metadata from the specified disk image. After importing the image, you then upload it using the ec2-import-volume command in the Amazon EC2 command-line interface (CLI) tools. For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

", - "ModifyImageAttribute": "

Modifies the specified attribute of the specified AMI. You can specify only one attribute at a time.

AWS Marketplace product codes cannot be modified. Images with an AWS Marketplace product code cannot be made public.

", - "ModifyInstanceAttribute": "

Modifies the specified attribute of the specified instance. You can specify only one attribute at a time.

To modify some attributes, the instance must be stopped. For more information, see Modifying Attributes of a Stopped Instance in the Amazon Elastic Compute Cloud User Guide.

", - "ModifyNetworkInterfaceAttribute": "

Modifies the specified network interface attribute. You can specify only one attribute at a time.

", - "ModifyReservedInstances": "

Modifies the Availability Zone, instance count, instance type, or network platform (EC2-Classic or EC2-VPC) of your Reserved Instances. The Reserved Instances to be modified must be identical, except for Availability Zone, network platform, and instance type.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "ModifySnapshotAttribute": "

Adds or removes permission settings for the specified snapshot. You may add or remove specified AWS account IDs from a snapshot's list of create volume permissions, but you cannot do both in a single API call. If you need to both add and remove account IDs for a snapshot, you must use multiple API calls.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

Snapshots with AWS Marketplace product codes cannot be made public.

", - "ModifySubnetAttribute": "

Modifies a subnet attribute.

", - "ModifyVolumeAttribute": "

Modifies a volume attribute.

By default, all I/O operations for the volume are suspended when the data on the volume is determined to be potentially inconsistent, to prevent undetectable, latent data corruption. The I/O access to the volume can be resumed by first enabling I/O access and then checking the data consistency on your volume.

You can change the default behavior to resume I/O operations. We recommend that you change this only for boot volumes or for volumes that are stateless or disposable.

", - "ModifyVpcAttribute": "

Modifies the specified attribute of the specified VPC.

", - "ModifyVpcEndpoint": "

Modifies attributes of a specified VPC endpoint. You can modify the policy associated with the endpoint, and you can add and remove route tables associated with the endpoint.

", - "MonitorInstances": "

Enables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "MoveAddressToVpc": "

Moves an Elastic IP address from the EC2-Classic platform to the EC2-VPC platform. The Elastic IP address must be allocated to your account, and it must not be associated with an instance. After the Elastic IP address is moved, it is no longer available for use in the EC2-Classic platform, unless you move it back using the RestoreAddressToClassic request. You cannot move an Elastic IP address that's allocated for use in the EC2-VPC platform to the EC2-Classic platform.

", - "PurchaseReservedInstancesOffering": "

Purchases a Reserved Instance for use with your account. With Amazon EC2 Reserved Instances, you obtain a capacity reservation for a certain instance configuration over a specified period of time and pay a lower hourly rate compared to on-Demand Instance pricing.

Use DescribeReservedInstancesOfferings to get a list of Reserved Instance offerings that match your specifications. After you've purchased a Reserved Instance, you can check for your new Reserved Instance with DescribeReservedInstances.

For more information, see Reserved Instances and Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "RebootInstances": "

Requests a reboot of one or more instances. This operation is asynchronous; it only queues a request to reboot the specified instances. The operation succeeds if the instances are valid and belong to you. Requests to reboot terminated instances are ignored.

If a Linux/Unix instance does not cleanly shut down within four minutes, Amazon EC2 performs a hard reboot.

For more information about troubleshooting, see Getting Console Output and Rebooting Instances in the Amazon Elastic Compute Cloud User Guide.

", - "RegisterImage": "

Registers an AMI. When you're creating an AMI, this is the final step you must complete before you can launch an instance from the AMI. For more information about creating AMIs, see Creating Your Own AMIs in the Amazon Elastic Compute Cloud User Guide.

For Amazon EBS-backed instances, CreateImage creates and registers the AMI in a single request, so you don't have to register the AMI yourself.

You can also use RegisterImage to create an Amazon EBS-backed Linux AMI from a snapshot of a root device volume. For more information, see Launching an Instance from a Snapshot in the Amazon Elastic Compute Cloud User Guide.

Some Linux distributions, such as Red Hat Enterprise Linux (RHEL) and SUSE Linux Enterprise Server (SLES), use the EC2 billingProduct code associated with an AMI to verify subscription status for package updates. Creating an AMI from an EBS snapshot does not maintain this billing code, and subsequent instances launched from such an AMI will not be able to connect to package update infrastructure.

Similarly, although you can create a Windows AMI from a snapshot, you can't successfully launch an instance from the AMI.

To create Windows AMIs or to create AMIs for Linux operating systems that must retain AMI billing codes to work properly, see CreateImage.

If needed, you can deregister an AMI at any time. Any modifications you make to an AMI backed by an instance store volume invalidates its registration. If you make changes to an image, deregister the previous image and register the new image.

You can't register an image where a secondary (non-root) snapshot has AWS Marketplace product codes.

", - "RejectVpcPeeringConnection": "

Rejects a VPC peering connection request. The VPC peering connection must be in the pending-acceptance state. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests. To delete an active VPC peering connection, or to delete a VPC peering connection request that you initiated, use DeleteVpcPeeringConnection.

", - "ReleaseAddress": "

Releases the specified Elastic IP address.

After releasing an Elastic IP address, it is released to the IP address pool and might be unavailable to you. Be sure to update your DNS records and any servers or devices that communicate with the address. If you attempt to release an Elastic IP address that you already released, you'll get an AuthFailure error if the address is already allocated to another AWS account.

[EC2-Classic, default VPC] Releasing an Elastic IP address automatically disassociates it from any instance that it's associated with. To disassociate an Elastic IP address without releasing it, use DisassociateAddress.

[Nondefault VPC] You must use DisassociateAddress to disassociate the Elastic IP address before you try to release it. Otherwise, Amazon EC2 returns an error (InvalidIPAddress.InUse).

", - "ReplaceNetworkAclAssociation": "

Changes which network ACL a subnet is associated with. By default when you create a subnet, it's automatically associated with the default network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceNetworkAclEntry": "

Replaces an entry (rule) in a network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRoute": "

Replaces an existing route within a route table in a VPC. You must provide only one of the following: Internet gateway or virtual private gateway, NAT instance, VPC peering connection, or network interface.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRouteTableAssociation": "

Changes the route table associated with a given subnet in a VPC. After the operation completes, the subnet uses the routes in the new route table it's associated with. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

You can also use ReplaceRouteTableAssociation to change which table is the main route table in the VPC. You just specify the main route table's association ID and the route table to be the new main route table.

", - "ReportInstanceStatus": "

Submits feedback about the status of an instance. The instance must be in the running state. If your experience with the instance differs from the instance status returned by DescribeInstanceStatus, use ReportInstanceStatus to report your experience with the instance. Amazon EC2 collects this information to improve the accuracy of status checks.

Use of this action does not change the value returned by DescribeInstanceStatus.

", - "RequestSpotFleet": "

Creates a Spot fleet request.

You can submit a single request that includes multiple launch specifications that vary by instance type, AMI, Availability Zone, or subnet.

By default, the Spot fleet requests Spot instances in the Spot pool where the price per unit is the lowest. Each launch specification can include its own instance weighting that reflects the value of the instance type to your application workload.

Alternatively, you can specify that the Spot fleet distribute the target capacity across the Spot pools included in its launch specifications. By ensuring that the Spot instances in your Spot fleet are in different Spot pools, you can improve the availability of your fleet.

For more information, see Spot Fleet Requests in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstances": "

Creates a Spot instance request. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot Instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

", - "ResetImageAttribute": "

Resets an attribute of an AMI to its default value.

The productCodes attribute can't be reset.

", - "ResetInstanceAttribute": "

Resets an attribute of an instance to its default value. To reset the kernel or ramdisk, the instance must be in a stopped state. To reset the SourceDestCheck, the instance can be either running or stopped.

The SourceDestCheck attribute controls whether source/destination checking is enabled. The default value is true, which means checking is enabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ResetNetworkInterfaceAttribute": "

Resets a network interface attribute. You can specify only one attribute at a time.

", - "ResetSnapshotAttribute": "

Resets permission settings for the specified snapshot.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "RestoreAddressToClassic": "

Restores an Elastic IP address that was previously moved to the EC2-VPC platform back to the EC2-Classic platform. You cannot move an Elastic IP address that was originally allocated for use in EC2-VPC. The Elastic IP address must not be associated with an instance or network interface.

", - "RevokeSecurityGroupEgress": "

Removes one or more egress rules from a security group for EC2-VPC. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be revoked.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RevokeSecurityGroupIngress": "

Removes one or more ingress rules from a security group. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be removed.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RunInstances": "

Launches the specified number of instances using an AMI for which you have permissions.

When you launch an instance, it enters the pending state. After the instance is ready for you, it enters the running state. To check the state of your instance, call DescribeInstances.

If you don't specify a security group when launching an instance, Amazon EC2 uses the default security group. For more information, see Security Groups in the Amazon Elastic Compute Cloud User Guide.

[EC2-VPC only accounts] If you don't specify a subnet in the request, we choose a default subnet from your default VPC for you.

[EC2-Classic accounts] If you're launching into EC2-Classic and you don't specify an Availability Zone, we choose one for you.

Linux instances have access to the public key of the key pair at boot. You can use this key to provide secure access to the instance. Amazon EC2 public images use this feature to provide secure access without passwords. For more information, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

You can provide optional user data when launching an instance. For more information, see Instance Metadata in the Amazon Elastic Compute Cloud User Guide.

If any of the AMIs have a product code attached for which the user has not subscribed, RunInstances fails.

T2 instance types can only be launched into a VPC. If you do not have a default VPC, or if you do not specify a subnet ID in the request, RunInstances fails.

For more information about troubleshooting, see What To Do If An Instance Immediately Terminates, and Troubleshooting Connecting to Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "StartInstances": "

Starts an Amazon EBS-backed AMI that you've previously stopped.

Instances that use Amazon EBS volumes as their root devices can be quickly stopped and started. When an instance is stopped, the compute resources are released and you are not billed for hourly instance usage. However, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. You can restart your instance at any time. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Performing this operation on an instance that uses an instance store as its root device returns an error.

For more information, see Stopping Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StopInstances": "

Stops an Amazon EBS-backed instance. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

You can't start or stop Spot Instances.

Instances that use Amazon EBS volumes as their root devices can be quickly stopped and started. When an instance is stopped, the compute resources are released and you are not billed for hourly instance usage. However, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. You can restart your instance at any time.

Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Performing this operation on an instance that uses an instance store as its root device returns an error.

You can stop, start, and terminate EBS-backed instances. You can only terminate instance store-backed instances. What happens to an instance differs if you stop it or terminate it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, the root device and any other devices attached during the instance launch are automatically deleted. For more information about the differences between stopping and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

For more information about troubleshooting, see Troubleshooting Stopping Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "TerminateInstances": "

Shuts down one or more instances. This operation is idempotent; if you terminate an instance more than once, each call succeeds.

Terminated instances remain visible after termination (for approximately one hour).

By default, Amazon EC2 deletes all EBS volumes that were attached when the instance launched. Volumes attached after instance launch continue running.

You can stop, start, and terminate EBS-backed instances. You can only terminate instance store-backed instances. What happens to an instance differs if you stop it or terminate it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, any attached EBS volumes with the DeleteOnTermination block device mapping parameter set to true are automatically deleted. For more information about the differences between stopping and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

For more information about troubleshooting, see Troubleshooting Terminating Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "UnassignPrivateIpAddresses": "

Unassigns one or more secondary private IP addresses from a network interface.

", - "UnmonitorInstances": "

Disables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

" - }, - "service": "Amazon Elastic Compute Cloud

Amazon Elastic Compute Cloud (Amazon EC2) provides resizable computing capacity in the Amazon Web Services (AWS) cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster.

", - "shapes": { - "AcceptVpcPeeringConnectionRequest": { - "base": null, - "refs": { - } - }, - "AcceptVpcPeeringConnectionResult": { - "base": null, - "refs": { - } - }, - "AccountAttribute": { - "base": "

Describes an account attribute.

", - "refs": { - "AccountAttributeList$member": null - } - }, - "AccountAttributeList": { - "base": null, - "refs": { - "DescribeAccountAttributesResult$AccountAttributes": "

Information about one or more account attributes.

" - } - }, - "AccountAttributeName": { - "base": null, - "refs": { - "AccountAttributeNameStringList$member": null - } - }, - "AccountAttributeNameStringList": { - "base": null, - "refs": { - "DescribeAccountAttributesRequest$AttributeNames": "

One or more account attribute names.

" - } - }, - "AccountAttributeValue": { - "base": "

Describes a value of an account attribute.

", - "refs": { - "AccountAttributeValueList$member": null - } - }, - "AccountAttributeValueList": { - "base": null, - "refs": { - "AccountAttribute$AttributeValues": "

One or more values for the account attribute.

" - } - }, - "ActiveInstance": { - "base": "

Describes a running instance in a Spot fleet.

", - "refs": { - "ActiveInstanceSet$member": null - } - }, - "ActiveInstanceSet": { - "base": null, - "refs": { - "DescribeSpotFleetInstancesResponse$ActiveInstances": "

The running instances. Note that this list is refreshed periodically and might be out of date.

" - } - }, - "Address": { - "base": "

Describes an Elastic IP address.

", - "refs": { - "AddressList$member": null - } - }, - "AddressList": { - "base": null, - "refs": { - "DescribeAddressesResult$Addresses": "

Information about one or more Elastic IP addresses.

" - } - }, - "AllocateAddressRequest": { - "base": null, - "refs": { - } - }, - "AllocateAddressResult": { - "base": null, - "refs": { - } - }, - "AllocationIdList": { - "base": null, - "refs": { - "DescribeAddressesRequest$AllocationIds": "

[EC2-VPC] One or more allocation IDs.

Default: Describes all your Elastic IP addresses.

" - } - }, - "AllocationStrategy": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$AllocationStrategy": "

Determines how to allocate the target capacity across the Spot pools specified by the Spot fleet request. The default is lowestPrice.

" - } - }, - "ArchitectureValues": { - "base": null, - "refs": { - "Image$Architecture": "

The architecture of the image.

", - "ImportInstanceLaunchSpecification$Architecture": "

The architecture of the instance.

", - "Instance$Architecture": "

The architecture of the image.

", - "RegisterImageRequest$Architecture": "

The architecture of the AMI.

Default: For Amazon EBS-backed AMIs, i386. For instance store-backed AMIs, the architecture specified in the manifest file.

" - } - }, - "AssignPrivateIpAddressesRequest": { - "base": null, - "refs": { - } - }, - "AssociateAddressRequest": { - "base": null, - "refs": { - } - }, - "AssociateAddressResult": { - "base": null, - "refs": { - } - }, - "AssociateDhcpOptionsRequest": { - "base": null, - "refs": { - } - }, - "AssociateRouteTableRequest": { - "base": null, - "refs": { - } - }, - "AssociateRouteTableResult": { - "base": null, - "refs": { - } - }, - "AttachClassicLinkVpcRequest": { - "base": null, - "refs": { - } - }, - "AttachClassicLinkVpcResult": { - "base": null, - "refs": { - } - }, - "AttachInternetGatewayRequest": { - "base": null, - "refs": { - } - }, - "AttachNetworkInterfaceRequest": { - "base": null, - "refs": { - } - }, - "AttachNetworkInterfaceResult": { - "base": null, - "refs": { - } - }, - "AttachVolumeRequest": { - "base": null, - "refs": { - } - }, - "AttachVpnGatewayRequest": { - "base": null, - "refs": { - } - }, - "AttachVpnGatewayResult": { - "base": null, - "refs": { - } - }, - "AttachmentStatus": { - "base": null, - "refs": { - "EbsInstanceBlockDevice$Status": "

The attachment state.

", - "InstanceNetworkInterfaceAttachment$Status": "

The attachment state.

", - "InternetGatewayAttachment$State": "

The current state of the attachment.

", - "NetworkInterfaceAttachment$Status": "

The attachment state.

", - "VpcAttachment$State": "

The current state of the attachment.

" - } - }, - "AttributeBooleanValue": { - "base": "

The value to use when a resource attribute accepts a Boolean value.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$SourceDestCheck": "

Indicates whether source/destination checking is enabled.

", - "DescribeVolumeAttributeResult$AutoEnableIO": "

The state of autoEnableIO attribute.

", - "DescribeVpcAttributeResult$EnableDnsSupport": "

Indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for your instances to their corresponding IP addresses; otherwise, it does not.

", - "DescribeVpcAttributeResult$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.

", - "InstanceAttribute$DisableApiTermination": "

If the value is true, you can't terminate the instance through the Amazon EC2 console, CLI, or API; otherwise, you can.

", - "InstanceAttribute$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O.

", - "InstanceAttribute$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$SourceDestCheck": "

Specifies whether source/destination checking is enabled. A value of true means that checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$DisableApiTermination": "

If the value is true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. You cannot use this paramater for Spot Instances.

", - "ModifyInstanceAttributeRequest$EbsOptimized": "

Specifies whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "ModifyNetworkInterfaceAttributeRequest$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ModifySubnetAttributeRequest$MapPublicIpOnLaunch": "

Specify true to indicate that instances launched into the specified subnet should be assigned public IP address.

", - "ModifyVolumeAttributeRequest$AutoEnableIO": "

Indicates whether the volume should be auto-enabled for I/O operations.

", - "ModifyVpcAttributeRequest$EnableDnsSupport": "

Indicates whether the DNS resolution is supported for the VPC. If enabled, queries to the Amazon provided DNS server at the 169.254.169.253 IP address, or the reserved IP address at the base of the VPC network range \"plus two\" will succeed. If disabled, the Amazon provided DNS service in the VPC that resolves public DNS hostnames to IP addresses is not enabled.

", - "ModifyVpcAttributeRequest$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If enabled, instances in the VPC get DNS hostnames; otherwise, they do not.

You can only enable DNS hostnames if you also enable DNS support.

" - } - }, - "AttributeValue": { - "base": "

The value to use for a resource attribute.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Description": "

The description of the network interface.

", - "ImageAttribute$KernelId": "

The kernel ID.

", - "ImageAttribute$RamdiskId": "

The RAM disk ID.

", - "ImageAttribute$Description": "

A description for the AMI.

", - "ImageAttribute$SriovNetSupport": null, - "InstanceAttribute$InstanceType": "

The instance type.

", - "InstanceAttribute$KernelId": "

The kernel ID.

", - "InstanceAttribute$RamdiskId": "

The RAM disk ID.

", - "InstanceAttribute$UserData": "

The Base64-encoded MIME user data.

", - "InstanceAttribute$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "InstanceAttribute$RootDeviceName": "

The name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "InstanceAttribute$SriovNetSupport": null, - "ModifyImageAttributeRequest$Description": "

A description for the AMI.

", - "ModifyInstanceAttributeRequest$InstanceType": "

Changes the instance type to the specified value. For more information, see Instance Types. If the instance type is not valid, the error returned is InvalidInstanceAttributeValue.

", - "ModifyInstanceAttributeRequest$Kernel": "

Changes the instance's kernel to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$Ramdisk": "

Changes the instance's RAM disk to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$InstanceInitiatedShutdownBehavior": "

Specifies whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "ModifyInstanceAttributeRequest$SriovNetSupport": "

Set to simple to enable enhanced networking for the instance.

There is no way to disable enhanced networking at this time.

This option is supported only for HVM instances. Specifying this option with a PV instance can make it unreachable.

", - "ModifyNetworkInterfaceAttributeRequest$Description": "

A description for the network interface.

", - "DhcpConfigurationValueList$member": null - } - }, - "AuthorizeSecurityGroupEgressRequest": { - "base": null, - "refs": { - } - }, - "AuthorizeSecurityGroupIngressRequest": { - "base": null, - "refs": { - } - }, - "AvailabilityZone": { - "base": "

Describes an Availability Zone.

", - "refs": { - "AvailabilityZoneList$member": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesResult$AvailabilityZones": "

Information about one or more Availability Zones.

" - } - }, - "AvailabilityZoneMessage": { - "base": "

Describes a message about an Availability Zone.

", - "refs": { - "AvailabilityZoneMessageList$member": null - } - }, - "AvailabilityZoneMessageList": { - "base": null, - "refs": { - "AvailabilityZone$Messages": "

Any messages about the Availability Zone.

" - } - }, - "AvailabilityZoneState": { - "base": null, - "refs": { - "AvailabilityZone$State": "

The state of the Availability Zone (available | impaired | unavailable).

" - } - }, - "BatchState": { - "base": null, - "refs": { - "CancelSpotFleetRequestsSuccessItem$CurrentSpotFleetRequestState": "

The current state of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$PreviousSpotFleetRequestState": "

The previous state of the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestState": "

The state of the Spot fleet request.

" - } - }, - "BlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "BlockDeviceMappingList$member": null, - "BlockDeviceMappingRequestList$member": null - } - }, - "BlockDeviceMappingList": { - "base": null, - "refs": { - "Image$BlockDeviceMappings": "

Any block device mapping entries.

", - "ImageAttribute$BlockDeviceMappings": "

One or more block device mapping entries.

", - "LaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

", - "SpotFleetLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

", - "RequestSpotLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "BlockDeviceMappingRequestList": { - "base": null, - "refs": { - "CreateImageRequest$BlockDeviceMappings": "

Information about one or more block device mappings.

", - "RegisterImageRequest$BlockDeviceMappings": "

One or more block device mapping entries.

", - "RunInstancesRequest$BlockDeviceMappings": "

The block device mapping.

" - } - }, - "Boolean": { - "base": null, - "refs": { - "AcceptVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AllocateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssignPrivateIpAddressesRequest$AllowReassignment": "

Indicates whether to allow an IP address that is already assigned to another network interface or instance to be reassigned to the specified network interface.

", - "AssociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateAddressRequest$AllowReassociation": "

[EC2-VPC] Allows an Elastic IP address that is already associated with an instance or network interface to be re-associated with the specified instance or network interface. Otherwise, the operation fails.

Default: false

", - "AssociateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "AttachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttributeBooleanValue$Value": "

Valid values are true or false.

", - "AuthorizeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AuthorizeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "BundleInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelBundleTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelConversionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelImportTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$TerminateInstances": "

Indicates whether to terminate instances for a Spot fleet request if it is canceled successfully.

", - "CancelSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ConfirmProductInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ConfirmProductInstanceResult$Return": "

The return value of the request. Returns true if the specified product code is owned by the requester and associated with the specified instance.

", - "CopyImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopySnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopySnapshotRequest$Encrypted": "

Specifies whether the destination snapshot should be encrypted. There is no way to create an unencrypted snapshot copy from an encrypted snapshot; however, you can encrypt a copy of an unencrypted snapshot with this flag. The default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$NoReboot": "

By default, this parameter is set to false, which means Amazon EC2 attempts to shut down the instance cleanly before image creation and then reboots the instance. When the parameter is set to true, Amazon EC2 doesn't shut down the instance before creating the image. When this option is used, file system integrity on the created image can't be guaranteed.

", - "CreateInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$Egress": "

Indicates whether this is an egress rule (rule is applied to traffic leaving the subnet).

", - "CreateNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreatePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "CreateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$Encrypted": "

Specifies whether the volume should be encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are automatically encrypted. There is no way to create an encrypted volume from an unencrypted snapshot or vice versa. If your AMI uses encrypted volumes, you can only launch it on supported instance types. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$Egress": "

Indicates whether the rule is an egress rule.

", - "DeleteNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeletePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DeleteVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeregisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAccountAttributesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAvailabilityZonesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeBundleTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeClassicLinkInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeConversionTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeCustomerGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImagesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportImageTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportSnapshotTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$IncludeAllInstances": "

When true, includes the health status for all instances. When false, includes the health status for running instances only.

Default: false

", - "DescribeInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInternetGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeKeyPairsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeMovingAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkAclsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfacesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePlacementGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePrefixListsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRegionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$IncludeMarketplace": "

Include Marketplace offerings in the response.

", - "DescribeReservedInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRouteTablesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotPriceHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSubnetsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointServicesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcPeeringConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DetachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$Force": "

Specifies whether to force a detachment.

", - "DetachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachVolumeRequest$Force": "

Forces detachment if the previous detachment attempt did not occur cleanly (for example, logging into an instance, unmounting the volume, and detaching normally). This option can lead to data loss or a corrupted file system. Use this option only as a last resort to detach a volume from a failed instance. The instance won't have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures.

", - "DetachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisassociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisassociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EbsBlockDevice$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "EbsBlockDevice$Encrypted": "

Indicates whether the EBS volume is encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption.

", - "EbsInstanceBlockDevice$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EbsInstanceBlockDeviceSpecification$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EnableVolumeIORequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "GetConsoleOutputRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetPasswordDataRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Image$Public": "

Indicates whether the image has public launch permissions. The value is true if this image has public launch permissions or false if it has only implicit and explicit launch permissions.

", - "ImportImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportInstanceLaunchSpecification$Monitoring": "

Indicates whether monitoring is enabled.

", - "ImportInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Instance$SourceDestCheck": "

Specifies whether to enable an instance launched in a VPC to perform NAT. This controls whether source/destination checking is enabled on the instance. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "Instance$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "InstanceNetworkInterface$SourceDestCheck": "

Indicates whether to validate network traffic to or from this network interface.

", - "InstanceNetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "InstanceNetworkInterfaceSpecification$DeleteOnTermination": "

If set to true, the interface is deleted when the instance is terminated. You can specify true only if creating a new network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to an instance you launch in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "InstancePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "LaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ModifyImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$ResetPolicy": "

Specify true to reset the policy document to the default policy. The default policy allows access to the service.

", - "ModifyVpcEndpointResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "MonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MoveAddressToVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "NetworkAcl$IsDefault": "

Indicates whether this is the default network ACL for the VPC.

", - "NetworkAclEntry$Egress": "

Indicates whether the rule is an egress rule (applied to traffic leaving the subnet).

", - "NetworkInterface$RequesterManaged": "

Indicates whether the network interface is being managed by AWS.

", - "NetworkInterface$SourceDestCheck": "

Indicates whether traffic to or from the instance is validated.

", - "NetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfaceAttachmentChanges$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfacePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "PriceSchedule$Active": "

The current price schedule, as determined by the term remaining for the Reserved Instance in the listing.

A specific price schedule is always in effect, but only one price schedule can be active at any time. Take, for example, a Reserved Instance listing that has five months remaining in its term. When you specify price schedules for five months and two months, this means that schedule 1, covering the first three months of the remaining term, will be active during months 5, 4, and 3. Then schedule 2, covering the last two months of the term, will be active for months 2 and 1.

", - "PrivateIpAddressSpecification$Primary": "

Indicates whether the private IP address is the primary private IP address. Only one IP address can be designated as primary.

", - "PurchaseReservedInstancesOfferingRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RebootInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RegisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RejectVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RejectVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ReleaseAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$Egress": "

Indicates whether to replace the egress rule.

Default: If no value is specified, we replace the ingress rule.

", - "ReplaceRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceRouteTableAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReportInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotFleetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReservedInstancesOffering$Marketplace": "

Indicates whether the offering is available through the Reserved Instance Marketplace (resale) or AWS. If it's a Reserved Instance Marketplace offering, this is true.

", - "ResetImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RestoreAddressToClassicRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RouteTableAssociation$Main": "

Indicates whether this is the main route table.

", - "RunInstancesMonitoringEnabled$Enabled": "

Indicates whether monitoring is enabled for the instance.

", - "RunInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RunInstancesRequest$DisableApiTermination": "

If you set this parameter to true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. If you set this parameter to true and then later want to be able to terminate the instance, you must first change the value of the disableApiTermination attribute to false using ModifyInstanceAttribute. Alternatively, if you set InstanceInitiatedShutdownBehavior to terminate, you can terminate the instance by running the shutdown command from the instance.

Default: false

", - "RunInstancesRequest$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "Snapshot$Encrypted": "

Indicates whether the snapshot is encrypted.

", - "SpotFleetLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "SpotFleetMonitoring$Enabled": "

Enables monitoring for the instance.

Default: false

", - "SpotFleetRequestConfigData$TerminateInstancesWithExpiration": "

Indicates whether running Spot instances should be terminated when the Spot fleet request expires.

", - "StartInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$Force": "

Forces the instances to stop. The instances do not have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures. This option is not recommended for Windows instances.

Default: false

", - "Subnet$DefaultForAz": "

Indicates whether this is the default subnet for the Availability Zone.

", - "Subnet$MapPublicIpOnLaunch": "

Indicates whether instances launched in this subnet receive a public IP address.

", - "TerminateInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "UnmonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Volume$Encrypted": "

Indicates whether the volume will be encrypted.

", - "VolumeAttachment$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "Vpc$IsDefault": "

Indicates whether the VPC is the default VPC.

", - "VpcClassicLink$ClassicLinkEnabled": "

Indicates whether the VPC is enabled for ClassicLink.

", - "VpnConnectionOptions$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

", - "VpnConnectionOptionsSpecification$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

", - "RequestSpotLaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

" - } - }, - "BundleIdStringList": { - "base": null, - "refs": { - "DescribeBundleTasksRequest$BundleIds": "

One or more bundle task IDs.

Default: Describes all your bundle tasks.

" - } - }, - "BundleInstanceRequest": { - "base": null, - "refs": { - } - }, - "BundleInstanceResult": { - "base": null, - "refs": { - } - }, - "BundleTask": { - "base": "

Describes a bundle task.

", - "refs": { - "BundleInstanceResult$BundleTask": "

Information about the bundle task.

", - "BundleTaskList$member": null, - "CancelBundleTaskResult$BundleTask": "

Information about the bundle task.

" - } - }, - "BundleTaskError": { - "base": "

Describes an error for BundleInstance.

", - "refs": { - "BundleTask$BundleTaskError": "

If the task fails, a description of the error.

" - } - }, - "BundleTaskList": { - "base": null, - "refs": { - "DescribeBundleTasksResult$BundleTasks": "

Information about one or more bundle tasks.

" - } - }, - "BundleTaskState": { - "base": null, - "refs": { - "BundleTask$State": "

The state of the task.

" - } - }, - "CancelBatchErrorCode": { - "base": null, - "refs": { - "CancelSpotFleetRequestsError$Code": "

The error code.

" - } - }, - "CancelBundleTaskRequest": { - "base": null, - "refs": { - } - }, - "CancelBundleTaskResult": { - "base": null, - "refs": { - } - }, - "CancelConversionRequest": { - "base": null, - "refs": { - } - }, - "CancelExportTaskRequest": { - "base": null, - "refs": { - } - }, - "CancelImportTaskRequest": { - "base": null, - "refs": { - } - }, - "CancelImportTaskResult": { - "base": null, - "refs": { - } - }, - "CancelReservedInstancesListingRequest": { - "base": null, - "refs": { - } - }, - "CancelReservedInstancesListingResult": { - "base": null, - "refs": { - } - }, - "CancelSpotFleetRequestsError": { - "base": "

Describes a Spot fleet error.

", - "refs": { - "CancelSpotFleetRequestsErrorItem$Error": "

The error.

" - } - }, - "CancelSpotFleetRequestsErrorItem": { - "base": "

Describes a Spot fleet request that was not successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsErrorSet$member": null - } - }, - "CancelSpotFleetRequestsErrorSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$UnsuccessfulFleetRequests": "

Information about the Spot fleet requests that are not successfully canceled.

" - } - }, - "CancelSpotFleetRequestsRequest": { - "base": "

Contains the parameters for CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsResponse": { - "base": "

Contains the output of CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsSuccessItem": { - "base": "

Describes a Spot fleet request that was successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsSuccessSet$member": null - } - }, - "CancelSpotFleetRequestsSuccessSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$SuccessfulFleetRequests": "

Information about the Spot fleet requests that are successfully canceled.

" - } - }, - "CancelSpotInstanceRequestState": { - "base": null, - "refs": { - "CancelledSpotInstanceRequest$State": "

The state of the Spot instance request.

" - } - }, - "CancelSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelSpotInstanceRequestsResult": { - "base": "

Contains the output of CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelledSpotInstanceRequest": { - "base": "

Describes a request to cancel a Spot instance.

", - "refs": { - "CancelledSpotInstanceRequestList$member": null - } - }, - "CancelledSpotInstanceRequestList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsResult$CancelledSpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "ClassicLinkInstance": { - "base": "

Describes a linked EC2-Classic instance.

", - "refs": { - "ClassicLinkInstanceList$member": null - } - }, - "ClassicLinkInstanceList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesResult$Instances": "

Information about one or more linked EC2-Classic instances.

" - } - }, - "ClientData": { - "base": "

Describes the client-specific data.

", - "refs": { - "ImportImageRequest$ClientData": "

The client-specific data.

", - "ImportSnapshotRequest$ClientData": "

The client-specific data.

" - } - }, - "ConfirmProductInstanceRequest": { - "base": null, - "refs": { - } - }, - "ConfirmProductInstanceResult": { - "base": null, - "refs": { - } - }, - "ContainerFormat": { - "base": null, - "refs": { - "ExportToS3Task$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

", - "ExportToS3TaskSpecification$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

" - } - }, - "ConversionIdStringList": { - "base": null, - "refs": { - "DescribeConversionTasksRequest$ConversionTaskIds": "

One or more conversion task IDs.

" - } - }, - "ConversionTask": { - "base": "

Describes a conversion task.

", - "refs": { - "DescribeConversionTaskList$member": null, - "ImportInstanceResult$ConversionTask": "

Information about the conversion task.

", - "ImportVolumeResult$ConversionTask": "

Information about the conversion task.

" - } - }, - "ConversionTaskState": { - "base": null, - "refs": { - "ConversionTask$State": "

The state of the conversion task.

" - } - }, - "CopyImageRequest": { - "base": null, - "refs": { - } - }, - "CopyImageResult": { - "base": null, - "refs": { - } - }, - "CopySnapshotRequest": { - "base": null, - "refs": { - } - }, - "CopySnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateCustomerGatewayRequest": { - "base": null, - "refs": { - } - }, - "CreateCustomerGatewayResult": { - "base": null, - "refs": { - } - }, - "CreateDhcpOptionsRequest": { - "base": null, - "refs": { - } - }, - "CreateDhcpOptionsResult": { - "base": null, - "refs": { - } - }, - "CreateFlowLogsRequest": { - "base": null, - "refs": { - } - }, - "CreateFlowLogsResult": { - "base": null, - "refs": { - } - }, - "CreateImageRequest": { - "base": null, - "refs": { - } - }, - "CreateImageResult": { - "base": null, - "refs": { - } - }, - "CreateInstanceExportTaskRequest": { - "base": null, - "refs": { - } - }, - "CreateInstanceExportTaskResult": { - "base": null, - "refs": { - } - }, - "CreateInternetGatewayRequest": { - "base": null, - "refs": { - } - }, - "CreateInternetGatewayResult": { - "base": null, - "refs": { - } - }, - "CreateKeyPairRequest": { - "base": null, - "refs": { - } - }, - "CreateNetworkAclEntryRequest": { - "base": null, - "refs": { - } - }, - "CreateNetworkAclRequest": { - "base": null, - "refs": { - } - }, - "CreateNetworkAclResult": { - "base": null, - "refs": { - } - }, - "CreateNetworkInterfaceRequest": { - "base": null, - "refs": { - } - }, - "CreateNetworkInterfaceResult": { - "base": null, - "refs": { - } - }, - "CreatePlacementGroupRequest": { - "base": null, - "refs": { - } - }, - "CreateReservedInstancesListingRequest": { - "base": null, - "refs": { - } - }, - "CreateReservedInstancesListingResult": { - "base": null, - "refs": { - } - }, - "CreateRouteRequest": { - "base": null, - "refs": { - } - }, - "CreateRouteResult": { - "base": null, - "refs": { - } - }, - "CreateRouteTableRequest": { - "base": null, - "refs": { - } - }, - "CreateRouteTableResult": { - "base": null, - "refs": { - } - }, - "CreateSecurityGroupRequest": { - "base": null, - "refs": { - } - }, - "CreateSecurityGroupResult": { - "base": null, - "refs": { - } - }, - "CreateSnapshotRequest": { - "base": null, - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSubnetRequest": { - "base": null, - "refs": { - } - }, - "CreateSubnetResult": { - "base": null, - "refs": { - } - }, - "CreateTagsRequest": { - "base": null, - "refs": { - } - }, - "CreateVolumePermission": { - "base": "

Describes the user or group to be added or removed from the permissions for a volume.

", - "refs": { - "CreateVolumePermissionList$member": null - } - }, - "CreateVolumePermissionList": { - "base": null, - "refs": { - "CreateVolumePermissionModifications$Add": "

Adds a specific AWS account ID or group to a volume's list of create volume permissions.

", - "CreateVolumePermissionModifications$Remove": "

Removes a specific AWS account ID or group from a volume's list of create volume permissions.

", - "DescribeSnapshotAttributeResult$CreateVolumePermissions": "

A list of permissions for creating volumes from the snapshot.

" - } - }, - "CreateVolumePermissionModifications": { - "base": "

Describes modifications to the permissions for a volume.

", - "refs": { - "ModifySnapshotAttributeRequest$CreateVolumePermission": "

A JSON representation of the snapshot attribute modification.

" - } - }, - "CreateVolumeRequest": { - "base": null, - "refs": { - } - }, - "CreateVpcEndpointRequest": { - "base": null, - "refs": { - } - }, - "CreateVpcEndpointResult": { - "base": null, - "refs": { - } - }, - "CreateVpcPeeringConnectionRequest": { - "base": null, - "refs": { - } - }, - "CreateVpcPeeringConnectionResult": { - "base": null, - "refs": { - } - }, - "CreateVpcRequest": { - "base": null, - "refs": { - } - }, - "CreateVpcResult": { - "base": null, - "refs": { - } - }, - "CreateVpnConnectionRequest": { - "base": null, - "refs": { - } - }, - "CreateVpnConnectionResult": { - "base": null, - "refs": { - } - }, - "CreateVpnConnectionRouteRequest": { - "base": null, - "refs": { - } - }, - "CreateVpnGatewayRequest": { - "base": null, - "refs": { - } - }, - "CreateVpnGatewayResult": { - "base": null, - "refs": { - } - }, - "CurrencyCodeValues": { - "base": null, - "refs": { - "PriceSchedule$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "PriceScheduleSpecification$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "ReservedInstanceLimitPrice$CurrencyCode": "

The currency in which the limitPrice amount is specified. At this time, the only supported currency is USD.

", - "ReservedInstances$CurrencyCode": "

The currency of the Reserved Instance. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

", - "ReservedInstancesOffering$CurrencyCode": "

The currency of the Reserved Instance offering you are purchasing. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

" - } - }, - "CustomerGateway": { - "base": "

Describes a customer gateway.

", - "refs": { - "CreateCustomerGatewayResult$CustomerGateway": "

Information about the customer gateway.

", - "CustomerGatewayList$member": null - } - }, - "CustomerGatewayIdStringList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysRequest$CustomerGatewayIds": "

One or more customer gateway IDs.

Default: Describes all your customer gateways.

" - } - }, - "CustomerGatewayList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysResult$CustomerGateways": "

Information about one or more customer gateways.

" - } - }, - "DatafeedSubscriptionState": { - "base": null, - "refs": { - "SpotDatafeedSubscription$State": "

The state of the Spot instance data feed subscription.

" - } - }, - "DateTime": { - "base": null, - "refs": { - "BundleTask$StartTime": "

The time this task started.

", - "BundleTask$UpdateTime": "

The time of the most recent update for the task.

", - "ClientData$UploadStart": "

The time that the disk upload starts.

", - "ClientData$UploadEnd": "

The time that the disk upload ends.

", - "DescribeSpotFleetRequestHistoryRequest$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$LastEvaluatedTime": "

The last date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). All records up to this time were retrieved.

If nextToken indicates that there are more results, this value is not present.

", - "DescribeSpotPriceHistoryRequest$StartTime": "

The date and time, up to the past 90 days, from which to start retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotPriceHistoryRequest$EndTime": "

The date and time, up to the current date, from which to stop retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "EbsInstanceBlockDevice$AttachTime": "

The time stamp when the attachment initiated.

", - "FlowLog$CreationTime": "

The date and time the flow log was created.

", - "GetConsoleOutputResult$Timestamp": "

The time the output was last updated.

", - "GetPasswordDataResult$Timestamp": "

The time the data was last updated.

", - "HistoryRecord$Timestamp": "

The date and time of the event, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "Instance$LaunchTime": "

The time the instance was launched.

", - "InstanceNetworkInterfaceAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "InstanceStatusDetails$ImpairedSince": "

The time when a status check failed. For an instance that was launched and impaired, this is the time when the instance was launched.

", - "InstanceStatusEvent$NotBefore": "

The earliest scheduled start time for the event.

", - "InstanceStatusEvent$NotAfter": "

The latest scheduled end time for the event.

", - "NetworkInterfaceAttachment$AttachTime": "

The timestamp indicating when the attachment initiated.

", - "ReportInstanceStatusRequest$StartTime": "

The time at which the reported instance health state began.

", - "ReportInstanceStatusRequest$EndTime": "

The time at which the reported instance health state ended.

", - "RequestSpotInstancesRequest$ValidFrom": "

The start date of the request. If this is a one-time request, the request becomes active at this date and time and remains active until all instances launch, the request expires, or the request is canceled. If the request is persistent, the request becomes active at this date and time and remains active until it expires or is canceled.

Default: The request is effective indefinitely.

", - "RequestSpotInstancesRequest$ValidUntil": "

The end date of the request. If this is a one-time request, the request remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date and time is reached.

Default: The request is effective indefinitely.

", - "ReservedInstances$Start": "

The date and time the Reserved Instance started.

", - "ReservedInstances$End": "

The time when the Reserved Instance expires.

", - "ReservedInstancesListing$CreateDate": "

The time the listing was created.

", - "ReservedInstancesListing$UpdateDate": "

The last modified timestamp of the listing.

", - "ReservedInstancesModification$CreateDate": "

The time when the modification request was created.

", - "ReservedInstancesModification$UpdateDate": "

The time when the modification request was last updated.

", - "ReservedInstancesModification$EffectiveDate": "

The time for the modification to become effective.

", - "Snapshot$StartTime": "

The time stamp when the snapshot was initiated.

", - "SpotFleetRequestConfigData$ValidFrom": "

The start date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The default is to start fulfilling the request immediately.

", - "SpotFleetRequestConfigData$ValidUntil": "

The end date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). At this point, no new Spot instance requests are placed or enabled to fulfill the request.

", - "SpotInstanceRequest$ValidFrom": "

The start date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). If this is a one-time request, the request becomes active at this date and time and remains active until all instances launch, the request expires, or the request is canceled. If the request is persistent, the request becomes active at this date and time and remains active until it expires or is canceled.

", - "SpotInstanceRequest$ValidUntil": "

The end date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). If this is a one-time request, the request remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date is reached.

", - "SpotInstanceRequest$CreateTime": "

The date and time when the Spot instance request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotInstanceStatus$UpdateTime": "

The date and time of the most recent status update, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotPrice$Timestamp": "

The date and time the request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "VgwTelemetry$LastStatusChange": "

The date and time of the last change in status.

", - "Volume$CreateTime": "

The time stamp when volume creation was initiated.

", - "VolumeAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "VolumeStatusEvent$NotBefore": "

The earliest start time of the event.

", - "VolumeStatusEvent$NotAfter": "

The latest end time of the event.

", - "VpcEndpoint$CreationTimestamp": "

The date and time the VPC endpoint was created.

", - "VpcPeeringConnection$ExpirationTime": "

The time that an unaccepted VPC peering connection will expire.

" - } - }, - "DeleteCustomerGatewayRequest": { - "base": null, - "refs": { - } - }, - "DeleteDhcpOptionsRequest": { - "base": null, - "refs": { - } - }, - "DeleteFlowLogsRequest": { - "base": null, - "refs": { - } - }, - "DeleteFlowLogsResult": { - "base": null, - "refs": { - } - }, - "DeleteInternetGatewayRequest": { - "base": null, - "refs": { - } - }, - "DeleteKeyPairRequest": { - "base": null, - "refs": { - } - }, - "DeleteNetworkAclEntryRequest": { - "base": null, - "refs": { - } - }, - "DeleteNetworkAclRequest": { - "base": null, - "refs": { - } - }, - "DeleteNetworkInterfaceRequest": { - "base": null, - "refs": { - } - }, - "DeletePlacementGroupRequest": { - "base": null, - "refs": { - } - }, - "DeleteRouteRequest": { - "base": null, - "refs": { - } - }, - "DeleteRouteTableRequest": { - "base": null, - "refs": { - } - }, - "DeleteSecurityGroupRequest": { - "base": null, - "refs": { - } - }, - "DeleteSnapshotRequest": { - "base": null, - "refs": { - } - }, - "DeleteSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DeleteSpotDatafeedSubscription.

", - "refs": { - } - }, - "DeleteSubnetRequest": { - "base": null, - "refs": { - } - }, - "DeleteTagsRequest": { - "base": null, - "refs": { - } - }, - "DeleteVolumeRequest": { - "base": null, - "refs": { - } - }, - "DeleteVpcEndpointsRequest": { - "base": null, - "refs": { - } - }, - "DeleteVpcEndpointsResult": { - "base": null, - "refs": { - } - }, - "DeleteVpcPeeringConnectionRequest": { - "base": null, - "refs": { - } - }, - "DeleteVpcPeeringConnectionResult": { - "base": null, - "refs": { - } - }, - "DeleteVpcRequest": { - "base": null, - "refs": { - } - }, - "DeleteVpnConnectionRequest": { - "base": null, - "refs": { - } - }, - "DeleteVpnConnectionRouteRequest": { - "base": null, - "refs": { - } - }, - "DeleteVpnGatewayRequest": { - "base": null, - "refs": { - } - }, - "DeregisterImageRequest": { - "base": null, - "refs": { - } - }, - "DescribeAccountAttributesRequest": { - "base": null, - "refs": { - } - }, - "DescribeAccountAttributesResult": { - "base": null, - "refs": { - } - }, - "DescribeAddressesRequest": { - "base": null, - "refs": { - } - }, - "DescribeAddressesResult": { - "base": null, - "refs": { - } - }, - "DescribeAvailabilityZonesRequest": { - "base": null, - "refs": { - } - }, - "DescribeAvailabilityZonesResult": { - "base": null, - "refs": { - } - }, - "DescribeBundleTasksRequest": { - "base": null, - "refs": { - } - }, - "DescribeBundleTasksResult": { - "base": null, - "refs": { - } - }, - "DescribeClassicLinkInstancesRequest": { - "base": null, - "refs": { - } - }, - "DescribeClassicLinkInstancesResult": { - "base": null, - "refs": { - } - }, - "DescribeConversionTaskList": { - "base": null, - "refs": { - "DescribeConversionTasksResult$ConversionTasks": "

Information about the conversion tasks.

" - } - }, - "DescribeConversionTasksRequest": { - "base": null, - "refs": { - } - }, - "DescribeConversionTasksResult": { - "base": null, - "refs": { - } - }, - "DescribeCustomerGatewaysRequest": { - "base": null, - "refs": { - } - }, - "DescribeCustomerGatewaysResult": { - "base": null, - "refs": { - } - }, - "DescribeDhcpOptionsRequest": { - "base": null, - "refs": { - } - }, - "DescribeDhcpOptionsResult": { - "base": null, - "refs": { - } - }, - "DescribeExportTasksRequest": { - "base": null, - "refs": { - } - }, - "DescribeExportTasksResult": { - "base": null, - "refs": { - } - }, - "DescribeFlowLogsRequest": { - "base": null, - "refs": { - } - }, - "DescribeFlowLogsResult": { - "base": null, - "refs": { - } - }, - "DescribeImageAttributeRequest": { - "base": null, - "refs": { - } - }, - "DescribeImagesRequest": { - "base": null, - "refs": { - } - }, - "DescribeImagesResult": { - "base": null, - "refs": { - } - }, - "DescribeImportImageTasksRequest": { - "base": null, - "refs": { - } - }, - "DescribeImportImageTasksResult": { - "base": null, - "refs": { - } - }, - "DescribeImportSnapshotTasksRequest": { - "base": null, - "refs": { - } - }, - "DescribeImportSnapshotTasksResult": { - "base": null, - "refs": { - } - }, - "DescribeInstanceAttributeRequest": { - "base": null, - "refs": { - } - }, - "DescribeInstanceStatusRequest": { - "base": null, - "refs": { - } - }, - "DescribeInstanceStatusResult": { - "base": null, - "refs": { - } - }, - "DescribeInstancesRequest": { - "base": null, - "refs": { - } - }, - "DescribeInstancesResult": { - "base": null, - "refs": { - } - }, - "DescribeInternetGatewaysRequest": { - "base": null, - "refs": { - } - }, - "DescribeInternetGatewaysResult": { - "base": null, - "refs": { - } - }, - "DescribeKeyPairsRequest": { - "base": null, - "refs": { - } - }, - "DescribeKeyPairsResult": { - "base": null, - "refs": { - } - }, - "DescribeMovingAddressesRequest": { - "base": null, - "refs": { - } - }, - "DescribeMovingAddressesResult": { - "base": null, - "refs": { - } - }, - "DescribeNetworkAclsRequest": { - "base": null, - "refs": { - } - }, - "DescribeNetworkAclsResult": { - "base": null, - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeRequest": { - "base": null, - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeResult": { - "base": null, - "refs": { - } - }, - "DescribeNetworkInterfacesRequest": { - "base": null, - "refs": { - } - }, - "DescribeNetworkInterfacesResult": { - "base": null, - "refs": { - } - }, - "DescribePlacementGroupsRequest": { - "base": null, - "refs": { - } - }, - "DescribePlacementGroupsResult": { - "base": null, - "refs": { - } - }, - "DescribePrefixListsRequest": { - "base": null, - "refs": { - } - }, - "DescribePrefixListsResult": { - "base": null, - "refs": { - } - }, - "DescribeRegionsRequest": { - "base": null, - "refs": { - } - }, - "DescribeRegionsResult": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesListingsRequest": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesListingsResult": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesModificationsRequest": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesModificationsResult": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesOfferingsRequest": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesOfferingsResult": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesRequest": { - "base": null, - "refs": { - } - }, - "DescribeReservedInstancesResult": { - "base": null, - "refs": { - } - }, - "DescribeRouteTablesRequest": { - "base": null, - "refs": { - } - }, - "DescribeRouteTablesResult": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupsRequest": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupsResult": { - "base": null, - "refs": { - } - }, - "DescribeSnapshotAttributeRequest": { - "base": null, - "refs": { - } - }, - "DescribeSnapshotAttributeResult": { - "base": null, - "refs": { - } - }, - "DescribeSnapshotsRequest": { - "base": null, - "refs": { - } - }, - "DescribeSnapshotsResult": { - "base": null, - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesRequest": { - "base": "

Contains the parameters for DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesResponse": { - "base": "

Contains the output of DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryResponse": { - "base": "

Contains the output of DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsResponse": { - "base": "

Contains the output of DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsResult": { - "base": "

Contains the output of DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryResult": { - "base": "

Contains the output of DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeSubnetsRequest": { - "base": null, - "refs": { - } - }, - "DescribeSubnetsResult": { - "base": null, - "refs": { - } - }, - "DescribeTagsRequest": { - "base": null, - "refs": { - } - }, - "DescribeTagsResult": { - "base": null, - "refs": { - } - }, - "DescribeVolumeAttributeRequest": { - "base": null, - "refs": { - } - }, - "DescribeVolumeAttributeResult": { - "base": null, - "refs": { - } - }, - "DescribeVolumeStatusRequest": { - "base": null, - "refs": { - } - }, - "DescribeVolumeStatusResult": { - "base": null, - "refs": { - } - }, - "DescribeVolumesRequest": { - "base": null, - "refs": { - } - }, - "DescribeVolumesResult": { - "base": null, - "refs": { - } - }, - "DescribeVpcAttributeRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpcAttributeResult": { - "base": null, - "refs": { - } - }, - "DescribeVpcClassicLinkRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpcClassicLinkResult": { - "base": null, - "refs": { - } - }, - "DescribeVpcEndpointServicesRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpcEndpointServicesResult": { - "base": null, - "refs": { - } - }, - "DescribeVpcEndpointsRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpcEndpointsResult": { - "base": null, - "refs": { - } - }, - "DescribeVpcPeeringConnectionsRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpcPeeringConnectionsResult": { - "base": null, - "refs": { - } - }, - "DescribeVpcsRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpcsResult": { - "base": null, - "refs": { - } - }, - "DescribeVpnConnectionsRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpnConnectionsResult": { - "base": null, - "refs": { - } - }, - "DescribeVpnGatewaysRequest": { - "base": null, - "refs": { - } - }, - "DescribeVpnGatewaysResult": { - "base": null, - "refs": { - } - }, - "DetachClassicLinkVpcRequest": { - "base": null, - "refs": { - } - }, - "DetachClassicLinkVpcResult": { - "base": null, - "refs": { - } - }, - "DetachInternetGatewayRequest": { - "base": null, - "refs": { - } - }, - "DetachNetworkInterfaceRequest": { - "base": null, - "refs": { - } - }, - "DetachVolumeRequest": { - "base": null, - "refs": { - } - }, - "DetachVpnGatewayRequest": { - "base": null, - "refs": { - } - }, - "DeviceType": { - "base": null, - "refs": { - "Image$RootDeviceType": "

The type of root device used by the AMI. The AMI can use an EBS volume or an instance store volume.

", - "Instance$RootDeviceType": "

The root device type used by the AMI. The AMI can use an EBS volume or an instance store volume.

" - } - }, - "DhcpConfiguration": { - "base": "

Describes a DHCP configuration option.

", - "refs": { - "DhcpConfigurationList$member": null - } - }, - "DhcpConfigurationList": { - "base": null, - "refs": { - "DhcpOptions$DhcpConfigurations": "

One or more DHCP options in the set.

" - } - }, - "DhcpOptions": { - "base": "

Describes a set of DHCP options.

", - "refs": { - "CreateDhcpOptionsResult$DhcpOptions": "

A set of DHCP options.

", - "DhcpOptionsList$member": null - } - }, - "DhcpOptionsIdStringList": { - "base": null, - "refs": { - "DescribeDhcpOptionsRequest$DhcpOptionsIds": "

The IDs of one or more DHCP options sets.

Default: Describes all your DHCP options sets.

" - } - }, - "DhcpOptionsList": { - "base": null, - "refs": { - "DescribeDhcpOptionsResult$DhcpOptions": "

Information about one or more DHCP options sets.

" - } - }, - "DisableVgwRoutePropagationRequest": { - "base": null, - "refs": { - } - }, - "DisableVpcClassicLinkRequest": { - "base": null, - "refs": { - } - }, - "DisableVpcClassicLinkResult": { - "base": null, - "refs": { - } - }, - "DisassociateAddressRequest": { - "base": null, - "refs": { - } - }, - "DisassociateRouteTableRequest": { - "base": null, - "refs": { - } - }, - "DiskImage": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImageList$member": null - } - }, - "DiskImageDescription": { - "base": "

Describes a disk image.

", - "refs": { - "ImportInstanceVolumeDetailItem$Image": "

The image.

", - "ImportVolumeTaskDetails$Image": "

The image.

" - } - }, - "DiskImageDetail": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImage$Image": "

Information about the disk image.

", - "ImportVolumeRequest$Image": "

The disk image.

" - } - }, - "DiskImageFormat": { - "base": null, - "refs": { - "DiskImageDescription$Format": "

The disk image format.

", - "DiskImageDetail$Format": "

The disk image format.

", - "ExportToS3Task$DiskImageFormat": "

The format for the exported image.

", - "ExportToS3TaskSpecification$DiskImageFormat": "

The format for the exported image.

" - } - }, - "DiskImageList": { - "base": null, - "refs": { - "ImportInstanceRequest$DiskImages": "

The disk image.

" - } - }, - "DiskImageVolumeDescription": { - "base": "

Describes a disk image volume.

", - "refs": { - "ImportInstanceVolumeDetailItem$Volume": "

The volume.

", - "ImportVolumeTaskDetails$Volume": "

The volume.

" - } - }, - "DomainType": { - "base": null, - "refs": { - "Address$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

", - "AllocateAddressRequest$Domain": "

Set to vpc to allocate the address for use with instances in a VPC.

Default: The address is for use with instances in EC2-Classic.

", - "AllocateAddressResult$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

" - } - }, - "Double": { - "base": null, - "refs": { - "ClientData$UploadSize": "

The size of the uploaded disk image, in GiB.

", - "PriceSchedule$Price": "

The fixed price for the term.

", - "PriceScheduleSpecification$Price": "

The fixed price for the term.

", - "PricingDetail$Price": "

The price per instance.

", - "RecurringCharge$Amount": "

The amount of the recurring charge.

", - "ReservedInstanceLimitPrice$Amount": "

Used for Reserved Instance Marketplace offerings. Specifies the limit price on the total order (instanceCount * price).

", - "SnapshotDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SnapshotTaskDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SpotFleetLaunchSpecification$WeightedCapacity": "

The number of units provided by the specified instance type. These are the same units that you chose to set the target capacity in terms (instances or a performance characteristic such as vCPUs, memory, or I/O).

If the target capacity divided by this value is not a whole number, we round the number of instances to the next whole number. If this value is not specified, the default is 1.

" - } - }, - "EbsBlockDevice": { - "base": "

Describes a block device for an EBS volume.

", - "refs": { - "BlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDevice": { - "base": "

Describes a parameter used to set up an EBS volume in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDeviceSpecification": { - "base": null, - "refs": { - "InstanceBlockDeviceMappingSpecification$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EnableVgwRoutePropagationRequest": { - "base": null, - "refs": { - } - }, - "EnableVolumeIORequest": { - "base": null, - "refs": { - } - }, - "EnableVpcClassicLinkRequest": { - "base": null, - "refs": { - } - }, - "EnableVpcClassicLinkResult": { - "base": null, - "refs": { - } - }, - "EventCode": { - "base": null, - "refs": { - "InstanceStatusEvent$Code": "

The event code.

" - } - }, - "EventInformation": { - "base": "

Describes a Spot fleet event.

", - "refs": { - "HistoryRecord$EventInformation": "

Information about the event.

" - } - }, - "EventType": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryRequest$EventType": "

The type of events to describe. By default, all events are described.

", - "HistoryRecord$EventType": "

The event type.

  • error - Indicates an error with the Spot fleet request.

  • fleetRequestChange - Indicates a change in the status or configuration of the Spot fleet request.

  • instanceChange - Indicates that an instance was launched or terminated.

" - } - }, - "ExecutableByStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ExecutableUsers": "

Scopes the images by users with explicit launch permissions. Specify an AWS account ID, self (the sender of the request), or all (public AMIs).

" - } - }, - "ExportEnvironment": { - "base": null, - "refs": { - "CreateInstanceExportTaskRequest$TargetEnvironment": "

The target virtualization environment.

", - "InstanceExportDetails$TargetEnvironment": "

The target virtualization environment.

" - } - }, - "ExportTask": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskResult$ExportTask": "

Information about the instance export task.

", - "ExportTaskList$member": null - } - }, - "ExportTaskIdStringList": { - "base": null, - "refs": { - "DescribeExportTasksRequest$ExportTaskIds": "

One or more export task IDs.

" - } - }, - "ExportTaskList": { - "base": null, - "refs": { - "DescribeExportTasksResult$ExportTasks": "

Information about the export tasks.

" - } - }, - "ExportTaskState": { - "base": null, - "refs": { - "ExportTask$State": "

The state of the export task.

" - } - }, - "ExportToS3Task": { - "base": "

Describes the format and location for an instance export task.

", - "refs": { - "ExportTask$ExportToS3Task": "

Information about the export task.

" - } - }, - "ExportToS3TaskSpecification": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskRequest$ExportToS3Task": "

The format and location for an instance export task.

" - } - }, - "Filter": { - "base": "

A filter name and value pair that is used to return a more specific list of results. Filters can be used to match a set of resources by various criteria, such as tags, attributes, or IDs.

", - "refs": { - "FilterList$member": null - } - }, - "FilterList": { - "base": null, - "refs": { - "DescribeAddressesRequest$Filters": "

One or more filters. Filter names and values are case-sensitive.

  • allocation-id - [EC2-VPC] The allocation ID for the address.

  • association-id - [EC2-VPC] The association ID for the address.

  • domain - Indicates whether the address is for use in EC2-Classic (standard) or in a VPC (vpc).

  • instance-id - The ID of the instance the address is associated with, if any.

  • network-interface-id - [EC2-VPC] The ID of the network interface that the address is associated with, if any.

  • network-interface-owner-id - The AWS account ID of the owner.

  • private-ip-address - [EC2-VPC] The private IP address associated with the Elastic IP address.

  • public-ip - The Elastic IP address.

", - "DescribeAvailabilityZonesRequest$Filters": "

One or more filters.

  • message - Information about the Availability Zone.

  • region-name - The name of the region for the Availability Zone (for example, us-east-1).

  • state - The state of the Availability Zone (available | impaired | unavailable).

  • zone-name - The name of the Availability Zone (for example, us-east-1a).

", - "DescribeBundleTasksRequest$Filters": "

One or more filters.

  • bundle-id - The ID of the bundle task.

  • error-code - If the task failed, the error code returned.

  • error-message - If the task failed, the error message returned.

  • instance-id - The ID of the instance.

  • progress - The level of task completion, as a percentage (for example, 20%).

  • s3-bucket - The Amazon S3 bucket to store the AMI.

  • s3-prefix - The beginning of the AMI name.

  • start-time - The time the task started (for example, 2013-09-15T17:15:20.000Z).

  • state - The state of the task (pending | waiting-for-shutdown | bundling | storing | cancelling | complete | failed).

  • update-time - The time of the most recent update for the task.

", - "DescribeClassicLinkInstancesRequest$Filters": "

One or more filters.

  • group-id - The ID of a VPC security group that's associated with the instance.

  • instance-id - The ID of the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC that the instance is linked to.

", - "DescribeConversionTasksRequest$Filters": "

One or more filters.

", - "DescribeCustomerGatewaysRequest$Filters": "

One or more filters.

  • bgp-asn - The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

  • customer-gateway-id - The ID of the customer gateway.

  • ip-address - The IP address of the customer gateway's Internet-routable external interface.

  • state - The state of the customer gateway (pending | available | deleting | deleted).

  • type - The type of customer gateway. Currently, the only supported type is ipsec.1.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeDhcpOptionsRequest$Filters": "

One or more filters.

  • dhcp-options-id - The ID of a set of DHCP options.

  • key - The key for one of the options (for example, domain-name).

  • value - The value for one of the options.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeFlowLogsRequest$Filter": "

One or more filters.

  • deliver-log-status - The status of the logs delivery (SUCCESS | FAILED).

  • flow-log-id - The ID of the flow log.

  • log-group-name - The name of the log group.

  • resource-id - The ID of the VPC, subnet, or network interface.

  • traffic-type - The type of traffic (ACCEPT | REJECT | ALL)

", - "DescribeImagesRequest$Filters": "

One or more filters.

  • architecture - The image architecture (i386 | x86_64).

  • block-device-mapping.delete-on-termination - A Boolean value that indicates whether the Amazon EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh).

  • block-device-mapping.snapshot-id - The ID of the snapshot used for the EBS volume.

  • block-device-mapping.volume-size - The volume size of the EBS volume, in GiB.

  • block-device-mapping.volume-type - The volume type of the EBS volume (gp2 | standard | io1).

  • description - The description of the image (provided during image creation).

  • hypervisor - The hypervisor type (ovm | xen).

  • image-id - The ID of the image.

  • image-type - The image type (machine | kernel | ramdisk).

  • is-public - A Boolean that indicates whether the image is public.

  • kernel-id - The kernel ID.

  • manifest-location - The location of the image manifest.

  • name - The name of the AMI (provided during image creation).

  • owner-alias - The AWS account alias (for example, amazon).

  • owner-id - The AWS account ID of the image owner.

  • platform - The platform. To only list Windows-based AMIs, use windows.

  • product-code - The product code.

  • product-code.type - The type of the product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • root-device-name - The name of the root device volume (for example, /dev/sda1).

  • root-device-type - The type of the root device volume (ebs | instance-store).

  • state - The state of the image (available | pending | failed).

  • state-reason-code - The reason code for the state change.

  • state-reason-message - The message for the state change.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • virtualization-type - The virtualization type (paravirtual | hvm).

", - "DescribeImportImageTasksRequest$Filters": "

One or more filters.

", - "DescribeImportSnapshotTasksRequest$Filters": "

One or more filters.

", - "DescribeInstanceStatusRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone of the instance.

  • event.code - The code for the scheduled event (instance-reboot | system-reboot | system-maintenance | instance-retirement | instance-stop).

  • event.description - A description of the event.

  • event.not-after - The latest end time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • event.not-before - The earliest start time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • instance-state-code - The code for the instance state, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-status.reachability - Filters on instance status where the name is reachability (passed | failed | initializing | insufficient-data).

  • instance-status.status - The status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

  • system-status.reachability - Filters on system status where the name is reachability (passed | failed | initializing | insufficient-data).

  • system-status.status - The system status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

", - "DescribeInstancesRequest$Filters": "

One or more filters.

  • architecture - The instance architecture (i386 | x86_64).

  • availability-zone - The Availability Zone of the instance.

  • block-device-mapping.attach-time - The attach time for an EBS volume mapped to the instance, for example, 2010-09-15T17:15:20.000Z.

  • block-device-mapping.delete-on-termination - A Boolean that indicates whether the EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh or xvdh).

  • block-device-mapping.status - The status for the EBS volume (attaching | attached | detaching | detached).

  • block-device-mapping.volume-id - The volume ID of the EBS volume.

  • client-token - The idempotency token you provided when you launched the instance.

  • dns-name - The public DNS name of the instance.

  • group-id - The ID of the security group for the instance. EC2-Classic only.

  • group-name - The name of the security group for the instance. EC2-Classic only.

  • hypervisor - The hypervisor type of the instance (ovm | xen).

  • iam-instance-profile.arn - The instance profile associated with the instance. Specified as an ARN.

  • image-id - The ID of the image used to launch the instance.

  • instance-id - The ID of the instance.

  • instance-lifecycle - Indicates whether this is a Spot Instance (spot).

  • instance-state-code - The state of the instance, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are: 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-type - The type of instance (for example, t2.micro).

  • instance.group-id - The ID of the security group for the instance.

  • instance.group-name - The name of the security group for the instance.

  • ip-address - The public IP address of the instance.

  • kernel-id - The kernel ID.

  • key-name - The name of the key pair used when the instance was launched.

  • launch-index - When launching multiple instances, this is the index for the instance in the launch group (for example, 0, 1, 2, and so on).

  • launch-time - The time when the instance was launched.

  • monitoring-state - Indicates whether monitoring is enabled for the instance (disabled | enabled).

  • owner-id - The AWS account ID of the instance owner.

  • placement-group-name - The name of the placement group for the instance.

  • platform - The platform. Use windows if you have Windows instances; otherwise, leave blank.

  • private-dns-name - The private DNS name of the instance.

  • private-ip-address - The private IP address of the instance.

  • product-code - The product code associated with the AMI used to launch the instance.

  • product-code.type - The type of product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • reason - The reason for the current state of the instance (for example, shows \"User Initiated [date]\" when you stop or terminate the instance). Similar to the state-reason-code filter.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • reservation-id - The ID of the instance's reservation. A reservation ID is created any time you launch an instance. A reservation ID has a one-to-one relationship with an instance launch request, but can be associated with more than one instance if you launch multiple instances using the same launch request. For example, if you launch one instance, you'll get one reservation ID. If you launch ten instances using the same launch request, you'll also get one reservation ID.

  • root-device-name - The name of the root device for the instance (for example, /dev/sda1 or /dev/xvda).

  • root-device-type - The type of root device that the instance uses (ebs | instance-store).

  • source-dest-check - Indicates whether the instance performs source/destination checking. A value of true means that checking is enabled, and false means checking is disabled. The value must be false for the instance to perform network address translation (NAT) in your VPC.

  • spot-instance-request-id - The ID of the Spot Instance request.

  • state-reason-code - The reason code for the state change.

  • state-reason-message - A message that describes the state change.

  • subnet-id - The ID of the subnet for the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource, where tag:key is the tag's key.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • tenancy - The tenancy of an instance (dedicated | default).

  • virtualization-type - The virtualization type of the instance (paravirtual | hvm).

  • vpc-id - The ID of the VPC that the instance is running in.

  • network-interface.description - The description of the network interface.

  • network-interface.subnet-id - The ID of the subnet for the network interface.

  • network-interface.vpc-id - The ID of the VPC for the network interface.

  • network-interface.network-interface.id - The ID of the network interface.

  • network-interface.owner-id - The ID of the owner of the network interface.

  • network-interface.availability-zone - The Availability Zone for the network interface.

  • network-interface.requester-id - The requester ID for the network interface.

  • network-interface.requester-managed - Indicates whether the network interface is being managed by AWS.

  • network-interface.status - The status of the network interface (available) | in-use).

  • network-interface.mac-address - The MAC address of the network interface.

  • network-interface-private-dns-name - The private DNS name of the network interface.

  • network-interface.source-dest-check - Whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • network-interface.group-id - The ID of a security group associated with the network interface.

  • network-interface.group-name - The name of a security group associated with the network interface.

  • network-interface.attachment.attachment-id - The ID of the interface attachment.

  • network-interface.attachment.instance-id - The ID of the instance to which the network interface is attached.

  • network-interface.attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • network-interface.addresses.private-ip-address - The private IP address associated with the network interface.

  • network-interface.attachment.device-index - The device index to which the network interface is attached.

  • network-interface.attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • network-interface.attachment.attach-time - The time that the network interface was attached to an instance.

  • network-interface.attachment.delete-on-termination - Specifies whether the attachment is deleted when an instance is terminated.

  • network-interface.addresses.primary - Specifies whether the IP address of the network interface is the primary private IP address.

  • network-interface.addresses.association.public-ip - The ID of the association of an Elastic IP address with a network interface.

  • network-interface.addresses.association.ip-owner-id - The owner ID of the private IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

", - "DescribeInternetGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (available). Present only if a VPC is attached.

  • attachment.vpc-id - The ID of an attached VPC.

  • internet-gateway-id - The ID of the Internet gateway.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeKeyPairsRequest$Filters": "

One or more filters.

  • fingerprint - The fingerprint of the key pair.

  • key-name - The name of the key pair.

", - "DescribeMovingAddressesRequest$Filters": "

One or more filters.

  • moving-status - The status of the Elastic IP address (MovingToVpc | RestoringToClassic).

", - "DescribeNetworkAclsRequest$Filters": "

One or more filters.

  • association.association-id - The ID of an association ID for the ACL.

  • association.network-acl-id - The ID of the network ACL involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • default - Indicates whether the ACL is the default network ACL for the VPC.

  • entry.cidr - The CIDR range specified in the entry.

  • entry.egress - Indicates whether the entry applies to egress traffic.

  • entry.icmp.code - The ICMP code specified in the entry, if any.

  • entry.icmp.type - The ICMP type specified in the entry, if any.

  • entry.port-range.from - The start of the port range specified in the entry.

  • entry.port-range.to - The end of the port range specified in the entry.

  • entry.protocol - The protocol specified in the entry (tcp | udp | icmp or a protocol number).

  • entry.rule-action - Allows or denies the matching traffic (allow | deny).

  • entry.rule-number - The number of an entry (in other words, rule) in the ACL's set of entries.

  • network-acl-id - The ID of the network ACL.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network ACL.

", - "DescribeNetworkInterfacesRequest$Filters": "

One or more filters.

  • addresses.private-ip-address - The private IP addresses associated with the network interface.

  • addresses.primary - Whether the private IP address is the primary IP address associated with the network interface.

  • addresses.association.public-ip - The association ID returned when the network interface was associated with the Elastic IP address.

  • addresses.association.owner-id - The owner ID of the addresses associated with the network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.public-dns-name - The public DNS name for the network interface.

  • attachment.attachment-id - The ID of the interface attachment.

  • attachment.instance-id - The ID of the instance to which the network interface is attached.

  • attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • attachment.device-index - The device index to which the network interface is attached.

  • attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • attachment.attach.time - The time that the network interface was attached to an instance.

  • attachment.delete-on-termination - Indicates whether the attachment is deleted when an instance is terminated.

  • availability-zone - The Availability Zone of the network interface.

  • description - The description of the network interface.

  • group-id - The ID of a security group associated with the network interface.

  • group-name - The name of a security group associated with the network interface.

  • mac-address - The MAC address of the network interface.

  • network-interface-id - The ID of the network interface.

  • owner-id - The AWS account ID of the network interface owner.

  • private-ip-address - The private IP address or addresses of the network interface.

  • private-dns-name - The private DNS name of the network interface.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • requester-managed - Indicates whether the network interface is being managed by an AWS service (for example, AWS Management Console, Auto Scaling, and so on).

  • source-desk-check - Indicates whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform Network Address Translation (NAT) in your VPC.

  • status - The status of the network interface. If the network interface is not attached to an instance, the status is available; if a network interface is attached to an instance the status is in-use.

  • subnet-id - The ID of the subnet for the network interface.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network interface.

", - "DescribePlacementGroupsRequest$Filters": "

One or more filters.

  • group-name - The name of the placement group.

  • state - The state of the placement group (pending | available | deleting | deleted).

  • strategy - The strategy of the placement group (cluster).

", - "DescribePrefixListsRequest$Filters": "

One or more filters.

  • prefix-list-id: The ID of a prefix list.

  • prefix-list-name: The name of a prefix list.

", - "DescribeRegionsRequest$Filters": "

One or more filters.

  • endpoint - The endpoint of the region (for example, ec2.us-east-1.amazonaws.com).

  • region-name - The name of the region (for example, us-east-1).

", - "DescribeReservedInstancesListingsRequest$Filters": "

One or more filters.

  • reserved-instances-id - The ID of the Reserved Instances.

  • reserved-instances-listing-id - The ID of the Reserved Instances listing.

  • status - The status of the Reserved Instance listing (pending | active | cancelled | closed).

  • status-message - The reason for the status.

", - "DescribeReservedInstancesModificationsRequest$Filters": "

One or more filters.

  • client-token - The idempotency token for the modification request.

  • create-date - The time when the modification request was created.

  • effective-date - The time when the modification becomes effective.

  • modification-result.reserved-instances-id - The ID for the Reserved Instances created as part of the modification request. This ID is only available when the status of the modification is fulfilled.

  • modification-result.target-configuration.availability-zone - The Availability Zone for the new Reserved Instances.

  • modification-result.target-configuration.instance-count - The number of new Reserved Instances.

  • modification-result.target-configuration.instance-type - The instance type of the new Reserved Instances.

  • modification-result.target-configuration.platform - The network platform of the new Reserved Instances (EC2-Classic | EC2-VPC).

  • reserved-instances-id - The ID of the Reserved Instances modified.

  • reserved-instances-modification-id - The ID of the modification request.

  • status - The status of the Reserved Instances modification request (processing | fulfilled | failed).

  • status-message - The reason for the status.

  • update-date - The time when the modification request was last updated.

", - "DescribeReservedInstancesOfferingsRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (for example, one year or three years), in seconds (31536000 | 94608000).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type on which the Reserved Instance can be used.

  • marketplace - Set to true to show only Reserved Instance Marketplace offerings. When this filter is not used, which is the default behavior, all offerings from AWS and Reserved Instance Marketplace are listed.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC. (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC))

  • reserved-instances-offering-id - The Reserved Instances offering ID.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeReservedInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (one year or three years), in seconds (31536000 | 94608000).

  • end - The time when the Reserved Instance expires (for example, 2015-08-07T11:54:42.000Z).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type on which the Reserved Instance can be used.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC. (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC)).

  • reserved-instances-id - The ID of the Reserved Instance.

  • start - The time at which the Reserved Instance purchase request was placed (for example, 2014-08-07T11:54:42.000Z).

  • state - The state of the Reserved Instance (payment-pending | active | payment-failed | retired).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeRouteTablesRequest$Filters": "

One or more filters.

  • association.route-table-association-id - The ID of an association ID for the route table.

  • association.route-table-id - The ID of the route table involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • association.main - Indicates whether the route table is the main route table for the VPC.

  • route-table-id - The ID of the route table.

  • route.destination-cidr-block - The CIDR range specified in a route in the table.

  • route.destination-prefix-list-id - The ID (prefix) of the AWS service specified in a route in the table.

  • route.gateway-id - The ID of a gateway specified in a route in the table.

  • route.instance-id - The ID of an instance specified in a route in the table.

  • route.origin - Describes how the route was created. CreateRouteTable indicates that the route was automatically created when the route table was created; CreateRoute indicates that the route was manually added to the route table; EnableVgwRoutePropagation indicates that the route was propagated by route propagation.

  • route.state - The state of a route in the route table (active | blackhole). The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, the specified NAT instance has been terminated, and so on).

  • route.vpc-peering-connection-id - The ID of a VPC peering connection specified in a route in the table.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the route table.

", - "DescribeSecurityGroupsRequest$Filters": "

One or more filters. If using multiple filters for rules, the results include security groups for which any combination of rules - not necessarily a single rule - match all filters.

  • description - The description of the security group.

  • egress.ip-permission.prefix-list-id - The ID (prefix) of the AWS service to which the security group allows access.

  • group-id - The ID of the security group.

  • group-name - The name of the security group.

  • ip-permission.cidr - A CIDR range that has been granted permission.

  • ip-permission.from-port - The start of port range for the TCP and UDP protocols, or an ICMP type number.

  • ip-permission.group-id - The ID of a security group that has been granted permission.

  • ip-permission.group-name - The name of a security group that has been granted permission.

  • ip-permission.protocol - The IP protocol for the permission (tcp | udp | icmp or a protocol number).

  • ip-permission.to-port - The end of port range for the TCP and UDP protocols, or an ICMP code.

  • ip-permission.user-id - The ID of an AWS account that has been granted permission.

  • owner-id - The AWS account ID of the owner of the security group.

  • tag-key - The key of a tag assigned to the security group.

  • tag-value - The value of a tag assigned to the security group.

  • vpc-id - The ID of the VPC specified when the security group was created.

", - "DescribeSnapshotsRequest$Filters": "

One or more filters.

  • description - A description of the snapshot.

  • owner-alias - The AWS account alias (for example, amazon) that owns the snapshot.

  • owner-id - The ID of the AWS account that owns the snapshot.

  • progress - The progress of the snapshot, as a percentage (for example, 80%).

  • snapshot-id - The snapshot ID.

  • start-time - The time stamp when the snapshot was initiated.

  • status - The status of the snapshot (pending | completed | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The ID of the volume the snapshot is for.

  • volume-size - The size of the volume, in GiB.

", - "DescribeSpotInstanceRequestsRequest$Filters": "

One or more filters.

  • availability-zone-group - The Availability Zone group.

  • create-time - The time stamp when the Spot instance request was created.

  • fault-code - The fault code related to the request.

  • fault-message - The fault message related to the request.

  • instance-id - The ID of the instance that fulfilled the request.

  • launch-group - The Spot instance launch group.

  • launch.block-device-mapping.delete-on-termination - Indicates whether the Amazon EBS volume is deleted on instance termination.

  • launch.block-device-mapping.device-name - The device name for the Amazon EBS volume (for example, /dev/sdh).

  • launch.block-device-mapping.snapshot-id - The ID of the snapshot used for the Amazon EBS volume.

  • launch.block-device-mapping.volume-size - The size of the Amazon EBS volume, in GiB.

  • launch.block-device-mapping.volume-type - The type of the Amazon EBS volume (gp2 | standard | io1).

  • launch.group-id - The security group for the instance.

  • launch.image-id - The ID of the AMI.

  • launch.instance-type - The type of instance (for example, m1.small).

  • launch.kernel-id - The kernel ID.

  • launch.key-name - The name of the key pair the instance launched with.

  • launch.monitoring-enabled - Whether monitoring is enabled for the Spot instance.

  • launch.ramdisk-id - The RAM disk ID.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.device-index - The index of the device for the network interface attachment on the instance.

  • network-interface.subnet-id - The ID of the subnet for the instance.

  • network-interface.description - A description of the network interface.

  • network-interface.private-ip-address - The primary private IP address of the network interface.

  • network-interface.delete-on-termination - Indicates whether the network interface is deleted when the instance is terminated.

  • network-interface.group-id - The ID of the security group associated with the network interface.

  • network-interface.group-name - The name of the security group associated with the network interface.

  • network-interface.addresses.primary - Indicates whether the IP address is the primary private IP address.

  • product-description - The product description associated with the instance (Linux/UNIX | Windows).

  • spot-instance-request-id - The Spot instance request ID.

  • spot-price - The maximum hourly price for any Spot instance launched to fulfill the request.

  • state - The state of the Spot instance request (open | active | closed | cancelled | failed). Spot bid status information can help you track your Amazon EC2 Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

  • status-code - The short code describing the most recent evaluation of your Spot instance request.

  • status-message - The message explaining the status of the Spot instance request.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of Spot instance request (one-time | persistent).

  • launched-availability-zone - The Availability Zone in which the bid is launched.

  • valid-from - The start date of the request.

  • valid-until - The end date of the request.

", - "DescribeSpotPriceHistoryRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone for which prices should be returned.

  • instance-type - The type of instance (for example, m1.small).

  • product-description - The product description for the Spot price (Linux/UNIX | SUSE Linux | Windows | Linux/UNIX (Amazon VPC) | SUSE Linux (Amazon VPC) | Windows (Amazon VPC)).

  • spot-price - The Spot price. The value must match exactly (or use wildcards; greater than or less than comparison is not supported).

  • timestamp - The timestamp of the Spot price history, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). You can use wildcards (* and ?). Greater than or less than comparison is not supported.

", - "DescribeSubnetsRequest$Filters": "

One or more filters.

  • availabilityZone - The Availability Zone for the subnet. You can also use availability-zone as the filter name.

  • available-ip-address-count - The number of IP addresses in the subnet that are available.

  • cidrBlock - The CIDR block of the subnet. The CIDR block you specify must exactly match the subnet's CIDR block for information to be returned for the subnet. You can also use cidr or cidr-block as the filter names.

  • defaultForAz - Indicates whether this is the default subnet for the Availability Zone. You can also use default-for-az as the filter name.

  • state - The state of the subnet (pending | available).

  • subnet-id - The ID of the subnet.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the subnet.

", - "DescribeTagsRequest$Filters": "

One or more filters.

  • key - The tag key.

  • resource-id - The resource ID.

  • resource-type - The resource type (customer-gateway | dhcp-options | image | instance | internet-gateway | network-acl | network-interface | reserved-instances | route-table | security-group | snapshot | spot-instances-request | subnet | volume | vpc | vpn-connection | vpn-gateway).

  • value - The tag value.

", - "DescribeVolumeStatusRequest$Filters": "

One or more filters.

  • action.code - The action code for the event (for example, enable-volume-io).

  • action.description - A description of the action.

  • action.event-id - The event ID associated with the action.

  • availability-zone - The Availability Zone of the instance.

  • event.description - A description of the event.

  • event.event-id - The event ID.

  • event.event-type - The event type (for io-enabled: passed | failed; for io-performance: io-performance:degraded | io-performance:severely-degraded | io-performance:stalled).

  • event.not-after - The latest end time for the event.

  • event.not-before - The earliest start time for the event.

  • volume-status.details-name - The cause for volume-status.status (io-enabled | io-performance).

  • volume-status.details-status - The status of volume-status.details-name (for io-enabled: passed | failed; for io-performance: normal | degraded | severely-degraded | stalled).

  • volume-status.status - The status of the volume (ok | impaired | warning | insufficient-data).

", - "DescribeVolumesRequest$Filters": "

One or more filters.

  • attachment.attach-time - The time stamp when the attachment initiated.

  • attachment.delete-on-termination - Whether the volume is deleted on instance termination.

  • attachment.device - The device name that is exposed to the instance (for example, /dev/sda1).

  • attachment.instance-id - The ID of the instance the volume is attached to.

  • attachment.status - The attachment state (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone in which the volume was created.

  • create-time - The time stamp when the volume was created.

  • encrypted - The encryption status of the volume.

  • size - The size of the volume, in GiB.

  • snapshot-id - The snapshot from which the volume was created.

  • status - The status of the volume (creating | available | in-use | deleting | deleted | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The volume ID.

  • volume-type - The Amazon EBS volume type. This can be gp2 for General Purpose (SSD) volumes, io1 for Provisioned IOPS (SSD) volumes, or standard for Magnetic volumes.

", - "DescribeVpcClassicLinkRequest$Filters": "

One or more filters.

  • is-classic-link-enabled - Whether the VPC is enabled for ClassicLink (true | false).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeVpcEndpointsRequest$Filters": "

One or more filters.

  • service-name: The name of the AWS service.

  • vpc-id: The ID of the VPC in which the endpoint resides.

  • vpc-endpoint-id: The ID of the endpoint.

  • vpc-endpoint-state: The state of the endpoint. (pending | available | deleting | deleted)

", - "DescribeVpcPeeringConnectionsRequest$Filters": "

One or more filters.

  • accepter-vpc-info.cidr-block - The CIDR block of the peer VPC.

  • accepter-vpc-info.owner-id - The AWS account ID of the owner of the peer VPC.

  • accepter-vpc-info.vpc-id - The ID of the peer VPC.

  • expiration-time - The expiration date and time for the VPC peering connection.

  • requester-vpc-info.cidr-block - The CIDR block of the requester's VPC.

  • requester-vpc-info.owner-id - The AWS account ID of the owner of the requester VPC.

  • requester-vpc-info.vpc-id - The ID of the requester VPC.

  • status-code - The status of the VPC peering connection (pending-acceptance | failed | expired | provisioning | active | deleted | rejected).

  • status-message - A message that provides more information about the status of the VPC peering connection, if applicable.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-peering-connection-id - The ID of the VPC peering connection.

", - "DescribeVpcsRequest$Filters": "

One or more filters.

  • cidr - The CIDR block of the VPC. The CIDR block you specify must exactly match the VPC's CIDR block for information to be returned for the VPC. Must contain the slash followed by one or two digits (for example, /28).

  • dhcp-options-id - The ID of a set of DHCP options.

  • isDefault - Indicates whether the VPC is the default VPC.

  • state - The state of the VPC (pending | available).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC.

", - "DescribeVpnConnectionsRequest$Filters": "

One or more filters.

  • customer-gateway-configuration - The configuration information for the customer gateway.

  • customer-gateway-id - The ID of a customer gateway associated with the VPN connection.

  • state - The state of the VPN connection (pending | available | deleting | deleted).

  • option.static-routes-only - Indicates whether the connection has static routes only. Used for devices that do not support Border Gateway Protocol (BGP).

  • route.destination-cidr-block - The destination CIDR block. This corresponds to the subnet used in a customer data center.

  • bgp-asn - The BGP Autonomous System Number (ASN) associated with a BGP device.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of VPN connection. Currently the only supported type is ipsec.1.

  • vpn-connection-id - The ID of the VPN connection.

  • vpn-gateway-id - The ID of a virtual private gateway associated with the VPN connection.

", - "DescribeVpnGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (attaching | attached | detaching | detached).

  • attachment.vpc-id - The ID of an attached VPC.

  • availability-zone - The Availability Zone for the virtual private gateway.

  • state - The state of the virtual private gateway (pending | available | deleting | deleted).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of virtual private gateway. Currently the only supported type is ipsec.1.

  • vpn-gateway-id - The ID of the virtual private gateway.

" - } - }, - "Float": { - "base": null, - "refs": { - "ReservedInstances$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstances$FixedPrice": "

The purchase price of the Reserved Instance.

", - "ReservedInstancesOffering$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstancesOffering$FixedPrice": "

The purchase price of the Reserved Instance.

" - } - }, - "FlowLog": { - "base": "

Describes a flow log.

", - "refs": { - "FlowLogSet$member": null - } - }, - "FlowLogSet": { - "base": null, - "refs": { - "DescribeFlowLogsResult$FlowLogs": "

Information about the flow logs.

" - } - }, - "FlowLogsResourceType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$ResourceType": "

The type of resource on which to create the flow log.

" - } - }, - "GatewayType": { - "base": null, - "refs": { - "CreateCustomerGatewayRequest$Type": "

The type of VPN connection that this customer gateway supports (ipsec.1).

", - "CreateVpnGatewayRequest$Type": "

The type of VPN connection this virtual private gateway supports.

", - "VpnConnection$Type": "

The type of VPN connection.

", - "VpnGateway$Type": "

The type of VPN connection the virtual private gateway supports.

" - } - }, - "GetConsoleOutputRequest": { - "base": null, - "refs": { - } - }, - "GetConsoleOutputResult": { - "base": null, - "refs": { - } - }, - "GetPasswordDataRequest": { - "base": null, - "refs": { - } - }, - "GetPasswordDataResult": { - "base": null, - "refs": { - } - }, - "GroupIdStringList": { - "base": null, - "refs": { - "AttachClassicLinkVpcRequest$Groups": "

The ID of one or more of the VPC's security groups. You cannot specify security groups from a different VPC.

", - "DescribeSecurityGroupsRequest$GroupIds": "

One or more security group IDs. Required for security groups in a nondefault VPC.

Default: Describes all your security groups.

", - "ModifyInstanceAttributeRequest$Groups": "

[EC2-VPC] Changes the security groups of the instance. You must specify at least one security group, even if it's just the default security group for the VPC. You must specify the security group ID, not the security group name.

" - } - }, - "GroupIdentifier": { - "base": "

Describes a security group.

", - "refs": { - "GroupIdentifierList$member": null - } - }, - "GroupIdentifierList": { - "base": null, - "refs": { - "ClassicLinkInstance$Groups": "

A list of security groups.

", - "DescribeNetworkInterfaceAttributeResult$Groups": "

The security groups associated with the network interface.

", - "Instance$SecurityGroups": "

One or more security groups for the instance.

", - "InstanceAttribute$Groups": "

The security groups associated with the instance.

", - "InstanceNetworkInterface$Groups": "

One or more security groups.

", - "LaunchSpecification$SecurityGroups": "

One or more security groups. To request an instance in a nondefault VPC, you must specify the ID of the security group. To request an instance in EC2-Classic or a default VPC, you can specify the name or the ID of the security group.

", - "NetworkInterface$Groups": "

Any security groups for the network interface.

", - "Reservation$Groups": "

One or more security groups.

", - "SpotFleetLaunchSpecification$SecurityGroups": "

One or more security groups. To request an instance in a nondefault VPC, you must specify the ID of the security group. To request an instance in EC2-Classic or a default VPC, you can specify the name or the ID of the security group.

" - } - }, - "GroupNameStringList": { - "base": null, - "refs": { - "DescribeSecurityGroupsRequest$GroupNames": "

[EC2-Classic and default VPC only] One or more security group names. You can specify either the security group name or the security group ID. For security groups in a nondefault VPC, use the group-name filter to describe security groups by name.

Default: Describes all your security groups.

", - "ModifySnapshotAttributeRequest$GroupNames": "

The group to modify for the snapshot.

" - } - }, - "HistoryRecord": { - "base": "

Describes an event in the history of the Spot fleet request.

", - "refs": { - "HistoryRecords$member": null - } - }, - "HistoryRecords": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryResponse$HistoryRecords": "

Information about the events in the history of the Spot fleet request.

" - } - }, - "HypervisorType": { - "base": null, - "refs": { - "Image$Hypervisor": "

The hypervisor type of the image.

", - "Instance$Hypervisor": "

The hypervisor type of the instance.

" - } - }, - "IamInstanceProfile": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "Instance$IamInstanceProfile": "

The IAM instance profile associated with the instance.

" - } - }, - "IamInstanceProfileSpecification": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "LaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RunInstancesRequest$IamInstanceProfile": "

The IAM instance profile.

", - "SpotFleetLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RequestSpotLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "IcmpTypeCode": { - "base": "

Describes the ICMP type and code.

", - "refs": { - "CreateNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying ICMP for the protocol.

", - "NetworkAclEntry$IcmpTypeCode": "

ICMP protocol: The ICMP type and code.

", - "ReplaceNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying 1 (ICMP) for the protocol.

" - } - }, - "Image": { - "base": "

Describes an image.

", - "refs": { - "ImageList$member": null - } - }, - "ImageAttribute": { - "base": "

Describes an image attribute.

", - "refs": { - } - }, - "ImageAttributeName": { - "base": null, - "refs": { - "DescribeImageAttributeRequest$Attribute": "

The AMI attribute.

Note: Depending on your account privileges, the blockDeviceMapping attribute may return a Client.AuthFailure error. If this happens, use DescribeImages to get information about the block device mapping for the AMI.

" - } - }, - "ImageDiskContainer": { - "base": "

Describes the disk container object for an import image task.

", - "refs": { - "ImageDiskContainerList$member": null - } - }, - "ImageDiskContainerList": { - "base": null, - "refs": { - "ImportImageRequest$DiskContainers": "

Information about the disk containers.

" - } - }, - "ImageIdStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ImageIds": "

One or more image IDs.

Default: Describes all images available to you.

" - } - }, - "ImageList": { - "base": null, - "refs": { - "DescribeImagesResult$Images": "

Information about one or more images.

" - } - }, - "ImageState": { - "base": null, - "refs": { - "Image$State": "

The current state of the AMI. If the state is available, the image is successfully registered and can be used to launch an instance.

" - } - }, - "ImageTypeValues": { - "base": null, - "refs": { - "Image$ImageType": "

The type of image.

" - } - }, - "ImportImageRequest": { - "base": null, - "refs": { - } - }, - "ImportImageResult": { - "base": null, - "refs": { - } - }, - "ImportImageTask": { - "base": "

Describes an import image task.

", - "refs": { - "ImportImageTaskList$member": null - } - }, - "ImportImageTaskList": { - "base": null, - "refs": { - "DescribeImportImageTasksResult$ImportImageTasks": "

A list of zero or more import image tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportInstanceLaunchSpecification": { - "base": "

Describes the launch specification for VM import.

", - "refs": { - "ImportInstanceRequest$LaunchSpecification": "

The launch specification.

" - } - }, - "ImportInstanceRequest": { - "base": null, - "refs": { - } - }, - "ImportInstanceResult": { - "base": null, - "refs": { - } - }, - "ImportInstanceTaskDetails": { - "base": "

Describes an import instance task.

", - "refs": { - "ConversionTask$ImportInstance": "

If the task is for importing an instance, this contains information about the import instance task.

" - } - }, - "ImportInstanceVolumeDetailItem": { - "base": "

Describes an import volume task.

", - "refs": { - "ImportInstanceVolumeDetailSet$member": null - } - }, - "ImportInstanceVolumeDetailSet": { - "base": null, - "refs": { - "ImportInstanceTaskDetails$Volumes": "

One or more volumes.

" - } - }, - "ImportKeyPairRequest": { - "base": null, - "refs": { - } - }, - "ImportKeyPairResult": { - "base": null, - "refs": { - } - }, - "ImportSnapshotRequest": { - "base": null, - "refs": { - } - }, - "ImportSnapshotResult": { - "base": null, - "refs": { - } - }, - "ImportSnapshotTask": { - "base": "

Describes an import snapshot task.

", - "refs": { - "ImportSnapshotTaskList$member": null - } - }, - "ImportSnapshotTaskList": { - "base": null, - "refs": { - "DescribeImportSnapshotTasksResult$ImportSnapshotTasks": "

A list of zero or more import snapshot tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportTaskIdList": { - "base": null, - "refs": { - "DescribeImportImageTasksRequest$ImportTaskIds": "

A list of import image task IDs.

", - "DescribeImportSnapshotTasksRequest$ImportTaskIds": "

A list of import snapshot task IDs.

" - } - }, - "ImportVolumeRequest": { - "base": null, - "refs": { - } - }, - "ImportVolumeResult": { - "base": null, - "refs": { - } - }, - "ImportVolumeTaskDetails": { - "base": "

Describes an import volume task.

", - "refs": { - "ConversionTask$ImportVolume": "

If the task is for importing a volume, this contains information about the import volume task.

" - } - }, - "Instance": { - "base": "

Describes an instance.

", - "refs": { - "InstanceList$member": null - } - }, - "InstanceAttribute": { - "base": "

Describes an instance attribute.

", - "refs": { - } - }, - "InstanceAttributeName": { - "base": null, - "refs": { - "DescribeInstanceAttributeRequest$Attribute": "

The instance attribute.

", - "ModifyInstanceAttributeRequest$Attribute": "

The name of the attribute.

", - "ResetInstanceAttributeRequest$Attribute": "

The attribute to reset.

" - } - }, - "InstanceBlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingList$member": null - } - }, - "InstanceBlockDeviceMappingList": { - "base": null, - "refs": { - "Instance$BlockDeviceMappings": "

Any block device mapping entries for the instance.

", - "InstanceAttribute$BlockDeviceMappings": "

The block device mapping of the instance.

" - } - }, - "InstanceBlockDeviceMappingSpecification": { - "base": "

Describes a block device mapping entry.

", - "refs": { - "InstanceBlockDeviceMappingSpecificationList$member": null - } - }, - "InstanceBlockDeviceMappingSpecificationList": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$BlockDeviceMappings": "

Modifies the DeleteOnTermination attribute for volumes that are currently attached. The volume must be owned by the caller. If no value is specified for DeleteOnTermination, the default is true and the volume is deleted when the instance is terminated.

To add instance store volumes to an Amazon EBS-backed instance, you must add them when you launch the instance. For more information, see Updating the Block Device Mapping when Launching an Instance in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "InstanceCount": { - "base": "

Describes a Reserved Instance listing state.

", - "refs": { - "InstanceCountList$member": null - } - }, - "InstanceCountList": { - "base": null, - "refs": { - "ReservedInstancesListing$InstanceCounts": "

The number of instances in this state.

" - } - }, - "InstanceExportDetails": { - "base": "

Describes an instance to export.

", - "refs": { - "ExportTask$InstanceExportDetails": "

Information about the instance to export.

" - } - }, - "InstanceIdStringList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesRequest$InstanceIds": "

One or more instance IDs. Must be instances linked to a VPC through ClassicLink.

", - "DescribeInstanceStatusRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

Constraints: Maximum 100 explicitly specified instance IDs.

", - "DescribeInstancesRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

", - "MonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "RebootInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "ReportInstanceStatusRequest$Instances": "

One or more instances.

", - "StartInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "StopInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "TerminateInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "UnmonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

" - } - }, - "InstanceLifecycleType": { - "base": null, - "refs": { - "Instance$InstanceLifecycle": "

Indicates whether this is a Spot Instance.

" - } - }, - "InstanceList": { - "base": null, - "refs": { - "Reservation$Instances": "

One or more instances.

" - } - }, - "InstanceMonitoring": { - "base": "

Describes the monitoring information of the instance.

", - "refs": { - "InstanceMonitoringList$member": null - } - }, - "InstanceMonitoringList": { - "base": null, - "refs": { - "MonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

", - "UnmonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

" - } - }, - "InstanceNetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceList$member": null - } - }, - "InstanceNetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "InstanceNetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "InstancePrivateIpAddress$Association": "

The association information for an Elastic IP address for the network interface.

" - } - }, - "InstanceNetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "InstanceNetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "InstanceNetworkInterfaceList": { - "base": null, - "refs": { - "Instance$NetworkInterfaces": "

[EC2-VPC] One or more network interfaces for the instance.

" - } - }, - "InstanceNetworkInterfaceSpecification": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceSpecificationList$member": null - } - }, - "InstanceNetworkInterfaceSpecificationList": { - "base": null, - "refs": { - "LaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RunInstancesRequest$NetworkInterfaces": "

One or more network interfaces.

", - "SpotFleetLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RequestSpotLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "InstancePrivateIpAddress": { - "base": "

Describes a private IP address.

", - "refs": { - "InstancePrivateIpAddressList$member": null - } - }, - "InstancePrivateIpAddressList": { - "base": null, - "refs": { - "InstanceNetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "InstanceState": { - "base": "

Describes the current state of the instance.

", - "refs": { - "Instance$State": "

The current state of the instance.

", - "InstanceStateChange$CurrentState": "

The current state of the instance.

", - "InstanceStateChange$PreviousState": "

The previous state of the instance.

", - "InstanceStatus$InstanceState": "

The intended state of the instance. DescribeInstanceStatus requires that an instance be in the running state.

" - } - }, - "InstanceStateChange": { - "base": "

Describes an instance state change.

", - "refs": { - "InstanceStateChangeList$member": null - } - }, - "InstanceStateChangeList": { - "base": null, - "refs": { - "StartInstancesResult$StartingInstances": "

Information about one or more started instances.

", - "StopInstancesResult$StoppingInstances": "

Information about one or more stopped instances.

", - "TerminateInstancesResult$TerminatingInstances": "

Information about one or more terminated instances.

" - } - }, - "InstanceStateName": { - "base": null, - "refs": { - "InstanceState$Name": "

The current state of the instance.

" - } - }, - "InstanceStatus": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatusList$member": null - } - }, - "InstanceStatusDetails": { - "base": "

Describes the instance status.

", - "refs": { - "InstanceStatusDetailsList$member": null - } - }, - "InstanceStatusDetailsList": { - "base": null, - "refs": { - "InstanceStatusSummary$Details": "

The system instance health or application instance health.

" - } - }, - "InstanceStatusEvent": { - "base": "

Describes a scheduled event for an instance.

", - "refs": { - "InstanceStatusEventList$member": null - } - }, - "InstanceStatusEventList": { - "base": null, - "refs": { - "InstanceStatus$Events": "

Any scheduled events associated with the instance.

" - } - }, - "InstanceStatusList": { - "base": null, - "refs": { - "DescribeInstanceStatusResult$InstanceStatuses": "

One or more instance status descriptions.

" - } - }, - "InstanceStatusSummary": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatus$SystemStatus": "

Reports impaired functionality that stems from issues related to the systems that support an instance, such as hardware failures and network connectivity problems.

", - "InstanceStatus$InstanceStatus": "

Reports impaired functionality that stems from issues internal to the instance, such as impaired reachability.

" - } - }, - "InstanceType": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$InstanceType": "

The instance type on which the Reserved Instance can be used. For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

", - "ImportInstanceLaunchSpecification$InstanceType": "

The instance type. For more information about the instance types that you can import, see Before You Get Started in the Amazon Elastic Compute Cloud User Guide.

", - "Instance$InstanceType": "

The instance type.

", - "InstanceTypeList$member": null, - "LaunchSpecification$InstanceType": "

The instance type.

", - "ReservedInstances$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$InstanceType": "

The instance type for the modified Reserved Instances.

", - "ReservedInstancesOffering$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "RunInstancesRequest$InstanceType": "

The instance type. For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

Default: m1.small

", - "SpotFleetLaunchSpecification$InstanceType": "

The instance type.

", - "SpotPrice$InstanceType": "

The instance type.

", - "RequestSpotLaunchSpecification$InstanceType": "

The instance type.

" - } - }, - "InstanceTypeList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$InstanceTypes": "

Filters the results by the specified instance types.

" - } - }, - "Integer": { - "base": null, - "refs": { - "AssignPrivateIpAddressesRequest$SecondaryPrivateIpAddressCount": "

The number of secondary IP addresses to assign to the network interface. You can't specify this parameter when also specifying private IP addresses.

", - "AttachNetworkInterfaceRequest$DeviceIndex": "

The index of the device for the network interface attachment.

", - "AuthorizeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "AuthorizeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "AuthorizeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "AuthorizeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "CreateCustomerGatewayRequest$BgpAsn": "

For devices that support BGP, the customer gateway's BGP ASN.

Default: 65000

", - "CreateNetworkAclEntryRequest$RuleNumber": "

The rule number for the entry (for example, 100). ACL entries are processed in ascending order by rule number.

Constraints: Positive integer from 1 to 32766

", - "CreateNetworkInterfaceRequest$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses to assign to a network interface. When you specify a number of secondary IP addresses, Amazon EC2 selects these IP addresses within the subnet range. You can't specify this option and specify more than one private IP address using privateIpAddresses.

The number of IP addresses you can assign to a network interface varies by instance type. For more information, see Private IP Addresses Per ENI Per Instance Type in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListingRequest$InstanceCount": "

The number of instances that are a part of a Reserved Instance account to be listed in the Reserved Instance Marketplace. This number should be less than or equal to the instance count associated with the Reserved Instance ID specified in this call.

", - "CreateVolumeRequest$Size": "

The size of the volume, in GiBs.

Constraints: 1-1024 for standard volumes, 1-16384 for gp2 volumes, and 4-16384 for io1 volumes. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "CreateVolumeRequest$Iops": "

Only valid for Provisioned IOPS (SSD) volumes. The number of I/O operations per second (IOPS) to provision for the volume, with a maximum ratio of 30 IOPS/GiB.

Constraint: Range is 100 to 20000 for Provisioned IOPS (SSD) volumes

", - "DeleteNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to delete.

", - "DescribeClassicLinkInstancesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the instance IDs parameter in the same request.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeFlowLogsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the flow log IDs parameter in the same request.

", - "DescribeImportImageTasksRequest$MaxResults": "

The maximum number of results to return in a single request.

", - "DescribeImportSnapshotTasksRequest$MaxResults": "

The maximum number of results to return in a single request.

", - "DescribeInstanceStatusRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the instance IDs parameter in the same request.

", - "DescribeInstancesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the instance IDs parameter in the same request.

", - "DescribeMovingAddressesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value outside of this range, an error is returned.

Default: If no value is provided, the default is 1000.

", - "DescribePrefixListsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribeReservedInstancesOfferingsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. The maximum is 100.

Default: 100

", - "DescribeReservedInstancesOfferingsRequest$MaxInstanceCount": "

The maximum number of instances to filter when searching for offerings.

Default: 20

", - "DescribeSnapshotsRequest$MaxResults": "

The maximum number of snapshot results returned by DescribeSnapshots in paginated output. When this parameter is used, DescribeSnapshots only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeSnapshots request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeSnapshots returns all results. You cannot specify this parameter and the snapshot IDs parameter in the same request.

", - "DescribeSpotFleetInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestsRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotPriceHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeTagsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned.

", - "DescribeVolumeStatusRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumeStatus in paginated output. When this parameter is used, the request only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumeStatus returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVolumesRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumes in paginated output. When this parameter is used, DescribeVolumes only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeVolumes request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumes returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVpcEndpointServicesRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeVpcEndpointsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "EbsBlockDevice$VolumeSize": "

The size of the volume, in GiB.

Constraints: 1-1024 for standard volumes, 1-16384 for gp2 volumes, and 4-16384 for io1 volumes. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "EbsBlockDevice$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For Provisioned IOPS (SSD) volumes, this represents the number of IOPS that are provisioned for the volume. For General Purpose (SSD) volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information on General Purpose (SSD) baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100 to 20000 for Provisioned IOPS (SSD) volumes and 3 to 10000 for General Purpose (SSD) volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create standard or gp2 volumes.

", - "IcmpTypeCode$Type": "

The ICMP code. A value of -1 means all codes for the specified ICMP type.

", - "IcmpTypeCode$Code": "

The ICMP type. A value of -1 means all types.

", - "Instance$AmiLaunchIndex": "

The AMI launch index, which can be used to find this instance in the launch group.

", - "InstanceCount$InstanceCount": "

The number of listed Reserved Instances in the state specified by the state.

", - "InstanceNetworkInterfaceAttachment$DeviceIndex": "

The index of the device on the instance for the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$DeviceIndex": "

The index of the device on the instance for the network interface attachment. If you are specifying a network interface in a RunInstances request, you must provide the device index.

", - "InstanceNetworkInterfaceSpecification$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses. You can't specify this option and specify more than one private IP address using the private IP addresses option.

", - "InstanceState$Code": "

The low byte represents the state. The high byte is an opaque internal value and should be ignored.

  • 0 : pending

  • 16 : running

  • 32 : shutting-down

  • 48 : terminated

  • 64 : stopping

  • 80 : stopped

", - "IpPermission$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "IpPermission$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code. A value of -1 indicates all ICMP codes for the specified ICMP type.

", - "NetworkAclEntry$RuleNumber": "

The rule number for the entry. ACL entries are processed in ascending order by rule number.

", - "NetworkInterfaceAttachment$DeviceIndex": "

The device index of the network interface attachment on the instance.

", - "PortRange$From": "

The first port in the range.

", - "PortRange$To": "

The last port in the range.

", - "PricingDetail$Count": "

The number of instances available for the price.

", - "PurchaseReservedInstancesOfferingRequest$InstanceCount": "

The number of Reserved Instances to purchase.

", - "ReplaceNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to replace.

", - "RequestSpotInstancesRequest$InstanceCount": "

The maximum number of Spot instances to launch.

Default: 1

", - "ReservedInstances$InstanceCount": "

The number of Reserved Instances purchased.

", - "ReservedInstancesConfiguration$InstanceCount": "

The number of modified Reserved Instances.

", - "RevokeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "RevokeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "RevokeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "RevokeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "RunInstancesRequest$MinCount": "

The minimum number of instances to launch. If you specify a minimum that is more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches no instances.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 General FAQ.

", - "RunInstancesRequest$MaxCount": "

The maximum number of instances to launch. If you specify more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches the largest possible number of instances above MinCount.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 General FAQ.

", - "Snapshot$VolumeSize": "

The size of the volume, in GiB.

", - "SpotFleetRequestConfigData$TargetCapacity": "

The number of units to request. You can choose to set the target capacity in terms of instances or a performance characteristic that is important to your application workload, such as vCPUs, memory, or I/O.

", - "Subnet$AvailableIpAddressCount": "

The number of unused IP addresses in the subnet. Note that the IP addresses for any stopped instances are considered unavailable.

", - "VgwTelemetry$AcceptedRouteCount": "

The number of accepted routes.

", - "Volume$Size": "

The size of the volume, in GiBs.

", - "Volume$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For Provisioned IOPS (SSD) volumes, this represents the number of IOPS that are provisioned for the volume. For General Purpose (SSD) volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information on General Purpose (SSD) baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100 to 20000 for Provisioned IOPS (SSD) volumes and 3 to 10000 for General Purpose (SSD) volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create standard or gp2 volumes.

" - } - }, - "InternetGateway": { - "base": "

Describes an Internet gateway.

", - "refs": { - "CreateInternetGatewayResult$InternetGateway": "

Information about the Internet gateway.

", - "InternetGatewayList$member": null - } - }, - "InternetGatewayAttachment": { - "base": "

Describes the attachment of a VPC to an Internet gateway.

", - "refs": { - "InternetGatewayAttachmentList$member": null - } - }, - "InternetGatewayAttachmentList": { - "base": null, - "refs": { - "InternetGateway$Attachments": "

Any VPCs attached to the Internet gateway.

" - } - }, - "InternetGatewayList": { - "base": null, - "refs": { - "DescribeInternetGatewaysResult$InternetGateways": "

Information about one or more Internet gateways.

" - } - }, - "IpPermission": { - "base": "

Describes a security group rule.

", - "refs": { - "IpPermissionList$member": null - } - }, - "IpPermissionList": { - "base": null, - "refs": { - "AuthorizeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "AuthorizeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. Can be used to specify multiple rules in a single command.

", - "RevokeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "RevokeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. You can't specify a source security group and a CIDR IP address range.

", - "SecurityGroup$IpPermissions": "

One or more inbound rules associated with the security group.

", - "SecurityGroup$IpPermissionsEgress": "

[EC2-VPC] One or more outbound rules associated with the security group.

" - } - }, - "IpRange": { - "base": "

Describes an IP range.

", - "refs": { - "IpRangeList$member": null - } - }, - "IpRangeList": { - "base": null, - "refs": { - "IpPermission$IpRanges": "

One or more IP ranges.

" - } - }, - "KeyNameStringList": { - "base": null, - "refs": { - "DescribeKeyPairsRequest$KeyNames": "

One or more key pair names.

Default: Describes all your key pairs.

" - } - }, - "KeyPair": { - "base": "

Describes a key pair.

", - "refs": { - } - }, - "KeyPairInfo": { - "base": "

Describes a key pair.

", - "refs": { - "KeyPairList$member": null - } - }, - "KeyPairList": { - "base": null, - "refs": { - "DescribeKeyPairsResult$KeyPairs": "

Information about one or more key pairs.

" - } - }, - "LaunchPermission": { - "base": "

Describes a launch permission.

", - "refs": { - "LaunchPermissionList$member": null - } - }, - "LaunchPermissionList": { - "base": null, - "refs": { - "ImageAttribute$LaunchPermissions": "

One or more launch permissions.

", - "LaunchPermissionModifications$Add": "

The AWS account ID to add to the list of launch permissions for the AMI.

", - "LaunchPermissionModifications$Remove": "

The AWS account ID to remove from the list of launch permissions for the AMI.

" - } - }, - "LaunchPermissionModifications": { - "base": "

Describes a launch permission modification.

", - "refs": { - "ModifyImageAttributeRequest$LaunchPermission": "

A launch permission modification.

" - } - }, - "LaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "SpotInstanceRequest$LaunchSpecification": "

Additional information for launching instances.

" - } - }, - "LaunchSpecsList": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$LaunchSpecifications": "

Information about the launch specifications for the Spot fleet request.

" - } - }, - "ListingState": { - "base": null, - "refs": { - "InstanceCount$State": "

The states of the listed Reserved Instances.

" - } - }, - "ListingStatus": { - "base": null, - "refs": { - "ReservedInstancesListing$Status": "

The status of the Reserved Instance listing.

" - } - }, - "Long": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$MinDuration": "

The minimum duration (in seconds) to filter when searching for offerings.

Default: 2592000 (1 month)

", - "DescribeReservedInstancesOfferingsRequest$MaxDuration": "

The maximum duration (in seconds) to filter when searching for offerings.

Default: 94608000 (3 years)

", - "DiskImageDescription$Size": "

The size of the disk image, in GiB.

", - "DiskImageDetail$Bytes": "

The size of the disk image, in GiB.

", - "DiskImageVolumeDescription$Size": "

The size of the volume, in GiB.

", - "ImportInstanceVolumeDetailItem$BytesConverted": "

The number of bytes converted so far.

", - "ImportVolumeTaskDetails$BytesConverted": "

The number of bytes converted so far.

", - "PriceSchedule$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "PriceScheduleSpecification$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "ReservedInstances$Duration": "

The duration of the Reserved Instance, in seconds.

", - "ReservedInstancesOffering$Duration": "

The duration of the Reserved Instance, in seconds.

", - "VolumeDetail$Size": "

The size of the volume, in GiB.

" - } - }, - "ModifyImageAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifyInstanceAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifyNetworkInterfaceAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifyReservedInstancesRequest": { - "base": null, - "refs": { - } - }, - "ModifyReservedInstancesResult": { - "base": null, - "refs": { - } - }, - "ModifySnapshotAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifySubnetAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifyVolumeAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifyVpcAttributeRequest": { - "base": null, - "refs": { - } - }, - "ModifyVpcEndpointRequest": { - "base": null, - "refs": { - } - }, - "ModifyVpcEndpointResult": { - "base": null, - "refs": { - } - }, - "MonitorInstancesRequest": { - "base": null, - "refs": { - } - }, - "MonitorInstancesResult": { - "base": null, - "refs": { - } - }, - "Monitoring": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "Instance$Monitoring": "

The monitoring information for the instance.

", - "InstanceMonitoring$Monitoring": "

The monitoring information.

" - } - }, - "MonitoringState": { - "base": null, - "refs": { - "Monitoring$State": "

Indicates whether monitoring is enabled for the instance.

" - } - }, - "MoveAddressToVpcRequest": { - "base": null, - "refs": { - } - }, - "MoveAddressToVpcResult": { - "base": null, - "refs": { - } - }, - "MoveStatus": { - "base": null, - "refs": { - "MovingAddressStatus$MoveStatus": "

The status of the Elastic IP address that's being moved to the EC2-VPC platform, or restored to the EC2-Classic platform.

" - } - }, - "MovingAddressStatus": { - "base": "

Describes the status of a moving Elastic IP address.

", - "refs": { - "MovingAddressStatusSet$member": null - } - }, - "MovingAddressStatusSet": { - "base": null, - "refs": { - "DescribeMovingAddressesResult$MovingAddressStatuses": "

The status for each Elastic IP address.

" - } - }, - "NetworkAcl": { - "base": "

Describes a network ACL.

", - "refs": { - "CreateNetworkAclResult$NetworkAcl": "

Information about the network ACL.

", - "NetworkAclList$member": null - } - }, - "NetworkAclAssociation": { - "base": "

Describes an association between a network ACL and a subnet.

", - "refs": { - "NetworkAclAssociationList$member": null - } - }, - "NetworkAclAssociationList": { - "base": null, - "refs": { - "NetworkAcl$Associations": "

Any associations between the network ACL and one or more subnets

" - } - }, - "NetworkAclEntry": { - "base": "

Describes an entry in a network ACL.

", - "refs": { - "NetworkAclEntryList$member": null - } - }, - "NetworkAclEntryList": { - "base": null, - "refs": { - "NetworkAcl$Entries": "

One or more entries (rules) in the network ACL.

" - } - }, - "NetworkAclList": { - "base": null, - "refs": { - "DescribeNetworkAclsResult$NetworkAcls": "

Information about one or more network ACLs.

" - } - }, - "NetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "CreateNetworkInterfaceResult$NetworkInterface": "

Information about the network interface.

", - "NetworkInterfaceList$member": null - } - }, - "NetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "NetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "NetworkInterfacePrivateIpAddress$Association": "

The association information for an Elastic IP address associated with the network interface.

" - } - }, - "NetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Attachment": "

The attachment (if any) of the network interface.

", - "NetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "NetworkInterfaceAttachmentChanges": { - "base": "

Describes an attachment change.

", - "refs": { - "ModifyNetworkInterfaceAttributeRequest$Attachment": "

Information about the interface attachment. If modifying the 'delete on termination' attribute, you must specify the ID of the interface attachment.

" - } - }, - "NetworkInterfaceAttribute": { - "base": null, - "refs": { - "DescribeNetworkInterfaceAttributeRequest$Attribute": "

The attribute of the network interface.

" - } - }, - "NetworkInterfaceIdList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesRequest$NetworkInterfaceIds": "

One or more network interface IDs.

Default: Describes all your network interfaces.

" - } - }, - "NetworkInterfaceList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesResult$NetworkInterfaces": "

Information about one or more network interfaces.

" - } - }, - "NetworkInterfacePrivateIpAddress": { - "base": "

Describes the private IP address of a network interface.

", - "refs": { - "NetworkInterfacePrivateIpAddressList$member": null - } - }, - "NetworkInterfacePrivateIpAddressList": { - "base": null, - "refs": { - "NetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "NetworkInterfaceStatus": { - "base": null, - "refs": { - "InstanceNetworkInterface$Status": "

The status of the network interface.

", - "NetworkInterface$Status": "

The status of the network interface.

" - } - }, - "OfferingTypeValues": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "DescribeReservedInstancesRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "ReservedInstances$OfferingType": "

The Reserved Instance offering type.

", - "ReservedInstancesOffering$OfferingType": "

The Reserved Instance offering type.

" - } - }, - "OperationType": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$OperationType": "

The operation type.

", - "ModifySnapshotAttributeRequest$OperationType": "

The type of operation to perform to the attribute.

" - } - }, - "OwnerStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$Owners": "

Filters the images by the owner. Specify an AWS account ID, amazon (owner is Amazon), aws-marketplace (owner is AWS Marketplace), self (owner is the sender of the request). Omitting this option returns all images for which you have launch permissions, regardless of ownership.

", - "DescribeSnapshotsRequest$OwnerIds": "

Returns the snapshots owned by the specified owner. Multiple owners can be specified.

" - } - }, - "PermissionGroup": { - "base": null, - "refs": { - "CreateVolumePermission$Group": "

The specific group that is to be added or removed from a volume's list of create volume permissions.

", - "LaunchPermission$Group": "

The name of the group.

" - } - }, - "Placement": { - "base": "

Describes the placement for the instance.

", - "refs": { - "ImportInstanceLaunchSpecification$Placement": "

The placement information for the instance.

", - "Instance$Placement": "

The location where the instance launched.

", - "RunInstancesRequest$Placement": "

The placement for the instance.

" - } - }, - "PlacementGroup": { - "base": "

Describes a placement group.

", - "refs": { - "PlacementGroupList$member": null - } - }, - "PlacementGroupList": { - "base": null, - "refs": { - "DescribePlacementGroupsResult$PlacementGroups": "

One or more placement groups.

" - } - }, - "PlacementGroupState": { - "base": null, - "refs": { - "PlacementGroup$State": "

The state of the placement group.

" - } - }, - "PlacementGroupStringList": { - "base": null, - "refs": { - "DescribePlacementGroupsRequest$GroupNames": "

One or more placement group names.

Default: Describes all your placement groups, or only those otherwise specified.

" - } - }, - "PlacementStrategy": { - "base": null, - "refs": { - "CreatePlacementGroupRequest$Strategy": "

The placement strategy.

", - "PlacementGroup$Strategy": "

The placement strategy.

" - } - }, - "PlatformValues": { - "base": null, - "refs": { - "Image$Platform": "

The value is Windows for Windows AMIs; otherwise blank.

", - "ImportInstanceRequest$Platform": "

The instance operating system.

", - "ImportInstanceTaskDetails$Platform": "

The instance operating system.

", - "Instance$Platform": "

The value is Windows for Windows instances; otherwise blank.

" - } - }, - "PortRange": { - "base": "

Describes a range of ports.

", - "refs": { - "CreateNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "NetworkAclEntry$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "ReplaceNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to. Required if specifying 6 (TCP) or 17 (UDP) for the protocol.

" - } - }, - "PrefixList": { - "base": "

Describes prefixes for AWS services.

", - "refs": { - "PrefixListSet$member": null - } - }, - "PrefixListId": { - "base": "

The ID of the prefix.

", - "refs": { - "PrefixListIdList$member": null - } - }, - "PrefixListIdList": { - "base": null, - "refs": { - "IpPermission$PrefixListIds": "

(Valid for AuthorizeSecurityGroupEgress, RevokeSecurityGroupEgress and DescribeSecurityGroups only) One or more prefix list IDs for an AWS service. In an AuthorizeSecurityGroupEgress request, this is the AWS service that you want to access through a VPC endpoint from instances associated with the security group.

" - } - }, - "PrefixListSet": { - "base": null, - "refs": { - "DescribePrefixListsResult$PrefixLists": "

All available prefix lists.

" - } - }, - "PriceSchedule": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleList$member": null - } - }, - "PriceScheduleList": { - "base": null, - "refs": { - "ReservedInstancesListing$PriceSchedules": "

The price of the Reserved Instance listing.

" - } - }, - "PriceScheduleSpecification": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleSpecificationList$member": null - } - }, - "PriceScheduleSpecificationList": { - "base": null, - "refs": { - "CreateReservedInstancesListingRequest$PriceSchedules": "

A list specifying the price of the Reserved Instance for each month remaining in the Reserved Instance term.

" - } - }, - "PricingDetail": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "PricingDetailsList$member": null - } - }, - "PricingDetailsList": { - "base": null, - "refs": { - "ReservedInstancesOffering$PricingDetails": "

The pricing details of the Reserved Instance offering.

" - } - }, - "PrivateIpAddressSpecification": { - "base": "

Describes a secondary private IP address for a network interface.

", - "refs": { - "PrivateIpAddressSpecificationList$member": null - } - }, - "PrivateIpAddressSpecificationList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$PrivateIpAddresses": "

One or more private IP addresses.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddresses": "

One or more private IP addresses to assign to the network interface. Only one private IP address can be designated as primary.

" - } - }, - "PrivateIpAddressStringList": { - "base": null, - "refs": { - "AssignPrivateIpAddressesRequest$PrivateIpAddresses": "

One or more IP addresses to be assigned as a secondary private IP address to the network interface. You can't specify this parameter when also specifying a number of secondary IP addresses.

If you don't specify an IP address, Amazon EC2 automatically selects an IP address within the subnet range.

", - "UnassignPrivateIpAddressesRequest$PrivateIpAddresses": "

The secondary private IP addresses to unassign from the network interface. You can specify this option multiple times to unassign more than one IP address.

" - } - }, - "ProductCode": { - "base": "

Describes a product code.

", - "refs": { - "ProductCodeList$member": null - } - }, - "ProductCodeList": { - "base": null, - "refs": { - "DescribeSnapshotAttributeResult$ProductCodes": "

A list of product codes.

", - "DescribeVolumeAttributeResult$ProductCodes": "

A list of product codes.

", - "Image$ProductCodes": "

Any product codes associated with the AMI.

", - "ImageAttribute$ProductCodes": "

One or more product codes.

", - "Instance$ProductCodes": "

The product codes attached to this instance.

", - "InstanceAttribute$ProductCodes": "

A list of product codes.

" - } - }, - "ProductCodeStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$ProductCodes": "

One or more product codes. After you add a product code to an AMI, it can't be removed. This is only valid when modifying the productCodes attribute.

" - } - }, - "ProductCodeValues": { - "base": null, - "refs": { - "ProductCode$ProductCodeType": "

The type of product code.

" - } - }, - "ProductDescriptionList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$ProductDescriptions": "

Filters the results by the specified basic product descriptions.

" - } - }, - "PropagatingVgw": { - "base": "

Describes a virtual private gateway propagating route.

", - "refs": { - "PropagatingVgwList$member": null - } - }, - "PropagatingVgwList": { - "base": null, - "refs": { - "RouteTable$PropagatingVgws": "

Any virtual private gateway (VGW) propagating routes.

" - } - }, - "PublicIpStringList": { - "base": null, - "refs": { - "DescribeAddressesRequest$PublicIps": "

[EC2-Classic] One or more Elastic IP addresses.

Default: Describes all your Elastic IP addresses.

" - } - }, - "PurchaseReservedInstancesOfferingRequest": { - "base": null, - "refs": { - } - }, - "PurchaseReservedInstancesOfferingResult": { - "base": null, - "refs": { - } - }, - "RIProductDescription": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ProductDescription": "

The Reserved Instance product platform description. Instances that include (Amazon VPC) in the description are for use with Amazon VPC.

", - "ReservedInstances$ProductDescription": "

The Reserved Instance product platform description.

", - "ReservedInstancesOffering$ProductDescription": "

The Reserved Instance product platform description.

", - "SpotInstanceRequest$ProductDescription": "

The product description associated with the Spot instance.

", - "SpotPrice$ProductDescription": "

A general description of the AMI.

" - } - }, - "ReasonCodesList": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$ReasonCodes": "

One or more reason codes that describes the health state of your instance.

  • instance-stuck-in-state: My instance is stuck in a state.

  • unresponsive: My instance is unresponsive.

  • not-accepting-credentials: My instance is not accepting my credentials.

  • password-not-available: A password is not available for my instance.

  • performance-network: My instance is experiencing performance problems which I believe are network related.

  • performance-instance-store: My instance is experiencing performance problems which I believe are related to the instance stores.

  • performance-ebs-volume: My instance is experiencing performance problems which I believe are related to an EBS volume.

  • performance-other: My instance is experiencing performance problems.

  • other: [explain using the description parameter]

" - } - }, - "RebootInstancesRequest": { - "base": null, - "refs": { - } - }, - "RecurringCharge": { - "base": "

Describes a recurring charge.

", - "refs": { - "RecurringChargesList$member": null - } - }, - "RecurringChargeFrequency": { - "base": null, - "refs": { - "RecurringCharge$Frequency": "

The frequency of the recurring charge.

" - } - }, - "RecurringChargesList": { - "base": null, - "refs": { - "ReservedInstances$RecurringCharges": "

The recurring charge tag assigned to the resource.

", - "ReservedInstancesOffering$RecurringCharges": "

The recurring charge tag assigned to the resource.

" - } - }, - "Region": { - "base": "

Describes a region.

", - "refs": { - "RegionList$member": null - } - }, - "RegionList": { - "base": null, - "refs": { - "DescribeRegionsResult$Regions": "

Information about one or more regions.

" - } - }, - "RegionNameStringList": { - "base": null, - "refs": { - "DescribeRegionsRequest$RegionNames": "

The names of one or more regions.

" - } - }, - "RegisterImageRequest": { - "base": null, - "refs": { - } - }, - "RegisterImageResult": { - "base": null, - "refs": { - } - }, - "RejectVpcPeeringConnectionRequest": { - "base": null, - "refs": { - } - }, - "RejectVpcPeeringConnectionResult": { - "base": null, - "refs": { - } - }, - "ReleaseAddressRequest": { - "base": null, - "refs": { - } - }, - "ReplaceNetworkAclAssociationRequest": { - "base": null, - "refs": { - } - }, - "ReplaceNetworkAclAssociationResult": { - "base": null, - "refs": { - } - }, - "ReplaceNetworkAclEntryRequest": { - "base": null, - "refs": { - } - }, - "ReplaceRouteRequest": { - "base": null, - "refs": { - } - }, - "ReplaceRouteTableAssociationRequest": { - "base": null, - "refs": { - } - }, - "ReplaceRouteTableAssociationResult": { - "base": null, - "refs": { - } - }, - "ReportInstanceReasonCodes": { - "base": null, - "refs": { - "ReasonCodesList$member": null - } - }, - "ReportInstanceStatusRequest": { - "base": null, - "refs": { - } - }, - "ReportStatusType": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$Status": "

The status of all instances listed.

" - } - }, - "RequestSpotFleetRequest": { - "base": "

Contains the parameters for RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotFleetResponse": { - "base": "

Contains the output of RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotInstancesRequest": { - "base": "

Contains the parameters for RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotInstancesResult": { - "base": "

Contains the output of RequestSpotInstances.

", - "refs": { - } - }, - "Reservation": { - "base": "

Describes a reservation.

", - "refs": { - "ReservationList$member": null - } - }, - "ReservationList": { - "base": null, - "refs": { - "DescribeInstancesResult$Reservations": "

One or more reservations.

" - } - }, - "ReservedInstanceLimitPrice": { - "base": "

Describes the limit price of a Reserved Instance offering.

", - "refs": { - "PurchaseReservedInstancesOfferingRequest$LimitPrice": "

Specified for Reserved Instance Marketplace offerings to limit the total order and ensure that the Reserved Instances are not purchased at unexpected prices.

" - } - }, - "ReservedInstanceState": { - "base": null, - "refs": { - "ReservedInstances$State": "

The state of the Reserved Instance purchase.

" - } - }, - "ReservedInstances": { - "base": "

Describes a Reserved Instance.

", - "refs": { - "ReservedInstancesList$member": null - } - }, - "ReservedInstancesConfiguration": { - "base": "

Describes the configuration settings for the modified Reserved Instances.

", - "refs": { - "ReservedInstancesConfigurationList$member": null, - "ReservedInstancesModificationResult$TargetConfiguration": "

The target Reserved Instances configurations supplied as part of the modification request.

" - } - }, - "ReservedInstancesConfigurationList": { - "base": null, - "refs": { - "ModifyReservedInstancesRequest$TargetConfigurations": "

The configuration settings for the Reserved Instances to modify.

" - } - }, - "ReservedInstancesId": { - "base": "

Describes the ID of a Reserved Instance.

", - "refs": { - "ReservedIntancesIds$member": null - } - }, - "ReservedInstancesIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesRequest$ReservedInstancesIds": "

One or more Reserved Instance IDs.

Default: Describes all your Reserved Instances, or only those otherwise specified.

", - "ModifyReservedInstancesRequest$ReservedInstancesIds": "

The IDs of the Reserved Instances to modify.

" - } - }, - "ReservedInstancesList": { - "base": null, - "refs": { - "DescribeReservedInstancesResult$ReservedInstances": "

A list of Reserved Instances.

" - } - }, - "ReservedInstancesListing": { - "base": "

Describes a Reserved Instance listing.

", - "refs": { - "ReservedInstancesListingList$member": null - } - }, - "ReservedInstancesListingList": { - "base": null, - "refs": { - "CancelReservedInstancesListingResult$ReservedInstancesListings": "

The Reserved Instance listing.

", - "CreateReservedInstancesListingResult$ReservedInstancesListings": "

Information about the Reserved Instances listing.

", - "DescribeReservedInstancesListingsResult$ReservedInstancesListings": "

Information about the Reserved Instance listing.

" - } - }, - "ReservedInstancesModification": { - "base": "

Describes a Reserved Instance modification.

", - "refs": { - "ReservedInstancesModificationList$member": null - } - }, - "ReservedInstancesModificationIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsRequest$ReservedInstancesModificationIds": "

IDs for the submitted modification request.

" - } - }, - "ReservedInstancesModificationList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsResult$ReservedInstancesModifications": "

The Reserved Instance modification information.

" - } - }, - "ReservedInstancesModificationResult": { - "base": null, - "refs": { - "ReservedInstancesModificationResultList$member": null - } - }, - "ReservedInstancesModificationResultList": { - "base": null, - "refs": { - "ReservedInstancesModification$ModificationResults": "

Contains target configurations along with their corresponding new Reserved Instance IDs.

" - } - }, - "ReservedInstancesOffering": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "ReservedInstancesOfferingList$member": null - } - }, - "ReservedInstancesOfferingIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ReservedInstancesOfferingIds": "

One or more Reserved Instances offering IDs.

" - } - }, - "ReservedInstancesOfferingList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsResult$ReservedInstancesOfferings": "

A list of Reserved Instances offerings.

" - } - }, - "ReservedIntancesIds": { - "base": null, - "refs": { - "ReservedInstancesModification$ReservedInstancesIds": "

The IDs of one or more Reserved Instances.

" - } - }, - "ResetImageAttributeName": { - "base": null, - "refs": { - "ResetImageAttributeRequest$Attribute": "

The attribute to reset (currently you can only reset the launch permission attribute).

" - } - }, - "ResetImageAttributeRequest": { - "base": null, - "refs": { - } - }, - "ResetInstanceAttributeRequest": { - "base": null, - "refs": { - } - }, - "ResetNetworkInterfaceAttributeRequest": { - "base": null, - "refs": { - } - }, - "ResetSnapshotAttributeRequest": { - "base": null, - "refs": { - } - }, - "ResourceIdList": { - "base": null, - "refs": { - "CreateTagsRequest$Resources": "

The IDs of one or more resources to tag. For example, ami-1a2b3c4d.

", - "DeleteTagsRequest$Resources": "

The ID of the resource. For example, ami-1a2b3c4d. You can specify more than one resource ID.

" - } - }, - "ResourceType": { - "base": null, - "refs": { - "TagDescription$ResourceType": "

The resource type.

" - } - }, - "RestorableByStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$RestorableByUserIds": "

One or more AWS accounts IDs that can create volumes from the snapshot.

" - } - }, - "RestoreAddressToClassicRequest": { - "base": null, - "refs": { - } - }, - "RestoreAddressToClassicResult": { - "base": null, - "refs": { - } - }, - "RevokeSecurityGroupEgressRequest": { - "base": null, - "refs": { - } - }, - "RevokeSecurityGroupIngressRequest": { - "base": null, - "refs": { - } - }, - "Route": { - "base": "

Describes a route in a route table.

", - "refs": { - "RouteList$member": null - } - }, - "RouteList": { - "base": null, - "refs": { - "RouteTable$Routes": "

The routes in the route table.

" - } - }, - "RouteOrigin": { - "base": null, - "refs": { - "Route$Origin": "

Describes how the route was created.

  • CreateRouteTable indicates that route was automatically created when the route table was created.
  • CreateRoute indicates that the route was manually added to the route table.
  • EnableVgwRoutePropagation indicates that the route was propagated by route propagation.
" - } - }, - "RouteState": { - "base": null, - "refs": { - "Route$State": "

The state of the route. The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, or the specified NAT instance has been terminated).

" - } - }, - "RouteTable": { - "base": "

Describes a route table.

", - "refs": { - "CreateRouteTableResult$RouteTable": "

Information about the route table.

", - "RouteTableList$member": null - } - }, - "RouteTableAssociation": { - "base": "

Describes an association between a route table and a subnet.

", - "refs": { - "RouteTableAssociationList$member": null - } - }, - "RouteTableAssociationList": { - "base": null, - "refs": { - "RouteTable$Associations": "

The associations between the route table and one or more subnets.

" - } - }, - "RouteTableList": { - "base": null, - "refs": { - "DescribeRouteTablesResult$RouteTables": "

Information about one or more route tables.

" - } - }, - "RuleAction": { - "base": null, - "refs": { - "CreateNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "NetworkAclEntry$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "ReplaceNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

" - } - }, - "RunInstancesMonitoringEnabled": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "LaunchSpecification$Monitoring": null, - "RunInstancesRequest$Monitoring": "

The monitoring for the instance.

", - "RequestSpotLaunchSpecification$Monitoring": null - } - }, - "RunInstancesRequest": { - "base": null, - "refs": { - } - }, - "S3Storage": { - "base": "

Describes the storage parameters for S3 and S3 buckets for an instance store-backed AMI.

", - "refs": { - "Storage$S3": "

An Amazon S3 storage location.

" - } - }, - "SecurityGroup": { - "base": "

Describes a security group

", - "refs": { - "SecurityGroupList$member": null - } - }, - "SecurityGroupIdStringList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$Groups": "

The IDs of one or more security groups.

", - "ImportInstanceLaunchSpecification$GroupIds": "

One or more security group IDs.

", - "InstanceNetworkInterfaceSpecification$Groups": "

The IDs of the security groups for the network interface. Applies only if creating a network interface when launching an instance.

", - "ModifyNetworkInterfaceAttributeRequest$Groups": "

Changes the security groups for the network interface. The new set of groups you specify replaces the current set. You must specify at least one group, even if it's just the default security group in the VPC. You must specify the ID of the security group, not the name.

", - "RunInstancesRequest$SecurityGroupIds": "

One or more security group IDs. You can create a security group using CreateSecurityGroup.

Default: Amazon EC2 uses the default security group.

" - } - }, - "SecurityGroupList": { - "base": null, - "refs": { - "DescribeSecurityGroupsResult$SecurityGroups": "

Information about one or more security groups.

" - } - }, - "SecurityGroupStringList": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$GroupNames": "

One or more security group names.

", - "RunInstancesRequest$SecurityGroups": "

[EC2-Classic, default VPC] One or more security group names. For a nondefault VPC, you must use security group IDs instead.

Default: Amazon EC2 uses the default security group.

" - } - }, - "ShutdownBehavior": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "RunInstancesRequest$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

Default: stop

" - } - }, - "Snapshot": { - "base": "

Describes a snapshot.

", - "refs": { - "SnapshotList$member": null - } - }, - "SnapshotAttributeName": { - "base": null, - "refs": { - "DescribeSnapshotAttributeRequest$Attribute": "

The snapshot attribute you would like to view.

", - "ModifySnapshotAttributeRequest$Attribute": "

The snapshot attribute to modify.

Only volume creation permissions may be modified at the customer level.

", - "ResetSnapshotAttributeRequest$Attribute": "

The attribute to reset. Currently, only the attribute for permission to create volumes can be reset.

" - } - }, - "SnapshotDetail": { - "base": "

Describes the snapshot created from the imported disk.

", - "refs": { - "SnapshotDetailList$member": null - } - }, - "SnapshotDetailList": { - "base": null, - "refs": { - "ImportImageResult$SnapshotDetails": "

Information about the snapshots.

", - "ImportImageTask$SnapshotDetails": "

Information about the snapshots.

" - } - }, - "SnapshotDiskContainer": { - "base": "

The disk container object for the import snapshot request.

", - "refs": { - "ImportSnapshotRequest$DiskContainer": "

Information about the disk container.

" - } - }, - "SnapshotIdStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$SnapshotIds": "

One or more snapshot IDs.

Default: Describes snapshots for which you have launch permissions.

" - } - }, - "SnapshotList": { - "base": null, - "refs": { - "DescribeSnapshotsResult$Snapshots": "

Information about the snapshots.

" - } - }, - "SnapshotState": { - "base": null, - "refs": { - "Snapshot$State": "

The snapshot state.

" - } - }, - "SnapshotTaskDetail": { - "base": "

Details about the import snapshot task.

", - "refs": { - "ImportSnapshotResult$SnapshotTaskDetail": "

Information about the import snapshot task.

", - "ImportSnapshotTask$SnapshotTaskDetail": "

Describes an import snapshot task.

" - } - }, - "SpotDatafeedSubscription": { - "base": "

Describes the data feed for a Spot instance.

", - "refs": { - "CreateSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

", - "DescribeSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

" - } - }, - "SpotFleetLaunchSpecification": { - "base": "

Describes the launch specification for one or more Spot instances.

", - "refs": { - "LaunchSpecsList$member": null - } - }, - "SpotFleetMonitoring": { - "base": "

Describes whether monitoring is enabled.

", - "refs": { - "SpotFleetLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "SpotFleetRequestConfig": { - "base": "

Describes a Spot fleet request.

", - "refs": { - "SpotFleetRequestConfigSet$member": null - } - }, - "SpotFleetRequestConfigData": { - "base": "

Describes the configuration of a Spot fleet request.

", - "refs": { - "RequestSpotFleetRequest$SpotFleetRequestConfig": "

The configuration for the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestConfig": "

Information about the configuration of the Spot fleet request.

" - } - }, - "SpotFleetRequestConfigSet": { - "base": null, - "refs": { - "DescribeSpotFleetRequestsResponse$SpotFleetRequestConfigs": "

Information about the configuration of your Spot fleet.

" - } - }, - "SpotInstanceRequest": { - "base": "

Describe a Spot instance request.

", - "refs": { - "SpotInstanceRequestList$member": null - } - }, - "SpotInstanceRequestIdList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

", - "DescribeSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

" - } - }, - "SpotInstanceRequestList": { - "base": null, - "refs": { - "DescribeSpotInstanceRequestsResult$SpotInstanceRequests": "

One or more Spot instance requests.

", - "RequestSpotInstancesResult$SpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "SpotInstanceState": { - "base": null, - "refs": { - "SpotInstanceRequest$State": "

The state of the Spot instance request. Spot bid status information can help you track your Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "SpotInstanceStateFault": { - "base": "

Describes a Spot instance state change.

", - "refs": { - "SpotDatafeedSubscription$Fault": "

The fault codes for the Spot instance request, if any.

", - "SpotInstanceRequest$Fault": "

The fault codes for the Spot instance request, if any.

" - } - }, - "SpotInstanceStatus": { - "base": "

Describes the status of a Spot instance request.

", - "refs": { - "SpotInstanceRequest$Status": "

The status code and status message describing the Spot instance request.

" - } - }, - "SpotInstanceType": { - "base": null, - "refs": { - "RequestSpotInstancesRequest$Type": "

The Spot instance request type.

Default: one-time

", - "SpotInstanceRequest$Type": "

The Spot instance request type.

" - } - }, - "SpotPlacement": { - "base": "

Describes Spot instance placement.

", - "refs": { - "LaunchSpecification$Placement": "

The placement information for the instance.

", - "SpotFleetLaunchSpecification$Placement": "

The placement information.

", - "RequestSpotLaunchSpecification$Placement": "

The placement information for the instance.

" - } - }, - "SpotPrice": { - "base": "

Describes the maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "refs": { - "SpotPriceHistoryList$member": null - } - }, - "SpotPriceHistoryList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryResult$SpotPriceHistory": "

The historical Spot prices.

" - } - }, - "StartInstancesRequest": { - "base": null, - "refs": { - } - }, - "StartInstancesResult": { - "base": null, - "refs": { - } - }, - "State": { - "base": null, - "refs": { - "VpcEndpoint$State": "

The state of the VPC endpoint.

" - } - }, - "StateReason": { - "base": "

Describes a state change.

", - "refs": { - "Image$StateReason": "

The reason for the state change.

", - "Instance$StateReason": "

The reason for the most recent state transition.

" - } - }, - "Status": { - "base": null, - "refs": { - "MoveAddressToVpcResult$Status": "

The status of the move of the IP address.

", - "RestoreAddressToClassicResult$Status": "

The move status for the IP address.

" - } - }, - "StatusName": { - "base": null, - "refs": { - "InstanceStatusDetails$Name": "

The type of instance status.

" - } - }, - "StatusType": { - "base": null, - "refs": { - "InstanceStatusDetails$Status": "

The status.

" - } - }, - "StopInstancesRequest": { - "base": null, - "refs": { - } - }, - "StopInstancesResult": { - "base": null, - "refs": { - } - }, - "Storage": { - "base": "

Describes the storage location for an instance store-backed AMI.

", - "refs": { - "BundleInstanceRequest$Storage": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "BundleTask$Storage": "

The Amazon S3 storage locations.

" - } - }, - "String": { - "base": null, - "refs": { - "AcceptVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "AccountAttribute$AttributeName": "

The name of the account attribute.

", - "AccountAttributeValue$AttributeValue": "

The value of the attribute.

", - "ActiveInstance$InstanceType": "

The instance type.

", - "ActiveInstance$InstanceId": "

The ID of the instance.

", - "ActiveInstance$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "Address$InstanceId": "

The ID of the instance that the address is associated with (if any).

", - "Address$PublicIp": "

The Elastic IP address.

", - "Address$AllocationId": "

The ID representing the allocation of the address for use with EC2-VPC.

", - "Address$AssociationId": "

The ID representing the association of the address with an instance in a VPC.

", - "Address$NetworkInterfaceId": "

The ID of the network interface.

", - "Address$NetworkInterfaceOwnerId": "

The ID of the AWS account that owns the network interface.

", - "Address$PrivateIpAddress": "

The private IP address associated with the Elastic IP address.

", - "AllocateAddressResult$PublicIp": "

The Elastic IP address.

", - "AllocateAddressResult$AllocationId": "

[EC2-VPC] The ID that AWS assigns to represent the allocation of the Elastic IP address for use with instances in a VPC.

", - "AllocationIdList$member": null, - "AssignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AssociateAddressRequest$InstanceId": "

The ID of the instance. This is required for EC2-Classic. For EC2-VPC, you can specify either the instance ID or the network interface ID, but not both. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "AssociateAddressRequest$PublicIp": "

The Elastic IP address. This is required for EC2-Classic.

", - "AssociateAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. This is required for EC2-VPC.

", - "AssociateAddressRequest$NetworkInterfaceId": "

[EC2-VPC] The ID of the network interface. If the instance has more than one network interface, you must specify a network interface ID.

", - "AssociateAddressRequest$PrivateIpAddress": "

[EC2-VPC] The primary or secondary private IP address to associate with the Elastic IP address. If no private IP address is specified, the Elastic IP address is associated with the primary private IP address.

", - "AssociateAddressResult$AssociationId": "

[EC2-VPC] The ID that represents the association of the Elastic IP address with an instance.

", - "AssociateDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set, or default to associate no DHCP options with the VPC.

", - "AssociateDhcpOptionsRequest$VpcId": "

The ID of the VPC.

", - "AssociateRouteTableRequest$SubnetId": "

The ID of the subnet.

", - "AssociateRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "AssociateRouteTableResult$AssociationId": "

The route table association ID (needed to disassociate the route table).

", - "AttachClassicLinkVpcRequest$InstanceId": "

The ID of an EC2-Classic instance to link to the ClassicLink-enabled VPC.

", - "AttachClassicLinkVpcRequest$VpcId": "

The ID of a ClassicLink-enabled VPC.

", - "AttachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "AttachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttachNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AttachNetworkInterfaceRequest$InstanceId": "

The ID of the instance.

", - "AttachNetworkInterfaceResult$AttachmentId": "

The ID of the network interface attachment.

", - "AttachVolumeRequest$VolumeId": "

The ID of the EBS volume. The volume and instance must be within the same Availability Zone.

", - "AttachVolumeRequest$InstanceId": "

The ID of the instance.

", - "AttachVolumeRequest$Device": "

The device name to expose to the instance (for example, /dev/sdh or xvdh).

", - "AttachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "AttachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttributeValue$Value": "

Valid values are case-sensitive and vary by action.

", - "AuthorizeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). Use -1 to specify all.

", - "AuthorizeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "AuthorizeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "AuthorizeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. For EC2-VPC, the source security group must be in the same VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic, default VPC] The AWS account number for the source security group. For EC2-VPC, the source security group must be in the same VPC. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). (VPC only) Use -1 to specify all.

", - "AuthorizeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "AvailabilityZone$ZoneName": "

The name of the Availability Zone.

", - "AvailabilityZone$RegionName": "

The name of the region.

", - "AvailabilityZoneMessage$Message": "

The message about the Availability Zone.

", - "BlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with 2 available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "BlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "BlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "BundleIdStringList$member": null, - "BundleInstanceRequest$InstanceId": "

The ID of the instance to bundle.

Type: String

Default: None

Required: Yes

", - "BundleTask$InstanceId": "

The ID of the instance associated with this bundle task.

", - "BundleTask$BundleId": "

The ID of the bundle task.

", - "BundleTask$Progress": "

The level of task completion, as a percent (for example, 20%).

", - "BundleTaskError$Code": "

The error code.

", - "BundleTaskError$Message": "

The error message.

", - "CancelBundleTaskRequest$BundleId": "

The ID of the bundle task.

", - "CancelConversionRequest$ConversionTaskId": "

The ID of the conversion task.

", - "CancelConversionRequest$ReasonMessage": "

The reason for canceling the conversion task.

", - "CancelExportTaskRequest$ExportTaskId": "

The ID of the export task. This is the ID returned by CreateInstanceExportTask.

", - "CancelImportTaskRequest$ImportTaskId": "

The ID of the import image or import snapshot task to be canceled.

", - "CancelImportTaskRequest$CancelReason": "

The reason for canceling the task.

", - "CancelImportTaskResult$ImportTaskId": "

The ID of the task being canceled.

", - "CancelImportTaskResult$State": "

The current state of the task being canceled.

", - "CancelImportTaskResult$PreviousState": "

The current state of the task being canceled.

", - "CancelReservedInstancesListingRequest$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "CancelSpotFleetRequestsError$Message": "

The description for the error code.

", - "CancelSpotFleetRequestsErrorItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelledSpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "ClassicLinkInstance$InstanceId": "

The ID of the instance.

", - "ClassicLinkInstance$VpcId": "

The ID of the VPC.

", - "ClientData$Comment": "

A user-defined comment about the disk upload.

", - "ConfirmProductInstanceRequest$ProductCode": "

The product code. This must be a product code that you own.

", - "ConfirmProductInstanceRequest$InstanceId": "

The ID of the instance.

", - "ConfirmProductInstanceResult$OwnerId": "

The AWS account ID of the instance owner. This is only present if the product code is attached to the instance.

", - "ConversionIdStringList$member": null, - "ConversionTask$ConversionTaskId": "

The ID of the conversion task.

", - "ConversionTask$ExpirationTime": "

The time when the task expires. If the upload isn't complete before the expiration time, we automatically cancel the task.

", - "ConversionTask$StatusMessage": "

The status message related to the conversion task.

", - "CopyImageRequest$SourceRegion": "

The name of the region that contains the AMI to copy.

", - "CopyImageRequest$SourceImageId": "

The ID of the AMI to copy.

", - "CopyImageRequest$Name": "

The name of the new AMI in the destination region.

", - "CopyImageRequest$Description": "

A description for the new AMI in the destination region.

", - "CopyImageRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "CopyImageResult$ImageId": "

The ID of the new AMI.

", - "CopySnapshotRequest$SourceRegion": "

The ID of the region that contains the snapshot to be copied.

", - "CopySnapshotRequest$SourceSnapshotId": "

The ID of the EBS snapshot to copy.

", - "CopySnapshotRequest$Description": "

A description for the EBS snapshot.

", - "CopySnapshotRequest$DestinationRegion": "

The destination region to use in the PresignedUrl parameter of a snapshot copy operation. This parameter is only valid for specifying the destination region in a PresignedUrl parameter, where it is required.

CopySnapshot sends the snapshot copy to the regional endpoint that you send the HTTP request to, such as ec2.us-east-1.amazonaws.com (in the AWS CLI, this is specified with the --region parameter or the default region in your AWS configuration file).

", - "CopySnapshotRequest$PresignedUrl": "

The pre-signed URL that facilitates copying an encrypted snapshot. This parameter is only required when copying an encrypted snapshot with the Amazon EC2 Query API; it is available as an optional parameter in all other cases. The PresignedUrl should use the snapshot source endpoint, the CopySnapshot action, and include the SourceRegion, SourceSnapshotId, and DestinationRegion parameters. The PresignedUrl must be signed using AWS Signature Version 4. Because EBS snapshots are stored in Amazon S3, the signing algorithm for this parameter uses the same logic that is described in Authenticating Requests by Using Query Parameters (AWS Signature Version 4) in the Amazon Simple Storage Service API Reference. An invalid or improperly signed PresignedUrl will cause the copy operation to fail asynchronously, and the snapshot will move to an error state.

", - "CopySnapshotRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when creating the snapshot copy. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopySnapshotResult$SnapshotId": "

The ID of the new snapshot.

", - "CreateCustomerGatewayRequest$PublicIp": "

The Internet-routable IP address for the customer gateway's outside interface. The address must be static.

", - "CreateFlowLogsRequest$LogGroupName": "

The name of the CloudWatch log group.

", - "CreateFlowLogsRequest$DeliverLogsPermissionArn": "

The ARN for the IAM role that's used to post flow logs to a CloudWatch Logs log group.

", - "CreateFlowLogsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateFlowLogsResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateImageRequest$InstanceId": "

The ID of the instance.

", - "CreateImageRequest$Name": "

A name for the new image.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "CreateImageRequest$Description": "

A description for the new image.

", - "CreateImageResult$ImageId": "

The ID of the new AMI.

", - "CreateInstanceExportTaskRequest$Description": "

A description for the conversion task or the resource being exported. The maximum length is 255 bytes.

", - "CreateInstanceExportTaskRequest$InstanceId": "

The ID of the instance.

", - "CreateKeyPairRequest$KeyName": "

A unique name for the key pair.

Constraints: Up to 255 ASCII characters

", - "CreateNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "CreateNetworkAclEntryRequest$Protocol": "

The protocol. A value of -1 means all protocols.

", - "CreateNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation (for example 172.16.0.0/24).

", - "CreateNetworkAclRequest$VpcId": "

The ID of the VPC.

", - "CreateNetworkInterfaceRequest$SubnetId": "

The ID of the subnet to associate with the network interface.

", - "CreateNetworkInterfaceRequest$Description": "

A description for the network interface.

", - "CreateNetworkInterfaceRequest$PrivateIpAddress": "

The primary private IP address of the network interface. If you don't specify an IP address, Amazon EC2 selects one for you from the subnet range. If you specify an IP address, you cannot indicate any IP addresses specified in privateIpAddresses as primary (only one IP address can be designated as primary).

", - "CreatePlacementGroupRequest$GroupName": "

A name for the placement group.

Constraints: Up to 255 ASCII characters

", - "CreateReservedInstancesListingRequest$ReservedInstancesId": "

The ID of the active Reserved Instance.

", - "CreateReservedInstancesListingRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "CreateRouteRequest$RouteTableId": "

The ID of the route table for the route.

", - "CreateRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. Routing decisions are based on the most specific match.

", - "CreateRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway attached to your VPC.

", - "CreateRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "CreateRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "CreateRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "CreateRouteTableRequest$VpcId": "

The ID of the VPC.

", - "CreateSecurityGroupRequest$GroupName": "

The name of the security group.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=&;{}!$*

", - "CreateSecurityGroupRequest$Description": "

A description for the security group. This is informational only.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=&;{}!$*

", - "CreateSecurityGroupRequest$VpcId": "

[EC2-VPC] The ID of the VPC. Required for EC2-VPC.

", - "CreateSecurityGroupResult$GroupId": "

The ID of the security group.

", - "CreateSnapshotRequest$VolumeId": "

The ID of the EBS volume.

", - "CreateSnapshotRequest$Description": "

A description for the snapshot.

", - "CreateSpotDatafeedSubscriptionRequest$Bucket": "

The Amazon S3 bucket in which to store the Spot instance data feed.

", - "CreateSpotDatafeedSubscriptionRequest$Prefix": "

A prefix for the data feed file names.

", - "CreateSubnetRequest$VpcId": "

The ID of the VPC.

", - "CreateSubnetRequest$CidrBlock": "

The network range for the subnet, in CIDR notation. For example, 10.0.0.0/24.

", - "CreateSubnetRequest$AvailabilityZone": "

The Availability Zone for the subnet.

Default: Amazon EC2 selects one for you (recommended).

", - "CreateVolumePermission$UserId": "

The specific AWS account ID that is to be added or removed from a volume's list of create volume permissions.

", - "CreateVolumeRequest$SnapshotId": "

The snapshot from which to create the volume.

", - "CreateVolumeRequest$AvailabilityZone": "

The Availability Zone in which to create the volume. Use DescribeAvailabilityZones to list the Availability Zones that are currently available to you.

", - "CreateVolumeRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) to use when creating the encrypted volume. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CreateVpcEndpointRequest$VpcId": "

The ID of the VPC in which the endpoint will be used.

", - "CreateVpcEndpointRequest$ServiceName": "

The AWS service name, in the form com.amazonaws.region.service. To get a list of available services, use the DescribeVpcEndpointServices request.

", - "CreateVpcEndpointRequest$PolicyDocument": "

A policy to attach to the endpoint that controls access to the service. The policy must be in valid JSON format. If this parameter is not specified, we attach a default policy that allows full access to the service.

", - "CreateVpcEndpointRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateVpcEndpointResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateVpcPeeringConnectionRequest$VpcId": "

The ID of the requester VPC.

", - "CreateVpcPeeringConnectionRequest$PeerVpcId": "

The ID of the VPC with which you are creating the VPC peering connection.

", - "CreateVpcPeeringConnectionRequest$PeerOwnerId": "

The AWS account ID of the owner of the peer VPC.

Default: Your AWS account ID

", - "CreateVpcRequest$CidrBlock": "

The network range for the VPC, in CIDR notation. For example, 10.0.0.0/16.

", - "CreateVpnConnectionRequest$Type": "

The type of VPN connection (ipsec.1).

", - "CreateVpnConnectionRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "CreateVpnConnectionRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "CreateVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "CreateVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "CreateVpnGatewayRequest$AvailabilityZone": "

The Availability Zone for the virtual private gateway.

", - "CustomerGateway$CustomerGatewayId": "

The ID of the customer gateway.

", - "CustomerGateway$State": "

The current state of the customer gateway (pending | available | deleting | deleted).

", - "CustomerGateway$Type": "

The type of VPN connection the customer gateway supports (ipsec.1).

", - "CustomerGateway$IpAddress": "

The Internet-routable IP address of the customer gateway's outside interface.

", - "CustomerGateway$BgpAsn": "

The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

", - "CustomerGatewayIdStringList$member": null, - "DeleteCustomerGatewayRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "DeleteDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set.

", - "DeleteInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DeleteKeyPairRequest$KeyName": "

The name of the key pair.

", - "DeleteNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkAclRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DeletePlacementGroupRequest$GroupName": "

The name of the placement group.

", - "DeleteRouteRequest$RouteTableId": "

The ID of the route table.

", - "DeleteRouteRequest$DestinationCidrBlock": "

The CIDR range for the route. The value you specify must match the CIDR for the route exactly.

", - "DeleteRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "DeleteSecurityGroupRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group. You can specify either the security group name or the security group ID.

", - "DeleteSecurityGroupRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "DeleteSnapshotRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DeleteSubnetRequest$SubnetId": "

The ID of the subnet.

", - "DeleteVolumeRequest$VolumeId": "

The ID of the volume.

", - "DeleteVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "DeleteVpcRequest$VpcId": "

The ID of the VPC.

", - "DeleteVpnConnectionRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "DeleteVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DeregisterImageRequest$ImageId": "

The ID of the AMI.

", - "DescribeClassicLinkInstancesRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeClassicLinkInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeFlowLogsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeFlowLogsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "DescribeImportImageTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportImageTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeImportSnapshotTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportSnapshotTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "DescribeInstanceStatusRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeInstanceStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeInstancesRequest$NextToken": "

The token to request the next page of results.

", - "DescribeInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeMovingAddressesRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeMovingAddressesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribeNetworkInterfaceAttributeResult$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribePrefixListsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribePrefixListsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesId": "

One or more Reserved Instance IDs.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesListingId": "

One or more Reserved Instance Listing IDs.

", - "DescribeReservedInstancesModificationsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesModificationsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeReservedInstancesOfferingsRequest$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "DescribeReservedInstancesOfferingsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesOfferingsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeSnapshotAttributeRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotAttributeResult$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotsRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeSnapshots request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeSnapshotsResult$NextToken": "

The NextToken value to include in a future DescribeSnapshots request. When the results of a DescribeSnapshots request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetInstancesRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetInstancesResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestHistoryRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestHistoryResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestsRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestsResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotPriceHistoryRequest$AvailabilityZone": "

Filters the results by the specified Availability Zone.

", - "DescribeSpotPriceHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotPriceHistoryResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeTagsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeTagsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return..

", - "DescribeVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "DescribeVolumeAttributeResult$VolumeId": "

The ID of the volume.

", - "DescribeVolumeStatusRequest$NextToken": "

The NextToken value to include in a future DescribeVolumeStatus request. When the results of the request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumeStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumesRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeVolumes request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeVolumesResult$NextToken": "

The NextToken value to include in a future DescribeVolumes request. When the results of a DescribeVolumes request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "DescribeVpcAttributeResult$VpcId": "

The ID of the VPC.

", - "DescribeVpcEndpointServicesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointServicesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeVpcEndpointsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DetachClassicLinkVpcRequest$InstanceId": "

The ID of the instance to unlink from the VPC.

", - "DetachClassicLinkVpcRequest$VpcId": "

The ID of the VPC to which the instance is linked.

", - "DetachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DetachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "DetachNetworkInterfaceRequest$AttachmentId": "

The ID of the attachment.

", - "DetachVolumeRequest$VolumeId": "

The ID of the volume.

", - "DetachVolumeRequest$InstanceId": "

The ID of the instance.

", - "DetachVolumeRequest$Device": "

The device name.

", - "DetachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DetachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "DhcpConfiguration$Key": "

The name of a DHCP option.

", - "DhcpOptions$DhcpOptionsId": "

The ID of the set of DHCP options.

", - "DhcpOptionsIdStringList$member": null, - "DisableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "DisableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "DisableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "DisassociateAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "DisassociateAddressRequest$AssociationId": "

[EC2-VPC] The association ID. Required for EC2-VPC.

", - "DisassociateRouteTableRequest$AssociationId": "

The association ID representing the current association between the route table and subnet.

", - "DiskImage$Description": "

A description of the disk image.

", - "DiskImageDescription$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

", - "DiskImageDescription$Checksum": "

The checksum computed for the disk image.

", - "DiskImageDetail$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3 and presented here as an Amazon S3 presigned URL. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

", - "DiskImageVolumeDescription$Id": "

The volume identifier.

", - "EbsBlockDevice$SnapshotId": "

The ID of the snapshot.

", - "EbsInstanceBlockDevice$VolumeId": "

The ID of the EBS volume.

", - "EbsInstanceBlockDeviceSpecification$VolumeId": "

The ID of the EBS volume.

", - "EnableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "EnableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "EnableVolumeIORequest$VolumeId": "

The ID of the volume.

", - "EnableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "EventInformation$InstanceId": "

The ID of the instance. This information is available only for instanceChange events.

", - "EventInformation$EventSubType": "

The event.

The following are the error events.

  • iamFleetRoleInvalid - Spot fleet did not have the required permissions either to launch or terminate an instance.

  • spotFleetRequestConfigurationInvalid - The configuration is not valid. For more information, see the description.

  • spotInstanceCountLimitExceeded - You've reached the limit on the number of Spot instances that you can launch.

The following are the fleetRequestChange events.

  • active - The Spot fleet has been validated and Amazon EC2 is attempting to maintain the target number of running Spot instances.

  • cancelled - The Spot fleet is canceled and has no running Spot instances. The Spot fleet will be deleted two days after its instances were terminated.

  • cancelled_running - The Spot fleet is canceled and will not launch additional Spot instances, but its existing Spot instances continue to run until they are interrupted or terminated.

  • cancelled_terminating - The Spot fleet is canceled and its Spot instances are terminating.

  • expired - The Spot fleet request has expired. A subsequent event indicates that the instances were terminated, if the request was created with TerminateInstancesWithExpiration set.

  • price_update - The bid price for a launch configuration was adjusted because it was too high. This change is permanent.

  • submitted - The Spot fleet request is being evaluated and Amazon EC2 is preparing to launch the target number of Spot instances.

The following are the instanceChange events.

  • launched - A bid was fulfilled and a new instance was launched.

  • terminated - An instance was terminated by the user.

", - "EventInformation$EventDescription": "

The description of the event.

", - "ExecutableByStringList$member": null, - "ExportTask$ExportTaskId": "

The ID of the export task.

", - "ExportTask$Description": "

A description of the resource being exported.

", - "ExportTask$StatusMessage": "

The status message related to the export task.

", - "ExportTaskIdStringList$member": null, - "ExportToS3Task$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3Task$S3Key": "

The encryption key for your S3 bucket.

", - "ExportToS3TaskSpecification$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3TaskSpecification$S3Prefix": "

The image is written to a single object in the S3 bucket at the S3 key s3prefix + exportTaskId + '.' + diskImageFormat.

", - "Filter$Name": "

The name of the filter. Filter names are case-sensitive.

", - "FlowLog$FlowLogId": "

The flow log ID.

", - "FlowLog$FlowLogStatus": "

The status of the flow log (ACTIVE).

", - "FlowLog$ResourceId": "

The ID of the resource on which the flow log was created.

", - "FlowLog$LogGroupName": "

The name of the flow log group.

", - "FlowLog$DeliverLogsStatus": "

The status of the logs delivery (SUCCESS | FAILED).

", - "FlowLog$DeliverLogsErrorMessage": "

Information about the error that occurred. Rate limited indicates that CloudWatch logs throttling has been applied for one or more network interfaces. Access error indicates that the IAM role associated with the flow log does not have sufficient permissions to publish to CloudWatch Logs. Unknown error indicates an internal error.

", - "FlowLog$DeliverLogsPermissionArn": "

The ARN of the IAM role that posts logs to CloudWatch Logs.

", - "GetConsoleOutputRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$Output": "

The console output, Base64 encoded.

", - "GetPasswordDataRequest$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$PasswordData": "

The password of the instance.

", - "GroupIdStringList$member": null, - "GroupIdentifier$GroupName": "

The name of the security group.

", - "GroupIdentifier$GroupId": "

The ID of the security group.

", - "GroupNameStringList$member": null, - "IamInstanceProfile$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfile$Id": "

The ID of the instance profile.

", - "IamInstanceProfileSpecification$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfileSpecification$Name": "

The name of the instance profile.

", - "Image$ImageId": "

The ID of the AMI.

", - "Image$ImageLocation": "

The location of the AMI.

", - "Image$OwnerId": "

The AWS account ID of the image owner.

", - "Image$CreationDate": "

The date and time the image was created.

", - "Image$KernelId": "

The kernel associated with the image, if any. Only applicable for machine images.

", - "Image$RamdiskId": "

The RAM disk associated with the image, if any. Only applicable for machine images.

", - "Image$SriovNetSupport": "

Specifies whether enhanced networking is enabled.

", - "Image$ImageOwnerAlias": "

The AWS account alias (for example, amazon, self) or the AWS account ID of the AMI owner.

", - "Image$Name": "

The name of the AMI that was provided during image creation.

", - "Image$Description": "

The description of the AMI that was provided during image creation.

", - "Image$RootDeviceName": "

The device name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "ImageAttribute$ImageId": "

The ID of the AMI.

", - "ImageDiskContainer$Description": "

The description of the disk image.

", - "ImageDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "ImageDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. The URL can either be a https URL (https://..) or an Amazon S3 URL (s3://..)

", - "ImageDiskContainer$DeviceName": "

The block device mapping for the disk.

", - "ImageDiskContainer$SnapshotId": "

The ID of the EBS snapshot to be used for importing the snapshot.

", - "ImageIdStringList$member": null, - "ImportImageRequest$Description": "

A description string for the import image task.

", - "ImportImageRequest$LicenseType": "

The license type to be used for the Amazon Machine Image (AMI) after importing.

Note: You may only use BYOL if you have existing licenses with rights to use these licenses in a third party cloud like AWS. For more information, see VM Import/Export Prerequisites in the Amazon Elastic Compute Cloud User Guide.

Valid values: AWS | BYOL

", - "ImportImageRequest$Hypervisor": "

The target hypervisor platform.

Valid values: xen

", - "ImportImageRequest$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageRequest$Platform": "

The operating system of the virtual machine.

Valid values: Windows | Linux

", - "ImportImageRequest$ClientToken": "

The token to enable idempotency for VM import requests.

", - "ImportImageRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportImageResult$ImportTaskId": "

The task ID of the import image task.

", - "ImportImageResult$Architecture": "

The architecture of the virtual machine.

", - "ImportImageResult$LicenseType": "

The license type of the virtual machine.

", - "ImportImageResult$Platform": "

The operating system of the virtual machine.

", - "ImportImageResult$Hypervisor": "

The target hypervisor of the import task.

", - "ImportImageResult$Description": "

A description of the import task.

", - "ImportImageResult$ImageId": "

The ID of the Amazon Machine Image (AMI) created by the import task.

", - "ImportImageResult$Progress": "

The progress of the task.

", - "ImportImageResult$StatusMessage": "

A detailed status message of the import task.

", - "ImportImageResult$Status": "

A brief status of the task.

", - "ImportImageTask$ImportTaskId": "

The ID of the import image task.

", - "ImportImageTask$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageTask$LicenseType": "

The license type of the virtual machine.

", - "ImportImageTask$Platform": "

The description string for the import image task.

", - "ImportImageTask$Hypervisor": "

The target hypervisor for the import task.

Valid values: xen

", - "ImportImageTask$Description": "

A description of the import task.

", - "ImportImageTask$ImageId": "

The ID of the Amazon Machine Image (AMI) of the imported virtual machine.

", - "ImportImageTask$Progress": "

The percentage of progress of the import image task.

", - "ImportImageTask$StatusMessage": "

A descriptive status message for the import image task.

", - "ImportImageTask$Status": "

A brief status for the import image task.

", - "ImportInstanceLaunchSpecification$AdditionalInfo": "

Reserved.

", - "ImportInstanceLaunchSpecification$SubnetId": "

[EC2-VPC] The ID of the subnet in which to launch the instance.

", - "ImportInstanceLaunchSpecification$PrivateIpAddress": "

[EC2-VPC] An available IP address from the IP address range of the subnet.

", - "ImportInstanceRequest$Description": "

A description for the instance being imported.

", - "ImportInstanceTaskDetails$InstanceId": "

The ID of the instance.

", - "ImportInstanceTaskDetails$Description": "

A description of the task.

", - "ImportInstanceVolumeDetailItem$AvailabilityZone": "

The Availability Zone where the resulting instance will reside.

", - "ImportInstanceVolumeDetailItem$Status": "

The status of the import of this particular disk image.

", - "ImportInstanceVolumeDetailItem$StatusMessage": "

The status information or errors related to the disk image.

", - "ImportInstanceVolumeDetailItem$Description": "

A description of the task.

", - "ImportKeyPairRequest$KeyName": "

A unique name for the key pair.

", - "ImportKeyPairResult$KeyName": "

The key pair name you provided.

", - "ImportKeyPairResult$KeyFingerprint": "

The MD5 public key fingerprint as specified in section 4 of RFC 4716.

", - "ImportSnapshotRequest$Description": "

The description string for the import snapshot task.

", - "ImportSnapshotRequest$ClientToken": "

Token to enable idempotency for VM import requests.

", - "ImportSnapshotRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportSnapshotResult$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotResult$Description": "

A description of the import snapshot task.

", - "ImportSnapshotTask$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotTask$Description": "

A description of the import snapshot task.

", - "ImportTaskIdList$member": null, - "ImportVolumeRequest$AvailabilityZone": "

The Availability Zone for the resulting EBS volume.

", - "ImportVolumeRequest$Description": "

A description of the volume.

", - "ImportVolumeTaskDetails$AvailabilityZone": "

The Availability Zone where the resulting volume will reside.

", - "ImportVolumeTaskDetails$Description": "

The description you provided when starting the import volume task.

", - "Instance$InstanceId": "

The ID of the instance.

", - "Instance$ImageId": "

The ID of the AMI used to launch the instance.

", - "Instance$PrivateDnsName": "

The private DNS name assigned to the instance. This DNS name can only be used inside the Amazon EC2 network. This name is not available until the instance enters the running state.

", - "Instance$PublicDnsName": "

The public DNS name assigned to the instance. This name is not available until the instance enters the running state.

", - "Instance$StateTransitionReason": "

The reason for the most recent state transition. This might be an empty string.

", - "Instance$KeyName": "

The name of the key pair, if this instance was launched with an associated key pair.

", - "Instance$KernelId": "

The kernel associated with this instance.

", - "Instance$RamdiskId": "

The RAM disk associated with this instance.

", - "Instance$SubnetId": "

The ID of the subnet in which the instance is running.

", - "Instance$VpcId": "

The ID of the VPC in which the instance is running.

", - "Instance$PrivateIpAddress": "

The private IP address assigned to the instance.

", - "Instance$PublicIpAddress": "

The public IP address assigned to the instance.

", - "Instance$RootDeviceName": "

The root device name (for example, /dev/sda1 or /dev/xvda).

", - "Instance$SpotInstanceRequestId": "

The ID of the Spot Instance request.

", - "Instance$ClientToken": "

The idempotency token you provided when you launched the instance.

", - "Instance$SriovNetSupport": "

Specifies whether enhanced networking is enabled.

", - "InstanceAttribute$InstanceId": "

The ID of the instance.

", - "InstanceBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$VirtualName": "

The virtual device name.

", - "InstanceBlockDeviceMappingSpecification$NoDevice": "

suppress the specified device included in the block device mapping.

", - "InstanceExportDetails$InstanceId": "

The ID of the resource being exported.

", - "InstanceIdStringList$member": null, - "InstanceMonitoring$InstanceId": "

The ID of the instance.

", - "InstanceNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterface$SubnetId": "

The ID of the subnet.

", - "InstanceNetworkInterface$VpcId": "

The ID of the VPC.

", - "InstanceNetworkInterface$Description": "

The description.

", - "InstanceNetworkInterface$OwnerId": "

The ID of the AWS account that created the network interface.

", - "InstanceNetworkInterface$MacAddress": "

The MAC address.

", - "InstanceNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "InstanceNetworkInterface$PrivateDnsName": "

The private DNS name.

", - "InstanceNetworkInterfaceAssociation$PublicIp": "

The public IP address or Elastic IP address bound to the network interface.

", - "InstanceNetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "InstanceNetworkInterfaceAssociation$IpOwnerId": "

The ID of the owner of the Elastic IP address.

", - "InstanceNetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterfaceSpecification$SubnetId": "

The ID of the subnet associated with the network string. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$Description": "

The description of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddress": "

The private IP address of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstancePrivateIpAddress$PrivateIpAddress": "

The private IP address of the network interface.

", - "InstancePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "InstanceStateChange$InstanceId": "

The ID of the instance.

", - "InstanceStatus$InstanceId": "

The ID of the instance.

", - "InstanceStatus$AvailabilityZone": "

The Availability Zone of the instance.

", - "InstanceStatusEvent$Description": "

A description of the event.

After a scheduled event is completed, it can still be described for up to a week. If the event has been completed, this description starts with the following text: [Completed].

", - "InternetGateway$InternetGatewayId": "

The ID of the Internet gateway.

", - "InternetGatewayAttachment$VpcId": "

The ID of the VPC.

", - "IpPermission$IpProtocol": "

The protocol.

When you call DescribeSecurityGroups, the protocol value returned is the number. Exception: For TCP, UDP, and ICMP, the value returned is the name (for example, tcp, udp, or icmp). For a list of protocol numbers, see Protocol Numbers. (VPC only) When you call AuthorizeSecurityGroupIngress, you can use -1 to specify all.

", - "IpRange$CidrIp": "

The CIDR range. You can either specify a CIDR range or a source security group, not both.

", - "KeyNameStringList$member": null, - "KeyPair$KeyName": "

The name of the key pair.

", - "KeyPair$KeyFingerprint": "

The SHA-1 digest of the DER encoded private key.

", - "KeyPair$KeyMaterial": "

An unencrypted PEM encoded RSA private key.

", - "KeyPairInfo$KeyName": "

The name of the key pair.

", - "KeyPairInfo$KeyFingerprint": "

If you used CreateKeyPair to create the key pair, this is the SHA-1 digest of the DER encoded private key. If you used ImportKeyPair to provide AWS the public key, this is the MD5 public key fingerprint as specified in section 4 of RFC4716.

", - "LaunchPermission$UserId": "

The AWS account ID.

", - "LaunchSpecification$ImageId": "

The ID of the AMI.

", - "LaunchSpecification$KeyName": "

The name of the key pair.

", - "LaunchSpecification$UserData": "

The Base64-encoded MIME user data to make available to the instances.

", - "LaunchSpecification$AddressingType": "

Deprecated.

", - "LaunchSpecification$KernelId": "

The ID of the kernel.

", - "LaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "LaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "ModifyImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ModifyImageAttributeRequest$Attribute": "

The name of the attribute to modify.

", - "ModifyImageAttributeRequest$Value": "

The value of the attribute being modified. This is only valid when modifying the description attribute.

", - "ModifyInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ModifyInstanceAttributeRequest$Value": "

A new value for the attribute. Use only with the kernel, ramdisk, userData, disableApiTermination, or instanceInitiatedShutdownBehavior attribute.

", - "ModifyNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ModifyReservedInstancesRequest$ClientToken": "

A unique, case-sensitive token you provide to ensure idempotency of your modification request. For more information, see Ensuring Idempotency.

", - "ModifyReservedInstancesResult$ReservedInstancesModificationId": "

The ID for the modification.

", - "ModifySnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ModifySubnetAttributeRequest$SubnetId": "

The ID of the subnet.

", - "ModifyVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "ModifyVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "ModifyVpcEndpointRequest$VpcEndpointId": "

The ID of the endpoint.

", - "ModifyVpcEndpointRequest$PolicyDocument": "

A policy document to attach to the endpoint. The policy must be in valid JSON format.

", - "MoveAddressToVpcRequest$PublicIp": "

The Elastic IP address.

", - "MoveAddressToVpcResult$AllocationId": "

The allocation ID for the Elastic IP address.

", - "MovingAddressStatus$PublicIp": "

The Elastic IP address.

", - "NetworkAcl$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAcl$VpcId": "

The ID of the VPC for the network ACL.

", - "NetworkAclAssociation$NetworkAclAssociationId": "

The ID of the association between a network ACL and a subnet.

", - "NetworkAclAssociation$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAclAssociation$SubnetId": "

The ID of the subnet.

", - "NetworkAclEntry$Protocol": "

The protocol. A value of -1 means all protocols.

", - "NetworkAclEntry$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "NetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "NetworkInterface$SubnetId": "

The ID of the subnet.

", - "NetworkInterface$VpcId": "

The ID of the VPC.

", - "NetworkInterface$AvailabilityZone": "

The Availability Zone.

", - "NetworkInterface$Description": "

A description.

", - "NetworkInterface$OwnerId": "

The AWS account ID of the owner of the network interface.

", - "NetworkInterface$RequesterId": "

The ID of the entity that launched the instance on your behalf (for example, AWS Management Console or Auto Scaling).

", - "NetworkInterface$MacAddress": "

The MAC address.

", - "NetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "NetworkInterface$PrivateDnsName": "

The private DNS name.

", - "NetworkInterfaceAssociation$PublicIp": "

The address of the Elastic IP address bound to the network interface.

", - "NetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "NetworkInterfaceAssociation$IpOwnerId": "

The ID of the Elastic IP address owner.

", - "NetworkInterfaceAssociation$AllocationId": "

The allocation ID.

", - "NetworkInterfaceAssociation$AssociationId": "

The association ID.

", - "NetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceAttachment$InstanceId": "

The ID of the instance.

", - "NetworkInterfaceAttachment$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "NetworkInterfaceAttachmentChanges$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceIdList$member": null, - "NetworkInterfacePrivateIpAddress$PrivateIpAddress": "

The private IP address.

", - "NetworkInterfacePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "OwnerStringList$member": null, - "Placement$AvailabilityZone": "

The Availability Zone of the instance.

", - "Placement$GroupName": "

The name of the placement group the instance is in (for cluster compute instances).

", - "PlacementGroup$GroupName": "

The name of the placement group.

", - "PlacementGroupStringList$member": null, - "PrefixList$PrefixListId": "

The ID of the prefix.

", - "PrefixList$PrefixListName": "

The name of the prefix.

", - "PrefixListId$PrefixListId": "

The ID of the prefix.

", - "PrivateIpAddressSpecification$PrivateIpAddress": "

The private IP addresses.

", - "PrivateIpAddressStringList$member": null, - "ProductCode$ProductCodeId": "

The product code.

", - "ProductCodeStringList$member": null, - "ProductDescriptionList$member": null, - "PropagatingVgw$GatewayId": "

The ID of the virtual private gateway (VGW).

", - "PublicIpStringList$member": null, - "PurchaseReservedInstancesOfferingRequest$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering to purchase.

", - "PurchaseReservedInstancesOfferingResult$ReservedInstancesId": "

The IDs of the purchased Reserved Instances.

", - "Region$RegionName": "

The name of the region.

", - "Region$Endpoint": "

The region service endpoint.

", - "RegionNameStringList$member": null, - "RegisterImageRequest$ImageLocation": "

The full path to your AMI manifest in Amazon S3 storage.

", - "RegisterImageRequest$Name": "

A name for your AMI.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "RegisterImageRequest$Description": "

A description for your AMI.

", - "RegisterImageRequest$KernelId": "

The ID of the kernel.

", - "RegisterImageRequest$RamdiskId": "

The ID of the RAM disk.

", - "RegisterImageRequest$RootDeviceName": "

The name of the root device (for example, /dev/sda1, or /dev/xvda).

", - "RegisterImageRequest$VirtualizationType": "

The type of virtualization.

Default: paravirtual

", - "RegisterImageRequest$SriovNetSupport": "

Set to simple to enable enhanced networking for the AMI and any instances that you launch from the AMI.

There is no way to disable enhanced networking at this time.

This option is supported only for HVM AMIs. Specifying this option with a PV AMI can make instances launched from the AMI unreachable.

", - "RegisterImageResult$ImageId": "

The ID of the newly registered AMI.

", - "RejectVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "ReleaseAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "ReleaseAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. Required for EC2-VPC.

", - "ReplaceNetworkAclAssociationRequest$AssociationId": "

The ID of the current association between the original network ACL and the subnet.

", - "ReplaceNetworkAclAssociationRequest$NetworkAclId": "

The ID of the new network ACL to associate with the subnet.

", - "ReplaceNetworkAclAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReplaceNetworkAclEntryRequest$NetworkAclId": "

The ID of the ACL.

", - "ReplaceNetworkAclEntryRequest$Protocol": "

The IP protocol. You can specify all or -1 to mean all protocols.

", - "ReplaceNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "ReplaceRouteRequest$RouteTableId": "

The ID of the route table.

", - "ReplaceRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. The value you provide must match the CIDR of an existing route in the table.

", - "ReplaceRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway.

", - "ReplaceRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC.

", - "ReplaceRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "ReplaceRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "ReplaceRouteTableAssociationRequest$AssociationId": "

The association ID.

", - "ReplaceRouteTableAssociationRequest$RouteTableId": "

The ID of the new route table to associate with the subnet.

", - "ReplaceRouteTableAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReportInstanceStatusRequest$Description": "

Descriptive text about the health state of your instance.

", - "RequestSpotFleetResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "RequestSpotInstancesRequest$SpotPrice": "

The maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "RequestSpotInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstancesRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

Default: Instances are launched and terminated individually

", - "RequestSpotInstancesRequest$AvailabilityZoneGroup": "

The user-specified name for a logical grouping of bids.

When you specify an Availability Zone group in a Spot Instance request, all Spot instances in the request are launched in the same Availability Zone. Instance proximity is maintained with this parameter, but the choice of Availability Zone is not. The group applies only to bids for Spot Instances of the same instance type. Any additional Spot instance requests that are specified with the same Availability Zone group name are launched in that same Availability Zone, as long as at least one instance from the group is still active.

If there is no active instance running in the Availability Zone group that you specify for a new Spot instance request (all instances are terminated, the bid is expired, or the bid falls below current market), then Amazon EC2 launches the instance in any Availability Zone where the constraint can be met. Consequently, the subsequent set of Spot instances could be placed in a different zone from the original request, even if you specified the same Availability Zone group.

Default: Instances are launched in any available Availability Zone.

", - "Reservation$ReservationId": "

The ID of the reservation.

", - "Reservation$OwnerId": "

The ID of the AWS account that owns the reservation.

", - "Reservation$RequesterId": "

The ID of the requester that launched the instances on your behalf (for example, AWS Management Console or Auto Scaling).

", - "ReservedInstances$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstances$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$AvailabilityZone": "

The Availability Zone for the modified Reserved Instances.

", - "ReservedInstancesConfiguration$Platform": "

The network platform of the modified Reserved Instances, which is either EC2-Classic or EC2-VPC.

", - "ReservedInstancesId$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesIdStringList$member": null, - "ReservedInstancesListing$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "ReservedInstancesListing$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesListing$StatusMessage": "

The reason for the current status of the Reserved Instance listing. The response can be blank.

", - "ReservedInstancesListing$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModification$ReservedInstancesModificationId": "

A unique ID for the Reserved Instance modification.

", - "ReservedInstancesModification$Status": "

The status of the Reserved Instances modification request.

", - "ReservedInstancesModification$StatusMessage": "

The reason for the status.

", - "ReservedInstancesModification$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModificationIdStringList$member": null, - "ReservedInstancesModificationResult$ReservedInstancesId": "

The ID for the Reserved Instances that were created as part of the modification request. This field is only available when the modification is fulfilled.

", - "ReservedInstancesOffering$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering.

", - "ReservedInstancesOffering$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesOfferingIdStringList$member": null, - "ResetImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ResetInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ResetNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ResetNetworkInterfaceAttributeRequest$SourceDestCheck": "

The source/destination checking attribute. Resets the value to true.

", - "ResetSnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ResourceIdList$member": null, - "RestorableByStringList$member": null, - "RestoreAddressToClassicRequest$PublicIp": "

The Elastic IP address.

", - "RestoreAddressToClassicResult$PublicIp": "

The Elastic IP address.

", - "RevokeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). Use -1 to specify all.

", - "RevokeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "RevokeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "RevokeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a security group in a nondefault VPC.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. For EC2-VPC, the source security group must be in the same VPC.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic, default VPC] The AWS account ID of the source security group. For EC2-VPC, the source security group must be in the same VPC. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). Use -1 to specify all.

", - "RevokeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "Route$DestinationCidrBlock": "

The CIDR block used for the destination match.

", - "Route$DestinationPrefixListId": "

The prefix of the AWS service.

", - "Route$GatewayId": "

The ID of a gateway attached to your VPC.

", - "Route$InstanceId": "

The ID of a NAT instance in your VPC.

", - "Route$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "Route$NetworkInterfaceId": "

The ID of the network interface.

", - "Route$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "RouteTable$RouteTableId": "

The ID of the route table.

", - "RouteTable$VpcId": "

The ID of the VPC.

", - "RouteTableAssociation$RouteTableAssociationId": "

The ID of the association between a route table and a subnet.

", - "RouteTableAssociation$RouteTableId": "

The ID of the route table.

", - "RouteTableAssociation$SubnetId": "

The ID of the subnet. A subnet ID is not returned for an implicit association.

", - "RunInstancesRequest$ImageId": "

The ID of the AMI, which you can get by calling DescribeImages.

", - "RunInstancesRequest$KeyName": "

The name of the key pair. You can create a key pair using CreateKeyPair or ImportKeyPair.

If you do not specify a key pair, you can't connect to the instance unless you choose an AMI that is configured to allow users another way to log in.

", - "RunInstancesRequest$UserData": "

The Base64-encoded MIME user data for the instances.

", - "RunInstancesRequest$KernelId": "

The ID of the kernel.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$RamdiskId": "

The ID of the RAM disk.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$SubnetId": "

[EC2-VPC] The ID of the subnet to launch the instance into.

", - "RunInstancesRequest$PrivateIpAddress": "

[EC2-VPC] The primary IP address. You must specify a value from the IP address range of the subnet.

Only one private IP address can be designated as primary. Therefore, you can't specify this parameter if PrivateIpAddresses.n.Primary is set to true and PrivateIpAddresses.n.PrivateIpAddress is set to an IP address.

Default: We select an IP address from the IP address range of the subnet.

", - "RunInstancesRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see Ensuring Idempotency.

Constraints: Maximum 64 ASCII characters

", - "RunInstancesRequest$AdditionalInfo": "

Reserved.

", - "S3Storage$Bucket": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "S3Storage$Prefix": "

The beginning of the file name of the AMI.

", - "S3Storage$AWSAccessKeyId": "

The access key ID of the owner of the bucket. Before you specify a value for your access key ID, review and follow the guidance in Best Practices for Managing AWS Access Keys.

", - "S3Storage$UploadPolicySignature": "

The signature of the Base64 encoded JSON document.

", - "SecurityGroup$OwnerId": "

The AWS account ID of the owner of the security group.

", - "SecurityGroup$GroupName": "

The name of the security group.

", - "SecurityGroup$GroupId": "

The ID of the security group.

", - "SecurityGroup$Description": "

A description of the security group.

", - "SecurityGroup$VpcId": "

[EC2-VPC] The ID of the VPC for the security group.

", - "SecurityGroupIdStringList$member": null, - "SecurityGroupStringList$member": null, - "Snapshot$SnapshotId": "

The ID of the snapshot. Each snapshot receives a unique identifier when it is created.

", - "Snapshot$VolumeId": "

The ID of the volume that was used to create the snapshot.

", - "Snapshot$StateMessage": "

Encrypted Amazon EBS snapshots are copied asynchronously. If a snapshot copy operation fails (for example, if the proper AWS Key Management Service (AWS KMS) permissions are not obtained) this field displays error state details to help you diagnose why the error occurred. This parameter is only returned by the DescribeSnapshots API operation.

", - "Snapshot$Progress": "

The progress of the snapshot, as a percentage.

", - "Snapshot$OwnerId": "

The AWS account ID of the EBS snapshot owner.

", - "Snapshot$Description": "

The description for the snapshot.

", - "Snapshot$OwnerAlias": "

The AWS account alias (for example, amazon, self) or AWS account ID that owns the snapshot.

", - "Snapshot$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the parent volume.

", - "Snapshot$DataEncryptionKeyId": "

The data encryption key identifier for the snapshot. This value is a unique identifier that corresponds to the data encryption key that was used to encrypt the original volume or snapshot copy. Because data encryption keys are inherited by volumes created from snapshots, and vice versa, if snapshots share the same data encryption key identifier, then they belong to the same volume/snapshot lineage. This parameter is only returned by the DescribeSnapshots API operation.

", - "SnapshotDetail$Description": "

A description for the snapshot.

", - "SnapshotDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotDetail$Url": "

The URL used to access the disk image.

", - "SnapshotDetail$DeviceName": "

The block device mapping for the snapshot.

", - "SnapshotDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotDetail$Progress": "

The percentage of progress for the task.

", - "SnapshotDetail$StatusMessage": "

A detailed status message for the snapshot creation.

", - "SnapshotDetail$Status": "

A brief status of the snapshot creation.

", - "SnapshotDiskContainer$Description": "

The description of the disk image being imported.

", - "SnapshotDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "SnapshotDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. It can either be a https URL (https://..) or an Amazon S3 URL (s3://..).

", - "SnapshotIdStringList$member": null, - "SnapshotTaskDetail$Description": "

The description of the snapshot.

", - "SnapshotTaskDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$Url": "

The URL of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotTaskDetail$Progress": "

The percentage of completion for the import snapshot task.

", - "SnapshotTaskDetail$StatusMessage": "

A detailed status message for the import snapshot task.

", - "SnapshotTaskDetail$Status": "

A brief status for the import snapshot task.

", - "SpotDatafeedSubscription$OwnerId": "

The AWS account ID of the account.

", - "SpotDatafeedSubscription$Bucket": "

The Amazon S3 bucket where the Spot instance data feed is located.

", - "SpotDatafeedSubscription$Prefix": "

The prefix that is prepended to data feed files.

", - "SpotFleetLaunchSpecification$ImageId": "

The ID of the AMI.

", - "SpotFleetLaunchSpecification$KeyName": "

The name of the key pair.

", - "SpotFleetLaunchSpecification$UserData": "

The Base64-encoded MIME user data to make available to the instances.

", - "SpotFleetLaunchSpecification$AddressingType": "

Deprecated.

", - "SpotFleetLaunchSpecification$KernelId": "

The ID of the kernel.

", - "SpotFleetLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "SpotFleetLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances.

", - "SpotFleetLaunchSpecification$SpotPrice": "

The bid price per unit hour for the specified instance type. If this value is not specified, the default is the Spot bid price specified for the fleet. To determine the bid price per unit hour, divide the Spot bid price by the value of WeightedCapacity.

", - "SpotFleetRequestConfig$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "SpotFleetRequestConfigData$ClientToken": "

A unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "SpotFleetRequestConfigData$SpotPrice": "

The bid price per unit hour.

", - "SpotFleetRequestConfigData$IamFleetRole": "

Grants the Spot fleet permission to terminate Spot instances on your behalf when you cancel its Spot fleet request using CancelSpotFleetRequests or when the Spot fleet request expires, if you set terminateInstancesWithExpiration.

", - "SpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "SpotInstanceRequest$SpotPrice": "

The maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "SpotInstanceRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

", - "SpotInstanceRequest$AvailabilityZoneGroup": "

The Availability Zone group. If you specify the same Availability Zone group for all Spot instance requests, all Spot instances are launched in the same Availability Zone.

", - "SpotInstanceRequest$InstanceId": "

The instance ID, if an instance has been launched to fulfill the Spot instance request.

", - "SpotInstanceRequest$LaunchedAvailabilityZone": "

The Availability Zone in which the bid is launched.

", - "SpotInstanceRequestIdList$member": null, - "SpotInstanceStateFault$Code": "

The reason code for the Spot instance state change.

", - "SpotInstanceStateFault$Message": "

The message for the Spot instance state change.

", - "SpotInstanceStatus$Code": "

The status code.

", - "SpotInstanceStatus$Message": "

The description for the status code.

", - "SpotPlacement$AvailabilityZone": "

The Availability Zone.

", - "SpotPlacement$GroupName": "

The name of the placement group (for cluster instances).

", - "SpotPrice$SpotPrice": "

The maximum price (bid) that you are willing to pay for a Spot instance.

", - "SpotPrice$AvailabilityZone": "

The Availability Zone.

", - "StartInstancesRequest$AdditionalInfo": "

Reserved.

", - "StateReason$Code": "

The reason code for the state change.

", - "StateReason$Message": "

The message for the state change.

  • Server.SpotInstanceTermination: A Spot Instance was terminated due to an increase in the market price.

  • Server.InternalError: An internal error occurred during instance launch, resulting in termination.

  • Server.InsufficientInstanceCapacity: There was insufficient instance capacity to satisfy the launch request.

  • Client.InternalError: A client error caused the instance to terminate on launch.

  • Client.InstanceInitiatedShutdown: The instance was shut down using the shutdown -h command from the instance.

  • Client.UserInitiatedShutdown: The instance was shut down using the Amazon EC2 API.

  • Client.VolumeLimitExceeded: The volume limit was exceeded.

  • Client.InvalidSnapshot.NotFound: The specified snapshot was not found.

", - "Subnet$SubnetId": "

The ID of the subnet.

", - "Subnet$VpcId": "

The ID of the VPC the subnet is in.

", - "Subnet$CidrBlock": "

The CIDR block assigned to the subnet.

", - "Subnet$AvailabilityZone": "

The Availability Zone of the subnet.

", - "SubnetIdStringList$member": null, - "Tag$Key": "

The key of the tag.

Constraints: Tag keys are case-sensitive and accept a maximum of 127 Unicode characters. May not begin with aws:

", - "Tag$Value": "

The value of the tag.

Constraints: Tag values are case-sensitive and accept a maximum of 255 Unicode characters.

", - "TagDescription$ResourceId": "

The ID of the resource. For example, ami-1a2b3c4d.

", - "TagDescription$Key": "

The tag key.

", - "TagDescription$Value": "

The tag value.

", - "UnassignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "UnsuccessfulItem$ResourceId": "

The ID of the resource.

", - "UnsuccessfulItemError$Code": "

The error code.

", - "UnsuccessfulItemError$Message": "

The error message accompanying the error code.

", - "UserBucket$S3Bucket": "

The name of the S3 bucket where the disk image is located.

", - "UserBucket$S3Key": "

The key for the disk image.

", - "UserBucketDetails$S3Bucket": "

The S3 bucket from which the disk image was created.

", - "UserBucketDetails$S3Key": "

The key from which the disk image was created.

", - "UserData$Data": "

The Base64-encoded MIME user data for the instance.

", - "UserGroupStringList$member": null, - "UserIdGroupPair$UserId": "

The ID of an AWS account. EC2-Classic only.

", - "UserIdGroupPair$GroupName": "

The name of the security group. In a request, use this parameter for a security group in EC2-Classic or a default VPC only. For a security group in a nondefault VPC, use GroupId.

", - "UserIdGroupPair$GroupId": "

The ID of the security group.

", - "UserIdStringList$member": null, - "ValueStringList$member": null, - "VgwTelemetry$OutsideIpAddress": "

The Internet-routable IP address of the virtual private gateway's outside interface.

", - "VgwTelemetry$StatusMessage": "

If an error occurs, a description of the error.

", - "Volume$VolumeId": "

The ID of the volume.

", - "Volume$SnapshotId": "

The snapshot from which the volume was created, if applicable.

", - "Volume$AvailabilityZone": "

The Availability Zone for the volume.

", - "Volume$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the volume.

", - "VolumeAttachment$VolumeId": "

The ID of the volume.

", - "VolumeAttachment$InstanceId": "

The ID of the instance.

", - "VolumeAttachment$Device": "

The device name.

", - "VolumeIdStringList$member": null, - "VolumeStatusAction$Code": "

The code identifying the operation, for example, enable-volume-io.

", - "VolumeStatusAction$Description": "

A description of the operation.

", - "VolumeStatusAction$EventType": "

The event type associated with this operation.

", - "VolumeStatusAction$EventId": "

The ID of the event associated with this operation.

", - "VolumeStatusDetails$Status": "

The intended status of the volume status.

", - "VolumeStatusEvent$EventType": "

The type of this event.

", - "VolumeStatusEvent$Description": "

A description of the event.

", - "VolumeStatusEvent$EventId": "

The ID of this event.

", - "VolumeStatusItem$VolumeId": "

The volume ID.

", - "VolumeStatusItem$AvailabilityZone": "

The Availability Zone of the volume.

", - "Vpc$VpcId": "

The ID of the VPC.

", - "Vpc$CidrBlock": "

The CIDR block for the VPC.

", - "Vpc$DhcpOptionsId": "

The ID of the set of DHCP options you've associated with the VPC (or default if the default options are associated with the VPC).

", - "VpcAttachment$VpcId": "

The ID of the VPC.

", - "VpcClassicLink$VpcId": "

The ID of the VPC.

", - "VpcClassicLinkIdList$member": null, - "VpcEndpoint$VpcEndpointId": "

The ID of the VPC endpoint.

", - "VpcEndpoint$VpcId": "

The ID of the VPC to which the endpoint is associated.

", - "VpcEndpoint$ServiceName": "

The name of the AWS service to which the endpoint is associated.

", - "VpcEndpoint$PolicyDocument": "

The policy document associated with the endpoint.

", - "VpcIdStringList$member": null, - "VpcPeeringConnection$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "VpcPeeringConnectionStateReason$Message": "

A message that provides more information about the status, if applicable.

", - "VpcPeeringConnectionVpcInfo$CidrBlock": "

The CIDR block for the VPC.

", - "VpcPeeringConnectionVpcInfo$OwnerId": "

The AWS account ID of the VPC owner.

", - "VpcPeeringConnectionVpcInfo$VpcId": "

The ID of the VPC.

", - "VpnConnection$VpnConnectionId": "

The ID of the VPN connection.

", - "VpnConnection$CustomerGatewayConfiguration": "

The configuration information for the VPN connection's customer gateway (in the native XML format). This element is always present in the CreateVpnConnection response; however, it's present in the DescribeVpnConnections response only if the VPN connection is in the pending or available state.

", - "VpnConnection$CustomerGatewayId": "

The ID of the customer gateway at your end of the VPN connection.

", - "VpnConnection$VpnGatewayId": "

The ID of the virtual private gateway at the AWS side of the VPN connection.

", - "VpnConnectionIdStringList$member": null, - "VpnGateway$VpnGatewayId": "

The ID of the virtual private gateway.

", - "VpnGateway$AvailabilityZone": "

The Availability Zone where the virtual private gateway was created.

", - "VpnGatewayIdStringList$member": null, - "VpnStaticRoute$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer data center.

", - "ZoneNameStringList$member": null, - "NewDhcpConfiguration$Key": null, - "RequestSpotLaunchSpecification$ImageId": "

The ID of the AMI.

", - "RequestSpotLaunchSpecification$KeyName": "

The name of the key pair.

", - "RequestSpotLaunchSpecification$UserData": "

The Base64-encoded MIME user data to make available to the instances.

", - "RequestSpotLaunchSpecification$AddressingType": "

Deprecated.

", - "RequestSpotLaunchSpecification$KernelId": "

The ID of the kernel.

", - "RequestSpotLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "RequestSpotLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

" - } - }, - "Subnet": { - "base": "

Describes a subnet.

", - "refs": { - "CreateSubnetResult$Subnet": "

Information about the subnet.

", - "SubnetList$member": null - } - }, - "SubnetIdStringList": { - "base": null, - "refs": { - "DescribeSubnetsRequest$SubnetIds": "

One or more subnet IDs.

Default: Describes all your subnets.

" - } - }, - "SubnetList": { - "base": null, - "refs": { - "DescribeSubnetsResult$Subnets": "

Information about one or more subnets.

" - } - }, - "SubnetState": { - "base": null, - "refs": { - "Subnet$State": "

The current state of the subnet.

" - } - }, - "SummaryStatus": { - "base": null, - "refs": { - "InstanceStatusSummary$Status": "

The status.

" - } - }, - "Tag": { - "base": "

Describes a tag.

", - "refs": { - "TagList$member": null - } - }, - "TagDescription": { - "base": "

Describes a tag.

", - "refs": { - "TagDescriptionList$member": null - } - }, - "TagDescriptionList": { - "base": null, - "refs": { - "DescribeTagsResult$Tags": "

A list of tags.

" - } - }, - "TagList": { - "base": null, - "refs": { - "ClassicLinkInstance$Tags": "

Any tags assigned to the instance.

", - "ConversionTask$Tags": "

Any tags assigned to the task.

", - "CreateTagsRequest$Tags": "

One or more tags. The value parameter is required, but if you don't want the tag to have a value, specify the parameter with no value, and we set the value to an empty string.

", - "CustomerGateway$Tags": "

Any tags assigned to the customer gateway.

", - "DeleteTagsRequest$Tags": "

One or more tags to delete. If you omit the value parameter, we delete the tag regardless of its value. If you specify this parameter with an empty string as the value, we delete the key only if its value is an empty string.

", - "DhcpOptions$Tags": "

Any tags assigned to the DHCP options set.

", - "Image$Tags": "

Any tags assigned to the image.

", - "Instance$Tags": "

Any tags assigned to the instance.

", - "InternetGateway$Tags": "

Any tags assigned to the Internet gateway.

", - "NetworkAcl$Tags": "

Any tags assigned to the network ACL.

", - "NetworkInterface$TagSet": "

Any tags assigned to the network interface.

", - "ReservedInstances$Tags": "

Any tags assigned to the resource.

", - "ReservedInstancesListing$Tags": "

Any tags assigned to the resource.

", - "RouteTable$Tags": "

Any tags assigned to the route table.

", - "SecurityGroup$Tags": "

Any tags assigned to the security group.

", - "Snapshot$Tags": "

Any tags assigned to the snapshot.

", - "SpotInstanceRequest$Tags": "

Any tags assigned to the resource.

", - "Subnet$Tags": "

Any tags assigned to the subnet.

", - "Volume$Tags": "

Any tags assigned to the volume.

", - "Vpc$Tags": "

Any tags assigned to the VPC.

", - "VpcClassicLink$Tags": "

Any tags assigned to the VPC.

", - "VpcPeeringConnection$Tags": "

Any tags assigned to the resource.

", - "VpnConnection$Tags": "

Any tags assigned to the VPN connection.

", - "VpnGateway$Tags": "

Any tags assigned to the virtual private gateway.

" - } - }, - "TelemetryStatus": { - "base": null, - "refs": { - "VgwTelemetry$Status": "

The status of the VPN tunnel.

" - } - }, - "Tenancy": { - "base": null, - "refs": { - "CreateVpcRequest$InstanceTenancy": "

The supported tenancy options for instances launched into the VPC. A value of default means that instances can be launched with any tenancy; a value of dedicated means all instances launched into the VPC are launched as dedicated tenancy instances regardless of the tenancy assigned to the instance at launch. Dedicated tenancy instances run on single-tenant hardware.

Default: default

", - "DescribeReservedInstancesOfferingsRequest$InstanceTenancy": "

The tenancy of the Reserved Instance offering. A Reserved Instance with dedicated tenancy runs on single-tenant hardware and can only be launched within a VPC.

Default: default

", - "Placement$Tenancy": "

The tenancy of the instance (if the instance is running in a VPC). An instance with a tenancy of dedicated runs on single-tenant hardware.

", - "ReservedInstances$InstanceTenancy": "

The tenancy of the reserved instance.

", - "ReservedInstancesOffering$InstanceTenancy": "

The tenancy of the reserved instance.

", - "Vpc$InstanceTenancy": "

The allowed tenancy of instances launched into the VPC.

" - } - }, - "TerminateInstancesRequest": { - "base": null, - "refs": { - } - }, - "TerminateInstancesResult": { - "base": null, - "refs": { - } - }, - "TrafficType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$TrafficType": "

The type of traffic to log.

", - "FlowLog$TrafficType": "

The type of traffic captured for the flow log.

" - } - }, - "UnassignPrivateIpAddressesRequest": { - "base": null, - "refs": { - } - }, - "UnmonitorInstancesRequest": { - "base": null, - "refs": { - } - }, - "UnmonitorInstancesResult": { - "base": null, - "refs": { - } - }, - "UnsuccessfulItem": { - "base": "

Information about items that were not successfully processed in a batch call.

", - "refs": { - "UnsuccessfulItemSet$member": null - } - }, - "UnsuccessfulItemError": { - "base": "

Information about the error that occurred. For more information about errors, see Error Codes.

", - "refs": { - "UnsuccessfulItem$Error": "

Information about the error.

" - } - }, - "UnsuccessfulItemSet": { - "base": null, - "refs": { - "CreateFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be created successfully.

", - "DeleteFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be deleted successfully.

", - "DeleteVpcEndpointsResult$Unsuccessful": "

Information about the endpoints that were not successfully deleted.

" - } - }, - "UserBucket": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "ImageDiskContainer$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotDiskContainer$UserBucket": null - } - }, - "UserBucketDetails": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "SnapshotDetail$UserBucket": null, - "SnapshotTaskDetail$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserData": { - "base": "

Describes the user data to be made available to an instance.

", - "refs": { - "ImportInstanceLaunchSpecification$UserData": "

The Base64-encoded MIME user data to be made available to the instance.

" - } - }, - "UserGroupStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserGroups": "

One or more user groups. This is only valid when modifying the launchPermission attribute.

" - } - }, - "UserIdGroupPair": { - "base": "

Describes a security group and AWS account ID pair.

", - "refs": { - "UserIdGroupPairList$member": null - } - }, - "UserIdGroupPairList": { - "base": null, - "refs": { - "IpPermission$UserIdGroupPairs": "

One or more security group and AWS account ID pairs.

" - } - }, - "UserIdStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserIds": "

One or more AWS account IDs. This is only valid when modifying the launchPermission attribute.

", - "ModifySnapshotAttributeRequest$UserIds": "

The account ID to modify for the snapshot.

" - } - }, - "ValueStringList": { - "base": null, - "refs": { - "CancelSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "CreateFlowLogsRequest$ResourceIds": "

One or more subnet, network interface, or VPC IDs.

", - "CreateFlowLogsResult$FlowLogIds": "

The IDs of the flow logs.

", - "CreateVpcEndpointRequest$RouteTableIds": "

One or more route table IDs.

", - "DeleteFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DeleteVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DescribeInternetGatewaysRequest$InternetGatewayIds": "

One or more Internet gateway IDs.

Default: Describes all your Internet gateways.

", - "DescribeMovingAddressesRequest$PublicIps": "

One or more Elastic IP addresses.

", - "DescribeNetworkAclsRequest$NetworkAclIds": "

One or more network ACL IDs.

Default: Describes all your network ACLs.

", - "DescribePrefixListsRequest$PrefixListIds": "

One or more prefix list IDs.

", - "DescribeRouteTablesRequest$RouteTableIds": "

One or more route table IDs.

Default: Describes all your route tables.

", - "DescribeSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "DescribeVpcEndpointServicesResult$ServiceNames": "

A list of supported AWS services.

", - "DescribeVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeVpcPeeringConnectionsRequest$VpcPeeringConnectionIds": "

One or more VPC peering connection IDs.

Default: Describes all your VPC peering connections.

", - "Filter$Values": "

One or more filter values. Filter values are case-sensitive.

", - "ModifyVpcEndpointRequest$AddRouteTableIds": "

One or more route tables IDs to associate with the endpoint.

", - "ModifyVpcEndpointRequest$RemoveRouteTableIds": "

One or more route table IDs to disassociate from the endpoint.

", - "PrefixList$Cidrs": "

The IP address range of the AWS service.

", - "VpcEndpoint$RouteTableIds": "

One or more route tables associated with the endpoint.

", - "NewDhcpConfiguration$Values": null, - "RequestSpotLaunchSpecification$SecurityGroups": null, - "RequestSpotLaunchSpecification$SecurityGroupIds": null - } - }, - "VgwTelemetry": { - "base": "

Describes telemetry for a VPN tunnel.

", - "refs": { - "VgwTelemetryList$member": null - } - }, - "VgwTelemetryList": { - "base": null, - "refs": { - "VpnConnection$VgwTelemetry": "

Information about the VPN tunnel.

" - } - }, - "VirtualizationType": { - "base": null, - "refs": { - "Image$VirtualizationType": "

The type of virtualization of the AMI.

", - "Instance$VirtualizationType": "

The virtualization type of the instance.

" - } - }, - "Volume": { - "base": "

Describes a volume.

", - "refs": { - "VolumeList$member": null - } - }, - "VolumeAttachment": { - "base": "

Describes volume attachment details.

", - "refs": { - "VolumeAttachmentList$member": null - } - }, - "VolumeAttachmentList": { - "base": null, - "refs": { - "Volume$Attachments": "

Information about the volume attachments.

" - } - }, - "VolumeAttachmentState": { - "base": null, - "refs": { - "VolumeAttachment$State": "

The attachment state of the volume.

" - } - }, - "VolumeAttributeName": { - "base": null, - "refs": { - "DescribeVolumeAttributeRequest$Attribute": "

The instance attribute.

" - } - }, - "VolumeDetail": { - "base": "

Describes an EBS volume.

", - "refs": { - "DiskImage$Volume": "

Information about the volume.

", - "ImportVolumeRequest$Volume": "

The volume size.

" - } - }, - "VolumeIdStringList": { - "base": null, - "refs": { - "DescribeVolumeStatusRequest$VolumeIds": "

One or more volume IDs.

Default: Describes all your volumes.

", - "DescribeVolumesRequest$VolumeIds": "

One or more volume IDs.

" - } - }, - "VolumeList": { - "base": null, - "refs": { - "DescribeVolumesResult$Volumes": "

Information about the volumes.

" - } - }, - "VolumeState": { - "base": null, - "refs": { - "Volume$State": "

The volume state.

" - } - }, - "VolumeStatusAction": { - "base": "

Describes a volume status operation code.

", - "refs": { - "VolumeStatusActionsList$member": null - } - }, - "VolumeStatusActionsList": { - "base": null, - "refs": { - "VolumeStatusItem$Actions": "

The details of the operation.

" - } - }, - "VolumeStatusDetails": { - "base": "

Describes a volume status.

", - "refs": { - "VolumeStatusDetailsList$member": null - } - }, - "VolumeStatusDetailsList": { - "base": null, - "refs": { - "VolumeStatusInfo$Details": "

The details of the volume status.

" - } - }, - "VolumeStatusEvent": { - "base": "

Describes a volume status event.

", - "refs": { - "VolumeStatusEventsList$member": null - } - }, - "VolumeStatusEventsList": { - "base": null, - "refs": { - "VolumeStatusItem$Events": "

A list of events associated with the volume.

" - } - }, - "VolumeStatusInfo": { - "base": "

Describes the status of a volume.

", - "refs": { - "VolumeStatusItem$VolumeStatus": "

The volume status.

" - } - }, - "VolumeStatusInfoStatus": { - "base": null, - "refs": { - "VolumeStatusInfo$Status": "

The status of the volume.

" - } - }, - "VolumeStatusItem": { - "base": "

Describes the volume status.

", - "refs": { - "VolumeStatusList$member": null - } - }, - "VolumeStatusList": { - "base": null, - "refs": { - "DescribeVolumeStatusResult$VolumeStatuses": "

A list of volumes.

" - } - }, - "VolumeStatusName": { - "base": null, - "refs": { - "VolumeStatusDetails$Name": "

The name of the volume status.

" - } - }, - "VolumeType": { - "base": null, - "refs": { - "CreateVolumeRequest$VolumeType": "

The volume type. This can be gp2 for General Purpose (SSD) volumes, io1 for Provisioned IOPS (SSD) volumes, or standard for Magnetic volumes.

Default: standard

", - "EbsBlockDevice$VolumeType": "

The volume type. gp2 for General Purpose (SSD) volumes, io1 for Provisioned IOPS (SSD) volumes, and standard for Magnetic volumes.

Default: standard

", - "Volume$VolumeType": "

The volume type. This can be gp2 for General Purpose (SSD) volumes, io1 for Provisioned IOPS (SSD) volumes, or standard for Magnetic volumes.

" - } - }, - "Vpc": { - "base": "

Describes a VPC.

", - "refs": { - "CreateVpcResult$Vpc": "

Information about the VPC.

", - "VpcList$member": null - } - }, - "VpcAttachment": { - "base": "

Describes an attachment between a virtual private gateway and a VPC.

", - "refs": { - "AttachVpnGatewayResult$VpcAttachment": "

Information about the attachment.

", - "VpcAttachmentList$member": null - } - }, - "VpcAttachmentList": { - "base": null, - "refs": { - "VpnGateway$VpcAttachments": "

Any VPCs attached to the virtual private gateway.

" - } - }, - "VpcAttributeName": { - "base": null, - "refs": { - "DescribeVpcAttributeRequest$Attribute": "

The VPC attribute.

" - } - }, - "VpcClassicLink": { - "base": "

Describes whether a VPC is enabled for ClassicLink.

", - "refs": { - "VpcClassicLinkList$member": null - } - }, - "VpcClassicLinkIdList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkRequest$VpcIds": "

One or more VPCs for which you want to describe the ClassicLink status.

" - } - }, - "VpcClassicLinkList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkResult$Vpcs": "

The ClassicLink status of one or more VPCs.

" - } - }, - "VpcEndpoint": { - "base": "

Describes a VPC endpoint.

", - "refs": { - "CreateVpcEndpointResult$VpcEndpoint": "

Information about the endpoint.

", - "VpcEndpointSet$member": null - } - }, - "VpcEndpointSet": { - "base": null, - "refs": { - "DescribeVpcEndpointsResult$VpcEndpoints": "

Information about the endpoints.

" - } - }, - "VpcIdStringList": { - "base": null, - "refs": { - "DescribeVpcsRequest$VpcIds": "

One or more VPC IDs.

Default: Describes all your VPCs.

" - } - }, - "VpcList": { - "base": null, - "refs": { - "DescribeVpcsResult$Vpcs": "

Information about one or more VPCs.

" - } - }, - "VpcPeeringConnection": { - "base": "

Describes a VPC peering connection.

", - "refs": { - "AcceptVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "CreateVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "VpcPeeringConnectionList$member": null - } - }, - "VpcPeeringConnectionList": { - "base": null, - "refs": { - "DescribeVpcPeeringConnectionsResult$VpcPeeringConnections": "

Information about the VPC peering connections.

" - } - }, - "VpcPeeringConnectionStateReason": { - "base": "

Describes the status of a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$Status": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionStateReasonCode": { - "base": null, - "refs": { - "VpcPeeringConnectionStateReason$Code": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionVpcInfo": { - "base": "

Describes a VPC in a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$AccepterVpcInfo": "

The information of the peer VPC.

", - "VpcPeeringConnection$RequesterVpcInfo": "

The information of the requester VPC.

" - } - }, - "VpcState": { - "base": null, - "refs": { - "Vpc$State": "

The current state of the VPC.

" - } - }, - "VpnConnection": { - "base": "

Describes a VPN connection.

", - "refs": { - "CreateVpnConnectionResult$VpnConnection": "

Information about the VPN connection.

", - "VpnConnectionList$member": null - } - }, - "VpnConnectionIdStringList": { - "base": null, - "refs": { - "DescribeVpnConnectionsRequest$VpnConnectionIds": "

One or more VPN connection IDs.

Default: Describes your VPN connections.

" - } - }, - "VpnConnectionList": { - "base": null, - "refs": { - "DescribeVpnConnectionsResult$VpnConnections": "

Information about one or more VPN connections.

" - } - }, - "VpnConnectionOptions": { - "base": "

Describes VPN connection options.

", - "refs": { - "VpnConnection$Options": "

The VPN connection options.

" - } - }, - "VpnConnectionOptionsSpecification": { - "base": "

Describes VPN connection options.

", - "refs": { - "CreateVpnConnectionRequest$Options": "

Indicates whether the VPN connection requires static routes. If you are creating a VPN connection for a device that does not support BGP, you must specify true.

Default: false

" - } - }, - "VpnGateway": { - "base": "

Describes a virtual private gateway.

", - "refs": { - "CreateVpnGatewayResult$VpnGateway": "

Information about the virtual private gateway.

", - "VpnGatewayList$member": null - } - }, - "VpnGatewayIdStringList": { - "base": null, - "refs": { - "DescribeVpnGatewaysRequest$VpnGatewayIds": "

One or more virtual private gateway IDs.

Default: Describes all your virtual private gateways.

" - } - }, - "VpnGatewayList": { - "base": null, - "refs": { - "DescribeVpnGatewaysResult$VpnGateways": "

Information about one or more virtual private gateways.

" - } - }, - "VpnState": { - "base": null, - "refs": { - "VpnConnection$State": "

The current state of the VPN connection.

", - "VpnGateway$State": "

The current state of the virtual private gateway.

", - "VpnStaticRoute$State": "

The current state of the static route.

" - } - }, - "VpnStaticRoute": { - "base": "

Describes a static route for a VPN connection.

", - "refs": { - "VpnStaticRouteList$member": null - } - }, - "VpnStaticRouteList": { - "base": null, - "refs": { - "VpnConnection$Routes": "

The static routes associated with the VPN connection.

" - } - }, - "VpnStaticRouteSource": { - "base": null, - "refs": { - "VpnStaticRoute$Source": "

Indicates how the routes were provided.

" - } - }, - "ZoneNameStringList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesRequest$ZoneNames": "

The names of one or more Availability Zones.

" - } - }, - "NewDhcpConfigurationList": { - "base": null, - "refs": { - "CreateDhcpOptionsRequest$DhcpConfigurations": "

A DHCP configuration option.

" - } - }, - "NewDhcpConfiguration": { - "base": null, - "refs": { - "NewDhcpConfigurationList$member": null - } - }, - "DhcpConfigurationValueList": { - "base": null, - "refs": { - "DhcpConfiguration$Values": "

One or more values for the DHCP option.

" - } - }, - "Blob": { - "base": null, - "refs": { - "ImportKeyPairRequest$PublicKeyMaterial": "

The public key. You must base64 encode the public key material before sending it to AWS.

", - "S3Storage$UploadPolicy": "

A Base64-encoded Amazon S3 upload policy that gives Amazon EC2 permission to upload items into Amazon S3 on your behalf.

", - "BlobAttributeValue$Value": null - } - }, - "BlobAttributeValue": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$UserData": "

Changes the instance's user data to the specified value.

" - } - }, - "RequestSpotLaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "RequestSpotInstancesRequest$LaunchSpecification": null - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/paginators-1.json deleted file mode 100644 index 740f2e36a..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/paginators-1.json +++ /dev/null @@ -1,125 +0,0 @@ -{ - "pagination": { - "DescribeAccountAttributes": { - "result_key": "AccountAttributes" - }, - "DescribeAddresses": { - "result_key": "Addresses" - }, - "DescribeAvailabilityZones": { - "result_key": "AvailabilityZones" - }, - "DescribeBundleTasks": { - "result_key": "BundleTasks" - }, - "DescribeConversionTasks": { - "result_key": "ConversionTasks" - }, - "DescribeCustomerGateways": { - "result_key": "CustomerGateways" - }, - "DescribeDhcpOptions": { - "result_key": "DhcpOptions" - }, - "DescribeExportTasks": { - "result_key": "ExportTasks" - }, - "DescribeImages": { - "result_key": "Images" - }, - "DescribeInstanceStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "InstanceStatuses" - }, - "DescribeInstances": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Reservations" - }, - "DescribeInternetGateways": { - "result_key": "InternetGateways" - }, - "DescribeKeyPairs": { - "result_key": "KeyPairs" - }, - "DescribeNetworkAcls": { - "result_key": "NetworkAcls" - }, - "DescribeNetworkInterfaces": { - "result_key": "NetworkInterfaces" - }, - "DescribePlacementGroups": { - "result_key": "PlacementGroups" - }, - "DescribeRegions": { - "result_key": "Regions" - }, - "DescribeReservedInstances": { - "result_key": "ReservedInstances" - }, - "DescribeReservedInstancesListings": { - "result_key": "ReservedInstancesListings" - }, - "DescribeReservedInstancesOfferings": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "ReservedInstancesOfferings" - }, - "DescribeReservedInstancesModifications": { - "input_token": "NextToken", - "output_token": "NextToken", - "result_key": "ReservedInstancesModifications" - }, - "DescribeRouteTables": { - "result_key": "RouteTables" - }, - "DescribeSecurityGroups": { - "result_key": "SecurityGroups" - }, - "DescribeSnapshots": { - "input_token": "NextToken", - "output_token": "NextToken", - "result_key": "Snapshots" - }, - "DescribeSpotInstanceRequests": { - "result_key": "SpotInstanceRequests" - }, - "DescribeSpotPriceHistory": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotPriceHistory" - }, - "DescribeSubnets": { - "result_key": "Subnets" - }, - "DescribeTags": { - "result_key": "Tags" - }, - "DescribeVolumeStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "VolumeStatuses" - }, - "DescribeVolumes": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Volumes" - }, - "DescribeVpcs": { - "result_key": "Vpcs" - }, - "DescribeVpnConnections": { - "result_key": "VpnConnections" - }, - "DescribeVpnGateways": { - "result_key": "VpnGateways" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/waiters-2.json deleted file mode 100644 index 0599f2422..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-04-15/waiters-2.json +++ /dev/null @@ -1,494 +0,0 @@ -{ - "version": 2, - "waiters": { - "InstanceExists": { - "delay": 5, - "maxAttempts": 40, - "operation": "DescribeInstances", - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidInstanceIDNotFound", - "state": "retry" - } - ] - }, - "BundleTaskComplete": { - "delay": 15, - "operation": "DescribeBundleTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "complete", - "matcher": "pathAll", - "state": "success", - "argument": "BundleTasks[].State" - }, - { - "expected": "failed", - "matcher": "pathAny", - "state": "failure", - "argument": "BundleTasks[].State" - } - ] - }, - "ConversionTaskCancelled": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskCompleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelled", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelling", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskDeleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "CustomerGatewayAvailable": { - "delay": 15, - "operation": "DescribeCustomerGateways", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - } - ] - }, - "ExportTaskCancelled": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ExportTaskCompleted": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ImageAvailable": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "Images[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "Images[].State", - "expected": "failed" - } - ] - }, - "InstanceRunning": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "running", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "shutting-down", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "InstanceStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].InstanceStatus.Status", - "expected": "ok" - } - ] - }, - "InstanceStopped": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "stopped", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "InstanceTerminated": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "terminated", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "KeyPairExists": { - "operation": "DescribeKeyPairs", - "delay": 5, - "maxAttempts": 6, - "acceptors": [ - { - "expected": true, - "matcher": "pathAll", - "state": "success", - "argument": "length(KeyPairs[].KeyName) > `0`" - }, - { - "expected": "InvalidKeyPairNotFound", - "matcher": "error", - "state": "retry" - } - ] - }, - "NetworkInterfaceAvailable": { - "operation": "DescribeNetworkInterfaces", - "delay": 20, - "maxAttempts": 10, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "NetworkInterfaces[].Status" - }, - { - "expected": "InvalidNetworkInterfaceIDNotFound", - "matcher": "error", - "state": "failure" - } - ] - }, - "PasswordDataAvailable": { - "operation": "GetPasswordData", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "path", - "argument": "length(PasswordData) > `0`", - "expected": true - } - ] - }, - "SnapshotCompleted": { - "delay": 15, - "operation": "DescribeSnapshots", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "Snapshots[].State" - } - ] - }, - "SpotInstanceRequestFulfilled": { - "operation": "DescribeSpotInstanceRequests", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "fulfilled" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "schedule-expired" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "canceled-before-fulfillment" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "bad-parameters" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "system-error" - } - ] - }, - "SubnetAvailable": { - "delay": 15, - "operation": "DescribeSubnets", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Subnets[].State" - } - ] - }, - "SystemStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].SystemStatus.Status", - "expected": "ok" - } - ] - }, - "VolumeAvailable": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VolumeDeleted": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "matcher": "error", - "expected": "InvalidVolumeNotFound", - "state": "success" - } - ] - }, - "VolumeInUse": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "in-use", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VpcAvailable": { - "delay": 15, - "operation": "DescribeVpcs", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Vpcs[].State" - } - ] - }, - "VpnConnectionAvailable": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpnConnectionDeleted": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/api-2.json deleted file mode 100644 index c5bb5d9bd..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/api-2.json +++ /dev/null @@ -1,13760 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"ec2-2015-10-01", - "apiVersion":"2015-10-01", - "endpointPrefix":"ec2", - "protocol":"ec2", - "serviceAbbreviation":"Amazon EC2", - "serviceFullName":"Amazon Elastic Compute Cloud", - "signatureVersion":"v4", - "xmlNamespace":"http://ec2.amazonaws.com/doc/2015-10-01" - }, - "operations":{ - "AcceptVpcPeeringConnection":{ - "name":"AcceptVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AcceptVpcPeeringConnectionRequest"}, - "output":{"shape":"AcceptVpcPeeringConnectionResult"} - }, - "AllocateAddress":{ - "name":"AllocateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateAddressRequest"}, - "output":{"shape":"AllocateAddressResult"} - }, - "AllocateHosts":{ - "name":"AllocateHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateHostsRequest"}, - "output":{"shape":"AllocateHostsResult"} - }, - "AssignPrivateIpAddresses":{ - "name":"AssignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssignPrivateIpAddressesRequest"} - }, - "AssociateAddress":{ - "name":"AssociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateAddressRequest"}, - "output":{"shape":"AssociateAddressResult"} - }, - "AssociateDhcpOptions":{ - "name":"AssociateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateDhcpOptionsRequest"} - }, - "AssociateRouteTable":{ - "name":"AssociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateRouteTableRequest"}, - "output":{"shape":"AssociateRouteTableResult"} - }, - "AttachClassicLinkVpc":{ - "name":"AttachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachClassicLinkVpcRequest"}, - "output":{"shape":"AttachClassicLinkVpcResult"} - }, - "AttachInternetGateway":{ - "name":"AttachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachInternetGatewayRequest"} - }, - "AttachNetworkInterface":{ - "name":"AttachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachNetworkInterfaceRequest"}, - "output":{"shape":"AttachNetworkInterfaceResult"} - }, - "AttachVolume":{ - "name":"AttachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVolumeRequest"}, - "output":{"shape":"VolumeAttachment"} - }, - "AttachVpnGateway":{ - "name":"AttachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVpnGatewayRequest"}, - "output":{"shape":"AttachVpnGatewayResult"} - }, - "AuthorizeSecurityGroupEgress":{ - "name":"AuthorizeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupEgressRequest"} - }, - "AuthorizeSecurityGroupIngress":{ - "name":"AuthorizeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupIngressRequest"} - }, - "BundleInstance":{ - "name":"BundleInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"BundleInstanceRequest"}, - "output":{"shape":"BundleInstanceResult"} - }, - "CancelBundleTask":{ - "name":"CancelBundleTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelBundleTaskRequest"}, - "output":{"shape":"CancelBundleTaskResult"} - }, - "CancelConversionTask":{ - "name":"CancelConversionTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelConversionRequest"} - }, - "CancelExportTask":{ - "name":"CancelExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelExportTaskRequest"} - }, - "CancelImportTask":{ - "name":"CancelImportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelImportTaskRequest"}, - "output":{"shape":"CancelImportTaskResult"} - }, - "CancelReservedInstancesListing":{ - "name":"CancelReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelReservedInstancesListingRequest"}, - "output":{"shape":"CancelReservedInstancesListingResult"} - }, - "CancelSpotFleetRequests":{ - "name":"CancelSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotFleetRequestsRequest"}, - "output":{"shape":"CancelSpotFleetRequestsResponse"} - }, - "CancelSpotInstanceRequests":{ - "name":"CancelSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotInstanceRequestsRequest"}, - "output":{"shape":"CancelSpotInstanceRequestsResult"} - }, - "ConfirmProductInstance":{ - "name":"ConfirmProductInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ConfirmProductInstanceRequest"}, - "output":{"shape":"ConfirmProductInstanceResult"} - }, - "CopyImage":{ - "name":"CopyImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyImageRequest"}, - "output":{"shape":"CopyImageResult"} - }, - "CopySnapshot":{ - "name":"CopySnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopySnapshotRequest"}, - "output":{"shape":"CopySnapshotResult"} - }, - "CreateCustomerGateway":{ - "name":"CreateCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateCustomerGatewayRequest"}, - "output":{"shape":"CreateCustomerGatewayResult"} - }, - "CreateDhcpOptions":{ - "name":"CreateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDhcpOptionsRequest"}, - "output":{"shape":"CreateDhcpOptionsResult"} - }, - "CreateFlowLogs":{ - "name":"CreateFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateFlowLogsRequest"}, - "output":{"shape":"CreateFlowLogsResult"} - }, - "CreateImage":{ - "name":"CreateImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateImageRequest"}, - "output":{"shape":"CreateImageResult"} - }, - "CreateInstanceExportTask":{ - "name":"CreateInstanceExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInstanceExportTaskRequest"}, - "output":{"shape":"CreateInstanceExportTaskResult"} - }, - "CreateInternetGateway":{ - "name":"CreateInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInternetGatewayRequest"}, - "output":{"shape":"CreateInternetGatewayResult"} - }, - "CreateKeyPair":{ - "name":"CreateKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateKeyPairRequest"}, - "output":{"shape":"KeyPair"} - }, - "CreateNatGateway":{ - "name":"CreateNatGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNatGatewayRequest"}, - "output":{"shape":"CreateNatGatewayResult"} - }, - "CreateNetworkAcl":{ - "name":"CreateNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclRequest"}, - "output":{"shape":"CreateNetworkAclResult"} - }, - "CreateNetworkAclEntry":{ - "name":"CreateNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclEntryRequest"} - }, - "CreateNetworkInterface":{ - "name":"CreateNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkInterfaceRequest"}, - "output":{"shape":"CreateNetworkInterfaceResult"} - }, - "CreatePlacementGroup":{ - "name":"CreatePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreatePlacementGroupRequest"} - }, - "CreateReservedInstancesListing":{ - "name":"CreateReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateReservedInstancesListingRequest"}, - "output":{"shape":"CreateReservedInstancesListingResult"} - }, - "CreateRoute":{ - "name":"CreateRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteRequest"}, - "output":{"shape":"CreateRouteResult"} - }, - "CreateRouteTable":{ - "name":"CreateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteTableRequest"}, - "output":{"shape":"CreateRouteTableResult"} - }, - "CreateSecurityGroup":{ - "name":"CreateSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSecurityGroupRequest"}, - "output":{"shape":"CreateSecurityGroupResult"} - }, - "CreateSnapshot":{ - "name":"CreateSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSnapshotRequest"}, - "output":{"shape":"Snapshot"} - }, - "CreateSpotDatafeedSubscription":{ - "name":"CreateSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"CreateSpotDatafeedSubscriptionResult"} - }, - "CreateSubnet":{ - "name":"CreateSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSubnetRequest"}, - "output":{"shape":"CreateSubnetResult"} - }, - "CreateTags":{ - "name":"CreateTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateTagsRequest"} - }, - "CreateVolume":{ - "name":"CreateVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVolumeRequest"}, - "output":{"shape":"Volume"} - }, - "CreateVpc":{ - "name":"CreateVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcRequest"}, - "output":{"shape":"CreateVpcResult"} - }, - "CreateVpcEndpoint":{ - "name":"CreateVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcEndpointRequest"}, - "output":{"shape":"CreateVpcEndpointResult"} - }, - "CreateVpcPeeringConnection":{ - "name":"CreateVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcPeeringConnectionRequest"}, - "output":{"shape":"CreateVpcPeeringConnectionResult"} - }, - "CreateVpnConnection":{ - "name":"CreateVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRequest"}, - "output":{"shape":"CreateVpnConnectionResult"} - }, - "CreateVpnConnectionRoute":{ - "name":"CreateVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRouteRequest"} - }, - "CreateVpnGateway":{ - "name":"CreateVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnGatewayRequest"}, - "output":{"shape":"CreateVpnGatewayResult"} - }, - "DeleteCustomerGateway":{ - "name":"DeleteCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteCustomerGatewayRequest"} - }, - "DeleteDhcpOptions":{ - "name":"DeleteDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDhcpOptionsRequest"} - }, - "DeleteFlowLogs":{ - "name":"DeleteFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteFlowLogsRequest"}, - "output":{"shape":"DeleteFlowLogsResult"} - }, - "DeleteInternetGateway":{ - "name":"DeleteInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteInternetGatewayRequest"} - }, - "DeleteKeyPair":{ - "name":"DeleteKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteKeyPairRequest"} - }, - "DeleteNatGateway":{ - "name":"DeleteNatGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNatGatewayRequest"}, - "output":{"shape":"DeleteNatGatewayResult"} - }, - "DeleteNetworkAcl":{ - "name":"DeleteNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclRequest"} - }, - "DeleteNetworkAclEntry":{ - "name":"DeleteNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclEntryRequest"} - }, - "DeleteNetworkInterface":{ - "name":"DeleteNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkInterfaceRequest"} - }, - "DeletePlacementGroup":{ - "name":"DeletePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeletePlacementGroupRequest"} - }, - "DeleteRoute":{ - "name":"DeleteRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteRequest"} - }, - "DeleteRouteTable":{ - "name":"DeleteRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteTableRequest"} - }, - "DeleteSecurityGroup":{ - "name":"DeleteSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSecurityGroupRequest"} - }, - "DeleteSnapshot":{ - "name":"DeleteSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSnapshotRequest"} - }, - "DeleteSpotDatafeedSubscription":{ - "name":"DeleteSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSpotDatafeedSubscriptionRequest"} - }, - "DeleteSubnet":{ - "name":"DeleteSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSubnetRequest"} - }, - "DeleteTags":{ - "name":"DeleteTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteTagsRequest"} - }, - "DeleteVolume":{ - "name":"DeleteVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVolumeRequest"} - }, - "DeleteVpc":{ - "name":"DeleteVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcRequest"} - }, - "DeleteVpcEndpoints":{ - "name":"DeleteVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcEndpointsRequest"}, - "output":{"shape":"DeleteVpcEndpointsResult"} - }, - "DeleteVpcPeeringConnection":{ - "name":"DeleteVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcPeeringConnectionRequest"}, - "output":{"shape":"DeleteVpcPeeringConnectionResult"} - }, - "DeleteVpnConnection":{ - "name":"DeleteVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRequest"} - }, - "DeleteVpnConnectionRoute":{ - "name":"DeleteVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRouteRequest"} - }, - "DeleteVpnGateway":{ - "name":"DeleteVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnGatewayRequest"} - }, - "DeregisterImage":{ - "name":"DeregisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeregisterImageRequest"} - }, - "DescribeAccountAttributes":{ - "name":"DescribeAccountAttributes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAccountAttributesRequest"}, - "output":{"shape":"DescribeAccountAttributesResult"} - }, - "DescribeAddresses":{ - "name":"DescribeAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAddressesRequest"}, - "output":{"shape":"DescribeAddressesResult"} - }, - "DescribeAvailabilityZones":{ - "name":"DescribeAvailabilityZones", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAvailabilityZonesRequest"}, - "output":{"shape":"DescribeAvailabilityZonesResult"} - }, - "DescribeBundleTasks":{ - "name":"DescribeBundleTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeBundleTasksRequest"}, - "output":{"shape":"DescribeBundleTasksResult"} - }, - "DescribeClassicLinkInstances":{ - "name":"DescribeClassicLinkInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeClassicLinkInstancesRequest"}, - "output":{"shape":"DescribeClassicLinkInstancesResult"} - }, - "DescribeConversionTasks":{ - "name":"DescribeConversionTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeConversionTasksRequest"}, - "output":{"shape":"DescribeConversionTasksResult"} - }, - "DescribeCustomerGateways":{ - "name":"DescribeCustomerGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeCustomerGatewaysRequest"}, - "output":{"shape":"DescribeCustomerGatewaysResult"} - }, - "DescribeDhcpOptions":{ - "name":"DescribeDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDhcpOptionsRequest"}, - "output":{"shape":"DescribeDhcpOptionsResult"} - }, - "DescribeExportTasks":{ - "name":"DescribeExportTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeExportTasksRequest"}, - "output":{"shape":"DescribeExportTasksResult"} - }, - "DescribeFlowLogs":{ - "name":"DescribeFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeFlowLogsRequest"}, - "output":{"shape":"DescribeFlowLogsResult"} - }, - "DescribeHosts":{ - "name":"DescribeHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostsRequest"}, - "output":{"shape":"DescribeHostsResult"} - }, - "DescribeIdFormat":{ - "name":"DescribeIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeIdFormatRequest"}, - "output":{"shape":"DescribeIdFormatResult"} - }, - "DescribeImageAttribute":{ - "name":"DescribeImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImageAttributeRequest"}, - "output":{"shape":"ImageAttribute"} - }, - "DescribeImages":{ - "name":"DescribeImages", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImagesRequest"}, - "output":{"shape":"DescribeImagesResult"} - }, - "DescribeImportImageTasks":{ - "name":"DescribeImportImageTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportImageTasksRequest"}, - "output":{"shape":"DescribeImportImageTasksResult"} - }, - "DescribeImportSnapshotTasks":{ - "name":"DescribeImportSnapshotTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportSnapshotTasksRequest"}, - "output":{"shape":"DescribeImportSnapshotTasksResult"} - }, - "DescribeInstanceAttribute":{ - "name":"DescribeInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceAttributeRequest"}, - "output":{"shape":"InstanceAttribute"} - }, - "DescribeInstanceStatus":{ - "name":"DescribeInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceStatusRequest"}, - "output":{"shape":"DescribeInstanceStatusResult"} - }, - "DescribeInstances":{ - "name":"DescribeInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstancesRequest"}, - "output":{"shape":"DescribeInstancesResult"} - }, - "DescribeInternetGateways":{ - "name":"DescribeInternetGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInternetGatewaysRequest"}, - "output":{"shape":"DescribeInternetGatewaysResult"} - }, - "DescribeKeyPairs":{ - "name":"DescribeKeyPairs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeKeyPairsRequest"}, - "output":{"shape":"DescribeKeyPairsResult"} - }, - "DescribeMovingAddresses":{ - "name":"DescribeMovingAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeMovingAddressesRequest"}, - "output":{"shape":"DescribeMovingAddressesResult"} - }, - "DescribeNatGateways":{ - "name":"DescribeNatGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNatGatewaysRequest"}, - "output":{"shape":"DescribeNatGatewaysResult"} - }, - "DescribeNetworkAcls":{ - "name":"DescribeNetworkAcls", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkAclsRequest"}, - "output":{"shape":"DescribeNetworkAclsResult"} - }, - "DescribeNetworkInterfaceAttribute":{ - "name":"DescribeNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfaceAttributeRequest"}, - "output":{"shape":"DescribeNetworkInterfaceAttributeResult"} - }, - "DescribeNetworkInterfaces":{ - "name":"DescribeNetworkInterfaces", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfacesRequest"}, - "output":{"shape":"DescribeNetworkInterfacesResult"} - }, - "DescribePlacementGroups":{ - "name":"DescribePlacementGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePlacementGroupsRequest"}, - "output":{"shape":"DescribePlacementGroupsResult"} - }, - "DescribePrefixLists":{ - "name":"DescribePrefixLists", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePrefixListsRequest"}, - "output":{"shape":"DescribePrefixListsResult"} - }, - "DescribeRegions":{ - "name":"DescribeRegions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRegionsRequest"}, - "output":{"shape":"DescribeRegionsResult"} - }, - "DescribeReservedInstances":{ - "name":"DescribeReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesRequest"}, - "output":{"shape":"DescribeReservedInstancesResult"} - }, - "DescribeReservedInstancesListings":{ - "name":"DescribeReservedInstancesListings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesListingsRequest"}, - "output":{"shape":"DescribeReservedInstancesListingsResult"} - }, - "DescribeReservedInstancesModifications":{ - "name":"DescribeReservedInstancesModifications", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesModificationsRequest"}, - "output":{"shape":"DescribeReservedInstancesModificationsResult"} - }, - "DescribeReservedInstancesOfferings":{ - "name":"DescribeReservedInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesOfferingsRequest"}, - "output":{"shape":"DescribeReservedInstancesOfferingsResult"} - }, - "DescribeRouteTables":{ - "name":"DescribeRouteTables", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRouteTablesRequest"}, - "output":{"shape":"DescribeRouteTablesResult"} - }, - "DescribeScheduledInstanceAvailability":{ - "name":"DescribeScheduledInstanceAvailability", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeScheduledInstanceAvailabilityRequest"}, - "output":{"shape":"DescribeScheduledInstanceAvailabilityResult"} - }, - "DescribeScheduledInstances":{ - "name":"DescribeScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeScheduledInstancesRequest"}, - "output":{"shape":"DescribeScheduledInstancesResult"} - }, - "DescribeSecurityGroupReferences":{ - "name":"DescribeSecurityGroupReferences", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupReferencesRequest"}, - "output":{"shape":"DescribeSecurityGroupReferencesResult"} - }, - "DescribeSecurityGroups":{ - "name":"DescribeSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupsRequest"}, - "output":{"shape":"DescribeSecurityGroupsResult"} - }, - "DescribeSnapshotAttribute":{ - "name":"DescribeSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotAttributeRequest"}, - "output":{"shape":"DescribeSnapshotAttributeResult"} - }, - "DescribeSnapshots":{ - "name":"DescribeSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotsRequest"}, - "output":{"shape":"DescribeSnapshotsResult"} - }, - "DescribeSpotDatafeedSubscription":{ - "name":"DescribeSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"DescribeSpotDatafeedSubscriptionResult"} - }, - "DescribeSpotFleetInstances":{ - "name":"DescribeSpotFleetInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetInstancesRequest"}, - "output":{"shape":"DescribeSpotFleetInstancesResponse"} - }, - "DescribeSpotFleetRequestHistory":{ - "name":"DescribeSpotFleetRequestHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestHistoryRequest"}, - "output":{"shape":"DescribeSpotFleetRequestHistoryResponse"} - }, - "DescribeSpotFleetRequests":{ - "name":"DescribeSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestsRequest"}, - "output":{"shape":"DescribeSpotFleetRequestsResponse"} - }, - "DescribeSpotInstanceRequests":{ - "name":"DescribeSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotInstanceRequestsRequest"}, - "output":{"shape":"DescribeSpotInstanceRequestsResult"} - }, - "DescribeSpotPriceHistory":{ - "name":"DescribeSpotPriceHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotPriceHistoryRequest"}, - "output":{"shape":"DescribeSpotPriceHistoryResult"} - }, - "DescribeStaleSecurityGroups":{ - "name":"DescribeStaleSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeStaleSecurityGroupsRequest"}, - "output":{"shape":"DescribeStaleSecurityGroupsResult"} - }, - "DescribeSubnets":{ - "name":"DescribeSubnets", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSubnetsRequest"}, - "output":{"shape":"DescribeSubnetsResult"} - }, - "DescribeTags":{ - "name":"DescribeTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeTagsRequest"}, - "output":{"shape":"DescribeTagsResult"} - }, - "DescribeVolumeAttribute":{ - "name":"DescribeVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeAttributeRequest"}, - "output":{"shape":"DescribeVolumeAttributeResult"} - }, - "DescribeVolumeStatus":{ - "name":"DescribeVolumeStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeStatusRequest"}, - "output":{"shape":"DescribeVolumeStatusResult"} - }, - "DescribeVolumes":{ - "name":"DescribeVolumes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumesRequest"}, - "output":{"shape":"DescribeVolumesResult"} - }, - "DescribeVpcAttribute":{ - "name":"DescribeVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcAttributeRequest"}, - "output":{"shape":"DescribeVpcAttributeResult"} - }, - "DescribeVpcClassicLink":{ - "name":"DescribeVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkRequest"}, - "output":{"shape":"DescribeVpcClassicLinkResult"} - }, - "DescribeVpcClassicLinkDnsSupport":{ - "name":"DescribeVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"DescribeVpcClassicLinkDnsSupportResult"} - }, - "DescribeVpcEndpointServices":{ - "name":"DescribeVpcEndpointServices", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointServicesRequest"}, - "output":{"shape":"DescribeVpcEndpointServicesResult"} - }, - "DescribeVpcEndpoints":{ - "name":"DescribeVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointsRequest"}, - "output":{"shape":"DescribeVpcEndpointsResult"} - }, - "DescribeVpcPeeringConnections":{ - "name":"DescribeVpcPeeringConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcPeeringConnectionsRequest"}, - "output":{"shape":"DescribeVpcPeeringConnectionsResult"} - }, - "DescribeVpcs":{ - "name":"DescribeVpcs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcsRequest"}, - "output":{"shape":"DescribeVpcsResult"} - }, - "DescribeVpnConnections":{ - "name":"DescribeVpnConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnConnectionsRequest"}, - "output":{"shape":"DescribeVpnConnectionsResult"} - }, - "DescribeVpnGateways":{ - "name":"DescribeVpnGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnGatewaysRequest"}, - "output":{"shape":"DescribeVpnGatewaysResult"} - }, - "DetachClassicLinkVpc":{ - "name":"DetachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachClassicLinkVpcRequest"}, - "output":{"shape":"DetachClassicLinkVpcResult"} - }, - "DetachInternetGateway":{ - "name":"DetachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachInternetGatewayRequest"} - }, - "DetachNetworkInterface":{ - "name":"DetachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachNetworkInterfaceRequest"} - }, - "DetachVolume":{ - "name":"DetachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVolumeRequest"}, - "output":{"shape":"VolumeAttachment"} - }, - "DetachVpnGateway":{ - "name":"DetachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVpnGatewayRequest"} - }, - "DisableVgwRoutePropagation":{ - "name":"DisableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVgwRoutePropagationRequest"} - }, - "DisableVpcClassicLink":{ - "name":"DisableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkRequest"}, - "output":{"shape":"DisableVpcClassicLinkResult"} - }, - "DisableVpcClassicLinkDnsSupport":{ - "name":"DisableVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"DisableVpcClassicLinkDnsSupportResult"} - }, - "DisassociateAddress":{ - "name":"DisassociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateAddressRequest"} - }, - "DisassociateRouteTable":{ - "name":"DisassociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateRouteTableRequest"} - }, - "EnableVgwRoutePropagation":{ - "name":"EnableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVgwRoutePropagationRequest"} - }, - "EnableVolumeIO":{ - "name":"EnableVolumeIO", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVolumeIORequest"} - }, - "EnableVpcClassicLink":{ - "name":"EnableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkRequest"}, - "output":{"shape":"EnableVpcClassicLinkResult"} - }, - "EnableVpcClassicLinkDnsSupport":{ - "name":"EnableVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"EnableVpcClassicLinkDnsSupportResult"} - }, - "GetConsoleOutput":{ - "name":"GetConsoleOutput", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleOutputRequest"}, - "output":{"shape":"GetConsoleOutputResult"} - }, - "GetConsoleScreenshot":{ - "name":"GetConsoleScreenshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleScreenshotRequest"}, - "output":{"shape":"GetConsoleScreenshotResult"} - }, - "GetPasswordData":{ - "name":"GetPasswordData", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetPasswordDataRequest"}, - "output":{"shape":"GetPasswordDataResult"} - }, - "ImportImage":{ - "name":"ImportImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportImageRequest"}, - "output":{"shape":"ImportImageResult"} - }, - "ImportInstance":{ - "name":"ImportInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportInstanceRequest"}, - "output":{"shape":"ImportInstanceResult"} - }, - "ImportKeyPair":{ - "name":"ImportKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportKeyPairRequest"}, - "output":{"shape":"ImportKeyPairResult"} - }, - "ImportSnapshot":{ - "name":"ImportSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportSnapshotRequest"}, - "output":{"shape":"ImportSnapshotResult"} - }, - "ImportVolume":{ - "name":"ImportVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportVolumeRequest"}, - "output":{"shape":"ImportVolumeResult"} - }, - "ModifyHosts":{ - "name":"ModifyHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyHostsRequest"}, - "output":{"shape":"ModifyHostsResult"} - }, - "ModifyIdFormat":{ - "name":"ModifyIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyIdFormatRequest"} - }, - "ModifyImageAttribute":{ - "name":"ModifyImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyImageAttributeRequest"} - }, - "ModifyInstanceAttribute":{ - "name":"ModifyInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstanceAttributeRequest"} - }, - "ModifyInstancePlacement":{ - "name":"ModifyInstancePlacement", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstancePlacementRequest"}, - "output":{"shape":"ModifyInstancePlacementResult"} - }, - "ModifyNetworkInterfaceAttribute":{ - "name":"ModifyNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyNetworkInterfaceAttributeRequest"} - }, - "ModifyReservedInstances":{ - "name":"ModifyReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyReservedInstancesRequest"}, - "output":{"shape":"ModifyReservedInstancesResult"} - }, - "ModifySnapshotAttribute":{ - "name":"ModifySnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySnapshotAttributeRequest"} - }, - "ModifySpotFleetRequest":{ - "name":"ModifySpotFleetRequest", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySpotFleetRequestRequest"}, - "output":{"shape":"ModifySpotFleetRequestResponse"} - }, - "ModifySubnetAttribute":{ - "name":"ModifySubnetAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySubnetAttributeRequest"} - }, - "ModifyVolumeAttribute":{ - "name":"ModifyVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVolumeAttributeRequest"} - }, - "ModifyVpcAttribute":{ - "name":"ModifyVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcAttributeRequest"} - }, - "ModifyVpcEndpoint":{ - "name":"ModifyVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcEndpointRequest"}, - "output":{"shape":"ModifyVpcEndpointResult"} - }, - "ModifyVpcPeeringConnectionOptions":{ - "name":"ModifyVpcPeeringConnectionOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcPeeringConnectionOptionsRequest"}, - "output":{"shape":"ModifyVpcPeeringConnectionOptionsResult"} - }, - "MonitorInstances":{ - "name":"MonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MonitorInstancesRequest"}, - "output":{"shape":"MonitorInstancesResult"} - }, - "MoveAddressToVpc":{ - "name":"MoveAddressToVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MoveAddressToVpcRequest"}, - "output":{"shape":"MoveAddressToVpcResult"} - }, - "PurchaseReservedInstancesOffering":{ - "name":"PurchaseReservedInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedInstancesOfferingRequest"}, - "output":{"shape":"PurchaseReservedInstancesOfferingResult"} - }, - "PurchaseScheduledInstances":{ - "name":"PurchaseScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseScheduledInstancesRequest"}, - "output":{"shape":"PurchaseScheduledInstancesResult"} - }, - "RebootInstances":{ - "name":"RebootInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootInstancesRequest"} - }, - "RegisterImage":{ - "name":"RegisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RegisterImageRequest"}, - "output":{"shape":"RegisterImageResult"} - }, - "RejectVpcPeeringConnection":{ - "name":"RejectVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RejectVpcPeeringConnectionRequest"}, - "output":{"shape":"RejectVpcPeeringConnectionResult"} - }, - "ReleaseAddress":{ - "name":"ReleaseAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseAddressRequest"} - }, - "ReleaseHosts":{ - "name":"ReleaseHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseHostsRequest"}, - "output":{"shape":"ReleaseHostsResult"} - }, - "ReplaceNetworkAclAssociation":{ - "name":"ReplaceNetworkAclAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclAssociationRequest"}, - "output":{"shape":"ReplaceNetworkAclAssociationResult"} - }, - "ReplaceNetworkAclEntry":{ - "name":"ReplaceNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclEntryRequest"} - }, - "ReplaceRoute":{ - "name":"ReplaceRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteRequest"} - }, - "ReplaceRouteTableAssociation":{ - "name":"ReplaceRouteTableAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteTableAssociationRequest"}, - "output":{"shape":"ReplaceRouteTableAssociationResult"} - }, - "ReportInstanceStatus":{ - "name":"ReportInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReportInstanceStatusRequest"} - }, - "RequestSpotFleet":{ - "name":"RequestSpotFleet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotFleetRequest"}, - "output":{"shape":"RequestSpotFleetResponse"} - }, - "RequestSpotInstances":{ - "name":"RequestSpotInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotInstancesRequest"}, - "output":{"shape":"RequestSpotInstancesResult"} - }, - "ResetImageAttribute":{ - "name":"ResetImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetImageAttributeRequest"} - }, - "ResetInstanceAttribute":{ - "name":"ResetInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetInstanceAttributeRequest"} - }, - "ResetNetworkInterfaceAttribute":{ - "name":"ResetNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetNetworkInterfaceAttributeRequest"} - }, - "ResetSnapshotAttribute":{ - "name":"ResetSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetSnapshotAttributeRequest"} - }, - "RestoreAddressToClassic":{ - "name":"RestoreAddressToClassic", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreAddressToClassicRequest"}, - "output":{"shape":"RestoreAddressToClassicResult"} - }, - "RevokeSecurityGroupEgress":{ - "name":"RevokeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupEgressRequest"} - }, - "RevokeSecurityGroupIngress":{ - "name":"RevokeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupIngressRequest"} - }, - "RunInstances":{ - "name":"RunInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunInstancesRequest"}, - "output":{"shape":"Reservation"} - }, - "RunScheduledInstances":{ - "name":"RunScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunScheduledInstancesRequest"}, - "output":{"shape":"RunScheduledInstancesResult"} - }, - "StartInstances":{ - "name":"StartInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StartInstancesRequest"}, - "output":{"shape":"StartInstancesResult"} - }, - "StopInstances":{ - "name":"StopInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StopInstancesRequest"}, - "output":{"shape":"StopInstancesResult"} - }, - "TerminateInstances":{ - "name":"TerminateInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"TerminateInstancesRequest"}, - "output":{"shape":"TerminateInstancesResult"} - }, - "UnassignPrivateIpAddresses":{ - "name":"UnassignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnassignPrivateIpAddressesRequest"} - }, - "UnmonitorInstances":{ - "name":"UnmonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnmonitorInstancesRequest"}, - "output":{"shape":"UnmonitorInstancesResult"} - } - }, - "shapes":{ - "AcceptVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "AcceptVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "AccountAttribute":{ - "type":"structure", - "members":{ - "AttributeName":{ - "shape":"String", - "locationName":"attributeName" - }, - "AttributeValues":{ - "shape":"AccountAttributeValueList", - "locationName":"attributeValueSet" - } - } - }, - "AccountAttributeList":{ - "type":"list", - "member":{ - "shape":"AccountAttribute", - "locationName":"item" - } - }, - "AccountAttributeName":{ - "type":"string", - "enum":[ - "supported-platforms", - "default-vpc" - ] - }, - "AccountAttributeNameStringList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeName", - "locationName":"attributeName" - } - }, - "AccountAttributeValue":{ - "type":"structure", - "members":{ - "AttributeValue":{ - "shape":"String", - "locationName":"attributeValue" - } - } - }, - "AccountAttributeValueList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeValue", - "locationName":"item" - } - }, - "ActiveInstance":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - } - } - }, - "ActiveInstanceSet":{ - "type":"list", - "member":{ - "shape":"ActiveInstance", - "locationName":"item" - } - }, - "Address":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "NetworkInterfaceOwnerId":{ - "shape":"String", - "locationName":"networkInterfaceOwnerId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "AddressList":{ - "type":"list", - "member":{ - "shape":"Address", - "locationName":"item" - } - }, - "Affinity":{ - "type":"string", - "enum":[ - "default", - "host" - ] - }, - "AllocateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Domain":{"shape":"DomainType"} - } - }, - "AllocateAddressResult":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - } - } - }, - "AllocateHostsRequest":{ - "type":"structure", - "required":[ - "InstanceType", - "Quantity", - "AvailabilityZone" - ], - "members":{ - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Quantity":{ - "shape":"Integer", - "locationName":"quantity" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "AllocateHostsResult":{ - "type":"structure", - "members":{ - "HostIds":{ - "shape":"ResponseHostIdList", - "locationName":"hostIdSet" - } - } - }, - "AllocationIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"AllocationId" - } - }, - "AllocationState":{ - "type":"string", - "enum":[ - "available", - "under-assessment", - "permanent-failure", - "released", - "released-permanent-failure" - ] - }, - "AllocationStrategy":{ - "type":"string", - "enum":[ - "lowestPrice", - "diversified" - ] - }, - "ArchitectureValues":{ - "type":"string", - "enum":[ - "i386", - "x86_64" - ] - }, - "AssignPrivateIpAddressesRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AllowReassignment":{ - "shape":"Boolean", - "locationName":"allowReassignment" - } - } - }, - "AssociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "AllowReassociation":{ - "shape":"Boolean", - "locationName":"allowReassociation" - } - } - }, - "AssociateAddressResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AssociateDhcpOptionsRequest":{ - "type":"structure", - "required":[ - "DhcpOptionsId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AssociateRouteTableRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "AssociateRouteTableResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AttachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId", - "Groups" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"SecurityGroupId" - } - } - }, - "AttachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "AttachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "AttachNetworkInterfaceRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "InstanceId", - "DeviceIndex" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - } - } - }, - "AttachNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - } - } - }, - "AttachVolumeRequest":{ - "type":"structure", - "required":[ - "VolumeId", - "InstanceId", - "Device" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"} - } - }, - "AttachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AttachVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpcAttachment":{ - "shape":"VpcAttachment", - "locationName":"attachment" - } - } - }, - "AttachmentStatus":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "AttributeBooleanValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Boolean", - "locationName":"value" - } - } - }, - "AttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "AuthorizeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "AuthorizeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "AutoPlacement":{ - "type":"string", - "enum":[ - "on", - "off" - ] - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "ZoneName":{ - "shape":"String", - "locationName":"zoneName" - }, - "State":{ - "shape":"AvailabilityZoneState", - "locationName":"zoneState" - }, - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Messages":{ - "shape":"AvailabilityZoneMessageList", - "locationName":"messageSet" - } - } - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"item" - } - }, - "AvailabilityZoneMessage":{ - "type":"structure", - "members":{ - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "AvailabilityZoneMessageList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZoneMessage", - "locationName":"item" - } - }, - "AvailabilityZoneState":{ - "type":"string", - "enum":[ - "available", - "information", - "impaired", - "unavailable" - ] - }, - "AvailableCapacity":{ - "type":"structure", - "members":{ - "AvailableInstanceCapacity":{ - "shape":"AvailableInstanceCapacityList", - "locationName":"availableInstanceCapacity" - }, - "AvailableVCpus":{ - "shape":"Integer", - "locationName":"availableVCpus" - } - } - }, - "AvailableInstanceCapacityList":{ - "type":"list", - "member":{ - "shape":"InstanceCapacity", - "locationName":"item" - } - }, - "BatchState":{ - "type":"string", - "enum":[ - "submitted", - "active", - "cancelled", - "failed", - "cancelled_running", - "cancelled_terminating", - "modifying" - ] - }, - "Blob":{"type":"blob"}, - "BlobAttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Blob", - "locationName":"value" - } - } - }, - "BlockDeviceMapping":{ - "type":"structure", - "members":{ - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsBlockDevice", - "locationName":"ebs" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "BlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"item" - } - }, - "BlockDeviceMappingRequestList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "Boolean":{"type":"boolean"}, - "BundleIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"BundleId" - } - }, - "BundleInstanceRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Storage" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "Storage":{"shape":"Storage"} - } - }, - "BundleInstanceResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "BundleTask":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "BundleId":{ - "shape":"String", - "locationName":"bundleId" - }, - "State":{ - "shape":"BundleTaskState", - "locationName":"state" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Storage":{ - "shape":"Storage", - "locationName":"storage" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "BundleTaskError":{ - "shape":"BundleTaskError", - "locationName":"error" - } - } - }, - "BundleTaskError":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "BundleTaskList":{ - "type":"list", - "member":{ - "shape":"BundleTask", - "locationName":"item" - } - }, - "BundleTaskState":{ - "type":"string", - "enum":[ - "pending", - "waiting-for-shutdown", - "bundling", - "storing", - "cancelling", - "complete", - "failed" - ] - }, - "CancelBatchErrorCode":{ - "type":"string", - "enum":[ - "fleetRequestIdDoesNotExist", - "fleetRequestIdMalformed", - "fleetRequestNotInCancellableState", - "unexpectedError" - ] - }, - "CancelBundleTaskRequest":{ - "type":"structure", - "required":["BundleId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleId":{"shape":"String"} - } - }, - "CancelBundleTaskResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "CancelConversionRequest":{ - "type":"structure", - "required":["ConversionTaskId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ReasonMessage":{ - "shape":"String", - "locationName":"reasonMessage" - } - } - }, - "CancelExportTaskRequest":{ - "type":"structure", - "required":["ExportTaskId"], - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - } - } - }, - "CancelImportTaskRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskId":{"shape":"String"}, - "CancelReason":{"shape":"String"} - } - }, - "CancelImportTaskResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "PreviousState":{ - "shape":"String", - "locationName":"previousState" - } - } - }, - "CancelReservedInstancesListingRequest":{ - "type":"structure", - "required":["ReservedInstancesListingId"], - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - } - } - }, - "CancelReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CancelSpotFleetRequestsError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"CancelBatchErrorCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "CancelSpotFleetRequestsErrorItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "Error" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "Error":{ - "shape":"CancelSpotFleetRequestsError", - "locationName":"error" - } - } - }, - "CancelSpotFleetRequestsErrorSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsErrorItem", - "locationName":"item" - } - }, - "CancelSpotFleetRequestsRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestIds", - "TerminateInstances" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "TerminateInstances":{ - "shape":"Boolean", - "locationName":"terminateInstances" - } - } - }, - "CancelSpotFleetRequestsResponse":{ - "type":"structure", - "members":{ - "UnsuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsErrorSet", - "locationName":"unsuccessfulFleetRequestSet" - }, - "SuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsSuccessSet", - "locationName":"successfulFleetRequestSet" - } - } - }, - "CancelSpotFleetRequestsSuccessItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "CurrentSpotFleetRequestState", - "PreviousSpotFleetRequestState" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "CurrentSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"currentSpotFleetRequestState" - }, - "PreviousSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"previousSpotFleetRequestState" - } - } - }, - "CancelSpotFleetRequestsSuccessSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsSuccessItem", - "locationName":"item" - } - }, - "CancelSpotInstanceRequestState":{ - "type":"string", - "enum":[ - "active", - "open", - "closed", - "cancelled", - "completed" - ] - }, - "CancelSpotInstanceRequestsRequest":{ - "type":"structure", - "required":["SpotInstanceRequestIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - } - } - }, - "CancelSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "CancelledSpotInstanceRequests":{ - "shape":"CancelledSpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "CancelledSpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "State":{ - "shape":"CancelSpotInstanceRequestState", - "locationName":"state" - } - } - }, - "CancelledSpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"CancelledSpotInstanceRequest", - "locationName":"item" - } - }, - "ClassicLinkDnsSupport":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkDnsSupported":{ - "shape":"Boolean", - "locationName":"classicLinkDnsSupported" - } - } - }, - "ClassicLinkDnsSupportList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkDnsSupport", - "locationName":"item" - } - }, - "ClassicLinkInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ClassicLinkInstanceList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkInstance", - "locationName":"item" - } - }, - "ClientData":{ - "type":"structure", - "members":{ - "UploadStart":{"shape":"DateTime"}, - "UploadEnd":{"shape":"DateTime"}, - "UploadSize":{"shape":"Double"}, - "Comment":{"shape":"String"} - } - }, - "ConfirmProductInstanceRequest":{ - "type":"structure", - "required":[ - "ProductCode", - "InstanceId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ProductCode":{"shape":"String"}, - "InstanceId":{"shape":"String"} - } - }, - "ConfirmProductInstanceResult":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ContainerFormat":{ - "type":"string", - "enum":["ova"] - }, - "ConversionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "ConversionTask":{ - "type":"structure", - "required":[ - "ConversionTaskId", - "State" - ], - "members":{ - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ExpirationTime":{ - "shape":"String", - "locationName":"expirationTime" - }, - "ImportInstance":{ - "shape":"ImportInstanceTaskDetails", - "locationName":"importInstance" - }, - "ImportVolume":{ - "shape":"ImportVolumeTaskDetails", - "locationName":"importVolume" - }, - "State":{ - "shape":"ConversionTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ConversionTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "CopyImageRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceImageId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceImageId":{"shape":"String"}, - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "ClientToken":{"shape":"String"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopyImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CopySnapshotRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceSnapshotId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceSnapshotId":{"shape":"String"}, - "Description":{"shape":"String"}, - "DestinationRegion":{ - "shape":"String", - "locationName":"destinationRegion" - }, - "PresignedUrl":{ - "shape":"String", - "locationName":"presignedUrl" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopySnapshotResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - } - } - }, - "CreateCustomerGatewayRequest":{ - "type":"structure", - "required":[ - "Type", - "PublicIp", - "BgpAsn" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "PublicIp":{ - "shape":"String", - "locationName":"IpAddress" - }, - "BgpAsn":{"shape":"Integer"} - } - }, - "CreateCustomerGatewayResult":{ - "type":"structure", - "members":{ - "CustomerGateway":{ - "shape":"CustomerGateway", - "locationName":"customerGateway" - } - } - }, - "CreateDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpConfigurations"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpConfigurations":{ - "shape":"NewDhcpConfigurationList", - "locationName":"dhcpConfiguration" - } - } - }, - "CreateDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptions", - "locationName":"dhcpOptions" - } - } - }, - "CreateFlowLogsRequest":{ - "type":"structure", - "required":[ - "ResourceIds", - "ResourceType", - "TrafficType", - "LogGroupName", - "DeliverLogsPermissionArn" - ], - "members":{ - "ResourceIds":{ - "shape":"ValueStringList", - "locationName":"ResourceId" - }, - "ResourceType":{"shape":"FlowLogsResourceType"}, - "TrafficType":{"shape":"TrafficType"}, - "LogGroupName":{"shape":"String"}, - "DeliverLogsPermissionArn":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"flowLogIdSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "CreateImageRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NoReboot":{ - "shape":"Boolean", - "locationName":"noReboot" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"blockDeviceMapping" - } - } - }, - "CreateImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CreateInstanceExportTaskRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "Description":{ - "shape":"String", - "locationName":"description" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - }, - "ExportToS3Task":{ - "shape":"ExportToS3TaskSpecification", - "locationName":"exportToS3" - } - } - }, - "CreateInstanceExportTaskResult":{ - "type":"structure", - "members":{ - "ExportTask":{ - "shape":"ExportTask", - "locationName":"exportTask" - } - } - }, - "CreateInternetGatewayRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateInternetGatewayResult":{ - "type":"structure", - "members":{ - "InternetGateway":{ - "shape":"InternetGateway", - "locationName":"internetGateway" - } - } - }, - "CreateKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "CreateNatGatewayRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "AllocationId" - ], - "members":{ - "SubnetId":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateNatGatewayResult":{ - "type":"structure", - "members":{ - "NatGateway":{ - "shape":"NatGateway", - "locationName":"natGateway" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "CreateNetworkAclRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateNetworkAclResult":{ - "type":"structure", - "members":{ - "NetworkAcl":{ - "shape":"NetworkAcl", - "locationName":"networkAcl" - } - } - }, - "CreateNetworkInterfaceRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "NetworkInterface":{ - "shape":"NetworkInterface", - "locationName":"networkInterface" - } - } - }, - "CreatePlacementGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Strategy" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - } - } - }, - "CreateReservedInstancesListingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesId", - "InstanceCount", - "PriceSchedules", - "ClientToken" - ], - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "PriceSchedules":{ - "shape":"PriceScheduleSpecificationList", - "locationName":"priceSchedules" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CreateRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "CreateRouteResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "CreateRouteTableRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateRouteTableResult":{ - "type":"structure", - "members":{ - "RouteTable":{ - "shape":"RouteTable", - "locationName":"routeTable" - } - } - }, - "CreateSecurityGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Description" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "Description":{ - "shape":"String", - "locationName":"GroupDescription" - }, - "VpcId":{"shape":"String"} - } - }, - "CreateSecurityGroupResult":{ - "type":"structure", - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "CreateSnapshotRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Description":{"shape":"String"} - } - }, - "CreateSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "required":["Bucket"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - } - } - }, - "CreateSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "CreateSubnetRequest":{ - "type":"structure", - "required":[ - "VpcId", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "CidrBlock":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateSubnetResult":{ - "type":"structure", - "members":{ - "Subnet":{ - "shape":"Subnet", - "locationName":"subnet" - } - } - }, - "CreateTagsRequest":{ - "type":"structure", - "required":[ - "Resources", - "Tags" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"ResourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"Tag" - } - } - }, - "CreateVolumePermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "CreateVolumePermissionList":{ - "type":"list", - "member":{ - "shape":"CreateVolumePermission", - "locationName":"item" - } - }, - "CreateVolumePermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"CreateVolumePermissionList"}, - "Remove":{"shape":"CreateVolumePermissionList"} - } - }, - "CreateVolumeRequest":{ - "type":"structure", - "required":["AvailabilityZone"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Size":{"shape":"Integer"}, - "SnapshotId":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "VolumeType":{"shape":"VolumeType"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{"shape":"String"} - } - }, - "CreateVpcEndpointRequest":{ - "type":"structure", - "required":[ - "VpcId", - "ServiceName" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "ServiceName":{"shape":"String"}, - "PolicyDocument":{"shape":"String"}, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "ClientToken":{"shape":"String"} - } - }, - "CreateVpcEndpointResult":{ - "type":"structure", - "members":{ - "VpcEndpoint":{ - "shape":"VpcEndpoint", - "locationName":"vpcEndpoint" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeerVpcId":{ - "shape":"String", - "locationName":"peerVpcId" - }, - "PeerOwnerId":{ - "shape":"String", - "locationName":"peerOwnerId" - } - } - }, - "CreateVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "CreateVpcRequest":{ - "type":"structure", - "required":["CidrBlock"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CidrBlock":{"shape":"String"}, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - } - } - }, - "CreateVpcResult":{ - "type":"structure", - "members":{ - "Vpc":{ - "shape":"Vpc", - "locationName":"vpc" - } - } - }, - "CreateVpnConnectionRequest":{ - "type":"structure", - "required":[ - "Type", - "CustomerGatewayId", - "VpnGatewayId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"String"}, - "CustomerGatewayId":{"shape":"String"}, - "VpnGatewayId":{"shape":"String"}, - "Options":{ - "shape":"VpnConnectionOptionsSpecification", - "locationName":"options" - } - } - }, - "CreateVpnConnectionResult":{ - "type":"structure", - "members":{ - "VpnConnection":{ - "shape":"VpnConnection", - "locationName":"vpnConnection" - } - } - }, - "CreateVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "CreateVpnGatewayRequest":{ - "type":"structure", - "required":["Type"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpnGateway":{ - "shape":"VpnGateway", - "locationName":"vpnGateway" - } - } - }, - "CurrencyCodeValues":{ - "type":"string", - "enum":["USD"] - }, - "CustomerGateway":{ - "type":"structure", - "members":{ - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "Type":{ - "shape":"String", - "locationName":"type" - }, - "IpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "BgpAsn":{ - "shape":"String", - "locationName":"bgpAsn" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "CustomerGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"CustomerGatewayId" - } - }, - "CustomerGatewayList":{ - "type":"list", - "member":{ - "shape":"CustomerGateway", - "locationName":"item" - } - }, - "DatafeedSubscriptionState":{ - "type":"string", - "enum":[ - "Active", - "Inactive" - ] - }, - "DateTime":{"type":"timestamp"}, - "DeleteCustomerGatewayRequest":{ - "type":"structure", - "required":["CustomerGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayId":{"shape":"String"} - } - }, - "DeleteDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpOptionsId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"} - } - }, - "DeleteFlowLogsRequest":{ - "type":"structure", - "required":["FlowLogIds"], - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - } - } - }, - "DeleteFlowLogsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteInternetGatewayRequest":{ - "type":"structure", - "required":["InternetGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - } - } - }, - "DeleteKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "DeleteNatGatewayRequest":{ - "type":"structure", - "required":["NatGatewayId"], - "members":{ - "NatGatewayId":{"shape":"String"} - } - }, - "DeleteNatGatewayResult":{ - "type":"structure", - "members":{ - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "DeleteNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Egress" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - } - } - }, - "DeleteNetworkAclRequest":{ - "type":"structure", - "required":["NetworkAclId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "DeleteNetworkInterfaceRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "DeletePlacementGroupRequest":{ - "type":"structure", - "required":["GroupName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "DeleteRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - } - } - }, - "DeleteRouteTableRequest":{ - "type":"structure", - "required":["RouteTableId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "DeleteSecurityGroupRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"} - } - }, - "DeleteSnapshotRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"} - } - }, - "DeleteSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DeleteSubnetRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{"shape":"String"} - } - }, - "DeleteTagsRequest":{ - "type":"structure", - "required":["Resources"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"resourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tag" - } - } - }, - "DeleteVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"} - } - }, - "DeleteVpcEndpointsRequest":{ - "type":"structure", - "required":["VpcEndpointIds"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - } - } - }, - "DeleteVpcEndpointsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "DeleteVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DeleteVpcRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRequest":{ - "type":"structure", - "required":["VpnConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "DeleteVpnGatewayRequest":{ - "type":"structure", - "required":["VpnGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"} - } - }, - "DeregisterImageRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"} - } - }, - "DescribeAccountAttributesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttributeNames":{ - "shape":"AccountAttributeNameStringList", - "locationName":"attributeName" - } - } - }, - "DescribeAccountAttributesResult":{ - "type":"structure", - "members":{ - "AccountAttributes":{ - "shape":"AccountAttributeList", - "locationName":"accountAttributeSet" - } - } - }, - "DescribeAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"PublicIpStringList", - "locationName":"PublicIp" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AllocationIds":{ - "shape":"AllocationIdList", - "locationName":"AllocationId" - } - } - }, - "DescribeAddressesResult":{ - "type":"structure", - "members":{ - "Addresses":{ - "shape":"AddressList", - "locationName":"addressesSet" - } - } - }, - "DescribeAvailabilityZonesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ZoneNames":{ - "shape":"ZoneNameStringList", - "locationName":"ZoneName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeAvailabilityZonesResult":{ - "type":"structure", - "members":{ - "AvailabilityZones":{ - "shape":"AvailabilityZoneList", - "locationName":"availabilityZoneInfo" - } - } - }, - "DescribeBundleTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleIds":{ - "shape":"BundleIdStringList", - "locationName":"BundleId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeBundleTasksResult":{ - "type":"structure", - "members":{ - "BundleTasks":{ - "shape":"BundleTaskList", - "locationName":"bundleInstanceTasksSet" - } - } - }, - "DescribeClassicLinkInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeClassicLinkInstancesResult":{ - "type":"structure", - "members":{ - "Instances":{ - "shape":"ClassicLinkInstanceList", - "locationName":"instancesSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeConversionTaskList":{ - "type":"list", - "member":{ - "shape":"ConversionTask", - "locationName":"item" - } - }, - "DescribeConversionTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "ConversionTaskIds":{ - "shape":"ConversionIdStringList", - "locationName":"conversionTaskId" - } - } - }, - "DescribeConversionTasksResult":{ - "type":"structure", - "members":{ - "ConversionTasks":{ - "shape":"DescribeConversionTaskList", - "locationName":"conversionTasks" - } - } - }, - "DescribeCustomerGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayIds":{ - "shape":"CustomerGatewayIdStringList", - "locationName":"CustomerGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeCustomerGatewaysResult":{ - "type":"structure", - "members":{ - "CustomerGateways":{ - "shape":"CustomerGatewayList", - "locationName":"customerGatewaySet" - } - } - }, - "DescribeDhcpOptionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsIds":{ - "shape":"DhcpOptionsIdStringList", - "locationName":"DhcpOptionsId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptionsList", - "locationName":"dhcpOptionsSet" - } - } - }, - "DescribeExportTasksRequest":{ - "type":"structure", - "members":{ - "ExportTaskIds":{ - "shape":"ExportTaskIdStringList", - "locationName":"exportTaskId" - } - } - }, - "DescribeExportTasksResult":{ - "type":"structure", - "members":{ - "ExportTasks":{ - "shape":"ExportTaskList", - "locationName":"exportTaskSet" - } - } - }, - "DescribeFlowLogsRequest":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - }, - "Filter":{"shape":"FilterList"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogs":{ - "shape":"FlowLogSet", - "locationName":"flowLogSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostsRequest":{ - "type":"structure", - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "Filter":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeHostsResult":{ - "type":"structure", - "members":{ - "Hosts":{ - "shape":"HostList", - "locationName":"hostSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeIdFormatRequest":{ - "type":"structure", - "members":{ - "Resource":{"shape":"String"} - } - }, - "DescribeIdFormatResult":{ - "type":"structure", - "members":{ - "Statuses":{ - "shape":"IdFormatList", - "locationName":"statusSet" - } - } - }, - "DescribeImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ImageAttributeName"} - } - }, - "DescribeImagesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageIds":{ - "shape":"ImageIdStringList", - "locationName":"ImageId" - }, - "Owners":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "ExecutableUsers":{ - "shape":"ExecutableByStringList", - "locationName":"ExecutableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeImagesResult":{ - "type":"structure", - "members":{ - "Images":{ - "shape":"ImageList", - "locationName":"imagesSet" - } - } - }, - "DescribeImportImageTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportImageTasksResult":{ - "type":"structure", - "members":{ - "ImportImageTasks":{ - "shape":"ImportImageTaskList", - "locationName":"importImageTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeImportSnapshotTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportSnapshotTasksResult":{ - "type":"structure", - "members":{ - "ImportSnapshotTasks":{ - "shape":"ImportSnapshotTaskList", - "locationName":"importSnapshotTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "DescribeInstanceStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "IncludeAllInstances":{ - "shape":"Boolean", - "locationName":"includeAllInstances" - } - } - }, - "DescribeInstanceStatusResult":{ - "type":"structure", - "members":{ - "InstanceStatuses":{ - "shape":"InstanceStatusList", - "locationName":"instanceStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeInstancesResult":{ - "type":"structure", - "members":{ - "Reservations":{ - "shape":"ReservationList", - "locationName":"reservationSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInternetGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayIds":{ - "shape":"ValueStringList", - "locationName":"internetGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeInternetGatewaysResult":{ - "type":"structure", - "members":{ - "InternetGateways":{ - "shape":"InternetGatewayList", - "locationName":"internetGatewaySet" - } - } - }, - "DescribeKeyPairsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyNames":{ - "shape":"KeyNameStringList", - "locationName":"KeyName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeKeyPairsResult":{ - "type":"structure", - "members":{ - "KeyPairs":{ - "shape":"KeyPairList", - "locationName":"keySet" - } - } - }, - "DescribeMovingAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"ValueStringList", - "locationName":"publicIp" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeMovingAddressesResult":{ - "type":"structure", - "members":{ - "MovingAddressStatuses":{ - "shape":"MovingAddressStatusSet", - "locationName":"movingAddressStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNatGatewaysRequest":{ - "type":"structure", - "members":{ - "NatGatewayIds":{ - "shape":"ValueStringList", - "locationName":"NatGatewayId" - }, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeNatGatewaysResult":{ - "type":"structure", - "members":{ - "NatGateways":{ - "shape":"NatGatewayList", - "locationName":"natGatewaySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNetworkAclsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclIds":{ - "shape":"ValueStringList", - "locationName":"NetworkAclId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeNetworkAclsResult":{ - "type":"structure", - "members":{ - "NetworkAcls":{ - "shape":"NetworkAclList", - "locationName":"networkAclSet" - } - } - }, - "DescribeNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Attribute":{ - "shape":"NetworkInterfaceAttribute", - "locationName":"attribute" - } - } - }, - "DescribeNetworkInterfaceAttributeResult":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - } - } - }, - "DescribeNetworkInterfacesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceIds":{ - "shape":"NetworkInterfaceIdList", - "locationName":"NetworkInterfaceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeNetworkInterfacesResult":{ - "type":"structure", - "members":{ - "NetworkInterfaces":{ - "shape":"NetworkInterfaceList", - "locationName":"networkInterfaceSet" - } - } - }, - "DescribePlacementGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"PlacementGroupStringList", - "locationName":"groupName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribePlacementGroupsResult":{ - "type":"structure", - "members":{ - "PlacementGroups":{ - "shape":"PlacementGroupList", - "locationName":"placementGroupSet" - } - } - }, - "DescribePrefixListsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "PrefixListIds":{ - "shape":"ValueStringList", - "locationName":"PrefixListId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribePrefixListsResult":{ - "type":"structure", - "members":{ - "PrefixLists":{ - "shape":"PrefixListSet", - "locationName":"prefixListSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeRegionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RegionNames":{ - "shape":"RegionNameStringList", - "locationName":"RegionName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRegionsResult":{ - "type":"structure", - "members":{ - "Regions":{ - "shape":"RegionList", - "locationName":"regionInfo" - } - } - }, - "DescribeReservedInstancesListingsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filters" - } - } - }, - "DescribeReservedInstancesListingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "DescribeReservedInstancesModificationsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationIds":{ - "shape":"ReservedInstancesModificationIdStringList", - "locationName":"ReservedInstancesModificationId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeReservedInstancesModificationsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModifications":{ - "shape":"ReservedInstancesModificationList", - "locationName":"reservedInstancesModificationsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesOfferingsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingIds":{ - "shape":"ReservedInstancesOfferingIdStringList", - "locationName":"ReservedInstancesOfferingId" - }, - "InstanceType":{"shape":"InstanceType"}, - "AvailabilityZone":{"shape":"String"}, - "ProductDescription":{"shape":"RIProductDescription"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "IncludeMarketplace":{"shape":"Boolean"}, - "MinDuration":{"shape":"Long"}, - "MaxDuration":{"shape":"Long"}, - "MaxInstanceCount":{"shape":"Integer"} - } - }, - "DescribeReservedInstancesOfferingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferings":{ - "shape":"ReservedInstancesOfferingList", - "locationName":"reservedInstancesOfferingsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - } - } - }, - "DescribeReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstances":{ - "shape":"ReservedInstancesList", - "locationName":"reservedInstancesSet" - } - } - }, - "DescribeRouteTablesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRouteTablesResult":{ - "type":"structure", - "members":{ - "RouteTables":{ - "shape":"RouteTableList", - "locationName":"routeTableSet" - } - } - }, - "DescribeScheduledInstanceAvailabilityRequest":{ - "type":"structure", - "required":[ - "Recurrence", - "FirstSlotStartTimeRange" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "Recurrence":{"shape":"ScheduledInstanceRecurrenceRequest"}, - "FirstSlotStartTimeRange":{"shape":"SlotDateTimeRangeRequest"}, - "MinSlotDurationInHours":{"shape":"Integer"}, - "MaxSlotDurationInHours":{"shape":"Integer"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeScheduledInstanceAvailabilityResult":{ - "type":"structure", - "members":{ - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "ScheduledInstanceAvailabilitySet":{ - "shape":"ScheduledInstanceAvailabilitySet", - "locationName":"scheduledInstanceAvailabilitySet" - } - } - }, - "DescribeScheduledInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ScheduledInstanceIds":{ - "shape":"ScheduledInstanceIdRequestSet", - "locationName":"ScheduledInstanceId" - }, - "SlotStartTimeRange":{"shape":"SlotStartTimeRangeRequest"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeScheduledInstancesResult":{ - "type":"structure", - "members":{ - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "ScheduledInstanceSet":{ - "shape":"ScheduledInstanceSet", - "locationName":"scheduledInstanceSet" - } - } - }, - "DescribeSecurityGroupReferencesRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "GroupId":{"shape":"GroupIds"} - } - }, - "DescribeSecurityGroupReferencesResult":{ - "type":"structure", - "members":{ - "SecurityGroupReferenceSet":{ - "shape":"SecurityGroupReferences", - "locationName":"securityGroupReferenceSet" - } - } - }, - "DescribeSecurityGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSecurityGroupsResult":{ - "type":"structure", - "members":{ - "SecurityGroups":{ - "shape":"SecurityGroupList", - "locationName":"securityGroupInfo" - } - } - }, - "DescribeSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "DescribeSnapshotAttributeResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "CreateVolumePermissions":{ - "shape":"CreateVolumePermissionList", - "locationName":"createVolumePermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeSnapshotsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotIds":{ - "shape":"SnapshotIdStringList", - "locationName":"SnapshotId" - }, - "OwnerIds":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "RestorableByUserIds":{ - "shape":"RestorableByStringList", - "locationName":"RestorableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeSnapshotsResult":{ - "type":"structure", - "members":{ - "Snapshots":{ - "shape":"SnapshotList", - "locationName":"snapshotSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DescribeSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "DescribeSpotFleetInstancesRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetInstancesResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "ActiveInstances" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "ActiveInstances":{ - "shape":"ActiveInstanceSet", - "locationName":"activeInstanceSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestHistoryRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestHistoryResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime", - "LastEvaluatedTime", - "HistoryRecords" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "LastEvaluatedTime":{ - "shape":"DateTime", - "locationName":"lastEvaluatedTime" - }, - "HistoryRecords":{ - "shape":"HistoryRecords", - "locationName":"historyRecordSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestsResponse":{ - "type":"structure", - "required":["SpotFleetRequestConfigs"], - "members":{ - "SpotFleetRequestConfigs":{ - "shape":"SpotFleetRequestConfigSet", - "locationName":"spotFleetRequestConfigSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotInstanceRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "DescribeSpotPriceHistoryRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "InstanceTypes":{ - "shape":"InstanceTypeList", - "locationName":"InstanceType" - }, - "ProductDescriptions":{ - "shape":"ProductDescriptionList", - "locationName":"ProductDescription" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotPriceHistoryResult":{ - "type":"structure", - "members":{ - "SpotPriceHistory":{ - "shape":"SpotPriceHistoryList", - "locationName":"spotPriceHistorySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeStaleSecurityGroupsRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "MaxResults":{"shape":"MaxResults"}, - "NextToken":{"shape":"NextToken"} - } - }, - "DescribeStaleSecurityGroupsResult":{ - "type":"structure", - "members":{ - "StaleSecurityGroupSet":{ - "shape":"StaleSecurityGroupSet", - "locationName":"staleSecurityGroupSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSubnetsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetIds":{ - "shape":"SubnetIdStringList", - "locationName":"SubnetId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSubnetsResult":{ - "type":"structure", - "members":{ - "Subnets":{ - "shape":"SubnetList", - "locationName":"subnetSet" - } - } - }, - "DescribeTagsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeTagsResult":{ - "type":"structure", - "members":{ - "Tags":{ - "shape":"TagDescriptionList", - "locationName":"tagSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Attribute":{"shape":"VolumeAttributeName"} - } - }, - "DescribeVolumeAttributeResult":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AutoEnableIO":{ - "shape":"AttributeBooleanValue", - "locationName":"autoEnableIO" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeVolumeStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeVolumeStatusResult":{ - "type":"structure", - "members":{ - "VolumeStatuses":{ - "shape":"VolumeStatusList", - "locationName":"volumeStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeVolumesResult":{ - "type":"structure", - "members":{ - "Volumes":{ - "shape":"VolumeList", - "locationName":"volumeSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcAttributeRequest":{ - "type":"structure", - "required":[ - "VpcId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "Attribute":{"shape":"VpcAttributeName"} - } - }, - "DescribeVpcAttributeResult":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsSupport" - }, - "EnableDnsHostnames":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsHostnames" - } - } - }, - "DescribeVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcIds":{"shape":"VpcClassicLinkIdList"}, - "MaxResults":{ - "shape":"MaxResults", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"NextToken", - "locationName":"nextToken" - } - } - }, - "DescribeVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"ClassicLinkDnsSupportList", - "locationName":"vpcs" - }, - "NextToken":{ - "shape":"NextToken", - "locationName":"nextToken" - } - } - }, - "DescribeVpcClassicLinkRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcClassicLinkIdList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcClassicLinkList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpcEndpointServicesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointServicesResult":{ - "type":"structure", - "members":{ - "ServiceNames":{ - "shape":"ValueStringList", - "locationName":"serviceNameSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcEndpointsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointsResult":{ - "type":"structure", - "members":{ - "VpcEndpoints":{ - "shape":"VpcEndpointSet", - "locationName":"vpcEndpointSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcPeeringConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionIds":{ - "shape":"ValueStringList", - "locationName":"VpcPeeringConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcPeeringConnectionsResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnections":{ - "shape":"VpcPeeringConnectionList", - "locationName":"vpcPeeringConnectionSet" - } - } - }, - "DescribeVpcsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcIdStringList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcsResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpnConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionIds":{ - "shape":"VpnConnectionIdStringList", - "locationName":"VpnConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnConnectionsResult":{ - "type":"structure", - "members":{ - "VpnConnections":{ - "shape":"VpnConnectionList", - "locationName":"vpnConnectionSet" - } - } - }, - "DescribeVpnGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayIds":{ - "shape":"VpnGatewayIdStringList", - "locationName":"VpnGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnGatewaysResult":{ - "type":"structure", - "members":{ - "VpnGateways":{ - "shape":"VpnGatewayList", - "locationName":"vpnGatewaySet" - } - } - }, - "DetachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DetachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachNetworkInterfaceRequest":{ - "type":"structure", - "required":["AttachmentId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "DetachVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"}, - "Force":{"shape":"Boolean"} - } - }, - "DetachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "DeviceType":{ - "type":"string", - "enum":[ - "ebs", - "instance-store" - ] - }, - "DhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"DhcpConfigurationValueList", - "locationName":"valueSet" - } - } - }, - "DhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"DhcpConfiguration", - "locationName":"item" - } - }, - "DhcpConfigurationValueList":{ - "type":"list", - "member":{ - "shape":"AttributeValue", - "locationName":"item" - } - }, - "DhcpOptions":{ - "type":"structure", - "members":{ - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "DhcpConfigurations":{ - "shape":"DhcpConfigurationList", - "locationName":"dhcpConfigurationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "DhcpOptionsIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DhcpOptionsId" - } - }, - "DhcpOptionsList":{ - "type":"list", - "member":{ - "shape":"DhcpOptions", - "locationName":"item" - } - }, - "DisableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DisableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisassociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AssociationId":{"shape":"String"} - } - }, - "DisassociateRouteTableRequest":{ - "type":"structure", - "required":["AssociationId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "DiskImage":{ - "type":"structure", - "members":{ - "Image":{"shape":"DiskImageDetail"}, - "Description":{"shape":"String"}, - "Volume":{"shape":"VolumeDetail"} - } - }, - "DiskImageDescription":{ - "type":"structure", - "required":[ - "Format", - "Size", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - }, - "Checksum":{ - "shape":"String", - "locationName":"checksum" - } - } - }, - "DiskImageDetail":{ - "type":"structure", - "required":[ - "Format", - "Bytes", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Bytes":{ - "shape":"Long", - "locationName":"bytes" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - } - } - }, - "DiskImageFormat":{ - "type":"string", - "enum":[ - "VMDK", - "RAW", - "VHD" - ] - }, - "DiskImageList":{ - "type":"list", - "member":{"shape":"DiskImage"} - }, - "DiskImageVolumeDescription":{ - "type":"structure", - "required":["Id"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "DomainType":{ - "type":"string", - "enum":[ - "vpc", - "standard" - ] - }, - "Double":{"type":"double"}, - "EbsBlockDevice":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - } - } - }, - "EbsInstanceBlockDevice":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EbsInstanceBlockDeviceSpecification":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EnableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "EnableVolumeIORequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - } - } - }, - "EnableVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcId":{"shape":"String"} - } - }, - "EnableVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EnableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "EnableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EventCode":{ - "type":"string", - "enum":[ - "instance-reboot", - "system-reboot", - "system-maintenance", - "instance-retirement", - "instance-stop" - ] - }, - "EventInformation":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "EventSubType":{ - "shape":"String", - "locationName":"eventSubType" - }, - "EventDescription":{ - "shape":"String", - "locationName":"eventDescription" - } - } - }, - "EventType":{ - "type":"string", - "enum":[ - "instanceChange", - "fleetRequestChange", - "error" - ] - }, - "ExcessCapacityTerminationPolicy":{ - "type":"string", - "enum":[ - "noTermination", - "default" - ] - }, - "ExecutableByStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExecutableBy" - } - }, - "ExportEnvironment":{ - "type":"string", - "enum":[ - "citrix", - "vmware", - "microsoft" - ] - }, - "ExportTask":{ - "type":"structure", - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "State":{ - "shape":"ExportTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceExportDetails":{ - "shape":"InstanceExportDetails", - "locationName":"instanceExport" - }, - "ExportToS3Task":{ - "shape":"ExportToS3Task", - "locationName":"exportToS3" - } - } - }, - "ExportTaskIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExportTaskId" - } - }, - "ExportTaskList":{ - "type":"list", - "member":{ - "shape":"ExportTask", - "locationName":"item" - } - }, - "ExportTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "ExportToS3Task":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "ExportToS3TaskSpecification":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Prefix":{ - "shape":"String", - "locationName":"s3Prefix" - } - } - }, - "Filter":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "FilterList":{ - "type":"list", - "member":{ - "shape":"Filter", - "locationName":"Filter" - } - }, - "FleetType":{ - "type":"string", - "enum":[ - "request", - "maintain" - ] - }, - "Float":{"type":"float"}, - "FlowLog":{ - "type":"structure", - "members":{ - "CreationTime":{ - "shape":"DateTime", - "locationName":"creationTime" - }, - "FlowLogId":{ - "shape":"String", - "locationName":"flowLogId" - }, - "FlowLogStatus":{ - "shape":"String", - "locationName":"flowLogStatus" - }, - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "TrafficType":{ - "shape":"TrafficType", - "locationName":"trafficType" - }, - "LogGroupName":{ - "shape":"String", - "locationName":"logGroupName" - }, - "DeliverLogsStatus":{ - "shape":"String", - "locationName":"deliverLogsStatus" - }, - "DeliverLogsErrorMessage":{ - "shape":"String", - "locationName":"deliverLogsErrorMessage" - }, - "DeliverLogsPermissionArn":{ - "shape":"String", - "locationName":"deliverLogsPermissionArn" - } - } - }, - "FlowLogSet":{ - "type":"list", - "member":{ - "shape":"FlowLog", - "locationName":"item" - } - }, - "FlowLogsResourceType":{ - "type":"string", - "enum":[ - "VPC", - "Subnet", - "NetworkInterface" - ] - }, - "GatewayType":{ - "type":"string", - "enum":["ipsec.1"] - }, - "GetConsoleOutputRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetConsoleOutputResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "Output":{ - "shape":"String", - "locationName":"output" - } - } - }, - "GetConsoleScreenshotRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "InstanceId":{"shape":"String"}, - "WakeUp":{"shape":"Boolean"} - } - }, - "GetConsoleScreenshotResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageData":{ - "shape":"String", - "locationName":"imageData" - } - } - }, - "GetPasswordDataRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetPasswordDataResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "PasswordData":{ - "shape":"String", - "locationName":"passwordData" - } - } - }, - "GroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"groupId" - } - }, - "GroupIdentifier":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "GroupIdentifierList":{ - "type":"list", - "member":{ - "shape":"GroupIdentifier", - "locationName":"item" - } - }, - "GroupIds":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "GroupNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"GroupName" - } - }, - "HistoryRecord":{ - "type":"structure", - "required":[ - "Timestamp", - "EventType", - "EventInformation" - ], - "members":{ - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "EventInformation":{ - "shape":"EventInformation", - "locationName":"eventInformation" - } - } - }, - "HistoryRecords":{ - "type":"list", - "member":{ - "shape":"HistoryRecord", - "locationName":"item" - } - }, - "Host":{ - "type":"structure", - "members":{ - "HostId":{ - "shape":"String", - "locationName":"hostId" - }, - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - }, - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "HostProperties":{ - "shape":"HostProperties", - "locationName":"hostProperties" - }, - "State":{ - "shape":"AllocationState", - "locationName":"state" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Instances":{ - "shape":"HostInstanceList", - "locationName":"instances" - }, - "AvailableCapacity":{ - "shape":"AvailableCapacity", - "locationName":"availableCapacity" - } - } - }, - "HostInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - } - } - }, - "HostInstanceList":{ - "type":"list", - "member":{ - "shape":"HostInstance", - "locationName":"item" - } - }, - "HostList":{ - "type":"list", - "member":{ - "shape":"Host", - "locationName":"item" - } - }, - "HostProperties":{ - "type":"structure", - "members":{ - "Sockets":{ - "shape":"Integer", - "locationName":"sockets" - }, - "Cores":{ - "shape":"Integer", - "locationName":"cores" - }, - "TotalVCpus":{ - "shape":"Integer", - "locationName":"totalVCpus" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - } - } - }, - "HostTenancy":{ - "type":"string", - "enum":[ - "dedicated", - "host" - ] - }, - "HypervisorType":{ - "type":"string", - "enum":[ - "ovm", - "xen" - ] - }, - "IamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "IamInstanceProfileSpecification":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Name":{ - "shape":"String", - "locationName":"name" - } - } - }, - "IcmpTypeCode":{ - "type":"structure", - "members":{ - "Type":{ - "shape":"Integer", - "locationName":"type" - }, - "Code":{ - "shape":"Integer", - "locationName":"code" - } - } - }, - "IdFormat":{ - "type":"structure", - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "UseLongIds":{ - "shape":"Boolean", - "locationName":"useLongIds" - }, - "Deadline":{ - "shape":"DateTime", - "locationName":"deadline" - } - } - }, - "IdFormatList":{ - "type":"list", - "member":{ - "shape":"IdFormat", - "locationName":"item" - } - }, - "Image":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "ImageLocation":{ - "shape":"String", - "locationName":"imageLocation" - }, - "State":{ - "shape":"ImageState", - "locationName":"imageState" - }, - "OwnerId":{ - "shape":"String", - "locationName":"imageOwnerId" - }, - "CreationDate":{ - "shape":"String", - "locationName":"creationDate" - }, - "Public":{ - "shape":"Boolean", - "locationName":"isPublic" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "ImageType":{ - "shape":"ImageTypeValues", - "locationName":"imageType" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "ImageOwnerAlias":{ - "shape":"String", - "locationName":"imageOwnerAlias" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - } - } - }, - "ImageAttribute":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "LaunchPermissions":{ - "shape":"LaunchPermissionList", - "locationName":"launchPermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - } - } - }, - "ImageAttributeName":{ - "type":"string", - "enum":[ - "description", - "kernel", - "ramdisk", - "launchPermission", - "productCodes", - "blockDeviceMapping", - "sriovNetSupport" - ] - }, - "ImageDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"}, - "DeviceName":{"shape":"String"}, - "SnapshotId":{"shape":"String"} - } - }, - "ImageDiskContainerList":{ - "type":"list", - "member":{ - "shape":"ImageDiskContainer", - "locationName":"item" - } - }, - "ImageIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImageId" - } - }, - "ImageList":{ - "type":"list", - "member":{ - "shape":"Image", - "locationName":"item" - } - }, - "ImageState":{ - "type":"string", - "enum":[ - "pending", - "available", - "invalid", - "deregistered", - "transient", - "failed", - "error" - ] - }, - "ImageTypeValues":{ - "type":"string", - "enum":[ - "machine", - "kernel", - "ramdisk" - ] - }, - "ImportImageRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainers":{ - "shape":"ImageDiskContainerList", - "locationName":"DiskContainer" - }, - "LicenseType":{"shape":"String"}, - "Hypervisor":{"shape":"String"}, - "Architecture":{"shape":"String"}, - "Platform":{"shape":"String"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportImageResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTaskList":{ - "type":"list", - "member":{ - "shape":"ImportImageTask", - "locationName":"item" - } - }, - "ImportInstanceLaunchSpecification":{ - "type":"structure", - "members":{ - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "GroupNames":{ - "shape":"SecurityGroupStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"GroupId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "UserData":{ - "shape":"UserData", - "locationName":"userData" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "Monitoring":{ - "shape":"Boolean", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "ImportInstanceRequest":{ - "type":"structure", - "required":["Platform"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "LaunchSpecification":{ - "shape":"ImportInstanceLaunchSpecification", - "locationName":"launchSpecification" - }, - "DiskImages":{ - "shape":"DiskImageList", - "locationName":"diskImage" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - } - } - }, - "ImportInstanceResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportInstanceTaskDetails":{ - "type":"structure", - "required":["Volumes"], - "members":{ - "Volumes":{ - "shape":"ImportInstanceVolumeDetailSet", - "locationName":"volumes" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailItem":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume", - "Status" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailSet":{ - "type":"list", - "member":{ - "shape":"ImportInstanceVolumeDetailItem", - "locationName":"item" - } - }, - "ImportKeyPairRequest":{ - "type":"structure", - "required":[ - "KeyName", - "PublicKeyMaterial" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "PublicKeyMaterial":{ - "shape":"Blob", - "locationName":"publicKeyMaterial" - } - } - }, - "ImportKeyPairResult":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "ImportSnapshotRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainer":{"shape":"SnapshotDiskContainer"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportSnapshotResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTaskList":{ - "type":"list", - "member":{ - "shape":"ImportSnapshotTask", - "locationName":"item" - } - }, - "ImportTaskIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImportTaskId" - } - }, - "ImportVolumeRequest":{ - "type":"structure", - "required":[ - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDetail", - "locationName":"image" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Volume":{ - "shape":"VolumeDetail", - "locationName":"volume" - } - } - }, - "ImportVolumeResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportVolumeTaskDetails":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - } - } - }, - "Instance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "State":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"dnsName" - }, - "StateTransitionReason":{ - "shape":"String", - "locationName":"reason" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "AmiLaunchIndex":{ - "shape":"Integer", - "locationName":"amiLaunchIndex" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "LaunchTime":{ - "shape":"DateTime", - "locationName":"launchTime" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PublicIpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "InstanceLifecycle":{ - "shape":"InstanceLifecycleType", - "locationName":"instanceLifecycle" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfile", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - } - } - }, - "InstanceAttribute":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"AttributeValue", - "locationName":"userData" - }, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "RootDeviceName":{ - "shape":"AttributeValue", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - } - } - }, - "InstanceAttributeName":{ - "type":"string", - "enum":[ - "instanceType", - "kernel", - "ramdisk", - "userData", - "disableApiTermination", - "instanceInitiatedShutdownBehavior", - "rootDeviceName", - "blockDeviceMapping", - "productCodes", - "sourceDestCheck", - "groupSet", - "ebsOptimized", - "sriovNetSupport" - ] - }, - "InstanceBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDevice", - "locationName":"ebs" - } - } - }, - "InstanceBlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMapping", - "locationName":"item" - } - }, - "InstanceBlockDeviceMappingSpecification":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDeviceSpecification", - "locationName":"ebs" - }, - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "InstanceBlockDeviceMappingSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMappingSpecification", - "locationName":"item" - } - }, - "InstanceCapacity":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "AvailableCapacity":{ - "shape":"Integer", - "locationName":"availableCapacity" - }, - "TotalCapacity":{ - "shape":"Integer", - "locationName":"totalCapacity" - } - } - }, - "InstanceCount":{ - "type":"structure", - "members":{ - "State":{ - "shape":"ListingState", - "locationName":"state" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - } - } - }, - "InstanceCountList":{ - "type":"list", - "member":{ - "shape":"InstanceCount", - "locationName":"item" - } - }, - "InstanceExportDetails":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - } - } - }, - "InstanceIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "InstanceIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"InstanceId" - } - }, - "InstanceLifecycleType":{ - "type":"string", - "enum":[ - "spot", - "scheduled" - ] - }, - "InstanceList":{ - "type":"list", - "member":{ - "shape":"Instance", - "locationName":"item" - } - }, - "InstanceMonitoring":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - } - } - }, - "InstanceMonitoringList":{ - "type":"list", - "member":{ - "shape":"InstanceMonitoring", - "locationName":"item" - } - }, - "InstanceNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"InstanceNetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - }, - "PrivateIpAddresses":{ - "shape":"InstancePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - } - } - }, - "InstanceNetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - } - } - }, - "InstanceNetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "InstanceNetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterface", - "locationName":"item" - } - }, - "InstanceNetworkInterfaceSpecification":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddressesSet", - "queryName":"PrivateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AssociatePublicIpAddress":{ - "shape":"Boolean", - "locationName":"associatePublicIpAddress" - } - } - }, - "InstanceNetworkInterfaceSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterfaceSpecification", - "locationName":"item" - } - }, - "InstancePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "InstancePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"InstancePrivateIpAddress", - "locationName":"item" - } - }, - "InstanceState":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"Integer", - "locationName":"code" - }, - "Name":{ - "shape":"InstanceStateName", - "locationName":"name" - } - } - }, - "InstanceStateChange":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CurrentState":{ - "shape":"InstanceState", - "locationName":"currentState" - }, - "PreviousState":{ - "shape":"InstanceState", - "locationName":"previousState" - } - } - }, - "InstanceStateChangeList":{ - "type":"list", - "member":{ - "shape":"InstanceStateChange", - "locationName":"item" - } - }, - "InstanceStateName":{ - "type":"string", - "enum":[ - "pending", - "running", - "shutting-down", - "terminated", - "stopping", - "stopped" - ] - }, - "InstanceStatus":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Events":{ - "shape":"InstanceStatusEventList", - "locationName":"eventsSet" - }, - "InstanceState":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "SystemStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"systemStatus" - }, - "InstanceStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"instanceStatus" - } - } - }, - "InstanceStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"StatusName", - "locationName":"name" - }, - "Status":{ - "shape":"StatusType", - "locationName":"status" - }, - "ImpairedSince":{ - "shape":"DateTime", - "locationName":"impairedSince" - } - } - }, - "InstanceStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusDetails", - "locationName":"item" - } - }, - "InstanceStatusEvent":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"EventCode", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - } - } - }, - "InstanceStatusEventList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusEvent", - "locationName":"item" - } - }, - "InstanceStatusList":{ - "type":"list", - "member":{ - "shape":"InstanceStatus", - "locationName":"item" - } - }, - "InstanceStatusSummary":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"SummaryStatus", - "locationName":"status" - }, - "Details":{ - "shape":"InstanceStatusDetailsList", - "locationName":"details" - } - } - }, - "InstanceType":{ - "type":"string", - "enum":[ - "t1.micro", - "m1.small", - "m1.medium", - "m1.large", - "m1.xlarge", - "m3.medium", - "m3.large", - "m3.xlarge", - "m3.2xlarge", - "m4.large", - "m4.xlarge", - "m4.2xlarge", - "m4.4xlarge", - "m4.10xlarge", - "t2.nano", - "t2.micro", - "t2.small", - "t2.medium", - "t2.large", - "m2.xlarge", - "m2.2xlarge", - "m2.4xlarge", - "cr1.8xlarge", - "x1.4xlarge", - "x1.8xlarge", - "x1.16xlarge", - "x1.32xlarge", - "i2.xlarge", - "i2.2xlarge", - "i2.4xlarge", - "i2.8xlarge", - "hi1.4xlarge", - "hs1.8xlarge", - "c1.medium", - "c1.xlarge", - "c3.large", - "c3.xlarge", - "c3.2xlarge", - "c3.4xlarge", - "c3.8xlarge", - "c4.large", - "c4.xlarge", - "c4.2xlarge", - "c4.4xlarge", - "c4.8xlarge", - "cc1.4xlarge", - "cc2.8xlarge", - "g2.2xlarge", - "g2.8xlarge", - "cg1.4xlarge", - "r3.large", - "r3.xlarge", - "r3.2xlarge", - "r3.4xlarge", - "r3.8xlarge", - "d2.xlarge", - "d2.2xlarge", - "d2.4xlarge", - "d2.8xlarge" - ] - }, - "InstanceTypeList":{ - "type":"list", - "member":{"shape":"InstanceType"} - }, - "Integer":{"type":"integer"}, - "InternetGateway":{ - "type":"structure", - "members":{ - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "Attachments":{ - "shape":"InternetGatewayAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "InternetGatewayAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "InternetGatewayAttachmentList":{ - "type":"list", - "member":{ - "shape":"InternetGatewayAttachment", - "locationName":"item" - } - }, - "InternetGatewayList":{ - "type":"list", - "member":{ - "shape":"InternetGateway", - "locationName":"item" - } - }, - "IpPermission":{ - "type":"structure", - "members":{ - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairList", - "locationName":"groups" - }, - "IpRanges":{ - "shape":"IpRangeList", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdList", - "locationName":"prefixListIds" - } - } - }, - "IpPermissionList":{ - "type":"list", - "member":{ - "shape":"IpPermission", - "locationName":"item" - } - }, - "IpRange":{ - "type":"structure", - "members":{ - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - } - } - }, - "IpRangeList":{ - "type":"list", - "member":{ - "shape":"IpRange", - "locationName":"item" - } - }, - "IpRanges":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "KeyNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"KeyName" - } - }, - "KeyPair":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - }, - "KeyMaterial":{ - "shape":"String", - "locationName":"keyMaterial" - } - } - }, - "KeyPairInfo":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "KeyPairList":{ - "type":"list", - "member":{ - "shape":"KeyPairInfo", - "locationName":"item" - } - }, - "LaunchPermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "LaunchPermissionList":{ - "type":"list", - "member":{ - "shape":"LaunchPermission", - "locationName":"item" - } - }, - "LaunchPermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"LaunchPermissionList"}, - "Remove":{"shape":"LaunchPermissionList"} - } - }, - "LaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - } - } - }, - "LaunchSpecsList":{ - "type":"list", - "member":{ - "shape":"SpotFleetLaunchSpecification", - "locationName":"item" - }, - "min":1 - }, - "ListingState":{ - "type":"string", - "enum":[ - "available", - "sold", - "cancelled", - "pending" - ] - }, - "ListingStatus":{ - "type":"string", - "enum":[ - "active", - "pending", - "cancelled", - "closed" - ] - }, - "Long":{"type":"long"}, - "MaxResults":{ - "type":"integer", - "max":255, - "min":5 - }, - "ModifyHostsRequest":{ - "type":"structure", - "required":[ - "HostIds", - "AutoPlacement" - ], - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - }, - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - } - } - }, - "ModifyHostsResult":{ - "type":"structure", - "members":{ - "Successful":{ - "shape":"ResponseHostIdList", - "locationName":"successful" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemList", - "locationName":"unsuccessful" - } - } - }, - "ModifyIdFormatRequest":{ - "type":"structure", - "required":[ - "Resource", - "UseLongIds" - ], - "members":{ - "Resource":{"shape":"String"}, - "UseLongIds":{"shape":"Boolean"} - } - }, - "ModifyImageAttributeRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"String"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "UserGroups":{ - "shape":"UserGroupStringList", - "locationName":"UserGroup" - }, - "ProductCodes":{ - "shape":"ProductCodeStringList", - "locationName":"ProductCode" - }, - "Value":{"shape":"String"}, - "LaunchPermission":{"shape":"LaunchPermissionModifications"}, - "Description":{"shape":"AttributeValue"} - } - }, - "ModifyInstanceAttributeRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - }, - "Value":{ - "shape":"String", - "locationName":"value" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingSpecificationList", - "locationName":"blockDeviceMapping" - }, - "SourceDestCheck":{"shape":"AttributeBooleanValue"}, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "Kernel":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "Ramdisk":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"BlobAttributeValue", - "locationName":"userData" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - } - } - }, - "ModifyInstancePlacementRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Tenancy":{ - "shape":"HostTenancy", - "locationName":"tenancy" - }, - "Affinity":{ - "shape":"Affinity", - "locationName":"affinity" - }, - "HostId":{ - "shape":"String", - "locationName":"hostId" - } - } - }, - "ModifyInstancePlacementResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifyNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachmentChanges", - "locationName":"attachment" - } - } - }, - "ModifyReservedInstancesRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesIds", - "TargetConfigurations" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "TargetConfigurations":{ - "shape":"ReservedInstancesConfigurationList", - "locationName":"ReservedInstancesConfigurationSetItemType" - } - } - }, - "ModifyReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - } - } - }, - "ModifySnapshotAttributeRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"UserGroup" - }, - "CreateVolumePermission":{"shape":"CreateVolumePermissionModifications"} - } - }, - "ModifySpotFleetRequestRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ExcessCapacityTerminationPolicy":{ - "shape":"ExcessCapacityTerminationPolicy", - "locationName":"excessCapacityTerminationPolicy" - } - } - }, - "ModifySpotFleetRequestResponse":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifySubnetAttributeRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "MapPublicIpOnLaunch":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "AutoEnableIO":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcAttributeRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{"shape":"AttributeBooleanValue"}, - "EnableDnsHostnames":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcEndpointRequest":{ - "type":"structure", - "required":["VpcEndpointId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointId":{"shape":"String"}, - "ResetPolicy":{"shape":"Boolean"}, - "PolicyDocument":{"shape":"String"}, - "AddRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"AddRouteTableId" - }, - "RemoveRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RemoveRouteTableId" - } - } - }, - "ModifyVpcEndpointResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifyVpcPeeringConnectionOptionsRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcPeeringConnectionId":{"shape":"String"}, - "RequesterPeeringConnectionOptions":{"shape":"PeeringConnectionOptionsRequest"}, - "AccepterPeeringConnectionOptions":{"shape":"PeeringConnectionOptionsRequest"} - } - }, - "ModifyVpcPeeringConnectionOptionsResult":{ - "type":"structure", - "members":{ - "RequesterPeeringConnectionOptions":{ - "shape":"PeeringConnectionOptions", - "locationName":"requesterPeeringConnectionOptions" - }, - "AccepterPeeringConnectionOptions":{ - "shape":"PeeringConnectionOptions", - "locationName":"accepterPeeringConnectionOptions" - } - } - }, - "MonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "MonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "Monitoring":{ - "type":"structure", - "members":{ - "State":{ - "shape":"MonitoringState", - "locationName":"state" - } - } - }, - "MonitoringState":{ - "type":"string", - "enum":[ - "disabled", - "disabling", - "enabled", - "pending" - ] - }, - "MoveAddressToVpcRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "MoveAddressToVpcResult":{ - "type":"structure", - "members":{ - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "Status":{ - "shape":"Status", - "locationName":"status" - } - } - }, - "MoveStatus":{ - "type":"string", - "enum":[ - "movingToVpc", - "restoringToClassic" - ] - }, - "MovingAddressStatus":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "MoveStatus":{ - "shape":"MoveStatus", - "locationName":"moveStatus" - } - } - }, - "MovingAddressStatusSet":{ - "type":"list", - "member":{ - "shape":"MovingAddressStatus", - "locationName":"item" - } - }, - "NatGateway":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "DeleteTime":{ - "shape":"DateTime", - "locationName":"deleteTime" - }, - "NatGatewayAddresses":{ - "shape":"NatGatewayAddressList", - "locationName":"natGatewayAddressSet" - }, - "State":{ - "shape":"NatGatewayState", - "locationName":"state" - }, - "FailureCode":{ - "shape":"String", - "locationName":"failureCode" - }, - "FailureMessage":{ - "shape":"String", - "locationName":"failureMessage" - }, - "ProvisionedBandwidth":{ - "shape":"ProvisionedBandwidth", - "locationName":"provisionedBandwidth" - } - } - }, - "NatGatewayAddress":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "PrivateIp":{ - "shape":"String", - "locationName":"privateIp" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "NatGatewayAddressList":{ - "type":"list", - "member":{ - "shape":"NatGatewayAddress", - "locationName":"item" - } - }, - "NatGatewayList":{ - "type":"list", - "member":{ - "shape":"NatGateway", - "locationName":"item" - } - }, - "NatGatewayState":{ - "type":"string", - "enum":[ - "pending", - "failed", - "available", - "deleting", - "deleted" - ] - }, - "NetworkAcl":{ - "type":"structure", - "members":{ - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"default" - }, - "Entries":{ - "shape":"NetworkAclEntryList", - "locationName":"entrySet" - }, - "Associations":{ - "shape":"NetworkAclAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "NetworkAclAssociation":{ - "type":"structure", - "members":{ - "NetworkAclAssociationId":{ - "shape":"String", - "locationName":"networkAclAssociationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - } - } - }, - "NetworkAclAssociationList":{ - "type":"list", - "member":{ - "shape":"NetworkAclAssociation", - "locationName":"item" - } - }, - "NetworkAclEntry":{ - "type":"structure", - "members":{ - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"icmpTypeCode" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "NetworkAclEntryList":{ - "type":"list", - "member":{ - "shape":"NetworkAclEntry", - "locationName":"item" - } - }, - "NetworkAclList":{ - "type":"list", - "member":{ - "shape":"NetworkAcl", - "locationName":"item" - } - }, - "NetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "RequesterManaged":{ - "shape":"Boolean", - "locationName":"requesterManaged" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - }, - "TagSet":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PrivateIpAddresses":{ - "shape":"NetworkInterfacePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - }, - "InterfaceType":{ - "shape":"NetworkInterfaceType", - "locationName":"interfaceType" - } - } - }, - "NetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "NetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttachmentChanges":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttribute":{ - "type":"string", - "enum":[ - "description", - "groupSet", - "sourceDestCheck", - "attachment" - ] - }, - "NetworkInterfaceIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "NetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"NetworkInterface", - "locationName":"item" - } - }, - "NetworkInterfacePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "NetworkInterfacePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"NetworkInterfacePrivateIpAddress", - "locationName":"item" - } - }, - "NetworkInterfaceStatus":{ - "type":"string", - "enum":[ - "available", - "attaching", - "in-use", - "detaching" - ] - }, - "NetworkInterfaceType":{ - "type":"string", - "enum":[ - "interface", - "natGateway" - ] - }, - "NewDhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "NewDhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"NewDhcpConfiguration", - "locationName":"item" - } - }, - "NextToken":{ - "type":"string", - "max":1024, - "min":1 - }, - "OccurrenceDayRequestSet":{ - "type":"list", - "member":{ - "shape":"Integer", - "locationName":"OccurenceDay" - } - }, - "OccurrenceDaySet":{ - "type":"list", - "member":{ - "shape":"Integer", - "locationName":"item" - } - }, - "OfferingTypeValues":{ - "type":"string", - "enum":[ - "Heavy Utilization", - "Medium Utilization", - "Light Utilization", - "No Upfront", - "Partial Upfront", - "All Upfront" - ] - }, - "OperationType":{ - "type":"string", - "enum":[ - "add", - "remove" - ] - }, - "OwnerStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"Owner" - } - }, - "PeeringConnectionOptions":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalClassicLinkToRemoteVpc" - }, - "AllowEgressFromLocalVpcToRemoteClassicLink":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalVpcToRemoteClassicLink" - } - } - }, - "PeeringConnectionOptionsRequest":{ - "type":"structure", - "required":[ - "AllowEgressFromLocalClassicLinkToRemoteVpc", - "AllowEgressFromLocalVpcToRemoteClassicLink" - ], - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{"shape":"Boolean"}, - "AllowEgressFromLocalVpcToRemoteClassicLink":{"shape":"Boolean"} - } - }, - "PermissionGroup":{ - "type":"string", - "enum":["all"] - }, - "Placement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Tenancy":{ - "shape":"Tenancy", - "locationName":"tenancy" - }, - "HostId":{ - "shape":"String", - "locationName":"hostId" - }, - "Affinity":{ - "shape":"String", - "locationName":"affinity" - } - } - }, - "PlacementGroup":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - }, - "State":{ - "shape":"PlacementGroupState", - "locationName":"state" - } - } - }, - "PlacementGroupList":{ - "type":"list", - "member":{ - "shape":"PlacementGroup", - "locationName":"item" - } - }, - "PlacementGroupState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "PlacementGroupStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PlacementStrategy":{ - "type":"string", - "enum":["cluster"] - }, - "PlatformValues":{ - "type":"string", - "enum":["Windows"] - }, - "PortRange":{ - "type":"structure", - "members":{ - "From":{ - "shape":"Integer", - "locationName":"from" - }, - "To":{ - "shape":"Integer", - "locationName":"to" - } - } - }, - "PrefixList":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - }, - "PrefixListName":{ - "shape":"String", - "locationName":"prefixListName" - }, - "Cidrs":{ - "shape":"ValueStringList", - "locationName":"cidrSet" - } - } - }, - "PrefixListId":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - } - } - }, - "PrefixListIdList":{ - "type":"list", - "member":{ - "shape":"PrefixListId", - "locationName":"item" - } - }, - "PrefixListIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "PrefixListSet":{ - "type":"list", - "member":{ - "shape":"PrefixList", - "locationName":"item" - } - }, - "PriceSchedule":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Active":{ - "shape":"Boolean", - "locationName":"active" - } - } - }, - "PriceScheduleList":{ - "type":"list", - "member":{ - "shape":"PriceSchedule", - "locationName":"item" - } - }, - "PriceScheduleSpecification":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "PriceScheduleSpecificationList":{ - "type":"list", - "member":{ - "shape":"PriceScheduleSpecification", - "locationName":"item" - } - }, - "PricingDetail":{ - "type":"structure", - "members":{ - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "Count":{ - "shape":"Integer", - "locationName":"count" - } - } - }, - "PricingDetailsList":{ - "type":"list", - "member":{ - "shape":"PricingDetail", - "locationName":"item" - } - }, - "PrivateIpAddressConfigSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesPrivateIpAddressConfig", - "locationName":"PrivateIpAddressConfigSet" - } - }, - "PrivateIpAddressSpecification":{ - "type":"structure", - "required":["PrivateIpAddress"], - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - } - } - }, - "PrivateIpAddressSpecificationList":{ - "type":"list", - "member":{ - "shape":"PrivateIpAddressSpecification", - "locationName":"item" - } - }, - "PrivateIpAddressStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PrivateIpAddress" - } - }, - "ProductCode":{ - "type":"structure", - "members":{ - "ProductCodeId":{ - "shape":"String", - "locationName":"productCode" - }, - "ProductCodeType":{ - "shape":"ProductCodeValues", - "locationName":"type" - } - } - }, - "ProductCodeList":{ - "type":"list", - "member":{ - "shape":"ProductCode", - "locationName":"item" - } - }, - "ProductCodeStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ProductCode" - } - }, - "ProductCodeValues":{ - "type":"string", - "enum":[ - "devpay", - "marketplace" - ] - }, - "ProductDescriptionList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PropagatingVgw":{ - "type":"structure", - "members":{ - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - } - } - }, - "PropagatingVgwList":{ - "type":"list", - "member":{ - "shape":"PropagatingVgw", - "locationName":"item" - } - }, - "ProvisionedBandwidth":{ - "type":"structure", - "members":{ - "Provisioned":{ - "shape":"String", - "locationName":"provisioned" - }, - "Requested":{ - "shape":"String", - "locationName":"requested" - }, - "RequestTime":{ - "shape":"DateTime", - "locationName":"requestTime" - }, - "ProvisionTime":{ - "shape":"DateTime", - "locationName":"provisionTime" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "PublicIpStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PublicIp" - } - }, - "PurchaseRequest":{ - "type":"structure", - "required":[ - "PurchaseToken", - "InstanceCount" - ], - "members":{ - "PurchaseToken":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"} - } - }, - "PurchaseRequestSet":{ - "type":"list", - "member":{ - "shape":"PurchaseRequest", - "locationName":"PurchaseRequest" - }, - "min":1 - }, - "PurchaseReservedInstancesOfferingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesOfferingId", - "InstanceCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingId":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"}, - "LimitPrice":{ - "shape":"ReservedInstanceLimitPrice", - "locationName":"limitPrice" - } - } - }, - "PurchaseReservedInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "PurchaseScheduledInstancesRequest":{ - "type":"structure", - "required":["PurchaseRequests"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ClientToken":{ - "shape":"String", - "idempotencyToken":true - }, - "PurchaseRequests":{ - "shape":"PurchaseRequestSet", - "locationName":"PurchaseRequest" - } - } - }, - "PurchaseScheduledInstancesResult":{ - "type":"structure", - "members":{ - "ScheduledInstanceSet":{ - "shape":"PurchasedScheduledInstanceSet", - "locationName":"scheduledInstanceSet" - } - } - }, - "PurchasedScheduledInstanceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstance", - "locationName":"item" - } - }, - "RIProductDescription":{ - "type":"string", - "enum":[ - "Linux/UNIX", - "Linux/UNIX (Amazon VPC)", - "Windows", - "Windows (Amazon VPC)" - ] - }, - "ReasonCodesList":{ - "type":"list", - "member":{ - "shape":"ReportInstanceReasonCodes", - "locationName":"item" - } - }, - "RebootInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"RecurringChargeFrequency", - "locationName":"frequency" - }, - "Amount":{ - "shape":"Double", - "locationName":"amount" - } - } - }, - "RecurringChargeFrequency":{ - "type":"string", - "enum":["Hourly"] - }, - "RecurringChargesList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"item" - } - }, - "Region":{ - "type":"structure", - "members":{ - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Endpoint":{ - "shape":"String", - "locationName":"regionEndpoint" - } - } - }, - "RegionList":{ - "type":"list", - "member":{ - "shape":"Region", - "locationName":"item" - } - }, - "RegionNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"RegionName" - } - }, - "RegisterImageRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageLocation":{"shape":"String"}, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"String", - "locationName":"virtualizationType" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - } - } - }, - "RegisterImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "RejectVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "RejectVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ReleaseAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"} - } - }, - "ReleaseHostsRequest":{ - "type":"structure", - "required":["HostIds"], - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - } - } - }, - "ReleaseHostsResult":{ - "type":"structure", - "members":{ - "Successful":{ - "shape":"ResponseHostIdList", - "locationName":"successful" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemList", - "locationName":"unsuccessful" - } - } - }, - "ReplaceNetworkAclAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "NetworkAclId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "ReplaceNetworkAclAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReplaceNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "ReplaceRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "ReplaceRouteTableAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "ReplaceRouteTableAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReportInstanceReasonCodes":{ - "type":"string", - "enum":[ - "instance-stuck-in-state", - "unresponsive", - "not-accepting-credentials", - "password-not-available", - "performance-network", - "performance-instance-store", - "performance-ebs-volume", - "performance-other", - "other" - ] - }, - "ReportInstanceStatusRequest":{ - "type":"structure", - "required":[ - "Instances", - "Status", - "ReasonCodes" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Instances":{ - "shape":"InstanceIdStringList", - "locationName":"instanceId" - }, - "Status":{ - "shape":"ReportStatusType", - "locationName":"status" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "ReasonCodes":{ - "shape":"ReasonCodesList", - "locationName":"reasonCode" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ReportStatusType":{ - "type":"string", - "enum":[ - "ok", - "impaired" - ] - }, - "RequestHostIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RequestSpotFleetRequest":{ - "type":"structure", - "required":["SpotFleetRequestConfig"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - } - } - }, - "RequestSpotFleetResponse":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - } - } - }, - "RequestSpotInstancesRequest":{ - "type":"structure", - "required":["SpotPrice"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "BlockDurationMinutes":{ - "shape":"Integer", - "locationName":"blockDurationMinutes" - }, - "LaunchSpecification":{"shape":"RequestSpotLaunchSpecification"} - } - }, - "RequestSpotInstancesResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "RequestSpotLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"ValueStringList", - "locationName":"SecurityGroup" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - }, - "SecurityGroupIds":{ - "shape":"ValueStringList", - "locationName":"SecurityGroupId" - } - } - }, - "Reservation":{ - "type":"structure", - "members":{ - "ReservationId":{ - "shape":"String", - "locationName":"reservationId" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Instances":{ - "shape":"InstanceList", - "locationName":"instancesSet" - } - } - }, - "ReservationList":{ - "type":"list", - "member":{ - "shape":"Reservation", - "locationName":"item" - } - }, - "ReservedInstanceLimitPrice":{ - "type":"structure", - "members":{ - "Amount":{ - "shape":"Double", - "locationName":"amount" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "ReservedInstanceState":{ - "type":"string", - "enum":[ - "payment-pending", - "active", - "payment-failed", - "retired" - ] - }, - "ReservedInstances":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Start":{ - "shape":"DateTime", - "locationName":"start" - }, - "End":{ - "shape":"DateTime", - "locationName":"end" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "State":{ - "shape":"ReservedInstanceState", - "locationName":"state" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - } - } - }, - "ReservedInstancesConfiguration":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - } - } - }, - "ReservedInstancesConfigurationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"item" - } - }, - "ReservedInstancesId":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "ReservedInstancesIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesId" - } - }, - "ReservedInstancesList":{ - "type":"list", - "member":{ - "shape":"ReservedInstances", - "locationName":"item" - } - }, - "ReservedInstancesListing":{ - "type":"structure", - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "Status":{ - "shape":"ListingStatus", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceCounts":{ - "shape":"InstanceCountList", - "locationName":"instanceCounts" - }, - "PriceSchedules":{ - "shape":"PriceScheduleList", - "locationName":"priceSchedules" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesListingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesListing", - "locationName":"item" - } - }, - "ReservedInstancesModification":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - }, - "ReservedInstancesIds":{ - "shape":"ReservedIntancesIds", - "locationName":"reservedInstancesSet" - }, - "ModificationResults":{ - "shape":"ReservedInstancesModificationResultList", - "locationName":"modificationResultSet" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "EffectiveDate":{ - "shape":"DateTime", - "locationName":"effectiveDate" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesModificationIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesModificationId" - } - }, - "ReservedInstancesModificationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModification", - "locationName":"item" - } - }, - "ReservedInstancesModificationResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "TargetConfiguration":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"targetConfiguration" - } - } - }, - "ReservedInstancesModificationResultList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModificationResult", - "locationName":"item" - } - }, - "ReservedInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferingId":{ - "shape":"String", - "locationName":"reservedInstancesOfferingId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - }, - "Marketplace":{ - "shape":"Boolean", - "locationName":"marketplace" - }, - "PricingDetails":{ - "shape":"PricingDetailsList", - "locationName":"pricingDetailsSet" - } - } - }, - "ReservedInstancesOfferingIdStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ReservedInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesOffering", - "locationName":"item" - } - }, - "ReservedIntancesIds":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesId", - "locationName":"item" - } - }, - "ResetImageAttributeName":{ - "type":"string", - "enum":["launchPermission"] - }, - "ResetImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ResetImageAttributeName"} - } - }, - "ResetInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "ResetNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SourceDestCheck":{ - "shape":"String", - "locationName":"sourceDestCheck" - } - } - }, - "ResetSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "ResourceIdList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ResourceType":{ - "type":"string", - "enum":[ - "customer-gateway", - "dhcp-options", - "image", - "instance", - "internet-gateway", - "network-acl", - "network-interface", - "reserved-instances", - "route-table", - "snapshot", - "spot-instances-request", - "subnet", - "security-group", - "volume", - "vpc", - "vpn-connection", - "vpn-gateway" - ] - }, - "ResponseHostIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RestorableByStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "RestoreAddressToClassicRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RestoreAddressToClassicResult":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"Status", - "locationName":"status" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RevokeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "RevokeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "Route":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "DestinationPrefixListId":{ - "shape":"String", - "locationName":"destinationPrefixListId" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - }, - "State":{ - "shape":"RouteState", - "locationName":"state" - }, - "Origin":{ - "shape":"RouteOrigin", - "locationName":"origin" - } - } - }, - "RouteList":{ - "type":"list", - "member":{ - "shape":"Route", - "locationName":"item" - } - }, - "RouteOrigin":{ - "type":"string", - "enum":[ - "CreateRouteTable", - "CreateRoute", - "EnableVgwRoutePropagation" - ] - }, - "RouteState":{ - "type":"string", - "enum":[ - "active", - "blackhole" - ] - }, - "RouteTable":{ - "type":"structure", - "members":{ - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Routes":{ - "shape":"RouteList", - "locationName":"routeSet" - }, - "Associations":{ - "shape":"RouteTableAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PropagatingVgws":{ - "shape":"PropagatingVgwList", - "locationName":"propagatingVgwSet" - } - } - }, - "RouteTableAssociation":{ - "type":"structure", - "members":{ - "RouteTableAssociationId":{ - "shape":"String", - "locationName":"routeTableAssociationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Main":{ - "shape":"Boolean", - "locationName":"main" - } - } - }, - "RouteTableAssociationList":{ - "type":"list", - "member":{ - "shape":"RouteTableAssociation", - "locationName":"item" - } - }, - "RouteTableList":{ - "type":"list", - "member":{ - "shape":"RouteTable", - "locationName":"item" - } - }, - "RuleAction":{ - "type":"string", - "enum":[ - "allow", - "deny" - ] - }, - "RunInstancesMonitoringEnabled":{ - "type":"structure", - "required":["Enabled"], - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "RunInstancesRequest":{ - "type":"structure", - "required":[ - "ImageId", - "MinCount", - "MaxCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "MinCount":{"shape":"Integer"}, - "MaxCount":{"shape":"Integer"}, - "KeyName":{"shape":"String"}, - "SecurityGroups":{ - "shape":"SecurityGroupStringList", - "locationName":"SecurityGroup" - }, - "SecurityGroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "InstanceType":{"shape":"InstanceType"}, - "Placement":{"shape":"Placement"}, - "KernelId":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"RunInstancesMonitoringEnabled"}, - "SubnetId":{"shape":"String"}, - "DisableApiTermination":{ - "shape":"Boolean", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - } - } - }, - "RunScheduledInstancesRequest":{ - "type":"structure", - "required":[ - "ScheduledInstanceId", - "LaunchSpecification" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ClientToken":{ - "shape":"String", - "idempotencyToken":true - }, - "InstanceCount":{"shape":"Integer"}, - "ScheduledInstanceId":{"shape":"String"}, - "LaunchSpecification":{"shape":"ScheduledInstancesLaunchSpecification"} - } - }, - "RunScheduledInstancesResult":{ - "type":"structure", - "members":{ - "InstanceIdSet":{ - "shape":"InstanceIdSet", - "locationName":"instanceIdSet" - } - } - }, - "S3Storage":{ - "type":"structure", - "members":{ - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "AWSAccessKeyId":{"shape":"String"}, - "UploadPolicy":{ - "shape":"Blob", - "locationName":"uploadPolicy" - }, - "UploadPolicySignature":{ - "shape":"String", - "locationName":"uploadPolicySignature" - } - } - }, - "ScheduledInstance":{ - "type":"structure", - "members":{ - "ScheduledInstanceId":{ - "shape":"String", - "locationName":"scheduledInstanceId" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "NetworkPlatform":{ - "shape":"String", - "locationName":"networkPlatform" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "SlotDurationInHours":{ - "shape":"Integer", - "locationName":"slotDurationInHours" - }, - "Recurrence":{ - "shape":"ScheduledInstanceRecurrence", - "locationName":"recurrence" - }, - "PreviousSlotEndTime":{ - "shape":"DateTime", - "locationName":"previousSlotEndTime" - }, - "NextSlotStartTime":{ - "shape":"DateTime", - "locationName":"nextSlotStartTime" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "TotalScheduledInstanceHours":{ - "shape":"Integer", - "locationName":"totalScheduledInstanceHours" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "TermStartDate":{ - "shape":"DateTime", - "locationName":"termStartDate" - }, - "TermEndDate":{ - "shape":"DateTime", - "locationName":"termEndDate" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - } - } - }, - "ScheduledInstanceAvailability":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "NetworkPlatform":{ - "shape":"String", - "locationName":"networkPlatform" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "PurchaseToken":{ - "shape":"String", - "locationName":"purchaseToken" - }, - "SlotDurationInHours":{ - "shape":"Integer", - "locationName":"slotDurationInHours" - }, - "Recurrence":{ - "shape":"ScheduledInstanceRecurrence", - "locationName":"recurrence" - }, - "FirstSlotStartTime":{ - "shape":"DateTime", - "locationName":"firstSlotStartTime" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "TotalScheduledInstanceHours":{ - "shape":"Integer", - "locationName":"totalScheduledInstanceHours" - }, - "AvailableInstanceCount":{ - "shape":"Integer", - "locationName":"availableInstanceCount" - }, - "MinTermDurationInDays":{ - "shape":"Integer", - "locationName":"minTermDurationInDays" - }, - "MaxTermDurationInDays":{ - "shape":"Integer", - "locationName":"maxTermDurationInDays" - } - } - }, - "ScheduledInstanceAvailabilitySet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstanceAvailability", - "locationName":"item" - } - }, - "ScheduledInstanceIdRequestSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ScheduledInstanceId" - } - }, - "ScheduledInstanceRecurrence":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"String", - "locationName":"frequency" - }, - "Interval":{ - "shape":"Integer", - "locationName":"interval" - }, - "OccurrenceDaySet":{ - "shape":"OccurrenceDaySet", - "locationName":"occurrenceDaySet" - }, - "OccurrenceRelativeToEnd":{ - "shape":"Boolean", - "locationName":"occurrenceRelativeToEnd" - }, - "OccurrenceUnit":{ - "shape":"String", - "locationName":"occurrenceUnit" - } - } - }, - "ScheduledInstanceRecurrenceRequest":{ - "type":"structure", - "members":{ - "Frequency":{"shape":"String"}, - "Interval":{"shape":"Integer"}, - "OccurrenceDays":{ - "shape":"OccurrenceDayRequestSet", - "locationName":"OccurrenceDay" - }, - "OccurrenceRelativeToEnd":{"shape":"Boolean"}, - "OccurrenceUnit":{"shape":"String"} - } - }, - "ScheduledInstanceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstance", - "locationName":"item" - } - }, - "ScheduledInstancesBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{"shape":"String"}, - "NoDevice":{"shape":"String"}, - "VirtualName":{"shape":"String"}, - "Ebs":{"shape":"ScheduledInstancesEbs"} - } - }, - "ScheduledInstancesBlockDeviceMappingSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesBlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "ScheduledInstancesEbs":{ - "type":"structure", - "members":{ - "SnapshotId":{"shape":"String"}, - "VolumeSize":{"shape":"Integer"}, - "DeleteOnTermination":{"shape":"Boolean"}, - "VolumeType":{"shape":"String"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{"shape":"Boolean"} - } - }, - "ScheduledInstancesIamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{"shape":"String"}, - "Name":{"shape":"String"} - } - }, - "ScheduledInstancesLaunchSpecification":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "ImageId":{"shape":"String"}, - "KeyName":{"shape":"String"}, - "SecurityGroupIds":{ - "shape":"ScheduledInstancesSecurityGroupIdSet", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "Placement":{"shape":"ScheduledInstancesPlacement"}, - "KernelId":{"shape":"String"}, - "InstanceType":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"ScheduledInstancesBlockDeviceMappingSet", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"ScheduledInstancesMonitoring"}, - "SubnetId":{"shape":"String"}, - "NetworkInterfaces":{ - "shape":"ScheduledInstancesNetworkInterfaceSet", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{"shape":"ScheduledInstancesIamInstanceProfile"}, - "EbsOptimized":{"shape":"Boolean"} - } - }, - "ScheduledInstancesMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{"shape":"Boolean"} - } - }, - "ScheduledInstancesNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{"shape":"String"}, - "DeviceIndex":{"shape":"Integer"}, - "SubnetId":{"shape":"String"}, - "Description":{"shape":"String"}, - "PrivateIpAddress":{"shape":"String"}, - "PrivateIpAddressConfigs":{ - "shape":"PrivateIpAddressConfigSet", - "locationName":"PrivateIpAddressConfig" - }, - "SecondaryPrivateIpAddressCount":{"shape":"Integer"}, - "AssociatePublicIpAddress":{"shape":"Boolean"}, - "Groups":{ - "shape":"ScheduledInstancesSecurityGroupIdSet", - "locationName":"Group" - }, - "DeleteOnTermination":{"shape":"Boolean"} - } - }, - "ScheduledInstancesNetworkInterfaceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesNetworkInterface", - "locationName":"NetworkInterface" - } - }, - "ScheduledInstancesPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{"shape":"String"}, - "GroupName":{"shape":"String"} - } - }, - "ScheduledInstancesPrivateIpAddressConfig":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{"shape":"String"}, - "Primary":{"shape":"Boolean"} - } - }, - "ScheduledInstancesSecurityGroupIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "Description":{ - "shape":"String", - "locationName":"groupDescription" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - }, - "IpPermissionsEgress":{ - "shape":"IpPermissionList", - "locationName":"ipPermissionsEgress" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SecurityGroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"SecurityGroup", - "locationName":"item" - } - }, - "SecurityGroupReference":{ - "type":"structure", - "required":[ - "GroupId", - "ReferencingVpcId" - ], - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "ReferencingVpcId":{ - "shape":"String", - "locationName":"referencingVpcId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "SecurityGroupReferences":{ - "type":"list", - "member":{ - "shape":"SecurityGroupReference", - "locationName":"item" - } - }, - "SecurityGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroup" - } - }, - "ShutdownBehavior":{ - "type":"string", - "enum":[ - "stop", - "terminate" - ] - }, - "SlotDateTimeRangeRequest":{ - "type":"structure", - "required":[ - "EarliestTime", - "LatestTime" - ], - "members":{ - "EarliestTime":{"shape":"DateTime"}, - "LatestTime":{"shape":"DateTime"} - } - }, - "SlotStartTimeRangeRequest":{ - "type":"structure", - "members":{ - "EarliestTime":{"shape":"DateTime"}, - "LatestTime":{"shape":"DateTime"} - } - }, - "Snapshot":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "State":{ - "shape":"SnapshotState", - "locationName":"status" - }, - "StateMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "OwnerAlias":{ - "shape":"String", - "locationName":"ownerAlias" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - }, - "DataEncryptionKeyId":{ - "shape":"String", - "locationName":"dataEncryptionKeyId" - } - } - }, - "SnapshotAttributeName":{ - "type":"string", - "enum":[ - "productCodes", - "createVolumePermission" - ] - }, - "SnapshotDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SnapshotDetailList":{ - "type":"list", - "member":{ - "shape":"SnapshotDetail", - "locationName":"item" - } - }, - "SnapshotDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"} - } - }, - "SnapshotIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SnapshotId" - } - }, - "SnapshotList":{ - "type":"list", - "member":{ - "shape":"Snapshot", - "locationName":"item" - } - }, - "SnapshotState":{ - "type":"string", - "enum":[ - "pending", - "completed", - "error" - ] - }, - "SnapshotTaskDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SpotDatafeedSubscription":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "State":{ - "shape":"DatafeedSubscriptionState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - } - } - }, - "SpotFleetLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "Monitoring":{ - "shape":"SpotFleetMonitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "WeightedCapacity":{ - "shape":"Double", - "locationName":"weightedCapacity" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - } - } - }, - "SpotFleetMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "SpotFleetRequestConfig":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "SpotFleetRequestState", - "SpotFleetRequestConfig", - "CreateTime" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "SpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"spotFleetRequestState" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - } - } - }, - "SpotFleetRequestConfigData":{ - "type":"structure", - "required":[ - "SpotPrice", - "TargetCapacity", - "IamFleetRole", - "LaunchSpecifications" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "TerminateInstancesWithExpiration":{ - "shape":"Boolean", - "locationName":"terminateInstancesWithExpiration" - }, - "IamFleetRole":{ - "shape":"String", - "locationName":"iamFleetRole" - }, - "LaunchSpecifications":{ - "shape":"LaunchSpecsList", - "locationName":"launchSpecifications" - }, - "ExcessCapacityTerminationPolicy":{ - "shape":"ExcessCapacityTerminationPolicy", - "locationName":"excessCapacityTerminationPolicy" - }, - "AllocationStrategy":{ - "shape":"AllocationStrategy", - "locationName":"allocationStrategy" - }, - "FulfilledCapacity":{ - "shape":"Double", - "locationName":"fulfilledCapacity" - }, - "Type":{ - "shape":"FleetType", - "locationName":"type" - } - } - }, - "SpotFleetRequestConfigSet":{ - "type":"list", - "member":{ - "shape":"SpotFleetRequestConfig", - "locationName":"item" - } - }, - "SpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "State":{ - "shape":"SpotInstanceState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - }, - "Status":{ - "shape":"SpotInstanceStatus", - "locationName":"status" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "LaunchSpecification":{ - "shape":"LaunchSpecification", - "locationName":"launchSpecification" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "BlockDurationMinutes":{ - "shape":"Integer", - "locationName":"blockDurationMinutes" - }, - "ActualBlockHourlyPrice":{ - "shape":"String", - "locationName":"actualBlockHourlyPrice" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "LaunchedAvailabilityZone":{ - "shape":"String", - "locationName":"launchedAvailabilityZone" - } - } - }, - "SpotInstanceRequestIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SpotInstanceRequestId" - } - }, - "SpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"SpotInstanceRequest", - "locationName":"item" - } - }, - "SpotInstanceState":{ - "type":"string", - "enum":[ - "open", - "active", - "closed", - "cancelled", - "failed" - ] - }, - "SpotInstanceStateFault":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceStatus":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceType":{ - "type":"string", - "enum":[ - "one-time", - "persistent" - ] - }, - "SpotPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "SpotPrice":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "SpotPriceHistoryList":{ - "type":"list", - "member":{ - "shape":"SpotPrice", - "locationName":"item" - } - }, - "StaleIpPermission":{ - "type":"structure", - "members":{ - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "IpRanges":{ - "shape":"IpRanges", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdSet", - "locationName":"prefixListIds" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairSet", - "locationName":"groups" - } - } - }, - "StaleIpPermissionSet":{ - "type":"list", - "member":{ - "shape":"StaleIpPermission", - "locationName":"item" - } - }, - "StaleSecurityGroup":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "StaleIpPermissions":{ - "shape":"StaleIpPermissionSet", - "locationName":"staleIpPermissions" - }, - "StaleIpPermissionsEgress":{ - "shape":"StaleIpPermissionSet", - "locationName":"staleIpPermissionsEgress" - } - } - }, - "StaleSecurityGroupSet":{ - "type":"list", - "member":{ - "shape":"StaleSecurityGroup", - "locationName":"item" - } - }, - "StartInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "StartInstancesResult":{ - "type":"structure", - "members":{ - "StartingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "State":{ - "type":"string", - "enum":[ - "Pending", - "Available", - "Deleting", - "Deleted" - ] - }, - "StateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "Status":{ - "type":"string", - "enum":[ - "MoveInProgress", - "InVpc", - "InClassic" - ] - }, - "StatusName":{ - "type":"string", - "enum":["reachability"] - }, - "StatusType":{ - "type":"string", - "enum":[ - "passed", - "failed", - "insufficient-data", - "initializing" - ] - }, - "StopInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "StopInstancesResult":{ - "type":"structure", - "members":{ - "StoppingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "Storage":{ - "type":"structure", - "members":{ - "S3":{"shape":"S3Storage"} - } - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "State":{ - "shape":"SubnetState", - "locationName":"state" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "AvailableIpAddressCount":{ - "shape":"Integer", - "locationName":"availableIpAddressCount" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "DefaultForAz":{ - "shape":"Boolean", - "locationName":"defaultForAz" - }, - "MapPublicIpOnLaunch":{ - "shape":"Boolean", - "locationName":"mapPublicIpOnLaunch" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SubnetIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetId" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"item" - } - }, - "SubnetState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "SummaryStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data", - "not-applicable", - "initializing" - ] - }, - "Tag":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescription":{ - "type":"structure", - "members":{ - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "ResourceType":{ - "shape":"ResourceType", - "locationName":"resourceType" - }, - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescriptionList":{ - "type":"list", - "member":{ - "shape":"TagDescription", - "locationName":"item" - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"item" - } - }, - "TelemetryStatus":{ - "type":"string", - "enum":[ - "UP", - "DOWN" - ] - }, - "Tenancy":{ - "type":"string", - "enum":[ - "default", - "dedicated", - "host" - ] - }, - "TerminateInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "TerminateInstancesResult":{ - "type":"structure", - "members":{ - "TerminatingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "TrafficType":{ - "type":"string", - "enum":[ - "ACCEPT", - "REJECT", - "ALL" - ] - }, - "UnassignPrivateIpAddressesRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "PrivateIpAddresses" - ], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - } - } - }, - "UnmonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "UnmonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "UnsuccessfulItem":{ - "type":"structure", - "required":["Error"], - "members":{ - "Error":{ - "shape":"UnsuccessfulItemError", - "locationName":"error" - }, - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - } - } - }, - "UnsuccessfulItemError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "UnsuccessfulItemList":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UnsuccessfulItemSet":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UserBucket":{ - "type":"structure", - "members":{ - "S3Bucket":{"shape":"String"}, - "S3Key":{"shape":"String"} - } - }, - "UserBucketDetails":{ - "type":"structure", - "members":{ - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "UserData":{ - "type":"structure", - "members":{ - "Data":{ - "shape":"String", - "locationName":"data" - } - } - }, - "UserGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserGroup" - } - }, - "UserIdGroupPair":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "PeeringStatus":{ - "shape":"String", - "locationName":"peeringStatus" - } - } - }, - "UserIdGroupPairList":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdGroupPairSet":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserId" - } - }, - "ValueStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "VgwTelemetry":{ - "type":"structure", - "members":{ - "OutsideIpAddress":{ - "shape":"String", - "locationName":"outsideIpAddress" - }, - "Status":{ - "shape":"TelemetryStatus", - "locationName":"status" - }, - "LastStatusChange":{ - "shape":"DateTime", - "locationName":"lastStatusChange" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "AcceptedRouteCount":{ - "shape":"Integer", - "locationName":"acceptedRouteCount" - } - } - }, - "VgwTelemetryList":{ - "type":"list", - "member":{ - "shape":"VgwTelemetry", - "locationName":"item" - } - }, - "VirtualizationType":{ - "type":"string", - "enum":[ - "hvm", - "paravirtual" - ] - }, - "Volume":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Size":{ - "shape":"Integer", - "locationName":"size" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "State":{ - "shape":"VolumeState", - "locationName":"status" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "Attachments":{ - "shape":"VolumeAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "VolumeAttachment":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Device":{ - "shape":"String", - "locationName":"device" - }, - "State":{ - "shape":"VolumeAttachmentState", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "VolumeAttachmentList":{ - "type":"list", - "member":{ - "shape":"VolumeAttachment", - "locationName":"item" - } - }, - "VolumeAttachmentState":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "VolumeAttributeName":{ - "type":"string", - "enum":[ - "autoEnableIO", - "productCodes" - ] - }, - "VolumeDetail":{ - "type":"structure", - "required":["Size"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - } - } - }, - "VolumeIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VolumeId" - } - }, - "VolumeList":{ - "type":"list", - "member":{ - "shape":"Volume", - "locationName":"item" - } - }, - "VolumeState":{ - "type":"string", - "enum":[ - "creating", - "available", - "in-use", - "deleting", - "deleted", - "error" - ] - }, - "VolumeStatusAction":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusActionsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusAction", - "locationName":"item" - } - }, - "VolumeStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"VolumeStatusName", - "locationName":"name" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "VolumeStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusDetails", - "locationName":"item" - } - }, - "VolumeStatusEvent":{ - "type":"structure", - "members":{ - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusEventsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusEvent", - "locationName":"item" - } - }, - "VolumeStatusInfo":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"VolumeStatusInfoStatus", - "locationName":"status" - }, - "Details":{ - "shape":"VolumeStatusDetailsList", - "locationName":"details" - } - } - }, - "VolumeStatusInfoStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data" - ] - }, - "VolumeStatusItem":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VolumeStatus":{ - "shape":"VolumeStatusInfo", - "locationName":"volumeStatus" - }, - "Events":{ - "shape":"VolumeStatusEventsList", - "locationName":"eventsSet" - }, - "Actions":{ - "shape":"VolumeStatusActionsList", - "locationName":"actionsSet" - } - } - }, - "VolumeStatusList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusItem", - "locationName":"item" - } - }, - "VolumeStatusName":{ - "type":"string", - "enum":[ - "io-enabled", - "io-performance" - ] - }, - "VolumeType":{ - "type":"string", - "enum":[ - "standard", - "io1", - "gp2", - "sc1", - "st1" - ] - }, - "Vpc":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"VpcState", - "locationName":"state" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"isDefault" - } - } - }, - "VpcAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "VpcAttachmentList":{ - "type":"list", - "member":{ - "shape":"VpcAttachment", - "locationName":"item" - } - }, - "VpcAttributeName":{ - "type":"string", - "enum":[ - "enableDnsSupport", - "enableDnsHostnames" - ] - }, - "VpcClassicLink":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkEnabled":{ - "shape":"Boolean", - "locationName":"classicLinkEnabled" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpcClassicLinkIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcClassicLinkList":{ - "type":"list", - "member":{ - "shape":"VpcClassicLink", - "locationName":"item" - } - }, - "VpcEndpoint":{ - "type":"structure", - "members":{ - "VpcEndpointId":{ - "shape":"String", - "locationName":"vpcEndpointId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ServiceName":{ - "shape":"String", - "locationName":"serviceName" - }, - "State":{ - "shape":"State", - "locationName":"state" - }, - "PolicyDocument":{ - "shape":"String", - "locationName":"policyDocument" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"routeTableIdSet" - }, - "CreationTimestamp":{ - "shape":"DateTime", - "locationName":"creationTimestamp" - } - } - }, - "VpcEndpointSet":{ - "type":"list", - "member":{ - "shape":"VpcEndpoint", - "locationName":"item" - } - }, - "VpcIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcList":{ - "type":"list", - "member":{ - "shape":"Vpc", - "locationName":"item" - } - }, - "VpcPeeringConnection":{ - "type":"structure", - "members":{ - "AccepterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"accepterVpcInfo" - }, - "ExpirationTime":{ - "shape":"DateTime", - "locationName":"expirationTime" - }, - "RequesterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"requesterVpcInfo" - }, - "Status":{ - "shape":"VpcPeeringConnectionStateReason", - "locationName":"status" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "VpcPeeringConnectionList":{ - "type":"list", - "member":{ - "shape":"VpcPeeringConnection", - "locationName":"item" - } - }, - "VpcPeeringConnectionOptionsDescription":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalClassicLinkToRemoteVpc" - }, - "AllowEgressFromLocalVpcToRemoteClassicLink":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalVpcToRemoteClassicLink" - } - } - }, - "VpcPeeringConnectionStateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"VpcPeeringConnectionStateReasonCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "VpcPeeringConnectionStateReasonCode":{ - "type":"string", - "enum":[ - "initiating-request", - "pending-acceptance", - "active", - "deleted", - "rejected", - "failed", - "expired", - "provisioning", - "deleting" - ] - }, - "VpcPeeringConnectionVpcInfo":{ - "type":"structure", - "members":{ - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeeringOptions":{ - "shape":"VpcPeeringConnectionOptionsDescription", - "locationName":"peeringOptions" - } - } - }, - "VpcState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "VpnConnection":{ - "type":"structure", - "members":{ - "VpnConnectionId":{ - "shape":"String", - "locationName":"vpnConnectionId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "CustomerGatewayConfiguration":{ - "shape":"String", - "locationName":"customerGatewayConfiguration" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VgwTelemetry":{ - "shape":"VgwTelemetryList", - "locationName":"vgwTelemetry" - }, - "Options":{ - "shape":"VpnConnectionOptions", - "locationName":"options" - }, - "Routes":{ - "shape":"VpnStaticRouteList", - "locationName":"routes" - } - } - }, - "VpnConnectionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnConnectionId" - } - }, - "VpnConnectionList":{ - "type":"list", - "member":{ - "shape":"VpnConnection", - "locationName":"item" - } - }, - "VpnConnectionOptions":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnConnectionOptionsSpecification":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnGateway":{ - "type":"structure", - "members":{ - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VpcAttachments":{ - "shape":"VpcAttachmentList", - "locationName":"attachments" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpnGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnGatewayId" - } - }, - "VpnGatewayList":{ - "type":"list", - "member":{ - "shape":"VpnGateway", - "locationName":"item" - } - }, - "VpnState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "VpnStaticRoute":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "Source":{ - "shape":"VpnStaticRouteSource", - "locationName":"source" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - } - } - }, - "VpnStaticRouteList":{ - "type":"list", - "member":{ - "shape":"VpnStaticRoute", - "locationName":"item" - } - }, - "VpnStaticRouteSource":{ - "type":"string", - "enum":["Static"] - }, - "ZoneNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ZoneName" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/docs-2.json deleted file mode 100644 index 8d9b760b9..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/docs-2.json +++ /dev/null @@ -1,6382 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon Elastic Compute Cloud

Amazon Elastic Compute Cloud (Amazon EC2) provides resizable computing capacity in the Amazon Web Services (AWS) cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster.

", - "operations": { - "AcceptVpcPeeringConnection": "

Accept a VPC peering connection request. To accept a request, the VPC peering connection must be in the pending-acceptance state, and you must be the owner of the peer VPC. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests.

", - "AllocateAddress": "

Acquires an Elastic IP address.

An Elastic IP address is for use either in the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "AllocateHosts": "

Allocates a Dedicated host to your account. At minimum you need to specify the instance size type, Availability Zone, and quantity of hosts you want to allocate.

", - "AssignPrivateIpAddresses": "

Assigns one or more secondary private IP addresses to the specified network interface. You can specify one or more specific secondary IP addresses, or you can specify the number of secondary IP addresses to be automatically assigned within the subnet's CIDR block range. The number of secondary IP addresses that you can assign to an instance varies by instance type. For information about instance types, see Instance Types in the Amazon Elastic Compute Cloud User Guide. For more information about Elastic IP addresses, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

AssignPrivateIpAddresses is available only in EC2-VPC.

", - "AssociateAddress": "

Associates an Elastic IP address with an instance or a network interface.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

[EC2-Classic, VPC in an EC2-VPC-only account] If the Elastic IP address is already associated with a different instance, it is disassociated from that instance and associated with the specified instance.

[VPC in an EC2-Classic account] If you don't specify a private IP address, the Elastic IP address is associated with the primary IP address. If the Elastic IP address is already associated with a different instance or a network interface, you get an error unless you allow reassociation.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

", - "AssociateDhcpOptions": "

Associates a set of DHCP options (that you've previously created) with the specified VPC, or associates no DHCP options with the VPC.

After you associate the options with the VPC, any existing instances and all new instances that you launch in that VPC use the options. You don't need to restart or relaunch the instances. They automatically pick up the changes within a few hours, depending on how frequently the instance renews its DHCP lease. You can explicitly renew the lease using the operating system on the instance.

For more information, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "AssociateRouteTable": "

Associates a subnet with a route table. The subnet and route table must be in the same VPC. This association causes traffic originating from the subnet to be routed according to the routes in the route table. The action returns an association ID, which you need in order to disassociate the route table from the subnet later. A route table can be associated with multiple subnets.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "AttachClassicLinkVpc": "

Links an EC2-Classic instance to a ClassicLink-enabled VPC through one or more of the VPC's security groups. You cannot link an EC2-Classic instance to more than one VPC at a time. You can only link an instance that's in the running state. An instance is automatically unlinked from a VPC when it's stopped - you can link it to the VPC again when you restart it.

After you've linked an instance, you cannot change the VPC security groups that are associated with it. To change the security groups, you must first unlink the instance, and then link it again.

Linking your instance to a VPC is sometimes referred to as attaching your instance.

", - "AttachInternetGateway": "

Attaches an Internet gateway to a VPC, enabling connectivity between the Internet and the VPC. For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "AttachNetworkInterface": "

Attaches a network interface to an instance.

", - "AttachVolume": "

Attaches an EBS volume to a running or stopped instance and exposes it to the instance with the specified device name.

Encrypted EBS volumes may only be attached to instances that support Amazon EBS encryption. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For a list of supported device names, see Attaching an EBS Volume to an Instance. Any device names that aren't reserved for instance store volumes can be used for EBS volumes. For more information, see Amazon EC2 Instance Store in the Amazon Elastic Compute Cloud User Guide.

If a volume has an AWS Marketplace product code:

  • The volume can be attached only to a stopped instance.

  • AWS Marketplace product codes are copied from the volume to the instance.

  • You must be subscribed to the product.

  • The instance type and operating system of the instance must support the product. For example, you can't detach a volume from a Windows instance and attach it to a Linux instance.

For an overview of the AWS Marketplace, see Introducing AWS Marketplace.

For more information about EBS volumes, see Attaching Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "AttachVpnGateway": "

Attaches a virtual private gateway to a VPC. For more information, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "AuthorizeSecurityGroupEgress": "

[EC2-VPC only] Adds one or more egress rules to a security group for use with a VPC. Specifically, this action permits instances to send traffic to one or more destination CIDR IP address ranges, or to one or more destination security groups for the same VPC. This action doesn't apply to security groups for use in EC2-Classic. For more information, see Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

You can have up to 50 rules per security group (covering both ingress and egress rules).

Each rule consists of the protocol (for example, TCP), plus either a CIDR range or a source group. For the TCP and UDP protocols, you must also specify the destination port or port range. For the ICMP protocol, you must also specify the ICMP type and code. You can use -1 for the type or code to mean all types or all codes.

Rule changes are propagated to affected instances as quickly as possible. However, a small delay might occur.

", - "AuthorizeSecurityGroupIngress": "

Adds one or more ingress rules to a security group.

EC2-Classic: You can have up to 100 rules per group.

EC2-VPC: You can have up to 50 rules per group (covering both ingress and egress rules).

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

[EC2-Classic] This action gives one or more CIDR IP address ranges permission to access a security group in your account, or gives one or more security groups (called the source groups) permission to access a security group for your account. A source group can be for your own AWS account, or another.

[EC2-VPC] This action gives one or more CIDR IP address ranges permission to access a security group in your VPC, or gives one or more other security groups (called the source groups) permission to access a security group for your VPC. The security groups must all be for the same VPC.

", - "BundleInstance": "

Bundles an Amazon instance store-backed Windows instance.

During bundling, only the root device volume (C:\\) is bundled. Data on other instance store volumes is not preserved.

This action is not applicable for Linux/Unix instances or Windows instances that are backed by Amazon EBS.

For more information, see Creating an Instance Store-Backed Windows AMI.

", - "CancelBundleTask": "

Cancels a bundling operation for an instance store-backed Windows instance.

", - "CancelConversionTask": "

Cancels an active conversion task. The task can be the import of an instance or volume. The action removes all artifacts of the conversion, including a partially uploaded volume or instance. If the conversion is complete or is in the process of transferring the final disk image, the command fails and returns an exception.

For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

", - "CancelExportTask": "

Cancels an active export task. The request removes all artifacts of the export, including any partially-created Amazon S3 objects. If the export task is complete or is in the process of transferring the final disk image, the command fails and returns an error.

", - "CancelImportTask": "

Cancels an in-process import virtual machine or import snapshot task.

", - "CancelReservedInstancesListing": "

Cancels the specified Reserved Instance listing in the Reserved Instance Marketplace.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CancelSpotFleetRequests": "

Cancels the specified Spot fleet requests.

After you cancel a Spot fleet request, the Spot fleet launches no new Spot instances. You must specify whether the Spot fleet should also terminate its Spot instances. If you terminate the instances, the Spot fleet request enters the cancelled_terminating state. Otherwise, the Spot fleet request enters the cancelled_running state and the instances continue to run until they are interrupted or you terminate them manually.

", - "CancelSpotInstanceRequests": "

Cancels one or more Spot instance requests. Spot instances are instances that Amazon EC2 starts on your behalf when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

Canceling a Spot instance request does not terminate running Spot instances associated with the request.

", - "ConfirmProductInstance": "

Determines whether a product code is associated with an instance. This action can only be used by the owner of the product code. It is useful when a product code owner needs to verify whether another user's instance is eligible for support.

", - "CopyImage": "

Initiates the copy of an AMI from the specified source region to the current region. You specify the destination region by using its endpoint when making the request.

For more information, see Copying AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshot": "

Copies a point-in-time snapshot of an EBS volume and stores it in Amazon S3. You can copy the snapshot within the same region or from one region to another. You can use the snapshot to create EBS volumes or Amazon Machine Images (AMIs). The snapshot is copied to the regional endpoint that you send the HTTP request to.

Copies of encrypted EBS snapshots remain encrypted. Copies of unencrypted snapshots remain unencrypted, unless the Encrypted flag is specified during the snapshot copy operation. By default, encrypted snapshot copies use the default AWS Key Management Service (AWS KMS) customer master key (CMK); however, you can specify a non-default CMK with the KmsKeyId parameter.

For more information, see Copying an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGateway": "

Provides information to AWS about your VPN customer gateway device. The customer gateway is the appliance at your end of the VPN connection. (The device on the AWS side of the VPN connection is the virtual private gateway.) You must provide the Internet-routable IP address of the customer gateway's external interface. The IP address must be static and may be behind a device performing network address translation (NAT).

For devices that use Border Gateway Protocol (BGP), you can also provide the device's BGP Autonomous System Number (ASN). You can use an existing ASN assigned to your network. If you don't have an ASN already, you can use a private ASN (in the 64512 - 65534 range).

Amazon EC2 supports all 2-byte ASN numbers in the range of 1 - 65534, with the exception of 7224, which is reserved in the us-east-1 region, and 9059, which is reserved in the eu-west-1 region.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

You cannot create more than one customer gateway with the same VPN type, IP address, and BGP ASN parameter values. If you run an identical request more than one time, the first request creates the customer gateway, and subsequent requests return information about the existing customer gateway. The subsequent requests do not create new customer gateway resources.

", - "CreateDhcpOptions": "

Creates a set of DHCP options for your VPC. After creating the set, you must associate it with the VPC, causing all existing and new instances that you launch in the VPC to use this set of DHCP options. The following are the individual DHCP options you can specify. For more information about the options, see RFC 2132.

  • domain-name-servers - The IP addresses of up to four domain name servers, or AmazonProvidedDNS. The default DHCP option set specifies AmazonProvidedDNS. If specifying more than one domain name server, specify the IP addresses in a single parameter, separated by commas.

  • domain-name - If you're using AmazonProvidedDNS in \"us-east-1\", specify \"ec2.internal\". If you're using AmazonProvidedDNS in another region, specify \"region.compute.internal\" (for example, \"ap-northeast-1.compute.internal\"). Otherwise, specify a domain name (for example, \"MyCompany.com\"). Important: Some Linux operating systems accept multiple domain names separated by spaces. However, Windows and other Linux operating systems treat the value as a single domain, which results in unexpected behavior. If your DHCP options set is associated with a VPC that has instances with multiple operating systems, specify only one domain name.

  • ntp-servers - The IP addresses of up to four Network Time Protocol (NTP) servers.

  • netbios-name-servers - The IP addresses of up to four NetBIOS name servers.

  • netbios-node-type - The NetBIOS node type (1, 2, 4, or 8). We recommend that you specify 2 (broadcast and multicast are not currently supported). For more information about these node types, see RFC 2132.

Your VPC automatically starts out with a set of DHCP options that includes only a DNS server that we provide (AmazonProvidedDNS). If you create a set of options, and if your VPC has an Internet gateway, make sure to set the domain-name-servers option either to AmazonProvidedDNS or to a domain name server of your choice. For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "CreateFlowLogs": "

Creates one or more flow logs to capture IP traffic for a specific network interface, subnet, or VPC. Flow logs are delivered to a specified log group in Amazon CloudWatch Logs. If you specify a VPC or subnet in the request, a log stream is created in CloudWatch Logs for each network interface in the subnet or VPC. Log streams can include information about accepted and rejected traffic to a network interface. You can view the data in your log streams using Amazon CloudWatch Logs.

In your request, you must also specify an IAM role that has permission to publish logs to CloudWatch Logs.

", - "CreateImage": "

Creates an Amazon EBS-backed AMI from an Amazon EBS-backed instance that is either running or stopped.

If you customized your instance with instance store volumes or EBS volumes in addition to the root device volume, the new AMI contains block device mapping information for those volumes. When you launch an instance from this new AMI, the instance automatically launches with those additional volumes.

For more information, see Creating Amazon EBS-Backed Linux AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateInstanceExportTask": "

Exports a running or stopped instance to an S3 bucket.

For information about the supported operating systems, image formats, and known limitations for the types of instances you can export, see Exporting EC2 Instances in the Amazon Elastic Compute Cloud User Guide.

", - "CreateInternetGateway": "

Creates an Internet gateway for use with a VPC. After creating the Internet gateway, you attach it to a VPC using AttachInternetGateway.

For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "CreateKeyPair": "

Creates a 2048-bit RSA key pair with the specified name. Amazon EC2 stores the public key and displays the private key for you to save to a file. The private key is returned as an unencrypted PEM encoded PKCS#8 private key. If a key with the specified name already exists, Amazon EC2 returns an error.

You can have up to five thousand key pairs per region.

The key pair returned to you is available only in the region in which you create it. To create a key pair that is available in all regions, use ImportKeyPair.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateNatGateway": "

Creates a NAT gateway in the specified subnet. A NAT gateway can be used to enable instances in a private subnet to connect to the Internet. This action creates a network interface in the specified subnet with a private IP address from the IP address range of the subnet. For more information, see NAT Gateways in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAcl": "

Creates a network ACL in a VPC. Network ACLs provide an optional layer of security (in addition to security groups) for the instances in your VPC.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAclEntry": "

Creates an entry (a rule) in a network ACL with the specified rule number. Each network ACL has a set of numbered ingress rules and a separate set of numbered egress rules. When determining whether a packet should be allowed in or out of a subnet associated with the ACL, we process the entries in the ACL according to the rule numbers, in ascending order. Each network ACL has a set of ingress rules and a separate set of egress rules.

We recommend that you leave room between the rule numbers (for example, 100, 110, 120, ...), and not number them one right after the other (for example, 101, 102, 103, ...). This makes it easier to add a rule between existing ones without having to renumber the rules.

After you add an entry, you can't modify it; you must either replace it, or create an entry and delete the old one.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkInterface": "

Creates a network interface in the specified subnet.

For more information about network interfaces, see Elastic Network Interfaces in the Amazon Elastic Compute Cloud User Guide.

", - "CreatePlacementGroup": "

Creates a placement group that you launch cluster instances into. You must give the group a name that's unique within the scope of your account.

For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListing": "

Creates a listing for Amazon EC2 Reserved Instances to be sold in the Reserved Instance Marketplace. You can submit one Reserved Instance listing at a time. To get a list of your Reserved Instances, you can use the DescribeReservedInstances operation.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

To sell your Reserved Instances, you must first register as a seller in the Reserved Instance Marketplace. After completing the registration process, you can create a Reserved Instance Marketplace listing of some or all of your Reserved Instances, and specify the upfront price to receive for them. Your Reserved Instance listings then become available for purchase. To view the details of your Reserved Instance listing, you can use the DescribeReservedInstancesListings operation.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CreateRoute": "

Creates a route in a route table within a VPC.

You must specify one of the following targets: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, or network interface.

When determining how to route traffic, we use the route with the most specific match. For example, let's say the traffic is destined for 192.0.2.3, and the route table includes the following two routes:

  • 192.0.2.0/24 (goes to some target A)

  • 192.0.2.0/28 (goes to some target B)

Both routes apply to the traffic destined for 192.0.2.3. However, the second route in the list covers a smaller number of IP addresses and is therefore more specific, so we use that route to determine where to target the traffic.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateRouteTable": "

Creates a route table for the specified VPC. After you create a route table, you can add routes and associate the table with a subnet.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateSecurityGroup": "

Creates a security group.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

EC2-Classic: You can have up to 500 security groups.

EC2-VPC: You can create up to 500 security groups per VPC.

When you create a security group, you specify a friendly name of your choice. You can have a security group for use in EC2-Classic with the same name as a security group for use in a VPC. However, you can't have two security groups for use in EC2-Classic with the same name or two security groups for use in a VPC with the same name.

You have a default security group for use in EC2-Classic and a default security group for use in your VPC. If you don't specify a security group when you launch an instance, the instance is launched into the appropriate default security group. A default security group includes a default rule that grants instances unrestricted network access to each other.

You can add or remove rules from your security groups using AuthorizeSecurityGroupIngress, AuthorizeSecurityGroupEgress, RevokeSecurityGroupIngress, and RevokeSecurityGroupEgress.

", - "CreateSnapshot": "

Creates a snapshot of an EBS volume and stores it in Amazon S3. You can use snapshots for backups, to make copies of EBS volumes, and to save data before shutting down an instance.

When a snapshot is created, any AWS Marketplace product codes that are associated with the source volume are propagated to the snapshot.

You can take a snapshot of an attached volume that is in use. However, snapshots only capture data that has been written to your EBS volume at the time the snapshot command is issued; this may exclude any data that has been cached by any applications or the operating system. If you can pause any file systems on the volume long enough to take a snapshot, your snapshot should be complete. However, if you cannot pause all file writes to the volume, you should unmount the volume from within the instance, issue the snapshot command, and then remount the volume to ensure a consistent and complete snapshot. You may remount and use your volume while the snapshot status is pending.

To create a snapshot for EBS volumes that serve as root devices, you should stop the instance before taking the snapshot.

Snapshots that are taken from encrypted volumes are automatically encrypted. Volumes that are created from encrypted snapshots are also automatically encrypted. Your encrypted volumes and any associated snapshots always remain protected.

For more information, see Amazon Elastic Block Store and Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSpotDatafeedSubscription": "

Creates a data feed for Spot instances, enabling you to view Spot instance usage logs. You can create one data feed per AWS account. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSubnet": "

Creates a subnet in an existing VPC.

When you create each subnet, you provide the VPC ID and the CIDR block you want for the subnet. After you create a subnet, you can't change its CIDR block. The subnet's CIDR block can be the same as the VPC's CIDR block (assuming you want only a single subnet in the VPC), or a subset of the VPC's CIDR block. If you create more than one subnet in a VPC, the subnets' CIDR blocks must not overlap. The smallest subnet (and VPC) you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses).

AWS reserves both the first four and the last IP address in each subnet's CIDR block. They're not available for use.

If you add more than one subnet to a VPC, they're set up in a star topology with a logical router in the middle.

If you launch an instance in a VPC using an Amazon EBS-backed AMI, the IP address doesn't change if you stop and restart the instance (unlike a similar instance launched outside a VPC, which gets a new IP address when restarted). It's therefore possible to have a subnet with no running instances (they're all stopped), but no remaining IP addresses available.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "CreateTags": "

Adds or overwrites one or more tags for the specified Amazon EC2 resource or resources. Each resource can have a maximum of 10 tags. Each tag consists of a key and optional value. Tag keys must be unique per resource.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide. For more information about creating IAM policies that control users' access to resources based on tags, see Supported Resource-Level Permissions for Amazon EC2 API Actions in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVolume": "

Creates an EBS volume that can be attached to an instance in the same Availability Zone. The volume is created in the regional endpoint that you send the HTTP request to. For more information see Regions and Endpoints.

You can create a new empty volume or restore a volume from an EBS snapshot. Any AWS Marketplace product codes from the snapshot are propagated to the volume.

You can create encrypted volumes with the Encrypted parameter. Encrypted volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are also automatically encrypted. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For more information, see Creating or Restoring an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpc": "

Creates a VPC with the specified CIDR block.

The smallest VPC you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses). To help you decide how big to make your VPC, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

By default, each instance you launch in the VPC has the default DHCP options, which includes only a default DNS server that we provide (AmazonProvidedDNS). For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

You can specify the instance tenancy value for the VPC when you create it. You can't change this value for the VPC after you create it. For more information, see Dedicated Instances in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpcEndpoint": "

Creates a VPC endpoint for a specified AWS service. An endpoint enables you to create a private connection between your VPC and another AWS service in your account. You can specify an endpoint policy to attach to the endpoint that will control access to the service from your VPC. You can also specify the VPC route tables that use the endpoint.

Currently, only endpoints to Amazon S3 are supported.

", - "CreateVpcPeeringConnection": "

Requests a VPC peering connection between two VPCs: a requester VPC that you own and a peer VPC with which to create the connection. The peer VPC can belong to another AWS account. The requester VPC and peer VPC cannot have overlapping CIDR blocks.

The owner of the peer VPC must accept the peering request to activate the peering connection. The VPC peering connection request expires after 7 days, after which it cannot be accepted or rejected.

A CreateVpcPeeringConnection request between VPCs with overlapping CIDR blocks results in the VPC peering connection having a status of failed.

", - "CreateVpnConnection": "

Creates a VPN connection between an existing virtual private gateway and a VPN customer gateway. The only supported connection type is ipsec.1.

The response includes information that you need to give to your network administrator to configure your customer gateway.

We strongly recommend that you use HTTPS when calling this operation because the response contains sensitive cryptographic information for configuring your customer gateway.

If you decide to shut down your VPN connection for any reason and later create a new VPN connection, you must reconfigure your customer gateway with the new information returned from this call.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnConnectionRoute": "

Creates a static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnGateway": "

Creates a virtual private gateway. A virtual private gateway is the endpoint on the VPC side of your VPN connection. You can create a virtual private gateway before creating the VPC itself.

For more information about virtual private gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DeleteCustomerGateway": "

Deletes the specified customer gateway. You must delete the VPN connection before you can delete the customer gateway.

", - "DeleteDhcpOptions": "

Deletes the specified set of DHCP options. You must disassociate the set of DHCP options before you can delete it. You can disassociate the set of DHCP options by associating either a new set of options or the default set of options with the VPC.

", - "DeleteFlowLogs": "

Deletes one or more flow logs.

", - "DeleteInternetGateway": "

Deletes the specified Internet gateway. You must detach the Internet gateway from the VPC before you can delete it.

", - "DeleteKeyPair": "

Deletes the specified key pair, by removing the public key from Amazon EC2.

", - "DeleteNatGateway": "

Deletes the specified NAT gateway. Deleting a NAT gateway disassociates its Elastic IP address, but does not release the address from your account. Deleting a NAT gateway does not delete any NAT gateway routes in your route tables.

", - "DeleteNetworkAcl": "

Deletes the specified network ACL. You can't delete the ACL if it's associated with any subnets. You can't delete the default network ACL.

", - "DeleteNetworkAclEntry": "

Deletes the specified ingress or egress entry (rule) from the specified network ACL.

", - "DeleteNetworkInterface": "

Deletes the specified network interface. You must detach the network interface before you can delete it.

", - "DeletePlacementGroup": "

Deletes the specified placement group. You must terminate all instances in the placement group before you can delete the placement group. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteRoute": "

Deletes the specified route from the specified route table.

", - "DeleteRouteTable": "

Deletes the specified route table. You must disassociate the route table from any subnets before you can delete it. You can't delete the main route table.

", - "DeleteSecurityGroup": "

Deletes a security group.

If you attempt to delete a security group that is associated with an instance, or is referenced by another security group, the operation fails with InvalidGroup.InUse in EC2-Classic or DependencyViolation in EC2-VPC.

", - "DeleteSnapshot": "

Deletes the specified snapshot.

When you make periodic snapshots of a volume, the snapshots are incremental, and only the blocks on the device that have changed since your last snapshot are saved in the new snapshot. When you delete a snapshot, only the data not needed for any other snapshot is removed. So regardless of which prior snapshots have been deleted, all active snapshots will have access to all the information needed to restore the volume.

You cannot delete a snapshot of the root device of an EBS volume used by a registered AMI. You must first de-register the AMI before you can delete the snapshot.

For more information, see Deleting an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteSpotDatafeedSubscription": "

Deletes the data feed for Spot instances.

", - "DeleteSubnet": "

Deletes the specified subnet. You must terminate all running instances in the subnet before you can delete the subnet.

", - "DeleteTags": "

Deletes the specified set of tags from the specified set of resources. This call is designed to follow a DescribeTags request.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVolume": "

Deletes the specified EBS volume. The volume must be in the available state (not attached to an instance).

The volume may remain in the deleting state for several minutes.

For more information, see Deleting an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVpc": "

Deletes the specified VPC. You must detach or delete all gateways and resources that are associated with the VPC before you can delete it. For example, you must terminate all instances running in the VPC, delete all security groups associated with the VPC (except the default one), delete all route tables associated with the VPC (except the default one), and so on.

", - "DeleteVpcEndpoints": "

Deletes one or more specified VPC endpoints. Deleting the endpoint also deletes the endpoint routes in the route tables that were associated with the endpoint.

", - "DeleteVpcPeeringConnection": "

Deletes a VPC peering connection. Either the owner of the requester VPC or the owner of the peer VPC can delete the VPC peering connection if it's in the active state. The owner of the requester VPC can delete a VPC peering connection in the pending-acceptance state.

", - "DeleteVpnConnection": "

Deletes the specified VPN connection.

If you're deleting the VPC and its associated components, we recommend that you detach the virtual private gateway from the VPC and delete the VPC before deleting the VPN connection. If you believe that the tunnel credentials for your VPN connection have been compromised, you can delete the VPN connection and create a new one that has new keys, without needing to delete the VPC or virtual private gateway. If you create a new VPN connection, you must reconfigure the customer gateway using the new configuration information returned with the new VPN connection ID.

", - "DeleteVpnConnectionRoute": "

Deletes the specified static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

", - "DeleteVpnGateway": "

Deletes the specified virtual private gateway. We recommend that before you delete a virtual private gateway, you detach it from the VPC and delete the VPN connection. Note that you don't need to delete the virtual private gateway if you plan to delete and recreate the VPN connection between your VPC and your network.

", - "DeregisterImage": "

Deregisters the specified AMI. After you deregister an AMI, it can't be used to launch new instances.

This command does not delete the AMI.

", - "DescribeAccountAttributes": "

Describes attributes of your AWS account. The following are the supported account attributes:

  • supported-platforms: Indicates whether your account can launch instances into EC2-Classic and EC2-VPC, or only into EC2-VPC.

  • default-vpc: The ID of the default VPC for your account, or none.

  • max-instances: The maximum number of On-Demand instances that you can run.

  • vpc-max-security-groups-per-interface: The maximum number of security groups that you can assign to a network interface.

  • max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-Classic.

  • vpc-max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-VPC.

", - "DescribeAddresses": "

Describes one or more of your Elastic IP addresses.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeAvailabilityZones": "

Describes one or more of the Availability Zones that are available to you. The results include zones only for the region you're currently using. If there is an event impacting an Availability Zone, you can use this request to view the state and any provided message for that Availability Zone.

For more information, see Regions and Availability Zones in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeBundleTasks": "

Describes one or more of your bundling tasks.

Completed bundle tasks are listed for only a limited time. If your bundle task is no longer in the list, you can still register an AMI from it. Just use RegisterImage with the Amazon S3 bucket name and image manifest name you provided to the bundle task.

", - "DescribeClassicLinkInstances": "

Describes one or more of your linked EC2-Classic instances. This request only returns information about EC2-Classic instances linked to a VPC through ClassicLink; you cannot use this request to return information about other instances.

", - "DescribeConversionTasks": "

Describes one or more of your conversion tasks. For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DescribeCustomerGateways": "

Describes one or more of your VPN customer gateways.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeDhcpOptions": "

Describes one or more of your DHCP options sets.

For more information about DHCP options sets, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeExportTasks": "

Describes one or more of your export tasks.

", - "DescribeFlowLogs": "

Describes one or more flow logs. To view the information in your flow logs (the log streams for the network interfaces), you must use the CloudWatch Logs console or the CloudWatch Logs API.

", - "DescribeHosts": "

Describes one or more of your Dedicated hosts.

The results describe only the Dedicated hosts in the region you're currently using. All listed instances consume capacity on your Dedicated host. Dedicated hosts that have recently been released will be listed with the state released.

", - "DescribeIdFormat": "

Describes the ID format settings for your resources on a per-region basis, for example, to view which resource types are enabled for longer IDs. This request only returns information about resource types whose ID formats can be modified; it does not return information about other resource types.

The following resource types support longer IDs: instance | reservation | snapshot | volume.

These settings apply to the IAM user who makes the request; they do not apply to the entire AWS account. By default, an IAM user defaults to the same settings as the root user, unless they explicitly override the settings by running the ModifyIdFormat command. Resources created with longer IDs are visible to all IAM users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "DescribeImageAttribute": "

Describes the specified attribute of the specified AMI. You can specify only one attribute at a time.

", - "DescribeImages": "

Describes one or more of the images (AMIs, AKIs, and ARIs) available to you. Images available to you include public images, private images that you own, and private images owned by other AWS accounts but for which you have explicit launch permissions.

Deregistered images are included in the returned results for an unspecified interval after deregistration.

", - "DescribeImportImageTasks": "

Displays details about an import virtual machine or import snapshot tasks that are already created.

", - "DescribeImportSnapshotTasks": "

Describes your import snapshot tasks.

", - "DescribeInstanceAttribute": "

Describes the specified attribute of the specified instance. You can specify only one attribute at a time. Valid attribute values are: instanceType | kernel | ramdisk | userData | disableApiTermination | instanceInitiatedShutdownBehavior | rootDeviceName | blockDeviceMapping | productCodes | sourceDestCheck | groupSet | ebsOptimized | sriovNetSupport

", - "DescribeInstanceStatus": "

Describes the status of one or more instances. By default, only running instances are described, unless specified otherwise.

Instance status includes the following components:

  • Status checks - Amazon EC2 performs status checks on running EC2 instances to identify hardware and software issues. For more information, see Status Checks for Your Instances and Troubleshooting Instances with Failed Status Checks in the Amazon Elastic Compute Cloud User Guide.

  • Scheduled events - Amazon EC2 can schedule events (such as reboot, stop, or terminate) for your instances related to hardware issues, software updates, or system maintenance. For more information, see Scheduled Events for Your Instances in the Amazon Elastic Compute Cloud User Guide.

  • Instance state - You can manage your instances from the moment you launch them through their termination. For more information, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeInstances": "

Describes one or more of your instances.

If you specify one or more instance IDs, Amazon EC2 returns information for those instances. If you do not specify instance IDs, Amazon EC2 returns information for all relevant instances. If you specify an instance ID that is not valid, an error is returned. If you specify an instance that you do not own, it is not included in the returned results.

Recently terminated instances might appear in the returned results. This interval is usually less than one hour.

", - "DescribeInternetGateways": "

Describes one or more of your Internet gateways.

", - "DescribeKeyPairs": "

Describes one or more of your key pairs.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeMovingAddresses": "

Describes your Elastic IP addresses that are being moved to the EC2-VPC platform, or that are being restored to the EC2-Classic platform. This request does not return information about any other Elastic IP addresses in your account.

", - "DescribeNatGateways": "

Describes one or more of the your NAT gateways.

", - "DescribeNetworkAcls": "

Describes one or more of your network ACLs.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "DescribeNetworkInterfaceAttribute": "

Describes a network interface attribute. You can specify only one attribute at a time.

", - "DescribeNetworkInterfaces": "

Describes one or more of your network interfaces.

", - "DescribePlacementGroups": "

Describes one or more of your placement groups. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribePrefixLists": "

Describes available AWS services in a prefix list format, which includes the prefix list name and prefix list ID of the service and the IP address range for the service. A prefix list ID is required for creating an outbound security group rule that allows traffic from a VPC to access an AWS service through a VPC endpoint.

", - "DescribeRegions": "

Describes one or more regions that are currently available to you.

For a list of the regions supported by Amazon EC2, see Regions and Endpoints.

", - "DescribeReservedInstances": "

Describes one or more of the Reserved Instances that you purchased.

For more information about Reserved Instances, see Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesListings": "

Describes your account's Reserved Instance listings in the Reserved Instance Marketplace.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

As a seller, you choose to list some or all of your Reserved Instances, and you specify the upfront price to receive for them. Your Reserved Instances are then listed in the Reserved Instance Marketplace and are available for purchase.

As a buyer, you specify the configuration of the Reserved Instance to purchase, and the Marketplace matches what you're searching for with what's available. The Marketplace first sells the lowest priced Reserved Instances to you, and continues to sell available Reserved Instance listings to you until your demand is met. You are charged based on the total price of all of the listings that you purchase.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesModifications": "

Describes the modifications made to your Reserved Instances. If no parameter is specified, information about all your Reserved Instances modification requests is returned. If a modification ID is specified, only information about the specific modification is returned.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesOfferings": "

Describes Reserved Instance offerings that are available for purchase. With Reserved Instances, you purchase the right to launch instances for a period of time. During that time period, you do not receive insufficient capacity errors, and you pay a lower usage rate than the rate charged for On-Demand instances for the actual time used.

If you have listed your own Reserved Instances for sale in the Reserved Instance Marketplace, they will be excluded from these results. This is to ensure that you do not purchase your own Reserved Instances.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeRouteTables": "

Describes one or more of your route tables.

Each subnet in your VPC must be associated with a route table. If a subnet is not explicitly associated with any route table, it is implicitly associated with the main route table. This command does not return the subnet ID for implicit associations.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "DescribeScheduledInstanceAvailability": "

Finds available schedules that meet the specified criteria.

You can search for an available schedule no more than 3 months in advance. You must meet the minimum required duration of 1,200 hours per year. For example, the minimum daily schedule is 4 hours, the minimum weekly schedule is 24 hours, and the minimum monthly schedule is 100 hours.

After you find a schedule that meets your needs, call PurchaseScheduledInstances to purchase Scheduled Instances with that schedule.

", - "DescribeScheduledInstances": "

Describes one or more of your Scheduled Instances.

", - "DescribeSecurityGroupReferences": "

[EC2-VPC only] Describes the VPCs on the other side of a VPC peering connection that are referencing the security groups you've specified in this request.

", - "DescribeSecurityGroups": "

Describes one or more of your security groups.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeSnapshotAttribute": "

Describes the specified attribute of the specified snapshot. You can specify only one attribute at a time.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSnapshots": "

Describes one or more of the EBS snapshots available to you. Available snapshots include public snapshots available for any AWS account to launch, private snapshots that you own, and private snapshots owned by another AWS account but for which you've been given explicit create volume permissions.

The create volume permissions fall into the following categories:

  • public: The owner of the snapshot granted create volume permissions for the snapshot to the all group. All AWS accounts have create volume permissions for these snapshots.

  • explicit: The owner of the snapshot granted create volume permissions to a specific AWS account.

  • implicit: An AWS account has implicit create volume permissions for all snapshots it owns.

The list of snapshots returned can be modified by specifying snapshot IDs, snapshot owners, or AWS accounts with create volume permissions. If no options are specified, Amazon EC2 returns all snapshots for which you have create volume permissions.

If you specify one or more snapshot IDs, only snapshots that have the specified IDs are returned. If you specify an invalid snapshot ID, an error is returned. If you specify a snapshot ID for which you do not have access, it is not included in the returned results.

If you specify one or more snapshot owners, only snapshots from the specified owners and for which you have access are returned. The results can include the AWS account IDs of the specified owners, amazon for snapshots owned by Amazon, or self for snapshots that you own.

If you specify a list of restorable users, only snapshots with create snapshot permissions for those users are returned. You can specify AWS account IDs (if you own the snapshots), self for snapshots for which you own or have explicit permissions, or all for public snapshots.

If you are describing a long list of snapshots, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeSnapshots request to retrieve the remaining results.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotDatafeedSubscription": "

Describes the data feed for Spot instances. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotFleetInstances": "

Describes the running instances for the specified Spot fleet.

", - "DescribeSpotFleetRequestHistory": "

Describes the events for the specified Spot fleet request during the specified time.

Spot fleet events are delayed by up to 30 seconds before they can be described. This ensures that you can query by the last evaluated time and not miss a recorded event.

", - "DescribeSpotFleetRequests": "

Describes your Spot fleet requests.

", - "DescribeSpotInstanceRequests": "

Describes the Spot instance requests that belong to your account. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

You can use DescribeSpotInstanceRequests to find a running Spot instance by examining the response. If the status of the Spot instance is fulfilled, the instance ID appears in the response and contains the identifier of the instance. Alternatively, you can use DescribeInstances with a filter to look for instances where the instance lifecycle is spot.

", - "DescribeSpotPriceHistory": "

Describes the Spot price history. The prices returned are listed in chronological order, from the oldest to the most recent, for up to the past 90 days. For more information, see Spot Instance Pricing History in the Amazon Elastic Compute Cloud User Guide.

When you specify a start and end time, this operation returns the prices of the instance types within the time range that you specified and the time when the price changed. The price is valid within the time period that you specified; the response merely indicates the last time that the price changed.

", - "DescribeStaleSecurityGroups": "

[EC2-VPC only] Describes the stale security group rules for security groups in a specified VPC. Rules are stale when they reference a deleted security group in a peer VPC, or a security group in a peer VPC for which the VPC peering connection has been deleted.

", - "DescribeSubnets": "

Describes one or more of your subnets.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeTags": "

Describes one or more of the tags for your EC2 resources.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeAttribute": "

Describes the specified attribute of the specified volume. You can specify only one attribute at a time.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeStatus": "

Describes the status of the specified volumes. Volume status provides the result of the checks performed on your volumes to determine events that can impair the performance of your volumes. The performance of a volume can be affected if an issue occurs on the volume's underlying host. If the volume's underlying host experiences a power outage or system issue, after the system is restored, there could be data inconsistencies on the volume. Volume events notify you if this occurs. Volume actions notify you if any action needs to be taken in response to the event.

The DescribeVolumeStatus operation provides the following information about the specified volumes:

Status: Reflects the current status of the volume. The possible values are ok, impaired , warning, or insufficient-data. If all checks pass, the overall status of the volume is ok. If the check fails, the overall status is impaired. If the status is insufficient-data, then the checks may still be taking place on your volume at the time. We recommend that you retry the request. For more information on volume status, see Monitoring the Status of Your Volumes.

Events: Reflect the cause of a volume status and may require you to take action. For example, if your volume returns an impaired status, then the volume event might be potential-data-inconsistency. This means that your volume has been affected by an issue with the underlying host, has all I/O operations disabled, and may have inconsistent data.

Actions: Reflect the actions you may have to take in response to an event. For example, if the status of the volume is impaired and the volume event shows potential-data-inconsistency, then the action shows enable-volume-io. This means that you may want to enable the I/O operations for the volume by calling the EnableVolumeIO action and then check the volume for data consistency.

Volume status is based on the volume status checks, and does not reflect the volume state. Therefore, volume status does not indicate volumes in the error state (for example, when a volume is incapable of accepting I/O.)

", - "DescribeVolumes": "

Describes the specified EBS volumes.

If you are describing a long list of volumes, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeVolumes request to retrieve the remaining results.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcAttribute": "

Describes the specified attribute of the specified VPC. You can specify only one attribute at a time.

", - "DescribeVpcClassicLink": "

Describes the ClassicLink status of one or more VPCs.

", - "DescribeVpcClassicLinkDnsSupport": "

Describes the ClassicLink DNS support status of one or more VPCs. If enabled, the DNS hostname of a linked EC2-Classic instance resolves to its private IP address when addressed from an instance in the VPC to which it's linked. Similarly, the DNS hostname of an instance in a VPC resolves to its private IP address when addressed from a linked EC2-Classic instance. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcEndpointServices": "

Describes all supported AWS services that can be specified when creating a VPC endpoint.

", - "DescribeVpcEndpoints": "

Describes one or more of your VPC endpoints.

", - "DescribeVpcPeeringConnections": "

Describes one or more of your VPC peering connections.

", - "DescribeVpcs": "

Describes one or more of your VPCs.

", - "DescribeVpnConnections": "

Describes one or more of your VPN connections.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeVpnGateways": "

Describes one or more of your virtual private gateways.

For more information about virtual private gateways, see Adding an IPsec Hardware VPN to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DetachClassicLinkVpc": "

Unlinks (detaches) a linked EC2-Classic instance from a VPC. After the instance has been unlinked, the VPC security groups are no longer associated with it. An instance is automatically unlinked from a VPC when it's stopped.

", - "DetachInternetGateway": "

Detaches an Internet gateway from a VPC, disabling connectivity between the Internet and the VPC. The VPC must not contain any running instances with Elastic IP addresses.

", - "DetachNetworkInterface": "

Detaches a network interface from an instance.

", - "DetachVolume": "

Detaches an EBS volume from an instance. Make sure to unmount any file systems on the device within your operating system before detaching the volume. Failure to do so results in the volume being stuck in a busy state while detaching.

If an Amazon EBS volume is the root device of an instance, it can't be detached while the instance is running. To detach the root volume, stop the instance first.

When a volume with an AWS Marketplace product code is detached from an instance, the product code is no longer associated with the instance.

For more information, see Detaching an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DetachVpnGateway": "

Detaches a virtual private gateway from a VPC. You do this if you're planning to turn off the VPC and not use it anymore. You can confirm a virtual private gateway has been completely detached from a VPC by describing the virtual private gateway (any attachments to the virtual private gateway are also described).

You must wait for the attachment's state to switch to detached before you can delete the VPC or attach a different VPC to the virtual private gateway.

", - "DisableVgwRoutePropagation": "

Disables a virtual private gateway (VGW) from propagating routes to a specified route table of a VPC.

", - "DisableVpcClassicLink": "

Disables ClassicLink for a VPC. You cannot disable ClassicLink for a VPC that has EC2-Classic instances linked to it.

", - "DisableVpcClassicLinkDnsSupport": "

Disables ClassicLink DNS support for a VPC. If disabled, DNS hostnames resolve to public IP addresses when addressed between a linked EC2-Classic instance and instances in the VPC to which it's linked. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "DisassociateAddress": "

Disassociates an Elastic IP address from the instance or network interface it's associated with.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

", - "DisassociateRouteTable": "

Disassociates a subnet from a route table.

After you perform this action, the subnet no longer uses the routes in the route table. Instead, it uses the routes in the VPC's main route table. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "EnableVgwRoutePropagation": "

Enables a virtual private gateway (VGW) to propagate routes to the specified route table of a VPC.

", - "EnableVolumeIO": "

Enables I/O operations for a volume that had I/O operations disabled because the data on the volume was potentially inconsistent.

", - "EnableVpcClassicLink": "

Enables a VPC for ClassicLink. You can then link EC2-Classic instances to your ClassicLink-enabled VPC to allow communication over private IP addresses. You cannot enable your VPC for ClassicLink if any of your VPC's route tables have existing routes for address ranges within the 10.0.0.0/8 IP address range, excluding local routes for VPCs in the 10.0.0.0/16 and 10.1.0.0/16 IP address ranges. For more information, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "EnableVpcClassicLinkDnsSupport": "

Enables a VPC to support DNS hostname resolution for ClassicLink. If enabled, the DNS hostname of a linked EC2-Classic instance resolves to its private IP address when addressed from an instance in the VPC to which it's linked. Similarly, the DNS hostname of an instance in a VPC resolves to its private IP address when addressed from a linked EC2-Classic instance. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "GetConsoleOutput": "

Gets the console output for the specified instance.

Instances do not have a physical monitor through which you can view their console output. They also lack physical controls that allow you to power up, reboot, or shut them down. To allow these actions, we provide them through the Amazon EC2 API and command line interface.

Instance console output is buffered and posted shortly after instance boot, reboot, and termination. Amazon EC2 preserves the most recent 64 KB output which is available for at least one hour after the most recent post.

For Linux instances, the instance console output displays the exact console output that would normally be displayed on a physical monitor attached to a computer. This output is buffered because the instance produces it and then posts it to a store where the instance's owner can retrieve it.

For Windows instances, the instance console output includes output from the EC2Config service.

", - "GetConsoleScreenshot": "

Retrieve a JPG-format screenshot of a running instance to help with troubleshooting.

The returned content is base64-encoded.

", - "GetPasswordData": "

Retrieves the encrypted administrator password for an instance running Windows.

The Windows password is generated at boot if the EC2Config service plugin, Ec2SetPassword, is enabled. This usually only happens the first time an AMI is launched, and then Ec2SetPassword is automatically disabled. The password is not generated for rebundled AMIs unless Ec2SetPassword is enabled before bundling.

The password is encrypted using the key pair that you specified when you launched the instance. You must provide the corresponding key pair file.

Password generation and encryption takes a few moments. We recommend that you wait up to 15 minutes after launching an instance before trying to retrieve the generated password.

", - "ImportImage": "

Import single or multi-volume disk images or EBS snapshots into an Amazon Machine Image (AMI).

", - "ImportInstance": "

Creates an import instance task using metadata from the specified disk image. ImportInstance only supports single-volume VMs. To import multi-volume VMs, use ImportImage. After importing the image, you then upload it using the ec2-import-volume command in the EC2 command line tools. For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "ImportKeyPair": "

Imports the public key from an RSA key pair that you created with a third-party tool. Compare this with CreateKeyPair, in which AWS creates the key pair and gives the keys to you (AWS keeps a copy of the public key). With ImportKeyPair, you create the key pair and give AWS just the public key. The private key is never transferred between you and AWS.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "ImportSnapshot": "

Imports a disk into an EBS snapshot.

", - "ImportVolume": "

Creates an import volume task using metadata from the specified disk image. After importing the image, you then upload it using the ec2-import-volume command in the Amazon EC2 command-line interface (CLI) tools. For more information, see Using the Command Line Tools to Import Your Virtual Machine to Amazon EC2 in the Amazon Elastic Compute Cloud User Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "ModifyHosts": "

Modify the auto-placement setting of a Dedicated host. When auto-placement is enabled, AWS will place instances that you launch with a tenancy of host, but without targeting a specific host ID, onto any available Dedicated host in your account which has auto-placement enabled. When auto-placement is disabled, you need to provide a host ID if you want the instance to launch onto a specific host. If no host ID is provided, the instance will be launched onto a suitable host which has auto-placement enabled.

", - "ModifyIdFormat": "

Modifies the ID format for the specified resource on a per-region basis. You can specify that resources should receive longer IDs (17-character IDs) when they are created. The following resource types support longer IDs: instance | reservation | snapshot | volume.

This setting applies to the IAM user who makes the request; it does not apply to the entire AWS account. By default, an IAM user defaults to the same settings as the root user. If you're using this action as the root user or as an IAM role that has permission to use this action, then these settings apply to the entire account, unless an IAM user explicitly overrides these settings for themselves. For more information, see Controlling Access to Longer ID Settings in the Amazon Elastic Compute Cloud User Guide.

Resources created with longer IDs are visible to all IAM users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "ModifyImageAttribute": "

Modifies the specified attribute of the specified AMI. You can specify only one attribute at a time.

AWS Marketplace product codes cannot be modified. Images with an AWS Marketplace product code cannot be made public.

", - "ModifyInstanceAttribute": "

Modifies the specified attribute of the specified instance. You can specify only one attribute at a time.

To modify some attributes, the instance must be stopped. For more information, see Modifying Attributes of a Stopped Instance in the Amazon Elastic Compute Cloud User Guide.

", - "ModifyInstancePlacement": "

Set the instance affinity value for a specific stopped instance and modify the instance tenancy setting.

Instance affinity is disabled by default. When instance affinity is host and it is not associated with a specific Dedicated host, the next time it is launched it will automatically be associated with the host it lands on. This relationship will persist if the instance is stopped/started, or rebooted.

You can modify the host ID associated with a stopped instance. If a stopped instance has a new host ID association, the instance will target that host when restarted.

You can modify the tenancy of a stopped instance with a tenancy of host or dedicated.

Affinity, hostID, and tenancy are not required parameters, but at least one of them must be specified in the request. Affinity and tenancy can be modified in the same request, but tenancy can only be modified on instances that are stopped.

", - "ModifyNetworkInterfaceAttribute": "

Modifies the specified network interface attribute. You can specify only one attribute at a time.

", - "ModifyReservedInstances": "

Modifies the Availability Zone, instance count, instance type, or network platform (EC2-Classic or EC2-VPC) of your Reserved Instances. The Reserved Instances to be modified must be identical, except for Availability Zone, network platform, and instance type.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "ModifySnapshotAttribute": "

Adds or removes permission settings for the specified snapshot. You may add or remove specified AWS account IDs from a snapshot's list of create volume permissions, but you cannot do both in a single API call. If you need to both add and remove account IDs for a snapshot, you must use multiple API calls.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

Snapshots with AWS Marketplace product codes cannot be made public.

", - "ModifySpotFleetRequest": "

Modifies the specified Spot fleet request.

While the Spot fleet request is being modified, it is in the modifying state.

To scale up your Spot fleet, increase its target capacity. The Spot fleet launches the additional Spot instances according to the allocation strategy for the Spot fleet request. If the allocation strategy is lowestPrice, the Spot fleet launches instances using the Spot pool with the lowest price. If the allocation strategy is diversified, the Spot fleet distributes the instances across the Spot pools.

To scale down your Spot fleet, decrease its target capacity. First, the Spot fleet cancels any open bids that exceed the new target capacity. You can request that the Spot fleet terminate Spot instances until the size of the fleet no longer exceeds the new target capacity. If the allocation strategy is lowestPrice, the Spot fleet terminates the instances with the highest price per unit. If the allocation strategy is diversified, the Spot fleet terminates instances across the Spot pools. Alternatively, you can request that the Spot fleet keep the fleet at its current size, but not replace any Spot instances that are interrupted or that you terminate manually.

", - "ModifySubnetAttribute": "

Modifies a subnet attribute.

", - "ModifyVolumeAttribute": "

Modifies a volume attribute.

By default, all I/O operations for the volume are suspended when the data on the volume is determined to be potentially inconsistent, to prevent undetectable, latent data corruption. The I/O access to the volume can be resumed by first enabling I/O access and then checking the data consistency on your volume.

You can change the default behavior to resume I/O operations. We recommend that you change this only for boot volumes or for volumes that are stateless or disposable.

", - "ModifyVpcAttribute": "

Modifies the specified attribute of the specified VPC.

", - "ModifyVpcEndpoint": "

Modifies attributes of a specified VPC endpoint. You can modify the policy associated with the endpoint, and you can add and remove route tables associated with the endpoint.

", - "ModifyVpcPeeringConnectionOptions": "

Modifies the VPC peering connection options on one side of a VPC peering connection. You can do the following:

  • Enable/disable communication over the peering connection between an EC2-Classic instance that's linked to your VPC (using ClassicLink) and instances in the peer VPC.

  • Enable/disable communication over the peering connection between instances in your VPC and an EC2-Classic instance that's linked to the peer VPC.

If the peered VPCs are in different accounts, each owner must initiate a separate request to enable or disable communication in either direction, depending on whether their VPC was the requester or accepter for the VPC peering connection. If the peered VPCs are in the same account, you can modify the requester and accepter options in the same request. To confirm which VPC is the accepter and requester for a VPC peering connection, use the DescribeVpcPeeringConnections command.

", - "MonitorInstances": "

Enables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "MoveAddressToVpc": "

Moves an Elastic IP address from the EC2-Classic platform to the EC2-VPC platform. The Elastic IP address must be allocated to your account for more than 24 hours, and it must not be associated with an instance. After the Elastic IP address is moved, it is no longer available for use in the EC2-Classic platform, unless you move it back using the RestoreAddressToClassic request. You cannot move an Elastic IP address that was originally allocated for use in the EC2-VPC platform to the EC2-Classic platform.

", - "PurchaseReservedInstancesOffering": "

Purchases a Reserved Instance for use with your account. With Reserved Instances, you obtain a capacity reservation for a certain instance configuration over a specified period of time and pay a lower hourly rate compared to On-Demand instance pricing.

Use DescribeReservedInstancesOfferings to get a list of Reserved Instance offerings that match your specifications. After you've purchased a Reserved Instance, you can check for your new Reserved Instance with DescribeReservedInstances.

For more information, see Reserved Instances and Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "PurchaseScheduledInstances": "

Purchases one or more Scheduled Instances with the specified schedule.

Scheduled Instances enable you to purchase Amazon EC2 compute capacity by the hour for a one-year term. Before you can purchase a Scheduled Instance, you must call DescribeScheduledInstanceAvailability to check for available schedules and obtain a purchase token. After you purchase a Scheduled Instance, you must call RunScheduledInstances during each scheduled time period.

After you purchase a Scheduled Instance, you can't cancel, modify, or resell your purchase.

", - "RebootInstances": "

Requests a reboot of one or more instances. This operation is asynchronous; it only queues a request to reboot the specified instances. The operation succeeds if the instances are valid and belong to you. Requests to reboot terminated instances are ignored.

If an instance does not cleanly shut down within four minutes, Amazon EC2 performs a hard reboot.

For more information about troubleshooting, see Getting Console Output and Rebooting Instances in the Amazon Elastic Compute Cloud User Guide.

", - "RegisterImage": "

Registers an AMI. When you're creating an AMI, this is the final step you must complete before you can launch an instance from the AMI. For more information about creating AMIs, see Creating Your Own AMIs in the Amazon Elastic Compute Cloud User Guide.

For Amazon EBS-backed instances, CreateImage creates and registers the AMI in a single request, so you don't have to register the AMI yourself.

You can also use RegisterImage to create an Amazon EBS-backed Linux AMI from a snapshot of a root device volume. For more information, see Launching an Instance from a Snapshot in the Amazon Elastic Compute Cloud User Guide.

Some Linux distributions, such as Red Hat Enterprise Linux (RHEL) and SUSE Linux Enterprise Server (SLES), use the EC2 billingProduct code associated with an AMI to verify subscription status for package updates. Creating an AMI from an EBS snapshot does not maintain this billing code, and subsequent instances launched from such an AMI will not be able to connect to package update infrastructure.

Similarly, although you can create a Windows AMI from a snapshot, you can't successfully launch an instance from the AMI.

To create Windows AMIs or to create AMIs for Linux operating systems that must retain AMI billing codes to work properly, see CreateImage.

If needed, you can deregister an AMI at any time. Any modifications you make to an AMI backed by an instance store volume invalidates its registration. If you make changes to an image, deregister the previous image and register the new image.

You can't register an image where a secondary (non-root) snapshot has AWS Marketplace product codes.

", - "RejectVpcPeeringConnection": "

Rejects a VPC peering connection request. The VPC peering connection must be in the pending-acceptance state. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests. To delete an active VPC peering connection, or to delete a VPC peering connection request that you initiated, use DeleteVpcPeeringConnection.

", - "ReleaseAddress": "

Releases the specified Elastic IP address.

After releasing an Elastic IP address, it is released to the IP address pool and might be unavailable to you. Be sure to update your DNS records and any servers or devices that communicate with the address. If you attempt to release an Elastic IP address that you already released, you'll get an AuthFailure error if the address is already allocated to another AWS account.

[EC2-Classic, default VPC] Releasing an Elastic IP address automatically disassociates it from any instance that it's associated with. To disassociate an Elastic IP address without releasing it, use DisassociateAddress.

[Nondefault VPC] You must use DisassociateAddress to disassociate the Elastic IP address before you try to release it. Otherwise, Amazon EC2 returns an error (InvalidIPAddress.InUse).

", - "ReleaseHosts": "

When you no longer want to use a Dedicated host it can be released. On-Demand billing is stopped and the host goes into released state. The host ID of Dedicated hosts that have been released can no longer be specified in another request, e.g., ModifyHosts. You must stop or terminate all instances on a host before it can be released.

When Dedicated hosts are released, it make take some time for them to stop counting toward your limit and you may receive capacity errors when trying to allocate new Dedicated hosts. Try waiting a few minutes, and then try again.

Released hosts will still appear in a DescribeHosts response.

", - "ReplaceNetworkAclAssociation": "

Changes which network ACL a subnet is associated with. By default when you create a subnet, it's automatically associated with the default network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceNetworkAclEntry": "

Replaces an entry (rule) in a network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRoute": "

Replaces an existing route within a route table in a VPC. You must provide only one of the following: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, or network interface.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRouteTableAssociation": "

Changes the route table associated with a given subnet in a VPC. After the operation completes, the subnet uses the routes in the new route table it's associated with. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

You can also use ReplaceRouteTableAssociation to change which table is the main route table in the VPC. You just specify the main route table's association ID and the route table to be the new main route table.

", - "ReportInstanceStatus": "

Submits feedback about the status of an instance. The instance must be in the running state. If your experience with the instance differs from the instance status returned by DescribeInstanceStatus, use ReportInstanceStatus to report your experience with the instance. Amazon EC2 collects this information to improve the accuracy of status checks.

Use of this action does not change the value returned by DescribeInstanceStatus.

", - "RequestSpotFleet": "

Creates a Spot fleet request.

You can submit a single request that includes multiple launch specifications that vary by instance type, AMI, Availability Zone, or subnet.

By default, the Spot fleet requests Spot instances in the Spot pool where the price per unit is the lowest. Each launch specification can include its own instance weighting that reflects the value of the instance type to your application workload.

Alternatively, you can specify that the Spot fleet distribute the target capacity across the Spot pools included in its launch specifications. By ensuring that the Spot instances in your Spot fleet are in different Spot pools, you can improve the availability of your fleet.

For more information, see Spot Fleet Requests in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstances": "

Creates a Spot instance request. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot Instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

", - "ResetImageAttribute": "

Resets an attribute of an AMI to its default value.

The productCodes attribute can't be reset.

", - "ResetInstanceAttribute": "

Resets an attribute of an instance to its default value. To reset the kernel or ramdisk, the instance must be in a stopped state. To reset the sourceDestCheck, the instance can be either running or stopped.

The sourceDestCheck attribute controls whether source/destination checking is enabled. The default value is true, which means checking is enabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ResetNetworkInterfaceAttribute": "

Resets a network interface attribute. You can specify only one attribute at a time.

", - "ResetSnapshotAttribute": "

Resets permission settings for the specified snapshot.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "RestoreAddressToClassic": "

Restores an Elastic IP address that was previously moved to the EC2-VPC platform back to the EC2-Classic platform. You cannot move an Elastic IP address that was originally allocated for use in EC2-VPC. The Elastic IP address must not be associated with an instance or network interface.

", - "RevokeSecurityGroupEgress": "

[EC2-VPC only] Removes one or more egress rules from a security group for EC2-VPC. This action doesn't apply to security groups for use in EC2-Classic. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be revoked.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RevokeSecurityGroupIngress": "

Removes one or more ingress rules from a security group. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be removed.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RunInstances": "

Launches the specified number of instances using an AMI for which you have permissions.

When you launch an instance, it enters the pending state. After the instance is ready for you, it enters the running state. To check the state of your instance, call DescribeInstances.

To ensure faster instance launches, break up large requests into smaller batches. For example, create five separate launch requests for 100 instances each instead of one launch request for 500 instances.

To tag your instance, ensure that it is running as CreateTags requires a resource ID. For more information about tagging, see Tagging Your Amazon EC2 Resources.

If you don't specify a security group when launching an instance, Amazon EC2 uses the default security group. For more information, see Security Groups in the Amazon Elastic Compute Cloud User Guide.

[EC2-VPC only accounts] If you don't specify a subnet in the request, we choose a default subnet from your default VPC for you.

[EC2-Classic accounts] If you're launching into EC2-Classic and you don't specify an Availability Zone, we choose one for you.

Linux instances have access to the public key of the key pair at boot. You can use this key to provide secure access to the instance. Amazon EC2 public images use this feature to provide secure access without passwords. For more information, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

You can provide optional user data when launching an instance. For more information, see Instance Metadata in the Amazon Elastic Compute Cloud User Guide.

If any of the AMIs have a product code attached for which the user has not subscribed, RunInstances fails.

Some instance types can only be launched into a VPC. If you do not have a default VPC, or if you do not specify a subnet ID in the request, RunInstances fails. For more information, see Instance Types Available Only in a VPC.

For more information about troubleshooting, see What To Do If An Instance Immediately Terminates, and Troubleshooting Connecting to Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "RunScheduledInstances": "

Launches the specified Scheduled Instances.

Before you can launch a Scheduled Instance, you must purchase it and obtain an identifier using PurchaseScheduledInstances.

You must launch a Scheduled Instance during its scheduled time period. You can't stop or reboot a Scheduled Instance, but you can terminate it as needed. If you terminate a Scheduled Instance before the current scheduled time period ends, you can launch it again after a few minutes. For more information, see Scheduled Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StartInstances": "

Starts an Amazon EBS-backed AMI that you've previously stopped.

Instances that use Amazon EBS volumes as their root devices can be quickly stopped and started. When an instance is stopped, the compute resources are released and you are not billed for hourly instance usage. However, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. You can restart your instance at any time. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Performing this operation on an instance that uses an instance store as its root device returns an error.

For more information, see Stopping Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StopInstances": "

Stops an Amazon EBS-backed instance.

We don't charge hourly usage for a stopped instance, or data transfer fees; however, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

You can't start or stop Spot instances, and you can't stop instance store-backed instances.

When you stop an instance, we shut it down. You can restart your instance at any time. Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Stopping an instance is different to rebooting or terminating it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, the root device and any other devices attached during the instance launch are automatically deleted. For more information about the differences between rebooting, stopping, and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

When you stop an instance, we attempt to shut it down forcibly after a short while. If your instance appears stuck in the stopping state after a period of time, there may be an issue with the underlying host computer. For more information, see Troubleshooting Stopping Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "TerminateInstances": "

Shuts down one or more instances. This operation is idempotent; if you terminate an instance more than once, each call succeeds.

Terminated instances remain visible after termination (for approximately one hour).

By default, Amazon EC2 deletes all EBS volumes that were attached when the instance launched. Volumes attached after instance launch continue running.

You can stop, start, and terminate EBS-backed instances. You can only terminate instance store-backed instances. What happens to an instance differs if you stop it or terminate it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, any attached EBS volumes with the DeleteOnTermination block device mapping parameter set to true are automatically deleted. For more information about the differences between stopping and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

For more information about troubleshooting, see Troubleshooting Terminating Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "UnassignPrivateIpAddresses": "

Unassigns one or more secondary private IP addresses from a network interface.

", - "UnmonitorInstances": "

Disables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

" - }, - "shapes": { - "AcceptVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for AcceptVpcPeeringConnection.

", - "refs": { - } - }, - "AcceptVpcPeeringConnectionResult": { - "base": "

Contains the output of AcceptVpcPeeringConnection.

", - "refs": { - } - }, - "AccountAttribute": { - "base": "

Describes an account attribute.

", - "refs": { - "AccountAttributeList$member": null - } - }, - "AccountAttributeList": { - "base": null, - "refs": { - "DescribeAccountAttributesResult$AccountAttributes": "

Information about one or more account attributes.

" - } - }, - "AccountAttributeName": { - "base": null, - "refs": { - "AccountAttributeNameStringList$member": null - } - }, - "AccountAttributeNameStringList": { - "base": null, - "refs": { - "DescribeAccountAttributesRequest$AttributeNames": "

One or more account attribute names.

" - } - }, - "AccountAttributeValue": { - "base": "

Describes a value of an account attribute.

", - "refs": { - "AccountAttributeValueList$member": null - } - }, - "AccountAttributeValueList": { - "base": null, - "refs": { - "AccountAttribute$AttributeValues": "

One or more values for the account attribute.

" - } - }, - "ActiveInstance": { - "base": "

Describes a running instance in a Spot fleet.

", - "refs": { - "ActiveInstanceSet$member": null - } - }, - "ActiveInstanceSet": { - "base": null, - "refs": { - "DescribeSpotFleetInstancesResponse$ActiveInstances": "

The running instances. Note that this list is refreshed periodically and might be out of date.

" - } - }, - "Address": { - "base": "

Describes an Elastic IP address.

", - "refs": { - "AddressList$member": null - } - }, - "AddressList": { - "base": null, - "refs": { - "DescribeAddressesResult$Addresses": "

Information about one or more Elastic IP addresses.

" - } - }, - "Affinity": { - "base": null, - "refs": { - "ModifyInstancePlacementRequest$Affinity": "

The new affinity setting for the instance.

" - } - }, - "AllocateAddressRequest": { - "base": "

Contains the parameters for AllocateAddress.

", - "refs": { - } - }, - "AllocateAddressResult": { - "base": "

Contains the output of AllocateAddress.

", - "refs": { - } - }, - "AllocateHostsRequest": { - "base": "

Contains the parameters for AllocateHosts.

", - "refs": { - } - }, - "AllocateHostsResult": { - "base": "

Contains the output of AllocateHosts.

", - "refs": { - } - }, - "AllocationIdList": { - "base": null, - "refs": { - "DescribeAddressesRequest$AllocationIds": "

[EC2-VPC] One or more allocation IDs.

Default: Describes all your Elastic IP addresses.

" - } - }, - "AllocationState": { - "base": null, - "refs": { - "Host$State": "

The Dedicated host's state.

" - } - }, - "AllocationStrategy": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$AllocationStrategy": "

Indicates how to allocate the target capacity across the Spot pools specified by the Spot fleet request. The default is lowestPrice.

" - } - }, - "ArchitectureValues": { - "base": null, - "refs": { - "Image$Architecture": "

The architecture of the image.

", - "ImportInstanceLaunchSpecification$Architecture": "

The architecture of the instance.

", - "Instance$Architecture": "

The architecture of the image.

", - "RegisterImageRequest$Architecture": "

The architecture of the AMI.

Default: For Amazon EBS-backed AMIs, i386. For instance store-backed AMIs, the architecture specified in the manifest file.

" - } - }, - "AssignPrivateIpAddressesRequest": { - "base": "

Contains the parameters for AssignPrivateIpAddresses.

", - "refs": { - } - }, - "AssociateAddressRequest": { - "base": "

Contains the parameters for AssociateAddress.

", - "refs": { - } - }, - "AssociateAddressResult": { - "base": "

Contains the output of AssociateAddress.

", - "refs": { - } - }, - "AssociateDhcpOptionsRequest": { - "base": "

Contains the parameters for AssociateDhcpOptions.

", - "refs": { - } - }, - "AssociateRouteTableRequest": { - "base": "

Contains the parameters for AssociateRouteTable.

", - "refs": { - } - }, - "AssociateRouteTableResult": { - "base": "

Contains the output of AssociateRouteTable.

", - "refs": { - } - }, - "AttachClassicLinkVpcRequest": { - "base": "

Contains the parameters for AttachClassicLinkVpc.

", - "refs": { - } - }, - "AttachClassicLinkVpcResult": { - "base": "

Contains the output of AttachClassicLinkVpc.

", - "refs": { - } - }, - "AttachInternetGatewayRequest": { - "base": "

Contains the parameters for AttachInternetGateway.

", - "refs": { - } - }, - "AttachNetworkInterfaceRequest": { - "base": "

Contains the parameters for AttachNetworkInterface.

", - "refs": { - } - }, - "AttachNetworkInterfaceResult": { - "base": "

Contains the output of AttachNetworkInterface.

", - "refs": { - } - }, - "AttachVolumeRequest": { - "base": "

Contains the parameters for AttachVolume.

", - "refs": { - } - }, - "AttachVpnGatewayRequest": { - "base": "

Contains the parameters for AttachVpnGateway.

", - "refs": { - } - }, - "AttachVpnGatewayResult": { - "base": "

Contains the output of AttachVpnGateway.

", - "refs": { - } - }, - "AttachmentStatus": { - "base": null, - "refs": { - "EbsInstanceBlockDevice$Status": "

The attachment state.

", - "InstanceNetworkInterfaceAttachment$Status": "

The attachment state.

", - "InternetGatewayAttachment$State": "

The current state of the attachment.

", - "NetworkInterfaceAttachment$Status": "

The attachment state.

", - "VpcAttachment$State": "

The current state of the attachment.

" - } - }, - "AttributeBooleanValue": { - "base": "

The value to use when a resource attribute accepts a Boolean value.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$SourceDestCheck": "

Indicates whether source/destination checking is enabled.

", - "DescribeVolumeAttributeResult$AutoEnableIO": "

The state of autoEnableIO attribute.

", - "DescribeVpcAttributeResult$EnableDnsSupport": "

Indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for your instances to their corresponding IP addresses; otherwise, it does not.

", - "DescribeVpcAttributeResult$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.

", - "InstanceAttribute$DisableApiTermination": "

If the value is true, you can't terminate the instance through the Amazon EC2 console, CLI, or API; otherwise, you can.

", - "InstanceAttribute$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O.

", - "InstanceAttribute$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$SourceDestCheck": "

Specifies whether source/destination checking is enabled. A value of true means that checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$DisableApiTermination": "

If the value is true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. You cannot use this paramater for Spot Instances.

", - "ModifyInstanceAttributeRequest$EbsOptimized": "

Specifies whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "ModifyNetworkInterfaceAttributeRequest$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ModifySubnetAttributeRequest$MapPublicIpOnLaunch": "

Specify true to indicate that instances launched into the specified subnet should be assigned public IP address.

", - "ModifyVolumeAttributeRequest$AutoEnableIO": "

Indicates whether the volume should be auto-enabled for I/O operations.

", - "ModifyVpcAttributeRequest$EnableDnsSupport": "

Indicates whether the DNS resolution is supported for the VPC. If enabled, queries to the Amazon provided DNS server at the 169.254.169.253 IP address, or the reserved IP address at the base of the VPC network range \"plus two\" will succeed. If disabled, the Amazon provided DNS service in the VPC that resolves public DNS hostnames to IP addresses is not enabled.

You cannot modify the DNS resolution and DNS hostnames attributes in the same request. Use separate requests for each attribute.

", - "ModifyVpcAttributeRequest$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If enabled, instances in the VPC get DNS hostnames; otherwise, they do not.

You cannot modify the DNS resolution and DNS hostnames attributes in the same request. Use separate requests for each attribute. You can only enable DNS hostnames if you've enabled DNS support.

" - } - }, - "AttributeValue": { - "base": "

The value to use for a resource attribute.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Description": "

The description of the network interface.

", - "DhcpConfigurationValueList$member": null, - "ImageAttribute$KernelId": "

The kernel ID.

", - "ImageAttribute$RamdiskId": "

The RAM disk ID.

", - "ImageAttribute$Description": "

A description for the AMI.

", - "ImageAttribute$SriovNetSupport": null, - "InstanceAttribute$InstanceType": "

The instance type.

", - "InstanceAttribute$KernelId": "

The kernel ID.

", - "InstanceAttribute$RamdiskId": "

The RAM disk ID.

", - "InstanceAttribute$UserData": "

The Base64-encoded MIME user data.

", - "InstanceAttribute$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "InstanceAttribute$RootDeviceName": "

The name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "InstanceAttribute$SriovNetSupport": null, - "ModifyImageAttributeRequest$Description": "

A description for the AMI.

", - "ModifyInstanceAttributeRequest$InstanceType": "

Changes the instance type to the specified value. For more information, see Instance Types. If the instance type is not valid, the error returned is InvalidInstanceAttributeValue.

", - "ModifyInstanceAttributeRequest$Kernel": "

Changes the instance's kernel to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$Ramdisk": "

Changes the instance's RAM disk to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$InstanceInitiatedShutdownBehavior": "

Specifies whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "ModifyInstanceAttributeRequest$SriovNetSupport": "

Set to simple to enable enhanced networking for the instance.

There is no way to disable enhanced networking at this time.

This option is supported only for HVM instances. Specifying this option with a PV instance can make it unreachable.

", - "ModifyNetworkInterfaceAttributeRequest$Description": "

A description for the network interface.

" - } - }, - "AuthorizeSecurityGroupEgressRequest": { - "base": "

Contains the parameters for AuthorizeSecurityGroupEgress.

", - "refs": { - } - }, - "AuthorizeSecurityGroupIngressRequest": { - "base": "

Contains the parameters for AuthorizeSecurityGroupIngress.

", - "refs": { - } - }, - "AutoPlacement": { - "base": null, - "refs": { - "AllocateHostsRequest$AutoPlacement": "

This is enabled by default. This property allows instances to be automatically placed onto available Dedicated hosts, when you are launching instances without specifying a host ID.

Default: Enabled

", - "Host$AutoPlacement": "

Whether auto-placement is on or off.

", - "ModifyHostsRequest$AutoPlacement": "

Specify whether to enable or disable auto-placement.

" - } - }, - "AvailabilityZone": { - "base": "

Describes an Availability Zone.

", - "refs": { - "AvailabilityZoneList$member": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesResult$AvailabilityZones": "

Information about one or more Availability Zones.

" - } - }, - "AvailabilityZoneMessage": { - "base": "

Describes a message about an Availability Zone.

", - "refs": { - "AvailabilityZoneMessageList$member": null - } - }, - "AvailabilityZoneMessageList": { - "base": null, - "refs": { - "AvailabilityZone$Messages": "

Any messages about the Availability Zone.

" - } - }, - "AvailabilityZoneState": { - "base": null, - "refs": { - "AvailabilityZone$State": "

The state of the Availability Zone.

" - } - }, - "AvailableCapacity": { - "base": "

The capacity information for instances launched onto the Dedicated host.

", - "refs": { - "Host$AvailableCapacity": "

The number of new instances that can be launched onto the Dedicated host.

" - } - }, - "AvailableInstanceCapacityList": { - "base": null, - "refs": { - "AvailableCapacity$AvailableInstanceCapacity": "

The total number of instances that the Dedicated host supports.

" - } - }, - "BatchState": { - "base": null, - "refs": { - "CancelSpotFleetRequestsSuccessItem$CurrentSpotFleetRequestState": "

The current state of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$PreviousSpotFleetRequestState": "

The previous state of the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestState": "

The state of the Spot fleet request.

" - } - }, - "Blob": { - "base": null, - "refs": { - "BlobAttributeValue$Value": null, - "ImportKeyPairRequest$PublicKeyMaterial": "

The public key. For API calls, the text must be base64-encoded. For command line tools, base64 encoding is performed for you.

", - "S3Storage$UploadPolicy": "

A base64-encoded Amazon S3 upload policy that gives Amazon EC2 permission to upload items into Amazon S3 on your behalf. For command line tools, base64 encoding is performed for you.

" - } - }, - "BlobAttributeValue": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$UserData": "

Changes the instance's user data to the specified base64-encoded value. For command line tools, base64 encoding is performed for you.

" - } - }, - "BlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "BlockDeviceMappingList$member": null, - "BlockDeviceMappingRequestList$member": null - } - }, - "BlockDeviceMappingList": { - "base": null, - "refs": { - "Image$BlockDeviceMappings": "

Any block device mapping entries.

", - "ImageAttribute$BlockDeviceMappings": "

One or more block device mapping entries.

", - "LaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

Although you can specify encrypted EBS volumes in this block device mapping for your Spot Instances, these volumes are not encrypted.

", - "RequestSpotLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

Although you can specify encrypted EBS volumes in this block device mapping for your Spot Instances, these volumes are not encrypted.

", - "SpotFleetLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "BlockDeviceMappingRequestList": { - "base": null, - "refs": { - "CreateImageRequest$BlockDeviceMappings": "

Information about one or more block device mappings.

", - "RegisterImageRequest$BlockDeviceMappings": "

One or more block device mapping entries.

", - "RunInstancesRequest$BlockDeviceMappings": "

The block device mapping.

" - } - }, - "Boolean": { - "base": null, - "refs": { - "AcceptVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AllocateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssignPrivateIpAddressesRequest$AllowReassignment": "

Indicates whether to allow an IP address that is already assigned to another network interface or instance to be reassigned to the specified network interface.

", - "AssociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateAddressRequest$AllowReassociation": "

[EC2-VPC] For a VPC in an EC2-Classic account, specify true to allow an Elastic IP address that is already associated with an instance or network interface to be reassociated with the specified instance or network interface. Otherwise, the operation fails. In a VPC in an EC2-VPC-only account, reassociation is automatic, therefore you can specify false to ensure the operation fails if the Elastic IP address is already associated with another resource.

", - "AssociateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "AttachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttributeBooleanValue$Value": "

Valid values are true or false.

", - "AuthorizeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AuthorizeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "BundleInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelBundleTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelConversionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelImportTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$TerminateInstances": "

Indicates whether to terminate instances for a Spot fleet request if it is canceled successfully.

", - "CancelSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ClassicLinkDnsSupport$ClassicLinkDnsSupported": "

Indicates whether ClassicLink DNS support is enabled for the VPC.

", - "ConfirmProductInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ConfirmProductInstanceResult$Return": "

The return value of the request. Returns true if the specified product code is owned by the requester and associated with the specified instance.

", - "CopyImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopyImageRequest$Encrypted": "

Specifies whether the destination snapshots of the copied image should be encrypted. The default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopySnapshotRequest$Encrypted": "

Specifies whether the destination snapshot should be encrypted. There is no way to create an unencrypted snapshot copy from an encrypted snapshot; however, you can encrypt a copy of an unencrypted snapshot with this flag. The default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$NoReboot": "

By default, Amazon EC2 attempts to shut down and reboot the instance before creating the image. If the 'No Reboot' option is set, Amazon EC2 doesn't shut down the instance before creating the image. When this option is used, file system integrity on the created image can't be guaranteed.

", - "CreateInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$Egress": "

Indicates whether this is an egress rule (rule is applied to traffic leaving the subnet).

", - "CreateNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreatePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "CreateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$Encrypted": "

Specifies whether the volume should be encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are automatically encrypted. There is no way to create an encrypted volume from an unencrypted snapshot or vice versa. If your AMI uses encrypted volumes, you can only launch it on supported instance types. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$Egress": "

Indicates whether the rule is an egress rule.

", - "DeleteNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeletePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DeleteVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeregisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAccountAttributesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAvailabilityZonesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeBundleTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeClassicLinkInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeConversionTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeCustomerGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImagesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportImageTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportSnapshotTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$IncludeAllInstances": "

When true, includes the health status for all instances. When false, includes the health status for running instances only.

Default: false

", - "DescribeInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInternetGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeKeyPairsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeMovingAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkAclsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfacesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePlacementGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePrefixListsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRegionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$IncludeMarketplace": "

Include Reserved Instance Marketplace offerings in the response.

", - "DescribeReservedInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRouteTablesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeScheduledInstanceAvailabilityRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupReferencesRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotPriceHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeStaleSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSubnetsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointServicesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcPeeringConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DetachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$Force": "

Specifies whether to force a detachment.

", - "DetachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachVolumeRequest$Force": "

Forces detachment if the previous detachment attempt did not occur cleanly (for example, logging into an instance, unmounting the volume, and detaching normally). This option can lead to data loss or a corrupted file system. Use this option only as a last resort to detach a volume from a failed instance. The instance won't have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures.

", - "DetachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkDnsSupportResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisassociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisassociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EbsBlockDevice$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "EbsBlockDevice$Encrypted": "

Indicates whether the EBS volume is encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption.

", - "EbsInstanceBlockDevice$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EbsInstanceBlockDeviceSpecification$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EnableVolumeIORequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkDnsSupportResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "EnableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "GetConsoleOutputRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetConsoleScreenshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetConsoleScreenshotRequest$WakeUp": "

When set to true, acts as keystroke input and wakes up an instance that's in standby or \"sleep\" mode.

", - "GetPasswordDataRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "IdFormat$UseLongIds": "

Indicates whether longer IDs (17-character IDs) are enabled for the resource.

", - "Image$Public": "

Indicates whether the image has public launch permissions. The value is true if this image has public launch permissions or false if it has only implicit and explicit launch permissions.

", - "ImportImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportInstanceLaunchSpecification$Monitoring": "

Indicates whether monitoring is enabled.

", - "ImportInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Instance$SourceDestCheck": "

Specifies whether to enable an instance launched in a VPC to perform NAT. This controls whether source/destination checking is enabled on the instance. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "Instance$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "InstanceNetworkInterface$SourceDestCheck": "

Indicates whether to validate network traffic to or from this network interface.

", - "InstanceNetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "InstanceNetworkInterfaceSpecification$DeleteOnTermination": "

If set to true, the interface is deleted when the instance is terminated. You can specify true only if creating a new network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to an instance you launch in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "InstancePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "LaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ModifyIdFormatRequest$UseLongIds": "

Indicate whether the resource should use longer IDs (17-character IDs).

", - "ModifyImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstancePlacementResult$Return": "

Is true if the request succeeds, and an error otherwise.

", - "ModifyNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySpotFleetRequestResponse$Return": "

Is true if the request succeeds, and an error otherwise.

", - "ModifyVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$ResetPolicy": "

Specify true to reset the policy document to the default policy. The default policy allows access to the service.

", - "ModifyVpcEndpointResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ModifyVpcPeeringConnectionOptionsRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MoveAddressToVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "NetworkAcl$IsDefault": "

Indicates whether this is the default network ACL for the VPC.

", - "NetworkAclEntry$Egress": "

Indicates whether the rule is an egress rule (applied to traffic leaving the subnet).

", - "NetworkInterface$RequesterManaged": "

Indicates whether the network interface is being managed by AWS.

", - "NetworkInterface$SourceDestCheck": "

Indicates whether traffic to or from the instance is validated.

", - "NetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfaceAttachmentChanges$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfacePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "PeeringConnectionOptions$AllowEgressFromLocalClassicLinkToRemoteVpc": "

If true, enables outbound communication from an EC2-Classic instance that's linked to a local VPC via ClassicLink to instances in a peer VPC.

", - "PeeringConnectionOptions$AllowEgressFromLocalVpcToRemoteClassicLink": "

If true, enables outbound communication from instances in a local VPC to an EC2-Classic instance that's linked to a peer VPC via ClassicLink.

", - "PeeringConnectionOptionsRequest$AllowEgressFromLocalClassicLinkToRemoteVpc": "

If true, enables outbound communication from an EC2-Classic instance that's linked to a local VPC via ClassicLink to instances in a peer VPC.

", - "PeeringConnectionOptionsRequest$AllowEgressFromLocalVpcToRemoteClassicLink": "

If true, enables outbound communication from instances in a local VPC to an EC2-Classic instance that's linked to a peer VPC via ClassicLink.

", - "PriceSchedule$Active": "

The current price schedule, as determined by the term remaining for the Reserved Instance in the listing.

A specific price schedule is always in effect, but only one price schedule can be active at any time. Take, for example, a Reserved Instance listing that has five months remaining in its term. When you specify price schedules for five months and two months, this means that schedule 1, covering the first three months of the remaining term, will be active during months 5, 4, and 3. Then schedule 2, covering the last two months of the term, will be active for months 2 and 1.

", - "PrivateIpAddressSpecification$Primary": "

Indicates whether the private IP address is the primary private IP address. Only one IP address can be designated as primary.

", - "PurchaseReservedInstancesOfferingRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "PurchaseScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RebootInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RegisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RejectVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RejectVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ReleaseAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$Egress": "

Indicates whether to replace the egress rule.

Default: If no value is specified, we replace the ingress rule.

", - "ReplaceRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceRouteTableAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReportInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotFleetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotLaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ReservedInstancesOffering$Marketplace": "

Indicates whether the offering is available through the Reserved Instance Marketplace (resale) or AWS. If it's a Reserved Instance Marketplace offering, this is true.

", - "ResetImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RestoreAddressToClassicRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RouteTableAssociation$Main": "

Indicates whether this is the main route table.

", - "RunInstancesMonitoringEnabled$Enabled": "

Indicates whether monitoring is enabled for the instance.

", - "RunInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RunInstancesRequest$DisableApiTermination": "

If you set this parameter to true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. If you set this parameter to true and then later want to be able to terminate the instance, you must first change the value of the disableApiTermination attribute to false using ModifyInstanceAttribute. Alternatively, if you set InstanceInitiatedShutdownBehavior to terminate, you can terminate the instance by running the shutdown command from the instance.

Default: false

", - "RunInstancesRequest$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "RunScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ScheduledInstanceRecurrence$OccurrenceRelativeToEnd": "

Indicates whether the occurrence is relative to the end of the specified week or month.

", - "ScheduledInstanceRecurrenceRequest$OccurrenceRelativeToEnd": "

Indicates whether the occurrence is relative to the end of the specified week or month. You can't specify this value with a daily schedule.

", - "ScheduledInstancesEbs$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "ScheduledInstancesEbs$Encrypted": "

Indicates whether the volume is encrypted. You can attached encrypted volumes only to instances that support them.

", - "ScheduledInstancesLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "ScheduledInstancesMonitoring$Enabled": "

Indicates whether monitoring is enabled.

", - "ScheduledInstancesNetworkInterface$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to instances launched in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "ScheduledInstancesNetworkInterface$DeleteOnTermination": "

Indicates whether to delete the interface when the instance is terminated.

", - "ScheduledInstancesPrivateIpAddressConfig$Primary": "

Indicates whether this is a primary IP address. Otherwise, this is a secondary IP address.

", - "Snapshot$Encrypted": "

Indicates whether the snapshot is encrypted.

", - "SpotFleetLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "SpotFleetMonitoring$Enabled": "

Enables monitoring for the instance.

Default: false

", - "SpotFleetRequestConfigData$TerminateInstancesWithExpiration": "

Indicates whether running Spot instances should be terminated when the Spot fleet request expires.

", - "StartInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$Force": "

Forces the instances to stop. The instances do not have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures. This option is not recommended for Windows instances.

Default: false

", - "Subnet$DefaultForAz": "

Indicates whether this is the default subnet for the Availability Zone.

", - "Subnet$MapPublicIpOnLaunch": "

Indicates whether instances launched in this subnet receive a public IP address.

", - "TerminateInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "UnmonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Volume$Encrypted": "

Indicates whether the volume will be encrypted.

", - "VolumeAttachment$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "Vpc$IsDefault": "

Indicates whether the VPC is the default VPC.

", - "VpcClassicLink$ClassicLinkEnabled": "

Indicates whether the VPC is enabled for ClassicLink.

", - "VpcPeeringConnectionOptionsDescription$AllowEgressFromLocalClassicLinkToRemoteVpc": "

Indicates whether a local ClassicLink connection can communicate with the peer VPC over the VPC peering connection.

", - "VpcPeeringConnectionOptionsDescription$AllowEgressFromLocalVpcToRemoteClassicLink": "

Indicates whether a local VPC can communicate with a ClassicLink connection in the peer VPC over the VPC peering connection.

", - "VpnConnectionOptions$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

", - "VpnConnectionOptionsSpecification$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

" - } - }, - "BundleIdStringList": { - "base": null, - "refs": { - "DescribeBundleTasksRequest$BundleIds": "

One or more bundle task IDs.

Default: Describes all your bundle tasks.

" - } - }, - "BundleInstanceRequest": { - "base": "

Contains the parameters for BundleInstance.

", - "refs": { - } - }, - "BundleInstanceResult": { - "base": "

Contains the output of BundleInstance.

", - "refs": { - } - }, - "BundleTask": { - "base": "

Describes a bundle task.

", - "refs": { - "BundleInstanceResult$BundleTask": "

Information about the bundle task.

", - "BundleTaskList$member": null, - "CancelBundleTaskResult$BundleTask": "

Information about the bundle task.

" - } - }, - "BundleTaskError": { - "base": "

Describes an error for BundleInstance.

", - "refs": { - "BundleTask$BundleTaskError": "

If the task fails, a description of the error.

" - } - }, - "BundleTaskList": { - "base": null, - "refs": { - "DescribeBundleTasksResult$BundleTasks": "

Information about one or more bundle tasks.

" - } - }, - "BundleTaskState": { - "base": null, - "refs": { - "BundleTask$State": "

The state of the task.

" - } - }, - "CancelBatchErrorCode": { - "base": null, - "refs": { - "CancelSpotFleetRequestsError$Code": "

The error code.

" - } - }, - "CancelBundleTaskRequest": { - "base": "

Contains the parameters for CancelBundleTask.

", - "refs": { - } - }, - "CancelBundleTaskResult": { - "base": "

Contains the output of CancelBundleTask.

", - "refs": { - } - }, - "CancelConversionRequest": { - "base": "

Contains the parameters for CancelConversionTask.

", - "refs": { - } - }, - "CancelExportTaskRequest": { - "base": "

Contains the parameters for CancelExportTask.

", - "refs": { - } - }, - "CancelImportTaskRequest": { - "base": "

Contains the parameters for CancelImportTask.

", - "refs": { - } - }, - "CancelImportTaskResult": { - "base": "

Contains the output for CancelImportTask.

", - "refs": { - } - }, - "CancelReservedInstancesListingRequest": { - "base": "

Contains the parameters for CancelReservedInstancesListing.

", - "refs": { - } - }, - "CancelReservedInstancesListingResult": { - "base": "

Contains the output of CancelReservedInstancesListing.

", - "refs": { - } - }, - "CancelSpotFleetRequestsError": { - "base": "

Describes a Spot fleet error.

", - "refs": { - "CancelSpotFleetRequestsErrorItem$Error": "

The error.

" - } - }, - "CancelSpotFleetRequestsErrorItem": { - "base": "

Describes a Spot fleet request that was not successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsErrorSet$member": null - } - }, - "CancelSpotFleetRequestsErrorSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$UnsuccessfulFleetRequests": "

Information about the Spot fleet requests that are not successfully canceled.

" - } - }, - "CancelSpotFleetRequestsRequest": { - "base": "

Contains the parameters for CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsResponse": { - "base": "

Contains the output of CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsSuccessItem": { - "base": "

Describes a Spot fleet request that was successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsSuccessSet$member": null - } - }, - "CancelSpotFleetRequestsSuccessSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$SuccessfulFleetRequests": "

Information about the Spot fleet requests that are successfully canceled.

" - } - }, - "CancelSpotInstanceRequestState": { - "base": null, - "refs": { - "CancelledSpotInstanceRequest$State": "

The state of the Spot instance request.

" - } - }, - "CancelSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelSpotInstanceRequestsResult": { - "base": "

Contains the output of CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelledSpotInstanceRequest": { - "base": "

Describes a request to cancel a Spot instance.

", - "refs": { - "CancelledSpotInstanceRequestList$member": null - } - }, - "CancelledSpotInstanceRequestList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsResult$CancelledSpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "ClassicLinkDnsSupport": { - "base": "

Describes the ClassicLink DNS support status of a VPC.

", - "refs": { - "ClassicLinkDnsSupportList$member": null - } - }, - "ClassicLinkDnsSupportList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkDnsSupportResult$Vpcs": "

Information about the ClassicLink DNS support status of the VPCs.

" - } - }, - "ClassicLinkInstance": { - "base": "

Describes a linked EC2-Classic instance.

", - "refs": { - "ClassicLinkInstanceList$member": null - } - }, - "ClassicLinkInstanceList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesResult$Instances": "

Information about one or more linked EC2-Classic instances.

" - } - }, - "ClientData": { - "base": "

Describes the client-specific data.

", - "refs": { - "ImportImageRequest$ClientData": "

The client-specific data.

", - "ImportSnapshotRequest$ClientData": "

The client-specific data.

" - } - }, - "ConfirmProductInstanceRequest": { - "base": "

Contains the parameters for ConfirmProductInstance.

", - "refs": { - } - }, - "ConfirmProductInstanceResult": { - "base": "

Contains the output of ConfirmProductInstance.

", - "refs": { - } - }, - "ContainerFormat": { - "base": null, - "refs": { - "ExportToS3Task$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

", - "ExportToS3TaskSpecification$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

" - } - }, - "ConversionIdStringList": { - "base": null, - "refs": { - "DescribeConversionTasksRequest$ConversionTaskIds": "

One or more conversion task IDs.

" - } - }, - "ConversionTask": { - "base": "

Describes a conversion task.

", - "refs": { - "DescribeConversionTaskList$member": null, - "ImportInstanceResult$ConversionTask": "

Information about the conversion task.

", - "ImportVolumeResult$ConversionTask": "

Information about the conversion task.

" - } - }, - "ConversionTaskState": { - "base": null, - "refs": { - "ConversionTask$State": "

The state of the conversion task.

" - } - }, - "CopyImageRequest": { - "base": "

Contains the parameters for CopyImage.

", - "refs": { - } - }, - "CopyImageResult": { - "base": "

Contains the output of CopyImage.

", - "refs": { - } - }, - "CopySnapshotRequest": { - "base": "

Contains the parameters for CopySnapshot.

", - "refs": { - } - }, - "CopySnapshotResult": { - "base": "

Contains the output of CopySnapshot.

", - "refs": { - } - }, - "CreateCustomerGatewayRequest": { - "base": "

Contains the parameters for CreateCustomerGateway.

", - "refs": { - } - }, - "CreateCustomerGatewayResult": { - "base": "

Contains the output of CreateCustomerGateway.

", - "refs": { - } - }, - "CreateDhcpOptionsRequest": { - "base": "

Contains the parameters for CreateDhcpOptions.

", - "refs": { - } - }, - "CreateDhcpOptionsResult": { - "base": "

Contains the output of CreateDhcpOptions.

", - "refs": { - } - }, - "CreateFlowLogsRequest": { - "base": "

Contains the parameters for CreateFlowLogs.

", - "refs": { - } - }, - "CreateFlowLogsResult": { - "base": "

Contains the output of CreateFlowLogs.

", - "refs": { - } - }, - "CreateImageRequest": { - "base": "

Contains the parameters for CreateImage.

", - "refs": { - } - }, - "CreateImageResult": { - "base": "

Contains the output of CreateImage.

", - "refs": { - } - }, - "CreateInstanceExportTaskRequest": { - "base": "

Contains the parameters for CreateInstanceExportTask.

", - "refs": { - } - }, - "CreateInstanceExportTaskResult": { - "base": "

Contains the output for CreateInstanceExportTask.

", - "refs": { - } - }, - "CreateInternetGatewayRequest": { - "base": "

Contains the parameters for CreateInternetGateway.

", - "refs": { - } - }, - "CreateInternetGatewayResult": { - "base": "

Contains the output of CreateInternetGateway.

", - "refs": { - } - }, - "CreateKeyPairRequest": { - "base": "

Contains the parameters for CreateKeyPair.

", - "refs": { - } - }, - "CreateNatGatewayRequest": { - "base": "

Contains the parameters for CreateNatGateway.

", - "refs": { - } - }, - "CreateNatGatewayResult": { - "base": "

Contains the output of CreateNatGateway.

", - "refs": { - } - }, - "CreateNetworkAclEntryRequest": { - "base": "

Contains the parameters for CreateNetworkAclEntry.

", - "refs": { - } - }, - "CreateNetworkAclRequest": { - "base": "

Contains the parameters for CreateNetworkAcl.

", - "refs": { - } - }, - "CreateNetworkAclResult": { - "base": "

Contains the output of CreateNetworkAcl.

", - "refs": { - } - }, - "CreateNetworkInterfaceRequest": { - "base": "

Contains the parameters for CreateNetworkInterface.

", - "refs": { - } - }, - "CreateNetworkInterfaceResult": { - "base": "

Contains the output of CreateNetworkInterface.

", - "refs": { - } - }, - "CreatePlacementGroupRequest": { - "base": "

Contains the parameters for CreatePlacementGroup.

", - "refs": { - } - }, - "CreateReservedInstancesListingRequest": { - "base": "

Contains the parameters for CreateReservedInstancesListing.

", - "refs": { - } - }, - "CreateReservedInstancesListingResult": { - "base": "

Contains the output of CreateReservedInstancesListing.

", - "refs": { - } - }, - "CreateRouteRequest": { - "base": "

Contains the parameters for CreateRoute.

", - "refs": { - } - }, - "CreateRouteResult": { - "base": "

Contains the output of CreateRoute.

", - "refs": { - } - }, - "CreateRouteTableRequest": { - "base": "

Contains the parameters for CreateRouteTable.

", - "refs": { - } - }, - "CreateRouteTableResult": { - "base": "

Contains the output of CreateRouteTable.

", - "refs": { - } - }, - "CreateSecurityGroupRequest": { - "base": "

Contains the parameters for CreateSecurityGroup.

", - "refs": { - } - }, - "CreateSecurityGroupResult": { - "base": "

Contains the output of CreateSecurityGroup.

", - "refs": { - } - }, - "CreateSnapshotRequest": { - "base": "

Contains the parameters for CreateSnapshot.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSubnetRequest": { - "base": "

Contains the parameters for CreateSubnet.

", - "refs": { - } - }, - "CreateSubnetResult": { - "base": "

Contains the output of CreateSubnet.

", - "refs": { - } - }, - "CreateTagsRequest": { - "base": "

Contains the parameters for CreateTags.

", - "refs": { - } - }, - "CreateVolumePermission": { - "base": "

Describes the user or group to be added or removed from the permissions for a volume.

", - "refs": { - "CreateVolumePermissionList$member": null - } - }, - "CreateVolumePermissionList": { - "base": null, - "refs": { - "CreateVolumePermissionModifications$Add": "

Adds a specific AWS account ID or group to a volume's list of create volume permissions.

", - "CreateVolumePermissionModifications$Remove": "

Removes a specific AWS account ID or group from a volume's list of create volume permissions.

", - "DescribeSnapshotAttributeResult$CreateVolumePermissions": "

A list of permissions for creating volumes from the snapshot.

" - } - }, - "CreateVolumePermissionModifications": { - "base": "

Describes modifications to the permissions for a volume.

", - "refs": { - "ModifySnapshotAttributeRequest$CreateVolumePermission": "

A JSON representation of the snapshot attribute modification.

" - } - }, - "CreateVolumeRequest": { - "base": "

Contains the parameters for CreateVolume.

", - "refs": { - } - }, - "CreateVpcEndpointRequest": { - "base": "

Contains the parameters for CreateVpcEndpoint.

", - "refs": { - } - }, - "CreateVpcEndpointResult": { - "base": "

Contains the output of CreateVpcEndpoint.

", - "refs": { - } - }, - "CreateVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for CreateVpcPeeringConnection.

", - "refs": { - } - }, - "CreateVpcPeeringConnectionResult": { - "base": "

Contains the output of CreateVpcPeeringConnection.

", - "refs": { - } - }, - "CreateVpcRequest": { - "base": "

Contains the parameters for CreateVpc.

", - "refs": { - } - }, - "CreateVpcResult": { - "base": "

Contains the output of CreateVpc.

", - "refs": { - } - }, - "CreateVpnConnectionRequest": { - "base": "

Contains the parameters for CreateVpnConnection.

", - "refs": { - } - }, - "CreateVpnConnectionResult": { - "base": "

Contains the output of CreateVpnConnection.

", - "refs": { - } - }, - "CreateVpnConnectionRouteRequest": { - "base": "

Contains the parameters for CreateVpnConnectionRoute.

", - "refs": { - } - }, - "CreateVpnGatewayRequest": { - "base": "

Contains the parameters for CreateVpnGateway.

", - "refs": { - } - }, - "CreateVpnGatewayResult": { - "base": "

Contains the output of CreateVpnGateway.

", - "refs": { - } - }, - "CurrencyCodeValues": { - "base": null, - "refs": { - "PriceSchedule$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "PriceScheduleSpecification$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "ReservedInstanceLimitPrice$CurrencyCode": "

The currency in which the limitPrice amount is specified. At this time, the only supported currency is USD.

", - "ReservedInstances$CurrencyCode": "

The currency of the Reserved Instance. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

", - "ReservedInstancesOffering$CurrencyCode": "

The currency of the Reserved Instance offering you are purchasing. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

" - } - }, - "CustomerGateway": { - "base": "

Describes a customer gateway.

", - "refs": { - "CreateCustomerGatewayResult$CustomerGateway": "

Information about the customer gateway.

", - "CustomerGatewayList$member": null - } - }, - "CustomerGatewayIdStringList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysRequest$CustomerGatewayIds": "

One or more customer gateway IDs.

Default: Describes all your customer gateways.

" - } - }, - "CustomerGatewayList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysResult$CustomerGateways": "

Information about one or more customer gateways.

" - } - }, - "DatafeedSubscriptionState": { - "base": null, - "refs": { - "SpotDatafeedSubscription$State": "

The state of the Spot instance data feed subscription.

" - } - }, - "DateTime": { - "base": null, - "refs": { - "BundleTask$StartTime": "

The time this task started.

", - "BundleTask$UpdateTime": "

The time of the most recent update for the task.

", - "ClientData$UploadStart": "

The time that the disk upload starts.

", - "ClientData$UploadEnd": "

The time that the disk upload ends.

", - "DescribeSpotFleetRequestHistoryRequest$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$LastEvaluatedTime": "

The last date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). All records up to this time were retrieved.

If nextToken indicates that there are more results, this value is not present.

", - "DescribeSpotPriceHistoryRequest$StartTime": "

The date and time, up to the past 90 days, from which to start retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotPriceHistoryRequest$EndTime": "

The date and time, up to the current date, from which to stop retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "EbsInstanceBlockDevice$AttachTime": "

The time stamp when the attachment initiated.

", - "FlowLog$CreationTime": "

The date and time the flow log was created.

", - "GetConsoleOutputResult$Timestamp": "

The time the output was last updated.

", - "GetPasswordDataResult$Timestamp": "

The time the data was last updated.

", - "HistoryRecord$Timestamp": "

The date and time of the event, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "IdFormat$Deadline": "

The date in UTC at which you are permanently switched over to using longer IDs. If a deadline is not yet available for this resource type, this field is not returned.

", - "Instance$LaunchTime": "

The time the instance was launched.

", - "InstanceNetworkInterfaceAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "InstanceStatusDetails$ImpairedSince": "

The time when a status check failed. For an instance that was launched and impaired, this is the time when the instance was launched.

", - "InstanceStatusEvent$NotBefore": "

The earliest scheduled start time for the event.

", - "InstanceStatusEvent$NotAfter": "

The latest scheduled end time for the event.

", - "NatGateway$CreateTime": "

The date and time the NAT gateway was created.

", - "NatGateway$DeleteTime": "

The date and time the NAT gateway was deleted, if applicable.

", - "NetworkInterfaceAttachment$AttachTime": "

The timestamp indicating when the attachment initiated.

", - "ProvisionedBandwidth$RequestTime": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$ProvisionTime": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ReportInstanceStatusRequest$StartTime": "

The time at which the reported instance health state began.

", - "ReportInstanceStatusRequest$EndTime": "

The time at which the reported instance health state ended.

", - "RequestSpotInstancesRequest$ValidFrom": "

The start date of the request. If this is a one-time request, the request becomes active at this date and time and remains active until all instances launch, the request expires, or the request is canceled. If the request is persistent, the request becomes active at this date and time and remains active until it expires or is canceled.

Default: The request is effective indefinitely.

", - "RequestSpotInstancesRequest$ValidUntil": "

The end date of the request. If this is a one-time request, the request remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date and time is reached.

Default: The request is effective indefinitely.

", - "ReservedInstances$Start": "

The date and time the Reserved Instance started.

", - "ReservedInstances$End": "

The time when the Reserved Instance expires.

", - "ReservedInstancesListing$CreateDate": "

The time the listing was created.

", - "ReservedInstancesListing$UpdateDate": "

The last modified timestamp of the listing.

", - "ReservedInstancesModification$CreateDate": "

The time when the modification request was created.

", - "ReservedInstancesModification$UpdateDate": "

The time when the modification request was last updated.

", - "ReservedInstancesModification$EffectiveDate": "

The time for the modification to become effective.

", - "ScheduledInstance$PreviousSlotEndTime": "

The time that the previous schedule ended or will end.

", - "ScheduledInstance$NextSlotStartTime": "

The time for the next schedule to start.

", - "ScheduledInstance$TermStartDate": "

The start date for the Scheduled Instance.

", - "ScheduledInstance$TermEndDate": "

The end date for the Scheduled Instance.

", - "ScheduledInstance$CreateDate": "

The date when the Scheduled Instance was purchased.

", - "ScheduledInstanceAvailability$FirstSlotStartTime": "

The time period for the first schedule to start.

", - "SlotDateTimeRangeRequest$EarliestTime": "

The earliest date and time, in UTC, for the Scheduled Instance to start.

", - "SlotDateTimeRangeRequest$LatestTime": "

The latest date and time, in UTC, for the Scheduled Instance to start. This value must be later than or equal to the earliest date and at most three months in the future.

", - "SlotStartTimeRangeRequest$EarliestTime": "

The earliest date and time, in UTC, for the Scheduled Instance to start.

", - "SlotStartTimeRangeRequest$LatestTime": "

The latest date and time, in UTC, for the Scheduled Instance to start.

", - "Snapshot$StartTime": "

The time stamp when the snapshot was initiated.

", - "SpotFleetRequestConfig$CreateTime": "

The creation date and time of the request.

", - "SpotFleetRequestConfigData$ValidFrom": "

The start date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The default is to start fulfilling the request immediately.

", - "SpotFleetRequestConfigData$ValidUntil": "

The end date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). At this point, no new Spot instance requests are placed or enabled to fulfill the request.

", - "SpotInstanceRequest$ValidFrom": "

The start date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The request becomes active at this date and time.

", - "SpotInstanceRequest$ValidUntil": "

The end date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). If this is a one-time request, it remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date is reached.

", - "SpotInstanceRequest$CreateTime": "

The date and time when the Spot instance request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotInstanceStatus$UpdateTime": "

The date and time of the most recent status update, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotPrice$Timestamp": "

The date and time the request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "VgwTelemetry$LastStatusChange": "

The date and time of the last change in status.

", - "Volume$CreateTime": "

The time stamp when volume creation was initiated.

", - "VolumeAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "VolumeStatusEvent$NotBefore": "

The earliest start time of the event.

", - "VolumeStatusEvent$NotAfter": "

The latest end time of the event.

", - "VpcEndpoint$CreationTimestamp": "

The date and time the VPC endpoint was created.

", - "VpcPeeringConnection$ExpirationTime": "

The time that an unaccepted VPC peering connection will expire.

" - } - }, - "DeleteCustomerGatewayRequest": { - "base": "

Contains the parameters for DeleteCustomerGateway.

", - "refs": { - } - }, - "DeleteDhcpOptionsRequest": { - "base": "

Contains the parameters for DeleteDhcpOptions.

", - "refs": { - } - }, - "DeleteFlowLogsRequest": { - "base": "

Contains the parameters for DeleteFlowLogs.

", - "refs": { - } - }, - "DeleteFlowLogsResult": { - "base": "

Contains the output of DeleteFlowLogs.

", - "refs": { - } - }, - "DeleteInternetGatewayRequest": { - "base": "

Contains the parameters for DeleteInternetGateway.

", - "refs": { - } - }, - "DeleteKeyPairRequest": { - "base": "

Contains the parameters for DeleteKeyPair.

", - "refs": { - } - }, - "DeleteNatGatewayRequest": { - "base": "

Contains the parameters for DeleteNatGateway.

", - "refs": { - } - }, - "DeleteNatGatewayResult": { - "base": "

Contains the output of DeleteNatGateway.

", - "refs": { - } - }, - "DeleteNetworkAclEntryRequest": { - "base": "

Contains the parameters for DeleteNetworkAclEntry.

", - "refs": { - } - }, - "DeleteNetworkAclRequest": { - "base": "

Contains the parameters for DeleteNetworkAcl.

", - "refs": { - } - }, - "DeleteNetworkInterfaceRequest": { - "base": "

Contains the parameters for DeleteNetworkInterface.

", - "refs": { - } - }, - "DeletePlacementGroupRequest": { - "base": "

Contains the parameters for DeletePlacementGroup.

", - "refs": { - } - }, - "DeleteRouteRequest": { - "base": "

Contains the parameters for DeleteRoute.

", - "refs": { - } - }, - "DeleteRouteTableRequest": { - "base": "

Contains the parameters for DeleteRouteTable.

", - "refs": { - } - }, - "DeleteSecurityGroupRequest": { - "base": "

Contains the parameters for DeleteSecurityGroup.

", - "refs": { - } - }, - "DeleteSnapshotRequest": { - "base": "

Contains the parameters for DeleteSnapshot.

", - "refs": { - } - }, - "DeleteSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DeleteSpotDatafeedSubscription.

", - "refs": { - } - }, - "DeleteSubnetRequest": { - "base": "

Contains the parameters for DeleteSubnet.

", - "refs": { - } - }, - "DeleteTagsRequest": { - "base": "

Contains the parameters for DeleteTags.

", - "refs": { - } - }, - "DeleteVolumeRequest": { - "base": "

Contains the parameters for DeleteVolume.

", - "refs": { - } - }, - "DeleteVpcEndpointsRequest": { - "base": "

Contains the parameters for DeleteVpcEndpoints.

", - "refs": { - } - }, - "DeleteVpcEndpointsResult": { - "base": "

Contains the output of DeleteVpcEndpoints.

", - "refs": { - } - }, - "DeleteVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for DeleteVpcPeeringConnection.

", - "refs": { - } - }, - "DeleteVpcPeeringConnectionResult": { - "base": "

Contains the output of DeleteVpcPeeringConnection.

", - "refs": { - } - }, - "DeleteVpcRequest": { - "base": "

Contains the parameters for DeleteVpc.

", - "refs": { - } - }, - "DeleteVpnConnectionRequest": { - "base": "

Contains the parameters for DeleteVpnConnection.

", - "refs": { - } - }, - "DeleteVpnConnectionRouteRequest": { - "base": "

Contains the parameters for DeleteVpnConnectionRoute.

", - "refs": { - } - }, - "DeleteVpnGatewayRequest": { - "base": "

Contains the parameters for DeleteVpnGateway.

", - "refs": { - } - }, - "DeregisterImageRequest": { - "base": "

Contains the parameters for DeregisterImage.

", - "refs": { - } - }, - "DescribeAccountAttributesRequest": { - "base": "

Contains the parameters for DescribeAccountAttributes.

", - "refs": { - } - }, - "DescribeAccountAttributesResult": { - "base": "

Contains the output of DescribeAccountAttributes.

", - "refs": { - } - }, - "DescribeAddressesRequest": { - "base": "

Contains the parameters for DescribeAddresses.

", - "refs": { - } - }, - "DescribeAddressesResult": { - "base": "

Contains the output of DescribeAddresses.

", - "refs": { - } - }, - "DescribeAvailabilityZonesRequest": { - "base": "

Contains the parameters for DescribeAvailabilityZones.

", - "refs": { - } - }, - "DescribeAvailabilityZonesResult": { - "base": "

Contains the output of DescribeAvailabiltyZones.

", - "refs": { - } - }, - "DescribeBundleTasksRequest": { - "base": "

Contains the parameters for DescribeBundleTasks.

", - "refs": { - } - }, - "DescribeBundleTasksResult": { - "base": "

Contains the output of DescribeBundleTasks.

", - "refs": { - } - }, - "DescribeClassicLinkInstancesRequest": { - "base": "

Contains the parameters for DescribeClassicLinkInstances.

", - "refs": { - } - }, - "DescribeClassicLinkInstancesResult": { - "base": "

Contains the output of DescribeClassicLinkInstances.

", - "refs": { - } - }, - "DescribeConversionTaskList": { - "base": null, - "refs": { - "DescribeConversionTasksResult$ConversionTasks": "

Information about the conversion tasks.

" - } - }, - "DescribeConversionTasksRequest": { - "base": "

Contains the parameters for DescribeConversionTasks.

", - "refs": { - } - }, - "DescribeConversionTasksResult": { - "base": "

Contains the output for DescribeConversionTasks.

", - "refs": { - } - }, - "DescribeCustomerGatewaysRequest": { - "base": "

Contains the parameters for DescribeCustomerGateways.

", - "refs": { - } - }, - "DescribeCustomerGatewaysResult": { - "base": "

Contains the output of DescribeCustomerGateways.

", - "refs": { - } - }, - "DescribeDhcpOptionsRequest": { - "base": "

Contains the parameters for DescribeDhcpOptions.

", - "refs": { - } - }, - "DescribeDhcpOptionsResult": { - "base": "

Contains the output of DescribeDhcpOptions.

", - "refs": { - } - }, - "DescribeExportTasksRequest": { - "base": "

Contains the parameters for DescribeExportTasks.

", - "refs": { - } - }, - "DescribeExportTasksResult": { - "base": "

Contains the output for DescribeExportTasks.

", - "refs": { - } - }, - "DescribeFlowLogsRequest": { - "base": "

Contains the parameters for DescribeFlowLogs.

", - "refs": { - } - }, - "DescribeFlowLogsResult": { - "base": "

Contains the output of DescribeFlowLogs.

", - "refs": { - } - }, - "DescribeHostsRequest": { - "base": "

Contains the parameters for DescribeHosts.

", - "refs": { - } - }, - "DescribeHostsResult": { - "base": "

Contains the output of DescribeHosts.

", - "refs": { - } - }, - "DescribeIdFormatRequest": { - "base": "

Contains the parameters for DescribeIdFormat.

", - "refs": { - } - }, - "DescribeIdFormatResult": { - "base": "

Contains the output of DescribeIdFormat.

", - "refs": { - } - }, - "DescribeImageAttributeRequest": { - "base": "

Contains the parameters for DescribeImageAttribute.

", - "refs": { - } - }, - "DescribeImagesRequest": { - "base": "

Contains the parameters for DescribeImages.

", - "refs": { - } - }, - "DescribeImagesResult": { - "base": "

Contains the output of DescribeImages.

", - "refs": { - } - }, - "DescribeImportImageTasksRequest": { - "base": "

Contains the parameters for DescribeImportImageTasks.

", - "refs": { - } - }, - "DescribeImportImageTasksResult": { - "base": "

Contains the output for DescribeImportImageTasks.

", - "refs": { - } - }, - "DescribeImportSnapshotTasksRequest": { - "base": "

Contains the parameters for DescribeImportSnapshotTasks.

", - "refs": { - } - }, - "DescribeImportSnapshotTasksResult": { - "base": "

Contains the output for DescribeImportSnapshotTasks.

", - "refs": { - } - }, - "DescribeInstanceAttributeRequest": { - "base": "

Contains the parameters for DescribeInstanceAttribute.

", - "refs": { - } - }, - "DescribeInstanceStatusRequest": { - "base": "

Contains the parameters for DescribeInstanceStatus.

", - "refs": { - } - }, - "DescribeInstanceStatusResult": { - "base": "

Contains the output of DescribeInstanceStatus.

", - "refs": { - } - }, - "DescribeInstancesRequest": { - "base": "

Contains the parameters for DescribeInstances.

", - "refs": { - } - }, - "DescribeInstancesResult": { - "base": "

Contains the output of DescribeInstances.

", - "refs": { - } - }, - "DescribeInternetGatewaysRequest": { - "base": "

Contains the parameters for DescribeInternetGateways.

", - "refs": { - } - }, - "DescribeInternetGatewaysResult": { - "base": "

Contains the output of DescribeInternetGateways.

", - "refs": { - } - }, - "DescribeKeyPairsRequest": { - "base": "

Contains the parameters for DescribeKeyPairs.

", - "refs": { - } - }, - "DescribeKeyPairsResult": { - "base": "

Contains the output of DescribeKeyPairs.

", - "refs": { - } - }, - "DescribeMovingAddressesRequest": { - "base": "

Contains the parameters for DescribeMovingAddresses.

", - "refs": { - } - }, - "DescribeMovingAddressesResult": { - "base": "

Contains the output of DescribeMovingAddresses.

", - "refs": { - } - }, - "DescribeNatGatewaysRequest": { - "base": "

Contains the parameters for DescribeNatGateways.

", - "refs": { - } - }, - "DescribeNatGatewaysResult": { - "base": "

Contains the output of DescribeNatGateways.

", - "refs": { - } - }, - "DescribeNetworkAclsRequest": { - "base": "

Contains the parameters for DescribeNetworkAcls.

", - "refs": { - } - }, - "DescribeNetworkAclsResult": { - "base": "

Contains the output of DescribeNetworkAcls.

", - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for DescribeNetworkInterfaceAttribute.

", - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeResult": { - "base": "

Contains the output of DescribeNetworkInterfaceAttribute.

", - "refs": { - } - }, - "DescribeNetworkInterfacesRequest": { - "base": "

Contains the parameters for DescribeNetworkInterfaces.

", - "refs": { - } - }, - "DescribeNetworkInterfacesResult": { - "base": "

Contains the output of DescribeNetworkInterfaces.

", - "refs": { - } - }, - "DescribePlacementGroupsRequest": { - "base": "

Contains the parameters for DescribePlacementGroups.

", - "refs": { - } - }, - "DescribePlacementGroupsResult": { - "base": "

Contains the output of DescribePlacementGroups.

", - "refs": { - } - }, - "DescribePrefixListsRequest": { - "base": "

Contains the parameters for DescribePrefixLists.

", - "refs": { - } - }, - "DescribePrefixListsResult": { - "base": "

Contains the output of DescribePrefixLists.

", - "refs": { - } - }, - "DescribeRegionsRequest": { - "base": "

Contains the parameters for DescribeRegions.

", - "refs": { - } - }, - "DescribeRegionsResult": { - "base": "

Contains the output of DescribeRegions.

", - "refs": { - } - }, - "DescribeReservedInstancesListingsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesListings.

", - "refs": { - } - }, - "DescribeReservedInstancesListingsResult": { - "base": "

Contains the output of DescribeReservedInstancesListings.

", - "refs": { - } - }, - "DescribeReservedInstancesModificationsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesModifications.

", - "refs": { - } - }, - "DescribeReservedInstancesModificationsResult": { - "base": "

Contains the output of DescribeReservedInstancesModifications.

", - "refs": { - } - }, - "DescribeReservedInstancesOfferingsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesOfferings.

", - "refs": { - } - }, - "DescribeReservedInstancesOfferingsResult": { - "base": "

Contains the output of DescribeReservedInstancesOfferings.

", - "refs": { - } - }, - "DescribeReservedInstancesRequest": { - "base": "

Contains the parameters for DescribeReservedInstances.

", - "refs": { - } - }, - "DescribeReservedInstancesResult": { - "base": "

Contains the output for DescribeReservedInstances.

", - "refs": { - } - }, - "DescribeRouteTablesRequest": { - "base": "

Contains the parameters for DescribeRouteTables.

", - "refs": { - } - }, - "DescribeRouteTablesResult": { - "base": "

Contains the output of DescribeRouteTables.

", - "refs": { - } - }, - "DescribeScheduledInstanceAvailabilityRequest": { - "base": "

Contains the parameters for DescribeScheduledInstanceAvailability.

", - "refs": { - } - }, - "DescribeScheduledInstanceAvailabilityResult": { - "base": "

Contains the output of DescribeScheduledInstanceAvailability.

", - "refs": { - } - }, - "DescribeScheduledInstancesRequest": { - "base": "

Contains the parameters for DescribeScheduledInstances.

", - "refs": { - } - }, - "DescribeScheduledInstancesResult": { - "base": "

Contains the output of DescribeScheduledInstances.

", - "refs": { - } - }, - "DescribeSecurityGroupReferencesRequest": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupReferencesResult": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupsRequest": { - "base": "

Contains the parameters for DescribeSecurityGroups.

", - "refs": { - } - }, - "DescribeSecurityGroupsResult": { - "base": "

Contains the output of DescribeSecurityGroups.

", - "refs": { - } - }, - "DescribeSnapshotAttributeRequest": { - "base": "

Contains the parameters for DescribeSnapshotAttribute.

", - "refs": { - } - }, - "DescribeSnapshotAttributeResult": { - "base": "

Contains the output of DescribeSnapshotAttribute.

", - "refs": { - } - }, - "DescribeSnapshotsRequest": { - "base": "

Contains the parameters for DescribeSnapshots.

", - "refs": { - } - }, - "DescribeSnapshotsResult": { - "base": "

Contains the output of DescribeSnapshots.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesRequest": { - "base": "

Contains the parameters for DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesResponse": { - "base": "

Contains the output of DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryResponse": { - "base": "

Contains the output of DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsResponse": { - "base": "

Contains the output of DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsResult": { - "base": "

Contains the output of DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryResult": { - "base": "

Contains the output of DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeStaleSecurityGroupsRequest": { - "base": null, - "refs": { - } - }, - "DescribeStaleSecurityGroupsResult": { - "base": null, - "refs": { - } - }, - "DescribeSubnetsRequest": { - "base": "

Contains the parameters for DescribeSubnets.

", - "refs": { - } - }, - "DescribeSubnetsResult": { - "base": "

Contains the output of DescribeSubnets.

", - "refs": { - } - }, - "DescribeTagsRequest": { - "base": "

Contains the parameters for DescribeTags.

", - "refs": { - } - }, - "DescribeTagsResult": { - "base": "

Contains the output of DescribeTags.

", - "refs": { - } - }, - "DescribeVolumeAttributeRequest": { - "base": "

Contains the parameters for DescribeVolumeAttribute.

", - "refs": { - } - }, - "DescribeVolumeAttributeResult": { - "base": "

Contains the output of DescribeVolumeAttribute.

", - "refs": { - } - }, - "DescribeVolumeStatusRequest": { - "base": "

Contains the parameters for DescribeVolumeStatus.

", - "refs": { - } - }, - "DescribeVolumeStatusResult": { - "base": "

Contains the output of DescribeVolumeStatus.

", - "refs": { - } - }, - "DescribeVolumesRequest": { - "base": "

Contains the parameters for DescribeVolumes.

", - "refs": { - } - }, - "DescribeVolumesResult": { - "base": "

Contains the output of DescribeVolumes.

", - "refs": { - } - }, - "DescribeVpcAttributeRequest": { - "base": "

Contains the parameters for DescribeVpcAttribute.

", - "refs": { - } - }, - "DescribeVpcAttributeResult": { - "base": "

Contains the output of DescribeVpcAttribute.

", - "refs": { - } - }, - "DescribeVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for DescribeVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DescribeVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of DescribeVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DescribeVpcClassicLinkRequest": { - "base": "

Contains the parameters for DescribeVpcClassicLink.

", - "refs": { - } - }, - "DescribeVpcClassicLinkResult": { - "base": "

Contains the output of DescribeVpcClassicLink.

", - "refs": { - } - }, - "DescribeVpcEndpointServicesRequest": { - "base": "

Contains the parameters for DescribeVpcEndpointServices.

", - "refs": { - } - }, - "DescribeVpcEndpointServicesResult": { - "base": "

Contains the output of DescribeVpcEndpointServices.

", - "refs": { - } - }, - "DescribeVpcEndpointsRequest": { - "base": "

Contains the parameters for DescribeVpcEndpoints.

", - "refs": { - } - }, - "DescribeVpcEndpointsResult": { - "base": "

Contains the output of DescribeVpcEndpoints.

", - "refs": { - } - }, - "DescribeVpcPeeringConnectionsRequest": { - "base": "

Contains the parameters for DescribeVpcPeeringConnections.

", - "refs": { - } - }, - "DescribeVpcPeeringConnectionsResult": { - "base": "

Contains the output of DescribeVpcPeeringConnections.

", - "refs": { - } - }, - "DescribeVpcsRequest": { - "base": "

Contains the parameters for DescribeVpcs.

", - "refs": { - } - }, - "DescribeVpcsResult": { - "base": "

Contains the output of DescribeVpcs.

", - "refs": { - } - }, - "DescribeVpnConnectionsRequest": { - "base": "

Contains the parameters for DescribeVpnConnections.

", - "refs": { - } - }, - "DescribeVpnConnectionsResult": { - "base": "

Contains the output of DescribeVpnConnections.

", - "refs": { - } - }, - "DescribeVpnGatewaysRequest": { - "base": "

Contains the parameters for DescribeVpnGateways.

", - "refs": { - } - }, - "DescribeVpnGatewaysResult": { - "base": "

Contains the output of DescribeVpnGateways.

", - "refs": { - } - }, - "DetachClassicLinkVpcRequest": { - "base": "

Contains the parameters for DetachClassicLinkVpc.

", - "refs": { - } - }, - "DetachClassicLinkVpcResult": { - "base": "

Contains the output of DetachClassicLinkVpc.

", - "refs": { - } - }, - "DetachInternetGatewayRequest": { - "base": "

Contains the parameters for DetachInternetGateway.

", - "refs": { - } - }, - "DetachNetworkInterfaceRequest": { - "base": "

Contains the parameters for DetachNetworkInterface.

", - "refs": { - } - }, - "DetachVolumeRequest": { - "base": "

Contains the parameters for DetachVolume.

", - "refs": { - } - }, - "DetachVpnGatewayRequest": { - "base": "

Contains the parameters for DetachVpnGateway.

", - "refs": { - } - }, - "DeviceType": { - "base": null, - "refs": { - "Image$RootDeviceType": "

The type of root device used by the AMI. The AMI can use an EBS volume or an instance store volume.

", - "Instance$RootDeviceType": "

The root device type used by the AMI. The AMI can use an EBS volume or an instance store volume.

" - } - }, - "DhcpConfiguration": { - "base": "

Describes a DHCP configuration option.

", - "refs": { - "DhcpConfigurationList$member": null - } - }, - "DhcpConfigurationList": { - "base": null, - "refs": { - "DhcpOptions$DhcpConfigurations": "

One or more DHCP options in the set.

" - } - }, - "DhcpConfigurationValueList": { - "base": null, - "refs": { - "DhcpConfiguration$Values": "

One or more values for the DHCP option.

" - } - }, - "DhcpOptions": { - "base": "

Describes a set of DHCP options.

", - "refs": { - "CreateDhcpOptionsResult$DhcpOptions": "

A set of DHCP options.

", - "DhcpOptionsList$member": null - } - }, - "DhcpOptionsIdStringList": { - "base": null, - "refs": { - "DescribeDhcpOptionsRequest$DhcpOptionsIds": "

The IDs of one or more DHCP options sets.

Default: Describes all your DHCP options sets.

" - } - }, - "DhcpOptionsList": { - "base": null, - "refs": { - "DescribeDhcpOptionsResult$DhcpOptions": "

Information about one or more DHCP options sets.

" - } - }, - "DisableVgwRoutePropagationRequest": { - "base": "

Contains the parameters for DisableVgwRoutePropagation.

", - "refs": { - } - }, - "DisableVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for DisableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DisableVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of DisableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DisableVpcClassicLinkRequest": { - "base": "

Contains the parameters for DisableVpcClassicLink.

", - "refs": { - } - }, - "DisableVpcClassicLinkResult": { - "base": "

Contains the output of DisableVpcClassicLink.

", - "refs": { - } - }, - "DisassociateAddressRequest": { - "base": "

Contains the parameters for DisassociateAddress.

", - "refs": { - } - }, - "DisassociateRouteTableRequest": { - "base": "

Contains the parameters for DisassociateRouteTable.

", - "refs": { - } - }, - "DiskImage": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImageList$member": null - } - }, - "DiskImageDescription": { - "base": "

Describes a disk image.

", - "refs": { - "ImportInstanceVolumeDetailItem$Image": "

The image.

", - "ImportVolumeTaskDetails$Image": "

The image.

" - } - }, - "DiskImageDetail": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImage$Image": "

Information about the disk image.

", - "ImportVolumeRequest$Image": "

The disk image.

" - } - }, - "DiskImageFormat": { - "base": null, - "refs": { - "DiskImageDescription$Format": "

The disk image format.

", - "DiskImageDetail$Format": "

The disk image format.

", - "ExportToS3Task$DiskImageFormat": "

The format for the exported image.

", - "ExportToS3TaskSpecification$DiskImageFormat": "

The format for the exported image.

" - } - }, - "DiskImageList": { - "base": null, - "refs": { - "ImportInstanceRequest$DiskImages": "

The disk image.

" - } - }, - "DiskImageVolumeDescription": { - "base": "

Describes a disk image volume.

", - "refs": { - "ImportInstanceVolumeDetailItem$Volume": "

The volume.

", - "ImportVolumeTaskDetails$Volume": "

The volume.

" - } - }, - "DomainType": { - "base": null, - "refs": { - "Address$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

", - "AllocateAddressRequest$Domain": "

Set to vpc to allocate the address for use with instances in a VPC.

Default: The address is for use with instances in EC2-Classic.

", - "AllocateAddressResult$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

" - } - }, - "Double": { - "base": null, - "refs": { - "ClientData$UploadSize": "

The size of the uploaded disk image, in GiB.

", - "PriceSchedule$Price": "

The fixed price for the term.

", - "PriceScheduleSpecification$Price": "

The fixed price for the term.

", - "PricingDetail$Price": "

The price per instance.

", - "RecurringCharge$Amount": "

The amount of the recurring charge.

", - "ReservedInstanceLimitPrice$Amount": "

Used for Reserved Instance Marketplace offerings. Specifies the limit price on the total order (instanceCount * price).

", - "SnapshotDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SnapshotTaskDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SpotFleetLaunchSpecification$WeightedCapacity": "

The number of units provided by the specified instance type. These are the same units that you chose to set the target capacity in terms (instances or a performance characteristic such as vCPUs, memory, or I/O).

If the target capacity divided by this value is not a whole number, we round the number of instances to the next whole number. If this value is not specified, the default is 1.

", - "SpotFleetRequestConfigData$FulfilledCapacity": "

The number of units fulfilled by this request compared to the set target capacity.

" - } - }, - "EbsBlockDevice": { - "base": "

Describes a block device for an EBS volume.

", - "refs": { - "BlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDevice": { - "base": "

Describes a parameter used to set up an EBS volume in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDeviceSpecification": { - "base": "

Describes information used to set up an EBS volume specified in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingSpecification$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EnableVgwRoutePropagationRequest": { - "base": "

Contains the parameters for EnableVgwRoutePropagation.

", - "refs": { - } - }, - "EnableVolumeIORequest": { - "base": "

Contains the parameters for EnableVolumeIO.

", - "refs": { - } - }, - "EnableVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for EnableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "EnableVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of EnableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "EnableVpcClassicLinkRequest": { - "base": "

Contains the parameters for EnableVpcClassicLink.

", - "refs": { - } - }, - "EnableVpcClassicLinkResult": { - "base": "

Contains the output of EnableVpcClassicLink.

", - "refs": { - } - }, - "EventCode": { - "base": null, - "refs": { - "InstanceStatusEvent$Code": "

The event code.

" - } - }, - "EventInformation": { - "base": "

Describes a Spot fleet event.

", - "refs": { - "HistoryRecord$EventInformation": "

Information about the event.

" - } - }, - "EventType": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryRequest$EventType": "

The type of events to describe. By default, all events are described.

", - "HistoryRecord$EventType": "

The event type.

  • error - Indicates an error with the Spot fleet request.

  • fleetRequestChange - Indicates a change in the status or configuration of the Spot fleet request.

  • instanceChange - Indicates that an instance was launched or terminated.

" - } - }, - "ExcessCapacityTerminationPolicy": { - "base": null, - "refs": { - "ModifySpotFleetRequestRequest$ExcessCapacityTerminationPolicy": "

Indicates whether running Spot instances should be terminated if the target capacity of the Spot fleet request is decreased below the current size of the Spot fleet.

", - "SpotFleetRequestConfigData$ExcessCapacityTerminationPolicy": "

Indicates whether running Spot instances should be terminated if the target capacity of the Spot fleet request is decreased below the current size of the Spot fleet.

" - } - }, - "ExecutableByStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ExecutableUsers": "

Scopes the images by users with explicit launch permissions. Specify an AWS account ID, self (the sender of the request), or all (public AMIs).

" - } - }, - "ExportEnvironment": { - "base": null, - "refs": { - "CreateInstanceExportTaskRequest$TargetEnvironment": "

The target virtualization environment.

", - "InstanceExportDetails$TargetEnvironment": "

The target virtualization environment.

" - } - }, - "ExportTask": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskResult$ExportTask": "

Information about the instance export task.

", - "ExportTaskList$member": null - } - }, - "ExportTaskIdStringList": { - "base": null, - "refs": { - "DescribeExportTasksRequest$ExportTaskIds": "

One or more export task IDs.

" - } - }, - "ExportTaskList": { - "base": null, - "refs": { - "DescribeExportTasksResult$ExportTasks": "

Information about the export tasks.

" - } - }, - "ExportTaskState": { - "base": null, - "refs": { - "ExportTask$State": "

The state of the export task.

" - } - }, - "ExportToS3Task": { - "base": "

Describes the format and location for an instance export task.

", - "refs": { - "ExportTask$ExportToS3Task": "

Information about the export task.

" - } - }, - "ExportToS3TaskSpecification": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskRequest$ExportToS3Task": "

The format and location for an instance export task.

" - } - }, - "Filter": { - "base": "

A filter name and value pair that is used to return a more specific list of results. Filters can be used to match a set of resources by various criteria, such as tags, attributes, or IDs.

", - "refs": { - "FilterList$member": null - } - }, - "FilterList": { - "base": null, - "refs": { - "DescribeAddressesRequest$Filters": "

One or more filters. Filter names and values are case-sensitive.

  • allocation-id - [EC2-VPC] The allocation ID for the address.

  • association-id - [EC2-VPC] The association ID for the address.

  • domain - Indicates whether the address is for use in EC2-Classic (standard) or in a VPC (vpc).

  • instance-id - The ID of the instance the address is associated with, if any.

  • network-interface-id - [EC2-VPC] The ID of the network interface that the address is associated with, if any.

  • network-interface-owner-id - The AWS account ID of the owner.

  • private-ip-address - [EC2-VPC] The private IP address associated with the Elastic IP address.

  • public-ip - The Elastic IP address.

", - "DescribeAvailabilityZonesRequest$Filters": "

One or more filters.

  • message - Information about the Availability Zone.

  • region-name - The name of the region for the Availability Zone (for example, us-east-1).

  • state - The state of the Availability Zone (available | information | impaired | unavailable).

  • zone-name - The name of the Availability Zone (for example, us-east-1a).

", - "DescribeBundleTasksRequest$Filters": "

One or more filters.

  • bundle-id - The ID of the bundle task.

  • error-code - If the task failed, the error code returned.

  • error-message - If the task failed, the error message returned.

  • instance-id - The ID of the instance.

  • progress - The level of task completion, as a percentage (for example, 20%).

  • s3-bucket - The Amazon S3 bucket to store the AMI.

  • s3-prefix - The beginning of the AMI name.

  • start-time - The time the task started (for example, 2013-09-15T17:15:20.000Z).

  • state - The state of the task (pending | waiting-for-shutdown | bundling | storing | cancelling | complete | failed).

  • update-time - The time of the most recent update for the task.

", - "DescribeClassicLinkInstancesRequest$Filters": "

One or more filters.

  • group-id - The ID of a VPC security group that's associated with the instance.

  • instance-id - The ID of the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC that the instance is linked to.

", - "DescribeConversionTasksRequest$Filters": "

One or more filters.

", - "DescribeCustomerGatewaysRequest$Filters": "

One or more filters.

  • bgp-asn - The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

  • customer-gateway-id - The ID of the customer gateway.

  • ip-address - The IP address of the customer gateway's Internet-routable external interface.

  • state - The state of the customer gateway (pending | available | deleting | deleted).

  • type - The type of customer gateway. Currently, the only supported type is ipsec.1.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeDhcpOptionsRequest$Filters": "

One or more filters.

  • dhcp-options-id - The ID of a set of DHCP options.

  • key - The key for one of the options (for example, domain-name).

  • value - The value for one of the options.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeFlowLogsRequest$Filter": "

One or more filters.

  • deliver-log-status - The status of the logs delivery (SUCCESS | FAILED).

  • flow-log-id - The ID of the flow log.

  • log-group-name - The name of the log group.

  • resource-id - The ID of the VPC, subnet, or network interface.

  • traffic-type - The type of traffic (ACCEPT | REJECT | ALL)

", - "DescribeHostsRequest$Filter": "

One or more filters.

  • instance-type - The instance type size that the Dedicated host is configured to support.

  • auto-placement - Whether auto-placement is enabled or disabled (on | off).

  • host-reservation-id - The ID of the reservation associated with this host.

  • client-token - The idempotency token you provided when you launched the instance

  • state- The allocation state of the Dedicated host (available | under-assessment | permanent-failure | released | released-permanent-failure).

  • availability-zone - The Availability Zone of the host.

", - "DescribeImagesRequest$Filters": "

One or more filters.

  • architecture - The image architecture (i386 | x86_64).

  • block-device-mapping.delete-on-termination - A Boolean value that indicates whether the Amazon EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh).

  • block-device-mapping.snapshot-id - The ID of the snapshot used for the EBS volume.

  • block-device-mapping.volume-size - The volume size of the EBS volume, in GiB.

  • block-device-mapping.volume-type - The volume type of the EBS volume (gp2 | io1 | st1 | sc1 | standard).

  • description - The description of the image (provided during image creation).

  • hypervisor - The hypervisor type (ovm | xen).

  • image-id - The ID of the image.

  • image-type - The image type (machine | kernel | ramdisk).

  • is-public - A Boolean that indicates whether the image is public.

  • kernel-id - The kernel ID.

  • manifest-location - The location of the image manifest.

  • name - The name of the AMI (provided during image creation).

  • owner-alias - The AWS account alias (for example, amazon).

  • owner-id - The AWS account ID of the image owner.

  • platform - The platform. To only list Windows-based AMIs, use windows.

  • product-code - The product code.

  • product-code.type - The type of the product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • root-device-name - The name of the root device volume (for example, /dev/sda1).

  • root-device-type - The type of the root device volume (ebs | instance-store).

  • state - The state of the image (available | pending | failed).

  • state-reason-code - The reason code for the state change.

  • state-reason-message - The message for the state change.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • virtualization-type - The virtualization type (paravirtual | hvm).

", - "DescribeImportImageTasksRequest$Filters": "

Filter tasks using the task-state filter and one of the following values: active, completed, deleting, deleted.

", - "DescribeImportSnapshotTasksRequest$Filters": "

One or more filters.

", - "DescribeInstanceStatusRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone of the instance.

  • event.code - The code for the scheduled event (instance-reboot | system-reboot | system-maintenance | instance-retirement | instance-stop).

  • event.description - A description of the event.

  • event.not-after - The latest end time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • event.not-before - The earliest start time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • instance-state-code - The code for the instance state, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-status.reachability - Filters on instance status where the name is reachability (passed | failed | initializing | insufficient-data).

  • instance-status.status - The status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

  • system-status.reachability - Filters on system status where the name is reachability (passed | failed | initializing | insufficient-data).

  • system-status.status - The system status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

", - "DescribeInstancesRequest$Filters": "

One or more filters.

  • affinity - The affinity setting for an instance running on a Dedicated host (default | host).

  • architecture - The instance architecture (i386 | x86_64).

  • availability-zone - The Availability Zone of the instance.

  • block-device-mapping.attach-time - The attach time for an EBS volume mapped to the instance, for example, 2010-09-15T17:15:20.000Z.

  • block-device-mapping.delete-on-termination - A Boolean that indicates whether the EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh or xvdh).

  • block-device-mapping.status - The status for the EBS volume (attaching | attached | detaching | detached).

  • block-device-mapping.volume-id - The volume ID of the EBS volume.

  • client-token - The idempotency token you provided when you launched the instance.

  • dns-name - The public DNS name of the instance.

  • group-id - The ID of the security group for the instance. EC2-Classic only.

  • group-name - The name of the security group for the instance. EC2-Classic only.

  • host-Id - The ID of the Dedicated host on which the instance is running, if applicable.

  • hypervisor - The hypervisor type of the instance (ovm | xen).

  • iam-instance-profile.arn - The instance profile associated with the instance. Specified as an ARN.

  • image-id - The ID of the image used to launch the instance.

  • instance-id - The ID of the instance.

  • instance-lifecycle - Indicates whether this is a Spot Instance or a Scheduled Instance (spot | scheduled).

  • instance-state-code - The state of the instance, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are: 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-type - The type of instance (for example, t2.micro).

  • instance.group-id - The ID of the security group for the instance.

  • instance.group-name - The name of the security group for the instance.

  • ip-address - The public IP address of the instance.

  • kernel-id - The kernel ID.

  • key-name - The name of the key pair used when the instance was launched.

  • launch-index - When launching multiple instances, this is the index for the instance in the launch group (for example, 0, 1, 2, and so on).

  • launch-time - The time when the instance was launched.

  • monitoring-state - Indicates whether monitoring is enabled for the instance (disabled | enabled).

  • owner-id - The AWS account ID of the instance owner.

  • placement-group-name - The name of the placement group for the instance.

  • platform - The platform. Use windows if you have Windows instances; otherwise, leave blank.

  • private-dns-name - The private DNS name of the instance.

  • private-ip-address - The private IP address of the instance.

  • product-code - The product code associated with the AMI used to launch the instance.

  • product-code.type - The type of product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • reason - The reason for the current state of the instance (for example, shows \"User Initiated [date]\" when you stop or terminate the instance). Similar to the state-reason-code filter.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • reservation-id - The ID of the instance's reservation. A reservation ID is created any time you launch an instance. A reservation ID has a one-to-one relationship with an instance launch request, but can be associated with more than one instance if you launch multiple instances using the same launch request. For example, if you launch one instance, you'll get one reservation ID. If you launch ten instances using the same launch request, you'll also get one reservation ID.

  • root-device-name - The name of the root device for the instance (for example, /dev/sda1 or /dev/xvda).

  • root-device-type - The type of root device that the instance uses (ebs | instance-store).

  • source-dest-check - Indicates whether the instance performs source/destination checking. A value of true means that checking is enabled, and false means checking is disabled. The value must be false for the instance to perform network address translation (NAT) in your VPC.

  • spot-instance-request-id - The ID of the Spot instance request.

  • state-reason-code - The reason code for the state change.

  • state-reason-message - A message that describes the state change.

  • subnet-id - The ID of the subnet for the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource, where tag:key is the tag's key.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • tenancy - The tenancy of an instance (dedicated | default | host).

  • virtualization-type - The virtualization type of the instance (paravirtual | hvm).

  • vpc-id - The ID of the VPC that the instance is running in.

  • network-interface.description - The description of the network interface.

  • network-interface.subnet-id - The ID of the subnet for the network interface.

  • network-interface.vpc-id - The ID of the VPC for the network interface.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.owner-id - The ID of the owner of the network interface.

  • network-interface.availability-zone - The Availability Zone for the network interface.

  • network-interface.requester-id - The requester ID for the network interface.

  • network-interface.requester-managed - Indicates whether the network interface is being managed by AWS.

  • network-interface.status - The status of the network interface (available) | in-use).

  • network-interface.mac-address - The MAC address of the network interface.

  • network-interface.private-dns-name - The private DNS name of the network interface.

  • network-interface.source-dest-check - Whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • network-interface.group-id - The ID of a security group associated with the network interface.

  • network-interface.group-name - The name of a security group associated with the network interface.

  • network-interface.attachment.attachment-id - The ID of the interface attachment.

  • network-interface.attachment.instance-id - The ID of the instance to which the network interface is attached.

  • network-interface.attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • network-interface.addresses.private-ip-address - The private IP address associated with the network interface.

  • network-interface.attachment.device-index - The device index to which the network interface is attached.

  • network-interface.attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • network-interface.attachment.attach-time - The time that the network interface was attached to an instance.

  • network-interface.attachment.delete-on-termination - Specifies whether the attachment is deleted when an instance is terminated.

  • network-interface.addresses.primary - Specifies whether the IP address of the network interface is the primary private IP address.

  • network-interface.addresses.association.public-ip - The ID of the association of an Elastic IP address with a network interface.

  • network-interface.addresses.association.ip-owner-id - The owner ID of the private IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

", - "DescribeInternetGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (available). Present only if a VPC is attached.

  • attachment.vpc-id - The ID of an attached VPC.

  • internet-gateway-id - The ID of the Internet gateway.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeKeyPairsRequest$Filters": "

One or more filters.

  • fingerprint - The fingerprint of the key pair.

  • key-name - The name of the key pair.

", - "DescribeMovingAddressesRequest$Filters": "

One or more filters.

  • moving-status - The status of the Elastic IP address (MovingToVpc | RestoringToClassic).

", - "DescribeNatGatewaysRequest$Filter": "

One or more filters.

  • nat-gateway-id - The ID of the NAT gateway.

  • state - The state of the NAT gateway (pending | failed | available | deleting | deleted).

  • subnet-id - The ID of the subnet in which the NAT gateway resides.

  • vpc-id - The ID of the VPC in which the NAT gateway resides.

", - "DescribeNetworkAclsRequest$Filters": "

One or more filters.

  • association.association-id - The ID of an association ID for the ACL.

  • association.network-acl-id - The ID of the network ACL involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • default - Indicates whether the ACL is the default network ACL for the VPC.

  • entry.cidr - The CIDR range specified in the entry.

  • entry.egress - Indicates whether the entry applies to egress traffic.

  • entry.icmp.code - The ICMP code specified in the entry, if any.

  • entry.icmp.type - The ICMP type specified in the entry, if any.

  • entry.port-range.from - The start of the port range specified in the entry.

  • entry.port-range.to - The end of the port range specified in the entry.

  • entry.protocol - The protocol specified in the entry (tcp | udp | icmp or a protocol number).

  • entry.rule-action - Allows or denies the matching traffic (allow | deny).

  • entry.rule-number - The number of an entry (in other words, rule) in the ACL's set of entries.

  • network-acl-id - The ID of the network ACL.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network ACL.

", - "DescribeNetworkInterfacesRequest$Filters": "

One or more filters.

  • addresses.private-ip-address - The private IP addresses associated with the network interface.

  • addresses.primary - Whether the private IP address is the primary IP address associated with the network interface.

  • addresses.association.public-ip - The association ID returned when the network interface was associated with the Elastic IP address.

  • addresses.association.owner-id - The owner ID of the addresses associated with the network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.public-dns-name - The public DNS name for the network interface.

  • attachment.attachment-id - The ID of the interface attachment.

  • attachment.attach.time - The time that the network interface was attached to an instance.

  • attachment.delete-on-termination - Indicates whether the attachment is deleted when an instance is terminated.

  • attachment.device-index - The device index to which the network interface is attached.

  • attachment.instance-id - The ID of the instance to which the network interface is attached.

  • attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • attachment.nat-gateway-id - The ID of the NAT gateway to which the network interface is attached.

  • attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone of the network interface.

  • description - The description of the network interface.

  • group-id - The ID of a security group associated with the network interface.

  • group-name - The name of a security group associated with the network interface.

  • mac-address - The MAC address of the network interface.

  • network-interface-id - The ID of the network interface.

  • owner-id - The AWS account ID of the network interface owner.

  • private-ip-address - The private IP address or addresses of the network interface.

  • private-dns-name - The private DNS name of the network interface.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • requester-managed - Indicates whether the network interface is being managed by an AWS service (for example, AWS Management Console, Auto Scaling, and so on).

  • source-desk-check - Indicates whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • status - The status of the network interface. If the network interface is not attached to an instance, the status is available; if a network interface is attached to an instance the status is in-use.

  • subnet-id - The ID of the subnet for the network interface.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network interface.

", - "DescribePlacementGroupsRequest$Filters": "

One or more filters.

  • group-name - The name of the placement group.

  • state - The state of the placement group (pending | available | deleting | deleted).

  • strategy - The strategy of the placement group (cluster).

", - "DescribePrefixListsRequest$Filters": "

One or more filters.

  • prefix-list-id: The ID of a prefix list.

  • prefix-list-name: The name of a prefix list.

", - "DescribeRegionsRequest$Filters": "

One or more filters.

  • endpoint - The endpoint of the region (for example, ec2.us-east-1.amazonaws.com).

  • region-name - The name of the region (for example, us-east-1).

", - "DescribeReservedInstancesListingsRequest$Filters": "

One or more filters.

  • reserved-instances-id - The ID of the Reserved Instances.

  • reserved-instances-listing-id - The ID of the Reserved Instances listing.

  • status - The status of the Reserved Instance listing (pending | active | cancelled | closed).

  • status-message - The reason for the status.

", - "DescribeReservedInstancesModificationsRequest$Filters": "

One or more filters.

  • client-token - The idempotency token for the modification request.

  • create-date - The time when the modification request was created.

  • effective-date - The time when the modification becomes effective.

  • modification-result.reserved-instances-id - The ID for the Reserved Instances created as part of the modification request. This ID is only available when the status of the modification is fulfilled.

  • modification-result.target-configuration.availability-zone - The Availability Zone for the new Reserved Instances.

  • modification-result.target-configuration.instance-count - The number of new Reserved Instances.

  • modification-result.target-configuration.instance-type - The instance type of the new Reserved Instances.

  • modification-result.target-configuration.platform - The network platform of the new Reserved Instances (EC2-Classic | EC2-VPC).

  • reserved-instances-id - The ID of the Reserved Instances modified.

  • reserved-instances-modification-id - The ID of the modification request.

  • status - The status of the Reserved Instances modification request (processing | fulfilled | failed).

  • status-message - The reason for the status.

  • update-date - The time when the modification request was last updated.

", - "DescribeReservedInstancesOfferingsRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (for example, one year or three years), in seconds (31536000 | 94608000).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type that is covered by the reservation.

  • marketplace - Set to true to show only Reserved Instance Marketplace offerings. When this filter is not used, which is the default behavior, all offerings from both AWS and the Reserved Instance Marketplace are listed.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC. (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC))

  • reserved-instances-offering-id - The Reserved Instances offering ID.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeReservedInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (one year or three years), in seconds (31536000 | 94608000).

  • end - The time when the Reserved Instance expires (for example, 2015-08-07T11:54:42.000Z).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type that is covered by the reservation.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC)).

  • reserved-instances-id - The ID of the Reserved Instance.

  • start - The time at which the Reserved Instance purchase request was placed (for example, 2014-08-07T11:54:42.000Z).

  • state - The state of the Reserved Instance (payment-pending | active | payment-failed | retired).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeRouteTablesRequest$Filters": "

One or more filters.

  • association.route-table-association-id - The ID of an association ID for the route table.

  • association.route-table-id - The ID of the route table involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • association.main - Indicates whether the route table is the main route table for the VPC (true | false).

  • route-table-id - The ID of the route table.

  • route.destination-cidr-block - The CIDR range specified in a route in the table.

  • route.destination-prefix-list-id - The ID (prefix) of the AWS service specified in a route in the table.

  • route.gateway-id - The ID of a gateway specified in a route in the table.

  • route.instance-id - The ID of an instance specified in a route in the table.

  • route.nat-gateway-id - The ID of a NAT gateway.

  • route.origin - Describes how the route was created. CreateRouteTable indicates that the route was automatically created when the route table was created; CreateRoute indicates that the route was manually added to the route table; EnableVgwRoutePropagation indicates that the route was propagated by route propagation.

  • route.state - The state of a route in the route table (active | blackhole). The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, the specified NAT instance has been terminated, and so on).

  • route.vpc-peering-connection-id - The ID of a VPC peering connection specified in a route in the table.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the route table.

", - "DescribeScheduledInstanceAvailabilityRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone (for example, us-west-2a).

  • instance-type - The instance type (for example, c4.large).

  • network-platform - The network platform (EC2-Classic or EC2-VPC).

  • platform - The platform (Linux/UNIX or Windows).

", - "DescribeScheduledInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone (for example, us-west-2a).

  • instance-type - The instance type (for example, c4.large).

  • network-platform - The network platform (EC2-Classic or EC2-VPC).

  • platform - The platform (Linux/UNIX or Windows).

", - "DescribeSecurityGroupsRequest$Filters": "

One or more filters. If using multiple filters for rules, the results include security groups for which any combination of rules - not necessarily a single rule - match all filters.

  • description - The description of the security group.

  • egress.ip-permission.prefix-list-id - The ID (prefix) of the AWS service to which the security group allows access.

  • group-id - The ID of the security group.

  • group-name - The name of the security group.

  • ip-permission.cidr - A CIDR range that has been granted permission.

  • ip-permission.from-port - The start of port range for the TCP and UDP protocols, or an ICMP type number.

  • ip-permission.group-id - The ID of a security group that has been granted permission.

  • ip-permission.group-name - The name of a security group that has been granted permission.

  • ip-permission.protocol - The IP protocol for the permission (tcp | udp | icmp or a protocol number).

  • ip-permission.to-port - The end of port range for the TCP and UDP protocols, or an ICMP code.

  • ip-permission.user-id - The ID of an AWS account that has been granted permission.

  • owner-id - The AWS account ID of the owner of the security group.

  • tag-key - The key of a tag assigned to the security group.

  • tag-value - The value of a tag assigned to the security group.

  • vpc-id - The ID of the VPC specified when the security group was created.

", - "DescribeSnapshotsRequest$Filters": "

One or more filters.

  • description - A description of the snapshot.

  • owner-alias - The AWS account alias (for example, amazon) that owns the snapshot.

  • owner-id - The ID of the AWS account that owns the snapshot.

  • progress - The progress of the snapshot, as a percentage (for example, 80%).

  • snapshot-id - The snapshot ID.

  • start-time - The time stamp when the snapshot was initiated.

  • status - The status of the snapshot (pending | completed | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The ID of the volume the snapshot is for.

  • volume-size - The size of the volume, in GiB.

", - "DescribeSpotInstanceRequestsRequest$Filters": "

One or more filters.

  • availability-zone-group - The Availability Zone group.

  • create-time - The time stamp when the Spot instance request was created.

  • fault-code - The fault code related to the request.

  • fault-message - The fault message related to the request.

  • instance-id - The ID of the instance that fulfilled the request.

  • launch-group - The Spot instance launch group.

  • launch.block-device-mapping.delete-on-termination - Indicates whether the Amazon EBS volume is deleted on instance termination.

  • launch.block-device-mapping.device-name - The device name for the Amazon EBS volume (for example, /dev/sdh).

  • launch.block-device-mapping.snapshot-id - The ID of the snapshot used for the Amazon EBS volume.

  • launch.block-device-mapping.volume-size - The size of the Amazon EBS volume, in GiB.

  • launch.block-device-mapping.volume-type - The type of the Amazon EBS volume: gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1for Cold HDD, or standard for Magnetic.

  • launch.group-id - The security group for the instance.

  • launch.image-id - The ID of the AMI.

  • launch.instance-type - The type of instance (for example, m3.medium).

  • launch.kernel-id - The kernel ID.

  • launch.key-name - The name of the key pair the instance launched with.

  • launch.monitoring-enabled - Whether monitoring is enabled for the Spot instance.

  • launch.ramdisk-id - The RAM disk ID.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.device-index - The index of the device for the network interface attachment on the instance.

  • network-interface.subnet-id - The ID of the subnet for the instance.

  • network-interface.description - A description of the network interface.

  • network-interface.private-ip-address - The primary private IP address of the network interface.

  • network-interface.delete-on-termination - Indicates whether the network interface is deleted when the instance is terminated.

  • network-interface.group-id - The ID of the security group associated with the network interface.

  • network-interface.group-name - The name of the security group associated with the network interface.

  • network-interface.addresses.primary - Indicates whether the IP address is the primary private IP address.

  • product-description - The product description associated with the instance (Linux/UNIX | Windows).

  • spot-instance-request-id - The Spot instance request ID.

  • spot-price - The maximum hourly price for any Spot instance launched to fulfill the request.

  • state - The state of the Spot instance request (open | active | closed | cancelled | failed). Spot bid status information can help you track your Amazon EC2 Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

  • status-code - The short code describing the most recent evaluation of your Spot instance request.

  • status-message - The message explaining the status of the Spot instance request.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of Spot instance request (one-time | persistent).

  • launched-availability-zone - The Availability Zone in which the bid is launched.

  • valid-from - The start date of the request.

  • valid-until - The end date of the request.

", - "DescribeSpotPriceHistoryRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone for which prices should be returned.

  • instance-type - The type of instance (for example, m3.medium).

  • product-description - The product description for the Spot price (Linux/UNIX | SUSE Linux | Windows | Linux/UNIX (Amazon VPC) | SUSE Linux (Amazon VPC) | Windows (Amazon VPC)).

  • spot-price - The Spot price. The value must match exactly (or use wildcards; greater than or less than comparison is not supported).

  • timestamp - The timestamp of the Spot price history, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). You can use wildcards (* and ?). Greater than or less than comparison is not supported.

", - "DescribeSubnetsRequest$Filters": "

One or more filters.

  • availabilityZone - The Availability Zone for the subnet. You can also use availability-zone as the filter name.

  • available-ip-address-count - The number of IP addresses in the subnet that are available.

  • cidrBlock - The CIDR block of the subnet. The CIDR block you specify must exactly match the subnet's CIDR block for information to be returned for the subnet. You can also use cidr or cidr-block as the filter names.

  • defaultForAz - Indicates whether this is the default subnet for the Availability Zone. You can also use default-for-az as the filter name.

  • state - The state of the subnet (pending | available).

  • subnet-id - The ID of the subnet.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the subnet.

", - "DescribeTagsRequest$Filters": "

One or more filters.

  • key - The tag key.

  • resource-id - The resource ID.

  • resource-type - The resource type (customer-gateway | dhcp-options | image | instance | internet-gateway | network-acl | network-interface | reserved-instances | route-table | security-group | snapshot | spot-instances-request | subnet | volume | vpc | vpn-connection | vpn-gateway).

  • value - The tag value.

", - "DescribeVolumeStatusRequest$Filters": "

One or more filters.

  • action.code - The action code for the event (for example, enable-volume-io).

  • action.description - A description of the action.

  • action.event-id - The event ID associated with the action.

  • availability-zone - The Availability Zone of the instance.

  • event.description - A description of the event.

  • event.event-id - The event ID.

  • event.event-type - The event type (for io-enabled: passed | failed; for io-performance: io-performance:degraded | io-performance:severely-degraded | io-performance:stalled).

  • event.not-after - The latest end time for the event.

  • event.not-before - The earliest start time for the event.

  • volume-status.details-name - The cause for volume-status.status (io-enabled | io-performance).

  • volume-status.details-status - The status of volume-status.details-name (for io-enabled: passed | failed; for io-performance: normal | degraded | severely-degraded | stalled).

  • volume-status.status - The status of the volume (ok | impaired | warning | insufficient-data).

", - "DescribeVolumesRequest$Filters": "

One or more filters.

  • attachment.attach-time - The time stamp when the attachment initiated.

  • attachment.delete-on-termination - Whether the volume is deleted on instance termination.

  • attachment.device - The device name that is exposed to the instance (for example, /dev/sda1).

  • attachment.instance-id - The ID of the instance the volume is attached to.

  • attachment.status - The attachment state (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone in which the volume was created.

  • create-time - The time stamp when the volume was created.

  • encrypted - The encryption status of the volume.

  • size - The size of the volume, in GiB.

  • snapshot-id - The snapshot from which the volume was created.

  • status - The status of the volume (creating | available | in-use | deleting | deleted | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The volume ID.

  • volume-type - The Amazon EBS volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

", - "DescribeVpcClassicLinkRequest$Filters": "

One or more filters.

  • is-classic-link-enabled - Whether the VPC is enabled for ClassicLink (true | false).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeVpcEndpointsRequest$Filters": "

One or more filters.

  • service-name: The name of the AWS service.

  • vpc-id: The ID of the VPC in which the endpoint resides.

  • vpc-endpoint-id: The ID of the endpoint.

  • vpc-endpoint-state: The state of the endpoint. (pending | available | deleting | deleted)

", - "DescribeVpcPeeringConnectionsRequest$Filters": "

One or more filters.

  • accepter-vpc-info.cidr-block - The CIDR block of the peer VPC.

  • accepter-vpc-info.owner-id - The AWS account ID of the owner of the peer VPC.

  • accepter-vpc-info.vpc-id - The ID of the peer VPC.

  • expiration-time - The expiration date and time for the VPC peering connection.

  • requester-vpc-info.cidr-block - The CIDR block of the requester's VPC.

  • requester-vpc-info.owner-id - The AWS account ID of the owner of the requester VPC.

  • requester-vpc-info.vpc-id - The ID of the requester VPC.

  • status-code - The status of the VPC peering connection (pending-acceptance | failed | expired | provisioning | active | deleted | rejected).

  • status-message - A message that provides more information about the status of the VPC peering connection, if applicable.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-peering-connection-id - The ID of the VPC peering connection.

", - "DescribeVpcsRequest$Filters": "

One or more filters.

  • cidr - The CIDR block of the VPC. The CIDR block you specify must exactly match the VPC's CIDR block for information to be returned for the VPC. Must contain the slash followed by one or two digits (for example, /28).

  • dhcp-options-id - The ID of a set of DHCP options.

  • isDefault - Indicates whether the VPC is the default VPC.

  • state - The state of the VPC (pending | available).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC.

", - "DescribeVpnConnectionsRequest$Filters": "

One or more filters.

  • customer-gateway-configuration - The configuration information for the customer gateway.

  • customer-gateway-id - The ID of a customer gateway associated with the VPN connection.

  • state - The state of the VPN connection (pending | available | deleting | deleted).

  • option.static-routes-only - Indicates whether the connection has static routes only. Used for devices that do not support Border Gateway Protocol (BGP).

  • route.destination-cidr-block - The destination CIDR block. This corresponds to the subnet used in a customer data center.

  • bgp-asn - The BGP Autonomous System Number (ASN) associated with a BGP device.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of VPN connection. Currently the only supported type is ipsec.1.

  • vpn-connection-id - The ID of the VPN connection.

  • vpn-gateway-id - The ID of a virtual private gateway associated with the VPN connection.

", - "DescribeVpnGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (attaching | attached | detaching | detached).

  • attachment.vpc-id - The ID of an attached VPC.

  • availability-zone - The Availability Zone for the virtual private gateway (if applicable).

  • state - The state of the virtual private gateway (pending | available | deleting | deleted).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of virtual private gateway. Currently the only supported type is ipsec.1.

  • vpn-gateway-id - The ID of the virtual private gateway.

" - } - }, - "FleetType": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$Type": "

The type of request. Indicates whether the fleet will only request the target capacity or also attempt to maintain it. When you request a certain target capacity, the fleet will only place the required bids. It will not attempt to replenish Spot instances if capacity is diminished, nor will it submit bids in alternative Spot pools if capacity is not available. When you want to maintain a certain target capacity, fleet will place the required bids to meet this target capacity. It will also automatically replenish any interrupted instances. Default: maintain.

" - } - }, - "Float": { - "base": null, - "refs": { - "ReservedInstances$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstances$FixedPrice": "

The purchase price of the Reserved Instance.

", - "ReservedInstancesOffering$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstancesOffering$FixedPrice": "

The purchase price of the Reserved Instance.

" - } - }, - "FlowLog": { - "base": "

Describes a flow log.

", - "refs": { - "FlowLogSet$member": null - } - }, - "FlowLogSet": { - "base": null, - "refs": { - "DescribeFlowLogsResult$FlowLogs": "

Information about the flow logs.

" - } - }, - "FlowLogsResourceType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$ResourceType": "

The type of resource on which to create the flow log.

" - } - }, - "GatewayType": { - "base": null, - "refs": { - "CreateCustomerGatewayRequest$Type": "

The type of VPN connection that this customer gateway supports (ipsec.1).

", - "CreateVpnGatewayRequest$Type": "

The type of VPN connection this virtual private gateway supports.

", - "VpnConnection$Type": "

The type of VPN connection.

", - "VpnGateway$Type": "

The type of VPN connection the virtual private gateway supports.

" - } - }, - "GetConsoleOutputRequest": { - "base": "

Contains the parameters for GetConsoleOutput.

", - "refs": { - } - }, - "GetConsoleOutputResult": { - "base": "

Contains the output of GetConsoleOutput.

", - "refs": { - } - }, - "GetConsoleScreenshotRequest": { - "base": "

Contains the parameters for the request.

", - "refs": { - } - }, - "GetConsoleScreenshotResult": { - "base": "

Contains the output of the request.

", - "refs": { - } - }, - "GetPasswordDataRequest": { - "base": "

Contains the parameters for GetPasswordData.

", - "refs": { - } - }, - "GetPasswordDataResult": { - "base": "

Contains the output of GetPasswordData.

", - "refs": { - } - }, - "GroupIdStringList": { - "base": null, - "refs": { - "AttachClassicLinkVpcRequest$Groups": "

The ID of one or more of the VPC's security groups. You cannot specify security groups from a different VPC.

", - "DescribeSecurityGroupsRequest$GroupIds": "

One or more security group IDs. Required for security groups in a nondefault VPC.

Default: Describes all your security groups.

", - "ModifyInstanceAttributeRequest$Groups": "

[EC2-VPC] Changes the security groups of the instance. You must specify at least one security group, even if it's just the default security group for the VPC. You must specify the security group ID, not the security group name.

" - } - }, - "GroupIdentifier": { - "base": "

Describes a security group.

", - "refs": { - "GroupIdentifierList$member": null - } - }, - "GroupIdentifierList": { - "base": null, - "refs": { - "ClassicLinkInstance$Groups": "

A list of security groups.

", - "DescribeNetworkInterfaceAttributeResult$Groups": "

The security groups associated with the network interface.

", - "Instance$SecurityGroups": "

One or more security groups for the instance.

", - "InstanceAttribute$Groups": "

The security groups associated with the instance.

", - "InstanceNetworkInterface$Groups": "

One or more security groups.

", - "LaunchSpecification$SecurityGroups": "

One or more security groups. When requesting instances in a VPC, you must specify the IDs of the security groups. When requesting instances in EC2-Classic, you can specify the names or the IDs of the security groups.

", - "NetworkInterface$Groups": "

Any security groups for the network interface.

", - "Reservation$Groups": "

[EC2-Classic only] One or more security groups.

", - "SpotFleetLaunchSpecification$SecurityGroups": "

One or more security groups. When requesting instances in a VPC, you must specify the IDs of the security groups. When requesting instances in EC2-Classic, you can specify the names or the IDs of the security groups.

" - } - }, - "GroupIds": { - "base": null, - "refs": { - "DescribeSecurityGroupReferencesRequest$GroupId": "

One or more security group IDs in your account.

" - } - }, - "GroupNameStringList": { - "base": null, - "refs": { - "DescribeSecurityGroupsRequest$GroupNames": "

[EC2-Classic and default VPC only] One or more security group names. You can specify either the security group name or the security group ID. For security groups in a nondefault VPC, use the group-name filter to describe security groups by name.

Default: Describes all your security groups.

", - "ModifySnapshotAttributeRequest$GroupNames": "

The group to modify for the snapshot.

" - } - }, - "HistoryRecord": { - "base": "

Describes an event in the history of the Spot fleet request.

", - "refs": { - "HistoryRecords$member": null - } - }, - "HistoryRecords": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryResponse$HistoryRecords": "

Information about the events in the history of the Spot fleet request.

" - } - }, - "Host": { - "base": "

Describes the properties of the Dedicated host.

", - "refs": { - "HostList$member": null - } - }, - "HostInstance": { - "base": "

Describes an instance running on a Dedicated host.

", - "refs": { - "HostInstanceList$member": null - } - }, - "HostInstanceList": { - "base": null, - "refs": { - "Host$Instances": "

The IDs and instance type that are currently running on the Dedicated host.

" - } - }, - "HostList": { - "base": null, - "refs": { - "DescribeHostsResult$Hosts": "

Information about the Dedicated hosts.

" - } - }, - "HostProperties": { - "base": "

Describes properties of a Dedicated host.

", - "refs": { - "Host$HostProperties": "

The hardware specifications of the Dedicated host.

" - } - }, - "HostTenancy": { - "base": null, - "refs": { - "ModifyInstancePlacementRequest$Tenancy": "

The tenancy of the instance that you are modifying.

" - } - }, - "HypervisorType": { - "base": null, - "refs": { - "Image$Hypervisor": "

The hypervisor type of the image.

", - "Instance$Hypervisor": "

The hypervisor type of the instance.

" - } - }, - "IamInstanceProfile": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "Instance$IamInstanceProfile": "

The IAM instance profile associated with the instance, if applicable.

" - } - }, - "IamInstanceProfileSpecification": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "LaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RequestSpotLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RunInstancesRequest$IamInstanceProfile": "

The IAM instance profile.

", - "SpotFleetLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "IcmpTypeCode": { - "base": "

Describes the ICMP type and code.

", - "refs": { - "CreateNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying ICMP for the protocol.

", - "NetworkAclEntry$IcmpTypeCode": "

ICMP protocol: The ICMP type and code.

", - "ReplaceNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying 1 (ICMP) for the protocol.

" - } - }, - "IdFormat": { - "base": "

Describes the ID format for a resource.

", - "refs": { - "IdFormatList$member": null - } - }, - "IdFormatList": { - "base": null, - "refs": { - "DescribeIdFormatResult$Statuses": "

Information about the ID format for the resource.

" - } - }, - "Image": { - "base": "

Describes an image.

", - "refs": { - "ImageList$member": null - } - }, - "ImageAttribute": { - "base": "

Describes an image attribute.

", - "refs": { - } - }, - "ImageAttributeName": { - "base": null, - "refs": { - "DescribeImageAttributeRequest$Attribute": "

The AMI attribute.

Note: Depending on your account privileges, the blockDeviceMapping attribute may return a Client.AuthFailure error. If this happens, use DescribeImages to get information about the block device mapping for the AMI.

" - } - }, - "ImageDiskContainer": { - "base": "

Describes the disk container object for an import image task.

", - "refs": { - "ImageDiskContainerList$member": null - } - }, - "ImageDiskContainerList": { - "base": null, - "refs": { - "ImportImageRequest$DiskContainers": "

Information about the disk containers.

" - } - }, - "ImageIdStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ImageIds": "

One or more image IDs.

Default: Describes all images available to you.

" - } - }, - "ImageList": { - "base": null, - "refs": { - "DescribeImagesResult$Images": "

Information about one or more images.

" - } - }, - "ImageState": { - "base": null, - "refs": { - "Image$State": "

The current state of the AMI. If the state is available, the image is successfully registered and can be used to launch an instance.

" - } - }, - "ImageTypeValues": { - "base": null, - "refs": { - "Image$ImageType": "

The type of image.

" - } - }, - "ImportImageRequest": { - "base": "

Contains the parameters for ImportImage.

", - "refs": { - } - }, - "ImportImageResult": { - "base": "

Contains the output for ImportImage.

", - "refs": { - } - }, - "ImportImageTask": { - "base": "

Describes an import image task.

", - "refs": { - "ImportImageTaskList$member": null - } - }, - "ImportImageTaskList": { - "base": null, - "refs": { - "DescribeImportImageTasksResult$ImportImageTasks": "

A list of zero or more import image tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportInstanceLaunchSpecification": { - "base": "

Describes the launch specification for VM import.

", - "refs": { - "ImportInstanceRequest$LaunchSpecification": "

The launch specification.

" - } - }, - "ImportInstanceRequest": { - "base": "

Contains the parameters for ImportInstance.

", - "refs": { - } - }, - "ImportInstanceResult": { - "base": "

Contains the output for ImportInstance.

", - "refs": { - } - }, - "ImportInstanceTaskDetails": { - "base": "

Describes an import instance task.

", - "refs": { - "ConversionTask$ImportInstance": "

If the task is for importing an instance, this contains information about the import instance task.

" - } - }, - "ImportInstanceVolumeDetailItem": { - "base": "

Describes an import volume task.

", - "refs": { - "ImportInstanceVolumeDetailSet$member": null - } - }, - "ImportInstanceVolumeDetailSet": { - "base": null, - "refs": { - "ImportInstanceTaskDetails$Volumes": "

One or more volumes.

" - } - }, - "ImportKeyPairRequest": { - "base": "

Contains the parameters for ImportKeyPair.

", - "refs": { - } - }, - "ImportKeyPairResult": { - "base": "

Contains the output of ImportKeyPair.

", - "refs": { - } - }, - "ImportSnapshotRequest": { - "base": "

Contains the parameters for ImportSnapshot.

", - "refs": { - } - }, - "ImportSnapshotResult": { - "base": "

Contains the output for ImportSnapshot.

", - "refs": { - } - }, - "ImportSnapshotTask": { - "base": "

Describes an import snapshot task.

", - "refs": { - "ImportSnapshotTaskList$member": null - } - }, - "ImportSnapshotTaskList": { - "base": null, - "refs": { - "DescribeImportSnapshotTasksResult$ImportSnapshotTasks": "

A list of zero or more import snapshot tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportTaskIdList": { - "base": null, - "refs": { - "DescribeImportImageTasksRequest$ImportTaskIds": "

A list of import image task IDs.

", - "DescribeImportSnapshotTasksRequest$ImportTaskIds": "

A list of import snapshot task IDs.

" - } - }, - "ImportVolumeRequest": { - "base": "

Contains the parameters for ImportVolume.

", - "refs": { - } - }, - "ImportVolumeResult": { - "base": "

Contains the output for ImportVolume.

", - "refs": { - } - }, - "ImportVolumeTaskDetails": { - "base": "

Describes an import volume task.

", - "refs": { - "ConversionTask$ImportVolume": "

If the task is for importing a volume, this contains information about the import volume task.

" - } - }, - "Instance": { - "base": "

Describes an instance.

", - "refs": { - "InstanceList$member": null - } - }, - "InstanceAttribute": { - "base": "

Describes an instance attribute.

", - "refs": { - } - }, - "InstanceAttributeName": { - "base": null, - "refs": { - "DescribeInstanceAttributeRequest$Attribute": "

The instance attribute.

", - "ModifyInstanceAttributeRequest$Attribute": "

The name of the attribute.

", - "ResetInstanceAttributeRequest$Attribute": "

The attribute to reset.

You can only reset the following attributes: kernel | ramdisk | sourceDestCheck. To change an instance attribute, use ModifyInstanceAttribute.

" - } - }, - "InstanceBlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingList$member": null - } - }, - "InstanceBlockDeviceMappingList": { - "base": null, - "refs": { - "Instance$BlockDeviceMappings": "

Any block device mapping entries for the instance.

", - "InstanceAttribute$BlockDeviceMappings": "

The block device mapping of the instance.

" - } - }, - "InstanceBlockDeviceMappingSpecification": { - "base": "

Describes a block device mapping entry.

", - "refs": { - "InstanceBlockDeviceMappingSpecificationList$member": null - } - }, - "InstanceBlockDeviceMappingSpecificationList": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$BlockDeviceMappings": "

Modifies the DeleteOnTermination attribute for volumes that are currently attached. The volume must be owned by the caller. If no value is specified for DeleteOnTermination, the default is true and the volume is deleted when the instance is terminated.

To add instance store volumes to an Amazon EBS-backed instance, you must add them when you launch the instance. For more information, see Updating the Block Device Mapping when Launching an Instance in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "InstanceCapacity": { - "base": "

Information about the instance type that the Dedicated host supports.

", - "refs": { - "AvailableInstanceCapacityList$member": null - } - }, - "InstanceCount": { - "base": "

Describes a Reserved Instance listing state.

", - "refs": { - "InstanceCountList$member": null - } - }, - "InstanceCountList": { - "base": null, - "refs": { - "ReservedInstancesListing$InstanceCounts": "

The number of instances in this state.

" - } - }, - "InstanceExportDetails": { - "base": "

Describes an instance to export.

", - "refs": { - "ExportTask$InstanceExportDetails": "

Information about the instance to export.

" - } - }, - "InstanceIdSet": { - "base": null, - "refs": { - "RunScheduledInstancesResult$InstanceIdSet": "

The IDs of the newly launched instances.

" - } - }, - "InstanceIdStringList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesRequest$InstanceIds": "

One or more instance IDs. Must be instances linked to a VPC through ClassicLink.

", - "DescribeInstanceStatusRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

Constraints: Maximum 100 explicitly specified instance IDs.

", - "DescribeInstancesRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

", - "MonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "RebootInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "ReportInstanceStatusRequest$Instances": "

One or more instances.

", - "StartInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "StopInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "TerminateInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "UnmonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

" - } - }, - "InstanceLifecycleType": { - "base": null, - "refs": { - "Instance$InstanceLifecycle": "

Indicates whether this is a Spot instance or a Scheduled Instance.

" - } - }, - "InstanceList": { - "base": null, - "refs": { - "Reservation$Instances": "

One or more instances.

" - } - }, - "InstanceMonitoring": { - "base": "

Describes the monitoring information of the instance.

", - "refs": { - "InstanceMonitoringList$member": null - } - }, - "InstanceMonitoringList": { - "base": null, - "refs": { - "MonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

", - "UnmonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

" - } - }, - "InstanceNetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceList$member": null - } - }, - "InstanceNetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "InstanceNetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "InstancePrivateIpAddress$Association": "

The association information for an Elastic IP address for the network interface.

" - } - }, - "InstanceNetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "InstanceNetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "InstanceNetworkInterfaceList": { - "base": null, - "refs": { - "Instance$NetworkInterfaces": "

[EC2-VPC] One or more network interfaces for the instance.

" - } - }, - "InstanceNetworkInterfaceSpecification": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceSpecificationList$member": null - } - }, - "InstanceNetworkInterfaceSpecificationList": { - "base": null, - "refs": { - "LaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RequestSpotLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RunInstancesRequest$NetworkInterfaces": "

One or more network interfaces.

", - "SpotFleetLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "InstancePrivateIpAddress": { - "base": "

Describes a private IP address.

", - "refs": { - "InstancePrivateIpAddressList$member": null - } - }, - "InstancePrivateIpAddressList": { - "base": null, - "refs": { - "InstanceNetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "InstanceState": { - "base": "

Describes the current state of the instance.

", - "refs": { - "Instance$State": "

The current state of the instance.

", - "InstanceStateChange$CurrentState": "

The current state of the instance.

", - "InstanceStateChange$PreviousState": "

The previous state of the instance.

", - "InstanceStatus$InstanceState": "

The intended state of the instance. DescribeInstanceStatus requires that an instance be in the running state.

" - } - }, - "InstanceStateChange": { - "base": "

Describes an instance state change.

", - "refs": { - "InstanceStateChangeList$member": null - } - }, - "InstanceStateChangeList": { - "base": null, - "refs": { - "StartInstancesResult$StartingInstances": "

Information about one or more started instances.

", - "StopInstancesResult$StoppingInstances": "

Information about one or more stopped instances.

", - "TerminateInstancesResult$TerminatingInstances": "

Information about one or more terminated instances.

" - } - }, - "InstanceStateName": { - "base": null, - "refs": { - "InstanceState$Name": "

The current state of the instance.

" - } - }, - "InstanceStatus": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatusList$member": null - } - }, - "InstanceStatusDetails": { - "base": "

Describes the instance status.

", - "refs": { - "InstanceStatusDetailsList$member": null - } - }, - "InstanceStatusDetailsList": { - "base": null, - "refs": { - "InstanceStatusSummary$Details": "

The system instance health or application instance health.

" - } - }, - "InstanceStatusEvent": { - "base": "

Describes a scheduled event for an instance.

", - "refs": { - "InstanceStatusEventList$member": null - } - }, - "InstanceStatusEventList": { - "base": null, - "refs": { - "InstanceStatus$Events": "

Any scheduled events associated with the instance.

" - } - }, - "InstanceStatusList": { - "base": null, - "refs": { - "DescribeInstanceStatusResult$InstanceStatuses": "

One or more instance status descriptions.

" - } - }, - "InstanceStatusSummary": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatus$SystemStatus": "

Reports impaired functionality that stems from issues related to the systems that support an instance, such as hardware failures and network connectivity problems.

", - "InstanceStatus$InstanceStatus": "

Reports impaired functionality that stems from issues internal to the instance, such as impaired reachability.

" - } - }, - "InstanceType": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$InstanceType": "

The instance type that the reservation will cover (for example, m1.small). For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

", - "ImportInstanceLaunchSpecification$InstanceType": "

The instance type. For more information about the instance types that you can import, see Before You Get Started in the Amazon Elastic Compute Cloud User Guide.

", - "Instance$InstanceType": "

The instance type.

", - "InstanceTypeList$member": null, - "LaunchSpecification$InstanceType": "

The instance type.

", - "RequestSpotLaunchSpecification$InstanceType": "

The instance type.

", - "ReservedInstances$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$InstanceType": "

The instance type for the modified Reserved Instances.

", - "ReservedInstancesOffering$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "RunInstancesRequest$InstanceType": "

The instance type. For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

Default: m1.small

", - "SpotFleetLaunchSpecification$InstanceType": "

The instance type.

", - "SpotPrice$InstanceType": "

The instance type.

" - } - }, - "InstanceTypeList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$InstanceTypes": "

Filters the results by the specified instance types.

" - } - }, - "Integer": { - "base": null, - "refs": { - "AllocateHostsRequest$Quantity": "

The number of Dedicated hosts you want to allocate to your account with these parameters.

", - "AssignPrivateIpAddressesRequest$SecondaryPrivateIpAddressCount": "

The number of secondary IP addresses to assign to the network interface. You can't specify this parameter when also specifying private IP addresses.

", - "AttachNetworkInterfaceRequest$DeviceIndex": "

The index of the device for the network interface attachment.

", - "AuthorizeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "AuthorizeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "AvailableCapacity$AvailableVCpus": "

The number of vCPUs available on the Dedicated host.

", - "CreateCustomerGatewayRequest$BgpAsn": "

For devices that support BGP, the customer gateway's BGP ASN.

Default: 65000

", - "CreateNetworkAclEntryRequest$RuleNumber": "

The rule number for the entry (for example, 100). ACL entries are processed in ascending order by rule number.

Constraints: Positive integer from 1 to 32766. The range 32767 to 65535 is reserved for internal use.

", - "CreateNetworkInterfaceRequest$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses to assign to a network interface. When you specify a number of secondary IP addresses, Amazon EC2 selects these IP addresses within the subnet range. You can't specify this option and specify more than one private IP address using privateIpAddresses.

The number of IP addresses you can assign to a network interface varies by instance type. For more information, see Private IP Addresses Per ENI Per Instance Type in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListingRequest$InstanceCount": "

The number of instances that are a part of a Reserved Instance account to be listed in the Reserved Instance Marketplace. This number should be less than or equal to the instance count associated with the Reserved Instance ID specified in this call.

", - "CreateVolumeRequest$Size": "

The size of the volume, in GiBs.

Constraints: 1-16384 for gp2, 4-16384 for io1, 500-16384 for st1, 500-16384 for sc1, and 1-1024 for standard. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "CreateVolumeRequest$Iops": "

Only valid for Provisioned IOPS SSD volumes. The number of I/O operations per second (IOPS) to provision for the volume, with a maximum ratio of 30 IOPS/GiB.

Constraint: Range is 100 to 20000 for Provisioned IOPS SSD volumes

", - "DeleteNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to delete.

", - "DescribeClassicLinkInstancesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the instance IDs parameter in the same request.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeFlowLogsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the flow log IDs parameter in the same request.

", - "DescribeHostsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error. You cannot specify this parameter and the host IDs parameter in the same request.

", - "DescribeImportImageTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeImportSnapshotTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeInstanceStatusRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter in the same call.

", - "DescribeInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter or tag filters in the same call.

", - "DescribeMovingAddressesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value outside of this range, an error is returned.

Default: If no value is provided, the default is 1000.

", - "DescribeNatGatewaysRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribePrefixListsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribeReservedInstancesOfferingsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. The maximum is 100.

Default: 100

", - "DescribeReservedInstancesOfferingsRequest$MaxInstanceCount": "

The maximum number of instances to filter when searching for offerings.

Default: 20

", - "DescribeScheduledInstanceAvailabilityRequest$MinSlotDurationInHours": "

The minimum available duration, in hours. The minimum required duration is 1,200 hours per year. For example, the minimum daily schedule is 4 hours, the minimum weekly schedule is 24 hours, and the minimum monthly schedule is 100 hours.

", - "DescribeScheduledInstanceAvailabilityRequest$MaxSlotDurationInHours": "

The maximum available duration, in hours. This value must be greater than MinSlotDurationInHours and less than 1,720.

", - "DescribeScheduledInstanceAvailabilityRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 300. The default value is 300. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeScheduledInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 300. The default value is 100. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSnapshotsRequest$MaxResults": "

The maximum number of snapshot results returned by DescribeSnapshots in paginated output. When this parameter is used, DescribeSnapshots only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeSnapshots request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeSnapshots returns all results. You cannot specify this parameter and the snapshot IDs parameter in the same request.

", - "DescribeSpotFleetInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestsRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotPriceHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeTagsRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeVolumeStatusRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumeStatus in paginated output. When this parameter is used, the request only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumeStatus returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVolumesRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumes in paginated output. When this parameter is used, DescribeVolumes only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeVolumes request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumes returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVpcEndpointServicesRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeVpcEndpointsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "EbsBlockDevice$VolumeSize": "

The size of the volume, in GiB.

Constraints: 1-16384 for General Purpose SSD (gp2), 4-16384 for Provisioned IOPS SSD (io1), 500-16384 for Throughput Optimized HDD (st1), 500-16384 for Cold HDD (sc1), and 1-1024 for Magnetic (standard) volumes. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "EbsBlockDevice$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For io1, this represents the number of IOPS that are provisioned for the volume. For gp2, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information on General Purpose SSD baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

", - "HostProperties$Sockets": "

The number of sockets on the Dedicated host.

", - "HostProperties$Cores": "

The number of cores on the Dedicated host.

", - "HostProperties$TotalVCpus": "

The number of vCPUs on the Dedicated host.

", - "IcmpTypeCode$Type": "

The ICMP code. A value of -1 means all codes for the specified ICMP type.

", - "IcmpTypeCode$Code": "

The ICMP type. A value of -1 means all types.

", - "Instance$AmiLaunchIndex": "

The AMI launch index, which can be used to find this instance in the launch group.

", - "InstanceCapacity$AvailableCapacity": "

The number of instances that can still be launched onto the Dedicated host.

", - "InstanceCapacity$TotalCapacity": "

The total number of instances that can be launched onto the Dedicated host.

", - "InstanceCount$InstanceCount": "

The number of listed Reserved Instances in the state specified by the state.

", - "InstanceNetworkInterfaceAttachment$DeviceIndex": "

The index of the device on the instance for the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$DeviceIndex": "

The index of the device on the instance for the network interface attachment. If you are specifying a network interface in a RunInstances request, you must provide the device index.

", - "InstanceNetworkInterfaceSpecification$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses. You can't specify this option and specify more than one private IP address using the private IP addresses option.

", - "InstanceState$Code": "

The low byte represents the state. The high byte is an opaque internal value and should be ignored.

  • 0 : pending

  • 16 : running

  • 32 : shutting-down

  • 48 : terminated

  • 64 : stopping

  • 80 : stopped

", - "IpPermission$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "IpPermission$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code. A value of -1 indicates all ICMP codes for the specified ICMP type.

", - "ModifySpotFleetRequestRequest$TargetCapacity": "

The size of the fleet.

", - "NetworkAclEntry$RuleNumber": "

The rule number for the entry. ACL entries are processed in ascending order by rule number.

", - "NetworkInterfaceAttachment$DeviceIndex": "

The device index of the network interface attachment on the instance.

", - "OccurrenceDayRequestSet$member": null, - "OccurrenceDaySet$member": null, - "PortRange$From": "

The first port in the range.

", - "PortRange$To": "

The last port in the range.

", - "PricingDetail$Count": "

The number of reservations available for the price.

", - "PurchaseRequest$InstanceCount": "

The number of instances.

", - "PurchaseReservedInstancesOfferingRequest$InstanceCount": "

The number of Reserved Instances to purchase.

", - "ReplaceNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to replace.

", - "RequestSpotInstancesRequest$InstanceCount": "

The maximum number of Spot instances to launch.

Default: 1

", - "RequestSpotInstancesRequest$BlockDurationMinutes": "

The required duration for the Spot instances (also known as Spot blocks), in minutes. This value must be a multiple of 60 (60, 120, 180, 240, 300, or 360).

The duration period starts as soon as your Spot instance receives its instance ID. At the end of the duration period, Amazon EC2 marks the Spot instance for termination and provides a Spot instance termination notice, which gives the instance a two-minute warning before it terminates.

Note that you can't specify an Availability Zone group or a launch group if you specify a duration.

", - "ReservedInstances$InstanceCount": "

The number of reservations purchased.

", - "ReservedInstancesConfiguration$InstanceCount": "

The number of modified Reserved Instances.

", - "RevokeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "RevokeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "RunInstancesRequest$MinCount": "

The minimum number of instances to launch. If you specify a minimum that is more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches no instances.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 General FAQ.

", - "RunInstancesRequest$MaxCount": "

The maximum number of instances to launch. If you specify more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches the largest possible number of instances above MinCount.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 FAQ.

", - "RunScheduledInstancesRequest$InstanceCount": "

The number of instances.

Default: 1

", - "ScheduledInstance$SlotDurationInHours": "

The number of hours in the schedule.

", - "ScheduledInstance$TotalScheduledInstanceHours": "

The total number of hours for a single instance for the entire term.

", - "ScheduledInstance$InstanceCount": "

The number of instances.

", - "ScheduledInstanceAvailability$SlotDurationInHours": "

The number of hours in the schedule.

", - "ScheduledInstanceAvailability$TotalScheduledInstanceHours": "

The total number of hours for a single instance for the entire term.

", - "ScheduledInstanceAvailability$AvailableInstanceCount": "

The number of available instances.

", - "ScheduledInstanceAvailability$MinTermDurationInDays": "

The minimum term. The only possible value is 365 days.

", - "ScheduledInstanceAvailability$MaxTermDurationInDays": "

The maximum term. The only possible value is 365 days.

", - "ScheduledInstanceRecurrence$Interval": "

The interval quantity. The interval unit depends on the value of frequency. For example, every 2 weeks or every 2 months.

", - "ScheduledInstanceRecurrenceRequest$Interval": "

The interval quantity. The interval unit depends on the value of Frequency. For example, every 2 weeks or every 2 months.

", - "ScheduledInstancesEbs$VolumeSize": "

The size of the volume, in GiB.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "ScheduledInstancesEbs$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For io1 volumes, this represents the number of IOPS that are provisioned for the volume. For gp2 volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information about gp2 baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

", - "ScheduledInstancesNetworkInterface$DeviceIndex": "

The index of the device for the network interface attachment.

", - "ScheduledInstancesNetworkInterface$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses.

", - "Snapshot$VolumeSize": "

The size of the volume, in GiB.

", - "SpotFleetRequestConfigData$TargetCapacity": "

The number of units to request. You can choose to set the target capacity in terms of instances or a performance characteristic that is important to your application workload, such as vCPUs, memory, or I/O.

", - "SpotInstanceRequest$BlockDurationMinutes": "

The duration for the Spot instance, in minutes.

", - "StaleIpPermission$FromPort": "

The start of the port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "StaleIpPermission$ToPort": "

The end of the port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "Subnet$AvailableIpAddressCount": "

The number of unused IP addresses in the subnet. Note that the IP addresses for any stopped instances are considered unavailable.

", - "VgwTelemetry$AcceptedRouteCount": "

The number of accepted routes.

", - "Volume$Size": "

The size of the volume, in GiBs.

", - "Volume$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For Provisioned IOPS SSD volumes, this represents the number of IOPS that are provisioned for the volume. For General Purpose SSD volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information on General Purpose SSD baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

" - } - }, - "InternetGateway": { - "base": "

Describes an Internet gateway.

", - "refs": { - "CreateInternetGatewayResult$InternetGateway": "

Information about the Internet gateway.

", - "InternetGatewayList$member": null - } - }, - "InternetGatewayAttachment": { - "base": "

Describes the attachment of a VPC to an Internet gateway.

", - "refs": { - "InternetGatewayAttachmentList$member": null - } - }, - "InternetGatewayAttachmentList": { - "base": null, - "refs": { - "InternetGateway$Attachments": "

Any VPCs attached to the Internet gateway.

" - } - }, - "InternetGatewayList": { - "base": null, - "refs": { - "DescribeInternetGatewaysResult$InternetGateways": "

Information about one or more Internet gateways.

" - } - }, - "IpPermission": { - "base": "

Describes a security group rule.

", - "refs": { - "IpPermissionList$member": null - } - }, - "IpPermissionList": { - "base": null, - "refs": { - "AuthorizeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "AuthorizeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. Can be used to specify multiple rules in a single command.

", - "RevokeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "RevokeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. You can't specify a source security group and a CIDR IP address range.

", - "SecurityGroup$IpPermissions": "

One or more inbound rules associated with the security group.

", - "SecurityGroup$IpPermissionsEgress": "

[EC2-VPC] One or more outbound rules associated with the security group.

" - } - }, - "IpRange": { - "base": "

Describes an IP range.

", - "refs": { - "IpRangeList$member": null - } - }, - "IpRangeList": { - "base": null, - "refs": { - "IpPermission$IpRanges": "

One or more IP ranges.

" - } - }, - "IpRanges": { - "base": null, - "refs": { - "StaleIpPermission$IpRanges": "

One or more IP ranges. Not applicable for stale security group rules.

" - } - }, - "KeyNameStringList": { - "base": null, - "refs": { - "DescribeKeyPairsRequest$KeyNames": "

One or more key pair names.

Default: Describes all your key pairs.

" - } - }, - "KeyPair": { - "base": "

Describes a key pair.

", - "refs": { - } - }, - "KeyPairInfo": { - "base": "

Describes a key pair.

", - "refs": { - "KeyPairList$member": null - } - }, - "KeyPairList": { - "base": null, - "refs": { - "DescribeKeyPairsResult$KeyPairs": "

Information about one or more key pairs.

" - } - }, - "LaunchPermission": { - "base": "

Describes a launch permission.

", - "refs": { - "LaunchPermissionList$member": null - } - }, - "LaunchPermissionList": { - "base": null, - "refs": { - "ImageAttribute$LaunchPermissions": "

One or more launch permissions.

", - "LaunchPermissionModifications$Add": "

The AWS account ID to add to the list of launch permissions for the AMI.

", - "LaunchPermissionModifications$Remove": "

The AWS account ID to remove from the list of launch permissions for the AMI.

" - } - }, - "LaunchPermissionModifications": { - "base": "

Describes a launch permission modification.

", - "refs": { - "ModifyImageAttributeRequest$LaunchPermission": "

A launch permission modification.

" - } - }, - "LaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "SpotInstanceRequest$LaunchSpecification": "

Additional information for launching instances.

" - } - }, - "LaunchSpecsList": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$LaunchSpecifications": "

Information about the launch specifications for the Spot fleet request.

" - } - }, - "ListingState": { - "base": null, - "refs": { - "InstanceCount$State": "

The states of the listed Reserved Instances.

" - } - }, - "ListingStatus": { - "base": null, - "refs": { - "ReservedInstancesListing$Status": "

The status of the Reserved Instance listing.

" - } - }, - "Long": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$MinDuration": "

The minimum duration (in seconds) to filter when searching for offerings.

Default: 2592000 (1 month)

", - "DescribeReservedInstancesOfferingsRequest$MaxDuration": "

The maximum duration (in seconds) to filter when searching for offerings.

Default: 94608000 (3 years)

", - "DiskImageDescription$Size": "

The size of the disk image, in GiB.

", - "DiskImageDetail$Bytes": "

The size of the disk image, in GiB.

", - "DiskImageVolumeDescription$Size": "

The size of the volume, in GiB.

", - "ImportInstanceVolumeDetailItem$BytesConverted": "

The number of bytes converted so far.

", - "ImportVolumeTaskDetails$BytesConverted": "

The number of bytes converted so far.

", - "PriceSchedule$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "PriceScheduleSpecification$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "ReservedInstances$Duration": "

The duration of the Reserved Instance, in seconds.

", - "ReservedInstancesOffering$Duration": "

The duration of the Reserved Instance, in seconds.

", - "VolumeDetail$Size": "

The size of the volume, in GiB.

" - } - }, - "MaxResults": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeVpcClassicLinkDnsSupportRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

" - } - }, - "ModifyHostsRequest": { - "base": "

Contains the parameters for ModifyHosts.

", - "refs": { - } - }, - "ModifyHostsResult": { - "base": "

Contains the output of ModifyHosts.

", - "refs": { - } - }, - "ModifyIdFormatRequest": { - "base": "

Contains the parameters of ModifyIdFormat.

", - "refs": { - } - }, - "ModifyImageAttributeRequest": { - "base": "

Contains the parameters for ModifyImageAttribute.

", - "refs": { - } - }, - "ModifyInstanceAttributeRequest": { - "base": "

Contains the parameters for ModifyInstanceAttribute.

", - "refs": { - } - }, - "ModifyInstancePlacementRequest": { - "base": "

Contains the parameters for ModifyInstancePlacement.

", - "refs": { - } - }, - "ModifyInstancePlacementResult": { - "base": "

Contains the output of ModifyInstancePlacement.

", - "refs": { - } - }, - "ModifyNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for ModifyNetworkInterfaceAttribute.

", - "refs": { - } - }, - "ModifyReservedInstancesRequest": { - "base": "

Contains the parameters for ModifyReservedInstances.

", - "refs": { - } - }, - "ModifyReservedInstancesResult": { - "base": "

Contains the output of ModifyReservedInstances.

", - "refs": { - } - }, - "ModifySnapshotAttributeRequest": { - "base": "

Contains the parameters for ModifySnapshotAttribute.

", - "refs": { - } - }, - "ModifySpotFleetRequestRequest": { - "base": "

Contains the parameters for ModifySpotFleetRequest.

", - "refs": { - } - }, - "ModifySpotFleetRequestResponse": { - "base": "

Contains the output of ModifySpotFleetRequest.

", - "refs": { - } - }, - "ModifySubnetAttributeRequest": { - "base": "

Contains the parameters for ModifySubnetAttribute.

", - "refs": { - } - }, - "ModifyVolumeAttributeRequest": { - "base": "

Contains the parameters for ModifyVolumeAttribute.

", - "refs": { - } - }, - "ModifyVpcAttributeRequest": { - "base": "

Contains the parameters for ModifyVpcAttribute.

", - "refs": { - } - }, - "ModifyVpcEndpointRequest": { - "base": "

Contains the parameters for ModifyVpcEndpoint.

", - "refs": { - } - }, - "ModifyVpcEndpointResult": { - "base": "

Contains the output of ModifyVpcEndpoint.

", - "refs": { - } - }, - "ModifyVpcPeeringConnectionOptionsRequest": { - "base": null, - "refs": { - } - }, - "ModifyVpcPeeringConnectionOptionsResult": { - "base": null, - "refs": { - } - }, - "MonitorInstancesRequest": { - "base": "

Contains the parameters for MonitorInstances.

", - "refs": { - } - }, - "MonitorInstancesResult": { - "base": "

Contains the output of MonitorInstances.

", - "refs": { - } - }, - "Monitoring": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "Instance$Monitoring": "

The monitoring information for the instance.

", - "InstanceMonitoring$Monitoring": "

The monitoring information.

" - } - }, - "MonitoringState": { - "base": null, - "refs": { - "Monitoring$State": "

Indicates whether monitoring is enabled for the instance.

" - } - }, - "MoveAddressToVpcRequest": { - "base": "

Contains the parameters for MoveAddressToVpc.

", - "refs": { - } - }, - "MoveAddressToVpcResult": { - "base": "

Contains the output of MoveAddressToVpc.

", - "refs": { - } - }, - "MoveStatus": { - "base": null, - "refs": { - "MovingAddressStatus$MoveStatus": "

The status of the Elastic IP address that's being moved to the EC2-VPC platform, or restored to the EC2-Classic platform.

" - } - }, - "MovingAddressStatus": { - "base": "

Describes the status of a moving Elastic IP address.

", - "refs": { - "MovingAddressStatusSet$member": null - } - }, - "MovingAddressStatusSet": { - "base": null, - "refs": { - "DescribeMovingAddressesResult$MovingAddressStatuses": "

The status for each Elastic IP address.

" - } - }, - "NatGateway": { - "base": "

Describes a NAT gateway.

", - "refs": { - "CreateNatGatewayResult$NatGateway": "

Information about the NAT gateway.

", - "NatGatewayList$member": null - } - }, - "NatGatewayAddress": { - "base": "

Describes the IP addresses and network interface associated with a NAT gateway.

", - "refs": { - "NatGatewayAddressList$member": null - } - }, - "NatGatewayAddressList": { - "base": null, - "refs": { - "NatGateway$NatGatewayAddresses": "

Information about the IP addresses and network interface associated with the NAT gateway.

" - } - }, - "NatGatewayList": { - "base": null, - "refs": { - "DescribeNatGatewaysResult$NatGateways": "

Information about the NAT gateways.

" - } - }, - "NatGatewayState": { - "base": null, - "refs": { - "NatGateway$State": "

The state of the NAT gateway.

  • pending: The NAT gateway is being created and is not ready to process traffic.

  • failed: The NAT gateway could not be created. Check the failureCode and failureMessage fields for the reason.

  • available: The NAT gateway is able to process traffic. This status remains until you delete the NAT gateway, and does not indicate the health of the NAT gateway.

  • deleting: The NAT gateway is in the process of being terminated and may still be processing traffic.

  • deleted: The NAT gateway has been terminated and is no longer processing traffic.

" - } - }, - "NetworkAcl": { - "base": "

Describes a network ACL.

", - "refs": { - "CreateNetworkAclResult$NetworkAcl": "

Information about the network ACL.

", - "NetworkAclList$member": null - } - }, - "NetworkAclAssociation": { - "base": "

Describes an association between a network ACL and a subnet.

", - "refs": { - "NetworkAclAssociationList$member": null - } - }, - "NetworkAclAssociationList": { - "base": null, - "refs": { - "NetworkAcl$Associations": "

Any associations between the network ACL and one or more subnets

" - } - }, - "NetworkAclEntry": { - "base": "

Describes an entry in a network ACL.

", - "refs": { - "NetworkAclEntryList$member": null - } - }, - "NetworkAclEntryList": { - "base": null, - "refs": { - "NetworkAcl$Entries": "

One or more entries (rules) in the network ACL.

" - } - }, - "NetworkAclList": { - "base": null, - "refs": { - "DescribeNetworkAclsResult$NetworkAcls": "

Information about one or more network ACLs.

" - } - }, - "NetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "CreateNetworkInterfaceResult$NetworkInterface": "

Information about the network interface.

", - "NetworkInterfaceList$member": null - } - }, - "NetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "NetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "NetworkInterfacePrivateIpAddress$Association": "

The association information for an Elastic IP address associated with the network interface.

" - } - }, - "NetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Attachment": "

The attachment (if any) of the network interface.

", - "NetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "NetworkInterfaceAttachmentChanges": { - "base": "

Describes an attachment change.

", - "refs": { - "ModifyNetworkInterfaceAttributeRequest$Attachment": "

Information about the interface attachment. If modifying the 'delete on termination' attribute, you must specify the ID of the interface attachment.

" - } - }, - "NetworkInterfaceAttribute": { - "base": null, - "refs": { - "DescribeNetworkInterfaceAttributeRequest$Attribute": "

The attribute of the network interface.

" - } - }, - "NetworkInterfaceIdList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesRequest$NetworkInterfaceIds": "

One or more network interface IDs.

Default: Describes all your network interfaces.

" - } - }, - "NetworkInterfaceList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesResult$NetworkInterfaces": "

Information about one or more network interfaces.

" - } - }, - "NetworkInterfacePrivateIpAddress": { - "base": "

Describes the private IP address of a network interface.

", - "refs": { - "NetworkInterfacePrivateIpAddressList$member": null - } - }, - "NetworkInterfacePrivateIpAddressList": { - "base": null, - "refs": { - "NetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "NetworkInterfaceStatus": { - "base": null, - "refs": { - "InstanceNetworkInterface$Status": "

The status of the network interface.

", - "NetworkInterface$Status": "

The status of the network interface.

" - } - }, - "NetworkInterfaceType": { - "base": null, - "refs": { - "NetworkInterface$InterfaceType": "

The type of interface.

" - } - }, - "NewDhcpConfiguration": { - "base": null, - "refs": { - "NewDhcpConfigurationList$member": null - } - }, - "NewDhcpConfigurationList": { - "base": null, - "refs": { - "CreateDhcpOptionsRequest$DhcpConfigurations": "

A DHCP configuration option.

" - } - }, - "NextToken": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcClassicLinkDnsSupportRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcClassicLinkDnsSupportResult$NextToken": "

The token to use when requesting the next set of items.

" - } - }, - "OccurrenceDayRequestSet": { - "base": null, - "refs": { - "ScheduledInstanceRecurrenceRequest$OccurrenceDays": "

The days. For a monthly schedule, this is one or more days of the month (1-31). For a weekly schedule, this is one or more days of the week (1-7, where 1 is Sunday). You can't specify this value with a daily schedule. If the occurrence is relative to the end of the month, you can specify only a single day.

" - } - }, - "OccurrenceDaySet": { - "base": null, - "refs": { - "ScheduledInstanceRecurrence$OccurrenceDaySet": "

The days. For a monthly schedule, this is one or more days of the month (1-31). For a weekly schedule, this is one or more days of the week (1-7, where 1 is Sunday).

" - } - }, - "OfferingTypeValues": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "DescribeReservedInstancesRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "ReservedInstances$OfferingType": "

The Reserved Instance offering type.

", - "ReservedInstancesOffering$OfferingType": "

The Reserved Instance offering type.

" - } - }, - "OperationType": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$OperationType": "

The operation type.

", - "ModifySnapshotAttributeRequest$OperationType": "

The type of operation to perform to the attribute.

" - } - }, - "OwnerStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$Owners": "

Filters the images by the owner. Specify an AWS account ID, amazon (owner is Amazon), aws-marketplace (owner is AWS Marketplace), self (owner is the sender of the request). Omitting this option returns all images for which you have launch permissions, regardless of ownership.

", - "DescribeSnapshotsRequest$OwnerIds": "

Returns the snapshots owned by the specified owner. Multiple owners can be specified.

" - } - }, - "PeeringConnectionOptions": { - "base": "

Describes the VPC peering connection options.

", - "refs": { - "ModifyVpcPeeringConnectionOptionsResult$RequesterPeeringConnectionOptions": "

Information about the VPC peering connection options for the requester VPC.

", - "ModifyVpcPeeringConnectionOptionsResult$AccepterPeeringConnectionOptions": "

Information about the VPC peering connection options for the accepter VPC.

" - } - }, - "PeeringConnectionOptionsRequest": { - "base": "

The VPC peering connection options.

", - "refs": { - "ModifyVpcPeeringConnectionOptionsRequest$RequesterPeeringConnectionOptions": "

The VPC peering connection options for the requester VPC.

", - "ModifyVpcPeeringConnectionOptionsRequest$AccepterPeeringConnectionOptions": "

The VPC peering connection options for the accepter VPC.

" - } - }, - "PermissionGroup": { - "base": null, - "refs": { - "CreateVolumePermission$Group": "

The specific group that is to be added or removed from a volume's list of create volume permissions.

", - "LaunchPermission$Group": "

The name of the group.

" - } - }, - "Placement": { - "base": "

Describes the placement for the instance.

", - "refs": { - "ImportInstanceLaunchSpecification$Placement": "

The placement information for the instance.

", - "Instance$Placement": "

The location where the instance launched, if applicable.

", - "RunInstancesRequest$Placement": "

The placement for the instance.

" - } - }, - "PlacementGroup": { - "base": "

Describes a placement group.

", - "refs": { - "PlacementGroupList$member": null - } - }, - "PlacementGroupList": { - "base": null, - "refs": { - "DescribePlacementGroupsResult$PlacementGroups": "

One or more placement groups.

" - } - }, - "PlacementGroupState": { - "base": null, - "refs": { - "PlacementGroup$State": "

The state of the placement group.

" - } - }, - "PlacementGroupStringList": { - "base": null, - "refs": { - "DescribePlacementGroupsRequest$GroupNames": "

One or more placement group names.

Default: Describes all your placement groups, or only those otherwise specified.

" - } - }, - "PlacementStrategy": { - "base": null, - "refs": { - "CreatePlacementGroupRequest$Strategy": "

The placement strategy.

", - "PlacementGroup$Strategy": "

The placement strategy.

" - } - }, - "PlatformValues": { - "base": null, - "refs": { - "Image$Platform": "

The value is Windows for Windows AMIs; otherwise blank.

", - "ImportInstanceRequest$Platform": "

The instance operating system.

", - "ImportInstanceTaskDetails$Platform": "

The instance operating system.

", - "Instance$Platform": "

The value is Windows for Windows instances; otherwise blank.

" - } - }, - "PortRange": { - "base": "

Describes a range of ports.

", - "refs": { - "CreateNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "NetworkAclEntry$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "ReplaceNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to. Required if specifying 6 (TCP) or 17 (UDP) for the protocol.

" - } - }, - "PrefixList": { - "base": "

Describes prefixes for AWS services.

", - "refs": { - "PrefixListSet$member": null - } - }, - "PrefixListId": { - "base": "

The ID of the prefix.

", - "refs": { - "PrefixListIdList$member": null - } - }, - "PrefixListIdList": { - "base": null, - "refs": { - "IpPermission$PrefixListIds": "

(Valid for AuthorizeSecurityGroupEgress, RevokeSecurityGroupEgress and DescribeSecurityGroups only) One or more prefix list IDs for an AWS service. In an AuthorizeSecurityGroupEgress request, this is the AWS service that you want to access through a VPC endpoint from instances associated with the security group.

" - } - }, - "PrefixListIdSet": { - "base": null, - "refs": { - "StaleIpPermission$PrefixListIds": "

One or more prefix list IDs for an AWS service. Not applicable for stale security group rules.

" - } - }, - "PrefixListSet": { - "base": null, - "refs": { - "DescribePrefixListsResult$PrefixLists": "

All available prefix lists.

" - } - }, - "PriceSchedule": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleList$member": null - } - }, - "PriceScheduleList": { - "base": null, - "refs": { - "ReservedInstancesListing$PriceSchedules": "

The price of the Reserved Instance listing.

" - } - }, - "PriceScheduleSpecification": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleSpecificationList$member": null - } - }, - "PriceScheduleSpecificationList": { - "base": null, - "refs": { - "CreateReservedInstancesListingRequest$PriceSchedules": "

A list specifying the price of the Reserved Instance for each month remaining in the Reserved Instance term.

" - } - }, - "PricingDetail": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "PricingDetailsList$member": null - } - }, - "PricingDetailsList": { - "base": null, - "refs": { - "ReservedInstancesOffering$PricingDetails": "

The pricing details of the Reserved Instance offering.

" - } - }, - "PrivateIpAddressConfigSet": { - "base": null, - "refs": { - "ScheduledInstancesNetworkInterface$PrivateIpAddressConfigs": "

The private IP addresses.

" - } - }, - "PrivateIpAddressSpecification": { - "base": "

Describes a secondary private IP address for a network interface.

", - "refs": { - "PrivateIpAddressSpecificationList$member": null - } - }, - "PrivateIpAddressSpecificationList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$PrivateIpAddresses": "

One or more private IP addresses.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddresses": "

One or more private IP addresses to assign to the network interface. Only one private IP address can be designated as primary.

" - } - }, - "PrivateIpAddressStringList": { - "base": null, - "refs": { - "AssignPrivateIpAddressesRequest$PrivateIpAddresses": "

One or more IP addresses to be assigned as a secondary private IP address to the network interface. You can't specify this parameter when also specifying a number of secondary IP addresses.

If you don't specify an IP address, Amazon EC2 automatically selects an IP address within the subnet range.

", - "UnassignPrivateIpAddressesRequest$PrivateIpAddresses": "

The secondary private IP addresses to unassign from the network interface. You can specify this option multiple times to unassign more than one IP address.

" - } - }, - "ProductCode": { - "base": "

Describes a product code.

", - "refs": { - "ProductCodeList$member": null - } - }, - "ProductCodeList": { - "base": null, - "refs": { - "DescribeSnapshotAttributeResult$ProductCodes": "

A list of product codes.

", - "DescribeVolumeAttributeResult$ProductCodes": "

A list of product codes.

", - "Image$ProductCodes": "

Any product codes associated with the AMI.

", - "ImageAttribute$ProductCodes": "

One or more product codes.

", - "Instance$ProductCodes": "

The product codes attached to this instance, if applicable.

", - "InstanceAttribute$ProductCodes": "

A list of product codes.

" - } - }, - "ProductCodeStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$ProductCodes": "

One or more product codes. After you add a product code to an AMI, it can't be removed. This is only valid when modifying the productCodes attribute.

" - } - }, - "ProductCodeValues": { - "base": null, - "refs": { - "ProductCode$ProductCodeType": "

The type of product code.

" - } - }, - "ProductDescriptionList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$ProductDescriptions": "

Filters the results by the specified basic product descriptions.

" - } - }, - "PropagatingVgw": { - "base": "

Describes a virtual private gateway propagating route.

", - "refs": { - "PropagatingVgwList$member": null - } - }, - "PropagatingVgwList": { - "base": null, - "refs": { - "RouteTable$PropagatingVgws": "

Any virtual private gateway (VGW) propagating routes.

" - } - }, - "ProvisionedBandwidth": { - "base": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "refs": { - "NatGateway$ProvisionedBandwidth": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

" - } - }, - "PublicIpStringList": { - "base": null, - "refs": { - "DescribeAddressesRequest$PublicIps": "

[EC2-Classic] One or more Elastic IP addresses.

Default: Describes all your Elastic IP addresses.

" - } - }, - "PurchaseRequest": { - "base": "

Describes a request to purchase Scheduled Instances.

", - "refs": { - "PurchaseRequestSet$member": null - } - }, - "PurchaseRequestSet": { - "base": null, - "refs": { - "PurchaseScheduledInstancesRequest$PurchaseRequests": "

One or more purchase requests.

" - } - }, - "PurchaseReservedInstancesOfferingRequest": { - "base": "

Contains the parameters for PurchaseReservedInstancesOffering.

", - "refs": { - } - }, - "PurchaseReservedInstancesOfferingResult": { - "base": "

Contains the output of PurchaseReservedInstancesOffering.

", - "refs": { - } - }, - "PurchaseScheduledInstancesRequest": { - "base": "

Contains the parameters for PurchaseScheduledInstances.

", - "refs": { - } - }, - "PurchaseScheduledInstancesResult": { - "base": "

Contains the output of PurchaseScheduledInstances.

", - "refs": { - } - }, - "PurchasedScheduledInstanceSet": { - "base": null, - "refs": { - "PurchaseScheduledInstancesResult$ScheduledInstanceSet": "

Information about the Scheduled Instances.

" - } - }, - "RIProductDescription": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ProductDescription": "

The Reserved Instance product platform description. Instances that include (Amazon VPC) in the description are for use with Amazon VPC.

", - "ReservedInstances$ProductDescription": "

The Reserved Instance product platform description.

", - "ReservedInstancesOffering$ProductDescription": "

The Reserved Instance product platform description.

", - "SpotInstanceRequest$ProductDescription": "

The product description associated with the Spot instance.

", - "SpotPrice$ProductDescription": "

A general description of the AMI.

" - } - }, - "ReasonCodesList": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$ReasonCodes": "

One or more reason codes that describes the health state of your instance.

  • instance-stuck-in-state: My instance is stuck in a state.

  • unresponsive: My instance is unresponsive.

  • not-accepting-credentials: My instance is not accepting my credentials.

  • password-not-available: A password is not available for my instance.

  • performance-network: My instance is experiencing performance problems which I believe are network related.

  • performance-instance-store: My instance is experiencing performance problems which I believe are related to the instance stores.

  • performance-ebs-volume: My instance is experiencing performance problems which I believe are related to an EBS volume.

  • performance-other: My instance is experiencing performance problems.

  • other: [explain using the description parameter]

" - } - }, - "RebootInstancesRequest": { - "base": "

Contains the parameters for RebootInstances.

", - "refs": { - } - }, - "RecurringCharge": { - "base": "

Describes a recurring charge.

", - "refs": { - "RecurringChargesList$member": null - } - }, - "RecurringChargeFrequency": { - "base": null, - "refs": { - "RecurringCharge$Frequency": "

The frequency of the recurring charge.

" - } - }, - "RecurringChargesList": { - "base": null, - "refs": { - "ReservedInstances$RecurringCharges": "

The recurring charge tag assigned to the resource.

", - "ReservedInstancesOffering$RecurringCharges": "

The recurring charge tag assigned to the resource.

" - } - }, - "Region": { - "base": "

Describes a region.

", - "refs": { - "RegionList$member": null - } - }, - "RegionList": { - "base": null, - "refs": { - "DescribeRegionsResult$Regions": "

Information about one or more regions.

" - } - }, - "RegionNameStringList": { - "base": null, - "refs": { - "DescribeRegionsRequest$RegionNames": "

The names of one or more regions.

" - } - }, - "RegisterImageRequest": { - "base": "

Contains the parameters for RegisterImage.

", - "refs": { - } - }, - "RegisterImageResult": { - "base": "

Contains the output of RegisterImage.

", - "refs": { - } - }, - "RejectVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for RejectVpcPeeringConnection.

", - "refs": { - } - }, - "RejectVpcPeeringConnectionResult": { - "base": "

Contains the output of RejectVpcPeeringConnection.

", - "refs": { - } - }, - "ReleaseAddressRequest": { - "base": "

Contains the parameters for ReleaseAddress.

", - "refs": { - } - }, - "ReleaseHostsRequest": { - "base": "

Contains the parameters for ReleaseHosts.

", - "refs": { - } - }, - "ReleaseHostsResult": { - "base": "

Contains the output of ReleaseHosts.

", - "refs": { - } - }, - "ReplaceNetworkAclAssociationRequest": { - "base": "

Contains the parameters for ReplaceNetworkAclAssociation.

", - "refs": { - } - }, - "ReplaceNetworkAclAssociationResult": { - "base": "

Contains the output of ReplaceNetworkAclAssociation.

", - "refs": { - } - }, - "ReplaceNetworkAclEntryRequest": { - "base": "

Contains the parameters for ReplaceNetworkAclEntry.

", - "refs": { - } - }, - "ReplaceRouteRequest": { - "base": "

Contains the parameters for ReplaceRoute.

", - "refs": { - } - }, - "ReplaceRouteTableAssociationRequest": { - "base": "

Contains the parameters for ReplaceRouteTableAssociation.

", - "refs": { - } - }, - "ReplaceRouteTableAssociationResult": { - "base": "

Contains the output of ReplaceRouteTableAssociation.

", - "refs": { - } - }, - "ReportInstanceReasonCodes": { - "base": null, - "refs": { - "ReasonCodesList$member": null - } - }, - "ReportInstanceStatusRequest": { - "base": "

Contains the parameters for ReportInstanceStatus.

", - "refs": { - } - }, - "ReportStatusType": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$Status": "

The status of all instances listed.

" - } - }, - "RequestHostIdList": { - "base": null, - "refs": { - "DescribeHostsRequest$HostIds": "

The IDs of the Dedicated hosts. The IDs are used for targeted instance launches.

", - "ModifyHostsRequest$HostIds": "

The host IDs of the Dedicated hosts you want to modify.

", - "ReleaseHostsRequest$HostIds": "

The IDs of the Dedicated hosts you want to release.

" - } - }, - "RequestSpotFleetRequest": { - "base": "

Contains the parameters for RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotFleetResponse": { - "base": "

Contains the output of RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotInstancesRequest": { - "base": "

Contains the parameters for RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotInstancesResult": { - "base": "

Contains the output of RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotLaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "RequestSpotInstancesRequest$LaunchSpecification": null - } - }, - "Reservation": { - "base": "

Describes a reservation.

", - "refs": { - "ReservationList$member": null - } - }, - "ReservationList": { - "base": null, - "refs": { - "DescribeInstancesResult$Reservations": "

Zero or more reservations.

" - } - }, - "ReservedInstanceLimitPrice": { - "base": "

Describes the limit price of a Reserved Instance offering.

", - "refs": { - "PurchaseReservedInstancesOfferingRequest$LimitPrice": "

Specified for Reserved Instance Marketplace offerings to limit the total order and ensure that the Reserved Instances are not purchased at unexpected prices.

" - } - }, - "ReservedInstanceState": { - "base": null, - "refs": { - "ReservedInstances$State": "

The state of the Reserved Instance purchase.

" - } - }, - "ReservedInstances": { - "base": "

Describes a Reserved Instance.

", - "refs": { - "ReservedInstancesList$member": null - } - }, - "ReservedInstancesConfiguration": { - "base": "

Describes the configuration settings for the modified Reserved Instances.

", - "refs": { - "ReservedInstancesConfigurationList$member": null, - "ReservedInstancesModificationResult$TargetConfiguration": "

The target Reserved Instances configurations supplied as part of the modification request.

" - } - }, - "ReservedInstancesConfigurationList": { - "base": null, - "refs": { - "ModifyReservedInstancesRequest$TargetConfigurations": "

The configuration settings for the Reserved Instances to modify.

" - } - }, - "ReservedInstancesId": { - "base": "

Describes the ID of a Reserved Instance.

", - "refs": { - "ReservedIntancesIds$member": null - } - }, - "ReservedInstancesIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesRequest$ReservedInstancesIds": "

One or more Reserved Instance IDs.

Default: Describes all your Reserved Instances, or only those otherwise specified.

", - "ModifyReservedInstancesRequest$ReservedInstancesIds": "

The IDs of the Reserved Instances to modify.

" - } - }, - "ReservedInstancesList": { - "base": null, - "refs": { - "DescribeReservedInstancesResult$ReservedInstances": "

A list of Reserved Instances.

" - } - }, - "ReservedInstancesListing": { - "base": "

Describes a Reserved Instance listing.

", - "refs": { - "ReservedInstancesListingList$member": null - } - }, - "ReservedInstancesListingList": { - "base": null, - "refs": { - "CancelReservedInstancesListingResult$ReservedInstancesListings": "

The Reserved Instance listing.

", - "CreateReservedInstancesListingResult$ReservedInstancesListings": "

Information about the Reserved Instance listing.

", - "DescribeReservedInstancesListingsResult$ReservedInstancesListings": "

Information about the Reserved Instance listing.

" - } - }, - "ReservedInstancesModification": { - "base": "

Describes a Reserved Instance modification.

", - "refs": { - "ReservedInstancesModificationList$member": null - } - }, - "ReservedInstancesModificationIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsRequest$ReservedInstancesModificationIds": "

IDs for the submitted modification request.

" - } - }, - "ReservedInstancesModificationList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsResult$ReservedInstancesModifications": "

The Reserved Instance modification information.

" - } - }, - "ReservedInstancesModificationResult": { - "base": "

Describes the modification request/s.

", - "refs": { - "ReservedInstancesModificationResultList$member": null - } - }, - "ReservedInstancesModificationResultList": { - "base": null, - "refs": { - "ReservedInstancesModification$ModificationResults": "

Contains target configurations along with their corresponding new Reserved Instance IDs.

" - } - }, - "ReservedInstancesOffering": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "ReservedInstancesOfferingList$member": null - } - }, - "ReservedInstancesOfferingIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ReservedInstancesOfferingIds": "

One or more Reserved Instances offering IDs.

" - } - }, - "ReservedInstancesOfferingList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsResult$ReservedInstancesOfferings": "

A list of Reserved Instances offerings.

" - } - }, - "ReservedIntancesIds": { - "base": null, - "refs": { - "ReservedInstancesModification$ReservedInstancesIds": "

The IDs of one or more Reserved Instances.

" - } - }, - "ResetImageAttributeName": { - "base": null, - "refs": { - "ResetImageAttributeRequest$Attribute": "

The attribute to reset (currently you can only reset the launch permission attribute).

" - } - }, - "ResetImageAttributeRequest": { - "base": "

Contains the parameters for ResetImageAttribute.

", - "refs": { - } - }, - "ResetInstanceAttributeRequest": { - "base": "

Contains the parameters for ResetInstanceAttribute.

", - "refs": { - } - }, - "ResetNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for ResetNetworkInterfaceAttribute.

", - "refs": { - } - }, - "ResetSnapshotAttributeRequest": { - "base": "

Contains the parameters for ResetSnapshotAttribute.

", - "refs": { - } - }, - "ResourceIdList": { - "base": null, - "refs": { - "CreateTagsRequest$Resources": "

The IDs of one or more resources to tag. For example, ami-1a2b3c4d.

", - "DeleteTagsRequest$Resources": "

The ID of the resource. For example, ami-1a2b3c4d. You can specify more than one resource ID.

" - } - }, - "ResourceType": { - "base": null, - "refs": { - "TagDescription$ResourceType": "

The resource type.

" - } - }, - "ResponseHostIdList": { - "base": null, - "refs": { - "AllocateHostsResult$HostIds": "

The ID of the allocated Dedicated host. This is used when you want to launch an instance onto a specific host.

", - "ModifyHostsResult$Successful": "

The IDs of the Dedicated hosts that were successfully modified.

", - "ReleaseHostsResult$Successful": "

The IDs of the Dedicated hosts that were successfully released.

" - } - }, - "RestorableByStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$RestorableByUserIds": "

One or more AWS accounts IDs that can create volumes from the snapshot.

" - } - }, - "RestoreAddressToClassicRequest": { - "base": "

Contains the parameters for RestoreAddressToClassic.

", - "refs": { - } - }, - "RestoreAddressToClassicResult": { - "base": "

Contains the output of RestoreAddressToClassic.

", - "refs": { - } - }, - "RevokeSecurityGroupEgressRequest": { - "base": "

Contains the parameters for RevokeSecurityGroupEgress.

", - "refs": { - } - }, - "RevokeSecurityGroupIngressRequest": { - "base": "

Contains the parameters for RevokeSecurityGroupIngress.

", - "refs": { - } - }, - "Route": { - "base": "

Describes a route in a route table.

", - "refs": { - "RouteList$member": null - } - }, - "RouteList": { - "base": null, - "refs": { - "RouteTable$Routes": "

The routes in the route table.

" - } - }, - "RouteOrigin": { - "base": null, - "refs": { - "Route$Origin": "

Describes how the route was created.

  • CreateRouteTable - The route was automatically created when the route table was created.

  • CreateRoute - The route was manually added to the route table.

  • EnableVgwRoutePropagation - The route was propagated by route propagation.

" - } - }, - "RouteState": { - "base": null, - "refs": { - "Route$State": "

The state of the route. The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, or the specified NAT instance has been terminated).

" - } - }, - "RouteTable": { - "base": "

Describes a route table.

", - "refs": { - "CreateRouteTableResult$RouteTable": "

Information about the route table.

", - "RouteTableList$member": null - } - }, - "RouteTableAssociation": { - "base": "

Describes an association between a route table and a subnet.

", - "refs": { - "RouteTableAssociationList$member": null - } - }, - "RouteTableAssociationList": { - "base": null, - "refs": { - "RouteTable$Associations": "

The associations between the route table and one or more subnets.

" - } - }, - "RouteTableList": { - "base": null, - "refs": { - "DescribeRouteTablesResult$RouteTables": "

Information about one or more route tables.

" - } - }, - "RuleAction": { - "base": null, - "refs": { - "CreateNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "NetworkAclEntry$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "ReplaceNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

" - } - }, - "RunInstancesMonitoringEnabled": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "LaunchSpecification$Monitoring": null, - "RequestSpotLaunchSpecification$Monitoring": null, - "RunInstancesRequest$Monitoring": "

The monitoring for the instance.

" - } - }, - "RunInstancesRequest": { - "base": "

Contains the parameters for RunInstances.

", - "refs": { - } - }, - "RunScheduledInstancesRequest": { - "base": "

Contains the parameters for RunScheduledInstances.

", - "refs": { - } - }, - "RunScheduledInstancesResult": { - "base": "

Contains the output of RunScheduledInstances.

", - "refs": { - } - }, - "S3Storage": { - "base": "

Describes the storage parameters for S3 and S3 buckets for an instance store-backed AMI.

", - "refs": { - "Storage$S3": "

An Amazon S3 storage location.

" - } - }, - "ScheduledInstance": { - "base": "

Describes a Scheduled Instance.

", - "refs": { - "PurchasedScheduledInstanceSet$member": null, - "ScheduledInstanceSet$member": null - } - }, - "ScheduledInstanceAvailability": { - "base": "

Describes a schedule that is available for your Scheduled Instances.

", - "refs": { - "ScheduledInstanceAvailabilitySet$member": null - } - }, - "ScheduledInstanceAvailabilitySet": { - "base": null, - "refs": { - "DescribeScheduledInstanceAvailabilityResult$ScheduledInstanceAvailabilitySet": "

Information about the available Scheduled Instances.

" - } - }, - "ScheduledInstanceIdRequestSet": { - "base": null, - "refs": { - "DescribeScheduledInstancesRequest$ScheduledInstanceIds": "

One or more Scheduled Instance IDs.

" - } - }, - "ScheduledInstanceRecurrence": { - "base": "

Describes the recurring schedule for a Scheduled Instance.

", - "refs": { - "ScheduledInstance$Recurrence": "

The schedule recurrence.

", - "ScheduledInstanceAvailability$Recurrence": "

The schedule recurrence.

" - } - }, - "ScheduledInstanceRecurrenceRequest": { - "base": "

Describes the recurring schedule for a Scheduled Instance.

", - "refs": { - "DescribeScheduledInstanceAvailabilityRequest$Recurrence": "

The schedule recurrence.

" - } - }, - "ScheduledInstanceSet": { - "base": null, - "refs": { - "DescribeScheduledInstancesResult$ScheduledInstanceSet": "

Information about the Scheduled Instances.

" - } - }, - "ScheduledInstancesBlockDeviceMapping": { - "base": "

Describes a block device mapping for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesBlockDeviceMappingSet$member": null - } - }, - "ScheduledInstancesBlockDeviceMappingSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "ScheduledInstancesEbs": { - "base": "

Describes an EBS volume for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesBlockDeviceMapping$Ebs": "

Parameters used to set up EBS volumes automatically when the instance is launched.

" - } - }, - "ScheduledInstancesIamInstanceProfile": { - "base": "

Describes an IAM instance profile for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "ScheduledInstancesLaunchSpecification": { - "base": "

Describes the launch specification for a Scheduled Instance.

If you are launching the Scheduled Instance in EC2-VPC, you must specify the ID of the subnet. You can specify the subnet using either SubnetId or NetworkInterface.

", - "refs": { - "RunScheduledInstancesRequest$LaunchSpecification": "

The launch specification.

" - } - }, - "ScheduledInstancesMonitoring": { - "base": "

Describes whether monitoring is enabled for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "ScheduledInstancesNetworkInterface": { - "base": "

Describes a network interface for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesNetworkInterfaceSet$member": null - } - }, - "ScheduledInstancesNetworkInterfaceSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "ScheduledInstancesPlacement": { - "base": "

Describes the placement for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$Placement": "

The placement information.

" - } - }, - "ScheduledInstancesPrivateIpAddressConfig": { - "base": "

Describes a private IP address for a Scheduled Instance.

", - "refs": { - "PrivateIpAddressConfigSet$member": null - } - }, - "ScheduledInstancesSecurityGroupIdSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$SecurityGroupIds": "

The IDs of one or more security groups.

", - "ScheduledInstancesNetworkInterface$Groups": "

The IDs of one or more security groups.

" - } - }, - "SecurityGroup": { - "base": "

Describes a security group

", - "refs": { - "SecurityGroupList$member": null - } - }, - "SecurityGroupIdStringList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$Groups": "

The IDs of one or more security groups.

", - "ImportInstanceLaunchSpecification$GroupIds": "

One or more security group IDs.

", - "InstanceNetworkInterfaceSpecification$Groups": "

The IDs of the security groups for the network interface. Applies only if creating a network interface when launching an instance.

", - "ModifyNetworkInterfaceAttributeRequest$Groups": "

Changes the security groups for the network interface. The new set of groups you specify replaces the current set. You must specify at least one group, even if it's just the default security group in the VPC. You must specify the ID of the security group, not the name.

", - "RunInstancesRequest$SecurityGroupIds": "

One or more security group IDs. You can create a security group using CreateSecurityGroup.

Default: Amazon EC2 uses the default security group.

" - } - }, - "SecurityGroupList": { - "base": null, - "refs": { - "DescribeSecurityGroupsResult$SecurityGroups": "

Information about one or more security groups.

" - } - }, - "SecurityGroupReference": { - "base": "

Describes a VPC with a security group that references your security group.

", - "refs": { - "SecurityGroupReferences$member": null - } - }, - "SecurityGroupReferences": { - "base": null, - "refs": { - "DescribeSecurityGroupReferencesResult$SecurityGroupReferenceSet": "

Information about the VPCs with the referencing security groups.

" - } - }, - "SecurityGroupStringList": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$GroupNames": "

One or more security group names.

", - "RunInstancesRequest$SecurityGroups": "

[EC2-Classic, default VPC] One or more security group names. For a nondefault VPC, you must use security group IDs instead.

Default: Amazon EC2 uses the default security group.

" - } - }, - "ShutdownBehavior": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "RunInstancesRequest$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

Default: stop

" - } - }, - "SlotDateTimeRangeRequest": { - "base": "

Describes the time period for a Scheduled Instance to start its first schedule. The time period must span less than one day.

", - "refs": { - "DescribeScheduledInstanceAvailabilityRequest$FirstSlotStartTimeRange": "

The time period for the first schedule to start.

" - } - }, - "SlotStartTimeRangeRequest": { - "base": "

Describes the time period for a Scheduled Instance to start its first schedule.

", - "refs": { - "DescribeScheduledInstancesRequest$SlotStartTimeRange": "

The time period for the first schedule to start.

" - } - }, - "Snapshot": { - "base": "

Describes a snapshot.

", - "refs": { - "SnapshotList$member": null - } - }, - "SnapshotAttributeName": { - "base": null, - "refs": { - "DescribeSnapshotAttributeRequest$Attribute": "

The snapshot attribute you would like to view.

", - "ModifySnapshotAttributeRequest$Attribute": "

The snapshot attribute to modify.

Only volume creation permissions may be modified at the customer level.

", - "ResetSnapshotAttributeRequest$Attribute": "

The attribute to reset. Currently, only the attribute for permission to create volumes can be reset.

" - } - }, - "SnapshotDetail": { - "base": "

Describes the snapshot created from the imported disk.

", - "refs": { - "SnapshotDetailList$member": null - } - }, - "SnapshotDetailList": { - "base": null, - "refs": { - "ImportImageResult$SnapshotDetails": "

Information about the snapshots.

", - "ImportImageTask$SnapshotDetails": "

Information about the snapshots.

" - } - }, - "SnapshotDiskContainer": { - "base": "

The disk container object for the import snapshot request.

", - "refs": { - "ImportSnapshotRequest$DiskContainer": "

Information about the disk container.

" - } - }, - "SnapshotIdStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$SnapshotIds": "

One or more snapshot IDs.

Default: Describes snapshots for which you have launch permissions.

" - } - }, - "SnapshotList": { - "base": null, - "refs": { - "DescribeSnapshotsResult$Snapshots": "

Information about the snapshots.

" - } - }, - "SnapshotState": { - "base": null, - "refs": { - "Snapshot$State": "

The snapshot state.

" - } - }, - "SnapshotTaskDetail": { - "base": "

Details about the import snapshot task.

", - "refs": { - "ImportSnapshotResult$SnapshotTaskDetail": "

Information about the import snapshot task.

", - "ImportSnapshotTask$SnapshotTaskDetail": "

Describes an import snapshot task.

" - } - }, - "SpotDatafeedSubscription": { - "base": "

Describes the data feed for a Spot instance.

", - "refs": { - "CreateSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

", - "DescribeSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

" - } - }, - "SpotFleetLaunchSpecification": { - "base": "

Describes the launch specification for one or more Spot instances.

", - "refs": { - "LaunchSpecsList$member": null - } - }, - "SpotFleetMonitoring": { - "base": "

Describes whether monitoring is enabled.

", - "refs": { - "SpotFleetLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "SpotFleetRequestConfig": { - "base": "

Describes a Spot fleet request.

", - "refs": { - "SpotFleetRequestConfigSet$member": null - } - }, - "SpotFleetRequestConfigData": { - "base": "

Describes the configuration of a Spot fleet request.

", - "refs": { - "RequestSpotFleetRequest$SpotFleetRequestConfig": "

The configuration for the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestConfig": "

Information about the configuration of the Spot fleet request.

" - } - }, - "SpotFleetRequestConfigSet": { - "base": null, - "refs": { - "DescribeSpotFleetRequestsResponse$SpotFleetRequestConfigs": "

Information about the configuration of your Spot fleet.

" - } - }, - "SpotInstanceRequest": { - "base": "

Describes a Spot instance request.

", - "refs": { - "SpotInstanceRequestList$member": null - } - }, - "SpotInstanceRequestIdList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

", - "DescribeSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

" - } - }, - "SpotInstanceRequestList": { - "base": null, - "refs": { - "DescribeSpotInstanceRequestsResult$SpotInstanceRequests": "

One or more Spot instance requests.

", - "RequestSpotInstancesResult$SpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "SpotInstanceState": { - "base": null, - "refs": { - "SpotInstanceRequest$State": "

The state of the Spot instance request. Spot bid status information can help you track your Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "SpotInstanceStateFault": { - "base": "

Describes a Spot instance state change.

", - "refs": { - "SpotDatafeedSubscription$Fault": "

The fault codes for the Spot instance request, if any.

", - "SpotInstanceRequest$Fault": "

The fault codes for the Spot instance request, if any.

" - } - }, - "SpotInstanceStatus": { - "base": "

Describes the status of a Spot instance request.

", - "refs": { - "SpotInstanceRequest$Status": "

The status code and status message describing the Spot instance request.

" - } - }, - "SpotInstanceType": { - "base": null, - "refs": { - "RequestSpotInstancesRequest$Type": "

The Spot instance request type.

Default: one-time

", - "SpotInstanceRequest$Type": "

The Spot instance request type.

" - } - }, - "SpotPlacement": { - "base": "

Describes Spot instance placement.

", - "refs": { - "LaunchSpecification$Placement": "

The placement information for the instance.

", - "RequestSpotLaunchSpecification$Placement": "

The placement information for the instance.

", - "SpotFleetLaunchSpecification$Placement": "

The placement information.

" - } - }, - "SpotPrice": { - "base": "

Describes the maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "refs": { - "SpotPriceHistoryList$member": null - } - }, - "SpotPriceHistoryList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryResult$SpotPriceHistory": "

The historical Spot prices.

" - } - }, - "StaleIpPermission": { - "base": "

Describes a stale rule in a security group.

", - "refs": { - "StaleIpPermissionSet$member": null - } - }, - "StaleIpPermissionSet": { - "base": null, - "refs": { - "StaleSecurityGroup$StaleIpPermissions": "

Information about the stale inbound rules in the security group.

", - "StaleSecurityGroup$StaleIpPermissionsEgress": "

Information about the stale outbound rules in the security group.

" - } - }, - "StaleSecurityGroup": { - "base": "

Describes a stale security group (a security group that contains stale rules).

", - "refs": { - "StaleSecurityGroupSet$member": null - } - }, - "StaleSecurityGroupSet": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsResult$StaleSecurityGroupSet": "

Information about the stale security groups.

" - } - }, - "StartInstancesRequest": { - "base": "

Contains the parameters for StartInstances.

", - "refs": { - } - }, - "StartInstancesResult": { - "base": "

Contains the output of StartInstances.

", - "refs": { - } - }, - "State": { - "base": null, - "refs": { - "VpcEndpoint$State": "

The state of the VPC endpoint.

" - } - }, - "StateReason": { - "base": "

Describes a state change.

", - "refs": { - "Image$StateReason": "

The reason for the state change.

", - "Instance$StateReason": "

The reason for the most recent state transition.

" - } - }, - "Status": { - "base": null, - "refs": { - "MoveAddressToVpcResult$Status": "

The status of the move of the IP address.

", - "RestoreAddressToClassicResult$Status": "

The move status for the IP address.

" - } - }, - "StatusName": { - "base": null, - "refs": { - "InstanceStatusDetails$Name": "

The type of instance status.

" - } - }, - "StatusType": { - "base": null, - "refs": { - "InstanceStatusDetails$Status": "

The status.

" - } - }, - "StopInstancesRequest": { - "base": "

Contains the parameters for StopInstances.

", - "refs": { - } - }, - "StopInstancesResult": { - "base": "

Contains the output of StopInstances.

", - "refs": { - } - }, - "Storage": { - "base": "

Describes the storage location for an instance store-backed AMI.

", - "refs": { - "BundleInstanceRequest$Storage": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "BundleTask$Storage": "

The Amazon S3 storage locations.

" - } - }, - "String": { - "base": null, - "refs": { - "AcceptVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "AccountAttribute$AttributeName": "

The name of the account attribute.

", - "AccountAttributeValue$AttributeValue": "

The value of the attribute.

", - "ActiveInstance$InstanceType": "

The instance type.

", - "ActiveInstance$InstanceId": "

The ID of the instance.

", - "ActiveInstance$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "Address$InstanceId": "

The ID of the instance that the address is associated with (if any).

", - "Address$PublicIp": "

The Elastic IP address.

", - "Address$AllocationId": "

The ID representing the allocation of the address for use with EC2-VPC.

", - "Address$AssociationId": "

The ID representing the association of the address with an instance in a VPC.

", - "Address$NetworkInterfaceId": "

The ID of the network interface.

", - "Address$NetworkInterfaceOwnerId": "

The ID of the AWS account that owns the network interface.

", - "Address$PrivateIpAddress": "

The private IP address associated with the Elastic IP address.

", - "AllocateAddressResult$PublicIp": "

The Elastic IP address.

", - "AllocateAddressResult$AllocationId": "

[EC2-VPC] The ID that AWS assigns to represent the allocation of the Elastic IP address for use with instances in a VPC.

", - "AllocateHostsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "AllocateHostsRequest$InstanceType": "

Specify the instance type that you want your Dedicated hosts to be configured for. When you specify the instance type, that is the only instance type that you can launch onto that host.

", - "AllocateHostsRequest$AvailabilityZone": "

The Availability Zone for the Dedicated hosts.

", - "AllocationIdList$member": null, - "AssignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AssociateAddressRequest$InstanceId": "

The ID of the instance. This is required for EC2-Classic. For EC2-VPC, you can specify either the instance ID or the network interface ID, but not both. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "AssociateAddressRequest$PublicIp": "

The Elastic IP address. This is required for EC2-Classic.

", - "AssociateAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. This is required for EC2-VPC.

", - "AssociateAddressRequest$NetworkInterfaceId": "

[EC2-VPC] The ID of the network interface. If the instance has more than one network interface, you must specify a network interface ID.

", - "AssociateAddressRequest$PrivateIpAddress": "

[EC2-VPC] The primary or secondary private IP address to associate with the Elastic IP address. If no private IP address is specified, the Elastic IP address is associated with the primary private IP address.

", - "AssociateAddressResult$AssociationId": "

[EC2-VPC] The ID that represents the association of the Elastic IP address with an instance.

", - "AssociateDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set, or default to associate no DHCP options with the VPC.

", - "AssociateDhcpOptionsRequest$VpcId": "

The ID of the VPC.

", - "AssociateRouteTableRequest$SubnetId": "

The ID of the subnet.

", - "AssociateRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "AssociateRouteTableResult$AssociationId": "

The route table association ID (needed to disassociate the route table).

", - "AttachClassicLinkVpcRequest$InstanceId": "

The ID of an EC2-Classic instance to link to the ClassicLink-enabled VPC.

", - "AttachClassicLinkVpcRequest$VpcId": "

The ID of a ClassicLink-enabled VPC.

", - "AttachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "AttachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttachNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AttachNetworkInterfaceRequest$InstanceId": "

The ID of the instance.

", - "AttachNetworkInterfaceResult$AttachmentId": "

The ID of the network interface attachment.

", - "AttachVolumeRequest$VolumeId": "

The ID of the EBS volume. The volume and instance must be within the same Availability Zone.

", - "AttachVolumeRequest$InstanceId": "

The ID of the instance.

", - "AttachVolumeRequest$Device": "

The device name to expose to the instance (for example, /dev/sdh or xvdh).

", - "AttachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "AttachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttributeValue$Value": "

Valid values are case-sensitive and vary by action.

", - "AuthorizeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name or number. We recommend that you specify the protocol in a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. We recommend that you specify the CIDR range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "AuthorizeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead. For EC2-VPC, the source security group must be in the same VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic] The AWS account number for the source security group, if the source security group is in a different account. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). (VPC only) Use -1 to specify all.

", - "AuthorizeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "AvailabilityZone$ZoneName": "

The name of the Availability Zone.

", - "AvailabilityZone$RegionName": "

The name of the region.

", - "AvailabilityZoneMessage$Message": "

The message about the Availability Zone.

", - "BlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with 2 available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "BlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "BlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "BundleIdStringList$member": null, - "BundleInstanceRequest$InstanceId": "

The ID of the instance to bundle.

Type: String

Default: None

Required: Yes

", - "BundleTask$InstanceId": "

The ID of the instance associated with this bundle task.

", - "BundleTask$BundleId": "

The ID of the bundle task.

", - "BundleTask$Progress": "

The level of task completion, as a percent (for example, 20%).

", - "BundleTaskError$Code": "

The error code.

", - "BundleTaskError$Message": "

The error message.

", - "CancelBundleTaskRequest$BundleId": "

The ID of the bundle task.

", - "CancelConversionRequest$ConversionTaskId": "

The ID of the conversion task.

", - "CancelConversionRequest$ReasonMessage": "

The reason for canceling the conversion task.

", - "CancelExportTaskRequest$ExportTaskId": "

The ID of the export task. This is the ID returned by CreateInstanceExportTask.

", - "CancelImportTaskRequest$ImportTaskId": "

The ID of the import image or import snapshot task to be canceled.

", - "CancelImportTaskRequest$CancelReason": "

The reason for canceling the task.

", - "CancelImportTaskResult$ImportTaskId": "

The ID of the task being canceled.

", - "CancelImportTaskResult$State": "

The current state of the task being canceled.

", - "CancelImportTaskResult$PreviousState": "

The current state of the task being canceled.

", - "CancelReservedInstancesListingRequest$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "CancelSpotFleetRequestsError$Message": "

The description for the error code.

", - "CancelSpotFleetRequestsErrorItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelledSpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "ClassicLinkDnsSupport$VpcId": "

The ID of the VPC.

", - "ClassicLinkInstance$InstanceId": "

The ID of the instance.

", - "ClassicLinkInstance$VpcId": "

The ID of the VPC.

", - "ClientData$Comment": "

A user-defined comment about the disk upload.

", - "ConfirmProductInstanceRequest$ProductCode": "

The product code. This must be a product code that you own.

", - "ConfirmProductInstanceRequest$InstanceId": "

The ID of the instance.

", - "ConfirmProductInstanceResult$OwnerId": "

The AWS account ID of the instance owner. This is only present if the product code is attached to the instance.

", - "ConversionIdStringList$member": null, - "ConversionTask$ConversionTaskId": "

The ID of the conversion task.

", - "ConversionTask$ExpirationTime": "

The time when the task expires. If the upload isn't complete before the expiration time, we automatically cancel the task.

", - "ConversionTask$StatusMessage": "

The status message related to the conversion task.

", - "CopyImageRequest$SourceRegion": "

The name of the region that contains the AMI to copy.

", - "CopyImageRequest$SourceImageId": "

The ID of the AMI to copy.

", - "CopyImageRequest$Name": "

The name of the new AMI in the destination region.

", - "CopyImageRequest$Description": "

A description for the new AMI in the destination region.

", - "CopyImageRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "CopyImageRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when encrypting the snapshots of an image during a copy operation. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopyImageResult$ImageId": "

The ID of the new AMI.

", - "CopySnapshotRequest$SourceRegion": "

The ID of the region that contains the snapshot to be copied.

", - "CopySnapshotRequest$SourceSnapshotId": "

The ID of the EBS snapshot to copy.

", - "CopySnapshotRequest$Description": "

A description for the EBS snapshot.

", - "CopySnapshotRequest$DestinationRegion": "

The destination region to use in the PresignedUrl parameter of a snapshot copy operation. This parameter is only valid for specifying the destination region in a PresignedUrl parameter, where it is required.

CopySnapshot sends the snapshot copy to the regional endpoint that you send the HTTP request to, such as ec2.us-east-1.amazonaws.com (in the AWS CLI, this is specified with the --region parameter or the default region in your AWS configuration file).

", - "CopySnapshotRequest$PresignedUrl": "

The pre-signed URL that facilitates copying an encrypted snapshot. This parameter is only required when copying an encrypted snapshot with the Amazon EC2 Query API; it is available as an optional parameter in all other cases. The PresignedUrl should use the snapshot source endpoint, the CopySnapshot action, and include the SourceRegion, SourceSnapshotId, and DestinationRegion parameters. The PresignedUrl must be signed using AWS Signature Version 4. Because EBS snapshots are stored in Amazon S3, the signing algorithm for this parameter uses the same logic that is described in Authenticating Requests by Using Query Parameters (AWS Signature Version 4) in the Amazon Simple Storage Service API Reference. An invalid or improperly signed PresignedUrl will cause the copy operation to fail asynchronously, and the snapshot will move to an error state.

", - "CopySnapshotRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when creating the snapshot copy. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopySnapshotResult$SnapshotId": "

The ID of the new snapshot.

", - "CreateCustomerGatewayRequest$PublicIp": "

The Internet-routable IP address for the customer gateway's outside interface. The address must be static.

", - "CreateFlowLogsRequest$LogGroupName": "

The name of the CloudWatch log group.

", - "CreateFlowLogsRequest$DeliverLogsPermissionArn": "

The ARN for the IAM role that's used to post flow logs to a CloudWatch Logs log group.

", - "CreateFlowLogsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateFlowLogsResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateImageRequest$InstanceId": "

The ID of the instance.

", - "CreateImageRequest$Name": "

A name for the new image.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "CreateImageRequest$Description": "

A description for the new image.

", - "CreateImageResult$ImageId": "

The ID of the new AMI.

", - "CreateInstanceExportTaskRequest$Description": "

A description for the conversion task or the resource being exported. The maximum length is 255 bytes.

", - "CreateInstanceExportTaskRequest$InstanceId": "

The ID of the instance.

", - "CreateKeyPairRequest$KeyName": "

A unique name for the key pair.

Constraints: Up to 255 ASCII characters

", - "CreateNatGatewayRequest$SubnetId": "

The subnet in which to create the NAT gateway.

", - "CreateNatGatewayRequest$AllocationId": "

The allocation ID of an Elastic IP address to associate with the NAT gateway. If the Elastic IP address is associated with another resource, you must first disassociate it.

", - "CreateNatGatewayRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

Constraint: Maximum 64 ASCII characters.

", - "CreateNatGatewayResult$ClientToken": "

Unique, case-sensitive identifier to ensure the idempotency of the request. Only returned if a client token was provided in the request.

", - "CreateNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "CreateNetworkAclEntryRequest$Protocol": "

The protocol. A value of -1 means all protocols.

", - "CreateNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation (for example 172.16.0.0/24).

", - "CreateNetworkAclRequest$VpcId": "

The ID of the VPC.

", - "CreateNetworkInterfaceRequest$SubnetId": "

The ID of the subnet to associate with the network interface.

", - "CreateNetworkInterfaceRequest$Description": "

A description for the network interface.

", - "CreateNetworkInterfaceRequest$PrivateIpAddress": "

The primary private IP address of the network interface. If you don't specify an IP address, Amazon EC2 selects one for you from the subnet range. If you specify an IP address, you cannot indicate any IP addresses specified in privateIpAddresses as primary (only one IP address can be designated as primary).

", - "CreatePlacementGroupRequest$GroupName": "

A name for the placement group.

Constraints: Up to 255 ASCII characters

", - "CreateReservedInstancesListingRequest$ReservedInstancesId": "

The ID of the active Reserved Instance.

", - "CreateReservedInstancesListingRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "CreateRouteRequest$RouteTableId": "

The ID of the route table for the route.

", - "CreateRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. Routing decisions are based on the most specific match.

", - "CreateRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway attached to your VPC.

", - "CreateRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "CreateRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "CreateRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "CreateRouteRequest$NatGatewayId": "

The ID of a NAT gateway.

", - "CreateRouteTableRequest$VpcId": "

The ID of the VPC.

", - "CreateSecurityGroupRequest$GroupName": "

The name of the security group.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=;{}!$*

", - "CreateSecurityGroupRequest$Description": "

A description for the security group. This is informational only.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=;{}!$*

", - "CreateSecurityGroupRequest$VpcId": "

[EC2-VPC] The ID of the VPC. Required for EC2-VPC.

", - "CreateSecurityGroupResult$GroupId": "

The ID of the security group.

", - "CreateSnapshotRequest$VolumeId": "

The ID of the EBS volume.

", - "CreateSnapshotRequest$Description": "

A description for the snapshot.

", - "CreateSpotDatafeedSubscriptionRequest$Bucket": "

The Amazon S3 bucket in which to store the Spot instance data feed.

", - "CreateSpotDatafeedSubscriptionRequest$Prefix": "

A prefix for the data feed file names.

", - "CreateSubnetRequest$VpcId": "

The ID of the VPC.

", - "CreateSubnetRequest$CidrBlock": "

The network range for the subnet, in CIDR notation. For example, 10.0.0.0/24.

", - "CreateSubnetRequest$AvailabilityZone": "

The Availability Zone for the subnet.

Default: AWS selects one for you. If you create more than one subnet in your VPC, we may not necessarily select a different zone for each subnet.

", - "CreateVolumePermission$UserId": "

The specific AWS account ID that is to be added or removed from a volume's list of create volume permissions.

", - "CreateVolumeRequest$SnapshotId": "

The snapshot from which to create the volume.

", - "CreateVolumeRequest$AvailabilityZone": "

The Availability Zone in which to create the volume. Use DescribeAvailabilityZones to list the Availability Zones that are currently available to you.

", - "CreateVolumeRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) to use when creating the encrypted volume. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CreateVpcEndpointRequest$VpcId": "

The ID of the VPC in which the endpoint will be used.

", - "CreateVpcEndpointRequest$ServiceName": "

The AWS service name, in the form com.amazonaws.region.service. To get a list of available services, use the DescribeVpcEndpointServices request.

", - "CreateVpcEndpointRequest$PolicyDocument": "

A policy to attach to the endpoint that controls access to the service. The policy must be in valid JSON format. If this parameter is not specified, we attach a default policy that allows full access to the service.

", - "CreateVpcEndpointRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateVpcEndpointResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateVpcPeeringConnectionRequest$VpcId": "

The ID of the requester VPC.

", - "CreateVpcPeeringConnectionRequest$PeerVpcId": "

The ID of the VPC with which you are creating the VPC peering connection.

", - "CreateVpcPeeringConnectionRequest$PeerOwnerId": "

The AWS account ID of the owner of the peer VPC.

Default: Your AWS account ID

", - "CreateVpcRequest$CidrBlock": "

The network range for the VPC, in CIDR notation. For example, 10.0.0.0/16.

", - "CreateVpnConnectionRequest$Type": "

The type of VPN connection (ipsec.1).

", - "CreateVpnConnectionRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "CreateVpnConnectionRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "CreateVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "CreateVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "CreateVpnGatewayRequest$AvailabilityZone": "

The Availability Zone for the virtual private gateway.

", - "CustomerGateway$CustomerGatewayId": "

The ID of the customer gateway.

", - "CustomerGateway$State": "

The current state of the customer gateway (pending | available | deleting | deleted).

", - "CustomerGateway$Type": "

The type of VPN connection the customer gateway supports (ipsec.1).

", - "CustomerGateway$IpAddress": "

The Internet-routable IP address of the customer gateway's outside interface.

", - "CustomerGateway$BgpAsn": "

The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

", - "CustomerGatewayIdStringList$member": null, - "DeleteCustomerGatewayRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "DeleteDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set.

", - "DeleteInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DeleteKeyPairRequest$KeyName": "

The name of the key pair.

", - "DeleteNatGatewayRequest$NatGatewayId": "

The ID of the NAT gateway.

", - "DeleteNatGatewayResult$NatGatewayId": "

The ID of the NAT gateway.

", - "DeleteNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkAclRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DeletePlacementGroupRequest$GroupName": "

The name of the placement group.

", - "DeleteRouteRequest$RouteTableId": "

The ID of the route table.

", - "DeleteRouteRequest$DestinationCidrBlock": "

The CIDR range for the route. The value you specify must match the CIDR for the route exactly.

", - "DeleteRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "DeleteSecurityGroupRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group. You can specify either the security group name or the security group ID.

", - "DeleteSecurityGroupRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "DeleteSnapshotRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DeleteSubnetRequest$SubnetId": "

The ID of the subnet.

", - "DeleteVolumeRequest$VolumeId": "

The ID of the volume.

", - "DeleteVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "DeleteVpcRequest$VpcId": "

The ID of the VPC.

", - "DeleteVpnConnectionRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "DeleteVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DeregisterImageRequest$ImageId": "

The ID of the AMI.

", - "DescribeClassicLinkInstancesRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeClassicLinkInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeFlowLogsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeFlowLogsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeHostsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeIdFormatRequest$Resource": "

The type of resource.

", - "DescribeImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "DescribeImportImageTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportImageTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeImportSnapshotTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportSnapshotTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "DescribeInstanceStatusRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeInstanceStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeInstancesRequest$NextToken": "

The token to request the next page of results.

", - "DescribeInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeMovingAddressesRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeMovingAddressesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNatGatewaysRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeNatGatewaysResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribeNetworkInterfaceAttributeResult$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribePrefixListsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribePrefixListsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesId": "

One or more Reserved Instance IDs.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesListingId": "

One or more Reserved Instance listing IDs.

", - "DescribeReservedInstancesModificationsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesModificationsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeReservedInstancesOfferingsRequest$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "DescribeReservedInstancesOfferingsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesOfferingsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeScheduledInstanceAvailabilityRequest$NextToken": "

The token for the next set of results.

", - "DescribeScheduledInstanceAvailabilityResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeScheduledInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeScheduledInstancesResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSnapshotAttributeRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotAttributeResult$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotsRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeSnapshots request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeSnapshotsResult$NextToken": "

The NextToken value to include in a future DescribeSnapshots request. When the results of a DescribeSnapshots request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetInstancesRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetInstancesResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestHistoryRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestHistoryResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestsRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestsResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotPriceHistoryRequest$AvailabilityZone": "

Filters the results by the specified Availability Zone.

", - "DescribeSpotPriceHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotPriceHistoryResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeStaleSecurityGroupsRequest$VpcId": "

The ID of the VPC.

", - "DescribeStaleSecurityGroupsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeTagsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeTagsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return..

", - "DescribeVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "DescribeVolumeAttributeResult$VolumeId": "

The ID of the volume.

", - "DescribeVolumeStatusRequest$NextToken": "

The NextToken value to include in a future DescribeVolumeStatus request. When the results of the request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumeStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumesRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeVolumes request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeVolumesResult$NextToken": "

The NextToken value to include in a future DescribeVolumes request. When the results of a DescribeVolumes request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "DescribeVpcAttributeResult$VpcId": "

The ID of the VPC.

", - "DescribeVpcEndpointServicesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointServicesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeVpcEndpointsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DetachClassicLinkVpcRequest$InstanceId": "

The ID of the instance to unlink from the VPC.

", - "DetachClassicLinkVpcRequest$VpcId": "

The ID of the VPC to which the instance is linked.

", - "DetachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DetachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "DetachNetworkInterfaceRequest$AttachmentId": "

The ID of the attachment.

", - "DetachVolumeRequest$VolumeId": "

The ID of the volume.

", - "DetachVolumeRequest$InstanceId": "

The ID of the instance.

", - "DetachVolumeRequest$Device": "

The device name.

", - "DetachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DetachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "DhcpConfiguration$Key": "

The name of a DHCP option.

", - "DhcpOptions$DhcpOptionsId": "

The ID of the set of DHCP options.

", - "DhcpOptionsIdStringList$member": null, - "DisableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "DisableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "DisableVpcClassicLinkDnsSupportRequest$VpcId": "

The ID of the VPC.

", - "DisableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "DisassociateAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "DisassociateAddressRequest$AssociationId": "

[EC2-VPC] The association ID. Required for EC2-VPC.

", - "DisassociateRouteTableRequest$AssociationId": "

The association ID representing the current association between the route table and subnet.

", - "DiskImage$Description": "

A description of the disk image.

", - "DiskImageDescription$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DiskImageDescription$Checksum": "

The checksum computed for the disk image.

", - "DiskImageDetail$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3 and presented here as an Amazon S3 presigned URL. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DiskImageVolumeDescription$Id": "

The volume identifier.

", - "EbsBlockDevice$SnapshotId": "

The ID of the snapshot.

", - "EbsInstanceBlockDevice$VolumeId": "

The ID of the EBS volume.

", - "EbsInstanceBlockDeviceSpecification$VolumeId": "

The ID of the EBS volume.

", - "EnableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "EnableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "EnableVolumeIORequest$VolumeId": "

The ID of the volume.

", - "EnableVpcClassicLinkDnsSupportRequest$VpcId": "

The ID of the VPC.

", - "EnableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "EventInformation$InstanceId": "

The ID of the instance. This information is available only for instanceChange events.

", - "EventInformation$EventSubType": "

The event.

The following are the error events.

  • iamFleetRoleInvalid - The Spot fleet did not have the required permissions either to launch or terminate an instance.

  • launchSpecTemporarilyBlacklisted - The configuration is not valid and several attempts to launch instances have failed. For more information, see the description of the event.

  • spotFleetRequestConfigurationInvalid - The configuration is not valid. For more information, see the description of the event.

  • spotInstanceCountLimitExceeded - You've reached the limit on the number of Spot instances that you can launch.

The following are the fleetRequestChange events.

  • active - The Spot fleet has been validated and Amazon EC2 is attempting to maintain the target number of running Spot instances.

  • cancelled - The Spot fleet is canceled and has no running Spot instances. The Spot fleet will be deleted two days after its instances were terminated.

  • cancelled_running - The Spot fleet is canceled and will not launch additional Spot instances, but its existing Spot instances continue to run until they are interrupted or terminated.

  • cancelled_terminating - The Spot fleet is canceled and its Spot instances are terminating.

  • expired - The Spot fleet request has expired. A subsequent event indicates that the instances were terminated, if the request was created with TerminateInstancesWithExpiration set.

  • modify_in_progress - A request to modify the Spot fleet request was accepted and is in progress.

  • modify_successful - The Spot fleet request was modified.

  • price_update - The bid price for a launch configuration was adjusted because it was too high. This change is permanent.

  • submitted - The Spot fleet request is being evaluated and Amazon EC2 is preparing to launch the target number of Spot instances.

The following are the instanceChange events.

  • launched - A bid was fulfilled and a new instance was launched.

  • terminated - An instance was terminated by the user.

", - "EventInformation$EventDescription": "

The description of the event.

", - "ExecutableByStringList$member": null, - "ExportTask$ExportTaskId": "

The ID of the export task.

", - "ExportTask$Description": "

A description of the resource being exported.

", - "ExportTask$StatusMessage": "

The status message related to the export task.

", - "ExportTaskIdStringList$member": null, - "ExportToS3Task$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3Task$S3Key": "

The encryption key for your S3 bucket.

", - "ExportToS3TaskSpecification$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3TaskSpecification$S3Prefix": "

The image is written to a single object in the S3 bucket at the S3 key s3prefix + exportTaskId + '.' + diskImageFormat.

", - "Filter$Name": "

The name of the filter. Filter names are case-sensitive.

", - "FlowLog$FlowLogId": "

The flow log ID.

", - "FlowLog$FlowLogStatus": "

The status of the flow log (ACTIVE).

", - "FlowLog$ResourceId": "

The ID of the resource on which the flow log was created.

", - "FlowLog$LogGroupName": "

The name of the flow log group.

", - "FlowLog$DeliverLogsStatus": "

The status of the logs delivery (SUCCESS | FAILED).

", - "FlowLog$DeliverLogsErrorMessage": "

Information about the error that occurred. Rate limited indicates that CloudWatch logs throttling has been applied for one or more network interfaces, or that you've reached the limit on the number of CloudWatch Logs log groups that you can create. Access error indicates that the IAM role associated with the flow log does not have sufficient permissions to publish to CloudWatch Logs. Unknown error indicates an internal error.

", - "FlowLog$DeliverLogsPermissionArn": "

The ARN of the IAM role that posts logs to CloudWatch Logs.

", - "GetConsoleOutputRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$Output": "

The console output, base64-encoded. If using a command line tool, the tools decode the output for you.

", - "GetConsoleScreenshotRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleScreenshotResult$InstanceId": "

The ID of the instance.

", - "GetConsoleScreenshotResult$ImageData": "

The data that comprises the image.

", - "GetPasswordDataRequest$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$PasswordData": "

The password of the instance.

", - "GroupIdStringList$member": null, - "GroupIdentifier$GroupName": "

The name of the security group.

", - "GroupIdentifier$GroupId": "

The ID of the security group.

", - "GroupIds$member": null, - "GroupNameStringList$member": null, - "Host$HostId": "

The ID of the Dedicated host.

", - "Host$HostReservationId": "

The reservation ID of the Dedicated host. This returns a null response if the Dedicated host doesn't have an associated reservation.

", - "Host$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "Host$AvailabilityZone": "

The Availability Zone of the Dedicated host.

", - "HostInstance$InstanceId": "

the IDs of instances that are running on the Dedicated host.

", - "HostInstance$InstanceType": "

The instance type size (for example, m3.medium) of the running instance.

", - "HostProperties$InstanceType": "

The instance type size that the Dedicated host supports (for example, m3.medium).

", - "IamInstanceProfile$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfile$Id": "

The ID of the instance profile.

", - "IamInstanceProfileSpecification$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfileSpecification$Name": "

The name of the instance profile.

", - "IdFormat$Resource": "

The type of resource.

", - "Image$ImageId": "

The ID of the AMI.

", - "Image$ImageLocation": "

The location of the AMI.

", - "Image$OwnerId": "

The AWS account ID of the image owner.

", - "Image$CreationDate": "

The date and time the image was created.

", - "Image$KernelId": "

The kernel associated with the image, if any. Only applicable for machine images.

", - "Image$RamdiskId": "

The RAM disk associated with the image, if any. Only applicable for machine images.

", - "Image$SriovNetSupport": "

Specifies whether enhanced networking is enabled.

", - "Image$ImageOwnerAlias": "

The AWS account alias (for example, amazon, self) or the AWS account ID of the AMI owner.

", - "Image$Name": "

The name of the AMI that was provided during image creation.

", - "Image$Description": "

The description of the AMI that was provided during image creation.

", - "Image$RootDeviceName": "

The device name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "ImageAttribute$ImageId": "

The ID of the AMI.

", - "ImageDiskContainer$Description": "

The description of the disk image.

", - "ImageDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "ImageDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. The URL can either be a https URL (https://..) or an Amazon S3 URL (s3://..)

", - "ImageDiskContainer$DeviceName": "

The block device mapping for the disk.

", - "ImageDiskContainer$SnapshotId": "

The ID of the EBS snapshot to be used for importing the snapshot.

", - "ImageIdStringList$member": null, - "ImportImageRequest$Description": "

A description string for the import image task.

", - "ImportImageRequest$LicenseType": "

The license type to be used for the Amazon Machine Image (AMI) after importing.

Note: You may only use BYOL if you have existing licenses with rights to use these licenses in a third party cloud like AWS. For more information, see VM Import/Export Prerequisites in the Amazon Elastic Compute Cloud User Guide.

Valid values: AWS | BYOL

", - "ImportImageRequest$Hypervisor": "

The target hypervisor platform.

Valid values: xen

", - "ImportImageRequest$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageRequest$Platform": "

The operating system of the virtual machine.

Valid values: Windows | Linux

", - "ImportImageRequest$ClientToken": "

The token to enable idempotency for VM import requests.

", - "ImportImageRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportImageResult$ImportTaskId": "

The task ID of the import image task.

", - "ImportImageResult$Architecture": "

The architecture of the virtual machine.

", - "ImportImageResult$LicenseType": "

The license type of the virtual machine.

", - "ImportImageResult$Platform": "

The operating system of the virtual machine.

", - "ImportImageResult$Hypervisor": "

The target hypervisor of the import task.

", - "ImportImageResult$Description": "

A description of the import task.

", - "ImportImageResult$ImageId": "

The ID of the Amazon Machine Image (AMI) created by the import task.

", - "ImportImageResult$Progress": "

The progress of the task.

", - "ImportImageResult$StatusMessage": "

A detailed status message of the import task.

", - "ImportImageResult$Status": "

A brief status of the task.

", - "ImportImageTask$ImportTaskId": "

The ID of the import image task.

", - "ImportImageTask$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageTask$LicenseType": "

The license type of the virtual machine.

", - "ImportImageTask$Platform": "

The description string for the import image task.

", - "ImportImageTask$Hypervisor": "

The target hypervisor for the import task.

Valid values: xen

", - "ImportImageTask$Description": "

A description of the import task.

", - "ImportImageTask$ImageId": "

The ID of the Amazon Machine Image (AMI) of the imported virtual machine.

", - "ImportImageTask$Progress": "

The percentage of progress of the import image task.

", - "ImportImageTask$StatusMessage": "

A descriptive status message for the import image task.

", - "ImportImageTask$Status": "

A brief status for the import image task.

", - "ImportInstanceLaunchSpecification$AdditionalInfo": "

Reserved.

", - "ImportInstanceLaunchSpecification$SubnetId": "

[EC2-VPC] The ID of the subnet in which to launch the instance.

", - "ImportInstanceLaunchSpecification$PrivateIpAddress": "

[EC2-VPC] An available IP address from the IP address range of the subnet.

", - "ImportInstanceRequest$Description": "

A description for the instance being imported.

", - "ImportInstanceTaskDetails$InstanceId": "

The ID of the instance.

", - "ImportInstanceTaskDetails$Description": "

A description of the task.

", - "ImportInstanceVolumeDetailItem$AvailabilityZone": "

The Availability Zone where the resulting instance will reside.

", - "ImportInstanceVolumeDetailItem$Status": "

The status of the import of this particular disk image.

", - "ImportInstanceVolumeDetailItem$StatusMessage": "

The status information or errors related to the disk image.

", - "ImportInstanceVolumeDetailItem$Description": "

A description of the task.

", - "ImportKeyPairRequest$KeyName": "

A unique name for the key pair.

", - "ImportKeyPairResult$KeyName": "

The key pair name you provided.

", - "ImportKeyPairResult$KeyFingerprint": "

The MD5 public key fingerprint as specified in section 4 of RFC 4716.

", - "ImportSnapshotRequest$Description": "

The description string for the import snapshot task.

", - "ImportSnapshotRequest$ClientToken": "

Token to enable idempotency for VM import requests.

", - "ImportSnapshotRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportSnapshotResult$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotResult$Description": "

A description of the import snapshot task.

", - "ImportSnapshotTask$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotTask$Description": "

A description of the import snapshot task.

", - "ImportTaskIdList$member": null, - "ImportVolumeRequest$AvailabilityZone": "

The Availability Zone for the resulting EBS volume.

", - "ImportVolumeRequest$Description": "

A description of the volume.

", - "ImportVolumeTaskDetails$AvailabilityZone": "

The Availability Zone where the resulting volume will reside.

", - "ImportVolumeTaskDetails$Description": "

The description you provided when starting the import volume task.

", - "Instance$InstanceId": "

The ID of the instance.

", - "Instance$ImageId": "

The ID of the AMI used to launch the instance.

", - "Instance$PrivateDnsName": "

The private DNS name assigned to the instance. This DNS name can only be used inside the Amazon EC2 network. This name is not available until the instance enters the running state. For EC2-VPC, this name is only available if you've enabled DNS hostnames for your VPC.

", - "Instance$PublicDnsName": "

The public DNS name assigned to the instance. This name is not available until the instance enters the running state. For EC2-VPC, this name is only available if you've enabled DNS hostnames for your VPC.

", - "Instance$StateTransitionReason": "

The reason for the most recent state transition. This might be an empty string.

", - "Instance$KeyName": "

The name of the key pair, if this instance was launched with an associated key pair.

", - "Instance$KernelId": "

The kernel associated with this instance, if applicable.

", - "Instance$RamdiskId": "

The RAM disk associated with this instance, if applicable.

", - "Instance$SubnetId": "

[EC2-VPC] The ID of the subnet in which the instance is running.

", - "Instance$VpcId": "

[EC2-VPC] The ID of the VPC in which the instance is running.

", - "Instance$PrivateIpAddress": "

The private IP address assigned to the instance.

", - "Instance$PublicIpAddress": "

The public IP address assigned to the instance, if applicable.

", - "Instance$RootDeviceName": "

The root device name (for example, /dev/sda1 or /dev/xvda).

", - "Instance$SpotInstanceRequestId": "

If the request is a Spot instance request, the ID of the request.

", - "Instance$ClientToken": "

The idempotency token you provided when you launched the instance, if applicable.

", - "Instance$SriovNetSupport": "

Specifies whether enhanced networking is enabled.

", - "InstanceAttribute$InstanceId": "

The ID of the instance.

", - "InstanceBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$VirtualName": "

The virtual device name.

", - "InstanceBlockDeviceMappingSpecification$NoDevice": "

suppress the specified device included in the block device mapping.

", - "InstanceCapacity$InstanceType": "

The instance type size supported by the Dedicated host.

", - "InstanceExportDetails$InstanceId": "

The ID of the resource being exported.

", - "InstanceIdSet$member": null, - "InstanceIdStringList$member": null, - "InstanceMonitoring$InstanceId": "

The ID of the instance.

", - "InstanceNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterface$SubnetId": "

The ID of the subnet.

", - "InstanceNetworkInterface$VpcId": "

The ID of the VPC.

", - "InstanceNetworkInterface$Description": "

The description.

", - "InstanceNetworkInterface$OwnerId": "

The ID of the AWS account that created the network interface.

", - "InstanceNetworkInterface$MacAddress": "

The MAC address.

", - "InstanceNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "InstanceNetworkInterface$PrivateDnsName": "

The private DNS name.

", - "InstanceNetworkInterfaceAssociation$PublicIp": "

The public IP address or Elastic IP address bound to the network interface.

", - "InstanceNetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "InstanceNetworkInterfaceAssociation$IpOwnerId": "

The ID of the owner of the Elastic IP address.

", - "InstanceNetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterfaceSpecification$SubnetId": "

The ID of the subnet associated with the network string. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$Description": "

The description of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddress": "

The private IP address of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstancePrivateIpAddress$PrivateIpAddress": "

The private IP address of the network interface.

", - "InstancePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "InstanceStateChange$InstanceId": "

The ID of the instance.

", - "InstanceStatus$InstanceId": "

The ID of the instance.

", - "InstanceStatus$AvailabilityZone": "

The Availability Zone of the instance.

", - "InstanceStatusEvent$Description": "

A description of the event.

After a scheduled event is completed, it can still be described for up to a week. If the event has been completed, this description starts with the following text: [Completed].

", - "InternetGateway$InternetGatewayId": "

The ID of the Internet gateway.

", - "InternetGatewayAttachment$VpcId": "

The ID of the VPC.

", - "IpPermission$IpProtocol": "

The IP protocol name (for tcp, udp, and icmp) or number (see Protocol Numbers).

[EC2-VPC only] When you authorize or revoke security group rules, you can use -1 to specify all.

", - "IpRange$CidrIp": "

The CIDR range. You can either specify a CIDR range or a source security group, not both.

", - "IpRanges$member": null, - "KeyNameStringList$member": null, - "KeyPair$KeyName": "

The name of the key pair.

", - "KeyPair$KeyFingerprint": "

The SHA-1 digest of the DER encoded private key.

", - "KeyPair$KeyMaterial": "

An unencrypted PEM encoded RSA private key.

", - "KeyPairInfo$KeyName": "

The name of the key pair.

", - "KeyPairInfo$KeyFingerprint": "

If you used CreateKeyPair to create the key pair, this is the SHA-1 digest of the DER encoded private key. If you used ImportKeyPair to provide AWS the public key, this is the MD5 public key fingerprint as specified in section 4 of RFC4716.

", - "LaunchPermission$UserId": "

The AWS account ID.

", - "LaunchSpecification$ImageId": "

The ID of the AMI.

", - "LaunchSpecification$KeyName": "

The name of the key pair.

", - "LaunchSpecification$UserData": "

The Base64-encoded MIME user data to make available to the instances.

", - "LaunchSpecification$AddressingType": "

Deprecated.

", - "LaunchSpecification$KernelId": "

The ID of the kernel.

", - "LaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "LaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "ModifyIdFormatRequest$Resource": "

The type of resource.

", - "ModifyImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ModifyImageAttributeRequest$Attribute": "

The name of the attribute to modify.

", - "ModifyImageAttributeRequest$Value": "

The value of the attribute being modified. This is only valid when modifying the description attribute.

", - "ModifyInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ModifyInstanceAttributeRequest$Value": "

A new value for the attribute. Use only with the kernel, ramdisk, userData, disableApiTermination, or instanceInitiatedShutdownBehavior attribute.

", - "ModifyInstancePlacementRequest$InstanceId": "

The ID of the instance that you are modifying.

", - "ModifyInstancePlacementRequest$HostId": "

The ID of the Dedicated host that the instance will have affinity with.

", - "ModifyNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ModifyReservedInstancesRequest$ClientToken": "

A unique, case-sensitive token you provide to ensure idempotency of your modification request. For more information, see Ensuring Idempotency.

", - "ModifyReservedInstancesResult$ReservedInstancesModificationId": "

The ID for the modification.

", - "ModifySnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ModifySpotFleetRequestRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "ModifySubnetAttributeRequest$SubnetId": "

The ID of the subnet.

", - "ModifyVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "ModifyVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "ModifyVpcEndpointRequest$VpcEndpointId": "

The ID of the endpoint.

", - "ModifyVpcEndpointRequest$PolicyDocument": "

A policy document to attach to the endpoint. The policy must be in valid JSON format.

", - "ModifyVpcPeeringConnectionOptionsRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "MoveAddressToVpcRequest$PublicIp": "

The Elastic IP address.

", - "MoveAddressToVpcResult$AllocationId": "

The allocation ID for the Elastic IP address.

", - "MovingAddressStatus$PublicIp": "

The Elastic IP address.

", - "NatGateway$VpcId": "

The ID of the VPC in which the NAT gateway is located.

", - "NatGateway$SubnetId": "

The ID of the subnet in which the NAT gateway is located.

", - "NatGateway$NatGatewayId": "

The ID of the NAT gateway.

", - "NatGateway$FailureCode": "

If the NAT gateway could not be created, specifies the error code for the failure. (InsufficientFreeAddressesInSubnet | Gateway.NotAttached | InvalidAllocationID.NotFound | Resource.AlreadyAssociated | InternalError | InvalidSubnetID.NotFound)

", - "NatGateway$FailureMessage": "

If the NAT gateway could not be created, specifies the error message for the failure, that corresponds to the error code.

  • For InsufficientFreeAddressesInSubnet: \"Subnet has insufficient free addresses to create this NAT gateway\"

  • For Gateway.NotAttached: \"Network vpc-xxxxxxxx has no Internet gateway attached\"

  • For InvalidAllocationID.NotFound: \"Elastic IP address eipalloc-xxxxxxxx could not be associated with this NAT gateway\"

  • For Resource.AlreadyAssociated: \"Elastic IP address eipalloc-xxxxxxxx is already associated\"

  • For InternalError: \"Network interface eni-xxxxxxxx, created and used internally by this NAT gateway is in an invalid state. Please try again.\"

  • For InvalidSubnetID.NotFound: \"The specified subnet subnet-xxxxxxxx does not exist or could not be found.\"

", - "NatGatewayAddress$PublicIp": "

The Elastic IP address associated with the NAT gateway.

", - "NatGatewayAddress$AllocationId": "

The allocation ID of the Elastic IP address that's associated with the NAT gateway.

", - "NatGatewayAddress$PrivateIp": "

The private IP address associated with the Elastic IP address.

", - "NatGatewayAddress$NetworkInterfaceId": "

The ID of the network interface associated with the NAT gateway.

", - "NetworkAcl$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAcl$VpcId": "

The ID of the VPC for the network ACL.

", - "NetworkAclAssociation$NetworkAclAssociationId": "

The ID of the association between a network ACL and a subnet.

", - "NetworkAclAssociation$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAclAssociation$SubnetId": "

The ID of the subnet.

", - "NetworkAclEntry$Protocol": "

The protocol. A value of -1 means all protocols.

", - "NetworkAclEntry$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "NetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "NetworkInterface$SubnetId": "

The ID of the subnet.

", - "NetworkInterface$VpcId": "

The ID of the VPC.

", - "NetworkInterface$AvailabilityZone": "

The Availability Zone.

", - "NetworkInterface$Description": "

A description.

", - "NetworkInterface$OwnerId": "

The AWS account ID of the owner of the network interface.

", - "NetworkInterface$RequesterId": "

The ID of the entity that launched the instance on your behalf (for example, AWS Management Console or Auto Scaling).

", - "NetworkInterface$MacAddress": "

The MAC address.

", - "NetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "NetworkInterface$PrivateDnsName": "

The private DNS name.

", - "NetworkInterfaceAssociation$PublicIp": "

The address of the Elastic IP address bound to the network interface.

", - "NetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "NetworkInterfaceAssociation$IpOwnerId": "

The ID of the Elastic IP address owner.

", - "NetworkInterfaceAssociation$AllocationId": "

The allocation ID.

", - "NetworkInterfaceAssociation$AssociationId": "

The association ID.

", - "NetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceAttachment$InstanceId": "

The ID of the instance.

", - "NetworkInterfaceAttachment$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "NetworkInterfaceAttachmentChanges$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceIdList$member": null, - "NetworkInterfacePrivateIpAddress$PrivateIpAddress": "

The private IP address.

", - "NetworkInterfacePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "NewDhcpConfiguration$Key": null, - "OwnerStringList$member": null, - "Placement$AvailabilityZone": "

The Availability Zone of the instance.

", - "Placement$GroupName": "

The name of the placement group the instance is in (for cluster compute instances).

", - "Placement$HostId": "

The ID of the Dedicted host on which the instance resides. This parameter is not support for the ImportInstance command.

", - "Placement$Affinity": "

The affinity setting for the instance on the Dedicated host. This parameter is not supported for the ImportInstance command.

", - "PlacementGroup$GroupName": "

The name of the placement group.

", - "PlacementGroupStringList$member": null, - "PrefixList$PrefixListId": "

The ID of the prefix.

", - "PrefixList$PrefixListName": "

The name of the prefix.

", - "PrefixListId$PrefixListId": "

The ID of the prefix.

", - "PrefixListIdSet$member": null, - "PrivateIpAddressSpecification$PrivateIpAddress": "

The private IP addresses.

", - "PrivateIpAddressStringList$member": null, - "ProductCode$ProductCodeId": "

The product code.

", - "ProductCodeStringList$member": null, - "ProductDescriptionList$member": null, - "PropagatingVgw$GatewayId": "

The ID of the virtual private gateway (VGW).

", - "ProvisionedBandwidth$Provisioned": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$Requested": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$Status": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "PublicIpStringList$member": null, - "PurchaseRequest$PurchaseToken": "

The purchase token.

", - "PurchaseReservedInstancesOfferingRequest$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering to purchase.

", - "PurchaseReservedInstancesOfferingResult$ReservedInstancesId": "

The IDs of the purchased Reserved Instances.

", - "PurchaseScheduledInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that ensures the idempotency of the request. For more information, see Ensuring Idempotency.

", - "Region$RegionName": "

The name of the region.

", - "Region$Endpoint": "

The region service endpoint.

", - "RegionNameStringList$member": null, - "RegisterImageRequest$ImageLocation": "

The full path to your AMI manifest in Amazon S3 storage.

", - "RegisterImageRequest$Name": "

A name for your AMI.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "RegisterImageRequest$Description": "

A description for your AMI.

", - "RegisterImageRequest$KernelId": "

The ID of the kernel.

", - "RegisterImageRequest$RamdiskId": "

The ID of the RAM disk.

", - "RegisterImageRequest$RootDeviceName": "

The name of the root device (for example, /dev/sda1, or /dev/xvda).

", - "RegisterImageRequest$VirtualizationType": "

The type of virtualization.

Default: paravirtual

", - "RegisterImageRequest$SriovNetSupport": "

Set to simple to enable enhanced networking for the AMI and any instances that you launch from the AMI.

There is no way to disable enhanced networking at this time.

This option is supported only for HVM AMIs. Specifying this option with a PV AMI can make instances launched from the AMI unreachable.

", - "RegisterImageResult$ImageId": "

The ID of the newly registered AMI.

", - "RejectVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "ReleaseAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "ReleaseAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. Required for EC2-VPC.

", - "ReplaceNetworkAclAssociationRequest$AssociationId": "

The ID of the current association between the original network ACL and the subnet.

", - "ReplaceNetworkAclAssociationRequest$NetworkAclId": "

The ID of the new network ACL to associate with the subnet.

", - "ReplaceNetworkAclAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReplaceNetworkAclEntryRequest$NetworkAclId": "

The ID of the ACL.

", - "ReplaceNetworkAclEntryRequest$Protocol": "

The IP protocol. You can specify all or -1 to mean all protocols.

", - "ReplaceNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "ReplaceRouteRequest$RouteTableId": "

The ID of the route table.

", - "ReplaceRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. The value you provide must match the CIDR of an existing route in the table.

", - "ReplaceRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway.

", - "ReplaceRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC.

", - "ReplaceRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "ReplaceRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "ReplaceRouteRequest$NatGatewayId": "

The ID of a NAT gateway.

", - "ReplaceRouteTableAssociationRequest$AssociationId": "

The association ID.

", - "ReplaceRouteTableAssociationRequest$RouteTableId": "

The ID of the new route table to associate with the subnet.

", - "ReplaceRouteTableAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReportInstanceStatusRequest$Description": "

Descriptive text about the health state of your instance.

", - "RequestHostIdList$member": null, - "RequestSpotFleetResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "RequestSpotInstancesRequest$SpotPrice": "

The maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "RequestSpotInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstancesRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

Default: Instances are launched and terminated individually

", - "RequestSpotInstancesRequest$AvailabilityZoneGroup": "

The user-specified name for a logical grouping of bids.

When you specify an Availability Zone group in a Spot Instance request, all Spot instances in the request are launched in the same Availability Zone. Instance proximity is maintained with this parameter, but the choice of Availability Zone is not. The group applies only to bids for Spot Instances of the same instance type. Any additional Spot instance requests that are specified with the same Availability Zone group name are launched in that same Availability Zone, as long as at least one instance from the group is still active.

If there is no active instance running in the Availability Zone group that you specify for a new Spot instance request (all instances are terminated, the bid is expired, or the bid falls below current market), then Amazon EC2 launches the instance in any Availability Zone where the constraint can be met. Consequently, the subsequent set of Spot instances could be placed in a different zone from the original request, even if you specified the same Availability Zone group.

Default: Instances are launched in any available Availability Zone.

", - "RequestSpotLaunchSpecification$ImageId": "

The ID of the AMI.

", - "RequestSpotLaunchSpecification$KeyName": "

The name of the key pair.

", - "RequestSpotLaunchSpecification$UserData": "

The Base64-encoded MIME user data to make available to the instances.

", - "RequestSpotLaunchSpecification$AddressingType": "

Deprecated.

", - "RequestSpotLaunchSpecification$KernelId": "

The ID of the kernel.

", - "RequestSpotLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "RequestSpotLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "Reservation$ReservationId": "

The ID of the reservation.

", - "Reservation$OwnerId": "

The ID of the AWS account that owns the reservation.

", - "Reservation$RequesterId": "

The ID of the requester that launched the instances on your behalf (for example, AWS Management Console or Auto Scaling).

", - "ReservedInstances$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstances$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$AvailabilityZone": "

The Availability Zone for the modified Reserved Instances.

", - "ReservedInstancesConfiguration$Platform": "

The network platform of the modified Reserved Instances, which is either EC2-Classic or EC2-VPC.

", - "ReservedInstancesId$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesIdStringList$member": null, - "ReservedInstancesListing$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "ReservedInstancesListing$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesListing$StatusMessage": "

The reason for the current status of the Reserved Instance listing. The response can be blank.

", - "ReservedInstancesListing$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModification$ReservedInstancesModificationId": "

A unique ID for the Reserved Instance modification.

", - "ReservedInstancesModification$Status": "

The status of the Reserved Instances modification request.

", - "ReservedInstancesModification$StatusMessage": "

The reason for the status.

", - "ReservedInstancesModification$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModificationIdStringList$member": null, - "ReservedInstancesModificationResult$ReservedInstancesId": "

The ID for the Reserved Instances that were created as part of the modification request. This field is only available when the modification is fulfilled.

", - "ReservedInstancesOffering$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering.

", - "ReservedInstancesOffering$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesOfferingIdStringList$member": null, - "ResetImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ResetInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ResetNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ResetNetworkInterfaceAttributeRequest$SourceDestCheck": "

The source/destination checking attribute. Resets the value to true.

", - "ResetSnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ResourceIdList$member": null, - "ResponseHostIdList$member": null, - "RestorableByStringList$member": null, - "RestoreAddressToClassicRequest$PublicIp": "

The Elastic IP address.

", - "RestoreAddressToClassicResult$PublicIp": "

The Elastic IP address.

", - "RevokeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name or number. We recommend that you specify the protocol in a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. We recommend that you specify the CIDR range in a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "RevokeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a security group in a nondefault VPC.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. For EC2-VPC, the source security group must be in the same VPC. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic] The AWS account ID of the source security group, if the source security group is in a different account. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). Use -1 to specify all.

", - "RevokeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "Route$DestinationCidrBlock": "

The CIDR block used for the destination match.

", - "Route$DestinationPrefixListId": "

The prefix of the AWS service.

", - "Route$GatewayId": "

The ID of a gateway attached to your VPC.

", - "Route$InstanceId": "

The ID of a NAT instance in your VPC.

", - "Route$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "Route$NetworkInterfaceId": "

The ID of the network interface.

", - "Route$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "Route$NatGatewayId": "

The ID of a NAT gateway.

", - "RouteTable$RouteTableId": "

The ID of the route table.

", - "RouteTable$VpcId": "

The ID of the VPC.

", - "RouteTableAssociation$RouteTableAssociationId": "

The ID of the association between a route table and a subnet.

", - "RouteTableAssociation$RouteTableId": "

The ID of the route table.

", - "RouteTableAssociation$SubnetId": "

The ID of the subnet. A subnet ID is not returned for an implicit association.

", - "RunInstancesRequest$ImageId": "

The ID of the AMI, which you can get by calling DescribeImages.

", - "RunInstancesRequest$KeyName": "

The name of the key pair. You can create a key pair using CreateKeyPair or ImportKeyPair.

If you do not specify a key pair, you can't connect to the instance unless you choose an AMI that is configured to allow users another way to log in.

", - "RunInstancesRequest$UserData": "

Data to configure the instance, or a script to run during instance launch. For more information, see Running Commands on Your Linux Instance at Launch (Linux) and Adding User Data (Windows). For API calls, the text must be base64-encoded. For command line tools, the encoding is performed for you, and you can load the text from a file.

", - "RunInstancesRequest$KernelId": "

The ID of the kernel.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$RamdiskId": "

The ID of the RAM disk.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$SubnetId": "

[EC2-VPC] The ID of the subnet to launch the instance into.

", - "RunInstancesRequest$PrivateIpAddress": "

[EC2-VPC] The primary IP address. You must specify a value from the IP address range of the subnet.

Only one private IP address can be designated as primary. Therefore, you can't specify this parameter if PrivateIpAddresses.n.Primary is set to true and PrivateIpAddresses.n.PrivateIpAddress is set to an IP address.

Default: We select an IP address from the IP address range of the subnet.

", - "RunInstancesRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see Ensuring Idempotency.

Constraints: Maximum 64 ASCII characters

", - "RunInstancesRequest$AdditionalInfo": "

Reserved.

", - "RunScheduledInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that ensures the idempotency of the request. For more information, see Ensuring Idempotency.

", - "RunScheduledInstancesRequest$ScheduledInstanceId": "

The Scheduled Instance ID.

", - "S3Storage$Bucket": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "S3Storage$Prefix": "

The beginning of the file name of the AMI.

", - "S3Storage$AWSAccessKeyId": "

The access key ID of the owner of the bucket. Before you specify a value for your access key ID, review and follow the guidance in Best Practices for Managing AWS Access Keys.

", - "S3Storage$UploadPolicySignature": "

The signature of the Base64 encoded JSON document.

", - "ScheduledInstance$ScheduledInstanceId": "

The Scheduled Instance ID.

", - "ScheduledInstance$InstanceType": "

The instance type.

", - "ScheduledInstance$Platform": "

The platform (Linux/UNIX or Windows).

", - "ScheduledInstance$NetworkPlatform": "

The network platform (EC2-Classic or EC2-VPC).

", - "ScheduledInstance$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstance$HourlyPrice": "

The hourly price for a single instance.

", - "ScheduledInstanceAvailability$InstanceType": "

The instance type. You can specify one of the C3, C4, M4, or R3 instance types.

", - "ScheduledInstanceAvailability$Platform": "

The platform (Linux/UNIX or Windows).

", - "ScheduledInstanceAvailability$NetworkPlatform": "

The network platform (EC2-Classic or EC2-VPC).

", - "ScheduledInstanceAvailability$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstanceAvailability$PurchaseToken": "

The purchase token. This token expires in two hours.

", - "ScheduledInstanceAvailability$HourlyPrice": "

The hourly price for a single instance.

", - "ScheduledInstanceIdRequestSet$member": null, - "ScheduledInstanceRecurrence$Frequency": "

The frequency (Daily, Weekly, or Monthly).

", - "ScheduledInstanceRecurrence$OccurrenceUnit": "

The unit for occurrenceDaySet (DayOfWeek or DayOfMonth).

", - "ScheduledInstanceRecurrenceRequest$Frequency": "

The frequency (Daily, Weekly, or Monthly).

", - "ScheduledInstanceRecurrenceRequest$OccurrenceUnit": "

The unit for OccurrenceDays (DayOfWeek or DayOfMonth). This value is required for a monthly schedule. You can't specify DayOfWeek with a weekly schedule. You can't specify this value with a daily schedule.

", - "ScheduledInstancesBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "ScheduledInstancesBlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "ScheduledInstancesBlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with two available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "ScheduledInstancesEbs$SnapshotId": "

The ID of the snapshot.

", - "ScheduledInstancesEbs$VolumeType": "

The volume type. gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, Throughput Optimized HDD for st1, Cold HDD for sc1, or standard for Magnetic.

Default: standard

", - "ScheduledInstancesIamInstanceProfile$Arn": "

The Amazon Resource Name (ARN).

", - "ScheduledInstancesIamInstanceProfile$Name": "

The name.

", - "ScheduledInstancesLaunchSpecification$ImageId": "

The ID of the Amazon Machine Image (AMI).

", - "ScheduledInstancesLaunchSpecification$KeyName": "

The name of the key pair.

", - "ScheduledInstancesLaunchSpecification$UserData": "

The base64-encoded MIME user data.

", - "ScheduledInstancesLaunchSpecification$KernelId": "

The ID of the kernel.

", - "ScheduledInstancesLaunchSpecification$InstanceType": "

The instance type.

", - "ScheduledInstancesLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "ScheduledInstancesLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances.

", - "ScheduledInstancesNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "ScheduledInstancesNetworkInterface$SubnetId": "

The ID of the subnet.

", - "ScheduledInstancesNetworkInterface$Description": "

The description.

", - "ScheduledInstancesNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "ScheduledInstancesPlacement$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstancesPlacement$GroupName": "

The name of the placement group.

", - "ScheduledInstancesPrivateIpAddressConfig$PrivateIpAddress": "

The IP address.

", - "ScheduledInstancesSecurityGroupIdSet$member": null, - "SecurityGroup$OwnerId": "

The AWS account ID of the owner of the security group.

", - "SecurityGroup$GroupName": "

The name of the security group.

", - "SecurityGroup$GroupId": "

The ID of the security group.

", - "SecurityGroup$Description": "

A description of the security group.

", - "SecurityGroup$VpcId": "

[EC2-VPC] The ID of the VPC for the security group.

", - "SecurityGroupIdStringList$member": null, - "SecurityGroupReference$GroupId": "

The ID of your security group.

", - "SecurityGroupReference$ReferencingVpcId": "

The ID of the VPC with the referencing security group.

", - "SecurityGroupReference$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "SecurityGroupStringList$member": null, - "Snapshot$SnapshotId": "

The ID of the snapshot. Each snapshot receives a unique identifier when it is created.

", - "Snapshot$VolumeId": "

The ID of the volume that was used to create the snapshot.

", - "Snapshot$StateMessage": "

Encrypted Amazon EBS snapshots are copied asynchronously. If a snapshot copy operation fails (for example, if the proper AWS Key Management Service (AWS KMS) permissions are not obtained) this field displays error state details to help you diagnose why the error occurred. This parameter is only returned by the DescribeSnapshots API operation.

", - "Snapshot$Progress": "

The progress of the snapshot, as a percentage.

", - "Snapshot$OwnerId": "

The AWS account ID of the EBS snapshot owner.

", - "Snapshot$Description": "

The description for the snapshot.

", - "Snapshot$OwnerAlias": "

The AWS account alias (for example, amazon, self) or AWS account ID that owns the snapshot.

", - "Snapshot$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the parent volume.

", - "Snapshot$DataEncryptionKeyId": "

The data encryption key identifier for the snapshot. This value is a unique identifier that corresponds to the data encryption key that was used to encrypt the original volume or snapshot copy. Because data encryption keys are inherited by volumes created from snapshots, and vice versa, if snapshots share the same data encryption key identifier, then they belong to the same volume/snapshot lineage. This parameter is only returned by the DescribeSnapshots API operation.

", - "SnapshotDetail$Description": "

A description for the snapshot.

", - "SnapshotDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotDetail$Url": "

The URL used to access the disk image.

", - "SnapshotDetail$DeviceName": "

The block device mapping for the snapshot.

", - "SnapshotDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotDetail$Progress": "

The percentage of progress for the task.

", - "SnapshotDetail$StatusMessage": "

A detailed status message for the snapshot creation.

", - "SnapshotDetail$Status": "

A brief status of the snapshot creation.

", - "SnapshotDiskContainer$Description": "

The description of the disk image being imported.

", - "SnapshotDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "SnapshotDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. It can either be a https URL (https://..) or an Amazon S3 URL (s3://..).

", - "SnapshotIdStringList$member": null, - "SnapshotTaskDetail$Description": "

The description of the snapshot.

", - "SnapshotTaskDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$Url": "

The URL of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotTaskDetail$Progress": "

The percentage of completion for the import snapshot task.

", - "SnapshotTaskDetail$StatusMessage": "

A detailed status message for the import snapshot task.

", - "SnapshotTaskDetail$Status": "

A brief status for the import snapshot task.

", - "SpotDatafeedSubscription$OwnerId": "

The AWS account ID of the account.

", - "SpotDatafeedSubscription$Bucket": "

The Amazon S3 bucket where the Spot instance data feed is located.

", - "SpotDatafeedSubscription$Prefix": "

The prefix that is prepended to data feed files.

", - "SpotFleetLaunchSpecification$ImageId": "

The ID of the AMI.

", - "SpotFleetLaunchSpecification$KeyName": "

The name of the key pair.

", - "SpotFleetLaunchSpecification$UserData": "

The Base64-encoded MIME user data to make available to the instances.

", - "SpotFleetLaunchSpecification$AddressingType": "

Deprecated.

", - "SpotFleetLaunchSpecification$KernelId": "

The ID of the kernel.

", - "SpotFleetLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "SpotFleetLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances. To specify multiple subnets, separate them using commas; for example, \"subnet-a61dafcf, subnet-65ea5f08\".

", - "SpotFleetLaunchSpecification$SpotPrice": "

The bid price per unit hour for the specified instance type. If this value is not specified, the default is the Spot bid price specified for the fleet. To determine the bid price per unit hour, divide the Spot bid price by the value of WeightedCapacity.

", - "SpotFleetRequestConfig$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "SpotFleetRequestConfigData$ClientToken": "

A unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "SpotFleetRequestConfigData$SpotPrice": "

The bid price per unit hour.

", - "SpotFleetRequestConfigData$IamFleetRole": "

Grants the Spot fleet permission to terminate Spot instances on your behalf when you cancel its Spot fleet request using CancelSpotFleetRequests or when the Spot fleet request expires, if you set terminateInstancesWithExpiration.

", - "SpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "SpotInstanceRequest$SpotPrice": "

The maximum hourly price (bid) for the Spot instance launched to fulfill the request.

", - "SpotInstanceRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

", - "SpotInstanceRequest$AvailabilityZoneGroup": "

The Availability Zone group. If you specify the same Availability Zone group for all Spot instance requests, all Spot instances are launched in the same Availability Zone.

", - "SpotInstanceRequest$InstanceId": "

The instance ID, if an instance has been launched to fulfill the Spot instance request.

", - "SpotInstanceRequest$ActualBlockHourlyPrice": "

If you specified a duration and your Spot instance request was fulfilled, this is the fixed hourly price in effect for the Spot instance while it runs.

", - "SpotInstanceRequest$LaunchedAvailabilityZone": "

The Availability Zone in which the bid is launched.

", - "SpotInstanceRequestIdList$member": null, - "SpotInstanceStateFault$Code": "

The reason code for the Spot instance state change.

", - "SpotInstanceStateFault$Message": "

The message for the Spot instance state change.

", - "SpotInstanceStatus$Code": "

The status code. For a list of status codes, see Spot Bid Status Codes in the Amazon Elastic Compute Cloud User Guide.

", - "SpotInstanceStatus$Message": "

The description for the status code.

", - "SpotPlacement$AvailabilityZone": "

The Availability Zone.

[Spot fleet only] To specify multiple Availability Zones, separate them using commas; for example, \"us-west-2a, us-west-2b\".

", - "SpotPlacement$GroupName": "

The name of the placement group (for cluster instances).

", - "SpotPrice$SpotPrice": "

The maximum price (bid) that you are willing to pay for a Spot instance.

", - "SpotPrice$AvailabilityZone": "

The Availability Zone.

", - "StaleIpPermission$IpProtocol": "

The IP protocol name (for tcp, udp, and icmp) or number (see Protocol Numbers).

", - "StaleSecurityGroup$GroupId": "

The ID of the security group.

", - "StaleSecurityGroup$GroupName": "

The name of the security group.

", - "StaleSecurityGroup$Description": "

The description of the security group.

", - "StaleSecurityGroup$VpcId": "

The ID of the VPC for the security group.

", - "StartInstancesRequest$AdditionalInfo": "

Reserved.

", - "StateReason$Code": "

The reason code for the state change.

", - "StateReason$Message": "

The message for the state change.

  • Server.SpotInstanceTermination: A Spot instance was terminated due to an increase in the market price.

  • Server.InternalError: An internal error occurred during instance launch, resulting in termination.

  • Server.InsufficientInstanceCapacity: There was insufficient instance capacity to satisfy the launch request.

  • Client.InternalError: A client error caused the instance to terminate on launch.

  • Client.InstanceInitiatedShutdown: The instance was shut down using the shutdown -h command from the instance.

  • Client.UserInitiatedShutdown: The instance was shut down using the Amazon EC2 API.

  • Client.VolumeLimitExceeded: The limit on the number of EBS volumes or total storage was exceeded. Decrease usage or request an increase in your limits.

  • Client.InvalidSnapshot.NotFound: The specified snapshot was not found.

", - "Subnet$SubnetId": "

The ID of the subnet.

", - "Subnet$VpcId": "

The ID of the VPC the subnet is in.

", - "Subnet$CidrBlock": "

The CIDR block assigned to the subnet.

", - "Subnet$AvailabilityZone": "

The Availability Zone of the subnet.

", - "SubnetIdStringList$member": null, - "Tag$Key": "

The key of the tag.

Constraints: Tag keys are case-sensitive and accept a maximum of 127 Unicode characters. May not begin with aws:

", - "Tag$Value": "

The value of the tag.

Constraints: Tag values are case-sensitive and accept a maximum of 255 Unicode characters.

", - "TagDescription$ResourceId": "

The ID of the resource. For example, ami-1a2b3c4d.

", - "TagDescription$Key": "

The tag key.

", - "TagDescription$Value": "

The tag value.

", - "UnassignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "UnsuccessfulItem$ResourceId": "

The ID of the resource.

", - "UnsuccessfulItemError$Code": "

The error code.

", - "UnsuccessfulItemError$Message": "

The error message accompanying the error code.

", - "UserBucket$S3Bucket": "

The name of the S3 bucket where the disk image is located.

", - "UserBucket$S3Key": "

The file name of the disk image.

", - "UserBucketDetails$S3Bucket": "

The S3 bucket from which the disk image was created.

", - "UserBucketDetails$S3Key": "

The file name of the disk image.

", - "UserData$Data": "

The Base64-encoded MIME user data for the instance.

", - "UserGroupStringList$member": null, - "UserIdGroupPair$UserId": "

The ID of an AWS account. For a referenced security group in another VPC, the account ID of the referenced security group is returned.

[EC2-Classic] Required when adding or removing rules that reference a security group in another AWS account.

", - "UserIdGroupPair$GroupName": "

The name of the security group. In a request, use this parameter for a security group in EC2-Classic or a default VPC only. For a security group in a nondefault VPC, use the security group ID.

", - "UserIdGroupPair$GroupId": "

The ID of the security group.

", - "UserIdGroupPair$VpcId": "

The ID of the VPC for the referenced security group, if applicable.

", - "UserIdGroupPair$VpcPeeringConnectionId": "

The ID of the VPC peering connection, if applicable.

", - "UserIdGroupPair$PeeringStatus": "

The status of a VPC peering connection, if applicable.

", - "UserIdStringList$member": null, - "ValueStringList$member": null, - "VgwTelemetry$OutsideIpAddress": "

The Internet-routable IP address of the virtual private gateway's outside interface.

", - "VgwTelemetry$StatusMessage": "

If an error occurs, a description of the error.

", - "Volume$VolumeId": "

The ID of the volume.

", - "Volume$SnapshotId": "

The snapshot from which the volume was created, if applicable.

", - "Volume$AvailabilityZone": "

The Availability Zone for the volume.

", - "Volume$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the volume.

", - "VolumeAttachment$VolumeId": "

The ID of the volume.

", - "VolumeAttachment$InstanceId": "

The ID of the instance.

", - "VolumeAttachment$Device": "

The device name.

", - "VolumeIdStringList$member": null, - "VolumeStatusAction$Code": "

The code identifying the operation, for example, enable-volume-io.

", - "VolumeStatusAction$Description": "

A description of the operation.

", - "VolumeStatusAction$EventType": "

The event type associated with this operation.

", - "VolumeStatusAction$EventId": "

The ID of the event associated with this operation.

", - "VolumeStatusDetails$Status": "

The intended status of the volume status.

", - "VolumeStatusEvent$EventType": "

The type of this event.

", - "VolumeStatusEvent$Description": "

A description of the event.

", - "VolumeStatusEvent$EventId": "

The ID of this event.

", - "VolumeStatusItem$VolumeId": "

The volume ID.

", - "VolumeStatusItem$AvailabilityZone": "

The Availability Zone of the volume.

", - "Vpc$VpcId": "

The ID of the VPC.

", - "Vpc$CidrBlock": "

The CIDR block for the VPC.

", - "Vpc$DhcpOptionsId": "

The ID of the set of DHCP options you've associated with the VPC (or default if the default options are associated with the VPC).

", - "VpcAttachment$VpcId": "

The ID of the VPC.

", - "VpcClassicLink$VpcId": "

The ID of the VPC.

", - "VpcClassicLinkIdList$member": null, - "VpcEndpoint$VpcEndpointId": "

The ID of the VPC endpoint.

", - "VpcEndpoint$VpcId": "

The ID of the VPC to which the endpoint is associated.

", - "VpcEndpoint$ServiceName": "

The name of the AWS service to which the endpoint is associated.

", - "VpcEndpoint$PolicyDocument": "

The policy document associated with the endpoint.

", - "VpcIdStringList$member": null, - "VpcPeeringConnection$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "VpcPeeringConnectionStateReason$Message": "

A message that provides more information about the status, if applicable.

", - "VpcPeeringConnectionVpcInfo$CidrBlock": "

The CIDR block for the VPC.

", - "VpcPeeringConnectionVpcInfo$OwnerId": "

The AWS account ID of the VPC owner.

", - "VpcPeeringConnectionVpcInfo$VpcId": "

The ID of the VPC.

", - "VpnConnection$VpnConnectionId": "

The ID of the VPN connection.

", - "VpnConnection$CustomerGatewayConfiguration": "

The configuration information for the VPN connection's customer gateway (in the native XML format). This element is always present in the CreateVpnConnection response; however, it's present in the DescribeVpnConnections response only if the VPN connection is in the pending or available state.

", - "VpnConnection$CustomerGatewayId": "

The ID of the customer gateway at your end of the VPN connection.

", - "VpnConnection$VpnGatewayId": "

The ID of the virtual private gateway at the AWS side of the VPN connection.

", - "VpnConnectionIdStringList$member": null, - "VpnGateway$VpnGatewayId": "

The ID of the virtual private gateway.

", - "VpnGateway$AvailabilityZone": "

The Availability Zone where the virtual private gateway was created, if applicable. This field may be empty or not returned.

", - "VpnGatewayIdStringList$member": null, - "VpnStaticRoute$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer data center.

", - "ZoneNameStringList$member": null - } - }, - "Subnet": { - "base": "

Describes a subnet.

", - "refs": { - "CreateSubnetResult$Subnet": "

Information about the subnet.

", - "SubnetList$member": null - } - }, - "SubnetIdStringList": { - "base": null, - "refs": { - "DescribeSubnetsRequest$SubnetIds": "

One or more subnet IDs.

Default: Describes all your subnets.

" - } - }, - "SubnetList": { - "base": null, - "refs": { - "DescribeSubnetsResult$Subnets": "

Information about one or more subnets.

" - } - }, - "SubnetState": { - "base": null, - "refs": { - "Subnet$State": "

The current state of the subnet.

" - } - }, - "SummaryStatus": { - "base": null, - "refs": { - "InstanceStatusSummary$Status": "

The status.

" - } - }, - "Tag": { - "base": "

Describes a tag.

", - "refs": { - "TagList$member": null - } - }, - "TagDescription": { - "base": "

Describes a tag.

", - "refs": { - "TagDescriptionList$member": null - } - }, - "TagDescriptionList": { - "base": null, - "refs": { - "DescribeTagsResult$Tags": "

A list of tags.

" - } - }, - "TagList": { - "base": null, - "refs": { - "ClassicLinkInstance$Tags": "

Any tags assigned to the instance.

", - "ConversionTask$Tags": "

Any tags assigned to the task.

", - "CreateTagsRequest$Tags": "

One or more tags. The value parameter is required, but if you don't want the tag to have a value, specify the parameter with no value, and we set the value to an empty string.

", - "CustomerGateway$Tags": "

Any tags assigned to the customer gateway.

", - "DeleteTagsRequest$Tags": "

One or more tags to delete. If you omit the value parameter, we delete the tag regardless of its value. If you specify this parameter with an empty string as the value, we delete the key only if its value is an empty string.

", - "DhcpOptions$Tags": "

Any tags assigned to the DHCP options set.

", - "Image$Tags": "

Any tags assigned to the image.

", - "Instance$Tags": "

Any tags assigned to the instance.

", - "InternetGateway$Tags": "

Any tags assigned to the Internet gateway.

", - "NetworkAcl$Tags": "

Any tags assigned to the network ACL.

", - "NetworkInterface$TagSet": "

Any tags assigned to the network interface.

", - "ReservedInstances$Tags": "

Any tags assigned to the resource.

", - "ReservedInstancesListing$Tags": "

Any tags assigned to the resource.

", - "RouteTable$Tags": "

Any tags assigned to the route table.

", - "SecurityGroup$Tags": "

Any tags assigned to the security group.

", - "Snapshot$Tags": "

Any tags assigned to the snapshot.

", - "SpotInstanceRequest$Tags": "

Any tags assigned to the resource.

", - "Subnet$Tags": "

Any tags assigned to the subnet.

", - "Volume$Tags": "

Any tags assigned to the volume.

", - "Vpc$Tags": "

Any tags assigned to the VPC.

", - "VpcClassicLink$Tags": "

Any tags assigned to the VPC.

", - "VpcPeeringConnection$Tags": "

Any tags assigned to the resource.

", - "VpnConnection$Tags": "

Any tags assigned to the VPN connection.

", - "VpnGateway$Tags": "

Any tags assigned to the virtual private gateway.

" - } - }, - "TelemetryStatus": { - "base": null, - "refs": { - "VgwTelemetry$Status": "

The status of the VPN tunnel.

" - } - }, - "Tenancy": { - "base": null, - "refs": { - "CreateVpcRequest$InstanceTenancy": "

The tenancy options for instances launched into the VPC. For default, instances are launched with shared tenancy by default. You can launch instances with any tenancy into a shared tenancy VPC. For dedicated, instances are launched as dedicated tenancy instances by default. You can only launch instances with a tenancy of dedicated or host into a dedicated tenancy VPC.

Important: The host value cannot be used with this parameter. Use the default or dedicated values only.

Default: default

", - "DescribeReservedInstancesOfferingsRequest$InstanceTenancy": "

The tenancy of the instances covered by the reservation. A Reserved Instance with a tenancy of dedicated is applied to instances that run in a VPC on single-tenant hardware (i.e., Dedicated Instances).

Default: default

", - "Placement$Tenancy": "

The tenancy of the instance (if the instance is running in a VPC). An instance with a tenancy of dedicated runs on single-tenant hardware. The host tenancy is not supported for the ImportInstance command.

", - "ReservedInstances$InstanceTenancy": "

The tenancy of the instance.

", - "ReservedInstancesOffering$InstanceTenancy": "

The tenancy of the instance.

", - "Vpc$InstanceTenancy": "

The allowed tenancy of instances launched into the VPC.

" - } - }, - "TerminateInstancesRequest": { - "base": "

Contains the parameters for TerminateInstances.

", - "refs": { - } - }, - "TerminateInstancesResult": { - "base": "

Contains the output of TerminateInstances.

", - "refs": { - } - }, - "TrafficType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$TrafficType": "

The type of traffic to log.

", - "FlowLog$TrafficType": "

The type of traffic captured for the flow log.

" - } - }, - "UnassignPrivateIpAddressesRequest": { - "base": "

Contains the parameters for UnassignPrivateIpAddresses.

", - "refs": { - } - }, - "UnmonitorInstancesRequest": { - "base": "

Contains the parameters for UnmonitorInstances.

", - "refs": { - } - }, - "UnmonitorInstancesResult": { - "base": "

Contains the output of UnmonitorInstances.

", - "refs": { - } - }, - "UnsuccessfulItem": { - "base": "

Information about items that were not successfully processed in a batch call.

", - "refs": { - "UnsuccessfulItemList$member": null, - "UnsuccessfulItemSet$member": null - } - }, - "UnsuccessfulItemError": { - "base": "

Information about the error that occurred. For more information about errors, see Error Codes.

", - "refs": { - "UnsuccessfulItem$Error": "

Information about the error.

" - } - }, - "UnsuccessfulItemList": { - "base": null, - "refs": { - "ModifyHostsResult$Unsuccessful": "

The IDs of the Dedicated hosts that could not be modified. Check whether the setting you requested can be used.

", - "ReleaseHostsResult$Unsuccessful": "

The IDs of the Dedicated hosts that could not be released, including an error message.

" - } - }, - "UnsuccessfulItemSet": { - "base": null, - "refs": { - "CreateFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be created successfully.

", - "DeleteFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be deleted successfully.

", - "DeleteVpcEndpointsResult$Unsuccessful": "

Information about the endpoints that were not successfully deleted.

" - } - }, - "UserBucket": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "ImageDiskContainer$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotDiskContainer$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserBucketDetails": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "SnapshotDetail$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotTaskDetail$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserData": { - "base": "

Describes the user data to be made available to an instance.

", - "refs": { - "ImportInstanceLaunchSpecification$UserData": "

The Base64-encoded MIME user data to be made available to the instance.

" - } - }, - "UserGroupStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserGroups": "

One or more user groups. This is only valid when modifying the launchPermission attribute.

" - } - }, - "UserIdGroupPair": { - "base": "

Describes a security group and AWS account ID pair.

", - "refs": { - "UserIdGroupPairList$member": null, - "UserIdGroupPairSet$member": null - } - }, - "UserIdGroupPairList": { - "base": null, - "refs": { - "IpPermission$UserIdGroupPairs": "

One or more security group and AWS account ID pairs.

" - } - }, - "UserIdGroupPairSet": { - "base": null, - "refs": { - "StaleIpPermission$UserIdGroupPairs": "

One or more security group pairs. Returns the ID of the referenced security group and VPC, and the ID and status of the VPC peering connection.

" - } - }, - "UserIdStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserIds": "

One or more AWS account IDs. This is only valid when modifying the launchPermission attribute.

", - "ModifySnapshotAttributeRequest$UserIds": "

The account ID to modify for the snapshot.

" - } - }, - "ValueStringList": { - "base": null, - "refs": { - "CancelSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "CreateFlowLogsRequest$ResourceIds": "

One or more subnet, network interface, or VPC IDs.

Constraints: Maximum of 1000 resources

", - "CreateFlowLogsResult$FlowLogIds": "

The IDs of the flow logs.

", - "CreateVpcEndpointRequest$RouteTableIds": "

One or more route table IDs.

", - "DeleteFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DeleteVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DescribeInternetGatewaysRequest$InternetGatewayIds": "

One or more Internet gateway IDs.

Default: Describes all your Internet gateways.

", - "DescribeMovingAddressesRequest$PublicIps": "

One or more Elastic IP addresses.

", - "DescribeNatGatewaysRequest$NatGatewayIds": "

One or more NAT gateway IDs.

", - "DescribeNetworkAclsRequest$NetworkAclIds": "

One or more network ACL IDs.

Default: Describes all your network ACLs.

", - "DescribePrefixListsRequest$PrefixListIds": "

One or more prefix list IDs.

", - "DescribeRouteTablesRequest$RouteTableIds": "

One or more route table IDs.

Default: Describes all your route tables.

", - "DescribeSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "DescribeVpcEndpointServicesResult$ServiceNames": "

A list of supported AWS services.

", - "DescribeVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeVpcPeeringConnectionsRequest$VpcPeeringConnectionIds": "

One or more VPC peering connection IDs.

Default: Describes all your VPC peering connections.

", - "Filter$Values": "

One or more filter values. Filter values are case-sensitive.

", - "ModifyVpcEndpointRequest$AddRouteTableIds": "

One or more route tables IDs to associate with the endpoint.

", - "ModifyVpcEndpointRequest$RemoveRouteTableIds": "

One or more route table IDs to disassociate from the endpoint.

", - "NewDhcpConfiguration$Values": null, - "PrefixList$Cidrs": "

The IP address range of the AWS service.

", - "RequestSpotLaunchSpecification$SecurityGroups": null, - "RequestSpotLaunchSpecification$SecurityGroupIds": null, - "VpcEndpoint$RouteTableIds": "

One or more route tables associated with the endpoint.

" - } - }, - "VgwTelemetry": { - "base": "

Describes telemetry for a VPN tunnel.

", - "refs": { - "VgwTelemetryList$member": null - } - }, - "VgwTelemetryList": { - "base": null, - "refs": { - "VpnConnection$VgwTelemetry": "

Information about the VPN tunnel.

" - } - }, - "VirtualizationType": { - "base": null, - "refs": { - "Image$VirtualizationType": "

The type of virtualization of the AMI.

", - "Instance$VirtualizationType": "

The virtualization type of the instance.

" - } - }, - "Volume": { - "base": "

Describes a volume.

", - "refs": { - "VolumeList$member": null - } - }, - "VolumeAttachment": { - "base": "

Describes volume attachment details.

", - "refs": { - "VolumeAttachmentList$member": null - } - }, - "VolumeAttachmentList": { - "base": null, - "refs": { - "Volume$Attachments": "

Information about the volume attachments.

" - } - }, - "VolumeAttachmentState": { - "base": null, - "refs": { - "VolumeAttachment$State": "

The attachment state of the volume.

" - } - }, - "VolumeAttributeName": { - "base": null, - "refs": { - "DescribeVolumeAttributeRequest$Attribute": "

The instance attribute.

" - } - }, - "VolumeDetail": { - "base": "

Describes an EBS volume.

", - "refs": { - "DiskImage$Volume": "

Information about the volume.

", - "ImportVolumeRequest$Volume": "

The volume size.

" - } - }, - "VolumeIdStringList": { - "base": null, - "refs": { - "DescribeVolumeStatusRequest$VolumeIds": "

One or more volume IDs.

Default: Describes all your volumes.

", - "DescribeVolumesRequest$VolumeIds": "

One or more volume IDs.

" - } - }, - "VolumeList": { - "base": null, - "refs": { - "DescribeVolumesResult$Volumes": "

Information about the volumes.

" - } - }, - "VolumeState": { - "base": null, - "refs": { - "Volume$State": "

The volume state.

" - } - }, - "VolumeStatusAction": { - "base": "

Describes a volume status operation code.

", - "refs": { - "VolumeStatusActionsList$member": null - } - }, - "VolumeStatusActionsList": { - "base": null, - "refs": { - "VolumeStatusItem$Actions": "

The details of the operation.

" - } - }, - "VolumeStatusDetails": { - "base": "

Describes a volume status.

", - "refs": { - "VolumeStatusDetailsList$member": null - } - }, - "VolumeStatusDetailsList": { - "base": null, - "refs": { - "VolumeStatusInfo$Details": "

The details of the volume status.

" - } - }, - "VolumeStatusEvent": { - "base": "

Describes a volume status event.

", - "refs": { - "VolumeStatusEventsList$member": null - } - }, - "VolumeStatusEventsList": { - "base": null, - "refs": { - "VolumeStatusItem$Events": "

A list of events associated with the volume.

" - } - }, - "VolumeStatusInfo": { - "base": "

Describes the status of a volume.

", - "refs": { - "VolumeStatusItem$VolumeStatus": "

The volume status.

" - } - }, - "VolumeStatusInfoStatus": { - "base": null, - "refs": { - "VolumeStatusInfo$Status": "

The status of the volume.

" - } - }, - "VolumeStatusItem": { - "base": "

Describes the volume status.

", - "refs": { - "VolumeStatusList$member": null - } - }, - "VolumeStatusList": { - "base": null, - "refs": { - "DescribeVolumeStatusResult$VolumeStatuses": "

A list of volumes.

" - } - }, - "VolumeStatusName": { - "base": null, - "refs": { - "VolumeStatusDetails$Name": "

The name of the volume status.

" - } - }, - "VolumeType": { - "base": null, - "refs": { - "CreateVolumeRequest$VolumeType": "

The volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

Default: standard

", - "EbsBlockDevice$VolumeType": "

The volume type: gp2, io1, st1, sc1, or standard.

Default: standard

", - "Volume$VolumeType": "

The volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

" - } - }, - "Vpc": { - "base": "

Describes a VPC.

", - "refs": { - "CreateVpcResult$Vpc": "

Information about the VPC.

", - "VpcList$member": null - } - }, - "VpcAttachment": { - "base": "

Describes an attachment between a virtual private gateway and a VPC.

", - "refs": { - "AttachVpnGatewayResult$VpcAttachment": "

Information about the attachment.

", - "VpcAttachmentList$member": null - } - }, - "VpcAttachmentList": { - "base": null, - "refs": { - "VpnGateway$VpcAttachments": "

Any VPCs attached to the virtual private gateway.

" - } - }, - "VpcAttributeName": { - "base": null, - "refs": { - "DescribeVpcAttributeRequest$Attribute": "

The VPC attribute.

" - } - }, - "VpcClassicLink": { - "base": "

Describes whether a VPC is enabled for ClassicLink.

", - "refs": { - "VpcClassicLinkList$member": null - } - }, - "VpcClassicLinkIdList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkDnsSupportRequest$VpcIds": "

One or more VPC IDs.

", - "DescribeVpcClassicLinkRequest$VpcIds": "

One or more VPCs for which you want to describe the ClassicLink status.

" - } - }, - "VpcClassicLinkList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkResult$Vpcs": "

The ClassicLink status of one or more VPCs.

" - } - }, - "VpcEndpoint": { - "base": "

Describes a VPC endpoint.

", - "refs": { - "CreateVpcEndpointResult$VpcEndpoint": "

Information about the endpoint.

", - "VpcEndpointSet$member": null - } - }, - "VpcEndpointSet": { - "base": null, - "refs": { - "DescribeVpcEndpointsResult$VpcEndpoints": "

Information about the endpoints.

" - } - }, - "VpcIdStringList": { - "base": null, - "refs": { - "DescribeVpcsRequest$VpcIds": "

One or more VPC IDs.

Default: Describes all your VPCs.

" - } - }, - "VpcList": { - "base": null, - "refs": { - "DescribeVpcsResult$Vpcs": "

Information about one or more VPCs.

" - } - }, - "VpcPeeringConnection": { - "base": "

Describes a VPC peering connection.

", - "refs": { - "AcceptVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "CreateVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "VpcPeeringConnectionList$member": null - } - }, - "VpcPeeringConnectionList": { - "base": null, - "refs": { - "DescribeVpcPeeringConnectionsResult$VpcPeeringConnections": "

Information about the VPC peering connections.

" - } - }, - "VpcPeeringConnectionOptionsDescription": { - "base": "

Describes the VPC peering connection options.

", - "refs": { - "VpcPeeringConnectionVpcInfo$PeeringOptions": "

Information about the VPC peering connection options for the accepter or requester VPC.

" - } - }, - "VpcPeeringConnectionStateReason": { - "base": "

Describes the status of a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$Status": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionStateReasonCode": { - "base": null, - "refs": { - "VpcPeeringConnectionStateReason$Code": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionVpcInfo": { - "base": "

Describes a VPC in a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$AccepterVpcInfo": "

Information about the accepter VPC. CIDR block information is not returned when creating a VPC peering connection, or when describing a VPC peering connection that's in the initiating-request or pending-acceptance state.

", - "VpcPeeringConnection$RequesterVpcInfo": "

Information about the requester VPC.

" - } - }, - "VpcState": { - "base": null, - "refs": { - "Vpc$State": "

The current state of the VPC.

" - } - }, - "VpnConnection": { - "base": "

Describes a VPN connection.

", - "refs": { - "CreateVpnConnectionResult$VpnConnection": "

Information about the VPN connection.

", - "VpnConnectionList$member": null - } - }, - "VpnConnectionIdStringList": { - "base": null, - "refs": { - "DescribeVpnConnectionsRequest$VpnConnectionIds": "

One or more VPN connection IDs.

Default: Describes your VPN connections.

" - } - }, - "VpnConnectionList": { - "base": null, - "refs": { - "DescribeVpnConnectionsResult$VpnConnections": "

Information about one or more VPN connections.

" - } - }, - "VpnConnectionOptions": { - "base": "

Describes VPN connection options.

", - "refs": { - "VpnConnection$Options": "

The VPN connection options.

" - } - }, - "VpnConnectionOptionsSpecification": { - "base": "

Describes VPN connection options.

", - "refs": { - "CreateVpnConnectionRequest$Options": "

Indicates whether the VPN connection requires static routes. If you are creating a VPN connection for a device that does not support BGP, you must specify true.

Default: false

" - } - }, - "VpnGateway": { - "base": "

Describes a virtual private gateway.

", - "refs": { - "CreateVpnGatewayResult$VpnGateway": "

Information about the virtual private gateway.

", - "VpnGatewayList$member": null - } - }, - "VpnGatewayIdStringList": { - "base": null, - "refs": { - "DescribeVpnGatewaysRequest$VpnGatewayIds": "

One or more virtual private gateway IDs.

Default: Describes all your virtual private gateways.

" - } - }, - "VpnGatewayList": { - "base": null, - "refs": { - "DescribeVpnGatewaysResult$VpnGateways": "

Information about one or more virtual private gateways.

" - } - }, - "VpnState": { - "base": null, - "refs": { - "VpnConnection$State": "

The current state of the VPN connection.

", - "VpnGateway$State": "

The current state of the virtual private gateway.

", - "VpnStaticRoute$State": "

The current state of the static route.

" - } - }, - "VpnStaticRoute": { - "base": "

Describes a static route for a VPN connection.

", - "refs": { - "VpnStaticRouteList$member": null - } - }, - "VpnStaticRouteList": { - "base": null, - "refs": { - "VpnConnection$Routes": "

The static routes associated with the VPN connection.

" - } - }, - "VpnStaticRouteSource": { - "base": null, - "refs": { - "VpnStaticRoute$Source": "

Indicates how the routes were provided.

" - } - }, - "ZoneNameStringList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesRequest$ZoneNames": "

The names of one or more Availability Zones.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/paginators-1.json deleted file mode 100644 index 9d04d89ab..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/paginators-1.json +++ /dev/null @@ -1,138 +0,0 @@ -{ - "pagination": { - "DescribeAccountAttributes": { - "result_key": "AccountAttributes" - }, - "DescribeAddresses": { - "result_key": "Addresses" - }, - "DescribeAvailabilityZones": { - "result_key": "AvailabilityZones" - }, - "DescribeBundleTasks": { - "result_key": "BundleTasks" - }, - "DescribeConversionTasks": { - "result_key": "ConversionTasks" - }, - "DescribeCustomerGateways": { - "result_key": "CustomerGateways" - }, - "DescribeDhcpOptions": { - "result_key": "DhcpOptions" - }, - "DescribeExportTasks": { - "result_key": "ExportTasks" - }, - "DescribeImages": { - "result_key": "Images" - }, - "DescribeInstanceStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "InstanceStatuses" - }, - "DescribeInstances": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Reservations" - }, - "DescribeInternetGateways": { - "result_key": "InternetGateways" - }, - "DescribeKeyPairs": { - "result_key": "KeyPairs" - }, - "DescribeNetworkAcls": { - "result_key": "NetworkAcls" - }, - "DescribeNetworkInterfaces": { - "result_key": "NetworkInterfaces" - }, - "DescribePlacementGroups": { - "result_key": "PlacementGroups" - }, - "DescribeRegions": { - "result_key": "Regions" - }, - "DescribeReservedInstances": { - "result_key": "ReservedInstances" - }, - "DescribeReservedInstancesListings": { - "result_key": "ReservedInstancesListings" - }, - "DescribeReservedInstancesOfferings": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "ReservedInstancesOfferings" - }, - "DescribeReservedInstancesModifications": { - "input_token": "NextToken", - "output_token": "NextToken", - "result_key": "ReservedInstancesModifications" - }, - "DescribeRouteTables": { - "result_key": "RouteTables" - }, - "DescribeSecurityGroups": { - "result_key": "SecurityGroups" - }, - "DescribeSnapshots": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Snapshots" - }, - "DescribeSpotInstanceRequests": { - "result_key": "SpotInstanceRequests" - }, - "DescribeSpotFleetRequests": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotFleetRequestConfigs" - }, - "DescribeSpotPriceHistory": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotPriceHistory" - }, - "DescribeSubnets": { - "result_key": "Subnets" - }, - "DescribeTags": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Tags" - }, - "DescribeVolumeStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "VolumeStatuses" - }, - "DescribeVolumes": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Volumes" - }, - "DescribeVpcs": { - "result_key": "Vpcs" - }, - "DescribeVpcPeeringConnections": { - "result_key": "VpcPeeringConnections" - }, - "DescribeVpnConnections": { - "result_key": "VpnConnections" - }, - "DescribeVpnGateways": { - "result_key": "VpnGateways" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/waiters-2.json deleted file mode 100644 index ecc9f1b6f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2015-10-01/waiters-2.json +++ /dev/null @@ -1,593 +0,0 @@ -{ - "version": 2, - "waiters": { - "InstanceExists": { - "delay": 5, - "maxAttempts": 40, - "operation": "DescribeInstances", - "acceptors": [ - { - "matcher": "path", - "expected": true, - "argument": "length(Reservations[]) > `0`", - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "BundleTaskComplete": { - "delay": 15, - "operation": "DescribeBundleTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "complete", - "matcher": "pathAll", - "state": "success", - "argument": "BundleTasks[].State" - }, - { - "expected": "failed", - "matcher": "pathAny", - "state": "failure", - "argument": "BundleTasks[].State" - } - ] - }, - "ConversionTaskCancelled": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskCompleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelled", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelling", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskDeleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "CustomerGatewayAvailable": { - "delay": 15, - "operation": "DescribeCustomerGateways", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - } - ] - }, - "ExportTaskCancelled": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ExportTaskCompleted": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ImageExists": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "matcher": "path", - "expected": true, - "argument": "length(Images[]) > `0`", - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidAMIID.NotFound", - "state": "retry" - } - ] - }, - "ImageAvailable": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "Images[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "Images[].State", - "expected": "failed" - } - ] - }, - "InstanceRunning": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "running", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "shutting-down", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "InstanceStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].InstanceStatus.Status", - "expected": "ok" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "InstanceStopped": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "stopped", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "InstanceTerminated": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "terminated", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "KeyPairExists": { - "operation": "DescribeKeyPairs", - "delay": 5, - "maxAttempts": 6, - "acceptors": [ - { - "expected": true, - "matcher": "pathAll", - "state": "success", - "argument": "length(KeyPairs[].KeyName) > `0`" - }, - { - "expected": "InvalidKeyPair.NotFound", - "matcher": "error", - "state": "retry" - } - ] - }, - "NatGatewayAvailable": { - "operation": "DescribeNatGateways", - "delay": 15, - "maxAttempts": 40, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "NatGateways[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "failed" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "deleting" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "deleted" - }, - { - "state": "retry", - "matcher": "error", - "expected": "NatGatewayNotFound" - } - ] - }, - "NetworkInterfaceAvailable": { - "operation": "DescribeNetworkInterfaces", - "delay": 20, - "maxAttempts": 10, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "NetworkInterfaces[].Status" - }, - { - "expected": "InvalidNetworkInterfaceID.NotFound", - "matcher": "error", - "state": "failure" - } - ] - }, - "PasswordDataAvailable": { - "operation": "GetPasswordData", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "path", - "argument": "length(PasswordData) > `0`", - "expected": true - } - ] - }, - "SnapshotCompleted": { - "delay": 15, - "operation": "DescribeSnapshots", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "Snapshots[].State" - } - ] - }, - "SpotInstanceRequestFulfilled": { - "operation": "DescribeSpotInstanceRequests", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "fulfilled" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "schedule-expired" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "canceled-before-fulfillment" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "bad-parameters" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "system-error" - } - ] - }, - "SubnetAvailable": { - "delay": 15, - "operation": "DescribeSubnets", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Subnets[].State" - } - ] - }, - "SystemStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].SystemStatus.Status", - "expected": "ok" - } - ] - }, - "VolumeAvailable": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VolumeDeleted": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "matcher": "error", - "expected": "InvalidVolume.NotFound", - "state": "success" - } - ] - }, - "VolumeInUse": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "in-use", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VpcAvailable": { - "delay": 15, - "operation": "DescribeVpcs", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Vpcs[].State" - } - ] - }, - "VpcExists": { - "operation": "DescribeVpcs", - "delay": 1, - "maxAttempts": 5, - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidVpcID.NotFound", - "state": "retry" - } - ] - }, - "VpnConnectionAvailable": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpnConnectionDeleted": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpcPeeringConnectionExists": { - "delay": 15, - "operation": "DescribeVpcPeeringConnections", - "maxAttempts": 40, - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidVpcPeeringConnectionID.NotFound", - "state": "retry" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/api-2.json deleted file mode 100644 index ef0bb74d3..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/api-2.json +++ /dev/null @@ -1,14191 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"ec2-2016-04-01", - "apiVersion":"2016-04-01", - "endpointPrefix":"ec2", - "protocol":"ec2", - "serviceAbbreviation":"Amazon EC2", - "serviceFullName":"Amazon Elastic Compute Cloud", - "signatureVersion":"v4", - "xmlNamespace":"http://ec2.amazonaws.com/doc/2016-04-01" - }, - "operations":{ - "AcceptVpcPeeringConnection":{ - "name":"AcceptVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AcceptVpcPeeringConnectionRequest"}, - "output":{"shape":"AcceptVpcPeeringConnectionResult"} - }, - "AllocateAddress":{ - "name":"AllocateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateAddressRequest"}, - "output":{"shape":"AllocateAddressResult"} - }, - "AllocateHosts":{ - "name":"AllocateHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateHostsRequest"}, - "output":{"shape":"AllocateHostsResult"} - }, - "AssignPrivateIpAddresses":{ - "name":"AssignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssignPrivateIpAddressesRequest"} - }, - "AssociateAddress":{ - "name":"AssociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateAddressRequest"}, - "output":{"shape":"AssociateAddressResult"} - }, - "AssociateDhcpOptions":{ - "name":"AssociateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateDhcpOptionsRequest"} - }, - "AssociateRouteTable":{ - "name":"AssociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateRouteTableRequest"}, - "output":{"shape":"AssociateRouteTableResult"} - }, - "AttachClassicLinkVpc":{ - "name":"AttachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachClassicLinkVpcRequest"}, - "output":{"shape":"AttachClassicLinkVpcResult"} - }, - "AttachInternetGateway":{ - "name":"AttachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachInternetGatewayRequest"} - }, - "AttachNetworkInterface":{ - "name":"AttachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachNetworkInterfaceRequest"}, - "output":{"shape":"AttachNetworkInterfaceResult"} - }, - "AttachVolume":{ - "name":"AttachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVolumeRequest"}, - "output":{"shape":"VolumeAttachment"} - }, - "AttachVpnGateway":{ - "name":"AttachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVpnGatewayRequest"}, - "output":{"shape":"AttachVpnGatewayResult"} - }, - "AuthorizeSecurityGroupEgress":{ - "name":"AuthorizeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupEgressRequest"} - }, - "AuthorizeSecurityGroupIngress":{ - "name":"AuthorizeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupIngressRequest"} - }, - "BundleInstance":{ - "name":"BundleInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"BundleInstanceRequest"}, - "output":{"shape":"BundleInstanceResult"} - }, - "CancelBundleTask":{ - "name":"CancelBundleTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelBundleTaskRequest"}, - "output":{"shape":"CancelBundleTaskResult"} - }, - "CancelConversionTask":{ - "name":"CancelConversionTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelConversionRequest"} - }, - "CancelExportTask":{ - "name":"CancelExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelExportTaskRequest"} - }, - "CancelImportTask":{ - "name":"CancelImportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelImportTaskRequest"}, - "output":{"shape":"CancelImportTaskResult"} - }, - "CancelReservedInstancesListing":{ - "name":"CancelReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelReservedInstancesListingRequest"}, - "output":{"shape":"CancelReservedInstancesListingResult"} - }, - "CancelSpotFleetRequests":{ - "name":"CancelSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotFleetRequestsRequest"}, - "output":{"shape":"CancelSpotFleetRequestsResponse"} - }, - "CancelSpotInstanceRequests":{ - "name":"CancelSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotInstanceRequestsRequest"}, - "output":{"shape":"CancelSpotInstanceRequestsResult"} - }, - "ConfirmProductInstance":{ - "name":"ConfirmProductInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ConfirmProductInstanceRequest"}, - "output":{"shape":"ConfirmProductInstanceResult"} - }, - "CopyImage":{ - "name":"CopyImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyImageRequest"}, - "output":{"shape":"CopyImageResult"} - }, - "CopySnapshot":{ - "name":"CopySnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopySnapshotRequest"}, - "output":{"shape":"CopySnapshotResult"} - }, - "CreateCustomerGateway":{ - "name":"CreateCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateCustomerGatewayRequest"}, - "output":{"shape":"CreateCustomerGatewayResult"} - }, - "CreateDhcpOptions":{ - "name":"CreateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDhcpOptionsRequest"}, - "output":{"shape":"CreateDhcpOptionsResult"} - }, - "CreateFlowLogs":{ - "name":"CreateFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateFlowLogsRequest"}, - "output":{"shape":"CreateFlowLogsResult"} - }, - "CreateImage":{ - "name":"CreateImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateImageRequest"}, - "output":{"shape":"CreateImageResult"} - }, - "CreateInstanceExportTask":{ - "name":"CreateInstanceExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInstanceExportTaskRequest"}, - "output":{"shape":"CreateInstanceExportTaskResult"} - }, - "CreateInternetGateway":{ - "name":"CreateInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInternetGatewayRequest"}, - "output":{"shape":"CreateInternetGatewayResult"} - }, - "CreateKeyPair":{ - "name":"CreateKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateKeyPairRequest"}, - "output":{"shape":"KeyPair"} - }, - "CreateNatGateway":{ - "name":"CreateNatGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNatGatewayRequest"}, - "output":{"shape":"CreateNatGatewayResult"} - }, - "CreateNetworkAcl":{ - "name":"CreateNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclRequest"}, - "output":{"shape":"CreateNetworkAclResult"} - }, - "CreateNetworkAclEntry":{ - "name":"CreateNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclEntryRequest"} - }, - "CreateNetworkInterface":{ - "name":"CreateNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkInterfaceRequest"}, - "output":{"shape":"CreateNetworkInterfaceResult"} - }, - "CreatePlacementGroup":{ - "name":"CreatePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreatePlacementGroupRequest"} - }, - "CreateReservedInstancesListing":{ - "name":"CreateReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateReservedInstancesListingRequest"}, - "output":{"shape":"CreateReservedInstancesListingResult"} - }, - "CreateRoute":{ - "name":"CreateRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteRequest"}, - "output":{"shape":"CreateRouteResult"} - }, - "CreateRouteTable":{ - "name":"CreateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteTableRequest"}, - "output":{"shape":"CreateRouteTableResult"} - }, - "CreateSecurityGroup":{ - "name":"CreateSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSecurityGroupRequest"}, - "output":{"shape":"CreateSecurityGroupResult"} - }, - "CreateSnapshot":{ - "name":"CreateSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSnapshotRequest"}, - "output":{"shape":"Snapshot"} - }, - "CreateSpotDatafeedSubscription":{ - "name":"CreateSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"CreateSpotDatafeedSubscriptionResult"} - }, - "CreateSubnet":{ - "name":"CreateSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSubnetRequest"}, - "output":{"shape":"CreateSubnetResult"} - }, - "CreateTags":{ - "name":"CreateTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateTagsRequest"} - }, - "CreateVolume":{ - "name":"CreateVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVolumeRequest"}, - "output":{"shape":"Volume"} - }, - "CreateVpc":{ - "name":"CreateVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcRequest"}, - "output":{"shape":"CreateVpcResult"} - }, - "CreateVpcEndpoint":{ - "name":"CreateVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcEndpointRequest"}, - "output":{"shape":"CreateVpcEndpointResult"} - }, - "CreateVpcPeeringConnection":{ - "name":"CreateVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcPeeringConnectionRequest"}, - "output":{"shape":"CreateVpcPeeringConnectionResult"} - }, - "CreateVpnConnection":{ - "name":"CreateVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRequest"}, - "output":{"shape":"CreateVpnConnectionResult"} - }, - "CreateVpnConnectionRoute":{ - "name":"CreateVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRouteRequest"} - }, - "CreateVpnGateway":{ - "name":"CreateVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnGatewayRequest"}, - "output":{"shape":"CreateVpnGatewayResult"} - }, - "DeleteCustomerGateway":{ - "name":"DeleteCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteCustomerGatewayRequest"} - }, - "DeleteDhcpOptions":{ - "name":"DeleteDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDhcpOptionsRequest"} - }, - "DeleteFlowLogs":{ - "name":"DeleteFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteFlowLogsRequest"}, - "output":{"shape":"DeleteFlowLogsResult"} - }, - "DeleteInternetGateway":{ - "name":"DeleteInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteInternetGatewayRequest"} - }, - "DeleteKeyPair":{ - "name":"DeleteKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteKeyPairRequest"} - }, - "DeleteNatGateway":{ - "name":"DeleteNatGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNatGatewayRequest"}, - "output":{"shape":"DeleteNatGatewayResult"} - }, - "DeleteNetworkAcl":{ - "name":"DeleteNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclRequest"} - }, - "DeleteNetworkAclEntry":{ - "name":"DeleteNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclEntryRequest"} - }, - "DeleteNetworkInterface":{ - "name":"DeleteNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkInterfaceRequest"} - }, - "DeletePlacementGroup":{ - "name":"DeletePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeletePlacementGroupRequest"} - }, - "DeleteRoute":{ - "name":"DeleteRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteRequest"} - }, - "DeleteRouteTable":{ - "name":"DeleteRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteTableRequest"} - }, - "DeleteSecurityGroup":{ - "name":"DeleteSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSecurityGroupRequest"} - }, - "DeleteSnapshot":{ - "name":"DeleteSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSnapshotRequest"} - }, - "DeleteSpotDatafeedSubscription":{ - "name":"DeleteSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSpotDatafeedSubscriptionRequest"} - }, - "DeleteSubnet":{ - "name":"DeleteSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSubnetRequest"} - }, - "DeleteTags":{ - "name":"DeleteTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteTagsRequest"} - }, - "DeleteVolume":{ - "name":"DeleteVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVolumeRequest"} - }, - "DeleteVpc":{ - "name":"DeleteVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcRequest"} - }, - "DeleteVpcEndpoints":{ - "name":"DeleteVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcEndpointsRequest"}, - "output":{"shape":"DeleteVpcEndpointsResult"} - }, - "DeleteVpcPeeringConnection":{ - "name":"DeleteVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcPeeringConnectionRequest"}, - "output":{"shape":"DeleteVpcPeeringConnectionResult"} - }, - "DeleteVpnConnection":{ - "name":"DeleteVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRequest"} - }, - "DeleteVpnConnectionRoute":{ - "name":"DeleteVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRouteRequest"} - }, - "DeleteVpnGateway":{ - "name":"DeleteVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnGatewayRequest"} - }, - "DeregisterImage":{ - "name":"DeregisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeregisterImageRequest"} - }, - "DescribeAccountAttributes":{ - "name":"DescribeAccountAttributes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAccountAttributesRequest"}, - "output":{"shape":"DescribeAccountAttributesResult"} - }, - "DescribeAddresses":{ - "name":"DescribeAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAddressesRequest"}, - "output":{"shape":"DescribeAddressesResult"} - }, - "DescribeAvailabilityZones":{ - "name":"DescribeAvailabilityZones", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAvailabilityZonesRequest"}, - "output":{"shape":"DescribeAvailabilityZonesResult"} - }, - "DescribeBundleTasks":{ - "name":"DescribeBundleTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeBundleTasksRequest"}, - "output":{"shape":"DescribeBundleTasksResult"} - }, - "DescribeClassicLinkInstances":{ - "name":"DescribeClassicLinkInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeClassicLinkInstancesRequest"}, - "output":{"shape":"DescribeClassicLinkInstancesResult"} - }, - "DescribeConversionTasks":{ - "name":"DescribeConversionTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeConversionTasksRequest"}, - "output":{"shape":"DescribeConversionTasksResult"} - }, - "DescribeCustomerGateways":{ - "name":"DescribeCustomerGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeCustomerGatewaysRequest"}, - "output":{"shape":"DescribeCustomerGatewaysResult"} - }, - "DescribeDhcpOptions":{ - "name":"DescribeDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDhcpOptionsRequest"}, - "output":{"shape":"DescribeDhcpOptionsResult"} - }, - "DescribeExportTasks":{ - "name":"DescribeExportTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeExportTasksRequest"}, - "output":{"shape":"DescribeExportTasksResult"} - }, - "DescribeFlowLogs":{ - "name":"DescribeFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeFlowLogsRequest"}, - "output":{"shape":"DescribeFlowLogsResult"} - }, - "DescribeHostReservationOfferings":{ - "name":"DescribeHostReservationOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostReservationOfferingsRequest"}, - "output":{"shape":"DescribeHostReservationOfferingsResult"} - }, - "DescribeHostReservations":{ - "name":"DescribeHostReservations", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostReservationsRequest"}, - "output":{"shape":"DescribeHostReservationsResult"} - }, - "DescribeHosts":{ - "name":"DescribeHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostsRequest"}, - "output":{"shape":"DescribeHostsResult"} - }, - "DescribeIdFormat":{ - "name":"DescribeIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeIdFormatRequest"}, - "output":{"shape":"DescribeIdFormatResult"} - }, - "DescribeIdentityIdFormat":{ - "name":"DescribeIdentityIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeIdentityIdFormatRequest"}, - "output":{"shape":"DescribeIdentityIdFormatResult"} - }, - "DescribeImageAttribute":{ - "name":"DescribeImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImageAttributeRequest"}, - "output":{"shape":"ImageAttribute"} - }, - "DescribeImages":{ - "name":"DescribeImages", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImagesRequest"}, - "output":{"shape":"DescribeImagesResult"} - }, - "DescribeImportImageTasks":{ - "name":"DescribeImportImageTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportImageTasksRequest"}, - "output":{"shape":"DescribeImportImageTasksResult"} - }, - "DescribeImportSnapshotTasks":{ - "name":"DescribeImportSnapshotTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportSnapshotTasksRequest"}, - "output":{"shape":"DescribeImportSnapshotTasksResult"} - }, - "DescribeInstanceAttribute":{ - "name":"DescribeInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceAttributeRequest"}, - "output":{"shape":"InstanceAttribute"} - }, - "DescribeInstanceStatus":{ - "name":"DescribeInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceStatusRequest"}, - "output":{"shape":"DescribeInstanceStatusResult"} - }, - "DescribeInstances":{ - "name":"DescribeInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstancesRequest"}, - "output":{"shape":"DescribeInstancesResult"} - }, - "DescribeInternetGateways":{ - "name":"DescribeInternetGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInternetGatewaysRequest"}, - "output":{"shape":"DescribeInternetGatewaysResult"} - }, - "DescribeKeyPairs":{ - "name":"DescribeKeyPairs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeKeyPairsRequest"}, - "output":{"shape":"DescribeKeyPairsResult"} - }, - "DescribeMovingAddresses":{ - "name":"DescribeMovingAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeMovingAddressesRequest"}, - "output":{"shape":"DescribeMovingAddressesResult"} - }, - "DescribeNatGateways":{ - "name":"DescribeNatGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNatGatewaysRequest"}, - "output":{"shape":"DescribeNatGatewaysResult"} - }, - "DescribeNetworkAcls":{ - "name":"DescribeNetworkAcls", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkAclsRequest"}, - "output":{"shape":"DescribeNetworkAclsResult"} - }, - "DescribeNetworkInterfaceAttribute":{ - "name":"DescribeNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfaceAttributeRequest"}, - "output":{"shape":"DescribeNetworkInterfaceAttributeResult"} - }, - "DescribeNetworkInterfaces":{ - "name":"DescribeNetworkInterfaces", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfacesRequest"}, - "output":{"shape":"DescribeNetworkInterfacesResult"} - }, - "DescribePlacementGroups":{ - "name":"DescribePlacementGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePlacementGroupsRequest"}, - "output":{"shape":"DescribePlacementGroupsResult"} - }, - "DescribePrefixLists":{ - "name":"DescribePrefixLists", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePrefixListsRequest"}, - "output":{"shape":"DescribePrefixListsResult"} - }, - "DescribeRegions":{ - "name":"DescribeRegions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRegionsRequest"}, - "output":{"shape":"DescribeRegionsResult"} - }, - "DescribeReservedInstances":{ - "name":"DescribeReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesRequest"}, - "output":{"shape":"DescribeReservedInstancesResult"} - }, - "DescribeReservedInstancesListings":{ - "name":"DescribeReservedInstancesListings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesListingsRequest"}, - "output":{"shape":"DescribeReservedInstancesListingsResult"} - }, - "DescribeReservedInstancesModifications":{ - "name":"DescribeReservedInstancesModifications", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesModificationsRequest"}, - "output":{"shape":"DescribeReservedInstancesModificationsResult"} - }, - "DescribeReservedInstancesOfferings":{ - "name":"DescribeReservedInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesOfferingsRequest"}, - "output":{"shape":"DescribeReservedInstancesOfferingsResult"} - }, - "DescribeRouteTables":{ - "name":"DescribeRouteTables", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRouteTablesRequest"}, - "output":{"shape":"DescribeRouteTablesResult"} - }, - "DescribeScheduledInstanceAvailability":{ - "name":"DescribeScheduledInstanceAvailability", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeScheduledInstanceAvailabilityRequest"}, - "output":{"shape":"DescribeScheduledInstanceAvailabilityResult"} - }, - "DescribeScheduledInstances":{ - "name":"DescribeScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeScheduledInstancesRequest"}, - "output":{"shape":"DescribeScheduledInstancesResult"} - }, - "DescribeSecurityGroupReferences":{ - "name":"DescribeSecurityGroupReferences", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupReferencesRequest"}, - "output":{"shape":"DescribeSecurityGroupReferencesResult"} - }, - "DescribeSecurityGroups":{ - "name":"DescribeSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupsRequest"}, - "output":{"shape":"DescribeSecurityGroupsResult"} - }, - "DescribeSnapshotAttribute":{ - "name":"DescribeSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotAttributeRequest"}, - "output":{"shape":"DescribeSnapshotAttributeResult"} - }, - "DescribeSnapshots":{ - "name":"DescribeSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotsRequest"}, - "output":{"shape":"DescribeSnapshotsResult"} - }, - "DescribeSpotDatafeedSubscription":{ - "name":"DescribeSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"DescribeSpotDatafeedSubscriptionResult"} - }, - "DescribeSpotFleetInstances":{ - "name":"DescribeSpotFleetInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetInstancesRequest"}, - "output":{"shape":"DescribeSpotFleetInstancesResponse"} - }, - "DescribeSpotFleetRequestHistory":{ - "name":"DescribeSpotFleetRequestHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestHistoryRequest"}, - "output":{"shape":"DescribeSpotFleetRequestHistoryResponse"} - }, - "DescribeSpotFleetRequests":{ - "name":"DescribeSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestsRequest"}, - "output":{"shape":"DescribeSpotFleetRequestsResponse"} - }, - "DescribeSpotInstanceRequests":{ - "name":"DescribeSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotInstanceRequestsRequest"}, - "output":{"shape":"DescribeSpotInstanceRequestsResult"} - }, - "DescribeSpotPriceHistory":{ - "name":"DescribeSpotPriceHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotPriceHistoryRequest"}, - "output":{"shape":"DescribeSpotPriceHistoryResult"} - }, - "DescribeStaleSecurityGroups":{ - "name":"DescribeStaleSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeStaleSecurityGroupsRequest"}, - "output":{"shape":"DescribeStaleSecurityGroupsResult"} - }, - "DescribeSubnets":{ - "name":"DescribeSubnets", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSubnetsRequest"}, - "output":{"shape":"DescribeSubnetsResult"} - }, - "DescribeTags":{ - "name":"DescribeTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeTagsRequest"}, - "output":{"shape":"DescribeTagsResult"} - }, - "DescribeVolumeAttribute":{ - "name":"DescribeVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeAttributeRequest"}, - "output":{"shape":"DescribeVolumeAttributeResult"} - }, - "DescribeVolumeStatus":{ - "name":"DescribeVolumeStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeStatusRequest"}, - "output":{"shape":"DescribeVolumeStatusResult"} - }, - "DescribeVolumes":{ - "name":"DescribeVolumes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumesRequest"}, - "output":{"shape":"DescribeVolumesResult"} - }, - "DescribeVpcAttribute":{ - "name":"DescribeVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcAttributeRequest"}, - "output":{"shape":"DescribeVpcAttributeResult"} - }, - "DescribeVpcClassicLink":{ - "name":"DescribeVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkRequest"}, - "output":{"shape":"DescribeVpcClassicLinkResult"} - }, - "DescribeVpcClassicLinkDnsSupport":{ - "name":"DescribeVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"DescribeVpcClassicLinkDnsSupportResult"} - }, - "DescribeVpcEndpointServices":{ - "name":"DescribeVpcEndpointServices", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointServicesRequest"}, - "output":{"shape":"DescribeVpcEndpointServicesResult"} - }, - "DescribeVpcEndpoints":{ - "name":"DescribeVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointsRequest"}, - "output":{"shape":"DescribeVpcEndpointsResult"} - }, - "DescribeVpcPeeringConnections":{ - "name":"DescribeVpcPeeringConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcPeeringConnectionsRequest"}, - "output":{"shape":"DescribeVpcPeeringConnectionsResult"} - }, - "DescribeVpcs":{ - "name":"DescribeVpcs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcsRequest"}, - "output":{"shape":"DescribeVpcsResult"} - }, - "DescribeVpnConnections":{ - "name":"DescribeVpnConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnConnectionsRequest"}, - "output":{"shape":"DescribeVpnConnectionsResult"} - }, - "DescribeVpnGateways":{ - "name":"DescribeVpnGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnGatewaysRequest"}, - "output":{"shape":"DescribeVpnGatewaysResult"} - }, - "DetachClassicLinkVpc":{ - "name":"DetachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachClassicLinkVpcRequest"}, - "output":{"shape":"DetachClassicLinkVpcResult"} - }, - "DetachInternetGateway":{ - "name":"DetachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachInternetGatewayRequest"} - }, - "DetachNetworkInterface":{ - "name":"DetachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachNetworkInterfaceRequest"} - }, - "DetachVolume":{ - "name":"DetachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVolumeRequest"}, - "output":{"shape":"VolumeAttachment"} - }, - "DetachVpnGateway":{ - "name":"DetachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVpnGatewayRequest"} - }, - "DisableVgwRoutePropagation":{ - "name":"DisableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVgwRoutePropagationRequest"} - }, - "DisableVpcClassicLink":{ - "name":"DisableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkRequest"}, - "output":{"shape":"DisableVpcClassicLinkResult"} - }, - "DisableVpcClassicLinkDnsSupport":{ - "name":"DisableVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"DisableVpcClassicLinkDnsSupportResult"} - }, - "DisassociateAddress":{ - "name":"DisassociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateAddressRequest"} - }, - "DisassociateRouteTable":{ - "name":"DisassociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateRouteTableRequest"} - }, - "EnableVgwRoutePropagation":{ - "name":"EnableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVgwRoutePropagationRequest"} - }, - "EnableVolumeIO":{ - "name":"EnableVolumeIO", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVolumeIORequest"} - }, - "EnableVpcClassicLink":{ - "name":"EnableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkRequest"}, - "output":{"shape":"EnableVpcClassicLinkResult"} - }, - "EnableVpcClassicLinkDnsSupport":{ - "name":"EnableVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"EnableVpcClassicLinkDnsSupportResult"} - }, - "GetConsoleOutput":{ - "name":"GetConsoleOutput", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleOutputRequest"}, - "output":{"shape":"GetConsoleOutputResult"} - }, - "GetConsoleScreenshot":{ - "name":"GetConsoleScreenshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleScreenshotRequest"}, - "output":{"shape":"GetConsoleScreenshotResult"} - }, - "GetHostReservationPurchasePreview":{ - "name":"GetHostReservationPurchasePreview", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetHostReservationPurchasePreviewRequest"}, - "output":{"shape":"GetHostReservationPurchasePreviewResult"} - }, - "GetPasswordData":{ - "name":"GetPasswordData", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetPasswordDataRequest"}, - "output":{"shape":"GetPasswordDataResult"} - }, - "ImportImage":{ - "name":"ImportImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportImageRequest"}, - "output":{"shape":"ImportImageResult"} - }, - "ImportInstance":{ - "name":"ImportInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportInstanceRequest"}, - "output":{"shape":"ImportInstanceResult"} - }, - "ImportKeyPair":{ - "name":"ImportKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportKeyPairRequest"}, - "output":{"shape":"ImportKeyPairResult"} - }, - "ImportSnapshot":{ - "name":"ImportSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportSnapshotRequest"}, - "output":{"shape":"ImportSnapshotResult"} - }, - "ImportVolume":{ - "name":"ImportVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportVolumeRequest"}, - "output":{"shape":"ImportVolumeResult"} - }, - "ModifyHosts":{ - "name":"ModifyHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyHostsRequest"}, - "output":{"shape":"ModifyHostsResult"} - }, - "ModifyIdFormat":{ - "name":"ModifyIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyIdFormatRequest"} - }, - "ModifyIdentityIdFormat":{ - "name":"ModifyIdentityIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyIdentityIdFormatRequest"} - }, - "ModifyImageAttribute":{ - "name":"ModifyImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyImageAttributeRequest"} - }, - "ModifyInstanceAttribute":{ - "name":"ModifyInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstanceAttributeRequest"} - }, - "ModifyInstancePlacement":{ - "name":"ModifyInstancePlacement", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstancePlacementRequest"}, - "output":{"shape":"ModifyInstancePlacementResult"} - }, - "ModifyNetworkInterfaceAttribute":{ - "name":"ModifyNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyNetworkInterfaceAttributeRequest"} - }, - "ModifyReservedInstances":{ - "name":"ModifyReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyReservedInstancesRequest"}, - "output":{"shape":"ModifyReservedInstancesResult"} - }, - "ModifySnapshotAttribute":{ - "name":"ModifySnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySnapshotAttributeRequest"} - }, - "ModifySpotFleetRequest":{ - "name":"ModifySpotFleetRequest", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySpotFleetRequestRequest"}, - "output":{"shape":"ModifySpotFleetRequestResponse"} - }, - "ModifySubnetAttribute":{ - "name":"ModifySubnetAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySubnetAttributeRequest"} - }, - "ModifyVolumeAttribute":{ - "name":"ModifyVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVolumeAttributeRequest"} - }, - "ModifyVpcAttribute":{ - "name":"ModifyVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcAttributeRequest"} - }, - "ModifyVpcEndpoint":{ - "name":"ModifyVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcEndpointRequest"}, - "output":{"shape":"ModifyVpcEndpointResult"} - }, - "ModifyVpcPeeringConnectionOptions":{ - "name":"ModifyVpcPeeringConnectionOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcPeeringConnectionOptionsRequest"}, - "output":{"shape":"ModifyVpcPeeringConnectionOptionsResult"} - }, - "MonitorInstances":{ - "name":"MonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MonitorInstancesRequest"}, - "output":{"shape":"MonitorInstancesResult"} - }, - "MoveAddressToVpc":{ - "name":"MoveAddressToVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MoveAddressToVpcRequest"}, - "output":{"shape":"MoveAddressToVpcResult"} - }, - "PurchaseHostReservation":{ - "name":"PurchaseHostReservation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseHostReservationRequest"}, - "output":{"shape":"PurchaseHostReservationResult"} - }, - "PurchaseReservedInstancesOffering":{ - "name":"PurchaseReservedInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedInstancesOfferingRequest"}, - "output":{"shape":"PurchaseReservedInstancesOfferingResult"} - }, - "PurchaseScheduledInstances":{ - "name":"PurchaseScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseScheduledInstancesRequest"}, - "output":{"shape":"PurchaseScheduledInstancesResult"} - }, - "RebootInstances":{ - "name":"RebootInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootInstancesRequest"} - }, - "RegisterImage":{ - "name":"RegisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RegisterImageRequest"}, - "output":{"shape":"RegisterImageResult"} - }, - "RejectVpcPeeringConnection":{ - "name":"RejectVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RejectVpcPeeringConnectionRequest"}, - "output":{"shape":"RejectVpcPeeringConnectionResult"} - }, - "ReleaseAddress":{ - "name":"ReleaseAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseAddressRequest"} - }, - "ReleaseHosts":{ - "name":"ReleaseHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseHostsRequest"}, - "output":{"shape":"ReleaseHostsResult"} - }, - "ReplaceNetworkAclAssociation":{ - "name":"ReplaceNetworkAclAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclAssociationRequest"}, - "output":{"shape":"ReplaceNetworkAclAssociationResult"} - }, - "ReplaceNetworkAclEntry":{ - "name":"ReplaceNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclEntryRequest"} - }, - "ReplaceRoute":{ - "name":"ReplaceRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteRequest"} - }, - "ReplaceRouteTableAssociation":{ - "name":"ReplaceRouteTableAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteTableAssociationRequest"}, - "output":{"shape":"ReplaceRouteTableAssociationResult"} - }, - "ReportInstanceStatus":{ - "name":"ReportInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReportInstanceStatusRequest"} - }, - "RequestSpotFleet":{ - "name":"RequestSpotFleet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotFleetRequest"}, - "output":{"shape":"RequestSpotFleetResponse"} - }, - "RequestSpotInstances":{ - "name":"RequestSpotInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotInstancesRequest"}, - "output":{"shape":"RequestSpotInstancesResult"} - }, - "ResetImageAttribute":{ - "name":"ResetImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetImageAttributeRequest"} - }, - "ResetInstanceAttribute":{ - "name":"ResetInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetInstanceAttributeRequest"} - }, - "ResetNetworkInterfaceAttribute":{ - "name":"ResetNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetNetworkInterfaceAttributeRequest"} - }, - "ResetSnapshotAttribute":{ - "name":"ResetSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetSnapshotAttributeRequest"} - }, - "RestoreAddressToClassic":{ - "name":"RestoreAddressToClassic", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreAddressToClassicRequest"}, - "output":{"shape":"RestoreAddressToClassicResult"} - }, - "RevokeSecurityGroupEgress":{ - "name":"RevokeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupEgressRequest"} - }, - "RevokeSecurityGroupIngress":{ - "name":"RevokeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupIngressRequest"} - }, - "RunInstances":{ - "name":"RunInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunInstancesRequest"}, - "output":{"shape":"Reservation"} - }, - "RunScheduledInstances":{ - "name":"RunScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunScheduledInstancesRequest"}, - "output":{"shape":"RunScheduledInstancesResult"} - }, - "StartInstances":{ - "name":"StartInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StartInstancesRequest"}, - "output":{"shape":"StartInstancesResult"} - }, - "StopInstances":{ - "name":"StopInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StopInstancesRequest"}, - "output":{"shape":"StopInstancesResult"} - }, - "TerminateInstances":{ - "name":"TerminateInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"TerminateInstancesRequest"}, - "output":{"shape":"TerminateInstancesResult"} - }, - "UnassignPrivateIpAddresses":{ - "name":"UnassignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnassignPrivateIpAddressesRequest"} - }, - "UnmonitorInstances":{ - "name":"UnmonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnmonitorInstancesRequest"}, - "output":{"shape":"UnmonitorInstancesResult"} - } - }, - "shapes":{ - "AcceptVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "AcceptVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "AccountAttribute":{ - "type":"structure", - "members":{ - "AttributeName":{ - "shape":"String", - "locationName":"attributeName" - }, - "AttributeValues":{ - "shape":"AccountAttributeValueList", - "locationName":"attributeValueSet" - } - } - }, - "AccountAttributeList":{ - "type":"list", - "member":{ - "shape":"AccountAttribute", - "locationName":"item" - } - }, - "AccountAttributeName":{ - "type":"string", - "enum":[ - "supported-platforms", - "default-vpc" - ] - }, - "AccountAttributeNameStringList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeName", - "locationName":"attributeName" - } - }, - "AccountAttributeValue":{ - "type":"structure", - "members":{ - "AttributeValue":{ - "shape":"String", - "locationName":"attributeValue" - } - } - }, - "AccountAttributeValueList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeValue", - "locationName":"item" - } - }, - "ActiveInstance":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - } - } - }, - "ActiveInstanceSet":{ - "type":"list", - "member":{ - "shape":"ActiveInstance", - "locationName":"item" - } - }, - "ActivityStatus":{ - "type":"string", - "enum":[ - "error", - "pending_fulfillment", - "pending_termination", - "fulfilled" - ] - }, - "Address":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "NetworkInterfaceOwnerId":{ - "shape":"String", - "locationName":"networkInterfaceOwnerId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "AddressList":{ - "type":"list", - "member":{ - "shape":"Address", - "locationName":"item" - } - }, - "Affinity":{ - "type":"string", - "enum":[ - "default", - "host" - ] - }, - "AllocateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Domain":{"shape":"DomainType"} - } - }, - "AllocateAddressResult":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - } - } - }, - "AllocateHostsRequest":{ - "type":"structure", - "required":[ - "InstanceType", - "Quantity", - "AvailabilityZone" - ], - "members":{ - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Quantity":{ - "shape":"Integer", - "locationName":"quantity" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "AllocateHostsResult":{ - "type":"structure", - "members":{ - "HostIds":{ - "shape":"ResponseHostIdList", - "locationName":"hostIdSet" - } - } - }, - "AllocationIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"AllocationId" - } - }, - "AllocationState":{ - "type":"string", - "enum":[ - "available", - "under-assessment", - "permanent-failure", - "released", - "released-permanent-failure" - ] - }, - "AllocationStrategy":{ - "type":"string", - "enum":[ - "lowestPrice", - "diversified" - ] - }, - "ArchitectureValues":{ - "type":"string", - "enum":[ - "i386", - "x86_64" - ] - }, - "AssignPrivateIpAddressesRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AllowReassignment":{ - "shape":"Boolean", - "locationName":"allowReassignment" - } - } - }, - "AssociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "AllowReassociation":{ - "shape":"Boolean", - "locationName":"allowReassociation" - } - } - }, - "AssociateAddressResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AssociateDhcpOptionsRequest":{ - "type":"structure", - "required":[ - "DhcpOptionsId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AssociateRouteTableRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "AssociateRouteTableResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AttachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId", - "Groups" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"SecurityGroupId" - } - } - }, - "AttachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "AttachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "AttachNetworkInterfaceRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "InstanceId", - "DeviceIndex" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - } - } - }, - "AttachNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - } - } - }, - "AttachVolumeRequest":{ - "type":"structure", - "required":[ - "VolumeId", - "InstanceId", - "Device" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"} - } - }, - "AttachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AttachVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpcAttachment":{ - "shape":"VpcAttachment", - "locationName":"attachment" - } - } - }, - "AttachmentStatus":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "AttributeBooleanValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Boolean", - "locationName":"value" - } - } - }, - "AttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "AuthorizeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "AuthorizeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "AutoPlacement":{ - "type":"string", - "enum":[ - "on", - "off" - ] - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "ZoneName":{ - "shape":"String", - "locationName":"zoneName" - }, - "State":{ - "shape":"AvailabilityZoneState", - "locationName":"zoneState" - }, - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Messages":{ - "shape":"AvailabilityZoneMessageList", - "locationName":"messageSet" - } - } - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"item" - } - }, - "AvailabilityZoneMessage":{ - "type":"structure", - "members":{ - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "AvailabilityZoneMessageList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZoneMessage", - "locationName":"item" - } - }, - "AvailabilityZoneState":{ - "type":"string", - "enum":[ - "available", - "information", - "impaired", - "unavailable" - ] - }, - "AvailableCapacity":{ - "type":"structure", - "members":{ - "AvailableInstanceCapacity":{ - "shape":"AvailableInstanceCapacityList", - "locationName":"availableInstanceCapacity" - }, - "AvailableVCpus":{ - "shape":"Integer", - "locationName":"availableVCpus" - } - } - }, - "AvailableInstanceCapacityList":{ - "type":"list", - "member":{ - "shape":"InstanceCapacity", - "locationName":"item" - } - }, - "BatchState":{ - "type":"string", - "enum":[ - "submitted", - "active", - "cancelled", - "failed", - "cancelled_running", - "cancelled_terminating", - "modifying" - ] - }, - "Blob":{"type":"blob"}, - "BlobAttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Blob", - "locationName":"value" - } - } - }, - "BlockDeviceMapping":{ - "type":"structure", - "members":{ - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsBlockDevice", - "locationName":"ebs" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "BlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"item" - } - }, - "BlockDeviceMappingRequestList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "Boolean":{"type":"boolean"}, - "BundleIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"BundleId" - } - }, - "BundleInstanceRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Storage" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "Storage":{"shape":"Storage"} - } - }, - "BundleInstanceResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "BundleTask":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "BundleId":{ - "shape":"String", - "locationName":"bundleId" - }, - "State":{ - "shape":"BundleTaskState", - "locationName":"state" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Storage":{ - "shape":"Storage", - "locationName":"storage" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "BundleTaskError":{ - "shape":"BundleTaskError", - "locationName":"error" - } - } - }, - "BundleTaskError":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "BundleTaskList":{ - "type":"list", - "member":{ - "shape":"BundleTask", - "locationName":"item" - } - }, - "BundleTaskState":{ - "type":"string", - "enum":[ - "pending", - "waiting-for-shutdown", - "bundling", - "storing", - "cancelling", - "complete", - "failed" - ] - }, - "CancelBatchErrorCode":{ - "type":"string", - "enum":[ - "fleetRequestIdDoesNotExist", - "fleetRequestIdMalformed", - "fleetRequestNotInCancellableState", - "unexpectedError" - ] - }, - "CancelBundleTaskRequest":{ - "type":"structure", - "required":["BundleId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleId":{"shape":"String"} - } - }, - "CancelBundleTaskResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "CancelConversionRequest":{ - "type":"structure", - "required":["ConversionTaskId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ReasonMessage":{ - "shape":"String", - "locationName":"reasonMessage" - } - } - }, - "CancelExportTaskRequest":{ - "type":"structure", - "required":["ExportTaskId"], - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - } - } - }, - "CancelImportTaskRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskId":{"shape":"String"}, - "CancelReason":{"shape":"String"} - } - }, - "CancelImportTaskResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "PreviousState":{ - "shape":"String", - "locationName":"previousState" - } - } - }, - "CancelReservedInstancesListingRequest":{ - "type":"structure", - "required":["ReservedInstancesListingId"], - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - } - } - }, - "CancelReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CancelSpotFleetRequestsError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"CancelBatchErrorCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "CancelSpotFleetRequestsErrorItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "Error" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "Error":{ - "shape":"CancelSpotFleetRequestsError", - "locationName":"error" - } - } - }, - "CancelSpotFleetRequestsErrorSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsErrorItem", - "locationName":"item" - } - }, - "CancelSpotFleetRequestsRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestIds", - "TerminateInstances" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "TerminateInstances":{ - "shape":"Boolean", - "locationName":"terminateInstances" - } - } - }, - "CancelSpotFleetRequestsResponse":{ - "type":"structure", - "members":{ - "UnsuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsErrorSet", - "locationName":"unsuccessfulFleetRequestSet" - }, - "SuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsSuccessSet", - "locationName":"successfulFleetRequestSet" - } - } - }, - "CancelSpotFleetRequestsSuccessItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "CurrentSpotFleetRequestState", - "PreviousSpotFleetRequestState" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "CurrentSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"currentSpotFleetRequestState" - }, - "PreviousSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"previousSpotFleetRequestState" - } - } - }, - "CancelSpotFleetRequestsSuccessSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsSuccessItem", - "locationName":"item" - } - }, - "CancelSpotInstanceRequestState":{ - "type":"string", - "enum":[ - "active", - "open", - "closed", - "cancelled", - "completed" - ] - }, - "CancelSpotInstanceRequestsRequest":{ - "type":"structure", - "required":["SpotInstanceRequestIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - } - } - }, - "CancelSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "CancelledSpotInstanceRequests":{ - "shape":"CancelledSpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "CancelledSpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "State":{ - "shape":"CancelSpotInstanceRequestState", - "locationName":"state" - } - } - }, - "CancelledSpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"CancelledSpotInstanceRequest", - "locationName":"item" - } - }, - "ClassicLinkDnsSupport":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkDnsSupported":{ - "shape":"Boolean", - "locationName":"classicLinkDnsSupported" - } - } - }, - "ClassicLinkDnsSupportList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkDnsSupport", - "locationName":"item" - } - }, - "ClassicLinkInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ClassicLinkInstanceList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkInstance", - "locationName":"item" - } - }, - "ClientData":{ - "type":"structure", - "members":{ - "UploadStart":{"shape":"DateTime"}, - "UploadEnd":{"shape":"DateTime"}, - "UploadSize":{"shape":"Double"}, - "Comment":{"shape":"String"} - } - }, - "ConfirmProductInstanceRequest":{ - "type":"structure", - "required":[ - "ProductCode", - "InstanceId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ProductCode":{"shape":"String"}, - "InstanceId":{"shape":"String"} - } - }, - "ConfirmProductInstanceResult":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ContainerFormat":{ - "type":"string", - "enum":["ova"] - }, - "ConversionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "ConversionTask":{ - "type":"structure", - "required":[ - "ConversionTaskId", - "State" - ], - "members":{ - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ExpirationTime":{ - "shape":"String", - "locationName":"expirationTime" - }, - "ImportInstance":{ - "shape":"ImportInstanceTaskDetails", - "locationName":"importInstance" - }, - "ImportVolume":{ - "shape":"ImportVolumeTaskDetails", - "locationName":"importVolume" - }, - "State":{ - "shape":"ConversionTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ConversionTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "CopyImageRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceImageId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceImageId":{"shape":"String"}, - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "ClientToken":{"shape":"String"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopyImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CopySnapshotRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceSnapshotId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceSnapshotId":{"shape":"String"}, - "Description":{"shape":"String"}, - "DestinationRegion":{ - "shape":"String", - "locationName":"destinationRegion" - }, - "PresignedUrl":{ - "shape":"String", - "locationName":"presignedUrl" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopySnapshotResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - } - } - }, - "CreateCustomerGatewayRequest":{ - "type":"structure", - "required":[ - "Type", - "PublicIp", - "BgpAsn" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "PublicIp":{ - "shape":"String", - "locationName":"IpAddress" - }, - "BgpAsn":{"shape":"Integer"} - } - }, - "CreateCustomerGatewayResult":{ - "type":"structure", - "members":{ - "CustomerGateway":{ - "shape":"CustomerGateway", - "locationName":"customerGateway" - } - } - }, - "CreateDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpConfigurations"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpConfigurations":{ - "shape":"NewDhcpConfigurationList", - "locationName":"dhcpConfiguration" - } - } - }, - "CreateDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptions", - "locationName":"dhcpOptions" - } - } - }, - "CreateFlowLogsRequest":{ - "type":"structure", - "required":[ - "ResourceIds", - "ResourceType", - "TrafficType", - "LogGroupName", - "DeliverLogsPermissionArn" - ], - "members":{ - "ResourceIds":{ - "shape":"ValueStringList", - "locationName":"ResourceId" - }, - "ResourceType":{"shape":"FlowLogsResourceType"}, - "TrafficType":{"shape":"TrafficType"}, - "LogGroupName":{"shape":"String"}, - "DeliverLogsPermissionArn":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"flowLogIdSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "CreateImageRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NoReboot":{ - "shape":"Boolean", - "locationName":"noReboot" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"blockDeviceMapping" - } - } - }, - "CreateImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CreateInstanceExportTaskRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "Description":{ - "shape":"String", - "locationName":"description" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - }, - "ExportToS3Task":{ - "shape":"ExportToS3TaskSpecification", - "locationName":"exportToS3" - } - } - }, - "CreateInstanceExportTaskResult":{ - "type":"structure", - "members":{ - "ExportTask":{ - "shape":"ExportTask", - "locationName":"exportTask" - } - } - }, - "CreateInternetGatewayRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateInternetGatewayResult":{ - "type":"structure", - "members":{ - "InternetGateway":{ - "shape":"InternetGateway", - "locationName":"internetGateway" - } - } - }, - "CreateKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "CreateNatGatewayRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "AllocationId" - ], - "members":{ - "SubnetId":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateNatGatewayResult":{ - "type":"structure", - "members":{ - "NatGateway":{ - "shape":"NatGateway", - "locationName":"natGateway" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "CreateNetworkAclRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateNetworkAclResult":{ - "type":"structure", - "members":{ - "NetworkAcl":{ - "shape":"NetworkAcl", - "locationName":"networkAcl" - } - } - }, - "CreateNetworkInterfaceRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "NetworkInterface":{ - "shape":"NetworkInterface", - "locationName":"networkInterface" - } - } - }, - "CreatePlacementGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Strategy" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - } - } - }, - "CreateReservedInstancesListingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesId", - "InstanceCount", - "PriceSchedules", - "ClientToken" - ], - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "PriceSchedules":{ - "shape":"PriceScheduleSpecificationList", - "locationName":"priceSchedules" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CreateRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "CreateRouteResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "CreateRouteTableRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateRouteTableResult":{ - "type":"structure", - "members":{ - "RouteTable":{ - "shape":"RouteTable", - "locationName":"routeTable" - } - } - }, - "CreateSecurityGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Description" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "Description":{ - "shape":"String", - "locationName":"GroupDescription" - }, - "VpcId":{"shape":"String"} - } - }, - "CreateSecurityGroupResult":{ - "type":"structure", - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "CreateSnapshotRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Description":{"shape":"String"} - } - }, - "CreateSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "required":["Bucket"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - } - } - }, - "CreateSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "CreateSubnetRequest":{ - "type":"structure", - "required":[ - "VpcId", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "CidrBlock":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateSubnetResult":{ - "type":"structure", - "members":{ - "Subnet":{ - "shape":"Subnet", - "locationName":"subnet" - } - } - }, - "CreateTagsRequest":{ - "type":"structure", - "required":[ - "Resources", - "Tags" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"ResourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"Tag" - } - } - }, - "CreateVolumePermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "CreateVolumePermissionList":{ - "type":"list", - "member":{ - "shape":"CreateVolumePermission", - "locationName":"item" - } - }, - "CreateVolumePermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"CreateVolumePermissionList"}, - "Remove":{"shape":"CreateVolumePermissionList"} - } - }, - "CreateVolumeRequest":{ - "type":"structure", - "required":["AvailabilityZone"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Size":{"shape":"Integer"}, - "SnapshotId":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "VolumeType":{"shape":"VolumeType"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{"shape":"String"} - } - }, - "CreateVpcEndpointRequest":{ - "type":"structure", - "required":[ - "VpcId", - "ServiceName" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "ServiceName":{"shape":"String"}, - "PolicyDocument":{"shape":"String"}, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "ClientToken":{"shape":"String"} - } - }, - "CreateVpcEndpointResult":{ - "type":"structure", - "members":{ - "VpcEndpoint":{ - "shape":"VpcEndpoint", - "locationName":"vpcEndpoint" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeerVpcId":{ - "shape":"String", - "locationName":"peerVpcId" - }, - "PeerOwnerId":{ - "shape":"String", - "locationName":"peerOwnerId" - } - } - }, - "CreateVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "CreateVpcRequest":{ - "type":"structure", - "required":["CidrBlock"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CidrBlock":{"shape":"String"}, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - } - } - }, - "CreateVpcResult":{ - "type":"structure", - "members":{ - "Vpc":{ - "shape":"Vpc", - "locationName":"vpc" - } - } - }, - "CreateVpnConnectionRequest":{ - "type":"structure", - "required":[ - "Type", - "CustomerGatewayId", - "VpnGatewayId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"String"}, - "CustomerGatewayId":{"shape":"String"}, - "VpnGatewayId":{"shape":"String"}, - "Options":{ - "shape":"VpnConnectionOptionsSpecification", - "locationName":"options" - } - } - }, - "CreateVpnConnectionResult":{ - "type":"structure", - "members":{ - "VpnConnection":{ - "shape":"VpnConnection", - "locationName":"vpnConnection" - } - } - }, - "CreateVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "CreateVpnGatewayRequest":{ - "type":"structure", - "required":["Type"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpnGateway":{ - "shape":"VpnGateway", - "locationName":"vpnGateway" - } - } - }, - "CurrencyCodeValues":{ - "type":"string", - "enum":["USD"] - }, - "CustomerGateway":{ - "type":"structure", - "members":{ - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "Type":{ - "shape":"String", - "locationName":"type" - }, - "IpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "BgpAsn":{ - "shape":"String", - "locationName":"bgpAsn" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "CustomerGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"CustomerGatewayId" - } - }, - "CustomerGatewayList":{ - "type":"list", - "member":{ - "shape":"CustomerGateway", - "locationName":"item" - } - }, - "DatafeedSubscriptionState":{ - "type":"string", - "enum":[ - "Active", - "Inactive" - ] - }, - "DateTime":{"type":"timestamp"}, - "DeleteCustomerGatewayRequest":{ - "type":"structure", - "required":["CustomerGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayId":{"shape":"String"} - } - }, - "DeleteDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpOptionsId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"} - } - }, - "DeleteFlowLogsRequest":{ - "type":"structure", - "required":["FlowLogIds"], - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - } - } - }, - "DeleteFlowLogsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteInternetGatewayRequest":{ - "type":"structure", - "required":["InternetGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - } - } - }, - "DeleteKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "DeleteNatGatewayRequest":{ - "type":"structure", - "required":["NatGatewayId"], - "members":{ - "NatGatewayId":{"shape":"String"} - } - }, - "DeleteNatGatewayResult":{ - "type":"structure", - "members":{ - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "DeleteNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Egress" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - } - } - }, - "DeleteNetworkAclRequest":{ - "type":"structure", - "required":["NetworkAclId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "DeleteNetworkInterfaceRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "DeletePlacementGroupRequest":{ - "type":"structure", - "required":["GroupName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "DeleteRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - } - } - }, - "DeleteRouteTableRequest":{ - "type":"structure", - "required":["RouteTableId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "DeleteSecurityGroupRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"} - } - }, - "DeleteSnapshotRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"} - } - }, - "DeleteSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DeleteSubnetRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{"shape":"String"} - } - }, - "DeleteTagsRequest":{ - "type":"structure", - "required":["Resources"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"resourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tag" - } - } - }, - "DeleteVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"} - } - }, - "DeleteVpcEndpointsRequest":{ - "type":"structure", - "required":["VpcEndpointIds"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - } - } - }, - "DeleteVpcEndpointsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "DeleteVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DeleteVpcRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRequest":{ - "type":"structure", - "required":["VpnConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "DeleteVpnGatewayRequest":{ - "type":"structure", - "required":["VpnGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"} - } - }, - "DeregisterImageRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"} - } - }, - "DescribeAccountAttributesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttributeNames":{ - "shape":"AccountAttributeNameStringList", - "locationName":"attributeName" - } - } - }, - "DescribeAccountAttributesResult":{ - "type":"structure", - "members":{ - "AccountAttributes":{ - "shape":"AccountAttributeList", - "locationName":"accountAttributeSet" - } - } - }, - "DescribeAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"PublicIpStringList", - "locationName":"PublicIp" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AllocationIds":{ - "shape":"AllocationIdList", - "locationName":"AllocationId" - } - } - }, - "DescribeAddressesResult":{ - "type":"structure", - "members":{ - "Addresses":{ - "shape":"AddressList", - "locationName":"addressesSet" - } - } - }, - "DescribeAvailabilityZonesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ZoneNames":{ - "shape":"ZoneNameStringList", - "locationName":"ZoneName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeAvailabilityZonesResult":{ - "type":"structure", - "members":{ - "AvailabilityZones":{ - "shape":"AvailabilityZoneList", - "locationName":"availabilityZoneInfo" - } - } - }, - "DescribeBundleTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleIds":{ - "shape":"BundleIdStringList", - "locationName":"BundleId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeBundleTasksResult":{ - "type":"structure", - "members":{ - "BundleTasks":{ - "shape":"BundleTaskList", - "locationName":"bundleInstanceTasksSet" - } - } - }, - "DescribeClassicLinkInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeClassicLinkInstancesResult":{ - "type":"structure", - "members":{ - "Instances":{ - "shape":"ClassicLinkInstanceList", - "locationName":"instancesSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeConversionTaskList":{ - "type":"list", - "member":{ - "shape":"ConversionTask", - "locationName":"item" - } - }, - "DescribeConversionTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "ConversionTaskIds":{ - "shape":"ConversionIdStringList", - "locationName":"conversionTaskId" - } - } - }, - "DescribeConversionTasksResult":{ - "type":"structure", - "members":{ - "ConversionTasks":{ - "shape":"DescribeConversionTaskList", - "locationName":"conversionTasks" - } - } - }, - "DescribeCustomerGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayIds":{ - "shape":"CustomerGatewayIdStringList", - "locationName":"CustomerGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeCustomerGatewaysResult":{ - "type":"structure", - "members":{ - "CustomerGateways":{ - "shape":"CustomerGatewayList", - "locationName":"customerGatewaySet" - } - } - }, - "DescribeDhcpOptionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsIds":{ - "shape":"DhcpOptionsIdStringList", - "locationName":"DhcpOptionsId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptionsList", - "locationName":"dhcpOptionsSet" - } - } - }, - "DescribeExportTasksRequest":{ - "type":"structure", - "members":{ - "ExportTaskIds":{ - "shape":"ExportTaskIdStringList", - "locationName":"exportTaskId" - } - } - }, - "DescribeExportTasksResult":{ - "type":"structure", - "members":{ - "ExportTasks":{ - "shape":"ExportTaskList", - "locationName":"exportTaskSet" - } - } - }, - "DescribeFlowLogsRequest":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - }, - "Filter":{"shape":"FilterList"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogs":{ - "shape":"FlowLogSet", - "locationName":"flowLogSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostReservationOfferingsRequest":{ - "type":"structure", - "members":{ - "OfferingId":{"shape":"String"}, - "MinDuration":{"shape":"Integer"}, - "MaxDuration":{"shape":"Integer"}, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeHostReservationOfferingsResult":{ - "type":"structure", - "members":{ - "OfferingSet":{ - "shape":"HostOfferingSet", - "locationName":"offeringSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostReservationsRequest":{ - "type":"structure", - "members":{ - "HostReservationIdSet":{"shape":"HostReservationIdSet"}, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeHostReservationsResult":{ - "type":"structure", - "members":{ - "HostReservationSet":{ - "shape":"HostReservationSet", - "locationName":"hostReservationSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostsRequest":{ - "type":"structure", - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "Filter":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeHostsResult":{ - "type":"structure", - "members":{ - "Hosts":{ - "shape":"HostList", - "locationName":"hostSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeIdFormatRequest":{ - "type":"structure", - "members":{ - "Resource":{"shape":"String"} - } - }, - "DescribeIdFormatResult":{ - "type":"structure", - "members":{ - "Statuses":{ - "shape":"IdFormatList", - "locationName":"statusSet" - } - } - }, - "DescribeIdentityIdFormatRequest":{ - "type":"structure", - "required":["PrincipalArn"], - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "PrincipalArn":{ - "shape":"String", - "locationName":"principalArn" - } - } - }, - "DescribeIdentityIdFormatResult":{ - "type":"structure", - "members":{ - "Statuses":{ - "shape":"IdFormatList", - "locationName":"statusSet" - } - } - }, - "DescribeImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ImageAttributeName"} - } - }, - "DescribeImagesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageIds":{ - "shape":"ImageIdStringList", - "locationName":"ImageId" - }, - "Owners":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "ExecutableUsers":{ - "shape":"ExecutableByStringList", - "locationName":"ExecutableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeImagesResult":{ - "type":"structure", - "members":{ - "Images":{ - "shape":"ImageList", - "locationName":"imagesSet" - } - } - }, - "DescribeImportImageTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportImageTasksResult":{ - "type":"structure", - "members":{ - "ImportImageTasks":{ - "shape":"ImportImageTaskList", - "locationName":"importImageTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeImportSnapshotTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportSnapshotTasksResult":{ - "type":"structure", - "members":{ - "ImportSnapshotTasks":{ - "shape":"ImportSnapshotTaskList", - "locationName":"importSnapshotTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "DescribeInstanceStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "IncludeAllInstances":{ - "shape":"Boolean", - "locationName":"includeAllInstances" - } - } - }, - "DescribeInstanceStatusResult":{ - "type":"structure", - "members":{ - "InstanceStatuses":{ - "shape":"InstanceStatusList", - "locationName":"instanceStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeInstancesResult":{ - "type":"structure", - "members":{ - "Reservations":{ - "shape":"ReservationList", - "locationName":"reservationSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInternetGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayIds":{ - "shape":"ValueStringList", - "locationName":"internetGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeInternetGatewaysResult":{ - "type":"structure", - "members":{ - "InternetGateways":{ - "shape":"InternetGatewayList", - "locationName":"internetGatewaySet" - } - } - }, - "DescribeKeyPairsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyNames":{ - "shape":"KeyNameStringList", - "locationName":"KeyName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeKeyPairsResult":{ - "type":"structure", - "members":{ - "KeyPairs":{ - "shape":"KeyPairList", - "locationName":"keySet" - } - } - }, - "DescribeMovingAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"ValueStringList", - "locationName":"publicIp" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeMovingAddressesResult":{ - "type":"structure", - "members":{ - "MovingAddressStatuses":{ - "shape":"MovingAddressStatusSet", - "locationName":"movingAddressStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNatGatewaysRequest":{ - "type":"structure", - "members":{ - "NatGatewayIds":{ - "shape":"ValueStringList", - "locationName":"NatGatewayId" - }, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeNatGatewaysResult":{ - "type":"structure", - "members":{ - "NatGateways":{ - "shape":"NatGatewayList", - "locationName":"natGatewaySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNetworkAclsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclIds":{ - "shape":"ValueStringList", - "locationName":"NetworkAclId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeNetworkAclsResult":{ - "type":"structure", - "members":{ - "NetworkAcls":{ - "shape":"NetworkAclList", - "locationName":"networkAclSet" - } - } - }, - "DescribeNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Attribute":{ - "shape":"NetworkInterfaceAttribute", - "locationName":"attribute" - } - } - }, - "DescribeNetworkInterfaceAttributeResult":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - } - } - }, - "DescribeNetworkInterfacesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceIds":{ - "shape":"NetworkInterfaceIdList", - "locationName":"NetworkInterfaceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeNetworkInterfacesResult":{ - "type":"structure", - "members":{ - "NetworkInterfaces":{ - "shape":"NetworkInterfaceList", - "locationName":"networkInterfaceSet" - } - } - }, - "DescribePlacementGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"PlacementGroupStringList", - "locationName":"groupName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribePlacementGroupsResult":{ - "type":"structure", - "members":{ - "PlacementGroups":{ - "shape":"PlacementGroupList", - "locationName":"placementGroupSet" - } - } - }, - "DescribePrefixListsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "PrefixListIds":{ - "shape":"ValueStringList", - "locationName":"PrefixListId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribePrefixListsResult":{ - "type":"structure", - "members":{ - "PrefixLists":{ - "shape":"PrefixListSet", - "locationName":"prefixListSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeRegionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RegionNames":{ - "shape":"RegionNameStringList", - "locationName":"RegionName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRegionsResult":{ - "type":"structure", - "members":{ - "Regions":{ - "shape":"RegionList", - "locationName":"regionInfo" - } - } - }, - "DescribeReservedInstancesListingsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filters" - } - } - }, - "DescribeReservedInstancesListingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "DescribeReservedInstancesModificationsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationIds":{ - "shape":"ReservedInstancesModificationIdStringList", - "locationName":"ReservedInstancesModificationId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeReservedInstancesModificationsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModifications":{ - "shape":"ReservedInstancesModificationList", - "locationName":"reservedInstancesModificationsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesOfferingsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingIds":{ - "shape":"ReservedInstancesOfferingIdStringList", - "locationName":"ReservedInstancesOfferingId" - }, - "InstanceType":{"shape":"InstanceType"}, - "AvailabilityZone":{"shape":"String"}, - "ProductDescription":{"shape":"RIProductDescription"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "IncludeMarketplace":{"shape":"Boolean"}, - "MinDuration":{"shape":"Long"}, - "MaxDuration":{"shape":"Long"}, - "MaxInstanceCount":{"shape":"Integer"} - } - }, - "DescribeReservedInstancesOfferingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferings":{ - "shape":"ReservedInstancesOfferingList", - "locationName":"reservedInstancesOfferingsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - } - } - }, - "DescribeReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstances":{ - "shape":"ReservedInstancesList", - "locationName":"reservedInstancesSet" - } - } - }, - "DescribeRouteTablesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRouteTablesResult":{ - "type":"structure", - "members":{ - "RouteTables":{ - "shape":"RouteTableList", - "locationName":"routeTableSet" - } - } - }, - "DescribeScheduledInstanceAvailabilityRequest":{ - "type":"structure", - "required":[ - "Recurrence", - "FirstSlotStartTimeRange" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "Recurrence":{"shape":"ScheduledInstanceRecurrenceRequest"}, - "FirstSlotStartTimeRange":{"shape":"SlotDateTimeRangeRequest"}, - "MinSlotDurationInHours":{"shape":"Integer"}, - "MaxSlotDurationInHours":{"shape":"Integer"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeScheduledInstanceAvailabilityResult":{ - "type":"structure", - "members":{ - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "ScheduledInstanceAvailabilitySet":{ - "shape":"ScheduledInstanceAvailabilitySet", - "locationName":"scheduledInstanceAvailabilitySet" - } - } - }, - "DescribeScheduledInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ScheduledInstanceIds":{ - "shape":"ScheduledInstanceIdRequestSet", - "locationName":"ScheduledInstanceId" - }, - "SlotStartTimeRange":{"shape":"SlotStartTimeRangeRequest"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeScheduledInstancesResult":{ - "type":"structure", - "members":{ - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "ScheduledInstanceSet":{ - "shape":"ScheduledInstanceSet", - "locationName":"scheduledInstanceSet" - } - } - }, - "DescribeSecurityGroupReferencesRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "GroupId":{"shape":"GroupIds"} - } - }, - "DescribeSecurityGroupReferencesResult":{ - "type":"structure", - "members":{ - "SecurityGroupReferenceSet":{ - "shape":"SecurityGroupReferences", - "locationName":"securityGroupReferenceSet" - } - } - }, - "DescribeSecurityGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSecurityGroupsResult":{ - "type":"structure", - "members":{ - "SecurityGroups":{ - "shape":"SecurityGroupList", - "locationName":"securityGroupInfo" - } - } - }, - "DescribeSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "DescribeSnapshotAttributeResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "CreateVolumePermissions":{ - "shape":"CreateVolumePermissionList", - "locationName":"createVolumePermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeSnapshotsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotIds":{ - "shape":"SnapshotIdStringList", - "locationName":"SnapshotId" - }, - "OwnerIds":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "RestorableByUserIds":{ - "shape":"RestorableByStringList", - "locationName":"RestorableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeSnapshotsResult":{ - "type":"structure", - "members":{ - "Snapshots":{ - "shape":"SnapshotList", - "locationName":"snapshotSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DescribeSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "DescribeSpotFleetInstancesRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetInstancesResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "ActiveInstances" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "ActiveInstances":{ - "shape":"ActiveInstanceSet", - "locationName":"activeInstanceSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestHistoryRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestHistoryResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime", - "LastEvaluatedTime", - "HistoryRecords" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "LastEvaluatedTime":{ - "shape":"DateTime", - "locationName":"lastEvaluatedTime" - }, - "HistoryRecords":{ - "shape":"HistoryRecords", - "locationName":"historyRecordSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestsResponse":{ - "type":"structure", - "required":["SpotFleetRequestConfigs"], - "members":{ - "SpotFleetRequestConfigs":{ - "shape":"SpotFleetRequestConfigSet", - "locationName":"spotFleetRequestConfigSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotInstanceRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "DescribeSpotPriceHistoryRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "InstanceTypes":{ - "shape":"InstanceTypeList", - "locationName":"InstanceType" - }, - "ProductDescriptions":{ - "shape":"ProductDescriptionList", - "locationName":"ProductDescription" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotPriceHistoryResult":{ - "type":"structure", - "members":{ - "SpotPriceHistory":{ - "shape":"SpotPriceHistoryList", - "locationName":"spotPriceHistorySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeStaleSecurityGroupsRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "MaxResults":{"shape":"MaxResults"}, - "NextToken":{"shape":"NextToken"} - } - }, - "DescribeStaleSecurityGroupsResult":{ - "type":"structure", - "members":{ - "StaleSecurityGroupSet":{ - "shape":"StaleSecurityGroupSet", - "locationName":"staleSecurityGroupSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSubnetsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetIds":{ - "shape":"SubnetIdStringList", - "locationName":"SubnetId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSubnetsResult":{ - "type":"structure", - "members":{ - "Subnets":{ - "shape":"SubnetList", - "locationName":"subnetSet" - } - } - }, - "DescribeTagsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeTagsResult":{ - "type":"structure", - "members":{ - "Tags":{ - "shape":"TagDescriptionList", - "locationName":"tagSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Attribute":{"shape":"VolumeAttributeName"} - } - }, - "DescribeVolumeAttributeResult":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AutoEnableIO":{ - "shape":"AttributeBooleanValue", - "locationName":"autoEnableIO" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeVolumeStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeVolumeStatusResult":{ - "type":"structure", - "members":{ - "VolumeStatuses":{ - "shape":"VolumeStatusList", - "locationName":"volumeStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeVolumesResult":{ - "type":"structure", - "members":{ - "Volumes":{ - "shape":"VolumeList", - "locationName":"volumeSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcAttributeRequest":{ - "type":"structure", - "required":[ - "VpcId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "Attribute":{"shape":"VpcAttributeName"} - } - }, - "DescribeVpcAttributeResult":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsSupport" - }, - "EnableDnsHostnames":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsHostnames" - } - } - }, - "DescribeVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcIds":{"shape":"VpcClassicLinkIdList"}, - "MaxResults":{ - "shape":"MaxResults", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"NextToken", - "locationName":"nextToken" - } - } - }, - "DescribeVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"ClassicLinkDnsSupportList", - "locationName":"vpcs" - }, - "NextToken":{ - "shape":"NextToken", - "locationName":"nextToken" - } - } - }, - "DescribeVpcClassicLinkRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcClassicLinkIdList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcClassicLinkList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpcEndpointServicesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointServicesResult":{ - "type":"structure", - "members":{ - "ServiceNames":{ - "shape":"ValueStringList", - "locationName":"serviceNameSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcEndpointsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointsResult":{ - "type":"structure", - "members":{ - "VpcEndpoints":{ - "shape":"VpcEndpointSet", - "locationName":"vpcEndpointSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcPeeringConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionIds":{ - "shape":"ValueStringList", - "locationName":"VpcPeeringConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcPeeringConnectionsResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnections":{ - "shape":"VpcPeeringConnectionList", - "locationName":"vpcPeeringConnectionSet" - } - } - }, - "DescribeVpcsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcIdStringList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcsResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpnConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionIds":{ - "shape":"VpnConnectionIdStringList", - "locationName":"VpnConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnConnectionsResult":{ - "type":"structure", - "members":{ - "VpnConnections":{ - "shape":"VpnConnectionList", - "locationName":"vpnConnectionSet" - } - } - }, - "DescribeVpnGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayIds":{ - "shape":"VpnGatewayIdStringList", - "locationName":"VpnGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnGatewaysResult":{ - "type":"structure", - "members":{ - "VpnGateways":{ - "shape":"VpnGatewayList", - "locationName":"vpnGatewaySet" - } - } - }, - "DetachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DetachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachNetworkInterfaceRequest":{ - "type":"structure", - "required":["AttachmentId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "DetachVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"}, - "Force":{"shape":"Boolean"} - } - }, - "DetachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "DeviceType":{ - "type":"string", - "enum":[ - "ebs", - "instance-store" - ] - }, - "DhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"DhcpConfigurationValueList", - "locationName":"valueSet" - } - } - }, - "DhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"DhcpConfiguration", - "locationName":"item" - } - }, - "DhcpConfigurationValueList":{ - "type":"list", - "member":{ - "shape":"AttributeValue", - "locationName":"item" - } - }, - "DhcpOptions":{ - "type":"structure", - "members":{ - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "DhcpConfigurations":{ - "shape":"DhcpConfigurationList", - "locationName":"dhcpConfigurationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "DhcpOptionsIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DhcpOptionsId" - } - }, - "DhcpOptionsList":{ - "type":"list", - "member":{ - "shape":"DhcpOptions", - "locationName":"item" - } - }, - "DisableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DisableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisassociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AssociationId":{"shape":"String"} - } - }, - "DisassociateRouteTableRequest":{ - "type":"structure", - "required":["AssociationId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "DiskImage":{ - "type":"structure", - "members":{ - "Image":{"shape":"DiskImageDetail"}, - "Description":{"shape":"String"}, - "Volume":{"shape":"VolumeDetail"} - } - }, - "DiskImageDescription":{ - "type":"structure", - "required":[ - "Format", - "Size", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - }, - "Checksum":{ - "shape":"String", - "locationName":"checksum" - } - } - }, - "DiskImageDetail":{ - "type":"structure", - "required":[ - "Format", - "Bytes", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Bytes":{ - "shape":"Long", - "locationName":"bytes" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - } - } - }, - "DiskImageFormat":{ - "type":"string", - "enum":[ - "VMDK", - "RAW", - "VHD" - ] - }, - "DiskImageList":{ - "type":"list", - "member":{"shape":"DiskImage"} - }, - "DiskImageVolumeDescription":{ - "type":"structure", - "required":["Id"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "DomainType":{ - "type":"string", - "enum":[ - "vpc", - "standard" - ] - }, - "Double":{"type":"double"}, - "EbsBlockDevice":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - } - } - }, - "EbsInstanceBlockDevice":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EbsInstanceBlockDeviceSpecification":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EnableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "EnableVolumeIORequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - } - } - }, - "EnableVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcId":{"shape":"String"} - } - }, - "EnableVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EnableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "EnableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EventCode":{ - "type":"string", - "enum":[ - "instance-reboot", - "system-reboot", - "system-maintenance", - "instance-retirement", - "instance-stop" - ] - }, - "EventInformation":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "EventSubType":{ - "shape":"String", - "locationName":"eventSubType" - }, - "EventDescription":{ - "shape":"String", - "locationName":"eventDescription" - } - } - }, - "EventType":{ - "type":"string", - "enum":[ - "instanceChange", - "fleetRequestChange", - "error" - ] - }, - "ExcessCapacityTerminationPolicy":{ - "type":"string", - "enum":[ - "noTermination", - "default" - ] - }, - "ExecutableByStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExecutableBy" - } - }, - "ExportEnvironment":{ - "type":"string", - "enum":[ - "citrix", - "vmware", - "microsoft" - ] - }, - "ExportTask":{ - "type":"structure", - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "State":{ - "shape":"ExportTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceExportDetails":{ - "shape":"InstanceExportDetails", - "locationName":"instanceExport" - }, - "ExportToS3Task":{ - "shape":"ExportToS3Task", - "locationName":"exportToS3" - } - } - }, - "ExportTaskIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExportTaskId" - } - }, - "ExportTaskList":{ - "type":"list", - "member":{ - "shape":"ExportTask", - "locationName":"item" - } - }, - "ExportTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "ExportToS3Task":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "ExportToS3TaskSpecification":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Prefix":{ - "shape":"String", - "locationName":"s3Prefix" - } - } - }, - "Filter":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "FilterList":{ - "type":"list", - "member":{ - "shape":"Filter", - "locationName":"Filter" - } - }, - "FleetType":{ - "type":"string", - "enum":[ - "request", - "maintain" - ] - }, - "Float":{"type":"float"}, - "FlowLog":{ - "type":"structure", - "members":{ - "CreationTime":{ - "shape":"DateTime", - "locationName":"creationTime" - }, - "FlowLogId":{ - "shape":"String", - "locationName":"flowLogId" - }, - "FlowLogStatus":{ - "shape":"String", - "locationName":"flowLogStatus" - }, - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "TrafficType":{ - "shape":"TrafficType", - "locationName":"trafficType" - }, - "LogGroupName":{ - "shape":"String", - "locationName":"logGroupName" - }, - "DeliverLogsStatus":{ - "shape":"String", - "locationName":"deliverLogsStatus" - }, - "DeliverLogsErrorMessage":{ - "shape":"String", - "locationName":"deliverLogsErrorMessage" - }, - "DeliverLogsPermissionArn":{ - "shape":"String", - "locationName":"deliverLogsPermissionArn" - } - } - }, - "FlowLogSet":{ - "type":"list", - "member":{ - "shape":"FlowLog", - "locationName":"item" - } - }, - "FlowLogsResourceType":{ - "type":"string", - "enum":[ - "VPC", - "Subnet", - "NetworkInterface" - ] - }, - "GatewayType":{ - "type":"string", - "enum":["ipsec.1"] - }, - "GetConsoleOutputRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetConsoleOutputResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "Output":{ - "shape":"String", - "locationName":"output" - } - } - }, - "GetConsoleScreenshotRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "InstanceId":{"shape":"String"}, - "WakeUp":{"shape":"Boolean"} - } - }, - "GetConsoleScreenshotResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageData":{ - "shape":"String", - "locationName":"imageData" - } - } - }, - "GetHostReservationPurchasePreviewRequest":{ - "type":"structure", - "required":[ - "OfferingId", - "HostIdSet" - ], - "members":{ - "OfferingId":{"shape":"String"}, - "HostIdSet":{"shape":"RequestHostIdSet"} - } - }, - "GetHostReservationPurchasePreviewResult":{ - "type":"structure", - "members":{ - "Purchase":{ - "shape":"PurchaseSet", - "locationName":"purchase" - }, - "TotalUpfrontPrice":{ - "shape":"String", - "locationName":"totalUpfrontPrice" - }, - "TotalHourlyPrice":{ - "shape":"String", - "locationName":"totalHourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "GetPasswordDataRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetPasswordDataResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "PasswordData":{ - "shape":"String", - "locationName":"passwordData" - } - } - }, - "GroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"groupId" - } - }, - "GroupIdentifier":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "GroupIdentifierList":{ - "type":"list", - "member":{ - "shape":"GroupIdentifier", - "locationName":"item" - } - }, - "GroupIds":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "GroupNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"GroupName" - } - }, - "HistoryRecord":{ - "type":"structure", - "required":[ - "Timestamp", - "EventType", - "EventInformation" - ], - "members":{ - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "EventInformation":{ - "shape":"EventInformation", - "locationName":"eventInformation" - } - } - }, - "HistoryRecords":{ - "type":"list", - "member":{ - "shape":"HistoryRecord", - "locationName":"item" - } - }, - "Host":{ - "type":"structure", - "members":{ - "HostId":{ - "shape":"String", - "locationName":"hostId" - }, - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - }, - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "HostProperties":{ - "shape":"HostProperties", - "locationName":"hostProperties" - }, - "State":{ - "shape":"AllocationState", - "locationName":"state" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Instances":{ - "shape":"HostInstanceList", - "locationName":"instances" - }, - "AvailableCapacity":{ - "shape":"AvailableCapacity", - "locationName":"availableCapacity" - } - } - }, - "HostInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - } - } - }, - "HostInstanceList":{ - "type":"list", - "member":{ - "shape":"HostInstance", - "locationName":"item" - } - }, - "HostList":{ - "type":"list", - "member":{ - "shape":"Host", - "locationName":"item" - } - }, - "HostOffering":{ - "type":"structure", - "members":{ - "OfferingId":{ - "shape":"String", - "locationName":"offeringId" - }, - "InstanceFamily":{ - "shape":"String", - "locationName":"instanceFamily" - }, - "PaymentOption":{ - "shape":"PaymentOption", - "locationName":"paymentOption" - }, - "UpfrontPrice":{ - "shape":"String", - "locationName":"upfrontPrice" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Duration":{ - "shape":"Integer", - "locationName":"duration" - } - } - }, - "HostOfferingSet":{ - "type":"list", - "member":{"shape":"HostOffering"} - }, - "HostProperties":{ - "type":"structure", - "members":{ - "Sockets":{ - "shape":"Integer", - "locationName":"sockets" - }, - "Cores":{ - "shape":"Integer", - "locationName":"cores" - }, - "TotalVCpus":{ - "shape":"Integer", - "locationName":"totalVCpus" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - } - } - }, - "HostReservation":{ - "type":"structure", - "members":{ - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "HostIdSet":{ - "shape":"ResponseHostIdSet", - "locationName":"hostIdSet" - }, - "OfferingId":{ - "shape":"String", - "locationName":"offeringId" - }, - "InstanceFamily":{ - "shape":"String", - "locationName":"instanceFamily" - }, - "PaymentOption":{ - "shape":"PaymentOption", - "locationName":"paymentOption" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "UpfrontPrice":{ - "shape":"String", - "locationName":"upfrontPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Count":{ - "shape":"Integer", - "locationName":"count" - }, - "Duration":{ - "shape":"Integer", - "locationName":"duration" - }, - "End":{ - "shape":"DateTime", - "locationName":"end" - }, - "Start":{ - "shape":"DateTime", - "locationName":"start" - }, - "State":{ - "shape":"ReservationState", - "locationName":"state" - } - } - }, - "HostReservationIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "HostReservationSet":{ - "type":"list", - "member":{"shape":"HostReservation"} - }, - "HostTenancy":{ - "type":"string", - "enum":[ - "dedicated", - "host" - ] - }, - "HypervisorType":{ - "type":"string", - "enum":[ - "ovm", - "xen" - ] - }, - "IamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "IamInstanceProfileSpecification":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Name":{ - "shape":"String", - "locationName":"name" - } - } - }, - "IcmpTypeCode":{ - "type":"structure", - "members":{ - "Type":{ - "shape":"Integer", - "locationName":"type" - }, - "Code":{ - "shape":"Integer", - "locationName":"code" - } - } - }, - "IdFormat":{ - "type":"structure", - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "UseLongIds":{ - "shape":"Boolean", - "locationName":"useLongIds" - }, - "Deadline":{ - "shape":"DateTime", - "locationName":"deadline" - } - } - }, - "IdFormatList":{ - "type":"list", - "member":{ - "shape":"IdFormat", - "locationName":"item" - } - }, - "Image":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "ImageLocation":{ - "shape":"String", - "locationName":"imageLocation" - }, - "State":{ - "shape":"ImageState", - "locationName":"imageState" - }, - "OwnerId":{ - "shape":"String", - "locationName":"imageOwnerId" - }, - "CreationDate":{ - "shape":"String", - "locationName":"creationDate" - }, - "Public":{ - "shape":"Boolean", - "locationName":"isPublic" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "ImageType":{ - "shape":"ImageTypeValues", - "locationName":"imageType" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"Boolean", - "locationName":"enaSupport" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "ImageOwnerAlias":{ - "shape":"String", - "locationName":"imageOwnerAlias" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - } - } - }, - "ImageAttribute":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "LaunchPermissions":{ - "shape":"LaunchPermissionList", - "locationName":"launchPermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - } - } - }, - "ImageAttributeName":{ - "type":"string", - "enum":[ - "description", - "kernel", - "ramdisk", - "launchPermission", - "productCodes", - "blockDeviceMapping", - "sriovNetSupport" - ] - }, - "ImageDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"}, - "DeviceName":{"shape":"String"}, - "SnapshotId":{"shape":"String"} - } - }, - "ImageDiskContainerList":{ - "type":"list", - "member":{ - "shape":"ImageDiskContainer", - "locationName":"item" - } - }, - "ImageIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImageId" - } - }, - "ImageList":{ - "type":"list", - "member":{ - "shape":"Image", - "locationName":"item" - } - }, - "ImageState":{ - "type":"string", - "enum":[ - "pending", - "available", - "invalid", - "deregistered", - "transient", - "failed", - "error" - ] - }, - "ImageTypeValues":{ - "type":"string", - "enum":[ - "machine", - "kernel", - "ramdisk" - ] - }, - "ImportImageRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainers":{ - "shape":"ImageDiskContainerList", - "locationName":"DiskContainer" - }, - "LicenseType":{"shape":"String"}, - "Hypervisor":{"shape":"String"}, - "Architecture":{"shape":"String"}, - "Platform":{"shape":"String"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportImageResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTaskList":{ - "type":"list", - "member":{ - "shape":"ImportImageTask", - "locationName":"item" - } - }, - "ImportInstanceLaunchSpecification":{ - "type":"structure", - "members":{ - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "GroupNames":{ - "shape":"SecurityGroupStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"GroupId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "UserData":{ - "shape":"UserData", - "locationName":"userData" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "Monitoring":{ - "shape":"Boolean", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "ImportInstanceRequest":{ - "type":"structure", - "required":["Platform"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "LaunchSpecification":{ - "shape":"ImportInstanceLaunchSpecification", - "locationName":"launchSpecification" - }, - "DiskImages":{ - "shape":"DiskImageList", - "locationName":"diskImage" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - } - } - }, - "ImportInstanceResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportInstanceTaskDetails":{ - "type":"structure", - "required":["Volumes"], - "members":{ - "Volumes":{ - "shape":"ImportInstanceVolumeDetailSet", - "locationName":"volumes" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailItem":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume", - "Status" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailSet":{ - "type":"list", - "member":{ - "shape":"ImportInstanceVolumeDetailItem", - "locationName":"item" - } - }, - "ImportKeyPairRequest":{ - "type":"structure", - "required":[ - "KeyName", - "PublicKeyMaterial" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "PublicKeyMaterial":{ - "shape":"Blob", - "locationName":"publicKeyMaterial" - } - } - }, - "ImportKeyPairResult":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "ImportSnapshotRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainer":{"shape":"SnapshotDiskContainer"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportSnapshotResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTaskList":{ - "type":"list", - "member":{ - "shape":"ImportSnapshotTask", - "locationName":"item" - } - }, - "ImportTaskIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImportTaskId" - } - }, - "ImportVolumeRequest":{ - "type":"structure", - "required":[ - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDetail", - "locationName":"image" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Volume":{ - "shape":"VolumeDetail", - "locationName":"volume" - } - } - }, - "ImportVolumeResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportVolumeTaskDetails":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - } - } - }, - "Instance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "State":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"dnsName" - }, - "StateTransitionReason":{ - "shape":"String", - "locationName":"reason" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "AmiLaunchIndex":{ - "shape":"Integer", - "locationName":"amiLaunchIndex" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "LaunchTime":{ - "shape":"DateTime", - "locationName":"launchTime" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PublicIpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "InstanceLifecycle":{ - "shape":"InstanceLifecycleType", - "locationName":"instanceLifecycle" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfile", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"Boolean", - "locationName":"enaSupport" - } - } - }, - "InstanceAttribute":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"AttributeValue", - "locationName":"userData" - }, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "RootDeviceName":{ - "shape":"AttributeValue", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enaSupport" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - } - } - }, - "InstanceAttributeName":{ - "type":"string", - "enum":[ - "instanceType", - "kernel", - "ramdisk", - "userData", - "disableApiTermination", - "instanceInitiatedShutdownBehavior", - "rootDeviceName", - "blockDeviceMapping", - "productCodes", - "sourceDestCheck", - "groupSet", - "ebsOptimized", - "sriovNetSupport", - "enaSupport" - ] - }, - "InstanceBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDevice", - "locationName":"ebs" - } - } - }, - "InstanceBlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMapping", - "locationName":"item" - } - }, - "InstanceBlockDeviceMappingSpecification":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDeviceSpecification", - "locationName":"ebs" - }, - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "InstanceBlockDeviceMappingSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMappingSpecification", - "locationName":"item" - } - }, - "InstanceCapacity":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "AvailableCapacity":{ - "shape":"Integer", - "locationName":"availableCapacity" - }, - "TotalCapacity":{ - "shape":"Integer", - "locationName":"totalCapacity" - } - } - }, - "InstanceCount":{ - "type":"structure", - "members":{ - "State":{ - "shape":"ListingState", - "locationName":"state" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - } - } - }, - "InstanceCountList":{ - "type":"list", - "member":{ - "shape":"InstanceCount", - "locationName":"item" - } - }, - "InstanceExportDetails":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - } - } - }, - "InstanceIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "InstanceIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"InstanceId" - } - }, - "InstanceLifecycleType":{ - "type":"string", - "enum":[ - "spot", - "scheduled" - ] - }, - "InstanceList":{ - "type":"list", - "member":{ - "shape":"Instance", - "locationName":"item" - } - }, - "InstanceMonitoring":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - } - } - }, - "InstanceMonitoringList":{ - "type":"list", - "member":{ - "shape":"InstanceMonitoring", - "locationName":"item" - } - }, - "InstanceNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"InstanceNetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - }, - "PrivateIpAddresses":{ - "shape":"InstancePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - } - } - }, - "InstanceNetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - } - } - }, - "InstanceNetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "InstanceNetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterface", - "locationName":"item" - } - }, - "InstanceNetworkInterfaceSpecification":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddressesSet", - "queryName":"PrivateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AssociatePublicIpAddress":{ - "shape":"Boolean", - "locationName":"associatePublicIpAddress" - } - } - }, - "InstanceNetworkInterfaceSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterfaceSpecification", - "locationName":"item" - } - }, - "InstancePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "InstancePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"InstancePrivateIpAddress", - "locationName":"item" - } - }, - "InstanceState":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"Integer", - "locationName":"code" - }, - "Name":{ - "shape":"InstanceStateName", - "locationName":"name" - } - } - }, - "InstanceStateChange":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CurrentState":{ - "shape":"InstanceState", - "locationName":"currentState" - }, - "PreviousState":{ - "shape":"InstanceState", - "locationName":"previousState" - } - } - }, - "InstanceStateChangeList":{ - "type":"list", - "member":{ - "shape":"InstanceStateChange", - "locationName":"item" - } - }, - "InstanceStateName":{ - "type":"string", - "enum":[ - "pending", - "running", - "shutting-down", - "terminated", - "stopping", - "stopped" - ] - }, - "InstanceStatus":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Events":{ - "shape":"InstanceStatusEventList", - "locationName":"eventsSet" - }, - "InstanceState":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "SystemStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"systemStatus" - }, - "InstanceStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"instanceStatus" - } - } - }, - "InstanceStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"StatusName", - "locationName":"name" - }, - "Status":{ - "shape":"StatusType", - "locationName":"status" - }, - "ImpairedSince":{ - "shape":"DateTime", - "locationName":"impairedSince" - } - } - }, - "InstanceStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusDetails", - "locationName":"item" - } - }, - "InstanceStatusEvent":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"EventCode", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - } - } - }, - "InstanceStatusEventList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusEvent", - "locationName":"item" - } - }, - "InstanceStatusList":{ - "type":"list", - "member":{ - "shape":"InstanceStatus", - "locationName":"item" - } - }, - "InstanceStatusSummary":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"SummaryStatus", - "locationName":"status" - }, - "Details":{ - "shape":"InstanceStatusDetailsList", - "locationName":"details" - } - } - }, - "InstanceType":{ - "type":"string", - "enum":[ - "t1.micro", - "t2.nano", - "t2.micro", - "t2.small", - "t2.medium", - "t2.large", - "m1.small", - "m1.medium", - "m1.large", - "m1.xlarge", - "m3.medium", - "m3.large", - "m3.xlarge", - "m3.2xlarge", - "m4.large", - "m4.xlarge", - "m4.2xlarge", - "m4.4xlarge", - "m4.10xlarge", - "m2.xlarge", - "m2.2xlarge", - "m2.4xlarge", - "cr1.8xlarge", - "r3.large", - "r3.xlarge", - "r3.2xlarge", - "r3.4xlarge", - "r3.8xlarge", - "x1.4xlarge", - "x1.8xlarge", - "x1.16xlarge", - "x1.32xlarge", - "i2.xlarge", - "i2.2xlarge", - "i2.4xlarge", - "i2.8xlarge", - "hi1.4xlarge", - "hs1.8xlarge", - "c1.medium", - "c1.xlarge", - "c3.large", - "c3.xlarge", - "c3.2xlarge", - "c3.4xlarge", - "c3.8xlarge", - "c4.large", - "c4.xlarge", - "c4.2xlarge", - "c4.4xlarge", - "c4.8xlarge", - "cc1.4xlarge", - "cc2.8xlarge", - "g2.2xlarge", - "g2.8xlarge", - "cg1.4xlarge", - "d2.xlarge", - "d2.2xlarge", - "d2.4xlarge", - "d2.8xlarge" - ] - }, - "InstanceTypeList":{ - "type":"list", - "member":{"shape":"InstanceType"} - }, - "Integer":{"type":"integer"}, - "InternetGateway":{ - "type":"structure", - "members":{ - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "Attachments":{ - "shape":"InternetGatewayAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "InternetGatewayAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "InternetGatewayAttachmentList":{ - "type":"list", - "member":{ - "shape":"InternetGatewayAttachment", - "locationName":"item" - } - }, - "InternetGatewayList":{ - "type":"list", - "member":{ - "shape":"InternetGateway", - "locationName":"item" - } - }, - "IpPermission":{ - "type":"structure", - "members":{ - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairList", - "locationName":"groups" - }, - "IpRanges":{ - "shape":"IpRangeList", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdList", - "locationName":"prefixListIds" - } - } - }, - "IpPermissionList":{ - "type":"list", - "member":{ - "shape":"IpPermission", - "locationName":"item" - } - }, - "IpRange":{ - "type":"structure", - "members":{ - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - } - } - }, - "IpRangeList":{ - "type":"list", - "member":{ - "shape":"IpRange", - "locationName":"item" - } - }, - "IpRanges":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "KeyNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"KeyName" - } - }, - "KeyPair":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - }, - "KeyMaterial":{ - "shape":"String", - "locationName":"keyMaterial" - } - } - }, - "KeyPairInfo":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "KeyPairList":{ - "type":"list", - "member":{ - "shape":"KeyPairInfo", - "locationName":"item" - } - }, - "LaunchPermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "LaunchPermissionList":{ - "type":"list", - "member":{ - "shape":"LaunchPermission", - "locationName":"item" - } - }, - "LaunchPermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"LaunchPermissionList"}, - "Remove":{"shape":"LaunchPermissionList"} - } - }, - "LaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - } - } - }, - "LaunchSpecsList":{ - "type":"list", - "member":{ - "shape":"SpotFleetLaunchSpecification", - "locationName":"item" - }, - "min":1 - }, - "ListingState":{ - "type":"string", - "enum":[ - "available", - "sold", - "cancelled", - "pending" - ] - }, - "ListingStatus":{ - "type":"string", - "enum":[ - "active", - "pending", - "cancelled", - "closed" - ] - }, - "Long":{"type":"long"}, - "MaxResults":{ - "type":"integer", - "max":255, - "min":5 - }, - "ModifyHostsRequest":{ - "type":"structure", - "required":[ - "HostIds", - "AutoPlacement" - ], - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - }, - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - } - } - }, - "ModifyHostsResult":{ - "type":"structure", - "members":{ - "Successful":{ - "shape":"ResponseHostIdList", - "locationName":"successful" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemList", - "locationName":"unsuccessful" - } - } - }, - "ModifyIdFormatRequest":{ - "type":"structure", - "required":[ - "Resource", - "UseLongIds" - ], - "members":{ - "Resource":{"shape":"String"}, - "UseLongIds":{"shape":"Boolean"} - } - }, - "ModifyIdentityIdFormatRequest":{ - "type":"structure", - "required":[ - "Resource", - "UseLongIds", - "PrincipalArn" - ], - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "UseLongIds":{ - "shape":"Boolean", - "locationName":"useLongIds" - }, - "PrincipalArn":{ - "shape":"String", - "locationName":"principalArn" - } - } - }, - "ModifyImageAttributeRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"String"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "UserGroups":{ - "shape":"UserGroupStringList", - "locationName":"UserGroup" - }, - "ProductCodes":{ - "shape":"ProductCodeStringList", - "locationName":"ProductCode" - }, - "Value":{"shape":"String"}, - "LaunchPermission":{"shape":"LaunchPermissionModifications"}, - "Description":{"shape":"AttributeValue"} - } - }, - "ModifyInstanceAttributeRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - }, - "Value":{ - "shape":"String", - "locationName":"value" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingSpecificationList", - "locationName":"blockDeviceMapping" - }, - "SourceDestCheck":{"shape":"AttributeBooleanValue"}, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "Kernel":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "Ramdisk":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"BlobAttributeValue", - "locationName":"userData" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enaSupport" - } - } - }, - "ModifyInstancePlacementRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Tenancy":{ - "shape":"HostTenancy", - "locationName":"tenancy" - }, - "Affinity":{ - "shape":"Affinity", - "locationName":"affinity" - }, - "HostId":{ - "shape":"String", - "locationName":"hostId" - } - } - }, - "ModifyInstancePlacementResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifyNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachmentChanges", - "locationName":"attachment" - } - } - }, - "ModifyReservedInstancesRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesIds", - "TargetConfigurations" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "TargetConfigurations":{ - "shape":"ReservedInstancesConfigurationList", - "locationName":"ReservedInstancesConfigurationSetItemType" - } - } - }, - "ModifyReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - } - } - }, - "ModifySnapshotAttributeRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"UserGroup" - }, - "CreateVolumePermission":{"shape":"CreateVolumePermissionModifications"} - } - }, - "ModifySpotFleetRequestRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ExcessCapacityTerminationPolicy":{ - "shape":"ExcessCapacityTerminationPolicy", - "locationName":"excessCapacityTerminationPolicy" - } - } - }, - "ModifySpotFleetRequestResponse":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifySubnetAttributeRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "MapPublicIpOnLaunch":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "AutoEnableIO":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcAttributeRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{"shape":"AttributeBooleanValue"}, - "EnableDnsHostnames":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcEndpointRequest":{ - "type":"structure", - "required":["VpcEndpointId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointId":{"shape":"String"}, - "ResetPolicy":{"shape":"Boolean"}, - "PolicyDocument":{"shape":"String"}, - "AddRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"AddRouteTableId" - }, - "RemoveRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RemoveRouteTableId" - } - } - }, - "ModifyVpcEndpointResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifyVpcPeeringConnectionOptionsRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcPeeringConnectionId":{"shape":"String"}, - "RequesterPeeringConnectionOptions":{"shape":"PeeringConnectionOptionsRequest"}, - "AccepterPeeringConnectionOptions":{"shape":"PeeringConnectionOptionsRequest"} - } - }, - "ModifyVpcPeeringConnectionOptionsResult":{ - "type":"structure", - "members":{ - "RequesterPeeringConnectionOptions":{ - "shape":"PeeringConnectionOptions", - "locationName":"requesterPeeringConnectionOptions" - }, - "AccepterPeeringConnectionOptions":{ - "shape":"PeeringConnectionOptions", - "locationName":"accepterPeeringConnectionOptions" - } - } - }, - "MonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "MonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "Monitoring":{ - "type":"structure", - "members":{ - "State":{ - "shape":"MonitoringState", - "locationName":"state" - } - } - }, - "MonitoringState":{ - "type":"string", - "enum":[ - "disabled", - "disabling", - "enabled", - "pending" - ] - }, - "MoveAddressToVpcRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "MoveAddressToVpcResult":{ - "type":"structure", - "members":{ - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "Status":{ - "shape":"Status", - "locationName":"status" - } - } - }, - "MoveStatus":{ - "type":"string", - "enum":[ - "movingToVpc", - "restoringToClassic" - ] - }, - "MovingAddressStatus":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "MoveStatus":{ - "shape":"MoveStatus", - "locationName":"moveStatus" - } - } - }, - "MovingAddressStatusSet":{ - "type":"list", - "member":{ - "shape":"MovingAddressStatus", - "locationName":"item" - } - }, - "NatGateway":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "DeleteTime":{ - "shape":"DateTime", - "locationName":"deleteTime" - }, - "NatGatewayAddresses":{ - "shape":"NatGatewayAddressList", - "locationName":"natGatewayAddressSet" - }, - "State":{ - "shape":"NatGatewayState", - "locationName":"state" - }, - "FailureCode":{ - "shape":"String", - "locationName":"failureCode" - }, - "FailureMessage":{ - "shape":"String", - "locationName":"failureMessage" - }, - "ProvisionedBandwidth":{ - "shape":"ProvisionedBandwidth", - "locationName":"provisionedBandwidth" - } - } - }, - "NatGatewayAddress":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "PrivateIp":{ - "shape":"String", - "locationName":"privateIp" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "NatGatewayAddressList":{ - "type":"list", - "member":{ - "shape":"NatGatewayAddress", - "locationName":"item" - } - }, - "NatGatewayList":{ - "type":"list", - "member":{ - "shape":"NatGateway", - "locationName":"item" - } - }, - "NatGatewayState":{ - "type":"string", - "enum":[ - "pending", - "failed", - "available", - "deleting", - "deleted" - ] - }, - "NetworkAcl":{ - "type":"structure", - "members":{ - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"default" - }, - "Entries":{ - "shape":"NetworkAclEntryList", - "locationName":"entrySet" - }, - "Associations":{ - "shape":"NetworkAclAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "NetworkAclAssociation":{ - "type":"structure", - "members":{ - "NetworkAclAssociationId":{ - "shape":"String", - "locationName":"networkAclAssociationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - } - } - }, - "NetworkAclAssociationList":{ - "type":"list", - "member":{ - "shape":"NetworkAclAssociation", - "locationName":"item" - } - }, - "NetworkAclEntry":{ - "type":"structure", - "members":{ - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"icmpTypeCode" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "NetworkAclEntryList":{ - "type":"list", - "member":{ - "shape":"NetworkAclEntry", - "locationName":"item" - } - }, - "NetworkAclList":{ - "type":"list", - "member":{ - "shape":"NetworkAcl", - "locationName":"item" - } - }, - "NetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "RequesterManaged":{ - "shape":"Boolean", - "locationName":"requesterManaged" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - }, - "TagSet":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PrivateIpAddresses":{ - "shape":"NetworkInterfacePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - }, - "InterfaceType":{ - "shape":"NetworkInterfaceType", - "locationName":"interfaceType" - } - } - }, - "NetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "NetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttachmentChanges":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttribute":{ - "type":"string", - "enum":[ - "description", - "groupSet", - "sourceDestCheck", - "attachment" - ] - }, - "NetworkInterfaceIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "NetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"NetworkInterface", - "locationName":"item" - } - }, - "NetworkInterfacePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "NetworkInterfacePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"NetworkInterfacePrivateIpAddress", - "locationName":"item" - } - }, - "NetworkInterfaceStatus":{ - "type":"string", - "enum":[ - "available", - "attaching", - "in-use", - "detaching" - ] - }, - "NetworkInterfaceType":{ - "type":"string", - "enum":[ - "interface", - "natGateway" - ] - }, - "NewDhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "NewDhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"NewDhcpConfiguration", - "locationName":"item" - } - }, - "NextToken":{ - "type":"string", - "max":1024, - "min":1 - }, - "OccurrenceDayRequestSet":{ - "type":"list", - "member":{ - "shape":"Integer", - "locationName":"OccurenceDay" - } - }, - "OccurrenceDaySet":{ - "type":"list", - "member":{ - "shape":"Integer", - "locationName":"item" - } - }, - "OfferingTypeValues":{ - "type":"string", - "enum":[ - "Heavy Utilization", - "Medium Utilization", - "Light Utilization", - "No Upfront", - "Partial Upfront", - "All Upfront" - ] - }, - "OperationType":{ - "type":"string", - "enum":[ - "add", - "remove" - ] - }, - "OwnerStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"Owner" - } - }, - "PaymentOption":{ - "type":"string", - "enum":[ - "AllUpfront", - "PartialUpfront", - "NoUpfront" - ] - }, - "PeeringConnectionOptions":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalClassicLinkToRemoteVpc" - }, - "AllowEgressFromLocalVpcToRemoteClassicLink":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalVpcToRemoteClassicLink" - }, - "AllowDnsResolutionFromRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowDnsResolutionFromRemoteVpc" - } - } - }, - "PeeringConnectionOptionsRequest":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{"shape":"Boolean"}, - "AllowEgressFromLocalVpcToRemoteClassicLink":{"shape":"Boolean"}, - "AllowDnsResolutionFromRemoteVpc":{"shape":"Boolean"} - } - }, - "PermissionGroup":{ - "type":"string", - "enum":["all"] - }, - "Placement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Tenancy":{ - "shape":"Tenancy", - "locationName":"tenancy" - }, - "HostId":{ - "shape":"String", - "locationName":"hostId" - }, - "Affinity":{ - "shape":"String", - "locationName":"affinity" - } - } - }, - "PlacementGroup":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - }, - "State":{ - "shape":"PlacementGroupState", - "locationName":"state" - } - } - }, - "PlacementGroupList":{ - "type":"list", - "member":{ - "shape":"PlacementGroup", - "locationName":"item" - } - }, - "PlacementGroupState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "PlacementGroupStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PlacementStrategy":{ - "type":"string", - "enum":["cluster"] - }, - "PlatformValues":{ - "type":"string", - "enum":["Windows"] - }, - "PortRange":{ - "type":"structure", - "members":{ - "From":{ - "shape":"Integer", - "locationName":"from" - }, - "To":{ - "shape":"Integer", - "locationName":"to" - } - } - }, - "PrefixList":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - }, - "PrefixListName":{ - "shape":"String", - "locationName":"prefixListName" - }, - "Cidrs":{ - "shape":"ValueStringList", - "locationName":"cidrSet" - } - } - }, - "PrefixListId":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - } - } - }, - "PrefixListIdList":{ - "type":"list", - "member":{ - "shape":"PrefixListId", - "locationName":"item" - } - }, - "PrefixListIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "PrefixListSet":{ - "type":"list", - "member":{ - "shape":"PrefixList", - "locationName":"item" - } - }, - "PriceSchedule":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Active":{ - "shape":"Boolean", - "locationName":"active" - } - } - }, - "PriceScheduleList":{ - "type":"list", - "member":{ - "shape":"PriceSchedule", - "locationName":"item" - } - }, - "PriceScheduleSpecification":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "PriceScheduleSpecificationList":{ - "type":"list", - "member":{ - "shape":"PriceScheduleSpecification", - "locationName":"item" - } - }, - "PricingDetail":{ - "type":"structure", - "members":{ - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "Count":{ - "shape":"Integer", - "locationName":"count" - } - } - }, - "PricingDetailsList":{ - "type":"list", - "member":{ - "shape":"PricingDetail", - "locationName":"item" - } - }, - "PrivateIpAddressConfigSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesPrivateIpAddressConfig", - "locationName":"PrivateIpAddressConfigSet" - } - }, - "PrivateIpAddressSpecification":{ - "type":"structure", - "required":["PrivateIpAddress"], - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - } - } - }, - "PrivateIpAddressSpecificationList":{ - "type":"list", - "member":{ - "shape":"PrivateIpAddressSpecification", - "locationName":"item" - } - }, - "PrivateIpAddressStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PrivateIpAddress" - } - }, - "ProductCode":{ - "type":"structure", - "members":{ - "ProductCodeId":{ - "shape":"String", - "locationName":"productCode" - }, - "ProductCodeType":{ - "shape":"ProductCodeValues", - "locationName":"type" - } - } - }, - "ProductCodeList":{ - "type":"list", - "member":{ - "shape":"ProductCode", - "locationName":"item" - } - }, - "ProductCodeStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ProductCode" - } - }, - "ProductCodeValues":{ - "type":"string", - "enum":[ - "devpay", - "marketplace" - ] - }, - "ProductDescriptionList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PropagatingVgw":{ - "type":"structure", - "members":{ - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - } - } - }, - "PropagatingVgwList":{ - "type":"list", - "member":{ - "shape":"PropagatingVgw", - "locationName":"item" - } - }, - "ProvisionedBandwidth":{ - "type":"structure", - "members":{ - "Provisioned":{ - "shape":"String", - "locationName":"provisioned" - }, - "Requested":{ - "shape":"String", - "locationName":"requested" - }, - "RequestTime":{ - "shape":"DateTime", - "locationName":"requestTime" - }, - "ProvisionTime":{ - "shape":"DateTime", - "locationName":"provisionTime" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "PublicIpStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PublicIp" - } - }, - "Purchase":{ - "type":"structure", - "members":{ - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "HostIdSet":{ - "shape":"ResponseHostIdSet", - "locationName":"hostIdSet" - }, - "InstanceFamily":{ - "shape":"String", - "locationName":"instanceFamily" - }, - "PaymentOption":{ - "shape":"PaymentOption", - "locationName":"paymentOption" - }, - "UpfrontPrice":{ - "shape":"String", - "locationName":"upfrontPrice" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Duration":{ - "shape":"Integer", - "locationName":"duration" - } - } - }, - "PurchaseHostReservationRequest":{ - "type":"structure", - "required":[ - "OfferingId", - "HostIdSet" - ], - "members":{ - "OfferingId":{"shape":"String"}, - "HostIdSet":{"shape":"RequestHostIdSet"}, - "LimitPrice":{"shape":"String"}, - "CurrencyCode":{"shape":"CurrencyCodeValues"}, - "ClientToken":{"shape":"String"} - } - }, - "PurchaseHostReservationResult":{ - "type":"structure", - "members":{ - "Purchase":{ - "shape":"PurchaseSet", - "locationName":"purchase" - }, - "TotalUpfrontPrice":{ - "shape":"String", - "locationName":"totalUpfrontPrice" - }, - "TotalHourlyPrice":{ - "shape":"String", - "locationName":"totalHourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "PurchaseRequest":{ - "type":"structure", - "required":[ - "PurchaseToken", - "InstanceCount" - ], - "members":{ - "PurchaseToken":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"} - } - }, - "PurchaseRequestSet":{ - "type":"list", - "member":{ - "shape":"PurchaseRequest", - "locationName":"PurchaseRequest" - }, - "min":1 - }, - "PurchaseReservedInstancesOfferingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesOfferingId", - "InstanceCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingId":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"}, - "LimitPrice":{ - "shape":"ReservedInstanceLimitPrice", - "locationName":"limitPrice" - } - } - }, - "PurchaseReservedInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "PurchaseScheduledInstancesRequest":{ - "type":"structure", - "required":["PurchaseRequests"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ClientToken":{ - "shape":"String", - "idempotencyToken":true - }, - "PurchaseRequests":{ - "shape":"PurchaseRequestSet", - "locationName":"PurchaseRequest" - } - } - }, - "PurchaseScheduledInstancesResult":{ - "type":"structure", - "members":{ - "ScheduledInstanceSet":{ - "shape":"PurchasedScheduledInstanceSet", - "locationName":"scheduledInstanceSet" - } - } - }, - "PurchaseSet":{ - "type":"list", - "member":{"shape":"Purchase"} - }, - "PurchasedScheduledInstanceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstance", - "locationName":"item" - } - }, - "RIProductDescription":{ - "type":"string", - "enum":[ - "Linux/UNIX", - "Linux/UNIX (Amazon VPC)", - "Windows", - "Windows (Amazon VPC)" - ] - }, - "ReasonCodesList":{ - "type":"list", - "member":{ - "shape":"ReportInstanceReasonCodes", - "locationName":"item" - } - }, - "RebootInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"RecurringChargeFrequency", - "locationName":"frequency" - }, - "Amount":{ - "shape":"Double", - "locationName":"amount" - } - } - }, - "RecurringChargeFrequency":{ - "type":"string", - "enum":["Hourly"] - }, - "RecurringChargesList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"item" - } - }, - "Region":{ - "type":"structure", - "members":{ - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Endpoint":{ - "shape":"String", - "locationName":"regionEndpoint" - } - } - }, - "RegionList":{ - "type":"list", - "member":{ - "shape":"Region", - "locationName":"item" - } - }, - "RegionNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"RegionName" - } - }, - "RegisterImageRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageLocation":{"shape":"String"}, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"String", - "locationName":"virtualizationType" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"Boolean", - "locationName":"enaSupport" - } - } - }, - "RegisterImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "RejectVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "RejectVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ReleaseAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"} - } - }, - "ReleaseHostsRequest":{ - "type":"structure", - "required":["HostIds"], - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - } - } - }, - "ReleaseHostsResult":{ - "type":"structure", - "members":{ - "Successful":{ - "shape":"ResponseHostIdList", - "locationName":"successful" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemList", - "locationName":"unsuccessful" - } - } - }, - "ReplaceNetworkAclAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "NetworkAclId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "ReplaceNetworkAclAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReplaceNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "ReplaceRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "ReplaceRouteTableAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "ReplaceRouteTableAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReportInstanceReasonCodes":{ - "type":"string", - "enum":[ - "instance-stuck-in-state", - "unresponsive", - "not-accepting-credentials", - "password-not-available", - "performance-network", - "performance-instance-store", - "performance-ebs-volume", - "performance-other", - "other" - ] - }, - "ReportInstanceStatusRequest":{ - "type":"structure", - "required":[ - "Instances", - "Status", - "ReasonCodes" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Instances":{ - "shape":"InstanceIdStringList", - "locationName":"instanceId" - }, - "Status":{ - "shape":"ReportStatusType", - "locationName":"status" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "ReasonCodes":{ - "shape":"ReasonCodesList", - "locationName":"reasonCode" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ReportStatusType":{ - "type":"string", - "enum":[ - "ok", - "impaired" - ] - }, - "RequestHostIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RequestHostIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RequestSpotFleetRequest":{ - "type":"structure", - "required":["SpotFleetRequestConfig"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - } - } - }, - "RequestSpotFleetResponse":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - } - } - }, - "RequestSpotInstancesRequest":{ - "type":"structure", - "required":["SpotPrice"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "BlockDurationMinutes":{ - "shape":"Integer", - "locationName":"blockDurationMinutes" - }, - "LaunchSpecification":{"shape":"RequestSpotLaunchSpecification"} - } - }, - "RequestSpotInstancesResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "RequestSpotLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"ValueStringList", - "locationName":"SecurityGroup" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - }, - "SecurityGroupIds":{ - "shape":"ValueStringList", - "locationName":"SecurityGroupId" - } - } - }, - "Reservation":{ - "type":"structure", - "members":{ - "ReservationId":{ - "shape":"String", - "locationName":"reservationId" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Instances":{ - "shape":"InstanceList", - "locationName":"instancesSet" - } - } - }, - "ReservationList":{ - "type":"list", - "member":{ - "shape":"Reservation", - "locationName":"item" - } - }, - "ReservationState":{ - "type":"string", - "enum":[ - "payment-pending", - "payment-failed", - "active", - "retired" - ] - }, - "ReservedInstanceLimitPrice":{ - "type":"structure", - "members":{ - "Amount":{ - "shape":"Double", - "locationName":"amount" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "ReservedInstanceState":{ - "type":"string", - "enum":[ - "payment-pending", - "active", - "payment-failed", - "retired" - ] - }, - "ReservedInstances":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Start":{ - "shape":"DateTime", - "locationName":"start" - }, - "End":{ - "shape":"DateTime", - "locationName":"end" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "State":{ - "shape":"ReservedInstanceState", - "locationName":"state" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - } - } - }, - "ReservedInstancesConfiguration":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - } - } - }, - "ReservedInstancesConfigurationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"item" - } - }, - "ReservedInstancesId":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "ReservedInstancesIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesId" - } - }, - "ReservedInstancesList":{ - "type":"list", - "member":{ - "shape":"ReservedInstances", - "locationName":"item" - } - }, - "ReservedInstancesListing":{ - "type":"structure", - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "Status":{ - "shape":"ListingStatus", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceCounts":{ - "shape":"InstanceCountList", - "locationName":"instanceCounts" - }, - "PriceSchedules":{ - "shape":"PriceScheduleList", - "locationName":"priceSchedules" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesListingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesListing", - "locationName":"item" - } - }, - "ReservedInstancesModification":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - }, - "ReservedInstancesIds":{ - "shape":"ReservedIntancesIds", - "locationName":"reservedInstancesSet" - }, - "ModificationResults":{ - "shape":"ReservedInstancesModificationResultList", - "locationName":"modificationResultSet" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "EffectiveDate":{ - "shape":"DateTime", - "locationName":"effectiveDate" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesModificationIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesModificationId" - } - }, - "ReservedInstancesModificationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModification", - "locationName":"item" - } - }, - "ReservedInstancesModificationResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "TargetConfiguration":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"targetConfiguration" - } - } - }, - "ReservedInstancesModificationResultList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModificationResult", - "locationName":"item" - } - }, - "ReservedInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferingId":{ - "shape":"String", - "locationName":"reservedInstancesOfferingId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - }, - "Marketplace":{ - "shape":"Boolean", - "locationName":"marketplace" - }, - "PricingDetails":{ - "shape":"PricingDetailsList", - "locationName":"pricingDetailsSet" - } - } - }, - "ReservedInstancesOfferingIdStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ReservedInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesOffering", - "locationName":"item" - } - }, - "ReservedIntancesIds":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesId", - "locationName":"item" - } - }, - "ResetImageAttributeName":{ - "type":"string", - "enum":["launchPermission"] - }, - "ResetImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ResetImageAttributeName"} - } - }, - "ResetInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "ResetNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SourceDestCheck":{ - "shape":"String", - "locationName":"sourceDestCheck" - } - } - }, - "ResetSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "ResourceIdList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ResourceType":{ - "type":"string", - "enum":[ - "customer-gateway", - "dhcp-options", - "image", - "instance", - "internet-gateway", - "network-acl", - "network-interface", - "reserved-instances", - "route-table", - "snapshot", - "spot-instances-request", - "subnet", - "security-group", - "volume", - "vpc", - "vpn-connection", - "vpn-gateway" - ] - }, - "ResponseHostIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "ResponseHostIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RestorableByStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "RestoreAddressToClassicRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RestoreAddressToClassicResult":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"Status", - "locationName":"status" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RevokeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "RevokeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "Route":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "DestinationPrefixListId":{ - "shape":"String", - "locationName":"destinationPrefixListId" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - }, - "State":{ - "shape":"RouteState", - "locationName":"state" - }, - "Origin":{ - "shape":"RouteOrigin", - "locationName":"origin" - } - } - }, - "RouteList":{ - "type":"list", - "member":{ - "shape":"Route", - "locationName":"item" - } - }, - "RouteOrigin":{ - "type":"string", - "enum":[ - "CreateRouteTable", - "CreateRoute", - "EnableVgwRoutePropagation" - ] - }, - "RouteState":{ - "type":"string", - "enum":[ - "active", - "blackhole" - ] - }, - "RouteTable":{ - "type":"structure", - "members":{ - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Routes":{ - "shape":"RouteList", - "locationName":"routeSet" - }, - "Associations":{ - "shape":"RouteTableAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PropagatingVgws":{ - "shape":"PropagatingVgwList", - "locationName":"propagatingVgwSet" - } - } - }, - "RouteTableAssociation":{ - "type":"structure", - "members":{ - "RouteTableAssociationId":{ - "shape":"String", - "locationName":"routeTableAssociationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Main":{ - "shape":"Boolean", - "locationName":"main" - } - } - }, - "RouteTableAssociationList":{ - "type":"list", - "member":{ - "shape":"RouteTableAssociation", - "locationName":"item" - } - }, - "RouteTableList":{ - "type":"list", - "member":{ - "shape":"RouteTable", - "locationName":"item" - } - }, - "RuleAction":{ - "type":"string", - "enum":[ - "allow", - "deny" - ] - }, - "RunInstancesMonitoringEnabled":{ - "type":"structure", - "required":["Enabled"], - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "RunInstancesRequest":{ - "type":"structure", - "required":[ - "ImageId", - "MinCount", - "MaxCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "MinCount":{"shape":"Integer"}, - "MaxCount":{"shape":"Integer"}, - "KeyName":{"shape":"String"}, - "SecurityGroups":{ - "shape":"SecurityGroupStringList", - "locationName":"SecurityGroup" - }, - "SecurityGroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "InstanceType":{"shape":"InstanceType"}, - "Placement":{"shape":"Placement"}, - "KernelId":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"RunInstancesMonitoringEnabled"}, - "SubnetId":{"shape":"String"}, - "DisableApiTermination":{ - "shape":"Boolean", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - } - } - }, - "RunScheduledInstancesRequest":{ - "type":"structure", - "required":[ - "ScheduledInstanceId", - "LaunchSpecification" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ClientToken":{ - "shape":"String", - "idempotencyToken":true - }, - "InstanceCount":{"shape":"Integer"}, - "ScheduledInstanceId":{"shape":"String"}, - "LaunchSpecification":{"shape":"ScheduledInstancesLaunchSpecification"} - } - }, - "RunScheduledInstancesResult":{ - "type":"structure", - "members":{ - "InstanceIdSet":{ - "shape":"InstanceIdSet", - "locationName":"instanceIdSet" - } - } - }, - "S3Storage":{ - "type":"structure", - "members":{ - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "AWSAccessKeyId":{"shape":"String"}, - "UploadPolicy":{ - "shape":"Blob", - "locationName":"uploadPolicy" - }, - "UploadPolicySignature":{ - "shape":"String", - "locationName":"uploadPolicySignature" - } - } - }, - "ScheduledInstance":{ - "type":"structure", - "members":{ - "ScheduledInstanceId":{ - "shape":"String", - "locationName":"scheduledInstanceId" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "NetworkPlatform":{ - "shape":"String", - "locationName":"networkPlatform" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "SlotDurationInHours":{ - "shape":"Integer", - "locationName":"slotDurationInHours" - }, - "Recurrence":{ - "shape":"ScheduledInstanceRecurrence", - "locationName":"recurrence" - }, - "PreviousSlotEndTime":{ - "shape":"DateTime", - "locationName":"previousSlotEndTime" - }, - "NextSlotStartTime":{ - "shape":"DateTime", - "locationName":"nextSlotStartTime" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "TotalScheduledInstanceHours":{ - "shape":"Integer", - "locationName":"totalScheduledInstanceHours" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "TermStartDate":{ - "shape":"DateTime", - "locationName":"termStartDate" - }, - "TermEndDate":{ - "shape":"DateTime", - "locationName":"termEndDate" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - } - } - }, - "ScheduledInstanceAvailability":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "NetworkPlatform":{ - "shape":"String", - "locationName":"networkPlatform" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "PurchaseToken":{ - "shape":"String", - "locationName":"purchaseToken" - }, - "SlotDurationInHours":{ - "shape":"Integer", - "locationName":"slotDurationInHours" - }, - "Recurrence":{ - "shape":"ScheduledInstanceRecurrence", - "locationName":"recurrence" - }, - "FirstSlotStartTime":{ - "shape":"DateTime", - "locationName":"firstSlotStartTime" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "TotalScheduledInstanceHours":{ - "shape":"Integer", - "locationName":"totalScheduledInstanceHours" - }, - "AvailableInstanceCount":{ - "shape":"Integer", - "locationName":"availableInstanceCount" - }, - "MinTermDurationInDays":{ - "shape":"Integer", - "locationName":"minTermDurationInDays" - }, - "MaxTermDurationInDays":{ - "shape":"Integer", - "locationName":"maxTermDurationInDays" - } - } - }, - "ScheduledInstanceAvailabilitySet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstanceAvailability", - "locationName":"item" - } - }, - "ScheduledInstanceIdRequestSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ScheduledInstanceId" - } - }, - "ScheduledInstanceRecurrence":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"String", - "locationName":"frequency" - }, - "Interval":{ - "shape":"Integer", - "locationName":"interval" - }, - "OccurrenceDaySet":{ - "shape":"OccurrenceDaySet", - "locationName":"occurrenceDaySet" - }, - "OccurrenceRelativeToEnd":{ - "shape":"Boolean", - "locationName":"occurrenceRelativeToEnd" - }, - "OccurrenceUnit":{ - "shape":"String", - "locationName":"occurrenceUnit" - } - } - }, - "ScheduledInstanceRecurrenceRequest":{ - "type":"structure", - "members":{ - "Frequency":{"shape":"String"}, - "Interval":{"shape":"Integer"}, - "OccurrenceDays":{ - "shape":"OccurrenceDayRequestSet", - "locationName":"OccurrenceDay" - }, - "OccurrenceRelativeToEnd":{"shape":"Boolean"}, - "OccurrenceUnit":{"shape":"String"} - } - }, - "ScheduledInstanceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstance", - "locationName":"item" - } - }, - "ScheduledInstancesBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{"shape":"String"}, - "NoDevice":{"shape":"String"}, - "VirtualName":{"shape":"String"}, - "Ebs":{"shape":"ScheduledInstancesEbs"} - } - }, - "ScheduledInstancesBlockDeviceMappingSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesBlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "ScheduledInstancesEbs":{ - "type":"structure", - "members":{ - "SnapshotId":{"shape":"String"}, - "VolumeSize":{"shape":"Integer"}, - "DeleteOnTermination":{"shape":"Boolean"}, - "VolumeType":{"shape":"String"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{"shape":"Boolean"} - } - }, - "ScheduledInstancesIamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{"shape":"String"}, - "Name":{"shape":"String"} - } - }, - "ScheduledInstancesLaunchSpecification":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "ImageId":{"shape":"String"}, - "KeyName":{"shape":"String"}, - "SecurityGroupIds":{ - "shape":"ScheduledInstancesSecurityGroupIdSet", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "Placement":{"shape":"ScheduledInstancesPlacement"}, - "KernelId":{"shape":"String"}, - "InstanceType":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"ScheduledInstancesBlockDeviceMappingSet", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"ScheduledInstancesMonitoring"}, - "SubnetId":{"shape":"String"}, - "NetworkInterfaces":{ - "shape":"ScheduledInstancesNetworkInterfaceSet", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{"shape":"ScheduledInstancesIamInstanceProfile"}, - "EbsOptimized":{"shape":"Boolean"} - } - }, - "ScheduledInstancesMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{"shape":"Boolean"} - } - }, - "ScheduledInstancesNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{"shape":"String"}, - "DeviceIndex":{"shape":"Integer"}, - "SubnetId":{"shape":"String"}, - "Description":{"shape":"String"}, - "PrivateIpAddress":{"shape":"String"}, - "PrivateIpAddressConfigs":{ - "shape":"PrivateIpAddressConfigSet", - "locationName":"PrivateIpAddressConfig" - }, - "SecondaryPrivateIpAddressCount":{"shape":"Integer"}, - "AssociatePublicIpAddress":{"shape":"Boolean"}, - "Groups":{ - "shape":"ScheduledInstancesSecurityGroupIdSet", - "locationName":"Group" - }, - "DeleteOnTermination":{"shape":"Boolean"} - } - }, - "ScheduledInstancesNetworkInterfaceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesNetworkInterface", - "locationName":"NetworkInterface" - } - }, - "ScheduledInstancesPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{"shape":"String"}, - "GroupName":{"shape":"String"} - } - }, - "ScheduledInstancesPrivateIpAddressConfig":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{"shape":"String"}, - "Primary":{"shape":"Boolean"} - } - }, - "ScheduledInstancesSecurityGroupIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "Description":{ - "shape":"String", - "locationName":"groupDescription" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - }, - "IpPermissionsEgress":{ - "shape":"IpPermissionList", - "locationName":"ipPermissionsEgress" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SecurityGroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"SecurityGroup", - "locationName":"item" - } - }, - "SecurityGroupReference":{ - "type":"structure", - "required":[ - "GroupId", - "ReferencingVpcId" - ], - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "ReferencingVpcId":{ - "shape":"String", - "locationName":"referencingVpcId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "SecurityGroupReferences":{ - "type":"list", - "member":{ - "shape":"SecurityGroupReference", - "locationName":"item" - } - }, - "SecurityGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroup" - } - }, - "ShutdownBehavior":{ - "type":"string", - "enum":[ - "stop", - "terminate" - ] - }, - "SlotDateTimeRangeRequest":{ - "type":"structure", - "required":[ - "EarliestTime", - "LatestTime" - ], - "members":{ - "EarliestTime":{"shape":"DateTime"}, - "LatestTime":{"shape":"DateTime"} - } - }, - "SlotStartTimeRangeRequest":{ - "type":"structure", - "members":{ - "EarliestTime":{"shape":"DateTime"}, - "LatestTime":{"shape":"DateTime"} - } - }, - "Snapshot":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "State":{ - "shape":"SnapshotState", - "locationName":"status" - }, - "StateMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "OwnerAlias":{ - "shape":"String", - "locationName":"ownerAlias" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - }, - "DataEncryptionKeyId":{ - "shape":"String", - "locationName":"dataEncryptionKeyId" - } - } - }, - "SnapshotAttributeName":{ - "type":"string", - "enum":[ - "productCodes", - "createVolumePermission" - ] - }, - "SnapshotDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SnapshotDetailList":{ - "type":"list", - "member":{ - "shape":"SnapshotDetail", - "locationName":"item" - } - }, - "SnapshotDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"} - } - }, - "SnapshotIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SnapshotId" - } - }, - "SnapshotList":{ - "type":"list", - "member":{ - "shape":"Snapshot", - "locationName":"item" - } - }, - "SnapshotState":{ - "type":"string", - "enum":[ - "pending", - "completed", - "error" - ] - }, - "SnapshotTaskDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SpotDatafeedSubscription":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "State":{ - "shape":"DatafeedSubscriptionState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - } - } - }, - "SpotFleetLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "Monitoring":{ - "shape":"SpotFleetMonitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "WeightedCapacity":{ - "shape":"Double", - "locationName":"weightedCapacity" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - } - } - }, - "SpotFleetMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "SpotFleetRequestConfig":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "SpotFleetRequestState", - "SpotFleetRequestConfig", - "CreateTime" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "SpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"spotFleetRequestState" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "ActivityStatus":{ - "shape":"ActivityStatus", - "locationName":"activityStatus" - } - } - }, - "SpotFleetRequestConfigData":{ - "type":"structure", - "required":[ - "SpotPrice", - "TargetCapacity", - "IamFleetRole", - "LaunchSpecifications" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "TerminateInstancesWithExpiration":{ - "shape":"Boolean", - "locationName":"terminateInstancesWithExpiration" - }, - "IamFleetRole":{ - "shape":"String", - "locationName":"iamFleetRole" - }, - "LaunchSpecifications":{ - "shape":"LaunchSpecsList", - "locationName":"launchSpecifications" - }, - "ExcessCapacityTerminationPolicy":{ - "shape":"ExcessCapacityTerminationPolicy", - "locationName":"excessCapacityTerminationPolicy" - }, - "AllocationStrategy":{ - "shape":"AllocationStrategy", - "locationName":"allocationStrategy" - }, - "FulfilledCapacity":{ - "shape":"Double", - "locationName":"fulfilledCapacity" - }, - "Type":{ - "shape":"FleetType", - "locationName":"type" - } - } - }, - "SpotFleetRequestConfigSet":{ - "type":"list", - "member":{ - "shape":"SpotFleetRequestConfig", - "locationName":"item" - } - }, - "SpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "State":{ - "shape":"SpotInstanceState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - }, - "Status":{ - "shape":"SpotInstanceStatus", - "locationName":"status" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "LaunchSpecification":{ - "shape":"LaunchSpecification", - "locationName":"launchSpecification" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "BlockDurationMinutes":{ - "shape":"Integer", - "locationName":"blockDurationMinutes" - }, - "ActualBlockHourlyPrice":{ - "shape":"String", - "locationName":"actualBlockHourlyPrice" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "LaunchedAvailabilityZone":{ - "shape":"String", - "locationName":"launchedAvailabilityZone" - } - } - }, - "SpotInstanceRequestIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SpotInstanceRequestId" - } - }, - "SpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"SpotInstanceRequest", - "locationName":"item" - } - }, - "SpotInstanceState":{ - "type":"string", - "enum":[ - "open", - "active", - "closed", - "cancelled", - "failed" - ] - }, - "SpotInstanceStateFault":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceStatus":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceType":{ - "type":"string", - "enum":[ - "one-time", - "persistent" - ] - }, - "SpotPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "SpotPrice":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "SpotPriceHistoryList":{ - "type":"list", - "member":{ - "shape":"SpotPrice", - "locationName":"item" - } - }, - "StaleIpPermission":{ - "type":"structure", - "members":{ - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "IpRanges":{ - "shape":"IpRanges", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdSet", - "locationName":"prefixListIds" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairSet", - "locationName":"groups" - } - } - }, - "StaleIpPermissionSet":{ - "type":"list", - "member":{ - "shape":"StaleIpPermission", - "locationName":"item" - } - }, - "StaleSecurityGroup":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "StaleIpPermissions":{ - "shape":"StaleIpPermissionSet", - "locationName":"staleIpPermissions" - }, - "StaleIpPermissionsEgress":{ - "shape":"StaleIpPermissionSet", - "locationName":"staleIpPermissionsEgress" - } - } - }, - "StaleSecurityGroupSet":{ - "type":"list", - "member":{ - "shape":"StaleSecurityGroup", - "locationName":"item" - } - }, - "StartInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "StartInstancesResult":{ - "type":"structure", - "members":{ - "StartingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "State":{ - "type":"string", - "enum":[ - "Pending", - "Available", - "Deleting", - "Deleted" - ] - }, - "StateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "Status":{ - "type":"string", - "enum":[ - "MoveInProgress", - "InVpc", - "InClassic" - ] - }, - "StatusName":{ - "type":"string", - "enum":["reachability"] - }, - "StatusType":{ - "type":"string", - "enum":[ - "passed", - "failed", - "insufficient-data", - "initializing" - ] - }, - "StopInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "StopInstancesResult":{ - "type":"structure", - "members":{ - "StoppingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "Storage":{ - "type":"structure", - "members":{ - "S3":{"shape":"S3Storage"} - } - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "State":{ - "shape":"SubnetState", - "locationName":"state" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "AvailableIpAddressCount":{ - "shape":"Integer", - "locationName":"availableIpAddressCount" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "DefaultForAz":{ - "shape":"Boolean", - "locationName":"defaultForAz" - }, - "MapPublicIpOnLaunch":{ - "shape":"Boolean", - "locationName":"mapPublicIpOnLaunch" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SubnetIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetId" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"item" - } - }, - "SubnetState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "SummaryStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data", - "not-applicable", - "initializing" - ] - }, - "Tag":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescription":{ - "type":"structure", - "members":{ - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "ResourceType":{ - "shape":"ResourceType", - "locationName":"resourceType" - }, - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescriptionList":{ - "type":"list", - "member":{ - "shape":"TagDescription", - "locationName":"item" - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"item" - } - }, - "TelemetryStatus":{ - "type":"string", - "enum":[ - "UP", - "DOWN" - ] - }, - "Tenancy":{ - "type":"string", - "enum":[ - "default", - "dedicated", - "host" - ] - }, - "TerminateInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "TerminateInstancesResult":{ - "type":"structure", - "members":{ - "TerminatingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "TrafficType":{ - "type":"string", - "enum":[ - "ACCEPT", - "REJECT", - "ALL" - ] - }, - "UnassignPrivateIpAddressesRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "PrivateIpAddresses" - ], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - } - } - }, - "UnmonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "UnmonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "UnsuccessfulItem":{ - "type":"structure", - "required":["Error"], - "members":{ - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "Error":{ - "shape":"UnsuccessfulItemError", - "locationName":"error" - } - } - }, - "UnsuccessfulItemError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "UnsuccessfulItemList":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UnsuccessfulItemSet":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UserBucket":{ - "type":"structure", - "members":{ - "S3Bucket":{"shape":"String"}, - "S3Key":{"shape":"String"} - } - }, - "UserBucketDetails":{ - "type":"structure", - "members":{ - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "UserData":{ - "type":"structure", - "members":{ - "Data":{ - "shape":"String", - "locationName":"data" - } - } - }, - "UserGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserGroup" - } - }, - "UserIdGroupPair":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "PeeringStatus":{ - "shape":"String", - "locationName":"peeringStatus" - } - } - }, - "UserIdGroupPairList":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdGroupPairSet":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserId" - } - }, - "ValueStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "VgwTelemetry":{ - "type":"structure", - "members":{ - "OutsideIpAddress":{ - "shape":"String", - "locationName":"outsideIpAddress" - }, - "Status":{ - "shape":"TelemetryStatus", - "locationName":"status" - }, - "LastStatusChange":{ - "shape":"DateTime", - "locationName":"lastStatusChange" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "AcceptedRouteCount":{ - "shape":"Integer", - "locationName":"acceptedRouteCount" - } - } - }, - "VgwTelemetryList":{ - "type":"list", - "member":{ - "shape":"VgwTelemetry", - "locationName":"item" - } - }, - "VirtualizationType":{ - "type":"string", - "enum":[ - "hvm", - "paravirtual" - ] - }, - "Volume":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Size":{ - "shape":"Integer", - "locationName":"size" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "State":{ - "shape":"VolumeState", - "locationName":"status" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "Attachments":{ - "shape":"VolumeAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "VolumeAttachment":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Device":{ - "shape":"String", - "locationName":"device" - }, - "State":{ - "shape":"VolumeAttachmentState", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "VolumeAttachmentList":{ - "type":"list", - "member":{ - "shape":"VolumeAttachment", - "locationName":"item" - } - }, - "VolumeAttachmentState":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "VolumeAttributeName":{ - "type":"string", - "enum":[ - "autoEnableIO", - "productCodes" - ] - }, - "VolumeDetail":{ - "type":"structure", - "required":["Size"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - } - } - }, - "VolumeIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VolumeId" - } - }, - "VolumeList":{ - "type":"list", - "member":{ - "shape":"Volume", - "locationName":"item" - } - }, - "VolumeState":{ - "type":"string", - "enum":[ - "creating", - "available", - "in-use", - "deleting", - "deleted", - "error" - ] - }, - "VolumeStatusAction":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusActionsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusAction", - "locationName":"item" - } - }, - "VolumeStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"VolumeStatusName", - "locationName":"name" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "VolumeStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusDetails", - "locationName":"item" - } - }, - "VolumeStatusEvent":{ - "type":"structure", - "members":{ - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusEventsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusEvent", - "locationName":"item" - } - }, - "VolumeStatusInfo":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"VolumeStatusInfoStatus", - "locationName":"status" - }, - "Details":{ - "shape":"VolumeStatusDetailsList", - "locationName":"details" - } - } - }, - "VolumeStatusInfoStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data" - ] - }, - "VolumeStatusItem":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VolumeStatus":{ - "shape":"VolumeStatusInfo", - "locationName":"volumeStatus" - }, - "Events":{ - "shape":"VolumeStatusEventsList", - "locationName":"eventsSet" - }, - "Actions":{ - "shape":"VolumeStatusActionsList", - "locationName":"actionsSet" - } - } - }, - "VolumeStatusList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusItem", - "locationName":"item" - } - }, - "VolumeStatusName":{ - "type":"string", - "enum":[ - "io-enabled", - "io-performance" - ] - }, - "VolumeType":{ - "type":"string", - "enum":[ - "standard", - "io1", - "gp2", - "sc1", - "st1" - ] - }, - "Vpc":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"VpcState", - "locationName":"state" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"isDefault" - } - } - }, - "VpcAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "VpcAttachmentList":{ - "type":"list", - "member":{ - "shape":"VpcAttachment", - "locationName":"item" - } - }, - "VpcAttributeName":{ - "type":"string", - "enum":[ - "enableDnsSupport", - "enableDnsHostnames" - ] - }, - "VpcClassicLink":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkEnabled":{ - "shape":"Boolean", - "locationName":"classicLinkEnabled" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpcClassicLinkIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcClassicLinkList":{ - "type":"list", - "member":{ - "shape":"VpcClassicLink", - "locationName":"item" - } - }, - "VpcEndpoint":{ - "type":"structure", - "members":{ - "VpcEndpointId":{ - "shape":"String", - "locationName":"vpcEndpointId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ServiceName":{ - "shape":"String", - "locationName":"serviceName" - }, - "State":{ - "shape":"State", - "locationName":"state" - }, - "PolicyDocument":{ - "shape":"String", - "locationName":"policyDocument" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"routeTableIdSet" - }, - "CreationTimestamp":{ - "shape":"DateTime", - "locationName":"creationTimestamp" - } - } - }, - "VpcEndpointSet":{ - "type":"list", - "member":{ - "shape":"VpcEndpoint", - "locationName":"item" - } - }, - "VpcIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcList":{ - "type":"list", - "member":{ - "shape":"Vpc", - "locationName":"item" - } - }, - "VpcPeeringConnection":{ - "type":"structure", - "members":{ - "AccepterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"accepterVpcInfo" - }, - "ExpirationTime":{ - "shape":"DateTime", - "locationName":"expirationTime" - }, - "RequesterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"requesterVpcInfo" - }, - "Status":{ - "shape":"VpcPeeringConnectionStateReason", - "locationName":"status" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "VpcPeeringConnectionList":{ - "type":"list", - "member":{ - "shape":"VpcPeeringConnection", - "locationName":"item" - } - }, - "VpcPeeringConnectionOptionsDescription":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalClassicLinkToRemoteVpc" - }, - "AllowEgressFromLocalVpcToRemoteClassicLink":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalVpcToRemoteClassicLink" - }, - "AllowDnsResolutionFromRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowDnsResolutionFromRemoteVpc" - } - } - }, - "VpcPeeringConnectionStateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"VpcPeeringConnectionStateReasonCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "VpcPeeringConnectionStateReasonCode":{ - "type":"string", - "enum":[ - "initiating-request", - "pending-acceptance", - "active", - "deleted", - "rejected", - "failed", - "expired", - "provisioning", - "deleting" - ] - }, - "VpcPeeringConnectionVpcInfo":{ - "type":"structure", - "members":{ - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeeringOptions":{ - "shape":"VpcPeeringConnectionOptionsDescription", - "locationName":"peeringOptions" - } - } - }, - "VpcState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "VpnConnection":{ - "type":"structure", - "members":{ - "VpnConnectionId":{ - "shape":"String", - "locationName":"vpnConnectionId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "CustomerGatewayConfiguration":{ - "shape":"String", - "locationName":"customerGatewayConfiguration" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VgwTelemetry":{ - "shape":"VgwTelemetryList", - "locationName":"vgwTelemetry" - }, - "Options":{ - "shape":"VpnConnectionOptions", - "locationName":"options" - }, - "Routes":{ - "shape":"VpnStaticRouteList", - "locationName":"routes" - } - } - }, - "VpnConnectionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnConnectionId" - } - }, - "VpnConnectionList":{ - "type":"list", - "member":{ - "shape":"VpnConnection", - "locationName":"item" - } - }, - "VpnConnectionOptions":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnConnectionOptionsSpecification":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnGateway":{ - "type":"structure", - "members":{ - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VpcAttachments":{ - "shape":"VpcAttachmentList", - "locationName":"attachments" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpnGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnGatewayId" - } - }, - "VpnGatewayList":{ - "type":"list", - "member":{ - "shape":"VpnGateway", - "locationName":"item" - } - }, - "VpnState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "VpnStaticRoute":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "Source":{ - "shape":"VpnStaticRouteSource", - "locationName":"source" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - } - } - }, - "VpnStaticRouteList":{ - "type":"list", - "member":{ - "shape":"VpnStaticRoute", - "locationName":"item" - } - }, - "VpnStaticRouteSource":{ - "type":"string", - "enum":["Static"] - }, - "ZoneNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ZoneName" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/docs-2.json deleted file mode 100644 index 86d816270..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/docs-2.json +++ /dev/null @@ -1,6582 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon Elastic Compute Cloud

Amazon Elastic Compute Cloud (Amazon EC2) provides resizable computing capacity in the Amazon Web Services (AWS) cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster.

", - "operations": { - "AcceptVpcPeeringConnection": "

Accept a VPC peering connection request. To accept a request, the VPC peering connection must be in the pending-acceptance state, and you must be the owner of the peer VPC. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests.

", - "AllocateAddress": "

Acquires an Elastic IP address.

An Elastic IP address is for use either in the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "AllocateHosts": "

Allocates a Dedicated Host to your account. At minimum you need to specify the instance size type, Availability Zone, and quantity of hosts you want to allocate.

", - "AssignPrivateIpAddresses": "

Assigns one or more secondary private IP addresses to the specified network interface. You can specify one or more specific secondary IP addresses, or you can specify the number of secondary IP addresses to be automatically assigned within the subnet's CIDR block range. The number of secondary IP addresses that you can assign to an instance varies by instance type. For information about instance types, see Instance Types in the Amazon Elastic Compute Cloud User Guide. For more information about Elastic IP addresses, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

AssignPrivateIpAddresses is available only in EC2-VPC.

", - "AssociateAddress": "

Associates an Elastic IP address with an instance or a network interface.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

[EC2-Classic, VPC in an EC2-VPC-only account] If the Elastic IP address is already associated with a different instance, it is disassociated from that instance and associated with the specified instance.

[VPC in an EC2-Classic account] If you don't specify a private IP address, the Elastic IP address is associated with the primary IP address. If the Elastic IP address is already associated with a different instance or a network interface, you get an error unless you allow reassociation.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error, and you may be charged for each time the Elastic IP address is remapped to the same instance. For more information, see the Elastic IP Addresses section of Amazon EC2 Pricing.

", - "AssociateDhcpOptions": "

Associates a set of DHCP options (that you've previously created) with the specified VPC, or associates no DHCP options with the VPC.

After you associate the options with the VPC, any existing instances and all new instances that you launch in that VPC use the options. You don't need to restart or relaunch the instances. They automatically pick up the changes within a few hours, depending on how frequently the instance renews its DHCP lease. You can explicitly renew the lease using the operating system on the instance.

For more information, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "AssociateRouteTable": "

Associates a subnet with a route table. The subnet and route table must be in the same VPC. This association causes traffic originating from the subnet to be routed according to the routes in the route table. The action returns an association ID, which you need in order to disassociate the route table from the subnet later. A route table can be associated with multiple subnets.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "AttachClassicLinkVpc": "

Links an EC2-Classic instance to a ClassicLink-enabled VPC through one or more of the VPC's security groups. You cannot link an EC2-Classic instance to more than one VPC at a time. You can only link an instance that's in the running state. An instance is automatically unlinked from a VPC when it's stopped - you can link it to the VPC again when you restart it.

After you've linked an instance, you cannot change the VPC security groups that are associated with it. To change the security groups, you must first unlink the instance, and then link it again.

Linking your instance to a VPC is sometimes referred to as attaching your instance.

", - "AttachInternetGateway": "

Attaches an Internet gateway to a VPC, enabling connectivity between the Internet and the VPC. For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "AttachNetworkInterface": "

Attaches a network interface to an instance.

", - "AttachVolume": "

Attaches an EBS volume to a running or stopped instance and exposes it to the instance with the specified device name.

Encrypted EBS volumes may only be attached to instances that support Amazon EBS encryption. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For a list of supported device names, see Attaching an EBS Volume to an Instance. Any device names that aren't reserved for instance store volumes can be used for EBS volumes. For more information, see Amazon EC2 Instance Store in the Amazon Elastic Compute Cloud User Guide.

If a volume has an AWS Marketplace product code:

  • The volume can be attached only to a stopped instance.

  • AWS Marketplace product codes are copied from the volume to the instance.

  • You must be subscribed to the product.

  • The instance type and operating system of the instance must support the product. For example, you can't detach a volume from a Windows instance and attach it to a Linux instance.

For an overview of the AWS Marketplace, see Introducing AWS Marketplace.

For more information about EBS volumes, see Attaching Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "AttachVpnGateway": "

Attaches a virtual private gateway to a VPC. For more information, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "AuthorizeSecurityGroupEgress": "

[EC2-VPC only] Adds one or more egress rules to a security group for use with a VPC. Specifically, this action permits instances to send traffic to one or more destination CIDR IP address ranges, or to one or more destination security groups for the same VPC. This action doesn't apply to security groups for use in EC2-Classic. For more information, see Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

You can have up to 50 rules per security group (covering both ingress and egress rules).

Each rule consists of the protocol (for example, TCP), plus either a CIDR range or a source group. For the TCP and UDP protocols, you must also specify the destination port or port range. For the ICMP protocol, you must also specify the ICMP type and code. You can use -1 for the type or code to mean all types or all codes.

Rule changes are propagated to affected instances as quickly as possible. However, a small delay might occur.

", - "AuthorizeSecurityGroupIngress": "

Adds one or more ingress rules to a security group.

EC2-Classic: You can have up to 100 rules per group.

EC2-VPC: You can have up to 50 rules per group (covering both ingress and egress rules).

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

[EC2-Classic] This action gives one or more CIDR IP address ranges permission to access a security group in your account, or gives one or more security groups (called the source groups) permission to access a security group for your account. A source group can be for your own AWS account, or another.

[EC2-VPC] This action gives one or more CIDR IP address ranges permission to access a security group in your VPC, or gives one or more other security groups (called the source groups) permission to access a security group for your VPC. The security groups must all be for the same VPC.

", - "BundleInstance": "

Bundles an Amazon instance store-backed Windows instance.

During bundling, only the root device volume (C:\\) is bundled. Data on other instance store volumes is not preserved.

This action is not applicable for Linux/Unix instances or Windows instances that are backed by Amazon EBS.

For more information, see Creating an Instance Store-Backed Windows AMI.

", - "CancelBundleTask": "

Cancels a bundling operation for an instance store-backed Windows instance.

", - "CancelConversionTask": "

Cancels an active conversion task. The task can be the import of an instance or volume. The action removes all artifacts of the conversion, including a partially uploaded volume or instance. If the conversion is complete or is in the process of transferring the final disk image, the command fails and returns an exception.

For more information, see Importing a Virtual Machine Using the Amazon EC2 CLI.

", - "CancelExportTask": "

Cancels an active export task. The request removes all artifacts of the export, including any partially-created Amazon S3 objects. If the export task is complete or is in the process of transferring the final disk image, the command fails and returns an error.

", - "CancelImportTask": "

Cancels an in-process import virtual machine or import snapshot task.

", - "CancelReservedInstancesListing": "

Cancels the specified Reserved Instance listing in the Reserved Instance Marketplace.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CancelSpotFleetRequests": "

Cancels the specified Spot fleet requests.

After you cancel a Spot fleet request, the Spot fleet launches no new Spot instances. You must specify whether the Spot fleet should also terminate its Spot instances. If you terminate the instances, the Spot fleet request enters the cancelled_terminating state. Otherwise, the Spot fleet request enters the cancelled_running state and the instances continue to run until they are interrupted or you terminate them manually.

", - "CancelSpotInstanceRequests": "

Cancels one or more Spot instance requests. Spot instances are instances that Amazon EC2 starts on your behalf when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

Canceling a Spot instance request does not terminate running Spot instances associated with the request.

", - "ConfirmProductInstance": "

Determines whether a product code is associated with an instance. This action can only be used by the owner of the product code. It is useful when a product code owner needs to verify whether another user's instance is eligible for support.

", - "CopyImage": "

Initiates the copy of an AMI from the specified source region to the current region. You specify the destination region by using its endpoint when making the request.

For more information, see Copying AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshot": "

Copies a point-in-time snapshot of an EBS volume and stores it in Amazon S3. You can copy the snapshot within the same region or from one region to another. You can use the snapshot to create EBS volumes or Amazon Machine Images (AMIs). The snapshot is copied to the regional endpoint that you send the HTTP request to.

Copies of encrypted EBS snapshots remain encrypted. Copies of unencrypted snapshots remain unencrypted, unless the Encrypted flag is specified during the snapshot copy operation. By default, encrypted snapshot copies use the default AWS Key Management Service (AWS KMS) customer master key (CMK); however, you can specify a non-default CMK with the KmsKeyId parameter.

To copy an encrypted snapshot that has been shared from another account, you must have permissions for the CMK used to encrypt the snapshot.

Snapshots created by the CopySnapshot action have an arbitrary volume ID that should not be used for any purpose.

For more information, see Copying an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGateway": "

Provides information to AWS about your VPN customer gateway device. The customer gateway is the appliance at your end of the VPN connection. (The device on the AWS side of the VPN connection is the virtual private gateway.) You must provide the Internet-routable IP address of the customer gateway's external interface. The IP address must be static and may be behind a device performing network address translation (NAT).

For devices that use Border Gateway Protocol (BGP), you can also provide the device's BGP Autonomous System Number (ASN). You can use an existing ASN assigned to your network. If you don't have an ASN already, you can use a private ASN (in the 64512 - 65534 range).

Amazon EC2 supports all 2-byte ASN numbers in the range of 1 - 65534, with the exception of 7224, which is reserved in the us-east-1 region, and 9059, which is reserved in the eu-west-1 region.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

You cannot create more than one customer gateway with the same VPN type, IP address, and BGP ASN parameter values. If you run an identical request more than one time, the first request creates the customer gateway, and subsequent requests return information about the existing customer gateway. The subsequent requests do not create new customer gateway resources.

", - "CreateDhcpOptions": "

Creates a set of DHCP options for your VPC. After creating the set, you must associate it with the VPC, causing all existing and new instances that you launch in the VPC to use this set of DHCP options. The following are the individual DHCP options you can specify. For more information about the options, see RFC 2132.

  • domain-name-servers - The IP addresses of up to four domain name servers, or AmazonProvidedDNS. The default DHCP option set specifies AmazonProvidedDNS. If specifying more than one domain name server, specify the IP addresses in a single parameter, separated by commas. If you want your instance to receive a custom DNS hostname as specified in domain-name, you must set domain-name-servers to a custom DNS server.

  • domain-name - If you're using AmazonProvidedDNS in \"us-east-1\", specify \"ec2.internal\". If you're using AmazonProvidedDNS in another region, specify \"region.compute.internal\" (for example, \"ap-northeast-1.compute.internal\"). Otherwise, specify a domain name (for example, \"MyCompany.com\"). This value is used to complete unqualified DNS hostnames. Important: Some Linux operating systems accept multiple domain names separated by spaces. However, Windows and other Linux operating systems treat the value as a single domain, which results in unexpected behavior. If your DHCP options set is associated with a VPC that has instances with multiple operating systems, specify only one domain name.

  • ntp-servers - The IP addresses of up to four Network Time Protocol (NTP) servers.

  • netbios-name-servers - The IP addresses of up to four NetBIOS name servers.

  • netbios-node-type - The NetBIOS node type (1, 2, 4, or 8). We recommend that you specify 2 (broadcast and multicast are not currently supported). For more information about these node types, see RFC 2132.

Your VPC automatically starts out with a set of DHCP options that includes only a DNS server that we provide (AmazonProvidedDNS). If you create a set of options, and if your VPC has an Internet gateway, make sure to set the domain-name-servers option either to AmazonProvidedDNS or to a domain name server of your choice. For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "CreateFlowLogs": "

Creates one or more flow logs to capture IP traffic for a specific network interface, subnet, or VPC. Flow logs are delivered to a specified log group in Amazon CloudWatch Logs. If you specify a VPC or subnet in the request, a log stream is created in CloudWatch Logs for each network interface in the subnet or VPC. Log streams can include information about accepted and rejected traffic to a network interface. You can view the data in your log streams using Amazon CloudWatch Logs.

In your request, you must also specify an IAM role that has permission to publish logs to CloudWatch Logs.

", - "CreateImage": "

Creates an Amazon EBS-backed AMI from an Amazon EBS-backed instance that is either running or stopped.

If you customized your instance with instance store volumes or EBS volumes in addition to the root device volume, the new AMI contains block device mapping information for those volumes. When you launch an instance from this new AMI, the instance automatically launches with those additional volumes.

For more information, see Creating Amazon EBS-Backed Linux AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateInstanceExportTask": "

Exports a running or stopped instance to an S3 bucket.

For information about the supported operating systems, image formats, and known limitations for the types of instances you can export, see Exporting an Instance as a VM Using VM Import/Export in the VM Import/Export User Guide.

", - "CreateInternetGateway": "

Creates an Internet gateway for use with a VPC. After creating the Internet gateway, you attach it to a VPC using AttachInternetGateway.

For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "CreateKeyPair": "

Creates a 2048-bit RSA key pair with the specified name. Amazon EC2 stores the public key and displays the private key for you to save to a file. The private key is returned as an unencrypted PEM encoded PKCS#8 private key. If a key with the specified name already exists, Amazon EC2 returns an error.

You can have up to five thousand key pairs per region.

The key pair returned to you is available only in the region in which you create it. To create a key pair that is available in all regions, use ImportKeyPair.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateNatGateway": "

Creates a NAT gateway in the specified subnet. A NAT gateway can be used to enable instances in a private subnet to connect to the Internet. This action creates a network interface in the specified subnet with a private IP address from the IP address range of the subnet. For more information, see NAT Gateways in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAcl": "

Creates a network ACL in a VPC. Network ACLs provide an optional layer of security (in addition to security groups) for the instances in your VPC.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAclEntry": "

Creates an entry (a rule) in a network ACL with the specified rule number. Each network ACL has a set of numbered ingress rules and a separate set of numbered egress rules. When determining whether a packet should be allowed in or out of a subnet associated with the ACL, we process the entries in the ACL according to the rule numbers, in ascending order. Each network ACL has a set of ingress rules and a separate set of egress rules.

We recommend that you leave room between the rule numbers (for example, 100, 110, 120, ...), and not number them one right after the other (for example, 101, 102, 103, ...). This makes it easier to add a rule between existing ones without having to renumber the rules.

After you add an entry, you can't modify it; you must either replace it, or create an entry and delete the old one.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkInterface": "

Creates a network interface in the specified subnet.

For more information about network interfaces, see Elastic Network Interfaces in the Amazon Elastic Compute Cloud User Guide.

", - "CreatePlacementGroup": "

Creates a placement group that you launch cluster instances into. You must give the group a name that's unique within the scope of your account.

For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListing": "

Creates a listing for Amazon EC2 Reserved Instances to be sold in the Reserved Instance Marketplace. You can submit one Reserved Instance listing at a time. To get a list of your Reserved Instances, you can use the DescribeReservedInstances operation.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

To sell your Reserved Instances, you must first register as a seller in the Reserved Instance Marketplace. After completing the registration process, you can create a Reserved Instance Marketplace listing of some or all of your Reserved Instances, and specify the upfront price to receive for them. Your Reserved Instance listings then become available for purchase. To view the details of your Reserved Instance listing, you can use the DescribeReservedInstancesListings operation.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CreateRoute": "

Creates a route in a route table within a VPC.

You must specify one of the following targets: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, or network interface.

When determining how to route traffic, we use the route with the most specific match. For example, let's say the traffic is destined for 192.0.2.3, and the route table includes the following two routes:

  • 192.0.2.0/24 (goes to some target A)

  • 192.0.2.0/28 (goes to some target B)

Both routes apply to the traffic destined for 192.0.2.3. However, the second route in the list covers a smaller number of IP addresses and is therefore more specific, so we use that route to determine where to target the traffic.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateRouteTable": "

Creates a route table for the specified VPC. After you create a route table, you can add routes and associate the table with a subnet.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateSecurityGroup": "

Creates a security group.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

EC2-Classic: You can have up to 500 security groups.

EC2-VPC: You can create up to 500 security groups per VPC.

When you create a security group, you specify a friendly name of your choice. You can have a security group for use in EC2-Classic with the same name as a security group for use in a VPC. However, you can't have two security groups for use in EC2-Classic with the same name or two security groups for use in a VPC with the same name.

You have a default security group for use in EC2-Classic and a default security group for use in your VPC. If you don't specify a security group when you launch an instance, the instance is launched into the appropriate default security group. A default security group includes a default rule that grants instances unrestricted network access to each other.

You can add or remove rules from your security groups using AuthorizeSecurityGroupIngress, AuthorizeSecurityGroupEgress, RevokeSecurityGroupIngress, and RevokeSecurityGroupEgress.

", - "CreateSnapshot": "

Creates a snapshot of an EBS volume and stores it in Amazon S3. You can use snapshots for backups, to make copies of EBS volumes, and to save data before shutting down an instance.

When a snapshot is created, any AWS Marketplace product codes that are associated with the source volume are propagated to the snapshot.

You can take a snapshot of an attached volume that is in use. However, snapshots only capture data that has been written to your EBS volume at the time the snapshot command is issued; this may exclude any data that has been cached by any applications or the operating system. If you can pause any file systems on the volume long enough to take a snapshot, your snapshot should be complete. However, if you cannot pause all file writes to the volume, you should unmount the volume from within the instance, issue the snapshot command, and then remount the volume to ensure a consistent and complete snapshot. You may remount and use your volume while the snapshot status is pending.

To create a snapshot for EBS volumes that serve as root devices, you should stop the instance before taking the snapshot.

Snapshots that are taken from encrypted volumes are automatically encrypted. Volumes that are created from encrypted snapshots are also automatically encrypted. Your encrypted volumes and any associated snapshots always remain protected.

For more information, see Amazon Elastic Block Store and Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSpotDatafeedSubscription": "

Creates a data feed for Spot instances, enabling you to view Spot instance usage logs. You can create one data feed per AWS account. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSubnet": "

Creates a subnet in an existing VPC.

When you create each subnet, you provide the VPC ID and the CIDR block you want for the subnet. After you create a subnet, you can't change its CIDR block. The subnet's CIDR block can be the same as the VPC's CIDR block (assuming you want only a single subnet in the VPC), or a subset of the VPC's CIDR block. If you create more than one subnet in a VPC, the subnets' CIDR blocks must not overlap. The smallest subnet (and VPC) you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses).

AWS reserves both the first four and the last IP address in each subnet's CIDR block. They're not available for use.

If you add more than one subnet to a VPC, they're set up in a star topology with a logical router in the middle.

If you launch an instance in a VPC using an Amazon EBS-backed AMI, the IP address doesn't change if you stop and restart the instance (unlike a similar instance launched outside a VPC, which gets a new IP address when restarted). It's therefore possible to have a subnet with no running instances (they're all stopped), but no remaining IP addresses available.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "CreateTags": "

Adds or overwrites one or more tags for the specified Amazon EC2 resource or resources. Each resource can have a maximum of 50 tags. Each tag consists of a key and optional value. Tag keys must be unique per resource.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide. For more information about creating IAM policies that control users' access to resources based on tags, see Supported Resource-Level Permissions for Amazon EC2 API Actions in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVolume": "

Creates an EBS volume that can be attached to an instance in the same Availability Zone. The volume is created in the regional endpoint that you send the HTTP request to. For more information see Regions and Endpoints.

You can create a new empty volume or restore a volume from an EBS snapshot. Any AWS Marketplace product codes from the snapshot are propagated to the volume.

You can create encrypted volumes with the Encrypted parameter. Encrypted volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are also automatically encrypted. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For more information, see Creating or Restoring an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpc": "

Creates a VPC with the specified CIDR block.

The smallest VPC you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses). To help you decide how big to make your VPC, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

By default, each instance you launch in the VPC has the default DHCP options, which includes only a default DNS server that we provide (AmazonProvidedDNS). For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

You can specify the instance tenancy value for the VPC when you create it. You can't change this value for the VPC after you create it. For more information, see Dedicated Instances in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpcEndpoint": "

Creates a VPC endpoint for a specified AWS service. An endpoint enables you to create a private connection between your VPC and another AWS service in your account. You can specify an endpoint policy to attach to the endpoint that will control access to the service from your VPC. You can also specify the VPC route tables that use the endpoint.

Currently, only endpoints to Amazon S3 are supported.

", - "CreateVpcPeeringConnection": "

Requests a VPC peering connection between two VPCs: a requester VPC that you own and a peer VPC with which to create the connection. The peer VPC can belong to another AWS account. The requester VPC and peer VPC cannot have overlapping CIDR blocks.

The owner of the peer VPC must accept the peering request to activate the peering connection. The VPC peering connection request expires after 7 days, after which it cannot be accepted or rejected.

A CreateVpcPeeringConnection request between VPCs with overlapping CIDR blocks results in the VPC peering connection having a status of failed.

", - "CreateVpnConnection": "

Creates a VPN connection between an existing virtual private gateway and a VPN customer gateway. The only supported connection type is ipsec.1.

The response includes information that you need to give to your network administrator to configure your customer gateway.

We strongly recommend that you use HTTPS when calling this operation because the response contains sensitive cryptographic information for configuring your customer gateway.

If you decide to shut down your VPN connection for any reason and later create a new VPN connection, you must reconfigure your customer gateway with the new information returned from this call.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnConnectionRoute": "

Creates a static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnGateway": "

Creates a virtual private gateway. A virtual private gateway is the endpoint on the VPC side of your VPN connection. You can create a virtual private gateway before creating the VPC itself.

For more information about virtual private gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DeleteCustomerGateway": "

Deletes the specified customer gateway. You must delete the VPN connection before you can delete the customer gateway.

", - "DeleteDhcpOptions": "

Deletes the specified set of DHCP options. You must disassociate the set of DHCP options before you can delete it. You can disassociate the set of DHCP options by associating either a new set of options or the default set of options with the VPC.

", - "DeleteFlowLogs": "

Deletes one or more flow logs.

", - "DeleteInternetGateway": "

Deletes the specified Internet gateway. You must detach the Internet gateway from the VPC before you can delete it.

", - "DeleteKeyPair": "

Deletes the specified key pair, by removing the public key from Amazon EC2.

", - "DeleteNatGateway": "

Deletes the specified NAT gateway. Deleting a NAT gateway disassociates its Elastic IP address, but does not release the address from your account. Deleting a NAT gateway does not delete any NAT gateway routes in your route tables.

", - "DeleteNetworkAcl": "

Deletes the specified network ACL. You can't delete the ACL if it's associated with any subnets. You can't delete the default network ACL.

", - "DeleteNetworkAclEntry": "

Deletes the specified ingress or egress entry (rule) from the specified network ACL.

", - "DeleteNetworkInterface": "

Deletes the specified network interface. You must detach the network interface before you can delete it.

", - "DeletePlacementGroup": "

Deletes the specified placement group. You must terminate all instances in the placement group before you can delete the placement group. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteRoute": "

Deletes the specified route from the specified route table.

", - "DeleteRouteTable": "

Deletes the specified route table. You must disassociate the route table from any subnets before you can delete it. You can't delete the main route table.

", - "DeleteSecurityGroup": "

Deletes a security group.

If you attempt to delete a security group that is associated with an instance, or is referenced by another security group, the operation fails with InvalidGroup.InUse in EC2-Classic or DependencyViolation in EC2-VPC.

", - "DeleteSnapshot": "

Deletes the specified snapshot.

When you make periodic snapshots of a volume, the snapshots are incremental, and only the blocks on the device that have changed since your last snapshot are saved in the new snapshot. When you delete a snapshot, only the data not needed for any other snapshot is removed. So regardless of which prior snapshots have been deleted, all active snapshots will have access to all the information needed to restore the volume.

You cannot delete a snapshot of the root device of an EBS volume used by a registered AMI. You must first de-register the AMI before you can delete the snapshot.

For more information, see Deleting an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteSpotDatafeedSubscription": "

Deletes the data feed for Spot instances.

", - "DeleteSubnet": "

Deletes the specified subnet. You must terminate all running instances in the subnet before you can delete the subnet.

", - "DeleteTags": "

Deletes the specified set of tags from the specified set of resources. This call is designed to follow a DescribeTags request.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVolume": "

Deletes the specified EBS volume. The volume must be in the available state (not attached to an instance).

The volume may remain in the deleting state for several minutes.

For more information, see Deleting an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVpc": "

Deletes the specified VPC. You must detach or delete all gateways and resources that are associated with the VPC before you can delete it. For example, you must terminate all instances running in the VPC, delete all security groups associated with the VPC (except the default one), delete all route tables associated with the VPC (except the default one), and so on.

", - "DeleteVpcEndpoints": "

Deletes one or more specified VPC endpoints. Deleting the endpoint also deletes the endpoint routes in the route tables that were associated with the endpoint.

", - "DeleteVpcPeeringConnection": "

Deletes a VPC peering connection. Either the owner of the requester VPC or the owner of the peer VPC can delete the VPC peering connection if it's in the active state. The owner of the requester VPC can delete a VPC peering connection in the pending-acceptance state.

", - "DeleteVpnConnection": "

Deletes the specified VPN connection.

If you're deleting the VPC and its associated components, we recommend that you detach the virtual private gateway from the VPC and delete the VPC before deleting the VPN connection. If you believe that the tunnel credentials for your VPN connection have been compromised, you can delete the VPN connection and create a new one that has new keys, without needing to delete the VPC or virtual private gateway. If you create a new VPN connection, you must reconfigure the customer gateway using the new configuration information returned with the new VPN connection ID.

", - "DeleteVpnConnectionRoute": "

Deletes the specified static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

", - "DeleteVpnGateway": "

Deletes the specified virtual private gateway. We recommend that before you delete a virtual private gateway, you detach it from the VPC and delete the VPN connection. Note that you don't need to delete the virtual private gateway if you plan to delete and recreate the VPN connection between your VPC and your network.

", - "DeregisterImage": "

Deregisters the specified AMI. After you deregister an AMI, it can't be used to launch new instances.

This command does not delete the AMI.

", - "DescribeAccountAttributes": "

Describes attributes of your AWS account. The following are the supported account attributes:

  • supported-platforms: Indicates whether your account can launch instances into EC2-Classic and EC2-VPC, or only into EC2-VPC.

  • default-vpc: The ID of the default VPC for your account, or none.

  • max-instances: The maximum number of On-Demand instances that you can run.

  • vpc-max-security-groups-per-interface: The maximum number of security groups that you can assign to a network interface.

  • max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-Classic.

  • vpc-max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-VPC.

", - "DescribeAddresses": "

Describes one or more of your Elastic IP addresses.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeAvailabilityZones": "

Describes one or more of the Availability Zones that are available to you. The results include zones only for the region you're currently using. If there is an event impacting an Availability Zone, you can use this request to view the state and any provided message for that Availability Zone.

For more information, see Regions and Availability Zones in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeBundleTasks": "

Describes one or more of your bundling tasks.

Completed bundle tasks are listed for only a limited time. If your bundle task is no longer in the list, you can still register an AMI from it. Just use RegisterImage with the Amazon S3 bucket name and image manifest name you provided to the bundle task.

", - "DescribeClassicLinkInstances": "

Describes one or more of your linked EC2-Classic instances. This request only returns information about EC2-Classic instances linked to a VPC through ClassicLink; you cannot use this request to return information about other instances.

", - "DescribeConversionTasks": "

Describes one or more of your conversion tasks. For more information, see the VM Import/Export User Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DescribeCustomerGateways": "

Describes one or more of your VPN customer gateways.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeDhcpOptions": "

Describes one or more of your DHCP options sets.

For more information about DHCP options sets, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeExportTasks": "

Describes one or more of your export tasks.

", - "DescribeFlowLogs": "

Describes one or more flow logs. To view the information in your flow logs (the log streams for the network interfaces), you must use the CloudWatch Logs console or the CloudWatch Logs API.

", - "DescribeHostReservationOfferings": "

Describes the Dedicated Host Reservations that are available to purchase.

The results describe all the Dedicated Host Reservation offerings, including offerings that may not match the instance family and region of your Dedicated Hosts. When purchasing an offering, ensure that the the instance family and region of the offering matches that of the Dedicated Host/s it will be associated with. For an overview of supported instance types, see Dedicated Hosts Overview in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeHostReservations": "

Describes Dedicated Host Reservations which are associated with Dedicated Hosts in your account.

", - "DescribeHosts": "

Describes one or more of your Dedicated Hosts.

The results describe only the Dedicated Hosts in the region you're currently using. All listed instances consume capacity on your Dedicated Host. Dedicated Hosts that have recently been released will be listed with the state released.

", - "DescribeIdFormat": "

Describes the ID format settings for your resources on a per-region basis, for example, to view which resource types are enabled for longer IDs. This request only returns information about resource types whose ID formats can be modified; it does not return information about other resource types.

The following resource types support longer IDs: instance | reservation | snapshot | volume.

These settings apply to the IAM user who makes the request; they do not apply to the entire AWS account. By default, an IAM user defaults to the same settings as the root user, unless they explicitly override the settings by running the ModifyIdFormat command. Resources created with longer IDs are visible to all IAM users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "DescribeIdentityIdFormat": "

Describes the ID format settings for resources for the specified IAM user, IAM role, or root user. For example, you can view the resource types that are enabled for longer IDs. This request only returns information about resource types whose ID formats can be modified; it does not return information about other resource types. For more information, see Resource IDs in the Amazon Elastic Compute Cloud User Guide.

The following resource types support longer IDs: instance | reservation | snapshot | volume.

These settings apply to the principal specified in the request. They do not apply to the principal that makes the request.

", - "DescribeImageAttribute": "

Describes the specified attribute of the specified AMI. You can specify only one attribute at a time.

", - "DescribeImages": "

Describes one or more of the images (AMIs, AKIs, and ARIs) available to you. Images available to you include public images, private images that you own, and private images owned by other AWS accounts but for which you have explicit launch permissions.

Deregistered images are included in the returned results for an unspecified interval after deregistration.

", - "DescribeImportImageTasks": "

Displays details about an import virtual machine or import snapshot tasks that are already created.

", - "DescribeImportSnapshotTasks": "

Describes your import snapshot tasks.

", - "DescribeInstanceAttribute": "

Describes the specified attribute of the specified instance. You can specify only one attribute at a time. Valid attribute values are: instanceType | kernel | ramdisk | userData | disableApiTermination | instanceInitiatedShutdownBehavior | rootDeviceName | blockDeviceMapping | productCodes | sourceDestCheck | groupSet | ebsOptimized | sriovNetSupport

", - "DescribeInstanceStatus": "

Describes the status of one or more instances. By default, only running instances are described, unless specified otherwise.

Instance status includes the following components:

  • Status checks - Amazon EC2 performs status checks on running EC2 instances to identify hardware and software issues. For more information, see Status Checks for Your Instances and Troubleshooting Instances with Failed Status Checks in the Amazon Elastic Compute Cloud User Guide.

  • Scheduled events - Amazon EC2 can schedule events (such as reboot, stop, or terminate) for your instances related to hardware issues, software updates, or system maintenance. For more information, see Scheduled Events for Your Instances in the Amazon Elastic Compute Cloud User Guide.

  • Instance state - You can manage your instances from the moment you launch them through their termination. For more information, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeInstances": "

Describes one or more of your instances.

If you specify one or more instance IDs, Amazon EC2 returns information for those instances. If you do not specify instance IDs, Amazon EC2 returns information for all relevant instances. If you specify an instance ID that is not valid, an error is returned. If you specify an instance that you do not own, it is not included in the returned results.

Recently terminated instances might appear in the returned results. This interval is usually less than one hour.

If you describe instances in the rare case where an Availability Zone is experiencing a service disruption and you specify instance IDs that are in the affected zone, or do not specify any instance IDs at all, the call fails. If you describe instances and specify only instance IDs that are in an unaffected zone, the call works normally.

", - "DescribeInternetGateways": "

Describes one or more of your Internet gateways.

", - "DescribeKeyPairs": "

Describes one or more of your key pairs.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeMovingAddresses": "

Describes your Elastic IP addresses that are being moved to the EC2-VPC platform, or that are being restored to the EC2-Classic platform. This request does not return information about any other Elastic IP addresses in your account.

", - "DescribeNatGateways": "

Describes one or more of the your NAT gateways.

", - "DescribeNetworkAcls": "

Describes one or more of your network ACLs.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "DescribeNetworkInterfaceAttribute": "

Describes a network interface attribute. You can specify only one attribute at a time.

", - "DescribeNetworkInterfaces": "

Describes one or more of your network interfaces.

", - "DescribePlacementGroups": "

Describes one or more of your placement groups. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribePrefixLists": "

Describes available AWS services in a prefix list format, which includes the prefix list name and prefix list ID of the service and the IP address range for the service. A prefix list ID is required for creating an outbound security group rule that allows traffic from a VPC to access an AWS service through a VPC endpoint.

", - "DescribeRegions": "

Describes one or more regions that are currently available to you.

For a list of the regions supported by Amazon EC2, see Regions and Endpoints.

", - "DescribeReservedInstances": "

Describes one or more of the Reserved Instances that you purchased.

For more information about Reserved Instances, see Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesListings": "

Describes your account's Reserved Instance listings in the Reserved Instance Marketplace.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

As a seller, you choose to list some or all of your Reserved Instances, and you specify the upfront price to receive for them. Your Reserved Instances are then listed in the Reserved Instance Marketplace and are available for purchase.

As a buyer, you specify the configuration of the Reserved Instance to purchase, and the Marketplace matches what you're searching for with what's available. The Marketplace first sells the lowest priced Reserved Instances to you, and continues to sell available Reserved Instance listings to you until your demand is met. You are charged based on the total price of all of the listings that you purchase.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesModifications": "

Describes the modifications made to your Reserved Instances. If no parameter is specified, information about all your Reserved Instances modification requests is returned. If a modification ID is specified, only information about the specific modification is returned.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesOfferings": "

Describes Reserved Instance offerings that are available for purchase. With Reserved Instances, you purchase the right to launch instances for a period of time. During that time period, you do not receive insufficient capacity errors, and you pay a lower usage rate than the rate charged for On-Demand instances for the actual time used.

If you have listed your own Reserved Instances for sale in the Reserved Instance Marketplace, they will be excluded from these results. This is to ensure that you do not purchase your own Reserved Instances.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeRouteTables": "

Describes one or more of your route tables.

Each subnet in your VPC must be associated with a route table. If a subnet is not explicitly associated with any route table, it is implicitly associated with the main route table. This command does not return the subnet ID for implicit associations.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "DescribeScheduledInstanceAvailability": "

Finds available schedules that meet the specified criteria.

You can search for an available schedule no more than 3 months in advance. You must meet the minimum required duration of 1,200 hours per year. For example, the minimum daily schedule is 4 hours, the minimum weekly schedule is 24 hours, and the minimum monthly schedule is 100 hours.

After you find a schedule that meets your needs, call PurchaseScheduledInstances to purchase Scheduled Instances with that schedule.

", - "DescribeScheduledInstances": "

Describes one or more of your Scheduled Instances.

", - "DescribeSecurityGroupReferences": "

[EC2-VPC only] Describes the VPCs on the other side of a VPC peering connection that are referencing the security groups you've specified in this request.

", - "DescribeSecurityGroups": "

Describes one or more of your security groups.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeSnapshotAttribute": "

Describes the specified attribute of the specified snapshot. You can specify only one attribute at a time.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSnapshots": "

Describes one or more of the EBS snapshots available to you. Available snapshots include public snapshots available for any AWS account to launch, private snapshots that you own, and private snapshots owned by another AWS account but for which you've been given explicit create volume permissions.

The create volume permissions fall into the following categories:

  • public: The owner of the snapshot granted create volume permissions for the snapshot to the all group. All AWS accounts have create volume permissions for these snapshots.

  • explicit: The owner of the snapshot granted create volume permissions to a specific AWS account.

  • implicit: An AWS account has implicit create volume permissions for all snapshots it owns.

The list of snapshots returned can be modified by specifying snapshot IDs, snapshot owners, or AWS accounts with create volume permissions. If no options are specified, Amazon EC2 returns all snapshots for which you have create volume permissions.

If you specify one or more snapshot IDs, only snapshots that have the specified IDs are returned. If you specify an invalid snapshot ID, an error is returned. If you specify a snapshot ID for which you do not have access, it is not included in the returned results.

If you specify one or more snapshot owners using the OwnerIds option, only snapshots from the specified owners and for which you have access are returned. The results can include the AWS account IDs of the specified owners, amazon for snapshots owned by Amazon, or self for snapshots that you own.

If you specify a list of restorable users, only snapshots with create snapshot permissions for those users are returned. You can specify AWS account IDs (if you own the snapshots), self for snapshots for which you own or have explicit permissions, or all for public snapshots.

If you are describing a long list of snapshots, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeSnapshots request to retrieve the remaining results.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotDatafeedSubscription": "

Describes the data feed for Spot instances. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotFleetInstances": "

Describes the running instances for the specified Spot fleet.

", - "DescribeSpotFleetRequestHistory": "

Describes the events for the specified Spot fleet request during the specified time.

Spot fleet events are delayed by up to 30 seconds before they can be described. This ensures that you can query by the last evaluated time and not miss a recorded event.

", - "DescribeSpotFleetRequests": "

Describes your Spot fleet requests.

", - "DescribeSpotInstanceRequests": "

Describes the Spot instance requests that belong to your account. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

You can use DescribeSpotInstanceRequests to find a running Spot instance by examining the response. If the status of the Spot instance is fulfilled, the instance ID appears in the response and contains the identifier of the instance. Alternatively, you can use DescribeInstances with a filter to look for instances where the instance lifecycle is spot.

", - "DescribeSpotPriceHistory": "

Describes the Spot price history. The prices returned are listed in chronological order, from the oldest to the most recent, for up to the past 90 days. For more information, see Spot Instance Pricing History in the Amazon Elastic Compute Cloud User Guide.

When you specify a start and end time, this operation returns the prices of the instance types within the time range that you specified and the time when the price changed. The price is valid within the time period that you specified; the response merely indicates the last time that the price changed.

", - "DescribeStaleSecurityGroups": "

[EC2-VPC only] Describes the stale security group rules for security groups in a specified VPC. Rules are stale when they reference a deleted security group in a peer VPC, or a security group in a peer VPC for which the VPC peering connection has been deleted.

", - "DescribeSubnets": "

Describes one or more of your subnets.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeTags": "

Describes one or more of the tags for your EC2 resources.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeAttribute": "

Describes the specified attribute of the specified volume. You can specify only one attribute at a time.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeStatus": "

Describes the status of the specified volumes. Volume status provides the result of the checks performed on your volumes to determine events that can impair the performance of your volumes. The performance of a volume can be affected if an issue occurs on the volume's underlying host. If the volume's underlying host experiences a power outage or system issue, after the system is restored, there could be data inconsistencies on the volume. Volume events notify you if this occurs. Volume actions notify you if any action needs to be taken in response to the event.

The DescribeVolumeStatus operation provides the following information about the specified volumes:

Status: Reflects the current status of the volume. The possible values are ok, impaired , warning, or insufficient-data. If all checks pass, the overall status of the volume is ok. If the check fails, the overall status is impaired. If the status is insufficient-data, then the checks may still be taking place on your volume at the time. We recommend that you retry the request. For more information on volume status, see Monitoring the Status of Your Volumes.

Events: Reflect the cause of a volume status and may require you to take action. For example, if your volume returns an impaired status, then the volume event might be potential-data-inconsistency. This means that your volume has been affected by an issue with the underlying host, has all I/O operations disabled, and may have inconsistent data.

Actions: Reflect the actions you may have to take in response to an event. For example, if the status of the volume is impaired and the volume event shows potential-data-inconsistency, then the action shows enable-volume-io. This means that you may want to enable the I/O operations for the volume by calling the EnableVolumeIO action and then check the volume for data consistency.

Volume status is based on the volume status checks, and does not reflect the volume state. Therefore, volume status does not indicate volumes in the error state (for example, when a volume is incapable of accepting I/O.)

", - "DescribeVolumes": "

Describes the specified EBS volumes.

If you are describing a long list of volumes, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeVolumes request to retrieve the remaining results.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcAttribute": "

Describes the specified attribute of the specified VPC. You can specify only one attribute at a time.

", - "DescribeVpcClassicLink": "

Describes the ClassicLink status of one or more VPCs.

", - "DescribeVpcClassicLinkDnsSupport": "

Describes the ClassicLink DNS support status of one or more VPCs. If enabled, the DNS hostname of a linked EC2-Classic instance resolves to its private IP address when addressed from an instance in the VPC to which it's linked. Similarly, the DNS hostname of an instance in a VPC resolves to its private IP address when addressed from a linked EC2-Classic instance. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcEndpointServices": "

Describes all supported AWS services that can be specified when creating a VPC endpoint.

", - "DescribeVpcEndpoints": "

Describes one or more of your VPC endpoints.

", - "DescribeVpcPeeringConnections": "

Describes one or more of your VPC peering connections.

", - "DescribeVpcs": "

Describes one or more of your VPCs.

", - "DescribeVpnConnections": "

Describes one or more of your VPN connections.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeVpnGateways": "

Describes one or more of your virtual private gateways.

For more information about virtual private gateways, see Adding an IPsec Hardware VPN to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DetachClassicLinkVpc": "

Unlinks (detaches) a linked EC2-Classic instance from a VPC. After the instance has been unlinked, the VPC security groups are no longer associated with it. An instance is automatically unlinked from a VPC when it's stopped.

", - "DetachInternetGateway": "

Detaches an Internet gateway from a VPC, disabling connectivity between the Internet and the VPC. The VPC must not contain any running instances with Elastic IP addresses.

", - "DetachNetworkInterface": "

Detaches a network interface from an instance.

", - "DetachVolume": "

Detaches an EBS volume from an instance. Make sure to unmount any file systems on the device within your operating system before detaching the volume. Failure to do so can result in the volume becoming stuck in the busy state while detaching. If this happens, detachment can be delayed indefinitely until you unmount the volume, force detachment, reboot the instance, or all three. If an EBS volume is the root device of an instance, it can't be detached while the instance is running. To detach the root volume, stop the instance first.

When a volume with an AWS Marketplace product code is detached from an instance, the product code is no longer associated with the instance.

For more information, see Detaching an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DetachVpnGateway": "

Detaches a virtual private gateway from a VPC. You do this if you're planning to turn off the VPC and not use it anymore. You can confirm a virtual private gateway has been completely detached from a VPC by describing the virtual private gateway (any attachments to the virtual private gateway are also described).

You must wait for the attachment's state to switch to detached before you can delete the VPC or attach a different VPC to the virtual private gateway.

", - "DisableVgwRoutePropagation": "

Disables a virtual private gateway (VGW) from propagating routes to a specified route table of a VPC.

", - "DisableVpcClassicLink": "

Disables ClassicLink for a VPC. You cannot disable ClassicLink for a VPC that has EC2-Classic instances linked to it.

", - "DisableVpcClassicLinkDnsSupport": "

Disables ClassicLink DNS support for a VPC. If disabled, DNS hostnames resolve to public IP addresses when addressed between a linked EC2-Classic instance and instances in the VPC to which it's linked. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "DisassociateAddress": "

Disassociates an Elastic IP address from the instance or network interface it's associated with.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

", - "DisassociateRouteTable": "

Disassociates a subnet from a route table.

After you perform this action, the subnet no longer uses the routes in the route table. Instead, it uses the routes in the VPC's main route table. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "EnableVgwRoutePropagation": "

Enables a virtual private gateway (VGW) to propagate routes to the specified route table of a VPC.

", - "EnableVolumeIO": "

Enables I/O operations for a volume that had I/O operations disabled because the data on the volume was potentially inconsistent.

", - "EnableVpcClassicLink": "

Enables a VPC for ClassicLink. You can then link EC2-Classic instances to your ClassicLink-enabled VPC to allow communication over private IP addresses. You cannot enable your VPC for ClassicLink if any of your VPC's route tables have existing routes for address ranges within the 10.0.0.0/8 IP address range, excluding local routes for VPCs in the 10.0.0.0/16 and 10.1.0.0/16 IP address ranges. For more information, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "EnableVpcClassicLinkDnsSupport": "

Enables a VPC to support DNS hostname resolution for ClassicLink. If enabled, the DNS hostname of a linked EC2-Classic instance resolves to its private IP address when addressed from an instance in the VPC to which it's linked. Similarly, the DNS hostname of an instance in a VPC resolves to its private IP address when addressed from a linked EC2-Classic instance. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "GetConsoleOutput": "

Gets the console output for the specified instance.

Instances do not have a physical monitor through which you can view their console output. They also lack physical controls that allow you to power up, reboot, or shut them down. To allow these actions, we provide them through the Amazon EC2 API and command line interface.

Instance console output is buffered and posted shortly after instance boot, reboot, and termination. Amazon EC2 preserves the most recent 64 KB output which is available for at least one hour after the most recent post.

For Linux instances, the instance console output displays the exact console output that would normally be displayed on a physical monitor attached to a computer. This output is buffered because the instance produces it and then posts it to a store where the instance's owner can retrieve it.

For Windows instances, the instance console output includes output from the EC2Config service.

", - "GetConsoleScreenshot": "

Retrieve a JPG-format screenshot of a running instance to help with troubleshooting.

The returned content is Base64-encoded.

", - "GetHostReservationPurchasePreview": "

Preview a reservation purchase with configurations that match those of your Dedicated Host. You must have active Dedicated Hosts in your account before you purchase a reservation.

This is a preview of the PurchaseHostReservation action and does not result in the offering being purchased.

", - "GetPasswordData": "

Retrieves the encrypted administrator password for an instance running Windows.

The Windows password is generated at boot if the EC2Config service plugin, Ec2SetPassword, is enabled. This usually only happens the first time an AMI is launched, and then Ec2SetPassword is automatically disabled. The password is not generated for rebundled AMIs unless Ec2SetPassword is enabled before bundling.

The password is encrypted using the key pair that you specified when you launched the instance. You must provide the corresponding key pair file.

Password generation and encryption takes a few moments. We recommend that you wait up to 15 minutes after launching an instance before trying to retrieve the generated password.

", - "ImportImage": "

Import single or multi-volume disk images or EBS snapshots into an Amazon Machine Image (AMI). For more information, see Importing a VM as an Image Using VM Import/Export in the VM Import/Export User Guide.

", - "ImportInstance": "

Creates an import instance task using metadata from the specified disk image. ImportInstance only supports single-volume VMs. To import multi-volume VMs, use ImportImage. For more information, see Importing a Virtual Machine Using the Amazon EC2 CLI.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "ImportKeyPair": "

Imports the public key from an RSA key pair that you created with a third-party tool. Compare this with CreateKeyPair, in which AWS creates the key pair and gives the keys to you (AWS keeps a copy of the public key). With ImportKeyPair, you create the key pair and give AWS just the public key. The private key is never transferred between you and AWS.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "ImportSnapshot": "

Imports a disk into an EBS snapshot.

", - "ImportVolume": "

Creates an import volume task using metadata from the specified disk image.For more information, see Importing Disks to Amazon EBS.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "ModifyHosts": "

Modify the auto-placement setting of a Dedicated Host. When auto-placement is enabled, AWS will place instances that you launch with a tenancy of host, but without targeting a specific host ID, onto any available Dedicated Host in your account which has auto-placement enabled. When auto-placement is disabled, you need to provide a host ID if you want the instance to launch onto a specific host. If no host ID is provided, the instance will be launched onto a suitable host which has auto-placement enabled.

", - "ModifyIdFormat": "

Modifies the ID format for the specified resource on a per-region basis. You can specify that resources should receive longer IDs (17-character IDs) when they are created. The following resource types support longer IDs: instance | reservation | snapshot | volume.

This setting applies to the IAM user who makes the request; it does not apply to the entire AWS account. By default, an IAM user defaults to the same settings as the root user. If you're using this action as the root user, then these settings apply to the entire account, unless an IAM user explicitly overrides these settings for themselves. For more information, see Resource IDs in the Amazon Elastic Compute Cloud User Guide.

Resources created with longer IDs are visible to all IAM roles and users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "ModifyIdentityIdFormat": "

Modifies the ID format of a resource for a specified IAM user, IAM role, or the root user for an account; or all IAM users, IAM roles, and the root user for an account. You can specify that resources should receive longer IDs (17-character IDs) when they are created.

The following resource types support longer IDs: instance | reservation | snapshot | volume. For more information, see Resource IDs in the Amazon Elastic Compute Cloud User Guide.

This setting applies to the principal specified in the request; it does not apply to the principal that makes the request.

Resources created with longer IDs are visible to all IAM roles and users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "ModifyImageAttribute": "

Modifies the specified attribute of the specified AMI. You can specify only one attribute at a time.

AWS Marketplace product codes cannot be modified. Images with an AWS Marketplace product code cannot be made public.

The SriovNetSupport enhanced networking attribute cannot be changed using this command. Instead, enable SriovNetSupport on an instance and create an AMI from the instance. This will result in an image with SriovNetSupport enabled.

", - "ModifyInstanceAttribute": "

Modifies the specified attribute of the specified instance. You can specify only one attribute at a time.

To modify some attributes, the instance must be stopped. For more information, see Modifying Attributes of a Stopped Instance in the Amazon Elastic Compute Cloud User Guide.

", - "ModifyInstancePlacement": "

Set the instance affinity value for a specific stopped instance and modify the instance tenancy setting.

Instance affinity is disabled by default. When instance affinity is host and it is not associated with a specific Dedicated Host, the next time it is launched it will automatically be associated with the host it lands on. This relationship will persist if the instance is stopped/started, or rebooted.

You can modify the host ID associated with a stopped instance. If a stopped instance has a new host ID association, the instance will target that host when restarted.

You can modify the tenancy of a stopped instance with a tenancy of host or dedicated.

Affinity, hostID, and tenancy are not required parameters, but at least one of them must be specified in the request. Affinity and tenancy can be modified in the same request, but tenancy can only be modified on instances that are stopped.

", - "ModifyNetworkInterfaceAttribute": "

Modifies the specified network interface attribute. You can specify only one attribute at a time.

", - "ModifyReservedInstances": "

Modifies the Availability Zone, instance count, instance type, or network platform (EC2-Classic or EC2-VPC) of your Reserved Instances. The Reserved Instances to be modified must be identical, except for Availability Zone, network platform, and instance type.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "ModifySnapshotAttribute": "

Adds or removes permission settings for the specified snapshot. You may add or remove specified AWS account IDs from a snapshot's list of create volume permissions, but you cannot do both in a single API call. If you need to both add and remove account IDs for a snapshot, you must use multiple API calls.

Encrypted snapshots and snapshots with AWS Marketplace product codes cannot be made public. Snapshots encrypted with your default CMK cannot be shared with other accounts.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "ModifySpotFleetRequest": "

Modifies the specified Spot fleet request.

While the Spot fleet request is being modified, it is in the modifying state.

To scale up your Spot fleet, increase its target capacity. The Spot fleet launches the additional Spot instances according to the allocation strategy for the Spot fleet request. If the allocation strategy is lowestPrice, the Spot fleet launches instances using the Spot pool with the lowest price. If the allocation strategy is diversified, the Spot fleet distributes the instances across the Spot pools.

To scale down your Spot fleet, decrease its target capacity. First, the Spot fleet cancels any open bids that exceed the new target capacity. You can request that the Spot fleet terminate Spot instances until the size of the fleet no longer exceeds the new target capacity. If the allocation strategy is lowestPrice, the Spot fleet terminates the instances with the highest price per unit. If the allocation strategy is diversified, the Spot fleet terminates instances across the Spot pools. Alternatively, you can request that the Spot fleet keep the fleet at its current size, but not replace any Spot instances that are interrupted or that you terminate manually.

", - "ModifySubnetAttribute": "

Modifies a subnet attribute.

", - "ModifyVolumeAttribute": "

Modifies a volume attribute.

By default, all I/O operations for the volume are suspended when the data on the volume is determined to be potentially inconsistent, to prevent undetectable, latent data corruption. The I/O access to the volume can be resumed by first enabling I/O access and then checking the data consistency on your volume.

You can change the default behavior to resume I/O operations. We recommend that you change this only for boot volumes or for volumes that are stateless or disposable.

", - "ModifyVpcAttribute": "

Modifies the specified attribute of the specified VPC.

", - "ModifyVpcEndpoint": "

Modifies attributes of a specified VPC endpoint. You can modify the policy associated with the endpoint, and you can add and remove route tables associated with the endpoint.

", - "ModifyVpcPeeringConnectionOptions": "

Modifies the VPC peering connection options on one side of a VPC peering connection. You can do the following:

  • Enable/disable communication over the peering connection between an EC2-Classic instance that's linked to your VPC (using ClassicLink) and instances in the peer VPC.

  • Enable/disable communication over the peering connection between instances in your VPC and an EC2-Classic instance that's linked to the peer VPC.

  • Enable/disable a local VPC to resolve public DNS hostnames to private IP addresses when queried from instances in the peer VPC.

If the peered VPCs are in different accounts, each owner must initiate a separate request to modify the peering connection options, depending on whether their VPC was the requester or accepter for the VPC peering connection. If the peered VPCs are in the same account, you can modify the requester and accepter options in the same request. To confirm which VPC is the accepter and requester for a VPC peering connection, use the DescribeVpcPeeringConnections command.

", - "MonitorInstances": "

Enables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "MoveAddressToVpc": "

Moves an Elastic IP address from the EC2-Classic platform to the EC2-VPC platform. The Elastic IP address must be allocated to your account for more than 24 hours, and it must not be associated with an instance. After the Elastic IP address is moved, it is no longer available for use in the EC2-Classic platform, unless you move it back using the RestoreAddressToClassic request. You cannot move an Elastic IP address that was originally allocated for use in the EC2-VPC platform to the EC2-Classic platform.

", - "PurchaseHostReservation": "

Purchase a reservation with configurations that match those of your Dedicated Host. You must have active Dedicated Hosts in your account before you purchase a reservation. This action results in the specified reservation being purchased and charged to your account.

", - "PurchaseReservedInstancesOffering": "

Purchases a Reserved Instance for use with your account. With Reserved Instances, you obtain a capacity reservation for a certain instance configuration over a specified period of time and pay a lower hourly rate compared to On-Demand instance pricing.

Use DescribeReservedInstancesOfferings to get a list of Reserved Instance offerings that match your specifications. After you've purchased a Reserved Instance, you can check for your new Reserved Instance with DescribeReservedInstances.

For more information, see Reserved Instances and Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "PurchaseScheduledInstances": "

Purchases one or more Scheduled Instances with the specified schedule.

Scheduled Instances enable you to purchase Amazon EC2 compute capacity by the hour for a one-year term. Before you can purchase a Scheduled Instance, you must call DescribeScheduledInstanceAvailability to check for available schedules and obtain a purchase token. After you purchase a Scheduled Instance, you must call RunScheduledInstances during each scheduled time period.

After you purchase a Scheduled Instance, you can't cancel, modify, or resell your purchase.

", - "RebootInstances": "

Requests a reboot of one or more instances. This operation is asynchronous; it only queues a request to reboot the specified instances. The operation succeeds if the instances are valid and belong to you. Requests to reboot terminated instances are ignored.

If an instance does not cleanly shut down within four minutes, Amazon EC2 performs a hard reboot.

For more information about troubleshooting, see Getting Console Output and Rebooting Instances in the Amazon Elastic Compute Cloud User Guide.

", - "RegisterImage": "

Registers an AMI. When you're creating an AMI, this is the final step you must complete before you can launch an instance from the AMI. For more information about creating AMIs, see Creating Your Own AMIs in the Amazon Elastic Compute Cloud User Guide.

For Amazon EBS-backed instances, CreateImage creates and registers the AMI in a single request, so you don't have to register the AMI yourself.

You can also use RegisterImage to create an Amazon EBS-backed Linux AMI from a snapshot of a root device volume. For more information, see Launching an Instance from a Snapshot in the Amazon Elastic Compute Cloud User Guide.

Some Linux distributions, such as Red Hat Enterprise Linux (RHEL) and SUSE Linux Enterprise Server (SLES), use the EC2 billingProduct code associated with an AMI to verify subscription status for package updates. Creating an AMI from an EBS snapshot does not maintain this billing code, and subsequent instances launched from such an AMI will not be able to connect to package update infrastructure.

Similarly, although you can create a Windows AMI from a snapshot, you can't successfully launch an instance from the AMI.

To create Windows AMIs or to create AMIs for Linux operating systems that must retain AMI billing codes to work properly, see CreateImage.

If needed, you can deregister an AMI at any time. Any modifications you make to an AMI backed by an instance store volume invalidates its registration. If you make changes to an image, deregister the previous image and register the new image.

You can't register an image where a secondary (non-root) snapshot has AWS Marketplace product codes.

", - "RejectVpcPeeringConnection": "

Rejects a VPC peering connection request. The VPC peering connection must be in the pending-acceptance state. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests. To delete an active VPC peering connection, or to delete a VPC peering connection request that you initiated, use DeleteVpcPeeringConnection.

", - "ReleaseAddress": "

Releases the specified Elastic IP address.

After releasing an Elastic IP address, it is released to the IP address pool and might be unavailable to you. Be sure to update your DNS records and any servers or devices that communicate with the address. If you attempt to release an Elastic IP address that you already released, you'll get an AuthFailure error if the address is already allocated to another AWS account.

[EC2-Classic, default VPC] Releasing an Elastic IP address automatically disassociates it from any instance that it's associated with. To disassociate an Elastic IP address without releasing it, use DisassociateAddress.

[Nondefault VPC] You must use DisassociateAddress to disassociate the Elastic IP address before you try to release it. Otherwise, Amazon EC2 returns an error (InvalidIPAddress.InUse).

", - "ReleaseHosts": "

When you no longer want to use an On-Demand Dedicated Host it can be released. On-Demand billing is stopped and the host goes into released state. The host ID of Dedicated Hosts that have been released can no longer be specified in another request, e.g., ModifyHosts. You must stop or terminate all instances on a host before it can be released.

When Dedicated Hosts are released, it make take some time for them to stop counting toward your limit and you may receive capacity errors when trying to allocate new Dedicated hosts. Try waiting a few minutes, and then try again.

Released hosts will still appear in a DescribeHosts response.

", - "ReplaceNetworkAclAssociation": "

Changes which network ACL a subnet is associated with. By default when you create a subnet, it's automatically associated with the default network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceNetworkAclEntry": "

Replaces an entry (rule) in a network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRoute": "

Replaces an existing route within a route table in a VPC. You must provide only one of the following: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, or network interface.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRouteTableAssociation": "

Changes the route table associated with a given subnet in a VPC. After the operation completes, the subnet uses the routes in the new route table it's associated with. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

You can also use ReplaceRouteTableAssociation to change which table is the main route table in the VPC. You just specify the main route table's association ID and the route table to be the new main route table.

", - "ReportInstanceStatus": "

Submits feedback about the status of an instance. The instance must be in the running state. If your experience with the instance differs from the instance status returned by DescribeInstanceStatus, use ReportInstanceStatus to report your experience with the instance. Amazon EC2 collects this information to improve the accuracy of status checks.

Use of this action does not change the value returned by DescribeInstanceStatus.

", - "RequestSpotFleet": "

Creates a Spot fleet request.

You can submit a single request that includes multiple launch specifications that vary by instance type, AMI, Availability Zone, or subnet.

By default, the Spot fleet requests Spot instances in the Spot pool where the price per unit is the lowest. Each launch specification can include its own instance weighting that reflects the value of the instance type to your application workload.

Alternatively, you can specify that the Spot fleet distribute the target capacity across the Spot pools included in its launch specifications. By ensuring that the Spot instances in your Spot fleet are in different Spot pools, you can improve the availability of your fleet.

For more information, see Spot Fleet Requests in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstances": "

Creates a Spot instance request. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot Instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

", - "ResetImageAttribute": "

Resets an attribute of an AMI to its default value.

The productCodes attribute can't be reset.

", - "ResetInstanceAttribute": "

Resets an attribute of an instance to its default value. To reset the kernel or ramdisk, the instance must be in a stopped state. To reset the sourceDestCheck, the instance can be either running or stopped.

The sourceDestCheck attribute controls whether source/destination checking is enabled. The default value is true, which means checking is enabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ResetNetworkInterfaceAttribute": "

Resets a network interface attribute. You can specify only one attribute at a time.

", - "ResetSnapshotAttribute": "

Resets permission settings for the specified snapshot.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "RestoreAddressToClassic": "

Restores an Elastic IP address that was previously moved to the EC2-VPC platform back to the EC2-Classic platform. You cannot move an Elastic IP address that was originally allocated for use in EC2-VPC. The Elastic IP address must not be associated with an instance or network interface.

", - "RevokeSecurityGroupEgress": "

[EC2-VPC only] Removes one or more egress rules from a security group for EC2-VPC. This action doesn't apply to security groups for use in EC2-Classic. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be revoked.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RevokeSecurityGroupIngress": "

Removes one or more ingress rules from a security group. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be removed.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RunInstances": "

Launches the specified number of instances using an AMI for which you have permissions.

When you launch an instance, it enters the pending state. After the instance is ready for you, it enters the running state. To check the state of your instance, call DescribeInstances.

To ensure faster instance launches, break up large requests into smaller batches. For example, create five separate launch requests for 100 instances each instead of one launch request for 500 instances.

To tag your instance, ensure that it is running as CreateTags requires a resource ID. For more information about tagging, see Tagging Your Amazon EC2 Resources.

If you don't specify a security group when launching an instance, Amazon EC2 uses the default security group. For more information, see Security Groups in the Amazon Elastic Compute Cloud User Guide.

[EC2-VPC only accounts] If you don't specify a subnet in the request, we choose a default subnet from your default VPC for you.

[EC2-Classic accounts] If you're launching into EC2-Classic and you don't specify an Availability Zone, we choose one for you.

Linux instances have access to the public key of the key pair at boot. You can use this key to provide secure access to the instance. Amazon EC2 public images use this feature to provide secure access without passwords. For more information, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

You can provide optional user data when launching an instance. For more information, see Instance Metadata in the Amazon Elastic Compute Cloud User Guide.

If any of the AMIs have a product code attached for which the user has not subscribed, RunInstances fails.

Some instance types can only be launched into a VPC. If you do not have a default VPC, or if you do not specify a subnet ID in the request, RunInstances fails. For more information, see Instance Types Available Only in a VPC.

For more information about troubleshooting, see What To Do If An Instance Immediately Terminates, and Troubleshooting Connecting to Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "RunScheduledInstances": "

Launches the specified Scheduled Instances.

Before you can launch a Scheduled Instance, you must purchase it and obtain an identifier using PurchaseScheduledInstances.

You must launch a Scheduled Instance during its scheduled time period. You can't stop or reboot a Scheduled Instance, but you can terminate it as needed. If you terminate a Scheduled Instance before the current scheduled time period ends, you can launch it again after a few minutes. For more information, see Scheduled Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StartInstances": "

Starts an Amazon EBS-backed AMI that you've previously stopped.

Instances that use Amazon EBS volumes as their root devices can be quickly stopped and started. When an instance is stopped, the compute resources are released and you are not billed for hourly instance usage. However, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. You can restart your instance at any time. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Performing this operation on an instance that uses an instance store as its root device returns an error.

For more information, see Stopping Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StopInstances": "

Stops an Amazon EBS-backed instance.

We don't charge hourly usage for a stopped instance, or data transfer fees; however, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

You can't start or stop Spot instances, and you can't stop instance store-backed instances.

When you stop an instance, we shut it down. You can restart your instance at any time. Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Stopping an instance is different to rebooting or terminating it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, the root device and any other devices attached during the instance launch are automatically deleted. For more information about the differences between rebooting, stopping, and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

When you stop an instance, we attempt to shut it down forcibly after a short while. If your instance appears stuck in the stopping state after a period of time, there may be an issue with the underlying host computer. For more information, see Troubleshooting Stopping Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "TerminateInstances": "

Shuts down one or more instances. This operation is idempotent; if you terminate an instance more than once, each call succeeds.

Terminated instances remain visible after termination (for approximately one hour).

By default, Amazon EC2 deletes all EBS volumes that were attached when the instance launched. Volumes attached after instance launch continue running.

You can stop, start, and terminate EBS-backed instances. You can only terminate instance store-backed instances. What happens to an instance differs if you stop it or terminate it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, any attached EBS volumes with the DeleteOnTermination block device mapping parameter set to true are automatically deleted. For more information about the differences between stopping and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

For more information about troubleshooting, see Troubleshooting Terminating Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "UnassignPrivateIpAddresses": "

Unassigns one or more secondary private IP addresses from a network interface.

", - "UnmonitorInstances": "

Disables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

" - }, - "shapes": { - "AcceptVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for AcceptVpcPeeringConnection.

", - "refs": { - } - }, - "AcceptVpcPeeringConnectionResult": { - "base": "

Contains the output of AcceptVpcPeeringConnection.

", - "refs": { - } - }, - "AccountAttribute": { - "base": "

Describes an account attribute.

", - "refs": { - "AccountAttributeList$member": null - } - }, - "AccountAttributeList": { - "base": null, - "refs": { - "DescribeAccountAttributesResult$AccountAttributes": "

Information about one or more account attributes.

" - } - }, - "AccountAttributeName": { - "base": null, - "refs": { - "AccountAttributeNameStringList$member": null - } - }, - "AccountAttributeNameStringList": { - "base": null, - "refs": { - "DescribeAccountAttributesRequest$AttributeNames": "

One or more account attribute names.

" - } - }, - "AccountAttributeValue": { - "base": "

Describes a value of an account attribute.

", - "refs": { - "AccountAttributeValueList$member": null - } - }, - "AccountAttributeValueList": { - "base": null, - "refs": { - "AccountAttribute$AttributeValues": "

One or more values for the account attribute.

" - } - }, - "ActiveInstance": { - "base": "

Describes a running instance in a Spot fleet.

", - "refs": { - "ActiveInstanceSet$member": null - } - }, - "ActiveInstanceSet": { - "base": null, - "refs": { - "DescribeSpotFleetInstancesResponse$ActiveInstances": "

The running instances. Note that this list is refreshed periodically and might be out of date.

" - } - }, - "ActivityStatus": { - "base": null, - "refs": { - "SpotFleetRequestConfig$ActivityStatus": "

The progress of the Spot fleet request. If there is an error, the status is error. After all bids are placed, the status is pending_fulfillment. If the size of the fleet is equal to or greater than its target capacity, the status is fulfilled. If the size of the fleet is decreased, the status is pending_termination while Spot instances are terminating.

" - } - }, - "Address": { - "base": "

Describes an Elastic IP address.

", - "refs": { - "AddressList$member": null - } - }, - "AddressList": { - "base": null, - "refs": { - "DescribeAddressesResult$Addresses": "

Information about one or more Elastic IP addresses.

" - } - }, - "Affinity": { - "base": null, - "refs": { - "ModifyInstancePlacementRequest$Affinity": "

The new affinity setting for the instance.

" - } - }, - "AllocateAddressRequest": { - "base": "

Contains the parameters for AllocateAddress.

", - "refs": { - } - }, - "AllocateAddressResult": { - "base": "

Contains the output of AllocateAddress.

", - "refs": { - } - }, - "AllocateHostsRequest": { - "base": "

Contains the parameters for AllocateHosts.

", - "refs": { - } - }, - "AllocateHostsResult": { - "base": "

Contains the output of AllocateHosts.

", - "refs": { - } - }, - "AllocationIdList": { - "base": null, - "refs": { - "DescribeAddressesRequest$AllocationIds": "

[EC2-VPC] One or more allocation IDs.

Default: Describes all your Elastic IP addresses.

" - } - }, - "AllocationState": { - "base": null, - "refs": { - "Host$State": "

The Dedicated Host's state.

" - } - }, - "AllocationStrategy": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$AllocationStrategy": "

Indicates how to allocate the target capacity across the Spot pools specified by the Spot fleet request. The default is lowestPrice.

" - } - }, - "ArchitectureValues": { - "base": null, - "refs": { - "Image$Architecture": "

The architecture of the image.

", - "ImportInstanceLaunchSpecification$Architecture": "

The architecture of the instance.

", - "Instance$Architecture": "

The architecture of the image.

", - "RegisterImageRequest$Architecture": "

The architecture of the AMI.

Default: For Amazon EBS-backed AMIs, i386. For instance store-backed AMIs, the architecture specified in the manifest file.

" - } - }, - "AssignPrivateIpAddressesRequest": { - "base": "

Contains the parameters for AssignPrivateIpAddresses.

", - "refs": { - } - }, - "AssociateAddressRequest": { - "base": "

Contains the parameters for AssociateAddress.

", - "refs": { - } - }, - "AssociateAddressResult": { - "base": "

Contains the output of AssociateAddress.

", - "refs": { - } - }, - "AssociateDhcpOptionsRequest": { - "base": "

Contains the parameters for AssociateDhcpOptions.

", - "refs": { - } - }, - "AssociateRouteTableRequest": { - "base": "

Contains the parameters for AssociateRouteTable.

", - "refs": { - } - }, - "AssociateRouteTableResult": { - "base": "

Contains the output of AssociateRouteTable.

", - "refs": { - } - }, - "AttachClassicLinkVpcRequest": { - "base": "

Contains the parameters for AttachClassicLinkVpc.

", - "refs": { - } - }, - "AttachClassicLinkVpcResult": { - "base": "

Contains the output of AttachClassicLinkVpc.

", - "refs": { - } - }, - "AttachInternetGatewayRequest": { - "base": "

Contains the parameters for AttachInternetGateway.

", - "refs": { - } - }, - "AttachNetworkInterfaceRequest": { - "base": "

Contains the parameters for AttachNetworkInterface.

", - "refs": { - } - }, - "AttachNetworkInterfaceResult": { - "base": "

Contains the output of AttachNetworkInterface.

", - "refs": { - } - }, - "AttachVolumeRequest": { - "base": "

Contains the parameters for AttachVolume.

", - "refs": { - } - }, - "AttachVpnGatewayRequest": { - "base": "

Contains the parameters for AttachVpnGateway.

", - "refs": { - } - }, - "AttachVpnGatewayResult": { - "base": "

Contains the output of AttachVpnGateway.

", - "refs": { - } - }, - "AttachmentStatus": { - "base": null, - "refs": { - "EbsInstanceBlockDevice$Status": "

The attachment state.

", - "InstanceNetworkInterfaceAttachment$Status": "

The attachment state.

", - "InternetGatewayAttachment$State": "

The current state of the attachment.

", - "NetworkInterfaceAttachment$Status": "

The attachment state.

", - "VpcAttachment$State": "

The current state of the attachment.

" - } - }, - "AttributeBooleanValue": { - "base": "

Describes a value for a resource attribute that is a Boolean value.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$SourceDestCheck": "

Indicates whether source/destination checking is enabled.

", - "DescribeVolumeAttributeResult$AutoEnableIO": "

The state of autoEnableIO attribute.

", - "DescribeVpcAttributeResult$EnableDnsSupport": "

Indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for your instances to their corresponding IP addresses; otherwise, it does not.

", - "DescribeVpcAttributeResult$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.

", - "InstanceAttribute$DisableApiTermination": "

If the value is true, you can't terminate the instance through the Amazon EC2 console, CLI, or API; otherwise, you can.

", - "InstanceAttribute$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O.

", - "InstanceAttribute$EnaSupport": "

Indicates whether enhanced networking with ENA is enabled.

", - "InstanceAttribute$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$SourceDestCheck": "

Specifies whether source/destination checking is enabled. A value of true means that checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$DisableApiTermination": "

If the value is true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. You cannot use this paramater for Spot Instances.

", - "ModifyInstanceAttributeRequest$EbsOptimized": "

Specifies whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "ModifyInstanceAttributeRequest$EnaSupport": "

Set to true to enable enhanced networking with ENA for the instance.

This option is supported only for HVM instances. Specifying this option with a PV instance can make it unreachable.

", - "ModifyNetworkInterfaceAttributeRequest$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ModifySubnetAttributeRequest$MapPublicIpOnLaunch": "

Specify true to indicate that instances launched into the specified subnet should be assigned public IP address.

", - "ModifyVolumeAttributeRequest$AutoEnableIO": "

Indicates whether the volume should be auto-enabled for I/O operations.

", - "ModifyVpcAttributeRequest$EnableDnsSupport": "

Indicates whether the DNS resolution is supported for the VPC. If enabled, queries to the Amazon provided DNS server at the 169.254.169.253 IP address, or the reserved IP address at the base of the VPC network range \"plus two\" will succeed. If disabled, the Amazon provided DNS service in the VPC that resolves public DNS hostnames to IP addresses is not enabled.

You cannot modify the DNS resolution and DNS hostnames attributes in the same request. Use separate requests for each attribute.

", - "ModifyVpcAttributeRequest$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If enabled, instances in the VPC get DNS hostnames; otherwise, they do not.

You cannot modify the DNS resolution and DNS hostnames attributes in the same request. Use separate requests for each attribute. You can only enable DNS hostnames if you've enabled DNS support.

" - } - }, - "AttributeValue": { - "base": "

Describes a value for a resource attribute that is a String.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Description": "

The description of the network interface.

", - "DhcpConfigurationValueList$member": null, - "ImageAttribute$KernelId": "

The kernel ID.

", - "ImageAttribute$RamdiskId": "

The RAM disk ID.

", - "ImageAttribute$Description": "

A description for the AMI.

", - "ImageAttribute$SriovNetSupport": "

Indicates whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "InstanceAttribute$InstanceType": "

The instance type.

", - "InstanceAttribute$KernelId": "

The kernel ID.

", - "InstanceAttribute$RamdiskId": "

The RAM disk ID.

", - "InstanceAttribute$UserData": "

The user data.

", - "InstanceAttribute$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "InstanceAttribute$RootDeviceName": "

The name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "InstanceAttribute$SriovNetSupport": "

Indicates whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "ModifyImageAttributeRequest$Description": "

A description for the AMI.

", - "ModifyInstanceAttributeRequest$InstanceType": "

Changes the instance type to the specified value. For more information, see Instance Types. If the instance type is not valid, the error returned is InvalidInstanceAttributeValue.

", - "ModifyInstanceAttributeRequest$Kernel": "

Changes the instance's kernel to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$Ramdisk": "

Changes the instance's RAM disk to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$InstanceInitiatedShutdownBehavior": "

Specifies whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "ModifyInstanceAttributeRequest$SriovNetSupport": "

Set to simple to enable enhanced networking with the Intel 82599 Virtual Function interface for the instance.

There is no way to disable enhanced networking with the Intel 82599 Virtual Function interface at this time.

This option is supported only for HVM instances. Specifying this option with a PV instance can make it unreachable.

", - "ModifyNetworkInterfaceAttributeRequest$Description": "

A description for the network interface.

" - } - }, - "AuthorizeSecurityGroupEgressRequest": { - "base": "

Contains the parameters for AuthorizeSecurityGroupEgress.

", - "refs": { - } - }, - "AuthorizeSecurityGroupIngressRequest": { - "base": "

Contains the parameters for AuthorizeSecurityGroupIngress.

", - "refs": { - } - }, - "AutoPlacement": { - "base": null, - "refs": { - "AllocateHostsRequest$AutoPlacement": "

This is enabled by default. This property allows instances to be automatically placed onto available Dedicated Hosts, when you are launching instances without specifying a host ID.

Default: Enabled

", - "Host$AutoPlacement": "

Whether auto-placement is on or off.

", - "ModifyHostsRequest$AutoPlacement": "

Specify whether to enable or disable auto-placement.

" - } - }, - "AvailabilityZone": { - "base": "

Describes an Availability Zone.

", - "refs": { - "AvailabilityZoneList$member": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesResult$AvailabilityZones": "

Information about one or more Availability Zones.

" - } - }, - "AvailabilityZoneMessage": { - "base": "

Describes a message about an Availability Zone.

", - "refs": { - "AvailabilityZoneMessageList$member": null - } - }, - "AvailabilityZoneMessageList": { - "base": null, - "refs": { - "AvailabilityZone$Messages": "

Any messages about the Availability Zone.

" - } - }, - "AvailabilityZoneState": { - "base": null, - "refs": { - "AvailabilityZone$State": "

The state of the Availability Zone.

" - } - }, - "AvailableCapacity": { - "base": "

The capacity information for instances launched onto the Dedicated Host.

", - "refs": { - "Host$AvailableCapacity": "

The number of new instances that can be launched onto the Dedicated Host.

" - } - }, - "AvailableInstanceCapacityList": { - "base": null, - "refs": { - "AvailableCapacity$AvailableInstanceCapacity": "

The total number of instances that the Dedicated Host supports.

" - } - }, - "BatchState": { - "base": null, - "refs": { - "CancelSpotFleetRequestsSuccessItem$CurrentSpotFleetRequestState": "

The current state of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$PreviousSpotFleetRequestState": "

The previous state of the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestState": "

The state of the Spot fleet request.

" - } - }, - "Blob": { - "base": null, - "refs": { - "BlobAttributeValue$Value": null, - "ImportKeyPairRequest$PublicKeyMaterial": "

The public key. For API calls, the text must be base64-encoded. For command line tools, base64 encoding is performed for you.

", - "S3Storage$UploadPolicy": "

An Amazon S3 upload policy that gives Amazon EC2 permission to upload items into Amazon S3 on your behalf.

" - } - }, - "BlobAttributeValue": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$UserData": "

Changes the instance's user data to the specified value. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

" - } - }, - "BlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "BlockDeviceMappingList$member": null, - "BlockDeviceMappingRequestList$member": null - } - }, - "BlockDeviceMappingList": { - "base": null, - "refs": { - "Image$BlockDeviceMappings": "

Any block device mapping entries.

", - "ImageAttribute$BlockDeviceMappings": "

One or more block device mapping entries.

", - "LaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

Although you can specify encrypted EBS volumes in this block device mapping for your Spot Instances, these volumes are not encrypted.

", - "RequestSpotLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

Although you can specify encrypted EBS volumes in this block device mapping for your Spot Instances, these volumes are not encrypted.

", - "SpotFleetLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "BlockDeviceMappingRequestList": { - "base": null, - "refs": { - "CreateImageRequest$BlockDeviceMappings": "

Information about one or more block device mappings.

", - "RegisterImageRequest$BlockDeviceMappings": "

One or more block device mapping entries.

", - "RunInstancesRequest$BlockDeviceMappings": "

The block device mapping.

Supplying both a snapshot ID and an encryption value as arguments for block-device mapping results in an error. This is because only blank volumes can be encrypted on start, and these are not created from a snapshot. If a snapshot is the basis for the volume, it contains data by definition and its encryption status cannot be changed using this action.

" - } - }, - "Boolean": { - "base": null, - "refs": { - "AcceptVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AllocateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssignPrivateIpAddressesRequest$AllowReassignment": "

Indicates whether to allow an IP address that is already assigned to another network interface or instance to be reassigned to the specified network interface.

", - "AssociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateAddressRequest$AllowReassociation": "

[EC2-VPC] For a VPC in an EC2-Classic account, specify true to allow an Elastic IP address that is already associated with an instance or network interface to be reassociated with the specified instance or network interface. Otherwise, the operation fails. In a VPC in an EC2-VPC-only account, reassociation is automatic, therefore you can specify false to ensure the operation fails if the Elastic IP address is already associated with another resource.

", - "AssociateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "AttachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttributeBooleanValue$Value": "

The attribute value. The valid values are true or false.

", - "AuthorizeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AuthorizeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "BundleInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelBundleTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelConversionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelImportTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$TerminateInstances": "

Indicates whether to terminate instances for a Spot fleet request if it is canceled successfully.

", - "CancelSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ClassicLinkDnsSupport$ClassicLinkDnsSupported": "

Indicates whether ClassicLink DNS support is enabled for the VPC.

", - "ConfirmProductInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ConfirmProductInstanceResult$Return": "

The return value of the request. Returns true if the specified product code is owned by the requester and associated with the specified instance.

", - "CopyImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopyImageRequest$Encrypted": "

Specifies whether the destination snapshots of the copied image should be encrypted. The default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopySnapshotRequest$Encrypted": "

Specifies whether the destination snapshot should be encrypted. You can encrypt a copy of an unencrypted snapshot using this flag, but you cannot use it to create an unencrypted copy from an encrypted snapshot. Your default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$NoReboot": "

By default, Amazon EC2 attempts to shut down and reboot the instance before creating the image. If the 'No Reboot' option is set, Amazon EC2 doesn't shut down the instance before creating the image. When this option is used, file system integrity on the created image can't be guaranteed.

", - "CreateInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$Egress": "

Indicates whether this is an egress rule (rule is applied to traffic leaving the subnet).

", - "CreateNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreatePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "CreateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$Encrypted": "

Specifies whether the volume should be encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are automatically encrypted. There is no way to create an encrypted volume from an unencrypted snapshot or vice versa. If your AMI uses encrypted volumes, you can only launch it on supported instance types. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$Egress": "

Indicates whether the rule is an egress rule.

", - "DeleteNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeletePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DeleteVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeregisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAccountAttributesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAvailabilityZonesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeBundleTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeClassicLinkInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeConversionTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeCustomerGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImagesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportImageTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportSnapshotTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$IncludeAllInstances": "

When true, includes the health status for all instances. When false, includes the health status for running instances only.

Default: false

", - "DescribeInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInternetGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeKeyPairsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeMovingAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkAclsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfacesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePlacementGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePrefixListsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRegionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$IncludeMarketplace": "

Include Reserved Instance Marketplace offerings in the response.

", - "DescribeReservedInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRouteTablesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeScheduledInstanceAvailabilityRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupReferencesRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotPriceHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeStaleSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSubnetsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointServicesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcPeeringConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DetachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$Force": "

Specifies whether to force a detachment.

", - "DetachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachVolumeRequest$Force": "

Forces detachment if the previous detachment attempt did not occur cleanly (for example, logging into an instance, unmounting the volume, and detaching normally). This option can lead to data loss or a corrupted file system. Use this option only as a last resort to detach a volume from a failed instance. The instance won't have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures.

", - "DetachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkDnsSupportResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisassociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisassociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EbsBlockDevice$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "EbsBlockDevice$Encrypted": "

Indicates whether the EBS volume is encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption.

", - "EbsInstanceBlockDevice$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EbsInstanceBlockDeviceSpecification$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EnableVolumeIORequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkDnsSupportResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "EnableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "GetConsoleOutputRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetConsoleScreenshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetConsoleScreenshotRequest$WakeUp": "

When set to true, acts as keystroke input and wakes up an instance that's in standby or \"sleep\" mode.

", - "GetPasswordDataRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "IdFormat$UseLongIds": "

Indicates whether longer IDs (17-character IDs) are enabled for the resource.

", - "Image$Public": "

Indicates whether the image has public launch permissions. The value is true if this image has public launch permissions or false if it has only implicit and explicit launch permissions.

", - "Image$EnaSupport": "

Specifies whether enhanced networking with ENA is enabled.

", - "ImportImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportInstanceLaunchSpecification$Monitoring": "

Indicates whether monitoring is enabled.

", - "ImportInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Instance$SourceDestCheck": "

Specifies whether to enable an instance launched in a VPC to perform NAT. This controls whether source/destination checking is enabled on the instance. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "Instance$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "Instance$EnaSupport": "

Specifies whether enhanced networking with ENA is enabled.

", - "InstanceNetworkInterface$SourceDestCheck": "

Indicates whether to validate network traffic to or from this network interface.

", - "InstanceNetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "InstanceNetworkInterfaceSpecification$DeleteOnTermination": "

If set to true, the interface is deleted when the instance is terminated. You can specify true only if creating a new network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to an instance you launch in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "InstancePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "LaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ModifyIdFormatRequest$UseLongIds": "

Indicate whether the resource should use longer IDs (17-character IDs).

", - "ModifyIdentityIdFormatRequest$UseLongIds": "

Indicates whether the resource should use longer IDs (17-character IDs)

", - "ModifyImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstancePlacementResult$Return": "

Is true if the request succeeds, and an error otherwise.

", - "ModifyNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySpotFleetRequestResponse$Return": "

Is true if the request succeeds, and an error otherwise.

", - "ModifyVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$ResetPolicy": "

Specify true to reset the policy document to the default policy. The default policy allows access to the service.

", - "ModifyVpcEndpointResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ModifyVpcPeeringConnectionOptionsRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MoveAddressToVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "NetworkAcl$IsDefault": "

Indicates whether this is the default network ACL for the VPC.

", - "NetworkAclEntry$Egress": "

Indicates whether the rule is an egress rule (applied to traffic leaving the subnet).

", - "NetworkInterface$RequesterManaged": "

Indicates whether the network interface is being managed by AWS.

", - "NetworkInterface$SourceDestCheck": "

Indicates whether traffic to or from the instance is validated.

", - "NetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfaceAttachmentChanges$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfacePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "PeeringConnectionOptions$AllowEgressFromLocalClassicLinkToRemoteVpc": "

If true, enables outbound communication from an EC2-Classic instance that's linked to a local VPC via ClassicLink to instances in a peer VPC.

", - "PeeringConnectionOptions$AllowEgressFromLocalVpcToRemoteClassicLink": "

If true, enables outbound communication from instances in a local VPC to an EC2-Classic instance that's linked to a peer VPC via ClassicLink.

", - "PeeringConnectionOptions$AllowDnsResolutionFromRemoteVpc": "

If true, enables a local VPC to resolve public DNS hostnames to private IP addresses when queried from instances in the peer VPC.

", - "PeeringConnectionOptionsRequest$AllowEgressFromLocalClassicLinkToRemoteVpc": "

If true, enables outbound communication from an EC2-Classic instance that's linked to a local VPC via ClassicLink to instances in a peer VPC.

", - "PeeringConnectionOptionsRequest$AllowEgressFromLocalVpcToRemoteClassicLink": "

If true, enables outbound communication from instances in a local VPC to an EC2-Classic instance that's linked to a peer VPC via ClassicLink.

", - "PeeringConnectionOptionsRequest$AllowDnsResolutionFromRemoteVpc": "

If true, enables a local VPC to resolve public DNS hostnames to private IP addresses when queried from instances in the peer VPC.

", - "PriceSchedule$Active": "

The current price schedule, as determined by the term remaining for the Reserved Instance in the listing.

A specific price schedule is always in effect, but only one price schedule can be active at any time. Take, for example, a Reserved Instance listing that has five months remaining in its term. When you specify price schedules for five months and two months, this means that schedule 1, covering the first three months of the remaining term, will be active during months 5, 4, and 3. Then schedule 2, covering the last two months of the term, will be active for months 2 and 1.

", - "PrivateIpAddressSpecification$Primary": "

Indicates whether the private IP address is the primary private IP address. Only one IP address can be designated as primary.

", - "PurchaseReservedInstancesOfferingRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "PurchaseScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RebootInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RegisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RegisterImageRequest$EnaSupport": "

Set to true to enable enhanced networking with ENA for the AMI and any instances that you launch from the AMI.

This option is supported only for HVM AMIs. Specifying this option with a PV AMI can make instances launched from the AMI unreachable.

", - "RejectVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RejectVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ReleaseAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$Egress": "

Indicates whether to replace the egress rule.

Default: If no value is specified, we replace the ingress rule.

", - "ReplaceRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceRouteTableAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReportInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotFleetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotLaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ReservedInstancesOffering$Marketplace": "

Indicates whether the offering is available through the Reserved Instance Marketplace (resale) or AWS. If it's a Reserved Instance Marketplace offering, this is true.

", - "ResetImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RestoreAddressToClassicRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RouteTableAssociation$Main": "

Indicates whether this is the main route table.

", - "RunInstancesMonitoringEnabled$Enabled": "

Indicates whether monitoring is enabled for the instance.

", - "RunInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RunInstancesRequest$DisableApiTermination": "

If you set this parameter to true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. If you set this parameter to true and then later want to be able to terminate the instance, you must first change the value of the disableApiTermination attribute to false using ModifyInstanceAttribute. Alternatively, if you set InstanceInitiatedShutdownBehavior to terminate, you can terminate the instance by running the shutdown command from the instance.

Default: false

", - "RunInstancesRequest$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "RunScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ScheduledInstanceRecurrence$OccurrenceRelativeToEnd": "

Indicates whether the occurrence is relative to the end of the specified week or month.

", - "ScheduledInstanceRecurrenceRequest$OccurrenceRelativeToEnd": "

Indicates whether the occurrence is relative to the end of the specified week or month. You can't specify this value with a daily schedule.

", - "ScheduledInstancesEbs$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "ScheduledInstancesEbs$Encrypted": "

Indicates whether the volume is encrypted. You can attached encrypted volumes only to instances that support them.

", - "ScheduledInstancesLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "ScheduledInstancesMonitoring$Enabled": "

Indicates whether monitoring is enabled.

", - "ScheduledInstancesNetworkInterface$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to instances launched in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "ScheduledInstancesNetworkInterface$DeleteOnTermination": "

Indicates whether to delete the interface when the instance is terminated.

", - "ScheduledInstancesPrivateIpAddressConfig$Primary": "

Indicates whether this is a primary IP address. Otherwise, this is a secondary IP address.

", - "Snapshot$Encrypted": "

Indicates whether the snapshot is encrypted.

", - "SpotFleetLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "SpotFleetMonitoring$Enabled": "

Enables monitoring for the instance.

Default: false

", - "SpotFleetRequestConfigData$TerminateInstancesWithExpiration": "

Indicates whether running Spot instances should be terminated when the Spot fleet request expires.

", - "StartInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$Force": "

Forces the instances to stop. The instances do not have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures. This option is not recommended for Windows instances.

Default: false

", - "Subnet$DefaultForAz": "

Indicates whether this is the default subnet for the Availability Zone.

", - "Subnet$MapPublicIpOnLaunch": "

Indicates whether instances launched in this subnet receive a public IP address.

", - "TerminateInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "UnmonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Volume$Encrypted": "

Indicates whether the volume will be encrypted.

", - "VolumeAttachment$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "Vpc$IsDefault": "

Indicates whether the VPC is the default VPC.

", - "VpcClassicLink$ClassicLinkEnabled": "

Indicates whether the VPC is enabled for ClassicLink.

", - "VpcPeeringConnectionOptionsDescription$AllowEgressFromLocalClassicLinkToRemoteVpc": "

Indicates whether a local ClassicLink connection can communicate with the peer VPC over the VPC peering connection.

", - "VpcPeeringConnectionOptionsDescription$AllowEgressFromLocalVpcToRemoteClassicLink": "

Indicates whether a local VPC can communicate with a ClassicLink connection in the peer VPC over the VPC peering connection.

", - "VpcPeeringConnectionOptionsDescription$AllowDnsResolutionFromRemoteVpc": "

Indicates whether a local VPC can resolve public DNS hostnames to private IP addresses when queried from instances in a peer VPC.

", - "VpnConnectionOptions$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

", - "VpnConnectionOptionsSpecification$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

" - } - }, - "BundleIdStringList": { - "base": null, - "refs": { - "DescribeBundleTasksRequest$BundleIds": "

One or more bundle task IDs.

Default: Describes all your bundle tasks.

" - } - }, - "BundleInstanceRequest": { - "base": "

Contains the parameters for BundleInstance.

", - "refs": { - } - }, - "BundleInstanceResult": { - "base": "

Contains the output of BundleInstance.

", - "refs": { - } - }, - "BundleTask": { - "base": "

Describes a bundle task.

", - "refs": { - "BundleInstanceResult$BundleTask": "

Information about the bundle task.

", - "BundleTaskList$member": null, - "CancelBundleTaskResult$BundleTask": "

Information about the bundle task.

" - } - }, - "BundleTaskError": { - "base": "

Describes an error for BundleInstance.

", - "refs": { - "BundleTask$BundleTaskError": "

If the task fails, a description of the error.

" - } - }, - "BundleTaskList": { - "base": null, - "refs": { - "DescribeBundleTasksResult$BundleTasks": "

Information about one or more bundle tasks.

" - } - }, - "BundleTaskState": { - "base": null, - "refs": { - "BundleTask$State": "

The state of the task.

" - } - }, - "CancelBatchErrorCode": { - "base": null, - "refs": { - "CancelSpotFleetRequestsError$Code": "

The error code.

" - } - }, - "CancelBundleTaskRequest": { - "base": "

Contains the parameters for CancelBundleTask.

", - "refs": { - } - }, - "CancelBundleTaskResult": { - "base": "

Contains the output of CancelBundleTask.

", - "refs": { - } - }, - "CancelConversionRequest": { - "base": "

Contains the parameters for CancelConversionTask.

", - "refs": { - } - }, - "CancelExportTaskRequest": { - "base": "

Contains the parameters for CancelExportTask.

", - "refs": { - } - }, - "CancelImportTaskRequest": { - "base": "

Contains the parameters for CancelImportTask.

", - "refs": { - } - }, - "CancelImportTaskResult": { - "base": "

Contains the output for CancelImportTask.

", - "refs": { - } - }, - "CancelReservedInstancesListingRequest": { - "base": "

Contains the parameters for CancelReservedInstancesListing.

", - "refs": { - } - }, - "CancelReservedInstancesListingResult": { - "base": "

Contains the output of CancelReservedInstancesListing.

", - "refs": { - } - }, - "CancelSpotFleetRequestsError": { - "base": "

Describes a Spot fleet error.

", - "refs": { - "CancelSpotFleetRequestsErrorItem$Error": "

The error.

" - } - }, - "CancelSpotFleetRequestsErrorItem": { - "base": "

Describes a Spot fleet request that was not successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsErrorSet$member": null - } - }, - "CancelSpotFleetRequestsErrorSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$UnsuccessfulFleetRequests": "

Information about the Spot fleet requests that are not successfully canceled.

" - } - }, - "CancelSpotFleetRequestsRequest": { - "base": "

Contains the parameters for CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsResponse": { - "base": "

Contains the output of CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsSuccessItem": { - "base": "

Describes a Spot fleet request that was successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsSuccessSet$member": null - } - }, - "CancelSpotFleetRequestsSuccessSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$SuccessfulFleetRequests": "

Information about the Spot fleet requests that are successfully canceled.

" - } - }, - "CancelSpotInstanceRequestState": { - "base": null, - "refs": { - "CancelledSpotInstanceRequest$State": "

The state of the Spot instance request.

" - } - }, - "CancelSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelSpotInstanceRequestsResult": { - "base": "

Contains the output of CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelledSpotInstanceRequest": { - "base": "

Describes a request to cancel a Spot instance.

", - "refs": { - "CancelledSpotInstanceRequestList$member": null - } - }, - "CancelledSpotInstanceRequestList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsResult$CancelledSpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "ClassicLinkDnsSupport": { - "base": "

Describes the ClassicLink DNS support status of a VPC.

", - "refs": { - "ClassicLinkDnsSupportList$member": null - } - }, - "ClassicLinkDnsSupportList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkDnsSupportResult$Vpcs": "

Information about the ClassicLink DNS support status of the VPCs.

" - } - }, - "ClassicLinkInstance": { - "base": "

Describes a linked EC2-Classic instance.

", - "refs": { - "ClassicLinkInstanceList$member": null - } - }, - "ClassicLinkInstanceList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesResult$Instances": "

Information about one or more linked EC2-Classic instances.

" - } - }, - "ClientData": { - "base": "

Describes the client-specific data.

", - "refs": { - "ImportImageRequest$ClientData": "

The client-specific data.

", - "ImportSnapshotRequest$ClientData": "

The client-specific data.

" - } - }, - "ConfirmProductInstanceRequest": { - "base": "

Contains the parameters for ConfirmProductInstance.

", - "refs": { - } - }, - "ConfirmProductInstanceResult": { - "base": "

Contains the output of ConfirmProductInstance.

", - "refs": { - } - }, - "ContainerFormat": { - "base": null, - "refs": { - "ExportToS3Task$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

", - "ExportToS3TaskSpecification$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

" - } - }, - "ConversionIdStringList": { - "base": null, - "refs": { - "DescribeConversionTasksRequest$ConversionTaskIds": "

One or more conversion task IDs.

" - } - }, - "ConversionTask": { - "base": "

Describes a conversion task.

", - "refs": { - "DescribeConversionTaskList$member": null, - "ImportInstanceResult$ConversionTask": "

Information about the conversion task.

", - "ImportVolumeResult$ConversionTask": "

Information about the conversion task.

" - } - }, - "ConversionTaskState": { - "base": null, - "refs": { - "ConversionTask$State": "

The state of the conversion task.

" - } - }, - "CopyImageRequest": { - "base": "

Contains the parameters for CopyImage.

", - "refs": { - } - }, - "CopyImageResult": { - "base": "

Contains the output of CopyImage.

", - "refs": { - } - }, - "CopySnapshotRequest": { - "base": "

Contains the parameters for CopySnapshot.

", - "refs": { - } - }, - "CopySnapshotResult": { - "base": "

Contains the output of CopySnapshot.

", - "refs": { - } - }, - "CreateCustomerGatewayRequest": { - "base": "

Contains the parameters for CreateCustomerGateway.

", - "refs": { - } - }, - "CreateCustomerGatewayResult": { - "base": "

Contains the output of CreateCustomerGateway.

", - "refs": { - } - }, - "CreateDhcpOptionsRequest": { - "base": "

Contains the parameters for CreateDhcpOptions.

", - "refs": { - } - }, - "CreateDhcpOptionsResult": { - "base": "

Contains the output of CreateDhcpOptions.

", - "refs": { - } - }, - "CreateFlowLogsRequest": { - "base": "

Contains the parameters for CreateFlowLogs.

", - "refs": { - } - }, - "CreateFlowLogsResult": { - "base": "

Contains the output of CreateFlowLogs.

", - "refs": { - } - }, - "CreateImageRequest": { - "base": "

Contains the parameters for CreateImage.

", - "refs": { - } - }, - "CreateImageResult": { - "base": "

Contains the output of CreateImage.

", - "refs": { - } - }, - "CreateInstanceExportTaskRequest": { - "base": "

Contains the parameters for CreateInstanceExportTask.

", - "refs": { - } - }, - "CreateInstanceExportTaskResult": { - "base": "

Contains the output for CreateInstanceExportTask.

", - "refs": { - } - }, - "CreateInternetGatewayRequest": { - "base": "

Contains the parameters for CreateInternetGateway.

", - "refs": { - } - }, - "CreateInternetGatewayResult": { - "base": "

Contains the output of CreateInternetGateway.

", - "refs": { - } - }, - "CreateKeyPairRequest": { - "base": "

Contains the parameters for CreateKeyPair.

", - "refs": { - } - }, - "CreateNatGatewayRequest": { - "base": "

Contains the parameters for CreateNatGateway.

", - "refs": { - } - }, - "CreateNatGatewayResult": { - "base": "

Contains the output of CreateNatGateway.

", - "refs": { - } - }, - "CreateNetworkAclEntryRequest": { - "base": "

Contains the parameters for CreateNetworkAclEntry.

", - "refs": { - } - }, - "CreateNetworkAclRequest": { - "base": "

Contains the parameters for CreateNetworkAcl.

", - "refs": { - } - }, - "CreateNetworkAclResult": { - "base": "

Contains the output of CreateNetworkAcl.

", - "refs": { - } - }, - "CreateNetworkInterfaceRequest": { - "base": "

Contains the parameters for CreateNetworkInterface.

", - "refs": { - } - }, - "CreateNetworkInterfaceResult": { - "base": "

Contains the output of CreateNetworkInterface.

", - "refs": { - } - }, - "CreatePlacementGroupRequest": { - "base": "

Contains the parameters for CreatePlacementGroup.

", - "refs": { - } - }, - "CreateReservedInstancesListingRequest": { - "base": "

Contains the parameters for CreateReservedInstancesListing.

", - "refs": { - } - }, - "CreateReservedInstancesListingResult": { - "base": "

Contains the output of CreateReservedInstancesListing.

", - "refs": { - } - }, - "CreateRouteRequest": { - "base": "

Contains the parameters for CreateRoute.

", - "refs": { - } - }, - "CreateRouteResult": { - "base": "

Contains the output of CreateRoute.

", - "refs": { - } - }, - "CreateRouteTableRequest": { - "base": "

Contains the parameters for CreateRouteTable.

", - "refs": { - } - }, - "CreateRouteTableResult": { - "base": "

Contains the output of CreateRouteTable.

", - "refs": { - } - }, - "CreateSecurityGroupRequest": { - "base": "

Contains the parameters for CreateSecurityGroup.

", - "refs": { - } - }, - "CreateSecurityGroupResult": { - "base": "

Contains the output of CreateSecurityGroup.

", - "refs": { - } - }, - "CreateSnapshotRequest": { - "base": "

Contains the parameters for CreateSnapshot.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSubnetRequest": { - "base": "

Contains the parameters for CreateSubnet.

", - "refs": { - } - }, - "CreateSubnetResult": { - "base": "

Contains the output of CreateSubnet.

", - "refs": { - } - }, - "CreateTagsRequest": { - "base": "

Contains the parameters for CreateTags.

", - "refs": { - } - }, - "CreateVolumePermission": { - "base": "

Describes the user or group to be added or removed from the permissions for a volume.

", - "refs": { - "CreateVolumePermissionList$member": null - } - }, - "CreateVolumePermissionList": { - "base": null, - "refs": { - "CreateVolumePermissionModifications$Add": "

Adds a specific AWS account ID or group to a volume's list of create volume permissions.

", - "CreateVolumePermissionModifications$Remove": "

Removes a specific AWS account ID or group from a volume's list of create volume permissions.

", - "DescribeSnapshotAttributeResult$CreateVolumePermissions": "

A list of permissions for creating volumes from the snapshot.

" - } - }, - "CreateVolumePermissionModifications": { - "base": "

Describes modifications to the permissions for a volume.

", - "refs": { - "ModifySnapshotAttributeRequest$CreateVolumePermission": "

A JSON representation of the snapshot attribute modification.

" - } - }, - "CreateVolumeRequest": { - "base": "

Contains the parameters for CreateVolume.

", - "refs": { - } - }, - "CreateVpcEndpointRequest": { - "base": "

Contains the parameters for CreateVpcEndpoint.

", - "refs": { - } - }, - "CreateVpcEndpointResult": { - "base": "

Contains the output of CreateVpcEndpoint.

", - "refs": { - } - }, - "CreateVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for CreateVpcPeeringConnection.

", - "refs": { - } - }, - "CreateVpcPeeringConnectionResult": { - "base": "

Contains the output of CreateVpcPeeringConnection.

", - "refs": { - } - }, - "CreateVpcRequest": { - "base": "

Contains the parameters for CreateVpc.

", - "refs": { - } - }, - "CreateVpcResult": { - "base": "

Contains the output of CreateVpc.

", - "refs": { - } - }, - "CreateVpnConnectionRequest": { - "base": "

Contains the parameters for CreateVpnConnection.

", - "refs": { - } - }, - "CreateVpnConnectionResult": { - "base": "

Contains the output of CreateVpnConnection.

", - "refs": { - } - }, - "CreateVpnConnectionRouteRequest": { - "base": "

Contains the parameters for CreateVpnConnectionRoute.

", - "refs": { - } - }, - "CreateVpnGatewayRequest": { - "base": "

Contains the parameters for CreateVpnGateway.

", - "refs": { - } - }, - "CreateVpnGatewayResult": { - "base": "

Contains the output of CreateVpnGateway.

", - "refs": { - } - }, - "CurrencyCodeValues": { - "base": null, - "refs": { - "GetHostReservationPurchasePreviewResult$CurrencyCode": "

The currency in which the totalUpfrontPrice and totalHourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "HostOffering$CurrencyCode": "

The currency of the offering.

", - "HostReservation$CurrencyCode": "

The currency in which the upfrontPrice and hourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "PriceSchedule$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "PriceScheduleSpecification$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "Purchase$CurrencyCode": "

The currency in which the UpfrontPrice and HourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "PurchaseHostReservationRequest$CurrencyCode": "

The currency in which the totalUpfrontPrice, LimitPrice, and totalHourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "PurchaseHostReservationResult$CurrencyCode": "

The currency in which the totalUpfrontPrice and totalHourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "ReservedInstanceLimitPrice$CurrencyCode": "

The currency in which the limitPrice amount is specified. At this time, the only supported currency is USD.

", - "ReservedInstances$CurrencyCode": "

The currency of the Reserved Instance. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

", - "ReservedInstancesOffering$CurrencyCode": "

The currency of the Reserved Instance offering you are purchasing. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

" - } - }, - "CustomerGateway": { - "base": "

Describes a customer gateway.

", - "refs": { - "CreateCustomerGatewayResult$CustomerGateway": "

Information about the customer gateway.

", - "CustomerGatewayList$member": null - } - }, - "CustomerGatewayIdStringList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysRequest$CustomerGatewayIds": "

One or more customer gateway IDs.

Default: Describes all your customer gateways.

" - } - }, - "CustomerGatewayList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysResult$CustomerGateways": "

Information about one or more customer gateways.

" - } - }, - "DatafeedSubscriptionState": { - "base": null, - "refs": { - "SpotDatafeedSubscription$State": "

The state of the Spot instance data feed subscription.

" - } - }, - "DateTime": { - "base": null, - "refs": { - "BundleTask$StartTime": "

The time this task started.

", - "BundleTask$UpdateTime": "

The time of the most recent update for the task.

", - "ClientData$UploadStart": "

The time that the disk upload starts.

", - "ClientData$UploadEnd": "

The time that the disk upload ends.

", - "DescribeSpotFleetRequestHistoryRequest$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$LastEvaluatedTime": "

The last date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). All records up to this time were retrieved.

If nextToken indicates that there are more results, this value is not present.

", - "DescribeSpotPriceHistoryRequest$StartTime": "

The date and time, up to the past 90 days, from which to start retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotPriceHistoryRequest$EndTime": "

The date and time, up to the current date, from which to stop retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "EbsInstanceBlockDevice$AttachTime": "

The time stamp when the attachment initiated.

", - "FlowLog$CreationTime": "

The date and time the flow log was created.

", - "GetConsoleOutputResult$Timestamp": "

The time the output was last updated.

", - "GetPasswordDataResult$Timestamp": "

The time the data was last updated.

", - "HistoryRecord$Timestamp": "

The date and time of the event, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "HostReservation$End": "

The date and time that the reservation ends.

", - "HostReservation$Start": "

The date and time that the reservation started.

", - "IdFormat$Deadline": "

The date in UTC at which you are permanently switched over to using longer IDs. If a deadline is not yet available for this resource type, this field is not returned.

", - "Instance$LaunchTime": "

The time the instance was launched.

", - "InstanceNetworkInterfaceAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "InstanceStatusDetails$ImpairedSince": "

The time when a status check failed. For an instance that was launched and impaired, this is the time when the instance was launched.

", - "InstanceStatusEvent$NotBefore": "

The earliest scheduled start time for the event.

", - "InstanceStatusEvent$NotAfter": "

The latest scheduled end time for the event.

", - "NatGateway$CreateTime": "

The date and time the NAT gateway was created.

", - "NatGateway$DeleteTime": "

The date and time the NAT gateway was deleted, if applicable.

", - "NetworkInterfaceAttachment$AttachTime": "

The timestamp indicating when the attachment initiated.

", - "ProvisionedBandwidth$RequestTime": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$ProvisionTime": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ReportInstanceStatusRequest$StartTime": "

The time at which the reported instance health state began.

", - "ReportInstanceStatusRequest$EndTime": "

The time at which the reported instance health state ended.

", - "RequestSpotInstancesRequest$ValidFrom": "

The start date of the request. If this is a one-time request, the request becomes active at this date and time and remains active until all instances launch, the request expires, or the request is canceled. If the request is persistent, the request becomes active at this date and time and remains active until it expires or is canceled.

Default: The request is effective indefinitely.

", - "RequestSpotInstancesRequest$ValidUntil": "

The end date of the request. If this is a one-time request, the request remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date and time is reached.

Default: The request is effective indefinitely.

", - "ReservedInstances$Start": "

The date and time the Reserved Instance started.

", - "ReservedInstances$End": "

The time when the Reserved Instance expires.

", - "ReservedInstancesListing$CreateDate": "

The time the listing was created.

", - "ReservedInstancesListing$UpdateDate": "

The last modified timestamp of the listing.

", - "ReservedInstancesModification$CreateDate": "

The time when the modification request was created.

", - "ReservedInstancesModification$UpdateDate": "

The time when the modification request was last updated.

", - "ReservedInstancesModification$EffectiveDate": "

The time for the modification to become effective.

", - "ScheduledInstance$PreviousSlotEndTime": "

The time that the previous schedule ended or will end.

", - "ScheduledInstance$NextSlotStartTime": "

The time for the next schedule to start.

", - "ScheduledInstance$TermStartDate": "

The start date for the Scheduled Instance.

", - "ScheduledInstance$TermEndDate": "

The end date for the Scheduled Instance.

", - "ScheduledInstance$CreateDate": "

The date when the Scheduled Instance was purchased.

", - "ScheduledInstanceAvailability$FirstSlotStartTime": "

The time period for the first schedule to start.

", - "SlotDateTimeRangeRequest$EarliestTime": "

The earliest date and time, in UTC, for the Scheduled Instance to start.

", - "SlotDateTimeRangeRequest$LatestTime": "

The latest date and time, in UTC, for the Scheduled Instance to start. This value must be later than or equal to the earliest date and at most three months in the future.

", - "SlotStartTimeRangeRequest$EarliestTime": "

The earliest date and time, in UTC, for the Scheduled Instance to start.

", - "SlotStartTimeRangeRequest$LatestTime": "

The latest date and time, in UTC, for the Scheduled Instance to start.

", - "Snapshot$StartTime": "

The time stamp when the snapshot was initiated.

", - "SpotFleetRequestConfig$CreateTime": "

The creation date and time of the request.

", - "SpotFleetRequestConfigData$ValidFrom": "

The start date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The default is to start fulfilling the request immediately.

", - "SpotFleetRequestConfigData$ValidUntil": "

The end date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). At this point, no new Spot instance requests are placed or enabled to fulfill the request.

", - "SpotInstanceRequest$ValidFrom": "

The start date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The request becomes active at this date and time.

", - "SpotInstanceRequest$ValidUntil": "

The end date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). If this is a one-time request, it remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date is reached.

", - "SpotInstanceRequest$CreateTime": "

The date and time when the Spot instance request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotInstanceStatus$UpdateTime": "

The date and time of the most recent status update, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotPrice$Timestamp": "

The date and time the request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "VgwTelemetry$LastStatusChange": "

The date and time of the last change in status.

", - "Volume$CreateTime": "

The time stamp when volume creation was initiated.

", - "VolumeAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "VolumeStatusEvent$NotBefore": "

The earliest start time of the event.

", - "VolumeStatusEvent$NotAfter": "

The latest end time of the event.

", - "VpcEndpoint$CreationTimestamp": "

The date and time the VPC endpoint was created.

", - "VpcPeeringConnection$ExpirationTime": "

The time that an unaccepted VPC peering connection will expire.

" - } - }, - "DeleteCustomerGatewayRequest": { - "base": "

Contains the parameters for DeleteCustomerGateway.

", - "refs": { - } - }, - "DeleteDhcpOptionsRequest": { - "base": "

Contains the parameters for DeleteDhcpOptions.

", - "refs": { - } - }, - "DeleteFlowLogsRequest": { - "base": "

Contains the parameters for DeleteFlowLogs.

", - "refs": { - } - }, - "DeleteFlowLogsResult": { - "base": "

Contains the output of DeleteFlowLogs.

", - "refs": { - } - }, - "DeleteInternetGatewayRequest": { - "base": "

Contains the parameters for DeleteInternetGateway.

", - "refs": { - } - }, - "DeleteKeyPairRequest": { - "base": "

Contains the parameters for DeleteKeyPair.

", - "refs": { - } - }, - "DeleteNatGatewayRequest": { - "base": "

Contains the parameters for DeleteNatGateway.

", - "refs": { - } - }, - "DeleteNatGatewayResult": { - "base": "

Contains the output of DeleteNatGateway.

", - "refs": { - } - }, - "DeleteNetworkAclEntryRequest": { - "base": "

Contains the parameters for DeleteNetworkAclEntry.

", - "refs": { - } - }, - "DeleteNetworkAclRequest": { - "base": "

Contains the parameters for DeleteNetworkAcl.

", - "refs": { - } - }, - "DeleteNetworkInterfaceRequest": { - "base": "

Contains the parameters for DeleteNetworkInterface.

", - "refs": { - } - }, - "DeletePlacementGroupRequest": { - "base": "

Contains the parameters for DeletePlacementGroup.

", - "refs": { - } - }, - "DeleteRouteRequest": { - "base": "

Contains the parameters for DeleteRoute.

", - "refs": { - } - }, - "DeleteRouteTableRequest": { - "base": "

Contains the parameters for DeleteRouteTable.

", - "refs": { - } - }, - "DeleteSecurityGroupRequest": { - "base": "

Contains the parameters for DeleteSecurityGroup.

", - "refs": { - } - }, - "DeleteSnapshotRequest": { - "base": "

Contains the parameters for DeleteSnapshot.

", - "refs": { - } - }, - "DeleteSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DeleteSpotDatafeedSubscription.

", - "refs": { - } - }, - "DeleteSubnetRequest": { - "base": "

Contains the parameters for DeleteSubnet.

", - "refs": { - } - }, - "DeleteTagsRequest": { - "base": "

Contains the parameters for DeleteTags.

", - "refs": { - } - }, - "DeleteVolumeRequest": { - "base": "

Contains the parameters for DeleteVolume.

", - "refs": { - } - }, - "DeleteVpcEndpointsRequest": { - "base": "

Contains the parameters for DeleteVpcEndpoints.

", - "refs": { - } - }, - "DeleteVpcEndpointsResult": { - "base": "

Contains the output of DeleteVpcEndpoints.

", - "refs": { - } - }, - "DeleteVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for DeleteVpcPeeringConnection.

", - "refs": { - } - }, - "DeleteVpcPeeringConnectionResult": { - "base": "

Contains the output of DeleteVpcPeeringConnection.

", - "refs": { - } - }, - "DeleteVpcRequest": { - "base": "

Contains the parameters for DeleteVpc.

", - "refs": { - } - }, - "DeleteVpnConnectionRequest": { - "base": "

Contains the parameters for DeleteVpnConnection.

", - "refs": { - } - }, - "DeleteVpnConnectionRouteRequest": { - "base": "

Contains the parameters for DeleteVpnConnectionRoute.

", - "refs": { - } - }, - "DeleteVpnGatewayRequest": { - "base": "

Contains the parameters for DeleteVpnGateway.

", - "refs": { - } - }, - "DeregisterImageRequest": { - "base": "

Contains the parameters for DeregisterImage.

", - "refs": { - } - }, - "DescribeAccountAttributesRequest": { - "base": "

Contains the parameters for DescribeAccountAttributes.

", - "refs": { - } - }, - "DescribeAccountAttributesResult": { - "base": "

Contains the output of DescribeAccountAttributes.

", - "refs": { - } - }, - "DescribeAddressesRequest": { - "base": "

Contains the parameters for DescribeAddresses.

", - "refs": { - } - }, - "DescribeAddressesResult": { - "base": "

Contains the output of DescribeAddresses.

", - "refs": { - } - }, - "DescribeAvailabilityZonesRequest": { - "base": "

Contains the parameters for DescribeAvailabilityZones.

", - "refs": { - } - }, - "DescribeAvailabilityZonesResult": { - "base": "

Contains the output of DescribeAvailabiltyZones.

", - "refs": { - } - }, - "DescribeBundleTasksRequest": { - "base": "

Contains the parameters for DescribeBundleTasks.

", - "refs": { - } - }, - "DescribeBundleTasksResult": { - "base": "

Contains the output of DescribeBundleTasks.

", - "refs": { - } - }, - "DescribeClassicLinkInstancesRequest": { - "base": "

Contains the parameters for DescribeClassicLinkInstances.

", - "refs": { - } - }, - "DescribeClassicLinkInstancesResult": { - "base": "

Contains the output of DescribeClassicLinkInstances.

", - "refs": { - } - }, - "DescribeConversionTaskList": { - "base": null, - "refs": { - "DescribeConversionTasksResult$ConversionTasks": "

Information about the conversion tasks.

" - } - }, - "DescribeConversionTasksRequest": { - "base": "

Contains the parameters for DescribeConversionTasks.

", - "refs": { - } - }, - "DescribeConversionTasksResult": { - "base": "

Contains the output for DescribeConversionTasks.

", - "refs": { - } - }, - "DescribeCustomerGatewaysRequest": { - "base": "

Contains the parameters for DescribeCustomerGateways.

", - "refs": { - } - }, - "DescribeCustomerGatewaysResult": { - "base": "

Contains the output of DescribeCustomerGateways.

", - "refs": { - } - }, - "DescribeDhcpOptionsRequest": { - "base": "

Contains the parameters for DescribeDhcpOptions.

", - "refs": { - } - }, - "DescribeDhcpOptionsResult": { - "base": "

Contains the output of DescribeDhcpOptions.

", - "refs": { - } - }, - "DescribeExportTasksRequest": { - "base": "

Contains the parameters for DescribeExportTasks.

", - "refs": { - } - }, - "DescribeExportTasksResult": { - "base": "

Contains the output for DescribeExportTasks.

", - "refs": { - } - }, - "DescribeFlowLogsRequest": { - "base": "

Contains the parameters for DescribeFlowLogs.

", - "refs": { - } - }, - "DescribeFlowLogsResult": { - "base": "

Contains the output of DescribeFlowLogs.

", - "refs": { - } - }, - "DescribeHostReservationOfferingsRequest": { - "base": null, - "refs": { - } - }, - "DescribeHostReservationOfferingsResult": { - "base": null, - "refs": { - } - }, - "DescribeHostReservationsRequest": { - "base": null, - "refs": { - } - }, - "DescribeHostReservationsResult": { - "base": null, - "refs": { - } - }, - "DescribeHostsRequest": { - "base": "

Contains the parameters for DescribeHosts.

", - "refs": { - } - }, - "DescribeHostsResult": { - "base": "

Contains the output of DescribeHosts.

", - "refs": { - } - }, - "DescribeIdFormatRequest": { - "base": "

Contains the parameters for DescribeIdFormat.

", - "refs": { - } - }, - "DescribeIdFormatResult": { - "base": "

Contains the output of DescribeIdFormat.

", - "refs": { - } - }, - "DescribeIdentityIdFormatRequest": { - "base": "

Contains the parameters for DescribeIdentityIdFormat.

", - "refs": { - } - }, - "DescribeIdentityIdFormatResult": { - "base": "

Contains the output of DescribeIdentityIdFormat.

", - "refs": { - } - }, - "DescribeImageAttributeRequest": { - "base": "

Contains the parameters for DescribeImageAttribute.

", - "refs": { - } - }, - "DescribeImagesRequest": { - "base": "

Contains the parameters for DescribeImages.

", - "refs": { - } - }, - "DescribeImagesResult": { - "base": "

Contains the output of DescribeImages.

", - "refs": { - } - }, - "DescribeImportImageTasksRequest": { - "base": "

Contains the parameters for DescribeImportImageTasks.

", - "refs": { - } - }, - "DescribeImportImageTasksResult": { - "base": "

Contains the output for DescribeImportImageTasks.

", - "refs": { - } - }, - "DescribeImportSnapshotTasksRequest": { - "base": "

Contains the parameters for DescribeImportSnapshotTasks.

", - "refs": { - } - }, - "DescribeImportSnapshotTasksResult": { - "base": "

Contains the output for DescribeImportSnapshotTasks.

", - "refs": { - } - }, - "DescribeInstanceAttributeRequest": { - "base": "

Contains the parameters for DescribeInstanceAttribute.

", - "refs": { - } - }, - "DescribeInstanceStatusRequest": { - "base": "

Contains the parameters for DescribeInstanceStatus.

", - "refs": { - } - }, - "DescribeInstanceStatusResult": { - "base": "

Contains the output of DescribeInstanceStatus.

", - "refs": { - } - }, - "DescribeInstancesRequest": { - "base": "

Contains the parameters for DescribeInstances.

", - "refs": { - } - }, - "DescribeInstancesResult": { - "base": "

Contains the output of DescribeInstances.

", - "refs": { - } - }, - "DescribeInternetGatewaysRequest": { - "base": "

Contains the parameters for DescribeInternetGateways.

", - "refs": { - } - }, - "DescribeInternetGatewaysResult": { - "base": "

Contains the output of DescribeInternetGateways.

", - "refs": { - } - }, - "DescribeKeyPairsRequest": { - "base": "

Contains the parameters for DescribeKeyPairs.

", - "refs": { - } - }, - "DescribeKeyPairsResult": { - "base": "

Contains the output of DescribeKeyPairs.

", - "refs": { - } - }, - "DescribeMovingAddressesRequest": { - "base": "

Contains the parameters for DescribeMovingAddresses.

", - "refs": { - } - }, - "DescribeMovingAddressesResult": { - "base": "

Contains the output of DescribeMovingAddresses.

", - "refs": { - } - }, - "DescribeNatGatewaysRequest": { - "base": "

Contains the parameters for DescribeNatGateways.

", - "refs": { - } - }, - "DescribeNatGatewaysResult": { - "base": "

Contains the output of DescribeNatGateways.

", - "refs": { - } - }, - "DescribeNetworkAclsRequest": { - "base": "

Contains the parameters for DescribeNetworkAcls.

", - "refs": { - } - }, - "DescribeNetworkAclsResult": { - "base": "

Contains the output of DescribeNetworkAcls.

", - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for DescribeNetworkInterfaceAttribute.

", - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeResult": { - "base": "

Contains the output of DescribeNetworkInterfaceAttribute.

", - "refs": { - } - }, - "DescribeNetworkInterfacesRequest": { - "base": "

Contains the parameters for DescribeNetworkInterfaces.

", - "refs": { - } - }, - "DescribeNetworkInterfacesResult": { - "base": "

Contains the output of DescribeNetworkInterfaces.

", - "refs": { - } - }, - "DescribePlacementGroupsRequest": { - "base": "

Contains the parameters for DescribePlacementGroups.

", - "refs": { - } - }, - "DescribePlacementGroupsResult": { - "base": "

Contains the output of DescribePlacementGroups.

", - "refs": { - } - }, - "DescribePrefixListsRequest": { - "base": "

Contains the parameters for DescribePrefixLists.

", - "refs": { - } - }, - "DescribePrefixListsResult": { - "base": "

Contains the output of DescribePrefixLists.

", - "refs": { - } - }, - "DescribeRegionsRequest": { - "base": "

Contains the parameters for DescribeRegions.

", - "refs": { - } - }, - "DescribeRegionsResult": { - "base": "

Contains the output of DescribeRegions.

", - "refs": { - } - }, - "DescribeReservedInstancesListingsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesListings.

", - "refs": { - } - }, - "DescribeReservedInstancesListingsResult": { - "base": "

Contains the output of DescribeReservedInstancesListings.

", - "refs": { - } - }, - "DescribeReservedInstancesModificationsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesModifications.

", - "refs": { - } - }, - "DescribeReservedInstancesModificationsResult": { - "base": "

Contains the output of DescribeReservedInstancesModifications.

", - "refs": { - } - }, - "DescribeReservedInstancesOfferingsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesOfferings.

", - "refs": { - } - }, - "DescribeReservedInstancesOfferingsResult": { - "base": "

Contains the output of DescribeReservedInstancesOfferings.

", - "refs": { - } - }, - "DescribeReservedInstancesRequest": { - "base": "

Contains the parameters for DescribeReservedInstances.

", - "refs": { - } - }, - "DescribeReservedInstancesResult": { - "base": "

Contains the output for DescribeReservedInstances.

", - "refs": { - } - }, - "DescribeRouteTablesRequest": { - "base": "

Contains the parameters for DescribeRouteTables.

", - "refs": { - } - }, - "DescribeRouteTablesResult": { - "base": "

Contains the output of DescribeRouteTables.

", - "refs": { - } - }, - "DescribeScheduledInstanceAvailabilityRequest": { - "base": "

Contains the parameters for DescribeScheduledInstanceAvailability.

", - "refs": { - } - }, - "DescribeScheduledInstanceAvailabilityResult": { - "base": "

Contains the output of DescribeScheduledInstanceAvailability.

", - "refs": { - } - }, - "DescribeScheduledInstancesRequest": { - "base": "

Contains the parameters for DescribeScheduledInstances.

", - "refs": { - } - }, - "DescribeScheduledInstancesResult": { - "base": "

Contains the output of DescribeScheduledInstances.

", - "refs": { - } - }, - "DescribeSecurityGroupReferencesRequest": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupReferencesResult": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupsRequest": { - "base": "

Contains the parameters for DescribeSecurityGroups.

", - "refs": { - } - }, - "DescribeSecurityGroupsResult": { - "base": "

Contains the output of DescribeSecurityGroups.

", - "refs": { - } - }, - "DescribeSnapshotAttributeRequest": { - "base": "

Contains the parameters for DescribeSnapshotAttribute.

", - "refs": { - } - }, - "DescribeSnapshotAttributeResult": { - "base": "

Contains the output of DescribeSnapshotAttribute.

", - "refs": { - } - }, - "DescribeSnapshotsRequest": { - "base": "

Contains the parameters for DescribeSnapshots.

", - "refs": { - } - }, - "DescribeSnapshotsResult": { - "base": "

Contains the output of DescribeSnapshots.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesRequest": { - "base": "

Contains the parameters for DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesResponse": { - "base": "

Contains the output of DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryResponse": { - "base": "

Contains the output of DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsResponse": { - "base": "

Contains the output of DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsResult": { - "base": "

Contains the output of DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryResult": { - "base": "

Contains the output of DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeStaleSecurityGroupsRequest": { - "base": null, - "refs": { - } - }, - "DescribeStaleSecurityGroupsResult": { - "base": null, - "refs": { - } - }, - "DescribeSubnetsRequest": { - "base": "

Contains the parameters for DescribeSubnets.

", - "refs": { - } - }, - "DescribeSubnetsResult": { - "base": "

Contains the output of DescribeSubnets.

", - "refs": { - } - }, - "DescribeTagsRequest": { - "base": "

Contains the parameters for DescribeTags.

", - "refs": { - } - }, - "DescribeTagsResult": { - "base": "

Contains the output of DescribeTags.

", - "refs": { - } - }, - "DescribeVolumeAttributeRequest": { - "base": "

Contains the parameters for DescribeVolumeAttribute.

", - "refs": { - } - }, - "DescribeVolumeAttributeResult": { - "base": "

Contains the output of DescribeVolumeAttribute.

", - "refs": { - } - }, - "DescribeVolumeStatusRequest": { - "base": "

Contains the parameters for DescribeVolumeStatus.

", - "refs": { - } - }, - "DescribeVolumeStatusResult": { - "base": "

Contains the output of DescribeVolumeStatus.

", - "refs": { - } - }, - "DescribeVolumesRequest": { - "base": "

Contains the parameters for DescribeVolumes.

", - "refs": { - } - }, - "DescribeVolumesResult": { - "base": "

Contains the output of DescribeVolumes.

", - "refs": { - } - }, - "DescribeVpcAttributeRequest": { - "base": "

Contains the parameters for DescribeVpcAttribute.

", - "refs": { - } - }, - "DescribeVpcAttributeResult": { - "base": "

Contains the output of DescribeVpcAttribute.

", - "refs": { - } - }, - "DescribeVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for DescribeVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DescribeVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of DescribeVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DescribeVpcClassicLinkRequest": { - "base": "

Contains the parameters for DescribeVpcClassicLink.

", - "refs": { - } - }, - "DescribeVpcClassicLinkResult": { - "base": "

Contains the output of DescribeVpcClassicLink.

", - "refs": { - } - }, - "DescribeVpcEndpointServicesRequest": { - "base": "

Contains the parameters for DescribeVpcEndpointServices.

", - "refs": { - } - }, - "DescribeVpcEndpointServicesResult": { - "base": "

Contains the output of DescribeVpcEndpointServices.

", - "refs": { - } - }, - "DescribeVpcEndpointsRequest": { - "base": "

Contains the parameters for DescribeVpcEndpoints.

", - "refs": { - } - }, - "DescribeVpcEndpointsResult": { - "base": "

Contains the output of DescribeVpcEndpoints.

", - "refs": { - } - }, - "DescribeVpcPeeringConnectionsRequest": { - "base": "

Contains the parameters for DescribeVpcPeeringConnections.

", - "refs": { - } - }, - "DescribeVpcPeeringConnectionsResult": { - "base": "

Contains the output of DescribeVpcPeeringConnections.

", - "refs": { - } - }, - "DescribeVpcsRequest": { - "base": "

Contains the parameters for DescribeVpcs.

", - "refs": { - } - }, - "DescribeVpcsResult": { - "base": "

Contains the output of DescribeVpcs.

", - "refs": { - } - }, - "DescribeVpnConnectionsRequest": { - "base": "

Contains the parameters for DescribeVpnConnections.

", - "refs": { - } - }, - "DescribeVpnConnectionsResult": { - "base": "

Contains the output of DescribeVpnConnections.

", - "refs": { - } - }, - "DescribeVpnGatewaysRequest": { - "base": "

Contains the parameters for DescribeVpnGateways.

", - "refs": { - } - }, - "DescribeVpnGatewaysResult": { - "base": "

Contains the output of DescribeVpnGateways.

", - "refs": { - } - }, - "DetachClassicLinkVpcRequest": { - "base": "

Contains the parameters for DetachClassicLinkVpc.

", - "refs": { - } - }, - "DetachClassicLinkVpcResult": { - "base": "

Contains the output of DetachClassicLinkVpc.

", - "refs": { - } - }, - "DetachInternetGatewayRequest": { - "base": "

Contains the parameters for DetachInternetGateway.

", - "refs": { - } - }, - "DetachNetworkInterfaceRequest": { - "base": "

Contains the parameters for DetachNetworkInterface.

", - "refs": { - } - }, - "DetachVolumeRequest": { - "base": "

Contains the parameters for DetachVolume.

", - "refs": { - } - }, - "DetachVpnGatewayRequest": { - "base": "

Contains the parameters for DetachVpnGateway.

", - "refs": { - } - }, - "DeviceType": { - "base": null, - "refs": { - "Image$RootDeviceType": "

The type of root device used by the AMI. The AMI can use an EBS volume or an instance store volume.

", - "Instance$RootDeviceType": "

The root device type used by the AMI. The AMI can use an EBS volume or an instance store volume.

" - } - }, - "DhcpConfiguration": { - "base": "

Describes a DHCP configuration option.

", - "refs": { - "DhcpConfigurationList$member": null - } - }, - "DhcpConfigurationList": { - "base": null, - "refs": { - "DhcpOptions$DhcpConfigurations": "

One or more DHCP options in the set.

" - } - }, - "DhcpConfigurationValueList": { - "base": null, - "refs": { - "DhcpConfiguration$Values": "

One or more values for the DHCP option.

" - } - }, - "DhcpOptions": { - "base": "

Describes a set of DHCP options.

", - "refs": { - "CreateDhcpOptionsResult$DhcpOptions": "

A set of DHCP options.

", - "DhcpOptionsList$member": null - } - }, - "DhcpOptionsIdStringList": { - "base": null, - "refs": { - "DescribeDhcpOptionsRequest$DhcpOptionsIds": "

The IDs of one or more DHCP options sets.

Default: Describes all your DHCP options sets.

" - } - }, - "DhcpOptionsList": { - "base": null, - "refs": { - "DescribeDhcpOptionsResult$DhcpOptions": "

Information about one or more DHCP options sets.

" - } - }, - "DisableVgwRoutePropagationRequest": { - "base": "

Contains the parameters for DisableVgwRoutePropagation.

", - "refs": { - } - }, - "DisableVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for DisableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DisableVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of DisableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DisableVpcClassicLinkRequest": { - "base": "

Contains the parameters for DisableVpcClassicLink.

", - "refs": { - } - }, - "DisableVpcClassicLinkResult": { - "base": "

Contains the output of DisableVpcClassicLink.

", - "refs": { - } - }, - "DisassociateAddressRequest": { - "base": "

Contains the parameters for DisassociateAddress.

", - "refs": { - } - }, - "DisassociateRouteTableRequest": { - "base": "

Contains the parameters for DisassociateRouteTable.

", - "refs": { - } - }, - "DiskImage": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImageList$member": null - } - }, - "DiskImageDescription": { - "base": "

Describes a disk image.

", - "refs": { - "ImportInstanceVolumeDetailItem$Image": "

The image.

", - "ImportVolumeTaskDetails$Image": "

The image.

" - } - }, - "DiskImageDetail": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImage$Image": "

Information about the disk image.

", - "ImportVolumeRequest$Image": "

The disk image.

" - } - }, - "DiskImageFormat": { - "base": null, - "refs": { - "DiskImageDescription$Format": "

The disk image format.

", - "DiskImageDetail$Format": "

The disk image format.

", - "ExportToS3Task$DiskImageFormat": "

The format for the exported image.

", - "ExportToS3TaskSpecification$DiskImageFormat": "

The format for the exported image.

" - } - }, - "DiskImageList": { - "base": null, - "refs": { - "ImportInstanceRequest$DiskImages": "

The disk image.

" - } - }, - "DiskImageVolumeDescription": { - "base": "

Describes a disk image volume.

", - "refs": { - "ImportInstanceVolumeDetailItem$Volume": "

The volume.

", - "ImportVolumeTaskDetails$Volume": "

The volume.

" - } - }, - "DomainType": { - "base": null, - "refs": { - "Address$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

", - "AllocateAddressRequest$Domain": "

Set to vpc to allocate the address for use with instances in a VPC.

Default: The address is for use with instances in EC2-Classic.

", - "AllocateAddressResult$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

" - } - }, - "Double": { - "base": null, - "refs": { - "ClientData$UploadSize": "

The size of the uploaded disk image, in GiB.

", - "PriceSchedule$Price": "

The fixed price for the term.

", - "PriceScheduleSpecification$Price": "

The fixed price for the term.

", - "PricingDetail$Price": "

The price per instance.

", - "RecurringCharge$Amount": "

The amount of the recurring charge.

", - "ReservedInstanceLimitPrice$Amount": "

Used for Reserved Instance Marketplace offerings. Specifies the limit price on the total order (instanceCount * price).

", - "SnapshotDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SnapshotTaskDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SpotFleetLaunchSpecification$WeightedCapacity": "

The number of units provided by the specified instance type. These are the same units that you chose to set the target capacity in terms (instances or a performance characteristic such as vCPUs, memory, or I/O).

If the target capacity divided by this value is not a whole number, we round the number of instances to the next whole number. If this value is not specified, the default is 1.

", - "SpotFleetRequestConfigData$FulfilledCapacity": "

The number of units fulfilled by this request compared to the set target capacity.

" - } - }, - "EbsBlockDevice": { - "base": "

Describes a block device for an EBS volume.

", - "refs": { - "BlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDevice": { - "base": "

Describes a parameter used to set up an EBS volume in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDeviceSpecification": { - "base": "

Describes information used to set up an EBS volume specified in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingSpecification$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EnableVgwRoutePropagationRequest": { - "base": "

Contains the parameters for EnableVgwRoutePropagation.

", - "refs": { - } - }, - "EnableVolumeIORequest": { - "base": "

Contains the parameters for EnableVolumeIO.

", - "refs": { - } - }, - "EnableVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for EnableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "EnableVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of EnableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "EnableVpcClassicLinkRequest": { - "base": "

Contains the parameters for EnableVpcClassicLink.

", - "refs": { - } - }, - "EnableVpcClassicLinkResult": { - "base": "

Contains the output of EnableVpcClassicLink.

", - "refs": { - } - }, - "EventCode": { - "base": null, - "refs": { - "InstanceStatusEvent$Code": "

The event code.

" - } - }, - "EventInformation": { - "base": "

Describes a Spot fleet event.

", - "refs": { - "HistoryRecord$EventInformation": "

Information about the event.

" - } - }, - "EventType": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryRequest$EventType": "

The type of events to describe. By default, all events are described.

", - "HistoryRecord$EventType": "

The event type.

  • error - Indicates an error with the Spot fleet request.

  • fleetRequestChange - Indicates a change in the status or configuration of the Spot fleet request.

  • instanceChange - Indicates that an instance was launched or terminated.

" - } - }, - "ExcessCapacityTerminationPolicy": { - "base": null, - "refs": { - "ModifySpotFleetRequestRequest$ExcessCapacityTerminationPolicy": "

Indicates whether running Spot instances should be terminated if the target capacity of the Spot fleet request is decreased below the current size of the Spot fleet.

", - "SpotFleetRequestConfigData$ExcessCapacityTerminationPolicy": "

Indicates whether running Spot instances should be terminated if the target capacity of the Spot fleet request is decreased below the current size of the Spot fleet.

" - } - }, - "ExecutableByStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ExecutableUsers": "

Scopes the images by users with explicit launch permissions. Specify an AWS account ID, self (the sender of the request), or all (public AMIs).

" - } - }, - "ExportEnvironment": { - "base": null, - "refs": { - "CreateInstanceExportTaskRequest$TargetEnvironment": "

The target virtualization environment.

", - "InstanceExportDetails$TargetEnvironment": "

The target virtualization environment.

" - } - }, - "ExportTask": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskResult$ExportTask": "

Information about the instance export task.

", - "ExportTaskList$member": null - } - }, - "ExportTaskIdStringList": { - "base": null, - "refs": { - "DescribeExportTasksRequest$ExportTaskIds": "

One or more export task IDs.

" - } - }, - "ExportTaskList": { - "base": null, - "refs": { - "DescribeExportTasksResult$ExportTasks": "

Information about the export tasks.

" - } - }, - "ExportTaskState": { - "base": null, - "refs": { - "ExportTask$State": "

The state of the export task.

" - } - }, - "ExportToS3Task": { - "base": "

Describes the format and location for an instance export task.

", - "refs": { - "ExportTask$ExportToS3Task": "

Information about the export task.

" - } - }, - "ExportToS3TaskSpecification": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskRequest$ExportToS3Task": "

The format and location for an instance export task.

" - } - }, - "Filter": { - "base": "

A filter name and value pair that is used to return a more specific list of results. Filters can be used to match a set of resources by various criteria, such as tags, attributes, or IDs.

", - "refs": { - "FilterList$member": null - } - }, - "FilterList": { - "base": null, - "refs": { - "DescribeAddressesRequest$Filters": "

One or more filters. Filter names and values are case-sensitive.

  • allocation-id - [EC2-VPC] The allocation ID for the address.

  • association-id - [EC2-VPC] The association ID for the address.

  • domain - Indicates whether the address is for use in EC2-Classic (standard) or in a VPC (vpc).

  • instance-id - The ID of the instance the address is associated with, if any.

  • network-interface-id - [EC2-VPC] The ID of the network interface that the address is associated with, if any.

  • network-interface-owner-id - The AWS account ID of the owner.

  • private-ip-address - [EC2-VPC] The private IP address associated with the Elastic IP address.

  • public-ip - The Elastic IP address.

", - "DescribeAvailabilityZonesRequest$Filters": "

One or more filters.

  • message - Information about the Availability Zone.

  • region-name - The name of the region for the Availability Zone (for example, us-east-1).

  • state - The state of the Availability Zone (available | information | impaired | unavailable).

  • zone-name - The name of the Availability Zone (for example, us-east-1a).

", - "DescribeBundleTasksRequest$Filters": "

One or more filters.

  • bundle-id - The ID of the bundle task.

  • error-code - If the task failed, the error code returned.

  • error-message - If the task failed, the error message returned.

  • instance-id - The ID of the instance.

  • progress - The level of task completion, as a percentage (for example, 20%).

  • s3-bucket - The Amazon S3 bucket to store the AMI.

  • s3-prefix - The beginning of the AMI name.

  • start-time - The time the task started (for example, 2013-09-15T17:15:20.000Z).

  • state - The state of the task (pending | waiting-for-shutdown | bundling | storing | cancelling | complete | failed).

  • update-time - The time of the most recent update for the task.

", - "DescribeClassicLinkInstancesRequest$Filters": "

One or more filters.

  • group-id - The ID of a VPC security group that's associated with the instance.

  • instance-id - The ID of the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC that the instance is linked to.

", - "DescribeConversionTasksRequest$Filters": "

One or more filters.

", - "DescribeCustomerGatewaysRequest$Filters": "

One or more filters.

  • bgp-asn - The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

  • customer-gateway-id - The ID of the customer gateway.

  • ip-address - The IP address of the customer gateway's Internet-routable external interface.

  • state - The state of the customer gateway (pending | available | deleting | deleted).

  • type - The type of customer gateway. Currently, the only supported type is ipsec.1.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeDhcpOptionsRequest$Filters": "

One or more filters.

  • dhcp-options-id - The ID of a set of DHCP options.

  • key - The key for one of the options (for example, domain-name).

  • value - The value for one of the options.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeFlowLogsRequest$Filter": "

One or more filters.

  • deliver-log-status - The status of the logs delivery (SUCCESS | FAILED).

  • flow-log-id - The ID of the flow log.

  • log-group-name - The name of the log group.

  • resource-id - The ID of the VPC, subnet, or network interface.

  • traffic-type - The type of traffic (ACCEPT | REJECT | ALL)

", - "DescribeHostReservationOfferingsRequest$Filter": "

One or more filters.

  • instance-family - The instance family of the offering (e.g., m4).

  • payment-option - The payment option (No Upfront | Partial Upfront | All Upfront).

", - "DescribeHostReservationsRequest$Filter": "

One or more filters.

  • instance-family - The instance family (e.g., m4).

  • payment-option - The payment option (No Upfront | Partial Upfront | All Upfront).

  • state - The state of the reservation (payment-pending | payment-failed | active | retired).

", - "DescribeHostsRequest$Filter": "

One or more filters.

  • instance-type - The instance type size that the Dedicated Host is configured to support.

  • auto-placement - Whether auto-placement is enabled or disabled (on | off).

  • host-reservation-id - The ID of the reservation assigned to this host.

  • client-token - The idempotency token you provided when you launched the instance

  • state- The allocation state of the Dedicated Host (available | under-assessment | permanent-failure | released | released-permanent-failure).

  • availability-zone - The Availability Zone of the host.

", - "DescribeImagesRequest$Filters": "

One or more filters.

  • architecture - The image architecture (i386 | x86_64).

  • block-device-mapping.delete-on-termination - A Boolean value that indicates whether the Amazon EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh).

  • block-device-mapping.snapshot-id - The ID of the snapshot used for the EBS volume.

  • block-device-mapping.volume-size - The volume size of the EBS volume, in GiB.

  • block-device-mapping.volume-type - The volume type of the EBS volume (gp2 | io1 | st1 | sc1 | standard).

  • description - The description of the image (provided during image creation).

  • hypervisor - The hypervisor type (ovm | xen).

  • image-id - The ID of the image.

  • image-type - The image type (machine | kernel | ramdisk).

  • is-public - A Boolean that indicates whether the image is public.

  • kernel-id - The kernel ID.

  • manifest-location - The location of the image manifest.

  • name - The name of the AMI (provided during image creation).

  • owner-alias - String value from an Amazon-maintained list (amazon | aws-marketplace | microsoft) of snapshot owners. Not to be confused with the user-configured AWS account alias, which is set from the IAM console.

  • owner-id - The AWS account ID of the image owner.

  • platform - The platform. To only list Windows-based AMIs, use windows.

  • product-code - The product code.

  • product-code.type - The type of the product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • root-device-name - The name of the root device volume (for example, /dev/sda1).

  • root-device-type - The type of the root device volume (ebs | instance-store).

  • state - The state of the image (available | pending | failed).

  • state-reason-code - The reason code for the state change.

  • state-reason-message - The message for the state change.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • virtualization-type - The virtualization type (paravirtual | hvm).

", - "DescribeImportImageTasksRequest$Filters": "

Filter tasks using the task-state filter and one of the following values: active, completed, deleting, deleted.

", - "DescribeImportSnapshotTasksRequest$Filters": "

One or more filters.

", - "DescribeInstanceStatusRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone of the instance.

  • event.code - The code for the scheduled event (instance-reboot | system-reboot | system-maintenance | instance-retirement | instance-stop).

  • event.description - A description of the event.

  • event.not-after - The latest end time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • event.not-before - The earliest start time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • instance-state-code - The code for the instance state, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-status.reachability - Filters on instance status where the name is reachability (passed | failed | initializing | insufficient-data).

  • instance-status.status - The status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

  • system-status.reachability - Filters on system status where the name is reachability (passed | failed | initializing | insufficient-data).

  • system-status.status - The system status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

", - "DescribeInstancesRequest$Filters": "

One or more filters.

  • affinity - The affinity setting for an instance running on a Dedicated Host (default | host).

  • architecture - The instance architecture (i386 | x86_64).

  • availability-zone - The Availability Zone of the instance.

  • block-device-mapping.attach-time - The attach time for an EBS volume mapped to the instance, for example, 2010-09-15T17:15:20.000Z.

  • block-device-mapping.delete-on-termination - A Boolean that indicates whether the EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh or xvdh).

  • block-device-mapping.status - The status for the EBS volume (attaching | attached | detaching | detached).

  • block-device-mapping.volume-id - The volume ID of the EBS volume.

  • client-token - The idempotency token you provided when you launched the instance.

  • dns-name - The public DNS name of the instance.

  • group-id - The ID of the security group for the instance. EC2-Classic only.

  • group-name - The name of the security group for the instance. EC2-Classic only.

  • host-id - The ID of the Dedicated Host on which the instance is running, if applicable.

  • hypervisor - The hypervisor type of the instance (ovm | xen).

  • iam-instance-profile.arn - The instance profile associated with the instance. Specified as an ARN.

  • image-id - The ID of the image used to launch the instance.

  • instance-id - The ID of the instance.

  • instance-lifecycle - Indicates whether this is a Spot Instance or a Scheduled Instance (spot | scheduled).

  • instance-state-code - The state of the instance, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are: 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-type - The type of instance (for example, t2.micro).

  • instance.group-id - The ID of the security group for the instance.

  • instance.group-name - The name of the security group for the instance.

  • ip-address - The public IP address of the instance.

  • kernel-id - The kernel ID.

  • key-name - The name of the key pair used when the instance was launched.

  • launch-index - When launching multiple instances, this is the index for the instance in the launch group (for example, 0, 1, 2, and so on).

  • launch-time - The time when the instance was launched.

  • monitoring-state - Indicates whether monitoring is enabled for the instance (disabled | enabled).

  • owner-id - The AWS account ID of the instance owner.

  • placement-group-name - The name of the placement group for the instance.

  • platform - The platform. Use windows if you have Windows instances; otherwise, leave blank.

  • private-dns-name - The private DNS name of the instance.

  • private-ip-address - The private IP address of the instance.

  • product-code - The product code associated with the AMI used to launch the instance.

  • product-code.type - The type of product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • reason - The reason for the current state of the instance (for example, shows \"User Initiated [date]\" when you stop or terminate the instance). Similar to the state-reason-code filter.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • reservation-id - The ID of the instance's reservation. A reservation ID is created any time you launch an instance. A reservation ID has a one-to-one relationship with an instance launch request, but can be associated with more than one instance if you launch multiple instances using the same launch request. For example, if you launch one instance, you'll get one reservation ID. If you launch ten instances using the same launch request, you'll also get one reservation ID.

  • root-device-name - The name of the root device for the instance (for example, /dev/sda1 or /dev/xvda).

  • root-device-type - The type of root device that the instance uses (ebs | instance-store).

  • source-dest-check - Indicates whether the instance performs source/destination checking. A value of true means that checking is enabled, and false means checking is disabled. The value must be false for the instance to perform network address translation (NAT) in your VPC.

  • spot-instance-request-id - The ID of the Spot instance request.

  • state-reason-code - The reason code for the state change.

  • state-reason-message - A message that describes the state change.

  • subnet-id - The ID of the subnet for the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource, where tag:key is the tag's key.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • tenancy - The tenancy of an instance (dedicated | default | host).

  • virtualization-type - The virtualization type of the instance (paravirtual | hvm).

  • vpc-id - The ID of the VPC that the instance is running in.

  • network-interface.description - The description of the network interface.

  • network-interface.subnet-id - The ID of the subnet for the network interface.

  • network-interface.vpc-id - The ID of the VPC for the network interface.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.owner-id - The ID of the owner of the network interface.

  • network-interface.availability-zone - The Availability Zone for the network interface.

  • network-interface.requester-id - The requester ID for the network interface.

  • network-interface.requester-managed - Indicates whether the network interface is being managed by AWS.

  • network-interface.status - The status of the network interface (available) | in-use).

  • network-interface.mac-address - The MAC address of the network interface.

  • network-interface.private-dns-name - The private DNS name of the network interface.

  • network-interface.source-dest-check - Whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • network-interface.group-id - The ID of a security group associated with the network interface.

  • network-interface.group-name - The name of a security group associated with the network interface.

  • network-interface.attachment.attachment-id - The ID of the interface attachment.

  • network-interface.attachment.instance-id - The ID of the instance to which the network interface is attached.

  • network-interface.attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • network-interface.addresses.private-ip-address - The private IP address associated with the network interface.

  • network-interface.attachment.device-index - The device index to which the network interface is attached.

  • network-interface.attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • network-interface.attachment.attach-time - The time that the network interface was attached to an instance.

  • network-interface.attachment.delete-on-termination - Specifies whether the attachment is deleted when an instance is terminated.

  • network-interface.addresses.primary - Specifies whether the IP address of the network interface is the primary private IP address.

  • network-interface.addresses.association.public-ip - The ID of the association of an Elastic IP address with a network interface.

  • network-interface.addresses.association.ip-owner-id - The owner ID of the private IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

", - "DescribeInternetGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (available). Present only if a VPC is attached.

  • attachment.vpc-id - The ID of an attached VPC.

  • internet-gateway-id - The ID of the Internet gateway.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeKeyPairsRequest$Filters": "

One or more filters.

  • fingerprint - The fingerprint of the key pair.

  • key-name - The name of the key pair.

", - "DescribeMovingAddressesRequest$Filters": "

One or more filters.

  • moving-status - The status of the Elastic IP address (MovingToVpc | RestoringToClassic).

", - "DescribeNatGatewaysRequest$Filter": "

One or more filters.

  • nat-gateway-id - The ID of the NAT gateway.

  • state - The state of the NAT gateway (pending | failed | available | deleting | deleted).

  • subnet-id - The ID of the subnet in which the NAT gateway resides.

  • vpc-id - The ID of the VPC in which the NAT gateway resides.

", - "DescribeNetworkAclsRequest$Filters": "

One or more filters.

  • association.association-id - The ID of an association ID for the ACL.

  • association.network-acl-id - The ID of the network ACL involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • default - Indicates whether the ACL is the default network ACL for the VPC.

  • entry.cidr - The CIDR range specified in the entry.

  • entry.egress - Indicates whether the entry applies to egress traffic.

  • entry.icmp.code - The ICMP code specified in the entry, if any.

  • entry.icmp.type - The ICMP type specified in the entry, if any.

  • entry.port-range.from - The start of the port range specified in the entry.

  • entry.port-range.to - The end of the port range specified in the entry.

  • entry.protocol - The protocol specified in the entry (tcp | udp | icmp or a protocol number).

  • entry.rule-action - Allows or denies the matching traffic (allow | deny).

  • entry.rule-number - The number of an entry (in other words, rule) in the ACL's set of entries.

  • network-acl-id - The ID of the network ACL.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network ACL.

", - "DescribeNetworkInterfacesRequest$Filters": "

One or more filters.

  • addresses.private-ip-address - The private IP addresses associated with the network interface.

  • addresses.primary - Whether the private IP address is the primary IP address associated with the network interface.

  • addresses.association.public-ip - The association ID returned when the network interface was associated with the Elastic IP address.

  • addresses.association.owner-id - The owner ID of the addresses associated with the network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.public-dns-name - The public DNS name for the network interface.

  • attachment.attachment-id - The ID of the interface attachment.

  • attachment.attach.time - The time that the network interface was attached to an instance.

  • attachment.delete-on-termination - Indicates whether the attachment is deleted when an instance is terminated.

  • attachment.device-index - The device index to which the network interface is attached.

  • attachment.instance-id - The ID of the instance to which the network interface is attached.

  • attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • attachment.nat-gateway-id - The ID of the NAT gateway to which the network interface is attached.

  • attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone of the network interface.

  • description - The description of the network interface.

  • group-id - The ID of a security group associated with the network interface.

  • group-name - The name of a security group associated with the network interface.

  • mac-address - The MAC address of the network interface.

  • network-interface-id - The ID of the network interface.

  • owner-id - The AWS account ID of the network interface owner.

  • private-ip-address - The private IP address or addresses of the network interface.

  • private-dns-name - The private DNS name of the network interface.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • requester-managed - Indicates whether the network interface is being managed by an AWS service (for example, AWS Management Console, Auto Scaling, and so on).

  • source-desk-check - Indicates whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • status - The status of the network interface. If the network interface is not attached to an instance, the status is available; if a network interface is attached to an instance the status is in-use.

  • subnet-id - The ID of the subnet for the network interface.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network interface.

", - "DescribePlacementGroupsRequest$Filters": "

One or more filters.

  • group-name - The name of the placement group.

  • state - The state of the placement group (pending | available | deleting | deleted).

  • strategy - The strategy of the placement group (cluster).

", - "DescribePrefixListsRequest$Filters": "

One or more filters.

  • prefix-list-id: The ID of a prefix list.

  • prefix-list-name: The name of a prefix list.

", - "DescribeRegionsRequest$Filters": "

One or more filters.

  • endpoint - The endpoint of the region (for example, ec2.us-east-1.amazonaws.com).

  • region-name - The name of the region (for example, us-east-1).

", - "DescribeReservedInstancesListingsRequest$Filters": "

One or more filters.

  • reserved-instances-id - The ID of the Reserved Instances.

  • reserved-instances-listing-id - The ID of the Reserved Instances listing.

  • status - The status of the Reserved Instance listing (pending | active | cancelled | closed).

  • status-message - The reason for the status.

", - "DescribeReservedInstancesModificationsRequest$Filters": "

One or more filters.

  • client-token - The idempotency token for the modification request.

  • create-date - The time when the modification request was created.

  • effective-date - The time when the modification becomes effective.

  • modification-result.reserved-instances-id - The ID for the Reserved Instances created as part of the modification request. This ID is only available when the status of the modification is fulfilled.

  • modification-result.target-configuration.availability-zone - The Availability Zone for the new Reserved Instances.

  • modification-result.target-configuration.instance-count - The number of new Reserved Instances.

  • modification-result.target-configuration.instance-type - The instance type of the new Reserved Instances.

  • modification-result.target-configuration.platform - The network platform of the new Reserved Instances (EC2-Classic | EC2-VPC).

  • reserved-instances-id - The ID of the Reserved Instances modified.

  • reserved-instances-modification-id - The ID of the modification request.

  • status - The status of the Reserved Instances modification request (processing | fulfilled | failed).

  • status-message - The reason for the status.

  • update-date - The time when the modification request was last updated.

", - "DescribeReservedInstancesOfferingsRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (for example, one year or three years), in seconds (31536000 | 94608000).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type that is covered by the reservation.

  • marketplace - Set to true to show only Reserved Instance Marketplace offerings. When this filter is not used, which is the default behavior, all offerings from both AWS and the Reserved Instance Marketplace are listed.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC. (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC))

  • reserved-instances-offering-id - The Reserved Instances offering ID.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeReservedInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (one year or three years), in seconds (31536000 | 94608000).

  • end - The time when the Reserved Instance expires (for example, 2015-08-07T11:54:42.000Z).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type that is covered by the reservation.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC)).

  • reserved-instances-id - The ID of the Reserved Instance.

  • start - The time at which the Reserved Instance purchase request was placed (for example, 2014-08-07T11:54:42.000Z).

  • state - The state of the Reserved Instance (payment-pending | active | payment-failed | retired).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeRouteTablesRequest$Filters": "

One or more filters.

  • association.route-table-association-id - The ID of an association ID for the route table.

  • association.route-table-id - The ID of the route table involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • association.main - Indicates whether the route table is the main route table for the VPC (true | false).

  • route-table-id - The ID of the route table.

  • route.destination-cidr-block - The CIDR range specified in a route in the table.

  • route.destination-prefix-list-id - The ID (prefix) of the AWS service specified in a route in the table.

  • route.gateway-id - The ID of a gateway specified in a route in the table.

  • route.instance-id - The ID of an instance specified in a route in the table.

  • route.nat-gateway-id - The ID of a NAT gateway.

  • route.origin - Describes how the route was created. CreateRouteTable indicates that the route was automatically created when the route table was created; CreateRoute indicates that the route was manually added to the route table; EnableVgwRoutePropagation indicates that the route was propagated by route propagation.

  • route.state - The state of a route in the route table (active | blackhole). The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, the specified NAT instance has been terminated, and so on).

  • route.vpc-peering-connection-id - The ID of a VPC peering connection specified in a route in the table.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the route table.

", - "DescribeScheduledInstanceAvailabilityRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone (for example, us-west-2a).

  • instance-type - The instance type (for example, c4.large).

  • network-platform - The network platform (EC2-Classic or EC2-VPC).

  • platform - The platform (Linux/UNIX or Windows).

", - "DescribeScheduledInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone (for example, us-west-2a).

  • instance-type - The instance type (for example, c4.large).

  • network-platform - The network platform (EC2-Classic or EC2-VPC).

  • platform - The platform (Linux/UNIX or Windows).

", - "DescribeSecurityGroupsRequest$Filters": "

One or more filters. If using multiple filters for rules, the results include security groups for which any combination of rules - not necessarily a single rule - match all filters.

  • description - The description of the security group.

  • egress.ip-permission.prefix-list-id - The ID (prefix) of the AWS service to which the security group allows access.

  • group-id - The ID of the security group.

  • group-name - The name of the security group.

  • ip-permission.cidr - A CIDR range that has been granted permission.

  • ip-permission.from-port - The start of port range for the TCP and UDP protocols, or an ICMP type number.

  • ip-permission.group-id - The ID of a security group that has been granted permission.

  • ip-permission.group-name - The name of a security group that has been granted permission.

  • ip-permission.protocol - The IP protocol for the permission (tcp | udp | icmp or a protocol number).

  • ip-permission.to-port - The end of port range for the TCP and UDP protocols, or an ICMP code.

  • ip-permission.user-id - The ID of an AWS account that has been granted permission.

  • owner-id - The AWS account ID of the owner of the security group.

  • tag-key - The key of a tag assigned to the security group.

  • tag-value - The value of a tag assigned to the security group.

  • vpc-id - The ID of the VPC specified when the security group was created.

", - "DescribeSnapshotsRequest$Filters": "

One or more filters.

  • description - A description of the snapshot.

  • owner-alias - Value from an Amazon-maintained list (amazon | aws-marketplace | microsoft) of snapshot owners. Not to be confused with the user-configured AWS account alias, which is set from the IAM consolew.

  • owner-id - The ID of the AWS account that owns the snapshot.

  • progress - The progress of the snapshot, as a percentage (for example, 80%).

  • snapshot-id - The snapshot ID.

  • start-time - The time stamp when the snapshot was initiated.

  • status - The status of the snapshot (pending | completed | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The ID of the volume the snapshot is for.

  • volume-size - The size of the volume, in GiB.

", - "DescribeSpotInstanceRequestsRequest$Filters": "

One or more filters.

  • availability-zone-group - The Availability Zone group.

  • create-time - The time stamp when the Spot instance request was created.

  • fault-code - The fault code related to the request.

  • fault-message - The fault message related to the request.

  • instance-id - The ID of the instance that fulfilled the request.

  • launch-group - The Spot instance launch group.

  • launch.block-device-mapping.delete-on-termination - Indicates whether the Amazon EBS volume is deleted on instance termination.

  • launch.block-device-mapping.device-name - The device name for the Amazon EBS volume (for example, /dev/sdh).

  • launch.block-device-mapping.snapshot-id - The ID of the snapshot used for the Amazon EBS volume.

  • launch.block-device-mapping.volume-size - The size of the Amazon EBS volume, in GiB.

  • launch.block-device-mapping.volume-type - The type of the Amazon EBS volume: gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1for Cold HDD, or standard for Magnetic.

  • launch.group-id - The security group for the instance.

  • launch.image-id - The ID of the AMI.

  • launch.instance-type - The type of instance (for example, m3.medium).

  • launch.kernel-id - The kernel ID.

  • launch.key-name - The name of the key pair the instance launched with.

  • launch.monitoring-enabled - Whether monitoring is enabled for the Spot instance.

  • launch.ramdisk-id - The RAM disk ID.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.device-index - The index of the device for the network interface attachment on the instance.

  • network-interface.subnet-id - The ID of the subnet for the instance.

  • network-interface.description - A description of the network interface.

  • network-interface.private-ip-address - The primary private IP address of the network interface.

  • network-interface.delete-on-termination - Indicates whether the network interface is deleted when the instance is terminated.

  • network-interface.group-id - The ID of the security group associated with the network interface.

  • network-interface.group-name - The name of the security group associated with the network interface.

  • network-interface.addresses.primary - Indicates whether the IP address is the primary private IP address.

  • product-description - The product description associated with the instance (Linux/UNIX | Windows).

  • spot-instance-request-id - The Spot instance request ID.

  • spot-price - The maximum hourly price for any Spot instance launched to fulfill the request.

  • state - The state of the Spot instance request (open | active | closed | cancelled | failed). Spot bid status information can help you track your Amazon EC2 Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

  • status-code - The short code describing the most recent evaluation of your Spot instance request.

  • status-message - The message explaining the status of the Spot instance request.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of Spot instance request (one-time | persistent).

  • launched-availability-zone - The Availability Zone in which the bid is launched.

  • valid-from - The start date of the request.

  • valid-until - The end date of the request.

", - "DescribeSpotPriceHistoryRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone for which prices should be returned.

  • instance-type - The type of instance (for example, m3.medium).

  • product-description - The product description for the Spot price (Linux/UNIX | SUSE Linux | Windows | Linux/UNIX (Amazon VPC) | SUSE Linux (Amazon VPC) | Windows (Amazon VPC)).

  • spot-price - The Spot price. The value must match exactly (or use wildcards; greater than or less than comparison is not supported).

  • timestamp - The timestamp of the Spot price history, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). You can use wildcards (* and ?). Greater than or less than comparison is not supported.

", - "DescribeSubnetsRequest$Filters": "

One or more filters.

  • availabilityZone - The Availability Zone for the subnet. You can also use availability-zone as the filter name.

  • available-ip-address-count - The number of IP addresses in the subnet that are available.

  • cidrBlock - The CIDR block of the subnet. The CIDR block you specify must exactly match the subnet's CIDR block for information to be returned for the subnet. You can also use cidr or cidr-block as the filter names.

  • defaultForAz - Indicates whether this is the default subnet for the Availability Zone. You can also use default-for-az as the filter name.

  • state - The state of the subnet (pending | available).

  • subnet-id - The ID of the subnet.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the subnet.

", - "DescribeTagsRequest$Filters": "

One or more filters.

  • key - The tag key.

  • resource-id - The resource ID.

  • resource-type - The resource type (customer-gateway | dhcp-options | image | instance | internet-gateway | network-acl | network-interface | reserved-instances | route-table | security-group | snapshot | spot-instances-request | subnet | volume | vpc | vpn-connection | vpn-gateway).

  • value - The tag value.

", - "DescribeVolumeStatusRequest$Filters": "

One or more filters.

  • action.code - The action code for the event (for example, enable-volume-io).

  • action.description - A description of the action.

  • action.event-id - The event ID associated with the action.

  • availability-zone - The Availability Zone of the instance.

  • event.description - A description of the event.

  • event.event-id - The event ID.

  • event.event-type - The event type (for io-enabled: passed | failed; for io-performance: io-performance:degraded | io-performance:severely-degraded | io-performance:stalled).

  • event.not-after - The latest end time for the event.

  • event.not-before - The earliest start time for the event.

  • volume-status.details-name - The cause for volume-status.status (io-enabled | io-performance).

  • volume-status.details-status - The status of volume-status.details-name (for io-enabled: passed | failed; for io-performance: normal | degraded | severely-degraded | stalled).

  • volume-status.status - The status of the volume (ok | impaired | warning | insufficient-data).

", - "DescribeVolumesRequest$Filters": "

One or more filters.

  • attachment.attach-time - The time stamp when the attachment initiated.

  • attachment.delete-on-termination - Whether the volume is deleted on instance termination.

  • attachment.device - The device name that is exposed to the instance (for example, /dev/sda1).

  • attachment.instance-id - The ID of the instance the volume is attached to.

  • attachment.status - The attachment state (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone in which the volume was created.

  • create-time - The time stamp when the volume was created.

  • encrypted - The encryption status of the volume.

  • size - The size of the volume, in GiB.

  • snapshot-id - The snapshot from which the volume was created.

  • status - The status of the volume (creating | available | in-use | deleting | deleted | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The volume ID.

  • volume-type - The Amazon EBS volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

", - "DescribeVpcClassicLinkRequest$Filters": "

One or more filters.

  • is-classic-link-enabled - Whether the VPC is enabled for ClassicLink (true | false).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeVpcEndpointsRequest$Filters": "

One or more filters.

  • service-name: The name of the AWS service.

  • vpc-id: The ID of the VPC in which the endpoint resides.

  • vpc-endpoint-id: The ID of the endpoint.

  • vpc-endpoint-state: The state of the endpoint. (pending | available | deleting | deleted)

", - "DescribeVpcPeeringConnectionsRequest$Filters": "

One or more filters.

  • accepter-vpc-info.cidr-block - The CIDR block of the peer VPC.

  • accepter-vpc-info.owner-id - The AWS account ID of the owner of the peer VPC.

  • accepter-vpc-info.vpc-id - The ID of the peer VPC.

  • expiration-time - The expiration date and time for the VPC peering connection.

  • requester-vpc-info.cidr-block - The CIDR block of the requester's VPC.

  • requester-vpc-info.owner-id - The AWS account ID of the owner of the requester VPC.

  • requester-vpc-info.vpc-id - The ID of the requester VPC.

  • status-code - The status of the VPC peering connection (pending-acceptance | failed | expired | provisioning | active | deleted | rejected).

  • status-message - A message that provides more information about the status of the VPC peering connection, if applicable.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-peering-connection-id - The ID of the VPC peering connection.

", - "DescribeVpcsRequest$Filters": "

One or more filters.

  • cidr - The CIDR block of the VPC. The CIDR block you specify must exactly match the VPC's CIDR block for information to be returned for the VPC. Must contain the slash followed by one or two digits (for example, /28).

  • dhcp-options-id - The ID of a set of DHCP options.

  • isDefault - Indicates whether the VPC is the default VPC.

  • state - The state of the VPC (pending | available).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC.

", - "DescribeVpnConnectionsRequest$Filters": "

One or more filters.

  • customer-gateway-configuration - The configuration information for the customer gateway.

  • customer-gateway-id - The ID of a customer gateway associated with the VPN connection.

  • state - The state of the VPN connection (pending | available | deleting | deleted).

  • option.static-routes-only - Indicates whether the connection has static routes only. Used for devices that do not support Border Gateway Protocol (BGP).

  • route.destination-cidr-block - The destination CIDR block. This corresponds to the subnet used in a customer data center.

  • bgp-asn - The BGP Autonomous System Number (ASN) associated with a BGP device.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of VPN connection. Currently the only supported type is ipsec.1.

  • vpn-connection-id - The ID of the VPN connection.

  • vpn-gateway-id - The ID of a virtual private gateway associated with the VPN connection.

", - "DescribeVpnGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (attaching | attached | detaching | detached).

  • attachment.vpc-id - The ID of an attached VPC.

  • availability-zone - The Availability Zone for the virtual private gateway (if applicable).

  • state - The state of the virtual private gateway (pending | available | deleting | deleted).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of virtual private gateway. Currently the only supported type is ipsec.1.

  • vpn-gateway-id - The ID of the virtual private gateway.

" - } - }, - "FleetType": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$Type": "

The type of request. Indicates whether the fleet will only request the target capacity or also attempt to maintain it. When you request a certain target capacity, the fleet will only place the required bids. It will not attempt to replenish Spot instances if capacity is diminished, nor will it submit bids in alternative Spot pools if capacity is not available. When you want to maintain a certain target capacity, fleet will place the required bids to meet this target capacity. It will also automatically replenish any interrupted instances. Default: maintain.

" - } - }, - "Float": { - "base": null, - "refs": { - "ReservedInstances$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstances$FixedPrice": "

The purchase price of the Reserved Instance.

", - "ReservedInstancesOffering$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstancesOffering$FixedPrice": "

The purchase price of the Reserved Instance.

" - } - }, - "FlowLog": { - "base": "

Describes a flow log.

", - "refs": { - "FlowLogSet$member": null - } - }, - "FlowLogSet": { - "base": null, - "refs": { - "DescribeFlowLogsResult$FlowLogs": "

Information about the flow logs.

" - } - }, - "FlowLogsResourceType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$ResourceType": "

The type of resource on which to create the flow log.

" - } - }, - "GatewayType": { - "base": null, - "refs": { - "CreateCustomerGatewayRequest$Type": "

The type of VPN connection that this customer gateway supports (ipsec.1).

", - "CreateVpnGatewayRequest$Type": "

The type of VPN connection this virtual private gateway supports.

", - "VpnConnection$Type": "

The type of VPN connection.

", - "VpnGateway$Type": "

The type of VPN connection the virtual private gateway supports.

" - } - }, - "GetConsoleOutputRequest": { - "base": "

Contains the parameters for GetConsoleOutput.

", - "refs": { - } - }, - "GetConsoleOutputResult": { - "base": "

Contains the output of GetConsoleOutput.

", - "refs": { - } - }, - "GetConsoleScreenshotRequest": { - "base": "

Contains the parameters for the request.

", - "refs": { - } - }, - "GetConsoleScreenshotResult": { - "base": "

Contains the output of the request.

", - "refs": { - } - }, - "GetHostReservationPurchasePreviewRequest": { - "base": null, - "refs": { - } - }, - "GetHostReservationPurchasePreviewResult": { - "base": null, - "refs": { - } - }, - "GetPasswordDataRequest": { - "base": "

Contains the parameters for GetPasswordData.

", - "refs": { - } - }, - "GetPasswordDataResult": { - "base": "

Contains the output of GetPasswordData.

", - "refs": { - } - }, - "GroupIdStringList": { - "base": null, - "refs": { - "AttachClassicLinkVpcRequest$Groups": "

The ID of one or more of the VPC's security groups. You cannot specify security groups from a different VPC.

", - "DescribeSecurityGroupsRequest$GroupIds": "

One or more security group IDs. Required for security groups in a nondefault VPC.

Default: Describes all your security groups.

", - "ModifyInstanceAttributeRequest$Groups": "

[EC2-VPC] Changes the security groups of the instance. You must specify at least one security group, even if it's just the default security group for the VPC. You must specify the security group ID, not the security group name.

" - } - }, - "GroupIdentifier": { - "base": "

Describes a security group.

", - "refs": { - "GroupIdentifierList$member": null - } - }, - "GroupIdentifierList": { - "base": null, - "refs": { - "ClassicLinkInstance$Groups": "

A list of security groups.

", - "DescribeNetworkInterfaceAttributeResult$Groups": "

The security groups associated with the network interface.

", - "Instance$SecurityGroups": "

One or more security groups for the instance.

", - "InstanceAttribute$Groups": "

The security groups associated with the instance.

", - "InstanceNetworkInterface$Groups": "

One or more security groups.

", - "LaunchSpecification$SecurityGroups": "

One or more security groups. When requesting instances in a VPC, you must specify the IDs of the security groups. When requesting instances in EC2-Classic, you can specify the names or the IDs of the security groups.

", - "NetworkInterface$Groups": "

Any security groups for the network interface.

", - "Reservation$Groups": "

[EC2-Classic only] One or more security groups.

", - "SpotFleetLaunchSpecification$SecurityGroups": "

One or more security groups. When requesting instances in a VPC, you must specify the IDs of the security groups. When requesting instances in EC2-Classic, you can specify the names or the IDs of the security groups.

" - } - }, - "GroupIds": { - "base": null, - "refs": { - "DescribeSecurityGroupReferencesRequest$GroupId": "

One or more security group IDs in your account.

" - } - }, - "GroupNameStringList": { - "base": null, - "refs": { - "DescribeSecurityGroupsRequest$GroupNames": "

[EC2-Classic and default VPC only] One or more security group names. You can specify either the security group name or the security group ID. For security groups in a nondefault VPC, use the group-name filter to describe security groups by name.

Default: Describes all your security groups.

", - "ModifySnapshotAttributeRequest$GroupNames": "

The group to modify for the snapshot.

" - } - }, - "HistoryRecord": { - "base": "

Describes an event in the history of the Spot fleet request.

", - "refs": { - "HistoryRecords$member": null - } - }, - "HistoryRecords": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryResponse$HistoryRecords": "

Information about the events in the history of the Spot fleet request.

" - } - }, - "Host": { - "base": "

Describes the properties of the Dedicated Host.

", - "refs": { - "HostList$member": null - } - }, - "HostInstance": { - "base": "

Describes an instance running on a Dedicated Host.

", - "refs": { - "HostInstanceList$member": null - } - }, - "HostInstanceList": { - "base": null, - "refs": { - "Host$Instances": "

The IDs and instance type that are currently running on the Dedicated Host.

" - } - }, - "HostList": { - "base": null, - "refs": { - "DescribeHostsResult$Hosts": "

Information about the Dedicated Hosts.

" - } - }, - "HostOffering": { - "base": "

Details about the Dedicated Host Reservation offering.

", - "refs": { - "HostOfferingSet$member": null - } - }, - "HostOfferingSet": { - "base": null, - "refs": { - "DescribeHostReservationOfferingsResult$OfferingSet": "

Information about the offerings.

" - } - }, - "HostProperties": { - "base": "

Describes properties of a Dedicated Host.

", - "refs": { - "Host$HostProperties": "

The hardware specifications of the Dedicated Host.

" - } - }, - "HostReservation": { - "base": "

Details about the Dedicated Host Reservation and associated Dedicated Hosts.

", - "refs": { - "HostReservationSet$member": null - } - }, - "HostReservationIdSet": { - "base": null, - "refs": { - "DescribeHostReservationsRequest$HostReservationIdSet": "

One or more host reservation IDs.

" - } - }, - "HostReservationSet": { - "base": null, - "refs": { - "DescribeHostReservationsResult$HostReservationSet": "

Details about the reservation's configuration.

" - } - }, - "HostTenancy": { - "base": null, - "refs": { - "ModifyInstancePlacementRequest$Tenancy": "

The tenancy of the instance that you are modifying.

" - } - }, - "HypervisorType": { - "base": null, - "refs": { - "Image$Hypervisor": "

The hypervisor type of the image.

", - "Instance$Hypervisor": "

The hypervisor type of the instance.

" - } - }, - "IamInstanceProfile": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "Instance$IamInstanceProfile": "

The IAM instance profile associated with the instance, if applicable.

" - } - }, - "IamInstanceProfileSpecification": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "LaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RequestSpotLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RunInstancesRequest$IamInstanceProfile": "

The IAM instance profile.

", - "SpotFleetLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "IcmpTypeCode": { - "base": "

Describes the ICMP type and code.

", - "refs": { - "CreateNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying ICMP for the protocol.

", - "NetworkAclEntry$IcmpTypeCode": "

ICMP protocol: The ICMP type and code.

", - "ReplaceNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying 1 (ICMP) for the protocol.

" - } - }, - "IdFormat": { - "base": "

Describes the ID format for a resource.

", - "refs": { - "IdFormatList$member": null - } - }, - "IdFormatList": { - "base": null, - "refs": { - "DescribeIdFormatResult$Statuses": "

Information about the ID format for the resource.

", - "DescribeIdentityIdFormatResult$Statuses": "

Information about the ID format for the resources.

" - } - }, - "Image": { - "base": "

Describes an image.

", - "refs": { - "ImageList$member": null - } - }, - "ImageAttribute": { - "base": "

Describes an image attribute.

", - "refs": { - } - }, - "ImageAttributeName": { - "base": null, - "refs": { - "DescribeImageAttributeRequest$Attribute": "

The AMI attribute.

Note: Depending on your account privileges, the blockDeviceMapping attribute may return a Client.AuthFailure error. If this happens, use DescribeImages to get information about the block device mapping for the AMI.

" - } - }, - "ImageDiskContainer": { - "base": "

Describes the disk container object for an import image task.

", - "refs": { - "ImageDiskContainerList$member": null - } - }, - "ImageDiskContainerList": { - "base": null, - "refs": { - "ImportImageRequest$DiskContainers": "

Information about the disk containers.

" - } - }, - "ImageIdStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ImageIds": "

One or more image IDs.

Default: Describes all images available to you.

" - } - }, - "ImageList": { - "base": null, - "refs": { - "DescribeImagesResult$Images": "

Information about one or more images.

" - } - }, - "ImageState": { - "base": null, - "refs": { - "Image$State": "

The current state of the AMI. If the state is available, the image is successfully registered and can be used to launch an instance.

" - } - }, - "ImageTypeValues": { - "base": null, - "refs": { - "Image$ImageType": "

The type of image.

" - } - }, - "ImportImageRequest": { - "base": "

Contains the parameters for ImportImage.

", - "refs": { - } - }, - "ImportImageResult": { - "base": "

Contains the output for ImportImage.

", - "refs": { - } - }, - "ImportImageTask": { - "base": "

Describes an import image task.

", - "refs": { - "ImportImageTaskList$member": null - } - }, - "ImportImageTaskList": { - "base": null, - "refs": { - "DescribeImportImageTasksResult$ImportImageTasks": "

A list of zero or more import image tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportInstanceLaunchSpecification": { - "base": "

Describes the launch specification for VM import.

", - "refs": { - "ImportInstanceRequest$LaunchSpecification": "

The launch specification.

" - } - }, - "ImportInstanceRequest": { - "base": "

Contains the parameters for ImportInstance.

", - "refs": { - } - }, - "ImportInstanceResult": { - "base": "

Contains the output for ImportInstance.

", - "refs": { - } - }, - "ImportInstanceTaskDetails": { - "base": "

Describes an import instance task.

", - "refs": { - "ConversionTask$ImportInstance": "

If the task is for importing an instance, this contains information about the import instance task.

" - } - }, - "ImportInstanceVolumeDetailItem": { - "base": "

Describes an import volume task.

", - "refs": { - "ImportInstanceVolumeDetailSet$member": null - } - }, - "ImportInstanceVolumeDetailSet": { - "base": null, - "refs": { - "ImportInstanceTaskDetails$Volumes": "

One or more volumes.

" - } - }, - "ImportKeyPairRequest": { - "base": "

Contains the parameters for ImportKeyPair.

", - "refs": { - } - }, - "ImportKeyPairResult": { - "base": "

Contains the output of ImportKeyPair.

", - "refs": { - } - }, - "ImportSnapshotRequest": { - "base": "

Contains the parameters for ImportSnapshot.

", - "refs": { - } - }, - "ImportSnapshotResult": { - "base": "

Contains the output for ImportSnapshot.

", - "refs": { - } - }, - "ImportSnapshotTask": { - "base": "

Describes an import snapshot task.

", - "refs": { - "ImportSnapshotTaskList$member": null - } - }, - "ImportSnapshotTaskList": { - "base": null, - "refs": { - "DescribeImportSnapshotTasksResult$ImportSnapshotTasks": "

A list of zero or more import snapshot tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportTaskIdList": { - "base": null, - "refs": { - "DescribeImportImageTasksRequest$ImportTaskIds": "

A list of import image task IDs.

", - "DescribeImportSnapshotTasksRequest$ImportTaskIds": "

A list of import snapshot task IDs.

" - } - }, - "ImportVolumeRequest": { - "base": "

Contains the parameters for ImportVolume.

", - "refs": { - } - }, - "ImportVolumeResult": { - "base": "

Contains the output for ImportVolume.

", - "refs": { - } - }, - "ImportVolumeTaskDetails": { - "base": "

Describes an import volume task.

", - "refs": { - "ConversionTask$ImportVolume": "

If the task is for importing a volume, this contains information about the import volume task.

" - } - }, - "Instance": { - "base": "

Describes an instance.

", - "refs": { - "InstanceList$member": null - } - }, - "InstanceAttribute": { - "base": "

Describes an instance attribute.

", - "refs": { - } - }, - "InstanceAttributeName": { - "base": null, - "refs": { - "DescribeInstanceAttributeRequest$Attribute": "

The instance attribute.

Note: The enaSupport attribute is not supported at this time.

", - "ModifyInstanceAttributeRequest$Attribute": "

The name of the attribute.

", - "ResetInstanceAttributeRequest$Attribute": "

The attribute to reset.

You can only reset the following attributes: kernel | ramdisk | sourceDestCheck. To change an instance attribute, use ModifyInstanceAttribute.

" - } - }, - "InstanceBlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingList$member": null - } - }, - "InstanceBlockDeviceMappingList": { - "base": null, - "refs": { - "Instance$BlockDeviceMappings": "

Any block device mapping entries for the instance.

", - "InstanceAttribute$BlockDeviceMappings": "

The block device mapping of the instance.

" - } - }, - "InstanceBlockDeviceMappingSpecification": { - "base": "

Describes a block device mapping entry.

", - "refs": { - "InstanceBlockDeviceMappingSpecificationList$member": null - } - }, - "InstanceBlockDeviceMappingSpecificationList": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$BlockDeviceMappings": "

Modifies the DeleteOnTermination attribute for volumes that are currently attached. The volume must be owned by the caller. If no value is specified for DeleteOnTermination, the default is true and the volume is deleted when the instance is terminated.

To add instance store volumes to an Amazon EBS-backed instance, you must add them when you launch the instance. For more information, see Updating the Block Device Mapping when Launching an Instance in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "InstanceCapacity": { - "base": "

Information about the instance type that the Dedicated Host supports.

", - "refs": { - "AvailableInstanceCapacityList$member": null - } - }, - "InstanceCount": { - "base": "

Describes a Reserved Instance listing state.

", - "refs": { - "InstanceCountList$member": null - } - }, - "InstanceCountList": { - "base": null, - "refs": { - "ReservedInstancesListing$InstanceCounts": "

The number of instances in this state.

" - } - }, - "InstanceExportDetails": { - "base": "

Describes an instance to export.

", - "refs": { - "ExportTask$InstanceExportDetails": "

Information about the instance to export.

" - } - }, - "InstanceIdSet": { - "base": null, - "refs": { - "RunScheduledInstancesResult$InstanceIdSet": "

The IDs of the newly launched instances.

" - } - }, - "InstanceIdStringList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesRequest$InstanceIds": "

One or more instance IDs. Must be instances linked to a VPC through ClassicLink.

", - "DescribeInstanceStatusRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

Constraints: Maximum 100 explicitly specified instance IDs.

", - "DescribeInstancesRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

", - "MonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "RebootInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "ReportInstanceStatusRequest$Instances": "

One or more instances.

", - "StartInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "StopInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "TerminateInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "UnmonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

" - } - }, - "InstanceLifecycleType": { - "base": null, - "refs": { - "Instance$InstanceLifecycle": "

Indicates whether this is a Spot instance or a Scheduled Instance.

" - } - }, - "InstanceList": { - "base": null, - "refs": { - "Reservation$Instances": "

One or more instances.

" - } - }, - "InstanceMonitoring": { - "base": "

Describes the monitoring information of the instance.

", - "refs": { - "InstanceMonitoringList$member": null - } - }, - "InstanceMonitoringList": { - "base": null, - "refs": { - "MonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

", - "UnmonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

" - } - }, - "InstanceNetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceList$member": null - } - }, - "InstanceNetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "InstanceNetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "InstancePrivateIpAddress$Association": "

The association information for an Elastic IP address for the network interface.

" - } - }, - "InstanceNetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "InstanceNetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "InstanceNetworkInterfaceList": { - "base": null, - "refs": { - "Instance$NetworkInterfaces": "

[EC2-VPC] One or more network interfaces for the instance.

" - } - }, - "InstanceNetworkInterfaceSpecification": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceSpecificationList$member": null - } - }, - "InstanceNetworkInterfaceSpecificationList": { - "base": null, - "refs": { - "LaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RequestSpotLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RunInstancesRequest$NetworkInterfaces": "

One or more network interfaces.

", - "SpotFleetLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "InstancePrivateIpAddress": { - "base": "

Describes a private IP address.

", - "refs": { - "InstancePrivateIpAddressList$member": null - } - }, - "InstancePrivateIpAddressList": { - "base": null, - "refs": { - "InstanceNetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "InstanceState": { - "base": "

Describes the current state of the instance.

", - "refs": { - "Instance$State": "

The current state of the instance.

", - "InstanceStateChange$CurrentState": "

The current state of the instance.

", - "InstanceStateChange$PreviousState": "

The previous state of the instance.

", - "InstanceStatus$InstanceState": "

The intended state of the instance. DescribeInstanceStatus requires that an instance be in the running state.

" - } - }, - "InstanceStateChange": { - "base": "

Describes an instance state change.

", - "refs": { - "InstanceStateChangeList$member": null - } - }, - "InstanceStateChangeList": { - "base": null, - "refs": { - "StartInstancesResult$StartingInstances": "

Information about one or more started instances.

", - "StopInstancesResult$StoppingInstances": "

Information about one or more stopped instances.

", - "TerminateInstancesResult$TerminatingInstances": "

Information about one or more terminated instances.

" - } - }, - "InstanceStateName": { - "base": null, - "refs": { - "InstanceState$Name": "

The current state of the instance.

" - } - }, - "InstanceStatus": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatusList$member": null - } - }, - "InstanceStatusDetails": { - "base": "

Describes the instance status.

", - "refs": { - "InstanceStatusDetailsList$member": null - } - }, - "InstanceStatusDetailsList": { - "base": null, - "refs": { - "InstanceStatusSummary$Details": "

The system instance health or application instance health.

" - } - }, - "InstanceStatusEvent": { - "base": "

Describes a scheduled event for an instance.

", - "refs": { - "InstanceStatusEventList$member": null - } - }, - "InstanceStatusEventList": { - "base": null, - "refs": { - "InstanceStatus$Events": "

Any scheduled events associated with the instance.

" - } - }, - "InstanceStatusList": { - "base": null, - "refs": { - "DescribeInstanceStatusResult$InstanceStatuses": "

One or more instance status descriptions.

" - } - }, - "InstanceStatusSummary": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatus$SystemStatus": "

Reports impaired functionality that stems from issues related to the systems that support an instance, such as hardware failures and network connectivity problems.

", - "InstanceStatus$InstanceStatus": "

Reports impaired functionality that stems from issues internal to the instance, such as impaired reachability.

" - } - }, - "InstanceType": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$InstanceType": "

The instance type that the reservation will cover (for example, m1.small). For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

", - "ImportInstanceLaunchSpecification$InstanceType": "

The instance type. For more information about the instance types that you can import, see Instance Types in the VM Import/Export User Guide.

", - "Instance$InstanceType": "

The instance type.

", - "InstanceTypeList$member": null, - "LaunchSpecification$InstanceType": "

The instance type.

", - "RequestSpotLaunchSpecification$InstanceType": "

The instance type.

", - "ReservedInstances$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$InstanceType": "

The instance type for the modified Reserved Instances.

", - "ReservedInstancesOffering$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "RunInstancesRequest$InstanceType": "

The instance type. For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

Default: m1.small

", - "SpotFleetLaunchSpecification$InstanceType": "

The instance type.

", - "SpotPrice$InstanceType": "

The instance type.

" - } - }, - "InstanceTypeList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$InstanceTypes": "

Filters the results by the specified instance types.

" - } - }, - "Integer": { - "base": null, - "refs": { - "AllocateHostsRequest$Quantity": "

The number of Dedicated Hosts you want to allocate to your account with these parameters.

", - "AssignPrivateIpAddressesRequest$SecondaryPrivateIpAddressCount": "

The number of secondary IP addresses to assign to the network interface. You can't specify this parameter when also specifying private IP addresses.

", - "AttachNetworkInterfaceRequest$DeviceIndex": "

The index of the device for the network interface attachment.

", - "AuthorizeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "AuthorizeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "AvailableCapacity$AvailableVCpus": "

The number of vCPUs available on the Dedicated Host.

", - "CreateCustomerGatewayRequest$BgpAsn": "

For devices that support BGP, the customer gateway's BGP ASN.

Default: 65000

", - "CreateNetworkAclEntryRequest$RuleNumber": "

The rule number for the entry (for example, 100). ACL entries are processed in ascending order by rule number.

Constraints: Positive integer from 1 to 32766. The range 32767 to 65535 is reserved for internal use.

", - "CreateNetworkInterfaceRequest$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses to assign to a network interface. When you specify a number of secondary IP addresses, Amazon EC2 selects these IP addresses within the subnet range. You can't specify this option and specify more than one private IP address using privateIpAddresses.

The number of IP addresses you can assign to a network interface varies by instance type. For more information, see Private IP Addresses Per ENI Per Instance Type in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListingRequest$InstanceCount": "

The number of instances that are a part of a Reserved Instance account to be listed in the Reserved Instance Marketplace. This number should be less than or equal to the instance count associated with the Reserved Instance ID specified in this call.

", - "CreateVolumeRequest$Size": "

The size of the volume, in GiBs.

Constraints: 1-16384 for gp2, 4-16384 for io1, 500-16384 for st1, 500-16384 for sc1, and 1-1024 for standard. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "CreateVolumeRequest$Iops": "

Only valid for Provisioned IOPS SSD volumes. The number of I/O operations per second (IOPS) to provision for the volume, with a maximum ratio of 30 IOPS/GiB.

Constraint: Range is 100 to 20000 for Provisioned IOPS SSD volumes

", - "DeleteNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to delete.

", - "DescribeClassicLinkInstancesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the instance IDs parameter in the same request.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeFlowLogsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the flow log IDs parameter in the same request.

", - "DescribeHostReservationOfferingsRequest$MinDuration": "

This is the minimum duration of the reservation you'd like to purchase, specified in seconds. Reservations are available in one-year and three-year terms. The number of seconds specified must be the number of seconds in a year (365x24x60x60) times one of the supported durations (1 or 3). For example, specify 31536000 for one year.

", - "DescribeHostReservationOfferingsRequest$MaxDuration": "

This is the maximum duration of the reservation you'd like to purchase, specified in seconds. Reservations are available in one-year and three-year terms. The number of seconds specified must be the number of seconds in a year (365x24x60x60) times one of the supported durations (1 or 3). For example, specify 94608000 for three years.

", - "DescribeHostReservationOfferingsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error.

", - "DescribeHostReservationsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error.

", - "DescribeHostsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error. You cannot specify this parameter and the host IDs parameter in the same request.

", - "DescribeImportImageTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeImportSnapshotTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeInstanceStatusRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter in the same call.

", - "DescribeInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter or tag filters in the same call.

", - "DescribeMovingAddressesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value outside of this range, an error is returned.

Default: If no value is provided, the default is 1000.

", - "DescribeNatGatewaysRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribePrefixListsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribeReservedInstancesOfferingsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. The maximum is 100.

Default: 100

", - "DescribeReservedInstancesOfferingsRequest$MaxInstanceCount": "

The maximum number of instances to filter when searching for offerings.

Default: 20

", - "DescribeScheduledInstanceAvailabilityRequest$MinSlotDurationInHours": "

The minimum available duration, in hours. The minimum required duration is 1,200 hours per year. For example, the minimum daily schedule is 4 hours, the minimum weekly schedule is 24 hours, and the minimum monthly schedule is 100 hours.

", - "DescribeScheduledInstanceAvailabilityRequest$MaxSlotDurationInHours": "

The maximum available duration, in hours. This value must be greater than MinSlotDurationInHours and less than 1,720.

", - "DescribeScheduledInstanceAvailabilityRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 300. The default value is 300. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeScheduledInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 300. The default value is 100. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSnapshotsRequest$MaxResults": "

The maximum number of snapshot results returned by DescribeSnapshots in paginated output. When this parameter is used, DescribeSnapshots only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeSnapshots request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeSnapshots returns all results. You cannot specify this parameter and the snapshot IDs parameter in the same request.

", - "DescribeSpotFleetInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestsRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotPriceHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeTagsRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeVolumeStatusRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumeStatus in paginated output. When this parameter is used, the request only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumeStatus returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVolumesRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumes in paginated output. When this parameter is used, DescribeVolumes only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeVolumes request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumes returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVpcEndpointServicesRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeVpcEndpointsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "EbsBlockDevice$VolumeSize": "

The size of the volume, in GiB.

Constraints: 1-16384 for General Purpose SSD (gp2), 4-16384 for Provisioned IOPS SSD (io1), 500-16384 for Throughput Optimized HDD (st1), 500-16384 for Cold HDD (sc1), and 1-1024 for Magnetic (standard) volumes. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "EbsBlockDevice$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For io1, this represents the number of IOPS that are provisioned for the volume. For gp2, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information about General Purpose SSD baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

", - "HostOffering$Duration": "

The duration of the offering (in seconds).

", - "HostProperties$Sockets": "

The number of sockets on the Dedicated Host.

", - "HostProperties$Cores": "

The number of cores on the Dedicated Host.

", - "HostProperties$TotalVCpus": "

The number of vCPUs on the Dedicated Host.

", - "HostReservation$Count": "

The number of Dedicated Hosts the reservation is associated with.

", - "HostReservation$Duration": "

The length of the reservation's term, specified in seconds. Can be 31536000 (1 year) | 94608000 (3 years).

", - "IcmpTypeCode$Type": "

The ICMP code. A value of -1 means all codes for the specified ICMP type.

", - "IcmpTypeCode$Code": "

The ICMP type. A value of -1 means all types.

", - "Instance$AmiLaunchIndex": "

The AMI launch index, which can be used to find this instance in the launch group.

", - "InstanceCapacity$AvailableCapacity": "

The number of instances that can still be launched onto the Dedicated Host.

", - "InstanceCapacity$TotalCapacity": "

The total number of instances that can be launched onto the Dedicated Host.

", - "InstanceCount$InstanceCount": "

The number of listed Reserved Instances in the state specified by the state.

", - "InstanceNetworkInterfaceAttachment$DeviceIndex": "

The index of the device on the instance for the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$DeviceIndex": "

The index of the device on the instance for the network interface attachment. If you are specifying a network interface in a RunInstances request, you must provide the device index.

", - "InstanceNetworkInterfaceSpecification$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses. You can't specify this option and specify more than one private IP address using the private IP addresses option.

", - "InstanceState$Code": "

The low byte represents the state. The high byte is an opaque internal value and should be ignored.

  • 0 : pending

  • 16 : running

  • 32 : shutting-down

  • 48 : terminated

  • 64 : stopping

  • 80 : stopped

", - "IpPermission$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "IpPermission$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code. A value of -1 indicates all ICMP codes for the specified ICMP type.

", - "ModifySpotFleetRequestRequest$TargetCapacity": "

The size of the fleet.

", - "NetworkAclEntry$RuleNumber": "

The rule number for the entry. ACL entries are processed in ascending order by rule number.

", - "NetworkInterfaceAttachment$DeviceIndex": "

The device index of the network interface attachment on the instance.

", - "OccurrenceDayRequestSet$member": null, - "OccurrenceDaySet$member": null, - "PortRange$From": "

The first port in the range.

", - "PortRange$To": "

The last port in the range.

", - "PricingDetail$Count": "

The number of reservations available for the price.

", - "Purchase$Duration": "

The duration of the reservation's term in seconds.

", - "PurchaseRequest$InstanceCount": "

The number of instances.

", - "PurchaseReservedInstancesOfferingRequest$InstanceCount": "

The number of Reserved Instances to purchase.

", - "ReplaceNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to replace.

", - "RequestSpotInstancesRequest$InstanceCount": "

The maximum number of Spot instances to launch.

Default: 1

", - "RequestSpotInstancesRequest$BlockDurationMinutes": "

The required duration for the Spot instances (also known as Spot blocks), in minutes. This value must be a multiple of 60 (60, 120, 180, 240, 300, or 360).

The duration period starts as soon as your Spot instance receives its instance ID. At the end of the duration period, Amazon EC2 marks the Spot instance for termination and provides a Spot instance termination notice, which gives the instance a two-minute warning before it terminates.

Note that you can't specify an Availability Zone group or a launch group if you specify a duration.

", - "ReservedInstances$InstanceCount": "

The number of reservations purchased.

", - "ReservedInstancesConfiguration$InstanceCount": "

The number of modified Reserved Instances.

", - "RevokeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "RevokeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "RunInstancesRequest$MinCount": "

The minimum number of instances to launch. If you specify a minimum that is more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches no instances.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 General FAQ.

", - "RunInstancesRequest$MaxCount": "

The maximum number of instances to launch. If you specify more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches the largest possible number of instances above MinCount.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 FAQ.

", - "RunScheduledInstancesRequest$InstanceCount": "

The number of instances.

Default: 1

", - "ScheduledInstance$SlotDurationInHours": "

The number of hours in the schedule.

", - "ScheduledInstance$TotalScheduledInstanceHours": "

The total number of hours for a single instance for the entire term.

", - "ScheduledInstance$InstanceCount": "

The number of instances.

", - "ScheduledInstanceAvailability$SlotDurationInHours": "

The number of hours in the schedule.

", - "ScheduledInstanceAvailability$TotalScheduledInstanceHours": "

The total number of hours for a single instance for the entire term.

", - "ScheduledInstanceAvailability$AvailableInstanceCount": "

The number of available instances.

", - "ScheduledInstanceAvailability$MinTermDurationInDays": "

The minimum term. The only possible value is 365 days.

", - "ScheduledInstanceAvailability$MaxTermDurationInDays": "

The maximum term. The only possible value is 365 days.

", - "ScheduledInstanceRecurrence$Interval": "

The interval quantity. The interval unit depends on the value of frequency. For example, every 2 weeks or every 2 months.

", - "ScheduledInstanceRecurrenceRequest$Interval": "

The interval quantity. The interval unit depends on the value of Frequency. For example, every 2 weeks or every 2 months.

", - "ScheduledInstancesEbs$VolumeSize": "

The size of the volume, in GiB.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "ScheduledInstancesEbs$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For io1 volumes, this represents the number of IOPS that are provisioned for the volume. For gp2 volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information about gp2 baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

", - "ScheduledInstancesNetworkInterface$DeviceIndex": "

The index of the device for the network interface attachment.

", - "ScheduledInstancesNetworkInterface$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses.

", - "Snapshot$VolumeSize": "

The size of the volume, in GiB.

", - "SpotFleetRequestConfigData$TargetCapacity": "

The number of units to request. You can choose to set the target capacity in terms of instances or a performance characteristic that is important to your application workload, such as vCPUs, memory, or I/O.

", - "SpotInstanceRequest$BlockDurationMinutes": "

The duration for the Spot instance, in minutes.

", - "StaleIpPermission$FromPort": "

The start of the port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "StaleIpPermission$ToPort": "

The end of the port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "Subnet$AvailableIpAddressCount": "

The number of unused IP addresses in the subnet. Note that the IP addresses for any stopped instances are considered unavailable.

", - "VgwTelemetry$AcceptedRouteCount": "

The number of accepted routes.

", - "Volume$Size": "

The size of the volume, in GiBs.

", - "Volume$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For Provisioned IOPS SSD volumes, this represents the number of IOPS that are provisioned for the volume. For General Purpose SSD volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information on General Purpose SSD baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

" - } - }, - "InternetGateway": { - "base": "

Describes an Internet gateway.

", - "refs": { - "CreateInternetGatewayResult$InternetGateway": "

Information about the Internet gateway.

", - "InternetGatewayList$member": null - } - }, - "InternetGatewayAttachment": { - "base": "

Describes the attachment of a VPC to an Internet gateway.

", - "refs": { - "InternetGatewayAttachmentList$member": null - } - }, - "InternetGatewayAttachmentList": { - "base": null, - "refs": { - "InternetGateway$Attachments": "

Any VPCs attached to the Internet gateway.

" - } - }, - "InternetGatewayList": { - "base": null, - "refs": { - "DescribeInternetGatewaysResult$InternetGateways": "

Information about one or more Internet gateways.

" - } - }, - "IpPermission": { - "base": "

Describes a security group rule.

", - "refs": { - "IpPermissionList$member": null - } - }, - "IpPermissionList": { - "base": null, - "refs": { - "AuthorizeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "AuthorizeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. Can be used to specify multiple rules in a single command.

", - "RevokeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "RevokeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. You can't specify a source security group and a CIDR IP address range.

", - "SecurityGroup$IpPermissions": "

One or more inbound rules associated with the security group.

", - "SecurityGroup$IpPermissionsEgress": "

[EC2-VPC] One or more outbound rules associated with the security group.

" - } - }, - "IpRange": { - "base": "

Describes an IP range.

", - "refs": { - "IpRangeList$member": null - } - }, - "IpRangeList": { - "base": null, - "refs": { - "IpPermission$IpRanges": "

One or more IP ranges.

" - } - }, - "IpRanges": { - "base": null, - "refs": { - "StaleIpPermission$IpRanges": "

One or more IP ranges. Not applicable for stale security group rules.

" - } - }, - "KeyNameStringList": { - "base": null, - "refs": { - "DescribeKeyPairsRequest$KeyNames": "

One or more key pair names.

Default: Describes all your key pairs.

" - } - }, - "KeyPair": { - "base": "

Describes a key pair.

", - "refs": { - } - }, - "KeyPairInfo": { - "base": "

Describes a key pair.

", - "refs": { - "KeyPairList$member": null - } - }, - "KeyPairList": { - "base": null, - "refs": { - "DescribeKeyPairsResult$KeyPairs": "

Information about one or more key pairs.

" - } - }, - "LaunchPermission": { - "base": "

Describes a launch permission.

", - "refs": { - "LaunchPermissionList$member": null - } - }, - "LaunchPermissionList": { - "base": null, - "refs": { - "ImageAttribute$LaunchPermissions": "

One or more launch permissions.

", - "LaunchPermissionModifications$Add": "

The AWS account ID to add to the list of launch permissions for the AMI.

", - "LaunchPermissionModifications$Remove": "

The AWS account ID to remove from the list of launch permissions for the AMI.

" - } - }, - "LaunchPermissionModifications": { - "base": "

Describes a launch permission modification.

", - "refs": { - "ModifyImageAttributeRequest$LaunchPermission": "

A launch permission modification.

" - } - }, - "LaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "SpotInstanceRequest$LaunchSpecification": "

Additional information for launching instances.

" - } - }, - "LaunchSpecsList": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$LaunchSpecifications": "

Information about the launch specifications for the Spot fleet request.

" - } - }, - "ListingState": { - "base": null, - "refs": { - "InstanceCount$State": "

The states of the listed Reserved Instances.

" - } - }, - "ListingStatus": { - "base": null, - "refs": { - "ReservedInstancesListing$Status": "

The status of the Reserved Instance listing.

" - } - }, - "Long": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$MinDuration": "

The minimum duration (in seconds) to filter when searching for offerings.

Default: 2592000 (1 month)

", - "DescribeReservedInstancesOfferingsRequest$MaxDuration": "

The maximum duration (in seconds) to filter when searching for offerings.

Default: 94608000 (3 years)

", - "DiskImageDescription$Size": "

The size of the disk image, in GiB.

", - "DiskImageDetail$Bytes": "

The size of the disk image, in GiB.

", - "DiskImageVolumeDescription$Size": "

The size of the volume, in GiB.

", - "ImportInstanceVolumeDetailItem$BytesConverted": "

The number of bytes converted so far.

", - "ImportVolumeTaskDetails$BytesConverted": "

The number of bytes converted so far.

", - "PriceSchedule$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "PriceScheduleSpecification$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "ReservedInstances$Duration": "

The duration of the Reserved Instance, in seconds.

", - "ReservedInstancesOffering$Duration": "

The duration of the Reserved Instance, in seconds.

", - "VolumeDetail$Size": "

The size of the volume, in GiB.

" - } - }, - "MaxResults": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeVpcClassicLinkDnsSupportRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

" - } - }, - "ModifyHostsRequest": { - "base": "

Contains the parameters for ModifyHosts.

", - "refs": { - } - }, - "ModifyHostsResult": { - "base": "

Contains the output of ModifyHosts.

", - "refs": { - } - }, - "ModifyIdFormatRequest": { - "base": "

Contains the parameters of ModifyIdFormat.

", - "refs": { - } - }, - "ModifyIdentityIdFormatRequest": { - "base": "

Contains the parameters of ModifyIdentityIdFormat.

", - "refs": { - } - }, - "ModifyImageAttributeRequest": { - "base": "

Contains the parameters for ModifyImageAttribute.

", - "refs": { - } - }, - "ModifyInstanceAttributeRequest": { - "base": "

Contains the parameters for ModifyInstanceAttribute.

", - "refs": { - } - }, - "ModifyInstancePlacementRequest": { - "base": "

Contains the parameters for ModifyInstancePlacement.

", - "refs": { - } - }, - "ModifyInstancePlacementResult": { - "base": "

Contains the output of ModifyInstancePlacement.

", - "refs": { - } - }, - "ModifyNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for ModifyNetworkInterfaceAttribute.

", - "refs": { - } - }, - "ModifyReservedInstancesRequest": { - "base": "

Contains the parameters for ModifyReservedInstances.

", - "refs": { - } - }, - "ModifyReservedInstancesResult": { - "base": "

Contains the output of ModifyReservedInstances.

", - "refs": { - } - }, - "ModifySnapshotAttributeRequest": { - "base": "

Contains the parameters for ModifySnapshotAttribute.

", - "refs": { - } - }, - "ModifySpotFleetRequestRequest": { - "base": "

Contains the parameters for ModifySpotFleetRequest.

", - "refs": { - } - }, - "ModifySpotFleetRequestResponse": { - "base": "

Contains the output of ModifySpotFleetRequest.

", - "refs": { - } - }, - "ModifySubnetAttributeRequest": { - "base": "

Contains the parameters for ModifySubnetAttribute.

", - "refs": { - } - }, - "ModifyVolumeAttributeRequest": { - "base": "

Contains the parameters for ModifyVolumeAttribute.

", - "refs": { - } - }, - "ModifyVpcAttributeRequest": { - "base": "

Contains the parameters for ModifyVpcAttribute.

", - "refs": { - } - }, - "ModifyVpcEndpointRequest": { - "base": "

Contains the parameters for ModifyVpcEndpoint.

", - "refs": { - } - }, - "ModifyVpcEndpointResult": { - "base": "

Contains the output of ModifyVpcEndpoint.

", - "refs": { - } - }, - "ModifyVpcPeeringConnectionOptionsRequest": { - "base": null, - "refs": { - } - }, - "ModifyVpcPeeringConnectionOptionsResult": { - "base": null, - "refs": { - } - }, - "MonitorInstancesRequest": { - "base": "

Contains the parameters for MonitorInstances.

", - "refs": { - } - }, - "MonitorInstancesResult": { - "base": "

Contains the output of MonitorInstances.

", - "refs": { - } - }, - "Monitoring": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "Instance$Monitoring": "

The monitoring information for the instance.

", - "InstanceMonitoring$Monitoring": "

The monitoring information.

" - } - }, - "MonitoringState": { - "base": null, - "refs": { - "Monitoring$State": "

Indicates whether monitoring is enabled for the instance.

" - } - }, - "MoveAddressToVpcRequest": { - "base": "

Contains the parameters for MoveAddressToVpc.

", - "refs": { - } - }, - "MoveAddressToVpcResult": { - "base": "

Contains the output of MoveAddressToVpc.

", - "refs": { - } - }, - "MoveStatus": { - "base": null, - "refs": { - "MovingAddressStatus$MoveStatus": "

The status of the Elastic IP address that's being moved to the EC2-VPC platform, or restored to the EC2-Classic platform.

" - } - }, - "MovingAddressStatus": { - "base": "

Describes the status of a moving Elastic IP address.

", - "refs": { - "MovingAddressStatusSet$member": null - } - }, - "MovingAddressStatusSet": { - "base": null, - "refs": { - "DescribeMovingAddressesResult$MovingAddressStatuses": "

The status for each Elastic IP address.

" - } - }, - "NatGateway": { - "base": "

Describes a NAT gateway.

", - "refs": { - "CreateNatGatewayResult$NatGateway": "

Information about the NAT gateway.

", - "NatGatewayList$member": null - } - }, - "NatGatewayAddress": { - "base": "

Describes the IP addresses and network interface associated with a NAT gateway.

", - "refs": { - "NatGatewayAddressList$member": null - } - }, - "NatGatewayAddressList": { - "base": null, - "refs": { - "NatGateway$NatGatewayAddresses": "

Information about the IP addresses and network interface associated with the NAT gateway.

" - } - }, - "NatGatewayList": { - "base": null, - "refs": { - "DescribeNatGatewaysResult$NatGateways": "

Information about the NAT gateways.

" - } - }, - "NatGatewayState": { - "base": null, - "refs": { - "NatGateway$State": "

The state of the NAT gateway.

  • pending: The NAT gateway is being created and is not ready to process traffic.

  • failed: The NAT gateway could not be created. Check the failureCode and failureMessage fields for the reason.

  • available: The NAT gateway is able to process traffic. This status remains until you delete the NAT gateway, and does not indicate the health of the NAT gateway.

  • deleting: The NAT gateway is in the process of being terminated and may still be processing traffic.

  • deleted: The NAT gateway has been terminated and is no longer processing traffic.

" - } - }, - "NetworkAcl": { - "base": "

Describes a network ACL.

", - "refs": { - "CreateNetworkAclResult$NetworkAcl": "

Information about the network ACL.

", - "NetworkAclList$member": null - } - }, - "NetworkAclAssociation": { - "base": "

Describes an association between a network ACL and a subnet.

", - "refs": { - "NetworkAclAssociationList$member": null - } - }, - "NetworkAclAssociationList": { - "base": null, - "refs": { - "NetworkAcl$Associations": "

Any associations between the network ACL and one or more subnets

" - } - }, - "NetworkAclEntry": { - "base": "

Describes an entry in a network ACL.

", - "refs": { - "NetworkAclEntryList$member": null - } - }, - "NetworkAclEntryList": { - "base": null, - "refs": { - "NetworkAcl$Entries": "

One or more entries (rules) in the network ACL.

" - } - }, - "NetworkAclList": { - "base": null, - "refs": { - "DescribeNetworkAclsResult$NetworkAcls": "

Information about one or more network ACLs.

" - } - }, - "NetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "CreateNetworkInterfaceResult$NetworkInterface": "

Information about the network interface.

", - "NetworkInterfaceList$member": null - } - }, - "NetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "NetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "NetworkInterfacePrivateIpAddress$Association": "

The association information for an Elastic IP address associated with the network interface.

" - } - }, - "NetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Attachment": "

The attachment (if any) of the network interface.

", - "NetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "NetworkInterfaceAttachmentChanges": { - "base": "

Describes an attachment change.

", - "refs": { - "ModifyNetworkInterfaceAttributeRequest$Attachment": "

Information about the interface attachment. If modifying the 'delete on termination' attribute, you must specify the ID of the interface attachment.

" - } - }, - "NetworkInterfaceAttribute": { - "base": null, - "refs": { - "DescribeNetworkInterfaceAttributeRequest$Attribute": "

The attribute of the network interface.

" - } - }, - "NetworkInterfaceIdList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesRequest$NetworkInterfaceIds": "

One or more network interface IDs.

Default: Describes all your network interfaces.

" - } - }, - "NetworkInterfaceList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesResult$NetworkInterfaces": "

Information about one or more network interfaces.

" - } - }, - "NetworkInterfacePrivateIpAddress": { - "base": "

Describes the private IP address of a network interface.

", - "refs": { - "NetworkInterfacePrivateIpAddressList$member": null - } - }, - "NetworkInterfacePrivateIpAddressList": { - "base": null, - "refs": { - "NetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "NetworkInterfaceStatus": { - "base": null, - "refs": { - "InstanceNetworkInterface$Status": "

The status of the network interface.

", - "NetworkInterface$Status": "

The status of the network interface.

" - } - }, - "NetworkInterfaceType": { - "base": null, - "refs": { - "NetworkInterface$InterfaceType": "

The type of interface.

" - } - }, - "NewDhcpConfiguration": { - "base": null, - "refs": { - "NewDhcpConfigurationList$member": null - } - }, - "NewDhcpConfigurationList": { - "base": null, - "refs": { - "CreateDhcpOptionsRequest$DhcpConfigurations": "

A DHCP configuration option.

" - } - }, - "NextToken": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcClassicLinkDnsSupportRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcClassicLinkDnsSupportResult$NextToken": "

The token to use when requesting the next set of items.

" - } - }, - "OccurrenceDayRequestSet": { - "base": null, - "refs": { - "ScheduledInstanceRecurrenceRequest$OccurrenceDays": "

The days. For a monthly schedule, this is one or more days of the month (1-31). For a weekly schedule, this is one or more days of the week (1-7, where 1 is Sunday). You can't specify this value with a daily schedule. If the occurrence is relative to the end of the month, you can specify only a single day.

" - } - }, - "OccurrenceDaySet": { - "base": null, - "refs": { - "ScheduledInstanceRecurrence$OccurrenceDaySet": "

The days. For a monthly schedule, this is one or more days of the month (1-31). For a weekly schedule, this is one or more days of the week (1-7, where 1 is Sunday).

" - } - }, - "OfferingTypeValues": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "DescribeReservedInstancesRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "ReservedInstances$OfferingType": "

The Reserved Instance offering type.

", - "ReservedInstancesOffering$OfferingType": "

The Reserved Instance offering type.

" - } - }, - "OperationType": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$OperationType": "

The operation type.

", - "ModifySnapshotAttributeRequest$OperationType": "

The type of operation to perform to the attribute.

" - } - }, - "OwnerStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$Owners": "

Filters the images by the owner. Specify an AWS account ID, self (owner is the sender of the request), or an AWS owner alias (valid values are amazon | aws-marketplace | microsoft). Omitting this option returns all images for which you have launch permissions, regardless of ownership.

", - "DescribeSnapshotsRequest$OwnerIds": "

Returns the snapshots owned by the specified owner. Multiple owners can be specified.

" - } - }, - "PaymentOption": { - "base": null, - "refs": { - "HostOffering$PaymentOption": "

The available payment option.

", - "HostReservation$PaymentOption": "

The payment option selected for this reservation.

", - "Purchase$PaymentOption": "

The payment option for the reservation.

" - } - }, - "PeeringConnectionOptions": { - "base": "

Describes the VPC peering connection options.

", - "refs": { - "ModifyVpcPeeringConnectionOptionsResult$RequesterPeeringConnectionOptions": "

Information about the VPC peering connection options for the requester VPC.

", - "ModifyVpcPeeringConnectionOptionsResult$AccepterPeeringConnectionOptions": "

Information about the VPC peering connection options for the accepter VPC.

" - } - }, - "PeeringConnectionOptionsRequest": { - "base": "

The VPC peering connection options.

", - "refs": { - "ModifyVpcPeeringConnectionOptionsRequest$RequesterPeeringConnectionOptions": "

The VPC peering connection options for the requester VPC.

", - "ModifyVpcPeeringConnectionOptionsRequest$AccepterPeeringConnectionOptions": "

The VPC peering connection options for the accepter VPC.

" - } - }, - "PermissionGroup": { - "base": null, - "refs": { - "CreateVolumePermission$Group": "

The specific group that is to be added or removed from a volume's list of create volume permissions.

", - "LaunchPermission$Group": "

The name of the group.

" - } - }, - "Placement": { - "base": "

Describes the placement for the instance.

", - "refs": { - "ImportInstanceLaunchSpecification$Placement": "

The placement information for the instance.

", - "Instance$Placement": "

The location where the instance launched, if applicable.

", - "RunInstancesRequest$Placement": "

The placement for the instance.

" - } - }, - "PlacementGroup": { - "base": "

Describes a placement group.

", - "refs": { - "PlacementGroupList$member": null - } - }, - "PlacementGroupList": { - "base": null, - "refs": { - "DescribePlacementGroupsResult$PlacementGroups": "

One or more placement groups.

" - } - }, - "PlacementGroupState": { - "base": null, - "refs": { - "PlacementGroup$State": "

The state of the placement group.

" - } - }, - "PlacementGroupStringList": { - "base": null, - "refs": { - "DescribePlacementGroupsRequest$GroupNames": "

One or more placement group names.

Default: Describes all your placement groups, or only those otherwise specified.

" - } - }, - "PlacementStrategy": { - "base": null, - "refs": { - "CreatePlacementGroupRequest$Strategy": "

The placement strategy.

", - "PlacementGroup$Strategy": "

The placement strategy.

" - } - }, - "PlatformValues": { - "base": null, - "refs": { - "Image$Platform": "

The value is Windows for Windows AMIs; otherwise blank.

", - "ImportInstanceRequest$Platform": "

The instance operating system.

", - "ImportInstanceTaskDetails$Platform": "

The instance operating system.

", - "Instance$Platform": "

The value is Windows for Windows instances; otherwise blank.

" - } - }, - "PortRange": { - "base": "

Describes a range of ports.

", - "refs": { - "CreateNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "NetworkAclEntry$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "ReplaceNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to. Required if specifying 6 (TCP) or 17 (UDP) for the protocol.

" - } - }, - "PrefixList": { - "base": "

Describes prefixes for AWS services.

", - "refs": { - "PrefixListSet$member": null - } - }, - "PrefixListId": { - "base": "

The ID of the prefix.

", - "refs": { - "PrefixListIdList$member": null - } - }, - "PrefixListIdList": { - "base": null, - "refs": { - "IpPermission$PrefixListIds": "

(Valid for AuthorizeSecurityGroupEgress, RevokeSecurityGroupEgress and DescribeSecurityGroups only) One or more prefix list IDs for an AWS service. In an AuthorizeSecurityGroupEgress request, this is the AWS service that you want to access through a VPC endpoint from instances associated with the security group.

" - } - }, - "PrefixListIdSet": { - "base": null, - "refs": { - "StaleIpPermission$PrefixListIds": "

One or more prefix list IDs for an AWS service. Not applicable for stale security group rules.

" - } - }, - "PrefixListSet": { - "base": null, - "refs": { - "DescribePrefixListsResult$PrefixLists": "

All available prefix lists.

" - } - }, - "PriceSchedule": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleList$member": null - } - }, - "PriceScheduleList": { - "base": null, - "refs": { - "ReservedInstancesListing$PriceSchedules": "

The price of the Reserved Instance listing.

" - } - }, - "PriceScheduleSpecification": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleSpecificationList$member": null - } - }, - "PriceScheduleSpecificationList": { - "base": null, - "refs": { - "CreateReservedInstancesListingRequest$PriceSchedules": "

A list specifying the price of the Reserved Instance for each month remaining in the Reserved Instance term.

" - } - }, - "PricingDetail": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "PricingDetailsList$member": null - } - }, - "PricingDetailsList": { - "base": null, - "refs": { - "ReservedInstancesOffering$PricingDetails": "

The pricing details of the Reserved Instance offering.

" - } - }, - "PrivateIpAddressConfigSet": { - "base": null, - "refs": { - "ScheduledInstancesNetworkInterface$PrivateIpAddressConfigs": "

The private IP addresses.

" - } - }, - "PrivateIpAddressSpecification": { - "base": "

Describes a secondary private IP address for a network interface.

", - "refs": { - "PrivateIpAddressSpecificationList$member": null - } - }, - "PrivateIpAddressSpecificationList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$PrivateIpAddresses": "

One or more private IP addresses.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddresses": "

One or more private IP addresses to assign to the network interface. Only one private IP address can be designated as primary.

" - } - }, - "PrivateIpAddressStringList": { - "base": null, - "refs": { - "AssignPrivateIpAddressesRequest$PrivateIpAddresses": "

One or more IP addresses to be assigned as a secondary private IP address to the network interface. You can't specify this parameter when also specifying a number of secondary IP addresses.

If you don't specify an IP address, Amazon EC2 automatically selects an IP address within the subnet range.

", - "UnassignPrivateIpAddressesRequest$PrivateIpAddresses": "

The secondary private IP addresses to unassign from the network interface. You can specify this option multiple times to unassign more than one IP address.

" - } - }, - "ProductCode": { - "base": "

Describes a product code.

", - "refs": { - "ProductCodeList$member": null - } - }, - "ProductCodeList": { - "base": null, - "refs": { - "DescribeSnapshotAttributeResult$ProductCodes": "

A list of product codes.

", - "DescribeVolumeAttributeResult$ProductCodes": "

A list of product codes.

", - "Image$ProductCodes": "

Any product codes associated with the AMI.

", - "ImageAttribute$ProductCodes": "

One or more product codes.

", - "Instance$ProductCodes": "

The product codes attached to this instance, if applicable.

", - "InstanceAttribute$ProductCodes": "

A list of product codes.

" - } - }, - "ProductCodeStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$ProductCodes": "

One or more product codes. After you add a product code to an AMI, it can't be removed. This is only valid when modifying the productCodes attribute.

" - } - }, - "ProductCodeValues": { - "base": null, - "refs": { - "ProductCode$ProductCodeType": "

The type of product code.

" - } - }, - "ProductDescriptionList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$ProductDescriptions": "

Filters the results by the specified basic product descriptions.

" - } - }, - "PropagatingVgw": { - "base": "

Describes a virtual private gateway propagating route.

", - "refs": { - "PropagatingVgwList$member": null - } - }, - "PropagatingVgwList": { - "base": null, - "refs": { - "RouteTable$PropagatingVgws": "

Any virtual private gateway (VGW) propagating routes.

" - } - }, - "ProvisionedBandwidth": { - "base": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "refs": { - "NatGateway$ProvisionedBandwidth": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

" - } - }, - "PublicIpStringList": { - "base": null, - "refs": { - "DescribeAddressesRequest$PublicIps": "

[EC2-Classic] One or more Elastic IP addresses.

Default: Describes all your Elastic IP addresses.

" - } - }, - "Purchase": { - "base": "

Describes the result of the purchase.

", - "refs": { - "PurchaseSet$member": null - } - }, - "PurchaseHostReservationRequest": { - "base": null, - "refs": { - } - }, - "PurchaseHostReservationResult": { - "base": null, - "refs": { - } - }, - "PurchaseRequest": { - "base": "

Describes a request to purchase Scheduled Instances.

", - "refs": { - "PurchaseRequestSet$member": null - } - }, - "PurchaseRequestSet": { - "base": null, - "refs": { - "PurchaseScheduledInstancesRequest$PurchaseRequests": "

One or more purchase requests.

" - } - }, - "PurchaseReservedInstancesOfferingRequest": { - "base": "

Contains the parameters for PurchaseReservedInstancesOffering.

", - "refs": { - } - }, - "PurchaseReservedInstancesOfferingResult": { - "base": "

Contains the output of PurchaseReservedInstancesOffering.

", - "refs": { - } - }, - "PurchaseScheduledInstancesRequest": { - "base": "

Contains the parameters for PurchaseScheduledInstances.

", - "refs": { - } - }, - "PurchaseScheduledInstancesResult": { - "base": "

Contains the output of PurchaseScheduledInstances.

", - "refs": { - } - }, - "PurchaseSet": { - "base": null, - "refs": { - "GetHostReservationPurchasePreviewResult$Purchase": "

The purchase information of the Dedicated Host Reservation and the Dedicated Hosts associated with it.

", - "PurchaseHostReservationResult$Purchase": "

Describes the details of the purchase.

" - } - }, - "PurchasedScheduledInstanceSet": { - "base": null, - "refs": { - "PurchaseScheduledInstancesResult$ScheduledInstanceSet": "

Information about the Scheduled Instances.

" - } - }, - "RIProductDescription": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ProductDescription": "

The Reserved Instance product platform description. Instances that include (Amazon VPC) in the description are for use with Amazon VPC.

", - "ReservedInstances$ProductDescription": "

The Reserved Instance product platform description.

", - "ReservedInstancesOffering$ProductDescription": "

The Reserved Instance product platform description.

", - "SpotInstanceRequest$ProductDescription": "

The product description associated with the Spot instance.

", - "SpotPrice$ProductDescription": "

A general description of the AMI.

" - } - }, - "ReasonCodesList": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$ReasonCodes": "

One or more reason codes that describes the health state of your instance.

  • instance-stuck-in-state: My instance is stuck in a state.

  • unresponsive: My instance is unresponsive.

  • not-accepting-credentials: My instance is not accepting my credentials.

  • password-not-available: A password is not available for my instance.

  • performance-network: My instance is experiencing performance problems which I believe are network related.

  • performance-instance-store: My instance is experiencing performance problems which I believe are related to the instance stores.

  • performance-ebs-volume: My instance is experiencing performance problems which I believe are related to an EBS volume.

  • performance-other: My instance is experiencing performance problems.

  • other: [explain using the description parameter]

" - } - }, - "RebootInstancesRequest": { - "base": "

Contains the parameters for RebootInstances.

", - "refs": { - } - }, - "RecurringCharge": { - "base": "

Describes a recurring charge.

", - "refs": { - "RecurringChargesList$member": null - } - }, - "RecurringChargeFrequency": { - "base": null, - "refs": { - "RecurringCharge$Frequency": "

The frequency of the recurring charge.

" - } - }, - "RecurringChargesList": { - "base": null, - "refs": { - "ReservedInstances$RecurringCharges": "

The recurring charge tag assigned to the resource.

", - "ReservedInstancesOffering$RecurringCharges": "

The recurring charge tag assigned to the resource.

" - } - }, - "Region": { - "base": "

Describes a region.

", - "refs": { - "RegionList$member": null - } - }, - "RegionList": { - "base": null, - "refs": { - "DescribeRegionsResult$Regions": "

Information about one or more regions.

" - } - }, - "RegionNameStringList": { - "base": null, - "refs": { - "DescribeRegionsRequest$RegionNames": "

The names of one or more regions.

" - } - }, - "RegisterImageRequest": { - "base": "

Contains the parameters for RegisterImage.

", - "refs": { - } - }, - "RegisterImageResult": { - "base": "

Contains the output of RegisterImage.

", - "refs": { - } - }, - "RejectVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for RejectVpcPeeringConnection.

", - "refs": { - } - }, - "RejectVpcPeeringConnectionResult": { - "base": "

Contains the output of RejectVpcPeeringConnection.

", - "refs": { - } - }, - "ReleaseAddressRequest": { - "base": "

Contains the parameters for ReleaseAddress.

", - "refs": { - } - }, - "ReleaseHostsRequest": { - "base": "

Contains the parameters for ReleaseHosts.

", - "refs": { - } - }, - "ReleaseHostsResult": { - "base": "

Contains the output of ReleaseHosts.

", - "refs": { - } - }, - "ReplaceNetworkAclAssociationRequest": { - "base": "

Contains the parameters for ReplaceNetworkAclAssociation.

", - "refs": { - } - }, - "ReplaceNetworkAclAssociationResult": { - "base": "

Contains the output of ReplaceNetworkAclAssociation.

", - "refs": { - } - }, - "ReplaceNetworkAclEntryRequest": { - "base": "

Contains the parameters for ReplaceNetworkAclEntry.

", - "refs": { - } - }, - "ReplaceRouteRequest": { - "base": "

Contains the parameters for ReplaceRoute.

", - "refs": { - } - }, - "ReplaceRouteTableAssociationRequest": { - "base": "

Contains the parameters for ReplaceRouteTableAssociation.

", - "refs": { - } - }, - "ReplaceRouteTableAssociationResult": { - "base": "

Contains the output of ReplaceRouteTableAssociation.

", - "refs": { - } - }, - "ReportInstanceReasonCodes": { - "base": null, - "refs": { - "ReasonCodesList$member": null - } - }, - "ReportInstanceStatusRequest": { - "base": "

Contains the parameters for ReportInstanceStatus.

", - "refs": { - } - }, - "ReportStatusType": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$Status": "

The status of all instances listed.

" - } - }, - "RequestHostIdList": { - "base": null, - "refs": { - "DescribeHostsRequest$HostIds": "

The IDs of the Dedicated Hosts. The IDs are used for targeted instance launches.

", - "ModifyHostsRequest$HostIds": "

The host IDs of the Dedicated Hosts you want to modify.

", - "ReleaseHostsRequest$HostIds": "

The IDs of the Dedicated Hosts you want to release.

" - } - }, - "RequestHostIdSet": { - "base": null, - "refs": { - "GetHostReservationPurchasePreviewRequest$HostIdSet": "

The ID/s of the Dedicated Host/s that the reservation will be associated with.

", - "PurchaseHostReservationRequest$HostIdSet": "

The ID/s of the Dedicated Host/s that the reservation will be associated with.

" - } - }, - "RequestSpotFleetRequest": { - "base": "

Contains the parameters for RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotFleetResponse": { - "base": "

Contains the output of RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotInstancesRequest": { - "base": "

Contains the parameters for RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotInstancesResult": { - "base": "

Contains the output of RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotLaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "RequestSpotInstancesRequest$LaunchSpecification": null - } - }, - "Reservation": { - "base": "

Describes a reservation.

", - "refs": { - "ReservationList$member": null - } - }, - "ReservationList": { - "base": null, - "refs": { - "DescribeInstancesResult$Reservations": "

Zero or more reservations.

" - } - }, - "ReservationState": { - "base": null, - "refs": { - "HostReservation$State": "

The state of the reservation.

" - } - }, - "ReservedInstanceLimitPrice": { - "base": "

Describes the limit price of a Reserved Instance offering.

", - "refs": { - "PurchaseReservedInstancesOfferingRequest$LimitPrice": "

Specified for Reserved Instance Marketplace offerings to limit the total order and ensure that the Reserved Instances are not purchased at unexpected prices.

" - } - }, - "ReservedInstanceState": { - "base": null, - "refs": { - "ReservedInstances$State": "

The state of the Reserved Instance purchase.

" - } - }, - "ReservedInstances": { - "base": "

Describes a Reserved Instance.

", - "refs": { - "ReservedInstancesList$member": null - } - }, - "ReservedInstancesConfiguration": { - "base": "

Describes the configuration settings for the modified Reserved Instances.

", - "refs": { - "ReservedInstancesConfigurationList$member": null, - "ReservedInstancesModificationResult$TargetConfiguration": "

The target Reserved Instances configurations supplied as part of the modification request.

" - } - }, - "ReservedInstancesConfigurationList": { - "base": null, - "refs": { - "ModifyReservedInstancesRequest$TargetConfigurations": "

The configuration settings for the Reserved Instances to modify.

" - } - }, - "ReservedInstancesId": { - "base": "

Describes the ID of a Reserved Instance.

", - "refs": { - "ReservedIntancesIds$member": null - } - }, - "ReservedInstancesIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesRequest$ReservedInstancesIds": "

One or more Reserved Instance IDs.

Default: Describes all your Reserved Instances, or only those otherwise specified.

", - "ModifyReservedInstancesRequest$ReservedInstancesIds": "

The IDs of the Reserved Instances to modify.

" - } - }, - "ReservedInstancesList": { - "base": null, - "refs": { - "DescribeReservedInstancesResult$ReservedInstances": "

A list of Reserved Instances.

" - } - }, - "ReservedInstancesListing": { - "base": "

Describes a Reserved Instance listing.

", - "refs": { - "ReservedInstancesListingList$member": null - } - }, - "ReservedInstancesListingList": { - "base": null, - "refs": { - "CancelReservedInstancesListingResult$ReservedInstancesListings": "

The Reserved Instance listing.

", - "CreateReservedInstancesListingResult$ReservedInstancesListings": "

Information about the Reserved Instance listing.

", - "DescribeReservedInstancesListingsResult$ReservedInstancesListings": "

Information about the Reserved Instance listing.

" - } - }, - "ReservedInstancesModification": { - "base": "

Describes a Reserved Instance modification.

", - "refs": { - "ReservedInstancesModificationList$member": null - } - }, - "ReservedInstancesModificationIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsRequest$ReservedInstancesModificationIds": "

IDs for the submitted modification request.

" - } - }, - "ReservedInstancesModificationList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsResult$ReservedInstancesModifications": "

The Reserved Instance modification information.

" - } - }, - "ReservedInstancesModificationResult": { - "base": "

Describes the modification request/s.

", - "refs": { - "ReservedInstancesModificationResultList$member": null - } - }, - "ReservedInstancesModificationResultList": { - "base": null, - "refs": { - "ReservedInstancesModification$ModificationResults": "

Contains target configurations along with their corresponding new Reserved Instance IDs.

" - } - }, - "ReservedInstancesOffering": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "ReservedInstancesOfferingList$member": null - } - }, - "ReservedInstancesOfferingIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ReservedInstancesOfferingIds": "

One or more Reserved Instances offering IDs.

" - } - }, - "ReservedInstancesOfferingList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsResult$ReservedInstancesOfferings": "

A list of Reserved Instances offerings.

" - } - }, - "ReservedIntancesIds": { - "base": null, - "refs": { - "ReservedInstancesModification$ReservedInstancesIds": "

The IDs of one or more Reserved Instances.

" - } - }, - "ResetImageAttributeName": { - "base": null, - "refs": { - "ResetImageAttributeRequest$Attribute": "

The attribute to reset (currently you can only reset the launch permission attribute).

" - } - }, - "ResetImageAttributeRequest": { - "base": "

Contains the parameters for ResetImageAttribute.

", - "refs": { - } - }, - "ResetInstanceAttributeRequest": { - "base": "

Contains the parameters for ResetInstanceAttribute.

", - "refs": { - } - }, - "ResetNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for ResetNetworkInterfaceAttribute.

", - "refs": { - } - }, - "ResetSnapshotAttributeRequest": { - "base": "

Contains the parameters for ResetSnapshotAttribute.

", - "refs": { - } - }, - "ResourceIdList": { - "base": null, - "refs": { - "CreateTagsRequest$Resources": "

The IDs of one or more resources to tag. For example, ami-1a2b3c4d.

", - "DeleteTagsRequest$Resources": "

The ID of the resource. For example, ami-1a2b3c4d. You can specify more than one resource ID.

" - } - }, - "ResourceType": { - "base": null, - "refs": { - "TagDescription$ResourceType": "

The resource type.

" - } - }, - "ResponseHostIdList": { - "base": null, - "refs": { - "AllocateHostsResult$HostIds": "

The ID of the allocated Dedicated Host. This is used when you want to launch an instance onto a specific host.

", - "ModifyHostsResult$Successful": "

The IDs of the Dedicated Hosts that were successfully modified.

", - "ReleaseHostsResult$Successful": "

The IDs of the Dedicated Hosts that were successfully released.

" - } - }, - "ResponseHostIdSet": { - "base": null, - "refs": { - "HostReservation$HostIdSet": "

The IDs of the Dedicated Hosts associated with the reservation.

", - "Purchase$HostIdSet": "

The IDs of the Dedicated Hosts associated with the reservation.

" - } - }, - "RestorableByStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$RestorableByUserIds": "

One or more AWS accounts IDs that can create volumes from the snapshot.

" - } - }, - "RestoreAddressToClassicRequest": { - "base": "

Contains the parameters for RestoreAddressToClassic.

", - "refs": { - } - }, - "RestoreAddressToClassicResult": { - "base": "

Contains the output of RestoreAddressToClassic.

", - "refs": { - } - }, - "RevokeSecurityGroupEgressRequest": { - "base": "

Contains the parameters for RevokeSecurityGroupEgress.

", - "refs": { - } - }, - "RevokeSecurityGroupIngressRequest": { - "base": "

Contains the parameters for RevokeSecurityGroupIngress.

", - "refs": { - } - }, - "Route": { - "base": "

Describes a route in a route table.

", - "refs": { - "RouteList$member": null - } - }, - "RouteList": { - "base": null, - "refs": { - "RouteTable$Routes": "

The routes in the route table.

" - } - }, - "RouteOrigin": { - "base": null, - "refs": { - "Route$Origin": "

Describes how the route was created.

  • CreateRouteTable - The route was automatically created when the route table was created.

  • CreateRoute - The route was manually added to the route table.

  • EnableVgwRoutePropagation - The route was propagated by route propagation.

" - } - }, - "RouteState": { - "base": null, - "refs": { - "Route$State": "

The state of the route. The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, or the specified NAT instance has been terminated).

" - } - }, - "RouteTable": { - "base": "

Describes a route table.

", - "refs": { - "CreateRouteTableResult$RouteTable": "

Information about the route table.

", - "RouteTableList$member": null - } - }, - "RouteTableAssociation": { - "base": "

Describes an association between a route table and a subnet.

", - "refs": { - "RouteTableAssociationList$member": null - } - }, - "RouteTableAssociationList": { - "base": null, - "refs": { - "RouteTable$Associations": "

The associations between the route table and one or more subnets.

" - } - }, - "RouteTableList": { - "base": null, - "refs": { - "DescribeRouteTablesResult$RouteTables": "

Information about one or more route tables.

" - } - }, - "RuleAction": { - "base": null, - "refs": { - "CreateNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "NetworkAclEntry$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "ReplaceNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

" - } - }, - "RunInstancesMonitoringEnabled": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "LaunchSpecification$Monitoring": null, - "RequestSpotLaunchSpecification$Monitoring": null, - "RunInstancesRequest$Monitoring": "

The monitoring for the instance.

" - } - }, - "RunInstancesRequest": { - "base": "

Contains the parameters for RunInstances.

", - "refs": { - } - }, - "RunScheduledInstancesRequest": { - "base": "

Contains the parameters for RunScheduledInstances.

", - "refs": { - } - }, - "RunScheduledInstancesResult": { - "base": "

Contains the output of RunScheduledInstances.

", - "refs": { - } - }, - "S3Storage": { - "base": "

Describes the storage parameters for S3 and S3 buckets for an instance store-backed AMI.

", - "refs": { - "Storage$S3": "

An Amazon S3 storage location.

" - } - }, - "ScheduledInstance": { - "base": "

Describes a Scheduled Instance.

", - "refs": { - "PurchasedScheduledInstanceSet$member": null, - "ScheduledInstanceSet$member": null - } - }, - "ScheduledInstanceAvailability": { - "base": "

Describes a schedule that is available for your Scheduled Instances.

", - "refs": { - "ScheduledInstanceAvailabilitySet$member": null - } - }, - "ScheduledInstanceAvailabilitySet": { - "base": null, - "refs": { - "DescribeScheduledInstanceAvailabilityResult$ScheduledInstanceAvailabilitySet": "

Information about the available Scheduled Instances.

" - } - }, - "ScheduledInstanceIdRequestSet": { - "base": null, - "refs": { - "DescribeScheduledInstancesRequest$ScheduledInstanceIds": "

One or more Scheduled Instance IDs.

" - } - }, - "ScheduledInstanceRecurrence": { - "base": "

Describes the recurring schedule for a Scheduled Instance.

", - "refs": { - "ScheduledInstance$Recurrence": "

The schedule recurrence.

", - "ScheduledInstanceAvailability$Recurrence": "

The schedule recurrence.

" - } - }, - "ScheduledInstanceRecurrenceRequest": { - "base": "

Describes the recurring schedule for a Scheduled Instance.

", - "refs": { - "DescribeScheduledInstanceAvailabilityRequest$Recurrence": "

The schedule recurrence.

" - } - }, - "ScheduledInstanceSet": { - "base": null, - "refs": { - "DescribeScheduledInstancesResult$ScheduledInstanceSet": "

Information about the Scheduled Instances.

" - } - }, - "ScheduledInstancesBlockDeviceMapping": { - "base": "

Describes a block device mapping for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesBlockDeviceMappingSet$member": null - } - }, - "ScheduledInstancesBlockDeviceMappingSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "ScheduledInstancesEbs": { - "base": "

Describes an EBS volume for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesBlockDeviceMapping$Ebs": "

Parameters used to set up EBS volumes automatically when the instance is launched.

" - } - }, - "ScheduledInstancesIamInstanceProfile": { - "base": "

Describes an IAM instance profile for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "ScheduledInstancesLaunchSpecification": { - "base": "

Describes the launch specification for a Scheduled Instance.

If you are launching the Scheduled Instance in EC2-VPC, you must specify the ID of the subnet. You can specify the subnet using either SubnetId or NetworkInterface.

", - "refs": { - "RunScheduledInstancesRequest$LaunchSpecification": "

The launch specification. You must match the instance type, Availability Zone, network, and platform of the schedule that you purchased.

" - } - }, - "ScheduledInstancesMonitoring": { - "base": "

Describes whether monitoring is enabled for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "ScheduledInstancesNetworkInterface": { - "base": "

Describes a network interface for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesNetworkInterfaceSet$member": null - } - }, - "ScheduledInstancesNetworkInterfaceSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "ScheduledInstancesPlacement": { - "base": "

Describes the placement for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$Placement": "

The placement information.

" - } - }, - "ScheduledInstancesPrivateIpAddressConfig": { - "base": "

Describes a private IP address for a Scheduled Instance.

", - "refs": { - "PrivateIpAddressConfigSet$member": null - } - }, - "ScheduledInstancesSecurityGroupIdSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$SecurityGroupIds": "

The IDs of one or more security groups.

", - "ScheduledInstancesNetworkInterface$Groups": "

The IDs of one or more security groups.

" - } - }, - "SecurityGroup": { - "base": "

Describes a security group

", - "refs": { - "SecurityGroupList$member": null - } - }, - "SecurityGroupIdStringList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$Groups": "

The IDs of one or more security groups.

", - "ImportInstanceLaunchSpecification$GroupIds": "

One or more security group IDs.

", - "InstanceNetworkInterfaceSpecification$Groups": "

The IDs of the security groups for the network interface. Applies only if creating a network interface when launching an instance.

", - "ModifyNetworkInterfaceAttributeRequest$Groups": "

Changes the security groups for the network interface. The new set of groups you specify replaces the current set. You must specify at least one group, even if it's just the default security group in the VPC. You must specify the ID of the security group, not the name.

", - "RunInstancesRequest$SecurityGroupIds": "

One or more security group IDs. You can create a security group using CreateSecurityGroup.

Default: Amazon EC2 uses the default security group.

" - } - }, - "SecurityGroupList": { - "base": null, - "refs": { - "DescribeSecurityGroupsResult$SecurityGroups": "

Information about one or more security groups.

" - } - }, - "SecurityGroupReference": { - "base": "

Describes a VPC with a security group that references your security group.

", - "refs": { - "SecurityGroupReferences$member": null - } - }, - "SecurityGroupReferences": { - "base": null, - "refs": { - "DescribeSecurityGroupReferencesResult$SecurityGroupReferenceSet": "

Information about the VPCs with the referencing security groups.

" - } - }, - "SecurityGroupStringList": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$GroupNames": "

One or more security group names.

", - "RunInstancesRequest$SecurityGroups": "

[EC2-Classic, default VPC] One or more security group names. For a nondefault VPC, you must use security group IDs instead.

Default: Amazon EC2 uses the default security group.

" - } - }, - "ShutdownBehavior": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "RunInstancesRequest$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

Default: stop

" - } - }, - "SlotDateTimeRangeRequest": { - "base": "

Describes the time period for a Scheduled Instance to start its first schedule. The time period must span less than one day.

", - "refs": { - "DescribeScheduledInstanceAvailabilityRequest$FirstSlotStartTimeRange": "

The time period for the first schedule to start.

" - } - }, - "SlotStartTimeRangeRequest": { - "base": "

Describes the time period for a Scheduled Instance to start its first schedule.

", - "refs": { - "DescribeScheduledInstancesRequest$SlotStartTimeRange": "

The time period for the first schedule to start.

" - } - }, - "Snapshot": { - "base": "

Describes a snapshot.

", - "refs": { - "SnapshotList$member": null - } - }, - "SnapshotAttributeName": { - "base": null, - "refs": { - "DescribeSnapshotAttributeRequest$Attribute": "

The snapshot attribute you would like to view.

", - "ModifySnapshotAttributeRequest$Attribute": "

The snapshot attribute to modify.

Only volume creation permissions may be modified at the customer level.

", - "ResetSnapshotAttributeRequest$Attribute": "

The attribute to reset. Currently, only the attribute for permission to create volumes can be reset.

" - } - }, - "SnapshotDetail": { - "base": "

Describes the snapshot created from the imported disk.

", - "refs": { - "SnapshotDetailList$member": null - } - }, - "SnapshotDetailList": { - "base": null, - "refs": { - "ImportImageResult$SnapshotDetails": "

Information about the snapshots.

", - "ImportImageTask$SnapshotDetails": "

Information about the snapshots.

" - } - }, - "SnapshotDiskContainer": { - "base": "

The disk container object for the import snapshot request.

", - "refs": { - "ImportSnapshotRequest$DiskContainer": "

Information about the disk container.

" - } - }, - "SnapshotIdStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$SnapshotIds": "

One or more snapshot IDs.

Default: Describes snapshots for which you have launch permissions.

" - } - }, - "SnapshotList": { - "base": null, - "refs": { - "DescribeSnapshotsResult$Snapshots": "

Information about the snapshots.

" - } - }, - "SnapshotState": { - "base": null, - "refs": { - "Snapshot$State": "

The snapshot state.

" - } - }, - "SnapshotTaskDetail": { - "base": "

Details about the import snapshot task.

", - "refs": { - "ImportSnapshotResult$SnapshotTaskDetail": "

Information about the import snapshot task.

", - "ImportSnapshotTask$SnapshotTaskDetail": "

Describes an import snapshot task.

" - } - }, - "SpotDatafeedSubscription": { - "base": "

Describes the data feed for a Spot instance.

", - "refs": { - "CreateSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

", - "DescribeSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

" - } - }, - "SpotFleetLaunchSpecification": { - "base": "

Describes the launch specification for one or more Spot instances.

", - "refs": { - "LaunchSpecsList$member": null - } - }, - "SpotFleetMonitoring": { - "base": "

Describes whether monitoring is enabled.

", - "refs": { - "SpotFleetLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "SpotFleetRequestConfig": { - "base": "

Describes a Spot fleet request.

", - "refs": { - "SpotFleetRequestConfigSet$member": null - } - }, - "SpotFleetRequestConfigData": { - "base": "

Describes the configuration of a Spot fleet request.

", - "refs": { - "RequestSpotFleetRequest$SpotFleetRequestConfig": "

The configuration for the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestConfig": "

Information about the configuration of the Spot fleet request.

" - } - }, - "SpotFleetRequestConfigSet": { - "base": null, - "refs": { - "DescribeSpotFleetRequestsResponse$SpotFleetRequestConfigs": "

Information about the configuration of your Spot fleet.

" - } - }, - "SpotInstanceRequest": { - "base": "

Describes a Spot instance request.

", - "refs": { - "SpotInstanceRequestList$member": null - } - }, - "SpotInstanceRequestIdList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

", - "DescribeSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

" - } - }, - "SpotInstanceRequestList": { - "base": null, - "refs": { - "DescribeSpotInstanceRequestsResult$SpotInstanceRequests": "

One or more Spot instance requests.

", - "RequestSpotInstancesResult$SpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "SpotInstanceState": { - "base": null, - "refs": { - "SpotInstanceRequest$State": "

The state of the Spot instance request. Spot bid status information can help you track your Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "SpotInstanceStateFault": { - "base": "

Describes a Spot instance state change.

", - "refs": { - "SpotDatafeedSubscription$Fault": "

The fault codes for the Spot instance request, if any.

", - "SpotInstanceRequest$Fault": "

The fault codes for the Spot instance request, if any.

" - } - }, - "SpotInstanceStatus": { - "base": "

Describes the status of a Spot instance request.

", - "refs": { - "SpotInstanceRequest$Status": "

The status code and status message describing the Spot instance request.

" - } - }, - "SpotInstanceType": { - "base": null, - "refs": { - "RequestSpotInstancesRequest$Type": "

The Spot instance request type.

Default: one-time

", - "SpotInstanceRequest$Type": "

The Spot instance request type.

" - } - }, - "SpotPlacement": { - "base": "

Describes Spot instance placement.

", - "refs": { - "LaunchSpecification$Placement": "

The placement information for the instance.

", - "RequestSpotLaunchSpecification$Placement": "

The placement information for the instance.

", - "SpotFleetLaunchSpecification$Placement": "

The placement information.

" - } - }, - "SpotPrice": { - "base": "

Describes the maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "refs": { - "SpotPriceHistoryList$member": null - } - }, - "SpotPriceHistoryList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryResult$SpotPriceHistory": "

The historical Spot prices.

" - } - }, - "StaleIpPermission": { - "base": "

Describes a stale rule in a security group.

", - "refs": { - "StaleIpPermissionSet$member": null - } - }, - "StaleIpPermissionSet": { - "base": null, - "refs": { - "StaleSecurityGroup$StaleIpPermissions": "

Information about the stale inbound rules in the security group.

", - "StaleSecurityGroup$StaleIpPermissionsEgress": "

Information about the stale outbound rules in the security group.

" - } - }, - "StaleSecurityGroup": { - "base": "

Describes a stale security group (a security group that contains stale rules).

", - "refs": { - "StaleSecurityGroupSet$member": null - } - }, - "StaleSecurityGroupSet": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsResult$StaleSecurityGroupSet": "

Information about the stale security groups.

" - } - }, - "StartInstancesRequest": { - "base": "

Contains the parameters for StartInstances.

", - "refs": { - } - }, - "StartInstancesResult": { - "base": "

Contains the output of StartInstances.

", - "refs": { - } - }, - "State": { - "base": null, - "refs": { - "VpcEndpoint$State": "

The state of the VPC endpoint.

" - } - }, - "StateReason": { - "base": "

Describes a state change.

", - "refs": { - "Image$StateReason": "

The reason for the state change.

", - "Instance$StateReason": "

The reason for the most recent state transition.

" - } - }, - "Status": { - "base": null, - "refs": { - "MoveAddressToVpcResult$Status": "

The status of the move of the IP address.

", - "RestoreAddressToClassicResult$Status": "

The move status for the IP address.

" - } - }, - "StatusName": { - "base": null, - "refs": { - "InstanceStatusDetails$Name": "

The type of instance status.

" - } - }, - "StatusType": { - "base": null, - "refs": { - "InstanceStatusDetails$Status": "

The status.

" - } - }, - "StopInstancesRequest": { - "base": "

Contains the parameters for StopInstances.

", - "refs": { - } - }, - "StopInstancesResult": { - "base": "

Contains the output of StopInstances.

", - "refs": { - } - }, - "Storage": { - "base": "

Describes the storage location for an instance store-backed AMI.

", - "refs": { - "BundleInstanceRequest$Storage": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "BundleTask$Storage": "

The Amazon S3 storage locations.

" - } - }, - "String": { - "base": null, - "refs": { - "AcceptVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "AccountAttribute$AttributeName": "

The name of the account attribute.

", - "AccountAttributeValue$AttributeValue": "

The value of the attribute.

", - "ActiveInstance$InstanceType": "

The instance type.

", - "ActiveInstance$InstanceId": "

The ID of the instance.

", - "ActiveInstance$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "Address$InstanceId": "

The ID of the instance that the address is associated with (if any).

", - "Address$PublicIp": "

The Elastic IP address.

", - "Address$AllocationId": "

The ID representing the allocation of the address for use with EC2-VPC.

", - "Address$AssociationId": "

The ID representing the association of the address with an instance in a VPC.

", - "Address$NetworkInterfaceId": "

The ID of the network interface.

", - "Address$NetworkInterfaceOwnerId": "

The ID of the AWS account that owns the network interface.

", - "Address$PrivateIpAddress": "

The private IP address associated with the Elastic IP address.

", - "AllocateAddressResult$PublicIp": "

The Elastic IP address.

", - "AllocateAddressResult$AllocationId": "

[EC2-VPC] The ID that AWS assigns to represent the allocation of the Elastic IP address for use with instances in a VPC.

", - "AllocateHostsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "AllocateHostsRequest$InstanceType": "

Specify the instance type that you want your Dedicated Hosts to be configured for. When you specify the instance type, that is the only instance type that you can launch onto that host.

", - "AllocateHostsRequest$AvailabilityZone": "

The Availability Zone for the Dedicated Hosts.

", - "AllocationIdList$member": null, - "AssignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AssociateAddressRequest$InstanceId": "

The ID of the instance. This is required for EC2-Classic. For EC2-VPC, you can specify either the instance ID or the network interface ID, but not both. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "AssociateAddressRequest$PublicIp": "

The Elastic IP address. This is required for EC2-Classic.

", - "AssociateAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. This is required for EC2-VPC.

", - "AssociateAddressRequest$NetworkInterfaceId": "

[EC2-VPC] The ID of the network interface. If the instance has more than one network interface, you must specify a network interface ID.

", - "AssociateAddressRequest$PrivateIpAddress": "

[EC2-VPC] The primary or secondary private IP address to associate with the Elastic IP address. If no private IP address is specified, the Elastic IP address is associated with the primary private IP address.

", - "AssociateAddressResult$AssociationId": "

[EC2-VPC] The ID that represents the association of the Elastic IP address with an instance.

", - "AssociateDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set, or default to associate no DHCP options with the VPC.

", - "AssociateDhcpOptionsRequest$VpcId": "

The ID of the VPC.

", - "AssociateRouteTableRequest$SubnetId": "

The ID of the subnet.

", - "AssociateRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "AssociateRouteTableResult$AssociationId": "

The route table association ID (needed to disassociate the route table).

", - "AttachClassicLinkVpcRequest$InstanceId": "

The ID of an EC2-Classic instance to link to the ClassicLink-enabled VPC.

", - "AttachClassicLinkVpcRequest$VpcId": "

The ID of a ClassicLink-enabled VPC.

", - "AttachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "AttachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttachNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AttachNetworkInterfaceRequest$InstanceId": "

The ID of the instance.

", - "AttachNetworkInterfaceResult$AttachmentId": "

The ID of the network interface attachment.

", - "AttachVolumeRequest$VolumeId": "

The ID of the EBS volume. The volume and instance must be within the same Availability Zone.

", - "AttachVolumeRequest$InstanceId": "

The ID of the instance.

", - "AttachVolumeRequest$Device": "

The device name to expose to the instance (for example, /dev/sdh or xvdh).

", - "AttachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "AttachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttributeValue$Value": "

The attribute value. Note that the value is case-sensitive.

", - "AuthorizeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name or number. We recommend that you specify the protocol in a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. We recommend that you specify the CIDR range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "AuthorizeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead. For EC2-VPC, the source security group must be in the same VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic] The AWS account number for the source security group, if the source security group is in a different account. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). (VPC only) Use -1 to specify all traffic. If you specify -1, traffic on all ports is allowed, regardless of any ports you specify.

", - "AuthorizeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "AvailabilityZone$ZoneName": "

The name of the Availability Zone.

", - "AvailabilityZone$RegionName": "

The name of the region.

", - "AvailabilityZoneMessage$Message": "

The message about the Availability Zone.

", - "BlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with 2 available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "BlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "BlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "BundleIdStringList$member": null, - "BundleInstanceRequest$InstanceId": "

The ID of the instance to bundle.

Type: String

Default: None

Required: Yes

", - "BundleTask$InstanceId": "

The ID of the instance associated with this bundle task.

", - "BundleTask$BundleId": "

The ID of the bundle task.

", - "BundleTask$Progress": "

The level of task completion, as a percent (for example, 20%).

", - "BundleTaskError$Code": "

The error code.

", - "BundleTaskError$Message": "

The error message.

", - "CancelBundleTaskRequest$BundleId": "

The ID of the bundle task.

", - "CancelConversionRequest$ConversionTaskId": "

The ID of the conversion task.

", - "CancelConversionRequest$ReasonMessage": "

The reason for canceling the conversion task.

", - "CancelExportTaskRequest$ExportTaskId": "

The ID of the export task. This is the ID returned by CreateInstanceExportTask.

", - "CancelImportTaskRequest$ImportTaskId": "

The ID of the import image or import snapshot task to be canceled.

", - "CancelImportTaskRequest$CancelReason": "

The reason for canceling the task.

", - "CancelImportTaskResult$ImportTaskId": "

The ID of the task being canceled.

", - "CancelImportTaskResult$State": "

The current state of the task being canceled.

", - "CancelImportTaskResult$PreviousState": "

The current state of the task being canceled.

", - "CancelReservedInstancesListingRequest$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "CancelSpotFleetRequestsError$Message": "

The description for the error code.

", - "CancelSpotFleetRequestsErrorItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelledSpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "ClassicLinkDnsSupport$VpcId": "

The ID of the VPC.

", - "ClassicLinkInstance$InstanceId": "

The ID of the instance.

", - "ClassicLinkInstance$VpcId": "

The ID of the VPC.

", - "ClientData$Comment": "

A user-defined comment about the disk upload.

", - "ConfirmProductInstanceRequest$ProductCode": "

The product code. This must be a product code that you own.

", - "ConfirmProductInstanceRequest$InstanceId": "

The ID of the instance.

", - "ConfirmProductInstanceResult$OwnerId": "

The AWS account ID of the instance owner. This is only present if the product code is attached to the instance.

", - "ConversionIdStringList$member": null, - "ConversionTask$ConversionTaskId": "

The ID of the conversion task.

", - "ConversionTask$ExpirationTime": "

The time when the task expires. If the upload isn't complete before the expiration time, we automatically cancel the task.

", - "ConversionTask$StatusMessage": "

The status message related to the conversion task.

", - "CopyImageRequest$SourceRegion": "

The name of the region that contains the AMI to copy.

", - "CopyImageRequest$SourceImageId": "

The ID of the AMI to copy.

", - "CopyImageRequest$Name": "

The name of the new AMI in the destination region.

", - "CopyImageRequest$Description": "

A description for the new AMI in the destination region.

", - "CopyImageRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "CopyImageRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when encrypting the snapshots of an image during a copy operation. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopyImageResult$ImageId": "

The ID of the new AMI.

", - "CopySnapshotRequest$SourceRegion": "

The ID of the region that contains the snapshot to be copied.

", - "CopySnapshotRequest$SourceSnapshotId": "

The ID of the EBS snapshot to copy.

", - "CopySnapshotRequest$Description": "

A description for the EBS snapshot.

", - "CopySnapshotRequest$DestinationRegion": "

The destination region to use in the PresignedUrl parameter of a snapshot copy operation. This parameter is only valid for specifying the destination region in a PresignedUrl parameter, where it is required.

CopySnapshot sends the snapshot copy to the regional endpoint that you send the HTTP request to, such as ec2.us-east-1.amazonaws.com (in the AWS CLI, this is specified with the --region parameter or the default region in your AWS configuration file).

", - "CopySnapshotRequest$PresignedUrl": "

The pre-signed URL that facilitates copying an encrypted snapshot. This parameter is only required when copying an encrypted snapshot with the Amazon EC2 Query API; it is available as an optional parameter in all other cases. The PresignedUrl should use the snapshot source endpoint, the CopySnapshot action, and include the SourceRegion, SourceSnapshotId, and DestinationRegion parameters. The PresignedUrl must be signed using AWS Signature Version 4. Because EBS snapshots are stored in Amazon S3, the signing algorithm for this parameter uses the same logic that is described in Authenticating Requests by Using Query Parameters (AWS Signature Version 4) in the Amazon Simple Storage Service API Reference. An invalid or improperly signed PresignedUrl will cause the copy operation to fail asynchronously, and the snapshot will move to an error state.

", - "CopySnapshotRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when creating the snapshot copy. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopySnapshotResult$SnapshotId": "

The ID of the new snapshot.

", - "CreateCustomerGatewayRequest$PublicIp": "

The Internet-routable IP address for the customer gateway's outside interface. The address must be static.

", - "CreateFlowLogsRequest$LogGroupName": "

The name of the CloudWatch log group.

", - "CreateFlowLogsRequest$DeliverLogsPermissionArn": "

The ARN for the IAM role that's used to post flow logs to a CloudWatch Logs log group.

", - "CreateFlowLogsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateFlowLogsResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateImageRequest$InstanceId": "

The ID of the instance.

", - "CreateImageRequest$Name": "

A name for the new image.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "CreateImageRequest$Description": "

A description for the new image.

", - "CreateImageResult$ImageId": "

The ID of the new AMI.

", - "CreateInstanceExportTaskRequest$Description": "

A description for the conversion task or the resource being exported. The maximum length is 255 bytes.

", - "CreateInstanceExportTaskRequest$InstanceId": "

The ID of the instance.

", - "CreateKeyPairRequest$KeyName": "

A unique name for the key pair.

Constraints: Up to 255 ASCII characters

", - "CreateNatGatewayRequest$SubnetId": "

The subnet in which to create the NAT gateway.

", - "CreateNatGatewayRequest$AllocationId": "

The allocation ID of an Elastic IP address to associate with the NAT gateway. If the Elastic IP address is associated with another resource, you must first disassociate it.

", - "CreateNatGatewayRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

Constraint: Maximum 64 ASCII characters.

", - "CreateNatGatewayResult$ClientToken": "

Unique, case-sensitive identifier to ensure the idempotency of the request. Only returned if a client token was provided in the request.

", - "CreateNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "CreateNetworkAclEntryRequest$Protocol": "

The protocol. A value of -1 means all protocols.

", - "CreateNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation (for example 172.16.0.0/24).

", - "CreateNetworkAclRequest$VpcId": "

The ID of the VPC.

", - "CreateNetworkInterfaceRequest$SubnetId": "

The ID of the subnet to associate with the network interface.

", - "CreateNetworkInterfaceRequest$Description": "

A description for the network interface.

", - "CreateNetworkInterfaceRequest$PrivateIpAddress": "

The primary private IP address of the network interface. If you don't specify an IP address, Amazon EC2 selects one for you from the subnet range. If you specify an IP address, you cannot indicate any IP addresses specified in privateIpAddresses as primary (only one IP address can be designated as primary).

", - "CreatePlacementGroupRequest$GroupName": "

A name for the placement group.

Constraints: Up to 255 ASCII characters

", - "CreateReservedInstancesListingRequest$ReservedInstancesId": "

The ID of the active Reserved Instance.

", - "CreateReservedInstancesListingRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "CreateRouteRequest$RouteTableId": "

The ID of the route table for the route.

", - "CreateRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. Routing decisions are based on the most specific match.

", - "CreateRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway attached to your VPC.

", - "CreateRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "CreateRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "CreateRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "CreateRouteRequest$NatGatewayId": "

The ID of a NAT gateway.

", - "CreateRouteTableRequest$VpcId": "

The ID of the VPC.

", - "CreateSecurityGroupRequest$GroupName": "

The name of the security group.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=&;{}!$*

", - "CreateSecurityGroupRequest$Description": "

A description for the security group. This is informational only.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=&;{}!$*

", - "CreateSecurityGroupRequest$VpcId": "

[EC2-VPC] The ID of the VPC. Required for EC2-VPC.

", - "CreateSecurityGroupResult$GroupId": "

The ID of the security group.

", - "CreateSnapshotRequest$VolumeId": "

The ID of the EBS volume.

", - "CreateSnapshotRequest$Description": "

A description for the snapshot.

", - "CreateSpotDatafeedSubscriptionRequest$Bucket": "

The Amazon S3 bucket in which to store the Spot instance data feed.

", - "CreateSpotDatafeedSubscriptionRequest$Prefix": "

A prefix for the data feed file names.

", - "CreateSubnetRequest$VpcId": "

The ID of the VPC.

", - "CreateSubnetRequest$CidrBlock": "

The network range for the subnet, in CIDR notation. For example, 10.0.0.0/24.

", - "CreateSubnetRequest$AvailabilityZone": "

The Availability Zone for the subnet.

Default: AWS selects one for you. If you create more than one subnet in your VPC, we may not necessarily select a different zone for each subnet.

", - "CreateVolumePermission$UserId": "

The specific AWS account ID that is to be added or removed from a volume's list of create volume permissions.

", - "CreateVolumeRequest$SnapshotId": "

The snapshot from which to create the volume.

", - "CreateVolumeRequest$AvailabilityZone": "

The Availability Zone in which to create the volume. Use DescribeAvailabilityZones to list the Availability Zones that are currently available to you.

", - "CreateVolumeRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) to use when creating the encrypted volume. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CreateVpcEndpointRequest$VpcId": "

The ID of the VPC in which the endpoint will be used.

", - "CreateVpcEndpointRequest$ServiceName": "

The AWS service name, in the form com.amazonaws.region.service . To get a list of available services, use the DescribeVpcEndpointServices request.

", - "CreateVpcEndpointRequest$PolicyDocument": "

A policy to attach to the endpoint that controls access to the service. The policy must be in valid JSON format. If this parameter is not specified, we attach a default policy that allows full access to the service.

", - "CreateVpcEndpointRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateVpcEndpointResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateVpcPeeringConnectionRequest$VpcId": "

The ID of the requester VPC.

", - "CreateVpcPeeringConnectionRequest$PeerVpcId": "

The ID of the VPC with which you are creating the VPC peering connection.

", - "CreateVpcPeeringConnectionRequest$PeerOwnerId": "

The AWS account ID of the owner of the peer VPC.

Default: Your AWS account ID

", - "CreateVpcRequest$CidrBlock": "

The network range for the VPC, in CIDR notation. For example, 10.0.0.0/16.

", - "CreateVpnConnectionRequest$Type": "

The type of VPN connection (ipsec.1).

", - "CreateVpnConnectionRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "CreateVpnConnectionRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "CreateVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "CreateVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "CreateVpnGatewayRequest$AvailabilityZone": "

The Availability Zone for the virtual private gateway.

", - "CustomerGateway$CustomerGatewayId": "

The ID of the customer gateway.

", - "CustomerGateway$State": "

The current state of the customer gateway (pending | available | deleting | deleted).

", - "CustomerGateway$Type": "

The type of VPN connection the customer gateway supports (ipsec.1).

", - "CustomerGateway$IpAddress": "

The Internet-routable IP address of the customer gateway's outside interface.

", - "CustomerGateway$BgpAsn": "

The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

", - "CustomerGatewayIdStringList$member": null, - "DeleteCustomerGatewayRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "DeleteDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set.

", - "DeleteInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DeleteKeyPairRequest$KeyName": "

The name of the key pair.

", - "DeleteNatGatewayRequest$NatGatewayId": "

The ID of the NAT gateway.

", - "DeleteNatGatewayResult$NatGatewayId": "

The ID of the NAT gateway.

", - "DeleteNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkAclRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DeletePlacementGroupRequest$GroupName": "

The name of the placement group.

", - "DeleteRouteRequest$RouteTableId": "

The ID of the route table.

", - "DeleteRouteRequest$DestinationCidrBlock": "

The CIDR range for the route. The value you specify must match the CIDR for the route exactly.

", - "DeleteRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "DeleteSecurityGroupRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group. You can specify either the security group name or the security group ID.

", - "DeleteSecurityGroupRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "DeleteSnapshotRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DeleteSubnetRequest$SubnetId": "

The ID of the subnet.

", - "DeleteVolumeRequest$VolumeId": "

The ID of the volume.

", - "DeleteVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "DeleteVpcRequest$VpcId": "

The ID of the VPC.

", - "DeleteVpnConnectionRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "DeleteVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DeregisterImageRequest$ImageId": "

The ID of the AMI.

", - "DescribeClassicLinkInstancesRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeClassicLinkInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeFlowLogsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeFlowLogsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostReservationOfferingsRequest$OfferingId": "

The ID of the reservation offering.

", - "DescribeHostReservationOfferingsRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeHostReservationOfferingsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostReservationsRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeHostReservationsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeHostsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "DescribeIdentityIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "DescribeIdentityIdFormatRequest$PrincipalArn": "

The ARN of the principal, which can be an IAM role, IAM user, or the root user.

", - "DescribeImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "DescribeImportImageTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportImageTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeImportSnapshotTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportSnapshotTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "DescribeInstanceStatusRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeInstanceStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeInstancesRequest$NextToken": "

The token to request the next page of results.

", - "DescribeInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeMovingAddressesRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeMovingAddressesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNatGatewaysRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeNatGatewaysResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribeNetworkInterfaceAttributeResult$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribePrefixListsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribePrefixListsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesId": "

One or more Reserved Instance IDs.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesListingId": "

One or more Reserved Instance listing IDs.

", - "DescribeReservedInstancesModificationsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesModificationsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeReservedInstancesOfferingsRequest$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "DescribeReservedInstancesOfferingsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesOfferingsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeScheduledInstanceAvailabilityRequest$NextToken": "

The token for the next set of results.

", - "DescribeScheduledInstanceAvailabilityResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeScheduledInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeScheduledInstancesResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSnapshotAttributeRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotAttributeResult$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotsRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeSnapshots request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeSnapshotsResult$NextToken": "

The NextToken value to include in a future DescribeSnapshots request. When the results of a DescribeSnapshots request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetInstancesRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetInstancesResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestHistoryRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestHistoryResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestsRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestsResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotPriceHistoryRequest$AvailabilityZone": "

Filters the results by the specified Availability Zone.

", - "DescribeSpotPriceHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotPriceHistoryResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeStaleSecurityGroupsRequest$VpcId": "

The ID of the VPC.

", - "DescribeStaleSecurityGroupsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeTagsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeTagsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return..

", - "DescribeVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "DescribeVolumeAttributeResult$VolumeId": "

The ID of the volume.

", - "DescribeVolumeStatusRequest$NextToken": "

The NextToken value to include in a future DescribeVolumeStatus request. When the results of the request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumeStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumesRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeVolumes request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeVolumesResult$NextToken": "

The NextToken value to include in a future DescribeVolumes request. When the results of a DescribeVolumes request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "DescribeVpcAttributeResult$VpcId": "

The ID of the VPC.

", - "DescribeVpcEndpointServicesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointServicesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeVpcEndpointsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DetachClassicLinkVpcRequest$InstanceId": "

The ID of the instance to unlink from the VPC.

", - "DetachClassicLinkVpcRequest$VpcId": "

The ID of the VPC to which the instance is linked.

", - "DetachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DetachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "DetachNetworkInterfaceRequest$AttachmentId": "

The ID of the attachment.

", - "DetachVolumeRequest$VolumeId": "

The ID of the volume.

", - "DetachVolumeRequest$InstanceId": "

The ID of the instance.

", - "DetachVolumeRequest$Device": "

The device name.

", - "DetachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DetachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "DhcpConfiguration$Key": "

The name of a DHCP option.

", - "DhcpOptions$DhcpOptionsId": "

The ID of the set of DHCP options.

", - "DhcpOptionsIdStringList$member": null, - "DisableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "DisableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "DisableVpcClassicLinkDnsSupportRequest$VpcId": "

The ID of the VPC.

", - "DisableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "DisassociateAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "DisassociateAddressRequest$AssociationId": "

[EC2-VPC] The association ID. Required for EC2-VPC.

", - "DisassociateRouteTableRequest$AssociationId": "

The association ID representing the current association between the route table and subnet.

", - "DiskImage$Description": "

A description of the disk image.

", - "DiskImageDescription$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DiskImageDescription$Checksum": "

The checksum computed for the disk image.

", - "DiskImageDetail$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3 and presented here as an Amazon S3 presigned URL. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DiskImageVolumeDescription$Id": "

The volume identifier.

", - "EbsBlockDevice$SnapshotId": "

The ID of the snapshot.

", - "EbsInstanceBlockDevice$VolumeId": "

The ID of the EBS volume.

", - "EbsInstanceBlockDeviceSpecification$VolumeId": "

The ID of the EBS volume.

", - "EnableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "EnableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "EnableVolumeIORequest$VolumeId": "

The ID of the volume.

", - "EnableVpcClassicLinkDnsSupportRequest$VpcId": "

The ID of the VPC.

", - "EnableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "EventInformation$InstanceId": "

The ID of the instance. This information is available only for instanceChange events.

", - "EventInformation$EventSubType": "

The event.

The following are the error events.

  • iamFleetRoleInvalid - The Spot fleet did not have the required permissions either to launch or terminate an instance.

  • launchSpecTemporarilyBlacklisted - The configuration is not valid and several attempts to launch instances have failed. For more information, see the description of the event.

  • spotFleetRequestConfigurationInvalid - The configuration is not valid. For more information, see the description of the event.

  • spotInstanceCountLimitExceeded - You've reached the limit on the number of Spot instances that you can launch.

The following are the fleetRequestChange events.

  • active - The Spot fleet has been validated and Amazon EC2 is attempting to maintain the target number of running Spot instances.

  • cancelled - The Spot fleet is canceled and has no running Spot instances. The Spot fleet will be deleted two days after its instances were terminated.

  • cancelled_running - The Spot fleet is canceled and will not launch additional Spot instances, but its existing Spot instances continue to run until they are interrupted or terminated.

  • cancelled_terminating - The Spot fleet is canceled and its Spot instances are terminating.

  • expired - The Spot fleet request has expired. A subsequent event indicates that the instances were terminated, if the request was created with TerminateInstancesWithExpiration set.

  • modify_in_progress - A request to modify the Spot fleet request was accepted and is in progress.

  • modify_successful - The Spot fleet request was modified.

  • price_update - The bid price for a launch configuration was adjusted because it was too high. This change is permanent.

  • submitted - The Spot fleet request is being evaluated and Amazon EC2 is preparing to launch the target number of Spot instances.

The following are the instanceChange events.

  • launched - A bid was fulfilled and a new instance was launched.

  • terminated - An instance was terminated by the user.

", - "EventInformation$EventDescription": "

The description of the event.

", - "ExecutableByStringList$member": null, - "ExportTask$ExportTaskId": "

The ID of the export task.

", - "ExportTask$Description": "

A description of the resource being exported.

", - "ExportTask$StatusMessage": "

The status message related to the export task.

", - "ExportTaskIdStringList$member": null, - "ExportToS3Task$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3Task$S3Key": "

The encryption key for your S3 bucket.

", - "ExportToS3TaskSpecification$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3TaskSpecification$S3Prefix": "

The image is written to a single object in the S3 bucket at the S3 key s3prefix + exportTaskId + '.' + diskImageFormat.

", - "Filter$Name": "

The name of the filter. Filter names are case-sensitive.

", - "FlowLog$FlowLogId": "

The flow log ID.

", - "FlowLog$FlowLogStatus": "

The status of the flow log (ACTIVE).

", - "FlowLog$ResourceId": "

The ID of the resource on which the flow log was created.

", - "FlowLog$LogGroupName": "

The name of the flow log group.

", - "FlowLog$DeliverLogsStatus": "

The status of the logs delivery (SUCCESS | FAILED).

", - "FlowLog$DeliverLogsErrorMessage": "

Information about the error that occurred. Rate limited indicates that CloudWatch logs throttling has been applied for one or more network interfaces, or that you've reached the limit on the number of CloudWatch Logs log groups that you can create. Access error indicates that the IAM role associated with the flow log does not have sufficient permissions to publish to CloudWatch Logs. Unknown error indicates an internal error.

", - "FlowLog$DeliverLogsPermissionArn": "

The ARN of the IAM role that posts logs to CloudWatch Logs.

", - "GetConsoleOutputRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$Output": "

The console output, Base64-encoded. If using a command line tool, the tool decodes the output for you.

", - "GetConsoleScreenshotRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleScreenshotResult$InstanceId": "

The ID of the instance.

", - "GetConsoleScreenshotResult$ImageData": "

The data that comprises the image.

", - "GetHostReservationPurchasePreviewRequest$OfferingId": "

The offering ID of the reservation.

", - "GetHostReservationPurchasePreviewResult$TotalUpfrontPrice": "

The potential total upfront price. This is billed immediately.

", - "GetHostReservationPurchasePreviewResult$TotalHourlyPrice": "

The potential total hourly price of the reservation per hour.

", - "GetPasswordDataRequest$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$PasswordData": "

The password of the instance.

", - "GroupIdStringList$member": null, - "GroupIdentifier$GroupName": "

The name of the security group.

", - "GroupIdentifier$GroupId": "

The ID of the security group.

", - "GroupIds$member": null, - "GroupNameStringList$member": null, - "Host$HostId": "

The ID of the Dedicated Host.

", - "Host$HostReservationId": "

The reservation ID of the Dedicated Host. This returns a null response if the Dedicated Host doesn't have an associated reservation.

", - "Host$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "Host$AvailabilityZone": "

The Availability Zone of the Dedicated Host.

", - "HostInstance$InstanceId": "

the IDs of instances that are running on the Dedicated Host.

", - "HostInstance$InstanceType": "

The instance type size (for example, m3.medium) of the running instance.

", - "HostOffering$OfferingId": "

The ID of the offering.

", - "HostOffering$InstanceFamily": "

The instance family of the offering.

", - "HostOffering$UpfrontPrice": "

The upfront price of the offering. Does not apply to No Upfront offerings.

", - "HostOffering$HourlyPrice": "

The hourly price of the offering.

", - "HostProperties$InstanceType": "

The instance type size that the Dedicated Host supports (for example, m3.medium).

", - "HostReservation$HostReservationId": "

The ID of the reservation that specifies the associated Dedicated Hosts.

", - "HostReservation$OfferingId": "

The ID of the reservation. This remains the same regardless of which Dedicated Hosts are associated with it.

", - "HostReservation$InstanceFamily": "

The instance family of the Dedicated Host Reservation. The instance family on the Dedicated Host must be the same in order for it to benefit from the reservation.

", - "HostReservation$HourlyPrice": "

The hourly price of the reservation.

", - "HostReservation$UpfrontPrice": "

The upfront price of the reservation.

", - "HostReservationIdSet$member": null, - "IamInstanceProfile$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfile$Id": "

The ID of the instance profile.

", - "IamInstanceProfileSpecification$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfileSpecification$Name": "

The name of the instance profile.

", - "IdFormat$Resource": "

The type of resource.

", - "Image$ImageId": "

The ID of the AMI.

", - "Image$ImageLocation": "

The location of the AMI.

", - "Image$OwnerId": "

The AWS account ID of the image owner.

", - "Image$CreationDate": "

The date and time the image was created.

", - "Image$KernelId": "

The kernel associated with the image, if any. Only applicable for machine images.

", - "Image$RamdiskId": "

The RAM disk associated with the image, if any. Only applicable for machine images.

", - "Image$SriovNetSupport": "

Specifies whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "Image$ImageOwnerAlias": "

The AWS account alias (for example, amazon, self) or the AWS account ID of the AMI owner.

", - "Image$Name": "

The name of the AMI that was provided during image creation.

", - "Image$Description": "

The description of the AMI that was provided during image creation.

", - "Image$RootDeviceName": "

The device name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "ImageAttribute$ImageId": "

The ID of the AMI.

", - "ImageDiskContainer$Description": "

The description of the disk image.

", - "ImageDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "ImageDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. The URL can either be a https URL (https://..) or an Amazon S3 URL (s3://..)

", - "ImageDiskContainer$DeviceName": "

The block device mapping for the disk.

", - "ImageDiskContainer$SnapshotId": "

The ID of the EBS snapshot to be used for importing the snapshot.

", - "ImageIdStringList$member": null, - "ImportImageRequest$Description": "

A description string for the import image task.

", - "ImportImageRequest$LicenseType": "

The license type to be used for the Amazon Machine Image (AMI) after importing.

Note: You may only use BYOL if you have existing licenses with rights to use these licenses in a third party cloud like AWS. For more information, see Prerequisites in the VM Import/Export User Guide.

Valid values: AWS | BYOL

", - "ImportImageRequest$Hypervisor": "

The target hypervisor platform.

Valid values: xen

", - "ImportImageRequest$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageRequest$Platform": "

The operating system of the virtual machine.

Valid values: Windows | Linux

", - "ImportImageRequest$ClientToken": "

The token to enable idempotency for VM import requests.

", - "ImportImageRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportImageResult$ImportTaskId": "

The task ID of the import image task.

", - "ImportImageResult$Architecture": "

The architecture of the virtual machine.

", - "ImportImageResult$LicenseType": "

The license type of the virtual machine.

", - "ImportImageResult$Platform": "

The operating system of the virtual machine.

", - "ImportImageResult$Hypervisor": "

The target hypervisor of the import task.

", - "ImportImageResult$Description": "

A description of the import task.

", - "ImportImageResult$ImageId": "

The ID of the Amazon Machine Image (AMI) created by the import task.

", - "ImportImageResult$Progress": "

The progress of the task.

", - "ImportImageResult$StatusMessage": "

A detailed status message of the import task.

", - "ImportImageResult$Status": "

A brief status of the task.

", - "ImportImageTask$ImportTaskId": "

The ID of the import image task.

", - "ImportImageTask$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageTask$LicenseType": "

The license type of the virtual machine.

", - "ImportImageTask$Platform": "

The description string for the import image task.

", - "ImportImageTask$Hypervisor": "

The target hypervisor for the import task.

Valid values: xen

", - "ImportImageTask$Description": "

A description of the import task.

", - "ImportImageTask$ImageId": "

The ID of the Amazon Machine Image (AMI) of the imported virtual machine.

", - "ImportImageTask$Progress": "

The percentage of progress of the import image task.

", - "ImportImageTask$StatusMessage": "

A descriptive status message for the import image task.

", - "ImportImageTask$Status": "

A brief status for the import image task.

", - "ImportInstanceLaunchSpecification$AdditionalInfo": "

Reserved.

", - "ImportInstanceLaunchSpecification$SubnetId": "

[EC2-VPC] The ID of the subnet in which to launch the instance.

", - "ImportInstanceLaunchSpecification$PrivateIpAddress": "

[EC2-VPC] An available IP address from the IP address range of the subnet.

", - "ImportInstanceRequest$Description": "

A description for the instance being imported.

", - "ImportInstanceTaskDetails$InstanceId": "

The ID of the instance.

", - "ImportInstanceTaskDetails$Description": "

A description of the task.

", - "ImportInstanceVolumeDetailItem$AvailabilityZone": "

The Availability Zone where the resulting instance will reside.

", - "ImportInstanceVolumeDetailItem$Status": "

The status of the import of this particular disk image.

", - "ImportInstanceVolumeDetailItem$StatusMessage": "

The status information or errors related to the disk image.

", - "ImportInstanceVolumeDetailItem$Description": "

A description of the task.

", - "ImportKeyPairRequest$KeyName": "

A unique name for the key pair.

", - "ImportKeyPairResult$KeyName": "

The key pair name you provided.

", - "ImportKeyPairResult$KeyFingerprint": "

The MD5 public key fingerprint as specified in section 4 of RFC 4716.

", - "ImportSnapshotRequest$Description": "

The description string for the import snapshot task.

", - "ImportSnapshotRequest$ClientToken": "

Token to enable idempotency for VM import requests.

", - "ImportSnapshotRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportSnapshotResult$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotResult$Description": "

A description of the import snapshot task.

", - "ImportSnapshotTask$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotTask$Description": "

A description of the import snapshot task.

", - "ImportTaskIdList$member": null, - "ImportVolumeRequest$AvailabilityZone": "

The Availability Zone for the resulting EBS volume.

", - "ImportVolumeRequest$Description": "

A description of the volume.

", - "ImportVolumeTaskDetails$AvailabilityZone": "

The Availability Zone where the resulting volume will reside.

", - "ImportVolumeTaskDetails$Description": "

The description you provided when starting the import volume task.

", - "Instance$InstanceId": "

The ID of the instance.

", - "Instance$ImageId": "

The ID of the AMI used to launch the instance.

", - "Instance$PrivateDnsName": "

The private DNS name assigned to the instance. This DNS name can only be used inside the Amazon EC2 network. This name is not available until the instance enters the running state. For EC2-VPC, this name is only available if you've enabled DNS hostnames for your VPC.

", - "Instance$PublicDnsName": "

The public DNS name assigned to the instance. This name is not available until the instance enters the running state. For EC2-VPC, this name is only available if you've enabled DNS hostnames for your VPC.

", - "Instance$StateTransitionReason": "

The reason for the most recent state transition. This might be an empty string.

", - "Instance$KeyName": "

The name of the key pair, if this instance was launched with an associated key pair.

", - "Instance$KernelId": "

The kernel associated with this instance, if applicable.

", - "Instance$RamdiskId": "

The RAM disk associated with this instance, if applicable.

", - "Instance$SubnetId": "

[EC2-VPC] The ID of the subnet in which the instance is running.

", - "Instance$VpcId": "

[EC2-VPC] The ID of the VPC in which the instance is running.

", - "Instance$PrivateIpAddress": "

The private IP address assigned to the instance.

", - "Instance$PublicIpAddress": "

The public IP address assigned to the instance, if applicable.

", - "Instance$RootDeviceName": "

The root device name (for example, /dev/sda1 or /dev/xvda).

", - "Instance$SpotInstanceRequestId": "

If the request is a Spot instance request, the ID of the request.

", - "Instance$ClientToken": "

The idempotency token you provided when you launched the instance, if applicable.

", - "Instance$SriovNetSupport": "

Specifies whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "InstanceAttribute$InstanceId": "

The ID of the instance.

", - "InstanceBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$VirtualName": "

The virtual device name.

", - "InstanceBlockDeviceMappingSpecification$NoDevice": "

suppress the specified device included in the block device mapping.

", - "InstanceCapacity$InstanceType": "

The instance type size supported by the Dedicated Host.

", - "InstanceExportDetails$InstanceId": "

The ID of the resource being exported.

", - "InstanceIdSet$member": null, - "InstanceIdStringList$member": null, - "InstanceMonitoring$InstanceId": "

The ID of the instance.

", - "InstanceNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterface$SubnetId": "

The ID of the subnet.

", - "InstanceNetworkInterface$VpcId": "

The ID of the VPC.

", - "InstanceNetworkInterface$Description": "

The description.

", - "InstanceNetworkInterface$OwnerId": "

The ID of the AWS account that created the network interface.

", - "InstanceNetworkInterface$MacAddress": "

The MAC address.

", - "InstanceNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "InstanceNetworkInterface$PrivateDnsName": "

The private DNS name.

", - "InstanceNetworkInterfaceAssociation$PublicIp": "

The public IP address or Elastic IP address bound to the network interface.

", - "InstanceNetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "InstanceNetworkInterfaceAssociation$IpOwnerId": "

The ID of the owner of the Elastic IP address.

", - "InstanceNetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterfaceSpecification$SubnetId": "

The ID of the subnet associated with the network string. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$Description": "

The description of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddress": "

The private IP address of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstancePrivateIpAddress$PrivateIpAddress": "

The private IP address of the network interface.

", - "InstancePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "InstanceStateChange$InstanceId": "

The ID of the instance.

", - "InstanceStatus$InstanceId": "

The ID of the instance.

", - "InstanceStatus$AvailabilityZone": "

The Availability Zone of the instance.

", - "InstanceStatusEvent$Description": "

A description of the event.

After a scheduled event is completed, it can still be described for up to a week. If the event has been completed, this description starts with the following text: [Completed].

", - "InternetGateway$InternetGatewayId": "

The ID of the Internet gateway.

", - "InternetGatewayAttachment$VpcId": "

The ID of the VPC.

", - "IpPermission$IpProtocol": "

The IP protocol name (for tcp, udp, and icmp) or number (see Protocol Numbers).

[EC2-VPC only] When you authorize or revoke security group rules, you can use -1 to specify all.

", - "IpRange$CidrIp": "

The CIDR range. You can either specify a CIDR range or a source security group, not both.

", - "IpRanges$member": null, - "KeyNameStringList$member": null, - "KeyPair$KeyName": "

The name of the key pair.

", - "KeyPair$KeyFingerprint": "

The SHA-1 digest of the DER encoded private key.

", - "KeyPair$KeyMaterial": "

An unencrypted PEM encoded RSA private key.

", - "KeyPairInfo$KeyName": "

The name of the key pair.

", - "KeyPairInfo$KeyFingerprint": "

If you used CreateKeyPair to create the key pair, this is the SHA-1 digest of the DER encoded private key. If you used ImportKeyPair to provide AWS the public key, this is the MD5 public key fingerprint as specified in section 4 of RFC4716.

", - "LaunchPermission$UserId": "

The AWS account ID.

", - "LaunchSpecification$ImageId": "

The ID of the AMI.

", - "LaunchSpecification$KeyName": "

The name of the key pair.

", - "LaunchSpecification$UserData": "

The user data to make available to the instances. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "LaunchSpecification$AddressingType": "

Deprecated.

", - "LaunchSpecification$KernelId": "

The ID of the kernel.

", - "LaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "LaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "ModifyIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "ModifyIdentityIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "ModifyIdentityIdFormatRequest$PrincipalArn": "

The ARN of the principal, which can be an IAM user, IAM role, or the root user. Specify all to modify the ID format for all IAM users, IAM roles, and the root user of the account.

", - "ModifyImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ModifyImageAttributeRequest$Attribute": "

The name of the attribute to modify.

", - "ModifyImageAttributeRequest$Value": "

The value of the attribute being modified. This is only valid when modifying the description attribute.

", - "ModifyInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ModifyInstanceAttributeRequest$Value": "

A new value for the attribute. Use only with the kernel, ramdisk, userData, disableApiTermination, or instanceInitiatedShutdownBehavior attribute.

", - "ModifyInstancePlacementRequest$InstanceId": "

The ID of the instance that you are modifying.

", - "ModifyInstancePlacementRequest$HostId": "

The ID of the Dedicated Host that the instance will have affinity with.

", - "ModifyNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ModifyReservedInstancesRequest$ClientToken": "

A unique, case-sensitive token you provide to ensure idempotency of your modification request. For more information, see Ensuring Idempotency.

", - "ModifyReservedInstancesResult$ReservedInstancesModificationId": "

The ID for the modification.

", - "ModifySnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ModifySpotFleetRequestRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "ModifySubnetAttributeRequest$SubnetId": "

The ID of the subnet.

", - "ModifyVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "ModifyVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "ModifyVpcEndpointRequest$VpcEndpointId": "

The ID of the endpoint.

", - "ModifyVpcEndpointRequest$PolicyDocument": "

A policy document to attach to the endpoint. The policy must be in valid JSON format.

", - "ModifyVpcPeeringConnectionOptionsRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "MoveAddressToVpcRequest$PublicIp": "

The Elastic IP address.

", - "MoveAddressToVpcResult$AllocationId": "

The allocation ID for the Elastic IP address.

", - "MovingAddressStatus$PublicIp": "

The Elastic IP address.

", - "NatGateway$VpcId": "

The ID of the VPC in which the NAT gateway is located.

", - "NatGateway$SubnetId": "

The ID of the subnet in which the NAT gateway is located.

", - "NatGateway$NatGatewayId": "

The ID of the NAT gateway.

", - "NatGateway$FailureCode": "

If the NAT gateway could not be created, specifies the error code for the failure. (InsufficientFreeAddressesInSubnet | Gateway.NotAttached | InvalidAllocationID.NotFound | Resource.AlreadyAssociated | InternalError | InvalidSubnetID.NotFound)

", - "NatGateway$FailureMessage": "

If the NAT gateway could not be created, specifies the error message for the failure, that corresponds to the error code.

  • For InsufficientFreeAddressesInSubnet: \"Subnet has insufficient free addresses to create this NAT gateway\"

  • For Gateway.NotAttached: \"Network vpc-xxxxxxxx has no Internet gateway attached\"

  • For InvalidAllocationID.NotFound: \"Elastic IP address eipalloc-xxxxxxxx could not be associated with this NAT gateway\"

  • For Resource.AlreadyAssociated: \"Elastic IP address eipalloc-xxxxxxxx is already associated\"

  • For InternalError: \"Network interface eni-xxxxxxxx, created and used internally by this NAT gateway is in an invalid state. Please try again.\"

  • For InvalidSubnetID.NotFound: \"The specified subnet subnet-xxxxxxxx does not exist or could not be found.\"

", - "NatGatewayAddress$PublicIp": "

The Elastic IP address associated with the NAT gateway.

", - "NatGatewayAddress$AllocationId": "

The allocation ID of the Elastic IP address that's associated with the NAT gateway.

", - "NatGatewayAddress$PrivateIp": "

The private IP address associated with the Elastic IP address.

", - "NatGatewayAddress$NetworkInterfaceId": "

The ID of the network interface associated with the NAT gateway.

", - "NetworkAcl$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAcl$VpcId": "

The ID of the VPC for the network ACL.

", - "NetworkAclAssociation$NetworkAclAssociationId": "

The ID of the association between a network ACL and a subnet.

", - "NetworkAclAssociation$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAclAssociation$SubnetId": "

The ID of the subnet.

", - "NetworkAclEntry$Protocol": "

The protocol. A value of -1 means all protocols.

", - "NetworkAclEntry$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "NetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "NetworkInterface$SubnetId": "

The ID of the subnet.

", - "NetworkInterface$VpcId": "

The ID of the VPC.

", - "NetworkInterface$AvailabilityZone": "

The Availability Zone.

", - "NetworkInterface$Description": "

A description.

", - "NetworkInterface$OwnerId": "

The AWS account ID of the owner of the network interface.

", - "NetworkInterface$RequesterId": "

The ID of the entity that launched the instance on your behalf (for example, AWS Management Console or Auto Scaling).

", - "NetworkInterface$MacAddress": "

The MAC address.

", - "NetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "NetworkInterface$PrivateDnsName": "

The private DNS name.

", - "NetworkInterfaceAssociation$PublicIp": "

The address of the Elastic IP address bound to the network interface.

", - "NetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "NetworkInterfaceAssociation$IpOwnerId": "

The ID of the Elastic IP address owner.

", - "NetworkInterfaceAssociation$AllocationId": "

The allocation ID.

", - "NetworkInterfaceAssociation$AssociationId": "

The association ID.

", - "NetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceAttachment$InstanceId": "

The ID of the instance.

", - "NetworkInterfaceAttachment$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "NetworkInterfaceAttachmentChanges$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceIdList$member": null, - "NetworkInterfacePrivateIpAddress$PrivateIpAddress": "

The private IP address.

", - "NetworkInterfacePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "NewDhcpConfiguration$Key": null, - "OwnerStringList$member": null, - "Placement$AvailabilityZone": "

The Availability Zone of the instance.

", - "Placement$GroupName": "

The name of the placement group the instance is in (for cluster compute instances).

", - "Placement$HostId": "

The ID of the Dedicted host on which the instance resides. This parameter is not support for the ImportInstance command.

", - "Placement$Affinity": "

The affinity setting for the instance on the Dedicated Host. This parameter is not supported for the ImportInstance command.

", - "PlacementGroup$GroupName": "

The name of the placement group.

", - "PlacementGroupStringList$member": null, - "PrefixList$PrefixListId": "

The ID of the prefix.

", - "PrefixList$PrefixListName": "

The name of the prefix.

", - "PrefixListId$PrefixListId": "

The ID of the prefix.

", - "PrefixListIdSet$member": null, - "PrivateIpAddressSpecification$PrivateIpAddress": "

The private IP addresses.

", - "PrivateIpAddressStringList$member": null, - "ProductCode$ProductCodeId": "

The product code.

", - "ProductCodeStringList$member": null, - "ProductDescriptionList$member": null, - "PropagatingVgw$GatewayId": "

The ID of the virtual private gateway (VGW).

", - "ProvisionedBandwidth$Provisioned": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$Requested": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$Status": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "PublicIpStringList$member": null, - "Purchase$HostReservationId": "

The ID of the reservation.

", - "Purchase$InstanceFamily": "

The instance family on the Dedicated Host that the reservation can be associated with.

", - "Purchase$UpfrontPrice": "

The upfront price of the reservation.

", - "Purchase$HourlyPrice": "

The hourly price of the reservation per hour.

", - "PurchaseHostReservationRequest$OfferingId": "

The ID of the offering.

", - "PurchaseHostReservationRequest$LimitPrice": "

The specified limit is checked against the total upfront cost of the reservation (calculated as the offering's upfront cost multiplied by the host count). If the total upfront cost is greater than the specified price limit, the request will fail. This is used to ensure that the purchase does not exceed the expected upfront cost of the purchase. At this time, the only supported currency is USD. For example, to indicate a limit price of USD 100, specify 100.00.

", - "PurchaseHostReservationRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "PurchaseHostReservationResult$TotalUpfrontPrice": "

The total amount that will be charged to your account when you purchase the reservation.

", - "PurchaseHostReservationResult$TotalHourlyPrice": "

The total hourly price of the reservation calculated per hour.

", - "PurchaseHostReservationResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide

", - "PurchaseRequest$PurchaseToken": "

The purchase token.

", - "PurchaseReservedInstancesOfferingRequest$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering to purchase.

", - "PurchaseReservedInstancesOfferingResult$ReservedInstancesId": "

The IDs of the purchased Reserved Instances.

", - "PurchaseScheduledInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that ensures the idempotency of the request. For more information, see Ensuring Idempotency.

", - "Region$RegionName": "

The name of the region.

", - "Region$Endpoint": "

The region service endpoint.

", - "RegionNameStringList$member": null, - "RegisterImageRequest$ImageLocation": "

The full path to your AMI manifest in Amazon S3 storage.

", - "RegisterImageRequest$Name": "

A name for your AMI.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "RegisterImageRequest$Description": "

A description for your AMI.

", - "RegisterImageRequest$KernelId": "

The ID of the kernel.

", - "RegisterImageRequest$RamdiskId": "

The ID of the RAM disk.

", - "RegisterImageRequest$RootDeviceName": "

The name of the root device (for example, /dev/sda1, or /dev/xvda).

", - "RegisterImageRequest$VirtualizationType": "

The type of virtualization.

Default: paravirtual

", - "RegisterImageRequest$SriovNetSupport": "

Set to simple to enable enhanced networking with the Intel 82599 Virtual Function interface for the AMI and any instances that you launch from the AMI.

There is no way to disable sriovNetSupport at this time.

This option is supported only for HVM AMIs. Specifying this option with a PV AMI can make instances launched from the AMI unreachable.

", - "RegisterImageResult$ImageId": "

The ID of the newly registered AMI.

", - "RejectVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "ReleaseAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "ReleaseAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. Required for EC2-VPC.

", - "ReplaceNetworkAclAssociationRequest$AssociationId": "

The ID of the current association between the original network ACL and the subnet.

", - "ReplaceNetworkAclAssociationRequest$NetworkAclId": "

The ID of the new network ACL to associate with the subnet.

", - "ReplaceNetworkAclAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReplaceNetworkAclEntryRequest$NetworkAclId": "

The ID of the ACL.

", - "ReplaceNetworkAclEntryRequest$Protocol": "

The IP protocol. You can specify all or -1 to mean all protocols.

", - "ReplaceNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "ReplaceRouteRequest$RouteTableId": "

The ID of the route table.

", - "ReplaceRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. The value you provide must match the CIDR of an existing route in the table.

", - "ReplaceRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway.

", - "ReplaceRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC.

", - "ReplaceRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "ReplaceRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "ReplaceRouteRequest$NatGatewayId": "

The ID of a NAT gateway.

", - "ReplaceRouteTableAssociationRequest$AssociationId": "

The association ID.

", - "ReplaceRouteTableAssociationRequest$RouteTableId": "

The ID of the new route table to associate with the subnet.

", - "ReplaceRouteTableAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReportInstanceStatusRequest$Description": "

Descriptive text about the health state of your instance.

", - "RequestHostIdList$member": null, - "RequestHostIdSet$member": null, - "RequestSpotFleetResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "RequestSpotInstancesRequest$SpotPrice": "

The maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "RequestSpotInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstancesRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

Default: Instances are launched and terminated individually

", - "RequestSpotInstancesRequest$AvailabilityZoneGroup": "

The user-specified name for a logical grouping of bids.

When you specify an Availability Zone group in a Spot Instance request, all Spot instances in the request are launched in the same Availability Zone. Instance proximity is maintained with this parameter, but the choice of Availability Zone is not. The group applies only to bids for Spot Instances of the same instance type. Any additional Spot instance requests that are specified with the same Availability Zone group name are launched in that same Availability Zone, as long as at least one instance from the group is still active.

If there is no active instance running in the Availability Zone group that you specify for a new Spot instance request (all instances are terminated, the bid is expired, or the bid falls below current market), then Amazon EC2 launches the instance in any Availability Zone where the constraint can be met. Consequently, the subsequent set of Spot instances could be placed in a different zone from the original request, even if you specified the same Availability Zone group.

Default: Instances are launched in any available Availability Zone.

", - "RequestSpotLaunchSpecification$ImageId": "

The ID of the AMI.

", - "RequestSpotLaunchSpecification$KeyName": "

The name of the key pair.

", - "RequestSpotLaunchSpecification$UserData": "

The user data to make available to the instances. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "RequestSpotLaunchSpecification$AddressingType": "

Deprecated.

", - "RequestSpotLaunchSpecification$KernelId": "

The ID of the kernel.

", - "RequestSpotLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "RequestSpotLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "Reservation$ReservationId": "

The ID of the reservation.

", - "Reservation$OwnerId": "

The ID of the AWS account that owns the reservation.

", - "Reservation$RequesterId": "

The ID of the requester that launched the instances on your behalf (for example, AWS Management Console or Auto Scaling).

", - "ReservedInstances$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstances$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$AvailabilityZone": "

The Availability Zone for the modified Reserved Instances.

", - "ReservedInstancesConfiguration$Platform": "

The network platform of the modified Reserved Instances, which is either EC2-Classic or EC2-VPC.

", - "ReservedInstancesId$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesIdStringList$member": null, - "ReservedInstancesListing$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "ReservedInstancesListing$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesListing$StatusMessage": "

The reason for the current status of the Reserved Instance listing. The response can be blank.

", - "ReservedInstancesListing$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModification$ReservedInstancesModificationId": "

A unique ID for the Reserved Instance modification.

", - "ReservedInstancesModification$Status": "

The status of the Reserved Instances modification request.

", - "ReservedInstancesModification$StatusMessage": "

The reason for the status.

", - "ReservedInstancesModification$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModificationIdStringList$member": null, - "ReservedInstancesModificationResult$ReservedInstancesId": "

The ID for the Reserved Instances that were created as part of the modification request. This field is only available when the modification is fulfilled.

", - "ReservedInstancesOffering$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering.

", - "ReservedInstancesOffering$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesOfferingIdStringList$member": null, - "ResetImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ResetInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ResetNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ResetNetworkInterfaceAttributeRequest$SourceDestCheck": "

The source/destination checking attribute. Resets the value to true.

", - "ResetSnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ResourceIdList$member": null, - "ResponseHostIdList$member": null, - "ResponseHostIdSet$member": null, - "RestorableByStringList$member": null, - "RestoreAddressToClassicRequest$PublicIp": "

The Elastic IP address.

", - "RestoreAddressToClassicResult$PublicIp": "

The Elastic IP address.

", - "RevokeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name or number. We recommend that you specify the protocol in a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. We recommend that you specify the CIDR range in a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "RevokeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a security group in a nondefault VPC.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. For EC2-VPC, the source security group must be in the same VPC. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic] The AWS account ID of the source security group, if the source security group is in a different account. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). Use -1 to specify all.

", - "RevokeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "Route$DestinationCidrBlock": "

The CIDR block used for the destination match.

", - "Route$DestinationPrefixListId": "

The prefix of the AWS service.

", - "Route$GatewayId": "

The ID of a gateway attached to your VPC.

", - "Route$InstanceId": "

The ID of a NAT instance in your VPC.

", - "Route$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "Route$NetworkInterfaceId": "

The ID of the network interface.

", - "Route$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "Route$NatGatewayId": "

The ID of a NAT gateway.

", - "RouteTable$RouteTableId": "

The ID of the route table.

", - "RouteTable$VpcId": "

The ID of the VPC.

", - "RouteTableAssociation$RouteTableAssociationId": "

The ID of the association between a route table and a subnet.

", - "RouteTableAssociation$RouteTableId": "

The ID of the route table.

", - "RouteTableAssociation$SubnetId": "

The ID of the subnet. A subnet ID is not returned for an implicit association.

", - "RunInstancesRequest$ImageId": "

The ID of the AMI, which you can get by calling DescribeImages.

", - "RunInstancesRequest$KeyName": "

The name of the key pair. You can create a key pair using CreateKeyPair or ImportKeyPair.

If you do not specify a key pair, you can't connect to the instance unless you choose an AMI that is configured to allow users another way to log in.

", - "RunInstancesRequest$UserData": "

The user data to make available to the instance. For more information, see Running Commands on Your Linux Instance at Launch (Linux) and Adding User Data (Windows). If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "RunInstancesRequest$KernelId": "

The ID of the kernel.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$RamdiskId": "

The ID of the RAM disk.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$SubnetId": "

[EC2-VPC] The ID of the subnet to launch the instance into.

", - "RunInstancesRequest$PrivateIpAddress": "

[EC2-VPC] The primary IP address. You must specify a value from the IP address range of the subnet.

Only one private IP address can be designated as primary. Therefore, you can't specify this parameter if PrivateIpAddresses.n.Primary is set to true and PrivateIpAddresses.n.PrivateIpAddress is set to an IP address.

Default: We select an IP address from the IP address range of the subnet.

", - "RunInstancesRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see Ensuring Idempotency.

Constraints: Maximum 64 ASCII characters

", - "RunInstancesRequest$AdditionalInfo": "

Reserved.

", - "RunScheduledInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that ensures the idempotency of the request. For more information, see Ensuring Idempotency.

", - "RunScheduledInstancesRequest$ScheduledInstanceId": "

The Scheduled Instance ID.

", - "S3Storage$Bucket": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "S3Storage$Prefix": "

The beginning of the file name of the AMI.

", - "S3Storage$AWSAccessKeyId": "

The access key ID of the owner of the bucket. Before you specify a value for your access key ID, review and follow the guidance in Best Practices for Managing AWS Access Keys.

", - "S3Storage$UploadPolicySignature": "

The signature of the JSON document.

", - "ScheduledInstance$ScheduledInstanceId": "

The Scheduled Instance ID.

", - "ScheduledInstance$InstanceType": "

The instance type.

", - "ScheduledInstance$Platform": "

The platform (Linux/UNIX or Windows).

", - "ScheduledInstance$NetworkPlatform": "

The network platform (EC2-Classic or EC2-VPC).

", - "ScheduledInstance$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstance$HourlyPrice": "

The hourly price for a single instance.

", - "ScheduledInstanceAvailability$InstanceType": "

The instance type. You can specify one of the C3, C4, M4, or R3 instance types.

", - "ScheduledInstanceAvailability$Platform": "

The platform (Linux/UNIX or Windows).

", - "ScheduledInstanceAvailability$NetworkPlatform": "

The network platform (EC2-Classic or EC2-VPC).

", - "ScheduledInstanceAvailability$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstanceAvailability$PurchaseToken": "

The purchase token. This token expires in two hours.

", - "ScheduledInstanceAvailability$HourlyPrice": "

The hourly price for a single instance.

", - "ScheduledInstanceIdRequestSet$member": null, - "ScheduledInstanceRecurrence$Frequency": "

The frequency (Daily, Weekly, or Monthly).

", - "ScheduledInstanceRecurrence$OccurrenceUnit": "

The unit for occurrenceDaySet (DayOfWeek or DayOfMonth).

", - "ScheduledInstanceRecurrenceRequest$Frequency": "

The frequency (Daily, Weekly, or Monthly).

", - "ScheduledInstanceRecurrenceRequest$OccurrenceUnit": "

The unit for OccurrenceDays (DayOfWeek or DayOfMonth). This value is required for a monthly schedule. You can't specify DayOfWeek with a weekly schedule. You can't specify this value with a daily schedule.

", - "ScheduledInstancesBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "ScheduledInstancesBlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "ScheduledInstancesBlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with two available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "ScheduledInstancesEbs$SnapshotId": "

The ID of the snapshot.

", - "ScheduledInstancesEbs$VolumeType": "

The volume type. gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, Throughput Optimized HDD for st1, Cold HDD for sc1, or standard for Magnetic.

Default: standard

", - "ScheduledInstancesIamInstanceProfile$Arn": "

The Amazon Resource Name (ARN).

", - "ScheduledInstancesIamInstanceProfile$Name": "

The name.

", - "ScheduledInstancesLaunchSpecification$ImageId": "

The ID of the Amazon Machine Image (AMI).

", - "ScheduledInstancesLaunchSpecification$KeyName": "

The name of the key pair.

", - "ScheduledInstancesLaunchSpecification$UserData": "

The base64-encoded MIME user data.

", - "ScheduledInstancesLaunchSpecification$KernelId": "

The ID of the kernel.

", - "ScheduledInstancesLaunchSpecification$InstanceType": "

The instance type.

", - "ScheduledInstancesLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "ScheduledInstancesLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances.

", - "ScheduledInstancesNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "ScheduledInstancesNetworkInterface$SubnetId": "

The ID of the subnet.

", - "ScheduledInstancesNetworkInterface$Description": "

The description.

", - "ScheduledInstancesNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "ScheduledInstancesPlacement$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstancesPlacement$GroupName": "

The name of the placement group.

", - "ScheduledInstancesPrivateIpAddressConfig$PrivateIpAddress": "

The IP address.

", - "ScheduledInstancesSecurityGroupIdSet$member": null, - "SecurityGroup$OwnerId": "

The AWS account ID of the owner of the security group.

", - "SecurityGroup$GroupName": "

The name of the security group.

", - "SecurityGroup$GroupId": "

The ID of the security group.

", - "SecurityGroup$Description": "

A description of the security group.

", - "SecurityGroup$VpcId": "

[EC2-VPC] The ID of the VPC for the security group.

", - "SecurityGroupIdStringList$member": null, - "SecurityGroupReference$GroupId": "

The ID of your security group.

", - "SecurityGroupReference$ReferencingVpcId": "

The ID of the VPC with the referencing security group.

", - "SecurityGroupReference$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "SecurityGroupStringList$member": null, - "Snapshot$SnapshotId": "

The ID of the snapshot. Each snapshot receives a unique identifier when it is created.

", - "Snapshot$VolumeId": "

The ID of the volume that was used to create the snapshot. Snapshots created by the CopySnapshot action have an arbitrary volume ID that should not be used for any purpose.

", - "Snapshot$StateMessage": "

Encrypted Amazon EBS snapshots are copied asynchronously. If a snapshot copy operation fails (for example, if the proper AWS Key Management Service (AWS KMS) permissions are not obtained) this field displays error state details to help you diagnose why the error occurred. This parameter is only returned by the DescribeSnapshots API operation.

", - "Snapshot$Progress": "

The progress of the snapshot, as a percentage.

", - "Snapshot$OwnerId": "

The AWS account ID of the EBS snapshot owner.

", - "Snapshot$Description": "

The description for the snapshot.

", - "Snapshot$OwnerAlias": "

Value from an Amazon-maintained list (amazon | aws-marketplace | microsoft) of snapshot owners. Not to be confused with the user-configured AWS account alias, which is set from the IAM console.

", - "Snapshot$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the parent volume.

", - "Snapshot$DataEncryptionKeyId": "

The data encryption key identifier for the snapshot. This value is a unique identifier that corresponds to the data encryption key that was used to encrypt the original volume or snapshot copy. Because data encryption keys are inherited by volumes created from snapshots, and vice versa, if snapshots share the same data encryption key identifier, then they belong to the same volume/snapshot lineage. This parameter is only returned by the DescribeSnapshots API operation.

", - "SnapshotDetail$Description": "

A description for the snapshot.

", - "SnapshotDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotDetail$Url": "

The URL used to access the disk image.

", - "SnapshotDetail$DeviceName": "

The block device mapping for the snapshot.

", - "SnapshotDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotDetail$Progress": "

The percentage of progress for the task.

", - "SnapshotDetail$StatusMessage": "

A detailed status message for the snapshot creation.

", - "SnapshotDetail$Status": "

A brief status of the snapshot creation.

", - "SnapshotDiskContainer$Description": "

The description of the disk image being imported.

", - "SnapshotDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "SnapshotDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. It can either be a https URL (https://..) or an Amazon S3 URL (s3://..).

", - "SnapshotIdStringList$member": null, - "SnapshotTaskDetail$Description": "

The description of the snapshot.

", - "SnapshotTaskDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$Url": "

The URL of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotTaskDetail$Progress": "

The percentage of completion for the import snapshot task.

", - "SnapshotTaskDetail$StatusMessage": "

A detailed status message for the import snapshot task.

", - "SnapshotTaskDetail$Status": "

A brief status for the import snapshot task.

", - "SpotDatafeedSubscription$OwnerId": "

The AWS account ID of the account.

", - "SpotDatafeedSubscription$Bucket": "

The Amazon S3 bucket where the Spot instance data feed is located.

", - "SpotDatafeedSubscription$Prefix": "

The prefix that is prepended to data feed files.

", - "SpotFleetLaunchSpecification$ImageId": "

The ID of the AMI.

", - "SpotFleetLaunchSpecification$KeyName": "

The name of the key pair.

", - "SpotFleetLaunchSpecification$UserData": "

The user data to make available to the instances. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "SpotFleetLaunchSpecification$AddressingType": "

Deprecated.

", - "SpotFleetLaunchSpecification$KernelId": "

The ID of the kernel.

", - "SpotFleetLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "SpotFleetLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances. To specify multiple subnets, separate them using commas; for example, \"subnet-a61dafcf, subnet-65ea5f08\".

", - "SpotFleetLaunchSpecification$SpotPrice": "

The bid price per unit hour for the specified instance type. If this value is not specified, the default is the Spot bid price specified for the fleet. To determine the bid price per unit hour, divide the Spot bid price by the value of WeightedCapacity.

", - "SpotFleetRequestConfig$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "SpotFleetRequestConfigData$ClientToken": "

A unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "SpotFleetRequestConfigData$SpotPrice": "

The bid price per unit hour.

", - "SpotFleetRequestConfigData$IamFleetRole": "

Grants the Spot fleet permission to terminate Spot instances on your behalf when you cancel its Spot fleet request using CancelSpotFleetRequests or when the Spot fleet request expires, if you set terminateInstancesWithExpiration.

", - "SpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "SpotInstanceRequest$SpotPrice": "

The maximum hourly price (bid) for the Spot instance launched to fulfill the request.

", - "SpotInstanceRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

", - "SpotInstanceRequest$AvailabilityZoneGroup": "

The Availability Zone group. If you specify the same Availability Zone group for all Spot instance requests, all Spot instances are launched in the same Availability Zone.

", - "SpotInstanceRequest$InstanceId": "

The instance ID, if an instance has been launched to fulfill the Spot instance request.

", - "SpotInstanceRequest$ActualBlockHourlyPrice": "

If you specified a duration and your Spot instance request was fulfilled, this is the fixed hourly price in effect for the Spot instance while it runs.

", - "SpotInstanceRequest$LaunchedAvailabilityZone": "

The Availability Zone in which the bid is launched.

", - "SpotInstanceRequestIdList$member": null, - "SpotInstanceStateFault$Code": "

The reason code for the Spot instance state change.

", - "SpotInstanceStateFault$Message": "

The message for the Spot instance state change.

", - "SpotInstanceStatus$Code": "

The status code. For a list of status codes, see Spot Bid Status Codes in the Amazon Elastic Compute Cloud User Guide.

", - "SpotInstanceStatus$Message": "

The description for the status code.

", - "SpotPlacement$AvailabilityZone": "

The Availability Zone.

[Spot fleet only] To specify multiple Availability Zones, separate them using commas; for example, \"us-west-2a, us-west-2b\".

", - "SpotPlacement$GroupName": "

The name of the placement group (for cluster instances).

", - "SpotPrice$SpotPrice": "

The maximum price (bid) that you are willing to pay for a Spot instance.

", - "SpotPrice$AvailabilityZone": "

The Availability Zone.

", - "StaleIpPermission$IpProtocol": "

The IP protocol name (for tcp, udp, and icmp) or number (see Protocol Numbers).

", - "StaleSecurityGroup$GroupId": "

The ID of the security group.

", - "StaleSecurityGroup$GroupName": "

The name of the security group.

", - "StaleSecurityGroup$Description": "

The description of the security group.

", - "StaleSecurityGroup$VpcId": "

The ID of the VPC for the security group.

", - "StartInstancesRequest$AdditionalInfo": "

Reserved.

", - "StateReason$Code": "

The reason code for the state change.

", - "StateReason$Message": "

The message for the state change.

  • Server.SpotInstanceTermination: A Spot instance was terminated due to an increase in the market price.

  • Server.InternalError: An internal error occurred during instance launch, resulting in termination.

  • Server.InsufficientInstanceCapacity: There was insufficient instance capacity to satisfy the launch request.

  • Client.InternalError: A client error caused the instance to terminate on launch.

  • Client.InstanceInitiatedShutdown: The instance was shut down using the shutdown -h command from the instance.

  • Client.UserInitiatedShutdown: The instance was shut down using the Amazon EC2 API.

  • Client.VolumeLimitExceeded: The limit on the number of EBS volumes or total storage was exceeded. Decrease usage or request an increase in your limits.

  • Client.InvalidSnapshot.NotFound: The specified snapshot was not found.

", - "Subnet$SubnetId": "

The ID of the subnet.

", - "Subnet$VpcId": "

The ID of the VPC the subnet is in.

", - "Subnet$CidrBlock": "

The CIDR block assigned to the subnet.

", - "Subnet$AvailabilityZone": "

The Availability Zone of the subnet.

", - "SubnetIdStringList$member": null, - "Tag$Key": "

The key of the tag.

Constraints: Tag keys are case-sensitive and accept a maximum of 127 Unicode characters. May not begin with aws:

", - "Tag$Value": "

The value of the tag.

Constraints: Tag values are case-sensitive and accept a maximum of 255 Unicode characters.

", - "TagDescription$ResourceId": "

The ID of the resource. For example, ami-1a2b3c4d.

", - "TagDescription$Key": "

The tag key.

", - "TagDescription$Value": "

The tag value.

", - "UnassignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "UnsuccessfulItem$ResourceId": "

The ID of the resource.

", - "UnsuccessfulItemError$Code": "

The error code.

", - "UnsuccessfulItemError$Message": "

The error message accompanying the error code.

", - "UserBucket$S3Bucket": "

The name of the S3 bucket where the disk image is located.

", - "UserBucket$S3Key": "

The file name of the disk image.

", - "UserBucketDetails$S3Bucket": "

The S3 bucket from which the disk image was created.

", - "UserBucketDetails$S3Key": "

The file name of the disk image.

", - "UserData$Data": "

The user data. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "UserGroupStringList$member": null, - "UserIdGroupPair$UserId": "

The ID of an AWS account. For a referenced security group in another VPC, the account ID of the referenced security group is returned.

[EC2-Classic] Required when adding or removing rules that reference a security group in another AWS account.

", - "UserIdGroupPair$GroupName": "

The name of the security group. In a request, use this parameter for a security group in EC2-Classic or a default VPC only. For a security group in a nondefault VPC, use the security group ID.

", - "UserIdGroupPair$GroupId": "

The ID of the security group.

", - "UserIdGroupPair$VpcId": "

The ID of the VPC for the referenced security group, if applicable.

", - "UserIdGroupPair$VpcPeeringConnectionId": "

The ID of the VPC peering connection, if applicable.

", - "UserIdGroupPair$PeeringStatus": "

The status of a VPC peering connection, if applicable.

", - "UserIdStringList$member": null, - "ValueStringList$member": null, - "VgwTelemetry$OutsideIpAddress": "

The Internet-routable IP address of the virtual private gateway's outside interface.

", - "VgwTelemetry$StatusMessage": "

If an error occurs, a description of the error.

", - "Volume$VolumeId": "

The ID of the volume.

", - "Volume$SnapshotId": "

The snapshot from which the volume was created, if applicable.

", - "Volume$AvailabilityZone": "

The Availability Zone for the volume.

", - "Volume$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the volume.

", - "VolumeAttachment$VolumeId": "

The ID of the volume.

", - "VolumeAttachment$InstanceId": "

The ID of the instance.

", - "VolumeAttachment$Device": "

The device name.

", - "VolumeIdStringList$member": null, - "VolumeStatusAction$Code": "

The code identifying the operation, for example, enable-volume-io.

", - "VolumeStatusAction$Description": "

A description of the operation.

", - "VolumeStatusAction$EventType": "

The event type associated with this operation.

", - "VolumeStatusAction$EventId": "

The ID of the event associated with this operation.

", - "VolumeStatusDetails$Status": "

The intended status of the volume status.

", - "VolumeStatusEvent$EventType": "

The type of this event.

", - "VolumeStatusEvent$Description": "

A description of the event.

", - "VolumeStatusEvent$EventId": "

The ID of this event.

", - "VolumeStatusItem$VolumeId": "

The volume ID.

", - "VolumeStatusItem$AvailabilityZone": "

The Availability Zone of the volume.

", - "Vpc$VpcId": "

The ID of the VPC.

", - "Vpc$CidrBlock": "

The CIDR block for the VPC.

", - "Vpc$DhcpOptionsId": "

The ID of the set of DHCP options you've associated with the VPC (or default if the default options are associated with the VPC).

", - "VpcAttachment$VpcId": "

The ID of the VPC.

", - "VpcClassicLink$VpcId": "

The ID of the VPC.

", - "VpcClassicLinkIdList$member": null, - "VpcEndpoint$VpcEndpointId": "

The ID of the VPC endpoint.

", - "VpcEndpoint$VpcId": "

The ID of the VPC to which the endpoint is associated.

", - "VpcEndpoint$ServiceName": "

The name of the AWS service to which the endpoint is associated.

", - "VpcEndpoint$PolicyDocument": "

The policy document associated with the endpoint.

", - "VpcIdStringList$member": null, - "VpcPeeringConnection$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "VpcPeeringConnectionStateReason$Message": "

A message that provides more information about the status, if applicable.

", - "VpcPeeringConnectionVpcInfo$CidrBlock": "

The CIDR block for the VPC.

", - "VpcPeeringConnectionVpcInfo$OwnerId": "

The AWS account ID of the VPC owner.

", - "VpcPeeringConnectionVpcInfo$VpcId": "

The ID of the VPC.

", - "VpnConnection$VpnConnectionId": "

The ID of the VPN connection.

", - "VpnConnection$CustomerGatewayConfiguration": "

The configuration information for the VPN connection's customer gateway (in the native XML format). This element is always present in the CreateVpnConnection response; however, it's present in the DescribeVpnConnections response only if the VPN connection is in the pending or available state.

", - "VpnConnection$CustomerGatewayId": "

The ID of the customer gateway at your end of the VPN connection.

", - "VpnConnection$VpnGatewayId": "

The ID of the virtual private gateway at the AWS side of the VPN connection.

", - "VpnConnectionIdStringList$member": null, - "VpnGateway$VpnGatewayId": "

The ID of the virtual private gateway.

", - "VpnGateway$AvailabilityZone": "

The Availability Zone where the virtual private gateway was created, if applicable. This field may be empty or not returned.

", - "VpnGatewayIdStringList$member": null, - "VpnStaticRoute$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer data center.

", - "ZoneNameStringList$member": null - } - }, - "Subnet": { - "base": "

Describes a subnet.

", - "refs": { - "CreateSubnetResult$Subnet": "

Information about the subnet.

", - "SubnetList$member": null - } - }, - "SubnetIdStringList": { - "base": null, - "refs": { - "DescribeSubnetsRequest$SubnetIds": "

One or more subnet IDs.

Default: Describes all your subnets.

" - } - }, - "SubnetList": { - "base": null, - "refs": { - "DescribeSubnetsResult$Subnets": "

Information about one or more subnets.

" - } - }, - "SubnetState": { - "base": null, - "refs": { - "Subnet$State": "

The current state of the subnet.

" - } - }, - "SummaryStatus": { - "base": null, - "refs": { - "InstanceStatusSummary$Status": "

The status.

" - } - }, - "Tag": { - "base": "

Describes a tag.

", - "refs": { - "TagList$member": null - } - }, - "TagDescription": { - "base": "

Describes a tag.

", - "refs": { - "TagDescriptionList$member": null - } - }, - "TagDescriptionList": { - "base": null, - "refs": { - "DescribeTagsResult$Tags": "

A list of tags.

" - } - }, - "TagList": { - "base": null, - "refs": { - "ClassicLinkInstance$Tags": "

Any tags assigned to the instance.

", - "ConversionTask$Tags": "

Any tags assigned to the task.

", - "CreateTagsRequest$Tags": "

One or more tags. The value parameter is required, but if you don't want the tag to have a value, specify the parameter with no value, and we set the value to an empty string.

", - "CustomerGateway$Tags": "

Any tags assigned to the customer gateway.

", - "DeleteTagsRequest$Tags": "

One or more tags to delete. If you omit the value parameter, we delete the tag regardless of its value. If you specify this parameter with an empty string as the value, we delete the key only if its value is an empty string.

", - "DhcpOptions$Tags": "

Any tags assigned to the DHCP options set.

", - "Image$Tags": "

Any tags assigned to the image.

", - "Instance$Tags": "

Any tags assigned to the instance.

", - "InternetGateway$Tags": "

Any tags assigned to the Internet gateway.

", - "NetworkAcl$Tags": "

Any tags assigned to the network ACL.

", - "NetworkInterface$TagSet": "

Any tags assigned to the network interface.

", - "ReservedInstances$Tags": "

Any tags assigned to the resource.

", - "ReservedInstancesListing$Tags": "

Any tags assigned to the resource.

", - "RouteTable$Tags": "

Any tags assigned to the route table.

", - "SecurityGroup$Tags": "

Any tags assigned to the security group.

", - "Snapshot$Tags": "

Any tags assigned to the snapshot.

", - "SpotInstanceRequest$Tags": "

Any tags assigned to the resource.

", - "Subnet$Tags": "

Any tags assigned to the subnet.

", - "Volume$Tags": "

Any tags assigned to the volume.

", - "Vpc$Tags": "

Any tags assigned to the VPC.

", - "VpcClassicLink$Tags": "

Any tags assigned to the VPC.

", - "VpcPeeringConnection$Tags": "

Any tags assigned to the resource.

", - "VpnConnection$Tags": "

Any tags assigned to the VPN connection.

", - "VpnGateway$Tags": "

Any tags assigned to the virtual private gateway.

" - } - }, - "TelemetryStatus": { - "base": null, - "refs": { - "VgwTelemetry$Status": "

The status of the VPN tunnel.

" - } - }, - "Tenancy": { - "base": null, - "refs": { - "CreateVpcRequest$InstanceTenancy": "

The tenancy options for instances launched into the VPC. For default, instances are launched with shared tenancy by default. You can launch instances with any tenancy into a shared tenancy VPC. For dedicated, instances are launched as dedicated tenancy instances by default. You can only launch instances with a tenancy of dedicated or host into a dedicated tenancy VPC.

Important: The host value cannot be used with this parameter. Use the default or dedicated values only.

Default: default

", - "DescribeReservedInstancesOfferingsRequest$InstanceTenancy": "

The tenancy of the instances covered by the reservation. A Reserved Instance with a tenancy of dedicated is applied to instances that run in a VPC on single-tenant hardware (i.e., Dedicated Instances).

Default: default

", - "Placement$Tenancy": "

The tenancy of the instance (if the instance is running in a VPC). An instance with a tenancy of dedicated runs on single-tenant hardware. The host tenancy is not supported for the ImportInstance command.

", - "ReservedInstances$InstanceTenancy": "

The tenancy of the instance.

", - "ReservedInstancesOffering$InstanceTenancy": "

The tenancy of the instance.

", - "Vpc$InstanceTenancy": "

The allowed tenancy of instances launched into the VPC.

" - } - }, - "TerminateInstancesRequest": { - "base": "

Contains the parameters for TerminateInstances.

", - "refs": { - } - }, - "TerminateInstancesResult": { - "base": "

Contains the output of TerminateInstances.

", - "refs": { - } - }, - "TrafficType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$TrafficType": "

The type of traffic to log.

", - "FlowLog$TrafficType": "

The type of traffic captured for the flow log.

" - } - }, - "UnassignPrivateIpAddressesRequest": { - "base": "

Contains the parameters for UnassignPrivateIpAddresses.

", - "refs": { - } - }, - "UnmonitorInstancesRequest": { - "base": "

Contains the parameters for UnmonitorInstances.

", - "refs": { - } - }, - "UnmonitorInstancesResult": { - "base": "

Contains the output of UnmonitorInstances.

", - "refs": { - } - }, - "UnsuccessfulItem": { - "base": "

Information about items that were not successfully processed in a batch call.

", - "refs": { - "UnsuccessfulItemList$member": null, - "UnsuccessfulItemSet$member": null - } - }, - "UnsuccessfulItemError": { - "base": "

Information about the error that occurred. For more information about errors, see Error Codes.

", - "refs": { - "UnsuccessfulItem$Error": "

Information about the error.

" - } - }, - "UnsuccessfulItemList": { - "base": null, - "refs": { - "ModifyHostsResult$Unsuccessful": "

The IDs of the Dedicated Hosts that could not be modified. Check whether the setting you requested can be used.

", - "ReleaseHostsResult$Unsuccessful": "

The IDs of the Dedicated Hosts that could not be released, including an error message.

" - } - }, - "UnsuccessfulItemSet": { - "base": null, - "refs": { - "CreateFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be created successfully.

", - "DeleteFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be deleted successfully.

", - "DeleteVpcEndpointsResult$Unsuccessful": "

Information about the endpoints that were not successfully deleted.

" - } - }, - "UserBucket": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "ImageDiskContainer$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotDiskContainer$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserBucketDetails": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "SnapshotDetail$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotTaskDetail$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserData": { - "base": "

Describes the user data for an instance.

", - "refs": { - "ImportInstanceLaunchSpecification$UserData": "

The user data to make available to the instance. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

" - } - }, - "UserGroupStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserGroups": "

One or more user groups. This is only valid when modifying the launchPermission attribute.

" - } - }, - "UserIdGroupPair": { - "base": "

Describes a security group and AWS account ID pair.

", - "refs": { - "UserIdGroupPairList$member": null, - "UserIdGroupPairSet$member": null - } - }, - "UserIdGroupPairList": { - "base": null, - "refs": { - "IpPermission$UserIdGroupPairs": "

One or more security group and AWS account ID pairs.

" - } - }, - "UserIdGroupPairSet": { - "base": null, - "refs": { - "StaleIpPermission$UserIdGroupPairs": "

One or more security group pairs. Returns the ID of the referenced security group and VPC, and the ID and status of the VPC peering connection.

" - } - }, - "UserIdStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserIds": "

One or more AWS account IDs. This is only valid when modifying the launchPermission attribute.

", - "ModifySnapshotAttributeRequest$UserIds": "

The account ID to modify for the snapshot.

" - } - }, - "ValueStringList": { - "base": null, - "refs": { - "CancelSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "CreateFlowLogsRequest$ResourceIds": "

One or more subnet, network interface, or VPC IDs.

Constraints: Maximum of 1000 resources

", - "CreateFlowLogsResult$FlowLogIds": "

The IDs of the flow logs.

", - "CreateVpcEndpointRequest$RouteTableIds": "

One or more route table IDs.

", - "DeleteFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DeleteVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DescribeInternetGatewaysRequest$InternetGatewayIds": "

One or more Internet gateway IDs.

Default: Describes all your Internet gateways.

", - "DescribeMovingAddressesRequest$PublicIps": "

One or more Elastic IP addresses.

", - "DescribeNatGatewaysRequest$NatGatewayIds": "

One or more NAT gateway IDs.

", - "DescribeNetworkAclsRequest$NetworkAclIds": "

One or more network ACL IDs.

Default: Describes all your network ACLs.

", - "DescribePrefixListsRequest$PrefixListIds": "

One or more prefix list IDs.

", - "DescribeRouteTablesRequest$RouteTableIds": "

One or more route table IDs.

Default: Describes all your route tables.

", - "DescribeSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "DescribeVpcEndpointServicesResult$ServiceNames": "

A list of supported AWS services.

", - "DescribeVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeVpcPeeringConnectionsRequest$VpcPeeringConnectionIds": "

One or more VPC peering connection IDs.

Default: Describes all your VPC peering connections.

", - "Filter$Values": "

One or more filter values. Filter values are case-sensitive.

", - "ModifyVpcEndpointRequest$AddRouteTableIds": "

One or more route tables IDs to associate with the endpoint.

", - "ModifyVpcEndpointRequest$RemoveRouteTableIds": "

One or more route table IDs to disassociate from the endpoint.

", - "NewDhcpConfiguration$Values": null, - "PrefixList$Cidrs": "

The IP address range of the AWS service.

", - "RequestSpotLaunchSpecification$SecurityGroups": null, - "RequestSpotLaunchSpecification$SecurityGroupIds": null, - "VpcEndpoint$RouteTableIds": "

One or more route tables associated with the endpoint.

" - } - }, - "VgwTelemetry": { - "base": "

Describes telemetry for a VPN tunnel.

", - "refs": { - "VgwTelemetryList$member": null - } - }, - "VgwTelemetryList": { - "base": null, - "refs": { - "VpnConnection$VgwTelemetry": "

Information about the VPN tunnel.

" - } - }, - "VirtualizationType": { - "base": null, - "refs": { - "Image$VirtualizationType": "

The type of virtualization of the AMI.

", - "Instance$VirtualizationType": "

The virtualization type of the instance.

" - } - }, - "Volume": { - "base": "

Describes a volume.

", - "refs": { - "VolumeList$member": null - } - }, - "VolumeAttachment": { - "base": "

Describes volume attachment details.

", - "refs": { - "VolumeAttachmentList$member": null - } - }, - "VolumeAttachmentList": { - "base": null, - "refs": { - "Volume$Attachments": "

Information about the volume attachments.

" - } - }, - "VolumeAttachmentState": { - "base": null, - "refs": { - "VolumeAttachment$State": "

The attachment state of the volume.

" - } - }, - "VolumeAttributeName": { - "base": null, - "refs": { - "DescribeVolumeAttributeRequest$Attribute": "

The instance attribute.

" - } - }, - "VolumeDetail": { - "base": "

Describes an EBS volume.

", - "refs": { - "DiskImage$Volume": "

Information about the volume.

", - "ImportVolumeRequest$Volume": "

The volume size.

" - } - }, - "VolumeIdStringList": { - "base": null, - "refs": { - "DescribeVolumeStatusRequest$VolumeIds": "

One or more volume IDs.

Default: Describes all your volumes.

", - "DescribeVolumesRequest$VolumeIds": "

One or more volume IDs.

" - } - }, - "VolumeList": { - "base": null, - "refs": { - "DescribeVolumesResult$Volumes": "

Information about the volumes.

" - } - }, - "VolumeState": { - "base": null, - "refs": { - "Volume$State": "

The volume state.

" - } - }, - "VolumeStatusAction": { - "base": "

Describes a volume status operation code.

", - "refs": { - "VolumeStatusActionsList$member": null - } - }, - "VolumeStatusActionsList": { - "base": null, - "refs": { - "VolumeStatusItem$Actions": "

The details of the operation.

" - } - }, - "VolumeStatusDetails": { - "base": "

Describes a volume status.

", - "refs": { - "VolumeStatusDetailsList$member": null - } - }, - "VolumeStatusDetailsList": { - "base": null, - "refs": { - "VolumeStatusInfo$Details": "

The details of the volume status.

" - } - }, - "VolumeStatusEvent": { - "base": "

Describes a volume status event.

", - "refs": { - "VolumeStatusEventsList$member": null - } - }, - "VolumeStatusEventsList": { - "base": null, - "refs": { - "VolumeStatusItem$Events": "

A list of events associated with the volume.

" - } - }, - "VolumeStatusInfo": { - "base": "

Describes the status of a volume.

", - "refs": { - "VolumeStatusItem$VolumeStatus": "

The volume status.

" - } - }, - "VolumeStatusInfoStatus": { - "base": null, - "refs": { - "VolumeStatusInfo$Status": "

The status of the volume.

" - } - }, - "VolumeStatusItem": { - "base": "

Describes the volume status.

", - "refs": { - "VolumeStatusList$member": null - } - }, - "VolumeStatusList": { - "base": null, - "refs": { - "DescribeVolumeStatusResult$VolumeStatuses": "

A list of volumes.

" - } - }, - "VolumeStatusName": { - "base": null, - "refs": { - "VolumeStatusDetails$Name": "

The name of the volume status.

" - } - }, - "VolumeType": { - "base": null, - "refs": { - "CreateVolumeRequest$VolumeType": "

The volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

Default: standard

", - "EbsBlockDevice$VolumeType": "

The volume type: gp2, io1, st1, sc1, or standard.

Default: standard

", - "Volume$VolumeType": "

The volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

" - } - }, - "Vpc": { - "base": "

Describes a VPC.

", - "refs": { - "CreateVpcResult$Vpc": "

Information about the VPC.

", - "VpcList$member": null - } - }, - "VpcAttachment": { - "base": "

Describes an attachment between a virtual private gateway and a VPC.

", - "refs": { - "AttachVpnGatewayResult$VpcAttachment": "

Information about the attachment.

", - "VpcAttachmentList$member": null - } - }, - "VpcAttachmentList": { - "base": null, - "refs": { - "VpnGateway$VpcAttachments": "

Any VPCs attached to the virtual private gateway.

" - } - }, - "VpcAttributeName": { - "base": null, - "refs": { - "DescribeVpcAttributeRequest$Attribute": "

The VPC attribute.

" - } - }, - "VpcClassicLink": { - "base": "

Describes whether a VPC is enabled for ClassicLink.

", - "refs": { - "VpcClassicLinkList$member": null - } - }, - "VpcClassicLinkIdList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkDnsSupportRequest$VpcIds": "

One or more VPC IDs.

", - "DescribeVpcClassicLinkRequest$VpcIds": "

One or more VPCs for which you want to describe the ClassicLink status.

" - } - }, - "VpcClassicLinkList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkResult$Vpcs": "

The ClassicLink status of one or more VPCs.

" - } - }, - "VpcEndpoint": { - "base": "

Describes a VPC endpoint.

", - "refs": { - "CreateVpcEndpointResult$VpcEndpoint": "

Information about the endpoint.

", - "VpcEndpointSet$member": null - } - }, - "VpcEndpointSet": { - "base": null, - "refs": { - "DescribeVpcEndpointsResult$VpcEndpoints": "

Information about the endpoints.

" - } - }, - "VpcIdStringList": { - "base": null, - "refs": { - "DescribeVpcsRequest$VpcIds": "

One or more VPC IDs.

Default: Describes all your VPCs.

" - } - }, - "VpcList": { - "base": null, - "refs": { - "DescribeVpcsResult$Vpcs": "

Information about one or more VPCs.

" - } - }, - "VpcPeeringConnection": { - "base": "

Describes a VPC peering connection.

", - "refs": { - "AcceptVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "CreateVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "VpcPeeringConnectionList$member": null - } - }, - "VpcPeeringConnectionList": { - "base": null, - "refs": { - "DescribeVpcPeeringConnectionsResult$VpcPeeringConnections": "

Information about the VPC peering connections.

" - } - }, - "VpcPeeringConnectionOptionsDescription": { - "base": "

Describes the VPC peering connection options.

", - "refs": { - "VpcPeeringConnectionVpcInfo$PeeringOptions": "

Information about the VPC peering connection options for the accepter or requester VPC.

" - } - }, - "VpcPeeringConnectionStateReason": { - "base": "

Describes the status of a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$Status": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionStateReasonCode": { - "base": null, - "refs": { - "VpcPeeringConnectionStateReason$Code": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionVpcInfo": { - "base": "

Describes a VPC in a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$AccepterVpcInfo": "

Information about the accepter VPC. CIDR block information is not returned when creating a VPC peering connection, or when describing a VPC peering connection that's in the initiating-request or pending-acceptance state.

", - "VpcPeeringConnection$RequesterVpcInfo": "

Information about the requester VPC.

" - } - }, - "VpcState": { - "base": null, - "refs": { - "Vpc$State": "

The current state of the VPC.

" - } - }, - "VpnConnection": { - "base": "

Describes a VPN connection.

", - "refs": { - "CreateVpnConnectionResult$VpnConnection": "

Information about the VPN connection.

", - "VpnConnectionList$member": null - } - }, - "VpnConnectionIdStringList": { - "base": null, - "refs": { - "DescribeVpnConnectionsRequest$VpnConnectionIds": "

One or more VPN connection IDs.

Default: Describes your VPN connections.

" - } - }, - "VpnConnectionList": { - "base": null, - "refs": { - "DescribeVpnConnectionsResult$VpnConnections": "

Information about one or more VPN connections.

" - } - }, - "VpnConnectionOptions": { - "base": "

Describes VPN connection options.

", - "refs": { - "VpnConnection$Options": "

The VPN connection options.

" - } - }, - "VpnConnectionOptionsSpecification": { - "base": "

Describes VPN connection options.

", - "refs": { - "CreateVpnConnectionRequest$Options": "

Indicates whether the VPN connection requires static routes. If you are creating a VPN connection for a device that does not support BGP, you must specify true.

Default: false

" - } - }, - "VpnGateway": { - "base": "

Describes a virtual private gateway.

", - "refs": { - "CreateVpnGatewayResult$VpnGateway": "

Information about the virtual private gateway.

", - "VpnGatewayList$member": null - } - }, - "VpnGatewayIdStringList": { - "base": null, - "refs": { - "DescribeVpnGatewaysRequest$VpnGatewayIds": "

One or more virtual private gateway IDs.

Default: Describes all your virtual private gateways.

" - } - }, - "VpnGatewayList": { - "base": null, - "refs": { - "DescribeVpnGatewaysResult$VpnGateways": "

Information about one or more virtual private gateways.

" - } - }, - "VpnState": { - "base": null, - "refs": { - "VpnConnection$State": "

The current state of the VPN connection.

", - "VpnGateway$State": "

The current state of the virtual private gateway.

", - "VpnStaticRoute$State": "

The current state of the static route.

" - } - }, - "VpnStaticRoute": { - "base": "

Describes a static route for a VPN connection.

", - "refs": { - "VpnStaticRouteList$member": null - } - }, - "VpnStaticRouteList": { - "base": null, - "refs": { - "VpnConnection$Routes": "

The static routes associated with the VPN connection.

" - } - }, - "VpnStaticRouteSource": { - "base": null, - "refs": { - "VpnStaticRoute$Source": "

Indicates how the routes were provided.

" - } - }, - "ZoneNameStringList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesRequest$ZoneNames": "

The names of one or more Availability Zones.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/examples-1.json deleted file mode 100644 index 3f584e9f4..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/examples-1.json +++ /dev/null @@ -1,3729 +0,0 @@ -{ - "version": "1.0", - "examples": { - "AllocateAddress": [ - { - "input": { - "Domain": "vpc" - }, - "output": { - "AllocationId": "eipalloc-64d5890a", - "Domain": "vpc", - "PublicIp": "203.0.113.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example allocates an Elastic IP address to use with an instance in a VPC.", - "id": "ec2-allocate-address-1", - "title": "To allocate an Elastic IP address for EC2-VPC" - }, - { - "output": { - "Domain": "standard", - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example allocates an Elastic IP address to use with an instance in EC2-Classic.", - "id": "ec2-allocate-address-2", - "title": "To allocate an Elastic IP address for EC2-Classic" - } - ], - "AssignPrivateIpAddresses": [ - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3", - "PrivateIpAddresses": [ - "10.0.0.82" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example assigns the specified secondary private IP address to the specified network interface.", - "id": "ec2-assign-private-ip-addresses-1", - "title": "To assign a specific secondary private IP address to an interface" - }, - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3", - "SecondaryPrivateIpAddressCount": 2 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example assigns two secondary private IP addresses to the specified network interface. Amazon EC2 automatically assigns these IP addresses from the available IP addresses in the CIDR block range of the subnet the network interface is associated with.", - "id": "ec2-assign-private-ip-addresses-2", - "title": "To assign secondary private IP addresses that Amazon EC2 selects to an interface" - } - ], - "AssociateAddress": [ - { - "input": { - "AllocationId": "eipalloc-64d5890a", - "InstanceId": "i-0b263919b6498b123" - }, - "output": { - "AssociationId": "eipassoc-2bebb745" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified Elastic IP address with the specified instance in a VPC.", - "id": "ec2-associate-address-1", - "title": "To associate an Elastic IP address in EC2-VPC" - }, - { - "input": { - "AllocationId": "eipalloc-64d5890a", - "NetworkInterfaceId": "eni-1a2b3c4d" - }, - "output": { - "AssociationId": "eipassoc-2bebb745" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified Elastic IP address with the specified network interface.", - "id": "ec2-associate-address-2", - "title": "To associate an Elastic IP address with a network interface" - }, - { - "input": { - "InstanceId": "i-07ffe74c7330ebf53", - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates an Elastic IP address with an instance in EC2-Classic.", - "id": "ec2-associate-address-3", - "title": "To associate an Elastic IP address in EC2-Classic" - } - ], - "AssociateDhcpOptions": [ - { - "input": { - "DhcpOptionsId": "dopt-d9070ebb", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified DHCP options set with the specified VPC.", - "id": "ec2-associate-dhcp-options-1", - "title": "To associate a DHCP options set with a VPC" - }, - { - "input": { - "DhcpOptionsId": "default", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the default DHCP options set with the specified VPC.", - "id": "ec2-associate-dhcp-options-2", - "title": "To associate the default DHCP options set with a VPC" - } - ], - "AssociateRouteTable": [ - { - "input": { - "RouteTableId": "rtb-22574640", - "SubnetId": "subnet-9d4a7b6" - }, - "output": { - "AssociationId": "rtbassoc-781d0d1a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified route table with the specified subnet.", - "id": "ec2-associate-route-table-1", - "title": "To associate a route table with a subnet" - } - ], - "AttachInternetGateway": [ - { - "input": { - "InternetGatewayId": "igw-c0a643a9", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example attaches the specified Internet gateway to the specified VPC.", - "id": "ec2-attach-internet-gateway-1", - "title": "To attach an Internet gateway to a VPC" - } - ], - "AttachNetworkInterface": [ - { - "input": { - "DeviceIndex": 1, - "InstanceId": "i-1234567890abcdef0", - "NetworkInterfaceId": "eni-e5aa89a3" - }, - "output": { - "AttachmentId": "eni-attach-66c4350a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example attaches the specified network interface to the specified instance.", - "id": "ec2-attach-network-interface-1", - "title": "To attach a network interface to an instance" - } - ], - "AttachVolume": [ - { - "input": { - "Device": "/dev/sdf", - "InstanceId": "i-01474ef662b89480", - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "AttachTime": "2016-08-29T18:52:32.724Z", - "Device": "/dev/sdf", - "InstanceId": "i-01474ef662b89480", - "State": "attaching", - "VolumeId": "vol-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example attaches a volume (``vol-1234567890abcdef0``) to an instance (``i-01474ef662b89480``) as ``/dev/sdf``.", - "id": "to-attach-a-volume-to-an-instance-1472499213109", - "title": "To attach a volume to an instance" - } - ], - "CancelSpotFleetRequests": [ - { - "input": { - "SpotFleetRequestIds": [ - "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - ], - "TerminateInstances": true - }, - "output": { - "SuccessfulFleetRequests": [ - { - "CurrentSpotFleetRequestState": "cancelled_running", - "PreviousSpotFleetRequestState": "active", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example cancels the specified Spot fleet request and terminates its associated Spot Instances.", - "id": "ec2-cancel-spot-fleet-requests-1", - "title": "To cancel a Spot fleet request" - }, - { - "input": { - "SpotFleetRequestIds": [ - "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - ], - "TerminateInstances": false - }, - "output": { - "SuccessfulFleetRequests": [ - { - "CurrentSpotFleetRequestState": "cancelled_terminating", - "PreviousSpotFleetRequestState": "active", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example cancels the specified Spot fleet request without terminating its associated Spot Instances.", - "id": "ec2-cancel-spot-fleet-requests-2", - "title": "To cancel a Spot fleet request without terminating its Spot Instances" - } - ], - "CancelSpotInstanceRequests": [ - { - "input": { - "SpotInstanceRequestIds": [ - "sir-08b93456" - ] - }, - "output": { - "CancelledSpotInstanceRequests": [ - { - "SpotInstanceRequestId": "sir-08b93456", - "State": "cancelled" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example cancels a Spot Instance request.", - "id": "ec2-cancel-spot-instance-requests-1", - "title": "To cancel Spot Instance requests" - } - ], - "ConfirmProductInstance": [ - { - "input": { - "InstanceId": "i-1234567890abcdef0", - "ProductCode": "774F4FF8" - }, - "output": { - "OwnerId": "123456789012" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example determines whether the specified product code is associated with the specified instance.", - "id": "to-confirm-the-product-instance-1472712108494", - "title": "To confirm the product instance" - } - ], - "CopySnapshot": [ - { - "input": { - "Description": "This is my copied snapshot.", - "DestinationRegion": "us-east-1", - "SourceRegion": "us-west-2", - "SourceSnapshotId": "snap-066877671789bd71b" - }, - "output": { - "SnapshotId": "snap-066877671789bd71b" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example copies a snapshot with the snapshot ID of ``snap-066877671789bd71b`` from the ``us-west-2`` region to the ``us-east-1`` region and adds a short description to identify the snapshot.", - "id": "to-copy-a-snapshot-1472502259774", - "title": "To copy a snapshot" - } - ], - "CreateCustomerGateway": [ - { - "input": { - "BgpAsn": 65534, - "PublicIp": "12.1.2.3", - "Type": "ipsec.1" - }, - "output": { - "CustomerGateway": { - "BgpAsn": "65534", - "CustomerGatewayId": "cgw-0e11f167", - "IpAddress": "12.1.2.3", - "State": "available", - "Type": "ipsec.1" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a customer gateway with the specified IP address for its outside interface.", - "id": "ec2-create-customer-gateway-1", - "title": "To create a customer gateway" - } - ], - "CreateDhcpOptions": [ - { - "input": { - "DhcpConfigurations": [ - { - "Key": "domain-name-servers", - "Values": [ - "10.2.5.1", - "10.2.5.2" - ] - } - ] - }, - "output": { - "DhcpOptions": { - "DhcpConfigurations": [ - { - "Key": "domain-name-servers", - "Values": [ - "10.2.5.2", - "10.2.5.1" - ] - } - ], - "DhcpOptionsId": "dopt-d9070ebb" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a DHCP options set.", - "id": "ec2-create-dhcp-options-1", - "title": "To create a DHCP options set" - } - ], - "CreateInternetGateway": [ - { - "output": { - "InternetGateway": { - "Attachments": [ - - ], - "InternetGatewayId": "igw-c0a643a9", - "Tags": [ - - ] - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates an Internet gateway.", - "id": "ec2-create-internet-gateway-1", - "title": "To create an Internet gateway" - } - ], - "CreateKeyPair": [ - { - "input": { - "KeyName": "my-key-pair" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a key pair named my-key-pair.", - "id": "ec2-create-key-pair-1", - "title": "To create a key pair" - } - ], - "CreateNatGateway": [ - { - "input": { - "AllocationId": "eipalloc-37fc1a52", - "SubnetId": "subnet-1a2b3c4d" - }, - "output": { - "NatGateway": { - "CreateTime": "2015-12-17T12:45:26.732Z", - "NatGatewayAddresses": [ - { - "AllocationId": "eipalloc-37fc1a52" - } - ], - "NatGatewayId": "nat-08d48af2a8e83edfd", - "State": "pending", - "SubnetId": "subnet-1a2b3c4d", - "VpcId": "vpc-1122aabb" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a NAT gateway in subnet subnet-1a2b3c4d and associates an Elastic IP address with the allocation ID eipalloc-37fc1a52 with the NAT gateway.", - "id": "ec2-create-nat-gateway-1", - "title": "To create a NAT gateway" - } - ], - "CreateNetworkAcl": [ - { - "input": { - "VpcId": "vpc-a01106c2" - }, - "output": { - "NetworkAcl": { - "Associations": [ - - ], - "Entries": [ - { - "CidrBlock": "0.0.0.0/0", - "Egress": true, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - }, - { - "CidrBlock": "0.0.0.0/0", - "Egress": false, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - } - ], - "IsDefault": false, - "NetworkAclId": "acl-5fb85d36", - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a network ACL for the specified VPC.", - "id": "ec2-create-network-acl-1", - "title": "To create a network ACL" - } - ], - "CreateNetworkAclEntry": [ - { - "input": { - "CidrBlock": "0.0.0.0/0", - "Egress": false, - "NetworkAclId": "acl-5fb85d36", - "PortRange": { - "From": 53, - "To": 53 - }, - "Protocol": "udp", - "RuleAction": "allow", - "RuleNumber": 100 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates an entry for the specified network ACL. The rule allows ingress traffic from anywhere (0.0.0.0/0) on UDP port 53 (DNS) into any associated subnet.", - "id": "ec2-create-network-acl-entry-1", - "title": "To create a network ACL entry" - } - ], - "CreateNetworkInterface": [ - { - "input": { - "Description": "my network interface", - "Groups": [ - "sg-903004f8" - ], - "PrivateIpAddress": "10.0.2.17", - "SubnetId": "subnet-9d4a7b6c" - }, - "output": { - "NetworkInterface": { - "AvailabilityZone": "us-east-1d", - "Description": "my network interface", - "Groups": [ - { - "GroupId": "sg-903004f8", - "GroupName": "default" - } - ], - "MacAddress": "02:1a:80:41:52:9c", - "NetworkInterfaceId": "eni-e5aa89a3", - "OwnerId": "123456789012", - "PrivateIpAddress": "10.0.2.17", - "PrivateIpAddresses": [ - { - "Primary": true, - "PrivateIpAddress": "10.0.2.17" - } - ], - "RequesterManaged": false, - "SourceDestCheck": true, - "Status": "pending", - "SubnetId": "subnet-9d4a7b6c", - "TagSet": [ - - ], - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a network interface for the specified subnet.", - "id": "ec2-create-network-interface-1", - "title": "To create a network interface" - } - ], - "CreatePlacementGroup": [ - { - "input": { - "GroupName": "my-cluster", - "Strategy": "cluster" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a placement group with the specified name.", - "id": "to-create-a-placement-group-1472712245768", - "title": "To create a placement group" - } - ], - "CreateRoute": [ - { - "input": { - "DestinationCidrBlock": "0.0.0.0/0", - "GatewayId": "igw-c0a643a9", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a route for the specified route table. The route matches all traffic (0.0.0.0/0) and routes it to the specified Internet gateway.", - "id": "ec2-create-route-1", - "title": "To create a route" - } - ], - "CreateRouteTable": [ - { - "input": { - "VpcId": "vpc-a01106c2" - }, - "output": { - "RouteTable": { - "Associations": [ - - ], - "PropagatingVgws": [ - - ], - "RouteTableId": "rtb-22574640", - "Routes": [ - { - "DestinationCidrBlock": "10.0.0.0/16", - "GatewayId": "local", - "State": "active" - } - ], - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a route table for the specified VPC.", - "id": "ec2-create-route-table-1", - "title": "To create a route table" - } - ], - "CreateSnapshot": [ - { - "input": { - "Description": "This is my root volume snapshot.", - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "Description": "This is my root volume snapshot.", - "OwnerId": "012345678910", - "SnapshotId": "snap-066877671789bd71b", - "StartTime": "2014-02-28T21:06:01.000Z", - "State": "pending", - "Tags": [ - - ], - "VolumeId": "vol-1234567890abcdef0", - "VolumeSize": 8 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a snapshot of the volume with a volume ID of ``vol-1234567890abcdef0`` and a short description to identify the snapshot.", - "id": "to-create-a-snapshot-1472502529790", - "title": "To create a snapshot" - } - ], - "CreateSpotDatafeedSubscription": [ - { - "input": { - "Bucket": "my-s3-bucket", - "Prefix": "spotdata" - }, - "output": { - "SpotDatafeedSubscription": { - "Bucket": "my-s3-bucket", - "OwnerId": "123456789012", - "Prefix": "spotdata", - "State": "Active" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot Instance data feed for your AWS account.", - "id": "ec2-create-spot-datafeed-subscription-1", - "title": "To create a Spot Instance datafeed" - } - ], - "CreateSubnet": [ - { - "input": { - "CidrBlock": "10.0.1.0/24", - "VpcId": "vpc-a01106c2" - }, - "output": { - "Subnet": { - "AvailabilityZone": "us-west-2c", - "AvailableIpAddressCount": 251, - "CidrBlock": "10.0.1.0/24", - "State": "pending", - "SubnetId": "subnet-9d4a7b6c", - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a subnet in the specified VPC with the specified CIDR block. We recommend that you let us select an Availability Zone for you.", - "id": "ec2-create-subnet-1", - "title": "To create a subnet" - } - ], - "CreateTags": [ - { - "input": { - "Resources": [ - "ami-78a54011" - ], - "Tags": [ - { - "Key": "Stack", - "Value": "production" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example adds the tag Stack=production to the specified image, or overwrites an existing tag for the AMI where the tag key is Stack.", - "id": "ec2-create-tags-1", - "title": "To add a tag to a resource" - } - ], - "CreateVolume": [ - { - "input": { - "AvailabilityZone": "us-east-1a", - "Size": 80, - "VolumeType": "gp2" - }, - "output": { - "AvailabilityZone": "us-east-1a", - "CreateTime": "2016-08-29T18:52:32.724Z", - "Encrypted": false, - "Iops": 240, - "Size": 80, - "SnapshotId": "", - "State": "creating", - "VolumeId": "vol-6b60b7c7", - "VolumeType": "gp2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates an 80 GiB General Purpose (SSD) volume in the Availability Zone ``us-east-1a``.", - "id": "to-create-a-new-volume-1472496724296", - "title": "To create a new volume" - }, - { - "input": { - "AvailabilityZone": "us-east-1a", - "Iops": 1000, - "SnapshotId": "snap-066877671789bd71b", - "VolumeType": "io1" - }, - "output": { - "Attachments": [ - - ], - "AvailabilityZone": "us-east-1a", - "CreateTime": "2016-08-29T18:52:32.724Z", - "Iops": 1000, - "Size": 500, - "SnapshotId": "snap-066877671789bd71b", - "State": "creating", - "Tags": [ - - ], - "VolumeId": "vol-1234567890abcdef0", - "VolumeType": "io1" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a new Provisioned IOPS (SSD) volume with 1000 provisioned IOPS from a snapshot in the Availability Zone ``us-east-1a``.", - "id": "to-create-a-new-provisioned-iops-ssd-volume-from-a-snapshot-1472498975176", - "title": "To create a new Provisioned IOPS (SSD) volume from a snapshot" - } - ], - "CreateVpc": [ - { - "input": { - "CidrBlock": "10.0.0.0/16" - }, - "output": { - "Vpc": { - "CidrBlock": "10.0.0.0/16", - "DhcpOptionsId": "dopt-7a8b9c2d", - "InstanceTenancy": "default", - "State": "pending", - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a VPC with the specified CIDR block.", - "id": "ec2-create-vpc-1", - "title": "To create a VPC" - } - ], - "DeleteCustomerGateway": [ - { - "input": { - "CustomerGatewayId": "cgw-0e11f167" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified customer gateway.", - "id": "ec2-delete-customer-gateway-1", - "title": "To delete a customer gateway" - } - ], - "DeleteDhcpOptions": [ - { - "input": { - "DhcpOptionsId": "dopt-d9070ebb" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified DHCP options set.", - "id": "ec2-delete-dhcp-options-1", - "title": "To delete a DHCP options set" - } - ], - "DeleteInternetGateway": [ - { - "input": { - "InternetGatewayId": "igw-c0a643a9" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified Internet gateway.", - "id": "ec2-delete-internet-gateway-1", - "title": "To delete an Internet gateway" - } - ], - "DeleteKeyPair": [ - { - "input": { - "KeyName": "my-key-pair" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified key pair.", - "id": "ec2-delete-key-pair-1", - "title": "To delete a key pair" - } - ], - "DeleteNatGateway": [ - { - "input": { - "NatGatewayId": "nat-04ae55e711cec5680" - }, - "output": { - "NatGatewayId": "nat-04ae55e711cec5680" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified NAT gateway.", - "id": "ec2-delete-nat-gateway-1", - "title": "To delete a NAT gateway" - } - ], - "DeleteNetworkAcl": [ - { - "input": { - "NetworkAclId": "acl-5fb85d36" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified network ACL.", - "id": "ec2-delete-network-acl-1", - "title": "To delete a network ACL" - } - ], - "DeleteNetworkAclEntry": [ - { - "input": { - "Egress": true, - "NetworkAclId": "acl-5fb85d36", - "RuleNumber": 100 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes ingress rule number 100 from the specified network ACL.", - "id": "ec2-delete-network-acl-entry-1", - "title": "To delete a network ACL entry" - } - ], - "DeleteNetworkInterface": [ - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified network interface.", - "id": "ec2-delete-network-interface-1", - "title": "To delete a network interface" - } - ], - "DeletePlacementGroup": [ - { - "input": { - "GroupName": "my-cluster" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified placement group.\n", - "id": "to-delete-a-placement-group-1472712349959", - "title": "To delete a placement group" - } - ], - "DeleteRoute": [ - { - "input": { - "DestinationCidrBlock": "0.0.0.0/0", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified route from the specified route table.", - "id": "ec2-delete-route-1", - "title": "To delete a route" - } - ], - "DeleteRouteTable": [ - { - "input": { - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified route table.", - "id": "ec2-delete-route-table-1", - "title": "To delete a route table" - } - ], - "DeleteSnapshot": [ - { - "input": { - "SnapshotId": "snap-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes a snapshot with the snapshot ID of ``snap-1234567890abcdef0``. If the command succeeds, no output is returned.", - "id": "to-delete-a-snapshot-1472503042567", - "title": "To delete a snapshot" - } - ], - "DeleteSpotDatafeedSubscription": [ - { - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes a Spot data feed subscription for the account.", - "id": "ec2-delete-spot-datafeed-subscription-1", - "title": "To cancel a Spot Instance data feed subscription" - } - ], - "DeleteSubnet": [ - { - "input": { - "SubnetId": "subnet-9d4a7b6c" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified subnet.", - "id": "ec2-delete-subnet-1", - "title": "To delete a subnet" - } - ], - "DeleteTags": [ - { - "input": { - "Resources": [ - "ami-78a54011" - ], - "Tags": [ - { - "Key": "Stack", - "Value": "test" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the tag Stack=test from the specified image.", - "id": "ec2-delete-tags-1", - "title": "To delete a tag from a resource" - } - ], - "DeleteVolume": [ - { - "input": { - "VolumeId": "vol-049df61146c4d7901" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes an available volume with the volume ID of ``vol-049df61146c4d7901``. If the command succeeds, no output is returned.", - "id": "to-delete-a-volume-1472503111160", - "title": "To delete a volume" - } - ], - "DeleteVpc": [ - { - "input": { - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified VPC.", - "id": "ec2-delete-vpc-1", - "title": "To delete a VPC" - } - ], - "DescribeAccountAttributes": [ - { - "input": { - "AttributeNames": [ - "supported-platforms" - ] - }, - "output": { - "AccountAttributes": [ - { - "AttributeName": "supported-platforms", - "AttributeValues": [ - { - "AttributeValue": "EC2" - }, - { - "AttributeValue": "VPC" - } - ] - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the supported-platforms attribute for your AWS account.", - "id": "ec2-describe-account-attributes-1", - "title": "To describe a single attribute for your AWS account" - }, - { - "output": { - "AccountAttributes": [ - { - "AttributeName": "supported-platforms", - "AttributeValues": [ - { - "AttributeValue": "EC2" - }, - { - "AttributeValue": "VPC" - } - ] - }, - { - "AttributeName": "vpc-max-security-groups-per-interface", - "AttributeValues": [ - { - "AttributeValue": "5" - } - ] - }, - { - "AttributeName": "max-elastic-ips", - "AttributeValues": [ - { - "AttributeValue": "5" - } - ] - }, - { - "AttributeName": "max-instances", - "AttributeValues": [ - { - "AttributeValue": "20" - } - ] - }, - { - "AttributeName": "vpc-max-elastic-ips", - "AttributeValues": [ - { - "AttributeValue": "5" - } - ] - }, - { - "AttributeName": "default-vpc", - "AttributeValues": [ - { - "AttributeValue": "none" - } - ] - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the attributes for your AWS account.", - "id": "ec2-describe-account-attributes-2", - "title": "To describe all attributes for your AWS account" - } - ], - "DescribeAddresses": [ - { - "output": { - "Addresses": [ - { - "Domain": "standard", - "InstanceId": "i-1234567890abcdef0", - "PublicIp": "198.51.100.0" - }, - { - "AllocationId": "eipalloc-12345678", - "AssociationId": "eipassoc-12345678", - "Domain": "vpc", - "InstanceId": "i-1234567890abcdef0", - "NetworkInterfaceId": "eni-12345678", - "NetworkInterfaceOwnerId": "123456789012", - "PrivateIpAddress": "10.0.1.241", - "PublicIp": "203.0.113.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes your Elastic IP addresses.", - "id": "ec2-describe-addresses-1", - "title": "To describe your Elastic IP addresses" - }, - { - "input": { - "Filters": [ - { - "Name": "domain", - "Values": [ - "vpc" - ] - } - ] - }, - "output": { - "Addresses": [ - { - "AllocationId": "eipalloc-12345678", - "AssociationId": "eipassoc-12345678", - "Domain": "vpc", - "InstanceId": "i-1234567890abcdef0", - "NetworkInterfaceId": "eni-12345678", - "NetworkInterfaceOwnerId": "123456789012", - "PrivateIpAddress": "10.0.1.241", - "PublicIp": "203.0.113.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes your Elastic IP addresses for use with instances in a VPC.", - "id": "ec2-describe-addresses-2", - "title": "To describe your Elastic IP addresses for EC2-VPC" - }, - { - "input": { - "Filters": [ - { - "Name": "domain", - "Values": [ - "standard" - ] - } - ] - }, - "output": { - "Addresses": [ - { - "Domain": "standard", - "InstanceId": "i-1234567890abcdef0", - "PublicIp": "198.51.100.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes your Elastic IP addresses for use with instances in EC2-Classic.", - "id": "ec2-describe-addresses-3", - "title": "To describe your Elastic IP addresses for EC2-Classic" - } - ], - "DescribeAvailabilityZones": [ - { - "output": { - "AvailabilityZones": [ - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1b" - }, - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1c" - }, - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1d" - }, - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1e" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the Availability Zones that are available to you. The response includes Availability Zones only for the current region.", - "id": "ec2-describe-availability-zones-1", - "title": "To describe your Availability Zones" - } - ], - "DescribeCustomerGateways": [ - { - "input": { - "CustomerGatewayIds": [ - "cgw-0e11f167" - ] - }, - "output": { - "CustomerGateways": [ - { - "BgpAsn": "65534", - "CustomerGatewayId": "cgw-0e11f167", - "IpAddress": "12.1.2.3", - "State": "available", - "Type": "ipsec.1" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified customer gateway.", - "id": "ec2-describe-customer-gateways-1", - "title": "To describe a customer gateway" - } - ], - "DescribeDhcpOptions": [ - { - "input": { - "DhcpOptionsIds": [ - "dopt-d9070ebb" - ] - }, - "output": { - "DhcpOptions": [ - { - "DhcpConfigurations": [ - { - "Key": "domain-name-servers", - "Values": [ - "10.2.5.2", - "10.2.5.1" - ] - } - ], - "DhcpOptionsId": "dopt-d9070ebb" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified DHCP options set.", - "id": "ec2-describe-dhcp-options-1", - "title": "To describe a DHCP options set" - } - ], - "DescribeInstanceAttribute": [ - { - "input": { - "Attribute": "instanceType", - "InstanceId": "i-1234567890abcdef0" - }, - "output": { - "InstanceId": "i-1234567890abcdef0", - "InstanceType": { - "Value": "t1.micro" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the instance type of the specified instance.\n", - "id": "to-describe-the-instance-type-1472712432132", - "title": "To describe the instance type" - }, - { - "input": { - "Attribute": "disableApiTermination", - "InstanceId": "i-1234567890abcdef0" - }, - "output": { - "DisableApiTermination": { - "Value": "false" - }, - "InstanceId": "i-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``disableApiTermination`` attribute of the specified instance.\n", - "id": "to-describe-the-disableapitermination-attribute-1472712533466", - "title": "To describe the disableApiTermination attribute" - }, - { - "input": { - "Attribute": "blockDeviceMapping", - "InstanceId": "i-1234567890abcdef0" - }, - "output": { - "BlockDeviceMappings": [ - { - "DeviceName": "/dev/sda1", - "Ebs": { - "AttachTime": "2013-05-17T22:42:34.000Z", - "DeleteOnTermination": true, - "Status": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - }, - { - "DeviceName": "/dev/sdf", - "Ebs": { - "AttachTime": "2013-09-10T23:07:00.000Z", - "DeleteOnTermination": false, - "Status": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - } - ], - "InstanceId": "i-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``blockDeviceMapping`` attribute of the specified instance.\n", - "id": "to-describe-the-block-device-mapping-for-an-instance-1472712645423", - "title": "To describe the block device mapping for an instance" - } - ], - "DescribeInternetGateways": [ - { - "input": { - "Filters": [ - { - "Name": "attachment.vpc-id", - "Values": [ - "vpc-a01106c2" - ] - } - ] - }, - "output": { - "InternetGateways": [ - { - "Attachments": [ - { - "State": "available", - "VpcId": "vpc-a01106c2" - } - ], - "InternetGatewayId": "igw-c0a643a9", - "Tags": [ - - ] - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the Internet gateway for the specified VPC.", - "id": "ec2-describe-internet-gateways-1", - "title": "To describe the Internet gateway for a VPC" - } - ], - "DescribeKeyPairs": [ - { - "input": { - "KeyNames": [ - "my-key-pair" - ] - }, - "output": { - "KeyPairs": [ - { - "KeyFingerprint": "1f:51:ae:28:bf:89:e9:d8:1f:25:5d:37:2d:7d:b8:ca:9f:f5:f1:6f", - "KeyName": "my-key-pair" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example displays the fingerprint for the specified key.", - "id": "ec2-describe-key-pairs-1", - "title": "To display a key pair" - } - ], - "DescribeMovingAddresses": [ - { - "output": { - "MovingAddressStatuses": [ - { - "MoveStatus": "MovingToVpc", - "PublicIp": "198.51.100.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all of your moving Elastic IP addresses.", - "id": "ec2-describe-moving-addresses-1", - "title": "To describe your moving addresses" - } - ], - "DescribeNatGateways": [ - { - "input": { - "Filters": [ - { - "Name": "vpc-id", - "Values": [ - "vpc-1a2b3c4d" - ] - } - ] - }, - "output": { - "NatGateways": [ - { - "CreateTime": "2015-12-01T12:26:55.983Z", - "NatGatewayAddresses": [ - { - "AllocationId": "eipalloc-89c620ec", - "NetworkInterfaceId": "eni-9dec76cd", - "PrivateIp": "10.0.0.149", - "PublicIp": "198.11.222.333" - } - ], - "NatGatewayId": "nat-05dba92075d71c408", - "State": "available", - "SubnetId": "subnet-847e4dc2", - "VpcId": "vpc-1a2b3c4d" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the NAT gateway for the specified VPC.", - "id": "ec2-describe-nat-gateways-1", - "title": "To describe a NAT gateway" - } - ], - "DescribeNetworkAcls": [ - { - "input": { - "NetworkAclIds": [ - "acl-5fb85d36" - ] - }, - "output": { - "NetworkAcls": [ - { - "Associations": [ - { - "NetworkAclAssociationId": "aclassoc-66ea5f0b", - "NetworkAclId": "acl-9aeb5ef7", - "SubnetId": "subnet-65ea5f08" - } - ], - "Entries": [ - { - "CidrBlock": "0.0.0.0/0", - "Egress": true, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - }, - { - "CidrBlock": "0.0.0.0/0", - "Egress": false, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - } - ], - "IsDefault": false, - "NetworkAclId": "acl-5fb85d36", - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified network ACL.", - "id": "ec2-", - "title": "To describe a network ACL" - } - ], - "DescribeNetworkInterfaceAttribute": [ - { - "input": { - "Attribute": "attachment", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "Attachment": { - "AttachTime": "2015-05-21T20:02:20.000Z", - "AttachmentId": "eni-attach-43348162", - "DeleteOnTermination": true, - "DeviceIndex": 0, - "InstanceId": "i-1234567890abcdef0", - "InstanceOwnerId": "123456789012", - "Status": "attached" - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the attachment attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-1", - "title": "To describe the attachment attribute of a network interface" - }, - { - "input": { - "Attribute": "description", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "Description": { - "Value": "My description" - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the description attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-2", - "title": "To describe the description attribute of a network interface" - }, - { - "input": { - "Attribute": "groupSet", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "Groups": [ - { - "GroupId": "sg-903004f8", - "GroupName": "my-security-group" - } - ], - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the groupSet attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-3", - "title": "To describe the groupSet attribute of a network interface" - }, - { - "input": { - "Attribute": "sourceDestCheck", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "NetworkInterfaceId": "eni-686ea200", - "SourceDestCheck": { - "Value": true - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the sourceDestCheck attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-4", - "title": "To describe the sourceDestCheck attribute of a network interface" - } - ], - "DescribeNetworkInterfaces": [ - { - "input": { - "NetworkInterfaceIds": [ - "eni-e5aa89a3" - ] - }, - "output": { - "NetworkInterfaces": [ - { - "Association": { - "AssociationId": "eipassoc-0fbb766a", - "IpOwnerId": "123456789012", - "PublicDnsName": "ec2-203-0-113-12.compute-1.amazonaws.com", - "PublicIp": "203.0.113.12" - }, - "Attachment": { - "AttachTime": "2013-11-30T23:36:42.000Z", - "AttachmentId": "eni-attach-66c4350a", - "DeleteOnTermination": false, - "DeviceIndex": 1, - "InstanceId": "i-1234567890abcdef0", - "InstanceOwnerId": "123456789012", - "Status": "attached" - }, - "AvailabilityZone": "us-east-1d", - "Description": "my network interface", - "Groups": [ - { - "GroupId": "sg-8637d3e3", - "GroupName": "default" - } - ], - "MacAddress": "02:2f:8f:b0:cf:75", - "NetworkInterfaceId": "eni-e5aa89a3", - "OwnerId": "123456789012", - "PrivateDnsName": "ip-10-0-1-17.ec2.internal", - "PrivateIpAddress": "10.0.1.17", - "PrivateIpAddresses": [ - { - "Association": { - "AssociationId": "eipassoc-0fbb766a", - "IpOwnerId": "123456789012", - "PublicDnsName": "ec2-203-0-113-12.compute-1.amazonaws.com", - "PublicIp": "203.0.113.12" - }, - "Primary": true, - "PrivateDnsName": "ip-10-0-1-17.ec2.internal", - "PrivateIpAddress": "10.0.1.17" - } - ], - "RequesterManaged": false, - "SourceDestCheck": true, - "Status": "in-use", - "SubnetId": "subnet-b61f49f0", - "TagSet": [ - - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "", - "id": "ec2-describe-network-interfaces-1", - "title": "To describe a network interface" - } - ], - "DescribeRegions": [ - { - "output": { - "Regions": [ - { - "Endpoint": "ec2.ap-south-1.amazonaws.com", - "RegionName": "ap-south-1" - }, - { - "Endpoint": "ec2.eu-west-1.amazonaws.com", - "RegionName": "eu-west-1" - }, - { - "Endpoint": "ec2.ap-southeast-1.amazonaws.com", - "RegionName": "ap-southeast-1" - }, - { - "Endpoint": "ec2.ap-southeast-2.amazonaws.com", - "RegionName": "ap-southeast-2" - }, - { - "Endpoint": "ec2.eu-central-1.amazonaws.com", - "RegionName": "eu-central-1" - }, - { - "Endpoint": "ec2.ap-northeast-2.amazonaws.com", - "RegionName": "ap-northeast-2" - }, - { - "Endpoint": "ec2.ap-northeast-1.amazonaws.com", - "RegionName": "ap-northeast-1" - }, - { - "Endpoint": "ec2.us-east-1.amazonaws.com", - "RegionName": "us-east-1" - }, - { - "Endpoint": "ec2.sa-east-1.amazonaws.com", - "RegionName": "sa-east-1" - }, - { - "Endpoint": "ec2.us-west-1.amazonaws.com", - "RegionName": "us-west-1" - }, - { - "Endpoint": "ec2.us-west-2.amazonaws.com", - "RegionName": "us-west-2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all the regions that are available to you.", - "id": "ec2-describe-regions-1", - "title": "To describe your regions" - } - ], - "DescribeRouteTables": [ - { - "input": { - "RouteTableIds": [ - "rtb-1f382e7d" - ] - }, - "output": { - "RouteTables": [ - { - "Associations": [ - { - "Main": true, - "RouteTableAssociationId": "rtbassoc-d8ccddba", - "RouteTableId": "rtb-1f382e7d" - } - ], - "PropagatingVgws": [ - - ], - "RouteTableId": "rtb-1f382e7d", - "Routes": [ - { - "DestinationCidrBlock": "10.0.0.0/16", - "GatewayId": "local", - "State": "active" - } - ], - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified route table.", - "id": "ec2-describe-route-tables-1", - "title": "To describe a route table" - } - ], - "DescribeScheduledInstanceAvailability": [ - { - "input": { - "FirstSlotStartTimeRange": { - "EarliestTime": "2016-01-31T00:00:00Z", - "LatestTime": "2016-01-31T04:00:00Z" - }, - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDays": [ - 1 - ] - } - }, - "output": { - "ScheduledInstanceAvailabilitySet": [ - { - "AvailabilityZone": "us-west-2b", - "AvailableInstanceCount": 20, - "FirstSlotStartTime": "2016-01-31T00:00:00Z", - "HourlyPrice": "0.095", - "InstanceType": "c4.large", - "MaxTermDurationInDays": 366, - "MinTermDurationInDays": 366, - "NetworkPlatform": "EC2-VPC", - "Platform": "Linux/UNIX", - "PurchaseToken": "eyJ2IjoiMSIsInMiOjEsImMiOi...", - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDaySet": [ - 1 - ], - "OccurrenceRelativeToEnd": false - }, - "SlotDurationInHours": 23, - "TotalScheduledInstanceHours": 1219 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes a schedule that occurs every week on Sunday, starting on the specified date. Note that the output contains a single schedule as an example.", - "id": "ec2-describe-scheduled-instance-availability-1", - "title": "To describe an available schedule" - } - ], - "DescribeScheduledInstances": [ - { - "input": { - "ScheduledInstanceIds": [ - "sci-1234-1234-1234-1234-123456789012" - ] - }, - "output": { - "ScheduledInstanceSet": [ - { - "AvailabilityZone": "us-west-2b", - "CreateDate": "2016-01-25T21:43:38.612Z", - "HourlyPrice": "0.095", - "InstanceCount": 1, - "InstanceType": "c4.large", - "NetworkPlatform": "EC2-VPC", - "NextSlotStartTime": "2016-01-31T09:00:00Z", - "Platform": "Linux/UNIX", - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDaySet": [ - 1 - ], - "OccurrenceRelativeToEnd": false, - "OccurrenceUnit": "" - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012", - "SlotDurationInHours": 32, - "TermEndDate": "2017-01-31T09:00:00Z", - "TermStartDate": "2016-01-31T09:00:00Z", - "TotalScheduledInstanceHours": 1696 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified Scheduled Instance.", - "id": "ec2-describe-scheduled-instances-1", - "title": "To describe your Scheduled Instances" - } - ], - "DescribeSnapshotAttribute": [ - { - "input": { - "Attribute": "createVolumePermission", - "SnapshotId": "snap-066877671789bd71b" - }, - "output": { - "CreateVolumePermissions": [ - - ], - "SnapshotId": "snap-066877671789bd71b" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``createVolumePermission`` attribute on a snapshot with the snapshot ID of ``snap-066877671789bd71b``.", - "id": "to-describe-snapshot-attributes-1472503199736", - "title": "To describe snapshot attributes" - } - ], - "DescribeSnapshots": [ - { - "input": { - "SnapshotIds": [ - "snap-1234567890abcdef0" - ] - }, - "output": { - "NextToken": "", - "Snapshots": [ - { - "Description": "This is my snapshot.", - "OwnerId": "012345678910", - "Progress": "100%", - "SnapshotId": "snap-1234567890abcdef0", - "StartTime": "2014-02-28T21:28:32.000Z", - "State": "completed", - "VolumeId": "vol-049df61146c4d7901", - "VolumeSize": 8 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes a snapshot with the snapshot ID of ``snap-1234567890abcdef0``.", - "id": "to-describe-a-snapshot-1472503807850", - "title": "To describe a snapshot" - }, - { - "input": { - "Filters": [ - { - "Name": "status", - "Values": [ - "pending" - ] - } - ], - "OwnerIds": [ - "012345678910" - ] - }, - "output": { - "NextToken": "", - "Snapshots": [ - { - "Description": "This is my copied snapshot.", - "OwnerId": "012345678910", - "Progress": "87%", - "SnapshotId": "snap-066877671789bd71b", - "StartTime": "2014-02-28T21:37:27.000Z", - "State": "pending", - "VolumeId": "vol-1234567890abcdef0", - "VolumeSize": 8 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all snapshots owned by the ID 012345678910 that are in the ``pending`` status.", - "id": "to-describe-snapshots-using-filters-1472503929793", - "title": "To describe snapshots using filters" - } - ], - "DescribeSpotDatafeedSubscription": [ - { - "output": { - "SpotDatafeedSubscription": { - "Bucket": "my-s3-bucket", - "OwnerId": "123456789012", - "Prefix": "spotdata", - "State": "Active" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the Spot Instance datafeed subscription for your AWS account.", - "id": "ec2-describe-spot-datafeed-subscription-1", - "title": "To describe the datafeed for your AWS account" - } - ], - "DescribeSpotFleetInstances": [ - { - "input": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "output": { - "ActiveInstances": [ - { - "InstanceId": "i-1234567890abcdef0", - "InstanceType": "m3.medium", - "SpotInstanceRequestId": "sir-08b93456" - } - ], - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example lists the Spot Instances associated with the specified Spot fleet.", - "id": "ec2-describe-spot-fleet-instances-1", - "title": "To describe the Spot Instances associated with a Spot fleet" - } - ], - "DescribeSpotFleetRequestHistory": [ - { - "input": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "StartTime": "2015-05-26T00:00:00Z" - }, - "output": { - "HistoryRecords": [ - { - "EventInformation": { - "EventSubType": "submitted" - }, - "EventType": "fleetRequestChange", - "Timestamp": "2015-05-26T23:17:20.697Z" - }, - { - "EventInformation": { - "EventSubType": "active" - }, - "EventType": "fleetRequestChange", - "Timestamp": "2015-05-26T23:17:20.873Z" - }, - { - "EventInformation": { - "EventSubType": "launched", - "InstanceId": "i-1234567890abcdef0" - }, - "EventType": "instanceChange", - "Timestamp": "2015-05-26T23:21:21.712Z" - }, - { - "EventInformation": { - "EventSubType": "launched", - "InstanceId": "i-1234567890abcdef1" - }, - "EventType": "instanceChange", - "Timestamp": "2015-05-26T23:21:21.816Z" - } - ], - "NextToken": "CpHNsscimcV5oH7bSbub03CI2Qms5+ypNpNm+53MNlR0YcXAkp0xFlfKf91yVxSExmbtma3awYxMFzNA663ZskT0AHtJ6TCb2Z8bQC2EnZgyELbymtWPfpZ1ZbauVg+P+TfGlWxWWB/Vr5dk5d4LfdgA/DRAHUrYgxzrEXAMPLE=", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "StartTime": "2015-05-26T00:00:00Z" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example returns the history for the specified Spot fleet starting at the specified time.", - "id": "ec2-describe-spot-fleet-request-history-1", - "title": "To describe Spot fleet history" - } - ], - "DescribeSpotFleetRequests": [ - { - "input": { - "SpotFleetRequestIds": [ - "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - ] - }, - "output": { - "SpotFleetRequestConfigs": [ - { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "EbsOptimized": false, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "cc2.8xlarge", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeleteOnTermination": false, - "DeviceIndex": 0, - "SecondaryPrivateIpAddressCount": 0, - "SubnetId": "subnet-a61dafcf" - } - ] - }, - { - "EbsOptimized": false, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "r3.8xlarge", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeleteOnTermination": false, - "DeviceIndex": 0, - "SecondaryPrivateIpAddressCount": 0, - "SubnetId": "subnet-a61dafcf" - } - ] - } - ], - "SpotPrice": "0.05", - "TargetCapacity": 20 - }, - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "SpotFleetRequestState": "active" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified Spot fleet request.", - "id": "ec2-describe-spot-fleet-requests-1", - "title": "To describe a Spot fleet request" - } - ], - "DescribeSpotInstanceRequests": [ - { - "input": { - "SpotInstanceRequestIds": [ - "sir-08b93456" - ] - }, - "output": { - "SpotInstanceRequests": [ - { - "CreateTime": "2014-04-30T18:14:55.000Z", - "InstanceId": "i-1234567890abcdef0", - "LaunchSpecification": { - "BlockDeviceMappings": [ - { - "DeviceName": "/dev/sda1", - "Ebs": { - "DeleteOnTermination": true, - "VolumeSize": 8, - "VolumeType": "standard" - } - } - ], - "EbsOptimized": false, - "ImageId": "ami-7aba833f", - "InstanceType": "m1.small", - "KeyName": "my-key-pair", - "SecurityGroups": [ - { - "GroupId": "sg-e38f24a7", - "GroupName": "my-security-group" - } - ] - }, - "LaunchedAvailabilityZone": "us-west-1b", - "ProductDescription": "Linux/UNIX", - "SpotInstanceRequestId": "sir-08b93456", - "SpotPrice": "0.010000", - "State": "active", - "Status": { - "Code": "fulfilled", - "Message": "Your Spot request is fulfilled.", - "UpdateTime": "2014-04-30T18:16:21.000Z" - }, - "Type": "one-time" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified Spot Instance request.", - "id": "ec2-describe-spot-instance-requests-1", - "title": "To describe a Spot Instance request" - } - ], - "DescribeSpotPriceHistory": [ - { - "input": { - "EndTime": "2014-01-06T08:09:10", - "InstanceTypes": [ - "m1.xlarge" - ], - "ProductDescriptions": [ - "Linux/UNIX (Amazon VPC)" - ], - "StartTime": "2014-01-06T07:08:09" - }, - "output": { - "SpotPriceHistory": [ - { - "AvailabilityZone": "us-west-1a", - "InstanceType": "m1.xlarge", - "ProductDescription": "Linux/UNIX (Amazon VPC)", - "SpotPrice": "0.080000", - "Timestamp": "2014-01-06T04:32:53.000Z" - }, - { - "AvailabilityZone": "us-west-1c", - "InstanceType": "m1.xlarge", - "ProductDescription": "Linux/UNIX (Amazon VPC)", - "SpotPrice": "0.080000", - "Timestamp": "2014-01-05T11:28:26.000Z" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example returns the Spot Price history for m1.xlarge, Linux/UNIX (Amazon VPC) instances for a particular day in January.", - "id": "ec2-describe-spot-price-history-1", - "title": "To describe Spot price history for Linux/UNIX (Amazon VPC)" - } - ], - "DescribeSubnets": [ - { - "input": { - "Filters": [ - { - "Name": "vpc-id", - "Values": [ - "vpc-a01106c2" - ] - } - ] - }, - "output": { - "Subnets": [ - { - "AvailabilityZone": "us-east-1c", - "AvailableIpAddressCount": 251, - "CidrBlock": "10.0.1.0/24", - "DefaultForAz": false, - "MapPublicIpOnLaunch": false, - "State": "available", - "SubnetId": "subnet-9d4a7b6c", - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the subnets for the specified VPC.", - "id": "ec2-describe-subnets-1", - "title": "To describe the subnets for a VPC" - } - ], - "DescribeTags": [ - { - "input": { - "Filters": [ - { - "Name": "resource-id", - "Values": [ - "i-1234567890abcdef8" - ] - } - ] - }, - "output": { - "Tags": [ - { - "Key": "Stack", - "ResourceId": "i-1234567890abcdef8", - "ResourceType": "instance", - "Value": "test" - }, - { - "Key": "Name", - "ResourceId": "i-1234567890abcdef8", - "ResourceType": "instance", - "Value": "Beta Server" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the tags for the specified instance.", - "id": "ec2-describe-tags-1", - "title": "To describe the tags for a single resource" - } - ], - "DescribeVolumeAttribute": [ - { - "input": { - "Attribute": "autoEnableIO", - "VolumeId": "vol-049df61146c4d7901" - }, - "output": { - "AutoEnableIO": { - "Value": false - }, - "VolumeId": "vol-049df61146c4d7901" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``autoEnableIo`` attribute of the volume with the ID ``vol-049df61146c4d7901``.", - "id": "to-describe-a-volume-attribute-1472505773492", - "title": "To describe a volume attribute" - } - ], - "DescribeVolumeStatus": [ - { - "input": { - "VolumeIds": [ - "vol-1234567890abcdef0" - ] - }, - "output": { - "VolumeStatuses": [ - { - "Actions": [ - - ], - "AvailabilityZone": "us-east-1a", - "Events": [ - - ], - "VolumeId": "vol-1234567890abcdef0", - "VolumeStatus": { - "Details": [ - { - "Name": "io-enabled", - "Status": "passed" - }, - { - "Name": "io-performance", - "Status": "not-applicable" - } - ], - "Status": "ok" - } - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the status for the volume ``vol-1234567890abcdef0``.", - "id": "to-describe-the-status-of-a-single-volume-1472507016193", - "title": "To describe the status of a single volume" - }, - { - "input": { - "Filters": [ - { - "Name": "volume-status.status", - "Values": [ - "impaired" - ] - } - ] - }, - "output": { - "VolumeStatuses": [ - - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the status for all volumes that are impaired. In this example output, there are no impaired volumes.", - "id": "to-describe-the-status-of-impaired-volumes-1472507239821", - "title": "To describe the status of impaired volumes" - } - ], - "DescribeVolumes": [ - { - "input": { - }, - "output": { - "NextToken": "", - "Volumes": [ - { - "Attachments": [ - { - "AttachTime": "2013-12-18T22:35:00.000Z", - "DeleteOnTermination": true, - "Device": "/dev/sda1", - "InstanceId": "i-1234567890abcdef0", - "State": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - ], - "AvailabilityZone": "us-east-1a", - "CreateTime": "2013-12-18T22:35:00.084Z", - "Size": 8, - "SnapshotId": "snap-1234567890abcdef0", - "State": "in-use", - "VolumeId": "vol-049df61146c4d7901", - "VolumeType": "standard" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all of your volumes in the default region.", - "id": "to-describe-all-volumes-1472506358883", - "title": "To describe all volumes" - }, - { - "input": { - "Filters": [ - { - "Name": "attachment.instance-id", - "Values": [ - "i-1234567890abcdef0" - ] - }, - { - "Name": "attachment.delete-on-termination", - "Values": [ - "true" - ] - } - ] - }, - "output": { - "Volumes": [ - { - "Attachments": [ - { - "AttachTime": "2013-12-18T22:35:00.000Z", - "DeleteOnTermination": true, - "Device": "/dev/sda1", - "InstanceId": "i-1234567890abcdef0", - "State": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - ], - "AvailabilityZone": "us-east-1a", - "CreateTime": "2013-12-18T22:35:00.084Z", - "Size": 8, - "SnapshotId": "snap-1234567890abcdef0", - "State": "in-use", - "VolumeId": "vol-049df61146c4d7901", - "VolumeType": "standard" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all volumes that are both attached to the instance with the ID i-1234567890abcdef0 and set to delete when the instance terminates.", - "id": "to-describe-volumes-that-are-attached-to-a-specific-instance-1472506613578", - "title": "To describe volumes that are attached to a specific instance" - } - ], - "DescribeVpcAttribute": [ - { - "input": { - "Attribute": "enableDnsSupport", - "VpcId": "vpc-a01106c2" - }, - "output": { - "EnableDnsSupport": { - "Value": true - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the enableDnsSupport attribute. This attribute indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for your instances to their corresponding IP addresses; otherwise, it does not.", - "id": "ec2-describe-vpc-attribute-1", - "title": "To describe the enableDnsSupport attribute" - }, - { - "input": { - "Attribute": "enableDnsHostnames", - "VpcId": "vpc-a01106c2" - }, - "output": { - "EnableDnsHostnames": { - "Value": true - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the enableDnsHostnames attribute. This attribute indicates whether the instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.", - "id": "ec2-describe-vpc-attribute-2", - "title": "To describe the enableDnsHostnames attribute" - } - ], - "DescribeVpcs": [ - { - "input": { - "VpcIds": [ - "vpc-a01106c2" - ] - }, - "output": { - "Vpcs": [ - { - "CidrBlock": "10.0.0.0/16", - "DhcpOptionsId": "dopt-7a8b9c2d", - "InstanceTenancy": "default", - "IsDefault": false, - "State": "available", - "Tags": [ - { - "Key": "Name", - "Value": "MyVPC" - } - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified VPC.", - "id": "ec2-describe-vpcs-1", - "title": "To describe a VPC" - } - ], - "DetachInternetGateway": [ - { - "input": { - "InternetGatewayId": "igw-c0a643a9", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example detaches the specified Internet gateway from the specified VPC.", - "id": "ec2-detach-internet-gateway-1", - "title": "To detach an Internet gateway from a VPC" - } - ], - "DetachNetworkInterface": [ - { - "input": { - "AttachmentId": "eni-attach-66c4350a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example detaches the specified network interface from its attached instance.", - "id": "ec2-detach-network-interface-1", - "title": "To detach a network interface from an instance" - } - ], - "DetachVolume": [ - { - "input": { - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "AttachTime": "2014-02-27T19:23:06.000Z", - "Device": "/dev/sdb", - "InstanceId": "i-1234567890abcdef0", - "State": "detaching", - "VolumeId": "vol-049df61146c4d7901" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example detaches the volume (``vol-049df61146c4d7901``) from the instance it is attached to.", - "id": "to-detach-a-volume-from-an-instance-1472507977694", - "title": "To detach a volume from an instance" - } - ], - "DisableVgwRoutePropagation": [ - { - "input": { - "GatewayId": "vgw-9a4cacf3", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disables the specified virtual private gateway from propagating static routes to the specified route table.", - "id": "ec2-disable-vgw-route-propagation-1", - "title": "To disable route propagation" - } - ], - "DisassociateAddress": [ - { - "input": { - "AssociationId": "eipassoc-2bebb745" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disassociates an Elastic IP address from an instance in a VPC.", - "id": "ec2-disassociate-address-1", - "title": "To disassociate an Elastic IP address in EC2-VPC" - }, - { - "input": { - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disassociates an Elastic IP address from an instance in EC2-Classic.", - "id": "ec2-disassociate-address-2", - "title": "To disassociate an Elastic IP addresses in EC2-Classic" - } - ], - "DisassociateRouteTable": [ - { - "input": { - "AssociationId": "rtbassoc-781d0d1a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disassociates the specified route table from its associated subnet.", - "id": "ec2-disassociate-route-table-1", - "title": "To disassociate a route table" - } - ], - "EnableVgwRoutePropagation": [ - { - "input": { - "GatewayId": "vgw-9a4cacf3", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example enables the specified virtual private gateway to propagate static routes to the specified route table.", - "id": "ec2-enable-vgw-route-propagation-1", - "title": "To enable route propagation" - } - ], - "EnableVolumeIO": [ - { - "input": { - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "Return": true - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example enables I/O on volume ``vol-1234567890abcdef0``.", - "id": "to-enable-io-for-a-volume-1472508114867", - "title": "To enable I/O for a volume" - } - ], - "ModifyNetworkInterfaceAttribute": [ - { - "input": { - "Attachment": { - "AttachmentId": "eni-attach-43348162", - "DeleteOnTermination": false - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the attachment attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-1", - "title": "To modify the attachment attribute of a network interface" - }, - { - "input": { - "Description": "My description", - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the description attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-2", - "title": "To modify the description attribute of a network interface" - }, - { - "input": { - "Groups": [ - "sg-903004f8", - "sg-1a2b3c4d" - ], - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example command modifies the groupSet attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-3", - "title": "To modify the groupSet attribute of a network interface" - }, - { - "input": { - "NetworkInterfaceId": "eni-686ea200", - "SourceDestCheck": false - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example command modifies the sourceDestCheck attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-4", - "title": "To modify the sourceDestCheck attribute of a network interface" - } - ], - "ModifySnapshotAttribute": [ - { - "input": { - "Attribute": "createVolumePermission", - "OperationType": "remove", - "SnapshotId": "snap-1234567890abcdef0", - "UserIds": [ - "123456789012" - ] - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies snapshot ``snap-1234567890abcdef0`` to remove the create volume permission for a user with the account ID ``123456789012``. If the command succeeds, no output is returned.", - "id": "to-modify-a-snapshot-attribute-1472508385907", - "title": "To modify a snapshot attribute" - }, - { - "input": { - "Attribute": "createVolumePermission", - "GroupNames": [ - "all" - ], - "OperationType": "add", - "SnapshotId": "snap-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example makes the snapshot ``snap-1234567890abcdef0`` public.", - "id": "to-make-a-snapshot-public-1472508470529", - "title": "To make a snapshot public" - } - ], - "ModifySpotFleetRequest": [ - { - "input": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "TargetCapacity": 20 - }, - "output": { - "Return": true - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example increases the target capacity of the specified Spot fleet request.", - "id": "ec2-modify-spot-fleet-request-1", - "title": "To increase the target capacity of a Spot fleet request" - }, - { - "input": { - "ExcessCapacityTerminationPolicy": "NoTermination ", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "TargetCapacity": 10 - }, - "output": { - "Return": true - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example decreases the target capacity of the specified Spot fleet request without terminating any Spot Instances as a result.", - "id": "ec2-modify-spot-fleet-request-2", - "title": "To decrease the target capacity of a Spot fleet request" - } - ], - "ModifySubnetAttribute": [ - { - "input": { - "MapPublicIpOnLaunch": true, - "SubnetId": "subnet-1a2b3c4d" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the specified subnet so that all instances launched into this subnet are assigned a public IP address.", - "id": "ec2-modify-subnet-attribute-1", - "title": "To change a subnet's public IP addressing behavior" - } - ], - "ModifyVolumeAttribute": [ - { - "input": { - "AutoEnableIO": { - "Value": true - }, - "DryRun": true, - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example sets the ``autoEnableIo`` attribute of the volume with the ID ``vol-1234567890abcdef0`` to ``true``. If the command succeeds, no output is returned.", - "id": "to-modify-a-volume-attribute-1472508596749", - "title": "To modify a volume attribute" - } - ], - "ModifyVpcAttribute": [ - { - "input": { - "EnableDnsSupport": { - "Value": false - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the enableDnsSupport attribute. This attribute indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for instances in the VPC to their corresponding IP addresses; otherwise, it does not.", - "id": "ec2-modify-vpc-attribute-1", - "title": "To modify the enableDnsSupport attribute" - }, - { - "input": { - "EnableDnsHostnames": { - "Value": false - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the enableDnsHostnames attribute. This attribute indicates whether instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.", - "id": "ec2-modify-vpc-attribute-2", - "title": "To modify the enableDnsHostnames attribute" - } - ], - "MoveAddressToVpc": [ - { - "input": { - "PublicIp": "54.123.4.56" - }, - "output": { - "Status": "MoveInProgress" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example moves the specified Elastic IP address to the EC2-VPC platform.", - "id": "ec2-move-address-to-vpc-1", - "title": "To move an address to EC2-VPC" - } - ], - "PurchaseScheduledInstances": [ - { - "input": { - "PurchaseRequests": [ - { - "InstanceCount": 1, - "PurchaseToken": "eyJ2IjoiMSIsInMiOjEsImMiOi..." - } - ] - }, - "output": { - "ScheduledInstanceSet": [ - { - "AvailabilityZone": "us-west-2b", - "CreateDate": "2016-01-25T21:43:38.612Z", - "HourlyPrice": "0.095", - "InstanceCount": 1, - "InstanceType": "c4.large", - "NetworkPlatform": "EC2-VPC", - "NextSlotStartTime": "2016-01-31T09:00:00Z", - "Platform": "Linux/UNIX", - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDaySet": [ - 1 - ], - "OccurrenceRelativeToEnd": false, - "OccurrenceUnit": "" - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012", - "SlotDurationInHours": 32, - "TermEndDate": "2017-01-31T09:00:00Z", - "TermStartDate": "2016-01-31T09:00:00Z", - "TotalScheduledInstanceHours": 1696 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example purchases a Scheduled Instance.", - "id": "ec2-purchase-scheduled-instances-1", - "title": "To purchase a Scheduled Instance" - } - ], - "ReleaseAddress": [ - { - "input": { - "AllocationId": "eipalloc-64d5890a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example releases an Elastic IP address for use with instances in a VPC.", - "id": "ec2-release-address-1", - "title": "To release an Elastic IP address for EC2-VPC" - }, - { - "input": { - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example releases an Elastic IP address for use with instances in EC2-Classic.", - "id": "ec2-release-address-2", - "title": "To release an Elastic IP addresses for EC2-Classic" - } - ], - "ReplaceNetworkAclAssociation": [ - { - "input": { - "AssociationId": "aclassoc-e5b95c8c", - "NetworkAclId": "acl-5fb85d36" - }, - "output": { - "NewAssociationId": "aclassoc-3999875b" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified network ACL with the subnet for the specified network ACL association.", - "id": "ec2-replace-network-acl-association-1", - "title": "To replace the network ACL associated with a subnet" - } - ], - "ReplaceNetworkAclEntry": [ - { - "input": { - "CidrBlock": "203.0.113.12/24", - "Egress": false, - "NetworkAclId": "acl-5fb85d36", - "PortRange": { - "From": 53, - "To": 53 - }, - "Protocol": "udp", - "RuleAction": "allow", - "RuleNumber": 100 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example replaces an entry for the specified network ACL. The new rule 100 allows ingress traffic from 203.0.113.12/24 on UDP port 53 (DNS) into any associated subnet.", - "id": "ec2-replace-network-acl-entry-1", - "title": "To replace a network ACL entry" - } - ], - "ReplaceRoute": [ - { - "input": { - "DestinationCidrBlock": "10.0.0.0/16", - "GatewayId": "vgw-9a4cacf3", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example replaces the specified route in the specified table table. The new route matches the specified CIDR and sends the traffic to the specified virtual private gateway.", - "id": "ec2-replace-route-1", - "title": "To replace a route" - } - ], - "ReplaceRouteTableAssociation": [ - { - "input": { - "AssociationId": "rtbassoc-781d0d1a", - "RouteTableId": "rtb-22574640" - }, - "output": { - "NewAssociationId": "rtbassoc-3a1f0f58" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified route table with the subnet for the specified route table association.", - "id": "ec2-replace-route-table-association-1", - "title": "To replace the route table associated with a subnet" - } - ], - "RequestSpotFleet": [ - { - "input": { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "SecurityGroups": [ - { - "GroupId": "sg-1a2b3c4d" - } - ], - "SubnetId": "subnet-1a2b3c4d, subnet-3c4d5e6f" - } - ], - "SpotPrice": "0.04", - "TargetCapacity": 2 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot fleet request with two launch specifications that differ only by subnet. The Spot fleet launches the instances in the specified subnet with the lowest price. If the instances are launched in a default VPC, they receive a public IP address by default. If the instances are launched in a nondefault VPC, they do not receive a public IP address by default. Note that you can't specify different subnets from the same Availability Zone in a Spot fleet request.", - "id": "ec2-request-spot-fleet-1", - "title": "To request a Spot fleet in the subnet with the lowest price" - }, - { - "input": { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "Placement": { - "AvailabilityZone": "us-west-2a, us-west-2b" - }, - "SecurityGroups": [ - { - "GroupId": "sg-1a2b3c4d" - } - ] - } - ], - "SpotPrice": "0.04", - "TargetCapacity": 2 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot fleet request with two launch specifications that differ only by Availability Zone. The Spot fleet launches the instances in the specified Availability Zone with the lowest price. If your account supports EC2-VPC only, Amazon EC2 launches the Spot instances in the default subnet of the Availability Zone. If your account supports EC2-Classic, Amazon EC2 launches the instances in EC2-Classic in the Availability Zone.", - "id": "ec2-request-spot-fleet-2", - "title": "To request a Spot fleet in the Availability Zone with the lowest price" - }, - { - "input": { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::880185128111:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeviceIndex": 0, - "Groups": [ - "sg-1a2b3c4d" - ], - "SubnetId": "subnet-1a2b3c4d" - } - ] - } - ], - "SpotPrice": "0.04", - "TargetCapacity": 2 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example assigns public addresses to instances launched in a nondefault VPC. Note that when you specify a network interface, you must include the subnet ID and security group ID using the network interface.", - "id": "ec2-request-spot-fleet-3", - "title": "To launch Spot instances in a subnet and assign them public IP addresses" - }, - { - "input": { - "SpotFleetRequestConfig": { - "AllocationStrategy": "diversified", - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "ImageId": "ami-1a2b3c4d", - "InstanceType": "c4.2xlarge", - "SubnetId": "subnet-1a2b3c4d" - }, - { - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.2xlarge", - "SubnetId": "subnet-1a2b3c4d" - }, - { - "ImageId": "ami-1a2b3c4d", - "InstanceType": "r3.2xlarge", - "SubnetId": "subnet-1a2b3c4d" - } - ], - "SpotPrice": "0.70", - "TargetCapacity": 30 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot fleet request that launches 30 instances using the diversified allocation strategy. The launch specifications differ by instance type. The Spot fleet distributes the instances across the launch specifications such that there are 10 instances of each type.", - "id": "ec2-request-spot-fleet-4", - "title": "To request a Spot fleet using the diversified allocation strategy" - } - ], - "RequestSpotInstances": [ - { - "input": { - "InstanceCount": 5, - "LaunchSpecification": { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "Placement": { - "AvailabilityZone": "us-west-2a" - }, - "SecurityGroupIds": [ - "sg-1a2b3c4d" - ] - }, - "SpotPrice": "0.03", - "Type": "one-time" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a one-time Spot Instance request for five instances in the specified Availability Zone. If your account supports EC2-VPC only, Amazon EC2 launches the instances in the default subnet of the specified Availability Zone. If your account supports EC2-Classic, Amazon EC2 launches the instances in EC2-Classic in the specified Availability Zone.", - "id": "ec2-request-spot-instances-1", - "title": "To create a one-time Spot Instance request" - }, - { - "input": { - "InstanceCount": 5, - "LaunchSpecification": { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "SecurityGroupIds": [ - "sg-1a2b3c4d" - ], - "SubnetId": "subnet-1a2b3c4d" - }, - "SpotPrice": "0.050", - "Type": "one-time" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example command creates a one-time Spot Instance request for five instances in the specified subnet. Amazon EC2 launches the instances in the specified subnet. If the VPC is a nondefault VPC, the instances do not receive a public IP address by default.", - "id": "ec2-request-spot-instances-2", - "title": "To create a one-time Spot Instance request" - } - ], - "ResetSnapshotAttribute": [ - { - "input": { - "Attribute": "createVolumePermission", - "SnapshotId": "snap-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example resets the create volume permissions for snapshot ``snap-1234567890abcdef0``. If the command succeeds, no output is returned.", - "id": "to-reset-a-snapshot-attribute-1472508825735", - "title": "To reset a snapshot attribute" - } - ], - "RestoreAddressToClassic": [ - { - "input": { - "PublicIp": "198.51.100.0" - }, - "output": { - "PublicIp": "198.51.100.0", - "Status": "MoveInProgress" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example restores the specified Elastic IP address to the EC2-Classic platform.", - "id": "ec2-restore-address-to-classic-1", - "title": "To restore an address to EC2-Classic" - } - ], - "RunScheduledInstances": [ - { - "input": { - "InstanceCount": 1, - "LaunchSpecification": { - "IamInstanceProfile": { - "Name": "my-iam-role" - }, - "ImageId": "ami-12345678", - "InstanceType": "c4.large", - "KeyName": "my-key-pair", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeviceIndex": 0, - "Groups": [ - "sg-12345678" - ], - "SubnetId": "subnet-12345678" - } - ] - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012" - }, - "output": { - "InstanceIdSet": [ - "i-1234567890abcdef0" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example launches the specified Scheduled Instance in a VPC.", - "id": "ec2-run-scheduled-instances-1", - "title": "To launch a Scheduled Instance in a VPC" - }, - { - "input": { - "InstanceCount": 1, - "LaunchSpecification": { - "IamInstanceProfile": { - "Name": "my-iam-role" - }, - "ImageId": "ami-12345678", - "InstanceType": "c4.large", - "KeyName": "my-key-pair", - "Placement": { - "AvailabilityZone": "us-west-2b" - }, - "SecurityGroupIds": [ - "sg-12345678" - ] - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012" - }, - "output": { - "InstanceIdSet": [ - "i-1234567890abcdef0" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example launches the specified Scheduled Instance in EC2-Classic.", - "id": "ec2-run-scheduled-instances-2", - "title": "To launch a Scheduled Instance in EC2-Classic" - } - ], - "UnassignPrivateIpAddresses": [ - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3", - "PrivateIpAddresses": [ - "10.0.0.82" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example unassigns the specified private IP address from the specified network interface.", - "id": "ec2-unassign-private-ip-addresses-1", - "title": "To unassign a secondary private IP address from a network interface" - } - ] - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/paginators-1.json deleted file mode 100644 index 9d04d89ab..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/paginators-1.json +++ /dev/null @@ -1,138 +0,0 @@ -{ - "pagination": { - "DescribeAccountAttributes": { - "result_key": "AccountAttributes" - }, - "DescribeAddresses": { - "result_key": "Addresses" - }, - "DescribeAvailabilityZones": { - "result_key": "AvailabilityZones" - }, - "DescribeBundleTasks": { - "result_key": "BundleTasks" - }, - "DescribeConversionTasks": { - "result_key": "ConversionTasks" - }, - "DescribeCustomerGateways": { - "result_key": "CustomerGateways" - }, - "DescribeDhcpOptions": { - "result_key": "DhcpOptions" - }, - "DescribeExportTasks": { - "result_key": "ExportTasks" - }, - "DescribeImages": { - "result_key": "Images" - }, - "DescribeInstanceStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "InstanceStatuses" - }, - "DescribeInstances": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Reservations" - }, - "DescribeInternetGateways": { - "result_key": "InternetGateways" - }, - "DescribeKeyPairs": { - "result_key": "KeyPairs" - }, - "DescribeNetworkAcls": { - "result_key": "NetworkAcls" - }, - "DescribeNetworkInterfaces": { - "result_key": "NetworkInterfaces" - }, - "DescribePlacementGroups": { - "result_key": "PlacementGroups" - }, - "DescribeRegions": { - "result_key": "Regions" - }, - "DescribeReservedInstances": { - "result_key": "ReservedInstances" - }, - "DescribeReservedInstancesListings": { - "result_key": "ReservedInstancesListings" - }, - "DescribeReservedInstancesOfferings": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "ReservedInstancesOfferings" - }, - "DescribeReservedInstancesModifications": { - "input_token": "NextToken", - "output_token": "NextToken", - "result_key": "ReservedInstancesModifications" - }, - "DescribeRouteTables": { - "result_key": "RouteTables" - }, - "DescribeSecurityGroups": { - "result_key": "SecurityGroups" - }, - "DescribeSnapshots": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Snapshots" - }, - "DescribeSpotInstanceRequests": { - "result_key": "SpotInstanceRequests" - }, - "DescribeSpotFleetRequests": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotFleetRequestConfigs" - }, - "DescribeSpotPriceHistory": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotPriceHistory" - }, - "DescribeSubnets": { - "result_key": "Subnets" - }, - "DescribeTags": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Tags" - }, - "DescribeVolumeStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "VolumeStatuses" - }, - "DescribeVolumes": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Volumes" - }, - "DescribeVpcs": { - "result_key": "Vpcs" - }, - "DescribeVpcPeeringConnections": { - "result_key": "VpcPeeringConnections" - }, - "DescribeVpnConnections": { - "result_key": "VpnConnections" - }, - "DescribeVpnGateways": { - "result_key": "VpnGateways" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/waiters-2.json deleted file mode 100644 index ecc9f1b6f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-04-01/waiters-2.json +++ /dev/null @@ -1,593 +0,0 @@ -{ - "version": 2, - "waiters": { - "InstanceExists": { - "delay": 5, - "maxAttempts": 40, - "operation": "DescribeInstances", - "acceptors": [ - { - "matcher": "path", - "expected": true, - "argument": "length(Reservations[]) > `0`", - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "BundleTaskComplete": { - "delay": 15, - "operation": "DescribeBundleTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "complete", - "matcher": "pathAll", - "state": "success", - "argument": "BundleTasks[].State" - }, - { - "expected": "failed", - "matcher": "pathAny", - "state": "failure", - "argument": "BundleTasks[].State" - } - ] - }, - "ConversionTaskCancelled": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskCompleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelled", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelling", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskDeleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "CustomerGatewayAvailable": { - "delay": 15, - "operation": "DescribeCustomerGateways", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - } - ] - }, - "ExportTaskCancelled": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ExportTaskCompleted": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ImageExists": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "matcher": "path", - "expected": true, - "argument": "length(Images[]) > `0`", - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidAMIID.NotFound", - "state": "retry" - } - ] - }, - "ImageAvailable": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "Images[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "Images[].State", - "expected": "failed" - } - ] - }, - "InstanceRunning": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "running", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "shutting-down", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "InstanceStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].InstanceStatus.Status", - "expected": "ok" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "InstanceStopped": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "stopped", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "InstanceTerminated": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "terminated", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "KeyPairExists": { - "operation": "DescribeKeyPairs", - "delay": 5, - "maxAttempts": 6, - "acceptors": [ - { - "expected": true, - "matcher": "pathAll", - "state": "success", - "argument": "length(KeyPairs[].KeyName) > `0`" - }, - { - "expected": "InvalidKeyPair.NotFound", - "matcher": "error", - "state": "retry" - } - ] - }, - "NatGatewayAvailable": { - "operation": "DescribeNatGateways", - "delay": 15, - "maxAttempts": 40, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "NatGateways[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "failed" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "deleting" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "deleted" - }, - { - "state": "retry", - "matcher": "error", - "expected": "NatGatewayNotFound" - } - ] - }, - "NetworkInterfaceAvailable": { - "operation": "DescribeNetworkInterfaces", - "delay": 20, - "maxAttempts": 10, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "NetworkInterfaces[].Status" - }, - { - "expected": "InvalidNetworkInterfaceID.NotFound", - "matcher": "error", - "state": "failure" - } - ] - }, - "PasswordDataAvailable": { - "operation": "GetPasswordData", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "path", - "argument": "length(PasswordData) > `0`", - "expected": true - } - ] - }, - "SnapshotCompleted": { - "delay": 15, - "operation": "DescribeSnapshots", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "Snapshots[].State" - } - ] - }, - "SpotInstanceRequestFulfilled": { - "operation": "DescribeSpotInstanceRequests", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "fulfilled" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "schedule-expired" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "canceled-before-fulfillment" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "bad-parameters" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "system-error" - } - ] - }, - "SubnetAvailable": { - "delay": 15, - "operation": "DescribeSubnets", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Subnets[].State" - } - ] - }, - "SystemStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].SystemStatus.Status", - "expected": "ok" - } - ] - }, - "VolumeAvailable": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VolumeDeleted": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "matcher": "error", - "expected": "InvalidVolume.NotFound", - "state": "success" - } - ] - }, - "VolumeInUse": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "in-use", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VpcAvailable": { - "delay": 15, - "operation": "DescribeVpcs", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Vpcs[].State" - } - ] - }, - "VpcExists": { - "operation": "DescribeVpcs", - "delay": 1, - "maxAttempts": 5, - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidVpcID.NotFound", - "state": "retry" - } - ] - }, - "VpnConnectionAvailable": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpnConnectionDeleted": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpcPeeringConnectionExists": { - "delay": 15, - "operation": "DescribeVpcPeeringConnections", - "maxAttempts": 40, - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidVpcPeeringConnectionID.NotFound", - "state": "retry" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/api-2.json deleted file mode 100755 index bc4f75696..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/api-2.json +++ /dev/null @@ -1,14415 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "uid":"ec2-2016-09-15", - "apiVersion":"2016-09-15", - "endpointPrefix":"ec2", - "protocol":"ec2", - "serviceAbbreviation":"Amazon EC2", - "serviceFullName":"Amazon Elastic Compute Cloud", - "signatureVersion":"v4", - "xmlNamespace":"http://ec2.amazonaws.com/doc/2016-09-15" - }, - "operations":{ - "AcceptReservedInstancesExchangeQuote":{ - "name":"AcceptReservedInstancesExchangeQuote", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AcceptReservedInstancesExchangeQuoteRequest"}, - "output":{"shape":"AcceptReservedInstancesExchangeQuoteResult"} - }, - "AcceptVpcPeeringConnection":{ - "name":"AcceptVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AcceptVpcPeeringConnectionRequest"}, - "output":{"shape":"AcceptVpcPeeringConnectionResult"} - }, - "AllocateAddress":{ - "name":"AllocateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateAddressRequest"}, - "output":{"shape":"AllocateAddressResult"} - }, - "AllocateHosts":{ - "name":"AllocateHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AllocateHostsRequest"}, - "output":{"shape":"AllocateHostsResult"} - }, - "AssignPrivateIpAddresses":{ - "name":"AssignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssignPrivateIpAddressesRequest"} - }, - "AssociateAddress":{ - "name":"AssociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateAddressRequest"}, - "output":{"shape":"AssociateAddressResult"} - }, - "AssociateDhcpOptions":{ - "name":"AssociateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateDhcpOptionsRequest"} - }, - "AssociateRouteTable":{ - "name":"AssociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AssociateRouteTableRequest"}, - "output":{"shape":"AssociateRouteTableResult"} - }, - "AttachClassicLinkVpc":{ - "name":"AttachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachClassicLinkVpcRequest"}, - "output":{"shape":"AttachClassicLinkVpcResult"} - }, - "AttachInternetGateway":{ - "name":"AttachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachInternetGatewayRequest"} - }, - "AttachNetworkInterface":{ - "name":"AttachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachNetworkInterfaceRequest"}, - "output":{"shape":"AttachNetworkInterfaceResult"} - }, - "AttachVolume":{ - "name":"AttachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVolumeRequest"}, - "output":{"shape":"VolumeAttachment"} - }, - "AttachVpnGateway":{ - "name":"AttachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachVpnGatewayRequest"}, - "output":{"shape":"AttachVpnGatewayResult"} - }, - "AuthorizeSecurityGroupEgress":{ - "name":"AuthorizeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupEgressRequest"} - }, - "AuthorizeSecurityGroupIngress":{ - "name":"AuthorizeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeSecurityGroupIngressRequest"} - }, - "BundleInstance":{ - "name":"BundleInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"BundleInstanceRequest"}, - "output":{"shape":"BundleInstanceResult"} - }, - "CancelBundleTask":{ - "name":"CancelBundleTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelBundleTaskRequest"}, - "output":{"shape":"CancelBundleTaskResult"} - }, - "CancelConversionTask":{ - "name":"CancelConversionTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelConversionRequest"} - }, - "CancelExportTask":{ - "name":"CancelExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelExportTaskRequest"} - }, - "CancelImportTask":{ - "name":"CancelImportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelImportTaskRequest"}, - "output":{"shape":"CancelImportTaskResult"} - }, - "CancelReservedInstancesListing":{ - "name":"CancelReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelReservedInstancesListingRequest"}, - "output":{"shape":"CancelReservedInstancesListingResult"} - }, - "CancelSpotFleetRequests":{ - "name":"CancelSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotFleetRequestsRequest"}, - "output":{"shape":"CancelSpotFleetRequestsResponse"} - }, - "CancelSpotInstanceRequests":{ - "name":"CancelSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CancelSpotInstanceRequestsRequest"}, - "output":{"shape":"CancelSpotInstanceRequestsResult"} - }, - "ConfirmProductInstance":{ - "name":"ConfirmProductInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ConfirmProductInstanceRequest"}, - "output":{"shape":"ConfirmProductInstanceResult"} - }, - "CopyImage":{ - "name":"CopyImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyImageRequest"}, - "output":{"shape":"CopyImageResult"} - }, - "CopySnapshot":{ - "name":"CopySnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopySnapshotRequest"}, - "output":{"shape":"CopySnapshotResult"} - }, - "CreateCustomerGateway":{ - "name":"CreateCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateCustomerGatewayRequest"}, - "output":{"shape":"CreateCustomerGatewayResult"} - }, - "CreateDhcpOptions":{ - "name":"CreateDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDhcpOptionsRequest"}, - "output":{"shape":"CreateDhcpOptionsResult"} - }, - "CreateFlowLogs":{ - "name":"CreateFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateFlowLogsRequest"}, - "output":{"shape":"CreateFlowLogsResult"} - }, - "CreateImage":{ - "name":"CreateImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateImageRequest"}, - "output":{"shape":"CreateImageResult"} - }, - "CreateInstanceExportTask":{ - "name":"CreateInstanceExportTask", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInstanceExportTaskRequest"}, - "output":{"shape":"CreateInstanceExportTaskResult"} - }, - "CreateInternetGateway":{ - "name":"CreateInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateInternetGatewayRequest"}, - "output":{"shape":"CreateInternetGatewayResult"} - }, - "CreateKeyPair":{ - "name":"CreateKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateKeyPairRequest"}, - "output":{"shape":"KeyPair"} - }, - "CreateNatGateway":{ - "name":"CreateNatGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNatGatewayRequest"}, - "output":{"shape":"CreateNatGatewayResult"} - }, - "CreateNetworkAcl":{ - "name":"CreateNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclRequest"}, - "output":{"shape":"CreateNetworkAclResult"} - }, - "CreateNetworkAclEntry":{ - "name":"CreateNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkAclEntryRequest"} - }, - "CreateNetworkInterface":{ - "name":"CreateNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateNetworkInterfaceRequest"}, - "output":{"shape":"CreateNetworkInterfaceResult"} - }, - "CreatePlacementGroup":{ - "name":"CreatePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreatePlacementGroupRequest"} - }, - "CreateReservedInstancesListing":{ - "name":"CreateReservedInstancesListing", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateReservedInstancesListingRequest"}, - "output":{"shape":"CreateReservedInstancesListingResult"} - }, - "CreateRoute":{ - "name":"CreateRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteRequest"}, - "output":{"shape":"CreateRouteResult"} - }, - "CreateRouteTable":{ - "name":"CreateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateRouteTableRequest"}, - "output":{"shape":"CreateRouteTableResult"} - }, - "CreateSecurityGroup":{ - "name":"CreateSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSecurityGroupRequest"}, - "output":{"shape":"CreateSecurityGroupResult"} - }, - "CreateSnapshot":{ - "name":"CreateSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSnapshotRequest"}, - "output":{"shape":"Snapshot"} - }, - "CreateSpotDatafeedSubscription":{ - "name":"CreateSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"CreateSpotDatafeedSubscriptionResult"} - }, - "CreateSubnet":{ - "name":"CreateSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateSubnetRequest"}, - "output":{"shape":"CreateSubnetResult"} - }, - "CreateTags":{ - "name":"CreateTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateTagsRequest"} - }, - "CreateVolume":{ - "name":"CreateVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVolumeRequest"}, - "output":{"shape":"Volume"} - }, - "CreateVpc":{ - "name":"CreateVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcRequest"}, - "output":{"shape":"CreateVpcResult"} - }, - "CreateVpcEndpoint":{ - "name":"CreateVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcEndpointRequest"}, - "output":{"shape":"CreateVpcEndpointResult"} - }, - "CreateVpcPeeringConnection":{ - "name":"CreateVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpcPeeringConnectionRequest"}, - "output":{"shape":"CreateVpcPeeringConnectionResult"} - }, - "CreateVpnConnection":{ - "name":"CreateVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRequest"}, - "output":{"shape":"CreateVpnConnectionResult"} - }, - "CreateVpnConnectionRoute":{ - "name":"CreateVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnConnectionRouteRequest"} - }, - "CreateVpnGateway":{ - "name":"CreateVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateVpnGatewayRequest"}, - "output":{"shape":"CreateVpnGatewayResult"} - }, - "DeleteCustomerGateway":{ - "name":"DeleteCustomerGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteCustomerGatewayRequest"} - }, - "DeleteDhcpOptions":{ - "name":"DeleteDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDhcpOptionsRequest"} - }, - "DeleteFlowLogs":{ - "name":"DeleteFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteFlowLogsRequest"}, - "output":{"shape":"DeleteFlowLogsResult"} - }, - "DeleteInternetGateway":{ - "name":"DeleteInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteInternetGatewayRequest"} - }, - "DeleteKeyPair":{ - "name":"DeleteKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteKeyPairRequest"} - }, - "DeleteNatGateway":{ - "name":"DeleteNatGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNatGatewayRequest"}, - "output":{"shape":"DeleteNatGatewayResult"} - }, - "DeleteNetworkAcl":{ - "name":"DeleteNetworkAcl", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclRequest"} - }, - "DeleteNetworkAclEntry":{ - "name":"DeleteNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkAclEntryRequest"} - }, - "DeleteNetworkInterface":{ - "name":"DeleteNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteNetworkInterfaceRequest"} - }, - "DeletePlacementGroup":{ - "name":"DeletePlacementGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeletePlacementGroupRequest"} - }, - "DeleteRoute":{ - "name":"DeleteRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteRequest"} - }, - "DeleteRouteTable":{ - "name":"DeleteRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRouteTableRequest"} - }, - "DeleteSecurityGroup":{ - "name":"DeleteSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSecurityGroupRequest"} - }, - "DeleteSnapshot":{ - "name":"DeleteSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSnapshotRequest"} - }, - "DeleteSpotDatafeedSubscription":{ - "name":"DeleteSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSpotDatafeedSubscriptionRequest"} - }, - "DeleteSubnet":{ - "name":"DeleteSubnet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteSubnetRequest"} - }, - "DeleteTags":{ - "name":"DeleteTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteTagsRequest"} - }, - "DeleteVolume":{ - "name":"DeleteVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVolumeRequest"} - }, - "DeleteVpc":{ - "name":"DeleteVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcRequest"} - }, - "DeleteVpcEndpoints":{ - "name":"DeleteVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcEndpointsRequest"}, - "output":{"shape":"DeleteVpcEndpointsResult"} - }, - "DeleteVpcPeeringConnection":{ - "name":"DeleteVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpcPeeringConnectionRequest"}, - "output":{"shape":"DeleteVpcPeeringConnectionResult"} - }, - "DeleteVpnConnection":{ - "name":"DeleteVpnConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRequest"} - }, - "DeleteVpnConnectionRoute":{ - "name":"DeleteVpnConnectionRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnConnectionRouteRequest"} - }, - "DeleteVpnGateway":{ - "name":"DeleteVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteVpnGatewayRequest"} - }, - "DeregisterImage":{ - "name":"DeregisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeregisterImageRequest"} - }, - "DescribeAccountAttributes":{ - "name":"DescribeAccountAttributes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAccountAttributesRequest"}, - "output":{"shape":"DescribeAccountAttributesResult"} - }, - "DescribeAddresses":{ - "name":"DescribeAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAddressesRequest"}, - "output":{"shape":"DescribeAddressesResult"} - }, - "DescribeAvailabilityZones":{ - "name":"DescribeAvailabilityZones", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAvailabilityZonesRequest"}, - "output":{"shape":"DescribeAvailabilityZonesResult"} - }, - "DescribeBundleTasks":{ - "name":"DescribeBundleTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeBundleTasksRequest"}, - "output":{"shape":"DescribeBundleTasksResult"} - }, - "DescribeClassicLinkInstances":{ - "name":"DescribeClassicLinkInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeClassicLinkInstancesRequest"}, - "output":{"shape":"DescribeClassicLinkInstancesResult"} - }, - "DescribeConversionTasks":{ - "name":"DescribeConversionTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeConversionTasksRequest"}, - "output":{"shape":"DescribeConversionTasksResult"} - }, - "DescribeCustomerGateways":{ - "name":"DescribeCustomerGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeCustomerGatewaysRequest"}, - "output":{"shape":"DescribeCustomerGatewaysResult"} - }, - "DescribeDhcpOptions":{ - "name":"DescribeDhcpOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDhcpOptionsRequest"}, - "output":{"shape":"DescribeDhcpOptionsResult"} - }, - "DescribeExportTasks":{ - "name":"DescribeExportTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeExportTasksRequest"}, - "output":{"shape":"DescribeExportTasksResult"} - }, - "DescribeFlowLogs":{ - "name":"DescribeFlowLogs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeFlowLogsRequest"}, - "output":{"shape":"DescribeFlowLogsResult"} - }, - "DescribeHostReservationOfferings":{ - "name":"DescribeHostReservationOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostReservationOfferingsRequest"}, - "output":{"shape":"DescribeHostReservationOfferingsResult"} - }, - "DescribeHostReservations":{ - "name":"DescribeHostReservations", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostReservationsRequest"}, - "output":{"shape":"DescribeHostReservationsResult"} - }, - "DescribeHosts":{ - "name":"DescribeHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeHostsRequest"}, - "output":{"shape":"DescribeHostsResult"} - }, - "DescribeIdFormat":{ - "name":"DescribeIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeIdFormatRequest"}, - "output":{"shape":"DescribeIdFormatResult"} - }, - "DescribeIdentityIdFormat":{ - "name":"DescribeIdentityIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeIdentityIdFormatRequest"}, - "output":{"shape":"DescribeIdentityIdFormatResult"} - }, - "DescribeImageAttribute":{ - "name":"DescribeImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImageAttributeRequest"}, - "output":{"shape":"ImageAttribute"} - }, - "DescribeImages":{ - "name":"DescribeImages", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImagesRequest"}, - "output":{"shape":"DescribeImagesResult"} - }, - "DescribeImportImageTasks":{ - "name":"DescribeImportImageTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportImageTasksRequest"}, - "output":{"shape":"DescribeImportImageTasksResult"} - }, - "DescribeImportSnapshotTasks":{ - "name":"DescribeImportSnapshotTasks", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeImportSnapshotTasksRequest"}, - "output":{"shape":"DescribeImportSnapshotTasksResult"} - }, - "DescribeInstanceAttribute":{ - "name":"DescribeInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceAttributeRequest"}, - "output":{"shape":"InstanceAttribute"} - }, - "DescribeInstanceStatus":{ - "name":"DescribeInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstanceStatusRequest"}, - "output":{"shape":"DescribeInstanceStatusResult"} - }, - "DescribeInstances":{ - "name":"DescribeInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInstancesRequest"}, - "output":{"shape":"DescribeInstancesResult"} - }, - "DescribeInternetGateways":{ - "name":"DescribeInternetGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeInternetGatewaysRequest"}, - "output":{"shape":"DescribeInternetGatewaysResult"} - }, - "DescribeKeyPairs":{ - "name":"DescribeKeyPairs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeKeyPairsRequest"}, - "output":{"shape":"DescribeKeyPairsResult"} - }, - "DescribeMovingAddresses":{ - "name":"DescribeMovingAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeMovingAddressesRequest"}, - "output":{"shape":"DescribeMovingAddressesResult"} - }, - "DescribeNatGateways":{ - "name":"DescribeNatGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNatGatewaysRequest"}, - "output":{"shape":"DescribeNatGatewaysResult"} - }, - "DescribeNetworkAcls":{ - "name":"DescribeNetworkAcls", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkAclsRequest"}, - "output":{"shape":"DescribeNetworkAclsResult"} - }, - "DescribeNetworkInterfaceAttribute":{ - "name":"DescribeNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfaceAttributeRequest"}, - "output":{"shape":"DescribeNetworkInterfaceAttributeResult"} - }, - "DescribeNetworkInterfaces":{ - "name":"DescribeNetworkInterfaces", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeNetworkInterfacesRequest"}, - "output":{"shape":"DescribeNetworkInterfacesResult"} - }, - "DescribePlacementGroups":{ - "name":"DescribePlacementGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePlacementGroupsRequest"}, - "output":{"shape":"DescribePlacementGroupsResult"} - }, - "DescribePrefixLists":{ - "name":"DescribePrefixLists", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribePrefixListsRequest"}, - "output":{"shape":"DescribePrefixListsResult"} - }, - "DescribeRegions":{ - "name":"DescribeRegions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRegionsRequest"}, - "output":{"shape":"DescribeRegionsResult"} - }, - "DescribeReservedInstances":{ - "name":"DescribeReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesRequest"}, - "output":{"shape":"DescribeReservedInstancesResult"} - }, - "DescribeReservedInstancesListings":{ - "name":"DescribeReservedInstancesListings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesListingsRequest"}, - "output":{"shape":"DescribeReservedInstancesListingsResult"} - }, - "DescribeReservedInstancesModifications":{ - "name":"DescribeReservedInstancesModifications", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesModificationsRequest"}, - "output":{"shape":"DescribeReservedInstancesModificationsResult"} - }, - "DescribeReservedInstancesOfferings":{ - "name":"DescribeReservedInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedInstancesOfferingsRequest"}, - "output":{"shape":"DescribeReservedInstancesOfferingsResult"} - }, - "DescribeRouteTables":{ - "name":"DescribeRouteTables", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRouteTablesRequest"}, - "output":{"shape":"DescribeRouteTablesResult"} - }, - "DescribeScheduledInstanceAvailability":{ - "name":"DescribeScheduledInstanceAvailability", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeScheduledInstanceAvailabilityRequest"}, - "output":{"shape":"DescribeScheduledInstanceAvailabilityResult"} - }, - "DescribeScheduledInstances":{ - "name":"DescribeScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeScheduledInstancesRequest"}, - "output":{"shape":"DescribeScheduledInstancesResult"} - }, - "DescribeSecurityGroupReferences":{ - "name":"DescribeSecurityGroupReferences", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupReferencesRequest"}, - "output":{"shape":"DescribeSecurityGroupReferencesResult"} - }, - "DescribeSecurityGroups":{ - "name":"DescribeSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSecurityGroupsRequest"}, - "output":{"shape":"DescribeSecurityGroupsResult"} - }, - "DescribeSnapshotAttribute":{ - "name":"DescribeSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotAttributeRequest"}, - "output":{"shape":"DescribeSnapshotAttributeResult"} - }, - "DescribeSnapshots":{ - "name":"DescribeSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSnapshotsRequest"}, - "output":{"shape":"DescribeSnapshotsResult"} - }, - "DescribeSpotDatafeedSubscription":{ - "name":"DescribeSpotDatafeedSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotDatafeedSubscriptionRequest"}, - "output":{"shape":"DescribeSpotDatafeedSubscriptionResult"} - }, - "DescribeSpotFleetInstances":{ - "name":"DescribeSpotFleetInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetInstancesRequest"}, - "output":{"shape":"DescribeSpotFleetInstancesResponse"} - }, - "DescribeSpotFleetRequestHistory":{ - "name":"DescribeSpotFleetRequestHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestHistoryRequest"}, - "output":{"shape":"DescribeSpotFleetRequestHistoryResponse"} - }, - "DescribeSpotFleetRequests":{ - "name":"DescribeSpotFleetRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotFleetRequestsRequest"}, - "output":{"shape":"DescribeSpotFleetRequestsResponse"} - }, - "DescribeSpotInstanceRequests":{ - "name":"DescribeSpotInstanceRequests", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotInstanceRequestsRequest"}, - "output":{"shape":"DescribeSpotInstanceRequestsResult"} - }, - "DescribeSpotPriceHistory":{ - "name":"DescribeSpotPriceHistory", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSpotPriceHistoryRequest"}, - "output":{"shape":"DescribeSpotPriceHistoryResult"} - }, - "DescribeStaleSecurityGroups":{ - "name":"DescribeStaleSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeStaleSecurityGroupsRequest"}, - "output":{"shape":"DescribeStaleSecurityGroupsResult"} - }, - "DescribeSubnets":{ - "name":"DescribeSubnets", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeSubnetsRequest"}, - "output":{"shape":"DescribeSubnetsResult"} - }, - "DescribeTags":{ - "name":"DescribeTags", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeTagsRequest"}, - "output":{"shape":"DescribeTagsResult"} - }, - "DescribeVolumeAttribute":{ - "name":"DescribeVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeAttributeRequest"}, - "output":{"shape":"DescribeVolumeAttributeResult"} - }, - "DescribeVolumeStatus":{ - "name":"DescribeVolumeStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumeStatusRequest"}, - "output":{"shape":"DescribeVolumeStatusResult"} - }, - "DescribeVolumes":{ - "name":"DescribeVolumes", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVolumesRequest"}, - "output":{"shape":"DescribeVolumesResult"} - }, - "DescribeVpcAttribute":{ - "name":"DescribeVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcAttributeRequest"}, - "output":{"shape":"DescribeVpcAttributeResult"} - }, - "DescribeVpcClassicLink":{ - "name":"DescribeVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkRequest"}, - "output":{"shape":"DescribeVpcClassicLinkResult"} - }, - "DescribeVpcClassicLinkDnsSupport":{ - "name":"DescribeVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"DescribeVpcClassicLinkDnsSupportResult"} - }, - "DescribeVpcEndpointServices":{ - "name":"DescribeVpcEndpointServices", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointServicesRequest"}, - "output":{"shape":"DescribeVpcEndpointServicesResult"} - }, - "DescribeVpcEndpoints":{ - "name":"DescribeVpcEndpoints", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcEndpointsRequest"}, - "output":{"shape":"DescribeVpcEndpointsResult"} - }, - "DescribeVpcPeeringConnections":{ - "name":"DescribeVpcPeeringConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcPeeringConnectionsRequest"}, - "output":{"shape":"DescribeVpcPeeringConnectionsResult"} - }, - "DescribeVpcs":{ - "name":"DescribeVpcs", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpcsRequest"}, - "output":{"shape":"DescribeVpcsResult"} - }, - "DescribeVpnConnections":{ - "name":"DescribeVpnConnections", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnConnectionsRequest"}, - "output":{"shape":"DescribeVpnConnectionsResult"} - }, - "DescribeVpnGateways":{ - "name":"DescribeVpnGateways", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeVpnGatewaysRequest"}, - "output":{"shape":"DescribeVpnGatewaysResult"} - }, - "DetachClassicLinkVpc":{ - "name":"DetachClassicLinkVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachClassicLinkVpcRequest"}, - "output":{"shape":"DetachClassicLinkVpcResult"} - }, - "DetachInternetGateway":{ - "name":"DetachInternetGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachInternetGatewayRequest"} - }, - "DetachNetworkInterface":{ - "name":"DetachNetworkInterface", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachNetworkInterfaceRequest"} - }, - "DetachVolume":{ - "name":"DetachVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVolumeRequest"}, - "output":{"shape":"VolumeAttachment"} - }, - "DetachVpnGateway":{ - "name":"DetachVpnGateway", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachVpnGatewayRequest"} - }, - "DisableVgwRoutePropagation":{ - "name":"DisableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVgwRoutePropagationRequest"} - }, - "DisableVpcClassicLink":{ - "name":"DisableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkRequest"}, - "output":{"shape":"DisableVpcClassicLinkResult"} - }, - "DisableVpcClassicLinkDnsSupport":{ - "name":"DisableVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"DisableVpcClassicLinkDnsSupportResult"} - }, - "DisassociateAddress":{ - "name":"DisassociateAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateAddressRequest"} - }, - "DisassociateRouteTable":{ - "name":"DisassociateRouteTable", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisassociateRouteTableRequest"} - }, - "EnableVgwRoutePropagation":{ - "name":"EnableVgwRoutePropagation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVgwRoutePropagationRequest"} - }, - "EnableVolumeIO":{ - "name":"EnableVolumeIO", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVolumeIORequest"} - }, - "EnableVpcClassicLink":{ - "name":"EnableVpcClassicLink", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkRequest"}, - "output":{"shape":"EnableVpcClassicLinkResult"} - }, - "EnableVpcClassicLinkDnsSupport":{ - "name":"EnableVpcClassicLinkDnsSupport", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableVpcClassicLinkDnsSupportRequest"}, - "output":{"shape":"EnableVpcClassicLinkDnsSupportResult"} - }, - "GetConsoleOutput":{ - "name":"GetConsoleOutput", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleOutputRequest"}, - "output":{"shape":"GetConsoleOutputResult"} - }, - "GetConsoleScreenshot":{ - "name":"GetConsoleScreenshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetConsoleScreenshotRequest"}, - "output":{"shape":"GetConsoleScreenshotResult"} - }, - "GetHostReservationPurchasePreview":{ - "name":"GetHostReservationPurchasePreview", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetHostReservationPurchasePreviewRequest"}, - "output":{"shape":"GetHostReservationPurchasePreviewResult"} - }, - "GetPasswordData":{ - "name":"GetPasswordData", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetPasswordDataRequest"}, - "output":{"shape":"GetPasswordDataResult"} - }, - "GetReservedInstancesExchangeQuote":{ - "name":"GetReservedInstancesExchangeQuote", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetReservedInstancesExchangeQuoteRequest"}, - "output":{"shape":"GetReservedInstancesExchangeQuoteResult"} - }, - "ImportImage":{ - "name":"ImportImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportImageRequest"}, - "output":{"shape":"ImportImageResult"} - }, - "ImportInstance":{ - "name":"ImportInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportInstanceRequest"}, - "output":{"shape":"ImportInstanceResult"} - }, - "ImportKeyPair":{ - "name":"ImportKeyPair", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportKeyPairRequest"}, - "output":{"shape":"ImportKeyPairResult"} - }, - "ImportSnapshot":{ - "name":"ImportSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportSnapshotRequest"}, - "output":{"shape":"ImportSnapshotResult"} - }, - "ImportVolume":{ - "name":"ImportVolume", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ImportVolumeRequest"}, - "output":{"shape":"ImportVolumeResult"} - }, - "ModifyHosts":{ - "name":"ModifyHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyHostsRequest"}, - "output":{"shape":"ModifyHostsResult"} - }, - "ModifyIdFormat":{ - "name":"ModifyIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyIdFormatRequest"} - }, - "ModifyIdentityIdFormat":{ - "name":"ModifyIdentityIdFormat", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyIdentityIdFormatRequest"} - }, - "ModifyImageAttribute":{ - "name":"ModifyImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyImageAttributeRequest"} - }, - "ModifyInstanceAttribute":{ - "name":"ModifyInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstanceAttributeRequest"} - }, - "ModifyInstancePlacement":{ - "name":"ModifyInstancePlacement", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyInstancePlacementRequest"}, - "output":{"shape":"ModifyInstancePlacementResult"} - }, - "ModifyNetworkInterfaceAttribute":{ - "name":"ModifyNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyNetworkInterfaceAttributeRequest"} - }, - "ModifyReservedInstances":{ - "name":"ModifyReservedInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyReservedInstancesRequest"}, - "output":{"shape":"ModifyReservedInstancesResult"} - }, - "ModifySnapshotAttribute":{ - "name":"ModifySnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySnapshotAttributeRequest"} - }, - "ModifySpotFleetRequest":{ - "name":"ModifySpotFleetRequest", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySpotFleetRequestRequest"}, - "output":{"shape":"ModifySpotFleetRequestResponse"} - }, - "ModifySubnetAttribute":{ - "name":"ModifySubnetAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifySubnetAttributeRequest"} - }, - "ModifyVolumeAttribute":{ - "name":"ModifyVolumeAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVolumeAttributeRequest"} - }, - "ModifyVpcAttribute":{ - "name":"ModifyVpcAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcAttributeRequest"} - }, - "ModifyVpcEndpoint":{ - "name":"ModifyVpcEndpoint", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcEndpointRequest"}, - "output":{"shape":"ModifyVpcEndpointResult"} - }, - "ModifyVpcPeeringConnectionOptions":{ - "name":"ModifyVpcPeeringConnectionOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyVpcPeeringConnectionOptionsRequest"}, - "output":{"shape":"ModifyVpcPeeringConnectionOptionsResult"} - }, - "MonitorInstances":{ - "name":"MonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MonitorInstancesRequest"}, - "output":{"shape":"MonitorInstancesResult"} - }, - "MoveAddressToVpc":{ - "name":"MoveAddressToVpc", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"MoveAddressToVpcRequest"}, - "output":{"shape":"MoveAddressToVpcResult"} - }, - "PurchaseHostReservation":{ - "name":"PurchaseHostReservation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseHostReservationRequest"}, - "output":{"shape":"PurchaseHostReservationResult"} - }, - "PurchaseReservedInstancesOffering":{ - "name":"PurchaseReservedInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedInstancesOfferingRequest"}, - "output":{"shape":"PurchaseReservedInstancesOfferingResult"} - }, - "PurchaseScheduledInstances":{ - "name":"PurchaseScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseScheduledInstancesRequest"}, - "output":{"shape":"PurchaseScheduledInstancesResult"} - }, - "RebootInstances":{ - "name":"RebootInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootInstancesRequest"} - }, - "RegisterImage":{ - "name":"RegisterImage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RegisterImageRequest"}, - "output":{"shape":"RegisterImageResult"} - }, - "RejectVpcPeeringConnection":{ - "name":"RejectVpcPeeringConnection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RejectVpcPeeringConnectionRequest"}, - "output":{"shape":"RejectVpcPeeringConnectionResult"} - }, - "ReleaseAddress":{ - "name":"ReleaseAddress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseAddressRequest"} - }, - "ReleaseHosts":{ - "name":"ReleaseHosts", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReleaseHostsRequest"}, - "output":{"shape":"ReleaseHostsResult"} - }, - "ReplaceNetworkAclAssociation":{ - "name":"ReplaceNetworkAclAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclAssociationRequest"}, - "output":{"shape":"ReplaceNetworkAclAssociationResult"} - }, - "ReplaceNetworkAclEntry":{ - "name":"ReplaceNetworkAclEntry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceNetworkAclEntryRequest"} - }, - "ReplaceRoute":{ - "name":"ReplaceRoute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteRequest"} - }, - "ReplaceRouteTableAssociation":{ - "name":"ReplaceRouteTableAssociation", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReplaceRouteTableAssociationRequest"}, - "output":{"shape":"ReplaceRouteTableAssociationResult"} - }, - "ReportInstanceStatus":{ - "name":"ReportInstanceStatus", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ReportInstanceStatusRequest"} - }, - "RequestSpotFleet":{ - "name":"RequestSpotFleet", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotFleetRequest"}, - "output":{"shape":"RequestSpotFleetResponse"} - }, - "RequestSpotInstances":{ - "name":"RequestSpotInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RequestSpotInstancesRequest"}, - "output":{"shape":"RequestSpotInstancesResult"} - }, - "ResetImageAttribute":{ - "name":"ResetImageAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetImageAttributeRequest"} - }, - "ResetInstanceAttribute":{ - "name":"ResetInstanceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetInstanceAttributeRequest"} - }, - "ResetNetworkInterfaceAttribute":{ - "name":"ResetNetworkInterfaceAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetNetworkInterfaceAttributeRequest"} - }, - "ResetSnapshotAttribute":{ - "name":"ResetSnapshotAttribute", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetSnapshotAttributeRequest"} - }, - "RestoreAddressToClassic":{ - "name":"RestoreAddressToClassic", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreAddressToClassicRequest"}, - "output":{"shape":"RestoreAddressToClassicResult"} - }, - "RevokeSecurityGroupEgress":{ - "name":"RevokeSecurityGroupEgress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupEgressRequest"} - }, - "RevokeSecurityGroupIngress":{ - "name":"RevokeSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeSecurityGroupIngressRequest"} - }, - "RunInstances":{ - "name":"RunInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunInstancesRequest"}, - "output":{"shape":"Reservation"} - }, - "RunScheduledInstances":{ - "name":"RunScheduledInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunScheduledInstancesRequest"}, - "output":{"shape":"RunScheduledInstancesResult"} - }, - "StartInstances":{ - "name":"StartInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StartInstancesRequest"}, - "output":{"shape":"StartInstancesResult"} - }, - "StopInstances":{ - "name":"StopInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StopInstancesRequest"}, - "output":{"shape":"StopInstancesResult"} - }, - "TerminateInstances":{ - "name":"TerminateInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"TerminateInstancesRequest"}, - "output":{"shape":"TerminateInstancesResult"} - }, - "UnassignPrivateIpAddresses":{ - "name":"UnassignPrivateIpAddresses", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnassignPrivateIpAddressesRequest"} - }, - "UnmonitorInstances":{ - "name":"UnmonitorInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UnmonitorInstancesRequest"}, - "output":{"shape":"UnmonitorInstancesResult"} - } - }, - "shapes":{ - "AcceptReservedInstancesExchangeQuoteRequest":{ - "type":"structure", - "required":["ReservedInstanceIds"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ReservedInstanceIds":{ - "shape":"ReservedInstanceIdSet", - "locationName":"ReservedInstanceId" - }, - "TargetConfigurations":{ - "shape":"TargetConfigurationRequestSet", - "locationName":"TargetConfiguration" - } - } - }, - "AcceptReservedInstancesExchangeQuoteResult":{ - "type":"structure", - "members":{ - "ExchangeId":{ - "shape":"String", - "locationName":"exchangeId" - } - } - }, - "AcceptVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "AcceptVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "AccountAttribute":{ - "type":"structure", - "members":{ - "AttributeName":{ - "shape":"String", - "locationName":"attributeName" - }, - "AttributeValues":{ - "shape":"AccountAttributeValueList", - "locationName":"attributeValueSet" - } - } - }, - "AccountAttributeList":{ - "type":"list", - "member":{ - "shape":"AccountAttribute", - "locationName":"item" - } - }, - "AccountAttributeName":{ - "type":"string", - "enum":[ - "supported-platforms", - "default-vpc" - ] - }, - "AccountAttributeNameStringList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeName", - "locationName":"attributeName" - } - }, - "AccountAttributeValue":{ - "type":"structure", - "members":{ - "AttributeValue":{ - "shape":"String", - "locationName":"attributeValue" - } - } - }, - "AccountAttributeValueList":{ - "type":"list", - "member":{ - "shape":"AccountAttributeValue", - "locationName":"item" - } - }, - "ActiveInstance":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - } - } - }, - "ActiveInstanceSet":{ - "type":"list", - "member":{ - "shape":"ActiveInstance", - "locationName":"item" - } - }, - "ActivityStatus":{ - "type":"string", - "enum":[ - "error", - "pending_fulfillment", - "pending_termination", - "fulfilled" - ] - }, - "Address":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "NetworkInterfaceOwnerId":{ - "shape":"String", - "locationName":"networkInterfaceOwnerId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "AddressList":{ - "type":"list", - "member":{ - "shape":"Address", - "locationName":"item" - } - }, - "Affinity":{ - "type":"string", - "enum":[ - "default", - "host" - ] - }, - "AllocateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Domain":{"shape":"DomainType"} - } - }, - "AllocateAddressResult":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "Domain":{ - "shape":"DomainType", - "locationName":"domain" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - } - } - }, - "AllocateHostsRequest":{ - "type":"structure", - "required":[ - "InstanceType", - "Quantity", - "AvailabilityZone" - ], - "members":{ - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Quantity":{ - "shape":"Integer", - "locationName":"quantity" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "AllocateHostsResult":{ - "type":"structure", - "members":{ - "HostIds":{ - "shape":"ResponseHostIdList", - "locationName":"hostIdSet" - } - } - }, - "AllocationIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"AllocationId" - } - }, - "AllocationState":{ - "type":"string", - "enum":[ - "available", - "under-assessment", - "permanent-failure", - "released", - "released-permanent-failure" - ] - }, - "AllocationStrategy":{ - "type":"string", - "enum":[ - "lowestPrice", - "diversified" - ] - }, - "ArchitectureValues":{ - "type":"string", - "enum":[ - "i386", - "x86_64" - ] - }, - "AssignPrivateIpAddressesRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AllowReassignment":{ - "shape":"Boolean", - "locationName":"allowReassignment" - } - } - }, - "AssociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "AllowReassociation":{ - "shape":"Boolean", - "locationName":"allowReassociation" - } - } - }, - "AssociateAddressResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AssociateDhcpOptionsRequest":{ - "type":"structure", - "required":[ - "DhcpOptionsId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AssociateRouteTableRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "AssociateRouteTableResult":{ - "type":"structure", - "members":{ - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "AttachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId", - "Groups" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"SecurityGroupId" - } - } - }, - "AttachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "AttachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "AttachNetworkInterfaceRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "InstanceId", - "DeviceIndex" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - } - } - }, - "AttachNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - } - } - }, - "AttachVolumeRequest":{ - "type":"structure", - "required":[ - "VolumeId", - "InstanceId", - "Device" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"} - } - }, - "AttachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "AttachVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpcAttachment":{ - "shape":"VpcAttachment", - "locationName":"attachment" - } - } - }, - "AttachmentStatus":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "AttributeBooleanValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Boolean", - "locationName":"value" - } - } - }, - "AttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "AuthorizeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "AuthorizeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "AutoPlacement":{ - "type":"string", - "enum":[ - "on", - "off" - ] - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "ZoneName":{ - "shape":"String", - "locationName":"zoneName" - }, - "State":{ - "shape":"AvailabilityZoneState", - "locationName":"zoneState" - }, - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Messages":{ - "shape":"AvailabilityZoneMessageList", - "locationName":"messageSet" - } - } - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"item" - } - }, - "AvailabilityZoneMessage":{ - "type":"structure", - "members":{ - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "AvailabilityZoneMessageList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZoneMessage", - "locationName":"item" - } - }, - "AvailabilityZoneState":{ - "type":"string", - "enum":[ - "available", - "information", - "impaired", - "unavailable" - ] - }, - "AvailableCapacity":{ - "type":"structure", - "members":{ - "AvailableInstanceCapacity":{ - "shape":"AvailableInstanceCapacityList", - "locationName":"availableInstanceCapacity" - }, - "AvailableVCpus":{ - "shape":"Integer", - "locationName":"availableVCpus" - } - } - }, - "AvailableInstanceCapacityList":{ - "type":"list", - "member":{ - "shape":"InstanceCapacity", - "locationName":"item" - } - }, - "BatchState":{ - "type":"string", - "enum":[ - "submitted", - "active", - "cancelled", - "failed", - "cancelled_running", - "cancelled_terminating", - "modifying" - ] - }, - "Blob":{"type":"blob"}, - "BlobAttributeValue":{ - "type":"structure", - "members":{ - "Value":{ - "shape":"Blob", - "locationName":"value" - } - } - }, - "BlockDeviceMapping":{ - "type":"structure", - "members":{ - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsBlockDevice", - "locationName":"ebs" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "BlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"item" - } - }, - "BlockDeviceMappingRequestList":{ - "type":"list", - "member":{ - "shape":"BlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "Boolean":{"type":"boolean"}, - "BundleIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"BundleId" - } - }, - "BundleInstanceRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Storage" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"}, - "Storage":{"shape":"Storage"} - } - }, - "BundleInstanceResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "BundleTask":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "BundleId":{ - "shape":"String", - "locationName":"bundleId" - }, - "State":{ - "shape":"BundleTaskState", - "locationName":"state" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Storage":{ - "shape":"Storage", - "locationName":"storage" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "BundleTaskError":{ - "shape":"BundleTaskError", - "locationName":"error" - } - } - }, - "BundleTaskError":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "BundleTaskList":{ - "type":"list", - "member":{ - "shape":"BundleTask", - "locationName":"item" - } - }, - "BundleTaskState":{ - "type":"string", - "enum":[ - "pending", - "waiting-for-shutdown", - "bundling", - "storing", - "cancelling", - "complete", - "failed" - ] - }, - "CancelBatchErrorCode":{ - "type":"string", - "enum":[ - "fleetRequestIdDoesNotExist", - "fleetRequestIdMalformed", - "fleetRequestNotInCancellableState", - "unexpectedError" - ] - }, - "CancelBundleTaskRequest":{ - "type":"structure", - "required":["BundleId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleId":{"shape":"String"} - } - }, - "CancelBundleTaskResult":{ - "type":"structure", - "members":{ - "BundleTask":{ - "shape":"BundleTask", - "locationName":"bundleInstanceTask" - } - } - }, - "CancelConversionRequest":{ - "type":"structure", - "required":["ConversionTaskId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ReasonMessage":{ - "shape":"String", - "locationName":"reasonMessage" - } - } - }, - "CancelExportTaskRequest":{ - "type":"structure", - "required":["ExportTaskId"], - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - } - } - }, - "CancelImportTaskRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskId":{"shape":"String"}, - "CancelReason":{"shape":"String"} - } - }, - "CancelImportTaskResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "PreviousState":{ - "shape":"String", - "locationName":"previousState" - } - } - }, - "CancelReservedInstancesListingRequest":{ - "type":"structure", - "required":["ReservedInstancesListingId"], - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - } - } - }, - "CancelReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CancelSpotFleetRequestsError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"CancelBatchErrorCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "CancelSpotFleetRequestsErrorItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "Error" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "Error":{ - "shape":"CancelSpotFleetRequestsError", - "locationName":"error" - } - } - }, - "CancelSpotFleetRequestsErrorSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsErrorItem", - "locationName":"item" - } - }, - "CancelSpotFleetRequestsRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestIds", - "TerminateInstances" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "TerminateInstances":{ - "shape":"Boolean", - "locationName":"terminateInstances" - } - } - }, - "CancelSpotFleetRequestsResponse":{ - "type":"structure", - "members":{ - "UnsuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsErrorSet", - "locationName":"unsuccessfulFleetRequestSet" - }, - "SuccessfulFleetRequests":{ - "shape":"CancelSpotFleetRequestsSuccessSet", - "locationName":"successfulFleetRequestSet" - } - } - }, - "CancelSpotFleetRequestsSuccessItem":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "CurrentSpotFleetRequestState", - "PreviousSpotFleetRequestState" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "CurrentSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"currentSpotFleetRequestState" - }, - "PreviousSpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"previousSpotFleetRequestState" - } - } - }, - "CancelSpotFleetRequestsSuccessSet":{ - "type":"list", - "member":{ - "shape":"CancelSpotFleetRequestsSuccessItem", - "locationName":"item" - } - }, - "CancelSpotInstanceRequestState":{ - "type":"string", - "enum":[ - "active", - "open", - "closed", - "cancelled", - "completed" - ] - }, - "CancelSpotInstanceRequestsRequest":{ - "type":"structure", - "required":["SpotInstanceRequestIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - } - } - }, - "CancelSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "CancelledSpotInstanceRequests":{ - "shape":"CancelledSpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "CancelledSpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "State":{ - "shape":"CancelSpotInstanceRequestState", - "locationName":"state" - } - } - }, - "CancelledSpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"CancelledSpotInstanceRequest", - "locationName":"item" - } - }, - "ClassicLinkDnsSupport":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkDnsSupported":{ - "shape":"Boolean", - "locationName":"classicLinkDnsSupported" - } - } - }, - "ClassicLinkDnsSupportList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkDnsSupport", - "locationName":"item" - } - }, - "ClassicLinkInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ClassicLinkInstanceList":{ - "type":"list", - "member":{ - "shape":"ClassicLinkInstance", - "locationName":"item" - } - }, - "ClientData":{ - "type":"structure", - "members":{ - "UploadStart":{"shape":"DateTime"}, - "UploadEnd":{"shape":"DateTime"}, - "UploadSize":{"shape":"Double"}, - "Comment":{"shape":"String"} - } - }, - "ConfirmProductInstanceRequest":{ - "type":"structure", - "required":[ - "ProductCode", - "InstanceId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ProductCode":{"shape":"String"}, - "InstanceId":{"shape":"String"} - } - }, - "ConfirmProductInstanceResult":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ContainerFormat":{ - "type":"string", - "enum":["ova"] - }, - "ConversionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "ConversionTask":{ - "type":"structure", - "required":[ - "ConversionTaskId", - "State" - ], - "members":{ - "ConversionTaskId":{ - "shape":"String", - "locationName":"conversionTaskId" - }, - "ExpirationTime":{ - "shape":"String", - "locationName":"expirationTime" - }, - "ImportInstance":{ - "shape":"ImportInstanceTaskDetails", - "locationName":"importInstance" - }, - "ImportVolume":{ - "shape":"ImportVolumeTaskDetails", - "locationName":"importVolume" - }, - "State":{ - "shape":"ConversionTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "ConversionTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "CopyImageRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceImageId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceImageId":{"shape":"String"}, - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "ClientToken":{"shape":"String"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopyImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CopySnapshotRequest":{ - "type":"structure", - "required":[ - "SourceRegion", - "SourceSnapshotId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SourceRegion":{"shape":"String"}, - "SourceSnapshotId":{"shape":"String"}, - "Description":{"shape":"String"}, - "DestinationRegion":{ - "shape":"String", - "locationName":"destinationRegion" - }, - "PresignedUrl":{ - "shape":"String", - "locationName":"presignedUrl" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "CopySnapshotResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - } - } - }, - "CreateCustomerGatewayRequest":{ - "type":"structure", - "required":[ - "Type", - "PublicIp", - "BgpAsn" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "PublicIp":{ - "shape":"String", - "locationName":"IpAddress" - }, - "BgpAsn":{"shape":"Integer"} - } - }, - "CreateCustomerGatewayResult":{ - "type":"structure", - "members":{ - "CustomerGateway":{ - "shape":"CustomerGateway", - "locationName":"customerGateway" - } - } - }, - "CreateDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpConfigurations"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpConfigurations":{ - "shape":"NewDhcpConfigurationList", - "locationName":"dhcpConfiguration" - } - } - }, - "CreateDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptions", - "locationName":"dhcpOptions" - } - } - }, - "CreateFlowLogsRequest":{ - "type":"structure", - "required":[ - "ResourceIds", - "ResourceType", - "TrafficType", - "LogGroupName", - "DeliverLogsPermissionArn" - ], - "members":{ - "ResourceIds":{ - "shape":"ValueStringList", - "locationName":"ResourceId" - }, - "ResourceType":{"shape":"FlowLogsResourceType"}, - "TrafficType":{"shape":"TrafficType"}, - "LogGroupName":{"shape":"String"}, - "DeliverLogsPermissionArn":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"flowLogIdSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "CreateImageRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Name" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NoReboot":{ - "shape":"Boolean", - "locationName":"noReboot" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"blockDeviceMapping" - } - } - }, - "CreateImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "CreateInstanceExportTaskRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "Description":{ - "shape":"String", - "locationName":"description" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - }, - "ExportToS3Task":{ - "shape":"ExportToS3TaskSpecification", - "locationName":"exportToS3" - } - } - }, - "CreateInstanceExportTaskResult":{ - "type":"structure", - "members":{ - "ExportTask":{ - "shape":"ExportTask", - "locationName":"exportTask" - } - } - }, - "CreateInternetGatewayRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateInternetGatewayResult":{ - "type":"structure", - "members":{ - "InternetGateway":{ - "shape":"InternetGateway", - "locationName":"internetGateway" - } - } - }, - "CreateKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "CreateNatGatewayRequest":{ - "type":"structure", - "required":[ - "SubnetId", - "AllocationId" - ], - "members":{ - "SubnetId":{"shape":"String"}, - "AllocationId":{"shape":"String"}, - "ClientToken":{"shape":"String"} - } - }, - "CreateNatGatewayResult":{ - "type":"structure", - "members":{ - "NatGateway":{ - "shape":"NatGateway", - "locationName":"natGateway" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "CreateNetworkAclRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateNetworkAclResult":{ - "type":"structure", - "members":{ - "NetworkAcl":{ - "shape":"NetworkAcl", - "locationName":"networkAcl" - } - } - }, - "CreateNetworkInterfaceRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "CreateNetworkInterfaceResult":{ - "type":"structure", - "members":{ - "NetworkInterface":{ - "shape":"NetworkInterface", - "locationName":"networkInterface" - } - } - }, - "CreatePlacementGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Strategy" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - } - } - }, - "CreateReservedInstancesListingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesId", - "InstanceCount", - "PriceSchedules", - "ClientToken" - ], - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "PriceSchedules":{ - "shape":"PriceScheduleSpecificationList", - "locationName":"priceSchedules" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateReservedInstancesListingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "CreateRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "CreateRouteResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "CreateRouteTableRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "CreateRouteTableResult":{ - "type":"structure", - "members":{ - "RouteTable":{ - "shape":"RouteTable", - "locationName":"routeTable" - } - } - }, - "CreateSecurityGroupRequest":{ - "type":"structure", - "required":[ - "GroupName", - "Description" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "Description":{ - "shape":"String", - "locationName":"GroupDescription" - }, - "VpcId":{"shape":"String"} - } - }, - "CreateSecurityGroupResult":{ - "type":"structure", - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "CreateSnapshotRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Description":{"shape":"String"} - } - }, - "CreateSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "required":["Bucket"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - } - } - }, - "CreateSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "CreateSubnetRequest":{ - "type":"structure", - "required":[ - "VpcId", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "CidrBlock":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateSubnetResult":{ - "type":"structure", - "members":{ - "Subnet":{ - "shape":"Subnet", - "locationName":"subnet" - } - } - }, - "CreateTagsRequest":{ - "type":"structure", - "required":[ - "Resources", - "Tags" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"ResourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"Tag" - } - } - }, - "CreateVolumePermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "CreateVolumePermissionList":{ - "type":"list", - "member":{ - "shape":"CreateVolumePermission", - "locationName":"item" - } - }, - "CreateVolumePermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"CreateVolumePermissionList"}, - "Remove":{"shape":"CreateVolumePermissionList"} - } - }, - "CreateVolumeRequest":{ - "type":"structure", - "required":["AvailabilityZone"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Size":{"shape":"Integer"}, - "SnapshotId":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "VolumeType":{"shape":"VolumeType"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{"shape":"String"} - } - }, - "CreateVpcEndpointRequest":{ - "type":"structure", - "required":[ - "VpcId", - "ServiceName" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "ServiceName":{"shape":"String"}, - "PolicyDocument":{"shape":"String"}, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "ClientToken":{"shape":"String"} - } - }, - "CreateVpcEndpointResult":{ - "type":"structure", - "members":{ - "VpcEndpoint":{ - "shape":"VpcEndpoint", - "locationName":"vpcEndpoint" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "CreateVpcPeeringConnectionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeerVpcId":{ - "shape":"String", - "locationName":"peerVpcId" - }, - "PeerOwnerId":{ - "shape":"String", - "locationName":"peerOwnerId" - } - } - }, - "CreateVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnection":{ - "shape":"VpcPeeringConnection", - "locationName":"vpcPeeringConnection" - } - } - }, - "CreateVpcRequest":{ - "type":"structure", - "required":["CidrBlock"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CidrBlock":{"shape":"String"}, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - } - } - }, - "CreateVpcResult":{ - "type":"structure", - "members":{ - "Vpc":{ - "shape":"Vpc", - "locationName":"vpc" - } - } - }, - "CreateVpnConnectionRequest":{ - "type":"structure", - "required":[ - "Type", - "CustomerGatewayId", - "VpnGatewayId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"String"}, - "CustomerGatewayId":{"shape":"String"}, - "VpnGatewayId":{"shape":"String"}, - "Options":{ - "shape":"VpnConnectionOptionsSpecification", - "locationName":"options" - } - } - }, - "CreateVpnConnectionResult":{ - "type":"structure", - "members":{ - "VpnConnection":{ - "shape":"VpnConnection", - "locationName":"vpnConnection" - } - } - }, - "CreateVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "CreateVpnGatewayRequest":{ - "type":"structure", - "required":["Type"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Type":{"shape":"GatewayType"}, - "AvailabilityZone":{"shape":"String"} - } - }, - "CreateVpnGatewayResult":{ - "type":"structure", - "members":{ - "VpnGateway":{ - "shape":"VpnGateway", - "locationName":"vpnGateway" - } - } - }, - "CurrencyCodeValues":{ - "type":"string", - "enum":["USD"] - }, - "CustomerGateway":{ - "type":"structure", - "members":{ - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "State":{ - "shape":"String", - "locationName":"state" - }, - "Type":{ - "shape":"String", - "locationName":"type" - }, - "IpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "BgpAsn":{ - "shape":"String", - "locationName":"bgpAsn" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "CustomerGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"CustomerGatewayId" - } - }, - "CustomerGatewayList":{ - "type":"list", - "member":{ - "shape":"CustomerGateway", - "locationName":"item" - } - }, - "DatafeedSubscriptionState":{ - "type":"string", - "enum":[ - "Active", - "Inactive" - ] - }, - "DateTime":{"type":"timestamp"}, - "DeleteCustomerGatewayRequest":{ - "type":"structure", - "required":["CustomerGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayId":{"shape":"String"} - } - }, - "DeleteDhcpOptionsRequest":{ - "type":"structure", - "required":["DhcpOptionsId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsId":{"shape":"String"} - } - }, - "DeleteFlowLogsRequest":{ - "type":"structure", - "required":["FlowLogIds"], - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - } - } - }, - "DeleteFlowLogsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteInternetGatewayRequest":{ - "type":"structure", - "required":["InternetGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - } - } - }, - "DeleteKeyPairRequest":{ - "type":"structure", - "required":["KeyName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{"shape":"String"} - } - }, - "DeleteNatGatewayRequest":{ - "type":"structure", - "required":["NatGatewayId"], - "members":{ - "NatGatewayId":{"shape":"String"} - } - }, - "DeleteNatGatewayResult":{ - "type":"structure", - "members":{ - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "DeleteNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Egress" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - } - } - }, - "DeleteNetworkAclRequest":{ - "type":"structure", - "required":["NetworkAclId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "DeleteNetworkInterfaceRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "DeletePlacementGroupRequest":{ - "type":"structure", - "required":["GroupName"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "DeleteRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - } - } - }, - "DeleteRouteTableRequest":{ - "type":"structure", - "required":["RouteTableId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "DeleteSecurityGroupRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"} - } - }, - "DeleteSnapshotRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"} - } - }, - "DeleteSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DeleteSubnetRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetId":{"shape":"String"} - } - }, - "DeleteTagsRequest":{ - "type":"structure", - "required":["Resources"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Resources":{ - "shape":"ResourceIdList", - "locationName":"resourceId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tag" - } - } - }, - "DeleteVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"} - } - }, - "DeleteVpcEndpointsRequest":{ - "type":"structure", - "required":["VpcEndpointIds"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - } - } - }, - "DeleteVpcEndpointsResult":{ - "type":"structure", - "members":{ - "Unsuccessful":{ - "shape":"UnsuccessfulItemSet", - "locationName":"unsuccessful" - } - } - }, - "DeleteVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "DeleteVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DeleteVpcRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRequest":{ - "type":"structure", - "required":["VpnConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionId":{"shape":"String"} - } - }, - "DeleteVpnConnectionRouteRequest":{ - "type":"structure", - "required":[ - "VpnConnectionId", - "DestinationCidrBlock" - ], - "members":{ - "VpnConnectionId":{"shape":"String"}, - "DestinationCidrBlock":{"shape":"String"} - } - }, - "DeleteVpnGatewayRequest":{ - "type":"structure", - "required":["VpnGatewayId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"} - } - }, - "DeregisterImageRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"} - } - }, - "DescribeAccountAttributesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttributeNames":{ - "shape":"AccountAttributeNameStringList", - "locationName":"attributeName" - } - } - }, - "DescribeAccountAttributesResult":{ - "type":"structure", - "members":{ - "AccountAttributes":{ - "shape":"AccountAttributeList", - "locationName":"accountAttributeSet" - } - } - }, - "DescribeAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"PublicIpStringList", - "locationName":"PublicIp" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AllocationIds":{ - "shape":"AllocationIdList", - "locationName":"AllocationId" - } - } - }, - "DescribeAddressesResult":{ - "type":"structure", - "members":{ - "Addresses":{ - "shape":"AddressList", - "locationName":"addressesSet" - } - } - }, - "DescribeAvailabilityZonesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ZoneNames":{ - "shape":"ZoneNameStringList", - "locationName":"ZoneName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeAvailabilityZonesResult":{ - "type":"structure", - "members":{ - "AvailabilityZones":{ - "shape":"AvailabilityZoneList", - "locationName":"availabilityZoneInfo" - } - } - }, - "DescribeBundleTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "BundleIds":{ - "shape":"BundleIdStringList", - "locationName":"BundleId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeBundleTasksResult":{ - "type":"structure", - "members":{ - "BundleTasks":{ - "shape":"BundleTaskList", - "locationName":"bundleInstanceTasksSet" - } - } - }, - "DescribeClassicLinkInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeClassicLinkInstancesResult":{ - "type":"structure", - "members":{ - "Instances":{ - "shape":"ClassicLinkInstanceList", - "locationName":"instancesSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeConversionTaskList":{ - "type":"list", - "member":{ - "shape":"ConversionTask", - "locationName":"item" - } - }, - "DescribeConversionTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ConversionTaskIds":{ - "shape":"ConversionIdStringList", - "locationName":"conversionTaskId" - } - } - }, - "DescribeConversionTasksResult":{ - "type":"structure", - "members":{ - "ConversionTasks":{ - "shape":"DescribeConversionTaskList", - "locationName":"conversionTasks" - } - } - }, - "DescribeCustomerGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "CustomerGatewayIds":{ - "shape":"CustomerGatewayIdStringList", - "locationName":"CustomerGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeCustomerGatewaysResult":{ - "type":"structure", - "members":{ - "CustomerGateways":{ - "shape":"CustomerGatewayList", - "locationName":"customerGatewaySet" - } - } - }, - "DescribeDhcpOptionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "DhcpOptionsIds":{ - "shape":"DhcpOptionsIdStringList", - "locationName":"DhcpOptionsId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeDhcpOptionsResult":{ - "type":"structure", - "members":{ - "DhcpOptions":{ - "shape":"DhcpOptionsList", - "locationName":"dhcpOptionsSet" - } - } - }, - "DescribeExportTasksRequest":{ - "type":"structure", - "members":{ - "ExportTaskIds":{ - "shape":"ExportTaskIdStringList", - "locationName":"exportTaskId" - } - } - }, - "DescribeExportTasksResult":{ - "type":"structure", - "members":{ - "ExportTasks":{ - "shape":"ExportTaskList", - "locationName":"exportTaskSet" - } - } - }, - "DescribeFlowLogsRequest":{ - "type":"structure", - "members":{ - "FlowLogIds":{ - "shape":"ValueStringList", - "locationName":"FlowLogId" - }, - "Filter":{"shape":"FilterList"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeFlowLogsResult":{ - "type":"structure", - "members":{ - "FlowLogs":{ - "shape":"FlowLogSet", - "locationName":"flowLogSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostReservationOfferingsRequest":{ - "type":"structure", - "members":{ - "OfferingId":{"shape":"String"}, - "MinDuration":{"shape":"Integer"}, - "MaxDuration":{"shape":"Integer"}, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeHostReservationOfferingsResult":{ - "type":"structure", - "members":{ - "OfferingSet":{ - "shape":"HostOfferingSet", - "locationName":"offeringSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostReservationsRequest":{ - "type":"structure", - "members":{ - "HostReservationIdSet":{"shape":"HostReservationIdSet"}, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeHostReservationsResult":{ - "type":"structure", - "members":{ - "HostReservationSet":{ - "shape":"HostReservationSet", - "locationName":"hostReservationSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeHostsRequest":{ - "type":"structure", - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "Filter":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeHostsResult":{ - "type":"structure", - "members":{ - "Hosts":{ - "shape":"HostList", - "locationName":"hostSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeIdFormatRequest":{ - "type":"structure", - "members":{ - "Resource":{"shape":"String"} - } - }, - "DescribeIdFormatResult":{ - "type":"structure", - "members":{ - "Statuses":{ - "shape":"IdFormatList", - "locationName":"statusSet" - } - } - }, - "DescribeIdentityIdFormatRequest":{ - "type":"structure", - "required":["PrincipalArn"], - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "PrincipalArn":{ - "shape":"String", - "locationName":"principalArn" - } - } - }, - "DescribeIdentityIdFormatResult":{ - "type":"structure", - "members":{ - "Statuses":{ - "shape":"IdFormatList", - "locationName":"statusSet" - } - } - }, - "DescribeImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ImageAttributeName"} - } - }, - "DescribeImagesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageIds":{ - "shape":"ImageIdStringList", - "locationName":"ImageId" - }, - "Owners":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "ExecutableUsers":{ - "shape":"ExecutableByStringList", - "locationName":"ExecutableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeImagesResult":{ - "type":"structure", - "members":{ - "Images":{ - "shape":"ImageList", - "locationName":"imagesSet" - } - } - }, - "DescribeImportImageTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportImageTasksResult":{ - "type":"structure", - "members":{ - "ImportImageTasks":{ - "shape":"ImportImageTaskList", - "locationName":"importImageTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeImportSnapshotTasksRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ImportTaskIds":{ - "shape":"ImportTaskIdList", - "locationName":"ImportTaskId" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeImportSnapshotTasksResult":{ - "type":"structure", - "members":{ - "ImportSnapshotTasks":{ - "shape":"ImportSnapshotTaskList", - "locationName":"importSnapshotTaskSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "DescribeInstanceStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "IncludeAllInstances":{ - "shape":"Boolean", - "locationName":"includeAllInstances" - } - } - }, - "DescribeInstanceStatusResult":{ - "type":"structure", - "members":{ - "InstanceStatuses":{ - "shape":"InstanceStatusList", - "locationName":"instanceStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeInstancesResult":{ - "type":"structure", - "members":{ - "Reservations":{ - "shape":"ReservationList", - "locationName":"reservationSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeInternetGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayIds":{ - "shape":"ValueStringList", - "locationName":"internetGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeInternetGatewaysResult":{ - "type":"structure", - "members":{ - "InternetGateways":{ - "shape":"InternetGatewayList", - "locationName":"internetGatewaySet" - } - } - }, - "DescribeKeyPairsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyNames":{ - "shape":"KeyNameStringList", - "locationName":"KeyName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeKeyPairsResult":{ - "type":"structure", - "members":{ - "KeyPairs":{ - "shape":"KeyPairList", - "locationName":"keySet" - } - } - }, - "DescribeMovingAddressesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIps":{ - "shape":"ValueStringList", - "locationName":"publicIp" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeMovingAddressesResult":{ - "type":"structure", - "members":{ - "MovingAddressStatuses":{ - "shape":"MovingAddressStatusSet", - "locationName":"movingAddressStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNatGatewaysRequest":{ - "type":"structure", - "members":{ - "NatGatewayIds":{ - "shape":"ValueStringList", - "locationName":"NatGatewayId" - }, - "Filter":{"shape":"FilterList"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeNatGatewaysResult":{ - "type":"structure", - "members":{ - "NatGateways":{ - "shape":"NatGatewayList", - "locationName":"natGatewaySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeNetworkAclsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclIds":{ - "shape":"ValueStringList", - "locationName":"NetworkAclId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeNetworkAclsResult":{ - "type":"structure", - "members":{ - "NetworkAcls":{ - "shape":"NetworkAclList", - "locationName":"networkAclSet" - } - } - }, - "DescribeNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Attribute":{ - "shape":"NetworkInterfaceAttribute", - "locationName":"attribute" - } - } - }, - "DescribeNetworkInterfaceAttributeResult":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - } - } - }, - "DescribeNetworkInterfacesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceIds":{ - "shape":"NetworkInterfaceIdList", - "locationName":"NetworkInterfaceId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"filter" - } - } - }, - "DescribeNetworkInterfacesResult":{ - "type":"structure", - "members":{ - "NetworkInterfaces":{ - "shape":"NetworkInterfaceList", - "locationName":"networkInterfaceSet" - } - } - }, - "DescribePlacementGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"PlacementGroupStringList", - "locationName":"groupName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribePlacementGroupsResult":{ - "type":"structure", - "members":{ - "PlacementGroups":{ - "shape":"PlacementGroupList", - "locationName":"placementGroupSet" - } - } - }, - "DescribePrefixListsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "PrefixListIds":{ - "shape":"ValueStringList", - "locationName":"PrefixListId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribePrefixListsResult":{ - "type":"structure", - "members":{ - "PrefixLists":{ - "shape":"PrefixListSet", - "locationName":"prefixListSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeRegionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RegionNames":{ - "shape":"RegionNameStringList", - "locationName":"RegionName" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRegionsResult":{ - "type":"structure", - "members":{ - "Regions":{ - "shape":"RegionList", - "locationName":"regionInfo" - } - } - }, - "DescribeReservedInstancesListingsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeReservedInstancesListingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesListings":{ - "shape":"ReservedInstancesListingList", - "locationName":"reservedInstancesListingsSet" - } - } - }, - "DescribeReservedInstancesModificationsRequest":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationIds":{ - "shape":"ReservedInstancesModificationIdStringList", - "locationName":"ReservedInstancesModificationId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeReservedInstancesModificationsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModifications":{ - "shape":"ReservedInstancesModificationList", - "locationName":"reservedInstancesModificationsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesOfferingsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingIds":{ - "shape":"ReservedInstancesOfferingIdStringList", - "locationName":"ReservedInstancesOfferingId" - }, - "InstanceType":{"shape":"InstanceType"}, - "AvailabilityZone":{"shape":"String"}, - "ProductDescription":{"shape":"RIProductDescription"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "IncludeMarketplace":{"shape":"Boolean"}, - "MinDuration":{"shape":"Long"}, - "MaxDuration":{"shape":"Long"}, - "MaxInstanceCount":{"shape":"Integer"}, - "OfferingClass":{"shape":"OfferingClassType"} - } - }, - "DescribeReservedInstancesOfferingsResult":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferings":{ - "shape":"ReservedInstancesOfferingList", - "locationName":"reservedInstancesOfferingsSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeReservedInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "OfferingClass":{"shape":"OfferingClassType"} - } - }, - "DescribeReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstances":{ - "shape":"ReservedInstancesList", - "locationName":"reservedInstancesSet" - } - } - }, - "DescribeRouteTablesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RouteTableId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeRouteTablesResult":{ - "type":"structure", - "members":{ - "RouteTables":{ - "shape":"RouteTableList", - "locationName":"routeTableSet" - } - } - }, - "DescribeScheduledInstanceAvailabilityRequest":{ - "type":"structure", - "required":[ - "Recurrence", - "FirstSlotStartTimeRange" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "Recurrence":{"shape":"ScheduledInstanceRecurrenceRequest"}, - "FirstSlotStartTimeRange":{"shape":"SlotDateTimeRangeRequest"}, - "MinSlotDurationInHours":{"shape":"Integer"}, - "MaxSlotDurationInHours":{"shape":"Integer"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeScheduledInstanceAvailabilityResult":{ - "type":"structure", - "members":{ - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "ScheduledInstanceAvailabilitySet":{ - "shape":"ScheduledInstanceAvailabilitySet", - "locationName":"scheduledInstanceAvailabilitySet" - } - } - }, - "DescribeScheduledInstancesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "ScheduledInstanceIds":{ - "shape":"ScheduledInstanceIdRequestSet", - "locationName":"ScheduledInstanceId" - }, - "SlotStartTimeRange":{"shape":"SlotStartTimeRangeRequest"}, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"}, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeScheduledInstancesResult":{ - "type":"structure", - "members":{ - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "ScheduledInstanceSet":{ - "shape":"ScheduledInstanceSet", - "locationName":"scheduledInstanceSet" - } - } - }, - "DescribeSecurityGroupReferencesRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "GroupId":{"shape":"GroupIds"} - } - }, - "DescribeSecurityGroupReferencesResult":{ - "type":"structure", - "members":{ - "SecurityGroupReferenceSet":{ - "shape":"SecurityGroupReferences", - "locationName":"securityGroupReferenceSet" - } - } - }, - "DescribeSecurityGroupsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSecurityGroupsResult":{ - "type":"structure", - "members":{ - "SecurityGroups":{ - "shape":"SecurityGroupList", - "locationName":"securityGroupInfo" - } - } - }, - "DescribeSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "DescribeSnapshotAttributeResult":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "CreateVolumePermissions":{ - "shape":"CreateVolumePermissionList", - "locationName":"createVolumePermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeSnapshotsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotIds":{ - "shape":"SnapshotIdStringList", - "locationName":"SnapshotId" - }, - "OwnerIds":{ - "shape":"OwnerStringList", - "locationName":"Owner" - }, - "RestorableByUserIds":{ - "shape":"RestorableByStringList", - "locationName":"RestorableBy" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeSnapshotsResult":{ - "type":"structure", - "members":{ - "Snapshots":{ - "shape":"SnapshotList", - "locationName":"snapshotSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotDatafeedSubscriptionRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "DescribeSpotDatafeedSubscriptionResult":{ - "type":"structure", - "members":{ - "SpotDatafeedSubscription":{ - "shape":"SpotDatafeedSubscription", - "locationName":"spotDatafeedSubscription" - } - } - }, - "DescribeSpotFleetInstancesRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetInstancesResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "ActiveInstances" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "ActiveInstances":{ - "shape":"ActiveInstanceSet", - "locationName":"activeInstanceSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestHistoryRequest":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestHistoryResponse":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "StartTime", - "LastEvaluatedTime", - "HistoryRecords" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "LastEvaluatedTime":{ - "shape":"DateTime", - "locationName":"lastEvaluatedTime" - }, - "HistoryRecords":{ - "shape":"HistoryRecords", - "locationName":"historyRecordSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotFleetRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestIds":{ - "shape":"ValueStringList", - "locationName":"spotFleetRequestId" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeSpotFleetRequestsResponse":{ - "type":"structure", - "required":["SpotFleetRequestConfigs"], - "members":{ - "SpotFleetRequestConfigs":{ - "shape":"SpotFleetRequestConfigSet", - "locationName":"spotFleetRequestConfigSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotInstanceRequestsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotInstanceRequestIds":{ - "shape":"SpotInstanceRequestIdList", - "locationName":"SpotInstanceRequestId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSpotInstanceRequestsResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "DescribeSpotPriceHistoryRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "InstanceTypes":{ - "shape":"InstanceTypeList", - "locationName":"InstanceType" - }, - "ProductDescriptions":{ - "shape":"ProductDescriptionList", - "locationName":"ProductDescription" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSpotPriceHistoryResult":{ - "type":"structure", - "members":{ - "SpotPriceHistory":{ - "shape":"SpotPriceHistoryList", - "locationName":"spotPriceHistorySet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeStaleSecurityGroupsRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcId":{"shape":"String"}, - "MaxResults":{"shape":"MaxResults"}, - "NextToken":{"shape":"NextToken"} - } - }, - "DescribeStaleSecurityGroupsResult":{ - "type":"structure", - "members":{ - "StaleSecurityGroupSet":{ - "shape":"StaleSecurityGroupSet", - "locationName":"staleSecurityGroupSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeSubnetsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SubnetIds":{ - "shape":"SubnetIdStringList", - "locationName":"SubnetId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeSubnetsResult":{ - "type":"structure", - "members":{ - "Subnets":{ - "shape":"SubnetList", - "locationName":"subnetSet" - } - } - }, - "DescribeTagsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeTagsResult":{ - "type":"structure", - "members":{ - "Tags":{ - "shape":"TagDescriptionList", - "locationName":"tagSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "Attribute":{"shape":"VolumeAttributeName"} - } - }, - "DescribeVolumeAttributeResult":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AutoEnableIO":{ - "shape":"AttributeBooleanValue", - "locationName":"autoEnableIO" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - } - } - }, - "DescribeVolumeStatusRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{"shape":"String"}, - "MaxResults":{"shape":"Integer"} - } - }, - "DescribeVolumeStatusResult":{ - "type":"structure", - "members":{ - "VolumeStatuses":{ - "shape":"VolumeStatusList", - "locationName":"volumeStatusSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVolumesRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeIds":{ - "shape":"VolumeIdStringList", - "locationName":"VolumeId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - }, - "MaxResults":{ - "shape":"Integer", - "locationName":"maxResults" - } - } - }, - "DescribeVolumesResult":{ - "type":"structure", - "members":{ - "Volumes":{ - "shape":"VolumeList", - "locationName":"volumeSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcAttributeRequest":{ - "type":"structure", - "required":[ - "VpcId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{"shape":"String"}, - "Attribute":{"shape":"VpcAttributeName"} - } - }, - "DescribeVpcAttributeResult":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsSupport" - }, - "EnableDnsHostnames":{ - "shape":"AttributeBooleanValue", - "locationName":"enableDnsHostnames" - } - } - }, - "DescribeVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcIds":{"shape":"VpcClassicLinkIdList"}, - "MaxResults":{ - "shape":"MaxResults", - "locationName":"maxResults" - }, - "NextToken":{ - "shape":"NextToken", - "locationName":"nextToken" - } - } - }, - "DescribeVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"ClassicLinkDnsSupportList", - "locationName":"vpcs" - }, - "NextToken":{ - "shape":"NextToken", - "locationName":"nextToken" - } - } - }, - "DescribeVpcClassicLinkRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcClassicLinkIdList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcClassicLinkList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpcEndpointServicesRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointServicesResult":{ - "type":"structure", - "members":{ - "ServiceNames":{ - "shape":"ValueStringList", - "locationName":"serviceNameSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcEndpointsRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointIds":{ - "shape":"ValueStringList", - "locationName":"VpcEndpointId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - }, - "MaxResults":{"shape":"Integer"}, - "NextToken":{"shape":"String"} - } - }, - "DescribeVpcEndpointsResult":{ - "type":"structure", - "members":{ - "VpcEndpoints":{ - "shape":"VpcEndpointSet", - "locationName":"vpcEndpointSet" - }, - "NextToken":{ - "shape":"String", - "locationName":"nextToken" - } - } - }, - "DescribeVpcPeeringConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionIds":{ - "shape":"ValueStringList", - "locationName":"VpcPeeringConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcPeeringConnectionsResult":{ - "type":"structure", - "members":{ - "VpcPeeringConnections":{ - "shape":"VpcPeeringConnectionList", - "locationName":"vpcPeeringConnectionSet" - } - } - }, - "DescribeVpcsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcIds":{ - "shape":"VpcIdStringList", - "locationName":"VpcId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpcsResult":{ - "type":"structure", - "members":{ - "Vpcs":{ - "shape":"VpcList", - "locationName":"vpcSet" - } - } - }, - "DescribeVpnConnectionsRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnConnectionIds":{ - "shape":"VpnConnectionIdStringList", - "locationName":"VpnConnectionId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnConnectionsResult":{ - "type":"structure", - "members":{ - "VpnConnections":{ - "shape":"VpnConnectionList", - "locationName":"vpnConnectionSet" - } - } - }, - "DescribeVpnGatewaysRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayIds":{ - "shape":"VpnGatewayIdStringList", - "locationName":"VpnGatewayId" - }, - "Filters":{ - "shape":"FilterList", - "locationName":"Filter" - } - } - }, - "DescribeVpnGatewaysResult":{ - "type":"structure", - "members":{ - "VpnGateways":{ - "shape":"VpnGatewayList", - "locationName":"vpnGatewaySet" - } - } - }, - "DetachClassicLinkVpcRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachClassicLinkVpcResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DetachInternetGatewayRequest":{ - "type":"structure", - "required":[ - "InternetGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DetachNetworkInterfaceRequest":{ - "type":"structure", - "required":["AttachmentId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "DetachVolumeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "InstanceId":{"shape":"String"}, - "Device":{"shape":"String"}, - "Force":{"shape":"Boolean"} - } - }, - "DetachVpnGatewayRequest":{ - "type":"structure", - "required":[ - "VpnGatewayId", - "VpcId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpnGatewayId":{"shape":"String"}, - "VpcId":{"shape":"String"} - } - }, - "DeviceType":{ - "type":"string", - "enum":[ - "ebs", - "instance-store" - ] - }, - "DhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"DhcpConfigurationValueList", - "locationName":"valueSet" - } - } - }, - "DhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"DhcpConfiguration", - "locationName":"item" - } - }, - "DhcpConfigurationValueList":{ - "type":"list", - "member":{ - "shape":"AttributeValue", - "locationName":"item" - } - }, - "DhcpOptions":{ - "type":"structure", - "members":{ - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "DhcpConfigurations":{ - "shape":"DhcpConfigurationList", - "locationName":"dhcpConfigurationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "DhcpOptionsIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DhcpOptionsId" - } - }, - "DhcpOptionsList":{ - "type":"list", - "member":{ - "shape":"DhcpOptions", - "locationName":"item" - } - }, - "DisableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcId":{"shape":"String"} - } - }, - "DisableVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "DisableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "DisassociateAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AssociationId":{"shape":"String"} - } - }, - "DisassociateRouteTableRequest":{ - "type":"structure", - "required":["AssociationId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "DiskImage":{ - "type":"structure", - "members":{ - "Image":{"shape":"DiskImageDetail"}, - "Description":{"shape":"String"}, - "Volume":{"shape":"VolumeDetail"} - } - }, - "DiskImageDescription":{ - "type":"structure", - "required":[ - "Format", - "Size", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - }, - "Checksum":{ - "shape":"String", - "locationName":"checksum" - } - } - }, - "DiskImageDetail":{ - "type":"structure", - "required":[ - "Format", - "Bytes", - "ImportManifestUrl" - ], - "members":{ - "Format":{ - "shape":"DiskImageFormat", - "locationName":"format" - }, - "Bytes":{ - "shape":"Long", - "locationName":"bytes" - }, - "ImportManifestUrl":{ - "shape":"String", - "locationName":"importManifestUrl" - } - } - }, - "DiskImageFormat":{ - "type":"string", - "enum":[ - "VMDK", - "RAW", - "VHD" - ] - }, - "DiskImageList":{ - "type":"list", - "member":{"shape":"DiskImage"} - }, - "DiskImageVolumeDescription":{ - "type":"structure", - "required":["Id"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "DomainType":{ - "type":"string", - "enum":[ - "vpc", - "standard" - ] - }, - "Double":{"type":"double"}, - "EbsBlockDevice":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - } - } - }, - "EbsInstanceBlockDevice":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EbsInstanceBlockDeviceSpecification":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "EnableVgwRoutePropagationRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "GatewayId" - ], - "members":{ - "RouteTableId":{"shape":"String"}, - "GatewayId":{"shape":"String"} - } - }, - "EnableVolumeIORequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - } - } - }, - "EnableVpcClassicLinkDnsSupportRequest":{ - "type":"structure", - "members":{ - "VpcId":{"shape":"String"} - } - }, - "EnableVpcClassicLinkDnsSupportResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EnableVpcClassicLinkRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - } - } - }, - "EnableVpcClassicLinkResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "EventCode":{ - "type":"string", - "enum":[ - "instance-reboot", - "system-reboot", - "system-maintenance", - "instance-retirement", - "instance-stop" - ] - }, - "EventInformation":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "EventSubType":{ - "shape":"String", - "locationName":"eventSubType" - }, - "EventDescription":{ - "shape":"String", - "locationName":"eventDescription" - } - } - }, - "EventType":{ - "type":"string", - "enum":[ - "instanceChange", - "fleetRequestChange", - "error" - ] - }, - "ExcessCapacityTerminationPolicy":{ - "type":"string", - "enum":[ - "noTermination", - "default" - ] - }, - "ExecutableByStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExecutableBy" - } - }, - "ExportEnvironment":{ - "type":"string", - "enum":[ - "citrix", - "vmware", - "microsoft" - ] - }, - "ExportTask":{ - "type":"structure", - "members":{ - "ExportTaskId":{ - "shape":"String", - "locationName":"exportTaskId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "State":{ - "shape":"ExportTaskState", - "locationName":"state" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceExportDetails":{ - "shape":"InstanceExportDetails", - "locationName":"instanceExport" - }, - "ExportToS3Task":{ - "shape":"ExportToS3Task", - "locationName":"exportToS3" - } - } - }, - "ExportTaskIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ExportTaskId" - } - }, - "ExportTaskList":{ - "type":"list", - "member":{ - "shape":"ExportTask", - "locationName":"item" - } - }, - "ExportTaskState":{ - "type":"string", - "enum":[ - "active", - "cancelling", - "cancelled", - "completed" - ] - }, - "ExportToS3Task":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "ExportToS3TaskSpecification":{ - "type":"structure", - "members":{ - "DiskImageFormat":{ - "shape":"DiskImageFormat", - "locationName":"diskImageFormat" - }, - "ContainerFormat":{ - "shape":"ContainerFormat", - "locationName":"containerFormat" - }, - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Prefix":{ - "shape":"String", - "locationName":"s3Prefix" - } - } - }, - "Filter":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "FilterList":{ - "type":"list", - "member":{ - "shape":"Filter", - "locationName":"Filter" - } - }, - "FleetType":{ - "type":"string", - "enum":[ - "request", - "maintain" - ] - }, - "Float":{"type":"float"}, - "FlowLog":{ - "type":"structure", - "members":{ - "CreationTime":{ - "shape":"DateTime", - "locationName":"creationTime" - }, - "FlowLogId":{ - "shape":"String", - "locationName":"flowLogId" - }, - "FlowLogStatus":{ - "shape":"String", - "locationName":"flowLogStatus" - }, - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "TrafficType":{ - "shape":"TrafficType", - "locationName":"trafficType" - }, - "LogGroupName":{ - "shape":"String", - "locationName":"logGroupName" - }, - "DeliverLogsStatus":{ - "shape":"String", - "locationName":"deliverLogsStatus" - }, - "DeliverLogsErrorMessage":{ - "shape":"String", - "locationName":"deliverLogsErrorMessage" - }, - "DeliverLogsPermissionArn":{ - "shape":"String", - "locationName":"deliverLogsPermissionArn" - } - } - }, - "FlowLogSet":{ - "type":"list", - "member":{ - "shape":"FlowLog", - "locationName":"item" - } - }, - "FlowLogsResourceType":{ - "type":"string", - "enum":[ - "VPC", - "Subnet", - "NetworkInterface" - ] - }, - "GatewayType":{ - "type":"string", - "enum":["ipsec.1"] - }, - "GetConsoleOutputRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetConsoleOutputResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "Output":{ - "shape":"String", - "locationName":"output" - } - } - }, - "GetConsoleScreenshotRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "InstanceId":{"shape":"String"}, - "WakeUp":{"shape":"Boolean"} - } - }, - "GetConsoleScreenshotResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageData":{ - "shape":"String", - "locationName":"imageData" - } - } - }, - "GetHostReservationPurchasePreviewRequest":{ - "type":"structure", - "required":[ - "OfferingId", - "HostIdSet" - ], - "members":{ - "OfferingId":{"shape":"String"}, - "HostIdSet":{"shape":"RequestHostIdSet"} - } - }, - "GetHostReservationPurchasePreviewResult":{ - "type":"structure", - "members":{ - "Purchase":{ - "shape":"PurchaseSet", - "locationName":"purchase" - }, - "TotalUpfrontPrice":{ - "shape":"String", - "locationName":"totalUpfrontPrice" - }, - "TotalHourlyPrice":{ - "shape":"String", - "locationName":"totalHourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "GetPasswordDataRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{"shape":"String"} - } - }, - "GetPasswordDataResult":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "PasswordData":{ - "shape":"String", - "locationName":"passwordData" - } - } - }, - "GetReservedInstancesExchangeQuoteRequest":{ - "type":"structure", - "required":["ReservedInstanceIds"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ReservedInstanceIds":{ - "shape":"ReservedInstanceIdSet", - "locationName":"ReservedInstanceId" - }, - "TargetConfigurations":{ - "shape":"TargetConfigurationRequestSet", - "locationName":"TargetConfiguration" - } - } - }, - "GetReservedInstancesExchangeQuoteResult":{ - "type":"structure", - "members":{ - "ReservedInstanceValueSet":{ - "shape":"ReservedInstanceReservationValueSet", - "locationName":"reservedInstanceValueSet" - }, - "ReservedInstanceValueRollup":{ - "shape":"ReservationValue", - "locationName":"reservedInstanceValueRollup" - }, - "TargetConfigurationValueSet":{ - "shape":"TargetReservationValueSet", - "locationName":"targetConfigurationValueSet" - }, - "TargetConfigurationValueRollup":{ - "shape":"ReservationValue", - "locationName":"targetConfigurationValueRollup" - }, - "PaymentDue":{ - "shape":"String", - "locationName":"paymentDue" - }, - "CurrencyCode":{ - "shape":"String", - "locationName":"currencyCode" - }, - "OutputReservedInstancesWillExpireAt":{ - "shape":"DateTime", - "locationName":"outputReservedInstancesWillExpireAt" - }, - "IsValidExchange":{ - "shape":"Boolean", - "locationName":"isValidExchange" - }, - "ValidationFailureReason":{ - "shape":"String", - "locationName":"validationFailureReason" - } - } - }, - "GroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"groupId" - } - }, - "GroupIdentifier":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - } - } - }, - "GroupIdentifierList":{ - "type":"list", - "member":{ - "shape":"GroupIdentifier", - "locationName":"item" - } - }, - "GroupIds":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "GroupNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"GroupName" - } - }, - "HistoryRecord":{ - "type":"structure", - "required":[ - "Timestamp", - "EventType", - "EventInformation" - ], - "members":{ - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "EventType":{ - "shape":"EventType", - "locationName":"eventType" - }, - "EventInformation":{ - "shape":"EventInformation", - "locationName":"eventInformation" - } - } - }, - "HistoryRecords":{ - "type":"list", - "member":{ - "shape":"HistoryRecord", - "locationName":"item" - } - }, - "Host":{ - "type":"structure", - "members":{ - "HostId":{ - "shape":"String", - "locationName":"hostId" - }, - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - }, - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "HostProperties":{ - "shape":"HostProperties", - "locationName":"hostProperties" - }, - "State":{ - "shape":"AllocationState", - "locationName":"state" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Instances":{ - "shape":"HostInstanceList", - "locationName":"instances" - }, - "AvailableCapacity":{ - "shape":"AvailableCapacity", - "locationName":"availableCapacity" - } - } - }, - "HostInstance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - } - } - }, - "HostInstanceList":{ - "type":"list", - "member":{ - "shape":"HostInstance", - "locationName":"item" - } - }, - "HostList":{ - "type":"list", - "member":{ - "shape":"Host", - "locationName":"item" - } - }, - "HostOffering":{ - "type":"structure", - "members":{ - "OfferingId":{ - "shape":"String", - "locationName":"offeringId" - }, - "InstanceFamily":{ - "shape":"String", - "locationName":"instanceFamily" - }, - "PaymentOption":{ - "shape":"PaymentOption", - "locationName":"paymentOption" - }, - "UpfrontPrice":{ - "shape":"String", - "locationName":"upfrontPrice" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Duration":{ - "shape":"Integer", - "locationName":"duration" - } - } - }, - "HostOfferingSet":{ - "type":"list", - "member":{"shape":"HostOffering"} - }, - "HostProperties":{ - "type":"structure", - "members":{ - "Sockets":{ - "shape":"Integer", - "locationName":"sockets" - }, - "Cores":{ - "shape":"Integer", - "locationName":"cores" - }, - "TotalVCpus":{ - "shape":"Integer", - "locationName":"totalVCpus" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - } - } - }, - "HostReservation":{ - "type":"structure", - "members":{ - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "HostIdSet":{ - "shape":"ResponseHostIdSet", - "locationName":"hostIdSet" - }, - "OfferingId":{ - "shape":"String", - "locationName":"offeringId" - }, - "InstanceFamily":{ - "shape":"String", - "locationName":"instanceFamily" - }, - "PaymentOption":{ - "shape":"PaymentOption", - "locationName":"paymentOption" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "UpfrontPrice":{ - "shape":"String", - "locationName":"upfrontPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Count":{ - "shape":"Integer", - "locationName":"count" - }, - "Duration":{ - "shape":"Integer", - "locationName":"duration" - }, - "End":{ - "shape":"DateTime", - "locationName":"end" - }, - "Start":{ - "shape":"DateTime", - "locationName":"start" - }, - "State":{ - "shape":"ReservationState", - "locationName":"state" - } - } - }, - "HostReservationIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "HostReservationSet":{ - "type":"list", - "member":{"shape":"HostReservation"} - }, - "HostTenancy":{ - "type":"string", - "enum":[ - "dedicated", - "host" - ] - }, - "HypervisorType":{ - "type":"string", - "enum":[ - "ovm", - "xen" - ] - }, - "IamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Id":{ - "shape":"String", - "locationName":"id" - } - } - }, - "IamInstanceProfileSpecification":{ - "type":"structure", - "members":{ - "Arn":{ - "shape":"String", - "locationName":"arn" - }, - "Name":{ - "shape":"String", - "locationName":"name" - } - } - }, - "IcmpTypeCode":{ - "type":"structure", - "members":{ - "Type":{ - "shape":"Integer", - "locationName":"type" - }, - "Code":{ - "shape":"Integer", - "locationName":"code" - } - } - }, - "IdFormat":{ - "type":"structure", - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "UseLongIds":{ - "shape":"Boolean", - "locationName":"useLongIds" - }, - "Deadline":{ - "shape":"DateTime", - "locationName":"deadline" - } - } - }, - "IdFormatList":{ - "type":"list", - "member":{ - "shape":"IdFormat", - "locationName":"item" - } - }, - "Image":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "ImageLocation":{ - "shape":"String", - "locationName":"imageLocation" - }, - "State":{ - "shape":"ImageState", - "locationName":"imageState" - }, - "OwnerId":{ - "shape":"String", - "locationName":"imageOwnerId" - }, - "CreationDate":{ - "shape":"String", - "locationName":"creationDate" - }, - "Public":{ - "shape":"Boolean", - "locationName":"isPublic" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "ImageType":{ - "shape":"ImageTypeValues", - "locationName":"imageType" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"Boolean", - "locationName":"enaSupport" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "ImageOwnerAlias":{ - "shape":"String", - "locationName":"imageOwnerAlias" - }, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - } - } - }, - "ImageAttribute":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "LaunchPermissions":{ - "shape":"LaunchPermissionList", - "locationName":"launchPermission" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - } - } - }, - "ImageAttributeName":{ - "type":"string", - "enum":[ - "description", - "kernel", - "ramdisk", - "launchPermission", - "productCodes", - "blockDeviceMapping", - "sriovNetSupport" - ] - }, - "ImageDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"}, - "DeviceName":{"shape":"String"}, - "SnapshotId":{"shape":"String"} - } - }, - "ImageDiskContainerList":{ - "type":"list", - "member":{ - "shape":"ImageDiskContainer", - "locationName":"item" - } - }, - "ImageIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImageId" - } - }, - "ImageList":{ - "type":"list", - "member":{ - "shape":"Image", - "locationName":"item" - } - }, - "ImageState":{ - "type":"string", - "enum":[ - "pending", - "available", - "invalid", - "deregistered", - "transient", - "failed", - "error" - ] - }, - "ImageTypeValues":{ - "type":"string", - "enum":[ - "machine", - "kernel", - "ramdisk" - ] - }, - "ImportImageRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainers":{ - "shape":"ImageDiskContainerList", - "locationName":"DiskContainer" - }, - "LicenseType":{"shape":"String"}, - "Hypervisor":{"shape":"String"}, - "Architecture":{"shape":"String"}, - "Platform":{"shape":"String"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportImageResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "Architecture":{ - "shape":"String", - "locationName":"architecture" - }, - "LicenseType":{ - "shape":"String", - "locationName":"licenseType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "Hypervisor":{ - "shape":"String", - "locationName":"hypervisor" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "SnapshotDetails":{ - "shape":"SnapshotDetailList", - "locationName":"snapshotDetailSet" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "ImportImageTaskList":{ - "type":"list", - "member":{ - "shape":"ImportImageTask", - "locationName":"item" - } - }, - "ImportInstanceLaunchSpecification":{ - "type":"structure", - "members":{ - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "GroupNames":{ - "shape":"SecurityGroupStringList", - "locationName":"GroupName" - }, - "GroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"GroupId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "UserData":{ - "shape":"UserData", - "locationName":"userData" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "Monitoring":{ - "shape":"Boolean", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - } - } - }, - "ImportInstanceRequest":{ - "type":"structure", - "required":["Platform"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "LaunchSpecification":{ - "shape":"ImportInstanceLaunchSpecification", - "locationName":"launchSpecification" - }, - "DiskImages":{ - "shape":"DiskImageList", - "locationName":"diskImage" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - } - } - }, - "ImportInstanceResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportInstanceTaskDetails":{ - "type":"structure", - "required":["Volumes"], - "members":{ - "Volumes":{ - "shape":"ImportInstanceVolumeDetailSet", - "locationName":"volumes" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailItem":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume", - "Status" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportInstanceVolumeDetailSet":{ - "type":"list", - "member":{ - "shape":"ImportInstanceVolumeDetailItem", - "locationName":"item" - } - }, - "ImportKeyPairRequest":{ - "type":"structure", - "required":[ - "KeyName", - "PublicKeyMaterial" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "PublicKeyMaterial":{ - "shape":"Blob", - "locationName":"publicKeyMaterial" - } - } - }, - "ImportKeyPairResult":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "ImportSnapshotRequest":{ - "type":"structure", - "members":{ - "DryRun":{"shape":"Boolean"}, - "Description":{"shape":"String"}, - "DiskContainer":{"shape":"SnapshotDiskContainer"}, - "ClientData":{"shape":"ClientData"}, - "ClientToken":{"shape":"String"}, - "RoleName":{"shape":"String"} - } - }, - "ImportSnapshotResult":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTask":{ - "type":"structure", - "members":{ - "ImportTaskId":{ - "shape":"String", - "locationName":"importTaskId" - }, - "SnapshotTaskDetail":{ - "shape":"SnapshotTaskDetail", - "locationName":"snapshotTaskDetail" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ImportSnapshotTaskList":{ - "type":"list", - "member":{ - "shape":"ImportSnapshotTask", - "locationName":"item" - } - }, - "ImportTaskIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ImportTaskId" - } - }, - "ImportVolumeRequest":{ - "type":"structure", - "required":[ - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Image":{ - "shape":"DiskImageDetail", - "locationName":"image" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Volume":{ - "shape":"VolumeDetail", - "locationName":"volume" - } - } - }, - "ImportVolumeResult":{ - "type":"structure", - "members":{ - "ConversionTask":{ - "shape":"ConversionTask", - "locationName":"conversionTask" - } - } - }, - "ImportVolumeTaskDetails":{ - "type":"structure", - "required":[ - "BytesConverted", - "AvailabilityZone", - "Image", - "Volume" - ], - "members":{ - "BytesConverted":{ - "shape":"Long", - "locationName":"bytesConverted" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Image":{ - "shape":"DiskImageDescription", - "locationName":"image" - }, - "Volume":{ - "shape":"DiskImageVolumeDescription", - "locationName":"volume" - } - } - }, - "Instance":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "State":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"dnsName" - }, - "StateTransitionReason":{ - "shape":"String", - "locationName":"reason" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "AmiLaunchIndex":{ - "shape":"Integer", - "locationName":"amiLaunchIndex" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "LaunchTime":{ - "shape":"DateTime", - "locationName":"launchTime" - }, - "Placement":{ - "shape":"Placement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "Platform":{ - "shape":"PlatformValues", - "locationName":"platform" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PublicIpAddress":{ - "shape":"String", - "locationName":"ipAddress" - }, - "StateReason":{ - "shape":"StateReason", - "locationName":"stateReason" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "RootDeviceType":{ - "shape":"DeviceType", - "locationName":"rootDeviceType" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"VirtualizationType", - "locationName":"virtualizationType" - }, - "InstanceLifecycle":{ - "shape":"InstanceLifecycleType", - "locationName":"instanceLifecycle" - }, - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Hypervisor":{ - "shape":"HypervisorType", - "locationName":"hypervisor" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfile", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"Boolean", - "locationName":"enaSupport" - } - } - }, - "InstanceAttribute":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "KernelId":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "RamdiskId":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"AttributeValue", - "locationName":"userData" - }, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "RootDeviceName":{ - "shape":"AttributeValue", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "ProductCodes":{ - "shape":"ProductCodeList", - "locationName":"productCodes" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enaSupport" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - } - } - }, - "InstanceAttributeName":{ - "type":"string", - "enum":[ - "instanceType", - "kernel", - "ramdisk", - "userData", - "disableApiTermination", - "instanceInitiatedShutdownBehavior", - "rootDeviceName", - "blockDeviceMapping", - "productCodes", - "sourceDestCheck", - "groupSet", - "ebsOptimized", - "sriovNetSupport", - "enaSupport" - ] - }, - "InstanceBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDevice", - "locationName":"ebs" - } - } - }, - "InstanceBlockDeviceMappingList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMapping", - "locationName":"item" - } - }, - "InstanceBlockDeviceMappingSpecification":{ - "type":"structure", - "members":{ - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "Ebs":{ - "shape":"EbsInstanceBlockDeviceSpecification", - "locationName":"ebs" - }, - "VirtualName":{ - "shape":"String", - "locationName":"virtualName" - }, - "NoDevice":{ - "shape":"String", - "locationName":"noDevice" - } - } - }, - "InstanceBlockDeviceMappingSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceBlockDeviceMappingSpecification", - "locationName":"item" - } - }, - "InstanceCapacity":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "AvailableCapacity":{ - "shape":"Integer", - "locationName":"availableCapacity" - }, - "TotalCapacity":{ - "shape":"Integer", - "locationName":"totalCapacity" - } - } - }, - "InstanceCount":{ - "type":"structure", - "members":{ - "State":{ - "shape":"ListingState", - "locationName":"state" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - } - } - }, - "InstanceCountList":{ - "type":"list", - "member":{ - "shape":"InstanceCount", - "locationName":"item" - } - }, - "InstanceExportDetails":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "TargetEnvironment":{ - "shape":"ExportEnvironment", - "locationName":"targetEnvironment" - } - } - }, - "InstanceIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "InstanceIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"InstanceId" - } - }, - "InstanceLifecycleType":{ - "type":"string", - "enum":[ - "spot", - "scheduled" - ] - }, - "InstanceList":{ - "type":"list", - "member":{ - "shape":"Instance", - "locationName":"item" - } - }, - "InstanceMonitoring":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Monitoring":{ - "shape":"Monitoring", - "locationName":"monitoring" - } - } - }, - "InstanceMonitoringList":{ - "type":"list", - "member":{ - "shape":"InstanceMonitoring", - "locationName":"item" - } - }, - "InstanceNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"InstanceNetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - }, - "PrivateIpAddresses":{ - "shape":"InstancePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - } - } - }, - "InstanceNetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - } - } - }, - "InstanceNetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "InstanceNetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterface", - "locationName":"item" - } - }, - "InstanceNetworkInterfaceSpecification":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressSpecificationList", - "locationName":"privateIpAddressesSet", - "queryName":"PrivateIpAddresses" - }, - "SecondaryPrivateIpAddressCount":{ - "shape":"Integer", - "locationName":"secondaryPrivateIpAddressCount" - }, - "AssociatePublicIpAddress":{ - "shape":"Boolean", - "locationName":"associatePublicIpAddress" - } - } - }, - "InstanceNetworkInterfaceSpecificationList":{ - "type":"list", - "member":{ - "shape":"InstanceNetworkInterfaceSpecification", - "locationName":"item" - } - }, - "InstancePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"InstanceNetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "InstancePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"InstancePrivateIpAddress", - "locationName":"item" - } - }, - "InstanceState":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"Integer", - "locationName":"code" - }, - "Name":{ - "shape":"InstanceStateName", - "locationName":"name" - } - } - }, - "InstanceStateChange":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CurrentState":{ - "shape":"InstanceState", - "locationName":"currentState" - }, - "PreviousState":{ - "shape":"InstanceState", - "locationName":"previousState" - } - } - }, - "InstanceStateChangeList":{ - "type":"list", - "member":{ - "shape":"InstanceStateChange", - "locationName":"item" - } - }, - "InstanceStateName":{ - "type":"string", - "enum":[ - "pending", - "running", - "shutting-down", - "terminated", - "stopping", - "stopped" - ] - }, - "InstanceStatus":{ - "type":"structure", - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Events":{ - "shape":"InstanceStatusEventList", - "locationName":"eventsSet" - }, - "InstanceState":{ - "shape":"InstanceState", - "locationName":"instanceState" - }, - "SystemStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"systemStatus" - }, - "InstanceStatus":{ - "shape":"InstanceStatusSummary", - "locationName":"instanceStatus" - } - } - }, - "InstanceStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"StatusName", - "locationName":"name" - }, - "Status":{ - "shape":"StatusType", - "locationName":"status" - }, - "ImpairedSince":{ - "shape":"DateTime", - "locationName":"impairedSince" - } - } - }, - "InstanceStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusDetails", - "locationName":"item" - } - }, - "InstanceStatusEvent":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"EventCode", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - } - } - }, - "InstanceStatusEventList":{ - "type":"list", - "member":{ - "shape":"InstanceStatusEvent", - "locationName":"item" - } - }, - "InstanceStatusList":{ - "type":"list", - "member":{ - "shape":"InstanceStatus", - "locationName":"item" - } - }, - "InstanceStatusSummary":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"SummaryStatus", - "locationName":"status" - }, - "Details":{ - "shape":"InstanceStatusDetailsList", - "locationName":"details" - } - } - }, - "InstanceType":{ - "type":"string", - "enum":[ - "t1.micro", - "t2.nano", - "t2.micro", - "t2.small", - "t2.medium", - "t2.large", - "m1.small", - "m1.medium", - "m1.large", - "m1.xlarge", - "m3.medium", - "m3.large", - "m3.xlarge", - "m3.2xlarge", - "m4.large", - "m4.xlarge", - "m4.2xlarge", - "m4.4xlarge", - "m4.10xlarge", - "m4.16xlarge", - "m2.xlarge", - "m2.2xlarge", - "m2.4xlarge", - "cr1.8xlarge", - "r3.large", - "r3.xlarge", - "r3.2xlarge", - "r3.4xlarge", - "r3.8xlarge", - "x1.16xlarge", - "x1.32xlarge", - "i2.xlarge", - "i2.2xlarge", - "i2.4xlarge", - "i2.8xlarge", - "hi1.4xlarge", - "hs1.8xlarge", - "c1.medium", - "c1.xlarge", - "c3.large", - "c3.xlarge", - "c3.2xlarge", - "c3.4xlarge", - "c3.8xlarge", - "c4.large", - "c4.xlarge", - "c4.2xlarge", - "c4.4xlarge", - "c4.8xlarge", - "cc1.4xlarge", - "cc2.8xlarge", - "g2.2xlarge", - "g2.8xlarge", - "cg1.4xlarge", - "p2.xlarge", - "p2.8xlarge", - "p2.16xlarge", - "d2.xlarge", - "d2.2xlarge", - "d2.4xlarge", - "d2.8xlarge" - ] - }, - "InstanceTypeList":{ - "type":"list", - "member":{"shape":"InstanceType"} - }, - "Integer":{"type":"integer"}, - "InternetGateway":{ - "type":"structure", - "members":{ - "InternetGatewayId":{ - "shape":"String", - "locationName":"internetGatewayId" - }, - "Attachments":{ - "shape":"InternetGatewayAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "InternetGatewayAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "InternetGatewayAttachmentList":{ - "type":"list", - "member":{ - "shape":"InternetGatewayAttachment", - "locationName":"item" - } - }, - "InternetGatewayList":{ - "type":"list", - "member":{ - "shape":"InternetGateway", - "locationName":"item" - } - }, - "IpPermission":{ - "type":"structure", - "members":{ - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairList", - "locationName":"groups" - }, - "IpRanges":{ - "shape":"IpRangeList", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdList", - "locationName":"prefixListIds" - } - } - }, - "IpPermissionList":{ - "type":"list", - "member":{ - "shape":"IpPermission", - "locationName":"item" - } - }, - "IpRange":{ - "type":"structure", - "members":{ - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - } - } - }, - "IpRangeList":{ - "type":"list", - "member":{ - "shape":"IpRange", - "locationName":"item" - } - }, - "IpRanges":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "KeyNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"KeyName" - } - }, - "KeyPair":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - }, - "KeyMaterial":{ - "shape":"String", - "locationName":"keyMaterial" - } - } - }, - "KeyPairInfo":{ - "type":"structure", - "members":{ - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "KeyFingerprint":{ - "shape":"String", - "locationName":"keyFingerprint" - } - } - }, - "KeyPairList":{ - "type":"list", - "member":{ - "shape":"KeyPairInfo", - "locationName":"item" - } - }, - "LaunchPermission":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "Group":{ - "shape":"PermissionGroup", - "locationName":"group" - } - } - }, - "LaunchPermissionList":{ - "type":"list", - "member":{ - "shape":"LaunchPermission", - "locationName":"item" - } - }, - "LaunchPermissionModifications":{ - "type":"structure", - "members":{ - "Add":{"shape":"LaunchPermissionList"}, - "Remove":{"shape":"LaunchPermissionList"} - } - }, - "LaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - } - } - }, - "LaunchSpecsList":{ - "type":"list", - "member":{ - "shape":"SpotFleetLaunchSpecification", - "locationName":"item" - }, - "min":1 - }, - "ListingState":{ - "type":"string", - "enum":[ - "available", - "sold", - "cancelled", - "pending" - ] - }, - "ListingStatus":{ - "type":"string", - "enum":[ - "active", - "pending", - "cancelled", - "closed" - ] - }, - "Long":{"type":"long"}, - "MaxResults":{ - "type":"integer", - "max":255, - "min":5 - }, - "ModifyHostsRequest":{ - "type":"structure", - "required":[ - "HostIds", - "AutoPlacement" - ], - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - }, - "AutoPlacement":{ - "shape":"AutoPlacement", - "locationName":"autoPlacement" - } - } - }, - "ModifyHostsResult":{ - "type":"structure", - "members":{ - "Successful":{ - "shape":"ResponseHostIdList", - "locationName":"successful" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemList", - "locationName":"unsuccessful" - } - } - }, - "ModifyIdFormatRequest":{ - "type":"structure", - "required":[ - "Resource", - "UseLongIds" - ], - "members":{ - "Resource":{"shape":"String"}, - "UseLongIds":{"shape":"Boolean"} - } - }, - "ModifyIdentityIdFormatRequest":{ - "type":"structure", - "required":[ - "Resource", - "UseLongIds", - "PrincipalArn" - ], - "members":{ - "Resource":{ - "shape":"String", - "locationName":"resource" - }, - "UseLongIds":{ - "shape":"Boolean", - "locationName":"useLongIds" - }, - "PrincipalArn":{ - "shape":"String", - "locationName":"principalArn" - } - } - }, - "ModifyImageAttributeRequest":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"String"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "UserGroups":{ - "shape":"UserGroupStringList", - "locationName":"UserGroup" - }, - "ProductCodes":{ - "shape":"ProductCodeStringList", - "locationName":"ProductCode" - }, - "Value":{"shape":"String"}, - "LaunchPermission":{"shape":"LaunchPermissionModifications"}, - "Description":{"shape":"AttributeValue"} - } - }, - "ModifyInstanceAttributeRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - }, - "Value":{ - "shape":"String", - "locationName":"value" - }, - "BlockDeviceMappings":{ - "shape":"InstanceBlockDeviceMappingSpecificationList", - "locationName":"blockDeviceMapping" - }, - "SourceDestCheck":{"shape":"AttributeBooleanValue"}, - "DisableApiTermination":{ - "shape":"AttributeBooleanValue", - "locationName":"disableApiTermination" - }, - "InstanceType":{ - "shape":"AttributeValue", - "locationName":"instanceType" - }, - "Kernel":{ - "shape":"AttributeValue", - "locationName":"kernel" - }, - "Ramdisk":{ - "shape":"AttributeValue", - "locationName":"ramdisk" - }, - "UserData":{ - "shape":"BlobAttributeValue", - "locationName":"userData" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"AttributeValue", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "Groups":{ - "shape":"GroupIdStringList", - "locationName":"GroupId" - }, - "EbsOptimized":{ - "shape":"AttributeBooleanValue", - "locationName":"ebsOptimized" - }, - "SriovNetSupport":{ - "shape":"AttributeValue", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"AttributeBooleanValue", - "locationName":"enaSupport" - } - } - }, - "ModifyInstancePlacementRequest":{ - "type":"structure", - "required":["InstanceId"], - "members":{ - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Tenancy":{ - "shape":"HostTenancy", - "locationName":"tenancy" - }, - "Affinity":{ - "shape":"Affinity", - "locationName":"affinity" - }, - "HostId":{ - "shape":"String", - "locationName":"hostId" - } - } - }, - "ModifyInstancePlacementResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifyNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "Description":{ - "shape":"AttributeValue", - "locationName":"description" - }, - "SourceDestCheck":{ - "shape":"AttributeBooleanValue", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachmentChanges", - "locationName":"attachment" - } - } - }, - "ModifyReservedInstancesRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesIds", - "TargetConfigurations" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "ReservedInstancesIds":{ - "shape":"ReservedInstancesIdStringList", - "locationName":"ReservedInstancesId" - }, - "TargetConfigurations":{ - "shape":"ReservedInstancesConfigurationList", - "locationName":"ReservedInstancesConfigurationSetItemType" - } - } - }, - "ModifyReservedInstancesResult":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - } - } - }, - "ModifySnapshotAttributeRequest":{ - "type":"structure", - "required":["SnapshotId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"}, - "OperationType":{"shape":"OperationType"}, - "UserIds":{ - "shape":"UserIdStringList", - "locationName":"UserId" - }, - "GroupNames":{ - "shape":"GroupNameStringList", - "locationName":"UserGroup" - }, - "CreateVolumePermission":{"shape":"CreateVolumePermissionModifications"} - } - }, - "ModifySpotFleetRequestRequest":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ExcessCapacityTerminationPolicy":{ - "shape":"ExcessCapacityTerminationPolicy", - "locationName":"excessCapacityTerminationPolicy" - } - } - }, - "ModifySpotFleetRequestResponse":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifySubnetAttributeRequest":{ - "type":"structure", - "required":["SubnetId"], - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "MapPublicIpOnLaunch":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVolumeAttributeRequest":{ - "type":"structure", - "required":["VolumeId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VolumeId":{"shape":"String"}, - "AutoEnableIO":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcAttributeRequest":{ - "type":"structure", - "required":["VpcId"], - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "EnableDnsSupport":{"shape":"AttributeBooleanValue"}, - "EnableDnsHostnames":{"shape":"AttributeBooleanValue"} - } - }, - "ModifyVpcEndpointRequest":{ - "type":"structure", - "required":["VpcEndpointId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcEndpointId":{"shape":"String"}, - "ResetPolicy":{"shape":"Boolean"}, - "PolicyDocument":{"shape":"String"}, - "AddRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"AddRouteTableId" - }, - "RemoveRouteTableIds":{ - "shape":"ValueStringList", - "locationName":"RemoveRouteTableId" - } - } - }, - "ModifyVpcEndpointResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ModifyVpcPeeringConnectionOptionsRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "VpcPeeringConnectionId":{"shape":"String"}, - "RequesterPeeringConnectionOptions":{"shape":"PeeringConnectionOptionsRequest"}, - "AccepterPeeringConnectionOptions":{"shape":"PeeringConnectionOptionsRequest"} - } - }, - "ModifyVpcPeeringConnectionOptionsResult":{ - "type":"structure", - "members":{ - "RequesterPeeringConnectionOptions":{ - "shape":"PeeringConnectionOptions", - "locationName":"requesterPeeringConnectionOptions" - }, - "AccepterPeeringConnectionOptions":{ - "shape":"PeeringConnectionOptions", - "locationName":"accepterPeeringConnectionOptions" - } - } - }, - "MonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "MonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "Monitoring":{ - "type":"structure", - "members":{ - "State":{ - "shape":"MonitoringState", - "locationName":"state" - } - } - }, - "MonitoringState":{ - "type":"string", - "enum":[ - "disabled", - "disabling", - "enabled", - "pending" - ] - }, - "MoveAddressToVpcRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "MoveAddressToVpcResult":{ - "type":"structure", - "members":{ - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "Status":{ - "shape":"Status", - "locationName":"status" - } - } - }, - "MoveStatus":{ - "type":"string", - "enum":[ - "movingToVpc", - "restoringToClassic" - ] - }, - "MovingAddressStatus":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "MoveStatus":{ - "shape":"MoveStatus", - "locationName":"moveStatus" - } - } - }, - "MovingAddressStatusSet":{ - "type":"list", - "member":{ - "shape":"MovingAddressStatus", - "locationName":"item" - } - }, - "NatGateway":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "DeleteTime":{ - "shape":"DateTime", - "locationName":"deleteTime" - }, - "NatGatewayAddresses":{ - "shape":"NatGatewayAddressList", - "locationName":"natGatewayAddressSet" - }, - "State":{ - "shape":"NatGatewayState", - "locationName":"state" - }, - "FailureCode":{ - "shape":"String", - "locationName":"failureCode" - }, - "FailureMessage":{ - "shape":"String", - "locationName":"failureMessage" - }, - "ProvisionedBandwidth":{ - "shape":"ProvisionedBandwidth", - "locationName":"provisionedBandwidth" - } - } - }, - "NatGatewayAddress":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "PrivateIp":{ - "shape":"String", - "locationName":"privateIp" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - } - } - }, - "NatGatewayAddressList":{ - "type":"list", - "member":{ - "shape":"NatGatewayAddress", - "locationName":"item" - } - }, - "NatGatewayList":{ - "type":"list", - "member":{ - "shape":"NatGateway", - "locationName":"item" - } - }, - "NatGatewayState":{ - "type":"string", - "enum":[ - "pending", - "failed", - "available", - "deleting", - "deleted" - ] - }, - "NetworkAcl":{ - "type":"structure", - "members":{ - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"default" - }, - "Entries":{ - "shape":"NetworkAclEntryList", - "locationName":"entrySet" - }, - "Associations":{ - "shape":"NetworkAclAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "NetworkAclAssociation":{ - "type":"structure", - "members":{ - "NetworkAclAssociationId":{ - "shape":"String", - "locationName":"networkAclAssociationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - } - } - }, - "NetworkAclAssociationList":{ - "type":"list", - "member":{ - "shape":"NetworkAclAssociation", - "locationName":"item" - } - }, - "NetworkAclEntry":{ - "type":"structure", - "members":{ - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"icmpTypeCode" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "NetworkAclEntryList":{ - "type":"list", - "member":{ - "shape":"NetworkAclEntry", - "locationName":"item" - } - }, - "NetworkAclList":{ - "type":"list", - "member":{ - "shape":"NetworkAcl", - "locationName":"item" - } - }, - "NetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "RequesterManaged":{ - "shape":"Boolean", - "locationName":"requesterManaged" - }, - "Status":{ - "shape":"NetworkInterfaceStatus", - "locationName":"status" - }, - "MacAddress":{ - "shape":"String", - "locationName":"macAddress" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "SourceDestCheck":{ - "shape":"Boolean", - "locationName":"sourceDestCheck" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Attachment":{ - "shape":"NetworkInterfaceAttachment", - "locationName":"attachment" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - }, - "TagSet":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PrivateIpAddresses":{ - "shape":"NetworkInterfacePrivateIpAddressList", - "locationName":"privateIpAddressesSet" - }, - "InterfaceType":{ - "shape":"NetworkInterfaceType", - "locationName":"interfaceType" - } - } - }, - "NetworkInterfaceAssociation":{ - "type":"structure", - "members":{ - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - }, - "PublicDnsName":{ - "shape":"String", - "locationName":"publicDnsName" - }, - "IpOwnerId":{ - "shape":"String", - "locationName":"ipOwnerId" - }, - "AllocationId":{ - "shape":"String", - "locationName":"allocationId" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - } - } - }, - "NetworkInterfaceAttachment":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "DeviceIndex":{ - "shape":"Integer", - "locationName":"deviceIndex" - }, - "Status":{ - "shape":"AttachmentStatus", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttachmentChanges":{ - "type":"structure", - "members":{ - "AttachmentId":{ - "shape":"String", - "locationName":"attachmentId" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "NetworkInterfaceAttribute":{ - "type":"string", - "enum":[ - "description", - "groupSet", - "sourceDestCheck", - "attachment" - ] - }, - "NetworkInterfaceIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "NetworkInterfaceList":{ - "type":"list", - "member":{ - "shape":"NetworkInterface", - "locationName":"item" - } - }, - "NetworkInterfacePrivateIpAddress":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "PrivateDnsName":{ - "shape":"String", - "locationName":"privateDnsName" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - }, - "Association":{ - "shape":"NetworkInterfaceAssociation", - "locationName":"association" - } - } - }, - "NetworkInterfacePrivateIpAddressList":{ - "type":"list", - "member":{ - "shape":"NetworkInterfacePrivateIpAddress", - "locationName":"item" - } - }, - "NetworkInterfaceStatus":{ - "type":"string", - "enum":[ - "available", - "attaching", - "in-use", - "detaching" - ] - }, - "NetworkInterfaceType":{ - "type":"string", - "enum":[ - "interface", - "natGateway" - ] - }, - "NewDhcpConfiguration":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Values":{ - "shape":"ValueStringList", - "locationName":"Value" - } - } - }, - "NewDhcpConfigurationList":{ - "type":"list", - "member":{ - "shape":"NewDhcpConfiguration", - "locationName":"item" - } - }, - "NextToken":{ - "type":"string", - "max":1024, - "min":1 - }, - "OccurrenceDayRequestSet":{ - "type":"list", - "member":{ - "shape":"Integer", - "locationName":"OccurenceDay" - } - }, - "OccurrenceDaySet":{ - "type":"list", - "member":{ - "shape":"Integer", - "locationName":"item" - } - }, - "OfferingClassType":{ - "type":"string", - "enum":[ - "standard", - "convertible" - ] - }, - "OfferingTypeValues":{ - "type":"string", - "enum":[ - "Heavy Utilization", - "Medium Utilization", - "Light Utilization", - "No Upfront", - "Partial Upfront", - "All Upfront" - ] - }, - "OperationType":{ - "type":"string", - "enum":[ - "add", - "remove" - ] - }, - "OwnerStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"Owner" - } - }, - "PaymentOption":{ - "type":"string", - "enum":[ - "AllUpfront", - "PartialUpfront", - "NoUpfront" - ] - }, - "PeeringConnectionOptions":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalClassicLinkToRemoteVpc" - }, - "AllowEgressFromLocalVpcToRemoteClassicLink":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalVpcToRemoteClassicLink" - }, - "AllowDnsResolutionFromRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowDnsResolutionFromRemoteVpc" - } - } - }, - "PeeringConnectionOptionsRequest":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{"shape":"Boolean"}, - "AllowEgressFromLocalVpcToRemoteClassicLink":{"shape":"Boolean"}, - "AllowDnsResolutionFromRemoteVpc":{"shape":"Boolean"} - } - }, - "PermissionGroup":{ - "type":"string", - "enum":["all"] - }, - "Placement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Tenancy":{ - "shape":"Tenancy", - "locationName":"tenancy" - }, - "HostId":{ - "shape":"String", - "locationName":"hostId" - }, - "Affinity":{ - "shape":"String", - "locationName":"affinity" - } - } - }, - "PlacementGroup":{ - "type":"structure", - "members":{ - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Strategy":{ - "shape":"PlacementStrategy", - "locationName":"strategy" - }, - "State":{ - "shape":"PlacementGroupState", - "locationName":"state" - } - } - }, - "PlacementGroupList":{ - "type":"list", - "member":{ - "shape":"PlacementGroup", - "locationName":"item" - } - }, - "PlacementGroupState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "PlacementGroupStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PlacementStrategy":{ - "type":"string", - "enum":["cluster"] - }, - "PlatformValues":{ - "type":"string", - "enum":["Windows"] - }, - "PortRange":{ - "type":"structure", - "members":{ - "From":{ - "shape":"Integer", - "locationName":"from" - }, - "To":{ - "shape":"Integer", - "locationName":"to" - } - } - }, - "PrefixList":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - }, - "PrefixListName":{ - "shape":"String", - "locationName":"prefixListName" - }, - "Cidrs":{ - "shape":"ValueStringList", - "locationName":"cidrSet" - } - } - }, - "PrefixListId":{ - "type":"structure", - "members":{ - "PrefixListId":{ - "shape":"String", - "locationName":"prefixListId" - } - } - }, - "PrefixListIdList":{ - "type":"list", - "member":{ - "shape":"PrefixListId", - "locationName":"item" - } - }, - "PrefixListIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "PrefixListSet":{ - "type":"list", - "member":{ - "shape":"PrefixList", - "locationName":"item" - } - }, - "PriceSchedule":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Active":{ - "shape":"Boolean", - "locationName":"active" - } - } - }, - "PriceScheduleList":{ - "type":"list", - "member":{ - "shape":"PriceSchedule", - "locationName":"item" - } - }, - "PriceScheduleSpecification":{ - "type":"structure", - "members":{ - "Term":{ - "shape":"Long", - "locationName":"term" - }, - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "PriceScheduleSpecificationList":{ - "type":"list", - "member":{ - "shape":"PriceScheduleSpecification", - "locationName":"item" - } - }, - "PricingDetail":{ - "type":"structure", - "members":{ - "Price":{ - "shape":"Double", - "locationName":"price" - }, - "Count":{ - "shape":"Integer", - "locationName":"count" - } - } - }, - "PricingDetailsList":{ - "type":"list", - "member":{ - "shape":"PricingDetail", - "locationName":"item" - } - }, - "PrivateIpAddressConfigSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesPrivateIpAddressConfig", - "locationName":"PrivateIpAddressConfigSet" - } - }, - "PrivateIpAddressSpecification":{ - "type":"structure", - "required":["PrivateIpAddress"], - "members":{ - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "Primary":{ - "shape":"Boolean", - "locationName":"primary" - } - } - }, - "PrivateIpAddressSpecificationList":{ - "type":"list", - "member":{ - "shape":"PrivateIpAddressSpecification", - "locationName":"item" - } - }, - "PrivateIpAddressStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PrivateIpAddress" - } - }, - "ProductCode":{ - "type":"structure", - "members":{ - "ProductCodeId":{ - "shape":"String", - "locationName":"productCode" - }, - "ProductCodeType":{ - "shape":"ProductCodeValues", - "locationName":"type" - } - } - }, - "ProductCodeList":{ - "type":"list", - "member":{ - "shape":"ProductCode", - "locationName":"item" - } - }, - "ProductCodeStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ProductCode" - } - }, - "ProductCodeValues":{ - "type":"string", - "enum":[ - "devpay", - "marketplace" - ] - }, - "ProductDescriptionList":{ - "type":"list", - "member":{"shape":"String"} - }, - "PropagatingVgw":{ - "type":"structure", - "members":{ - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - } - } - }, - "PropagatingVgwList":{ - "type":"list", - "member":{ - "shape":"PropagatingVgw", - "locationName":"item" - } - }, - "ProvisionedBandwidth":{ - "type":"structure", - "members":{ - "Provisioned":{ - "shape":"String", - "locationName":"provisioned" - }, - "Requested":{ - "shape":"String", - "locationName":"requested" - }, - "RequestTime":{ - "shape":"DateTime", - "locationName":"requestTime" - }, - "ProvisionTime":{ - "shape":"DateTime", - "locationName":"provisionTime" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "PublicIpStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"PublicIp" - } - }, - "Purchase":{ - "type":"structure", - "members":{ - "HostReservationId":{ - "shape":"String", - "locationName":"hostReservationId" - }, - "HostIdSet":{ - "shape":"ResponseHostIdSet", - "locationName":"hostIdSet" - }, - "InstanceFamily":{ - "shape":"String", - "locationName":"instanceFamily" - }, - "PaymentOption":{ - "shape":"PaymentOption", - "locationName":"paymentOption" - }, - "UpfrontPrice":{ - "shape":"String", - "locationName":"upfrontPrice" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "Duration":{ - "shape":"Integer", - "locationName":"duration" - } - } - }, - "PurchaseHostReservationRequest":{ - "type":"structure", - "required":[ - "OfferingId", - "HostIdSet" - ], - "members":{ - "OfferingId":{"shape":"String"}, - "HostIdSet":{"shape":"RequestHostIdSet"}, - "LimitPrice":{"shape":"String"}, - "CurrencyCode":{"shape":"CurrencyCodeValues"}, - "ClientToken":{"shape":"String"} - } - }, - "PurchaseHostReservationResult":{ - "type":"structure", - "members":{ - "Purchase":{ - "shape":"PurchaseSet", - "locationName":"purchase" - }, - "TotalUpfrontPrice":{ - "shape":"String", - "locationName":"totalUpfrontPrice" - }, - "TotalHourlyPrice":{ - "shape":"String", - "locationName":"totalHourlyPrice" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "PurchaseRequest":{ - "type":"structure", - "required":[ - "PurchaseToken", - "InstanceCount" - ], - "members":{ - "PurchaseToken":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"} - } - }, - "PurchaseRequestSet":{ - "type":"list", - "member":{ - "shape":"PurchaseRequest", - "locationName":"PurchaseRequest" - }, - "min":1 - }, - "PurchaseReservedInstancesOfferingRequest":{ - "type":"structure", - "required":[ - "ReservedInstancesOfferingId", - "InstanceCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ReservedInstancesOfferingId":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"}, - "LimitPrice":{ - "shape":"ReservedInstanceLimitPrice", - "locationName":"limitPrice" - } - } - }, - "PurchaseReservedInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "PurchaseScheduledInstancesRequest":{ - "type":"structure", - "required":["PurchaseRequests"], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ClientToken":{ - "shape":"String", - "idempotencyToken":true - }, - "PurchaseRequests":{ - "shape":"PurchaseRequestSet", - "locationName":"PurchaseRequest" - } - } - }, - "PurchaseScheduledInstancesResult":{ - "type":"structure", - "members":{ - "ScheduledInstanceSet":{ - "shape":"PurchasedScheduledInstanceSet", - "locationName":"scheduledInstanceSet" - } - } - }, - "PurchaseSet":{ - "type":"list", - "member":{"shape":"Purchase"} - }, - "PurchasedScheduledInstanceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstance", - "locationName":"item" - } - }, - "RIProductDescription":{ - "type":"string", - "enum":[ - "Linux/UNIX", - "Linux/UNIX (Amazon VPC)", - "Windows", - "Windows (Amazon VPC)" - ] - }, - "ReasonCodesList":{ - "type":"list", - "member":{ - "shape":"ReportInstanceReasonCodes", - "locationName":"item" - } - }, - "RebootInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"RecurringChargeFrequency", - "locationName":"frequency" - }, - "Amount":{ - "shape":"Double", - "locationName":"amount" - } - } - }, - "RecurringChargeFrequency":{ - "type":"string", - "enum":["Hourly"] - }, - "RecurringChargesList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"item" - } - }, - "Region":{ - "type":"structure", - "members":{ - "RegionName":{ - "shape":"String", - "locationName":"regionName" - }, - "Endpoint":{ - "shape":"String", - "locationName":"regionEndpoint" - } - } - }, - "RegionList":{ - "type":"list", - "member":{ - "shape":"Region", - "locationName":"item" - } - }, - "RegionNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"RegionName" - } - }, - "RegisterImageRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageLocation":{"shape":"String"}, - "Name":{ - "shape":"String", - "locationName":"name" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Architecture":{ - "shape":"ArchitectureValues", - "locationName":"architecture" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "RootDeviceName":{ - "shape":"String", - "locationName":"rootDeviceName" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "VirtualizationType":{ - "shape":"String", - "locationName":"virtualizationType" - }, - "SriovNetSupport":{ - "shape":"String", - "locationName":"sriovNetSupport" - }, - "EnaSupport":{ - "shape":"Boolean", - "locationName":"enaSupport" - } - } - }, - "RegisterImageResult":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - } - } - }, - "RejectVpcPeeringConnectionRequest":{ - "type":"structure", - "required":["VpcPeeringConnectionId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "RejectVpcPeeringConnectionResult":{ - "type":"structure", - "members":{ - "Return":{ - "shape":"Boolean", - "locationName":"return" - } - } - }, - "ReleaseAddressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{"shape":"String"}, - "AllocationId":{"shape":"String"} - } - }, - "ReleaseHostsRequest":{ - "type":"structure", - "required":["HostIds"], - "members":{ - "HostIds":{ - "shape":"RequestHostIdList", - "locationName":"hostId" - } - } - }, - "ReleaseHostsResult":{ - "type":"structure", - "members":{ - "Successful":{ - "shape":"ResponseHostIdList", - "locationName":"successful" - }, - "Unsuccessful":{ - "shape":"UnsuccessfulItemList", - "locationName":"unsuccessful" - } - } - }, - "ReplaceNetworkAclAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "NetworkAclId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - } - } - }, - "ReplaceNetworkAclAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReplaceNetworkAclEntryRequest":{ - "type":"structure", - "required":[ - "NetworkAclId", - "RuleNumber", - "Protocol", - "RuleAction", - "Egress", - "CidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkAclId":{ - "shape":"String", - "locationName":"networkAclId" - }, - "RuleNumber":{ - "shape":"Integer", - "locationName":"ruleNumber" - }, - "Protocol":{ - "shape":"String", - "locationName":"protocol" - }, - "RuleAction":{ - "shape":"RuleAction", - "locationName":"ruleAction" - }, - "Egress":{ - "shape":"Boolean", - "locationName":"egress" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "IcmpTypeCode":{ - "shape":"IcmpTypeCode", - "locationName":"Icmp" - }, - "PortRange":{ - "shape":"PortRange", - "locationName":"portRange" - } - } - }, - "ReplaceRouteRequest":{ - "type":"structure", - "required":[ - "RouteTableId", - "DestinationCidrBlock" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - } - } - }, - "ReplaceRouteTableAssociationRequest":{ - "type":"structure", - "required":[ - "AssociationId", - "RouteTableId" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "AssociationId":{ - "shape":"String", - "locationName":"associationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - } - } - }, - "ReplaceRouteTableAssociationResult":{ - "type":"structure", - "members":{ - "NewAssociationId":{ - "shape":"String", - "locationName":"newAssociationId" - } - } - }, - "ReportInstanceReasonCodes":{ - "type":"string", - "enum":[ - "instance-stuck-in-state", - "unresponsive", - "not-accepting-credentials", - "password-not-available", - "performance-network", - "performance-instance-store", - "performance-ebs-volume", - "performance-other", - "other" - ] - }, - "ReportInstanceStatusRequest":{ - "type":"structure", - "required":[ - "Instances", - "Status", - "ReasonCodes" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "Instances":{ - "shape":"InstanceIdStringList", - "locationName":"instanceId" - }, - "Status":{ - "shape":"ReportStatusType", - "locationName":"status" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "EndTime":{ - "shape":"DateTime", - "locationName":"endTime" - }, - "ReasonCodes":{ - "shape":"ReasonCodesList", - "locationName":"reasonCode" - }, - "Description":{ - "shape":"String", - "locationName":"description" - } - } - }, - "ReportStatusType":{ - "type":"string", - "enum":[ - "ok", - "impaired" - ] - }, - "RequestHostIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RequestHostIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RequestSpotFleetRequest":{ - "type":"structure", - "required":["SpotFleetRequestConfig"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - } - } - }, - "RequestSpotFleetResponse":{ - "type":"structure", - "required":["SpotFleetRequestId"], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - } - } - }, - "RequestSpotInstancesRequest":{ - "type":"structure", - "required":["SpotPrice"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "BlockDurationMinutes":{ - "shape":"Integer", - "locationName":"blockDurationMinutes" - }, - "LaunchSpecification":{"shape":"RequestSpotLaunchSpecification"} - } - }, - "RequestSpotInstancesResult":{ - "type":"structure", - "members":{ - "SpotInstanceRequests":{ - "shape":"SpotInstanceRequestList", - "locationName":"spotInstanceRequestSet" - } - } - }, - "RequestSpotLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"ValueStringList", - "locationName":"SecurityGroup" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "Monitoring":{ - "shape":"RunInstancesMonitoringEnabled", - "locationName":"monitoring" - }, - "SecurityGroupIds":{ - "shape":"ValueStringList", - "locationName":"SecurityGroupId" - } - } - }, - "Reservation":{ - "type":"structure", - "members":{ - "ReservationId":{ - "shape":"String", - "locationName":"reservationId" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "RequesterId":{ - "shape":"String", - "locationName":"requesterId" - }, - "Groups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "Instances":{ - "shape":"InstanceList", - "locationName":"instancesSet" - } - } - }, - "ReservationList":{ - "type":"list", - "member":{ - "shape":"Reservation", - "locationName":"item" - } - }, - "ReservationState":{ - "type":"string", - "enum":[ - "payment-pending", - "payment-failed", - "active", - "retired" - ] - }, - "ReservationValue":{ - "type":"structure", - "members":{ - "RemainingTotalValue":{ - "shape":"String", - "locationName":"remainingTotalValue" - }, - "RemainingUpfrontValue":{ - "shape":"String", - "locationName":"remainingUpfrontValue" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - } - } - }, - "ReservedInstanceIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstanceId" - } - }, - "ReservedInstanceLimitPrice":{ - "type":"structure", - "members":{ - "Amount":{ - "shape":"Double", - "locationName":"amount" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - } - } - }, - "ReservedInstanceReservationValue":{ - "type":"structure", - "members":{ - "ReservedInstanceId":{ - "shape":"String", - "locationName":"reservedInstanceId" - }, - "ReservationValue":{ - "shape":"ReservationValue", - "locationName":"reservationValue" - } - } - }, - "ReservedInstanceReservationValueSet":{ - "type":"list", - "member":{ - "shape":"ReservedInstanceReservationValue", - "locationName":"item" - } - }, - "ReservedInstanceState":{ - "type":"string", - "enum":[ - "payment-pending", - "active", - "payment-failed", - "retired" - ] - }, - "ReservedInstances":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Start":{ - "shape":"DateTime", - "locationName":"start" - }, - "End":{ - "shape":"DateTime", - "locationName":"end" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "State":{ - "shape":"ReservedInstanceState", - "locationName":"state" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - }, - "OfferingClass":{ - "shape":"OfferingClassType", - "locationName":"offeringClass" - }, - "Scope":{ - "shape":"scope", - "locationName":"scope" - } - } - }, - "ReservedInstancesConfiguration":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Scope":{ - "shape":"scope", - "locationName":"scope" - } - } - }, - "ReservedInstancesConfigurationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"item" - } - }, - "ReservedInstancesId":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - } - } - }, - "ReservedInstancesIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesId" - } - }, - "ReservedInstancesList":{ - "type":"list", - "member":{ - "shape":"ReservedInstances", - "locationName":"item" - } - }, - "ReservedInstancesListing":{ - "type":"structure", - "members":{ - "ReservedInstancesListingId":{ - "shape":"String", - "locationName":"reservedInstancesListingId" - }, - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "Status":{ - "shape":"ListingStatus", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "InstanceCounts":{ - "shape":"InstanceCountList", - "locationName":"instanceCounts" - }, - "PriceSchedules":{ - "shape":"PriceScheduleList", - "locationName":"priceSchedules" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesListingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesListing", - "locationName":"item" - } - }, - "ReservedInstancesModification":{ - "type":"structure", - "members":{ - "ReservedInstancesModificationId":{ - "shape":"String", - "locationName":"reservedInstancesModificationId" - }, - "ReservedInstancesIds":{ - "shape":"ReservedIntancesIds", - "locationName":"reservedInstancesSet" - }, - "ModificationResults":{ - "shape":"ReservedInstancesModificationResultList", - "locationName":"modificationResultSet" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - }, - "UpdateDate":{ - "shape":"DateTime", - "locationName":"updateDate" - }, - "EffectiveDate":{ - "shape":"DateTime", - "locationName":"effectiveDate" - }, - "Status":{ - "shape":"String", - "locationName":"status" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - } - } - }, - "ReservedInstancesModificationIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReservedInstancesModificationId" - } - }, - "ReservedInstancesModificationList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModification", - "locationName":"item" - } - }, - "ReservedInstancesModificationResult":{ - "type":"structure", - "members":{ - "ReservedInstancesId":{ - "shape":"String", - "locationName":"reservedInstancesId" - }, - "TargetConfiguration":{ - "shape":"ReservedInstancesConfiguration", - "locationName":"targetConfiguration" - } - } - }, - "ReservedInstancesModificationResultList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesModificationResult", - "locationName":"item" - } - }, - "ReservedInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedInstancesOfferingId":{ - "shape":"String", - "locationName":"reservedInstancesOfferingId" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "Duration":{ - "shape":"Long", - "locationName":"duration" - }, - "UsagePrice":{ - "shape":"Float", - "locationName":"usagePrice" - }, - "FixedPrice":{ - "shape":"Float", - "locationName":"fixedPrice" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "CurrencyCode":{ - "shape":"CurrencyCodeValues", - "locationName":"currencyCode" - }, - "OfferingType":{ - "shape":"OfferingTypeValues", - "locationName":"offeringType" - }, - "RecurringCharges":{ - "shape":"RecurringChargesList", - "locationName":"recurringCharges" - }, - "Marketplace":{ - "shape":"Boolean", - "locationName":"marketplace" - }, - "PricingDetails":{ - "shape":"PricingDetailsList", - "locationName":"pricingDetailsSet" - }, - "OfferingClass":{ - "shape":"OfferingClassType", - "locationName":"offeringClass" - }, - "Scope":{ - "shape":"scope", - "locationName":"scope" - } - } - }, - "ReservedInstancesOfferingIdStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ReservedInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesOffering", - "locationName":"item" - } - }, - "ReservedIntancesIds":{ - "type":"list", - "member":{ - "shape":"ReservedInstancesId", - "locationName":"item" - } - }, - "ResetImageAttributeName":{ - "type":"string", - "enum":["launchPermission"] - }, - "ResetImageAttributeRequest":{ - "type":"structure", - "required":[ - "ImageId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "Attribute":{"shape":"ResetImageAttributeName"} - } - }, - "ResetInstanceAttributeRequest":{ - "type":"structure", - "required":[ - "InstanceId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Attribute":{ - "shape":"InstanceAttributeName", - "locationName":"attribute" - } - } - }, - "ResetNetworkInterfaceAttributeRequest":{ - "type":"structure", - "required":["NetworkInterfaceId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "SourceDestCheck":{ - "shape":"String", - "locationName":"sourceDestCheck" - } - } - }, - "ResetSnapshotAttributeRequest":{ - "type":"structure", - "required":[ - "SnapshotId", - "Attribute" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "SnapshotId":{"shape":"String"}, - "Attribute":{"shape":"SnapshotAttributeName"} - } - }, - "ResourceIdList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ResourceType":{ - "type":"string", - "enum":[ - "customer-gateway", - "dhcp-options", - "image", - "instance", - "internet-gateway", - "network-acl", - "network-interface", - "reserved-instances", - "route-table", - "snapshot", - "spot-instances-request", - "subnet", - "security-group", - "volume", - "vpc", - "vpn-connection", - "vpn-gateway" - ] - }, - "ResponseHostIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "ResponseHostIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "RestorableByStringList":{ - "type":"list", - "member":{"shape":"String"} - }, - "RestoreAddressToClassicRequest":{ - "type":"structure", - "required":["PublicIp"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RestoreAddressToClassicResult":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"Status", - "locationName":"status" - }, - "PublicIp":{ - "shape":"String", - "locationName":"publicIp" - } - } - }, - "RevokeSecurityGroupEgressRequest":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "SourceSecurityGroupName":{ - "shape":"String", - "locationName":"sourceSecurityGroupName" - }, - "SourceSecurityGroupOwnerId":{ - "shape":"String", - "locationName":"sourceSecurityGroupOwnerId" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "CidrIp":{ - "shape":"String", - "locationName":"cidrIp" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - } - } - }, - "RevokeSecurityGroupIngressRequest":{ - "type":"structure", - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "GroupName":{"shape":"String"}, - "GroupId":{"shape":"String"}, - "SourceSecurityGroupName":{"shape":"String"}, - "SourceSecurityGroupOwnerId":{"shape":"String"}, - "IpProtocol":{"shape":"String"}, - "FromPort":{"shape":"Integer"}, - "ToPort":{"shape":"Integer"}, - "CidrIp":{"shape":"String"}, - "IpPermissions":{"shape":"IpPermissionList"} - } - }, - "Route":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "DestinationPrefixListId":{ - "shape":"String", - "locationName":"destinationPrefixListId" - }, - "GatewayId":{ - "shape":"String", - "locationName":"gatewayId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "InstanceOwnerId":{ - "shape":"String", - "locationName":"instanceOwnerId" - }, - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "NatGatewayId":{ - "shape":"String", - "locationName":"natGatewayId" - }, - "State":{ - "shape":"RouteState", - "locationName":"state" - }, - "Origin":{ - "shape":"RouteOrigin", - "locationName":"origin" - } - } - }, - "RouteList":{ - "type":"list", - "member":{ - "shape":"Route", - "locationName":"item" - } - }, - "RouteOrigin":{ - "type":"string", - "enum":[ - "CreateRouteTable", - "CreateRoute", - "EnableVgwRoutePropagation" - ] - }, - "RouteState":{ - "type":"string", - "enum":[ - "active", - "blackhole" - ] - }, - "RouteTable":{ - "type":"structure", - "members":{ - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Routes":{ - "shape":"RouteList", - "locationName":"routeSet" - }, - "Associations":{ - "shape":"RouteTableAssociationList", - "locationName":"associationSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "PropagatingVgws":{ - "shape":"PropagatingVgwList", - "locationName":"propagatingVgwSet" - } - } - }, - "RouteTableAssociation":{ - "type":"structure", - "members":{ - "RouteTableAssociationId":{ - "shape":"String", - "locationName":"routeTableAssociationId" - }, - "RouteTableId":{ - "shape":"String", - "locationName":"routeTableId" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "Main":{ - "shape":"Boolean", - "locationName":"main" - } - } - }, - "RouteTableAssociationList":{ - "type":"list", - "member":{ - "shape":"RouteTableAssociation", - "locationName":"item" - } - }, - "RouteTableList":{ - "type":"list", - "member":{ - "shape":"RouteTable", - "locationName":"item" - } - }, - "RuleAction":{ - "type":"string", - "enum":[ - "allow", - "deny" - ] - }, - "RunInstancesMonitoringEnabled":{ - "type":"structure", - "required":["Enabled"], - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "RunInstancesRequest":{ - "type":"structure", - "required":[ - "ImageId", - "MinCount", - "MaxCount" - ], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "ImageId":{"shape":"String"}, - "MinCount":{"shape":"Integer"}, - "MaxCount":{"shape":"Integer"}, - "KeyName":{"shape":"String"}, - "SecurityGroups":{ - "shape":"SecurityGroupStringList", - "locationName":"SecurityGroup" - }, - "SecurityGroupIds":{ - "shape":"SecurityGroupIdStringList", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "InstanceType":{"shape":"InstanceType"}, - "Placement":{"shape":"Placement"}, - "KernelId":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingRequestList", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"RunInstancesMonitoringEnabled"}, - "SubnetId":{"shape":"String"}, - "DisableApiTermination":{ - "shape":"Boolean", - "locationName":"disableApiTermination" - }, - "InstanceInitiatedShutdownBehavior":{ - "shape":"ShutdownBehavior", - "locationName":"instanceInitiatedShutdownBehavior" - }, - "PrivateIpAddress":{ - "shape":"String", - "locationName":"privateIpAddress" - }, - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterface" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - } - } - }, - "RunScheduledInstancesRequest":{ - "type":"structure", - "required":[ - "ScheduledInstanceId", - "LaunchSpecification" - ], - "members":{ - "DryRun":{"shape":"Boolean"}, - "ClientToken":{ - "shape":"String", - "idempotencyToken":true - }, - "InstanceCount":{"shape":"Integer"}, - "ScheduledInstanceId":{"shape":"String"}, - "LaunchSpecification":{"shape":"ScheduledInstancesLaunchSpecification"} - } - }, - "RunScheduledInstancesResult":{ - "type":"structure", - "members":{ - "InstanceIdSet":{ - "shape":"InstanceIdSet", - "locationName":"instanceIdSet" - } - } - }, - "S3Storage":{ - "type":"structure", - "members":{ - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "AWSAccessKeyId":{"shape":"String"}, - "UploadPolicy":{ - "shape":"Blob", - "locationName":"uploadPolicy" - }, - "UploadPolicySignature":{ - "shape":"String", - "locationName":"uploadPolicySignature" - } - } - }, - "ScheduledInstance":{ - "type":"structure", - "members":{ - "ScheduledInstanceId":{ - "shape":"String", - "locationName":"scheduledInstanceId" - }, - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "NetworkPlatform":{ - "shape":"String", - "locationName":"networkPlatform" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "SlotDurationInHours":{ - "shape":"Integer", - "locationName":"slotDurationInHours" - }, - "Recurrence":{ - "shape":"ScheduledInstanceRecurrence", - "locationName":"recurrence" - }, - "PreviousSlotEndTime":{ - "shape":"DateTime", - "locationName":"previousSlotEndTime" - }, - "NextSlotStartTime":{ - "shape":"DateTime", - "locationName":"nextSlotStartTime" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "TotalScheduledInstanceHours":{ - "shape":"Integer", - "locationName":"totalScheduledInstanceHours" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - }, - "TermStartDate":{ - "shape":"DateTime", - "locationName":"termStartDate" - }, - "TermEndDate":{ - "shape":"DateTime", - "locationName":"termEndDate" - }, - "CreateDate":{ - "shape":"DateTime", - "locationName":"createDate" - } - } - }, - "ScheduledInstanceAvailability":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"String", - "locationName":"instanceType" - }, - "Platform":{ - "shape":"String", - "locationName":"platform" - }, - "NetworkPlatform":{ - "shape":"String", - "locationName":"networkPlatform" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "PurchaseToken":{ - "shape":"String", - "locationName":"purchaseToken" - }, - "SlotDurationInHours":{ - "shape":"Integer", - "locationName":"slotDurationInHours" - }, - "Recurrence":{ - "shape":"ScheduledInstanceRecurrence", - "locationName":"recurrence" - }, - "FirstSlotStartTime":{ - "shape":"DateTime", - "locationName":"firstSlotStartTime" - }, - "HourlyPrice":{ - "shape":"String", - "locationName":"hourlyPrice" - }, - "TotalScheduledInstanceHours":{ - "shape":"Integer", - "locationName":"totalScheduledInstanceHours" - }, - "AvailableInstanceCount":{ - "shape":"Integer", - "locationName":"availableInstanceCount" - }, - "MinTermDurationInDays":{ - "shape":"Integer", - "locationName":"minTermDurationInDays" - }, - "MaxTermDurationInDays":{ - "shape":"Integer", - "locationName":"maxTermDurationInDays" - } - } - }, - "ScheduledInstanceAvailabilitySet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstanceAvailability", - "locationName":"item" - } - }, - "ScheduledInstanceIdRequestSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ScheduledInstanceId" - } - }, - "ScheduledInstanceRecurrence":{ - "type":"structure", - "members":{ - "Frequency":{ - "shape":"String", - "locationName":"frequency" - }, - "Interval":{ - "shape":"Integer", - "locationName":"interval" - }, - "OccurrenceDaySet":{ - "shape":"OccurrenceDaySet", - "locationName":"occurrenceDaySet" - }, - "OccurrenceRelativeToEnd":{ - "shape":"Boolean", - "locationName":"occurrenceRelativeToEnd" - }, - "OccurrenceUnit":{ - "shape":"String", - "locationName":"occurrenceUnit" - } - } - }, - "ScheduledInstanceRecurrenceRequest":{ - "type":"structure", - "members":{ - "Frequency":{"shape":"String"}, - "Interval":{"shape":"Integer"}, - "OccurrenceDays":{ - "shape":"OccurrenceDayRequestSet", - "locationName":"OccurrenceDay" - }, - "OccurrenceRelativeToEnd":{"shape":"Boolean"}, - "OccurrenceUnit":{"shape":"String"} - } - }, - "ScheduledInstanceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstance", - "locationName":"item" - } - }, - "ScheduledInstancesBlockDeviceMapping":{ - "type":"structure", - "members":{ - "DeviceName":{"shape":"String"}, - "NoDevice":{"shape":"String"}, - "VirtualName":{"shape":"String"}, - "Ebs":{"shape":"ScheduledInstancesEbs"} - } - }, - "ScheduledInstancesBlockDeviceMappingSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesBlockDeviceMapping", - "locationName":"BlockDeviceMapping" - } - }, - "ScheduledInstancesEbs":{ - "type":"structure", - "members":{ - "SnapshotId":{"shape":"String"}, - "VolumeSize":{"shape":"Integer"}, - "DeleteOnTermination":{"shape":"Boolean"}, - "VolumeType":{"shape":"String"}, - "Iops":{"shape":"Integer"}, - "Encrypted":{"shape":"Boolean"} - } - }, - "ScheduledInstancesIamInstanceProfile":{ - "type":"structure", - "members":{ - "Arn":{"shape":"String"}, - "Name":{"shape":"String"} - } - }, - "ScheduledInstancesLaunchSpecification":{ - "type":"structure", - "required":["ImageId"], - "members":{ - "ImageId":{"shape":"String"}, - "KeyName":{"shape":"String"}, - "SecurityGroupIds":{ - "shape":"ScheduledInstancesSecurityGroupIdSet", - "locationName":"SecurityGroupId" - }, - "UserData":{"shape":"String"}, - "Placement":{"shape":"ScheduledInstancesPlacement"}, - "KernelId":{"shape":"String"}, - "InstanceType":{"shape":"String"}, - "RamdiskId":{"shape":"String"}, - "BlockDeviceMappings":{ - "shape":"ScheduledInstancesBlockDeviceMappingSet", - "locationName":"BlockDeviceMapping" - }, - "Monitoring":{"shape":"ScheduledInstancesMonitoring"}, - "SubnetId":{"shape":"String"}, - "NetworkInterfaces":{ - "shape":"ScheduledInstancesNetworkInterfaceSet", - "locationName":"NetworkInterface" - }, - "IamInstanceProfile":{"shape":"ScheduledInstancesIamInstanceProfile"}, - "EbsOptimized":{"shape":"Boolean"} - } - }, - "ScheduledInstancesMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{"shape":"Boolean"} - } - }, - "ScheduledInstancesNetworkInterface":{ - "type":"structure", - "members":{ - "NetworkInterfaceId":{"shape":"String"}, - "DeviceIndex":{"shape":"Integer"}, - "SubnetId":{"shape":"String"}, - "Description":{"shape":"String"}, - "PrivateIpAddress":{"shape":"String"}, - "PrivateIpAddressConfigs":{ - "shape":"PrivateIpAddressConfigSet", - "locationName":"PrivateIpAddressConfig" - }, - "SecondaryPrivateIpAddressCount":{"shape":"Integer"}, - "AssociatePublicIpAddress":{"shape":"Boolean"}, - "Groups":{ - "shape":"ScheduledInstancesSecurityGroupIdSet", - "locationName":"Group" - }, - "DeleteOnTermination":{"shape":"Boolean"} - } - }, - "ScheduledInstancesNetworkInterfaceSet":{ - "type":"list", - "member":{ - "shape":"ScheduledInstancesNetworkInterface", - "locationName":"NetworkInterface" - } - }, - "ScheduledInstancesPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{"shape":"String"}, - "GroupName":{"shape":"String"} - } - }, - "ScheduledInstancesPrivateIpAddressConfig":{ - "type":"structure", - "members":{ - "PrivateIpAddress":{"shape":"String"}, - "Primary":{"shape":"Boolean"} - } - }, - "ScheduledInstancesSecurityGroupIdSet":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "Description":{ - "shape":"String", - "locationName":"groupDescription" - }, - "IpPermissions":{ - "shape":"IpPermissionList", - "locationName":"ipPermissions" - }, - "IpPermissionsEgress":{ - "shape":"IpPermissionList", - "locationName":"ipPermissionsEgress" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SecurityGroupIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroupId" - } - }, - "SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"SecurityGroup", - "locationName":"item" - } - }, - "SecurityGroupReference":{ - "type":"structure", - "required":[ - "GroupId", - "ReferencingVpcId" - ], - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "ReferencingVpcId":{ - "shape":"String", - "locationName":"referencingVpcId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "SecurityGroupReferences":{ - "type":"list", - "member":{ - "shape":"SecurityGroupReference", - "locationName":"item" - } - }, - "SecurityGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SecurityGroup" - } - }, - "ShutdownBehavior":{ - "type":"string", - "enum":[ - "stop", - "terminate" - ] - }, - "SlotDateTimeRangeRequest":{ - "type":"structure", - "required":[ - "EarliestTime", - "LatestTime" - ], - "members":{ - "EarliestTime":{"shape":"DateTime"}, - "LatestTime":{"shape":"DateTime"} - } - }, - "SlotStartTimeRangeRequest":{ - "type":"structure", - "members":{ - "EarliestTime":{"shape":"DateTime"}, - "LatestTime":{"shape":"DateTime"} - } - }, - "Snapshot":{ - "type":"structure", - "members":{ - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "State":{ - "shape":"SnapshotState", - "locationName":"status" - }, - "StateMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "StartTime":{ - "shape":"DateTime", - "locationName":"startTime" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VolumeSize":{ - "shape":"Integer", - "locationName":"volumeSize" - }, - "OwnerAlias":{ - "shape":"String", - "locationName":"ownerAlias" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - }, - "DataEncryptionKeyId":{ - "shape":"String", - "locationName":"dataEncryptionKeyId" - } - } - }, - "SnapshotAttributeName":{ - "type":"string", - "enum":[ - "productCodes", - "createVolumePermission" - ] - }, - "SnapshotDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "DeviceName":{ - "shape":"String", - "locationName":"deviceName" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SnapshotDetailList":{ - "type":"list", - "member":{ - "shape":"SnapshotDetail", - "locationName":"item" - } - }, - "SnapshotDiskContainer":{ - "type":"structure", - "members":{ - "Description":{"shape":"String"}, - "Format":{"shape":"String"}, - "Url":{"shape":"String"}, - "UserBucket":{"shape":"UserBucket"} - } - }, - "SnapshotIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SnapshotId" - } - }, - "SnapshotList":{ - "type":"list", - "member":{ - "shape":"Snapshot", - "locationName":"item" - } - }, - "SnapshotState":{ - "type":"string", - "enum":[ - "pending", - "completed", - "error" - ] - }, - "SnapshotTaskDetail":{ - "type":"structure", - "members":{ - "DiskImageSize":{ - "shape":"Double", - "locationName":"diskImageSize" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "Format":{ - "shape":"String", - "locationName":"format" - }, - "Url":{ - "shape":"String", - "locationName":"url" - }, - "UserBucket":{ - "shape":"UserBucketDetails", - "locationName":"userBucket" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "Progress":{ - "shape":"String", - "locationName":"progress" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "SpotDatafeedSubscription":{ - "type":"structure", - "members":{ - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "Bucket":{ - "shape":"String", - "locationName":"bucket" - }, - "Prefix":{ - "shape":"String", - "locationName":"prefix" - }, - "State":{ - "shape":"DatafeedSubscriptionState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - } - } - }, - "SpotFleetLaunchSpecification":{ - "type":"structure", - "members":{ - "ImageId":{ - "shape":"String", - "locationName":"imageId" - }, - "KeyName":{ - "shape":"String", - "locationName":"keyName" - }, - "SecurityGroups":{ - "shape":"GroupIdentifierList", - "locationName":"groupSet" - }, - "UserData":{ - "shape":"String", - "locationName":"userData" - }, - "AddressingType":{ - "shape":"String", - "locationName":"addressingType" - }, - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "Placement":{ - "shape":"SpotPlacement", - "locationName":"placement" - }, - "KernelId":{ - "shape":"String", - "locationName":"kernelId" - }, - "RamdiskId":{ - "shape":"String", - "locationName":"ramdiskId" - }, - "BlockDeviceMappings":{ - "shape":"BlockDeviceMappingList", - "locationName":"blockDeviceMapping" - }, - "Monitoring":{ - "shape":"SpotFleetMonitoring", - "locationName":"monitoring" - }, - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "NetworkInterfaces":{ - "shape":"InstanceNetworkInterfaceSpecificationList", - "locationName":"networkInterfaceSet" - }, - "IamInstanceProfile":{ - "shape":"IamInstanceProfileSpecification", - "locationName":"iamInstanceProfile" - }, - "EbsOptimized":{ - "shape":"Boolean", - "locationName":"ebsOptimized" - }, - "WeightedCapacity":{ - "shape":"Double", - "locationName":"weightedCapacity" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - } - } - }, - "SpotFleetMonitoring":{ - "type":"structure", - "members":{ - "Enabled":{ - "shape":"Boolean", - "locationName":"enabled" - } - } - }, - "SpotFleetRequestConfig":{ - "type":"structure", - "required":[ - "SpotFleetRequestId", - "SpotFleetRequestState", - "SpotFleetRequestConfig", - "CreateTime" - ], - "members":{ - "SpotFleetRequestId":{ - "shape":"String", - "locationName":"spotFleetRequestId" - }, - "SpotFleetRequestState":{ - "shape":"BatchState", - "locationName":"spotFleetRequestState" - }, - "SpotFleetRequestConfig":{ - "shape":"SpotFleetRequestConfigData", - "locationName":"spotFleetRequestConfig" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "ActivityStatus":{ - "shape":"ActivityStatus", - "locationName":"activityStatus" - } - } - }, - "SpotFleetRequestConfigData":{ - "type":"structure", - "required":[ - "SpotPrice", - "TargetCapacity", - "IamFleetRole", - "LaunchSpecifications" - ], - "members":{ - "ClientToken":{ - "shape":"String", - "locationName":"clientToken" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "TargetCapacity":{ - "shape":"Integer", - "locationName":"targetCapacity" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "TerminateInstancesWithExpiration":{ - "shape":"Boolean", - "locationName":"terminateInstancesWithExpiration" - }, - "IamFleetRole":{ - "shape":"String", - "locationName":"iamFleetRole" - }, - "LaunchSpecifications":{ - "shape":"LaunchSpecsList", - "locationName":"launchSpecifications" - }, - "ExcessCapacityTerminationPolicy":{ - "shape":"ExcessCapacityTerminationPolicy", - "locationName":"excessCapacityTerminationPolicy" - }, - "AllocationStrategy":{ - "shape":"AllocationStrategy", - "locationName":"allocationStrategy" - }, - "FulfilledCapacity":{ - "shape":"Double", - "locationName":"fulfilledCapacity" - }, - "Type":{ - "shape":"FleetType", - "locationName":"type" - } - } - }, - "SpotFleetRequestConfigSet":{ - "type":"list", - "member":{ - "shape":"SpotFleetRequestConfig", - "locationName":"item" - } - }, - "SpotInstanceRequest":{ - "type":"structure", - "members":{ - "SpotInstanceRequestId":{ - "shape":"String", - "locationName":"spotInstanceRequestId" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Type":{ - "shape":"SpotInstanceType", - "locationName":"type" - }, - "State":{ - "shape":"SpotInstanceState", - "locationName":"state" - }, - "Fault":{ - "shape":"SpotInstanceStateFault", - "locationName":"fault" - }, - "Status":{ - "shape":"SpotInstanceStatus", - "locationName":"status" - }, - "ValidFrom":{ - "shape":"DateTime", - "locationName":"validFrom" - }, - "ValidUntil":{ - "shape":"DateTime", - "locationName":"validUntil" - }, - "LaunchGroup":{ - "shape":"String", - "locationName":"launchGroup" - }, - "AvailabilityZoneGroup":{ - "shape":"String", - "locationName":"availabilityZoneGroup" - }, - "LaunchSpecification":{ - "shape":"LaunchSpecification", - "locationName":"launchSpecification" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "BlockDurationMinutes":{ - "shape":"Integer", - "locationName":"blockDurationMinutes" - }, - "ActualBlockHourlyPrice":{ - "shape":"String", - "locationName":"actualBlockHourlyPrice" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "LaunchedAvailabilityZone":{ - "shape":"String", - "locationName":"launchedAvailabilityZone" - } - } - }, - "SpotInstanceRequestIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SpotInstanceRequestId" - } - }, - "SpotInstanceRequestList":{ - "type":"list", - "member":{ - "shape":"SpotInstanceRequest", - "locationName":"item" - } - }, - "SpotInstanceState":{ - "type":"string", - "enum":[ - "open", - "active", - "closed", - "cancelled", - "failed" - ] - }, - "SpotInstanceStateFault":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceStatus":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "UpdateTime":{ - "shape":"DateTime", - "locationName":"updateTime" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "SpotInstanceType":{ - "type":"string", - "enum":[ - "one-time", - "persistent" - ] - }, - "SpotPlacement":{ - "type":"structure", - "members":{ - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - } - } - }, - "SpotPrice":{ - "type":"structure", - "members":{ - "InstanceType":{ - "shape":"InstanceType", - "locationName":"instanceType" - }, - "ProductDescription":{ - "shape":"RIProductDescription", - "locationName":"productDescription" - }, - "SpotPrice":{ - "shape":"String", - "locationName":"spotPrice" - }, - "Timestamp":{ - "shape":"DateTime", - "locationName":"timestamp" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - } - } - }, - "SpotPriceHistoryList":{ - "type":"list", - "member":{ - "shape":"SpotPrice", - "locationName":"item" - } - }, - "StaleIpPermission":{ - "type":"structure", - "members":{ - "FromPort":{ - "shape":"Integer", - "locationName":"fromPort" - }, - "IpProtocol":{ - "shape":"String", - "locationName":"ipProtocol" - }, - "IpRanges":{ - "shape":"IpRanges", - "locationName":"ipRanges" - }, - "PrefixListIds":{ - "shape":"PrefixListIdSet", - "locationName":"prefixListIds" - }, - "ToPort":{ - "shape":"Integer", - "locationName":"toPort" - }, - "UserIdGroupPairs":{ - "shape":"UserIdGroupPairSet", - "locationName":"groups" - } - } - }, - "StaleIpPermissionSet":{ - "type":"list", - "member":{ - "shape":"StaleIpPermission", - "locationName":"item" - } - }, - "StaleSecurityGroup":{ - "type":"structure", - "required":["GroupId"], - "members":{ - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "StaleIpPermissions":{ - "shape":"StaleIpPermissionSet", - "locationName":"staleIpPermissions" - }, - "StaleIpPermissionsEgress":{ - "shape":"StaleIpPermissionSet", - "locationName":"staleIpPermissionsEgress" - } - } - }, - "StaleSecurityGroupSet":{ - "type":"list", - "member":{ - "shape":"StaleSecurityGroup", - "locationName":"item" - } - }, - "StartInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "AdditionalInfo":{ - "shape":"String", - "locationName":"additionalInfo" - }, - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - } - } - }, - "StartInstancesResult":{ - "type":"structure", - "members":{ - "StartingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "State":{ - "type":"string", - "enum":[ - "Pending", - "Available", - "Deleting", - "Deleted" - ] - }, - "StateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "Status":{ - "type":"string", - "enum":[ - "MoveInProgress", - "InVpc", - "InClassic" - ] - }, - "StatusName":{ - "type":"string", - "enum":["reachability"] - }, - "StatusType":{ - "type":"string", - "enum":[ - "passed", - "failed", - "insufficient-data", - "initializing" - ] - }, - "StopInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - }, - "Force":{ - "shape":"Boolean", - "locationName":"force" - } - } - }, - "StopInstancesResult":{ - "type":"structure", - "members":{ - "StoppingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "Storage":{ - "type":"structure", - "members":{ - "S3":{"shape":"S3Storage"} - } - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetId":{ - "shape":"String", - "locationName":"subnetId" - }, - "State":{ - "shape":"SubnetState", - "locationName":"state" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "AvailableIpAddressCount":{ - "shape":"Integer", - "locationName":"availableIpAddressCount" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "DefaultForAz":{ - "shape":"Boolean", - "locationName":"defaultForAz" - }, - "MapPublicIpOnLaunch":{ - "shape":"Boolean", - "locationName":"mapPublicIpOnLaunch" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "SubnetIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetId" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"item" - } - }, - "SubnetState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "SummaryStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data", - "not-applicable", - "initializing" - ] - }, - "Tag":{ - "type":"structure", - "members":{ - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescription":{ - "type":"structure", - "members":{ - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - }, - "ResourceType":{ - "shape":"ResourceType", - "locationName":"resourceType" - }, - "Key":{ - "shape":"String", - "locationName":"key" - }, - "Value":{ - "shape":"String", - "locationName":"value" - } - } - }, - "TagDescriptionList":{ - "type":"list", - "member":{ - "shape":"TagDescription", - "locationName":"item" - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"item" - } - }, - "TargetConfiguration":{ - "type":"structure", - "members":{ - "OfferingId":{ - "shape":"String", - "locationName":"offeringId" - }, - "InstanceCount":{ - "shape":"Integer", - "locationName":"instanceCount" - } - } - }, - "TargetConfigurationRequest":{ - "type":"structure", - "required":["OfferingId"], - "members":{ - "OfferingId":{"shape":"String"}, - "InstanceCount":{"shape":"Integer"} - } - }, - "TargetConfigurationRequestSet":{ - "type":"list", - "member":{ - "shape":"TargetConfigurationRequest", - "locationName":"TargetConfigurationRequest" - } - }, - "TargetReservationValue":{ - "type":"structure", - "members":{ - "TargetConfiguration":{ - "shape":"TargetConfiguration", - "locationName":"targetConfiguration" - }, - "ReservationValue":{ - "shape":"ReservationValue", - "locationName":"reservationValue" - } - } - }, - "TargetReservationValueSet":{ - "type":"list", - "member":{ - "shape":"TargetReservationValue", - "locationName":"item" - } - }, - "TelemetryStatus":{ - "type":"string", - "enum":[ - "UP", - "DOWN" - ] - }, - "Tenancy":{ - "type":"string", - "enum":[ - "default", - "dedicated", - "host" - ] - }, - "TerminateInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "TerminateInstancesResult":{ - "type":"structure", - "members":{ - "TerminatingInstances":{ - "shape":"InstanceStateChangeList", - "locationName":"instancesSet" - } - } - }, - "TrafficType":{ - "type":"string", - "enum":[ - "ACCEPT", - "REJECT", - "ALL" - ] - }, - "UnassignPrivateIpAddressesRequest":{ - "type":"structure", - "required":[ - "NetworkInterfaceId", - "PrivateIpAddresses" - ], - "members":{ - "NetworkInterfaceId":{ - "shape":"String", - "locationName":"networkInterfaceId" - }, - "PrivateIpAddresses":{ - "shape":"PrivateIpAddressStringList", - "locationName":"privateIpAddress" - } - } - }, - "UnmonitorInstancesRequest":{ - "type":"structure", - "required":["InstanceIds"], - "members":{ - "DryRun":{ - "shape":"Boolean", - "locationName":"dryRun" - }, - "InstanceIds":{ - "shape":"InstanceIdStringList", - "locationName":"InstanceId" - } - } - }, - "UnmonitorInstancesResult":{ - "type":"structure", - "members":{ - "InstanceMonitorings":{ - "shape":"InstanceMonitoringList", - "locationName":"instancesSet" - } - } - }, - "UnsuccessfulItem":{ - "type":"structure", - "required":["Error"], - "members":{ - "Error":{ - "shape":"UnsuccessfulItemError", - "locationName":"error" - }, - "ResourceId":{ - "shape":"String", - "locationName":"resourceId" - } - } - }, - "UnsuccessfulItemError":{ - "type":"structure", - "required":[ - "Code", - "Message" - ], - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "UnsuccessfulItemList":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UnsuccessfulItemSet":{ - "type":"list", - "member":{ - "shape":"UnsuccessfulItem", - "locationName":"item" - } - }, - "UserBucket":{ - "type":"structure", - "members":{ - "S3Bucket":{"shape":"String"}, - "S3Key":{"shape":"String"} - } - }, - "UserBucketDetails":{ - "type":"structure", - "members":{ - "S3Bucket":{ - "shape":"String", - "locationName":"s3Bucket" - }, - "S3Key":{ - "shape":"String", - "locationName":"s3Key" - } - } - }, - "UserData":{ - "type":"structure", - "members":{ - "Data":{ - "shape":"String", - "locationName":"data" - } - } - }, - "UserGroupStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserGroup" - } - }, - "UserIdGroupPair":{ - "type":"structure", - "members":{ - "UserId":{ - "shape":"String", - "locationName":"userId" - }, - "GroupName":{ - "shape":"String", - "locationName":"groupName" - }, - "GroupId":{ - "shape":"String", - "locationName":"groupId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - }, - "PeeringStatus":{ - "shape":"String", - "locationName":"peeringStatus" - } - } - }, - "UserIdGroupPairList":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdGroupPairSet":{ - "type":"list", - "member":{ - "shape":"UserIdGroupPair", - "locationName":"item" - } - }, - "UserIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"UserId" - } - }, - "ValueStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"item" - } - }, - "VgwTelemetry":{ - "type":"structure", - "members":{ - "OutsideIpAddress":{ - "shape":"String", - "locationName":"outsideIpAddress" - }, - "Status":{ - "shape":"TelemetryStatus", - "locationName":"status" - }, - "LastStatusChange":{ - "shape":"DateTime", - "locationName":"lastStatusChange" - }, - "StatusMessage":{ - "shape":"String", - "locationName":"statusMessage" - }, - "AcceptedRouteCount":{ - "shape":"Integer", - "locationName":"acceptedRouteCount" - } - } - }, - "VgwTelemetryList":{ - "type":"list", - "member":{ - "shape":"VgwTelemetry", - "locationName":"item" - } - }, - "VirtualizationType":{ - "type":"string", - "enum":[ - "hvm", - "paravirtual" - ] - }, - "Volume":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "Size":{ - "shape":"Integer", - "locationName":"size" - }, - "SnapshotId":{ - "shape":"String", - "locationName":"snapshotId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "State":{ - "shape":"VolumeState", - "locationName":"status" - }, - "CreateTime":{ - "shape":"DateTime", - "locationName":"createTime" - }, - "Attachments":{ - "shape":"VolumeAttachmentList", - "locationName":"attachmentSet" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VolumeType":{ - "shape":"VolumeType", - "locationName":"volumeType" - }, - "Iops":{ - "shape":"Integer", - "locationName":"iops" - }, - "Encrypted":{ - "shape":"Boolean", - "locationName":"encrypted" - }, - "KmsKeyId":{ - "shape":"String", - "locationName":"kmsKeyId" - } - } - }, - "VolumeAttachment":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "InstanceId":{ - "shape":"String", - "locationName":"instanceId" - }, - "Device":{ - "shape":"String", - "locationName":"device" - }, - "State":{ - "shape":"VolumeAttachmentState", - "locationName":"status" - }, - "AttachTime":{ - "shape":"DateTime", - "locationName":"attachTime" - }, - "DeleteOnTermination":{ - "shape":"Boolean", - "locationName":"deleteOnTermination" - } - } - }, - "VolumeAttachmentList":{ - "type":"list", - "member":{ - "shape":"VolumeAttachment", - "locationName":"item" - } - }, - "VolumeAttachmentState":{ - "type":"string", - "enum":[ - "attaching", - "attached", - "detaching", - "detached" - ] - }, - "VolumeAttributeName":{ - "type":"string", - "enum":[ - "autoEnableIO", - "productCodes" - ] - }, - "VolumeDetail":{ - "type":"structure", - "required":["Size"], - "members":{ - "Size":{ - "shape":"Long", - "locationName":"size" - } - } - }, - "VolumeIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VolumeId" - } - }, - "VolumeList":{ - "type":"list", - "member":{ - "shape":"Volume", - "locationName":"item" - } - }, - "VolumeState":{ - "type":"string", - "enum":[ - "creating", - "available", - "in-use", - "deleting", - "deleted", - "error" - ] - }, - "VolumeStatusAction":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"String", - "locationName":"code" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusActionsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusAction", - "locationName":"item" - } - }, - "VolumeStatusDetails":{ - "type":"structure", - "members":{ - "Name":{ - "shape":"VolumeStatusName", - "locationName":"name" - }, - "Status":{ - "shape":"String", - "locationName":"status" - } - } - }, - "VolumeStatusDetailsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusDetails", - "locationName":"item" - } - }, - "VolumeStatusEvent":{ - "type":"structure", - "members":{ - "EventType":{ - "shape":"String", - "locationName":"eventType" - }, - "Description":{ - "shape":"String", - "locationName":"description" - }, - "NotBefore":{ - "shape":"DateTime", - "locationName":"notBefore" - }, - "NotAfter":{ - "shape":"DateTime", - "locationName":"notAfter" - }, - "EventId":{ - "shape":"String", - "locationName":"eventId" - } - } - }, - "VolumeStatusEventsList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusEvent", - "locationName":"item" - } - }, - "VolumeStatusInfo":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"VolumeStatusInfoStatus", - "locationName":"status" - }, - "Details":{ - "shape":"VolumeStatusDetailsList", - "locationName":"details" - } - } - }, - "VolumeStatusInfoStatus":{ - "type":"string", - "enum":[ - "ok", - "impaired", - "insufficient-data" - ] - }, - "VolumeStatusItem":{ - "type":"structure", - "members":{ - "VolumeId":{ - "shape":"String", - "locationName":"volumeId" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VolumeStatus":{ - "shape":"VolumeStatusInfo", - "locationName":"volumeStatus" - }, - "Events":{ - "shape":"VolumeStatusEventsList", - "locationName":"eventsSet" - }, - "Actions":{ - "shape":"VolumeStatusActionsList", - "locationName":"actionsSet" - } - } - }, - "VolumeStatusList":{ - "type":"list", - "member":{ - "shape":"VolumeStatusItem", - "locationName":"item" - } - }, - "VolumeStatusName":{ - "type":"string", - "enum":[ - "io-enabled", - "io-performance" - ] - }, - "VolumeType":{ - "type":"string", - "enum":[ - "standard", - "io1", - "gp2", - "sc1", - "st1" - ] - }, - "Vpc":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"VpcState", - "locationName":"state" - }, - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "DhcpOptionsId":{ - "shape":"String", - "locationName":"dhcpOptionsId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "InstanceTenancy":{ - "shape":"Tenancy", - "locationName":"instanceTenancy" - }, - "IsDefault":{ - "shape":"Boolean", - "locationName":"isDefault" - } - } - }, - "VpcAttachment":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "State":{ - "shape":"AttachmentStatus", - "locationName":"state" - } - } - }, - "VpcAttachmentList":{ - "type":"list", - "member":{ - "shape":"VpcAttachment", - "locationName":"item" - } - }, - "VpcAttributeName":{ - "type":"string", - "enum":[ - "enableDnsSupport", - "enableDnsHostnames" - ] - }, - "VpcClassicLink":{ - "type":"structure", - "members":{ - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ClassicLinkEnabled":{ - "shape":"Boolean", - "locationName":"classicLinkEnabled" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpcClassicLinkIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcClassicLinkList":{ - "type":"list", - "member":{ - "shape":"VpcClassicLink", - "locationName":"item" - } - }, - "VpcEndpoint":{ - "type":"structure", - "members":{ - "VpcEndpointId":{ - "shape":"String", - "locationName":"vpcEndpointId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "ServiceName":{ - "shape":"String", - "locationName":"serviceName" - }, - "State":{ - "shape":"State", - "locationName":"state" - }, - "PolicyDocument":{ - "shape":"String", - "locationName":"policyDocument" - }, - "RouteTableIds":{ - "shape":"ValueStringList", - "locationName":"routeTableIdSet" - }, - "CreationTimestamp":{ - "shape":"DateTime", - "locationName":"creationTimestamp" - } - } - }, - "VpcEndpointSet":{ - "type":"list", - "member":{ - "shape":"VpcEndpoint", - "locationName":"item" - } - }, - "VpcIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcId" - } - }, - "VpcList":{ - "type":"list", - "member":{ - "shape":"Vpc", - "locationName":"item" - } - }, - "VpcPeeringConnection":{ - "type":"structure", - "members":{ - "AccepterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"accepterVpcInfo" - }, - "ExpirationTime":{ - "shape":"DateTime", - "locationName":"expirationTime" - }, - "RequesterVpcInfo":{ - "shape":"VpcPeeringConnectionVpcInfo", - "locationName":"requesterVpcInfo" - }, - "Status":{ - "shape":"VpcPeeringConnectionStateReason", - "locationName":"status" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VpcPeeringConnectionId":{ - "shape":"String", - "locationName":"vpcPeeringConnectionId" - } - } - }, - "VpcPeeringConnectionList":{ - "type":"list", - "member":{ - "shape":"VpcPeeringConnection", - "locationName":"item" - } - }, - "VpcPeeringConnectionOptionsDescription":{ - "type":"structure", - "members":{ - "AllowEgressFromLocalClassicLinkToRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalClassicLinkToRemoteVpc" - }, - "AllowEgressFromLocalVpcToRemoteClassicLink":{ - "shape":"Boolean", - "locationName":"allowEgressFromLocalVpcToRemoteClassicLink" - }, - "AllowDnsResolutionFromRemoteVpc":{ - "shape":"Boolean", - "locationName":"allowDnsResolutionFromRemoteVpc" - } - } - }, - "VpcPeeringConnectionStateReason":{ - "type":"structure", - "members":{ - "Code":{ - "shape":"VpcPeeringConnectionStateReasonCode", - "locationName":"code" - }, - "Message":{ - "shape":"String", - "locationName":"message" - } - } - }, - "VpcPeeringConnectionStateReasonCode":{ - "type":"string", - "enum":[ - "initiating-request", - "pending-acceptance", - "active", - "deleted", - "rejected", - "failed", - "expired", - "provisioning", - "deleting" - ] - }, - "VpcPeeringConnectionVpcInfo":{ - "type":"structure", - "members":{ - "CidrBlock":{ - "shape":"String", - "locationName":"cidrBlock" - }, - "OwnerId":{ - "shape":"String", - "locationName":"ownerId" - }, - "VpcId":{ - "shape":"String", - "locationName":"vpcId" - }, - "PeeringOptions":{ - "shape":"VpcPeeringConnectionOptionsDescription", - "locationName":"peeringOptions" - } - } - }, - "VpcState":{ - "type":"string", - "enum":[ - "pending", - "available" - ] - }, - "VpnConnection":{ - "type":"structure", - "members":{ - "VpnConnectionId":{ - "shape":"String", - "locationName":"vpnConnectionId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "CustomerGatewayConfiguration":{ - "shape":"String", - "locationName":"customerGatewayConfiguration" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "CustomerGatewayId":{ - "shape":"String", - "locationName":"customerGatewayId" - }, - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - }, - "VgwTelemetry":{ - "shape":"VgwTelemetryList", - "locationName":"vgwTelemetry" - }, - "Options":{ - "shape":"VpnConnectionOptions", - "locationName":"options" - }, - "Routes":{ - "shape":"VpnStaticRouteList", - "locationName":"routes" - } - } - }, - "VpnConnectionIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnConnectionId" - } - }, - "VpnConnectionList":{ - "type":"list", - "member":{ - "shape":"VpnConnection", - "locationName":"item" - } - }, - "VpnConnectionOptions":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnConnectionOptionsSpecification":{ - "type":"structure", - "members":{ - "StaticRoutesOnly":{ - "shape":"Boolean", - "locationName":"staticRoutesOnly" - } - } - }, - "VpnGateway":{ - "type":"structure", - "members":{ - "VpnGatewayId":{ - "shape":"String", - "locationName":"vpnGatewayId" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - }, - "Type":{ - "shape":"GatewayType", - "locationName":"type" - }, - "AvailabilityZone":{ - "shape":"String", - "locationName":"availabilityZone" - }, - "VpcAttachments":{ - "shape":"VpcAttachmentList", - "locationName":"attachments" - }, - "Tags":{ - "shape":"TagList", - "locationName":"tagSet" - } - } - }, - "VpnGatewayIdStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpnGatewayId" - } - }, - "VpnGatewayList":{ - "type":"list", - "member":{ - "shape":"VpnGateway", - "locationName":"item" - } - }, - "VpnState":{ - "type":"string", - "enum":[ - "pending", - "available", - "deleting", - "deleted" - ] - }, - "VpnStaticRoute":{ - "type":"structure", - "members":{ - "DestinationCidrBlock":{ - "shape":"String", - "locationName":"destinationCidrBlock" - }, - "Source":{ - "shape":"VpnStaticRouteSource", - "locationName":"source" - }, - "State":{ - "shape":"VpnState", - "locationName":"state" - } - } - }, - "VpnStaticRouteList":{ - "type":"list", - "member":{ - "shape":"VpnStaticRoute", - "locationName":"item" - } - }, - "VpnStaticRouteSource":{ - "type":"string", - "enum":["Static"] - }, - "ZoneNameStringList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ZoneName" - } - }, - "scope":{ - "type":"string", - "enum":[ - "Availability Zone", - "Region" - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/docs-2.json deleted file mode 100755 index bd83c9ee9..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/docs-2.json +++ /dev/null @@ -1,6696 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon Elastic Compute Cloud

Amazon Elastic Compute Cloud (Amazon EC2) provides resizable computing capacity in the Amazon Web Services (AWS) cloud. Using Amazon EC2 eliminates your need to invest in hardware up front, so you can develop and deploy applications faster.

", - "operations": { - "AcceptReservedInstancesExchangeQuote": "

Purchases Convertible Reserved Instance offerings described in the GetReservedInstancesExchangeQuote call.

", - "AcceptVpcPeeringConnection": "

Accept a VPC peering connection request. To accept a request, the VPC peering connection must be in the pending-acceptance state, and you must be the owner of the peer VPC. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests.

", - "AllocateAddress": "

Acquires an Elastic IP address.

An Elastic IP address is for use either in the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "AllocateHosts": "

Allocates a Dedicated Host to your account. At minimum you need to specify the instance size type, Availability Zone, and quantity of hosts you want to allocate.

", - "AssignPrivateIpAddresses": "

Assigns one or more secondary private IP addresses to the specified network interface. You can specify one or more specific secondary IP addresses, or you can specify the number of secondary IP addresses to be automatically assigned within the subnet's CIDR block range. The number of secondary IP addresses that you can assign to an instance varies by instance type. For information about instance types, see Instance Types in the Amazon Elastic Compute Cloud User Guide. For more information about Elastic IP addresses, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

AssignPrivateIpAddresses is available only in EC2-VPC.

", - "AssociateAddress": "

Associates an Elastic IP address with an instance or a network interface.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

[EC2-Classic, VPC in an EC2-VPC-only account] If the Elastic IP address is already associated with a different instance, it is disassociated from that instance and associated with the specified instance.

[VPC in an EC2-Classic account] If you don't specify a private IP address, the Elastic IP address is associated with the primary IP address. If the Elastic IP address is already associated with a different instance or a network interface, you get an error unless you allow reassociation.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error, and you may be charged for each time the Elastic IP address is remapped to the same instance. For more information, see the Elastic IP Addresses section of Amazon EC2 Pricing.

", - "AssociateDhcpOptions": "

Associates a set of DHCP options (that you've previously created) with the specified VPC, or associates no DHCP options with the VPC.

After you associate the options with the VPC, any existing instances and all new instances that you launch in that VPC use the options. You don't need to restart or relaunch the instances. They automatically pick up the changes within a few hours, depending on how frequently the instance renews its DHCP lease. You can explicitly renew the lease using the operating system on the instance.

For more information, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "AssociateRouteTable": "

Associates a subnet with a route table. The subnet and route table must be in the same VPC. This association causes traffic originating from the subnet to be routed according to the routes in the route table. The action returns an association ID, which you need in order to disassociate the route table from the subnet later. A route table can be associated with multiple subnets.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "AttachClassicLinkVpc": "

Links an EC2-Classic instance to a ClassicLink-enabled VPC through one or more of the VPC's security groups. You cannot link an EC2-Classic instance to more than one VPC at a time. You can only link an instance that's in the running state. An instance is automatically unlinked from a VPC when it's stopped - you can link it to the VPC again when you restart it.

After you've linked an instance, you cannot change the VPC security groups that are associated with it. To change the security groups, you must first unlink the instance, and then link it again.

Linking your instance to a VPC is sometimes referred to as attaching your instance.

", - "AttachInternetGateway": "

Attaches an Internet gateway to a VPC, enabling connectivity between the Internet and the VPC. For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "AttachNetworkInterface": "

Attaches a network interface to an instance.

", - "AttachVolume": "

Attaches an EBS volume to a running or stopped instance and exposes it to the instance with the specified device name.

Encrypted EBS volumes may only be attached to instances that support Amazon EBS encryption. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For a list of supported device names, see Attaching an EBS Volume to an Instance. Any device names that aren't reserved for instance store volumes can be used for EBS volumes. For more information, see Amazon EC2 Instance Store in the Amazon Elastic Compute Cloud User Guide.

If a volume has an AWS Marketplace product code:

  • The volume can be attached only to a stopped instance.

  • AWS Marketplace product codes are copied from the volume to the instance.

  • You must be subscribed to the product.

  • The instance type and operating system of the instance must support the product. For example, you can't detach a volume from a Windows instance and attach it to a Linux instance.

For an overview of the AWS Marketplace, see Introducing AWS Marketplace.

For more information about EBS volumes, see Attaching Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "AttachVpnGateway": "

Attaches a virtual private gateway to a VPC. For more information, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "AuthorizeSecurityGroupEgress": "

[EC2-VPC only] Adds one or more egress rules to a security group for use with a VPC. Specifically, this action permits instances to send traffic to one or more destination CIDR IP address ranges, or to one or more destination security groups for the same VPC. This action doesn't apply to security groups for use in EC2-Classic. For more information, see Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

You can have up to 50 rules per security group (covering both ingress and egress rules).

Each rule consists of the protocol (for example, TCP), plus either a CIDR range or a source group. For the TCP and UDP protocols, you must also specify the destination port or port range. For the ICMP protocol, you must also specify the ICMP type and code. You can use -1 for the type or code to mean all types or all codes.

Rule changes are propagated to affected instances as quickly as possible. However, a small delay might occur.

", - "AuthorizeSecurityGroupIngress": "

Adds one or more ingress rules to a security group.

EC2-Classic: You can have up to 100 rules per group.

EC2-VPC: You can have up to 50 rules per group (covering both ingress and egress rules).

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

[EC2-Classic] This action gives one or more CIDR IP address ranges permission to access a security group in your account, or gives one or more security groups (called the source groups) permission to access a security group for your account. A source group can be for your own AWS account, or another.

[EC2-VPC] This action gives one or more CIDR IP address ranges permission to access a security group in your VPC, or gives one or more other security groups (called the source groups) permission to access a security group for your VPC. The security groups must all be for the same VPC.

", - "BundleInstance": "

Bundles an Amazon instance store-backed Windows instance.

During bundling, only the root device volume (C:\\) is bundled. Data on other instance store volumes is not preserved.

This action is not applicable for Linux/Unix instances or Windows instances that are backed by Amazon EBS.

For more information, see Creating an Instance Store-Backed Windows AMI.

", - "CancelBundleTask": "

Cancels a bundling operation for an instance store-backed Windows instance.

", - "CancelConversionTask": "

Cancels an active conversion task. The task can be the import of an instance or volume. The action removes all artifacts of the conversion, including a partially uploaded volume or instance. If the conversion is complete or is in the process of transferring the final disk image, the command fails and returns an exception.

For more information, see Importing a Virtual Machine Using the Amazon EC2 CLI.

", - "CancelExportTask": "

Cancels an active export task. The request removes all artifacts of the export, including any partially-created Amazon S3 objects. If the export task is complete or is in the process of transferring the final disk image, the command fails and returns an error.

", - "CancelImportTask": "

Cancels an in-process import virtual machine or import snapshot task.

", - "CancelReservedInstancesListing": "

Cancels the specified Reserved Instance listing in the Reserved Instance Marketplace.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CancelSpotFleetRequests": "

Cancels the specified Spot fleet requests.

After you cancel a Spot fleet request, the Spot fleet launches no new Spot instances. You must specify whether the Spot fleet should also terminate its Spot instances. If you terminate the instances, the Spot fleet request enters the cancelled_terminating state. Otherwise, the Spot fleet request enters the cancelled_running state and the instances continue to run until they are interrupted or you terminate them manually.

", - "CancelSpotInstanceRequests": "

Cancels one or more Spot instance requests. Spot instances are instances that Amazon EC2 starts on your behalf when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

Canceling a Spot instance request does not terminate running Spot instances associated with the request.

", - "ConfirmProductInstance": "

Determines whether a product code is associated with an instance. This action can only be used by the owner of the product code. It is useful when a product code owner needs to verify whether another user's instance is eligible for support.

", - "CopyImage": "

Initiates the copy of an AMI from the specified source region to the current region. You specify the destination region by using its endpoint when making the request.

For more information, see Copying AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshot": "

Copies a point-in-time snapshot of an EBS volume and stores it in Amazon S3. You can copy the snapshot within the same region or from one region to another. You can use the snapshot to create EBS volumes or Amazon Machine Images (AMIs). The snapshot is copied to the regional endpoint that you send the HTTP request to.

Copies of encrypted EBS snapshots remain encrypted. Copies of unencrypted snapshots remain unencrypted, unless the Encrypted flag is specified during the snapshot copy operation. By default, encrypted snapshot copies use the default AWS Key Management Service (AWS KMS) customer master key (CMK); however, you can specify a non-default CMK with the KmsKeyId parameter.

To copy an encrypted snapshot that has been shared from another account, you must have permissions for the CMK used to encrypt the snapshot.

Snapshots created by the CopySnapshot action have an arbitrary volume ID that should not be used for any purpose.

For more information, see Copying an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGateway": "

Provides information to AWS about your VPN customer gateway device. The customer gateway is the appliance at your end of the VPN connection. (The device on the AWS side of the VPN connection is the virtual private gateway.) You must provide the Internet-routable IP address of the customer gateway's external interface. The IP address must be static and may be behind a device performing network address translation (NAT).

For devices that use Border Gateway Protocol (BGP), you can also provide the device's BGP Autonomous System Number (ASN). You can use an existing ASN assigned to your network. If you don't have an ASN already, you can use a private ASN (in the 64512 - 65534 range).

Amazon EC2 supports all 2-byte ASN numbers in the range of 1 - 65534, with the exception of 7224, which is reserved in the us-east-1 region, and 9059, which is reserved in the eu-west-1 region.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

You cannot create more than one customer gateway with the same VPN type, IP address, and BGP ASN parameter values. If you run an identical request more than one time, the first request creates the customer gateway, and subsequent requests return information about the existing customer gateway. The subsequent requests do not create new customer gateway resources.

", - "CreateDhcpOptions": "

Creates a set of DHCP options for your VPC. After creating the set, you must associate it with the VPC, causing all existing and new instances that you launch in the VPC to use this set of DHCP options. The following are the individual DHCP options you can specify. For more information about the options, see RFC 2132.

  • domain-name-servers - The IP addresses of up to four domain name servers, or AmazonProvidedDNS. The default DHCP option set specifies AmazonProvidedDNS. If specifying more than one domain name server, specify the IP addresses in a single parameter, separated by commas. If you want your instance to receive a custom DNS hostname as specified in domain-name, you must set domain-name-servers to a custom DNS server.

  • domain-name - If you're using AmazonProvidedDNS in \"us-east-1\", specify \"ec2.internal\". If you're using AmazonProvidedDNS in another region, specify \"region.compute.internal\" (for example, \"ap-northeast-1.compute.internal\"). Otherwise, specify a domain name (for example, \"MyCompany.com\"). This value is used to complete unqualified DNS hostnames. Important: Some Linux operating systems accept multiple domain names separated by spaces. However, Windows and other Linux operating systems treat the value as a single domain, which results in unexpected behavior. If your DHCP options set is associated with a VPC that has instances with multiple operating systems, specify only one domain name.

  • ntp-servers - The IP addresses of up to four Network Time Protocol (NTP) servers.

  • netbios-name-servers - The IP addresses of up to four NetBIOS name servers.

  • netbios-node-type - The NetBIOS node type (1, 2, 4, or 8). We recommend that you specify 2 (broadcast and multicast are not currently supported). For more information about these node types, see RFC 2132.

Your VPC automatically starts out with a set of DHCP options that includes only a DNS server that we provide (AmazonProvidedDNS). If you create a set of options, and if your VPC has an Internet gateway, make sure to set the domain-name-servers option either to AmazonProvidedDNS or to a domain name server of your choice. For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "CreateFlowLogs": "

Creates one or more flow logs to capture IP traffic for a specific network interface, subnet, or VPC. Flow logs are delivered to a specified log group in Amazon CloudWatch Logs. If you specify a VPC or subnet in the request, a log stream is created in CloudWatch Logs for each network interface in the subnet or VPC. Log streams can include information about accepted and rejected traffic to a network interface. You can view the data in your log streams using Amazon CloudWatch Logs.

In your request, you must also specify an IAM role that has permission to publish logs to CloudWatch Logs.

", - "CreateImage": "

Creates an Amazon EBS-backed AMI from an Amazon EBS-backed instance that is either running or stopped.

If you customized your instance with instance store volumes or EBS volumes in addition to the root device volume, the new AMI contains block device mapping information for those volumes. When you launch an instance from this new AMI, the instance automatically launches with those additional volumes.

For more information, see Creating Amazon EBS-Backed Linux AMIs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateInstanceExportTask": "

Exports a running or stopped instance to an S3 bucket.

For information about the supported operating systems, image formats, and known limitations for the types of instances you can export, see Exporting an Instance as a VM Using VM Import/Export in the VM Import/Export User Guide.

", - "CreateInternetGateway": "

Creates an Internet gateway for use with a VPC. After creating the Internet gateway, you attach it to a VPC using AttachInternetGateway.

For more information about your VPC and Internet gateway, see the Amazon Virtual Private Cloud User Guide.

", - "CreateKeyPair": "

Creates a 2048-bit RSA key pair with the specified name. Amazon EC2 stores the public key and displays the private key for you to save to a file. The private key is returned as an unencrypted PEM encoded PKCS#8 private key. If a key with the specified name already exists, Amazon EC2 returns an error.

You can have up to five thousand key pairs per region.

The key pair returned to you is available only in the region in which you create it. To create a key pair that is available in all regions, use ImportKeyPair.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "CreateNatGateway": "

Creates a NAT gateway in the specified subnet. A NAT gateway can be used to enable instances in a private subnet to connect to the Internet. This action creates a network interface in the specified subnet with a private IP address from the IP address range of the subnet. For more information, see NAT Gateways in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAcl": "

Creates a network ACL in a VPC. Network ACLs provide an optional layer of security (in addition to security groups) for the instances in your VPC.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkAclEntry": "

Creates an entry (a rule) in a network ACL with the specified rule number. Each network ACL has a set of numbered ingress rules and a separate set of numbered egress rules. When determining whether a packet should be allowed in or out of a subnet associated with the ACL, we process the entries in the ACL according to the rule numbers, in ascending order. Each network ACL has a set of ingress rules and a separate set of egress rules.

We recommend that you leave room between the rule numbers (for example, 100, 110, 120, ...), and not number them one right after the other (for example, 101, 102, 103, ...). This makes it easier to add a rule between existing ones without having to renumber the rules.

After you add an entry, you can't modify it; you must either replace it, or create an entry and delete the old one.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkInterface": "

Creates a network interface in the specified subnet.

For more information about network interfaces, see Elastic Network Interfaces in the Amazon Elastic Compute Cloud User Guide.

", - "CreatePlacementGroup": "

Creates a placement group that you launch cluster instances into. You must give the group a name that's unique within the scope of your account.

For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListing": "

Creates a listing for Amazon EC2 Standard Reserved Instances to be sold in the Reserved Instance Marketplace. You can submit one Standard Reserved Instance listing at a time. To get a list of your Standard Reserved Instances, you can use the DescribeReservedInstances operation.

The Reserved Instance Marketplace matches sellers who want to resell Standard Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

To sell your Standard Reserved Instances, you must first register as a seller in the Reserved Instance Marketplace. After completing the registration process, you can create a Reserved Instance Marketplace listing of some or all of your Standard Reserved Instances, and specify the upfront price to receive for them. Your Standard Reserved Instance listings then become available for purchase. To view the details of your Standard Reserved Instance listing, you can use the DescribeReservedInstancesListings operation.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "CreateRoute": "

Creates a route in a route table within a VPC.

You must specify one of the following targets: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, or network interface.

When determining how to route traffic, we use the route with the most specific match. For example, let's say the traffic is destined for 192.0.2.3, and the route table includes the following two routes:

  • 192.0.2.0/24 (goes to some target A)

  • 192.0.2.0/28 (goes to some target B)

Both routes apply to the traffic destined for 192.0.2.3. However, the second route in the list covers a smaller number of IP addresses and is therefore more specific, so we use that route to determine where to target the traffic.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateRouteTable": "

Creates a route table for the specified VPC. After you create a route table, you can add routes and associate the table with a subnet.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "CreateSecurityGroup": "

Creates a security group.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

EC2-Classic: You can have up to 500 security groups.

EC2-VPC: You can create up to 500 security groups per VPC.

When you create a security group, you specify a friendly name of your choice. You can have a security group for use in EC2-Classic with the same name as a security group for use in a VPC. However, you can't have two security groups for use in EC2-Classic with the same name or two security groups for use in a VPC with the same name.

You have a default security group for use in EC2-Classic and a default security group for use in your VPC. If you don't specify a security group when you launch an instance, the instance is launched into the appropriate default security group. A default security group includes a default rule that grants instances unrestricted network access to each other.

You can add or remove rules from your security groups using AuthorizeSecurityGroupIngress, AuthorizeSecurityGroupEgress, RevokeSecurityGroupIngress, and RevokeSecurityGroupEgress.

", - "CreateSnapshot": "

Creates a snapshot of an EBS volume and stores it in Amazon S3. You can use snapshots for backups, to make copies of EBS volumes, and to save data before shutting down an instance.

When a snapshot is created, any AWS Marketplace product codes that are associated with the source volume are propagated to the snapshot.

You can take a snapshot of an attached volume that is in use. However, snapshots only capture data that has been written to your EBS volume at the time the snapshot command is issued; this may exclude any data that has been cached by any applications or the operating system. If you can pause any file systems on the volume long enough to take a snapshot, your snapshot should be complete. However, if you cannot pause all file writes to the volume, you should unmount the volume from within the instance, issue the snapshot command, and then remount the volume to ensure a consistent and complete snapshot. You may remount and use your volume while the snapshot status is pending.

To create a snapshot for EBS volumes that serve as root devices, you should stop the instance before taking the snapshot.

Snapshots that are taken from encrypted volumes are automatically encrypted. Volumes that are created from encrypted snapshots are also automatically encrypted. Your encrypted volumes and any associated snapshots always remain protected.

For more information, see Amazon Elastic Block Store and Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSpotDatafeedSubscription": "

Creates a data feed for Spot instances, enabling you to view Spot instance usage logs. You can create one data feed per AWS account. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "CreateSubnet": "

Creates a subnet in an existing VPC.

When you create each subnet, you provide the VPC ID and the CIDR block you want for the subnet. After you create a subnet, you can't change its CIDR block. The subnet's CIDR block can be the same as the VPC's CIDR block (assuming you want only a single subnet in the VPC), or a subset of the VPC's CIDR block. If you create more than one subnet in a VPC, the subnets' CIDR blocks must not overlap. The smallest subnet (and VPC) you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses).

AWS reserves both the first four and the last IP address in each subnet's CIDR block. They're not available for use.

If you add more than one subnet to a VPC, they're set up in a star topology with a logical router in the middle.

If you launch an instance in a VPC using an Amazon EBS-backed AMI, the IP address doesn't change if you stop and restart the instance (unlike a similar instance launched outside a VPC, which gets a new IP address when restarted). It's therefore possible to have a subnet with no running instances (they're all stopped), but no remaining IP addresses available.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "CreateTags": "

Adds or overwrites one or more tags for the specified Amazon EC2 resource or resources. Each resource can have a maximum of 50 tags. Each tag consists of a key and optional value. Tag keys must be unique per resource.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide. For more information about creating IAM policies that control users' access to resources based on tags, see Supported Resource-Level Permissions for Amazon EC2 API Actions in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVolume": "

Creates an EBS volume that can be attached to an instance in the same Availability Zone. The volume is created in the regional endpoint that you send the HTTP request to. For more information see Regions and Endpoints.

You can create a new empty volume or restore a volume from an EBS snapshot. Any AWS Marketplace product codes from the snapshot are propagated to the volume.

You can create encrypted volumes with the Encrypted parameter. Encrypted volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are also automatically encrypted. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

For more information, see Creating or Restoring an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpc": "

Creates a VPC with the specified CIDR block.

The smallest VPC you can create uses a /28 netmask (16 IP addresses), and the largest uses a /16 netmask (65,536 IP addresses). To help you decide how big to make your VPC, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

By default, each instance you launch in the VPC has the default DHCP options, which includes only a default DNS server that we provide (AmazonProvidedDNS). For more information about DHCP options, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

You can specify the instance tenancy value for the VPC when you create it. You can't change this value for the VPC after you create it. For more information, see Dedicated Instances in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpcEndpoint": "

Creates a VPC endpoint for a specified AWS service. An endpoint enables you to create a private connection between your VPC and another AWS service in your account. You can specify an endpoint policy to attach to the endpoint that will control access to the service from your VPC. You can also specify the VPC route tables that use the endpoint.

Currently, only endpoints to Amazon S3 are supported.

", - "CreateVpcPeeringConnection": "

Requests a VPC peering connection between two VPCs: a requester VPC that you own and a peer VPC with which to create the connection. The peer VPC can belong to another AWS account. The requester VPC and peer VPC cannot have overlapping CIDR blocks.

The owner of the peer VPC must accept the peering request to activate the peering connection. The VPC peering connection request expires after 7 days, after which it cannot be accepted or rejected.

A CreateVpcPeeringConnection request between VPCs with overlapping CIDR blocks results in the VPC peering connection having a status of failed.

", - "CreateVpnConnection": "

Creates a VPN connection between an existing virtual private gateway and a VPN customer gateway. The only supported connection type is ipsec.1.

The response includes information that you need to give to your network administrator to configure your customer gateway.

We strongly recommend that you use HTTPS when calling this operation because the response contains sensitive cryptographic information for configuring your customer gateway.

If you decide to shut down your VPN connection for any reason and later create a new VPN connection, you must reconfigure your customer gateway with the new information returned from this call.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnConnectionRoute": "

Creates a static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "CreateVpnGateway": "

Creates a virtual private gateway. A virtual private gateway is the endpoint on the VPC side of your VPN connection. You can create a virtual private gateway before creating the VPC itself.

For more information about virtual private gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DeleteCustomerGateway": "

Deletes the specified customer gateway. You must delete the VPN connection before you can delete the customer gateway.

", - "DeleteDhcpOptions": "

Deletes the specified set of DHCP options. You must disassociate the set of DHCP options before you can delete it. You can disassociate the set of DHCP options by associating either a new set of options or the default set of options with the VPC.

", - "DeleteFlowLogs": "

Deletes one or more flow logs.

", - "DeleteInternetGateway": "

Deletes the specified Internet gateway. You must detach the Internet gateway from the VPC before you can delete it.

", - "DeleteKeyPair": "

Deletes the specified key pair, by removing the public key from Amazon EC2.

", - "DeleteNatGateway": "

Deletes the specified NAT gateway. Deleting a NAT gateway disassociates its Elastic IP address, but does not release the address from your account. Deleting a NAT gateway does not delete any NAT gateway routes in your route tables.

", - "DeleteNetworkAcl": "

Deletes the specified network ACL. You can't delete the ACL if it's associated with any subnets. You can't delete the default network ACL.

", - "DeleteNetworkAclEntry": "

Deletes the specified ingress or egress entry (rule) from the specified network ACL.

", - "DeleteNetworkInterface": "

Deletes the specified network interface. You must detach the network interface before you can delete it.

", - "DeletePlacementGroup": "

Deletes the specified placement group. You must terminate all instances in the placement group before you can delete the placement group. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteRoute": "

Deletes the specified route from the specified route table.

", - "DeleteRouteTable": "

Deletes the specified route table. You must disassociate the route table from any subnets before you can delete it. You can't delete the main route table.

", - "DeleteSecurityGroup": "

Deletes a security group.

If you attempt to delete a security group that is associated with an instance, or is referenced by another security group, the operation fails with InvalidGroup.InUse in EC2-Classic or DependencyViolation in EC2-VPC.

", - "DeleteSnapshot": "

Deletes the specified snapshot.

When you make periodic snapshots of a volume, the snapshots are incremental, and only the blocks on the device that have changed since your last snapshot are saved in the new snapshot. When you delete a snapshot, only the data not needed for any other snapshot is removed. So regardless of which prior snapshots have been deleted, all active snapshots will have access to all the information needed to restore the volume.

You cannot delete a snapshot of the root device of an EBS volume used by a registered AMI. You must first de-register the AMI before you can delete the snapshot.

For more information, see Deleting an Amazon EBS Snapshot in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteSpotDatafeedSubscription": "

Deletes the data feed for Spot instances.

", - "DeleteSubnet": "

Deletes the specified subnet. You must terminate all running instances in the subnet before you can delete the subnet.

", - "DeleteTags": "

Deletes the specified set of tags from the specified set of resources. This call is designed to follow a DescribeTags request.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVolume": "

Deletes the specified EBS volume. The volume must be in the available state (not attached to an instance).

The volume may remain in the deleting state for several minutes.

For more information, see Deleting an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DeleteVpc": "

Deletes the specified VPC. You must detach or delete all gateways and resources that are associated with the VPC before you can delete it. For example, you must terminate all instances running in the VPC, delete all security groups associated with the VPC (except the default one), delete all route tables associated with the VPC (except the default one), and so on.

", - "DeleteVpcEndpoints": "

Deletes one or more specified VPC endpoints. Deleting the endpoint also deletes the endpoint routes in the route tables that were associated with the endpoint.

", - "DeleteVpcPeeringConnection": "

Deletes a VPC peering connection. Either the owner of the requester VPC or the owner of the peer VPC can delete the VPC peering connection if it's in the active state. The owner of the requester VPC can delete a VPC peering connection in the pending-acceptance state.

", - "DeleteVpnConnection": "

Deletes the specified VPN connection.

If you're deleting the VPC and its associated components, we recommend that you detach the virtual private gateway from the VPC and delete the VPC before deleting the VPN connection. If you believe that the tunnel credentials for your VPN connection have been compromised, you can delete the VPN connection and create a new one that has new keys, without needing to delete the VPC or virtual private gateway. If you create a new VPN connection, you must reconfigure the customer gateway using the new configuration information returned with the new VPN connection ID.

", - "DeleteVpnConnectionRoute": "

Deletes the specified static route associated with a VPN connection between an existing virtual private gateway and a VPN customer gateway. The static route allows traffic to be routed from the virtual private gateway to the VPN customer gateway.

", - "DeleteVpnGateway": "

Deletes the specified virtual private gateway. We recommend that before you delete a virtual private gateway, you detach it from the VPC and delete the VPN connection. Note that you don't need to delete the virtual private gateway if you plan to delete and recreate the VPN connection between your VPC and your network.

", - "DeregisterImage": "

Deregisters the specified AMI. After you deregister an AMI, it can't be used to launch new instances.

This command does not delete the AMI.

", - "DescribeAccountAttributes": "

Describes attributes of your AWS account. The following are the supported account attributes:

  • supported-platforms: Indicates whether your account can launch instances into EC2-Classic and EC2-VPC, or only into EC2-VPC.

  • default-vpc: The ID of the default VPC for your account, or none.

  • max-instances: The maximum number of On-Demand instances that you can run.

  • vpc-max-security-groups-per-interface: The maximum number of security groups that you can assign to a network interface.

  • max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-Classic.

  • vpc-max-elastic-ips: The maximum number of Elastic IP addresses that you can allocate for use with EC2-VPC.

", - "DescribeAddresses": "

Describes one or more of your Elastic IP addresses.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeAvailabilityZones": "

Describes one or more of the Availability Zones that are available to you. The results include zones only for the region you're currently using. If there is an event impacting an Availability Zone, you can use this request to view the state and any provided message for that Availability Zone.

For more information, see Regions and Availability Zones in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeBundleTasks": "

Describes one or more of your bundling tasks.

Completed bundle tasks are listed for only a limited time. If your bundle task is no longer in the list, you can still register an AMI from it. Just use RegisterImage with the Amazon S3 bucket name and image manifest name you provided to the bundle task.

", - "DescribeClassicLinkInstances": "

Describes one or more of your linked EC2-Classic instances. This request only returns information about EC2-Classic instances linked to a VPC through ClassicLink; you cannot use this request to return information about other instances.

", - "DescribeConversionTasks": "

Describes one or more of your conversion tasks. For more information, see the VM Import/Export User Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DescribeCustomerGateways": "

Describes one or more of your VPN customer gateways.

For more information about VPN customer gateways, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeDhcpOptions": "

Describes one or more of your DHCP options sets.

For more information about DHCP options sets, see DHCP Options Sets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeExportTasks": "

Describes one or more of your export tasks.

", - "DescribeFlowLogs": "

Describes one or more flow logs. To view the information in your flow logs (the log streams for the network interfaces), you must use the CloudWatch Logs console or the CloudWatch Logs API.

", - "DescribeHostReservationOfferings": "

Describes the Dedicated Host Reservations that are available to purchase.

The results describe all the Dedicated Host Reservation offerings, including offerings that may not match the instance family and region of your Dedicated Hosts. When purchasing an offering, ensure that the the instance family and region of the offering matches that of the Dedicated Host/s it will be associated with. For an overview of supported instance types, see Dedicated Hosts Overview in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeHostReservations": "

Describes Dedicated Host Reservations which are associated with Dedicated Hosts in your account.

", - "DescribeHosts": "

Describes one or more of your Dedicated Hosts.

The results describe only the Dedicated Hosts in the region you're currently using. All listed instances consume capacity on your Dedicated Host. Dedicated Hosts that have recently been released will be listed with the state released.

", - "DescribeIdFormat": "

Describes the ID format settings for your resources on a per-region basis, for example, to view which resource types are enabled for longer IDs. This request only returns information about resource types whose ID formats can be modified; it does not return information about other resource types.

The following resource types support longer IDs: instance | reservation | snapshot | volume.

These settings apply to the IAM user who makes the request; they do not apply to the entire AWS account. By default, an IAM user defaults to the same settings as the root user, unless they explicitly override the settings by running the ModifyIdFormat command. Resources created with longer IDs are visible to all IAM users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "DescribeIdentityIdFormat": "

Describes the ID format settings for resources for the specified IAM user, IAM role, or root user. For example, you can view the resource types that are enabled for longer IDs. This request only returns information about resource types whose ID formats can be modified; it does not return information about other resource types. For more information, see Resource IDs in the Amazon Elastic Compute Cloud User Guide.

The following resource types support longer IDs: instance | reservation | snapshot | volume.

These settings apply to the principal specified in the request. They do not apply to the principal that makes the request.

", - "DescribeImageAttribute": "

Describes the specified attribute of the specified AMI. You can specify only one attribute at a time.

", - "DescribeImages": "

Describes one or more of the images (AMIs, AKIs, and ARIs) available to you. Images available to you include public images, private images that you own, and private images owned by other AWS accounts but for which you have explicit launch permissions.

Deregistered images are included in the returned results for an unspecified interval after deregistration.

", - "DescribeImportImageTasks": "

Displays details about an import virtual machine or import snapshot tasks that are already created.

", - "DescribeImportSnapshotTasks": "

Describes your import snapshot tasks.

", - "DescribeInstanceAttribute": "

Describes the specified attribute of the specified instance. You can specify only one attribute at a time. Valid attribute values are: instanceType | kernel | ramdisk | userData | disableApiTermination | instanceInitiatedShutdownBehavior | rootDeviceName | blockDeviceMapping | productCodes | sourceDestCheck | groupSet | ebsOptimized | sriovNetSupport

", - "DescribeInstanceStatus": "

Describes the status of one or more instances. By default, only running instances are described, unless specified otherwise.

Instance status includes the following components:

  • Status checks - Amazon EC2 performs status checks on running EC2 instances to identify hardware and software issues. For more information, see Status Checks for Your Instances and Troubleshooting Instances with Failed Status Checks in the Amazon Elastic Compute Cloud User Guide.

  • Scheduled events - Amazon EC2 can schedule events (such as reboot, stop, or terminate) for your instances related to hardware issues, software updates, or system maintenance. For more information, see Scheduled Events for Your Instances in the Amazon Elastic Compute Cloud User Guide.

  • Instance state - You can manage your instances from the moment you launch them through their termination. For more information, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeInstances": "

Describes one or more of your instances.

If you specify one or more instance IDs, Amazon EC2 returns information for those instances. If you do not specify instance IDs, Amazon EC2 returns information for all relevant instances. If you specify an instance ID that is not valid, an error is returned. If you specify an instance that you do not own, it is not included in the returned results.

Recently terminated instances might appear in the returned results. This interval is usually less than one hour.

If you describe instances in the rare case where an Availability Zone is experiencing a service disruption and you specify instance IDs that are in the affected zone, or do not specify any instance IDs at all, the call fails. If you describe instances and specify only instance IDs that are in an unaffected zone, the call works normally.

", - "DescribeInternetGateways": "

Describes one or more of your Internet gateways.

", - "DescribeKeyPairs": "

Describes one or more of your key pairs.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeMovingAddresses": "

Describes your Elastic IP addresses that are being moved to the EC2-VPC platform, or that are being restored to the EC2-Classic platform. This request does not return information about any other Elastic IP addresses in your account.

", - "DescribeNatGateways": "

Describes one or more of the your NAT gateways.

", - "DescribeNetworkAcls": "

Describes one or more of your network ACLs.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "DescribeNetworkInterfaceAttribute": "

Describes a network interface attribute. You can specify only one attribute at a time.

", - "DescribeNetworkInterfaces": "

Describes one or more of your network interfaces.

", - "DescribePlacementGroups": "

Describes one or more of your placement groups. For more information about placement groups and cluster instances, see Cluster Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribePrefixLists": "

Describes available AWS services in a prefix list format, which includes the prefix list name and prefix list ID of the service and the IP address range for the service. A prefix list ID is required for creating an outbound security group rule that allows traffic from a VPC to access an AWS service through a VPC endpoint.

", - "DescribeRegions": "

Describes one or more regions that are currently available to you.

For a list of the regions supported by Amazon EC2, see Regions and Endpoints.

", - "DescribeReservedInstances": "

Describes one or more of the Reserved Instances that you purchased.

For more information about Reserved Instances, see Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesListings": "

Describes your account's Reserved Instance listings in the Reserved Instance Marketplace.

The Reserved Instance Marketplace matches sellers who want to resell Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

As a seller, you choose to list some or all of your Reserved Instances, and you specify the upfront price to receive for them. Your Reserved Instances are then listed in the Reserved Instance Marketplace and are available for purchase.

As a buyer, you specify the configuration of the Reserved Instance to purchase, and the Marketplace matches what you're searching for with what's available. The Marketplace first sells the lowest priced Reserved Instances to you, and continues to sell available Reserved Instance listings to you until your demand is met. You are charged based on the total price of all of the listings that you purchase.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesModifications": "

Describes the modifications made to your Reserved Instances. If no parameter is specified, information about all your Reserved Instances modification requests is returned. If a modification ID is specified, only information about the specific modification is returned.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeReservedInstancesOfferings": "

Describes Reserved Instance offerings that are available for purchase. With Reserved Instances, you purchase the right to launch instances for a period of time. During that time period, you do not receive insufficient capacity errors, and you pay a lower usage rate than the rate charged for On-Demand instances for the actual time used.

If you have listed your own Reserved Instances for sale in the Reserved Instance Marketplace, they will be excluded from these results. This is to ensure that you do not purchase your own Reserved Instances.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeRouteTables": "

Describes one or more of your route tables.

Each subnet in your VPC must be associated with a route table. If a subnet is not explicitly associated with any route table, it is implicitly associated with the main route table. This command does not return the subnet ID for implicit associations.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "DescribeScheduledInstanceAvailability": "

Finds available schedules that meet the specified criteria.

You can search for an available schedule no more than 3 months in advance. You must meet the minimum required duration of 1,200 hours per year. For example, the minimum daily schedule is 4 hours, the minimum weekly schedule is 24 hours, and the minimum monthly schedule is 100 hours.

After you find a schedule that meets your needs, call PurchaseScheduledInstances to purchase Scheduled Instances with that schedule.

", - "DescribeScheduledInstances": "

Describes one or more of your Scheduled Instances.

", - "DescribeSecurityGroupReferences": "

[EC2-VPC only] Describes the VPCs on the other side of a VPC peering connection that are referencing the security groups you've specified in this request.

", - "DescribeSecurityGroups": "

Describes one or more of your security groups.

A security group is for use with instances either in the EC2-Classic platform or in a specific VPC. For more information, see Amazon EC2 Security Groups in the Amazon Elastic Compute Cloud User Guide and Security Groups for Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeSnapshotAttribute": "

Describes the specified attribute of the specified snapshot. You can specify only one attribute at a time.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSnapshots": "

Describes one or more of the EBS snapshots available to you. Available snapshots include public snapshots available for any AWS account to launch, private snapshots that you own, and private snapshots owned by another AWS account but for which you've been given explicit create volume permissions.

The create volume permissions fall into the following categories:

  • public: The owner of the snapshot granted create volume permissions for the snapshot to the all group. All AWS accounts have create volume permissions for these snapshots.

  • explicit: The owner of the snapshot granted create volume permissions to a specific AWS account.

  • implicit: An AWS account has implicit create volume permissions for all snapshots it owns.

The list of snapshots returned can be modified by specifying snapshot IDs, snapshot owners, or AWS accounts with create volume permissions. If no options are specified, Amazon EC2 returns all snapshots for which you have create volume permissions.

If you specify one or more snapshot IDs, only snapshots that have the specified IDs are returned. If you specify an invalid snapshot ID, an error is returned. If you specify a snapshot ID for which you do not have access, it is not included in the returned results.

If you specify one or more snapshot owners using the OwnerIds option, only snapshots from the specified owners and for which you have access are returned. The results can include the AWS account IDs of the specified owners, amazon for snapshots owned by Amazon, or self for snapshots that you own.

If you specify a list of restorable users, only snapshots with create snapshot permissions for those users are returned. You can specify AWS account IDs (if you own the snapshots), self for snapshots for which you own or have explicit permissions, or all for public snapshots.

If you are describing a long list of snapshots, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeSnapshots request to retrieve the remaining results.

For more information about EBS snapshots, see Amazon EBS Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotDatafeedSubscription": "

Describes the data feed for Spot instances. For more information, see Spot Instance Data Feed in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeSpotFleetInstances": "

Describes the running instances for the specified Spot fleet.

", - "DescribeSpotFleetRequestHistory": "

Describes the events for the specified Spot fleet request during the specified time.

Spot fleet events are delayed by up to 30 seconds before they can be described. This ensures that you can query by the last evaluated time and not miss a recorded event.

", - "DescribeSpotFleetRequests": "

Describes your Spot fleet requests.

Spot fleet requests are deleted 48 hours after they are canceled and their instances are terminated.

", - "DescribeSpotInstanceRequests": "

Describes the Spot instance requests that belong to your account. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

You can use DescribeSpotInstanceRequests to find a running Spot instance by examining the response. If the status of the Spot instance is fulfilled, the instance ID appears in the response and contains the identifier of the instance. Alternatively, you can use DescribeInstances with a filter to look for instances where the instance lifecycle is spot.

Spot instance requests are deleted 4 hours after they are canceled and their instances are terminated.

", - "DescribeSpotPriceHistory": "

Describes the Spot price history. The prices returned are listed in chronological order, from the oldest to the most recent, for up to the past 90 days. For more information, see Spot Instance Pricing History in the Amazon Elastic Compute Cloud User Guide.

When you specify a start and end time, this operation returns the prices of the instance types within the time range that you specified and the time when the price changed. The price is valid within the time period that you specified; the response merely indicates the last time that the price changed.

", - "DescribeStaleSecurityGroups": "

[EC2-VPC only] Describes the stale security group rules for security groups in a specified VPC. Rules are stale when they reference a deleted security group in a peer VPC, or a security group in a peer VPC for which the VPC peering connection has been deleted.

", - "DescribeSubnets": "

Describes one or more of your subnets.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", - "DescribeTags": "

Describes one or more of the tags for your EC2 resources.

For more information about tags, see Tagging Your Resources in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeAttribute": "

Describes the specified attribute of the specified volume. You can specify only one attribute at a time.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVolumeStatus": "

Describes the status of the specified volumes. Volume status provides the result of the checks performed on your volumes to determine events that can impair the performance of your volumes. The performance of a volume can be affected if an issue occurs on the volume's underlying host. If the volume's underlying host experiences a power outage or system issue, after the system is restored, there could be data inconsistencies on the volume. Volume events notify you if this occurs. Volume actions notify you if any action needs to be taken in response to the event.

The DescribeVolumeStatus operation provides the following information about the specified volumes:

Status: Reflects the current status of the volume. The possible values are ok, impaired , warning, or insufficient-data. If all checks pass, the overall status of the volume is ok. If the check fails, the overall status is impaired. If the status is insufficient-data, then the checks may still be taking place on your volume at the time. We recommend that you retry the request. For more information on volume status, see Monitoring the Status of Your Volumes.

Events: Reflect the cause of a volume status and may require you to take action. For example, if your volume returns an impaired status, then the volume event might be potential-data-inconsistency. This means that your volume has been affected by an issue with the underlying host, has all I/O operations disabled, and may have inconsistent data.

Actions: Reflect the actions you may have to take in response to an event. For example, if the status of the volume is impaired and the volume event shows potential-data-inconsistency, then the action shows enable-volume-io. This means that you may want to enable the I/O operations for the volume by calling the EnableVolumeIO action and then check the volume for data consistency.

Volume status is based on the volume status checks, and does not reflect the volume state. Therefore, volume status does not indicate volumes in the error state (for example, when a volume is incapable of accepting I/O.)

", - "DescribeVolumes": "

Describes the specified EBS volumes.

If you are describing a long list of volumes, you can paginate the output to make the list more manageable. The MaxResults parameter sets the maximum number of results returned in a single page. If the list of results exceeds your MaxResults value, then that number of results is returned along with a NextToken value that can be passed to a subsequent DescribeVolumes request to retrieve the remaining results.

For more information about EBS volumes, see Amazon EBS Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcAttribute": "

Describes the specified attribute of the specified VPC. You can specify only one attribute at a time.

", - "DescribeVpcClassicLink": "

Describes the ClassicLink status of one or more VPCs.

", - "DescribeVpcClassicLinkDnsSupport": "

Describes the ClassicLink DNS support status of one or more VPCs. If enabled, the DNS hostname of a linked EC2-Classic instance resolves to its private IP address when addressed from an instance in the VPC to which it's linked. Similarly, the DNS hostname of an instance in a VPC resolves to its private IP address when addressed from a linked EC2-Classic instance. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "DescribeVpcEndpointServices": "

Describes all supported AWS services that can be specified when creating a VPC endpoint.

", - "DescribeVpcEndpoints": "

Describes one or more of your VPC endpoints.

", - "DescribeVpcPeeringConnections": "

Describes one or more of your VPC peering connections.

", - "DescribeVpcs": "

Describes one or more of your VPCs.

", - "DescribeVpnConnections": "

Describes one or more of your VPN connections.

For more information about VPN connections, see Adding a Hardware Virtual Private Gateway to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DescribeVpnGateways": "

Describes one or more of your virtual private gateways.

For more information about virtual private gateways, see Adding an IPsec Hardware VPN to Your VPC in the Amazon Virtual Private Cloud User Guide.

", - "DetachClassicLinkVpc": "

Unlinks (detaches) a linked EC2-Classic instance from a VPC. After the instance has been unlinked, the VPC security groups are no longer associated with it. An instance is automatically unlinked from a VPC when it's stopped.

", - "DetachInternetGateway": "

Detaches an Internet gateway from a VPC, disabling connectivity between the Internet and the VPC. The VPC must not contain any running instances with Elastic IP addresses.

", - "DetachNetworkInterface": "

Detaches a network interface from an instance.

", - "DetachVolume": "

Detaches an EBS volume from an instance. Make sure to unmount any file systems on the device within your operating system before detaching the volume. Failure to do so can result in the volume becoming stuck in the busy state while detaching. If this happens, detachment can be delayed indefinitely until you unmount the volume, force detachment, reboot the instance, or all three. If an EBS volume is the root device of an instance, it can't be detached while the instance is running. To detach the root volume, stop the instance first.

When a volume with an AWS Marketplace product code is detached from an instance, the product code is no longer associated with the instance.

For more information, see Detaching an Amazon EBS Volume in the Amazon Elastic Compute Cloud User Guide.

", - "DetachVpnGateway": "

Detaches a virtual private gateway from a VPC. You do this if you're planning to turn off the VPC and not use it anymore. You can confirm a virtual private gateway has been completely detached from a VPC by describing the virtual private gateway (any attachments to the virtual private gateway are also described).

You must wait for the attachment's state to switch to detached before you can delete the VPC or attach a different VPC to the virtual private gateway.

", - "DisableVgwRoutePropagation": "

Disables a virtual private gateway (VGW) from propagating routes to a specified route table of a VPC.

", - "DisableVpcClassicLink": "

Disables ClassicLink for a VPC. You cannot disable ClassicLink for a VPC that has EC2-Classic instances linked to it.

", - "DisableVpcClassicLinkDnsSupport": "

Disables ClassicLink DNS support for a VPC. If disabled, DNS hostnames resolve to public IP addresses when addressed between a linked EC2-Classic instance and instances in the VPC to which it's linked. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "DisassociateAddress": "

Disassociates an Elastic IP address from the instance or network interface it's associated with.

An Elastic IP address is for use in either the EC2-Classic platform or in a VPC. For more information, see Elastic IP Addresses in the Amazon Elastic Compute Cloud User Guide.

This is an idempotent operation. If you perform the operation more than once, Amazon EC2 doesn't return an error.

", - "DisassociateRouteTable": "

Disassociates a subnet from a route table.

After you perform this action, the subnet no longer uses the routes in the route table. Instead, it uses the routes in the VPC's main route table. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "EnableVgwRoutePropagation": "

Enables a virtual private gateway (VGW) to propagate routes to the specified route table of a VPC.

", - "EnableVolumeIO": "

Enables I/O operations for a volume that had I/O operations disabled because the data on the volume was potentially inconsistent.

", - "EnableVpcClassicLink": "

Enables a VPC for ClassicLink. You can then link EC2-Classic instances to your ClassicLink-enabled VPC to allow communication over private IP addresses. You cannot enable your VPC for ClassicLink if any of your VPC's route tables have existing routes for address ranges within the 10.0.0.0/8 IP address range, excluding local routes for VPCs in the 10.0.0.0/16 and 10.1.0.0/16 IP address ranges. For more information, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "EnableVpcClassicLinkDnsSupport": "

Enables a VPC to support DNS hostname resolution for ClassicLink. If enabled, the DNS hostname of a linked EC2-Classic instance resolves to its private IP address when addressed from an instance in the VPC to which it's linked. Similarly, the DNS hostname of an instance in a VPC resolves to its private IP address when addressed from a linked EC2-Classic instance. For more information about ClassicLink, see ClassicLink in the Amazon Elastic Compute Cloud User Guide.

", - "GetConsoleOutput": "

Gets the console output for the specified instance.

Instances do not have a physical monitor through which you can view their console output. They also lack physical controls that allow you to power up, reboot, or shut them down. To allow these actions, we provide them through the Amazon EC2 API and command line interface.

Instance console output is buffered and posted shortly after instance boot, reboot, and termination. Amazon EC2 preserves the most recent 64 KB output which is available for at least one hour after the most recent post.

For Linux instances, the instance console output displays the exact console output that would normally be displayed on a physical monitor attached to a computer. This output is buffered because the instance produces it and then posts it to a store where the instance's owner can retrieve it.

For Windows instances, the instance console output includes output from the EC2Config service.

", - "GetConsoleScreenshot": "

Retrieve a JPG-format screenshot of a running instance to help with troubleshooting.

The returned content is Base64-encoded.

", - "GetHostReservationPurchasePreview": "

Preview a reservation purchase with configurations that match those of your Dedicated Host. You must have active Dedicated Hosts in your account before you purchase a reservation.

This is a preview of the PurchaseHostReservation action and does not result in the offering being purchased.

", - "GetPasswordData": "

Retrieves the encrypted administrator password for an instance running Windows.

The Windows password is generated at boot if the EC2Config service plugin, Ec2SetPassword, is enabled. This usually only happens the first time an AMI is launched, and then Ec2SetPassword is automatically disabled. The password is not generated for rebundled AMIs unless Ec2SetPassword is enabled before bundling.

The password is encrypted using the key pair that you specified when you launched the instance. You must provide the corresponding key pair file.

Password generation and encryption takes a few moments. We recommend that you wait up to 15 minutes after launching an instance before trying to retrieve the generated password.

", - "GetReservedInstancesExchangeQuote": "

Returns details about the values and term of your specified Convertible Reserved Instances. When an offering ID is specified it returns information about whether the exchange is valid and can be performed.

", - "ImportImage": "

Import single or multi-volume disk images or EBS snapshots into an Amazon Machine Image (AMI). For more information, see Importing a VM as an Image Using VM Import/Export in the VM Import/Export User Guide.

", - "ImportInstance": "

Creates an import instance task using metadata from the specified disk image. ImportInstance only supports single-volume VMs. To import multi-volume VMs, use ImportImage. For more information, see Importing a Virtual Machine Using the Amazon EC2 CLI.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "ImportKeyPair": "

Imports the public key from an RSA key pair that you created with a third-party tool. Compare this with CreateKeyPair, in which AWS creates the key pair and gives the keys to you (AWS keeps a copy of the public key). With ImportKeyPair, you create the key pair and give AWS just the public key. The private key is never transferred between you and AWS.

For more information about key pairs, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

", - "ImportSnapshot": "

Imports a disk into an EBS snapshot.

", - "ImportVolume": "

Creates an import volume task using metadata from the specified disk image.For more information, see Importing Disks to Amazon EBS.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "ModifyHosts": "

Modify the auto-placement setting of a Dedicated Host. When auto-placement is enabled, AWS will place instances that you launch with a tenancy of host, but without targeting a specific host ID, onto any available Dedicated Host in your account which has auto-placement enabled. When auto-placement is disabled, you need to provide a host ID if you want the instance to launch onto a specific host. If no host ID is provided, the instance will be launched onto a suitable host which has auto-placement enabled.

", - "ModifyIdFormat": "

Modifies the ID format for the specified resource on a per-region basis. You can specify that resources should receive longer IDs (17-character IDs) when they are created. The following resource types support longer IDs: instance | reservation | snapshot | volume.

This setting applies to the IAM user who makes the request; it does not apply to the entire AWS account. By default, an IAM user defaults to the same settings as the root user. If you're using this action as the root user, then these settings apply to the entire account, unless an IAM user explicitly overrides these settings for themselves. For more information, see Resource IDs in the Amazon Elastic Compute Cloud User Guide.

Resources created with longer IDs are visible to all IAM roles and users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "ModifyIdentityIdFormat": "

Modifies the ID format of a resource for a specified IAM user, IAM role, or the root user for an account; or all IAM users, IAM roles, and the root user for an account. You can specify that resources should receive longer IDs (17-character IDs) when they are created.

The following resource types support longer IDs: instance | reservation | snapshot | volume. For more information, see Resource IDs in the Amazon Elastic Compute Cloud User Guide.

This setting applies to the principal specified in the request; it does not apply to the principal that makes the request.

Resources created with longer IDs are visible to all IAM roles and users, regardless of these settings and provided that they have permission to use the relevant Describe command for the resource type.

", - "ModifyImageAttribute": "

Modifies the specified attribute of the specified AMI. You can specify only one attribute at a time.

AWS Marketplace product codes cannot be modified. Images with an AWS Marketplace product code cannot be made public.

The SriovNetSupport enhanced networking attribute cannot be changed using this command. Instead, enable SriovNetSupport on an instance and create an AMI from the instance. This will result in an image with SriovNetSupport enabled.

", - "ModifyInstanceAttribute": "

Modifies the specified attribute of the specified instance. You can specify only one attribute at a time.

To modify some attributes, the instance must be stopped. For more information, see Modifying Attributes of a Stopped Instance in the Amazon Elastic Compute Cloud User Guide.

", - "ModifyInstancePlacement": "

Set the instance affinity value for a specific stopped instance and modify the instance tenancy setting.

Instance affinity is disabled by default. When instance affinity is host and it is not associated with a specific Dedicated Host, the next time it is launched it will automatically be associated with the host it lands on. This relationship will persist if the instance is stopped/started, or rebooted.

You can modify the host ID associated with a stopped instance. If a stopped instance has a new host ID association, the instance will target that host when restarted.

You can modify the tenancy of a stopped instance with a tenancy of host or dedicated.

Affinity, hostID, and tenancy are not required parameters, but at least one of them must be specified in the request. Affinity and tenancy can be modified in the same request, but tenancy can only be modified on instances that are stopped.

", - "ModifyNetworkInterfaceAttribute": "

Modifies the specified network interface attribute. You can specify only one attribute at a time.

", - "ModifyReservedInstances": "

Modifies the Availability Zone, instance count, instance type, or network platform (EC2-Classic or EC2-VPC) of your Standard Reserved Instances. The Reserved Instances to be modified must be identical, except for Availability Zone, network platform, and instance type.

For more information, see Modifying Reserved Instances in the Amazon Elastic Compute Cloud User Guide.

", - "ModifySnapshotAttribute": "

Adds or removes permission settings for the specified snapshot. You may add or remove specified AWS account IDs from a snapshot's list of create volume permissions, but you cannot do both in a single API call. If you need to both add and remove account IDs for a snapshot, you must use multiple API calls.

Encrypted snapshots and snapshots with AWS Marketplace product codes cannot be made public. Snapshots encrypted with your default CMK cannot be shared with other accounts.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "ModifySpotFleetRequest": "

Modifies the specified Spot fleet request.

While the Spot fleet request is being modified, it is in the modifying state.

To scale up your Spot fleet, increase its target capacity. The Spot fleet launches the additional Spot instances according to the allocation strategy for the Spot fleet request. If the allocation strategy is lowestPrice, the Spot fleet launches instances using the Spot pool with the lowest price. If the allocation strategy is diversified, the Spot fleet distributes the instances across the Spot pools.

To scale down your Spot fleet, decrease its target capacity. First, the Spot fleet cancels any open bids that exceed the new target capacity. You can request that the Spot fleet terminate Spot instances until the size of the fleet no longer exceeds the new target capacity. If the allocation strategy is lowestPrice, the Spot fleet terminates the instances with the highest price per unit. If the allocation strategy is diversified, the Spot fleet terminates instances across the Spot pools. Alternatively, you can request that the Spot fleet keep the fleet at its current size, but not replace any Spot instances that are interrupted or that you terminate manually.

", - "ModifySubnetAttribute": "

Modifies a subnet attribute.

", - "ModifyVolumeAttribute": "

Modifies a volume attribute.

By default, all I/O operations for the volume are suspended when the data on the volume is determined to be potentially inconsistent, to prevent undetectable, latent data corruption. The I/O access to the volume can be resumed by first enabling I/O access and then checking the data consistency on your volume.

You can change the default behavior to resume I/O operations. We recommend that you change this only for boot volumes or for volumes that are stateless or disposable.

", - "ModifyVpcAttribute": "

Modifies the specified attribute of the specified VPC.

", - "ModifyVpcEndpoint": "

Modifies attributes of a specified VPC endpoint. You can modify the policy associated with the endpoint, and you can add and remove route tables associated with the endpoint.

", - "ModifyVpcPeeringConnectionOptions": "

Modifies the VPC peering connection options on one side of a VPC peering connection. You can do the following:

  • Enable/disable communication over the peering connection between an EC2-Classic instance that's linked to your VPC (using ClassicLink) and instances in the peer VPC.

  • Enable/disable communication over the peering connection between instances in your VPC and an EC2-Classic instance that's linked to the peer VPC.

  • Enable/disable a local VPC to resolve public DNS hostnames to private IP addresses when queried from instances in the peer VPC.

If the peered VPCs are in different accounts, each owner must initiate a separate request to modify the peering connection options, depending on whether their VPC was the requester or accepter for the VPC peering connection. If the peered VPCs are in the same account, you can modify the requester and accepter options in the same request. To confirm which VPC is the accepter and requester for a VPC peering connection, use the DescribeVpcPeeringConnections command.

", - "MonitorInstances": "

Enables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

", - "MoveAddressToVpc": "

Moves an Elastic IP address from the EC2-Classic platform to the EC2-VPC platform. The Elastic IP address must be allocated to your account for more than 24 hours, and it must not be associated with an instance. After the Elastic IP address is moved, it is no longer available for use in the EC2-Classic platform, unless you move it back using the RestoreAddressToClassic request. You cannot move an Elastic IP address that was originally allocated for use in the EC2-VPC platform to the EC2-Classic platform.

", - "PurchaseHostReservation": "

Purchase a reservation with configurations that match those of your Dedicated Host. You must have active Dedicated Hosts in your account before you purchase a reservation. This action results in the specified reservation being purchased and charged to your account.

", - "PurchaseReservedInstancesOffering": "

Purchases a Reserved Instance for use with your account. With Reserved Instances, you pay a lower hourly rate compared to On-Demand instance pricing.

Use DescribeReservedInstancesOfferings to get a list of Reserved Instance offerings that match your specifications. After you've purchased a Reserved Instance, you can check for your new Reserved Instance with DescribeReservedInstances.

For more information, see Reserved Instances and Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", - "PurchaseScheduledInstances": "

Purchases one or more Scheduled Instances with the specified schedule.

Scheduled Instances enable you to purchase Amazon EC2 compute capacity by the hour for a one-year term. Before you can purchase a Scheduled Instance, you must call DescribeScheduledInstanceAvailability to check for available schedules and obtain a purchase token. After you purchase a Scheduled Instance, you must call RunScheduledInstances during each scheduled time period.

After you purchase a Scheduled Instance, you can't cancel, modify, or resell your purchase.

", - "RebootInstances": "

Requests a reboot of one or more instances. This operation is asynchronous; it only queues a request to reboot the specified instances. The operation succeeds if the instances are valid and belong to you. Requests to reboot terminated instances are ignored.

If an instance does not cleanly shut down within four minutes, Amazon EC2 performs a hard reboot.

For more information about troubleshooting, see Getting Console Output and Rebooting Instances in the Amazon Elastic Compute Cloud User Guide.

", - "RegisterImage": "

Registers an AMI. When you're creating an AMI, this is the final step you must complete before you can launch an instance from the AMI. For more information about creating AMIs, see Creating Your Own AMIs in the Amazon Elastic Compute Cloud User Guide.

For Amazon EBS-backed instances, CreateImage creates and registers the AMI in a single request, so you don't have to register the AMI yourself.

You can also use RegisterImage to create an Amazon EBS-backed Linux AMI from a snapshot of a root device volume. For more information, see Launching an Instance from a Snapshot in the Amazon Elastic Compute Cloud User Guide.

Some Linux distributions, such as Red Hat Enterprise Linux (RHEL) and SUSE Linux Enterprise Server (SLES), use the EC2 billingProduct code associated with an AMI to verify subscription status for package updates. Creating an AMI from an EBS snapshot does not maintain this billing code, and subsequent instances launched from such an AMI will not be able to connect to package update infrastructure.

Similarly, although you can create a Windows AMI from a snapshot, you can't successfully launch an instance from the AMI.

To create Windows AMIs or to create AMIs for Linux operating systems that must retain AMI billing codes to work properly, see CreateImage.

If needed, you can deregister an AMI at any time. Any modifications you make to an AMI backed by an instance store volume invalidates its registration. If you make changes to an image, deregister the previous image and register the new image.

You can't register an image where a secondary (non-root) snapshot has AWS Marketplace product codes.

", - "RejectVpcPeeringConnection": "

Rejects a VPC peering connection request. The VPC peering connection must be in the pending-acceptance state. Use the DescribeVpcPeeringConnections request to view your outstanding VPC peering connection requests. To delete an active VPC peering connection, or to delete a VPC peering connection request that you initiated, use DeleteVpcPeeringConnection.

", - "ReleaseAddress": "

Releases the specified Elastic IP address.

After releasing an Elastic IP address, it is released to the IP address pool and might be unavailable to you. Be sure to update your DNS records and any servers or devices that communicate with the address. If you attempt to release an Elastic IP address that you already released, you'll get an AuthFailure error if the address is already allocated to another AWS account.

[EC2-Classic, default VPC] Releasing an Elastic IP address automatically disassociates it from any instance that it's associated with. To disassociate an Elastic IP address without releasing it, use DisassociateAddress.

[Nondefault VPC] You must use DisassociateAddress to disassociate the Elastic IP address before you try to release it. Otherwise, Amazon EC2 returns an error (InvalidIPAddress.InUse).

", - "ReleaseHosts": "

When you no longer want to use an On-Demand Dedicated Host it can be released. On-Demand billing is stopped and the host goes into released state. The host ID of Dedicated Hosts that have been released can no longer be specified in another request, e.g., ModifyHosts. You must stop or terminate all instances on a host before it can be released.

When Dedicated Hosts are released, it make take some time for them to stop counting toward your limit and you may receive capacity errors when trying to allocate new Dedicated hosts. Try waiting a few minutes, and then try again.

Released hosts will still appear in a DescribeHosts response.

", - "ReplaceNetworkAclAssociation": "

Changes which network ACL a subnet is associated with. By default when you create a subnet, it's automatically associated with the default network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceNetworkAclEntry": "

Replaces an entry (rule) in a network ACL. For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRoute": "

Replaces an existing route within a route table in a VPC. You must provide only one of the following: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, or network interface.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", - "ReplaceRouteTableAssociation": "

Changes the route table associated with a given subnet in a VPC. After the operation completes, the subnet uses the routes in the new route table it's associated with. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

You can also use ReplaceRouteTableAssociation to change which table is the main route table in the VPC. You just specify the main route table's association ID and the route table to be the new main route table.

", - "ReportInstanceStatus": "

Submits feedback about the status of an instance. The instance must be in the running state. If your experience with the instance differs from the instance status returned by DescribeInstanceStatus, use ReportInstanceStatus to report your experience with the instance. Amazon EC2 collects this information to improve the accuracy of status checks.

Use of this action does not change the value returned by DescribeInstanceStatus.

", - "RequestSpotFleet": "

Creates a Spot fleet request.

You can submit a single request that includes multiple launch specifications that vary by instance type, AMI, Availability Zone, or subnet.

By default, the Spot fleet requests Spot instances in the Spot pool where the price per unit is the lowest. Each launch specification can include its own instance weighting that reflects the value of the instance type to your application workload.

Alternatively, you can specify that the Spot fleet distribute the target capacity across the Spot pools included in its launch specifications. By ensuring that the Spot instances in your Spot fleet are in different Spot pools, you can improve the availability of your fleet.

For more information, see Spot Fleet Requests in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstances": "

Creates a Spot instance request. Spot instances are instances that Amazon EC2 launches when the bid price that you specify exceeds the current Spot price. Amazon EC2 periodically sets the Spot price based on available Spot Instance capacity and current Spot instance requests. For more information, see Spot Instance Requests in the Amazon Elastic Compute Cloud User Guide.

", - "ResetImageAttribute": "

Resets an attribute of an AMI to its default value.

The productCodes attribute can't be reset.

", - "ResetInstanceAttribute": "

Resets an attribute of an instance to its default value. To reset the kernel or ramdisk, the instance must be in a stopped state. To reset the sourceDestCheck, the instance can be either running or stopped.

The sourceDestCheck attribute controls whether source/destination checking is enabled. The default value is true, which means checking is enabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ResetNetworkInterfaceAttribute": "

Resets a network interface attribute. You can specify only one attribute at a time.

", - "ResetSnapshotAttribute": "

Resets permission settings for the specified snapshot.

For more information on modifying snapshot permissions, see Sharing Snapshots in the Amazon Elastic Compute Cloud User Guide.

", - "RestoreAddressToClassic": "

Restores an Elastic IP address that was previously moved to the EC2-VPC platform back to the EC2-Classic platform. You cannot move an Elastic IP address that was originally allocated for use in EC2-VPC. The Elastic IP address must not be associated with an instance or network interface.

", - "RevokeSecurityGroupEgress": "

[EC2-VPC only] Removes one or more egress rules from a security group for EC2-VPC. This action doesn't apply to security groups for use in EC2-Classic. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be revoked.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RevokeSecurityGroupIngress": "

Removes one or more ingress rules from a security group. The values that you specify in the revoke request (for example, ports) must match the existing rule's values for the rule to be removed.

Each rule consists of the protocol and the CIDR range or source security group. For the TCP and UDP protocols, you must also specify the destination port or range of ports. For the ICMP protocol, you must also specify the ICMP type and code.

Rule changes are propagated to instances within the security group as quickly as possible. However, a small delay might occur.

", - "RunInstances": "

Launches the specified number of instances using an AMI for which you have permissions.

When you launch an instance, it enters the pending state. After the instance is ready for you, it enters the running state. To check the state of your instance, call DescribeInstances.

To ensure faster instance launches, break up large requests into smaller batches. For example, create five separate launch requests for 100 instances each instead of one launch request for 500 instances.

To tag your instance, ensure that it is running as CreateTags requires a resource ID. For more information about tagging, see Tagging Your Amazon EC2 Resources.

If you don't specify a security group when launching an instance, Amazon EC2 uses the default security group. For more information, see Security Groups in the Amazon Elastic Compute Cloud User Guide.

[EC2-VPC only accounts] If you don't specify a subnet in the request, we choose a default subnet from your default VPC for you.

[EC2-Classic accounts] If you're launching into EC2-Classic and you don't specify an Availability Zone, we choose one for you.

Linux instances have access to the public key of the key pair at boot. You can use this key to provide secure access to the instance. Amazon EC2 public images use this feature to provide secure access without passwords. For more information, see Key Pairs in the Amazon Elastic Compute Cloud User Guide.

You can provide optional user data when launching an instance. For more information, see Instance Metadata in the Amazon Elastic Compute Cloud User Guide.

If any of the AMIs have a product code attached for which the user has not subscribed, RunInstances fails.

Some instance types can only be launched into a VPC. If you do not have a default VPC, or if you do not specify a subnet ID in the request, RunInstances fails. For more information, see Instance Types Available Only in a VPC.

For more information about troubleshooting, see What To Do If An Instance Immediately Terminates, and Troubleshooting Connecting to Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "RunScheduledInstances": "

Launches the specified Scheduled Instances.

Before you can launch a Scheduled Instance, you must purchase it and obtain an identifier using PurchaseScheduledInstances.

You must launch a Scheduled Instance during its scheduled time period. You can't stop or reboot a Scheduled Instance, but you can terminate it as needed. If you terminate a Scheduled Instance before the current scheduled time period ends, you can launch it again after a few minutes. For more information, see Scheduled Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StartInstances": "

Starts an Amazon EBS-backed AMI that you've previously stopped.

Instances that use Amazon EBS volumes as their root devices can be quickly stopped and started. When an instance is stopped, the compute resources are released and you are not billed for hourly instance usage. However, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. You can restart your instance at any time. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Performing this operation on an instance that uses an instance store as its root device returns an error.

For more information, see Stopping Instances in the Amazon Elastic Compute Cloud User Guide.

", - "StopInstances": "

Stops an Amazon EBS-backed instance.

We don't charge hourly usage for a stopped instance, or data transfer fees; however, your root partition Amazon EBS volume remains, continues to persist your data, and you are charged for Amazon EBS volume usage. Each time you transition an instance from stopped to started, Amazon EC2 charges a full instance hour, even if transitions happen multiple times within a single hour.

You can't start or stop Spot instances, and you can't stop instance store-backed instances.

When you stop an instance, we shut it down. You can restart your instance at any time. Before stopping an instance, make sure it is in a state from which it can be restarted. Stopping an instance does not preserve data stored in RAM.

Stopping an instance is different to rebooting or terminating it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, the root device and any other devices attached during the instance launch are automatically deleted. For more information about the differences between rebooting, stopping, and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

When you stop an instance, we attempt to shut it down forcibly after a short while. If your instance appears stuck in the stopping state after a period of time, there may be an issue with the underlying host computer. For more information, see Troubleshooting Stopping Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "TerminateInstances": "

Shuts down one or more instances. This operation is idempotent; if you terminate an instance more than once, each call succeeds.

If you specify multiple instances and the request fails (for example, because of a single incorrect instance ID), none of the instances are terminated.

Terminated instances remain visible after termination (for approximately one hour).

By default, Amazon EC2 deletes all EBS volumes that were attached when the instance launched. Volumes attached after instance launch continue running.

You can stop, start, and terminate EBS-backed instances. You can only terminate instance store-backed instances. What happens to an instance differs if you stop it or terminate it. For example, when you stop an instance, the root device and any other devices attached to the instance persist. When you terminate an instance, any attached EBS volumes with the DeleteOnTermination block device mapping parameter set to true are automatically deleted. For more information about the differences between stopping and terminating instances, see Instance Lifecycle in the Amazon Elastic Compute Cloud User Guide.

For more information about troubleshooting, see Troubleshooting Terminating Your Instance in the Amazon Elastic Compute Cloud User Guide.

", - "UnassignPrivateIpAddresses": "

Unassigns one or more secondary private IP addresses from a network interface.

", - "UnmonitorInstances": "

Disables monitoring for a running instance. For more information about monitoring instances, see Monitoring Your Instances and Volumes in the Amazon Elastic Compute Cloud User Guide.

" - }, - "shapes": { - "AcceptReservedInstancesExchangeQuoteRequest": { - "base": "

Contains the parameters for accepting the quote.

", - "refs": { - } - }, - "AcceptReservedInstancesExchangeQuoteResult": { - "base": "

The result of the exchange and whether it was successful.

", - "refs": { - } - }, - "AcceptVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for AcceptVpcPeeringConnection.

", - "refs": { - } - }, - "AcceptVpcPeeringConnectionResult": { - "base": "

Contains the output of AcceptVpcPeeringConnection.

", - "refs": { - } - }, - "AccountAttribute": { - "base": "

Describes an account attribute.

", - "refs": { - "AccountAttributeList$member": null - } - }, - "AccountAttributeList": { - "base": null, - "refs": { - "DescribeAccountAttributesResult$AccountAttributes": "

Information about one or more account attributes.

" - } - }, - "AccountAttributeName": { - "base": null, - "refs": { - "AccountAttributeNameStringList$member": null - } - }, - "AccountAttributeNameStringList": { - "base": null, - "refs": { - "DescribeAccountAttributesRequest$AttributeNames": "

One or more account attribute names.

" - } - }, - "AccountAttributeValue": { - "base": "

Describes a value of an account attribute.

", - "refs": { - "AccountAttributeValueList$member": null - } - }, - "AccountAttributeValueList": { - "base": null, - "refs": { - "AccountAttribute$AttributeValues": "

One or more values for the account attribute.

" - } - }, - "ActiveInstance": { - "base": "

Describes a running instance in a Spot fleet.

", - "refs": { - "ActiveInstanceSet$member": null - } - }, - "ActiveInstanceSet": { - "base": null, - "refs": { - "DescribeSpotFleetInstancesResponse$ActiveInstances": "

The running instances. Note that this list is refreshed periodically and might be out of date.

" - } - }, - "ActivityStatus": { - "base": null, - "refs": { - "SpotFleetRequestConfig$ActivityStatus": "

The progress of the Spot fleet request. If there is an error, the status is error. After all bids are placed, the status is pending_fulfillment. If the size of the fleet is equal to or greater than its target capacity, the status is fulfilled. If the size of the fleet is decreased, the status is pending_termination while Spot instances are terminating.

" - } - }, - "Address": { - "base": "

Describes an Elastic IP address.

", - "refs": { - "AddressList$member": null - } - }, - "AddressList": { - "base": null, - "refs": { - "DescribeAddressesResult$Addresses": "

Information about one or more Elastic IP addresses.

" - } - }, - "Affinity": { - "base": null, - "refs": { - "ModifyInstancePlacementRequest$Affinity": "

The new affinity setting for the instance.

" - } - }, - "AllocateAddressRequest": { - "base": "

Contains the parameters for AllocateAddress.

", - "refs": { - } - }, - "AllocateAddressResult": { - "base": "

Contains the output of AllocateAddress.

", - "refs": { - } - }, - "AllocateHostsRequest": { - "base": "

Contains the parameters for AllocateHosts.

", - "refs": { - } - }, - "AllocateHostsResult": { - "base": "

Contains the output of AllocateHosts.

", - "refs": { - } - }, - "AllocationIdList": { - "base": null, - "refs": { - "DescribeAddressesRequest$AllocationIds": "

[EC2-VPC] One or more allocation IDs.

Default: Describes all your Elastic IP addresses.

" - } - }, - "AllocationState": { - "base": null, - "refs": { - "Host$State": "

The Dedicated Host's state.

" - } - }, - "AllocationStrategy": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$AllocationStrategy": "

Indicates how to allocate the target capacity across the Spot pools specified by the Spot fleet request. The default is lowestPrice.

" - } - }, - "ArchitectureValues": { - "base": null, - "refs": { - "Image$Architecture": "

The architecture of the image.

", - "ImportInstanceLaunchSpecification$Architecture": "

The architecture of the instance.

", - "Instance$Architecture": "

The architecture of the image.

", - "RegisterImageRequest$Architecture": "

The architecture of the AMI.

Default: For Amazon EBS-backed AMIs, i386. For instance store-backed AMIs, the architecture specified in the manifest file.

" - } - }, - "AssignPrivateIpAddressesRequest": { - "base": "

Contains the parameters for AssignPrivateIpAddresses.

", - "refs": { - } - }, - "AssociateAddressRequest": { - "base": "

Contains the parameters for AssociateAddress.

", - "refs": { - } - }, - "AssociateAddressResult": { - "base": "

Contains the output of AssociateAddress.

", - "refs": { - } - }, - "AssociateDhcpOptionsRequest": { - "base": "

Contains the parameters for AssociateDhcpOptions.

", - "refs": { - } - }, - "AssociateRouteTableRequest": { - "base": "

Contains the parameters for AssociateRouteTable.

", - "refs": { - } - }, - "AssociateRouteTableResult": { - "base": "

Contains the output of AssociateRouteTable.

", - "refs": { - } - }, - "AttachClassicLinkVpcRequest": { - "base": "

Contains the parameters for AttachClassicLinkVpc.

", - "refs": { - } - }, - "AttachClassicLinkVpcResult": { - "base": "

Contains the output of AttachClassicLinkVpc.

", - "refs": { - } - }, - "AttachInternetGatewayRequest": { - "base": "

Contains the parameters for AttachInternetGateway.

", - "refs": { - } - }, - "AttachNetworkInterfaceRequest": { - "base": "

Contains the parameters for AttachNetworkInterface.

", - "refs": { - } - }, - "AttachNetworkInterfaceResult": { - "base": "

Contains the output of AttachNetworkInterface.

", - "refs": { - } - }, - "AttachVolumeRequest": { - "base": "

Contains the parameters for AttachVolume.

", - "refs": { - } - }, - "AttachVpnGatewayRequest": { - "base": "

Contains the parameters for AttachVpnGateway.

", - "refs": { - } - }, - "AttachVpnGatewayResult": { - "base": "

Contains the output of AttachVpnGateway.

", - "refs": { - } - }, - "AttachmentStatus": { - "base": null, - "refs": { - "EbsInstanceBlockDevice$Status": "

The attachment state.

", - "InstanceNetworkInterfaceAttachment$Status": "

The attachment state.

", - "InternetGatewayAttachment$State": "

The current state of the attachment.

", - "NetworkInterfaceAttachment$Status": "

The attachment state.

", - "VpcAttachment$State": "

The current state of the attachment.

" - } - }, - "AttributeBooleanValue": { - "base": "

Describes a value for a resource attribute that is a Boolean value.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$SourceDestCheck": "

Indicates whether source/destination checking is enabled.

", - "DescribeVolumeAttributeResult$AutoEnableIO": "

The state of autoEnableIO attribute.

", - "DescribeVpcAttributeResult$EnableDnsSupport": "

Indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for your instances to their corresponding IP addresses; otherwise, it does not.

", - "DescribeVpcAttributeResult$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.

", - "InstanceAttribute$DisableApiTermination": "

If the value is true, you can't terminate the instance through the Amazon EC2 console, CLI, or API; otherwise, you can.

", - "InstanceAttribute$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O.

", - "InstanceAttribute$EnaSupport": "

Indicates whether enhanced networking with ENA is enabled.

", - "InstanceAttribute$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$SourceDestCheck": "

Specifies whether source/destination checking is enabled. A value of true means that checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT.

", - "ModifyInstanceAttributeRequest$DisableApiTermination": "

If the value is true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. You cannot use this paramater for Spot Instances.

", - "ModifyInstanceAttributeRequest$EbsOptimized": "

Specifies whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "ModifyInstanceAttributeRequest$EnaSupport": "

Set to true to enable enhanced networking with ENA for the instance.

This option is supported only for HVM instances. Specifying this option with a PV instance can make it unreachable.

", - "ModifyNetworkInterfaceAttributeRequest$SourceDestCheck": "

Indicates whether source/destination checking is enabled. A value of true means checking is enabled, and false means checking is disabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "ModifySubnetAttributeRequest$MapPublicIpOnLaunch": "

Specify true to indicate that instances launched into the specified subnet should be assigned public IP address.

", - "ModifyVolumeAttributeRequest$AutoEnableIO": "

Indicates whether the volume should be auto-enabled for I/O operations.

", - "ModifyVpcAttributeRequest$EnableDnsSupport": "

Indicates whether the DNS resolution is supported for the VPC. If enabled, queries to the Amazon provided DNS server at the 169.254.169.253 IP address, or the reserved IP address at the base of the VPC network range \"plus two\" will succeed. If disabled, the Amazon provided DNS service in the VPC that resolves public DNS hostnames to IP addresses is not enabled.

You cannot modify the DNS resolution and DNS hostnames attributes in the same request. Use separate requests for each attribute.

", - "ModifyVpcAttributeRequest$EnableDnsHostnames": "

Indicates whether the instances launched in the VPC get DNS hostnames. If enabled, instances in the VPC get DNS hostnames; otherwise, they do not.

You cannot modify the DNS resolution and DNS hostnames attributes in the same request. Use separate requests for each attribute. You can only enable DNS hostnames if you've enabled DNS support.

" - } - }, - "AttributeValue": { - "base": "

Describes a value for a resource attribute that is a String.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Description": "

The description of the network interface.

", - "DhcpConfigurationValueList$member": null, - "ImageAttribute$KernelId": "

The kernel ID.

", - "ImageAttribute$RamdiskId": "

The RAM disk ID.

", - "ImageAttribute$Description": "

A description for the AMI.

", - "ImageAttribute$SriovNetSupport": "

Indicates whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "InstanceAttribute$InstanceType": "

The instance type.

", - "InstanceAttribute$KernelId": "

The kernel ID.

", - "InstanceAttribute$RamdiskId": "

The RAM disk ID.

", - "InstanceAttribute$UserData": "

The user data.

", - "InstanceAttribute$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "InstanceAttribute$RootDeviceName": "

The name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "InstanceAttribute$SriovNetSupport": "

Indicates whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "ModifyImageAttributeRequest$Description": "

A description for the AMI.

", - "ModifyInstanceAttributeRequest$InstanceType": "

Changes the instance type to the specified value. For more information, see Instance Types. If the instance type is not valid, the error returned is InvalidInstanceAttributeValue.

", - "ModifyInstanceAttributeRequest$Kernel": "

Changes the instance's kernel to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$Ramdisk": "

Changes the instance's RAM disk to the specified value. We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB.

", - "ModifyInstanceAttributeRequest$InstanceInitiatedShutdownBehavior": "

Specifies whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "ModifyInstanceAttributeRequest$SriovNetSupport": "

Set to simple to enable enhanced networking with the Intel 82599 Virtual Function interface for the instance.

There is no way to disable enhanced networking with the Intel 82599 Virtual Function interface at this time.

This option is supported only for HVM instances. Specifying this option with a PV instance can make it unreachable.

", - "ModifyNetworkInterfaceAttributeRequest$Description": "

A description for the network interface.

" - } - }, - "AuthorizeSecurityGroupEgressRequest": { - "base": "

Contains the parameters for AuthorizeSecurityGroupEgress.

", - "refs": { - } - }, - "AuthorizeSecurityGroupIngressRequest": { - "base": "

Contains the parameters for AuthorizeSecurityGroupIngress.

", - "refs": { - } - }, - "AutoPlacement": { - "base": null, - "refs": { - "AllocateHostsRequest$AutoPlacement": "

This is enabled by default. This property allows instances to be automatically placed onto available Dedicated Hosts, when you are launching instances without specifying a host ID.

Default: Enabled

", - "Host$AutoPlacement": "

Whether auto-placement is on or off.

", - "ModifyHostsRequest$AutoPlacement": "

Specify whether to enable or disable auto-placement.

" - } - }, - "AvailabilityZone": { - "base": "

Describes an Availability Zone.

", - "refs": { - "AvailabilityZoneList$member": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesResult$AvailabilityZones": "

Information about one or more Availability Zones.

" - } - }, - "AvailabilityZoneMessage": { - "base": "

Describes a message about an Availability Zone.

", - "refs": { - "AvailabilityZoneMessageList$member": null - } - }, - "AvailabilityZoneMessageList": { - "base": null, - "refs": { - "AvailabilityZone$Messages": "

Any messages about the Availability Zone.

" - } - }, - "AvailabilityZoneState": { - "base": null, - "refs": { - "AvailabilityZone$State": "

The state of the Availability Zone.

" - } - }, - "AvailableCapacity": { - "base": "

The capacity information for instances launched onto the Dedicated Host.

", - "refs": { - "Host$AvailableCapacity": "

The number of new instances that can be launched onto the Dedicated Host.

" - } - }, - "AvailableInstanceCapacityList": { - "base": null, - "refs": { - "AvailableCapacity$AvailableInstanceCapacity": "

The total number of instances that the Dedicated Host supports.

" - } - }, - "BatchState": { - "base": null, - "refs": { - "CancelSpotFleetRequestsSuccessItem$CurrentSpotFleetRequestState": "

The current state of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$PreviousSpotFleetRequestState": "

The previous state of the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestState": "

The state of the Spot fleet request.

" - } - }, - "Blob": { - "base": null, - "refs": { - "BlobAttributeValue$Value": null, - "ImportKeyPairRequest$PublicKeyMaterial": "

The public key. For API calls, the text must be base64-encoded. For command line tools, base64 encoding is performed for you.

", - "S3Storage$UploadPolicy": "

An Amazon S3 upload policy that gives Amazon EC2 permission to upload items into Amazon S3 on your behalf.

" - } - }, - "BlobAttributeValue": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$UserData": "

Changes the instance's user data to the specified value. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

" - } - }, - "BlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "BlockDeviceMappingList$member": null, - "BlockDeviceMappingRequestList$member": null - } - }, - "BlockDeviceMappingList": { - "base": null, - "refs": { - "Image$BlockDeviceMappings": "

Any block device mapping entries.

", - "ImageAttribute$BlockDeviceMappings": "

One or more block device mapping entries.

", - "LaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

Although you can specify encrypted EBS volumes in this block device mapping for your Spot Instances, these volumes are not encrypted.

", - "RequestSpotLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

Although you can specify encrypted EBS volumes in this block device mapping for your Spot Instances, these volumes are not encrypted.

", - "SpotFleetLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "BlockDeviceMappingRequestList": { - "base": null, - "refs": { - "CreateImageRequest$BlockDeviceMappings": "

Information about one or more block device mappings.

", - "RegisterImageRequest$BlockDeviceMappings": "

One or more block device mapping entries.

", - "RunInstancesRequest$BlockDeviceMappings": "

The block device mapping.

Supplying both a snapshot ID and an encryption value as arguments for block-device mapping results in an error. This is because only blank volumes can be encrypted on start, and these are not created from a snapshot. If a snapshot is the basis for the volume, it contains data by definition and its encryption status cannot be changed using this action.

" - } - }, - "Boolean": { - "base": null, - "refs": { - "AcceptReservedInstancesExchangeQuoteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AcceptVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AllocateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssignPrivateIpAddressesRequest$AllowReassignment": "

Indicates whether to allow an IP address that is already assigned to another network interface or instance to be reassigned to the specified network interface.

", - "AssociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateAddressRequest$AllowReassociation": "

[EC2-VPC] For a VPC in an EC2-Classic account, specify true to allow an Elastic IP address that is already associated with an instance or network interface to be reassociated with the specified instance or network interface. Otherwise, the operation fails. In a VPC in an EC2-VPC-only account, reassociation is automatic, therefore you can specify false to ensure the operation fails if the Elastic IP address is already associated with another resource.

", - "AssociateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AssociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "AttachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AttributeBooleanValue$Value": "

The attribute value. The valid values are true or false.

", - "AuthorizeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "AuthorizeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "BundleInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelBundleTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelConversionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelImportTaskRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CancelSpotFleetRequestsRequest$TerminateInstances": "

Indicates whether to terminate instances for a Spot fleet request if it is canceled successfully.

", - "CancelSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ClassicLinkDnsSupport$ClassicLinkDnsSupported": "

Indicates whether ClassicLink DNS support is enabled for the VPC.

", - "ConfirmProductInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ConfirmProductInstanceResult$Return": "

The return value of the request. Returns true if the specified product code is owned by the requester and associated with the specified instance.

", - "CopyImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopyImageRequest$Encrypted": "

Specifies whether the destination snapshots of the copied image should be encrypted. The default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CopySnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CopySnapshotRequest$Encrypted": "

Specifies whether the destination snapshot should be encrypted. You can encrypt a copy of an unencrypted snapshot using this flag, but you cannot use it to create an unencrypted copy from an encrypted snapshot. Your default CMK for EBS is used unless a non-default AWS Key Management Service (AWS KMS) CMK is specified with KmsKeyId. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateImageRequest$NoReboot": "

By default, Amazon EC2 attempts to shut down and reboot the instance before creating the image. If the 'No Reboot' option is set, Amazon EC2 doesn't shut down the instance before creating the image. When this option is used, file system integrity on the created image can't be guaranteed.

", - "CreateInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkAclEntryRequest$Egress": "

Indicates whether this is an egress rule (rule is applied to traffic leaving the subnet).

", - "CreateNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreatePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateRouteResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "CreateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVolumeRequest$Encrypted": "

Specifies whether the volume should be encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption. Volumes that are created from encrypted snapshots are automatically encrypted. There is no way to create an encrypted volume from an unencrypted snapshot or vice versa. If your AMI uses encrypted volumes, you can only launch it on supported instance types. For more information, see Amazon EBS Encryption in the Amazon Elastic Compute Cloud User Guide.

", - "CreateVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "CreateVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteCustomerGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkAclEntryRequest$Egress": "

Indicates whether the rule is an egress rule.

", - "DeleteNetworkAclRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeletePlacementGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSecurityGroupRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteSubnetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DeleteVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeleteVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DeregisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAccountAttributesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeAvailabilityZonesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeBundleTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeClassicLinkInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeConversionTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeCustomerGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeDhcpOptionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImagesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportImageTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeImportSnapshotTasksRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInstanceStatusRequest$IncludeAllInstances": "

When true, includes the health status for all instances. When false, includes the health status for running instances only.

Default: false

", - "DescribeInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeInternetGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeKeyPairsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeMovingAddressesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkAclsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeNetworkInterfacesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePlacementGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribePrefixListsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRegionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeReservedInstancesOfferingsRequest$IncludeMarketplace": "

Include Reserved Instance Marketplace offerings in the response.

", - "DescribeReservedInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeRouteTablesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeScheduledInstanceAvailabilityRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupReferencesRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSnapshotsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotDatafeedSubscriptionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotFleetRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotInstanceRequestsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSpotPriceHistoryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeStaleSecurityGroupsRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeSubnetsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeTagsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumeStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVolumesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointServicesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcEndpointsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcPeeringConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpcsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnConnectionsRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DescribeVpnGatewaysRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachClassicLinkVpcResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DetachInternetGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachNetworkInterfaceRequest$Force": "

Specifies whether to force a detachment.

", - "DetachVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DetachVolumeRequest$Force": "

Forces detachment if the previous detachment attempt did not occur cleanly (for example, logging into an instance, unmounting the volume, and detaching normally). This option can lead to data loss or a corrupted file system. Use this option only as a last resort to detach a volume from a failed instance. The instance won't have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures.

", - "DetachVpnGatewayRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkDnsSupportResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "DisassociateAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "DisassociateRouteTableRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EbsBlockDevice$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "EbsBlockDevice$Encrypted": "

Indicates whether the EBS volume is encrypted. Encrypted Amazon EBS volumes may only be attached to instances that support Amazon EBS encryption.

", - "EbsInstanceBlockDevice$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EbsInstanceBlockDeviceSpecification$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "EnableVolumeIORequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkDnsSupportResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "EnableVpcClassicLinkRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "EnableVpcClassicLinkResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "GetConsoleOutputRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetConsoleScreenshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetConsoleScreenshotRequest$WakeUp": "

When set to true, acts as keystroke input and wakes up an instance that's in standby or \"sleep\" mode.

", - "GetPasswordDataRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetReservedInstancesExchangeQuoteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "GetReservedInstancesExchangeQuoteResult$IsValidExchange": "

If true, the exchange is valid. If false, the exchange cannot be performed.

", - "IdFormat$UseLongIds": "

Indicates whether longer IDs (17-character IDs) are enabled for the resource.

", - "Image$Public": "

Indicates whether the image has public launch permissions. The value is true if this image has public launch permissions or false if it has only implicit and explicit launch permissions.

", - "Image$EnaSupport": "

Specifies whether enhanced networking with ENA is enabled.

", - "ImportImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportInstanceLaunchSpecification$Monitoring": "

Indicates whether monitoring is enabled.

", - "ImportInstanceRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportKeyPairRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportSnapshotRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ImportVolumeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Instance$SourceDestCheck": "

Specifies whether to enable an instance launched in a VPC to perform NAT. This controls whether source/destination checking is enabled on the instance. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", - "Instance$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

", - "Instance$EnaSupport": "

Specifies whether enhanced networking with ENA is enabled.

", - "InstanceNetworkInterface$SourceDestCheck": "

Indicates whether to validate network traffic to or from this network interface.

", - "InstanceNetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "InstanceNetworkInterfaceSpecification$DeleteOnTermination": "

If set to true, the interface is deleted when the instance is terminated. You can specify true only if creating a new network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to an instance you launch in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "InstancePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "LaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ModifyIdFormatRequest$UseLongIds": "

Indicate whether the resource should use longer IDs (17-character IDs).

", - "ModifyIdentityIdFormatRequest$UseLongIds": "

Indicates whether the resource should use longer IDs (17-character IDs)

", - "ModifyImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyInstancePlacementResult$Return": "

Is true if the request succeeds, and an error otherwise.

", - "ModifyNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifySpotFleetRequestResponse$Return": "

Is true if the request succeeds, and an error otherwise.

", - "ModifyVolumeAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ModifyVpcEndpointRequest$ResetPolicy": "

Specify true to reset the policy document to the default policy. The default policy allows access to the service.

", - "ModifyVpcEndpointResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ModifyVpcPeeringConnectionOptionsRequest$DryRun": "

Checks whether you have the required permissions for the operation, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "MoveAddressToVpcRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "NetworkAcl$IsDefault": "

Indicates whether this is the default network ACL for the VPC.

", - "NetworkAclEntry$Egress": "

Indicates whether the rule is an egress rule (applied to traffic leaving the subnet).

", - "NetworkInterface$RequesterManaged": "

Indicates whether the network interface is being managed by AWS.

", - "NetworkInterface$SourceDestCheck": "

Indicates whether traffic to or from the instance is validated.

", - "NetworkInterfaceAttachment$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfaceAttachmentChanges$DeleteOnTermination": "

Indicates whether the network interface is deleted when the instance is terminated.

", - "NetworkInterfacePrivateIpAddress$Primary": "

Indicates whether this IP address is the primary private IP address of the network interface.

", - "PeeringConnectionOptions$AllowEgressFromLocalClassicLinkToRemoteVpc": "

If true, enables outbound communication from an EC2-Classic instance that's linked to a local VPC via ClassicLink to instances in a peer VPC.

", - "PeeringConnectionOptions$AllowEgressFromLocalVpcToRemoteClassicLink": "

If true, enables outbound communication from instances in a local VPC to an EC2-Classic instance that's linked to a peer VPC via ClassicLink.

", - "PeeringConnectionOptions$AllowDnsResolutionFromRemoteVpc": "

If true, enables a local VPC to resolve public DNS hostnames to private IP addresses when queried from instances in the peer VPC.

", - "PeeringConnectionOptionsRequest$AllowEgressFromLocalClassicLinkToRemoteVpc": "

If true, enables outbound communication from an EC2-Classic instance that's linked to a local VPC via ClassicLink to instances in a peer VPC.

", - "PeeringConnectionOptionsRequest$AllowEgressFromLocalVpcToRemoteClassicLink": "

If true, enables outbound communication from instances in a local VPC to an EC2-Classic instance that's linked to a peer VPC via ClassicLink.

", - "PeeringConnectionOptionsRequest$AllowDnsResolutionFromRemoteVpc": "

If true, enables a local VPC to resolve public DNS hostnames to private IP addresses when queried from instances in the peer VPC.

", - "PriceSchedule$Active": "

The current price schedule, as determined by the term remaining for the Reserved Instance in the listing.

A specific price schedule is always in effect, but only one price schedule can be active at any time. Take, for example, a Reserved Instance listing that has five months remaining in its term. When you specify price schedules for five months and two months, this means that schedule 1, covering the first three months of the remaining term, will be active during months 5, 4, and 3. Then schedule 2, covering the last two months of the term, will be active for months 2 and 1.

", - "PrivateIpAddressSpecification$Primary": "

Indicates whether the private IP address is the primary private IP address. Only one IP address can be designated as primary.

", - "PurchaseReservedInstancesOfferingRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "PurchaseScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RebootInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RegisterImageRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RegisterImageRequest$EnaSupport": "

Set to true to enable enhanced networking with ENA for the AMI and any instances that you launch from the AMI.

This option is supported only for HVM AMIs. Specifying this option with a PV AMI can make instances launched from the AMI unreachable.

", - "RejectVpcPeeringConnectionRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RejectVpcPeeringConnectionResult$Return": "

Returns true if the request succeeds; otherwise, it returns an error.

", - "ReleaseAddressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceNetworkAclEntryRequest$Egress": "

Indicates whether to replace the egress rule.

Default: If no value is specified, we replace the ingress rule.

", - "ReplaceRouteRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReplaceRouteTableAssociationRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ReportInstanceStatusRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotFleetRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RequestSpotLaunchSpecification$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "ReservedInstancesOffering$Marketplace": "

Indicates whether the offering is available through the Reserved Instance Marketplace (resale) or AWS. If it's a Reserved Instance Marketplace offering, this is true.

", - "ResetImageAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetInstanceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetNetworkInterfaceAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ResetSnapshotAttributeRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RestoreAddressToClassicRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupEgressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RevokeSecurityGroupIngressRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RouteTableAssociation$Main": "

Indicates whether this is the main route table.

", - "RunInstancesMonitoringEnabled$Enabled": "

Indicates whether monitoring is enabled for the instance.

", - "RunInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "RunInstancesRequest$DisableApiTermination": "

If you set this parameter to true, you can't terminate the instance using the Amazon EC2 console, CLI, or API; otherwise, you can. If you set this parameter to true and then later want to be able to terminate the instance, you must first change the value of the disableApiTermination attribute to false using ModifyInstanceAttribute. Alternatively, if you set InstanceInitiatedShutdownBehavior to terminate, you can terminate the instance by running the shutdown command from the instance.

Default: false

", - "RunInstancesRequest$EbsOptimized": "

Indicates whether the instance is optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "RunScheduledInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "ScheduledInstanceRecurrence$OccurrenceRelativeToEnd": "

Indicates whether the occurrence is relative to the end of the specified week or month.

", - "ScheduledInstanceRecurrenceRequest$OccurrenceRelativeToEnd": "

Indicates whether the occurrence is relative to the end of the specified week or month. You can't specify this value with a daily schedule.

", - "ScheduledInstancesEbs$DeleteOnTermination": "

Indicates whether the volume is deleted on instance termination.

", - "ScheduledInstancesEbs$Encrypted": "

Indicates whether the volume is encrypted. You can attached encrypted volumes only to instances that support them.

", - "ScheduledInstancesLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS-optimized instance.

Default: false

", - "ScheduledInstancesMonitoring$Enabled": "

Indicates whether monitoring is enabled.

", - "ScheduledInstancesNetworkInterface$AssociatePublicIpAddress": "

Indicates whether to assign a public IP address to instances launched in a VPC. The public IP address can only be assigned to a network interface for eth0, and can only be assigned to a new network interface, not an existing one. You cannot specify more than one network interface in the request. If launching into a default subnet, the default value is true.

", - "ScheduledInstancesNetworkInterface$DeleteOnTermination": "

Indicates whether to delete the interface when the instance is terminated.

", - "ScheduledInstancesPrivateIpAddressConfig$Primary": "

Indicates whether this is a primary IP address. Otherwise, this is a secondary IP address.

", - "Snapshot$Encrypted": "

Indicates whether the snapshot is encrypted.

", - "SpotFleetLaunchSpecification$EbsOptimized": "

Indicates whether the instances are optimized for EBS I/O. This optimization provides dedicated throughput to Amazon EBS and an optimized configuration stack to provide optimal EBS I/O performance. This optimization isn't available with all instance types. Additional usage charges apply when using an EBS Optimized instance.

Default: false

", - "SpotFleetMonitoring$Enabled": "

Enables monitoring for the instance.

Default: false

", - "SpotFleetRequestConfigData$TerminateInstancesWithExpiration": "

Indicates whether running Spot instances should be terminated when the Spot fleet request expires.

", - "StartInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "StopInstancesRequest$Force": "

Forces the instances to stop. The instances do not have an opportunity to flush file system caches or file system metadata. If you use this option, you must perform file system check and repair procedures. This option is not recommended for Windows instances.

Default: false

", - "Subnet$DefaultForAz": "

Indicates whether this is the default subnet for the Availability Zone.

", - "Subnet$MapPublicIpOnLaunch": "

Indicates whether instances launched in this subnet receive a public IP address.

", - "TerminateInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "UnmonitorInstancesRequest$DryRun": "

Checks whether you have the required permissions for the action, without actually making the request, and provides an error response. If you have the required permissions, the error response is DryRunOperation. Otherwise, it is UnauthorizedOperation.

", - "Volume$Encrypted": "

Indicates whether the volume will be encrypted.

", - "VolumeAttachment$DeleteOnTermination": "

Indicates whether the EBS volume is deleted on instance termination.

", - "Vpc$IsDefault": "

Indicates whether the VPC is the default VPC.

", - "VpcClassicLink$ClassicLinkEnabled": "

Indicates whether the VPC is enabled for ClassicLink.

", - "VpcPeeringConnectionOptionsDescription$AllowEgressFromLocalClassicLinkToRemoteVpc": "

Indicates whether a local ClassicLink connection can communicate with the peer VPC over the VPC peering connection.

", - "VpcPeeringConnectionOptionsDescription$AllowEgressFromLocalVpcToRemoteClassicLink": "

Indicates whether a local VPC can communicate with a ClassicLink connection in the peer VPC over the VPC peering connection.

", - "VpcPeeringConnectionOptionsDescription$AllowDnsResolutionFromRemoteVpc": "

Indicates whether a local VPC can resolve public DNS hostnames to private IP addresses when queried from instances in a peer VPC.

", - "VpnConnectionOptions$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

", - "VpnConnectionOptionsSpecification$StaticRoutesOnly": "

Indicates whether the VPN connection uses static routes only. Static routes must be used for devices that don't support BGP.

" - } - }, - "BundleIdStringList": { - "base": null, - "refs": { - "DescribeBundleTasksRequest$BundleIds": "

One or more bundle task IDs.

Default: Describes all your bundle tasks.

" - } - }, - "BundleInstanceRequest": { - "base": "

Contains the parameters for BundleInstance.

", - "refs": { - } - }, - "BundleInstanceResult": { - "base": "

Contains the output of BundleInstance.

", - "refs": { - } - }, - "BundleTask": { - "base": "

Describes a bundle task.

", - "refs": { - "BundleInstanceResult$BundleTask": "

Information about the bundle task.

", - "BundleTaskList$member": null, - "CancelBundleTaskResult$BundleTask": "

Information about the bundle task.

" - } - }, - "BundleTaskError": { - "base": "

Describes an error for BundleInstance.

", - "refs": { - "BundleTask$BundleTaskError": "

If the task fails, a description of the error.

" - } - }, - "BundleTaskList": { - "base": null, - "refs": { - "DescribeBundleTasksResult$BundleTasks": "

Information about one or more bundle tasks.

" - } - }, - "BundleTaskState": { - "base": null, - "refs": { - "BundleTask$State": "

The state of the task.

" - } - }, - "CancelBatchErrorCode": { - "base": null, - "refs": { - "CancelSpotFleetRequestsError$Code": "

The error code.

" - } - }, - "CancelBundleTaskRequest": { - "base": "

Contains the parameters for CancelBundleTask.

", - "refs": { - } - }, - "CancelBundleTaskResult": { - "base": "

Contains the output of CancelBundleTask.

", - "refs": { - } - }, - "CancelConversionRequest": { - "base": "

Contains the parameters for CancelConversionTask.

", - "refs": { - } - }, - "CancelExportTaskRequest": { - "base": "

Contains the parameters for CancelExportTask.

", - "refs": { - } - }, - "CancelImportTaskRequest": { - "base": "

Contains the parameters for CancelImportTask.

", - "refs": { - } - }, - "CancelImportTaskResult": { - "base": "

Contains the output for CancelImportTask.

", - "refs": { - } - }, - "CancelReservedInstancesListingRequest": { - "base": "

Contains the parameters for CancelReservedInstancesListing.

", - "refs": { - } - }, - "CancelReservedInstancesListingResult": { - "base": "

Contains the output of CancelReservedInstancesListing.

", - "refs": { - } - }, - "CancelSpotFleetRequestsError": { - "base": "

Describes a Spot fleet error.

", - "refs": { - "CancelSpotFleetRequestsErrorItem$Error": "

The error.

" - } - }, - "CancelSpotFleetRequestsErrorItem": { - "base": "

Describes a Spot fleet request that was not successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsErrorSet$member": null - } - }, - "CancelSpotFleetRequestsErrorSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$UnsuccessfulFleetRequests": "

Information about the Spot fleet requests that are not successfully canceled.

" - } - }, - "CancelSpotFleetRequestsRequest": { - "base": "

Contains the parameters for CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsResponse": { - "base": "

Contains the output of CancelSpotFleetRequests.

", - "refs": { - } - }, - "CancelSpotFleetRequestsSuccessItem": { - "base": "

Describes a Spot fleet request that was successfully canceled.

", - "refs": { - "CancelSpotFleetRequestsSuccessSet$member": null - } - }, - "CancelSpotFleetRequestsSuccessSet": { - "base": null, - "refs": { - "CancelSpotFleetRequestsResponse$SuccessfulFleetRequests": "

Information about the Spot fleet requests that are successfully canceled.

" - } - }, - "CancelSpotInstanceRequestState": { - "base": null, - "refs": { - "CancelledSpotInstanceRequest$State": "

The state of the Spot instance request.

" - } - }, - "CancelSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelSpotInstanceRequestsResult": { - "base": "

Contains the output of CancelSpotInstanceRequests.

", - "refs": { - } - }, - "CancelledSpotInstanceRequest": { - "base": "

Describes a request to cancel a Spot instance.

", - "refs": { - "CancelledSpotInstanceRequestList$member": null - } - }, - "CancelledSpotInstanceRequestList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsResult$CancelledSpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "ClassicLinkDnsSupport": { - "base": "

Describes the ClassicLink DNS support status of a VPC.

", - "refs": { - "ClassicLinkDnsSupportList$member": null - } - }, - "ClassicLinkDnsSupportList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkDnsSupportResult$Vpcs": "

Information about the ClassicLink DNS support status of the VPCs.

" - } - }, - "ClassicLinkInstance": { - "base": "

Describes a linked EC2-Classic instance.

", - "refs": { - "ClassicLinkInstanceList$member": null - } - }, - "ClassicLinkInstanceList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesResult$Instances": "

Information about one or more linked EC2-Classic instances.

" - } - }, - "ClientData": { - "base": "

Describes the client-specific data.

", - "refs": { - "ImportImageRequest$ClientData": "

The client-specific data.

", - "ImportSnapshotRequest$ClientData": "

The client-specific data.

" - } - }, - "ConfirmProductInstanceRequest": { - "base": "

Contains the parameters for ConfirmProductInstance.

", - "refs": { - } - }, - "ConfirmProductInstanceResult": { - "base": "

Contains the output of ConfirmProductInstance.

", - "refs": { - } - }, - "ContainerFormat": { - "base": null, - "refs": { - "ExportToS3Task$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

", - "ExportToS3TaskSpecification$ContainerFormat": "

The container format used to combine disk images with metadata (such as OVF). If absent, only the disk image is exported.

" - } - }, - "ConversionIdStringList": { - "base": null, - "refs": { - "DescribeConversionTasksRequest$ConversionTaskIds": "

One or more conversion task IDs.

" - } - }, - "ConversionTask": { - "base": "

Describes a conversion task.

", - "refs": { - "DescribeConversionTaskList$member": null, - "ImportInstanceResult$ConversionTask": "

Information about the conversion task.

", - "ImportVolumeResult$ConversionTask": "

Information about the conversion task.

" - } - }, - "ConversionTaskState": { - "base": null, - "refs": { - "ConversionTask$State": "

The state of the conversion task.

" - } - }, - "CopyImageRequest": { - "base": "

Contains the parameters for CopyImage.

", - "refs": { - } - }, - "CopyImageResult": { - "base": "

Contains the output of CopyImage.

", - "refs": { - } - }, - "CopySnapshotRequest": { - "base": "

Contains the parameters for CopySnapshot.

", - "refs": { - } - }, - "CopySnapshotResult": { - "base": "

Contains the output of CopySnapshot.

", - "refs": { - } - }, - "CreateCustomerGatewayRequest": { - "base": "

Contains the parameters for CreateCustomerGateway.

", - "refs": { - } - }, - "CreateCustomerGatewayResult": { - "base": "

Contains the output of CreateCustomerGateway.

", - "refs": { - } - }, - "CreateDhcpOptionsRequest": { - "base": "

Contains the parameters for CreateDhcpOptions.

", - "refs": { - } - }, - "CreateDhcpOptionsResult": { - "base": "

Contains the output of CreateDhcpOptions.

", - "refs": { - } - }, - "CreateFlowLogsRequest": { - "base": "

Contains the parameters for CreateFlowLogs.

", - "refs": { - } - }, - "CreateFlowLogsResult": { - "base": "

Contains the output of CreateFlowLogs.

", - "refs": { - } - }, - "CreateImageRequest": { - "base": "

Contains the parameters for CreateImage.

", - "refs": { - } - }, - "CreateImageResult": { - "base": "

Contains the output of CreateImage.

", - "refs": { - } - }, - "CreateInstanceExportTaskRequest": { - "base": "

Contains the parameters for CreateInstanceExportTask.

", - "refs": { - } - }, - "CreateInstanceExportTaskResult": { - "base": "

Contains the output for CreateInstanceExportTask.

", - "refs": { - } - }, - "CreateInternetGatewayRequest": { - "base": "

Contains the parameters for CreateInternetGateway.

", - "refs": { - } - }, - "CreateInternetGatewayResult": { - "base": "

Contains the output of CreateInternetGateway.

", - "refs": { - } - }, - "CreateKeyPairRequest": { - "base": "

Contains the parameters for CreateKeyPair.

", - "refs": { - } - }, - "CreateNatGatewayRequest": { - "base": "

Contains the parameters for CreateNatGateway.

", - "refs": { - } - }, - "CreateNatGatewayResult": { - "base": "

Contains the output of CreateNatGateway.

", - "refs": { - } - }, - "CreateNetworkAclEntryRequest": { - "base": "

Contains the parameters for CreateNetworkAclEntry.

", - "refs": { - } - }, - "CreateNetworkAclRequest": { - "base": "

Contains the parameters for CreateNetworkAcl.

", - "refs": { - } - }, - "CreateNetworkAclResult": { - "base": "

Contains the output of CreateNetworkAcl.

", - "refs": { - } - }, - "CreateNetworkInterfaceRequest": { - "base": "

Contains the parameters for CreateNetworkInterface.

", - "refs": { - } - }, - "CreateNetworkInterfaceResult": { - "base": "

Contains the output of CreateNetworkInterface.

", - "refs": { - } - }, - "CreatePlacementGroupRequest": { - "base": "

Contains the parameters for CreatePlacementGroup.

", - "refs": { - } - }, - "CreateReservedInstancesListingRequest": { - "base": "

Contains the parameters for CreateReservedInstancesListing.

", - "refs": { - } - }, - "CreateReservedInstancesListingResult": { - "base": "

Contains the output of CreateReservedInstancesListing.

", - "refs": { - } - }, - "CreateRouteRequest": { - "base": "

Contains the parameters for CreateRoute.

", - "refs": { - } - }, - "CreateRouteResult": { - "base": "

Contains the output of CreateRoute.

", - "refs": { - } - }, - "CreateRouteTableRequest": { - "base": "

Contains the parameters for CreateRouteTable.

", - "refs": { - } - }, - "CreateRouteTableResult": { - "base": "

Contains the output of CreateRouteTable.

", - "refs": { - } - }, - "CreateSecurityGroupRequest": { - "base": "

Contains the parameters for CreateSecurityGroup.

", - "refs": { - } - }, - "CreateSecurityGroupResult": { - "base": "

Contains the output of CreateSecurityGroup.

", - "refs": { - } - }, - "CreateSnapshotRequest": { - "base": "

Contains the parameters for CreateSnapshot.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of CreateSpotDatafeedSubscription.

", - "refs": { - } - }, - "CreateSubnetRequest": { - "base": "

Contains the parameters for CreateSubnet.

", - "refs": { - } - }, - "CreateSubnetResult": { - "base": "

Contains the output of CreateSubnet.

", - "refs": { - } - }, - "CreateTagsRequest": { - "base": "

Contains the parameters for CreateTags.

", - "refs": { - } - }, - "CreateVolumePermission": { - "base": "

Describes the user or group to be added or removed from the permissions for a volume.

", - "refs": { - "CreateVolumePermissionList$member": null - } - }, - "CreateVolumePermissionList": { - "base": null, - "refs": { - "CreateVolumePermissionModifications$Add": "

Adds a specific AWS account ID or group to a volume's list of create volume permissions.

", - "CreateVolumePermissionModifications$Remove": "

Removes a specific AWS account ID or group from a volume's list of create volume permissions.

", - "DescribeSnapshotAttributeResult$CreateVolumePermissions": "

A list of permissions for creating volumes from the snapshot.

" - } - }, - "CreateVolumePermissionModifications": { - "base": "

Describes modifications to the permissions for a volume.

", - "refs": { - "ModifySnapshotAttributeRequest$CreateVolumePermission": "

A JSON representation of the snapshot attribute modification.

" - } - }, - "CreateVolumeRequest": { - "base": "

Contains the parameters for CreateVolume.

", - "refs": { - } - }, - "CreateVpcEndpointRequest": { - "base": "

Contains the parameters for CreateVpcEndpoint.

", - "refs": { - } - }, - "CreateVpcEndpointResult": { - "base": "

Contains the output of CreateVpcEndpoint.

", - "refs": { - } - }, - "CreateVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for CreateVpcPeeringConnection.

", - "refs": { - } - }, - "CreateVpcPeeringConnectionResult": { - "base": "

Contains the output of CreateVpcPeeringConnection.

", - "refs": { - } - }, - "CreateVpcRequest": { - "base": "

Contains the parameters for CreateVpc.

", - "refs": { - } - }, - "CreateVpcResult": { - "base": "

Contains the output of CreateVpc.

", - "refs": { - } - }, - "CreateVpnConnectionRequest": { - "base": "

Contains the parameters for CreateVpnConnection.

", - "refs": { - } - }, - "CreateVpnConnectionResult": { - "base": "

Contains the output of CreateVpnConnection.

", - "refs": { - } - }, - "CreateVpnConnectionRouteRequest": { - "base": "

Contains the parameters for CreateVpnConnectionRoute.

", - "refs": { - } - }, - "CreateVpnGatewayRequest": { - "base": "

Contains the parameters for CreateVpnGateway.

", - "refs": { - } - }, - "CreateVpnGatewayResult": { - "base": "

Contains the output of CreateVpnGateway.

", - "refs": { - } - }, - "CurrencyCodeValues": { - "base": null, - "refs": { - "GetHostReservationPurchasePreviewResult$CurrencyCode": "

The currency in which the totalUpfrontPrice and totalHourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "HostOffering$CurrencyCode": "

The currency of the offering.

", - "HostReservation$CurrencyCode": "

The currency in which the upfrontPrice and hourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "PriceSchedule$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "PriceScheduleSpecification$CurrencyCode": "

The currency for transacting the Reserved Instance resale. At this time, the only supported currency is USD.

", - "Purchase$CurrencyCode": "

The currency in which the UpfrontPrice and HourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "PurchaseHostReservationRequest$CurrencyCode": "

The currency in which the totalUpfrontPrice, LimitPrice, and totalHourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "PurchaseHostReservationResult$CurrencyCode": "

The currency in which the totalUpfrontPrice and totalHourlyPrice amounts are specified. At this time, the only supported currency is USD.

", - "ReservedInstanceLimitPrice$CurrencyCode": "

The currency in which the limitPrice amount is specified. At this time, the only supported currency is USD.

", - "ReservedInstances$CurrencyCode": "

The currency of the Reserved Instance. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

", - "ReservedInstancesOffering$CurrencyCode": "

The currency of the Reserved Instance offering you are purchasing. It's specified using ISO 4217 standard currency codes. At this time, the only supported currency is USD.

" - } - }, - "CustomerGateway": { - "base": "

Describes a customer gateway.

", - "refs": { - "CreateCustomerGatewayResult$CustomerGateway": "

Information about the customer gateway.

", - "CustomerGatewayList$member": null - } - }, - "CustomerGatewayIdStringList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysRequest$CustomerGatewayIds": "

One or more customer gateway IDs.

Default: Describes all your customer gateways.

" - } - }, - "CustomerGatewayList": { - "base": null, - "refs": { - "DescribeCustomerGatewaysResult$CustomerGateways": "

Information about one or more customer gateways.

" - } - }, - "DatafeedSubscriptionState": { - "base": null, - "refs": { - "SpotDatafeedSubscription$State": "

The state of the Spot instance data feed subscription.

" - } - }, - "DateTime": { - "base": null, - "refs": { - "BundleTask$StartTime": "

The time this task started.

", - "BundleTask$UpdateTime": "

The time of the most recent update for the task.

", - "ClientData$UploadStart": "

The time that the disk upload starts.

", - "ClientData$UploadEnd": "

The time that the disk upload ends.

", - "DescribeSpotFleetRequestHistoryRequest$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$StartTime": "

The starting date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotFleetRequestHistoryResponse$LastEvaluatedTime": "

The last date and time for the events, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). All records up to this time were retrieved.

If nextToken indicates that there are more results, this value is not present.

", - "DescribeSpotPriceHistoryRequest$StartTime": "

The date and time, up to the past 90 days, from which to start retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "DescribeSpotPriceHistoryRequest$EndTime": "

The date and time, up to the current date, from which to stop retrieving the price history data, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "EbsInstanceBlockDevice$AttachTime": "

The time stamp when the attachment initiated.

", - "FlowLog$CreationTime": "

The date and time the flow log was created.

", - "GetConsoleOutputResult$Timestamp": "

The time the output was last updated.

", - "GetPasswordDataResult$Timestamp": "

The time the data was last updated.

", - "GetReservedInstancesExchangeQuoteResult$OutputReservedInstancesWillExpireAt": "

The new end date of the reservation term.

", - "HistoryRecord$Timestamp": "

The date and time of the event, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "HostReservation$End": "

The date and time that the reservation ends.

", - "HostReservation$Start": "

The date and time that the reservation started.

", - "IdFormat$Deadline": "

The date in UTC at which you are permanently switched over to using longer IDs. If a deadline is not yet available for this resource type, this field is not returned.

", - "Instance$LaunchTime": "

The time the instance was launched.

", - "InstanceNetworkInterfaceAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "InstanceStatusDetails$ImpairedSince": "

The time when a status check failed. For an instance that was launched and impaired, this is the time when the instance was launched.

", - "InstanceStatusEvent$NotBefore": "

The earliest scheduled start time for the event.

", - "InstanceStatusEvent$NotAfter": "

The latest scheduled end time for the event.

", - "NatGateway$CreateTime": "

The date and time the NAT gateway was created.

", - "NatGateway$DeleteTime": "

The date and time the NAT gateway was deleted, if applicable.

", - "NetworkInterfaceAttachment$AttachTime": "

The timestamp indicating when the attachment initiated.

", - "ProvisionedBandwidth$RequestTime": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$ProvisionTime": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ReportInstanceStatusRequest$StartTime": "

The time at which the reported instance health state began.

", - "ReportInstanceStatusRequest$EndTime": "

The time at which the reported instance health state ended.

", - "RequestSpotInstancesRequest$ValidFrom": "

The start date of the request. If this is a one-time request, the request becomes active at this date and time and remains active until all instances launch, the request expires, or the request is canceled. If the request is persistent, the request becomes active at this date and time and remains active until it expires or is canceled.

Default: The request is effective indefinitely.

", - "RequestSpotInstancesRequest$ValidUntil": "

The end date of the request. If this is a one-time request, the request remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date and time is reached.

Default: The request is effective indefinitely.

", - "ReservedInstances$Start": "

The date and time the Reserved Instance started.

", - "ReservedInstances$End": "

The time when the Reserved Instance expires.

", - "ReservedInstancesListing$CreateDate": "

The time the listing was created.

", - "ReservedInstancesListing$UpdateDate": "

The last modified timestamp of the listing.

", - "ReservedInstancesModification$CreateDate": "

The time when the modification request was created.

", - "ReservedInstancesModification$UpdateDate": "

The time when the modification request was last updated.

", - "ReservedInstancesModification$EffectiveDate": "

The time for the modification to become effective.

", - "ScheduledInstance$PreviousSlotEndTime": "

The time that the previous schedule ended or will end.

", - "ScheduledInstance$NextSlotStartTime": "

The time for the next schedule to start.

", - "ScheduledInstance$TermStartDate": "

The start date for the Scheduled Instance.

", - "ScheduledInstance$TermEndDate": "

The end date for the Scheduled Instance.

", - "ScheduledInstance$CreateDate": "

The date when the Scheduled Instance was purchased.

", - "ScheduledInstanceAvailability$FirstSlotStartTime": "

The time period for the first schedule to start.

", - "SlotDateTimeRangeRequest$EarliestTime": "

The earliest date and time, in UTC, for the Scheduled Instance to start.

", - "SlotDateTimeRangeRequest$LatestTime": "

The latest date and time, in UTC, for the Scheduled Instance to start. This value must be later than or equal to the earliest date and at most three months in the future.

", - "SlotStartTimeRangeRequest$EarliestTime": "

The earliest date and time, in UTC, for the Scheduled Instance to start.

", - "SlotStartTimeRangeRequest$LatestTime": "

The latest date and time, in UTC, for the Scheduled Instance to start.

", - "Snapshot$StartTime": "

The time stamp when the snapshot was initiated.

", - "SpotFleetRequestConfig$CreateTime": "

The creation date and time of the request.

", - "SpotFleetRequestConfigData$ValidFrom": "

The start date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The default is to start fulfilling the request immediately.

", - "SpotFleetRequestConfigData$ValidUntil": "

The end date and time of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). At this point, no new Spot instance requests are placed or enabled to fulfill the request.

", - "SpotInstanceRequest$ValidFrom": "

The start date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). The request becomes active at this date and time.

", - "SpotInstanceRequest$ValidUntil": "

The end date of the request, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). If this is a one-time request, it remains active until all instances launch, the request is canceled, or this date is reached. If the request is persistent, it remains active until it is canceled or this date is reached.

", - "SpotInstanceRequest$CreateTime": "

The date and time when the Spot instance request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotInstanceStatus$UpdateTime": "

The date and time of the most recent status update, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "SpotPrice$Timestamp": "

The date and time the request was created, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ).

", - "VgwTelemetry$LastStatusChange": "

The date and time of the last change in status.

", - "Volume$CreateTime": "

The time stamp when volume creation was initiated.

", - "VolumeAttachment$AttachTime": "

The time stamp when the attachment initiated.

", - "VolumeStatusEvent$NotBefore": "

The earliest start time of the event.

", - "VolumeStatusEvent$NotAfter": "

The latest end time of the event.

", - "VpcEndpoint$CreationTimestamp": "

The date and time the VPC endpoint was created.

", - "VpcPeeringConnection$ExpirationTime": "

The time that an unaccepted VPC peering connection will expire.

" - } - }, - "DeleteCustomerGatewayRequest": { - "base": "

Contains the parameters for DeleteCustomerGateway.

", - "refs": { - } - }, - "DeleteDhcpOptionsRequest": { - "base": "

Contains the parameters for DeleteDhcpOptions.

", - "refs": { - } - }, - "DeleteFlowLogsRequest": { - "base": "

Contains the parameters for DeleteFlowLogs.

", - "refs": { - } - }, - "DeleteFlowLogsResult": { - "base": "

Contains the output of DeleteFlowLogs.

", - "refs": { - } - }, - "DeleteInternetGatewayRequest": { - "base": "

Contains the parameters for DeleteInternetGateway.

", - "refs": { - } - }, - "DeleteKeyPairRequest": { - "base": "

Contains the parameters for DeleteKeyPair.

", - "refs": { - } - }, - "DeleteNatGatewayRequest": { - "base": "

Contains the parameters for DeleteNatGateway.

", - "refs": { - } - }, - "DeleteNatGatewayResult": { - "base": "

Contains the output of DeleteNatGateway.

", - "refs": { - } - }, - "DeleteNetworkAclEntryRequest": { - "base": "

Contains the parameters for DeleteNetworkAclEntry.

", - "refs": { - } - }, - "DeleteNetworkAclRequest": { - "base": "

Contains the parameters for DeleteNetworkAcl.

", - "refs": { - } - }, - "DeleteNetworkInterfaceRequest": { - "base": "

Contains the parameters for DeleteNetworkInterface.

", - "refs": { - } - }, - "DeletePlacementGroupRequest": { - "base": "

Contains the parameters for DeletePlacementGroup.

", - "refs": { - } - }, - "DeleteRouteRequest": { - "base": "

Contains the parameters for DeleteRoute.

", - "refs": { - } - }, - "DeleteRouteTableRequest": { - "base": "

Contains the parameters for DeleteRouteTable.

", - "refs": { - } - }, - "DeleteSecurityGroupRequest": { - "base": "

Contains the parameters for DeleteSecurityGroup.

", - "refs": { - } - }, - "DeleteSnapshotRequest": { - "base": "

Contains the parameters for DeleteSnapshot.

", - "refs": { - } - }, - "DeleteSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DeleteSpotDatafeedSubscription.

", - "refs": { - } - }, - "DeleteSubnetRequest": { - "base": "

Contains the parameters for DeleteSubnet.

", - "refs": { - } - }, - "DeleteTagsRequest": { - "base": "

Contains the parameters for DeleteTags.

", - "refs": { - } - }, - "DeleteVolumeRequest": { - "base": "

Contains the parameters for DeleteVolume.

", - "refs": { - } - }, - "DeleteVpcEndpointsRequest": { - "base": "

Contains the parameters for DeleteVpcEndpoints.

", - "refs": { - } - }, - "DeleteVpcEndpointsResult": { - "base": "

Contains the output of DeleteVpcEndpoints.

", - "refs": { - } - }, - "DeleteVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for DeleteVpcPeeringConnection.

", - "refs": { - } - }, - "DeleteVpcPeeringConnectionResult": { - "base": "

Contains the output of DeleteVpcPeeringConnection.

", - "refs": { - } - }, - "DeleteVpcRequest": { - "base": "

Contains the parameters for DeleteVpc.

", - "refs": { - } - }, - "DeleteVpnConnectionRequest": { - "base": "

Contains the parameters for DeleteVpnConnection.

", - "refs": { - } - }, - "DeleteVpnConnectionRouteRequest": { - "base": "

Contains the parameters for DeleteVpnConnectionRoute.

", - "refs": { - } - }, - "DeleteVpnGatewayRequest": { - "base": "

Contains the parameters for DeleteVpnGateway.

", - "refs": { - } - }, - "DeregisterImageRequest": { - "base": "

Contains the parameters for DeregisterImage.

", - "refs": { - } - }, - "DescribeAccountAttributesRequest": { - "base": "

Contains the parameters for DescribeAccountAttributes.

", - "refs": { - } - }, - "DescribeAccountAttributesResult": { - "base": "

Contains the output of DescribeAccountAttributes.

", - "refs": { - } - }, - "DescribeAddressesRequest": { - "base": "

Contains the parameters for DescribeAddresses.

", - "refs": { - } - }, - "DescribeAddressesResult": { - "base": "

Contains the output of DescribeAddresses.

", - "refs": { - } - }, - "DescribeAvailabilityZonesRequest": { - "base": "

Contains the parameters for DescribeAvailabilityZones.

", - "refs": { - } - }, - "DescribeAvailabilityZonesResult": { - "base": "

Contains the output of DescribeAvailabiltyZones.

", - "refs": { - } - }, - "DescribeBundleTasksRequest": { - "base": "

Contains the parameters for DescribeBundleTasks.

", - "refs": { - } - }, - "DescribeBundleTasksResult": { - "base": "

Contains the output of DescribeBundleTasks.

", - "refs": { - } - }, - "DescribeClassicLinkInstancesRequest": { - "base": "

Contains the parameters for DescribeClassicLinkInstances.

", - "refs": { - } - }, - "DescribeClassicLinkInstancesResult": { - "base": "

Contains the output of DescribeClassicLinkInstances.

", - "refs": { - } - }, - "DescribeConversionTaskList": { - "base": null, - "refs": { - "DescribeConversionTasksResult$ConversionTasks": "

Information about the conversion tasks.

" - } - }, - "DescribeConversionTasksRequest": { - "base": "

Contains the parameters for DescribeConversionTasks.

", - "refs": { - } - }, - "DescribeConversionTasksResult": { - "base": "

Contains the output for DescribeConversionTasks.

", - "refs": { - } - }, - "DescribeCustomerGatewaysRequest": { - "base": "

Contains the parameters for DescribeCustomerGateways.

", - "refs": { - } - }, - "DescribeCustomerGatewaysResult": { - "base": "

Contains the output of DescribeCustomerGateways.

", - "refs": { - } - }, - "DescribeDhcpOptionsRequest": { - "base": "

Contains the parameters for DescribeDhcpOptions.

", - "refs": { - } - }, - "DescribeDhcpOptionsResult": { - "base": "

Contains the output of DescribeDhcpOptions.

", - "refs": { - } - }, - "DescribeExportTasksRequest": { - "base": "

Contains the parameters for DescribeExportTasks.

", - "refs": { - } - }, - "DescribeExportTasksResult": { - "base": "

Contains the output for DescribeExportTasks.

", - "refs": { - } - }, - "DescribeFlowLogsRequest": { - "base": "

Contains the parameters for DescribeFlowLogs.

", - "refs": { - } - }, - "DescribeFlowLogsResult": { - "base": "

Contains the output of DescribeFlowLogs.

", - "refs": { - } - }, - "DescribeHostReservationOfferingsRequest": { - "base": null, - "refs": { - } - }, - "DescribeHostReservationOfferingsResult": { - "base": null, - "refs": { - } - }, - "DescribeHostReservationsRequest": { - "base": null, - "refs": { - } - }, - "DescribeHostReservationsResult": { - "base": null, - "refs": { - } - }, - "DescribeHostsRequest": { - "base": "

Contains the parameters for DescribeHosts.

", - "refs": { - } - }, - "DescribeHostsResult": { - "base": "

Contains the output of DescribeHosts.

", - "refs": { - } - }, - "DescribeIdFormatRequest": { - "base": "

Contains the parameters for DescribeIdFormat.

", - "refs": { - } - }, - "DescribeIdFormatResult": { - "base": "

Contains the output of DescribeIdFormat.

", - "refs": { - } - }, - "DescribeIdentityIdFormatRequest": { - "base": "

Contains the parameters for DescribeIdentityIdFormat.

", - "refs": { - } - }, - "DescribeIdentityIdFormatResult": { - "base": "

Contains the output of DescribeIdentityIdFormat.

", - "refs": { - } - }, - "DescribeImageAttributeRequest": { - "base": "

Contains the parameters for DescribeImageAttribute.

", - "refs": { - } - }, - "DescribeImagesRequest": { - "base": "

Contains the parameters for DescribeImages.

", - "refs": { - } - }, - "DescribeImagesResult": { - "base": "

Contains the output of DescribeImages.

", - "refs": { - } - }, - "DescribeImportImageTasksRequest": { - "base": "

Contains the parameters for DescribeImportImageTasks.

", - "refs": { - } - }, - "DescribeImportImageTasksResult": { - "base": "

Contains the output for DescribeImportImageTasks.

", - "refs": { - } - }, - "DescribeImportSnapshotTasksRequest": { - "base": "

Contains the parameters for DescribeImportSnapshotTasks.

", - "refs": { - } - }, - "DescribeImportSnapshotTasksResult": { - "base": "

Contains the output for DescribeImportSnapshotTasks.

", - "refs": { - } - }, - "DescribeInstanceAttributeRequest": { - "base": "

Contains the parameters for DescribeInstanceAttribute.

", - "refs": { - } - }, - "DescribeInstanceStatusRequest": { - "base": "

Contains the parameters for DescribeInstanceStatus.

", - "refs": { - } - }, - "DescribeInstanceStatusResult": { - "base": "

Contains the output of DescribeInstanceStatus.

", - "refs": { - } - }, - "DescribeInstancesRequest": { - "base": "

Contains the parameters for DescribeInstances.

", - "refs": { - } - }, - "DescribeInstancesResult": { - "base": "

Contains the output of DescribeInstances.

", - "refs": { - } - }, - "DescribeInternetGatewaysRequest": { - "base": "

Contains the parameters for DescribeInternetGateways.

", - "refs": { - } - }, - "DescribeInternetGatewaysResult": { - "base": "

Contains the output of DescribeInternetGateways.

", - "refs": { - } - }, - "DescribeKeyPairsRequest": { - "base": "

Contains the parameters for DescribeKeyPairs.

", - "refs": { - } - }, - "DescribeKeyPairsResult": { - "base": "

Contains the output of DescribeKeyPairs.

", - "refs": { - } - }, - "DescribeMovingAddressesRequest": { - "base": "

Contains the parameters for DescribeMovingAddresses.

", - "refs": { - } - }, - "DescribeMovingAddressesResult": { - "base": "

Contains the output of DescribeMovingAddresses.

", - "refs": { - } - }, - "DescribeNatGatewaysRequest": { - "base": "

Contains the parameters for DescribeNatGateways.

", - "refs": { - } - }, - "DescribeNatGatewaysResult": { - "base": "

Contains the output of DescribeNatGateways.

", - "refs": { - } - }, - "DescribeNetworkAclsRequest": { - "base": "

Contains the parameters for DescribeNetworkAcls.

", - "refs": { - } - }, - "DescribeNetworkAclsResult": { - "base": "

Contains the output of DescribeNetworkAcls.

", - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for DescribeNetworkInterfaceAttribute.

", - "refs": { - } - }, - "DescribeNetworkInterfaceAttributeResult": { - "base": "

Contains the output of DescribeNetworkInterfaceAttribute.

", - "refs": { - } - }, - "DescribeNetworkInterfacesRequest": { - "base": "

Contains the parameters for DescribeNetworkInterfaces.

", - "refs": { - } - }, - "DescribeNetworkInterfacesResult": { - "base": "

Contains the output of DescribeNetworkInterfaces.

", - "refs": { - } - }, - "DescribePlacementGroupsRequest": { - "base": "

Contains the parameters for DescribePlacementGroups.

", - "refs": { - } - }, - "DescribePlacementGroupsResult": { - "base": "

Contains the output of DescribePlacementGroups.

", - "refs": { - } - }, - "DescribePrefixListsRequest": { - "base": "

Contains the parameters for DescribePrefixLists.

", - "refs": { - } - }, - "DescribePrefixListsResult": { - "base": "

Contains the output of DescribePrefixLists.

", - "refs": { - } - }, - "DescribeRegionsRequest": { - "base": "

Contains the parameters for DescribeRegions.

", - "refs": { - } - }, - "DescribeRegionsResult": { - "base": "

Contains the output of DescribeRegions.

", - "refs": { - } - }, - "DescribeReservedInstancesListingsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesListings.

", - "refs": { - } - }, - "DescribeReservedInstancesListingsResult": { - "base": "

Contains the output of DescribeReservedInstancesListings.

", - "refs": { - } - }, - "DescribeReservedInstancesModificationsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesModifications.

", - "refs": { - } - }, - "DescribeReservedInstancesModificationsResult": { - "base": "

Contains the output of DescribeReservedInstancesModifications.

", - "refs": { - } - }, - "DescribeReservedInstancesOfferingsRequest": { - "base": "

Contains the parameters for DescribeReservedInstancesOfferings.

", - "refs": { - } - }, - "DescribeReservedInstancesOfferingsResult": { - "base": "

Contains the output of DescribeReservedInstancesOfferings.

", - "refs": { - } - }, - "DescribeReservedInstancesRequest": { - "base": "

Contains the parameters for DescribeReservedInstances.

", - "refs": { - } - }, - "DescribeReservedInstancesResult": { - "base": "

Contains the output for DescribeReservedInstances.

", - "refs": { - } - }, - "DescribeRouteTablesRequest": { - "base": "

Contains the parameters for DescribeRouteTables.

", - "refs": { - } - }, - "DescribeRouteTablesResult": { - "base": "

Contains the output of DescribeRouteTables.

", - "refs": { - } - }, - "DescribeScheduledInstanceAvailabilityRequest": { - "base": "

Contains the parameters for DescribeScheduledInstanceAvailability.

", - "refs": { - } - }, - "DescribeScheduledInstanceAvailabilityResult": { - "base": "

Contains the output of DescribeScheduledInstanceAvailability.

", - "refs": { - } - }, - "DescribeScheduledInstancesRequest": { - "base": "

Contains the parameters for DescribeScheduledInstances.

", - "refs": { - } - }, - "DescribeScheduledInstancesResult": { - "base": "

Contains the output of DescribeScheduledInstances.

", - "refs": { - } - }, - "DescribeSecurityGroupReferencesRequest": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupReferencesResult": { - "base": null, - "refs": { - } - }, - "DescribeSecurityGroupsRequest": { - "base": "

Contains the parameters for DescribeSecurityGroups.

", - "refs": { - } - }, - "DescribeSecurityGroupsResult": { - "base": "

Contains the output of DescribeSecurityGroups.

", - "refs": { - } - }, - "DescribeSnapshotAttributeRequest": { - "base": "

Contains the parameters for DescribeSnapshotAttribute.

", - "refs": { - } - }, - "DescribeSnapshotAttributeResult": { - "base": "

Contains the output of DescribeSnapshotAttribute.

", - "refs": { - } - }, - "DescribeSnapshotsRequest": { - "base": "

Contains the parameters for DescribeSnapshots.

", - "refs": { - } - }, - "DescribeSnapshotsResult": { - "base": "

Contains the output of DescribeSnapshots.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionRequest": { - "base": "

Contains the parameters for DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotDatafeedSubscriptionResult": { - "base": "

Contains the output of DescribeSpotDatafeedSubscription.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesRequest": { - "base": "

Contains the parameters for DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetInstancesResponse": { - "base": "

Contains the output of DescribeSpotFleetInstances.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestHistoryResponse": { - "base": "

Contains the output of DescribeSpotFleetRequestHistory.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotFleetRequestsResponse": { - "base": "

Contains the output of DescribeSpotFleetRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsRequest": { - "base": "

Contains the parameters for DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotInstanceRequestsResult": { - "base": "

Contains the output of DescribeSpotInstanceRequests.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryRequest": { - "base": "

Contains the parameters for DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeSpotPriceHistoryResult": { - "base": "

Contains the output of DescribeSpotPriceHistory.

", - "refs": { - } - }, - "DescribeStaleSecurityGroupsRequest": { - "base": null, - "refs": { - } - }, - "DescribeStaleSecurityGroupsResult": { - "base": null, - "refs": { - } - }, - "DescribeSubnetsRequest": { - "base": "

Contains the parameters for DescribeSubnets.

", - "refs": { - } - }, - "DescribeSubnetsResult": { - "base": "

Contains the output of DescribeSubnets.

", - "refs": { - } - }, - "DescribeTagsRequest": { - "base": "

Contains the parameters for DescribeTags.

", - "refs": { - } - }, - "DescribeTagsResult": { - "base": "

Contains the output of DescribeTags.

", - "refs": { - } - }, - "DescribeVolumeAttributeRequest": { - "base": "

Contains the parameters for DescribeVolumeAttribute.

", - "refs": { - } - }, - "DescribeVolumeAttributeResult": { - "base": "

Contains the output of DescribeVolumeAttribute.

", - "refs": { - } - }, - "DescribeVolumeStatusRequest": { - "base": "

Contains the parameters for DescribeVolumeStatus.

", - "refs": { - } - }, - "DescribeVolumeStatusResult": { - "base": "

Contains the output of DescribeVolumeStatus.

", - "refs": { - } - }, - "DescribeVolumesRequest": { - "base": "

Contains the parameters for DescribeVolumes.

", - "refs": { - } - }, - "DescribeVolumesResult": { - "base": "

Contains the output of DescribeVolumes.

", - "refs": { - } - }, - "DescribeVpcAttributeRequest": { - "base": "

Contains the parameters for DescribeVpcAttribute.

", - "refs": { - } - }, - "DescribeVpcAttributeResult": { - "base": "

Contains the output of DescribeVpcAttribute.

", - "refs": { - } - }, - "DescribeVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for DescribeVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DescribeVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of DescribeVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DescribeVpcClassicLinkRequest": { - "base": "

Contains the parameters for DescribeVpcClassicLink.

", - "refs": { - } - }, - "DescribeVpcClassicLinkResult": { - "base": "

Contains the output of DescribeVpcClassicLink.

", - "refs": { - } - }, - "DescribeVpcEndpointServicesRequest": { - "base": "

Contains the parameters for DescribeVpcEndpointServices.

", - "refs": { - } - }, - "DescribeVpcEndpointServicesResult": { - "base": "

Contains the output of DescribeVpcEndpointServices.

", - "refs": { - } - }, - "DescribeVpcEndpointsRequest": { - "base": "

Contains the parameters for DescribeVpcEndpoints.

", - "refs": { - } - }, - "DescribeVpcEndpointsResult": { - "base": "

Contains the output of DescribeVpcEndpoints.

", - "refs": { - } - }, - "DescribeVpcPeeringConnectionsRequest": { - "base": "

Contains the parameters for DescribeVpcPeeringConnections.

", - "refs": { - } - }, - "DescribeVpcPeeringConnectionsResult": { - "base": "

Contains the output of DescribeVpcPeeringConnections.

", - "refs": { - } - }, - "DescribeVpcsRequest": { - "base": "

Contains the parameters for DescribeVpcs.

", - "refs": { - } - }, - "DescribeVpcsResult": { - "base": "

Contains the output of DescribeVpcs.

", - "refs": { - } - }, - "DescribeVpnConnectionsRequest": { - "base": "

Contains the parameters for DescribeVpnConnections.

", - "refs": { - } - }, - "DescribeVpnConnectionsResult": { - "base": "

Contains the output of DescribeVpnConnections.

", - "refs": { - } - }, - "DescribeVpnGatewaysRequest": { - "base": "

Contains the parameters for DescribeVpnGateways.

", - "refs": { - } - }, - "DescribeVpnGatewaysResult": { - "base": "

Contains the output of DescribeVpnGateways.

", - "refs": { - } - }, - "DetachClassicLinkVpcRequest": { - "base": "

Contains the parameters for DetachClassicLinkVpc.

", - "refs": { - } - }, - "DetachClassicLinkVpcResult": { - "base": "

Contains the output of DetachClassicLinkVpc.

", - "refs": { - } - }, - "DetachInternetGatewayRequest": { - "base": "

Contains the parameters for DetachInternetGateway.

", - "refs": { - } - }, - "DetachNetworkInterfaceRequest": { - "base": "

Contains the parameters for DetachNetworkInterface.

", - "refs": { - } - }, - "DetachVolumeRequest": { - "base": "

Contains the parameters for DetachVolume.

", - "refs": { - } - }, - "DetachVpnGatewayRequest": { - "base": "

Contains the parameters for DetachVpnGateway.

", - "refs": { - } - }, - "DeviceType": { - "base": null, - "refs": { - "Image$RootDeviceType": "

The type of root device used by the AMI. The AMI can use an EBS volume or an instance store volume.

", - "Instance$RootDeviceType": "

The root device type used by the AMI. The AMI can use an EBS volume or an instance store volume.

" - } - }, - "DhcpConfiguration": { - "base": "

Describes a DHCP configuration option.

", - "refs": { - "DhcpConfigurationList$member": null - } - }, - "DhcpConfigurationList": { - "base": null, - "refs": { - "DhcpOptions$DhcpConfigurations": "

One or more DHCP options in the set.

" - } - }, - "DhcpConfigurationValueList": { - "base": null, - "refs": { - "DhcpConfiguration$Values": "

One or more values for the DHCP option.

" - } - }, - "DhcpOptions": { - "base": "

Describes a set of DHCP options.

", - "refs": { - "CreateDhcpOptionsResult$DhcpOptions": "

A set of DHCP options.

", - "DhcpOptionsList$member": null - } - }, - "DhcpOptionsIdStringList": { - "base": null, - "refs": { - "DescribeDhcpOptionsRequest$DhcpOptionsIds": "

The IDs of one or more DHCP options sets.

Default: Describes all your DHCP options sets.

" - } - }, - "DhcpOptionsList": { - "base": null, - "refs": { - "DescribeDhcpOptionsResult$DhcpOptions": "

Information about one or more DHCP options sets.

" - } - }, - "DisableVgwRoutePropagationRequest": { - "base": "

Contains the parameters for DisableVgwRoutePropagation.

", - "refs": { - } - }, - "DisableVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for DisableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DisableVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of DisableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "DisableVpcClassicLinkRequest": { - "base": "

Contains the parameters for DisableVpcClassicLink.

", - "refs": { - } - }, - "DisableVpcClassicLinkResult": { - "base": "

Contains the output of DisableVpcClassicLink.

", - "refs": { - } - }, - "DisassociateAddressRequest": { - "base": "

Contains the parameters for DisassociateAddress.

", - "refs": { - } - }, - "DisassociateRouteTableRequest": { - "base": "

Contains the parameters for DisassociateRouteTable.

", - "refs": { - } - }, - "DiskImage": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImageList$member": null - } - }, - "DiskImageDescription": { - "base": "

Describes a disk image.

", - "refs": { - "ImportInstanceVolumeDetailItem$Image": "

The image.

", - "ImportVolumeTaskDetails$Image": "

The image.

" - } - }, - "DiskImageDetail": { - "base": "

Describes a disk image.

", - "refs": { - "DiskImage$Image": "

Information about the disk image.

", - "ImportVolumeRequest$Image": "

The disk image.

" - } - }, - "DiskImageFormat": { - "base": null, - "refs": { - "DiskImageDescription$Format": "

The disk image format.

", - "DiskImageDetail$Format": "

The disk image format.

", - "ExportToS3Task$DiskImageFormat": "

The format for the exported image.

", - "ExportToS3TaskSpecification$DiskImageFormat": "

The format for the exported image.

" - } - }, - "DiskImageList": { - "base": null, - "refs": { - "ImportInstanceRequest$DiskImages": "

The disk image.

" - } - }, - "DiskImageVolumeDescription": { - "base": "

Describes a disk image volume.

", - "refs": { - "ImportInstanceVolumeDetailItem$Volume": "

The volume.

", - "ImportVolumeTaskDetails$Volume": "

The volume.

" - } - }, - "DomainType": { - "base": null, - "refs": { - "Address$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

", - "AllocateAddressRequest$Domain": "

Set to vpc to allocate the address for use with instances in a VPC.

Default: The address is for use with instances in EC2-Classic.

", - "AllocateAddressResult$Domain": "

Indicates whether this Elastic IP address is for use with instances in EC2-Classic (standard) or instances in a VPC (vpc).

" - } - }, - "Double": { - "base": null, - "refs": { - "ClientData$UploadSize": "

The size of the uploaded disk image, in GiB.

", - "PriceSchedule$Price": "

The fixed price for the term.

", - "PriceScheduleSpecification$Price": "

The fixed price for the term.

", - "PricingDetail$Price": "

The price per instance.

", - "RecurringCharge$Amount": "

The amount of the recurring charge.

", - "ReservedInstanceLimitPrice$Amount": "

Used for Reserved Instance Marketplace offerings. Specifies the limit price on the total order (instanceCount * price).

", - "SnapshotDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SnapshotTaskDetail$DiskImageSize": "

The size of the disk in the snapshot, in GiB.

", - "SpotFleetLaunchSpecification$WeightedCapacity": "

The number of units provided by the specified instance type. These are the same units that you chose to set the target capacity in terms (instances or a performance characteristic such as vCPUs, memory, or I/O).

If the target capacity divided by this value is not a whole number, we round the number of instances to the next whole number. If this value is not specified, the default is 1.

", - "SpotFleetRequestConfigData$FulfilledCapacity": "

The number of units fulfilled by this request compared to the set target capacity.

" - } - }, - "EbsBlockDevice": { - "base": "

Describes a block device for an EBS volume.

", - "refs": { - "BlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDevice": { - "base": "

Describes a parameter used to set up an EBS volume in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMapping$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EbsInstanceBlockDeviceSpecification": { - "base": "

Describes information used to set up an EBS volume specified in a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingSpecification$Ebs": "

Parameters used to automatically set up EBS volumes when the instance is launched.

" - } - }, - "EnableVgwRoutePropagationRequest": { - "base": "

Contains the parameters for EnableVgwRoutePropagation.

", - "refs": { - } - }, - "EnableVolumeIORequest": { - "base": "

Contains the parameters for EnableVolumeIO.

", - "refs": { - } - }, - "EnableVpcClassicLinkDnsSupportRequest": { - "base": "

Contains the parameters for EnableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "EnableVpcClassicLinkDnsSupportResult": { - "base": "

Contains the output of EnableVpcClassicLinkDnsSupport.

", - "refs": { - } - }, - "EnableVpcClassicLinkRequest": { - "base": "

Contains the parameters for EnableVpcClassicLink.

", - "refs": { - } - }, - "EnableVpcClassicLinkResult": { - "base": "

Contains the output of EnableVpcClassicLink.

", - "refs": { - } - }, - "EventCode": { - "base": null, - "refs": { - "InstanceStatusEvent$Code": "

The event code.

" - } - }, - "EventInformation": { - "base": "

Describes a Spot fleet event.

", - "refs": { - "HistoryRecord$EventInformation": "

Information about the event.

" - } - }, - "EventType": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryRequest$EventType": "

The type of events to describe. By default, all events are described.

", - "HistoryRecord$EventType": "

The event type.

  • error - Indicates an error with the Spot fleet request.

  • fleetRequestChange - Indicates a change in the status or configuration of the Spot fleet request.

  • instanceChange - Indicates that an instance was launched or terminated.

" - } - }, - "ExcessCapacityTerminationPolicy": { - "base": null, - "refs": { - "ModifySpotFleetRequestRequest$ExcessCapacityTerminationPolicy": "

Indicates whether running Spot instances should be terminated if the target capacity of the Spot fleet request is decreased below the current size of the Spot fleet.

", - "SpotFleetRequestConfigData$ExcessCapacityTerminationPolicy": "

Indicates whether running Spot instances should be terminated if the target capacity of the Spot fleet request is decreased below the current size of the Spot fleet.

" - } - }, - "ExecutableByStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ExecutableUsers": "

Scopes the images by users with explicit launch permissions. Specify an AWS account ID, self (the sender of the request), or all (public AMIs).

" - } - }, - "ExportEnvironment": { - "base": null, - "refs": { - "CreateInstanceExportTaskRequest$TargetEnvironment": "

The target virtualization environment.

", - "InstanceExportDetails$TargetEnvironment": "

The target virtualization environment.

" - } - }, - "ExportTask": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskResult$ExportTask": "

Information about the instance export task.

", - "ExportTaskList$member": null - } - }, - "ExportTaskIdStringList": { - "base": null, - "refs": { - "DescribeExportTasksRequest$ExportTaskIds": "

One or more export task IDs.

" - } - }, - "ExportTaskList": { - "base": null, - "refs": { - "DescribeExportTasksResult$ExportTasks": "

Information about the export tasks.

" - } - }, - "ExportTaskState": { - "base": null, - "refs": { - "ExportTask$State": "

The state of the export task.

" - } - }, - "ExportToS3Task": { - "base": "

Describes the format and location for an instance export task.

", - "refs": { - "ExportTask$ExportToS3Task": "

Information about the export task.

" - } - }, - "ExportToS3TaskSpecification": { - "base": "

Describes an instance export task.

", - "refs": { - "CreateInstanceExportTaskRequest$ExportToS3Task": "

The format and location for an instance export task.

" - } - }, - "Filter": { - "base": "

A filter name and value pair that is used to return a more specific list of results. Filters can be used to match a set of resources by various criteria, such as tags, attributes, or IDs.

", - "refs": { - "FilterList$member": null - } - }, - "FilterList": { - "base": null, - "refs": { - "DescribeAddressesRequest$Filters": "

One or more filters. Filter names and values are case-sensitive.

  • allocation-id - [EC2-VPC] The allocation ID for the address.

  • association-id - [EC2-VPC] The association ID for the address.

  • domain - Indicates whether the address is for use in EC2-Classic (standard) or in a VPC (vpc).

  • instance-id - The ID of the instance the address is associated with, if any.

  • network-interface-id - [EC2-VPC] The ID of the network interface that the address is associated with, if any.

  • network-interface-owner-id - The AWS account ID of the owner.

  • private-ip-address - [EC2-VPC] The private IP address associated with the Elastic IP address.

  • public-ip - The Elastic IP address.

", - "DescribeAvailabilityZonesRequest$Filters": "

One or more filters.

  • message - Information about the Availability Zone.

  • region-name - The name of the region for the Availability Zone (for example, us-east-1).

  • state - The state of the Availability Zone (available | information | impaired | unavailable).

  • zone-name - The name of the Availability Zone (for example, us-east-1a).

", - "DescribeBundleTasksRequest$Filters": "

One or more filters.

  • bundle-id - The ID of the bundle task.

  • error-code - If the task failed, the error code returned.

  • error-message - If the task failed, the error message returned.

  • instance-id - The ID of the instance.

  • progress - The level of task completion, as a percentage (for example, 20%).

  • s3-bucket - The Amazon S3 bucket to store the AMI.

  • s3-prefix - The beginning of the AMI name.

  • start-time - The time the task started (for example, 2013-09-15T17:15:20.000Z).

  • state - The state of the task (pending | waiting-for-shutdown | bundling | storing | cancelling | complete | failed).

  • update-time - The time of the most recent update for the task.

", - "DescribeClassicLinkInstancesRequest$Filters": "

One or more filters.

  • group-id - The ID of a VPC security group that's associated with the instance.

  • instance-id - The ID of the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC that the instance is linked to.

", - "DescribeCustomerGatewaysRequest$Filters": "

One or more filters.

  • bgp-asn - The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

  • customer-gateway-id - The ID of the customer gateway.

  • ip-address - The IP address of the customer gateway's Internet-routable external interface.

  • state - The state of the customer gateway (pending | available | deleting | deleted).

  • type - The type of customer gateway. Currently, the only supported type is ipsec.1.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeDhcpOptionsRequest$Filters": "

One or more filters.

  • dhcp-options-id - The ID of a set of DHCP options.

  • key - The key for one of the options (for example, domain-name).

  • value - The value for one of the options.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeFlowLogsRequest$Filter": "

One or more filters.

  • deliver-log-status - The status of the logs delivery (SUCCESS | FAILED).

  • flow-log-id - The ID of the flow log.

  • log-group-name - The name of the log group.

  • resource-id - The ID of the VPC, subnet, or network interface.

  • traffic-type - The type of traffic (ACCEPT | REJECT | ALL)

", - "DescribeHostReservationOfferingsRequest$Filter": "

One or more filters.

  • instance-family - The instance family of the offering (e.g., m4).

  • payment-option - The payment option (No Upfront | Partial Upfront | All Upfront).

", - "DescribeHostReservationsRequest$Filter": "

One or more filters.

  • instance-family - The instance family (e.g., m4).

  • payment-option - The payment option (No Upfront | Partial Upfront | All Upfront).

  • state - The state of the reservation (payment-pending | payment-failed | active | retired).

", - "DescribeHostsRequest$Filter": "

One or more filters.

  • instance-type - The instance type size that the Dedicated Host is configured to support.

  • auto-placement - Whether auto-placement is enabled or disabled (on | off).

  • host-reservation-id - The ID of the reservation assigned to this host.

  • client-token - The idempotency token you provided when you launched the instance

  • state- The allocation state of the Dedicated Host (available | under-assessment | permanent-failure | released | released-permanent-failure).

  • availability-zone - The Availability Zone of the host.

", - "DescribeImagesRequest$Filters": "

One or more filters.

  • architecture - The image architecture (i386 | x86_64).

  • block-device-mapping.delete-on-termination - A Boolean value that indicates whether the Amazon EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh).

  • block-device-mapping.snapshot-id - The ID of the snapshot used for the EBS volume.

  • block-device-mapping.volume-size - The volume size of the EBS volume, in GiB.

  • block-device-mapping.volume-type - The volume type of the EBS volume (gp2 | io1 | st1 | sc1 | standard).

  • description - The description of the image (provided during image creation).

  • hypervisor - The hypervisor type (ovm | xen).

  • image-id - The ID of the image.

  • image-type - The image type (machine | kernel | ramdisk).

  • is-public - A Boolean that indicates whether the image is public.

  • kernel-id - The kernel ID.

  • manifest-location - The location of the image manifest.

  • name - The name of the AMI (provided during image creation).

  • owner-alias - String value from an Amazon-maintained list (amazon | aws-marketplace | microsoft) of snapshot owners. Not to be confused with the user-configured AWS account alias, which is set from the IAM console.

  • owner-id - The AWS account ID of the image owner.

  • platform - The platform. To only list Windows-based AMIs, use windows.

  • product-code - The product code.

  • product-code.type - The type of the product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • root-device-name - The name of the root device volume (for example, /dev/sda1).

  • root-device-type - The type of the root device volume (ebs | instance-store).

  • state - The state of the image (available | pending | failed).

  • state-reason-code - The reason code for the state change.

  • state-reason-message - The message for the state change.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • virtualization-type - The virtualization type (paravirtual | hvm).

", - "DescribeImportImageTasksRequest$Filters": "

Filter tasks using the task-state filter and one of the following values: active, completed, deleting, deleted.

", - "DescribeImportSnapshotTasksRequest$Filters": "

One or more filters.

", - "DescribeInstanceStatusRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone of the instance.

  • event.code - The code for the scheduled event (instance-reboot | system-reboot | system-maintenance | instance-retirement | instance-stop).

  • event.description - A description of the event.

  • event.not-after - The latest end time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • event.not-before - The earliest start time for the scheduled event (for example, 2014-09-15T17:15:20.000Z).

  • instance-state-code - The code for the instance state, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-status.reachability - Filters on instance status where the name is reachability (passed | failed | initializing | insufficient-data).

  • instance-status.status - The status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

  • system-status.reachability - Filters on system status where the name is reachability (passed | failed | initializing | insufficient-data).

  • system-status.status - The system status of the instance (ok | impaired | initializing | insufficient-data | not-applicable).

", - "DescribeInstancesRequest$Filters": "

One or more filters.

  • affinity - The affinity setting for an instance running on a Dedicated Host (default | host).

  • architecture - The instance architecture (i386 | x86_64).

  • availability-zone - The Availability Zone of the instance.

  • block-device-mapping.attach-time - The attach time for an EBS volume mapped to the instance, for example, 2010-09-15T17:15:20.000Z.

  • block-device-mapping.delete-on-termination - A Boolean that indicates whether the EBS volume is deleted on instance termination.

  • block-device-mapping.device-name - The device name for the EBS volume (for example, /dev/sdh or xvdh).

  • block-device-mapping.status - The status for the EBS volume (attaching | attached | detaching | detached).

  • block-device-mapping.volume-id - The volume ID of the EBS volume.

  • client-token - The idempotency token you provided when you launched the instance.

  • dns-name - The public DNS name of the instance.

  • group-id - The ID of the security group for the instance. EC2-Classic only.

  • group-name - The name of the security group for the instance. EC2-Classic only.

  • host-id - The ID of the Dedicated Host on which the instance is running, if applicable.

  • hypervisor - The hypervisor type of the instance (ovm | xen).

  • iam-instance-profile.arn - The instance profile associated with the instance. Specified as an ARN.

  • image-id - The ID of the image used to launch the instance.

  • instance-id - The ID of the instance.

  • instance-lifecycle - Indicates whether this is a Spot Instance or a Scheduled Instance (spot | scheduled).

  • instance-state-code - The state of the instance, as a 16-bit unsigned integer. The high byte is an opaque internal value and should be ignored. The low byte is set based on the state represented. The valid values are: 0 (pending), 16 (running), 32 (shutting-down), 48 (terminated), 64 (stopping), and 80 (stopped).

  • instance-state-name - The state of the instance (pending | running | shutting-down | terminated | stopping | stopped).

  • instance-type - The type of instance (for example, t2.micro).

  • instance.group-id - The ID of the security group for the instance.

  • instance.group-name - The name of the security group for the instance.

  • ip-address - The public IP address of the instance.

  • kernel-id - The kernel ID.

  • key-name - The name of the key pair used when the instance was launched.

  • launch-index - When launching multiple instances, this is the index for the instance in the launch group (for example, 0, 1, 2, and so on).

  • launch-time - The time when the instance was launched.

  • monitoring-state - Indicates whether monitoring is enabled for the instance (disabled | enabled).

  • owner-id - The AWS account ID of the instance owner.

  • placement-group-name - The name of the placement group for the instance.

  • platform - The platform. Use windows if you have Windows instances; otherwise, leave blank.

  • private-dns-name - The private DNS name of the instance.

  • private-ip-address - The private IP address of the instance.

  • product-code - The product code associated with the AMI used to launch the instance.

  • product-code.type - The type of product code (devpay | marketplace).

  • ramdisk-id - The RAM disk ID.

  • reason - The reason for the current state of the instance (for example, shows \"User Initiated [date]\" when you stop or terminate the instance). Similar to the state-reason-code filter.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • reservation-id - The ID of the instance's reservation. A reservation ID is created any time you launch an instance. A reservation ID has a one-to-one relationship with an instance launch request, but can be associated with more than one instance if you launch multiple instances using the same launch request. For example, if you launch one instance, you'll get one reservation ID. If you launch ten instances using the same launch request, you'll also get one reservation ID.

  • root-device-name - The name of the root device for the instance (for example, /dev/sda1 or /dev/xvda).

  • root-device-type - The type of root device that the instance uses (ebs | instance-store).

  • source-dest-check - Indicates whether the instance performs source/destination checking. A value of true means that checking is enabled, and false means checking is disabled. The value must be false for the instance to perform network address translation (NAT) in your VPC.

  • spot-instance-request-id - The ID of the Spot instance request.

  • state-reason-code - The reason code for the state change.

  • state-reason-message - A message that describes the state change.

  • subnet-id - The ID of the subnet for the instance.

  • tag:key=value - The key/value combination of a tag assigned to the resource, where tag:key is the tag's key.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • tenancy - The tenancy of an instance (dedicated | default | host).

  • virtualization-type - The virtualization type of the instance (paravirtual | hvm).

  • vpc-id - The ID of the VPC that the instance is running in.

  • network-interface.description - The description of the network interface.

  • network-interface.subnet-id - The ID of the subnet for the network interface.

  • network-interface.vpc-id - The ID of the VPC for the network interface.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.owner-id - The ID of the owner of the network interface.

  • network-interface.availability-zone - The Availability Zone for the network interface.

  • network-interface.requester-id - The requester ID for the network interface.

  • network-interface.requester-managed - Indicates whether the network interface is being managed by AWS.

  • network-interface.status - The status of the network interface (available) | in-use).

  • network-interface.mac-address - The MAC address of the network interface.

  • network-interface.private-dns-name - The private DNS name of the network interface.

  • network-interface.source-dest-check - Whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • network-interface.group-id - The ID of a security group associated with the network interface.

  • network-interface.group-name - The name of a security group associated with the network interface.

  • network-interface.attachment.attachment-id - The ID of the interface attachment.

  • network-interface.attachment.instance-id - The ID of the instance to which the network interface is attached.

  • network-interface.attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • network-interface.addresses.private-ip-address - The private IP address associated with the network interface.

  • network-interface.attachment.device-index - The device index to which the network interface is attached.

  • network-interface.attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • network-interface.attachment.attach-time - The time that the network interface was attached to an instance.

  • network-interface.attachment.delete-on-termination - Specifies whether the attachment is deleted when an instance is terminated.

  • network-interface.addresses.primary - Specifies whether the IP address of the network interface is the primary private IP address.

  • network-interface.addresses.association.public-ip - The ID of the association of an Elastic IP address with a network interface.

  • network-interface.addresses.association.ip-owner-id - The owner ID of the private IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

", - "DescribeInternetGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (available). Present only if a VPC is attached.

  • attachment.vpc-id - The ID of an attached VPC.

  • internet-gateway-id - The ID of the Internet gateway.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeKeyPairsRequest$Filters": "

One or more filters.

  • fingerprint - The fingerprint of the key pair.

  • key-name - The name of the key pair.

", - "DescribeMovingAddressesRequest$Filters": "

One or more filters.

  • moving-status - The status of the Elastic IP address (MovingToVpc | RestoringToClassic).

", - "DescribeNatGatewaysRequest$Filter": "

One or more filters.

  • nat-gateway-id - The ID of the NAT gateway.

  • state - The state of the NAT gateway (pending | failed | available | deleting | deleted).

  • subnet-id - The ID of the subnet in which the NAT gateway resides.

  • vpc-id - The ID of the VPC in which the NAT gateway resides.

", - "DescribeNetworkAclsRequest$Filters": "

One or more filters.

  • association.association-id - The ID of an association ID for the ACL.

  • association.network-acl-id - The ID of the network ACL involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • default - Indicates whether the ACL is the default network ACL for the VPC.

  • entry.cidr - The CIDR range specified in the entry.

  • entry.egress - Indicates whether the entry applies to egress traffic.

  • entry.icmp.code - The ICMP code specified in the entry, if any.

  • entry.icmp.type - The ICMP type specified in the entry, if any.

  • entry.port-range.from - The start of the port range specified in the entry.

  • entry.port-range.to - The end of the port range specified in the entry.

  • entry.protocol - The protocol specified in the entry (tcp | udp | icmp or a protocol number).

  • entry.rule-action - Allows or denies the matching traffic (allow | deny).

  • entry.rule-number - The number of an entry (in other words, rule) in the ACL's set of entries.

  • network-acl-id - The ID of the network ACL.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network ACL.

", - "DescribeNetworkInterfacesRequest$Filters": "

One or more filters.

  • addresses.private-ip-address - The private IP addresses associated with the network interface.

  • addresses.primary - Whether the private IP address is the primary IP address associated with the network interface.

  • addresses.association.public-ip - The association ID returned when the network interface was associated with the Elastic IP address.

  • addresses.association.owner-id - The owner ID of the addresses associated with the network interface.

  • association.association-id - The association ID returned when the network interface was associated with an IP address.

  • association.allocation-id - The allocation ID returned when you allocated the Elastic IP address for your network interface.

  • association.ip-owner-id - The owner of the Elastic IP address associated with the network interface.

  • association.public-ip - The address of the Elastic IP address bound to the network interface.

  • association.public-dns-name - The public DNS name for the network interface.

  • attachment.attachment-id - The ID of the interface attachment.

  • attachment.attach.time - The time that the network interface was attached to an instance.

  • attachment.delete-on-termination - Indicates whether the attachment is deleted when an instance is terminated.

  • attachment.device-index - The device index to which the network interface is attached.

  • attachment.instance-id - The ID of the instance to which the network interface is attached.

  • attachment.instance-owner-id - The owner ID of the instance to which the network interface is attached.

  • attachment.nat-gateway-id - The ID of the NAT gateway to which the network interface is attached.

  • attachment.status - The status of the attachment (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone of the network interface.

  • description - The description of the network interface.

  • group-id - The ID of a security group associated with the network interface.

  • group-name - The name of a security group associated with the network interface.

  • mac-address - The MAC address of the network interface.

  • network-interface-id - The ID of the network interface.

  • owner-id - The AWS account ID of the network interface owner.

  • private-ip-address - The private IP address or addresses of the network interface.

  • private-dns-name - The private DNS name of the network interface.

  • requester-id - The ID of the entity that launched the instance on your behalf (for example, AWS Management Console, Auto Scaling, and so on).

  • requester-managed - Indicates whether the network interface is being managed by an AWS service (for example, AWS Management Console, Auto Scaling, and so on).

  • source-desk-check - Indicates whether the network interface performs source/destination checking. A value of true means checking is enabled, and false means checking is disabled. The value must be false for the network interface to perform network address translation (NAT) in your VPC.

  • status - The status of the network interface. If the network interface is not attached to an instance, the status is available; if a network interface is attached to an instance the status is in-use.

  • subnet-id - The ID of the subnet for the network interface.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the network interface.

", - "DescribePlacementGroupsRequest$Filters": "

One or more filters.

  • group-name - The name of the placement group.

  • state - The state of the placement group (pending | available | deleting | deleted).

  • strategy - The strategy of the placement group (cluster).

", - "DescribePrefixListsRequest$Filters": "

One or more filters.

  • prefix-list-id: The ID of a prefix list.

  • prefix-list-name: The name of a prefix list.

", - "DescribeRegionsRequest$Filters": "

One or more filters.

  • endpoint - The endpoint of the region (for example, ec2.us-east-1.amazonaws.com).

  • region-name - The name of the region (for example, us-east-1).

", - "DescribeReservedInstancesListingsRequest$Filters": "

One or more filters.

  • reserved-instances-id - The ID of the Reserved Instances.

  • reserved-instances-listing-id - The ID of the Reserved Instances listing.

  • status - The status of the Reserved Instance listing (pending | active | cancelled | closed).

  • status-message - The reason for the status.

", - "DescribeReservedInstancesModificationsRequest$Filters": "

One or more filters.

  • client-token - The idempotency token for the modification request.

  • create-date - The time when the modification request was created.

  • effective-date - The time when the modification becomes effective.

  • modification-result.reserved-instances-id - The ID for the Reserved Instances created as part of the modification request. This ID is only available when the status of the modification is fulfilled.

  • modification-result.target-configuration.availability-zone - The Availability Zone for the new Reserved Instances.

  • modification-result.target-configuration.instance-count - The number of new Reserved Instances.

  • modification-result.target-configuration.instance-type - The instance type of the new Reserved Instances.

  • modification-result.target-configuration.platform - The network platform of the new Reserved Instances (EC2-Classic | EC2-VPC).

  • reserved-instances-id - The ID of the Reserved Instances modified.

  • reserved-instances-modification-id - The ID of the modification request.

  • status - The status of the Reserved Instances modification request (processing | fulfilled | failed).

  • status-message - The reason for the status.

  • update-date - The time when the modification request was last updated.

", - "DescribeReservedInstancesOfferingsRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (for example, one year or three years), in seconds (31536000 | 94608000).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type that is covered by the reservation.

  • marketplace - Set to true to show only Reserved Instance Marketplace offerings. When this filter is not used, which is the default behavior, all offerings from both AWS and the Reserved Instance Marketplace are listed.

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC. (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC))

  • reserved-instances-offering-id - The Reserved Instances offering ID.

  • scope - The scope of the Reserved Instance (Availability Zone or Region).

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeReservedInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone where the Reserved Instance can be used.

  • duration - The duration of the Reserved Instance (one year or three years), in seconds (31536000 | 94608000).

  • end - The time when the Reserved Instance expires (for example, 2015-08-07T11:54:42.000Z).

  • fixed-price - The purchase price of the Reserved Instance (for example, 9800.0).

  • instance-type - The instance type that is covered by the reservation.

  • scope - The scope of the Reserved Instance (Region or Availability Zone).

  • product-description - The Reserved Instance product platform description. Instances that include (Amazon VPC) in the product platform description will only be displayed to EC2-Classic account holders and are for use with Amazon VPC (Linux/UNIX | Linux/UNIX (Amazon VPC) | SUSE Linux | SUSE Linux (Amazon VPC) | Red Hat Enterprise Linux | Red Hat Enterprise Linux (Amazon VPC) | Windows | Windows (Amazon VPC) | Windows with SQL Server Standard | Windows with SQL Server Standard (Amazon VPC) | Windows with SQL Server Web | Windows with SQL Server Web (Amazon VPC) | Windows with SQL Server Enterprise | Windows with SQL Server Enterprise (Amazon VPC)).

  • reserved-instances-id - The ID of the Reserved Instance.

  • start - The time at which the Reserved Instance purchase request was placed (for example, 2014-08-07T11:54:42.000Z).

  • state - The state of the Reserved Instance (payment-pending | active | payment-failed | retired).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • usage-price - The usage price of the Reserved Instance, per hour (for example, 0.84).

", - "DescribeRouteTablesRequest$Filters": "

One or more filters.

  • association.route-table-association-id - The ID of an association ID for the route table.

  • association.route-table-id - The ID of the route table involved in the association.

  • association.subnet-id - The ID of the subnet involved in the association.

  • association.main - Indicates whether the route table is the main route table for the VPC (true | false).

  • route-table-id - The ID of the route table.

  • route.destination-cidr-block - The CIDR range specified in a route in the table.

  • route.destination-prefix-list-id - The ID (prefix) of the AWS service specified in a route in the table.

  • route.gateway-id - The ID of a gateway specified in a route in the table.

  • route.instance-id - The ID of an instance specified in a route in the table.

  • route.nat-gateway-id - The ID of a NAT gateway.

  • route.origin - Describes how the route was created. CreateRouteTable indicates that the route was automatically created when the route table was created; CreateRoute indicates that the route was manually added to the route table; EnableVgwRoutePropagation indicates that the route was propagated by route propagation.

  • route.state - The state of a route in the route table (active | blackhole). The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, the specified NAT instance has been terminated, and so on).

  • route.vpc-peering-connection-id - The ID of a VPC peering connection specified in a route in the table.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the route table.

", - "DescribeScheduledInstanceAvailabilityRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone (for example, us-west-2a).

  • instance-type - The instance type (for example, c4.large).

  • network-platform - The network platform (EC2-Classic or EC2-VPC).

  • platform - The platform (Linux/UNIX or Windows).

", - "DescribeScheduledInstancesRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone (for example, us-west-2a).

  • instance-type - The instance type (for example, c4.large).

  • network-platform - The network platform (EC2-Classic or EC2-VPC).

  • platform - The platform (Linux/UNIX or Windows).

", - "DescribeSecurityGroupsRequest$Filters": "

One or more filters. If using multiple filters for rules, the results include security groups for which any combination of rules - not necessarily a single rule - match all filters.

  • description - The description of the security group.

  • egress.ip-permission.prefix-list-id - The ID (prefix) of the AWS service to which the security group allows access.

  • group-id - The ID of the security group.

  • group-name - The name of the security group.

  • ip-permission.cidr - A CIDR range that has been granted permission.

  • ip-permission.from-port - The start of port range for the TCP and UDP protocols, or an ICMP type number.

  • ip-permission.group-id - The ID of a security group that has been granted permission.

  • ip-permission.group-name - The name of a security group that has been granted permission.

  • ip-permission.protocol - The IP protocol for the permission (tcp | udp | icmp or a protocol number).

  • ip-permission.to-port - The end of port range for the TCP and UDP protocols, or an ICMP code.

  • ip-permission.user-id - The ID of an AWS account that has been granted permission.

  • owner-id - The AWS account ID of the owner of the security group.

  • tag-key - The key of a tag assigned to the security group.

  • tag-value - The value of a tag assigned to the security group.

  • vpc-id - The ID of the VPC specified when the security group was created.

", - "DescribeSnapshotsRequest$Filters": "

One or more filters.

  • description - A description of the snapshot.

  • owner-alias - Value from an Amazon-maintained list (amazon | aws-marketplace | microsoft) of snapshot owners. Not to be confused with the user-configured AWS account alias, which is set from the IAM consolew.

  • owner-id - The ID of the AWS account that owns the snapshot.

  • progress - The progress of the snapshot, as a percentage (for example, 80%).

  • snapshot-id - The snapshot ID.

  • start-time - The time stamp when the snapshot was initiated.

  • status - The status of the snapshot (pending | completed | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The ID of the volume the snapshot is for.

  • volume-size - The size of the volume, in GiB.

", - "DescribeSpotInstanceRequestsRequest$Filters": "

One or more filters.

  • availability-zone-group - The Availability Zone group.

  • create-time - The time stamp when the Spot instance request was created.

  • fault-code - The fault code related to the request.

  • fault-message - The fault message related to the request.

  • instance-id - The ID of the instance that fulfilled the request.

  • launch-group - The Spot instance launch group.

  • launch.block-device-mapping.delete-on-termination - Indicates whether the Amazon EBS volume is deleted on instance termination.

  • launch.block-device-mapping.device-name - The device name for the Amazon EBS volume (for example, /dev/sdh).

  • launch.block-device-mapping.snapshot-id - The ID of the snapshot used for the Amazon EBS volume.

  • launch.block-device-mapping.volume-size - The size of the Amazon EBS volume, in GiB.

  • launch.block-device-mapping.volume-type - The type of the Amazon EBS volume: gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1for Cold HDD, or standard for Magnetic.

  • launch.group-id - The security group for the instance.

  • launch.image-id - The ID of the AMI.

  • launch.instance-type - The type of instance (for example, m3.medium).

  • launch.kernel-id - The kernel ID.

  • launch.key-name - The name of the key pair the instance launched with.

  • launch.monitoring-enabled - Whether monitoring is enabled for the Spot instance.

  • launch.ramdisk-id - The RAM disk ID.

  • network-interface.network-interface-id - The ID of the network interface.

  • network-interface.device-index - The index of the device for the network interface attachment on the instance.

  • network-interface.subnet-id - The ID of the subnet for the instance.

  • network-interface.description - A description of the network interface.

  • network-interface.private-ip-address - The primary private IP address of the network interface.

  • network-interface.delete-on-termination - Indicates whether the network interface is deleted when the instance is terminated.

  • network-interface.group-id - The ID of the security group associated with the network interface.

  • network-interface.group-name - The name of the security group associated with the network interface.

  • network-interface.addresses.primary - Indicates whether the IP address is the primary private IP address.

  • product-description - The product description associated with the instance (Linux/UNIX | Windows).

  • spot-instance-request-id - The Spot instance request ID.

  • spot-price - The maximum hourly price for any Spot instance launched to fulfill the request.

  • state - The state of the Spot instance request (open | active | closed | cancelled | failed). Spot bid status information can help you track your Amazon EC2 Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

  • status-code - The short code describing the most recent evaluation of your Spot instance request.

  • status-message - The message explaining the status of the Spot instance request.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of Spot instance request (one-time | persistent).

  • launched-availability-zone - The Availability Zone in which the bid is launched.

  • valid-from - The start date of the request.

  • valid-until - The end date of the request.

", - "DescribeSpotPriceHistoryRequest$Filters": "

One or more filters.

  • availability-zone - The Availability Zone for which prices should be returned.

  • instance-type - The type of instance (for example, m3.medium).

  • product-description - The product description for the Spot price (Linux/UNIX | SUSE Linux | Windows | Linux/UNIX (Amazon VPC) | SUSE Linux (Amazon VPC) | Windows (Amazon VPC)).

  • spot-price - The Spot price. The value must match exactly (or use wildcards; greater than or less than comparison is not supported).

  • timestamp - The timestamp of the Spot price history, in UTC format (for example, YYYY-MM-DDTHH:MM:SSZ). You can use wildcards (* and ?). Greater than or less than comparison is not supported.

", - "DescribeSubnetsRequest$Filters": "

One or more filters.

  • availabilityZone - The Availability Zone for the subnet. You can also use availability-zone as the filter name.

  • available-ip-address-count - The number of IP addresses in the subnet that are available.

  • cidrBlock - The CIDR block of the subnet. The CIDR block you specify must exactly match the subnet's CIDR block for information to be returned for the subnet. You can also use cidr or cidr-block as the filter names.

  • defaultForAz - Indicates whether this is the default subnet for the Availability Zone. You can also use default-for-az as the filter name.

  • state - The state of the subnet (pending | available).

  • subnet-id - The ID of the subnet.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC for the subnet.

", - "DescribeTagsRequest$Filters": "

One or more filters.

  • key - The tag key.

  • resource-id - The resource ID.

  • resource-type - The resource type (customer-gateway | dhcp-options | image | instance | internet-gateway | network-acl | network-interface | reserved-instances | route-table | security-group | snapshot | spot-instances-request | subnet | volume | vpc | vpn-connection | vpn-gateway).

  • value - The tag value.

", - "DescribeVolumeStatusRequest$Filters": "

One or more filters.

  • action.code - The action code for the event (for example, enable-volume-io).

  • action.description - A description of the action.

  • action.event-id - The event ID associated with the action.

  • availability-zone - The Availability Zone of the instance.

  • event.description - A description of the event.

  • event.event-id - The event ID.

  • event.event-type - The event type (for io-enabled: passed | failed; for io-performance: io-performance:degraded | io-performance:severely-degraded | io-performance:stalled).

  • event.not-after - The latest end time for the event.

  • event.not-before - The earliest start time for the event.

  • volume-status.details-name - The cause for volume-status.status (io-enabled | io-performance).

  • volume-status.details-status - The status of volume-status.details-name (for io-enabled: passed | failed; for io-performance: normal | degraded | severely-degraded | stalled).

  • volume-status.status - The status of the volume (ok | impaired | warning | insufficient-data).

", - "DescribeVolumesRequest$Filters": "

One or more filters.

  • attachment.attach-time - The time stamp when the attachment initiated.

  • attachment.delete-on-termination - Whether the volume is deleted on instance termination.

  • attachment.device - The device name that is exposed to the instance (for example, /dev/sda1).

  • attachment.instance-id - The ID of the instance the volume is attached to.

  • attachment.status - The attachment state (attaching | attached | detaching | detached).

  • availability-zone - The Availability Zone in which the volume was created.

  • create-time - The time stamp when the volume was created.

  • encrypted - The encryption status of the volume.

  • size - The size of the volume, in GiB.

  • snapshot-id - The snapshot from which the volume was created.

  • status - The status of the volume (creating | available | in-use | deleting | deleted | error).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • volume-id - The volume ID.

  • volume-type - The Amazon EBS volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

", - "DescribeVpcClassicLinkRequest$Filters": "

One or more filters.

  • is-classic-link-enabled - Whether the VPC is enabled for ClassicLink (true | false).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

", - "DescribeVpcEndpointsRequest$Filters": "

One or more filters.

  • service-name: The name of the AWS service.

  • vpc-id: The ID of the VPC in which the endpoint resides.

  • vpc-endpoint-id: The ID of the endpoint.

  • vpc-endpoint-state: The state of the endpoint. (pending | available | deleting | deleted)

", - "DescribeVpcPeeringConnectionsRequest$Filters": "

One or more filters.

  • accepter-vpc-info.cidr-block - The CIDR block of the peer VPC.

  • accepter-vpc-info.owner-id - The AWS account ID of the owner of the peer VPC.

  • accepter-vpc-info.vpc-id - The ID of the peer VPC.

  • expiration-time - The expiration date and time for the VPC peering connection.

  • requester-vpc-info.cidr-block - The CIDR block of the requester's VPC.

  • requester-vpc-info.owner-id - The AWS account ID of the owner of the requester VPC.

  • requester-vpc-info.vpc-id - The ID of the requester VPC.

  • status-code - The status of the VPC peering connection (pending-acceptance | failed | expired | provisioning | active | deleted | rejected).

  • status-message - A message that provides more information about the status of the VPC peering connection, if applicable.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-peering-connection-id - The ID of the VPC peering connection.

", - "DescribeVpcsRequest$Filters": "

One or more filters.

  • cidr - The CIDR block of the VPC. The CIDR block you specify must exactly match the VPC's CIDR block for information to be returned for the VPC. Must contain the slash followed by one or two digits (for example, /28).

  • dhcp-options-id - The ID of a set of DHCP options.

  • isDefault - Indicates whether the VPC is the default VPC.

  • state - The state of the VPC (pending | available).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • vpc-id - The ID of the VPC.

", - "DescribeVpnConnectionsRequest$Filters": "

One or more filters.

  • customer-gateway-configuration - The configuration information for the customer gateway.

  • customer-gateway-id - The ID of a customer gateway associated with the VPN connection.

  • state - The state of the VPN connection (pending | available | deleting | deleted).

  • option.static-routes-only - Indicates whether the connection has static routes only. Used for devices that do not support Border Gateway Protocol (BGP).

  • route.destination-cidr-block - The destination CIDR block. This corresponds to the subnet used in a customer data center.

  • bgp-asn - The BGP Autonomous System Number (ASN) associated with a BGP device.

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of VPN connection. Currently the only supported type is ipsec.1.

  • vpn-connection-id - The ID of the VPN connection.

  • vpn-gateway-id - The ID of a virtual private gateway associated with the VPN connection.

", - "DescribeVpnGatewaysRequest$Filters": "

One or more filters.

  • attachment.state - The current state of the attachment between the gateway and the VPC (attaching | attached | detaching | detached).

  • attachment.vpc-id - The ID of an attached VPC.

  • availability-zone - The Availability Zone for the virtual private gateway (if applicable).

  • state - The state of the virtual private gateway (pending | available | deleting | deleted).

  • tag:key=value - The key/value combination of a tag assigned to the resource.

  • tag-key - The key of a tag assigned to the resource. This filter is independent of the tag-value filter. For example, if you use both the filter \"tag-key=Purpose\" and the filter \"tag-value=X\", you get any resources assigned both the tag key Purpose (regardless of what the tag's value is), and the tag value X (regardless of what the tag's key is). If you want to list only resources where Purpose is X, see the tag:key=value filter.

  • tag-value - The value of a tag assigned to the resource. This filter is independent of the tag-key filter.

  • type - The type of virtual private gateway. Currently the only supported type is ipsec.1.

  • vpn-gateway-id - The ID of the virtual private gateway.

" - } - }, - "FleetType": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$Type": "

The type of request. Indicates whether the fleet will only request the target capacity or also attempt to maintain it. When you request a certain target capacity, the fleet will only place the required bids. It will not attempt to replenish Spot instances if capacity is diminished, nor will it submit bids in alternative Spot pools if capacity is not available. When you want to maintain a certain target capacity, fleet will place the required bids to meet this target capacity. It will also automatically replenish any interrupted instances. Default: maintain.

" - } - }, - "Float": { - "base": null, - "refs": { - "ReservedInstances$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstances$FixedPrice": "

The purchase price of the Reserved Instance.

", - "ReservedInstancesOffering$UsagePrice": "

The usage price of the Reserved Instance, per hour.

", - "ReservedInstancesOffering$FixedPrice": "

The purchase price of the Reserved Instance.

" - } - }, - "FlowLog": { - "base": "

Describes a flow log.

", - "refs": { - "FlowLogSet$member": null - } - }, - "FlowLogSet": { - "base": null, - "refs": { - "DescribeFlowLogsResult$FlowLogs": "

Information about the flow logs.

" - } - }, - "FlowLogsResourceType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$ResourceType": "

The type of resource on which to create the flow log.

" - } - }, - "GatewayType": { - "base": null, - "refs": { - "CreateCustomerGatewayRequest$Type": "

The type of VPN connection that this customer gateway supports (ipsec.1).

", - "CreateVpnGatewayRequest$Type": "

The type of VPN connection this virtual private gateway supports.

", - "VpnConnection$Type": "

The type of VPN connection.

", - "VpnGateway$Type": "

The type of VPN connection the virtual private gateway supports.

" - } - }, - "GetConsoleOutputRequest": { - "base": "

Contains the parameters for GetConsoleOutput.

", - "refs": { - } - }, - "GetConsoleOutputResult": { - "base": "

Contains the output of GetConsoleOutput.

", - "refs": { - } - }, - "GetConsoleScreenshotRequest": { - "base": "

Contains the parameters for the request.

", - "refs": { - } - }, - "GetConsoleScreenshotResult": { - "base": "

Contains the output of the request.

", - "refs": { - } - }, - "GetHostReservationPurchasePreviewRequest": { - "base": null, - "refs": { - } - }, - "GetHostReservationPurchasePreviewResult": { - "base": null, - "refs": { - } - }, - "GetPasswordDataRequest": { - "base": "

Contains the parameters for GetPasswordData.

", - "refs": { - } - }, - "GetPasswordDataResult": { - "base": "

Contains the output of GetPasswordData.

", - "refs": { - } - }, - "GetReservedInstancesExchangeQuoteRequest": { - "base": "

Contains the parameters for GetReservedInstanceExchangeQuote.

", - "refs": { - } - }, - "GetReservedInstancesExchangeQuoteResult": { - "base": "

Contains the output of GetReservedInstancesExchangeQuote.

", - "refs": { - } - }, - "GroupIdStringList": { - "base": null, - "refs": { - "AttachClassicLinkVpcRequest$Groups": "

The ID of one or more of the VPC's security groups. You cannot specify security groups from a different VPC.

", - "DescribeSecurityGroupsRequest$GroupIds": "

One or more security group IDs. Required for security groups in a nondefault VPC.

Default: Describes all your security groups.

", - "ModifyInstanceAttributeRequest$Groups": "

[EC2-VPC] Changes the security groups of the instance. You must specify at least one security group, even if it's just the default security group for the VPC. You must specify the security group ID, not the security group name.

" - } - }, - "GroupIdentifier": { - "base": "

Describes a security group.

", - "refs": { - "GroupIdentifierList$member": null - } - }, - "GroupIdentifierList": { - "base": null, - "refs": { - "ClassicLinkInstance$Groups": "

A list of security groups.

", - "DescribeNetworkInterfaceAttributeResult$Groups": "

The security groups associated with the network interface.

", - "Instance$SecurityGroups": "

One or more security groups for the instance.

", - "InstanceAttribute$Groups": "

The security groups associated with the instance.

", - "InstanceNetworkInterface$Groups": "

One or more security groups.

", - "LaunchSpecification$SecurityGroups": "

One or more security groups. When requesting instances in a VPC, you must specify the IDs of the security groups. When requesting instances in EC2-Classic, you can specify the names or the IDs of the security groups.

", - "NetworkInterface$Groups": "

Any security groups for the network interface.

", - "Reservation$Groups": "

[EC2-Classic only] One or more security groups.

", - "SpotFleetLaunchSpecification$SecurityGroups": "

One or more security groups. When requesting instances in a VPC, you must specify the IDs of the security groups. When requesting instances in EC2-Classic, you can specify the names or the IDs of the security groups.

" - } - }, - "GroupIds": { - "base": null, - "refs": { - "DescribeSecurityGroupReferencesRequest$GroupId": "

One or more security group IDs in your account.

" - } - }, - "GroupNameStringList": { - "base": null, - "refs": { - "DescribeSecurityGroupsRequest$GroupNames": "

[EC2-Classic and default VPC only] One or more security group names. You can specify either the security group name or the security group ID. For security groups in a nondefault VPC, use the group-name filter to describe security groups by name.

Default: Describes all your security groups.

", - "ModifySnapshotAttributeRequest$GroupNames": "

The group to modify for the snapshot.

" - } - }, - "HistoryRecord": { - "base": "

Describes an event in the history of the Spot fleet request.

", - "refs": { - "HistoryRecords$member": null - } - }, - "HistoryRecords": { - "base": null, - "refs": { - "DescribeSpotFleetRequestHistoryResponse$HistoryRecords": "

Information about the events in the history of the Spot fleet request.

" - } - }, - "Host": { - "base": "

Describes the properties of the Dedicated Host.

", - "refs": { - "HostList$member": null - } - }, - "HostInstance": { - "base": "

Describes an instance running on a Dedicated Host.

", - "refs": { - "HostInstanceList$member": null - } - }, - "HostInstanceList": { - "base": null, - "refs": { - "Host$Instances": "

The IDs and instance type that are currently running on the Dedicated Host.

" - } - }, - "HostList": { - "base": null, - "refs": { - "DescribeHostsResult$Hosts": "

Information about the Dedicated Hosts.

" - } - }, - "HostOffering": { - "base": "

Details about the Dedicated Host Reservation offering.

", - "refs": { - "HostOfferingSet$member": null - } - }, - "HostOfferingSet": { - "base": null, - "refs": { - "DescribeHostReservationOfferingsResult$OfferingSet": "

Information about the offerings.

" - } - }, - "HostProperties": { - "base": "

Describes properties of a Dedicated Host.

", - "refs": { - "Host$HostProperties": "

The hardware specifications of the Dedicated Host.

" - } - }, - "HostReservation": { - "base": "

Details about the Dedicated Host Reservation and associated Dedicated Hosts.

", - "refs": { - "HostReservationSet$member": null - } - }, - "HostReservationIdSet": { - "base": null, - "refs": { - "DescribeHostReservationsRequest$HostReservationIdSet": "

One or more host reservation IDs.

" - } - }, - "HostReservationSet": { - "base": null, - "refs": { - "DescribeHostReservationsResult$HostReservationSet": "

Details about the reservation's configuration.

" - } - }, - "HostTenancy": { - "base": null, - "refs": { - "ModifyInstancePlacementRequest$Tenancy": "

The tenancy of the instance that you are modifying.

" - } - }, - "HypervisorType": { - "base": null, - "refs": { - "Image$Hypervisor": "

The hypervisor type of the image.

", - "Instance$Hypervisor": "

The hypervisor type of the instance.

" - } - }, - "IamInstanceProfile": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "Instance$IamInstanceProfile": "

The IAM instance profile associated with the instance, if applicable.

" - } - }, - "IamInstanceProfileSpecification": { - "base": "

Describes an IAM instance profile.

", - "refs": { - "LaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RequestSpotLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

", - "RunInstancesRequest$IamInstanceProfile": "

The IAM instance profile.

", - "SpotFleetLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "IcmpTypeCode": { - "base": "

Describes the ICMP type and code.

", - "refs": { - "CreateNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying ICMP for the protocol.

", - "NetworkAclEntry$IcmpTypeCode": "

ICMP protocol: The ICMP type and code.

", - "ReplaceNetworkAclEntryRequest$IcmpTypeCode": "

ICMP protocol: The ICMP type and code. Required if specifying 1 (ICMP) for the protocol.

" - } - }, - "IdFormat": { - "base": "

Describes the ID format for a resource.

", - "refs": { - "IdFormatList$member": null - } - }, - "IdFormatList": { - "base": null, - "refs": { - "DescribeIdFormatResult$Statuses": "

Information about the ID format for the resource.

", - "DescribeIdentityIdFormatResult$Statuses": "

Information about the ID format for the resources.

" - } - }, - "Image": { - "base": "

Describes an image.

", - "refs": { - "ImageList$member": null - } - }, - "ImageAttribute": { - "base": "

Describes an image attribute.

", - "refs": { - } - }, - "ImageAttributeName": { - "base": null, - "refs": { - "DescribeImageAttributeRequest$Attribute": "

The AMI attribute.

Note: Depending on your account privileges, the blockDeviceMapping attribute may return a Client.AuthFailure error. If this happens, use DescribeImages to get information about the block device mapping for the AMI.

" - } - }, - "ImageDiskContainer": { - "base": "

Describes the disk container object for an import image task.

", - "refs": { - "ImageDiskContainerList$member": null - } - }, - "ImageDiskContainerList": { - "base": null, - "refs": { - "ImportImageRequest$DiskContainers": "

Information about the disk containers.

" - } - }, - "ImageIdStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$ImageIds": "

One or more image IDs.

Default: Describes all images available to you.

" - } - }, - "ImageList": { - "base": null, - "refs": { - "DescribeImagesResult$Images": "

Information about one or more images.

" - } - }, - "ImageState": { - "base": null, - "refs": { - "Image$State": "

The current state of the AMI. If the state is available, the image is successfully registered and can be used to launch an instance.

" - } - }, - "ImageTypeValues": { - "base": null, - "refs": { - "Image$ImageType": "

The type of image.

" - } - }, - "ImportImageRequest": { - "base": "

Contains the parameters for ImportImage.

", - "refs": { - } - }, - "ImportImageResult": { - "base": "

Contains the output for ImportImage.

", - "refs": { - } - }, - "ImportImageTask": { - "base": "

Describes an import image task.

", - "refs": { - "ImportImageTaskList$member": null - } - }, - "ImportImageTaskList": { - "base": null, - "refs": { - "DescribeImportImageTasksResult$ImportImageTasks": "

A list of zero or more import image tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportInstanceLaunchSpecification": { - "base": "

Describes the launch specification for VM import.

", - "refs": { - "ImportInstanceRequest$LaunchSpecification": "

The launch specification.

" - } - }, - "ImportInstanceRequest": { - "base": "

Contains the parameters for ImportInstance.

", - "refs": { - } - }, - "ImportInstanceResult": { - "base": "

Contains the output for ImportInstance.

", - "refs": { - } - }, - "ImportInstanceTaskDetails": { - "base": "

Describes an import instance task.

", - "refs": { - "ConversionTask$ImportInstance": "

If the task is for importing an instance, this contains information about the import instance task.

" - } - }, - "ImportInstanceVolumeDetailItem": { - "base": "

Describes an import volume task.

", - "refs": { - "ImportInstanceVolumeDetailSet$member": null - } - }, - "ImportInstanceVolumeDetailSet": { - "base": null, - "refs": { - "ImportInstanceTaskDetails$Volumes": "

One or more volumes.

" - } - }, - "ImportKeyPairRequest": { - "base": "

Contains the parameters for ImportKeyPair.

", - "refs": { - } - }, - "ImportKeyPairResult": { - "base": "

Contains the output of ImportKeyPair.

", - "refs": { - } - }, - "ImportSnapshotRequest": { - "base": "

Contains the parameters for ImportSnapshot.

", - "refs": { - } - }, - "ImportSnapshotResult": { - "base": "

Contains the output for ImportSnapshot.

", - "refs": { - } - }, - "ImportSnapshotTask": { - "base": "

Describes an import snapshot task.

", - "refs": { - "ImportSnapshotTaskList$member": null - } - }, - "ImportSnapshotTaskList": { - "base": null, - "refs": { - "DescribeImportSnapshotTasksResult$ImportSnapshotTasks": "

A list of zero or more import snapshot tasks that are currently active or were completed or canceled in the previous 7 days.

" - } - }, - "ImportTaskIdList": { - "base": null, - "refs": { - "DescribeImportImageTasksRequest$ImportTaskIds": "

A list of import image task IDs.

", - "DescribeImportSnapshotTasksRequest$ImportTaskIds": "

A list of import snapshot task IDs.

" - } - }, - "ImportVolumeRequest": { - "base": "

Contains the parameters for ImportVolume.

", - "refs": { - } - }, - "ImportVolumeResult": { - "base": "

Contains the output for ImportVolume.

", - "refs": { - } - }, - "ImportVolumeTaskDetails": { - "base": "

Describes an import volume task.

", - "refs": { - "ConversionTask$ImportVolume": "

If the task is for importing a volume, this contains information about the import volume task.

" - } - }, - "Instance": { - "base": "

Describes an instance.

", - "refs": { - "InstanceList$member": null - } - }, - "InstanceAttribute": { - "base": "

Describes an instance attribute.

", - "refs": { - } - }, - "InstanceAttributeName": { - "base": null, - "refs": { - "DescribeInstanceAttributeRequest$Attribute": "

The instance attribute.

Note: The enaSupport attribute is not supported at this time.

", - "ModifyInstanceAttributeRequest$Attribute": "

The name of the attribute.

", - "ResetInstanceAttributeRequest$Attribute": "

The attribute to reset.

You can only reset the following attributes: kernel | ramdisk | sourceDestCheck. To change an instance attribute, use ModifyInstanceAttribute.

" - } - }, - "InstanceBlockDeviceMapping": { - "base": "

Describes a block device mapping.

", - "refs": { - "InstanceBlockDeviceMappingList$member": null - } - }, - "InstanceBlockDeviceMappingList": { - "base": null, - "refs": { - "Instance$BlockDeviceMappings": "

Any block device mapping entries for the instance.

", - "InstanceAttribute$BlockDeviceMappings": "

The block device mapping of the instance.

" - } - }, - "InstanceBlockDeviceMappingSpecification": { - "base": "

Describes a block device mapping entry.

", - "refs": { - "InstanceBlockDeviceMappingSpecificationList$member": null - } - }, - "InstanceBlockDeviceMappingSpecificationList": { - "base": null, - "refs": { - "ModifyInstanceAttributeRequest$BlockDeviceMappings": "

Modifies the DeleteOnTermination attribute for volumes that are currently attached. The volume must be owned by the caller. If no value is specified for DeleteOnTermination, the default is true and the volume is deleted when the instance is terminated.

To add instance store volumes to an Amazon EBS-backed instance, you must add them when you launch the instance. For more information, see Updating the Block Device Mapping when Launching an Instance in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "InstanceCapacity": { - "base": "

Information about the instance type that the Dedicated Host supports.

", - "refs": { - "AvailableInstanceCapacityList$member": null - } - }, - "InstanceCount": { - "base": "

Describes a Reserved Instance listing state.

", - "refs": { - "InstanceCountList$member": null - } - }, - "InstanceCountList": { - "base": null, - "refs": { - "ReservedInstancesListing$InstanceCounts": "

The number of instances in this state.

" - } - }, - "InstanceExportDetails": { - "base": "

Describes an instance to export.

", - "refs": { - "ExportTask$InstanceExportDetails": "

Information about the instance to export.

" - } - }, - "InstanceIdSet": { - "base": null, - "refs": { - "RunScheduledInstancesResult$InstanceIdSet": "

The IDs of the newly launched instances.

" - } - }, - "InstanceIdStringList": { - "base": null, - "refs": { - "DescribeClassicLinkInstancesRequest$InstanceIds": "

One or more instance IDs. Must be instances linked to a VPC through ClassicLink.

", - "DescribeInstanceStatusRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

Constraints: Maximum 100 explicitly specified instance IDs.

", - "DescribeInstancesRequest$InstanceIds": "

One or more instance IDs.

Default: Describes all your instances.

", - "MonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "RebootInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "ReportInstanceStatusRequest$Instances": "

One or more instances.

", - "StartInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "StopInstancesRequest$InstanceIds": "

One or more instance IDs.

", - "TerminateInstancesRequest$InstanceIds": "

One or more instance IDs.

Constraints: Up to 1000 instance IDs. We recommend breaking up this request into smaller batches.

", - "UnmonitorInstancesRequest$InstanceIds": "

One or more instance IDs.

" - } - }, - "InstanceLifecycleType": { - "base": null, - "refs": { - "Instance$InstanceLifecycle": "

Indicates whether this is a Spot instance or a Scheduled Instance.

" - } - }, - "InstanceList": { - "base": null, - "refs": { - "Reservation$Instances": "

One or more instances.

" - } - }, - "InstanceMonitoring": { - "base": "

Describes the monitoring information of the instance.

", - "refs": { - "InstanceMonitoringList$member": null - } - }, - "InstanceMonitoringList": { - "base": null, - "refs": { - "MonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

", - "UnmonitorInstancesResult$InstanceMonitorings": "

Monitoring information for one or more instances.

" - } - }, - "InstanceNetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceList$member": null - } - }, - "InstanceNetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "InstanceNetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "InstancePrivateIpAddress$Association": "

The association information for an Elastic IP address for the network interface.

" - } - }, - "InstanceNetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "InstanceNetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "InstanceNetworkInterfaceList": { - "base": null, - "refs": { - "Instance$NetworkInterfaces": "

[EC2-VPC] One or more network interfaces for the instance.

" - } - }, - "InstanceNetworkInterfaceSpecification": { - "base": "

Describes a network interface.

", - "refs": { - "InstanceNetworkInterfaceSpecificationList$member": null - } - }, - "InstanceNetworkInterfaceSpecificationList": { - "base": null, - "refs": { - "LaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RequestSpotLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

", - "RunInstancesRequest$NetworkInterfaces": "

One or more network interfaces.

", - "SpotFleetLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "InstancePrivateIpAddress": { - "base": "

Describes a private IP address.

", - "refs": { - "InstancePrivateIpAddressList$member": null - } - }, - "InstancePrivateIpAddressList": { - "base": null, - "refs": { - "InstanceNetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "InstanceState": { - "base": "

Describes the current state of the instance.

", - "refs": { - "Instance$State": "

The current state of the instance.

", - "InstanceStateChange$CurrentState": "

The current state of the instance.

", - "InstanceStateChange$PreviousState": "

The previous state of the instance.

", - "InstanceStatus$InstanceState": "

The intended state of the instance. DescribeInstanceStatus requires that an instance be in the running state.

" - } - }, - "InstanceStateChange": { - "base": "

Describes an instance state change.

", - "refs": { - "InstanceStateChangeList$member": null - } - }, - "InstanceStateChangeList": { - "base": null, - "refs": { - "StartInstancesResult$StartingInstances": "

Information about one or more started instances.

", - "StopInstancesResult$StoppingInstances": "

Information about one or more stopped instances.

", - "TerminateInstancesResult$TerminatingInstances": "

Information about one or more terminated instances.

" - } - }, - "InstanceStateName": { - "base": null, - "refs": { - "InstanceState$Name": "

The current state of the instance.

" - } - }, - "InstanceStatus": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatusList$member": null - } - }, - "InstanceStatusDetails": { - "base": "

Describes the instance status.

", - "refs": { - "InstanceStatusDetailsList$member": null - } - }, - "InstanceStatusDetailsList": { - "base": null, - "refs": { - "InstanceStatusSummary$Details": "

The system instance health or application instance health.

" - } - }, - "InstanceStatusEvent": { - "base": "

Describes a scheduled event for an instance.

", - "refs": { - "InstanceStatusEventList$member": null - } - }, - "InstanceStatusEventList": { - "base": null, - "refs": { - "InstanceStatus$Events": "

Any scheduled events associated with the instance.

" - } - }, - "InstanceStatusList": { - "base": null, - "refs": { - "DescribeInstanceStatusResult$InstanceStatuses": "

One or more instance status descriptions.

" - } - }, - "InstanceStatusSummary": { - "base": "

Describes the status of an instance.

", - "refs": { - "InstanceStatus$SystemStatus": "

Reports impaired functionality that stems from issues related to the systems that support an instance, such as hardware failures and network connectivity problems.

", - "InstanceStatus$InstanceStatus": "

Reports impaired functionality that stems from issues internal to the instance, such as impaired reachability.

" - } - }, - "InstanceType": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$InstanceType": "

The instance type that the reservation will cover (for example, m1.small). For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

", - "ImportInstanceLaunchSpecification$InstanceType": "

The instance type. For more information about the instance types that you can import, see Instance Types in the VM Import/Export User Guide.

", - "Instance$InstanceType": "

The instance type.

", - "InstanceTypeList$member": null, - "LaunchSpecification$InstanceType": "

The instance type.

", - "RequestSpotLaunchSpecification$InstanceType": "

The instance type.

", - "ReservedInstances$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$InstanceType": "

The instance type for the modified Reserved Instances.

", - "ReservedInstancesOffering$InstanceType": "

The instance type on which the Reserved Instance can be used.

", - "RunInstancesRequest$InstanceType": "

The instance type. For more information, see Instance Types in the Amazon Elastic Compute Cloud User Guide.

Default: m1.small

", - "SpotFleetLaunchSpecification$InstanceType": "

The instance type.

", - "SpotPrice$InstanceType": "

The instance type.

" - } - }, - "InstanceTypeList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$InstanceTypes": "

Filters the results by the specified instance types.

" - } - }, - "Integer": { - "base": null, - "refs": { - "AllocateHostsRequest$Quantity": "

The number of Dedicated Hosts you want to allocate to your account with these parameters.

", - "AssignPrivateIpAddressesRequest$SecondaryPrivateIpAddressCount": "

The number of secondary IP addresses to assign to the network interface. You can't specify this parameter when also specifying private IP addresses.

", - "AttachNetworkInterfaceRequest$DeviceIndex": "

The index of the device for the network interface attachment.

", - "AuthorizeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "AuthorizeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "AvailableCapacity$AvailableVCpus": "

The number of vCPUs available on the Dedicated Host.

", - "CreateCustomerGatewayRequest$BgpAsn": "

For devices that support BGP, the customer gateway's BGP ASN.

Default: 65000

", - "CreateNetworkAclEntryRequest$RuleNumber": "

The rule number for the entry (for example, 100). ACL entries are processed in ascending order by rule number.

Constraints: Positive integer from 1 to 32766. The range 32767 to 65535 is reserved for internal use.

", - "CreateNetworkInterfaceRequest$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses to assign to a network interface. When you specify a number of secondary IP addresses, Amazon EC2 selects these IP addresses within the subnet range. You can't specify this option and specify more than one private IP address using privateIpAddresses.

The number of IP addresses you can assign to a network interface varies by instance type. For more information, see Private IP Addresses Per ENI Per Instance Type in the Amazon Elastic Compute Cloud User Guide.

", - "CreateReservedInstancesListingRequest$InstanceCount": "

The number of instances that are a part of a Reserved Instance account to be listed in the Reserved Instance Marketplace. This number should be less than or equal to the instance count associated with the Reserved Instance ID specified in this call.

", - "CreateVolumeRequest$Size": "

The size of the volume, in GiBs.

Constraints: 1-16384 for gp2, 4-16384 for io1, 500-16384 for st1, 500-16384 for sc1, and 1-1024 for standard. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "CreateVolumeRequest$Iops": "

Only valid for Provisioned IOPS SSD volumes. The number of I/O operations per second (IOPS) to provision for the volume, with a maximum ratio of 30 IOPS/GiB.

Constraint: Range is 100 to 20000 for Provisioned IOPS SSD volumes

", - "DeleteNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to delete.

", - "DescribeClassicLinkInstancesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the instance IDs parameter in the same request.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeFlowLogsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. You cannot specify this parameter and the flow log IDs parameter in the same request.

", - "DescribeHostReservationOfferingsRequest$MinDuration": "

This is the minimum duration of the reservation you'd like to purchase, specified in seconds. Reservations are available in one-year and three-year terms. The number of seconds specified must be the number of seconds in a year (365x24x60x60) times one of the supported durations (1 or 3). For example, specify 31536000 for one year.

", - "DescribeHostReservationOfferingsRequest$MaxDuration": "

This is the maximum duration of the reservation you'd like to purchase, specified in seconds. Reservations are available in one-year and three-year terms. The number of seconds specified must be the number of seconds in a year (365x24x60x60) times one of the supported durations (1 or 3). For example, specify 94608000 for three years.

", - "DescribeHostReservationOfferingsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error.

", - "DescribeHostReservationsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error.

", - "DescribeHostsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results can be seen by sending another request with the returned nextToken value. This value can be between 5 and 500; if maxResults is given a larger value than 500, you will receive an error. You cannot specify this parameter and the host IDs parameter in the same request.

", - "DescribeImportImageTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeImportSnapshotTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeInstanceStatusRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter in the same call.

", - "DescribeInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter or tag filters in the same call.

", - "DescribeMovingAddressesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value outside of this range, an error is returned.

Default: If no value is provided, the default is 1000.

", - "DescribeNatGatewaysRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribePrefixListsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value specified is greater than 1000, we return only 1000 items.

", - "DescribeReservedInstancesOfferingsRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. The maximum is 100.

Default: 100

", - "DescribeReservedInstancesOfferingsRequest$MaxInstanceCount": "

The maximum number of instances to filter when searching for offerings.

Default: 20

", - "DescribeScheduledInstanceAvailabilityRequest$MinSlotDurationInHours": "

The minimum available duration, in hours. The minimum required duration is 1,200 hours per year. For example, the minimum daily schedule is 4 hours, the minimum weekly schedule is 24 hours, and the minimum monthly schedule is 100 hours.

", - "DescribeScheduledInstanceAvailabilityRequest$MaxSlotDurationInHours": "

The maximum available duration, in hours. This value must be greater than MinSlotDurationInHours and less than 1,720.

", - "DescribeScheduledInstanceAvailabilityRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 300. The default value is 300. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeScheduledInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 300. The default value is 100. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSnapshotsRequest$MaxResults": "

The maximum number of snapshot results returned by DescribeSnapshots in paginated output. When this parameter is used, DescribeSnapshots only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeSnapshots request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeSnapshots returns all results. You cannot specify this parameter and the snapshot IDs parameter in the same request.

", - "DescribeSpotFleetInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotFleetRequestsRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeSpotPriceHistoryRequest$MaxResults": "

The maximum number of results to return in a single call. Specify a value between 1 and 1000. The default value is 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeTagsRequest$MaxResults": "

The maximum number of results to return in a single call. This value can be between 5 and 1000. To retrieve the remaining results, make another call with the returned NextToken value.

", - "DescribeVolumeStatusRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumeStatus in paginated output. When this parameter is used, the request only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumeStatus returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVolumesRequest$MaxResults": "

The maximum number of volume results returned by DescribeVolumes in paginated output. When this parameter is used, DescribeVolumes only returns MaxResults results in a single page along with a NextToken response element. The remaining results of the initial request can be seen by sending another DescribeVolumes request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value larger than 1000, only 1000 results are returned. If this parameter is not used, then DescribeVolumes returns all results. You cannot specify this parameter and the volume IDs parameter in the same request.

", - "DescribeVpcEndpointServicesRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "DescribeVpcEndpointsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

Constraint: If the value is greater than 1000, we return only 1000 items.

", - "EbsBlockDevice$VolumeSize": "

The size of the volume, in GiB.

Constraints: 1-16384 for General Purpose SSD (gp2), 4-16384 for Provisioned IOPS SSD (io1), 500-16384 for Throughput Optimized HDD (st1), 500-16384 for Cold HDD (sc1), and 1-1024 for Magnetic (standard) volumes. If you specify a snapshot, the volume size must be equal to or larger than the snapshot size.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "EbsBlockDevice$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For io1, this represents the number of IOPS that are provisioned for the volume. For gp2, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information about General Purpose SSD baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

", - "HostOffering$Duration": "

The duration of the offering (in seconds).

", - "HostProperties$Sockets": "

The number of sockets on the Dedicated Host.

", - "HostProperties$Cores": "

The number of cores on the Dedicated Host.

", - "HostProperties$TotalVCpus": "

The number of vCPUs on the Dedicated Host.

", - "HostReservation$Count": "

The number of Dedicated Hosts the reservation is associated with.

", - "HostReservation$Duration": "

The length of the reservation's term, specified in seconds. Can be 31536000 (1 year) | 94608000 (3 years).

", - "IcmpTypeCode$Type": "

The ICMP code. A value of -1 means all codes for the specified ICMP type.

", - "IcmpTypeCode$Code": "

The ICMP type. A value of -1 means all types.

", - "Instance$AmiLaunchIndex": "

The AMI launch index, which can be used to find this instance in the launch group.

", - "InstanceCapacity$AvailableCapacity": "

The number of instances that can still be launched onto the Dedicated Host.

", - "InstanceCapacity$TotalCapacity": "

The total number of instances that can be launched onto the Dedicated Host.

", - "InstanceCount$InstanceCount": "

The number of listed Reserved Instances in the state specified by the state.

", - "InstanceNetworkInterfaceAttachment$DeviceIndex": "

The index of the device on the instance for the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$DeviceIndex": "

The index of the device on the instance for the network interface attachment. If you are specifying a network interface in a RunInstances request, you must provide the device index.

", - "InstanceNetworkInterfaceSpecification$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses. You can't specify this option and specify more than one private IP address using the private IP addresses option. You cannot specify this option if you're launching more than one instance in a RunInstances request.

", - "InstanceState$Code": "

The low byte represents the state. The high byte is an opaque internal value and should be ignored.

  • 0 : pending

  • 16 : running

  • 32 : shutting-down

  • 48 : terminated

  • 64 : stopping

  • 80 : stopped

", - "IpPermission$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "IpPermission$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code. A value of -1 indicates all ICMP codes for the specified ICMP type.

", - "ModifySpotFleetRequestRequest$TargetCapacity": "

The size of the fleet.

", - "NetworkAclEntry$RuleNumber": "

The rule number for the entry. ACL entries are processed in ascending order by rule number.

", - "NetworkInterfaceAttachment$DeviceIndex": "

The device index of the network interface attachment on the instance.

", - "OccurrenceDayRequestSet$member": null, - "OccurrenceDaySet$member": null, - "PortRange$From": "

The first port in the range.

", - "PortRange$To": "

The last port in the range.

", - "PricingDetail$Count": "

The number of reservations available for the price.

", - "Purchase$Duration": "

The duration of the reservation's term in seconds.

", - "PurchaseRequest$InstanceCount": "

The number of instances.

", - "PurchaseReservedInstancesOfferingRequest$InstanceCount": "

The number of Reserved Instances to purchase.

", - "ReplaceNetworkAclEntryRequest$RuleNumber": "

The rule number of the entry to replace.

", - "RequestSpotInstancesRequest$InstanceCount": "

The maximum number of Spot instances to launch.

Default: 1

", - "RequestSpotInstancesRequest$BlockDurationMinutes": "

The required duration for the Spot instances (also known as Spot blocks), in minutes. This value must be a multiple of 60 (60, 120, 180, 240, 300, or 360).

The duration period starts as soon as your Spot instance receives its instance ID. At the end of the duration period, Amazon EC2 marks the Spot instance for termination and provides a Spot instance termination notice, which gives the instance a two-minute warning before it terminates.

Note that you can't specify an Availability Zone group or a launch group if you specify a duration.

", - "ReservedInstances$InstanceCount": "

The number of reservations purchased.

", - "ReservedInstancesConfiguration$InstanceCount": "

The number of modified Reserved Instances.

", - "RevokeSecurityGroupEgressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP type number. We recommend that you specify the port range in a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$FromPort": "

The start of port range for the TCP and UDP protocols, or an ICMP type number. For the ICMP type number, use -1 to specify all ICMP types.

", - "RevokeSecurityGroupIngressRequest$ToPort": "

The end of port range for the TCP and UDP protocols, or an ICMP code number. For the ICMP code number, use -1 to specify all ICMP codes for the ICMP type.

", - "RunInstancesRequest$MinCount": "

The minimum number of instances to launch. If you specify a minimum that is more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches no instances.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 General FAQ.

", - "RunInstancesRequest$MaxCount": "

The maximum number of instances to launch. If you specify more instances than Amazon EC2 can launch in the target Availability Zone, Amazon EC2 launches the largest possible number of instances above MinCount.

Constraints: Between 1 and the maximum number you're allowed for the specified instance type. For more information about the default limits, and how to request an increase, see How many instances can I run in Amazon EC2 in the Amazon EC2 FAQ.

", - "RunScheduledInstancesRequest$InstanceCount": "

The number of instances.

Default: 1

", - "ScheduledInstance$SlotDurationInHours": "

The number of hours in the schedule.

", - "ScheduledInstance$TotalScheduledInstanceHours": "

The total number of hours for a single instance for the entire term.

", - "ScheduledInstance$InstanceCount": "

The number of instances.

", - "ScheduledInstanceAvailability$SlotDurationInHours": "

The number of hours in the schedule.

", - "ScheduledInstanceAvailability$TotalScheduledInstanceHours": "

The total number of hours for a single instance for the entire term.

", - "ScheduledInstanceAvailability$AvailableInstanceCount": "

The number of available instances.

", - "ScheduledInstanceAvailability$MinTermDurationInDays": "

The minimum term. The only possible value is 365 days.

", - "ScheduledInstanceAvailability$MaxTermDurationInDays": "

The maximum term. The only possible value is 365 days.

", - "ScheduledInstanceRecurrence$Interval": "

The interval quantity. The interval unit depends on the value of frequency. For example, every 2 weeks or every 2 months.

", - "ScheduledInstanceRecurrenceRequest$Interval": "

The interval quantity. The interval unit depends on the value of Frequency. For example, every 2 weeks or every 2 months.

", - "ScheduledInstancesEbs$VolumeSize": "

The size of the volume, in GiB.

Default: If you're creating the volume from a snapshot and don't specify a volume size, the default is the snapshot size.

", - "ScheduledInstancesEbs$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For io1 volumes, this represents the number of IOPS that are provisioned for the volume. For gp2 volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information about gp2 baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

", - "ScheduledInstancesNetworkInterface$DeviceIndex": "

The index of the device for the network interface attachment.

", - "ScheduledInstancesNetworkInterface$SecondaryPrivateIpAddressCount": "

The number of secondary private IP addresses.

", - "Snapshot$VolumeSize": "

The size of the volume, in GiB.

", - "SpotFleetRequestConfigData$TargetCapacity": "

The number of units to request. You can choose to set the target capacity in terms of instances or a performance characteristic that is important to your application workload, such as vCPUs, memory, or I/O.

", - "SpotInstanceRequest$BlockDurationMinutes": "

The duration for the Spot instance, in minutes.

", - "StaleIpPermission$FromPort": "

The start of the port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "StaleIpPermission$ToPort": "

The end of the port range for the TCP and UDP protocols, or an ICMP type number. A value of -1 indicates all ICMP types.

", - "Subnet$AvailableIpAddressCount": "

The number of unused IP addresses in the subnet. Note that the IP addresses for any stopped instances are considered unavailable.

", - "TargetConfiguration$InstanceCount": "

The number of instances the Convertible Reserved Instance offering can be applied to. This parameter is reserved and cannot be specified in a request

", - "TargetConfigurationRequest$InstanceCount": "

The number of instances the Covertible Reserved Instance offering can be applied to. This parameter is reserved and cannot be specified in a request

", - "VgwTelemetry$AcceptedRouteCount": "

The number of accepted routes.

", - "Volume$Size": "

The size of the volume, in GiBs.

", - "Volume$Iops": "

The number of I/O operations per second (IOPS) that the volume supports. For Provisioned IOPS SSD volumes, this represents the number of IOPS that are provisioned for the volume. For General Purpose SSD volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. For more information on General Purpose SSD baseline performance, I/O credits, and bursting, see Amazon EBS Volume Types in the Amazon Elastic Compute Cloud User Guide.

Constraint: Range is 100-20000 IOPS for io1 volumes and 100-10000 IOPS for gp2 volumes.

Condition: This parameter is required for requests to create io1 volumes; it is not used in requests to create gp2, st1, sc1, or standard volumes.

" - } - }, - "InternetGateway": { - "base": "

Describes an Internet gateway.

", - "refs": { - "CreateInternetGatewayResult$InternetGateway": "

Information about the Internet gateway.

", - "InternetGatewayList$member": null - } - }, - "InternetGatewayAttachment": { - "base": "

Describes the attachment of a VPC to an Internet gateway.

", - "refs": { - "InternetGatewayAttachmentList$member": null - } - }, - "InternetGatewayAttachmentList": { - "base": null, - "refs": { - "InternetGateway$Attachments": "

Any VPCs attached to the Internet gateway.

" - } - }, - "InternetGatewayList": { - "base": null, - "refs": { - "DescribeInternetGatewaysResult$InternetGateways": "

Information about one or more Internet gateways.

" - } - }, - "IpPermission": { - "base": "

Describes a security group rule.

", - "refs": { - "IpPermissionList$member": null - } - }, - "IpPermissionList": { - "base": null, - "refs": { - "AuthorizeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "AuthorizeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. Can be used to specify multiple rules in a single command.

", - "RevokeSecurityGroupEgressRequest$IpPermissions": "

A set of IP permissions. You can't specify a destination security group and a CIDR IP address range.

", - "RevokeSecurityGroupIngressRequest$IpPermissions": "

A set of IP permissions. You can't specify a source security group and a CIDR IP address range.

", - "SecurityGroup$IpPermissions": "

One or more inbound rules associated with the security group.

", - "SecurityGroup$IpPermissionsEgress": "

[EC2-VPC] One or more outbound rules associated with the security group.

" - } - }, - "IpRange": { - "base": "

Describes an IP range.

", - "refs": { - "IpRangeList$member": null - } - }, - "IpRangeList": { - "base": null, - "refs": { - "IpPermission$IpRanges": "

One or more IP ranges.

" - } - }, - "IpRanges": { - "base": null, - "refs": { - "StaleIpPermission$IpRanges": "

One or more IP ranges. Not applicable for stale security group rules.

" - } - }, - "KeyNameStringList": { - "base": null, - "refs": { - "DescribeKeyPairsRequest$KeyNames": "

One or more key pair names.

Default: Describes all your key pairs.

" - } - }, - "KeyPair": { - "base": "

Describes a key pair.

", - "refs": { - } - }, - "KeyPairInfo": { - "base": "

Describes a key pair.

", - "refs": { - "KeyPairList$member": null - } - }, - "KeyPairList": { - "base": null, - "refs": { - "DescribeKeyPairsResult$KeyPairs": "

Information about one or more key pairs.

" - } - }, - "LaunchPermission": { - "base": "

Describes a launch permission.

", - "refs": { - "LaunchPermissionList$member": null - } - }, - "LaunchPermissionList": { - "base": null, - "refs": { - "ImageAttribute$LaunchPermissions": "

One or more launch permissions.

", - "LaunchPermissionModifications$Add": "

The AWS account ID to add to the list of launch permissions for the AMI.

", - "LaunchPermissionModifications$Remove": "

The AWS account ID to remove from the list of launch permissions for the AMI.

" - } - }, - "LaunchPermissionModifications": { - "base": "

Describes a launch permission modification.

", - "refs": { - "ModifyImageAttributeRequest$LaunchPermission": "

A launch permission modification.

" - } - }, - "LaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "SpotInstanceRequest$LaunchSpecification": "

Additional information for launching instances.

" - } - }, - "LaunchSpecsList": { - "base": null, - "refs": { - "SpotFleetRequestConfigData$LaunchSpecifications": "

Information about the launch specifications for the Spot fleet request.

" - } - }, - "ListingState": { - "base": null, - "refs": { - "InstanceCount$State": "

The states of the listed Reserved Instances.

" - } - }, - "ListingStatus": { - "base": null, - "refs": { - "ReservedInstancesListing$Status": "

The status of the Reserved Instance listing.

" - } - }, - "Long": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$MinDuration": "

The minimum duration (in seconds) to filter when searching for offerings.

Default: 2592000 (1 month)

", - "DescribeReservedInstancesOfferingsRequest$MaxDuration": "

The maximum duration (in seconds) to filter when searching for offerings.

Default: 94608000 (3 years)

", - "DiskImageDescription$Size": "

The size of the disk image, in GiB.

", - "DiskImageDetail$Bytes": "

The size of the disk image, in GiB.

", - "DiskImageVolumeDescription$Size": "

The size of the volume, in GiB.

", - "ImportInstanceVolumeDetailItem$BytesConverted": "

The number of bytes converted so far.

", - "ImportVolumeTaskDetails$BytesConverted": "

The number of bytes converted so far.

", - "PriceSchedule$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "PriceScheduleSpecification$Term": "

The number of months remaining in the reservation. For example, 2 is the second to the last month before the capacity reservation expires.

", - "ReservedInstances$Duration": "

The duration of the Reserved Instance, in seconds.

", - "ReservedInstancesOffering$Duration": "

The duration of the Reserved Instance, in seconds.

", - "VolumeDetail$Size": "

The size of the volume, in GiB.

" - } - }, - "MaxResults": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeVpcClassicLinkDnsSupportRequest$MaxResults": "

The maximum number of items to return for this request. The request returns a token that you can specify in a subsequent call to get the next set of results.

" - } - }, - "ModifyHostsRequest": { - "base": "

Contains the parameters for ModifyHosts.

", - "refs": { - } - }, - "ModifyHostsResult": { - "base": "

Contains the output of ModifyHosts.

", - "refs": { - } - }, - "ModifyIdFormatRequest": { - "base": "

Contains the parameters of ModifyIdFormat.

", - "refs": { - } - }, - "ModifyIdentityIdFormatRequest": { - "base": "

Contains the parameters of ModifyIdentityIdFormat.

", - "refs": { - } - }, - "ModifyImageAttributeRequest": { - "base": "

Contains the parameters for ModifyImageAttribute.

", - "refs": { - } - }, - "ModifyInstanceAttributeRequest": { - "base": "

Contains the parameters for ModifyInstanceAttribute.

", - "refs": { - } - }, - "ModifyInstancePlacementRequest": { - "base": "

Contains the parameters for ModifyInstancePlacement.

", - "refs": { - } - }, - "ModifyInstancePlacementResult": { - "base": "

Contains the output of ModifyInstancePlacement.

", - "refs": { - } - }, - "ModifyNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for ModifyNetworkInterfaceAttribute.

", - "refs": { - } - }, - "ModifyReservedInstancesRequest": { - "base": "

Contains the parameters for ModifyReservedInstances.

", - "refs": { - } - }, - "ModifyReservedInstancesResult": { - "base": "

Contains the output of ModifyReservedInstances.

", - "refs": { - } - }, - "ModifySnapshotAttributeRequest": { - "base": "

Contains the parameters for ModifySnapshotAttribute.

", - "refs": { - } - }, - "ModifySpotFleetRequestRequest": { - "base": "

Contains the parameters for ModifySpotFleetRequest.

", - "refs": { - } - }, - "ModifySpotFleetRequestResponse": { - "base": "

Contains the output of ModifySpotFleetRequest.

", - "refs": { - } - }, - "ModifySubnetAttributeRequest": { - "base": "

Contains the parameters for ModifySubnetAttribute.

", - "refs": { - } - }, - "ModifyVolumeAttributeRequest": { - "base": "

Contains the parameters for ModifyVolumeAttribute.

", - "refs": { - } - }, - "ModifyVpcAttributeRequest": { - "base": "

Contains the parameters for ModifyVpcAttribute.

", - "refs": { - } - }, - "ModifyVpcEndpointRequest": { - "base": "

Contains the parameters for ModifyVpcEndpoint.

", - "refs": { - } - }, - "ModifyVpcEndpointResult": { - "base": "

Contains the output of ModifyVpcEndpoint.

", - "refs": { - } - }, - "ModifyVpcPeeringConnectionOptionsRequest": { - "base": null, - "refs": { - } - }, - "ModifyVpcPeeringConnectionOptionsResult": { - "base": null, - "refs": { - } - }, - "MonitorInstancesRequest": { - "base": "

Contains the parameters for MonitorInstances.

", - "refs": { - } - }, - "MonitorInstancesResult": { - "base": "

Contains the output of MonitorInstances.

", - "refs": { - } - }, - "Monitoring": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "Instance$Monitoring": "

The monitoring information for the instance.

", - "InstanceMonitoring$Monitoring": "

The monitoring information.

" - } - }, - "MonitoringState": { - "base": null, - "refs": { - "Monitoring$State": "

Indicates whether monitoring is enabled for the instance.

" - } - }, - "MoveAddressToVpcRequest": { - "base": "

Contains the parameters for MoveAddressToVpc.

", - "refs": { - } - }, - "MoveAddressToVpcResult": { - "base": "

Contains the output of MoveAddressToVpc.

", - "refs": { - } - }, - "MoveStatus": { - "base": null, - "refs": { - "MovingAddressStatus$MoveStatus": "

The status of the Elastic IP address that's being moved to the EC2-VPC platform, or restored to the EC2-Classic platform.

" - } - }, - "MovingAddressStatus": { - "base": "

Describes the status of a moving Elastic IP address.

", - "refs": { - "MovingAddressStatusSet$member": null - } - }, - "MovingAddressStatusSet": { - "base": null, - "refs": { - "DescribeMovingAddressesResult$MovingAddressStatuses": "

The status for each Elastic IP address.

" - } - }, - "NatGateway": { - "base": "

Describes a NAT gateway.

", - "refs": { - "CreateNatGatewayResult$NatGateway": "

Information about the NAT gateway.

", - "NatGatewayList$member": null - } - }, - "NatGatewayAddress": { - "base": "

Describes the IP addresses and network interface associated with a NAT gateway.

", - "refs": { - "NatGatewayAddressList$member": null - } - }, - "NatGatewayAddressList": { - "base": null, - "refs": { - "NatGateway$NatGatewayAddresses": "

Information about the IP addresses and network interface associated with the NAT gateway.

" - } - }, - "NatGatewayList": { - "base": null, - "refs": { - "DescribeNatGatewaysResult$NatGateways": "

Information about the NAT gateways.

" - } - }, - "NatGatewayState": { - "base": null, - "refs": { - "NatGateway$State": "

The state of the NAT gateway.

  • pending: The NAT gateway is being created and is not ready to process traffic.

  • failed: The NAT gateway could not be created. Check the failureCode and failureMessage fields for the reason.

  • available: The NAT gateway is able to process traffic. This status remains until you delete the NAT gateway, and does not indicate the health of the NAT gateway.

  • deleting: The NAT gateway is in the process of being terminated and may still be processing traffic.

  • deleted: The NAT gateway has been terminated and is no longer processing traffic.

" - } - }, - "NetworkAcl": { - "base": "

Describes a network ACL.

", - "refs": { - "CreateNetworkAclResult$NetworkAcl": "

Information about the network ACL.

", - "NetworkAclList$member": null - } - }, - "NetworkAclAssociation": { - "base": "

Describes an association between a network ACL and a subnet.

", - "refs": { - "NetworkAclAssociationList$member": null - } - }, - "NetworkAclAssociationList": { - "base": null, - "refs": { - "NetworkAcl$Associations": "

Any associations between the network ACL and one or more subnets

" - } - }, - "NetworkAclEntry": { - "base": "

Describes an entry in a network ACL.

", - "refs": { - "NetworkAclEntryList$member": null - } - }, - "NetworkAclEntryList": { - "base": null, - "refs": { - "NetworkAcl$Entries": "

One or more entries (rules) in the network ACL.

" - } - }, - "NetworkAclList": { - "base": null, - "refs": { - "DescribeNetworkAclsResult$NetworkAcls": "

Information about one or more network ACLs.

" - } - }, - "NetworkInterface": { - "base": "

Describes a network interface.

", - "refs": { - "CreateNetworkInterfaceResult$NetworkInterface": "

Information about the network interface.

", - "NetworkInterfaceList$member": null - } - }, - "NetworkInterfaceAssociation": { - "base": "

Describes association information for an Elastic IP address.

", - "refs": { - "NetworkInterface$Association": "

The association information for an Elastic IP associated with the network interface.

", - "NetworkInterfacePrivateIpAddress$Association": "

The association information for an Elastic IP address associated with the network interface.

" - } - }, - "NetworkInterfaceAttachment": { - "base": "

Describes a network interface attachment.

", - "refs": { - "DescribeNetworkInterfaceAttributeResult$Attachment": "

The attachment (if any) of the network interface.

", - "NetworkInterface$Attachment": "

The network interface attachment.

" - } - }, - "NetworkInterfaceAttachmentChanges": { - "base": "

Describes an attachment change.

", - "refs": { - "ModifyNetworkInterfaceAttributeRequest$Attachment": "

Information about the interface attachment. If modifying the 'delete on termination' attribute, you must specify the ID of the interface attachment.

" - } - }, - "NetworkInterfaceAttribute": { - "base": null, - "refs": { - "DescribeNetworkInterfaceAttributeRequest$Attribute": "

The attribute of the network interface.

" - } - }, - "NetworkInterfaceIdList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesRequest$NetworkInterfaceIds": "

One or more network interface IDs.

Default: Describes all your network interfaces.

" - } - }, - "NetworkInterfaceList": { - "base": null, - "refs": { - "DescribeNetworkInterfacesResult$NetworkInterfaces": "

Information about one or more network interfaces.

" - } - }, - "NetworkInterfacePrivateIpAddress": { - "base": "

Describes the private IP address of a network interface.

", - "refs": { - "NetworkInterfacePrivateIpAddressList$member": null - } - }, - "NetworkInterfacePrivateIpAddressList": { - "base": null, - "refs": { - "NetworkInterface$PrivateIpAddresses": "

The private IP addresses associated with the network interface.

" - } - }, - "NetworkInterfaceStatus": { - "base": null, - "refs": { - "InstanceNetworkInterface$Status": "

The status of the network interface.

", - "NetworkInterface$Status": "

The status of the network interface.

" - } - }, - "NetworkInterfaceType": { - "base": null, - "refs": { - "NetworkInterface$InterfaceType": "

The type of interface.

" - } - }, - "NewDhcpConfiguration": { - "base": null, - "refs": { - "NewDhcpConfigurationList$member": null - } - }, - "NewDhcpConfigurationList": { - "base": null, - "refs": { - "CreateDhcpOptionsRequest$DhcpConfigurations": "

A DHCP configuration option.

" - } - }, - "NextToken": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcClassicLinkDnsSupportRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcClassicLinkDnsSupportResult$NextToken": "

The token to use when requesting the next set of items.

" - } - }, - "OccurrenceDayRequestSet": { - "base": null, - "refs": { - "ScheduledInstanceRecurrenceRequest$OccurrenceDays": "

The days. For a monthly schedule, this is one or more days of the month (1-31). For a weekly schedule, this is one or more days of the week (1-7, where 1 is Sunday). You can't specify this value with a daily schedule. If the occurrence is relative to the end of the month, you can specify only a single day.

" - } - }, - "OccurrenceDaySet": { - "base": null, - "refs": { - "ScheduledInstanceRecurrence$OccurrenceDaySet": "

The days. For a monthly schedule, this is one or more days of the month (1-31). For a weekly schedule, this is one or more days of the week (1-7, where 1 is Sunday).

" - } - }, - "OfferingClassType": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$OfferingClass": "

The offering class of the Reserved Instance. Can be standard or convertible.

", - "DescribeReservedInstancesRequest$OfferingClass": "

Describes whether the Reserved Instance is Standard or Convertible.

", - "ReservedInstances$OfferingClass": "

The offering class of the Reserved Instance.

", - "ReservedInstancesOffering$OfferingClass": "

If convertible it can be exchanged for Reserved Instances of the same or higher monetary value, with different configurations. If standard, it is not possible to perform an exchange.

" - } - }, - "OfferingTypeValues": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "DescribeReservedInstancesRequest$OfferingType": "

The Reserved Instance offering type. If you are using tools that predate the 2011-11-01 API version, you only have access to the Medium Utilization Reserved Instance offering type.

", - "ReservedInstances$OfferingType": "

The Reserved Instance offering type.

", - "ReservedInstancesOffering$OfferingType": "

The Reserved Instance offering type.

" - } - }, - "OperationType": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$OperationType": "

The operation type.

", - "ModifySnapshotAttributeRequest$OperationType": "

The type of operation to perform to the attribute.

" - } - }, - "OwnerStringList": { - "base": null, - "refs": { - "DescribeImagesRequest$Owners": "

Filters the images by the owner. Specify an AWS account ID, self (owner is the sender of the request), or an AWS owner alias (valid values are amazon | aws-marketplace | microsoft). Omitting this option returns all images for which you have launch permissions, regardless of ownership.

", - "DescribeSnapshotsRequest$OwnerIds": "

Returns the snapshots owned by the specified owner. Multiple owners can be specified.

" - } - }, - "PaymentOption": { - "base": null, - "refs": { - "HostOffering$PaymentOption": "

The available payment option.

", - "HostReservation$PaymentOption": "

The payment option selected for this reservation.

", - "Purchase$PaymentOption": "

The payment option for the reservation.

" - } - }, - "PeeringConnectionOptions": { - "base": "

Describes the VPC peering connection options.

", - "refs": { - "ModifyVpcPeeringConnectionOptionsResult$RequesterPeeringConnectionOptions": "

Information about the VPC peering connection options for the requester VPC.

", - "ModifyVpcPeeringConnectionOptionsResult$AccepterPeeringConnectionOptions": "

Information about the VPC peering connection options for the accepter VPC.

" - } - }, - "PeeringConnectionOptionsRequest": { - "base": "

The VPC peering connection options.

", - "refs": { - "ModifyVpcPeeringConnectionOptionsRequest$RequesterPeeringConnectionOptions": "

The VPC peering connection options for the requester VPC.

", - "ModifyVpcPeeringConnectionOptionsRequest$AccepterPeeringConnectionOptions": "

The VPC peering connection options for the accepter VPC.

" - } - }, - "PermissionGroup": { - "base": null, - "refs": { - "CreateVolumePermission$Group": "

The specific group that is to be added or removed from a volume's list of create volume permissions.

", - "LaunchPermission$Group": "

The name of the group.

" - } - }, - "Placement": { - "base": "

Describes the placement for the instance.

", - "refs": { - "ImportInstanceLaunchSpecification$Placement": "

The placement information for the instance.

", - "Instance$Placement": "

The location where the instance launched, if applicable.

", - "RunInstancesRequest$Placement": "

The placement for the instance.

" - } - }, - "PlacementGroup": { - "base": "

Describes a placement group.

", - "refs": { - "PlacementGroupList$member": null - } - }, - "PlacementGroupList": { - "base": null, - "refs": { - "DescribePlacementGroupsResult$PlacementGroups": "

One or more placement groups.

" - } - }, - "PlacementGroupState": { - "base": null, - "refs": { - "PlacementGroup$State": "

The state of the placement group.

" - } - }, - "PlacementGroupStringList": { - "base": null, - "refs": { - "DescribePlacementGroupsRequest$GroupNames": "

One or more placement group names.

Default: Describes all your placement groups, or only those otherwise specified.

" - } - }, - "PlacementStrategy": { - "base": null, - "refs": { - "CreatePlacementGroupRequest$Strategy": "

The placement strategy.

", - "PlacementGroup$Strategy": "

The placement strategy.

" - } - }, - "PlatformValues": { - "base": null, - "refs": { - "Image$Platform": "

The value is Windows for Windows AMIs; otherwise blank.

", - "ImportInstanceRequest$Platform": "

The instance operating system.

", - "ImportInstanceTaskDetails$Platform": "

The instance operating system.

", - "Instance$Platform": "

The value is Windows for Windows instances; otherwise blank.

" - } - }, - "PortRange": { - "base": "

Describes a range of ports.

", - "refs": { - "CreateNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "NetworkAclEntry$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to.

", - "ReplaceNetworkAclEntryRequest$PortRange": "

TCP or UDP protocols: The range of ports the rule applies to. Required if specifying 6 (TCP) or 17 (UDP) for the protocol.

" - } - }, - "PrefixList": { - "base": "

Describes prefixes for AWS services.

", - "refs": { - "PrefixListSet$member": null - } - }, - "PrefixListId": { - "base": "

The ID of the prefix.

", - "refs": { - "PrefixListIdList$member": null - } - }, - "PrefixListIdList": { - "base": null, - "refs": { - "IpPermission$PrefixListIds": "

(Valid for AuthorizeSecurityGroupEgress, RevokeSecurityGroupEgress and DescribeSecurityGroups only) One or more prefix list IDs for an AWS service. In an AuthorizeSecurityGroupEgress request, this is the AWS service that you want to access through a VPC endpoint from instances associated with the security group.

" - } - }, - "PrefixListIdSet": { - "base": null, - "refs": { - "StaleIpPermission$PrefixListIds": "

One or more prefix list IDs for an AWS service. Not applicable for stale security group rules.

" - } - }, - "PrefixListSet": { - "base": null, - "refs": { - "DescribePrefixListsResult$PrefixLists": "

All available prefix lists.

" - } - }, - "PriceSchedule": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleList$member": null - } - }, - "PriceScheduleList": { - "base": null, - "refs": { - "ReservedInstancesListing$PriceSchedules": "

The price of the Reserved Instance listing.

" - } - }, - "PriceScheduleSpecification": { - "base": "

Describes the price for a Reserved Instance.

", - "refs": { - "PriceScheduleSpecificationList$member": null - } - }, - "PriceScheduleSpecificationList": { - "base": null, - "refs": { - "CreateReservedInstancesListingRequest$PriceSchedules": "

A list specifying the price of the Standard Reserved Instance for each month remaining in the Reserved Instance term.

" - } - }, - "PricingDetail": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "PricingDetailsList$member": null - } - }, - "PricingDetailsList": { - "base": null, - "refs": { - "ReservedInstancesOffering$PricingDetails": "

The pricing details of the Reserved Instance offering.

" - } - }, - "PrivateIpAddressConfigSet": { - "base": null, - "refs": { - "ScheduledInstancesNetworkInterface$PrivateIpAddressConfigs": "

The private IP addresses.

" - } - }, - "PrivateIpAddressSpecification": { - "base": "

Describes a secondary private IP address for a network interface.

", - "refs": { - "PrivateIpAddressSpecificationList$member": null - } - }, - "PrivateIpAddressSpecificationList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$PrivateIpAddresses": "

One or more private IP addresses.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddresses": "

One or more private IP addresses to assign to the network interface. Only one private IP address can be designated as primary. You cannot specify this option if you're launching more than one instance in a RunInstances request.

" - } - }, - "PrivateIpAddressStringList": { - "base": null, - "refs": { - "AssignPrivateIpAddressesRequest$PrivateIpAddresses": "

One or more IP addresses to be assigned as a secondary private IP address to the network interface. You can't specify this parameter when also specifying a number of secondary IP addresses.

If you don't specify an IP address, Amazon EC2 automatically selects an IP address within the subnet range.

", - "UnassignPrivateIpAddressesRequest$PrivateIpAddresses": "

The secondary private IP addresses to unassign from the network interface. You can specify this option multiple times to unassign more than one IP address.

" - } - }, - "ProductCode": { - "base": "

Describes a product code.

", - "refs": { - "ProductCodeList$member": null - } - }, - "ProductCodeList": { - "base": null, - "refs": { - "DescribeSnapshotAttributeResult$ProductCodes": "

A list of product codes.

", - "DescribeVolumeAttributeResult$ProductCodes": "

A list of product codes.

", - "Image$ProductCodes": "

Any product codes associated with the AMI.

", - "ImageAttribute$ProductCodes": "

One or more product codes.

", - "Instance$ProductCodes": "

The product codes attached to this instance, if applicable.

", - "InstanceAttribute$ProductCodes": "

A list of product codes.

" - } - }, - "ProductCodeStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$ProductCodes": "

One or more product codes. After you add a product code to an AMI, it can't be removed. This is only valid when modifying the productCodes attribute.

" - } - }, - "ProductCodeValues": { - "base": null, - "refs": { - "ProductCode$ProductCodeType": "

The type of product code.

" - } - }, - "ProductDescriptionList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryRequest$ProductDescriptions": "

Filters the results by the specified basic product descriptions.

" - } - }, - "PropagatingVgw": { - "base": "

Describes a virtual private gateway propagating route.

", - "refs": { - "PropagatingVgwList$member": null - } - }, - "PropagatingVgwList": { - "base": null, - "refs": { - "RouteTable$PropagatingVgws": "

Any virtual private gateway (VGW) propagating routes.

" - } - }, - "ProvisionedBandwidth": { - "base": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "refs": { - "NatGateway$ProvisionedBandwidth": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

" - } - }, - "PublicIpStringList": { - "base": null, - "refs": { - "DescribeAddressesRequest$PublicIps": "

[EC2-Classic] One or more Elastic IP addresses.

Default: Describes all your Elastic IP addresses.

" - } - }, - "Purchase": { - "base": "

Describes the result of the purchase.

", - "refs": { - "PurchaseSet$member": null - } - }, - "PurchaseHostReservationRequest": { - "base": null, - "refs": { - } - }, - "PurchaseHostReservationResult": { - "base": null, - "refs": { - } - }, - "PurchaseRequest": { - "base": "

Describes a request to purchase Scheduled Instances.

", - "refs": { - "PurchaseRequestSet$member": null - } - }, - "PurchaseRequestSet": { - "base": null, - "refs": { - "PurchaseScheduledInstancesRequest$PurchaseRequests": "

One or more purchase requests.

" - } - }, - "PurchaseReservedInstancesOfferingRequest": { - "base": "

Contains the parameters for PurchaseReservedInstancesOffering.

", - "refs": { - } - }, - "PurchaseReservedInstancesOfferingResult": { - "base": "

Contains the output of PurchaseReservedInstancesOffering.

", - "refs": { - } - }, - "PurchaseScheduledInstancesRequest": { - "base": "

Contains the parameters for PurchaseScheduledInstances.

", - "refs": { - } - }, - "PurchaseScheduledInstancesResult": { - "base": "

Contains the output of PurchaseScheduledInstances.

", - "refs": { - } - }, - "PurchaseSet": { - "base": null, - "refs": { - "GetHostReservationPurchasePreviewResult$Purchase": "

The purchase information of the Dedicated Host Reservation and the Dedicated Hosts associated with it.

", - "PurchaseHostReservationResult$Purchase": "

Describes the details of the purchase.

" - } - }, - "PurchasedScheduledInstanceSet": { - "base": null, - "refs": { - "PurchaseScheduledInstancesResult$ScheduledInstanceSet": "

Information about the Scheduled Instances.

" - } - }, - "RIProductDescription": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ProductDescription": "

The Reserved Instance product platform description. Instances that include (Amazon VPC) in the description are for use with Amazon VPC.

", - "ReservedInstances$ProductDescription": "

The Reserved Instance product platform description.

", - "ReservedInstancesOffering$ProductDescription": "

The Reserved Instance product platform description.

", - "SpotInstanceRequest$ProductDescription": "

The product description associated with the Spot instance.

", - "SpotPrice$ProductDescription": "

A general description of the AMI.

" - } - }, - "ReasonCodesList": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$ReasonCodes": "

One or more reason codes that describes the health state of your instance.

  • instance-stuck-in-state: My instance is stuck in a state.

  • unresponsive: My instance is unresponsive.

  • not-accepting-credentials: My instance is not accepting my credentials.

  • password-not-available: A password is not available for my instance.

  • performance-network: My instance is experiencing performance problems which I believe are network related.

  • performance-instance-store: My instance is experiencing performance problems which I believe are related to the instance stores.

  • performance-ebs-volume: My instance is experiencing performance problems which I believe are related to an EBS volume.

  • performance-other: My instance is experiencing performance problems.

  • other: [explain using the description parameter]

" - } - }, - "RebootInstancesRequest": { - "base": "

Contains the parameters for RebootInstances.

", - "refs": { - } - }, - "RecurringCharge": { - "base": "

Describes a recurring charge.

", - "refs": { - "RecurringChargesList$member": null - } - }, - "RecurringChargeFrequency": { - "base": null, - "refs": { - "RecurringCharge$Frequency": "

The frequency of the recurring charge.

" - } - }, - "RecurringChargesList": { - "base": null, - "refs": { - "ReservedInstances$RecurringCharges": "

The recurring charge tag assigned to the resource.

", - "ReservedInstancesOffering$RecurringCharges": "

The recurring charge tag assigned to the resource.

" - } - }, - "Region": { - "base": "

Describes a region.

", - "refs": { - "RegionList$member": null - } - }, - "RegionList": { - "base": null, - "refs": { - "DescribeRegionsResult$Regions": "

Information about one or more regions.

" - } - }, - "RegionNameStringList": { - "base": null, - "refs": { - "DescribeRegionsRequest$RegionNames": "

The names of one or more regions.

" - } - }, - "RegisterImageRequest": { - "base": "

Contains the parameters for RegisterImage.

", - "refs": { - } - }, - "RegisterImageResult": { - "base": "

Contains the output of RegisterImage.

", - "refs": { - } - }, - "RejectVpcPeeringConnectionRequest": { - "base": "

Contains the parameters for RejectVpcPeeringConnection.

", - "refs": { - } - }, - "RejectVpcPeeringConnectionResult": { - "base": "

Contains the output of RejectVpcPeeringConnection.

", - "refs": { - } - }, - "ReleaseAddressRequest": { - "base": "

Contains the parameters for ReleaseAddress.

", - "refs": { - } - }, - "ReleaseHostsRequest": { - "base": "

Contains the parameters for ReleaseHosts.

", - "refs": { - } - }, - "ReleaseHostsResult": { - "base": "

Contains the output of ReleaseHosts.

", - "refs": { - } - }, - "ReplaceNetworkAclAssociationRequest": { - "base": "

Contains the parameters for ReplaceNetworkAclAssociation.

", - "refs": { - } - }, - "ReplaceNetworkAclAssociationResult": { - "base": "

Contains the output of ReplaceNetworkAclAssociation.

", - "refs": { - } - }, - "ReplaceNetworkAclEntryRequest": { - "base": "

Contains the parameters for ReplaceNetworkAclEntry.

", - "refs": { - } - }, - "ReplaceRouteRequest": { - "base": "

Contains the parameters for ReplaceRoute.

", - "refs": { - } - }, - "ReplaceRouteTableAssociationRequest": { - "base": "

Contains the parameters for ReplaceRouteTableAssociation.

", - "refs": { - } - }, - "ReplaceRouteTableAssociationResult": { - "base": "

Contains the output of ReplaceRouteTableAssociation.

", - "refs": { - } - }, - "ReportInstanceReasonCodes": { - "base": null, - "refs": { - "ReasonCodesList$member": null - } - }, - "ReportInstanceStatusRequest": { - "base": "

Contains the parameters for ReportInstanceStatus.

", - "refs": { - } - }, - "ReportStatusType": { - "base": null, - "refs": { - "ReportInstanceStatusRequest$Status": "

The status of all instances listed.

" - } - }, - "RequestHostIdList": { - "base": null, - "refs": { - "DescribeHostsRequest$HostIds": "

The IDs of the Dedicated Hosts. The IDs are used for targeted instance launches.

", - "ModifyHostsRequest$HostIds": "

The host IDs of the Dedicated Hosts you want to modify.

", - "ReleaseHostsRequest$HostIds": "

The IDs of the Dedicated Hosts you want to release.

" - } - }, - "RequestHostIdSet": { - "base": null, - "refs": { - "GetHostReservationPurchasePreviewRequest$HostIdSet": "

The ID/s of the Dedicated Host/s that the reservation will be associated with.

", - "PurchaseHostReservationRequest$HostIdSet": "

The ID/s of the Dedicated Host/s that the reservation will be associated with.

" - } - }, - "RequestSpotFleetRequest": { - "base": "

Contains the parameters for RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotFleetResponse": { - "base": "

Contains the output of RequestSpotFleet.

", - "refs": { - } - }, - "RequestSpotInstancesRequest": { - "base": "

Contains the parameters for RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotInstancesResult": { - "base": "

Contains the output of RequestSpotInstances.

", - "refs": { - } - }, - "RequestSpotLaunchSpecification": { - "base": "

Describes the launch specification for an instance.

", - "refs": { - "RequestSpotInstancesRequest$LaunchSpecification": null - } - }, - "Reservation": { - "base": "

Describes a reservation.

", - "refs": { - "ReservationList$member": null - } - }, - "ReservationList": { - "base": null, - "refs": { - "DescribeInstancesResult$Reservations": "

Zero or more reservations.

" - } - }, - "ReservationState": { - "base": null, - "refs": { - "HostReservation$State": "

The state of the reservation.

" - } - }, - "ReservationValue": { - "base": "

The cost associated with the Reserved Instance.

", - "refs": { - "GetReservedInstancesExchangeQuoteResult$ReservedInstanceValueRollup": null, - "GetReservedInstancesExchangeQuoteResult$TargetConfigurationValueRollup": null, - "ReservedInstanceReservationValue$ReservationValue": "

The total value of the Convertible Reserved Instance that you are exchanging.

", - "TargetReservationValue$ReservationValue": "

The total value of the Convertible Reserved Instances that make up the exchange. This is the sum of the list value, remaining upfront price, and additional upfront cost of the exchange.

" - } - }, - "ReservedInstanceIdSet": { - "base": null, - "refs": { - "AcceptReservedInstancesExchangeQuoteRequest$ReservedInstanceIds": "

The IDs of the Convertible Reserved Instances that you want to exchange for other Convertible Reserved Instances of the same or higher value.

", - "GetReservedInstancesExchangeQuoteRequest$ReservedInstanceIds": "

The ID/s of the Convertible Reserved Instances you want to exchange.

" - } - }, - "ReservedInstanceLimitPrice": { - "base": "

Describes the limit price of a Reserved Instance offering.

", - "refs": { - "PurchaseReservedInstancesOfferingRequest$LimitPrice": "

Specified for Reserved Instance Marketplace offerings to limit the total order and ensure that the Reserved Instances are not purchased at unexpected prices.

" - } - }, - "ReservedInstanceReservationValue": { - "base": "

The total value of the Convertible Reserved Instance.

", - "refs": { - "ReservedInstanceReservationValueSet$member": null - } - }, - "ReservedInstanceReservationValueSet": { - "base": null, - "refs": { - "GetReservedInstancesExchangeQuoteResult$ReservedInstanceValueSet": "

The configuration of your Convertible Reserved Instances.

" - } - }, - "ReservedInstanceState": { - "base": null, - "refs": { - "ReservedInstances$State": "

The state of the Reserved Instance purchase.

" - } - }, - "ReservedInstances": { - "base": "

Describes a Reserved Instance.

", - "refs": { - "ReservedInstancesList$member": null - } - }, - "ReservedInstancesConfiguration": { - "base": "

Describes the configuration settings for the modified Reserved Instances.

", - "refs": { - "ReservedInstancesConfigurationList$member": null, - "ReservedInstancesModificationResult$TargetConfiguration": "

The target Reserved Instances configurations supplied as part of the modification request.

" - } - }, - "ReservedInstancesConfigurationList": { - "base": null, - "refs": { - "ModifyReservedInstancesRequest$TargetConfigurations": "

The configuration settings for the Reserved Instances to modify.

" - } - }, - "ReservedInstancesId": { - "base": "

Describes the ID of a Reserved Instance.

", - "refs": { - "ReservedIntancesIds$member": null - } - }, - "ReservedInstancesIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesRequest$ReservedInstancesIds": "

One or more Reserved Instance IDs.

Default: Describes all your Reserved Instances, or only those otherwise specified.

", - "ModifyReservedInstancesRequest$ReservedInstancesIds": "

The IDs of the Reserved Instances to modify.

" - } - }, - "ReservedInstancesList": { - "base": null, - "refs": { - "DescribeReservedInstancesResult$ReservedInstances": "

A list of Reserved Instances.

" - } - }, - "ReservedInstancesListing": { - "base": "

Describes a Reserved Instance listing.

", - "refs": { - "ReservedInstancesListingList$member": null - } - }, - "ReservedInstancesListingList": { - "base": null, - "refs": { - "CancelReservedInstancesListingResult$ReservedInstancesListings": "

The Reserved Instance listing.

", - "CreateReservedInstancesListingResult$ReservedInstancesListings": "

Information about the Standard Reserved Instance listing.

", - "DescribeReservedInstancesListingsResult$ReservedInstancesListings": "

Information about the Reserved Instance listing.

" - } - }, - "ReservedInstancesModification": { - "base": "

Describes a Reserved Instance modification.

", - "refs": { - "ReservedInstancesModificationList$member": null - } - }, - "ReservedInstancesModificationIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsRequest$ReservedInstancesModificationIds": "

IDs for the submitted modification request.

" - } - }, - "ReservedInstancesModificationList": { - "base": null, - "refs": { - "DescribeReservedInstancesModificationsResult$ReservedInstancesModifications": "

The Reserved Instance modification information.

" - } - }, - "ReservedInstancesModificationResult": { - "base": "

Describes the modification request/s.

", - "refs": { - "ReservedInstancesModificationResultList$member": null - } - }, - "ReservedInstancesModificationResultList": { - "base": null, - "refs": { - "ReservedInstancesModification$ModificationResults": "

Contains target configurations along with their corresponding new Reserved Instance IDs.

" - } - }, - "ReservedInstancesOffering": { - "base": "

Describes a Reserved Instance offering.

", - "refs": { - "ReservedInstancesOfferingList$member": null - } - }, - "ReservedInstancesOfferingIdStringList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsRequest$ReservedInstancesOfferingIds": "

One or more Reserved Instances offering IDs.

" - } - }, - "ReservedInstancesOfferingList": { - "base": null, - "refs": { - "DescribeReservedInstancesOfferingsResult$ReservedInstancesOfferings": "

A list of Reserved Instances offerings.

" - } - }, - "ReservedIntancesIds": { - "base": null, - "refs": { - "ReservedInstancesModification$ReservedInstancesIds": "

The IDs of one or more Reserved Instances.

" - } - }, - "ResetImageAttributeName": { - "base": null, - "refs": { - "ResetImageAttributeRequest$Attribute": "

The attribute to reset (currently you can only reset the launch permission attribute).

" - } - }, - "ResetImageAttributeRequest": { - "base": "

Contains the parameters for ResetImageAttribute.

", - "refs": { - } - }, - "ResetInstanceAttributeRequest": { - "base": "

Contains the parameters for ResetInstanceAttribute.

", - "refs": { - } - }, - "ResetNetworkInterfaceAttributeRequest": { - "base": "

Contains the parameters for ResetNetworkInterfaceAttribute.

", - "refs": { - } - }, - "ResetSnapshotAttributeRequest": { - "base": "

Contains the parameters for ResetSnapshotAttribute.

", - "refs": { - } - }, - "ResourceIdList": { - "base": null, - "refs": { - "CreateTagsRequest$Resources": "

The IDs of one or more resources to tag. For example, ami-1a2b3c4d.

", - "DeleteTagsRequest$Resources": "

The ID of the resource. For example, ami-1a2b3c4d. You can specify more than one resource ID.

" - } - }, - "ResourceType": { - "base": null, - "refs": { - "TagDescription$ResourceType": "

The resource type.

" - } - }, - "ResponseHostIdList": { - "base": null, - "refs": { - "AllocateHostsResult$HostIds": "

The ID of the allocated Dedicated Host. This is used when you want to launch an instance onto a specific host.

", - "ModifyHostsResult$Successful": "

The IDs of the Dedicated Hosts that were successfully modified.

", - "ReleaseHostsResult$Successful": "

The IDs of the Dedicated Hosts that were successfully released.

" - } - }, - "ResponseHostIdSet": { - "base": null, - "refs": { - "HostReservation$HostIdSet": "

The IDs of the Dedicated Hosts associated with the reservation.

", - "Purchase$HostIdSet": "

The IDs of the Dedicated Hosts associated with the reservation.

" - } - }, - "RestorableByStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$RestorableByUserIds": "

One or more AWS accounts IDs that can create volumes from the snapshot.

" - } - }, - "RestoreAddressToClassicRequest": { - "base": "

Contains the parameters for RestoreAddressToClassic.

", - "refs": { - } - }, - "RestoreAddressToClassicResult": { - "base": "

Contains the output of RestoreAddressToClassic.

", - "refs": { - } - }, - "RevokeSecurityGroupEgressRequest": { - "base": "

Contains the parameters for RevokeSecurityGroupEgress.

", - "refs": { - } - }, - "RevokeSecurityGroupIngressRequest": { - "base": "

Contains the parameters for RevokeSecurityGroupIngress.

", - "refs": { - } - }, - "Route": { - "base": "

Describes a route in a route table.

", - "refs": { - "RouteList$member": null - } - }, - "RouteList": { - "base": null, - "refs": { - "RouteTable$Routes": "

The routes in the route table.

" - } - }, - "RouteOrigin": { - "base": null, - "refs": { - "Route$Origin": "

Describes how the route was created.

  • CreateRouteTable - The route was automatically created when the route table was created.

  • CreateRoute - The route was manually added to the route table.

  • EnableVgwRoutePropagation - The route was propagated by route propagation.

" - } - }, - "RouteState": { - "base": null, - "refs": { - "Route$State": "

The state of the route. The blackhole state indicates that the route's target isn't available (for example, the specified gateway isn't attached to the VPC, or the specified NAT instance has been terminated).

" - } - }, - "RouteTable": { - "base": "

Describes a route table.

", - "refs": { - "CreateRouteTableResult$RouteTable": "

Information about the route table.

", - "RouteTableList$member": null - } - }, - "RouteTableAssociation": { - "base": "

Describes an association between a route table and a subnet.

", - "refs": { - "RouteTableAssociationList$member": null - } - }, - "RouteTableAssociationList": { - "base": null, - "refs": { - "RouteTable$Associations": "

The associations between the route table and one or more subnets.

" - } - }, - "RouteTableList": { - "base": null, - "refs": { - "DescribeRouteTablesResult$RouteTables": "

Information about one or more route tables.

" - } - }, - "RuleAction": { - "base": null, - "refs": { - "CreateNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "NetworkAclEntry$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

", - "ReplaceNetworkAclEntryRequest$RuleAction": "

Indicates whether to allow or deny the traffic that matches the rule.

" - } - }, - "RunInstancesMonitoringEnabled": { - "base": "

Describes the monitoring for the instance.

", - "refs": { - "LaunchSpecification$Monitoring": null, - "RequestSpotLaunchSpecification$Monitoring": null, - "RunInstancesRequest$Monitoring": "

The monitoring for the instance.

" - } - }, - "RunInstancesRequest": { - "base": "

Contains the parameters for RunInstances.

", - "refs": { - } - }, - "RunScheduledInstancesRequest": { - "base": "

Contains the parameters for RunScheduledInstances.

", - "refs": { - } - }, - "RunScheduledInstancesResult": { - "base": "

Contains the output of RunScheduledInstances.

", - "refs": { - } - }, - "S3Storage": { - "base": "

Describes the storage parameters for S3 and S3 buckets for an instance store-backed AMI.

", - "refs": { - "Storage$S3": "

An Amazon S3 storage location.

" - } - }, - "ScheduledInstance": { - "base": "

Describes a Scheduled Instance.

", - "refs": { - "PurchasedScheduledInstanceSet$member": null, - "ScheduledInstanceSet$member": null - } - }, - "ScheduledInstanceAvailability": { - "base": "

Describes a schedule that is available for your Scheduled Instances.

", - "refs": { - "ScheduledInstanceAvailabilitySet$member": null - } - }, - "ScheduledInstanceAvailabilitySet": { - "base": null, - "refs": { - "DescribeScheduledInstanceAvailabilityResult$ScheduledInstanceAvailabilitySet": "

Information about the available Scheduled Instances.

" - } - }, - "ScheduledInstanceIdRequestSet": { - "base": null, - "refs": { - "DescribeScheduledInstancesRequest$ScheduledInstanceIds": "

One or more Scheduled Instance IDs.

" - } - }, - "ScheduledInstanceRecurrence": { - "base": "

Describes the recurring schedule for a Scheduled Instance.

", - "refs": { - "ScheduledInstance$Recurrence": "

The schedule recurrence.

", - "ScheduledInstanceAvailability$Recurrence": "

The schedule recurrence.

" - } - }, - "ScheduledInstanceRecurrenceRequest": { - "base": "

Describes the recurring schedule for a Scheduled Instance.

", - "refs": { - "DescribeScheduledInstanceAvailabilityRequest$Recurrence": "

The schedule recurrence.

" - } - }, - "ScheduledInstanceSet": { - "base": null, - "refs": { - "DescribeScheduledInstancesResult$ScheduledInstanceSet": "

Information about the Scheduled Instances.

" - } - }, - "ScheduledInstancesBlockDeviceMapping": { - "base": "

Describes a block device mapping for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesBlockDeviceMappingSet$member": null - } - }, - "ScheduledInstancesBlockDeviceMappingSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$BlockDeviceMappings": "

One or more block device mapping entries.

" - } - }, - "ScheduledInstancesEbs": { - "base": "

Describes an EBS volume for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesBlockDeviceMapping$Ebs": "

Parameters used to set up EBS volumes automatically when the instance is launched.

" - } - }, - "ScheduledInstancesIamInstanceProfile": { - "base": "

Describes an IAM instance profile for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$IamInstanceProfile": "

The IAM instance profile.

" - } - }, - "ScheduledInstancesLaunchSpecification": { - "base": "

Describes the launch specification for a Scheduled Instance.

If you are launching the Scheduled Instance in EC2-VPC, you must specify the ID of the subnet. You can specify the subnet using either SubnetId or NetworkInterface.

", - "refs": { - "RunScheduledInstancesRequest$LaunchSpecification": "

The launch specification. You must match the instance type, Availability Zone, network, and platform of the schedule that you purchased.

" - } - }, - "ScheduledInstancesMonitoring": { - "base": "

Describes whether monitoring is enabled for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "ScheduledInstancesNetworkInterface": { - "base": "

Describes a network interface for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesNetworkInterfaceSet$member": null - } - }, - "ScheduledInstancesNetworkInterfaceSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$NetworkInterfaces": "

One or more network interfaces.

" - } - }, - "ScheduledInstancesPlacement": { - "base": "

Describes the placement for a Scheduled Instance.

", - "refs": { - "ScheduledInstancesLaunchSpecification$Placement": "

The placement information.

" - } - }, - "ScheduledInstancesPrivateIpAddressConfig": { - "base": "

Describes a private IP address for a Scheduled Instance.

", - "refs": { - "PrivateIpAddressConfigSet$member": null - } - }, - "ScheduledInstancesSecurityGroupIdSet": { - "base": null, - "refs": { - "ScheduledInstancesLaunchSpecification$SecurityGroupIds": "

The IDs of one or more security groups.

", - "ScheduledInstancesNetworkInterface$Groups": "

The IDs of one or more security groups.

" - } - }, - "SecurityGroup": { - "base": "

Describes a security group

", - "refs": { - "SecurityGroupList$member": null - } - }, - "SecurityGroupIdStringList": { - "base": null, - "refs": { - "CreateNetworkInterfaceRequest$Groups": "

The IDs of one or more security groups.

", - "ImportInstanceLaunchSpecification$GroupIds": "

One or more security group IDs.

", - "InstanceNetworkInterfaceSpecification$Groups": "

The IDs of the security groups for the network interface. Applies only if creating a network interface when launching an instance.

", - "ModifyNetworkInterfaceAttributeRequest$Groups": "

Changes the security groups for the network interface. The new set of groups you specify replaces the current set. You must specify at least one group, even if it's just the default security group in the VPC. You must specify the ID of the security group, not the name.

", - "RunInstancesRequest$SecurityGroupIds": "

One or more security group IDs. You can create a security group using CreateSecurityGroup.

Default: Amazon EC2 uses the default security group.

" - } - }, - "SecurityGroupList": { - "base": null, - "refs": { - "DescribeSecurityGroupsResult$SecurityGroups": "

Information about one or more security groups.

" - } - }, - "SecurityGroupReference": { - "base": "

Describes a VPC with a security group that references your security group.

", - "refs": { - "SecurityGroupReferences$member": null - } - }, - "SecurityGroupReferences": { - "base": null, - "refs": { - "DescribeSecurityGroupReferencesResult$SecurityGroupReferenceSet": "

Information about the VPCs with the referencing security groups.

" - } - }, - "SecurityGroupStringList": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$GroupNames": "

One or more security group names.

", - "RunInstancesRequest$SecurityGroups": "

[EC2-Classic, default VPC] One or more security group names. For a nondefault VPC, you must use security group IDs instead.

Default: Amazon EC2 uses the default security group.

" - } - }, - "ShutdownBehavior": { - "base": null, - "refs": { - "ImportInstanceLaunchSpecification$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

", - "RunInstancesRequest$InstanceInitiatedShutdownBehavior": "

Indicates whether an instance stops or terminates when you initiate shutdown from the instance (using the operating system command for system shutdown).

Default: stop

" - } - }, - "SlotDateTimeRangeRequest": { - "base": "

Describes the time period for a Scheduled Instance to start its first schedule. The time period must span less than one day.

", - "refs": { - "DescribeScheduledInstanceAvailabilityRequest$FirstSlotStartTimeRange": "

The time period for the first schedule to start.

" - } - }, - "SlotStartTimeRangeRequest": { - "base": "

Describes the time period for a Scheduled Instance to start its first schedule.

", - "refs": { - "DescribeScheduledInstancesRequest$SlotStartTimeRange": "

The time period for the first schedule to start.

" - } - }, - "Snapshot": { - "base": "

Describes a snapshot.

", - "refs": { - "SnapshotList$member": null - } - }, - "SnapshotAttributeName": { - "base": null, - "refs": { - "DescribeSnapshotAttributeRequest$Attribute": "

The snapshot attribute you would like to view.

", - "ModifySnapshotAttributeRequest$Attribute": "

The snapshot attribute to modify.

Only volume creation permissions may be modified at the customer level.

", - "ResetSnapshotAttributeRequest$Attribute": "

The attribute to reset. Currently, only the attribute for permission to create volumes can be reset.

" - } - }, - "SnapshotDetail": { - "base": "

Describes the snapshot created from the imported disk.

", - "refs": { - "SnapshotDetailList$member": null - } - }, - "SnapshotDetailList": { - "base": null, - "refs": { - "ImportImageResult$SnapshotDetails": "

Information about the snapshots.

", - "ImportImageTask$SnapshotDetails": "

Information about the snapshots.

" - } - }, - "SnapshotDiskContainer": { - "base": "

The disk container object for the import snapshot request.

", - "refs": { - "ImportSnapshotRequest$DiskContainer": "

Information about the disk container.

" - } - }, - "SnapshotIdStringList": { - "base": null, - "refs": { - "DescribeSnapshotsRequest$SnapshotIds": "

One or more snapshot IDs.

Default: Describes snapshots for which you have launch permissions.

" - } - }, - "SnapshotList": { - "base": null, - "refs": { - "DescribeSnapshotsResult$Snapshots": "

Information about the snapshots.

" - } - }, - "SnapshotState": { - "base": null, - "refs": { - "Snapshot$State": "

The snapshot state.

" - } - }, - "SnapshotTaskDetail": { - "base": "

Details about the import snapshot task.

", - "refs": { - "ImportSnapshotResult$SnapshotTaskDetail": "

Information about the import snapshot task.

", - "ImportSnapshotTask$SnapshotTaskDetail": "

Describes an import snapshot task.

" - } - }, - "SpotDatafeedSubscription": { - "base": "

Describes the data feed for a Spot instance.

", - "refs": { - "CreateSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

", - "DescribeSpotDatafeedSubscriptionResult$SpotDatafeedSubscription": "

The Spot instance data feed subscription.

" - } - }, - "SpotFleetLaunchSpecification": { - "base": "

Describes the launch specification for one or more Spot instances.

", - "refs": { - "LaunchSpecsList$member": null - } - }, - "SpotFleetMonitoring": { - "base": "

Describes whether monitoring is enabled.

", - "refs": { - "SpotFleetLaunchSpecification$Monitoring": "

Enable or disable monitoring for the instances.

" - } - }, - "SpotFleetRequestConfig": { - "base": "

Describes a Spot fleet request.

", - "refs": { - "SpotFleetRequestConfigSet$member": null - } - }, - "SpotFleetRequestConfigData": { - "base": "

Describes the configuration of a Spot fleet request.

", - "refs": { - "RequestSpotFleetRequest$SpotFleetRequestConfig": "

The configuration for the Spot fleet request.

", - "SpotFleetRequestConfig$SpotFleetRequestConfig": "

Information about the configuration of the Spot fleet request.

" - } - }, - "SpotFleetRequestConfigSet": { - "base": null, - "refs": { - "DescribeSpotFleetRequestsResponse$SpotFleetRequestConfigs": "

Information about the configuration of your Spot fleet.

" - } - }, - "SpotInstanceRequest": { - "base": "

Describes a Spot instance request.

", - "refs": { - "SpotInstanceRequestList$member": null - } - }, - "SpotInstanceRequestIdList": { - "base": null, - "refs": { - "CancelSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

", - "DescribeSpotInstanceRequestsRequest$SpotInstanceRequestIds": "

One or more Spot instance request IDs.

" - } - }, - "SpotInstanceRequestList": { - "base": null, - "refs": { - "DescribeSpotInstanceRequestsResult$SpotInstanceRequests": "

One or more Spot instance requests.

", - "RequestSpotInstancesResult$SpotInstanceRequests": "

One or more Spot instance requests.

" - } - }, - "SpotInstanceState": { - "base": null, - "refs": { - "SpotInstanceRequest$State": "

The state of the Spot instance request. Spot bid status information can help you track your Spot instance requests. For more information, see Spot Bid Status in the Amazon Elastic Compute Cloud User Guide.

" - } - }, - "SpotInstanceStateFault": { - "base": "

Describes a Spot instance state change.

", - "refs": { - "SpotDatafeedSubscription$Fault": "

The fault codes for the Spot instance request, if any.

", - "SpotInstanceRequest$Fault": "

The fault codes for the Spot instance request, if any.

" - } - }, - "SpotInstanceStatus": { - "base": "

Describes the status of a Spot instance request.

", - "refs": { - "SpotInstanceRequest$Status": "

The status code and status message describing the Spot instance request.

" - } - }, - "SpotInstanceType": { - "base": null, - "refs": { - "RequestSpotInstancesRequest$Type": "

The Spot instance request type.

Default: one-time

", - "SpotInstanceRequest$Type": "

The Spot instance request type.

" - } - }, - "SpotPlacement": { - "base": "

Describes Spot instance placement.

", - "refs": { - "LaunchSpecification$Placement": "

The placement information for the instance.

", - "RequestSpotLaunchSpecification$Placement": "

The placement information for the instance.

", - "SpotFleetLaunchSpecification$Placement": "

The placement information.

" - } - }, - "SpotPrice": { - "base": "

Describes the maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "refs": { - "SpotPriceHistoryList$member": null - } - }, - "SpotPriceHistoryList": { - "base": null, - "refs": { - "DescribeSpotPriceHistoryResult$SpotPriceHistory": "

The historical Spot prices.

" - } - }, - "StaleIpPermission": { - "base": "

Describes a stale rule in a security group.

", - "refs": { - "StaleIpPermissionSet$member": null - } - }, - "StaleIpPermissionSet": { - "base": null, - "refs": { - "StaleSecurityGroup$StaleIpPermissions": "

Information about the stale inbound rules in the security group.

", - "StaleSecurityGroup$StaleIpPermissionsEgress": "

Information about the stale outbound rules in the security group.

" - } - }, - "StaleSecurityGroup": { - "base": "

Describes a stale security group (a security group that contains stale rules).

", - "refs": { - "StaleSecurityGroupSet$member": null - } - }, - "StaleSecurityGroupSet": { - "base": null, - "refs": { - "DescribeStaleSecurityGroupsResult$StaleSecurityGroupSet": "

Information about the stale security groups.

" - } - }, - "StartInstancesRequest": { - "base": "

Contains the parameters for StartInstances.

", - "refs": { - } - }, - "StartInstancesResult": { - "base": "

Contains the output of StartInstances.

", - "refs": { - } - }, - "State": { - "base": null, - "refs": { - "VpcEndpoint$State": "

The state of the VPC endpoint.

" - } - }, - "StateReason": { - "base": "

Describes a state change.

", - "refs": { - "Image$StateReason": "

The reason for the state change.

", - "Instance$StateReason": "

The reason for the most recent state transition.

" - } - }, - "Status": { - "base": null, - "refs": { - "MoveAddressToVpcResult$Status": "

The status of the move of the IP address.

", - "RestoreAddressToClassicResult$Status": "

The move status for the IP address.

" - } - }, - "StatusName": { - "base": null, - "refs": { - "InstanceStatusDetails$Name": "

The type of instance status.

" - } - }, - "StatusType": { - "base": null, - "refs": { - "InstanceStatusDetails$Status": "

The status.

" - } - }, - "StopInstancesRequest": { - "base": "

Contains the parameters for StopInstances.

", - "refs": { - } - }, - "StopInstancesResult": { - "base": "

Contains the output of StopInstances.

", - "refs": { - } - }, - "Storage": { - "base": "

Describes the storage location for an instance store-backed AMI.

", - "refs": { - "BundleInstanceRequest$Storage": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "BundleTask$Storage": "

The Amazon S3 storage locations.

" - } - }, - "String": { - "base": null, - "refs": { - "AcceptReservedInstancesExchangeQuoteResult$ExchangeId": "

The ID of the successful exchange.

", - "AcceptVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "AccountAttribute$AttributeName": "

The name of the account attribute.

", - "AccountAttributeValue$AttributeValue": "

The value of the attribute.

", - "ActiveInstance$InstanceType": "

The instance type.

", - "ActiveInstance$InstanceId": "

The ID of the instance.

", - "ActiveInstance$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "Address$InstanceId": "

The ID of the instance that the address is associated with (if any).

", - "Address$PublicIp": "

The Elastic IP address.

", - "Address$AllocationId": "

The ID representing the allocation of the address for use with EC2-VPC.

", - "Address$AssociationId": "

The ID representing the association of the address with an instance in a VPC.

", - "Address$NetworkInterfaceId": "

The ID of the network interface.

", - "Address$NetworkInterfaceOwnerId": "

The ID of the AWS account that owns the network interface.

", - "Address$PrivateIpAddress": "

The private IP address associated with the Elastic IP address.

", - "AllocateAddressResult$PublicIp": "

The Elastic IP address.

", - "AllocateAddressResult$AllocationId": "

[EC2-VPC] The ID that AWS assigns to represent the allocation of the Elastic IP address for use with instances in a VPC.

", - "AllocateHostsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "AllocateHostsRequest$InstanceType": "

Specify the instance type that you want your Dedicated Hosts to be configured for. When you specify the instance type, that is the only instance type that you can launch onto that host.

", - "AllocateHostsRequest$AvailabilityZone": "

The Availability Zone for the Dedicated Hosts.

", - "AllocationIdList$member": null, - "AssignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AssociateAddressRequest$InstanceId": "

The ID of the instance. This is required for EC2-Classic. For EC2-VPC, you can specify either the instance ID or the network interface ID, but not both. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "AssociateAddressRequest$PublicIp": "

The Elastic IP address. This is required for EC2-Classic.

", - "AssociateAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. This is required for EC2-VPC.

", - "AssociateAddressRequest$NetworkInterfaceId": "

[EC2-VPC] The ID of the network interface. If the instance has more than one network interface, you must specify a network interface ID.

", - "AssociateAddressRequest$PrivateIpAddress": "

[EC2-VPC] The primary or secondary private IP address to associate with the Elastic IP address. If no private IP address is specified, the Elastic IP address is associated with the primary private IP address.

", - "AssociateAddressResult$AssociationId": "

[EC2-VPC] The ID that represents the association of the Elastic IP address with an instance.

", - "AssociateDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set, or default to associate no DHCP options with the VPC.

", - "AssociateDhcpOptionsRequest$VpcId": "

The ID of the VPC.

", - "AssociateRouteTableRequest$SubnetId": "

The ID of the subnet.

", - "AssociateRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "AssociateRouteTableResult$AssociationId": "

The route table association ID (needed to disassociate the route table).

", - "AttachClassicLinkVpcRequest$InstanceId": "

The ID of an EC2-Classic instance to link to the ClassicLink-enabled VPC.

", - "AttachClassicLinkVpcRequest$VpcId": "

The ID of a ClassicLink-enabled VPC.

", - "AttachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "AttachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttachNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "AttachNetworkInterfaceRequest$InstanceId": "

The ID of the instance.

", - "AttachNetworkInterfaceResult$AttachmentId": "

The ID of the network interface attachment.

", - "AttachVolumeRequest$VolumeId": "

The ID of the EBS volume. The volume and instance must be within the same Availability Zone.

", - "AttachVolumeRequest$InstanceId": "

The ID of the instance.

", - "AttachVolumeRequest$Device": "

The device name to expose to the instance (for example, /dev/sdh or xvdh).

", - "AttachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "AttachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "AttributeValue$Value": "

The attribute value. Note that the value is case-sensitive.

", - "AuthorizeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To authorize outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name or number. We recommend that you specify the protocol in a set of IP permissions instead.

", - "AuthorizeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. We recommend that you specify the CIDR range in a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "AuthorizeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead. For EC2-VPC, the source security group must be in the same VPC.

", - "AuthorizeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic] The AWS account number for the source security group, if the source security group is in a different account. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. Creates rules that grant full ICMP, UDP, and TCP access. To create a rule with a specific IP protocol and port range, use a set of IP permissions instead.

", - "AuthorizeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). (VPC only) Use -1 to specify all traffic. If you specify -1, traffic on all ports is allowed, regardless of any ports you specify.

", - "AuthorizeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "AvailabilityZone$ZoneName": "

The name of the Availability Zone.

", - "AvailabilityZone$RegionName": "

The name of the region.

", - "AvailabilityZoneMessage$Message": "

The message about the Availability Zone.

", - "BlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with 2 available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "BlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "BlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "BundleIdStringList$member": null, - "BundleInstanceRequest$InstanceId": "

The ID of the instance to bundle.

Type: String

Default: None

Required: Yes

", - "BundleTask$InstanceId": "

The ID of the instance associated with this bundle task.

", - "BundleTask$BundleId": "

The ID of the bundle task.

", - "BundleTask$Progress": "

The level of task completion, as a percent (for example, 20%).

", - "BundleTaskError$Code": "

The error code.

", - "BundleTaskError$Message": "

The error message.

", - "CancelBundleTaskRequest$BundleId": "

The ID of the bundle task.

", - "CancelConversionRequest$ConversionTaskId": "

The ID of the conversion task.

", - "CancelConversionRequest$ReasonMessage": "

The reason for canceling the conversion task.

", - "CancelExportTaskRequest$ExportTaskId": "

The ID of the export task. This is the ID returned by CreateInstanceExportTask.

", - "CancelImportTaskRequest$ImportTaskId": "

The ID of the import image or import snapshot task to be canceled.

", - "CancelImportTaskRequest$CancelReason": "

The reason for canceling the task.

", - "CancelImportTaskResult$ImportTaskId": "

The ID of the task being canceled.

", - "CancelImportTaskResult$State": "

The current state of the task being canceled.

", - "CancelImportTaskResult$PreviousState": "

The current state of the task being canceled.

", - "CancelReservedInstancesListingRequest$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "CancelSpotFleetRequestsError$Message": "

The description for the error code.

", - "CancelSpotFleetRequestsErrorItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelSpotFleetRequestsSuccessItem$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "CancelledSpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "ClassicLinkDnsSupport$VpcId": "

The ID of the VPC.

", - "ClassicLinkInstance$InstanceId": "

The ID of the instance.

", - "ClassicLinkInstance$VpcId": "

The ID of the VPC.

", - "ClientData$Comment": "

A user-defined comment about the disk upload.

", - "ConfirmProductInstanceRequest$ProductCode": "

The product code. This must be a product code that you own.

", - "ConfirmProductInstanceRequest$InstanceId": "

The ID of the instance.

", - "ConfirmProductInstanceResult$OwnerId": "

The AWS account ID of the instance owner. This is only present if the product code is attached to the instance.

", - "ConversionIdStringList$member": null, - "ConversionTask$ConversionTaskId": "

The ID of the conversion task.

", - "ConversionTask$ExpirationTime": "

The time when the task expires. If the upload isn't complete before the expiration time, we automatically cancel the task.

", - "ConversionTask$StatusMessage": "

The status message related to the conversion task.

", - "CopyImageRequest$SourceRegion": "

The name of the region that contains the AMI to copy.

", - "CopyImageRequest$SourceImageId": "

The ID of the AMI to copy.

", - "CopyImageRequest$Name": "

The name of the new AMI in the destination region.

", - "CopyImageRequest$Description": "

A description for the new AMI in the destination region.

", - "CopyImageRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "CopyImageRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when encrypting the snapshots of an image during a copy operation. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopyImageResult$ImageId": "

The ID of the new AMI.

", - "CopySnapshotRequest$SourceRegion": "

The ID of the region that contains the snapshot to be copied.

", - "CopySnapshotRequest$SourceSnapshotId": "

The ID of the EBS snapshot to copy.

", - "CopySnapshotRequest$Description": "

A description for the EBS snapshot.

", - "CopySnapshotRequest$DestinationRegion": "

The destination region to use in the PresignedUrl parameter of a snapshot copy operation. This parameter is only valid for specifying the destination region in a PresignedUrl parameter, where it is required.

CopySnapshot sends the snapshot copy to the regional endpoint that you send the HTTP request to, such as ec2.us-east-1.amazonaws.com (in the AWS CLI, this is specified with the --region parameter or the default region in your AWS configuration file).

", - "CopySnapshotRequest$PresignedUrl": "

The pre-signed URL that facilitates copying an encrypted snapshot. This parameter is only required when copying an encrypted snapshot with the Amazon EC2 Query API; it is available as an optional parameter in all other cases. The PresignedUrl should use the snapshot source endpoint, the CopySnapshot action, and include the SourceRegion, SourceSnapshotId, and DestinationRegion parameters. The PresignedUrl must be signed using AWS Signature Version 4. Because EBS snapshots are stored in Amazon S3, the signing algorithm for this parameter uses the same logic that is described in Authenticating Requests by Using Query Parameters (AWS Signature Version 4) in the Amazon Simple Storage Service API Reference. An invalid or improperly signed PresignedUrl will cause the copy operation to fail asynchronously, and the snapshot will move to an error state.

", - "CopySnapshotRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) CMK to use when creating the snapshot copy. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. The specified CMK must exist in the region that the snapshot is being copied to. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CopySnapshotResult$SnapshotId": "

The ID of the new snapshot.

", - "CreateCustomerGatewayRequest$PublicIp": "

The Internet-routable IP address for the customer gateway's outside interface. The address must be static.

", - "CreateFlowLogsRequest$LogGroupName": "

The name of the CloudWatch log group.

", - "CreateFlowLogsRequest$DeliverLogsPermissionArn": "

The ARN for the IAM role that's used to post flow logs to a CloudWatch Logs log group.

", - "CreateFlowLogsRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateFlowLogsResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateImageRequest$InstanceId": "

The ID of the instance.

", - "CreateImageRequest$Name": "

A name for the new image.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "CreateImageRequest$Description": "

A description for the new image.

", - "CreateImageResult$ImageId": "

The ID of the new AMI.

", - "CreateInstanceExportTaskRequest$Description": "

A description for the conversion task or the resource being exported. The maximum length is 255 bytes.

", - "CreateInstanceExportTaskRequest$InstanceId": "

The ID of the instance.

", - "CreateKeyPairRequest$KeyName": "

A unique name for the key pair.

Constraints: Up to 255 ASCII characters

", - "CreateNatGatewayRequest$SubnetId": "

The subnet in which to create the NAT gateway.

", - "CreateNatGatewayRequest$AllocationId": "

The allocation ID of an Elastic IP address to associate with the NAT gateway. If the Elastic IP address is associated with another resource, you must first disassociate it.

", - "CreateNatGatewayRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

Constraint: Maximum 64 ASCII characters.

", - "CreateNatGatewayResult$ClientToken": "

Unique, case-sensitive identifier to ensure the idempotency of the request. Only returned if a client token was provided in the request.

", - "CreateNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "CreateNetworkAclEntryRequest$Protocol": "

The protocol. A value of -1 means all protocols.

", - "CreateNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation (for example 172.16.0.0/24).

", - "CreateNetworkAclRequest$VpcId": "

The ID of the VPC.

", - "CreateNetworkInterfaceRequest$SubnetId": "

The ID of the subnet to associate with the network interface.

", - "CreateNetworkInterfaceRequest$Description": "

A description for the network interface.

", - "CreateNetworkInterfaceRequest$PrivateIpAddress": "

The primary private IP address of the network interface. If you don't specify an IP address, Amazon EC2 selects one for you from the subnet range. If you specify an IP address, you cannot indicate any IP addresses specified in privateIpAddresses as primary (only one IP address can be designated as primary).

", - "CreatePlacementGroupRequest$GroupName": "

A name for the placement group.

Constraints: Up to 255 ASCII characters

", - "CreateReservedInstancesListingRequest$ReservedInstancesId": "

The ID of the active Standard Reserved Instance.

", - "CreateReservedInstancesListingRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "CreateRouteRequest$RouteTableId": "

The ID of the route table for the route.

", - "CreateRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. Routing decisions are based on the most specific match.

", - "CreateRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway attached to your VPC.

", - "CreateRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC. The operation fails if you specify an instance ID unless exactly one network interface is attached.

", - "CreateRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "CreateRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "CreateRouteRequest$NatGatewayId": "

The ID of a NAT gateway.

", - "CreateRouteTableRequest$VpcId": "

The ID of the VPC.

", - "CreateSecurityGroupRequest$GroupName": "

The name of the security group.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=&;{}!$*

", - "CreateSecurityGroupRequest$Description": "

A description for the security group. This is informational only.

Constraints: Up to 255 characters in length

Constraints for EC2-Classic: ASCII characters

Constraints for EC2-VPC: a-z, A-Z, 0-9, spaces, and ._-:/()#,@[]+=&;{}!$*

", - "CreateSecurityGroupRequest$VpcId": "

[EC2-VPC] The ID of the VPC. Required for EC2-VPC.

", - "CreateSecurityGroupResult$GroupId": "

The ID of the security group.

", - "CreateSnapshotRequest$VolumeId": "

The ID of the EBS volume.

", - "CreateSnapshotRequest$Description": "

A description for the snapshot.

", - "CreateSpotDatafeedSubscriptionRequest$Bucket": "

The Amazon S3 bucket in which to store the Spot instance data feed.

", - "CreateSpotDatafeedSubscriptionRequest$Prefix": "

A prefix for the data feed file names.

", - "CreateSubnetRequest$VpcId": "

The ID of the VPC.

", - "CreateSubnetRequest$CidrBlock": "

The network range for the subnet, in CIDR notation. For example, 10.0.0.0/24.

", - "CreateSubnetRequest$AvailabilityZone": "

The Availability Zone for the subnet.

Default: AWS selects one for you. If you create more than one subnet in your VPC, we may not necessarily select a different zone for each subnet.

", - "CreateVolumePermission$UserId": "

The specific AWS account ID that is to be added or removed from a volume's list of create volume permissions.

", - "CreateVolumeRequest$SnapshotId": "

The snapshot from which to create the volume.

", - "CreateVolumeRequest$AvailabilityZone": "

The Availability Zone in which to create the volume. Use DescribeAvailabilityZones to list the Availability Zones that are currently available to you.

", - "CreateVolumeRequest$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) to use when creating the encrypted volume. This parameter is only required if you want to use a non-default CMK; if this parameter is not specified, the default CMK for EBS is used. The ARN contains the arn:aws:kms namespace, followed by the region of the CMK, the AWS account ID of the CMK owner, the key namespace, and then the CMK ID. For example, arn:aws:kms:us-east-1:012345678910:key/abcd1234-a123-456a-a12b-a123b4cd56ef. If a KmsKeyId is specified, the Encrypted flag must also be set.

", - "CreateVpcEndpointRequest$VpcId": "

The ID of the VPC in which the endpoint will be used.

", - "CreateVpcEndpointRequest$ServiceName": "

The AWS service name, in the form com.amazonaws.region.service . To get a list of available services, use the DescribeVpcEndpointServices request.

", - "CreateVpcEndpointRequest$PolicyDocument": "

A policy to attach to the endpoint that controls access to the service. The policy must be in valid JSON format. If this parameter is not specified, we attach a default policy that allows full access to the service.

", - "CreateVpcEndpointRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency.

", - "CreateVpcEndpointResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", - "CreateVpcPeeringConnectionRequest$VpcId": "

The ID of the requester VPC.

", - "CreateVpcPeeringConnectionRequest$PeerVpcId": "

The ID of the VPC with which you are creating the VPC peering connection.

", - "CreateVpcPeeringConnectionRequest$PeerOwnerId": "

The AWS account ID of the owner of the peer VPC.

Default: Your AWS account ID

", - "CreateVpcRequest$CidrBlock": "

The network range for the VPC, in CIDR notation. For example, 10.0.0.0/16.

", - "CreateVpnConnectionRequest$Type": "

The type of VPN connection (ipsec.1).

", - "CreateVpnConnectionRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "CreateVpnConnectionRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "CreateVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "CreateVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "CreateVpnGatewayRequest$AvailabilityZone": "

The Availability Zone for the virtual private gateway.

", - "CustomerGateway$CustomerGatewayId": "

The ID of the customer gateway.

", - "CustomerGateway$State": "

The current state of the customer gateway (pending | available | deleting | deleted).

", - "CustomerGateway$Type": "

The type of VPN connection the customer gateway supports (ipsec.1).

", - "CustomerGateway$IpAddress": "

The Internet-routable IP address of the customer gateway's outside interface.

", - "CustomerGateway$BgpAsn": "

The customer gateway's Border Gateway Protocol (BGP) Autonomous System Number (ASN).

", - "CustomerGatewayIdStringList$member": null, - "DeleteCustomerGatewayRequest$CustomerGatewayId": "

The ID of the customer gateway.

", - "DeleteDhcpOptionsRequest$DhcpOptionsId": "

The ID of the DHCP options set.

", - "DeleteInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DeleteKeyPairRequest$KeyName": "

The name of the key pair.

", - "DeleteNatGatewayRequest$NatGatewayId": "

The ID of the NAT gateway.

", - "DeleteNatGatewayResult$NatGatewayId": "

The ID of the NAT gateway.

", - "DeleteNetworkAclEntryRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkAclRequest$NetworkAclId": "

The ID of the network ACL.

", - "DeleteNetworkInterfaceRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DeletePlacementGroupRequest$GroupName": "

The name of the placement group.

", - "DeleteRouteRequest$RouteTableId": "

The ID of the route table.

", - "DeleteRouteRequest$DestinationCidrBlock": "

The CIDR range for the route. The value you specify must match the CIDR for the route exactly.

", - "DeleteRouteTableRequest$RouteTableId": "

The ID of the route table.

", - "DeleteSecurityGroupRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group. You can specify either the security group name or the security group ID.

", - "DeleteSecurityGroupRequest$GroupId": "

The ID of the security group. Required for a nondefault VPC.

", - "DeleteSnapshotRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DeleteSubnetRequest$SubnetId": "

The ID of the subnet.

", - "DeleteVolumeRequest$VolumeId": "

The ID of the volume.

", - "DeleteVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "DeleteVpcRequest$VpcId": "

The ID of the VPC.

", - "DeleteVpnConnectionRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$VpnConnectionId": "

The ID of the VPN connection.

", - "DeleteVpnConnectionRouteRequest$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer network.

", - "DeleteVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DeregisterImageRequest$ImageId": "

The ID of the AMI.

", - "DescribeClassicLinkInstancesRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeClassicLinkInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeFlowLogsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeFlowLogsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostReservationOfferingsRequest$OfferingId": "

The ID of the reservation offering.

", - "DescribeHostReservationOfferingsRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeHostReservationOfferingsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostReservationsRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeHostReservationsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeHostsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeHostsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "DescribeIdentityIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "DescribeIdentityIdFormatRequest$PrincipalArn": "

The ARN of the principal, which can be an IAM role, IAM user, or the root user.

", - "DescribeImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "DescribeImportImageTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportImageTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeImportSnapshotTasksRequest$NextToken": "

A token that indicates the next page of results.

", - "DescribeImportSnapshotTasksResult$NextToken": "

The token to use to get the next page of results. This value is null when there are no more results to return.

", - "DescribeInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "DescribeInstanceStatusRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeInstanceStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeInstancesRequest$NextToken": "

The token to request the next page of results.

", - "DescribeInstancesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeMovingAddressesRequest$NextToken": "

The token to use to retrieve the next page of results.

", - "DescribeMovingAddressesResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNatGatewaysRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeNatGatewaysResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribeNetworkInterfaceAttributeResult$NetworkInterfaceId": "

The ID of the network interface.

", - "DescribePrefixListsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribePrefixListsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesId": "

One or more Reserved Instance IDs.

", - "DescribeReservedInstancesListingsRequest$ReservedInstancesListingId": "

One or more Reserved Instance listing IDs.

", - "DescribeReservedInstancesModificationsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesModificationsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeReservedInstancesOfferingsRequest$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "DescribeReservedInstancesOfferingsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeReservedInstancesOfferingsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeScheduledInstanceAvailabilityRequest$NextToken": "

The token for the next set of results.

", - "DescribeScheduledInstanceAvailabilityResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeScheduledInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeScheduledInstancesResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSnapshotAttributeRequest$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotAttributeResult$SnapshotId": "

The ID of the EBS snapshot.

", - "DescribeSnapshotsRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeSnapshots request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeSnapshotsResult$NextToken": "

The NextToken value to include in a future DescribeSnapshots request. When the results of a DescribeSnapshots request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetInstancesRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetInstancesResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetInstancesResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestHistoryRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestHistoryResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "DescribeSpotFleetRequestHistoryResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotFleetRequestsRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotFleetRequestsResponse$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeSpotPriceHistoryRequest$AvailabilityZone": "

Filters the results by the specified Availability Zone.

", - "DescribeSpotPriceHistoryRequest$NextToken": "

The token for the next set of results.

", - "DescribeSpotPriceHistoryResult$NextToken": "

The token required to retrieve the next set of results. This value is null when there are no more results to return.

", - "DescribeStaleSecurityGroupsRequest$VpcId": "

The ID of the VPC.

", - "DescribeStaleSecurityGroupsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeTagsRequest$NextToken": "

The token to retrieve the next page of results.

", - "DescribeTagsResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return..

", - "DescribeVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "DescribeVolumeAttributeResult$VolumeId": "

The ID of the volume.

", - "DescribeVolumeStatusRequest$NextToken": "

The NextToken value to include in a future DescribeVolumeStatus request. When the results of the request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumeStatusResult$NextToken": "

The token to use to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVolumesRequest$NextToken": "

The NextToken value returned from a previous paginated DescribeVolumes request where MaxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the NextToken value. This value is null when there are no more results to return.

", - "DescribeVolumesResult$NextToken": "

The NextToken value to include in a future DescribeVolumes request. When the results of a DescribeVolumes request exceed MaxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "DescribeVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "DescribeVpcAttributeResult$VpcId": "

The ID of the VPC.

", - "DescribeVpcEndpointServicesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointServicesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeVpcEndpointsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a prior call.)

", - "DescribeVpcEndpointsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DetachClassicLinkVpcRequest$InstanceId": "

The ID of the instance to unlink from the VPC.

", - "DetachClassicLinkVpcRequest$VpcId": "

The ID of the VPC to which the instance is linked.

", - "DetachInternetGatewayRequest$InternetGatewayId": "

The ID of the Internet gateway.

", - "DetachInternetGatewayRequest$VpcId": "

The ID of the VPC.

", - "DetachNetworkInterfaceRequest$AttachmentId": "

The ID of the attachment.

", - "DetachVolumeRequest$VolumeId": "

The ID of the volume.

", - "DetachVolumeRequest$InstanceId": "

The ID of the instance.

", - "DetachVolumeRequest$Device": "

The device name.

", - "DetachVpnGatewayRequest$VpnGatewayId": "

The ID of the virtual private gateway.

", - "DetachVpnGatewayRequest$VpcId": "

The ID of the VPC.

", - "DhcpConfiguration$Key": "

The name of a DHCP option.

", - "DhcpOptions$DhcpOptionsId": "

The ID of the set of DHCP options.

", - "DhcpOptionsIdStringList$member": null, - "DisableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "DisableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "DisableVpcClassicLinkDnsSupportRequest$VpcId": "

The ID of the VPC.

", - "DisableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "DisassociateAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "DisassociateAddressRequest$AssociationId": "

[EC2-VPC] The association ID. Required for EC2-VPC.

", - "DisassociateRouteTableRequest$AssociationId": "

The association ID representing the current association between the route table and subnet.

", - "DiskImage$Description": "

A description of the disk image.

", - "DiskImageDescription$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DiskImageDescription$Checksum": "

The checksum computed for the disk image.

", - "DiskImageDetail$ImportManifestUrl": "

A presigned URL for the import manifest stored in Amazon S3 and presented here as an Amazon S3 presigned URL. For information about creating a presigned URL for an Amazon S3 object, read the \"Query String Request Authentication Alternative\" section of the Authenticating REST Requests topic in the Amazon Simple Storage Service Developer Guide.

For information about the import manifest referenced by this API action, see VM Import Manifest.

", - "DiskImageVolumeDescription$Id": "

The volume identifier.

", - "EbsBlockDevice$SnapshotId": "

The ID of the snapshot.

", - "EbsInstanceBlockDevice$VolumeId": "

The ID of the EBS volume.

", - "EbsInstanceBlockDeviceSpecification$VolumeId": "

The ID of the EBS volume.

", - "EnableVgwRoutePropagationRequest$RouteTableId": "

The ID of the route table.

", - "EnableVgwRoutePropagationRequest$GatewayId": "

The ID of the virtual private gateway.

", - "EnableVolumeIORequest$VolumeId": "

The ID of the volume.

", - "EnableVpcClassicLinkDnsSupportRequest$VpcId": "

The ID of the VPC.

", - "EnableVpcClassicLinkRequest$VpcId": "

The ID of the VPC.

", - "EventInformation$InstanceId": "

The ID of the instance. This information is available only for instanceChange events.

", - "EventInformation$EventSubType": "

The event.

The following are the error events.

  • iamFleetRoleInvalid - The Spot fleet did not have the required permissions either to launch or terminate an instance.

  • launchSpecTemporarilyBlacklisted - The configuration is not valid and several attempts to launch instances have failed. For more information, see the description of the event.

  • spotFleetRequestConfigurationInvalid - The configuration is not valid. For more information, see the description of the event.

  • spotInstanceCountLimitExceeded - You've reached the limit on the number of Spot instances that you can launch.

The following are the fleetRequestChange events.

  • active - The Spot fleet has been validated and Amazon EC2 is attempting to maintain the target number of running Spot instances.

  • cancelled - The Spot fleet is canceled and has no running Spot instances. The Spot fleet will be deleted two days after its instances were terminated.

  • cancelled_running - The Spot fleet is canceled and will not launch additional Spot instances, but its existing Spot instances continue to run until they are interrupted or terminated.

  • cancelled_terminating - The Spot fleet is canceled and its Spot instances are terminating.

  • expired - The Spot fleet request has expired. A subsequent event indicates that the instances were terminated, if the request was created with TerminateInstancesWithExpiration set.

  • modify_in_progress - A request to modify the Spot fleet request was accepted and is in progress.

  • modify_successful - The Spot fleet request was modified.

  • price_update - The bid price for a launch configuration was adjusted because it was too high. This change is permanent.

  • submitted - The Spot fleet request is being evaluated and Amazon EC2 is preparing to launch the target number of Spot instances.

The following are the instanceChange events.

  • launched - A bid was fulfilled and a new instance was launched.

  • terminated - An instance was terminated by the user.

", - "EventInformation$EventDescription": "

The description of the event.

", - "ExecutableByStringList$member": null, - "ExportTask$ExportTaskId": "

The ID of the export task.

", - "ExportTask$Description": "

A description of the resource being exported.

", - "ExportTask$StatusMessage": "

The status message related to the export task.

", - "ExportTaskIdStringList$member": null, - "ExportToS3Task$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3Task$S3Key": "

The encryption key for your S3 bucket.

", - "ExportToS3TaskSpecification$S3Bucket": "

The S3 bucket for the destination image. The destination bucket must exist and grant WRITE and READ_ACP permissions to the AWS account vm-import-export@amazon.com.

", - "ExportToS3TaskSpecification$S3Prefix": "

The image is written to a single object in the S3 bucket at the S3 key s3prefix + exportTaskId + '.' + diskImageFormat.

", - "Filter$Name": "

The name of the filter. Filter names are case-sensitive.

", - "FlowLog$FlowLogId": "

The flow log ID.

", - "FlowLog$FlowLogStatus": "

The status of the flow log (ACTIVE).

", - "FlowLog$ResourceId": "

The ID of the resource on which the flow log was created.

", - "FlowLog$LogGroupName": "

The name of the flow log group.

", - "FlowLog$DeliverLogsStatus": "

The status of the logs delivery (SUCCESS | FAILED).

", - "FlowLog$DeliverLogsErrorMessage": "

Information about the error that occurred. Rate limited indicates that CloudWatch logs throttling has been applied for one or more network interfaces, or that you've reached the limit on the number of CloudWatch Logs log groups that you can create. Access error indicates that the IAM role associated with the flow log does not have sufficient permissions to publish to CloudWatch Logs. Unknown error indicates an internal error.

", - "FlowLog$DeliverLogsPermissionArn": "

The ARN of the IAM role that posts logs to CloudWatch Logs.

", - "GetConsoleOutputRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$InstanceId": "

The ID of the instance.

", - "GetConsoleOutputResult$Output": "

The console output, Base64-encoded. If using a command line tool, the tool decodes the output for you.

", - "GetConsoleScreenshotRequest$InstanceId": "

The ID of the instance.

", - "GetConsoleScreenshotResult$InstanceId": "

The ID of the instance.

", - "GetConsoleScreenshotResult$ImageData": "

The data that comprises the image.

", - "GetHostReservationPurchasePreviewRequest$OfferingId": "

The offering ID of the reservation.

", - "GetHostReservationPurchasePreviewResult$TotalUpfrontPrice": "

The potential total upfront price. This is billed immediately.

", - "GetHostReservationPurchasePreviewResult$TotalHourlyPrice": "

The potential total hourly price of the reservation per hour.

", - "GetPasswordDataRequest$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$InstanceId": "

The ID of the Windows instance.

", - "GetPasswordDataResult$PasswordData": "

The password of the instance.

", - "GetReservedInstancesExchangeQuoteResult$PaymentDue": "

The total true upfront charge for the exchange.

", - "GetReservedInstancesExchangeQuoteResult$CurrencyCode": "

The currency of the transaction.

", - "GetReservedInstancesExchangeQuoteResult$ValidationFailureReason": "

Describes the reason why the exchange can not be completed.

", - "GroupIdStringList$member": null, - "GroupIdentifier$GroupName": "

The name of the security group.

", - "GroupIdentifier$GroupId": "

The ID of the security group.

", - "GroupIds$member": null, - "GroupNameStringList$member": null, - "Host$HostId": "

The ID of the Dedicated Host.

", - "Host$HostReservationId": "

The reservation ID of the Dedicated Host. This returns a null response if the Dedicated Host doesn't have an associated reservation.

", - "Host$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "Host$AvailabilityZone": "

The Availability Zone of the Dedicated Host.

", - "HostInstance$InstanceId": "

the IDs of instances that are running on the Dedicated Host.

", - "HostInstance$InstanceType": "

The instance type size (for example, m3.medium) of the running instance.

", - "HostOffering$OfferingId": "

The ID of the offering.

", - "HostOffering$InstanceFamily": "

The instance family of the offering.

", - "HostOffering$UpfrontPrice": "

The upfront price of the offering. Does not apply to No Upfront offerings.

", - "HostOffering$HourlyPrice": "

The hourly price of the offering.

", - "HostProperties$InstanceType": "

The instance type size that the Dedicated Host supports (for example, m3.medium).

", - "HostReservation$HostReservationId": "

The ID of the reservation that specifies the associated Dedicated Hosts.

", - "HostReservation$OfferingId": "

The ID of the reservation. This remains the same regardless of which Dedicated Hosts are associated with it.

", - "HostReservation$InstanceFamily": "

The instance family of the Dedicated Host Reservation. The instance family on the Dedicated Host must be the same in order for it to benefit from the reservation.

", - "HostReservation$HourlyPrice": "

The hourly price of the reservation.

", - "HostReservation$UpfrontPrice": "

The upfront price of the reservation.

", - "HostReservationIdSet$member": null, - "IamInstanceProfile$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfile$Id": "

The ID of the instance profile.

", - "IamInstanceProfileSpecification$Arn": "

The Amazon Resource Name (ARN) of the instance profile.

", - "IamInstanceProfileSpecification$Name": "

The name of the instance profile.

", - "IdFormat$Resource": "

The type of resource.

", - "Image$ImageId": "

The ID of the AMI.

", - "Image$ImageLocation": "

The location of the AMI.

", - "Image$OwnerId": "

The AWS account ID of the image owner.

", - "Image$CreationDate": "

The date and time the image was created.

", - "Image$KernelId": "

The kernel associated with the image, if any. Only applicable for machine images.

", - "Image$RamdiskId": "

The RAM disk associated with the image, if any. Only applicable for machine images.

", - "Image$SriovNetSupport": "

Specifies whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "Image$ImageOwnerAlias": "

The AWS account alias (for example, amazon, self) or the AWS account ID of the AMI owner.

", - "Image$Name": "

The name of the AMI that was provided during image creation.

", - "Image$Description": "

The description of the AMI that was provided during image creation.

", - "Image$RootDeviceName": "

The device name of the root device (for example, /dev/sda1 or /dev/xvda).

", - "ImageAttribute$ImageId": "

The ID of the AMI.

", - "ImageDiskContainer$Description": "

The description of the disk image.

", - "ImageDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "ImageDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. The URL can either be a https URL (https://..) or an Amazon S3 URL (s3://..)

", - "ImageDiskContainer$DeviceName": "

The block device mapping for the disk.

", - "ImageDiskContainer$SnapshotId": "

The ID of the EBS snapshot to be used for importing the snapshot.

", - "ImageIdStringList$member": null, - "ImportImageRequest$Description": "

A description string for the import image task.

", - "ImportImageRequest$LicenseType": "

The license type to be used for the Amazon Machine Image (AMI) after importing.

Note: You may only use BYOL if you have existing licenses with rights to use these licenses in a third party cloud like AWS. For more information, see Prerequisites in the VM Import/Export User Guide.

Valid values: AWS | BYOL

", - "ImportImageRequest$Hypervisor": "

The target hypervisor platform.

Valid values: xen

", - "ImportImageRequest$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageRequest$Platform": "

The operating system of the virtual machine.

Valid values: Windows | Linux

", - "ImportImageRequest$ClientToken": "

The token to enable idempotency for VM import requests.

", - "ImportImageRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportImageResult$ImportTaskId": "

The task ID of the import image task.

", - "ImportImageResult$Architecture": "

The architecture of the virtual machine.

", - "ImportImageResult$LicenseType": "

The license type of the virtual machine.

", - "ImportImageResult$Platform": "

The operating system of the virtual machine.

", - "ImportImageResult$Hypervisor": "

The target hypervisor of the import task.

", - "ImportImageResult$Description": "

A description of the import task.

", - "ImportImageResult$ImageId": "

The ID of the Amazon Machine Image (AMI) created by the import task.

", - "ImportImageResult$Progress": "

The progress of the task.

", - "ImportImageResult$StatusMessage": "

A detailed status message of the import task.

", - "ImportImageResult$Status": "

A brief status of the task.

", - "ImportImageTask$ImportTaskId": "

The ID of the import image task.

", - "ImportImageTask$Architecture": "

The architecture of the virtual machine.

Valid values: i386 | x86_64

", - "ImportImageTask$LicenseType": "

The license type of the virtual machine.

", - "ImportImageTask$Platform": "

The description string for the import image task.

", - "ImportImageTask$Hypervisor": "

The target hypervisor for the import task.

Valid values: xen

", - "ImportImageTask$Description": "

A description of the import task.

", - "ImportImageTask$ImageId": "

The ID of the Amazon Machine Image (AMI) of the imported virtual machine.

", - "ImportImageTask$Progress": "

The percentage of progress of the import image task.

", - "ImportImageTask$StatusMessage": "

A descriptive status message for the import image task.

", - "ImportImageTask$Status": "

A brief status for the import image task.

", - "ImportInstanceLaunchSpecification$AdditionalInfo": "

Reserved.

", - "ImportInstanceLaunchSpecification$SubnetId": "

[EC2-VPC] The ID of the subnet in which to launch the instance.

", - "ImportInstanceLaunchSpecification$PrivateIpAddress": "

[EC2-VPC] An available IP address from the IP address range of the subnet.

", - "ImportInstanceRequest$Description": "

A description for the instance being imported.

", - "ImportInstanceTaskDetails$InstanceId": "

The ID of the instance.

", - "ImportInstanceTaskDetails$Description": "

A description of the task.

", - "ImportInstanceVolumeDetailItem$AvailabilityZone": "

The Availability Zone where the resulting instance will reside.

", - "ImportInstanceVolumeDetailItem$Status": "

The status of the import of this particular disk image.

", - "ImportInstanceVolumeDetailItem$StatusMessage": "

The status information or errors related to the disk image.

", - "ImportInstanceVolumeDetailItem$Description": "

A description of the task.

", - "ImportKeyPairRequest$KeyName": "

A unique name for the key pair.

", - "ImportKeyPairResult$KeyName": "

The key pair name you provided.

", - "ImportKeyPairResult$KeyFingerprint": "

The MD5 public key fingerprint as specified in section 4 of RFC 4716.

", - "ImportSnapshotRequest$Description": "

The description string for the import snapshot task.

", - "ImportSnapshotRequest$ClientToken": "

Token to enable idempotency for VM import requests.

", - "ImportSnapshotRequest$RoleName": "

The name of the role to use when not using the default role, 'vmimport'.

", - "ImportSnapshotResult$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotResult$Description": "

A description of the import snapshot task.

", - "ImportSnapshotTask$ImportTaskId": "

The ID of the import snapshot task.

", - "ImportSnapshotTask$Description": "

A description of the import snapshot task.

", - "ImportTaskIdList$member": null, - "ImportVolumeRequest$AvailabilityZone": "

The Availability Zone for the resulting EBS volume.

", - "ImportVolumeRequest$Description": "

A description of the volume.

", - "ImportVolumeTaskDetails$AvailabilityZone": "

The Availability Zone where the resulting volume will reside.

", - "ImportVolumeTaskDetails$Description": "

The description you provided when starting the import volume task.

", - "Instance$InstanceId": "

The ID of the instance.

", - "Instance$ImageId": "

The ID of the AMI used to launch the instance.

", - "Instance$PrivateDnsName": "

The private DNS name assigned to the instance. This DNS name can only be used inside the Amazon EC2 network. This name is not available until the instance enters the running state. For EC2-VPC, this name is only available if you've enabled DNS hostnames for your VPC.

", - "Instance$PublicDnsName": "

The public DNS name assigned to the instance. This name is not available until the instance enters the running state. For EC2-VPC, this name is only available if you've enabled DNS hostnames for your VPC.

", - "Instance$StateTransitionReason": "

The reason for the most recent state transition. This might be an empty string.

", - "Instance$KeyName": "

The name of the key pair, if this instance was launched with an associated key pair.

", - "Instance$KernelId": "

The kernel associated with this instance, if applicable.

", - "Instance$RamdiskId": "

The RAM disk associated with this instance, if applicable.

", - "Instance$SubnetId": "

[EC2-VPC] The ID of the subnet in which the instance is running.

", - "Instance$VpcId": "

[EC2-VPC] The ID of the VPC in which the instance is running.

", - "Instance$PrivateIpAddress": "

The private IP address assigned to the instance.

", - "Instance$PublicIpAddress": "

The public IP address assigned to the instance, if applicable.

", - "Instance$RootDeviceName": "

The root device name (for example, /dev/sda1 or /dev/xvda).

", - "Instance$SpotInstanceRequestId": "

If the request is a Spot instance request, the ID of the request.

", - "Instance$ClientToken": "

The idempotency token you provided when you launched the instance, if applicable.

", - "Instance$SriovNetSupport": "

Specifies whether enhanced networking with the Intel 82599 Virtual Function interface is enabled.

", - "InstanceAttribute$InstanceId": "

The ID of the instance.

", - "InstanceBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "InstanceBlockDeviceMappingSpecification$VirtualName": "

The virtual device name.

", - "InstanceBlockDeviceMappingSpecification$NoDevice": "

suppress the specified device included in the block device mapping.

", - "InstanceCapacity$InstanceType": "

The instance type size supported by the Dedicated Host.

", - "InstanceExportDetails$InstanceId": "

The ID of the resource being exported.

", - "InstanceIdSet$member": null, - "InstanceIdStringList$member": null, - "InstanceMonitoring$InstanceId": "

The ID of the instance.

", - "InstanceNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterface$SubnetId": "

The ID of the subnet.

", - "InstanceNetworkInterface$VpcId": "

The ID of the VPC.

", - "InstanceNetworkInterface$Description": "

The description.

", - "InstanceNetworkInterface$OwnerId": "

The ID of the AWS account that created the network interface.

", - "InstanceNetworkInterface$MacAddress": "

The MAC address.

", - "InstanceNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "InstanceNetworkInterface$PrivateDnsName": "

The private DNS name.

", - "InstanceNetworkInterfaceAssociation$PublicIp": "

The public IP address or Elastic IP address bound to the network interface.

", - "InstanceNetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "InstanceNetworkInterfaceAssociation$IpOwnerId": "

The ID of the owner of the Elastic IP address.

", - "InstanceNetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "InstanceNetworkInterfaceSpecification$NetworkInterfaceId": "

The ID of the network interface.

", - "InstanceNetworkInterfaceSpecification$SubnetId": "

The ID of the subnet associated with the network string. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$Description": "

The description of the network interface. Applies only if creating a network interface when launching an instance.

", - "InstanceNetworkInterfaceSpecification$PrivateIpAddress": "

The private IP address of the network interface. Applies only if creating a network interface when launching an instance. You cannot specify this option if you're launching more than one instance in a RunInstances request.

", - "InstancePrivateIpAddress$PrivateIpAddress": "

The private IP address of the network interface.

", - "InstancePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "InstanceStateChange$InstanceId": "

The ID of the instance.

", - "InstanceStatus$InstanceId": "

The ID of the instance.

", - "InstanceStatus$AvailabilityZone": "

The Availability Zone of the instance.

", - "InstanceStatusEvent$Description": "

A description of the event.

After a scheduled event is completed, it can still be described for up to a week. If the event has been completed, this description starts with the following text: [Completed].

", - "InternetGateway$InternetGatewayId": "

The ID of the Internet gateway.

", - "InternetGatewayAttachment$VpcId": "

The ID of the VPC.

", - "IpPermission$IpProtocol": "

The IP protocol name (for tcp, udp, and icmp) or number (see Protocol Numbers).

[EC2-VPC only] When you authorize or revoke security group rules, you can use -1 to specify all.

", - "IpRange$CidrIp": "

The CIDR range. You can either specify a CIDR range or a source security group, not both.

", - "IpRanges$member": null, - "KeyNameStringList$member": null, - "KeyPair$KeyName": "

The name of the key pair.

", - "KeyPair$KeyFingerprint": "

The SHA-1 digest of the DER encoded private key.

", - "KeyPair$KeyMaterial": "

An unencrypted PEM encoded RSA private key.

", - "KeyPairInfo$KeyName": "

The name of the key pair.

", - "KeyPairInfo$KeyFingerprint": "

If you used CreateKeyPair to create the key pair, this is the SHA-1 digest of the DER encoded private key. If you used ImportKeyPair to provide AWS the public key, this is the MD5 public key fingerprint as specified in section 4 of RFC4716.

", - "LaunchPermission$UserId": "

The AWS account ID.

", - "LaunchSpecification$ImageId": "

The ID of the AMI.

", - "LaunchSpecification$KeyName": "

The name of the key pair.

", - "LaunchSpecification$UserData": "

The user data to make available to the instances. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "LaunchSpecification$AddressingType": "

Deprecated.

", - "LaunchSpecification$KernelId": "

The ID of the kernel.

", - "LaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "LaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "ModifyIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "ModifyIdentityIdFormatRequest$Resource": "

The type of resource: instance | reservation | snapshot | volume

", - "ModifyIdentityIdFormatRequest$PrincipalArn": "

The ARN of the principal, which can be an IAM user, IAM role, or the root user. Specify all to modify the ID format for all IAM users, IAM roles, and the root user of the account.

", - "ModifyImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ModifyImageAttributeRequest$Attribute": "

The name of the attribute to modify.

", - "ModifyImageAttributeRequest$Value": "

The value of the attribute being modified. This is only valid when modifying the description attribute.

", - "ModifyInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ModifyInstanceAttributeRequest$Value": "

A new value for the attribute. Use only with the kernel, ramdisk, userData, disableApiTermination, or instanceInitiatedShutdownBehavior attribute.

", - "ModifyInstancePlacementRequest$InstanceId": "

The ID of the instance that you are modifying.

", - "ModifyInstancePlacementRequest$HostId": "

The ID of the Dedicated Host that the instance will have affinity with.

", - "ModifyNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ModifyReservedInstancesRequest$ClientToken": "

A unique, case-sensitive token you provide to ensure idempotency of your modification request. For more information, see Ensuring Idempotency.

", - "ModifyReservedInstancesResult$ReservedInstancesModificationId": "

The ID for the modification.

", - "ModifySnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ModifySpotFleetRequestRequest$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "ModifySubnetAttributeRequest$SubnetId": "

The ID of the subnet.

", - "ModifyVolumeAttributeRequest$VolumeId": "

The ID of the volume.

", - "ModifyVpcAttributeRequest$VpcId": "

The ID of the VPC.

", - "ModifyVpcEndpointRequest$VpcEndpointId": "

The ID of the endpoint.

", - "ModifyVpcEndpointRequest$PolicyDocument": "

A policy document to attach to the endpoint. The policy must be in valid JSON format.

", - "ModifyVpcPeeringConnectionOptionsRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "MoveAddressToVpcRequest$PublicIp": "

The Elastic IP address.

", - "MoveAddressToVpcResult$AllocationId": "

The allocation ID for the Elastic IP address.

", - "MovingAddressStatus$PublicIp": "

The Elastic IP address.

", - "NatGateway$VpcId": "

The ID of the VPC in which the NAT gateway is located.

", - "NatGateway$SubnetId": "

The ID of the subnet in which the NAT gateway is located.

", - "NatGateway$NatGatewayId": "

The ID of the NAT gateway.

", - "NatGateway$FailureCode": "

If the NAT gateway could not be created, specifies the error code for the failure. (InsufficientFreeAddressesInSubnet | Gateway.NotAttached | InvalidAllocationID.NotFound | Resource.AlreadyAssociated | InternalError | InvalidSubnetID.NotFound)

", - "NatGateway$FailureMessage": "

If the NAT gateway could not be created, specifies the error message for the failure, that corresponds to the error code.

  • For InsufficientFreeAddressesInSubnet: \"Subnet has insufficient free addresses to create this NAT gateway\"

  • For Gateway.NotAttached: \"Network vpc-xxxxxxxx has no Internet gateway attached\"

  • For InvalidAllocationID.NotFound: \"Elastic IP address eipalloc-xxxxxxxx could not be associated with this NAT gateway\"

  • For Resource.AlreadyAssociated: \"Elastic IP address eipalloc-xxxxxxxx is already associated\"

  • For InternalError: \"Network interface eni-xxxxxxxx, created and used internally by this NAT gateway is in an invalid state. Please try again.\"

  • For InvalidSubnetID.NotFound: \"The specified subnet subnet-xxxxxxxx does not exist or could not be found.\"

", - "NatGatewayAddress$PublicIp": "

The Elastic IP address associated with the NAT gateway.

", - "NatGatewayAddress$AllocationId": "

The allocation ID of the Elastic IP address that's associated with the NAT gateway.

", - "NatGatewayAddress$PrivateIp": "

The private IP address associated with the Elastic IP address.

", - "NatGatewayAddress$NetworkInterfaceId": "

The ID of the network interface associated with the NAT gateway.

", - "NetworkAcl$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAcl$VpcId": "

The ID of the VPC for the network ACL.

", - "NetworkAclAssociation$NetworkAclAssociationId": "

The ID of the association between a network ACL and a subnet.

", - "NetworkAclAssociation$NetworkAclId": "

The ID of the network ACL.

", - "NetworkAclAssociation$SubnetId": "

The ID of the subnet.

", - "NetworkAclEntry$Protocol": "

The protocol. A value of -1 means all protocols.

", - "NetworkAclEntry$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "NetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "NetworkInterface$SubnetId": "

The ID of the subnet.

", - "NetworkInterface$VpcId": "

The ID of the VPC.

", - "NetworkInterface$AvailabilityZone": "

The Availability Zone.

", - "NetworkInterface$Description": "

A description.

", - "NetworkInterface$OwnerId": "

The AWS account ID of the owner of the network interface.

", - "NetworkInterface$RequesterId": "

The ID of the entity that launched the instance on your behalf (for example, AWS Management Console or Auto Scaling).

", - "NetworkInterface$MacAddress": "

The MAC address.

", - "NetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "NetworkInterface$PrivateDnsName": "

The private DNS name.

", - "NetworkInterfaceAssociation$PublicIp": "

The address of the Elastic IP address bound to the network interface.

", - "NetworkInterfaceAssociation$PublicDnsName": "

The public DNS name.

", - "NetworkInterfaceAssociation$IpOwnerId": "

The ID of the Elastic IP address owner.

", - "NetworkInterfaceAssociation$AllocationId": "

The allocation ID.

", - "NetworkInterfaceAssociation$AssociationId": "

The association ID.

", - "NetworkInterfaceAttachment$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceAttachment$InstanceId": "

The ID of the instance.

", - "NetworkInterfaceAttachment$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "NetworkInterfaceAttachmentChanges$AttachmentId": "

The ID of the network interface attachment.

", - "NetworkInterfaceIdList$member": null, - "NetworkInterfacePrivateIpAddress$PrivateIpAddress": "

The private IP address.

", - "NetworkInterfacePrivateIpAddress$PrivateDnsName": "

The private DNS name.

", - "NewDhcpConfiguration$Key": null, - "OwnerStringList$member": null, - "Placement$AvailabilityZone": "

The Availability Zone of the instance.

", - "Placement$GroupName": "

The name of the placement group the instance is in (for cluster compute instances).

", - "Placement$HostId": "

The ID of the Dedicted host on which the instance resides. This parameter is not support for the ImportInstance command.

", - "Placement$Affinity": "

The affinity setting for the instance on the Dedicated Host. This parameter is not supported for the ImportInstance command.

", - "PlacementGroup$GroupName": "

The name of the placement group.

", - "PlacementGroupStringList$member": null, - "PrefixList$PrefixListId": "

The ID of the prefix.

", - "PrefixList$PrefixListName": "

The name of the prefix.

", - "PrefixListId$PrefixListId": "

The ID of the prefix.

", - "PrefixListIdSet$member": null, - "PrivateIpAddressSpecification$PrivateIpAddress": "

The private IP addresses.

", - "PrivateIpAddressStringList$member": null, - "ProductCode$ProductCodeId": "

The product code.

", - "ProductCodeStringList$member": null, - "ProductDescriptionList$member": null, - "PropagatingVgw$GatewayId": "

The ID of the virtual private gateway (VGW).

", - "ProvisionedBandwidth$Provisioned": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$Requested": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "ProvisionedBandwidth$Status": "

Reserved. If you need to sustain traffic greater than the documented limits, contact us through the Support Center.

", - "PublicIpStringList$member": null, - "Purchase$HostReservationId": "

The ID of the reservation.

", - "Purchase$InstanceFamily": "

The instance family on the Dedicated Host that the reservation can be associated with.

", - "Purchase$UpfrontPrice": "

The upfront price of the reservation.

", - "Purchase$HourlyPrice": "

The hourly price of the reservation per hour.

", - "PurchaseHostReservationRequest$OfferingId": "

The ID of the offering.

", - "PurchaseHostReservationRequest$LimitPrice": "

The specified limit is checked against the total upfront cost of the reservation (calculated as the offering's upfront cost multiplied by the host count). If the total upfront cost is greater than the specified price limit, the request will fail. This is used to ensure that the purchase does not exceed the expected upfront cost of the purchase. At this time, the only supported currency is USD. For example, to indicate a limit price of USD 100, specify 100.00.

", - "PurchaseHostReservationRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "PurchaseHostReservationResult$TotalUpfrontPrice": "

The total amount that will be charged to your account when you purchase the reservation.

", - "PurchaseHostReservationResult$TotalHourlyPrice": "

The total hourly price of the reservation calculated per hour.

", - "PurchaseHostReservationResult$ClientToken": "

Unique, case-sensitive identifier you provide to ensure idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide

", - "PurchaseRequest$PurchaseToken": "

The purchase token.

", - "PurchaseReservedInstancesOfferingRequest$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering to purchase.

", - "PurchaseReservedInstancesOfferingResult$ReservedInstancesId": "

The IDs of the purchased Reserved Instances.

", - "PurchaseScheduledInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that ensures the idempotency of the request. For more information, see Ensuring Idempotency.

", - "Region$RegionName": "

The name of the region.

", - "Region$Endpoint": "

The region service endpoint.

", - "RegionNameStringList$member": null, - "RegisterImageRequest$ImageLocation": "

The full path to your AMI manifest in Amazon S3 storage.

", - "RegisterImageRequest$Name": "

A name for your AMI.

Constraints: 3-128 alphanumeric characters, parentheses (()), square brackets ([]), spaces ( ), periods (.), slashes (/), dashes (-), single quotes ('), at-signs (@), or underscores(_)

", - "RegisterImageRequest$Description": "

A description for your AMI.

", - "RegisterImageRequest$KernelId": "

The ID of the kernel.

", - "RegisterImageRequest$RamdiskId": "

The ID of the RAM disk.

", - "RegisterImageRequest$RootDeviceName": "

The name of the root device (for example, /dev/sda1, or /dev/xvda).

", - "RegisterImageRequest$VirtualizationType": "

The type of virtualization.

Default: paravirtual

", - "RegisterImageRequest$SriovNetSupport": "

Set to simple to enable enhanced networking with the Intel 82599 Virtual Function interface for the AMI and any instances that you launch from the AMI.

There is no way to disable sriovNetSupport at this time.

This option is supported only for HVM AMIs. Specifying this option with a PV AMI can make instances launched from the AMI unreachable.

", - "RegisterImageResult$ImageId": "

The ID of the newly registered AMI.

", - "RejectVpcPeeringConnectionRequest$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "ReleaseAddressRequest$PublicIp": "

[EC2-Classic] The Elastic IP address. Required for EC2-Classic.

", - "ReleaseAddressRequest$AllocationId": "

[EC2-VPC] The allocation ID. Required for EC2-VPC.

", - "ReplaceNetworkAclAssociationRequest$AssociationId": "

The ID of the current association between the original network ACL and the subnet.

", - "ReplaceNetworkAclAssociationRequest$NetworkAclId": "

The ID of the new network ACL to associate with the subnet.

", - "ReplaceNetworkAclAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReplaceNetworkAclEntryRequest$NetworkAclId": "

The ID of the ACL.

", - "ReplaceNetworkAclEntryRequest$Protocol": "

The IP protocol. You can specify all or -1 to mean all protocols.

", - "ReplaceNetworkAclEntryRequest$CidrBlock": "

The network range to allow or deny, in CIDR notation.

", - "ReplaceRouteRequest$RouteTableId": "

The ID of the route table.

", - "ReplaceRouteRequest$DestinationCidrBlock": "

The CIDR address block used for the destination match. The value you provide must match the CIDR of an existing route in the table.

", - "ReplaceRouteRequest$GatewayId": "

The ID of an Internet gateway or virtual private gateway.

", - "ReplaceRouteRequest$InstanceId": "

The ID of a NAT instance in your VPC.

", - "ReplaceRouteRequest$NetworkInterfaceId": "

The ID of a network interface.

", - "ReplaceRouteRequest$VpcPeeringConnectionId": "

The ID of a VPC peering connection.

", - "ReplaceRouteRequest$NatGatewayId": "

The ID of a NAT gateway.

", - "ReplaceRouteTableAssociationRequest$AssociationId": "

The association ID.

", - "ReplaceRouteTableAssociationRequest$RouteTableId": "

The ID of the new route table to associate with the subnet.

", - "ReplaceRouteTableAssociationResult$NewAssociationId": "

The ID of the new association.

", - "ReportInstanceStatusRequest$Description": "

Descriptive text about the health state of your instance.

", - "RequestHostIdList$member": null, - "RequestHostIdSet$member": null, - "RequestSpotFleetResponse$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "RequestSpotInstancesRequest$SpotPrice": "

The maximum hourly price (bid) for any Spot instance launched to fulfill the request.

", - "RequestSpotInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that you provide to ensure the idempotency of the request. For more information, see How to Ensure Idempotency in the Amazon Elastic Compute Cloud User Guide.

", - "RequestSpotInstancesRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

Default: Instances are launched and terminated individually

", - "RequestSpotInstancesRequest$AvailabilityZoneGroup": "

The user-specified name for a logical grouping of bids.

When you specify an Availability Zone group in a Spot Instance request, all Spot instances in the request are launched in the same Availability Zone. Instance proximity is maintained with this parameter, but the choice of Availability Zone is not. The group applies only to bids for Spot Instances of the same instance type. Any additional Spot instance requests that are specified with the same Availability Zone group name are launched in that same Availability Zone, as long as at least one instance from the group is still active.

If there is no active instance running in the Availability Zone group that you specify for a new Spot instance request (all instances are terminated, the bid is expired, or the bid falls below current market), then Amazon EC2 launches the instance in any Availability Zone where the constraint can be met. Consequently, the subsequent set of Spot instances could be placed in a different zone from the original request, even if you specified the same Availability Zone group.

Default: Instances are launched in any available Availability Zone.

", - "RequestSpotLaunchSpecification$ImageId": "

The ID of the AMI.

", - "RequestSpotLaunchSpecification$KeyName": "

The name of the key pair.

", - "RequestSpotLaunchSpecification$UserData": "

The user data to make available to the instances. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "RequestSpotLaunchSpecification$AddressingType": "

Deprecated.

", - "RequestSpotLaunchSpecification$KernelId": "

The ID of the kernel.

", - "RequestSpotLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "RequestSpotLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instance.

", - "Reservation$ReservationId": "

The ID of the reservation.

", - "Reservation$OwnerId": "

The ID of the AWS account that owns the reservation.

", - "Reservation$RequesterId": "

The ID of the requester that launched the instances on your behalf (for example, AWS Management Console or Auto Scaling).

", - "ReservationValue$RemainingTotalValue": "

The balance of the total value (the sum of remainingUpfrontValue + hourlyPrice * number of hours remaining).

", - "ReservationValue$RemainingUpfrontValue": "

The remaining upfront cost of the reservation.

", - "ReservationValue$HourlyPrice": "

The hourly rate of the reservation.

", - "ReservedInstanceIdSet$member": null, - "ReservedInstanceReservationValue$ReservedInstanceId": "

The ID of the Convertible Reserved Instance that you are exchanging.

", - "ReservedInstances$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstances$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesConfiguration$AvailabilityZone": "

The Availability Zone for the modified Reserved Instances.

", - "ReservedInstancesConfiguration$Platform": "

The network platform of the modified Reserved Instances, which is either EC2-Classic or EC2-VPC.

", - "ReservedInstancesId$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesIdStringList$member": null, - "ReservedInstancesListing$ReservedInstancesListingId": "

The ID of the Reserved Instance listing.

", - "ReservedInstancesListing$ReservedInstancesId": "

The ID of the Reserved Instance.

", - "ReservedInstancesListing$StatusMessage": "

The reason for the current status of the Reserved Instance listing. The response can be blank.

", - "ReservedInstancesListing$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModification$ReservedInstancesModificationId": "

A unique ID for the Reserved Instance modification.

", - "ReservedInstancesModification$Status": "

The status of the Reserved Instances modification request.

", - "ReservedInstancesModification$StatusMessage": "

The reason for the status.

", - "ReservedInstancesModification$ClientToken": "

A unique, case-sensitive key supplied by the client to ensure that the request is idempotent. For more information, see Ensuring Idempotency.

", - "ReservedInstancesModificationIdStringList$member": null, - "ReservedInstancesModificationResult$ReservedInstancesId": "

The ID for the Reserved Instances that were created as part of the modification request. This field is only available when the modification is fulfilled.

", - "ReservedInstancesOffering$ReservedInstancesOfferingId": "

The ID of the Reserved Instance offering. This is the offering ID used in GetReservedInstancesExchangeQuote to confirm that an exchange can be made.

", - "ReservedInstancesOffering$AvailabilityZone": "

The Availability Zone in which the Reserved Instance can be used.

", - "ReservedInstancesOfferingIdStringList$member": null, - "ResetImageAttributeRequest$ImageId": "

The ID of the AMI.

", - "ResetInstanceAttributeRequest$InstanceId": "

The ID of the instance.

", - "ResetNetworkInterfaceAttributeRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "ResetNetworkInterfaceAttributeRequest$SourceDestCheck": "

The source/destination checking attribute. Resets the value to true.

", - "ResetSnapshotAttributeRequest$SnapshotId": "

The ID of the snapshot.

", - "ResourceIdList$member": null, - "ResponseHostIdList$member": null, - "ResponseHostIdSet$member": null, - "RestorableByStringList$member": null, - "RestoreAddressToClassicRequest$PublicIp": "

The Elastic IP address.

", - "RestoreAddressToClassicResult$PublicIp": "

The Elastic IP address.

", - "RevokeSecurityGroupEgressRequest$GroupId": "

The ID of the security group.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupName": "

The name of a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$SourceSecurityGroupOwnerId": "

The AWS account number for a destination security group. To revoke outbound access to a destination security group, we recommend that you use a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$IpProtocol": "

The IP protocol name or number. We recommend that you specify the protocol in a set of IP permissions instead.

", - "RevokeSecurityGroupEgressRequest$CidrIp": "

The CIDR IP address range. We recommend that you specify the CIDR range in a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$GroupName": "

[EC2-Classic, default VPC] The name of the security group.

", - "RevokeSecurityGroupIngressRequest$GroupId": "

The ID of the security group. Required for a security group in a nondefault VPC.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupName": "

[EC2-Classic, default VPC] The name of the source security group. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the start of the port range, the IP protocol, and the end of the port range. For EC2-VPC, the source security group must be in the same VPC. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$SourceSecurityGroupOwnerId": "

[EC2-Classic] The AWS account ID of the source security group, if the source security group is in a different account. You can't specify this parameter in combination with the following parameters: the CIDR IP address range, the IP protocol, the start of the port range, and the end of the port range. To revoke a specific rule for an IP protocol and port range, use a set of IP permissions instead.

", - "RevokeSecurityGroupIngressRequest$IpProtocol": "

The IP protocol name (tcp, udp, icmp) or number (see Protocol Numbers). Use -1 to specify all.

", - "RevokeSecurityGroupIngressRequest$CidrIp": "

The CIDR IP address range. You can't specify this parameter when specifying a source security group.

", - "Route$DestinationCidrBlock": "

The CIDR block used for the destination match.

", - "Route$DestinationPrefixListId": "

The prefix of the AWS service.

", - "Route$GatewayId": "

The ID of a gateway attached to your VPC.

", - "Route$InstanceId": "

The ID of a NAT instance in your VPC.

", - "Route$InstanceOwnerId": "

The AWS account ID of the owner of the instance.

", - "Route$NetworkInterfaceId": "

The ID of the network interface.

", - "Route$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "Route$NatGatewayId": "

The ID of a NAT gateway.

", - "RouteTable$RouteTableId": "

The ID of the route table.

", - "RouteTable$VpcId": "

The ID of the VPC.

", - "RouteTableAssociation$RouteTableAssociationId": "

The ID of the association between a route table and a subnet.

", - "RouteTableAssociation$RouteTableId": "

The ID of the route table.

", - "RouteTableAssociation$SubnetId": "

The ID of the subnet. A subnet ID is not returned for an implicit association.

", - "RunInstancesRequest$ImageId": "

The ID of the AMI, which you can get by calling DescribeImages.

", - "RunInstancesRequest$KeyName": "

The name of the key pair. You can create a key pair using CreateKeyPair or ImportKeyPair.

If you do not specify a key pair, you can't connect to the instance unless you choose an AMI that is configured to allow users another way to log in.

", - "RunInstancesRequest$UserData": "

The user data to make available to the instance. For more information, see Running Commands on Your Linux Instance at Launch (Linux) and Adding User Data (Windows). If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "RunInstancesRequest$KernelId": "

The ID of the kernel.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$RamdiskId": "

The ID of the RAM disk.

We recommend that you use PV-GRUB instead of kernels and RAM disks. For more information, see PV-GRUB in the Amazon Elastic Compute Cloud User Guide.

", - "RunInstancesRequest$SubnetId": "

[EC2-VPC] The ID of the subnet to launch the instance into.

", - "RunInstancesRequest$PrivateIpAddress": "

[EC2-VPC] The primary IP address. You must specify a value from the IP address range of the subnet.

Only one private IP address can be designated as primary. Therefore, you can't specify this parameter if PrivateIpAddresses.n.Primary is set to true and PrivateIpAddresses.n.PrivateIpAddress is set to an IP address.

You cannot specify this option if you're launching more than one instance in the request.

Default: We select an IP address from the IP address range of the subnet.

", - "RunInstancesRequest$ClientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. For more information, see Ensuring Idempotency.

Constraints: Maximum 64 ASCII characters

", - "RunInstancesRequest$AdditionalInfo": "

Reserved.

", - "RunScheduledInstancesRequest$ClientToken": "

Unique, case-sensitive identifier that ensures the idempotency of the request. For more information, see Ensuring Idempotency.

", - "RunScheduledInstancesRequest$ScheduledInstanceId": "

The Scheduled Instance ID.

", - "S3Storage$Bucket": "

The bucket in which to store the AMI. You can specify a bucket that you already own or a new bucket that Amazon EC2 creates on your behalf. If you specify a bucket that belongs to someone else, Amazon EC2 returns an error.

", - "S3Storage$Prefix": "

The beginning of the file name of the AMI.

", - "S3Storage$AWSAccessKeyId": "

The access key ID of the owner of the bucket. Before you specify a value for your access key ID, review and follow the guidance in Best Practices for Managing AWS Access Keys.

", - "S3Storage$UploadPolicySignature": "

The signature of the JSON document.

", - "ScheduledInstance$ScheduledInstanceId": "

The Scheduled Instance ID.

", - "ScheduledInstance$InstanceType": "

The instance type.

", - "ScheduledInstance$Platform": "

The platform (Linux/UNIX or Windows).

", - "ScheduledInstance$NetworkPlatform": "

The network platform (EC2-Classic or EC2-VPC).

", - "ScheduledInstance$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstance$HourlyPrice": "

The hourly price for a single instance.

", - "ScheduledInstanceAvailability$InstanceType": "

The instance type. You can specify one of the C3, C4, M4, or R3 instance types.

", - "ScheduledInstanceAvailability$Platform": "

The platform (Linux/UNIX or Windows).

", - "ScheduledInstanceAvailability$NetworkPlatform": "

The network platform (EC2-Classic or EC2-VPC).

", - "ScheduledInstanceAvailability$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstanceAvailability$PurchaseToken": "

The purchase token. This token expires in two hours.

", - "ScheduledInstanceAvailability$HourlyPrice": "

The hourly price for a single instance.

", - "ScheduledInstanceIdRequestSet$member": null, - "ScheduledInstanceRecurrence$Frequency": "

The frequency (Daily, Weekly, or Monthly).

", - "ScheduledInstanceRecurrence$OccurrenceUnit": "

The unit for occurrenceDaySet (DayOfWeek or DayOfMonth).

", - "ScheduledInstanceRecurrenceRequest$Frequency": "

The frequency (Daily, Weekly, or Monthly).

", - "ScheduledInstanceRecurrenceRequest$OccurrenceUnit": "

The unit for OccurrenceDays (DayOfWeek or DayOfMonth). This value is required for a monthly schedule. You can't specify DayOfWeek with a weekly schedule. You can't specify this value with a daily schedule.

", - "ScheduledInstancesBlockDeviceMapping$DeviceName": "

The device name exposed to the instance (for example, /dev/sdh or xvdh).

", - "ScheduledInstancesBlockDeviceMapping$NoDevice": "

Suppresses the specified device included in the block device mapping of the AMI.

", - "ScheduledInstancesBlockDeviceMapping$VirtualName": "

The virtual device name (ephemeralN). Instance store volumes are numbered starting from 0. An instance type with two available instance store volumes can specify mappings for ephemeral0 and ephemeral1.The number of available instance store volumes depends on the instance type. After you connect to the instance, you must mount the volume.

Constraints: For M3 instances, you must specify instance store volumes in the block device mapping for the instance. When you launch an M3 instance, we ignore any instance store volumes specified in the block device mapping for the AMI.

", - "ScheduledInstancesEbs$SnapshotId": "

The ID of the snapshot.

", - "ScheduledInstancesEbs$VolumeType": "

The volume type. gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, Throughput Optimized HDD for st1, Cold HDD for sc1, or standard for Magnetic.

Default: standard

", - "ScheduledInstancesIamInstanceProfile$Arn": "

The Amazon Resource Name (ARN).

", - "ScheduledInstancesIamInstanceProfile$Name": "

The name.

", - "ScheduledInstancesLaunchSpecification$ImageId": "

The ID of the Amazon Machine Image (AMI).

", - "ScheduledInstancesLaunchSpecification$KeyName": "

The name of the key pair.

", - "ScheduledInstancesLaunchSpecification$UserData": "

The base64-encoded MIME user data.

", - "ScheduledInstancesLaunchSpecification$KernelId": "

The ID of the kernel.

", - "ScheduledInstancesLaunchSpecification$InstanceType": "

The instance type.

", - "ScheduledInstancesLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "ScheduledInstancesLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances.

", - "ScheduledInstancesNetworkInterface$NetworkInterfaceId": "

The ID of the network interface.

", - "ScheduledInstancesNetworkInterface$SubnetId": "

The ID of the subnet.

", - "ScheduledInstancesNetworkInterface$Description": "

The description.

", - "ScheduledInstancesNetworkInterface$PrivateIpAddress": "

The IP address of the network interface within the subnet.

", - "ScheduledInstancesPlacement$AvailabilityZone": "

The Availability Zone.

", - "ScheduledInstancesPlacement$GroupName": "

The name of the placement group.

", - "ScheduledInstancesPrivateIpAddressConfig$PrivateIpAddress": "

The IP address.

", - "ScheduledInstancesSecurityGroupIdSet$member": null, - "SecurityGroup$OwnerId": "

The AWS account ID of the owner of the security group.

", - "SecurityGroup$GroupName": "

The name of the security group.

", - "SecurityGroup$GroupId": "

The ID of the security group.

", - "SecurityGroup$Description": "

A description of the security group.

", - "SecurityGroup$VpcId": "

[EC2-VPC] The ID of the VPC for the security group.

", - "SecurityGroupIdStringList$member": null, - "SecurityGroupReference$GroupId": "

The ID of your security group.

", - "SecurityGroupReference$ReferencingVpcId": "

The ID of the VPC with the referencing security group.

", - "SecurityGroupReference$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "SecurityGroupStringList$member": null, - "Snapshot$SnapshotId": "

The ID of the snapshot. Each snapshot receives a unique identifier when it is created.

", - "Snapshot$VolumeId": "

The ID of the volume that was used to create the snapshot. Snapshots created by the CopySnapshot action have an arbitrary volume ID that should not be used for any purpose.

", - "Snapshot$StateMessage": "

Encrypted Amazon EBS snapshots are copied asynchronously. If a snapshot copy operation fails (for example, if the proper AWS Key Management Service (AWS KMS) permissions are not obtained) this field displays error state details to help you diagnose why the error occurred. This parameter is only returned by the DescribeSnapshots API operation.

", - "Snapshot$Progress": "

The progress of the snapshot, as a percentage.

", - "Snapshot$OwnerId": "

The AWS account ID of the EBS snapshot owner.

", - "Snapshot$Description": "

The description for the snapshot.

", - "Snapshot$OwnerAlias": "

Value from an Amazon-maintained list (amazon | aws-marketplace | microsoft) of snapshot owners. Not to be confused with the user-configured AWS account alias, which is set from the IAM console.

", - "Snapshot$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the parent volume.

", - "Snapshot$DataEncryptionKeyId": "

The data encryption key identifier for the snapshot. This value is a unique identifier that corresponds to the data encryption key that was used to encrypt the original volume or snapshot copy. Because data encryption keys are inherited by volumes created from snapshots, and vice versa, if snapshots share the same data encryption key identifier, then they belong to the same volume/snapshot lineage. This parameter is only returned by the DescribeSnapshots API operation.

", - "SnapshotDetail$Description": "

A description for the snapshot.

", - "SnapshotDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotDetail$Url": "

The URL used to access the disk image.

", - "SnapshotDetail$DeviceName": "

The block device mapping for the snapshot.

", - "SnapshotDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotDetail$Progress": "

The percentage of progress for the task.

", - "SnapshotDetail$StatusMessage": "

A detailed status message for the snapshot creation.

", - "SnapshotDetail$Status": "

A brief status of the snapshot creation.

", - "SnapshotDiskContainer$Description": "

The description of the disk image being imported.

", - "SnapshotDiskContainer$Format": "

The format of the disk image being imported.

Valid values: RAW | VHD | VMDK | OVA

", - "SnapshotDiskContainer$Url": "

The URL to the Amazon S3-based disk image being imported. It can either be a https URL (https://..) or an Amazon S3 URL (s3://..).

", - "SnapshotIdStringList$member": null, - "SnapshotTaskDetail$Description": "

The description of the snapshot.

", - "SnapshotTaskDetail$Format": "

The format of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$Url": "

The URL of the disk image from which the snapshot is created.

", - "SnapshotTaskDetail$SnapshotId": "

The snapshot ID of the disk being imported.

", - "SnapshotTaskDetail$Progress": "

The percentage of completion for the import snapshot task.

", - "SnapshotTaskDetail$StatusMessage": "

A detailed status message for the import snapshot task.

", - "SnapshotTaskDetail$Status": "

A brief status for the import snapshot task.

", - "SpotDatafeedSubscription$OwnerId": "

The AWS account ID of the account.

", - "SpotDatafeedSubscription$Bucket": "

The Amazon S3 bucket where the Spot instance data feed is located.

", - "SpotDatafeedSubscription$Prefix": "

The prefix that is prepended to data feed files.

", - "SpotFleetLaunchSpecification$ImageId": "

The ID of the AMI.

", - "SpotFleetLaunchSpecification$KeyName": "

The name of the key pair.

", - "SpotFleetLaunchSpecification$UserData": "

The user data to make available to the instances. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "SpotFleetLaunchSpecification$AddressingType": "

Deprecated.

", - "SpotFleetLaunchSpecification$KernelId": "

The ID of the kernel.

", - "SpotFleetLaunchSpecification$RamdiskId": "

The ID of the RAM disk.

", - "SpotFleetLaunchSpecification$SubnetId": "

The ID of the subnet in which to launch the instances. To specify multiple subnets, separate them using commas; for example, \"subnet-a61dafcf, subnet-65ea5f08\".

", - "SpotFleetLaunchSpecification$SpotPrice": "

The bid price per unit hour for the specified instance type. If this value is not specified, the default is the Spot bid price specified for the fleet. To determine the bid price per unit hour, divide the Spot bid price by the value of WeightedCapacity.

", - "SpotFleetRequestConfig$SpotFleetRequestId": "

The ID of the Spot fleet request.

", - "SpotFleetRequestConfigData$ClientToken": "

A unique, case-sensitive identifier you provide to ensure idempotency of your listings. This helps avoid duplicate listings. For more information, see Ensuring Idempotency.

", - "SpotFleetRequestConfigData$SpotPrice": "

The bid price per unit hour.

", - "SpotFleetRequestConfigData$IamFleetRole": "

Grants the Spot fleet permission to terminate Spot instances on your behalf when you cancel its Spot fleet request using CancelSpotFleetRequests or when the Spot fleet request expires, if you set terminateInstancesWithExpiration.

", - "SpotInstanceRequest$SpotInstanceRequestId": "

The ID of the Spot instance request.

", - "SpotInstanceRequest$SpotPrice": "

The maximum hourly price (bid) for the Spot instance launched to fulfill the request.

", - "SpotInstanceRequest$LaunchGroup": "

The instance launch group. Launch groups are Spot instances that launch together and terminate together.

", - "SpotInstanceRequest$AvailabilityZoneGroup": "

The Availability Zone group. If you specify the same Availability Zone group for all Spot instance requests, all Spot instances are launched in the same Availability Zone.

", - "SpotInstanceRequest$InstanceId": "

The instance ID, if an instance has been launched to fulfill the Spot instance request.

", - "SpotInstanceRequest$ActualBlockHourlyPrice": "

If you specified a duration and your Spot instance request was fulfilled, this is the fixed hourly price in effect for the Spot instance while it runs.

", - "SpotInstanceRequest$LaunchedAvailabilityZone": "

The Availability Zone in which the bid is launched.

", - "SpotInstanceRequestIdList$member": null, - "SpotInstanceStateFault$Code": "

The reason code for the Spot instance state change.

", - "SpotInstanceStateFault$Message": "

The message for the Spot instance state change.

", - "SpotInstanceStatus$Code": "

The status code. For a list of status codes, see Spot Bid Status Codes in the Amazon Elastic Compute Cloud User Guide.

", - "SpotInstanceStatus$Message": "

The description for the status code.

", - "SpotPlacement$AvailabilityZone": "

The Availability Zone.

[Spot fleet only] To specify multiple Availability Zones, separate them using commas; for example, \"us-west-2a, us-west-2b\".

", - "SpotPlacement$GroupName": "

The name of the placement group (for cluster instances).

", - "SpotPrice$SpotPrice": "

The maximum price (bid) that you are willing to pay for a Spot instance.

", - "SpotPrice$AvailabilityZone": "

The Availability Zone.

", - "StaleIpPermission$IpProtocol": "

The IP protocol name (for tcp, udp, and icmp) or number (see Protocol Numbers).

", - "StaleSecurityGroup$GroupId": "

The ID of the security group.

", - "StaleSecurityGroup$GroupName": "

The name of the security group.

", - "StaleSecurityGroup$Description": "

The description of the security group.

", - "StaleSecurityGroup$VpcId": "

The ID of the VPC for the security group.

", - "StartInstancesRequest$AdditionalInfo": "

Reserved.

", - "StateReason$Code": "

The reason code for the state change.

", - "StateReason$Message": "

The message for the state change.

  • Server.SpotInstanceTermination: A Spot instance was terminated due to an increase in the market price.

  • Server.InternalError: An internal error occurred during instance launch, resulting in termination.

  • Server.InsufficientInstanceCapacity: There was insufficient instance capacity to satisfy the launch request.

  • Client.InternalError: A client error caused the instance to terminate on launch.

  • Client.InstanceInitiatedShutdown: The instance was shut down using the shutdown -h command from the instance.

  • Client.UserInitiatedShutdown: The instance was shut down using the Amazon EC2 API.

  • Client.VolumeLimitExceeded: The limit on the number of EBS volumes or total storage was exceeded. Decrease usage or request an increase in your limits.

  • Client.InvalidSnapshot.NotFound: The specified snapshot was not found.

", - "Subnet$SubnetId": "

The ID of the subnet.

", - "Subnet$VpcId": "

The ID of the VPC the subnet is in.

", - "Subnet$CidrBlock": "

The CIDR block assigned to the subnet.

", - "Subnet$AvailabilityZone": "

The Availability Zone of the subnet.

", - "SubnetIdStringList$member": null, - "Tag$Key": "

The key of the tag.

Constraints: Tag keys are case-sensitive and accept a maximum of 127 Unicode characters. May not begin with aws:

", - "Tag$Value": "

The value of the tag.

Constraints: Tag values are case-sensitive and accept a maximum of 255 Unicode characters.

", - "TagDescription$ResourceId": "

The ID of the resource. For example, ami-1a2b3c4d.

", - "TagDescription$Key": "

The tag key.

", - "TagDescription$Value": "

The tag value.

", - "TargetConfiguration$OfferingId": "

The ID of the Convertible Reserved Instance offering.

", - "TargetConfigurationRequest$OfferingId": "

The Convertible Reserved Instance offering ID. If this isn't included in the request, the response lists your current Convertible Reserved Instance/s and their value/s.

", - "UnassignPrivateIpAddressesRequest$NetworkInterfaceId": "

The ID of the network interface.

", - "UnsuccessfulItem$ResourceId": "

The ID of the resource.

", - "UnsuccessfulItemError$Code": "

The error code.

", - "UnsuccessfulItemError$Message": "

The error message accompanying the error code.

", - "UserBucket$S3Bucket": "

The name of the S3 bucket where the disk image is located.

", - "UserBucket$S3Key": "

The file name of the disk image.

", - "UserBucketDetails$S3Bucket": "

The S3 bucket from which the disk image was created.

", - "UserBucketDetails$S3Key": "

The file name of the disk image.

", - "UserData$Data": "

The user data. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

", - "UserGroupStringList$member": null, - "UserIdGroupPair$UserId": "

The ID of an AWS account. For a referenced security group in another VPC, the account ID of the referenced security group is returned.

[EC2-Classic] Required when adding or removing rules that reference a security group in another AWS account.

", - "UserIdGroupPair$GroupName": "

The name of the security group. In a request, use this parameter for a security group in EC2-Classic or a default VPC only. For a security group in a nondefault VPC, use the security group ID.

", - "UserIdGroupPair$GroupId": "

The ID of the security group.

", - "UserIdGroupPair$VpcId": "

The ID of the VPC for the referenced security group, if applicable.

", - "UserIdGroupPair$VpcPeeringConnectionId": "

The ID of the VPC peering connection, if applicable.

", - "UserIdGroupPair$PeeringStatus": "

The status of a VPC peering connection, if applicable.

", - "UserIdStringList$member": null, - "ValueStringList$member": null, - "VgwTelemetry$OutsideIpAddress": "

The Internet-routable IP address of the virtual private gateway's outside interface.

", - "VgwTelemetry$StatusMessage": "

If an error occurs, a description of the error.

", - "Volume$VolumeId": "

The ID of the volume.

", - "Volume$SnapshotId": "

The snapshot from which the volume was created, if applicable.

", - "Volume$AvailabilityZone": "

The Availability Zone for the volume.

", - "Volume$KmsKeyId": "

The full ARN of the AWS Key Management Service (AWS KMS) customer master key (CMK) that was used to protect the volume encryption key for the volume.

", - "VolumeAttachment$VolumeId": "

The ID of the volume.

", - "VolumeAttachment$InstanceId": "

The ID of the instance.

", - "VolumeAttachment$Device": "

The device name.

", - "VolumeIdStringList$member": null, - "VolumeStatusAction$Code": "

The code identifying the operation, for example, enable-volume-io.

", - "VolumeStatusAction$Description": "

A description of the operation.

", - "VolumeStatusAction$EventType": "

The event type associated with this operation.

", - "VolumeStatusAction$EventId": "

The ID of the event associated with this operation.

", - "VolumeStatusDetails$Status": "

The intended status of the volume status.

", - "VolumeStatusEvent$EventType": "

The type of this event.

", - "VolumeStatusEvent$Description": "

A description of the event.

", - "VolumeStatusEvent$EventId": "

The ID of this event.

", - "VolumeStatusItem$VolumeId": "

The volume ID.

", - "VolumeStatusItem$AvailabilityZone": "

The Availability Zone of the volume.

", - "Vpc$VpcId": "

The ID of the VPC.

", - "Vpc$CidrBlock": "

The CIDR block for the VPC.

", - "Vpc$DhcpOptionsId": "

The ID of the set of DHCP options you've associated with the VPC (or default if the default options are associated with the VPC).

", - "VpcAttachment$VpcId": "

The ID of the VPC.

", - "VpcClassicLink$VpcId": "

The ID of the VPC.

", - "VpcClassicLinkIdList$member": null, - "VpcEndpoint$VpcEndpointId": "

The ID of the VPC endpoint.

", - "VpcEndpoint$VpcId": "

The ID of the VPC to which the endpoint is associated.

", - "VpcEndpoint$ServiceName": "

The name of the AWS service to which the endpoint is associated.

", - "VpcEndpoint$PolicyDocument": "

The policy document associated with the endpoint.

", - "VpcIdStringList$member": null, - "VpcPeeringConnection$VpcPeeringConnectionId": "

The ID of the VPC peering connection.

", - "VpcPeeringConnectionStateReason$Message": "

A message that provides more information about the status, if applicable.

", - "VpcPeeringConnectionVpcInfo$CidrBlock": "

The CIDR block for the VPC.

", - "VpcPeeringConnectionVpcInfo$OwnerId": "

The AWS account ID of the VPC owner.

", - "VpcPeeringConnectionVpcInfo$VpcId": "

The ID of the VPC.

", - "VpnConnection$VpnConnectionId": "

The ID of the VPN connection.

", - "VpnConnection$CustomerGatewayConfiguration": "

The configuration information for the VPN connection's customer gateway (in the native XML format). This element is always present in the CreateVpnConnection response; however, it's present in the DescribeVpnConnections response only if the VPN connection is in the pending or available state.

", - "VpnConnection$CustomerGatewayId": "

The ID of the customer gateway at your end of the VPN connection.

", - "VpnConnection$VpnGatewayId": "

The ID of the virtual private gateway at the AWS side of the VPN connection.

", - "VpnConnectionIdStringList$member": null, - "VpnGateway$VpnGatewayId": "

The ID of the virtual private gateway.

", - "VpnGateway$AvailabilityZone": "

The Availability Zone where the virtual private gateway was created, if applicable. This field may be empty or not returned.

", - "VpnGatewayIdStringList$member": null, - "VpnStaticRoute$DestinationCidrBlock": "

The CIDR block associated with the local subnet of the customer data center.

", - "ZoneNameStringList$member": null - } - }, - "Subnet": { - "base": "

Describes a subnet.

", - "refs": { - "CreateSubnetResult$Subnet": "

Information about the subnet.

", - "SubnetList$member": null - } - }, - "SubnetIdStringList": { - "base": null, - "refs": { - "DescribeSubnetsRequest$SubnetIds": "

One or more subnet IDs.

Default: Describes all your subnets.

" - } - }, - "SubnetList": { - "base": null, - "refs": { - "DescribeSubnetsResult$Subnets": "

Information about one or more subnets.

" - } - }, - "SubnetState": { - "base": null, - "refs": { - "Subnet$State": "

The current state of the subnet.

" - } - }, - "SummaryStatus": { - "base": null, - "refs": { - "InstanceStatusSummary$Status": "

The status.

" - } - }, - "Tag": { - "base": "

Describes a tag.

", - "refs": { - "TagList$member": null - } - }, - "TagDescription": { - "base": "

Describes a tag.

", - "refs": { - "TagDescriptionList$member": null - } - }, - "TagDescriptionList": { - "base": null, - "refs": { - "DescribeTagsResult$Tags": "

A list of tags.

" - } - }, - "TagList": { - "base": null, - "refs": { - "ClassicLinkInstance$Tags": "

Any tags assigned to the instance.

", - "ConversionTask$Tags": "

Any tags assigned to the task.

", - "CreateTagsRequest$Tags": "

One or more tags. The value parameter is required, but if you don't want the tag to have a value, specify the parameter with no value, and we set the value to an empty string.

", - "CustomerGateway$Tags": "

Any tags assigned to the customer gateway.

", - "DeleteTagsRequest$Tags": "

One or more tags to delete. If you omit the value parameter, we delete the tag regardless of its value. If you specify this parameter with an empty string as the value, we delete the key only if its value is an empty string.

", - "DhcpOptions$Tags": "

Any tags assigned to the DHCP options set.

", - "Image$Tags": "

Any tags assigned to the image.

", - "Instance$Tags": "

Any tags assigned to the instance.

", - "InternetGateway$Tags": "

Any tags assigned to the Internet gateway.

", - "NetworkAcl$Tags": "

Any tags assigned to the network ACL.

", - "NetworkInterface$TagSet": "

Any tags assigned to the network interface.

", - "ReservedInstances$Tags": "

Any tags assigned to the resource.

", - "ReservedInstancesListing$Tags": "

Any tags assigned to the resource.

", - "RouteTable$Tags": "

Any tags assigned to the route table.

", - "SecurityGroup$Tags": "

Any tags assigned to the security group.

", - "Snapshot$Tags": "

Any tags assigned to the snapshot.

", - "SpotInstanceRequest$Tags": "

Any tags assigned to the resource.

", - "Subnet$Tags": "

Any tags assigned to the subnet.

", - "Volume$Tags": "

Any tags assigned to the volume.

", - "Vpc$Tags": "

Any tags assigned to the VPC.

", - "VpcClassicLink$Tags": "

Any tags assigned to the VPC.

", - "VpcPeeringConnection$Tags": "

Any tags assigned to the resource.

", - "VpnConnection$Tags": "

Any tags assigned to the VPN connection.

", - "VpnGateway$Tags": "

Any tags assigned to the virtual private gateway.

" - } - }, - "TargetConfiguration": { - "base": "

Information about the Convertible Reserved Instance offering.

", - "refs": { - "TargetReservationValue$TargetConfiguration": "

The configuration of the Convertible Reserved Instances that make up the exchange.

" - } - }, - "TargetConfigurationRequest": { - "base": "

Details about the target configuration.

", - "refs": { - "TargetConfigurationRequestSet$member": null - } - }, - "TargetConfigurationRequestSet": { - "base": null, - "refs": { - "AcceptReservedInstancesExchangeQuoteRequest$TargetConfigurations": "

The configurations of the Convertible Reserved Instance offerings you are purchasing in this exchange.

", - "GetReservedInstancesExchangeQuoteRequest$TargetConfigurations": "

The configuration requirements of the Convertible Reserved Instances you want in exchange for your current Convertible Reserved Instances.

" - } - }, - "TargetReservationValue": { - "base": "

The total value of the new Convertible Reserved Instances.

", - "refs": { - "TargetReservationValueSet$member": null - } - }, - "TargetReservationValueSet": { - "base": null, - "refs": { - "GetReservedInstancesExchangeQuoteResult$TargetConfigurationValueSet": "

The values of the target Convertible Reserved Instances.

" - } - }, - "TelemetryStatus": { - "base": null, - "refs": { - "VgwTelemetry$Status": "

The status of the VPN tunnel.

" - } - }, - "Tenancy": { - "base": null, - "refs": { - "CreateVpcRequest$InstanceTenancy": "

The tenancy options for instances launched into the VPC. For default, instances are launched with shared tenancy by default. You can launch instances with any tenancy into a shared tenancy VPC. For dedicated, instances are launched as dedicated tenancy instances by default. You can only launch instances with a tenancy of dedicated or host into a dedicated tenancy VPC.

Important: The host value cannot be used with this parameter. Use the default or dedicated values only.

Default: default

", - "DescribeReservedInstancesOfferingsRequest$InstanceTenancy": "

The tenancy of the instances covered by the reservation. A Reserved Instance with a tenancy of dedicated is applied to instances that run in a VPC on single-tenant hardware (i.e., Dedicated Instances).

Default: default

", - "Placement$Tenancy": "

The tenancy of the instance (if the instance is running in a VPC). An instance with a tenancy of dedicated runs on single-tenant hardware. The host tenancy is not supported for the ImportInstance command.

", - "ReservedInstances$InstanceTenancy": "

The tenancy of the instance.

", - "ReservedInstancesOffering$InstanceTenancy": "

The tenancy of the instance.

", - "Vpc$InstanceTenancy": "

The allowed tenancy of instances launched into the VPC.

" - } - }, - "TerminateInstancesRequest": { - "base": "

Contains the parameters for TerminateInstances.

", - "refs": { - } - }, - "TerminateInstancesResult": { - "base": "

Contains the output of TerminateInstances.

", - "refs": { - } - }, - "TrafficType": { - "base": null, - "refs": { - "CreateFlowLogsRequest$TrafficType": "

The type of traffic to log.

", - "FlowLog$TrafficType": "

The type of traffic captured for the flow log.

" - } - }, - "UnassignPrivateIpAddressesRequest": { - "base": "

Contains the parameters for UnassignPrivateIpAddresses.

", - "refs": { - } - }, - "UnmonitorInstancesRequest": { - "base": "

Contains the parameters for UnmonitorInstances.

", - "refs": { - } - }, - "UnmonitorInstancesResult": { - "base": "

Contains the output of UnmonitorInstances.

", - "refs": { - } - }, - "UnsuccessfulItem": { - "base": "

Information about items that were not successfully processed in a batch call.

", - "refs": { - "UnsuccessfulItemList$member": null, - "UnsuccessfulItemSet$member": null - } - }, - "UnsuccessfulItemError": { - "base": "

Information about the error that occurred. For more information about errors, see Error Codes.

", - "refs": { - "UnsuccessfulItem$Error": "

Information about the error.

" - } - }, - "UnsuccessfulItemList": { - "base": null, - "refs": { - "ModifyHostsResult$Unsuccessful": "

The IDs of the Dedicated Hosts that could not be modified. Check whether the setting you requested can be used.

", - "ReleaseHostsResult$Unsuccessful": "

The IDs of the Dedicated Hosts that could not be released, including an error message.

" - } - }, - "UnsuccessfulItemSet": { - "base": null, - "refs": { - "CreateFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be created successfully.

", - "DeleteFlowLogsResult$Unsuccessful": "

Information about the flow logs that could not be deleted successfully.

", - "DeleteVpcEndpointsResult$Unsuccessful": "

Information about the endpoints that were not successfully deleted.

" - } - }, - "UserBucket": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "ImageDiskContainer$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotDiskContainer$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserBucketDetails": { - "base": "

Describes the S3 bucket for the disk image.

", - "refs": { - "SnapshotDetail$UserBucket": "

The S3 bucket for the disk image.

", - "SnapshotTaskDetail$UserBucket": "

The S3 bucket for the disk image.

" - } - }, - "UserData": { - "base": "

Describes the user data for an instance.

", - "refs": { - "ImportInstanceLaunchSpecification$UserData": "

The user data to make available to the instance. If you are using an AWS SDK or command line tool, Base64-encoding is performed for you, and you can load the text from a file. Otherwise, you must provide Base64-encoded text.

" - } - }, - "UserGroupStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserGroups": "

One or more user groups. This is only valid when modifying the launchPermission attribute.

" - } - }, - "UserIdGroupPair": { - "base": "

Describes a security group and AWS account ID pair.

", - "refs": { - "UserIdGroupPairList$member": null, - "UserIdGroupPairSet$member": null - } - }, - "UserIdGroupPairList": { - "base": null, - "refs": { - "IpPermission$UserIdGroupPairs": "

One or more security group and AWS account ID pairs.

" - } - }, - "UserIdGroupPairSet": { - "base": null, - "refs": { - "StaleIpPermission$UserIdGroupPairs": "

One or more security group pairs. Returns the ID of the referenced security group and VPC, and the ID and status of the VPC peering connection.

" - } - }, - "UserIdStringList": { - "base": null, - "refs": { - "ModifyImageAttributeRequest$UserIds": "

One or more AWS account IDs. This is only valid when modifying the launchPermission attribute.

", - "ModifySnapshotAttributeRequest$UserIds": "

The account ID to modify for the snapshot.

" - } - }, - "ValueStringList": { - "base": null, - "refs": { - "CancelSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "CreateFlowLogsRequest$ResourceIds": "

One or more subnet, network interface, or VPC IDs.

Constraints: Maximum of 1000 resources

", - "CreateFlowLogsResult$FlowLogIds": "

The IDs of the flow logs.

", - "CreateVpcEndpointRequest$RouteTableIds": "

One or more route table IDs.

", - "DeleteFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DeleteVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeFlowLogsRequest$FlowLogIds": "

One or more flow log IDs.

", - "DescribeInternetGatewaysRequest$InternetGatewayIds": "

One or more Internet gateway IDs.

Default: Describes all your Internet gateways.

", - "DescribeMovingAddressesRequest$PublicIps": "

One or more Elastic IP addresses.

", - "DescribeNatGatewaysRequest$NatGatewayIds": "

One or more NAT gateway IDs.

", - "DescribeNetworkAclsRequest$NetworkAclIds": "

One or more network ACL IDs.

Default: Describes all your network ACLs.

", - "DescribePrefixListsRequest$PrefixListIds": "

One or more prefix list IDs.

", - "DescribeRouteTablesRequest$RouteTableIds": "

One or more route table IDs.

Default: Describes all your route tables.

", - "DescribeSpotFleetRequestsRequest$SpotFleetRequestIds": "

The IDs of the Spot fleet requests.

", - "DescribeVpcEndpointServicesResult$ServiceNames": "

A list of supported AWS services.

", - "DescribeVpcEndpointsRequest$VpcEndpointIds": "

One or more endpoint IDs.

", - "DescribeVpcPeeringConnectionsRequest$VpcPeeringConnectionIds": "

One or more VPC peering connection IDs.

Default: Describes all your VPC peering connections.

", - "Filter$Values": "

One or more filter values. Filter values are case-sensitive.

", - "ModifyVpcEndpointRequest$AddRouteTableIds": "

One or more route tables IDs to associate with the endpoint.

", - "ModifyVpcEndpointRequest$RemoveRouteTableIds": "

One or more route table IDs to disassociate from the endpoint.

", - "NewDhcpConfiguration$Values": null, - "PrefixList$Cidrs": "

The IP address range of the AWS service.

", - "RequestSpotLaunchSpecification$SecurityGroups": null, - "RequestSpotLaunchSpecification$SecurityGroupIds": null, - "VpcEndpoint$RouteTableIds": "

One or more route tables associated with the endpoint.

" - } - }, - "VgwTelemetry": { - "base": "

Describes telemetry for a VPN tunnel.

", - "refs": { - "VgwTelemetryList$member": null - } - }, - "VgwTelemetryList": { - "base": null, - "refs": { - "VpnConnection$VgwTelemetry": "

Information about the VPN tunnel.

" - } - }, - "VirtualizationType": { - "base": null, - "refs": { - "Image$VirtualizationType": "

The type of virtualization of the AMI.

", - "Instance$VirtualizationType": "

The virtualization type of the instance.

" - } - }, - "Volume": { - "base": "

Describes a volume.

", - "refs": { - "VolumeList$member": null - } - }, - "VolumeAttachment": { - "base": "

Describes volume attachment details.

", - "refs": { - "VolumeAttachmentList$member": null - } - }, - "VolumeAttachmentList": { - "base": null, - "refs": { - "Volume$Attachments": "

Information about the volume attachments.

" - } - }, - "VolumeAttachmentState": { - "base": null, - "refs": { - "VolumeAttachment$State": "

The attachment state of the volume.

" - } - }, - "VolumeAttributeName": { - "base": null, - "refs": { - "DescribeVolumeAttributeRequest$Attribute": "

The instance attribute.

" - } - }, - "VolumeDetail": { - "base": "

Describes an EBS volume.

", - "refs": { - "DiskImage$Volume": "

Information about the volume.

", - "ImportVolumeRequest$Volume": "

The volume size.

" - } - }, - "VolumeIdStringList": { - "base": null, - "refs": { - "DescribeVolumeStatusRequest$VolumeIds": "

One or more volume IDs.

Default: Describes all your volumes.

", - "DescribeVolumesRequest$VolumeIds": "

One or more volume IDs.

" - } - }, - "VolumeList": { - "base": null, - "refs": { - "DescribeVolumesResult$Volumes": "

Information about the volumes.

" - } - }, - "VolumeState": { - "base": null, - "refs": { - "Volume$State": "

The volume state.

" - } - }, - "VolumeStatusAction": { - "base": "

Describes a volume status operation code.

", - "refs": { - "VolumeStatusActionsList$member": null - } - }, - "VolumeStatusActionsList": { - "base": null, - "refs": { - "VolumeStatusItem$Actions": "

The details of the operation.

" - } - }, - "VolumeStatusDetails": { - "base": "

Describes a volume status.

", - "refs": { - "VolumeStatusDetailsList$member": null - } - }, - "VolumeStatusDetailsList": { - "base": null, - "refs": { - "VolumeStatusInfo$Details": "

The details of the volume status.

" - } - }, - "VolumeStatusEvent": { - "base": "

Describes a volume status event.

", - "refs": { - "VolumeStatusEventsList$member": null - } - }, - "VolumeStatusEventsList": { - "base": null, - "refs": { - "VolumeStatusItem$Events": "

A list of events associated with the volume.

" - } - }, - "VolumeStatusInfo": { - "base": "

Describes the status of a volume.

", - "refs": { - "VolumeStatusItem$VolumeStatus": "

The volume status.

" - } - }, - "VolumeStatusInfoStatus": { - "base": null, - "refs": { - "VolumeStatusInfo$Status": "

The status of the volume.

" - } - }, - "VolumeStatusItem": { - "base": "

Describes the volume status.

", - "refs": { - "VolumeStatusList$member": null - } - }, - "VolumeStatusList": { - "base": null, - "refs": { - "DescribeVolumeStatusResult$VolumeStatuses": "

A list of volumes.

" - } - }, - "VolumeStatusName": { - "base": null, - "refs": { - "VolumeStatusDetails$Name": "

The name of the volume status.

" - } - }, - "VolumeType": { - "base": null, - "refs": { - "CreateVolumeRequest$VolumeType": "

The volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

Default: standard

", - "EbsBlockDevice$VolumeType": "

The volume type: gp2, io1, st1, sc1, or standard.

Default: standard

", - "Volume$VolumeType": "

The volume type. This can be gp2 for General Purpose SSD, io1 for Provisioned IOPS SSD, st1 for Throughput Optimized HDD, sc1 for Cold HDD, or standard for Magnetic volumes.

" - } - }, - "Vpc": { - "base": "

Describes a VPC.

", - "refs": { - "CreateVpcResult$Vpc": "

Information about the VPC.

", - "VpcList$member": null - } - }, - "VpcAttachment": { - "base": "

Describes an attachment between a virtual private gateway and a VPC.

", - "refs": { - "AttachVpnGatewayResult$VpcAttachment": "

Information about the attachment.

", - "VpcAttachmentList$member": null - } - }, - "VpcAttachmentList": { - "base": null, - "refs": { - "VpnGateway$VpcAttachments": "

Any VPCs attached to the virtual private gateway.

" - } - }, - "VpcAttributeName": { - "base": null, - "refs": { - "DescribeVpcAttributeRequest$Attribute": "

The VPC attribute.

" - } - }, - "VpcClassicLink": { - "base": "

Describes whether a VPC is enabled for ClassicLink.

", - "refs": { - "VpcClassicLinkList$member": null - } - }, - "VpcClassicLinkIdList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkDnsSupportRequest$VpcIds": "

One or more VPC IDs.

", - "DescribeVpcClassicLinkRequest$VpcIds": "

One or more VPCs for which you want to describe the ClassicLink status.

" - } - }, - "VpcClassicLinkList": { - "base": null, - "refs": { - "DescribeVpcClassicLinkResult$Vpcs": "

The ClassicLink status of one or more VPCs.

" - } - }, - "VpcEndpoint": { - "base": "

Describes a VPC endpoint.

", - "refs": { - "CreateVpcEndpointResult$VpcEndpoint": "

Information about the endpoint.

", - "VpcEndpointSet$member": null - } - }, - "VpcEndpointSet": { - "base": null, - "refs": { - "DescribeVpcEndpointsResult$VpcEndpoints": "

Information about the endpoints.

" - } - }, - "VpcIdStringList": { - "base": null, - "refs": { - "DescribeVpcsRequest$VpcIds": "

One or more VPC IDs.

Default: Describes all your VPCs.

" - } - }, - "VpcList": { - "base": null, - "refs": { - "DescribeVpcsResult$Vpcs": "

Information about one or more VPCs.

" - } - }, - "VpcPeeringConnection": { - "base": "

Describes a VPC peering connection.

", - "refs": { - "AcceptVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "CreateVpcPeeringConnectionResult$VpcPeeringConnection": "

Information about the VPC peering connection.

", - "VpcPeeringConnectionList$member": null - } - }, - "VpcPeeringConnectionList": { - "base": null, - "refs": { - "DescribeVpcPeeringConnectionsResult$VpcPeeringConnections": "

Information about the VPC peering connections.

" - } - }, - "VpcPeeringConnectionOptionsDescription": { - "base": "

Describes the VPC peering connection options.

", - "refs": { - "VpcPeeringConnectionVpcInfo$PeeringOptions": "

Information about the VPC peering connection options for the accepter or requester VPC.

" - } - }, - "VpcPeeringConnectionStateReason": { - "base": "

Describes the status of a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$Status": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionStateReasonCode": { - "base": null, - "refs": { - "VpcPeeringConnectionStateReason$Code": "

The status of the VPC peering connection.

" - } - }, - "VpcPeeringConnectionVpcInfo": { - "base": "

Describes a VPC in a VPC peering connection.

", - "refs": { - "VpcPeeringConnection$AccepterVpcInfo": "

Information about the accepter VPC. CIDR block information is not returned when creating a VPC peering connection, or when describing a VPC peering connection that's in the initiating-request or pending-acceptance state.

", - "VpcPeeringConnection$RequesterVpcInfo": "

Information about the requester VPC.

" - } - }, - "VpcState": { - "base": null, - "refs": { - "Vpc$State": "

The current state of the VPC.

" - } - }, - "VpnConnection": { - "base": "

Describes a VPN connection.

", - "refs": { - "CreateVpnConnectionResult$VpnConnection": "

Information about the VPN connection.

", - "VpnConnectionList$member": null - } - }, - "VpnConnectionIdStringList": { - "base": null, - "refs": { - "DescribeVpnConnectionsRequest$VpnConnectionIds": "

One or more VPN connection IDs.

Default: Describes your VPN connections.

" - } - }, - "VpnConnectionList": { - "base": null, - "refs": { - "DescribeVpnConnectionsResult$VpnConnections": "

Information about one or more VPN connections.

" - } - }, - "VpnConnectionOptions": { - "base": "

Describes VPN connection options.

", - "refs": { - "VpnConnection$Options": "

The VPN connection options.

" - } - }, - "VpnConnectionOptionsSpecification": { - "base": "

Describes VPN connection options.

", - "refs": { - "CreateVpnConnectionRequest$Options": "

Indicates whether the VPN connection requires static routes. If you are creating a VPN connection for a device that does not support BGP, you must specify true.

Default: false

" - } - }, - "VpnGateway": { - "base": "

Describes a virtual private gateway.

", - "refs": { - "CreateVpnGatewayResult$VpnGateway": "

Information about the virtual private gateway.

", - "VpnGatewayList$member": null - } - }, - "VpnGatewayIdStringList": { - "base": null, - "refs": { - "DescribeVpnGatewaysRequest$VpnGatewayIds": "

One or more virtual private gateway IDs.

Default: Describes all your virtual private gateways.

" - } - }, - "VpnGatewayList": { - "base": null, - "refs": { - "DescribeVpnGatewaysResult$VpnGateways": "

Information about one or more virtual private gateways.

" - } - }, - "VpnState": { - "base": null, - "refs": { - "VpnConnection$State": "

The current state of the VPN connection.

", - "VpnGateway$State": "

The current state of the virtual private gateway.

", - "VpnStaticRoute$State": "

The current state of the static route.

" - } - }, - "VpnStaticRoute": { - "base": "

Describes a static route for a VPN connection.

", - "refs": { - "VpnStaticRouteList$member": null - } - }, - "VpnStaticRouteList": { - "base": null, - "refs": { - "VpnConnection$Routes": "

The static routes associated with the VPN connection.

" - } - }, - "VpnStaticRouteSource": { - "base": null, - "refs": { - "VpnStaticRoute$Source": "

Indicates how the routes were provided.

" - } - }, - "ZoneNameStringList": { - "base": null, - "refs": { - "DescribeAvailabilityZonesRequest$ZoneNames": "

The names of one or more Availability Zones.

" - } - }, - "scope": { - "base": null, - "refs": { - "ReservedInstances$Scope": "

The scope of the Reserved Instance.

", - "ReservedInstancesConfiguration$Scope": "

Whether the Reserved Instance is standard or convertible.

", - "ReservedInstancesOffering$Scope": "

Whether the Reserved Instance is applied to instances in a region or an Availability Zone.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/examples-1.json deleted file mode 100755 index f6a8719f2..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/examples-1.json +++ /dev/null @@ -1,3740 +0,0 @@ -{ - "version": "1.0", - "examples": { - "AllocateAddress": [ - { - "input": { - "Domain": "vpc" - }, - "output": { - "AllocationId": "eipalloc-64d5890a", - "Domain": "vpc", - "PublicIp": "203.0.113.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example allocates an Elastic IP address to use with an instance in a VPC.", - "id": "ec2-allocate-address-1", - "title": "To allocate an Elastic IP address for EC2-VPC" - }, - { - "output": { - "Domain": "standard", - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example allocates an Elastic IP address to use with an instance in EC2-Classic.", - "id": "ec2-allocate-address-2", - "title": "To allocate an Elastic IP address for EC2-Classic" - } - ], - "AssignPrivateIpAddresses": [ - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3", - "PrivateIpAddresses": [ - "10.0.0.82" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example assigns the specified secondary private IP address to the specified network interface.", - "id": "ec2-assign-private-ip-addresses-1", - "title": "To assign a specific secondary private IP address to an interface" - }, - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3", - "SecondaryPrivateIpAddressCount": 2 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example assigns two secondary private IP addresses to the specified network interface. Amazon EC2 automatically assigns these IP addresses from the available IP addresses in the CIDR block range of the subnet the network interface is associated with.", - "id": "ec2-assign-private-ip-addresses-2", - "title": "To assign secondary private IP addresses that Amazon EC2 selects to an interface" - } - ], - "AssociateAddress": [ - { - "input": { - "AllocationId": "eipalloc-64d5890a", - "InstanceId": "i-0b263919b6498b123" - }, - "output": { - "AssociationId": "eipassoc-2bebb745" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified Elastic IP address with the specified instance in a VPC.", - "id": "ec2-associate-address-1", - "title": "To associate an Elastic IP address in EC2-VPC" - }, - { - "input": { - "AllocationId": "eipalloc-64d5890a", - "NetworkInterfaceId": "eni-1a2b3c4d" - }, - "output": { - "AssociationId": "eipassoc-2bebb745" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified Elastic IP address with the specified network interface.", - "id": "ec2-associate-address-2", - "title": "To associate an Elastic IP address with a network interface" - }, - { - "input": { - "InstanceId": "i-07ffe74c7330ebf53", - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates an Elastic IP address with an instance in EC2-Classic.", - "id": "ec2-associate-address-3", - "title": "To associate an Elastic IP address in EC2-Classic" - } - ], - "AssociateDhcpOptions": [ - { - "input": { - "DhcpOptionsId": "dopt-d9070ebb", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified DHCP options set with the specified VPC.", - "id": "ec2-associate-dhcp-options-1", - "title": "To associate a DHCP options set with a VPC" - }, - { - "input": { - "DhcpOptionsId": "default", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the default DHCP options set with the specified VPC.", - "id": "ec2-associate-dhcp-options-2", - "title": "To associate the default DHCP options set with a VPC" - } - ], - "AssociateRouteTable": [ - { - "input": { - "RouteTableId": "rtb-22574640", - "SubnetId": "subnet-9d4a7b6" - }, - "output": { - "AssociationId": "rtbassoc-781d0d1a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified route table with the specified subnet.", - "id": "ec2-associate-route-table-1", - "title": "To associate a route table with a subnet" - } - ], - "AttachInternetGateway": [ - { - "input": { - "InternetGatewayId": "igw-c0a643a9", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example attaches the specified Internet gateway to the specified VPC.", - "id": "ec2-attach-internet-gateway-1", - "title": "To attach an Internet gateway to a VPC" - } - ], - "AttachNetworkInterface": [ - { - "input": { - "DeviceIndex": 1, - "InstanceId": "i-1234567890abcdef0", - "NetworkInterfaceId": "eni-e5aa89a3" - }, - "output": { - "AttachmentId": "eni-attach-66c4350a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example attaches the specified network interface to the specified instance.", - "id": "ec2-attach-network-interface-1", - "title": "To attach a network interface to an instance" - } - ], - "AttachVolume": [ - { - "input": { - "Device": "/dev/sdf", - "InstanceId": "i-01474ef662b89480", - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "AttachTime": "2016-08-29T18:52:32.724Z", - "Device": "/dev/sdf", - "InstanceId": "i-01474ef662b89480", - "State": "attaching", - "VolumeId": "vol-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example attaches a volume (``vol-1234567890abcdef0``) to an instance (``i-01474ef662b89480``) as ``/dev/sdf``.", - "id": "to-attach-a-volume-to-an-instance-1472499213109", - "title": "To attach a volume to an instance" - } - ], - "CancelSpotFleetRequests": [ - { - "input": { - "SpotFleetRequestIds": [ - "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - ], - "TerminateInstances": true - }, - "output": { - "SuccessfulFleetRequests": [ - { - "CurrentSpotFleetRequestState": "cancelled_running", - "PreviousSpotFleetRequestState": "active", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example cancels the specified Spot fleet request and terminates its associated Spot Instances.", - "id": "ec2-cancel-spot-fleet-requests-1", - "title": "To cancel a Spot fleet request" - }, - { - "input": { - "SpotFleetRequestIds": [ - "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - ], - "TerminateInstances": false - }, - "output": { - "SuccessfulFleetRequests": [ - { - "CurrentSpotFleetRequestState": "cancelled_terminating", - "PreviousSpotFleetRequestState": "active", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example cancels the specified Spot fleet request without terminating its associated Spot Instances.", - "id": "ec2-cancel-spot-fleet-requests-2", - "title": "To cancel a Spot fleet request without terminating its Spot Instances" - } - ], - "CancelSpotInstanceRequests": [ - { - "input": { - "SpotInstanceRequestIds": [ - "sir-08b93456" - ] - }, - "output": { - "CancelledSpotInstanceRequests": [ - { - "SpotInstanceRequestId": "sir-08b93456", - "State": "cancelled" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example cancels a Spot Instance request.", - "id": "ec2-cancel-spot-instance-requests-1", - "title": "To cancel Spot Instance requests" - } - ], - "ConfirmProductInstance": [ - { - "input": { - "InstanceId": "i-1234567890abcdef0", - "ProductCode": "774F4FF8" - }, - "output": { - "OwnerId": "123456789012" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example determines whether the specified product code is associated with the specified instance.", - "id": "to-confirm-the-product-instance-1472712108494", - "title": "To confirm the product instance" - } - ], - "CopySnapshot": [ - { - "input": { - "Description": "This is my copied snapshot.", - "DestinationRegion": "us-east-1", - "SourceRegion": "us-west-2", - "SourceSnapshotId": "snap-066877671789bd71b" - }, - "output": { - "SnapshotId": "snap-066877671789bd71b" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example copies a snapshot with the snapshot ID of ``snap-066877671789bd71b`` from the ``us-west-2`` region to the ``us-east-1`` region and adds a short description to identify the snapshot.", - "id": "to-copy-a-snapshot-1472502259774", - "title": "To copy a snapshot" - } - ], - "CreateCustomerGateway": [ - { - "input": { - "BgpAsn": 65534, - "PublicIp": "12.1.2.3", - "Type": "ipsec.1" - }, - "output": { - "CustomerGateway": { - "BgpAsn": "65534", - "CustomerGatewayId": "cgw-0e11f167", - "IpAddress": "12.1.2.3", - "State": "available", - "Type": "ipsec.1" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a customer gateway with the specified IP address for its outside interface.", - "id": "ec2-create-customer-gateway-1", - "title": "To create a customer gateway" - } - ], - "CreateDhcpOptions": [ - { - "input": { - "DhcpConfigurations": [ - { - "Key": "domain-name-servers", - "Values": [ - "10.2.5.1", - "10.2.5.2" - ] - } - ] - }, - "output": { - "DhcpOptions": { - "DhcpConfigurations": [ - { - "Key": "domain-name-servers", - "Values": [ - { - "Value": "10.2.5.2" - }, - { - "Value": "10.2.5.1" - } - ] - } - ], - "DhcpOptionsId": "dopt-d9070ebb" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a DHCP options set.", - "id": "ec2-create-dhcp-options-1", - "title": "To create a DHCP options set" - } - ], - "CreateInternetGateway": [ - { - "output": { - "InternetGateway": { - "Attachments": [ - - ], - "InternetGatewayId": "igw-c0a643a9", - "Tags": [ - - ] - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates an Internet gateway.", - "id": "ec2-create-internet-gateway-1", - "title": "To create an Internet gateway" - } - ], - "CreateKeyPair": [ - { - "input": { - "KeyName": "my-key-pair" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a key pair named my-key-pair.", - "id": "ec2-create-key-pair-1", - "title": "To create a key pair" - } - ], - "CreateNatGateway": [ - { - "input": { - "AllocationId": "eipalloc-37fc1a52", - "SubnetId": "subnet-1a2b3c4d" - }, - "output": { - "NatGateway": { - "CreateTime": "2015-12-17T12:45:26.732Z", - "NatGatewayAddresses": [ - { - "AllocationId": "eipalloc-37fc1a52" - } - ], - "NatGatewayId": "nat-08d48af2a8e83edfd", - "State": "pending", - "SubnetId": "subnet-1a2b3c4d", - "VpcId": "vpc-1122aabb" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a NAT gateway in subnet subnet-1a2b3c4d and associates an Elastic IP address with the allocation ID eipalloc-37fc1a52 with the NAT gateway.", - "id": "ec2-create-nat-gateway-1", - "title": "To create a NAT gateway" - } - ], - "CreateNetworkAcl": [ - { - "input": { - "VpcId": "vpc-a01106c2" - }, - "output": { - "NetworkAcl": { - "Associations": [ - - ], - "Entries": [ - { - "CidrBlock": "0.0.0.0/0", - "Egress": true, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - }, - { - "CidrBlock": "0.0.0.0/0", - "Egress": false, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - } - ], - "IsDefault": false, - "NetworkAclId": "acl-5fb85d36", - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a network ACL for the specified VPC.", - "id": "ec2-create-network-acl-1", - "title": "To create a network ACL" - } - ], - "CreateNetworkAclEntry": [ - { - "input": { - "CidrBlock": "0.0.0.0/0", - "Egress": false, - "NetworkAclId": "acl-5fb85d36", - "PortRange": { - "From": 53, - "To": 53 - }, - "Protocol": "udp", - "RuleAction": "allow", - "RuleNumber": 100 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates an entry for the specified network ACL. The rule allows ingress traffic from anywhere (0.0.0.0/0) on UDP port 53 (DNS) into any associated subnet.", - "id": "ec2-create-network-acl-entry-1", - "title": "To create a network ACL entry" - } - ], - "CreateNetworkInterface": [ - { - "input": { - "Description": "my network interface", - "Groups": [ - "sg-903004f8" - ], - "PrivateIpAddress": "10.0.2.17", - "SubnetId": "subnet-9d4a7b6c" - }, - "output": { - "NetworkInterface": { - "AvailabilityZone": "us-east-1d", - "Description": "my network interface", - "Groups": [ - { - "GroupId": "sg-903004f8", - "GroupName": "default" - } - ], - "MacAddress": "02:1a:80:41:52:9c", - "NetworkInterfaceId": "eni-e5aa89a3", - "OwnerId": "123456789012", - "PrivateIpAddress": "10.0.2.17", - "PrivateIpAddresses": [ - { - "Primary": true, - "PrivateIpAddress": "10.0.2.17" - } - ], - "RequesterManaged": false, - "SourceDestCheck": true, - "Status": "pending", - "SubnetId": "subnet-9d4a7b6c", - "TagSet": [ - - ], - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a network interface for the specified subnet.", - "id": "ec2-create-network-interface-1", - "title": "To create a network interface" - } - ], - "CreatePlacementGroup": [ - { - "input": { - "GroupName": "my-cluster", - "Strategy": "cluster" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a placement group with the specified name.", - "id": "to-create-a-placement-group-1472712245768", - "title": "To create a placement group" - } - ], - "CreateRoute": [ - { - "input": { - "DestinationCidrBlock": "0.0.0.0/0", - "GatewayId": "igw-c0a643a9", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a route for the specified route table. The route matches all traffic (0.0.0.0/0) and routes it to the specified Internet gateway.", - "id": "ec2-create-route-1", - "title": "To create a route" - } - ], - "CreateRouteTable": [ - { - "input": { - "VpcId": "vpc-a01106c2" - }, - "output": { - "RouteTable": { - "Associations": [ - - ], - "PropagatingVgws": [ - - ], - "RouteTableId": "rtb-22574640", - "Routes": [ - { - "DestinationCidrBlock": "10.0.0.0/16", - "GatewayId": "local", - "State": "active" - } - ], - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a route table for the specified VPC.", - "id": "ec2-create-route-table-1", - "title": "To create a route table" - } - ], - "CreateSnapshot": [ - { - "input": { - "Description": "This is my root volume snapshot.", - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "Description": "This is my root volume snapshot.", - "OwnerId": "012345678910", - "SnapshotId": "snap-066877671789bd71b", - "StartTime": "2014-02-28T21:06:01.000Z", - "State": "pending", - "Tags": [ - - ], - "VolumeId": "vol-1234567890abcdef0", - "VolumeSize": 8 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a snapshot of the volume with a volume ID of ``vol-1234567890abcdef0`` and a short description to identify the snapshot.", - "id": "to-create-a-snapshot-1472502529790", - "title": "To create a snapshot" - } - ], - "CreateSpotDatafeedSubscription": [ - { - "input": { - "Bucket": "my-s3-bucket", - "Prefix": "spotdata" - }, - "output": { - "SpotDatafeedSubscription": { - "Bucket": "my-s3-bucket", - "OwnerId": "123456789012", - "Prefix": "spotdata", - "State": "Active" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot Instance data feed for your AWS account.", - "id": "ec2-create-spot-datafeed-subscription-1", - "title": "To create a Spot Instance datafeed" - } - ], - "CreateSubnet": [ - { - "input": { - "CidrBlock": "10.0.1.0/24", - "VpcId": "vpc-a01106c2" - }, - "output": { - "Subnet": { - "AvailabilityZone": "us-west-2c", - "AvailableIpAddressCount": 251, - "CidrBlock": "10.0.1.0/24", - "State": "pending", - "SubnetId": "subnet-9d4a7b6c", - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a subnet in the specified VPC with the specified CIDR block. We recommend that you let us select an Availability Zone for you.", - "id": "ec2-create-subnet-1", - "title": "To create a subnet" - } - ], - "CreateTags": [ - { - "input": { - "Resources": [ - "ami-78a54011" - ], - "Tags": [ - { - "Key": "Stack", - "Value": "production" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example adds the tag Stack=production to the specified image, or overwrites an existing tag for the AMI where the tag key is Stack.", - "id": "ec2-create-tags-1", - "title": "To add a tag to a resource" - } - ], - "CreateVolume": [ - { - "input": { - "AvailabilityZone": "us-east-1a", - "Size": 80, - "VolumeType": "gp2" - }, - "output": { - "AvailabilityZone": "us-east-1a", - "CreateTime": "2016-08-29T18:52:32.724Z", - "Encrypted": false, - "Iops": 240, - "Size": 80, - "SnapshotId": "", - "State": "creating", - "VolumeId": "vol-6b60b7c7", - "VolumeType": "gp2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates an 80 GiB General Purpose (SSD) volume in the Availability Zone ``us-east-1a``.", - "id": "to-create-a-new-volume-1472496724296", - "title": "To create a new volume" - }, - { - "input": { - "AvailabilityZone": "us-east-1a", - "Iops": 1000, - "SnapshotId": "snap-066877671789bd71b", - "VolumeType": "io1" - }, - "output": { - "Attachments": [ - - ], - "AvailabilityZone": "us-east-1a", - "CreateTime": "2016-08-29T18:52:32.724Z", - "Iops": 1000, - "Size": 500, - "SnapshotId": "snap-066877671789bd71b", - "State": "creating", - "Tags": [ - - ], - "VolumeId": "vol-1234567890abcdef0", - "VolumeType": "io1" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a new Provisioned IOPS (SSD) volume with 1000 provisioned IOPS from a snapshot in the Availability Zone ``us-east-1a``.", - "id": "to-create-a-new-provisioned-iops-ssd-volume-from-a-snapshot-1472498975176", - "title": "To create a new Provisioned IOPS (SSD) volume from a snapshot" - } - ], - "CreateVpc": [ - { - "input": { - "CidrBlock": "10.0.0.0/16" - }, - "output": { - "Vpc": { - "CidrBlock": "10.0.0.0/16", - "DhcpOptionsId": "dopt-7a8b9c2d", - "InstanceTenancy": "default", - "State": "pending", - "VpcId": "vpc-a01106c2" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a VPC with the specified CIDR block.", - "id": "ec2-create-vpc-1", - "title": "To create a VPC" - } - ], - "DeleteCustomerGateway": [ - { - "input": { - "CustomerGatewayId": "cgw-0e11f167" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified customer gateway.", - "id": "ec2-delete-customer-gateway-1", - "title": "To delete a customer gateway" - } - ], - "DeleteDhcpOptions": [ - { - "input": { - "DhcpOptionsId": "dopt-d9070ebb" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified DHCP options set.", - "id": "ec2-delete-dhcp-options-1", - "title": "To delete a DHCP options set" - } - ], - "DeleteInternetGateway": [ - { - "input": { - "InternetGatewayId": "igw-c0a643a9" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified Internet gateway.", - "id": "ec2-delete-internet-gateway-1", - "title": "To delete an Internet gateway" - } - ], - "DeleteKeyPair": [ - { - "input": { - "KeyName": "my-key-pair" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified key pair.", - "id": "ec2-delete-key-pair-1", - "title": "To delete a key pair" - } - ], - "DeleteNatGateway": [ - { - "input": { - "NatGatewayId": "nat-04ae55e711cec5680" - }, - "output": { - "NatGatewayId": "nat-04ae55e711cec5680" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified NAT gateway.", - "id": "ec2-delete-nat-gateway-1", - "title": "To delete a NAT gateway" - } - ], - "DeleteNetworkAcl": [ - { - "input": { - "NetworkAclId": "acl-5fb85d36" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified network ACL.", - "id": "ec2-delete-network-acl-1", - "title": "To delete a network ACL" - } - ], - "DeleteNetworkAclEntry": [ - { - "input": { - "Egress": true, - "NetworkAclId": "acl-5fb85d36", - "RuleNumber": 100 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes ingress rule number 100 from the specified network ACL.", - "id": "ec2-delete-network-acl-entry-1", - "title": "To delete a network ACL entry" - } - ], - "DeleteNetworkInterface": [ - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified network interface.", - "id": "ec2-delete-network-interface-1", - "title": "To delete a network interface" - } - ], - "DeletePlacementGroup": [ - { - "input": { - "GroupName": "my-cluster" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified placement group.\n", - "id": "to-delete-a-placement-group-1472712349959", - "title": "To delete a placement group" - } - ], - "DeleteRoute": [ - { - "input": { - "DestinationCidrBlock": "0.0.0.0/0", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified route from the specified route table.", - "id": "ec2-delete-route-1", - "title": "To delete a route" - } - ], - "DeleteRouteTable": [ - { - "input": { - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified route table.", - "id": "ec2-delete-route-table-1", - "title": "To delete a route table" - } - ], - "DeleteSnapshot": [ - { - "input": { - "SnapshotId": "snap-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes a snapshot with the snapshot ID of ``snap-1234567890abcdef0``. If the command succeeds, no output is returned.", - "id": "to-delete-a-snapshot-1472503042567", - "title": "To delete a snapshot" - } - ], - "DeleteSpotDatafeedSubscription": [ - { - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes a Spot data feed subscription for the account.", - "id": "ec2-delete-spot-datafeed-subscription-1", - "title": "To cancel a Spot Instance data feed subscription" - } - ], - "DeleteSubnet": [ - { - "input": { - "SubnetId": "subnet-9d4a7b6c" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified subnet.", - "id": "ec2-delete-subnet-1", - "title": "To delete a subnet" - } - ], - "DeleteTags": [ - { - "input": { - "Resources": [ - "ami-78a54011" - ], - "Tags": [ - { - "Key": "Stack", - "Value": "test" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the tag Stack=test from the specified image.", - "id": "ec2-delete-tags-1", - "title": "To delete a tag from a resource" - } - ], - "DeleteVolume": [ - { - "input": { - "VolumeId": "vol-049df61146c4d7901" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes an available volume with the volume ID of ``vol-049df61146c4d7901``. If the command succeeds, no output is returned.", - "id": "to-delete-a-volume-1472503111160", - "title": "To delete a volume" - } - ], - "DeleteVpc": [ - { - "input": { - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example deletes the specified VPC.", - "id": "ec2-delete-vpc-1", - "title": "To delete a VPC" - } - ], - "DescribeAccountAttributes": [ - { - "input": { - "AttributeNames": [ - "supported-platforms" - ] - }, - "output": { - "AccountAttributes": [ - { - "AttributeName": "supported-platforms", - "AttributeValues": [ - { - "AttributeValue": "EC2" - }, - { - "AttributeValue": "VPC" - } - ] - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the supported-platforms attribute for your AWS account.", - "id": "ec2-describe-account-attributes-1", - "title": "To describe a single attribute for your AWS account" - }, - { - "output": { - "AccountAttributes": [ - { - "AttributeName": "supported-platforms", - "AttributeValues": [ - { - "AttributeValue": "EC2" - }, - { - "AttributeValue": "VPC" - } - ] - }, - { - "AttributeName": "vpc-max-security-groups-per-interface", - "AttributeValues": [ - { - "AttributeValue": "5" - } - ] - }, - { - "AttributeName": "max-elastic-ips", - "AttributeValues": [ - { - "AttributeValue": "5" - } - ] - }, - { - "AttributeName": "max-instances", - "AttributeValues": [ - { - "AttributeValue": "20" - } - ] - }, - { - "AttributeName": "vpc-max-elastic-ips", - "AttributeValues": [ - { - "AttributeValue": "5" - } - ] - }, - { - "AttributeName": "default-vpc", - "AttributeValues": [ - { - "AttributeValue": "none" - } - ] - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the attributes for your AWS account.", - "id": "ec2-describe-account-attributes-2", - "title": "To describe all attributes for your AWS account" - } - ], - "DescribeAddresses": [ - { - "output": { - "Addresses": [ - { - "Domain": "standard", - "InstanceId": "i-1234567890abcdef0", - "PublicIp": "198.51.100.0" - }, - { - "AllocationId": "eipalloc-12345678", - "AssociationId": "eipassoc-12345678", - "Domain": "vpc", - "InstanceId": "i-1234567890abcdef0", - "NetworkInterfaceId": "eni-12345678", - "NetworkInterfaceOwnerId": "123456789012", - "PrivateIpAddress": "10.0.1.241", - "PublicIp": "203.0.113.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes your Elastic IP addresses.", - "id": "ec2-describe-addresses-1", - "title": "To describe your Elastic IP addresses" - }, - { - "input": { - "Filters": [ - { - "Name": "domain", - "Values": [ - "vpc" - ] - } - ] - }, - "output": { - "Addresses": [ - { - "AllocationId": "eipalloc-12345678", - "AssociationId": "eipassoc-12345678", - "Domain": "vpc", - "InstanceId": "i-1234567890abcdef0", - "NetworkInterfaceId": "eni-12345678", - "NetworkInterfaceOwnerId": "123456789012", - "PrivateIpAddress": "10.0.1.241", - "PublicIp": "203.0.113.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes your Elastic IP addresses for use with instances in a VPC.", - "id": "ec2-describe-addresses-2", - "title": "To describe your Elastic IP addresses for EC2-VPC" - }, - { - "input": { - "Filters": [ - { - "Name": "domain", - "Values": [ - "standard" - ] - } - ] - }, - "output": { - "Addresses": [ - { - "Domain": "standard", - "InstanceId": "i-1234567890abcdef0", - "PublicIp": "198.51.100.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes your Elastic IP addresses for use with instances in EC2-Classic.", - "id": "ec2-describe-addresses-3", - "title": "To describe your Elastic IP addresses for EC2-Classic" - } - ], - "DescribeAvailabilityZones": [ - { - "output": { - "AvailabilityZones": [ - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1b" - }, - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1c" - }, - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1d" - }, - { - "Messages": [ - - ], - "RegionName": "us-east-1", - "State": "available", - "ZoneName": "us-east-1e" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the Availability Zones that are available to you. The response includes Availability Zones only for the current region.", - "id": "ec2-describe-availability-zones-1", - "title": "To describe your Availability Zones" - } - ], - "DescribeCustomerGateways": [ - { - "input": { - "CustomerGatewayIds": [ - "cgw-0e11f167" - ] - }, - "output": { - "CustomerGateways": [ - { - "BgpAsn": "65534", - "CustomerGatewayId": "cgw-0e11f167", - "IpAddress": "12.1.2.3", - "State": "available", - "Type": "ipsec.1" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified customer gateway.", - "id": "ec2-describe-customer-gateways-1", - "title": "To describe a customer gateway" - } - ], - "DescribeDhcpOptions": [ - { - "input": { - "DhcpOptionsIds": [ - "dopt-d9070ebb" - ] - }, - "output": { - "DhcpOptions": [ - { - "DhcpConfigurations": [ - { - "Key": "domain-name-servers", - "Values": [ - { - "Value": "10.2.5.2" - }, - { - "Value": "10.2.5.1" - } - ] - } - ], - "DhcpOptionsId": "dopt-d9070ebb" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified DHCP options set.", - "id": "ec2-describe-dhcp-options-1", - "title": "To describe a DHCP options set" - } - ], - "DescribeInstanceAttribute": [ - { - "input": { - "Attribute": "instanceType", - "InstanceId": "i-1234567890abcdef0" - }, - "output": { - "InstanceId": "i-1234567890abcdef0", - "InstanceType": { - "Value": "t1.micro" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the instance type of the specified instance.\n", - "id": "to-describe-the-instance-type-1472712432132", - "title": "To describe the instance type" - }, - { - "input": { - "Attribute": "disableApiTermination", - "InstanceId": "i-1234567890abcdef0" - }, - "output": { - "DisableApiTermination": { - "Value": "false" - }, - "InstanceId": "i-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``disableApiTermination`` attribute of the specified instance.\n", - "id": "to-describe-the-disableapitermination-attribute-1472712533466", - "title": "To describe the disableApiTermination attribute" - }, - { - "input": { - "Attribute": "blockDeviceMapping", - "InstanceId": "i-1234567890abcdef0" - }, - "output": { - "BlockDeviceMappings": [ - { - "DeviceName": "/dev/sda1", - "Ebs": { - "AttachTime": "2013-05-17T22:42:34.000Z", - "DeleteOnTermination": true, - "Status": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - }, - { - "DeviceName": "/dev/sdf", - "Ebs": { - "AttachTime": "2013-09-10T23:07:00.000Z", - "DeleteOnTermination": false, - "Status": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - } - ], - "InstanceId": "i-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``blockDeviceMapping`` attribute of the specified instance.\n", - "id": "to-describe-the-block-device-mapping-for-an-instance-1472712645423", - "title": "To describe the block device mapping for an instance" - } - ], - "DescribeInternetGateways": [ - { - "input": { - "Filters": [ - { - "Name": "attachment.vpc-id", - "Values": [ - "vpc-a01106c2" - ] - } - ] - }, - "output": { - "InternetGateways": [ - { - "Attachments": [ - { - "State": "available", - "VpcId": "vpc-a01106c2" - } - ], - "InternetGatewayId": "igw-c0a643a9", - "Tags": [ - - ] - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the Internet gateway for the specified VPC.", - "id": "ec2-describe-internet-gateways-1", - "title": "To describe the Internet gateway for a VPC" - } - ], - "DescribeKeyPairs": [ - { - "input": { - "KeyNames": [ - "my-key-pair" - ] - }, - "output": { - "KeyPairs": [ - { - "KeyFingerprint": "1f:51:ae:28:bf:89:e9:d8:1f:25:5d:37:2d:7d:b8:ca:9f:f5:f1:6f", - "KeyName": "my-key-pair" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example displays the fingerprint for the specified key.", - "id": "ec2-describe-key-pairs-1", - "title": "To display a key pair" - } - ], - "DescribeMovingAddresses": [ - { - "output": { - "MovingAddressStatuses": [ - { - "MoveStatus": "MovingToVpc", - "PublicIp": "198.51.100.0" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all of your moving Elastic IP addresses.", - "id": "ec2-describe-moving-addresses-1", - "title": "To describe your moving addresses" - } - ], - "DescribeNatGateways": [ - { - "input": { - "Filter": [ - { - "Name": "vpc-id", - "Values": [ - "vpc-1a2b3c4d" - ] - } - ] - }, - "output": { - "NatGateways": [ - { - "CreateTime": "2015-12-01T12:26:55.983Z", - "NatGatewayAddresses": [ - { - "AllocationId": "eipalloc-89c620ec", - "NetworkInterfaceId": "eni-9dec76cd", - "PrivateIp": "10.0.0.149", - "PublicIp": "198.11.222.333" - } - ], - "NatGatewayId": "nat-05dba92075d71c408", - "State": "available", - "SubnetId": "subnet-847e4dc2", - "VpcId": "vpc-1a2b3c4d" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the NAT gateway for the specified VPC.", - "id": "ec2-describe-nat-gateways-1", - "title": "To describe a NAT gateway" - } - ], - "DescribeNetworkAcls": [ - { - "input": { - "NetworkAclIds": [ - "acl-5fb85d36" - ] - }, - "output": { - "NetworkAcls": [ - { - "Associations": [ - { - "NetworkAclAssociationId": "aclassoc-66ea5f0b", - "NetworkAclId": "acl-9aeb5ef7", - "SubnetId": "subnet-65ea5f08" - } - ], - "Entries": [ - { - "CidrBlock": "0.0.0.0/0", - "Egress": true, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - }, - { - "CidrBlock": "0.0.0.0/0", - "Egress": false, - "Protocol": "-1", - "RuleAction": "deny", - "RuleNumber": 32767 - } - ], - "IsDefault": false, - "NetworkAclId": "acl-5fb85d36", - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified network ACL.", - "id": "ec2-", - "title": "To describe a network ACL" - } - ], - "DescribeNetworkInterfaceAttribute": [ - { - "input": { - "Attribute": "attachment", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "Attachment": { - "AttachTime": "2015-05-21T20:02:20.000Z", - "AttachmentId": "eni-attach-43348162", - "DeleteOnTermination": true, - "DeviceIndex": 0, - "InstanceId": "i-1234567890abcdef0", - "InstanceOwnerId": "123456789012", - "Status": "attached" - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the attachment attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-1", - "title": "To describe the attachment attribute of a network interface" - }, - { - "input": { - "Attribute": "description", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "Description": { - "Value": "My description" - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the description attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-2", - "title": "To describe the description attribute of a network interface" - }, - { - "input": { - "Attribute": "groupSet", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "Groups": [ - { - "GroupId": "sg-903004f8", - "GroupName": "my-security-group" - } - ], - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the groupSet attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-3", - "title": "To describe the groupSet attribute of a network interface" - }, - { - "input": { - "Attribute": "sourceDestCheck", - "NetworkInterfaceId": "eni-686ea200" - }, - "output": { - "NetworkInterfaceId": "eni-686ea200", - "SourceDestCheck": { - "Value": true - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the sourceDestCheck attribute of the specified network interface.", - "id": "ec2-describe-network-interface-attribute-4", - "title": "To describe the sourceDestCheck attribute of a network interface" - } - ], - "DescribeNetworkInterfaces": [ - { - "input": { - "NetworkInterfaceIds": [ - "eni-e5aa89a3" - ] - }, - "output": { - "NetworkInterfaces": [ - { - "Association": { - "AssociationId": "eipassoc-0fbb766a", - "IpOwnerId": "123456789012", - "PublicDnsName": "ec2-203-0-113-12.compute-1.amazonaws.com", - "PublicIp": "203.0.113.12" - }, - "Attachment": { - "AttachTime": "2013-11-30T23:36:42.000Z", - "AttachmentId": "eni-attach-66c4350a", - "DeleteOnTermination": false, - "DeviceIndex": 1, - "InstanceId": "i-1234567890abcdef0", - "InstanceOwnerId": "123456789012", - "Status": "attached" - }, - "AvailabilityZone": "us-east-1d", - "Description": "my network interface", - "Groups": [ - { - "GroupId": "sg-8637d3e3", - "GroupName": "default" - } - ], - "MacAddress": "02:2f:8f:b0:cf:75", - "NetworkInterfaceId": "eni-e5aa89a3", - "OwnerId": "123456789012", - "PrivateDnsName": "ip-10-0-1-17.ec2.internal", - "PrivateIpAddress": "10.0.1.17", - "PrivateIpAddresses": [ - { - "Association": { - "AssociationId": "eipassoc-0fbb766a", - "IpOwnerId": "123456789012", - "PublicDnsName": "ec2-203-0-113-12.compute-1.amazonaws.com", - "PublicIp": "203.0.113.12" - }, - "Primary": true, - "PrivateDnsName": "ip-10-0-1-17.ec2.internal", - "PrivateIpAddress": "10.0.1.17" - } - ], - "RequesterManaged": false, - "SourceDestCheck": true, - "Status": "in-use", - "SubnetId": "subnet-b61f49f0", - "TagSet": [ - - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "", - "id": "ec2-describe-network-interfaces-1", - "title": "To describe a network interface" - } - ], - "DescribeRegions": [ - { - "output": { - "Regions": [ - { - "Endpoint": "ec2.ap-south-1.amazonaws.com", - "RegionName": "ap-south-1" - }, - { - "Endpoint": "ec2.eu-west-1.amazonaws.com", - "RegionName": "eu-west-1" - }, - { - "Endpoint": "ec2.ap-southeast-1.amazonaws.com", - "RegionName": "ap-southeast-1" - }, - { - "Endpoint": "ec2.ap-southeast-2.amazonaws.com", - "RegionName": "ap-southeast-2" - }, - { - "Endpoint": "ec2.eu-central-1.amazonaws.com", - "RegionName": "eu-central-1" - }, - { - "Endpoint": "ec2.ap-northeast-2.amazonaws.com", - "RegionName": "ap-northeast-2" - }, - { - "Endpoint": "ec2.ap-northeast-1.amazonaws.com", - "RegionName": "ap-northeast-1" - }, - { - "Endpoint": "ec2.us-east-1.amazonaws.com", - "RegionName": "us-east-1" - }, - { - "Endpoint": "ec2.sa-east-1.amazonaws.com", - "RegionName": "sa-east-1" - }, - { - "Endpoint": "ec2.us-west-1.amazonaws.com", - "RegionName": "us-west-1" - }, - { - "Endpoint": "ec2.us-west-2.amazonaws.com", - "RegionName": "us-west-2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all the regions that are available to you.", - "id": "ec2-describe-regions-1", - "title": "To describe your regions" - } - ], - "DescribeRouteTables": [ - { - "input": { - "RouteTableIds": [ - "rtb-1f382e7d" - ] - }, - "output": { - "RouteTables": [ - { - "Associations": [ - { - "Main": true, - "RouteTableAssociationId": "rtbassoc-d8ccddba", - "RouteTableId": "rtb-1f382e7d" - } - ], - "PropagatingVgws": [ - - ], - "RouteTableId": "rtb-1f382e7d", - "Routes": [ - { - "DestinationCidrBlock": "10.0.0.0/16", - "GatewayId": "local", - "State": "active" - } - ], - "Tags": [ - - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified route table.", - "id": "ec2-describe-route-tables-1", - "title": "To describe a route table" - } - ], - "DescribeScheduledInstanceAvailability": [ - { - "input": { - "FirstSlotStartTimeRange": { - "EarliestTime": "2016-01-31T00:00:00Z", - "LatestTime": "2016-01-31T04:00:00Z" - }, - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDays": [ - 1 - ] - } - }, - "output": { - "ScheduledInstanceAvailabilitySet": [ - { - "AvailabilityZone": "us-west-2b", - "AvailableInstanceCount": 20, - "FirstSlotStartTime": "2016-01-31T00:00:00Z", - "HourlyPrice": "0.095", - "InstanceType": "c4.large", - "MaxTermDurationInDays": 366, - "MinTermDurationInDays": 366, - "NetworkPlatform": "EC2-VPC", - "Platform": "Linux/UNIX", - "PurchaseToken": "eyJ2IjoiMSIsInMiOjEsImMiOi...", - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDaySet": [ - 1 - ], - "OccurrenceRelativeToEnd": false - }, - "SlotDurationInHours": 23, - "TotalScheduledInstanceHours": 1219 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes a schedule that occurs every week on Sunday, starting on the specified date. Note that the output contains a single schedule as an example.", - "id": "ec2-describe-scheduled-instance-availability-1", - "title": "To describe an available schedule" - } - ], - "DescribeScheduledInstances": [ - { - "input": { - "ScheduledInstanceIds": [ - "sci-1234-1234-1234-1234-123456789012" - ] - }, - "output": { - "ScheduledInstanceSet": [ - { - "AvailabilityZone": "us-west-2b", - "CreateDate": "2016-01-25T21:43:38.612Z", - "HourlyPrice": "0.095", - "InstanceCount": 1, - "InstanceType": "c4.large", - "NetworkPlatform": "EC2-VPC", - "NextSlotStartTime": "2016-01-31T09:00:00Z", - "Platform": "Linux/UNIX", - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDaySet": [ - 1 - ], - "OccurrenceRelativeToEnd": false, - "OccurrenceUnit": "" - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012", - "SlotDurationInHours": 32, - "TermEndDate": "2017-01-31T09:00:00Z", - "TermStartDate": "2016-01-31T09:00:00Z", - "TotalScheduledInstanceHours": 1696 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified Scheduled Instance.", - "id": "ec2-describe-scheduled-instances-1", - "title": "To describe your Scheduled Instances" - } - ], - "DescribeSnapshotAttribute": [ - { - "input": { - "Attribute": "createVolumePermission", - "SnapshotId": "snap-066877671789bd71b" - }, - "output": { - "CreateVolumePermissions": [ - - ], - "SnapshotId": "snap-066877671789bd71b" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``createVolumePermission`` attribute on a snapshot with the snapshot ID of ``snap-066877671789bd71b``.", - "id": "to-describe-snapshot-attributes-1472503199736", - "title": "To describe snapshot attributes" - } - ], - "DescribeSnapshots": [ - { - "input": { - "SnapshotIds": [ - "snap-1234567890abcdef0" - ] - }, - "output": { - "NextToken": "", - "Snapshots": [ - { - "Description": "This is my snapshot.", - "OwnerId": "012345678910", - "Progress": "100%", - "SnapshotId": "snap-1234567890abcdef0", - "StartTime": "2014-02-28T21:28:32.000Z", - "State": "completed", - "VolumeId": "vol-049df61146c4d7901", - "VolumeSize": 8 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes a snapshot with the snapshot ID of ``snap-1234567890abcdef0``.", - "id": "to-describe-a-snapshot-1472503807850", - "title": "To describe a snapshot" - }, - { - "input": { - "Filters": [ - { - "Name": "status", - "Values": [ - "pending" - ] - } - ], - "OwnerIds": [ - "012345678910" - ] - }, - "output": { - "NextToken": "", - "Snapshots": [ - { - "Description": "This is my copied snapshot.", - "OwnerId": "012345678910", - "Progress": "87%", - "SnapshotId": "snap-066877671789bd71b", - "StartTime": "2014-02-28T21:37:27.000Z", - "State": "pending", - "VolumeId": "vol-1234567890abcdef0", - "VolumeSize": 8 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all snapshots owned by the ID 012345678910 that are in the ``pending`` status.", - "id": "to-describe-snapshots-using-filters-1472503929793", - "title": "To describe snapshots using filters" - } - ], - "DescribeSpotDatafeedSubscription": [ - { - "output": { - "SpotDatafeedSubscription": { - "Bucket": "my-s3-bucket", - "OwnerId": "123456789012", - "Prefix": "spotdata", - "State": "Active" - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the Spot Instance datafeed subscription for your AWS account.", - "id": "ec2-describe-spot-datafeed-subscription-1", - "title": "To describe the datafeed for your AWS account" - } - ], - "DescribeSpotFleetInstances": [ - { - "input": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "output": { - "ActiveInstances": [ - { - "InstanceId": "i-1234567890abcdef0", - "InstanceType": "m3.medium", - "SpotInstanceRequestId": "sir-08b93456" - } - ], - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example lists the Spot Instances associated with the specified Spot fleet.", - "id": "ec2-describe-spot-fleet-instances-1", - "title": "To describe the Spot Instances associated with a Spot fleet" - } - ], - "DescribeSpotFleetRequestHistory": [ - { - "input": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "StartTime": "2015-05-26T00:00:00Z" - }, - "output": { - "HistoryRecords": [ - { - "EventInformation": { - "EventSubType": "submitted" - }, - "EventType": "fleetRequestChange", - "Timestamp": "2015-05-26T23:17:20.697Z" - }, - { - "EventInformation": { - "EventSubType": "active" - }, - "EventType": "fleetRequestChange", - "Timestamp": "2015-05-26T23:17:20.873Z" - }, - { - "EventInformation": { - "EventSubType": "launched", - "InstanceId": "i-1234567890abcdef0" - }, - "EventType": "instanceChange", - "Timestamp": "2015-05-26T23:21:21.712Z" - }, - { - "EventInformation": { - "EventSubType": "launched", - "InstanceId": "i-1234567890abcdef1" - }, - "EventType": "instanceChange", - "Timestamp": "2015-05-26T23:21:21.816Z" - } - ], - "NextToken": "CpHNsscimcV5oH7bSbub03CI2Qms5+ypNpNm+53MNlR0YcXAkp0xFlfKf91yVxSExmbtma3awYxMFzNA663ZskT0AHtJ6TCb2Z8bQC2EnZgyELbymtWPfpZ1ZbauVg+P+TfGlWxWWB/Vr5dk5d4LfdgA/DRAHUrYgxzrEXAMPLE=", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "StartTime": "2015-05-26T00:00:00Z" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example returns the history for the specified Spot fleet starting at the specified time.", - "id": "ec2-describe-spot-fleet-request-history-1", - "title": "To describe Spot fleet history" - } - ], - "DescribeSpotFleetRequests": [ - { - "input": { - "SpotFleetRequestIds": [ - "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - ] - }, - "output": { - "SpotFleetRequestConfigs": [ - { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "EbsOptimized": false, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "cc2.8xlarge", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeleteOnTermination": false, - "DeviceIndex": 0, - "SecondaryPrivateIpAddressCount": 0, - "SubnetId": "subnet-a61dafcf" - } - ] - }, - { - "EbsOptimized": false, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "r3.8xlarge", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeleteOnTermination": false, - "DeviceIndex": 0, - "SecondaryPrivateIpAddressCount": 0, - "SubnetId": "subnet-a61dafcf" - } - ] - } - ], - "SpotPrice": "0.05", - "TargetCapacity": 20 - }, - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "SpotFleetRequestState": "active" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified Spot fleet request.", - "id": "ec2-describe-spot-fleet-requests-1", - "title": "To describe a Spot fleet request" - } - ], - "DescribeSpotInstanceRequests": [ - { - "input": { - "SpotInstanceRequestIds": [ - "sir-08b93456" - ] - }, - "output": { - "SpotInstanceRequests": [ - { - "CreateTime": "2014-04-30T18:14:55.000Z", - "InstanceId": "i-1234567890abcdef0", - "LaunchSpecification": { - "BlockDeviceMappings": [ - { - "DeviceName": "/dev/sda1", - "Ebs": { - "DeleteOnTermination": true, - "VolumeSize": 8, - "VolumeType": "standard" - } - } - ], - "EbsOptimized": false, - "ImageId": "ami-7aba833f", - "InstanceType": "m1.small", - "KeyName": "my-key-pair", - "SecurityGroups": [ - { - "GroupId": "sg-e38f24a7", - "GroupName": "my-security-group" - } - ] - }, - "LaunchedAvailabilityZone": "us-west-1b", - "ProductDescription": "Linux/UNIX", - "SpotInstanceRequestId": "sir-08b93456", - "SpotPrice": "0.010000", - "State": "active", - "Status": { - "Code": "fulfilled", - "Message": "Your Spot request is fulfilled.", - "UpdateTime": "2014-04-30T18:16:21.000Z" - }, - "Type": "one-time" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified Spot Instance request.", - "id": "ec2-describe-spot-instance-requests-1", - "title": "To describe a Spot Instance request" - } - ], - "DescribeSpotPriceHistory": [ - { - "input": { - "EndTime": "2014-01-06T08:09:10", - "InstanceTypes": [ - "m1.xlarge" - ], - "ProductDescriptions": [ - "Linux/UNIX (Amazon VPC)" - ], - "StartTime": "2014-01-06T07:08:09" - }, - "output": { - "SpotPriceHistory": [ - { - "AvailabilityZone": "us-west-1a", - "InstanceType": "m1.xlarge", - "ProductDescription": "Linux/UNIX (Amazon VPC)", - "SpotPrice": "0.080000", - "Timestamp": "2014-01-06T04:32:53.000Z" - }, - { - "AvailabilityZone": "us-west-1c", - "InstanceType": "m1.xlarge", - "ProductDescription": "Linux/UNIX (Amazon VPC)", - "SpotPrice": "0.080000", - "Timestamp": "2014-01-05T11:28:26.000Z" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example returns the Spot Price history for m1.xlarge, Linux/UNIX (Amazon VPC) instances for a particular day in January.", - "id": "ec2-describe-spot-price-history-1", - "title": "To describe Spot price history for Linux/UNIX (Amazon VPC)" - } - ], - "DescribeSubnets": [ - { - "input": { - "Filters": [ - { - "Name": "vpc-id", - "Values": [ - "vpc-a01106c2" - ] - } - ] - }, - "output": { - "Subnets": [ - { - "AvailabilityZone": "us-east-1c", - "AvailableIpAddressCount": 251, - "CidrBlock": "10.0.1.0/24", - "DefaultForAz": false, - "MapPublicIpOnLaunch": false, - "State": "available", - "SubnetId": "subnet-9d4a7b6c", - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the subnets for the specified VPC.", - "id": "ec2-describe-subnets-1", - "title": "To describe the subnets for a VPC" - } - ], - "DescribeTags": [ - { - "input": { - "Filters": [ - { - "Name": "resource-id", - "Values": [ - "i-1234567890abcdef8" - ] - } - ] - }, - "output": { - "Tags": [ - { - "Key": "Stack", - "ResourceId": "i-1234567890abcdef8", - "ResourceType": "instance", - "Value": "test" - }, - { - "Key": "Name", - "ResourceId": "i-1234567890abcdef8", - "ResourceType": "instance", - "Value": "Beta Server" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the tags for the specified instance.", - "id": "ec2-describe-tags-1", - "title": "To describe the tags for a single resource" - } - ], - "DescribeVolumeAttribute": [ - { - "input": { - "Attribute": "autoEnableIO", - "VolumeId": "vol-049df61146c4d7901" - }, - "output": { - "AutoEnableIO": { - "Value": false - }, - "VolumeId": "vol-049df61146c4d7901" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the ``autoEnableIo`` attribute of the volume with the ID ``vol-049df61146c4d7901``.", - "id": "to-describe-a-volume-attribute-1472505773492", - "title": "To describe a volume attribute" - } - ], - "DescribeVolumeStatus": [ - { - "input": { - "VolumeIds": [ - "vol-1234567890abcdef0" - ] - }, - "output": { - "VolumeStatuses": [ - { - "Actions": [ - - ], - "AvailabilityZone": "us-east-1a", - "Events": [ - - ], - "VolumeId": "vol-1234567890abcdef0", - "VolumeStatus": { - "Details": [ - { - "Name": "io-enabled", - "Status": "passed" - }, - { - "Name": "io-performance", - "Status": "not-applicable" - } - ], - "Status": "ok" - } - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the status for the volume ``vol-1234567890abcdef0``.", - "id": "to-describe-the-status-of-a-single-volume-1472507016193", - "title": "To describe the status of a single volume" - }, - { - "input": { - "Filters": [ - { - "Name": "volume-status.status", - "Values": [ - "impaired" - ] - } - ] - }, - "output": { - "VolumeStatuses": [ - - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the status for all volumes that are impaired. In this example output, there are no impaired volumes.", - "id": "to-describe-the-status-of-impaired-volumes-1472507239821", - "title": "To describe the status of impaired volumes" - } - ], - "DescribeVolumes": [ - { - "input": { - }, - "output": { - "NextToken": "", - "Volumes": [ - { - "Attachments": [ - { - "AttachTime": "2013-12-18T22:35:00.000Z", - "DeleteOnTermination": true, - "Device": "/dev/sda1", - "InstanceId": "i-1234567890abcdef0", - "State": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - ], - "AvailabilityZone": "us-east-1a", - "CreateTime": "2013-12-18T22:35:00.084Z", - "Size": 8, - "SnapshotId": "snap-1234567890abcdef0", - "State": "in-use", - "VolumeId": "vol-049df61146c4d7901", - "VolumeType": "standard" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all of your volumes in the default region.", - "id": "to-describe-all-volumes-1472506358883", - "title": "To describe all volumes" - }, - { - "input": { - "Filters": [ - { - "Name": "attachment.instance-id", - "Values": [ - "i-1234567890abcdef0" - ] - }, - { - "Name": "attachment.delete-on-termination", - "Values": [ - "true" - ] - } - ] - }, - "output": { - "Volumes": [ - { - "Attachments": [ - { - "AttachTime": "2013-12-18T22:35:00.000Z", - "DeleteOnTermination": true, - "Device": "/dev/sda1", - "InstanceId": "i-1234567890abcdef0", - "State": "attached", - "VolumeId": "vol-049df61146c4d7901" - } - ], - "AvailabilityZone": "us-east-1a", - "CreateTime": "2013-12-18T22:35:00.084Z", - "Size": 8, - "SnapshotId": "snap-1234567890abcdef0", - "State": "in-use", - "VolumeId": "vol-049df61146c4d7901", - "VolumeType": "standard" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes all volumes that are both attached to the instance with the ID i-1234567890abcdef0 and set to delete when the instance terminates.", - "id": "to-describe-volumes-that-are-attached-to-a-specific-instance-1472506613578", - "title": "To describe volumes that are attached to a specific instance" - } - ], - "DescribeVpcAttribute": [ - { - "input": { - "Attribute": "enableDnsSupport", - "VpcId": "vpc-a01106c2" - }, - "output": { - "EnableDnsSupport": { - "Value": true - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the enableDnsSupport attribute. This attribute indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for your instances to their corresponding IP addresses; otherwise, it does not.", - "id": "ec2-describe-vpc-attribute-1", - "title": "To describe the enableDnsSupport attribute" - }, - { - "input": { - "Attribute": "enableDnsHostnames", - "VpcId": "vpc-a01106c2" - }, - "output": { - "EnableDnsHostnames": { - "Value": true - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the enableDnsHostnames attribute. This attribute indicates whether the instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.", - "id": "ec2-describe-vpc-attribute-2", - "title": "To describe the enableDnsHostnames attribute" - } - ], - "DescribeVpcs": [ - { - "input": { - "VpcIds": [ - "vpc-a01106c2" - ] - }, - "output": { - "Vpcs": [ - { - "CidrBlock": "10.0.0.0/16", - "DhcpOptionsId": "dopt-7a8b9c2d", - "InstanceTenancy": "default", - "IsDefault": false, - "State": "available", - "Tags": [ - { - "Key": "Name", - "Value": "MyVPC" - } - ], - "VpcId": "vpc-a01106c2" - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example describes the specified VPC.", - "id": "ec2-describe-vpcs-1", - "title": "To describe a VPC" - } - ], - "DetachInternetGateway": [ - { - "input": { - "InternetGatewayId": "igw-c0a643a9", - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example detaches the specified Internet gateway from the specified VPC.", - "id": "ec2-detach-internet-gateway-1", - "title": "To detach an Internet gateway from a VPC" - } - ], - "DetachNetworkInterface": [ - { - "input": { - "AttachmentId": "eni-attach-66c4350a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example detaches the specified network interface from its attached instance.", - "id": "ec2-detach-network-interface-1", - "title": "To detach a network interface from an instance" - } - ], - "DetachVolume": [ - { - "input": { - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - "AttachTime": "2014-02-27T19:23:06.000Z", - "Device": "/dev/sdb", - "InstanceId": "i-1234567890abcdef0", - "State": "detaching", - "VolumeId": "vol-049df61146c4d7901" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example detaches the volume (``vol-049df61146c4d7901``) from the instance it is attached to.", - "id": "to-detach-a-volume-from-an-instance-1472507977694", - "title": "To detach a volume from an instance" - } - ], - "DisableVgwRoutePropagation": [ - { - "input": { - "GatewayId": "vgw-9a4cacf3", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disables the specified virtual private gateway from propagating static routes to the specified route table.", - "id": "ec2-disable-vgw-route-propagation-1", - "title": "To disable route propagation" - } - ], - "DisassociateAddress": [ - { - "input": { - "AssociationId": "eipassoc-2bebb745" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disassociates an Elastic IP address from an instance in a VPC.", - "id": "ec2-disassociate-address-1", - "title": "To disassociate an Elastic IP address in EC2-VPC" - }, - { - "input": { - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disassociates an Elastic IP address from an instance in EC2-Classic.", - "id": "ec2-disassociate-address-2", - "title": "To disassociate an Elastic IP addresses in EC2-Classic" - } - ], - "DisassociateRouteTable": [ - { - "input": { - "AssociationId": "rtbassoc-781d0d1a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example disassociates the specified route table from its associated subnet.", - "id": "ec2-disassociate-route-table-1", - "title": "To disassociate a route table" - } - ], - "EnableVgwRoutePropagation": [ - { - "input": { - "GatewayId": "vgw-9a4cacf3", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example enables the specified virtual private gateway to propagate static routes to the specified route table.", - "id": "ec2-enable-vgw-route-propagation-1", - "title": "To enable route propagation" - } - ], - "EnableVolumeIO": [ - { - "input": { - "VolumeId": "vol-1234567890abcdef0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example enables I/O on volume ``vol-1234567890abcdef0``.", - "id": "to-enable-io-for-a-volume-1472508114867", - "title": "To enable I/O for a volume" - } - ], - "ModifyNetworkInterfaceAttribute": [ - { - "input": { - "Attachment": { - "AttachmentId": "eni-attach-43348162", - "DeleteOnTermination": false - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the attachment attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-1", - "title": "To modify the attachment attribute of a network interface" - }, - { - "input": { - "Description": { - "Value": "My description" - }, - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the description attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-2", - "title": "To modify the description attribute of a network interface" - }, - { - "input": { - "Groups": [ - "sg-903004f8", - "sg-1a2b3c4d" - ], - "NetworkInterfaceId": "eni-686ea200" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example command modifies the groupSet attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-3", - "title": "To modify the groupSet attribute of a network interface" - }, - { - "input": { - "NetworkInterfaceId": "eni-686ea200", - "SourceDestCheck": { - "Value": false - } - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example command modifies the sourceDestCheck attribute of the specified network interface.", - "id": "ec2-modify-network-interface-attribute-4", - "title": "To modify the sourceDestCheck attribute of a network interface" - } - ], - "ModifySnapshotAttribute": [ - { - "input": { - "Attribute": "createVolumePermission", - "OperationType": "remove", - "SnapshotId": "snap-1234567890abcdef0", - "UserIds": [ - "123456789012" - ] - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies snapshot ``snap-1234567890abcdef0`` to remove the create volume permission for a user with the account ID ``123456789012``. If the command succeeds, no output is returned.", - "id": "to-modify-a-snapshot-attribute-1472508385907", - "title": "To modify a snapshot attribute" - }, - { - "input": { - "Attribute": "createVolumePermission", - "GroupNames": [ - "all" - ], - "OperationType": "add", - "SnapshotId": "snap-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example makes the snapshot ``snap-1234567890abcdef0`` public.", - "id": "to-make-a-snapshot-public-1472508470529", - "title": "To make a snapshot public" - } - ], - "ModifySpotFleetRequest": [ - { - "input": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "TargetCapacity": 20 - }, - "output": { - "Return": true - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example increases the target capacity of the specified Spot fleet request.", - "id": "ec2-modify-spot-fleet-request-1", - "title": "To increase the target capacity of a Spot fleet request" - }, - { - "input": { - "ExcessCapacityTerminationPolicy": "NoTermination ", - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE", - "TargetCapacity": 10 - }, - "output": { - "Return": true - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example decreases the target capacity of the specified Spot fleet request without terminating any Spot Instances as a result.", - "id": "ec2-modify-spot-fleet-request-2", - "title": "To decrease the target capacity of a Spot fleet request" - } - ], - "ModifySubnetAttribute": [ - { - "input": { - "MapPublicIpOnLaunch": { - "Value": true - }, - "SubnetId": "subnet-1a2b3c4d" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the specified subnet so that all instances launched into this subnet are assigned a public IP address.", - "id": "ec2-modify-subnet-attribute-1", - "title": "To change a subnet's public IP addressing behavior" - } - ], - "ModifyVolumeAttribute": [ - { - "input": { - "AutoEnableIO": { - "Value": true - }, - "DryRun": true, - "VolumeId": "vol-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example sets the ``autoEnableIo`` attribute of the volume with the ID ``vol-1234567890abcdef0`` to ``true``. If the command succeeds, no output is returned.", - "id": "to-modify-a-volume-attribute-1472508596749", - "title": "To modify a volume attribute" - } - ], - "ModifyVpcAttribute": [ - { - "input": { - "EnableDnsSupport": { - "Value": false - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the enableDnsSupport attribute. This attribute indicates whether DNS resolution is enabled for the VPC. If this attribute is true, the Amazon DNS server resolves DNS hostnames for instances in the VPC to their corresponding IP addresses; otherwise, it does not.", - "id": "ec2-modify-vpc-attribute-1", - "title": "To modify the enableDnsSupport attribute" - }, - { - "input": { - "EnableDnsHostnames": { - "Value": false - }, - "VpcId": "vpc-a01106c2" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example modifies the enableDnsHostnames attribute. This attribute indicates whether instances launched in the VPC get DNS hostnames. If this attribute is true, instances in the VPC get DNS hostnames; otherwise, they do not.", - "id": "ec2-modify-vpc-attribute-2", - "title": "To modify the enableDnsHostnames attribute" - } - ], - "MoveAddressToVpc": [ - { - "input": { - "PublicIp": "54.123.4.56" - }, - "output": { - "Status": "MoveInProgress" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example moves the specified Elastic IP address to the EC2-VPC platform.", - "id": "ec2-move-address-to-vpc-1", - "title": "To move an address to EC2-VPC" - } - ], - "PurchaseScheduledInstances": [ - { - "input": { - "PurchaseRequests": [ - { - "InstanceCount": 1, - "PurchaseToken": "eyJ2IjoiMSIsInMiOjEsImMiOi..." - } - ] - }, - "output": { - "ScheduledInstanceSet": [ - { - "AvailabilityZone": "us-west-2b", - "CreateDate": "2016-01-25T21:43:38.612Z", - "HourlyPrice": "0.095", - "InstanceCount": 1, - "InstanceType": "c4.large", - "NetworkPlatform": "EC2-VPC", - "NextSlotStartTime": "2016-01-31T09:00:00Z", - "Platform": "Linux/UNIX", - "Recurrence": { - "Frequency": "Weekly", - "Interval": 1, - "OccurrenceDaySet": [ - 1 - ], - "OccurrenceRelativeToEnd": false, - "OccurrenceUnit": "" - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012", - "SlotDurationInHours": 32, - "TermEndDate": "2017-01-31T09:00:00Z", - "TermStartDate": "2016-01-31T09:00:00Z", - "TotalScheduledInstanceHours": 1696 - } - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example purchases a Scheduled Instance.", - "id": "ec2-purchase-scheduled-instances-1", - "title": "To purchase a Scheduled Instance" - } - ], - "ReleaseAddress": [ - { - "input": { - "AllocationId": "eipalloc-64d5890a" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example releases an Elastic IP address for use with instances in a VPC.", - "id": "ec2-release-address-1", - "title": "To release an Elastic IP address for EC2-VPC" - }, - { - "input": { - "PublicIp": "198.51.100.0" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example releases an Elastic IP address for use with instances in EC2-Classic.", - "id": "ec2-release-address-2", - "title": "To release an Elastic IP addresses for EC2-Classic" - } - ], - "ReplaceNetworkAclAssociation": [ - { - "input": { - "AssociationId": "aclassoc-e5b95c8c", - "NetworkAclId": "acl-5fb85d36" - }, - "output": { - "NewAssociationId": "aclassoc-3999875b" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified network ACL with the subnet for the specified network ACL association.", - "id": "ec2-replace-network-acl-association-1", - "title": "To replace the network ACL associated with a subnet" - } - ], - "ReplaceNetworkAclEntry": [ - { - "input": { - "CidrBlock": "203.0.113.12/24", - "Egress": false, - "NetworkAclId": "acl-5fb85d36", - "PortRange": { - "From": 53, - "To": 53 - }, - "Protocol": "udp", - "RuleAction": "allow", - "RuleNumber": 100 - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example replaces an entry for the specified network ACL. The new rule 100 allows ingress traffic from 203.0.113.12/24 on UDP port 53 (DNS) into any associated subnet.", - "id": "ec2-replace-network-acl-entry-1", - "title": "To replace a network ACL entry" - } - ], - "ReplaceRoute": [ - { - "input": { - "DestinationCidrBlock": "10.0.0.0/16", - "GatewayId": "vgw-9a4cacf3", - "RouteTableId": "rtb-22574640" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example replaces the specified route in the specified table table. The new route matches the specified CIDR and sends the traffic to the specified virtual private gateway.", - "id": "ec2-replace-route-1", - "title": "To replace a route" - } - ], - "ReplaceRouteTableAssociation": [ - { - "input": { - "AssociationId": "rtbassoc-781d0d1a", - "RouteTableId": "rtb-22574640" - }, - "output": { - "NewAssociationId": "rtbassoc-3a1f0f58" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example associates the specified route table with the subnet for the specified route table association.", - "id": "ec2-replace-route-table-association-1", - "title": "To replace the route table associated with a subnet" - } - ], - "RequestSpotFleet": [ - { - "input": { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "SecurityGroups": [ - { - "GroupId": "sg-1a2b3c4d" - } - ], - "SubnetId": "subnet-1a2b3c4d, subnet-3c4d5e6f" - } - ], - "SpotPrice": "0.04", - "TargetCapacity": 2 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot fleet request with two launch specifications that differ only by subnet. The Spot fleet launches the instances in the specified subnet with the lowest price. If the instances are launched in a default VPC, they receive a public IP address by default. If the instances are launched in a nondefault VPC, they do not receive a public IP address by default. Note that you can't specify different subnets from the same Availability Zone in a Spot fleet request.", - "id": "ec2-request-spot-fleet-1", - "title": "To request a Spot fleet in the subnet with the lowest price" - }, - { - "input": { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "Placement": { - "AvailabilityZone": "us-west-2a, us-west-2b" - }, - "SecurityGroups": [ - { - "GroupId": "sg-1a2b3c4d" - } - ] - } - ], - "SpotPrice": "0.04", - "TargetCapacity": 2 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot fleet request with two launch specifications that differ only by Availability Zone. The Spot fleet launches the instances in the specified Availability Zone with the lowest price. If your account supports EC2-VPC only, Amazon EC2 launches the Spot instances in the default subnet of the Availability Zone. If your account supports EC2-Classic, Amazon EC2 launches the instances in EC2-Classic in the Availability Zone.", - "id": "ec2-request-spot-fleet-2", - "title": "To request a Spot fleet in the Availability Zone with the lowest price" - }, - { - "input": { - "SpotFleetRequestConfig": { - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::880185128111:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeviceIndex": 0, - "Groups": [ - "sg-1a2b3c4d" - ], - "SubnetId": "subnet-1a2b3c4d" - } - ] - } - ], - "SpotPrice": "0.04", - "TargetCapacity": 2 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example assigns public addresses to instances launched in a nondefault VPC. Note that when you specify a network interface, you must include the subnet ID and security group ID using the network interface.", - "id": "ec2-request-spot-fleet-3", - "title": "To launch Spot instances in a subnet and assign them public IP addresses" - }, - { - "input": { - "SpotFleetRequestConfig": { - "AllocationStrategy": "diversified", - "IamFleetRole": "arn:aws:iam::123456789012:role/my-spot-fleet-role", - "LaunchSpecifications": [ - { - "ImageId": "ami-1a2b3c4d", - "InstanceType": "c4.2xlarge", - "SubnetId": "subnet-1a2b3c4d" - }, - { - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.2xlarge", - "SubnetId": "subnet-1a2b3c4d" - }, - { - "ImageId": "ami-1a2b3c4d", - "InstanceType": "r3.2xlarge", - "SubnetId": "subnet-1a2b3c4d" - } - ], - "SpotPrice": "0.70", - "TargetCapacity": 30 - } - }, - "output": { - "SpotFleetRequestId": "sfr-73fbd2ce-aa30-494c-8788-1cee4EXAMPLE" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a Spot fleet request that launches 30 instances using the diversified allocation strategy. The launch specifications differ by instance type. The Spot fleet distributes the instances across the launch specifications such that there are 10 instances of each type.", - "id": "ec2-request-spot-fleet-4", - "title": "To request a Spot fleet using the diversified allocation strategy" - } - ], - "RequestSpotInstances": [ - { - "input": { - "InstanceCount": 5, - "LaunchSpecification": { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "KeyName": "my-key-pair", - "Placement": { - "AvailabilityZone": "us-west-2a" - }, - "SecurityGroupIds": [ - "sg-1a2b3c4d" - ] - }, - "SpotPrice": "0.03", - "Type": "one-time" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example creates a one-time Spot Instance request for five instances in the specified Availability Zone. If your account supports EC2-VPC only, Amazon EC2 launches the instances in the default subnet of the specified Availability Zone. If your account supports EC2-Classic, Amazon EC2 launches the instances in EC2-Classic in the specified Availability Zone.", - "id": "ec2-request-spot-instances-1", - "title": "To create a one-time Spot Instance request" - }, - { - "input": { - "InstanceCount": 5, - "LaunchSpecification": { - "IamInstanceProfile": { - "Arn": "arn:aws:iam::123456789012:instance-profile/my-iam-role" - }, - "ImageId": "ami-1a2b3c4d", - "InstanceType": "m3.medium", - "SecurityGroupIds": [ - "sg-1a2b3c4d" - ], - "SubnetId": "subnet-1a2b3c4d" - }, - "SpotPrice": "0.050", - "Type": "one-time" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example command creates a one-time Spot Instance request for five instances in the specified subnet. Amazon EC2 launches the instances in the specified subnet. If the VPC is a nondefault VPC, the instances do not receive a public IP address by default.", - "id": "ec2-request-spot-instances-2", - "title": "To create a one-time Spot Instance request" - } - ], - "ResetSnapshotAttribute": [ - { - "input": { - "Attribute": "createVolumePermission", - "SnapshotId": "snap-1234567890abcdef0" - }, - "output": { - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example resets the create volume permissions for snapshot ``snap-1234567890abcdef0``. If the command succeeds, no output is returned.", - "id": "to-reset-a-snapshot-attribute-1472508825735", - "title": "To reset a snapshot attribute" - } - ], - "RestoreAddressToClassic": [ - { - "input": { - "PublicIp": "198.51.100.0" - }, - "output": { - "PublicIp": "198.51.100.0", - "Status": "MoveInProgress" - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example restores the specified Elastic IP address to the EC2-Classic platform.", - "id": "ec2-restore-address-to-classic-1", - "title": "To restore an address to EC2-Classic" - } - ], - "RunScheduledInstances": [ - { - "input": { - "InstanceCount": 1, - "LaunchSpecification": { - "IamInstanceProfile": { - "Name": "my-iam-role" - }, - "ImageId": "ami-12345678", - "InstanceType": "c4.large", - "KeyName": "my-key-pair", - "NetworkInterfaces": [ - { - "AssociatePublicIpAddress": true, - "DeviceIndex": 0, - "Groups": [ - "sg-12345678" - ], - "SubnetId": "subnet-12345678" - } - ] - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012" - }, - "output": { - "InstanceIdSet": [ - "i-1234567890abcdef0" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example launches the specified Scheduled Instance in a VPC.", - "id": "ec2-run-scheduled-instances-1", - "title": "To launch a Scheduled Instance in a VPC" - }, - { - "input": { - "InstanceCount": 1, - "LaunchSpecification": { - "IamInstanceProfile": { - "Name": "my-iam-role" - }, - "ImageId": "ami-12345678", - "InstanceType": "c4.large", - "KeyName": "my-key-pair", - "Placement": { - "AvailabilityZone": "us-west-2b" - }, - "SecurityGroupIds": [ - "sg-12345678" - ] - }, - "ScheduledInstanceId": "sci-1234-1234-1234-1234-123456789012" - }, - "output": { - "InstanceIdSet": [ - "i-1234567890abcdef0" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example launches the specified Scheduled Instance in EC2-Classic.", - "id": "ec2-run-scheduled-instances-2", - "title": "To launch a Scheduled Instance in EC2-Classic" - } - ], - "UnassignPrivateIpAddresses": [ - { - "input": { - "NetworkInterfaceId": "eni-e5aa89a3", - "PrivateIpAddresses": [ - "10.0.0.82" - ] - }, - "comments": { - "input": { - }, - "output": { - } - }, - "description": "This example unassigns the specified private IP address from the specified network interface.", - "id": "ec2-unassign-private-ip-addresses-1", - "title": "To unassign a secondary private IP address from a network interface" - } - ] - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/paginators-1.json deleted file mode 100755 index 9d04d89ab..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/paginators-1.json +++ /dev/null @@ -1,138 +0,0 @@ -{ - "pagination": { - "DescribeAccountAttributes": { - "result_key": "AccountAttributes" - }, - "DescribeAddresses": { - "result_key": "Addresses" - }, - "DescribeAvailabilityZones": { - "result_key": "AvailabilityZones" - }, - "DescribeBundleTasks": { - "result_key": "BundleTasks" - }, - "DescribeConversionTasks": { - "result_key": "ConversionTasks" - }, - "DescribeCustomerGateways": { - "result_key": "CustomerGateways" - }, - "DescribeDhcpOptions": { - "result_key": "DhcpOptions" - }, - "DescribeExportTasks": { - "result_key": "ExportTasks" - }, - "DescribeImages": { - "result_key": "Images" - }, - "DescribeInstanceStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "InstanceStatuses" - }, - "DescribeInstances": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Reservations" - }, - "DescribeInternetGateways": { - "result_key": "InternetGateways" - }, - "DescribeKeyPairs": { - "result_key": "KeyPairs" - }, - "DescribeNetworkAcls": { - "result_key": "NetworkAcls" - }, - "DescribeNetworkInterfaces": { - "result_key": "NetworkInterfaces" - }, - "DescribePlacementGroups": { - "result_key": "PlacementGroups" - }, - "DescribeRegions": { - "result_key": "Regions" - }, - "DescribeReservedInstances": { - "result_key": "ReservedInstances" - }, - "DescribeReservedInstancesListings": { - "result_key": "ReservedInstancesListings" - }, - "DescribeReservedInstancesOfferings": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "ReservedInstancesOfferings" - }, - "DescribeReservedInstancesModifications": { - "input_token": "NextToken", - "output_token": "NextToken", - "result_key": "ReservedInstancesModifications" - }, - "DescribeRouteTables": { - "result_key": "RouteTables" - }, - "DescribeSecurityGroups": { - "result_key": "SecurityGroups" - }, - "DescribeSnapshots": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Snapshots" - }, - "DescribeSpotInstanceRequests": { - "result_key": "SpotInstanceRequests" - }, - "DescribeSpotFleetRequests": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotFleetRequestConfigs" - }, - "DescribeSpotPriceHistory": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "SpotPriceHistory" - }, - "DescribeSubnets": { - "result_key": "Subnets" - }, - "DescribeTags": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Tags" - }, - "DescribeVolumeStatus": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "VolumeStatuses" - }, - "DescribeVolumes": { - "input_token": "NextToken", - "output_token": "NextToken", - "limit_key": "MaxResults", - "result_key": "Volumes" - }, - "DescribeVpcs": { - "result_key": "Vpcs" - }, - "DescribeVpcPeeringConnections": { - "result_key": "VpcPeeringConnections" - }, - "DescribeVpnConnections": { - "result_key": "VpnConnections" - }, - "DescribeVpnGateways": { - "result_key": "VpnGateways" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/waiters-2.json deleted file mode 100755 index ecc9f1b6f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-09-15/waiters-2.json +++ /dev/null @@ -1,593 +0,0 @@ -{ - "version": 2, - "waiters": { - "InstanceExists": { - "delay": 5, - "maxAttempts": 40, - "operation": "DescribeInstances", - "acceptors": [ - { - "matcher": "path", - "expected": true, - "argument": "length(Reservations[]) > `0`", - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "BundleTaskComplete": { - "delay": 15, - "operation": "DescribeBundleTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "complete", - "matcher": "pathAll", - "state": "success", - "argument": "BundleTasks[].State" - }, - { - "expected": "failed", - "matcher": "pathAny", - "state": "failure", - "argument": "BundleTasks[].State" - } - ] - }, - "ConversionTaskCancelled": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskCompleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelled", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - }, - { - "expected": "cancelling", - "matcher": "pathAny", - "state": "failure", - "argument": "ConversionTasks[].State" - } - ] - }, - "ConversionTaskDeleted": { - "delay": 15, - "operation": "DescribeConversionTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "ConversionTasks[].State" - } - ] - }, - "CustomerGatewayAvailable": { - "delay": 15, - "operation": "DescribeCustomerGateways", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "CustomerGateways[].State" - } - ] - }, - "ExportTaskCancelled": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "cancelled", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ExportTaskCompleted": { - "delay": 15, - "operation": "DescribeExportTasks", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "ExportTasks[].State" - } - ] - }, - "ImageExists": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "matcher": "path", - "expected": true, - "argument": "length(Images[]) > `0`", - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidAMIID.NotFound", - "state": "retry" - } - ] - }, - "ImageAvailable": { - "operation": "DescribeImages", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "Images[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "Images[].State", - "expected": "failed" - } - ] - }, - "InstanceRunning": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "running", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "shutting-down", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "InstanceStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].InstanceStatus.Status", - "expected": "ok" - }, - { - "matcher": "error", - "expected": "InvalidInstanceID.NotFound", - "state": "retry" - } - ] - }, - "InstanceStopped": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "stopped", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "terminated", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "InstanceTerminated": { - "delay": 15, - "operation": "DescribeInstances", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "terminated", - "matcher": "pathAll", - "state": "success", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - }, - { - "expected": "stopping", - "matcher": "pathAny", - "state": "failure", - "argument": "Reservations[].Instances[].State.Name" - } - ] - }, - "KeyPairExists": { - "operation": "DescribeKeyPairs", - "delay": 5, - "maxAttempts": 6, - "acceptors": [ - { - "expected": true, - "matcher": "pathAll", - "state": "success", - "argument": "length(KeyPairs[].KeyName) > `0`" - }, - { - "expected": "InvalidKeyPair.NotFound", - "matcher": "error", - "state": "retry" - } - ] - }, - "NatGatewayAvailable": { - "operation": "DescribeNatGateways", - "delay": 15, - "maxAttempts": 40, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "NatGateways[].State", - "expected": "available" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "failed" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "deleting" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "NatGateways[].State", - "expected": "deleted" - }, - { - "state": "retry", - "matcher": "error", - "expected": "NatGatewayNotFound" - } - ] - }, - "NetworkInterfaceAvailable": { - "operation": "DescribeNetworkInterfaces", - "delay": 20, - "maxAttempts": 10, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "NetworkInterfaces[].Status" - }, - { - "expected": "InvalidNetworkInterfaceID.NotFound", - "matcher": "error", - "state": "failure" - } - ] - }, - "PasswordDataAvailable": { - "operation": "GetPasswordData", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "path", - "argument": "length(PasswordData) > `0`", - "expected": true - } - ] - }, - "SnapshotCompleted": { - "delay": 15, - "operation": "DescribeSnapshots", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "completed", - "matcher": "pathAll", - "state": "success", - "argument": "Snapshots[].State" - } - ] - }, - "SpotInstanceRequestFulfilled": { - "operation": "DescribeSpotInstanceRequests", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "fulfilled" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "schedule-expired" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "canceled-before-fulfillment" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "bad-parameters" - }, - { - "state": "failure", - "matcher": "pathAny", - "argument": "SpotInstanceRequests[].Status.Code", - "expected": "system-error" - } - ] - }, - "SubnetAvailable": { - "delay": 15, - "operation": "DescribeSubnets", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Subnets[].State" - } - ] - }, - "SystemStatusOk": { - "operation": "DescribeInstanceStatus", - "maxAttempts": 40, - "delay": 15, - "acceptors": [ - { - "state": "success", - "matcher": "pathAll", - "argument": "InstanceStatuses[].SystemStatus.Status", - "expected": "ok" - } - ] - }, - "VolumeAvailable": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VolumeDeleted": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "matcher": "error", - "expected": "InvalidVolume.NotFound", - "state": "success" - } - ] - }, - "VolumeInUse": { - "delay": 15, - "operation": "DescribeVolumes", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "in-use", - "matcher": "pathAll", - "state": "success", - "argument": "Volumes[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "Volumes[].State" - } - ] - }, - "VpcAvailable": { - "delay": 15, - "operation": "DescribeVpcs", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "Vpcs[].State" - } - ] - }, - "VpcExists": { - "operation": "DescribeVpcs", - "delay": 1, - "maxAttempts": 5, - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidVpcID.NotFound", - "state": "retry" - } - ] - }, - "VpnConnectionAvailable": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "available", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleting", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - }, - { - "expected": "deleted", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpnConnectionDeleted": { - "delay": 15, - "operation": "DescribeVpnConnections", - "maxAttempts": 40, - "acceptors": [ - { - "expected": "deleted", - "matcher": "pathAll", - "state": "success", - "argument": "VpnConnections[].State" - }, - { - "expected": "pending", - "matcher": "pathAny", - "state": "failure", - "argument": "VpnConnections[].State" - } - ] - }, - "VpcPeeringConnectionExists": { - "delay": 15, - "operation": "DescribeVpcPeeringConnections", - "maxAttempts": 40, - "acceptors": [ - { - "matcher": "status", - "expected": 200, - "state": "success" - }, - { - "matcher": "error", - "expected": "InvalidVpcPeeringConnectionID.NotFound", - "state": "retry" - } - ] - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/api-2.json index 952a1ec89..394318850 100755 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/api-2.json @@ -11670,6 +11670,7 @@ "i3.4xlarge", "i3.8xlarge", "i3.16xlarge", + "i3.metal", "hi1.4xlarge", "hs1.8xlarge", "c1.medium", @@ -11690,6 +11691,12 @@ "c5.4xlarge", "c5.9xlarge", "c5.18xlarge", + "c5d.large", + "c5d.xlarge", + "c5d.2xlarge", + "c5d.4xlarge", + "c5d.9xlarge", + "c5d.18xlarge", "cc1.4xlarge", "cc2.8xlarge", "g2.2xlarge", @@ -11716,6 +11723,12 @@ "m5.4xlarge", "m5.12xlarge", "m5.24xlarge", + "m5d.large", + "m5d.xlarge", + "m5d.2xlarge", + "m5d.4xlarge", + "m5d.12xlarge", + "m5d.24xlarge", "h1.2xlarge", "h1.4xlarge", "h1.8xlarge", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/docs-2.json index 6a73dcb5d..f9ab54929 100755 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ec2/2016-11-15/docs-2.json @@ -29,7 +29,7 @@ "CancelImportTask": "

Cancels an in-process import virtual machine or import snapshot task.

", "CancelReservedInstancesListing": "

Cancels the specified Reserved Instance listing in the Reserved Instance Marketplace.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", "CancelSpotFleetRequests": "

Cancels the specified Spot Fleet requests.

After you cancel a Spot Fleet request, the Spot Fleet launches no new Spot Instances. You must specify whether the Spot Fleet should also terminate its Spot Instances. If you terminate the instances, the Spot Fleet request enters the cancelled_terminating state. Otherwise, the Spot Fleet request enters the cancelled_running state and the instances continue to run until they are interrupted or you terminate them manually.

", - "CancelSpotInstanceRequests": "

Cancels one or more Spot Instance requests. Spot Instances are instances that Amazon EC2 starts on your behalf when the maximum price that you specify exceeds the current Spot price. For more information, see Spot Instance Requests in the Amazon EC2 User Guide for Linux Instances.

Canceling a Spot Instance request does not terminate running Spot Instances associated with the request.

", + "CancelSpotInstanceRequests": "

Cancels one or more Spot Instance requests.

Canceling a Spot Instance request does not terminate running Spot Instances associated with the request.

", "ConfirmProductInstance": "

Determines whether a product code is associated with an instance. This action can only be used by the owner of the product code. It is useful when a product code owner must verify whether another user's instance is eligible for support.

", "CopyFpgaImage": "

Copies the specified Amazon FPGA Image (AFI) to the current region.

", "CopyImage": "

Initiates the copy of an AMI from the specified source region to the current region. You specify the destination region by using its endpoint when making the request.

For more information about the prerequisites and limits when copying an AMI, see Copying an AMI in the Amazon Elastic Compute Cloud User Guide.

", @@ -52,7 +52,7 @@ "CreateNetworkAcl": "

Creates a network ACL in a VPC. Network ACLs provide an optional layer of security (in addition to security groups) for the instances in your VPC.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", "CreateNetworkAclEntry": "

Creates an entry (a rule) in a network ACL with the specified rule number. Each network ACL has a set of numbered ingress rules and a separate set of numbered egress rules. When determining whether a packet should be allowed in or out of a subnet associated with the ACL, we process the entries in the ACL according to the rule numbers, in ascending order. Each network ACL has a set of ingress rules and a separate set of egress rules.

We recommend that you leave room between the rule numbers (for example, 100, 110, 120, ...), and not number them one right after the other (for example, 101, 102, 103, ...). This makes it easier to add a rule between existing ones without having to renumber the rules.

After you add an entry, you can't modify it; you must either replace it, or create an entry and delete the old one.

For more information about network ACLs, see Network ACLs in the Amazon Virtual Private Cloud User Guide.

", "CreateNetworkInterface": "

Creates a network interface in the specified subnet.

For more information about network interfaces, see Elastic Network Interfaces in the Amazon Virtual Private Cloud User Guide.

", - "CreateNetworkInterfacePermission": "

Grants an AWS authorized partner account permission to attach the specified network interface to an instance in their account.

You can grant permission to a single AWS account only, and only one account at a time.

", + "CreateNetworkInterfacePermission": "

Grants an AWS-authorized account permission to attach the specified network interface to an instance in their account.

You can grant permission to a single AWS account only, and only one account at a time.

", "CreatePlacementGroup": "

Creates a placement group in which to launch instances. The strategy of the placement group determines how the instances are organized within the group.

A cluster placement group is a logical grouping of instances within a single Availability Zone that benefit from low network latency, high network throughput. A spread placement group places instances on distinct hardware.

For more information, see Placement Groups in the Amazon Elastic Compute Cloud User Guide.

", "CreateReservedInstancesListing": "

Creates a listing for Amazon EC2 Standard Reserved Instances to be sold in the Reserved Instance Marketplace. You can submit one Standard Reserved Instance listing at a time. To get a list of your Standard Reserved Instances, you can use the DescribeReservedInstances operation.

Only Standard Reserved Instances with a capacity reservation can be sold in the Reserved Instance Marketplace. Convertible Reserved Instances and Standard Reserved Instances with a regional benefit cannot be sold.

The Reserved Instance Marketplace matches sellers who want to resell Standard Reserved Instance capacity that they no longer need with buyers who want to purchase additional capacity. Reserved Instances bought and sold through the Reserved Instance Marketplace work like any other Reserved Instances.

To sell your Standard Reserved Instances, you must first register as a seller in the Reserved Instance Marketplace. After completing the registration process, you can create a Reserved Instance Marketplace listing of some or all of your Standard Reserved Instances, and specify the upfront price to receive for them. Your Standard Reserved Instance listings then become available for purchase. To view the details of your Standard Reserved Instance listing, you can use the DescribeReservedInstancesListings operation.

For more information, see Reserved Instance Marketplace in the Amazon Elastic Compute Cloud User Guide.

", "CreateRoute": "

Creates a route in a route table within a VPC.

You must specify one of the following targets: Internet gateway or virtual private gateway, NAT instance, NAT gateway, VPC peering connection, network interface, or egress-only Internet gateway.

When determining how to route traffic, we use the route with the most specific match. For example, traffic is destined for the IPv4 address 192.0.2.3, and the route table includes the following two IPv4 routes:

  • 192.0.2.0/24 (goes to some target A)

  • 192.0.2.0/28 (goes to some target B)

Both routes apply to the traffic destined for 192.0.2.3. However, the second route in the list covers a smaller number of IP addresses and is therefore more specific, so we use that route to determine where to target the traffic.

For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

", @@ -165,7 +165,7 @@ "DescribeSpotFleetInstances": "

Describes the running instances for the specified Spot Fleet.

", "DescribeSpotFleetRequestHistory": "

Describes the events for the specified Spot Fleet request during the specified time.

Spot Fleet events are delayed by up to 30 seconds before they can be described. This ensures that you can query by the last evaluated time and not miss a recorded event.

", "DescribeSpotFleetRequests": "

Describes your Spot Fleet requests.

Spot Fleet requests are deleted 48 hours after they are canceled and their instances are terminated.

", - "DescribeSpotInstanceRequests": "

Describes the Spot Instance requests that belong to your account. Spot Instances are instances that Amazon EC2 launches when the Spot price that you specify exceeds the current Spot price. For more information, see Spot Instance Requests in the Amazon EC2 User Guide for Linux Instances.

You can use DescribeSpotInstanceRequests to find a running Spot Instance by examining the response. If the status of the Spot Instance is fulfilled, the instance ID appears in the response and contains the identifier of the instance. Alternatively, you can use DescribeInstances with a filter to look for instances where the instance lifecycle is spot.

Spot Instance requests are deleted four hours after they are canceled and their instances are terminated.

", + "DescribeSpotInstanceRequests": "

Describes the specified Spot Instance requests.

You can use DescribeSpotInstanceRequests to find a running Spot Instance by examining the response. If the status of the Spot Instance is fulfilled, the instance ID appears in the response and contains the identifier of the instance. Alternatively, you can use DescribeInstances with a filter to look for instances where the instance lifecycle is spot.

Spot Instance requests are deleted four hours after they are canceled and their instances are terminated.

", "DescribeSpotPriceHistory": "

Describes the Spot price history. For more information, see Spot Instance Pricing History in the Amazon EC2 User Guide for Linux Instances.

When you specify a start and end time, this operation returns the prices of the instance types within the time range that you specified and the time when the price changed. The price is valid within the time period that you specified; the response merely indicates the last time that the price changed.

", "DescribeStaleSecurityGroups": "

[EC2-VPC only] Describes the stale security group rules for security groups in a specified VPC. Rules are stale when they reference a deleted security group in a peer VPC, or a security group in a peer VPC for which the VPC peering connection has been deleted.

", "DescribeSubnets": "

Describes one or more of your subnets.

For more information about subnets, see Your VPC and Subnets in the Amazon Virtual Private Cloud User Guide.

", @@ -257,7 +257,7 @@ "ReplaceRouteTableAssociation": "

Changes the route table associated with a given subnet in a VPC. After the operation completes, the subnet uses the routes in the new route table it's associated with. For more information about route tables, see Route Tables in the Amazon Virtual Private Cloud User Guide.

You can also use ReplaceRouteTableAssociation to change which table is the main route table in the VPC. You just specify the main route table's association ID and the route table to be the new main route table.

", "ReportInstanceStatus": "

Submits feedback about the status of an instance. The instance must be in the running state. If your experience with the instance differs from the instance status returned by DescribeInstanceStatus, use ReportInstanceStatus to report your experience with the instance. Amazon EC2 collects this information to improve the accuracy of status checks.

Use of this action does not change the value returned by DescribeInstanceStatus.

", "RequestSpotFleet": "

Creates a Spot Fleet request.

The Spot Fleet request specifies the total target capacity and the On-Demand target capacity. Amazon EC2 calculates the difference between the total capacity and On-Demand capacity, and launches the difference as Spot capacity.

You can submit a single request that includes multiple launch specifications that vary by instance type, AMI, Availability Zone, or subnet.

By default, the Spot Fleet requests Spot Instances in the Spot pool where the price per unit is the lowest. Each launch specification can include its own instance weighting that reflects the value of the instance type to your application workload.

Alternatively, you can specify that the Spot Fleet distribute the target capacity across the Spot pools included in its launch specifications. By ensuring that the Spot Instances in your Spot Fleet are in different Spot pools, you can improve the availability of your fleet.

You can specify tags for the Spot Instances. You cannot tag other resource types in a Spot Fleet request because only the instance resource type is supported.

For more information, see Spot Fleet Requests in the Amazon EC2 User Guide for Linux Instances.

", - "RequestSpotInstances": "

Creates a Spot Instance request. Spot Instances are instances that Amazon EC2 launches when the maximum price that you specify exceeds the current Spot price. For more information, see Spot Instance Requests in the Amazon EC2 User Guide for Linux Instances.

", + "RequestSpotInstances": "

Creates a Spot Instance request.

For more information, see Spot Instance Requests in the Amazon EC2 User Guide for Linux Instances.

", "ResetFpgaImageAttribute": "

Resets the specified attribute of the specified Amazon FPGA Image (AFI) to its default value. You can only reset the load permission attribute.

", "ResetImageAttribute": "

Resets an attribute of an AMI to its default value.

The productCodes attribute can't be reset.

", "ResetInstanceAttribute": "

Resets an attribute of an instance to its default value. To reset the kernel or ramdisk, the instance must be in a stopped state. To reset the sourceDestCheck, the instance can be either running or stopped.

The sourceDestCheck attribute controls whether source/destination checking is enabled. The default value is true, which means checking is enabled. This value must be false for a NAT instance to perform NAT. For more information, see NAT Instances in the Amazon Virtual Private Cloud User Guide.

", @@ -4591,7 +4591,7 @@ "DescribeImportSnapshotTasksRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value.

", "DescribeInstanceCreditSpecificationsRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter in the same call.

", "DescribeInstanceStatusRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter in the same call.

", - "DescribeInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter or tag filters in the same call.

", + "DescribeInstancesRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000. You cannot specify this parameter and the instance IDs parameter in the same call.

", "DescribeLaunchTemplateVersionsRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000.

", "DescribeLaunchTemplatesRequest$MaxResults": "

The maximum number of results to return in a single call. To retrieve the remaining results, make another call with the returned NextToken value. This value can be between 5 and 1000.

", "DescribeMovingAddressesRequest$MaxResults": "

The maximum number of results to return for the request in a single page. The remaining results of the initial request can be seen by sending another request with the returned NextToken value. This value can be between 5 and 1000; if MaxResults is given a value outside of this range, an error is returned.

Default: If no value is provided, the default is 1000.

", @@ -6464,7 +6464,7 @@ "LaunchTemplateTagSpecificationRequest$ResourceType": "

The type of resource to tag. Currently, the resource types that support tagging on creation are instance and volume.

", "SpotFleetTagSpecification$ResourceType": "

The type of resource. Currently, the only resource type that is supported is instance.

", "TagDescription$ResourceType": "

The resource type.

", - "TagSpecification$ResourceType": "

The type of resource to tag. Currently, the resource types that support tagging on creation are instance and volume.

" + "TagSpecification$ResourceType": "

The type of resource to tag. Currently, the resource types that support tagging on creation are instance, snapshot, and volume.

" } }, "ResponseError": { @@ -8267,7 +8267,7 @@ "StaleSecurityGroup$VpcId": "

The ID of the VPC for the security group.

", "StartInstancesRequest$AdditionalInfo": "

Reserved.

", "StateReason$Code": "

The reason code for the state change.

", - "StateReason$Message": "

The message for the state change.

  • Server.InsufficientInstanceCapacity: There was insufficient instance capacity to satisfy the launch request.

  • Server.InternalError: An internal error occurred during instance launch, resulting in termination.

  • Server.ScheduledStop: The instance was stopped due to a scheduled retirement.

  • Server.SpotInstanceTermination: A Spot Instance was terminated due to an increase in the Spot price.

  • Client.InternalError: A client error caused the instance to terminate on launch.

  • Client.InstanceInitiatedShutdown: The instance was shut down using the shutdown -h command from the instance.

  • Client.InstanceTerminated: The instance was terminated or rebooted during AMI creation.

  • Client.UserInitiatedShutdown: The instance was shut down using the Amazon EC2 API.

  • Client.VolumeLimitExceeded: The limit on the number of EBS volumes or total storage was exceeded. Decrease usage or request an increase in your limits.

  • Client.InvalidSnapshot.NotFound: The specified snapshot was not found.

", + "StateReason$Message": "

The message for the state change.

  • Server.InsufficientInstanceCapacity: There was insufficient capacity available to satisfy the launch request.

  • Server.InternalError: An internal error caused the instance to terminate during launch.

  • Server.ScheduledStop: The instance was stopped due to a scheduled retirement.

  • Server.SpotInstanceShutdown: The instance was stopped because the number of Spot requests with a maximum price equal to or higher than the Spot price exceeded available capacity or because of an increase in the Spot price.

  • Server.SpotInstanceTermination: The instance was terminated because the number of Spot requests with a maximum price equal to or higher than the Spot price exceeded available capacity or because of an increase in the Spot price.

  • Client.InstanceInitiatedShutdown: The instance was shut down using the shutdown -h command from the instance.

  • Client.InstanceTerminated: The instance was terminated or rebooted during AMI creation.

  • Client.InternalError: A client error caused the instance to terminate during launch.

  • Client.InvalidSnapshot.NotFound: The specified snapshot was not found.

  • Client.UserInitiatedShutdown: The instance was shut down using the Amazon EC2 API.

  • Client.VolumeLimitExceeded: The limit on the number of EBS volumes or total storage was exceeded. Decrease usage or request an increase in your account limits.

", "StorageLocation$Bucket": "

The name of the S3 bucket.

", "StorageLocation$Key": "

The key.

", "Subnet$AvailabilityZone": "

The Availability Zone of the subnet.

", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/api-2.json index fd42d019b..92a33d3e4 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/api-2.json @@ -790,8 +790,7 @@ "type":"structure", "required":[ "serviceName", - "taskDefinition", - "desiredCount" + "taskDefinition" ], "members":{ "cluster":{"shape":"String"}, @@ -808,7 +807,8 @@ "placementConstraints":{"shape":"PlacementConstraints"}, "placementStrategy":{"shape":"PlacementStrategies"}, "networkConfiguration":{"shape":"NetworkConfiguration"}, - "healthCheckGracePeriodSeconds":{"shape":"BoxedInteger"} + "healthCheckGracePeriodSeconds":{"shape":"BoxedInteger"}, + "schedulingStrategy":{"shape":"SchedulingStrategy"} } }, "CreateServiceResponse":{ @@ -849,7 +849,8 @@ "required":["service"], "members":{ "cluster":{"shape":"String"}, - "service":{"shape":"String"} + "service":{"shape":"String"}, + "force":{"shape":"BoxedBoolean"} } }, "DeleteServiceResponse":{ @@ -1187,7 +1188,8 @@ "cluster":{"shape":"String"}, "nextToken":{"shape":"String"}, "maxResults":{"shape":"BoxedInteger"}, - "launchType":{"shape":"LaunchType"} + "launchType":{"shape":"LaunchType"}, + "schedulingStrategy":{"shape":"SchedulingStrategy"} } }, "ListServicesResponse":{ @@ -1515,6 +1517,13 @@ "failures":{"shape":"Failures"} } }, + "SchedulingStrategy":{ + "type":"string", + "enum":[ + "REPLICA", + "DAEMON" + ] + }, "ServerException":{ "type":"structure", "members":{ @@ -1546,7 +1555,8 @@ "placementConstraints":{"shape":"PlacementConstraints"}, "placementStrategy":{"shape":"PlacementStrategies"}, "networkConfiguration":{"shape":"NetworkConfiguration"}, - "healthCheckGracePeriodSeconds":{"shape":"BoxedInteger"} + "healthCheckGracePeriodSeconds":{"shape":"BoxedInteger"}, + "schedulingStrategy":{"shape":"SchedulingStrategy"} } }, "ServiceEvent":{ @@ -1581,7 +1591,9 @@ "type":"structure", "members":{ "registryArn":{"shape":"String"}, - "port":{"shape":"BoxedInteger"} + "port":{"shape":"BoxedInteger"}, + "containerName":{"shape":"String"}, + "containerPort":{"shape":"BoxedInteger"} } }, "Services":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/docs-2.json index 4992b008f..9bf18a18e 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ecs/2014-11-13/docs-2.json @@ -3,7 +3,7 @@ "service": "

Amazon Elastic Container Service (Amazon ECS) is a highly scalable, fast, container management service that makes it easy to run, stop, and manage Docker containers on a cluster. You can host your cluster on a serverless infrastructure that is managed by Amazon ECS by launching your services or tasks using the Fargate launch type. For more control, you can host your tasks on a cluster of Amazon Elastic Compute Cloud (Amazon EC2) instances that you manage by using the EC2 launch type. For more information about launch types, see Amazon ECS Launch Types.

Amazon ECS lets you launch and stop container-based applications with simple API calls, allows you to get the state of your cluster from a centralized service, and gives you access to many familiar Amazon EC2 features.

You can use Amazon ECS to schedule the placement of containers across your cluster based on your resource needs, isolation policies, and availability requirements. Amazon ECS eliminates the need for you to operate your own cluster management and configuration management systems or worry about scaling your management infrastructure.

", "operations": { "CreateCluster": "

Creates a new Amazon ECS cluster. By default, your account receives a default cluster when you launch your first container instance. However, you can create your own cluster with a unique name with the CreateCluster action.

When you call the CreateCluster API operation, Amazon ECS attempts to create the service-linked role for your account so that required resources in other AWS services can be managed on your behalf. However, if the IAM user that makes the call does not have permissions to create the service-linked role, it is not created. For more information, see Using Service-Linked Roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide.

", - "CreateService": "

Runs and maintains a desired number of tasks from a specified task definition. If the number of tasks running in a service drops below desiredCount, Amazon ECS spawns another copy of the task in the specified cluster. To update an existing service, see UpdateService.

In addition to maintaining the desired count of tasks in your service, you can optionally run your service behind a load balancer. The load balancer distributes traffic across the tasks that are associated with the service. For more information, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide.

You can optionally specify a deployment configuration for your service. During a deployment, the service scheduler uses the minimumHealthyPercent and maximumPercent parameters to determine the deployment strategy. The deployment is triggered by changing the task definition or the desired count of a service with an UpdateService operation.

The minimumHealthyPercent represents a lower limit on the number of your service's tasks that must remain in the RUNNING state during a deployment, as a percentage of the desiredCount (rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has a desiredCount of four tasks and a minimumHealthyPercent of 50%, the scheduler can stop two existing tasks to free up cluster capacity before starting two new tasks. Tasks for services that do not use a load balancer are considered healthy if they are in the RUNNING state. Tasks for services that do use a load balancer are considered healthy if they are in the RUNNING state and the container instance they are hosted on is reported as healthy by the load balancer. The default value for minimumHealthyPercent is 50% in the console and 100% for the AWS CLI, the AWS SDKs, and the APIs.

The maximumPercent parameter represents an upper limit on the number of your service's tasks that are allowed in the RUNNING or PENDING state during a deployment, as a percentage of the desiredCount (rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service has a desiredCount of four tasks and a maximumPercent value of 200%, the scheduler can start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The default value for maximumPercent is 200%.

When the service scheduler launches new tasks, it determines task placement in your cluster using the following logic:

  • Determine which of the container instances in your cluster can support your service's task definition (for example, they have the required CPU, memory, ports, and container instance attributes).

  • By default, the service scheduler attempts to balance tasks across Availability Zones in this manner (although you can choose a different placement strategy) with the placementStrategy parameter):

    • Sort the valid container instances by the fewest number of running tasks for this service in the same Availability Zone as the instance. For example, if zone A has one running service task and zones B and C each have zero, valid container instances in either zone B or C are considered optimal for placement.

    • Place the new service task on a valid container instance in an optimal Availability Zone (based on the previous steps), favoring container instances with the fewest number of running tasks for this service.

", + "CreateService": "

Runs and maintains a desired number of tasks from a specified task definition. If the number of tasks running in a service drops below desiredCount, Amazon ECS spawns another copy of the task in the specified cluster. To update an existing service, see UpdateService.

In addition to maintaining the desired count of tasks in your service, you can optionally run your service behind a load balancer. The load balancer distributes traffic across the tasks that are associated with the service. For more information, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide.

You can optionally specify a deployment configuration for your service. During a deployment, the service scheduler uses the minimumHealthyPercent and maximumPercent parameters to determine the deployment strategy. The deployment is triggered by changing the task definition or the desired count of a service with an UpdateService operation.

The minimumHealthyPercent represents a lower limit on the number of your service's tasks that must remain in the RUNNING state during a deployment, as a percentage of the desiredCount (rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has a desiredCount of four tasks and a minimumHealthyPercent of 50%, the scheduler can stop two existing tasks to free up cluster capacity before starting two new tasks. Tasks for services that do not use a load balancer are considered healthy if they are in the RUNNING state. Tasks for services that do use a load balancer are considered healthy if they are in the RUNNING state and the container instance they are hosted on is reported as healthy by the load balancer. The default value for a replica service for minimumHealthyPercent is 50% in the console and 100% for the AWS CLI, the AWS SDKs, and the APIs. The default value for a daemon service for minimumHealthyPercent is 0% for the AWS CLI, the AWS SDKs, and the APIs and 50% for the console.

The maximumPercent parameter represents an upper limit on the number of your service's tasks that are allowed in the RUNNING or PENDING state during a deployment, as a percentage of the desiredCount (rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your replica service has a desiredCount of four tasks and a maximumPercent value of 200%, the scheduler can start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The default value for a replica service for maximumPercent is 200%. If you are using a daemon service type, the maximumPercent should remain at 100%, which is the default value.

When the service scheduler launches new tasks, it determines task placement in your cluster using the following logic:

  • Determine which of the container instances in your cluster can support your service's task definition (for example, they have the required CPU, memory, ports, and container instance attributes).

  • By default, the service scheduler attempts to balance tasks across Availability Zones in this manner (although you can choose a different placement strategy) with the placementStrategy parameter):

    • Sort the valid container instances, giving priority to instances that have the fewest number of running tasks for this service in their respective Availability Zone. For example, if zone A has one running service task and zones B and C each have zero, valid container instances in either zone B or C are considered optimal for placement.

    • Place the new service task on a valid container instance in an optimal Availability Zone (based on the previous steps), favoring container instances with the fewest number of running tasks for this service.

", "DeleteAttributes": "

Deletes one or more custom attributes from an Amazon ECS resource.

", "DeleteCluster": "

Deletes the specified cluster. You must deregister all container instances from this cluster before you may delete it. You can list the container instances in a cluster with ListContainerInstances and deregister them with DeregisterContainerInstance.

", "DeleteService": "

Deletes a specified service within a cluster. You can delete a service if you have no running tasks in it and the desired task count is zero. If the service is actively maintaining tasks, you cannot delete it, and you must update the service to a desired task count of zero. For more information, see UpdateService.

When you delete a service, if there are still running tasks that require cleanup, the service status moves from ACTIVE to DRAINING, and the service is no longer visible in the console or in ListServices API operations. After the tasks have stopped, then the service status moves from DRAINING to INACTIVE. Services in the DRAINING or INACTIVE status can still be viewed with DescribeServices API operations. However, in the future, INACTIVE services may be cleaned up and purged from Amazon ECS record keeping, and DescribeServices API operations on those services return a ServiceNotFoundException error.

", @@ -132,6 +132,7 @@ "ContainerDefinition$disableNetworking": "

When this parameter is true, networking is disabled within the container. This parameter maps to NetworkDisabled in the Create a container section of the Docker Remote API.

This parameter is not supported for Windows containers.

", "ContainerDefinition$privileged": "

When this parameter is true, the container is given elevated privileges on the host container instance (similar to the root user). This parameter maps to Privileged in the Create a container section of the Docker Remote API and the --privileged option to docker run.

This parameter is not supported for Windows containers or tasks using the Fargate launch type.

", "ContainerDefinition$readonlyRootFilesystem": "

When this parameter is true, the container is given read-only access to its root file system. This parameter maps to ReadonlyRootfs in the Create a container section of the Docker Remote API and the --read-only option to docker run.

This parameter is not supported for Windows containers.

", + "DeleteServiceRequest$force": "

If true, allows you to delete a service even if it has not been scaled down to zero tasks. It is only necessary to use this if the service is using the REPLICA scheduling strategy.

", "DeregisterContainerInstanceRequest$force": "

Forces the deregistration of the container instance. If you have tasks running on the container instance when you deregister it with the force option, these tasks remain running until you terminate the instance or the tasks stop through some other means, but they are orphaned (no longer monitored or accounted for by Amazon ECS). If an orphaned task on your container instance is part of an Amazon ECS service, then the service scheduler starts another copy of that task, on a different container instance if possible.

Any containers in orphaned service tasks that are registered with a Classic Load Balancer or an Application Load Balancer target group are deregistered. They begin connection draining according to the settings on the load balancer or target group.

", "LinuxParameters$initProcessEnabled": "

Run an init process inside the container that forwards signals and reaps processes. This parameter maps to the --init option to docker run. This parameter requires version 1.25 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command: sudo docker version | grep \"Server API version\"

", "MountPoint$readOnly": "

If this value is true, the container has read-only access to the volume. If this value is false, then the container can write to the volume. The default value is false.

", @@ -156,7 +157,7 @@ "HealthCheck$timeout": "

The time period in seconds to wait for a health check to succeed before it is considered a failure. You may specify between 2 and 60 seconds. The default value is 5 seconds.

", "HealthCheck$retries": "

The number of times to retry a failed health check before the container is considered unhealthy. You may specify between 1 and 10 retries. The default value is 3 retries.

", "HealthCheck$startPeriod": "

The optional grace period within which to provide containers time to bootstrap before failed health checks count towards the maximum number of retries. You may specify between 0 and 300 seconds. The startPeriod is disabled by default.

If a health check succeeds within the startPeriod, then the container is considered healthy and any subsequent failures count toward the maximum number of retries.

", - "LinuxParameters$sharedMemorySize": "

The value for the size of the /dev/shm volume. This parameter maps to the --shm-size option to docker run.

", + "LinuxParameters$sharedMemorySize": "

The value for the size (in MiB) of the /dev/shm volume. This parameter maps to the --shm-size option to docker run.

If you are using tasks that use the Fargate launch type, the sharedMemorySize parameter is not supported.

", "ListAttributesRequest$maxResults": "

The maximum number of cluster results returned by ListAttributes in paginated output. When this parameter is used, ListAttributes only returns maxResults results in a single page along with a nextToken response element. The remaining results of the initial request can be seen by sending another ListAttributes request with the returned nextToken value. This value can be between 1 and 100. If this parameter is not used, then ListAttributes returns up to 100 results and a nextToken value if applicable.

", "ListClustersRequest$maxResults": "

The maximum number of cluster results returned by ListClusters in paginated output. When this parameter is used, ListClusters only returns maxResults results in a single page along with a nextToken response element. The remaining results of the initial request can be seen by sending another ListClusters request with the returned nextToken value. This value can be between 1 and 100. If this parameter is not used, then ListClusters returns up to 100 results and a nextToken value if applicable.

", "ListContainerInstancesRequest$maxResults": "

The maximum number of container instance results returned by ListContainerInstances in paginated output. When this parameter is used, ListContainerInstances only returns maxResults results in a single page along with a nextToken response element. The remaining results of the initial request can be seen by sending another ListContainerInstances request with the returned nextToken value. This value can be between 1 and 100. If this parameter is not used, then ListContainerInstances returns up to 100 results and a nextToken value if applicable.

", @@ -171,7 +172,8 @@ "PortMapping$hostPort": "

The port number on the container instance to reserve for your container.

If using containers in a task with the awsvpc or host network mode, the hostPort can either be left blank or set to the same value as the containerPort.

If using containers in a task with the bridge network mode, you can specify a non-reserved host port for your container port mapping, or you can omit the hostPort (or set it to 0) while specifying a containerPort and your container automatically receives a port in the ephemeral port range for your container instance operating system and Docker version.

The default ephemeral port range for Docker version 1.6.0 and later is listed on the instance under /proc/sys/net/ipv4/ip_local_port_range; if this kernel parameter is unavailable, the default ephemeral port range from 49153 through 65535 is used. You should not attempt to specify a host port in the ephemeral port range as these are reserved for automatic assignment. In general, ports below 32768 are outside of the ephemeral port range.

The default ephemeral port range from 49153 through 65535 is always used for Docker versions before 1.6.0.

The default reserved ports are 22 for SSH, the Docker ports 2375 and 2376, and the Amazon ECS container agent ports 51678 and 51679. Any host port that was previously specified in a running task is also reserved while the task is running (after a task stops, the host port is released). The current reserved ports are displayed in the remainingResources of DescribeContainerInstances output, and a container instance may have up to 100 reserved ports at a time, including the default reserved ports (automatically assigned ports do not count toward the 100 reserved ports limit).

", "RunTaskRequest$count": "

The number of instantiations of the specified task to place on your cluster. You can specify up to 10 tasks per call.

", "Service$healthCheckGracePeriodSeconds": "

The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started.

", - "ServiceRegistry$port": "

The port value used if your Service Discovery service specified an SRV record.

", + "ServiceRegistry$port": "

The port value used if your service discovery service specified an SRV record. This field is required if both the awsvpc network mode and SRV records are used.

", + "ServiceRegistry$containerPort": "

The port value, already specified in the task definition, to be used for your service discovery service. If the task definition your service task specifies uses the bridge or host network mode, you must specify a containerName and containerPort combination from the task definition. If the task definition your service task specifies uses the awsvpc network mode and a type SRV DNS record is used, you must specify either a containerName and containerPort combination or a port value, but not both.

", "SubmitContainerStateChangeRequest$exitCode": "

The exit code returned for the state change request.

", "UpdateServiceRequest$desiredCount": "

The number of instantiations of the task to place and keep running in your service.

", "UpdateServiceRequest$healthCheckGracePeriodSeconds": "

The period of time, in seconds, that the Amazon ECS service scheduler should ignore unhealthy Elastic Load Balancing target health checks after a task has first started. This is only valid if your service is configured to use a load balancer. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 1,800 seconds during which the ECS service scheduler ignores the Elastic Load Balancing health check status. This grace period can prevent the ECS service scheduler from marking tasks as unhealthy and stopping them before they have time to come up.

" @@ -586,7 +588,7 @@ "Service$runningCount": "

The number of tasks in the cluster that are in the RUNNING state.

", "Service$pendingCount": "

The number of tasks in the cluster that are in the PENDING state.

", "TaskDefinition$revision": "

The revision of the task in a particular family. The revision is a version number of a task definition in a family. When you register a task definition for the first time, the revision is 1; each time you register a new revision of a task definition in the same family, the revision value always increases by one (even if you have deregistered previous revisions in this family).

", - "Tmpfs$size": "

The size of the tmpfs volume.

", + "Tmpfs$size": "

The size (in MiB) of the tmpfs volume.

", "Ulimit$softLimit": "

The soft limit for the ulimit type.

", "Ulimit$hardLimit": "

The hard limit for the ulimit type.

" } @@ -699,7 +701,7 @@ } }, "LoadBalancer": { - "base": "

Details on a load balancer that is used with a service.

", + "base": "

Details on a load balancer that is used with a service.

Services with tasks that use the awsvpc network mode (for example, those with the Fargate launch type) only support Application Load Balancers and Network Load Balancers; Classic Load Balancers are not supported. Also, when you create any target groups for these services, you must choose ip as the target type, not instance, because tasks that use the awsvpc network mode are associated with an elastic network interface, not an Amazon EC2 instance.

", "refs": { "LoadBalancers$member": null } @@ -707,8 +709,8 @@ "LoadBalancers": { "base": null, "refs": { - "CreateServiceRequest$loadBalancers": "

A load balancer object representing the load balancer to use with your service. Currently, you are limited to one load balancer or target group per service. After you create a service, the load balancer name or target group ARN, container name, and container port specified in the service definition are immutable.

For Classic Load Balancers, this object must contain the load balancer name, the container name (as it appears in a container definition), and the container port to access from the load balancer. When a task from this service is placed on a container instance, the container instance is registered with the load balancer specified here.

For Application Load Balancers and Network Load Balancers, this object must contain the load balancer target group ARN, the container name (as it appears in a container definition), and the container port to access from the load balancer. When a task from this service is placed on a container instance, the container instance and port combination is registered as a target in the target group specified here.

", - "Service$loadBalancers": "

A list of Elastic Load Balancing load balancer objects, containing the load balancer name, the container name (as it appears in a container definition), and the container port to access from the load balancer.

" + "CreateServiceRequest$loadBalancers": "

A load balancer object representing the load balancer to use with your service. Currently, you are limited to one load balancer or target group per service. After you create a service, the load balancer name or target group ARN, container name, and container port specified in the service definition are immutable.

For Classic Load Balancers, this object must contain the load balancer name, the container name (as it appears in a container definition), and the container port to access from the load balancer. When a task from this service is placed on a container instance, the container instance is registered with the load balancer specified here.

For Application Load Balancers and Network Load Balancers, this object must contain the load balancer target group ARN, the container name (as it appears in a container definition), and the container port to access from the load balancer. When a task from this service is placed on a container instance, the container instance and port combination is registered as a target in the target group specified here.

Services with tasks that use the awsvpc network mode (for example, those with the Fargate launch type) only support Application Load Balancers and Network Load Balancers; Classic Load Balancers are not supported. Also, when you create any target groups for these services, you must choose ip as the target type, not instance, because tasks that use the awsvpc network mode are associated with an elastic network interface, not an Amazon EC2 instance.

", + "Service$loadBalancers": "

A list of Elastic Load Balancing load balancer objects, containing the load balancer name, the container name (as it appears in a container definition), and the container port to access from the load balancer.

Services with tasks that use the awsvpc network mode (for example, those with the Fargate launch type) only support Application Load Balancers and Network Load Balancers; Classic Load Balancers are not supported. Also, when you create any target groups for these services, you must choose ip as the target type, not instance, because tasks that use the awsvpc network mode are associated with an elastic network interface, not an Amazon EC2 instance.

" } }, "LogConfiguration": { @@ -910,7 +912,7 @@ "Resources": { "base": null, "refs": { - "ContainerInstance$remainingResources": "

For CPU and memory resource types, this parameter describes the remaining CPU and memory on the that has not already been allocated to tasks (and is therefore available for new tasks). For port resource types, this parameter describes the ports that were reserved by the Amazon ECS container agent (at instance registration time) and any task containers that have reserved port mappings on the host (with the host or bridge network mode). Any port that is not specified here is available for new tasks.

", + "ContainerInstance$remainingResources": "

For CPU and memory resource types, this parameter describes the remaining CPU and memory that has not already been allocated to tasks and is therefore available for new tasks. For port resource types, this parameter describes the ports that were reserved by the Amazon ECS container agent (at instance registration time) and any task containers that have reserved port mappings on the host (with the host or bridge network mode). Any port that is not specified here is available for new tasks.

", "ContainerInstance$registeredResources": "

For CPU and memory resource types, this parameter describes the amount of each resource that was available on the container instance when the container agent registered it with Amazon ECS; this value represents the total amount of CPU and memory that can be allocated on this container instance to tasks. For port resource types, this parameter describes the ports that were reserved by the Amazon ECS container agent when it registered the container instance with Amazon ECS.

", "RegisterContainerInstanceRequest$totalResources": "

The resources available on the instance.

" } @@ -925,6 +927,14 @@ "refs": { } }, + "SchedulingStrategy": { + "base": null, + "refs": { + "CreateServiceRequest$schedulingStrategy": "

The scheduling strategy to use for the service. For more information, see Services.

There are two service scheduler strategies available:

  • REPLICA-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions.

  • DAEMON-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. When using this strategy, there is no need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies.

    Fargate tasks do not support the DAEMON scheduling strategy.

", + "ListServicesRequest$schedulingStrategy": "

The scheduling strategy for services to list.

", + "Service$schedulingStrategy": "

The scheduling strategy to use for the service. For more information, see Services.

There are two service scheduler strategies available:

  • REPLICA-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions.

  • DAEMON-The daemon scheduling strategy deploys exactly one task on each container instance in your cluster. When using this strategy, do not specify a desired number of tasks or any task placement strategies.

    Fargate tasks do not support the DAEMON scheduling strategy.

" + } + }, "ServerException": { "base": "

These errors are usually caused by a server issue.

", "refs": { @@ -964,7 +974,7 @@ "ServiceRegistries": { "base": null, "refs": { - "CreateServiceRequest$serviceRegistries": "

The details of the service discovery registries you want to assign to this service. For more information, see Service Discovery.

", + "CreateServiceRequest$serviceRegistries": "

The details of the service discovery registries you want to assign to this service. For more information, see Service Discovery.

Service discovery is supported for Fargate tasks if using platform version v1.1.0 or later. For more information, see AWS Fargate Platform Versions.

", "Service$serviceRegistries": "

" } }, @@ -1034,7 +1044,7 @@ "Container$reason": "

A short (255 max characters) human-readable string to provide additional details about a running or stopped container.

", "ContainerDefinition$name": "

The name of a container. If you are linking multiple containers together in a task definition, the name of one container can be entered in the links of another container to connect the containers. Up to 255 letters (uppercase and lowercase), numbers, hyphens, and underscores are allowed. This parameter maps to name in the Create a container section of the Docker Remote API and the --name option to docker run.

", "ContainerDefinition$image": "

The image used to start a container. This string is passed directly to the Docker daemon. Images in the Docker Hub registry are available by default. Other repositories are specified with either repository-url/image:tag or repository-url/image@digest . Up to 255 letters (uppercase and lowercase), numbers, hyphens, underscores, colons, periods, forward slashes, and number signs are allowed. This parameter maps to Image in the Create a container section of the Docker Remote API and the IMAGE parameter of docker run.

  • When a new task starts, the Amazon ECS container agent pulls the latest version of the specified image and tag for the container to use. However, subsequent updates to a repository image are not propagated to already running tasks.

  • Images in Amazon ECR repositories can be specified by either using the full registry/repository:tag or registry/repository@digest. For example, 012345678910.dkr.ecr.<region-name>.amazonaws.com/<repository-name>:latest or 012345678910.dkr.ecr.<region-name>.amazonaws.com/<repository-name>@sha256:94afd1f2e64d908bc90dbca0035a5b567EXAMPLE.

  • Images in official repositories on Docker Hub use a single name (for example, ubuntu or mongo).

  • Images in other repositories on Docker Hub are qualified with an organization name (for example, amazon/amazon-ecs-agent).

  • Images in other online repositories are qualified further by a domain name (for example, quay.io/assemblyline/ubuntu).

", - "ContainerDefinition$hostname": "

The hostname to use for your container. This parameter maps to Hostname in the Create a container section of the Docker Remote API and the --hostname option to docker run.

", + "ContainerDefinition$hostname": "

The hostname to use for your container. This parameter maps to Hostname in the Create a container section of the Docker Remote API and the --hostname option to docker run.

The hostname parameter is not supported if using the awsvpc networkMode.

", "ContainerDefinition$user": "

The user name to use inside the container. This parameter maps to User in the Create a container section of the Docker Remote API and the --user option to docker run.

This parameter is not supported for Windows containers.

", "ContainerDefinition$workingDirectory": "

The working directory in which to run commands inside the container. This parameter maps to WorkingDir in the Create a container section of the Docker Remote API and the --workdir option to docker run.

", "ContainerInstance$containerInstanceArn": "

The Amazon Resource Name (ARN) of the container instance. The ARN contains the arn:aws:ecs namespace, followed by the region of the container instance, the AWS account ID of the container instance owner, the container-instance namespace, and then the container instance ID. For example, arn:aws:ecs:region:aws_account_id:container-instance/container_instance_ID .

", @@ -1048,7 +1058,7 @@ "CreateServiceRequest$cluster": "

The short name or full Amazon Resource Name (ARN) of the cluster on which to run your service. If you do not specify a cluster, the default cluster is assumed.

", "CreateServiceRequest$serviceName": "

The name of your service. Up to 255 letters (uppercase and lowercase), numbers, hyphens, and underscores are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a region or across multiple regions.

", "CreateServiceRequest$taskDefinition": "

The family and revision (family:revision) or full ARN of the task definition to run in your service. If a revision is not specified, the latest ACTIVE revision is used.

", - "CreateServiceRequest$clientToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request. Up to 32 ASCII characters are allowed.

", + "CreateServiceRequest$clientToken": "

Unique, case-sensitive identifier that you provide to ensure the idempotency of the request. Up to 32 ASCII characters are allowed.

", "CreateServiceRequest$platformVersion": "

The platform version on which to run your service. If one is not specified, the latest version is used by default.

", "CreateServiceRequest$role": "

The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition does not use the awsvpc network mode. If you specify the role parameter, you must also specify a load balancer object with the loadBalancers parameter.

If your account has already created the Amazon ECS service-linked role, that role is used by default for your service unless you specify a role here. The service-linked role is required if your task definition uses the awsvpc network mode, in which case you should not specify a role here. For more information, see Using Service-Linked Roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide.

If your specified role has a path other than /, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the name bar has a path of /foo/ then you would specify /foo/bar as the role name. For more information, see Friendly Names and Paths in the IAM User Guide.

", "DeleteAttributesRequest$cluster": "

The short name or full Amazon Resource Name (ARN) of the cluster that contains the resource to delete attributes. If you do not specify a cluster, the default cluster is assumed.

", @@ -1108,7 +1118,7 @@ "ListTasksRequest$startedBy": "

The startedBy value with which to filter the task results. Specifying a startedBy value limits the results to tasks that were started with that value.

", "ListTasksRequest$serviceName": "

The name of the service with which to filter the ListTasks results. Specifying a serviceName limits the results to tasks that belong to that service.

", "ListTasksResponse$nextToken": "

The nextToken value to include in a future ListTasks request. When the results of a ListTasks request exceed maxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", - "LoadBalancer$targetGroupArn": "

The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group associated with a service.

", + "LoadBalancer$targetGroupArn": "

The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group associated with a service.

If your service's task definition uses the awsvpc network mode (which is required for the Fargate launch type), you must choose ip as the target type, not instance, because tasks that use the awsvpc network mode are associated with an elastic network interface, not an Amazon EC2 instance.

", "LoadBalancer$loadBalancerName": "

The name of a load balancer.

", "LoadBalancer$containerName": "

The name of the container (as it appears in a container definition) to associate with the load balancer.

", "LogConfigurationOptionsMap$key": null, @@ -1131,7 +1141,7 @@ "RegisterTaskDefinitionRequest$executionRoleArn": "

The Amazon Resource Name (ARN) of the task execution role that the Amazon ECS container agent and the Docker daemon can assume.

", "RegisterTaskDefinitionRequest$cpu": "

The number of CPU units used by the task. It can be expressed as an integer using CPU units, for example 1024, or as a string using vCPUs, for example 1 vCPU or 1 vcpu, in a task definition but will be converted to an integer indicating the CPU units when the task definition is registered.

Task-level CPU and memory parameters are ignored for Windows containers. We recommend specifying container-level resources for Windows containers.

If using the EC2 launch type, this field is optional. Supported values are between 128 CPU units (0.125 vCPUs) and 10240 CPU units (10 vCPUs).

If using the Fargate launch type, this field is required and you must use one of the following values, which determines your range of supported values for the memory parameter:

  • 256 (.25 vCPU) - Available memory values: 512 (0.5 GB), 1024 (1 GB), 2048 (2 GB)

  • 512 (.5 vCPU) - Available memory values: 1024 (1 GB), 2048 (2 GB), 3072 (3 GB), 4096 (4 GB)

  • 1024 (1 vCPU) - Available memory values: 2048 (2 GB), 3072 (3 GB), 4096 (4 GB), 5120 (5 GB), 6144 (6 GB), 7168 (7 GB), 8192 (8 GB)

  • 2048 (2 vCPU) - Available memory values: Between 4096 (4 GB) and 16384 (16 GB) in increments of 1024 (1 GB)

  • 4096 (4 vCPU) - Available memory values: Between 8192 (8 GB) and 30720 (30 GB) in increments of 1024 (1 GB)

", "RegisterTaskDefinitionRequest$memory": "

The amount of memory (in MiB) used by the task. It can be expressed as an integer using MiB, for example 1024, or as a string using GB, for example 1GB or 1 GB, in a task definition but will be converted to an integer indicating the MiB when the task definition is registered.

Task-level CPU and memory parameters are ignored for Windows containers. We recommend specifying container-level resources for Windows containers.

If using the EC2 launch type, this field is optional.

If using the Fargate launch type, this field is required and you must use one of the following values, which determines your range of supported values for the cpu parameter:

  • 512 (0.5 GB), 1024 (1 GB), 2048 (2 GB) - Available cpu values: 256 (.25 vCPU)

  • 1024 (1 GB), 2048 (2 GB), 3072 (3 GB), 4096 (4 GB) - Available cpu values: 512 (.5 vCPU)

  • 2048 (2 GB), 3072 (3 GB), 4096 (4 GB), 5120 (5 GB), 6144 (6 GB), 7168 (7 GB), 8192 (8 GB) - Available cpu values: 1024 (1 vCPU)

  • Between 4096 (4 GB) and 16384 (16 GB) in increments of 1024 (1 GB) - Available cpu values: 2048 (2 vCPU)

  • Between 8192 (8 GB) and 30720 (30 GB) in increments of 1024 (1 GB) - Available cpu values: 4096 (4 vCPU)

", - "Resource$name": "

The name of the resource, such as cpu, memory, ports, or a user-defined resource.

", + "Resource$name": "

The name of the resource, such as CPU, MEMORY, PORTS, PORTS_UDP, or a user-defined resource.

", "Resource$type": "

The type of the resource, such as INTEGER, DOUBLE, LONG, or STRINGSET.

", "RunTaskRequest$cluster": "

The short name or full Amazon Resource Name (ARN) of the cluster on which to run your task. If you do not specify a cluster, the default cluster is assumed.

", "RunTaskRequest$taskDefinition": "

The family and revision (family:revision) or full ARN of the task definition to run. If a revision is not specified, the latest ACTIVE revision is used.

", @@ -1148,7 +1158,8 @@ "Service$roleArn": "

The ARN of the IAM role associated with the service that allows the Amazon ECS container agent to register container instances with an Elastic Load Balancing load balancer.

", "ServiceEvent$id": "

The ID string of the event.

", "ServiceEvent$message": "

The event message.

", - "ServiceRegistry$registryArn": "

The Amazon Resource Name (ARN) of the Service Registry. The currently supported service registry is Amazon Route 53 Auto Naming Service. For more information, see Service.

", + "ServiceRegistry$registryArn": "

The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is Amazon Route 53 Auto Naming. For more information, see Service.

", + "ServiceRegistry$containerName": "

The container name value, already specified in the task definition, to be used for your service discovery service. If the task definition that your service task specifies uses the bridge or host network mode, you must specify a containerName and containerPort combination from the task definition. If the task definition that your service task specifies uses the awsvpc network mode and a type SRV DNS record is used, you must specify either a containerName and containerPort combination or a port value, but not both.

", "StartTaskRequest$cluster": "

The short name or full Amazon Resource Name (ARN) of the cluster on which to start your task. If you do not specify a cluster, the default cluster is assumed.

", "StartTaskRequest$taskDefinition": "

The family and revision (family:revision) or full ARN of the task definition to start. If a revision is not specified, the latest ACTIVE revision is used.

", "StartTaskRequest$startedBy": "

An optional tag specified when a task is started. For example if you automatically trigger a task to run a batch process job, you could apply a unique identifier for that job to your task with the startedBy parameter. You can then identify which tasks belong to that job by filtering the results of a ListTasks call with the startedBy value. Up to 36 letters (uppercase and lowercase), numbers, hyphens, and underscores are allowed.

If a task is started by an Amazon ECS service, then the startedBy parameter contains the deployment ID of the service that starts it.

", @@ -1217,7 +1228,7 @@ "ContainerDefinition$dockerSecurityOptions": "

A list of strings to provide custom labels for SELinux and AppArmor multi-level security systems. This field is not valid for containers in tasks using the Fargate launch type.

This parameter maps to SecurityOpt in the Create a container section of the Docker Remote API and the --security-opt option to docker run.

The Amazon ECS container agent running on a container instance must register with the ECS_SELINUX_CAPABLE=true or ECS_APPARMOR_CAPABLE=true environment variables before containers placed on that instance can use these security options. For more information, see Amazon ECS Container Agent Configuration in the Amazon Elastic Container Service Developer Guide.

This parameter is not supported for Windows containers.

", "ContainerOverride$command": "

The command to send to the container that overrides the default command from the Docker image or the task definition. You must also specify a container name.

", "DescribeClustersRequest$clusters": "

A list of up to 100 cluster names or full cluster Amazon Resource Name (ARN) entries. If you do not specify a cluster, the default cluster is assumed.

", - "DescribeContainerInstancesRequest$containerInstances": "

A list of container instance IDs or full ARN entries.

", + "DescribeContainerInstancesRequest$containerInstances": "

A list of up to 100 container instance IDs or full Amazon Resource Name (ARN) entries.

", "DescribeServicesRequest$services": "

A list of services to describe. You may specify up to 10 services to describe in a single operation.

", "DescribeTasksRequest$tasks": "

A list of up to 100 task IDs or full ARN entries.

", "HealthCheck$command": "

A string array representing the command that the container runs to determine if it is healthy. The string array must start with CMD to execute the command arguments directly, or CMD-SHELL to run the command with the container's default shell. For example:

[ \"CMD-SHELL\", \"curl -f http://localhost/ || exit 1\" ]

An exit code of 0 indicates success, and non-zero exit code indicates failure. For more information, see HealthCheck in the Create a container section of the Docker Remote API.

", @@ -1360,7 +1371,7 @@ "TmpfsList": { "base": null, "refs": { - "LinuxParameters$tmpfs": "

The container path, mount options, and size of the tmpfs mount. This parameter maps to the --tmpfs option to docker run.

" + "LinuxParameters$tmpfs": "

The container path, mount options, and size (in MiB) of the tmpfs mount. This parameter maps to the --tmpfs option to docker run.

If you are using tasks that use the Fargate launch type, the tmpfs parameter is not supported.

" } }, "TransportProtocol": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/api-2.json new file mode 100644 index 000000000..8494169c1 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/api-2.json @@ -0,0 +1,301 @@ +{ + "version":"2.0", + "metadata":{ + "apiVersion":"2017-11-01", + "endpointPrefix":"eks", + "jsonVersion":"1.1", + "protocol":"rest-json", + "serviceAbbreviation":"Amazon EKS", + "serviceFullName":"Amazon Elastic Container Service for Kubernetes", + "serviceId":"EKS", + "signatureVersion":"v4", + "signingName":"eks", + "uid":"eks-2017-11-01" + }, + "operations":{ + "CreateCluster":{ + "name":"CreateCluster", + "http":{ + "method":"POST", + "requestUri":"/clusters" + }, + "input":{"shape":"CreateClusterRequest"}, + "output":{"shape":"CreateClusterResponse"}, + "errors":[ + {"shape":"ResourceInUseException"}, + {"shape":"ResourceLimitExceededException"}, + {"shape":"InvalidParameterException"}, + {"shape":"ClientException"}, + {"shape":"ServerException"}, + {"shape":"ServiceUnavailableException"}, + {"shape":"UnsupportedAvailabilityZoneException"} + ] + }, + "DeleteCluster":{ + "name":"DeleteCluster", + "http":{ + "method":"DELETE", + "requestUri":"/clusters/{name}" + }, + "input":{"shape":"DeleteClusterRequest"}, + "output":{"shape":"DeleteClusterResponse"}, + "errors":[ + {"shape":"ResourceInUseException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"ClientException"}, + {"shape":"ServerException"}, + {"shape":"ServiceUnavailableException"} + ] + }, + "DescribeCluster":{ + "name":"DescribeCluster", + "http":{ + "method":"GET", + "requestUri":"/clusters/{name}" + }, + "input":{"shape":"DescribeClusterRequest"}, + "output":{"shape":"DescribeClusterResponse"}, + "errors":[ + {"shape":"ResourceNotFoundException"}, + {"shape":"ClientException"}, + {"shape":"ServerException"}, + {"shape":"ServiceUnavailableException"} + ] + }, + "ListClusters":{ + "name":"ListClusters", + "http":{ + "method":"GET", + "requestUri":"/clusters" + }, + "input":{"shape":"ListClustersRequest"}, + "output":{"shape":"ListClustersResponse"}, + "errors":[ + {"shape":"InvalidParameterException"}, + {"shape":"ClientException"}, + {"shape":"ServerException"}, + {"shape":"ServiceUnavailableException"} + ] + } + }, + "shapes":{ + "Certificate":{ + "type":"structure", + "members":{ + "data":{"shape":"String"} + } + }, + "ClientException":{ + "type":"structure", + "members":{ + "clusterName":{"shape":"String"}, + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":400}, + "exception":true + }, + "Cluster":{ + "type":"structure", + "members":{ + "name":{"shape":"String"}, + "arn":{"shape":"String"}, + "createdAt":{"shape":"Timestamp"}, + "version":{"shape":"String"}, + "endpoint":{"shape":"String"}, + "roleArn":{"shape":"String"}, + "resourcesVpcConfig":{"shape":"VpcConfigResponse"}, + "status":{"shape":"ClusterStatus"}, + "certificateAuthority":{"shape":"Certificate"}, + "clientRequestToken":{"shape":"String"} + } + }, + "ClusterName":{ + "type":"string", + "max":255, + "min":1, + "pattern":"[A-Za-z0-9\\-_]*" + }, + "ClusterStatus":{ + "type":"string", + "enum":[ + "CREATING", + "ACTIVE", + "DELETING", + "FAILED" + ] + }, + "CreateClusterRequest":{ + "type":"structure", + "required":[ + "name", + "roleArn", + "resourcesVpcConfig" + ], + "members":{ + "name":{"shape":"ClusterName"}, + "version":{"shape":"String"}, + "roleArn":{"shape":"String"}, + "resourcesVpcConfig":{"shape":"VpcConfigRequest"}, + "clientRequestToken":{ + "shape":"String", + "idempotencyToken":true + } + } + }, + "CreateClusterResponse":{ + "type":"structure", + "members":{ + "cluster":{"shape":"Cluster"} + } + }, + "DeleteClusterRequest":{ + "type":"structure", + "required":["name"], + "members":{ + "name":{ + "shape":"String", + "location":"uri", + "locationName":"name" + } + } + }, + "DeleteClusterResponse":{ + "type":"structure", + "members":{ + "cluster":{"shape":"Cluster"} + } + }, + "DescribeClusterRequest":{ + "type":"structure", + "required":["name"], + "members":{ + "name":{ + "shape":"String", + "location":"uri", + "locationName":"name" + } + } + }, + "DescribeClusterResponse":{ + "type":"structure", + "members":{ + "cluster":{"shape":"Cluster"} + } + }, + "InvalidParameterException":{ + "type":"structure", + "members":{ + "clusterName":{"shape":"String"}, + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":400}, + "exception":true + }, + "ListClustersRequest":{ + "type":"structure", + "members":{ + "maxResults":{ + "shape":"ListClustersRequestMaxResults", + "location":"querystring", + "locationName":"maxResults" + }, + "nextToken":{ + "shape":"String", + "location":"querystring", + "locationName":"nextToken" + } + } + }, + "ListClustersRequestMaxResults":{ + "type":"integer", + "box":true, + "max":100, + "min":1 + }, + "ListClustersResponse":{ + "type":"structure", + "members":{ + "clusters":{"shape":"StringList"}, + "nextToken":{"shape":"String"} + } + }, + "ResourceInUseException":{ + "type":"structure", + "members":{ + "clusterName":{"shape":"String"}, + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":409}, + "exception":true + }, + "ResourceLimitExceededException":{ + "type":"structure", + "members":{ + "clusterName":{"shape":"String"}, + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":400}, + "exception":true + }, + "ResourceNotFoundException":{ + "type":"structure", + "members":{ + "clusterName":{"shape":"String"}, + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":404}, + "exception":true + }, + "ServerException":{ + "type":"structure", + "members":{ + "clusterName":{"shape":"String"}, + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":500}, + "exception":true, + "fault":true + }, + "ServiceUnavailableException":{ + "type":"structure", + "members":{ + "message":{"shape":"String"} + }, + "error":{"httpStatusCode":503}, + "exception":true, + "fault":true + }, + "String":{"type":"string"}, + "StringList":{ + "type":"list", + "member":{"shape":"String"} + }, + "Timestamp":{"type":"timestamp"}, + "UnsupportedAvailabilityZoneException":{ + "type":"structure", + "members":{ + "message":{"shape":"String"}, + "clusterName":{"shape":"String"}, + "validZones":{"shape":"StringList"} + }, + "error":{"httpStatusCode":400}, + "exception":true + }, + "VpcConfigRequest":{ + "type":"structure", + "required":["subnetIds"], + "members":{ + "subnetIds":{"shape":"StringList"}, + "securityGroupIds":{"shape":"StringList"} + } + }, + "VpcConfigResponse":{ + "type":"structure", + "members":{ + "subnetIds":{"shape":"StringList"}, + "securityGroupIds":{"shape":"StringList"}, + "vpcId":{"shape":"String"} + } + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/docs-2.json new file mode 100644 index 000000000..7ba7b4edc --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/docs-2.json @@ -0,0 +1,189 @@ +{ + "version": "2.0", + "service": "

Amazon Elastic Container Service for Kubernetes (Amazon EKS) is a managed service that makes it easy for you to run Kubernetes on AWS without needing to stand up or maintain your own Kubernetes control plane. Kubernetes is an open-source system for automating the deployment, scaling, and management of containerized applications.

Amazon EKS runs three Kubernetes control plane instances across three Availability Zones to ensure high availability. Amazon EKS automatically detects and replaces unhealthy control plane instances, and it provides automated version upgrades and patching for them.

Amazon EKS is also integrated with many AWS services to provide scalability and security for your applications, including the following:

  • Elastic Load Balancing for load distribution

  • IAM for authentication

  • Amazon VPC for isolation

Amazon EKS runs up to date versions of the open-source Kubernetes software, so you can use all the existing plugins and tooling from the Kubernetes community. Applications running on Amazon EKS are fully compatible with applications running on any standard Kubernetes environment, whether running in on-premises data centers or public clouds. This means that you can easily migrate any standard Kubernetes application to Amazon EKS without any code modification required.

", + "operations": { + "CreateCluster": "

Creates an Amazon EKS control plane.

The Amazon EKS control plane consists of control plane instances that run the Kubernetes software, like etcd and the API server. The control plane runs in an account managed by AWS, and the Kubernetes API is exposed via the Amazon EKS API server endpoint.

Amazon EKS worker nodes run in your AWS account and connect to your cluster's control plane via the Kubernetes API server endpoint and a certificate file that is created for your cluster.

The cluster control plane is provisioned across multiple Availability Zones and fronted by an Elastic Load Balancing Network Load Balancer. Amazon EKS also provisions elastic network interfaces in your VPC subnets to provide connectivity from the control plane instances to the worker nodes (for example, to support kubectl exec, logs, and proxy data flows).

After you create an Amazon EKS cluster, you must configure your Kubernetes tooling to communicate with the API server and launch worker nodes into your cluster. For more information, see Managing Cluster Authentication and Launching Amazon EKS Worker Nodesin the Amazon EKS User Guide.

", + "DeleteCluster": "

Deletes the Amazon EKS cluster control plane.

If you have active services in your cluster that are associated with a load balancer, you must delete those services before deleting the cluster so that the load balancers are deleted properly. Otherwise, you can have orphaned resources in your VPC that prevent you from being able to delete the VPC. For more information, see Deleting a Cluster in the Amazon EKS User Guide.

", + "DescribeCluster": "

Returns descriptive information about an Amazon EKS cluster.

The API server endpoint and certificate authority data returned by this operation are required for kubelet and kubectl to communicate with your Kubernetes API server. For more information, see Create a kubeconfig for Amazon EKS.

The API server endpoint and certificate authority data are not available until the cluster reaches the ACTIVE state.

", + "ListClusters": "

Lists the Amazon EKS clusters in your AWS account in the specified region.

" + }, + "shapes": { + "Certificate": { + "base": "

An object representing the certificate-authority-data for your cluster.

", + "refs": { + "Cluster$certificateAuthority": "

The certificate-authority-data for your cluster.

" + } + }, + "ClientException": { + "base": "

These errors are usually caused by a client action, such as using an action or resource on behalf of a user that doesn't have permissions to use the action or resource, or specifying an identifier that is not valid.

", + "refs": { + } + }, + "Cluster": { + "base": "

An object representing an Amazon EKS cluster.

", + "refs": { + "CreateClusterResponse$cluster": "

The full description of your new cluster.

", + "DeleteClusterResponse$cluster": "

The full description of the cluster to delete.

", + "DescribeClusterResponse$cluster": "

The full description of your specified cluster.

" + } + }, + "ClusterName": { + "base": null, + "refs": { + "CreateClusterRequest$name": "

The unique name to give to your cluster.

" + } + }, + "ClusterStatus": { + "base": null, + "refs": { + "Cluster$status": "

The current status of the cluster.

" + } + }, + "CreateClusterRequest": { + "base": null, + "refs": { + } + }, + "CreateClusterResponse": { + "base": null, + "refs": { + } + }, + "DeleteClusterRequest": { + "base": null, + "refs": { + } + }, + "DeleteClusterResponse": { + "base": null, + "refs": { + } + }, + "DescribeClusterRequest": { + "base": null, + "refs": { + } + }, + "DescribeClusterResponse": { + "base": null, + "refs": { + } + }, + "InvalidParameterException": { + "base": "

The specified parameter is invalid. Review the available parameters for the API request.

", + "refs": { + } + }, + "ListClustersRequest": { + "base": null, + "refs": { + } + }, + "ListClustersRequestMaxResults": { + "base": null, + "refs": { + "ListClustersRequest$maxResults": "

The maximum number of cluster results returned by ListClusters in paginated output. When this parameter is used, ListClusters only returns maxResults results in a single page along with a nextToken response element. The remaining results of the initial request can be seen by sending another ListClusters request with the returned nextToken value. This value can be between 1 and 100. If this parameter is not used, then ListClusters returns up to 100 results and a nextToken value if applicable.

" + } + }, + "ListClustersResponse": { + "base": null, + "refs": { + } + }, + "ResourceInUseException": { + "base": "

The specified resource is in use.

", + "refs": { + } + }, + "ResourceLimitExceededException": { + "base": "

You have encountered a service limit on the specified resource.

", + "refs": { + } + }, + "ResourceNotFoundException": { + "base": "

The specified resource could not be found. You can view your available clusters with ListClusters. Amazon EKS clusters are region-specific.

", + "refs": { + } + }, + "ServerException": { + "base": "

These errors are usually caused by a server-side issue.

", + "refs": { + } + }, + "ServiceUnavailableException": { + "base": "

The service is unavailable, back off and retry the operation.

", + "refs": { + } + }, + "String": { + "base": null, + "refs": { + "Certificate$data": "

The base64 encoded certificate data required to communicate with your cluster. Add this to the certificate-authority-data section of the kubeconfig file for your cluster.

", + "ClientException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "ClientException$message": null, + "Cluster$name": "

The name of the cluster.

", + "Cluster$arn": "

The Amazon Resource Name (ARN) of the cluster.

", + "Cluster$version": "

The Kubernetes server version for the cluster.

", + "Cluster$endpoint": "

The endpoint for your Kubernetes API server.

", + "Cluster$roleArn": "

The Amazon Resource Name (ARN) of the IAM role that provides permissions for the Kubernetes control plane to make calls to AWS API operations on your behalf.

", + "Cluster$clientRequestToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", + "CreateClusterRequest$version": "

The desired Kubernetes version for your cluster. If you do not specify a value here, the latest version available in Amazon EKS is used.

", + "CreateClusterRequest$roleArn": "

The Amazon Resource Name (ARN) of the IAM role that provides permissions for Amazon EKS to make calls to other AWS API operations on your behalf. For more information, see Amazon EKS Service IAM Role in the Amazon EKS User Guide

", + "CreateClusterRequest$clientRequestToken": "

Unique, case-sensitive identifier you provide to ensure the idempotency of the request.

", + "DeleteClusterRequest$name": "

The name of the cluster to delete.

", + "DescribeClusterRequest$name": "

The name of the cluster to describe.

", + "InvalidParameterException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "InvalidParameterException$message": null, + "ListClustersRequest$nextToken": "

The nextToken value returned from a previous paginated ListClusters request where maxResults was used and the results exceeded the value of that parameter. Pagination continues from the end of the previous results that returned the nextToken value.

This token should be treated as an opaque identifier that is only used to retrieve the next items in a list and not for other programmatic purposes.

", + "ListClustersResponse$nextToken": "

The nextToken value to include in a future ListClusters request. When the results of a ListClusters request exceed maxResults, this value can be used to retrieve the next page of results. This value is null when there are no more results to return.

", + "ResourceInUseException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "ResourceInUseException$message": null, + "ResourceLimitExceededException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "ResourceLimitExceededException$message": null, + "ResourceNotFoundException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "ResourceNotFoundException$message": null, + "ServerException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "ServerException$message": null, + "ServiceUnavailableException$message": null, + "StringList$member": null, + "UnsupportedAvailabilityZoneException$message": null, + "UnsupportedAvailabilityZoneException$clusterName": "

The Amazon EKS cluster associated with the exception.

", + "VpcConfigResponse$vpcId": "

The VPC associated with your cluster.

" + } + }, + "StringList": { + "base": null, + "refs": { + "ListClustersResponse$clusters": "

A list of all of the clusters for your account in the specified region.

", + "UnsupportedAvailabilityZoneException$validZones": "

The supported Availability Zones for your account. Choose subnets in these Availability Zones for your cluster.

", + "VpcConfigRequest$subnetIds": "

Specify subnets for your Amazon EKS worker nodes. Amazon EKS creates cross-account elastic network interfaces in these subnets to allow communication between your worker nodes and the Kubernetes control plane.

", + "VpcConfigRequest$securityGroupIds": "

Specify one or more security groups for the cross-account elastic network interfaces that Amazon EKS creates to use to allow communication between your worker nodes and the Kubernetes control plane.

", + "VpcConfigResponse$subnetIds": "

The subnets associated with your cluster.

", + "VpcConfigResponse$securityGroupIds": "

The security groups associated with the cross-account elastic network interfaces that are used to allow communication between your worker nodes and the Kubernetes control plane.

" + } + }, + "Timestamp": { + "base": null, + "refs": { + "Cluster$createdAt": "

The Unix epoch time stamp in seconds for when the cluster was created.

" + } + }, + "UnsupportedAvailabilityZoneException": { + "base": "

At least one of your specified cluster subnets is in an Availability Zone that does not support Amazon EKS. The exception output will specify the supported Availability Zones for your account, from which you can choose subnets for your cluster.

", + "refs": { + } + }, + "VpcConfigRequest": { + "base": "

An object representing an Amazon EKS cluster VPC configuration request.

", + "refs": { + "CreateClusterRequest$resourcesVpcConfig": "

The VPC subnets and security groups used by the cluster control plane. Amazon EKS VPC resources have specific requirements to work properly with Kubernetes. For more information, see Cluster VPC Considerations and Cluster Security Group Considerations in the Amazon EKS User Guide.

" + } + }, + "VpcConfigResponse": { + "base": "

An object representing an Amazon EKS cluster VPC configuration response.

", + "refs": { + "Cluster$resourcesVpcConfig": "

The VPC subnets and security groups used by the cluster control plane. Amazon EKS VPC resources have specific requirements to work properly with Kubernetes. For more information, see Cluster VPC Considerations and Cluster Security Group Considerations in the Amazon EKS User Guide.

" + } + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/examples-1.json new file mode 100644 index 000000000..6a83da723 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/examples-1.json @@ -0,0 +1,114 @@ +{ + "version": "1.0", + "examples": { + "CreateCluster": [ + { + "input": { + "version": "1.10", + "name": "prod", + "clientRequestToken": "1d2129a1-3d38-460a-9756-e5b91fddb951", + "resourcesVpcConfig": { + "securityGroupIds": [ + "sg-6979fe18" + ], + "subnetIds": [ + "subnet-6782e71e", + "subnet-e7e761ac" + ] + }, + "roleArn": "arn:aws:iam::012345678910:role/eks-service-role-AWSServiceRoleForAmazonEKS-J7ONKE3BQ4PI" + }, + "output": { + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "The following example creates an Amazon EKS cluster called prod.", + "id": "to-create-a-new-cluster-1527868185648", + "title": "To create a new cluster" + } + ], + "DeleteCluster": [ + { + "input": { + "name": "devel" + }, + "output": { + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "This example command deletes a cluster named `devel` in your default region.", + "id": "to-delete-a-cluster-1527868641252", + "title": "To delete a cluster" + } + ], + "DescribeCluster": [ + { + "input": { + "name": "devel" + }, + "output": { + "cluster": { + "version": "1.10", + "name": "devel", + "arn": "arn:aws:eks:us-west-2:012345678910:cluster/devel", + "certificateAuthority": { + "data": "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUN5RENDQWJDZ0F3SUJBZ0lCQURBTkJna3Foa2lHOXcwQkFRc0ZBREFWTVJNd0VRWURWUVFERXdwcmRXSmwKY201bGRHVnpNQjRYRFRFNE1EVXpNVEl6TVRFek1Wb1hEVEk0TURVeU9ESXpNVEV6TVZvd0ZURVRNQkVHQTFVRQpBeE1LYTNWaVpYSnVaWFJsY3pDQ0FTSXdEUVlKS29aSWh2Y05BUUVCQlFBRGdnRVBBRENDQVFvQ2dnRUJBTTZWCjVUaG4rdFcySm9Xa2hQMzRlVUZMNitaRXJOZGIvWVdrTmtDdWNGS2RaaXl2TjlMVmdvUmV2MjlFVFZlN1ZGbSsKUTJ3ZURyRXJiQyt0dVlibkFuN1ZLYmE3ay9hb1BHekZMdmVnb0t6b0M1N2NUdGVwZzRIazRlK2tIWHNaME10MApyb3NzcjhFM1ROeExETnNJTThGL1cwdjhsTGNCbWRPcjQyV2VuTjFHZXJnaDNSZ2wzR3JIazBnNTU0SjFWenJZCm9hTi8zODFUczlOTFF2QTBXb0xIcjBFRlZpTFdSZEoyZ3lXaC9ybDVyOFNDOHZaQXg1YW1BU0hVd01aTFpWRC8KTDBpOW4wRVM0MkpVdzQyQmxHOEdpd3NhTkJWV3lUTHZKclNhRXlDSHFtVVZaUTFDZkFXUjl0L3JleVVOVXM3TApWV1FqM3BFbk9RMitMSWJrc0RzQ0F3RUFBYU1qTUNFd0RnWURWUjBQQVFIL0JBUURBZ0trTUE4R0ExVWRFd0VCCi93UUZNQU1CQWY4d0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFNZ3RsQ1dIQ2U2YzVHMXl2YlFTS0Q4K2hUalkKSm1NSG56L2EvRGt0WG9YUjFVQzIrZUgzT1BZWmVjRVZZZHVaSlZCckNNQ2VWR0ZkeWdBYlNLc1FxWDg0S2RXbAp1MU5QaERDSmEyRHliN2pVMUV6VThTQjFGZUZ5ZFE3a0hNS1E1blpBRVFQOTY4S01hSGUrSm0yQ2x1UFJWbEJVCjF4WlhTS1gzTVZ0K1Q0SU1EV2d6c3JRSjVuQkRjdEtLcUZtM3pKdVVubHo5ZEpVckdscEltMjVJWXJDckxYUFgKWkUwRUtRNWEzMHhkVWNrTHRGQkQrOEtBdFdqSS9yZUZPNzM1YnBMdVoyOTBaNm42QlF3elRrS0p4cnhVc3QvOAppNGsxcnlsaUdWMm5SSjBUYjNORkczNHgrYWdzYTRoSTFPbU90TFM0TmgvRXJxT3lIUXNDc2hEQUtKUT0KLS0tLS1FTkQgQ0VSVElGSUNBVEUtLS0tLQo=" + }, + "createdAt": 1527807879.988, + "endpoint": "https://A0DCCD80A04F01705DD065655C30CC3D.yl4.us-west-2.eks.amazonaws.com", + "resourcesVpcConfig": { + "securityGroupIds": [ + "sg-6979fe18" + ], + "subnetIds": [ + "subnet-6782e71e", + "subnet-e7e761ac" + ], + "vpcId": "vpc-950809ec" + }, + "roleArn": "arn:aws:iam::012345678910:role/eks-service-role-AWSServiceRoleForAmazonEKS-J7ONKE3BQ4PI", + "status": "ACTIVE" + } + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "This example command provides a description of the specified cluster in your default region.", + "id": "to-describe-a-cluster-1527868708512", + "title": "To describe a cluster" + } + ], + "ListClusters": [ + { + "input": { + }, + "output": { + "clusters": [ + "devel", + "prod" + ] + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "This example command lists all of your available clusters in your default region.", + "id": "to-list-your-available-clusters-1527868801040", + "title": "To list your available clusters" + } + ] + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/paginators-1.json similarity index 100% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/paginators-1.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/eks/2017-11-01/paginators-1.json diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/api-2.json index afa28d2dc..71588a6e7 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/api-2.json @@ -1419,7 +1419,8 @@ "Info", "Warning", "Degraded", - "Severe" + "Severe", + "Suspended" ] }, "EnvironmentId":{"type":"string"}, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/docs-2.json index 61aafe21a..62b7f122a 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/docs-2.json @@ -1,2026 +1,2026 @@ { - "version": "2.0", - "service": "AWS Elastic Beanstalk

AWS Elastic Beanstalk makes it easy for you to create, deploy, and manage scalable, fault-tolerant applications running on the Amazon Web Services cloud.

For more information about this product, go to the AWS Elastic Beanstalk details page. The location of the latest AWS Elastic Beanstalk WSDL is http://elasticbeanstalk.s3.amazonaws.com/doc/2010-12-01/AWSElasticBeanstalk.wsdl. To install the Software Development Kits (SDKs), Integrated Development Environment (IDE) Toolkits, and command line tools that enable you to access the API, go to Tools for Amazon Web Services.

Endpoints

For a list of region-specific endpoints that AWS Elastic Beanstalk supports, go to Regions and Endpoints in the Amazon Web Services Glossary.

", - "operations": { - "AbortEnvironmentUpdate": "

Cancels in-progress environment configuration update or application version deployment.

", - "ApplyEnvironmentManagedAction": "

Applies a scheduled managed action immediately. A managed action can be applied only if its status is Scheduled. Get the status and action ID of a managed action with DescribeEnvironmentManagedActions.

", - "CheckDNSAvailability": "

Checks if the specified CNAME is available.

", - "ComposeEnvironments": "

Create or update a group of environments that each run a separate component of a single application. Takes a list of version labels that specify application source bundles for each of the environments to create or update. The name of each environment and other required information must be included in the source bundles in an environment manifest named env.yaml. See Compose Environments for details.

", - "CreateApplication": "

Creates an application that has one configuration template named default and no application versions.

", - "CreateApplicationVersion": "

Creates an application version for the specified application. You can create an application version from a source bundle in Amazon S3, a commit in AWS CodeCommit, or the output of an AWS CodeBuild build as follows:

Specify a commit in an AWS CodeCommit repository with SourceBuildInformation.

Specify a build in an AWS CodeBuild with SourceBuildInformation and BuildConfiguration.

Specify a source bundle in S3 with SourceBundle

Omit both SourceBuildInformation and SourceBundle to use the default sample application.

Once you create an application version with a specified Amazon S3 bucket and key location, you cannot change that Amazon S3 location. If you change the Amazon S3 location, you receive an exception when you attempt to launch an environment from the application version.

", - "CreateConfigurationTemplate": "

Creates a configuration template. Templates are associated with a specific application and are used to deploy different versions of the application with the same configuration settings.

Related Topics

", - "CreateEnvironment": "

Launches an environment for the specified application using the specified configuration.

", - "CreatePlatformVersion": "

Create a new version of your custom platform.

", - "CreateStorageLocation": "

Creates a bucket in Amazon S3 to store application versions, logs, and other files used by Elastic Beanstalk environments. The Elastic Beanstalk console and EB CLI call this API the first time you create an environment in a region. If the storage location already exists, CreateStorageLocation still returns the bucket name but does not create a new bucket.

", - "DeleteApplication": "

Deletes the specified application along with all associated versions and configurations. The application versions will not be deleted from your Amazon S3 bucket.

You cannot delete an application that has a running environment.

", - "DeleteApplicationVersion": "

Deletes the specified version from the specified application.

You cannot delete an application version that is associated with a running environment.

", - "DeleteConfigurationTemplate": "

Deletes the specified configuration template.

When you launch an environment using a configuration template, the environment gets a copy of the template. You can delete or modify the environment's copy of the template without affecting the running environment.

", - "DeleteEnvironmentConfiguration": "

Deletes the draft configuration associated with the running environment.

Updating a running environment with any configuration changes creates a draft configuration set. You can get the draft configuration using DescribeConfigurationSettings while the update is in progress or if the update fails. The DeploymentStatus for the draft configuration indicates whether the deployment is in process or has failed. The draft configuration remains in existence until it is deleted with this action.

", - "DeletePlatformVersion": "

Deletes the specified version of a custom platform.

", - "DescribeAccountAttributes": "

Returns attributes related to AWS Elastic Beanstalk that are associated with the calling AWS account.

The result currently has one set of attributes—resource quotas.

", - "DescribeApplicationVersions": "

Retrieve a list of application versions.

", - "DescribeApplications": "

Returns the descriptions of existing applications.

", - "DescribeConfigurationOptions": "

Describes the configuration options that are used in a particular configuration template or environment, or that a specified solution stack defines. The description includes the values the options, their default values, and an indication of the required action on a running environment if an option value is changed.

", - "DescribeConfigurationSettings": "

Returns a description of the settings for the specified configuration set, that is, either a configuration template or the configuration set associated with a running environment.

When describing the settings for the configuration set associated with a running environment, it is possible to receive two sets of setting descriptions. One is the deployed configuration set, and the other is a draft configuration of an environment that is either in the process of deployment or that failed to deploy.

Related Topics

", - "DescribeEnvironmentHealth": "

Returns information about the overall health of the specified environment. The DescribeEnvironmentHealth operation is only available with AWS Elastic Beanstalk Enhanced Health.

", - "DescribeEnvironmentManagedActionHistory": "

Lists an environment's completed and failed managed actions.

", - "DescribeEnvironmentManagedActions": "

Lists an environment's upcoming and in-progress managed actions.

", - "DescribeEnvironmentResources": "

Returns AWS resources for this environment.

", - "DescribeEnvironments": "

Returns descriptions for existing environments.

", - "DescribeEvents": "

Returns list of event descriptions matching criteria up to the last 6 weeks.

This action returns the most recent 1,000 events from the specified NextToken.

", - "DescribeInstancesHealth": "

Retrives detailed information about the health of instances in your AWS Elastic Beanstalk. This operation requires enhanced health reporting.

", - "DescribePlatformVersion": "

Describes the version of the platform.

", - "ListAvailableSolutionStacks": "

Returns a list of the available solution stack names, with the public version first and then in reverse chronological order.

", - "ListPlatformVersions": "

Lists the available platforms.

", - "ListTagsForResource": "

Returns the tags applied to an AWS Elastic Beanstalk resource. The response contains a list of tag key-value pairs.

Currently, Elastic Beanstalk only supports tagging of Elastic Beanstalk environments. For details about environment tagging, see Tagging Resources in Your Elastic Beanstalk Environment.

", - "RebuildEnvironment": "

Deletes and recreates all of the AWS resources (for example: the Auto Scaling group, load balancer, etc.) for a specified environment and forces a restart.

", - "RequestEnvironmentInfo": "

Initiates a request to compile the specified type of information of the deployed environment.

Setting the InfoType to tail compiles the last lines from the application server log files of every Amazon EC2 instance in your environment.

Setting the InfoType to bundle compresses the application server log files for every Amazon EC2 instance into a .zip file. Legacy and .NET containers do not support bundle logs.

Use RetrieveEnvironmentInfo to obtain the set of logs.

Related Topics

", - "RestartAppServer": "

Causes the environment to restart the application container server running on each Amazon EC2 instance.

", - "RetrieveEnvironmentInfo": "

Retrieves the compiled information from a RequestEnvironmentInfo request.

Related Topics

", - "SwapEnvironmentCNAMEs": "

Swaps the CNAMEs of two environments.

", - "TerminateEnvironment": "

Terminates the specified environment.

", - "UpdateApplication": "

Updates the specified application to have the specified properties.

If a property (for example, description) is not provided, the value remains unchanged. To clear these properties, specify an empty string.

", - "UpdateApplicationResourceLifecycle": "

Modifies lifecycle settings for an application.

", - "UpdateApplicationVersion": "

Updates the specified application version to have the specified properties.

If a property (for example, description) is not provided, the value remains unchanged. To clear properties, specify an empty string.

", - "UpdateConfigurationTemplate": "

Updates the specified configuration template to have the specified properties or configuration option values.

If a property (for example, ApplicationName) is not provided, its value remains unchanged. To clear such properties, specify an empty string.

Related Topics

", - "UpdateEnvironment": "

Updates the environment description, deploys a new application version, updates the configuration settings to an entirely new configuration template, or updates select configuration option values in the running environment.

Attempting to update both the release and configuration is not allowed and AWS Elastic Beanstalk returns an InvalidParameterCombination error.

When updating the configuration settings to a new template or individual settings, a draft configuration is created and DescribeConfigurationSettings for this environment returns two setting descriptions with different DeploymentStatus values.

", - "UpdateTagsForResource": "

Update the list of tags applied to an AWS Elastic Beanstalk resource. Two lists can be passed: TagsToAdd for tags to add or update, and TagsToRemove.

Currently, Elastic Beanstalk only supports tagging of Elastic Beanstalk environments. For details about environment tagging, see Tagging Resources in Your Elastic Beanstalk Environment.

If you create a custom IAM user policy to control permission to this operation, specify one of the following two virtual actions (or both) instead of the API operation name:

elasticbeanstalk:AddTags

Controls permission to call UpdateTagsForResource and pass a list of tags to add in the TagsToAdd parameter.

elasticbeanstalk:RemoveTags

Controls permission to call UpdateTagsForResource and pass a list of tag keys to remove in the TagsToRemove parameter.

For details about creating a custom user policy, see Creating a Custom User Policy.

", - "ValidateConfigurationSettings": "

Takes a set of configuration settings and either a configuration template or environment, and determines whether those values are valid.

This action returns a list of messages indicating any errors or warnings associated with the selection of option values.

" + "version":"2.0", + "service":"AWS Elastic Beanstalk

AWS Elastic Beanstalk makes it easy for you to create, deploy, and manage scalable, fault-tolerant applications running on the Amazon Web Services cloud.

For more information about this product, go to the AWS Elastic Beanstalk details page. The location of the latest AWS Elastic Beanstalk WSDL is http://elasticbeanstalk.s3.amazonaws.com/doc/2010-12-01/AWSElasticBeanstalk.wsdl. To install the Software Development Kits (SDKs), Integrated Development Environment (IDE) Toolkits, and command line tools that enable you to access the API, go to Tools for Amazon Web Services.

Endpoints

For a list of region-specific endpoints that AWS Elastic Beanstalk supports, go to Regions and Endpoints in the Amazon Web Services Glossary.

", + "operations":{ + "AbortEnvironmentUpdate":"

Cancels in-progress environment configuration update or application version deployment.

", + "ApplyEnvironmentManagedAction":"

Applies a scheduled managed action immediately. A managed action can be applied only if its status is Scheduled. Get the status and action ID of a managed action with DescribeEnvironmentManagedActions.

", + "CheckDNSAvailability":"

Checks if the specified CNAME is available.

", + "ComposeEnvironments":"

Create or update a group of environments that each run a separate component of a single application. Takes a list of version labels that specify application source bundles for each of the environments to create or update. The name of each environment and other required information must be included in the source bundles in an environment manifest named env.yaml. See Compose Environments for details.

", + "CreateApplication":"

Creates an application that has one configuration template named default and no application versions.

", + "CreateApplicationVersion":"

Creates an application version for the specified application. You can create an application version from a source bundle in Amazon S3, a commit in AWS CodeCommit, or the output of an AWS CodeBuild build as follows:

Specify a commit in an AWS CodeCommit repository with SourceBuildInformation.

Specify a build in an AWS CodeBuild with SourceBuildInformation and BuildConfiguration.

Specify a source bundle in S3 with SourceBundle

Omit both SourceBuildInformation and SourceBundle to use the default sample application.

Once you create an application version with a specified Amazon S3 bucket and key location, you cannot change that Amazon S3 location. If you change the Amazon S3 location, you receive an exception when you attempt to launch an environment from the application version.

", + "CreateConfigurationTemplate":"

Creates a configuration template. Templates are associated with a specific application and are used to deploy different versions of the application with the same configuration settings.

Related Topics

", + "CreateEnvironment":"

Launches an environment for the specified application using the specified configuration.

", + "CreatePlatformVersion":"

Create a new version of your custom platform.

", + "CreateStorageLocation":"

Creates a bucket in Amazon S3 to store application versions, logs, and other files used by Elastic Beanstalk environments. The Elastic Beanstalk console and EB CLI call this API the first time you create an environment in a region. If the storage location already exists, CreateStorageLocation still returns the bucket name but does not create a new bucket.

", + "DeleteApplication":"

Deletes the specified application along with all associated versions and configurations. The application versions will not be deleted from your Amazon S3 bucket.

You cannot delete an application that has a running environment.

", + "DeleteApplicationVersion":"

Deletes the specified version from the specified application.

You cannot delete an application version that is associated with a running environment.

", + "DeleteConfigurationTemplate":"

Deletes the specified configuration template.

When you launch an environment using a configuration template, the environment gets a copy of the template. You can delete or modify the environment's copy of the template without affecting the running environment.

", + "DeleteEnvironmentConfiguration":"

Deletes the draft configuration associated with the running environment.

Updating a running environment with any configuration changes creates a draft configuration set. You can get the draft configuration using DescribeConfigurationSettings while the update is in progress or if the update fails. The DeploymentStatus for the draft configuration indicates whether the deployment is in process or has failed. The draft configuration remains in existence until it is deleted with this action.

", + "DeletePlatformVersion":"

Deletes the specified version of a custom platform.

", + "DescribeAccountAttributes":"

Returns attributes related to AWS Elastic Beanstalk that are associated with the calling AWS account.

The result currently has one set of attributes—resource quotas.

", + "DescribeApplicationVersions":"

Retrieve a list of application versions.

", + "DescribeApplications":"

Returns the descriptions of existing applications.

", + "DescribeConfigurationOptions":"

Describes the configuration options that are used in a particular configuration template or environment, or that a specified solution stack defines. The description includes the values the options, their default values, and an indication of the required action on a running environment if an option value is changed.

", + "DescribeConfigurationSettings":"

Returns a description of the settings for the specified configuration set, that is, either a configuration template or the configuration set associated with a running environment.

When describing the settings for the configuration set associated with a running environment, it is possible to receive two sets of setting descriptions. One is the deployed configuration set, and the other is a draft configuration of an environment that is either in the process of deployment or that failed to deploy.

Related Topics

", + "DescribeEnvironmentHealth":"

Returns information about the overall health of the specified environment. The DescribeEnvironmentHealth operation is only available with AWS Elastic Beanstalk Enhanced Health.

", + "DescribeEnvironmentManagedActionHistory":"

Lists an environment's completed and failed managed actions.

", + "DescribeEnvironmentManagedActions":"

Lists an environment's upcoming and in-progress managed actions.

", + "DescribeEnvironmentResources":"

Returns AWS resources for this environment.

", + "DescribeEnvironments":"

Returns descriptions for existing environments.

", + "DescribeEvents":"

Returns list of event descriptions matching criteria up to the last 6 weeks.

This action returns the most recent 1,000 events from the specified NextToken.

", + "DescribeInstancesHealth":"

Retrieves detailed information about the health of instances in your AWS Elastic Beanstalk. This operation requires enhanced health reporting.

", + "DescribePlatformVersion":"

Describes the version of the platform.

", + "ListAvailableSolutionStacks":"

Returns a list of the available solution stack names, with the public version first and then in reverse chronological order.

", + "ListPlatformVersions":"

Lists the available platforms.

", + "ListTagsForResource":"

Returns the tags applied to an AWS Elastic Beanstalk resource. The response contains a list of tag key-value pairs.

Currently, Elastic Beanstalk only supports tagging of Elastic Beanstalk environments. For details about environment tagging, see Tagging Resources in Your Elastic Beanstalk Environment.

", + "RebuildEnvironment":"

Deletes and recreates all of the AWS resources (for example: the Auto Scaling group, load balancer, etc.) for a specified environment and forces a restart.

", + "RequestEnvironmentInfo":"

Initiates a request to compile the specified type of information of the deployed environment.

Setting the InfoType to tail compiles the last lines from the application server log files of every Amazon EC2 instance in your environment.

Setting the InfoType to bundle compresses the application server log files for every Amazon EC2 instance into a .zip file. Legacy and .NET containers do not support bundle logs.

Use RetrieveEnvironmentInfo to obtain the set of logs.

Related Topics

", + "RestartAppServer":"

Causes the environment to restart the application container server running on each Amazon EC2 instance.

", + "RetrieveEnvironmentInfo":"

Retrieves the compiled information from a RequestEnvironmentInfo request.

Related Topics

", + "SwapEnvironmentCNAMEs":"

Swaps the CNAMEs of two environments.

", + "TerminateEnvironment":"

Terminates the specified environment.

", + "UpdateApplication":"

Updates the specified application to have the specified properties.

If a property (for example, description) is not provided, the value remains unchanged. To clear these properties, specify an empty string.

", + "UpdateApplicationResourceLifecycle":"

Modifies lifecycle settings for an application.

", + "UpdateApplicationVersion":"

Updates the specified application version to have the specified properties.

If a property (for example, description) is not provided, the value remains unchanged. To clear properties, specify an empty string.

", + "UpdateConfigurationTemplate":"

Updates the specified configuration template to have the specified properties or configuration option values.

If a property (for example, ApplicationName) is not provided, its value remains unchanged. To clear such properties, specify an empty string.

Related Topics

", + "UpdateEnvironment":"

Updates the environment description, deploys a new application version, updates the configuration settings to an entirely new configuration template, or updates select configuration option values in the running environment.

Attempting to update both the release and configuration is not allowed and AWS Elastic Beanstalk returns an InvalidParameterCombination error.

When updating the configuration settings to a new template or individual settings, a draft configuration is created and DescribeConfigurationSettings for this environment returns two setting descriptions with different DeploymentStatus values.

", + "UpdateTagsForResource":"

Update the list of tags applied to an AWS Elastic Beanstalk resource. Two lists can be passed: TagsToAdd for tags to add or update, and TagsToRemove.

Currently, Elastic Beanstalk only supports tagging of Elastic Beanstalk environments. For details about environment tagging, see Tagging Resources in Your Elastic Beanstalk Environment.

If you create a custom IAM user policy to control permission to this operation, specify one of the following two virtual actions (or both) instead of the API operation name:

elasticbeanstalk:AddTags

Controls permission to call UpdateTagsForResource and pass a list of tags to add in the TagsToAdd parameter.

elasticbeanstalk:RemoveTags

Controls permission to call UpdateTagsForResource and pass a list of tag keys to remove in the TagsToRemove parameter.

For details about creating a custom user policy, see Creating a Custom User Policy.

", + "ValidateConfigurationSettings":"

Takes a set of configuration settings and either a configuration template or environment, and determines whether those values are valid.

This action returns a list of messages indicating any errors or warnings associated with the selection of option values.

" }, - "shapes": { - "ARN": { - "base": null, - "refs": { - "Builder$ARN": "

The ARN of the builder.

" - } - }, - "AbortEnvironmentUpdateMessage": { - "base": "

", - "refs": { - } - }, - "AbortableOperationInProgress": { - "base": null, - "refs": { - "EnvironmentDescription$AbortableOperationInProgress": "

Indicates if there is an in-progress environment configuration update or application version deployment that you can cancel.

true: There is an update in progress.

false: There are no updates currently in progress.

" - } - }, - "ActionHistoryStatus": { - "base": null, - "refs": { - "ManagedActionHistoryItem$Status": "

The status of the action.

" + "shapes":{ + "ARN":{ + "base":null, + "refs":{ + "Builder$ARN":"

The ARN of the builder.

" + } + }, + "AbortEnvironmentUpdateMessage":{ + "base":"

", + "refs":{ + } + }, + "AbortableOperationInProgress":{ + "base":null, + "refs":{ + "EnvironmentDescription$AbortableOperationInProgress":"

Indicates if there is an in-progress environment configuration update or application version deployment that you can cancel.

true: There is an update in progress.

false: There are no updates currently in progress.

" + } + }, + "ActionHistoryStatus":{ + "base":null, + "refs":{ + "ManagedActionHistoryItem$Status":"

The status of the action.

" } - }, - "ActionStatus": { - "base": null, - "refs": { - "DescribeEnvironmentManagedActionsRequest$Status": "

To show only actions with a particular status, specify a status.

", - "ManagedAction$Status": "

The status of the managed action. If the action is Scheduled, you can apply it immediately with ApplyEnvironmentManagedAction.

" + }, + "ActionStatus":{ + "base":null, + "refs":{ + "DescribeEnvironmentManagedActionsRequest$Status":"

To show only actions with a particular status, specify a status.

", + "ManagedAction$Status":"

The status of the managed action. If the action is Scheduled, you can apply it immediately with ApplyEnvironmentManagedAction.

" } }, - "ActionType": { - "base": null, - "refs": { - "ApplyEnvironmentManagedActionResult$ActionType": "

The type of managed action.

", - "ManagedAction$ActionType": "

The type of managed action.

", - "ManagedActionHistoryItem$ActionType": "

The type of the managed action.

" + "ActionType":{ + "base":null, + "refs":{ + "ApplyEnvironmentManagedActionResult$ActionType":"

The type of managed action.

", + "ManagedAction$ActionType":"

The type of managed action.

", + "ManagedActionHistoryItem$ActionType":"

The type of the managed action.

" } }, - "ApplicationArn": { - "base": null, - "refs": { - "ApplicationDescription$ApplicationArn": "

The Amazon Resource Name (ARN) of the application.

" + "ApplicationArn":{ + "base":null, + "refs":{ + "ApplicationDescription$ApplicationArn":"

The Amazon Resource Name (ARN) of the application.

" } }, - "ApplicationDescription": { - "base": "

Describes the properties of an application.

", - "refs": { - "ApplicationDescriptionList$member": null, - "ApplicationDescriptionMessage$Application": "

The ApplicationDescription of the application.

" + "ApplicationDescription":{ + "base":"

Describes the properties of an application.

", + "refs":{ + "ApplicationDescriptionList$member":null, + "ApplicationDescriptionMessage$Application":"

The ApplicationDescription of the application.

" } }, - "ApplicationDescriptionList": { - "base": null, - "refs": { - "ApplicationDescriptionsMessage$Applications": "

This parameter contains a list of ApplicationDescription.

" + "ApplicationDescriptionList":{ + "base":null, + "refs":{ + "ApplicationDescriptionsMessage$Applications":"

This parameter contains a list of ApplicationDescription.

" } }, - "ApplicationDescriptionMessage": { - "base": "

Result message containing a single description of an application.

", - "refs": { + "ApplicationDescriptionMessage":{ + "base":"

Result message containing a single description of an application.

", + "refs":{ } }, - "ApplicationDescriptionsMessage": { - "base": "

Result message containing a list of application descriptions.

", - "refs": { + "ApplicationDescriptionsMessage":{ + "base":"

Result message containing a list of application descriptions.

", + "refs":{ } }, - "ApplicationMetrics": { - "base": "

Application request metrics for an AWS Elastic Beanstalk environment.

", - "refs": { - "DescribeEnvironmentHealthResult$ApplicationMetrics": "

Application request metrics for the environment.

", - "SingleInstanceHealth$ApplicationMetrics": "

Request metrics from your application.

" + "ApplicationMetrics":{ + "base":"

Application request metrics for an AWS Elastic Beanstalk environment.

", + "refs":{ + "DescribeEnvironmentHealthResult$ApplicationMetrics":"

Application request metrics for the environment.

", + "SingleInstanceHealth$ApplicationMetrics":"

Request metrics from your application.

" } }, - "ApplicationName": { - "base": null, - "refs": { - "ApplicationDescription$ApplicationName": "

The name of the application.

", - "ApplicationNamesList$member": null, - "ApplicationResourceLifecycleDescriptionMessage$ApplicationName": "

The name of the application.

", - "ApplicationVersionDescription$ApplicationName": "

The name of the application to which the application version belongs.

", - "ComposeEnvironmentsMessage$ApplicationName": "

The name of the application to which the specified source bundles belong.

", - "ConfigurationSettingsDescription$ApplicationName": "

The name of the application associated with this configuration set.

", - "CreateApplicationMessage$ApplicationName": "

The name of the application.

Constraint: This name must be unique within your account. If the specified name already exists, the action returns an InvalidParameterValue error.

", - "CreateApplicationVersionMessage$ApplicationName": "

The name of the application. If no application is found with this name, and AutoCreateApplication is false, returns an InvalidParameterValue error.

", - "CreateConfigurationTemplateMessage$ApplicationName": "

The name of the application to associate with this configuration template. If no application is found with this name, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", - "CreateEnvironmentMessage$ApplicationName": "

The name of the application that contains the version to be deployed.

If no application is found with this name, CreateEnvironment returns an InvalidParameterValue error.

", - "DeleteApplicationMessage$ApplicationName": "

The name of the application to delete.

", - "DeleteApplicationVersionMessage$ApplicationName": "

The name of the application to which the version belongs.

", - "DeleteConfigurationTemplateMessage$ApplicationName": "

The name of the application to delete the configuration template from.

", - "DeleteEnvironmentConfigurationMessage$ApplicationName": "

The name of the application the environment is associated with.

", - "DescribeApplicationVersionsMessage$ApplicationName": "

Specify an application name to show only application versions for that application.

", - "DescribeConfigurationOptionsMessage$ApplicationName": "

The name of the application associated with the configuration template or environment. Only needed if you want to describe the configuration options associated with either the configuration template or environment.

", - "DescribeConfigurationSettingsMessage$ApplicationName": "

The application for the environment or configuration template.

", - "DescribeEnvironmentsMessage$ApplicationName": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that are associated with this application.

", - "DescribeEventsMessage$ApplicationName": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those associated with this application.

", - "EnvironmentDescription$ApplicationName": "

The name of the application associated with this environment.

", - "EventDescription$ApplicationName": "

The application associated with the event.

", - "SourceConfiguration$ApplicationName": "

The name of the application associated with the configuration.

", - "UpdateApplicationMessage$ApplicationName": "

The name of the application to update. If no such application is found, UpdateApplication returns an InvalidParameterValue error.

", - "UpdateApplicationResourceLifecycleMessage$ApplicationName": "

The name of the application.

", - "UpdateApplicationVersionMessage$ApplicationName": "

The name of the application associated with this version.

If no application is found with this name, UpdateApplication returns an InvalidParameterValue error.

", - "UpdateConfigurationTemplateMessage$ApplicationName": "

The name of the application associated with the configuration template to update.

If no application is found with this name, UpdateConfigurationTemplate returns an InvalidParameterValue error.

", - "UpdateEnvironmentMessage$ApplicationName": "

The name of the application with which the environment is associated.

", - "ValidateConfigurationSettingsMessage$ApplicationName": "

The name of the application that the configuration template or environment belongs to.

" + "ApplicationName":{ + "base":null, + "refs":{ + "ApplicationDescription$ApplicationName":"

The name of the application.

", + "ApplicationNamesList$member":null, + "ApplicationResourceLifecycleDescriptionMessage$ApplicationName":"

The name of the application.

", + "ApplicationVersionDescription$ApplicationName":"

The name of the application to which the application version belongs.

", + "ComposeEnvironmentsMessage$ApplicationName":"

The name of the application to which the specified source bundles belong.

", + "ConfigurationSettingsDescription$ApplicationName":"

The name of the application associated with this configuration set.

", + "CreateApplicationMessage$ApplicationName":"

The name of the application.

Constraint: This name must be unique within your account. If the specified name already exists, the action returns an InvalidParameterValue error.

", + "CreateApplicationVersionMessage$ApplicationName":"

The name of the application. If no application is found with this name, and AutoCreateApplication is false, returns an InvalidParameterValue error.

", + "CreateConfigurationTemplateMessage$ApplicationName":"

The name of the application to associate with this configuration template. If no application is found with this name, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", + "CreateEnvironmentMessage$ApplicationName":"

The name of the application that contains the version to be deployed.

If no application is found with this name, CreateEnvironment returns an InvalidParameterValue error.

", + "DeleteApplicationMessage$ApplicationName":"

The name of the application to delete.

", + "DeleteApplicationVersionMessage$ApplicationName":"

The name of the application to which the version belongs.

", + "DeleteConfigurationTemplateMessage$ApplicationName":"

The name of the application to delete the configuration template from.

", + "DeleteEnvironmentConfigurationMessage$ApplicationName":"

The name of the application the environment is associated with.

", + "DescribeApplicationVersionsMessage$ApplicationName":"

Specify an application name to show only application versions for that application.

", + "DescribeConfigurationOptionsMessage$ApplicationName":"

The name of the application associated with the configuration template or environment. Only needed if you want to describe the configuration options associated with either the configuration template or environment.

", + "DescribeConfigurationSettingsMessage$ApplicationName":"

The application for the environment or configuration template.

", + "DescribeEnvironmentsMessage$ApplicationName":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that are associated with this application.

", + "DescribeEventsMessage$ApplicationName":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those associated with this application.

", + "EnvironmentDescription$ApplicationName":"

The name of the application associated with this environment.

", + "EventDescription$ApplicationName":"

The application associated with the event.

", + "SourceConfiguration$ApplicationName":"

The name of the application associated with the configuration.

", + "UpdateApplicationMessage$ApplicationName":"

The name of the application to update. If no such application is found, UpdateApplication returns an InvalidParameterValue error.

", + "UpdateApplicationResourceLifecycleMessage$ApplicationName":"

The name of the application.

", + "UpdateApplicationVersionMessage$ApplicationName":"

The name of the application associated with this version.

If no application is found with this name, UpdateApplication returns an InvalidParameterValue error.

", + "UpdateConfigurationTemplateMessage$ApplicationName":"

The name of the application associated with the configuration template to update.

If no application is found with this name, UpdateConfigurationTemplate returns an InvalidParameterValue error.

", + "UpdateEnvironmentMessage$ApplicationName":"

The name of the application with which the environment is associated.

", + "ValidateConfigurationSettingsMessage$ApplicationName":"

The name of the application that the configuration template or environment belongs to.

" } }, - "ApplicationNamesList": { - "base": null, - "refs": { - "DescribeApplicationsMessage$ApplicationNames": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to only include those with the specified names.

" + "ApplicationNamesList":{ + "base":null, + "refs":{ + "DescribeApplicationsMessage$ApplicationNames":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to only include those with the specified names.

" } }, - "ApplicationResourceLifecycleConfig": { - "base": "

The resource lifecycle configuration for an application. Defines lifecycle settings for resources that belong to the application, and the service role that Elastic Beanstalk assumes in order to apply lifecycle settings. The version lifecycle configuration defines lifecycle settings for application versions.

", - "refs": { - "ApplicationDescription$ResourceLifecycleConfig": "

The lifecycle settings for the application.

", - "ApplicationResourceLifecycleDescriptionMessage$ResourceLifecycleConfig": "

The lifecycle configuration.

", - "CreateApplicationMessage$ResourceLifecycleConfig": "

Specify an application resource lifecycle configuration to prevent your application from accumulating too many versions.

", - "UpdateApplicationResourceLifecycleMessage$ResourceLifecycleConfig": "

The lifecycle configuration.

" + "ApplicationResourceLifecycleConfig":{ + "base":"

The resource lifecycle configuration for an application. Defines lifecycle settings for resources that belong to the application, and the service role that Elastic Beanstalk assumes in order to apply lifecycle settings. The version lifecycle configuration defines lifecycle settings for application versions.

", + "refs":{ + "ApplicationDescription$ResourceLifecycleConfig":"

The lifecycle settings for the application.

", + "ApplicationResourceLifecycleDescriptionMessage$ResourceLifecycleConfig":"

The lifecycle configuration.

", + "CreateApplicationMessage$ResourceLifecycleConfig":"

Specify an application resource lifecycle configuration to prevent your application from accumulating too many versions.

", + "UpdateApplicationResourceLifecycleMessage$ResourceLifecycleConfig":"

The lifecycle configuration.

" } }, - "ApplicationResourceLifecycleDescriptionMessage": { - "base": null, - "refs": { + "ApplicationResourceLifecycleDescriptionMessage":{ + "base":null, + "refs":{ } }, - "ApplicationVersionArn": { - "base": null, - "refs": { - "ApplicationVersionDescription$ApplicationVersionArn": "

The Amazon Resource Name (ARN) of the application version.

" + "ApplicationVersionArn":{ + "base":null, + "refs":{ + "ApplicationVersionDescription$ApplicationVersionArn":"

The Amazon Resource Name (ARN) of the application version.

" } }, - "ApplicationVersionDescription": { - "base": "

Describes the properties of an application version.

", - "refs": { - "ApplicationVersionDescriptionList$member": null, - "ApplicationVersionDescriptionMessage$ApplicationVersion": "

The ApplicationVersionDescription of the application version.

" + "ApplicationVersionDescription":{ + "base":"

Describes the properties of an application version.

", + "refs":{ + "ApplicationVersionDescriptionList$member":null, + "ApplicationVersionDescriptionMessage$ApplicationVersion":"

The ApplicationVersionDescription of the application version.

" } }, - "ApplicationVersionDescriptionList": { - "base": null, - "refs": { - "ApplicationVersionDescriptionsMessage$ApplicationVersions": "

List of ApplicationVersionDescription objects sorted in order of creation.

" + "ApplicationVersionDescriptionList":{ + "base":null, + "refs":{ + "ApplicationVersionDescriptionsMessage$ApplicationVersions":"

List of ApplicationVersionDescription objects sorted in order of creation.

" } }, - "ApplicationVersionDescriptionMessage": { - "base": "

Result message wrapping a single description of an application version.

", - "refs": { + "ApplicationVersionDescriptionMessage":{ + "base":"

Result message wrapping a single description of an application version.

", + "refs":{ } }, - "ApplicationVersionDescriptionsMessage": { - "base": "

Result message wrapping a list of application version descriptions.

", - "refs": { + "ApplicationVersionDescriptionsMessage":{ + "base":"

Result message wrapping a list of application version descriptions.

", + "refs":{ } }, - "ApplicationVersionLifecycleConfig": { - "base": "

The application version lifecycle settings for an application. Defines the rules that Elastic Beanstalk applies to an application's versions in order to avoid hitting the per-region limit for application versions.

When Elastic Beanstalk deletes an application version from its database, you can no longer deploy that version to an environment. The source bundle remains in S3 unless you configure the rule to delete it.

", - "refs": { - "ApplicationResourceLifecycleConfig$VersionLifecycleConfig": "

The application version lifecycle configuration.

" + "ApplicationVersionLifecycleConfig":{ + "base":"

The application version lifecycle settings for an application. Defines the rules that Elastic Beanstalk applies to an application's versions in order to avoid hitting the per-region limit for application versions.

When Elastic Beanstalk deletes an application version from its database, you can no longer deploy that version to an environment. The source bundle remains in S3 unless you configure the rule to delete it.

", + "refs":{ + "ApplicationResourceLifecycleConfig$VersionLifecycleConfig":"

The application version lifecycle configuration.

" } }, - "ApplicationVersionProccess": { - "base": null, - "refs": { - "CreateApplicationVersionMessage$Process": "

Preprocesses and validates the environment manifest (env.yaml) and configuration files (*.config files in the .ebextensions folder) in the source bundle. Validating configuration files can identify issues prior to deploying the application version to an environment.

The Process option validates Elastic Beanstalk configuration files. It doesn't validate your application's configuration files, like proxy server or Docker configuration.

" + "ApplicationVersionProccess":{ + "base":null, + "refs":{ + "CreateApplicationVersionMessage$Process":"

Pre-processes and validates the environment manifest (env.yaml) and configuration files (*.config files in the .ebextensions folder) in the source bundle. Validating configuration files can identify issues prior to deploying the application version to an environment.

You must turn processing on for application versions that you create using AWS CodeBuild or AWS CodeCommit. For application versions built from a source bundle in Amazon S3, processing is optional.

The Process option validates Elastic Beanstalk configuration files. It doesn't validate your application's configuration files, like proxy server or Docker configuration.

" } }, - "ApplicationVersionStatus": { - "base": null, - "refs": { - "ApplicationVersionDescription$Status": "

The processing status of the application version.

" + "ApplicationVersionStatus":{ + "base":null, + "refs":{ + "ApplicationVersionDescription$Status":"

The processing status of the application version. Reflects the state of the application version during its creation. Many of the values are only applicable if you specified True for the Process parameter of the CreateApplicationVersion action. The following list describes the possible values.

  • Unprocessed – Application version wasn't pre-processed or validated. Elastic Beanstalk will validate configuration files during deployment of the application version to an environment.

  • Processing – Elastic Beanstalk is currently processing the application version.

  • Building – Application version is currently undergoing an AWS CodeBuild build.

  • Processed – Elastic Beanstalk was successfully pre-processed and validated.

  • Failed – Either the AWS CodeBuild build failed or configuration files didn't pass validation. This application version isn't usable.

" } }, - "ApplyEnvironmentManagedActionRequest": { - "base": "

Request to execute a scheduled managed action immediately.

", - "refs": { + "ApplyEnvironmentManagedActionRequest":{ + "base":"

Request to execute a scheduled managed action immediately.

", + "refs":{ } }, - "ApplyEnvironmentManagedActionResult": { - "base": "

The result message containing information about the managed action.

", - "refs": { + "ApplyEnvironmentManagedActionResult":{ + "base":"

The result message containing information about the managed action.

", + "refs":{ } }, - "AutoCreateApplication": { - "base": null, - "refs": { - "CreateApplicationVersionMessage$AutoCreateApplication": "

Set to true to create an application with the specified name if it doesn't already exist.

" + "AutoCreateApplication":{ + "base":null, + "refs":{ + "CreateApplicationVersionMessage$AutoCreateApplication":"

Set to true to create an application with the specified name if it doesn't already exist.

" } }, - "AutoScalingGroup": { - "base": "

Describes an Auto Scaling launch configuration.

", - "refs": { - "AutoScalingGroupList$member": null + "AutoScalingGroup":{ + "base":"

Describes an Auto Scaling launch configuration.

", + "refs":{ + "AutoScalingGroupList$member":null } }, - "AutoScalingGroupList": { - "base": null, - "refs": { - "EnvironmentResourceDescription$AutoScalingGroups": "

The AutoScalingGroups used by this environment.

" + "AutoScalingGroupList":{ + "base":null, + "refs":{ + "EnvironmentResourceDescription$AutoScalingGroups":"

The AutoScalingGroups used by this environment.

" } }, - "AvailableSolutionStackDetailsList": { - "base": null, - "refs": { - "ListAvailableSolutionStacksResultMessage$SolutionStackDetails": "

A list of available solution stacks and their SolutionStackDescription.

" + "AvailableSolutionStackDetailsList":{ + "base":null, + "refs":{ + "ListAvailableSolutionStacksResultMessage$SolutionStackDetails":"

A list of available solution stacks and their SolutionStackDescription.

" } }, - "AvailableSolutionStackNamesList": { - "base": null, - "refs": { - "ListAvailableSolutionStacksResultMessage$SolutionStacks": "

A list of available solution stacks.

" + "AvailableSolutionStackNamesList":{ + "base":null, + "refs":{ + "ListAvailableSolutionStacksResultMessage$SolutionStacks":"

A list of available solution stacks.

" } }, - "BoxedBoolean": { - "base": null, - "refs": { - "MaxAgeRule$Enabled": "

Specify true to apply the rule, or false to disable it.

", - "MaxAgeRule$DeleteSourceFromS3": "

Set to true to delete a version's source bundle from Amazon S3 when Elastic Beanstalk deletes the application version.

", - "MaxCountRule$Enabled": "

Specify true to apply the rule, or false to disable it.

", - "MaxCountRule$DeleteSourceFromS3": "

Set to true to delete a version's source bundle from Amazon S3 when Elastic Beanstalk deletes the application version.

" + "BoxedBoolean":{ + "base":null, + "refs":{ + "MaxAgeRule$Enabled":"

Specify true to apply the rule, or false to disable it.

", + "MaxAgeRule$DeleteSourceFromS3":"

Set to true to delete a version's source bundle from Amazon S3 when Elastic Beanstalk deletes the application version.

", + "MaxCountRule$Enabled":"

Specify true to apply the rule, or false to disable it.

", + "MaxCountRule$DeleteSourceFromS3":"

Set to true to delete a version's source bundle from Amazon S3 when Elastic Beanstalk deletes the application version.

" } }, - "BoxedInt": { - "base": null, - "refs": { - "BuildConfiguration$TimeoutInMinutes": "

How long in minutes, from 5 to 480 (8 hours), for AWS CodeBuild to wait until timing out any related build that does not get marked as completed. The default is 60 minutes.

", - "MaxAgeRule$MaxAgeInDays": "

Specify the number of days to retain an application versions.

", - "MaxCountRule$MaxCount": "

Specify the maximum number of application versions to retain.

", - "ResourceQuota$Maximum": "

The maximum number of instances of this Elastic Beanstalk resource type that an AWS account can use.

" + "BoxedInt":{ + "base":null, + "refs":{ + "BuildConfiguration$TimeoutInMinutes":"

How long in minutes, from 5 to 480 (8 hours), for AWS CodeBuild to wait until timing out any related build that does not get marked as completed. The default is 60 minutes.

", + "MaxAgeRule$MaxAgeInDays":"

Specify the number of days to retain an application versions.

", + "MaxCountRule$MaxCount":"

Specify the maximum number of application versions to retain.

", + "ResourceQuota$Maximum":"

The maximum number of instances of this Elastic Beanstalk resource type that an AWS account can use.

" } }, - "BuildConfiguration": { - "base": "

Settings for an AWS CodeBuild build.

", - "refs": { - "CreateApplicationVersionMessage$BuildConfiguration": "

Settings for an AWS CodeBuild build.

" + "BuildConfiguration":{ + "base":"

Settings for an AWS CodeBuild build.

", + "refs":{ + "CreateApplicationVersionMessage$BuildConfiguration":"

Settings for an AWS CodeBuild build.

" } }, - "Builder": { - "base": "

The builder used to build the custom platform.

", - "refs": { - "CreatePlatformVersionResult$Builder": "

The builder used to create the custom platform.

" + "Builder":{ + "base":"

The builder used to build the custom platform.

", + "refs":{ + "CreatePlatformVersionResult$Builder":"

The builder used to create the custom platform.

" } }, - "CPUUtilization": { - "base": "

CPU utilization metrics for an instance.

", - "refs": { - "SystemStatus$CPUUtilization": "

CPU utilization metrics for the instance.

" + "CPUUtilization":{ + "base":"

CPU utilization metrics for an instance.

", + "refs":{ + "SystemStatus$CPUUtilization":"

CPU utilization metrics for the instance.

" } }, - "Cause": { - "base": null, - "refs": { - "Causes$member": null + "Cause":{ + "base":null, + "refs":{ + "Causes$member":null } }, - "Causes": { - "base": null, - "refs": { - "DescribeEnvironmentHealthResult$Causes": "

Descriptions of the data that contributed to the environment's current health status.

", - "SingleInstanceHealth$Causes": "

Represents the causes, which provide more information about the current health status.

" + "Causes":{ + "base":null, + "refs":{ + "DescribeEnvironmentHealthResult$Causes":"

Descriptions of the data that contributed to the environment's current health status.

", + "SingleInstanceHealth$Causes":"

Represents the causes, which provide more information about the current health status.

" } }, - "CheckDNSAvailabilityMessage": { - "base": "

Results message indicating whether a CNAME is available.

", - "refs": { + "CheckDNSAvailabilityMessage":{ + "base":"

Results message indicating whether a CNAME is available.

", + "refs":{ } }, - "CheckDNSAvailabilityResultMessage": { - "base": "

Indicates if the specified CNAME is available.

", - "refs": { + "CheckDNSAvailabilityResultMessage":{ + "base":"

Indicates if the specified CNAME is available.

", + "refs":{ } }, - "CnameAvailability": { - "base": null, - "refs": { - "CheckDNSAvailabilityResultMessage$Available": "

Indicates if the specified CNAME is available:

  • true : The CNAME is available.

  • false : The CNAME is not available.

" + "CnameAvailability":{ + "base":null, + "refs":{ + "CheckDNSAvailabilityResultMessage$Available":"

Indicates if the specified CNAME is available:

  • true : The CNAME is available.

  • false : The CNAME is not available.

" } }, - "CodeBuildNotInServiceRegionException": { - "base": "

AWS CodeBuild is not available in the specified region.

", - "refs": { + "CodeBuildNotInServiceRegionException":{ + "base":"

AWS CodeBuild is not available in the specified region.

", + "refs":{ } }, - "ComposeEnvironmentsMessage": { - "base": "

Request to create or update a group of environments.

", - "refs": { + "ComposeEnvironmentsMessage":{ + "base":"

Request to create or update a group of environments.

", + "refs":{ } }, - "ComputeType": { - "base": null, - "refs": { - "BuildConfiguration$ComputeType": "

Information about the compute resources the build project will use.

  • BUILD_GENERAL1_SMALL: Use up to 3 GB memory and 2 vCPUs for builds

  • BUILD_GENERAL1_MEDIUM: Use up to 7 GB memory and 4 vCPUs for builds

  • BUILD_GENERAL1_LARGE: Use up to 15 GB memory and 8 vCPUs for builds

" + "ComputeType":{ + "base":null, + "refs":{ + "BuildConfiguration$ComputeType":"

Information about the compute resources the build project will use.

  • BUILD_GENERAL1_SMALL: Use up to 3 GB memory and 2 vCPUs for builds

  • BUILD_GENERAL1_MEDIUM: Use up to 7 GB memory and 4 vCPUs for builds

  • BUILD_GENERAL1_LARGE: Use up to 15 GB memory and 8 vCPUs for builds

" } }, - "ConfigurationDeploymentStatus": { - "base": null, - "refs": { - "ConfigurationSettingsDescription$DeploymentStatus": "

If this configuration set is associated with an environment, the DeploymentStatus parameter indicates the deployment status of this configuration set:

  • null: This configuration is not associated with a running environment.

  • pending: This is a draft configuration that is not deployed to the associated environment but is in the process of deploying.

  • deployed: This is the configuration that is currently deployed to the associated running environment.

  • failed: This is a draft configuration that failed to successfully deploy.

" + "ConfigurationDeploymentStatus":{ + "base":null, + "refs":{ + "ConfigurationSettingsDescription$DeploymentStatus":"

If this configuration set is associated with an environment, the DeploymentStatus parameter indicates the deployment status of this configuration set:

  • null: This configuration is not associated with a running environment.

  • pending: This is a draft configuration that is not deployed to the associated environment but is in the process of deploying.

  • deployed: This is the configuration that is currently deployed to the associated running environment.

  • failed: This is a draft configuration that failed to successfully deploy.

" } }, - "ConfigurationOptionDefaultValue": { - "base": null, - "refs": { - "ConfigurationOptionDescription$DefaultValue": "

The default value for this configuration option.

" + "ConfigurationOptionDefaultValue":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$DefaultValue":"

The default value for this configuration option.

" } }, - "ConfigurationOptionDescription": { - "base": "

Describes the possible values for a configuration option.

", - "refs": { - "ConfigurationOptionDescriptionsList$member": null + "ConfigurationOptionDescription":{ + "base":"

Describes the possible values for a configuration option.

", + "refs":{ + "ConfigurationOptionDescriptionsList$member":null } }, - "ConfigurationOptionDescriptionsList": { - "base": null, - "refs": { - "ConfigurationOptionsDescription$Options": "

A list of ConfigurationOptionDescription.

" + "ConfigurationOptionDescriptionsList":{ + "base":null, + "refs":{ + "ConfigurationOptionsDescription$Options":"

A list of ConfigurationOptionDescription.

" } }, - "ConfigurationOptionName": { - "base": null, - "refs": { - "ConfigurationOptionDescription$Name": "

The name of the configuration option.

", - "ConfigurationOptionSetting$OptionName": "

The name of the configuration option.

", - "OptionSpecification$OptionName": "

The name of the configuration option.

", - "ValidationMessage$OptionName": "

The name of the option.

" + "ConfigurationOptionName":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$Name":"

The name of the configuration option.

", + "ConfigurationOptionSetting$OptionName":"

The name of the configuration option.

", + "OptionSpecification$OptionName":"

The name of the configuration option.

", + "ValidationMessage$OptionName":"

The name of the option.

" } }, - "ConfigurationOptionPossibleValue": { - "base": null, - "refs": { - "ConfigurationOptionPossibleValues$member": null + "ConfigurationOptionPossibleValue":{ + "base":null, + "refs":{ + "ConfigurationOptionPossibleValues$member":null } }, - "ConfigurationOptionPossibleValues": { - "base": null, - "refs": { - "ConfigurationOptionDescription$ValueOptions": "

If specified, values for the configuration option are selected from this list.

" + "ConfigurationOptionPossibleValues":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$ValueOptions":"

If specified, values for the configuration option are selected from this list.

" } }, - "ConfigurationOptionSetting": { - "base": "

A specification identifying an individual configuration option along with its current value. For a list of possible option values, go to Option Values in the AWS Elastic Beanstalk Developer Guide.

", - "refs": { - "ConfigurationOptionSettingsList$member": null + "ConfigurationOptionSetting":{ + "base":"

A specification identifying an individual configuration option along with its current value. For a list of possible option values, go to Option Values in the AWS Elastic Beanstalk Developer Guide.

", + "refs":{ + "ConfigurationOptionSettingsList$member":null } }, - "ConfigurationOptionSettingsList": { - "base": null, - "refs": { - "ConfigurationSettingsDescription$OptionSettings": "

A list of the configuration options and their values in this configuration set.

", - "CreateConfigurationTemplateMessage$OptionSettings": "

If specified, AWS Elastic Beanstalk sets the specified configuration option to the requested value. The new value overrides the value obtained from the solution stack or the source configuration template.

", - "CreateEnvironmentMessage$OptionSettings": "

If specified, AWS Elastic Beanstalk sets the specified configuration options to the requested value in the configuration set for the new environment. These override the values obtained from the solution stack or the configuration template.

", - "CreatePlatformVersionRequest$OptionSettings": "

The configuration option settings to apply to the builder environment.

", - "UpdateConfigurationTemplateMessage$OptionSettings": "

A list of configuration option settings to update with the new specified option value.

", - "UpdateEnvironmentMessage$OptionSettings": "

If specified, AWS Elastic Beanstalk updates the configuration set associated with the running environment and sets the specified configuration options to the requested value.

", - "ValidateConfigurationSettingsMessage$OptionSettings": "

A list of the options and desired values to evaluate.

" + "ConfigurationOptionSettingsList":{ + "base":null, + "refs":{ + "ConfigurationSettingsDescription$OptionSettings":"

A list of the configuration options and their values in this configuration set.

", + "CreateConfigurationTemplateMessage$OptionSettings":"

If specified, AWS Elastic Beanstalk sets the specified configuration option to the requested value. The new value overrides the value obtained from the solution stack or the source configuration template.

", + "CreateEnvironmentMessage$OptionSettings":"

If specified, AWS Elastic Beanstalk sets the specified configuration options to the requested value in the configuration set for the new environment. These override the values obtained from the solution stack or the configuration template.

", + "CreatePlatformVersionRequest$OptionSettings":"

The configuration option settings to apply to the builder environment.

", + "UpdateConfigurationTemplateMessage$OptionSettings":"

A list of configuration option settings to update with the new specified option value.

", + "UpdateEnvironmentMessage$OptionSettings":"

If specified, AWS Elastic Beanstalk updates the configuration set associated with the running environment and sets the specified configuration options to the requested value.

", + "ValidateConfigurationSettingsMessage$OptionSettings":"

A list of the options and desired values to evaluate.

" } }, - "ConfigurationOptionSeverity": { - "base": null, - "refs": { - "ConfigurationOptionDescription$ChangeSeverity": "

An indication of which action is required if the value for this configuration option changes:

  • NoInterruption : There is no interruption to the environment or application availability.

  • RestartEnvironment : The environment is entirely restarted, all AWS resources are deleted and recreated, and the environment is unavailable during the process.

  • RestartApplicationServer : The environment is available the entire time. However, a short application outage occurs when the application servers on the running Amazon EC2 instances are restarted.

" + "ConfigurationOptionSeverity":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$ChangeSeverity":"

An indication of which action is required if the value for this configuration option changes:

  • NoInterruption : There is no interruption to the environment or application availability.

  • RestartEnvironment : The environment is entirely restarted, all AWS resources are deleted and recreated, and the environment is unavailable during the process.

  • RestartApplicationServer : The environment is available the entire time. However, a short application outage occurs when the application servers on the running Amazon EC2 instances are restarted.

" } }, - "ConfigurationOptionValue": { - "base": null, - "refs": { - "ConfigurationOptionSetting$Value": "

The current value for the configuration option.

" + "ConfigurationOptionValue":{ + "base":null, + "refs":{ + "ConfigurationOptionSetting$Value":"

The current value for the configuration option.

" } }, - "ConfigurationOptionValueType": { - "base": null, - "refs": { - "ConfigurationOptionDescription$ValueType": "

An indication of which type of values this option has and whether it is allowable to select one or more than one of the possible values:

  • Scalar : Values for this option are a single selection from the possible values, or an unformatted string, or numeric value governed by the MIN/MAX/Regex constraints.

  • List : Values for this option are multiple selections from the possible values.

  • Boolean : Values for this option are either true or false .

  • Json : Values for this option are a JSON representation of a ConfigDocument.

" + "ConfigurationOptionValueType":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$ValueType":"

An indication of which type of values this option has and whether it is allowable to select one or more than one of the possible values:

  • Scalar : Values for this option are a single selection from the possible values, or an unformatted string, or numeric value governed by the MIN/MAX/Regex constraints.

  • List : Values for this option are multiple selections from the possible values.

  • Boolean : Values for this option are either true or false .

  • Json : Values for this option are a JSON representation of a ConfigDocument.

" } }, - "ConfigurationOptionsDescription": { - "base": "

Describes the settings for a specified configuration set.

", - "refs": { + "ConfigurationOptionsDescription":{ + "base":"

Describes the settings for a specified configuration set.

", + "refs":{ } }, - "ConfigurationSettingsDescription": { - "base": "

Describes the settings for a configuration set.

", - "refs": { - "ConfigurationSettingsDescriptionList$member": null + "ConfigurationSettingsDescription":{ + "base":"

Describes the settings for a configuration set.

", + "refs":{ + "ConfigurationSettingsDescriptionList$member":null } }, - "ConfigurationSettingsDescriptionList": { - "base": null, - "refs": { - "ConfigurationSettingsDescriptions$ConfigurationSettings": "

A list of ConfigurationSettingsDescription.

" + "ConfigurationSettingsDescriptionList":{ + "base":null, + "refs":{ + "ConfigurationSettingsDescriptions$ConfigurationSettings":"

A list of ConfigurationSettingsDescription.

" } }, - "ConfigurationSettingsDescriptions": { - "base": "

The results from a request to change the configuration settings of an environment.

", - "refs": { + "ConfigurationSettingsDescriptions":{ + "base":"

The results from a request to change the configuration settings of an environment.

", + "refs":{ } }, - "ConfigurationSettingsValidationMessages": { - "base": "

Provides a list of validation messages.

", - "refs": { + "ConfigurationSettingsValidationMessages":{ + "base":"

Provides a list of validation messages.

", + "refs":{ } }, - "ConfigurationTemplateName": { - "base": null, - "refs": { - "ConfigurationSettingsDescription$TemplateName": "

If not null, the name of the configuration template for this configuration set.

", - "ConfigurationTemplateNamesList$member": null, - "CreateConfigurationTemplateMessage$TemplateName": "

The name of the configuration template.

Constraint: This name must be unique per application.

Default: If a configuration template already exists with this name, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", - "CreateEnvironmentMessage$TemplateName": "

The name of the configuration template to use in deployment. If no configuration template is found with this name, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", - "DeleteConfigurationTemplateMessage$TemplateName": "

The name of the configuration template to delete.

", - "DescribeConfigurationOptionsMessage$TemplateName": "

The name of the configuration template whose configuration options you want to describe.

", - "DescribeConfigurationSettingsMessage$TemplateName": "

The name of the configuration template to describe.

Conditional: You must specify either this parameter or an EnvironmentName, but not both. If you specify both, AWS Elastic Beanstalk returns an InvalidParameterCombination error. If you do not specify either, AWS Elastic Beanstalk returns a MissingRequiredParameter error.

", - "DescribeEventsMessage$TemplateName": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those that are associated with this environment configuration.

", - "EnvironmentDescription$TemplateName": "

The name of the configuration template used to originally launch this environment.

", - "EventDescription$TemplateName": "

The name of the configuration associated with this event.

", - "SourceConfiguration$TemplateName": "

The name of the configuration template.

", - "UpdateConfigurationTemplateMessage$TemplateName": "

The name of the configuration template to update.

If no configuration template is found with this name, UpdateConfigurationTemplate returns an InvalidParameterValue error.

", - "UpdateEnvironmentMessage$TemplateName": "

If this parameter is specified, AWS Elastic Beanstalk deploys this configuration template to the environment. If no such configuration template is found, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", - "ValidateConfigurationSettingsMessage$TemplateName": "

The name of the configuration template to validate the settings against.

Condition: You cannot specify both this and an environment name.

" + "ConfigurationTemplateName":{ + "base":null, + "refs":{ + "ConfigurationSettingsDescription$TemplateName":"

If not null, the name of the configuration template for this configuration set.

", + "ConfigurationTemplateNamesList$member":null, + "CreateConfigurationTemplateMessage$TemplateName":"

The name of the configuration template.

Constraint: This name must be unique per application.

Default: If a configuration template already exists with this name, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", + "CreateEnvironmentMessage$TemplateName":"

The name of the configuration template to use in deployment. If no configuration template is found with this name, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", + "DeleteConfigurationTemplateMessage$TemplateName":"

The name of the configuration template to delete.

", + "DescribeConfigurationOptionsMessage$TemplateName":"

The name of the configuration template whose configuration options you want to describe.

", + "DescribeConfigurationSettingsMessage$TemplateName":"

The name of the configuration template to describe.

Conditional: You must specify either this parameter or an EnvironmentName, but not both. If you specify both, AWS Elastic Beanstalk returns an InvalidParameterCombination error. If you do not specify either, AWS Elastic Beanstalk returns a MissingRequiredParameter error.

", + "DescribeEventsMessage$TemplateName":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those that are associated with this environment configuration.

", + "EnvironmentDescription$TemplateName":"

The name of the configuration template used to originally launch this environment.

", + "EventDescription$TemplateName":"

The name of the configuration associated with this event.

", + "SourceConfiguration$TemplateName":"

The name of the configuration template.

", + "UpdateConfigurationTemplateMessage$TemplateName":"

The name of the configuration template to update.

If no configuration template is found with this name, UpdateConfigurationTemplate returns an InvalidParameterValue error.

", + "UpdateEnvironmentMessage$TemplateName":"

If this parameter is specified, AWS Elastic Beanstalk deploys this configuration template to the environment. If no such configuration template is found, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", + "ValidateConfigurationSettingsMessage$TemplateName":"

The name of the configuration template to validate the settings against.

Condition: You cannot specify both this and an environment name.

" } }, - "ConfigurationTemplateNamesList": { - "base": null, - "refs": { - "ApplicationDescription$ConfigurationTemplates": "

The names of the configuration templates associated with this application.

" + "ConfigurationTemplateNamesList":{ + "base":null, + "refs":{ + "ApplicationDescription$ConfigurationTemplates":"

The names of the configuration templates associated with this application.

" } }, - "CreateApplicationMessage": { - "base": "

Request to create an application.

", - "refs": { + "CreateApplicationMessage":{ + "base":"

Request to create an application.

", + "refs":{ } }, - "CreateApplicationVersionMessage": { - "base": "

", - "refs": { + "CreateApplicationVersionMessage":{ + "base":"

", + "refs":{ } }, - "CreateConfigurationTemplateMessage": { - "base": "

Request to create a configuration template.

", - "refs": { + "CreateConfigurationTemplateMessage":{ + "base":"

Request to create a configuration template.

", + "refs":{ } }, - "CreateEnvironmentMessage": { - "base": "

", - "refs": { + "CreateEnvironmentMessage":{ + "base":"

", + "refs":{ } }, - "CreatePlatformVersionRequest": { - "base": "

Request to create a new platform version.

", - "refs": { + "CreatePlatformVersionRequest":{ + "base":"

Request to create a new platform version.

", + "refs":{ } }, - "CreatePlatformVersionResult": { - "base": null, - "refs": { + "CreatePlatformVersionResult":{ + "base":null, + "refs":{ } }, - "CreateStorageLocationResultMessage": { - "base": "

Results of a CreateStorageLocationResult call.

", - "refs": { + "CreateStorageLocationResultMessage":{ + "base":"

Results of a CreateStorageLocationResult call.

", + "refs":{ } }, - "CreationDate": { - "base": null, - "refs": { - "ApplicationDescription$DateCreated": "

The date when the application was created.

", - "ApplicationVersionDescription$DateCreated": "

The creation date of the application version.

", - "ConfigurationSettingsDescription$DateCreated": "

The date (in UTC time) when this configuration set was created.

", - "EnvironmentDescription$DateCreated": "

The creation date for this environment.

", - "PlatformDescription$DateCreated": "

The date when the platform was created.

" + "CreationDate":{ + "base":null, + "refs":{ + "ApplicationDescription$DateCreated":"

The date when the application was created.

", + "ApplicationVersionDescription$DateCreated":"

The creation date of the application version.

", + "ConfigurationSettingsDescription$DateCreated":"

The date (in UTC time) when this configuration set was created.

", + "EnvironmentDescription$DateCreated":"

The creation date for this environment.

", + "PlatformDescription$DateCreated":"

The date when the platform was created.

" } }, - "CustomAmi": { - "base": "

A custom AMI available to platforms.

", - "refs": { - "CustomAmiList$member": null + "CustomAmi":{ + "base":"

A custom AMI available to platforms.

", + "refs":{ + "CustomAmiList$member":null } }, - "CustomAmiList": { - "base": null, - "refs": { - "PlatformDescription$CustomAmiList": "

The custom AMIs supported by the platform.

" + "CustomAmiList":{ + "base":null, + "refs":{ + "PlatformDescription$CustomAmiList":"

The custom AMIs supported by the platform.

" } }, - "DNSCname": { - "base": null, - "refs": { - "CheckDNSAvailabilityResultMessage$FullyQualifiedCNAME": "

The fully qualified CNAME to reserve when CreateEnvironment is called with the provided prefix.

", - "EnvironmentDescription$CNAME": "

The URL to the CNAME for this environment.

" + "DNSCname":{ + "base":null, + "refs":{ + "CheckDNSAvailabilityResultMessage$FullyQualifiedCNAME":"

The fully qualified CNAME to reserve when CreateEnvironment is called with the provided prefix.

", + "EnvironmentDescription$CNAME":"

The URL to the CNAME for this environment.

" } }, - "DNSCnamePrefix": { - "base": null, - "refs": { - "CheckDNSAvailabilityMessage$CNAMEPrefix": "

The prefix used when this CNAME is reserved.

", - "CreateEnvironmentMessage$CNAMEPrefix": "

If specified, the environment attempts to use this value as the prefix for the CNAME. If not specified, the CNAME is generated automatically by appending a random alphanumeric string to the environment name.

" + "DNSCnamePrefix":{ + "base":null, + "refs":{ + "CheckDNSAvailabilityMessage$CNAMEPrefix":"

The prefix used when this CNAME is reserved.

", + "CreateEnvironmentMessage$CNAMEPrefix":"

If specified, the environment attempts to use this value as the prefix for the CNAME. If not specified, the CNAME is generated automatically by appending a random alphanumeric string to the environment name.

" } }, - "DeleteApplicationMessage": { - "base": "

Request to delete an application.

", - "refs": { + "DeleteApplicationMessage":{ + "base":"

Request to delete an application.

", + "refs":{ } }, - "DeleteApplicationVersionMessage": { - "base": "

Request to delete an application version.

", - "refs": { + "DeleteApplicationVersionMessage":{ + "base":"

Request to delete an application version.

", + "refs":{ } }, - "DeleteConfigurationTemplateMessage": { - "base": "

Request to delete a configuration template.

", - "refs": { + "DeleteConfigurationTemplateMessage":{ + "base":"

Request to delete a configuration template.

", + "refs":{ } }, - "DeleteEnvironmentConfigurationMessage": { - "base": "

Request to delete a draft environment configuration.

", - "refs": { + "DeleteEnvironmentConfigurationMessage":{ + "base":"

Request to delete a draft environment configuration.

", + "refs":{ } }, - "DeletePlatformVersionRequest": { - "base": null, - "refs": { + "DeletePlatformVersionRequest":{ + "base":null, + "refs":{ } }, - "DeletePlatformVersionResult": { - "base": null, - "refs": { + "DeletePlatformVersionResult":{ + "base":null, + "refs":{ } }, - "DeleteSourceBundle": { - "base": null, - "refs": { - "DeleteApplicationVersionMessage$DeleteSourceBundle": "

Set to true to delete the source bundle from your storage bucket. Otherwise, the application version is deleted only from Elastic Beanstalk and the source bundle remains in Amazon S3.

" + "DeleteSourceBundle":{ + "base":null, + "refs":{ + "DeleteApplicationVersionMessage$DeleteSourceBundle":"

Set to true to delete the source bundle from your storage bucket. Otherwise, the application version is deleted only from Elastic Beanstalk and the source bundle remains in Amazon S3.

" } }, - "Deployment": { - "base": "

Information about an application version deployment.

", - "refs": { - "SingleInstanceHealth$Deployment": "

Information about the most recent deployment to an instance.

" + "Deployment":{ + "base":"

Information about an application version deployment.

", + "refs":{ + "SingleInstanceHealth$Deployment":"

Information about the most recent deployment to an instance.

" } }, - "DeploymentTimestamp": { - "base": null, - "refs": { - "Deployment$DeploymentTime": "

For in-progress deployments, the time that the deployment started.

For completed deployments, the time that the deployment ended.

" + "DeploymentTimestamp":{ + "base":null, + "refs":{ + "Deployment$DeploymentTime":"

For in-progress deployments, the time that the deployment started.

For completed deployments, the time that the deployment ended.

" } }, - "DescribeAccountAttributesResult": { - "base": null, - "refs": { + "DescribeAccountAttributesResult":{ + "base":null, + "refs":{ } }, - "DescribeApplicationVersionsMessage": { - "base": "

Request to describe application versions.

", - "refs": { + "DescribeApplicationVersionsMessage":{ + "base":"

Request to describe application versions.

", + "refs":{ } }, - "DescribeApplicationsMessage": { - "base": "

Request to describe one or more applications.

", - "refs": { + "DescribeApplicationsMessage":{ + "base":"

Request to describe one or more applications.

", + "refs":{ } }, - "DescribeConfigurationOptionsMessage": { - "base": "

Result message containing a list of application version descriptions.

", - "refs": { + "DescribeConfigurationOptionsMessage":{ + "base":"

Result message containing a list of application version descriptions.

", + "refs":{ } }, - "DescribeConfigurationSettingsMessage": { - "base": "

Result message containing all of the configuration settings for a specified solution stack or configuration template.

", - "refs": { + "DescribeConfigurationSettingsMessage":{ + "base":"

Result message containing all of the configuration settings for a specified solution stack or configuration template.

", + "refs":{ } }, - "DescribeEnvironmentHealthRequest": { - "base": "

See the example below to learn how to create a request body.

", - "refs": { + "DescribeEnvironmentHealthRequest":{ + "base":"

See the example below to learn how to create a request body.

", + "refs":{ } }, - "DescribeEnvironmentHealthResult": { - "base": "

Health details for an AWS Elastic Beanstalk environment.

", - "refs": { + "DescribeEnvironmentHealthResult":{ + "base":"

Health details for an AWS Elastic Beanstalk environment.

", + "refs":{ } }, - "DescribeEnvironmentManagedActionHistoryRequest": { - "base": "

Request to list completed and failed managed actions.

", - "refs": { + "DescribeEnvironmentManagedActionHistoryRequest":{ + "base":"

Request to list completed and failed managed actions.

", + "refs":{ } }, - "DescribeEnvironmentManagedActionHistoryResult": { - "base": "

A result message containing a list of completed and failed managed actions.

", - "refs": { + "DescribeEnvironmentManagedActionHistoryResult":{ + "base":"

A result message containing a list of completed and failed managed actions.

", + "refs":{ } }, - "DescribeEnvironmentManagedActionsRequest": { - "base": "

Request to list an environment's upcoming and in-progress managed actions.

", - "refs": { + "DescribeEnvironmentManagedActionsRequest":{ + "base":"

Request to list an environment's upcoming and in-progress managed actions.

", + "refs":{ } }, - "DescribeEnvironmentManagedActionsResult": { - "base": "

The result message containing a list of managed actions.

", - "refs": { + "DescribeEnvironmentManagedActionsResult":{ + "base":"

The result message containing a list of managed actions.

", + "refs":{ } }, - "DescribeEnvironmentResourcesMessage": { - "base": "

Request to describe the resources in an environment.

", - "refs": { + "DescribeEnvironmentResourcesMessage":{ + "base":"

Request to describe the resources in an environment.

", + "refs":{ } }, - "DescribeEnvironmentsMessage": { - "base": "

Request to describe one or more environments.

", - "refs": { + "DescribeEnvironmentsMessage":{ + "base":"

Request to describe one or more environments.

", + "refs":{ } }, - "DescribeEventsMessage": { - "base": "

Request to retrieve a list of events for an environment.

", - "refs": { + "DescribeEventsMessage":{ + "base":"

Request to retrieve a list of events for an environment.

", + "refs":{ } }, - "DescribeInstancesHealthRequest": { - "base": "

Parameters for a call to DescribeInstancesHealth.

", - "refs": { + "DescribeInstancesHealthRequest":{ + "base":"

Parameters for a call to DescribeInstancesHealth.

", + "refs":{ } }, - "DescribeInstancesHealthResult": { - "base": "

Detailed health information about the Amazon EC2 instances in an AWS Elastic Beanstalk environment.

", - "refs": { + "DescribeInstancesHealthResult":{ + "base":"

Detailed health information about the Amazon EC2 instances in an AWS Elastic Beanstalk environment.

", + "refs":{ } }, - "DescribePlatformVersionRequest": { - "base": null, - "refs": { + "DescribePlatformVersionRequest":{ + "base":null, + "refs":{ } }, - "DescribePlatformVersionResult": { - "base": null, - "refs": { + "DescribePlatformVersionResult":{ + "base":null, + "refs":{ } }, - "Description": { - "base": null, - "refs": { - "ApplicationDescription$Description": "

User-defined description of the application.

", - "ApplicationVersionDescription$Description": "

The description of the application version.

", - "ConfigurationSettingsDescription$Description": "

Describes this configuration set.

", - "CreateApplicationMessage$Description": "

Describes the application.

", - "CreateApplicationVersionMessage$Description": "

Describes this version.

", - "CreateConfigurationTemplateMessage$Description": "

Describes this configuration.

", - "CreateEnvironmentMessage$Description": "

Describes this environment.

", - "EnvironmentDescription$Description": "

Describes this environment.

", - "PlatformDescription$Description": "

The description of the platform.

", - "UpdateApplicationMessage$Description": "

A new description for the application.

Default: If not specified, AWS Elastic Beanstalk does not update the description.

", - "UpdateApplicationVersionMessage$Description": "

A new description for this version.

", - "UpdateConfigurationTemplateMessage$Description": "

A new description for the configuration.

", - "UpdateEnvironmentMessage$Description": "

If this parameter is specified, AWS Elastic Beanstalk updates the description of this environment.

" + "Description":{ + "base":null, + "refs":{ + "ApplicationDescription$Description":"

User-defined description of the application.

", + "ApplicationVersionDescription$Description":"

The description of the application version.

", + "ConfigurationSettingsDescription$Description":"

Describes this configuration set.

", + "CreateApplicationMessage$Description":"

Describes the application.

", + "CreateApplicationVersionMessage$Description":"

Describes this version.

", + "CreateConfigurationTemplateMessage$Description":"

Describes this configuration.

", + "CreateEnvironmentMessage$Description":"

Describes this environment.

", + "EnvironmentDescription$Description":"

Describes this environment.

", + "PlatformDescription$Description":"

The description of the platform.

", + "UpdateApplicationMessage$Description":"

A new description for the application.

Default: If not specified, AWS Elastic Beanstalk does not update the description.

", + "UpdateApplicationVersionMessage$Description":"

A new description for this version.

", + "UpdateConfigurationTemplateMessage$Description":"

A new description for the configuration.

", + "UpdateEnvironmentMessage$Description":"

If this parameter is specified, AWS Elastic Beanstalk updates the description of this environment.

" } }, - "Ec2InstanceId": { - "base": null, - "refs": { - "EnvironmentInfoDescription$Ec2InstanceId": "

The Amazon EC2 Instance ID for this information.

" + "Ec2InstanceId":{ + "base":null, + "refs":{ + "EnvironmentInfoDescription$Ec2InstanceId":"

The Amazon EC2 Instance ID for this information.

" } }, - "ElasticBeanstalkServiceException": { - "base": "

A generic service exception has occurred.

", - "refs": { + "ElasticBeanstalkServiceException":{ + "base":"

A generic service exception has occurred.

", + "refs":{ } }, - "EndpointURL": { - "base": null, - "refs": { - "EnvironmentDescription$EndpointURL": "

For load-balanced, autoscaling environments, the URL to the LoadBalancer. For single-instance environments, the IP address of the instance.

" + "EndpointURL":{ + "base":null, + "refs":{ + "EnvironmentDescription$EndpointURL":"

For load-balanced, autoscaling environments, the URL to the LoadBalancer. For single-instance environments, the IP address of the instance.

" } }, - "EnvironmentArn": { - "base": null, - "refs": { - "EnvironmentDescription$EnvironmentArn": "

The environment's Amazon Resource Name (ARN), which can be used in other API requests that require an ARN.

" + "EnvironmentArn":{ + "base":null, + "refs":{ + "EnvironmentDescription$EnvironmentArn":"

The environment's Amazon Resource Name (ARN), which can be used in other API requests that require an ARN.

" } }, - "EnvironmentDescription": { - "base": "

Describes the properties of an environment.

", - "refs": { - "EnvironmentDescriptionsList$member": null + "EnvironmentDescription":{ + "base":"

Describes the properties of an environment.

", + "refs":{ + "EnvironmentDescriptionsList$member":null } }, - "EnvironmentDescriptionsList": { - "base": null, - "refs": { - "EnvironmentDescriptionsMessage$Environments": "

Returns an EnvironmentDescription list.

" + "EnvironmentDescriptionsList":{ + "base":null, + "refs":{ + "EnvironmentDescriptionsMessage$Environments":"

Returns an EnvironmentDescription list.

" } }, - "EnvironmentDescriptionsMessage": { - "base": "

Result message containing a list of environment descriptions.

", - "refs": { + "EnvironmentDescriptionsMessage":{ + "base":"

Result message containing a list of environment descriptions.

", + "refs":{ } }, - "EnvironmentHealth": { - "base": null, - "refs": { - "DescribeEnvironmentHealthResult$Status": "

The environment's operational status. Ready, Launching, Updating, Terminating, or Terminated.

", - "EnvironmentDescription$Health": "

Describes the health status of the environment. AWS Elastic Beanstalk indicates the failure levels for a running environment:

  • Red: Indicates the environment is not responsive. Occurs when three or more consecutive failures occur for an environment.

  • Yellow: Indicates that something is wrong. Occurs when two consecutive failures occur for an environment.

  • Green: Indicates the environment is healthy and fully functional.

  • Grey: Default health for a new environment. The environment is not fully launched and health checks have not started or health checks are suspended during an UpdateEnvironment or RestartEnvironement request.

Default: Grey

" + "EnvironmentHealth":{ + "base":null, + "refs":{ + "DescribeEnvironmentHealthResult$Status":"

The environment's operational status. Ready, Launching, Updating, Terminating, or Terminated.

", + "EnvironmentDescription$Health":"

Describes the health status of the environment. AWS Elastic Beanstalk indicates the failure levels for a running environment:

  • Red: Indicates the environment is not responsive. Occurs when three or more consecutive failures occur for an environment.

  • Yellow: Indicates that something is wrong. Occurs when two consecutive failures occur for an environment.

  • Green: Indicates the environment is healthy and fully functional.

  • Grey: Default health for a new environment. The environment is not fully launched and health checks have not started or health checks are suspended during an UpdateEnvironment or RestartEnvironement request.

Default: Grey

" } }, - "EnvironmentHealthAttribute": { - "base": null, - "refs": { - "EnvironmentHealthAttributes$member": null + "EnvironmentHealthAttribute":{ + "base":null, + "refs":{ + "EnvironmentHealthAttributes$member":null } }, - "EnvironmentHealthAttributes": { - "base": null, - "refs": { - "DescribeEnvironmentHealthRequest$AttributeNames": "

Specify the response elements to return. To retrieve all attributes, set to All. If no attribute names are specified, returns the name of the environment.

" + "EnvironmentHealthAttributes":{ + "base":null, + "refs":{ + "DescribeEnvironmentHealthRequest$AttributeNames":"

Specify the response elements to return. To retrieve all attributes, set to All. If no attribute names are specified, returns the name of the environment.

" } }, - "EnvironmentHealthStatus": { - "base": null, - "refs": { - "EnvironmentDescription$HealthStatus": "

Returns the health status of the application running in your environment. For more information, see Health Colors and Statuses.

" + "EnvironmentHealthStatus":{ + "base":null, + "refs":{ + "EnvironmentDescription$HealthStatus":"

Returns the health status of the application running in your environment. For more information, see Health Colors and Statuses.

" } }, - "EnvironmentId": { - "base": null, - "refs": { - "AbortEnvironmentUpdateMessage$EnvironmentId": "

This specifies the ID of the environment with the in-progress update that you want to cancel.

", - "CreateConfigurationTemplateMessage$EnvironmentId": "

The ID of the environment used with this configuration template.

", - "DescribeEnvironmentHealthRequest$EnvironmentId": "

Specify the environment by ID.

You must specify either this or an EnvironmentName, or both.

", - "DescribeEnvironmentManagedActionHistoryRequest$EnvironmentId": "

The environment ID of the target environment.

", - "DescribeEnvironmentResourcesMessage$EnvironmentId": "

The ID of the environment to retrieve AWS resource usage data.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "DescribeEventsMessage$EnvironmentId": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those associated with this environment.

", - "DescribeInstancesHealthRequest$EnvironmentId": "

Specify the AWS Elastic Beanstalk environment by ID.

", - "EnvironmentDescription$EnvironmentId": "

The ID of this environment.

", - "EnvironmentIdList$member": null, - "RebuildEnvironmentMessage$EnvironmentId": "

The ID of the environment to rebuild.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "RequestEnvironmentInfoMessage$EnvironmentId": "

The ID of the environment of the requested data.

If no such environment is found, RequestEnvironmentInfo returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "RestartAppServerMessage$EnvironmentId": "

The ID of the environment to restart the server for.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "RetrieveEnvironmentInfoMessage$EnvironmentId": "

The ID of the data's environment.

If no such environment is found, returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "SwapEnvironmentCNAMEsMessage$SourceEnvironmentId": "

The ID of the source environment.

Condition: You must specify at least the SourceEnvironmentID or the SourceEnvironmentName. You may also specify both. If you specify the SourceEnvironmentId, you must specify the DestinationEnvironmentId.

", - "SwapEnvironmentCNAMEsMessage$DestinationEnvironmentId": "

The ID of the destination environment.

Condition: You must specify at least the DestinationEnvironmentID or the DestinationEnvironmentName. You may also specify both. You must specify the SourceEnvironmentId with the DestinationEnvironmentId.

", - "TerminateEnvironmentMessage$EnvironmentId": "

The ID of the environment to terminate.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "UpdateEnvironmentMessage$EnvironmentId": "

The ID of the environment to update.

If no environment with this ID exists, AWS Elastic Beanstalk returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

" + "EnvironmentId":{ + "base":null, + "refs":{ + "AbortEnvironmentUpdateMessage$EnvironmentId":"

This specifies the ID of the environment with the in-progress update that you want to cancel.

", + "CreateConfigurationTemplateMessage$EnvironmentId":"

The ID of the environment used with this configuration template.

", + "DescribeEnvironmentHealthRequest$EnvironmentId":"

Specify the environment by ID.

You must specify either this or an EnvironmentName, or both.

", + "DescribeEnvironmentManagedActionHistoryRequest$EnvironmentId":"

The environment ID of the target environment.

", + "DescribeEnvironmentResourcesMessage$EnvironmentId":"

The ID of the environment to retrieve AWS resource usage data.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "DescribeEventsMessage$EnvironmentId":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those associated with this environment.

", + "DescribeInstancesHealthRequest$EnvironmentId":"

Specify the AWS Elastic Beanstalk environment by ID.

", + "EnvironmentDescription$EnvironmentId":"

The ID of this environment.

", + "EnvironmentIdList$member":null, + "RebuildEnvironmentMessage$EnvironmentId":"

The ID of the environment to rebuild.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "RequestEnvironmentInfoMessage$EnvironmentId":"

The ID of the environment of the requested data.

If no such environment is found, RequestEnvironmentInfo returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "RestartAppServerMessage$EnvironmentId":"

The ID of the environment to restart the server for.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "RetrieveEnvironmentInfoMessage$EnvironmentId":"

The ID of the data's environment.

If no such environment is found, returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "SwapEnvironmentCNAMEsMessage$SourceEnvironmentId":"

The ID of the source environment.

Condition: You must specify at least the SourceEnvironmentID or the SourceEnvironmentName. You may also specify both. If you specify the SourceEnvironmentId, you must specify the DestinationEnvironmentId.

", + "SwapEnvironmentCNAMEsMessage$DestinationEnvironmentId":"

The ID of the destination environment.

Condition: You must specify at least the DestinationEnvironmentID or the DestinationEnvironmentName. You may also specify both. You must specify the SourceEnvironmentId with the DestinationEnvironmentId.

", + "TerminateEnvironmentMessage$EnvironmentId":"

The ID of the environment to terminate.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "UpdateEnvironmentMessage$EnvironmentId":"

The ID of the environment to update.

If no environment with this ID exists, AWS Elastic Beanstalk returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentName, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

" } }, - "EnvironmentIdList": { - "base": null, - "refs": { - "DescribeEnvironmentsMessage$EnvironmentIds": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that have the specified IDs.

" + "EnvironmentIdList":{ + "base":null, + "refs":{ + "DescribeEnvironmentsMessage$EnvironmentIds":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that have the specified IDs.

" } }, - "EnvironmentInfoDescription": { - "base": "

The information retrieved from the Amazon EC2 instances.

", - "refs": { - "EnvironmentInfoDescriptionList$member": null + "EnvironmentInfoDescription":{ + "base":"

The information retrieved from the Amazon EC2 instances.

", + "refs":{ + "EnvironmentInfoDescriptionList$member":null } }, - "EnvironmentInfoDescriptionList": { - "base": null, - "refs": { - "RetrieveEnvironmentInfoResultMessage$EnvironmentInfo": "

The EnvironmentInfoDescription of the environment.

" + "EnvironmentInfoDescriptionList":{ + "base":null, + "refs":{ + "RetrieveEnvironmentInfoResultMessage$EnvironmentInfo":"

The EnvironmentInfoDescription of the environment.

" } }, - "EnvironmentInfoType": { - "base": null, - "refs": { - "EnvironmentInfoDescription$InfoType": "

The type of information retrieved.

", - "RequestEnvironmentInfoMessage$InfoType": "

The type of information to request.

", - "RetrieveEnvironmentInfoMessage$InfoType": "

The type of information to retrieve.

" + "EnvironmentInfoType":{ + "base":null, + "refs":{ + "EnvironmentInfoDescription$InfoType":"

The type of information retrieved.

", + "RequestEnvironmentInfoMessage$InfoType":"

The type of information to request.

", + "RetrieveEnvironmentInfoMessage$InfoType":"

The type of information to retrieve.

" } }, - "EnvironmentLink": { - "base": "

A link to another environment, defined in the environment's manifest. Links provide connection information in system properties that can be used to connect to another environment in the same group. See Environment Manifest (env.yaml) for details.

", - "refs": { - "EnvironmentLinks$member": null + "EnvironmentLink":{ + "base":"

A link to another environment, defined in the environment's manifest. Links provide connection information in system properties that can be used to connect to another environment in the same group. See Environment Manifest (env.yaml) for details.

", + "refs":{ + "EnvironmentLinks$member":null } }, - "EnvironmentLinks": { - "base": null, - "refs": { - "EnvironmentDescription$EnvironmentLinks": "

A list of links to other environments in the same group.

" + "EnvironmentLinks":{ + "base":null, + "refs":{ + "EnvironmentDescription$EnvironmentLinks":"

A list of links to other environments in the same group.

" } }, - "EnvironmentName": { - "base": null, - "refs": { - "AbortEnvironmentUpdateMessage$EnvironmentName": "

This specifies the name of the environment with the in-progress update that you want to cancel.

", - "ConfigurationSettingsDescription$EnvironmentName": "

If not null, the name of the environment for this configuration set.

", - "CreateEnvironmentMessage$EnvironmentName": "

A unique name for the deployment environment. Used in the application URL.

Constraint: Must be from 4 to 40 characters in length. The name can contain only letters, numbers, and hyphens. It cannot start or end with a hyphen. This name must be unique within a region in your account. If the specified name already exists in the region, AWS Elastic Beanstalk returns an InvalidParameterValue error.

Default: If the CNAME parameter is not specified, the environment name becomes part of the CNAME, and therefore part of the visible URL for your application.

", - "CreatePlatformVersionRequest$EnvironmentName": "

The name of the builder environment.

", - "DeleteEnvironmentConfigurationMessage$EnvironmentName": "

The name of the environment to delete the draft configuration from.

", - "DescribeConfigurationOptionsMessage$EnvironmentName": "

The name of the environment whose configuration options you want to describe.

", - "DescribeConfigurationSettingsMessage$EnvironmentName": "

The name of the environment to describe.

Condition: You must specify either this or a TemplateName, but not both. If you specify both, AWS Elastic Beanstalk returns an InvalidParameterCombination error. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "DescribeEnvironmentHealthRequest$EnvironmentName": "

Specify the environment by name.

You must specify either this or an EnvironmentName, or both.

", - "DescribeEnvironmentHealthResult$EnvironmentName": "

The environment's name.

", - "DescribeEnvironmentManagedActionHistoryRequest$EnvironmentName": "

The name of the target environment.

", - "DescribeEnvironmentResourcesMessage$EnvironmentName": "

The name of the environment to retrieve AWS resource usage data.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "DescribeEventsMessage$EnvironmentName": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those associated with this environment.

", - "DescribeInstancesHealthRequest$EnvironmentName": "

Specify the AWS Elastic Beanstalk environment by name.

", - "EnvironmentDescription$EnvironmentName": "

The name of this environment.

", - "EnvironmentNamesList$member": null, - "EnvironmentResourceDescription$EnvironmentName": "

The name of the environment.

", - "EventDescription$EnvironmentName": "

The name of the environment associated with this event.

", - "RebuildEnvironmentMessage$EnvironmentName": "

The name of the environment to rebuild.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "RequestEnvironmentInfoMessage$EnvironmentName": "

The name of the environment of the requested data.

If no such environment is found, RequestEnvironmentInfo returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "RestartAppServerMessage$EnvironmentName": "

The name of the environment to restart the server for.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "RetrieveEnvironmentInfoMessage$EnvironmentName": "

The name of the data's environment.

If no such environment is found, returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "SwapEnvironmentCNAMEsMessage$SourceEnvironmentName": "

The name of the source environment.

Condition: You must specify at least the SourceEnvironmentID or the SourceEnvironmentName. You may also specify both. If you specify the SourceEnvironmentName, you must specify the DestinationEnvironmentName.

", - "SwapEnvironmentCNAMEsMessage$DestinationEnvironmentName": "

The name of the destination environment.

Condition: You must specify at least the DestinationEnvironmentID or the DestinationEnvironmentName. You may also specify both. You must specify the SourceEnvironmentName with the DestinationEnvironmentName.

", - "TerminateEnvironmentMessage$EnvironmentName": "

The name of the environment to terminate.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "UpdateEnvironmentMessage$EnvironmentName": "

The name of the environment to update. If no environment with this name exists, AWS Elastic Beanstalk returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", - "ValidateConfigurationSettingsMessage$EnvironmentName": "

The name of the environment to validate the settings against.

Condition: You cannot specify both this and a configuration template name.

" + "EnvironmentName":{ + "base":null, + "refs":{ + "AbortEnvironmentUpdateMessage$EnvironmentName":"

This specifies the name of the environment with the in-progress update that you want to cancel.

", + "ConfigurationSettingsDescription$EnvironmentName":"

If not null, the name of the environment for this configuration set.

", + "CreateEnvironmentMessage$EnvironmentName":"

A unique name for the deployment environment. Used in the application URL.

Constraint: Must be from 4 to 40 characters in length. The name can contain only letters, numbers, and hyphens. It cannot start or end with a hyphen. This name must be unique within a region in your account. If the specified name already exists in the region, AWS Elastic Beanstalk returns an InvalidParameterValue error.

Default: If the CNAME parameter is not specified, the environment name becomes part of the CNAME, and therefore part of the visible URL for your application.

", + "CreatePlatformVersionRequest$EnvironmentName":"

The name of the builder environment.

", + "DeleteEnvironmentConfigurationMessage$EnvironmentName":"

The name of the environment to delete the draft configuration from.

", + "DescribeConfigurationOptionsMessage$EnvironmentName":"

The name of the environment whose configuration options you want to describe.

", + "DescribeConfigurationSettingsMessage$EnvironmentName":"

The name of the environment to describe.

Condition: You must specify either this or a TemplateName, but not both. If you specify both, AWS Elastic Beanstalk returns an InvalidParameterCombination error. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "DescribeEnvironmentHealthRequest$EnvironmentName":"

Specify the environment by name.

You must specify either this or an EnvironmentName, or both.

", + "DescribeEnvironmentHealthResult$EnvironmentName":"

The environment's name.

", + "DescribeEnvironmentManagedActionHistoryRequest$EnvironmentName":"

The name of the target environment.

", + "DescribeEnvironmentResourcesMessage$EnvironmentName":"

The name of the environment to retrieve AWS resource usage data.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "DescribeEventsMessage$EnvironmentName":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those associated with this environment.

", + "DescribeInstancesHealthRequest$EnvironmentName":"

Specify the AWS Elastic Beanstalk environment by name.

", + "EnvironmentDescription$EnvironmentName":"

The name of this environment.

", + "EnvironmentNamesList$member":null, + "EnvironmentResourceDescription$EnvironmentName":"

The name of the environment.

", + "EventDescription$EnvironmentName":"

The name of the environment associated with this event.

", + "RebuildEnvironmentMessage$EnvironmentName":"

The name of the environment to rebuild.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "RequestEnvironmentInfoMessage$EnvironmentName":"

The name of the environment of the requested data.

If no such environment is found, RequestEnvironmentInfo returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "RestartAppServerMessage$EnvironmentName":"

The name of the environment to restart the server for.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "RetrieveEnvironmentInfoMessage$EnvironmentName":"

The name of the data's environment.

If no such environment is found, returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "SwapEnvironmentCNAMEsMessage$SourceEnvironmentName":"

The name of the source environment.

Condition: You must specify at least the SourceEnvironmentID or the SourceEnvironmentName. You may also specify both. If you specify the SourceEnvironmentName, you must specify the DestinationEnvironmentName.

", + "SwapEnvironmentCNAMEsMessage$DestinationEnvironmentName":"

The name of the destination environment.

Condition: You must specify at least the DestinationEnvironmentID or the DestinationEnvironmentName. You may also specify both. You must specify the SourceEnvironmentName with the DestinationEnvironmentName.

", + "TerminateEnvironmentMessage$EnvironmentName":"

The name of the environment to terminate.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "UpdateEnvironmentMessage$EnvironmentName":"

The name of the environment to update. If no environment with this name exists, AWS Elastic Beanstalk returns an InvalidParameterValue error.

Condition: You must specify either this or an EnvironmentId, or both. If you do not specify either, AWS Elastic Beanstalk returns MissingRequiredParameter error.

", + "ValidateConfigurationSettingsMessage$EnvironmentName":"

The name of the environment to validate the settings against.

Condition: You cannot specify both this and a configuration template name.

" } }, - "EnvironmentNamesList": { - "base": null, - "refs": { - "DescribeEnvironmentsMessage$EnvironmentNames": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that have the specified names.

" + "EnvironmentNamesList":{ + "base":null, + "refs":{ + "DescribeEnvironmentsMessage$EnvironmentNames":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that have the specified names.

" } }, - "EnvironmentResourceDescription": { - "base": "

Describes the AWS resources in use by this environment. This data is live.

", - "refs": { - "EnvironmentResourceDescriptionsMessage$EnvironmentResources": "

A list of EnvironmentResourceDescription.

" + "EnvironmentResourceDescription":{ + "base":"

Describes the AWS resources in use by this environment. This data is live.

", + "refs":{ + "EnvironmentResourceDescriptionsMessage$EnvironmentResources":"

A list of EnvironmentResourceDescription.

" } }, - "EnvironmentResourceDescriptionsMessage": { - "base": "

Result message containing a list of environment resource descriptions.

", - "refs": { + "EnvironmentResourceDescriptionsMessage":{ + "base":"

Result message containing a list of environment resource descriptions.

", + "refs":{ } }, - "EnvironmentResourcesDescription": { - "base": "

Describes the AWS resources in use by this environment. This data is not live data.

", - "refs": { - "EnvironmentDescription$Resources": "

The description of the AWS resources used by this environment.

" + "EnvironmentResourcesDescription":{ + "base":"

Describes the AWS resources in use by this environment. This data is not live data.

", + "refs":{ + "EnvironmentDescription$Resources":"

The description of the AWS resources used by this environment.

" } }, - "EnvironmentStatus": { - "base": null, - "refs": { - "EnvironmentDescription$Status": "

The current operational status of the environment:

  • Launching: Environment is in the process of initial deployment.

  • Updating: Environment is in the process of updating its configuration settings or application version.

  • Ready: Environment is available to have an action performed on it, such as update or terminate.

  • Terminating: Environment is in the shut-down process.

  • Terminated: Environment is not running.

" + "EnvironmentStatus":{ + "base":null, + "refs":{ + "EnvironmentDescription$Status":"

The current operational status of the environment:

  • Launching: Environment is in the process of initial deployment.

  • Updating: Environment is in the process of updating its configuration settings or application version.

  • Ready: Environment is available to have an action performed on it, such as update or terminate.

  • Terminating: Environment is in the shut-down process.

  • Terminated: Environment is not running.

" } }, - "EnvironmentTier": { - "base": "

Describes the properties of an environment tier

", - "refs": { - "CreateEnvironmentMessage$Tier": "

This specifies the tier to use for creating this environment.

", - "EnvironmentDescription$Tier": "

Describes the current tier of this environment.

", - "UpdateEnvironmentMessage$Tier": "

This specifies the tier to use to update the environment.

Condition: At this time, if you change the tier version, name, or type, AWS Elastic Beanstalk returns InvalidParameterValue error.

" + "EnvironmentTier":{ + "base":"

Describes the properties of an environment tier

", + "refs":{ + "CreateEnvironmentMessage$Tier":"

This specifies the tier to use for creating this environment.

", + "EnvironmentDescription$Tier":"

Describes the current tier of this environment.

", + "UpdateEnvironmentMessage$Tier":"

This specifies the tier to use to update the environment.

Condition: At this time, if you change the tier version, name, or type, AWS Elastic Beanstalk returns InvalidParameterValue error.

" } }, - "EventDate": { - "base": null, - "refs": { - "EventDescription$EventDate": "

The date when the event occurred.

" + "EventDate":{ + "base":null, + "refs":{ + "EventDescription$EventDate":"

The date when the event occurred.

" } }, - "EventDescription": { - "base": "

Describes an event.

", - "refs": { - "EventDescriptionList$member": null + "EventDescription":{ + "base":"

Describes an event.

", + "refs":{ + "EventDescriptionList$member":null } }, - "EventDescriptionList": { - "base": null, - "refs": { - "EventDescriptionsMessage$Events": "

A list of EventDescription.

" + "EventDescriptionList":{ + "base":null, + "refs":{ + "EventDescriptionsMessage$Events":"

A list of EventDescription.

" } }, - "EventDescriptionsMessage": { - "base": "

Result message wrapping a list of event descriptions.

", - "refs": { + "EventDescriptionsMessage":{ + "base":"

Result message wrapping a list of event descriptions.

", + "refs":{ } }, - "EventMessage": { - "base": null, - "refs": { - "EventDescription$Message": "

The event message.

" + "EventMessage":{ + "base":null, + "refs":{ + "EventDescription$Message":"

The event message.

" } }, - "EventSeverity": { - "base": null, - "refs": { - "DescribeEventsMessage$Severity": "

If specified, limits the events returned from this call to include only those with the specified severity or higher.

", - "EventDescription$Severity": "

The severity level of this event.

" + "EventSeverity":{ + "base":null, + "refs":{ + "DescribeEventsMessage$Severity":"

If specified, limits the events returned from this call to include only those with the specified severity or higher.

", + "EventDescription$Severity":"

The severity level of this event.

" } }, - "ExceptionMessage": { - "base": null, - "refs": { - "ElasticBeanstalkServiceException$message": "

The exception error message.

" + "ExceptionMessage":{ + "base":null, + "refs":{ + "ElasticBeanstalkServiceException$message":"

The exception error message.

" } }, - "FailureType": { - "base": null, - "refs": { - "ManagedActionHistoryItem$FailureType": "

If the action failed, the type of failure.

" + "FailureType":{ + "base":null, + "refs":{ + "ManagedActionHistoryItem$FailureType":"

If the action failed, the type of failure.

" } }, - "FileTypeExtension": { - "base": null, - "refs": { - "SolutionStackFileTypeList$member": null + "FileTypeExtension":{ + "base":null, + "refs":{ + "SolutionStackFileTypeList$member":null } }, - "ForceTerminate": { - "base": null, - "refs": { - "TerminateEnvironmentMessage$ForceTerminate": "

Terminates the target environment even if another environment in the same group is dependent on it.

" + "ForceTerminate":{ + "base":null, + "refs":{ + "TerminateEnvironmentMessage$ForceTerminate":"

Terminates the target environment even if another environment in the same group is dependent on it.

" } }, - "GroupName": { - "base": null, - "refs": { - "ComposeEnvironmentsMessage$GroupName": "

The name of the group to which the target environments belong. Specify a group name only if the environment name defined in each target environment's manifest ends with a + (plus) character. See Environment Manifest (env.yaml) for details.

", - "CreateEnvironmentMessage$GroupName": "

The name of the group to which the target environment belongs. Specify a group name only if the environment's name is specified in an environment manifest and not with the environment name parameter. See Environment Manifest (env.yaml) for details.

", - "UpdateEnvironmentMessage$GroupName": "

The name of the group to which the target environment belongs. Specify a group name only if the environment's name is specified in an environment manifest and not with the environment name or environment ID parameters. See Environment Manifest (env.yaml) for details.

" + "GroupName":{ + "base":null, + "refs":{ + "ComposeEnvironmentsMessage$GroupName":"

The name of the group to which the target environments belong. Specify a group name only if the environment name defined in each target environment's manifest ends with a + (plus) character. See Environment Manifest (env.yaml) for details.

", + "CreateEnvironmentMessage$GroupName":"

The name of the group to which the target environment belongs. Specify a group name only if the environment's name is specified in an environment manifest and not with the environment name parameter. See Environment Manifest (env.yaml) for details.

", + "UpdateEnvironmentMessage$GroupName":"

The name of the group to which the target environment belongs. Specify a group name only if the environment's name is specified in an environment manifest and not with the environment name or environment ID parameters. See Environment Manifest (env.yaml) for details.

" } }, - "ImageId": { - "base": null, - "refs": { - "CustomAmi$ImageId": "

THe ID of the image used to create the custom AMI.

" + "ImageId":{ + "base":null, + "refs":{ + "CustomAmi$ImageId":"

THe ID of the image used to create the custom AMI.

" } }, - "IncludeDeleted": { - "base": null, - "refs": { - "DescribeEnvironmentsMessage$IncludeDeleted": "

Indicates whether to include deleted environments:

true: Environments that have been deleted after IncludedDeletedBackTo are displayed.

false: Do not include deleted environments.

" + "IncludeDeleted":{ + "base":null, + "refs":{ + "DescribeEnvironmentsMessage$IncludeDeleted":"

Indicates whether to include deleted environments:

true: Environments that have been deleted after IncludedDeletedBackTo are displayed.

false: Do not include deleted environments.

" } }, - "IncludeDeletedBackTo": { - "base": null, - "refs": { - "DescribeEnvironmentsMessage$IncludedDeletedBackTo": "

If specified when IncludeDeleted is set to true, then environments deleted after this date are displayed.

" + "IncludeDeletedBackTo":{ + "base":null, + "refs":{ + "DescribeEnvironmentsMessage$IncludedDeletedBackTo":"

If specified when IncludeDeleted is set to true, then environments deleted after this date are displayed.

" } }, - "Instance": { - "base": "

The description of an Amazon EC2 instance.

", - "refs": { - "InstanceList$member": null + "Instance":{ + "base":"

The description of an Amazon EC2 instance.

", + "refs":{ + "InstanceList$member":null } }, - "InstanceHealthList": { - "base": null, - "refs": { - "DescribeInstancesHealthResult$InstanceHealthList": "

Detailed health information about each instance.

" + "InstanceHealthList":{ + "base":null, + "refs":{ + "DescribeInstancesHealthResult$InstanceHealthList":"

Detailed health information about each instance.

" } }, - "InstanceHealthSummary": { - "base": "

Represents summary information about the health of an instance. For more information, see Health Colors and Statuses.

", - "refs": { - "DescribeEnvironmentHealthResult$InstancesHealth": "

Summary health information for the instances in the environment.

" + "InstanceHealthSummary":{ + "base":"

Represents summary information about the health of an instance. For more information, see Health Colors and Statuses.

", + "refs":{ + "DescribeEnvironmentHealthResult$InstancesHealth":"

Summary health information for the instances in the environment.

" } }, - "InstanceId": { - "base": null, - "refs": { - "SingleInstanceHealth$InstanceId": "

The ID of the Amazon EC2 instance.

" + "InstanceId":{ + "base":null, + "refs":{ + "SingleInstanceHealth$InstanceId":"

The ID of the Amazon EC2 instance.

" } }, - "InstanceList": { - "base": null, - "refs": { - "EnvironmentResourceDescription$Instances": "

The Amazon EC2 instances used by this environment.

" + "InstanceList":{ + "base":null, + "refs":{ + "EnvironmentResourceDescription$Instances":"

The Amazon EC2 instances used by this environment.

" } }, - "InstancesHealthAttribute": { - "base": null, - "refs": { - "InstancesHealthAttributes$member": null + "InstancesHealthAttribute":{ + "base":null, + "refs":{ + "InstancesHealthAttributes$member":null } }, - "InstancesHealthAttributes": { - "base": null, - "refs": { - "DescribeInstancesHealthRequest$AttributeNames": "

Specifies the response elements you wish to receive. To retrieve all attributes, set to All. If no attribute names are specified, returns a list of instances.

" + "InstancesHealthAttributes":{ + "base":null, + "refs":{ + "DescribeInstancesHealthRequest$AttributeNames":"

Specifies the response elements you wish to receive. To retrieve all attributes, set to All. If no attribute names are specified, returns a list of instances.

" } }, - "InsufficientPrivilegesException": { - "base": "

The specified account does not have sufficient privileges for one or more AWS services.

", - "refs": { + "InsufficientPrivilegesException":{ + "base":"

The specified account does not have sufficient privileges for one or more AWS services.

", + "refs":{ } }, - "Integer": { - "base": null, - "refs": { - "DescribeEnvironmentManagedActionHistoryRequest$MaxItems": "

The maximum number of items to return for a single request.

", - "Listener$Port": "

The port that is used by the Listener.

" + "Integer":{ + "base":null, + "refs":{ + "DescribeEnvironmentManagedActionHistoryRequest$MaxItems":"

The maximum number of items to return for a single request.

", + "Listener$Port":"

The port that is used by the Listener.

" } }, - "InvalidRequestException": { - "base": "

One or more input parameters is not valid. Please correct the input parameters and try the operation again.

", - "refs": { + "InvalidRequestException":{ + "base":"

One or more input parameters is not valid. Please correct the input parameters and try the operation again.

", + "refs":{ } }, - "Latency": { - "base": "

Represents the average latency for the slowest X percent of requests over the last 10 seconds.

", - "refs": { - "ApplicationMetrics$Latency": "

Represents the average latency for the slowest X percent of requests over the last 10 seconds. Latencies are in seconds with one millisecond resolution.

" + "Latency":{ + "base":"

Represents the average latency for the slowest X percent of requests over the last 10 seconds.

", + "refs":{ + "ApplicationMetrics$Latency":"

Represents the average latency for the slowest X percent of requests over the last 10 seconds. Latencies are in seconds with one millisecond resolution.

" } }, - "LaunchConfiguration": { - "base": "

Describes an Auto Scaling launch configuration.

", - "refs": { - "LaunchConfigurationList$member": null + "LaunchConfiguration":{ + "base":"

Describes an Auto Scaling launch configuration.

", + "refs":{ + "LaunchConfigurationList$member":null } }, - "LaunchConfigurationList": { - "base": null, - "refs": { - "EnvironmentResourceDescription$LaunchConfigurations": "

The Auto Scaling launch configurations in use by this environment.

" + "LaunchConfigurationList":{ + "base":null, + "refs":{ + "EnvironmentResourceDescription$LaunchConfigurations":"

The Auto Scaling launch configurations in use by this environment.

" } }, - "LaunchedAt": { - "base": null, - "refs": { - "SingleInstanceHealth$LaunchedAt": "

The time at which the EC2 instance was launched.

" + "LaunchedAt":{ + "base":null, + "refs":{ + "SingleInstanceHealth$LaunchedAt":"

The time at which the EC2 instance was launched.

" } }, - "ListAvailableSolutionStacksResultMessage": { - "base": "

A list of available AWS Elastic Beanstalk solution stacks.

", - "refs": { + "ListAvailableSolutionStacksResultMessage":{ + "base":"

A list of available AWS Elastic Beanstalk solution stacks.

", + "refs":{ } }, - "ListPlatformVersionsRequest": { - "base": null, - "refs": { + "ListPlatformVersionsRequest":{ + "base":null, + "refs":{ } }, - "ListPlatformVersionsResult": { - "base": null, - "refs": { + "ListPlatformVersionsResult":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceMessage": { - "base": null, - "refs": { + "ListTagsForResourceMessage":{ + "base":null, + "refs":{ } }, - "Listener": { - "base": "

Describes the properties of a Listener for the LoadBalancer.

", - "refs": { - "LoadBalancerListenersDescription$member": null + "Listener":{ + "base":"

Describes the properties of a Listener for the LoadBalancer.

", + "refs":{ + "LoadBalancerListenersDescription$member":null } }, - "LoadAverage": { - "base": null, - "refs": { - "SystemStatus$LoadAverage": "

Load average in the last 1-minute, 5-minute, and 15-minute periods. For more information, see Operating System Metrics.

" + "LoadAverage":{ + "base":null, + "refs":{ + "SystemStatus$LoadAverage":"

Load average in the last 1-minute, 5-minute, and 15-minute periods. For more information, see Operating System Metrics.

" } }, - "LoadAverageValue": { - "base": null, - "refs": { - "LoadAverage$member": null + "LoadAverageValue":{ + "base":null, + "refs":{ + "LoadAverage$member":null } }, - "LoadBalancer": { - "base": "

Describes a LoadBalancer.

", - "refs": { - "LoadBalancerList$member": null + "LoadBalancer":{ + "base":"

Describes a LoadBalancer.

", + "refs":{ + "LoadBalancerList$member":null } }, - "LoadBalancerDescription": { - "base": "

Describes the details of a LoadBalancer.

", - "refs": { - "EnvironmentResourcesDescription$LoadBalancer": "

Describes the LoadBalancer.

" + "LoadBalancerDescription":{ + "base":"

Describes the details of a LoadBalancer.

", + "refs":{ + "EnvironmentResourcesDescription$LoadBalancer":"

Describes the LoadBalancer.

" } }, - "LoadBalancerList": { - "base": null, - "refs": { - "EnvironmentResourceDescription$LoadBalancers": "

The LoadBalancers in use by this environment.

" + "LoadBalancerList":{ + "base":null, + "refs":{ + "EnvironmentResourceDescription$LoadBalancers":"

The LoadBalancers in use by this environment.

" } }, - "LoadBalancerListenersDescription": { - "base": null, - "refs": { - "LoadBalancerDescription$Listeners": "

A list of Listeners used by the LoadBalancer.

" + "LoadBalancerListenersDescription":{ + "base":null, + "refs":{ + "LoadBalancerDescription$Listeners":"

A list of Listeners used by the LoadBalancer.

" } }, - "Maintainer": { - "base": null, - "refs": { - "PlatformDescription$Maintainer": "

Information about the maintainer of the platform.

" + "Maintainer":{ + "base":null, + "refs":{ + "PlatformDescription$Maintainer":"

Information about the maintainer of the platform.

" } }, - "ManagedAction": { - "base": "

The record of an upcoming or in-progress managed action.

", - "refs": { - "ManagedActions$member": null + "ManagedAction":{ + "base":"

The record of an upcoming or in-progress managed action.

", + "refs":{ + "ManagedActions$member":null } }, - "ManagedActionHistoryItem": { - "base": "

The record of a completed or failed managed action.

", - "refs": { - "ManagedActionHistoryItems$member": null + "ManagedActionHistoryItem":{ + "base":"

The record of a completed or failed managed action.

", + "refs":{ + "ManagedActionHistoryItems$member":null } }, - "ManagedActionHistoryItems": { - "base": null, - "refs": { - "DescribeEnvironmentManagedActionHistoryResult$ManagedActionHistoryItems": "

A list of completed and failed managed actions.

" + "ManagedActionHistoryItems":{ + "base":null, + "refs":{ + "DescribeEnvironmentManagedActionHistoryResult$ManagedActionHistoryItems":"

A list of completed and failed managed actions.

" } }, - "ManagedActionInvalidStateException": { - "base": "

Cannot modify the managed action in its current state.

", - "refs": { + "ManagedActionInvalidStateException":{ + "base":"

Cannot modify the managed action in its current state.

", + "refs":{ } }, - "ManagedActions": { - "base": null, - "refs": { - "DescribeEnvironmentManagedActionsResult$ManagedActions": "

A list of upcoming and in-progress managed actions.

" + "ManagedActions":{ + "base":null, + "refs":{ + "DescribeEnvironmentManagedActionsResult$ManagedActions":"

A list of upcoming and in-progress managed actions.

" } }, - "MaxAgeRule": { - "base": "

A lifecycle rule that deletes application versions after the specified number of days.

", - "refs": { - "ApplicationVersionLifecycleConfig$MaxAgeRule": "

Specify a max age rule to restrict the length of time that application versions are retained for an application.

" + "MaxAgeRule":{ + "base":"

A lifecycle rule that deletes application versions after the specified number of days.

", + "refs":{ + "ApplicationVersionLifecycleConfig$MaxAgeRule":"

Specify a max age rule to restrict the length of time that application versions are retained for an application.

" } }, - "MaxCountRule": { - "base": "

A lifecycle rule that deletes the oldest application version when the maximum count is exceeded.

", - "refs": { - "ApplicationVersionLifecycleConfig$MaxCountRule": "

Specify a max count rule to restrict the number of application versions that are retained for an application.

" + "MaxCountRule":{ + "base":"

A lifecycle rule that deletes the oldest application version when the maximum count is exceeded.

", + "refs":{ + "ApplicationVersionLifecycleConfig$MaxCountRule":"

Specify a max count rule to restrict the number of application versions that are retained for an application.

" } }, - "MaxRecords": { - "base": null, - "refs": { - "DescribeApplicationVersionsMessage$MaxRecords": "

For a paginated request. Specify a maximum number of application versions to include in each response.

If no MaxRecords is specified, all available application versions are retrieved in a single response.

", - "DescribeEnvironmentsMessage$MaxRecords": "

For a paginated request. Specify a maximum number of environments to include in each response.

If no MaxRecords is specified, all available environments are retrieved in a single response.

", - "DescribeEventsMessage$MaxRecords": "

Specifies the maximum number of events that can be returned, beginning with the most recent event.

" + "MaxRecords":{ + "base":null, + "refs":{ + "DescribeApplicationVersionsMessage$MaxRecords":"

For a paginated request. Specify a maximum number of application versions to include in each response.

If no MaxRecords is specified, all available application versions are retrieved in a single response.

", + "DescribeEnvironmentsMessage$MaxRecords":"

For a paginated request. Specify a maximum number of environments to include in each response.

If no MaxRecords is specified, all available environments are retrieved in a single response.

", + "DescribeEventsMessage$MaxRecords":"

Specifies the maximum number of events that can be returned, beginning with the most recent event.

" } }, - "Message": { - "base": null, - "refs": { - "EnvironmentInfoDescription$Message": "

The retrieved information.

" + "Message":{ + "base":null, + "refs":{ + "EnvironmentInfoDescription$Message":"

The retrieved information.

" } }, - "NextToken": { - "base": null, - "refs": { - "DescribeInstancesHealthRequest$NextToken": "

Specify the pagination token returned by a previous call.

", - "DescribeInstancesHealthResult$NextToken": "

Pagination token for the next page of results, if available.

" + "NextToken":{ + "base":null, + "refs":{ + "DescribeInstancesHealthRequest$NextToken":"

Specify the pagination token returned by a previous call.

", + "DescribeInstancesHealthResult$NextToken":"

Pagination token for the next page of results, if available.

" } }, - "NonEmptyString": { - "base": null, - "refs": { - "BuildConfiguration$CodeBuildServiceRole": "

The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that enables AWS CodeBuild to interact with dependent AWS services on behalf of the AWS account.

", - "BuildConfiguration$Image": "

The ID of the Docker image to use for this build project.

" + "NonEmptyString":{ + "base":null, + "refs":{ + "BuildConfiguration$CodeBuildServiceRole":"

The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that enables AWS CodeBuild to interact with dependent AWS services on behalf of the AWS account.

", + "BuildConfiguration$Image":"

The ID of the Docker image to use for this build project.

" } }, - "NullableDouble": { - "base": null, - "refs": { - "CPUUtilization$User": "

Percentage of time that the CPU has spent in the User state over the last 10 seconds.

", - "CPUUtilization$Nice": "

Percentage of time that the CPU has spent in the Nice state over the last 10 seconds.

", - "CPUUtilization$System": "

Percentage of time that the CPU has spent in the System state over the last 10 seconds.

", - "CPUUtilization$Idle": "

Percentage of time that the CPU has spent in the Idle state over the last 10 seconds.

", - "CPUUtilization$IOWait": "

Percentage of time that the CPU has spent in the I/O Wait state over the last 10 seconds.

", - "CPUUtilization$IRQ": "

Percentage of time that the CPU has spent in the IRQ state over the last 10 seconds.

", - "CPUUtilization$SoftIRQ": "

Percentage of time that the CPU has spent in the SoftIRQ state over the last 10 seconds.

", - "Latency$P999": "

The average latency for the slowest 0.1 percent of requests over the last 10 seconds.

", - "Latency$P99": "

The average latency for the slowest 1 percent of requests over the last 10 seconds.

", - "Latency$P95": "

The average latency for the slowest 5 percent of requests over the last 10 seconds.

", - "Latency$P90": "

The average latency for the slowest 10 percent of requests over the last 10 seconds.

", - "Latency$P85": "

The average latency for the slowest 15 percent of requests over the last 10 seconds.

", - "Latency$P75": "

The average latency for the slowest 25 percent of requests over the last 10 seconds.

", - "Latency$P50": "

The average latency for the slowest 50 percent of requests over the last 10 seconds.

", - "Latency$P10": "

The average latency for the slowest 90 percent of requests over the last 10 seconds.

" + "NullableDouble":{ + "base":null, + "refs":{ + "CPUUtilization$User":"

Percentage of time that the CPU has spent in the User state over the last 10 seconds.

", + "CPUUtilization$Nice":"

Percentage of time that the CPU has spent in the Nice state over the last 10 seconds.

", + "CPUUtilization$System":"

Percentage of time that the CPU has spent in the System state over the last 10 seconds.

", + "CPUUtilization$Idle":"

Percentage of time that the CPU has spent in the Idle state over the last 10 seconds.

", + "CPUUtilization$IOWait":"

Percentage of time that the CPU has spent in the I/O Wait state over the last 10 seconds.

", + "CPUUtilization$IRQ":"

Percentage of time that the CPU has spent in the IRQ state over the last 10 seconds.

", + "CPUUtilization$SoftIRQ":"

Percentage of time that the CPU has spent in the SoftIRQ state over the last 10 seconds.

", + "Latency$P999":"

The average latency for the slowest 0.1 percent of requests over the last 10 seconds.

", + "Latency$P99":"

The average latency for the slowest 1 percent of requests over the last 10 seconds.

", + "Latency$P95":"

The average latency for the slowest 5 percent of requests over the last 10 seconds.

", + "Latency$P90":"

The average latency for the slowest 10 percent of requests over the last 10 seconds.

", + "Latency$P85":"

The average latency for the slowest 15 percent of requests over the last 10 seconds.

", + "Latency$P75":"

The average latency for the slowest 25 percent of requests over the last 10 seconds.

", + "Latency$P50":"

The average latency for the slowest 50 percent of requests over the last 10 seconds.

", + "Latency$P10":"

The average latency for the slowest 90 percent of requests over the last 10 seconds.

" } }, - "NullableInteger": { - "base": null, - "refs": { - "ApplicationMetrics$Duration": "

The amount of time that the metrics cover (usually 10 seconds). For example, you might have 5 requests (request_count) within the most recent time slice of 10 seconds (duration).

", - "InstanceHealthSummary$NoData": "

Grey. AWS Elastic Beanstalk and the health agent are reporting no data on an instance.

", - "InstanceHealthSummary$Unknown": "

Grey. AWS Elastic Beanstalk and the health agent are reporting an insufficient amount of data on an instance.

", - "InstanceHealthSummary$Pending": "

Grey. An operation is in progress on an instance within the command timeout.

", - "InstanceHealthSummary$Ok": "

Green. An instance is passing health checks and the health agent is not reporting any problems.

", - "InstanceHealthSummary$Info": "

Green. An operation is in progress on an instance.

", - "InstanceHealthSummary$Warning": "

Yellow. The health agent is reporting a moderate number of request failures or other issues for an instance or environment.

", - "InstanceHealthSummary$Degraded": "

Red. The health agent is reporting a high number of request failures or other issues for an instance or environment.

", - "InstanceHealthSummary$Severe": "

Red. The health agent is reporting a very high number of request failures or other issues for an instance or environment.

", - "StatusCodes$Status2xx": "

The percentage of requests over the last 10 seconds that resulted in a 2xx (200, 201, etc.) status code.

", - "StatusCodes$Status3xx": "

The percentage of requests over the last 10 seconds that resulted in a 3xx (300, 301, etc.) status code.

", - "StatusCodes$Status4xx": "

The percentage of requests over the last 10 seconds that resulted in a 4xx (400, 401, etc.) status code.

", - "StatusCodes$Status5xx": "

The percentage of requests over the last 10 seconds that resulted in a 5xx (500, 501, etc.) status code.

" + "NullableInteger":{ + "base":null, + "refs":{ + "ApplicationMetrics$Duration":"

The amount of time that the metrics cover (usually 10 seconds). For example, you might have 5 requests (request_count) within the most recent time slice of 10 seconds (duration).

", + "InstanceHealthSummary$NoData":"

Grey. AWS Elastic Beanstalk and the health agent are reporting no data on an instance.

", + "InstanceHealthSummary$Unknown":"

Grey. AWS Elastic Beanstalk and the health agent are reporting an insufficient amount of data on an instance.

", + "InstanceHealthSummary$Pending":"

Grey. An operation is in progress on an instance within the command timeout.

", + "InstanceHealthSummary$Ok":"

Green. An instance is passing health checks and the health agent is not reporting any problems.

", + "InstanceHealthSummary$Info":"

Green. An operation is in progress on an instance.

", + "InstanceHealthSummary$Warning":"

Yellow. The health agent is reporting a moderate number of request failures or other issues for an instance or environment.

", + "InstanceHealthSummary$Degraded":"

Red. The health agent is reporting a high number of request failures or other issues for an instance or environment.

", + "InstanceHealthSummary$Severe":"

Red. The health agent is reporting a very high number of request failures or other issues for an instance or environment.

", + "StatusCodes$Status2xx":"

The percentage of requests over the last 10 seconds that resulted in a 2xx (200, 201, etc.) status code.

", + "StatusCodes$Status3xx":"

The percentage of requests over the last 10 seconds that resulted in a 3xx (300, 301, etc.) status code.

", + "StatusCodes$Status4xx":"

The percentage of requests over the last 10 seconds that resulted in a 4xx (400, 401, etc.) status code.

", + "StatusCodes$Status5xx":"

The percentage of requests over the last 10 seconds that resulted in a 5xx (500, 501, etc.) status code.

" } }, - "NullableLong": { - "base": null, - "refs": { - "Deployment$DeploymentId": "

The ID of the deployment. This number increases by one each time that you deploy source code or change instance configuration settings.

" + "NullableLong":{ + "base":null, + "refs":{ + "Deployment$DeploymentId":"

The ID of the deployment. This number increases by one each time that you deploy source code or change instance configuration settings.

" } }, - "OperatingSystemName": { - "base": null, - "refs": { - "PlatformDescription$OperatingSystemName": "

The operating system used by the platform.

", - "PlatformSummary$OperatingSystemName": "

The operating system used by the platform.

" + "OperatingSystemName":{ + "base":null, + "refs":{ + "PlatformDescription$OperatingSystemName":"

The operating system used by the platform.

", + "PlatformSummary$OperatingSystemName":"

The operating system used by the platform.

" } }, - "OperatingSystemVersion": { - "base": null, - "refs": { - "PlatformDescription$OperatingSystemVersion": "

The version of the operating system used by the platform.

", - "PlatformSummary$OperatingSystemVersion": "

The version of the operating system used by the platform.

" + "OperatingSystemVersion":{ + "base":null, + "refs":{ + "PlatformDescription$OperatingSystemVersion":"

The version of the operating system used by the platform.

", + "PlatformSummary$OperatingSystemVersion":"

The version of the operating system used by the platform.

" } }, - "OperationInProgressException": { - "base": "

Unable to perform the specified operation because another operation that effects an element in this activity is already in progress.

", - "refs": { + "OperationInProgressException":{ + "base":"

Unable to perform the specified operation because another operation that effects an element in this activity is already in progress.

", + "refs":{ } }, - "OptionNamespace": { - "base": null, - "refs": { - "ConfigurationOptionDescription$Namespace": "

A unique namespace identifying the option's associated AWS resource.

", - "ConfigurationOptionSetting$Namespace": "

A unique namespace identifying the option's associated AWS resource.

", - "OptionSpecification$Namespace": "

A unique namespace identifying the option's associated AWS resource.

", - "ValidationMessage$Namespace": "

The namespace to which the option belongs.

" + "OptionNamespace":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$Namespace":"

A unique namespace identifying the option's associated AWS resource.

", + "ConfigurationOptionSetting$Namespace":"

A unique namespace identifying the option's associated AWS resource.

", + "OptionSpecification$Namespace":"

A unique namespace identifying the option's associated AWS resource.

", + "ValidationMessage$Namespace":"

The namespace to which the option belongs.

" } }, - "OptionRestrictionMaxLength": { - "base": null, - "refs": { - "ConfigurationOptionDescription$MaxLength": "

If specified, the configuration option must be a string value no longer than this value.

" + "OptionRestrictionMaxLength":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$MaxLength":"

If specified, the configuration option must be a string value no longer than this value.

" } }, - "OptionRestrictionMaxValue": { - "base": null, - "refs": { - "ConfigurationOptionDescription$MaxValue": "

If specified, the configuration option must be a numeric value less than this value.

" + "OptionRestrictionMaxValue":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$MaxValue":"

If specified, the configuration option must be a numeric value less than this value.

" } }, - "OptionRestrictionMinValue": { - "base": null, - "refs": { - "ConfigurationOptionDescription$MinValue": "

If specified, the configuration option must be a numeric value greater than this value.

" + "OptionRestrictionMinValue":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$MinValue":"

If specified, the configuration option must be a numeric value greater than this value.

" } }, - "OptionRestrictionRegex": { - "base": "

A regular expression representing a restriction on a string configuration option value.

", - "refs": { - "ConfigurationOptionDescription$Regex": "

If specified, the configuration option must be a string value that satisfies this regular expression.

" + "OptionRestrictionRegex":{ + "base":"

A regular expression representing a restriction on a string configuration option value.

", + "refs":{ + "ConfigurationOptionDescription$Regex":"

If specified, the configuration option must be a string value that satisfies this regular expression.

" } }, - "OptionSpecification": { - "base": "

A specification identifying an individual configuration option.

", - "refs": { - "OptionsSpecifierList$member": null + "OptionSpecification":{ + "base":"

A specification identifying an individual configuration option.

", + "refs":{ + "OptionsSpecifierList$member":null } }, - "OptionsSpecifierList": { - "base": null, - "refs": { - "CreateEnvironmentMessage$OptionsToRemove": "

A list of custom user-defined configuration options to remove from the configuration set for this new environment.

", - "DescribeConfigurationOptionsMessage$Options": "

If specified, restricts the descriptions to only the specified options.

", - "UpdateConfigurationTemplateMessage$OptionsToRemove": "

A list of configuration options to remove from the configuration set.

Constraint: You can remove only UserDefined configuration options.

", - "UpdateEnvironmentMessage$OptionsToRemove": "

A list of custom user-defined configuration options to remove from the configuration set for this environment.

" + "OptionsSpecifierList":{ + "base":null, + "refs":{ + "CreateEnvironmentMessage$OptionsToRemove":"

A list of custom user-defined configuration options to remove from the configuration set for this new environment.

", + "DescribeConfigurationOptionsMessage$Options":"

If specified, restricts the descriptions to only the specified options.

", + "UpdateConfigurationTemplateMessage$OptionsToRemove":"

A list of configuration options to remove from the configuration set.

Constraint: You can remove only UserDefined configuration options.

", + "UpdateEnvironmentMessage$OptionsToRemove":"

A list of custom user-defined configuration options to remove from the configuration set for this environment.

" } }, - "PlatformArn": { - "base": null, - "refs": { - "ConfigurationOptionsDescription$PlatformArn": "

The ARN of the platform.

", - "ConfigurationSettingsDescription$PlatformArn": "

The ARN of the platform.

", - "CreateConfigurationTemplateMessage$PlatformArn": "

The ARN of the custom platform.

", - "CreateEnvironmentMessage$PlatformArn": "

The ARN of the platform.

", - "DeletePlatformVersionRequest$PlatformArn": "

The ARN of the version of the custom platform.

", - "DescribeConfigurationOptionsMessage$PlatformArn": "

The ARN of the custom platform.

", - "DescribeEventsMessage$PlatformArn": "

The ARN of the version of the custom platform.

", - "DescribePlatformVersionRequest$PlatformArn": "

The ARN of the version of the platform.

", - "EnvironmentDescription$PlatformArn": "

The ARN of the platform.

", - "EventDescription$PlatformArn": "

The ARN of the platform.

", - "PlatformDescription$PlatformArn": "

The ARN of the platform.

", - "PlatformSummary$PlatformArn": "

The ARN of the platform.

", - "UpdateEnvironmentMessage$PlatformArn": "

The ARN of the platform, if used.

" + "PlatformArn":{ + "base":null, + "refs":{ + "ConfigurationOptionsDescription$PlatformArn":"

The ARN of the platform.

", + "ConfigurationSettingsDescription$PlatformArn":"

The ARN of the platform.

", + "CreateConfigurationTemplateMessage$PlatformArn":"

The ARN of the custom platform.

", + "CreateEnvironmentMessage$PlatformArn":"

The ARN of the platform.

", + "DeletePlatformVersionRequest$PlatformArn":"

The ARN of the version of the custom platform.

", + "DescribeConfigurationOptionsMessage$PlatformArn":"

The ARN of the custom platform.

", + "DescribeEventsMessage$PlatformArn":"

The ARN of the version of the custom platform.

", + "DescribePlatformVersionRequest$PlatformArn":"

The ARN of the version of the platform.

", + "EnvironmentDescription$PlatformArn":"

The ARN of the platform.

", + "EventDescription$PlatformArn":"

The ARN of the platform.

", + "PlatformDescription$PlatformArn":"

The ARN of the platform.

", + "PlatformSummary$PlatformArn":"

The ARN of the platform.

", + "UpdateEnvironmentMessage$PlatformArn":"

The ARN of the platform, if used.

" } }, - "PlatformCategory": { - "base": null, - "refs": { - "PlatformDescription$PlatformCategory": "

The category of the platform.

", - "PlatformSummary$PlatformCategory": "

The category of platform.

" + "PlatformCategory":{ + "base":null, + "refs":{ + "PlatformDescription$PlatformCategory":"

The category of the platform.

", + "PlatformSummary$PlatformCategory":"

The category of platform.

" } }, - "PlatformDescription": { - "base": "

Detailed information about a platform.

", - "refs": { - "DescribePlatformVersionResult$PlatformDescription": "

Detailed information about the version of the platform.

" + "PlatformDescription":{ + "base":"

Detailed information about a platform.

", + "refs":{ + "DescribePlatformVersionResult$PlatformDescription":"

Detailed information about the version of the platform.

" } }, - "PlatformFilter": { - "base": "

Specify criteria to restrict the results when listing custom platforms.

The filter is evaluated as the expression:

Type Operator Values[i]

", - "refs": { - "PlatformFilters$member": null + "PlatformFilter":{ + "base":"

Specify criteria to restrict the results when listing custom platforms.

The filter is evaluated as the expression:

Type Operator Values[i]

", + "refs":{ + "PlatformFilters$member":null } }, - "PlatformFilterOperator": { - "base": null, - "refs": { - "PlatformFilter$Operator": "

The operator to apply to the Type with each of the Values.

Valid Values: = (equal to) | != (not equal to) | < (less than) | <= (less than or equal to) | > (greater than) | >= (greater than or equal to) | contains | begins_with | ends_with

" + "PlatformFilterOperator":{ + "base":null, + "refs":{ + "PlatformFilter$Operator":"

The operator to apply to the Type with each of the Values.

Valid Values: = (equal to) | != (not equal to) | < (less than) | <= (less than or equal to) | > (greater than) | >= (greater than or equal to) | contains | begins_with | ends_with

" } }, - "PlatformFilterType": { - "base": null, - "refs": { - "PlatformFilter$Type": "

The custom platform attribute to which the filter values are applied.

Valid Values: PlatformName | PlatformVersion | PlatformStatus | PlatformOwner

" + "PlatformFilterType":{ + "base":null, + "refs":{ + "PlatformFilter$Type":"

The custom platform attribute to which the filter values are applied.

Valid Values: PlatformName | PlatformVersion | PlatformStatus | PlatformOwner

" } }, - "PlatformFilterValue": { - "base": null, - "refs": { - "PlatformFilterValueList$member": null + "PlatformFilterValue":{ + "base":null, + "refs":{ + "PlatformFilterValueList$member":null } }, - "PlatformFilterValueList": { - "base": null, - "refs": { - "PlatformFilter$Values": "

The list of values applied to the custom platform attribute.

" + "PlatformFilterValueList":{ + "base":null, + "refs":{ + "PlatformFilter$Values":"

The list of values applied to the custom platform attribute.

" } }, - "PlatformFilters": { - "base": null, - "refs": { - "ListPlatformVersionsRequest$Filters": "

List only the platforms where the platform member value relates to one of the supplied values.

" + "PlatformFilters":{ + "base":null, + "refs":{ + "ListPlatformVersionsRequest$Filters":"

List only the platforms where the platform member value relates to one of the supplied values.

" } }, - "PlatformFramework": { - "base": "

A framework supported by the custom platform.

", - "refs": { - "PlatformFrameworks$member": null + "PlatformFramework":{ + "base":"

A framework supported by the custom platform.

", + "refs":{ + "PlatformFrameworks$member":null } }, - "PlatformFrameworks": { - "base": null, - "refs": { - "PlatformDescription$Frameworks": "

The frameworks supported by the platform.

" + "PlatformFrameworks":{ + "base":null, + "refs":{ + "PlatformDescription$Frameworks":"

The frameworks supported by the platform.

" } }, - "PlatformMaxRecords": { - "base": null, - "refs": { - "ListPlatformVersionsRequest$MaxRecords": "

The maximum number of platform values returned in one call.

" + "PlatformMaxRecords":{ + "base":null, + "refs":{ + "ListPlatformVersionsRequest$MaxRecords":"

The maximum number of platform values returned in one call.

" } }, - "PlatformName": { - "base": null, - "refs": { - "CreatePlatformVersionRequest$PlatformName": "

The name of your custom platform.

", - "PlatformDescription$PlatformName": "

The name of the platform.

" + "PlatformName":{ + "base":null, + "refs":{ + "CreatePlatformVersionRequest$PlatformName":"

The name of your custom platform.

", + "PlatformDescription$PlatformName":"

The name of the platform.

" } }, - "PlatformOwner": { - "base": null, - "refs": { - "PlatformDescription$PlatformOwner": "

The AWS account ID of the person who created the platform.

", - "PlatformSummary$PlatformOwner": "

The AWS account ID of the person who created the platform.

" + "PlatformOwner":{ + "base":null, + "refs":{ + "PlatformDescription$PlatformOwner":"

The AWS account ID of the person who created the platform.

", + "PlatformSummary$PlatformOwner":"

The AWS account ID of the person who created the platform.

" } }, - "PlatformProgrammingLanguage": { - "base": "

A programming language supported by the platform.

", - "refs": { - "PlatformProgrammingLanguages$member": null + "PlatformProgrammingLanguage":{ + "base":"

A programming language supported by the platform.

", + "refs":{ + "PlatformProgrammingLanguages$member":null } }, - "PlatformProgrammingLanguages": { - "base": null, - "refs": { - "PlatformDescription$ProgrammingLanguages": "

The programming languages supported by the platform.

" + "PlatformProgrammingLanguages":{ + "base":null, + "refs":{ + "PlatformDescription$ProgrammingLanguages":"

The programming languages supported by the platform.

" } }, - "PlatformStatus": { - "base": null, - "refs": { - "PlatformDescription$PlatformStatus": "

The status of the platform.

", - "PlatformSummary$PlatformStatus": "

The status of the platform. You can create an environment from the platform once it is ready.

" + "PlatformStatus":{ + "base":null, + "refs":{ + "PlatformDescription$PlatformStatus":"

The status of the platform.

", + "PlatformSummary$PlatformStatus":"

The status of the platform. You can create an environment from the platform once it is ready.

" } }, - "PlatformSummary": { - "base": "

Detailed information about a platform.

", - "refs": { - "CreatePlatformVersionResult$PlatformSummary": "

Detailed information about the new version of the custom platform.

", - "DeletePlatformVersionResult$PlatformSummary": "

Detailed information about the version of the custom platform.

", - "PlatformSummaryList$member": null + "PlatformSummary":{ + "base":"

Detailed information about a platform.

", + "refs":{ + "CreatePlatformVersionResult$PlatformSummary":"

Detailed information about the new version of the custom platform.

", + "DeletePlatformVersionResult$PlatformSummary":"

Detailed information about the version of the custom platform.

", + "PlatformSummaryList$member":null } }, - "PlatformSummaryList": { - "base": null, - "refs": { - "ListPlatformVersionsResult$PlatformSummaryList": "

Detailed information about the platforms.

" + "PlatformSummaryList":{ + "base":null, + "refs":{ + "ListPlatformVersionsResult$PlatformSummaryList":"

Detailed information about the platforms.

" } }, - "PlatformVersion": { - "base": null, - "refs": { - "CreatePlatformVersionRequest$PlatformVersion": "

The number, such as 1.0.2, for the new platform version.

", - "PlatformDescription$PlatformVersion": "

The version of the platform.

" + "PlatformVersion":{ + "base":null, + "refs":{ + "CreatePlatformVersionRequest$PlatformVersion":"

The number, such as 1.0.2, for the new platform version.

", + "PlatformDescription$PlatformVersion":"

The version of the platform.

" } }, - "PlatformVersionStillReferencedException": { - "base": "

You cannot delete the platform version because there are still environments running on it.

", - "refs": { + "PlatformVersionStillReferencedException":{ + "base":"

You cannot delete the platform version because there are still environments running on it.

", + "refs":{ } }, - "Queue": { - "base": "

Describes a queue.

", - "refs": { - "QueueList$member": null + "Queue":{ + "base":"

Describes a queue.

", + "refs":{ + "QueueList$member":null } }, - "QueueList": { - "base": null, - "refs": { - "EnvironmentResourceDescription$Queues": "

The queues used by this environment.

" + "QueueList":{ + "base":null, + "refs":{ + "EnvironmentResourceDescription$Queues":"

The queues used by this environment.

" } }, - "RebuildEnvironmentMessage": { - "base": "

", - "refs": { + "RebuildEnvironmentMessage":{ + "base":"

", + "refs":{ } }, - "RefreshedAt": { - "base": null, - "refs": { - "DescribeEnvironmentHealthResult$RefreshedAt": "

The date and time that the health information was retrieved.

", - "DescribeInstancesHealthResult$RefreshedAt": "

The date and time that the health information was retrieved.

" + "RefreshedAt":{ + "base":null, + "refs":{ + "DescribeEnvironmentHealthResult$RefreshedAt":"

The date and time that the health information was retrieved.

", + "DescribeInstancesHealthResult$RefreshedAt":"

The date and time that the health information was retrieved.

" } }, - "RegexLabel": { - "base": null, - "refs": { - "OptionRestrictionRegex$Label": "

A unique name representing this regular expression.

" + "RegexLabel":{ + "base":null, + "refs":{ + "OptionRestrictionRegex$Label":"

A unique name representing this regular expression.

" } }, - "RegexPattern": { - "base": null, - "refs": { - "OptionRestrictionRegex$Pattern": "

The regular expression pattern that a string configuration option value with this restriction must match.

" + "RegexPattern":{ + "base":null, + "refs":{ + "OptionRestrictionRegex$Pattern":"

The regular expression pattern that a string configuration option value with this restriction must match.

" } }, - "RequestCount": { - "base": null, - "refs": { - "ApplicationMetrics$RequestCount": "

Average number of requests handled by the web server per second over the last 10 seconds.

" + "RequestCount":{ + "base":null, + "refs":{ + "ApplicationMetrics$RequestCount":"

Average number of requests handled by the web server per second over the last 10 seconds.

" } }, - "RequestEnvironmentInfoMessage": { - "base": "

Request to retrieve logs from an environment and store them in your Elastic Beanstalk storage bucket.

", - "refs": { + "RequestEnvironmentInfoMessage":{ + "base":"

Request to retrieve logs from an environment and store them in your Elastic Beanstalk storage bucket.

", + "refs":{ } }, - "RequestId": { - "base": null, - "refs": { - "DescribeEventsMessage$RequestId": "

If specified, AWS Elastic Beanstalk restricts the described events to include only those associated with this request ID.

", - "EventDescription$RequestId": "

The web service request ID for the activity of this event.

" + "RequestId":{ + "base":null, + "refs":{ + "DescribeEventsMessage$RequestId":"

If specified, AWS Elastic Beanstalk restricts the described events to include only those associated with this request ID.

", + "EventDescription$RequestId":"

The web service request ID for the activity of this event.

" } }, - "ResourceArn": { - "base": null, - "refs": { - "ListTagsForResourceMessage$ResourceArn": "

The Amazon Resource Name (ARN) of the resouce for which a tag list is requested.

Must be the ARN of an Elastic Beanstalk environment.

", - "ResourceTagsDescriptionMessage$ResourceArn": "

The Amazon Resource Name (ARN) of the resouce for which a tag list was requested.

", - "UpdateTagsForResourceMessage$ResourceArn": "

The Amazon Resource Name (ARN) of the resouce to be updated.

Must be the ARN of an Elastic Beanstalk environment.

" + "ResourceArn":{ + "base":null, + "refs":{ + "ListTagsForResourceMessage$ResourceArn":"

The Amazon Resource Name (ARN) of the resouce for which a tag list is requested.

Must be the ARN of an Elastic Beanstalk environment.

", + "ResourceTagsDescriptionMessage$ResourceArn":"

The Amazon Resource Name (ARN) of the resouce for which a tag list was requested.

", + "UpdateTagsForResourceMessage$ResourceArn":"

The Amazon Resource Name (ARN) of the resouce to be updated.

Must be the ARN of an Elastic Beanstalk environment.

" } }, - "ResourceId": { - "base": null, - "refs": { - "AutoScalingGroup$Name": "

The name of the AutoScalingGroup .

", - "Instance$Id": "

The ID of the Amazon EC2 instance.

", - "LaunchConfiguration$Name": "

The name of the launch configuration.

", - "LoadBalancer$Name": "

The name of the LoadBalancer.

", - "Trigger$Name": "

The name of the trigger.

" + "ResourceId":{ + "base":null, + "refs":{ + "AutoScalingGroup$Name":"

The name of the AutoScalingGroup .

", + "Instance$Id":"

The ID of the Amazon EC2 instance.

", + "LaunchConfiguration$Name":"

The name of the launch configuration.

", + "LoadBalancer$Name":"

The name of the LoadBalancer.

", + "Trigger$Name":"

The name of the trigger.

" } }, - "ResourceName": { - "base": null, - "refs": { - "ConfigurationOptionSetting$ResourceName": "

A unique resource name for a time-based scaling configuration option.

", - "OptionSpecification$ResourceName": "

A unique resource name for a time-based scaling configuration option.

" + "ResourceName":{ + "base":null, + "refs":{ + "ConfigurationOptionSetting$ResourceName":"

A unique resource name for a time-based scaling configuration option.

", + "OptionSpecification$ResourceName":"

A unique resource name for a time-based scaling configuration option.

" } }, - "ResourceNotFoundException": { - "base": "

A resource doesn't exist for the specified Amazon Resource Name (ARN).

", - "refs": { + "ResourceNotFoundException":{ + "base":"

A resource doesn't exist for the specified Amazon Resource Name (ARN).

", + "refs":{ } }, - "ResourceQuota": { - "base": "

The AWS Elastic Beanstalk quota information for a single resource type in an AWS account. It reflects the resource's limits for this account.

", - "refs": { - "ResourceQuotas$ApplicationQuota": "

The quota for applications in the AWS account.

", - "ResourceQuotas$ApplicationVersionQuota": "

The quota for application versions in the AWS account.

", - "ResourceQuotas$EnvironmentQuota": "

The quota for environments in the AWS account.

", - "ResourceQuotas$ConfigurationTemplateQuota": "

The quota for configuration templates in the AWS account.

", - "ResourceQuotas$CustomPlatformQuota": "

The quota for custom platforms in the AWS account.

" + "ResourceQuota":{ + "base":"

The AWS Elastic Beanstalk quota information for a single resource type in an AWS account. It reflects the resource's limits for this account.

", + "refs":{ + "ResourceQuotas$ApplicationQuota":"

The quota for applications in the AWS account.

", + "ResourceQuotas$ApplicationVersionQuota":"

The quota for application versions in the AWS account.

", + "ResourceQuotas$EnvironmentQuota":"

The quota for environments in the AWS account.

", + "ResourceQuotas$ConfigurationTemplateQuota":"

The quota for configuration templates in the AWS account.

", + "ResourceQuotas$CustomPlatformQuota":"

The quota for custom platforms in the AWS account.

" } }, - "ResourceQuotas": { - "base": "

A set of per-resource AWS Elastic Beanstalk quotas associated with an AWS account. They reflect Elastic Beanstalk resource limits for this account.

", - "refs": { - "DescribeAccountAttributesResult$ResourceQuotas": "

The Elastic Beanstalk resource quotas associated with the calling AWS account.

" + "ResourceQuotas":{ + "base":"

A set of per-resource AWS Elastic Beanstalk quotas associated with an AWS account. They reflect Elastic Beanstalk resource limits for this account.

", + "refs":{ + "DescribeAccountAttributesResult$ResourceQuotas":"

The Elastic Beanstalk resource quotas associated with the calling AWS account.

" } }, - "ResourceTagsDescriptionMessage": { - "base": null, - "refs": { + "ResourceTagsDescriptionMessage":{ + "base":null, + "refs":{ } }, - "ResourceTypeNotSupportedException": { - "base": "

The type of the specified Amazon Resource Name (ARN) isn't supported for this operation.

", - "refs": { + "ResourceTypeNotSupportedException":{ + "base":"

The type of the specified Amazon Resource Name (ARN) isn't supported for this operation.

", + "refs":{ } }, - "RestartAppServerMessage": { - "base": "

", - "refs": { + "RestartAppServerMessage":{ + "base":"

", + "refs":{ } }, - "RetrieveEnvironmentInfoMessage": { - "base": "

Request to download logs retrieved with RequestEnvironmentInfo.

", - "refs": { + "RetrieveEnvironmentInfoMessage":{ + "base":"

Request to download logs retrieved with RequestEnvironmentInfo.

", + "refs":{ } }, - "RetrieveEnvironmentInfoResultMessage": { - "base": "

Result message containing a description of the requested environment info.

", - "refs": { + "RetrieveEnvironmentInfoResultMessage":{ + "base":"

Result message containing a description of the requested environment info.

", + "refs":{ } }, - "S3Bucket": { - "base": null, - "refs": { - "CreateStorageLocationResultMessage$S3Bucket": "

The name of the Amazon S3 bucket created.

", - "S3Location$S3Bucket": "

The Amazon S3 bucket where the data is located.

" + "S3Bucket":{ + "base":null, + "refs":{ + "CreateStorageLocationResultMessage$S3Bucket":"

The name of the Amazon S3 bucket created.

", + "S3Location$S3Bucket":"

The Amazon S3 bucket where the data is located.

" } }, - "S3Key": { - "base": null, - "refs": { - "S3Location$S3Key": "

The Amazon S3 key where the data is located.

" + "S3Key":{ + "base":null, + "refs":{ + "S3Location$S3Key":"

The Amazon S3 key where the data is located.

" } }, - "S3Location": { - "base": "

The bucket and key of an item stored in Amazon S3.

", - "refs": { - "ApplicationVersionDescription$SourceBundle": "

The storage location of the application version's source bundle in Amazon S3.

", - "CreateApplicationVersionMessage$SourceBundle": "

The Amazon S3 bucket and key that identify the location of the source bundle for this version.

The Amazon S3 bucket must be in the same region as the environment.

Specify a source bundle in S3 or a commit in an AWS CodeCommit repository (with SourceBuildInformation), but not both. If neither SourceBundle nor SourceBuildInformation are provided, Elastic Beanstalk uses a sample application.

", - "CreatePlatformVersionRequest$PlatformDefinitionBundle": "

The location of the platform definition archive in Amazon S3.

" + "S3Location":{ + "base":"

The bucket and key of an item stored in Amazon S3.

", + "refs":{ + "ApplicationVersionDescription$SourceBundle":"

The storage location of the application version's source bundle in Amazon S3.

", + "CreateApplicationVersionMessage$SourceBundle":"

The Amazon S3 bucket and key that identify the location of the source bundle for this version.

The Amazon S3 bucket must be in the same region as the environment.

Specify a source bundle in S3 or a commit in an AWS CodeCommit repository (with SourceBuildInformation), but not both. If neither SourceBundle nor SourceBuildInformation are provided, Elastic Beanstalk uses a sample application.

", + "CreatePlatformVersionRequest$PlatformDefinitionBundle":"

The location of the platform definition archive in Amazon S3.

" } }, - "S3LocationNotInServiceRegionException": { - "base": "

The specified S3 bucket does not belong to the S3 region in which the service is running. The following regions are supported:

  • IAD/us-east-1

  • PDX/us-west-2

  • DUB/eu-west-1

", - "refs": { + "S3LocationNotInServiceRegionException":{ + "base":"

The specified S3 bucket does not belong to the S3 region in which the service is running. The following regions are supported:

  • IAD/us-east-1

  • PDX/us-west-2

  • DUB/eu-west-1

", + "refs":{ } }, - "S3SubscriptionRequiredException": { - "base": "

The specified account does not have a subscription to Amazon S3.

", - "refs": { + "S3SubscriptionRequiredException":{ + "base":"

The specified account does not have a subscription to Amazon S3.

", + "refs":{ } }, - "SampleTimestamp": { - "base": null, - "refs": { - "EnvironmentInfoDescription$SampleTimestamp": "

The time stamp when this information was retrieved.

" + "SampleTimestamp":{ + "base":null, + "refs":{ + "EnvironmentInfoDescription$SampleTimestamp":"

The time stamp when this information was retrieved.

" } }, - "SingleInstanceHealth": { - "base": "

Detailed health information about an Amazon EC2 instance in your Elastic Beanstalk environment.

", - "refs": { - "InstanceHealthList$member": null + "SingleInstanceHealth":{ + "base":"

Detailed health information about an Amazon EC2 instance in your Elastic Beanstalk environment.

", + "refs":{ + "InstanceHealthList$member":null } }, - "SolutionStackDescription": { - "base": "

Describes the solution stack.

", - "refs": { - "AvailableSolutionStackDetailsList$member": null + "SolutionStackDescription":{ + "base":"

Describes the solution stack.

", + "refs":{ + "AvailableSolutionStackDetailsList$member":null } }, - "SolutionStackFileTypeList": { - "base": null, - "refs": { - "SolutionStackDescription$PermittedFileTypes": "

The permitted file types allowed for a solution stack.

" + "SolutionStackFileTypeList":{ + "base":null, + "refs":{ + "SolutionStackDescription$PermittedFileTypes":"

The permitted file types allowed for a solution stack.

" } }, - "SolutionStackName": { - "base": null, - "refs": { - "AvailableSolutionStackNamesList$member": null, - "ConfigurationOptionsDescription$SolutionStackName": "

The name of the solution stack these configuration options belong to.

", - "ConfigurationSettingsDescription$SolutionStackName": "

The name of the solution stack this configuration set uses.

", - "CreateConfigurationTemplateMessage$SolutionStackName": "

The name of the solution stack used by this configuration. The solution stack specifies the operating system, architecture, and application server for a configuration template. It determines the set of configuration options as well as the possible and default values.

Use ListAvailableSolutionStacks to obtain a list of available solution stacks.

A solution stack name or a source configuration parameter must be specified, otherwise AWS Elastic Beanstalk returns an InvalidParameterValue error.

If a solution stack name is not specified and the source configuration parameter is specified, AWS Elastic Beanstalk uses the same solution stack as the source configuration template.

", - "CreateEnvironmentMessage$SolutionStackName": "

This is an alternative to specifying a template name. If specified, AWS Elastic Beanstalk sets the configuration values to the default values associated with the specified solution stack.

", - "DescribeConfigurationOptionsMessage$SolutionStackName": "

The name of the solution stack whose configuration options you want to describe.

", - "EnvironmentDescription$SolutionStackName": "

The name of the SolutionStack deployed with this environment.

", - "PlatformDescription$SolutionStackName": "

The name of the solution stack used by the platform.

", - "SolutionStackDescription$SolutionStackName": "

The name of the solution stack.

", - "UpdateEnvironmentMessage$SolutionStackName": "

This specifies the platform version that the environment will run after the environment is updated.

" + "SolutionStackName":{ + "base":null, + "refs":{ + "AvailableSolutionStackNamesList$member":null, + "ConfigurationOptionsDescription$SolutionStackName":"

The name of the solution stack these configuration options belong to.

", + "ConfigurationSettingsDescription$SolutionStackName":"

The name of the solution stack this configuration set uses.

", + "CreateConfigurationTemplateMessage$SolutionStackName":"

The name of the solution stack used by this configuration. The solution stack specifies the operating system, architecture, and application server for a configuration template. It determines the set of configuration options as well as the possible and default values.

Use ListAvailableSolutionStacks to obtain a list of available solution stacks.

A solution stack name or a source configuration parameter must be specified, otherwise AWS Elastic Beanstalk returns an InvalidParameterValue error.

If a solution stack name is not specified and the source configuration parameter is specified, AWS Elastic Beanstalk uses the same solution stack as the source configuration template.

", + "CreateEnvironmentMessage$SolutionStackName":"

This is an alternative to specifying a template name. If specified, AWS Elastic Beanstalk sets the configuration values to the default values associated with the specified solution stack.

For a list of current solution stacks, see Elastic Beanstalk Supported Platforms.

", + "DescribeConfigurationOptionsMessage$SolutionStackName":"

The name of the solution stack whose configuration options you want to describe.

", + "EnvironmentDescription$SolutionStackName":"

The name of the SolutionStack deployed with this environment.

", + "PlatformDescription$SolutionStackName":"

The name of the solution stack used by the platform.

", + "SolutionStackDescription$SolutionStackName":"

The name of the solution stack.

", + "UpdateEnvironmentMessage$SolutionStackName":"

This specifies the platform version that the environment will run after the environment is updated.

" } }, - "SourceBuildInformation": { - "base": "

Location of the source code for an application version.

", - "refs": { - "ApplicationVersionDescription$SourceBuildInformation": "

If the version's source code was retrieved from AWS CodeCommit, the location of the source code for the application version.

", - "CreateApplicationVersionMessage$SourceBuildInformation": "

Specify a commit in an AWS CodeCommit Git repository to use as the source code for the application version.

" + "SourceBuildInformation":{ + "base":"

Location of the source code for an application version.

", + "refs":{ + "ApplicationVersionDescription$SourceBuildInformation":"

If the version's source code was retrieved from AWS CodeCommit, the location of the source code for the application version.

", + "CreateApplicationVersionMessage$SourceBuildInformation":"

Specify a commit in an AWS CodeCommit Git repository to use as the source code for the application version.

" } }, - "SourceBundleDeletionException": { - "base": "

Unable to delete the Amazon S3 source bundle associated with the application version. The application version was deleted successfully.

", - "refs": { + "SourceBundleDeletionException":{ + "base":"

Unable to delete the Amazon S3 source bundle associated with the application version. The application version was deleted successfully.

", + "refs":{ } }, - "SourceConfiguration": { - "base": "

A specification for an environment configuration

", - "refs": { - "CreateConfigurationTemplateMessage$SourceConfiguration": "

If specified, AWS Elastic Beanstalk uses the configuration values from the specified configuration template to create a new configuration.

Values specified in the OptionSettings parameter of this call overrides any values obtained from the SourceConfiguration.

If no configuration template is found, returns an InvalidParameterValue error.

Constraint: If both the solution stack name parameter and the source configuration parameters are specified, the solution stack of the source configuration template must match the specified solution stack name or else AWS Elastic Beanstalk returns an InvalidParameterCombination error.

" + "SourceConfiguration":{ + "base":"

A specification for an environment configuration

", + "refs":{ + "CreateConfigurationTemplateMessage$SourceConfiguration":"

If specified, AWS Elastic Beanstalk uses the configuration values from the specified configuration template to create a new configuration.

Values specified in the OptionSettings parameter of this call overrides any values obtained from the SourceConfiguration.

If no configuration template is found, returns an InvalidParameterValue error.

Constraint: If both the solution stack name parameter and the source configuration parameters are specified, the solution stack of the source configuration template must match the specified solution stack name or else AWS Elastic Beanstalk returns an InvalidParameterCombination error.

" } }, - "SourceLocation": { - "base": null, - "refs": { - "SourceBuildInformation$SourceLocation": "

The location of the source code, as a formatted string, depending on the value of SourceRepository

  • For CodeCommit, the format is the repository name and commit ID, separated by a forward slash. For example, my-git-repo/265cfa0cf6af46153527f55d6503ec030551f57a.

  • For S3, the format is the S3 bucket name and object key, separated by a forward slash. For example, my-s3-bucket/Folders/my-source-file.

" + "SourceLocation":{ + "base":null, + "refs":{ + "SourceBuildInformation$SourceLocation":"

The location of the source code, as a formatted string, depending on the value of SourceRepository

  • For CodeCommit, the format is the repository name and commit ID, separated by a forward slash. For example, my-git-repo/265cfa0cf6af46153527f55d6503ec030551f57a.

  • For S3, the format is the S3 bucket name and object key, separated by a forward slash. For example, my-s3-bucket/Folders/my-source-file.

" } }, - "SourceRepository": { - "base": null, - "refs": { - "SourceBuildInformation$SourceRepository": "

Location where the repository is stored.

  • CodeCommit

  • S3

" + "SourceRepository":{ + "base":null, + "refs":{ + "SourceBuildInformation$SourceRepository":"

Location where the repository is stored.

  • CodeCommit

  • S3

" } }, - "SourceType": { - "base": null, - "refs": { - "SourceBuildInformation$SourceType": "

The type of repository.

  • Git

  • Zip

" + "SourceType":{ + "base":null, + "refs":{ + "SourceBuildInformation$SourceType":"

The type of repository.

  • Git

  • Zip

" } }, - "StatusCodes": { - "base": "

Represents the percentage of requests over the last 10 seconds that resulted in each type of status code response. For more information, see Status Code Definitions.

", - "refs": { - "ApplicationMetrics$StatusCodes": "

Represents the percentage of requests over the last 10 seconds that resulted in each type of status code response.

" + "StatusCodes":{ + "base":"

Represents the percentage of requests over the last 10 seconds that resulted in each type of status code response. For more information, see Status Code Definitions.

", + "refs":{ + "ApplicationMetrics$StatusCodes":"

Represents the percentage of requests over the last 10 seconds that resulted in each type of status code response.

" } }, - "String": { - "base": null, - "refs": { - "ApplicationResourceLifecycleConfig$ServiceRole": "

The ARN of an IAM service role that Elastic Beanstalk has permission to assume.

", - "ApplicationVersionDescription$BuildArn": "

Reference to the artifact from the AWS CodeBuild build.

", - "ApplyEnvironmentManagedActionRequest$EnvironmentName": "

The name of the target environment.

", - "ApplyEnvironmentManagedActionRequest$EnvironmentId": "

The environment ID of the target environment.

", - "ApplyEnvironmentManagedActionRequest$ActionId": "

The action ID of the scheduled managed action to execute.

", - "ApplyEnvironmentManagedActionResult$ActionId": "

The action ID of the managed action.

", - "ApplyEnvironmentManagedActionResult$ActionDescription": "

A description of the managed action.

", - "ApplyEnvironmentManagedActionResult$Status": "

The status of the managed action.

", - "BuildConfiguration$ArtifactName": "

The name of the artifact of the CodeBuild build. If provided, Elastic Beanstalk stores the build artifact in the S3 location S3-bucket/resources/application-name/codebuild/codebuild-version-label-artifact-name.zip. If not provided, Elastic Beanstalk stores the build artifact in the S3 location S3-bucket/resources/application-name/codebuild/codebuild-version-label.zip.

", - "Deployment$VersionLabel": "

The version label of the application version in the deployment.

", - "Deployment$Status": "

The status of the deployment:

  • In Progress : The deployment is in progress.

  • Deployed : The deployment succeeded.

  • Failed : The deployment failed.

", - "DescribeEnvironmentHealthResult$HealthStatus": "

The health status of the environment. For example, Ok.

", - "DescribeEnvironmentHealthResult$Color": "

The health color of the environment.

", - "DescribeEnvironmentManagedActionHistoryRequest$NextToken": "

The pagination token returned by a previous request.

", - "DescribeEnvironmentManagedActionHistoryResult$NextToken": "

A pagination token that you pass to DescribeEnvironmentManagedActionHistory to get the next page of results.

", - "DescribeEnvironmentManagedActionsRequest$EnvironmentName": "

The name of the target environment.

", - "DescribeEnvironmentManagedActionsRequest$EnvironmentId": "

The environment ID of the target environment.

", - "EnvironmentLink$LinkName": "

The name of the link.

", - "EnvironmentLink$EnvironmentName": "

The name of the linked environment (the dependency).

", - "EnvironmentTier$Name": "

The name of this environment tier.

", - "EnvironmentTier$Type": "

The type of this environment tier.

", - "EnvironmentTier$Version": "

The version of this environment tier.

", - "Listener$Protocol": "

The protocol that is used by the Listener.

", - "LoadBalancerDescription$LoadBalancerName": "

The name of the LoadBalancer.

", - "LoadBalancerDescription$Domain": "

The domain name of the LoadBalancer.

", - "ManagedAction$ActionId": "

A unique identifier for the managed action.

", - "ManagedAction$ActionDescription": "

A description of the managed action.

", - "ManagedActionHistoryItem$ActionId": "

A unique identifier for the managed action.

", - "ManagedActionHistoryItem$ActionDescription": "

A description of the managed action.

", - "ManagedActionHistoryItem$FailureDescription": "

If the action failed, a description of the failure.

", - "PlatformFramework$Name": "

The name of the framework.

", - "PlatformFramework$Version": "

The version of the framework.

", - "PlatformProgrammingLanguage$Name": "

The name of the programming language.

", - "PlatformProgrammingLanguage$Version": "

The version of the programming language.

", - "Queue$Name": "

The name of the queue.

", - "Queue$URL": "

The URL of the queue.

", - "SingleInstanceHealth$HealthStatus": "

Returns the health status of the specified instance. For more information, see Health Colors and Statuses.

", - "SingleInstanceHealth$Color": "

Represents the color indicator that gives you information about the health of the EC2 instance. For more information, see Health Colors and Statuses.

", - "SingleInstanceHealth$AvailabilityZone": "

The availability zone in which the instance runs.

", - "SingleInstanceHealth$InstanceType": "

The instance's type.

" + "String":{ + "base":null, + "refs":{ + "ApplicationResourceLifecycleConfig$ServiceRole":"

The ARN of an IAM service role that Elastic Beanstalk has permission to assume.

The ServiceRole property is required the first time that you provide a VersionLifecycleConfig for the application in one of the supporting calls (CreateApplication or UpdateApplicationResourceLifecycle). After you provide it once, in either one of the calls, Elastic Beanstalk persists the Service Role with the application, and you don't need to specify it again in subsequent UpdateApplicationResourceLifecycle calls. You can, however, specify it in subsequent calls to change the Service Role to another value.

", + "ApplicationVersionDescription$BuildArn":"

Reference to the artifact from the AWS CodeBuild build.

", + "ApplyEnvironmentManagedActionRequest$EnvironmentName":"

The name of the target environment.

", + "ApplyEnvironmentManagedActionRequest$EnvironmentId":"

The environment ID of the target environment.

", + "ApplyEnvironmentManagedActionRequest$ActionId":"

The action ID of the scheduled managed action to execute.

", + "ApplyEnvironmentManagedActionResult$ActionId":"

The action ID of the managed action.

", + "ApplyEnvironmentManagedActionResult$ActionDescription":"

A description of the managed action.

", + "ApplyEnvironmentManagedActionResult$Status":"

The status of the managed action.

", + "BuildConfiguration$ArtifactName":"

The name of the artifact of the CodeBuild build. If provided, Elastic Beanstalk stores the build artifact in the S3 location S3-bucket/resources/application-name/codebuild/codebuild-version-label-artifact-name.zip. If not provided, Elastic Beanstalk stores the build artifact in the S3 location S3-bucket/resources/application-name/codebuild/codebuild-version-label.zip.

", + "Deployment$VersionLabel":"

The version label of the application version in the deployment.

", + "Deployment$Status":"

The status of the deployment:

  • In Progress : The deployment is in progress.

  • Deployed : The deployment succeeded.

  • Failed : The deployment failed.

", + "DescribeEnvironmentHealthResult$HealthStatus":"

The health status of the environment. For example, Ok.

", + "DescribeEnvironmentHealthResult$Color":"

The health color of the environment.

", + "DescribeEnvironmentManagedActionHistoryRequest$NextToken":"

The pagination token returned by a previous request.

", + "DescribeEnvironmentManagedActionHistoryResult$NextToken":"

A pagination token that you pass to DescribeEnvironmentManagedActionHistory to get the next page of results.

", + "DescribeEnvironmentManagedActionsRequest$EnvironmentName":"

The name of the target environment.

", + "DescribeEnvironmentManagedActionsRequest$EnvironmentId":"

The environment ID of the target environment.

", + "EnvironmentLink$LinkName":"

The name of the link.

", + "EnvironmentLink$EnvironmentName":"

The name of the linked environment (the dependency).

", + "EnvironmentTier$Name":"

The name of this environment tier.

", + "EnvironmentTier$Type":"

The type of this environment tier.

", + "EnvironmentTier$Version":"

The version of this environment tier. When you don't set a value to it, Elastic Beanstalk uses the latest compatible worker tier version.

This member is deprecated. Any specific version that you set may become out of date. We recommend leaving it unspecified.

", + "Listener$Protocol":"

The protocol that is used by the Listener.

", + "LoadBalancerDescription$LoadBalancerName":"

The name of the LoadBalancer.

", + "LoadBalancerDescription$Domain":"

The domain name of the LoadBalancer.

", + "ManagedAction$ActionId":"

A unique identifier for the managed action.

", + "ManagedAction$ActionDescription":"

A description of the managed action.

", + "ManagedActionHistoryItem$ActionId":"

A unique identifier for the managed action.

", + "ManagedActionHistoryItem$ActionDescription":"

A description of the managed action.

", + "ManagedActionHistoryItem$FailureDescription":"

If the action failed, a description of the failure.

", + "PlatformFramework$Name":"

The name of the framework.

", + "PlatformFramework$Version":"

The version of the framework.

", + "PlatformProgrammingLanguage$Name":"

The name of the programming language.

", + "PlatformProgrammingLanguage$Version":"

The version of the programming language.

", + "Queue$Name":"

The name of the queue.

", + "Queue$URL":"

The URL of the queue.

", + "SingleInstanceHealth$HealthStatus":"

Returns the health status of the specified instance. For more information, see Health Colors and Statuses.

", + "SingleInstanceHealth$Color":"

Represents the color indicator that gives you information about the health of the EC2 instance. For more information, see Health Colors and Statuses.

", + "SingleInstanceHealth$AvailabilityZone":"

The availability zone in which the instance runs.

", + "SingleInstanceHealth$InstanceType":"

The instance's type.

" } }, - "SupportedAddon": { - "base": null, - "refs": { - "SupportedAddonList$member": null + "SupportedAddon":{ + "base":null, + "refs":{ + "SupportedAddonList$member":null } }, - "SupportedAddonList": { - "base": null, - "refs": { - "PlatformDescription$SupportedAddonList": "

The additions supported by the platform.

", - "PlatformSummary$SupportedAddonList": "

The additions associated with the platform.

" + "SupportedAddonList":{ + "base":null, + "refs":{ + "PlatformDescription$SupportedAddonList":"

The additions supported by the platform.

", + "PlatformSummary$SupportedAddonList":"

The additions associated with the platform.

" } }, - "SupportedTier": { - "base": null, - "refs": { - "SupportedTierList$member": null + "SupportedTier":{ + "base":null, + "refs":{ + "SupportedTierList$member":null } }, - "SupportedTierList": { - "base": null, - "refs": { - "PlatformDescription$SupportedTierList": "

The tiers supported by the platform.

", - "PlatformSummary$SupportedTierList": "

The tiers in which the platform runs.

" + "SupportedTierList":{ + "base":null, + "refs":{ + "PlatformDescription$SupportedTierList":"

The tiers supported by the platform.

", + "PlatformSummary$SupportedTierList":"

The tiers in which the platform runs.

" } }, - "SwapEnvironmentCNAMEsMessage": { - "base": "

Swaps the CNAMEs of two environments.

", - "refs": { + "SwapEnvironmentCNAMEsMessage":{ + "base":"

Swaps the CNAMEs of two environments.

", + "refs":{ } }, - "SystemStatus": { - "base": "

CPU utilization and load average metrics for an Amazon EC2 instance.

", - "refs": { - "SingleInstanceHealth$System": "

Operating system metrics from the instance.

" + "SystemStatus":{ + "base":"

CPU utilization and load average metrics for an Amazon EC2 instance.

", + "refs":{ + "SingleInstanceHealth$System":"

Operating system metrics from the instance.

" } }, - "Tag": { - "base": "

Describes a tag applied to a resource in an environment.

", - "refs": { - "TagList$member": null, - "Tags$member": null + "Tag":{ + "base":"

Describes a tag applied to a resource in an environment.

", + "refs":{ + "TagList$member":null, + "Tags$member":null } }, - "TagKey": { - "base": null, - "refs": { - "Tag$Key": "

The key of the tag.

", - "TagKeyList$member": null + "TagKey":{ + "base":null, + "refs":{ + "Tag$Key":"

The key of the tag.

", + "TagKeyList$member":null } }, - "TagKeyList": { - "base": null, - "refs": { - "UpdateTagsForResourceMessage$TagsToRemove": "

A list of tag keys to remove.

If a tag key doesn't exist, it is silently ignored.

" + "TagKeyList":{ + "base":null, + "refs":{ + "UpdateTagsForResourceMessage$TagsToRemove":"

A list of tag keys to remove.

If a tag key doesn't exist, it is silently ignored.

" } }, - "TagList": { - "base": null, - "refs": { - "ResourceTagsDescriptionMessage$ResourceTags": "

A list of tag key-value pairs.

", - "UpdateTagsForResourceMessage$TagsToAdd": "

A list of tags to add or update.

If a key of an existing tag is added, the tag's value is updated.

" + "TagList":{ + "base":null, + "refs":{ + "ResourceTagsDescriptionMessage$ResourceTags":"

A list of tag key-value pairs.

", + "UpdateTagsForResourceMessage$TagsToAdd":"

A list of tags to add or update.

If a key of an existing tag is added, the tag's value is updated.

" } }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

The value of the tag.

" + "TagValue":{ + "base":null, + "refs":{ + "Tag$Value":"

The value of the tag.

" } }, - "Tags": { - "base": null, - "refs": { - "CreateEnvironmentMessage$Tags": "

This specifies the tags applied to resources in the environment.

" + "Tags":{ + "base":null, + "refs":{ + "CreateEnvironmentMessage$Tags":"

This specifies the tags applied to resources in the environment.

" } }, - "TerminateEnvForce": { - "base": null, - "refs": { - "DeleteApplicationMessage$TerminateEnvByForce": "

When set to true, running environments will be terminated before deleting the application.

" + "TerminateEnvForce":{ + "base":null, + "refs":{ + "DeleteApplicationMessage$TerminateEnvByForce":"

When set to true, running environments will be terminated before deleting the application.

" } }, - "TerminateEnvironmentMessage": { - "base": "

Request to terminate an environment.

", - "refs": { + "TerminateEnvironmentMessage":{ + "base":"

Request to terminate an environment.

", + "refs":{ } }, - "TerminateEnvironmentResources": { - "base": null, - "refs": { - "TerminateEnvironmentMessage$TerminateResources": "

Indicates whether the associated AWS resources should shut down when the environment is terminated:

  • true: The specified environment as well as the associated AWS resources, such as Auto Scaling group and LoadBalancer, are terminated.

  • false: AWS Elastic Beanstalk resource management is removed from the environment, but the AWS resources continue to operate.

For more information, see the AWS Elastic Beanstalk User Guide.

Default: true

Valid Values: true | false

" + "TerminateEnvironmentResources":{ + "base":null, + "refs":{ + "TerminateEnvironmentMessage$TerminateResources":"

Indicates whether the associated AWS resources should shut down when the environment is terminated:

  • true: The specified environment as well as the associated AWS resources, such as Auto Scaling group and LoadBalancer, are terminated.

  • false: AWS Elastic Beanstalk resource management is removed from the environment, but the AWS resources continue to operate.

For more information, see the AWS Elastic Beanstalk User Guide.

Default: true

Valid Values: true | false

" } }, - "TimeFilterEnd": { - "base": null, - "refs": { - "DescribeEventsMessage$EndTime": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those that occur up to, but not including, the EndTime.

" + "TimeFilterEnd":{ + "base":null, + "refs":{ + "DescribeEventsMessage$EndTime":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those that occur up to, but not including, the EndTime.

" } }, - "TimeFilterStart": { - "base": null, - "refs": { - "DescribeEventsMessage$StartTime": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those that occur on or after this time.

" + "TimeFilterStart":{ + "base":null, + "refs":{ + "DescribeEventsMessage$StartTime":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those that occur on or after this time.

" } }, - "Timestamp": { - "base": null, - "refs": { - "ManagedAction$WindowStartTime": "

The start time of the maintenance window in which the managed action will execute.

", - "ManagedActionHistoryItem$ExecutedTime": "

The date and time that the action started executing.

", - "ManagedActionHistoryItem$FinishedTime": "

The date and time that the action finished executing.

" + "Timestamp":{ + "base":null, + "refs":{ + "ManagedAction$WindowStartTime":"

The start time of the maintenance window in which the managed action will execute.

", + "ManagedActionHistoryItem$ExecutedTime":"

The date and time that the action started executing.

", + "ManagedActionHistoryItem$FinishedTime":"

The date and time that the action finished executing.

" } }, - "Token": { - "base": null, - "refs": { - "ApplicationVersionDescriptionsMessage$NextToken": "

In a paginated request, the token that you can pass in a subsequent request to get the next response page.

", - "DescribeApplicationVersionsMessage$NextToken": "

For a paginated request. Specify a token from a previous response page to retrieve the next response page. All other parameter values must be identical to the ones specified in the initial request.

If no NextToken is specified, the first page is retrieved.

", - "DescribeEnvironmentsMessage$NextToken": "

For a paginated request. Specify a token from a previous response page to retrieve the next response page. All other parameter values must be identical to the ones specified in the initial request.

If no NextToken is specified, the first page is retrieved.

", - "DescribeEventsMessage$NextToken": "

Pagination token. If specified, the events return the next batch of results.

", - "EnvironmentDescriptionsMessage$NextToken": "

In a paginated request, the token that you can pass in a subsequent request to get the next response page.

", - "EventDescriptionsMessage$NextToken": "

If returned, this indicates that there are more results to obtain. Use this token in the next DescribeEvents call to get the next batch of events.

", - "ListPlatformVersionsRequest$NextToken": "

The starting index into the remaining list of platforms. Use the NextToken value from a previous ListPlatformVersion call.

", - "ListPlatformVersionsResult$NextToken": "

The starting index into the remaining list of platforms. if this value is not null, you can use it in a subsequent ListPlatformVersion call.

" + "Token":{ + "base":null, + "refs":{ + "ApplicationVersionDescriptionsMessage$NextToken":"

In a paginated request, the token that you can pass in a subsequent request to get the next response page.

", + "DescribeApplicationVersionsMessage$NextToken":"

For a paginated request. Specify a token from a previous response page to retrieve the next response page. All other parameter values must be identical to the ones specified in the initial request.

If no NextToken is specified, the first page is retrieved.

", + "DescribeEnvironmentsMessage$NextToken":"

For a paginated request. Specify a token from a previous response page to retrieve the next response page. All other parameter values must be identical to the ones specified in the initial request.

If no NextToken is specified, the first page is retrieved.

", + "DescribeEventsMessage$NextToken":"

Pagination token. If specified, the events return the next batch of results.

", + "EnvironmentDescriptionsMessage$NextToken":"

In a paginated request, the token that you can pass in a subsequent request to get the next response page.

", + "EventDescriptionsMessage$NextToken":"

If returned, this indicates that there are more results to obtain. Use this token in the next DescribeEvents call to get the next batch of events.

", + "ListPlatformVersionsRequest$NextToken":"

The starting index into the remaining list of platforms. Use the NextToken value from a previous ListPlatformVersion call.

", + "ListPlatformVersionsResult$NextToken":"

The starting index into the remaining list of platforms. if this value is not null, you can use it in a subsequent ListPlatformVersion call.

" } }, - "TooManyApplicationVersionsException": { - "base": "

The specified account has reached its limit of application versions.

", - "refs": { + "TooManyApplicationVersionsException":{ + "base":"

The specified account has reached its limit of application versions.

", + "refs":{ } }, - "TooManyApplicationsException": { - "base": "

The specified account has reached its limit of applications.

", - "refs": { + "TooManyApplicationsException":{ + "base":"

The specified account has reached its limit of applications.

", + "refs":{ } }, - "TooManyBucketsException": { - "base": "

The specified account has reached its limit of Amazon S3 buckets.

", - "refs": { + "TooManyBucketsException":{ + "base":"

The specified account has reached its limit of Amazon S3 buckets.

", + "refs":{ } }, - "TooManyConfigurationTemplatesException": { - "base": "

The specified account has reached its limit of configuration templates.

", - "refs": { + "TooManyConfigurationTemplatesException":{ + "base":"

The specified account has reached its limit of configuration templates.

", + "refs":{ } }, - "TooManyEnvironmentsException": { - "base": "

The specified account has reached its limit of environments.

", - "refs": { + "TooManyEnvironmentsException":{ + "base":"

The specified account has reached its limit of environments.

", + "refs":{ } }, - "TooManyPlatformsException": { - "base": "

You have exceeded the maximum number of allowed platforms associated with the account.

", - "refs": { + "TooManyPlatformsException":{ + "base":"

You have exceeded the maximum number of allowed platforms associated with the account.

", + "refs":{ } }, - "TooManyTagsException": { - "base": "

The number of tags in the resource would exceed the number of tags that each resource can have.

To calculate this, the operation considers both the number of tags the resource already has and the tags this operation would add if it succeeded.

", - "refs": { + "TooManyTagsException":{ + "base":"

The number of tags in the resource would exceed the number of tags that each resource can have.

To calculate this, the operation considers both the number of tags the resource already has and the tags this operation would add if it succeeded.

", + "refs":{ } }, - "Trigger": { - "base": "

Describes a trigger.

", - "refs": { - "TriggerList$member": null + "Trigger":{ + "base":"

Describes a trigger.

", + "refs":{ + "TriggerList$member":null } }, - "TriggerList": { - "base": null, - "refs": { - "EnvironmentResourceDescription$Triggers": "

The AutoScaling triggers in use by this environment.

" + "TriggerList":{ + "base":null, + "refs":{ + "EnvironmentResourceDescription$Triggers":"

The AutoScaling triggers in use by this environment.

" } }, - "UpdateApplicationMessage": { - "base": "

Request to update an application.

", - "refs": { + "UpdateApplicationMessage":{ + "base":"

Request to update an application.

", + "refs":{ } }, - "UpdateApplicationResourceLifecycleMessage": { - "base": null, - "refs": { + "UpdateApplicationResourceLifecycleMessage":{ + "base":null, + "refs":{ } }, - "UpdateApplicationVersionMessage": { - "base": "

", - "refs": { + "UpdateApplicationVersionMessage":{ + "base":"

", + "refs":{ } }, - "UpdateConfigurationTemplateMessage": { - "base": "

The result message containing the options for the specified solution stack.

", - "refs": { + "UpdateConfigurationTemplateMessage":{ + "base":"

The result message containing the options for the specified solution stack.

", + "refs":{ } }, - "UpdateDate": { - "base": null, - "refs": { - "ApplicationDescription$DateUpdated": "

The date when the application was last modified.

", - "ApplicationVersionDescription$DateUpdated": "

The last modified date of the application version.

", - "ConfigurationSettingsDescription$DateUpdated": "

The date (in UTC time) when this configuration set was last modified.

", - "EnvironmentDescription$DateUpdated": "

The last modified date for this environment.

", - "PlatformDescription$DateUpdated": "

The date when the platform was last updated.

" + "UpdateDate":{ + "base":null, + "refs":{ + "ApplicationDescription$DateUpdated":"

The date when the application was last modified.

", + "ApplicationVersionDescription$DateUpdated":"

The last modified date of the application version.

", + "ConfigurationSettingsDescription$DateUpdated":"

The date (in UTC time) when this configuration set was last modified.

", + "EnvironmentDescription$DateUpdated":"

The last modified date for this environment.

", + "PlatformDescription$DateUpdated":"

The date when the platform was last updated.

" } }, - "UpdateEnvironmentMessage": { - "base": "

Request to update an environment.

", - "refs": { + "UpdateEnvironmentMessage":{ + "base":"

Request to update an environment.

", + "refs":{ } }, - "UpdateTagsForResourceMessage": { - "base": null, - "refs": { + "UpdateTagsForResourceMessage":{ + "base":null, + "refs":{ } }, - "UserDefinedOption": { - "base": null, - "refs": { - "ConfigurationOptionDescription$UserDefined": "

An indication of whether the user defined this configuration option:

  • true : This configuration option was defined by the user. It is a valid choice for specifying if this as an Option to Remove when updating configuration settings.

  • false : This configuration was not defined by the user.

Constraint: You can remove only UserDefined options from a configuration.

Valid Values: true | false

" + "UserDefinedOption":{ + "base":null, + "refs":{ + "ConfigurationOptionDescription$UserDefined":"

An indication of whether the user defined this configuration option:

  • true : This configuration option was defined by the user. It is a valid choice for specifying if this as an Option to Remove when updating configuration settings.

  • false : This configuration was not defined by the user.

Constraint: You can remove only UserDefined options from a configuration.

Valid Values: true | false

" } }, - "ValidateConfigurationSettingsMessage": { - "base": "

A list of validation messages for a specified configuration template.

", - "refs": { + "ValidateConfigurationSettingsMessage":{ + "base":"

A list of validation messages for a specified configuration template.

", + "refs":{ } }, - "ValidationMessage": { - "base": "

An error or warning for a desired configuration option value.

", - "refs": { - "ValidationMessagesList$member": null + "ValidationMessage":{ + "base":"

An error or warning for a desired configuration option value.

", + "refs":{ + "ValidationMessagesList$member":null } }, - "ValidationMessageString": { - "base": null, - "refs": { - "ValidationMessage$Message": "

A message describing the error or warning.

" + "ValidationMessageString":{ + "base":null, + "refs":{ + "ValidationMessage$Message":"

A message describing the error or warning.

" } }, - "ValidationMessagesList": { - "base": null, - "refs": { - "ConfigurationSettingsValidationMessages$Messages": "

A list of ValidationMessage.

" + "ValidationMessagesList":{ + "base":null, + "refs":{ + "ConfigurationSettingsValidationMessages$Messages":"

A list of ValidationMessage.

" } }, - "ValidationSeverity": { - "base": null, - "refs": { - "ValidationMessage$Severity": "

An indication of the severity of this message:

  • error: This message indicates that this is not a valid setting for an option.

  • warning: This message is providing information you should take into account.

" + "ValidationSeverity":{ + "base":null, + "refs":{ + "ValidationMessage$Severity":"

An indication of the severity of this message:

  • error: This message indicates that this is not a valid setting for an option.

  • warning: This message is providing information you should take into account.

" } }, - "VersionLabel": { - "base": null, - "refs": { - "ApplicationVersionDescription$VersionLabel": "

A unique identifier for the application version.

", - "CreateApplicationVersionMessage$VersionLabel": "

A label identifying this version.

Constraint: Must be unique per application. If an application version already exists with this label for the specified application, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", - "CreateEnvironmentMessage$VersionLabel": "

The name of the application version to deploy.

If the specified application has no associated application versions, AWS Elastic Beanstalk UpdateEnvironment returns an InvalidParameterValue error.

Default: If not specified, AWS Elastic Beanstalk attempts to launch the sample application in the container.

", - "DeleteApplicationVersionMessage$VersionLabel": "

The label of the version to delete.

", - "DescribeEnvironmentsMessage$VersionLabel": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that are associated with this application version.

", - "DescribeEventsMessage$VersionLabel": "

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those associated with this application version.

", - "EnvironmentDescription$VersionLabel": "

The application version deployed in this environment.

", - "EventDescription$VersionLabel": "

The release label for the application version associated with this event.

", - "UpdateApplicationVersionMessage$VersionLabel": "

The name of the version to update.

If no application version is found with this label, UpdateApplication returns an InvalidParameterValue error.

", - "UpdateEnvironmentMessage$VersionLabel": "

If this parameter is specified, AWS Elastic Beanstalk deploys the named application version to the environment. If no such application version is found, returns an InvalidParameterValue error.

", - "VersionLabels$member": null, - "VersionLabelsList$member": null + "VersionLabel":{ + "base":null, + "refs":{ + "ApplicationVersionDescription$VersionLabel":"

A unique identifier for the application version.

", + "CreateApplicationVersionMessage$VersionLabel":"

A label identifying this version.

Constraint: Must be unique per application. If an application version already exists with this label for the specified application, AWS Elastic Beanstalk returns an InvalidParameterValue error.

", + "CreateEnvironmentMessage$VersionLabel":"

The name of the application version to deploy.

If the specified application has no associated application versions, AWS Elastic Beanstalk UpdateEnvironment returns an InvalidParameterValue error.

Default: If not specified, AWS Elastic Beanstalk attempts to launch the sample application in the container.

", + "DeleteApplicationVersionMessage$VersionLabel":"

The label of the version to delete.

", + "DescribeEnvironmentsMessage$VersionLabel":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to include only those that are associated with this application version.

", + "DescribeEventsMessage$VersionLabel":"

If specified, AWS Elastic Beanstalk restricts the returned descriptions to those associated with this application version.

", + "EnvironmentDescription$VersionLabel":"

The application version deployed in this environment.

", + "EventDescription$VersionLabel":"

The release label for the application version associated with this event.

", + "UpdateApplicationVersionMessage$VersionLabel":"

The name of the version to update.

If no application version is found with this label, UpdateApplication returns an InvalidParameterValue error.

", + "UpdateEnvironmentMessage$VersionLabel":"

If this parameter is specified, AWS Elastic Beanstalk deploys the named application version to the environment. If no such application version is found, returns an InvalidParameterValue error.

", + "VersionLabels$member":null, + "VersionLabelsList$member":null } }, - "VersionLabels": { - "base": null, - "refs": { - "ComposeEnvironmentsMessage$VersionLabels": "

A list of version labels, specifying one or more application source bundles that belong to the target application. Each source bundle must include an environment manifest that specifies the name of the environment and the name of the solution stack to use, and optionally can specify environment links to create.

" + "VersionLabels":{ + "base":null, + "refs":{ + "ComposeEnvironmentsMessage$VersionLabels":"

A list of version labels, specifying one or more application source bundles that belong to the target application. Each source bundle must include an environment manifest that specifies the name of the environment and the name of the solution stack to use, and optionally can specify environment links to create.

" } }, - "VersionLabelsList": { - "base": null, - "refs": { - "ApplicationDescription$Versions": "

The names of the versions for this application.

", - "DescribeApplicationVersionsMessage$VersionLabels": "

Specify a version label to show a specific application version.

" + "VersionLabelsList":{ + "base":null, + "refs":{ + "ApplicationDescription$Versions":"

The names of the versions for this application.

", + "DescribeApplicationVersionsMessage$VersionLabels":"

Specify a version label to show a specific application version.

" } }, - "VirtualizationType": { - "base": null, - "refs": { - "CustomAmi$VirtualizationType": "

The type of virtualization used to create the custom AMI.

" + "VirtualizationType":{ + "base":null, + "refs":{ + "CustomAmi$VirtualizationType":"

The type of virtualization used to create the custom AMI.

" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/examples-1.json index 0fded6281..8a8ad3918 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticbeanstalk/2010-12-01/examples-1.json @@ -1,695 +1,695 @@ { - "version": "1.0", - "examples": { - "AbortEnvironmentUpdate": [ + "version":"1.0", + "examples":{ + "AbortEnvironmentUpdate":[ { - "input": { - "EnvironmentName": "my-env" + "input":{ + "EnvironmentName":"my-env" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following code aborts a running application version deployment for an environment named my-env:", - "id": "to-abort-a-deployment-1456267848227", - "title": "To abort a deployment" + "description":"The following code aborts a running application version deployment for an environment named my-env:", + "id":"to-abort-a-deployment-1456267848227", + "title":"To abort a deployment" } ], - "CheckDNSAvailability": [ + "CheckDNSAvailability":[ { - "input": { - "CNAMEPrefix": "my-cname" + "input":{ + "CNAMEPrefix":"my-cname" }, - "output": { - "Available": true, - "FullyQualifiedCNAME": "my-cname.us-west-2.elasticbeanstalk.com" + "output":{ + "Available":true, + "FullyQualifiedCNAME":"my-cname.us-west-2.elasticbeanstalk.com" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation checks the availability of the subdomain my-cname:", - "id": "to-check-the-availability-of-a-cname-1456268589537", - "title": "To check the availability of a CNAME" + "description":"The following operation checks the availability of the subdomain my-cname:", + "id":"to-check-the-availability-of-a-cname-1456268589537", + "title":"To check the availability of a CNAME" } ], - "CreateApplication": [ + "CreateApplication":[ { - "input": { - "ApplicationName": "my-app", - "Description": "my application" + "input":{ + "ApplicationName":"my-app", + "Description":"my application" }, - "output": { - "Application": { - "ApplicationName": "my-app", - "ConfigurationTemplates": [ + "output":{ + "Application":{ + "ApplicationName":"my-app", + "ConfigurationTemplates":[ ], - "DateCreated": "2015-02-12T18:32:21.181Z", - "DateUpdated": "2015-02-12T18:32:21.181Z", - "Description": "my application" + "DateCreated":"2015-02-12T18:32:21.181Z", + "DateUpdated":"2015-02-12T18:32:21.181Z", + "Description":"my application" } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation creates a new application named my-app:", - "id": "to-create-a-new-application-1456268895683", - "title": "To create a new application" + "description":"The following operation creates a new application named my-app:", + "id":"to-create-a-new-application-1456268895683", + "title":"To create a new application" } ], - "CreateApplicationVersion": [ + "CreateApplicationVersion":[ { - "input": { - "ApplicationName": "my-app", - "AutoCreateApplication": true, - "Description": "my-app-v1", - "Process": true, - "SourceBundle": { - "S3Bucket": "my-bucket", - "S3Key": "sample.war" + "input":{ + "ApplicationName":"my-app", + "AutoCreateApplication":true, + "Description":"my-app-v1", + "Process":true, + "SourceBundle":{ + "S3Bucket":"my-bucket", + "S3Key":"sample.war" }, - "VersionLabel": "v1" + "VersionLabel":"v1" }, - "output": { - "ApplicationVersion": { - "ApplicationName": "my-app", - "DateCreated": "2015-02-03T23:01:25.412Z", - "DateUpdated": "2015-02-03T23:01:25.412Z", - "Description": "my-app-v1", - "SourceBundle": { - "S3Bucket": "my-bucket", - "S3Key": "sample.war" + "output":{ + "ApplicationVersion":{ + "ApplicationName":"my-app", + "DateCreated":"2015-02-03T23:01:25.412Z", + "DateUpdated":"2015-02-03T23:01:25.412Z", + "Description":"my-app-v1", + "SourceBundle":{ + "S3Bucket":"my-bucket", + "S3Key":"sample.war" }, - "VersionLabel": "v1" + "VersionLabel":"v1" } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation creates a new version (v1) of an application named my-app:", - "id": "to-create-a-new-application-1456268895683", - "title": "To create a new application" + "description":"The following operation creates a new version (v1) of an application named my-app:", + "id":"to-create-a-new-application-1456268895683", + "title":"To create a new application" } ], - "CreateConfigurationTemplate": [ + "CreateConfigurationTemplate":[ { - "input": { - "ApplicationName": "my-app", - "EnvironmentId": "e-rpqsewtp2j", - "TemplateName": "my-app-v1" + "input":{ + "ApplicationName":"my-app", + "EnvironmentId":"e-rpqsewtp2j", + "TemplateName":"my-app-v1" }, - "output": { - "ApplicationName": "my-app", - "DateCreated": "2015-08-12T18:40:39Z", - "DateUpdated": "2015-08-12T18:40:39Z", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "TemplateName": "my-app-v1" + "output":{ + "ApplicationName":"my-app", + "DateCreated":"2015-08-12T18:40:39Z", + "DateUpdated":"2015-08-12T18:40:39Z", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "TemplateName":"my-app-v1" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation creates a configuration template named my-app-v1 from the settings applied to an environment with the id e-rpqsewtp2j:", - "id": "to-create-a-configuration-template-1456269283586", - "title": "To create a configuration template" + "description":"The following operation creates a configuration template named my-app-v1 from the settings applied to an environment with the id e-rpqsewtp2j:", + "id":"to-create-a-configuration-template-1456269283586", + "title":"To create a configuration template" } ], - "CreateEnvironment": [ + "CreateEnvironment":[ { - "input": { - "ApplicationName": "my-app", - "CNAMEPrefix": "my-app", - "EnvironmentName": "my-env", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "VersionLabel": "v1" + "input":{ + "ApplicationName":"my-app", + "CNAMEPrefix":"my-app", + "EnvironmentName":"my-env", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "VersionLabel":"v1" }, - "output": { - "ApplicationName": "my-app", - "CNAME": "my-app.elasticbeanstalk.com", - "DateCreated": "2015-02-03T23:04:54.479Z", - "DateUpdated": "2015-02-03T23:04:54.479Z", - "EnvironmentId": "e-izqpassy4h", - "EnvironmentName": "my-env", - "Health": "Grey", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "Status": "Launching", - "Tier": { - "Name": "WebServer", - "Type": "Standard", - "Version": " " + "output":{ + "ApplicationName":"my-app", + "CNAME":"my-app.elasticbeanstalk.com", + "DateCreated":"2015-02-03T23:04:54.479Z", + "DateUpdated":"2015-02-03T23:04:54.479Z", + "EnvironmentId":"e-izqpassy4h", + "EnvironmentName":"my-env", + "Health":"Grey", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "Status":"Launching", + "Tier":{ + "Name":"WebServer", + "Type":"Standard", + "Version":" " }, - "VersionLabel": "v1" + "VersionLabel":"v1" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation creates a new environment for version v1 of a java application named my-app:", - "id": "to-create-a-new-environment-for-an-application-1456269380396", - "title": "To create a new environment for an application" + "description":"The following operation creates a new environment for version v1 of a java application named my-app:", + "id":"to-create-a-new-environment-for-an-application-1456269380396", + "title":"To create a new environment for an application" } ], - "CreateStorageLocation": [ + "CreateStorageLocation":[ { - "output": { - "S3Bucket": "elasticbeanstalk-us-west-2-0123456789012" + "output":{ + "S3Bucket":"elasticbeanstalk-us-west-2-0123456789012" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation creates a new environment for version v1 of a java application named my-app:", - "id": "to-create-a-new-environment-for-an-application-1456269380396", - "title": "To create a new environment for an application" + "description":"The following operation creates a new environment for version v1 of a java application named my-app:", + "id":"to-create-a-new-environment-for-an-application-1456269380396", + "title":"To create a new environment for an application" } ], - "DeleteApplication": [ + "DeleteApplication":[ { - "input": { - "ApplicationName": "my-app" + "input":{ + "ApplicationName":"my-app" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation deletes an application named my-app:", - "id": "to-delete-an-application-1456269699366", - "title": "To delete an application" + "description":"The following operation deletes an application named my-app:", + "id":"to-delete-an-application-1456269699366", + "title":"To delete an application" } ], - "DeleteApplicationVersion": [ + "DeleteApplicationVersion":[ { - "input": { - "ApplicationName": "my-app", - "DeleteSourceBundle": true, - "VersionLabel": "22a0-stage-150819_182129" + "input":{ + "ApplicationName":"my-app", + "DeleteSourceBundle":true, + "VersionLabel":"22a0-stage-150819_182129" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation deletes an application version named 22a0-stage-150819_182129 for an application named my-app:", - "id": "to-delete-an-application-version-1456269792956", - "title": "To delete an application version" + "description":"The following operation deletes an application version named 22a0-stage-150819_182129 for an application named my-app:", + "id":"to-delete-an-application-version-1456269792956", + "title":"To delete an application version" } ], - "DeleteConfigurationTemplate": [ + "DeleteConfigurationTemplate":[ { - "input": { - "ApplicationName": "my-app", - "TemplateName": "my-template" + "input":{ + "ApplicationName":"my-app", + "TemplateName":"my-template" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation deletes a configuration template named my-template for an application named my-app:", - "id": "to-delete-a-configuration-template-1456269836701", - "title": "To delete a configuration template" + "description":"The following operation deletes a configuration template named my-template for an application named my-app:", + "id":"to-delete-a-configuration-template-1456269836701", + "title":"To delete a configuration template" } ], - "DeleteEnvironmentConfiguration": [ + "DeleteEnvironmentConfiguration":[ { - "input": { - "ApplicationName": "my-app", - "EnvironmentName": "my-env" + "input":{ + "ApplicationName":"my-app", + "EnvironmentName":"my-env" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation deletes a draft configuration for an environment named my-env:", - "id": "to-delete-a-draft-configuration-1456269886654", - "title": "To delete a draft configuration" + "description":"The following operation deletes a draft configuration for an environment named my-env:", + "id":"to-delete-a-draft-configuration-1456269886654", + "title":"To delete a draft configuration" } ], - "DescribeApplicationVersions": [ + "DescribeApplicationVersions":[ { - "input": { - "ApplicationName": "my-app", - "VersionLabels": [ + "input":{ + "ApplicationName":"my-app", + "VersionLabels":[ "v2" ] }, - "output": { - "ApplicationVersions": [ + "output":{ + "ApplicationVersions":[ { - "ApplicationName": "my-app", - "DateCreated": "2015-07-23T01:32:26.079Z", - "DateUpdated": "2015-07-23T01:32:26.079Z", - "Description": "update cover page", - "SourceBundle": { - "S3Bucket": "elasticbeanstalk-us-west-2-015321684451", - "S3Key": "my-app/5026-stage-150723_224258.war" + "ApplicationName":"my-app", + "DateCreated":"2015-07-23T01:32:26.079Z", + "DateUpdated":"2015-07-23T01:32:26.079Z", + "Description":"update cover page", + "SourceBundle":{ + "S3Bucket":"elasticbeanstalk-us-west-2-015321684451", + "S3Key":"my-app/5026-stage-150723_224258.war" }, - "VersionLabel": "v2" + "VersionLabel":"v2" }, { - "ApplicationName": "my-app", - "DateCreated": "2015-07-23T22:26:10.816Z", - "DateUpdated": "2015-07-23T22:26:10.816Z", - "Description": "initial version", - "SourceBundle": { - "S3Bucket": "elasticbeanstalk-us-west-2-015321684451", - "S3Key": "my-app/5026-stage-150723_222618.war" + "ApplicationName":"my-app", + "DateCreated":"2015-07-23T22:26:10.816Z", + "DateUpdated":"2015-07-23T22:26:10.816Z", + "Description":"initial version", + "SourceBundle":{ + "S3Bucket":"elasticbeanstalk-us-west-2-015321684451", + "S3Key":"my-app/5026-stage-150723_222618.war" }, - "VersionLabel": "v1" + "VersionLabel":"v1" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves information about an application version labeled v2:", - "id": "to-view-information-about-an-application-version-1456269947428", - "title": "To view information about an application version" + "description":"The following operation retrieves information about an application version labeled v2:", + "id":"to-view-information-about-an-application-version-1456269947428", + "title":"To view information about an application version" } ], - "DescribeApplications": [ + "DescribeApplications":[ { - "input": { + "input":{ }, - "output": { - "Applications": [ + "output":{ + "Applications":[ { - "ApplicationName": "ruby", - "ConfigurationTemplates": [ + "ApplicationName":"ruby", + "ConfigurationTemplates":[ ], - "DateCreated": "2015-08-13T21:05:44.376Z", - "DateUpdated": "2015-08-13T21:05:44.376Z", - "Versions": [ + "DateCreated":"2015-08-13T21:05:44.376Z", + "DateUpdated":"2015-08-13T21:05:44.376Z", + "Versions":[ "Sample Application" ] }, { - "ApplicationName": "pythonsample", - "ConfigurationTemplates": [ + "ApplicationName":"pythonsample", + "ConfigurationTemplates":[ ], - "DateCreated": "2015-08-13T19:05:43.637Z", - "DateUpdated": "2015-08-13T19:05:43.637Z", - "Description": "Application created from the EB CLI using \"eb init\"", - "Versions": [ + "DateCreated":"2015-08-13T19:05:43.637Z", + "DateUpdated":"2015-08-13T19:05:43.637Z", + "Description":"Application created from the EB CLI using \"eb init\"", + "Versions":[ "Sample Application" ] }, { - "ApplicationName": "nodejs-example", - "ConfigurationTemplates": [ + "ApplicationName":"nodejs-example", + "ConfigurationTemplates":[ ], - "DateCreated": "2015-08-06T17:50:02.486Z", - "DateUpdated": "2015-08-06T17:50:02.486Z", - "Versions": [ + "DateCreated":"2015-08-06T17:50:02.486Z", + "DateUpdated":"2015-08-06T17:50:02.486Z", + "Versions":[ "add elasticache", "First Release" ] } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves information about applications in the current region:", - "id": "to-view-a-list-of-applications-1456270027373", - "title": "To view a list of applications" + "description":"The following operation retrieves information about applications in the current region:", + "id":"to-view-a-list-of-applications-1456270027373", + "title":"To view a list of applications" } ], - "DescribeConfigurationOptions": [ + "DescribeConfigurationOptions":[ { - "input": { - "ApplicationName": "my-app", - "EnvironmentName": "my-env" + "input":{ + "ApplicationName":"my-app", + "EnvironmentName":"my-env" }, - "output": { - "Options": [ + "output":{ + "Options":[ { - "ChangeSeverity": "NoInterruption", - "DefaultValue": "30", - "MaxValue": 300, - "MinValue": 5, - "Name": "Interval", - "Namespace": "aws:elb:healthcheck", - "UserDefined": false, - "ValueType": "Scalar" + "ChangeSeverity":"NoInterruption", + "DefaultValue":"30", + "MaxValue":300, + "MinValue":5, + "Name":"Interval", + "Namespace":"aws:elb:healthcheck", + "UserDefined":false, + "ValueType":"Scalar" }, { - "ChangeSeverity": "NoInterruption", - "DefaultValue": "2000000", - "MinValue": 0, - "Name": "LowerThreshold", - "Namespace": "aws:autoscaling:trigger", - "UserDefined": false, - "ValueType": "Scalar" + "ChangeSeverity":"NoInterruption", + "DefaultValue":"2000000", + "MinValue":0, + "Name":"LowerThreshold", + "Namespace":"aws:autoscaling:trigger", + "UserDefined":false, + "ValueType":"Scalar" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves descriptions of all available configuration options for an environment named my-env:", - "id": "to-view-configuration-options-for-an-environment-1456276763917", - "title": "To view configuration options for an environment" + "description":"The following operation retrieves descriptions of all available configuration options for an environment named my-env:", + "id":"to-view-configuration-options-for-an-environment-1456276763917", + "title":"To view configuration options for an environment" } ], - "DescribeConfigurationSettings": [ + "DescribeConfigurationSettings":[ { - "input": { - "ApplicationName": "my-app", - "EnvironmentName": "my-env" + "input":{ + "ApplicationName":"my-app", + "EnvironmentName":"my-env" }, - "output": { - "ConfigurationSettings": [ + "output":{ + "ConfigurationSettings":[ { - "ApplicationName": "my-app", - "DateCreated": "2015-08-13T19:16:25Z", - "DateUpdated": "2015-08-13T23:30:07Z", - "DeploymentStatus": "deployed", - "Description": "Environment created from the EB CLI using \"eb create\"", - "EnvironmentName": "my-env", - "OptionSettings": [ + "ApplicationName":"my-app", + "DateCreated":"2015-08-13T19:16:25Z", + "DateUpdated":"2015-08-13T23:30:07Z", + "DeploymentStatus":"deployed", + "Description":"Environment created from the EB CLI using \"eb create\"", + "EnvironmentName":"my-env", + "OptionSettings":[ { - "Namespace": "aws:autoscaling:asg", - "OptionName": "Availability Zones", - "ResourceName": "AWSEBAutoScalingGroup", - "Value": "Any" + "Namespace":"aws:autoscaling:asg", + "OptionName":"Availability Zones", + "ResourceName":"AWSEBAutoScalingGroup", + "Value":"Any" }, { - "Namespace": "aws:autoscaling:asg", - "OptionName": "Cooldown", - "ResourceName": "AWSEBAutoScalingGroup", - "Value": "360" + "Namespace":"aws:autoscaling:asg", + "OptionName":"Cooldown", + "ResourceName":"AWSEBAutoScalingGroup", + "Value":"360" }, { - "Namespace": "aws:elb:policies", - "OptionName": "ConnectionDrainingTimeout", - "ResourceName": "AWSEBLoadBalancer", - "Value": "20" + "Namespace":"aws:elb:policies", + "OptionName":"ConnectionDrainingTimeout", + "ResourceName":"AWSEBLoadBalancer", + "Value":"20" }, { - "Namespace": "aws:elb:policies", - "OptionName": "ConnectionSettingIdleTimeout", - "ResourceName": "AWSEBLoadBalancer", - "Value": "60" + "Namespace":"aws:elb:policies", + "OptionName":"ConnectionSettingIdleTimeout", + "ResourceName":"AWSEBLoadBalancer", + "Value":"60" } ], - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8" + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { - "abbreviated": "Output is abbreviated" + "output":{ + "abbreviated":"Output is abbreviated" } }, - "description": "The following operation retrieves configuration settings for an environment named my-env:", - "id": "to-view-configurations-settings-for-an-environment-1456276924537", - "title": "To view configurations settings for an environment" + "description":"The following operation retrieves configuration settings for an environment named my-env:", + "id":"to-view-configurations-settings-for-an-environment-1456276924537", + "title":"To view configurations settings for an environment" } ], - "DescribeEnvironmentHealth": [ + "DescribeEnvironmentHealth":[ { - "input": { - "AttributeNames": [ + "input":{ + "AttributeNames":[ "All" ], - "EnvironmentName": "my-env" + "EnvironmentName":"my-env" }, - "output": { - "ApplicationMetrics": { - "Duration": 10, - "Latency": { - "P10": 0.001, - "P50": 0.001, - "P75": 0.002, - "P85": 0.003, - "P90": 0.003, - "P95": 0.004, - "P99": 0.004, - "P999": 0.004 + "output":{ + "ApplicationMetrics":{ + "Duration":10, + "Latency":{ + "P10":0.001, + "P50":0.001, + "P75":0.002, + "P85":0.003, + "P90":0.003, + "P95":0.004, + "P99":0.004, + "P999":0.004 }, - "RequestCount": 45, - "StatusCodes": { - "Status2xx": 45, - "Status3xx": 0, - "Status4xx": 0, - "Status5xx": 0 + "RequestCount":45, + "StatusCodes":{ + "Status2xx":45, + "Status3xx":0, + "Status4xx":0, + "Status5xx":0 } }, - "Causes": [ + "Causes":[ ], - "Color": "Green", - "EnvironmentName": "my-env", - "HealthStatus": "Ok", - "InstancesHealth": { - "Degraded": 0, - "Info": 0, - "NoData": 0, - "Ok": 1, - "Pending": 0, - "Severe": 0, - "Unknown": 0, - "Warning": 0 + "Color":"Green", + "EnvironmentName":"my-env", + "HealthStatus":"Ok", + "InstancesHealth":{ + "Degraded":0, + "Info":0, + "NoData":0, + "Ok":1, + "Pending":0, + "Severe":0, + "Unknown":0, + "Warning":0 }, - "RefreshedAt": "2015-08-20T21:09:18Z" + "RefreshedAt":"2015-08-20T21:09:18Z" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves overall health information for an environment named my-env:", - "id": "to-view-environment-health-1456277109510", - "title": "To view environment health" + "description":"The following operation retrieves overall health information for an environment named my-env:", + "id":"to-view-environment-health-1456277109510", + "title":"To view environment health" } ], - "DescribeEnvironmentResources": [ + "DescribeEnvironmentResources":[ { - "input": { - "EnvironmentName": "my-env" + "input":{ + "EnvironmentName":"my-env" }, - "output": { - "EnvironmentResources": { - "AutoScalingGroups": [ + "output":{ + "EnvironmentResources":{ + "AutoScalingGroups":[ { - "Name": "awseb-e-qu3fyyjyjs-stack-AWSEBAutoScalingGroup-QSB2ZO88SXZT" + "Name":"awseb-e-qu3fyyjyjs-stack-AWSEBAutoScalingGroup-QSB2ZO88SXZT" } ], - "EnvironmentName": "my-env", - "Instances": [ + "EnvironmentName":"my-env", + "Instances":[ { - "Id": "i-0c91c786" + "Id":"i-0c91c786" } ], - "LaunchConfigurations": [ + "LaunchConfigurations":[ { - "Name": "awseb-e-qu3fyyjyjs-stack-AWSEBAutoScalingLaunchConfiguration-1UUVQIBC96TQ2" + "Name":"awseb-e-qu3fyyjyjs-stack-AWSEBAutoScalingLaunchConfiguration-1UUVQIBC96TQ2" } ], - "LoadBalancers": [ + "LoadBalancers":[ { - "Name": "awseb-e-q-AWSEBLoa-1EEPZ0K98BIF0" + "Name":"awseb-e-q-AWSEBLoa-1EEPZ0K98BIF0" } ], - "Queues": [ + "Queues":[ ], - "Triggers": [ + "Triggers":[ ] } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves information about resources in an environment named my-env:", - "id": "to-view-information-about-the-aws-resources-in-your-environment-1456277206232", - "title": "To view information about the AWS resources in your environment" + "description":"The following operation retrieves information about resources in an environment named my-env:", + "id":"to-view-information-about-the-aws-resources-in-your-environment-1456277206232", + "title":"To view information about the AWS resources in your environment" } ], - "DescribeEnvironments": [ + "DescribeEnvironments":[ { - "input": { - "EnvironmentNames": [ + "input":{ + "EnvironmentNames":[ "my-env" ] }, - "output": { - "Environments": [ + "output":{ + "Environments":[ { - "AbortableOperationInProgress": false, - "ApplicationName": "my-app", - "CNAME": "my-env.elasticbeanstalk.com", - "DateCreated": "2015-08-07T20:48:49.599Z", - "DateUpdated": "2015-08-12T18:16:55.019Z", - "EndpointURL": "awseb-e-w-AWSEBLoa-1483140XB0Q4L-109QXY8121.us-west-2.elb.amazonaws.com", - "EnvironmentId": "e-rpqsewtp2j", - "EnvironmentName": "my-env", - "Health": "Green", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "Status": "Ready", - "Tier": { - "Name": "WebServer", - "Type": "Standard", - "Version": " " + "AbortableOperationInProgress":false, + "ApplicationName":"my-app", + "CNAME":"my-env.elasticbeanstalk.com", + "DateCreated":"2015-08-07T20:48:49.599Z", + "DateUpdated":"2015-08-12T18:16:55.019Z", + "EndpointURL":"awseb-e-w-AWSEBLoa-1483140XB0Q4L-109QXY8121.us-west-2.elb.amazonaws.com", + "EnvironmentId":"e-rpqsewtp2j", + "EnvironmentName":"my-env", + "Health":"Green", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "Status":"Ready", + "Tier":{ + "Name":"WebServer", + "Type":"Standard", + "Version":" " }, - "VersionLabel": "7f58-stage-150812_025409" + "VersionLabel":"7f58-stage-150812_025409" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves information about an environment named my-env:", - "id": "to-view-information-about-an-environment-1456277288662", - "title": "To view information about an environment" + "description":"The following operation retrieves information about an environment named my-env:", + "id":"to-view-information-about-an-environment-1456277288662", + "title":"To view information about an environment" } ], - "DescribeEvents": [ + "DescribeEvents":[ { - "input": { - "EnvironmentName": "my-env" + "input":{ + "EnvironmentName":"my-env" }, - "output": { - "Events": [ + "output":{ + "Events":[ { - "ApplicationName": "my-app", - "EnvironmentName": "my-env", - "EventDate": "2015-08-20T07:06:53.535Z", - "Message": "Environment health has transitioned from Info to Ok.", - "Severity": "INFO" + "ApplicationName":"my-app", + "EnvironmentName":"my-env", + "EventDate":"2015-08-20T07:06:53.535Z", + "Message":"Environment health has transitioned from Info to Ok.", + "Severity":"INFO" }, { - "ApplicationName": "my-app", - "EnvironmentName": "my-env", - "EventDate": "2015-08-20T07:06:02.049Z", - "Message": "Environment update completed successfully.", - "RequestId": "b7f3960b-4709-11e5-ba1e-07e16200da41", - "Severity": "INFO" + "ApplicationName":"my-app", + "EnvironmentName":"my-env", + "EventDate":"2015-08-20T07:06:02.049Z", + "Message":"Environment update completed successfully.", + "RequestId":"b7f3960b-4709-11e5-ba1e-07e16200da41", + "Severity":"INFO" }, { - "ApplicationName": "my-app", - "EnvironmentName": "my-env", - "EventDate": "2015-08-13T19:16:27.561Z", - "Message": "Using elasticbeanstalk-us-west-2-012445113685 as Amazon S3 storage bucket for environment data.", - "RequestId": "ca8dfbf6-41ef-11e5-988b-651aa638f46b", - "Severity": "INFO" + "ApplicationName":"my-app", + "EnvironmentName":"my-env", + "EventDate":"2015-08-13T19:16:27.561Z", + "Message":"Using elasticbeanstalk-us-west-2-012445113685 as Amazon S3 storage bucket for environment data.", + "RequestId":"ca8dfbf6-41ef-11e5-988b-651aa638f46b", + "Severity":"INFO" }, { - "ApplicationName": "my-app", - "EnvironmentName": "my-env", - "EventDate": "2015-08-13T19:16:26.581Z", - "Message": "createEnvironment is starting.", - "RequestId": "cdfba8f6-41ef-11e5-988b-65638f41aa6b", - "Severity": "INFO" + "ApplicationName":"my-app", + "EnvironmentName":"my-env", + "EventDate":"2015-08-13T19:16:26.581Z", + "Message":"createEnvironment is starting.", + "RequestId":"cdfba8f6-41ef-11e5-988b-65638f41aa6b", + "Severity":"INFO" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves events for an environment named my-env:", - "id": "to-view-events-for-an-environment-1456277367589", - "title": "To view events for an environment" + "description":"The following operation retrieves events for an environment named my-env:", + "id":"to-view-events-for-an-environment-1456277367589", + "title":"To view events for an environment" } ], - "DescribeInstancesHealth": [ + "DescribeInstancesHealth":[ { - "input": { - "AttributeNames": [ + "input":{ + "AttributeNames":[ "All" ], - "EnvironmentName": "my-env" + "EnvironmentName":"my-env" }, - "output": { - "InstanceHealthList": [ + "output":{ + "InstanceHealthList":[ { - "ApplicationMetrics": { - "Duration": 10, - "Latency": { - "P10": 0, - "P50": 0.001, - "P75": 0.002, - "P85": 0.003, - "P90": 0.004, - "P95": 0.005, - "P99": 0.006, - "P999": 0.006 + "ApplicationMetrics":{ + "Duration":10, + "Latency":{ + "P10":0, + "P50":0.001, + "P75":0.002, + "P85":0.003, + "P90":0.004, + "P95":0.005, + "P99":0.006, + "P999":0.006 }, - "RequestCount": 48, - "StatusCodes": { - "Status2xx": 47, - "Status3xx": 0, - "Status4xx": 1, - "Status5xx": 0 + "RequestCount":48, + "StatusCodes":{ + "Status2xx":47, + "Status3xx":0, + "Status4xx":1, + "Status5xx":0 } }, - "Causes": [ + "Causes":[ ], - "Color": "Green", - "HealthStatus": "Ok", - "InstanceId": "i-08691cc7", - "LaunchedAt": "2015-08-13T19:17:09Z", - "System": { - "CPUUtilization": { - "IOWait": 0.2, - "IRQ": 0, - "Idle": 97.8, - "Nice": 0.1, - "SoftIRQ": 0.1, - "System": 0.3, - "User": 1.5 + "Color":"Green", + "HealthStatus":"Ok", + "InstanceId":"i-08691cc7", + "LaunchedAt":"2015-08-13T19:17:09Z", + "System":{ + "CPUUtilization":{ + "IOWait":0.2, + "IRQ":0, + "Idle":97.8, + "Nice":0.1, + "SoftIRQ":0.1, + "System":0.3, + "User":1.5 }, - "LoadAverage": [ + "LoadAverage":[ 0, 0.02, 0.05 @@ -697,31 +697,31 @@ } } ], - "RefreshedAt": "2015-08-20T21:09:08Z" + "RefreshedAt":"2015-08-20T21:09:08Z" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves health information for instances in an environment named my-env:", - "id": "to-view-environment-health-1456277424757", - "title": "To view environment health" + "description":"The following operation retrieves health information for instances in an environment named my-env:", + "id":"to-view-environment-health-1456277424757", + "title":"To view environment health" } ], - "ListAvailableSolutionStacks": [ + "ListAvailableSolutionStacks":[ { - "output": { - "SolutionStackDetails": [ + "output":{ + "SolutionStackDetails":[ { - "PermittedFileTypes": [ + "PermittedFileTypes":[ "zip" ], - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Node.js" + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Node.js" } ], - "SolutionStacks": [ + "SolutionStacks":[ "64bit Amazon Linux 2015.03 v2.0.0 running Node.js", "64bit Amazon Linux 2015.03 v2.0.0 running PHP 5.6", "64bit Amazon Linux 2015.03 v2.0.0 running PHP 5.5", @@ -752,160 +752,160 @@ "64bit Debian jessie v2.0.0 running Python 3.4 (Preconfigured - Docker)" ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation lists solution stacks for all currently available platform configurations and any that you have used in the past:", - "id": "to-view-solution-stacks-1456277504811", - "title": "To view solution stacks" + "description":"The following operation lists solution stacks for all currently available platform configurations and any that you have used in the past:", + "id":"to-view-solution-stacks-1456277504811", + "title":"To view solution stacks" } ], - "RebuildEnvironment": [ + "RebuildEnvironment":[ { - "input": { - "EnvironmentName": "my-env" + "input":{ + "EnvironmentName":"my-env" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation terminates and recreates the resources in an environment named my-env:", - "id": "to-rebuild-an-environment-1456277600918", - "title": "To rebuild an environment" + "description":"The following operation terminates and recreates the resources in an environment named my-env:", + "id":"to-rebuild-an-environment-1456277600918", + "title":"To rebuild an environment" } ], - "RequestEnvironmentInfo": [ + "RequestEnvironmentInfo":[ { - "input": { - "EnvironmentName": "my-env", - "InfoType": "tail" + "input":{ + "EnvironmentName":"my-env", + "InfoType":"tail" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation requests logs from an environment named my-env:", - "id": "to-request-tailed-logs-1456277657045", - "title": "To request tailed logs" + "description":"The following operation requests logs from an environment named my-env:", + "id":"to-request-tailed-logs-1456277657045", + "title":"To request tailed logs" } ], - "RestartAppServer": [ + "RestartAppServer":[ { - "input": { - "EnvironmentName": "my-env" + "input":{ + "EnvironmentName":"my-env" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation restarts application servers on all instances in an environment named my-env:", - "id": "to-restart-application-servers-1456277739302", - "title": "To restart application servers" + "description":"The following operation restarts application servers on all instances in an environment named my-env:", + "id":"to-restart-application-servers-1456277739302", + "title":"To restart application servers" } ], - "RetrieveEnvironmentInfo": [ + "RetrieveEnvironmentInfo":[ { - "input": { - "EnvironmentName": "my-env", - "InfoType": "tail" + "input":{ + "EnvironmentName":"my-env", + "InfoType":"tail" }, - "output": { - "EnvironmentInfo": [ + "output":{ + "EnvironmentInfo":[ { - "Ec2InstanceId": "i-09c1c867", - "InfoType": "tail", - "Message": "https://elasticbeanstalk-us-west-2-0123456789012.s3.amazonaws.com/resources/environments/logs/tail/e-fyqyju3yjs/i-09c1c867/TailLogs-1440109397703.out?AWSAccessKeyId=AKGPT4J56IAJ2EUBL5CQ&Expires=1440195891&Signature=n%2BEalOV6A2HIOx4Rcfb7LT16bBM%3D", - "SampleTimestamp": "2015-08-20T22:23:17.703Z" + "Ec2InstanceId":"i-09c1c867", + "InfoType":"tail", + "Message":"https://elasticbeanstalk-us-west-2-0123456789012.s3.amazonaws.com/resources/environments/logs/tail/e-fyqyju3yjs/i-09c1c867/TailLogs-1440109397703.out?AWSAccessKeyId=AKGPT4J56IAJ2EUBL5CQ&Expires=1440195891&Signature=n%2BEalOV6A2HIOx4Rcfb7LT16bBM%3D", + "SampleTimestamp":"2015-08-20T22:23:17.703Z" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation retrieves a link to logs from an environment named my-env:", - "id": "to-retrieve-tailed-logs-1456277792734", - "title": "To retrieve tailed logs" + "description":"The following operation retrieves a link to logs from an environment named my-env:", + "id":"to-retrieve-tailed-logs-1456277792734", + "title":"To retrieve tailed logs" } ], - "SwapEnvironmentCNAMEs": [ + "SwapEnvironmentCNAMEs":[ { - "input": { - "DestinationEnvironmentName": "my-env-green", - "SourceEnvironmentName": "my-env-blue" + "input":{ + "DestinationEnvironmentName":"my-env-green", + "SourceEnvironmentName":"my-env-blue" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation swaps the assigned subdomains of two environments:", - "id": "to-swap-environment-cnames-1456277839438", - "title": "To swap environment CNAMES" + "description":"The following operation swaps the assigned subdomains of two environments:", + "id":"to-swap-environment-cnames-1456277839438", + "title":"To swap environment CNAMES" } ], - "TerminateEnvironment": [ + "TerminateEnvironment":[ { - "input": { - "EnvironmentName": "my-env" + "input":{ + "EnvironmentName":"my-env" }, - "output": { - "AbortableOperationInProgress": false, - "ApplicationName": "my-app", - "CNAME": "my-env.elasticbeanstalk.com", - "DateCreated": "2015-08-12T18:52:53.622Z", - "DateUpdated": "2015-08-12T19:05:54.744Z", - "EndpointURL": "awseb-e-f-AWSEBLoa-1I9XUMP4-8492WNUP202574.us-west-2.elb.amazonaws.com", - "EnvironmentId": "e-fh2eravpns", - "EnvironmentName": "my-env", - "Health": "Grey", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "Status": "Terminating", - "Tier": { - "Name": "WebServer", - "Type": "Standard", - "Version": " " + "output":{ + "AbortableOperationInProgress":false, + "ApplicationName":"my-app", + "CNAME":"my-env.elasticbeanstalk.com", + "DateCreated":"2015-08-12T18:52:53.622Z", + "DateUpdated":"2015-08-12T19:05:54.744Z", + "EndpointURL":"awseb-e-f-AWSEBLoa-1I9XUMP4-8492WNUP202574.us-west-2.elb.amazonaws.com", + "EnvironmentId":"e-fh2eravpns", + "EnvironmentName":"my-env", + "Health":"Grey", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "Status":"Terminating", + "Tier":{ + "Name":"WebServer", + "Type":"Standard", + "Version":" " } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation terminates an Elastic Beanstalk environment named my-env:", - "id": "to-terminate-an-environment-1456277888556", - "title": "To terminate an environment" + "description":"The following operation terminates an Elastic Beanstalk environment named my-env:", + "id":"to-terminate-an-environment-1456277888556", + "title":"To terminate an environment" } ], - "UpdateApplication": [ + "UpdateApplication":[ { - "input": { - "ApplicationName": "my-app", - "Description": "my Elastic Beanstalk application" + "input":{ + "ApplicationName":"my-app", + "Description":"my Elastic Beanstalk application" }, - "output": { - "Application": { - "ApplicationName": "my-app", - "ConfigurationTemplates": [ + "output":{ + "Application":{ + "ApplicationName":"my-app", + "ConfigurationTemplates":[ ], - "DateCreated": "2015-08-13T19:15:50.449Z", - "DateUpdated": "2015-08-20T22:34:56.195Z", - "Description": "my Elastic Beanstalk application", - "Versions": [ + "DateCreated":"2015-08-13T19:15:50.449Z", + "DateUpdated":"2015-08-20T22:34:56.195Z", + "Description":"my Elastic Beanstalk application", + "Versions":[ "2fba-stage-150819_234450", "bf07-stage-150820_214945", "93f8", @@ -914,195 +914,195 @@ ] } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation updates the description of an application named my-app:", - "id": "to-change-an-applications-description-1456277957075", - "title": "To change an application's description" + "description":"The following operation updates the description of an application named my-app:", + "id":"to-change-an-applications-description-1456277957075", + "title":"To change an application's description" } ], - "UpdateApplicationVersion": [ + "UpdateApplicationVersion":[ { - "input": { - "ApplicationName": "my-app", - "Description": "new description", - "VersionLabel": "22a0-stage-150819_185942" + "input":{ + "ApplicationName":"my-app", + "Description":"new description", + "VersionLabel":"22a0-stage-150819_185942" }, - "output": { - "ApplicationVersion": { - "ApplicationName": "my-app", - "DateCreated": "2015-08-19T18:59:17.646Z", - "DateUpdated": "2015-08-20T22:53:28.871Z", - "Description": "new description", - "SourceBundle": { - "S3Bucket": "elasticbeanstalk-us-west-2-0123456789012", - "S3Key": "my-app/22a0-stage-150819_185942.war" + "output":{ + "ApplicationVersion":{ + "ApplicationName":"my-app", + "DateCreated":"2015-08-19T18:59:17.646Z", + "DateUpdated":"2015-08-20T22:53:28.871Z", + "Description":"new description", + "SourceBundle":{ + "S3Bucket":"elasticbeanstalk-us-west-2-0123456789012", + "S3Key":"my-app/22a0-stage-150819_185942.war" }, - "VersionLabel": "22a0-stage-150819_185942" + "VersionLabel":"22a0-stage-150819_185942" } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation updates the description of an application version named 22a0-stage-150819_185942:", - "id": "to-change-an-application-versions-description-1456278019237", - "title": "To change an application version's description" + "description":"The following operation updates the description of an application version named 22a0-stage-150819_185942:", + "id":"to-change-an-application-versions-description-1456278019237", + "title":"To change an application version's description" } ], - "UpdateConfigurationTemplate": [ + "UpdateConfigurationTemplate":[ { - "input": { - "ApplicationName": "my-app", - "OptionsToRemove": [ + "input":{ + "ApplicationName":"my-app", + "OptionsToRemove":[ { - "Namespace": "aws:elasticbeanstalk:healthreporting:system", - "OptionName": "ConfigDocument" + "Namespace":"aws:elasticbeanstalk:healthreporting:system", + "OptionName":"ConfigDocument" } ], - "TemplateName": "my-template" + "TemplateName":"my-template" }, - "output": { - "ApplicationName": "my-app", - "DateCreated": "2015-08-20T22:39:31Z", - "DateUpdated": "2015-08-20T22:43:11Z", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "TemplateName": "my-template" + "output":{ + "ApplicationName":"my-app", + "DateCreated":"2015-08-20T22:39:31Z", + "DateUpdated":"2015-08-20T22:43:11Z", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "TemplateName":"my-template" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation removes the configured CloudWatch custom health metrics configuration ConfigDocument from a saved configuration template named my-template:", - "id": "to-update-a-configuration-template-1456278075300", - "title": "To update a configuration template" + "description":"The following operation removes the configured CloudWatch custom health metrics configuration ConfigDocument from a saved configuration template named my-template:", + "id":"to-update-a-configuration-template-1456278075300", + "title":"To update a configuration template" } ], - "UpdateEnvironment": [ + "UpdateEnvironment":[ { - "input": { - "EnvironmentName": "my-env", - "VersionLabel": "v2" + "input":{ + "EnvironmentName":"my-env", + "VersionLabel":"v2" }, - "output": { - "ApplicationName": "my-app", - "CNAME": "my-env.elasticbeanstalk.com", - "DateCreated": "2015-02-03T23:04:54.453Z", - "DateUpdated": "2015-02-03T23:12:29.119Z", - "EndpointURL": "awseb-e-i-AWSEBLoa-1RDLX6TC9VUAO-0123456789.us-west-2.elb.amazonaws.com", - "EnvironmentId": "e-szqipays4h", - "EnvironmentName": "my-env", - "Health": "Grey", - "SolutionStackName": "64bit Amazon Linux running Tomcat 7", - "Status": "Updating", - "Tier": { - "Name": "WebServer", - "Type": "Standard", - "Version": " " + "output":{ + "ApplicationName":"my-app", + "CNAME":"my-env.elasticbeanstalk.com", + "DateCreated":"2015-02-03T23:04:54.453Z", + "DateUpdated":"2015-02-03T23:12:29.119Z", + "EndpointURL":"awseb-e-i-AWSEBLoa-1RDLX6TC9VUAO-0123456789.us-west-2.elb.amazonaws.com", + "EnvironmentId":"e-szqipays4h", + "EnvironmentName":"my-env", + "Health":"Grey", + "SolutionStackName":"64bit Amazon Linux running Tomcat 7", + "Status":"Updating", + "Tier":{ + "Name":"WebServer", + "Type":"Standard", + "Version":" " }, - "VersionLabel": "v2" + "VersionLabel":"v2" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation updates an environment named \"my-env\" to version \"v2\" of the application to which it belongs:", - "id": "to-update-an-environment-to-a-new-version-1456278210718", - "title": "To update an environment to a new version" + "description":"The following operation updates an environment named \"my-env\" to version \"v2\" of the application to which it belongs:", + "id":"to-update-an-environment-to-a-new-version-1456278210718", + "title":"To update an environment to a new version" }, { - "input": { - "EnvironmentName": "my-env", - "OptionSettings": [ + "input":{ + "EnvironmentName":"my-env", + "OptionSettings":[ { - "Namespace": "aws:elb:healthcheck", - "OptionName": "Interval", - "Value": "15" + "Namespace":"aws:elb:healthcheck", + "OptionName":"Interval", + "Value":"15" }, { - "Namespace": "aws:elb:healthcheck", - "OptionName": "Timeout", - "Value": "8" + "Namespace":"aws:elb:healthcheck", + "OptionName":"Timeout", + "Value":"8" }, { - "Namespace": "aws:elb:healthcheck", - "OptionName": "HealthyThreshold", - "Value": "2" + "Namespace":"aws:elb:healthcheck", + "OptionName":"HealthyThreshold", + "Value":"2" }, { - "Namespace": "aws:elb:healthcheck", - "OptionName": "UnhealthyThreshold", - "Value": "3" + "Namespace":"aws:elb:healthcheck", + "OptionName":"UnhealthyThreshold", + "Value":"3" } ] }, - "output": { - "AbortableOperationInProgress": true, - "ApplicationName": "my-app", - "CNAME": "my-env.elasticbeanstalk.com", - "DateCreated": "2015-08-07T20:48:49.599Z", - "DateUpdated": "2015-08-12T18:15:23.804Z", - "EndpointURL": "awseb-e-w-AWSEBLoa-14XB83101Q4L-104QXY80921.sa-east-1.elb.amazonaws.com", - "EnvironmentId": "e-wtp2rpqsej", - "EnvironmentName": "my-env", - "Health": "Grey", - "SolutionStackName": "64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", - "Status": "Updating", - "Tier": { - "Name": "WebServer", - "Type": "Standard", - "Version": " " + "output":{ + "AbortableOperationInProgress":true, + "ApplicationName":"my-app", + "CNAME":"my-env.elasticbeanstalk.com", + "DateCreated":"2015-08-07T20:48:49.599Z", + "DateUpdated":"2015-08-12T18:15:23.804Z", + "EndpointURL":"awseb-e-w-AWSEBLoa-14XB83101Q4L-104QXY80921.sa-east-1.elb.amazonaws.com", + "EnvironmentId":"e-wtp2rpqsej", + "EnvironmentName":"my-env", + "Health":"Grey", + "SolutionStackName":"64bit Amazon Linux 2015.03 v2.0.0 running Tomcat 8 Java 8", + "Status":"Updating", + "Tier":{ + "Name":"WebServer", + "Type":"Standard", + "Version":" " }, - "VersionLabel": "7f58-stage-150812_025409" + "VersionLabel":"7f58-stage-150812_025409" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation configures several options in the aws:elb:loadbalancer namespace:", - "id": "to-configure-option-settings-1456278286349", - "title": "To configure option settings" + "description":"The following operation configures several options in the aws:elb:loadbalancer namespace:", + "id":"to-configure-option-settings-1456278286349", + "title":"To configure option settings" } ], - "ValidateConfigurationSettings": [ + "ValidateConfigurationSettings":[ { - "input": { - "ApplicationName": "my-app", - "EnvironmentName": "my-env", - "OptionSettings": [ + "input":{ + "ApplicationName":"my-app", + "EnvironmentName":"my-env", + "OptionSettings":[ { - "Namespace": "aws:elasticbeanstalk:healthreporting:system", - "OptionName": "ConfigDocument", - "Value": "{\"CloudWatchMetrics\": {\"Environment\": {\"ApplicationLatencyP99.9\": null,\"InstancesSevere\": 60,\"ApplicationLatencyP90\": 60,\"ApplicationLatencyP99\": null,\"ApplicationLatencyP95\": 60,\"InstancesUnknown\": 60,\"ApplicationLatencyP85\": 60,\"InstancesInfo\": null,\"ApplicationRequests2xx\": null,\"InstancesDegraded\": null,\"InstancesWarning\": 60,\"ApplicationLatencyP50\": 60,\"ApplicationRequestsTotal\": null,\"InstancesNoData\": null,\"InstancesPending\": 60,\"ApplicationLatencyP10\": null,\"ApplicationRequests5xx\": null,\"ApplicationLatencyP75\": null,\"InstancesOk\": 60,\"ApplicationRequests3xx\": null,\"ApplicationRequests4xx\": null},\"Instance\": {\"ApplicationLatencyP99.9\": null,\"ApplicationLatencyP90\": 60,\"ApplicationLatencyP99\": null,\"ApplicationLatencyP95\": null,\"ApplicationLatencyP85\": null,\"CPUUser\": 60,\"ApplicationRequests2xx\": null,\"CPUIdle\": null,\"ApplicationLatencyP50\": null,\"ApplicationRequestsTotal\": 60,\"RootFilesystemUtil\": null,\"LoadAverage1min\": null,\"CPUIrq\": null,\"CPUNice\": 60,\"CPUIowait\": 60,\"ApplicationLatencyP10\": null,\"LoadAverage5min\": null,\"ApplicationRequests5xx\": null,\"ApplicationLatencyP75\": 60,\"CPUSystem\": 60,\"ApplicationRequests3xx\": 60,\"ApplicationRequests4xx\": null,\"InstanceHealth\": null,\"CPUSoftirq\": 60}},\"Version\": 1}" + "Namespace":"aws:elasticbeanstalk:healthreporting:system", + "OptionName":"ConfigDocument", + "Value":"{\"CloudWatchMetrics\": {\"Environment\": {\"ApplicationLatencyP99.9\": null,\"InstancesSevere\": 60,\"ApplicationLatencyP90\": 60,\"ApplicationLatencyP99\": null,\"ApplicationLatencyP95\": 60,\"InstancesUnknown\": 60,\"ApplicationLatencyP85\": 60,\"InstancesInfo\": null,\"ApplicationRequests2xx\": null,\"InstancesDegraded\": null,\"InstancesWarning\": 60,\"ApplicationLatencyP50\": 60,\"ApplicationRequestsTotal\": null,\"InstancesNoData\": null,\"InstancesPending\": 60,\"ApplicationLatencyP10\": null,\"ApplicationRequests5xx\": null,\"ApplicationLatencyP75\": null,\"InstancesOk\": 60,\"ApplicationRequests3xx\": null,\"ApplicationRequests4xx\": null},\"Instance\": {\"ApplicationLatencyP99.9\": null,\"ApplicationLatencyP90\": 60,\"ApplicationLatencyP99\": null,\"ApplicationLatencyP95\": null,\"ApplicationLatencyP85\": null,\"CPUUser\": 60,\"ApplicationRequests2xx\": null,\"CPUIdle\": null,\"ApplicationLatencyP50\": null,\"ApplicationRequestsTotal\": 60,\"RootFilesystemUtil\": null,\"LoadAverage1min\": null,\"CPUIrq\": null,\"CPUNice\": 60,\"CPUIowait\": 60,\"ApplicationLatencyP10\": null,\"LoadAverage5min\": null,\"ApplicationRequests5xx\": null,\"ApplicationLatencyP75\": 60,\"CPUSystem\": 60,\"ApplicationRequests3xx\": 60,\"ApplicationRequests4xx\": null,\"InstanceHealth\": null,\"CPUSoftirq\": 60}},\"Version\": 1}" } ] }, - "output": { - "Messages": [ + "output":{ + "Messages":[ ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following operation validates a CloudWatch custom metrics config document:", - "id": "to-validate-configuration-settings-1456278393654", - "title": "To validate configuration settings" + "description":"The following operation validates a CloudWatch custom metrics config document:", + "id":"to-validate-configuration-settings-1456278393654", + "title":"To validate configuration settings" } ] } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/api-2.json index 60d9ffc20..07b629b0d 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/api-2.json @@ -71,7 +71,9 @@ {"shape":"CertificateNotFoundException"}, {"shape":"UnsupportedProtocolException"}, {"shape":"TooManyRegistrationsForTargetIdException"}, - {"shape":"TooManyTargetsException"} + {"shape":"TooManyTargetsException"}, + {"shape":"TooManyActionsException"}, + {"shape":"InvalidLoadBalancerActionException"} ] }, "CreateLoadBalancer":{ @@ -122,7 +124,10 @@ {"shape":"TargetGroupNotFoundException"}, {"shape":"InvalidConfigurationRequestException"}, {"shape":"TooManyRegistrationsForTargetIdException"}, - {"shape":"TooManyTargetsException"} + {"shape":"TooManyTargetsException"}, + {"shape":"UnsupportedProtocolException"}, + {"shape":"TooManyActionsException"}, + {"shape":"InvalidLoadBalancerActionException"} ] }, "CreateTargetGroup":{ @@ -261,7 +266,8 @@ }, "errors":[ {"shape":"ListenerNotFoundException"}, - {"shape":"LoadBalancerNotFoundException"} + {"shape":"LoadBalancerNotFoundException"}, + {"shape":"UnsupportedProtocolException"} ] }, "DescribeLoadBalancerAttributes":{ @@ -307,7 +313,8 @@ }, "errors":[ {"shape":"ListenerNotFoundException"}, - {"shape":"RuleNotFoundException"} + {"shape":"RuleNotFoundException"}, + {"shape":"UnsupportedProtocolException"} ] }, "DescribeSSLPolicies":{ @@ -415,7 +422,9 @@ {"shape":"InvalidConfigurationRequestException"}, {"shape":"UnsupportedProtocolException"}, {"shape":"TooManyRegistrationsForTargetIdException"}, - {"shape":"TooManyTargetsException"} + {"shape":"TooManyTargetsException"}, + {"shape":"TooManyActionsException"}, + {"shape":"InvalidLoadBalancerActionException"} ] }, "ModifyLoadBalancerAttributes":{ @@ -452,7 +461,10 @@ {"shape":"OperationNotPermittedException"}, {"shape":"TooManyRegistrationsForTargetIdException"}, {"shape":"TooManyTargetsException"}, - {"shape":"TargetGroupNotFoundException"} + {"shape":"TargetGroupNotFoundException"}, + {"shape":"UnsupportedProtocolException"}, + {"shape":"TooManyActionsException"}, + {"shape":"InvalidLoadBalancerActionException"} ] }, "ModifyTargetGroup":{ @@ -615,18 +627,27 @@ "shapes":{ "Action":{ "type":"structure", - "required":[ - "Type", - "TargetGroupArn" - ], + "required":["Type"], "members":{ "Type":{"shape":"ActionTypeEnum"}, - "TargetGroupArn":{"shape":"TargetGroupArn"} + "TargetGroupArn":{"shape":"TargetGroupArn"}, + "AuthenticateOidcConfig":{"shape":"AuthenticateOidcActionConfig"}, + "AuthenticateCognitoConfig":{"shape":"AuthenticateCognitoActionConfig"}, + "Order":{"shape":"ActionOrder"} } }, + "ActionOrder":{ + "type":"integer", + "max":50000, + "min":1 + }, "ActionTypeEnum":{ "type":"string", - "enum":["forward"] + "enum":[ + "forward", + "authenticate-oidc", + "authenticate-cognito" + ] }, "Actions":{ "type":"list", @@ -677,6 +698,93 @@ }, "exception":true }, + "AuthenticateCognitoActionAuthenticationRequestExtraParams":{ + "type":"map", + "key":{"shape":"AuthenticateCognitoActionAuthenticationRequestParamName"}, + "value":{"shape":"AuthenticateCognitoActionAuthenticationRequestParamValue"} + }, + "AuthenticateCognitoActionAuthenticationRequestParamName":{"type":"string"}, + "AuthenticateCognitoActionAuthenticationRequestParamValue":{"type":"string"}, + "AuthenticateCognitoActionConditionalBehaviorEnum":{ + "type":"string", + "enum":[ + "deny", + "allow", + "authenticate" + ] + }, + "AuthenticateCognitoActionConfig":{ + "type":"structure", + "required":[ + "UserPoolArn", + "UserPoolClientId", + "UserPoolDomain" + ], + "members":{ + "UserPoolArn":{"shape":"AuthenticateCognitoActionUserPoolArn"}, + "UserPoolClientId":{"shape":"AuthenticateCognitoActionUserPoolClientId"}, + "UserPoolDomain":{"shape":"AuthenticateCognitoActionUserPoolDomain"}, + "SessionCookieName":{"shape":"AuthenticateCognitoActionSessionCookieName"}, + "Scope":{"shape":"AuthenticateCognitoActionScope"}, + "SessionTimeout":{"shape":"AuthenticateCognitoActionSessionTimeout"}, + "AuthenticationRequestExtraParams":{"shape":"AuthenticateCognitoActionAuthenticationRequestExtraParams"}, + "OnUnauthenticatedRequest":{"shape":"AuthenticateCognitoActionConditionalBehaviorEnum"} + } + }, + "AuthenticateCognitoActionScope":{"type":"string"}, + "AuthenticateCognitoActionSessionCookieName":{"type":"string"}, + "AuthenticateCognitoActionSessionTimeout":{"type":"long"}, + "AuthenticateCognitoActionUserPoolArn":{"type":"string"}, + "AuthenticateCognitoActionUserPoolClientId":{"type":"string"}, + "AuthenticateCognitoActionUserPoolDomain":{"type":"string"}, + "AuthenticateOidcActionAuthenticationRequestExtraParams":{ + "type":"map", + "key":{"shape":"AuthenticateOidcActionAuthenticationRequestParamName"}, + "value":{"shape":"AuthenticateOidcActionAuthenticationRequestParamValue"} + }, + "AuthenticateOidcActionAuthenticationRequestParamName":{"type":"string"}, + "AuthenticateOidcActionAuthenticationRequestParamValue":{"type":"string"}, + "AuthenticateOidcActionAuthorizationEndpoint":{"type":"string"}, + "AuthenticateOidcActionClientId":{"type":"string"}, + "AuthenticateOidcActionClientSecret":{"type":"string"}, + "AuthenticateOidcActionConditionalBehaviorEnum":{ + "type":"string", + "enum":[ + "deny", + "allow", + "authenticate" + ] + }, + "AuthenticateOidcActionConfig":{ + "type":"structure", + "required":[ + "Issuer", + "AuthorizationEndpoint", + "TokenEndpoint", + "UserInfoEndpoint", + "ClientId", + "ClientSecret" + ], + "members":{ + "Issuer":{"shape":"AuthenticateOidcActionIssuer"}, + "AuthorizationEndpoint":{"shape":"AuthenticateOidcActionAuthorizationEndpoint"}, + "TokenEndpoint":{"shape":"AuthenticateOidcActionTokenEndpoint"}, + "UserInfoEndpoint":{"shape":"AuthenticateOidcActionUserInfoEndpoint"}, + "ClientId":{"shape":"AuthenticateOidcActionClientId"}, + "ClientSecret":{"shape":"AuthenticateOidcActionClientSecret"}, + "SessionCookieName":{"shape":"AuthenticateOidcActionSessionCookieName"}, + "Scope":{"shape":"AuthenticateOidcActionScope"}, + "SessionTimeout":{"shape":"AuthenticateOidcActionSessionTimeout"}, + "AuthenticationRequestExtraParams":{"shape":"AuthenticateOidcActionAuthenticationRequestExtraParams"}, + "OnUnauthenticatedRequest":{"shape":"AuthenticateOidcActionConditionalBehaviorEnum"} + } + }, + "AuthenticateOidcActionIssuer":{"type":"string"}, + "AuthenticateOidcActionScope":{"type":"string"}, + "AuthenticateOidcActionSessionCookieName":{"type":"string"}, + "AuthenticateOidcActionSessionTimeout":{"type":"long"}, + "AuthenticateOidcActionTokenEndpoint":{"type":"string"}, + "AuthenticateOidcActionUserInfoEndpoint":{"type":"string"}, "AvailabilityZone":{ "type":"structure", "members":{ @@ -1159,6 +1267,17 @@ }, "exception":true }, + "InvalidLoadBalancerActionException":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"InvalidLoadBalancerAction", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, "InvalidSchemeException":{ "type":"structure", "members":{ @@ -1692,10 +1811,7 @@ }, "SetSubnetsInput":{ "type":"structure", - "required":[ - "LoadBalancerArn", - "Subnets" - ], + "required":["LoadBalancerArn"], "members":{ "LoadBalancerArn":{"shape":"LoadBalancerArn"}, "Subnets":{"shape":"Subnets"}, @@ -1941,6 +2057,17 @@ "ip" ] }, + "TooManyActionsException":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"TooManyActions", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, "TooManyCertificatesException":{ "type":"structure", "members":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/docs-2.json index 87add7b22..a0e2d902d 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/elasticloadbalancingv2/2015-12-01/docs-2.json @@ -6,7 +6,7 @@ "AddTags": "

Adds the specified tags to the specified Elastic Load Balancing resource. You can tag your Application Load Balancers, Network Load Balancers, and your target groups.

Each tag consists of a key and an optional value. If a resource already has a tag with the same key, AddTags updates its value.

To list the current tags for your resources, use DescribeTags. To remove tags from your resources, use RemoveTags.

", "CreateListener": "

Creates a listener for the specified Application Load Balancer or Network Load Balancer.

To update a listener, use ModifyListener. When you are finished with a listener, you can delete it using DeleteListener. If you are finished with both the listener and the load balancer, you can delete them both using DeleteLoadBalancer.

This operation is idempotent, which means that it completes at most one time. If you attempt to create multiple listeners with the same settings, each call succeeds.

For more information, see Listeners for Your Application Load Balancers in the Application Load Balancers Guide and Listeners for Your Network Load Balancers in the Network Load Balancers Guide.

", "CreateLoadBalancer": "

Creates an Application Load Balancer or a Network Load Balancer.

When you create a load balancer, you can specify security groups, public subnets, IP address type, and tags. Otherwise, you could do so later using SetSecurityGroups, SetSubnets, SetIpAddressType, and AddTags.

To create listeners for your load balancer, use CreateListener. To describe your current load balancers, see DescribeLoadBalancers. When you are finished with a load balancer, you can delete it using DeleteLoadBalancer.

For limit information, see Limits for Your Application Load Balancer in the Application Load Balancers Guide and Limits for Your Network Load Balancer in the Network Load Balancers Guide.

This operation is idempotent, which means that it completes at most one time. If you attempt to create multiple load balancers with the same settings, each call succeeds.

For more information, see Application Load Balancers in the Application Load Balancers Guide and Network Load Balancers in the Network Load Balancers Guide.

", - "CreateRule": "

Creates a rule for the specified listener. The listener must be associated with an Application Load Balancer.

Rules are evaluated in priority order, from the lowest value to the highest value. When the condition for a rule is met, the specified action is taken. If no conditions are met, the action for the default rule is taken. For more information, see Listener Rules in the Application Load Balancers Guide.

To view your current rules, use DescribeRules. To update a rule, use ModifyRule. To set the priorities of your rules, use SetRulePriorities. To delete a rule, use DeleteRule.

", + "CreateRule": "

Creates a rule for the specified listener. The listener must be associated with an Application Load Balancer.

Rules are evaluated in priority order, from the lowest value to the highest value. When the conditions for a rule are met, its actions are performed. If the conditions for no rules are met, the actions for the default rule are performed. For more information, see Listener Rules in the Application Load Balancers Guide.

To view your current rules, use DescribeRules. To update a rule, use ModifyRule. To set the priorities of your rules, use SetRulePriorities. To delete a rule, use DeleteRule.

", "CreateTargetGroup": "

Creates a target group.

To register targets with the target group, use RegisterTargets. To update the health check settings for the target group, use ModifyTargetGroup. To monitor the health of targets in the target group, use DescribeTargetHealth.

To route traffic to the targets in a target group, specify the target group in an action using CreateListener or CreateRule.

To delete a target group, use DeleteTargetGroup.

This operation is idempotent, which means that it completes at most one time. If you attempt to create multiple target groups with the same settings, each call succeeds.

For more information, see Target Groups for Your Application Load Balancers in the Application Load Balancers Guide or Target Groups for Your Network Load Balancers in the Network Load Balancers Guide.

", "DeleteListener": "

Deletes the specified listener.

Alternatively, your listener is deleted when you delete the load balancer it is attached to using DeleteLoadBalancer.

", "DeleteLoadBalancer": "

Deletes the specified Application Load Balancer or Network Load Balancer and its attached listeners.

You can't delete a load balancer if deletion protection is enabled. If the load balancer does not exist or has already been deleted, the call succeeds.

Deleting a load balancer does not affect its registered targets. For example, your EC2 instances continue to run and are still registered to their target groups. If you no longer need these EC2 instances, you can stop or terminate them.

", @@ -16,17 +16,17 @@ "DescribeAccountLimits": "

Describes the current Elastic Load Balancing resource limits for your AWS account.

For more information, see Limits for Your Application Load Balancers in the Application Load Balancer Guide or Limits for Your Network Load Balancers in the Network Load Balancers Guide.

", "DescribeListenerCertificates": "

Describes the certificates for the specified secure listener.

", "DescribeListeners": "

Describes the specified listeners or the listeners for the specified Application Load Balancer or Network Load Balancer. You must specify either a load balancer or one or more listeners.

", - "DescribeLoadBalancerAttributes": "

Describes the attributes for the specified Application Load Balancer or Network Load Balancer.

", + "DescribeLoadBalancerAttributes": "

Describes the attributes for the specified Application Load Balancer or Network Load Balancer.

For more information, see Load Balancer Attributes in the Application Load Balancers Guide or Load Balancer Attributes in the Network Load Balancers Guide.

", "DescribeLoadBalancers": "

Describes the specified load balancers or all of your load balancers.

To describe the listeners for a load balancer, use DescribeListeners. To describe the attributes for a load balancer, use DescribeLoadBalancerAttributes.

", "DescribeRules": "

Describes the specified rules or the rules for the specified listener. You must specify either a listener or one or more rules.

", "DescribeSSLPolicies": "

Describes the specified policies or all policies used for SSL negotiation.

For more information, see Security Policies in the Application Load Balancers Guide.

", "DescribeTags": "

Describes the tags for the specified resources. You can describe the tags for one or more Application Load Balancers, Network Load Balancers, and target groups.

", - "DescribeTargetGroupAttributes": "

Describes the attributes for the specified target group.

", + "DescribeTargetGroupAttributes": "

Describes the attributes for the specified target group.

For more information, see Target Group Attributes in the Application Load Balancers Guide or Target Group Attributes in the Network Load Balancers Guide.

", "DescribeTargetGroups": "

Describes the specified target groups or all of your target groups. By default, all target groups are described. Alternatively, you can specify one of the following to filter the results: the ARN of the load balancer, the names of one or more target groups, or the ARNs of one or more target groups.

To describe the targets for a target group, use DescribeTargetHealth. To describe the attributes of a target group, use DescribeTargetGroupAttributes.

", "DescribeTargetHealth": "

Describes the health of the specified targets or all of your targets.

", "ModifyListener": "

Modifies the specified properties of the specified listener.

Any properties that you do not specify retain their current values. However, changing the protocol from HTTPS to HTTP removes the security policy and SSL certificate properties. If you change the protocol from HTTP to HTTPS, you must add the security policy and server certificate.

", "ModifyLoadBalancerAttributes": "

Modifies the specified attributes of the specified Application Load Balancer or Network Load Balancer.

If any of the specified attributes can't be modified as requested, the call fails. Any existing attributes that you do not modify retain their current values.

", - "ModifyRule": "

Modifies the specified rule.

Any existing properties that you do not modify retain their current values.

To modify the default action, use ModifyListener.

", + "ModifyRule": "

Modifies the specified rule.

Any existing properties that you do not modify retain their current values.

To modify the actions for the default rule, use ModifyListener.

", "ModifyTargetGroup": "

Modifies the health checks used when evaluating the health state of the targets in the specified target group.

To monitor the health of the targets, use DescribeTargetHealth.

", "ModifyTargetGroupAttributes": "

Modifies the specified attributes of the specified target group.

", "RegisterTargets": "

Registers the specified targets with the specified target group.

You can register targets by instance ID or by IP address. If the target is an EC2 instance, it must be in the running state when you register it.

By default, the load balancer routes requests to registered targets using the protocol and port for the target group. Alternatively, you can override the port for a target when you register it. You can register each EC2 instance or IP address with the same target group multiple times using different ports.

With a Network Load Balancer, you cannot register instances by instance ID if they have the following instance types: C1, CC1, CC2, CG1, CG2, CR1, CS1, G1, G2, HI1, HS1, M1, M2, M3, and T1. You can register instances of these types by IP address.

To remove a target from a target group, use DeregisterTargets.

", @@ -44,20 +44,26 @@ "Actions$member": null } }, + "ActionOrder": { + "base": null, + "refs": { + "Action$Order": "

The order for the action. This value is required for rules with multiple actions. The action with the lowest value for order is performed first. The forward action must be performed last.

" + } + }, "ActionTypeEnum": { "base": null, "refs": { - "Action$Type": "

The type of action.

" + "Action$Type": "

The type of action. Each rule must include one forward action.

" } }, "Actions": { "base": null, "refs": { - "CreateListenerInput$DefaultActions": "

The default action for the listener. For Application Load Balancers, the protocol of the specified target group must be HTTP or HTTPS. For Network Load Balancers, the protocol of the specified target group must be TCP.

", - "CreateRuleInput$Actions": "

An action. Each action has the type forward and specifies a target group.

", + "CreateListenerInput$DefaultActions": "

The actions for the default rule. The rule must include one forward action.

If the action type is forward, you can specify a single target group. The protocol of the target group must be HTTP or HTTPS for an Application Load Balancer or TCP for a Network Load Balancer.

If the action type is authenticate-oidc, you can use an identity provider that is OpenID Connect (OIDC) compliant to authenticate users as they access your application.

If the action type is authenticate-cognito, you can use Amazon Cognito to authenticate users as they access your application.

", + "CreateRuleInput$Actions": "

The actions. Each rule must include one forward action.

If the action type is forward, you can specify a single target group.

If the action type is authenticate-oidc, you can use an identity provider that is OpenID Connect (OIDC) compliant to authenticate users as they access your application.

If the action type is authenticate-cognito, you can use Amazon Cognito to authenticate users as they access your application.

", "Listener$DefaultActions": "

The default actions for the listener.

", - "ModifyListenerInput$DefaultActions": "

The default action. For Application Load Balancers, the protocol of the specified target group must be HTTP or HTTPS. For Network Load Balancers, the protocol of the specified target group must be TCP.

", - "ModifyRuleInput$Actions": "

The actions. The target group must use the HTTP or HTTPS protocol.

", + "ModifyListenerInput$DefaultActions": "

The actions for the default rule. The rule must include one forward action.

If the action type is forward, you can specify a single target group. The protocol of the target group must be HTTP or HTTPS for an Application Load Balancer or TCP for a Network Load Balancer.

If the action type is authenticate-oidc, you can use an identity provider that is OpenID Connect (OIDC) compliant to authenticate users as they access your application.

If the action type is authenticate-cognito, you can use Amazon Cognito to authenticate users as they access your application.

", + "ModifyRuleInput$Actions": "

The actions.

If the action type is forward, you can specify a single target group.

If the action type is authenticate-oidc, you can use an identity provider that is OpenID Connect (OIDC) compliant to authenticate users as they access your application.

If the action type is authenticate-cognito, you can use Amazon Cognito to authenticate users as they access your application.

", "Rule$Actions": "

The actions.

" } }, @@ -93,6 +99,156 @@ "refs": { } }, + "AuthenticateCognitoActionAuthenticationRequestExtraParams": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$AuthenticationRequestExtraParams": "

The query parameters (up to 10) to include in the redirect request to the authorization endpoint.

" + } + }, + "AuthenticateCognitoActionAuthenticationRequestParamName": { + "base": null, + "refs": { + "AuthenticateCognitoActionAuthenticationRequestExtraParams$key": null + } + }, + "AuthenticateCognitoActionAuthenticationRequestParamValue": { + "base": null, + "refs": { + "AuthenticateCognitoActionAuthenticationRequestExtraParams$value": null + } + }, + "AuthenticateCognitoActionConditionalBehaviorEnum": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$OnUnauthenticatedRequest": "

The behavior if the user is not authenticated. The following are possible values:

  • deny - Return an HTTP 401 Unauthorized error.

  • allow - Allow the request to be forwarded to the target.

  • authenticate - Redirect the request to the IdP authorization endpoint. This is the default value.

" + } + }, + "AuthenticateCognitoActionConfig": { + "base": "

Request parameters to use when integrating with Amazon Cognito to authenticate users.

", + "refs": { + "Action$AuthenticateCognitoConfig": "

[HTTPS listener] Information for using Amazon Cognito to authenticate users. Specify only when Type is authenticate-cognito.

" + } + }, + "AuthenticateCognitoActionScope": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$Scope": "

The set of user claims to be requested from the IdP. The default is openid.

To verify which scope values your IdP supports and how to separate multiple values, see the documentation for your IdP.

" + } + }, + "AuthenticateCognitoActionSessionCookieName": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$SessionCookieName": "

The name of the cookie used to maintain session information. The default is AWSELBAuthSessionCookie.

" + } + }, + "AuthenticateCognitoActionSessionTimeout": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$SessionTimeout": "

The maximum duration of the authentication session, in seconds. The default is 604800 seconds (7 days).

" + } + }, + "AuthenticateCognitoActionUserPoolArn": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$UserPoolArn": "

The Amazon Resource Name (ARN) of the Amazon Cognito user pool.

" + } + }, + "AuthenticateCognitoActionUserPoolClientId": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$UserPoolClientId": "

The ID of the Amazon Cognito user pool client.

" + } + }, + "AuthenticateCognitoActionUserPoolDomain": { + "base": null, + "refs": { + "AuthenticateCognitoActionConfig$UserPoolDomain": "

The domain prefix or fully-qualified domain name of the Amazon Cognito user pool.

" + } + }, + "AuthenticateOidcActionAuthenticationRequestExtraParams": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$AuthenticationRequestExtraParams": "

The query parameters (up to 10) to include in the redirect request to the authorization endpoint.

" + } + }, + "AuthenticateOidcActionAuthenticationRequestParamName": { + "base": null, + "refs": { + "AuthenticateOidcActionAuthenticationRequestExtraParams$key": null + } + }, + "AuthenticateOidcActionAuthenticationRequestParamValue": { + "base": null, + "refs": { + "AuthenticateOidcActionAuthenticationRequestExtraParams$value": null + } + }, + "AuthenticateOidcActionAuthorizationEndpoint": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$AuthorizationEndpoint": "

The authorization endpoint of the IdP. This must be a full URL, including the HTTPS protocol, the domain, and the path.

" + } + }, + "AuthenticateOidcActionClientId": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$ClientId": "

The OAuth 2.0 client identifier.

" + } + }, + "AuthenticateOidcActionClientSecret": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$ClientSecret": "

The OAuth 2.0 client secret.

" + } + }, + "AuthenticateOidcActionConditionalBehaviorEnum": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$OnUnauthenticatedRequest": "

The behavior if the user is not authenticated. The following are possible values:

  • deny - Return an HTTP 401 Unauthorized error.

  • allow - Allow the request to be forwarded to the target.

  • authenticate - Redirect the request to the IdP authorization endpoint. This is the default value.

" + } + }, + "AuthenticateOidcActionConfig": { + "base": "

Request parameters when using an identity provider (IdP) that is compliant with OpenID Connect (OIDC) to authenticate users.

", + "refs": { + "Action$AuthenticateOidcConfig": "

[HTTPS listener] Information about an identity provider that is compliant with OpenID Connect (OIDC). Specify only when Type is authenticate-oidc.

" + } + }, + "AuthenticateOidcActionIssuer": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$Issuer": "

The OIDC issuer identifier of the IdP. This must be a full URL, including the HTTPS protocol, the domain, and the path.

" + } + }, + "AuthenticateOidcActionScope": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$Scope": "

The set of user claims to be requested from the IdP. The default is openid.

To verify which scope values your IdP supports and how to separate multiple values, see the documentation for your IdP.

" + } + }, + "AuthenticateOidcActionSessionCookieName": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$SessionCookieName": "

The name of the cookie used to maintain session information. The default is AWSELBAuthSessionCookie.

" + } + }, + "AuthenticateOidcActionSessionTimeout": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$SessionTimeout": "

The maximum duration of the authentication session, in seconds. The default is 604800 seconds (7 days).

" + } + }, + "AuthenticateOidcActionTokenEndpoint": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$TokenEndpoint": "

The token endpoint of the IdP. This must be a full URL, including the HTTPS protocol, the domain, and the path.

" + } + }, + "AuthenticateOidcActionUserInfoEndpoint": { + "base": null, + "refs": { + "AuthenticateOidcActionConfig$UserInfoEndpoint": "

The user info endpoint of the IdP. This must be a full URL, including the HTTPS protocol, the domain, and the path.

" + } + }, "AvailabilityZone": { "base": "

Information about an Availability Zone.

", "refs": { @@ -134,10 +290,10 @@ "refs": { "AddListenerCertificatesInput$Certificates": "

The certificate to add. You can specify one certificate per call.

", "AddListenerCertificatesOutput$Certificates": "

Information about the certificates.

", - "CreateListenerInput$Certificates": "

[HTTPS listeners] The SSL server certificate. You must provide exactly one certificate.

", + "CreateListenerInput$Certificates": "

[HTTPS listeners] The default SSL server certificate. You must provide exactly one certificate. To create a certificate list, use AddListenerCertificates.

", "DescribeListenerCertificatesOutput$Certificates": "

Information about the certificates.

", "Listener$Certificates": "

The SSL server certificate. You must provide a certificate if the protocol is HTTPS.

", - "ModifyListenerInput$Certificates": "

The default SSL server certificate.

", + "ModifyListenerInput$Certificates": "

[HTTPS listeners] The default SSL server certificate. You must provide exactly one certificate. To create a certificate list, use AddListenerCertificates.

", "RemoveListenerCertificatesInput$Certificates": "

The certificate to remove. You can specify one certificate per call.

" } }, @@ -477,6 +633,11 @@ "refs": { } }, + "InvalidLoadBalancerActionException": { + "base": "

The requested action is not valid.

", + "refs": { + } + }, "InvalidSchemeException": { "base": "

The requested scheme is not valid.

", "refs": { @@ -572,7 +733,7 @@ "refs": { "CreateListenerOutput$Listeners": "

Information about the listener.

", "DescribeListenersOutput$Listeners": "

Information about the listeners.

", - "ModifyListenerOutput$Listeners": "

Information about the modified listeners.

" + "ModifyListenerOutput$Listeners": "

Information about the modified listener.

" } }, "LoadBalancer": { @@ -626,7 +787,7 @@ "LoadBalancerAttributeKey": { "base": null, "refs": { - "LoadBalancerAttribute$Key": "

The name of the attribute.

  • access_logs.s3.enabled - [Application Load Balancers] Indicates whether access logs stored in Amazon S3 are enabled. The value is true or false.

  • access_logs.s3.bucket - [Application Load Balancers] The name of the S3 bucket for the access logs. This attribute is required if access logs in Amazon S3 are enabled. The bucket must exist in the same region as the load balancer and have a bucket policy that grants Elastic Load Balancing permission to write to the bucket.

  • access_logs.s3.prefix - [Application Load Balancers] The prefix for the location in the S3 bucket. If you don't specify a prefix, the access logs are stored in the root of the bucket.

  • deletion_protection.enabled - Indicates whether deletion protection is enabled. The value is true or false.

  • idle_timeout.timeout_seconds - [Application Load Balancers] The idle timeout value, in seconds. The valid range is 1-4000. The default is 60 seconds.

  • load_balancing.cross_zone.enabled - [Network Load Balancers] Indicates whether cross-zone load balancing is enabled. The value is true or false. The default is false.

  • routing.http2.enabled - [Application Load Balancers] Indicates whether HTTP/2 is enabled. The value is true or false. The default is true.

" + "LoadBalancerAttribute$Key": "

The name of the attribute.

The following attributes are supported by both Application Load Balancers and Network Load Balancers:

  • deletion_protection.enabled - Indicates whether deletion protection is enabled. The value is true or false. The default is false.

The following attributes are supported by only Application Load Balancers:

  • access_logs.s3.enabled - Indicates whether access logs are enabled. The value is true or false. The default is false.

  • access_logs.s3.bucket - The name of the S3 bucket for the access logs. This attribute is required if access logs are enabled. The bucket must exist in the same region as the load balancer and have a bucket policy that grants Elastic Load Balancing permission to write to the bucket.

  • access_logs.s3.prefix - The prefix for the location in the S3 bucket for the access logs.

  • idle_timeout.timeout_seconds - The idle timeout value, in seconds. The valid range is 1-4000 seconds. The default is 60 seconds.

  • routing.http2.enabled - Indicates whether HTTP/2 is enabled. The value is true or false. The default is true.

The following attributes are supported by only Network Load Balancers:

  • load_balancing.cross_zone.enabled - Indicates whether cross-zone load balancing is enabled. The value is true or false. The default is false.

" } }, "LoadBalancerAttributeValue": { @@ -646,7 +807,7 @@ "LoadBalancerName": { "base": null, "refs": { - "CreateLoadBalancerInput$Name": "

The name of the load balancer.

This name must be unique per region per account, can have a maximum of 32 characters, must contain only alphanumeric characters or hyphens, and must not begin or end with a hyphen.

", + "CreateLoadBalancerInput$Name": "

The name of the load balancer.

This name must be unique per region per account, can have a maximum of 32 characters, must contain only alphanumeric characters or hyphens, must not begin or end with a hyphen, and must not begin with \"internal-\".

", "LoadBalancer$LoadBalancerName": "

The name of the load balancer.

", "LoadBalancerNames$member": null } @@ -920,7 +1081,7 @@ "base": null, "refs": { "CreateRuleInput$Conditions": "

The conditions. Each condition specifies a field name and a single value.

If the field name is host-header, you can specify a single host name (for example, my.example.com). A host name is case insensitive, can be up to 128 characters in length, and can contain any of the following characters. Note that you can include up to three wildcard characters.

  • A-Z, a-z, 0-9

  • - .

  • * (matches 0 or more characters)

  • ? (matches exactly 1 character)

If the field name is path-pattern, you can specify a single path pattern. A path pattern is case sensitive, can be up to 128 characters in length, and can contain any of the following characters. Note that you can include up to three wildcard characters.

  • A-Z, a-z, 0-9

  • _ - . $ / ~ \" ' @ : +

  • & (using &amp;)

  • * (matches 0 or more characters)

  • ? (matches exactly 1 character)

", - "ModifyRuleInput$Conditions": "

The conditions.

", + "ModifyRuleInput$Conditions": "

The conditions. Each condition specifies a field name and a single value.

If the field name is host-header, you can specify a single host name (for example, my.example.com). A host name is case insensitive, can be up to 128 characters in length, and can contain any of the following characters. Note that you can include up to three wildcard characters.

  • A-Z, a-z, 0-9

  • - .

  • * (matches 0 or more characters)

  • ? (matches exactly 1 character)

If the field name is path-pattern, you can specify a single path pattern. A path pattern is case sensitive, can be up to 128 characters in length, and can contain any of the following characters. Note that you can include up to three wildcard characters.

  • A-Z, a-z, 0-9

  • _ - . $ / ~ \" ' @ : +

  • & (using &amp;)

  • * (matches 0 or more characters)

  • ? (matches exactly 1 character)

", "Rule$Conditions": "

The conditions.

" } }, @@ -932,7 +1093,7 @@ "RulePriority": { "base": null, "refs": { - "CreateRuleInput$Priority": "

The priority for the rule. A listener can't have multiple rules with the same priority.

", + "CreateRuleInput$Priority": "

The rule priority. A listener can't have multiple rules with the same priority.

", "RulePriorityPair$Priority": "

The rule priority.

" } }, @@ -953,7 +1114,7 @@ "refs": { "CreateRuleOutput$Rules": "

Information about the rule.

", "DescribeRulesOutput$Rules": "

Information about the rules.

", - "ModifyRuleOutput$Rules": "

Information about the rule.

", + "ModifyRuleOutput$Rules": "

Information about the modified rule.

", "SetRulePrioritiesOutput$Rules": "

Information about the rules.

" } }, @@ -1034,7 +1195,7 @@ "refs": { "CreateListenerInput$SslPolicy": "

[HTTPS listeners] The security policy that defines which ciphers and protocols are supported. The default is the current predefined security policy.

", "Listener$SslPolicy": "

The security policy that defines which ciphers and protocols are supported. The default is the current predefined security policy.

", - "ModifyListenerInput$SslPolicy": "

The security policy that defines which protocols and ciphers are supported. For more information, see Security Policies in the Application Load Balancers Guide.

", + "ModifyListenerInput$SslPolicy": "

[HTTPS listeners] The security policy that defines which protocols and ciphers are supported. For more information, see Security Policies in the Application Load Balancers Guide.

", "SslPolicy$Name": "

The name of the policy.

", "SslPolicyNames$member": null } @@ -1177,7 +1338,7 @@ "TargetGroupArn": { "base": null, "refs": { - "Action$TargetGroupArn": "

The Amazon Resource Name (ARN) of the target group.

", + "Action$TargetGroupArn": "

The Amazon Resource Name (ARN) of the target group. Specify only when Type is forward.

For a default rule, the protocol of the target group must be HTTP or HTTPS for an Application Load Balancer or TCP for a Network Load Balancer.

", "DeleteTargetGroupInput$TargetGroupArn": "

The Amazon Resource Name (ARN) of the target group.

", "DeregisterTargetsInput$TargetGroupArn": "

The Amazon Resource Name (ARN) of the target group.

", "DescribeTargetGroupAttributesInput$TargetGroupArn": "

The Amazon Resource Name (ARN) of the target group.

", @@ -1209,7 +1370,7 @@ "TargetGroupAttributeKey": { "base": null, "refs": { - "TargetGroupAttribute$Key": "

The name of the attribute.

  • deregistration_delay.timeout_seconds - The amount time for Elastic Load Balancing to wait before changing the state of a deregistering target from draining to unused. The range is 0-3600 seconds. The default value is 300 seconds.

  • proxy_protocol_v2.enabled - [Network Load Balancers] Indicates whether Proxy Protocol version 2 is enabled.

  • stickiness.enabled - [Application Load Balancers] Indicates whether sticky sessions are enabled. The value is true or false.

  • stickiness.type - [Application Load Balancers] The type of sticky sessions. The possible value is lb_cookie.

  • stickiness.lb_cookie.duration_seconds - [Application Load Balancers] The time period, in seconds, during which requests from a client should be routed to the same target. After this time period expires, the load balancer-generated cookie is considered stale. The range is 1 second to 1 week (604800 seconds). The default value is 1 day (86400 seconds).

" + "TargetGroupAttribute$Key": "

The name of the attribute.

The following attributes are supported by both Application Load Balancers and Network Load Balancers:

  • deregistration_delay.timeout_seconds - The amount of time, in seconds, for Elastic Load Balancing to wait before changing the state of a deregistering target from draining to unused. The range is 0-3600 seconds. The default value is 300 seconds.

The following attributes are supported by only Application Load Balancers:

  • slow_start.duration_seconds - The time period, in seconds, during which a newly registered target receives a linearly increasing share of the traffic to the target group. After this time period ends, the target receives its full share of traffic. The range is 30-900 seconds (15 minutes). Slow start mode is disabled by default.

  • stickiness.enabled - Indicates whether sticky sessions are enabled. The value is true or false. The default is false.

  • stickiness.type - The type of sticky sessions. The possible value is lb_cookie.

  • stickiness.lb_cookie.duration_seconds - The time period, in seconds, during which requests from a client should be routed to the same target. After this time period expires, the load balancer-generated cookie is considered stale. The range is 1 second to 1 week (604800 seconds). The default value is 1 day (86400 seconds).

The following attributes are supported by only Network Load Balancers:

  • proxy_protocol_v2.enabled - Indicates whether Proxy Protocol version 2 is enabled. The value is true or false. The default is false.

" } }, "TargetGroupAttributeValue": { @@ -1250,7 +1411,7 @@ "refs": { "CreateTargetGroupOutput$TargetGroups": "

Information about the target group.

", "DescribeTargetGroupsOutput$TargetGroups": "

Information about the target groups.

", - "ModifyTargetGroupOutput$TargetGroups": "

Information about the target group.

" + "ModifyTargetGroupOutput$TargetGroups": "

Information about the modified target group.

" } }, "TargetHealth": { @@ -1296,6 +1457,11 @@ "TargetGroup$TargetType": "

The type of target that you must specify when registering targets with this target group. The possible values are instance (targets are specified by instance ID) or ip (targets are specified by IP address).

" } }, + "TooManyActionsException": { + "base": "

You've reached the limit on the number of actions per rule.

", + "refs": { + } + }, "TooManyCertificatesException": { "base": "

You've reached the limit on the number of certificates per load balancer.

", "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/docs-2.json index 8a2da5674..0b8343c24 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/docs-2.json @@ -1,20 +1,20 @@ { "version": "2.0", - "service": "Amazon Simple Email Service

This is the API Reference for Amazon Simple Email Service (Amazon SES). This documentation is intended to be used in conjunction with the Amazon SES Developer Guide.

For a list of Amazon SES endpoints to use in service requests, see Regions and Amazon SES in the Amazon SES Developer Guide.

", + "service": "Amazon Simple Email Service

This document contains reference information for the Amazon Simple Email Service (Amazon SES) API, version 2010-12-01. This document is best used in conjunction with the Amazon SES Developer Guide.

For a list of Amazon SES endpoints to use in service requests, see Regions and Amazon SES in the Amazon SES Developer Guide.

", "operations": { "CloneReceiptRuleSet": "

Creates a receipt rule set by cloning an existing one. All receipt rules and configurations are copied to the new receipt rule set and are completely independent of the source rule set.

For information about setting up rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "CreateConfigurationSet": "

Creates a configuration set.

Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "CreateConfigurationSetEventDestination": "

Creates a configuration set event destination.

When you create or update an event destination, you must provide one, and only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).

An event destination is the AWS service to which Amazon SES publishes the email sending events associated with a configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "CreateConfigurationSetTrackingOptions": "

Creates an association between a configuration set and a custom domain for open and click event tracking.

By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using configuration sets, see Configuring Custom Domains to Handle Open and Click Tracking in the Amazon SES Developer Guide.

", - "CreateCustomVerificationEmailTemplate": "

Creates a new custom verification email template.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "CreateConfigurationSetEventDestination": "

Creates a configuration set event destination.

When you create or update an event destination, you must provide one, and only one, destination. The destination can be CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).

An event destination is the AWS service to which Amazon SES publishes the email sending events associated with a configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "CreateConfigurationSetTrackingOptions": "

Creates an association between a configuration set and a custom domain for open and click event tracking.

By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.

", + "CreateCustomVerificationEmailTemplate": "

Creates a new custom verification email template.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "CreateReceiptFilter": "

Creates a new IP address filter.

For information about setting up IP address filters, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "CreateReceiptRule": "

Creates a receipt rule.

For information about setting up receipt rules, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "CreateReceiptRuleSet": "

Creates an empty receipt rule set.

For information about setting up receipt rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "CreateTemplate": "

Creates an email template. Email templates enable you to send personalized email to one or more destinations in a single API operation. For more information, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "DeleteConfigurationSet": "

Deletes a configuration set. Configuration sets enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "DeleteConfigurationSetEventDestination": "

Deletes a configuration set event destination. Configuration set event destinations are associated with configuration sets, which enable you to publish email sending events. For information about using configuration sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "DeleteConfigurationSetTrackingOptions": "

Deletes an association between a configuration set and a custom domain for open and click event tracking.

By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using configuration sets, see Configuring Custom Domains to Handle Open and Click Tracking in the Amazon SES Developer Guide.

Deleting this kind of association will result in emails sent using the specified configuration set to capture open and click events using the standard, Amazon SES-operated domains.

", - "DeleteCustomVerificationEmailTemplate": "

Deletes an existing custom verification email template.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "DeleteConfigurationSetTrackingOptions": "

Deletes an association between a configuration set and a custom domain for open and click event tracking.

By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.

Deleting this kind of association will result in emails sent using the specified configuration set to capture open and click events using the standard, Amazon SES-operated domains.

", + "DeleteCustomVerificationEmailTemplate": "

Deletes an existing custom verification email template.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "DeleteIdentity": "

Deletes the specified identity (an email address or a domain) from the list of verified identities.

You can execute this operation no more than once per second.

", "DeleteIdentityPolicy": "

Deletes the specified sending authorization policy for the given identity (an email address or a domain). This API returns successfully even if a policy with the specified name does not exist.

This API is for the identity owner only. If you have not verified the identity, this API will return an error.

Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "DeleteReceiptFilter": "

Deletes the specified IP address filter.

For information about managing IP address filters, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", @@ -26,32 +26,32 @@ "DescribeConfigurationSet": "

Returns the details of the specified configuration set. For information about using configuration sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "DescribeReceiptRule": "

Returns the details of the specified receipt rule.

For information about setting up receipt rules, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "DescribeReceiptRuleSet": "

Returns the details of the specified receipt rule set.

For information about managing receipt rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "GetAccountSendingEnabled": "

Returns the email sending status of the Amazon SES account.

You can execute this operation no more than once per second.

", - "GetCustomVerificationEmailTemplate": "

Returns the custom email verification template for the template name you specify.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "GetAccountSendingEnabled": "

Returns the email sending status of the Amazon SES account for the current region.

You can execute this operation no more than once per second.

", + "GetCustomVerificationEmailTemplate": "

Returns the custom email verification template for the template name you specify.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "GetIdentityDkimAttributes": "

Returns the current status of Easy DKIM signing for an entity. For domain name identities, this operation also returns the DKIM tokens that are required for Easy DKIM signing, and whether Amazon SES has successfully verified that these tokens have been published.

This operation takes a list of identities as input and returns the following information for each:

  • Whether Easy DKIM signing is enabled or disabled.

  • A set of DKIM tokens that represent the identity. If the identity is an email address, the tokens represent the domain of that address.

  • Whether Amazon SES has successfully verified the DKIM tokens published in the domain's DNS. This information is only returned for domain name identities, not for email addresses.

This operation is throttled at one request per second and can only get DKIM attributes for up to 100 identities at a time.

For more information about creating DNS records using DKIM tokens, go to the Amazon SES Developer Guide.

", "GetIdentityMailFromDomainAttributes": "

Returns the custom MAIL FROM attributes for a list of identities (email addresses : domains).

This operation is throttled at one request per second and can only get custom MAIL FROM attributes for up to 100 identities at a time.

", "GetIdentityNotificationAttributes": "

Given a list of verified identities (email addresses and/or domains), returns a structure describing identity notification attributes.

This operation is throttled at one request per second and can only get notification attributes for up to 100 identities at a time.

For more information about using notifications with Amazon SES, see the Amazon SES Developer Guide.

", "GetIdentityPolicies": "

Returns the requested sending authorization policies for the given identity (an email address or a domain). The policies are returned as a map of policy names to policy contents. You can retrieve a maximum of 20 policies at a time.

This API is for the identity owner only. If you have not verified the identity, this API will return an error.

Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "GetIdentityVerificationAttributes": "

Given a list of identities (email addresses and/or domains), returns the verification status and (for domain identities) the verification token for each identity.

The verification status of an email address is \"Pending\" until the email address owner clicks the link within the verification email that Amazon SES sent to that address. If the email address owner clicks the link within 24 hours, the verification status of the email address changes to \"Success\". If the link is not clicked within 24 hours, the verification status changes to \"Failed.\" In that case, if you still want to verify the email address, you must restart the verification process from the beginning.

For domain identities, the domain's verification status is \"Pending\" as Amazon SES searches for the required TXT record in the DNS settings of the domain. When Amazon SES detects the record, the domain's verification status changes to \"Success\". If Amazon SES is unable to detect the record within 72 hours, the domain's verification status changes to \"Failed.\" In that case, if you still want to verify the domain, you must restart the verification process from the beginning.

This operation is throttled at one request per second and can only get verification attributes for up to 100 identities at a time.

", "GetSendQuota": "

Provides the sending limits for the Amazon SES account.

You can execute this operation no more than once per second.

", - "GetSendStatistics": "

Provides sending statistics for the Amazon SES account. The result is a list of data points, representing the last two weeks of sending activity. Each data point in the list contains statistics for a 15-minute period of time.

You can execute this operation no more than once per second.

", + "GetSendStatistics": "

Provides sending statistics for the current AWS Region. The result is a list of data points, representing the last two weeks of sending activity. Each data point in the list contains statistics for a 15-minute period of time.

You can execute this operation no more than once per second.

", "GetTemplate": "

Displays the template object (which includes the Subject line, HTML part and text part) for the template you specify.

You can execute this operation no more than once per second.

", - "ListConfigurationSets": "

Provides a list of the configuration sets associated with your Amazon SES account. For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.

You can execute this operation no more than once per second. This operation will return up to 1,000 configuration sets each time it is run. If your Amazon SES account has more than 1,000 configuration sets, this operation will also return a NextToken element. You can then execute the ListConfigurationSets operation again, passing the NextToken parameter and the value of the NextToken element to retrieve additional results.

", - "ListCustomVerificationEmailTemplates": "

Lists the existing custom verification email templates for your account.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "ListIdentities": "

Returns a list containing all of the identities (email addresses and domains) for your AWS account, regardless of verification status.

You can execute this operation no more than once per second.

", + "ListConfigurationSets": "

Provides a list of the configuration sets associated with your Amazon SES account in the current AWS Region. For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.

You can execute this operation no more than once per second. This operation will return up to 1,000 configuration sets each time it is run. If your Amazon SES account has more than 1,000 configuration sets, this operation will also return a NextToken element. You can then execute the ListConfigurationSets operation again, passing the NextToken parameter and the value of the NextToken element to retrieve additional results.

", + "ListCustomVerificationEmailTemplates": "

Lists the existing custom verification email templates for your account in the current AWS Region.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "ListIdentities": "

Returns a list containing all of the identities (email addresses and domains) for your AWS account in the current AWS Region, regardless of verification status.

You can execute this operation no more than once per second.

", "ListIdentityPolicies": "

Returns a list of sending authorization policies that are attached to the given identity (an email address or a domain). This API returns only a list. If you want the actual policy content, you can use GetIdentityPolicies.

This API is for the identity owner only. If you have not verified the identity, this API will return an error.

Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "ListReceiptFilters": "

Lists the IP address filters associated with your AWS account.

For information about managing IP address filters, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "ListReceiptRuleSets": "

Lists the receipt rule sets that exist under your AWS account. If there are additional receipt rule sets to be retrieved, you will receive a NextToken that you can provide to the next call to ListReceiptRuleSets to retrieve the additional entries.

For information about managing receipt rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", - "ListTemplates": "

Lists the email templates present in your Amazon SES account.

You can execute this operation no more than once per second.

", + "ListReceiptFilters": "

Lists the IP address filters associated with your AWS account in the current AWS Region.

For information about managing IP address filters, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "ListReceiptRuleSets": "

Lists the receipt rule sets that exist under your AWS account in the current AWS Region. If there are additional receipt rule sets to be retrieved, you will receive a NextToken that you can provide to the next call to ListReceiptRuleSets to retrieve the additional entries.

For information about managing receipt rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "ListTemplates": "

Lists the email templates present in your Amazon SES account in the current AWS Region.

You can execute this operation no more than once per second.

", "ListVerifiedEmailAddresses": "

Deprecated. Use the ListIdentities operation to list the email addresses and domains associated with your account.

", "PutIdentityPolicy": "

Adds or updates a sending authorization policy for the specified identity (an email address or a domain).

This API is for the identity owner only. If you have not verified the identity, this API will return an error.

Sending authorization is a feature that enables an identity owner to authorize other senders to use its identities. For information about using sending authorization, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "ReorderReceiptRuleSet": "

Reorders the receipt rules within a receipt rule set.

All of the rules in the rule set must be represented in this request. That is, this API will return an error if the reorder request doesn't explicitly position all of the rules.

For information about managing receipt rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "SendBounce": "

Generates and sends a bounce message to the sender of an email you received through Amazon SES. You can only use this API on an email up to 24 hours after you receive it.

You cannot use this API to send generic bounces for mail that was not received by Amazon SES.

For information about receiving email through Amazon SES, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "SendBulkTemplatedEmail": "

Composes an email message to multiple destinations. The message body is created using an email template.

In order to send email using the SendBulkTemplatedEmail operation, your call to the API must meet the following requirements:

  • The call must refer to an existing email template. You can create email templates using the CreateTemplate operation.

  • The message must be sent from a verified email address or domain.

  • If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

  • The total size of the message, including attachments, must be less than 10 MB.

  • Each Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.

", - "SendCustomVerificationEmail": "

Adds an email address to the list of identities for your Amazon SES account and attempts to verify it. As a result of executing this operation, a customized verification email is sent to the specified address.

To use this operation, you must first create a custom verification email template. For more information about creating and using custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "SendCustomVerificationEmail": "

Adds an email address to the list of identities for your Amazon SES account in the current AWS Region and attempts to verify it. As a result of executing this operation, a customized verification email is sent to the specified address.

To use this operation, you must first create a custom verification email template. For more information about creating and using custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "SendEmail": "

Composes an email message and immediately queues it for sending. In order to send email using the SendEmail operation, your message must meet the following requirements:

  • The message must be sent from a verified email address or domain. If you attempt to send email using a non-verified address or domain, the operation will result in an \"Email address not verified\" error.

  • If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

  • The total size of the message, including attachments, must be smaller than 10 MB.

  • The message must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.

  • The message may not include more than 50 recipients, across the To:, CC: and BCC: fields. If you need to send an email message to a larger audience, you can divide your recipient list into groups of 50 or fewer, and then call the SendEmail operation several times to send the message to each group.

For every message that you send, the total number of recipients (including each recipient in the To:, CC: and BCC: fields) is counted against the maximum number of emails you can send in a 24-hour period (your sending quota). For more information about sending quotas in Amazon SES, see Managing Your Amazon SES Sending Limits in the Amazon SES Developer Guide.

", "SendRawEmail": "

Composes an email message and immediately queues it for sending. When calling this operation, you may specify the message headers as well as the content. The SendRawEmail operation is particularly useful for sending multipart MIME emails (such as those that contain both a plain-text and an HTML version).

In order to send email using the SendRawEmail operation, your message must meet the following requirements:

  • The message must be sent from a verified email address or domain. If you attempt to send email using a non-verified address or domain, the operation will result in an \"Email address not verified\" error.

  • If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

  • The total size of the message, including attachments, must be smaller than 10 MB.

  • The message must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.

  • The message may not include more than 50 recipients, across the To:, CC: and BCC: fields. If you need to send an email message to a larger audience, you can divide your recipient list into groups of 50 or fewer, and then call the SendRawEmail operation several times to send the message to each group.

For every message that you send, the total number of recipients (including each recipient in the To:, CC: and BCC: fields) is counted against the maximum number of emails you can send in a 24-hour period (your sending quota). For more information about sending quotas in Amazon SES, see Managing Your Amazon SES Sending Limits in the Amazon SES Developer Guide.

Additionally, keep the following considerations in mind when using the SendRawEmail operation:

  • Although you can customize the message headers when using the SendRawEmail operation, Amazon SES will automatically apply its own Message-ID and Date headers; if you passed these headers when creating the message, they will be overwritten by the values that Amazon SES provides.

  • If you are using sending authorization to send on behalf of another user, SendRawEmail enables you to specify the cross-account identity for the email's Source, From, and Return-Path parameters in one of two ways: you can pass optional parameters SourceArn, FromArn, and/or ReturnPathArn to the API, or you can include the following X-headers in the header of your raw email:

    • X-SES-SOURCE-ARN

    • X-SES-FROM-ARN

    • X-SES-RETURN-PATH-ARN

    Do not include these X-headers in the DKIM signature; Amazon SES will remove them before sending the email.

    For most common sending authorization scenarios, we recommend that you specify the SourceIdentityArn parameter and not the FromIdentityArn or ReturnPathIdentityArn parameters. If you only specify the SourceIdentityArn parameter, Amazon SES will set the From and Return Path addresses to the identity specified in SourceIdentityArn. For more information about sending authorization, see the Using Sending Authorization with Amazon SES in the Amazon SES Developer Guide.

", - "SendTemplatedEmail": "

Composes an email message using an email template and immediately queues it for sending.

In order to send email using the SendTemplatedEmail operation, your call to the API must meet the following requirements:

  • The call must refer to an existing email template. You can create email templates using the CreateTemplate operation.

  • The message must be sent from a verified email address or domain.

  • If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

  • The total size of the message, including attachments, must be less than 10 MB.

  • Calls to the SendTemplatedEmail operation may only include one Destination parameter. A destination is a set of recipients who will receive the same version of the email. The Destination parameter can include up to 50 recipients, across the To:, CC: and BCC: fields.

  • The Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.

", + "SendTemplatedEmail": "

Composes an email message using an email template and immediately queues it for sending.

In order to send email using the SendTemplatedEmail operation, your call to the API must meet the following requirements:

  • The call must refer to an existing email template. You can create email templates using the CreateTemplate operation.

  • The message must be sent from a verified email address or domain.

  • If your account is still in the Amazon SES sandbox, you may only send to verified addresses or domains, or to email addresses associated with the Amazon SES Mailbox Simulator. For more information, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

  • The total size of the message, including attachments, must be less than 10 MB.

  • Calls to the SendTemplatedEmail operation may only include one Destination parameter. A destination is a set of recipients who will receive the same version of the email. The Destination parameter can include up to 50 recipients, across the To:, CC: and BCC: fields.

  • The Destination parameter must include at least one recipient email address. The recipient address can be a To: address, a CC: address, or a BCC: address. If a recipient email address is invalid (that is, it is not in the format UserName@[SubDomain.]Domain.TopLevelDomain), the entire message will be rejected, even if the message contains other recipients that are valid.

If your call to the SendTemplatedEmail operation includes all of the required parameters, Amazon SES accepts it and returns a Message ID. However, if Amazon SES can't render the email because the template contains errors, it doesn't send the email. Additionally, because it already accepted the message, Amazon SES doesn't return a message stating that it was unable to send the email.

For these reasons, we highly recommend that you set up Amazon SES to send you notifications when Rendering Failure events occur. For more information, see Sending Personalized Email Using the Amazon SES API in the Amazon Simple Email Service Developer Guide.

", "SetActiveReceiptRuleSet": "

Sets the specified receipt rule set as the active receipt rule set.

To disable your email-receiving through Amazon SES completely, you can call this API with RuleSetName set to null.

For information about managing receipt rule sets, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "SetIdentityDkimEnabled": "

Enables or disables Easy DKIM signing of email sent from an identity:

  • If Easy DKIM signing is enabled for a domain name identity (such as example.com), then Amazon SES will DKIM-sign all email sent by addresses under that domain name (for example, user@example.com).

  • If Easy DKIM signing is enabled for an email address, then Amazon SES will DKIM-sign all email sent by that email address.

For email addresses (for example, user@example.com), you can only enable Easy DKIM signing if the corresponding domain (in this case, example.com) has been set up for Easy DKIM using the AWS Console or the VerifyDomainDkim operation.

You can execute this operation no more than once per second.

For more information about Easy DKIM signing, go to the Amazon SES Developer Guide.

", "SetIdentityFeedbackForwardingEnabled": "

Given an identity (an email address or a domain), enables or disables whether Amazon SES forwards bounce and complaint notifications as email. Feedback forwarding can only be disabled when Amazon Simple Notification Service (Amazon SNS) topics are specified for both bounces and complaints.

Feedback forwarding does not apply to delivery notifications. Delivery notifications are only available through Amazon SNS.

You can execute this operation no more than once per second.

For more information about using notifications with Amazon SES, see the Amazon SES Developer Guide.

", @@ -60,18 +60,18 @@ "SetIdentityNotificationTopic": "

Given an identity (an email address or a domain), sets the Amazon Simple Notification Service (Amazon SNS) topic to which Amazon SES will publish bounce, complaint, and/or delivery notifications for emails sent with that identity as the Source.

Unless feedback forwarding is enabled, you must specify Amazon SNS topics for bounce and complaint notifications. For more information, see SetIdentityFeedbackForwardingEnabled.

You can execute this operation no more than once per second.

For more information about feedback notification, see the Amazon SES Developer Guide.

", "SetReceiptRulePosition": "

Sets the position of the specified receipt rule in the receipt rule set.

For information about managing receipt rules, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "TestRenderTemplate": "

Creates a preview of the MIME content of an email when provided with a template and a set of replacement data.

You can execute this operation no more than once per second.

", - "UpdateAccountSendingEnabled": "

Enables or disables email sending across your entire Amazon SES account. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending across your Amazon SES account when reputation metrics (such as your bounce on complaint rate) reach certain thresholds.

You can execute this operation no more than once per second.

", + "UpdateAccountSendingEnabled": "

Enables or disables email sending across your entire Amazon SES account in the current AWS Region. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending across your Amazon SES account in a given AWS Region when reputation metrics (such as your bounce or complaint rates) reach certain thresholds.

You can execute this operation no more than once per second.

", "UpdateConfigurationSetEventDestination": "

Updates the event destination of a configuration set. Event destinations are associated with configuration sets, which enable you to publish email sending events to Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS). For information about using configuration sets, see Monitoring Your Amazon SES Sending Activity in the Amazon SES Developer Guide.

When you create or update an event destination, you must provide one, and only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis Firehose, or Amazon Simple Notification Service (Amazon SNS).

You can execute this operation no more than once per second.

", - "UpdateConfigurationSetReputationMetricsEnabled": "

Enables or disables the publishing of reputation metrics for emails sent using a specific configuration set. Reputation metrics include bounce and complaint rates. These metrics are published to Amazon CloudWatch. By using Amazon CloudWatch, you can create alarms when bounce or complaint rates exceed a certain threshold.

You can execute this operation no more than once per second.

", - "UpdateConfigurationSetSendingEnabled": "

Enables or disables email sending for messages sent using a specific configuration set. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending for a configuration set when the reputation metrics for that configuration set (such as your bounce on complaint rate) reach certain thresholds.

You can execute this operation no more than once per second.

", - "UpdateConfigurationSetTrackingOptions": "

Modifies an association between a configuration set and a custom domain for open and click event tracking.

By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using configuration sets, see Configuring Custom Domains to Handle Open and Click Tracking in the Amazon SES Developer Guide.

", - "UpdateCustomVerificationEmailTemplate": "

Updates an existing custom verification email template.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "UpdateConfigurationSetReputationMetricsEnabled": "

Enables or disables the publishing of reputation metrics for emails sent using a specific configuration set in a given AWS Region. Reputation metrics include bounce and complaint rates. These metrics are published to Amazon CloudWatch. By using CloudWatch, you can create alarms when bounce or complaint rates exceed certain thresholds.

You can execute this operation no more than once per second.

", + "UpdateConfigurationSetSendingEnabled": "

Enables or disables email sending for messages sent using a specific configuration set in a given AWS Region. You can use this operation in conjunction with Amazon CloudWatch alarms to temporarily pause email sending for a configuration set when the reputation metrics for that configuration set (such as your bounce on complaint rate) exceed certain thresholds.

You can execute this operation no more than once per second.

", + "UpdateConfigurationSetTrackingOptions": "

Modifies an association between a configuration set and a custom domain for open and click event tracking.

By default, images and links used for tracking open and click events are hosted on domains operated by Amazon SES. You can configure a subdomain of your own to handle these events. For information about using custom domains, see the Amazon SES Developer Guide.

", + "UpdateCustomVerificationEmailTemplate": "

Updates an existing custom verification email template.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "UpdateReceiptRule": "

Updates a receipt rule.

For information about managing receipt rules, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "UpdateTemplate": "

Updates an email template. Email templates enable you to send personalized email to one or more destinations in a single API operation. For more information, see the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "VerifyDomainDkim": "

Returns a set of DKIM tokens for a domain. DKIM tokens are character strings that represent your domain's identity. Using these tokens, you will need to create DNS CNAME records that point to DKIM public keys hosted by Amazon SES. Amazon Web Services will eventually detect that you have updated your DNS records; this detection process may take up to 72 hours. Upon successful detection, Amazon SES will be able to DKIM-sign email originating from that domain.

You can execute this operation no more than once per second.

To enable or disable Easy DKIM signing for a domain, use the SetIdentityDkimEnabled operation.

For more information about creating DNS records using DKIM tokens, go to the Amazon SES Developer Guide.

", - "VerifyDomainIdentity": "

Adds a domain to the list of identities for your Amazon SES account and attempts to verify it. For more information about verifying domains, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", + "VerifyDomainIdentity": "

Adds a domain to the list of identities for your Amazon SES account in the current AWS Region and attempts to verify it. For more information about verifying domains, see Verifying Email Addresses and Domains in the Amazon SES Developer Guide.

You can execute this operation no more than once per second.

", "VerifyEmailAddress": "

Deprecated. Use the VerifyEmailIdentity operation to verify a new email address.

", - "VerifyEmailIdentity": "

Adds an email address to the list of identities for your Amazon SES account and attempts to verify it. As a result of executing this operation, a verification email is sent to the specified address.

You can execute this operation no more than once per second.

" + "VerifyEmailIdentity": "

Adds an email address to the list of identities for your Amazon SES account in the current AWS region and attempts to verify it. As a result of executing this operation, a verification email is sent to the specified address.

You can execute this operation no more than once per second.

" }, "shapes": { "AccountSendingPausedException": { @@ -136,7 +136,7 @@ "LambdaAction$TopicArn": "

The Amazon Resource Name (ARN) of the Amazon SNS topic to notify when the Lambda action is taken. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.

", "LambdaAction$FunctionArn": "

The Amazon Resource Name (ARN) of the AWS Lambda function. An example of an AWS Lambda function ARN is arn:aws:lambda:us-west-2:account-id:function:MyFunction. For more information about AWS Lambda, see the AWS Lambda Developer Guide.

", "S3Action$TopicArn": "

The ARN of the Amazon SNS topic to notify when the message is saved to the Amazon S3 bucket. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.

", - "S3Action$KmsKeyArn": "

The customer master key that Amazon SES should use to encrypt your emails before saving them to the Amazon S3 bucket. You can use the default master key or a custom master key you created in AWS KMS as follows:

  • To use the default master key, provide an ARN in the form of arn:aws:kms:REGION:ACCOUNT-ID-WITHOUT-HYPHENS:alias/aws/ses. For example, if your AWS account ID is 123456789012 and you want to use the default master key in the US West (Oregon) region, the ARN of the default master key would be arn:aws:kms:us-west-2:123456789012:alias/aws/ses. If you use the default master key, you don't need to perform any extra steps to give Amazon SES permission to use the key.

  • To use a custom master key you created in AWS KMS, provide the ARN of the master key and ensure that you add a statement to your key's policy to give Amazon SES permission to use it. For more information about giving permissions, see the Amazon SES Developer Guide.

For more information about key policies, see the AWS KMS Developer Guide. If you do not specify a master key, Amazon SES will not encrypt your emails.

Your mail is encrypted by Amazon SES using the Amazon S3 encryption client before the mail is submitted to Amazon S3 for storage. It is not encrypted using Amazon S3 server-side encryption. This means that you must use the Amazon S3 encryption client to decrypt the email after retrieving it from Amazon S3, as the service has no access to use your AWS KMS keys for decryption. This encryption client is currently available with the AWS Java SDK and AWS Ruby SDK only. For more information about client-side encryption using AWS KMS master keys, see the Amazon S3 Developer Guide.

", + "S3Action$KmsKeyArn": "

The customer master key that Amazon SES should use to encrypt your emails before saving them to the Amazon S3 bucket. You can use the default master key or a custom master key you created in AWS KMS as follows:

  • To use the default master key, provide an ARN in the form of arn:aws:kms:REGION:ACCOUNT-ID-WITHOUT-HYPHENS:alias/aws/ses. For example, if your AWS account ID is 123456789012 and you want to use the default master key in the US West (Oregon) region, the ARN of the default master key would be arn:aws:kms:us-west-2:123456789012:alias/aws/ses. If you use the default master key, you don't need to perform any extra steps to give Amazon SES permission to use the key.

  • To use a custom master key you created in AWS KMS, provide the ARN of the master key and ensure that you add a statement to your key's policy to give Amazon SES permission to use it. For more information about giving permissions, see the Amazon SES Developer Guide.

For more information about key policies, see the AWS KMS Developer Guide. If you do not specify a master key, Amazon SES will not encrypt your emails.

Your mail is encrypted by Amazon SES using the Amazon S3 encryption client before the mail is submitted to Amazon S3 for storage. It is not encrypted using Amazon S3 server-side encryption. This means that you must use the Amazon S3 encryption client to decrypt the email after retrieving it from Amazon S3, as the service has no access to use your AWS KMS keys for decryption. This encryption client is currently available with the AWS SDK for Java and AWS SDK for Ruby only. For more information about client-side encryption using AWS KMS master keys, see the Amazon S3 Developer Guide.

", "SNSAction$TopicArn": "

The Amazon Resource Name (ARN) of the Amazon SNS topic to notify. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.

", "SNSDestination$TopicARN": "

The ARN of the Amazon SNS topic that email sending events will be published to. An example of an Amazon SNS topic ARN is arn:aws:sns:us-west-2:123456789012:MyTopic. For more information about Amazon SNS topics, see the Amazon SNS Developer Guide.

", "SendBounceRequest$BounceSenderArn": "

This parameter is used only for sending authorization. It is the ARN of the identity that is associated with the sending authorization policy that permits you to use the address in the \"From\" header of the bounce. For more information about sending authorization, see the Amazon SES Developer Guide.

", @@ -696,7 +696,7 @@ "base": null, "refs": { "EventDestination$Enabled": "

Sets whether Amazon SES publishes events to this destination when you send an email with the associated configuration set. Set to true to enable publishing to this destination; set to false to prevent publishing to this destination. The default value is false.

", - "GetAccountSendingEnabledResponse$Enabled": "

Describes whether email sending is enabled or disabled for your Amazon SES account.

", + "GetAccountSendingEnabledResponse$Enabled": "

Describes whether email sending is enabled or disabled for your Amazon SES account in the current AWS Region.

", "IdentityDkimAttributes$DkimEnabled": "

True if DKIM signing is enabled for email sent from the identity; false otherwise. The default value is true.

", "IdentityNotificationAttributes$ForwardingEnabled": "

Describes whether Amazon SES will forward bounce and complaint notifications as email. true indicates that Amazon SES will forward bounce and complaint notifications as email, while false indicates that bounce and complaint notifications will be published only to the specified bounce and complaint Amazon SNS topics.

", "IdentityNotificationAttributes$HeadersInBounceNotificationsEnabled": "

Describes whether Amazon SES includes the original email headers in Amazon SNS notifications of type Bounce. A value of true specifies that Amazon SES will include headers in bounce notifications, and a value of false specifies that Amazon SES will not include headers in bounce notifications.

", @@ -709,7 +709,7 @@ "SetIdentityDkimEnabledRequest$DkimEnabled": "

Sets whether DKIM signing is enabled for an identity. Set to true to enable DKIM signing for this identity; false to disable it.

", "SetIdentityFeedbackForwardingEnabledRequest$ForwardingEnabled": "

Sets whether Amazon SES will forward bounce and complaint notifications as email. true specifies that Amazon SES will forward bounce and complaint notifications as email, in addition to any Amazon SNS topic publishing otherwise specified. false specifies that Amazon SES will publish bounce and complaint notifications only through Amazon SNS. This value can only be set to false when Amazon SNS topics are set for both Bounce and Complaint notification types.

", "SetIdentityHeadersInNotificationsEnabledRequest$Enabled": "

Sets whether Amazon SES includes the original email headers in Amazon SNS notifications of the specified notification type. A value of true specifies that Amazon SES will include headers in notifications, and a value of false specifies that Amazon SES will not include headers in notifications.

This value can only be set when NotificationType is already set to use a particular Amazon SNS topic.

", - "UpdateAccountSendingEnabledRequest$Enabled": "

Describes whether email sending is enabled or disabled for your Amazon SES account.

", + "UpdateAccountSendingEnabledRequest$Enabled": "

Describes whether email sending is enabled or disabled for your Amazon SES account in the current AWS Region.

", "UpdateConfigurationSetReputationMetricsEnabledRequest$Enabled": "

Describes whether or not Amazon SES will publish reputation metrics for the configuration set, such as bounce and complaint rates, to Amazon CloudWatch.

", "UpdateConfigurationSetSendingEnabledRequest$Enabled": "

Describes whether email sending is enabled or disabled for the configuration set.

" } @@ -824,7 +824,7 @@ } }, "GetAccountSendingEnabledResponse": { - "base": "

Represents a request to return the email sending status for your Amazon SES account.

", + "base": "

Represents a request to return the email sending status for your Amazon SES account in the current AWS Region.

", "refs": { } }, @@ -1032,7 +1032,7 @@ } }, "InvalidTemplateException": { - "base": "

Indicates that a template could not be created because it contained invalid JSON.

", + "base": "

Indicates that the template that you specified could not be rendered. This issue may occur when a template refers to a partial that does not exist.

", "refs": { } }, @@ -1087,7 +1087,7 @@ } }, "ListCustomVerificationEmailTemplatesRequest": { - "base": "

Represents a request to list the existing custom verification email templates for your account.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

", + "base": "

Represents a request to list the existing custom verification email templates for your account.

For more information about custom verification email templates, see Using Custom Verification Email Templates in the Amazon SES Developer Guide.

", "refs": { } }, @@ -1856,7 +1856,7 @@ } }, "TrackingOptions": { - "base": "

A domain that is used to redirect email recipients to an Amazon SES-operated domain. This domain captures open and click events generated by Amazon SES emails.

For more information, see Configuring Custom Domains to Handle Open and Click Tracking in the Amazon SES Developer Guide.

", + "base": "

A domain that is used to redirect email recipients to an Amazon SES-operated domain. This domain captures open and click events generated by Amazon SES emails.

For more information, see Configuring Custom Domains to Handle Open and Click Tracking in the Amazon SES Developer Guide.

", "refs": { "CreateConfigurationSetTrackingOptionsRequest$TrackingOptions": null, "DescribeConfigurationSetResponse$TrackingOptions": "

The name of the custom open and click tracking domain associated with the configuration set.

", @@ -1999,7 +1999,7 @@ "WorkmailAction": { "base": "

When included in a receipt rule, this action calls Amazon WorkMail and, optionally, publishes a notification to Amazon Simple Notification Service (Amazon SNS). You will typically not use this action directly because Amazon WorkMail adds the rule automatically during its setup procedure.

For information using a receipt rule to call Amazon WorkMail, see the Amazon SES Developer Guide.

", "refs": { - "ReceiptAction$WorkmailAction": "

Calls Amazon WorkMail and, optionally, publishes a notification to Amazon SNS.

" + "ReceiptAction$WorkmailAction": "

Calls Amazon WorkMail and, optionally, publishes a notification to Amazon Amazon SNS.

" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/smoke.json b/vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/smoke.json similarity index 63% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/smoke.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/smoke.json index 068b23492..deaee0181 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/smoke.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/email/2010-12-01/smoke.json @@ -3,14 +3,14 @@ "defaultRegion": "us-west-2", "testCases": [ { - "operationName": "DescribeDBEngineVersions", + "operationName": "ListIdentities", "input": {}, "errorExpectedFromService": false }, { - "operationName": "DescribeDBInstances", + "operationName": "VerifyEmailIdentity", "input": { - "DBInstanceIdentifier": "fake-id" + "EmailAddress": "fake_email" }, "errorExpectedFromService": true } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/api-2.json deleted file mode 100644 index c9980d9f5..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/api-2.json +++ /dev/null @@ -1,643 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2014-02-03", - "endpointPrefix":"events", - "jsonVersion":"1.1", - "serviceFullName":"Amazon CloudWatch Events", - "signatureVersion":"v4", - "targetPrefix":"AWSEvents", - "protocol":"json" - }, - "operations":{ - "DeleteRule":{ - "name":"DeleteRule", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRuleRequest"}, - "errors":[ - { - "shape":"ConcurrentModificationException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "DescribeRule":{ - "name":"DescribeRule", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRuleRequest"}, - "output":{"shape":"DescribeRuleResponse"}, - "errors":[ - { - "shape":"ResourceNotFoundException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "DisableRule":{ - "name":"DisableRule", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DisableRuleRequest"}, - "errors":[ - { - "shape":"ResourceNotFoundException", - "exception":true - }, - { - "shape":"ConcurrentModificationException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "EnableRule":{ - "name":"EnableRule", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"EnableRuleRequest"}, - "errors":[ - { - "shape":"ResourceNotFoundException", - "exception":true - }, - { - "shape":"ConcurrentModificationException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "ListRuleNamesByTarget":{ - "name":"ListRuleNamesByTarget", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListRuleNamesByTargetRequest"}, - "output":{"shape":"ListRuleNamesByTargetResponse"}, - "errors":[ - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "ListRules":{ - "name":"ListRules", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListRulesRequest"}, - "output":{"shape":"ListRulesResponse"}, - "errors":[ - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "ListTargetsByRule":{ - "name":"ListTargetsByRule", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListTargetsByRuleRequest"}, - "output":{"shape":"ListTargetsByRuleResponse"}, - "errors":[ - { - "shape":"ResourceNotFoundException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "PutEvents":{ - "name":"PutEvents", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PutEventsRequest"}, - "output":{"shape":"PutEventsResponse"}, - "errors":[ - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "PutRule":{ - "name":"PutRule", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PutRuleRequest"}, - "output":{"shape":"PutRuleResponse"}, - "errors":[ - { - "shape":"InvalidEventPatternException", - "exception":true - }, - { - "shape":"LimitExceededException", - "exception":true - }, - { - "shape":"ConcurrentModificationException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "PutTargets":{ - "name":"PutTargets", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PutTargetsRequest"}, - "output":{"shape":"PutTargetsResponse"}, - "errors":[ - { - "shape":"ResourceNotFoundException", - "exception":true - }, - { - "shape":"ConcurrentModificationException", - "exception":true - }, - { - "shape":"LimitExceededException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "RemoveTargets":{ - "name":"RemoveTargets", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveTargetsRequest"}, - "output":{"shape":"RemoveTargetsResponse"}, - "errors":[ - { - "shape":"ResourceNotFoundException", - "exception":true - }, - { - "shape":"ConcurrentModificationException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - }, - "TestEventPattern":{ - "name":"TestEventPattern", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"TestEventPatternRequest"}, - "output":{"shape":"TestEventPatternResponse"}, - "errors":[ - { - "shape":"InvalidEventPatternException", - "exception":true - }, - { - "shape":"InternalException", - "exception":true, - "fault":true - } - ] - } - }, - "shapes":{ - "Boolean":{"type":"boolean"}, - "ConcurrentModificationException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "DeleteRuleRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "Name":{"shape":"RuleName"} - } - }, - "DescribeRuleRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "Name":{"shape":"RuleName"} - } - }, - "DescribeRuleResponse":{ - "type":"structure", - "members":{ - "Name":{"shape":"RuleName"}, - "Arn":{"shape":"RuleArn"}, - "EventPattern":{"shape":"EventPattern"}, - "ScheduleExpression":{"shape":"ScheduleExpression"}, - "State":{"shape":"RuleState"}, - "Description":{"shape":"RuleDescription"}, - "RoleArn":{"shape":"RoleArn"} - } - }, - "DisableRuleRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "Name":{"shape":"RuleName"} - } - }, - "EnableRuleRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "Name":{"shape":"RuleName"} - } - }, - "ErrorCode":{"type":"string"}, - "ErrorMessage":{"type":"string"}, - "EventId":{"type":"string"}, - "EventPattern":{ - "type":"string", - "max":2048 - }, - "EventResource":{"type":"string"}, - "EventResourceList":{ - "type":"list", - "member":{"shape":"EventResource"} - }, - "EventTime":{"type":"timestamp"}, - "Integer":{"type":"integer"}, - "InternalException":{ - "type":"structure", - "members":{ - }, - "exception":true, - "fault":true - }, - "InvalidEventPatternException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "LimitExceededException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "LimitMax100":{ - "type":"integer", - "min":1, - "max":100 - }, - "ListRuleNamesByTargetRequest":{ - "type":"structure", - "required":["TargetArn"], - "members":{ - "TargetArn":{"shape":"TargetArn"}, - "NextToken":{"shape":"NextToken"}, - "Limit":{"shape":"LimitMax100"} - } - }, - "ListRuleNamesByTargetResponse":{ - "type":"structure", - "members":{ - "RuleNames":{"shape":"RuleNameList"}, - "NextToken":{"shape":"NextToken"} - } - }, - "ListRulesRequest":{ - "type":"structure", - "members":{ - "NamePrefix":{"shape":"RuleName"}, - "NextToken":{"shape":"NextToken"}, - "Limit":{"shape":"LimitMax100"} - } - }, - "ListRulesResponse":{ - "type":"structure", - "members":{ - "Rules":{"shape":"RuleResponseList"}, - "NextToken":{"shape":"NextToken"} - } - }, - "ListTargetsByRuleRequest":{ - "type":"structure", - "required":["Rule"], - "members":{ - "Rule":{"shape":"RuleName"}, - "NextToken":{"shape":"NextToken"}, - "Limit":{"shape":"LimitMax100"} - } - }, - "ListTargetsByRuleResponse":{ - "type":"structure", - "members":{ - "Targets":{"shape":"TargetList"}, - "NextToken":{"shape":"NextToken"} - } - }, - "NextToken":{ - "type":"string", - "min":1, - "max":2048 - }, - "PutEventsRequest":{ - "type":"structure", - "required":["Entries"], - "members":{ - "Entries":{"shape":"PutEventsRequestEntryList"} - } - }, - "PutEventsRequestEntry":{ - "type":"structure", - "members":{ - "Time":{"shape":"EventTime"}, - "Source":{"shape":"String"}, - "Resources":{"shape":"EventResourceList"}, - "DetailType":{"shape":"String"}, - "Detail":{"shape":"String"} - } - }, - "PutEventsRequestEntryList":{ - "type":"list", - "member":{"shape":"PutEventsRequestEntry"}, - "min":1, - "max":10 - }, - "PutEventsResponse":{ - "type":"structure", - "members":{ - "FailedEntryCount":{"shape":"Integer"}, - "Entries":{"shape":"PutEventsResultEntryList"} - } - }, - "PutEventsResultEntry":{ - "type":"structure", - "members":{ - "EventId":{"shape":"EventId"}, - "ErrorCode":{"shape":"ErrorCode"}, - "ErrorMessage":{"shape":"ErrorMessage"} - } - }, - "PutEventsResultEntryList":{ - "type":"list", - "member":{"shape":"PutEventsResultEntry"} - }, - "PutRuleRequest":{ - "type":"structure", - "required":["Name"], - "members":{ - "Name":{"shape":"RuleName"}, - "ScheduleExpression":{"shape":"ScheduleExpression"}, - "EventPattern":{"shape":"EventPattern"}, - "State":{"shape":"RuleState"}, - "Description":{"shape":"RuleDescription"}, - "RoleArn":{"shape":"RoleArn"} - } - }, - "PutRuleResponse":{ - "type":"structure", - "members":{ - "RuleArn":{"shape":"RuleArn"} - } - }, - "PutTargetsRequest":{ - "type":"structure", - "required":[ - "Rule", - "Targets" - ], - "members":{ - "Rule":{"shape":"RuleName"}, - "Targets":{"shape":"TargetList"} - } - }, - "PutTargetsResponse":{ - "type":"structure", - "members":{ - "FailedEntryCount":{"shape":"Integer"}, - "FailedEntries":{"shape":"PutTargetsResultEntryList"} - } - }, - "PutTargetsResultEntry":{ - "type":"structure", - "members":{ - "TargetId":{"shape":"TargetId"}, - "ErrorCode":{"shape":"ErrorCode"}, - "ErrorMessage":{"shape":"ErrorMessage"} - } - }, - "PutTargetsResultEntryList":{ - "type":"list", - "member":{"shape":"PutTargetsResultEntry"} - }, - "RemoveTargetsRequest":{ - "type":"structure", - "required":[ - "Rule", - "Ids" - ], - "members":{ - "Rule":{"shape":"RuleName"}, - "Ids":{"shape":"TargetIdList"} - } - }, - "RemoveTargetsResponse":{ - "type":"structure", - "members":{ - "FailedEntryCount":{"shape":"Integer"}, - "FailedEntries":{"shape":"RemoveTargetsResultEntryList"} - } - }, - "RemoveTargetsResultEntry":{ - "type":"structure", - "members":{ - "TargetId":{"shape":"TargetId"}, - "ErrorCode":{"shape":"ErrorCode"}, - "ErrorMessage":{"shape":"ErrorMessage"} - } - }, - "RemoveTargetsResultEntryList":{ - "type":"list", - "member":{"shape":"RemoveTargetsResultEntry"} - }, - "ResourceNotFoundException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "RoleArn":{ - "type":"string", - "min":1, - "max":1600 - }, - "Rule":{ - "type":"structure", - "members":{ - "Name":{"shape":"RuleName"}, - "Arn":{"shape":"RuleArn"}, - "EventPattern":{"shape":"EventPattern"}, - "State":{"shape":"RuleState"}, - "Description":{"shape":"RuleDescription"}, - "ScheduleExpression":{"shape":"ScheduleExpression"}, - "RoleArn":{"shape":"RoleArn"} - } - }, - "RuleArn":{ - "type":"string", - "min":1, - "max":1600 - }, - "RuleDescription":{ - "type":"string", - "max":512 - }, - "RuleName":{ - "type":"string", - "min":1, - "max":64, - "pattern":"[\\.\\-_A-Za-z0-9]+" - }, - "RuleNameList":{ - "type":"list", - "member":{"shape":"RuleName"} - }, - "RuleResponseList":{ - "type":"list", - "member":{"shape":"Rule"} - }, - "RuleState":{ - "type":"string", - "enum":[ - "ENABLED", - "DISABLED" - ] - }, - "ScheduleExpression":{ - "type":"string", - "max":256 - }, - "String":{"type":"string"}, - "Target":{ - "type":"structure", - "required":[ - "Id", - "Arn" - ], - "members":{ - "Id":{"shape":"TargetId"}, - "Arn":{"shape":"TargetArn"}, - "Input":{"shape":"TargetInput"}, - "InputPath":{"shape":"TargetInputPath"} - } - }, - "TargetArn":{ - "type":"string", - "min":1, - "max":1600 - }, - "TargetId":{ - "type":"string", - "min":1, - "max":64, - "pattern":"[\\.\\-_A-Za-z0-9]+" - }, - "TargetIdList":{ - "type":"list", - "member":{"shape":"TargetId"}, - "min":1, - "max":100 - }, - "TargetInput":{ - "type":"string", - "max":8192 - }, - "TargetInputPath":{ - "type":"string", - "max":256 - }, - "TargetList":{ - "type":"list", - "member":{"shape":"Target"} - }, - "TestEventPatternRequest":{ - "type":"structure", - "required":[ - "EventPattern", - "Event" - ], - "members":{ - "EventPattern":{"shape":"EventPattern"}, - "Event":{"shape":"String"} - } - }, - "TestEventPatternResponse":{ - "type":"structure", - "members":{ - "Result":{"shape":"Boolean"} - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/docs-2.json deleted file mode 100644 index c6eaf8627..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/docs-2.json +++ /dev/null @@ -1,411 +0,0 @@ -{ - "version": "2.0", - "operations": { - "DeleteRule": "

Deletes a rule. You must remove all targets from a rule using RemoveTargets before you can delete the rule.

Note: When you make a change with this action, incoming events might still continue to match to the deleted rule. Please allow a short period of time for changes to take effect.

", - "DescribeRule": "

Describes the details of the specified rule.

", - "DisableRule": "

Disables a rule. A disabled rule won't match any events, and won't self-trigger if it has a schedule expression.

Note: When you make a change with this action, incoming events might still continue to match to the disabled rule. Please allow a short period of time for changes to take effect.

", - "EnableRule": "

Enables a rule. If the rule does not exist, the operation fails.

Note: When you make a change with this action, incoming events might not immediately start matching to a newly enabled rule. Please allow a short period of time for changes to take effect.

", - "ListRuleNamesByTarget": "

Lists the names of the rules that the given target is put to. Using this action, you can find out which of the rules in Amazon CloudWatch Events can invoke a specific target in your account. If you have more rules in your account than the given limit, the results will be paginated. In that case, use the next token returned in the response and repeat the ListRulesByTarget action until the NextToken in the response is returned as null.

", - "ListRules": "

Lists the Amazon CloudWatch Events rules in your account. You can either list all the rules or you can provide a prefix to match to the rule names. If you have more rules in your account than the given limit, the results will be paginated. In that case, use the next token returned in the response and repeat the ListRules action until the NextToken in the response is returned as null.

", - "ListTargetsByRule": "

Lists of targets assigned to the rule.

", - "PutEvents": "

Sends custom events to Amazon CloudWatch Events so that they can be matched to rules.

", - "PutRule": "

Creates or updates a rule. Rules are enabled by default, or based on value of the State parameter. You can disable a rule using DisableRule.

Note: When you make a change with this action, incoming events might not immediately start matching to new or updated rules. Please allow a short period of time for changes to take effect.

A rule must contain at least an EventPattern or ScheduleExpression. Rules with EventPatterns are triggered when a matching event is observed. Rules with ScheduleExpressions self-trigger based on the given schedule. A rule can have both an EventPattern and a ScheduleExpression, in which case the rule will trigger on matching events as well as on a schedule.

Note: Most services in AWS treat : or / as the same character in Amazon Resource Names (ARNs). However, CloudWatch Events uses an exact match in event patterns and rules. Be sure to use the correct ARN characters when creating event patterns so that they match the ARN syntax in the event you want to match.

", - "PutTargets": "

Adds target(s) to a rule. Updates the target(s) if they are already associated with the role. In other words, if there is already a target with the given target ID, then the target associated with that ID is updated.

Note: When you make a change with this action, when the associated rule triggers, new or updated targets might not be immediately invoked. Please allow a short period of time for changes to take effect.

", - "RemoveTargets": "

Removes target(s) from a rule so that when the rule is triggered, those targets will no longer be invoked.

Note: When you make a change with this action, when the associated rule triggers, removed targets might still continue to be invoked. Please allow a short period of time for changes to take effect.

", - "TestEventPattern": "

Tests whether an event pattern matches the provided event.

Note: Most services in AWS treat : or / as the same character in Amazon Resource Names (ARNs). However, CloudWatch Events uses an exact match in event patterns and rules. Be sure to use the correct ARN characters when creating event patterns so that they match the ARN syntax in the event you want to match.

" - }, - "service": "

Amazon CloudWatch Events helps you to respond to state changes in your AWS resources. When your resources change state they automatically send events into an event stream. You can create rules that match selected events in the stream and route them to targets to take action. You can also use rules to take action on a pre-determined schedule. For example, you can configure rules to:

  • Automatically invoke an AWS Lambda function to update DNS entries when an event notifies you that Amazon EC2 instance enters the running state.
  • Direct specific API records from CloudTrail to an Amazon Kinesis stream for detailed analysis of potential security or availability risks.
  • Periodically invoke a built-in target to create a snapshot of an Amazon EBS volume.

For more information about Amazon CloudWatch Events features, see the Amazon CloudWatch Developer Guide.

", - "shapes": { - "Boolean": { - "base": null, - "refs": { - "TestEventPatternResponse$Result": "

Indicates whether the event matches the event pattern.

" - } - }, - "ConcurrentModificationException": { - "base": "

This exception occurs if there is concurrent modification on rule or target.

", - "refs": { - } - }, - "DeleteRuleRequest": { - "base": "

Container for the parameters to the DeleteRule operation.

", - "refs": { - } - }, - "DescribeRuleRequest": { - "base": "

Container for the parameters to the DescribeRule operation.

", - "refs": { - } - }, - "DescribeRuleResponse": { - "base": "

The result of the DescribeRule operation.

", - "refs": { - } - }, - "DisableRuleRequest": { - "base": "

Container for the parameters to the DisableRule operation.

", - "refs": { - } - }, - "EnableRuleRequest": { - "base": "

Container for the parameters to the EnableRule operation.

", - "refs": { - } - }, - "ErrorCode": { - "base": null, - "refs": { - "PutEventsResultEntry$ErrorCode": "

The error code representing why the event submission failed on this entry.

", - "PutTargetsResultEntry$ErrorCode": "

The error code representing why the target submission failed on this entry.

", - "RemoveTargetsResultEntry$ErrorCode": "

The error code representing why the target removal failed on this entry.

" - } - }, - "ErrorMessage": { - "base": null, - "refs": { - "PutEventsResultEntry$ErrorMessage": "

The error message explaining why the event submission failed on this entry.

", - "PutTargetsResultEntry$ErrorMessage": "

The error message explaining why the target submission failed on this entry.

", - "RemoveTargetsResultEntry$ErrorMessage": "

The error message explaining why the target removal failed on this entry.

" - } - }, - "EventId": { - "base": null, - "refs": { - "PutEventsResultEntry$EventId": "

The ID of the event submitted to Amazon CloudWatch Events.

" - } - }, - "EventPattern": { - "base": null, - "refs": { - "DescribeRuleResponse$EventPattern": "

The event pattern.

", - "PutRuleRequest$EventPattern": "

The event pattern.

", - "Rule$EventPattern": "

The event pattern of the rule.

", - "TestEventPatternRequest$EventPattern": "

The event pattern you want to test.

" - } - }, - "EventResource": { - "base": null, - "refs": { - "EventResourceList$member": null - } - }, - "EventResourceList": { - "base": null, - "refs": { - "PutEventsRequestEntry$Resources": "

AWS resources, identified by Amazon Resource Name (ARN), which the event primarily concerns. Any number, including zero, may be present.

" - } - }, - "EventTime": { - "base": null, - "refs": { - "PutEventsRequestEntry$Time": "

Timestamp of event, per RFC3339. If no timestamp is provided, the timestamp of the PutEvents call will be used.

" - } - }, - "Integer": { - "base": null, - "refs": { - "PutEventsResponse$FailedEntryCount": "

The number of failed entries.

", - "PutTargetsResponse$FailedEntryCount": "

The number of failed entries.

", - "RemoveTargetsResponse$FailedEntryCount": "

The number of failed entries.

" - } - }, - "InternalException": { - "base": "

This exception occurs due to unexpected causes.

", - "refs": { - } - }, - "InvalidEventPatternException": { - "base": "

The event pattern is invalid.

", - "refs": { - } - }, - "LimitExceededException": { - "base": "

This exception occurs if you try to create more rules or add more targets to a rule than allowed by default.

", - "refs": { - } - }, - "LimitMax100": { - "base": null, - "refs": { - "ListRuleNamesByTargetRequest$Limit": "

The maximum number of results to return.

", - "ListRulesRequest$Limit": "

The maximum number of results to return.

", - "ListTargetsByRuleRequest$Limit": "

The maximum number of results to return.

" - } - }, - "ListRuleNamesByTargetRequest": { - "base": "

Container for the parameters to the ListRuleNamesByTarget operation.

", - "refs": { - } - }, - "ListRuleNamesByTargetResponse": { - "base": "

The result of the ListRuleNamesByTarget operation.

", - "refs": { - } - }, - "ListRulesRequest": { - "base": "

Container for the parameters to the ListRules operation.

", - "refs": { - } - }, - "ListRulesResponse": { - "base": "

The result of the ListRules operation.

", - "refs": { - } - }, - "ListTargetsByRuleRequest": { - "base": "

Container for the parameters to the ListTargetsByRule operation.

", - "refs": { - } - }, - "ListTargetsByRuleResponse": { - "base": "

The result of the ListTargetsByRule operation.

", - "refs": { - } - }, - "NextToken": { - "base": null, - "refs": { - "ListRuleNamesByTargetRequest$NextToken": "

The token returned by a previous call to indicate that there is more data available.

", - "ListRuleNamesByTargetResponse$NextToken": "

Indicates that there are additional results to retrieve.

", - "ListRulesRequest$NextToken": "

The token returned by a previous call to indicate that there is more data available.

", - "ListRulesResponse$NextToken": "

Indicates that there are additional results to retrieve.

", - "ListTargetsByRuleRequest$NextToken": "

The token returned by a previous call to indicate that there is more data available.

", - "ListTargetsByRuleResponse$NextToken": "

Indicates that there are additional results to retrieve.

" - } - }, - "PutEventsRequest": { - "base": "

Container for the parameters to the PutEvents operation.

", - "refs": { - } - }, - "PutEventsRequestEntry": { - "base": "

Contains information about the event to be used in the PutEvents action.

", - "refs": { - "PutEventsRequestEntryList$member": null - } - }, - "PutEventsRequestEntryList": { - "base": null, - "refs": { - "PutEventsRequest$Entries": "

The entry that defines an event in your system. You can specify several parameters for the entry such as the source and type of the event, resources associated with the event, and so on.

" - } - }, - "PutEventsResponse": { - "base": "

The result of the PutEvents operation.

", - "refs": { - } - }, - "PutEventsResultEntry": { - "base": "

A PutEventsResult contains a list of PutEventsResultEntry.

", - "refs": { - "PutEventsResultEntryList$member": null - } - }, - "PutEventsResultEntryList": { - "base": null, - "refs": { - "PutEventsResponse$Entries": "

A list of successfully and unsuccessfully ingested events results. If the ingestion was successful, the entry will have the event ID in it. If not, then the ErrorCode and ErrorMessage can be used to identify the problem with the entry.

" - } - }, - "PutRuleRequest": { - "base": "

Container for the parameters to the PutRule operation.

", - "refs": { - } - }, - "PutRuleResponse": { - "base": "

The result of the PutRule operation.

", - "refs": { - } - }, - "PutTargetsRequest": { - "base": "

Container for the parameters to the PutTargets operation.

", - "refs": { - } - }, - "PutTargetsResponse": { - "base": "

The result of the PutTargets operation.

", - "refs": { - } - }, - "PutTargetsResultEntry": { - "base": "

A PutTargetsResult contains a list of PutTargetsResultEntry.

", - "refs": { - "PutTargetsResultEntryList$member": null - } - }, - "PutTargetsResultEntryList": { - "base": null, - "refs": { - "PutTargetsResponse$FailedEntries": "

An array of failed target entries.

" - } - }, - "RemoveTargetsRequest": { - "base": "

Container for the parameters to the RemoveTargets operation.

", - "refs": { - } - }, - "RemoveTargetsResponse": { - "base": "

The result of the RemoveTargets operation.

", - "refs": { - } - }, - "RemoveTargetsResultEntry": { - "base": "

The ID of the target requested to be removed from the rule by Amazon CloudWatch Events.

", - "refs": { - "RemoveTargetsResultEntryList$member": null - } - }, - "RemoveTargetsResultEntryList": { - "base": null, - "refs": { - "RemoveTargetsResponse$FailedEntries": "

An array of failed target entries.

" - } - }, - "ResourceNotFoundException": { - "base": "

The rule does not exist.

", - "refs": { - } - }, - "RoleArn": { - "base": null, - "refs": { - "DescribeRuleResponse$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role associated with the rule.

", - "PutRuleRequest$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role associated with the rule.

", - "Rule$RoleArn": "

The Amazon Resource Name (ARN) associated with the role that is used for target invocation.

" - } - }, - "Rule": { - "base": "

Contains information about a rule in Amazon CloudWatch Events. A ListRulesResult contains a list of Rules.

", - "refs": { - "RuleResponseList$member": null - } - }, - "RuleArn": { - "base": null, - "refs": { - "DescribeRuleResponse$Arn": "

The Amazon Resource Name (ARN) associated with the rule.

", - "PutRuleResponse$RuleArn": "

The Amazon Resource Name (ARN) that identifies the rule.

", - "Rule$Arn": "

The Amazon Resource Name (ARN) of the rule.

" - } - }, - "RuleDescription": { - "base": null, - "refs": { - "DescribeRuleResponse$Description": "

The rule's description.

", - "PutRuleRequest$Description": "

A description of the rule.

", - "Rule$Description": "

The description of the rule.

" - } - }, - "RuleName": { - "base": null, - "refs": { - "DeleteRuleRequest$Name": "

The name of the rule to be deleted.

", - "DescribeRuleRequest$Name": "

The name of the rule you want to describe details for.

", - "DescribeRuleResponse$Name": "

The rule's name.

", - "DisableRuleRequest$Name": "

The name of the rule you want to disable.

", - "EnableRuleRequest$Name": "

The name of the rule that you want to enable.

", - "ListRulesRequest$NamePrefix": "

The prefix matching the rule name.

", - "ListTargetsByRuleRequest$Rule": "

The name of the rule whose targets you want to list.

", - "PutRuleRequest$Name": "

The name of the rule that you are creating or updating.

", - "PutTargetsRequest$Rule": "

The name of the rule you want to add targets to.

", - "RemoveTargetsRequest$Rule": "

The name of the rule you want to remove targets from.

", - "Rule$Name": "

The rule's name.

", - "RuleNameList$member": null - } - }, - "RuleNameList": { - "base": null, - "refs": { - "ListRuleNamesByTargetResponse$RuleNames": "

List of rules names that can invoke the given target.

" - } - }, - "RuleResponseList": { - "base": null, - "refs": { - "ListRulesResponse$Rules": "

List of rules matching the specified criteria.

" - } - }, - "RuleState": { - "base": null, - "refs": { - "DescribeRuleResponse$State": "

Specifies whether the rule is enabled or disabled.

", - "PutRuleRequest$State": "

Indicates whether the rule is enabled or disabled.

", - "Rule$State": "

The rule's state.

" - } - }, - "ScheduleExpression": { - "base": null, - "refs": { - "DescribeRuleResponse$ScheduleExpression": "

The scheduling expression. For example, \"cron(0 20 * * ? *)\", \"rate(5 minutes)\".

", - "PutRuleRequest$ScheduleExpression": "

The scheduling expression. For example, \"cron(0 20 * * ? *)\", \"rate(5 minutes)\".

", - "Rule$ScheduleExpression": "

The scheduling expression. For example, \"cron(0 20 * * ? *)\", \"rate(5 minutes)\".

" - } - }, - "String": { - "base": null, - "refs": { - "PutEventsRequestEntry$Source": "

The source of the event.

", - "PutEventsRequestEntry$DetailType": "

Free-form string used to decide what fields to expect in the event detail.

", - "PutEventsRequestEntry$Detail": "

In the JSON sense, an object containing fields, which may also contain nested sub-objects. No constraints are imposed on its contents.

", - "TestEventPatternRequest$Event": "

The event in the JSON format to test against the event pattern.

" - } - }, - "Target": { - "base": "

Targets are the resources that can be invoked when a rule is triggered. For example, AWS Lambda functions, Amazon Kinesis streams, and built-in targets.

Input and InputPath are mutually-exclusive and optional parameters of a target. When a rule is triggered due to a matched event, if for a target:

  • Neither Input nor InputPath is specified, then the entire event is passed to the target in JSON form.
  • InputPath is specified in the form of JSONPath (e.g. $.detail), then only the part of the event specified in the path is passed to the target (e.g. only the detail part of the event is passed).
  • Input is specified in the form of a valid JSON, then the matched event is overridden with this constant.
", - "refs": { - "TargetList$member": null - } - }, - "TargetArn": { - "base": null, - "refs": { - "ListRuleNamesByTargetRequest$TargetArn": "

The Amazon Resource Name (ARN) of the target resource that you want to list the rules for.

", - "Target$Arn": "

The Amazon Resource Name (ARN) associated of the target.

" - } - }, - "TargetId": { - "base": null, - "refs": { - "PutTargetsResultEntry$TargetId": "

The ID of the target submitted to Amazon CloudWatch Events.

", - "RemoveTargetsResultEntry$TargetId": "

The ID of the target requested to be removed by Amazon CloudWatch Events.

", - "Target$Id": "

The unique target assignment ID.

", - "TargetIdList$member": null - } - }, - "TargetIdList": { - "base": null, - "refs": { - "RemoveTargetsRequest$Ids": "

The list of target IDs to remove from the rule.

" - } - }, - "TargetInput": { - "base": null, - "refs": { - "Target$Input": "

Valid JSON text passed to the target. For more information about JSON text, see The JavaScript Object Notation (JSON) Data Interchange Format.

" - } - }, - "TargetInputPath": { - "base": null, - "refs": { - "Target$InputPath": "

The value of the JSONPath that is used for extracting part of the matched event when passing it to the target. For more information about JSON paths, see JSONPath.

" - } - }, - "TargetList": { - "base": null, - "refs": { - "ListTargetsByRuleResponse$Targets": "

Lists the targets assigned to the rule.

", - "PutTargetsRequest$Targets": "

List of targets you want to update or add to the rule.

" - } - }, - "TestEventPatternRequest": { - "base": "

Container for the parameters to the TestEventPattern operation.

", - "refs": { - } - }, - "TestEventPatternResponse": { - "base": "

The result of the TestEventPattern operation.

", - "refs": { - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/api-2.json index c4a2fab03..c08f601d0 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/api-2.json @@ -1191,7 +1191,8 @@ "members":{ "JobName":{"shape":"NameString"}, "Arguments":{"shape":"GenericMap"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "NotificationProperty":{"shape":"NotificationProperty"} } }, "ActionList":{ @@ -1844,7 +1845,8 @@ "Connections":{"shape":"ConnectionsList"}, "MaxRetries":{"shape":"MaxRetries"}, "AllocatedCapacity":{"shape":"IntegerValue"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "NotificationProperty":{"shape":"NotificationProperty"} } }, "CreateJobResponse":{ @@ -2255,6 +2257,7 @@ "MaxConcurrentRuns":{"shape":"MaxConcurrentRuns"} } }, + "ExecutionTime":{"type":"integer"}, "FieldType":{"type":"string"}, "FilterString":{ "type":"string", @@ -2845,7 +2848,8 @@ "Connections":{"shape":"ConnectionsList"}, "MaxRetries":{"shape":"MaxRetries"}, "AllocatedCapacity":{"shape":"IntegerValue"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "NotificationProperty":{"shape":"NotificationProperty"} } }, "JobBookmarkEntry":{ @@ -2886,8 +2890,9 @@ "ErrorMessage":{"shape":"ErrorString"}, "PredecessorRuns":{"shape":"PredecessorList"}, "AllocatedCapacity":{"shape":"IntegerValue"}, - "ExecutionTime":{"shape":"IntegerValue"}, - "Timeout":{"shape":"Timeout"} + "ExecutionTime":{"shape":"ExecutionTime"}, + "Timeout":{"shape":"Timeout"}, + "NotificationProperty":{"shape":"NotificationProperty"} } }, "JobRunList":{ @@ -2918,7 +2923,8 @@ "Connections":{"shape":"ConnectionsList"}, "MaxRetries":{"shape":"MaxRetries"}, "AllocatedCapacity":{"shape":"IntegerValue"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "NotificationProperty":{"shape":"NotificationProperty"} } }, "JsonClassifier":{ @@ -3065,6 +3071,17 @@ "type":"integer", "min":0 }, + "NotificationProperty":{ + "type":"structure", + "members":{ + "NotifyDelayAfter":{"shape":"NotifyDelayAfter"} + } + }, + "NotifyDelayAfter":{ + "type":"integer", + "box":true, + "min":1 + }, "OperationTimeoutException":{ "type":"structure", "members":{ @@ -3369,7 +3386,8 @@ "JobRunId":{"shape":"IdString"}, "Arguments":{"shape":"GenericMap"}, "AllocatedCapacity":{"shape":"IntegerValue"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "NotificationProperty":{"shape":"NotificationProperty"} } }, "StartJobRunResponse":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/docs-2.json index f9cc632af..d6c6dca1b 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/glue/2017-03-31/docs-2.json @@ -989,6 +989,12 @@ "JobUpdate$ExecutionProperty": "

An ExecutionProperty specifying the maximum number of concurrent runs allowed for this job.

" } }, + "ExecutionTime": { + "base": null, + "refs": { + "JobRun$ExecutionTime": "

The amount of time (in seconds) that the job run consumed resources.

" + } + }, "FieldType": { "base": null, "refs": { @@ -1446,7 +1452,6 @@ "JobBookmarkEntry$Run": "

The run ID number.

", "JobBookmarkEntry$Attempt": "

The attempt ID number.

", "JobRun$AllocatedCapacity": "

The number of AWS Glue data processing units (DPUs) allocated to this JobRun. From 2 to 100 DPUs can be allocated; the default is 10. A DPU is a relative measure of processing power that consists of 4 vCPUs of compute capacity and 16 GB of memory. For more information, see the AWS Glue pricing page.

", - "JobRun$ExecutionTime": "

The amount of time (in seconds) that the job run consumed resources.

", "JobUpdate$AllocatedCapacity": "

The number of AWS Glue data processing units (DPUs) to allocate to this Job. From 2 to 100 DPUs can be allocated; the default is 10. A DPU is a relative measure of processing power that consists of 4 vCPUs of compute capacity and 16 GB of memory. For more information, see the AWS Glue pricing page.

", "StartJobRunRequest$AllocatedCapacity": "

The number of AWS Glue data processing units (DPUs) to allocate to this JobRun. From 2 to 100 DPUs can be allocated; the default is 10. A DPU is a relative measure of processing power that consists of 4 vCPUs of compute capacity and 16 GB of memory. For more information, see the AWS Glue pricing page.

" } @@ -1869,6 +1874,23 @@ "TableInput$Retention": "

Retention time for this table.

" } }, + "NotificationProperty": { + "base": "

Specifies configuration properties of a notification.

", + "refs": { + "Action$NotificationProperty": "

Specifies configuration properties of a job run notification.

", + "CreateJobRequest$NotificationProperty": "

Specifies configuration properties of a job notification.

", + "Job$NotificationProperty": "

Specifies configuration properties of a job notification.

", + "JobRun$NotificationProperty": "

Specifies configuration properties of a job run notification.

", + "JobUpdate$NotificationProperty": "

Specifies configuration properties of a job notification.

", + "StartJobRunRequest$NotificationProperty": "

Specifies configuration properties of a job run notification.

" + } + }, + "NotifyDelayAfter": { + "base": null, + "refs": { + "NotificationProperty$NotifyDelayAfter": "

After a job run starts, the number of minutes to wait before sending a job run delay notification.

" + } + }, "OperationTimeoutException": { "base": "

The operation timed out.

", "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/api-2.json deleted file mode 100644 index d1a6d91b2..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/api-2.json +++ /dev/null @@ -1,1426 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2015-08-18", - "endpointPrefix":"inspector", - "jsonVersion":"1.1", - "protocol":"json", - "serviceFullName":"Amazon Inspector", - "signatureVersion":"v4", - "targetPrefix":"InspectorService" - }, - "operations":{ - "AddAttributesToFindings":{ - "name":"AddAttributesToFindings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddAttributesToFindingsRequest"}, - "output":{"shape":"AddAttributesToFindingsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "AttachAssessmentAndRulesPackage":{ - "name":"AttachAssessmentAndRulesPackage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AttachAssessmentAndRulesPackageRequest"}, - "output":{"shape":"AttachAssessmentAndRulesPackageResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "CreateApplication":{ - "name":"CreateApplication", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateApplicationRequest"}, - "output":{"shape":"CreateApplicationResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "CreateAssessment":{ - "name":"CreateAssessment", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateAssessmentRequest"}, - "output":{"shape":"CreateAssessmentResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "CreateResourceGroup":{ - "name":"CreateResourceGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateResourceGroupRequest"}, - "output":{"shape":"CreateResourceGroupResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"} - ] - }, - "DeleteApplication":{ - "name":"DeleteApplication", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteApplicationRequest"}, - "output":{"shape":"DeleteApplicationResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"OperationInProgressException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DeleteAssessment":{ - "name":"DeleteAssessment", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteAssessmentRequest"}, - "output":{"shape":"DeleteAssessmentResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"OperationInProgressException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DeleteRun":{ - "name":"DeleteRun", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteRunRequest"}, - "output":{"shape":"DeleteRunResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DescribeApplication":{ - "name":"DescribeApplication", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeApplicationRequest"}, - "output":{"shape":"DescribeApplicationResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DescribeAssessment":{ - "name":"DescribeAssessment", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeAssessmentRequest"}, - "output":{"shape":"DescribeAssessmentResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DescribeCrossAccountAccessRole":{ - "name":"DescribeCrossAccountAccessRole", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "output":{"shape":"DescribeCrossAccountAccessRoleResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"AccessDeniedException"} - ] - }, - "DescribeFinding":{ - "name":"DescribeFinding", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeFindingRequest"}, - "output":{"shape":"DescribeFindingResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DescribeResourceGroup":{ - "name":"DescribeResourceGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeResourceGroupRequest"}, - "output":{"shape":"DescribeResourceGroupResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DescribeRulesPackage":{ - "name":"DescribeRulesPackage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRulesPackageRequest"}, - "output":{"shape":"DescribeRulesPackageResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DescribeRun":{ - "name":"DescribeRun", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeRunRequest"}, - "output":{"shape":"DescribeRunResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "DetachAssessmentAndRulesPackage":{ - "name":"DetachAssessmentAndRulesPackage", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DetachAssessmentAndRulesPackageRequest"}, - "output":{"shape":"DetachAssessmentAndRulesPackageResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "GetAssessmentTelemetry":{ - "name":"GetAssessmentTelemetry", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"GetAssessmentTelemetryRequest"}, - "output":{"shape":"GetAssessmentTelemetryResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListApplications":{ - "name":"ListApplications", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListApplicationsRequest"}, - "output":{"shape":"ListApplicationsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"} - ] - }, - "ListAssessmentAgents":{ - "name":"ListAssessmentAgents", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListAssessmentAgentsRequest"}, - "output":{"shape":"ListAssessmentAgentsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListAssessments":{ - "name":"ListAssessments", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListAssessmentsRequest"}, - "output":{"shape":"ListAssessmentsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListAttachedAssessments":{ - "name":"ListAttachedAssessments", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListAttachedAssessmentsRequest"}, - "output":{"shape":"ListAttachedAssessmentsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListAttachedRulesPackages":{ - "name":"ListAttachedRulesPackages", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListAttachedRulesPackagesRequest"}, - "output":{"shape":"ListAttachedRulesPackagesResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListFindings":{ - "name":"ListFindings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListFindingsRequest"}, - "output":{"shape":"ListFindingsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListRulesPackages":{ - "name":"ListRulesPackages", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListRulesPackagesRequest"}, - "output":{"shape":"ListRulesPackagesResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"} - ] - }, - "ListRuns":{ - "name":"ListRuns", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListRunsRequest"}, - "output":{"shape":"ListRunsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListTagsForResourceRequest"}, - "output":{"shape":"ListTagsForResourceResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "LocalizeText":{ - "name":"LocalizeText", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"LocalizeTextRequest"}, - "output":{"shape":"LocalizeTextResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "PreviewAgentsForResourceGroup":{ - "name":"PreviewAgentsForResourceGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PreviewAgentsForResourceGroupRequest"}, - "output":{"shape":"PreviewAgentsForResourceGroupResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"}, - {"shape":"InvalidCrossAccountRoleException"} - ] - }, - "RegisterCrossAccountAccessRole":{ - "name":"RegisterCrossAccountAccessRole", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RegisterCrossAccountAccessRoleRequest"}, - "output":{"shape":"RegisterCrossAccountAccessRoleResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"InvalidCrossAccountRoleException"} - ] - }, - "RemoveAttributesFromFindings":{ - "name":"RemoveAttributesFromFindings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveAttributesFromFindingsRequest"}, - "output":{"shape":"RemoveAttributesFromFindingsResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "RunAssessment":{ - "name":"RunAssessment", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RunAssessmentRequest"}, - "output":{"shape":"RunAssessmentResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "SetTagsForResource":{ - "name":"SetTagsForResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"SetTagsForResourceRequest"}, - "output":{"shape":"SetTagsForResourceResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "StartDataCollection":{ - "name":"StartDataCollection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StartDataCollectionRequest"}, - "output":{"shape":"StartDataCollectionResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"}, - {"shape":"InvalidCrossAccountRoleException"} - ] - }, - "StopDataCollection":{ - "name":"StopDataCollection", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"StopDataCollectionRequest"}, - "output":{"shape":"StopDataCollectionResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "UpdateApplication":{ - "name":"UpdateApplication", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UpdateApplicationRequest"}, - "output":{"shape":"UpdateApplicationResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - }, - "UpdateAssessment":{ - "name":"UpdateAssessment", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"UpdateAssessmentRequest"}, - "output":{"shape":"UpdateAssessmentResponse"}, - "errors":[ - {"shape":"InternalException"}, - {"shape":"InvalidInputException"}, - {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} - ] - } - }, - "shapes":{ - "AccessDeniedException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "AddAttributesToFindingsRequest":{ - "type":"structure", - "required":[ - "findingArns", - "attributes" - ], - "members":{ - "findingArns":{"shape":"ArnList"}, - "attributes":{"shape":"AttributeList"} - } - }, - "AddAttributesToFindingsResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "Agent":{ - "type":"structure", - "members":{ - "agentId":{"shape":"AgentId"}, - "assessmentArn":{"shape":"Arn"}, - "agentHealth":{"shape":"AgentHealth"}, - "agentHealthCode":{"shape":"AgentHealthCode"}, - "agentHealthDetails":{"shape":"AgentHealthDetails"}, - "autoScalingGroup":{"shape":"AutoScalingGroup"}, - "accountId":{"shape":"AwsAccount"}, - "telemetry":{"shape":"TelemetryList"} - } - }, - "AgentHealth":{"type":"string"}, - "AgentHealthCode":{"type":"string"}, - "AgentHealthDetails":{"type":"string"}, - "AgentHealthList":{ - "type":"list", - "member":{"shape":"AgentHealth"} - }, - "AgentId":{"type":"string"}, - "AgentList":{ - "type":"list", - "member":{"shape":"Agent"} - }, - "AgentPreview":{ - "type":"structure", - "members":{ - "agentId":{"shape":"AgentId"}, - "autoScalingGroup":{"shape":"AutoScalingGroup"} - } - }, - "AgentPreviewList":{ - "type":"list", - "member":{"shape":"AgentPreview"} - }, - "AgentsFilter":{ - "type":"structure", - "members":{ - "agentHealthList":{"shape":"AgentHealthList"} - } - }, - "Application":{ - "type":"structure", - "members":{ - "applicationArn":{"shape":"Arn"}, - "applicationName":{"shape":"Name"}, - "resourceGroupArn":{"shape":"Arn"} - } - }, - "ApplicationsFilter":{ - "type":"structure", - "members":{ - "applicationNamePatterns":{"shape":"NamePatternList"} - } - }, - "Arn":{"type":"string"}, - "ArnList":{ - "type":"list", - "member":{"shape":"Arn"} - }, - "Assessment":{ - "type":"structure", - "members":{ - "assessmentArn":{"shape":"Arn"}, - "assessmentName":{"shape":"Name"}, - "applicationArn":{"shape":"Arn"}, - "assessmentState":{"shape":"AssessmentState"}, - "failureMessage":{"shape":"FailureMessage"}, - "dataCollected":{"shape":"Bool"}, - "startTime":{"shape":"Timestamp"}, - "endTime":{"shape":"Timestamp"}, - "durationInSeconds":{"shape":"Duration"}, - "userAttributesForFindings":{"shape":"AttributeList"} - } - }, - "AssessmentState":{"type":"string"}, - "AssessmentStateList":{ - "type":"list", - "member":{"shape":"AssessmentState"} - }, - "AssessmentsFilter":{ - "type":"structure", - "members":{ - "assessmentNamePatterns":{"shape":"NamePatternList"}, - "assessmentStates":{"shape":"AssessmentStateList"}, - "dataCollected":{"shape":"Bool"}, - "startTimeRange":{"shape":"TimestampRange"}, - "endTimeRange":{"shape":"TimestampRange"}, - "durationRange":{"shape":"DurationRange"} - } - }, - "AttachAssessmentAndRulesPackageRequest":{ - "type":"structure", - "required":[ - "assessmentArn", - "rulesPackageArn" - ], - "members":{ - "assessmentArn":{"shape":"Arn"}, - "rulesPackageArn":{"shape":"Arn"} - } - }, - "AttachAssessmentAndRulesPackageResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "Attribute":{ - "type":"structure", - "members":{ - "key":{"shape":"AttributeKey"}, - "value":{"shape":"AttributeValue"} - } - }, - "AttributeKey":{"type":"string"}, - "AttributeKeyList":{ - "type":"list", - "member":{"shape":"AttributeKey"} - }, - "AttributeList":{ - "type":"list", - "member":{"shape":"Attribute"} - }, - "AttributeValue":{"type":"string"}, - "AutoScalingGroup":{"type":"string"}, - "AwsAccount":{"type":"string"}, - "Bool":{"type":"boolean"}, - "CreateApplicationRequest":{ - "type":"structure", - "required":[ - "applicationName", - "resourceGroupArn" - ], - "members":{ - "applicationName":{"shape":"Name"}, - "resourceGroupArn":{"shape":"Arn"} - } - }, - "CreateApplicationResponse":{ - "type":"structure", - "members":{ - "applicationArn":{"shape":"Arn"} - } - }, - "CreateAssessmentRequest":{ - "type":"structure", - "required":[ - "applicationArn", - "assessmentName", - "durationInSeconds" - ], - "members":{ - "applicationArn":{"shape":"Arn"}, - "assessmentName":{"shape":"Name"}, - "durationInSeconds":{"shape":"Duration"}, - "userAttributesForFindings":{"shape":"AttributeList"} - } - }, - "CreateAssessmentResponse":{ - "type":"structure", - "members":{ - "assessmentArn":{"shape":"Arn"} - } - }, - "CreateResourceGroupRequest":{ - "type":"structure", - "required":["resourceGroupTags"], - "members":{ - "resourceGroupTags":{"shape":"ResourceGroupTags"} - } - }, - "CreateResourceGroupResponse":{ - "type":"structure", - "members":{ - "resourceGroupArn":{"shape":"Arn"} - } - }, - "DeleteApplicationRequest":{ - "type":"structure", - "required":["applicationArn"], - "members":{ - "applicationArn":{"shape":"Arn"} - } - }, - "DeleteApplicationResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "DeleteAssessmentRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"} - } - }, - "DeleteAssessmentResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "DeleteRunRequest":{ - "type":"structure", - "required":["runArn"], - "members":{ - "runArn":{"shape":"Arn"} - } - }, - "DeleteRunResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "DescribeApplicationRequest":{ - "type":"structure", - "required":["applicationArn"], - "members":{ - "applicationArn":{"shape":"Arn"} - } - }, - "DescribeApplicationResponse":{ - "type":"structure", - "members":{ - "application":{"shape":"Application"} - } - }, - "DescribeAssessmentRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"} - } - }, - "DescribeAssessmentResponse":{ - "type":"structure", - "members":{ - "assessment":{"shape":"Assessment"} - } - }, - "DescribeCrossAccountAccessRoleResponse":{ - "type":"structure", - "members":{ - "roleArn":{"shape":"Arn"}, - "valid":{"shape":"Bool"} - } - }, - "DescribeFindingRequest":{ - "type":"structure", - "required":["findingArn"], - "members":{ - "findingArn":{"shape":"Arn"} - } - }, - "DescribeFindingResponse":{ - "type":"structure", - "members":{ - "finding":{"shape":"Finding"} - } - }, - "DescribeResourceGroupRequest":{ - "type":"structure", - "required":["resourceGroupArn"], - "members":{ - "resourceGroupArn":{"shape":"Arn"} - } - }, - "DescribeResourceGroupResponse":{ - "type":"structure", - "members":{ - "resourceGroup":{"shape":"ResourceGroup"} - } - }, - "DescribeRulesPackageRequest":{ - "type":"structure", - "required":["rulesPackageArn"], - "members":{ - "rulesPackageArn":{"shape":"Arn"} - } - }, - "DescribeRulesPackageResponse":{ - "type":"structure", - "members":{ - "rulesPackage":{"shape":"RulesPackage"} - } - }, - "DescribeRunRequest":{ - "type":"structure", - "required":["runArn"], - "members":{ - "runArn":{"shape":"Arn"} - } - }, - "DescribeRunResponse":{ - "type":"structure", - "members":{ - "run":{"shape":"Run"} - } - }, - "DetachAssessmentAndRulesPackageRequest":{ - "type":"structure", - "required":[ - "assessmentArn", - "rulesPackageArn" - ], - "members":{ - "assessmentArn":{"shape":"Arn"}, - "rulesPackageArn":{"shape":"Arn"} - } - }, - "DetachAssessmentAndRulesPackageResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "Duration":{"type":"integer"}, - "DurationRange":{ - "type":"structure", - "members":{ - "minimum":{"shape":"Duration"}, - "maximum":{"shape":"Duration"} - } - }, - "FailureMessage":{"type":"string"}, - "Finding":{ - "type":"structure", - "members":{ - "findingArn":{"shape":"Arn"}, - "runArn":{"shape":"Arn"}, - "rulesPackageArn":{"shape":"Arn"}, - "ruleName":{"shape":"Name"}, - "agentId":{"shape":"AgentId"}, - "autoScalingGroup":{"shape":"AutoScalingGroup"}, - "severity":{"shape":"Severity"}, - "finding":{"shape":"LocalizedText"}, - "description":{"shape":"LocalizedText"}, - "recommendation":{"shape":"LocalizedText"}, - "attributes":{"shape":"AttributeList"}, - "userAttributes":{"shape":"AttributeList"} - } - }, - "FindingsFilter":{ - "type":"structure", - "members":{ - "rulesPackageArns":{"shape":"ArnList"}, - "ruleNames":{"shape":"NameList"}, - "severities":{"shape":"SeverityList"}, - "attributes":{"shape":"AttributeList"}, - "userAttributes":{"shape":"AttributeList"} - } - }, - "GetAssessmentTelemetryRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"} - } - }, - "GetAssessmentTelemetryResponse":{ - "type":"structure", - "members":{ - "telemetry":{"shape":"TelemetryList"} - } - }, - "Integer":{"type":"integer"}, - "InternalException":{ - "type":"structure", - "members":{ - }, - "exception":true, - "fault":true - }, - "InvalidCrossAccountRoleException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "InvalidInputException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "ListApplicationsRequest":{ - "type":"structure", - "members":{ - "filter":{"shape":"ApplicationsFilter"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListApplicationsResponse":{ - "type":"structure", - "members":{ - "applicationArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListAssessmentAgentsRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"}, - "filter":{"shape":"AgentsFilter"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListAssessmentAgentsResponse":{ - "type":"structure", - "members":{ - "agentList":{"shape":"AgentList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListAssessmentsRequest":{ - "type":"structure", - "members":{ - "applicationArns":{"shape":"ArnList"}, - "filter":{"shape":"AssessmentsFilter"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListAssessmentsResponse":{ - "type":"structure", - "members":{ - "assessmentArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListAttachedAssessmentsRequest":{ - "type":"structure", - "required":["rulesPackageArn"], - "members":{ - "rulesPackageArn":{"shape":"Arn"}, - "filter":{"shape":"AssessmentsFilter"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListAttachedAssessmentsResponse":{ - "type":"structure", - "members":{ - "assessmentArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListAttachedRulesPackagesRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListAttachedRulesPackagesResponse":{ - "type":"structure", - "members":{ - "rulesPackageArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListFindingsRequest":{ - "type":"structure", - "members":{ - "runArns":{"shape":"ArnList"}, - "filter":{"shape":"FindingsFilter"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListFindingsResponse":{ - "type":"structure", - "members":{ - "findingArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListRulesPackagesRequest":{ - "type":"structure", - "members":{ - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListRulesPackagesResponse":{ - "type":"structure", - "members":{ - "rulesPackageArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListRunsRequest":{ - "type":"structure", - "members":{ - "assessmentArns":{"shape":"ArnList"}, - "filter":{"shape":"RunsFilter"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "ListRunsResponse":{ - "type":"structure", - "members":{ - "runArnList":{"shape":"ArnList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "ListTagsForResourceRequest":{ - "type":"structure", - "required":["resourceArn"], - "members":{ - "resourceArn":{"shape":"Arn"} - } - }, - "ListTagsForResourceResponse":{ - "type":"structure", - "members":{ - "tagList":{"shape":"TagList"} - } - }, - "Locale":{"type":"string"}, - "LocalizeTextRequest":{ - "type":"structure", - "required":[ - "localizedTexts", - "locale" - ], - "members":{ - "localizedTexts":{"shape":"LocalizedTextList"}, - "locale":{"shape":"Locale"} - } - }, - "LocalizeTextResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"}, - "results":{"shape":"TextList"} - } - }, - "LocalizedFacility":{"type":"string"}, - "LocalizedText":{ - "type":"structure", - "members":{ - "key":{"shape":"LocalizedTextKey"}, - "parameters":{"shape":"ParameterList"} - } - }, - "LocalizedTextId":{"type":"string"}, - "LocalizedTextKey":{ - "type":"structure", - "members":{ - "facility":{"shape":"LocalizedFacility"}, - "id":{"shape":"LocalizedTextId"} - } - }, - "LocalizedTextList":{ - "type":"list", - "member":{"shape":"LocalizedText"} - }, - "Long":{"type":"long"}, - "Message":{"type":"string"}, - "MessageType":{"type":"string"}, - "MessageTypeTelemetry":{ - "type":"structure", - "members":{ - "messageType":{"shape":"MessageType"}, - "count":{"shape":"Long"}, - "dataSize":{"shape":"Long"} - } - }, - "MessageTypeTelemetryList":{ - "type":"list", - "member":{"shape":"MessageTypeTelemetry"} - }, - "Name":{"type":"string"}, - "NameList":{ - "type":"list", - "member":{"shape":"Name"} - }, - "NamePattern":{"type":"string"}, - "NamePatternList":{ - "type":"list", - "member":{"shape":"NamePattern"} - }, - "NoSuchEntityException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "OperationInProgressException":{ - "type":"structure", - "members":{ - }, - "exception":true - }, - "PaginationToken":{"type":"string"}, - "Parameter":{ - "type":"structure", - "members":{ - "name":{"shape":"ParameterName"}, - "value":{"shape":"ParameterValue"} - } - }, - "ParameterList":{ - "type":"list", - "member":{"shape":"Parameter"} - }, - "ParameterName":{"type":"string"}, - "ParameterValue":{"type":"string"}, - "PreviewAgentsForResourceGroupRequest":{ - "type":"structure", - "required":["resourceGroupArn"], - "members":{ - "resourceGroupArn":{"shape":"Arn"}, - "nextToken":{"shape":"PaginationToken"}, - "maxResults":{"shape":"Integer"} - } - }, - "PreviewAgentsForResourceGroupResponse":{ - "type":"structure", - "members":{ - "agentPreviewList":{"shape":"AgentPreviewList"}, - "nextToken":{"shape":"PaginationToken"} - } - }, - "RegisterCrossAccountAccessRoleRequest":{ - "type":"structure", - "required":["roleArn"], - "members":{ - "roleArn":{"shape":"Arn"} - } - }, - "RegisterCrossAccountAccessRoleResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "RemoveAttributesFromFindingsRequest":{ - "type":"structure", - "required":[ - "findingArns", - "attributeKeys" - ], - "members":{ - "findingArns":{"shape":"ArnList"}, - "attributeKeys":{"shape":"AttributeKeyList"} - } - }, - "RemoveAttributesFromFindingsResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "ResourceGroup":{ - "type":"structure", - "members":{ - "resourceGroupArn":{"shape":"Arn"}, - "resourceGroupTags":{"shape":"ResourceGroupTags"} - } - }, - "ResourceGroupTags":{"type":"string"}, - "RulesPackage":{ - "type":"structure", - "members":{ - "rulesPackageArn":{"shape":"Arn"}, - "rulesPackageName":{"shape":"Name"}, - "version":{"shape":"Version"}, - "provider":{"shape":"Name"}, - "description":{"shape":"LocalizedText"} - } - }, - "Run":{ - "type":"structure", - "members":{ - "runArn":{"shape":"Arn"}, - "runName":{"shape":"Name"}, - "assessmentArn":{"shape":"Arn"}, - "runState":{"shape":"RunState"}, - "rulesPackages":{"shape":"ArnList"}, - "creationTime":{"shape":"Timestamp"}, - "completionTime":{"shape":"Timestamp"} - } - }, - "RunAssessmentRequest":{ - "type":"structure", - "required":[ - "assessmentArn", - "runName" - ], - "members":{ - "assessmentArn":{"shape":"Arn"}, - "runName":{"shape":"Name"} - } - }, - "RunAssessmentResponse":{ - "type":"structure", - "members":{ - "runArn":{"shape":"Arn"} - } - }, - "RunState":{"type":"string"}, - "RunStateList":{ - "type":"list", - "member":{"shape":"RunState"} - }, - "RunsFilter":{ - "type":"structure", - "members":{ - "runNamePatterns":{"shape":"NamePatternList"}, - "runStates":{"shape":"RunStateList"}, - "rulesPackages":{"shape":"ArnList"}, - "creationTime":{"shape":"TimestampRange"}, - "completionTime":{"shape":"TimestampRange"} - } - }, - "SetTagsForResourceRequest":{ - "type":"structure", - "required":["resourceArn"], - "members":{ - "resourceArn":{"shape":"Arn"}, - "tags":{"shape":"TagList"} - } - }, - "SetTagsForResourceResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "Severity":{"type":"string"}, - "SeverityList":{ - "type":"list", - "member":{"shape":"Severity"} - }, - "StartDataCollectionRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"} - } - }, - "StartDataCollectionResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "StopDataCollectionRequest":{ - "type":"structure", - "required":["assessmentArn"], - "members":{ - "assessmentArn":{"shape":"Arn"} - } - }, - "StopDataCollectionResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "Tag":{ - "type":"structure", - "members":{ - "Key":{"shape":"TagKey"}, - "Value":{"shape":"TagValue"} - } - }, - "TagKey":{"type":"string"}, - "TagList":{ - "type":"list", - "member":{"shape":"Tag"} - }, - "TagValue":{"type":"string"}, - "Telemetry":{ - "type":"structure", - "members":{ - "status":{"shape":"TelemetryStatus"}, - "messageTypeTelemetries":{"shape":"MessageTypeTelemetryList"} - } - }, - "TelemetryList":{ - "type":"list", - "member":{"shape":"Telemetry"} - }, - "TelemetryStatus":{"type":"string"}, - "Text":{"type":"string"}, - "TextList":{ - "type":"list", - "member":{"shape":"Text"} - }, - "Timestamp":{"type":"timestamp"}, - "TimestampRange":{ - "type":"structure", - "members":{ - "minimum":{"shape":"Timestamp"}, - "maximum":{"shape":"Timestamp"} - } - }, - "UpdateApplicationRequest":{ - "type":"structure", - "required":[ - "applicationArn", - "applicationName", - "resourceGroupArn" - ], - "members":{ - "applicationArn":{"shape":"Arn"}, - "applicationName":{"shape":"Name"}, - "resourceGroupArn":{"shape":"Arn"} - } - }, - "UpdateApplicationResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "UpdateAssessmentRequest":{ - "type":"structure", - "required":[ - "assessmentArn", - "assessmentName", - "durationInSeconds" - ], - "members":{ - "assessmentArn":{"shape":"Arn"}, - "assessmentName":{"shape":"Name"}, - "durationInSeconds":{"shape":"Duration"} - } - }, - "UpdateAssessmentResponse":{ - "type":"structure", - "members":{ - "message":{"shape":"Message"} - } - }, - "Version":{"type":"string"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/docs-2.json deleted file mode 100644 index a7ef77a7e..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/docs-2.json +++ /dev/null @@ -1,1016 +0,0 @@ -{ - "version": "2.0", - "service": "Amazon Inspector

Amazon Inspector enables you to analyze the behavior of the applications you run in AWS and to identify potential security issues. For more information, see Amazon Inspector User Guide.

", - "operations": { - "AddAttributesToFindings": "

Assigns attributes (key and value pair) to the findings specified by the findings' ARNs.

", - "AttachAssessmentAndRulesPackage": "

Attaches the rules package specified by the rules package ARN to the assessment specified by the assessment ARN.

", - "CreateApplication": "

Creates a new application using the resource group ARN generated by CreateResourceGroup. You can create up to 50 applications per AWS account. You can run up to 500 concurrent agents per AWS account. For more information, see Inspector Applications.

", - "CreateAssessment": "

Creates an assessment for the application specified by the application ARN. You can create up to 500 assessments per AWS account.

", - "CreateResourceGroup": "

Creates a resource group using the specified set of tags (key and value pairs) that are used to select the EC2 instances to be included in an Inspector application. The created resource group is then used to create an Inspector application.

", - "DeleteApplication": "

Deletes the application specified by the application ARN.

", - "DeleteAssessment": "

Deletes the assessment specified by the assessment ARN.

", - "DeleteRun": "

Deletes the assessment run specified by the run ARN.

", - "DescribeApplication": "

Describes the application specified by the application ARN.

", - "DescribeAssessment": "

Describes the assessment specified by the assessment ARN.

", - "DescribeCrossAccountAccessRole": "

Describes the IAM role that enables Inspector to access your AWS account.

", - "DescribeFinding": "

Describes the finding specified by the finding ARN.

", - "DescribeResourceGroup": "

Describes the resource group specified by the resource group ARN.

", - "DescribeRulesPackage": "

Describes the rules package specified by the rules package ARN.

", - "DescribeRun": "

Describes the assessment run specified by the run ARN.

", - "DetachAssessmentAndRulesPackage": "

Detaches the rules package specified by the rules package ARN from the assessment specified by the assessment ARN.

", - "GetAssessmentTelemetry": "

Returns the metadata about the telemetry (application behavioral data) for the assessment specified by the assessment ARN.

", - "ListApplications": "

Lists the ARNs of the applications within this AWS account. For more information about applications, see Inspector Applications.

", - "ListAssessmentAgents": "

Lists the agents of the assessment specified by the assessment ARN.

", - "ListAssessments": "

Lists the assessments corresponding to applications specified by the applications' ARNs.

", - "ListAttachedAssessments": "

Lists the assessments attached to the rules package specified by the rules package ARN.

", - "ListAttachedRulesPackages": "

Lists the rules packages attached to the assessment specified by the assessment ARN.

", - "ListFindings": "

Lists findings generated by the assessment run specified by the run ARNs.

", - "ListRulesPackages": "

Lists all available Inspector rules packages.

", - "ListRuns": "

Lists the assessment runs associated with the assessments specified by the assessment ARNs.

", - "ListTagsForResource": "

Lists all tags associated with a resource.

", - "LocalizeText": "

Translates a textual identifier into a user-readable text in a specified locale.

", - "PreviewAgentsForResourceGroup": "

Previews the agents installed on the EC2 instances that are included in the application created with the specified resource group.

", - "RegisterCrossAccountAccessRole": "

Register the role that Inspector uses to list your EC2 instances during the assessment.

", - "RemoveAttributesFromFindings": "

Removes the entire attribute (key and value pair) from the findings specified by the finding ARNs where an attribute with the specified key exists.

", - "RunAssessment": "

Starts the analysis of the application’s behavior against selected rule packages for the assessment specified by the assessment ARN.

", - "SetTagsForResource": "

Sets tags (key and value pairs) to the assessment specified by the assessment ARN.

", - "StartDataCollection": "

Starts data collection for the assessment specified by the assessment ARN. For this API to function properly, you must not exceed the limit of running up to 500 concurrent agents per AWS account.

", - "StopDataCollection": "

Stop data collection for the assessment specified by the assessment ARN.

", - "UpdateApplication": "

Updates application specified by the application ARN.

", - "UpdateAssessment": "

Updates the assessment specified by the assessment ARN.

" - }, - "shapes": { - "AccessDeniedException": { - "base": null, - "refs": { - } - }, - "AddAttributesToFindingsRequest": { - "base": null, - "refs": { - } - }, - "AddAttributesToFindingsResponse": { - "base": null, - "refs": { - } - }, - "Agent": { - "base": "

Contains information about an Inspector agent. This data type is used as a response element in the ListAssessmentAgents action.

", - "refs": { - "AgentList$member": null - } - }, - "AgentHealth": { - "base": null, - "refs": { - "Agent$agentHealth": "

The current health state of the agent. Values can be set to HEALTHY or UNHEALTHY.

", - "AgentHealthList$member": null - } - }, - "AgentHealthCode": { - "base": null, - "refs": { - "Agent$agentHealthCode": "

The detailed health state of the agent. Values can be set to RUNNING, HEALTHY, UNHEALTHY, UNKNOWN, BLACKLISTED, SHUTDOWN, THROTTLED.

" - } - }, - "AgentHealthDetails": { - "base": null, - "refs": { - "Agent$agentHealthDetails": "

The description for the agent health code.

" - } - }, - "AgentHealthList": { - "base": null, - "refs": { - "AgentsFilter$agentHealthList": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the agentHealth property of the Agent data type.

" - } - }, - "AgentId": { - "base": null, - "refs": { - "Agent$agentId": "

The EC2 instance ID where the agent is installed.

", - "AgentPreview$agentId": "

The id of the EC2 instance where the agent is intalled.

", - "Finding$agentId": "

The EC2 instance ID where the agent is installed that is used during the assessment that generates the finding.

" - } - }, - "AgentList": { - "base": null, - "refs": { - "ListAssessmentAgentsResponse$agentList": "

A list of ARNs specifying the agents returned by the action.

" - } - }, - "AgentPreview": { - "base": "

This data type is used as a response element in the PreviewAgentsForResourceGroup action.

", - "refs": { - "AgentPreviewList$member": null - } - }, - "AgentPreviewList": { - "base": null, - "refs": { - "PreviewAgentsForResourceGroupResponse$agentPreviewList": "

The resulting list of agents.

" - } - }, - "AgentsFilter": { - "base": "

This data type is used as a response element in the ListAssessmentAgents action.

", - "refs": { - "ListAssessmentAgentsRequest$filter": "

You can use this parameter to specify a subset of data to be included in the action's response.

For a record to match a filter, all specified filter attributes must match. When multiple values are specified for a filter attribute, any of the values can match.

" - } - }, - "Application": { - "base": "

Contains information about an Inspector application.

This data type is used as the response element in the DescribeApplication action.

", - "refs": { - "DescribeApplicationResponse$application": "

Information about the application.

" - } - }, - "ApplicationsFilter": { - "base": "

This data type is used as the request parameter in the ListApplications action.

", - "refs": { - "ListApplicationsRequest$filter": "

You can use this parameter to specify a subset of data to be included in the action's response.

For a record to match a filter, all specified filter attributes must match. When multiple values are specified for a filter attribute, any of the values can match.

" - } - }, - "Arn": { - "base": null, - "refs": { - "Agent$assessmentArn": "

The ARN of the assessment that is associated with the agent.

", - "Application$applicationArn": "

The ARN specifying the Inspector application.

", - "Application$resourceGroupArn": "

The ARN specifying the resource group that is associated with the application.

", - "ArnList$member": null, - "Assessment$assessmentArn": "

The ARN of the assessment.

", - "Assessment$applicationArn": "

The ARN of the application that corresponds to this assessment.

", - "AttachAssessmentAndRulesPackageRequest$assessmentArn": "

The ARN specifying the assessment to which you want to attach a rules package.

", - "AttachAssessmentAndRulesPackageRequest$rulesPackageArn": "

The ARN specifying the rules package that you want to attach to the assessment.

", - "CreateApplicationRequest$resourceGroupArn": "

The ARN specifying the resource group that is used to create the application.

", - "CreateApplicationResponse$applicationArn": "

The ARN specifying the application that is created.

", - "CreateAssessmentRequest$applicationArn": "

The ARN specifying the application for which you want to create an assessment.

", - "CreateAssessmentResponse$assessmentArn": "

The ARN specifying the assessment that is created.

", - "CreateResourceGroupResponse$resourceGroupArn": "

The ARN specifying the resource group that is created.

", - "DeleteApplicationRequest$applicationArn": "

The ARN specifying the application that you want to delete.

", - "DeleteAssessmentRequest$assessmentArn": "

The ARN specifying the assessment that you want to delete.

", - "DeleteRunRequest$runArn": "

The ARN specifying the assessment run that you want to delete.

", - "DescribeApplicationRequest$applicationArn": "

The ARN specifying the application that you want to describe.

", - "DescribeAssessmentRequest$assessmentArn": "

The ARN specifying the assessment that you want to describe.

", - "DescribeCrossAccountAccessRoleResponse$roleArn": "

The ARN specifying the IAM role that Inspector uses to access your AWS account.

", - "DescribeFindingRequest$findingArn": "

The ARN specifying the finding that you want to describe.

", - "DescribeResourceGroupRequest$resourceGroupArn": "

The ARN specifying the resource group that you want to describe.

", - "DescribeRulesPackageRequest$rulesPackageArn": "

The ARN specifying the rules package that you want to describe.

", - "DescribeRunRequest$runArn": "

The ARN specifying the assessment run that you want to describe.

", - "DetachAssessmentAndRulesPackageRequest$assessmentArn": "

The ARN specifying the assessment from which you want to detach a rules package.

", - "DetachAssessmentAndRulesPackageRequest$rulesPackageArn": "

The ARN specifying the rules package that you want to detach from the assessment.

", - "Finding$findingArn": "

The ARN specifying the finding.

", - "Finding$runArn": "

The ARN of the assessment run that generated the finding.

", - "Finding$rulesPackageArn": "

The ARN of the rules package that is used to generate the finding.

", - "GetAssessmentTelemetryRequest$assessmentArn": "

The ARN specifying the assessment the telemetry of which you want to obtain.

", - "ListAssessmentAgentsRequest$assessmentArn": "

The ARN specifying the assessment whose agents you want to list.

", - "ListAttachedAssessmentsRequest$rulesPackageArn": "

The ARN specifying the rules package whose assessments you want to list.

", - "ListAttachedRulesPackagesRequest$assessmentArn": "

The ARN specifying the assessment whose rules packages you want to list.

", - "ListTagsForResourceRequest$resourceArn": "

The ARN specifying the resource whose tags you want to list.

", - "PreviewAgentsForResourceGroupRequest$resourceGroupArn": "

The ARN of the resource group that is used to create an application.

", - "RegisterCrossAccountAccessRoleRequest$roleArn": "The ARN of the IAM role that Inspector uses to list your EC2 instances during the assessment.", - "ResourceGroup$resourceGroupArn": "

The ARN of the resource group.

", - "RulesPackage$rulesPackageArn": "

The ARN of the rules package.

", - "Run$runArn": "

The ARN of the run.

", - "Run$assessmentArn": "

The ARN of the assessment that is associated with the run.

", - "RunAssessmentRequest$assessmentArn": "

The ARN of the assessment that you want to run.

", - "RunAssessmentResponse$runArn": "

The ARN specifying the run of the assessment.

", - "SetTagsForResourceRequest$resourceArn": "

The ARN of the assessment that you want to set tags to.

", - "StartDataCollectionRequest$assessmentArn": "

The ARN of the assessment for which you want to start the data collection process.

", - "StopDataCollectionRequest$assessmentArn": "

The ARN of the assessment for which you want to stop the data collection process.

", - "UpdateApplicationRequest$applicationArn": "

Application ARN that you want to update.

", - "UpdateApplicationRequest$resourceGroupArn": "

The resource group ARN that you want to update.

", - "UpdateAssessmentRequest$assessmentArn": "

Asessment ARN that you want to update.

" - } - }, - "ArnList": { - "base": null, - "refs": { - "AddAttributesToFindingsRequest$findingArns": "

The ARNs specifying the findings that you want to assign attributes to.

", - "FindingsFilter$rulesPackageArns": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the rulesPackageArn property of the Finding data type.

", - "ListApplicationsResponse$applicationArnList": "

A list of ARNs specifying the applications returned by the action.

", - "ListAssessmentsRequest$applicationArns": "

A list of ARNs specifying the applications the assessments of which you want to list.

", - "ListAssessmentsResponse$assessmentArnList": "

A list of ARNs specifying the assessments returned by the action.

", - "ListAttachedAssessmentsResponse$assessmentArnList": "

A list of ARNs specifying the assessments returned by the action.

", - "ListAttachedRulesPackagesResponse$rulesPackageArnList": "

A list of ARNs specifying the rules packages returned by the action.

", - "ListFindingsRequest$runArns": "

The ARNs of the assessment runs that generate the findings that you want to list.

", - "ListFindingsResponse$findingArnList": "

A list of ARNs specifying the findings returned by the action.

", - "ListRulesPackagesResponse$rulesPackageArnList": "

The list of ARNs specifying the rules packages returned by the action.

", - "ListRunsRequest$assessmentArns": "

The ARNs specifying the assessments whose runs you want to list.

", - "ListRunsResponse$runArnList": "

A list of ARNs specifying the assessment runs returned by the action.

", - "RemoveAttributesFromFindingsRequest$findingArns": "

The ARNs specifying the findings that you want to remove attributes from.

", - "Run$rulesPackages": "

Rules packages selected for the run of the assessment.

", - "RunsFilter$rulesPackages": "

For a record to match a filter, the value specified for this data type property must match a list of values of the rulesPackages property of the Run data type.

" - } - }, - "Assessment": { - "base": "

Contains information about an Inspector assessment.

This data type is used as the response element in the DescribeAssessment action.

", - "refs": { - "DescribeAssessmentResponse$assessment": "

Information about the assessment.

" - } - }, - "AssessmentState": { - "base": null, - "refs": { - "Assessment$assessmentState": "

The state of the assessment. Values can be set to Created, Collecting Data, Stopping, and Completed.

", - "AssessmentStateList$member": null - } - }, - "AssessmentStateList": { - "base": null, - "refs": { - "AssessmentsFilter$assessmentStates": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the assessmentState property of the Assessment data type.

" - } - }, - "AssessmentsFilter": { - "base": "

This data type is used as the request parameter in the ListAssessments and ListAttachedAssessments actions.

", - "refs": { - "ListAssessmentsRequest$filter": "

You can use this parameter to specify a subset of data to be included in the action's response.

For a record to match a filter, all specified filter attributes must match. When multiple values are specified for a filter attribute, any of the values can match.

", - "ListAttachedAssessmentsRequest$filter": "

You can use this parameter to specify a subset of data to be included in the action's response.

For a record to match a filter, all specified filter attributes must match. When multiple values are specified for a filter attribute, any of the values can match.

" - } - }, - "AttachAssessmentAndRulesPackageRequest": { - "base": null, - "refs": { - } - }, - "AttachAssessmentAndRulesPackageResponse": { - "base": null, - "refs": { - } - }, - "Attribute": { - "base": "

This data type is used as a response element in the AddAttributesToFindings action and a request parameter in the CreateAssessment action.

", - "refs": { - "AttributeList$member": null - } - }, - "AttributeKey": { - "base": null, - "refs": { - "Attribute$key": "

The attribute key.

", - "AttributeKeyList$member": null - } - }, - "AttributeKeyList": { - "base": null, - "refs": { - "RemoveAttributesFromFindingsRequest$attributeKeys": "

The array of attribute keys that you want to remove from specified findings.

" - } - }, - "AttributeList": { - "base": null, - "refs": { - "AddAttributesToFindingsRequest$attributes": "

The array of attributes that you want to assign to specified findings.

", - "Assessment$userAttributesForFindings": "

The user-defined attributes that are assigned to every generated finding.

", - "CreateAssessmentRequest$userAttributesForFindings": "

The user-defined attributes that are assigned to every finding generated by running this assessment.

", - "Finding$attributes": "

The system-defined attributes for the finding.

", - "Finding$userAttributes": "

The user-defined attributes that are assigned to the finding.

", - "FindingsFilter$attributes": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the attributes property of the Finding data type.

", - "FindingsFilter$userAttributes": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the userAttributes property of the Finding data type.

" - } - }, - "AttributeValue": { - "base": null, - "refs": { - "Attribute$value": "

The value assigned to the attribute key.

" - } - }, - "AutoScalingGroup": { - "base": null, - "refs": { - "Agent$autoScalingGroup": "

This data type property is currently not used.

", - "AgentPreview$autoScalingGroup": "

The autoscaling group for the EC2 instance where the agent is installed.

", - "Finding$autoScalingGroup": "

The autoscaling group of the EC2 instance where the agent is installed that is used during the assessment that generates the finding.

" - } - }, - "AwsAccount": { - "base": null, - "refs": { - "Agent$accountId": "

AWS account of the EC2 instance where the agent is installed.

" - } - }, - "Bool": { - "base": null, - "refs": { - "Assessment$dataCollected": "

Boolean value (true or false) specifying whether the data collection process is completed.

", - "AssessmentsFilter$dataCollected": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the dataCollected property of the Assessment data type.

", - "DescribeCrossAccountAccessRoleResponse$valid": "

A Boolean value that specifies whether the IAM role has the necessary policies attached to enable Inspector to access your AWS account.

" - } - }, - "CreateApplicationRequest": { - "base": null, - "refs": { - } - }, - "CreateApplicationResponse": { - "base": null, - "refs": { - } - }, - "CreateAssessmentRequest": { - "base": null, - "refs": { - } - }, - "CreateAssessmentResponse": { - "base": null, - "refs": { - } - }, - "CreateResourceGroupRequest": { - "base": null, - "refs": { - } - }, - "CreateResourceGroupResponse": { - "base": null, - "refs": { - } - }, - "DeleteApplicationRequest": { - "base": null, - "refs": { - } - }, - "DeleteApplicationResponse": { - "base": null, - "refs": { - } - }, - "DeleteAssessmentRequest": { - "base": null, - "refs": { - } - }, - "DeleteAssessmentResponse": { - "base": null, - "refs": { - } - }, - "DeleteRunRequest": { - "base": null, - "refs": { - } - }, - "DeleteRunResponse": { - "base": null, - "refs": { - } - }, - "DescribeApplicationRequest": { - "base": null, - "refs": { - } - }, - "DescribeApplicationResponse": { - "base": null, - "refs": { - } - }, - "DescribeAssessmentRequest": { - "base": null, - "refs": { - } - }, - "DescribeAssessmentResponse": { - "base": null, - "refs": { - } - }, - "DescribeCrossAccountAccessRoleResponse": { - "base": null, - "refs": { - } - }, - "DescribeFindingRequest": { - "base": null, - "refs": { - } - }, - "DescribeFindingResponse": { - "base": null, - "refs": { - } - }, - "DescribeResourceGroupRequest": { - "base": null, - "refs": { - } - }, - "DescribeResourceGroupResponse": { - "base": null, - "refs": { - } - }, - "DescribeRulesPackageRequest": { - "base": null, - "refs": { - } - }, - "DescribeRulesPackageResponse": { - "base": null, - "refs": { - } - }, - "DescribeRunRequest": { - "base": null, - "refs": { - } - }, - "DescribeRunResponse": { - "base": null, - "refs": { - } - }, - "DetachAssessmentAndRulesPackageRequest": { - "base": null, - "refs": { - } - }, - "DetachAssessmentAndRulesPackageResponse": { - "base": null, - "refs": { - } - }, - "Duration": { - "base": null, - "refs": { - "Assessment$durationInSeconds": "

The assessment duration in seconds. The default value is 3600 seconds (one hour). The maximum value is 86400 seconds (one day).

", - "CreateAssessmentRequest$durationInSeconds": "

The duration of the assessment in seconds. The default value is 3600 seconds (one hour). The maximum value is 86400 seconds (one day).

", - "DurationRange$minimum": "

The minimum value of the duration range. Must be greater than zero.

", - "DurationRange$maximum": "

The maximum value of the duration range. Must be less than or equal to 604800 seconds (1 week).

", - "UpdateAssessmentRequest$durationInSeconds": "

Assessment duration in seconds that you want to update. The default value is 3600 seconds (one hour). The maximum value is 86400 seconds (one day).

" - } - }, - "DurationRange": { - "base": "

This data type is used in the AssessmentsFilter data type.

", - "refs": { - "AssessmentsFilter$durationRange": "

For a record to match a filter, the value specified for this data type property must inclusively match any value between the specified minimum and maximum values of the durationInSeconds property of the Assessment data type.

" - } - }, - "FailureMessage": { - "base": null, - "refs": { - "Assessment$failureMessage": "

This data type property is not currently used.

" - } - }, - "Finding": { - "base": "

Contains information about an Inspector finding.

This data type is used as the response element in the DescribeFinding action.

", - "refs": { - "DescribeFindingResponse$finding": "

Information about the finding.

" - } - }, - "FindingsFilter": { - "base": "

This data type is used as a request parameter in the ListFindings action.

", - "refs": { - "ListFindingsRequest$filter": "

You can use this parameter to specify a subset of data to be included in the action's response.

For a record to match a filter, all specified filter attributes must match. When multiple values are specified for a filter attribute, any of the values can match.

" - } - }, - "GetAssessmentTelemetryRequest": { - "base": null, - "refs": { - } - }, - "GetAssessmentTelemetryResponse": { - "base": null, - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "ListApplicationsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListAssessmentAgentsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListAssessmentsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListAttachedAssessmentsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListAttachedRulesPackagesRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListFindingsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListRulesPackagesRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "ListRunsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", - "PreviewAgentsForResourceGroupRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

" - } - }, - "InternalException": { - "base": null, - "refs": { - } - }, - "InvalidCrossAccountRoleException": { - "base": null, - "refs": { - } - }, - "InvalidInputException": { - "base": null, - "refs": { - } - }, - "ListApplicationsRequest": { - "base": null, - "refs": { - } - }, - "ListApplicationsResponse": { - "base": null, - "refs": { - } - }, - "ListAssessmentAgentsRequest": { - "base": null, - "refs": { - } - }, - "ListAssessmentAgentsResponse": { - "base": null, - "refs": { - } - }, - "ListAssessmentsRequest": { - "base": null, - "refs": { - } - }, - "ListAssessmentsResponse": { - "base": null, - "refs": { - } - }, - "ListAttachedAssessmentsRequest": { - "base": null, - "refs": { - } - }, - "ListAttachedAssessmentsResponse": { - "base": null, - "refs": { - } - }, - "ListAttachedRulesPackagesRequest": { - "base": null, - "refs": { - } - }, - "ListAttachedRulesPackagesResponse": { - "base": null, - "refs": { - } - }, - "ListFindingsRequest": { - "base": null, - "refs": { - } - }, - "ListFindingsResponse": { - "base": null, - "refs": { - } - }, - "ListRulesPackagesRequest": { - "base": null, - "refs": { - } - }, - "ListRulesPackagesResponse": { - "base": null, - "refs": { - } - }, - "ListRunsRequest": { - "base": null, - "refs": { - } - }, - "ListRunsResponse": { - "base": null, - "refs": { - } - }, - "ListTagsForResourceRequest": { - "base": null, - "refs": { - } - }, - "ListTagsForResourceResponse": { - "base": null, - "refs": { - } - }, - "Locale": { - "base": null, - "refs": { - "LocalizeTextRequest$locale": "

The locale that you want to translate a textual identifier into.

" - } - }, - "LocalizeTextRequest": { - "base": null, - "refs": { - } - }, - "LocalizeTextResponse": { - "base": null, - "refs": { - } - }, - "LocalizedFacility": { - "base": null, - "refs": { - "LocalizedTextKey$facility": "

The module response source of the text.

" - } - }, - "LocalizedText": { - "base": "

The textual identifier. This data type is used as the request parameter in the LocalizeText action.

", - "refs": { - "Finding$finding": "

A short description that identifies the finding.

", - "Finding$description": "

The description of the finding.

", - "Finding$recommendation": "

The recommendation for the finding.

", - "LocalizedTextList$member": null, - "RulesPackage$description": "

The description of the rules package.

" - } - }, - "LocalizedTextId": { - "base": null, - "refs": { - "LocalizedTextKey$id": "

Part of the module response source of the text.

" - } - }, - "LocalizedTextKey": { - "base": "

This data type is used in the LocalizedText data type.

", - "refs": { - "LocalizedText$key": "

The facility and id properties of the LocalizedTextKey data type.

" - } - }, - "LocalizedTextList": { - "base": null, - "refs": { - "LocalizeTextRequest$localizedTexts": "

A list of textual identifiers.

" - } - }, - "Long": { - "base": null, - "refs": { - "MessageTypeTelemetry$count": "

The number of times that the behavioral data is collected by the agent during an assessment.

", - "MessageTypeTelemetry$dataSize": "

The total size of the behavioral data that is collected by the agent during an assessment.

" - } - }, - "Message": { - "base": null, - "refs": { - "AddAttributesToFindingsResponse$message": "

Confirmation details of the action performed.

", - "AttachAssessmentAndRulesPackageResponse$message": "

Confirmation details of the action performed.

", - "DeleteApplicationResponse$message": "

Confirmation details of the action performed.

", - "DeleteAssessmentResponse$message": "

Confirmation details of the action performed.

", - "DeleteRunResponse$message": "

Confirmation details of the action performed.

", - "DetachAssessmentAndRulesPackageResponse$message": "

Confirmation details of the action performed.

", - "LocalizeTextResponse$message": "

Confirmation details of the action performed.

", - "RegisterCrossAccountAccessRoleResponse$message": "

Confirmation details of the action performed.

", - "RemoveAttributesFromFindingsResponse$message": "

Confirmation details of the action performed.

", - "SetTagsForResourceResponse$message": "

Confirmation details of the action performed.

", - "StartDataCollectionResponse$message": "

Confirmation details of the action performed.

", - "StopDataCollectionResponse$message": "

Confirmation details of the action performed.

", - "UpdateApplicationResponse$message": "

Confirmation details of the action performed.

", - "UpdateAssessmentResponse$message": "

Confirmation details of the action performed.

" - } - }, - "MessageType": { - "base": null, - "refs": { - "MessageTypeTelemetry$messageType": "

A specific type of behavioral data that is collected by the agent.

" - } - }, - "MessageTypeTelemetry": { - "base": "

This data type is used in the Telemetry data type.

This is metadata about the behavioral data collected by the Inspector agent on your EC2 instances during an assessment and passed to the Inspector service for analysis.

", - "refs": { - "MessageTypeTelemetryList$member": null - } - }, - "MessageTypeTelemetryList": { - "base": null, - "refs": { - "Telemetry$messageTypeTelemetries": "

Counts of individual metrics received by Inspector from the agent.

" - } - }, - "Name": { - "base": null, - "refs": { - "Application$applicationName": "

The name of the Inspector application.

", - "Assessment$assessmentName": "

The name of the assessment.

", - "CreateApplicationRequest$applicationName": "

The user-defined name identifying the application that you want to create. The name must be unique within the AWS account.

", - "CreateAssessmentRequest$assessmentName": "

The user-defined name identifying the assessment that you want to create. You can create several assessments for an application. The names of the assessments corresponding to a particular application must be unique.

", - "Finding$ruleName": "

The rule name that is used to generate the finding.

", - "NameList$member": null, - "RulesPackage$rulesPackageName": "

The name of the rules package.

", - "RulesPackage$provider": "

The provider of the rules package.

", - "Run$runName": "

The auto-generated name for the run.

", - "RunAssessmentRequest$runName": "

A name specifying the run of the assessment.

", - "UpdateApplicationRequest$applicationName": "

Application name that you want to update.

", - "UpdateAssessmentRequest$assessmentName": "

Assessment name that you want to update.

" - } - }, - "NameList": { - "base": null, - "refs": { - "FindingsFilter$ruleNames": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the ruleName property of the Finding data type.

" - } - }, - "NamePattern": { - "base": null, - "refs": { - "NamePatternList$member": null - } - }, - "NamePatternList": { - "base": null, - "refs": { - "ApplicationsFilter$applicationNamePatterns": "

For a record to match a filter, an explicit value or a string containing a wildcard specified for this data type property must match the value of the applicationName property of the Application data type.

", - "AssessmentsFilter$assessmentNamePatterns": "

For a record to match a filter, an explicit value or a string containing a wildcard specified for this data type property must match the value of the assessmentName property of the Assessment data type.

", - "RunsFilter$runNamePatterns": "

For a record to match a filter, an explicit value or a string containing a wildcard specified for this data type property must match the value of the runName property of the Run data type.

" - } - }, - "NoSuchEntityException": { - "base": null, - "refs": { - } - }, - "OperationInProgressException": { - "base": null, - "refs": { - } - }, - "PaginationToken": { - "base": null, - "refs": { - "ListApplicationsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListApplications action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListApplicationsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListAssessmentAgentsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListAssessmentAgents action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListAssessmentAgentsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListAssessmentsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListAssessments action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListAssessmentsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListAttachedAssessmentsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListAttachedAssessments action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListAttachedAssessmentsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListAttachedRulesPackagesRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListAttachedRulesPackages action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListAttachedRulesPackagesResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListFindingsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListFindings action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListFindingsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListRulesPackagesRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListRulesPackages action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListRulesPackagesResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "ListRunsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the ListRuns action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "ListRunsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

", - "PreviewAgentsForResourceGroupRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to 'null' on your first call to the PreviewAgentsForResourceGroup action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from previous response to continue listing data.

", - "PreviewAgentsForResourceGroupResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to 'null'.

" - } - }, - "Parameter": { - "base": "

This data type is used in the LocalizedText data type.

", - "refs": { - "ParameterList$member": null - } - }, - "ParameterList": { - "base": null, - "refs": { - "LocalizedText$parameters": "

Values for the dynamic elements of the string specified by the textual identifier.

" - } - }, - "ParameterName": { - "base": null, - "refs": { - "Parameter$name": "

The name of the variable that is being replaced.

" - } - }, - "ParameterValue": { - "base": null, - "refs": { - "Parameter$value": "

The value assigned to the variable that is being replaced.

" - } - }, - "PreviewAgentsForResourceGroupRequest": { - "base": null, - "refs": { - } - }, - "PreviewAgentsForResourceGroupResponse": { - "base": null, - "refs": { - } - }, - "RegisterCrossAccountAccessRoleRequest": { - "base": null, - "refs": { - } - }, - "RegisterCrossAccountAccessRoleResponse": { - "base": null, - "refs": { - } - }, - "RemoveAttributesFromFindingsRequest": { - "base": null, - "refs": { - } - }, - "RemoveAttributesFromFindingsResponse": { - "base": null, - "refs": { - } - }, - "ResourceGroup": { - "base": "

Contains information about a resource group. The resource group defines a set of tags that, when queried, identify the AWS resources that comprise the application.

This data type is used as the response element in the DescribeResourceGroup action.

", - "refs": { - "DescribeResourceGroupResponse$resourceGroup": "

Information about the resource group.

" - } - }, - "ResourceGroupTags": { - "base": null, - "refs": { - "CreateResourceGroupRequest$resourceGroupTags": "

A collection of keys and an array of possible values in JSON format.

For example, [{ \"key1\" : [\"Value1\",\"Value2\"]},{\"Key2\": [\"Value3\"]}]

", - "ResourceGroup$resourceGroupTags": "

The tags (key and value pairs) of the resource group.

This data type property is used in the CreateResourceGroup action.

A collection of keys and an array of possible values in JSON format.

For example, [{ \"key1\" : [\"Value1\",\"Value2\"]},{\"Key2\": [\"Value3\"]}]

" - } - }, - "RulesPackage": { - "base": "

Contains information about an Inspector rules package.

This data type is used as the response element in the DescribeRulesPackage action.

", - "refs": { - "DescribeRulesPackageResponse$rulesPackage": "

Information about the rules package.

" - } - }, - "Run": { - "base": "

A snapshot of an Inspector assessment that contains the assessment's findings.

This data type is used as the response element in the DescribeRun action.

", - "refs": { - "DescribeRunResponse$run": "

Information about the assessment run.

" - } - }, - "RunAssessmentRequest": { - "base": null, - "refs": { - } - }, - "RunAssessmentResponse": { - "base": null, - "refs": { - } - }, - "RunState": { - "base": null, - "refs": { - "Run$runState": "

The state of the run. Values can be set to DataCollectionComplete, EvaluatingPolicies, EvaluatingPoliciesErrorCanRetry, Completed, Failed, TombStoned.

", - "RunStateList$member": null - } - }, - "RunStateList": { - "base": null, - "refs": { - "RunsFilter$runStates": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the runState property of the Run data type.

" - } - }, - "RunsFilter": { - "base": "

This data type is used as the request parameter in the ListRuns action.

", - "refs": { - "ListRunsRequest$filter": "

You can use this parameter to specify a subset of data to be included in the action's response.

For a record to match a filter, all specified filter attributes must match. When multiple values are specified for a filter attribute, any of the values can match.

" - } - }, - "SetTagsForResourceRequest": { - "base": null, - "refs": { - } - }, - "SetTagsForResourceResponse": { - "base": null, - "refs": { - } - }, - "Severity": { - "base": null, - "refs": { - "Finding$severity": "

The finding severity. Values can be set to High, Medium, Low, and Informational.

", - "SeverityList$member": null - } - }, - "SeverityList": { - "base": null, - "refs": { - "FindingsFilter$severities": "

For a record to match a filter, the value specified for this data type property must be the exact match of the value of the severity property of the Finding data type.

" - } - }, - "StartDataCollectionRequest": { - "base": null, - "refs": { - } - }, - "StartDataCollectionResponse": { - "base": null, - "refs": { - } - }, - "StopDataCollectionRequest": { - "base": null, - "refs": { - } - }, - "StopDataCollectionResponse": { - "base": null, - "refs": { - } - }, - "Tag": { - "base": "

A key and value pair.

This data type is used as a request parameter in the SetTagsForResource action and a response element in the ListTagsForResource action.

", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": null, - "refs": { - "Tag$Key": "

The tag key.

" - } - }, - "TagList": { - "base": null, - "refs": { - "ListTagsForResourceResponse$tagList": "

A collection of key and value pairs.

", - "SetTagsForResourceRequest$tags": "

A collection of key and value pairs that you want to set to an assessment.

" - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

The value assigned to a tag key.

" - } - }, - "Telemetry": { - "base": "

The metadata about the Inspector application data metrics collected by the agent.

This data type is used as the response element in the GetAssessmentTelemetry action.

", - "refs": { - "TelemetryList$member": null - } - }, - "TelemetryList": { - "base": null, - "refs": { - "Agent$telemetry": "

The Inspector application data metrics collected by the agent.

", - "GetAssessmentTelemetryResponse$telemetry": "

Telemetry details.

" - } - }, - "TelemetryStatus": { - "base": null, - "refs": { - "Telemetry$status": "

The category of the individual metrics that together constitute the telemetry that Inspector received from the agent.

" - } - }, - "Text": { - "base": null, - "refs": { - "TextList$member": null - } - }, - "TextList": { - "base": null, - "refs": { - "LocalizeTextResponse$results": "

The resulting list of user-readable texts.

" - } - }, - "Timestamp": { - "base": null, - "refs": { - "Assessment$startTime": "

The assessment start time.

", - "Assessment$endTime": "

The assessment end time.

", - "Run$creationTime": "

Run creation time that corresponds to the data collection completion time or failure.

", - "Run$completionTime": "

Run completion time that corresponds to the rules packages evaluation completion time or failure.

", - "TimestampRange$minimum": "

The minimum value of the timestamp range.

", - "TimestampRange$maximum": "

The maximum value of the timestamp range.

" - } - }, - "TimestampRange": { - "base": "

This data type is used in the AssessmentsFilter and RunsFilter data types.

", - "refs": { - "AssessmentsFilter$startTimeRange": "

For a record to match a filter, the value specified for this data type property must inclusively match any value between the specified minimum and maximum values of the startTime property of the Assessment data type.

", - "AssessmentsFilter$endTimeRange": "

For a record to match a filter, the value specified for this data type property must inclusively match any value between the specified minimum and maximum values of the endTime property of the Assessment data type.

", - "RunsFilter$creationTime": "

For a record to match a filter, the value specified for this data type property must inclusively match any value between the specified minimum and maximum values of the creationTime property of the Run data type.

", - "RunsFilter$completionTime": "

For a record to match a filter, the value specified for this data type property must inclusively match any value between the specified minimum and maximum values of the completionTime property of the Run data type.

" - } - }, - "UpdateApplicationRequest": { - "base": null, - "refs": { - } - }, - "UpdateApplicationResponse": { - "base": null, - "refs": { - } - }, - "UpdateAssessmentRequest": { - "base": null, - "refs": { - } - }, - "UpdateAssessmentResponse": { - "base": null, - "refs": { - } - }, - "Version": { - "base": null, - "refs": { - "RulesPackage$version": "

The version id of the rules package.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2015-08-18/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/api-2.json index 7ffa9c16f..4cd732101 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/api-2.json @@ -39,7 +39,8 @@ {"shape":"InvalidInputException"}, {"shape":"LimitExceededException"}, {"shape":"AccessDeniedException"}, - {"shape":"NoSuchEntityException"} + {"shape":"NoSuchEntityException"}, + {"shape":"InvalidCrossAccountRoleException"} ] }, "CreateAssessmentTemplate":{ @@ -58,6 +59,22 @@ {"shape":"NoSuchEntityException"} ] }, + "CreateExclusionsPreview":{ + "name":"CreateExclusionsPreview", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"CreateExclusionsPreviewRequest"}, + "output":{"shape":"CreateExclusionsPreviewResponse"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"PreviewGenerationInProgressException"}, + {"shape":"InternalException"}, + {"shape":"AccessDeniedException"}, + {"shape":"NoSuchEntityException"} + ] + }, "CreateResourceGroup":{ "name":"CreateResourceGroup", "http":{ @@ -168,6 +185,19 @@ {"shape":"InternalException"} ] }, + "DescribeExclusions":{ + "name":"DescribeExclusions", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeExclusionsRequest"}, + "output":{"shape":"DescribeExclusionsResponse"}, + "errors":[ + {"shape":"InternalException"}, + {"shape":"InvalidInputException"} + ] + }, "DescribeFindings":{ "name":"DescribeFindings", "http":{ @@ -224,6 +254,21 @@ {"shape":"UnsupportedFeatureException"} ] }, + "GetExclusionsPreview":{ + "name":"GetExclusionsPreview", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"GetExclusionsPreviewRequest"}, + "output":{"shape":"GetExclusionsPreviewResponse"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"InternalException"}, + {"shape":"AccessDeniedException"}, + {"shape":"NoSuchEntityException"} + ] + }, "GetTelemetryMetadata":{ "name":"GetTelemetryMetadata", "http":{ @@ -313,6 +358,21 @@ {"shape":"NoSuchEntityException"} ] }, + "ListExclusions":{ + "name":"ListExclusions", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListExclusionsRequest"}, + "output":{"shape":"ListExclusionsResponse"}, + "errors":[ + {"shape":"InternalException"}, + {"shape":"InvalidInputException"}, + {"shape":"AccessDeniedException"}, + {"shape":"NoSuchEntityException"} + ] + }, "ListFindings":{ "name":"ListFindings", "http":{ @@ -864,7 +924,6 @@ "required":[ "arn", "name", - "resourceGroupArn", "createdAt", "updatedAt" ], @@ -999,13 +1058,16 @@ "max":10, "min":1 }, + "BatchDescribeExclusionsArnList":{ + "type":"list", + "member":{"shape":"Arn"}, + "max":100, + "min":1 + }, "Bool":{"type":"boolean"}, "CreateAssessmentTargetRequest":{ "type":"structure", - "required":[ - "assessmentTargetName", - "resourceGroupArn" - ], + "required":["assessmentTargetName"], "members":{ "assessmentTargetName":{"shape":"AssessmentTargetName"}, "resourceGroupArn":{"shape":"Arn"} @@ -1041,6 +1103,20 @@ "assessmentTemplateArn":{"shape":"Arn"} } }, + "CreateExclusionsPreviewRequest":{ + "type":"structure", + "required":["assessmentTemplateArn"], + "members":{ + "assessmentTemplateArn":{"shape":"Arn"} + } + }, + "CreateExclusionsPreviewResponse":{ + "type":"structure", + "required":["previewToken"], + "members":{ + "previewToken":{"shape":"UUID"} + } + }, "CreateResourceGroupRequest":{ "type":"structure", "required":["resourceGroupTags"], @@ -1143,6 +1219,25 @@ "registeredAt":{"shape":"Timestamp"} } }, + "DescribeExclusionsRequest":{ + "type":"structure", + "required":["exclusionArns"], + "members":{ + "exclusionArns":{"shape":"BatchDescribeExclusionsArnList"}, + "locale":{"shape":"Locale"} + } + }, + "DescribeExclusionsResponse":{ + "type":"structure", + "required":[ + "exclusions", + "failedItems" + ], + "members":{ + "exclusions":{"shape":"ExclusionMap"}, + "failedItems":{"shape":"FailedItems"} + } + }, "DescribeFindingsRequest":{ "type":"structure", "required":["findingArns"], @@ -1228,6 +1323,53 @@ "max":50, "min":1 }, + "Exclusion":{ + "type":"structure", + "required":[ + "arn", + "title", + "description", + "recommendation", + "scopes" + ], + "members":{ + "arn":{"shape":"Arn"}, + "title":{"shape":"Text"}, + "description":{"shape":"Text"}, + "recommendation":{"shape":"Text"}, + "scopes":{"shape":"ScopeList"}, + "attributes":{"shape":"AttributeList"} + } + }, + "ExclusionMap":{ + "type":"map", + "key":{"shape":"Arn"}, + "value":{"shape":"Exclusion"}, + "max":100, + "min":1 + }, + "ExclusionPreview":{ + "type":"structure", + "required":[ + "title", + "description", + "recommendation", + "scopes" + ], + "members":{ + "title":{"shape":"Text"}, + "description":{"shape":"Text"}, + "recommendation":{"shape":"Text"}, + "scopes":{"shape":"ScopeList"}, + "attributes":{"shape":"AttributeList"} + } + }, + "ExclusionPreviewList":{ + "type":"list", + "member":{"shape":"ExclusionPreview"}, + "max":100, + "min":0 + }, "FailedItemDetails":{ "type":"structure", "required":[ @@ -1337,6 +1479,29 @@ "url":{"shape":"Url"} } }, + "GetExclusionsPreviewRequest":{ + "type":"structure", + "required":[ + "assessmentTemplateArn", + "previewToken" + ], + "members":{ + "assessmentTemplateArn":{"shape":"Arn"}, + "previewToken":{"shape":"UUID"}, + "nextToken":{"shape":"PaginationToken"}, + "maxResults":{"shape":"ListMaxResults"}, + "locale":{"shape":"Locale"} + } + }, + "GetExclusionsPreviewResponse":{ + "type":"structure", + "required":["previewStatus"], + "members":{ + "previewStatus":{"shape":"PreviewStatus"}, + "exclusionPreviews":{"shape":"ExclusionPreviewList"}, + "nextToken":{"shape":"PaginationToken"} + } + }, "GetTelemetryMetadataRequest":{ "type":"structure", "required":["assessmentRunArn"], @@ -1612,6 +1777,23 @@ "nextToken":{"shape":"PaginationToken"} } }, + "ListExclusionsRequest":{ + "type":"structure", + "required":["assessmentRunArn"], + "members":{ + "assessmentRunArn":{"shape":"Arn"}, + "nextToken":{"shape":"PaginationToken"}, + "maxResults":{"shape":"ListMaxResults"} + } + }, + "ListExclusionsResponse":{ + "type":"structure", + "required":["exclusionArns"], + "members":{ + "exclusionArns":{"shape":"ListReturnedArnList"}, + "nextToken":{"shape":"PaginationToken"} + } + }, "ListFindingsRequest":{ "type":"structure", "members":{ @@ -1755,6 +1937,21 @@ "nextToken":{"shape":"PaginationToken"} } }, + "PreviewGenerationInProgressException":{ + "type":"structure", + "required":["message"], + "members":{ + "message":{"shape":"ErrorMessage"} + }, + "exception":true + }, + "PreviewStatus":{ + "type":"string", + "enum":[ + "WORK_IN_PROGRESS", + "COMPLETED" + ] + }, "ProviderName":{ "type":"string", "max":1000, @@ -1877,6 +2074,26 @@ "max":1000, "min":0 }, + "Scope":{ + "type":"structure", + "members":{ + "key":{"shape":"ScopeType"}, + "value":{"shape":"ScopeValue"} + } + }, + "ScopeList":{ + "type":"list", + "member":{"shape":"Scope"}, + "min":1 + }, + "ScopeType":{ + "type":"string", + "enum":[ + "INSTANCE_ID", + "RULES_PACKAGE_ARN" + ] + }, + "ScopeValue":{"type":"string"}, "ServiceName":{ "type":"string", "max":128, @@ -2023,6 +2240,10 @@ "endDate":{"shape":"Timestamp"} } }, + "UUID":{ + "type":"string", + "pattern":"[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}" + }, "UnsubscribeFromEventRequest":{ "type":"structure", "required":[ @@ -2052,8 +2273,7 @@ "type":"structure", "required":[ "assessmentTargetArn", - "assessmentTargetName", - "resourceGroupArn" + "assessmentTargetName" ], "members":{ "assessmentTargetArn":{"shape":"Arn"}, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/docs-2.json index 831a3851c..6bd76e85f 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/docs-2.json @@ -5,6 +5,7 @@ "AddAttributesToFindings": "

Assigns attributes (key and value pairs) to the findings that are specified by the ARNs of the findings.

", "CreateAssessmentTarget": "

Creates a new assessment target using the ARN of the resource group that is generated by CreateResourceGroup. If the service-linked role isn’t already registered, also creates and registers a service-linked role to grant Amazon Inspector access to AWS Services needed to perform security assessments. You can create up to 50 assessment targets per AWS account. You can run up to 500 concurrent agents per AWS account. For more information, see Amazon Inspector Assessment Targets.

", "CreateAssessmentTemplate": "

Creates an assessment template for the assessment target that is specified by the ARN of the assessment target. If the service-linked role isn’t already registered, also creates and registers a service-linked role to grant Amazon Inspector access to AWS Services needed to perform security assessments.

", + "CreateExclusionsPreview": "

Starts the generation of an exclusions preview for the specified assessment template. The exclusions preview lists the potential exclusions (ExclusionPreview) that Inspector can detect before it runs the assessment.

", "CreateResourceGroup": "

Creates a resource group using the specified set of tags (key and value pairs) that are used to select the EC2 instances to be included in an Amazon Inspector assessment target. The created resource group is then used to create an Amazon Inspector assessment target. For more information, see CreateAssessmentTarget.

", "DeleteAssessmentRun": "

Deletes the assessment run that is specified by the ARN of the assessment run.

", "DeleteAssessmentTarget": "

Deletes the assessment target that is specified by the ARN of the assessment target.

", @@ -13,16 +14,19 @@ "DescribeAssessmentTargets": "

Describes the assessment targets that are specified by the ARNs of the assessment targets.

", "DescribeAssessmentTemplates": "

Describes the assessment templates that are specified by the ARNs of the assessment templates.

", "DescribeCrossAccountAccessRole": "

Describes the IAM role that enables Amazon Inspector to access your AWS account.

", + "DescribeExclusions": "

Describes the exclusions that are specified by the exclusions' ARNs.

", "DescribeFindings": "

Describes the findings that are specified by the ARNs of the findings.

", "DescribeResourceGroups": "

Describes the resource groups that are specified by the ARNs of the resource groups.

", "DescribeRulesPackages": "

Describes the rules packages that are specified by the ARNs of the rules packages.

", "GetAssessmentReport": "

Produces an assessment report that includes detailed and comprehensive results of a specified assessment run.

", + "GetExclusionsPreview": "

Retrieves the exclusions preview (a list of ExclusionPreview objects) specified by the preview token. You can obtain the preview token by running the CreateExclusionsPreview API.

", "GetTelemetryMetadata": "

Information about the data that is collected for the specified assessment run.

", "ListAssessmentRunAgents": "

Lists the agents of the assessment runs that are specified by the ARNs of the assessment runs.

", "ListAssessmentRuns": "

Lists the assessment runs that correspond to the assessment templates that are specified by the ARNs of the assessment templates.

", "ListAssessmentTargets": "

Lists the ARNs of the assessment targets within this AWS account. For more information about assessment targets, see Amazon Inspector Assessment Targets.

", "ListAssessmentTemplates": "

Lists the assessment templates that correspond to the assessment targets that are specified by the ARNs of the assessment targets.

", "ListEventSubscriptions": "

Lists all the event subscriptions for the assessment template that is specified by the ARN of the assessment template. For more information, see SubscribeToEvent and UnsubscribeFromEvent.

", + "ListExclusions": "

List exclusions that are generated by the assessment run.

", "ListFindings": "

Lists findings that are generated by the assessment runs that are specified by the ARNs of the assessment runs.

", "ListRulesPackages": "

Lists all available Amazon Inspector rules packages.

", "ListTagsForResource": "

Lists all tags associated with an assessment template.

", @@ -170,27 +174,33 @@ "AssessmentTarget$resourceGroupArn": "

The ARN that specifies the resource group that is associated with the assessment target.

", "AssessmentTemplate$arn": "

The ARN of the assessment template.

", "AssessmentTemplate$assessmentTargetArn": "

The ARN of the assessment target that corresponds to this assessment template.

", - "AssessmentTemplate$lastAssessmentRunArn": "

The Amazon Resource Name (ARN) of the most recent assessment run associated with this assessment template. This value exists only when the value of assessmentRunCount is greater than zero.

", + "AssessmentTemplate$lastAssessmentRunArn": "

The Amazon Resource Name (ARN) of the most recent assessment run associated with this assessment template. This value exists only when the value of assessmentRunCount is greaterpa than zero.

", "AssessmentTemplateRulesPackageArnList$member": null, "BatchDescribeArnList$member": null, + "BatchDescribeExclusionsArnList$member": null, "CreateAssessmentTargetRequest$resourceGroupArn": "

The ARN that specifies the resource group that is used to create the assessment target.

", "CreateAssessmentTargetResponse$assessmentTargetArn": "

The ARN that specifies the assessment target that is created.

", "CreateAssessmentTemplateRequest$assessmentTargetArn": "

The ARN that specifies the assessment target for which you want to create the assessment template.

", "CreateAssessmentTemplateResponse$assessmentTemplateArn": "

The ARN that specifies the assessment template that is created.

", + "CreateExclusionsPreviewRequest$assessmentTemplateArn": "

The ARN that specifies the assessment template for which you want to create an exclusions preview.

", "CreateResourceGroupResponse$resourceGroupArn": "

The ARN that specifies the resource group that is created.

", "DeleteAssessmentRunRequest$assessmentRunArn": "

The ARN that specifies the assessment run that you want to delete.

", "DeleteAssessmentTargetRequest$assessmentTargetArn": "

The ARN that specifies the assessment target that you want to delete.

", "DeleteAssessmentTemplateRequest$assessmentTemplateArn": "

The ARN that specifies the assessment template that you want to delete.

", "DescribeCrossAccountAccessRoleResponse$roleArn": "

The ARN that specifies the IAM role that Amazon Inspector uses to access your AWS account.

", + "Exclusion$arn": "

The ARN that specifies the exclusion.

", + "ExclusionMap$key": null, "FailedItems$key": null, "FilterRulesPackageArnList$member": null, "Finding$arn": "

The ARN that specifies the finding.

", "GetAssessmentReportRequest$assessmentRunArn": "

The ARN that specifies the assessment run for which you want to generate a report.

", + "GetExclusionsPreviewRequest$assessmentTemplateArn": "

The ARN that specifies the assessment template for which the exclusions preview was requested.

", "GetTelemetryMetadataRequest$assessmentRunArn": "

The ARN that specifies the assessment run that has the telemetry data that you want to obtain.

", "InspectorServiceAttributes$assessmentRunArn": "

The ARN of the assessment run during which the finding is generated.

", "InspectorServiceAttributes$rulesPackageArn": "

The ARN of the rules package that is used to generate the finding.

", "ListAssessmentRunAgentsRequest$assessmentRunArn": "

The ARN that specifies the assessment run whose agents you want to list.

", "ListEventSubscriptionsRequest$resourceArn": "

The ARN of the assessment template for which you want to list the existing event subscriptions.

", + "ListExclusionsRequest$assessmentRunArn": "

The ARN of the assessment run that generated the exclusions that you want to list.

", "ListParentArnList$member": null, "ListReturnedArnList$member": null, "ListTagsForResourceRequest$resourceArn": "

The ARN that specifies the assessment template whose tags you want to list.

", @@ -246,7 +256,7 @@ "base": null, "refs": { "AssessmentRun$durationInSeconds": "

The duration of the assessment run.

", - "AssessmentTemplate$durationInSeconds": "

The duration in seconds specified for this assessment tempate. The default value is 3600 seconds (one hour). The maximum value is 86400 seconds (one day).

", + "AssessmentTemplate$durationInSeconds": "

The duration in seconds specified for this assessment template. The default value is 3600 seconds (one hour). The maximum value is 86400 seconds (one day).

", "CreateAssessmentTemplateRequest$durationInSeconds": "

The duration of the assessment run in seconds. The default value is 3600 seconds (one hour).

", "DurationRange$minSeconds": "

The minimum value of the duration range. Must be greater than zero.

", "DurationRange$maxSeconds": "

The maximum value of the duration range. Must be less than or equal to 604800 seconds (1 week).

" @@ -419,6 +429,8 @@ "AttributeList": { "base": null, "refs": { + "Exclusion$attributes": "

The system-defined attributes for the exclusion.

", + "ExclusionPreview$attributes": "

The system-defined attributes for the exclusion preview.

", "Finding$attributes": "

The system-defined attributes for the finding.

", "FindingFilter$attributes": "

For a record to match a filter, the list of values that are specified for this data type property must be contained in the list of values of the attributes property of the Finding data type.

", "FindingFilter$userAttributes": "

For a record to match a filter, the value that is specified for this data type property must be contained in the list of values of the userAttributes property of the Finding data type.

" @@ -456,6 +468,12 @@ "DescribeRulesPackagesRequest$rulesPackageArns": "

The ARN that specifies the rules package that you want to describe.

" } }, + "BatchDescribeExclusionsArnList": { + "base": null, + "refs": { + "DescribeExclusionsRequest$exclusionArns": "

The list of ARNs that specify the exclusions that you want to describe.

" + } + }, "Bool": { "base": null, "refs": { @@ -497,6 +515,16 @@ "refs": { } }, + "CreateExclusionsPreviewRequest": { + "base": null, + "refs": { + } + }, + "CreateExclusionsPreviewResponse": { + "base": null, + "refs": { + } + }, "CreateResourceGroupRequest": { "base": null, "refs": { @@ -557,6 +585,16 @@ "refs": { } }, + "DescribeExclusionsRequest": { + "base": null, + "refs": { + } + }, + "DescribeExclusionsResponse": { + "base": null, + "refs": { + } + }, "DescribeFindingsRequest": { "base": null, "refs": { @@ -605,6 +643,7 @@ "InvalidInputException$message": "

Details of the exception error.

", "LimitExceededException$message": "

Details of the exception error.

", "NoSuchEntityException$message": "

Details of the exception error.

", + "PreviewGenerationInProgressException$message": null, "UnsupportedFeatureException$message": null } }, @@ -620,6 +659,30 @@ "Subscription$eventSubscriptions": "

The list of existing event subscriptions.

" } }, + "Exclusion": { + "base": "

Contains information about what was excluded from an assessment run.

", + "refs": { + "ExclusionMap$value": null + } + }, + "ExclusionMap": { + "base": null, + "refs": { + "DescribeExclusionsResponse$exclusions": "

Information about the exclusions.

" + } + }, + "ExclusionPreview": { + "base": "

Contains information about what is excluded from an assessment run given the current state of the assessment template.

", + "refs": { + "ExclusionPreviewList$member": null + } + }, + "ExclusionPreviewList": { + "base": null, + "refs": { + "GetExclusionsPreviewResponse$exclusionPreviews": "

Information about the exclusions included in the preview.

" + } + }, "FailedItemDetails": { "base": "

Includes details about the failed items.

", "refs": { @@ -639,6 +702,7 @@ "DescribeAssessmentRunsResponse$failedItems": "

Assessment run details that cannot be described. An error code is provided for each failed item.

", "DescribeAssessmentTargetsResponse$failedItems": "

Assessment target details that cannot be described. An error code is provided for each failed item.

", "DescribeAssessmentTemplatesResponse$failedItems": "

Assessment template details that cannot be described. An error code is provided for each failed item.

", + "DescribeExclusionsResponse$failedItems": "

Exclusion details that cannot be described. An error code is provided for each failed item.

", "DescribeFindingsResponse$failedItems": "

Finding details that cannot be described. An error code is provided for each failed item.

", "DescribeResourceGroupsResponse$failedItems": "

Resource group details that cannot be described. An error code is provided for each failed item.

", "DescribeRulesPackagesResponse$failedItems": "

Rules package details that cannot be described. An error code is provided for each failed item.

", @@ -693,6 +757,16 @@ "refs": { } }, + "GetExclusionsPreviewRequest": { + "base": null, + "refs": { + } + }, + "GetExclusionsPreviewResponse": { + "base": null, + "refs": { + } + }, "GetTelemetryMetadataRequest": { "base": null, "refs": { @@ -844,6 +918,16 @@ "refs": { } }, + "ListExclusionsRequest": { + "base": null, + "refs": { + } + }, + "ListExclusionsResponse": { + "base": null, + "refs": { + } + }, "ListFindingsRequest": { "base": null, "refs": { @@ -857,10 +941,12 @@ "ListMaxResults": { "base": null, "refs": { + "GetExclusionsPreviewRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 100. The maximum value is 500.

", "ListAssessmentRunAgentsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items that you want in the response. The default value is 10. The maximum value is 500.

", "ListAssessmentRunsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items that you want in the response. The default value is 10. The maximum value is 500.

", "ListAssessmentTargetsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", "ListAssessmentTemplatesRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", + "ListExclusionsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 100. The maximum value is 500.

", "ListFindingsRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

", "ListRulesPackagesRequest$maxResults": "

You can use this parameter to indicate the maximum number of items you want in the response. The default value is 10. The maximum value is 500.

" } @@ -879,6 +965,7 @@ "ListAssessmentRunsResponse$assessmentRunArns": "

A list of ARNs that specifies the assessment runs that are returned by the action.

", "ListAssessmentTargetsResponse$assessmentTargetArns": "

A list of ARNs that specifies the assessment targets that are returned by the action.

", "ListAssessmentTemplatesResponse$assessmentTemplateArns": "

A list of ARNs that specifies the assessment templates returned by the action.

", + "ListExclusionsResponse$exclusionArns": "

A list of exclusions' ARNs returned by the action.

", "ListFindingsResponse$findingArns": "

A list of ARNs that specifies the findings returned by the action.

", "ListRulesPackagesResponse$rulesPackageArns": "

The list of ARNs that specifies the rules packages returned by the action.

" } @@ -906,8 +993,10 @@ "Locale": { "base": null, "refs": { + "DescribeExclusionsRequest$locale": "

The locale into which you want to translate the exclusion's title, description, and recommendation.

", "DescribeFindingsRequest$locale": "

The locale into which you want to translate a finding description, recommendation, and the short description that identifies the finding.

", - "DescribeRulesPackagesRequest$locale": "

The locale that you want to translate a rules package description into.

" + "DescribeRulesPackagesRequest$locale": "

The locale that you want to translate a rules package description into.

", + "GetExclusionsPreviewRequest$locale": "

The locale into which you want to translate the exclusion's title, description, and recommendation.

" } }, "Long": { @@ -972,6 +1061,8 @@ "PaginationToken": { "base": null, "refs": { + "GetExclusionsPreviewRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the GetExclusionsPreviewRequest action. Subsequent calls to the action fill nextToken in the request with the value of nextToken from the previous response to continue listing data.

", + "GetExclusionsPreviewResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameters is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", "ListAssessmentRunAgentsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListAssessmentRunAgents action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from the previous response to continue listing data.

", "ListAssessmentRunAgentsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", "ListAssessmentRunsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListAssessmentRuns action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from the previous response to continue listing data.

", @@ -982,6 +1073,8 @@ "ListAssessmentTemplatesResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", "ListEventSubscriptionsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListEventSubscriptions action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from the previous response to continue listing data.

", "ListEventSubscriptionsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", + "ListExclusionsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListExclusionsRequest action. Subsequent calls to the action fill nextToken in the request with the value of nextToken from the previous response to continue listing data.

", + "ListExclusionsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameters is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", "ListFindingsRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListFindings action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from the previous response to continue listing data.

", "ListFindingsResponse$nextToken": "

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", "ListRulesPackagesRequest$nextToken": "

You can use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListRulesPackages action. Subsequent calls to the action fill nextToken in the request with the value of NextToken from the previous response to continue listing data.

", @@ -1006,6 +1099,17 @@ "refs": { } }, + "PreviewGenerationInProgressException": { + "base": "

The request is rejected. The specified assessment template is currently generating an exclusions preview.

", + "refs": { + } + }, + "PreviewStatus": { + "base": null, + "refs": { + "GetExclusionsPreviewResponse$previewStatus": "

Specifies the status of the request to generate an exclusions preview.

" + } + }, "ProviderName": { "base": null, "refs": { @@ -1100,6 +1204,31 @@ "RulesPackage$name": "

The name of the rules package.

" } }, + "Scope": { + "base": "

This data type contains key-value pairs that identify various Amazon resources.

", + "refs": { + "ScopeList$member": null + } + }, + "ScopeList": { + "base": null, + "refs": { + "Exclusion$scopes": "

The AWS resources for which the exclusion pertains.

", + "ExclusionPreview$scopes": "

The AWS resources for which the exclusion preview pertains.

" + } + }, + "ScopeType": { + "base": null, + "refs": { + "Scope$key": "

The type of the scope.

" + } + }, + "ScopeValue": { + "base": null, + "refs": { + "Scope$value": "

The resource identifier for the specified scope type.

" + } + }, "ServiceName": { "base": null, "refs": { @@ -1206,6 +1335,12 @@ "Text": { "base": null, "refs": { + "Exclusion$title": "

The name of the exclusion.

", + "Exclusion$description": "

The description of the exclusion.

", + "Exclusion$recommendation": "

The recommendation for the exclusion.

", + "ExclusionPreview$title": "

The name of the exclusion preview.

", + "ExclusionPreview$description": "

The description of the exclusion preview.

", + "ExclusionPreview$recommendation": "

The recommendation for the exclusion preview.

", "Finding$title": "

The name of the finding.

", "Finding$description": "

The description of the finding.

", "Finding$recommendation": "

The recommendation for the finding.

", @@ -1242,6 +1377,13 @@ "FindingFilter$creationTimeRange": "

The time range during which the finding is generated.

" } }, + "UUID": { + "base": null, + "refs": { + "CreateExclusionsPreviewResponse$previewToken": "

Specifies the unique identifier of the requested exclusions preview. You can use the unique identifier to retrieve the exclusions preview when running the GetExclusionsPreview API.

", + "GetExclusionsPreviewRequest$previewToken": "

The unique identifier associated of the exclusions preview.

" + } + }, "UnsubscribeFromEventRequest": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/paginators-1.json index 0f2aaeeb2..e702bf0d8 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/inspector/2016-02-16/paginators-1.json @@ -1,5 +1,10 @@ { "pagination": { + "GetExclusionsPreview": { + "input_token": "nextToken", + "output_token": "nextToken", + "limit_key": "maxResults" + }, "ListAssessmentRunAgents": { "input_token": "nextToken", "output_token": "nextToken", @@ -25,6 +30,11 @@ "output_token": "nextToken", "limit_key": "maxResults" }, + "ListExclusions": { + "input_token": "nextToken", + "output_token": "nextToken", + "limit_key": "maxResults" + }, "ListFindings": { "input_token": "nextToken", "output_token": "nextToken", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/api-2.json index 81fd362e2..eb9aa90ec 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/api-2.json @@ -143,6 +143,22 @@ {"shape":"ServiceUnavailableException"} ] }, + "CancelJobExecution":{ + "name":"CancelJobExecution", + "http":{ + "method":"PUT", + "requestUri":"/things/{thingName}/jobs/{jobId}/cancel" + }, + "input":{"shape":"CancelJobExecutionRequest"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"InvalidStateTransitionException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"ThrottlingException"}, + {"shape":"ServiceUnavailableException"}, + {"shape":"VersionConflictException"} + ] + }, "ClearDefaultAuthorizer":{ "name":"ClearDefaultAuthorizer", "http":{ @@ -437,6 +453,37 @@ {"shape":"ResourceNotFoundException"} ] }, + "DeleteJob":{ + "name":"DeleteJob", + "http":{ + "method":"DELETE", + "requestUri":"/jobs/{jobId}" + }, + "input":{"shape":"DeleteJobRequest"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"InvalidStateTransitionException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"LimitExceededException"}, + {"shape":"ThrottlingException"}, + {"shape":"ServiceUnavailableException"} + ] + }, + "DeleteJobExecution":{ + "name":"DeleteJobExecution", + "http":{ + "method":"DELETE", + "requestUri":"/things/{thingName}/jobs/{jobId}/executionNumber/{executionNumber}" + }, + "input":{"shape":"DeleteJobExecutionRequest"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"InvalidStateTransitionException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"ThrottlingException"}, + {"shape":"ServiceUnavailableException"} + ] + }, "DeleteOTAUpdate":{ "name":"DeleteOTAUpdate", "http":{ @@ -2348,6 +2395,32 @@ } } }, + "CancelJobExecutionRequest":{ + "type":"structure", + "required":[ + "jobId", + "thingName" + ], + "members":{ + "jobId":{ + "shape":"JobId", + "location":"uri", + "locationName":"jobId" + }, + "thingName":{ + "shape":"ThingName", + "location":"uri", + "locationName":"thingName" + }, + "force":{ + "shape":"ForceFlag", + "location":"querystring", + "locationName":"force" + }, + "expectedVersion":{"shape":"ExpectedVersion"}, + "statusDetails":{"shape":"DetailsMap"} + } + }, "CancelJobRequest":{ "type":"structure", "required":["jobId"], @@ -2357,7 +2430,12 @@ "location":"uri", "locationName":"jobId" }, - "comment":{"shape":"Comment"} + "comment":{"shape":"Comment"}, + "force":{ + "shape":"ForceFlag", + "location":"querystring", + "locationName":"force" + } } }, "CancelJobResponse":{ @@ -2939,6 +3017,52 @@ "error":{"httpStatusCode":409}, "exception":true }, + "DeleteJobExecutionRequest":{ + "type":"structure", + "required":[ + "jobId", + "thingName", + "executionNumber" + ], + "members":{ + "jobId":{ + "shape":"JobId", + "location":"uri", + "locationName":"jobId" + }, + "thingName":{ + "shape":"ThingName", + "location":"uri", + "locationName":"thingName" + }, + "executionNumber":{ + "shape":"ExecutionNumber", + "location":"uri", + "locationName":"executionNumber" + }, + "force":{ + "shape":"ForceFlag", + "location":"querystring", + "locationName":"force" + } + } + }, + "DeleteJobRequest":{ + "type":"structure", + "required":["jobId"], + "members":{ + "jobId":{ + "shape":"JobId", + "location":"uri", + "locationName":"jobId" + }, + "force":{ + "shape":"ForceFlag", + "location":"querystring", + "locationName":"force" + } + } + }, "DeleteOTAUpdateRequest":{ "type":"structure", "required":["otaUpdateId"], @@ -3632,6 +3756,7 @@ ] }, "ExecutionNumber":{"type":"long"}, + "ExpectedVersion":{"type":"long"}, "ExpiresInSec":{ "type":"long", "max":3600, @@ -3668,6 +3793,8 @@ }, "Flag":{"type":"boolean"}, "ForceDelete":{"type":"boolean"}, + "ForceFlag":{"type":"boolean"}, + "Forced":{"type":"boolean"}, "FunctionArn":{"type":"string"}, "GEMaxResults":{ "type":"integer", @@ -3934,6 +4061,14 @@ "error":{"httpStatusCode":400}, "exception":true }, + "InvalidStateTransitionException":{ + "type":"structure", + "members":{ + "message":{"shape":"errorMessage"} + }, + "error":{"httpStatusCode":409}, + "exception":true + }, "IotAnalyticsAction":{ "type":"structure", "members":{ @@ -3952,6 +4087,7 @@ "jobId":{"shape":"JobId"}, "targetSelection":{"shape":"TargetSelection"}, "status":{"shape":"JobStatus"}, + "forceCanceled":{"shape":"Forced"}, "comment":{"shape":"Comment"}, "targets":{"shape":"JobTargets"}, "description":{"shape":"JobDescription"}, @@ -3990,12 +4126,14 @@ "members":{ "jobId":{"shape":"JobId"}, "status":{"shape":"JobExecutionStatus"}, + "forceCanceled":{"shape":"Forced"}, "statusDetails":{"shape":"JobExecutionStatusDetails"}, "thingArn":{"shape":"ThingArn"}, "queuedAt":{"shape":"DateType"}, "startedAt":{"shape":"DateType"}, "lastUpdatedAt":{"shape":"DateType"}, - "executionNumber":{"shape":"ExecutionNumber"} + "executionNumber":{"shape":"ExecutionNumber"}, + "versionNumber":{"shape":"VersionNumber"} } }, "JobExecutionStatus":{ @@ -4078,7 +4216,8 @@ "enum":[ "IN_PROGRESS", "CANCELED", - "COMPLETED" + "COMPLETED", + "DELETION_IN_PROGRESS" ] }, "JobSummary":{ @@ -6428,6 +6567,7 @@ "error":{"httpStatusCode":409}, "exception":true }, + "VersionNumber":{"type":"long"}, "VersionsLimitExceededException":{ "type":"structure", "members":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/docs-2.json index 8fcc3cf50..98fa8a570 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/iot/2015-05-28/docs-2.json @@ -10,6 +10,7 @@ "AttachThingPrincipal": "

Attaches the specified principal to the specified thing.

", "CancelCertificateTransfer": "

Cancels a pending transfer for the specified certificate.

Note Only the transfer source account can use this operation to cancel a transfer. (Transfer destinations can use RejectCertificateTransfer instead.) After transfer, AWS IoT returns the certificate to the source account in the INACTIVE state. After the destination account has accepted the transfer, the transfer cannot be cancelled.

After a certificate transfer is cancelled, the status of the certificate changes from PENDING_TRANSFER to INACTIVE.

", "CancelJob": "

Cancels a job.

", + "CancelJobExecution": "

Cancels the execution of a job for a given thing.

", "ClearDefaultAuthorizer": "

Clears the default authorizer.

", "CreateAuthorizer": "

Creates an authorizer.

", "CreateCertificateFromCsr": "

Creates an X.509 certificate using the specified certificate signing request.

Note: The CSR must include a public key that is either an RSA key with a length of at least 2048 bits or an ECC key from NIST P-256 or NIST P-384 curves.

Note: Reusing the same certificate signing request (CSR) results in a distinct certificate.

You can create multiple certificates in a batch by creating a directory, copying multiple .csr files into that directory, and then specifying that directory on the command line. The following commands show how to create a batch of certificates given a batch of CSRs.

Assuming a set of CSRs are located inside of the directory my-csr-directory:

On Linux and OS X, the command is:

$ ls my-csr-directory/ | xargs -I {} aws iot create-certificate-from-csr --certificate-signing-request file://my-csr-directory/{}

This command lists all of the CSRs in my-csr-directory and pipes each CSR file name to the aws iot create-certificate-from-csr AWS CLI command to create a certificate for the corresponding CSR.

The aws iot create-certificate-from-csr part of the command can also be run in parallel to speed up the certificate creation process:

$ ls my-csr-directory/ | xargs -P 10 -I {} aws iot create-certificate-from-csr --certificate-signing-request file://my-csr-directory/{}

On Windows PowerShell, the command to create certificates for all CSRs in my-csr-directory is:

> ls -Name my-csr-directory | %{aws iot create-certificate-from-csr --certificate-signing-request file://my-csr-directory/$_}

On a Windows command prompt, the command to create certificates for all CSRs in my-csr-directory is:

> forfiles /p my-csr-directory /c \"cmd /c aws iot create-certificate-from-csr --certificate-signing-request file://@path\"

", @@ -27,6 +28,8 @@ "DeleteAuthorizer": "

Deletes an authorizer.

", "DeleteCACertificate": "

Deletes a registered CA certificate.

", "DeleteCertificate": "

Deletes the specified certificate.

A certificate cannot be deleted if it has a policy attached to it or if its status is set to ACTIVE. To delete a certificate, first use the DetachPrincipalPolicy API to detach all policies. Next, use the UpdateCertificate API to set the certificate to the INACTIVE status.

", + "DeleteJob": "

Deletes a job and its related job executions.

Deleting a job may take time, depending on the number of job executions created for the job and various other factors. While the job is being deleted, the status of the job will be shown as \"DELETION_IN_PROGRESS\". Attempting to delete or cancel a job whose status is already \"DELETION_IN_PROGRESS\" will result in an error.

Only 10 jobs may have status \"DELETION_IN_PROGRESS\" at the same time, or a LimitExceededException will occur.

", + "DeleteJobExecution": "

Deletes a job execution.

", "DeleteOTAUpdate": "

Delete an OTA update.

", "DeletePolicy": "

Deletes the specified policy.

A policy cannot be deleted if it has non-default versions or it is attached to any certificate.

To delete a policy, use the DeletePolicyVersion API to delete all non-default versions of the policy; use the DetachPrincipalPolicy API to detach the policy from any certificate; and then use the DeletePolicy API to delete the policy.

When a policy is deleted using DeletePolicy, its default version is deleted with it.

", "DeletePolicyVersion": "

Deletes the specified version of the specified policy. You cannot delete the default version of a policy using this API. To delete the default version of a policy, use DeletePolicy. To find out which version of a policy is marked as the default version, use ListPolicyVersions.

", @@ -59,7 +62,7 @@ "DetachThingPrincipal": "

Detaches the specified principal from the specified thing.

", "DisableTopicRule": "

Disables the rule.

", "EnableTopicRule": "

Enables the rule.

", - "GetEffectivePolicies": "

Gets effective policies.

", + "GetEffectivePolicies": "

Gets a list of the policies that have an effect on the authorization behavior of the specified device when it connects to the AWS IoT device gateway.

", "GetIndexingConfiguration": "

Gets the search configuration.

", "GetJobDocument": "

Gets a job document.

", "GetLoggingOptions": "

Gets the logging options.

", @@ -112,8 +115,8 @@ "SetV2LoggingOptions": "

Sets the logging options for the V2 logging service.

", "StartThingRegistrationTask": "

Creates a bulk thing provisioning task.

", "StopThingRegistrationTask": "

Cancels a bulk thing provisioning task.

", - "TestAuthorization": "

Test custom authorization.

", - "TestInvokeAuthorizer": "

Invoke the specified custom authorizer for testing purposes.

", + "TestAuthorization": "

Tests if a specified principal is authorized to perform an AWS IoT action on a specified resource. Use this to test and debug the authorization behavior of devices that connect to the AWS IoT device gateway.

", + "TestInvokeAuthorizer": "

Tests a custom authorization behavior by invoking a specified custom authorizer. Use this to test and debug the custom authorization behavior of devices that connect to the AWS IoT device gateway.

", "TransferCertificate": "

Transfers the specified certificate to the specified AWS account.

You can cancel the transfer until it is acknowledged by the recipient.

No notification is sent to the transfer destination's account. It is up to the caller to notify the transfer target.

The certificate being transferred must not be in the ACTIVE state. You can use the UpdateCertificate API to deactivate it.

The certificate must not have any policies attached to it. You can use the DetachPrincipalPolicy API to detach them.

", "UpdateAuthorizer": "

Updates an authorizer.

", "UpdateCACertificate": "

Updates a registered CA certificate.

", @@ -394,7 +397,7 @@ "GetLoggingOptionsResponse$roleArn": "

The ARN of the IAM role that grants access.

", "GetV2LoggingOptionsResponse$roleArn": "

The IAM role ARN AWS IoT uses to write to your CloudWatch logs.

", "IotAnalyticsAction$channelArn": "

(deprecated) The ARN of the IoT Analytics channel to which message data will be sent.

", - "IotAnalyticsAction$roleArn": "

The ARN of the role which has a policy that grants IoT permission to send message data via IoT Analytics (iotanalytics:BatchPutMessage).

", + "IotAnalyticsAction$roleArn": "

The ARN of the role which has a policy that grants IoT Analytics permission to send message data via IoT Analytics (iotanalytics:BatchPutMessage).

", "KinesisAction$roleArn": "

The ARN of the IAM role that grants access to the Amazon Kinesis stream.

", "LoggingOptionsPayload$roleArn": "

The ARN of the IAM role that grants access.

", "RepublishAction$roleArn": "

The ARN of the IAM role that grants access.

", @@ -469,6 +472,11 @@ "refs": { } }, + "CancelJobExecutionRequest": { + "base": null, + "refs": { + } + }, "CancelJobRequest": { "base": null, "refs": { @@ -927,6 +935,16 @@ "refs": { } }, + "DeleteJobExecutionRequest": { + "base": null, + "refs": { + } + }, + "DeleteJobRequest": { + "base": null, + "refs": { + } + }, "DeleteOTAUpdateRequest": { "base": null, "refs": { @@ -1231,6 +1249,7 @@ "DetailsMap": { "base": null, "refs": { + "CancelJobExecutionRequest$statusDetails": "

A collection of name/value pairs that describe the status of the job execution. If not specified, the statusDetails are unchanged. You can specify at most 10 name/value pairs.

", "JobExecutionStatusDetails$detailsMap": "

The job execution status.

" } }, @@ -1370,11 +1389,18 @@ "ExecutionNumber": { "base": null, "refs": { + "DeleteJobExecutionRequest$executionNumber": "

The ID of the job execution to be deleted. The executionNumber refers to the execution of a particular job on a particular device.

Note that once a job execution is deleted, the executionNumber may be reused by IoT, so be sure you get and use the correct value here.

", "DescribeJobExecutionRequest$executionNumber": "

A string (consisting of the digits \"0\" through \"9\" which is used to specify a particular job execution on a particular device.

", "JobExecution$executionNumber": "

A string (consisting of the digits \"0\" through \"9\") which identifies this particular job execution on this particular device. It can be used in commands which return or update job execution information.

", "JobExecutionSummary$executionNumber": "

A string (consisting of the digits \"0\" through \"9\") which identifies this particular job execution on this particular device. It can be used later in commands which return or update job execution information.

" } }, + "ExpectedVersion": { + "base": null, + "refs": { + "CancelJobExecutionRequest$expectedVersion": "

(Optional) The expected current version of the job execution. Each time you update the job execution, its version is incremented. If the version of the job execution stored in Jobs does not match, the update is rejected with a VersionMismatch error, and an ErrorResponse that contains the current job execution status data is returned. (This makes it unnecessary to perform a separate DescribeJobExecution request in order to obtain the job execution status data.)

" + } + }, "ExpiresInSec": { "base": null, "refs": { @@ -1430,6 +1456,22 @@ "DeleteCertificateRequest$forceDelete": "

Forces a certificate request to be deleted.

" } }, + "ForceFlag": { + "base": null, + "refs": { + "CancelJobExecutionRequest$force": "

(Optional) If true the job execution will be canceled if it has status IN_PROGRESS or QUEUED, otherwise the job execution will be canceled only if it has status QUEUED. If you attempt to cancel a job execution that is IN_PROGRESS, and you do not set force to true, then an InvalidStateTransitionException will be thrown. The default is false.

Canceling a job execution which is \"IN_PROGRESS\", will cause the device to be unable to update the job execution status. Use caution and ensure that the device is able to recover to a valid state.

", + "CancelJobRequest$force": "

(Optional) If true job executions with status \"IN_PROGRESS\" and \"QUEUED\" are canceled, otherwise only job executions with status \"QUEUED\" are canceled. The default is false.

Canceling a job which is \"IN_PROGRESS\", will cause a device which is executing the job to be unable to update the job execution status. Use caution and ensure that each device executing a job which is canceled is able to recover to a valid state.

", + "DeleteJobExecutionRequest$force": "

(Optional) When true, you can delete a job execution which is \"IN_PROGRESS\". Otherwise, you can only delete a job execution which is in a terminal state (\"SUCCEEDED\", \"FAILED\", \"REJECTED\", \"REMOVED\" or \"CANCELED\") or an exception will occur. The default is false.

Deleting a job execution which is \"IN_PROGRESS\", will cause the device to be unable to access job information or update the job execution status. Use caution and ensure that the device is able to recover to a valid state.

", + "DeleteJobRequest$force": "

(Optional) When true, you can delete a job which is \"IN_PROGRESS\". Otherwise, you can only delete a job which is in a terminal state (\"COMPLETED\" or \"CANCELED\") or an exception will occur. The default is false.

Deleting a job which is \"IN_PROGRESS\", will cause a device which is executing the job to be unable to access job information or update the job execution status. Use caution and ensure that each device executing a job which is deleted is able to recover to a valid state.

" + } + }, + "Forced": { + "base": null, + "refs": { + "Job$forceCanceled": "

Will be true if the job was canceled with the optional force parameter set to true.

", + "JobExecution$forceCanceled": "

Will be true if the job execution was canceled with the optional force parameter set to true.

" + } + }, "FunctionArn": { "base": null, "refs": { @@ -1650,8 +1692,13 @@ "refs": { } }, + "InvalidStateTransitionException": { + "base": "

An attempt was made to change to an invalid state, for example by deleting a job or a job execution which is \"IN_PROGRESS\" without setting the force parameter.

", + "refs": { + } + }, "IotAnalyticsAction": { - "base": "

Sends message data to an AWS IoT Analytics channel.

", + "base": "

Sends messge data to an AWS IoT Analytics channel.

", "refs": { "Action$iotAnalytics": "

Sends message data to an AWS IoT Analytics channel.

" } @@ -1790,10 +1837,13 @@ "refs": { "AssociateTargetsWithJobRequest$jobId": "

The unique identifier you assigned to this job when it was created.

", "AssociateTargetsWithJobResponse$jobId": "

The unique identifier you assigned to this job when it was created.

", + "CancelJobExecutionRequest$jobId": "

The ID of the job to be canceled.

", "CancelJobRequest$jobId": "

The unique identifier you assigned to this job when it was created.

", "CancelJobResponse$jobId": "

The unique identifier you assigned to this job when it was created.

", "CreateJobRequest$jobId": "

A job identifier which must be unique for your AWS account. We recommend using a UUID. Alpha-numeric characters, \"-\" and \"_\" are valid for use here.

", "CreateJobResponse$jobId": "

The unique identifier you assigned to this job.

", + "DeleteJobExecutionRequest$jobId": "

The ID of the job whose execution on a particular device will be deleted.

", + "DeleteJobRequest$jobId": "

The ID of the job to be deleted.

After a job deletion is completed, you may reuse this jobId when you create a new job. However, this is not recommended, and you must ensure that your devices are not using the jobId to refer to the deleted job.

", "DescribeJobExecutionRequest$jobId": "

The unique identifier you assigned to this job when it was created.

", "DescribeJobRequest$jobId": "

The unique identifier you assigned to this job when it was created.

", "GetJobDocumentRequest$jobId": "

The unique identifier you assigned to this job when it was created.

", @@ -1898,7 +1948,7 @@ } }, "LimitExceededException": { - "base": "

The number of attached entities exceeds the limit.

", + "base": "

A limit has been exceeded.

", "refs": { } }, @@ -2301,7 +2351,7 @@ "MessageFormat": { "base": null, "refs": { - "SnsAction$messageFormat": "

The message format of the message to publish. Optional. Accepted values are \"JSON\" and \"RAW\". The default value of the attribute is \"RAW\". SNS uses this setting to determine if the payload should be parsed and relevant platform-specific bits of the payload should be extracted. To read more about SNS message formats, see http://docs.aws.amazon.com/sns/latest/dg/json-formats.html refer to their official documentation.

" + "SnsAction$messageFormat": "

(Optional) The message format of the message to publish. Accepted values are \"JSON\" and \"RAW\". The default value of the attribute is \"RAW\". SNS uses this setting to determine if the payload should be parsed and relevant platform-specific bits of the payload should be extracted. To read more about SNS message formats, see http://docs.aws.amazon.com/sns/latest/dg/json-formats.html refer to their official documentation.

" } }, "MetricName": { @@ -2719,7 +2769,7 @@ "ProcessingTargetNameList": { "base": null, "refs": { - "JobProcessDetails$processingTargets": "

The devices on which the job is executing.

" + "JobProcessDetails$processingTargets": "

The target devices to which the job execution is being rolled out. This value will be null after the job execution has finished rolling out to all the target devices.

" } }, "PublicKey": { @@ -3593,8 +3643,10 @@ "refs": { "AddThingToThingGroupRequest$thingName": "

The name of the thing to add to a group.

", "AttachThingPrincipalRequest$thingName": "

The name of the thing.

", + "CancelJobExecutionRequest$thingName": "

The name of the thing whose execution of the job will be canceled.

", "CreateThingRequest$thingName": "

The name of the thing to create.

", "CreateThingResponse$thingName": "

The name of the new thing.

", + "DeleteJobExecutionRequest$thingName": "

The name of the thing whose job execution will be deleted.

", "DeleteThingRequest$thingName": "

The name of the thing to delete.

", "DescribeJobExecutionRequest$thingName": "

The name of the thing on which the job execution is running.

", "DescribeThingRequest$thingName": "

The name of the thing.

", @@ -3900,10 +3952,16 @@ } }, "VersionConflictException": { - "base": "

An exception thrown when the version of a thing passed to a command is different than the version specified with the --version parameter.

", + "base": "

An exception thrown when the version of an entity specified with the expectedVersion parameter does not match the latest version in the system.

", "refs": { } }, + "VersionNumber": { + "base": null, + "refs": { + "JobExecution$versionNumber": "

The version of the job execution. Job execution versions are incremented each time they are updated by a device.

" + } + }, "VersionsLimitExceededException": { "base": "

The number of policy versions exceeds the limit.

", "refs": { @@ -3923,6 +3981,7 @@ "InvalidQueryException$message": null, "InvalidRequestException$message": "

The message for the exception.

", "InvalidResponseException$message": null, + "InvalidStateTransitionException$message": null, "LimitExceededException$message": "

The message for the exception.

", "MalformedPolicyException$message": "

The message for the exception.

", "NotConfiguredException$message": null, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/api-2.json index 645c350c7..a9497068e 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/api-2.json @@ -367,6 +367,23 @@ {"shape":"ThrottlingException"} ] }, + "ListTagsForResource":{ + "name":"ListTagsForResource", + "http":{ + "method":"GET", + "requestUri":"/tags" + }, + "input":{"shape":"ListTagsForResourceRequest"}, + "output":{"shape":"ListTagsForResourceResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"InternalFailureException"}, + {"shape":"ServiceUnavailableException"}, + {"shape":"ThrottlingException"}, + {"shape":"LimitExceededException"}, + {"shape":"ResourceNotFoundException"} + ] + }, "PutLoggingOptions":{ "name":"PutLoggingOptions", "http":{ @@ -429,6 +446,42 @@ {"shape":"ThrottlingException"} ] }, + "TagResource":{ + "name":"TagResource", + "http":{ + "method":"POST", + "requestUri":"/tags", + "responseCode":204 + }, + "input":{"shape":"TagResourceRequest"}, + "output":{"shape":"TagResourceResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"InternalFailureException"}, + {"shape":"ServiceUnavailableException"}, + {"shape":"ThrottlingException"}, + {"shape":"LimitExceededException"}, + {"shape":"ResourceNotFoundException"} + ] + }, + "UntagResource":{ + "name":"UntagResource", + "http":{ + "method":"DELETE", + "requestUri":"/tags", + "responseCode":204 + }, + "input":{"shape":"UntagResourceRequest"}, + "output":{"shape":"UntagResourceResponse"}, + "errors":[ + {"shape":"InvalidRequestException"}, + {"shape":"InternalFailureException"}, + {"shape":"ServiceUnavailableException"}, + {"shape":"ThrottlingException"}, + {"shape":"LimitExceededException"}, + {"shape":"ResourceNotFoundException"} + ] + }, "UpdateChannel":{ "name":"UpdateChannel", "http":{ @@ -641,7 +694,8 @@ "required":["channelName"], "members":{ "channelName":{"shape":"ChannelName"}, - "retentionPeriod":{"shape":"RetentionPeriod"} + "retentionPeriod":{"shape":"RetentionPeriod"}, + "tags":{"shape":"TagList"} } }, "CreateChannelResponse":{ @@ -672,7 +726,8 @@ "members":{ "datasetName":{"shape":"DatasetName"}, "actions":{"shape":"DatasetActions"}, - "triggers":{"shape":"DatasetTriggers"} + "triggers":{"shape":"DatasetTriggers"}, + "tags":{"shape":"TagList"} } }, "CreateDatasetResponse":{ @@ -687,7 +742,8 @@ "required":["datastoreName"], "members":{ "datastoreName":{"shape":"DatastoreName"}, - "retentionPeriod":{"shape":"RetentionPeriod"} + "retentionPeriod":{"shape":"RetentionPeriod"}, + "tags":{"shape":"TagList"} } }, "CreateDatastoreResponse":{ @@ -706,7 +762,8 @@ ], "members":{ "pipelineName":{"shape":"PipelineName"}, - "pipelineActivities":{"shape":"PipelineActivities"} + "pipelineActivities":{"shape":"PipelineActivities"}, + "tags":{"shape":"TagList"} } }, "CreatePipelineResponse":{ @@ -1213,6 +1270,23 @@ "nextToken":{"shape":"NextToken"} } }, + "ListTagsForResourceRequest":{ + "type":"structure", + "required":["resourceArn"], + "members":{ + "resourceArn":{ + "shape":"ResourceArn", + "location":"querystring", + "locationName":"resourceArn" + } + } + }, + "ListTagsForResourceResponse":{ + "type":"structure", + "members":{ + "tags":{"shape":"TagList"} + } + }, "LogResult":{"type":"string"}, "LoggingEnabled":{"type":"boolean"}, "LoggingLevel":{ @@ -1394,6 +1468,11 @@ "error":{"httpStatusCode":409}, "exception":true }, + "ResourceArn":{ + "type":"string", + "max":2048, + "min":20 + }, "ResourceNotFoundException":{ "type":"structure", "members":{ @@ -1524,6 +1603,59 @@ } }, "StartTime":{"type":"timestamp"}, + "Tag":{ + "type":"structure", + "required":[ + "key", + "value" + ], + "members":{ + "key":{"shape":"TagKey"}, + "value":{"shape":"TagValue"} + } + }, + "TagKey":{ + "type":"string", + "max":256, + "min":1 + }, + "TagKeyList":{ + "type":"list", + "member":{"shape":"TagKey"}, + "max":50, + "min":1 + }, + "TagList":{ + "type":"list", + "member":{"shape":"Tag"}, + "max":50, + "min":1 + }, + "TagResourceRequest":{ + "type":"structure", + "required":[ + "resourceArn", + "tags" + ], + "members":{ + "resourceArn":{ + "shape":"ResourceArn", + "location":"querystring", + "locationName":"resourceArn" + }, + "tags":{"shape":"TagList"} + } + }, + "TagResourceResponse":{ + "type":"structure", + "members":{ + } + }, + "TagValue":{ + "type":"string", + "max":256, + "min":1 + }, "ThrottlingException":{ "type":"structure", "members":{ @@ -1534,6 +1666,30 @@ }, "Timestamp":{"type":"timestamp"}, "UnlimitedRetentionPeriod":{"type":"boolean"}, + "UntagResourceRequest":{ + "type":"structure", + "required":[ + "resourceArn", + "tagKeys" + ], + "members":{ + "resourceArn":{ + "shape":"ResourceArn", + "location":"querystring", + "locationName":"resourceArn" + }, + "tagKeys":{ + "shape":"TagKeyList", + "location":"querystring", + "locationName":"tagKeys" + } + } + }, + "UntagResourceResponse":{ + "type":"structure", + "members":{ + } + }, "UpdateChannelRequest":{ "type":"structure", "required":["channelName"], diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/docs-2.json index 39457ac82..49622e7e3 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/iotanalytics/2017-11-27/docs-2.json @@ -1,6 +1,6 @@ { "version": "2.0", - "service": "

AWS IoT Analytics provides advanced data analysis for AWS IoT. It allows you to collect large amounts of device data, process messages, store them, and then query the data and run sophisticated analytics to make accurate decisions in your IoT applications and machine learning use cases. AWS IoT Analytics enables advanced data exploration through integration with Jupyter Notebooks and data visualization through integration with Amazon QuickSight.

Traditional analytics and business intelligence tools are designed to process structured data. IoT data often comes from devices that record noisy processes (such as temperature, motion, or sound). As a result, the data from these devices can have significant gaps, corrupted messages, and false readings that must be cleaned up before analysis can occur. Also, IoT data is often only meaningful in the context of other data from external sources.

AWS IoT Analytics automates each of the steps required to analyze data from IoT devices. AWS IoT Analytics filters, transforms, and enriches IoT data before storing it in a time-series data store for analysis. You can set up the service to collect only the data you need from your devices, apply mathematical transforms to process the data, and enrich the data with device-specific metadata such as device type and location before storing it. Then, you can analyze your data by running queries using the built-in SQL query engine, or perform more complex analytics and machine learning inference. AWS IoT Analytics includes models for common IoT use cases so you can answer questions like which devices are about to fail or which customers are at risk of abandoning their wearable devices.

", + "service": "

AWS IoT Analytics allows you to collect large amounts of device data, process messages, and store them. You can then query the data and run sophisticated analytics on it. AWS IoT Analytics enables advanced data exploration through integration with Jupyter Notebooks and data visualization through integration with Amazon QuickSight.

Traditional analytics and business intelligence tools are designed to process structured data. IoT data often comes from devices that record noisy processes (such as temperature, motion, or sound). As a result the data from these devices can have significant gaps, corrupted messages, and false readings that must be cleaned up before analysis can occur. Also, IoT data is often only meaningful in the context of other data from external sources.

AWS IoT Analytics automates the steps required to analyze data from IoT devices. AWS IoT Analytics filters, transforms, and enriches IoT data before storing it in a time-series data store for analysis. You can set up the service to collect only the data you need from your devices, apply mathematical transforms to process the data, and enrich the data with device-specific metadata such as device type and location before storing it. Then, you can analyze your data by running queries using the built-in SQL query engine, or perform more complex analytics and machine learning inference. AWS IoT Analytics includes pre-built models for common IoT use cases so you can answer questions like which devices are about to fail or which customers are at risk of abandoning their wearable devices.

", "operations": { "BatchPutMessage": "

Sends messages to a channel.

", "CancelPipelineReprocessing": "

Cancels the reprocessing of data through the pipeline.

", @@ -24,10 +24,13 @@ "ListDatasets": "

Retrieves information about data sets.

", "ListDatastores": "

Retrieves a list of data stores.

", "ListPipelines": "

Retrieves a list of pipelines.

", - "PutLoggingOptions": "

Sets or updates the AWS IoT Analytics logging options.

", + "ListTagsForResource": "

Lists the tags (metadata) which you have assigned to the resource.

", + "PutLoggingOptions": "

Sets or updates the AWS IoT Analytics logging options.

Note that if you update the value of any loggingOptions field, it takes up to one minute for the change to take effect. Also, if you change the policy attached to the role you specified in the roleArn field (for example, to correct an invalid policy) it takes up to 5 minutes for that change to take effect.

", "RunPipelineActivity": "

Simulates the results of running a pipeline activity on a message payload.

", "SampleChannelData": "

Retrieves a sample of messages from the specified channel ingested during the specified timeframe. Up to 10 messages can be retrieved.

", "StartPipelineReprocessing": "

Starts the reprocessing of raw message data through the pipeline.

", + "TagResource": "

Adds to or modifies the tags of the given resource. Tags are metadata which can be used to manage a resource.

", + "UntagResource": "

Removes the given tags (metadata) from the resource.

", "UpdateChannel": "

Updates the settings of a channel.

", "UpdateDataset": "

Updates the settings of a data set.

", "UpdateDatastore": "

Updates the settings of a data store.

", @@ -590,6 +593,16 @@ "refs": { } }, + "ListTagsForResourceRequest": { + "base": null, + "refs": { + } + }, + "ListTagsForResourceResponse": { + "base": null, + "refs": { + } + }, "LogResult": { "base": null, "refs": { @@ -652,7 +665,7 @@ "base": null, "refs": { "BatchPutMessageErrorEntry$messageId": "

The ID of the message that caused the error. (See the value corresponding to the \"messageId\" key in the message object.)

", - "Message$messageId": "

The ID you wish to assign to the message.

" + "Message$messageId": "

The ID you wish to assign to the message. Each \"messageId\" must be unique within each batch sent.

" } }, "MessagePayload": { @@ -798,6 +811,14 @@ "refs": { } }, + "ResourceArn": { + "base": null, + "refs": { + "ListTagsForResourceRequest$resourceArn": "

The ARN of the resource whose tags you want to list.

", + "TagResourceRequest$resourceArn": "

The ARN of the resource whose tags will be modified.

", + "UntagResourceRequest$resourceArn": "

The ARN of the resource whose tags will be removed.

" + } + }, "ResourceNotFoundException": { "base": "

A resource with the specified name could not be found.

", "refs": { @@ -902,6 +923,52 @@ "StartPipelineReprocessingRequest$startTime": "

The start time (inclusive) of raw message data that is reprocessed.

" } }, + "Tag": { + "base": "

A set of key/value pairs which are used to manage the resource.

", + "refs": { + "TagList$member": null + } + }, + "TagKey": { + "base": null, + "refs": { + "Tag$key": "

The tag's key.

", + "TagKeyList$member": null + } + }, + "TagKeyList": { + "base": null, + "refs": { + "UntagResourceRequest$tagKeys": "

The keys of those tags which will be removed.

" + } + }, + "TagList": { + "base": null, + "refs": { + "CreateChannelRequest$tags": "

Metadata which can be used to manage the channel.

", + "CreateDatasetRequest$tags": "

Metadata which can be used to manage the data set.

", + "CreateDatastoreRequest$tags": "

Metadata which can be used to manage the data store.

", + "CreatePipelineRequest$tags": "

Metadata which can be used to manage the pipeline.

", + "ListTagsForResourceResponse$tags": "

The tags (metadata) which you have assigned to the resource.

", + "TagResourceRequest$tags": "

The new or modified tags for the resource.

" + } + }, + "TagResourceRequest": { + "base": null, + "refs": { + } + }, + "TagResourceResponse": { + "base": null, + "refs": { + } + }, + "TagValue": { + "base": null, + "refs": { + "Tag$value": "

The tag's value.

" + } + }, "ThrottlingException": { "base": "

The request was denied due to request throttling.

", "refs": { @@ -936,6 +1003,16 @@ "RetentionPeriod$unlimited": "

If true, message data is kept indefinitely.

" } }, + "UntagResourceRequest": { + "base": null, + "refs": { + } + }, + "UntagResourceResponse": { + "base": null, + "refs": { + } + }, "UpdateChannelRequest": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/api-2.json deleted file mode 100644 index a8f192ac0..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/api-2.json +++ /dev/null @@ -1,668 +0,0 @@ -{ - "metadata":{ - "uid":"lambda-2014-11-11", - "apiVersion":"2014-11-11", - "endpointPrefix":"lambda", - "serviceFullName":"AWS Lambda", - "signatureVersion":"v4", - "protocol":"rest-json" - }, - "operations":{ - "AddEventSource":{ - "name":"AddEventSource", - "http":{ - "method":"POST", - "requestUri":"/2014-11-13/event-source-mappings/" - }, - "input":{"shape":"AddEventSourceRequest"}, - "output":{"shape":"EventSourceConfiguration"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"InvalidParameterValueException", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "DeleteFunction":{ - "name":"DeleteFunction", - "http":{ - "method":"DELETE", - "requestUri":"/2014-11-13/functions/{FunctionName}", - "responseCode":204 - }, - "input":{"shape":"DeleteFunctionRequest"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - } - ] - }, - "GetEventSource":{ - "name":"GetEventSource", - "http":{ - "method":"GET", - "requestUri":"/2014-11-13/event-source-mappings/{UUID}", - "responseCode":200 - }, - "input":{"shape":"GetEventSourceRequest"}, - "output":{"shape":"EventSourceConfiguration"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"InvalidParameterValueException", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "GetFunction":{ - "name":"GetFunction", - "http":{ - "method":"GET", - "requestUri":"/2014-11-13/functions/{FunctionName}", - "responseCode":200 - }, - "input":{"shape":"GetFunctionRequest"}, - "output":{"shape":"GetFunctionResponse"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - } - ] - }, - "GetFunctionConfiguration":{ - "name":"GetFunctionConfiguration", - "http":{ - "method":"GET", - "requestUri":"/2014-11-13/functions/{FunctionName}/configuration", - "responseCode":200 - }, - "input":{"shape":"GetFunctionConfigurationRequest"}, - "output":{"shape":"FunctionConfiguration"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - } - ] - }, - "InvokeAsync":{ - "name":"InvokeAsync", - "http":{ - "method":"POST", - "requestUri":"/2014-11-13/functions/{FunctionName}/invoke-async/", - "responseCode":202 - }, - "input":{"shape":"InvokeAsyncRequest"}, - "output":{"shape":"InvokeAsyncResponse"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"InvalidRequestContentException", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListEventSources":{ - "name":"ListEventSources", - "http":{ - "method":"GET", - "requestUri":"/2014-11-13/event-source-mappings/", - "responseCode":200 - }, - "input":{"shape":"ListEventSourcesRequest"}, - "output":{"shape":"ListEventSourcesResponse"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"InvalidParameterValueException", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "ListFunctions":{ - "name":"ListFunctions", - "http":{ - "method":"GET", - "requestUri":"/2014-11-13/functions/", - "responseCode":200 - }, - "input":{"shape":"ListFunctionsRequest"}, - "output":{"shape":"ListFunctionsResponse"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - } - ] - }, - "RemoveEventSource":{ - "name":"RemoveEventSource", - "http":{ - "method":"DELETE", - "requestUri":"/2014-11-13/event-source-mappings/{UUID}", - "responseCode":204 - }, - "input":{"shape":"RemoveEventSourceRequest"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"InvalidParameterValueException", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UpdateFunctionConfiguration":{ - "name":"UpdateFunctionConfiguration", - "http":{ - "method":"PUT", - "requestUri":"/2014-11-13/functions/{FunctionName}/configuration", - "responseCode":200 - }, - "input":{"shape":"UpdateFunctionConfigurationRequest"}, - "output":{"shape":"FunctionConfiguration"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - }, - { - "shape":"InvalidParameterValueException", - "error":{"httpStatusCode":400}, - "exception":true - } - ] - }, - "UploadFunction":{ - "name":"UploadFunction", - "http":{ - "method":"PUT", - "requestUri":"/2014-11-13/functions/{FunctionName}", - "responseCode":201 - }, - "input":{"shape":"UploadFunctionRequest"}, - "output":{"shape":"FunctionConfiguration"}, - "errors":[ - { - "shape":"ServiceException", - "error":{"httpStatusCode":500}, - "exception":true - }, - { - "shape":"InvalidParameterValueException", - "error":{"httpStatusCode":400}, - "exception":true - }, - { - "shape":"ResourceNotFoundException", - "error":{"httpStatusCode":404}, - "exception":true - } - ] - } - }, - "shapes":{ - "AddEventSourceRequest":{ - "type":"structure", - "required":[ - "EventSource", - "FunctionName", - "Role" - ], - "members":{ - "EventSource":{"shape":"String"}, - "FunctionName":{"shape":"FunctionName"}, - "Role":{"shape":"RoleArn"}, - "BatchSize":{"shape":"Integer"}, - "Parameters":{"shape":"Map"} - } - }, - "Blob":{ - "type":"blob", - "streaming":true - }, - "DeleteFunctionRequest":{ - "type":"structure", - "required":["FunctionName"], - "members":{ - "FunctionName":{ - "shape":"FunctionName", - "location":"uri", - "locationName":"FunctionName" - } - } - }, - "Description":{ - "type":"string", - "min":0, - "max":256 - }, - "EventSourceConfiguration":{ - "type":"structure", - "members":{ - "UUID":{"shape":"String"}, - "BatchSize":{"shape":"Integer"}, - "EventSource":{"shape":"String"}, - "FunctionName":{"shape":"FunctionName"}, - "Parameters":{"shape":"Map"}, - "Role":{"shape":"RoleArn"}, - "LastModified":{"shape":"Timestamp"}, - "IsActive":{"shape":"Boolean"}, - "Status":{"shape":"String"} - } - }, - "EventSourceList":{ - "type":"list", - "member":{"shape":"EventSourceConfiguration"} - }, - "FunctionArn":{ - "type":"string", - "pattern":"arn:aws:lambda:[a-z]{2}-[a-z]+-\\d{1}:\\d{12}:function:[a-zA-Z0-9-_]+(\\/[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12})?" - }, - "FunctionCodeLocation":{ - "type":"structure", - "members":{ - "RepositoryType":{"shape":"String"}, - "Location":{"shape":"String"} - } - }, - "FunctionConfiguration":{ - "type":"structure", - "members":{ - "FunctionName":{"shape":"FunctionName"}, - "FunctionARN":{"shape":"FunctionArn"}, - "ConfigurationId":{"shape":"String"}, - "Runtime":{"shape":"Runtime"}, - "Role":{"shape":"RoleArn"}, - "Handler":{"shape":"Handler"}, - "Mode":{"shape":"Mode"}, - "CodeSize":{"shape":"Long"}, - "Description":{"shape":"Description"}, - "Timeout":{"shape":"Timeout"}, - "MemorySize":{"shape":"MemorySize"}, - "LastModified":{"shape":"Timestamp"} - } - }, - "FunctionList":{ - "type":"list", - "member":{"shape":"FunctionConfiguration"} - }, - "FunctionName":{ - "type":"string", - "min":1, - "max":64, - "pattern":"[a-zA-Z0-9-_]+" - }, - "GetEventSourceRequest":{ - "type":"structure", - "required":["UUID"], - "members":{ - "UUID":{ - "shape":"String", - "location":"uri", - "locationName":"UUID" - } - } - }, - "GetFunctionConfigurationRequest":{ - "type":"structure", - "required":["FunctionName"], - "members":{ - "FunctionName":{ - "shape":"FunctionName", - "location":"uri", - "locationName":"FunctionName" - } - } - }, - "GetFunctionRequest":{ - "type":"structure", - "required":["FunctionName"], - "members":{ - "FunctionName":{ - "shape":"FunctionName", - "location":"uri", - "locationName":"FunctionName" - } - } - }, - "GetFunctionResponse":{ - "type":"structure", - "members":{ - "Configuration":{"shape":"FunctionConfiguration"}, - "Code":{"shape":"FunctionCodeLocation"} - } - }, - "Handler":{ - "type":"string", - "pattern":"[a-zA-Z0-9./\\-_]+" - }, - "HttpStatus":{"type":"integer"}, - "Integer":{"type":"integer"}, - "InvalidParameterValueException":{ - "type":"structure", - "members":{ - "Type":{"shape":"String"}, - "message":{"shape":"String"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvalidRequestContentException":{ - "type":"structure", - "members":{ - "Type":{"shape":"String"}, - "message":{"shape":"String"} - }, - "error":{"httpStatusCode":400}, - "exception":true - }, - "InvokeAsyncRequest":{ - "type":"structure", - "required":[ - "FunctionName", - "InvokeArgs" - ], - "members":{ - "FunctionName":{ - "shape":"FunctionName", - "location":"uri", - "locationName":"FunctionName" - }, - "InvokeArgs":{"shape":"Blob"} - }, - "payload":"InvokeArgs" - }, - "InvokeAsyncResponse":{ - "type":"structure", - "members":{ - "Status":{ - "shape":"HttpStatus", - "location":"statusCode" - } - } - }, - "ListEventSourcesRequest":{ - "type":"structure", - "members":{ - "EventSourceArn":{ - "shape":"String", - "location":"querystring", - "locationName":"EventSource" - }, - "FunctionName":{ - "shape":"FunctionName", - "location":"querystring", - "locationName":"FunctionName" - }, - "Marker":{ - "shape":"String", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"MaxListItems", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListEventSourcesResponse":{ - "type":"structure", - "members":{ - "NextMarker":{"shape":"String"}, - "EventSources":{"shape":"EventSourceList"} - } - }, - "ListFunctionsRequest":{ - "type":"structure", - "members":{ - "Marker":{ - "shape":"String", - "location":"querystring", - "locationName":"Marker" - }, - "MaxItems":{ - "shape":"MaxListItems", - "location":"querystring", - "locationName":"MaxItems" - } - } - }, - "ListFunctionsResponse":{ - "type":"structure", - "members":{ - "NextMarker":{"shape":"String"}, - "Functions":{"shape":"FunctionList"} - } - }, - "Long":{"type":"long"}, - "Map":{ - "type":"map", - "key":{"shape":"String"}, - "value":{"shape":"String"} - }, - "MaxListItems":{ - "type":"integer", - "min":1, - "max":10000 - }, - "MemorySize":{ - "type":"integer", - "min":128, - "max":1024 - }, - "Mode":{ - "type":"string", - "enum":["event"] - }, - "RemoveEventSourceRequest":{ - "type":"structure", - "required":["UUID"], - "members":{ - "UUID":{ - "shape":"String", - "location":"uri", - "locationName":"UUID" - } - } - }, - "ResourceNotFoundException":{ - "type":"structure", - "members":{ - "Type":{"shape":"String"}, - "Message":{"shape":"String"} - }, - "error":{"httpStatusCode":404}, - "exception":true - }, - "RoleArn":{ - "type":"string", - "pattern":"arn:aws:iam::\\d{12}:role/?[a-zA-Z_0-9+=,.@\\-_/]+" - }, - "Runtime":{ - "type":"string", - "enum":["nodejs"] - }, - "ServiceException":{ - "type":"structure", - "members":{ - "Type":{"shape":"String"}, - "Message":{"shape":"String"} - }, - "error":{"httpStatusCode":500}, - "exception":true - }, - "String":{"type":"string"}, - "Timeout":{ - "type":"integer", - "min":1, - "max":60 - }, - "Timestamp":{"type":"string"}, - "UpdateFunctionConfigurationRequest":{ - "type":"structure", - "required":["FunctionName"], - "members":{ - "FunctionName":{ - "shape":"FunctionName", - "location":"uri", - "locationName":"FunctionName" - }, - "Role":{ - "shape":"RoleArn", - "location":"querystring", - "locationName":"Role" - }, - "Handler":{ - "shape":"Handler", - "location":"querystring", - "locationName":"Handler" - }, - "Description":{ - "shape":"Description", - "location":"querystring", - "locationName":"Description" - }, - "Timeout":{ - "shape":"Timeout", - "location":"querystring", - "locationName":"Timeout" - }, - "MemorySize":{ - "shape":"MemorySize", - "location":"querystring", - "locationName":"MemorySize" - } - } - }, - "UploadFunctionRequest":{ - "type":"structure", - "required":[ - "FunctionName", - "FunctionZip", - "Runtime", - "Role", - "Handler", - "Mode" - ], - "members":{ - "FunctionName":{ - "shape":"FunctionName", - "location":"uri", - "locationName":"FunctionName" - }, - "FunctionZip":{"shape":"Blob"}, - "Runtime":{ - "shape":"Runtime", - "location":"querystring", - "locationName":"Runtime" - }, - "Role":{ - "shape":"RoleArn", - "location":"querystring", - "locationName":"Role" - }, - "Handler":{ - "shape":"Handler", - "location":"querystring", - "locationName":"Handler" - }, - "Mode":{ - "shape":"Mode", - "location":"querystring", - "locationName":"Mode" - }, - "Description":{ - "shape":"Description", - "location":"querystring", - "locationName":"Description" - }, - "Timeout":{ - "shape":"Timeout", - "location":"querystring", - "locationName":"Timeout" - }, - "MemorySize":{ - "shape":"MemorySize", - "location":"querystring", - "locationName":"MemorySize" - } - }, - "payload":"FunctionZip" - }, - "Boolean":{"type":"boolean"} - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/docs-2.json deleted file mode 100644 index 4ef27761b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/docs-2.json +++ /dev/null @@ -1,303 +0,0 @@ -{ - "operations": { - "AddEventSource": "

Identifies a stream as an event source for an AWS Lambda function. It can be either an Amazon Kinesis stream or a Amazon DynamoDB stream. AWS Lambda invokes the specified function when records are posted to the stream.

This is the pull model, where AWS Lambda invokes the function. For more information, go to AWS Lambda: How it Works in the AWS Lambda Developer Guide.

This association between an Amazon Kinesis stream and an AWS Lambda function is called the event source mapping. You provide the configuration information (for example, which stream to read from and which AWS Lambda function to invoke) for the event source mapping in the request body.

Each event source, such as a Kinesis stream, can only be associated with one AWS Lambda function. If you call AddEventSource for an event source that is already mapped to another AWS Lambda function, the existing mapping is updated to call the new function instead of the old one.

This operation requires permission for the iam:PassRole action for the IAM role. It also requires permission for the lambda:AddEventSource action.

", - "DeleteFunction": "

Deletes the specified Lambda function code and configuration.

This operation requires permission for the lambda:DeleteFunction action.

", - "GetEventSource": "

Returns configuration information for the specified event source mapping (see AddEventSource).

This operation requires permission for the lambda:GetEventSource action.

", - "GetFunction": "

Returns the configuration information of the Lambda function and a presigned URL link to the .zip file you uploaded with UploadFunction so you can download the .zip file. Note that the URL is valid for up to 10 minutes. The configuration information is the same information you provided as parameters when uploading the function.

This operation requires permission for the lambda:GetFunction action.

", - "GetFunctionConfiguration": "

Returns the configuration information of the Lambda function. This the same information you provided as parameters when uploading the function by using UploadFunction.

This operation requires permission for the lambda:GetFunctionConfiguration operation.

", - "InvokeAsync": "

Submits an invocation request to AWS Lambda. Upon receiving the request, Lambda executes the specified function asynchronously. To see the logs generated by the Lambda function execution, see the CloudWatch logs console.

This operation requires permission for the lambda:InvokeFunction action.

", - "ListEventSources": "

Returns a list of event source mappings you created using the AddEventSource (see AddEventSource), where you identify a stream as event source. This list does not include Amazon S3 event sources.

For each mapping, the API returns configuration information. You can optionally specify filters to retrieve specific event source mappings.

This operation requires permission for the lambda:ListEventSources action.

", - "ListFunctions": "

Returns a list of your Lambda functions. For each function, the response includes the function configuration information. You must use GetFunction to retrieve the code for your function.

This operation requires permission for the lambda:ListFunctions action.

", - "RemoveEventSource": "

Removes an event source mapping. This means AWS Lambda will no longer invoke the function for events in the associated source.

This operation requires permission for the lambda:RemoveEventSource action.

", - "UpdateFunctionConfiguration": "

Updates the configuration parameters for the specified Lambda function by using the values provided in the request. You provide only the parameters you want to change. This operation must only be used on an existing Lambda function and cannot be used to update the function's code.

This operation requires permission for the lambda:UpdateFunctionConfiguration action.

", - "UploadFunction": "

Creates a new Lambda function or updates an existing function. The function metadata is created from the request parameters, and the code for the function is provided by a .zip file in the request body. If the function name already exists, the existing Lambda function is updated with the new code and metadata.

This operation requires permission for the lambda:UploadFunction action.

" - }, - "service": "AWS Lambda

Overview

This is the AWS Lambda API Reference. The AWS Lambda Developer Guide provides additional information. For the service overview, go to What is AWS Lambda, and for information about how the service works, go to AWS LambdaL How it Works in the AWS Lambda Developer Guide.

", - "shapes": { - "AddEventSourceRequest": { - "base": null, - "refs": { - } - }, - "Blob": { - "base": null, - "refs": { - "InvokeAsyncRequest$InvokeArgs": "

JSON that you want to provide to your Lambda function as input.

", - "UploadFunctionRequest$FunctionZip": "

A .zip file containing your packaged source code. For more information about creating a .zip file, go to AWS LambdaL How it Works in the AWS Lambda Developer Guide.

" - } - }, - "DeleteFunctionRequest": { - "base": null, - "refs": { - } - }, - "Description": { - "base": null, - "refs": { - "FunctionConfiguration$Description": "

The user-provided description.

", - "UpdateFunctionConfigurationRequest$Description": "

A short user-defined function description. Lambda does not use this value. Assign a meaningful description as you see fit.

", - "UploadFunctionRequest$Description": "

A short, user-defined function description. Lambda does not use this value. Assign a meaningful description as you see fit.

" - } - }, - "EventSourceConfiguration": { - "base": "

Describes mapping between an Amazon Kinesis stream and a Lambda function.

", - "refs": { - "EventSourceList$member": null - } - }, - "EventSourceList": { - "base": null, - "refs": { - "ListEventSourcesResponse$EventSources": "

An arrary of EventSourceConfiguration objects.

" - } - }, - "FunctionArn": { - "base": null, - "refs": { - "FunctionConfiguration$FunctionARN": "

The Amazon Resource Name (ARN) assigned to the function.

" - } - }, - "FunctionCodeLocation": { - "base": "

The object for the Lambda function location.

", - "refs": { - "GetFunctionResponse$Code": null - } - }, - "FunctionConfiguration": { - "base": "

A complex type that describes function metadata.

", - "refs": { - "FunctionList$member": null, - "GetFunctionResponse$Configuration": null - } - }, - "FunctionList": { - "base": null, - "refs": { - "ListFunctionsResponse$Functions": "

A list of Lambda functions.

" - } - }, - "FunctionName": { - "base": null, - "refs": { - "AddEventSourceRequest$FunctionName": "

The Lambda function to invoke when AWS Lambda detects an event on the stream.

", - "DeleteFunctionRequest$FunctionName": "

The Lambda function to delete.

", - "EventSourceConfiguration$FunctionName": "

The Lambda function to invoke when AWS Lambda detects an event on the stream.

", - "FunctionConfiguration$FunctionName": "

The name of the function.

", - "GetFunctionConfigurationRequest$FunctionName": "

The name of the Lambda function for which you want to retrieve the configuration information.

", - "GetFunctionRequest$FunctionName": "

The Lambda function name.

", - "InvokeAsyncRequest$FunctionName": "

The Lambda function name.

", - "ListEventSourcesRequest$FunctionName": "

The name of the AWS Lambda function.

", - "UpdateFunctionConfigurationRequest$FunctionName": "

The name of the Lambda function.

", - "UploadFunctionRequest$FunctionName": "

The name you want to assign to the function you are uploading. The function names appear in the console and are returned in the ListFunctions API. Function names are used to specify functions to other AWS Lambda APIs, such as InvokeAsync.

" - } - }, - "GetEventSourceRequest": { - "base": null, - "refs": { - } - }, - "GetFunctionConfigurationRequest": { - "base": null, - "refs": { - } - }, - "GetFunctionRequest": { - "base": null, - "refs": { - } - }, - "GetFunctionResponse": { - "base": "

This response contains the object for AWS Lambda function location (see API_FunctionCodeLocation

", - "refs": { - } - }, - "Handler": { - "base": null, - "refs": { - "FunctionConfiguration$Handler": "

The function Lambda calls to begin executing your function.

", - "UpdateFunctionConfigurationRequest$Handler": "

The function that Lambda calls to begin executing your function. For Node.js, it is the module-name.export value in your function.

", - "UploadFunctionRequest$Handler": "

The function that Lambda calls to begin execution. For Node.js, it is the module-name.export value in your function.

" - } - }, - "HttpStatus": { - "base": null, - "refs": { - "InvokeAsyncResponse$Status": "

It will be 202 upon success.

" - } - }, - "Integer": { - "base": null, - "refs": { - "AddEventSourceRequest$BatchSize": "

The largest number of records that AWS Lambda will give to your function in a single event. The default is 100 records.

", - "EventSourceConfiguration$BatchSize": "

The largest number of records that AWS Lambda will POST in the invocation request to your function.

" - } - }, - "InvalidParameterValueException": { - "base": "

One of the parameters in the request is invalid. For example, if you provided an IAM role for AWS Lambda to assume in the UploadFunction or the UpdateFunctionConfiguration API, that AWS Lambda is unable to assume you will get this exception.

", - "refs": { - } - }, - "InvalidRequestContentException": { - "base": "

The request body could not be parsed as JSON.

", - "refs": { - } - }, - "InvokeAsyncRequest": { - "base": null, - "refs": { - } - }, - "InvokeAsyncResponse": { - "base": "

Upon success, it returns empty response. Otherwise, throws an exception.

", - "refs": { - } - }, - "ListEventSourcesRequest": { - "base": null, - "refs": { - } - }, - "ListEventSourcesResponse": { - "base": "

Contains a list of event sources (see API_EventSourceConfiguration)

", - "refs": { - } - }, - "ListFunctionsRequest": { - "base": null, - "refs": { - } - }, - "ListFunctionsResponse": { - "base": "

Contains a list of AWS Lambda function configurations (see API_FunctionConfiguration.

", - "refs": { - } - }, - "Long": { - "base": null, - "refs": { - "FunctionConfiguration$CodeSize": "

The size, in bytes, of the function .zip file you uploaded.

" - } - }, - "Map": { - "base": null, - "refs": { - "AddEventSourceRequest$Parameters": "

A map (key-value pairs) defining the configuration for AWS Lambda to use when reading the event source. Currently, AWS Lambda supports only the InitialPositionInStream key. The valid values are: \"TRIM_HORIZON\" and \"LATEST\". The default value is \"TRIM_HORIZON\". For more information, go to ShardIteratorType in the Amazon Kinesis Service API Reference.

", - "EventSourceConfiguration$Parameters": "

The map (key-value pairs) defining the configuration for AWS Lambda to use when reading the event source.

" - } - }, - "MaxListItems": { - "base": null, - "refs": { - "ListEventSourcesRequest$MaxItems": "

Optional integer. Specifies the maximum number of event sources to return in response. This value must be greater than 0.

", - "ListFunctionsRequest$MaxItems": "

Optional integer. Specifies the maximum number of AWS Lambda functions to return in response. This parameter value must be greater than 0.

" - } - }, - "MemorySize": { - "base": null, - "refs": { - "FunctionConfiguration$MemorySize": "

The memory size, in MB, you configured for the function. Must be a multiple of 64 MB.

", - "UpdateFunctionConfigurationRequest$MemorySize": "

The amount of memory, in MB, your Lambda function is given. Lambda uses this memory size to infer the amount of CPU allocated to your function. Your function use-case determines your CPU and memory requirements. For example, a database operation might need less memory compared to an image processing function. The default value is 128 MB. The value must be a multiple of 64 MB.

", - "UploadFunctionRequest$MemorySize": "

The amount of memory, in MB, your Lambda function is given. Lambda uses this memory size to infer the amount of CPU allocated to your function. Your function use-case determines your CPU and memory requirements. For example, database operation might need less memory compared to image processing function. The default value is 128 MB. The value must be a multiple of 64 MB.

" - } - }, - "Mode": { - "base": null, - "refs": { - "FunctionConfiguration$Mode": "

The type of the Lambda function you uploaded.

", - "UploadFunctionRequest$Mode": "

How the Lambda function will be invoked. Lambda supports only the \"event\" mode.

" - } - }, - "RemoveEventSourceRequest": { - "base": null, - "refs": { - } - }, - "ResourceNotFoundException": { - "base": "

The function or the event source specified in the request does not exist.

", - "refs": { - } - }, - "RoleArn": { - "base": null, - "refs": { - "AddEventSourceRequest$Role": "

The ARN of the IAM role (invocation role) that AWS Lambda can assume to read from the stream and invoke the function.

", - "EventSourceConfiguration$Role": "

The ARN of the IAM role (invocation role) that AWS Lambda can assume to read from the stream and invoke the function.

", - "FunctionConfiguration$Role": "

The Amazon Resource Name (ARN) of the IAM role that Lambda assumes when it executes your function to access any other Amazon Web Services (AWS) resources.

", - "UpdateFunctionConfigurationRequest$Role": "

The Amazon Resource Name (ARN) of the IAM role that Lambda will assume when it executes your function.

", - "UploadFunctionRequest$Role": "

The Amazon Resource Name (ARN) of the IAM role that Lambda assumes when it executes your function to access any other Amazon Web Services (AWS) resources.

" - } - }, - "Runtime": { - "base": null, - "refs": { - "FunctionConfiguration$Runtime": "

The runtime environment for the Lambda function.

", - "UploadFunctionRequest$Runtime": "

The runtime environment for the Lambda function you are uploading. Currently, Lambda supports only \"nodejs\" as the runtime.

" - } - }, - "ServiceException": { - "base": "

The AWS Lambda service encountered an internal error.

", - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "AddEventSourceRequest$EventSource": "

The Amazon Resource Name (ARN) of the Amazon Kinesis stream that is the event source. Any record added to this stream causes AWS Lambda to invoke your Lambda function. AWS Lambda POSTs the Amazon Kinesis event, containing records, to your Lambda function as JSON.

", - "EventSourceConfiguration$UUID": "

The AWS Lambda assigned opaque identifier for the mapping.

", - "EventSourceConfiguration$EventSource": "

The Amazon Resource Name (ARN) of the Amazon Kinesis stream that is the source of events.

", - "EventSourceConfiguration$Status": "

The description of the health of the event source mapping. Valid values are: \"PENDING\", \"OK\", and \"PROBLEM:message\". Initially this staus is \"PENDING\". When AWS Lambda begins processing events, it changes the status to \"OK\".

", - "FunctionCodeLocation$RepositoryType": "

The repository from which you can download the function.

", - "FunctionCodeLocation$Location": "

The presigned URL you can use to download the function's .zip file that you previously uploaded. The URL is valid for up to 10 minutes.

", - "FunctionConfiguration$ConfigurationId": "

A Lambda-assigned unique identifier for the current function code and related configuration.

", - "GetEventSourceRequest$UUID": "

The AWS Lambda assigned ID of the event source mapping.

", - "InvalidParameterValueException$Type": null, - "InvalidParameterValueException$message": null, - "InvalidRequestContentException$Type": null, - "InvalidRequestContentException$message": null, - "ListEventSourcesRequest$EventSourceArn": "

The Amazon Resource Name (ARN) of the Amazon Kinesis stream.

", - "ListEventSourcesRequest$Marker": "

Optional string. An opaque pagination token returned from a previous ListEventSources operation. If present, specifies to continue the list from where the returning call left off.

", - "ListEventSourcesResponse$NextMarker": "

A string, present if there are more event source mappings.

", - "ListFunctionsRequest$Marker": "

Optional string. An opaque pagination token returned from a previous ListFunctions operation. If present, indicates where to continue the listing.

", - "ListFunctionsResponse$NextMarker": "

A string, present if there are more functions.

", - "Map$key": null, - "Map$value": null, - "RemoveEventSourceRequest$UUID": "

The event source mapping ID.

", - "ResourceNotFoundException$Type": null, - "ResourceNotFoundException$Message": null, - "ServiceException$Type": null, - "ServiceException$Message": null - } - }, - "Timeout": { - "base": null, - "refs": { - "FunctionConfiguration$Timeout": "

The function execution time at which Lambda should terminate the function. Because the execution time has cost implications, we recommend you set this value based on your expected execution time. The default is 3 seconds.

", - "UpdateFunctionConfigurationRequest$Timeout": "

The function execution time at which Lambda should terminate the function. Because the execution time has cost implications, we recommend you set this value based on your expected execution time. The default is 3 seconds.

", - "UploadFunctionRequest$Timeout": "

The function execution time at which Lambda should terminate the function. Because the execution time has cost implications, we recommend you set this value based on your expected execution time. The default is 3 seconds.

" - } - }, - "Timestamp": { - "base": null, - "refs": { - "EventSourceConfiguration$LastModified": "

The UTC time string indicating the last time the event mapping was updated.

", - "FunctionConfiguration$LastModified": "

The timestamp of the last time you updated the function.

" - } - }, - "UpdateFunctionConfigurationRequest": { - "base": null, - "refs": { - } - }, - "UploadFunctionRequest": { - "base": null, - "refs": { - } - }, - "Boolean": { - "base": null, - "refs": { - "EventSourceConfiguration$IsActive": "

Indicates whether the event source mapping is currently honored. Events are only processes if IsActive is true.

" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/paginators-1.json deleted file mode 100644 index deaf07d38..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2014-11-11/paginators-1.json +++ /dev/null @@ -1,16 +0,0 @@ -{ - "pagination": { - "ListEventSources": { - "input_token": "Marker", - "output_token": "NextMarker", - "limit_key": "MaxItems", - "result_key": "EventSources" - }, - "ListFunctions": { - "input_token": "Marker", - "output_token": "NextMarker", - "limit_key": "MaxItems", - "result_key": "Functions" - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/api-2.json index e5a5b3acd..be410f909 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/api-2.json @@ -108,7 +108,8 @@ {"shape":"ServiceException"}, {"shape":"ResourceNotFoundException"}, {"shape":"InvalidParameterValueException"}, - {"shape":"TooManyRequestsException"} + {"shape":"TooManyRequestsException"}, + {"shape":"ResourceInUseException"} ] }, "DeleteFunction":{ @@ -473,7 +474,8 @@ {"shape":"ResourceNotFoundException"}, {"shape":"InvalidParameterValueException"}, {"shape":"TooManyRequestsException"}, - {"shape":"ResourceConflictException"} + {"shape":"ResourceConflictException"}, + {"shape":"ResourceInUseException"} ] }, "UpdateFunctionCode":{ @@ -663,8 +665,7 @@ "type":"structure", "required":[ "EventSourceArn", - "FunctionName", - "StartingPosition" + "FunctionName" ], "members":{ "EventSourceArn":{"shape":"Arn"}, @@ -1526,6 +1527,15 @@ "error":{"httpStatusCode":409}, "exception":true }, + "ResourceInUseException":{ + "type":"structure", + "members":{ + "Type":{"shape":"String"}, + "Message":{"shape":"String"} + }, + "error":{"httpStatusCode":400}, + "exception":true + }, "ResourceNotFoundException":{ "type":"structure", "members":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/docs-2.json index 1d188cb78..1484d3915 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/lambda/2015-03-31/docs-2.json @@ -4,7 +4,7 @@ "operations": { "AddPermission": "

Adds a permission to the resource policy associated with the specified AWS Lambda function. You use resource policies to grant permissions to event sources that use push model. In a push model, event sources (such as Amazon S3 and custom applications) invoke your Lambda function. Each permission you add to the resource policy allows an event source, permission to invoke the Lambda function.

For information about the push model, see Lambda Functions.

If you are using versioning, the permissions you add are specific to the Lambda function version or alias you specify in the AddPermission request via the Qualifier parameter. For more information about versioning, see AWS Lambda Function Versioning and Aliases.

This operation requires permission for the lambda:AddPermission action.

", "CreateAlias": "

Creates an alias that points to the specified Lambda function version. For more information, see Introduction to AWS Lambda Aliases.

Alias names are unique for a given function. This requires permission for the lambda:CreateAlias action.

", - "CreateEventSourceMapping": "

Identifies a stream as an event source for a Lambda function. It can be either an Amazon Kinesis stream or an Amazon DynamoDB stream. AWS Lambda invokes the specified function when records are posted to the stream.

This association between a stream source and a Lambda function is called the event source mapping.

You provide mapping information (for example, which stream to read from and which Lambda function to invoke) in the request body.

Each event source, such as an Amazon Kinesis or a DynamoDB stream, can be associated with multiple AWS Lambda functions. A given Lambda function can be associated with multiple AWS event sources.

If you are using versioning, you can specify a specific function version or an alias via the function name parameter. For more information about versioning, see AWS Lambda Function Versioning and Aliases.

This operation requires permission for the lambda:CreateEventSourceMapping action.

", + "CreateEventSourceMapping": "

Identifies a poll-based event source for a Lambda function. It can be either an Amazon Kinesis or DynamoDB stream, or an Amazon SQS queue. AWS Lambda invokes the specified function when records are posted to the event source.

This association between a poll-based source and a Lambda function is called the event source mapping.

You provide mapping information (for example, which stream or SQS queue to read from and which Lambda function to invoke) in the request body.

Amazon Kinesis or DynamoDB stream event sources can be associated with multiple AWS Lambda functions and a given Lambda function can be associated with multiple AWS event sources. For Amazon SQS, you can configure multiple queues as event sources for a single Lambda function, but an SQS queue can be mapped only to a single Lambda function.

If you are using versioning, you can specify a specific function version or an alias via the function name parameter. For more information about versioning, see AWS Lambda Function Versioning and Aliases.

This operation requires permission for the lambda:CreateEventSourceMapping action.

", "CreateFunction": "

Creates a new Lambda function. The function metadata is created from the request parameters, and the code for the function is provided by a .zip file in the request body. If the function name already exists, the operation will fail. Note that the function name is case-sensitive.

If you are using versioning, you can also publish a version of the Lambda function you are creating using the Publish parameter. For more information about versioning, see AWS Lambda Function Versioning and Aliases.

This operation requires permission for the lambda:CreateFunction action.

", "DeleteAlias": "

Deletes the specified Lambda function alias. For more information, see Introduction to AWS Lambda Aliases.

This requires permission for the lambda:DeleteAlias action.

", "DeleteEventSourceMapping": "

Removes an event source mapping. This means AWS Lambda will no longer invoke the function for events in the associated source.

This operation requires permission for the lambda:DeleteEventSourceMapping action.

", @@ -108,15 +108,15 @@ "base": null, "refs": { "AddPermissionRequest$SourceArn": "

This is optional; however, when granting permission to invoke your function, you should specify this field with the Amazon Resource Name (ARN) as its value. This ensures that only events generated from the specified source can invoke the function.

If you add a permission without providing the source ARN, any AWS account that creates a mapping to your function ARN can send events to invoke your Lambda function.

", - "CreateEventSourceMappingRequest$EventSourceArn": "

The Amazon Resource Name (ARN) of the Amazon Kinesis or the Amazon DynamoDB stream that is the event source. Any record added to this stream could cause AWS Lambda to invoke your Lambda function, it depends on the BatchSize. AWS Lambda POSTs the Amazon Kinesis event, containing records, to your Lambda function as JSON.

", - "EventSourceMappingConfiguration$EventSourceArn": "

The Amazon Resource Name (ARN) of the Amazon Kinesis stream that is the source of events.

", - "ListEventSourceMappingsRequest$EventSourceArn": "

The Amazon Resource Name (ARN) of the Amazon Kinesis stream. (This parameter is optional.)

" + "CreateEventSourceMappingRequest$EventSourceArn": "

The Amazon Resource Name (ARN) of the event source. Any record added to this source could cause AWS Lambda to invoke your Lambda function, it depends on the BatchSize. AWS Lambda POSTs the event's records to your Lambda function as JSON.

", + "EventSourceMappingConfiguration$EventSourceArn": "

The Amazon Resource Name (ARN) of the Amazon Kinesis or DynamoDB stream or the SQS queue that is the source of events.

", + "ListEventSourceMappingsRequest$EventSourceArn": "

The Amazon Resource Name (ARN) of the Amazon Kinesis or DynamoDB stream, or an SQS queue. (This parameter is optional.)

" } }, "BatchSize": { "base": null, "refs": { - "CreateEventSourceMappingRequest$BatchSize": "

The largest number of records that AWS Lambda will retrieve from your event source at the time of invoking your function. Your function receives an event with all the retrieved records. The default is 100 records.

", + "CreateEventSourceMappingRequest$BatchSize": "

The largest number of records that AWS Lambda will retrieve from your event source at the time of invoking your function. Your function receives an event with all the retrieved records. The default for Amazon Kinesis and Amazon DynamoDB is 100 records. For SQS, the default is 1.

", "EventSourceMappingConfiguration$BatchSize": "

The largest number of records that AWS Lambda will retrieve from your event source at the time of invoking your function. Your function receives an event with all the retrieved records.

", "UpdateEventSourceMappingRequest$BatchSize": "

The maximum number of stream records that can be sent to your Lambda function for a single invocation.

" } @@ -283,7 +283,7 @@ } }, "EventSourceMappingConfiguration": { - "base": "

Describes mapping between an Amazon Kinesis stream and a Lambda function.

", + "base": "

Describes mapping between an Amazon Kinesis or DynamoDB stream or an Amazon SQS queue and a Lambda function.

", "refs": { "EventSourceMappingsList$member": null } @@ -310,7 +310,7 @@ "base": null, "refs": { "AliasConfiguration$AliasArn": "

Lambda function ARN that is qualified using the alias name as the suffix. For example, if you create an alias called BETA that points to a helloworld function version, the ARN is arn:aws:lambda:aws-regions:acct-id:function:helloworld:BETA.

", - "EventSourceMappingConfiguration$FunctionArn": "

The Lambda function to invoke when AWS Lambda detects an event on the stream.

", + "EventSourceMappingConfiguration$FunctionArn": "

The Lambda function to invoke when AWS Lambda detects an event on the poll-based source.

", "FunctionConfiguration$MasterArn": "

Returns the ARN (Amazon Resource Name) of the master function.

", "ListTagsRequest$Resource": "

The ARN (Amazon Resource Name) of the function. For more information, see Tagging Lambda Functions in the AWS Lambda Developer Guide.

", "TagResourceRequest$Resource": "

The ARN (Amazon Resource Name) of the Lambda function. For more information, see Tagging Lambda Functions in the AWS Lambda Developer Guide.

", @@ -701,6 +701,11 @@ "refs": { } }, + "ResourceInUseException": { + "base": "

The operation conflicts with the resource's availability. For example, you attempted to update an EventSoure Mapping in CREATING, or tried to delete a EventSoure mapping currently in the UPDATING state.

", + "refs": { + } + }, "ResourceNotFoundException": { "base": "

The resource (for example, a Lambda function or access policy statement) specified in the request does not exist.

", "refs": { @@ -851,6 +856,8 @@ "RequestTooLargeException$message": null, "ResourceConflictException$Type": "

", "ResourceConflictException$message": "

", + "ResourceInUseException$Type": null, + "ResourceInUseException$Message": null, "ResourceNotFoundException$Type": null, "ResourceNotFoundException$Message": null, "ServiceException$Type": null, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/api-2.json new file mode 100644 index 000000000..b5ea29b2c --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/api-2.json @@ -0,0 +1,365 @@ +{ + "version":"2.0", + "metadata":{ + "apiVersion":"2017-12-19", + "endpointPrefix":"macie", + "jsonVersion":"1.1", + "protocol":"json", + "serviceFullName":"Amazon Macie", + "serviceId":"Macie", + "signatureVersion":"v4", + "targetPrefix":"MacieService", + "uid":"macie-2017-12-19" + }, + "operations":{ + "AssociateMemberAccount":{ + "name":"AssociateMemberAccount", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"AssociateMemberAccountRequest"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"LimitExceededException"}, + {"shape":"InternalException"} + ] + }, + "AssociateS3Resources":{ + "name":"AssociateS3Resources", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"AssociateS3ResourcesRequest"}, + "output":{"shape":"AssociateS3ResourcesResult"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"AccessDeniedException"}, + {"shape":"LimitExceededException"}, + {"shape":"InternalException"} + ] + }, + "DisassociateMemberAccount":{ + "name":"DisassociateMemberAccount", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DisassociateMemberAccountRequest"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"InternalException"} + ] + }, + "DisassociateS3Resources":{ + "name":"DisassociateS3Resources", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DisassociateS3ResourcesRequest"}, + "output":{"shape":"DisassociateS3ResourcesResult"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"AccessDeniedException"}, + {"shape":"InternalException"} + ] + }, + "ListMemberAccounts":{ + "name":"ListMemberAccounts", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListMemberAccountsRequest"}, + "output":{"shape":"ListMemberAccountsResult"}, + "errors":[ + {"shape":"InternalException"}, + {"shape":"InvalidInputException"} + ] + }, + "ListS3Resources":{ + "name":"ListS3Resources", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListS3ResourcesRequest"}, + "output":{"shape":"ListS3ResourcesResult"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"AccessDeniedException"}, + {"shape":"InternalException"} + ] + }, + "UpdateS3Resources":{ + "name":"UpdateS3Resources", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"UpdateS3ResourcesRequest"}, + "output":{"shape":"UpdateS3ResourcesResult"}, + "errors":[ + {"shape":"InvalidInputException"}, + {"shape":"AccessDeniedException"}, + {"shape":"InternalException"} + ] + } + }, + "shapes":{ + "AWSAccountId":{ + "type":"string", + "pattern":"[0-9]{12}" + }, + "AccessDeniedException":{ + "type":"structure", + "members":{ + "message":{"shape":"ExceptionMessage"}, + "resourceType":{"shape":"ResourceType"} + }, + "exception":true + }, + "AssociateMemberAccountRequest":{ + "type":"structure", + "required":["memberAccountId"], + "members":{ + "memberAccountId":{"shape":"AWSAccountId"} + } + }, + "AssociateS3ResourcesRequest":{ + "type":"structure", + "required":["s3Resources"], + "members":{ + "memberAccountId":{"shape":"AWSAccountId"}, + "s3Resources":{"shape":"S3ResourcesClassification"} + } + }, + "AssociateS3ResourcesResult":{ + "type":"structure", + "members":{ + "failedS3Resources":{"shape":"FailedS3Resources"} + } + }, + "BucketName":{ + "type":"string", + "max":500 + }, + "ClassificationType":{ + "type":"structure", + "required":[ + "oneTime", + "continuous" + ], + "members":{ + "oneTime":{"shape":"S3OneTimeClassificationType"}, + "continuous":{"shape":"S3ContinuousClassificationType"} + } + }, + "ClassificationTypeUpdate":{ + "type":"structure", + "members":{ + "oneTime":{"shape":"S3OneTimeClassificationType"}, + "continuous":{"shape":"S3ContinuousClassificationType"} + } + }, + "DisassociateMemberAccountRequest":{ + "type":"structure", + "required":["memberAccountId"], + "members":{ + "memberAccountId":{"shape":"AWSAccountId"} + } + }, + "DisassociateS3ResourcesRequest":{ + "type":"structure", + "required":["associatedS3Resources"], + "members":{ + "memberAccountId":{"shape":"AWSAccountId"}, + "associatedS3Resources":{"shape":"S3Resources"} + } + }, + "DisassociateS3ResourcesResult":{ + "type":"structure", + "members":{ + "failedS3Resources":{"shape":"FailedS3Resources"} + } + }, + "ErrorCode":{ + "type":"string", + "max":10 + }, + "ExceptionMessage":{ + "type":"string", + "max":10000 + }, + "FailedS3Resource":{ + "type":"structure", + "members":{ + "failedItem":{"shape":"S3Resource"}, + "errorCode":{"shape":"ErrorCode"}, + "errorMessage":{"shape":"ExceptionMessage"} + } + }, + "FailedS3Resources":{ + "type":"list", + "member":{"shape":"FailedS3Resource"} + }, + "FieldName":{ + "type":"string", + "max":1000 + }, + "InternalException":{ + "type":"structure", + "members":{ + "errorCode":{"shape":"ErrorCode"}, + "message":{"shape":"ExceptionMessage"} + }, + "exception":true, + "fault":true + }, + "InvalidInputException":{ + "type":"structure", + "members":{ + "errorCode":{"shape":"ErrorCode"}, + "message":{"shape":"ExceptionMessage"}, + "fieldName":{"shape":"FieldName"} + }, + "exception":true + }, + "LimitExceededException":{ + "type":"structure", + "members":{ + "errorCode":{"shape":"ErrorCode"}, + "message":{"shape":"ExceptionMessage"}, + "resourceType":{"shape":"ResourceType"} + }, + "exception":true + }, + "ListMemberAccountsRequest":{ + "type":"structure", + "members":{ + "nextToken":{"shape":"NextToken"}, + "maxResults":{"shape":"MaxResults"} + } + }, + "ListMemberAccountsResult":{ + "type":"structure", + "members":{ + "memberAccounts":{"shape":"MemberAccounts"}, + "nextToken":{"shape":"NextToken"} + } + }, + "ListS3ResourcesRequest":{ + "type":"structure", + "members":{ + "memberAccountId":{"shape":"AWSAccountId"}, + "nextToken":{"shape":"NextToken"}, + "maxResults":{"shape":"MaxResults"} + } + }, + "ListS3ResourcesResult":{ + "type":"structure", + "members":{ + "s3Resources":{"shape":"S3ResourcesClassification"}, + "nextToken":{"shape":"NextToken"} + } + }, + "MaxResults":{ + "type":"integer", + "box":true, + "max":250 + }, + "MemberAccount":{ + "type":"structure", + "members":{ + "accountId":{"shape":"AWSAccountId"} + } + }, + "MemberAccounts":{ + "type":"list", + "member":{"shape":"MemberAccount"} + }, + "NextToken":{ + "type":"string", + "max":500 + }, + "Prefix":{ + "type":"string", + "max":10000 + }, + "ResourceType":{ + "type":"string", + "max":1000 + }, + "S3ContinuousClassificationType":{ + "type":"string", + "enum":["FULL"] + }, + "S3OneTimeClassificationType":{ + "type":"string", + "enum":[ + "FULL", + "NONE" + ] + }, + "S3Resource":{ + "type":"structure", + "required":["bucketName"], + "members":{ + "bucketName":{"shape":"BucketName"}, + "prefix":{"shape":"Prefix"} + } + }, + "S3ResourceClassification":{ + "type":"structure", + "required":[ + "bucketName", + "classificationType" + ], + "members":{ + "bucketName":{"shape":"BucketName"}, + "prefix":{"shape":"Prefix"}, + "classificationType":{"shape":"ClassificationType"} + } + }, + "S3ResourceClassificationUpdate":{ + "type":"structure", + "required":[ + "bucketName", + "classificationTypeUpdate" + ], + "members":{ + "bucketName":{"shape":"BucketName"}, + "prefix":{"shape":"Prefix"}, + "classificationTypeUpdate":{"shape":"ClassificationTypeUpdate"} + } + }, + "S3Resources":{ + "type":"list", + "member":{"shape":"S3Resource"} + }, + "S3ResourcesClassification":{ + "type":"list", + "member":{"shape":"S3ResourceClassification"} + }, + "S3ResourcesClassificationUpdate":{ + "type":"list", + "member":{"shape":"S3ResourceClassificationUpdate"} + }, + "UpdateS3ResourcesRequest":{ + "type":"structure", + "required":["s3ResourcesUpdate"], + "members":{ + "memberAccountId":{"shape":"AWSAccountId"}, + "s3ResourcesUpdate":{"shape":"S3ResourcesClassificationUpdate"} + } + }, + "UpdateS3ResourcesResult":{ + "type":"structure", + "members":{ + "failedS3Resources":{"shape":"FailedS3Resources"} + } + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/docs-2.json new file mode 100644 index 000000000..f0cb74fbb --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/docs-2.json @@ -0,0 +1,261 @@ +{ + "version":"2.0", + "service":"Amazon Macie

Amazon Macie is a security service that uses machine learning to automatically discover, classify, and protect sensitive data in AWS. Macie recognizes sensitive data such as personally identifiable information (PII) or intellectual property, and provides you with dashboards and alerts that give visibility into how this data is being accessed or moved. For more information, see the Macie User Guide.

", + "operations":{ + "AssociateMemberAccount":"

Associates a specified AWS account with Amazon Macie as a member account.

", + "AssociateS3Resources":"

Associates specified S3 resources with Amazon Macie for monitoring and data classification. If memberAccountId isn't specified, the action associates specified S3 resources with Macie for the current master account. If memberAccountId is specified, the action associates specified S3 resources with Macie for the specified member account.

", + "DisassociateMemberAccount":"

Removes the specified member account from Amazon Macie.

", + "DisassociateS3Resources":"

Removes specified S3 resources from being monitored by Amazon Macie. If memberAccountId isn't specified, the action removes specified S3 resources from Macie for the current master account. If memberAccountId is specified, the action removes specified S3 resources from Macie for the specified member account.

", + "ListMemberAccounts":"

Lists all Amazon Macie member accounts for the current Amazon Macie master account.

", + "ListS3Resources":"

Lists all the S3 resources associated with Amazon Macie. If memberAccountId isn't specified, the action lists the S3 resources associated with Amazon Macie for the current master account. If memberAccountId is specified, the action lists the S3 resources associated with Amazon Macie for the specified member account.

", + "UpdateS3Resources":"

Updates the classification types for the specified S3 resources. If memberAccountId isn't specified, the action updates the classification types of the S3 resources associated with Amazon Macie for the current master account. If memberAccountId is specified, the action updates the classification types of the S3 resources associated with Amazon Macie for the specified member account.

" + }, + "shapes":{ + "AWSAccountId":{ + "base":null, + "refs":{ + "AssociateMemberAccountRequest$memberAccountId":"

The ID of the AWS account that you want to associate with Amazon Macie as a member account.

", + "AssociateS3ResourcesRequest$memberAccountId":"

The ID of the Amazon Macie member account whose resources you want to associate with Macie.

", + "DisassociateMemberAccountRequest$memberAccountId":"

The ID of the member account that you want to remove from Amazon Macie.

", + "DisassociateS3ResourcesRequest$memberAccountId":"

The ID of the Amazon Macie member account whose resources you want to remove from being monitored by Amazon Macie.

", + "ListS3ResourcesRequest$memberAccountId":"

The Amazon Macie member account ID whose associated S3 resources you want to list.

", + "MemberAccount$accountId":"

The AWS account ID of the Amazon Macie member account.

", + "UpdateS3ResourcesRequest$memberAccountId":"

The AWS ID of the Amazon Macie member account whose S3 resources' classification types you want to update.

" + } + }, + "AccessDeniedException":{ + "base":"

You do not have required permissions to access the requested resource.

", + "refs":{ + } + }, + "AssociateMemberAccountRequest":{ + "base":null, + "refs":{ + } + }, + "AssociateS3ResourcesRequest":{ + "base":null, + "refs":{ + } + }, + "AssociateS3ResourcesResult":{ + "base":null, + "refs":{ + } + }, + "BucketName":{ + "base":null, + "refs":{ + "S3Resource$bucketName":"

The name of the S3 bucket.

", + "S3ResourceClassification$bucketName":"

The name of the S3 bucket that you want to associate with Amazon Macie.

", + "S3ResourceClassificationUpdate$bucketName":"

The name of the S3 bucket whose classification types you want to update.

" + } + }, + "ClassificationType":{ + "base":"

The classification type that Amazon Macie applies to the associated S3 resources.

", + "refs":{ + "S3ResourceClassification$classificationType":"

The classification type that you want to specify for the resource associated with Amazon Macie.

" + } + }, + "ClassificationTypeUpdate":{ + "base":"

The classification type that Amazon Macie applies to the associated S3 resources. At least one of the classification types (oneTime or continuous) must be specified.

", + "refs":{ + "S3ResourceClassificationUpdate$classificationTypeUpdate":"

The classification type that you want to update for the resource associated with Amazon Macie.

" + } + }, + "DisassociateMemberAccountRequest":{ + "base":null, + "refs":{ + } + }, + "DisassociateS3ResourcesRequest":{ + "base":null, + "refs":{ + } + }, + "DisassociateS3ResourcesResult":{ + "base":null, + "refs":{ + } + }, + "ErrorCode":{ + "base":"Error code for the exception", + "refs":{ + "FailedS3Resource$errorCode":"

The status code of a failed item.

", + "InternalException$errorCode":null, + "InvalidInputException$errorCode":null, + "LimitExceededException$errorCode":null + } + }, + "ExceptionMessage":{ + "base":null, + "refs":{ + "AccessDeniedException$message":null, + "FailedS3Resource$errorMessage":"

The error message of a failed item.

", + "InternalException$message":null, + "InvalidInputException$message":null, + "LimitExceededException$message":null + } + }, + "FailedS3Resource":{ + "base":"

Includes details about the failed S3 resources.

", + "refs":{ + "FailedS3Resources$member":null + } + }, + "FailedS3Resources":{ + "base":null, + "refs":{ + "AssociateS3ResourcesResult$failedS3Resources":"

S3 resources that couldn't be associated with Amazon Macie. An error code and an error message are provided for each failed item.

", + "DisassociateS3ResourcesResult$failedS3Resources":"

S3 resources that couldn't be removed from being monitored and classified by Amazon Macie. An error code and an error message are provided for each failed item.

", + "UpdateS3ResourcesResult$failedS3Resources":"

The S3 resources whose classification types can't be updated. An error code and an error message are provided for each failed item.

" + } + }, + "FieldName":{ + "base":"Field that has invalid input", + "refs":{ + "InvalidInputException$fieldName":null + } + }, + "InternalException":{ + "base":"

Internal server error.

", + "refs":{ + } + }, + "InvalidInputException":{ + "base":"

The request was rejected because an invalid or out-of-range value was supplied for an input parameter.

", + "refs":{ + } + }, + "LimitExceededException":{ + "base":"

The request was rejected because it attempted to create resources beyond the current AWS account limits. The error code describes the limit exceeded.

", + "refs":{ + } + }, + "ListMemberAccountsRequest":{ + "base":null, + "refs":{ + } + }, + "ListMemberAccountsResult":{ + "base":null, + "refs":{ + } + }, + "ListS3ResourcesRequest":{ + "base":null, + "refs":{ + } + }, + "ListS3ResourcesResult":{ + "base":null, + "refs":{ + } + }, + "MaxResults":{ + "base":null, + "refs":{ + "ListMemberAccountsRequest$maxResults":"

Use this parameter to indicate the maximum number of items that you want in the response. The default value is 250.

", + "ListS3ResourcesRequest$maxResults":"

Use this parameter to indicate the maximum number of items that you want in the response. The default value is 250.

" + } + }, + "MemberAccount":{ + "base":"

Contains information about the Amazon Macie member account.

", + "refs":{ + "MemberAccounts$member":null + } + }, + "MemberAccounts":{ + "base":null, + "refs":{ + "ListMemberAccountsResult$memberAccounts":"

A list of the Amazon Macie member accounts returned by the action. The current master account is also included in this list.

" + } + }, + "NextToken":{ + "base":null, + "refs":{ + "ListMemberAccountsRequest$nextToken":"

Use this parameter when paginating results. Set the value of this parameter to null on your first call to the ListMemberAccounts action. Subsequent calls to the action fill nextToken in the request with the value of nextToken from the previous response to continue listing data.

", + "ListMemberAccountsResult$nextToken":"

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

", + "ListS3ResourcesRequest$nextToken":"

Use this parameter when paginating results. Set its value to null on your first call to the ListS3Resources action. Subsequent calls to the action fill nextToken in the request with the value of nextToken from the previous response to continue listing data.

", + "ListS3ResourcesResult$nextToken":"

When a response is generated, if there is more data to be listed, this parameter is present in the response and contains the value to use for the nextToken parameter in a subsequent pagination request. If there is no more data to be listed, this parameter is set to null.

" + } + }, + "Prefix":{ + "base":null, + "refs":{ + "S3Resource$prefix":"

The prefix of the S3 bucket.

", + "S3ResourceClassification$prefix":"

The prefix of the S3 bucket that you want to associate with Amazon Macie.

", + "S3ResourceClassificationUpdate$prefix":"

The prefix of the S3 bucket whose classification types you want to update.

" + } + }, + "ResourceType":{ + "base":"Resource type that caused the exception", + "refs":{ + "AccessDeniedException$resourceType":null, + "LimitExceededException$resourceType":null + } + }, + "S3ContinuousClassificationType":{ + "base":null, + "refs":{ + "ClassificationType$continuous":"

A continuous classification of the objects that are added to a specified S3 bucket. Amazon Macie begins performing continuous classification after a bucket is successfully associated with Amazon Macie.

", + "ClassificationTypeUpdate$continuous":"

A continuous classification of the objects that are added to a specified S3 bucket. Amazon Macie begins performing continuous classification after a bucket is successfully associated with Amazon Macie.

" + } + }, + "S3OneTimeClassificationType":{ + "base":null, + "refs":{ + "ClassificationType$oneTime":"

A one-time classification of all of the existing objects in a specified S3 bucket.

", + "ClassificationTypeUpdate$oneTime":"

A one-time classification of all of the existing objects in a specified S3 bucket.

" + } + }, + "S3Resource":{ + "base":"

Contains information about the S3 resource. This data type is used as a request parameter in the DisassociateS3Resources action and can be used as a response parameter in the AssociateS3Resources and UpdateS3Resources actions.

", + "refs":{ + "FailedS3Resource$failedItem":"

The failed S3 resources.

", + "S3Resources$member":null + } + }, + "S3ResourceClassification":{ + "base":"

The S3 resources that you want to associate with Amazon Macie for monitoring and data classification. This data type is used as a request parameter in the AssociateS3Resources action and a response parameter in the ListS3Resources action.

", + "refs":{ + "S3ResourcesClassification$member":null + } + }, + "S3ResourceClassificationUpdate":{ + "base":"

The S3 resources whose classification types you want to update. This data type is used as a request parameter in the UpdateS3Resources action.

", + "refs":{ + "S3ResourcesClassificationUpdate$member":null + } + }, + "S3Resources":{ + "base":null, + "refs":{ + "DisassociateS3ResourcesRequest$associatedS3Resources":"

The S3 resources (buckets or prefixes) that you want to remove from being monitored and classified by Amazon Macie.

" + } + }, + "S3ResourcesClassification":{ + "base":null, + "refs":{ + "AssociateS3ResourcesRequest$s3Resources":"

The S3 resources that you want to associate with Amazon Macie for monitoring and data classification.

", + "ListS3ResourcesResult$s3Resources":"

A list of the associated S3 resources returned by the action.

" + } + }, + "S3ResourcesClassificationUpdate":{ + "base":null, + "refs":{ + "UpdateS3ResourcesRequest$s3ResourcesUpdate":"

The S3 resources whose classification types you want to update.

" + } + }, + "UpdateS3ResourcesRequest":{ + "base":null, + "refs":{ + } + }, + "UpdateS3ResourcesResult":{ + "base":null, + "refs":{ + } + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/examples-1.json similarity index 100% rename from vendor/github.com/aws/aws-sdk-go/models/apis/events/2014-02-03/examples-1.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/examples-1.json diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/paginators-1.json new file mode 100644 index 000000000..0b3465fca --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/macie/2017-12-19/paginators-1.json @@ -0,0 +1,14 @@ +{ + "pagination":{ + "ListMemberAccounts":{ + "input_token":"nextToken", + "output_token":"nextToken", + "limit_key":"maxResults" + }, + "ListS3Resources":{ + "input_token":"nextToken", + "output_token":"nextToken", + "limit_key":"maxResults" + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/api-2.json index 75be4657b..53998d656 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/api-2.json @@ -741,7 +741,7 @@ "locationName": "audioDescriptionBroadcasterMix" }, "Bitrate": { - "shape": "__integer", + "shape": "__integerMin6000Max1024000", "locationName": "bitrate" }, "CodecProfile": { @@ -761,7 +761,7 @@ "locationName": "rawFormat" }, "SampleRate": { - "shape": "__integer", + "shape": "__integerMin8000Max96000", "locationName": "sampleRate" }, "Specification": { @@ -772,7 +772,11 @@ "shape": "AacVbrQuality", "locationName": "vbrQuality" } - } + }, + "required": [ + "CodingMode", + "SampleRate" + ] }, "AacSpecification": { "type": "string", @@ -837,7 +841,7 @@ "type": "structure", "members": { "Bitrate": { - "shape": "__integer", + "shape": "__integerMin64000Max640000", "locationName": "bitrate" }, "BitstreamMode": { @@ -849,7 +853,7 @@ "locationName": "codingMode" }, "Dialnorm": { - "shape": "__integer", + "shape": "__integerMin1Max31", "locationName": "dialnorm" }, "DynamicRangeCompressionProfile": { @@ -865,7 +869,7 @@ "locationName": "metadataControl" }, "SampleRate": { - "shape": "__integer", + "shape": "__integerMin48000Max48000", "locationName": "sampleRate" } } @@ -882,15 +886,15 @@ "type": "structure", "members": { "BitDepth": { - "shape": "__integer", + "shape": "__integerMin16Max24", "locationName": "bitDepth" }, "Channels": { - "shape": "__integer", + "shape": "__integerMin1Max2", "locationName": "channels" }, "SampleRate": { - "shape": "__integer", + "shape": "__integerMin8000Max192000", "locationName": "sampleRate" } } @@ -899,7 +903,7 @@ "type": "structure", "members": { "SourceAncillaryChannelNumber": { - "shape": "__integer", + "shape": "__integerMin1Max4", "locationName": "sourceAncillaryChannelNumber" } } @@ -954,7 +958,10 @@ "shape": "WavSettings", "locationName": "wavSettings" } - } + }, + "required": [ + "Codec" + ] }, "AudioDefaultSelection": { "type": "string", @@ -975,7 +982,7 @@ "locationName": "audioSourceName" }, "AudioType": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "audioType" }, "AudioTypeControl": { @@ -986,6 +993,10 @@ "shape": "AudioCodecSettings", "locationName": "codecSettings" }, + "CustomLanguageCode": { + "shape": "__stringMin3Max3PatternAZaZ3", + "locationName": "customLanguageCode" + }, "LanguageCode": { "shape": "LanguageCode", "locationName": "languageCode" @@ -999,10 +1010,13 @@ "locationName": "remixSettings" }, "StreamName": { - "shape": "__string", + "shape": "__stringPatternWS", "locationName": "streamName" } - } + }, + "required": [ + "CodecSettings" + ] }, "AudioLanguageCodeControl": { "type": "string", @@ -1051,7 +1065,7 @@ "locationName": "algorithmControl" }, "CorrectionGateLevel": { - "shape": "__integer", + "shape": "__integerMinNegative70Max0", "locationName": "correctionGateLevel" }, "LoudnessLogging": { @@ -1063,7 +1077,7 @@ "locationName": "peakCalculation" }, "TargetLkfs": { - "shape": "__double", + "shape": "__doubleMinNegative59Max0", "locationName": "targetLkfs" } } @@ -1071,12 +1085,16 @@ "AudioSelector": { "type": "structure", "members": { + "CustomLanguageCode": { + "shape": "__stringMin3Max3PatternAZaZ3", + "locationName": "customLanguageCode" + }, "DefaultSelection": { "shape": "AudioDefaultSelection", "locationName": "defaultSelection" }, "ExternalAudioFileInput": { - "shape": "__string", + "shape": "__stringPatternS3MM2VVMMPPEEGGAAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MMAAAACCAAIIFFFFMMPP2AACC3EECC3DDTTSSEE", "locationName": "externalAudioFileInput" }, "LanguageCode": { @@ -1084,15 +1102,15 @@ "locationName": "languageCode" }, "Offset": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "offset" }, "Pids": { - "shape": "ListOf__integer", + "shape": "__listOf__integerMin1Max2147483647", "locationName": "pids" }, "ProgramSelection": { - "shape": "__integer", + "shape": "__integerMin0Max8", "locationName": "programSelection" }, "RemixSettings": { @@ -1104,7 +1122,7 @@ "locationName": "selectorType" }, "Tracks": { - "shape": "ListOf__integer", + "shape": "__listOf__integerMin1Max2147483647", "locationName": "tracks" } } @@ -1113,10 +1131,13 @@ "type": "structure", "members": { "AudioSelectorNames": { - "shape": "ListOf__string", + "shape": "__listOf__stringMin1", "locationName": "audioSelectorNames" } - } + }, + "required": [ + "AudioSelectorNames" + ] }, "AudioSelectorType": { "type": "string", @@ -1137,7 +1158,7 @@ "type": "structure", "members": { "AvailBlankingImage": { - "shape": "__string", + "shape": "__stringMin14PatternS3BmpBMPPngPNG", "locationName": "availBlankingImage" } } @@ -1167,7 +1188,7 @@ "locationName": "backgroundColor" }, "BackgroundOpacity": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "backgroundOpacity" }, "FontColor": { @@ -1175,15 +1196,15 @@ "locationName": "fontColor" }, "FontOpacity": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "fontOpacity" }, "FontResolution": { - "shape": "__integer", + "shape": "__integerMin96Max600", "locationName": "fontResolution" }, "FontSize": { - "shape": "__integer", + "shape": "__integerMin0Max96", "locationName": "fontSize" }, "OutlineColor": { @@ -1191,7 +1212,7 @@ "locationName": "outlineColor" }, "OutlineSize": { - "shape": "__integer", + "shape": "__integerMin0Max10", "locationName": "outlineSize" }, "ShadowColor": { @@ -1199,15 +1220,15 @@ "locationName": "shadowColor" }, "ShadowOpacity": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "shadowOpacity" }, "ShadowXOffset": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "shadowXOffset" }, "ShadowYOffset": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "shadowYOffset" }, "TeletextSpacing": { @@ -1215,14 +1236,20 @@ "locationName": "teletextSpacing" }, "XPosition": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "xPosition" }, "YPosition": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "yPosition" } - } + }, + "required": [ + "OutlineColor", + "Alignment", + "OutlineSize", + "FontOpacity" + ] }, "BurninSubtitleAlignment": { "type": "string", @@ -1298,9 +1325,13 @@ "type": "structure", "members": { "CaptionSelectorName": { - "shape": "__string", + "shape": "__stringMin1", "locationName": "captionSelectorName" }, + "CustomLanguageCode": { + "shape": "__stringMin3Max3PatternAZaZ3", + "locationName": "customLanguageCode" + }, "DestinationSettings": { "shape": "CaptionDestinationSettings", "locationName": "destinationSettings" @@ -1313,11 +1344,19 @@ "shape": "__string", "locationName": "languageDescription" } - } + }, + "required": [ + "DestinationSettings", + "CaptionSelectorName" + ] }, "CaptionDescriptionPreset": { "type": "structure", "members": { + "CustomLanguageCode": { + "shape": "__stringMin3Max3PatternAZaZ3", + "locationName": "customLanguageCode" + }, "DestinationSettings": { "shape": "CaptionDestinationSettings", "locationName": "destinationSettings" @@ -1330,7 +1369,10 @@ "shape": "__string", "locationName": "languageDescription" } - } + }, + "required": [ + "DestinationSettings" + ] }, "CaptionDestinationSettings": { "type": "structure", @@ -1359,7 +1401,10 @@ "shape": "TtmlDestinationSettings", "locationName": "ttmlDestinationSettings" } - } + }, + "required": [ + "DestinationType" + ] }, "CaptionDestinationType": { "type": "string", @@ -1377,6 +1422,10 @@ "CaptionSelector": { "type": "structure", "members": { + "CustomLanguageCode": { + "shape": "__stringMin3Max3PatternAZaZ3", + "locationName": "customLanguageCode" + }, "LanguageCode": { "shape": "LanguageCode", "locationName": "languageCode" @@ -1385,7 +1434,10 @@ "shape": "CaptionSourceSettings", "locationName": "sourceSettings" } - } + }, + "required": [ + "SourceSettings" + ] }, "CaptionSourceSettings": { "type": "structure", @@ -1414,7 +1466,10 @@ "shape": "TeletextSourceSettings", "locationName": "teletextSourceSettings" } - } + }, + "required": [ + "SourceType" + ] }, "CaptionSourceType": { "type": "string", @@ -1434,16 +1489,187 @@ "type": "structure", "members": { "OutputChannels": { - "shape": "ListOfOutputChannelMapping", + "shape": "__listOfOutputChannelMapping", "locationName": "outputChannels" } - } + }, + "required": [ + "OutputChannels" + ] + }, + "CmafClientCache": { + "type": "string", + "enum": [ + "DISABLED", + "ENABLED" + ] + }, + "CmafCodecSpecification": { + "type": "string", + "enum": [ + "RFC_6381", + "RFC_4281" + ] + }, + "CmafEncryptionSettings": { + "type": "structure", + "members": { + "ConstantInitializationVector": { + "shape": "__stringMin32Max32Pattern09aFAF32", + "locationName": "constantInitializationVector" + }, + "EncryptionMethod": { + "shape": "CmafEncryptionType", + "locationName": "encryptionMethod" + }, + "InitializationVectorInManifest": { + "shape": "CmafInitializationVectorInManifest", + "locationName": "initializationVectorInManifest" + }, + "StaticKeyProvider": { + "shape": "StaticKeyProvider", + "locationName": "staticKeyProvider" + }, + "Type": { + "shape": "CmafKeyProviderType", + "locationName": "type" + } + }, + "required": [ + "Type" + ] + }, + "CmafEncryptionType": { + "type": "string", + "enum": [ + "SAMPLE_AES" + ] + }, + "CmafGroupSettings": { + "type": "structure", + "members": { + "BaseUrl": { + "shape": "__string", + "locationName": "baseUrl" + }, + "ClientCache": { + "shape": "CmafClientCache", + "locationName": "clientCache" + }, + "CodecSpecification": { + "shape": "CmafCodecSpecification", + "locationName": "codecSpecification" + }, + "Destination": { + "shape": "__stringPatternS3", + "locationName": "destination" + }, + "Encryption": { + "shape": "CmafEncryptionSettings", + "locationName": "encryption" + }, + "FragmentLength": { + "shape": "__integerMin1Max2147483647", + "locationName": "fragmentLength" + }, + "ManifestCompression": { + "shape": "CmafManifestCompression", + "locationName": "manifestCompression" + }, + "ManifestDurationFormat": { + "shape": "CmafManifestDurationFormat", + "locationName": "manifestDurationFormat" + }, + "MinBufferTime": { + "shape": "__integerMin0Max2147483647", + "locationName": "minBufferTime" + }, + "SegmentControl": { + "shape": "CmafSegmentControl", + "locationName": "segmentControl" + }, + "SegmentLength": { + "shape": "__integerMin1Max2147483647", + "locationName": "segmentLength" + }, + "StreamInfResolution": { + "shape": "CmafStreamInfResolution", + "locationName": "streamInfResolution" + }, + "WriteDashManifest": { + "shape": "CmafWriteDASHManifest", + "locationName": "writeDashManifest" + }, + "WriteHlsManifest": { + "shape": "CmafWriteHLSManifest", + "locationName": "writeHlsManifest" + } + }, + "required": [ + "FragmentLength", + "SegmentLength" + ] + }, + "CmafInitializationVectorInManifest": { + "type": "string", + "enum": [ + "INCLUDE", + "EXCLUDE" + ] + }, + "CmafKeyProviderType": { + "type": "string", + "enum": [ + "STATIC_KEY" + ] + }, + "CmafManifestCompression": { + "type": "string", + "enum": [ + "GZIP", + "NONE" + ] + }, + "CmafManifestDurationFormat": { + "type": "string", + "enum": [ + "FLOATING_POINT", + "INTEGER" + ] + }, + "CmafSegmentControl": { + "type": "string", + "enum": [ + "SINGLE_FILE", + "SEGMENTED_FILES" + ] + }, + "CmafStreamInfResolution": { + "type": "string", + "enum": [ + "INCLUDE", + "EXCLUDE" + ] + }, + "CmafWriteDASHManifest": { + "type": "string", + "enum": [ + "DISABLED", + "ENABLED" + ] + }, + "CmafWriteHLSManifest": { + "type": "string", + "enum": [ + "DISABLED", + "ENABLED" + ] }, "ColorCorrector": { "type": "structure", "members": { "Brightness": { - "shape": "__integer", + "shape": "__integerMin1Max100", "locationName": "brightness" }, "ColorSpaceConversion": { @@ -1451,7 +1677,7 @@ "locationName": "colorSpaceConversion" }, "Contrast": { - "shape": "__integer", + "shape": "__integerMin1Max100", "locationName": "contrast" }, "Hdr10Metadata": { @@ -1459,11 +1685,11 @@ "locationName": "hdr10Metadata" }, "Hue": { - "shape": "__integer", + "shape": "__integerMinNegative180Max180", "locationName": "hue" }, "Saturation": { - "shape": "__integer", + "shape": "__integerMin1Max100", "locationName": "saturation" } } @@ -1542,7 +1768,10 @@ "shape": "Mp4Settings", "locationName": "mp4Settings" } - } + }, + "required": [ + "Container" + ] }, "ContainerType": { "type": "string", @@ -1551,6 +1780,7 @@ "ISMV", "M2TS", "M3U8", + "CMFC", "MOV", "MP4", "MPD", @@ -1583,10 +1813,14 @@ "locationName": "settings" }, "UserMetadata": { - "shape": "MapOf__string", + "shape": "__mapOf__string", "locationName": "userMetadata" } - } + }, + "required": [ + "Role", + "Settings" + ] }, "CreateJobResponse": { "type": "structure", @@ -1620,7 +1854,11 @@ "shape": "JobTemplateSettings", "locationName": "settings" } - } + }, + "required": [ + "Settings", + "Name" + ] }, "CreateJobTemplateResponse": { "type": "structure", @@ -1650,7 +1888,11 @@ "shape": "PresetSettings", "locationName": "settings" } - } + }, + "required": [ + "Settings", + "Name" + ] }, "CreatePresetResponse": { "type": "structure", @@ -1672,7 +1914,10 @@ "shape": "__string", "locationName": "name" } - } + }, + "required": [ + "Name" + ] }, "CreateQueueResponse": { "type": "structure", @@ -1690,7 +1935,10 @@ "shape": "SpekeKeyProvider", "locationName": "spekeKeyProvider" } - } + }, + "required": [ + "SpekeKeyProvider" + ] }, "DashIsoGroupSettings": { "type": "structure", @@ -1700,7 +1948,7 @@ "locationName": "baseUrl" }, "Destination": { - "shape": "__string", + "shape": "__stringPatternS3", "locationName": "destination" }, "Encryption": { @@ -1708,7 +1956,7 @@ "locationName": "encryption" }, "FragmentLength": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "fragmentLength" }, "HbbtvCompliance": { @@ -1716,7 +1964,7 @@ "locationName": "hbbtvCompliance" }, "MinBufferTime": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "minBufferTime" }, "SegmentControl": { @@ -1724,10 +1972,14 @@ "locationName": "segmentControl" }, "SegmentLength": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "segmentLength" } - } + }, + "required": [ + "SegmentLength", + "FragmentLength" + ] }, "DashIsoHbbtvCompliance": { "type": "string", @@ -1855,7 +2107,7 @@ "type": "structure", "members": { "Endpoints": { - "shape": "ListOfEndpoint", + "shape": "__listOfEndpoint", "locationName": "endpoints" }, "NextToken": { @@ -1875,18 +2127,23 @@ "type": "structure", "members": { "NetworkId": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "networkId" }, "NetworkName": { - "shape": "__string", + "shape": "__stringMin1Max256", "locationName": "networkName" }, "NitInterval": { - "shape": "__integer", + "shape": "__integerMin25Max10000", "locationName": "nitInterval" } - } + }, + "required": [ + "NetworkName", + "NitInterval", + "NetworkId" + ] }, "DvbSdtSettings": { "type": "structure", @@ -1896,15 +2153,15 @@ "locationName": "outputSdt" }, "SdtInterval": { - "shape": "__integer", + "shape": "__integerMin25Max2000", "locationName": "sdtInterval" }, "ServiceName": { - "shape": "__string", + "shape": "__stringMin1Max256", "locationName": "serviceName" }, "ServiceProviderName": { - "shape": "__string", + "shape": "__stringMin1Max256", "locationName": "serviceProviderName" } } @@ -1921,7 +2178,7 @@ "locationName": "backgroundColor" }, "BackgroundOpacity": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "backgroundOpacity" }, "FontColor": { @@ -1929,15 +2186,15 @@ "locationName": "fontColor" }, "FontOpacity": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "fontOpacity" }, "FontResolution": { - "shape": "__integer", + "shape": "__integerMin96Max600", "locationName": "fontResolution" }, "FontSize": { - "shape": "__integer", + "shape": "__integerMin0Max96", "locationName": "fontSize" }, "OutlineColor": { @@ -1945,7 +2202,7 @@ "locationName": "outlineColor" }, "OutlineSize": { - "shape": "__integer", + "shape": "__integerMin0Max10", "locationName": "outlineSize" }, "ShadowColor": { @@ -1953,15 +2210,15 @@ "locationName": "shadowColor" }, "ShadowOpacity": { - "shape": "__integer", + "shape": "__integerMin0Max255", "locationName": "shadowOpacity" }, "ShadowXOffset": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "shadowXOffset" }, "ShadowYOffset": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "shadowYOffset" }, "TeletextSpacing": { @@ -1969,20 +2226,26 @@ "locationName": "teletextSpacing" }, "XPosition": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "xPosition" }, "YPosition": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "yPosition" } - } + }, + "required": [ + "OutlineColor", + "Alignment", + "OutlineSize", + "FontOpacity" + ] }, "DvbSubSourceSettings": { "type": "structure", "members": { "Pid": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "pid" } } @@ -2043,10 +2306,13 @@ "type": "structure", "members": { "TdtInterval": { - "shape": "__integer", + "shape": "__integerMin1000Max30000", "locationName": "tdtInterval" } - } + }, + "required": [ + "TdtInterval" + ] }, "Eac3AttenuationControl": { "type": "string", @@ -2145,7 +2411,7 @@ "locationName": "attenuationControl" }, "Bitrate": { - "shape": "__integer", + "shape": "__integerMin64000Max640000", "locationName": "bitrate" }, "BitstreamMode": { @@ -2161,7 +2427,7 @@ "locationName": "dcFilter" }, "Dialnorm": { - "shape": "__integer", + "shape": "__integerMin1Max31", "locationName": "dialnorm" }, "DynamicRangeCompressionLine": { @@ -2181,19 +2447,19 @@ "locationName": "lfeFilter" }, "LoRoCenterMixLevel": { - "shape": "__double", + "shape": "__doubleMinNegative60Max3", "locationName": "loRoCenterMixLevel" }, "LoRoSurroundMixLevel": { - "shape": "__double", + "shape": "__doubleMinNegative60MaxNegative1", "locationName": "loRoSurroundMixLevel" }, "LtRtCenterMixLevel": { - "shape": "__double", + "shape": "__doubleMinNegative60Max3", "locationName": "ltRtCenterMixLevel" }, "LtRtSurroundMixLevel": { - "shape": "__double", + "shape": "__doubleMinNegative60MaxNegative1", "locationName": "ltRtSurroundMixLevel" }, "MetadataControl": { @@ -2209,7 +2475,7 @@ "locationName": "phaseControl" }, "SampleRate": { - "shape": "__integer", + "shape": "__integerMin48000Max48000", "locationName": "sampleRate" }, "StereoDownmix": { @@ -2266,11 +2532,11 @@ "locationName": "convert608To708" }, "Source608ChannelNumber": { - "shape": "__integer", + "shape": "__integerMin1Max4", "locationName": "source608ChannelNumber" }, "Source608TrackNumber": { - "shape": "__integer", + "shape": "__integerMin1Max1", "locationName": "source608TrackNumber" } } @@ -2313,7 +2579,7 @@ "type": "structure", "members": { "Destination": { - "shape": "__string", + "shape": "__stringPatternS3", "locationName": "destination" } } @@ -2333,14 +2599,17 @@ "locationName": "convert608To708" }, "SourceFile": { - "shape": "__string", + "shape": "__stringMin14PatternS3SccSCCTtmlTTMLDfxpDFXPStlSTLSrtSRTSmiSMI", "locationName": "sourceFile" }, "TimeDelta": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "timeDelta" } - } + }, + "required": [ + "SourceFile" + ] }, "ForbiddenException": { "type": "structure", @@ -2359,19 +2628,19 @@ "type": "structure", "members": { "FramerateDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateDenominator" }, "FramerateNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateNumerator" }, "MaxCaptures": { - "shape": "__integer", + "shape": "__integerMin1Max10000000", "locationName": "maxCaptures" }, "Quality": { - "shape": "__integer", + "shape": "__integerMin1Max100", "locationName": "quality" } } @@ -2611,7 +2880,7 @@ "locationName": "adaptiveQuantization" }, "Bitrate": { - "shape": "__integer", + "shape": "__integerMin1000Max1152000000", "locationName": "bitrate" }, "CodecLevel": { @@ -2643,11 +2912,11 @@ "locationName": "framerateConversionAlgorithm" }, "FramerateDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateDenominator" }, "FramerateNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateNumerator" }, "GopBReference": { @@ -2655,11 +2924,11 @@ "locationName": "gopBReference" }, "GopClosedCadence": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "gopClosedCadence" }, "GopSize": { - "shape": "__double", + "shape": "__doubleMin0", "locationName": "gopSize" }, "GopSizeUnits": { @@ -2667,11 +2936,11 @@ "locationName": "gopSizeUnits" }, "HrdBufferInitialFillPercentage": { - "shape": "__integer", + "shape": "__integerMin0Max100", "locationName": "hrdBufferInitialFillPercentage" }, "HrdBufferSize": { - "shape": "__integer", + "shape": "__integerMin0Max1152000000", "locationName": "hrdBufferSize" }, "InterlaceMode": { @@ -2679,19 +2948,19 @@ "locationName": "interlaceMode" }, "MaxBitrate": { - "shape": "__integer", + "shape": "__integerMin1000Max1152000000", "locationName": "maxBitrate" }, "MinIInterval": { - "shape": "__integer", + "shape": "__integerMin0Max30", "locationName": "minIInterval" }, "NumberBFramesBetweenReferenceFrames": { - "shape": "__integer", + "shape": "__integerMin0Max7", "locationName": "numberBFramesBetweenReferenceFrames" }, "NumberReferenceFrames": { - "shape": "__integer", + "shape": "__integerMin1Max6", "locationName": "numberReferenceFrames" }, "ParControl": { @@ -2699,11 +2968,11 @@ "locationName": "parControl" }, "ParDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parDenominator" }, "ParNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parNumerator" }, "QualityTuningLevel": { @@ -2723,7 +2992,7 @@ "locationName": "sceneChangeDetect" }, "Slices": { - "shape": "__integer", + "shape": "__integerMin1Max32", "locationName": "slices" }, "SlowPal": { @@ -2731,7 +3000,7 @@ "locationName": "slowPal" }, "Softness": { - "shape": "__integer", + "shape": "__integerMin0Max128", "locationName": "softness" }, "SpatialAdaptiveQuantization": { @@ -2943,7 +3212,7 @@ "locationName": "alternateTransferFunctionSei" }, "Bitrate": { - "shape": "__integer", + "shape": "__integerMin1000Max1466400000", "locationName": "bitrate" }, "CodecLevel": { @@ -2967,11 +3236,11 @@ "locationName": "framerateConversionAlgorithm" }, "FramerateDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateDenominator" }, "FramerateNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateNumerator" }, "GopBReference": { @@ -2979,11 +3248,11 @@ "locationName": "gopBReference" }, "GopClosedCadence": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "gopClosedCadence" }, "GopSize": { - "shape": "__double", + "shape": "__doubleMin0", "locationName": "gopSize" }, "GopSizeUnits": { @@ -2991,11 +3260,11 @@ "locationName": "gopSizeUnits" }, "HrdBufferInitialFillPercentage": { - "shape": "__integer", + "shape": "__integerMin0Max100", "locationName": "hrdBufferInitialFillPercentage" }, "HrdBufferSize": { - "shape": "__integer", + "shape": "__integerMin0Max1466400000", "locationName": "hrdBufferSize" }, "InterlaceMode": { @@ -3003,19 +3272,19 @@ "locationName": "interlaceMode" }, "MaxBitrate": { - "shape": "__integer", + "shape": "__integerMin1000Max1466400000", "locationName": "maxBitrate" }, "MinIInterval": { - "shape": "__integer", + "shape": "__integerMin0Max30", "locationName": "minIInterval" }, "NumberBFramesBetweenReferenceFrames": { - "shape": "__integer", + "shape": "__integerMin0Max7", "locationName": "numberBFramesBetweenReferenceFrames" }, "NumberReferenceFrames": { - "shape": "__integer", + "shape": "__integerMin1Max6", "locationName": "numberReferenceFrames" }, "ParControl": { @@ -3023,11 +3292,11 @@ "locationName": "parControl" }, "ParDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parDenominator" }, "ParNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parNumerator" }, "QualityTuningLevel": { @@ -3047,7 +3316,7 @@ "locationName": "sceneChangeDetect" }, "Slices": { - "shape": "__integer", + "shape": "__integerMin1Max32", "locationName": "slices" }, "SlowPal": { @@ -3077,6 +3346,10 @@ "UnregisteredSeiTimecode": { "shape": "H265UnregisteredSeiTimecode", "locationName": "unregisteredSeiTimecode" + }, + "WriteMp4PackagingType": { + "shape": "H265WriteMp4PackagingType", + "locationName": "writeMp4PackagingType" } } }, @@ -3130,58 +3403,69 @@ "ENABLED" ] }, + "H265WriteMp4PackagingType": { + "type": "string", + "enum": [ + "HVC1", + "HEV1" + ] + }, "Hdr10Metadata": { "type": "structure", "members": { "BluePrimaryX": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "bluePrimaryX" }, "BluePrimaryY": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "bluePrimaryY" }, "GreenPrimaryX": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "greenPrimaryX" }, "GreenPrimaryY": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "greenPrimaryY" }, "MaxContentLightLevel": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "maxContentLightLevel" }, "MaxFrameAverageLightLevel": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "maxFrameAverageLightLevel" }, "MaxLuminance": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "maxLuminance" }, "MinLuminance": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "minLuminance" }, "RedPrimaryX": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "redPrimaryX" }, "RedPrimaryY": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "redPrimaryY" }, "WhitePointX": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "whitePointX" }, "WhitePointY": { - "shape": "__integer", + "shape": "__integerMin0Max50000", "locationName": "whitePointY" } - } + }, + "required": [ + "MaxContentLightLevel", + "MaxFrameAverageLightLevel" + ] }, "HlsAdMarkers": { "type": "string", @@ -3203,9 +3487,13 @@ "type": "structure", "members": { "CaptionChannel": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "captionChannel" }, + "CustomLanguageCode": { + "shape": "__stringMin3Max3PatternAZaZ3", + "locationName": "customLanguageCode" + }, "LanguageCode": { "shape": "LanguageCode", "locationName": "languageCode" @@ -3249,7 +3537,7 @@ "type": "structure", "members": { "ConstantInitializationVector": { - "shape": "__string", + "shape": "__stringMin32Max32Pattern09aFAF32", "locationName": "constantInitializationVector" }, "EncryptionMethod": { @@ -3272,7 +3560,10 @@ "shape": "HlsKeyProviderType", "locationName": "type" } - } + }, + "required": [ + "Type" + ] }, "HlsEncryptionType": { "type": "string", @@ -3285,7 +3576,7 @@ "type": "structure", "members": { "AdMarkers": { - "shape": "ListOfHlsAdMarkers", + "shape": "__listOfHlsAdMarkers", "locationName": "adMarkers" }, "BaseUrl": { @@ -3293,7 +3584,7 @@ "locationName": "baseUrl" }, "CaptionLanguageMappings": { - "shape": "ListOfHlsCaptionLanguageMapping", + "shape": "__listOfHlsCaptionLanguageMapping", "locationName": "captionLanguageMappings" }, "CaptionLanguageSetting": { @@ -3309,7 +3600,7 @@ "locationName": "codecSpecification" }, "Destination": { - "shape": "__string", + "shape": "__stringPatternS3", "locationName": "destination" }, "DirectoryStructure": { @@ -3329,7 +3620,7 @@ "locationName": "manifestDurationFormat" }, "MinSegmentLength": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "minSegmentLength" }, "OutputSelection": { @@ -3341,7 +3632,7 @@ "locationName": "programDateTime" }, "ProgramDateTimePeriod": { - "shape": "__integer", + "shape": "__integerMin0Max3600", "locationName": "programDateTimePeriod" }, "SegmentControl": { @@ -3349,11 +3640,11 @@ "locationName": "segmentControl" }, "SegmentLength": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "segmentLength" }, "SegmentsPerSubdirectory": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "segmentsPerSubdirectory" }, "StreamInfResolution": { @@ -3365,14 +3656,18 @@ "locationName": "timedMetadataId3Frame" }, "TimedMetadataId3Period": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "timedMetadataId3Period" }, "TimestampDeltaMilliseconds": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "timestampDeltaMilliseconds" } - } + }, + "required": [ + "MinSegmentLength", + "SegmentLength" + ] }, "HlsIFrameOnlyManifest": { "type": "string", @@ -3474,37 +3769,44 @@ "type": "structure", "members": { "Id3": { - "shape": "__string", + "shape": "__stringPatternAZaZ0902", "locationName": "id3" }, "Timecode": { - "shape": "__string", + "shape": "__stringPattern010920405090509092", "locationName": "timecode" } - } + }, + "required": [ + "Timecode", + "Id3" + ] }, "ImageInserter": { "type": "structure", "members": { "InsertableImages": { - "shape": "ListOfInsertableImage", + "shape": "__listOfInsertableImage", "locationName": "insertableImages" } - } + }, + "required": [ + "InsertableImages" + ] }, "Input": { "type": "structure", "members": { "AudioSelectorGroups": { - "shape": "MapOfAudioSelectorGroup", + "shape": "__mapOfAudioSelectorGroup", "locationName": "audioSelectorGroups" }, "AudioSelectors": { - "shape": "MapOfAudioSelector", + "shape": "__mapOfAudioSelector", "locationName": "audioSelectors" }, "CaptionSelectors": { - "shape": "MapOfCaptionSelector", + "shape": "__mapOfCaptionSelector", "locationName": "captionSelectors" }, "DeblockFilter": { @@ -3516,7 +3818,7 @@ "locationName": "denoiseFilter" }, "FileInput": { - "shape": "__string", + "shape": "__stringPatternS3MM2VVMMPPEEGGAAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MM", "locationName": "fileInput" }, "FilterEnable": { @@ -3524,15 +3826,15 @@ "locationName": "filterEnable" }, "FilterStrength": { - "shape": "__integer", + "shape": "__integerMinNegative5Max5", "locationName": "filterStrength" }, "InputClippings": { - "shape": "ListOfInputClipping", + "shape": "__listOfInputClipping", "locationName": "inputClippings" }, "ProgramNumber": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "programNumber" }, "PsiControl": { @@ -3547,17 +3849,20 @@ "shape": "VideoSelector", "locationName": "videoSelector" } - } + }, + "required": [ + "FileInput" + ] }, "InputClipping": { "type": "structure", "members": { "EndTimecode": { - "shape": "__string", + "shape": "__stringPattern010920405090509092", "locationName": "endTimecode" }, "StartTimecode": { - "shape": "__string", + "shape": "__stringPattern010920405090509092", "locationName": "startTimecode" } } @@ -3595,15 +3900,15 @@ "type": "structure", "members": { "AudioSelectorGroups": { - "shape": "MapOfAudioSelectorGroup", + "shape": "__mapOfAudioSelectorGroup", "locationName": "audioSelectorGroups" }, "AudioSelectors": { - "shape": "MapOfAudioSelector", + "shape": "__mapOfAudioSelector", "locationName": "audioSelectors" }, "CaptionSelectors": { - "shape": "MapOfCaptionSelector", + "shape": "__mapOfCaptionSelector", "locationName": "captionSelectors" }, "DeblockFilter": { @@ -3619,15 +3924,15 @@ "locationName": "filterEnable" }, "FilterStrength": { - "shape": "__integer", + "shape": "__integerMinNegative5Max5", "locationName": "filterStrength" }, "InputClippings": { - "shape": "ListOfInputClipping", + "shape": "__listOfInputClipping", "locationName": "inputClippings" }, "ProgramNumber": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "programNumber" }, "PsiControl": { @@ -3656,50 +3961,57 @@ "type": "structure", "members": { "Duration": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "duration" }, "FadeIn": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "fadeIn" }, "FadeOut": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "fadeOut" }, "Height": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "height" }, "ImageInserterInput": { - "shape": "__string", + "shape": "__stringMin14PatternS3BmpBMPPngPNGTgaTGA", "locationName": "imageInserterInput" }, "ImageX": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "imageX" }, "ImageY": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "imageY" }, "Layer": { - "shape": "__integer", + "shape": "__integerMin0Max99", "locationName": "layer" }, "Opacity": { - "shape": "__integer", + "shape": "__integerMin0Max100", "locationName": "opacity" }, "StartTime": { - "shape": "__string", + "shape": "__stringPattern01D20305D205D", "locationName": "startTime" }, "Width": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "width" } - } + }, + "required": [ + "ImageY", + "ImageX", + "ImageInserterInput", + "Opacity", + "Layer" + ] }, "InternalServerErrorException": { "type": "structure", @@ -3722,7 +4034,7 @@ "locationName": "arn" }, "CreatedAt": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "createdAt" }, "ErrorCode": { @@ -3742,7 +4054,7 @@ "locationName": "jobTemplate" }, "OutputGroupDetails": { - "shape": "ListOfOutputGroupDetail", + "shape": "__listOfOutputGroupDetail", "locationName": "outputGroupDetails" }, "Queue": { @@ -3766,16 +4078,20 @@ "locationName": "timing" }, "UserMetadata": { - "shape": "MapOf__string", + "shape": "__mapOf__string", "locationName": "userMetadata" } - } + }, + "required": [ + "Role", + "Settings" + ] }, "JobSettings": { "type": "structure", "members": { "AdAvailOffset": { - "shape": "__integer", + "shape": "__integerMinNegative1000Max1000", "locationName": "adAvailOffset" }, "AvailBlanking": { @@ -3783,7 +4099,7 @@ "locationName": "availBlanking" }, "Inputs": { - "shape": "ListOfInput", + "shape": "__listOfInput", "locationName": "inputs" }, "NielsenConfiguration": { @@ -3791,7 +4107,7 @@ "locationName": "nielsenConfiguration" }, "OutputGroups": { - "shape": "ListOfOutputGroup", + "shape": "__listOfOutputGroup", "locationName": "outputGroups" }, "TimecodeConfig": { @@ -3802,7 +4118,11 @@ "shape": "TimedMetadataInsertion", "locationName": "timedMetadataInsertion" } - } + }, + "required": [ + "OutputGroups", + "Inputs" + ] }, "JobStatus": { "type": "string", @@ -3826,7 +4146,7 @@ "locationName": "category" }, "CreatedAt": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "createdAt" }, "Description": { @@ -3834,7 +4154,7 @@ "locationName": "description" }, "LastUpdated": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "lastUpdated" }, "Name": { @@ -3853,7 +4173,11 @@ "shape": "Type", "locationName": "type" } - } + }, + "required": [ + "Settings", + "Name" + ] }, "JobTemplateListBy": { "type": "string", @@ -3867,7 +4191,7 @@ "type": "structure", "members": { "AdAvailOffset": { - "shape": "__integer", + "shape": "__integerMinNegative1000Max1000", "locationName": "adAvailOffset" }, "AvailBlanking": { @@ -3875,7 +4199,7 @@ "locationName": "availBlanking" }, "Inputs": { - "shape": "ListOfInputTemplate", + "shape": "__listOfInputTemplate", "locationName": "inputs" }, "NielsenConfiguration": { @@ -3883,7 +4207,7 @@ "locationName": "nielsenConfiguration" }, "OutputGroups": { - "shape": "ListOfOutputGroup", + "shape": "__listOfOutputGroup", "locationName": "outputGroups" }, "TimecodeConfig": { @@ -3894,7 +4218,10 @@ "shape": "TimedMetadataInsertion", "locationName": "timedMetadataInsertion" } - } + }, + "required": [ + "OutputGroups" + ] }, "LanguageCode": { "type": "string", @@ -4126,7 +4453,7 @@ "type": "structure", "members": { "JobTemplates": { - "shape": "ListOfJobTemplate", + "shape": "__listOfJobTemplate", "locationName": "jobTemplates" }, "NextToken": { @@ -4169,7 +4496,7 @@ "type": "structure", "members": { "Jobs": { - "shape": "ListOfJob", + "shape": "__listOfJob", "locationName": "jobs" }, "NextToken": { @@ -4178,138 +4505,6 @@ } } }, - "ListOfAudioDescription": { - "type": "list", - "member": { - "shape": "AudioDescription" - } - }, - "ListOfCaptionDescription": { - "type": "list", - "member": { - "shape": "CaptionDescription" - } - }, - "ListOfCaptionDescriptionPreset": { - "type": "list", - "member": { - "shape": "CaptionDescriptionPreset" - } - }, - "ListOfEndpoint": { - "type": "list", - "member": { - "shape": "Endpoint" - } - }, - "ListOfHlsAdMarkers": { - "type": "list", - "member": { - "shape": "HlsAdMarkers" - } - }, - "ListOfHlsCaptionLanguageMapping": { - "type": "list", - "member": { - "shape": "HlsCaptionLanguageMapping" - } - }, - "ListOfId3Insertion": { - "type": "list", - "member": { - "shape": "Id3Insertion" - } - }, - "ListOfInput": { - "type": "list", - "member": { - "shape": "Input" - } - }, - "ListOfInputClipping": { - "type": "list", - "member": { - "shape": "InputClipping" - } - }, - "ListOfInputTemplate": { - "type": "list", - "member": { - "shape": "InputTemplate" - } - }, - "ListOfInsertableImage": { - "type": "list", - "member": { - "shape": "InsertableImage" - } - }, - "ListOfJob": { - "type": "list", - "member": { - "shape": "Job" - } - }, - "ListOfJobTemplate": { - "type": "list", - "member": { - "shape": "JobTemplate" - } - }, - "ListOfOutput": { - "type": "list", - "member": { - "shape": "Output" - } - }, - "ListOfOutputChannelMapping": { - "type": "list", - "member": { - "shape": "OutputChannelMapping" - } - }, - "ListOfOutputDetail": { - "type": "list", - "member": { - "shape": "OutputDetail" - } - }, - "ListOfOutputGroup": { - "type": "list", - "member": { - "shape": "OutputGroup" - } - }, - "ListOfOutputGroupDetail": { - "type": "list", - "member": { - "shape": "OutputGroupDetail" - } - }, - "ListOfPreset": { - "type": "list", - "member": { - "shape": "Preset" - } - }, - "ListOfQueue": { - "type": "list", - "member": { - "shape": "Queue" - } - }, - "ListOf__integer": { - "type": "list", - "member": { - "shape": "__integer" - } - }, - "ListOf__string": { - "type": "list", - "member": { - "shape": "__string" - } - }, "ListPresetsRequest": { "type": "structure", "members": { @@ -4348,7 +4543,7 @@ "locationName": "nextToken" }, "Presets": { - "shape": "ListOfPreset", + "shape": "__listOfPreset", "locationName": "presets" } } @@ -4386,7 +4581,7 @@ "locationName": "nextToken" }, "Queues": { - "shape": "ListOfQueue", + "shape": "__listOfQueue", "locationName": "queues" } } @@ -4480,15 +4675,15 @@ "locationName": "audioBufferModel" }, "AudioFramesPerPes": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "audioFramesPerPes" }, "AudioPids": { - "shape": "ListOf__integer", + "shape": "__listOf__integerMin32Max8182", "locationName": "audioPids" }, "Bitrate": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "bitrate" }, "BufferModel": { @@ -4504,7 +4699,7 @@ "locationName": "dvbSdtSettings" }, "DvbSubPids": { - "shape": "ListOf__integer", + "shape": "__listOf__integerMin32Max8182", "locationName": "dvbSubPids" }, "DvbTdtSettings": { @@ -4512,7 +4707,7 @@ "locationName": "dvbTdtSettings" }, "DvbTeletextPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "dvbTeletextPid" }, "EbpAudioInterval": { @@ -4528,15 +4723,15 @@ "locationName": "esRateInPes" }, "FragmentTime": { - "shape": "__double", + "shape": "__doubleMin0", "locationName": "fragmentTime" }, "MaxPcrInterval": { - "shape": "__integer", + "shape": "__integerMin0Max500", "locationName": "maxPcrInterval" }, "MinEbpInterval": { - "shape": "__integer", + "shape": "__integerMin0Max10000", "locationName": "minEbpInterval" }, "NielsenId3": { @@ -4544,11 +4739,11 @@ "locationName": "nielsenId3" }, "NullPacketBitrate": { - "shape": "__double", + "shape": "__doubleMin0", "locationName": "nullPacketBitrate" }, "PatInterval": { - "shape": "__integer", + "shape": "__integerMin0Max1000", "locationName": "patInterval" }, "PcrControl": { @@ -4556,23 +4751,23 @@ "locationName": "pcrControl" }, "PcrPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "pcrPid" }, "PmtInterval": { - "shape": "__integer", + "shape": "__integerMin0Max1000", "locationName": "pmtInterval" }, "PmtPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "pmtPid" }, "PrivateMetadataPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "privateMetadataPid" }, "ProgramNumber": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "programNumber" }, "RateMode": { @@ -4580,7 +4775,7 @@ "locationName": "rateMode" }, "Scte35Pid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "scte35Pid" }, "Scte35Source": { @@ -4596,19 +4791,19 @@ "locationName": "segmentationStyle" }, "SegmentationTime": { - "shape": "__double", + "shape": "__doubleMin0", "locationName": "segmentationTime" }, "TimedMetadataPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "timedMetadataPid" }, "TransportStreamId": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "transportStreamId" }, "VideoPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "videoPid" } } @@ -4638,11 +4833,11 @@ "type": "structure", "members": { "AudioFramesPerPes": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "audioFramesPerPes" }, "AudioPids": { - "shape": "ListOf__integer", + "shape": "__listOf__integerMin32Max8182", "locationName": "audioPids" }, "NielsenId3": { @@ -4650,7 +4845,7 @@ "locationName": "nielsenId3" }, "PatInterval": { - "shape": "__integer", + "shape": "__integerMin0Max1000", "locationName": "patInterval" }, "PcrControl": { @@ -4658,27 +4853,27 @@ "locationName": "pcrControl" }, "PcrPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "pcrPid" }, "PmtInterval": { - "shape": "__integer", + "shape": "__integerMin0Max1000", "locationName": "pmtInterval" }, "PmtPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "pmtPid" }, "PrivateMetadataPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "privateMetadataPid" }, "ProgramNumber": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "programNumber" }, "Scte35Pid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "scte35Pid" }, "Scte35Source": { @@ -4690,55 +4885,19 @@ "locationName": "timedMetadata" }, "TimedMetadataPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "timedMetadataPid" }, "TransportStreamId": { - "shape": "__integer", + "shape": "__integerMin0Max65535", "locationName": "transportStreamId" }, "VideoPid": { - "shape": "__integer", + "shape": "__integerMin32Max8182", "locationName": "videoPid" } } }, - "MapOfAudioSelector": { - "type": "map", - "key": { - "shape": "__string" - }, - "value": { - "shape": "AudioSelector" - } - }, - "MapOfAudioSelectorGroup": { - "type": "map", - "key": { - "shape": "__string" - }, - "value": { - "shape": "AudioSelectorGroup" - } - }, - "MapOfCaptionSelector": { - "type": "map", - "key": { - "shape": "__string" - }, - "value": { - "shape": "CaptionSelector" - } - }, - "MapOf__string": { - "type": "map", - "key": { - "shape": "__string" - }, - "value": { - "shape": "__string" - } - }, "MovClapAtom": { "type": "string", "enum": [ @@ -4803,15 +4962,15 @@ "type": "structure", "members": { "Bitrate": { - "shape": "__integer", + "shape": "__integerMin32000Max384000", "locationName": "bitrate" }, "Channels": { - "shape": "__integer", + "shape": "__integerMin1Max2", "locationName": "channels" }, "SampleRate": { - "shape": "__integer", + "shape": "__integerMin32000Max48000", "locationName": "sampleRate" } } @@ -4961,7 +5120,7 @@ "locationName": "adaptiveQuantization" }, "Bitrate": { - "shape": "__integer", + "shape": "__integerMin1000Max288000000", "locationName": "bitrate" }, "CodecLevel": { @@ -4981,19 +5140,19 @@ "locationName": "framerateConversionAlgorithm" }, "FramerateDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max1001", "locationName": "framerateDenominator" }, "FramerateNumerator": { - "shape": "__integer", + "shape": "__integerMin24Max60000", "locationName": "framerateNumerator" }, "GopClosedCadence": { - "shape": "__integer", + "shape": "__integerMin0Max2147483647", "locationName": "gopClosedCadence" }, "GopSize": { - "shape": "__double", + "shape": "__doubleMin0", "locationName": "gopSize" }, "GopSizeUnits": { @@ -5001,11 +5160,11 @@ "locationName": "gopSizeUnits" }, "HrdBufferInitialFillPercentage": { - "shape": "__integer", + "shape": "__integerMin0Max100", "locationName": "hrdBufferInitialFillPercentage" }, "HrdBufferSize": { - "shape": "__integer", + "shape": "__integerMin0Max47185920", "locationName": "hrdBufferSize" }, "InterlaceMode": { @@ -5017,15 +5176,15 @@ "locationName": "intraDcPrecision" }, "MaxBitrate": { - "shape": "__integer", + "shape": "__integerMin1000Max300000000", "locationName": "maxBitrate" }, "MinIInterval": { - "shape": "__integer", + "shape": "__integerMin0Max30", "locationName": "minIInterval" }, "NumberBFramesBetweenReferenceFrames": { - "shape": "__integer", + "shape": "__integerMin0Max7", "locationName": "numberBFramesBetweenReferenceFrames" }, "ParControl": { @@ -5033,11 +5192,11 @@ "locationName": "parControl" }, "ParDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parDenominator" }, "ParNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parNumerator" }, "QualityTuningLevel": { @@ -5057,7 +5216,7 @@ "locationName": "slowPal" }, "Softness": { - "shape": "__integer", + "shape": "__integerMin0Max128", "locationName": "softness" }, "SpatialAdaptiveQuantization": { @@ -5128,7 +5287,10 @@ "shape": "SpekeKeyProvider", "locationName": "spekeKeyProvider" } - } + }, + "required": [ + "SpekeKeyProvider" + ] }, "MsSmoothGroupSettings": { "type": "structure", @@ -5138,7 +5300,7 @@ "locationName": "audioDeduplication" }, "Destination": { - "shape": "__string", + "shape": "__stringPatternS3", "locationName": "destination" }, "Encryption": { @@ -5146,14 +5308,17 @@ "locationName": "encryption" }, "FragmentLength": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "fragmentLength" }, "ManifestEncoding": { "shape": "MsSmoothManifestEncoding", "locationName": "manifestEncoding" } - } + }, + "required": [ + "FragmentLength" + ] }, "MsSmoothManifestEncoding": { "type": "string", @@ -5166,7 +5331,7 @@ "type": "structure", "members": { "BreakoutCode": { - "shape": "__integer", + "shape": "__integerMin0Max9", "locationName": "breakoutCode" }, "DistributorId": { @@ -5190,7 +5355,10 @@ "shape": "NoiseReducerSpatialFilterSettings", "locationName": "spatialFilterSettings" } - } + }, + "required": [ + "Filter" + ] }, "NoiseReducerFilter": { "type": "string", @@ -5208,7 +5376,7 @@ "type": "structure", "members": { "Strength": { - "shape": "__integer", + "shape": "__integerMin0Max3", "locationName": "strength" } } @@ -5217,15 +5385,15 @@ "type": "structure", "members": { "PostFilterSharpenStrength": { - "shape": "__integer", + "shape": "__integerMin0Max3", "locationName": "postFilterSharpenStrength" }, "Speed": { - "shape": "__integer", + "shape": "__integerMinNegative2Max3", "locationName": "speed" }, "Strength": { - "shape": "__integer", + "shape": "__integerMin0Max16", "locationName": "strength" } } @@ -5254,11 +5422,11 @@ "type": "structure", "members": { "AudioDescriptions": { - "shape": "ListOfAudioDescription", + "shape": "__listOfAudioDescription", "locationName": "audioDescriptions" }, "CaptionDescriptions": { - "shape": "ListOfCaptionDescription", + "shape": "__listOfCaptionDescription", "locationName": "captionDescriptions" }, "ContainerSettings": { @@ -5270,7 +5438,7 @@ "locationName": "extension" }, "NameModifier": { - "shape": "__string", + "shape": "__stringMin1", "locationName": "nameModifier" }, "OutputSettings": { @@ -5278,7 +5446,7 @@ "locationName": "outputSettings" }, "Preset": { - "shape": "__string", + "shape": "__stringMin0", "locationName": "preset" }, "VideoDescription": { @@ -5291,10 +5459,13 @@ "type": "structure", "members": { "InputChannels": { - "shape": "ListOf__integer", + "shape": "__listOf__integerMinNegative60Max6", "locationName": "inputChannels" } - } + }, + "required": [ + "InputChannels" + ] }, "OutputDetail": { "type": "structure", @@ -5325,16 +5496,20 @@ "locationName": "outputGroupSettings" }, "Outputs": { - "shape": "ListOfOutput", + "shape": "__listOfOutput", "locationName": "outputs" } - } + }, + "required": [ + "Outputs", + "OutputGroupSettings" + ] }, "OutputGroupDetail": { "type": "structure", "members": { "OutputDetails": { - "shape": "ListOfOutputDetail", + "shape": "__listOfOutputDetail", "locationName": "outputDetails" } } @@ -5342,6 +5517,10 @@ "OutputGroupSettings": { "type": "structure", "members": { + "CmafGroupSettings": { + "shape": "CmafGroupSettings", + "locationName": "cmafGroupSettings" + }, "DashIsoGroupSettings": { "shape": "DashIsoGroupSettings", "locationName": "dashIsoGroupSettings" @@ -5362,7 +5541,10 @@ "shape": "OutputGroupType", "locationName": "type" } - } + }, + "required": [ + "Type" + ] }, "OutputGroupType": { "type": "string", @@ -5370,7 +5552,8 @@ "HLS_GROUP_SETTINGS", "DASH_ISO_GROUP_SETTINGS", "FILE_GROUP_SETTINGS", - "MS_SMOOTH_GROUP_SETTINGS" + "MS_SMOOTH_GROUP_SETTINGS", + "CMAF_GROUP_SETTINGS" ] }, "OutputSdt": { @@ -5403,7 +5586,7 @@ "locationName": "category" }, "CreatedAt": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "createdAt" }, "Description": { @@ -5411,7 +5594,7 @@ "locationName": "description" }, "LastUpdated": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "lastUpdated" }, "Name": { @@ -5426,7 +5609,11 @@ "shape": "Type", "locationName": "type" } - } + }, + "required": [ + "Settings", + "Name" + ] }, "PresetListBy": { "type": "string", @@ -5440,11 +5627,11 @@ "type": "structure", "members": { "AudioDescriptions": { - "shape": "ListOfAudioDescription", + "shape": "__listOfAudioDescription", "locationName": "audioDescriptions" }, "CaptionDescriptions": { - "shape": "ListOfCaptionDescriptionPreset", + "shape": "__listOfCaptionDescriptionPreset", "locationName": "captionDescriptions" }, "ContainerSettings": { @@ -5513,11 +5700,11 @@ "locationName": "framerateConversionAlgorithm" }, "FramerateDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateDenominator" }, "FramerateNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "framerateNumerator" }, "InterlaceMode": { @@ -5529,11 +5716,11 @@ "locationName": "parControl" }, "ParDenominator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parDenominator" }, "ParNumerator": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "parNumerator" }, "SlowPal": { @@ -5568,7 +5755,7 @@ "locationName": "arn" }, "CreatedAt": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "createdAt" }, "Description": { @@ -5576,7 +5763,7 @@ "locationName": "description" }, "LastUpdated": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "lastUpdated" }, "Name": { @@ -5591,7 +5778,10 @@ "shape": "Type", "locationName": "type" } - } + }, + "required": [ + "Name" + ] }, "QueueListBy": { "type": "string", @@ -5611,22 +5801,28 @@ "type": "structure", "members": { "Height": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "height" }, "Width": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "width" }, "X": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "x" }, "Y": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "y" } - } + }, + "required": [ + "X", + "Y", + "Height", + "Width" + ] }, "RemixSettings": { "type": "structure", @@ -5636,14 +5832,19 @@ "locationName": "channelMapping" }, "ChannelsIn": { - "shape": "__integer", + "shape": "__integerMin1Max16", "locationName": "channelsIn" }, "ChannelsOut": { - "shape": "__integer", + "shape": "__integerMin1Max8", "locationName": "channelsOut" } - } + }, + "required": [ + "ChannelsOut", + "ChannelMapping", + "ChannelsIn" + ] }, "RespondToAfd": { "type": "string", @@ -5686,41 +5887,50 @@ "locationName": "resourceId" }, "SystemIds": { - "shape": "ListOf__string", + "shape": "__listOf__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12", "locationName": "systemIds" }, "Url": { - "shape": "__string", + "shape": "__stringPatternHttps", "locationName": "url" } - } + }, + "required": [ + "ResourceId", + "SystemIds", + "Url" + ] }, "StaticKeyProvider": { "type": "structure", "members": { "KeyFormat": { - "shape": "__string", + "shape": "__stringPatternIdentityAZaZ26AZaZ09163", "locationName": "keyFormat" }, "KeyFormatVersions": { - "shape": "__string", + "shape": "__stringPatternDD", "locationName": "keyFormatVersions" }, "StaticKeyValue": { - "shape": "__string", + "shape": "__stringPatternAZaZ0932", "locationName": "staticKeyValue" }, "Url": { "shape": "__string", "locationName": "url" } - } + }, + "required": [ + "Url", + "StaticKeyValue" + ] }, "TeletextDestinationSettings": { "type": "structure", "members": { "PageNumber": { - "shape": "__string", + "shape": "__stringMin3Max3Pattern1809aFAF09aEAE", "locationName": "pageNumber" } } @@ -5729,7 +5939,7 @@ "type": "structure", "members": { "PageNumber": { - "shape": "__string", + "shape": "__stringMin3Max3Pattern1809aFAF09aEAE", "locationName": "pageNumber" } } @@ -5738,7 +5948,7 @@ "type": "structure", "members": { "FontSize": { - "shape": "__integer", + "shape": "__integerMin10Max48", "locationName": "fontSize" }, "Position": { @@ -5746,7 +5956,7 @@ "locationName": "position" }, "Prefix": { - "shape": "__string", + "shape": "__stringPattern", "locationName": "prefix" } } @@ -5769,7 +5979,7 @@ "type": "structure", "members": { "Anchor": { - "shape": "__string", + "shape": "__stringPattern010920405090509092", "locationName": "anchor" }, "Source": { @@ -5777,11 +5987,11 @@ "locationName": "source" }, "Start": { - "shape": "__string", + "shape": "__stringPattern010920405090509092", "locationName": "start" }, "TimestampOffset": { - "shape": "__string", + "shape": "__stringPattern0940191020191209301", "locationName": "timestampOffset" } } @@ -5805,24 +6015,27 @@ "type": "structure", "members": { "Id3Insertions": { - "shape": "ListOfId3Insertion", + "shape": "__listOfId3Insertion", "locationName": "id3Insertions" } - } + }, + "required": [ + "Id3Insertions" + ] }, "Timing": { "type": "structure", "members": { "FinishTime": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "finishTime" }, "StartTime": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "startTime" }, "SubmitTime": { - "shape": "__timestamp", + "shape": "__timestampIso8601", "locationName": "submitTime" } } @@ -6002,7 +6215,10 @@ "shape": "ProresSettings", "locationName": "proresSettings" } - } + }, + "required": [ + "Codec" + ] }, "VideoDescription": { "type": "structure", @@ -6032,11 +6248,11 @@ "locationName": "dropFrameTimecode" }, "FixedAfd": { - "shape": "__integer", + "shape": "__integerMin0Max15", "locationName": "fixedAfd" }, "Height": { - "shape": "__integer", + "shape": "__integerMin32Max2160", "locationName": "height" }, "Position": { @@ -6052,7 +6268,7 @@ "locationName": "scalingBehavior" }, "Sharpness": { - "shape": "__integer", + "shape": "__integerMin0Max100", "locationName": "sharpness" }, "TimecodeInsertion": { @@ -6064,10 +6280,13 @@ "locationName": "videoPreprocessors" }, "Width": { - "shape": "__integer", + "shape": "__integerMin32Max4096", "locationName": "width" } - } + }, + "required": [ + "CodecSettings" + ] }, "VideoDetail": { "type": "structure", @@ -6123,11 +6342,11 @@ "locationName": "hdr10Metadata" }, "Pid": { - "shape": "__integer", + "shape": "__integerMin1Max2147483647", "locationName": "pid" }, "ProgramNumber": { - "shape": "__integer", + "shape": "__integerMinNegative2147483648Max2147483647", "locationName": "programNumber" } } @@ -6139,19 +6358,30 @@ "PIC_TIMING_SEI" ] }, + "WavFormat": { + "type": "string", + "enum": [ + "RIFF", + "RF64" + ] + }, "WavSettings": { "type": "structure", "members": { "BitDepth": { - "shape": "__integer", + "shape": "__integerMin16Max24", "locationName": "bitDepth" }, "Channels": { - "shape": "__integer", + "shape": "__integerMin1Max8", "locationName": "channels" }, + "Format": { + "shape": "WavFormat", + "locationName": "format" + }, "SampleRate": { - "shape": "__integer", + "shape": "__integerMin8000Max192000", "locationName": "sampleRate" } } @@ -6162,14 +6392,637 @@ "__double": { "type": "double" }, + "__doubleMin0": { + "type": "double" + }, + "__doubleMinNegative59Max0": { + "type": "double" + }, + "__doubleMinNegative60Max3": { + "type": "double" + }, + "__doubleMinNegative60MaxNegative1": { + "type": "double" + }, "__integer": { "type": "integer" }, + "__integerMin0Max10": { + "type": "integer", + "min": 0, + "max": 10 + }, + "__integerMin0Max100": { + "type": "integer", + "min": 0, + "max": 100 + }, + "__integerMin0Max1000": { + "type": "integer", + "min": 0, + "max": 1000 + }, + "__integerMin0Max10000": { + "type": "integer", + "min": 0, + "max": 10000 + }, + "__integerMin0Max1152000000": { + "type": "integer", + "min": 0, + "max": 1152000000 + }, + "__integerMin0Max128": { + "type": "integer", + "min": 0, + "max": 128 + }, + "__integerMin0Max1466400000": { + "type": "integer", + "min": 0, + "max": 1466400000 + }, + "__integerMin0Max15": { + "type": "integer", + "min": 0, + "max": 15 + }, + "__integerMin0Max16": { + "type": "integer", + "min": 0, + "max": 16 + }, + "__integerMin0Max2147483647": { + "type": "integer", + "min": 0, + "max": 2147483647 + }, + "__integerMin0Max255": { + "type": "integer", + "min": 0, + "max": 255 + }, + "__integerMin0Max3": { + "type": "integer", + "min": 0, + "max": 3 + }, + "__integerMin0Max30": { + "type": "integer", + "min": 0, + "max": 30 + }, + "__integerMin0Max3600": { + "type": "integer", + "min": 0, + "max": 3600 + }, + "__integerMin0Max47185920": { + "type": "integer", + "min": 0, + "max": 47185920 + }, + "__integerMin0Max500": { + "type": "integer", + "min": 0, + "max": 500 + }, + "__integerMin0Max50000": { + "type": "integer", + "min": 0, + "max": 50000 + }, + "__integerMin0Max65535": { + "type": "integer", + "min": 0, + "max": 65535 + }, + "__integerMin0Max7": { + "type": "integer", + "min": 0, + "max": 7 + }, + "__integerMin0Max8": { + "type": "integer", + "min": 0, + "max": 8 + }, + "__integerMin0Max9": { + "type": "integer", + "min": 0, + "max": 9 + }, + "__integerMin0Max96": { + "type": "integer", + "min": 0, + "max": 96 + }, + "__integerMin0Max99": { + "type": "integer", + "min": 0, + "max": 99 + }, + "__integerMin1000Max1152000000": { + "type": "integer", + "min": 1000, + "max": 1152000000 + }, + "__integerMin1000Max1466400000": { + "type": "integer", + "min": 1000, + "max": 1466400000 + }, + "__integerMin1000Max288000000": { + "type": "integer", + "min": 1000, + "max": 288000000 + }, + "__integerMin1000Max30000": { + "type": "integer", + "min": 1000, + "max": 30000 + }, + "__integerMin1000Max300000000": { + "type": "integer", + "min": 1000, + "max": 300000000 + }, + "__integerMin10Max48": { + "type": "integer", + "min": 10, + "max": 48 + }, + "__integerMin16Max24": { + "type": "integer", + "min": 16, + "max": 24 + }, + "__integerMin1Max1": { + "type": "integer", + "min": 1, + "max": 1 + }, + "__integerMin1Max100": { + "type": "integer", + "min": 1, + "max": 100 + }, + "__integerMin1Max10000000": { + "type": "integer", + "min": 1, + "max": 10000000 + }, + "__integerMin1Max1001": { + "type": "integer", + "min": 1, + "max": 1001 + }, + "__integerMin1Max16": { + "type": "integer", + "min": 1, + "max": 16 + }, + "__integerMin1Max2": { + "type": "integer", + "min": 1, + "max": 2 + }, + "__integerMin1Max2147483647": { + "type": "integer", + "min": 1, + "max": 2147483647 + }, + "__integerMin1Max31": { + "type": "integer", + "min": 1, + "max": 31 + }, + "__integerMin1Max32": { + "type": "integer", + "min": 1, + "max": 32 + }, + "__integerMin1Max4": { + "type": "integer", + "min": 1, + "max": 4 + }, + "__integerMin1Max6": { + "type": "integer", + "min": 1, + "max": 6 + }, + "__integerMin1Max8": { + "type": "integer", + "min": 1, + "max": 8 + }, + "__integerMin24Max60000": { + "type": "integer", + "min": 24, + "max": 60000 + }, + "__integerMin25Max10000": { + "type": "integer", + "min": 25, + "max": 10000 + }, + "__integerMin25Max2000": { + "type": "integer", + "min": 25, + "max": 2000 + }, + "__integerMin32000Max384000": { + "type": "integer", + "min": 32000, + "max": 384000 + }, + "__integerMin32000Max48000": { + "type": "integer", + "min": 32000, + "max": 48000 + }, + "__integerMin32Max2160": { + "type": "integer", + "min": 32, + "max": 2160 + }, + "__integerMin32Max4096": { + "type": "integer", + "min": 32, + "max": 4096 + }, + "__integerMin32Max8182": { + "type": "integer", + "min": 32, + "max": 8182 + }, + "__integerMin48000Max48000": { + "type": "integer", + "min": 48000, + "max": 48000 + }, + "__integerMin6000Max1024000": { + "type": "integer", + "min": 6000, + "max": 1024000 + }, + "__integerMin64000Max640000": { + "type": "integer", + "min": 64000, + "max": 640000 + }, + "__integerMin8000Max192000": { + "type": "integer", + "min": 8000, + "max": 192000 + }, + "__integerMin8000Max96000": { + "type": "integer", + "min": 8000, + "max": 96000 + }, + "__integerMin96Max600": { + "type": "integer", + "min": 96, + "max": 600 + }, + "__integerMinNegative1000Max1000": { + "type": "integer", + "min": -1000, + "max": 1000 + }, + "__integerMinNegative180Max180": { + "type": "integer", + "min": -180, + "max": 180 + }, + "__integerMinNegative2147483648Max2147483647": { + "type": "integer", + "min": -2147483648, + "max": 2147483647 + }, + "__integerMinNegative2Max3": { + "type": "integer", + "min": -2, + "max": 3 + }, + "__integerMinNegative5Max5": { + "type": "integer", + "min": -5, + "max": 5 + }, + "__integerMinNegative60Max6": { + "type": "integer", + "min": -60, + "max": 6 + }, + "__integerMinNegative70Max0": { + "type": "integer", + "min": -70, + "max": 0 + }, + "__listOfAudioDescription": { + "type": "list", + "member": { + "shape": "AudioDescription" + } + }, + "__listOfCaptionDescription": { + "type": "list", + "member": { + "shape": "CaptionDescription" + } + }, + "__listOfCaptionDescriptionPreset": { + "type": "list", + "member": { + "shape": "CaptionDescriptionPreset" + } + }, + "__listOfEndpoint": { + "type": "list", + "member": { + "shape": "Endpoint" + } + }, + "__listOfHlsAdMarkers": { + "type": "list", + "member": { + "shape": "HlsAdMarkers" + } + }, + "__listOfHlsCaptionLanguageMapping": { + "type": "list", + "member": { + "shape": "HlsCaptionLanguageMapping" + } + }, + "__listOfId3Insertion": { + "type": "list", + "member": { + "shape": "Id3Insertion" + } + }, + "__listOfInput": { + "type": "list", + "member": { + "shape": "Input" + } + }, + "__listOfInputClipping": { + "type": "list", + "member": { + "shape": "InputClipping" + } + }, + "__listOfInputTemplate": { + "type": "list", + "member": { + "shape": "InputTemplate" + } + }, + "__listOfInsertableImage": { + "type": "list", + "member": { + "shape": "InsertableImage" + } + }, + "__listOfJob": { + "type": "list", + "member": { + "shape": "Job" + } + }, + "__listOfJobTemplate": { + "type": "list", + "member": { + "shape": "JobTemplate" + } + }, + "__listOfOutput": { + "type": "list", + "member": { + "shape": "Output" + } + }, + "__listOfOutputChannelMapping": { + "type": "list", + "member": { + "shape": "OutputChannelMapping" + } + }, + "__listOfOutputDetail": { + "type": "list", + "member": { + "shape": "OutputDetail" + } + }, + "__listOfOutputGroup": { + "type": "list", + "member": { + "shape": "OutputGroup" + } + }, + "__listOfOutputGroupDetail": { + "type": "list", + "member": { + "shape": "OutputGroupDetail" + } + }, + "__listOfPreset": { + "type": "list", + "member": { + "shape": "Preset" + } + }, + "__listOfQueue": { + "type": "list", + "member": { + "shape": "Queue" + } + }, + "__listOf__integerMin1Max2147483647": { + "type": "list", + "member": { + "shape": "__integerMin1Max2147483647" + } + }, + "__listOf__integerMin32Max8182": { + "type": "list", + "member": { + "shape": "__integerMin32Max8182" + } + }, + "__listOf__integerMinNegative60Max6": { + "type": "list", + "member": { + "shape": "__integerMinNegative60Max6" + } + }, + "__listOf__stringMin1": { + "type": "list", + "member": { + "shape": "__stringMin1" + } + }, + "__listOf__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12": { + "type": "list", + "member": { + "shape": "__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12" + } + }, + "__long": { + "type": "long" + }, + "__mapOfAudioSelector": { + "type": "map", + "key": { + "shape": "__string" + }, + "value": { + "shape": "AudioSelector" + } + }, + "__mapOfAudioSelectorGroup": { + "type": "map", + "key": { + "shape": "__string" + }, + "value": { + "shape": "AudioSelectorGroup" + } + }, + "__mapOfCaptionSelector": { + "type": "map", + "key": { + "shape": "__string" + }, + "value": { + "shape": "CaptionSelector" + } + }, + "__mapOf__string": { + "type": "map", + "key": { + "shape": "__string" + }, + "value": { + "shape": "__string" + } + }, "__string": { "type": "string" }, - "__timestamp": { - "type": "timestamp" + "__stringMin0": { + "type": "string", + "min": 0 + }, + "__stringMin1": { + "type": "string", + "min": 1 + }, + "__stringMin14PatternS3BmpBMPPngPNG": { + "type": "string", + "min": 14, + "pattern": "^(s3:\\/\\/)(.*?)\\.(bmp|BMP|png|PNG)$" + }, + "__stringMin14PatternS3BmpBMPPngPNGTgaTGA": { + "type": "string", + "min": 14, + "pattern": "^(s3:\\/\\/)(.*?)\\.(bmp|BMP|png|PNG|tga|TGA)$" + }, + "__stringMin14PatternS3SccSCCTtmlTTMLDfxpDFXPStlSTLSrtSRTSmiSMI": { + "type": "string", + "min": 14, + "pattern": "^(s3:\\/\\/)(.*?)\\.(scc|SCC|ttml|TTML|dfxp|DFXP|stl|STL|srt|SRT|smi|SMI)$" + }, + "__stringMin1Max256": { + "type": "string", + "min": 1, + "max": 256 + }, + "__stringMin32Max32Pattern09aFAF32": { + "type": "string", + "min": 32, + "max": 32, + "pattern": "^[0-9a-fA-F]{32}$" + }, + "__stringMin3Max3Pattern1809aFAF09aEAE": { + "type": "string", + "min": 3, + "max": 3, + "pattern": "^[1-8][0-9a-fA-F][0-9a-eA-E]$" + }, + "__stringMin3Max3PatternAZaZ3": { + "type": "string", + "min": 3, + "max": 3, + "pattern": "^[A-Za-z]{3}$" + }, + "__stringPattern": { + "type": "string", + "pattern": "^[ -~]+$" + }, + "__stringPattern010920405090509092": { + "type": "string", + "pattern": "^([01][0-9]|2[0-4]):[0-5][0-9]:[0-5][0-9][:;][0-9]{2}$" + }, + "__stringPattern01D20305D205D": { + "type": "string", + "pattern": "^((([0-1]\\d)|(2[0-3]))(:[0-5]\\d){2}([:;][0-5]\\d))$" + }, + "__stringPattern0940191020191209301": { + "type": "string", + "pattern": "^([0-9]{4})-(0[1-9]|1[0-2])-(0[1-9]|[12][0-9]|3[01])$" + }, + "__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12": { + "type": "string", + "pattern": "^[0-9a-fA-F]{8}-[0-9a-fA-F]{4}-[0-9a-fA-F]{4}-[0-9a-fA-F]{4}-[0-9a-fA-F]{12}$" + }, + "__stringPatternAZaZ0902": { + "type": "string", + "pattern": "^[A-Za-z0-9+\\/]+={0,2}$" + }, + "__stringPatternAZaZ0932": { + "type": "string", + "pattern": "^[A-Za-z0-9]{32}$" + }, + "__stringPatternDD": { + "type": "string", + "pattern": "^(\\d+(\\/\\d+)*)$" + }, + "__stringPatternHttps": { + "type": "string", + "pattern": "^https:\\/\\/" + }, + "__stringPatternIdentityAZaZ26AZaZ09163": { + "type": "string", + "pattern": "^(identity|[A-Za-z]{2,6}(\\.[A-Za-z0-9-]{1,63})+)$" + }, + "__stringPatternS3": { + "type": "string", + "pattern": "^s3:\\/\\/" + }, + "__stringPatternS3MM2VVMMPPEEGGAAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MM": { + "type": "string", + "pattern": "^(s3:\\/\\/)([^\\/]+\\/)+([^\\/\\.]+|(([^\\/]*)\\.([mM]2[vV]|[mM][pP][eE][gG]|[aA][vV][iI]|[mM][pP]4|[fF][lL][vV]|[mM][pP][tT]|[mM][pP][gG]|[mM]4[vV]|[tT][rR][pP]|[fF]4[vV]|[mM]2[tT][sS]|[tT][sS]|264|[hH]264|[mM][kK][vV]|[mM][oO][vV]|[mM][tT][sS]|[mM]2[tT]|[wW][mM][vV]|[aA][sS][fF]|[vV][oO][bB]|3[gG][pP]|3[gG][pP][pP]|[mM][xX][fF]|[dD][iI][vV][xX]|[xX][vV][iI][dD]|[rR][aA][wW]|[dD][vV]|[gG][xX][fF]|[mM]1[vV]|3[gG]2|[vV][mM][fF]|[mM]3[uU]8|[lL][cC][hH]|[gG][xX][fF]_[mM][pP][eE][gG]2|[mM][xX][fF]_[mM][pP][eE][gG]2|[mM][xX][fF][hH][dD]|[wW][aA][vV]|[yY]4[mM])))$" + }, + "__stringPatternS3MM2VVMMPPEEGGAAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MMAAAACCAAIIFFFFMMPP2AACC3EECC3DDTTSSEE": { + "type": "string", + "pattern": "^(s3:\\/\\/)([^\\/]+\\/)+([^\\/\\.]+|(([^\\/]*)\\.([mM]2[vV]|[mM][pP][eE][gG]|[aA][vV][iI]|[mM][pP]4|[fF][lL][vV]|[mM][pP][tT]|[mM][pP][gG]|[mM]4[vV]|[tT][rR][pP]|[fF]4[vV]|[mM]2[tT][sS]|[tT][sS]|264|[hH]264|[mM][kK][vV]|[mM][oO][vV]|[mM][tT][sS]|[mM]2[tT]|[wW][mM][vV]|[aA][sS][fF]|[vV][oO][bB]|3[gG][pP]|3[gG][pP][pP]|[mM][xX][fF]|[dD][iI][vV][xX]|[xX][vV][iI][dD]|[rR][aA][wW]|[dD][vV]|[gG][xX][fF]|[mM]1[vV]|3[gG]2|[vV][mM][fF]|[mM]3[uU]8|[lL][cC][hH]|[gG][xX][fF]_[mM][pP][eE][gG]2|[mM][xX][fF]_[mM][pP][eE][gG]2|[mM][xX][fF][hH][dD]|[wW][aA][vV]|[yY]4[mM]|[aA][aA][cC]|[aA][iI][fF][fF]|[mM][pP]2|[aA][cC]3|[eE][cC]3|[dD][tT][sS][eE])))$" + }, + "__stringPatternWS": { + "type": "string", + "pattern": "^[\\w\\s]*$" + }, + "__timestampIso8601": { + "type": "timestamp", + "timestampFormat": "iso8601" + }, + "__timestampUnix": { + "type": "timestamp", + "timestampFormat": "unixTimestamp" } } } \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/docs-2.json index bf0715d33..2b7504b0f 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/mediaconvert/2017-08-29/docs-2.json @@ -55,7 +55,7 @@ } }, "AacSettings": { - "base": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AAC.", + "base": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode (rateControlMode) to \"VBR\" or \"CBR\". In VBR mode, you control the audio quality with the setting VBR quality (vbrQuality). In CBR mode, you use the setting Bitrate (bitrate). Defaults and valid values depend on the rate control mode.", "refs": { "AudioCodecSettings$AacSettings": null } @@ -109,7 +109,7 @@ } }, "AfdSignaling": { - "base": "This setting only applies to H.264 and MPEG2 outputs. Use Insert AFD signaling (AfdSignaling) to whether there are AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data.", + "base": "This setting only applies to H.264 and MPEG2 outputs. Use Insert AFD signaling (AfdSignaling) to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data.", "refs": { "VideoDescription$AfdSignaling": null } @@ -145,7 +145,7 @@ } }, "AudioDefaultSelection": { - "base": "When an \"Audio Description\":#audio_description specifies an AudioSelector or AudioSelectorGroup for which no matching source is found in the input, then the audio selector marked as DEFAULT will be used. If none are marked as default, silence will be inserted for the duration of the input.", + "base": "Enable this setting on one audio selector to set it as the default for the job. The service uses this default for outputs where it can't find the specified input audio. If you don't set a default, those outputs have no audio.", "refs": { "AudioSelector$DefaultSelection": null } @@ -153,7 +153,7 @@ "AudioDescription": { "base": "Description of audio output", "refs": { - "ListOfAudioDescription$member": null + "__listOfAudioDescription$member": null } }, "AudioLanguageCodeControl": { @@ -195,13 +195,13 @@ "AudioSelector": { "base": "Selector for Audio", "refs": { - "MapOfAudioSelector$member": null + "__mapOfAudioSelector$member": null } }, "AudioSelectorGroup": { "base": "Group of Audio Selectors", "refs": { - "MapOfAudioSelectorGroup$member": null + "__mapOfAudioSelectorGroup$member": null } }, "AudioSelectorType": { @@ -265,7 +265,7 @@ } }, "BurninSubtitleTeletextSpacing": { - "base": "Controls whether a fixed grid size or proportional font spacing will be used to generate the output subtitles bitmap. Only applicable for Teletext inputs and DVB-Sub/Burn-in outputs.", + "base": "Only applies to jobs with input captions in Teletext or STL formats. Specify whether the spacing between letters in your captions is set by the captions grid or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read if the captions are closed caption.", "refs": { "BurninDestinationSettings$TeletextSpacing": null } @@ -283,13 +283,13 @@ "CaptionDescription": { "base": "Description of Caption output", "refs": { - "ListOfCaptionDescription$member": null + "__listOfCaptionDescription$member": null } }, "CaptionDescriptionPreset": { "base": "Caption Description for preset", "refs": { - "ListOfCaptionDescriptionPreset$member": null + "__listOfCaptionDescriptionPreset$member": null } }, "CaptionDestinationSettings": { @@ -306,9 +306,9 @@ } }, "CaptionSelector": { - "base": "Caption inputs to be mapped to caption outputs.", + "base": "Set up captions in your outputs by first selecting them from your input here.", "refs": { - "MapOfCaptionSelector$member": null + "__mapOfCaptionSelector$member": null } }, "CaptionSourceSettings": { @@ -329,6 +329,84 @@ "RemixSettings$ChannelMapping": null } }, + "CmafClientCache": { + "base": "When set to ENABLED, sets #EXT-X-ALLOW-CACHE:no tag, which prevents client from saving media segments for later replay.", + "refs": { + "CmafGroupSettings$ClientCache": null + } + }, + "CmafCodecSpecification": { + "base": "Specification to use (RFC-6381 or the default RFC-4281) during m3u8 playlist generation.", + "refs": { + "CmafGroupSettings$CodecSpecification": null + } + }, + "CmafEncryptionSettings": { + "base": "Settings for CMAF encryption", + "refs": { + "CmafGroupSettings$Encryption": "DRM settings." + } + }, + "CmafEncryptionType": { + "base": "Encrypts the segments with the given encryption scheme. Leave blank to disable. Selecting 'Disabled' in the web interface also disables encryption.", + "refs": { + "CmafEncryptionSettings$EncryptionMethod": null + } + }, + "CmafGroupSettings": { + "base": "Required when you set (Type) under (OutputGroups)>(OutputGroupSettings) to CMAF_GROUP_SETTINGS. Each output in a CMAF Output Group may only contain a single video, audio, or caption output.", + "refs": { + "OutputGroupSettings$CmafGroupSettings": null + } + }, + "CmafInitializationVectorInManifest": { + "base": "The Initialization Vector is a 128-bit number used in conjunction with the key for encrypting blocks. If set to INCLUDE, Initialization Vector is listed in the manifest. Otherwise Initialization Vector is not in the manifest.", + "refs": { + "CmafEncryptionSettings$InitializationVectorInManifest": null + } + }, + "CmafKeyProviderType": { + "base": "Indicates which type of key provider is used for encryption.", + "refs": { + "CmafEncryptionSettings$Type": null + } + }, + "CmafManifestCompression": { + "base": "When set to GZIP, compresses HLS playlist.", + "refs": { + "CmafGroupSettings$ManifestCompression": null + } + }, + "CmafManifestDurationFormat": { + "base": "Indicates whether the output manifest should use floating point values for segment duration.", + "refs": { + "CmafGroupSettings$ManifestDurationFormat": null + } + }, + "CmafSegmentControl": { + "base": "When set to SINGLE_FILE, a single output file is generated, which is internally segmented using the Fragment Length and Segment Length. When set to SEGMENTED_FILES, separate segment files will be created.", + "refs": { + "CmafGroupSettings$SegmentControl": null + } + }, + "CmafStreamInfResolution": { + "base": "Include or exclude RESOLUTION attribute for video in EXT-X-STREAM-INF tag of variant manifest.", + "refs": { + "CmafGroupSettings$StreamInfResolution": null + } + }, + "CmafWriteDASHManifest": { + "base": "When set to ENABLED, a DASH MPD manifest will be generated for this output.", + "refs": { + "CmafGroupSettings$WriteDashManifest": null + } + }, + "CmafWriteHLSManifest": { + "base": "When set to ENABLED, an Apple HLS manifest will be generated for this output.", + "refs": { + "CmafGroupSettings$WriteHlsManifest": null + } + }, "ColorCorrector": { "base": "Settings for color correction.", "refs": { @@ -342,7 +420,7 @@ } }, "ColorSpace": { - "base": "Specifies the colorspace of an input. This setting works in tandem with \"Color Corrector\":#color_corrector > color_space_conversion to determine if any conversion will be performed.", + "base": "If your input video has accurate color space metadata, or if you don't know about color space, leave this set to the default value FOLLOW. The service will automatically detect your input color space. If your input video has metadata indicating the wrong color space, or if your input video is missing color space metadata that should be there, specify the accurate color space here. If you choose HDR10, you can also correct inaccurate color space coefficients, using the HDR master display information controls. You must also set Color space usage (ColorSpaceUsage) to FORCE for the service to use these values.", "refs": { "VideoSelector$ColorSpace": null } @@ -354,7 +432,7 @@ } }, "ColorSpaceUsage": { - "base": "There are two sources for color metadata, the input file and the job configuration. This enum controls which takes precedence. FORCE: System will use color metadata supplied by user, if any. If the user does not supply color metadata the system will use data from the source. FALLBACK: System will use color metadata from the source. If source has no color metadata, the system will use user-supplied color metadata values if available.", + "base": "There are two sources for color metadata, the input file and the job configuration (in the Color space and HDR master display informaiton settings). The Color space usage setting controls which takes precedence. FORCE: The system will use color metadata supplied by user, if any. If the user does not supply color metadata, the system will use data from the source. FALLBACK: The system will use color metadata from the source. If source has no color metadata, the system will use user-supplied color metadata values if available.", "refs": { "VideoSelector$ColorSpaceUsage": null } @@ -566,7 +644,7 @@ } }, "DvbSubtitleTeletextSpacing": { - "base": "Controls whether a fixed grid size or proportional font spacing will be used to generate the output subtitles bitmap. Only applicable for Teletext inputs and DVB-Sub/Burn-in outputs.", + "base": "Only applies to jobs with input captions in Teletext or STL formats. Specify whether the spacing between letters in your captions is set by the captions grid or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read if the captions are closed caption.", "refs": { "DvbSubDestinationSettings$TeletextSpacing": null } @@ -682,7 +760,7 @@ "Endpoint": { "base": "Describes account specific API endpoint", "refs": { - "ListOfEndpoint$member": null + "__listOfEndpoint$member": null } }, "ExceptionBody": { @@ -808,7 +886,7 @@ } }, "H264FramerateControl": { - "base": "Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Using the console, do this by choosing INITIALIZE_FROM_SOURCE for Framerate.", + "base": "If you are using the console, use the Framerate setting to specify the framerate for this output. If you want to keep the same framerate as the input video, choose Follow source. If you want to do framerate conversion, choose a framerate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your framerate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Choose SPECIFIED if you want the service to use the framerate you specify in the settings FramerateNumerator and FramerateDenominator.", "refs": { "H264Settings$FramerateControl": null } @@ -832,7 +910,7 @@ } }, "H264InterlaceMode": { - "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type. - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\". - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", + "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type, as follows.\n - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\".\n - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", "refs": { "H264Settings$InterlaceMode": null } @@ -850,7 +928,7 @@ } }, "H264RateControlMode": { - "base": "Rate control mode. CQ uses constant quantizer (qp), ABR (average bitrate) does not write HRD parameters.", + "base": "Use this setting to specify whether this output has a variable bitrate (VBR) or constant bitrate (CBR).", "refs": { "H264Settings$RateControlMode": null } @@ -940,7 +1018,7 @@ } }, "H265FramerateControl": { - "base": "Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Using the console, do this by choosing INITIALIZE_FROM_SOURCE for Framerate.", + "base": "If you are using the console, use the Framerate setting to specify the framerate for this output. If you want to keep the same framerate as the input video, choose Follow source. If you want to do framerate conversion, choose a framerate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your framerate as a fraction. If you are creating your transcoding job sepecification as a JSON file without the console, use FramerateControl to specify which value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Choose SPECIFIED if you want the service to use the framerate you specify in the settings FramerateNumerator and FramerateDenominator.", "refs": { "H265Settings$FramerateControl": null } @@ -964,7 +1042,7 @@ } }, "H265InterlaceMode": { - "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type. - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\". - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", + "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type.\n - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\".\n - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", "refs": { "H265Settings$InterlaceMode": null } @@ -982,7 +1060,7 @@ } }, "H265RateControlMode": { - "base": "Rate control mode. CQ uses constant quantizer (qp), ABR (average bitrate) does not write HRD parameters.", + "base": "Use this setting to specify whether this output has a variable bitrate (VBR) or constant bitrate (CBR).", "refs": { "H265Settings$RateControlMode": null } @@ -1047,8 +1125,14 @@ "H265Settings$UnregisteredSeiTimecode": null } }, + "H265WriteMp4PackagingType": { + "base": "If HVC1, output that is H.265 will be marked as HVC1 and adhere to the ISO-IECJTC1-SC29_N13798_Text_ISOIEC_FDIS_14496-15_3rd_E spec which states that parameter set NAL units will be stored in the sample headers but not in the samples directly. If HEV1, then H.265 will be marked as HEV1 and parameter set NAL units will be written into the samples.", + "refs": { + "H265Settings$WriteMp4PackagingType": null + } + }, "Hdr10Metadata": { - "base": "Use the HDR master display (Hdr10Metadata) settings to provide values for HDR color. These values vary depending on the input video and must be provided by a color grader. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", + "base": "Use the HDR master display (Hdr10Metadata) settings to correct HDR metadata or to provide missing metadata. These values vary depending on the input video and must be provided by a color grader. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that these settings are not color correction. Note that if you are creating HDR outputs inside of an HLS CMAF package, to comply with the Apple specification, you must use the HVC1 for H.265 setting.", "refs": { "ColorCorrector$Hdr10Metadata": null, "VideoSelector$Hdr10Metadata": null @@ -1057,7 +1141,7 @@ "HlsAdMarkers": { "base": null, "refs": { - "ListOfHlsAdMarkers$member": null + "__listOfHlsAdMarkers$member": null } }, "HlsAudioTrackType": { @@ -1069,7 +1153,7 @@ "HlsCaptionLanguageMapping": { "base": "Caption Language Mapping", "refs": { - "ListOfHlsCaptionLanguageMapping$member": null + "__listOfHlsCaptionLanguageMapping$member": null } }, "HlsCaptionLanguageSetting": { @@ -1181,9 +1265,9 @@ } }, "Id3Insertion": { - "base": "To insert ID3 tags in your output, specify two values. Use ID3 tag (Id3) to specify the base 64 encoded string and use Timecode (TimeCode) to specify the time when the tag should be inserted. To insert multiple ID3 tags in your output, create mulitple instances of ID3 insertion (Id3Insertion).", + "base": "To insert ID3 tags in your output, specify two values. Use ID3 tag (Id3) to specify the base 64 encoded string and use Timecode (TimeCode) to specify the time when the tag should be inserted. To insert multiple ID3 tags in your output, create multiple instances of ID3 insertion (Id3Insertion).", "refs": { - "ListOfId3Insertion$member": null + "__listOfId3Insertion$member": null } }, "ImageInserter": { @@ -1195,13 +1279,13 @@ "Input": { "base": "Specifies media input", "refs": { - "ListOfInput$member": null + "__listOfInput$member": null } }, "InputClipping": { - "base": "Include one instance of (InputClipping) for each input clip.", + "base": "To transcode only portions of your input (clips), include one Input clipping (one instance of InputClipping in the JSON job file) for each input clip. All input clips you specify will be included in every output of the job.", "refs": { - "ListOfInputClipping$member": null + "__listOfInputClipping$member": null } }, "InputDeblockFilter": { @@ -1235,11 +1319,11 @@ "InputTemplate": { "base": "Specified video input in a template.", "refs": { - "ListOfInputTemplate$member": null + "__listOfInputTemplate$member": null } }, "InputTimecodeSource": { - "base": "Use Timecode source (InputTimecodeSource) to specify how timecode information from your input is adjusted and encoded in all outputs for the job. Default is embedded. Set to Embedded (EMBEDDED) to use the timecode that is in the input video. If no embedded timecode is in the source, will set the timecode for the first frame to 00:00:00:00. Set to Start at 0 (ZEROBASED) to set the timecode of the initial frame to 00:00:00:00. Set to Specified start (SPECIFIEDSTART) to provide the initial timecode yourself the setting (Start).", + "base": "Timecode source under input settings (InputTimecodeSource) only affects the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Use this setting to specify whether the service counts frames by timecodes embedded in the video (EMBEDDED) or by starting the first frame at zero (ZEROBASED). In both cases, the timecode format is HH:MM:SS:FF or HH:MM:SS;FF, where FF is the frame number. Only set this to EMBEDDED if your source video has embedded timecodes.", "refs": { "Input$TimecodeSource": null, "InputTemplate$TimecodeSource": null @@ -1248,7 +1332,7 @@ "InsertableImage": { "base": "Settings for Insertable Image", "refs": { - "ListOfInsertableImage$member": null + "__listOfInsertableImage$member": null } }, "InternalServerErrorException": { @@ -1261,7 +1345,7 @@ "refs": { "CreateJobResponse$Job": null, "GetJobResponse$Job": null, - "ListOfJob$member": null + "__listOfJob$member": null } }, "JobSettings": { @@ -1283,8 +1367,8 @@ "refs": { "CreateJobTemplateResponse$JobTemplate": null, "GetJobTemplateResponse$JobTemplate": null, - "ListOfJobTemplate$member": null, - "UpdateJobTemplateResponse$JobTemplate": null + "UpdateJobTemplateResponse$JobTemplate": null, + "__listOfJobTemplate$member": null } }, "JobTemplateListBy": { @@ -1302,7 +1386,7 @@ } }, "LanguageCode": { - "base": "Code to specify the language, following the specification \"ISO 639-2 three-digit code\":http://www.loc.gov/standards/iso639-2/", + "base": "Specify the language, using the ISO 639-2 three-letter code listed at https://www.loc.gov/standards/iso639-2/php/code_list.php.", "refs": { "AudioDescription$LanguageCode": "Indicates the language of the audio output track. The ISO 639 language specified in the 'Language Code' drop down will be used when 'Follow Input Language Code' is not selected or when 'Follow Input Language Code' is selected but there is no ISO 639 language code specified by the input.", "AudioSelector$LanguageCode": "Selects a specific language code from within an audio source.", @@ -1332,147 +1416,6 @@ "refs": { } }, - "ListOfAudioDescription": { - "base": null, - "refs": { - "Output$AudioDescriptions": "(AudioDescriptions) contains groups of audio encoding settings organized by audio codec. Include one instance of (AudioDescriptions) per output. (AudioDescriptions) can contain multiple groups of encoding settings.", - "PresetSettings$AudioDescriptions": "(AudioDescriptions) contains groups of audio encoding settings organized by audio codec. Include one instance of (AudioDescriptions) per output. (AudioDescriptions) can contain multiple groups of encoding settings." - } - }, - "ListOfCaptionDescription": { - "base": null, - "refs": { - "Output$CaptionDescriptions": "(CaptionDescriptions) contains groups of captions settings. For each output that has captions, include one instance of (CaptionDescriptions). (CaptionDescriptions) can contain multiple groups of captions settings." - } - }, - "ListOfCaptionDescriptionPreset": { - "base": null, - "refs": { - "PresetSettings$CaptionDescriptions": "Caption settings for this preset. There can be multiple caption settings in a single output." - } - }, - "ListOfEndpoint": { - "base": null, - "refs": { - "DescribeEndpointsResponse$Endpoints": "List of endpoints" - } - }, - "ListOfHlsAdMarkers": { - "base": null, - "refs": { - "HlsGroupSettings$AdMarkers": "Choose one or more ad marker types to pass SCTE35 signals through to this group of Apple HLS outputs." - } - }, - "ListOfHlsCaptionLanguageMapping": { - "base": null, - "refs": { - "HlsGroupSettings$CaptionLanguageMappings": "Language to be used on Caption outputs" - } - }, - "ListOfId3Insertion": { - "base": null, - "refs": { - "TimedMetadataInsertion$Id3Insertions": "Id3Insertions contains the array of Id3Insertion instances." - } - }, - "ListOfInput": { - "base": null, - "refs": { - "JobSettings$Inputs": "Use Inputs (inputs) to define source file used in the transcode job. There can be multiple inputs add in a job. These inputs will be concantenated together to create the output." - } - }, - "ListOfInputClipping": { - "base": null, - "refs": { - "Input$InputClippings": "(InputClippings) contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them.", - "InputTemplate$InputClippings": "(InputClippings) contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them." - } - }, - "ListOfInputTemplate": { - "base": null, - "refs": { - "JobTemplateSettings$Inputs": "Use Inputs (inputs) to define the source file used in the transcode job. There can only be one input in a job template. Using the API, you can include multiple inputs when referencing a job template." - } - }, - "ListOfInsertableImage": { - "base": null, - "refs": { - "ImageInserter$InsertableImages": "Image to insert. Must be 32 bit windows BMP, PNG, or TGA file. Must not be larger than the output frames." - } - }, - "ListOfJob": { - "base": null, - "refs": { - "ListJobsResponse$Jobs": "List of jobs" - } - }, - "ListOfJobTemplate": { - "base": null, - "refs": { - "ListJobTemplatesResponse$JobTemplates": "List of Job templates." - } - }, - "ListOfOutput": { - "base": null, - "refs": { - "OutputGroup$Outputs": "This object holds groups of encoding settings, one group of settings per output." - } - }, - "ListOfOutputChannelMapping": { - "base": null, - "refs": { - "ChannelMapping$OutputChannels": "List of output channels" - } - }, - "ListOfOutputDetail": { - "base": null, - "refs": { - "OutputGroupDetail$OutputDetails": "Details about the output" - } - }, - "ListOfOutputGroup": { - "base": null, - "refs": { - "JobSettings$OutputGroups": "**!!**(OutputGroups) contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in (OutputGroups) is a group of settings that apply to the whole group. This required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings", - "JobTemplateSettings$OutputGroups": "**!!**(OutputGroups) contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in (OutputGroups) is a group of settings that apply to the whole group. This required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings" - } - }, - "ListOfOutputGroupDetail": { - "base": null, - "refs": { - "Job$OutputGroupDetails": "List of output group details" - } - }, - "ListOfPreset": { - "base": null, - "refs": { - "ListPresetsResponse$Presets": "List of presets" - } - }, - "ListOfQueue": { - "base": null, - "refs": { - "ListQueuesResponse$Queues": "List of queues" - } - }, - "ListOf__integer": { - "base": null, - "refs": { - "AudioSelector$Pids": "Selects a specific PID from within an audio source (e.g. 257 selects PID 0x101).", - "AudioSelector$Tracks": "Identify the channel to include in this selector by entering the 1-based track index. To combine several tracks, enter a comma-separated list, e.g. \"1,2,3\" for tracks 1-3.", - "M2tsSettings$AudioPids": "Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation.", - "M2tsSettings$DvbSubPids": "Packet Identifier (PID) for input source DVB Subtitle data to this output. Multiple values are accepted, and can be entered in ranges and/or by comma separation.", - "M3u8Settings$AudioPids": "Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation.", - "OutputChannelMapping$InputChannels": "List of input channels" - } - }, - "ListOf__string": { - "base": null, - "refs": { - "AudioSelectorGroup$AudioSelectorNames": "Name of an \"Audio Selector\":#inputs-audio_selector within the same input to include in the group. Audio selector names are standardized, based on their order within the input (e.g. \"Audio Selector 1\"). The audio_selector_name parameter can be repeated to add any number of audio selectors to the group.", - "SpekeKeyProvider$SystemIds": "Relates to SPEKE implementation. DRM system identifiers. DASH output groups support a max of two system ids. Other group types support one system id." - } - }, "ListPresetsRequest": { "base": "You can send list presets requests with an empty body. Optionally, you can filter the response by category by specifying it in your request body. You can also optionally specify the maximum number, up to twenty, of queues to be returned.", "refs": { @@ -1589,34 +1532,6 @@ "ContainerSettings$M3u8Settings": null } }, - "MapOfAudioSelector": { - "base": null, - "refs": { - "Input$AudioSelectors": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use mutiple Audio selectors per input.", - "InputTemplate$AudioSelectors": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use mutiple Audio selectors per input." - } - }, - "MapOfAudioSelectorGroup": { - "base": null, - "refs": { - "Input$AudioSelectorGroups": "Specifies set of audio selectors within an input to combine. An input may have multiple audio selector groups. See \"Audio Selector Group\":#inputs-audio_selector_group for more information.", - "InputTemplate$AudioSelectorGroups": "Specifies set of audio selectors within an input to combine. An input may have multiple audio selector groups. See \"Audio Selector Group\":#inputs-audio_selector_group for more information." - } - }, - "MapOfCaptionSelector": { - "base": null, - "refs": { - "Input$CaptionSelectors": "Use Captions selectors (CaptionSelectors) to specify the captions data from the input that you will use in your outputs. You can use mutiple captions selectors per input.", - "InputTemplate$CaptionSelectors": "Use Captions selectors (CaptionSelectors) to specify the captions data from the input that you will use in your outputs. You can use mutiple captions selectors per input." - } - }, - "MapOf__string": { - "base": null, - "refs": { - "CreateJobRequest$UserMetadata": "User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs.", - "Job$UserMetadata": "User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs." - } - }, "MovClapAtom": { "base": "When enabled, include 'clap' atom if appropriate for the video output settings.", "refs": { @@ -1702,7 +1617,7 @@ } }, "Mpeg2FramerateControl": { - "base": "Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Using the console, do this by choosing INITIALIZE_FROM_SOURCE for Framerate.", + "base": "If you are using the console, use the Framerate setting to specify the framerate for this output. If you want to keep the same framerate as the input video, choose Follow source. If you want to do framerate conversion, choose a framerate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your framerate as a fraction. If you are creating your transcoding job sepecification as a JSON file without the console, use FramerateControl to specify which value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Choose SPECIFIED if you want the service to use the framerate you specify in the settings FramerateNumerator and FramerateDenominator.", "refs": { "Mpeg2Settings$FramerateControl": null } @@ -1720,7 +1635,7 @@ } }, "Mpeg2InterlaceMode": { - "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type. - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\". - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", + "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type.\n - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\".\n - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", "refs": { "Mpeg2Settings$InterlaceMode": null } @@ -1829,7 +1744,7 @@ } }, "NoiseReducerFilter": { - "base": "Use Noise reducer filter (NoiseReducerFilter) to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer (NoiseReducer). * Bilateral is an edge preserving noise reduction filter * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) are convolution filters * Conserve is a min/max noise reduction filter * Spatial is frequency-domain filter based on JND principles.", + "base": "Use Noise reducer filter (NoiseReducerFilter) to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer (NoiseReducer). * Bilateral is an edge preserving noise reduction filter. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) are convolution filters. * Conserve is a min/max noise reduction filter. * Spatial is a frequency-domain filter based on JND principles.", "refs": { "NoiseReducer$Filter": null } @@ -1863,31 +1778,31 @@ "Output": { "base": "An output object describes the settings for a single output file or stream in an output group.", "refs": { - "ListOfOutput$member": null + "__listOfOutput$member": null } }, "OutputChannelMapping": { "base": "OutputChannel mapping settings.", "refs": { - "ListOfOutputChannelMapping$member": null + "__listOfOutputChannelMapping$member": null } }, "OutputDetail": { "base": "Details regarding output", "refs": { - "ListOfOutputDetail$member": null + "__listOfOutputDetail$member": null } }, "OutputGroup": { "base": "Group of outputs", "refs": { - "ListOfOutputGroup$member": null + "__listOfOutputGroup$member": null } }, "OutputGroupDetail": { "base": "Contains details about the output groups specified in the job settings.", "refs": { - "ListOfOutputGroupDetail$member": null + "__listOfOutputGroupDetail$member": null } }, "OutputGroupSettings": { @@ -1897,7 +1812,7 @@ } }, "OutputGroupType": { - "base": "Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming)", + "base": "Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming, CMAF)", "refs": { "OutputGroupSettings$Type": null } @@ -1919,8 +1834,8 @@ "refs": { "CreatePresetResponse$Preset": null, "GetPresetResponse$Preset": null, - "ListOfPreset$member": null, - "UpdatePresetResponse$Preset": null + "UpdatePresetResponse$Preset": null, + "__listOfPreset$member": null } }, "PresetListBy": { @@ -1944,7 +1859,7 @@ } }, "ProresFramerateControl": { - "base": "Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Using the console, do this by choosing INITIALIZE_FROM_SOURCE for Framerate.", + "base": "If you are using the console, use the Framerate setting to specify the framerate for this output. If you want to keep the same framerate as the input video, choose Follow source. If you want to do framerate conversion, choose a framerate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your framerate as a fraction. If you are creating your transcoding job sepecification as a JSON file without the console, use FramerateControl to specify which value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the framerate from the input. Choose SPECIFIED if you want the service to use the framerate you specify in the settings FramerateNumerator and FramerateDenominator.", "refs": { "ProresSettings$FramerateControl": null } @@ -1956,7 +1871,7 @@ } }, "ProresInterlaceMode": { - "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type. - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\". - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", + "base": "Use Interlace mode (InterlaceMode) to choose the scan line type for the output. * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce interlaced output with the entire output having the same field polarity (top or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, behavior depends on the input scan type.\n - If the source is interlaced, the output will be interlaced with the same polarity as the source (it will follow the source). The output could therefore be a mix of \"top field first\" and \"bottom field first\".\n - If the source is progressive, the output will be interlaced with \"top field first\" or \"bottom field first\" polarity, depending on which of the Follow options you chose.", "refs": { "ProresSettings$InterlaceMode": null } @@ -1990,8 +1905,8 @@ "refs": { "CreateQueueResponse$Queue": null, "GetQueueResponse$Queue": null, - "ListOfQueue$member": null, - "UpdateQueueResponse$Queue": null + "UpdateQueueResponse$Queue": null, + "__listOfQueue$member": null } }, "QueueListBy": { @@ -2015,10 +1930,10 @@ } }, "RemixSettings": { - "base": "Use Manual audio remixing (RemixSettings) to adjust audio levels for each output channel. With audio remixing, you can output more or fewer audio channels than your input audio source provides.", + "base": "Use Manual audio remixing (RemixSettings) to adjust audio levels for each audio channel in each output of your job. With audio remixing, you can output more or fewer audio channels than your input audio source provides.", "refs": { "AudioDescription$RemixSettings": "Advanced audio remixing settings.", - "AudioSelector$RemixSettings": "Advanced audio remixing settings." + "AudioSelector$RemixSettings": "Use these settings to reorder the audio channels of one input to match those of another input. This allows you to combine the two files into a single output, one after the other." } }, "RespondToAfd": { @@ -2056,6 +1971,7 @@ "StaticKeyProvider": { "base": "Settings for use with a SPEKE key provider.", "refs": { + "CmafEncryptionSettings$StaticKeyProvider": null, "HlsEncryptionSettings$StaticKeyProvider": null } }, @@ -2084,20 +2000,20 @@ } }, "TimecodeConfig": { - "base": "Contains settings used to acquire and adjust timecode information from inputs.", + "base": "These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping.", "refs": { "JobSettings$TimecodeConfig": "Contains settings used to acquire and adjust timecode information from inputs.", "JobTemplateSettings$TimecodeConfig": "Contains settings used to acquire and adjust timecode information from inputs." } }, "TimecodeSource": { - "base": "Use Timecode source (TimecodeSource) to set how timecodes are handled within this input. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 (ZEROBASED) instead. * Start at 0 (ZEROBASED) - Set the timecode of the initial frame to 00:00:00:00. * Specified Start (SPECIFIEDSTART) - Set the timecode of the initial frame to a value other than zero. You use Start timecode (Start) to provide this value.", + "base": "Use Source (TimecodeSource) to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 (ZEROBASED) instead. * Start at 0 (ZEROBASED) - Set the timecode of the initial frame to 00:00:00:00. * Specified Start (SPECIFIEDSTART) - Set the timecode of the initial frame to a value other than zero. You use Start timecode (Start) to provide this value.", "refs": { "TimecodeConfig$Source": null } }, "TimedMetadata": { - "base": "If PASSTHROUGH, inserts ID3 timed metadata from the timed_metadata REST command into this output.", + "base": "Applies only to HLS outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output.", "refs": { "M3u8Settings$TimedMetadata": null } @@ -2209,25 +2125,26 @@ } }, "VideoTimecodeInsertion": { - "base": "Enable Timecode insertion to include timecode information in this output. Do this in the API by setting (VideoTimecodeInsertion) to (PIC_TIMING_SEI). To get timecodes to appear correctly in your output, also set up the timecode configuration for your job in the input settings. Only enable Timecode insertion when the input framerate is identical to output framerate. Disable this setting to remove the timecode from the output. Default is disabled.", + "base": "Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input framerate is identical to the output framerate. To include timecodes in this output, set Timecode insertion (VideoTimecodeInsertion) to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration (TimecodeConfig). In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings (InputTimecodeSource) does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration (TimecodeSource) does.", "refs": { "VideoDescription$TimecodeInsertion": null } }, + "WavFormat": { + "base": "The service defaults to using RIFF for WAV outputs. If your output audio is likely to exceed 4 GB in file size, or if you otherwise need the extended support of the RF64 format, set your output WAV file format to RF64.", + "refs": { + "WavSettings$Format": null + } + }, "WavSettings": { "base": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value WAV.", "refs": { "AudioCodecSettings$WavSettings": null } }, - "__double": { + "__doubleMin0": { "base": null, "refs": { - "AudioNormalizationSettings$TargetLkfs": "Target LKFS(loudness) to adjust volume to. If no value is entered, a default value will be used according to the chosen algorithm. The CALM Act (1770-1) recommends a target of -24 LKFS. The EBU R-128 specification (1770-2) recommends a target of -23 LKFS.", - "Eac3Settings$LoRoCenterMixLevel": "Left only/Right only center mix level. Only used for 3/2 coding mode.\nValid values: 3.0, 1.5, 0.0, -1.5 -3.0 -4.5 -6.0 -60", - "Eac3Settings$LoRoSurroundMixLevel": "Left only/Right only surround mix level. Only used for 3/2 coding mode.\nValid values: -1.5 -3.0 -4.5 -6.0 -60", - "Eac3Settings$LtRtCenterMixLevel": "Left total/Right total center mix level. Only used for 3/2 coding mode.\nValid values: 3.0, 1.5, 0.0, -1.5 -3.0 -4.5 -6.0 -60", - "Eac3Settings$LtRtSurroundMixLevel": "Left total/Right total surround mix level. Only used for 3/2 coding mode.\nValid values: -1.5 -3.0 -4.5 -6.0 -60", "H264Settings$GopSize": "GOP Length (keyframe interval) in frames or seconds. Must be greater than zero.", "H265Settings$GopSize": "GOP Length (keyframe interval) in frames or seconds. Must be greater than zero.", "M2tsSettings$FragmentTime": "The length in seconds of each fragment. Only used with EBP markers.", @@ -2236,215 +2153,736 @@ "Mpeg2Settings$GopSize": "GOP Length (keyframe interval) in frames or seconds. Must be greater than zero." } }, + "__doubleMinNegative59Max0": { + "base": null, + "refs": { + "AudioNormalizationSettings$TargetLkfs": "Target LKFS(loudness) to adjust volume to. If no value is entered, a default value will be used according to the chosen algorithm. The CALM Act (1770-1) recommends a target of -24 LKFS. The EBU R-128 specification (1770-2) recommends a target of -23 LKFS." + } + }, + "__doubleMinNegative60Max3": { + "base": null, + "refs": { + "Eac3Settings$LoRoCenterMixLevel": "Left only/Right only center mix level. Only used for 3/2 coding mode.\nValid values: 3.0, 1.5, 0.0, -1.5 -3.0 -4.5 -6.0 -60", + "Eac3Settings$LtRtCenterMixLevel": "Left total/Right total center mix level. Only used for 3/2 coding mode.\nValid values: 3.0, 1.5, 0.0, -1.5 -3.0 -4.5 -6.0 -60" + } + }, + "__doubleMinNegative60MaxNegative1": { + "base": null, + "refs": { + "Eac3Settings$LoRoSurroundMixLevel": "Left only/Right only surround mix level. Only used for 3/2 coding mode.\nValid values: -1.5 -3.0 -4.5 -6.0 -60", + "Eac3Settings$LtRtSurroundMixLevel": "Left total/Right total surround mix level. Only used for 3/2 coding mode.\nValid values: -1.5 -3.0 -4.5 -6.0 -60" + } + }, "__integer": { "base": null, "refs": { - "AacSettings$Bitrate": "Average bitrate in bits/second. Valid values depend on rate control mode and profile.", - "AacSettings$SampleRate": "Sample rate in Hz. Valid values depend on rate control mode and profile.", - "Ac3Settings$Bitrate": "Average bitrate in bits/second. Valid bitrates depend on the coding mode.", - "Ac3Settings$Dialnorm": "Sets the dialnorm for the output. If blank and input audio is Dolby Digital, dialnorm will be passed through.", - "Ac3Settings$SampleRate": "Sample rate in hz. Sample rate is always 48000.", - "AiffSettings$BitDepth": "Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track.", - "AiffSettings$Channels": "Set Channels to specify the number of channels in this output audio track. Choosing Mono in the console will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2.", - "AiffSettings$SampleRate": "Sample rate in hz.", - "AncillarySourceSettings$SourceAncillaryChannelNumber": "Specifies the 608 channel number in the ancillary data track from which to extract captions. Unused for passthrough.", - "AudioDescription$AudioType": "Applies only if Follow Input Audio Type is unchecked (false). A number between 0 and 255. The following are defined in ISO-IEC 13818-1: 0 = Undefined, 1 = Clean Effects, 2 = Hearing Impaired, 3 = Visually Impaired Commentary, 4-255 = Reserved.", - "AudioNormalizationSettings$CorrectionGateLevel": "Content measuring above this level will be corrected to the target level. Content measuring below this level will not be corrected. Gating only applies when not using real_time_correction.", - "AudioSelector$Offset": "Specifies a time delta in milliseconds to offset the audio from the input video.", - "AudioSelector$ProgramSelection": "Applies only when input streams contain Dolby E. Enter the program ID (according to the metadata in the audio) of the Dolby E program to extract from the specified track. One program extracted per audio selector. To select multiple programs, create multiple selectors with the same Track and different Program numbers. \"All channels\" means to ignore the program IDs and include all the channels in this selector; useful if metadata is known to be incorrect.", - "BurninDestinationSettings$BackgroundOpacity": "Specifies the opacity of the background rectangle. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", - "BurninDestinationSettings$FontOpacity": "Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.\nAll burn-in and DVB-Sub font settings must match.", - "BurninDestinationSettings$FontResolution": "Font resolution in DPI (dots per inch); default is 96 dpi.\nAll burn-in and DVB-Sub font settings must match.", - "BurninDestinationSettings$FontSize": "A positive integer indicates the exact font size in points. Set to 0 for automatic font size selection. All burn-in and DVB-Sub font settings must match.", + "DescribeEndpointsRequest$MaxResults": "Optional. Max number of endpoints, up to twenty, that will be returned at one time.", + "Job$ErrorCode": "Error code for the job", + "ListJobTemplatesRequest$MaxResults": "Optional. Number of job templates, up to twenty, that will be returned at one time.", + "ListJobsRequest$MaxResults": "Optional. Number of jobs, up to twenty, that will be returned at one time.", + "ListPresetsRequest$MaxResults": "Optional. Number of presets, up to twenty, that will be returned at one time", + "ListQueuesRequest$MaxResults": "Optional. Number of queues, up to twenty, that will be returned at one time.", + "OutputDetail$DurationInMs": "Duration in milliseconds", + "VideoDetail$HeightInPx": "Height in pixels for the output", + "VideoDetail$WidthInPx": "Width in pixels for the output" + } + }, + "__integerMin0Max10": { + "base": null, + "refs": { "BurninDestinationSettings$OutlineSize": "Specifies font outline size in pixels. This option is not valid for source captions that are either 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.", - "BurninDestinationSettings$ShadowOpacity": "Specifies the opacity of the shadow. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", - "BurninDestinationSettings$ShadowXOffset": "Specifies the horizontal offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match.", - "BurninDestinationSettings$ShadowYOffset": "Specifies the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$OutlineSize": "Specifies font outline size in pixels. This option is not valid for source captions that are either 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match." + } + }, + "__integerMin0Max100": { + "base": null, + "refs": { + "H264Settings$HrdBufferInitialFillPercentage": "Percentage of the buffer that should initially be filled (HRD buffer model).", + "H265Settings$HrdBufferInitialFillPercentage": "Percentage of the buffer that should initially be filled (HRD buffer model).", + "InsertableImage$Opacity": "Use Opacity (Opacity) to specify how much of the underlying video shows through the inserted image. 0 is transparent and 100 is fully opaque. Default is 50.", + "Mpeg2Settings$HrdBufferInitialFillPercentage": "Percentage of the buffer that should initially be filled (HRD buffer model).", + "VideoDescription$Sharpness": "Use Sharpness (Sharpness)setting to specify the strength of anti-aliasing. This setting changes the width of the anti-alias filter kernel used for scaling. Sharpness only applies if your output resolution is different from your input resolution, and if you set Anti-alias (AntiAlias) to ENABLED. 0 is the softest setting, 100 the sharpest, and 50 recommended for most content." + } + }, + "__integerMin0Max1000": { + "base": null, + "refs": { + "M2tsSettings$PatInterval": "The number of milliseconds between instances of this table in the output transport stream.", + "M2tsSettings$PmtInterval": "The number of milliseconds between instances of this table in the output transport stream.", + "M3u8Settings$PatInterval": "The number of milliseconds between instances of this table in the output transport stream.", + "M3u8Settings$PmtInterval": "The number of milliseconds between instances of this table in the output transport stream." + } + }, + "__integerMin0Max10000": { + "base": null, + "refs": { + "M2tsSettings$MinEbpInterval": "When set, enforces that Encoder Boundary Points do not come within the specified time interval of each other by looking ahead at input video. If another EBP is going to come in within the specified time interval, the current EBP is not emitted, and the segment is \"stretched\" to the next marker. The lookahead value does not add latency to the system. The Live Event must be configured elsewhere to create sufficient latency to make the lookahead accurate." + } + }, + "__integerMin0Max1152000000": { + "base": null, + "refs": { + "H264Settings$HrdBufferSize": "Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000." + } + }, + "__integerMin0Max128": { + "base": null, + "refs": { + "H264Settings$Softness": "Softness. Selects quantizer matrix, larger values reduce high-frequency content in the encoded image.", + "Mpeg2Settings$Softness": "Softness. Selects quantizer matrix, larger values reduce high-frequency content in the encoded image." + } + }, + "__integerMin0Max1466400000": { + "base": null, + "refs": { + "H265Settings$HrdBufferSize": "Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000." + } + }, + "__integerMin0Max15": { + "base": null, + "refs": { + "VideoDescription$FixedAfd": "Applies only if you set AFD Signaling(AfdSignaling) to Fixed (FIXED). Use Fixed (FixedAfd) to specify a four-bit AFD value which the service will write on all frames of this video output." + } + }, + "__integerMin0Max16": { + "base": null, + "refs": { + "NoiseReducerSpatialFilterSettings$Strength": "Relative strength of noise reducing filter. Higher values produce stronger filtering." + } + }, + "__integerMin0Max2147483647": { + "base": null, + "refs": { "BurninDestinationSettings$XPosition": "Specifies the horizontal position of the caption relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.", "BurninDestinationSettings$YPosition": "Specifies the vertical position of the caption relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.", - "ColorCorrector$Brightness": "Brightness level.", - "ColorCorrector$Contrast": "Contrast level.", - "ColorCorrector$Hue": "Hue in degrees.", - "ColorCorrector$Saturation": "Saturation level.", - "DashIsoGroupSettings$FragmentLength": "Length of fragments to generate (in seconds). Fragment length must be compatible with GOP size and Framerate. Note that fragments will end on the next keyframe after this number of seconds, so actual fragment length may be longer. When Emit Single File is checked, the fragmentation is internal to a single output file and it does not cause the creation of many output files as in other output types.", + "CmafGroupSettings$MinBufferTime": "Minimum time of initially buffered media that is needed to ensure smooth playout.", "DashIsoGroupSettings$MinBufferTime": "Minimum time of initially buffered media that is needed to ensure smooth playout.", - "DashIsoGroupSettings$SegmentLength": "Length of mpd segments to create (in seconds). Note that segments will end on the next keyframe after this number of seconds, so actual segment length may be longer. When Emit Single File is checked, the segmentation is internal to a single output file and it does not cause the creation of many output files as in other output types.", - "DescribeEndpointsRequest$MaxResults": "Optional. Max number of endpoints, up to twenty, that will be returned at one time.", - "DvbNitSettings$NetworkId": "The numeric value placed in the Network Information Table (NIT).", - "DvbNitSettings$NitInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "DvbSdtSettings$SdtInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "DvbSubDestinationSettings$BackgroundOpacity": "Specifies the opacity of the background rectangle. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$FontOpacity": "Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.\nAll burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$FontResolution": "Font resolution in DPI (dots per inch); default is 96 dpi.\nAll burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$FontSize": "A positive integer indicates the exact font size in points. Set to 0 for automatic font size selection. All burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$OutlineSize": "Specifies font outline size in pixels. This option is not valid for source captions that are either 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$ShadowOpacity": "Specifies the opacity of the shadow. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$ShadowXOffset": "Specifies the horizontal offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match.", - "DvbSubDestinationSettings$ShadowYOffset": "Specifies the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match.", "DvbSubDestinationSettings$XPosition": "Specifies the horizontal position of the caption relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.", "DvbSubDestinationSettings$YPosition": "Specifies the vertical position of the caption relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output. This option is not valid for source captions that are STL, 608/embedded or teletext. These source settings are already pre-defined by the caption stream. All burn-in and DVB-Sub font settings must match.", - "DvbSubSourceSettings$Pid": "When using DVB-Sub with Burn-In or SMPTE-TT, use this PID for the source content. Unused for DVB-Sub passthrough. All DVB-Sub content is passed through, regardless of selectors.", - "DvbTdtSettings$TdtInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "Eac3Settings$Bitrate": "Average bitrate in bits/second. Valid bitrates depend on the coding mode.", - "Eac3Settings$Dialnorm": "Sets the dialnorm for the output. If blank and input audio is Dolby Digital Plus, dialnorm will be passed through.", - "Eac3Settings$SampleRate": "Sample rate in hz. Sample rate is always 48000.", - "EmbeddedSourceSettings$Source608ChannelNumber": "Specifies the 608/708 channel number within the video track from which to extract captions. Unused for passthrough.", - "EmbeddedSourceSettings$Source608TrackNumber": "Specifies the video track index used for extracting captions. The system only supports one input video track, so this should always be set to '1'.", - "FileSourceSettings$TimeDelta": "Specifies a time delta in seconds to offset the captions from the source file.", - "FrameCaptureSettings$FramerateDenominator": "Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.n.jpg where n is the 0-based sequence number of each Capture.", - "FrameCaptureSettings$FramerateNumerator": "Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.NNNNNNN.jpg where N is the 0-based frame sequence number zero padded to 7 decimal places.", - "FrameCaptureSettings$MaxCaptures": "Maximum number of captures (encoded jpg output files).", - "FrameCaptureSettings$Quality": "JPEG Quality - a higher value equals higher quality.", - "H264Settings$Bitrate": "Average bitrate in bits/second. Required for VBR, CBR, and ABR. Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.", - "H264Settings$FramerateDenominator": "When you use the API for transcode jobs that use framerate conversion, specify the framerate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use framerate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.", - "H264Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps.", "H264Settings$GopClosedCadence": "Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.", - "H264Settings$HrdBufferInitialFillPercentage": "Percentage of the buffer that should initially be filled (HRD buffer model).", - "H264Settings$HrdBufferSize": "Size of buffer (HRD buffer model). Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m.", - "H264Settings$MaxBitrate": "Maximum bitrate in bits/second (for VBR mode only). Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m.", - "H264Settings$MinIInterval": "Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. This setting is only used when Scene Change Detect is enabled. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1", - "H264Settings$NumberBFramesBetweenReferenceFrames": "Number of B-frames between reference frames.", - "H264Settings$NumberReferenceFrames": "Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.", - "H264Settings$ParDenominator": "Pixel Aspect Ratio denominator.", - "H264Settings$ParNumerator": "Pixel Aspect Ratio numerator.", - "H264Settings$Slices": "Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.", - "H264Settings$Softness": "Softness. Selects quantizer matrix, larger values reduce high-frequency content in the encoded image.", - "H265Settings$Bitrate": "Average bitrate in bits/second. Required for VBR, CBR, and ABR. Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.", - "H265Settings$FramerateDenominator": "Framerate denominator.", - "H265Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps.", "H265Settings$GopClosedCadence": "Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.", - "H265Settings$HrdBufferInitialFillPercentage": "Percentage of the buffer that should initially be filled (HRD buffer model).", - "H265Settings$HrdBufferSize": "Size of buffer (HRD buffer model). Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m.", - "H265Settings$MaxBitrate": "Maximum bitrate in bits/second (for VBR mode only). Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m.", - "H265Settings$MinIInterval": "Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. This setting is only used when Scene Change Detect is enabled. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1", - "H265Settings$NumberBFramesBetweenReferenceFrames": "Number of B-frames between reference frames.", - "H265Settings$NumberReferenceFrames": "Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.", - "H265Settings$ParDenominator": "Pixel Aspect Ratio denominator.", - "H265Settings$ParNumerator": "Pixel Aspect Ratio numerator.", - "H265Settings$Slices": "Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.", - "Hdr10Metadata$BluePrimaryX": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$BluePrimaryY": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$GreenPrimaryX": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$GreenPrimaryY": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$MaxContentLightLevel": "Maximum light level among all samples in the coded video sequence, in units of candelas per square meter.", - "Hdr10Metadata$MaxFrameAverageLightLevel": "Maximum average light level of any frame in the coded video sequence, in units of candelas per square meter.", "Hdr10Metadata$MaxLuminance": "Nominal maximum mastering display luminance in units of of 0.0001 candelas per square meter.", "Hdr10Metadata$MinLuminance": "Nominal minimum mastering display luminance in units of of 0.0001 candelas per square meter", - "Hdr10Metadata$RedPrimaryX": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$RedPrimaryY": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$WhitePointX": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "Hdr10Metadata$WhitePointY": "HDR Master Display Information comes from the color grader and the color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate.", - "HlsCaptionLanguageMapping$CaptionChannel": "Caption channel.", "HlsGroupSettings$MinSegmentLength": "When set, Minimum Segment Size is enforced by looking ahead and back within the specified range for a nearby avail and extending the segment size if needed.", - "HlsGroupSettings$ProgramDateTimePeriod": "Period of insertion of EXT-X-PROGRAM-DATE-TIME entry, in seconds.", + "M2tsSettings$AudioFramesPerPes": "The number of audio frames to insert for each PES packet.", + "M2tsSettings$Bitrate": "The output bitrate of the transport stream in bits per second. Setting to 0 lets the muxer automatically determine the appropriate bitrate. Other common values are 3750000, 7500000, and 15000000.", + "M3u8Settings$AudioFramesPerPes": "The number of audio frames to insert for each PES packet.", + "Mpeg2Settings$GopClosedCadence": "Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting." + } + }, + "__integerMin0Max255": { + "base": null, + "refs": { + "AudioDescription$AudioType": "Applies only if Follow Input Audio Type is unchecked (false). A number between 0 and 255. The following are defined in ISO-IEC 13818-1: 0 = Undefined, 1 = Clean Effects, 2 = Hearing Impaired, 3 = Visually Impaired Commentary, 4-255 = Reserved.", + "BurninDestinationSettings$BackgroundOpacity": "Specifies the opacity of the background rectangle. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", + "BurninDestinationSettings$FontOpacity": "Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.\nAll burn-in and DVB-Sub font settings must match.", + "BurninDestinationSettings$ShadowOpacity": "Specifies the opacity of the shadow. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$BackgroundOpacity": "Specifies the opacity of the background rectangle. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$FontOpacity": "Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.\nAll burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$ShadowOpacity": "Specifies the opacity of the shadow. 255 is opaque; 0 is transparent. Leaving this parameter blank is equivalent to setting it to 0 (transparent). All burn-in and DVB-Sub font settings must match." + } + }, + "__integerMin0Max3": { + "base": null, + "refs": { + "NoiseReducerFilterSettings$Strength": "Relative strength of noise reducing filter. Higher values produce stronger filtering.", + "NoiseReducerSpatialFilterSettings$PostFilterSharpenStrength": "Specify strength of post noise reduction sharpening filter, with 0 disabling the filter and 3 enabling it at maximum strength." + } + }, + "__integerMin0Max30": { + "base": null, + "refs": { + "H264Settings$MinIInterval": "Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. This setting is only used when Scene Change Detect is enabled. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1", + "H265Settings$MinIInterval": "Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. This setting is only used when Scene Change Detect is enabled. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1", + "Mpeg2Settings$MinIInterval": "Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. This setting is only used when Scene Change Detect is enabled. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1" + } + }, + "__integerMin0Max3600": { + "base": null, + "refs": { + "HlsGroupSettings$ProgramDateTimePeriod": "Period of insertion of EXT-X-PROGRAM-DATE-TIME entry, in seconds." + } + }, + "__integerMin0Max47185920": { + "base": null, + "refs": { + "Mpeg2Settings$HrdBufferSize": "Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000." + } + }, + "__integerMin0Max500": { + "base": null, + "refs": { + "M2tsSettings$MaxPcrInterval": "Maximum time in milliseconds between Program Clock References (PCRs) inserted into the transport stream." + } + }, + "__integerMin0Max50000": { + "base": null, + "refs": { + "Hdr10Metadata$BluePrimaryX": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$BluePrimaryY": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$GreenPrimaryX": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$GreenPrimaryY": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$RedPrimaryX": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$RedPrimaryY": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$WhitePointX": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.", + "Hdr10Metadata$WhitePointY": "HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction." + } + }, + "__integerMin0Max65535": { + "base": null, + "refs": { + "DvbNitSettings$NetworkId": "The numeric value placed in the Network Information Table (NIT).", + "Hdr10Metadata$MaxContentLightLevel": "Maximum light level among all samples in the coded video sequence, in units of candelas per square meter.", + "Hdr10Metadata$MaxFrameAverageLightLevel": "Maximum average light level of any frame in the coded video sequence, in units of candelas per square meter.", + "M2tsSettings$ProgramNumber": "The value of the program number field in the Program Map Table.", + "M2tsSettings$TransportStreamId": "The value of the transport stream ID field in the Program Map Table.", + "M3u8Settings$ProgramNumber": "The value of the program number field in the Program Map Table.", + "M3u8Settings$TransportStreamId": "The value of the transport stream ID field in the Program Map Table." + } + }, + "__integerMin0Max7": { + "base": null, + "refs": { + "H264Settings$NumberBFramesBetweenReferenceFrames": "Number of B-frames between reference frames.", + "H265Settings$NumberBFramesBetweenReferenceFrames": "Number of B-frames between reference frames.", + "Mpeg2Settings$NumberBFramesBetweenReferenceFrames": "Number of B-frames between reference frames." + } + }, + "__integerMin0Max8": { + "base": null, + "refs": { + "AudioSelector$ProgramSelection": "Use this setting for input streams that contain Dolby E, to have the service extract specific program data from the track. To select multiple programs, create multiple selectors with the same Track and different Program numbers. In the console, this setting is visible when you set Selector type to Track. Choose the program number from the dropdown list. If you are sending a JSON file, provide the program ID, which is part of the audio metadata. If your input file has incorrect metadata, you can choose All channels instead of a program number to have the service ignore the program IDs and include all the programs in the track." + } + }, + "__integerMin0Max9": { + "base": null, + "refs": { + "NielsenConfiguration$BreakoutCode": "Use Nielsen Configuration (NielsenConfiguration) to set the Nielsen measurement system breakout code. Supported values are 0, 3, 7, and 9." + } + }, + "__integerMin0Max96": { + "base": null, + "refs": { + "BurninDestinationSettings$FontSize": "A positive integer indicates the exact font size in points. Set to 0 for automatic font size selection. All burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$FontSize": "A positive integer indicates the exact font size in points. Set to 0 for automatic font size selection. All burn-in and DVB-Sub font settings must match." + } + }, + "__integerMin0Max99": { + "base": null, + "refs": { + "InsertableImage$Layer": "Use Layer (Layer) to specify how overlapping inserted images appear. Images with higher values of layer appear on top of images with lower values of layer." + } + }, + "__integerMin1000Max1152000000": { + "base": null, + "refs": { + "H264Settings$Bitrate": "Average bitrate in bits/second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.", + "H264Settings$MaxBitrate": "Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000." + } + }, + "__integerMin1000Max1466400000": { + "base": null, + "refs": { + "H265Settings$Bitrate": "Average bitrate in bits/second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.", + "H265Settings$MaxBitrate": "Maximum bitrate in bits/second." + } + }, + "__integerMin1000Max288000000": { + "base": null, + "refs": { + "Mpeg2Settings$Bitrate": "Average bitrate in bits/second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000." + } + }, + "__integerMin1000Max30000": { + "base": null, + "refs": { + "DvbTdtSettings$TdtInterval": "The number of milliseconds between instances of this table in the output transport stream." + } + }, + "__integerMin1000Max300000000": { + "base": null, + "refs": { + "Mpeg2Settings$MaxBitrate": "Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000." + } + }, + "__integerMin10Max48": { + "base": null, + "refs": { + "TimecodeBurnin$FontSize": "Use Font Size (FontSize) to set the font size of any burned-in timecode. Valid values are 10, 16, 32, 48." + } + }, + "__integerMin16Max24": { + "base": null, + "refs": { + "AiffSettings$BitDepth": "Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track.", + "WavSettings$BitDepth": "Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track." + } + }, + "__integerMin1Max1": { + "base": null, + "refs": { + "EmbeddedSourceSettings$Source608TrackNumber": "Specifies the video track index used for extracting captions. The system only supports one input video track, so this should always be set to '1'." + } + }, + "__integerMin1Max100": { + "base": null, + "refs": { + "ColorCorrector$Brightness": "Brightness level.", + "ColorCorrector$Contrast": "Contrast level.", + "ColorCorrector$Saturation": "Saturation level.", + "FrameCaptureSettings$Quality": "JPEG Quality - a higher value equals higher quality." + } + }, + "__integerMin1Max10000000": { + "base": null, + "refs": { + "FrameCaptureSettings$MaxCaptures": "Maximum number of captures (encoded jpg output files)." + } + }, + "__integerMin1Max1001": { + "base": null, + "refs": { + "Mpeg2Settings$FramerateDenominator": "Framerate denominator." + } + }, + "__integerMin1Max16": { + "base": null, + "refs": { + "RemixSettings$ChannelsIn": "Specify the number of audio channels from your input that you want to use in your output. With remixing, you might combine or split the data in these channels, so the number of channels in your final output might be different." + } + }, + "__integerMin1Max2": { + "base": null, + "refs": { + "AiffSettings$Channels": "Set Channels to specify the number of channels in this output audio track. Choosing Mono in the console will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2.", + "Mp2Settings$Channels": "Set Channels to specify the number of channels in this output audio track. Choosing Mono in the console will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2." + } + }, + "__integerMin1Max2147483647": { + "base": null, + "refs": { + "CmafGroupSettings$FragmentLength": "Length of fragments to generate (in seconds). Fragment length must be compatible with GOP size and Framerate. Note that fragments will end on the next keyframe after this number of seconds, so actual fragment length may be longer. When Emit Single File is checked, the fragmentation is internal to a single output file and it does not cause the creation of many output files as in other output types.", + "CmafGroupSettings$SegmentLength": "Use this setting to specify the length, in seconds, of each individual CMAF segment. This value applies to the whole package; that is, to every output in the output group. Note that segments end on the first keyframe after this number of seconds, so the actual segment length might be slightly longer. If you set Segment control (CmafSegmentControl) to single file, the service puts the content of each output in a single file that has metadata that marks these segments. If you set it to segmented files, the service creates multiple files for each output, each with the content of one segment.", + "DashIsoGroupSettings$FragmentLength": "Length of fragments to generate (in seconds). Fragment length must be compatible with GOP size and Framerate. Note that fragments will end on the next keyframe after this number of seconds, so actual fragment length may be longer. When Emit Single File is checked, the fragmentation is internal to a single output file and it does not cause the creation of many output files as in other output types.", + "DashIsoGroupSettings$SegmentLength": "Length of mpd segments to create (in seconds). Note that segments will end on the next keyframe after this number of seconds, so actual segment length may be longer. When Emit Single File is checked, the segmentation is internal to a single output file and it does not cause the creation of many output files as in other output types.", + "DvbSubSourceSettings$Pid": "When using DVB-Sub with Burn-In or SMPTE-TT, use this PID for the source content. Unused for DVB-Sub passthrough. All DVB-Sub content is passed through, regardless of selectors.", + "FrameCaptureSettings$FramerateDenominator": "Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.n.jpg where n is the 0-based sequence number of each Capture.", + "FrameCaptureSettings$FramerateNumerator": "Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.NNNNNNN.jpg where N is the 0-based frame sequence number zero padded to 7 decimal places.", + "H264Settings$FramerateDenominator": "When you use the API for transcode jobs that use framerate conversion, specify the framerate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use framerate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.", + "H264Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps.", + "H264Settings$ParDenominator": "Pixel Aspect Ratio denominator.", + "H264Settings$ParNumerator": "Pixel Aspect Ratio numerator.", + "H265Settings$FramerateDenominator": "Framerate denominator.", + "H265Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps.", + "H265Settings$ParDenominator": "Pixel Aspect Ratio denominator.", + "H265Settings$ParNumerator": "Pixel Aspect Ratio numerator.", "HlsGroupSettings$SegmentLength": "Length of MPEG-2 Transport Stream segments to create (in seconds). Note that segments will end on the next keyframe after this number of seconds, so actual segment length may be longer.", "HlsGroupSettings$SegmentsPerSubdirectory": "Number of segments to write to a subdirectory before starting a new one. directoryStructure must be SINGLE_DIRECTORY for this setting to have an effect.", + "Input$ProgramNumber": "Use Program (programNumber) to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default.", + "InputTemplate$ProgramNumber": "Use Program (programNumber) to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default.", + "Mpeg2Settings$ParDenominator": "Pixel Aspect Ratio denominator.", + "Mpeg2Settings$ParNumerator": "Pixel Aspect Ratio numerator.", + "MsSmoothGroupSettings$FragmentLength": "Use Fragment length (FragmentLength) to specify the mp4 fragment sizes in seconds. Fragment length must be compatible with GOP size and framerate.", + "ProresSettings$FramerateDenominator": "Framerate denominator.", + "ProresSettings$FramerateNumerator": "When you use the API for transcode jobs that use framerate conversion, specify the framerate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator.", + "ProresSettings$ParDenominator": "Pixel Aspect Ratio denominator.", + "ProresSettings$ParNumerator": "Pixel Aspect Ratio numerator.", + "VideoSelector$Pid": "Use PID (Pid) to select specific video data from an input file. Specify this value as an integer; the system automatically converts it to the hexidecimal value. For example, 257 selects PID 0x101. A PID, or packet identifier, is an identifier for a set of data in an MPEG-2 transport stream container.", + "__listOf__integerMin1Max2147483647$member": null + } + }, + "__integerMin1Max31": { + "base": null, + "refs": { + "Ac3Settings$Dialnorm": "Sets the dialnorm for the output. If blank and input audio is Dolby Digital, dialnorm will be passed through.", + "Eac3Settings$Dialnorm": "Sets the dialnorm for the output. If blank and input audio is Dolby Digital Plus, dialnorm will be passed through." + } + }, + "__integerMin1Max32": { + "base": null, + "refs": { + "H264Settings$Slices": "Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.", + "H265Settings$Slices": "Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures." + } + }, + "__integerMin1Max4": { + "base": null, + "refs": { + "AncillarySourceSettings$SourceAncillaryChannelNumber": "Specifies the 608 channel number in the ancillary data track from which to extract captions. Unused for passthrough.", + "EmbeddedSourceSettings$Source608ChannelNumber": "Specifies the 608/708 channel number within the video track from which to extract captions. Unused for passthrough." + } + }, + "__integerMin1Max6": { + "base": null, + "refs": { + "H264Settings$NumberReferenceFrames": "Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.", + "H265Settings$NumberReferenceFrames": "Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding." + } + }, + "__integerMin1Max8": { + "base": null, + "refs": { + "RemixSettings$ChannelsOut": "Specify the number of channels in this output after remixing. Valid values: 1, 2, 4, 6, 8", + "WavSettings$Channels": "Set Channels to specify the number of channels in this output audio track. With WAV, valid values 1, 2, 4, and 8. In the console, these values are Mono, Stereo, 4-Channel, and 8-Channel, respectively." + } + }, + "__integerMin24Max60000": { + "base": null, + "refs": { + "Mpeg2Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps." + } + }, + "__integerMin25Max10000": { + "base": null, + "refs": { + "DvbNitSettings$NitInterval": "The number of milliseconds between instances of this table in the output transport stream." + } + }, + "__integerMin25Max2000": { + "base": null, + "refs": { + "DvbSdtSettings$SdtInterval": "The number of milliseconds between instances of this table in the output transport stream." + } + }, + "__integerMin32000Max384000": { + "base": null, + "refs": { + "Mp2Settings$Bitrate": "Average bitrate in bits/second." + } + }, + "__integerMin32000Max48000": { + "base": null, + "refs": { + "Mp2Settings$SampleRate": "Sample rate in hz." + } + }, + "__integerMin32Max2160": { + "base": null, + "refs": { + "VideoDescription$Height": "Use the Height (Height) setting to define the video resolution height for this output. Specify in pixels. If you don't provide a value here, the service will use the input height." + } + }, + "__integerMin32Max4096": { + "base": null, + "refs": { + "VideoDescription$Width": "Use Width (Width) to define the video resolution width, in pixels, for this output. If you don't provide a value here, the service will use the input width." + } + }, + "__integerMin32Max8182": { + "base": null, + "refs": { + "M2tsSettings$DvbTeletextPid": "Packet Identifier (PID) for input source DVB Teletext data to this output.", + "M2tsSettings$PcrPid": "Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport stream. When no value is given, the encoder will assign the same value as the Video PID.", + "M2tsSettings$PmtPid": "Packet Identifier (PID) for the Program Map Table (PMT) in the transport stream.", + "M2tsSettings$PrivateMetadataPid": "Packet Identifier (PID) of the private metadata stream in the transport stream.", + "M2tsSettings$Scte35Pid": "Packet Identifier (PID) of the SCTE-35 stream in the transport stream.", + "M2tsSettings$TimedMetadataPid": "Packet Identifier (PID) of the timed metadata stream in the transport stream.", + "M2tsSettings$VideoPid": "Packet Identifier (PID) of the elementary video stream in the transport stream.", + "M3u8Settings$PcrPid": "Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport stream. When no value is given, the encoder will assign the same value as the Video PID.", + "M3u8Settings$PmtPid": "Packet Identifier (PID) for the Program Map Table (PMT) in the transport stream.", + "M3u8Settings$PrivateMetadataPid": "Packet Identifier (PID) of the private metadata stream in the transport stream.", + "M3u8Settings$Scte35Pid": "Packet Identifier (PID) of the SCTE-35 stream in the transport stream.", + "M3u8Settings$TimedMetadataPid": "Packet Identifier (PID) of the timed metadata stream in the transport stream.", + "M3u8Settings$VideoPid": "Packet Identifier (PID) of the elementary video stream in the transport stream.", + "__listOf__integerMin32Max8182$member": null + } + }, + "__integerMin48000Max48000": { + "base": null, + "refs": { + "Ac3Settings$SampleRate": "Sample rate in hz. Sample rate is always 48000.", + "Eac3Settings$SampleRate": "Sample rate in hz. Sample rate is always 48000." + } + }, + "__integerMin6000Max1024000": { + "base": null, + "refs": { + "AacSettings$Bitrate": "Average bitrate in bits/second. Defaults and valid values depend on rate control mode and profile." + } + }, + "__integerMin64000Max640000": { + "base": null, + "refs": { + "Ac3Settings$Bitrate": "Average bitrate in bits/second. Valid bitrates depend on the coding mode.", + "Eac3Settings$Bitrate": "Average bitrate in bits/second. Valid bitrates depend on the coding mode." + } + }, + "__integerMin8000Max192000": { + "base": null, + "refs": { + "AiffSettings$SampleRate": "Sample rate in hz.", + "WavSettings$SampleRate": "Sample rate in Hz." + } + }, + "__integerMin8000Max96000": { + "base": null, + "refs": { + "AacSettings$SampleRate": "Sample rate in Hz. Valid values depend on rate control mode and profile." + } + }, + "__integerMin96Max600": { + "base": null, + "refs": { + "BurninDestinationSettings$FontResolution": "Font resolution in DPI (dots per inch); default is 96 dpi.\nAll burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$FontResolution": "Font resolution in DPI (dots per inch); default is 96 dpi.\nAll burn-in and DVB-Sub font settings must match." + } + }, + "__integerMinNegative1000Max1000": { + "base": null, + "refs": { + "JobSettings$AdAvailOffset": "When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.", + "JobTemplateSettings$AdAvailOffset": "When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time." + } + }, + "__integerMinNegative180Max180": { + "base": null, + "refs": { + "ColorCorrector$Hue": "Hue in degrees." + } + }, + "__integerMinNegative2147483648Max2147483647": { + "base": null, + "refs": { + "AudioSelector$Offset": "Specifies a time delta in milliseconds to offset the audio from the input video.", + "BurninDestinationSettings$ShadowXOffset": "Specifies the horizontal offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match.", + "BurninDestinationSettings$ShadowYOffset": "Specifies the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$ShadowXOffset": "Specifies the horizontal offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left. All burn-in and DVB-Sub font settings must match.", + "DvbSubDestinationSettings$ShadowYOffset": "Specifies the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. All burn-in and DVB-Sub font settings must match.", + "FileSourceSettings$TimeDelta": "Specifies a time delta in seconds to offset the captions from the source file.", + "HlsCaptionLanguageMapping$CaptionChannel": "Caption channel.", "HlsGroupSettings$TimedMetadataId3Period": "Timed Metadata interval in seconds.", "HlsGroupSettings$TimestampDeltaMilliseconds": "Provides an extra millisecond delta offset to fine tune the timestamps.", - "Input$FilterStrength": "Use Filter strength (FilterStrength) to adjust the magnitude the input filter settings (Deblock and Denoise). The range is -5 to 5. Default is 0.", - "Input$ProgramNumber": "Use Program (programNumber) to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default.", - "InputTemplate$FilterStrength": "Use Filter strength (FilterStrength) to adjust the magnitude the input filter settings (Deblock and Denoise). The range is -5 to 5. Default is 0.", - "InputTemplate$ProgramNumber": "Use Program (programNumber) to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default.", "InsertableImage$Duration": "Use Duration (Duration) to set the time, in milliseconds, for the image to remain on the output video.", "InsertableImage$FadeIn": "Use Fade in (FadeIut) to set the length, in milliseconds, of the inserted image fade in. If you don't specify a value for Fade in, the image will appear abruptly at the Start time.", "InsertableImage$FadeOut": "Use Fade out (FadeOut) to set the length, in milliseconds, of the inserted image fade out. If you don't specify a value for Fade out, the image will disappear abruptly at the end of the inserted image duration.", "InsertableImage$Height": "Specify the Height (Height) of the inserted image. Use a value that is less than or equal to the video resolution height. Leave this setting blank to use the native height of the image.", "InsertableImage$ImageX": "Use Left (ImageX) to set the distance, in pixels, between the inserted image and the left edge of the frame. Required for BMP, PNG and TGA input.", "InsertableImage$ImageY": "Use Top (ImageY) to set the distance, in pixels, between the inserted image and the top edge of the video frame. Required for BMP, PNG and TGA input.", - "InsertableImage$Layer": "Use Layer (Layer) to specify how overlapping inserted images appear. Images with higher values of layer appear on top of images with lower values of layer.", - "InsertableImage$Opacity": "Use Opacity (Opacity) to specify how much of the underlying video shows through the inserted image. 0 is transparent and 100 is fully opaque. Default is 50.", "InsertableImage$Width": "Specify the Width (Width) of the inserted image. Use a value that is less than or equal to the video resolution width. Leave this setting blank to use the native width of the image.", - "Job$ErrorCode": "Error code for the job", - "JobSettings$AdAvailOffset": "When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.", - "JobTemplateSettings$AdAvailOffset": "When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.", - "ListJobTemplatesRequest$MaxResults": "Optional. Number of job templates, up to twenty, that will be returned at one time.", - "ListJobsRequest$MaxResults": "Optional. Number of jobs, up to twenty, that will be returned at one time.", - "ListOf__integer$member": null, - "ListPresetsRequest$MaxResults": "Optional. Number of presets, up to twenty, that will be returned at one time", - "ListQueuesRequest$MaxResults": "Optional. Number of queues, up to twenty, that will be returned at one time.", - "M2tsSettings$AudioFramesPerPes": "The number of audio frames to insert for each PES packet.", - "M2tsSettings$Bitrate": "The output bitrate of the transport stream in bits per second. Setting to 0 lets the muxer automatically determine the appropriate bitrate. Other common values are 3750000, 7500000, and 15000000.", - "M2tsSettings$DvbTeletextPid": "Packet Identifier (PID) for input source DVB Teletext data to this output.", - "M2tsSettings$MaxPcrInterval": "Maximum time in milliseconds between Program Clock References (PCRs) inserted into the transport stream.", - "M2tsSettings$MinEbpInterval": "When set, enforces that Encoder Boundary Points do not come within the specified time interval of each other by looking ahead at input video. If another EBP is going to come in within the specified time interval, the current EBP is not emitted, and the segment is \"stretched\" to the next marker. The lookahead value does not add latency to the system. The Live Event must be configured elsewhere to create sufficient latency to make the lookahead accurate.", - "M2tsSettings$PatInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "M2tsSettings$PcrPid": "Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport stream. When no value is given, the encoder will assign the same value as the Video PID.", - "M2tsSettings$PmtInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "M2tsSettings$PmtPid": "Packet Identifier (PID) for the Program Map Table (PMT) in the transport stream.", - "M2tsSettings$PrivateMetadataPid": "Packet Identifier (PID) of the private metadata stream in the transport stream.", - "M2tsSettings$ProgramNumber": "The value of the program number field in the Program Map Table.", - "M2tsSettings$Scte35Pid": "Packet Identifier (PID) of the SCTE-35 stream in the transport stream.", - "M2tsSettings$TimedMetadataPid": "Packet Identifier (PID) of the timed metadata stream in the transport stream.", - "M2tsSettings$TransportStreamId": "The value of the transport stream ID field in the Program Map Table.", - "M2tsSettings$VideoPid": "Packet Identifier (PID) of the elementary video stream in the transport stream.", - "M3u8Settings$AudioFramesPerPes": "The number of audio frames to insert for each PES packet.", - "M3u8Settings$PatInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "M3u8Settings$PcrPid": "Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport stream. When no value is given, the encoder will assign the same value as the Video PID.", - "M3u8Settings$PmtInterval": "The number of milliseconds between instances of this table in the output transport stream.", - "M3u8Settings$PmtPid": "Packet Identifier (PID) for the Program Map Table (PMT) in the transport stream.", - "M3u8Settings$PrivateMetadataPid": "Packet Identifier (PID) of the private metadata stream in the transport stream.", - "M3u8Settings$ProgramNumber": "The value of the program number field in the Program Map Table.", - "M3u8Settings$Scte35Pid": "Packet Identifier (PID) of the SCTE-35 stream in the transport stream.", - "M3u8Settings$TimedMetadataPid": "Packet Identifier (PID) of the timed metadata stream in the transport stream.", - "M3u8Settings$TransportStreamId": "The value of the transport stream ID field in the Program Map Table.", - "M3u8Settings$VideoPid": "Packet Identifier (PID) of the elementary video stream in the transport stream.", - "Mp2Settings$Bitrate": "Average bitrate in bits/second.", - "Mp2Settings$Channels": "Set Channels to specify the number of channels in this output audio track. Choosing Mono in the console will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2.", - "Mp2Settings$SampleRate": "Sample rate in hz.", - "Mpeg2Settings$Bitrate": "Average bitrate in bits/second. Required for VBR, CBR, and ABR. Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.", - "Mpeg2Settings$FramerateDenominator": "Framerate denominator.", - "Mpeg2Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps.", - "Mpeg2Settings$GopClosedCadence": "Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.", - "Mpeg2Settings$HrdBufferInitialFillPercentage": "Percentage of the buffer that should initially be filled (HRD buffer model).", - "Mpeg2Settings$HrdBufferSize": "Size of buffer (HRD buffer model). Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m.", - "Mpeg2Settings$MaxBitrate": "Maximum bitrate in bits/second (for VBR mode only). Five megabits can be entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 or 0.5m.", - "Mpeg2Settings$MinIInterval": "Enforces separation between repeated (cadence) I-frames and I-frames inserted by Scene Change Detection. If a scene change I-frame is within I-interval frames of a cadence I-frame, the GOP is shrunk and/or stretched to the scene change I-frame. GOP stretch requires enabling lookahead as well as setting I-interval. The normal cadence resumes for the next GOP. This setting is only used when Scene Change Detect is enabled. Note: Maximum GOP stretch = GOP size + Min-I-interval - 1", - "Mpeg2Settings$NumberBFramesBetweenReferenceFrames": "Number of B-frames between reference frames.", - "Mpeg2Settings$ParDenominator": "Pixel Aspect Ratio denominator.", - "Mpeg2Settings$ParNumerator": "Pixel Aspect Ratio numerator.", - "Mpeg2Settings$Softness": "Softness. Selects quantizer matrix, larger values reduce high-frequency content in the encoded image.", - "MsSmoothGroupSettings$FragmentLength": "Use Fragment length (FragmentLength) to specify the mp4 fragment sizes in seconds. Fragment length must be compatible with GOP size and framerate.", - "NielsenConfiguration$BreakoutCode": "Use Nielsen Configuration (NielsenConfiguration) to set the Nielsen measurement system breakout code. Supported values are 0, 3, 7, and 9.", - "NoiseReducerFilterSettings$Strength": "Relative strength of noise reducing filter. Higher values produce stronger filtering.", - "NoiseReducerSpatialFilterSettings$PostFilterSharpenStrength": "Specify strength of post noise reduction sharpening filter, with 0 disabling the filter and 3 enabling it at maximum strength.", - "NoiseReducerSpatialFilterSettings$Speed": "The speed of the filter, from -2 (lower speed) to 3 (higher speed), with 0 being the nominal value.", - "NoiseReducerSpatialFilterSettings$Strength": "Relative strength of noise reducing filter. Higher values produce stronger filtering.", - "OutputDetail$DurationInMs": "Duration in milliseconds", - "ProresSettings$FramerateDenominator": "Framerate denominator.", - "ProresSettings$FramerateNumerator": "When you use the API for transcode jobs that use framerate conversion, specify the framerate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator.", - "ProresSettings$ParDenominator": "Pixel Aspect Ratio denominator.", - "ProresSettings$ParNumerator": "Pixel Aspect Ratio numerator.", "Rectangle$Height": "Height of rectangle in pixels.", "Rectangle$Width": "Width of rectangle in pixels.", "Rectangle$X": "The distance, in pixels, between the rectangle and the left edge of the video frame.", "Rectangle$Y": "The distance, in pixels, between the rectangle and the top edge of the video frame.", - "RemixSettings$ChannelsIn": "Specify the number of audio channels from your input that you want to use in your output. With remixing, you might combine or split the data in these channels, so the number of channels in your final output might be different.", - "RemixSettings$ChannelsOut": "Specify the number of channels in this output after remixing. Valid values: 1, 2, 4, 6, 8", - "TimecodeBurnin$FontSize": "Use Font Size (FontSize) to set the font size of any burned-in timecode. Valid values are 10, 16, 32, 48.", - "VideoDescription$FixedAfd": "Applies only if you set AFD Signaling(AfdSignaling) to Fixed (FIXED). Use Fixed (FixedAfd) to specify a four-bit AFD value which the service will write on all frames of this video output.", - "VideoDescription$Height": "Use the Height (Height) setting to define the video resolution height for this output. Specify in pixels. If you don't provide a value here, the service will use the input height.", - "VideoDescription$Sharpness": "Use Sharpness (Sharpness)setting to specify the strength of anti-aliasing. This setting changes the width of the anti-alias filter kernel used for scaling. Sharpness only applies if your output resolution is different from your input resolution, and if you set Anti-alias (AntiAlias) to ENABLED. 0 is the softest setting, 100 the sharpest, and 50 recommended for most content.", - "VideoDescription$Width": "Use Width (Width) to define the video resolution width, in pixels, for this output. If you don't provide a value here, the service will use the input width.", - "VideoDetail$HeightInPx": "Height in pixels for the output", - "VideoDetail$WidthInPx": "Width in pixels for the output", - "VideoSelector$Pid": "Use PID (Pid) to select specific video data from an input file. Specify this value as an integer; the system automatically converts it to the hexidecimal value. For example, 257 selects PID 0x101. A PID, or packet identifier, is an identifier for a set of data in an MPEG-2 transport stream container.", - "VideoSelector$ProgramNumber": "Selects a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported.", - "WavSettings$BitDepth": "Specify Bit depth (BitDepth), in bits per sample, to choose the encoding quality for this audio track.", - "WavSettings$Channels": "Set Channels to specify the number of channels in this output audio track. With WAV, valid values 1, 2, 4, and 8. In the console, these values are Mono, Stereo, 4-Channel, and 8-Channel, respectively.", - "WavSettings$SampleRate": "Sample rate in Hz." + "VideoSelector$ProgramNumber": "Selects a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported." + } + }, + "__integerMinNegative2Max3": { + "base": null, + "refs": { + "NoiseReducerSpatialFilterSettings$Speed": "The speed of the filter, from -2 (lower speed) to 3 (higher speed), with 0 being the nominal value." + } + }, + "__integerMinNegative5Max5": { + "base": null, + "refs": { + "Input$FilterStrength": "Use Filter strength (FilterStrength) to adjust the magnitude the input filter settings (Deblock and Denoise). The range is -5 to 5. Default is 0.", + "InputTemplate$FilterStrength": "Use Filter strength (FilterStrength) to adjust the magnitude the input filter settings (Deblock and Denoise). The range is -5 to 5. Default is 0." + } + }, + "__integerMinNegative60Max6": { + "base": null, + "refs": { + "__listOf__integerMinNegative60Max6$member": null + } + }, + "__integerMinNegative70Max0": { + "base": null, + "refs": { + "AudioNormalizationSettings$CorrectionGateLevel": "Content measuring above this level will be corrected to the target level. Content measuring below this level will not be corrected. Gating only applies when not using real_time_correction." + } + }, + "__listOfAudioDescription": { + "base": null, + "refs": { + "Output$AudioDescriptions": "(AudioDescriptions) contains groups of audio encoding settings organized by audio codec. Include one instance of (AudioDescriptions) per output. (AudioDescriptions) can contain multiple groups of encoding settings.", + "PresetSettings$AudioDescriptions": "(AudioDescriptions) contains groups of audio encoding settings organized by audio codec. Include one instance of (AudioDescriptions) per output. (AudioDescriptions) can contain multiple groups of encoding settings." + } + }, + "__listOfCaptionDescription": { + "base": null, + "refs": { + "Output$CaptionDescriptions": "(CaptionDescriptions) contains groups of captions settings. For each output that has captions, include one instance of (CaptionDescriptions). (CaptionDescriptions) can contain multiple groups of captions settings." + } + }, + "__listOfCaptionDescriptionPreset": { + "base": null, + "refs": { + "PresetSettings$CaptionDescriptions": "Caption settings for this preset. There can be multiple caption settings in a single output." + } + }, + "__listOfEndpoint": { + "base": null, + "refs": { + "DescribeEndpointsResponse$Endpoints": "List of endpoints" + } + }, + "__listOfHlsAdMarkers": { + "base": null, + "refs": { + "HlsGroupSettings$AdMarkers": "Choose one or more ad marker types to pass SCTE35 signals through to this group of Apple HLS outputs." + } + }, + "__listOfHlsCaptionLanguageMapping": { + "base": null, + "refs": { + "HlsGroupSettings$CaptionLanguageMappings": "Language to be used on Caption outputs" + } + }, + "__listOfId3Insertion": { + "base": null, + "refs": { + "TimedMetadataInsertion$Id3Insertions": "Id3Insertions contains the array of Id3Insertion instances." + } + }, + "__listOfInput": { + "base": null, + "refs": { + "JobSettings$Inputs": "Use Inputs (inputs) to define source file used in the transcode job. There can be multiple inputs add in a job. These inputs will be concantenated together to create the output." + } + }, + "__listOfInputClipping": { + "base": null, + "refs": { + "Input$InputClippings": "(InputClippings) contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them.", + "InputTemplate$InputClippings": "(InputClippings) contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them." + } + }, + "__listOfInputTemplate": { + "base": null, + "refs": { + "JobTemplateSettings$Inputs": "Use Inputs (inputs) to define the source file used in the transcode job. There can only be one input in a job template. Using the API, you can include multiple inputs when referencing a job template." + } + }, + "__listOfInsertableImage": { + "base": null, + "refs": { + "ImageInserter$InsertableImages": "Image to insert. Must be 32 bit windows BMP, PNG, or TGA file. Must not be larger than the output frames." + } + }, + "__listOfJob": { + "base": null, + "refs": { + "ListJobsResponse$Jobs": "List of jobs" + } + }, + "__listOfJobTemplate": { + "base": null, + "refs": { + "ListJobTemplatesResponse$JobTemplates": "List of Job templates." + } + }, + "__listOfOutput": { + "base": null, + "refs": { + "OutputGroup$Outputs": "This object holds groups of encoding settings, one group of settings per output." + } + }, + "__listOfOutputChannelMapping": { + "base": null, + "refs": { + "ChannelMapping$OutputChannels": "List of output channels" + } + }, + "__listOfOutputDetail": { + "base": null, + "refs": { + "OutputGroupDetail$OutputDetails": "Details about the output" + } + }, + "__listOfOutputGroup": { + "base": null, + "refs": { + "JobSettings$OutputGroups": "(OutputGroups) contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in (OutputGroups) is a group of settings that apply to the whole group. This required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings", + "JobTemplateSettings$OutputGroups": "(OutputGroups) contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in (OutputGroups) is a group of settings that apply to the whole group. This required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings" + } + }, + "__listOfOutputGroupDetail": { + "base": null, + "refs": { + "Job$OutputGroupDetails": "List of output group details" + } + }, + "__listOfPreset": { + "base": null, + "refs": { + "ListPresetsResponse$Presets": "List of presets" + } + }, + "__listOfQueue": { + "base": null, + "refs": { + "ListQueuesResponse$Queues": "List of queues" + } + }, + "__listOf__integerMin1Max2147483647": { + "base": null, + "refs": { + "AudioSelector$Pids": "Selects a specific PID from within an audio source (e.g. 257 selects PID 0x101).", + "AudioSelector$Tracks": "Identify a track from the input audio to include in this selector by entering the track index number. To include several tracks in a single audio selector, specify multiple tracks as follows. Using the console, enter a comma-separated list. For examle, type \"1,2,3\" to include tracks 1 through 3. Specifying directly in your JSON job file, provide the track numbers in an array. For example, \"tracks\": [1,2,3]." + } + }, + "__listOf__integerMin32Max8182": { + "base": null, + "refs": { + "M2tsSettings$AudioPids": "Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation.", + "M2tsSettings$DvbSubPids": "Packet Identifier (PID) for input source DVB Subtitle data to this output. Multiple values are accepted, and can be entered in ranges and/or by comma separation.", + "M3u8Settings$AudioPids": "Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation." + } + }, + "__listOf__integerMinNegative60Max6": { + "base": null, + "refs": { + "OutputChannelMapping$InputChannels": "List of input channels" + } + }, + "__listOf__stringMin1": { + "base": null, + "refs": { + "AudioSelectorGroup$AudioSelectorNames": "Name of an Audio Selector within the same input to include in the group. Audio selector names are standardized, based on their order within the input (e.g., \"Audio Selector 1\"). The audio selector name parameter can be repeated to add any number of audio selectors to the group." + } + }, + "__listOf__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12": { + "base": null, + "refs": { + "SpekeKeyProvider$SystemIds": "Relates to SPEKE implementation. DRM system identifiers. DASH output groups support a max of two system ids. Other group types support one system id." + } + }, + "__mapOfAudioSelector": { + "base": null, + "refs": { + "Input$AudioSelectors": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use mutiple Audio selectors per input.", + "InputTemplate$AudioSelectors": "Use Audio selectors (AudioSelectors) to specify a track or set of tracks from the input that you will use in your outputs. You can use mutiple Audio selectors per input." + } + }, + "__mapOfAudioSelectorGroup": { + "base": null, + "refs": { + "Input$AudioSelectorGroups": "Specifies set of audio selectors within an input to combine. An input may have multiple audio selector groups. See \"Audio Selector Group\":#inputs-audio_selector_group for more information.", + "InputTemplate$AudioSelectorGroups": "Specifies set of audio selectors within an input to combine. An input may have multiple audio selector groups. See \"Audio Selector Group\":#inputs-audio_selector_group for more information." + } + }, + "__mapOfCaptionSelector": { + "base": null, + "refs": { + "Input$CaptionSelectors": "Use Captions selectors (CaptionSelectors) to specify the captions data from the input that you will use in your outputs. You can use mutiple captions selectors per input.", + "InputTemplate$CaptionSelectors": "Use Captions selectors (CaptionSelectors) to specify the captions data from the input that you will use in your outputs. You can use mutiple captions selectors per input." + } + }, + "__mapOf__string": { + "base": null, + "refs": { + "CreateJobRequest$UserMetadata": "User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs.", + "Job$UserMetadata": "User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs." } }, "__string": { "base": null, "refs": { "AudioDescription$AudioSourceName": "Specifies which audio data to use from each input. In the simplest case, specify an \"Audio Selector\":#inputs-audio_selector by name based on its order within each input. For example if you specify \"Audio Selector 3\", then the third audio selector will be used from each input. If an input does not have an \"Audio Selector 3\", then the audio selector marked as \"default\" in that input will be used. If there is no audio selector marked as \"default\", silence will be inserted for the duration of that input. Alternatively, an \"Audio Selector Group\":#inputs-audio_selector_group name may be specified, with similar default/silence behavior. If no audio_source_name is specified, then \"Audio Selector 1\" will be chosen automatically.", - "AudioDescription$StreamName": "Used for MS Smooth and Apple HLS outputs. Indicates the name displayed by the player (eg. English, or Director Commentary). Alphanumeric characters, spaces, and underscore are legal.", - "AudioSelector$ExternalAudioFileInput": "Specifies audio data from an external file source.", - "AvailBlanking$AvailBlankingImage": "Blanking image to be used. Leave empty for solid black. Only bmp and png images are supported.", "CancelJobRequest$Id": "The Job ID of the job to be cancelled.", - "CaptionDescription$CaptionSelectorName": "Specifies which \"Caption Selector\":#inputs-caption_selector to use from each input when generating captions. The name should be of the format \"Caption Selector \", which denotes that the Nth Caption Selector will be used from each input.", "CaptionDescription$LanguageDescription": "Human readable information to indicate captions available for players (eg. English, or Spanish). Alphanumeric characters, spaces, and underscore are legal.", "CaptionDescriptionPreset$LanguageDescription": "Human readable information to indicate captions available for players (eg. English, or Spanish). Alphanumeric characters, spaces, and underscore are legal.", + "CmafGroupSettings$BaseUrl": "A partial URI prefix that will be put in the manifest file at the top level BaseURL element. Can be used if streams are delivered from a different URL than the manifest file.", "CreateJobRequest$ClientRequestToken": "Idempotency token for CreateJob operation.", "CreateJobRequest$JobTemplate": "When you create a job, you can either specify a job template or specify the transcoding settings individually", "CreateJobRequest$Queue": "Optional. When you create a job, you can specify a queue to send it to. If you don't specify, the job will go to the default queue. For more about queues, see the User Guide topic at http://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html.", @@ -2459,37 +2897,22 @@ "CreateQueueRequest$Description": "Optional. A description of the queue you are creating.", "CreateQueueRequest$Name": "The name of the queue you are creating.", "DashIsoGroupSettings$BaseUrl": "A partial URI prefix that will be put in the manifest (.mpd) file at the top level BaseURL element. Can be used if streams are delivered from a different URL than the manifest file.", - "DashIsoGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", "DeleteJobTemplateRequest$Name": "The name of the job template to be deleted.", "DeletePresetRequest$Name": "The name of the preset to be deleted.", "DeleteQueueRequest$Name": "The name of the queue to be deleted.", "DescribeEndpointsRequest$NextToken": "Use this string, provided with the response to a previous request, to request the next batch of endpoints.", "DescribeEndpointsResponse$NextToken": "Use this string to request the next batch of endpoints.", - "DvbNitSettings$NetworkName": "The network name text placed in the network_name_descriptor inside the Network Information Table. Maximum length is 256 characters.", - "DvbSdtSettings$ServiceName": "The service name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.", - "DvbSdtSettings$ServiceProviderName": "The service provider name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.", "Endpoint$Url": "URL of endpoint", "ExceptionBody$Message": null, - "FileGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", - "FileSourceSettings$SourceFile": "External caption file used for loading captions. Accepted file extensions are 'scc', 'ttml', 'dfxp', 'stl', 'srt', and 'smi'.", "GetJobRequest$Id": "the job ID of the job.", "GetJobTemplateRequest$Name": "The name of the job template.", "GetPresetRequest$Name": "The name of the preset.", "GetQueueRequest$Name": "The name of the queue.", "HlsCaptionLanguageMapping$LanguageDescription": "Caption language description.", - "HlsEncryptionSettings$ConstantInitializationVector": "This is a 128-bit, 16-byte hex value represented by a 32-character text string. If this parameter is not set then the Initialization Vector will follow the segment number by default.", "HlsGroupSettings$BaseUrl": "A partial URI prefix that will be prepended to each output in the media .m3u8 file. Can be used if base manifest is delivered from a different URL than the main .m3u8 file.", - "HlsGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", "HlsSettings$AudioGroupId": "Specifies the group to which the audio Rendition belongs.", "HlsSettings$AudioRenditionSets": "List all the audio groups that are used with the video output stream. Input all the audio GROUP-IDs that are associated to the video, separate by ','.", "HlsSettings$SegmentModifier": "String concatenated to end of segment filenames. Accepts \"Format Identifiers\":#format_identifier_parameters.", - "Id3Insertion$Id3": "Use ID3 tag (Id3) to provide a tag value in base64-encode format.", - "Id3Insertion$Timecode": "Provide a Timecode (TimeCode) in HH:MM:SS:FF or HH:MM:SS;FF format.", - "Input$FileInput": "Use Input (fileInput) to define the source file used in the transcode job. There can be multiple inputs in a job. These inputs are concantenated, in the order they are specified in the job, to create the output.", - "InputClipping$EndTimecode": "Set End timecode (EndTimecode) to the end of the portion of the input you are clipping. The frame corresponding to the End timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. When choosing this value, take into account your setting for Input timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to begin five minutes into the video, use 01:00:05:00.", - "InputClipping$StartTimecode": "Set Start timecode (StartTimecode) to the beginning of the portion of the input you are clipping. The frame corresponding to the Start timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. When choosing this value, take into account your setting for Input timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to begin five minutes into the video, use 01:00:05:00.", - "InsertableImage$ImageInserterInput": "Use Image location (imageInserterInput) to specify the Amazon S3 location of the image to be inserted into the output. Use a 32 bit BMP, PNG, or TGA file that fits inside the video frame.", - "InsertableImage$StartTime": "Use Start time (StartTime) to specify the video timecode when the image is inserted in the output. This must be in timecode format (HH:MM:SS:FF)", "Job$Arn": "An identifier for this resource that is unique within all of AWS.", "Job$ErrorMessage": "Error message of Job", "Job$Id": "A portion of the job's ARN, unique within your AWS Elemental MediaConvert resources", @@ -2507,19 +2930,14 @@ "ListJobsRequest$NextToken": "Use this string, provided with the response to a previous request, to request the next batch of jobs.", "ListJobsRequest$Queue": "Provide a queue name to get back only jobs from that queue.", "ListJobsResponse$NextToken": "Use this string to request the next batch of jobs.", - "ListOf__string$member": null, "ListPresetsRequest$Category": "Optionally, specify a preset category to limit responses to only presets from that category.", "ListPresetsRequest$NextToken": "Use this string, provided with the response to a previous request, to request the next batch of presets.", "ListPresetsResponse$NextToken": "Use this string to request the next batch of presets.", "ListQueuesRequest$NextToken": "Use this string, provided with the response to a previous request, to request the next batch of queues.", "ListQueuesResponse$NextToken": "Use this string to request the next batch of queues.", - "MapOf__string$member": null, "Mp4Settings$Mp4MajorBrand": "Overrides the \"Major Brand\" field in the output file. Usually not necessary to specify.", - "MsSmoothGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", "NielsenConfiguration$DistributorId": "Use Distributor ID (DistributorID) to specify the distributor ID that is assigned to your organization by Neilsen.", "Output$Extension": "Use Extension (Extension) to specify the file extension for outputs in File output groups. If you do not specify a value, the service will use default extensions by container type as follows * MPEG-2 transport stream, m2ts * Quicktime, mov * MXF container, mxf * MPEG-4 container, mp4 * No Container, the service will use codec extensions (e.g. AAC, H265, H265, AC3)", - "Output$NameModifier": "Use Name modifier (NameModifier) to have the service add a string to the end of each output filename. You specify the base filename as part of your destination URI. When you create multiple outputs in the same output group, Name modifier (NameModifier) is required. Name modifier also accepts format identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ or $Time$ in one output, you must use them in the same way in all outputs of the output group.", - "Output$Preset": "Use Preset (Preset) to specifiy a preset for your transcoding settings. Provide the system or custom preset name. You can specify either Preset (Preset) or Container settings (ContainerSettings), but not both.", "OutputGroup$CustomName": "Use Custom Group Name (CustomName) to specify a name for the output group. This value is displayed on the console and can make your job settings JSON more human-readable. It does not affect your outputs. Use up to twelve characters that are either letters, numbers, spaces, or underscores.", "OutputGroup$Name": "Name of the output group", "Preset$Arn": "An identifier for this resource that is unique within all of AWS.", @@ -2530,17 +2948,7 @@ "Queue$Description": "An optional description you create for each queue.", "Queue$Name": "A name you create for each queue. Each name must be unique within your account.", "SpekeKeyProvider$ResourceId": "The SPEKE-compliant server uses Resource ID (ResourceId) to identify content.", - "SpekeKeyProvider$Url": "Use URL (Url) to specify the SPEKE-compliant server that will provide keys for content.", - "StaticKeyProvider$KeyFormat": "Relates to DRM implementation. Sets the value of the KEYFORMAT attribute. Must be 'identity' or a reverse DNS string. May be omitted to indicate an implicit value of 'identity'.", - "StaticKeyProvider$KeyFormatVersions": "Relates to DRM implementation. Either a single positive integer version value or a slash delimited list of version values (1/2/3).", - "StaticKeyProvider$StaticKeyValue": "Relates to DRM implementation. Use a 32-character hexidecimal string to specify Key Value (StaticKeyValue).", "StaticKeyProvider$Url": "Relates to DRM implementation. The location of the license server used for protecting content.", - "TeletextDestinationSettings$PageNumber": "Set pageNumber to the Teletext page number for the destination captions for this output. This value must be a three-digit hexadecimal string; strings ending in -FF are invalid. If you are passing through the entire set of Teletext data, do not use this field.", - "TeletextSourceSettings$PageNumber": "Use Page Number (PageNumber) to specify the three-digit hexadecimal page number that will be used for Teletext captions. Do not use this setting if you are passing through teletext from the input source to output.", - "TimecodeBurnin$Prefix": "Use Prefix (Prefix) to place ASCII characters before any burned-in timecode. For example, a prefix of \"EZ-\" will result in the timecode \"EZ-00:00:00:00\". Provide either the characters themselves or the ASCII code equivalents. The supported range of characters is 0x20 through 0x7e. This includes letters, numbers, and all special characters represented on a standard English keyboard.", - "TimecodeConfig$Anchor": "If you use an editing platform that relies on an anchor timecode, use Anchor Timecode (Anchor) to specify a timecode that will match the input video frame to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). This setting ignores framerate conversion. System behavior for Anchor Timecode varies depending on your setting for Timecode source (TimecodeSource). * If Timecode source (TimecodeSource) is set to Specified Start (specifiedstart), the first input frame is the specified value in Start Timecode (Start). Anchor Timecode (Anchor) and Start Timecode (Start) are used calculate output timecode. * If Timecode source (TimecodeSource) is set to Start at 0 (zerobased) the first frame is 00:00:00:00. * If Timecode source (TimecodeSource) is set to Embedded (embedded), the first frame is the timecode value on the first input frame of the input.", - "TimecodeConfig$Start": "Only use when you set Timecode Source (TimecodeSource) to Specified Start (SPECIFIEDSTART). Use Start timecode (Start) to specify the timecode for the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF).", - "TimecodeConfig$TimestampOffset": "Only applies to outputs that support program-date-time stamp. Use Time stamp offset (TimestampOffset) to overwrite the timecode date without affecting the time and frame number. Provide the new date as a string in the format \"yyyy-mm-dd\". To use Time stamp offset, you must also enable Insert program-date-time (InsertProgramDateTime) in the output settings.", "UpdateJobTemplateRequest$Category": "The new category for the job template, if you are changing it.", "UpdateJobTemplateRequest$Description": "The new description for the job template, if you are changing it.", "UpdateJobTemplateRequest$Name": "The name of the job template you are modifying", @@ -2549,10 +2957,168 @@ "UpdatePresetRequest$Description": "The new description for the preset, if you are changing it.", "UpdatePresetRequest$Name": "The name of the preset you are modifying.", "UpdateQueueRequest$Description": "The new description for the queue, if you are changing it.", - "UpdateQueueRequest$Name": "The name of the queue you are modifying." + "UpdateQueueRequest$Name": "The name of the queue you are modifying.", + "__mapOf__string$member": null } }, - "__timestamp": { + "__stringMin0": { + "base": null, + "refs": { + "Output$Preset": "Use Preset (Preset) to specifiy a preset for your transcoding settings. Provide the system or custom preset name. You can specify either Preset (Preset) or Container settings (ContainerSettings), but not both." + } + }, + "__stringMin1": { + "base": null, + "refs": { + "CaptionDescription$CaptionSelectorName": "Specifies which \"Caption Selector\":#inputs-caption_selector to use from each input when generating captions. The name should be of the format \"Caption Selector \", which denotes that the Nth Caption Selector will be used from each input.", + "Output$NameModifier": "Use Name modifier (NameModifier) to have the service add a string to the end of each output filename. You specify the base filename as part of your destination URI. When you create multiple outputs in the same output group, Name modifier (NameModifier) is required. Name modifier also accepts format identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ or $Time$ in one output, you must use them in the same way in all outputs of the output group.", + "__listOf__stringMin1$member": null + } + }, + "__stringMin14PatternS3BmpBMPPngPNG": { + "base": null, + "refs": { + "AvailBlanking$AvailBlankingImage": "Blanking image to be used. Leave empty for solid black. Only bmp and png images are supported." + } + }, + "__stringMin14PatternS3BmpBMPPngPNGTgaTGA": { + "base": null, + "refs": { + "InsertableImage$ImageInserterInput": "Use Image location (imageInserterInput) to specify the Amazon S3 location of the image to be inserted into the output. Use a 32 bit BMP, PNG, or TGA file that fits inside the video frame." + } + }, + "__stringMin14PatternS3SccSCCTtmlTTMLDfxpDFXPStlSTLSrtSRTSmiSMI": { + "base": null, + "refs": { + "FileSourceSettings$SourceFile": "External caption file used for loading captions. Accepted file extensions are 'scc', 'ttml', 'dfxp', 'stl', 'srt', and 'smi'." + } + }, + "__stringMin1Max256": { + "base": null, + "refs": { + "DvbNitSettings$NetworkName": "The network name text placed in the network_name_descriptor inside the Network Information Table. Maximum length is 256 characters.", + "DvbSdtSettings$ServiceName": "The service name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.", + "DvbSdtSettings$ServiceProviderName": "The service provider name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters." + } + }, + "__stringMin32Max32Pattern09aFAF32": { + "base": null, + "refs": { + "CmafEncryptionSettings$ConstantInitializationVector": "This is a 128-bit, 16-byte hex value represented by a 32-character text string. If this parameter is not set then the Initialization Vector will follow the segment number by default.", + "HlsEncryptionSettings$ConstantInitializationVector": "This is a 128-bit, 16-byte hex value represented by a 32-character text string. If this parameter is not set then the Initialization Vector will follow the segment number by default." + } + }, + "__stringMin3Max3Pattern1809aFAF09aEAE": { + "base": null, + "refs": { + "TeletextDestinationSettings$PageNumber": "Set pageNumber to the Teletext page number for the destination captions for this output. This value must be a three-digit hexadecimal string; strings ending in -FF are invalid. If you are passing through the entire set of Teletext data, do not use this field.", + "TeletextSourceSettings$PageNumber": "Use Page Number (PageNumber) to specify the three-digit hexadecimal page number that will be used for Teletext captions. Do not use this setting if you are passing through teletext from the input source to output." + } + }, + "__stringMin3Max3PatternAZaZ3": { + "base": null, + "refs": { + "AudioDescription$CustomLanguageCode": "Specify the language for this audio output track, using the ISO 639-2 or ISO 639-3 three-letter language code. The language specified will be used when 'Follow Input Language Code' is not selected or when 'Follow Input Language Code' is selected but there is no ISO 639 language code specified by the input.", + "AudioSelector$CustomLanguageCode": "Selects a specific language code from within an audio source, using the ISO 639-2 or ISO 639-3 three-letter language code", + "CaptionDescription$CustomLanguageCode": "Indicates the language of the caption output track, using the ISO 639-2 or ISO 639-3 three-letter language code", + "CaptionDescriptionPreset$CustomLanguageCode": "Indicates the language of the caption output track, using the ISO 639-2 or ISO 639-3 three-letter language code", + "CaptionSelector$CustomLanguageCode": "The specific language to extract from source, using the ISO 639-2 or ISO 639-3 three-letter language code. If input is SCTE-27, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub and output is Burn-in or SMPTE-TT, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub that is being passed through, omit this field (and PID field); there is no way to extract a specific language with pass-through captions.", + "HlsCaptionLanguageMapping$CustomLanguageCode": "Specify the language for this caption channel, using the ISO 639-2 or ISO 639-3 three-letter language code" + } + }, + "__stringPattern": { + "base": null, + "refs": { + "TimecodeBurnin$Prefix": "Use Prefix (Prefix) to place ASCII characters before any burned-in timecode. For example, a prefix of \"EZ-\" will result in the timecode \"EZ-00:00:00:00\". Provide either the characters themselves or the ASCII code equivalents. The supported range of characters is 0x20 through 0x7e. This includes letters, numbers, and all special characters represented on a standard English keyboard." + } + }, + "__stringPattern010920405090509092": { + "base": null, + "refs": { + "Id3Insertion$Timecode": "Provide a Timecode (TimeCode) in HH:MM:SS:FF or HH:MM:SS;FF format.", + "InputClipping$EndTimecode": "Set End timecode (EndTimecode) to the end of the portion of the input you are clipping. The frame corresponding to the End timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for timecode source under input settings (InputTimecodeSource). For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to end six minutes into the video, use 01:06:00:00.", + "InputClipping$StartTimecode": "Set Start timecode (StartTimecode) to the beginning of the portion of the input you are clipping. The frame corresponding to the Start timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for Input timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to begin five minutes into the video, use 01:05:00:00.", + "TimecodeConfig$Anchor": "If you use an editing platform that relies on an anchor timecode, use Anchor Timecode (Anchor) to specify a timecode that will match the input video frame to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). This setting ignores framerate conversion. System behavior for Anchor Timecode varies depending on your setting for Source (TimecodeSource). * If Source (TimecodeSource) is set to Specified Start (SPECIFIEDSTART), the first input frame is the specified value in Start Timecode (Start). Anchor Timecode (Anchor) and Start Timecode (Start) are used calculate output timecode. * If Source (TimecodeSource) is set to Start at 0 (ZEROBASED) the first frame is 00:00:00:00. * If Source (TimecodeSource) is set to Embedded (EMBEDDED), the first frame is the timecode value on the first input frame of the input.", + "TimecodeConfig$Start": "Only use when you set Source (TimecodeSource) to Specified start (SPECIFIEDSTART). Use Start timecode (Start) to specify the timecode for the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF)." + } + }, + "__stringPattern01D20305D205D": { + "base": null, + "refs": { + "InsertableImage$StartTime": "Use Start time (StartTime) to specify the video timecode when the image is inserted in the output. This must be in timecode (HH:MM:SS:FF or HH:MM:SS;FF) format." + } + }, + "__stringPattern0940191020191209301": { + "base": null, + "refs": { + "TimecodeConfig$TimestampOffset": "Only applies to outputs that support program-date-time stamp. Use Timestamp offset (TimestampOffset) to overwrite the timecode date without affecting the time and frame number. Provide the new date as a string in the format \"yyyy-mm-dd\". To use Time stamp offset, you must also enable Insert program-date-time (InsertProgramDateTime) in the output settings. For example, if the date part of your timecodes is 2002-1-25 and you want to change it to one year later, set Timestamp offset (TimestampOffset) to 2003-1-25." + } + }, + "__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12": { + "base": null, + "refs": { + "__listOf__stringPattern09aFAF809aFAF409aFAF409aFAF409aFAF12$member": null + } + }, + "__stringPatternAZaZ0902": { + "base": null, + "refs": { + "Id3Insertion$Id3": "Use ID3 tag (Id3) to provide a tag value in base64-encode format." + } + }, + "__stringPatternAZaZ0932": { + "base": null, + "refs": { + "StaticKeyProvider$StaticKeyValue": "Relates to DRM implementation. Use a 32-character hexidecimal string to specify Key Value (StaticKeyValue)." + } + }, + "__stringPatternDD": { + "base": null, + "refs": { + "StaticKeyProvider$KeyFormatVersions": "Relates to DRM implementation. Either a single positive integer version value or a slash delimited list of version values (1/2/3)." + } + }, + "__stringPatternHttps": { + "base": null, + "refs": { + "SpekeKeyProvider$Url": "Use URL (Url) to specify the SPEKE-compliant server that will provide keys for content." + } + }, + "__stringPatternIdentityAZaZ26AZaZ09163": { + "base": null, + "refs": { + "StaticKeyProvider$KeyFormat": "Relates to DRM implementation. Sets the value of the KEYFORMAT attribute. Must be 'identity' or a reverse DNS string. May be omitted to indicate an implicit value of 'identity'." + } + }, + "__stringPatternS3": { + "base": null, + "refs": { + "CmafGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", + "DashIsoGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", + "FileGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", + "HlsGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.", + "MsSmoothGroupSettings$Destination": "Use Destination (Destination) to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file." + } + }, + "__stringPatternS3MM2VVMMPPEEGGAAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MM": { + "base": null, + "refs": { + "Input$FileInput": "Use Input (fileInput) to define the source file used in the transcode job. There can be multiple inputs in a job. These inputs are concantenated, in the order they are specified in the job, to create the output." + } + }, + "__stringPatternS3MM2VVMMPPEEGGAAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMOOVVMMTTSSMM2TTWWMMVVAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8LLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MMAAAACCAAIIFFFFMMPP2AACC3EECC3DDTTSSEE": { + "base": null, + "refs": { + "AudioSelector$ExternalAudioFileInput": "Specifies audio data from an external file source." + } + }, + "__stringPatternWS": { + "base": null, + "refs": { + "AudioDescription$StreamName": "Used for MS Smooth and Apple HLS outputs. Indicates the name displayed by the player (eg. English, or Director Commentary). Alphanumeric characters, spaces, and underscore are legal." + } + }, + "__timestampIso8601": { "base": null, "refs": { "Job$CreatedAt": "The time, in Unix epoch format in seconds, when the job got created.", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/api-2.json index 41f388cb0..f7e1aac27 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/api-2.json @@ -27,10 +27,10 @@ }, "errors": [ { - "shape": "UnprocessableEntityException" + "shape": "BadRequestException" }, { - "shape": "BadRequestException" + "shape": "UnprocessableEntityException" }, { "shape": "InternalServerErrorException" @@ -237,6 +237,46 @@ } ] }, + "DeleteReservation": { + "name": "DeleteReservation", + "http": { + "method": "DELETE", + "requestUri": "/prod/reservations/{reservationId}", + "responseCode": 200 + }, + "input": { + "shape": "DeleteReservationRequest" + }, + "output": { + "shape": "DeleteReservationResponse" + }, + "errors": [ + { + "shape": "BadRequestException" + }, + { + "shape": "InternalServerErrorException" + }, + { + "shape": "ForbiddenException" + }, + { + "shape": "BadGatewayException" + }, + { + "shape": "NotFoundException" + }, + { + "shape": "GatewayTimeoutException" + }, + { + "shape": "TooManyRequestsException" + }, + { + "shape": "ConflictException" + } + ] + }, "DescribeChannel": { "name": "DescribeChannel", "http": { @@ -348,6 +388,80 @@ } ] }, + "DescribeOffering": { + "name": "DescribeOffering", + "http": { + "method": "GET", + "requestUri": "/prod/offerings/{offeringId}", + "responseCode": 200 + }, + "input": { + "shape": "DescribeOfferingRequest" + }, + "output": { + "shape": "DescribeOfferingResponse" + }, + "errors": [ + { + "shape": "BadRequestException" + }, + { + "shape": "InternalServerErrorException" + }, + { + "shape": "ForbiddenException" + }, + { + "shape": "BadGatewayException" + }, + { + "shape": "NotFoundException" + }, + { + "shape": "GatewayTimeoutException" + }, + { + "shape": "TooManyRequestsException" + } + ] + }, + "DescribeReservation": { + "name": "DescribeReservation", + "http": { + "method": "GET", + "requestUri": "/prod/reservations/{reservationId}", + "responseCode": 200 + }, + "input": { + "shape": "DescribeReservationRequest" + }, + "output": { + "shape": "DescribeReservationResponse" + }, + "errors": [ + { + "shape": "BadRequestException" + }, + { + "shape": "InternalServerErrorException" + }, + { + "shape": "ForbiddenException" + }, + { + "shape": "BadGatewayException" + }, + { + "shape": "NotFoundException" + }, + { + "shape": "GatewayTimeoutException" + }, + { + "shape": "TooManyRequestsException" + } + ] + }, "ListChannels": { "name": "ListChannels", "http": { @@ -450,6 +564,114 @@ } ] }, + "ListOfferings": { + "name": "ListOfferings", + "http": { + "method": "GET", + "requestUri": "/prod/offerings", + "responseCode": 200 + }, + "input": { + "shape": "ListOfferingsRequest" + }, + "output": { + "shape": "ListOfferingsResponse" + }, + "errors": [ + { + "shape": "BadRequestException" + }, + { + "shape": "InternalServerErrorException" + }, + { + "shape": "ForbiddenException" + }, + { + "shape": "BadGatewayException" + }, + { + "shape": "GatewayTimeoutException" + }, + { + "shape": "TooManyRequestsException" + } + ] + }, + "ListReservations": { + "name": "ListReservations", + "http": { + "method": "GET", + "requestUri": "/prod/reservations", + "responseCode": 200 + }, + "input": { + "shape": "ListReservationsRequest" + }, + "output": { + "shape": "ListReservationsResponse" + }, + "errors": [ + { + "shape": "BadRequestException" + }, + { + "shape": "InternalServerErrorException" + }, + { + "shape": "ForbiddenException" + }, + { + "shape": "BadGatewayException" + }, + { + "shape": "GatewayTimeoutException" + }, + { + "shape": "TooManyRequestsException" + } + ] + }, + "PurchaseOffering": { + "name": "PurchaseOffering", + "http": { + "method": "POST", + "requestUri": "/prod/offerings/{offeringId}/purchase", + "responseCode": 201 + }, + "input": { + "shape": "PurchaseOfferingRequest" + }, + "output": { + "shape": "PurchaseOfferingResponse" + }, + "errors": [ + { + "shape": "BadRequestException" + }, + { + "shape": "InternalServerErrorException" + }, + { + "shape": "ForbiddenException" + }, + { + "shape": "BadGatewayException" + }, + { + "shape": "NotFoundException" + }, + { + "shape": "GatewayTimeoutException" + }, + { + "shape": "TooManyRequestsException" + }, + { + "shape": "ConflictException" + } + ] + }, "StartChannel": { "name": "StartChannel", "http": { @@ -545,10 +767,10 @@ }, "errors": [ { - "shape": "UnprocessableEntityException" + "shape": "BadRequestException" }, { - "shape": "BadRequestException" + "shape": "UnprocessableEntityException" }, { "shape": "InternalServerErrorException" @@ -1545,6 +1767,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -1626,6 +1852,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -1676,6 +1906,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -1715,6 +1949,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -1902,6 +2140,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -1956,6 +2198,92 @@ "members": { } }, + "DeleteReservationRequest": { + "type": "structure", + "members": { + "ReservationId": { + "shape": "__string", + "location": "uri", + "locationName": "reservationId" + } + }, + "required": [ + "ReservationId" + ] + }, + "DeleteReservationResponse": { + "type": "structure", + "members": { + "Arn": { + "shape": "__string", + "locationName": "arn" + }, + "Count": { + "shape": "__integer", + "locationName": "count" + }, + "CurrencyCode": { + "shape": "__string", + "locationName": "currencyCode" + }, + "Duration": { + "shape": "__integer", + "locationName": "duration" + }, + "DurationUnits": { + "shape": "OfferingDurationUnits", + "locationName": "durationUnits" + }, + "End": { + "shape": "__string", + "locationName": "end" + }, + "FixedPrice": { + "shape": "__double", + "locationName": "fixedPrice" + }, + "Name": { + "shape": "__string", + "locationName": "name" + }, + "OfferingDescription": { + "shape": "__string", + "locationName": "offeringDescription" + }, + "OfferingId": { + "shape": "__string", + "locationName": "offeringId" + }, + "OfferingType": { + "shape": "OfferingType", + "locationName": "offeringType" + }, + "Region": { + "shape": "__string", + "locationName": "region" + }, + "ReservationId": { + "shape": "__string", + "locationName": "reservationId" + }, + "ResourceSpecification": { + "shape": "ReservationResourceSpecification", + "locationName": "resourceSpecification" + }, + "Start": { + "shape": "__string", + "locationName": "start" + }, + "State": { + "shape": "ReservationState", + "locationName": "state" + }, + "UsagePrice": { + "shape": "__double", + "locationName": "usagePrice" + } + } + }, "DescribeChannelRequest": { "type": "structure", "members": { @@ -2000,6 +2328,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -2110,6 +2442,154 @@ } } }, + "DescribeOfferingRequest": { + "type": "structure", + "members": { + "OfferingId": { + "shape": "__string", + "location": "uri", + "locationName": "offeringId" + } + }, + "required": [ + "OfferingId" + ] + }, + "DescribeOfferingResponse": { + "type": "structure", + "members": { + "Arn": { + "shape": "__string", + "locationName": "arn" + }, + "CurrencyCode": { + "shape": "__string", + "locationName": "currencyCode" + }, + "Duration": { + "shape": "__integer", + "locationName": "duration" + }, + "DurationUnits": { + "shape": "OfferingDurationUnits", + "locationName": "durationUnits" + }, + "FixedPrice": { + "shape": "__double", + "locationName": "fixedPrice" + }, + "OfferingDescription": { + "shape": "__string", + "locationName": "offeringDescription" + }, + "OfferingId": { + "shape": "__string", + "locationName": "offeringId" + }, + "OfferingType": { + "shape": "OfferingType", + "locationName": "offeringType" + }, + "Region": { + "shape": "__string", + "locationName": "region" + }, + "ResourceSpecification": { + "shape": "ReservationResourceSpecification", + "locationName": "resourceSpecification" + }, + "UsagePrice": { + "shape": "__double", + "locationName": "usagePrice" + } + } + }, + "DescribeReservationRequest": { + "type": "structure", + "members": { + "ReservationId": { + "shape": "__string", + "location": "uri", + "locationName": "reservationId" + } + }, + "required": [ + "ReservationId" + ] + }, + "DescribeReservationResponse": { + "type": "structure", + "members": { + "Arn": { + "shape": "__string", + "locationName": "arn" + }, + "Count": { + "shape": "__integer", + "locationName": "count" + }, + "CurrencyCode": { + "shape": "__string", + "locationName": "currencyCode" + }, + "Duration": { + "shape": "__integer", + "locationName": "duration" + }, + "DurationUnits": { + "shape": "OfferingDurationUnits", + "locationName": "durationUnits" + }, + "End": { + "shape": "__string", + "locationName": "end" + }, + "FixedPrice": { + "shape": "__double", + "locationName": "fixedPrice" + }, + "Name": { + "shape": "__string", + "locationName": "name" + }, + "OfferingDescription": { + "shape": "__string", + "locationName": "offeringDescription" + }, + "OfferingId": { + "shape": "__string", + "locationName": "offeringId" + }, + "OfferingType": { + "shape": "OfferingType", + "locationName": "offeringType" + }, + "Region": { + "shape": "__string", + "locationName": "region" + }, + "ReservationId": { + "shape": "__string", + "locationName": "reservationId" + }, + "ResourceSpecification": { + "shape": "ReservationResourceSpecification", + "locationName": "resourceSpecification" + }, + "Start": { + "shape": "__string", + "locationName": "start" + }, + "State": { + "shape": "ReservationState", + "locationName": "state" + }, + "UsagePrice": { + "shape": "__double", + "locationName": "usagePrice" + } + } + }, "DvbNitSettings": { "type": "structure", "members": { @@ -4046,6 +4526,173 @@ } } }, + "ListOfferingsRequest": { + "type": "structure", + "members": { + "ChannelConfiguration": { + "shape": "__string", + "location": "querystring", + "locationName": "channelConfiguration" + }, + "Codec": { + "shape": "__string", + "location": "querystring", + "locationName": "codec" + }, + "MaxResults": { + "shape": "MaxResults", + "location": "querystring", + "locationName": "maxResults" + }, + "MaximumBitrate": { + "shape": "__string", + "location": "querystring", + "locationName": "maximumBitrate" + }, + "MaximumFramerate": { + "shape": "__string", + "location": "querystring", + "locationName": "maximumFramerate" + }, + "NextToken": { + "shape": "__string", + "location": "querystring", + "locationName": "nextToken" + }, + "Resolution": { + "shape": "__string", + "location": "querystring", + "locationName": "resolution" + }, + "ResourceType": { + "shape": "__string", + "location": "querystring", + "locationName": "resourceType" + }, + "SpecialFeature": { + "shape": "__string", + "location": "querystring", + "locationName": "specialFeature" + }, + "VideoQuality": { + "shape": "__string", + "location": "querystring", + "locationName": "videoQuality" + } + } + }, + "ListOfferingsResponse": { + "type": "structure", + "members": { + "NextToken": { + "shape": "__string", + "locationName": "nextToken" + }, + "Offerings": { + "shape": "__listOfOffering", + "locationName": "offerings" + } + } + }, + "ListOfferingsResultModel": { + "type": "structure", + "members": { + "NextToken": { + "shape": "__string", + "locationName": "nextToken" + }, + "Offerings": { + "shape": "__listOfOffering", + "locationName": "offerings" + } + } + }, + "ListReservationsRequest": { + "type": "structure", + "members": { + "Codec": { + "shape": "__string", + "location": "querystring", + "locationName": "codec" + }, + "MaxResults": { + "shape": "MaxResults", + "location": "querystring", + "locationName": "maxResults" + }, + "MaximumBitrate": { + "shape": "__string", + "location": "querystring", + "locationName": "maximumBitrate" + }, + "MaximumFramerate": { + "shape": "__string", + "location": "querystring", + "locationName": "maximumFramerate" + }, + "NextToken": { + "shape": "__string", + "location": "querystring", + "locationName": "nextToken" + }, + "Resolution": { + "shape": "__string", + "location": "querystring", + "locationName": "resolution" + }, + "ResourceType": { + "shape": "__string", + "location": "querystring", + "locationName": "resourceType" + }, + "SpecialFeature": { + "shape": "__string", + "location": "querystring", + "locationName": "specialFeature" + }, + "VideoQuality": { + "shape": "__string", + "location": "querystring", + "locationName": "videoQuality" + } + } + }, + "ListReservationsResponse": { + "type": "structure", + "members": { + "NextToken": { + "shape": "__string", + "locationName": "nextToken" + }, + "Reservations": { + "shape": "__listOfReservation", + "locationName": "reservations" + } + } + }, + "ListReservationsResultModel": { + "type": "structure", + "members": { + "NextToken": { + "shape": "__string", + "locationName": "nextToken" + }, + "Reservations": { + "shape": "__listOfReservation", + "locationName": "reservations" + } + } + }, + "LogLevel": { + "type": "string", + "enum": [ + "ERROR", + "WARNING", + "INFO", + "DEBUG", + "DISABLED" + ] + }, "M2tsAbsentInputAudioBehavior": { "type": "string", "enum": [ @@ -4610,6 +5257,67 @@ "httpStatusCode": 404 } }, + "Offering": { + "type": "structure", + "members": { + "Arn": { + "shape": "__string", + "locationName": "arn" + }, + "CurrencyCode": { + "shape": "__string", + "locationName": "currencyCode" + }, + "Duration": { + "shape": "__integer", + "locationName": "duration" + }, + "DurationUnits": { + "shape": "OfferingDurationUnits", + "locationName": "durationUnits" + }, + "FixedPrice": { + "shape": "__double", + "locationName": "fixedPrice" + }, + "OfferingDescription": { + "shape": "__string", + "locationName": "offeringDescription" + }, + "OfferingId": { + "shape": "__string", + "locationName": "offeringId" + }, + "OfferingType": { + "shape": "OfferingType", + "locationName": "offeringType" + }, + "Region": { + "shape": "__string", + "locationName": "region" + }, + "ResourceSpecification": { + "shape": "ReservationResourceSpecification", + "locationName": "resourceSpecification" + }, + "UsagePrice": { + "shape": "__double", + "locationName": "usagePrice" + } + } + }, + "OfferingDurationUnits": { + "type": "string", + "enum": [ + "MONTHS" + ] + }, + "OfferingType": { + "type": "string", + "enum": [ + "NO_UPFRONT" + ] + }, "Output": { "type": "structure", "members": { @@ -4757,6 +5465,68 @@ "members": { } }, + "PurchaseOffering": { + "type": "structure", + "members": { + "Count": { + "shape": "__integerMin1", + "locationName": "count" + }, + "Name": { + "shape": "__string", + "locationName": "name" + }, + "RequestId": { + "shape": "__string", + "locationName": "requestId", + "idempotencyToken": true + } + } + }, + "PurchaseOfferingRequest": { + "type": "structure", + "members": { + "Count": { + "shape": "__integerMin1", + "locationName": "count" + }, + "Name": { + "shape": "__string", + "locationName": "name" + }, + "OfferingId": { + "shape": "__string", + "location": "uri", + "locationName": "offeringId" + }, + "RequestId": { + "shape": "__string", + "locationName": "requestId", + "idempotencyToken": true + } + }, + "required": [ + "OfferingId" + ] + }, + "PurchaseOfferingResponse": { + "type": "structure", + "members": { + "Reservation": { + "shape": "Reservation", + "locationName": "reservation" + } + } + }, + "PurchaseOfferingResultModel": { + "type": "structure", + "members": { + "Reservation": { + "shape": "Reservation", + "locationName": "reservation" + } + } + }, "RemixSettings": { "type": "structure", "members": { @@ -4777,6 +5547,176 @@ "ChannelMappings" ] }, + "Reservation": { + "type": "structure", + "members": { + "Arn": { + "shape": "__string", + "locationName": "arn" + }, + "Count": { + "shape": "__integer", + "locationName": "count" + }, + "CurrencyCode": { + "shape": "__string", + "locationName": "currencyCode" + }, + "Duration": { + "shape": "__integer", + "locationName": "duration" + }, + "DurationUnits": { + "shape": "OfferingDurationUnits", + "locationName": "durationUnits" + }, + "End": { + "shape": "__string", + "locationName": "end" + }, + "FixedPrice": { + "shape": "__double", + "locationName": "fixedPrice" + }, + "Name": { + "shape": "__string", + "locationName": "name" + }, + "OfferingDescription": { + "shape": "__string", + "locationName": "offeringDescription" + }, + "OfferingId": { + "shape": "__string", + "locationName": "offeringId" + }, + "OfferingType": { + "shape": "OfferingType", + "locationName": "offeringType" + }, + "Region": { + "shape": "__string", + "locationName": "region" + }, + "ReservationId": { + "shape": "__string", + "locationName": "reservationId" + }, + "ResourceSpecification": { + "shape": "ReservationResourceSpecification", + "locationName": "resourceSpecification" + }, + "Start": { + "shape": "__string", + "locationName": "start" + }, + "State": { + "shape": "ReservationState", + "locationName": "state" + }, + "UsagePrice": { + "shape": "__double", + "locationName": "usagePrice" + } + } + }, + "ReservationCodec": { + "type": "string", + "enum": [ + "MPEG2", + "AVC", + "HEVC", + "AUDIO" + ] + }, + "ReservationMaximumBitrate": { + "type": "string", + "enum": [ + "MAX_10_MBPS", + "MAX_20_MBPS", + "MAX_50_MBPS" + ] + }, + "ReservationMaximumFramerate": { + "type": "string", + "enum": [ + "MAX_30_FPS", + "MAX_60_FPS" + ] + }, + "ReservationResolution": { + "type": "string", + "enum": [ + "SD", + "HD", + "UHD" + ] + }, + "ReservationResourceSpecification": { + "type": "structure", + "members": { + "Codec": { + "shape": "ReservationCodec", + "locationName": "codec" + }, + "MaximumBitrate": { + "shape": "ReservationMaximumBitrate", + "locationName": "maximumBitrate" + }, + "MaximumFramerate": { + "shape": "ReservationMaximumFramerate", + "locationName": "maximumFramerate" + }, + "Resolution": { + "shape": "ReservationResolution", + "locationName": "resolution" + }, + "ResourceType": { + "shape": "ReservationResourceType", + "locationName": "resourceType" + }, + "SpecialFeature": { + "shape": "ReservationSpecialFeature", + "locationName": "specialFeature" + }, + "VideoQuality": { + "shape": "ReservationVideoQuality", + "locationName": "videoQuality" + } + } + }, + "ReservationResourceType": { + "type": "string", + "enum": [ + "INPUT", + "OUTPUT", + "CHANNEL" + ] + }, + "ReservationSpecialFeature": { + "type": "string", + "enum": [ + "ADVANCED_AUDIO", + "AUDIO_NORMALIZATION" + ] + }, + "ReservationState": { + "type": "string", + "enum": [ + "ACTIVE", + "EXPIRED", + "CANCELED", + "DELETED" + ] + }, + "ReservationVideoQuality": { + "type": "string", + "enum": [ + "STANDARD", + "ENHANCED", + "PREMIUM" + ] + }, "ResourceConflict": { "type": "structure", "members": { @@ -5094,6 +6034,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -5125,6 +6069,7 @@ } }, "required": [ + "KeyProviderServer", "StaticKeyValue" ] }, @@ -5172,6 +6117,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -5352,6 +6301,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -5386,6 +6339,10 @@ "shape": "InputSpecification", "locationName": "inputSpecification" }, + "LogLevel": { + "shape": "LogLevel", + "locationName": "logLevel" + }, "Name": { "shape": "__string", "locationName": "name" @@ -5978,6 +6935,12 @@ "shape": "InputWhitelistRuleCidr" } }, + "__listOfOffering": { + "type": "list", + "member": { + "shape": "Offering" + } + }, "__listOfOutput": { "type": "list", "member": { @@ -6002,6 +6965,12 @@ "shape": "OutputGroup" } }, + "__listOfReservation": { + "type": "list", + "member": { + "shape": "Reservation" + } + }, "__listOfValidationError": { "type": "list", "member": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/docs-2.json index 0c7e3d3c8..50a0537c4 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/docs-2.json @@ -8,12 +8,18 @@ "DeleteChannel": "Starts deletion of channel. The associated outputs are also deleted.", "DeleteInput": "Deletes the input end point", "DeleteInputSecurityGroup": "Deletes an Input Security Group", + "DeleteReservation": "Delete an expired reservation.", "DescribeChannel": "Gets details about a channel", "DescribeInput": "Produces details about an input", "DescribeInputSecurityGroup": "Produces a summary of an Input Security Group", + "DescribeOffering": "Get details for an offering.", + "DescribeReservation": "Get details for a reservation.", "ListChannels": "Produces list of channels that have been created", "ListInputSecurityGroups": "Produces a list of Input Security Groups for an account", "ListInputs": "Produces list of inputs that have been created", + "ListOfferings": "List offerings available for purchase.", + "ListReservations": "List purchased reservations.", + "PurchaseOffering": "Purchase an offering and create a reservation.", "StartChannel": "Starts an existing channel", "StopChannel": "Stops a running channel", "UpdateChannel": "Updates a channel.", @@ -1207,6 +1213,25 @@ "refs": { } }, + "ListOfferingsResultModel": { + "base": "ListOfferings response", + "refs": { + } + }, + "ListReservationsResultModel": { + "base": "ListReservations response", + "refs": { + } + }, + "LogLevel": { + "base": null, + "refs": { + "Channel$LogLevel": "The log level being written to CloudWatch Logs.", + "ChannelSummary$LogLevel": "The log level being written to CloudWatch Logs.", + "CreateChannel$LogLevel": "The log level to write to CloudWatch Logs.", + "UpdateChannel$LogLevel": "The log level to write to CloudWatch Logs." + } + }, "M2tsAbsentInputAudioBehavior": { "base": null, "refs": { @@ -1387,6 +1412,26 @@ "refs": { } }, + "Offering": { + "base": "Reserved resources available for purchase", + "refs": { + "__listOfOffering$member": null + } + }, + "OfferingDurationUnits": { + "base": "Units for duration, e.g. 'MONTHS'", + "refs": { + "Offering$DurationUnits": "Units for duration, e.g. 'MONTHS'", + "Reservation$DurationUnits": "Units for duration, e.g. 'MONTHS'" + } + }, + "OfferingType": { + "base": "Offering type, e.g. 'NO_UPFRONT'", + "refs": { + "Offering$OfferingType": "Offering type, e.g. 'NO_UPFRONT'", + "Reservation$OfferingType": "Offering type, e.g. 'NO_UPFRONT'" + } + }, "Output": { "base": "Output settings. There can be multiple outputs within a group.", "refs": { @@ -1439,12 +1484,84 @@ "AudioCodecSettings$PassThroughSettings": null } }, + "PurchaseOffering": { + "base": "PurchaseOffering request", + "refs": { + } + }, + "PurchaseOfferingResultModel": { + "base": "PurchaseOffering response", + "refs": { + } + }, "RemixSettings": { "base": null, "refs": { "AudioDescription$RemixSettings": "Settings that control how input audio channels are remixed into the output audio channels." } }, + "Reservation": { + "base": "Reserved resources available to use", + "refs": { + "PurchaseOfferingResultModel$Reservation": null, + "__listOfReservation$member": null + } + }, + "ReservationCodec": { + "base": "Codec, 'MPEG2', 'AVC', 'HEVC', or 'AUDIO'", + "refs": { + "ReservationResourceSpecification$Codec": "Codec, e.g. 'AVC'" + } + }, + "ReservationMaximumBitrate": { + "base": "Maximum bitrate in megabits per second", + "refs": { + "ReservationResourceSpecification$MaximumBitrate": "Maximum bitrate, e.g. 'MAX_20_MBPS'" + } + }, + "ReservationMaximumFramerate": { + "base": "Maximum framerate in frames per second (Outputs only)", + "refs": { + "ReservationResourceSpecification$MaximumFramerate": "Maximum framerate, e.g. 'MAX_30_FPS' (Outputs only)" + } + }, + "ReservationResolution": { + "base": "Resolution based on lines of vertical resolution; SD is less than 720 lines, HD is 720 to 1080 lines, UHD is greater than 1080 lines\n", + "refs": { + "ReservationResourceSpecification$Resolution": "Resolution, e.g. 'HD'" + } + }, + "ReservationResourceSpecification": { + "base": "Resource configuration (codec, resolution, bitrate, ...)", + "refs": { + "Offering$ResourceSpecification": "Resource configuration details", + "Reservation$ResourceSpecification": "Resource configuration details" + } + }, + "ReservationResourceType": { + "base": "Resource type, 'INPUT', 'OUTPUT', or 'CHANNEL'", + "refs": { + "ReservationResourceSpecification$ResourceType": "Resource type, 'INPUT', 'OUTPUT', or 'CHANNEL'" + } + }, + "ReservationSpecialFeature": { + "base": "Special features, 'ADVANCED_AUDIO' or 'AUDIO_NORMALIZATION'", + "refs": { + "ReservationResourceSpecification$SpecialFeature": "Special feature, e.g. 'AUDIO_NORMALIZATION' (Channels only)" + } + }, + "ReservationState": { + "base": "Current reservation state", + "refs": { + "Reservation$State": "Current state of reservation, e.g. 'ACTIVE'" + } + }, + "ReservationVideoQuality": { + "base": "Video quality, e.g. 'STANDARD' (Outputs only)", + "refs": { + "ReservationResourceSpecification$VideoQuality": "Video quality, e.g. 'STANDARD' (Outputs only)" + } + }, "ResourceConflict": { "base": null, "refs": { @@ -1802,7 +1919,11 @@ "Eac3Settings$LtRtCenterMixLevel": "Left total/Right total center mix level. Only used for 3/2 coding mode.", "Eac3Settings$LtRtSurroundMixLevel": "Left total/Right total surround mix level. Only used for 3/2 coding mode.", "Mp2Settings$Bitrate": "Average bitrate in bits/second.", - "Mp2Settings$SampleRate": "Sample rate in Hz." + "Mp2Settings$SampleRate": "Sample rate in Hz.", + "Offering$FixedPrice": "One-time charge for each reserved resource, e.g. '0.0' for a NO_UPFRONT offering", + "Offering$UsagePrice": "Recurring usage charge for each reserved resource, e.g. '157.0'", + "Reservation$FixedPrice": "One-time charge for each reserved resource, e.g. '0.0' for a NO_UPFRONT offering", + "Reservation$UsagePrice": "Recurring usage charge for each reserved resource, e.g. '157.0'" } }, "__doubleMin0": { @@ -1837,6 +1958,9 @@ "H264Settings$FramerateDenominator": "Framerate denominator.", "H264Settings$FramerateNumerator": "Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 fps.", "H264Settings$ParNumerator": "Pixel Aspect Ratio numerator.", + "Offering$Duration": "Lease duration, e.g. '12'", + "Reservation$Count": "Number of reserved resources", + "Reservation$Duration": "Lease duration, e.g. '12'", "VideoDescription$Height": "Output video height (in pixels). Leave blank to use source video height. If left blank, width must also be unspecified.", "VideoDescription$Width": "Output video width (in pixels). Leave out to use source video width. If left out, height must also be left out. Display aspect ratio is always preserved by letterboxing or pillarboxing when necessary." } @@ -2010,6 +2134,7 @@ "HlsGroupSettings$SegmentLength": "Length of MPEG-2 Transport Stream segments to create (in seconds). Note that segments will end on the next keyframe after this number of seconds, so actual segment length may be longer.", "HlsGroupSettings$SegmentsPerSubdirectory": "Number of segments to write to a subdirectory before starting a new one. directoryStructure must be subdirectoryPerStream for this setting to have an effect.", "MsSmoothGroupSettings$FragmentLength": "Length of mp4 fragments to generate (in seconds). Fragment length must be compatible with GOP size and framerate.", + "PurchaseOffering$Count": "Number of resources", "RtmpOutputSettings$ConnectionRetryInterval": "Number of seconds to wait before retrying a connection to the Flash Media server if the connection is lost.", "Scte27SourceSettings$Pid": "The pid field is used in conjunction with the caption selector languageCode field as follows:\n - Specify PID and Language: Extracts captions from that PID; the language is \"informational\".\n - Specify PID and omit Language: Extracts the specified PID.\n - Omit PID and specify Language: Extracts the specified language, whichever PID that happens to be.\n - Omit PID and omit Language: Valid only if source is DVB-Sub that is being passed through; all languages will be passed through." } @@ -2261,6 +2386,12 @@ "InputSecurityGroupWhitelistRequest$WhitelistRules": "List of IPv4 CIDR addresses to whitelist" } }, + "__listOfOffering": { + "base": null, + "refs": { + "ListOfferingsResultModel$Offerings": "List of offerings" + } + }, "__listOfOutput": { "base": null, "refs": { @@ -2288,6 +2419,12 @@ "EncoderSettings$OutputGroups": null } }, + "__listOfReservation": { + "base": null, + "refs": { + "ListReservationsResultModel$Reservations": "List of reservations" + } + }, "__listOfValidationError": { "base": null, "refs": { @@ -2384,6 +2521,8 @@ "ListChannelsResultModel$NextToken": null, "ListInputSecurityGroupsResultModel$NextToken": null, "ListInputsResultModel$NextToken": null, + "ListOfferingsResultModel$NextToken": "Token to retrieve the next page of results", + "ListReservationsResultModel$NextToken": "Token to retrieve the next page of results", "M2tsSettings$AribCaptionsPid": "Packet Identifier (PID) for ARIB Captions in the transport stream. Can be entered as a decimal or hexadecimal value. Valid values are 32 (or 0x20)..8182 (or 0x1ff6).", "M2tsSettings$AudioPids": "Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation. Can be entered as decimal or hexadecimal values. Each PID specified must be in the range of 32 (or 0x20)..8182 (or 0x1ff6).", "M2tsSettings$DvbSubPids": "Packet Identifier (PID) for input source DVB Subtitle data to this output. Multiple values are accepted, and can be entered in ranges and/or by comma separation. Can be entered as decimal or hexadecimal values. Each PID specified must be in the range of 32 (or 0x20)..8182 (or 0x1ff6).", @@ -2409,6 +2548,11 @@ "MsSmoothGroupSettings$EventId": "MS Smooth event ID to be sent to the IIS server.\n\nShould only be specified if eventIdMode is set to useConfigured.", "MsSmoothGroupSettings$TimestampOffset": "Timestamp offset for the event. Only used if timestampOffsetMode is set to useConfiguredOffset.", "MsSmoothOutputSettings$NameModifier": "String concatenated to the end of the destination filename. Required for multiple outputs of the same type.", + "Offering$Arn": "Unique offering ARN, e.g. 'arn:aws:medialive:us-west-2:123456789012:offering:87654321'", + "Offering$CurrencyCode": "Currency code for usagePrice and fixedPrice in ISO-4217 format, e.g. 'USD'", + "Offering$OfferingDescription": "Offering description, e.g. 'HD AVC output at 10-20 Mbps, 30 fps, and standard VQ in US West (Oregon)'", + "Offering$OfferingId": "Unique offering ID, e.g. '87654321'", + "Offering$Region": "AWS region, e.g. 'us-west-2'", "Output$VideoDescriptionName": "The name of the VideoDescription used as the source for this output.", "OutputDestination$Id": "User-specified id. This is used in an output group or an output.", "OutputDestinationSettings$PasswordParam": "key used to extract the password from EC2 Parameter store", @@ -2416,6 +2560,17 @@ "OutputDestinationSettings$Url": "A URL specifying a destination", "OutputDestinationSettings$Username": "username for destination", "OutputLocationRef$DestinationRefId": null, + "PurchaseOffering$Name": "Name for the new reservation", + "PurchaseOffering$RequestId": "Unique request ID to be specified. This is needed to prevent retries from creating multiple resources.", + "Reservation$Arn": "Unique reservation ARN, e.g. 'arn:aws:medialive:us-west-2:123456789012:reservation:1234567'", + "Reservation$CurrencyCode": "Currency code for usagePrice and fixedPrice in ISO-4217 format, e.g. 'USD'", + "Reservation$End": "Reservation UTC end date and time in ISO-8601 format, e.g. '2019-03-01T00:00:00'", + "Reservation$Name": "User specified reservation name", + "Reservation$OfferingDescription": "Offering description, e.g. 'HD AVC output at 10-20 Mbps, 30 fps, and standard VQ in US West (Oregon)'", + "Reservation$OfferingId": "Unique offering ID, e.g. '87654321'", + "Reservation$Region": "AWS region, e.g. 'us-west-2'", + "Reservation$ReservationId": "Unique reservation ID, e.g. '1234567'", + "Reservation$Start": "Reservation UTC start date and time in ISO-8601 format, e.g. '2018-03-01T00:00:00'", "ResourceConflict$Message": null, "ResourceNotFound$Message": null, "StandardHlsSettings$AudioRenditionSets": "List all the audio groups that are used with the video output stream. Input all the audio GROUP-IDs that are associated to the video, separate by ','.", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/paginators-1.json index 1335d9663..c9bc58df4 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/medialive/2017-10-14/paginators-1.json @@ -17,6 +17,18 @@ "output_token": "NextToken", "limit_key": "MaxResults", "result_key": "InputSecurityGroups" + }, + "ListOfferings": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults", + "result_key": "Offerings" + }, + "ListReservations": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults", + "result_key": "Reservations" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/api-2.json new file mode 100644 index 000000000..757ba1355 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/api-2.json @@ -0,0 +1,266 @@ +{ + "metadata" : { + "apiVersion" : "2018-04-23", + "endpointPrefix" : "api.mediatailor", + "signingName" : "mediatailor", + "serviceFullName" : "AWS MediaTailor", + "serviceId" : "MediaTailor", + "protocol" : "rest-json", + "jsonVersion" : "1.1", + "uid" : "mediatailor-2018-04-23", + "signatureVersion" : "v4", + "serviceAbbreviation": "MediaTailor" + }, + "operations" : { + "DeletePlaybackConfiguration" : { + "name" : "DeletePlaybackConfiguration", + "http" : { + "method" : "DELETE", + "requestUri" : "/playbackConfiguration/{Name}", + "responseCode" : 204 + }, + "input" : { + "shape" : "DeletePlaybackConfigurationRequest" + }, + "errors" : [ ] + }, + "GetPlaybackConfiguration" : { + "name" : "GetPlaybackConfiguration", + "http" : { + "method" : "GET", + "requestUri" : "/playbackConfiguration/{Name}", + "responseCode" : 200 + }, + "input" : { + "shape" : "GetPlaybackConfigurationRequest" + }, + "output" : { + "shape" : "GetPlaybackConfigurationResponse" + }, + "errors" : [ ] + }, + "ListPlaybackConfigurations" : { + "name" : "ListPlaybackConfigurations", + "http" : { + "method" : "GET", + "requestUri" : "/playbackConfigurations", + "responseCode" : 200 + }, + "input" : { + "shape" : "ListPlaybackConfigurationsRequest" + }, + "output" : { + "shape" : "ListPlaybackConfigurationsResponse" + }, + "errors" : [ ] + }, + "PutPlaybackConfiguration" : { + "name" : "PutPlaybackConfiguration", + "http" : { + "method" : "PUT", + "requestUri" : "/playbackConfiguration", + "responseCode" : 200 + }, + "input" : { + "shape" : "PutPlaybackConfigurationRequest" + }, + "output" : { + "shape" : "PutPlaybackConfigurationResponse" + }, + "errors" : [ ] + } + }, + "shapes" : { + "CdnConfiguration" : { + "type" : "structure", + "members" : { + "AdSegmentUrlPrefix" : { + "shape" : "__string" + }, + "ContentSegmentUrlPrefix" : { + "shape" : "__string" + } + } + }, + "HlsConfiguration": { + "type" : "structure", + "members" : { + "ManifestEndpointPrefix" : { + "shape" : "__string" + } + } + }, + "DeletePlaybackConfigurationRequest" : { + "type" : "structure", + "members" : { + "Name" : { + "shape" : "__string", + "location" : "uri", + "locationName" : "Name" + } + }, + "required" : [ "Name" ] + }, + "GetPlaybackConfigurationRequest" : { + "type" : "structure", + "members" : { + "Name" : { + "shape" : "__string", + "location" : "uri", + "locationName" : "Name" + } + }, + "required" : [ "Name" ] + }, + "GetPlaybackConfigurationResponse" : { + "type" : "structure", + "members" : { + "AdDecisionServerUrl" : { + "shape" : "__string" + }, + "CdnConfiguration" : { + "shape" : "CdnConfiguration" + }, + "HlsConfiguration" : { + "shape" : "HlsConfiguration" + }, + "Name" : { + "shape" : "__string" + }, + "PlaybackEndpointPrefix" : { + "shape" : "__string" + }, + "SessionInitializationEndpointPrefix" : { + "shape" : "__string" + }, + "SlateAdUrl" : { + "shape" : "__string" + }, + "VideoContentSourceUrl" : { + "shape" : "__string" + } + } + }, + "PlaybackConfiguration" : { + "type" : "structure", + "members" : { + "AdDecisionServerUrl" : { + "shape" : "__string" + }, + "CdnConfiguration" : { + "shape" : "CdnConfiguration" + }, + "Name" : { + "shape" : "__string" + }, + "SlateAdUrl" : { + "shape" : "__string" + }, + "VideoContentSourceUrl" : { + "shape" : "__string" + } + } + }, + "ListPlaybackConfigurationsRequest" : { + "type" : "structure", + "members" : { + "MaxResults" : { + "shape" : "__integerMin1Max100", + "location" : "querystring", + "locationName" : "MaxResults" + }, + "NextToken" : { + "shape" : "__string", + "location" : "querystring", + "locationName" : "NextToken" + } + } + }, + "ListPlaybackConfigurationsResponse" : { + "type" : "structure", + "members" : { + "Items" : { + "shape" : "__listOfPlaybackConfigurations" + }, + "NextToken" : { + "shape" : "__string" + } + } + }, + "PutPlaybackConfigurationRequest" : { + "type" : "structure", + "members" : { + "AdDecisionServerUrl" : { + "shape" : "__string" + }, + "CdnConfiguration" : { + "shape" : "CdnConfiguration" + }, + "Name" : { + "shape" : "__string" + }, + "SlateAdUrl" : { + "shape" : "__string" + }, + "VideoContentSourceUrl" : { + "shape" : "__string" + } + } + }, + "PutPlaybackConfigurationResponse" : { + "type" : "structure", + "members" : { + "AdDecisionServerUrl" : { + "shape" : "__string" + }, + "CdnConfiguration" : { + "shape" : "CdnConfiguration" + }, + "HlsConfiguration" : { + "shape" : "HlsConfiguration" + }, + "Name" : { + "shape" : "__string" + }, + "PlaybackEndpointPrefix" : { + "shape" : "__string" + }, + "SessionInitializationEndpointPrefix" : { + "shape" : "__string" + }, + "SlateAdUrl" : { + "shape" : "__string" + }, + "VideoContentSourceUrl" : { + "shape" : "__string" + } + } + }, + "__boolean" : { + "type" : "boolean" + }, + "__double" : { + "type" : "double" + }, + "__integer" : { + "type" : "integer" + }, + "__listOfPlaybackConfigurations" : { + "type" : "list", + "member" : { + "shape" : "PlaybackConfiguration" + } + }, + "__long" : { + "type" : "long" + }, + "__string" : { + "type" : "string" + }, + "__integerMin1Max100" : { + "type": "integer", + "min": 1, + "max": 100 + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/docs-2.json new file mode 100644 index 000000000..6c8bc2b6d --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/docs-2.json @@ -0,0 +1,67 @@ +{ + "version" : "2.0", + "service" : "

Use the AWS Elemental MediaTailor SDK to configure scalable ad insertion for your live and VOD content. With AWS Elemental MediaTailor, you can serve targeted ads to viewers while maintaining broadcast quality in over-the-top (OTT) video applications. For information about using the service, including detailed information about the settings covered in this guide, see the AWS Elemental MediaTailor User Guide.

Through the SDK, you manage AWS Elemental MediaTailor configurations the same as you do through the console. For example, you specify ad insertion behavior and mapping information for the origin server and the ad decision server (ADS).

", + "operations" : { + "DeletePlaybackConfiguration" : "

Deletes the configuration for the specified name.

", + "GetPlaybackConfiguration" : "

Returns the configuration for the specified name.

", + "ListPlaybackConfigurations" : "

Returns a list of the configurations defined in AWS Elemental MediaTailor. You can specify a max number of configurations to return at a time. The default max is 50. Results are returned in pagefuls. If AWS Elemental MediaTailor has more configurations than the specified max, it provides parameters in the response that you can use to retrieve the next pageful.

", + "PutPlaybackConfiguration" : "

Adds a new configuration to AWS Elemental MediaTailor.

" + }, + "shapes" : { + "CdnConfiguration" : { + "base" : "

The configuration for using a content delivery network (CDN), like Amazon CloudFront, for content and ad segment management.

", + "refs" : { + "GetPlaybackConfigurationResponse$CdnConfiguration" : "

The configuration for using a content delivery network (CDN), like Amazon CloudFront, for content and ad segment management.

", + "PutPlaybackConfigurationRequest$CdnConfiguration" : "

The configuration for using a content delivery network (CDN), like Amazon CloudFront, for content and ad segment management.

" + } + }, + "PlaybackConfiguration" : { + "refs" : { + "__listOfPlaybackConfigurations$member" : "

The AWSMediaTailor configuration.

" + } + }, + "GetPlaybackConfigurationResponse" : { + "base" : null, + "refs" : { } + }, + "HlsConfiguration" : { + "base" : "

The configuration for HLS content.

", + "refs" : { + "GetPlaybackConfigurationResponse$HlsConfiguration" : "

The configuration for HLS content.

" + } + }, + "ListPlaybackConfigurationsResponse" : { + "base" : null, + "refs" : { } + }, + "PutPlaybackConfigurationRequest" : { + "base" : null, + "refs" : { } + }, + "__listOfPlaybackConfigurations" : { + "base" : null, + "refs" : { + "ListPlaybackConfigurationsResponse$Items" : "

Array of playback configurations. This may be all of the available configurations or a subset, depending on the settings you provide and on the total number of configurations stored.

" + } + }, + "__string" : { + "base" : null, + "refs" : { + "CdnConfiguration$AdSegmentUrlPrefix" : "

A non-default content delivery network (CDN) to serve ad segments. By default, AWS Elemental MediaTailor uses Amazon CloudFront with default cache settings as its CDN for ad segments. To set up an alternate CDN, create a rule in your CDN for the following origin: ads.mediatailor.<region>.amazonaws.com. Then specify the rule's name in this AdSegmentUrlPrefix. When AWS Elemental MediaTailor serves a manifest, it reports your CDN as the source for ad segments.

", + "CdnConfiguration$ContentSegmentUrlPrefix" : "

A content delivery network (CDN) to cache content segments, so that content requests don’t always have to go to the origin server. First, create a rule in your CDN for the content segment origin server. Then specify the rule's name in this ContentSegmentUrlPrefix. When AWS Elemental MediaTailor serves a manifest, it reports your CDN as the source for content segments.

", + "GetPlaybackConfigurationResponse$AdDecisionServerUrl" : "

The URL for the ad decision server (ADS). This includes the specification of static parameters and placeholders for dynamic parameters. AWS Elemental MediaTailor substitutes player-specific and session-specific parameters as needed when calling the ADS. Alternately, for testing, you can provide a static VAST URL. The maximum length is 25000 characters.

", + "GetPlaybackConfigurationResponse$Name" : "

The identifier for the configuration.

", + "GetPlaybackConfigurationResponse$PlaybackEndpointPrefix" : "

The URL that the player accesses to get a manifest from AWS Elemental MediaTailor. This session will use server-side reporting.

", + "GetPlaybackConfigurationResponse$SessionInitializationEndpointPrefix" : "

The URL that the player uses to initialize a session that uses client-side reporting.

", + "GetPlaybackConfigurationResponse$SlateAdUrl" : "

URL for a high-quality video asset to transcode and use to fill in time that's not used by ads. AWS Elemental MediaTailor shows the slate to fill in gaps in media content. Configuring the slate is optional for non-VPAID configurations. For VPAID, the slate is required because AWS Elemental MediaTailor provides it in the slots designated for dynamic ad content. The slate must be a high-quality asset that contains both audio and video.

", + "GetPlaybackConfigurationResponse$VideoContentSourceUrl" : "

The URL prefix for the master playlist for the stream, minus the asset ID. The maximum length is 512 characters.

", + "HlsConfiguration$ManifestEndpointPrefix" : "

The URL that is used to initiate a playback session for devices that support Apple HLS. The session uses server-side reporting.

", + "ListPlaybackConfigurationsResponse$NextToken" : "

Pagination token returned by the GET list request when results overrun the meximum allowed. Use the token to fetch the next page of results.

", + "PutPlaybackConfigurationRequest$AdDecisionServerUrl" : "

The URL for the ad decision server (ADS). This includes the specification of static parameters and placeholders for dynamic parameters. AWS Elemental MediaTailor substitutes player-specific and session-specific parameters as needed when calling the ADS. Alternately, for testing you can provide a static VAST URL. The maximum length is 25000 characters.

", + "PutPlaybackConfigurationRequest$Name" : "

The identifier for the configuration.

", + "PutPlaybackConfigurationRequest$SlateAdUrl" : "

The URL for a high-quality video asset to transcode and use to fill in time that's not used by ads. AWS Elemental MediaTailor shows the slate to fill in gaps in media content. Configuring the slate is optional for non-VPAID configurations. For VPAID, the slate is required because AWS Elemental MediaTailor provides it in the slots that are designated for dynamic ad content. The slate must be a high-quality asset that contains both audio and video.

", + "PutPlaybackConfigurationRequest$VideoContentSourceUrl" : "

The URL prefix for the master playlist for the stream, minus the asset ID. The maximum length is 512 characters.

" + } + } + } +} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/paginators-1.json new file mode 100644 index 000000000..f3b7195d8 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/mediatailor/2018-04-23/paginators-1.json @@ -0,0 +1,3 @@ +{ + "pagination" : { } +} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/api-2.json similarity index 65% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/api-2.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/api-2.json index 6115af1f0..96fc58c72 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/api-2.json @@ -1,17 +1,32 @@ { "version":"2.0", "metadata":{ - "apiVersion":"2014-09-01", + "apiVersion":"2014-10-31", "endpointPrefix":"rds", "protocol":"query", - "serviceAbbreviation":"Amazon RDS", - "serviceFullName":"Amazon Relational Database Service", - "serviceId":"RDS", + "serviceAbbreviation":"Amazon Neptune", + "serviceFullName":"Amazon Neptune", + "serviceId":"Neptune", "signatureVersion":"v4", - "uid":"rds-2014-09-01", - "xmlNamespace":"http://rds.amazonaws.com/doc/2014-09-01/" + "signingName":"rds", + "uid":"neptune-2014-10-31", + "xmlNamespace":"http://rds.amazonaws.com/doc/2014-10-31/" }, "operations":{ + "AddRoleToDBCluster":{ + "name":"AddRoleToDBCluster", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"AddRoleToDBClusterMessage"}, + "errors":[ + {"shape":"DBClusterNotFoundFault"}, + {"shape":"DBClusterRoleAlreadyExistsFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"DBClusterRoleQuotaExceededFault"} + ] + }, "AddSourceIdentifierToSubscription":{ "name":"AddSourceIdentifierToSubscription", "http":{ @@ -37,25 +52,60 @@ "input":{"shape":"AddTagsToResourceMessage"}, "errors":[ {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} + {"shape":"DBSnapshotNotFoundFault"}, + {"shape":"DBClusterNotFoundFault"} ] }, - "AuthorizeDBSecurityGroupIngress":{ - "name":"AuthorizeDBSecurityGroupIngress", + "ApplyPendingMaintenanceAction":{ + "name":"ApplyPendingMaintenanceAction", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"AuthorizeDBSecurityGroupIngressMessage"}, + "input":{"shape":"ApplyPendingMaintenanceActionMessage"}, "output":{ - "shape":"AuthorizeDBSecurityGroupIngressResult", - "resultWrapper":"AuthorizeDBSecurityGroupIngressResult" + "shape":"ApplyPendingMaintenanceActionResult", + "resultWrapper":"ApplyPendingMaintenanceActionResult" }, "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"AuthorizationAlreadyExistsFault"}, - {"shape":"AuthorizationQuotaExceededFault"} + {"shape":"ResourceNotFoundFault"} + ] + }, + "CopyDBClusterParameterGroup":{ + "name":"CopyDBClusterParameterGroup", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"CopyDBClusterParameterGroupMessage"}, + "output":{ + "shape":"CopyDBClusterParameterGroupResult", + "resultWrapper":"CopyDBClusterParameterGroupResult" + }, + "errors":[ + {"shape":"DBParameterGroupNotFoundFault"}, + {"shape":"DBParameterGroupQuotaExceededFault"}, + {"shape":"DBParameterGroupAlreadyExistsFault"} + ] + }, + "CopyDBClusterSnapshot":{ + "name":"CopyDBClusterSnapshot", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"CopyDBClusterSnapshotMessage"}, + "output":{ + "shape":"CopyDBClusterSnapshotResult", + "resultWrapper":"CopyDBClusterSnapshotResult" + }, + "errors":[ + {"shape":"DBClusterSnapshotAlreadyExistsFault"}, + {"shape":"DBClusterSnapshotNotFoundFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"InvalidDBClusterSnapshotStateFault"}, + {"shape":"SnapshotQuotaExceededFault"}, + {"shape":"KMSKeyNotAccessibleFault"} ] }, "CopyDBParameterGroup":{ @@ -75,39 +125,68 @@ {"shape":"DBParameterGroupQuotaExceededFault"} ] }, - "CopyDBSnapshot":{ - "name":"CopyDBSnapshot", + "CreateDBCluster":{ + "name":"CreateDBCluster", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"CopyDBSnapshotMessage"}, + "input":{"shape":"CreateDBClusterMessage"}, "output":{ - "shape":"CopyDBSnapshotResult", - "resultWrapper":"CopyDBSnapshotResult" + "shape":"CreateDBClusterResult", + "resultWrapper":"CreateDBClusterResult" }, "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"SnapshotQuotaExceededFault"} + {"shape":"DBClusterAlreadyExistsFault"}, + {"shape":"InsufficientStorageClusterCapacityFault"}, + {"shape":"DBClusterQuotaExceededFault"}, + {"shape":"StorageQuotaExceededFault"}, + {"shape":"DBSubnetGroupNotFoundFault"}, + {"shape":"InvalidVPCNetworkStateFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"InvalidDBSubnetGroupStateFault"}, + {"shape":"InvalidSubnet"}, + {"shape":"InvalidDBInstanceStateFault"}, + {"shape":"DBClusterParameterGroupNotFoundFault"}, + {"shape":"KMSKeyNotAccessibleFault"}, + {"shape":"DBClusterNotFoundFault"}, + {"shape":"DBInstanceNotFoundFault"}, + {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"} ] }, - "CopyOptionGroup":{ - "name":"CopyOptionGroup", + "CreateDBClusterParameterGroup":{ + "name":"CreateDBClusterParameterGroup", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"CopyOptionGroupMessage"}, + "input":{"shape":"CreateDBClusterParameterGroupMessage"}, "output":{ - "shape":"CopyOptionGroupResult", - "resultWrapper":"CopyOptionGroupResult" + "shape":"CreateDBClusterParameterGroupResult", + "resultWrapper":"CreateDBClusterParameterGroupResult" }, "errors":[ - {"shape":"OptionGroupAlreadyExistsFault"}, - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"OptionGroupQuotaExceededFault"} + {"shape":"DBParameterGroupQuotaExceededFault"}, + {"shape":"DBParameterGroupAlreadyExistsFault"} + ] + }, + "CreateDBClusterSnapshot":{ + "name":"CreateDBClusterSnapshot", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"CreateDBClusterSnapshotMessage"}, + "output":{ + "shape":"CreateDBClusterSnapshotResult", + "resultWrapper":"CreateDBClusterSnapshotResult" + }, + "errors":[ + {"shape":"DBClusterSnapshotAlreadyExistsFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"DBClusterNotFoundFault"}, + {"shape":"SnapshotQuotaExceededFault"}, + {"shape":"InvalidDBClusterSnapshotStateFault"} ] }, "CreateDBInstance":{ @@ -130,43 +209,16 @@ {"shape":"StorageQuotaExceededFault"}, {"shape":"DBSubnetGroupNotFoundFault"}, {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, + {"shape":"InvalidDBClusterStateFault"}, {"shape":"InvalidSubnet"}, {"shape":"InvalidVPCNetworkStateFault"}, {"shape":"ProvisionedIopsNotAvailableInAZFault"}, {"shape":"OptionGroupNotFoundFault"}, + {"shape":"DBClusterNotFoundFault"}, {"shape":"StorageTypeNotSupportedFault"}, - {"shape":"AuthorizationNotFoundFault"} - ] - }, - "CreateDBInstanceReadReplica":{ - "name":"CreateDBInstanceReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceReadReplicaMessage"}, - "output":{ - "shape":"CreateDBInstanceReadReplicaResult", - "resultWrapper":"CreateDBInstanceReadReplicaResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"DBSubnetGroupNotAllowedFault"}, - {"shape":"InvalidDBSubnetGroupFault"}, - {"shape":"StorageTypeNotSupportedFault"} + {"shape":"AuthorizationNotFoundFault"}, + {"shape":"KMSKeyNotAccessibleFault"}, + {"shape":"DomainNotFoundFault"} ] }, "CreateDBParameterGroup":{ @@ -185,41 +237,6 @@ {"shape":"DBParameterGroupAlreadyExistsFault"} ] }, - "CreateDBSecurityGroup":{ - "name":"CreateDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSecurityGroupMessage"}, - "output":{ - "shape":"CreateDBSecurityGroupResult", - "resultWrapper":"CreateDBSecurityGroupResult" - }, - "errors":[ - {"shape":"DBSecurityGroupAlreadyExistsFault"}, - {"shape":"DBSecurityGroupQuotaExceededFault"}, - {"shape":"DBSecurityGroupNotSupportedFault"} - ] - }, - "CreateDBSnapshot":{ - "name":"CreateDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSnapshotMessage"}, - "output":{ - "shape":"CreateDBSnapshotResult", - "resultWrapper":"CreateDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, "CreateDBSubnetGroup":{ "name":"CreateDBSubnetGroup", "http":{ @@ -260,20 +277,51 @@ {"shape":"SourceNotFoundFault"} ] }, - "CreateOptionGroup":{ - "name":"CreateOptionGroup", + "DeleteDBCluster":{ + "name":"DeleteDBCluster", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"CreateOptionGroupMessage"}, + "input":{"shape":"DeleteDBClusterMessage"}, "output":{ - "shape":"CreateOptionGroupResult", - "resultWrapper":"CreateOptionGroupResult" + "shape":"DeleteDBClusterResult", + "resultWrapper":"DeleteDBClusterResult" }, "errors":[ - {"shape":"OptionGroupAlreadyExistsFault"}, - {"shape":"OptionGroupQuotaExceededFault"} + {"shape":"DBClusterNotFoundFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"DBClusterSnapshotAlreadyExistsFault"}, + {"shape":"SnapshotQuotaExceededFault"}, + {"shape":"InvalidDBClusterSnapshotStateFault"} + ] + }, + "DeleteDBClusterParameterGroup":{ + "name":"DeleteDBClusterParameterGroup", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DeleteDBClusterParameterGroupMessage"}, + "errors":[ + {"shape":"InvalidDBParameterGroupStateFault"}, + {"shape":"DBParameterGroupNotFoundFault"} + ] + }, + "DeleteDBClusterSnapshot":{ + "name":"DeleteDBClusterSnapshot", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DeleteDBClusterSnapshotMessage"}, + "output":{ + "shape":"DeleteDBClusterSnapshotResult", + "resultWrapper":"DeleteDBClusterSnapshotResult" + }, + "errors":[ + {"shape":"InvalidDBClusterSnapshotStateFault"}, + {"shape":"DBClusterSnapshotNotFoundFault"} ] }, "DeleteDBInstance":{ @@ -291,7 +339,8 @@ {"shape":"DBInstanceNotFoundFault"}, {"shape":"InvalidDBInstanceStateFault"}, {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"SnapshotQuotaExceededFault"} + {"shape":"SnapshotQuotaExceededFault"}, + {"shape":"InvalidDBClusterStateFault"} ] }, "DeleteDBParameterGroup":{ @@ -306,34 +355,6 @@ {"shape":"DBParameterGroupNotFoundFault"} ] }, - "DeleteDBSecurityGroup":{ - "name":"DeleteDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSecurityGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DeleteDBSnapshot":{ - "name":"DeleteDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSnapshotMessage"}, - "output":{ - "shape":"DeleteDBSnapshotResult", - "resultWrapper":"DeleteDBSnapshotResult" - }, - "errors":[ - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, "DeleteDBSubnetGroup":{ "name":"DeleteDBSubnetGroup", "http":{ @@ -363,16 +384,79 @@ {"shape":"InvalidEventSubscriptionStateFault"} ] }, - "DeleteOptionGroup":{ - "name":"DeleteOptionGroup", + "DescribeDBClusterParameterGroups":{ + "name":"DescribeDBClusterParameterGroups", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"DeleteOptionGroupMessage"}, + "input":{"shape":"DescribeDBClusterParameterGroupsMessage"}, + "output":{ + "shape":"DBClusterParameterGroupsMessage", + "resultWrapper":"DescribeDBClusterParameterGroupsResult" + }, "errors":[ - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"InvalidOptionGroupStateFault"} + {"shape":"DBParameterGroupNotFoundFault"} + ] + }, + "DescribeDBClusterParameters":{ + "name":"DescribeDBClusterParameters", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDBClusterParametersMessage"}, + "output":{ + "shape":"DBClusterParameterGroupDetails", + "resultWrapper":"DescribeDBClusterParametersResult" + }, + "errors":[ + {"shape":"DBParameterGroupNotFoundFault"} + ] + }, + "DescribeDBClusterSnapshotAttributes":{ + "name":"DescribeDBClusterSnapshotAttributes", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDBClusterSnapshotAttributesMessage"}, + "output":{ + "shape":"DescribeDBClusterSnapshotAttributesResult", + "resultWrapper":"DescribeDBClusterSnapshotAttributesResult" + }, + "errors":[ + {"shape":"DBClusterSnapshotNotFoundFault"} + ] + }, + "DescribeDBClusterSnapshots":{ + "name":"DescribeDBClusterSnapshots", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDBClusterSnapshotsMessage"}, + "output":{ + "shape":"DBClusterSnapshotMessage", + "resultWrapper":"DescribeDBClusterSnapshotsResult" + }, + "errors":[ + {"shape":"DBClusterSnapshotNotFoundFault"} + ] + }, + "DescribeDBClusters":{ + "name":"DescribeDBClusters", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDBClustersMessage"}, + "output":{ + "shape":"DBClusterMessage", + "resultWrapper":"DescribeDBClustersResult" + }, + "errors":[ + {"shape":"DBClusterNotFoundFault"} ] }, "DescribeDBEngineVersions":{ @@ -402,21 +486,6 @@ {"shape":"DBInstanceNotFoundFault"} ] }, - "DescribeDBLogFiles":{ - "name":"DescribeDBLogFiles", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBLogFilesMessage"}, - "output":{ - "shape":"DescribeDBLogFilesResponse", - "resultWrapper":"DescribeDBLogFilesResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"} - ] - }, "DescribeDBParameterGroups":{ "name":"DescribeDBParameterGroups", "http":{ @@ -447,36 +516,6 @@ {"shape":"DBParameterGroupNotFoundFault"} ] }, - "DescribeDBSecurityGroups":{ - "name":"DescribeDBSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSecurityGroupsMessage"}, - "output":{ - "shape":"DBSecurityGroupMessage", - "resultWrapper":"DescribeDBSecurityGroupsResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DescribeDBSnapshots":{ - "name":"DescribeDBSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSnapshotsMessage"}, - "output":{ - "shape":"DBSnapshotMessage", - "resultWrapper":"DescribeDBSnapshotsResult" - }, - "errors":[ - {"shape":"DBSnapshotNotFoundFault"} - ] - }, "DescribeDBSubnetGroups":{ "name":"DescribeDBSubnetGroups", "http":{ @@ -492,6 +531,18 @@ {"shape":"DBSubnetGroupNotFoundFault"} ] }, + "DescribeEngineDefaultClusterParameters":{ + "name":"DescribeEngineDefaultClusterParameters", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeEngineDefaultClusterParametersMessage"}, + "output":{ + "shape":"DescribeEngineDefaultClusterParametersResult", + "resultWrapper":"DescribeEngineDefaultClusterParametersResult" + } + }, "DescribeEngineDefaultParameters":{ "name":"DescribeEngineDefaultParameters", "http":{ @@ -543,33 +594,6 @@ "resultWrapper":"DescribeEventsResult" } }, - "DescribeOptionGroupOptions":{ - "name":"DescribeOptionGroupOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupOptionsMessage"}, - "output":{ - "shape":"OptionGroupOptionsMessage", - "resultWrapper":"DescribeOptionGroupOptionsResult" - } - }, - "DescribeOptionGroups":{ - "name":"DescribeOptionGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupsMessage"}, - "output":{ - "shape":"OptionGroups", - "resultWrapper":"DescribeOptionGroupsResult" - }, - "errors":[ - {"shape":"OptionGroupNotFoundFault"} - ] - }, "DescribeOrderableDBInstanceOptions":{ "name":"DescribeOrderableDBInstanceOptions", "http":{ @@ -582,50 +606,52 @@ "resultWrapper":"DescribeOrderableDBInstanceOptionsResult" } }, - "DescribeReservedDBInstances":{ - "name":"DescribeReservedDBInstances", + "DescribePendingMaintenanceActions":{ + "name":"DescribePendingMaintenanceActions", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"DescribeReservedDBInstancesMessage"}, + "input":{"shape":"DescribePendingMaintenanceActionsMessage"}, "output":{ - "shape":"ReservedDBInstanceMessage", - "resultWrapper":"DescribeReservedDBInstancesResult" + "shape":"PendingMaintenanceActionsMessage", + "resultWrapper":"DescribePendingMaintenanceActionsResult" }, "errors":[ - {"shape":"ReservedDBInstanceNotFoundFault"} + {"shape":"ResourceNotFoundFault"} ] }, - "DescribeReservedDBInstancesOfferings":{ - "name":"DescribeReservedDBInstancesOfferings", + "DescribeValidDBInstanceModifications":{ + "name":"DescribeValidDBInstanceModifications", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"DescribeReservedDBInstancesOfferingsMessage"}, + "input":{"shape":"DescribeValidDBInstanceModificationsMessage"}, "output":{ - "shape":"ReservedDBInstancesOfferingMessage", - "resultWrapper":"DescribeReservedDBInstancesOfferingsResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"} - ] - }, - "DownloadDBLogFilePortion":{ - "name":"DownloadDBLogFilePortion", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DownloadDBLogFilePortionMessage"}, - "output":{ - "shape":"DownloadDBLogFilePortionDetails", - "resultWrapper":"DownloadDBLogFilePortionResult" + "shape":"DescribeValidDBInstanceModificationsResult", + "resultWrapper":"DescribeValidDBInstanceModificationsResult" }, "errors":[ {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBLogFileNotFoundFault"} + {"shape":"InvalidDBInstanceStateFault"} + ] + }, + "FailoverDBCluster":{ + "name":"FailoverDBCluster", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"FailoverDBClusterMessage"}, + "output":{ + "shape":"FailoverDBClusterResult", + "resultWrapper":"FailoverDBClusterResult" + }, + "errors":[ + {"shape":"DBClusterNotFoundFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"InvalidDBInstanceStateFault"} ] }, "ListTagsForResource":{ @@ -641,7 +667,66 @@ }, "errors":[ {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} + {"shape":"DBSnapshotNotFoundFault"}, + {"shape":"DBClusterNotFoundFault"} + ] + }, + "ModifyDBCluster":{ + "name":"ModifyDBCluster", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ModifyDBClusterMessage"}, + "output":{ + "shape":"ModifyDBClusterResult", + "resultWrapper":"ModifyDBClusterResult" + }, + "errors":[ + {"shape":"DBClusterNotFoundFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"StorageQuotaExceededFault"}, + {"shape":"DBSubnetGroupNotFoundFault"}, + {"shape":"InvalidVPCNetworkStateFault"}, + {"shape":"InvalidDBSubnetGroupStateFault"}, + {"shape":"InvalidSubnet"}, + {"shape":"DBClusterParameterGroupNotFoundFault"}, + {"shape":"InvalidDBSecurityGroupStateFault"}, + {"shape":"InvalidDBInstanceStateFault"}, + {"shape":"DBClusterAlreadyExistsFault"} + ] + }, + "ModifyDBClusterParameterGroup":{ + "name":"ModifyDBClusterParameterGroup", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ModifyDBClusterParameterGroupMessage"}, + "output":{ + "shape":"DBClusterParameterGroupNameMessage", + "resultWrapper":"ModifyDBClusterParameterGroupResult" + }, + "errors":[ + {"shape":"DBParameterGroupNotFoundFault"}, + {"shape":"InvalidDBParameterGroupStateFault"} + ] + }, + "ModifyDBClusterSnapshotAttribute":{ + "name":"ModifyDBClusterSnapshotAttribute", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ModifyDBClusterSnapshotAttributeMessage"}, + "output":{ + "shape":"ModifyDBClusterSnapshotAttributeResult", + "resultWrapper":"ModifyDBClusterSnapshotAttributeResult" + }, + "errors":[ + {"shape":"DBClusterSnapshotNotFoundFault"}, + {"shape":"InvalidDBClusterSnapshotStateFault"}, + {"shape":"SharedSnapshotQuotaExceededFault"} ] }, "ModifyDBInstance":{ @@ -669,7 +754,9 @@ {"shape":"OptionGroupNotFoundFault"}, {"shape":"DBUpgradeDependencyFailureFault"}, {"shape":"StorageTypeNotSupportedFault"}, - {"shape":"AuthorizationNotFoundFault"} + {"shape":"AuthorizationNotFoundFault"}, + {"shape":"CertificateNotFoundFault"}, + {"shape":"DomainNotFoundFault"} ] }, "ModifyDBParameterGroup":{ @@ -727,53 +814,20 @@ {"shape":"SubscriptionCategoryNotFoundFault"} ] }, - "ModifyOptionGroup":{ - "name":"ModifyOptionGroup", + "PromoteReadReplicaDBCluster":{ + "name":"PromoteReadReplicaDBCluster", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"ModifyOptionGroupMessage"}, + "input":{"shape":"PromoteReadReplicaDBClusterMessage"}, "output":{ - "shape":"ModifyOptionGroupResult", - "resultWrapper":"ModifyOptionGroupResult" + "shape":"PromoteReadReplicaDBClusterResult", + "resultWrapper":"PromoteReadReplicaDBClusterResult" }, "errors":[ - {"shape":"InvalidOptionGroupStateFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "PromoteReadReplica":{ - "name":"PromoteReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PromoteReadReplicaMessage"}, - "output":{ - "shape":"PromoteReadReplicaResult", - "resultWrapper":"PromoteReadReplicaResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "PurchaseReservedDBInstancesOffering":{ - "name":"PurchaseReservedDBInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedDBInstancesOfferingMessage"}, - "output":{ - "shape":"PurchaseReservedDBInstancesOfferingResult", - "resultWrapper":"PurchaseReservedDBInstancesOfferingResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"}, - {"shape":"ReservedDBInstanceAlreadyExistsFault"}, - {"shape":"ReservedDBInstanceQuotaExceededFault"} + {"shape":"DBClusterNotFoundFault"}, + {"shape":"InvalidDBClusterStateFault"} ] }, "RebootDBInstance":{ @@ -792,6 +846,19 @@ {"shape":"DBInstanceNotFoundFault"} ] }, + "RemoveRoleFromDBCluster":{ + "name":"RemoveRoleFromDBCluster", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"RemoveRoleFromDBClusterMessage"}, + "errors":[ + {"shape":"DBClusterNotFoundFault"}, + {"shape":"DBClusterRoleNotFoundFault"}, + {"shape":"InvalidDBClusterStateFault"} + ] + }, "RemoveSourceIdentifierFromSubscription":{ "name":"RemoveSourceIdentifierFromSubscription", "http":{ @@ -817,7 +884,24 @@ "input":{"shape":"RemoveTagsFromResourceMessage"}, "errors":[ {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} + {"shape":"DBSnapshotNotFoundFault"}, + {"shape":"DBClusterNotFoundFault"} + ] + }, + "ResetDBClusterParameterGroup":{ + "name":"ResetDBClusterParameterGroup", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ResetDBClusterParameterGroupMessage"}, + "output":{ + "shape":"DBClusterParameterGroupNameMessage", + "resultWrapper":"ResetDBClusterParameterGroupResult" + }, + "errors":[ + {"shape":"InvalidDBParameterGroupStateFault"}, + {"shape":"DBParameterGroupNotFoundFault"} ] }, "ResetDBParameterGroup":{ @@ -836,84 +920,80 @@ {"shape":"DBParameterGroupNotFoundFault"} ] }, - "RestoreDBInstanceFromDBSnapshot":{ - "name":"RestoreDBInstanceFromDBSnapshot", + "RestoreDBClusterFromSnapshot":{ + "name":"RestoreDBClusterFromSnapshot", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"RestoreDBInstanceFromDBSnapshotMessage"}, + "input":{"shape":"RestoreDBClusterFromSnapshotMessage"}, "output":{ - "shape":"RestoreDBInstanceFromDBSnapshotResult", - "resultWrapper":"RestoreDBInstanceFromDBSnapshotResult" + "shape":"RestoreDBClusterFromSnapshotResult", + "resultWrapper":"RestoreDBClusterFromSnapshotResult" }, "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, + {"shape":"DBClusterAlreadyExistsFault"}, + {"shape":"DBClusterQuotaExceededFault"}, + {"shape":"StorageQuotaExceededFault"}, + {"shape":"DBSubnetGroupNotFoundFault"}, {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, + {"shape":"DBClusterSnapshotNotFoundFault"}, + {"shape":"InsufficientDBClusterCapacityFault"}, + {"shape":"InsufficientStorageClusterCapacityFault"}, {"shape":"InvalidDBSnapshotStateFault"}, + {"shape":"InvalidDBClusterSnapshotStateFault"}, {"shape":"StorageQuotaExceededFault"}, {"shape":"InvalidVPCNetworkStateFault"}, {"shape":"InvalidRestoreFault"}, {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, {"shape":"OptionGroupNotFoundFault"}, - {"shape":"StorageTypeNotSupportedFault"}, - {"shape":"AuthorizationNotFoundFault"} + {"shape":"KMSKeyNotAccessibleFault"} ] }, - "RestoreDBInstanceToPointInTime":{ - "name":"RestoreDBInstanceToPointInTime", + "RestoreDBClusterToPointInTime":{ + "name":"RestoreDBClusterToPointInTime", "http":{ "method":"POST", "requestUri":"/" }, - "input":{"shape":"RestoreDBInstanceToPointInTimeMessage"}, + "input":{"shape":"RestoreDBClusterToPointInTimeMessage"}, "output":{ - "shape":"RestoreDBInstanceToPointInTimeResult", - "resultWrapper":"RestoreDBInstanceToPointInTimeResult" + "shape":"RestoreDBClusterToPointInTimeResult", + "resultWrapper":"RestoreDBClusterToPointInTimeResult" }, "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"PointInTimeRestoreNotEnabledFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, + {"shape":"DBClusterAlreadyExistsFault"}, + {"shape":"DBClusterNotFoundFault"}, + {"shape":"DBClusterQuotaExceededFault"}, + {"shape":"DBClusterSnapshotNotFoundFault"}, {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, + {"shape":"InsufficientDBClusterCapacityFault"}, + {"shape":"InsufficientStorageClusterCapacityFault"}, + {"shape":"InvalidDBClusterSnapshotStateFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"InvalidDBSnapshotStateFault"}, + {"shape":"InvalidRestoreFault"}, {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, + {"shape":"InvalidVPCNetworkStateFault"}, + {"shape":"KMSKeyNotAccessibleFault"}, {"shape":"OptionGroupNotFoundFault"}, - {"shape":"StorageTypeNotSupportedFault"}, - {"shape":"AuthorizationNotFoundFault"} - ] - }, - "RevokeDBSecurityGroupIngress":{ - "name":"RevokeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"RevokeDBSecurityGroupIngressResult", - "resultWrapper":"RevokeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"AuthorizationNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"} + {"shape":"StorageQuotaExceededFault"} ] } }, "shapes":{ + "AddRoleToDBClusterMessage":{ + "type":"structure", + "required":[ + "DBClusterIdentifier", + "RoleArn" + ], + "members":{ + "DBClusterIdentifier":{"shape":"String"}, + "RoleArn":{"shape":"String"} + } + }, "AddSourceIdentifierToSubscriptionMessage":{ "type":"structure", "required":[ @@ -949,16 +1029,31 @@ "pending-reboot" ] }, - "AuthorizationAlreadyExistsFault":{ + "ApplyPendingMaintenanceActionMessage":{ + "type":"structure", + "required":[ + "ResourceIdentifier", + "ApplyAction", + "OptInType" + ], + "members":{ + "ResourceIdentifier":{"shape":"String"}, + "ApplyAction":{"shape":"String"}, + "OptInType":{"shape":"String"} + } + }, + "ApplyPendingMaintenanceActionResult":{ "type":"structure", "members":{ - }, - "error":{ - "code":"AuthorizationAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true + "ResourcePendingMaintenanceActions":{"shape":"ResourcePendingMaintenanceActions"} + } + }, + "AttributeValueList":{ + "type":"list", + "member":{ + "shape":"String", + "locationName":"AttributeValue" + } }, "AuthorizationNotFoundFault":{ "type":"structure", @@ -971,34 +1066,6 @@ }, "exception":true }, - "AuthorizationQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "AuthorizeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, "AvailabilityZone":{ "type":"structure", "members":{ @@ -1013,8 +1080,26 @@ "locationName":"AvailabilityZone" } }, + "AvailabilityZones":{ + "type":"list", + "member":{ + "shape":"String", + "locationName":"AvailabilityZone" + } + }, "Boolean":{"type":"boolean"}, "BooleanOptional":{"type":"boolean"}, + "CertificateNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"CertificateNotFound", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, "CharacterSet":{ "type":"structure", "members":{ @@ -1022,6 +1107,54 @@ "CharacterSetDescription":{"shape":"String"} } }, + "CloudwatchLogsExportConfiguration":{ + "type":"structure", + "members":{ + "EnableLogTypes":{"shape":"LogTypeList"}, + "DisableLogTypes":{"shape":"LogTypeList"} + } + }, + "CopyDBClusterParameterGroupMessage":{ + "type":"structure", + "required":[ + "SourceDBClusterParameterGroupIdentifier", + "TargetDBClusterParameterGroupIdentifier", + "TargetDBClusterParameterGroupDescription" + ], + "members":{ + "SourceDBClusterParameterGroupIdentifier":{"shape":"String"}, + "TargetDBClusterParameterGroupIdentifier":{"shape":"String"}, + "TargetDBClusterParameterGroupDescription":{"shape":"String"}, + "Tags":{"shape":"TagList"} + } + }, + "CopyDBClusterParameterGroupResult":{ + "type":"structure", + "members":{ + "DBClusterParameterGroup":{"shape":"DBClusterParameterGroup"} + } + }, + "CopyDBClusterSnapshotMessage":{ + "type":"structure", + "required":[ + "SourceDBClusterSnapshotIdentifier", + "TargetDBClusterSnapshotIdentifier" + ], + "members":{ + "SourceDBClusterSnapshotIdentifier":{"shape":"String"}, + "TargetDBClusterSnapshotIdentifier":{"shape":"String"}, + "KmsKeyId":{"shape":"String"}, + "PreSignedUrl":{"shape":"String"}, + "CopyTags":{"shape":"BooleanOptional"}, + "Tags":{"shape":"TagList"} + } + }, + "CopyDBClusterSnapshotResult":{ + "type":"structure", + "members":{ + "DBClusterSnapshot":{"shape":"DBClusterSnapshot"} + } + }, "CopyDBParameterGroupMessage":{ "type":"structure", "required":[ @@ -1042,53 +1175,87 @@ "DBParameterGroup":{"shape":"DBParameterGroup"} } }, - "CopyDBSnapshotMessage":{ + "CreateDBClusterMessage":{ "type":"structure", "required":[ - "SourceDBSnapshotIdentifier", - "TargetDBSnapshotIdentifier" + "DBClusterIdentifier", + "Engine" ], "members":{ - "SourceDBSnapshotIdentifier":{"shape":"String"}, - "TargetDBSnapshotIdentifier":{"shape":"String"}, + "AvailabilityZones":{"shape":"AvailabilityZones"}, + "BackupRetentionPeriod":{"shape":"IntegerOptional"}, + "CharacterSetName":{"shape":"String"}, + "DatabaseName":{"shape":"String"}, + "DBClusterIdentifier":{"shape":"String"}, + "DBClusterParameterGroupName":{"shape":"String"}, + "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, + "DBSubnetGroupName":{"shape":"String"}, + "Engine":{"shape":"String"}, + "EngineVersion":{"shape":"String"}, + "Port":{"shape":"IntegerOptional"}, + "MasterUsername":{"shape":"String"}, + "MasterUserPassword":{"shape":"String"}, + "OptionGroupName":{"shape":"String"}, + "PreferredBackupWindow":{"shape":"String"}, + "PreferredMaintenanceWindow":{"shape":"String"}, + "ReplicationSourceIdentifier":{"shape":"String"}, + "Tags":{"shape":"TagList"}, + "StorageEncrypted":{"shape":"BooleanOptional"}, + "KmsKeyId":{"shape":"String"}, + "PreSignedUrl":{"shape":"String"}, + "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"} + } + }, + "CreateDBClusterParameterGroupMessage":{ + "type":"structure", + "required":[ + "DBClusterParameterGroupName", + "DBParameterGroupFamily", + "Description" + ], + "members":{ + "DBClusterParameterGroupName":{"shape":"String"}, + "DBParameterGroupFamily":{"shape":"String"}, + "Description":{"shape":"String"}, "Tags":{"shape":"TagList"} } }, - "CopyDBSnapshotResult":{ + "CreateDBClusterParameterGroupResult":{ "type":"structure", "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} + "DBClusterParameterGroup":{"shape":"DBClusterParameterGroup"} } }, - "CopyOptionGroupMessage":{ + "CreateDBClusterResult":{ + "type":"structure", + "members":{ + "DBCluster":{"shape":"DBCluster"} + } + }, + "CreateDBClusterSnapshotMessage":{ "type":"structure", "required":[ - "SourceOptionGroupIdentifier", - "TargetOptionGroupIdentifier", - "TargetOptionGroupDescription" + "DBClusterSnapshotIdentifier", + "DBClusterIdentifier" ], "members":{ - "SourceOptionGroupIdentifier":{"shape":"String"}, - "TargetOptionGroupIdentifier":{"shape":"String"}, - "TargetOptionGroupDescription":{"shape":"String"}, + "DBClusterSnapshotIdentifier":{"shape":"String"}, + "DBClusterIdentifier":{"shape":"String"}, "Tags":{"shape":"TagList"} } }, - "CopyOptionGroupResult":{ + "CreateDBClusterSnapshotResult":{ "type":"structure", "members":{ - "OptionGroup":{"shape":"OptionGroup"} + "DBClusterSnapshot":{"shape":"DBClusterSnapshot"} } }, "CreateDBInstanceMessage":{ "type":"structure", "required":[ "DBInstanceIdentifier", - "AllocatedStorage", "DBInstanceClass", - "Engine", - "MasterUsername", - "MasterUserPassword" + "Engine" ], "members":{ "DBName":{"shape":"String"}, @@ -1114,38 +1281,28 @@ "Iops":{"shape":"IntegerOptional"}, "OptionGroupName":{"shape":"String"}, "CharacterSetName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, + "PubliclyAccessible":{ + "shape":"BooleanOptional", + "deprecated":true + }, "Tags":{"shape":"TagList"}, + "DBClusterIdentifier":{"shape":"String"}, "StorageType":{"shape":"String"}, "TdeCredentialArn":{"shape":"String"}, - "TdeCredentialPassword":{"shape":"String"} - } - }, - "CreateDBInstanceReadReplicaMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "SourceDBInstanceIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SourceDBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "Tags":{"shape":"TagList"}, - "DBSubnetGroupName":{"shape":"String"}, - "StorageType":{"shape":"String"} - } - }, - "CreateDBInstanceReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} + "TdeCredentialPassword":{"shape":"String"}, + "StorageEncrypted":{"shape":"BooleanOptional"}, + "KmsKeyId":{"shape":"String"}, + "Domain":{"shape":"String"}, + "CopyTagsToSnapshot":{"shape":"BooleanOptional"}, + "MonitoringInterval":{"shape":"IntegerOptional"}, + "MonitoringRoleArn":{"shape":"String"}, + "DomainIAMRoleName":{"shape":"String"}, + "PromotionTier":{"shape":"IntegerOptional"}, + "Timezone":{"shape":"String"}, + "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, + "EnablePerformanceInsights":{"shape":"BooleanOptional"}, + "PerformanceInsightsKMSKeyId":{"shape":"String"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} } }, "CreateDBInstanceResult":{ @@ -1174,42 +1331,6 @@ "DBParameterGroup":{"shape":"DBParameterGroup"} } }, - "CreateDBSecurityGroupMessage":{ - "type":"structure", - "required":[ - "DBSecurityGroupName", - "DBSecurityGroupDescription" - ], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBSecurityGroupResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "CreateDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBSnapshotIdentifier", - "DBInstanceIdentifier" - ], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, "CreateDBSubnetGroupMessage":{ "type":"structure", "required":[ @@ -1252,28 +1373,304 @@ "EventSubscription":{"shape":"EventSubscription"} } }, - "CreateOptionGroupMessage":{ + "DBCluster":{ "type":"structure", - "required":[ - "OptionGroupName", - "EngineName", - "MajorEngineVersion", - "OptionGroupDescription" - ], "members":{ - "OptionGroupName":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"}, - "Tags":{"shape":"TagList"} + "AllocatedStorage":{"shape":"IntegerOptional"}, + "AvailabilityZones":{"shape":"AvailabilityZones"}, + "BackupRetentionPeriod":{"shape":"IntegerOptional"}, + "CharacterSetName":{"shape":"String"}, + "DatabaseName":{"shape":"String"}, + "DBClusterIdentifier":{"shape":"String"}, + "DBClusterParameterGroup":{"shape":"String"}, + "DBSubnetGroup":{"shape":"String"}, + "Status":{"shape":"String"}, + "PercentProgress":{"shape":"String"}, + "EarliestRestorableTime":{"shape":"TStamp"}, + "Endpoint":{"shape":"String"}, + "ReaderEndpoint":{"shape":"String"}, + "MultiAZ":{"shape":"Boolean"}, + "Engine":{"shape":"String"}, + "EngineVersion":{"shape":"String"}, + "LatestRestorableTime":{"shape":"TStamp"}, + "Port":{"shape":"IntegerOptional"}, + "MasterUsername":{"shape":"String"}, + "DBClusterOptionGroupMemberships":{"shape":"DBClusterOptionGroupMemberships"}, + "PreferredBackupWindow":{"shape":"String"}, + "PreferredMaintenanceWindow":{"shape":"String"}, + "ReplicationSourceIdentifier":{"shape":"String"}, + "ReadReplicaIdentifiers":{"shape":"ReadReplicaIdentifierList"}, + "DBClusterMembers":{"shape":"DBClusterMemberList"}, + "VpcSecurityGroups":{"shape":"VpcSecurityGroupMembershipList"}, + "HostedZoneId":{"shape":"String"}, + "StorageEncrypted":{"shape":"Boolean"}, + "KmsKeyId":{"shape":"String"}, + "DbClusterResourceId":{"shape":"String"}, + "DBClusterArn":{"shape":"String"}, + "AssociatedRoles":{"shape":"DBClusterRoles"}, + "IAMDatabaseAuthenticationEnabled":{"shape":"Boolean"}, + "CloneGroupId":{"shape":"String"}, + "ClusterCreateTime":{"shape":"TStamp"} + }, + "wrapper":true + }, + "DBClusterAlreadyExistsFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterAlreadyExistsFault", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, + "DBClusterList":{ + "type":"list", + "member":{ + "shape":"DBCluster", + "locationName":"DBCluster" } }, - "CreateOptionGroupResult":{ + "DBClusterMember":{ "type":"structure", "members":{ - "OptionGroup":{"shape":"OptionGroup"} + "DBInstanceIdentifier":{"shape":"String"}, + "IsClusterWriter":{"shape":"Boolean"}, + "DBClusterParameterGroupStatus":{"shape":"String"}, + "PromotionTier":{"shape":"IntegerOptional"} + }, + "wrapper":true + }, + "DBClusterMemberList":{ + "type":"list", + "member":{ + "shape":"DBClusterMember", + "locationName":"DBClusterMember" } }, + "DBClusterMessage":{ + "type":"structure", + "members":{ + "Marker":{"shape":"String"}, + "DBClusters":{"shape":"DBClusterList"} + } + }, + "DBClusterNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterNotFoundFault", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, + "DBClusterOptionGroupMemberships":{ + "type":"list", + "member":{ + "shape":"DBClusterOptionGroupStatus", + "locationName":"DBClusterOptionGroup" + } + }, + "DBClusterOptionGroupStatus":{ + "type":"structure", + "members":{ + "DBClusterOptionGroupName":{"shape":"String"}, + "Status":{"shape":"String"} + } + }, + "DBClusterParameterGroup":{ + "type":"structure", + "members":{ + "DBClusterParameterGroupName":{"shape":"String"}, + "DBParameterGroupFamily":{"shape":"String"}, + "Description":{"shape":"String"}, + "DBClusterParameterGroupArn":{"shape":"String"} + }, + "wrapper":true + }, + "DBClusterParameterGroupDetails":{ + "type":"structure", + "members":{ + "Parameters":{"shape":"ParametersList"}, + "Marker":{"shape":"String"} + } + }, + "DBClusterParameterGroupList":{ + "type":"list", + "member":{ + "shape":"DBClusterParameterGroup", + "locationName":"DBClusterParameterGroup" + } + }, + "DBClusterParameterGroupNameMessage":{ + "type":"structure", + "members":{ + "DBClusterParameterGroupName":{"shape":"String"} + } + }, + "DBClusterParameterGroupNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterParameterGroupNotFound", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, + "DBClusterParameterGroupsMessage":{ + "type":"structure", + "members":{ + "Marker":{"shape":"String"}, + "DBClusterParameterGroups":{"shape":"DBClusterParameterGroupList"} + } + }, + "DBClusterQuotaExceededFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterQuotaExceededFault", + "httpStatusCode":403, + "senderFault":true + }, + "exception":true + }, + "DBClusterRole":{ + "type":"structure", + "members":{ + "RoleArn":{"shape":"String"}, + "Status":{"shape":"String"} + } + }, + "DBClusterRoleAlreadyExistsFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterRoleAlreadyExists", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, + "DBClusterRoleNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterRoleNotFound", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, + "DBClusterRoleQuotaExceededFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterRoleQuotaExceeded", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, + "DBClusterRoles":{ + "type":"list", + "member":{ + "shape":"DBClusterRole", + "locationName":"DBClusterRole" + } + }, + "DBClusterSnapshot":{ + "type":"structure", + "members":{ + "AvailabilityZones":{"shape":"AvailabilityZones"}, + "DBClusterSnapshotIdentifier":{"shape":"String"}, + "DBClusterIdentifier":{"shape":"String"}, + "SnapshotCreateTime":{"shape":"TStamp"}, + "Engine":{"shape":"String"}, + "AllocatedStorage":{"shape":"Integer"}, + "Status":{"shape":"String"}, + "Port":{"shape":"Integer"}, + "VpcId":{"shape":"String"}, + "ClusterCreateTime":{"shape":"TStamp"}, + "MasterUsername":{"shape":"String"}, + "EngineVersion":{"shape":"String"}, + "LicenseModel":{"shape":"String"}, + "SnapshotType":{"shape":"String"}, + "PercentProgress":{"shape":"Integer"}, + "StorageEncrypted":{"shape":"Boolean"}, + "KmsKeyId":{"shape":"String"}, + "DBClusterSnapshotArn":{"shape":"String"}, + "SourceDBClusterSnapshotArn":{"shape":"String"}, + "IAMDatabaseAuthenticationEnabled":{"shape":"Boolean"} + }, + "wrapper":true + }, + "DBClusterSnapshotAlreadyExistsFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterSnapshotAlreadyExistsFault", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, + "DBClusterSnapshotAttribute":{ + "type":"structure", + "members":{ + "AttributeName":{"shape":"String"}, + "AttributeValues":{"shape":"AttributeValueList"} + } + }, + "DBClusterSnapshotAttributeList":{ + "type":"list", + "member":{ + "shape":"DBClusterSnapshotAttribute", + "locationName":"DBClusterSnapshotAttribute" + } + }, + "DBClusterSnapshotAttributesResult":{ + "type":"structure", + "members":{ + "DBClusterSnapshotIdentifier":{"shape":"String"}, + "DBClusterSnapshotAttributes":{"shape":"DBClusterSnapshotAttributeList"} + }, + "wrapper":true + }, + "DBClusterSnapshotList":{ + "type":"list", + "member":{ + "shape":"DBClusterSnapshot", + "locationName":"DBClusterSnapshot" + } + }, + "DBClusterSnapshotMessage":{ + "type":"structure", + "members":{ + "Marker":{"shape":"String"}, + "DBClusterSnapshots":{"shape":"DBClusterSnapshotList"} + } + }, + "DBClusterSnapshotNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DBClusterSnapshotNotFoundFault", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, "DBEngineVersion":{ "type":"structure", "members":{ @@ -1283,7 +1680,12 @@ "DBEngineDescription":{"shape":"String"}, "DBEngineVersionDescription":{"shape":"String"}, "DefaultCharacterSet":{"shape":"CharacterSet"}, - "SupportedCharacterSets":{"shape":"SupportedCharacterSetsList"} + "SupportedCharacterSets":{"shape":"SupportedCharacterSetsList"}, + "ValidUpgradeTarget":{"shape":"ValidUpgradeTargetList"}, + "SupportedTimezones":{"shape":"SupportedTimezonesList"}, + "ExportableLogTypes":{"shape":"LogTypeList"}, + "SupportsLogExportsToCloudwatchLogs":{"shape":"Boolean"}, + "SupportsReadReplica":{"shape":"Boolean"} } }, "DBEngineVersionList":{ @@ -1327,15 +1729,37 @@ "AutoMinorVersionUpgrade":{"shape":"Boolean"}, "ReadReplicaSourceDBInstanceIdentifier":{"shape":"String"}, "ReadReplicaDBInstanceIdentifiers":{"shape":"ReadReplicaDBInstanceIdentifierList"}, + "ReadReplicaDBClusterIdentifiers":{"shape":"ReadReplicaDBClusterIdentifierList"}, "LicenseModel":{"shape":"String"}, "Iops":{"shape":"IntegerOptional"}, "OptionGroupMemberships":{"shape":"OptionGroupMembershipList"}, "CharacterSetName":{"shape":"String"}, "SecondaryAvailabilityZone":{"shape":"String"}, - "PubliclyAccessible":{"shape":"Boolean"}, + "PubliclyAccessible":{ + "shape":"Boolean", + "deprecated":true + }, "StatusInfos":{"shape":"DBInstanceStatusInfoList"}, "StorageType":{"shape":"String"}, - "TdeCredentialArn":{"shape":"String"} + "TdeCredentialArn":{"shape":"String"}, + "DbInstancePort":{"shape":"Integer"}, + "DBClusterIdentifier":{"shape":"String"}, + "StorageEncrypted":{"shape":"Boolean"}, + "KmsKeyId":{"shape":"String"}, + "DbiResourceId":{"shape":"String"}, + "CACertificateIdentifier":{"shape":"String"}, + "DomainMemberships":{"shape":"DomainMembershipList"}, + "CopyTagsToSnapshot":{"shape":"Boolean"}, + "MonitoringInterval":{"shape":"IntegerOptional"}, + "EnhancedMonitoringResourceArn":{"shape":"String"}, + "MonitoringRoleArn":{"shape":"String"}, + "PromotionTier":{"shape":"IntegerOptional"}, + "DBInstanceArn":{"shape":"String"}, + "Timezone":{"shape":"String"}, + "IAMDatabaseAuthenticationEnabled":{"shape":"Boolean"}, + "PerformanceInsightsEnabled":{"shape":"BooleanOptional"}, + "PerformanceInsightsKMSKeyId":{"shape":"String"}, + "EnabledCloudwatchLogsExports":{"shape":"LogTypeList"} }, "wrapper":true }, @@ -1391,23 +1815,13 @@ "locationName":"DBInstanceStatusInfo" } }, - "DBLogFileNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBLogFileNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, "DBParameterGroup":{ "type":"structure", "members":{ "DBParameterGroupName":{"shape":"String"}, "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"} + "Description":{"shape":"String"}, + "DBParameterGroupArn":{"shape":"String"} }, "wrapper":true }, @@ -1485,29 +1899,6 @@ "DBParameterGroups":{"shape":"DBParameterGroupList"} } }, - "DBSecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{"shape":"String"}, - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "EC2SecurityGroups":{"shape":"EC2SecurityGroupList"}, - "IPRanges":{"shape":"IPRangeList"} - }, - "wrapper":true - }, - "DBSecurityGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, "DBSecurityGroupMembership":{ "type":"structure", "members":{ @@ -1522,13 +1913,6 @@ "locationName":"DBSecurityGroup" } }, - "DBSecurityGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroups"} - } - }, "DBSecurityGroupNameList":{ "type":"list", "member":{ @@ -1547,61 +1931,6 @@ }, "exception":true }, - "DBSecurityGroupNotSupportedFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotSupported", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"QuotaExceeded.DBSecurityGroup", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroups":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroup", - "locationName":"DBSecurityGroup" - } - }, - "DBSnapshot":{ - "type":"structure", - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "SnapshotCreateTime":{"shape":"TStamp"}, - "Engine":{"shape":"String"}, - "AllocatedStorage":{"shape":"Integer"}, - "Status":{"shape":"String"}, - "Port":{"shape":"Integer"}, - "AvailabilityZone":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "MasterUsername":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "PercentProgress":{"shape":"Integer"}, - "SourceRegion":{"shape":"String"}, - "StorageType":{"shape":"String"}, - "TdeCredentialArn":{"shape":"String"} - }, - "wrapper":true - }, "DBSnapshotAlreadyExistsFault":{ "type":"structure", "members":{ @@ -1613,20 +1942,6 @@ }, "exception":true }, - "DBSnapshotList":{ - "type":"list", - "member":{ - "shape":"DBSnapshot", - "locationName":"DBSnapshot" - } - }, - "DBSnapshotMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSnapshots":{"shape":"DBSnapshotList"} - } - }, "DBSnapshotNotFoundFault":{ "type":"structure", "members":{ @@ -1645,7 +1960,8 @@ "DBSubnetGroupDescription":{"shape":"String"}, "VpcId":{"shape":"String"}, "SubnetGroupStatus":{"shape":"String"}, - "Subnets":{"shape":"SubnetList"} + "Subnets":{"shape":"SubnetList"}, + "DBSubnetGroupArn":{"shape":"String"} }, "wrapper":true }, @@ -1678,17 +1994,6 @@ "DBSubnetGroups":{"shape":"DBSubnetGroups"} } }, - "DBSubnetGroupNotAllowedFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupNotAllowedFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, "DBSubnetGroupNotFoundFault":{ "type":"structure", "members":{ @@ -1740,6 +2045,41 @@ }, "exception":true }, + "DeleteDBClusterMessage":{ + "type":"structure", + "required":["DBClusterIdentifier"], + "members":{ + "DBClusterIdentifier":{"shape":"String"}, + "SkipFinalSnapshot":{"shape":"Boolean"}, + "FinalDBSnapshotIdentifier":{"shape":"String"} + } + }, + "DeleteDBClusterParameterGroupMessage":{ + "type":"structure", + "required":["DBClusterParameterGroupName"], + "members":{ + "DBClusterParameterGroupName":{"shape":"String"} + } + }, + "DeleteDBClusterResult":{ + "type":"structure", + "members":{ + "DBCluster":{"shape":"DBCluster"} + } + }, + "DeleteDBClusterSnapshotMessage":{ + "type":"structure", + "required":["DBClusterSnapshotIdentifier"], + "members":{ + "DBClusterSnapshotIdentifier":{"shape":"String"} + } + }, + "DeleteDBClusterSnapshotResult":{ + "type":"structure", + "members":{ + "DBClusterSnapshot":{"shape":"DBClusterSnapshot"} + } + }, "DeleteDBInstanceMessage":{ "type":"structure", "required":["DBInstanceIdentifier"], @@ -1762,26 +2102,6 @@ "DBParameterGroupName":{"shape":"String"} } }, - "DeleteDBSecurityGroupMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"} - } - }, - "DeleteDBSnapshotMessage":{ - "type":"structure", - "required":["DBSnapshotIdentifier"], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, "DeleteDBSubnetGroupMessage":{ "type":"structure", "required":["DBSubnetGroupName"], @@ -1802,11 +2122,59 @@ "EventSubscription":{"shape":"EventSubscription"} } }, - "DeleteOptionGroupMessage":{ + "DescribeDBClusterParameterGroupsMessage":{ "type":"structure", - "required":["OptionGroupName"], "members":{ - "OptionGroupName":{"shape":"String"} + "DBClusterParameterGroupName":{"shape":"String"}, + "Filters":{"shape":"FilterList"}, + "MaxRecords":{"shape":"IntegerOptional"}, + "Marker":{"shape":"String"} + } + }, + "DescribeDBClusterParametersMessage":{ + "type":"structure", + "required":["DBClusterParameterGroupName"], + "members":{ + "DBClusterParameterGroupName":{"shape":"String"}, + "Source":{"shape":"String"}, + "Filters":{"shape":"FilterList"}, + "MaxRecords":{"shape":"IntegerOptional"}, + "Marker":{"shape":"String"} + } + }, + "DescribeDBClusterSnapshotAttributesMessage":{ + "type":"structure", + "required":["DBClusterSnapshotIdentifier"], + "members":{ + "DBClusterSnapshotIdentifier":{"shape":"String"} + } + }, + "DescribeDBClusterSnapshotAttributesResult":{ + "type":"structure", + "members":{ + "DBClusterSnapshotAttributesResult":{"shape":"DBClusterSnapshotAttributesResult"} + } + }, + "DescribeDBClusterSnapshotsMessage":{ + "type":"structure", + "members":{ + "DBClusterIdentifier":{"shape":"String"}, + "DBClusterSnapshotIdentifier":{"shape":"String"}, + "SnapshotType":{"shape":"String"}, + "Filters":{"shape":"FilterList"}, + "MaxRecords":{"shape":"IntegerOptional"}, + "Marker":{"shape":"String"}, + "IncludeShared":{"shape":"Boolean"}, + "IncludePublic":{"shape":"Boolean"} + } + }, + "DescribeDBClustersMessage":{ + "type":"structure", + "members":{ + "DBClusterIdentifier":{"shape":"String"}, + "Filters":{"shape":"FilterList"}, + "MaxRecords":{"shape":"IntegerOptional"}, + "Marker":{"shape":"String"} } }, "DescribeDBEngineVersionsMessage":{ @@ -1819,7 +2187,8 @@ "MaxRecords":{"shape":"IntegerOptional"}, "Marker":{"shape":"String"}, "DefaultOnly":{"shape":"Boolean"}, - "ListSupportedCharacterSets":{"shape":"BooleanOptional"} + "ListSupportedCharacterSets":{"shape":"BooleanOptional"}, + "ListSupportedTimezones":{"shape":"BooleanOptional"} } }, "DescribeDBInstancesMessage":{ @@ -1831,41 +2200,6 @@ "Marker":{"shape":"String"} } }, - "DescribeDBLogFilesDetails":{ - "type":"structure", - "members":{ - "LogFileName":{"shape":"String"}, - "LastWritten":{"shape":"Long"}, - "Size":{"shape":"Long"} - } - }, - "DescribeDBLogFilesList":{ - "type":"list", - "member":{ - "shape":"DescribeDBLogFilesDetails", - "locationName":"DescribeDBLogFilesDetails" - } - }, - "DescribeDBLogFilesMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "FilenameContains":{"shape":"String"}, - "FileLastWritten":{"shape":"Long"}, - "FileSize":{"shape":"Long"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBLogFilesResponse":{ - "type":"structure", - "members":{ - "DescribeDBLogFiles":{"shape":"DescribeDBLogFilesList"}, - "Marker":{"shape":"String"} - } - }, "DescribeDBParameterGroupsMessage":{ "type":"structure", "members":{ @@ -1886,26 +2220,6 @@ "Marker":{"shape":"String"} } }, - "DescribeDBSecurityGroupsMessage":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSnapshotsMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, "DescribeDBSubnetGroupsMessage":{ "type":"structure", "members":{ @@ -1915,6 +2229,22 @@ "Marker":{"shape":"String"} } }, + "DescribeEngineDefaultClusterParametersMessage":{ + "type":"structure", + "required":["DBParameterGroupFamily"], + "members":{ + "DBParameterGroupFamily":{"shape":"String"}, + "Filters":{"shape":"FilterList"}, + "MaxRecords":{"shape":"IntegerOptional"}, + "Marker":{"shape":"String"} + } + }, + "DescribeEngineDefaultClusterParametersResult":{ + "type":"structure", + "members":{ + "EngineDefaults":{"shape":"EngineDefaults"} + } + }, "DescribeEngineDefaultParametersMessage":{ "type":"structure", "required":["DBParameterGroupFamily"], @@ -1961,28 +2291,6 @@ "Marker":{"shape":"String"} } }, - "DescribeOptionGroupOptionsMessage":{ - "type":"structure", - "required":["EngineName"], - "members":{ - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupsMessage":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "Marker":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"} - } - }, "DescribeOrderableDBInstanceOptionsMessage":{ "type":"structure", "required":["Engine"], @@ -1997,78 +2305,77 @@ "Marker":{"shape":"String"} } }, - "DescribeReservedDBInstancesMessage":{ + "DescribePendingMaintenanceActionsMessage":{ "type":"structure", "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, + "ResourceIdentifier":{"shape":"String"}, "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesOfferingsMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "Double":{"type":"double"}, - "DownloadDBLogFilePortionDetails":{ - "type":"structure", - "members":{ - "LogFileData":{"shape":"String"}, "Marker":{"shape":"String"}, - "AdditionalDataPending":{"shape":"Boolean"} + "MaxRecords":{"shape":"IntegerOptional"} } }, - "DownloadDBLogFilePortionMessage":{ + "DescribeValidDBInstanceModificationsMessage":{ "type":"structure", - "required":[ - "DBInstanceIdentifier", - "LogFileName" - ], + "required":["DBInstanceIdentifier"], "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "LogFileName":{"shape":"String"}, - "Marker":{"shape":"String"}, - "NumberOfLines":{"shape":"Integer"} + "DBInstanceIdentifier":{"shape":"String"} } }, - "EC2SecurityGroup":{ + "DescribeValidDBInstanceModificationsResult":{ "type":"structure", "members":{ + "ValidDBInstanceModificationsMessage":{"shape":"ValidDBInstanceModificationsMessage"} + } + }, + "DomainMembership":{ + "type":"structure", + "members":{ + "Domain":{"shape":"String"}, "Status":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} + "FQDN":{"shape":"String"}, + "IAMRoleName":{"shape":"String"} } }, - "EC2SecurityGroupList":{ + "DomainMembershipList":{ "type":"list", "member":{ - "shape":"EC2SecurityGroup", - "locationName":"EC2SecurityGroup" + "shape":"DomainMembership", + "locationName":"DomainMembership" + } + }, + "DomainNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"DomainNotFoundFault", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, + "Double":{"type":"double"}, + "DoubleOptional":{"type":"double"}, + "DoubleRange":{ + "type":"structure", + "members":{ + "From":{"shape":"Double"}, + "To":{"shape":"Double"} + } + }, + "DoubleRangeList":{ + "type":"list", + "member":{ + "shape":"DoubleRange", + "locationName":"DoubleRange" } }, "Endpoint":{ "type":"structure", "members":{ "Address":{"shape":"String"}, - "Port":{"shape":"Integer"} + "Port":{"shape":"Integer"}, + "HostedZoneId":{"shape":"String"} } }, "EngineDefaults":{ @@ -2087,7 +2394,8 @@ "SourceType":{"shape":"SourceType"}, "Message":{"shape":"String"}, "EventCategories":{"shape":"EventCategoriesList"}, - "Date":{"shape":"TStamp"} + "Date":{"shape":"TStamp"}, + "SourceArn":{"shape":"String"} } }, "EventCategoriesList":{ @@ -2136,7 +2444,8 @@ "SourceType":{"shape":"String"}, "SourceIdsList":{"shape":"SourceIdsList"}, "EventCategoriesList":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"Boolean"} + "Enabled":{"shape":"Boolean"}, + "EventSubscriptionArn":{"shape":"String"} }, "wrapper":true }, @@ -2172,6 +2481,19 @@ "Events":{"shape":"EventList"} } }, + "FailoverDBClusterMessage":{ + "type":"structure", + "members":{ + "DBClusterIdentifier":{"shape":"String"}, + "TargetDBInstanceIdentifier":{"shape":"String"} + } + }, + "FailoverDBClusterResult":{ + "type":"structure", + "members":{ + "DBCluster":{"shape":"DBCluster"} + } + }, "Filter":{ "type":"structure", "required":[ @@ -2197,20 +2519,6 @@ "locationName":"Value" } }, - "IPRange":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "CIDRIP":{"shape":"String"} - } - }, - "IPRangeList":{ - "type":"list", - "member":{ - "shape":"IPRange", - "locationName":"IPRange" - } - }, "InstanceQuotaExceededFault":{ "type":"structure", "members":{ @@ -2222,6 +2530,17 @@ }, "exception":true }, + "InsufficientDBClusterCapacityFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"InsufficientDBClusterCapacityFault", + "httpStatusCode":403, + "senderFault":true + }, + "exception":true + }, "InsufficientDBInstanceCapacityFault":{ "type":"structure", "members":{ @@ -2233,8 +2552,41 @@ }, "exception":true }, + "InsufficientStorageClusterCapacityFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"InsufficientStorageClusterCapacity", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, "Integer":{"type":"integer"}, "IntegerOptional":{"type":"integer"}, + "InvalidDBClusterSnapshotStateFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"InvalidDBClusterSnapshotStateFault", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, + "InvalidDBClusterStateFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"InvalidDBClusterStateFault", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, "InvalidDBInstanceStateFault":{ "type":"structure", "members":{ @@ -2279,17 +2631,6 @@ }, "exception":true }, - "InvalidDBSubnetGroupFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetGroupFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, "InvalidDBSubnetGroupStateFault":{ "type":"structure", "members":{ @@ -2323,17 +2664,6 @@ }, "exception":true }, - "InvalidOptionGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidOptionGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, "InvalidRestoreFault":{ "type":"structure", "members":{ @@ -2367,6 +2697,17 @@ }, "exception":true }, + "KMSKeyNotAccessibleFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"KMSKeyNotAccessibleFault", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, "KeyList":{ "type":"list", "member":{"shape":"String"} @@ -2379,7 +2720,65 @@ "Filters":{"shape":"FilterList"} } }, - "Long":{"type":"long"}, + "LogTypeList":{ + "type":"list", + "member":{"shape":"String"} + }, + "ModifyDBClusterMessage":{ + "type":"structure", + "required":["DBClusterIdentifier"], + "members":{ + "DBClusterIdentifier":{"shape":"String"}, + "NewDBClusterIdentifier":{"shape":"String"}, + "ApplyImmediately":{"shape":"Boolean"}, + "BackupRetentionPeriod":{"shape":"IntegerOptional"}, + "DBClusterParameterGroupName":{"shape":"String"}, + "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, + "Port":{"shape":"IntegerOptional"}, + "MasterUserPassword":{"shape":"String"}, + "OptionGroupName":{"shape":"String"}, + "PreferredBackupWindow":{"shape":"String"}, + "PreferredMaintenanceWindow":{"shape":"String"}, + "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, + "EngineVersion":{"shape":"String"} + } + }, + "ModifyDBClusterParameterGroupMessage":{ + "type":"structure", + "required":[ + "DBClusterParameterGroupName", + "Parameters" + ], + "members":{ + "DBClusterParameterGroupName":{"shape":"String"}, + "Parameters":{"shape":"ParametersList"} + } + }, + "ModifyDBClusterResult":{ + "type":"structure", + "members":{ + "DBCluster":{"shape":"DBCluster"} + } + }, + "ModifyDBClusterSnapshotAttributeMessage":{ + "type":"structure", + "required":[ + "DBClusterSnapshotIdentifier", + "AttributeName" + ], + "members":{ + "DBClusterSnapshotIdentifier":{"shape":"String"}, + "AttributeName":{"shape":"String"}, + "ValuesToAdd":{"shape":"AttributeValueList"}, + "ValuesToRemove":{"shape":"AttributeValueList"} + } + }, + "ModifyDBClusterSnapshotAttributeResult":{ + "type":"structure", + "members":{ + "DBClusterSnapshotAttributesResult":{"shape":"DBClusterSnapshotAttributesResult"} + } + }, "ModifyDBInstanceMessage":{ "type":"structure", "required":["DBInstanceIdentifier"], @@ -2387,6 +2786,7 @@ "DBInstanceIdentifier":{"shape":"String"}, "AllocatedStorage":{"shape":"IntegerOptional"}, "DBInstanceClass":{"shape":"String"}, + "DBSubnetGroupName":{"shape":"String"}, "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, "ApplyImmediately":{"shape":"Boolean"}, @@ -2399,12 +2799,29 @@ "EngineVersion":{"shape":"String"}, "AllowMajorVersionUpgrade":{"shape":"Boolean"}, "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, + "LicenseModel":{"shape":"String"}, "Iops":{"shape":"IntegerOptional"}, "OptionGroupName":{"shape":"String"}, "NewDBInstanceIdentifier":{"shape":"String"}, "StorageType":{"shape":"String"}, "TdeCredentialArn":{"shape":"String"}, - "TdeCredentialPassword":{"shape":"String"} + "TdeCredentialPassword":{"shape":"String"}, + "CACertificateIdentifier":{"shape":"String"}, + "Domain":{"shape":"String"}, + "CopyTagsToSnapshot":{"shape":"BooleanOptional"}, + "MonitoringInterval":{"shape":"IntegerOptional"}, + "DBPortNumber":{"shape":"IntegerOptional"}, + "PubliclyAccessible":{ + "shape":"BooleanOptional", + "deprecated":true + }, + "MonitoringRoleArn":{"shape":"String"}, + "DomainIAMRoleName":{"shape":"String"}, + "PromotionTier":{"shape":"IntegerOptional"}, + "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, + "EnablePerformanceInsights":{"shape":"BooleanOptional"}, + "PerformanceInsightsKMSKeyId":{"shape":"String"}, + "CloudwatchLogsExportConfiguration":{"shape":"CloudwatchLogsExportConfiguration"} } }, "ModifyDBInstanceResult":{ @@ -2459,77 +2876,6 @@ "EventSubscription":{"shape":"EventSubscription"} } }, - "ModifyOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionsToInclude":{"shape":"OptionConfigurationList"}, - "OptionsToRemove":{"shape":"OptionNamesList"}, - "ApplyImmediately":{"shape":"Boolean"} - } - }, - "ModifyOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "Option":{ - "type":"structure", - "members":{ - "OptionName":{"shape":"String"}, - "OptionDescription":{"shape":"String"}, - "Persistent":{"shape":"Boolean"}, - "Permanent":{"shape":"Boolean"}, - "Port":{"shape":"IntegerOptional"}, - "OptionSettings":{"shape":"OptionSettingConfigurationList"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupMembershipList"} - } - }, - "OptionConfiguration":{ - "type":"structure", - "required":["OptionName"], - "members":{ - "OptionName":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupIdList"}, - "OptionSettings":{"shape":"OptionSettingsList"} - } - }, - "OptionConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionConfiguration", - "locationName":"OptionConfiguration" - } - }, - "OptionGroup":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "Options":{"shape":"OptionsList"}, - "AllowsVpcAndNonVpcInstanceMemberships":{"shape":"Boolean"}, - "VpcId":{"shape":"String"} - }, - "wrapper":true - }, - "OptionGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupAlreadyExistsFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, "OptionGroupMembership":{ "type":"structure", "members":{ @@ -2555,125 +2901,6 @@ }, "exception":true }, - "OptionGroupOption":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "MinimumRequiredMinorEngineVersion":{"shape":"String"}, - "PortRequired":{"shape":"Boolean"}, - "DefaultPort":{"shape":"IntegerOptional"}, - "OptionsDependedOn":{"shape":"OptionsDependedOn"}, - "Persistent":{"shape":"Boolean"}, - "Permanent":{"shape":"Boolean"}, - "OptionGroupOptionSettings":{"shape":"OptionGroupOptionSettingsList"} - } - }, - "OptionGroupOptionSetting":{ - "type":"structure", - "members":{ - "SettingName":{"shape":"String"}, - "SettingDescription":{"shape":"String"}, - "DefaultValue":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"} - } - }, - "OptionGroupOptionSettingsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOptionSetting", - "locationName":"OptionGroupOptionSetting" - } - }, - "OptionGroupOptionsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOption", - "locationName":"OptionGroupOption" - } - }, - "OptionGroupOptionsMessage":{ - "type":"structure", - "members":{ - "OptionGroupOptions":{"shape":"OptionGroupOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroups":{ - "type":"structure", - "members":{ - "OptionGroupsList":{"shape":"OptionGroupsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupsList":{ - "type":"list", - "member":{ - "shape":"OptionGroup", - "locationName":"OptionGroup" - } - }, - "OptionNamesList":{ - "type":"list", - "member":{"shape":"String"} - }, - "OptionSetting":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Value":{"shape":"String"}, - "DefaultValue":{"shape":"String"}, - "Description":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "DataType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"}, - "IsCollection":{"shape":"Boolean"} - } - }, - "OptionSettingConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionSetting", - "locationName":"OptionSetting" - } - }, - "OptionSettingsList":{ - "type":"list", - "member":{ - "shape":"OptionSetting", - "locationName":"OptionSetting" - } - }, - "OptionsDependedOn":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"OptionName" - } - }, - "OptionsList":{ - "type":"list", - "member":{ - "shape":"Option", - "locationName":"Option" - } - }, "OrderableDBInstanceOption":{ "type":"structure", "members":{ @@ -2685,8 +2912,18 @@ "MultiAZCapable":{"shape":"Boolean"}, "ReadReplicaCapable":{"shape":"Boolean"}, "Vpc":{"shape":"Boolean"}, + "SupportsStorageEncryption":{"shape":"Boolean"}, "StorageType":{"shape":"String"}, - "SupportsIops":{"shape":"Boolean"} + "SupportsIops":{"shape":"Boolean"}, + "SupportsEnhancedMonitoring":{"shape":"Boolean"}, + "SupportsIAMDatabaseAuthentication":{"shape":"Boolean"}, + "SupportsPerformanceInsights":{"shape":"Boolean"}, + "MinStorageSize":{"shape":"IntegerOptional"}, + "MaxStorageSize":{"shape":"IntegerOptional"}, + "MinIopsPerDbInstance":{"shape":"IntegerOptional"}, + "MaxIopsPerDbInstance":{"shape":"IntegerOptional"}, + "MinIopsPerGib":{"shape":"DoubleOptional"}, + "MaxIopsPerGib":{"shape":"DoubleOptional"} }, "wrapper":true }, @@ -2726,6 +2963,45 @@ "locationName":"Parameter" } }, + "PendingCloudwatchLogsExports":{ + "type":"structure", + "members":{ + "LogTypesToEnable":{"shape":"LogTypeList"}, + "LogTypesToDisable":{"shape":"LogTypeList"} + } + }, + "PendingMaintenanceAction":{ + "type":"structure", + "members":{ + "Action":{"shape":"String"}, + "AutoAppliedAfterDate":{"shape":"TStamp"}, + "ForcedApplyDate":{"shape":"TStamp"}, + "OptInStatus":{"shape":"String"}, + "CurrentApplyDate":{"shape":"TStamp"}, + "Description":{"shape":"String"} + } + }, + "PendingMaintenanceActionDetails":{ + "type":"list", + "member":{ + "shape":"PendingMaintenanceAction", + "locationName":"PendingMaintenanceAction" + } + }, + "PendingMaintenanceActions":{ + "type":"list", + "member":{ + "shape":"ResourcePendingMaintenanceActions", + "locationName":"ResourcePendingMaintenanceActions" + } + }, + "PendingMaintenanceActionsMessage":{ + "type":"structure", + "members":{ + "PendingMaintenanceActions":{"shape":"PendingMaintenanceActions"}, + "Marker":{"shape":"String"} + } + }, "PendingModifiedValues":{ "type":"structure", "members":{ @@ -2736,35 +3012,26 @@ "BackupRetentionPeriod":{"shape":"IntegerOptional"}, "MultiAZ":{"shape":"BooleanOptional"}, "EngineVersion":{"shape":"String"}, + "LicenseModel":{"shape":"String"}, "Iops":{"shape":"IntegerOptional"}, "DBInstanceIdentifier":{"shape":"String"}, - "StorageType":{"shape":"String"} + "StorageType":{"shape":"String"}, + "CACertificateIdentifier":{"shape":"String"}, + "DBSubnetGroupName":{"shape":"String"}, + "PendingCloudwatchLogsExports":{"shape":"PendingCloudwatchLogsExports"} } }, - "PointInTimeRestoreNotEnabledFault":{ + "PromoteReadReplicaDBClusterMessage":{ "type":"structure", + "required":["DBClusterIdentifier"], "members":{ - }, - "error":{ - "code":"PointInTimeRestoreNotEnabled", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PromoteReadReplicaMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"} + "DBClusterIdentifier":{"shape":"String"} } }, - "PromoteReadReplicaResult":{ + "PromoteReadReplicaDBClusterResult":{ "type":"structure", "members":{ - "DBInstance":{"shape":"DBInstance"} + "DBCluster":{"shape":"DBCluster"} } }, "ProvisionedIopsNotAvailableInAZFault":{ @@ -2778,20 +3045,26 @@ }, "exception":true }, - "PurchaseReservedDBInstancesOfferingMessage":{ + "Range":{ "type":"structure", - "required":["ReservedDBInstancesOfferingId"], "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "ReservedDBInstanceId":{"shape":"String"}, - "DBInstanceCount":{"shape":"IntegerOptional"}, - "Tags":{"shape":"TagList"} + "From":{"shape":"Integer"}, + "To":{"shape":"Integer"}, + "Step":{"shape":"IntegerOptional"} } }, - "PurchaseReservedDBInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedDBInstance":{"shape":"ReservedDBInstance"} + "RangeList":{ + "type":"list", + "member":{ + "shape":"Range", + "locationName":"Range" + } + }, + "ReadReplicaDBClusterIdentifierList":{ + "type":"list", + "member":{ + "shape":"String", + "locationName":"ReadReplicaDBClusterIdentifier" } }, "ReadReplicaDBInstanceIdentifierList":{ @@ -2801,6 +3074,13 @@ "locationName":"ReadReplicaDBInstanceIdentifier" } }, + "ReadReplicaIdentifierList":{ + "type":"list", + "member":{ + "shape":"String", + "locationName":"ReadReplicaIdentifier" + } + }, "RebootDBInstanceMessage":{ "type":"structure", "required":["DBInstanceIdentifier"], @@ -2815,19 +3095,15 @@ "DBInstance":{"shape":"DBInstance"} } }, - "RecurringCharge":{ + "RemoveRoleFromDBClusterMessage":{ "type":"structure", + "required":[ + "DBClusterIdentifier", + "RoleArn" + ], "members":{ - "RecurringChargeAmount":{"shape":"Double"}, - "RecurringChargeFrequency":{"shape":"String"} - }, - "wrapper":true - }, - "RecurringChargeList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"RecurringCharge" + "DBClusterIdentifier":{"shape":"String"}, + "RoleArn":{"shape":"String"} } }, "RemoveSourceIdentifierFromSubscriptionMessage":{ @@ -2858,114 +3134,15 @@ "TagKeys":{"shape":"KeyList"} } }, - "ReservedDBInstance":{ + "ResetDBClusterParameterGroupMessage":{ "type":"structure", + "required":["DBClusterParameterGroupName"], "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "StartTime":{"shape":"TStamp"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "DBInstanceCount":{"shape":"Integer"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "State":{"shape":"String"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceAlreadyExists", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstance", - "locationName":"ReservedDBInstance" + "DBClusterParameterGroupName":{"shape":"String"}, + "ResetAllParameters":{"shape":"Boolean"}, + "Parameters":{"shape":"ParametersList"} } }, - "ReservedDBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstances":{"shape":"ReservedDBInstanceList"} - } - }, - "ReservedDBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstancesOffering", - "locationName":"ReservedDBInstancesOffering" - } - }, - "ReservedDBInstancesOfferingMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstancesOfferings":{"shape":"ReservedDBInstancesOfferingList"} - } - }, - "ReservedDBInstancesOfferingNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstancesOfferingNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, "ResetDBParameterGroupMessage":{ "type":"structure", "required":["DBParameterGroupName"], @@ -2975,89 +3152,79 @@ "Parameters":{"shape":"ParametersList"} } }, - "RestoreDBInstanceFromDBSnapshotMessage":{ + "ResourceNotFoundFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"ResourceNotFoundFault", + "httpStatusCode":404, + "senderFault":true + }, + "exception":true + }, + "ResourcePendingMaintenanceActions":{ + "type":"structure", + "members":{ + "ResourceIdentifier":{"shape":"String"}, + "PendingMaintenanceActionDetails":{"shape":"PendingMaintenanceActionDetails"} + }, + "wrapper":true + }, + "RestoreDBClusterFromSnapshotMessage":{ "type":"structure", "required":[ - "DBInstanceIdentifier", - "DBSnapshotIdentifier" + "DBClusterIdentifier", + "SnapshotIdentifier", + "Engine" ], "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, + "AvailabilityZones":{"shape":"AvailabilityZones"}, + "DBClusterIdentifier":{"shape":"String"}, + "SnapshotIdentifier":{"shape":"String"}, "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, + "EngineVersion":{"shape":"String"}, + "Port":{"shape":"IntegerOptional"}, + "DBSubnetGroupName":{"shape":"String"}, + "DatabaseName":{"shape":"String"}, "OptionGroupName":{"shape":"String"}, + "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, "Tags":{"shape":"TagList"}, - "StorageType":{"shape":"String"}, - "TdeCredentialArn":{"shape":"String"}, - "TdeCredentialPassword":{"shape":"String"} + "KmsKeyId":{"shape":"String"}, + "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"} } }, - "RestoreDBInstanceFromDBSnapshotResult":{ + "RestoreDBClusterFromSnapshotResult":{ "type":"structure", "members":{ - "DBInstance":{"shape":"DBInstance"} + "DBCluster":{"shape":"DBCluster"} } }, - "RestoreDBInstanceToPointInTimeMessage":{ + "RestoreDBClusterToPointInTimeMessage":{ "type":"structure", "required":[ - "SourceDBInstanceIdentifier", - "TargetDBInstanceIdentifier" + "DBClusterIdentifier", + "SourceDBClusterIdentifier" ], "members":{ - "SourceDBInstanceIdentifier":{"shape":"String"}, - "TargetDBInstanceIdentifier":{"shape":"String"}, - "RestoreTime":{"shape":"TStamp"}, + "DBClusterIdentifier":{"shape":"String"}, + "RestoreType":{"shape":"String"}, + "SourceDBClusterIdentifier":{"shape":"String"}, + "RestoreToTime":{"shape":"TStamp"}, "UseLatestRestorableTime":{"shape":"Boolean"}, - "DBInstanceClass":{"shape":"String"}, "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, "OptionGroupName":{"shape":"String"}, + "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, "Tags":{"shape":"TagList"}, - "StorageType":{"shape":"String"}, - "TdeCredentialArn":{"shape":"String"}, - "TdeCredentialPassword":{"shape":"String"} + "KmsKeyId":{"shape":"String"}, + "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"} } }, - "RestoreDBInstanceToPointInTimeResult":{ + "RestoreDBClusterToPointInTimeResult":{ "type":"structure", "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RevokeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "RevokeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} + "DBCluster":{"shape":"DBCluster"} } }, "SNSInvalidTopicFault":{ @@ -3093,6 +3260,17 @@ }, "exception":true }, + "SharedSnapshotQuotaExceededFault":{ + "type":"structure", + "members":{ + }, + "error":{ + "code":"SharedSnapshotQuotaExceeded", + "httpStatusCode":400, + "senderFault":true + }, + "exception":true + }, "SnapshotQuotaExceededFault":{ "type":"structure", "members":{ @@ -3128,7 +3306,9 @@ "db-instance", "db-parameter-group", "db-security-group", - "db-snapshot" + "db-snapshot", + "db-cluster", + "db-cluster-snapshot" ] }, "StorageQuotaExceededFault":{ @@ -3227,6 +3407,13 @@ "locationName":"CharacterSet" } }, + "SupportedTimezonesList":{ + "type":"list", + "member":{ + "shape":"Timezone", + "locationName":"Timezone" + } + }, "TStamp":{"type":"timestamp"}, "Tag":{ "type":"structure", @@ -3248,6 +3435,52 @@ "TagList":{"shape":"TagList"} } }, + "Timezone":{ + "type":"structure", + "members":{ + "TimezoneName":{"shape":"String"} + } + }, + "UpgradeTarget":{ + "type":"structure", + "members":{ + "Engine":{"shape":"String"}, + "EngineVersion":{"shape":"String"}, + "Description":{"shape":"String"}, + "AutoUpgrade":{"shape":"Boolean"}, + "IsMajorVersionUpgrade":{"shape":"Boolean"} + } + }, + "ValidDBInstanceModificationsMessage":{ + "type":"structure", + "members":{ + "Storage":{"shape":"ValidStorageOptionsList"} + }, + "wrapper":true + }, + "ValidStorageOptions":{ + "type":"structure", + "members":{ + "StorageType":{"shape":"String"}, + "StorageSize":{"shape":"RangeList"}, + "ProvisionedIops":{"shape":"RangeList"}, + "IopsToStorageRatio":{"shape":"DoubleRangeList"} + } + }, + "ValidStorageOptionsList":{ + "type":"list", + "member":{ + "shape":"ValidStorageOptions", + "locationName":"ValidStorageOptions" + } + }, + "ValidUpgradeTargetList":{ + "type":"list", + "member":{ + "shape":"UpgradeTarget", + "locationName":"UpgradeTarget" + } + }, "VpcSecurityGroupIdList":{ "type":"list", "member":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/docs-2.json new file mode 100644 index 000000000..1bc087854 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/docs-2.json @@ -0,0 +1,2063 @@ +{ + "version":"2.0", + "service":"Amazon Neptune

Amazon Neptune is a fast, reliable, fully-managed graph database service that makes it easy to build and run applications that work with highly connected datasets. The core of Amazon Neptune is a purpose-built, high-performance graph database engine optimized for storing billions of relationships and querying the graph with milliseconds latency. Amazon Neptune supports popular graph models Property Graph and W3C's RDF, and their respective query languages Apache TinkerPop Gremlin and SPARQL, allowing you to easily build queries that efficiently navigate highly connected datasets. Neptune powers graph use cases such as recommendation engines, fraud detection, knowledge graphs, drug discovery, and network security.

This interface reference for Amazon Neptune contains documentation for a programming or command line interface you can use to manage Amazon Neptune. Note that Amazon Neptune is asynchronous, which means that some interfaces might require techniques such as polling or callback functions to determine when a command has been applied. In this reference, the parameter descriptions indicate whether a command is applied immediately, on the next instance reboot, or during the maintenance window. The reference structure is as follows, and we list following some related topics from the user guide.

Amazon Neptune API Reference

", + "operations":{ + "AddRoleToDBCluster":"

Associates an Identity and Access Management (IAM) role from an Neptune DB cluster.

", + "AddSourceIdentifierToSubscription":"

Adds a source identifier to an existing event notification subscription.

", + "AddTagsToResource":"

Adds metadata tags to an Amazon Neptune resource. These tags can also be used with cost allocation reporting to track cost associated with Amazon Neptune resources, or used in a Condition statement in an IAM policy for Amazon Neptune.

", + "ApplyPendingMaintenanceAction":"

Applies a pending maintenance action to a resource (for example, to a DB instance).

", + "CopyDBClusterParameterGroup":"

Copies the specified DB cluster parameter group.

", + "CopyDBClusterSnapshot":"

Copies a snapshot of a DB cluster.

To copy a DB cluster snapshot from a shared manual DB cluster snapshot, SourceDBClusterSnapshotIdentifier must be the Amazon Resource Name (ARN) of the shared DB cluster snapshot.

You can copy an encrypted DB cluster snapshot from another AWS Region. In that case, the AWS Region where you call the CopyDBClusterSnapshot action is the destination AWS Region for the encrypted DB cluster snapshot to be copied to. To copy an encrypted DB cluster snapshot from another AWS Region, you must provide the following values:

  • KmsKeyId - The AWS Key Management System (AWS KMS) key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region.

  • PreSignedUrl - A URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot action to be called in the source AWS Region where the DB cluster snapshot is copied from. The pre-signed URL must be a valid request for the CopyDBClusterSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB cluster snapshot to be copied.

    The pre-signed URL request must contain the following parameter values:

    • KmsKeyId - The KMS key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. This is the same identifier for both the CopyDBClusterSnapshot action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

    • DestinationRegion - The name of the AWS Region that the DB cluster snapshot will be created in.

    • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster snapshot from the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:neptune-cluster1-snapshot-20161115.

    To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

  • TargetDBClusterSnapshotIdentifier - The identifier for the new copy of the DB cluster snapshot in the destination AWS Region.

  • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the ARN format for the source AWS Region and is the same value as the SourceDBClusterSnapshotIdentifier in the pre-signed URL.

To cancel the copy operation once it is in progress, delete the target DB cluster snapshot identified by TargetDBClusterSnapshotIdentifier while that DB cluster snapshot is in \"copying\" status.

", + "CopyDBParameterGroup":"

Copies the specified DB parameter group.

", + "CreateDBCluster":"

Creates a new Amazon Neptune DB cluster.

You can use the ReplicationSourceIdentifier parameter to create the DB cluster as a Read Replica of another DB cluster or Amazon Neptune DB instance. For cross-region replication where the DB cluster identified by ReplicationSourceIdentifier is encrypted, you must also specify the PreSignedUrl parameter.

", + "CreateDBClusterParameterGroup":"

Creates a new DB cluster parameter group.

Parameters in a DB cluster parameter group apply to all of the instances in a DB cluster.

A DB cluster parameter group is initially created with the default parameters for the database engine used by instances in the DB cluster. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBClusterParameterGroup. Once you've created a DB cluster parameter group, you need to associate it with your DB cluster using ModifyDBCluster. When you associate a new DB cluster parameter group with a running DB cluster, you need to reboot the DB instances in the DB cluster without failover for the new DB cluster parameter group and associated settings to take effect.

After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon Neptune to fully complete the create action before the DB cluster parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon Neptune console or the DescribeDBClusterParameters command to verify that your DB cluster parameter group has been created or modified.

", + "CreateDBClusterSnapshot":"

Creates a snapshot of a DB cluster.

", + "CreateDBInstance":"

Creates a new DB instance.

", + "CreateDBParameterGroup":"

Creates a new DB parameter group.

A DB parameter group is initially created with the default parameters for the database engine used by the DB instance. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBParameterGroup. Once you've created a DB parameter group, you need to associate it with your DB instance using ModifyDBInstance. When you associate a new DB parameter group with a running DB instance, you need to reboot the DB instance without failover for the new DB parameter group and associated settings to take effect.

After you create a DB parameter group, you should wait at least 5 minutes before creating your first DB instance that uses that DB parameter group as the default parameter group. This allows Amazon Neptune to fully complete the create action before the parameter group is used as the default for a new DB instance. This is especially important for parameters that are critical when creating the default database for a DB instance, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon Neptune console or the DescribeDBParameters command to verify that your DB parameter group has been created or modified.

", + "CreateDBSubnetGroup":"

Creates a new DB subnet group. DB subnet groups must contain at least one subnet in at least two AZs in the AWS Region.

", + "CreateEventSubscription":"

Creates an event notification subscription. This action requires a topic ARN (Amazon Resource Name) created by either the Neptune console, the SNS console, or the SNS API. To obtain an ARN with SNS, you must create a topic in Amazon SNS and subscribe to the topic. The ARN is displayed in the SNS console.

You can specify the type of source (SourceType) you want to be notified of, provide a list of Neptune sources (SourceIds) that triggers the events, and provide a list of event categories (EventCategories) for events you want to be notified of. For example, you can specify SourceType = db-instance, SourceIds = mydbinstance1, mydbinstance2 and EventCategories = Availability, Backup.

If you specify both the SourceType and SourceIds, such as SourceType = db-instance and SourceIdentifier = myDBInstance1, you are notified of all the db-instance events for the specified source. If you specify a SourceType but do not specify a SourceIdentifier, you receive notice of the events for that source type for all your Neptune sources. If you do not specify either the SourceType nor the SourceIdentifier, you are notified of events generated from all Neptune sources belonging to your customer account.

", + "DeleteDBCluster":"

The DeleteDBCluster action deletes a previously provisioned DB cluster. When you delete a DB cluster, all automated backups for that DB cluster are deleted and can't be recovered. Manual DB cluster snapshots of the specified DB cluster are not deleted.

", + "DeleteDBClusterParameterGroup":"

Deletes a specified DB cluster parameter group. The DB cluster parameter group to be deleted can't be associated with any DB clusters.

", + "DeleteDBClusterSnapshot":"

Deletes a DB cluster snapshot. If the snapshot is being copied, the copy operation is terminated.

The DB cluster snapshot must be in the available state to be deleted.

", + "DeleteDBInstance":"

The DeleteDBInstance action deletes a previously provisioned DB instance. When you delete a DB instance, all automated backups for that instance are deleted and can't be recovered. Manual DB snapshots of the DB instance to be deleted by DeleteDBInstance are not deleted.

If you request a final DB snapshot the status of the Amazon Neptune DB instance is deleting until the DB snapshot is created. The API action DescribeDBInstance is used to monitor the status of this operation. The action can't be canceled or reverted once submitted.

Note that when a DB instance is in a failure state and has a status of failed, incompatible-restore, or incompatible-network, you can only delete it when the SkipFinalSnapshot parameter is set to true.

If the specified DB instance is part of a DB cluster, you can't delete the DB instance if both of the following conditions are true:

  • The DB cluster is a Read Replica of another DB cluster.

  • The DB instance is the only instance in the DB cluster.

To delete a DB instance in this case, first call the PromoteReadReplicaDBCluster API action to promote the DB cluster so it's no longer a Read Replica. After the promotion completes, then call the DeleteDBInstance API action to delete the final instance in the DB cluster.

", + "DeleteDBParameterGroup":"

Deletes a specified DBParameterGroup. The DBParameterGroup to be deleted can't be associated with any DB instances.

", + "DeleteDBSubnetGroup":"

Deletes a DB subnet group.

The specified database subnet group must not be associated with any DB instances.

", + "DeleteEventSubscription":"

Deletes an event notification subscription.

", + "DescribeDBClusterParameterGroups":"

Returns a list of DBClusterParameterGroup descriptions. If a DBClusterParameterGroupName parameter is specified, the list will contain only the description of the specified DB cluster parameter group.

", + "DescribeDBClusterParameters":"

Returns the detailed parameter list for a particular DB cluster parameter group.

", + "DescribeDBClusterSnapshotAttributes":"

Returns a list of DB cluster snapshot attribute names and values for a manual DB cluster snapshot.

When sharing snapshots with other AWS accounts, DescribeDBClusterSnapshotAttributes returns the restore attribute and a list of IDs for the AWS accounts that are authorized to copy or restore the manual DB cluster snapshot. If all is included in the list of values for the restore attribute, then the manual DB cluster snapshot is public and can be copied or restored by all AWS accounts.

To add or remove access for an AWS account to copy or restore a manual DB cluster snapshot, or to make the manual DB cluster snapshot public or private, use the ModifyDBClusterSnapshotAttribute API action.

", + "DescribeDBClusterSnapshots":"

Returns information about DB cluster snapshots. This API action supports pagination.

", + "DescribeDBClusters":"

Returns information about provisioned DB clusters. This API supports pagination.

", + "DescribeDBEngineVersions":"

Returns a list of the available DB engines.

", + "DescribeDBInstances":"

Returns information about provisioned instances. This API supports pagination.

", + "DescribeDBParameterGroups":"

Returns a list of DBParameterGroup descriptions. If a DBParameterGroupName is specified, the list will contain only the description of the specified DB parameter group.

", + "DescribeDBParameters":"

Returns the detailed parameter list for a particular DB parameter group.

", + "DescribeDBSubnetGroups":"

Returns a list of DBSubnetGroup descriptions. If a DBSubnetGroupName is specified, the list will contain only the descriptions of the specified DBSubnetGroup.

For an overview of CIDR ranges, go to the Wikipedia Tutorial.

", + "DescribeEngineDefaultClusterParameters":"

Returns the default engine and system parameter information for the cluster database engine.

", + "DescribeEngineDefaultParameters":"

Returns the default engine and system parameter information for the specified database engine.

", + "DescribeEventCategories":"

Displays a list of categories for all event source types, or, if specified, for a specified source type.

", + "DescribeEventSubscriptions":"

Lists all the subscription descriptions for a customer account. The description for a subscription includes SubscriptionName, SNSTopicARN, CustomerID, SourceType, SourceID, CreationTime, and Status.

If you specify a SubscriptionName, lists the description for that subscription.

", + "DescribeEvents":"

Returns events related to DB instances, DB security groups, DB snapshots, and DB parameter groups for the past 14 days. Events specific to a particular DB instance, DB security group, database snapshot, or DB parameter group can be obtained by providing the name as a parameter. By default, the past hour of events are returned.

", + "DescribeOrderableDBInstanceOptions":"

Returns a list of orderable DB instance options for the specified engine.

", + "DescribePendingMaintenanceActions":"

Returns a list of resources (for example, DB instances) that have at least one pending maintenance action.

", + "DescribeValidDBInstanceModifications":"

You can call DescribeValidDBInstanceModifications to learn what modifications you can make to your DB instance. You can use this information when you call ModifyDBInstance.

", + "FailoverDBCluster":"

Forces a failover for a DB cluster.

A failover for a DB cluster promotes one of the Read Replicas (read-only instances) in the DB cluster to be the primary instance (the cluster writer).

Amazon Neptune will automatically fail over to a Read Replica, if one exists, when the primary instance fails. You can force a failover when you want to simulate a failure of a primary instance for testing. Because each instance in a DB cluster has its own endpoint address, you will need to clean up and re-establish any existing connections that use those endpoint addresses when the failover is complete.

", + "ListTagsForResource":"

Lists all tags on an Amazon Neptune resource.

", + "ModifyDBCluster":"

Modify a setting for a DB cluster. You can change one or more database configuration parameters by specifying these parameters and the new values in the request.

", + "ModifyDBClusterParameterGroup":"

Modifies the parameters of a DB cluster parameter group. To modify more than one parameter, submit a list of the following: ParameterName, ParameterValue, and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

Changes to dynamic parameters are applied immediately. Changes to static parameters require a reboot without failover to the DB cluster associated with the parameter group before the change can take effect.

After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon Neptune to fully complete the create action before the parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon Neptune console or the DescribeDBClusterParameters command to verify that your DB cluster parameter group has been created or modified.

", + "ModifyDBClusterSnapshotAttribute":"

Adds an attribute and values to, or removes an attribute and values from, a manual DB cluster snapshot.

To share a manual DB cluster snapshot with other AWS accounts, specify restore as the AttributeName and use the ValuesToAdd parameter to add a list of IDs of the AWS accounts that are authorized to restore the manual DB cluster snapshot. Use the value all to make the manual DB cluster snapshot public, which means that it can be copied or restored by all AWS accounts. Do not add the all value for any manual DB cluster snapshots that contain private information that you don't want available to all AWS accounts. If a manual DB cluster snapshot is encrypted, it can be shared, but only by specifying a list of authorized AWS account IDs for the ValuesToAdd parameter. You can't use all as a value for that parameter in this case.

To view which AWS accounts have access to copy or restore a manual DB cluster snapshot, or whether a manual DB cluster snapshot public or private, use the DescribeDBClusterSnapshotAttributes API action.

", + "ModifyDBInstance":"

Modifies settings for a DB instance. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. To learn what modifications you can make to your DB instance, call DescribeValidDBInstanceModifications before you call ModifyDBInstance.

", + "ModifyDBParameterGroup":"

Modifies the parameters of a DB parameter group. To modify more than one parameter, submit a list of the following: ParameterName, ParameterValue, and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

Changes to dynamic parameters are applied immediately. Changes to static parameters require a reboot without failover to the DB instance associated with the parameter group before the change can take effect.

After you modify a DB parameter group, you should wait at least 5 minutes before creating your first DB instance that uses that DB parameter group as the default parameter group. This allows Amazon Neptune to fully complete the modify action before the parameter group is used as the default for a new DB instance. This is especially important for parameters that are critical when creating the default database for a DB instance, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon Neptune console or the DescribeDBParameters command to verify that your DB parameter group has been created or modified.

", + "ModifyDBSubnetGroup":"

Modifies an existing DB subnet group. DB subnet groups must contain at least one subnet in at least two AZs in the AWS Region.

", + "ModifyEventSubscription":"

Modifies an existing event notification subscription. Note that you can't modify the source identifiers using this call; to change source identifiers for a subscription, use the AddSourceIdentifierToSubscription and RemoveSourceIdentifierFromSubscription calls.

You can see a list of the event categories for a given SourceType by using the DescribeEventCategories action.

", + "PromoteReadReplicaDBCluster":"

Promotes a Read Replica DB cluster to a standalone DB cluster.

", + "RebootDBInstance":"

You might need to reboot your DB instance, usually for maintenance reasons. For example, if you make certain modifications, or if you change the DB parameter group associated with the DB instance, you must reboot the instance for the changes to take effect.

Rebooting a DB instance restarts the database engine service. Rebooting a DB instance results in a momentary outage, during which the DB instance status is set to rebooting.

", + "RemoveRoleFromDBCluster":"

Disassociates an Identity and Access Management (IAM) role from a DB cluster.

", + "RemoveSourceIdentifierFromSubscription":"

Removes a source identifier from an existing event notification subscription.

", + "RemoveTagsFromResource":"

Removes metadata tags from an Amazon Neptune resource.

", + "ResetDBClusterParameterGroup":"

Modifies the parameters of a DB cluster parameter group to the default value. To reset specific parameters submit a list of the following: ParameterName and ApplyMethod. To reset the entire DB cluster parameter group, specify the DBClusterParameterGroupName and ResetAllParameters parameters.

When resetting the entire group, dynamic parameters are updated immediately and static parameters are set to pending-reboot to take effect on the next DB instance restart or RebootDBInstance request. You must call RebootDBInstance for every DB instance in your DB cluster that you want the updated static parameter to apply to.

", + "ResetDBParameterGroup":"

Modifies the parameters of a DB parameter group to the engine/system default value. To reset specific parameters, provide a list of the following: ParameterName and ApplyMethod. To reset the entire DB parameter group, specify the DBParameterGroup name and ResetAllParameters parameters. When resetting the entire group, dynamic parameters are updated immediately and static parameters are set to pending-reboot to take effect on the next DB instance restart or RebootDBInstance request.

", + "RestoreDBClusterFromSnapshot":"

Creates a new DB cluster from a DB snapshot or DB cluster snapshot.

If a DB snapshot is specified, the target DB cluster is created from the source DB snapshot with a default configuration and default security group.

If a DB cluster snapshot is specified, the target DB cluster is created from the source DB cluster restore point with the same configuration as the original source DB cluster, except that the new DB cluster is created with the default security group.

", + "RestoreDBClusterToPointInTime":"

Restores a DB cluster to an arbitrary point in time. Users can restore to any point in time before LatestRestorableTime for up to BackupRetentionPeriod days. The target DB cluster is created from the source DB cluster with the same configuration as the original DB cluster, except that the new DB cluster is created with the default DB security group.

This action only restores the DB cluster, not the DB instances for that DB cluster. You must invoke the CreateDBInstance action to create DB instances for the restored DB cluster, specifying the identifier of the restored DB cluster in DBClusterIdentifier. You can create DB instances only after the RestoreDBClusterToPointInTime action has completed and the DB cluster is available.

" + }, + "shapes":{ + "AddRoleToDBClusterMessage":{ + "base":null, + "refs":{ + } + }, + "AddSourceIdentifierToSubscriptionMessage":{ + "base":"

", + "refs":{ + } + }, + "AddSourceIdentifierToSubscriptionResult":{ + "base":null, + "refs":{ + } + }, + "AddTagsToResourceMessage":{ + "base":"

", + "refs":{ + } + }, + "ApplyMethod":{ + "base":null, + "refs":{ + "Parameter$ApplyMethod":"

Indicates when to apply parameter updates.

" + } + }, + "ApplyPendingMaintenanceActionMessage":{ + "base":"

", + "refs":{ + } + }, + "ApplyPendingMaintenanceActionResult":{ + "base":null, + "refs":{ + } + }, + "AttributeValueList":{ + "base":null, + "refs":{ + "DBClusterSnapshotAttribute$AttributeValues":"

The value(s) for the manual DB cluster snapshot attribute.

If the AttributeName field is set to restore, then this element returns a list of IDs of the AWS accounts that are authorized to copy or restore the manual DB cluster snapshot. If a value of all is in the list, then the manual DB cluster snapshot is public and available for any AWS account to copy or restore.

", + "ModifyDBClusterSnapshotAttributeMessage$ValuesToAdd":"

A list of DB cluster snapshot attributes to add to the attribute specified by AttributeName.

To authorize other AWS accounts to copy or restore a manual DB cluster snapshot, set this list to include one or more AWS account IDs, or all to make the manual DB cluster snapshot restorable by any AWS account. Do not add the all value for any manual DB cluster snapshots that contain private information that you don't want available to all AWS accounts.

", + "ModifyDBClusterSnapshotAttributeMessage$ValuesToRemove":"

A list of DB cluster snapshot attributes to remove from the attribute specified by AttributeName.

To remove authorization for other AWS accounts to copy or restore a manual DB cluster snapshot, set this list to include one or more AWS account identifiers, or all to remove authorization for any AWS account to copy or restore the DB cluster snapshot. If you specify all, an AWS account whose account ID is explicitly added to the restore attribute can still copy or restore a manual DB cluster snapshot.

" + } + }, + "AuthorizationNotFoundFault":{ + "base":"

Specified CIDRIP or EC2 security group is not authorized for the specified DB security group.

Neptune may not also be authorized via IAM to perform necessary actions on your behalf.

", + "refs":{ + } + }, + "AvailabilityZone":{ + "base":"

Contains Availability Zone information.

This data type is used as an element in the following data type:

", + "refs":{ + "AvailabilityZoneList$member":null, + "Subnet$SubnetAvailabilityZone":null + } + }, + "AvailabilityZoneList":{ + "base":null, + "refs":{ + "OrderableDBInstanceOption$AvailabilityZones":"

A list of Availability Zones for a DB instance.

" + } + }, + "AvailabilityZones":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$AvailabilityZones":"

A list of EC2 Availability Zones that instances in the DB cluster can be created in.

", + "DBCluster$AvailabilityZones":"

Provides the list of EC2 Availability Zones that instances in the DB cluster can be created in.

", + "DBClusterSnapshot$AvailabilityZones":"

Provides the list of EC2 Availability Zones that instances in the DB cluster snapshot can be restored in.

", + "RestoreDBClusterFromSnapshotMessage$AvailabilityZones":"

Provides the list of EC2 Availability Zones that instances in the restored DB cluster can be created in.

" + } + }, + "Boolean":{ + "base":null, + "refs":{ + "DBCluster$MultiAZ":"

Specifies whether the DB cluster has instances in multiple Availability Zones.

", + "DBCluster$StorageEncrypted":"

Specifies whether the DB cluster is encrypted.

", + "DBCluster$IAMDatabaseAuthenticationEnabled":"

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", + "DBClusterMember$IsClusterWriter":"

Value that is true if the cluster member is the primary instance for the DB cluster and false otherwise.

", + "DBClusterSnapshot$StorageEncrypted":"

Specifies whether the DB cluster snapshot is encrypted.

", + "DBClusterSnapshot$IAMDatabaseAuthenticationEnabled":"

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", + "DBEngineVersion$SupportsLogExportsToCloudwatchLogs":"

A value that indicates whether the engine version supports exporting the log types specified by ExportableLogTypes to CloudWatch Logs.

", + "DBEngineVersion$SupportsReadReplica":"

Indicates whether the database engine version supports read replicas.

", + "DBInstance$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment.

", + "DBInstance$AutoMinorVersionUpgrade":"

Indicates that minor version patches are applied automatically.

", + "DBInstance$PubliclyAccessible":"

This parameter is not supported.

", + "DBInstance$StorageEncrypted":"

Specifies whether the DB instance is encrypted.

", + "DBInstance$CopyTagsToSnapshot":"

Specifies whether tags are copied from the DB instance to snapshots of the DB instance.

", + "DBInstance$IAMDatabaseAuthenticationEnabled":"

True if AWS Identity and Access Management (IAM) authentication is enabled, and otherwise false.

", + "DBInstanceStatusInfo$Normal":"

Boolean value that is true if the instance is operating normally, or false if the instance is in an error state.

", + "DeleteDBClusterMessage$SkipFinalSnapshot":"

Determines whether a final DB cluster snapshot is created before the DB cluster is deleted. If true is specified, no DB cluster snapshot is created. If false is specified, a DB cluster snapshot is created before the DB cluster is deleted.

You must specify a FinalDBSnapshotIdentifier parameter if SkipFinalSnapshot is false.

Default: false

", + "DeleteDBInstanceMessage$SkipFinalSnapshot":"

Determines whether a final DB snapshot is created before the DB instance is deleted. If true is specified, no DBSnapshot is created. If false is specified, a DB snapshot is created before the DB instance is deleted.

Note that when a DB instance is in a failure state and has a status of 'failed', 'incompatible-restore', or 'incompatible-network', it can only be deleted when the SkipFinalSnapshot parameter is set to \"true\".

Specify true when deleting a Read Replica.

The FinalDBSnapshotIdentifier parameter must be specified if SkipFinalSnapshot is false.

Default: false

", + "DescribeDBClusterSnapshotsMessage$IncludeShared":"

True to include shared manual DB cluster snapshots from other AWS accounts that this AWS account has been given permission to copy or restore, and otherwise false. The default is false.

You can give an AWS account permission to restore a manual DB cluster snapshot from another AWS account by the ModifyDBClusterSnapshotAttribute API action.

", + "DescribeDBClusterSnapshotsMessage$IncludePublic":"

True to include manual DB cluster snapshots that are public and can be copied or restored by any AWS account, and otherwise false. The default is false. The default is false.

You can share a manual DB cluster snapshot as public by using the ModifyDBClusterSnapshotAttribute API action.

", + "DescribeDBEngineVersionsMessage$DefaultOnly":"

Indicates that only the default version of the specified engine or engine and major version combination is returned.

", + "EventSubscription$Enabled":"

A Boolean value indicating if the subscription is enabled. True indicates the subscription is enabled.

", + "ModifyDBClusterMessage$ApplyImmediately":"

A value that specifies whether the modifications in this request and any pending modifications are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow setting for the DB cluster. If this parameter is set to false, changes to the DB cluster are applied during the next maintenance window.

The ApplyImmediately parameter only affects the NewDBClusterIdentifier and MasterUserPassword values. If you set the ApplyImmediately parameter value to false, then changes to the NewDBClusterIdentifier and MasterUserPassword values are applied during the next maintenance window. All other changes are applied immediately, regardless of the value of the ApplyImmediately parameter.

Default: false

", + "ModifyDBInstanceMessage$ApplyImmediately":"

Specifies whether the modifications in this request and any pending modifications are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow setting for the DB instance.

If this parameter is set to false, changes to the DB instance are applied during the next maintenance window. Some parameter changes can cause an outage and are applied on the next call to RebootDBInstance, or the next failure reboot.

Default: false

", + "ModifyDBInstanceMessage$AllowMajorVersionUpgrade":"

Indicates that major version upgrades are allowed. Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible.

Constraints: This parameter must be set to true when specifying a value for the EngineVersion parameter that is a different major version than the DB instance's current version.

", + "OrderableDBInstanceOption$MultiAZCapable":"

Indicates whether a DB instance is Multi-AZ capable.

", + "OrderableDBInstanceOption$ReadReplicaCapable":"

Indicates whether a DB instance can have a Read Replica.

", + "OrderableDBInstanceOption$Vpc":"

Indicates whether a DB instance is in a VPC.

", + "OrderableDBInstanceOption$SupportsStorageEncryption":"

Indicates whether a DB instance supports encrypted storage.

", + "OrderableDBInstanceOption$SupportsIops":"

Indicates whether a DB instance supports provisioned IOPS.

", + "OrderableDBInstanceOption$SupportsEnhancedMonitoring":"

Indicates whether a DB instance supports Enhanced Monitoring at intervals from 1 to 60 seconds.

", + "OrderableDBInstanceOption$SupportsIAMDatabaseAuthentication":"

Indicates whether a DB instance supports IAM database authentication.

", + "OrderableDBInstanceOption$SupportsPerformanceInsights":"

True if a DB instance supports Performance Insights, otherwise false.

", + "Parameter$IsModifiable":"

Indicates whether (true) or not (false) the parameter can be modified. Some parameters have security or operational implications that prevent them from being changed.

", + "ResetDBClusterParameterGroupMessage$ResetAllParameters":"

A value that is set to true to reset all parameters in the DB cluster parameter group to their default values, and false otherwise. You can't use this parameter if there is a list of parameter names specified for the Parameters parameter.

", + "ResetDBParameterGroupMessage$ResetAllParameters":"

Specifies whether (true) or not (false) to reset all parameters in the DB parameter group to default values.

Default: true

", + "RestoreDBClusterToPointInTimeMessage$UseLatestRestorableTime":"

A value that is set to true to restore the DB cluster to the latest restorable backup time, and false otherwise.

Default: false

Constraints: Cannot be specified if RestoreToTime parameter is provided.

", + "UpgradeTarget$AutoUpgrade":"

A value that indicates whether the target version is applied to any source DB instances that have AutoMinorVersionUpgrade set to true.

", + "UpgradeTarget$IsMajorVersionUpgrade":"

A value that indicates whether a database engine is upgraded to a major version.

" + } + }, + "BooleanOptional":{ + "base":null, + "refs":{ + "CopyDBClusterSnapshotMessage$CopyTags":"

True to copy all tags from the source DB cluster snapshot to the target DB cluster snapshot, and otherwise false. The default is false.

", + "CreateDBClusterMessage$StorageEncrypted":"

Specifies whether the DB cluster is encrypted.

", + "CreateDBClusterMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "CreateDBInstanceMessage$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment. You can't set the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", + "CreateDBInstanceMessage$AutoMinorVersionUpgrade":"

Indicates that minor engine upgrades are applied automatically to the DB instance during the maintenance window.

Default: true

", + "CreateDBInstanceMessage$PubliclyAccessible":"

This parameter is not supported.

", + "CreateDBInstanceMessage$StorageEncrypted":"

Specifies whether the DB instance is encrypted.

Not applicable. The encryption for DB instances is managed by the DB cluster. For more information, see CreateDBCluster.

Default: false

", + "CreateDBInstanceMessage$CopyTagsToSnapshot":"

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", + "CreateDBInstanceMessage$EnableIAMDatabaseAuthentication":"

True to enable AWS Identity and Access Management (IAM) authentication for Neptune.

Default: false

", + "CreateDBInstanceMessage$EnablePerformanceInsights":"

True to enable Performance Insights for the DB instance, and otherwise false.

", + "CreateEventSubscriptionMessage$Enabled":"

A Boolean value; set to true to activate the subscription, set to false to create the subscription but not active it.

", + "DBInstance$PerformanceInsightsEnabled":"

True if Performance Insights is enabled for the DB instance, and otherwise false.

", + "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets":"

If this parameter is specified and the requested engine supports the CharacterSetName parameter for CreateDBInstance, the response includes a list of supported character sets for each engine version.

", + "DescribeDBEngineVersionsMessage$ListSupportedTimezones":"

If this parameter is specified and the requested engine supports the TimeZone parameter for CreateDBInstance, the response includes a list of supported time zones for each engine version.

", + "DescribeOrderableDBInstanceOptionsMessage$Vpc":"

The VPC filter value. Specify this parameter to show only the available VPC or non-VPC offerings.

", + "ModifyDBClusterMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "ModifyDBInstanceMessage$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment. Changing this parameter doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

", + "ModifyDBInstanceMessage$AutoMinorVersionUpgrade":"

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window. Changing this parameter doesn't result in an outage except in the following case and the change is asynchronously applied as soon as possible. An outage will result if this parameter is set to true during the maintenance window, and a newer minor version is available, and Neptune has enabled auto patching for that engine version.

", + "ModifyDBInstanceMessage$CopyTagsToSnapshot":"

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", + "ModifyDBInstanceMessage$PubliclyAccessible":"

This parameter is not supported.

", + "ModifyDBInstanceMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

Not applicable. Mapping AWS IAM accounts to database accounts is managed by the DB cluster. For more information, see ModifyDBCluster.

Default: false

", + "ModifyDBInstanceMessage$EnablePerformanceInsights":"

True to enable Performance Insights for the DB instance, and otherwise false.

", + "ModifyEventSubscriptionMessage$Enabled":"

A Boolean value; set to true to activate the subscription.

", + "PendingModifiedValues$MultiAZ":"

Indicates that the Single-AZ DB instance is to change to a Multi-AZ deployment.

", + "RebootDBInstanceMessage$ForceFailover":"

When true, the reboot is conducted through a MultiAZ failover.

Constraint: You can't specify true if the instance is not configured for MultiAZ.

", + "RestoreDBClusterFromSnapshotMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "RestoreDBClusterToPointInTimeMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

" + } + }, + "CertificateNotFoundFault":{ + "base":"

CertificateIdentifier does not refer to an existing certificate.

", + "refs":{ + } + }, + "CharacterSet":{ + "base":"

This data type is used as a response element in the action DescribeDBEngineVersions.

", + "refs":{ + "DBEngineVersion$DefaultCharacterSet":"

The default character set for new instances of this engine version, if the CharacterSetName parameter of the CreateDBInstance API is not specified.

", + "SupportedCharacterSetsList$member":null + } + }, + "CloudwatchLogsExportConfiguration":{ + "base":"

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB instance or DB cluster.

", + "refs":{ + "ModifyDBInstanceMessage$CloudwatchLogsExportConfiguration":"

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB instance or DB cluster.

" + } + }, + "CopyDBClusterParameterGroupMessage":{ + "base":null, + "refs":{ + } + }, + "CopyDBClusterParameterGroupResult":{ + "base":null, + "refs":{ + } + }, + "CopyDBClusterSnapshotMessage":{ + "base":"

", + "refs":{ + } + }, + "CopyDBClusterSnapshotResult":{ + "base":null, + "refs":{ + } + }, + "CopyDBParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "CopyDBParameterGroupResult":{ + "base":null, + "refs":{ + } + }, + "CreateDBClusterMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateDBClusterParameterGroupResult":{ + "base":null, + "refs":{ + } + }, + "CreateDBClusterResult":{ + "base":null, + "refs":{ + } + }, + "CreateDBClusterSnapshotMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateDBClusterSnapshotResult":{ + "base":null, + "refs":{ + } + }, + "CreateDBInstanceMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateDBInstanceResult":{ + "base":null, + "refs":{ + } + }, + "CreateDBParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateDBParameterGroupResult":{ + "base":null, + "refs":{ + } + }, + "CreateDBSubnetGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateDBSubnetGroupResult":{ + "base":null, + "refs":{ + } + }, + "CreateEventSubscriptionMessage":{ + "base":"

", + "refs":{ + } + }, + "CreateEventSubscriptionResult":{ + "base":null, + "refs":{ + } + }, + "DBCluster":{ + "base":"

Contains the details of an Amazon Neptune DB cluster.

This data type is used as a response element in the DescribeDBClusters action.

", + "refs":{ + "CreateDBClusterResult$DBCluster":null, + "DBClusterList$member":null, + "DeleteDBClusterResult$DBCluster":null, + "FailoverDBClusterResult$DBCluster":null, + "ModifyDBClusterResult$DBCluster":null, + "PromoteReadReplicaDBClusterResult$DBCluster":null, + "RestoreDBClusterFromSnapshotResult$DBCluster":null, + "RestoreDBClusterToPointInTimeResult$DBCluster":null + } + }, + "DBClusterAlreadyExistsFault":{ + "base":"

User already has a DB cluster with the given identifier.

", + "refs":{ + } + }, + "DBClusterList":{ + "base":null, + "refs":{ + "DBClusterMessage$DBClusters":"

Contains a list of DB clusters for the user.

" + } + }, + "DBClusterMember":{ + "base":"

Contains information about an instance that is part of a DB cluster.

", + "refs":{ + "DBClusterMemberList$member":null + } + }, + "DBClusterMemberList":{ + "base":null, + "refs":{ + "DBCluster$DBClusterMembers":"

Provides the list of instances that make up the DB cluster.

" + } + }, + "DBClusterMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBClusters action.

", + "refs":{ + } + }, + "DBClusterNotFoundFault":{ + "base":"

DBClusterIdentifier does not refer to an existing DB cluster.

", + "refs":{ + } + }, + "DBClusterOptionGroupMemberships":{ + "base":null, + "refs":{ + "DBCluster$DBClusterOptionGroupMemberships":"

Provides the list of option group memberships for this DB cluster.

" + } + }, + "DBClusterOptionGroupStatus":{ + "base":"

Contains status information for a DB cluster option group.

", + "refs":{ + "DBClusterOptionGroupMemberships$member":null + } + }, + "DBClusterParameterGroup":{ + "base":"

Contains the details of an Amazon Neptune DB cluster parameter group.

This data type is used as a response element in the DescribeDBClusterParameterGroups action.

", + "refs":{ + "CopyDBClusterParameterGroupResult$DBClusterParameterGroup":null, + "CreateDBClusterParameterGroupResult$DBClusterParameterGroup":null, + "DBClusterParameterGroupList$member":null + } + }, + "DBClusterParameterGroupDetails":{ + "base":"

Provides details about a DB cluster parameter group including the parameters in the DB cluster parameter group.

", + "refs":{ + } + }, + "DBClusterParameterGroupList":{ + "base":null, + "refs":{ + "DBClusterParameterGroupsMessage$DBClusterParameterGroups":"

A list of DB cluster parameter groups.

" + } + }, + "DBClusterParameterGroupNameMessage":{ + "base":"

", + "refs":{ + } + }, + "DBClusterParameterGroupNotFoundFault":{ + "base":"

DBClusterParameterGroupName does not refer to an existing DB Cluster parameter group.

", + "refs":{ + } + }, + "DBClusterParameterGroupsMessage":{ + "base":"

", + "refs":{ + } + }, + "DBClusterQuotaExceededFault":{ + "base":"

User attempted to create a new DB cluster and the user has already reached the maximum allowed DB cluster quota.

", + "refs":{ + } + }, + "DBClusterRole":{ + "base":"

Describes an AWS Identity and Access Management (IAM) role that is associated with a DB cluster.

", + "refs":{ + "DBClusterRoles$member":null + } + }, + "DBClusterRoleAlreadyExistsFault":{ + "base":"

The specified IAM role Amazon Resource Name (ARN) is already associated with the specified DB cluster.

", + "refs":{ + } + }, + "DBClusterRoleNotFoundFault":{ + "base":"

The specified IAM role Amazon Resource Name (ARN) is not associated with the specified DB cluster.

", + "refs":{ + } + }, + "DBClusterRoleQuotaExceededFault":{ + "base":"

You have exceeded the maximum number of IAM roles that can be associated with the specified DB cluster.

", + "refs":{ + } + }, + "DBClusterRoles":{ + "base":null, + "refs":{ + "DBCluster$AssociatedRoles":"

Provides a list of the AWS Identity and Access Management (IAM) roles that are associated with the DB cluster. IAM roles that are associated with a DB cluster grant permission for the DB cluster to access other AWS services on your behalf.

" + } + }, + "DBClusterSnapshot":{ + "base":"

Contains the details for an Amazon Neptune DB cluster snapshot

This data type is used as a response element in the DescribeDBClusterSnapshots action.

", + "refs":{ + "CopyDBClusterSnapshotResult$DBClusterSnapshot":null, + "CreateDBClusterSnapshotResult$DBClusterSnapshot":null, + "DBClusterSnapshotList$member":null, + "DeleteDBClusterSnapshotResult$DBClusterSnapshot":null + } + }, + "DBClusterSnapshotAlreadyExistsFault":{ + "base":"

User already has a DB cluster snapshot with the given identifier.

", + "refs":{ + } + }, + "DBClusterSnapshotAttribute":{ + "base":"

Contains the name and values of a manual DB cluster snapshot attribute.

Manual DB cluster snapshot attributes are used to authorize other AWS accounts to restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", + "refs":{ + "DBClusterSnapshotAttributeList$member":null + } + }, + "DBClusterSnapshotAttributeList":{ + "base":null, + "refs":{ + "DBClusterSnapshotAttributesResult$DBClusterSnapshotAttributes":"

The list of attributes and values for the manual DB cluster snapshot.

" + } + }, + "DBClusterSnapshotAttributesResult":{ + "base":"

Contains the results of a successful call to the DescribeDBClusterSnapshotAttributes API action.

Manual DB cluster snapshot attributes are used to authorize other AWS accounts to copy or restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", + "refs":{ + "DescribeDBClusterSnapshotAttributesResult$DBClusterSnapshotAttributesResult":null, + "ModifyDBClusterSnapshotAttributeResult$DBClusterSnapshotAttributesResult":null + } + }, + "DBClusterSnapshotList":{ + "base":null, + "refs":{ + "DBClusterSnapshotMessage$DBClusterSnapshots":"

Provides a list of DB cluster snapshots for the user.

" + } + }, + "DBClusterSnapshotMessage":{ + "base":"

Provides a list of DB cluster snapshots for the user as the result of a call to the DescribeDBClusterSnapshots action.

", + "refs":{ + } + }, + "DBClusterSnapshotNotFoundFault":{ + "base":"

DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot.

", + "refs":{ + } + }, + "DBEngineVersion":{ + "base":"

This data type is used as a response element in the action DescribeDBEngineVersions.

", + "refs":{ + "DBEngineVersionList$member":null + } + }, + "DBEngineVersionList":{ + "base":null, + "refs":{ + "DBEngineVersionMessage$DBEngineVersions":"

A list of DBEngineVersion elements.

" + } + }, + "DBEngineVersionMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBEngineVersions action.

", + "refs":{ + } + }, + "DBInstance":{ + "base":"

Contains the details of an Amazon Neptune DB instance.

This data type is used as a response element in the DescribeDBInstances action.

", + "refs":{ + "CreateDBInstanceResult$DBInstance":null, + "DBInstanceList$member":null, + "DeleteDBInstanceResult$DBInstance":null, + "ModifyDBInstanceResult$DBInstance":null, + "RebootDBInstanceResult$DBInstance":null + } + }, + "DBInstanceAlreadyExistsFault":{ + "base":"

User already has a DB instance with the given identifier.

", + "refs":{ + } + }, + "DBInstanceList":{ + "base":null, + "refs":{ + "DBInstanceMessage$DBInstances":"

A list of DBInstance instances.

" + } + }, + "DBInstanceMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBInstances action.

", + "refs":{ + } + }, + "DBInstanceNotFoundFault":{ + "base":"

DBInstanceIdentifier does not refer to an existing DB instance.

", + "refs":{ + } + }, + "DBInstanceStatusInfo":{ + "base":"

Provides a list of status information for a DB instance.

", + "refs":{ + "DBInstanceStatusInfoList$member":null + } + }, + "DBInstanceStatusInfoList":{ + "base":null, + "refs":{ + "DBInstance$StatusInfos":"

The status of a Read Replica. If the instance is not a Read Replica, this is blank.

" + } + }, + "DBParameterGroup":{ + "base":"

Contains the details of an Amazon Neptune DB parameter group.

This data type is used as a response element in the DescribeDBParameterGroups action.

", + "refs":{ + "CopyDBParameterGroupResult$DBParameterGroup":null, + "CreateDBParameterGroupResult$DBParameterGroup":null, + "DBParameterGroupList$member":null + } + }, + "DBParameterGroupAlreadyExistsFault":{ + "base":"

A DB parameter group with the same name exists.

", + "refs":{ + } + }, + "DBParameterGroupDetails":{ + "base":"

Contains the result of a successful invocation of the DescribeDBParameters action.

", + "refs":{ + } + }, + "DBParameterGroupList":{ + "base":null, + "refs":{ + "DBParameterGroupsMessage$DBParameterGroups":"

A list of DBParameterGroup instances.

" + } + }, + "DBParameterGroupNameMessage":{ + "base":"

Contains the result of a successful invocation of the ModifyDBParameterGroup or ResetDBParameterGroup action.

", + "refs":{ + } + }, + "DBParameterGroupNotFoundFault":{ + "base":"

DBParameterGroupName does not refer to an existing DB parameter group.

", + "refs":{ + } + }, + "DBParameterGroupQuotaExceededFault":{ + "base":"

Request would result in user exceeding the allowed number of DB parameter groups.

", + "refs":{ + } + }, + "DBParameterGroupStatus":{ + "base":"

The status of the DB parameter group.

This data type is used as a response element in the following actions:

", + "refs":{ + "DBParameterGroupStatusList$member":null + } + }, + "DBParameterGroupStatusList":{ + "base":null, + "refs":{ + "DBInstance$DBParameterGroups":"

Provides the list of DB parameter groups applied to this DB instance.

" + } + }, + "DBParameterGroupsMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBParameterGroups action.

", + "refs":{ + } + }, + "DBSecurityGroupMembership":{ + "base":"

This data type is used as a response element in the following actions:

", + "refs":{ + "DBSecurityGroupMembershipList$member":null + } + }, + "DBSecurityGroupMembershipList":{ + "base":null, + "refs":{ + "DBInstance$DBSecurityGroups":"

Provides List of DB security group elements containing only DBSecurityGroup.Name and DBSecurityGroup.Status subelements.

" + } + }, + "DBSecurityGroupNameList":{ + "base":null, + "refs":{ + "CreateDBInstanceMessage$DBSecurityGroups":"

A list of DB security groups to associate with this DB instance.

Default: The default DB security group for the database engine.

", + "ModifyDBInstanceMessage$DBSecurityGroups":"

A list of DB security groups to authorize on this DB instance. Changing this setting doesn't result in an outage and the change is asynchronously applied as soon as possible.

Constraints:

  • If supplied, must match existing DBSecurityGroups.

" + } + }, + "DBSecurityGroupNotFoundFault":{ + "base":"

DBSecurityGroupName does not refer to an existing DB security group.

", + "refs":{ + } + }, + "DBSnapshotAlreadyExistsFault":{ + "base":"

DBSnapshotIdentifier is already used by an existing snapshot.

", + "refs":{ + } + }, + "DBSnapshotNotFoundFault":{ + "base":"

DBSnapshotIdentifier does not refer to an existing DB snapshot.

", + "refs":{ + } + }, + "DBSubnetGroup":{ + "base":"

Contains the details of an Amazon Neptune DB subnet group.

This data type is used as a response element in the DescribeDBSubnetGroups action.

", + "refs":{ + "CreateDBSubnetGroupResult$DBSubnetGroup":null, + "DBInstance$DBSubnetGroup":"

Specifies information on the subnet group associated with the DB instance, including the name, description, and subnets in the subnet group.

", + "DBSubnetGroups$member":null, + "ModifyDBSubnetGroupResult$DBSubnetGroup":null + } + }, + "DBSubnetGroupAlreadyExistsFault":{ + "base":"

DBSubnetGroupName is already used by an existing DB subnet group.

", + "refs":{ + } + }, + "DBSubnetGroupDoesNotCoverEnoughAZs":{ + "base":"

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", + "refs":{ + } + }, + "DBSubnetGroupMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBSubnetGroups action.

", + "refs":{ + } + }, + "DBSubnetGroupNotFoundFault":{ + "base":"

DBSubnetGroupName does not refer to an existing DB subnet group.

", + "refs":{ + } + }, + "DBSubnetGroupQuotaExceededFault":{ + "base":"

Request would result in user exceeding the allowed number of DB subnet groups.

", + "refs":{ + } + }, + "DBSubnetGroups":{ + "base":null, + "refs":{ + "DBSubnetGroupMessage$DBSubnetGroups":"

A list of DBSubnetGroup instances.

" + } + }, + "DBSubnetQuotaExceededFault":{ + "base":"

Request would result in user exceeding the allowed number of subnets in a DB subnet groups.

", + "refs":{ + } + }, + "DBUpgradeDependencyFailureFault":{ + "base":"

The DB upgrade failed because a resource the DB depends on could not be modified.

", + "refs":{ + } + }, + "DeleteDBClusterMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteDBClusterResult":{ + "base":null, + "refs":{ + } + }, + "DeleteDBClusterSnapshotMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteDBClusterSnapshotResult":{ + "base":null, + "refs":{ + } + }, + "DeleteDBInstanceMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteDBInstanceResult":{ + "base":null, + "refs":{ + } + }, + "DeleteDBParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteDBSubnetGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteEventSubscriptionMessage":{ + "base":"

", + "refs":{ + } + }, + "DeleteEventSubscriptionResult":{ + "base":null, + "refs":{ + } + }, + "DescribeDBClusterParameterGroupsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBClusterParametersMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBClusterSnapshotAttributesMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBClusterSnapshotAttributesResult":{ + "base":null, + "refs":{ + } + }, + "DescribeDBClusterSnapshotsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBClustersMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBEngineVersionsMessage":{ + "base":null, + "refs":{ + } + }, + "DescribeDBInstancesMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBParameterGroupsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeDBParametersMessage":{ + "base":null, + "refs":{ + } + }, + "DescribeDBSubnetGroupsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeEngineDefaultClusterParametersMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeEngineDefaultClusterParametersResult":{ + "base":null, + "refs":{ + } + }, + "DescribeEngineDefaultParametersMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeEngineDefaultParametersResult":{ + "base":null, + "refs":{ + } + }, + "DescribeEventCategoriesMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeEventSubscriptionsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeEventsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeOrderableDBInstanceOptionsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribePendingMaintenanceActionsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeValidDBInstanceModificationsMessage":{ + "base":"

", + "refs":{ + } + }, + "DescribeValidDBInstanceModificationsResult":{ + "base":null, + "refs":{ + } + }, + "DomainMembership":{ + "base":"

An Active Directory Domain membership record associated with the DB instance.

", + "refs":{ + "DomainMembershipList$member":null + } + }, + "DomainMembershipList":{ + "base":"

List of Active Directory Domain membership records associated with a DB instance.

", + "refs":{ + "DBInstance$DomainMemberships":"

Not supported

" + } + }, + "DomainNotFoundFault":{ + "base":"

Domain does not refer to an existing Active Directory Domain.

", + "refs":{ + } + }, + "Double":{ + "base":null, + "refs":{ + "DoubleRange$From":"

The minimum value in the range.

", + "DoubleRange$To":"

The maximum value in the range.

" + } + }, + "DoubleOptional":{ + "base":null, + "refs":{ + "OrderableDBInstanceOption$MinIopsPerGib":"

Minimum provisioned IOPS per GiB for a DB instance.

", + "OrderableDBInstanceOption$MaxIopsPerGib":"

Maximum provisioned IOPS per GiB for a DB instance.

" + } + }, + "DoubleRange":{ + "base":"

A range of double values.

", + "refs":{ + "DoubleRangeList$member":null + } + }, + "DoubleRangeList":{ + "base":null, + "refs":{ + "ValidStorageOptions$IopsToStorageRatio":"

The valid range of Provisioned IOPS to gibibytes of storage multiplier. For example, 3-10, which means that provisioned IOPS can be between 3 and 10 times storage.

" + } + }, + "Endpoint":{ + "base":"

This data type is used as a response element in the following actions:

", + "refs":{ + "DBInstance$Endpoint":"

Specifies the connection endpoint.

" + } + }, + "EngineDefaults":{ + "base":"

Contains the result of a successful invocation of the DescribeEngineDefaultParameters action.

", + "refs":{ + "DescribeEngineDefaultClusterParametersResult$EngineDefaults":null, + "DescribeEngineDefaultParametersResult$EngineDefaults":null + } + }, + "Event":{ + "base":"

This data type is used as a response element in the DescribeEvents action.

", + "refs":{ + "EventList$member":null + } + }, + "EventCategoriesList":{ + "base":null, + "refs":{ + "CreateEventSubscriptionMessage$EventCategories":"

A list of event categories for a SourceType that you want to subscribe to. You can see a list of the categories for a given SourceType by using the DescribeEventCategories action.

", + "DescribeEventsMessage$EventCategories":"

A list of event categories that trigger notifications for a event notification subscription.

", + "Event$EventCategories":"

Specifies the category for the event.

", + "EventCategoriesMap$EventCategories":"

The event categories for the specified source type

", + "EventSubscription$EventCategoriesList":"

A list of event categories for the event notification subscription.

", + "ModifyEventSubscriptionMessage$EventCategories":"

A list of event categories for a SourceType that you want to subscribe to. You can see a list of the categories for a given SourceType by using the DescribeEventCategories action.

" + } + }, + "EventCategoriesMap":{ + "base":"

Contains the results of a successful invocation of the DescribeEventCategories action.

", + "refs":{ + "EventCategoriesMapList$member":null + } + }, + "EventCategoriesMapList":{ + "base":null, + "refs":{ + "EventCategoriesMessage$EventCategoriesMapList":"

A list of EventCategoriesMap data types.

" + } + }, + "EventCategoriesMessage":{ + "base":"

Data returned from the DescribeEventCategories action.

", + "refs":{ + } + }, + "EventList":{ + "base":null, + "refs":{ + "EventsMessage$Events":"

A list of Event instances.

" + } + }, + "EventSubscription":{ + "base":"

Contains the results of a successful invocation of the DescribeEventSubscriptions action.

", + "refs":{ + "AddSourceIdentifierToSubscriptionResult$EventSubscription":null, + "CreateEventSubscriptionResult$EventSubscription":null, + "DeleteEventSubscriptionResult$EventSubscription":null, + "EventSubscriptionsList$member":null, + "ModifyEventSubscriptionResult$EventSubscription":null, + "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription":null + } + }, + "EventSubscriptionQuotaExceededFault":{ + "base":null, + "refs":{ + } + }, + "EventSubscriptionsList":{ + "base":null, + "refs":{ + "EventSubscriptionsMessage$EventSubscriptionsList":"

A list of EventSubscriptions data types.

" + } + }, + "EventSubscriptionsMessage":{ + "base":"

Data returned by the DescribeEventSubscriptions action.

", + "refs":{ + } + }, + "EventsMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeEvents action.

", + "refs":{ + } + }, + "FailoverDBClusterMessage":{ + "base":"

", + "refs":{ + } + }, + "FailoverDBClusterResult":{ + "base":null, + "refs":{ + } + }, + "Filter":{ + "base":"

This type is not currently supported.

", + "refs":{ + "FilterList$member":null + } + }, + "FilterList":{ + "base":null, + "refs":{ + "DescribeDBClusterParameterGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClusterParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClusterSnapshotsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClustersMessage$Filters":"

A filter that specifies one or more DB clusters to describe.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include information about the DB clusters identified by these ARNs.

", + "DescribeDBEngineVersionsMessage$Filters":"

Not currently supported.

", + "DescribeDBInstancesMessage$Filters":"

A filter that specifies one or more DB instances to describe.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include information about the DB instances associated with the DB clusters identified by these ARNs.

  • db-instance-id - Accepts DB instance identifiers and DB instance Amazon Resource Names (ARNs). The results list will only include information about the DB instances identified by these ARNs.

", + "DescribeDBParameterGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBSubnetGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEngineDefaultClusterParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEngineDefaultParametersMessage$Filters":"

Not currently supported.

", + "DescribeEventCategoriesMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEventSubscriptionsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEventsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeOrderableDBInstanceOptionsMessage$Filters":"

This parameter is not currently supported.

", + "DescribePendingMaintenanceActionsMessage$Filters":"

A filter that specifies one or more resources to return pending maintenance actions for.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include pending maintenance actions for the DB clusters identified by these ARNs.

  • db-instance-id - Accepts DB instance identifiers and DB instance ARNs. The results list will only include pending maintenance actions for the DB instances identified by these ARNs.

", + "ListTagsForResourceMessage$Filters":"

This parameter is not currently supported.

" + } + }, + "FilterValueList":{ + "base":null, + "refs":{ + "Filter$Values":"

This parameter is not currently supported.

" + } + }, + "InstanceQuotaExceededFault":{ + "base":"

Request would result in user exceeding the allowed number of DB instances.

", + "refs":{ + } + }, + "InsufficientDBClusterCapacityFault":{ + "base":"

The DB cluster does not have enough capacity for the current operation.

", + "refs":{ + } + }, + "InsufficientDBInstanceCapacityFault":{ + "base":"

Specified DB instance class is not available in the specified Availability Zone.

", + "refs":{ + } + }, + "InsufficientStorageClusterCapacityFault":{ + "base":"

There is insufficient storage available for the current action. You may be able to resolve this error by updating your subnet group to use different Availability Zones that have more storage available.

", + "refs":{ + } + }, + "Integer":{ + "base":null, + "refs":{ + "DBClusterSnapshot$AllocatedStorage":"

Specifies the allocated storage size in gibibytes (GiB).

", + "DBClusterSnapshot$Port":"

Specifies the port that the DB cluster was listening on at the time of the snapshot.

", + "DBClusterSnapshot$PercentProgress":"

Specifies the percentage of the estimated data that has been transferred.

", + "DBInstance$AllocatedStorage":"

Specifies the allocated storage size specified in gibibytes.

", + "DBInstance$BackupRetentionPeriod":"

Specifies the number of days for which automatic DB snapshots are retained.

", + "DBInstance$DbInstancePort":"

Specifies the port that the DB instance listens on. If the DB instance is part of a DB cluster, this can be a different port than the DB cluster port.

", + "Endpoint$Port":"

Specifies the port that the database engine is listening on.

", + "Range$From":"

The minimum value in the range.

", + "Range$To":"

The maximum value in the range.

" + } + }, + "IntegerOptional":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$BackupRetentionPeriod":"

The number of days for which automated backups are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", + "CreateDBClusterMessage$Port":"

The port number on which the instances in the DB cluster accept connections.

Default: 8182

", + "CreateDBInstanceMessage$AllocatedStorage":"

The amount of storage (in gibibytes) to allocate for the DB instance.

Type: Integer

Not applicable. Neptune cluster volumes automatically grow as the amount of data in your database increases, though you are only charged for the space that you use in a Neptune cluster volume.

", + "CreateDBInstanceMessage$BackupRetentionPeriod":"

The number of days for which automated backups are retained.

Not applicable. The retention period for automated backups is managed by the DB cluster. For more information, see CreateDBCluster.

Default: 1

Constraints:

  • Must be a value from 0 to 35

  • Cannot be set to 0 if the DB instance is a source to Read Replicas

", + "CreateDBInstanceMessage$Port":"

The port number on which the database accepts connections.

Not applicable. The port is managed by the DB cluster. For more information, see CreateDBCluster.

Default: 8182

Type: Integer

", + "CreateDBInstanceMessage$Iops":"

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance.

", + "CreateDBInstanceMessage$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", + "CreateDBInstanceMessage$PromotionTier":"

A value that specifies the order in which an Read Replica is promoted to the primary instance after a failure of the existing primary instance.

Default: 1

Valid Values: 0 - 15

", + "DBCluster$AllocatedStorage":"

AllocatedStorage always returns 1, because Neptune DB cluster storage size is not fixed, but instead automatically adjusts as needed.

", + "DBCluster$BackupRetentionPeriod":"

Specifies the number of days for which automatic DB snapshots are retained.

", + "DBCluster$Port":"

Specifies the port that the database engine is listening on.

", + "DBClusterMember$PromotionTier":"

A value that specifies the order in which a Read Replica is promoted to the primary instance after a failure of the existing primary instance.

", + "DBInstance$Iops":"

Specifies the Provisioned IOPS (I/O operations per second) value.

", + "DBInstance$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance.

", + "DBInstance$PromotionTier":"

A value that specifies the order in which a Read Replica is promoted to the primary instance after a failure of the existing primary instance.

", + "DescribeDBClusterParameterGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClusterParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClusterSnapshotsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClustersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBEngineVersionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBInstancesMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBParameterGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBSubnetGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEngineDefaultClusterParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEngineDefaultParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEventSubscriptionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEventsMessage$Duration":"

The number of minutes to retrieve events for.

Default: 60

", + "DescribeEventsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeOrderableDBInstanceOptionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribePendingMaintenanceActionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "ModifyDBClusterMessage$BackupRetentionPeriod":"

The number of days for which automated backups are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", + "ModifyDBClusterMessage$Port":"

The port number on which the DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", + "ModifyDBInstanceMessage$AllocatedStorage":"

The new amount of storage (in gibibytes) to allocate for the DB instance.

Not applicable. Storage is managed by the DB Cluster.

", + "ModifyDBInstanceMessage$BackupRetentionPeriod":"

The number of days to retain automated backups. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Not applicable. The retention period for automated backups is managed by the DB cluster. For more information, see ModifyDBCluster.

Default: Uses existing setting

", + "ModifyDBInstanceMessage$Iops":"

The new Provisioned IOPS (I/O operations per second) value for the instance.

Changing this setting doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

Default: Uses existing setting

", + "ModifyDBInstanceMessage$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", + "ModifyDBInstanceMessage$DBPortNumber":"

The port number on which the database accepts connections.

The value of the DBPortNumber parameter must not match any of the port values specified for options in the option group for the DB instance.

Your database will restart when you change the DBPortNumber value regardless of the value of the ApplyImmediately parameter.

Default: 8182

", + "ModifyDBInstanceMessage$PromotionTier":"

A value that specifies the order in which a Read Replica is promoted to the primary instance after a failure of the existing primary instance.

Default: 1

Valid Values: 0 - 15

", + "OrderableDBInstanceOption$MinStorageSize":"

Minimum storage size for a DB instance.

", + "OrderableDBInstanceOption$MaxStorageSize":"

Maximum storage size for a DB instance.

", + "OrderableDBInstanceOption$MinIopsPerDbInstance":"

Minimum total provisioned IOPS for a DB instance.

", + "OrderableDBInstanceOption$MaxIopsPerDbInstance":"

Maximum total provisioned IOPS for a DB instance.

", + "PendingModifiedValues$AllocatedStorage":"

Contains the new AllocatedStorage size for the DB instance that will be applied or is currently being applied.

", + "PendingModifiedValues$Port":"

Specifies the pending port for the DB instance.

", + "PendingModifiedValues$BackupRetentionPeriod":"

Specifies the pending number of days for which automated backups are retained.

", + "PendingModifiedValues$Iops":"

Specifies the new Provisioned IOPS value for the DB instance that will be applied or is currently being applied.

", + "Range$Step":"

The step value for the range. For example, if you have a range of 5,000 to 10,000, with a step value of 1,000, the valid values start at 5,000 and step up by 1,000. Even though 7,500 is within the range, it isn't a valid value for the range. The valid values are 5,000, 6,000, 7,000, 8,000...

", + "RestoreDBClusterFromSnapshotMessage$Port":"

The port number on which the new DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$Port":"

The port number on which the new DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

" + } + }, + "InvalidDBClusterSnapshotStateFault":{ + "base":"

The supplied value is not a valid DB cluster snapshot state.

", + "refs":{ + } + }, + "InvalidDBClusterStateFault":{ + "base":"

The DB cluster is not in a valid state.

", + "refs":{ + } + }, + "InvalidDBInstanceStateFault":{ + "base":"

The specified DB instance is not in the available state.

", + "refs":{ + } + }, + "InvalidDBParameterGroupStateFault":{ + "base":"

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you cannot delete it when the parameter group is in this state.

", + "refs":{ + } + }, + "InvalidDBSecurityGroupStateFault":{ + "base":"

The state of the DB security group does not allow deletion.

", + "refs":{ + } + }, + "InvalidDBSnapshotStateFault":{ + "base":"

The state of the DB snapshot does not allow deletion.

", + "refs":{ + } + }, + "InvalidDBSubnetGroupStateFault":{ + "base":"

The DB subnet group cannot be deleted because it is in use.

", + "refs":{ + } + }, + "InvalidDBSubnetStateFault":{ + "base":"

The DB subnet is not in the available state.

", + "refs":{ + } + }, + "InvalidEventSubscriptionStateFault":{ + "base":null, + "refs":{ + } + }, + "InvalidRestoreFault":{ + "base":"

Cannot restore from vpc backup to non-vpc DB instance.

", + "refs":{ + } + }, + "InvalidSubnet":{ + "base":"

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", + "refs":{ + } + }, + "InvalidVPCNetworkStateFault":{ + "base":"

DB subnet group does not cover all Availability Zones after it is created because users' change.

", + "refs":{ + } + }, + "KMSKeyNotAccessibleFault":{ + "base":"

Error accessing KMS key.

", + "refs":{ + } + }, + "KeyList":{ + "base":null, + "refs":{ + "RemoveTagsFromResourceMessage$TagKeys":"

The tag key (name) of the tag to be removed.

" + } + }, + "ListTagsForResourceMessage":{ + "base":"

", + "refs":{ + } + }, + "LogTypeList":{ + "base":null, + "refs":{ + "CloudwatchLogsExportConfiguration$EnableLogTypes":"

The list of log types to enable.

", + "CloudwatchLogsExportConfiguration$DisableLogTypes":"

The list of log types to disable.

", + "CreateDBInstanceMessage$EnableCloudwatchLogsExports":"

The list of log types that need to be enabled for exporting to CloudWatch Logs.

", + "DBEngineVersion$ExportableLogTypes":"

The types of logs that the database engine has available for export to CloudWatch Logs.

", + "DBInstance$EnabledCloudwatchLogsExports":"

A list of log types that this DB instance is configured to export to CloudWatch Logs.

", + "PendingCloudwatchLogsExports$LogTypesToEnable":"

Log types that are in the process of being deactivated. After they are deactivated, these log types aren't exported to CloudWatch Logs.

", + "PendingCloudwatchLogsExports$LogTypesToDisable":"

Log types that are in the process of being enabled. After they are enabled, these log types are exported to CloudWatch Logs.

" + } + }, + "ModifyDBClusterMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyDBClusterResult":{ + "base":null, + "refs":{ + } + }, + "ModifyDBClusterSnapshotAttributeMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyDBClusterSnapshotAttributeResult":{ + "base":null, + "refs":{ + } + }, + "ModifyDBInstanceMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyDBInstanceResult":{ + "base":null, + "refs":{ + } + }, + "ModifyDBParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyDBSubnetGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyDBSubnetGroupResult":{ + "base":null, + "refs":{ + } + }, + "ModifyEventSubscriptionMessage":{ + "base":"

", + "refs":{ + } + }, + "ModifyEventSubscriptionResult":{ + "base":null, + "refs":{ + } + }, + "OptionGroupMembership":{ + "base":"

Provides information on the option groups the DB instance is a member of.

", + "refs":{ + "OptionGroupMembershipList$member":null + } + }, + "OptionGroupMembershipList":{ + "base":null, + "refs":{ + "DBInstance$OptionGroupMemberships":"

Provides the list of option group memberships for this DB instance.

" + } + }, + "OptionGroupNotFoundFault":{ + "base":null, + "refs":{ + } + }, + "OrderableDBInstanceOption":{ + "base":"

Contains a list of available options for a DB instance.

This data type is used as a response element in the DescribeOrderableDBInstanceOptions action.

", + "refs":{ + "OrderableDBInstanceOptionsList$member":null + } + }, + "OrderableDBInstanceOptionsList":{ + "base":null, + "refs":{ + "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions":"

An OrderableDBInstanceOption structure containing information about orderable options for the DB instance.

" + } + }, + "OrderableDBInstanceOptionsMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeOrderableDBInstanceOptions action.

", + "refs":{ + } + }, + "Parameter":{ + "base":"

This data type is used as a request parameter in the ModifyDBParameterGroup and ResetDBParameterGroup actions.

This data type is used as a response element in the DescribeEngineDefaultParameters and DescribeDBParameters actions.

", + "refs":{ + "ParametersList$member":null + } + }, + "ParametersList":{ + "base":null, + "refs":{ + "DBClusterParameterGroupDetails$Parameters":"

Provides a list of parameters for the DB cluster parameter group.

", + "DBParameterGroupDetails$Parameters":"

A list of Parameter values.

", + "EngineDefaults$Parameters":"

Contains a list of engine default parameters.

", + "ModifyDBClusterParameterGroupMessage$Parameters":"

A list of parameters in the DB cluster parameter group to modify.

", + "ModifyDBParameterGroupMessage$Parameters":"

An array of parameter names, values, and the apply method for the parameter update. At least one parameter name, value, and apply method must be supplied; subsequent arguments are optional. A maximum of 20 parameters can be modified in a single request.

Valid Values (for the application method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when you reboot the DB instance without failover.

", + "ResetDBClusterParameterGroupMessage$Parameters":"

A list of parameter names in the DB cluster parameter group to reset to the default values. You can't use this parameter if the ResetAllParameters parameter is set to true.

", + "ResetDBParameterGroupMessage$Parameters":"

To reset the entire DB parameter group, specify the DBParameterGroup name and ResetAllParameters parameters. To reset specific parameters, provide a list of the following: ParameterName and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

Valid Values (for Apply method): pending-reboot

" + } + }, + "PendingCloudwatchLogsExports":{ + "base":"

A list of the log types whose configuration is still pending. In other words, these log types are in the process of being activated or deactivated.

", + "refs":{ + "PendingModifiedValues$PendingCloudwatchLogsExports":null + } + }, + "PendingMaintenanceAction":{ + "base":"

Provides information about a pending maintenance action for a resource.

", + "refs":{ + "PendingMaintenanceActionDetails$member":null + } + }, + "PendingMaintenanceActionDetails":{ + "base":null, + "refs":{ + "ResourcePendingMaintenanceActions$PendingMaintenanceActionDetails":"

A list that provides details about the pending maintenance actions for the resource.

" + } + }, + "PendingMaintenanceActions":{ + "base":null, + "refs":{ + "PendingMaintenanceActionsMessage$PendingMaintenanceActions":"

A list of the pending maintenance actions for the resource.

" + } + }, + "PendingMaintenanceActionsMessage":{ + "base":"

Data returned from the DescribePendingMaintenanceActions action.

", + "refs":{ + } + }, + "PendingModifiedValues":{ + "base":"

This data type is used as a response element in the ModifyDBInstance action.

", + "refs":{ + "DBInstance$PendingModifiedValues":"

Specifies that changes to the DB instance are pending. This element is only included when changes are pending. Specific changes are identified by subelements.

" + } + }, + "PromoteReadReplicaDBClusterMessage":{ + "base":"

", + "refs":{ + } + }, + "PromoteReadReplicaDBClusterResult":{ + "base":null, + "refs":{ + } + }, + "ProvisionedIopsNotAvailableInAZFault":{ + "base":"

Provisioned IOPS not available in the specified Availability Zone.

", + "refs":{ + } + }, + "Range":{ + "base":"

A range of integer values.

", + "refs":{ + "RangeList$member":null + } + }, + "RangeList":{ + "base":null, + "refs":{ + "ValidStorageOptions$StorageSize":"

The valid range of storage in gibibytes. For example, 100 to 16384.

", + "ValidStorageOptions$ProvisionedIops":"

The valid range of provisioned IOPS. For example, 1000-20000.

" + } + }, + "ReadReplicaDBClusterIdentifierList":{ + "base":null, + "refs":{ + "DBInstance$ReadReplicaDBClusterIdentifiers":"

Contains one or more identifiers of DB clusters that are Read Replicas of this DB instance.

" + } + }, + "ReadReplicaDBInstanceIdentifierList":{ + "base":null, + "refs":{ + "DBInstance$ReadReplicaDBInstanceIdentifiers":"

Contains one or more identifiers of the Read Replicas associated with this DB instance.

" + } + }, + "ReadReplicaIdentifierList":{ + "base":null, + "refs":{ + "DBCluster$ReadReplicaIdentifiers":"

Contains one or more identifiers of the Read Replicas associated with this DB cluster.

" + } + }, + "RebootDBInstanceMessage":{ + "base":"

", + "refs":{ + } + }, + "RebootDBInstanceResult":{ + "base":null, + "refs":{ + } + }, + "RemoveRoleFromDBClusterMessage":{ + "base":null, + "refs":{ + } + }, + "RemoveSourceIdentifierFromSubscriptionMessage":{ + "base":"

", + "refs":{ + } + }, + "RemoveSourceIdentifierFromSubscriptionResult":{ + "base":null, + "refs":{ + } + }, + "RemoveTagsFromResourceMessage":{ + "base":"

", + "refs":{ + } + }, + "ResetDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "ResetDBParameterGroupMessage":{ + "base":"

", + "refs":{ + } + }, + "ResourceNotFoundFault":{ + "base":"

The specified resource ID was not found.

", + "refs":{ + } + }, + "ResourcePendingMaintenanceActions":{ + "base":"

Describes the pending maintenance actions for a resource.

", + "refs":{ + "ApplyPendingMaintenanceActionResult$ResourcePendingMaintenanceActions":null, + "PendingMaintenanceActions$member":null + } + }, + "RestoreDBClusterFromSnapshotMessage":{ + "base":"

", + "refs":{ + } + }, + "RestoreDBClusterFromSnapshotResult":{ + "base":null, + "refs":{ + } + }, + "RestoreDBClusterToPointInTimeMessage":{ + "base":"

", + "refs":{ + } + }, + "RestoreDBClusterToPointInTimeResult":{ + "base":null, + "refs":{ + } + }, + "SNSInvalidTopicFault":{ + "base":null, + "refs":{ + } + }, + "SNSNoAuthorizationFault":{ + "base":null, + "refs":{ + } + }, + "SNSTopicArnNotFoundFault":{ + "base":null, + "refs":{ + } + }, + "SharedSnapshotQuotaExceededFault":{ + "base":"

You have exceeded the maximum number of accounts that you can share a manual DB snapshot with.

", + "refs":{ + } + }, + "SnapshotQuotaExceededFault":{ + "base":"

Request would result in user exceeding the allowed number of DB snapshots.

", + "refs":{ + } + }, + "SourceIdsList":{ + "base":null, + "refs":{ + "CreateEventSubscriptionMessage$SourceIds":"

The list of identifiers of the event sources for which events are returned. If not specified, then all sources are included in the response. An identifier must begin with a letter and must contain only ASCII letters, digits, and hyphens; it can't end with a hyphen or contain two consecutive hyphens.

Constraints:

  • If SourceIds are supplied, SourceType must also be provided.

  • If the source type is a DB instance, then a DBInstanceIdentifier must be supplied.

  • If the source type is a DB security group, a DBSecurityGroupName must be supplied.

  • If the source type is a DB parameter group, a DBParameterGroupName must be supplied.

  • If the source type is a DB snapshot, a DBSnapshotIdentifier must be supplied.

", + "EventSubscription$SourceIdsList":"

A list of source IDs for the event notification subscription.

" + } + }, + "SourceNotFoundFault":{ + "base":null, + "refs":{ + } + }, + "SourceType":{ + "base":null, + "refs":{ + "DescribeEventsMessage$SourceType":"

The event source to retrieve events for. If no value is specified, all events are returned.

", + "Event$SourceType":"

Specifies the source type for this event.

" + } + }, + "StorageQuotaExceededFault":{ + "base":"

Request would result in user exceeding the allowed amount of storage available across all DB instances.

", + "refs":{ + } + }, + "StorageTypeNotSupportedFault":{ + "base":"

StorageType specified cannot be associated with the DB Instance.

", + "refs":{ + } + }, + "String":{ + "base":null, + "refs":{ + "AddRoleToDBClusterMessage$DBClusterIdentifier":"

The name of the DB cluster to associate the IAM role with.

", + "AddRoleToDBClusterMessage$RoleArn":"

The Amazon Resource Name (ARN) of the IAM role to associate with the Neptune DB cluster, for example arn:aws:iam::123456789012:role/NeptuneAccessRole.

", + "AddSourceIdentifierToSubscriptionMessage$SubscriptionName":"

The name of the event notification subscription you want to add a source identifier to.

", + "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier":"

The identifier of the event source to be added.

Constraints:

  • If the source type is a DB instance, then a DBInstanceIdentifier must be supplied.

  • If the source type is a DB security group, a DBSecurityGroupName must be supplied.

  • If the source type is a DB parameter group, a DBParameterGroupName must be supplied.

  • If the source type is a DB snapshot, a DBSnapshotIdentifier must be supplied.

", + "AddTagsToResourceMessage$ResourceName":"

The Amazon Neptune resource that the tags are added to. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an Amazon Resource Name (ARN).

", + "ApplyPendingMaintenanceActionMessage$ResourceIdentifier":"

The Amazon Resource Name (ARN) of the resource that the pending maintenance action applies to. For information about creating an ARN, see Constructing an Amazon Resource Name (ARN).

", + "ApplyPendingMaintenanceActionMessage$ApplyAction":"

The pending maintenance action to apply to this resource.

Valid values: system-update, db-upgrade

", + "ApplyPendingMaintenanceActionMessage$OptInType":"

A value that specifies the type of opt-in request, or undoes an opt-in request. An opt-in request of type immediate can't be undone.

Valid values:

  • immediate - Apply the maintenance action immediately.

  • next-maintenance - Apply the maintenance action during the next maintenance window for the resource.

  • undo-opt-in - Cancel any existing next-maintenance opt-in requests.

", + "AttributeValueList$member":null, + "AvailabilityZone$Name":"

The name of the availability zone.

", + "AvailabilityZones$member":null, + "CharacterSet$CharacterSetName":"

The name of the character set.

", + "CharacterSet$CharacterSetDescription":"

The description of the character set.

", + "CopyDBClusterParameterGroupMessage$SourceDBClusterParameterGroupIdentifier":"

The identifier or Amazon Resource Name (ARN) for the source DB cluster parameter group. For information about creating an ARN, see Constructing an Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid DB cluster parameter group.

  • If the source DB cluster parameter group is in the same AWS Region as the copy, specify a valid DB parameter group identifier, for example my-db-cluster-param-group, or a valid ARN.

  • If the source DB parameter group is in a different AWS Region than the copy, specify a valid DB cluster parameter group ARN, for example arn:aws:rds:us-east-1:123456789012:cluster-pg:custom-cluster-group1.

", + "CopyDBClusterParameterGroupMessage$TargetDBClusterParameterGroupIdentifier":"

The identifier for the copied DB cluster parameter group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-cluster-param-group1

", + "CopyDBClusterParameterGroupMessage$TargetDBClusterParameterGroupDescription":"

A description for the copied DB cluster parameter group.

", + "CopyDBClusterSnapshotMessage$SourceDBClusterSnapshotIdentifier":"

The identifier of the DB cluster snapshot to copy. This parameter is not case-sensitive.

You can't copy an encrypted, shared DB cluster snapshot from one AWS Region to another.

Constraints:

  • Must specify a valid system snapshot in the \"available\" state.

  • If the source snapshot is in the same AWS Region as the copy, specify a valid DB snapshot identifier.

  • If the source snapshot is in a different AWS Region than the copy, specify a valid DB cluster snapshot ARN.

Example: my-cluster-snapshot1

", + "CopyDBClusterSnapshotMessage$TargetDBClusterSnapshotIdentifier":"

The identifier of the new DB cluster snapshot to create from the source DB cluster snapshot. This parameter is not case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster-snapshot2

", + "CopyDBClusterSnapshotMessage$KmsKeyId":"

The AWS AWS KMS key ID for an encrypted DB cluster snapshot. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you copy an unencrypted DB cluster snapshot and specify a value for the KmsKeyId parameter, Amazon Neptune encrypts the target DB cluster snapshot using the specified KMS encryption key.

If you copy an encrypted DB cluster snapshot from your AWS account, you can specify a value for KmsKeyId to encrypt the copy with a new KMS encryption key. If you don't specify a value for KmsKeyId, then the copy of the DB cluster snapshot is encrypted with the same KMS key as the source DB cluster snapshot.

If you copy an encrypted DB cluster snapshot that is shared from another AWS account, then you must specify a value for KmsKeyId.

To copy an encrypted DB cluster snapshot to another AWS Region, you must set KmsKeyId to the KMS key ID you want to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

", + "CopyDBClusterSnapshotMessage$PreSignedUrl":"

The URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot API action in the AWS Region that contains the source DB cluster snapshot to copy. The PreSignedUrl parameter must be used when copying an encrypted DB cluster snapshot from another AWS Region.

The pre-signed URL must be a valid request for the CopyDBSClusterSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB cluster snapshot to be copied. The pre-signed URL request must contain the following parameter values:

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. This is the same identifier for both the CopyDBClusterSnapshot action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

  • DestinationRegion - The name of the AWS Region that the DB cluster snapshot will be created in.

  • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster snapshot from the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:neptune-cluster1-snapshot-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", + "CopyDBParameterGroupMessage$SourceDBParameterGroupIdentifier":"

The identifier or ARN for the source DB parameter group. For information about creating an ARN, see Constructing an Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid DB parameter group.

  • Must specify a valid DB parameter group identifier, for example my-db-param-group, or a valid ARN.

", + "CopyDBParameterGroupMessage$TargetDBParameterGroupIdentifier":"

The identifier for the copied DB parameter group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-db-parameter-group

", + "CopyDBParameterGroupMessage$TargetDBParameterGroupDescription":"

A description for the copied DB parameter group.

", + "CreateDBClusterMessage$CharacterSetName":"

A value that indicates that the DB cluster should be associated with the specified CharacterSet.

", + "CreateDBClusterMessage$DatabaseName":"

The name for your database of up to 64 alpha-numeric characters. If you do not provide a name, Amazon Neptune will not create a database in the DB cluster you are creating.

", + "CreateDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", + "CreateDBClusterMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to associate with this DB cluster. If this argument is omitted, the default is used.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "CreateDBClusterMessage$DBSubnetGroupName":"

A DB subnet group to associate with this DB cluster.

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", + "CreateDBClusterMessage$Engine":"

The name of the database engine to be used for this DB cluster.

Valid Values: neptune

", + "CreateDBClusterMessage$EngineVersion":"

The version number of the database engine to use.

Example: 1.0.1

", + "CreateDBClusterMessage$MasterUsername":"

The name of the master user for the DB cluster.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", + "CreateDBClusterMessage$MasterUserPassword":"

The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", + "CreateDBClusterMessage$OptionGroupName":"

A value that indicates that the DB cluster should be associated with the specified option group.

Permanent options can't be removed from an option group. The option group can't be removed from a DB cluster once it is associated with a DB cluster.

", + "CreateDBClusterMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon Neptune User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "CreateDBClusterMessage$PreferredMaintenanceWindow":"

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon Neptune User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "CreateDBClusterMessage$ReplicationSourceIdentifier":"

The Amazon Resource Name (ARN) of the source DB instance or DB cluster if this DB cluster is created as a Read Replica.

", + "CreateDBClusterMessage$KmsKeyId":"

The AWS KMS key identifier for an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

If an encryption key is not specified in KmsKeyId:

  • If ReplicationSourceIdentifier identifies an encrypted source, then Amazon Neptune will use the encryption key used to encrypt the source. Otherwise, Amazon Neptune will use your default encryption key.

  • If the StorageEncrypted parameter is true and ReplicationSourceIdentifier is not specified, then Amazon Neptune will use your default encryption key.

AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

If you create a Read Replica of an encrypted DB cluster in another AWS Region, you must set KmsKeyId to a KMS key ID that is valid in the destination AWS Region. This key is used to encrypt the Read Replica in that AWS Region.

", + "CreateDBClusterMessage$PreSignedUrl":"

A URL that contains a Signature Version 4 signed request for the CreateDBCluster action to be called in the source AWS Region where the DB cluster is replicated from. You only need to specify PreSignedUrl when you are performing cross-region replication from an encrypted DB cluster.

The pre-signed URL must be a valid request for the CreateDBCluster API action that can be executed in the source AWS Region that contains the encrypted DB cluster to be copied.

The pre-signed URL request must contain the following parameter values:

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB cluster in the destination AWS Region. This should refer to the same KMS key for both the CreateDBCluster action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

  • DestinationRegion - The name of the AWS Region that Read Replica will be created in.

  • ReplicationSourceIdentifier - The DB cluster identifier for the encrypted DB cluster to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster from the us-west-2 AWS Region, then your ReplicationSourceIdentifier would look like Example: arn:aws:rds:us-west-2:123456789012:cluster:neptune-cluster1.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", + "CreateDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group.

Constraints:

  • Must match the name of an existing DBClusterParameterGroup.

This value is stored as a lowercase string.

", + "CreateDBClusterParameterGroupMessage$DBParameterGroupFamily":"

The DB cluster parameter group family name. A DB cluster parameter group can be associated with one and only one DB cluster parameter group family, and can be applied only to a DB cluster running a database engine and engine version compatible with that DB cluster parameter group family.

", + "CreateDBClusterParameterGroupMessage$Description":"

The description for the DB cluster parameter group.

", + "CreateDBClusterSnapshotMessage$DBClusterSnapshotIdentifier":"

The identifier of the DB cluster snapshot. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1-snapshot1

", + "CreateDBClusterSnapshotMessage$DBClusterIdentifier":"

The identifier of the DB cluster to create a snapshot for. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

Example: my-cluster1

", + "CreateDBInstanceMessage$DBName":"

The database name.

Type: String

", + "CreateDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", + "CreateDBInstanceMessage$DBInstanceClass":"

The compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions.

", + "CreateDBInstanceMessage$Engine":"

The name of the database engine to be used for this instance.

Valid Values: neptune

", + "CreateDBInstanceMessage$MasterUsername":"

The name for the master user. Not used.

", + "CreateDBInstanceMessage$MasterUserPassword":"

The password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Not used.

", + "CreateDBInstanceMessage$AvailabilityZone":"

The EC2 Availability Zone that the DB instance is created in.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

Constraint: The AvailabilityZone parameter can't be specified if the MultiAZ parameter is set to true. The specified Availability Zone must be in the same AWS Region as the current endpoint.

", + "CreateDBInstanceMessage$DBSubnetGroupName":"

A DB subnet group to associate with this DB instance.

If there is no DB subnet group, then it is a non-VPC DB instance.

", + "CreateDBInstanceMessage$PreferredMaintenanceWindow":"

The time range each week during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "CreateDBInstanceMessage$DBParameterGroupName":"

The name of the DB parameter group to associate with this DB instance. If this argument is omitted, the default DBParameterGroup for the specified engine is used.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "CreateDBInstanceMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created.

Not applicable. The daily time range for creating automated backups is managed by the DB cluster. For more information, see CreateDBCluster.

", + "CreateDBInstanceMessage$EngineVersion":"

The version number of the database engine to use.

", + "CreateDBInstanceMessage$LicenseModel":"

License model information for this DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", + "CreateDBInstanceMessage$OptionGroupName":"

Indicates that the DB instance should be associated with the specified option group.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "CreateDBInstanceMessage$CharacterSetName":"

Indicates that the DB instance should be associated with the specified CharacterSet.

Not applicable. The character set is managed by the DB cluster. For more information, see CreateDBCluster.

", + "CreateDBInstanceMessage$DBClusterIdentifier":"

The identifier of the DB cluster that the instance will belong to.

For information on creating a DB cluster, see CreateDBCluster.

Type: String

", + "CreateDBInstanceMessage$StorageType":"

Specifies the storage type to be associated with the DB instance.

Not applicable. Storage is managed by the DB Cluster.

", + "CreateDBInstanceMessage$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "CreateDBInstanceMessage$TdeCredentialPassword":"

The password for the given ARN from the key store in order to access the device.

", + "CreateDBInstanceMessage$KmsKeyId":"

The AWS KMS key identifier for an encrypted DB instance.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB instance with the same AWS account that owns the KMS encryption key used to encrypt the new DB instance, then you can use the KMS key alias instead of the ARN for the KM encryption key.

Not applicable. The KMS key identifier is managed by the DB cluster. For more information, see CreateDBCluster.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon Neptune will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", + "CreateDBInstanceMessage$Domain":"

Specify the Active Directory Domain to create the instance in.

", + "CreateDBInstanceMessage$MonitoringRoleArn":"

The ARN for the IAM role that permits Neptune to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", + "CreateDBInstanceMessage$DomainIAMRoleName":"

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", + "CreateDBInstanceMessage$Timezone":"

The time zone of the DB instance.

", + "CreateDBInstanceMessage$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "CreateDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

This value is stored as a lowercase string.

", + "CreateDBParameterGroupMessage$DBParameterGroupFamily":"

The DB parameter group family name. A DB parameter group can be associated with one and only one DB parameter group family, and can be applied only to a DB instance running a database engine and engine version compatible with that DB parameter group family.

", + "CreateDBParameterGroupMessage$Description":"

The description for the DB parameter group.

", + "CreateDBSubnetGroupMessage$DBSubnetGroupName":"

The name for the DB subnet group. This value is stored as a lowercase string.

Constraints: Must contain no more than 255 letters, numbers, periods, underscores, spaces, or hyphens. Must not be default.

Example: mySubnetgroup

", + "CreateDBSubnetGroupMessage$DBSubnetGroupDescription":"

The description for the DB subnet group.

", + "CreateEventSubscriptionMessage$SubscriptionName":"

The name of the subscription.

Constraints: The name must be less than 255 characters.

", + "CreateEventSubscriptionMessage$SnsTopicArn":"

The Amazon Resource Name (ARN) of the SNS topic created for event notification. The ARN is created by Amazon SNS when you create a topic and subscribe to it.

", + "CreateEventSubscriptionMessage$SourceType":"

The type of source that is generating the events. For example, if you want to be notified of events generated by a DB instance, you would set this parameter to db-instance. if this value is not specified, all events are returned.

Valid values: db-instance | db-cluster | db-parameter-group | db-security-group | db-snapshot | db-cluster-snapshot

", + "DBCluster$CharacterSetName":"

If present, specifies the name of the character set that this cluster is associated with.

", + "DBCluster$DatabaseName":"

Contains the name of the initial database of this DB cluster that was provided at create time, if one was specified when the DB cluster was created. This same name is returned for the life of the DB cluster.

", + "DBCluster$DBClusterIdentifier":"

Contains a user-supplied DB cluster identifier. This identifier is the unique key that identifies a DB cluster.

", + "DBCluster$DBClusterParameterGroup":"

Specifies the name of the DB cluster parameter group for the DB cluster.

", + "DBCluster$DBSubnetGroup":"

Specifies information on the subnet group associated with the DB cluster, including the name, description, and subnets in the subnet group.

", + "DBCluster$Status":"

Specifies the current state of this DB cluster.

", + "DBCluster$PercentProgress":"

Specifies the progress of the operation as a percentage.

", + "DBCluster$Endpoint":"

Specifies the connection endpoint for the primary instance of the DB cluster.

", + "DBCluster$ReaderEndpoint":"

The reader endpoint for the DB cluster. The reader endpoint for a DB cluster load-balances connections across the Read Replicas that are available in a DB cluster. As clients request new connections to the reader endpoint, Neptune distributes the connection requests among the Read Replicas in the DB cluster. This functionality can help balance your read workload across multiple Read Replicas in your DB cluster.

If a failover occurs, and the Read Replica that you are connected to is promoted to be the primary instance, your connection is dropped. To continue sending your read workload to other Read Replicas in the cluster, you can then reconnect to the reader endpoint.

", + "DBCluster$Engine":"

Provides the name of the database engine to be used for this DB cluster.

", + "DBCluster$EngineVersion":"

Indicates the database engine version.

", + "DBCluster$MasterUsername":"

Contains the master username for the DB cluster.

", + "DBCluster$PreferredBackupWindow":"

Specifies the daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod.

", + "DBCluster$PreferredMaintenanceWindow":"

Specifies the weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

", + "DBCluster$ReplicationSourceIdentifier":"

Contains the identifier of the source DB cluster if this DB cluster is a Read Replica.

", + "DBCluster$HostedZoneId":"

Specifies the ID that Amazon Route 53 assigns when you create a hosted zone.

", + "DBCluster$KmsKeyId":"

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB cluster.

", + "DBCluster$DbClusterResourceId":"

The AWS Region-unique, immutable identifier for the DB cluster. This identifier is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB cluster is accessed.

", + "DBCluster$DBClusterArn":"

The Amazon Resource Name (ARN) for the DB cluster.

", + "DBCluster$CloneGroupId":"

Identifies the clone group to which the DB cluster is associated.

", + "DBClusterMember$DBInstanceIdentifier":"

Specifies the instance identifier for this member of the DB cluster.

", + "DBClusterMember$DBClusterParameterGroupStatus":"

Specifies the status of the DB cluster parameter group for this member of the DB cluster.

", + "DBClusterMessage$Marker":"

A pagination token that can be used in a subsequent DescribeDBClusters request.

", + "DBClusterOptionGroupStatus$DBClusterOptionGroupName":"

Specifies the name of the DB cluster option group.

", + "DBClusterOptionGroupStatus$Status":"

Specifies the status of the DB cluster option group.

", + "DBClusterParameterGroup$DBClusterParameterGroupName":"

Provides the name of the DB cluster parameter group.

", + "DBClusterParameterGroup$DBParameterGroupFamily":"

Provides the name of the DB parameter group family that this DB cluster parameter group is compatible with.

", + "DBClusterParameterGroup$Description":"

Provides the customer-specified description for this DB cluster parameter group.

", + "DBClusterParameterGroup$DBClusterParameterGroupArn":"

The Amazon Resource Name (ARN) for the DB cluster parameter group.

", + "DBClusterParameterGroupDetails$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DBClusterParameterGroupNameMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group.

Constraints:

  • Must be 1 to 255 letters or numbers.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

This value is stored as a lowercase string.

", + "DBClusterParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBClusterRole$RoleArn":"

The Amazon Resource Name (ARN) of the IAM role that is associated with the DB cluster.

", + "DBClusterRole$Status":"

Describes the state of association between the IAM role and the DB cluster. The Status property returns one of the following values:

  • ACTIVE - the IAM role ARN is associated with the DB cluster and can be used to access other AWS services on your behalf.

  • PENDING - the IAM role ARN is being associated with the DB cluster.

  • INVALID - the IAM role ARN is associated with the DB cluster, but the DB cluster is unable to assume the IAM role in order to access other AWS services on your behalf.

", + "DBClusterSnapshot$DBClusterSnapshotIdentifier":"

Specifies the identifier for the DB cluster snapshot.

", + "DBClusterSnapshot$DBClusterIdentifier":"

Specifies the DB cluster identifier of the DB cluster that this DB cluster snapshot was created from.

", + "DBClusterSnapshot$Engine":"

Specifies the name of the database engine.

", + "DBClusterSnapshot$Status":"

Specifies the status of this DB cluster snapshot.

", + "DBClusterSnapshot$VpcId":"

Provides the VPC ID associated with the DB cluster snapshot.

", + "DBClusterSnapshot$MasterUsername":"

Provides the master username for the DB cluster snapshot.

", + "DBClusterSnapshot$EngineVersion":"

Provides the version of the database engine for this DB cluster snapshot.

", + "DBClusterSnapshot$LicenseModel":"

Provides the license model information for this DB cluster snapshot.

", + "DBClusterSnapshot$SnapshotType":"

Provides the type of the DB cluster snapshot.

", + "DBClusterSnapshot$KmsKeyId":"

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB cluster snapshot.

", + "DBClusterSnapshot$DBClusterSnapshotArn":"

The Amazon Resource Name (ARN) for the DB cluster snapshot.

", + "DBClusterSnapshot$SourceDBClusterSnapshotArn":"

If the DB cluster snapshot was copied from a source DB cluster snapshot, the Amazon Resource Name (ARN) for the source DB cluster snapshot, otherwise, a null value.

", + "DBClusterSnapshotAttribute$AttributeName":"

The name of the manual DB cluster snapshot attribute.

The attribute named restore refers to the list of AWS accounts that have permission to copy or restore the manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", + "DBClusterSnapshotAttributesResult$DBClusterSnapshotIdentifier":"

The identifier of the manual DB cluster snapshot that the attributes apply to.

", + "DBClusterSnapshotMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBEngineVersion$Engine":"

The name of the database engine.

", + "DBEngineVersion$EngineVersion":"

The version number of the database engine.

", + "DBEngineVersion$DBParameterGroupFamily":"

The name of the DB parameter group family for the database engine.

", + "DBEngineVersion$DBEngineDescription":"

The description of the database engine.

", + "DBEngineVersion$DBEngineVersionDescription":"

The description of the database engine version.

", + "DBEngineVersionMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBInstance$DBInstanceIdentifier":"

Contains a user-supplied database identifier. This identifier is the unique key that identifies a DB instance.

", + "DBInstance$DBInstanceClass":"

Contains the name of the compute and memory capacity class of the DB instance.

", + "DBInstance$Engine":"

Provides the name of the database engine to be used for this DB instance.

", + "DBInstance$DBInstanceStatus":"

Specifies the current state of this database.

", + "DBInstance$MasterUsername":"

Contains the master username for the DB instance.

", + "DBInstance$DBName":"

The database name.

", + "DBInstance$PreferredBackupWindow":"

Specifies the daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod.

", + "DBInstance$AvailabilityZone":"

Specifies the name of the Availability Zone the DB instance is located in.

", + "DBInstance$PreferredMaintenanceWindow":"

Specifies the weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

", + "DBInstance$EngineVersion":"

Indicates the database engine version.

", + "DBInstance$ReadReplicaSourceDBInstanceIdentifier":"

Contains the identifier of the source DB instance if this DB instance is a Read Replica.

", + "DBInstance$LicenseModel":"

License model information for this DB instance.

", + "DBInstance$CharacterSetName":"

If present, specifies the name of the character set that this instance is associated with.

", + "DBInstance$SecondaryAvailabilityZone":"

If present, specifies the name of the secondary Availability Zone for a DB instance with multi-AZ support.

", + "DBInstance$StorageType":"

Specifies the storage type associated with DB instance.

", + "DBInstance$TdeCredentialArn":"

The ARN from the key store with which the instance is associated for TDE encryption.

", + "DBInstance$DBClusterIdentifier":"

If the DB instance is a member of a DB cluster, contains the name of the DB cluster that the DB instance is a member of.

", + "DBInstance$KmsKeyId":"

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB instance.

", + "DBInstance$DbiResourceId":"

The AWS Region-unique, immutable identifier for the DB instance. This identifier is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB instance is accessed.

", + "DBInstance$CACertificateIdentifier":"

The identifier of the CA certificate for this DB instance.

", + "DBInstance$EnhancedMonitoringResourceArn":"

The Amazon Resource Name (ARN) of the Amazon CloudWatch Logs log stream that receives the Enhanced Monitoring metrics data for the DB instance.

", + "DBInstance$MonitoringRoleArn":"

The ARN for the IAM role that permits Neptune to send Enhanced Monitoring metrics to Amazon CloudWatch Logs.

", + "DBInstance$DBInstanceArn":"

The Amazon Resource Name (ARN) for the DB instance.

", + "DBInstance$Timezone":"

Not supported.

", + "DBInstance$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "DBInstanceMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DBInstanceStatusInfo$StatusType":"

This value is currently \"read replication.\"

", + "DBInstanceStatusInfo$Status":"

Status of the DB instance. For a StatusType of read replica, the values can be replicating, error, stopped, or terminated.

", + "DBInstanceStatusInfo$Message":"

Details of the error if there is an error for the instance. If the instance is not in an error state, this value is blank.

", + "DBParameterGroup$DBParameterGroupName":"

Provides the name of the DB parameter group.

", + "DBParameterGroup$DBParameterGroupFamily":"

Provides the name of the DB parameter group family that this DB parameter group is compatible with.

", + "DBParameterGroup$Description":"

Provides the customer-specified description for this DB parameter group.

", + "DBParameterGroup$DBParameterGroupArn":"

The Amazon Resource Name (ARN) for the DB parameter group.

", + "DBParameterGroupDetails$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBParameterGroupNameMessage$DBParameterGroupName":"

Provides the name of the DB parameter group.

", + "DBParameterGroupStatus$DBParameterGroupName":"

The name of the DP parameter group.

", + "DBParameterGroupStatus$ParameterApplyStatus":"

The status of parameter updates.

", + "DBParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBSecurityGroupMembership$DBSecurityGroupName":"

The name of the DB security group.

", + "DBSecurityGroupMembership$Status":"

The status of the DB security group.

", + "DBSecurityGroupNameList$member":null, + "DBSubnetGroup$DBSubnetGroupName":"

The name of the DB subnet group.

", + "DBSubnetGroup$DBSubnetGroupDescription":"

Provides the description of the DB subnet group.

", + "DBSubnetGroup$VpcId":"

Provides the VpcId of the DB subnet group.

", + "DBSubnetGroup$SubnetGroupStatus":"

Provides the status of the DB subnet group.

", + "DBSubnetGroup$DBSubnetGroupArn":"

The Amazon Resource Name (ARN) for the DB subnet group.

", + "DBSubnetGroupMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DeleteDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier for the DB cluster to be deleted. This parameter isn't case-sensitive.

Constraints:

  • Must match an existing DBClusterIdentifier.

", + "DeleteDBClusterMessage$FinalDBSnapshotIdentifier":"

The DB cluster snapshot identifier of the new DB cluster snapshot created when SkipFinalSnapshot is set to false.

Specifying this parameter and also setting the SkipFinalShapshot parameter to true results in an error.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "DeleteDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group.

Constraints:

  • Must be the name of an existing DB cluster parameter group.

  • You can't delete a default DB cluster parameter group.

  • Cannot be associated with any DB clusters.

", + "DeleteDBClusterSnapshotMessage$DBClusterSnapshotIdentifier":"

The identifier of the DB cluster snapshot to delete.

Constraints: Must be the name of an existing DB cluster snapshot in the available state.

", + "DeleteDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier for the DB instance to be deleted. This parameter isn't case-sensitive.

Constraints:

  • Must match the name of an existing DB instance.

", + "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier":"

The DBSnapshotIdentifier of the new DBSnapshot created when SkipFinalSnapshot is set to false.

Specifying this parameter and also setting the SkipFinalShapshot parameter to true results in an error.

Constraints:

  • Must be 1 to 255 letters or numbers.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Cannot be specified when deleting a Read Replica.

", + "DeleteDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • Must be the name of an existing DB parameter group

  • You can't delete a default DB parameter group

  • Cannot be associated with any DB instances

", + "DeleteDBSubnetGroupMessage$DBSubnetGroupName":"

The name of the database subnet group to delete.

You can't delete the default subnet group.

Constraints:

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", + "DeleteEventSubscriptionMessage$SubscriptionName":"

The name of the event notification subscription you want to delete.

", + "DescribeDBClusterParameterGroupsMessage$DBClusterParameterGroupName":"

The name of a specific DB cluster parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "DescribeDBClusterParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClusterParametersMessage$DBClusterParameterGroupName":"

The name of a specific DB cluster parameter group to return parameter details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "DescribeDBClusterParametersMessage$Source":"

A value that indicates to return only parameters for a specific source. Parameter sources can be engine, service, or customer.

", + "DescribeDBClusterParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClusterSnapshotAttributesMessage$DBClusterSnapshotIdentifier":"

The identifier for the DB cluster snapshot to describe the attributes for.

", + "DescribeDBClusterSnapshotsMessage$DBClusterIdentifier":"

The ID of the DB cluster to retrieve the list of DB cluster snapshots for. This parameter can't be used in conjunction with the DBClusterSnapshotIdentifier parameter. This parameter is not case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBCluster.

", + "DescribeDBClusterSnapshotsMessage$DBClusterSnapshotIdentifier":"

A specific DB cluster snapshot identifier to describe. This parameter can't be used in conjunction with the DBClusterIdentifier parameter. This value is stored as a lowercase string.

Constraints:

  • If supplied, must match the identifier of an existing DBClusterSnapshot.

  • If this identifier is for an automated snapshot, the SnapshotType parameter must also be specified.

", + "DescribeDBClusterSnapshotsMessage$SnapshotType":"

The type of DB cluster snapshots to be returned. You can specify one of the following values:

  • automated - Return all DB cluster snapshots that have been automatically taken by Amazon Neptune for my AWS account.

  • manual - Return all DB cluster snapshots that have been taken by my AWS account.

  • shared - Return all manual DB cluster snapshots that have been shared to my AWS account.

  • public - Return all DB cluster snapshots that have been marked as public.

If you don't specify a SnapshotType value, then both automated and manual DB cluster snapshots are returned. You can include shared DB cluster snapshots with these results by setting the IncludeShared parameter to true. You can include public DB cluster snapshots with these results by setting the IncludePublic parameter to true.

The IncludeShared and IncludePublic parameters don't apply for SnapshotType values of manual or automated. The IncludePublic parameter doesn't apply when SnapshotType is set to shared. The IncludeShared parameter doesn't apply when SnapshotType is set to public.

", + "DescribeDBClusterSnapshotsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClustersMessage$DBClusterIdentifier":"

The user-supplied DB cluster identifier. If this parameter is specified, information from only the specific DB cluster is returned. This parameter isn't case-sensitive.

Constraints:

  • If supplied, must match an existing DBClusterIdentifier.

", + "DescribeDBClustersMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBEngineVersionsMessage$Engine":"

The database engine to return.

", + "DescribeDBEngineVersionsMessage$EngineVersion":"

The database engine version to return.

Example: 5.1.49

", + "DescribeDBEngineVersionsMessage$DBParameterGroupFamily":"

The name of a specific DB parameter group family to return details for.

Constraints:

  • If supplied, must match an existing DBParameterGroupFamily.

", + "DescribeDBEngineVersionsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBInstancesMessage$DBInstanceIdentifier":"

The user-supplied instance identifier. If this parameter is specified, information from only the specific DB instance is returned. This parameter isn't case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBInstance.

", + "DescribeDBInstancesMessage$Marker":"

An optional pagination token provided by a previous DescribeDBInstances request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBParameterGroupsMessage$DBParameterGroupName":"

The name of a specific DB parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "DescribeDBParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBParametersMessage$DBParameterGroupName":"

The name of a specific DB parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBParameterGroup.

", + "DescribeDBParametersMessage$Source":"

The parameter types to return.

Default: All parameter types returned

Valid Values: user | system | engine-default

", + "DescribeDBParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeDBParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBSubnetGroupsMessage$DBSubnetGroupName":"

The name of the DB subnet group to return details for.

", + "DescribeDBSubnetGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBSubnetGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeEngineDefaultClusterParametersMessage$DBParameterGroupFamily":"

The name of the DB cluster parameter group family to return engine parameter information for.

", + "DescribeEngineDefaultClusterParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeEngineDefaultClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily":"

The name of the DB parameter group family.

", + "DescribeEngineDefaultParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeEngineDefaultParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeEventCategoriesMessage$SourceType":"

The type of source that is generating the events.

Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot

", + "DescribeEventSubscriptionsMessage$SubscriptionName":"

The name of the event notification subscription you want to describe.

", + "DescribeEventSubscriptionsMessage$Marker":"

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DescribeEventsMessage$SourceIdentifier":"

The identifier of the event source for which events are returned. If not specified, then all sources are included in the response.

Constraints:

  • If SourceIdentifier is supplied, SourceType must also be provided.

  • If the source type is DBInstance, then a DBInstanceIdentifier must be supplied.

  • If the source type is DBSecurityGroup, a DBSecurityGroupName must be supplied.

  • If the source type is DBParameterGroup, a DBParameterGroupName must be supplied.

  • If the source type is DBSnapshot, a DBSnapshotIdentifier must be supplied.

  • Cannot end with a hyphen or contain two consecutive hyphens.

", + "DescribeEventsMessage$Marker":"

An optional pagination token provided by a previous DescribeEvents request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeOrderableDBInstanceOptionsMessage$Engine":"

The name of the engine to retrieve DB instance options for.

", + "DescribeOrderableDBInstanceOptionsMessage$EngineVersion":"

The engine version filter value. Specify this parameter to show only the available offerings matching the specified engine version.

", + "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass":"

The DB instance class filter value. Specify this parameter to show only the available offerings matching the specified DB instance class.

", + "DescribeOrderableDBInstanceOptionsMessage$LicenseModel":"

The license model filter value. Specify this parameter to show only the available offerings matching the specified license model.

", + "DescribeOrderableDBInstanceOptionsMessage$Marker":"

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DescribePendingMaintenanceActionsMessage$ResourceIdentifier":"

The ARN of a resource to return pending maintenance actions for.

", + "DescribePendingMaintenanceActionsMessage$Marker":"

An optional pagination token provided by a previous DescribePendingMaintenanceActions request. If this parameter is specified, the response includes only records beyond the marker, up to a number of records specified by MaxRecords.

", + "DescribeValidDBInstanceModificationsMessage$DBInstanceIdentifier":"

The customer identifier or the ARN of your DB instance.

", + "DomainMembership$Domain":"

The identifier of the Active Directory Domain.

", + "DomainMembership$Status":"

The status of the DB instance's Active Directory Domain membership, such as joined, pending-join, failed etc).

", + "DomainMembership$FQDN":"

The fully qualified domain name of the Active Directory Domain.

", + "DomainMembership$IAMRoleName":"

The name of the IAM role to be used when making API calls to the Directory Service.

", + "Endpoint$Address":"

Specifies the DNS address of the DB instance.

", + "Endpoint$HostedZoneId":"

Specifies the ID that Amazon Route 53 assigns when you create a hosted zone.

", + "EngineDefaults$DBParameterGroupFamily":"

Specifies the name of the DB parameter group family that the engine default parameters apply to.

", + "EngineDefaults$Marker":"

An optional pagination token provided by a previous EngineDefaults request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "Event$SourceIdentifier":"

Provides the identifier for the source of the event.

", + "Event$Message":"

Provides the text of this event.

", + "Event$SourceArn":"

The Amazon Resource Name (ARN) for the event.

", + "EventCategoriesList$member":null, + "EventCategoriesMap$SourceType":"

The source type that the returned categories belong to

", + "EventSubscription$CustomerAwsId":"

The AWS customer account associated with the event notification subscription.

", + "EventSubscription$CustSubscriptionId":"

The event notification subscription Id.

", + "EventSubscription$SnsTopicArn":"

The topic ARN of the event notification subscription.

", + "EventSubscription$Status":"

The status of the event notification subscription.

Constraints:

Can be one of the following: creating | modifying | deleting | active | no-permission | topic-not-exist

The status \"no-permission\" indicates that Neptune no longer has permission to post to the SNS topic. The status \"topic-not-exist\" indicates that the topic was deleted after the subscription was created.

", + "EventSubscription$SubscriptionCreationTime":"

The time the event notification subscription was created.

", + "EventSubscription$SourceType":"

The source type for the event notification subscription.

", + "EventSubscription$EventSubscriptionArn":"

The Amazon Resource Name (ARN) for the event subscription.

", + "EventSubscriptionsMessage$Marker":"

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "EventsMessage$Marker":"

An optional pagination token provided by a previous Events request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "FailoverDBClusterMessage$DBClusterIdentifier":"

A DB cluster identifier to force a failover for. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

", + "FailoverDBClusterMessage$TargetDBInstanceIdentifier":"

The name of the instance to promote to the primary instance.

You must specify the instance identifier for an Read Replica in the DB cluster. For example, mydbcluster-replica1.

", + "Filter$Name":"

This parameter is not currently supported.

", + "FilterValueList$member":null, + "KeyList$member":null, + "ListTagsForResourceMessage$ResourceName":"

The Amazon Neptune resource with tags to be listed. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an Amazon Resource Name (ARN).

", + "LogTypeList$member":null, + "ModifyDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier for the cluster being modified. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

", + "ModifyDBClusterMessage$NewDBClusterIdentifier":"

The new DB cluster identifier for the DB cluster when renaming a DB cluster. This value is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • The first character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-cluster2

", + "ModifyDBClusterMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to use for the DB cluster.

", + "ModifyDBClusterMessage$MasterUserPassword":"

The new password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", + "ModifyDBClusterMessage$OptionGroupName":"

A value that indicates that the DB cluster should be associated with the specified option group. Changing this parameter doesn't result in an outage except in the following case, and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If the parameter change results in an option group that enables OEM, this change can cause a brief (sub-second) period during which new connections are rejected but existing connections are not interrupted.

Permanent options can't be removed from an option group. The option group can't be removed from a DB cluster once it is associated with a DB cluster.

", + "ModifyDBClusterMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "ModifyDBClusterMessage$PreferredMaintenanceWindow":"

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "ModifyDBClusterMessage$EngineVersion":"

The version number of the database engine to which you want to upgrade. Changing this parameter results in an outage. The change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true.

For a list of valid engine versions, see CreateDBInstance, or call DescribeDBEngineVersions.

", + "ModifyDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to modify.

", + "ModifyDBClusterSnapshotAttributeMessage$DBClusterSnapshotIdentifier":"

The identifier for the DB cluster snapshot to modify the attributes for.

", + "ModifyDBClusterSnapshotAttributeMessage$AttributeName":"

The name of the DB cluster snapshot attribute to modify.

To manage authorization for other AWS accounts to copy or restore a manual DB cluster snapshot, set this value to restore.

", + "ModifyDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier. This value is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "ModifyDBInstanceMessage$DBInstanceClass":"

The new compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions.

If you modify the DB instance class, an outage occurs during the change. The change is applied during the next maintenance window, unless ApplyImmediately is specified as true for this request.

Default: Uses existing setting

", + "ModifyDBInstanceMessage$DBSubnetGroupName":"

The new DB subnet group for the DB instance. You can use this parameter to move your DB instance to a different VPC.

Changing the subnet group causes an outage during the change. The change is applied during the next maintenance window, unless you specify true for the ApplyImmediately parameter.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetGroup

", + "ModifyDBInstanceMessage$MasterUserPassword":"

The new password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Not applicable.

Default: Uses existing setting

", + "ModifyDBInstanceMessage$DBParameterGroupName":"

The name of the DB parameter group to apply to the DB instance. Changing this setting doesn't result in an outage. The parameter group name itself is changed immediately, but the actual parameter changes are not applied until you reboot the instance without failover. The db instance will NOT be rebooted automatically and the parameter changes will NOT be applied during the next maintenance window.

Default: Uses existing setting

Constraints: The DB parameter group must be in the same DB parameter group family as this DB instance.

", + "ModifyDBInstanceMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled.

Not applicable. The daily time range for creating automated backups is managed by the DB cluster. For more information, see ModifyDBCluster.

Constraints:

  • Must be in the format hh24:mi-hh24:mi

  • Must be in Universal Time Coordinated (UTC)

  • Must not conflict with the preferred maintenance window

  • Must be at least 30 minutes

", + "ModifyDBInstanceMessage$PreferredMaintenanceWindow":"

The weekly time range (in UTC) during which system maintenance can occur, which might result in an outage. Changing this parameter doesn't result in an outage, except in the following situation, and the change is asynchronously applied as soon as possible. If there are pending actions that cause a reboot, and the maintenance window is changed to include the current time, then changing this parameter will cause a reboot of the DB instance. If moving this window to the current time, there must be at least 30 minutes between the current time and end of the window to ensure pending changes are applied.

Default: Uses existing setting

Format: ddd:hh24:mi-ddd:hh24:mi

Valid Days: Mon | Tue | Wed | Thu | Fri | Sat | Sun

Constraints: Must be at least 30 minutes

", + "ModifyDBInstanceMessage$EngineVersion":"

The version number of the database engine to upgrade to. Changing this parameter results in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

For major version upgrades, if a nondefault DB parameter group is currently in use, a new DB parameter group in the DB parameter group family for the new engine version must be specified. The new DB parameter group can be the default for that DB parameter group family.

", + "ModifyDBInstanceMessage$LicenseModel":"

The license model for the DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", + "ModifyDBInstanceMessage$OptionGroupName":"

Indicates that the DB instance should be associated with the specified option group. Changing this parameter doesn't result in an outage except in the following case and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If the parameter change results in an option group that enables OEM, this change can cause a brief (sub-second) period during which new connections are rejected but existing connections are not interrupted.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "ModifyDBInstanceMessage$NewDBInstanceIdentifier":"

The new DB instance identifier for the DB instance when renaming a DB instance. When you change the DB instance identifier, an instance reboot will occur immediately if you set Apply Immediately to true, or will occur during the next maintenance window if Apply Immediately to false. This value is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • The first character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", + "ModifyDBInstanceMessage$StorageType":"

Specifies the storage type to be associated with the DB instance.

If you specify Provisioned IOPS (io1), you must also include a value for the Iops parameter.

If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon Neptune operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a Read Replica for the instance, and creating a DB snapshot of the instance.

Valid values: standard | gp2 | io1

Default: io1 if the Iops parameter is specified, otherwise standard

", + "ModifyDBInstanceMessage$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "ModifyDBInstanceMessage$TdeCredentialPassword":"

The password for the given ARN from the key store in order to access the device.

", + "ModifyDBInstanceMessage$CACertificateIdentifier":"

Indicates the certificate that needs to be associated with the instance.

", + "ModifyDBInstanceMessage$Domain":"

Not supported.

", + "ModifyDBInstanceMessage$MonitoringRoleArn":"

The ARN for the IAM role that permits Neptune to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", + "ModifyDBInstanceMessage$DomainIAMRoleName":"

Not supported

", + "ModifyDBInstanceMessage$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "ModifyDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • If supplied, must match the name of an existing DBParameterGroup.

", + "ModifyDBSubnetGroupMessage$DBSubnetGroupName":"

The name for the DB subnet group. This value is stored as a lowercase string. You can't modify the default subnet group.

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", + "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription":"

The description for the DB subnet group.

", + "ModifyEventSubscriptionMessage$SubscriptionName":"

The name of the event notification subscription.

", + "ModifyEventSubscriptionMessage$SnsTopicArn":"

The Amazon Resource Name (ARN) of the SNS topic created for event notification. The ARN is created by Amazon SNS when you create a topic and subscribe to it.

", + "ModifyEventSubscriptionMessage$SourceType":"

The type of source that is generating the events. For example, if you want to be notified of events generated by a DB instance, you would set this parameter to db-instance. if this value is not specified, all events are returned.

Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot

", + "OptionGroupMembership$OptionGroupName":"

The name of the option group that the instance belongs to.

", + "OptionGroupMembership$Status":"

The status of the DB instance's option group membership. Valid values are: in-sync, pending-apply, pending-removal, pending-maintenance-apply, pending-maintenance-removal, applying, removing, and failed.

", + "OrderableDBInstanceOption$Engine":"

The engine type of a DB instance.

", + "OrderableDBInstanceOption$EngineVersion":"

The engine version of a DB instance.

", + "OrderableDBInstanceOption$DBInstanceClass":"

The DB instance class for a DB instance.

", + "OrderableDBInstanceOption$LicenseModel":"

The license model for a DB instance.

", + "OrderableDBInstanceOption$StorageType":"

Indicates the storage type for a DB instance.

", + "OrderableDBInstanceOptionsMessage$Marker":"

An optional pagination token provided by a previous OrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "Parameter$ParameterName":"

Specifies the name of the parameter.

", + "Parameter$ParameterValue":"

Specifies the value of the parameter.

", + "Parameter$Description":"

Provides a description of the parameter.

", + "Parameter$Source":"

Indicates the source of the parameter value.

", + "Parameter$ApplyType":"

Specifies the engine specific parameters type.

", + "Parameter$DataType":"

Specifies the valid data type for the parameter.

", + "Parameter$AllowedValues":"

Specifies the valid range of values for the parameter.

", + "Parameter$MinimumEngineVersion":"

The earliest engine version to which the parameter can apply.

", + "PendingMaintenanceAction$Action":"

The type of pending maintenance action that is available for the resource.

", + "PendingMaintenanceAction$OptInStatus":"

Indicates the type of opt-in request that has been received for the resource.

", + "PendingMaintenanceAction$Description":"

A description providing more detail about the maintenance action.

", + "PendingMaintenanceActionsMessage$Marker":"

An optional pagination token provided by a previous DescribePendingMaintenanceActions request. If this parameter is specified, the response includes only records beyond the marker, up to a number of records specified by MaxRecords.

", + "PendingModifiedValues$DBInstanceClass":"

Contains the new DBInstanceClass for the DB instance that will be applied or is currently being applied.

", + "PendingModifiedValues$MasterUserPassword":"

Contains the pending or currently-in-progress change of the master credentials for the DB instance.

", + "PendingModifiedValues$EngineVersion":"

Indicates the database engine version.

", + "PendingModifiedValues$LicenseModel":"

The license model for the DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", + "PendingModifiedValues$DBInstanceIdentifier":"

Contains the new DBInstanceIdentifier for the DB instance that will be applied or is currently being applied.

", + "PendingModifiedValues$StorageType":"

Specifies the storage type to be associated with the DB instance.

", + "PendingModifiedValues$CACertificateIdentifier":"

Specifies the identifier of the CA certificate for the DB instance.

", + "PendingModifiedValues$DBSubnetGroupName":"

The new DB subnet group for the DB instance.

", + "PromoteReadReplicaDBClusterMessage$DBClusterIdentifier":"

The identifier of the DB cluster Read Replica to promote. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster Read Replica.

Example: my-cluster-replica1

", + "ReadReplicaDBClusterIdentifierList$member":null, + "ReadReplicaDBInstanceIdentifierList$member":null, + "ReadReplicaIdentifierList$member":null, + "RebootDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "RemoveRoleFromDBClusterMessage$DBClusterIdentifier":"

The name of the DB cluster to disassociate the IAM role from.

", + "RemoveRoleFromDBClusterMessage$RoleArn":"

The Amazon Resource Name (ARN) of the IAM role to disassociate from the DB cluster, for example arn:aws:iam::123456789012:role/NeptuneAccessRole.

", + "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName":"

The name of the event notification subscription you want to remove a source identifier from.

", + "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier":"

The source identifier to be removed from the subscription, such as the DB instance identifier for a DB instance or the name of a security group.

", + "RemoveTagsFromResourceMessage$ResourceName":"

The Amazon Neptune resource that the tags are removed from. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an Amazon Resource Name (ARN).

", + "ResetDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to reset.

", + "ResetDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • Must match the name of an existing DBParameterGroup.

", + "ResourcePendingMaintenanceActions$ResourceIdentifier":"

The ARN of the resource that has pending maintenance actions.

", + "RestoreDBClusterFromSnapshotMessage$DBClusterIdentifier":"

The name of the DB cluster to create from the DB snapshot or DB cluster snapshot. This parameter isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", + "RestoreDBClusterFromSnapshotMessage$SnapshotIdentifier":"

The identifier for the DB snapshot or DB cluster snapshot to restore from.

You can use either the name or the Amazon Resource Name (ARN) to specify a DB cluster snapshot. However, you can use only the ARN to specify a DB snapshot.

Constraints:

  • Must match the identifier of an existing Snapshot.

", + "RestoreDBClusterFromSnapshotMessage$Engine":"

The database engine to use for the new DB cluster.

Default: The same as source

Constraint: Must be compatible with the engine of the source

", + "RestoreDBClusterFromSnapshotMessage$EngineVersion":"

The version of the database engine to use for the new DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$DBSubnetGroupName":"

The name of the DB subnet group to use for the new DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBClusterFromSnapshotMessage$DatabaseName":"

The database name for the restored DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$OptionGroupName":"

The name of the option group to use for the restored DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$KmsKeyId":"

The AWS KMS key identifier to use when restoring an encrypted DB cluster from a DB snapshot or DB cluster snapshot.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are restoring a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

If you do not specify a value for the KmsKeyId parameter, then the following will occur:

  • If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is encrypted, then the restored DB cluster is encrypted using the KMS key that was used to encrypt the DB snapshot or DB cluster snapshot.

  • If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is not encrypted, then the restored DB cluster is not encrypted.

", + "RestoreDBClusterToPointInTimeMessage$DBClusterIdentifier":"

The name of the new DB cluster to be created.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "RestoreDBClusterToPointInTimeMessage$RestoreType":"

The type of restore to be performed. You can specify one of the following values:

  • full-copy - The new DB cluster is restored as a full copy of the source DB cluster.

  • copy-on-write - The new DB cluster is restored as a clone of the source DB cluster.

Constraints: You can't specify copy-on-write if the engine version of the source DB cluster is earlier than 1.11.

If you don't specify a RestoreType value, then the new DB cluster is restored as a full copy of the source DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$SourceDBClusterIdentifier":"

The identifier of the source DB cluster from which to restore.

Constraints:

  • Must match the identifier of an existing DBCluster.

", + "RestoreDBClusterToPointInTimeMessage$DBSubnetGroupName":"

The DB subnet group name to use for the new DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBClusterToPointInTimeMessage$OptionGroupName":"

The name of the option group for the new DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$KmsKeyId":"

The AWS KMS key identifier to use when restoring an encrypted DB cluster from an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are restoring a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

You can restore to a new DB cluster and encrypt the new DB cluster with a KMS key that is different than the KMS key used to encrypt the source DB cluster. The new DB cluster is encrypted with the KMS key identified by the KmsKeyId parameter.

If you do not specify a value for the KmsKeyId parameter, then the following will occur:

  • If the DB cluster is encrypted, then the restored DB cluster is encrypted using the KMS key that was used to encrypt the source DB cluster.

  • If the DB cluster is not encrypted, then the restored DB cluster is not encrypted.

If DBClusterIdentifier refers to a DB cluster that is not encrypted, then the restore request is rejected.

", + "SourceIdsList$member":null, + "Subnet$SubnetIdentifier":"

Specifies the identifier of the subnet.

", + "Subnet$SubnetStatus":"

Specifies the status of the subnet.

", + "SubnetIdentifierList$member":null, + "Tag$Key":"

A key is the required name of the tag. The string value can be from 1 to 128 Unicode characters in length and can't be prefixed with \"aws:\" or \"rds:\". The string can only contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-' (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").

", + "Tag$Value":"

A value is the optional value of the tag. The string value can be from 1 to 256 Unicode characters in length and can't be prefixed with \"aws:\" or \"rds:\". The string can only contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-' (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").

", + "Timezone$TimezoneName":"

The name of the time zone.

", + "UpgradeTarget$Engine":"

The name of the upgrade target database engine.

", + "UpgradeTarget$EngineVersion":"

The version number of the upgrade target database engine.

", + "UpgradeTarget$Description":"

The version of the database engine that a DB instance can be upgraded to.

", + "ValidStorageOptions$StorageType":"

The valid storage types for your DB instance. For example, gp2, io1.

", + "VpcSecurityGroupIdList$member":null, + "VpcSecurityGroupMembership$VpcSecurityGroupId":"

The name of the VPC security group.

", + "VpcSecurityGroupMembership$Status":"

The status of the VPC security group.

" + } + }, + "Subnet":{ + "base":"

This data type is used as a response element in the DescribeDBSubnetGroups action.

", + "refs":{ + "SubnetList$member":null + } + }, + "SubnetAlreadyInUse":{ + "base":"

The DB subnet is already in use in the Availability Zone.

", + "refs":{ + } + }, + "SubnetIdentifierList":{ + "base":null, + "refs":{ + "CreateDBSubnetGroupMessage$SubnetIds":"

The EC2 Subnet IDs for the DB subnet group.

", + "ModifyDBSubnetGroupMessage$SubnetIds":"

The EC2 subnet IDs for the DB subnet group.

" + } + }, + "SubnetList":{ + "base":null, + "refs":{ + "DBSubnetGroup$Subnets":"

Contains a list of Subnet elements.

" + } + }, + "SubscriptionAlreadyExistFault":{ + "base":null, + "refs":{ + } + }, + "SubscriptionCategoryNotFoundFault":{ + "base":null, + "refs":{ + } + }, + "SubscriptionNotFoundFault":{ + "base":null, + "refs":{ + } + }, + "SupportedCharacterSetsList":{ + "base":null, + "refs":{ + "DBEngineVersion$SupportedCharacterSets":"

A list of the character sets supported by this engine for the CharacterSetName parameter of the CreateDBInstance action.

" + } + }, + "SupportedTimezonesList":{ + "base":null, + "refs":{ + "DBEngineVersion$SupportedTimezones":"

A list of the time zones supported by this engine for the Timezone parameter of the CreateDBInstance action.

" + } + }, + "TStamp":{ + "base":null, + "refs":{ + "DBCluster$EarliestRestorableTime":"

Specifies the earliest time to which a database can be restored with point-in-time restore.

", + "DBCluster$LatestRestorableTime":"

Specifies the latest time to which a database can be restored with point-in-time restore.

", + "DBCluster$ClusterCreateTime":"

Specifies the time when the DB cluster was created, in Universal Coordinated Time (UTC).

", + "DBClusterSnapshot$SnapshotCreateTime":"

Provides the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", + "DBClusterSnapshot$ClusterCreateTime":"

Specifies the time when the DB cluster was created, in Universal Coordinated Time (UTC).

", + "DBInstance$InstanceCreateTime":"

Provides the date and time the DB instance was created.

", + "DBInstance$LatestRestorableTime":"

Specifies the latest time to which a database can be restored with point-in-time restore.

", + "DescribeEventsMessage$StartTime":"

The beginning of the time interval to retrieve events for, specified in ISO 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia page.

Example: 2009-07-08T18:00Z

", + "DescribeEventsMessage$EndTime":"

The end of the time interval for which to retrieve events, specified in ISO 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia page.

Example: 2009-07-08T18:00Z

", + "Event$Date":"

Specifies the date and time of the event.

", + "PendingMaintenanceAction$AutoAppliedAfterDate":"

The date of the maintenance window when the action is applied. The maintenance action is applied to the resource during its first maintenance window after this date. If this date is specified, any next-maintenance opt-in requests are ignored.

", + "PendingMaintenanceAction$ForcedApplyDate":"

The date when the maintenance action is automatically applied. The maintenance action is applied to the resource on this date regardless of the maintenance window for the resource. If this date is specified, any immediate opt-in requests are ignored.

", + "PendingMaintenanceAction$CurrentApplyDate":"

The effective date when the pending maintenance action is applied to the resource. This date takes into account opt-in requests received from the ApplyPendingMaintenanceAction API, the AutoAppliedAfterDate, and the ForcedApplyDate. This value is blank if an opt-in request has not been received and nothing has been specified as AutoAppliedAfterDate or ForcedApplyDate.

", + "RestoreDBClusterToPointInTimeMessage$RestoreToTime":"

The date and time to restore the DB cluster to.

Valid Values: Value must be a time in Universal Coordinated Time (UTC) format

Constraints:

  • Must be before the latest restorable time for the DB instance

  • Must be specified if UseLatestRestorableTime parameter is not provided

  • Cannot be specified if UseLatestRestorableTime parameter is true

  • Cannot be specified if RestoreType parameter is copy-on-write

Example: 2015-03-07T23:45:00Z

" + } + }, + "Tag":{ + "base":"

Metadata assigned to an Amazon Neptune resource consisting of a key-value pair.

", + "refs":{ + "TagList$member":null + } + }, + "TagList":{ + "base":"

A list of tags. For more information, see Tagging Amazon Neptune Resources.

", + "refs":{ + "AddTagsToResourceMessage$Tags":"

The tags to be assigned to the Amazon Neptune resource.

", + "CopyDBClusterParameterGroupMessage$Tags":null, + "CopyDBClusterSnapshotMessage$Tags":null, + "CopyDBParameterGroupMessage$Tags":null, + "CreateDBClusterMessage$Tags":null, + "CreateDBClusterParameterGroupMessage$Tags":null, + "CreateDBClusterSnapshotMessage$Tags":"

The tags to be assigned to the DB cluster snapshot.

", + "CreateDBInstanceMessage$Tags":null, + "CreateDBParameterGroupMessage$Tags":null, + "CreateDBSubnetGroupMessage$Tags":null, + "CreateEventSubscriptionMessage$Tags":null, + "RestoreDBClusterFromSnapshotMessage$Tags":"

The tags to be assigned to the restored DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$Tags":null, + "TagListMessage$TagList":"

List of tags returned by the ListTagsForResource operation.

" + } + }, + "TagListMessage":{ + "base":"

", + "refs":{ + } + }, + "Timezone":{ + "base":"

A time zone associated with a DBInstance. This data type is an element in the response to the DescribeDBInstances, and the DescribeDBEngineVersions actions.

", + "refs":{ + "SupportedTimezonesList$member":null + } + }, + "UpgradeTarget":{ + "base":"

The version of the database engine that a DB instance can be upgraded to.

", + "refs":{ + "ValidUpgradeTargetList$member":null + } + }, + "ValidDBInstanceModificationsMessage":{ + "base":"

Information about valid modifications that you can make to your DB instance. Contains the result of a successful call to the DescribeValidDBInstanceModifications action. You can use this information when you call ModifyDBInstance.

", + "refs":{ + "DescribeValidDBInstanceModificationsResult$ValidDBInstanceModificationsMessage":null + } + }, + "ValidStorageOptions":{ + "base":"

Information about valid modifications that you can make to your DB instance. Contains the result of a successful call to the DescribeValidDBInstanceModifications action.

", + "refs":{ + "ValidStorageOptionsList$member":null + } + }, + "ValidStorageOptionsList":{ + "base":null, + "refs":{ + "ValidDBInstanceModificationsMessage$Storage":"

Valid storage options for your DB instance.

" + } + }, + "ValidUpgradeTargetList":{ + "base":null, + "refs":{ + "DBEngineVersion$ValidUpgradeTarget":"

A list of engine versions that this database engine version can be upgraded to.

" + } + }, + "VpcSecurityGroupIdList":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to associate with this DB cluster.

", + "CreateDBInstanceMessage$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to associate with this DB instance.

Not applicable. The associated list of EC2 VPC security groups is managed by the DB cluster. For more information, see CreateDBCluster.

Default: The default EC2 VPC security group for the DB subnet group's VPC.

", + "ModifyDBClusterMessage$VpcSecurityGroupIds":"

A list of VPC security groups that the DB cluster will belong to.

", + "ModifyDBInstanceMessage$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to authorize on this DB instance. This change is asynchronously applied as soon as possible.

Not applicable. The associated list of EC2 VPC security groups is managed by the DB cluster. For more information, see ModifyDBCluster.

Constraints:

  • If supplied, must match existing VpcSecurityGroupIds.

", + "RestoreDBClusterFromSnapshotMessage$VpcSecurityGroupIds":"

A list of VPC security groups that the new DB cluster will belong to.

", + "RestoreDBClusterToPointInTimeMessage$VpcSecurityGroupIds":"

A list of VPC security groups that the new DB cluster belongs to.

" + } + }, + "VpcSecurityGroupMembership":{ + "base":"

This data type is used as a response element for queries on VPC security group membership.

", + "refs":{ + "VpcSecurityGroupMembershipList$member":null + } + }, + "VpcSecurityGroupMembershipList":{ + "base":null, + "refs":{ + "DBCluster$VpcSecurityGroups":"

Provides a list of VPC security groups that the DB cluster belongs to.

", + "DBInstance$VpcSecurityGroups":"

Provides a list of VPC security group elements that the DB instance belongs to.

" + } + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/examples-1.json new file mode 100644 index 000000000..faff76894 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/examples-1.json @@ -0,0 +1,5 @@ +{ + "version":"1.0", + "examples":{ + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/paginators-1.json similarity index 61% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/paginators-1.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/paginators-1.json index 2461b481b..2a4588640 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/paginators-1.json @@ -24,18 +24,6 @@ "output_token": "Marker", "result_key": "Parameters" }, - "DescribeDBSecurityGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSecurityGroups" - }, - "DescribeDBSnapshots": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSnapshots" - }, "DescribeDBSubnetGroups": { "input_token": "Marker", "limit_key": "MaxRecords", @@ -60,36 +48,12 @@ "output_token": "Marker", "result_key": "Events" }, - "DescribeOptionGroupOptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OptionGroupOptions" - }, - "DescribeOptionGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OptionGroupsList" - }, "DescribeOrderableDBInstanceOptions": { "input_token": "Marker", "limit_key": "MaxRecords", "output_token": "Marker", "result_key": "OrderableDBInstanceOptions" }, - "DescribeReservedDBInstances": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "ReservedDBInstances" - }, - "DescribeReservedDBInstancesOfferings": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "ReservedDBInstancesOfferings" - }, "ListTagsForResource": { "result_key": "TagList" } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/waiters-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/waiters-2.json similarity index 86% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/waiters-2.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/waiters-2.json index b01500797..e75f03b2a 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/waiters-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/neptune/2014-10-31/waiters-2.json @@ -41,18 +41,6 @@ "matcher": "pathAny", "state": "failure", "argument": "DBInstances[].DBInstanceStatus" - }, - { - "expected": "incompatible-parameters", - "matcher": "pathAny", - "state": "failure", - "argument": "DBInstances[].DBInstanceStatus" - }, - { - "expected": "incompatible-restore", - "matcher": "pathAny", - "state": "failure", - "argument": "DBInstances[].DBInstanceStatus" } ] }, @@ -67,6 +55,11 @@ "state": "success", "argument": "DBInstances[].DBInstanceStatus" }, + { + "expected": "DBInstanceNotFound", + "matcher": "error", + "state": "success" + }, { "expected": "creating", "matcher": "pathAny", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/api-2.json new file mode 100644 index 000000000..8188b6ac3 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/api-2.json @@ -0,0 +1,253 @@ +{ + "version":"2.0", + "metadata":{ + "apiVersion":"2018-02-27", + "endpointPrefix":"pi", + "jsonVersion":"1.1", + "protocol":"json", + "serviceAbbreviation":"AWS PI", + "serviceFullName":"AWS Performance Insights", + "serviceId":"PI", + "signatureVersion":"v4", + "signingName":"pi", + "targetPrefix":"PerformanceInsightsv20180227", + "uid":"pi-2018-02-27" + }, + "operations":{ + "DescribeDimensionKeys":{ + "name":"DescribeDimensionKeys", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDimensionKeysRequest"}, + "output":{"shape":"DescribeDimensionKeysResponse"}, + "errors":[ + {"shape":"InvalidArgumentException"}, + {"shape":"InternalServiceError"}, + {"shape":"NotAuthorizedException"} + ] + }, + "GetResourceMetrics":{ + "name":"GetResourceMetrics", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"GetResourceMetricsRequest"}, + "output":{"shape":"GetResourceMetricsResponse"}, + "errors":[ + {"shape":"InvalidArgumentException"}, + {"shape":"InternalServiceError"}, + {"shape":"NotAuthorizedException"} + ] + } + }, + "shapes":{ + "DataPoint":{ + "type":"structure", + "required":[ + "Timestamp", + "Value" + ], + "members":{ + "Timestamp":{"shape":"ISOTimestamp"}, + "Value":{"shape":"Double"} + } + }, + "DataPointsList":{ + "type":"list", + "member":{"shape":"DataPoint"} + }, + "DescribeDimensionKeysRequest":{ + "type":"structure", + "required":[ + "ServiceType", + "Identifier", + "StartTime", + "EndTime", + "Metric", + "GroupBy" + ], + "members":{ + "ServiceType":{"shape":"ServiceType"}, + "Identifier":{"shape":"String"}, + "StartTime":{"shape":"ISOTimestamp"}, + "EndTime":{"shape":"ISOTimestamp"}, + "Metric":{"shape":"String"}, + "PeriodInSeconds":{"shape":"Integer"}, + "GroupBy":{"shape":"DimensionGroup"}, + "PartitionBy":{"shape":"DimensionGroup"}, + "Filter":{"shape":"MetricQueryFilterMap"}, + "MaxResults":{"shape":"MaxResults"}, + "NextToken":{"shape":"String"} + } + }, + "DescribeDimensionKeysResponse":{ + "type":"structure", + "members":{ + "AlignedStartTime":{"shape":"ISOTimestamp"}, + "AlignedEndTime":{"shape":"ISOTimestamp"}, + "PartitionKeys":{"shape":"ResponsePartitionKeyList"}, + "Keys":{"shape":"DimensionKeyDescriptionList"}, + "NextToken":{"shape":"String"} + } + }, + "DimensionGroup":{ + "type":"structure", + "required":["Group"], + "members":{ + "Group":{"shape":"String"}, + "Dimensions":{"shape":"StringList"}, + "Limit":{"shape":"Limit"} + } + }, + "DimensionKeyDescription":{ + "type":"structure", + "members":{ + "Dimensions":{"shape":"DimensionMap"}, + "Total":{"shape":"Double"}, + "Partitions":{"shape":"MetricValuesList"} + } + }, + "DimensionKeyDescriptionList":{ + "type":"list", + "member":{"shape":"DimensionKeyDescription"} + }, + "DimensionMap":{ + "type":"map", + "key":{"shape":"String"}, + "value":{"shape":"String"} + }, + "Double":{"type":"double"}, + "GetResourceMetricsRequest":{ + "type":"structure", + "required":[ + "ServiceType", + "Identifier", + "MetricQueries", + "StartTime", + "EndTime" + ], + "members":{ + "ServiceType":{"shape":"ServiceType"}, + "Identifier":{"shape":"String"}, + "MetricQueries":{"shape":"MetricQueryList"}, + "StartTime":{"shape":"ISOTimestamp"}, + "EndTime":{"shape":"ISOTimestamp"}, + "PeriodInSeconds":{"shape":"Integer"}, + "MaxResults":{"shape":"MaxResults"}, + "NextToken":{"shape":"String"} + } + }, + "GetResourceMetricsResponse":{ + "type":"structure", + "members":{ + "AlignedStartTime":{"shape":"ISOTimestamp"}, + "AlignedEndTime":{"shape":"ISOTimestamp"}, + "Identifier":{"shape":"String"}, + "MetricList":{"shape":"MetricKeyDataPointsList"}, + "NextToken":{"shape":"String"} + } + }, + "ISOTimestamp":{"type":"timestamp"}, + "Integer":{"type":"integer"}, + "InternalServiceError":{ + "type":"structure", + "members":{ + "Message":{"shape":"String"} + }, + "exception":true, + "fault":true + }, + "InvalidArgumentException":{ + "type":"structure", + "members":{ + "Message":{"shape":"String"} + }, + "exception":true + }, + "Limit":{ + "type":"integer", + "max":10, + "min":1 + }, + "MaxResults":{ + "type":"integer", + "max":20, + "min":0 + }, + "MetricKeyDataPoints":{ + "type":"structure", + "members":{ + "Key":{"shape":"ResponseResourceMetricKey"}, + "DataPoints":{"shape":"DataPointsList"} + } + }, + "MetricKeyDataPointsList":{ + "type":"list", + "member":{"shape":"MetricKeyDataPoints"} + }, + "MetricQuery":{ + "type":"structure", + "required":["Metric"], + "members":{ + "Metric":{"shape":"String"}, + "GroupBy":{"shape":"DimensionGroup"}, + "Filter":{"shape":"MetricQueryFilterMap"} + } + }, + "MetricQueryFilterMap":{ + "type":"map", + "key":{"shape":"String"}, + "value":{"shape":"String"} + }, + "MetricQueryList":{ + "type":"list", + "member":{"shape":"MetricQuery"}, + "max":15, + "min":1 + }, + "MetricValuesList":{ + "type":"list", + "member":{"shape":"Double"} + }, + "NotAuthorizedException":{ + "type":"structure", + "members":{ + "Message":{"shape":"String"} + }, + "exception":true + }, + "ResponsePartitionKey":{ + "type":"structure", + "required":["Dimensions"], + "members":{ + "Dimensions":{"shape":"DimensionMap"} + } + }, + "ResponsePartitionKeyList":{ + "type":"list", + "member":{"shape":"ResponsePartitionKey"} + }, + "ResponseResourceMetricKey":{ + "type":"structure", + "required":["Metric"], + "members":{ + "Metric":{"shape":"String"}, + "Dimensions":{"shape":"DimensionMap"} + } + }, + "ServiceType":{ + "type":"string", + "enum":["RDS"] + }, + "String":{"type":"string"}, + "StringList":{ + "type":"list", + "member":{"shape":"String"}, + "max":10, + "min":1 + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/docs-2.json new file mode 100644 index 000000000..a1d3ef405 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/docs-2.json @@ -0,0 +1,219 @@ +{ + "version": "2.0", + "service": "

AWS Performance Insights enables you to monitor and explore different dimensions of database load based on data captured from a running RDS instance. The guide provides detailed information about Performance Insights data types, parameters and errors. For more information about Performance Insights capabilities see Using Amazon RDS Performance Insights in the Amazon RDS User Guide.

The AWS Performance Insights API provides visibility into the performance of your RDS instance, when Performance Insights is enabled for supported engine types. While Amazon CloudWatch provides the authoritative source for AWS service vended monitoring metrics, AWS Performance Insights offers a domain-specific view of database load measured as Average Active Sessions and provided to API consumers as a 2-dimensional time-series dataset. The time dimension of the data provides DB load data for each time point in the queried time range, and each time point decomposes overall load in relation to the requested dimensions, such as SQL, Wait-event, User or Host, measured at that time point.

", + "operations": { + "DescribeDimensionKeys": "

For a specific time period, retrieve the top N dimension keys for a metric.

", + "GetResourceMetrics": "

Retrieve Performance Insights metrics for a set of data sources, over a time period. You can provide specific dimension groups and dimensions, and provide aggregation and filtering criteria for each group.

" + }, + "shapes": { + "DataPoint": { + "base": "

A timestamp, and a single numerical value, which together represent a measurement at a particular point in time.

", + "refs": { + "DataPointsList$member": null + } + }, + "DataPointsList": { + "base": null, + "refs": { + "MetricKeyDataPoints$DataPoints": "

An array of timestamp-value pairs, representing measurements over a period of time.

" + } + }, + "DescribeDimensionKeysRequest": { + "base": null, + "refs": { + } + }, + "DescribeDimensionKeysResponse": { + "base": null, + "refs": { + } + }, + "DimensionGroup": { + "base": "

A logical grouping of Performance Insights metrics for a related subject area. For example, the db.sql dimension group consists of the following dimensions: db.sql.id, db.sql.db_id, db.sql.statement, and db.sql.tokenized_id.

", + "refs": { + "DescribeDimensionKeysRequest$GroupBy": "

A specification for how to aggregate the data points from a query result. You must specify a valid dimension group. Performance Insights will return all of the dimensions within that group, unless you provide the names of specific dimensions within that group. You can also request that Performance Insights return a limited number of values for a dimension.

", + "DescribeDimensionKeysRequest$PartitionBy": "

For each dimension specified in GroupBy, specify a secondary dimension to further subdivide the partition keys in the response.

", + "MetricQuery$GroupBy": "

A specification for how to aggregate the data points from a query result. You must specify a valid dimension group. Performance Insights will return all of the dimensions within that group, unless you provide the names of specific dimensions within that group. You can also request that Performance Insights return a limited number of values for a dimension.

" + } + }, + "DimensionKeyDescription": { + "base": "

An array of descriptions and aggregated values for each dimension within a dimension group.

", + "refs": { + "DimensionKeyDescriptionList$member": null + } + }, + "DimensionKeyDescriptionList": { + "base": null, + "refs": { + "DescribeDimensionKeysResponse$Keys": "

The dimension keys that were requested.

" + } + }, + "DimensionMap": { + "base": null, + "refs": { + "DimensionKeyDescription$Dimensions": "

A map of name-value pairs for the dimensions in the group.

", + "ResponsePartitionKey$Dimensions": "

A dimension map that contains the dimension(s) for this partition.

", + "ResponseResourceMetricKey$Dimensions": "

The valid dimensions for the metric.

" + } + }, + "Double": { + "base": null, + "refs": { + "DataPoint$Value": "

The actual value associated with a particular Timestamp.

", + "DimensionKeyDescription$Total": "

The aggregated metric value for the dimension(s), over the requested time range.

", + "MetricValuesList$member": null + } + }, + "GetResourceMetricsRequest": { + "base": null, + "refs": { + } + }, + "GetResourceMetricsResponse": { + "base": null, + "refs": { + } + }, + "ISOTimestamp": { + "base": null, + "refs": { + "DataPoint$Timestamp": "

The time, in epoch format, associated with a particular Value.

", + "DescribeDimensionKeysRequest$StartTime": "

The date and time specifying the beginning of the requested time series data. You can't specify a StartTime that's earlier than 7 days ago. The value specified is inclusive - data points equal to or greater than StartTime will be returned.

The value for StartTime must be earlier than the value for EndTime.

", + "DescribeDimensionKeysRequest$EndTime": "

The date and time specifying the end of the requested time series data. The value specified is exclusive - data points less than (but not equal to) EndTime will be returned.

The value for EndTime must be later than the value for StartTime.

", + "DescribeDimensionKeysResponse$AlignedStartTime": "

The start time for the returned dimension keys, after alignment to a granular boundary (as specified by PeriodInSeconds). AlignedStartTime will be less than or equal to the value of the user-specified StartTime.

", + "DescribeDimensionKeysResponse$AlignedEndTime": "

The end time for the returned dimension keys, after alignment to a granular boundary (as specified by PeriodInSeconds). AlignedEndTime will be greater than or equal to the value of the user-specified Endtime.

", + "GetResourceMetricsRequest$StartTime": "

The date and time specifying the beginning of the requested time series data. You can't specify a StartTime that's earlier than 7 days ago. The value specified is inclusive - data points equal to or greater than StartTime will be returned.

The value for StartTime must be earlier than the value for EndTime.

", + "GetResourceMetricsRequest$EndTime": "

The date and time specifiying the end of the requested time series data. The value specified is exclusive - data points less than (but not equal to) EndTime will be returned.

The value for EndTime must be later than the value for StartTime.

", + "GetResourceMetricsResponse$AlignedStartTime": "

The start time for the returned metrics, after alignment to a granular boundary (as specified by PeriodInSeconds). AlignedStartTime will be less than or equal to the value of the user-specified StartTime.

", + "GetResourceMetricsResponse$AlignedEndTime": "

The end time for the returned metrics, after alignment to a granular boundary (as specified by PeriodInSeconds). AlignedEndTime will be greater than or equal to the value of the user-specified Endtime.

" + } + }, + "Integer": { + "base": null, + "refs": { + "DescribeDimensionKeysRequest$PeriodInSeconds": "

The granularity, in seconds, of the data points returned from Performance Insights. A period can be as short as one second, or as long as one day (86400 seconds). Valid values are:

  • 1 (one second)

  • 60 (one minute)

  • 300 (five minutes)

  • 3600 (one hour)

  • 86400 (twenty-four hours)

If you don't specify PeriodInSeconds, then Performance Insights will choose a value for you, with a goal of returning roughly 100-200 data points in the response.

", + "GetResourceMetricsRequest$PeriodInSeconds": "

The granularity, in seconds, of the data points returned from Performance Insights. A period can be as short as one second, or as long as one day (86400 seconds). Valid values are:

  • 1 (one second)

  • 60 (one minute)

  • 300 (five minutes)

  • 3600 (one hour)

  • 86400 (twenty-four hours)

If you don't specify PeriodInSeconds, then Performance Insights will choose a value for you, with a goal of returning roughly 100-200 data points in the response.

" + } + }, + "InternalServiceError": { + "base": "

The request failed due to an unknown error.

", + "refs": { + } + }, + "InvalidArgumentException": { + "base": "

One of the arguments provided is invalid for this request.

", + "refs": { + } + }, + "Limit": { + "base": null, + "refs": { + "DimensionGroup$Limit": "

The maximum number of items to fetch for this dimension group.

" + } + }, + "MaxResults": { + "base": null, + "refs": { + "DescribeDimensionKeysRequest$MaxResults": "

The maximum number of items to return in the response. If more items exist than the specified MaxRecords value, a pagination token is included in the response so that the remaining results can be retrieved.

", + "GetResourceMetricsRequest$MaxResults": "

The maximum number of items to return in the response. If more items exist than the specified MaxRecords value, a pagination token is included in the response so that the remaining results can be retrieved.

" + } + }, + "MetricKeyDataPoints": { + "base": "

A time-ordered series of data points, correpsonding to a dimension of a Performance Insights metric.

", + "refs": { + "MetricKeyDataPointsList$member": null + } + }, + "MetricKeyDataPointsList": { + "base": null, + "refs": { + "GetResourceMetricsResponse$MetricList": "

An array of metric results,, where each array element contains all of the data points for a particular dimension.

" + } + }, + "MetricQuery": { + "base": "

A single query to be processed. You must provide the metric to query. If no other parameters are specified, Performance Insights returns all of the data points for that metric. You can optionally request that the data points be aggregated by dimension group ( GroupBy), and return only those data points that match your criteria (Filter).

", + "refs": { + "MetricQueryList$member": null + } + }, + "MetricQueryFilterMap": { + "base": null, + "refs": { + "DescribeDimensionKeysRequest$Filter": "

One or more filters to apply in the request. Restrictions:

  • Any number of filters by the same dimension, as specified in the GroupBy or Partition parameters.

  • A single filter for any other dimension in this dimension group.

", + "MetricQuery$Filter": "

One or more filters to apply in the request. Restrictions:

  • Any number of filters by the same dimension, as specified in the GroupBy parameter.

  • A single filter for any other dimension in this dimension group.

" + } + }, + "MetricQueryList": { + "base": null, + "refs": { + "GetResourceMetricsRequest$MetricQueries": "

An array of one or more queries to perform. Each query must specify a Performance Insights metric, and can optionally specify aggregation and filtering criteria.

" + } + }, + "MetricValuesList": { + "base": null, + "refs": { + "DimensionKeyDescription$Partitions": "

If PartitionBy was specified, PartitionKeys contains the dimensions that were.

" + } + }, + "NotAuthorizedException": { + "base": "

The user is not authorized to perform this request.

", + "refs": { + } + }, + "ResponsePartitionKey": { + "base": "

If PartitionBy was specified in a DescribeDimensionKeys request, the dimensions are returned in an array. Each element in the array specifies one dimension.

", + "refs": { + "ResponsePartitionKeyList$member": null + } + }, + "ResponsePartitionKeyList": { + "base": null, + "refs": { + "DescribeDimensionKeysResponse$PartitionKeys": "

If PartitionBy was present in the request, PartitionKeys contains the breakdown of dimension keys by the specified partitions.

" + } + }, + "ResponseResourceMetricKey": { + "base": "

An object describing a Performance Insights metric and one or more dimensions for that metric.

", + "refs": { + "MetricKeyDataPoints$Key": "

The dimension(s) to which the data points apply.

" + } + }, + "ServiceType": { + "base": null, + "refs": { + "DescribeDimensionKeysRequest$ServiceType": "

The AWS service for which Performance Insights will return metrics. The only valid value for ServiceType is: RDS

", + "GetResourceMetricsRequest$ServiceType": "

The AWS service for which Performance Insights will return metrics. The only valid value for ServiceType is: RDS

" + } + }, + "String": { + "base": null, + "refs": { + "DescribeDimensionKeysRequest$Identifier": "

An immutable, AWS Region-unique identifier for a data source. Performance Insights gathers metrics from this data source.

To use an Amazon RDS instance as a data source, you specify its DbiResourceId value - for example: db-FAIHNTYBKTGAUSUZQYPDS2GW4A

", + "DescribeDimensionKeysRequest$Metric": "

The name of a Performance Insights metric to be measured.

Valid values for Metric are:

  • db.load.avg - a scaled representation of the number of active sessions for the database engine.

  • db.sampledload.avg - the raw number of active sessions for the database engine.

", + "DescribeDimensionKeysRequest$NextToken": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the token, up to the value specified by MaxRecords.

", + "DescribeDimensionKeysResponse$NextToken": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the token, up to the value specified by MaxRecords.

", + "DimensionGroup$Group": "

The name of the dimension group. Valid values are:

  • db.user

  • db.host

  • db.sql

  • db.sql_tokenized

  • db.wait_event

  • db.wait_event_type

", + "DimensionMap$key": null, + "DimensionMap$value": null, + "GetResourceMetricsRequest$Identifier": "

An immutable, AWS Region-unique identifier for a data source. Performance Insights gathers metrics from this data source.

To use an Amazon RDS instance as a data source, you specify its DbiResourceId value - for example: db-FAIHNTYBKTGAUSUZQYPDS2GW4A

", + "GetResourceMetricsRequest$NextToken": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the token, up to the value specified by MaxRecords.

", + "GetResourceMetricsResponse$Identifier": "

An immutable, AWS Region-unique identifier for a data source. Performance Insights gathers metrics from this data source.

To use an Amazon RDS instance as a data source, you specify its DbiResourceId value - for example: db-FAIHNTYBKTGAUSUZQYPDS2GW4A

", + "GetResourceMetricsResponse$NextToken": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the token, up to the value specified by MaxRecords.

", + "InternalServiceError$Message": null, + "InvalidArgumentException$Message": null, + "MetricQuery$Metric": "

The name of a Performance Insights metric to be measured.

Valid values for Metric are:

  • db.load.avg - a scaled representation of the number of active sessions for the database engine.

  • db.sampledload.avg - the raw number of active sessions for the database engine.

", + "MetricQueryFilterMap$key": null, + "MetricQueryFilterMap$value": null, + "NotAuthorizedException$Message": null, + "ResponseResourceMetricKey$Metric": "

The name of a Performance Insights metric to be measured.

Valid values for Metric are:

  • db.load.avg - a scaled representation of the number of active sessions for the database engine.

  • db.sampledload.avg - the raw number of active sessions for the database engine.

", + "StringList$member": null + } + }, + "StringList": { + "base": null, + "refs": { + "DimensionGroup$Dimensions": "

A list of specific dimensions from a dimension group. If this parameter is not present, then it signifies that all of the dimensions in the group were requested, or are present in the response.

Valid values for elements in the Dimensions array are:

  • db.user.id

  • db.user.name

  • db.host.id

  • db.host.name

  • db.sql.id

  • db.sql.db_id

  • db.sql.statement

  • db.sql.tokenized_id

  • db.sql_tokenized.id

  • db.sql_tokenized.db_id

  • db.sql_tokenized.statement

  • db.wait_event.name

  • db.wait_event.type

  • db.wait_event_type.name

" + } + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/examples-1.json similarity index 100% rename from vendor/github.com/aws/aws-sdk-go/models/apis/cloudfront/2015-09-17/examples-1.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/examples-1.json diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/paginators-1.json new file mode 100644 index 000000000..5677bd8e4 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/pi/2018-02-27/paginators-1.json @@ -0,0 +1,4 @@ +{ + "pagination": { + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/api-2.json index 15f501283..ec6fb708d 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/api-2.json @@ -508,6 +508,7 @@ "Penelope", "Chantal", "Celine", + "Lea", "Mathieu", "Dora", "Karl", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/docs-2.json index a8462652c..f0846e0b7 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/docs-2.json @@ -301,7 +301,7 @@ } }, "TextLengthExceededException": { - "base": "

The value of the \"Text\" parameter is longer than the accepted limits. The limit for input text is a maximum of 3000 characters total, of which no more than 1500 can be billed characters. SSML tags are not counted as billed characters.

", + "base": "

The value of the \"Text\" parameter is longer than the accepted limits. The limit for input text is a maximum of 6000 characters total, of which no more than 3000 can be billed characters. SSML tags are not counted as billed characters.

", "refs": { } }, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/smoke.json b/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/smoke.json new file mode 100644 index 000000000..13bd72715 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/polly/2016-06-10/smoke.json @@ -0,0 +1,11 @@ +{ + "version": 1, + "defaultRegion": "us-west-2", + "testCases": [ + { + "operationName": "DescribeVoices", + "input": {}, + "errorExpectedFromService": false + } + ] +} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/api-2.json deleted file mode 100644 index f7b1614c7..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/api-2.json +++ /dev/null @@ -1,2903 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2013-01-10", - "endpointPrefix":"rds", - "protocol":"query", - "serviceAbbreviation":"Amazon RDS", - "serviceFullName":"Amazon Relational Database Service", - "serviceId":"RDS", - "signatureVersion":"v4", - "uid":"rds-2013-01-10", - "xmlNamespace":"http://rds.amazonaws.com/doc/2013-01-10/" - }, - "operations":{ - "AddSourceIdentifierToSubscription":{ - "name":"AddSourceIdentifierToSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddSourceIdentifierToSubscriptionMessage"}, - "output":{ - "shape":"AddSourceIdentifierToSubscriptionResult", - "resultWrapper":"AddSourceIdentifierToSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "AddTagsToResource":{ - "name":"AddTagsToResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddTagsToResourceMessage"}, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "AuthorizeDBSecurityGroupIngress":{ - "name":"AuthorizeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"AuthorizeDBSecurityGroupIngressResult", - "resultWrapper":"AuthorizeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"AuthorizationAlreadyExistsFault"}, - {"shape":"AuthorizationQuotaExceededFault"} - ] - }, - "CopyDBSnapshot":{ - "name":"CopyDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyDBSnapshotMessage"}, - "output":{ - "shape":"CopyDBSnapshotResult", - "resultWrapper":"CopyDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "CreateDBInstance":{ - "name":"CreateDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceMessage"}, - "output":{ - "shape":"CreateDBInstanceResult", - "resultWrapper":"CreateDBInstanceResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "CreateDBInstanceReadReplica":{ - "name":"CreateDBInstanceReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceReadReplicaMessage"}, - "output":{ - "shape":"CreateDBInstanceReadReplicaResult", - "resultWrapper":"CreateDBInstanceReadReplicaResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "CreateDBParameterGroup":{ - "name":"CreateDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBParameterGroupMessage"}, - "output":{ - "shape":"CreateDBParameterGroupResult", - "resultWrapper":"CreateDBParameterGroupResult" - }, - "errors":[ - {"shape":"DBParameterGroupQuotaExceededFault"}, - {"shape":"DBParameterGroupAlreadyExistsFault"} - ] - }, - "CreateDBSecurityGroup":{ - "name":"CreateDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSecurityGroupMessage"}, - "output":{ - "shape":"CreateDBSecurityGroupResult", - "resultWrapper":"CreateDBSecurityGroupResult" - }, - "errors":[ - {"shape":"DBSecurityGroupAlreadyExistsFault"}, - {"shape":"DBSecurityGroupQuotaExceededFault"}, - {"shape":"DBSecurityGroupNotSupportedFault"} - ] - }, - "CreateDBSnapshot":{ - "name":"CreateDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSnapshotMessage"}, - "output":{ - "shape":"CreateDBSnapshotResult", - "resultWrapper":"CreateDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "CreateDBSubnetGroup":{ - "name":"CreateDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSubnetGroupMessage"}, - "output":{ - "shape":"CreateDBSubnetGroupResult", - "resultWrapper":"CreateDBSubnetGroupResult" - }, - "errors":[ - {"shape":"DBSubnetGroupAlreadyExistsFault"}, - {"shape":"DBSubnetGroupQuotaExceededFault"}, - {"shape":"DBSubnetQuotaExceededFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"} - ] - }, - "CreateEventSubscription":{ - "name":"CreateEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateEventSubscriptionMessage"}, - "output":{ - "shape":"CreateEventSubscriptionResult", - "resultWrapper":"CreateEventSubscriptionResult" - }, - "errors":[ - {"shape":"EventSubscriptionQuotaExceededFault"}, - {"shape":"SubscriptionAlreadyExistFault"}, - {"shape":"SNSInvalidTopicFault"}, - {"shape":"SNSNoAuthorizationFault"}, - {"shape":"SNSTopicArnNotFoundFault"}, - {"shape":"SubscriptionCategoryNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "CreateOptionGroup":{ - "name":"CreateOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateOptionGroupMessage"}, - "output":{ - "shape":"CreateOptionGroupResult", - "resultWrapper":"CreateOptionGroupResult" - }, - "errors":[ - {"shape":"OptionGroupAlreadyExistsFault"}, - {"shape":"OptionGroupQuotaExceededFault"} - ] - }, - "DeleteDBInstance":{ - "name":"DeleteDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBInstanceMessage"}, - "output":{ - "shape":"DeleteDBInstanceResult", - "resultWrapper":"DeleteDBInstanceResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "DeleteDBParameterGroup":{ - "name":"DeleteDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBParameterGroupMessage"}, - "errors":[ - {"shape":"InvalidDBParameterGroupStateFault"}, - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DeleteDBSecurityGroup":{ - "name":"DeleteDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSecurityGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DeleteDBSnapshot":{ - "name":"DeleteDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSnapshotMessage"}, - "output":{ - "shape":"DeleteDBSnapshotResult", - "resultWrapper":"DeleteDBSnapshotResult" - }, - "errors":[ - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "DeleteDBSubnetGroup":{ - "name":"DeleteDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSubnetGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSubnetGroupStateFault"}, - {"shape":"InvalidDBSubnetStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"} - ] - }, - "DeleteEventSubscription":{ - "name":"DeleteEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteEventSubscriptionMessage"}, - "output":{ - "shape":"DeleteEventSubscriptionResult", - "resultWrapper":"DeleteEventSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"InvalidEventSubscriptionStateFault"} - ] - }, - "DeleteOptionGroup":{ - "name":"DeleteOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteOptionGroupMessage"}, - "errors":[ - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"InvalidOptionGroupStateFault"} - ] - }, - "DescribeDBEngineVersions":{ - "name":"DescribeDBEngineVersions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBEngineVersionsMessage"}, - "output":{ - "shape":"DBEngineVersionMessage", - "resultWrapper":"DescribeDBEngineVersionsResult" - } - }, - "DescribeDBInstances":{ - "name":"DescribeDBInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBInstancesMessage"}, - "output":{ - "shape":"DBInstanceMessage", - "resultWrapper":"DescribeDBInstancesResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "DescribeDBParameterGroups":{ - "name":"DescribeDBParameterGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBParameterGroupsMessage"}, - "output":{ - "shape":"DBParameterGroupsMessage", - "resultWrapper":"DescribeDBParameterGroupsResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DescribeDBParameters":{ - "name":"DescribeDBParameters", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBParametersMessage"}, - "output":{ - "shape":"DBParameterGroupDetails", - "resultWrapper":"DescribeDBParametersResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DescribeDBSecurityGroups":{ - "name":"DescribeDBSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSecurityGroupsMessage"}, - "output":{ - "shape":"DBSecurityGroupMessage", - "resultWrapper":"DescribeDBSecurityGroupsResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DescribeDBSnapshots":{ - "name":"DescribeDBSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSnapshotsMessage"}, - "output":{ - "shape":"DBSnapshotMessage", - "resultWrapper":"DescribeDBSnapshotsResult" - }, - "errors":[ - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "DescribeDBSubnetGroups":{ - "name":"DescribeDBSubnetGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSubnetGroupsMessage"}, - "output":{ - "shape":"DBSubnetGroupMessage", - "resultWrapper":"DescribeDBSubnetGroupsResult" - }, - "errors":[ - {"shape":"DBSubnetGroupNotFoundFault"} - ] - }, - "DescribeEngineDefaultParameters":{ - "name":"DescribeEngineDefaultParameters", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEngineDefaultParametersMessage"}, - "output":{ - "shape":"DescribeEngineDefaultParametersResult", - "resultWrapper":"DescribeEngineDefaultParametersResult" - } - }, - "DescribeEventCategories":{ - "name":"DescribeEventCategories", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventCategoriesMessage"}, - "output":{ - "shape":"EventCategoriesMessage", - "resultWrapper":"DescribeEventCategoriesResult" - } - }, - "DescribeEventSubscriptions":{ - "name":"DescribeEventSubscriptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventSubscriptionsMessage"}, - "output":{ - "shape":"EventSubscriptionsMessage", - "resultWrapper":"DescribeEventSubscriptionsResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"} - ] - }, - "DescribeEvents":{ - "name":"DescribeEvents", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventsMessage"}, - "output":{ - "shape":"EventsMessage", - "resultWrapper":"DescribeEventsResult" - } - }, - "DescribeOptionGroupOptions":{ - "name":"DescribeOptionGroupOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupOptionsMessage"}, - "output":{ - "shape":"OptionGroupOptionsMessage", - "resultWrapper":"DescribeOptionGroupOptionsResult" - } - }, - "DescribeOptionGroups":{ - "name":"DescribeOptionGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupsMessage"}, - "output":{ - "shape":"OptionGroups", - "resultWrapper":"DescribeOptionGroupsResult" - }, - "errors":[ - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "DescribeOrderableDBInstanceOptions":{ - "name":"DescribeOrderableDBInstanceOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOrderableDBInstanceOptionsMessage"}, - "output":{ - "shape":"OrderableDBInstanceOptionsMessage", - "resultWrapper":"DescribeOrderableDBInstanceOptionsResult" - } - }, - "DescribeReservedDBInstances":{ - "name":"DescribeReservedDBInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedDBInstancesMessage"}, - "output":{ - "shape":"ReservedDBInstanceMessage", - "resultWrapper":"DescribeReservedDBInstancesResult" - }, - "errors":[ - {"shape":"ReservedDBInstanceNotFoundFault"} - ] - }, - "DescribeReservedDBInstancesOfferings":{ - "name":"DescribeReservedDBInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedDBInstancesOfferingsMessage"}, - "output":{ - "shape":"ReservedDBInstancesOfferingMessage", - "resultWrapper":"DescribeReservedDBInstancesOfferingsResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListTagsForResourceMessage"}, - "output":{ - "shape":"TagListMessage", - "resultWrapper":"ListTagsForResourceResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "ModifyDBInstance":{ - "name":"ModifyDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBInstanceMessage"}, - "output":{ - "shape":"ModifyDBInstanceResult", - "resultWrapper":"ModifyDBInstanceResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"DBUpgradeDependencyFailureFault"} - ] - }, - "ModifyDBParameterGroup":{ - "name":"ModifyDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBParameterGroupMessage"}, - "output":{ - "shape":"DBParameterGroupNameMessage", - "resultWrapper":"ModifyDBParameterGroupResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"InvalidDBParameterGroupStateFault"} - ] - }, - "ModifyDBSubnetGroup":{ - "name":"ModifyDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBSubnetGroupMessage"}, - "output":{ - "shape":"ModifyDBSubnetGroupResult", - "resultWrapper":"ModifyDBSubnetGroupResult" - }, - "errors":[ - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetQuotaExceededFault"}, - {"shape":"SubnetAlreadyInUse"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"} - ] - }, - "ModifyEventSubscription":{ - "name":"ModifyEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyEventSubscriptionMessage"}, - "output":{ - "shape":"ModifyEventSubscriptionResult", - "resultWrapper":"ModifyEventSubscriptionResult" - }, - "errors":[ - {"shape":"EventSubscriptionQuotaExceededFault"}, - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SNSInvalidTopicFault"}, - {"shape":"SNSNoAuthorizationFault"}, - {"shape":"SNSTopicArnNotFoundFault"}, - {"shape":"SubscriptionCategoryNotFoundFault"} - ] - }, - "ModifyOptionGroup":{ - "name":"ModifyOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyOptionGroupMessage"}, - "output":{ - "shape":"ModifyOptionGroupResult", - "resultWrapper":"ModifyOptionGroupResult" - }, - "errors":[ - {"shape":"InvalidOptionGroupStateFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "PromoteReadReplica":{ - "name":"PromoteReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PromoteReadReplicaMessage"}, - "output":{ - "shape":"PromoteReadReplicaResult", - "resultWrapper":"PromoteReadReplicaResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "PurchaseReservedDBInstancesOffering":{ - "name":"PurchaseReservedDBInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedDBInstancesOfferingMessage"}, - "output":{ - "shape":"PurchaseReservedDBInstancesOfferingResult", - "resultWrapper":"PurchaseReservedDBInstancesOfferingResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"}, - {"shape":"ReservedDBInstanceAlreadyExistsFault"}, - {"shape":"ReservedDBInstanceQuotaExceededFault"} - ] - }, - "RebootDBInstance":{ - "name":"RebootDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootDBInstanceMessage"}, - "output":{ - "shape":"RebootDBInstanceResult", - "resultWrapper":"RebootDBInstanceResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "RemoveSourceIdentifierFromSubscription":{ - "name":"RemoveSourceIdentifierFromSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveSourceIdentifierFromSubscriptionMessage"}, - "output":{ - "shape":"RemoveSourceIdentifierFromSubscriptionResult", - "resultWrapper":"RemoveSourceIdentifierFromSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "RemoveTagsFromResource":{ - "name":"RemoveTagsFromResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveTagsFromResourceMessage"}, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "ResetDBParameterGroup":{ - "name":"ResetDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetDBParameterGroupMessage"}, - "output":{ - "shape":"DBParameterGroupNameMessage", - "resultWrapper":"ResetDBParameterGroupResult" - }, - "errors":[ - {"shape":"InvalidDBParameterGroupStateFault"}, - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "RestoreDBInstanceFromDBSnapshot":{ - "name":"RestoreDBInstanceFromDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreDBInstanceFromDBSnapshotMessage"}, - "output":{ - "shape":"RestoreDBInstanceFromDBSnapshotResult", - "resultWrapper":"RestoreDBInstanceFromDBSnapshotResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "RestoreDBInstanceToPointInTime":{ - "name":"RestoreDBInstanceToPointInTime", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreDBInstanceToPointInTimeMessage"}, - "output":{ - "shape":"RestoreDBInstanceToPointInTimeResult", - "resultWrapper":"RestoreDBInstanceToPointInTimeResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"PointInTimeRestoreNotEnabledFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "RevokeDBSecurityGroupIngress":{ - "name":"RevokeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"RevokeDBSecurityGroupIngressResult", - "resultWrapper":"RevokeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"AuthorizationNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"} - ] - } - }, - "shapes":{ - "AddSourceIdentifierToSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SourceIdentifier" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SourceIdentifier":{"shape":"String"} - } - }, - "AddSourceIdentifierToSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "AddTagsToResourceMessage":{ - "type":"structure", - "required":[ - "ResourceName", - "Tags" - ], - "members":{ - "ResourceName":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "ApplyMethod":{ - "type":"string", - "enum":[ - "immediate", - "pending-reboot" - ] - }, - "AuthorizationAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizationNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "AuthorizationQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "AuthorizeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "ProvisionedIopsCapable":{"shape":"Boolean"} - }, - "wrapper":true - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"AvailabilityZone" - } - }, - "Boolean":{"type":"boolean"}, - "BooleanOptional":{"type":"boolean"}, - "CharacterSet":{ - "type":"structure", - "members":{ - "CharacterSetName":{"shape":"String"}, - "CharacterSetDescription":{"shape":"String"} - } - }, - "CopyDBSnapshotMessage":{ - "type":"structure", - "required":[ - "SourceDBSnapshotIdentifier", - "TargetDBSnapshotIdentifier" - ], - "members":{ - "SourceDBSnapshotIdentifier":{"shape":"String"}, - "TargetDBSnapshotIdentifier":{"shape":"String"} - } - }, - "CopyDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "CreateDBInstanceMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "AllocatedStorage", - "DBInstanceClass", - "Engine", - "MasterUsername", - "MasterUserPassword" - ], - "members":{ - "DBName":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "DBInstanceClass":{"shape":"String"}, - "Engine":{"shape":"String"}, - "MasterUsername":{"shape":"String"}, - "MasterUserPassword":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "DBParameterGroupName":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "CharacterSetName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"} - } - }, - "CreateDBInstanceReadReplicaMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "SourceDBInstanceIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SourceDBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"} - } - }, - "CreateDBInstanceReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "CreateDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "CreateDBParameterGroupMessage":{ - "type":"structure", - "required":[ - "DBParameterGroupName", - "DBParameterGroupFamily", - "Description" - ], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"} - } - }, - "CreateDBParameterGroupResult":{ - "type":"structure", - "members":{ - "DBParameterGroup":{"shape":"DBParameterGroup"} - } - }, - "CreateDBSecurityGroupMessage":{ - "type":"structure", - "required":[ - "DBSecurityGroupName", - "DBSecurityGroupDescription" - ], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"} - } - }, - "CreateDBSecurityGroupResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "CreateDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBSnapshotIdentifier", - "DBInstanceIdentifier" - ], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"} - } - }, - "CreateDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "CreateDBSubnetGroupMessage":{ - "type":"structure", - "required":[ - "DBSubnetGroupName", - "DBSubnetGroupDescription", - "SubnetIds" - ], - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "SubnetIds":{"shape":"SubnetIdentifierList"} - } - }, - "CreateDBSubnetGroupResult":{ - "type":"structure", - "members":{ - "DBSubnetGroup":{"shape":"DBSubnetGroup"} - } - }, - "CreateEventSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SnsTopicArn" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "SourceIds":{"shape":"SourceIdsList"}, - "Enabled":{"shape":"BooleanOptional"} - } - }, - "CreateEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "CreateOptionGroupMessage":{ - "type":"structure", - "required":[ - "OptionGroupName", - "EngineName", - "MajorEngineVersion", - "OptionGroupDescription" - ], - "members":{ - "OptionGroupName":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"} - } - }, - "CreateOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "DBEngineVersion":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "DBEngineDescription":{"shape":"String"}, - "DBEngineVersionDescription":{"shape":"String"}, - "DefaultCharacterSet":{"shape":"CharacterSet"}, - "SupportedCharacterSets":{"shape":"SupportedCharacterSetsList"} - } - }, - "DBEngineVersionList":{ - "type":"list", - "member":{ - "shape":"DBEngineVersion", - "locationName":"DBEngineVersion" - } - }, - "DBEngineVersionMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBEngineVersions":{"shape":"DBEngineVersionList"} - } - }, - "DBInstance":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Engine":{"shape":"String"}, - "DBInstanceStatus":{"shape":"String"}, - "MasterUsername":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Endpoint":{"shape":"Endpoint"}, - "AllocatedStorage":{"shape":"Integer"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "PreferredBackupWindow":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"Integer"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroups":{"shape":"VpcSecurityGroupMembershipList"}, - "DBParameterGroups":{"shape":"DBParameterGroupStatusList"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroup":{"shape":"DBSubnetGroup"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "PendingModifiedValues":{"shape":"PendingModifiedValues"}, - "LatestRestorableTime":{"shape":"TStamp"}, - "MultiAZ":{"shape":"Boolean"}, - "EngineVersion":{"shape":"String"}, - "AutoMinorVersionUpgrade":{"shape":"Boolean"}, - "ReadReplicaSourceDBInstanceIdentifier":{"shape":"String"}, - "ReadReplicaDBInstanceIdentifiers":{"shape":"ReadReplicaDBInstanceIdentifierList"}, - "LicenseModel":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupMembership":{"shape":"OptionGroupMembership"}, - "CharacterSetName":{"shape":"String"}, - "SecondaryAvailabilityZone":{"shape":"String"}, - "PubliclyAccessible":{"shape":"Boolean"} - }, - "wrapper":true - }, - "DBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBInstanceAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBInstanceList":{ - "type":"list", - "member":{ - "shape":"DBInstance", - "locationName":"DBInstance" - } - }, - "DBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBInstances":{"shape":"DBInstanceList"} - } - }, - "DBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroup":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"} - }, - "wrapper":true - }, - "DBParameterGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupDetails":{ - "type":"structure", - "members":{ - "Parameters":{"shape":"ParametersList"}, - "Marker":{"shape":"String"} - } - }, - "DBParameterGroupList":{ - "type":"list", - "member":{ - "shape":"DBParameterGroup", - "locationName":"DBParameterGroup" - } - }, - "DBParameterGroupNameMessage":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"} - } - }, - "DBParameterGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupStatus":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "ParameterApplyStatus":{"shape":"String"} - } - }, - "DBParameterGroupStatusList":{ - "type":"list", - "member":{ - "shape":"DBParameterGroupStatus", - "locationName":"DBParameterGroup" - } - }, - "DBParameterGroupsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBParameterGroups":{"shape":"DBParameterGroupList"} - } - }, - "DBSecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{"shape":"String"}, - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "EC2SecurityGroups":{"shape":"EC2SecurityGroupList"}, - "IPRanges":{"shape":"IPRangeList"} - }, - "wrapper":true - }, - "DBSecurityGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupMembership":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "DBSecurityGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroupMembership", - "locationName":"DBSecurityGroup" - } - }, - "DBSecurityGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroups"} - } - }, - "DBSecurityGroupNameList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DBSecurityGroupName" - } - }, - "DBSecurityGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupNotSupportedFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotSupported", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"QuotaExceeded.DBSecurityGroup", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroups":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroup", - "locationName":"DBSecurityGroup" - } - }, - "DBSnapshot":{ - "type":"structure", - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "SnapshotCreateTime":{"shape":"TStamp"}, - "Engine":{"shape":"String"}, - "AllocatedStorage":{"shape":"Integer"}, - "Status":{"shape":"String"}, - "Port":{"shape":"Integer"}, - "AvailabilityZone":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "MasterUsername":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"} - }, - "wrapper":true - }, - "DBSnapshotAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSnapshotAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSnapshotList":{ - "type":"list", - "member":{ - "shape":"DBSnapshot", - "locationName":"DBSnapshot" - } - }, - "DBSnapshotMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSnapshots":{"shape":"DBSnapshotList"} - } - }, - "DBSnapshotNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSnapshotNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroup":{ - "type":"structure", - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "SubnetGroupStatus":{"shape":"String"}, - "Subnets":{"shape":"SubnetList"} - }, - "wrapper":true - }, - "DBSubnetGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupDoesNotCoverEnoughAZs":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupDoesNotCoverEnoughAZs", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSubnetGroups":{"shape":"DBSubnetGroups"} - } - }, - "DBSubnetGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroups":{ - "type":"list", - "member":{ - "shape":"DBSubnetGroup", - "locationName":"DBSubnetGroup" - } - }, - "DBSubnetQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBUpgradeDependencyFailureFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBUpgradeDependencyFailure", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DeleteDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SkipFinalSnapshot":{"shape":"Boolean"}, - "FinalDBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "DeleteDBParameterGroupMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"} - } - }, - "DeleteDBSecurityGroupMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"} - } - }, - "DeleteDBSnapshotMessage":{ - "type":"structure", - "required":["DBSnapshotIdentifier"], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "DeleteDBSubnetGroupMessage":{ - "type":"structure", - "required":["DBSubnetGroupName"], - "members":{ - "DBSubnetGroupName":{"shape":"String"} - } - }, - "DeleteEventSubscriptionMessage":{ - "type":"structure", - "required":["SubscriptionName"], - "members":{ - "SubscriptionName":{"shape":"String"} - } - }, - "DeleteEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "DeleteOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"} - } - }, - "DescribeDBEngineVersionsMessage":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"}, - "DefaultOnly":{"shape":"Boolean"}, - "ListSupportedCharacterSets":{"shape":"BooleanOptional"} - } - }, - "DescribeDBInstancesMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBParameterGroupsMessage":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBParametersMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Source":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSecurityGroupsMessage":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSnapshotsMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSubnetGroupsMessage":{ - "type":"structure", - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEngineDefaultParametersMessage":{ - "type":"structure", - "required":["DBParameterGroupFamily"], - "members":{ - "DBParameterGroupFamily":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEngineDefaultParametersResult":{ - "type":"structure", - "members":{ - "EngineDefaults":{"shape":"EngineDefaults"} - } - }, - "DescribeEventCategoriesMessage":{ - "type":"structure", - "members":{ - "SourceType":{"shape":"String"} - } - }, - "DescribeEventSubscriptionsMessage":{ - "type":"structure", - "members":{ - "SubscriptionName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEventsMessage":{ - "type":"structure", - "members":{ - "SourceIdentifier":{"shape":"String"}, - "SourceType":{"shape":"SourceType"}, - "StartTime":{"shape":"TStamp"}, - "EndTime":{"shape":"TStamp"}, - "Duration":{"shape":"IntegerOptional"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupOptionsMessage":{ - "type":"structure", - "required":["EngineName"], - "members":{ - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupsMessage":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Marker":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"} - } - }, - "DescribeOrderableDBInstanceOptionsMessage":{ - "type":"structure", - "required":["Engine"], - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "Vpc":{"shape":"BooleanOptional"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesOfferingsMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "Double":{"type":"double"}, - "EC2SecurityGroup":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "EC2SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"EC2SecurityGroup", - "locationName":"EC2SecurityGroup" - } - }, - "Endpoint":{ - "type":"structure", - "members":{ - "Address":{"shape":"String"}, - "Port":{"shape":"Integer"} - } - }, - "EngineDefaults":{ - "type":"structure", - "members":{ - "DBParameterGroupFamily":{"shape":"String"}, - "Marker":{"shape":"String"}, - "Parameters":{"shape":"ParametersList"} - }, - "wrapper":true - }, - "Event":{ - "type":"structure", - "members":{ - "SourceIdentifier":{"shape":"String"}, - "SourceType":{"shape":"SourceType"}, - "Message":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Date":{"shape":"TStamp"} - } - }, - "EventCategoriesList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"EventCategory" - } - }, - "EventCategoriesMap":{ - "type":"structure", - "members":{ - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"} - }, - "wrapper":true - }, - "EventCategoriesMapList":{ - "type":"list", - "member":{ - "shape":"EventCategoriesMap", - "locationName":"EventCategoriesMap" - } - }, - "EventCategoriesMessage":{ - "type":"structure", - "members":{ - "EventCategoriesMapList":{"shape":"EventCategoriesMapList"} - } - }, - "EventList":{ - "type":"list", - "member":{ - "shape":"Event", - "locationName":"Event" - } - }, - "EventSubscription":{ - "type":"structure", - "members":{ - "Id":{"shape":"String"}, - "CustomerAwsId":{"shape":"String"}, - "CustSubscriptionId":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "Status":{"shape":"String"}, - "SubscriptionCreationTime":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "SourceIdsList":{"shape":"SourceIdsList"}, - "EventCategoriesList":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"Boolean"} - }, - "wrapper":true - }, - "EventSubscriptionQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"EventSubscriptionQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "EventSubscriptionsList":{ - "type":"list", - "member":{ - "shape":"EventSubscription", - "locationName":"EventSubscription" - } - }, - "EventSubscriptionsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "EventSubscriptionsList":{"shape":"EventSubscriptionsList"} - } - }, - "EventsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "Events":{"shape":"EventList"} - } - }, - "IPRange":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "CIDRIP":{"shape":"String"} - } - }, - "IPRangeList":{ - "type":"list", - "member":{ - "shape":"IPRange", - "locationName":"IPRange" - } - }, - "InstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InsufficientDBInstanceCapacityFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InsufficientDBInstanceCapacity", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "Integer":{"type":"integer"}, - "IntegerOptional":{"type":"integer"}, - "InvalidDBInstanceStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBInstanceState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBParameterGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBParameterGroupState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSecurityGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSecurityGroupState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSnapshotStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSnapshotState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidEventSubscriptionStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidEventSubscriptionState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidOptionGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidOptionGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidRestoreFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidRestoreFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidSubnet":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidSubnet", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidVPCNetworkStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidVPCNetworkStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "KeyList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ListTagsForResourceMessage":{ - "type":"structure", - "required":["ResourceName"], - "members":{ - "ResourceName":{"shape":"String"} - } - }, - "ModifyDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "DBInstanceClass":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, - "ApplyImmediately":{"shape":"Boolean"}, - "MasterUserPassword":{"shape":"String"}, - "DBParameterGroupName":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "AllowMajorVersionUpgrade":{"shape":"Boolean"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "NewDBInstanceIdentifier":{"shape":"String"} - } - }, - "ModifyDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "ModifyDBParameterGroupMessage":{ - "type":"structure", - "required":[ - "DBParameterGroupName", - "Parameters" - ], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Parameters":{"shape":"ParametersList"} - } - }, - "ModifyDBSubnetGroupMessage":{ - "type":"structure", - "required":[ - "DBSubnetGroupName", - "SubnetIds" - ], - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "SubnetIds":{"shape":"SubnetIdentifierList"} - } - }, - "ModifyDBSubnetGroupResult":{ - "type":"structure", - "members":{ - "DBSubnetGroup":{"shape":"DBSubnetGroup"} - } - }, - "ModifyEventSubscriptionMessage":{ - "type":"structure", - "required":["SubscriptionName"], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"BooleanOptional"} - } - }, - "ModifyEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "ModifyOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionsToInclude":{"shape":"OptionConfigurationList"}, - "OptionsToRemove":{"shape":"OptionNamesList"}, - "ApplyImmediately":{"shape":"Boolean"} - } - }, - "ModifyOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "Option":{ - "type":"structure", - "members":{ - "OptionName":{"shape":"String"}, - "OptionDescription":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupMembershipList"} - } - }, - "OptionConfiguration":{ - "type":"structure", - "required":["OptionName"], - "members":{ - "OptionName":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupIdList"} - } - }, - "OptionConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionConfiguration", - "locationName":"OptionConfiguration" - } - }, - "OptionGroup":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "Options":{"shape":"OptionsList"}, - "AllowsVpcAndNonVpcInstanceMemberships":{"shape":"Boolean"}, - "VpcId":{"shape":"String"} - }, - "wrapper":true - }, - "OptionGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupAlreadyExistsFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroupMembership":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "OptionGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "OptionGroupOption":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "MinimumRequiredMinorEngineVersion":{"shape":"String"}, - "PortRequired":{"shape":"Boolean"}, - "DefaultPort":{"shape":"IntegerOptional"}, - "OptionsDependedOn":{"shape":"OptionsDependedOn"} - } - }, - "OptionGroupOptionsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOption", - "locationName":"OptionGroupOption" - } - }, - "OptionGroupOptionsMessage":{ - "type":"structure", - "members":{ - "OptionGroupOptions":{"shape":"OptionGroupOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroups":{ - "type":"structure", - "members":{ - "OptionGroupsList":{"shape":"OptionGroupsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupsList":{ - "type":"list", - "member":{ - "shape":"OptionGroup", - "locationName":"OptionGroup" - } - }, - "OptionNamesList":{ - "type":"list", - "member":{"shape":"String"} - }, - "OptionsDependedOn":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"OptionName" - } - }, - "OptionsList":{ - "type":"list", - "member":{ - "shape":"Option", - "locationName":"Option" - } - }, - "OrderableDBInstanceOption":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "AvailabilityZones":{"shape":"AvailabilityZoneList"}, - "MultiAZCapable":{"shape":"Boolean"}, - "ReadReplicaCapable":{"shape":"Boolean"}, - "Vpc":{"shape":"Boolean"} - }, - "wrapper":true - }, - "OrderableDBInstanceOptionsList":{ - "type":"list", - "member":{ - "shape":"OrderableDBInstanceOption", - "locationName":"OrderableDBInstanceOption" - } - }, - "OrderableDBInstanceOptionsMessage":{ - "type":"structure", - "members":{ - "OrderableDBInstanceOptions":{"shape":"OrderableDBInstanceOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "Parameter":{ - "type":"structure", - "members":{ - "ParameterName":{"shape":"String"}, - "ParameterValue":{"shape":"String"}, - "Description":{"shape":"String"}, - "Source":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "DataType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"}, - "MinimumEngineVersion":{"shape":"String"}, - "ApplyMethod":{"shape":"ApplyMethod"} - } - }, - "ParametersList":{ - "type":"list", - "member":{ - "shape":"Parameter", - "locationName":"Parameter" - } - }, - "PendingModifiedValues":{ - "type":"structure", - "members":{ - "DBInstanceClass":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "MasterUserPassword":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "DBInstanceIdentifier":{"shape":"String"} - } - }, - "PointInTimeRestoreNotEnabledFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"PointInTimeRestoreNotEnabled", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PromoteReadReplicaMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"} - } - }, - "PromoteReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "ProvisionedIopsNotAvailableInAZFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ProvisionedIopsNotAvailableInAZFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PurchaseReservedDBInstancesOfferingMessage":{ - "type":"structure", - "required":["ReservedDBInstancesOfferingId"], - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "ReservedDBInstanceId":{"shape":"String"}, - "DBInstanceCount":{"shape":"IntegerOptional"} - } - }, - "PurchaseReservedDBInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedDBInstance":{"shape":"ReservedDBInstance"} - } - }, - "ReadReplicaDBInstanceIdentifierList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReadReplicaDBInstanceIdentifier" - } - }, - "RebootDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "ForceFailover":{"shape":"BooleanOptional"} - } - }, - "RebootDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "RecurringChargeAmount":{"shape":"Double"}, - "RecurringChargeFrequency":{"shape":"String"} - }, - "wrapper":true - }, - "RecurringChargeList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"RecurringCharge" - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SourceIdentifier" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SourceIdentifier":{"shape":"String"} - } - }, - "RemoveSourceIdentifierFromSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "RemoveTagsFromResourceMessage":{ - "type":"structure", - "required":[ - "ResourceName", - "TagKeys" - ], - "members":{ - "ResourceName":{"shape":"String"}, - "TagKeys":{"shape":"KeyList"} - } - }, - "ReservedDBInstance":{ - "type":"structure", - "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "StartTime":{"shape":"TStamp"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "DBInstanceCount":{"shape":"Integer"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "State":{"shape":"String"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceAlreadyExists", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstance", - "locationName":"ReservedDBInstance" - } - }, - "ReservedDBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstances":{"shape":"ReservedDBInstanceList"} - } - }, - "ReservedDBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstancesOffering", - "locationName":"ReservedDBInstancesOffering" - } - }, - "ReservedDBInstancesOfferingMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstancesOfferings":{"shape":"ReservedDBInstancesOfferingList"} - } - }, - "ReservedDBInstancesOfferingNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstancesOfferingNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ResetDBParameterGroupMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "ResetAllParameters":{"shape":"Boolean"}, - "Parameters":{"shape":"ParametersList"} - } - }, - "RestoreDBInstanceFromDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "DBSnapshotIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"} - } - }, - "RestoreDBInstanceFromDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RestoreDBInstanceToPointInTimeMessage":{ - "type":"structure", - "required":[ - "SourceDBInstanceIdentifier", - "TargetDBInstanceIdentifier" - ], - "members":{ - "SourceDBInstanceIdentifier":{"shape":"String"}, - "TargetDBInstanceIdentifier":{"shape":"String"}, - "RestoreTime":{"shape":"TStamp"}, - "UseLatestRestorableTime":{"shape":"Boolean"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"} - } - }, - "RestoreDBInstanceToPointInTimeResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RevokeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "RevokeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "SNSInvalidTopicFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSInvalidTopic", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SNSNoAuthorizationFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSNoAuthorization", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SNSTopicArnNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSTopicArnNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SnapshotQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SnapshotQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SourceIdsList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SourceId" - } - }, - "SourceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SourceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SourceType":{ - "type":"string", - "enum":[ - "db-instance", - "db-parameter-group", - "db-security-group", - "db-snapshot" - ] - }, - "StorageQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"StorageQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetIdentifier":{"shape":"String"}, - "SubnetAvailabilityZone":{"shape":"AvailabilityZone"}, - "SubnetStatus":{"shape":"String"} - } - }, - "SubnetAlreadyInUse":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubnetAlreadyInUse", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SubnetIdentifierList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetIdentifier" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"Subnet" - } - }, - "SubscriptionAlreadyExistFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionAlreadyExist", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SubscriptionCategoryNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionCategoryNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SubscriptionNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SupportedCharacterSetsList":{ - "type":"list", - "member":{ - "shape":"CharacterSet", - "locationName":"CharacterSet" - } - }, - "TStamp":{"type":"timestamp"}, - "Tag":{ - "type":"structure", - "members":{ - "Key":{"shape":"String"}, - "Value":{"shape":"String"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagListMessage":{ - "type":"structure", - "members":{ - "TagList":{"shape":"TagList"} - } - }, - "VpcSecurityGroupIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcSecurityGroupId" - } - }, - "VpcSecurityGroupMembership":{ - "type":"structure", - "members":{ - "VpcSecurityGroupId":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "VpcSecurityGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"VpcSecurityGroupMembership", - "locationName":"VpcSecurityGroupMembership" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/docs-2.json deleted file mode 100644 index c3d02ac6f..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/docs-2.json +++ /dev/null @@ -1,1681 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "AddSourceIdentifierToSubscription": null, - "AddTagsToResource": null, - "AuthorizeDBSecurityGroupIngress": null, - "CopyDBSnapshot": null, - "CreateDBInstance": null, - "CreateDBInstanceReadReplica": null, - "CreateDBParameterGroup": null, - "CreateDBSecurityGroup": null, - "CreateDBSnapshot": null, - "CreateDBSubnetGroup": null, - "CreateEventSubscription": null, - "CreateOptionGroup": null, - "DeleteDBInstance": null, - "DeleteDBParameterGroup": null, - "DeleteDBSecurityGroup": null, - "DeleteDBSnapshot": null, - "DeleteDBSubnetGroup": null, - "DeleteEventSubscription": null, - "DeleteOptionGroup": null, - "DescribeDBEngineVersions": null, - "DescribeDBInstances": null, - "DescribeDBParameterGroups": null, - "DescribeDBParameters": null, - "DescribeDBSecurityGroups": null, - "DescribeDBSnapshots": null, - "DescribeDBSubnetGroups": null, - "DescribeEngineDefaultParameters": null, - "DescribeEventCategories": null, - "DescribeEventSubscriptions": null, - "DescribeEvents": null, - "DescribeOptionGroupOptions": null, - "DescribeOptionGroups": null, - "DescribeOrderableDBInstanceOptions": null, - "DescribeReservedDBInstances": null, - "DescribeReservedDBInstancesOfferings": null, - "ListTagsForResource": null, - "ModifyDBInstance": null, - "ModifyDBParameterGroup": null, - "ModifyDBSubnetGroup": null, - "ModifyEventSubscription": null, - "ModifyOptionGroup": null, - "PromoteReadReplica": null, - "PurchaseReservedDBInstancesOffering": null, - "RebootDBInstance": null, - "RemoveSourceIdentifierFromSubscription": null, - "RemoveTagsFromResource": null, - "ResetDBParameterGroup": null, - "RestoreDBInstanceFromDBSnapshot": null, - "RestoreDBInstanceToPointInTime": null, - "RevokeDBSecurityGroupIngress": null - }, - "shapes": { - "AddSourceIdentifierToSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "AddSourceIdentifierToSubscriptionResult": { - "base": null, - "refs": { - } - }, - "AddTagsToResourceMessage": { - "base": null, - "refs": { - } - }, - "ApplyMethod": { - "base": null, - "refs": { - "Parameter$ApplyMethod": null - } - }, - "AuthorizationAlreadyExistsFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group is already authorized for the specified DB security group.

", - "refs": { - } - }, - "AuthorizationNotFoundFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group isn't authorized for the specified DB security group.

RDS also may not be authorized by using IAM to perform necessary actions on your behalf.

", - "refs": { - } - }, - "AuthorizationQuotaExceededFault": { - "base": "

The DB security group authorization quota has been reached.

", - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "AvailabilityZone": { - "base": null, - "refs": { - "AvailabilityZoneList$member": null, - "Subnet$SubnetAvailabilityZone": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "OrderableDBInstanceOption$AvailabilityZones": null - } - }, - "Boolean": { - "base": null, - "refs": { - "AvailabilityZone$ProvisionedIopsCapable": null, - "DBInstance$MultiAZ": null, - "DBInstance$AutoMinorVersionUpgrade": null, - "DBInstance$PubliclyAccessible": null, - "DeleteDBInstanceMessage$SkipFinalSnapshot": null, - "DescribeDBEngineVersionsMessage$DefaultOnly": null, - "EventSubscription$Enabled": null, - "ModifyDBInstanceMessage$ApplyImmediately": null, - "ModifyDBInstanceMessage$AllowMajorVersionUpgrade": null, - "ModifyOptionGroupMessage$ApplyImmediately": null, - "OptionGroup$AllowsVpcAndNonVpcInstanceMemberships": null, - "OptionGroupOption$PortRequired": null, - "OrderableDBInstanceOption$MultiAZCapable": null, - "OrderableDBInstanceOption$ReadReplicaCapable": null, - "OrderableDBInstanceOption$Vpc": null, - "Parameter$IsModifiable": null, - "ReservedDBInstance$MultiAZ": null, - "ReservedDBInstancesOffering$MultiAZ": null, - "ResetDBParameterGroupMessage$ResetAllParameters": null, - "RestoreDBInstanceToPointInTimeMessage$UseLatestRestorableTime": null - } - }, - "BooleanOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$MultiAZ": null, - "CreateDBInstanceMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceMessage$PubliclyAccessible": null, - "CreateDBInstanceReadReplicaMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceReadReplicaMessage$PubliclyAccessible": null, - "CreateEventSubscriptionMessage$Enabled": null, - "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets": null, - "DescribeOrderableDBInstanceOptionsMessage$Vpc": null, - "DescribeReservedDBInstancesMessage$MultiAZ": null, - "DescribeReservedDBInstancesOfferingsMessage$MultiAZ": null, - "ModifyDBInstanceMessage$MultiAZ": null, - "ModifyDBInstanceMessage$AutoMinorVersionUpgrade": null, - "ModifyEventSubscriptionMessage$Enabled": null, - "PendingModifiedValues$MultiAZ": null, - "RebootDBInstanceMessage$ForceFailover": null, - "RestoreDBInstanceFromDBSnapshotMessage$MultiAZ": null, - "RestoreDBInstanceFromDBSnapshotMessage$PubliclyAccessible": null, - "RestoreDBInstanceFromDBSnapshotMessage$AutoMinorVersionUpgrade": null, - "RestoreDBInstanceToPointInTimeMessage$MultiAZ": null, - "RestoreDBInstanceToPointInTimeMessage$PubliclyAccessible": null, - "RestoreDBInstanceToPointInTimeMessage$AutoMinorVersionUpgrade": null - } - }, - "CharacterSet": { - "base": null, - "refs": { - "DBEngineVersion$DefaultCharacterSet": null, - "SupportedCharacterSetsList$member": null - } - }, - "CopyDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CopyDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceResult": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupResult": { - "base": null, - "refs": { - } - }, - "DBEngineVersion": { - "base": null, - "refs": { - "DBEngineVersionList$member": null - } - }, - "DBEngineVersionList": { - "base": null, - "refs": { - "DBEngineVersionMessage$DBEngineVersions": null - } - }, - "DBEngineVersionMessage": { - "base": null, - "refs": { - } - }, - "DBInstance": { - "base": null, - "refs": { - "CreateDBInstanceReadReplicaResult$DBInstance": null, - "CreateDBInstanceResult$DBInstance": null, - "DBInstanceList$member": null, - "DeleteDBInstanceResult$DBInstance": null, - "ModifyDBInstanceResult$DBInstance": null, - "PromoteReadReplicaResult$DBInstance": null, - "RebootDBInstanceResult$DBInstance": null, - "RestoreDBInstanceFromDBSnapshotResult$DBInstance": null, - "RestoreDBInstanceToPointInTimeResult$DBInstance": null - } - }, - "DBInstanceAlreadyExistsFault": { - "base": "

The user already has a DB instance with the given identifier.

", - "refs": { - } - }, - "DBInstanceList": { - "base": null, - "refs": { - "DBInstanceMessage$DBInstances": null - } - }, - "DBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DBInstanceNotFoundFault": { - "base": "

DBInstanceIdentifier doesn't refer to an existing DB instance.

", - "refs": { - } - }, - "DBParameterGroup": { - "base": null, - "refs": { - "CreateDBParameterGroupResult$DBParameterGroup": null, - "DBParameterGroupList$member": null - } - }, - "DBParameterGroupAlreadyExistsFault": { - "base": "

A DB parameter group with the same name exists.

", - "refs": { - } - }, - "DBParameterGroupDetails": { - "base": null, - "refs": { - } - }, - "DBParameterGroupList": { - "base": null, - "refs": { - "DBParameterGroupsMessage$DBParameterGroups": null - } - }, - "DBParameterGroupNameMessage": { - "base": null, - "refs": { - } - }, - "DBParameterGroupNotFoundFault": { - "base": "

DBParameterGroupName doesn't refer to an existing DB parameter group.

", - "refs": { - } - }, - "DBParameterGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB parameter groups.

", - "refs": { - } - }, - "DBParameterGroupStatus": { - "base": null, - "refs": { - "DBParameterGroupStatusList$member": null - } - }, - "DBParameterGroupStatusList": { - "base": null, - "refs": { - "DBInstance$DBParameterGroups": null - } - }, - "DBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroup": { - "base": null, - "refs": { - "AuthorizeDBSecurityGroupIngressResult$DBSecurityGroup": null, - "CreateDBSecurityGroupResult$DBSecurityGroup": null, - "DBSecurityGroups$member": null, - "RevokeDBSecurityGroupIngressResult$DBSecurityGroup": null - } - }, - "DBSecurityGroupAlreadyExistsFault": { - "base": "

A DB security group with the name specified in DBSecurityGroupName already exists.

", - "refs": { - } - }, - "DBSecurityGroupMembership": { - "base": null, - "refs": { - "DBSecurityGroupMembershipList$member": null - } - }, - "DBSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$DBSecurityGroups": null, - "Option$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroupNameList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$DBSecurityGroups": null, - "ModifyDBInstanceMessage$DBSecurityGroups": null, - "OptionConfiguration$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupNotFoundFault": { - "base": "

DBSecurityGroupName doesn't refer to an existing DB security group.

", - "refs": { - } - }, - "DBSecurityGroupNotSupportedFault": { - "base": "

A DB security group isn't allowed for this action.

", - "refs": { - } - }, - "DBSecurityGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB security groups.

", - "refs": { - } - }, - "DBSecurityGroups": { - "base": null, - "refs": { - "DBSecurityGroupMessage$DBSecurityGroups": null - } - }, - "DBSnapshot": { - "base": null, - "refs": { - "CopyDBSnapshotResult$DBSnapshot": null, - "CreateDBSnapshotResult$DBSnapshot": null, - "DBSnapshotList$member": null, - "DeleteDBSnapshotResult$DBSnapshot": null - } - }, - "DBSnapshotAlreadyExistsFault": { - "base": "

DBSnapshotIdentifier is already used by an existing snapshot.

", - "refs": { - } - }, - "DBSnapshotList": { - "base": null, - "refs": { - "DBSnapshotMessage$DBSnapshots": null - } - }, - "DBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DBSnapshotNotFoundFault": { - "base": "

DBSnapshotIdentifier doesn't refer to an existing DB snapshot.

", - "refs": { - } - }, - "DBSubnetGroup": { - "base": null, - "refs": { - "CreateDBSubnetGroupResult$DBSubnetGroup": null, - "DBInstance$DBSubnetGroup": null, - "DBSubnetGroups$member": null, - "ModifyDBSubnetGroupResult$DBSubnetGroup": null - } - }, - "DBSubnetGroupAlreadyExistsFault": { - "base": "

DBSubnetGroupName is already used by an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupDoesNotCoverEnoughAZs": { - "base": "

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", - "refs": { - } - }, - "DBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSubnetGroupNotFoundFault": { - "base": "

DBSubnetGroupName doesn't refer to an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB subnet groups.

", - "refs": { - } - }, - "DBSubnetGroups": { - "base": null, - "refs": { - "DBSubnetGroupMessage$DBSubnetGroups": null - } - }, - "DBSubnetQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of subnets in a DB subnet groups.

", - "refs": { - } - }, - "DBUpgradeDependencyFailureFault": { - "base": "

The DB upgrade failed because a resource the DB depends on can't be modified.

", - "refs": { - } - }, - "DeleteDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBInstanceResult": { - "base": null, - "refs": { - } - }, - "DeleteDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "DeleteDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "DeleteOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBEngineVersionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSecurityGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSnapshotsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSubnetGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersResult": { - "base": null, - "refs": { - } - }, - "DescribeEventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesOfferingsMessage": { - "base": null, - "refs": { - } - }, - "Double": { - "base": null, - "refs": { - "RecurringCharge$RecurringChargeAmount": null, - "ReservedDBInstance$FixedPrice": null, - "ReservedDBInstance$UsagePrice": null, - "ReservedDBInstancesOffering$FixedPrice": null, - "ReservedDBInstancesOffering$UsagePrice": null - } - }, - "EC2SecurityGroup": { - "base": null, - "refs": { - "EC2SecurityGroupList$member": null - } - }, - "EC2SecurityGroupList": { - "base": null, - "refs": { - "DBSecurityGroup$EC2SecurityGroups": null - } - }, - "Endpoint": { - "base": null, - "refs": { - "DBInstance$Endpoint": null - } - }, - "EngineDefaults": { - "base": null, - "refs": { - "DescribeEngineDefaultParametersResult$EngineDefaults": null - } - }, - "Event": { - "base": null, - "refs": { - "EventList$member": null - } - }, - "EventCategoriesList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$EventCategories": null, - "DescribeEventsMessage$EventCategories": null, - "Event$EventCategories": null, - "EventCategoriesMap$EventCategories": null, - "EventSubscription$EventCategoriesList": null, - "ModifyEventSubscriptionMessage$EventCategories": null - } - }, - "EventCategoriesMap": { - "base": null, - "refs": { - "EventCategoriesMapList$member": null - } - }, - "EventCategoriesMapList": { - "base": null, - "refs": { - "EventCategoriesMessage$EventCategoriesMapList": null - } - }, - "EventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "EventList": { - "base": null, - "refs": { - "EventsMessage$Events": null - } - }, - "EventSubscription": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionResult$EventSubscription": null, - "CreateEventSubscriptionResult$EventSubscription": null, - "DeleteEventSubscriptionResult$EventSubscription": null, - "EventSubscriptionsList$member": null, - "ModifyEventSubscriptionResult$EventSubscription": null, - "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription": null - } - }, - "EventSubscriptionQuotaExceededFault": { - "base": "

You have reached the maximum number of event subscriptions.

", - "refs": { - } - }, - "EventSubscriptionsList": { - "base": null, - "refs": { - "EventSubscriptionsMessage$EventSubscriptionsList": null - } - }, - "EventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "EventsMessage": { - "base": null, - "refs": { - } - }, - "IPRange": { - "base": null, - "refs": { - "IPRangeList$member": null - } - }, - "IPRangeList": { - "base": null, - "refs": { - "DBSecurityGroup$IPRanges": null - } - }, - "InstanceQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB instances.

", - "refs": { - } - }, - "InsufficientDBInstanceCapacityFault": { - "base": "

The specified DB instance class isn't available in the specified Availability Zone.

", - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "DBInstance$AllocatedStorage": null, - "DBInstance$BackupRetentionPeriod": null, - "DBSnapshot$AllocatedStorage": null, - "DBSnapshot$Port": null, - "Endpoint$Port": null, - "ReservedDBInstance$Duration": null, - "ReservedDBInstance$DBInstanceCount": null, - "ReservedDBInstancesOffering$Duration": null - } - }, - "IntegerOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$AllocatedStorage": null, - "CreateDBInstanceMessage$BackupRetentionPeriod": null, - "CreateDBInstanceMessage$Port": null, - "CreateDBInstanceMessage$Iops": null, - "CreateDBInstanceReadReplicaMessage$Port": null, - "CreateDBInstanceReadReplicaMessage$Iops": null, - "DBInstance$Iops": null, - "DBSnapshot$Iops": null, - "DescribeDBEngineVersionsMessage$MaxRecords": null, - "DescribeDBInstancesMessage$MaxRecords": null, - "DescribeDBParameterGroupsMessage$MaxRecords": null, - "DescribeDBParametersMessage$MaxRecords": null, - "DescribeDBSecurityGroupsMessage$MaxRecords": null, - "DescribeDBSnapshotsMessage$MaxRecords": null, - "DescribeDBSubnetGroupsMessage$MaxRecords": null, - "DescribeEngineDefaultParametersMessage$MaxRecords": null, - "DescribeEventSubscriptionsMessage$MaxRecords": null, - "DescribeEventsMessage$Duration": null, - "DescribeEventsMessage$MaxRecords": null, - "DescribeOptionGroupOptionsMessage$MaxRecords": null, - "DescribeOptionGroupsMessage$MaxRecords": null, - "DescribeOrderableDBInstanceOptionsMessage$MaxRecords": null, - "DescribeReservedDBInstancesMessage$MaxRecords": null, - "DescribeReservedDBInstancesOfferingsMessage$MaxRecords": null, - "ModifyDBInstanceMessage$AllocatedStorage": null, - "ModifyDBInstanceMessage$BackupRetentionPeriod": null, - "ModifyDBInstanceMessage$Iops": null, - "Option$Port": null, - "OptionConfiguration$Port": null, - "OptionGroupOption$DefaultPort": null, - "PendingModifiedValues$AllocatedStorage": null, - "PendingModifiedValues$Port": null, - "PendingModifiedValues$BackupRetentionPeriod": null, - "PendingModifiedValues$Iops": null, - "PromoteReadReplicaMessage$BackupRetentionPeriod": null, - "PurchaseReservedDBInstancesOfferingMessage$DBInstanceCount": null, - "RestoreDBInstanceFromDBSnapshotMessage$Port": null, - "RestoreDBInstanceFromDBSnapshotMessage$Iops": null, - "RestoreDBInstanceToPointInTimeMessage$Port": null, - "RestoreDBInstanceToPointInTimeMessage$Iops": null - } - }, - "InvalidDBInstanceStateFault": { - "base": "

The specified DB instance isn't in the available state.

", - "refs": { - } - }, - "InvalidDBParameterGroupStateFault": { - "base": "

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you can't delete it when the parameter group is in this state.

", - "refs": { - } - }, - "InvalidDBSecurityGroupStateFault": { - "base": "

The state of the DB security group doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSnapshotStateFault": { - "base": "

The state of the DB snapshot doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSubnetGroupStateFault": { - "base": "

The DB subnet group cannot be deleted because it's in use.

", - "refs": { - } - }, - "InvalidDBSubnetStateFault": { - "base": "

The DB subnet isn't in the available state.

", - "refs": { - } - }, - "InvalidEventSubscriptionStateFault": { - "base": "

This error can occur if someone else is modifying a subscription. You should retry the action.

", - "refs": { - } - }, - "InvalidOptionGroupStateFault": { - "base": "

The option group isn't in the available state.

", - "refs": { - } - }, - "InvalidRestoreFault": { - "base": "

Cannot restore from VPC backup to non-VPC DB instance.

", - "refs": { - } - }, - "InvalidSubnet": { - "base": "

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", - "refs": { - } - }, - "InvalidVPCNetworkStateFault": { - "base": "

The DB subnet group doesn't cover all Availability Zones after it's created because of users' change.

", - "refs": { - } - }, - "KeyList": { - "base": null, - "refs": { - "RemoveTagsFromResourceMessage$TagKeys": null - } - }, - "ListTagsForResourceMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBInstanceResult": { - "base": null, - "refs": { - } - }, - "ModifyDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupResult": { - "base": null, - "refs": { - } - }, - "Option": { - "base": null, - "refs": { - "OptionsList$member": null - } - }, - "OptionConfiguration": { - "base": null, - "refs": { - "OptionConfigurationList$member": null - } - }, - "OptionConfigurationList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToInclude": null - } - }, - "OptionGroup": { - "base": null, - "refs": { - "CreateOptionGroupResult$OptionGroup": null, - "ModifyOptionGroupResult$OptionGroup": null, - "OptionGroupsList$member": null - } - }, - "OptionGroupAlreadyExistsFault": { - "base": "

The option group you are trying to create already exists.

", - "refs": { - } - }, - "OptionGroupMembership": { - "base": null, - "refs": { - "DBInstance$OptionGroupMembership": null - } - }, - "OptionGroupNotFoundFault": { - "base": "

The specified option group could not be found.

", - "refs": { - } - }, - "OptionGroupOption": { - "base": null, - "refs": { - "OptionGroupOptionsList$member": null - } - }, - "OptionGroupOptionsList": { - "base": null, - "refs": { - "OptionGroupOptionsMessage$OptionGroupOptions": null - } - }, - "OptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "OptionGroupQuotaExceededFault": { - "base": "

The quota of 20 option groups was exceeded for this AWS account.

", - "refs": { - } - }, - "OptionGroups": { - "base": null, - "refs": { - } - }, - "OptionGroupsList": { - "base": null, - "refs": { - "OptionGroups$OptionGroupsList": null - } - }, - "OptionNamesList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToRemove": null - } - }, - "OptionsDependedOn": { - "base": null, - "refs": { - "OptionGroupOption$OptionsDependedOn": null - } - }, - "OptionsList": { - "base": null, - "refs": { - "OptionGroup$Options": null - } - }, - "OrderableDBInstanceOption": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsList$member": null - } - }, - "OrderableDBInstanceOptionsList": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions": null - } - }, - "OrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "Parameter": { - "base": null, - "refs": { - "ParametersList$member": null - } - }, - "ParametersList": { - "base": null, - "refs": { - "DBParameterGroupDetails$Parameters": null, - "EngineDefaults$Parameters": null, - "ModifyDBParameterGroupMessage$Parameters": null, - "ResetDBParameterGroupMessage$Parameters": null - } - }, - "PendingModifiedValues": { - "base": null, - "refs": { - "DBInstance$PendingModifiedValues": null - } - }, - "PointInTimeRestoreNotEnabledFault": { - "base": "

SourceDBInstanceIdentifier refers to a DB instance with BackupRetentionPeriod equal to 0.

", - "refs": { - } - }, - "PromoteReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "PromoteReadReplicaResult": { - "base": null, - "refs": { - } - }, - "ProvisionedIopsNotAvailableInAZFault": { - "base": "

Provisioned IOPS not available in the specified Availability Zone.

", - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingResult": { - "base": null, - "refs": { - } - }, - "ReadReplicaDBInstanceIdentifierList": { - "base": null, - "refs": { - "DBInstance$ReadReplicaDBInstanceIdentifiers": null - } - }, - "RebootDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "RebootDBInstanceResult": { - "base": null, - "refs": { - } - }, - "RecurringCharge": { - "base": null, - "refs": { - "RecurringChargeList$member": null - } - }, - "RecurringChargeList": { - "base": null, - "refs": { - "ReservedDBInstance$RecurringCharges": null, - "ReservedDBInstancesOffering$RecurringCharges": null - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "RemoveSourceIdentifierFromSubscriptionResult": { - "base": null, - "refs": { - } - }, - "RemoveTagsFromResourceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstance": { - "base": null, - "refs": { - "PurchaseReservedDBInstancesOfferingResult$ReservedDBInstance": null, - "ReservedDBInstanceList$member": null - } - }, - "ReservedDBInstanceAlreadyExistsFault": { - "base": "

User already has a reservation with the given identifier.

", - "refs": { - } - }, - "ReservedDBInstanceList": { - "base": null, - "refs": { - "ReservedDBInstanceMessage$ReservedDBInstances": null - } - }, - "ReservedDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstanceNotFoundFault": { - "base": "

The specified reserved DB Instance not found.

", - "refs": { - } - }, - "ReservedDBInstanceQuotaExceededFault": { - "base": "

Request would exceed the user's DB Instance quota.

", - "refs": { - } - }, - "ReservedDBInstancesOffering": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingList$member": null - } - }, - "ReservedDBInstancesOfferingList": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferings": null - } - }, - "ReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstancesOfferingNotFoundFault": { - "base": "

Specified offering does not exist.

", - "refs": { - } - }, - "ResetDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeResult": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "SNSInvalidTopicFault": { - "base": "

SNS has responded that there is a problem with the SND topic specified.

", - "refs": { - } - }, - "SNSNoAuthorizationFault": { - "base": "

You do not have permission to publish to the SNS topic ARN.

", - "refs": { - } - }, - "SNSTopicArnNotFoundFault": { - "base": "

The SNS topic ARN does not exist.

", - "refs": { - } - }, - "SnapshotQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB snapshots.

", - "refs": { - } - }, - "SourceIdsList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$SourceIds": null, - "EventSubscription$SourceIdsList": null - } - }, - "SourceNotFoundFault": { - "base": "

The requested source could not be found.

", - "refs": { - } - }, - "SourceType": { - "base": null, - "refs": { - "DescribeEventsMessage$SourceType": null, - "Event$SourceType": null - } - }, - "StorageQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed amount of storage available across all DB instances.

", - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionMessage$SubscriptionName": null, - "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier": null, - "AddTagsToResourceMessage$ResourceName": null, - "AuthorizeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$CIDRIP": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "AvailabilityZone$Name": null, - "CharacterSet$CharacterSetName": null, - "CharacterSet$CharacterSetDescription": null, - "CopyDBSnapshotMessage$SourceDBSnapshotIdentifier": null, - "CopyDBSnapshotMessage$TargetDBSnapshotIdentifier": null, - "CreateDBInstanceMessage$DBName": null, - "CreateDBInstanceMessage$DBInstanceIdentifier": null, - "CreateDBInstanceMessage$DBInstanceClass": null, - "CreateDBInstanceMessage$Engine": null, - "CreateDBInstanceMessage$MasterUsername": null, - "CreateDBInstanceMessage$MasterUserPassword": null, - "CreateDBInstanceMessage$AvailabilityZone": null, - "CreateDBInstanceMessage$DBSubnetGroupName": null, - "CreateDBInstanceMessage$PreferredMaintenanceWindow": null, - "CreateDBInstanceMessage$DBParameterGroupName": null, - "CreateDBInstanceMessage$PreferredBackupWindow": null, - "CreateDBInstanceMessage$EngineVersion": null, - "CreateDBInstanceMessage$LicenseModel": null, - "CreateDBInstanceMessage$OptionGroupName": null, - "CreateDBInstanceMessage$CharacterSetName": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$SourceDBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceClass": null, - "CreateDBInstanceReadReplicaMessage$AvailabilityZone": null, - "CreateDBInstanceReadReplicaMessage$OptionGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupFamily": null, - "CreateDBParameterGroupMessage$Description": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupName": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupDescription": null, - "CreateDBSnapshotMessage$DBSnapshotIdentifier": null, - "CreateDBSnapshotMessage$DBInstanceIdentifier": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupName": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "CreateEventSubscriptionMessage$SubscriptionName": null, - "CreateEventSubscriptionMessage$SnsTopicArn": null, - "CreateEventSubscriptionMessage$SourceType": null, - "CreateOptionGroupMessage$OptionGroupName": null, - "CreateOptionGroupMessage$EngineName": null, - "CreateOptionGroupMessage$MajorEngineVersion": null, - "CreateOptionGroupMessage$OptionGroupDescription": null, - "DBEngineVersion$Engine": null, - "DBEngineVersion$EngineVersion": null, - "DBEngineVersion$DBParameterGroupFamily": null, - "DBEngineVersion$DBEngineDescription": null, - "DBEngineVersion$DBEngineVersionDescription": null, - "DBEngineVersionMessage$Marker": null, - "DBInstance$DBInstanceIdentifier": null, - "DBInstance$DBInstanceClass": null, - "DBInstance$Engine": null, - "DBInstance$DBInstanceStatus": null, - "DBInstance$MasterUsername": null, - "DBInstance$DBName": null, - "DBInstance$PreferredBackupWindow": null, - "DBInstance$AvailabilityZone": null, - "DBInstance$PreferredMaintenanceWindow": null, - "DBInstance$EngineVersion": null, - "DBInstance$ReadReplicaSourceDBInstanceIdentifier": null, - "DBInstance$LicenseModel": null, - "DBInstance$CharacterSetName": null, - "DBInstance$SecondaryAvailabilityZone": null, - "DBInstanceMessage$Marker": null, - "DBParameterGroup$DBParameterGroupName": null, - "DBParameterGroup$DBParameterGroupFamily": null, - "DBParameterGroup$Description": null, - "DBParameterGroupDetails$Marker": null, - "DBParameterGroupNameMessage$DBParameterGroupName": null, - "DBParameterGroupStatus$DBParameterGroupName": null, - "DBParameterGroupStatus$ParameterApplyStatus": null, - "DBParameterGroupsMessage$Marker": null, - "DBSecurityGroup$OwnerId": null, - "DBSecurityGroup$DBSecurityGroupName": null, - "DBSecurityGroup$DBSecurityGroupDescription": null, - "DBSecurityGroup$VpcId": null, - "DBSecurityGroupMembership$DBSecurityGroupName": null, - "DBSecurityGroupMembership$Status": null, - "DBSecurityGroupMessage$Marker": null, - "DBSecurityGroupNameList$member": null, - "DBSnapshot$DBSnapshotIdentifier": null, - "DBSnapshot$DBInstanceIdentifier": null, - "DBSnapshot$Engine": null, - "DBSnapshot$Status": null, - "DBSnapshot$AvailabilityZone": null, - "DBSnapshot$VpcId": null, - "DBSnapshot$MasterUsername": null, - "DBSnapshot$EngineVersion": null, - "DBSnapshot$LicenseModel": null, - "DBSnapshot$SnapshotType": null, - "DBSnapshotMessage$Marker": null, - "DBSubnetGroup$DBSubnetGroupName": null, - "DBSubnetGroup$DBSubnetGroupDescription": null, - "DBSubnetGroup$VpcId": null, - "DBSubnetGroup$SubnetGroupStatus": null, - "DBSubnetGroupMessage$Marker": null, - "DeleteDBInstanceMessage$DBInstanceIdentifier": null, - "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier": null, - "DeleteDBParameterGroupMessage$DBParameterGroupName": null, - "DeleteDBSecurityGroupMessage$DBSecurityGroupName": null, - "DeleteDBSnapshotMessage$DBSnapshotIdentifier": null, - "DeleteDBSubnetGroupMessage$DBSubnetGroupName": null, - "DeleteEventSubscriptionMessage$SubscriptionName": null, - "DeleteOptionGroupMessage$OptionGroupName": null, - "DescribeDBEngineVersionsMessage$Engine": null, - "DescribeDBEngineVersionsMessage$EngineVersion": null, - "DescribeDBEngineVersionsMessage$DBParameterGroupFamily": null, - "DescribeDBEngineVersionsMessage$Marker": null, - "DescribeDBInstancesMessage$DBInstanceIdentifier": null, - "DescribeDBInstancesMessage$Marker": null, - "DescribeDBParameterGroupsMessage$DBParameterGroupName": null, - "DescribeDBParameterGroupsMessage$Marker": null, - "DescribeDBParametersMessage$DBParameterGroupName": null, - "DescribeDBParametersMessage$Source": null, - "DescribeDBParametersMessage$Marker": null, - "DescribeDBSecurityGroupsMessage$DBSecurityGroupName": null, - "DescribeDBSecurityGroupsMessage$Marker": null, - "DescribeDBSnapshotsMessage$DBInstanceIdentifier": null, - "DescribeDBSnapshotsMessage$DBSnapshotIdentifier": null, - "DescribeDBSnapshotsMessage$SnapshotType": null, - "DescribeDBSnapshotsMessage$Marker": null, - "DescribeDBSubnetGroupsMessage$DBSubnetGroupName": null, - "DescribeDBSubnetGroupsMessage$Marker": null, - "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily": null, - "DescribeEngineDefaultParametersMessage$Marker": null, - "DescribeEventCategoriesMessage$SourceType": null, - "DescribeEventSubscriptionsMessage$SubscriptionName": null, - "DescribeEventSubscriptionsMessage$Marker": null, - "DescribeEventsMessage$SourceIdentifier": null, - "DescribeEventsMessage$Marker": null, - "DescribeOptionGroupOptionsMessage$EngineName": null, - "DescribeOptionGroupOptionsMessage$MajorEngineVersion": null, - "DescribeOptionGroupOptionsMessage$Marker": null, - "DescribeOptionGroupsMessage$OptionGroupName": null, - "DescribeOptionGroupsMessage$Marker": null, - "DescribeOptionGroupsMessage$EngineName": null, - "DescribeOptionGroupsMessage$MajorEngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$Engine": null, - "DescribeOrderableDBInstanceOptionsMessage$EngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass": null, - "DescribeOrderableDBInstanceOptionsMessage$LicenseModel": null, - "DescribeOrderableDBInstanceOptionsMessage$Marker": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstanceId": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesMessage$Duration": null, - "DescribeReservedDBInstancesMessage$ProductDescription": null, - "DescribeReservedDBInstancesMessage$OfferingType": null, - "DescribeReservedDBInstancesMessage$Marker": null, - "DescribeReservedDBInstancesOfferingsMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesOfferingsMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesOfferingsMessage$Duration": null, - "DescribeReservedDBInstancesOfferingsMessage$ProductDescription": null, - "DescribeReservedDBInstancesOfferingsMessage$OfferingType": null, - "DescribeReservedDBInstancesOfferingsMessage$Marker": null, - "EC2SecurityGroup$Status": null, - "EC2SecurityGroup$EC2SecurityGroupName": null, - "EC2SecurityGroup$EC2SecurityGroupId": null, - "EC2SecurityGroup$EC2SecurityGroupOwnerId": null, - "Endpoint$Address": null, - "EngineDefaults$DBParameterGroupFamily": null, - "EngineDefaults$Marker": null, - "Event$SourceIdentifier": null, - "Event$Message": null, - "EventCategoriesList$member": null, - "EventCategoriesMap$SourceType": null, - "EventSubscription$Id": null, - "EventSubscription$CustomerAwsId": null, - "EventSubscription$CustSubscriptionId": null, - "EventSubscription$SnsTopicArn": null, - "EventSubscription$Status": null, - "EventSubscription$SubscriptionCreationTime": null, - "EventSubscription$SourceType": null, - "EventSubscriptionsMessage$Marker": null, - "EventsMessage$Marker": null, - "IPRange$Status": null, - "IPRange$CIDRIP": null, - "KeyList$member": null, - "ListTagsForResourceMessage$ResourceName": null, - "ModifyDBInstanceMessage$DBInstanceIdentifier": null, - "ModifyDBInstanceMessage$DBInstanceClass": null, - "ModifyDBInstanceMessage$MasterUserPassword": null, - "ModifyDBInstanceMessage$DBParameterGroupName": null, - "ModifyDBInstanceMessage$PreferredBackupWindow": null, - "ModifyDBInstanceMessage$PreferredMaintenanceWindow": null, - "ModifyDBInstanceMessage$EngineVersion": null, - "ModifyDBInstanceMessage$OptionGroupName": null, - "ModifyDBInstanceMessage$NewDBInstanceIdentifier": null, - "ModifyDBParameterGroupMessage$DBParameterGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "ModifyEventSubscriptionMessage$SubscriptionName": null, - "ModifyEventSubscriptionMessage$SnsTopicArn": null, - "ModifyEventSubscriptionMessage$SourceType": null, - "ModifyOptionGroupMessage$OptionGroupName": null, - "Option$OptionName": null, - "Option$OptionDescription": null, - "OptionConfiguration$OptionName": null, - "OptionGroup$OptionGroupName": null, - "OptionGroup$OptionGroupDescription": null, - "OptionGroup$EngineName": null, - "OptionGroup$MajorEngineVersion": null, - "OptionGroup$VpcId": null, - "OptionGroupMembership$OptionGroupName": null, - "OptionGroupMembership$Status": null, - "OptionGroupOption$Name": null, - "OptionGroupOption$Description": null, - "OptionGroupOption$EngineName": null, - "OptionGroupOption$MajorEngineVersion": null, - "OptionGroupOption$MinimumRequiredMinorEngineVersion": null, - "OptionGroupOptionsMessage$Marker": null, - "OptionGroups$Marker": null, - "OptionNamesList$member": null, - "OptionsDependedOn$member": null, - "OrderableDBInstanceOption$Engine": null, - "OrderableDBInstanceOption$EngineVersion": null, - "OrderableDBInstanceOption$DBInstanceClass": null, - "OrderableDBInstanceOption$LicenseModel": null, - "OrderableDBInstanceOptionsMessage$Marker": null, - "Parameter$ParameterName": null, - "Parameter$ParameterValue": null, - "Parameter$Description": null, - "Parameter$Source": null, - "Parameter$ApplyType": null, - "Parameter$DataType": null, - "Parameter$AllowedValues": null, - "Parameter$MinimumEngineVersion": null, - "PendingModifiedValues$DBInstanceClass": null, - "PendingModifiedValues$MasterUserPassword": null, - "PendingModifiedValues$EngineVersion": null, - "PendingModifiedValues$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$PreferredBackupWindow": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferingId": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstanceId": null, - "ReadReplicaDBInstanceIdentifierList$member": null, - "RebootDBInstanceMessage$DBInstanceIdentifier": null, - "RecurringCharge$RecurringChargeFrequency": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier": null, - "RemoveTagsFromResourceMessage$ResourceName": null, - "ReservedDBInstance$ReservedDBInstanceId": null, - "ReservedDBInstance$ReservedDBInstancesOfferingId": null, - "ReservedDBInstance$DBInstanceClass": null, - "ReservedDBInstance$CurrencyCode": null, - "ReservedDBInstance$ProductDescription": null, - "ReservedDBInstance$OfferingType": null, - "ReservedDBInstance$State": null, - "ReservedDBInstanceMessage$Marker": null, - "ReservedDBInstancesOffering$ReservedDBInstancesOfferingId": null, - "ReservedDBInstancesOffering$DBInstanceClass": null, - "ReservedDBInstancesOffering$CurrencyCode": null, - "ReservedDBInstancesOffering$ProductDescription": null, - "ReservedDBInstancesOffering$OfferingType": null, - "ReservedDBInstancesOfferingMessage$Marker": null, - "ResetDBParameterGroupMessage$DBParameterGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSnapshotIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceClass": null, - "RestoreDBInstanceFromDBSnapshotMessage$AvailabilityZone": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSubnetGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$LicenseModel": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBName": null, - "RestoreDBInstanceFromDBSnapshotMessage$Engine": null, - "RestoreDBInstanceFromDBSnapshotMessage$OptionGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$SourceDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$TargetDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$DBInstanceClass": null, - "RestoreDBInstanceToPointInTimeMessage$AvailabilityZone": null, - "RestoreDBInstanceToPointInTimeMessage$DBSubnetGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$LicenseModel": null, - "RestoreDBInstanceToPointInTimeMessage$DBName": null, - "RestoreDBInstanceToPointInTimeMessage$Engine": null, - "RestoreDBInstanceToPointInTimeMessage$OptionGroupName": null, - "RevokeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$CIDRIP": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "SourceIdsList$member": null, - "Subnet$SubnetIdentifier": null, - "Subnet$SubnetStatus": null, - "SubnetIdentifierList$member": null, - "Tag$Key": null, - "Tag$Value": null, - "VpcSecurityGroupIdList$member": null, - "VpcSecurityGroupMembership$VpcSecurityGroupId": null, - "VpcSecurityGroupMembership$Status": null - } - }, - "Subnet": { - "base": null, - "refs": { - "SubnetList$member": null - } - }, - "SubnetAlreadyInUse": { - "base": "

The DB subnet is already in use in the Availability Zone.

", - "refs": { - } - }, - "SubnetIdentifierList": { - "base": null, - "refs": { - "CreateDBSubnetGroupMessage$SubnetIds": null, - "ModifyDBSubnetGroupMessage$SubnetIds": null - } - }, - "SubnetList": { - "base": null, - "refs": { - "DBSubnetGroup$Subnets": null - } - }, - "SubscriptionAlreadyExistFault": { - "base": "

The supplied subscription name already exists.

", - "refs": { - } - }, - "SubscriptionCategoryNotFoundFault": { - "base": "

The supplied category does not exist.

", - "refs": { - } - }, - "SubscriptionNotFoundFault": { - "base": "

The subscription name does not exist.

", - "refs": { - } - }, - "SupportedCharacterSetsList": { - "base": null, - "refs": { - "DBEngineVersion$SupportedCharacterSets": null - } - }, - "TStamp": { - "base": null, - "refs": { - "DBInstance$InstanceCreateTime": null, - "DBInstance$LatestRestorableTime": null, - "DBSnapshot$SnapshotCreateTime": null, - "DBSnapshot$InstanceCreateTime": null, - "DescribeEventsMessage$StartTime": null, - "DescribeEventsMessage$EndTime": null, - "Event$Date": null, - "ReservedDBInstance$StartTime": null, - "RestoreDBInstanceToPointInTimeMessage$RestoreTime": null - } - }, - "Tag": { - "base": null, - "refs": { - "TagList$member": null - } - }, - "TagList": { - "base": null, - "refs": { - "AddTagsToResourceMessage$Tags": null, - "TagListMessage$TagList": null - } - }, - "TagListMessage": { - "base": null, - "refs": { - } - }, - "VpcSecurityGroupIdList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$VpcSecurityGroupIds": null, - "ModifyDBInstanceMessage$VpcSecurityGroupIds": null, - "OptionConfiguration$VpcSecurityGroupMemberships": null - } - }, - "VpcSecurityGroupMembership": { - "base": null, - "refs": { - "VpcSecurityGroupMembershipList$member": null - } - }, - "VpcSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$VpcSecurityGroups": null, - "Option$VpcSecurityGroupMemberships": null - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/smoke.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/smoke.json deleted file mode 100644 index 068b23492..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-01-10/smoke.json +++ /dev/null @@ -1,18 +0,0 @@ -{ - "version": 1, - "defaultRegion": "us-west-2", - "testCases": [ - { - "operationName": "DescribeDBEngineVersions", - "input": {}, - "errorExpectedFromService": false - }, - { - "operationName": "DescribeDBInstances", - "input": { - "DBInstanceIdentifier": "fake-id" - }, - "errorExpectedFromService": true - } - ] -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/api-2.json deleted file mode 100644 index 76c0b639d..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/api-2.json +++ /dev/null @@ -1,3059 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2013-02-12", - "endpointPrefix":"rds", - "protocol":"query", - "serviceAbbreviation":"Amazon RDS", - "serviceFullName":"Amazon Relational Database Service", - "serviceId":"RDS", - "signatureVersion":"v4", - "uid":"rds-2013-02-12", - "xmlNamespace":"http://rds.amazonaws.com/doc/2013-02-12/" - }, - "operations":{ - "AddSourceIdentifierToSubscription":{ - "name":"AddSourceIdentifierToSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddSourceIdentifierToSubscriptionMessage"}, - "output":{ - "shape":"AddSourceIdentifierToSubscriptionResult", - "resultWrapper":"AddSourceIdentifierToSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "AddTagsToResource":{ - "name":"AddTagsToResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddTagsToResourceMessage"}, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "AuthorizeDBSecurityGroupIngress":{ - "name":"AuthorizeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"AuthorizeDBSecurityGroupIngressResult", - "resultWrapper":"AuthorizeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"AuthorizationAlreadyExistsFault"}, - {"shape":"AuthorizationQuotaExceededFault"} - ] - }, - "CopyDBSnapshot":{ - "name":"CopyDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyDBSnapshotMessage"}, - "output":{ - "shape":"CopyDBSnapshotResult", - "resultWrapper":"CopyDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "CreateDBInstance":{ - "name":"CreateDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceMessage"}, - "output":{ - "shape":"CreateDBInstanceResult", - "resultWrapper":"CreateDBInstanceResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "CreateDBInstanceReadReplica":{ - "name":"CreateDBInstanceReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceReadReplicaMessage"}, - "output":{ - "shape":"CreateDBInstanceReadReplicaResult", - "resultWrapper":"CreateDBInstanceReadReplicaResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "CreateDBParameterGroup":{ - "name":"CreateDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBParameterGroupMessage"}, - "output":{ - "shape":"CreateDBParameterGroupResult", - "resultWrapper":"CreateDBParameterGroupResult" - }, - "errors":[ - {"shape":"DBParameterGroupQuotaExceededFault"}, - {"shape":"DBParameterGroupAlreadyExistsFault"} - ] - }, - "CreateDBSecurityGroup":{ - "name":"CreateDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSecurityGroupMessage"}, - "output":{ - "shape":"CreateDBSecurityGroupResult", - "resultWrapper":"CreateDBSecurityGroupResult" - }, - "errors":[ - {"shape":"DBSecurityGroupAlreadyExistsFault"}, - {"shape":"DBSecurityGroupQuotaExceededFault"}, - {"shape":"DBSecurityGroupNotSupportedFault"} - ] - }, - "CreateDBSnapshot":{ - "name":"CreateDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSnapshotMessage"}, - "output":{ - "shape":"CreateDBSnapshotResult", - "resultWrapper":"CreateDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "CreateDBSubnetGroup":{ - "name":"CreateDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSubnetGroupMessage"}, - "output":{ - "shape":"CreateDBSubnetGroupResult", - "resultWrapper":"CreateDBSubnetGroupResult" - }, - "errors":[ - {"shape":"DBSubnetGroupAlreadyExistsFault"}, - {"shape":"DBSubnetGroupQuotaExceededFault"}, - {"shape":"DBSubnetQuotaExceededFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"} - ] - }, - "CreateEventSubscription":{ - "name":"CreateEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateEventSubscriptionMessage"}, - "output":{ - "shape":"CreateEventSubscriptionResult", - "resultWrapper":"CreateEventSubscriptionResult" - }, - "errors":[ - {"shape":"EventSubscriptionQuotaExceededFault"}, - {"shape":"SubscriptionAlreadyExistFault"}, - {"shape":"SNSInvalidTopicFault"}, - {"shape":"SNSNoAuthorizationFault"}, - {"shape":"SNSTopicArnNotFoundFault"}, - {"shape":"SubscriptionCategoryNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "CreateOptionGroup":{ - "name":"CreateOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateOptionGroupMessage"}, - "output":{ - "shape":"CreateOptionGroupResult", - "resultWrapper":"CreateOptionGroupResult" - }, - "errors":[ - {"shape":"OptionGroupAlreadyExistsFault"}, - {"shape":"OptionGroupQuotaExceededFault"} - ] - }, - "DeleteDBInstance":{ - "name":"DeleteDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBInstanceMessage"}, - "output":{ - "shape":"DeleteDBInstanceResult", - "resultWrapper":"DeleteDBInstanceResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "DeleteDBParameterGroup":{ - "name":"DeleteDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBParameterGroupMessage"}, - "errors":[ - {"shape":"InvalidDBParameterGroupStateFault"}, - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DeleteDBSecurityGroup":{ - "name":"DeleteDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSecurityGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DeleteDBSnapshot":{ - "name":"DeleteDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSnapshotMessage"}, - "output":{ - "shape":"DeleteDBSnapshotResult", - "resultWrapper":"DeleteDBSnapshotResult" - }, - "errors":[ - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "DeleteDBSubnetGroup":{ - "name":"DeleteDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSubnetGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSubnetGroupStateFault"}, - {"shape":"InvalidDBSubnetStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"} - ] - }, - "DeleteEventSubscription":{ - "name":"DeleteEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteEventSubscriptionMessage"}, - "output":{ - "shape":"DeleteEventSubscriptionResult", - "resultWrapper":"DeleteEventSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"InvalidEventSubscriptionStateFault"} - ] - }, - "DeleteOptionGroup":{ - "name":"DeleteOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteOptionGroupMessage"}, - "errors":[ - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"InvalidOptionGroupStateFault"} - ] - }, - "DescribeDBEngineVersions":{ - "name":"DescribeDBEngineVersions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBEngineVersionsMessage"}, - "output":{ - "shape":"DBEngineVersionMessage", - "resultWrapper":"DescribeDBEngineVersionsResult" - } - }, - "DescribeDBInstances":{ - "name":"DescribeDBInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBInstancesMessage"}, - "output":{ - "shape":"DBInstanceMessage", - "resultWrapper":"DescribeDBInstancesResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "DescribeDBLogFiles":{ - "name":"DescribeDBLogFiles", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBLogFilesMessage"}, - "output":{ - "shape":"DescribeDBLogFilesResponse", - "resultWrapper":"DescribeDBLogFilesResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "DescribeDBParameterGroups":{ - "name":"DescribeDBParameterGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBParameterGroupsMessage"}, - "output":{ - "shape":"DBParameterGroupsMessage", - "resultWrapper":"DescribeDBParameterGroupsResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DescribeDBParameters":{ - "name":"DescribeDBParameters", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBParametersMessage"}, - "output":{ - "shape":"DBParameterGroupDetails", - "resultWrapper":"DescribeDBParametersResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DescribeDBSecurityGroups":{ - "name":"DescribeDBSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSecurityGroupsMessage"}, - "output":{ - "shape":"DBSecurityGroupMessage", - "resultWrapper":"DescribeDBSecurityGroupsResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DescribeDBSnapshots":{ - "name":"DescribeDBSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSnapshotsMessage"}, - "output":{ - "shape":"DBSnapshotMessage", - "resultWrapper":"DescribeDBSnapshotsResult" - }, - "errors":[ - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "DescribeDBSubnetGroups":{ - "name":"DescribeDBSubnetGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSubnetGroupsMessage"}, - "output":{ - "shape":"DBSubnetGroupMessage", - "resultWrapper":"DescribeDBSubnetGroupsResult" - }, - "errors":[ - {"shape":"DBSubnetGroupNotFoundFault"} - ] - }, - "DescribeEngineDefaultParameters":{ - "name":"DescribeEngineDefaultParameters", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEngineDefaultParametersMessage"}, - "output":{ - "shape":"DescribeEngineDefaultParametersResult", - "resultWrapper":"DescribeEngineDefaultParametersResult" - } - }, - "DescribeEventCategories":{ - "name":"DescribeEventCategories", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventCategoriesMessage"}, - "output":{ - "shape":"EventCategoriesMessage", - "resultWrapper":"DescribeEventCategoriesResult" - } - }, - "DescribeEventSubscriptions":{ - "name":"DescribeEventSubscriptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventSubscriptionsMessage"}, - "output":{ - "shape":"EventSubscriptionsMessage", - "resultWrapper":"DescribeEventSubscriptionsResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"} - ] - }, - "DescribeEvents":{ - "name":"DescribeEvents", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventsMessage"}, - "output":{ - "shape":"EventsMessage", - "resultWrapper":"DescribeEventsResult" - } - }, - "DescribeOptionGroupOptions":{ - "name":"DescribeOptionGroupOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupOptionsMessage"}, - "output":{ - "shape":"OptionGroupOptionsMessage", - "resultWrapper":"DescribeOptionGroupOptionsResult" - } - }, - "DescribeOptionGroups":{ - "name":"DescribeOptionGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupsMessage"}, - "output":{ - "shape":"OptionGroups", - "resultWrapper":"DescribeOptionGroupsResult" - }, - "errors":[ - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "DescribeOrderableDBInstanceOptions":{ - "name":"DescribeOrderableDBInstanceOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOrderableDBInstanceOptionsMessage"}, - "output":{ - "shape":"OrderableDBInstanceOptionsMessage", - "resultWrapper":"DescribeOrderableDBInstanceOptionsResult" - } - }, - "DescribeReservedDBInstances":{ - "name":"DescribeReservedDBInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedDBInstancesMessage"}, - "output":{ - "shape":"ReservedDBInstanceMessage", - "resultWrapper":"DescribeReservedDBInstancesResult" - }, - "errors":[ - {"shape":"ReservedDBInstanceNotFoundFault"} - ] - }, - "DescribeReservedDBInstancesOfferings":{ - "name":"DescribeReservedDBInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedDBInstancesOfferingsMessage"}, - "output":{ - "shape":"ReservedDBInstancesOfferingMessage", - "resultWrapper":"DescribeReservedDBInstancesOfferingsResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"} - ] - }, - "DownloadDBLogFilePortion":{ - "name":"DownloadDBLogFilePortion", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DownloadDBLogFilePortionMessage"}, - "output":{ - "shape":"DownloadDBLogFilePortionDetails", - "resultWrapper":"DownloadDBLogFilePortionResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBLogFileNotFoundFault"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListTagsForResourceMessage"}, - "output":{ - "shape":"TagListMessage", - "resultWrapper":"ListTagsForResourceResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "ModifyDBInstance":{ - "name":"ModifyDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBInstanceMessage"}, - "output":{ - "shape":"ModifyDBInstanceResult", - "resultWrapper":"ModifyDBInstanceResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"DBUpgradeDependencyFailureFault"} - ] - }, - "ModifyDBParameterGroup":{ - "name":"ModifyDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBParameterGroupMessage"}, - "output":{ - "shape":"DBParameterGroupNameMessage", - "resultWrapper":"ModifyDBParameterGroupResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"InvalidDBParameterGroupStateFault"} - ] - }, - "ModifyDBSubnetGroup":{ - "name":"ModifyDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBSubnetGroupMessage"}, - "output":{ - "shape":"ModifyDBSubnetGroupResult", - "resultWrapper":"ModifyDBSubnetGroupResult" - }, - "errors":[ - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetQuotaExceededFault"}, - {"shape":"SubnetAlreadyInUse"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"} - ] - }, - "ModifyEventSubscription":{ - "name":"ModifyEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyEventSubscriptionMessage"}, - "output":{ - "shape":"ModifyEventSubscriptionResult", - "resultWrapper":"ModifyEventSubscriptionResult" - }, - "errors":[ - {"shape":"EventSubscriptionQuotaExceededFault"}, - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SNSInvalidTopicFault"}, - {"shape":"SNSNoAuthorizationFault"}, - {"shape":"SNSTopicArnNotFoundFault"}, - {"shape":"SubscriptionCategoryNotFoundFault"} - ] - }, - "ModifyOptionGroup":{ - "name":"ModifyOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyOptionGroupMessage"}, - "output":{ - "shape":"ModifyOptionGroupResult", - "resultWrapper":"ModifyOptionGroupResult" - }, - "errors":[ - {"shape":"InvalidOptionGroupStateFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "PromoteReadReplica":{ - "name":"PromoteReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PromoteReadReplicaMessage"}, - "output":{ - "shape":"PromoteReadReplicaResult", - "resultWrapper":"PromoteReadReplicaResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "PurchaseReservedDBInstancesOffering":{ - "name":"PurchaseReservedDBInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedDBInstancesOfferingMessage"}, - "output":{ - "shape":"PurchaseReservedDBInstancesOfferingResult", - "resultWrapper":"PurchaseReservedDBInstancesOfferingResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"}, - {"shape":"ReservedDBInstanceAlreadyExistsFault"}, - {"shape":"ReservedDBInstanceQuotaExceededFault"} - ] - }, - "RebootDBInstance":{ - "name":"RebootDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootDBInstanceMessage"}, - "output":{ - "shape":"RebootDBInstanceResult", - "resultWrapper":"RebootDBInstanceResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "RemoveSourceIdentifierFromSubscription":{ - "name":"RemoveSourceIdentifierFromSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveSourceIdentifierFromSubscriptionMessage"}, - "output":{ - "shape":"RemoveSourceIdentifierFromSubscriptionResult", - "resultWrapper":"RemoveSourceIdentifierFromSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "RemoveTagsFromResource":{ - "name":"RemoveTagsFromResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveTagsFromResourceMessage"}, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "ResetDBParameterGroup":{ - "name":"ResetDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetDBParameterGroupMessage"}, - "output":{ - "shape":"DBParameterGroupNameMessage", - "resultWrapper":"ResetDBParameterGroupResult" - }, - "errors":[ - {"shape":"InvalidDBParameterGroupStateFault"}, - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "RestoreDBInstanceFromDBSnapshot":{ - "name":"RestoreDBInstanceFromDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreDBInstanceFromDBSnapshotMessage"}, - "output":{ - "shape":"RestoreDBInstanceFromDBSnapshotResult", - "resultWrapper":"RestoreDBInstanceFromDBSnapshotResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "RestoreDBInstanceToPointInTime":{ - "name":"RestoreDBInstanceToPointInTime", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreDBInstanceToPointInTimeMessage"}, - "output":{ - "shape":"RestoreDBInstanceToPointInTimeResult", - "resultWrapper":"RestoreDBInstanceToPointInTimeResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"PointInTimeRestoreNotEnabledFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "RevokeDBSecurityGroupIngress":{ - "name":"RevokeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"RevokeDBSecurityGroupIngressResult", - "resultWrapper":"RevokeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"AuthorizationNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"} - ] - } - }, - "shapes":{ - "AddSourceIdentifierToSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SourceIdentifier" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SourceIdentifier":{"shape":"String"} - } - }, - "AddSourceIdentifierToSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "AddTagsToResourceMessage":{ - "type":"structure", - "required":[ - "ResourceName", - "Tags" - ], - "members":{ - "ResourceName":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "ApplyMethod":{ - "type":"string", - "enum":[ - "immediate", - "pending-reboot" - ] - }, - "AuthorizationAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizationNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "AuthorizationQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "AuthorizeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "ProvisionedIopsCapable":{"shape":"Boolean"} - }, - "wrapper":true - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"AvailabilityZone" - } - }, - "Boolean":{"type":"boolean"}, - "BooleanOptional":{"type":"boolean"}, - "CharacterSet":{ - "type":"structure", - "members":{ - "CharacterSetName":{"shape":"String"}, - "CharacterSetDescription":{"shape":"String"} - } - }, - "CopyDBSnapshotMessage":{ - "type":"structure", - "required":[ - "SourceDBSnapshotIdentifier", - "TargetDBSnapshotIdentifier" - ], - "members":{ - "SourceDBSnapshotIdentifier":{"shape":"String"}, - "TargetDBSnapshotIdentifier":{"shape":"String"} - } - }, - "CopyDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "CreateDBInstanceMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "AllocatedStorage", - "DBInstanceClass", - "Engine", - "MasterUsername", - "MasterUserPassword" - ], - "members":{ - "DBName":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "DBInstanceClass":{"shape":"String"}, - "Engine":{"shape":"String"}, - "MasterUsername":{"shape":"String"}, - "MasterUserPassword":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "DBParameterGroupName":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "CharacterSetName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"} - } - }, - "CreateDBInstanceReadReplicaMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "SourceDBInstanceIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SourceDBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"} - } - }, - "CreateDBInstanceReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "CreateDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "CreateDBParameterGroupMessage":{ - "type":"structure", - "required":[ - "DBParameterGroupName", - "DBParameterGroupFamily", - "Description" - ], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"} - } - }, - "CreateDBParameterGroupResult":{ - "type":"structure", - "members":{ - "DBParameterGroup":{"shape":"DBParameterGroup"} - } - }, - "CreateDBSecurityGroupMessage":{ - "type":"structure", - "required":[ - "DBSecurityGroupName", - "DBSecurityGroupDescription" - ], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"} - } - }, - "CreateDBSecurityGroupResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "CreateDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBSnapshotIdentifier", - "DBInstanceIdentifier" - ], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"} - } - }, - "CreateDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "CreateDBSubnetGroupMessage":{ - "type":"structure", - "required":[ - "DBSubnetGroupName", - "DBSubnetGroupDescription", - "SubnetIds" - ], - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "SubnetIds":{"shape":"SubnetIdentifierList"} - } - }, - "CreateDBSubnetGroupResult":{ - "type":"structure", - "members":{ - "DBSubnetGroup":{"shape":"DBSubnetGroup"} - } - }, - "CreateEventSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SnsTopicArn" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "SourceIds":{"shape":"SourceIdsList"}, - "Enabled":{"shape":"BooleanOptional"} - } - }, - "CreateEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "CreateOptionGroupMessage":{ - "type":"structure", - "required":[ - "OptionGroupName", - "EngineName", - "MajorEngineVersion", - "OptionGroupDescription" - ], - "members":{ - "OptionGroupName":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"} - } - }, - "CreateOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "DBEngineVersion":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "DBEngineDescription":{"shape":"String"}, - "DBEngineVersionDescription":{"shape":"String"}, - "DefaultCharacterSet":{"shape":"CharacterSet"}, - "SupportedCharacterSets":{"shape":"SupportedCharacterSetsList"} - } - }, - "DBEngineVersionList":{ - "type":"list", - "member":{ - "shape":"DBEngineVersion", - "locationName":"DBEngineVersion" - } - }, - "DBEngineVersionMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBEngineVersions":{"shape":"DBEngineVersionList"} - } - }, - "DBInstance":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Engine":{"shape":"String"}, - "DBInstanceStatus":{"shape":"String"}, - "MasterUsername":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Endpoint":{"shape":"Endpoint"}, - "AllocatedStorage":{"shape":"Integer"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "PreferredBackupWindow":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"Integer"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroups":{"shape":"VpcSecurityGroupMembershipList"}, - "DBParameterGroups":{"shape":"DBParameterGroupStatusList"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroup":{"shape":"DBSubnetGroup"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "PendingModifiedValues":{"shape":"PendingModifiedValues"}, - "LatestRestorableTime":{"shape":"TStamp"}, - "MultiAZ":{"shape":"Boolean"}, - "EngineVersion":{"shape":"String"}, - "AutoMinorVersionUpgrade":{"shape":"Boolean"}, - "ReadReplicaSourceDBInstanceIdentifier":{"shape":"String"}, - "ReadReplicaDBInstanceIdentifiers":{"shape":"ReadReplicaDBInstanceIdentifierList"}, - "LicenseModel":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupMemberships":{"shape":"OptionGroupMembershipList"}, - "CharacterSetName":{"shape":"String"}, - "SecondaryAvailabilityZone":{"shape":"String"}, - "PubliclyAccessible":{"shape":"Boolean"} - }, - "wrapper":true - }, - "DBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBInstanceAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBInstanceList":{ - "type":"list", - "member":{ - "shape":"DBInstance", - "locationName":"DBInstance" - } - }, - "DBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBInstances":{"shape":"DBInstanceList"} - } - }, - "DBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBLogFileNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBLogFileNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroup":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"} - }, - "wrapper":true - }, - "DBParameterGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupDetails":{ - "type":"structure", - "members":{ - "Parameters":{"shape":"ParametersList"}, - "Marker":{"shape":"String"} - } - }, - "DBParameterGroupList":{ - "type":"list", - "member":{ - "shape":"DBParameterGroup", - "locationName":"DBParameterGroup" - } - }, - "DBParameterGroupNameMessage":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"} - } - }, - "DBParameterGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupStatus":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "ParameterApplyStatus":{"shape":"String"} - } - }, - "DBParameterGroupStatusList":{ - "type":"list", - "member":{ - "shape":"DBParameterGroupStatus", - "locationName":"DBParameterGroup" - } - }, - "DBParameterGroupsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBParameterGroups":{"shape":"DBParameterGroupList"} - } - }, - "DBSecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{"shape":"String"}, - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "EC2SecurityGroups":{"shape":"EC2SecurityGroupList"}, - "IPRanges":{"shape":"IPRangeList"} - }, - "wrapper":true - }, - "DBSecurityGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupMembership":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "DBSecurityGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroupMembership", - "locationName":"DBSecurityGroup" - } - }, - "DBSecurityGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroups"} - } - }, - "DBSecurityGroupNameList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DBSecurityGroupName" - } - }, - "DBSecurityGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupNotSupportedFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotSupported", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"QuotaExceeded.DBSecurityGroup", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroups":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroup", - "locationName":"DBSecurityGroup" - } - }, - "DBSnapshot":{ - "type":"structure", - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "SnapshotCreateTime":{"shape":"TStamp"}, - "Engine":{"shape":"String"}, - "AllocatedStorage":{"shape":"Integer"}, - "Status":{"shape":"String"}, - "Port":{"shape":"Integer"}, - "AvailabilityZone":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "MasterUsername":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"} - }, - "wrapper":true - }, - "DBSnapshotAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSnapshotAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSnapshotList":{ - "type":"list", - "member":{ - "shape":"DBSnapshot", - "locationName":"DBSnapshot" - } - }, - "DBSnapshotMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSnapshots":{"shape":"DBSnapshotList"} - } - }, - "DBSnapshotNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSnapshotNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroup":{ - "type":"structure", - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "SubnetGroupStatus":{"shape":"String"}, - "Subnets":{"shape":"SubnetList"} - }, - "wrapper":true - }, - "DBSubnetGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupDoesNotCoverEnoughAZs":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupDoesNotCoverEnoughAZs", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSubnetGroups":{"shape":"DBSubnetGroups"} - } - }, - "DBSubnetGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroups":{ - "type":"list", - "member":{ - "shape":"DBSubnetGroup", - "locationName":"DBSubnetGroup" - } - }, - "DBSubnetQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBUpgradeDependencyFailureFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBUpgradeDependencyFailure", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DeleteDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SkipFinalSnapshot":{"shape":"Boolean"}, - "FinalDBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "DeleteDBParameterGroupMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"} - } - }, - "DeleteDBSecurityGroupMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"} - } - }, - "DeleteDBSnapshotMessage":{ - "type":"structure", - "required":["DBSnapshotIdentifier"], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "DeleteDBSubnetGroupMessage":{ - "type":"structure", - "required":["DBSubnetGroupName"], - "members":{ - "DBSubnetGroupName":{"shape":"String"} - } - }, - "DeleteEventSubscriptionMessage":{ - "type":"structure", - "required":["SubscriptionName"], - "members":{ - "SubscriptionName":{"shape":"String"} - } - }, - "DeleteEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "DeleteOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"} - } - }, - "DescribeDBEngineVersionsMessage":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"}, - "DefaultOnly":{"shape":"Boolean"}, - "ListSupportedCharacterSets":{"shape":"BooleanOptional"} - } - }, - "DescribeDBInstancesMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBLogFilesDetails":{ - "type":"structure", - "members":{ - "LogFileName":{"shape":"String"}, - "LastWritten":{"shape":"Long"}, - "Size":{"shape":"Long"} - } - }, - "DescribeDBLogFilesList":{ - "type":"list", - "member":{ - "shape":"DescribeDBLogFilesDetails", - "locationName":"DescribeDBLogFilesDetails" - } - }, - "DescribeDBLogFilesMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "FilenameContains":{"shape":"String"}, - "FileLastWritten":{"shape":"Long"}, - "FileSize":{"shape":"Long"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBLogFilesResponse":{ - "type":"structure", - "members":{ - "DescribeDBLogFiles":{"shape":"DescribeDBLogFilesList"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBParameterGroupsMessage":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBParametersMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Source":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSecurityGroupsMessage":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSnapshotsMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSubnetGroupsMessage":{ - "type":"structure", - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEngineDefaultParametersMessage":{ - "type":"structure", - "required":["DBParameterGroupFamily"], - "members":{ - "DBParameterGroupFamily":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEngineDefaultParametersResult":{ - "type":"structure", - "members":{ - "EngineDefaults":{"shape":"EngineDefaults"} - } - }, - "DescribeEventCategoriesMessage":{ - "type":"structure", - "members":{ - "SourceType":{"shape":"String"} - } - }, - "DescribeEventSubscriptionsMessage":{ - "type":"structure", - "members":{ - "SubscriptionName":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEventsMessage":{ - "type":"structure", - "members":{ - "SourceIdentifier":{"shape":"String"}, - "SourceType":{"shape":"SourceType"}, - "StartTime":{"shape":"TStamp"}, - "EndTime":{"shape":"TStamp"}, - "Duration":{"shape":"IntegerOptional"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupOptionsMessage":{ - "type":"structure", - "required":["EngineName"], - "members":{ - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupsMessage":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Marker":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"} - } - }, - "DescribeOrderableDBInstanceOptionsMessage":{ - "type":"structure", - "required":["Engine"], - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "Vpc":{"shape":"BooleanOptional"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesOfferingsMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "Double":{"type":"double"}, - "DownloadDBLogFilePortionDetails":{ - "type":"structure", - "members":{ - "LogFileData":{"shape":"String"}, - "Marker":{"shape":"String"}, - "AdditionalDataPending":{"shape":"Boolean"} - } - }, - "DownloadDBLogFilePortionMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "LogFileName" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "LogFileName":{"shape":"String"}, - "Marker":{"shape":"String"}, - "NumberOfLines":{"shape":"Integer"} - } - }, - "EC2SecurityGroup":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "EC2SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"EC2SecurityGroup", - "locationName":"EC2SecurityGroup" - } - }, - "Endpoint":{ - "type":"structure", - "members":{ - "Address":{"shape":"String"}, - "Port":{"shape":"Integer"} - } - }, - "EngineDefaults":{ - "type":"structure", - "members":{ - "DBParameterGroupFamily":{"shape":"String"}, - "Marker":{"shape":"String"}, - "Parameters":{"shape":"ParametersList"} - }, - "wrapper":true - }, - "Event":{ - "type":"structure", - "members":{ - "SourceIdentifier":{"shape":"String"}, - "SourceType":{"shape":"SourceType"}, - "Message":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Date":{"shape":"TStamp"} - } - }, - "EventCategoriesList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"EventCategory" - } - }, - "EventCategoriesMap":{ - "type":"structure", - "members":{ - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"} - }, - "wrapper":true - }, - "EventCategoriesMapList":{ - "type":"list", - "member":{ - "shape":"EventCategoriesMap", - "locationName":"EventCategoriesMap" - } - }, - "EventCategoriesMessage":{ - "type":"structure", - "members":{ - "EventCategoriesMapList":{"shape":"EventCategoriesMapList"} - } - }, - "EventList":{ - "type":"list", - "member":{ - "shape":"Event", - "locationName":"Event" - } - }, - "EventSubscription":{ - "type":"structure", - "members":{ - "CustomerAwsId":{"shape":"String"}, - "CustSubscriptionId":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "Status":{"shape":"String"}, - "SubscriptionCreationTime":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "SourceIdsList":{"shape":"SourceIdsList"}, - "EventCategoriesList":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"Boolean"} - }, - "wrapper":true - }, - "EventSubscriptionQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"EventSubscriptionQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "EventSubscriptionsList":{ - "type":"list", - "member":{ - "shape":"EventSubscription", - "locationName":"EventSubscription" - } - }, - "EventSubscriptionsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "EventSubscriptionsList":{"shape":"EventSubscriptionsList"} - } - }, - "EventsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "Events":{"shape":"EventList"} - } - }, - "IPRange":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "CIDRIP":{"shape":"String"} - } - }, - "IPRangeList":{ - "type":"list", - "member":{ - "shape":"IPRange", - "locationName":"IPRange" - } - }, - "InstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InsufficientDBInstanceCapacityFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InsufficientDBInstanceCapacity", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "Integer":{"type":"integer"}, - "IntegerOptional":{"type":"integer"}, - "InvalidDBInstanceStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBInstanceState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBParameterGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBParameterGroupState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSecurityGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSecurityGroupState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSnapshotStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSnapshotState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidEventSubscriptionStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidEventSubscriptionState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidOptionGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidOptionGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidRestoreFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidRestoreFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidSubnet":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidSubnet", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidVPCNetworkStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidVPCNetworkStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "KeyList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ListTagsForResourceMessage":{ - "type":"structure", - "required":["ResourceName"], - "members":{ - "ResourceName":{"shape":"String"} - } - }, - "Long":{"type":"long"}, - "ModifyDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "DBInstanceClass":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, - "ApplyImmediately":{"shape":"Boolean"}, - "MasterUserPassword":{"shape":"String"}, - "DBParameterGroupName":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "AllowMajorVersionUpgrade":{"shape":"Boolean"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "NewDBInstanceIdentifier":{"shape":"String"} - } - }, - "ModifyDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "ModifyDBParameterGroupMessage":{ - "type":"structure", - "required":[ - "DBParameterGroupName", - "Parameters" - ], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Parameters":{"shape":"ParametersList"} - } - }, - "ModifyDBSubnetGroupMessage":{ - "type":"structure", - "required":[ - "DBSubnetGroupName", - "SubnetIds" - ], - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "SubnetIds":{"shape":"SubnetIdentifierList"} - } - }, - "ModifyDBSubnetGroupResult":{ - "type":"structure", - "members":{ - "DBSubnetGroup":{"shape":"DBSubnetGroup"} - } - }, - "ModifyEventSubscriptionMessage":{ - "type":"structure", - "required":["SubscriptionName"], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"BooleanOptional"} - } - }, - "ModifyEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "ModifyOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionsToInclude":{"shape":"OptionConfigurationList"}, - "OptionsToRemove":{"shape":"OptionNamesList"}, - "ApplyImmediately":{"shape":"Boolean"} - } - }, - "ModifyOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "Option":{ - "type":"structure", - "members":{ - "OptionName":{"shape":"String"}, - "OptionDescription":{"shape":"String"}, - "Persistent":{"shape":"Boolean"}, - "Port":{"shape":"IntegerOptional"}, - "OptionSettings":{"shape":"OptionSettingConfigurationList"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupMembershipList"} - } - }, - "OptionConfiguration":{ - "type":"structure", - "required":["OptionName"], - "members":{ - "OptionName":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupIdList"}, - "OptionSettings":{"shape":"OptionSettingsList"} - } - }, - "OptionConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionConfiguration", - "locationName":"OptionConfiguration" - } - }, - "OptionGroup":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "Options":{"shape":"OptionsList"}, - "AllowsVpcAndNonVpcInstanceMemberships":{"shape":"Boolean"}, - "VpcId":{"shape":"String"} - }, - "wrapper":true - }, - "OptionGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupAlreadyExistsFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroupMembership":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "OptionGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"OptionGroupMembership", - "locationName":"OptionGroupMembership" - } - }, - "OptionGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "OptionGroupOption":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "MinimumRequiredMinorEngineVersion":{"shape":"String"}, - "PortRequired":{"shape":"Boolean"}, - "DefaultPort":{"shape":"IntegerOptional"}, - "OptionsDependedOn":{"shape":"OptionsDependedOn"}, - "Persistent":{"shape":"Boolean"}, - "OptionGroupOptionSettings":{"shape":"OptionGroupOptionSettingsList"} - } - }, - "OptionGroupOptionSetting":{ - "type":"structure", - "members":{ - "SettingName":{"shape":"String"}, - "SettingDescription":{"shape":"String"}, - "DefaultValue":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"} - } - }, - "OptionGroupOptionSettingsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOptionSetting", - "locationName":"OptionGroupOptionSetting" - } - }, - "OptionGroupOptionsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOption", - "locationName":"OptionGroupOption" - } - }, - "OptionGroupOptionsMessage":{ - "type":"structure", - "members":{ - "OptionGroupOptions":{"shape":"OptionGroupOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroups":{ - "type":"structure", - "members":{ - "OptionGroupsList":{"shape":"OptionGroupsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupsList":{ - "type":"list", - "member":{ - "shape":"OptionGroup", - "locationName":"OptionGroup" - } - }, - "OptionNamesList":{ - "type":"list", - "member":{"shape":"String"} - }, - "OptionSetting":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Value":{"shape":"String"}, - "DefaultValue":{"shape":"String"}, - "Description":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "DataType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"}, - "IsCollection":{"shape":"Boolean"} - } - }, - "OptionSettingConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionSetting", - "locationName":"OptionSetting" - } - }, - "OptionSettingsList":{ - "type":"list", - "member":{ - "shape":"OptionSetting", - "locationName":"OptionSetting" - } - }, - "OptionsDependedOn":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"OptionName" - } - }, - "OptionsList":{ - "type":"list", - "member":{ - "shape":"Option", - "locationName":"Option" - } - }, - "OrderableDBInstanceOption":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "AvailabilityZones":{"shape":"AvailabilityZoneList"}, - "MultiAZCapable":{"shape":"Boolean"}, - "ReadReplicaCapable":{"shape":"Boolean"}, - "Vpc":{"shape":"Boolean"} - }, - "wrapper":true - }, - "OrderableDBInstanceOptionsList":{ - "type":"list", - "member":{ - "shape":"OrderableDBInstanceOption", - "locationName":"OrderableDBInstanceOption" - } - }, - "OrderableDBInstanceOptionsMessage":{ - "type":"structure", - "members":{ - "OrderableDBInstanceOptions":{"shape":"OrderableDBInstanceOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "Parameter":{ - "type":"structure", - "members":{ - "ParameterName":{"shape":"String"}, - "ParameterValue":{"shape":"String"}, - "Description":{"shape":"String"}, - "Source":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "DataType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"}, - "MinimumEngineVersion":{"shape":"String"}, - "ApplyMethod":{"shape":"ApplyMethod"} - } - }, - "ParametersList":{ - "type":"list", - "member":{ - "shape":"Parameter", - "locationName":"Parameter" - } - }, - "PendingModifiedValues":{ - "type":"structure", - "members":{ - "DBInstanceClass":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "MasterUserPassword":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "DBInstanceIdentifier":{"shape":"String"} - } - }, - "PointInTimeRestoreNotEnabledFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"PointInTimeRestoreNotEnabled", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PromoteReadReplicaMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"} - } - }, - "PromoteReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "ProvisionedIopsNotAvailableInAZFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ProvisionedIopsNotAvailableInAZFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PurchaseReservedDBInstancesOfferingMessage":{ - "type":"structure", - "required":["ReservedDBInstancesOfferingId"], - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "ReservedDBInstanceId":{"shape":"String"}, - "DBInstanceCount":{"shape":"IntegerOptional"} - } - }, - "PurchaseReservedDBInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedDBInstance":{"shape":"ReservedDBInstance"} - } - }, - "ReadReplicaDBInstanceIdentifierList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReadReplicaDBInstanceIdentifier" - } - }, - "RebootDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "ForceFailover":{"shape":"BooleanOptional"} - } - }, - "RebootDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "RecurringChargeAmount":{"shape":"Double"}, - "RecurringChargeFrequency":{"shape":"String"} - }, - "wrapper":true - }, - "RecurringChargeList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"RecurringCharge" - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SourceIdentifier" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SourceIdentifier":{"shape":"String"} - } - }, - "RemoveSourceIdentifierFromSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "RemoveTagsFromResourceMessage":{ - "type":"structure", - "required":[ - "ResourceName", - "TagKeys" - ], - "members":{ - "ResourceName":{"shape":"String"}, - "TagKeys":{"shape":"KeyList"} - } - }, - "ReservedDBInstance":{ - "type":"structure", - "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "StartTime":{"shape":"TStamp"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "DBInstanceCount":{"shape":"Integer"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "State":{"shape":"String"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceAlreadyExists", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstance", - "locationName":"ReservedDBInstance" - } - }, - "ReservedDBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstances":{"shape":"ReservedDBInstanceList"} - } - }, - "ReservedDBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstancesOffering", - "locationName":"ReservedDBInstancesOffering" - } - }, - "ReservedDBInstancesOfferingMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstancesOfferings":{"shape":"ReservedDBInstancesOfferingList"} - } - }, - "ReservedDBInstancesOfferingNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstancesOfferingNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ResetDBParameterGroupMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "ResetAllParameters":{"shape":"Boolean"}, - "Parameters":{"shape":"ParametersList"} - } - }, - "RestoreDBInstanceFromDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "DBSnapshotIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"} - } - }, - "RestoreDBInstanceFromDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RestoreDBInstanceToPointInTimeMessage":{ - "type":"structure", - "required":[ - "SourceDBInstanceIdentifier", - "TargetDBInstanceIdentifier" - ], - "members":{ - "SourceDBInstanceIdentifier":{"shape":"String"}, - "TargetDBInstanceIdentifier":{"shape":"String"}, - "RestoreTime":{"shape":"TStamp"}, - "UseLatestRestorableTime":{"shape":"Boolean"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"} - } - }, - "RestoreDBInstanceToPointInTimeResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RevokeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "RevokeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "SNSInvalidTopicFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSInvalidTopic", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SNSNoAuthorizationFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSNoAuthorization", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SNSTopicArnNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSTopicArnNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SnapshotQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SnapshotQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SourceIdsList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SourceId" - } - }, - "SourceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SourceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SourceType":{ - "type":"string", - "enum":[ - "db-instance", - "db-parameter-group", - "db-security-group", - "db-snapshot" - ] - }, - "StorageQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"StorageQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetIdentifier":{"shape":"String"}, - "SubnetAvailabilityZone":{"shape":"AvailabilityZone"}, - "SubnetStatus":{"shape":"String"} - } - }, - "SubnetAlreadyInUse":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubnetAlreadyInUse", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SubnetIdentifierList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetIdentifier" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"Subnet" - } - }, - "SubscriptionAlreadyExistFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionAlreadyExist", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SubscriptionCategoryNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionCategoryNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SubscriptionNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SupportedCharacterSetsList":{ - "type":"list", - "member":{ - "shape":"CharacterSet", - "locationName":"CharacterSet" - } - }, - "TStamp":{"type":"timestamp"}, - "Tag":{ - "type":"structure", - "members":{ - "Key":{"shape":"String"}, - "Value":{"shape":"String"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagListMessage":{ - "type":"structure", - "members":{ - "TagList":{"shape":"TagList"} - } - }, - "VpcSecurityGroupIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcSecurityGroupId" - } - }, - "VpcSecurityGroupMembership":{ - "type":"structure", - "members":{ - "VpcSecurityGroupId":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "VpcSecurityGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"VpcSecurityGroupMembership", - "locationName":"VpcSecurityGroupMembership" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/docs-2.json deleted file mode 100644 index 3d2d55e05..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/docs-2.json +++ /dev/null @@ -1,1796 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "AddSourceIdentifierToSubscription": null, - "AddTagsToResource": null, - "AuthorizeDBSecurityGroupIngress": null, - "CopyDBSnapshot": null, - "CreateDBInstance": null, - "CreateDBInstanceReadReplica": null, - "CreateDBParameterGroup": null, - "CreateDBSecurityGroup": null, - "CreateDBSnapshot": null, - "CreateDBSubnetGroup": null, - "CreateEventSubscription": null, - "CreateOptionGroup": null, - "DeleteDBInstance": null, - "DeleteDBParameterGroup": null, - "DeleteDBSecurityGroup": null, - "DeleteDBSnapshot": null, - "DeleteDBSubnetGroup": null, - "DeleteEventSubscription": null, - "DeleteOptionGroup": null, - "DescribeDBEngineVersions": null, - "DescribeDBInstances": null, - "DescribeDBLogFiles": null, - "DescribeDBParameterGroups": null, - "DescribeDBParameters": null, - "DescribeDBSecurityGroups": null, - "DescribeDBSnapshots": null, - "DescribeDBSubnetGroups": null, - "DescribeEngineDefaultParameters": null, - "DescribeEventCategories": null, - "DescribeEventSubscriptions": null, - "DescribeEvents": null, - "DescribeOptionGroupOptions": null, - "DescribeOptionGroups": null, - "DescribeOrderableDBInstanceOptions": null, - "DescribeReservedDBInstances": null, - "DescribeReservedDBInstancesOfferings": null, - "DownloadDBLogFilePortion": null, - "ListTagsForResource": null, - "ModifyDBInstance": null, - "ModifyDBParameterGroup": null, - "ModifyDBSubnetGroup": null, - "ModifyEventSubscription": null, - "ModifyOptionGroup": null, - "PromoteReadReplica": null, - "PurchaseReservedDBInstancesOffering": null, - "RebootDBInstance": null, - "RemoveSourceIdentifierFromSubscription": null, - "RemoveTagsFromResource": null, - "ResetDBParameterGroup": null, - "RestoreDBInstanceFromDBSnapshot": null, - "RestoreDBInstanceToPointInTime": null, - "RevokeDBSecurityGroupIngress": null - }, - "shapes": { - "AddSourceIdentifierToSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "AddSourceIdentifierToSubscriptionResult": { - "base": null, - "refs": { - } - }, - "AddTagsToResourceMessage": { - "base": null, - "refs": { - } - }, - "ApplyMethod": { - "base": null, - "refs": { - "Parameter$ApplyMethod": null - } - }, - "AuthorizationAlreadyExistsFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group is already authorized for the specified DB security group.

", - "refs": { - } - }, - "AuthorizationNotFoundFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group isn't authorized for the specified DB security group.

RDS also may not be authorized by using IAM to perform necessary actions on your behalf.

", - "refs": { - } - }, - "AuthorizationQuotaExceededFault": { - "base": "

The DB security group authorization quota has been reached.

", - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "AvailabilityZone": { - "base": null, - "refs": { - "AvailabilityZoneList$member": null, - "Subnet$SubnetAvailabilityZone": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "OrderableDBInstanceOption$AvailabilityZones": null - } - }, - "Boolean": { - "base": null, - "refs": { - "AvailabilityZone$ProvisionedIopsCapable": null, - "DBInstance$MultiAZ": null, - "DBInstance$AutoMinorVersionUpgrade": null, - "DBInstance$PubliclyAccessible": null, - "DeleteDBInstanceMessage$SkipFinalSnapshot": null, - "DescribeDBEngineVersionsMessage$DefaultOnly": null, - "DownloadDBLogFilePortionDetails$AdditionalDataPending": null, - "EventSubscription$Enabled": null, - "ModifyDBInstanceMessage$ApplyImmediately": null, - "ModifyDBInstanceMessage$AllowMajorVersionUpgrade": null, - "ModifyOptionGroupMessage$ApplyImmediately": null, - "Option$Persistent": null, - "OptionGroup$AllowsVpcAndNonVpcInstanceMemberships": null, - "OptionGroupOption$PortRequired": null, - "OptionGroupOption$Persistent": null, - "OptionGroupOptionSetting$IsModifiable": null, - "OptionSetting$IsModifiable": null, - "OptionSetting$IsCollection": null, - "OrderableDBInstanceOption$MultiAZCapable": null, - "OrderableDBInstanceOption$ReadReplicaCapable": null, - "OrderableDBInstanceOption$Vpc": null, - "Parameter$IsModifiable": null, - "ReservedDBInstance$MultiAZ": null, - "ReservedDBInstancesOffering$MultiAZ": null, - "ResetDBParameterGroupMessage$ResetAllParameters": null, - "RestoreDBInstanceToPointInTimeMessage$UseLatestRestorableTime": null - } - }, - "BooleanOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$MultiAZ": null, - "CreateDBInstanceMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceMessage$PubliclyAccessible": null, - "CreateDBInstanceReadReplicaMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceReadReplicaMessage$PubliclyAccessible": null, - "CreateEventSubscriptionMessage$Enabled": null, - "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets": null, - "DescribeOrderableDBInstanceOptionsMessage$Vpc": null, - "DescribeReservedDBInstancesMessage$MultiAZ": null, - "DescribeReservedDBInstancesOfferingsMessage$MultiAZ": null, - "ModifyDBInstanceMessage$MultiAZ": null, - "ModifyDBInstanceMessage$AutoMinorVersionUpgrade": null, - "ModifyEventSubscriptionMessage$Enabled": null, - "PendingModifiedValues$MultiAZ": null, - "RebootDBInstanceMessage$ForceFailover": null, - "RestoreDBInstanceFromDBSnapshotMessage$MultiAZ": null, - "RestoreDBInstanceFromDBSnapshotMessage$PubliclyAccessible": null, - "RestoreDBInstanceFromDBSnapshotMessage$AutoMinorVersionUpgrade": null, - "RestoreDBInstanceToPointInTimeMessage$MultiAZ": null, - "RestoreDBInstanceToPointInTimeMessage$PubliclyAccessible": null, - "RestoreDBInstanceToPointInTimeMessage$AutoMinorVersionUpgrade": null - } - }, - "CharacterSet": { - "base": null, - "refs": { - "DBEngineVersion$DefaultCharacterSet": null, - "SupportedCharacterSetsList$member": null - } - }, - "CopyDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CopyDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceResult": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupResult": { - "base": null, - "refs": { - } - }, - "DBEngineVersion": { - "base": null, - "refs": { - "DBEngineVersionList$member": null - } - }, - "DBEngineVersionList": { - "base": null, - "refs": { - "DBEngineVersionMessage$DBEngineVersions": null - } - }, - "DBEngineVersionMessage": { - "base": null, - "refs": { - } - }, - "DBInstance": { - "base": null, - "refs": { - "CreateDBInstanceReadReplicaResult$DBInstance": null, - "CreateDBInstanceResult$DBInstance": null, - "DBInstanceList$member": null, - "DeleteDBInstanceResult$DBInstance": null, - "ModifyDBInstanceResult$DBInstance": null, - "PromoteReadReplicaResult$DBInstance": null, - "RebootDBInstanceResult$DBInstance": null, - "RestoreDBInstanceFromDBSnapshotResult$DBInstance": null, - "RestoreDBInstanceToPointInTimeResult$DBInstance": null - } - }, - "DBInstanceAlreadyExistsFault": { - "base": "

The user already has a DB instance with the given identifier.

", - "refs": { - } - }, - "DBInstanceList": { - "base": null, - "refs": { - "DBInstanceMessage$DBInstances": null - } - }, - "DBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DBInstanceNotFoundFault": { - "base": "

DBInstanceIdentifier doesn't refer to an existing DB instance.

", - "refs": { - } - }, - "DBLogFileNotFoundFault": { - "base": "

LogFileName doesn't refer to an existing DB log file.

", - "refs": { - } - }, - "DBParameterGroup": { - "base": null, - "refs": { - "CreateDBParameterGroupResult$DBParameterGroup": null, - "DBParameterGroupList$member": null - } - }, - "DBParameterGroupAlreadyExistsFault": { - "base": "

A DB parameter group with the same name exists.

", - "refs": { - } - }, - "DBParameterGroupDetails": { - "base": null, - "refs": { - } - }, - "DBParameterGroupList": { - "base": null, - "refs": { - "DBParameterGroupsMessage$DBParameterGroups": null - } - }, - "DBParameterGroupNameMessage": { - "base": null, - "refs": { - } - }, - "DBParameterGroupNotFoundFault": { - "base": "

DBParameterGroupName doesn't refer to an existing DB parameter group.

", - "refs": { - } - }, - "DBParameterGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB parameter groups.

", - "refs": { - } - }, - "DBParameterGroupStatus": { - "base": null, - "refs": { - "DBParameterGroupStatusList$member": null - } - }, - "DBParameterGroupStatusList": { - "base": null, - "refs": { - "DBInstance$DBParameterGroups": null - } - }, - "DBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroup": { - "base": null, - "refs": { - "AuthorizeDBSecurityGroupIngressResult$DBSecurityGroup": null, - "CreateDBSecurityGroupResult$DBSecurityGroup": null, - "DBSecurityGroups$member": null, - "RevokeDBSecurityGroupIngressResult$DBSecurityGroup": null - } - }, - "DBSecurityGroupAlreadyExistsFault": { - "base": "

A DB security group with the name specified in DBSecurityGroupName already exists.

", - "refs": { - } - }, - "DBSecurityGroupMembership": { - "base": null, - "refs": { - "DBSecurityGroupMembershipList$member": null - } - }, - "DBSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$DBSecurityGroups": null, - "Option$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroupNameList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$DBSecurityGroups": null, - "ModifyDBInstanceMessage$DBSecurityGroups": null, - "OptionConfiguration$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupNotFoundFault": { - "base": "

DBSecurityGroupName doesn't refer to an existing DB security group.

", - "refs": { - } - }, - "DBSecurityGroupNotSupportedFault": { - "base": "

A DB security group isn't allowed for this action.

", - "refs": { - } - }, - "DBSecurityGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB security groups.

", - "refs": { - } - }, - "DBSecurityGroups": { - "base": null, - "refs": { - "DBSecurityGroupMessage$DBSecurityGroups": null - } - }, - "DBSnapshot": { - "base": null, - "refs": { - "CopyDBSnapshotResult$DBSnapshot": null, - "CreateDBSnapshotResult$DBSnapshot": null, - "DBSnapshotList$member": null, - "DeleteDBSnapshotResult$DBSnapshot": null - } - }, - "DBSnapshotAlreadyExistsFault": { - "base": "

DBSnapshotIdentifier is already used by an existing snapshot.

", - "refs": { - } - }, - "DBSnapshotList": { - "base": null, - "refs": { - "DBSnapshotMessage$DBSnapshots": null - } - }, - "DBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DBSnapshotNotFoundFault": { - "base": "

DBSnapshotIdentifier doesn't refer to an existing DB snapshot.

", - "refs": { - } - }, - "DBSubnetGroup": { - "base": null, - "refs": { - "CreateDBSubnetGroupResult$DBSubnetGroup": null, - "DBInstance$DBSubnetGroup": null, - "DBSubnetGroups$member": null, - "ModifyDBSubnetGroupResult$DBSubnetGroup": null - } - }, - "DBSubnetGroupAlreadyExistsFault": { - "base": "

DBSubnetGroupName is already used by an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupDoesNotCoverEnoughAZs": { - "base": "

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", - "refs": { - } - }, - "DBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSubnetGroupNotFoundFault": { - "base": "

DBSubnetGroupName doesn't refer to an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB subnet groups.

", - "refs": { - } - }, - "DBSubnetGroups": { - "base": null, - "refs": { - "DBSubnetGroupMessage$DBSubnetGroups": null - } - }, - "DBSubnetQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of subnets in a DB subnet groups.

", - "refs": { - } - }, - "DBUpgradeDependencyFailureFault": { - "base": "

The DB upgrade failed because a resource the DB depends on can't be modified.

", - "refs": { - } - }, - "DeleteDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBInstanceResult": { - "base": null, - "refs": { - } - }, - "DeleteDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "DeleteDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "DeleteOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBEngineVersionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBLogFilesDetails": { - "base": null, - "refs": { - "DescribeDBLogFilesList$member": null - } - }, - "DescribeDBLogFilesList": { - "base": null, - "refs": { - "DescribeDBLogFilesResponse$DescribeDBLogFiles": null - } - }, - "DescribeDBLogFilesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBLogFilesResponse": { - "base": null, - "refs": { - } - }, - "DescribeDBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSecurityGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSnapshotsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSubnetGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersResult": { - "base": null, - "refs": { - } - }, - "DescribeEventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesOfferingsMessage": { - "base": null, - "refs": { - } - }, - "Double": { - "base": null, - "refs": { - "RecurringCharge$RecurringChargeAmount": null, - "ReservedDBInstance$FixedPrice": null, - "ReservedDBInstance$UsagePrice": null, - "ReservedDBInstancesOffering$FixedPrice": null, - "ReservedDBInstancesOffering$UsagePrice": null - } - }, - "DownloadDBLogFilePortionDetails": { - "base": null, - "refs": { - } - }, - "DownloadDBLogFilePortionMessage": { - "base": null, - "refs": { - } - }, - "EC2SecurityGroup": { - "base": null, - "refs": { - "EC2SecurityGroupList$member": null - } - }, - "EC2SecurityGroupList": { - "base": null, - "refs": { - "DBSecurityGroup$EC2SecurityGroups": null - } - }, - "Endpoint": { - "base": null, - "refs": { - "DBInstance$Endpoint": null - } - }, - "EngineDefaults": { - "base": null, - "refs": { - "DescribeEngineDefaultParametersResult$EngineDefaults": null - } - }, - "Event": { - "base": null, - "refs": { - "EventList$member": null - } - }, - "EventCategoriesList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$EventCategories": null, - "DescribeEventsMessage$EventCategories": null, - "Event$EventCategories": null, - "EventCategoriesMap$EventCategories": null, - "EventSubscription$EventCategoriesList": null, - "ModifyEventSubscriptionMessage$EventCategories": null - } - }, - "EventCategoriesMap": { - "base": null, - "refs": { - "EventCategoriesMapList$member": null - } - }, - "EventCategoriesMapList": { - "base": null, - "refs": { - "EventCategoriesMessage$EventCategoriesMapList": null - } - }, - "EventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "EventList": { - "base": null, - "refs": { - "EventsMessage$Events": null - } - }, - "EventSubscription": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionResult$EventSubscription": null, - "CreateEventSubscriptionResult$EventSubscription": null, - "DeleteEventSubscriptionResult$EventSubscription": null, - "EventSubscriptionsList$member": null, - "ModifyEventSubscriptionResult$EventSubscription": null, - "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription": null - } - }, - "EventSubscriptionQuotaExceededFault": { - "base": "

You have reached the maximum number of event subscriptions.

", - "refs": { - } - }, - "EventSubscriptionsList": { - "base": null, - "refs": { - "EventSubscriptionsMessage$EventSubscriptionsList": null - } - }, - "EventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "EventsMessage": { - "base": null, - "refs": { - } - }, - "IPRange": { - "base": null, - "refs": { - "IPRangeList$member": null - } - }, - "IPRangeList": { - "base": null, - "refs": { - "DBSecurityGroup$IPRanges": null - } - }, - "InstanceQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB instances.

", - "refs": { - } - }, - "InsufficientDBInstanceCapacityFault": { - "base": "

The specified DB instance class isn't available in the specified Availability Zone.

", - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "DBInstance$AllocatedStorage": null, - "DBInstance$BackupRetentionPeriod": null, - "DBSnapshot$AllocatedStorage": null, - "DBSnapshot$Port": null, - "DownloadDBLogFilePortionMessage$NumberOfLines": null, - "Endpoint$Port": null, - "ReservedDBInstance$Duration": null, - "ReservedDBInstance$DBInstanceCount": null, - "ReservedDBInstancesOffering$Duration": null - } - }, - "IntegerOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$AllocatedStorage": null, - "CreateDBInstanceMessage$BackupRetentionPeriod": null, - "CreateDBInstanceMessage$Port": null, - "CreateDBInstanceMessage$Iops": null, - "CreateDBInstanceReadReplicaMessage$Port": null, - "CreateDBInstanceReadReplicaMessage$Iops": null, - "DBInstance$Iops": null, - "DBSnapshot$Iops": null, - "DescribeDBEngineVersionsMessage$MaxRecords": null, - "DescribeDBInstancesMessage$MaxRecords": null, - "DescribeDBLogFilesMessage$MaxRecords": null, - "DescribeDBParameterGroupsMessage$MaxRecords": null, - "DescribeDBParametersMessage$MaxRecords": null, - "DescribeDBSecurityGroupsMessage$MaxRecords": null, - "DescribeDBSnapshotsMessage$MaxRecords": null, - "DescribeDBSubnetGroupsMessage$MaxRecords": null, - "DescribeEngineDefaultParametersMessage$MaxRecords": null, - "DescribeEventSubscriptionsMessage$MaxRecords": null, - "DescribeEventsMessage$Duration": null, - "DescribeEventsMessage$MaxRecords": null, - "DescribeOptionGroupOptionsMessage$MaxRecords": null, - "DescribeOptionGroupsMessage$MaxRecords": null, - "DescribeOrderableDBInstanceOptionsMessage$MaxRecords": null, - "DescribeReservedDBInstancesMessage$MaxRecords": null, - "DescribeReservedDBInstancesOfferingsMessage$MaxRecords": null, - "ModifyDBInstanceMessage$AllocatedStorage": null, - "ModifyDBInstanceMessage$BackupRetentionPeriod": null, - "ModifyDBInstanceMessage$Iops": null, - "Option$Port": null, - "OptionConfiguration$Port": null, - "OptionGroupOption$DefaultPort": null, - "PendingModifiedValues$AllocatedStorage": null, - "PendingModifiedValues$Port": null, - "PendingModifiedValues$BackupRetentionPeriod": null, - "PendingModifiedValues$Iops": null, - "PromoteReadReplicaMessage$BackupRetentionPeriod": null, - "PurchaseReservedDBInstancesOfferingMessage$DBInstanceCount": null, - "RestoreDBInstanceFromDBSnapshotMessage$Port": null, - "RestoreDBInstanceFromDBSnapshotMessage$Iops": null, - "RestoreDBInstanceToPointInTimeMessage$Port": null, - "RestoreDBInstanceToPointInTimeMessage$Iops": null - } - }, - "InvalidDBInstanceStateFault": { - "base": "

The specified DB instance isn't in the available state.

", - "refs": { - } - }, - "InvalidDBParameterGroupStateFault": { - "base": "

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you can't delete it when the parameter group is in this state.

", - "refs": { - } - }, - "InvalidDBSecurityGroupStateFault": { - "base": "

The state of the DB security group doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSnapshotStateFault": { - "base": "

The state of the DB snapshot doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSubnetGroupStateFault": { - "base": "

The DB subnet group cannot be deleted because it's in use.

", - "refs": { - } - }, - "InvalidDBSubnetStateFault": { - "base": "

The DB subnet isn't in the available state.

", - "refs": { - } - }, - "InvalidEventSubscriptionStateFault": { - "base": "

This error can occur if someone else is modifying a subscription. You should retry the action.

", - "refs": { - } - }, - "InvalidOptionGroupStateFault": { - "base": "

The option group isn't in the available state.

", - "refs": { - } - }, - "InvalidRestoreFault": { - "base": "

Cannot restore from VPC backup to non-VPC DB instance.

", - "refs": { - } - }, - "InvalidSubnet": { - "base": "

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", - "refs": { - } - }, - "InvalidVPCNetworkStateFault": { - "base": "

The DB subnet group doesn't cover all Availability Zones after it's created because of users' change.

", - "refs": { - } - }, - "KeyList": { - "base": null, - "refs": { - "RemoveTagsFromResourceMessage$TagKeys": null - } - }, - "ListTagsForResourceMessage": { - "base": null, - "refs": { - } - }, - "Long": { - "base": null, - "refs": { - "DescribeDBLogFilesDetails$LastWritten": null, - "DescribeDBLogFilesDetails$Size": null, - "DescribeDBLogFilesMessage$FileLastWritten": null, - "DescribeDBLogFilesMessage$FileSize": null - } - }, - "ModifyDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBInstanceResult": { - "base": null, - "refs": { - } - }, - "ModifyDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupResult": { - "base": null, - "refs": { - } - }, - "Option": { - "base": null, - "refs": { - "OptionsList$member": null - } - }, - "OptionConfiguration": { - "base": null, - "refs": { - "OptionConfigurationList$member": null - } - }, - "OptionConfigurationList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToInclude": null - } - }, - "OptionGroup": { - "base": null, - "refs": { - "CreateOptionGroupResult$OptionGroup": null, - "ModifyOptionGroupResult$OptionGroup": null, - "OptionGroupsList$member": null - } - }, - "OptionGroupAlreadyExistsFault": { - "base": "

The option group you are trying to create already exists.

", - "refs": { - } - }, - "OptionGroupMembership": { - "base": null, - "refs": { - "OptionGroupMembershipList$member": null - } - }, - "OptionGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$OptionGroupMemberships": null - } - }, - "OptionGroupNotFoundFault": { - "base": "

The specified option group could not be found.

", - "refs": { - } - }, - "OptionGroupOption": { - "base": null, - "refs": { - "OptionGroupOptionsList$member": null - } - }, - "OptionGroupOptionSetting": { - "base": null, - "refs": { - "OptionGroupOptionSettingsList$member": null - } - }, - "OptionGroupOptionSettingsList": { - "base": null, - "refs": { - "OptionGroupOption$OptionGroupOptionSettings": null - } - }, - "OptionGroupOptionsList": { - "base": null, - "refs": { - "OptionGroupOptionsMessage$OptionGroupOptions": null - } - }, - "OptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "OptionGroupQuotaExceededFault": { - "base": "

The quota of 20 option groups was exceeded for this AWS account.

", - "refs": { - } - }, - "OptionGroups": { - "base": null, - "refs": { - } - }, - "OptionGroupsList": { - "base": null, - "refs": { - "OptionGroups$OptionGroupsList": null - } - }, - "OptionNamesList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToRemove": null - } - }, - "OptionSetting": { - "base": null, - "refs": { - "OptionSettingConfigurationList$member": null, - "OptionSettingsList$member": null - } - }, - "OptionSettingConfigurationList": { - "base": null, - "refs": { - "Option$OptionSettings": null - } - }, - "OptionSettingsList": { - "base": null, - "refs": { - "OptionConfiguration$OptionSettings": null - } - }, - "OptionsDependedOn": { - "base": null, - "refs": { - "OptionGroupOption$OptionsDependedOn": null - } - }, - "OptionsList": { - "base": null, - "refs": { - "OptionGroup$Options": null - } - }, - "OrderableDBInstanceOption": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsList$member": null - } - }, - "OrderableDBInstanceOptionsList": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions": null - } - }, - "OrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "Parameter": { - "base": null, - "refs": { - "ParametersList$member": null - } - }, - "ParametersList": { - "base": null, - "refs": { - "DBParameterGroupDetails$Parameters": null, - "EngineDefaults$Parameters": null, - "ModifyDBParameterGroupMessage$Parameters": null, - "ResetDBParameterGroupMessage$Parameters": null - } - }, - "PendingModifiedValues": { - "base": null, - "refs": { - "DBInstance$PendingModifiedValues": null - } - }, - "PointInTimeRestoreNotEnabledFault": { - "base": "

SourceDBInstanceIdentifier refers to a DB instance with BackupRetentionPeriod equal to 0.

", - "refs": { - } - }, - "PromoteReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "PromoteReadReplicaResult": { - "base": null, - "refs": { - } - }, - "ProvisionedIopsNotAvailableInAZFault": { - "base": "

Provisioned IOPS not available in the specified Availability Zone.

", - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingResult": { - "base": null, - "refs": { - } - }, - "ReadReplicaDBInstanceIdentifierList": { - "base": null, - "refs": { - "DBInstance$ReadReplicaDBInstanceIdentifiers": null - } - }, - "RebootDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "RebootDBInstanceResult": { - "base": null, - "refs": { - } - }, - "RecurringCharge": { - "base": null, - "refs": { - "RecurringChargeList$member": null - } - }, - "RecurringChargeList": { - "base": null, - "refs": { - "ReservedDBInstance$RecurringCharges": null, - "ReservedDBInstancesOffering$RecurringCharges": null - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "RemoveSourceIdentifierFromSubscriptionResult": { - "base": null, - "refs": { - } - }, - "RemoveTagsFromResourceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstance": { - "base": null, - "refs": { - "PurchaseReservedDBInstancesOfferingResult$ReservedDBInstance": null, - "ReservedDBInstanceList$member": null - } - }, - "ReservedDBInstanceAlreadyExistsFault": { - "base": "

User already has a reservation with the given identifier.

", - "refs": { - } - }, - "ReservedDBInstanceList": { - "base": null, - "refs": { - "ReservedDBInstanceMessage$ReservedDBInstances": null - } - }, - "ReservedDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstanceNotFoundFault": { - "base": "

The specified reserved DB Instance not found.

", - "refs": { - } - }, - "ReservedDBInstanceQuotaExceededFault": { - "base": "

Request would exceed the user's DB Instance quota.

", - "refs": { - } - }, - "ReservedDBInstancesOffering": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingList$member": null - } - }, - "ReservedDBInstancesOfferingList": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferings": null - } - }, - "ReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstancesOfferingNotFoundFault": { - "base": "

Specified offering does not exist.

", - "refs": { - } - }, - "ResetDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeResult": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "SNSInvalidTopicFault": { - "base": "

SNS has responded that there is a problem with the SND topic specified.

", - "refs": { - } - }, - "SNSNoAuthorizationFault": { - "base": "

You do not have permission to publish to the SNS topic ARN.

", - "refs": { - } - }, - "SNSTopicArnNotFoundFault": { - "base": "

The SNS topic ARN does not exist.

", - "refs": { - } - }, - "SnapshotQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB snapshots.

", - "refs": { - } - }, - "SourceIdsList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$SourceIds": null, - "EventSubscription$SourceIdsList": null - } - }, - "SourceNotFoundFault": { - "base": "

The requested source could not be found.

", - "refs": { - } - }, - "SourceType": { - "base": null, - "refs": { - "DescribeEventsMessage$SourceType": null, - "Event$SourceType": null - } - }, - "StorageQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed amount of storage available across all DB instances.

", - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionMessage$SubscriptionName": null, - "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier": null, - "AddTagsToResourceMessage$ResourceName": null, - "AuthorizeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$CIDRIP": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "AvailabilityZone$Name": null, - "CharacterSet$CharacterSetName": null, - "CharacterSet$CharacterSetDescription": null, - "CopyDBSnapshotMessage$SourceDBSnapshotIdentifier": null, - "CopyDBSnapshotMessage$TargetDBSnapshotIdentifier": null, - "CreateDBInstanceMessage$DBName": null, - "CreateDBInstanceMessage$DBInstanceIdentifier": null, - "CreateDBInstanceMessage$DBInstanceClass": null, - "CreateDBInstanceMessage$Engine": null, - "CreateDBInstanceMessage$MasterUsername": null, - "CreateDBInstanceMessage$MasterUserPassword": null, - "CreateDBInstanceMessage$AvailabilityZone": null, - "CreateDBInstanceMessage$DBSubnetGroupName": null, - "CreateDBInstanceMessage$PreferredMaintenanceWindow": null, - "CreateDBInstanceMessage$DBParameterGroupName": null, - "CreateDBInstanceMessage$PreferredBackupWindow": null, - "CreateDBInstanceMessage$EngineVersion": null, - "CreateDBInstanceMessage$LicenseModel": null, - "CreateDBInstanceMessage$OptionGroupName": null, - "CreateDBInstanceMessage$CharacterSetName": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$SourceDBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceClass": null, - "CreateDBInstanceReadReplicaMessage$AvailabilityZone": null, - "CreateDBInstanceReadReplicaMessage$OptionGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupFamily": null, - "CreateDBParameterGroupMessage$Description": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupName": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupDescription": null, - "CreateDBSnapshotMessage$DBSnapshotIdentifier": null, - "CreateDBSnapshotMessage$DBInstanceIdentifier": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupName": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "CreateEventSubscriptionMessage$SubscriptionName": null, - "CreateEventSubscriptionMessage$SnsTopicArn": null, - "CreateEventSubscriptionMessage$SourceType": null, - "CreateOptionGroupMessage$OptionGroupName": null, - "CreateOptionGroupMessage$EngineName": null, - "CreateOptionGroupMessage$MajorEngineVersion": null, - "CreateOptionGroupMessage$OptionGroupDescription": null, - "DBEngineVersion$Engine": null, - "DBEngineVersion$EngineVersion": null, - "DBEngineVersion$DBParameterGroupFamily": null, - "DBEngineVersion$DBEngineDescription": null, - "DBEngineVersion$DBEngineVersionDescription": null, - "DBEngineVersionMessage$Marker": null, - "DBInstance$DBInstanceIdentifier": null, - "DBInstance$DBInstanceClass": null, - "DBInstance$Engine": null, - "DBInstance$DBInstanceStatus": null, - "DBInstance$MasterUsername": null, - "DBInstance$DBName": null, - "DBInstance$PreferredBackupWindow": null, - "DBInstance$AvailabilityZone": null, - "DBInstance$PreferredMaintenanceWindow": null, - "DBInstance$EngineVersion": null, - "DBInstance$ReadReplicaSourceDBInstanceIdentifier": null, - "DBInstance$LicenseModel": null, - "DBInstance$CharacterSetName": null, - "DBInstance$SecondaryAvailabilityZone": null, - "DBInstanceMessage$Marker": null, - "DBParameterGroup$DBParameterGroupName": null, - "DBParameterGroup$DBParameterGroupFamily": null, - "DBParameterGroup$Description": null, - "DBParameterGroupDetails$Marker": null, - "DBParameterGroupNameMessage$DBParameterGroupName": null, - "DBParameterGroupStatus$DBParameterGroupName": null, - "DBParameterGroupStatus$ParameterApplyStatus": null, - "DBParameterGroupsMessage$Marker": null, - "DBSecurityGroup$OwnerId": null, - "DBSecurityGroup$DBSecurityGroupName": null, - "DBSecurityGroup$DBSecurityGroupDescription": null, - "DBSecurityGroup$VpcId": null, - "DBSecurityGroupMembership$DBSecurityGroupName": null, - "DBSecurityGroupMembership$Status": null, - "DBSecurityGroupMessage$Marker": null, - "DBSecurityGroupNameList$member": null, - "DBSnapshot$DBSnapshotIdentifier": null, - "DBSnapshot$DBInstanceIdentifier": null, - "DBSnapshot$Engine": null, - "DBSnapshot$Status": null, - "DBSnapshot$AvailabilityZone": null, - "DBSnapshot$VpcId": null, - "DBSnapshot$MasterUsername": null, - "DBSnapshot$EngineVersion": null, - "DBSnapshot$LicenseModel": null, - "DBSnapshot$SnapshotType": null, - "DBSnapshot$OptionGroupName": null, - "DBSnapshotMessage$Marker": null, - "DBSubnetGroup$DBSubnetGroupName": null, - "DBSubnetGroup$DBSubnetGroupDescription": null, - "DBSubnetGroup$VpcId": null, - "DBSubnetGroup$SubnetGroupStatus": null, - "DBSubnetGroupMessage$Marker": null, - "DeleteDBInstanceMessage$DBInstanceIdentifier": null, - "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier": null, - "DeleteDBParameterGroupMessage$DBParameterGroupName": null, - "DeleteDBSecurityGroupMessage$DBSecurityGroupName": null, - "DeleteDBSnapshotMessage$DBSnapshotIdentifier": null, - "DeleteDBSubnetGroupMessage$DBSubnetGroupName": null, - "DeleteEventSubscriptionMessage$SubscriptionName": null, - "DeleteOptionGroupMessage$OptionGroupName": null, - "DescribeDBEngineVersionsMessage$Engine": null, - "DescribeDBEngineVersionsMessage$EngineVersion": null, - "DescribeDBEngineVersionsMessage$DBParameterGroupFamily": null, - "DescribeDBEngineVersionsMessage$Marker": null, - "DescribeDBInstancesMessage$DBInstanceIdentifier": null, - "DescribeDBInstancesMessage$Marker": null, - "DescribeDBLogFilesDetails$LogFileName": null, - "DescribeDBLogFilesMessage$DBInstanceIdentifier": null, - "DescribeDBLogFilesMessage$FilenameContains": null, - "DescribeDBLogFilesMessage$Marker": null, - "DescribeDBLogFilesResponse$Marker": null, - "DescribeDBParameterGroupsMessage$DBParameterGroupName": null, - "DescribeDBParameterGroupsMessage$Marker": null, - "DescribeDBParametersMessage$DBParameterGroupName": null, - "DescribeDBParametersMessage$Source": null, - "DescribeDBParametersMessage$Marker": null, - "DescribeDBSecurityGroupsMessage$DBSecurityGroupName": null, - "DescribeDBSecurityGroupsMessage$Marker": null, - "DescribeDBSnapshotsMessage$DBInstanceIdentifier": null, - "DescribeDBSnapshotsMessage$DBSnapshotIdentifier": null, - "DescribeDBSnapshotsMessage$SnapshotType": null, - "DescribeDBSnapshotsMessage$Marker": null, - "DescribeDBSubnetGroupsMessage$DBSubnetGroupName": null, - "DescribeDBSubnetGroupsMessage$Marker": null, - "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily": null, - "DescribeEngineDefaultParametersMessage$Marker": null, - "DescribeEventCategoriesMessage$SourceType": null, - "DescribeEventSubscriptionsMessage$SubscriptionName": null, - "DescribeEventSubscriptionsMessage$Marker": null, - "DescribeEventsMessage$SourceIdentifier": null, - "DescribeEventsMessage$Marker": null, - "DescribeOptionGroupOptionsMessage$EngineName": null, - "DescribeOptionGroupOptionsMessage$MajorEngineVersion": null, - "DescribeOptionGroupOptionsMessage$Marker": null, - "DescribeOptionGroupsMessage$OptionGroupName": null, - "DescribeOptionGroupsMessage$Marker": null, - "DescribeOptionGroupsMessage$EngineName": null, - "DescribeOptionGroupsMessage$MajorEngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$Engine": null, - "DescribeOrderableDBInstanceOptionsMessage$EngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass": null, - "DescribeOrderableDBInstanceOptionsMessage$LicenseModel": null, - "DescribeOrderableDBInstanceOptionsMessage$Marker": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstanceId": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesMessage$Duration": null, - "DescribeReservedDBInstancesMessage$ProductDescription": null, - "DescribeReservedDBInstancesMessage$OfferingType": null, - "DescribeReservedDBInstancesMessage$Marker": null, - "DescribeReservedDBInstancesOfferingsMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesOfferingsMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesOfferingsMessage$Duration": null, - "DescribeReservedDBInstancesOfferingsMessage$ProductDescription": null, - "DescribeReservedDBInstancesOfferingsMessage$OfferingType": null, - "DescribeReservedDBInstancesOfferingsMessage$Marker": null, - "DownloadDBLogFilePortionDetails$LogFileData": null, - "DownloadDBLogFilePortionDetails$Marker": null, - "DownloadDBLogFilePortionMessage$DBInstanceIdentifier": null, - "DownloadDBLogFilePortionMessage$LogFileName": null, - "DownloadDBLogFilePortionMessage$Marker": null, - "EC2SecurityGroup$Status": null, - "EC2SecurityGroup$EC2SecurityGroupName": null, - "EC2SecurityGroup$EC2SecurityGroupId": null, - "EC2SecurityGroup$EC2SecurityGroupOwnerId": null, - "Endpoint$Address": null, - "EngineDefaults$DBParameterGroupFamily": null, - "EngineDefaults$Marker": null, - "Event$SourceIdentifier": null, - "Event$Message": null, - "EventCategoriesList$member": null, - "EventCategoriesMap$SourceType": null, - "EventSubscription$CustomerAwsId": null, - "EventSubscription$CustSubscriptionId": null, - "EventSubscription$SnsTopicArn": null, - "EventSubscription$Status": null, - "EventSubscription$SubscriptionCreationTime": null, - "EventSubscription$SourceType": null, - "EventSubscriptionsMessage$Marker": null, - "EventsMessage$Marker": null, - "IPRange$Status": null, - "IPRange$CIDRIP": null, - "KeyList$member": null, - "ListTagsForResourceMessage$ResourceName": null, - "ModifyDBInstanceMessage$DBInstanceIdentifier": null, - "ModifyDBInstanceMessage$DBInstanceClass": null, - "ModifyDBInstanceMessage$MasterUserPassword": null, - "ModifyDBInstanceMessage$DBParameterGroupName": null, - "ModifyDBInstanceMessage$PreferredBackupWindow": null, - "ModifyDBInstanceMessage$PreferredMaintenanceWindow": null, - "ModifyDBInstanceMessage$EngineVersion": null, - "ModifyDBInstanceMessage$OptionGroupName": null, - "ModifyDBInstanceMessage$NewDBInstanceIdentifier": null, - "ModifyDBParameterGroupMessage$DBParameterGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "ModifyEventSubscriptionMessage$SubscriptionName": null, - "ModifyEventSubscriptionMessage$SnsTopicArn": null, - "ModifyEventSubscriptionMessage$SourceType": null, - "ModifyOptionGroupMessage$OptionGroupName": null, - "Option$OptionName": null, - "Option$OptionDescription": null, - "OptionConfiguration$OptionName": null, - "OptionGroup$OptionGroupName": null, - "OptionGroup$OptionGroupDescription": null, - "OptionGroup$EngineName": null, - "OptionGroup$MajorEngineVersion": null, - "OptionGroup$VpcId": null, - "OptionGroupMembership$OptionGroupName": null, - "OptionGroupMembership$Status": null, - "OptionGroupOption$Name": null, - "OptionGroupOption$Description": null, - "OptionGroupOption$EngineName": null, - "OptionGroupOption$MajorEngineVersion": null, - "OptionGroupOption$MinimumRequiredMinorEngineVersion": null, - "OptionGroupOptionSetting$SettingName": null, - "OptionGroupOptionSetting$SettingDescription": null, - "OptionGroupOptionSetting$DefaultValue": null, - "OptionGroupOptionSetting$ApplyType": null, - "OptionGroupOptionSetting$AllowedValues": null, - "OptionGroupOptionsMessage$Marker": null, - "OptionGroups$Marker": null, - "OptionNamesList$member": null, - "OptionSetting$Name": null, - "OptionSetting$Value": null, - "OptionSetting$DefaultValue": null, - "OptionSetting$Description": null, - "OptionSetting$ApplyType": null, - "OptionSetting$DataType": null, - "OptionSetting$AllowedValues": null, - "OptionsDependedOn$member": null, - "OrderableDBInstanceOption$Engine": null, - "OrderableDBInstanceOption$EngineVersion": null, - "OrderableDBInstanceOption$DBInstanceClass": null, - "OrderableDBInstanceOption$LicenseModel": null, - "OrderableDBInstanceOptionsMessage$Marker": null, - "Parameter$ParameterName": null, - "Parameter$ParameterValue": null, - "Parameter$Description": null, - "Parameter$Source": null, - "Parameter$ApplyType": null, - "Parameter$DataType": null, - "Parameter$AllowedValues": null, - "Parameter$MinimumEngineVersion": null, - "PendingModifiedValues$DBInstanceClass": null, - "PendingModifiedValues$MasterUserPassword": null, - "PendingModifiedValues$EngineVersion": null, - "PendingModifiedValues$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$PreferredBackupWindow": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferingId": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstanceId": null, - "ReadReplicaDBInstanceIdentifierList$member": null, - "RebootDBInstanceMessage$DBInstanceIdentifier": null, - "RecurringCharge$RecurringChargeFrequency": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier": null, - "RemoveTagsFromResourceMessage$ResourceName": null, - "ReservedDBInstance$ReservedDBInstanceId": null, - "ReservedDBInstance$ReservedDBInstancesOfferingId": null, - "ReservedDBInstance$DBInstanceClass": null, - "ReservedDBInstance$CurrencyCode": null, - "ReservedDBInstance$ProductDescription": null, - "ReservedDBInstance$OfferingType": null, - "ReservedDBInstance$State": null, - "ReservedDBInstanceMessage$Marker": null, - "ReservedDBInstancesOffering$ReservedDBInstancesOfferingId": null, - "ReservedDBInstancesOffering$DBInstanceClass": null, - "ReservedDBInstancesOffering$CurrencyCode": null, - "ReservedDBInstancesOffering$ProductDescription": null, - "ReservedDBInstancesOffering$OfferingType": null, - "ReservedDBInstancesOfferingMessage$Marker": null, - "ResetDBParameterGroupMessage$DBParameterGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSnapshotIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceClass": null, - "RestoreDBInstanceFromDBSnapshotMessage$AvailabilityZone": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSubnetGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$LicenseModel": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBName": null, - "RestoreDBInstanceFromDBSnapshotMessage$Engine": null, - "RestoreDBInstanceFromDBSnapshotMessage$OptionGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$SourceDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$TargetDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$DBInstanceClass": null, - "RestoreDBInstanceToPointInTimeMessage$AvailabilityZone": null, - "RestoreDBInstanceToPointInTimeMessage$DBSubnetGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$LicenseModel": null, - "RestoreDBInstanceToPointInTimeMessage$DBName": null, - "RestoreDBInstanceToPointInTimeMessage$Engine": null, - "RestoreDBInstanceToPointInTimeMessage$OptionGroupName": null, - "RevokeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$CIDRIP": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "SourceIdsList$member": null, - "Subnet$SubnetIdentifier": null, - "Subnet$SubnetStatus": null, - "SubnetIdentifierList$member": null, - "Tag$Key": null, - "Tag$Value": null, - "VpcSecurityGroupIdList$member": null, - "VpcSecurityGroupMembership$VpcSecurityGroupId": null, - "VpcSecurityGroupMembership$Status": null - } - }, - "Subnet": { - "base": null, - "refs": { - "SubnetList$member": null - } - }, - "SubnetAlreadyInUse": { - "base": "

The DB subnet is already in use in the Availability Zone.

", - "refs": { - } - }, - "SubnetIdentifierList": { - "base": null, - "refs": { - "CreateDBSubnetGroupMessage$SubnetIds": null, - "ModifyDBSubnetGroupMessage$SubnetIds": null - } - }, - "SubnetList": { - "base": null, - "refs": { - "DBSubnetGroup$Subnets": null - } - }, - "SubscriptionAlreadyExistFault": { - "base": "

The supplied subscription name already exists.

", - "refs": { - } - }, - "SubscriptionCategoryNotFoundFault": { - "base": "

The supplied category does not exist.

", - "refs": { - } - }, - "SubscriptionNotFoundFault": { - "base": "

The subscription name does not exist.

", - "refs": { - } - }, - "SupportedCharacterSetsList": { - "base": null, - "refs": { - "DBEngineVersion$SupportedCharacterSets": null - } - }, - "TStamp": { - "base": null, - "refs": { - "DBInstance$InstanceCreateTime": null, - "DBInstance$LatestRestorableTime": null, - "DBSnapshot$SnapshotCreateTime": null, - "DBSnapshot$InstanceCreateTime": null, - "DescribeEventsMessage$StartTime": null, - "DescribeEventsMessage$EndTime": null, - "Event$Date": null, - "ReservedDBInstance$StartTime": null, - "RestoreDBInstanceToPointInTimeMessage$RestoreTime": null - } - }, - "Tag": { - "base": null, - "refs": { - "TagList$member": null - } - }, - "TagList": { - "base": null, - "refs": { - "AddTagsToResourceMessage$Tags": null, - "TagListMessage$TagList": null - } - }, - "TagListMessage": { - "base": null, - "refs": { - } - }, - "VpcSecurityGroupIdList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$VpcSecurityGroupIds": null, - "ModifyDBInstanceMessage$VpcSecurityGroupIds": null, - "OptionConfiguration$VpcSecurityGroupMemberships": null - } - }, - "VpcSecurityGroupMembership": { - "base": null, - "refs": { - "VpcSecurityGroupMembershipList$member": null - } - }, - "VpcSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$VpcSecurityGroups": null, - "Option$VpcSecurityGroupMemberships": null - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/paginators-1.json deleted file mode 100644 index c51d8d15b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-02-12/paginators-1.json +++ /dev/null @@ -1,110 +0,0 @@ -{ - "pagination": { - "DescribeDBEngineVersions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBEngineVersions" - }, - "DescribeDBInstances": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBInstances" - }, - "DescribeDBLogFiles": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DescribeDBLogFiles" - }, - "DescribeDBParameterGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBParameterGroups" - }, - "DescribeDBParameters": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "Parameters" - }, - "DescribeDBSecurityGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSecurityGroups" - }, - "DescribeDBSnapshots": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSnapshots" - }, - "DescribeDBSubnetGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSubnetGroups" - }, - "DescribeEngineDefaultParameters": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "EngineDefaults.Marker", - "result_key": "EngineDefaults.Parameters" - }, - "DescribeEventSubscriptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "EventSubscriptionsList" - }, - "DescribeEvents": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "Events" - }, - "DescribeOptionGroupOptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OptionGroupOptions" - }, - "DescribeOptionGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OptionGroupsList" - }, - "DescribeOrderableDBInstanceOptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OrderableDBInstanceOptions" - }, - "DescribeReservedDBInstances": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "ReservedDBInstances" - }, - "DescribeReservedDBInstancesOfferings": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "ReservedDBInstancesOfferings" - }, - "DownloadDBLogFilePortion": { - "input_token": "Marker", - "limit_key": "NumberOfLines", - "more_results": "AdditionalDataPending", - "output_token": "Marker", - "result_key": "LogFileData" - }, - "ListTagsForResource": { - "result_key": "TagList" - } - } -} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/api-2.json deleted file mode 100644 index c7e975155..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/api-2.json +++ /dev/null @@ -1,3160 +0,0 @@ -{ - "version":"2.0", - "metadata":{ - "apiVersion":"2013-09-09", - "endpointPrefix":"rds", - "protocol":"query", - "serviceAbbreviation":"Amazon RDS", - "serviceFullName":"Amazon Relational Database Service", - "serviceId":"RDS", - "signatureVersion":"v4", - "uid":"rds-2013-09-09", - "xmlNamespace":"http://rds.amazonaws.com/doc/2013-09-09/" - }, - "operations":{ - "AddSourceIdentifierToSubscription":{ - "name":"AddSourceIdentifierToSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddSourceIdentifierToSubscriptionMessage"}, - "output":{ - "shape":"AddSourceIdentifierToSubscriptionResult", - "resultWrapper":"AddSourceIdentifierToSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "AddTagsToResource":{ - "name":"AddTagsToResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AddTagsToResourceMessage"}, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "AuthorizeDBSecurityGroupIngress":{ - "name":"AuthorizeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"AuthorizeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"AuthorizeDBSecurityGroupIngressResult", - "resultWrapper":"AuthorizeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"AuthorizationAlreadyExistsFault"}, - {"shape":"AuthorizationQuotaExceededFault"} - ] - }, - "CopyDBSnapshot":{ - "name":"CopyDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CopyDBSnapshotMessage"}, - "output":{ - "shape":"CopyDBSnapshotResult", - "resultWrapper":"CopyDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "CreateDBInstance":{ - "name":"CreateDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceMessage"}, - "output":{ - "shape":"CreateDBInstanceResult", - "resultWrapper":"CreateDBInstanceResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "CreateDBInstanceReadReplica":{ - "name":"CreateDBInstanceReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBInstanceReadReplicaMessage"}, - "output":{ - "shape":"CreateDBInstanceReadReplicaResult", - "resultWrapper":"CreateDBInstanceReadReplicaResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"DBSubnetGroupNotAllowedFault"}, - {"shape":"InvalidDBSubnetGroupFault"} - ] - }, - "CreateDBParameterGroup":{ - "name":"CreateDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBParameterGroupMessage"}, - "output":{ - "shape":"CreateDBParameterGroupResult", - "resultWrapper":"CreateDBParameterGroupResult" - }, - "errors":[ - {"shape":"DBParameterGroupQuotaExceededFault"}, - {"shape":"DBParameterGroupAlreadyExistsFault"} - ] - }, - "CreateDBSecurityGroup":{ - "name":"CreateDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSecurityGroupMessage"}, - "output":{ - "shape":"CreateDBSecurityGroupResult", - "resultWrapper":"CreateDBSecurityGroupResult" - }, - "errors":[ - {"shape":"DBSecurityGroupAlreadyExistsFault"}, - {"shape":"DBSecurityGroupQuotaExceededFault"}, - {"shape":"DBSecurityGroupNotSupportedFault"} - ] - }, - "CreateDBSnapshot":{ - "name":"CreateDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSnapshotMessage"}, - "output":{ - "shape":"CreateDBSnapshotResult", - "resultWrapper":"CreateDBSnapshotResult" - }, - "errors":[ - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "CreateDBSubnetGroup":{ - "name":"CreateDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateDBSubnetGroupMessage"}, - "output":{ - "shape":"CreateDBSubnetGroupResult", - "resultWrapper":"CreateDBSubnetGroupResult" - }, - "errors":[ - {"shape":"DBSubnetGroupAlreadyExistsFault"}, - {"shape":"DBSubnetGroupQuotaExceededFault"}, - {"shape":"DBSubnetQuotaExceededFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"} - ] - }, - "CreateEventSubscription":{ - "name":"CreateEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateEventSubscriptionMessage"}, - "output":{ - "shape":"CreateEventSubscriptionResult", - "resultWrapper":"CreateEventSubscriptionResult" - }, - "errors":[ - {"shape":"EventSubscriptionQuotaExceededFault"}, - {"shape":"SubscriptionAlreadyExistFault"}, - {"shape":"SNSInvalidTopicFault"}, - {"shape":"SNSNoAuthorizationFault"}, - {"shape":"SNSTopicArnNotFoundFault"}, - {"shape":"SubscriptionCategoryNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "CreateOptionGroup":{ - "name":"CreateOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"CreateOptionGroupMessage"}, - "output":{ - "shape":"CreateOptionGroupResult", - "resultWrapper":"CreateOptionGroupResult" - }, - "errors":[ - {"shape":"OptionGroupAlreadyExistsFault"}, - {"shape":"OptionGroupQuotaExceededFault"} - ] - }, - "DeleteDBInstance":{ - "name":"DeleteDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBInstanceMessage"}, - "output":{ - "shape":"DeleteDBInstanceResult", - "resultWrapper":"DeleteDBInstanceResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBSnapshotAlreadyExistsFault"}, - {"shape":"SnapshotQuotaExceededFault"} - ] - }, - "DeleteDBParameterGroup":{ - "name":"DeleteDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBParameterGroupMessage"}, - "errors":[ - {"shape":"InvalidDBParameterGroupStateFault"}, - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DeleteDBSecurityGroup":{ - "name":"DeleteDBSecurityGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSecurityGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DeleteDBSnapshot":{ - "name":"DeleteDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSnapshotMessage"}, - "output":{ - "shape":"DeleteDBSnapshotResult", - "resultWrapper":"DeleteDBSnapshotResult" - }, - "errors":[ - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "DeleteDBSubnetGroup":{ - "name":"DeleteDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteDBSubnetGroupMessage"}, - "errors":[ - {"shape":"InvalidDBSubnetGroupStateFault"}, - {"shape":"InvalidDBSubnetStateFault"}, - {"shape":"DBSubnetGroupNotFoundFault"} - ] - }, - "DeleteEventSubscription":{ - "name":"DeleteEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteEventSubscriptionMessage"}, - "output":{ - "shape":"DeleteEventSubscriptionResult", - "resultWrapper":"DeleteEventSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"InvalidEventSubscriptionStateFault"} - ] - }, - "DeleteOptionGroup":{ - "name":"DeleteOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DeleteOptionGroupMessage"}, - "errors":[ - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"InvalidOptionGroupStateFault"} - ] - }, - "DescribeDBEngineVersions":{ - "name":"DescribeDBEngineVersions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBEngineVersionsMessage"}, - "output":{ - "shape":"DBEngineVersionMessage", - "resultWrapper":"DescribeDBEngineVersionsResult" - } - }, - "DescribeDBInstances":{ - "name":"DescribeDBInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBInstancesMessage"}, - "output":{ - "shape":"DBInstanceMessage", - "resultWrapper":"DescribeDBInstancesResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "DescribeDBLogFiles":{ - "name":"DescribeDBLogFiles", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBLogFilesMessage"}, - "output":{ - "shape":"DescribeDBLogFilesResponse", - "resultWrapper":"DescribeDBLogFilesResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "DescribeDBParameterGroups":{ - "name":"DescribeDBParameterGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBParameterGroupsMessage"}, - "output":{ - "shape":"DBParameterGroupsMessage", - "resultWrapper":"DescribeDBParameterGroupsResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DescribeDBParameters":{ - "name":"DescribeDBParameters", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBParametersMessage"}, - "output":{ - "shape":"DBParameterGroupDetails", - "resultWrapper":"DescribeDBParametersResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "DescribeDBSecurityGroups":{ - "name":"DescribeDBSecurityGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSecurityGroupsMessage"}, - "output":{ - "shape":"DBSecurityGroupMessage", - "resultWrapper":"DescribeDBSecurityGroupsResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"} - ] - }, - "DescribeDBSnapshots":{ - "name":"DescribeDBSnapshots", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSnapshotsMessage"}, - "output":{ - "shape":"DBSnapshotMessage", - "resultWrapper":"DescribeDBSnapshotsResult" - }, - "errors":[ - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "DescribeDBSubnetGroups":{ - "name":"DescribeDBSubnetGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeDBSubnetGroupsMessage"}, - "output":{ - "shape":"DBSubnetGroupMessage", - "resultWrapper":"DescribeDBSubnetGroupsResult" - }, - "errors":[ - {"shape":"DBSubnetGroupNotFoundFault"} - ] - }, - "DescribeEngineDefaultParameters":{ - "name":"DescribeEngineDefaultParameters", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEngineDefaultParametersMessage"}, - "output":{ - "shape":"DescribeEngineDefaultParametersResult", - "resultWrapper":"DescribeEngineDefaultParametersResult" - } - }, - "DescribeEventCategories":{ - "name":"DescribeEventCategories", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventCategoriesMessage"}, - "output":{ - "shape":"EventCategoriesMessage", - "resultWrapper":"DescribeEventCategoriesResult" - } - }, - "DescribeEventSubscriptions":{ - "name":"DescribeEventSubscriptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventSubscriptionsMessage"}, - "output":{ - "shape":"EventSubscriptionsMessage", - "resultWrapper":"DescribeEventSubscriptionsResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"} - ] - }, - "DescribeEvents":{ - "name":"DescribeEvents", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeEventsMessage"}, - "output":{ - "shape":"EventsMessage", - "resultWrapper":"DescribeEventsResult" - } - }, - "DescribeOptionGroupOptions":{ - "name":"DescribeOptionGroupOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupOptionsMessage"}, - "output":{ - "shape":"OptionGroupOptionsMessage", - "resultWrapper":"DescribeOptionGroupOptionsResult" - } - }, - "DescribeOptionGroups":{ - "name":"DescribeOptionGroups", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOptionGroupsMessage"}, - "output":{ - "shape":"OptionGroups", - "resultWrapper":"DescribeOptionGroupsResult" - }, - "errors":[ - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "DescribeOrderableDBInstanceOptions":{ - "name":"DescribeOrderableDBInstanceOptions", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeOrderableDBInstanceOptionsMessage"}, - "output":{ - "shape":"OrderableDBInstanceOptionsMessage", - "resultWrapper":"DescribeOrderableDBInstanceOptionsResult" - } - }, - "DescribeReservedDBInstances":{ - "name":"DescribeReservedDBInstances", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedDBInstancesMessage"}, - "output":{ - "shape":"ReservedDBInstanceMessage", - "resultWrapper":"DescribeReservedDBInstancesResult" - }, - "errors":[ - {"shape":"ReservedDBInstanceNotFoundFault"} - ] - }, - "DescribeReservedDBInstancesOfferings":{ - "name":"DescribeReservedDBInstancesOfferings", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DescribeReservedDBInstancesOfferingsMessage"}, - "output":{ - "shape":"ReservedDBInstancesOfferingMessage", - "resultWrapper":"DescribeReservedDBInstancesOfferingsResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"} - ] - }, - "DownloadDBLogFilePortion":{ - "name":"DownloadDBLogFilePortion", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"DownloadDBLogFilePortionMessage"}, - "output":{ - "shape":"DownloadDBLogFilePortionDetails", - "resultWrapper":"DownloadDBLogFilePortionResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBLogFileNotFoundFault"} - ] - }, - "ListTagsForResource":{ - "name":"ListTagsForResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ListTagsForResourceMessage"}, - "output":{ - "shape":"TagListMessage", - "resultWrapper":"ListTagsForResourceResult" - }, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "ModifyDBInstance":{ - "name":"ModifyDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBInstanceMessage"}, - "output":{ - "shape":"ModifyDBInstanceResult", - "resultWrapper":"ModifyDBInstanceResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"}, - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"}, - {"shape":"DBUpgradeDependencyFailureFault"} - ] - }, - "ModifyDBParameterGroup":{ - "name":"ModifyDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBParameterGroupMessage"}, - "output":{ - "shape":"DBParameterGroupNameMessage", - "resultWrapper":"ModifyDBParameterGroupResult" - }, - "errors":[ - {"shape":"DBParameterGroupNotFoundFault"}, - {"shape":"InvalidDBParameterGroupStateFault"} - ] - }, - "ModifyDBSubnetGroup":{ - "name":"ModifyDBSubnetGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyDBSubnetGroupMessage"}, - "output":{ - "shape":"ModifyDBSubnetGroupResult", - "resultWrapper":"ModifyDBSubnetGroupResult" - }, - "errors":[ - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetQuotaExceededFault"}, - {"shape":"SubnetAlreadyInUse"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"} - ] - }, - "ModifyEventSubscription":{ - "name":"ModifyEventSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyEventSubscriptionMessage"}, - "output":{ - "shape":"ModifyEventSubscriptionResult", - "resultWrapper":"ModifyEventSubscriptionResult" - }, - "errors":[ - {"shape":"EventSubscriptionQuotaExceededFault"}, - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SNSInvalidTopicFault"}, - {"shape":"SNSNoAuthorizationFault"}, - {"shape":"SNSTopicArnNotFoundFault"}, - {"shape":"SubscriptionCategoryNotFoundFault"} - ] - }, - "ModifyOptionGroup":{ - "name":"ModifyOptionGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ModifyOptionGroupMessage"}, - "output":{ - "shape":"ModifyOptionGroupResult", - "resultWrapper":"ModifyOptionGroupResult" - }, - "errors":[ - {"shape":"InvalidOptionGroupStateFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "PromoteReadReplica":{ - "name":"PromoteReadReplica", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PromoteReadReplicaMessage"}, - "output":{ - "shape":"PromoteReadReplicaResult", - "resultWrapper":"PromoteReadReplicaResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "PurchaseReservedDBInstancesOffering":{ - "name":"PurchaseReservedDBInstancesOffering", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"PurchaseReservedDBInstancesOfferingMessage"}, - "output":{ - "shape":"PurchaseReservedDBInstancesOfferingResult", - "resultWrapper":"PurchaseReservedDBInstancesOfferingResult" - }, - "errors":[ - {"shape":"ReservedDBInstancesOfferingNotFoundFault"}, - {"shape":"ReservedDBInstanceAlreadyExistsFault"}, - {"shape":"ReservedDBInstanceQuotaExceededFault"} - ] - }, - "RebootDBInstance":{ - "name":"RebootDBInstance", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RebootDBInstanceMessage"}, - "output":{ - "shape":"RebootDBInstanceResult", - "resultWrapper":"RebootDBInstanceResult" - }, - "errors":[ - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"DBInstanceNotFoundFault"} - ] - }, - "RemoveSourceIdentifierFromSubscription":{ - "name":"RemoveSourceIdentifierFromSubscription", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveSourceIdentifierFromSubscriptionMessage"}, - "output":{ - "shape":"RemoveSourceIdentifierFromSubscriptionResult", - "resultWrapper":"RemoveSourceIdentifierFromSubscriptionResult" - }, - "errors":[ - {"shape":"SubscriptionNotFoundFault"}, - {"shape":"SourceNotFoundFault"} - ] - }, - "RemoveTagsFromResource":{ - "name":"RemoveTagsFromResource", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RemoveTagsFromResourceMessage"}, - "errors":[ - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"DBSnapshotNotFoundFault"} - ] - }, - "ResetDBParameterGroup":{ - "name":"ResetDBParameterGroup", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"ResetDBParameterGroupMessage"}, - "output":{ - "shape":"DBParameterGroupNameMessage", - "resultWrapper":"ResetDBParameterGroupResult" - }, - "errors":[ - {"shape":"InvalidDBParameterGroupStateFault"}, - {"shape":"DBParameterGroupNotFoundFault"} - ] - }, - "RestoreDBInstanceFromDBSnapshot":{ - "name":"RestoreDBInstanceFromDBSnapshot", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreDBInstanceFromDBSnapshotMessage"}, - "output":{ - "shape":"RestoreDBInstanceFromDBSnapshotResult", - "resultWrapper":"RestoreDBInstanceFromDBSnapshotResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBSnapshotNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBSnapshotStateFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "RestoreDBInstanceToPointInTime":{ - "name":"RestoreDBInstanceToPointInTime", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RestoreDBInstanceToPointInTimeMessage"}, - "output":{ - "shape":"RestoreDBInstanceToPointInTimeResult", - "resultWrapper":"RestoreDBInstanceToPointInTimeResult" - }, - "errors":[ - {"shape":"DBInstanceAlreadyExistsFault"}, - {"shape":"DBInstanceNotFoundFault"}, - {"shape":"InstanceQuotaExceededFault"}, - {"shape":"InsufficientDBInstanceCapacityFault"}, - {"shape":"InvalidDBInstanceStateFault"}, - {"shape":"PointInTimeRestoreNotEnabledFault"}, - {"shape":"StorageQuotaExceededFault"}, - {"shape":"InvalidVPCNetworkStateFault"}, - {"shape":"InvalidRestoreFault"}, - {"shape":"DBSubnetGroupNotFoundFault"}, - {"shape":"DBSubnetGroupDoesNotCoverEnoughAZs"}, - {"shape":"InvalidSubnet"}, - {"shape":"ProvisionedIopsNotAvailableInAZFault"}, - {"shape":"OptionGroupNotFoundFault"} - ] - }, - "RevokeDBSecurityGroupIngress":{ - "name":"RevokeDBSecurityGroupIngress", - "http":{ - "method":"POST", - "requestUri":"/" - }, - "input":{"shape":"RevokeDBSecurityGroupIngressMessage"}, - "output":{ - "shape":"RevokeDBSecurityGroupIngressResult", - "resultWrapper":"RevokeDBSecurityGroupIngressResult" - }, - "errors":[ - {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"AuthorizationNotFoundFault"}, - {"shape":"InvalidDBSecurityGroupStateFault"} - ] - } - }, - "shapes":{ - "AddSourceIdentifierToSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SourceIdentifier" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SourceIdentifier":{"shape":"String"} - } - }, - "AddSourceIdentifierToSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "AddTagsToResourceMessage":{ - "type":"structure", - "required":[ - "ResourceName", - "Tags" - ], - "members":{ - "ResourceName":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "ApplyMethod":{ - "type":"string", - "enum":[ - "immediate", - "pending-reboot" - ] - }, - "AuthorizationAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizationNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "AuthorizationQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"AuthorizationQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "AuthorizeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "AuthorizeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "AvailabilityZone":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "ProvisionedIopsCapable":{"shape":"Boolean"} - }, - "wrapper":true - }, - "AvailabilityZoneList":{ - "type":"list", - "member":{ - "shape":"AvailabilityZone", - "locationName":"AvailabilityZone" - } - }, - "Boolean":{"type":"boolean"}, - "BooleanOptional":{"type":"boolean"}, - "CharacterSet":{ - "type":"structure", - "members":{ - "CharacterSetName":{"shape":"String"}, - "CharacterSetDescription":{"shape":"String"} - } - }, - "CopyDBSnapshotMessage":{ - "type":"structure", - "required":[ - "SourceDBSnapshotIdentifier", - "TargetDBSnapshotIdentifier" - ], - "members":{ - "SourceDBSnapshotIdentifier":{"shape":"String"}, - "TargetDBSnapshotIdentifier":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CopyDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "CreateDBInstanceMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "AllocatedStorage", - "DBInstanceClass", - "Engine", - "MasterUsername", - "MasterUserPassword" - ], - "members":{ - "DBName":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "DBInstanceClass":{"shape":"String"}, - "Engine":{"shape":"String"}, - "MasterUsername":{"shape":"String"}, - "MasterUserPassword":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "DBParameterGroupName":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "CharacterSetName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBInstanceReadReplicaMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "SourceDBInstanceIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SourceDBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "AvailabilityZone":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "Tags":{"shape":"TagList"}, - "DBSubnetGroupName":{"shape":"String"} - } - }, - "CreateDBInstanceReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "CreateDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "CreateDBParameterGroupMessage":{ - "type":"structure", - "required":[ - "DBParameterGroupName", - "DBParameterGroupFamily", - "Description" - ], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBParameterGroupResult":{ - "type":"structure", - "members":{ - "DBParameterGroup":{"shape":"DBParameterGroup"} - } - }, - "CreateDBSecurityGroupMessage":{ - "type":"structure", - "required":[ - "DBSecurityGroupName", - "DBSecurityGroupDescription" - ], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBSecurityGroupResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "CreateDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBSnapshotIdentifier", - "DBInstanceIdentifier" - ], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "CreateDBSubnetGroupMessage":{ - "type":"structure", - "required":[ - "DBSubnetGroupName", - "DBSubnetGroupDescription", - "SubnetIds" - ], - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "SubnetIds":{"shape":"SubnetIdentifierList"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateDBSubnetGroupResult":{ - "type":"structure", - "members":{ - "DBSubnetGroup":{"shape":"DBSubnetGroup"} - } - }, - "CreateEventSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SnsTopicArn" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "SourceIds":{"shape":"SourceIdsList"}, - "Enabled":{"shape":"BooleanOptional"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "CreateOptionGroupMessage":{ - "type":"structure", - "required":[ - "OptionGroupName", - "EngineName", - "MajorEngineVersion", - "OptionGroupDescription" - ], - "members":{ - "OptionGroupName":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "CreateOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "DBEngineVersion":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "DBEngineDescription":{"shape":"String"}, - "DBEngineVersionDescription":{"shape":"String"}, - "DefaultCharacterSet":{"shape":"CharacterSet"}, - "SupportedCharacterSets":{"shape":"SupportedCharacterSetsList"} - } - }, - "DBEngineVersionList":{ - "type":"list", - "member":{ - "shape":"DBEngineVersion", - "locationName":"DBEngineVersion" - } - }, - "DBEngineVersionMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBEngineVersions":{"shape":"DBEngineVersionList"} - } - }, - "DBInstance":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Engine":{"shape":"String"}, - "DBInstanceStatus":{"shape":"String"}, - "MasterUsername":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Endpoint":{"shape":"Endpoint"}, - "AllocatedStorage":{"shape":"Integer"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "PreferredBackupWindow":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"Integer"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroups":{"shape":"VpcSecurityGroupMembershipList"}, - "DBParameterGroups":{"shape":"DBParameterGroupStatusList"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroup":{"shape":"DBSubnetGroup"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "PendingModifiedValues":{"shape":"PendingModifiedValues"}, - "LatestRestorableTime":{"shape":"TStamp"}, - "MultiAZ":{"shape":"Boolean"}, - "EngineVersion":{"shape":"String"}, - "AutoMinorVersionUpgrade":{"shape":"Boolean"}, - "ReadReplicaSourceDBInstanceIdentifier":{"shape":"String"}, - "ReadReplicaDBInstanceIdentifiers":{"shape":"ReadReplicaDBInstanceIdentifierList"}, - "LicenseModel":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupMemberships":{"shape":"OptionGroupMembershipList"}, - "CharacterSetName":{"shape":"String"}, - "SecondaryAvailabilityZone":{"shape":"String"}, - "PubliclyAccessible":{"shape":"Boolean"}, - "StatusInfos":{"shape":"DBInstanceStatusInfoList"} - }, - "wrapper":true - }, - "DBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBInstanceAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBInstanceList":{ - "type":"list", - "member":{ - "shape":"DBInstance", - "locationName":"DBInstance" - } - }, - "DBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBInstances":{"shape":"DBInstanceList"} - } - }, - "DBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBInstanceStatusInfo":{ - "type":"structure", - "members":{ - "StatusType":{"shape":"String"}, - "Normal":{"shape":"Boolean"}, - "Status":{"shape":"String"}, - "Message":{"shape":"String"} - } - }, - "DBInstanceStatusInfoList":{ - "type":"list", - "member":{ - "shape":"DBInstanceStatusInfo", - "locationName":"DBInstanceStatusInfo" - } - }, - "DBLogFileNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBLogFileNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroup":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Description":{"shape":"String"} - }, - "wrapper":true - }, - "DBParameterGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupDetails":{ - "type":"structure", - "members":{ - "Parameters":{"shape":"ParametersList"}, - "Marker":{"shape":"String"} - } - }, - "DBParameterGroupList":{ - "type":"list", - "member":{ - "shape":"DBParameterGroup", - "locationName":"DBParameterGroup" - } - }, - "DBParameterGroupNameMessage":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"} - } - }, - "DBParameterGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBParameterGroupQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBParameterGroupStatus":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "ParameterApplyStatus":{"shape":"String"} - } - }, - "DBParameterGroupStatusList":{ - "type":"list", - "member":{ - "shape":"DBParameterGroupStatus", - "locationName":"DBParameterGroup" - } - }, - "DBParameterGroupsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBParameterGroups":{"shape":"DBParameterGroupList"} - } - }, - "DBSecurityGroup":{ - "type":"structure", - "members":{ - "OwnerId":{"shape":"String"}, - "DBSecurityGroupName":{"shape":"String"}, - "DBSecurityGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "EC2SecurityGroups":{"shape":"EC2SecurityGroupList"}, - "IPRanges":{"shape":"IPRangeList"} - }, - "wrapper":true - }, - "DBSecurityGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupMembership":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "DBSecurityGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroupMembership", - "locationName":"DBSecurityGroup" - } - }, - "DBSecurityGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroups"} - } - }, - "DBSecurityGroupNameList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"DBSecurityGroupName" - } - }, - "DBSecurityGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupNotSupportedFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSecurityGroupNotSupported", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"QuotaExceeded.DBSecurityGroup", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSecurityGroups":{ - "type":"list", - "member":{ - "shape":"DBSecurityGroup", - "locationName":"DBSecurityGroup" - } - }, - "DBSnapshot":{ - "type":"structure", - "members":{ - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceIdentifier":{"shape":"String"}, - "SnapshotCreateTime":{"shape":"TStamp"}, - "Engine":{"shape":"String"}, - "AllocatedStorage":{"shape":"Integer"}, - "Status":{"shape":"String"}, - "Port":{"shape":"Integer"}, - "AvailabilityZone":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "InstanceCreateTime":{"shape":"TStamp"}, - "MasterUsername":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "PercentProgress":{"shape":"Integer"}, - "SourceRegion":{"shape":"String"} - }, - "wrapper":true - }, - "DBSnapshotAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSnapshotAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSnapshotList":{ - "type":"list", - "member":{ - "shape":"DBSnapshot", - "locationName":"DBSnapshot" - } - }, - "DBSnapshotMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSnapshots":{"shape":"DBSnapshotList"} - } - }, - "DBSnapshotNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSnapshotNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroup":{ - "type":"structure", - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "VpcId":{"shape":"String"}, - "SubnetGroupStatus":{"shape":"String"}, - "Subnets":{"shape":"SubnetList"} - }, - "wrapper":true - }, - "DBSubnetGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupAlreadyExists", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupDoesNotCoverEnoughAZs":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupDoesNotCoverEnoughAZs", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "DBSubnetGroups":{"shape":"DBSubnetGroups"} - } - }, - "DBSubnetGroupNotAllowedFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupNotAllowedFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetGroupQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBSubnetGroups":{ - "type":"list", - "member":{ - "shape":"DBSubnetGroup", - "locationName":"DBSubnetGroup" - } - }, - "DBSubnetQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBSubnetQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DBUpgradeDependencyFailureFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"DBUpgradeDependencyFailure", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "DeleteDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "SkipFinalSnapshot":{"shape":"Boolean"}, - "FinalDBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "DeleteDBParameterGroupMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"} - } - }, - "DeleteDBSecurityGroupMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"} - } - }, - "DeleteDBSnapshotMessage":{ - "type":"structure", - "required":["DBSnapshotIdentifier"], - "members":{ - "DBSnapshotIdentifier":{"shape":"String"} - } - }, - "DeleteDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBSnapshot":{"shape":"DBSnapshot"} - } - }, - "DeleteDBSubnetGroupMessage":{ - "type":"structure", - "required":["DBSubnetGroupName"], - "members":{ - "DBSubnetGroupName":{"shape":"String"} - } - }, - "DeleteEventSubscriptionMessage":{ - "type":"structure", - "required":["SubscriptionName"], - "members":{ - "SubscriptionName":{"shape":"String"} - } - }, - "DeleteEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "DeleteOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"} - } - }, - "DescribeDBEngineVersionsMessage":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBParameterGroupFamily":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"}, - "DefaultOnly":{"shape":"Boolean"}, - "ListSupportedCharacterSets":{"shape":"BooleanOptional"} - } - }, - "DescribeDBInstancesMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBLogFilesDetails":{ - "type":"structure", - "members":{ - "LogFileName":{"shape":"String"}, - "LastWritten":{"shape":"Long"}, - "Size":{"shape":"Long"} - } - }, - "DescribeDBLogFilesList":{ - "type":"list", - "member":{ - "shape":"DescribeDBLogFilesDetails", - "locationName":"DescribeDBLogFilesDetails" - } - }, - "DescribeDBLogFilesMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "FilenameContains":{"shape":"String"}, - "FileLastWritten":{"shape":"Long"}, - "FileSize":{"shape":"Long"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBLogFilesResponse":{ - "type":"structure", - "members":{ - "DescribeDBLogFiles":{"shape":"DescribeDBLogFilesList"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBParameterGroupsMessage":{ - "type":"structure", - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBParametersMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Source":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSecurityGroupsMessage":{ - "type":"structure", - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSnapshotsMessage":{ - "type":"structure", - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "SnapshotType":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeDBSubnetGroupsMessage":{ - "type":"structure", - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEngineDefaultParametersMessage":{ - "type":"structure", - "required":["DBParameterGroupFamily"], - "members":{ - "DBParameterGroupFamily":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEngineDefaultParametersResult":{ - "type":"structure", - "members":{ - "EngineDefaults":{"shape":"EngineDefaults"} - } - }, - "DescribeEventCategoriesMessage":{ - "type":"structure", - "members":{ - "SourceType":{"shape":"String"}, - "Filters":{"shape":"FilterList"} - } - }, - "DescribeEventSubscriptionsMessage":{ - "type":"structure", - "members":{ - "SubscriptionName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeEventsMessage":{ - "type":"structure", - "members":{ - "SourceIdentifier":{"shape":"String"}, - "SourceType":{"shape":"SourceType"}, - "StartTime":{"shape":"TStamp"}, - "EndTime":{"shape":"TStamp"}, - "Duration":{"shape":"IntegerOptional"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupOptionsMessage":{ - "type":"structure", - "required":["EngineName"], - "members":{ - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeOptionGroupsMessage":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Filters":{"shape":"FilterList"}, - "Marker":{"shape":"String"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"} - } - }, - "DescribeOrderableDBInstanceOptionsMessage":{ - "type":"structure", - "required":["Engine"], - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "Vpc":{"shape":"BooleanOptional"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "DescribeReservedDBInstancesOfferingsMessage":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "Filters":{"shape":"FilterList"}, - "MaxRecords":{"shape":"IntegerOptional"}, - "Marker":{"shape":"String"} - } - }, - "Double":{"type":"double"}, - "DownloadDBLogFilePortionDetails":{ - "type":"structure", - "members":{ - "LogFileData":{"shape":"String"}, - "Marker":{"shape":"String"}, - "AdditionalDataPending":{"shape":"Boolean"} - } - }, - "DownloadDBLogFilePortionMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "LogFileName" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "LogFileName":{"shape":"String"}, - "Marker":{"shape":"String"}, - "NumberOfLines":{"shape":"Integer"} - } - }, - "EC2SecurityGroup":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "EC2SecurityGroupList":{ - "type":"list", - "member":{ - "shape":"EC2SecurityGroup", - "locationName":"EC2SecurityGroup" - } - }, - "Endpoint":{ - "type":"structure", - "members":{ - "Address":{"shape":"String"}, - "Port":{"shape":"Integer"} - } - }, - "EngineDefaults":{ - "type":"structure", - "members":{ - "DBParameterGroupFamily":{"shape":"String"}, - "Marker":{"shape":"String"}, - "Parameters":{"shape":"ParametersList"} - }, - "wrapper":true - }, - "Event":{ - "type":"structure", - "members":{ - "SourceIdentifier":{"shape":"String"}, - "SourceType":{"shape":"SourceType"}, - "Message":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Date":{"shape":"TStamp"} - } - }, - "EventCategoriesList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"EventCategory" - } - }, - "EventCategoriesMap":{ - "type":"structure", - "members":{ - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"} - }, - "wrapper":true - }, - "EventCategoriesMapList":{ - "type":"list", - "member":{ - "shape":"EventCategoriesMap", - "locationName":"EventCategoriesMap" - } - }, - "EventCategoriesMessage":{ - "type":"structure", - "members":{ - "EventCategoriesMapList":{"shape":"EventCategoriesMapList"} - } - }, - "EventList":{ - "type":"list", - "member":{ - "shape":"Event", - "locationName":"Event" - } - }, - "EventSubscription":{ - "type":"structure", - "members":{ - "CustomerAwsId":{"shape":"String"}, - "CustSubscriptionId":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "Status":{"shape":"String"}, - "SubscriptionCreationTime":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "SourceIdsList":{"shape":"SourceIdsList"}, - "EventCategoriesList":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"Boolean"} - }, - "wrapper":true - }, - "EventSubscriptionQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"EventSubscriptionQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "EventSubscriptionsList":{ - "type":"list", - "member":{ - "shape":"EventSubscription", - "locationName":"EventSubscription" - } - }, - "EventSubscriptionsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "EventSubscriptionsList":{"shape":"EventSubscriptionsList"} - } - }, - "EventsMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "Events":{"shape":"EventList"} - } - }, - "Filter":{ - "type":"structure", - "required":[ - "Name", - "Values" - ], - "members":{ - "Name":{"shape":"String"}, - "Values":{"shape":"FilterValueList"} - } - }, - "FilterList":{ - "type":"list", - "member":{ - "shape":"Filter", - "locationName":"Filter" - } - }, - "FilterValueList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"Value" - } - }, - "IPRange":{ - "type":"structure", - "members":{ - "Status":{"shape":"String"}, - "CIDRIP":{"shape":"String"} - } - }, - "IPRangeList":{ - "type":"list", - "member":{ - "shape":"IPRange", - "locationName":"IPRange" - } - }, - "InstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InsufficientDBInstanceCapacityFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InsufficientDBInstanceCapacity", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "Integer":{"type":"integer"}, - "IntegerOptional":{"type":"integer"}, - "InvalidDBInstanceStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBInstanceState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBParameterGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBParameterGroupState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSecurityGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSecurityGroupState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSnapshotStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSnapshotState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetGroupFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetGroupFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidDBSubnetStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidDBSubnetStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidEventSubscriptionStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidEventSubscriptionState", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidOptionGroupStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidOptionGroupStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidRestoreFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidRestoreFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidSubnet":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidSubnet", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "InvalidVPCNetworkStateFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"InvalidVPCNetworkStateFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "KeyList":{ - "type":"list", - "member":{"shape":"String"} - }, - "ListTagsForResourceMessage":{ - "type":"structure", - "required":["ResourceName"], - "members":{ - "ResourceName":{"shape":"String"}, - "Filters":{"shape":"FilterList"} - } - }, - "Long":{"type":"long"}, - "ModifyDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "DBInstanceClass":{"shape":"String"}, - "DBSecurityGroups":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupIds":{"shape":"VpcSecurityGroupIdList"}, - "ApplyImmediately":{"shape":"Boolean"}, - "MasterUserPassword":{"shape":"String"}, - "DBParameterGroupName":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"}, - "PreferredMaintenanceWindow":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "AllowMajorVersionUpgrade":{"shape":"Boolean"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "NewDBInstanceIdentifier":{"shape":"String"} - } - }, - "ModifyDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "ModifyDBParameterGroupMessage":{ - "type":"structure", - "required":[ - "DBParameterGroupName", - "Parameters" - ], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "Parameters":{"shape":"ParametersList"} - } - }, - "ModifyDBSubnetGroupMessage":{ - "type":"structure", - "required":[ - "DBSubnetGroupName", - "SubnetIds" - ], - "members":{ - "DBSubnetGroupName":{"shape":"String"}, - "DBSubnetGroupDescription":{"shape":"String"}, - "SubnetIds":{"shape":"SubnetIdentifierList"} - } - }, - "ModifyDBSubnetGroupResult":{ - "type":"structure", - "members":{ - "DBSubnetGroup":{"shape":"DBSubnetGroup"} - } - }, - "ModifyEventSubscriptionMessage":{ - "type":"structure", - "required":["SubscriptionName"], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SnsTopicArn":{"shape":"String"}, - "SourceType":{"shape":"String"}, - "EventCategories":{"shape":"EventCategoriesList"}, - "Enabled":{"shape":"BooleanOptional"} - } - }, - "ModifyEventSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "ModifyOptionGroupMessage":{ - "type":"structure", - "required":["OptionGroupName"], - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionsToInclude":{"shape":"OptionConfigurationList"}, - "OptionsToRemove":{"shape":"OptionNamesList"}, - "ApplyImmediately":{"shape":"Boolean"} - } - }, - "ModifyOptionGroupResult":{ - "type":"structure", - "members":{ - "OptionGroup":{"shape":"OptionGroup"} - } - }, - "Option":{ - "type":"structure", - "members":{ - "OptionName":{"shape":"String"}, - "OptionDescription":{"shape":"String"}, - "Persistent":{"shape":"Boolean"}, - "Permanent":{"shape":"Boolean"}, - "Port":{"shape":"IntegerOptional"}, - "OptionSettings":{"shape":"OptionSettingConfigurationList"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupMembershipList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupMembershipList"} - } - }, - "OptionConfiguration":{ - "type":"structure", - "required":["OptionName"], - "members":{ - "OptionName":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "DBSecurityGroupMemberships":{"shape":"DBSecurityGroupNameList"}, - "VpcSecurityGroupMemberships":{"shape":"VpcSecurityGroupIdList"}, - "OptionSettings":{"shape":"OptionSettingsList"} - } - }, - "OptionConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionConfiguration", - "locationName":"OptionConfiguration" - } - }, - "OptionGroup":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "OptionGroupDescription":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "Options":{"shape":"OptionsList"}, - "AllowsVpcAndNonVpcInstanceMemberships":{"shape":"Boolean"}, - "VpcId":{"shape":"String"} - }, - "wrapper":true - }, - "OptionGroupAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupAlreadyExistsFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroupMembership":{ - "type":"structure", - "members":{ - "OptionGroupName":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "OptionGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"OptionGroupMembership", - "locationName":"OptionGroupMembership" - } - }, - "OptionGroupNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupNotFoundFault", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "OptionGroupOption":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Description":{"shape":"String"}, - "EngineName":{"shape":"String"}, - "MajorEngineVersion":{"shape":"String"}, - "MinimumRequiredMinorEngineVersion":{"shape":"String"}, - "PortRequired":{"shape":"Boolean"}, - "DefaultPort":{"shape":"IntegerOptional"}, - "OptionsDependedOn":{"shape":"OptionsDependedOn"}, - "Persistent":{"shape":"Boolean"}, - "Permanent":{"shape":"Boolean"}, - "OptionGroupOptionSettings":{"shape":"OptionGroupOptionSettingsList"} - } - }, - "OptionGroupOptionSetting":{ - "type":"structure", - "members":{ - "SettingName":{"shape":"String"}, - "SettingDescription":{"shape":"String"}, - "DefaultValue":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"} - } - }, - "OptionGroupOptionSettingsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOptionSetting", - "locationName":"OptionGroupOptionSetting" - } - }, - "OptionGroupOptionsList":{ - "type":"list", - "member":{ - "shape":"OptionGroupOption", - "locationName":"OptionGroupOption" - } - }, - "OptionGroupOptionsMessage":{ - "type":"structure", - "members":{ - "OptionGroupOptions":{"shape":"OptionGroupOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"OptionGroupQuotaExceededFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "OptionGroups":{ - "type":"structure", - "members":{ - "OptionGroupsList":{"shape":"OptionGroupsList"}, - "Marker":{"shape":"String"} - } - }, - "OptionGroupsList":{ - "type":"list", - "member":{ - "shape":"OptionGroup", - "locationName":"OptionGroup" - } - }, - "OptionNamesList":{ - "type":"list", - "member":{"shape":"String"} - }, - "OptionSetting":{ - "type":"structure", - "members":{ - "Name":{"shape":"String"}, - "Value":{"shape":"String"}, - "DefaultValue":{"shape":"String"}, - "Description":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "DataType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"}, - "IsCollection":{"shape":"Boolean"} - } - }, - "OptionSettingConfigurationList":{ - "type":"list", - "member":{ - "shape":"OptionSetting", - "locationName":"OptionSetting" - } - }, - "OptionSettingsList":{ - "type":"list", - "member":{ - "shape":"OptionSetting", - "locationName":"OptionSetting" - } - }, - "OptionsDependedOn":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"OptionName" - } - }, - "OptionsList":{ - "type":"list", - "member":{ - "shape":"Option", - "locationName":"Option" - } - }, - "OrderableDBInstanceOption":{ - "type":"structure", - "members":{ - "Engine":{"shape":"String"}, - "EngineVersion":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "LicenseModel":{"shape":"String"}, - "AvailabilityZones":{"shape":"AvailabilityZoneList"}, - "MultiAZCapable":{"shape":"Boolean"}, - "ReadReplicaCapable":{"shape":"Boolean"}, - "Vpc":{"shape":"Boolean"} - }, - "wrapper":true - }, - "OrderableDBInstanceOptionsList":{ - "type":"list", - "member":{ - "shape":"OrderableDBInstanceOption", - "locationName":"OrderableDBInstanceOption" - } - }, - "OrderableDBInstanceOptionsMessage":{ - "type":"structure", - "members":{ - "OrderableDBInstanceOptions":{"shape":"OrderableDBInstanceOptionsList"}, - "Marker":{"shape":"String"} - } - }, - "Parameter":{ - "type":"structure", - "members":{ - "ParameterName":{"shape":"String"}, - "ParameterValue":{"shape":"String"}, - "Description":{"shape":"String"}, - "Source":{"shape":"String"}, - "ApplyType":{"shape":"String"}, - "DataType":{"shape":"String"}, - "AllowedValues":{"shape":"String"}, - "IsModifiable":{"shape":"Boolean"}, - "MinimumEngineVersion":{"shape":"String"}, - "ApplyMethod":{"shape":"ApplyMethod"} - } - }, - "ParametersList":{ - "type":"list", - "member":{ - "shape":"Parameter", - "locationName":"Parameter" - } - }, - "PendingModifiedValues":{ - "type":"structure", - "members":{ - "DBInstanceClass":{"shape":"String"}, - "AllocatedStorage":{"shape":"IntegerOptional"}, - "MasterUserPassword":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "EngineVersion":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "DBInstanceIdentifier":{"shape":"String"} - } - }, - "PointInTimeRestoreNotEnabledFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"PointInTimeRestoreNotEnabled", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PromoteReadReplicaMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "BackupRetentionPeriod":{"shape":"IntegerOptional"}, - "PreferredBackupWindow":{"shape":"String"} - } - }, - "PromoteReadReplicaResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "ProvisionedIopsNotAvailableInAZFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ProvisionedIopsNotAvailableInAZFault", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "PurchaseReservedDBInstancesOfferingMessage":{ - "type":"structure", - "required":["ReservedDBInstancesOfferingId"], - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "ReservedDBInstanceId":{"shape":"String"}, - "DBInstanceCount":{"shape":"IntegerOptional"}, - "Tags":{"shape":"TagList"} - } - }, - "PurchaseReservedDBInstancesOfferingResult":{ - "type":"structure", - "members":{ - "ReservedDBInstance":{"shape":"ReservedDBInstance"} - } - }, - "ReadReplicaDBInstanceIdentifierList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"ReadReplicaDBInstanceIdentifier" - } - }, - "RebootDBInstanceMessage":{ - "type":"structure", - "required":["DBInstanceIdentifier"], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "ForceFailover":{"shape":"BooleanOptional"} - } - }, - "RebootDBInstanceResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RecurringCharge":{ - "type":"structure", - "members":{ - "RecurringChargeAmount":{"shape":"Double"}, - "RecurringChargeFrequency":{"shape":"String"} - }, - "wrapper":true - }, - "RecurringChargeList":{ - "type":"list", - "member":{ - "shape":"RecurringCharge", - "locationName":"RecurringCharge" - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage":{ - "type":"structure", - "required":[ - "SubscriptionName", - "SourceIdentifier" - ], - "members":{ - "SubscriptionName":{"shape":"String"}, - "SourceIdentifier":{"shape":"String"} - } - }, - "RemoveSourceIdentifierFromSubscriptionResult":{ - "type":"structure", - "members":{ - "EventSubscription":{"shape":"EventSubscription"} - } - }, - "RemoveTagsFromResourceMessage":{ - "type":"structure", - "required":[ - "ResourceName", - "TagKeys" - ], - "members":{ - "ResourceName":{"shape":"String"}, - "TagKeys":{"shape":"KeyList"} - } - }, - "ReservedDBInstance":{ - "type":"structure", - "members":{ - "ReservedDBInstanceId":{"shape":"String"}, - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "StartTime":{"shape":"TStamp"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "DBInstanceCount":{"shape":"Integer"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "State":{"shape":"String"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstanceAlreadyExistsFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceAlreadyExists", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstance", - "locationName":"ReservedDBInstance" - } - }, - "ReservedDBInstanceMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstances":{"shape":"ReservedDBInstanceList"} - } - }, - "ReservedDBInstanceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstanceQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstanceQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "ReservedDBInstancesOffering":{ - "type":"structure", - "members":{ - "ReservedDBInstancesOfferingId":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Duration":{"shape":"Integer"}, - "FixedPrice":{"shape":"Double"}, - "UsagePrice":{"shape":"Double"}, - "CurrencyCode":{"shape":"String"}, - "ProductDescription":{"shape":"String"}, - "OfferingType":{"shape":"String"}, - "MultiAZ":{"shape":"Boolean"}, - "RecurringCharges":{"shape":"RecurringChargeList"} - }, - "wrapper":true - }, - "ReservedDBInstancesOfferingList":{ - "type":"list", - "member":{ - "shape":"ReservedDBInstancesOffering", - "locationName":"ReservedDBInstancesOffering" - } - }, - "ReservedDBInstancesOfferingMessage":{ - "type":"structure", - "members":{ - "Marker":{"shape":"String"}, - "ReservedDBInstancesOfferings":{"shape":"ReservedDBInstancesOfferingList"} - } - }, - "ReservedDBInstancesOfferingNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"ReservedDBInstancesOfferingNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "ResetDBParameterGroupMessage":{ - "type":"structure", - "required":["DBParameterGroupName"], - "members":{ - "DBParameterGroupName":{"shape":"String"}, - "ResetAllParameters":{"shape":"Boolean"}, - "Parameters":{"shape":"ParametersList"} - } - }, - "RestoreDBInstanceFromDBSnapshotMessage":{ - "type":"structure", - "required":[ - "DBInstanceIdentifier", - "DBSnapshotIdentifier" - ], - "members":{ - "DBInstanceIdentifier":{"shape":"String"}, - "DBSnapshotIdentifier":{"shape":"String"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "RestoreDBInstanceFromDBSnapshotResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RestoreDBInstanceToPointInTimeMessage":{ - "type":"structure", - "required":[ - "SourceDBInstanceIdentifier", - "TargetDBInstanceIdentifier" - ], - "members":{ - "SourceDBInstanceIdentifier":{"shape":"String"}, - "TargetDBInstanceIdentifier":{"shape":"String"}, - "RestoreTime":{"shape":"TStamp"}, - "UseLatestRestorableTime":{"shape":"Boolean"}, - "DBInstanceClass":{"shape":"String"}, - "Port":{"shape":"IntegerOptional"}, - "AvailabilityZone":{"shape":"String"}, - "DBSubnetGroupName":{"shape":"String"}, - "MultiAZ":{"shape":"BooleanOptional"}, - "PubliclyAccessible":{"shape":"BooleanOptional"}, - "AutoMinorVersionUpgrade":{"shape":"BooleanOptional"}, - "LicenseModel":{"shape":"String"}, - "DBName":{"shape":"String"}, - "Engine":{"shape":"String"}, - "Iops":{"shape":"IntegerOptional"}, - "OptionGroupName":{"shape":"String"}, - "Tags":{"shape":"TagList"} - } - }, - "RestoreDBInstanceToPointInTimeResult":{ - "type":"structure", - "members":{ - "DBInstance":{"shape":"DBInstance"} - } - }, - "RevokeDBSecurityGroupIngressMessage":{ - "type":"structure", - "required":["DBSecurityGroupName"], - "members":{ - "DBSecurityGroupName":{"shape":"String"}, - "CIDRIP":{"shape":"String"}, - "EC2SecurityGroupName":{"shape":"String"}, - "EC2SecurityGroupId":{"shape":"String"}, - "EC2SecurityGroupOwnerId":{"shape":"String"} - } - }, - "RevokeDBSecurityGroupIngressResult":{ - "type":"structure", - "members":{ - "DBSecurityGroup":{"shape":"DBSecurityGroup"} - } - }, - "SNSInvalidTopicFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSInvalidTopic", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SNSNoAuthorizationFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSNoAuthorization", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SNSTopicArnNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SNSTopicArnNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SnapshotQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SnapshotQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SourceIdsList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SourceId" - } - }, - "SourceNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SourceNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SourceType":{ - "type":"string", - "enum":[ - "db-instance", - "db-parameter-group", - "db-security-group", - "db-snapshot" - ] - }, - "StorageQuotaExceededFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"StorageQuotaExceeded", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "String":{"type":"string"}, - "Subnet":{ - "type":"structure", - "members":{ - "SubnetIdentifier":{"shape":"String"}, - "SubnetAvailabilityZone":{"shape":"AvailabilityZone"}, - "SubnetStatus":{"shape":"String"} - } - }, - "SubnetAlreadyInUse":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubnetAlreadyInUse", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SubnetIdentifierList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"SubnetIdentifier" - } - }, - "SubnetList":{ - "type":"list", - "member":{ - "shape":"Subnet", - "locationName":"Subnet" - } - }, - "SubscriptionAlreadyExistFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionAlreadyExist", - "httpStatusCode":400, - "senderFault":true - }, - "exception":true - }, - "SubscriptionCategoryNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionCategoryNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SubscriptionNotFoundFault":{ - "type":"structure", - "members":{ - }, - "error":{ - "code":"SubscriptionNotFound", - "httpStatusCode":404, - "senderFault":true - }, - "exception":true - }, - "SupportedCharacterSetsList":{ - "type":"list", - "member":{ - "shape":"CharacterSet", - "locationName":"CharacterSet" - } - }, - "TStamp":{"type":"timestamp"}, - "Tag":{ - "type":"structure", - "members":{ - "Key":{"shape":"String"}, - "Value":{"shape":"String"} - } - }, - "TagList":{ - "type":"list", - "member":{ - "shape":"Tag", - "locationName":"Tag" - } - }, - "TagListMessage":{ - "type":"structure", - "members":{ - "TagList":{"shape":"TagList"} - } - }, - "VpcSecurityGroupIdList":{ - "type":"list", - "member":{ - "shape":"String", - "locationName":"VpcSecurityGroupId" - } - }, - "VpcSecurityGroupMembership":{ - "type":"structure", - "members":{ - "VpcSecurityGroupId":{"shape":"String"}, - "Status":{"shape":"String"} - } - }, - "VpcSecurityGroupMembershipList":{ - "type":"list", - "member":{ - "shape":"VpcSecurityGroupMembership", - "locationName":"VpcSecurityGroupMembership" - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/docs-2.json deleted file mode 100644 index 9ccecc63b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/docs-2.json +++ /dev/null @@ -1,1876 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "AddSourceIdentifierToSubscription": null, - "AddTagsToResource": null, - "AuthorizeDBSecurityGroupIngress": null, - "CopyDBSnapshot": null, - "CreateDBInstance": null, - "CreateDBInstanceReadReplica": null, - "CreateDBParameterGroup": null, - "CreateDBSecurityGroup": null, - "CreateDBSnapshot": null, - "CreateDBSubnetGroup": null, - "CreateEventSubscription": null, - "CreateOptionGroup": null, - "DeleteDBInstance": null, - "DeleteDBParameterGroup": null, - "DeleteDBSecurityGroup": null, - "DeleteDBSnapshot": null, - "DeleteDBSubnetGroup": null, - "DeleteEventSubscription": null, - "DeleteOptionGroup": null, - "DescribeDBEngineVersions": null, - "DescribeDBInstances": null, - "DescribeDBLogFiles": null, - "DescribeDBParameterGroups": null, - "DescribeDBParameters": null, - "DescribeDBSecurityGroups": null, - "DescribeDBSnapshots": null, - "DescribeDBSubnetGroups": null, - "DescribeEngineDefaultParameters": null, - "DescribeEventCategories": null, - "DescribeEventSubscriptions": null, - "DescribeEvents": null, - "DescribeOptionGroupOptions": null, - "DescribeOptionGroups": null, - "DescribeOrderableDBInstanceOptions": null, - "DescribeReservedDBInstances": null, - "DescribeReservedDBInstancesOfferings": null, - "DownloadDBLogFilePortion": null, - "ListTagsForResource": null, - "ModifyDBInstance": null, - "ModifyDBParameterGroup": null, - "ModifyDBSubnetGroup": null, - "ModifyEventSubscription": null, - "ModifyOptionGroup": null, - "PromoteReadReplica": null, - "PurchaseReservedDBInstancesOffering": null, - "RebootDBInstance": null, - "RemoveSourceIdentifierFromSubscription": null, - "RemoveTagsFromResource": null, - "ResetDBParameterGroup": null, - "RestoreDBInstanceFromDBSnapshot": null, - "RestoreDBInstanceToPointInTime": null, - "RevokeDBSecurityGroupIngress": null - }, - "shapes": { - "AddSourceIdentifierToSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "AddSourceIdentifierToSubscriptionResult": { - "base": null, - "refs": { - } - }, - "AddTagsToResourceMessage": { - "base": null, - "refs": { - } - }, - "ApplyMethod": { - "base": null, - "refs": { - "Parameter$ApplyMethod": null - } - }, - "AuthorizationAlreadyExistsFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group is already authorized for the specified DB security group.

", - "refs": { - } - }, - "AuthorizationNotFoundFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group isn't authorized for the specified DB security group.

RDS also may not be authorized by using IAM to perform necessary actions on your behalf.

", - "refs": { - } - }, - "AuthorizationQuotaExceededFault": { - "base": "

The DB security group authorization quota has been reached.

", - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "AvailabilityZone": { - "base": null, - "refs": { - "AvailabilityZoneList$member": null, - "Subnet$SubnetAvailabilityZone": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "OrderableDBInstanceOption$AvailabilityZones": null - } - }, - "Boolean": { - "base": null, - "refs": { - "AvailabilityZone$ProvisionedIopsCapable": null, - "DBInstance$MultiAZ": null, - "DBInstance$AutoMinorVersionUpgrade": null, - "DBInstance$PubliclyAccessible": null, - "DBInstanceStatusInfo$Normal": null, - "DeleteDBInstanceMessage$SkipFinalSnapshot": null, - "DescribeDBEngineVersionsMessage$DefaultOnly": null, - "DownloadDBLogFilePortionDetails$AdditionalDataPending": null, - "EventSubscription$Enabled": null, - "ModifyDBInstanceMessage$ApplyImmediately": null, - "ModifyDBInstanceMessage$AllowMajorVersionUpgrade": null, - "ModifyOptionGroupMessage$ApplyImmediately": null, - "Option$Persistent": null, - "Option$Permanent": null, - "OptionGroup$AllowsVpcAndNonVpcInstanceMemberships": null, - "OptionGroupOption$PortRequired": null, - "OptionGroupOption$Persistent": null, - "OptionGroupOption$Permanent": null, - "OptionGroupOptionSetting$IsModifiable": null, - "OptionSetting$IsModifiable": null, - "OptionSetting$IsCollection": null, - "OrderableDBInstanceOption$MultiAZCapable": null, - "OrderableDBInstanceOption$ReadReplicaCapable": null, - "OrderableDBInstanceOption$Vpc": null, - "Parameter$IsModifiable": null, - "ReservedDBInstance$MultiAZ": null, - "ReservedDBInstancesOffering$MultiAZ": null, - "ResetDBParameterGroupMessage$ResetAllParameters": null, - "RestoreDBInstanceToPointInTimeMessage$UseLatestRestorableTime": null - } - }, - "BooleanOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$MultiAZ": null, - "CreateDBInstanceMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceMessage$PubliclyAccessible": null, - "CreateDBInstanceReadReplicaMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceReadReplicaMessage$PubliclyAccessible": null, - "CreateEventSubscriptionMessage$Enabled": null, - "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets": null, - "DescribeOrderableDBInstanceOptionsMessage$Vpc": null, - "DescribeReservedDBInstancesMessage$MultiAZ": null, - "DescribeReservedDBInstancesOfferingsMessage$MultiAZ": null, - "ModifyDBInstanceMessage$MultiAZ": null, - "ModifyDBInstanceMessage$AutoMinorVersionUpgrade": null, - "ModifyEventSubscriptionMessage$Enabled": null, - "PendingModifiedValues$MultiAZ": null, - "RebootDBInstanceMessage$ForceFailover": null, - "RestoreDBInstanceFromDBSnapshotMessage$MultiAZ": null, - "RestoreDBInstanceFromDBSnapshotMessage$PubliclyAccessible": null, - "RestoreDBInstanceFromDBSnapshotMessage$AutoMinorVersionUpgrade": null, - "RestoreDBInstanceToPointInTimeMessage$MultiAZ": null, - "RestoreDBInstanceToPointInTimeMessage$PubliclyAccessible": null, - "RestoreDBInstanceToPointInTimeMessage$AutoMinorVersionUpgrade": null - } - }, - "CharacterSet": { - "base": null, - "refs": { - "DBEngineVersion$DefaultCharacterSet": null, - "SupportedCharacterSetsList$member": null - } - }, - "CopyDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CopyDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceResult": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupResult": { - "base": null, - "refs": { - } - }, - "DBEngineVersion": { - "base": null, - "refs": { - "DBEngineVersionList$member": null - } - }, - "DBEngineVersionList": { - "base": null, - "refs": { - "DBEngineVersionMessage$DBEngineVersions": null - } - }, - "DBEngineVersionMessage": { - "base": null, - "refs": { - } - }, - "DBInstance": { - "base": null, - "refs": { - "CreateDBInstanceReadReplicaResult$DBInstance": null, - "CreateDBInstanceResult$DBInstance": null, - "DBInstanceList$member": null, - "DeleteDBInstanceResult$DBInstance": null, - "ModifyDBInstanceResult$DBInstance": null, - "PromoteReadReplicaResult$DBInstance": null, - "RebootDBInstanceResult$DBInstance": null, - "RestoreDBInstanceFromDBSnapshotResult$DBInstance": null, - "RestoreDBInstanceToPointInTimeResult$DBInstance": null - } - }, - "DBInstanceAlreadyExistsFault": { - "base": "

The user already has a DB instance with the given identifier.

", - "refs": { - } - }, - "DBInstanceList": { - "base": null, - "refs": { - "DBInstanceMessage$DBInstances": null - } - }, - "DBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DBInstanceNotFoundFault": { - "base": "

DBInstanceIdentifier doesn't refer to an existing DB instance.

", - "refs": { - } - }, - "DBInstanceStatusInfo": { - "base": null, - "refs": { - "DBInstanceStatusInfoList$member": null - } - }, - "DBInstanceStatusInfoList": { - "base": null, - "refs": { - "DBInstance$StatusInfos": null - } - }, - "DBLogFileNotFoundFault": { - "base": "

LogFileName doesn't refer to an existing DB log file.

", - "refs": { - } - }, - "DBParameterGroup": { - "base": null, - "refs": { - "CreateDBParameterGroupResult$DBParameterGroup": null, - "DBParameterGroupList$member": null - } - }, - "DBParameterGroupAlreadyExistsFault": { - "base": "

A DB parameter group with the same name exists.

", - "refs": { - } - }, - "DBParameterGroupDetails": { - "base": null, - "refs": { - } - }, - "DBParameterGroupList": { - "base": null, - "refs": { - "DBParameterGroupsMessage$DBParameterGroups": null - } - }, - "DBParameterGroupNameMessage": { - "base": null, - "refs": { - } - }, - "DBParameterGroupNotFoundFault": { - "base": "

DBParameterGroupName doesn't refer to an existing DB parameter group.

", - "refs": { - } - }, - "DBParameterGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB parameter groups.

", - "refs": { - } - }, - "DBParameterGroupStatus": { - "base": null, - "refs": { - "DBParameterGroupStatusList$member": null - } - }, - "DBParameterGroupStatusList": { - "base": null, - "refs": { - "DBInstance$DBParameterGroups": null - } - }, - "DBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroup": { - "base": null, - "refs": { - "AuthorizeDBSecurityGroupIngressResult$DBSecurityGroup": null, - "CreateDBSecurityGroupResult$DBSecurityGroup": null, - "DBSecurityGroups$member": null, - "RevokeDBSecurityGroupIngressResult$DBSecurityGroup": null - } - }, - "DBSecurityGroupAlreadyExistsFault": { - "base": "

A DB security group with the name specified in DBSecurityGroupName already exists.

", - "refs": { - } - }, - "DBSecurityGroupMembership": { - "base": null, - "refs": { - "DBSecurityGroupMembershipList$member": null - } - }, - "DBSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$DBSecurityGroups": null, - "Option$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroupNameList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$DBSecurityGroups": null, - "ModifyDBInstanceMessage$DBSecurityGroups": null, - "OptionConfiguration$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupNotFoundFault": { - "base": "

DBSecurityGroupName doesn't refer to an existing DB security group.

", - "refs": { - } - }, - "DBSecurityGroupNotSupportedFault": { - "base": "

A DB security group isn't allowed for this action.

", - "refs": { - } - }, - "DBSecurityGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB security groups.

", - "refs": { - } - }, - "DBSecurityGroups": { - "base": null, - "refs": { - "DBSecurityGroupMessage$DBSecurityGroups": null - } - }, - "DBSnapshot": { - "base": null, - "refs": { - "CopyDBSnapshotResult$DBSnapshot": null, - "CreateDBSnapshotResult$DBSnapshot": null, - "DBSnapshotList$member": null, - "DeleteDBSnapshotResult$DBSnapshot": null - } - }, - "DBSnapshotAlreadyExistsFault": { - "base": "

DBSnapshotIdentifier is already used by an existing snapshot.

", - "refs": { - } - }, - "DBSnapshotList": { - "base": null, - "refs": { - "DBSnapshotMessage$DBSnapshots": null - } - }, - "DBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DBSnapshotNotFoundFault": { - "base": "

DBSnapshotIdentifier doesn't refer to an existing DB snapshot.

", - "refs": { - } - }, - "DBSubnetGroup": { - "base": null, - "refs": { - "CreateDBSubnetGroupResult$DBSubnetGroup": null, - "DBInstance$DBSubnetGroup": null, - "DBSubnetGroups$member": null, - "ModifyDBSubnetGroupResult$DBSubnetGroup": null - } - }, - "DBSubnetGroupAlreadyExistsFault": { - "base": "

DBSubnetGroupName is already used by an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupDoesNotCoverEnoughAZs": { - "base": "

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", - "refs": { - } - }, - "DBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSubnetGroupNotAllowedFault": { - "base": "

The DBSubnetGroup shouldn't be specified while creating read replicas that lie in the same region as the source instance.

", - "refs": { - } - }, - "DBSubnetGroupNotFoundFault": { - "base": "

DBSubnetGroupName doesn't refer to an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB subnet groups.

", - "refs": { - } - }, - "DBSubnetGroups": { - "base": null, - "refs": { - "DBSubnetGroupMessage$DBSubnetGroups": null - } - }, - "DBSubnetQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of subnets in a DB subnet groups.

", - "refs": { - } - }, - "DBUpgradeDependencyFailureFault": { - "base": "

The DB upgrade failed because a resource the DB depends on can't be modified.

", - "refs": { - } - }, - "DeleteDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBInstanceResult": { - "base": null, - "refs": { - } - }, - "DeleteDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "DeleteDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "DeleteOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBEngineVersionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBLogFilesDetails": { - "base": null, - "refs": { - "DescribeDBLogFilesList$member": null - } - }, - "DescribeDBLogFilesList": { - "base": null, - "refs": { - "DescribeDBLogFilesResponse$DescribeDBLogFiles": null - } - }, - "DescribeDBLogFilesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBLogFilesResponse": { - "base": null, - "refs": { - } - }, - "DescribeDBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSecurityGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSnapshotsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSubnetGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersResult": { - "base": null, - "refs": { - } - }, - "DescribeEventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesOfferingsMessage": { - "base": null, - "refs": { - } - }, - "Double": { - "base": null, - "refs": { - "RecurringCharge$RecurringChargeAmount": null, - "ReservedDBInstance$FixedPrice": null, - "ReservedDBInstance$UsagePrice": null, - "ReservedDBInstancesOffering$FixedPrice": null, - "ReservedDBInstancesOffering$UsagePrice": null - } - }, - "DownloadDBLogFilePortionDetails": { - "base": null, - "refs": { - } - }, - "DownloadDBLogFilePortionMessage": { - "base": null, - "refs": { - } - }, - "EC2SecurityGroup": { - "base": null, - "refs": { - "EC2SecurityGroupList$member": null - } - }, - "EC2SecurityGroupList": { - "base": null, - "refs": { - "DBSecurityGroup$EC2SecurityGroups": null - } - }, - "Endpoint": { - "base": null, - "refs": { - "DBInstance$Endpoint": null - } - }, - "EngineDefaults": { - "base": null, - "refs": { - "DescribeEngineDefaultParametersResult$EngineDefaults": null - } - }, - "Event": { - "base": null, - "refs": { - "EventList$member": null - } - }, - "EventCategoriesList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$EventCategories": null, - "DescribeEventsMessage$EventCategories": null, - "Event$EventCategories": null, - "EventCategoriesMap$EventCategories": null, - "EventSubscription$EventCategoriesList": null, - "ModifyEventSubscriptionMessage$EventCategories": null - } - }, - "EventCategoriesMap": { - "base": null, - "refs": { - "EventCategoriesMapList$member": null - } - }, - "EventCategoriesMapList": { - "base": null, - "refs": { - "EventCategoriesMessage$EventCategoriesMapList": null - } - }, - "EventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "EventList": { - "base": null, - "refs": { - "EventsMessage$Events": null - } - }, - "EventSubscription": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionResult$EventSubscription": null, - "CreateEventSubscriptionResult$EventSubscription": null, - "DeleteEventSubscriptionResult$EventSubscription": null, - "EventSubscriptionsList$member": null, - "ModifyEventSubscriptionResult$EventSubscription": null, - "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription": null - } - }, - "EventSubscriptionQuotaExceededFault": { - "base": "

You have reached the maximum number of event subscriptions.

", - "refs": { - } - }, - "EventSubscriptionsList": { - "base": null, - "refs": { - "EventSubscriptionsMessage$EventSubscriptionsList": null - } - }, - "EventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "EventsMessage": { - "base": null, - "refs": { - } - }, - "Filter": { - "base": null, - "refs": { - "FilterList$member": null - } - }, - "FilterList": { - "base": null, - "refs": { - "DescribeDBEngineVersionsMessage$Filters": null, - "DescribeDBInstancesMessage$Filters": null, - "DescribeDBLogFilesMessage$Filters": null, - "DescribeDBParameterGroupsMessage$Filters": null, - "DescribeDBParametersMessage$Filters": null, - "DescribeDBSecurityGroupsMessage$Filters": null, - "DescribeDBSnapshotsMessage$Filters": null, - "DescribeDBSubnetGroupsMessage$Filters": null, - "DescribeEngineDefaultParametersMessage$Filters": null, - "DescribeEventCategoriesMessage$Filters": null, - "DescribeEventSubscriptionsMessage$Filters": null, - "DescribeEventsMessage$Filters": null, - "DescribeOptionGroupOptionsMessage$Filters": null, - "DescribeOptionGroupsMessage$Filters": null, - "DescribeOrderableDBInstanceOptionsMessage$Filters": null, - "DescribeReservedDBInstancesMessage$Filters": null, - "DescribeReservedDBInstancesOfferingsMessage$Filters": null, - "ListTagsForResourceMessage$Filters": null - } - }, - "FilterValueList": { - "base": null, - "refs": { - "Filter$Values": null - } - }, - "IPRange": { - "base": null, - "refs": { - "IPRangeList$member": null - } - }, - "IPRangeList": { - "base": null, - "refs": { - "DBSecurityGroup$IPRanges": null - } - }, - "InstanceQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB instances.

", - "refs": { - } - }, - "InsufficientDBInstanceCapacityFault": { - "base": "

The specified DB instance class isn't available in the specified Availability Zone.

", - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "DBInstance$AllocatedStorage": null, - "DBInstance$BackupRetentionPeriod": null, - "DBSnapshot$AllocatedStorage": null, - "DBSnapshot$Port": null, - "DBSnapshot$PercentProgress": null, - "DownloadDBLogFilePortionMessage$NumberOfLines": null, - "Endpoint$Port": null, - "ReservedDBInstance$Duration": null, - "ReservedDBInstance$DBInstanceCount": null, - "ReservedDBInstancesOffering$Duration": null - } - }, - "IntegerOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$AllocatedStorage": null, - "CreateDBInstanceMessage$BackupRetentionPeriod": null, - "CreateDBInstanceMessage$Port": null, - "CreateDBInstanceMessage$Iops": null, - "CreateDBInstanceReadReplicaMessage$Port": null, - "CreateDBInstanceReadReplicaMessage$Iops": null, - "DBInstance$Iops": null, - "DBSnapshot$Iops": null, - "DescribeDBEngineVersionsMessage$MaxRecords": null, - "DescribeDBInstancesMessage$MaxRecords": null, - "DescribeDBLogFilesMessage$MaxRecords": null, - "DescribeDBParameterGroupsMessage$MaxRecords": null, - "DescribeDBParametersMessage$MaxRecords": null, - "DescribeDBSecurityGroupsMessage$MaxRecords": null, - "DescribeDBSnapshotsMessage$MaxRecords": null, - "DescribeDBSubnetGroupsMessage$MaxRecords": null, - "DescribeEngineDefaultParametersMessage$MaxRecords": null, - "DescribeEventSubscriptionsMessage$MaxRecords": null, - "DescribeEventsMessage$Duration": null, - "DescribeEventsMessage$MaxRecords": null, - "DescribeOptionGroupOptionsMessage$MaxRecords": null, - "DescribeOptionGroupsMessage$MaxRecords": null, - "DescribeOrderableDBInstanceOptionsMessage$MaxRecords": null, - "DescribeReservedDBInstancesMessage$MaxRecords": null, - "DescribeReservedDBInstancesOfferingsMessage$MaxRecords": null, - "ModifyDBInstanceMessage$AllocatedStorage": null, - "ModifyDBInstanceMessage$BackupRetentionPeriod": null, - "ModifyDBInstanceMessage$Iops": null, - "Option$Port": null, - "OptionConfiguration$Port": null, - "OptionGroupOption$DefaultPort": null, - "PendingModifiedValues$AllocatedStorage": null, - "PendingModifiedValues$Port": null, - "PendingModifiedValues$BackupRetentionPeriod": null, - "PendingModifiedValues$Iops": null, - "PromoteReadReplicaMessage$BackupRetentionPeriod": null, - "PurchaseReservedDBInstancesOfferingMessage$DBInstanceCount": null, - "RestoreDBInstanceFromDBSnapshotMessage$Port": null, - "RestoreDBInstanceFromDBSnapshotMessage$Iops": null, - "RestoreDBInstanceToPointInTimeMessage$Port": null, - "RestoreDBInstanceToPointInTimeMessage$Iops": null - } - }, - "InvalidDBInstanceStateFault": { - "base": "

The specified DB instance isn't in the available state.

", - "refs": { - } - }, - "InvalidDBParameterGroupStateFault": { - "base": "

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you can't delete it when the parameter group is in this state.

", - "refs": { - } - }, - "InvalidDBSecurityGroupStateFault": { - "base": "

The state of the DB security group doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSnapshotStateFault": { - "base": "

The state of the DB snapshot doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSubnetGroupFault": { - "base": "

The DBSubnetGroup doesn't belong to the same VPC as that of an existing cross-region read replica of the same source instance.

", - "refs": { - } - }, - "InvalidDBSubnetGroupStateFault": { - "base": "

The DB subnet group cannot be deleted because it's in use.

", - "refs": { - } - }, - "InvalidDBSubnetStateFault": { - "base": "

The DB subnet isn't in the available state.

", - "refs": { - } - }, - "InvalidEventSubscriptionStateFault": { - "base": "

This error can occur if someone else is modifying a subscription. You should retry the action.

", - "refs": { - } - }, - "InvalidOptionGroupStateFault": { - "base": "

The option group isn't in the available state.

", - "refs": { - } - }, - "InvalidRestoreFault": { - "base": "

Cannot restore from VPC backup to non-VPC DB instance.

", - "refs": { - } - }, - "InvalidSubnet": { - "base": "

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", - "refs": { - } - }, - "InvalidVPCNetworkStateFault": { - "base": "

The DB subnet group doesn't cover all Availability Zones after it's created because of users' change.

", - "refs": { - } - }, - "KeyList": { - "base": null, - "refs": { - "RemoveTagsFromResourceMessage$TagKeys": null - } - }, - "ListTagsForResourceMessage": { - "base": null, - "refs": { - } - }, - "Long": { - "base": null, - "refs": { - "DescribeDBLogFilesDetails$LastWritten": null, - "DescribeDBLogFilesDetails$Size": null, - "DescribeDBLogFilesMessage$FileLastWritten": null, - "DescribeDBLogFilesMessage$FileSize": null - } - }, - "ModifyDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBInstanceResult": { - "base": null, - "refs": { - } - }, - "ModifyDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupResult": { - "base": null, - "refs": { - } - }, - "Option": { - "base": null, - "refs": { - "OptionsList$member": null - } - }, - "OptionConfiguration": { - "base": null, - "refs": { - "OptionConfigurationList$member": null - } - }, - "OptionConfigurationList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToInclude": null - } - }, - "OptionGroup": { - "base": null, - "refs": { - "CreateOptionGroupResult$OptionGroup": null, - "ModifyOptionGroupResult$OptionGroup": null, - "OptionGroupsList$member": null - } - }, - "OptionGroupAlreadyExistsFault": { - "base": "

The option group you are trying to create already exists.

", - "refs": { - } - }, - "OptionGroupMembership": { - "base": null, - "refs": { - "OptionGroupMembershipList$member": null - } - }, - "OptionGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$OptionGroupMemberships": null - } - }, - "OptionGroupNotFoundFault": { - "base": "

The specified option group could not be found.

", - "refs": { - } - }, - "OptionGroupOption": { - "base": null, - "refs": { - "OptionGroupOptionsList$member": null - } - }, - "OptionGroupOptionSetting": { - "base": null, - "refs": { - "OptionGroupOptionSettingsList$member": null - } - }, - "OptionGroupOptionSettingsList": { - "base": null, - "refs": { - "OptionGroupOption$OptionGroupOptionSettings": null - } - }, - "OptionGroupOptionsList": { - "base": null, - "refs": { - "OptionGroupOptionsMessage$OptionGroupOptions": null - } - }, - "OptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "OptionGroupQuotaExceededFault": { - "base": "

The quota of 20 option groups was exceeded for this AWS account.

", - "refs": { - } - }, - "OptionGroups": { - "base": null, - "refs": { - } - }, - "OptionGroupsList": { - "base": null, - "refs": { - "OptionGroups$OptionGroupsList": null - } - }, - "OptionNamesList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToRemove": null - } - }, - "OptionSetting": { - "base": null, - "refs": { - "OptionSettingConfigurationList$member": null, - "OptionSettingsList$member": null - } - }, - "OptionSettingConfigurationList": { - "base": null, - "refs": { - "Option$OptionSettings": null - } - }, - "OptionSettingsList": { - "base": null, - "refs": { - "OptionConfiguration$OptionSettings": null - } - }, - "OptionsDependedOn": { - "base": null, - "refs": { - "OptionGroupOption$OptionsDependedOn": null - } - }, - "OptionsList": { - "base": null, - "refs": { - "OptionGroup$Options": null - } - }, - "OrderableDBInstanceOption": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsList$member": null - } - }, - "OrderableDBInstanceOptionsList": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions": null - } - }, - "OrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "Parameter": { - "base": null, - "refs": { - "ParametersList$member": null - } - }, - "ParametersList": { - "base": null, - "refs": { - "DBParameterGroupDetails$Parameters": null, - "EngineDefaults$Parameters": null, - "ModifyDBParameterGroupMessage$Parameters": null, - "ResetDBParameterGroupMessage$Parameters": null - } - }, - "PendingModifiedValues": { - "base": null, - "refs": { - "DBInstance$PendingModifiedValues": null - } - }, - "PointInTimeRestoreNotEnabledFault": { - "base": "

SourceDBInstanceIdentifier refers to a DB instance with BackupRetentionPeriod equal to 0.

", - "refs": { - } - }, - "PromoteReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "PromoteReadReplicaResult": { - "base": null, - "refs": { - } - }, - "ProvisionedIopsNotAvailableInAZFault": { - "base": "

Provisioned IOPS not available in the specified Availability Zone.

", - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingResult": { - "base": null, - "refs": { - } - }, - "ReadReplicaDBInstanceIdentifierList": { - "base": null, - "refs": { - "DBInstance$ReadReplicaDBInstanceIdentifiers": null - } - }, - "RebootDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "RebootDBInstanceResult": { - "base": null, - "refs": { - } - }, - "RecurringCharge": { - "base": null, - "refs": { - "RecurringChargeList$member": null - } - }, - "RecurringChargeList": { - "base": null, - "refs": { - "ReservedDBInstance$RecurringCharges": null, - "ReservedDBInstancesOffering$RecurringCharges": null - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "RemoveSourceIdentifierFromSubscriptionResult": { - "base": null, - "refs": { - } - }, - "RemoveTagsFromResourceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstance": { - "base": null, - "refs": { - "PurchaseReservedDBInstancesOfferingResult$ReservedDBInstance": null, - "ReservedDBInstanceList$member": null - } - }, - "ReservedDBInstanceAlreadyExistsFault": { - "base": "

User already has a reservation with the given identifier.

", - "refs": { - } - }, - "ReservedDBInstanceList": { - "base": null, - "refs": { - "ReservedDBInstanceMessage$ReservedDBInstances": null - } - }, - "ReservedDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstanceNotFoundFault": { - "base": "

The specified reserved DB Instance not found.

", - "refs": { - } - }, - "ReservedDBInstanceQuotaExceededFault": { - "base": "

Request would exceed the user's DB Instance quota.

", - "refs": { - } - }, - "ReservedDBInstancesOffering": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingList$member": null - } - }, - "ReservedDBInstancesOfferingList": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferings": null - } - }, - "ReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstancesOfferingNotFoundFault": { - "base": "

Specified offering does not exist.

", - "refs": { - } - }, - "ResetDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeResult": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "SNSInvalidTopicFault": { - "base": "

SNS has responded that there is a problem with the SND topic specified.

", - "refs": { - } - }, - "SNSNoAuthorizationFault": { - "base": "

You do not have permission to publish to the SNS topic ARN.

", - "refs": { - } - }, - "SNSTopicArnNotFoundFault": { - "base": "

The SNS topic ARN does not exist.

", - "refs": { - } - }, - "SnapshotQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB snapshots.

", - "refs": { - } - }, - "SourceIdsList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$SourceIds": null, - "EventSubscription$SourceIdsList": null - } - }, - "SourceNotFoundFault": { - "base": "

The requested source could not be found.

", - "refs": { - } - }, - "SourceType": { - "base": null, - "refs": { - "DescribeEventsMessage$SourceType": null, - "Event$SourceType": null - } - }, - "StorageQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed amount of storage available across all DB instances.

", - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionMessage$SubscriptionName": null, - "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier": null, - "AddTagsToResourceMessage$ResourceName": null, - "AuthorizeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$CIDRIP": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "AvailabilityZone$Name": null, - "CharacterSet$CharacterSetName": null, - "CharacterSet$CharacterSetDescription": null, - "CopyDBSnapshotMessage$SourceDBSnapshotIdentifier": null, - "CopyDBSnapshotMessage$TargetDBSnapshotIdentifier": null, - "CreateDBInstanceMessage$DBName": null, - "CreateDBInstanceMessage$DBInstanceIdentifier": null, - "CreateDBInstanceMessage$DBInstanceClass": null, - "CreateDBInstanceMessage$Engine": null, - "CreateDBInstanceMessage$MasterUsername": null, - "CreateDBInstanceMessage$MasterUserPassword": null, - "CreateDBInstanceMessage$AvailabilityZone": null, - "CreateDBInstanceMessage$DBSubnetGroupName": null, - "CreateDBInstanceMessage$PreferredMaintenanceWindow": null, - "CreateDBInstanceMessage$DBParameterGroupName": null, - "CreateDBInstanceMessage$PreferredBackupWindow": null, - "CreateDBInstanceMessage$EngineVersion": null, - "CreateDBInstanceMessage$LicenseModel": null, - "CreateDBInstanceMessage$OptionGroupName": null, - "CreateDBInstanceMessage$CharacterSetName": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$SourceDBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceClass": null, - "CreateDBInstanceReadReplicaMessage$AvailabilityZone": null, - "CreateDBInstanceReadReplicaMessage$OptionGroupName": null, - "CreateDBInstanceReadReplicaMessage$DBSubnetGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupFamily": null, - "CreateDBParameterGroupMessage$Description": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupName": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupDescription": null, - "CreateDBSnapshotMessage$DBSnapshotIdentifier": null, - "CreateDBSnapshotMessage$DBInstanceIdentifier": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupName": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "CreateEventSubscriptionMessage$SubscriptionName": null, - "CreateEventSubscriptionMessage$SnsTopicArn": null, - "CreateEventSubscriptionMessage$SourceType": null, - "CreateOptionGroupMessage$OptionGroupName": null, - "CreateOptionGroupMessage$EngineName": null, - "CreateOptionGroupMessage$MajorEngineVersion": null, - "CreateOptionGroupMessage$OptionGroupDescription": null, - "DBEngineVersion$Engine": null, - "DBEngineVersion$EngineVersion": null, - "DBEngineVersion$DBParameterGroupFamily": null, - "DBEngineVersion$DBEngineDescription": null, - "DBEngineVersion$DBEngineVersionDescription": null, - "DBEngineVersionMessage$Marker": null, - "DBInstance$DBInstanceIdentifier": null, - "DBInstance$DBInstanceClass": null, - "DBInstance$Engine": null, - "DBInstance$DBInstanceStatus": null, - "DBInstance$MasterUsername": null, - "DBInstance$DBName": null, - "DBInstance$PreferredBackupWindow": null, - "DBInstance$AvailabilityZone": null, - "DBInstance$PreferredMaintenanceWindow": null, - "DBInstance$EngineVersion": null, - "DBInstance$ReadReplicaSourceDBInstanceIdentifier": null, - "DBInstance$LicenseModel": null, - "DBInstance$CharacterSetName": null, - "DBInstance$SecondaryAvailabilityZone": null, - "DBInstanceMessage$Marker": null, - "DBInstanceStatusInfo$StatusType": null, - "DBInstanceStatusInfo$Status": null, - "DBInstanceStatusInfo$Message": null, - "DBParameterGroup$DBParameterGroupName": null, - "DBParameterGroup$DBParameterGroupFamily": null, - "DBParameterGroup$Description": null, - "DBParameterGroupDetails$Marker": null, - "DBParameterGroupNameMessage$DBParameterGroupName": null, - "DBParameterGroupStatus$DBParameterGroupName": null, - "DBParameterGroupStatus$ParameterApplyStatus": null, - "DBParameterGroupsMessage$Marker": null, - "DBSecurityGroup$OwnerId": null, - "DBSecurityGroup$DBSecurityGroupName": null, - "DBSecurityGroup$DBSecurityGroupDescription": null, - "DBSecurityGroup$VpcId": null, - "DBSecurityGroupMembership$DBSecurityGroupName": null, - "DBSecurityGroupMembership$Status": null, - "DBSecurityGroupMessage$Marker": null, - "DBSecurityGroupNameList$member": null, - "DBSnapshot$DBSnapshotIdentifier": null, - "DBSnapshot$DBInstanceIdentifier": null, - "DBSnapshot$Engine": null, - "DBSnapshot$Status": null, - "DBSnapshot$AvailabilityZone": null, - "DBSnapshot$VpcId": null, - "DBSnapshot$MasterUsername": null, - "DBSnapshot$EngineVersion": null, - "DBSnapshot$LicenseModel": null, - "DBSnapshot$SnapshotType": null, - "DBSnapshot$OptionGroupName": null, - "DBSnapshot$SourceRegion": null, - "DBSnapshotMessage$Marker": null, - "DBSubnetGroup$DBSubnetGroupName": null, - "DBSubnetGroup$DBSubnetGroupDescription": null, - "DBSubnetGroup$VpcId": null, - "DBSubnetGroup$SubnetGroupStatus": null, - "DBSubnetGroupMessage$Marker": null, - "DeleteDBInstanceMessage$DBInstanceIdentifier": null, - "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier": null, - "DeleteDBParameterGroupMessage$DBParameterGroupName": null, - "DeleteDBSecurityGroupMessage$DBSecurityGroupName": null, - "DeleteDBSnapshotMessage$DBSnapshotIdentifier": null, - "DeleteDBSubnetGroupMessage$DBSubnetGroupName": null, - "DeleteEventSubscriptionMessage$SubscriptionName": null, - "DeleteOptionGroupMessage$OptionGroupName": null, - "DescribeDBEngineVersionsMessage$Engine": null, - "DescribeDBEngineVersionsMessage$EngineVersion": null, - "DescribeDBEngineVersionsMessage$DBParameterGroupFamily": null, - "DescribeDBEngineVersionsMessage$Marker": null, - "DescribeDBInstancesMessage$DBInstanceIdentifier": null, - "DescribeDBInstancesMessage$Marker": null, - "DescribeDBLogFilesDetails$LogFileName": null, - "DescribeDBLogFilesMessage$DBInstanceIdentifier": null, - "DescribeDBLogFilesMessage$FilenameContains": null, - "DescribeDBLogFilesMessage$Marker": null, - "DescribeDBLogFilesResponse$Marker": null, - "DescribeDBParameterGroupsMessage$DBParameterGroupName": null, - "DescribeDBParameterGroupsMessage$Marker": null, - "DescribeDBParametersMessage$DBParameterGroupName": null, - "DescribeDBParametersMessage$Source": null, - "DescribeDBParametersMessage$Marker": null, - "DescribeDBSecurityGroupsMessage$DBSecurityGroupName": null, - "DescribeDBSecurityGroupsMessage$Marker": null, - "DescribeDBSnapshotsMessage$DBInstanceIdentifier": null, - "DescribeDBSnapshotsMessage$DBSnapshotIdentifier": null, - "DescribeDBSnapshotsMessage$SnapshotType": null, - "DescribeDBSnapshotsMessage$Marker": null, - "DescribeDBSubnetGroupsMessage$DBSubnetGroupName": null, - "DescribeDBSubnetGroupsMessage$Marker": null, - "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily": null, - "DescribeEngineDefaultParametersMessage$Marker": null, - "DescribeEventCategoriesMessage$SourceType": null, - "DescribeEventSubscriptionsMessage$SubscriptionName": null, - "DescribeEventSubscriptionsMessage$Marker": null, - "DescribeEventsMessage$SourceIdentifier": null, - "DescribeEventsMessage$Marker": null, - "DescribeOptionGroupOptionsMessage$EngineName": null, - "DescribeOptionGroupOptionsMessage$MajorEngineVersion": null, - "DescribeOptionGroupOptionsMessage$Marker": null, - "DescribeOptionGroupsMessage$OptionGroupName": null, - "DescribeOptionGroupsMessage$Marker": null, - "DescribeOptionGroupsMessage$EngineName": null, - "DescribeOptionGroupsMessage$MajorEngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$Engine": null, - "DescribeOrderableDBInstanceOptionsMessage$EngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass": null, - "DescribeOrderableDBInstanceOptionsMessage$LicenseModel": null, - "DescribeOrderableDBInstanceOptionsMessage$Marker": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstanceId": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesMessage$Duration": null, - "DescribeReservedDBInstancesMessage$ProductDescription": null, - "DescribeReservedDBInstancesMessage$OfferingType": null, - "DescribeReservedDBInstancesMessage$Marker": null, - "DescribeReservedDBInstancesOfferingsMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesOfferingsMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesOfferingsMessage$Duration": null, - "DescribeReservedDBInstancesOfferingsMessage$ProductDescription": null, - "DescribeReservedDBInstancesOfferingsMessage$OfferingType": null, - "DescribeReservedDBInstancesOfferingsMessage$Marker": null, - "DownloadDBLogFilePortionDetails$LogFileData": null, - "DownloadDBLogFilePortionDetails$Marker": null, - "DownloadDBLogFilePortionMessage$DBInstanceIdentifier": null, - "DownloadDBLogFilePortionMessage$LogFileName": null, - "DownloadDBLogFilePortionMessage$Marker": null, - "EC2SecurityGroup$Status": null, - "EC2SecurityGroup$EC2SecurityGroupName": null, - "EC2SecurityGroup$EC2SecurityGroupId": null, - "EC2SecurityGroup$EC2SecurityGroupOwnerId": null, - "Endpoint$Address": null, - "EngineDefaults$DBParameterGroupFamily": null, - "EngineDefaults$Marker": null, - "Event$SourceIdentifier": null, - "Event$Message": null, - "EventCategoriesList$member": null, - "EventCategoriesMap$SourceType": null, - "EventSubscription$CustomerAwsId": null, - "EventSubscription$CustSubscriptionId": null, - "EventSubscription$SnsTopicArn": null, - "EventSubscription$Status": null, - "EventSubscription$SubscriptionCreationTime": null, - "EventSubscription$SourceType": null, - "EventSubscriptionsMessage$Marker": null, - "EventsMessage$Marker": null, - "Filter$Name": null, - "FilterValueList$member": null, - "IPRange$Status": null, - "IPRange$CIDRIP": null, - "KeyList$member": null, - "ListTagsForResourceMessage$ResourceName": null, - "ModifyDBInstanceMessage$DBInstanceIdentifier": null, - "ModifyDBInstanceMessage$DBInstanceClass": null, - "ModifyDBInstanceMessage$MasterUserPassword": null, - "ModifyDBInstanceMessage$DBParameterGroupName": null, - "ModifyDBInstanceMessage$PreferredBackupWindow": null, - "ModifyDBInstanceMessage$PreferredMaintenanceWindow": null, - "ModifyDBInstanceMessage$EngineVersion": null, - "ModifyDBInstanceMessage$OptionGroupName": null, - "ModifyDBInstanceMessage$NewDBInstanceIdentifier": null, - "ModifyDBParameterGroupMessage$DBParameterGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "ModifyEventSubscriptionMessage$SubscriptionName": null, - "ModifyEventSubscriptionMessage$SnsTopicArn": null, - "ModifyEventSubscriptionMessage$SourceType": null, - "ModifyOptionGroupMessage$OptionGroupName": null, - "Option$OptionName": null, - "Option$OptionDescription": null, - "OptionConfiguration$OptionName": null, - "OptionGroup$OptionGroupName": null, - "OptionGroup$OptionGroupDescription": null, - "OptionGroup$EngineName": null, - "OptionGroup$MajorEngineVersion": null, - "OptionGroup$VpcId": null, - "OptionGroupMembership$OptionGroupName": null, - "OptionGroupMembership$Status": null, - "OptionGroupOption$Name": null, - "OptionGroupOption$Description": null, - "OptionGroupOption$EngineName": null, - "OptionGroupOption$MajorEngineVersion": null, - "OptionGroupOption$MinimumRequiredMinorEngineVersion": null, - "OptionGroupOptionSetting$SettingName": null, - "OptionGroupOptionSetting$SettingDescription": null, - "OptionGroupOptionSetting$DefaultValue": null, - "OptionGroupOptionSetting$ApplyType": null, - "OptionGroupOptionSetting$AllowedValues": null, - "OptionGroupOptionsMessage$Marker": null, - "OptionGroups$Marker": null, - "OptionNamesList$member": null, - "OptionSetting$Name": null, - "OptionSetting$Value": null, - "OptionSetting$DefaultValue": null, - "OptionSetting$Description": null, - "OptionSetting$ApplyType": null, - "OptionSetting$DataType": null, - "OptionSetting$AllowedValues": null, - "OptionsDependedOn$member": null, - "OrderableDBInstanceOption$Engine": null, - "OrderableDBInstanceOption$EngineVersion": null, - "OrderableDBInstanceOption$DBInstanceClass": null, - "OrderableDBInstanceOption$LicenseModel": null, - "OrderableDBInstanceOptionsMessage$Marker": null, - "Parameter$ParameterName": null, - "Parameter$ParameterValue": null, - "Parameter$Description": null, - "Parameter$Source": null, - "Parameter$ApplyType": null, - "Parameter$DataType": null, - "Parameter$AllowedValues": null, - "Parameter$MinimumEngineVersion": null, - "PendingModifiedValues$DBInstanceClass": null, - "PendingModifiedValues$MasterUserPassword": null, - "PendingModifiedValues$EngineVersion": null, - "PendingModifiedValues$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$PreferredBackupWindow": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferingId": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstanceId": null, - "ReadReplicaDBInstanceIdentifierList$member": null, - "RebootDBInstanceMessage$DBInstanceIdentifier": null, - "RecurringCharge$RecurringChargeFrequency": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier": null, - "RemoveTagsFromResourceMessage$ResourceName": null, - "ReservedDBInstance$ReservedDBInstanceId": null, - "ReservedDBInstance$ReservedDBInstancesOfferingId": null, - "ReservedDBInstance$DBInstanceClass": null, - "ReservedDBInstance$CurrencyCode": null, - "ReservedDBInstance$ProductDescription": null, - "ReservedDBInstance$OfferingType": null, - "ReservedDBInstance$State": null, - "ReservedDBInstanceMessage$Marker": null, - "ReservedDBInstancesOffering$ReservedDBInstancesOfferingId": null, - "ReservedDBInstancesOffering$DBInstanceClass": null, - "ReservedDBInstancesOffering$CurrencyCode": null, - "ReservedDBInstancesOffering$ProductDescription": null, - "ReservedDBInstancesOffering$OfferingType": null, - "ReservedDBInstancesOfferingMessage$Marker": null, - "ResetDBParameterGroupMessage$DBParameterGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSnapshotIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceClass": null, - "RestoreDBInstanceFromDBSnapshotMessage$AvailabilityZone": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSubnetGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$LicenseModel": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBName": null, - "RestoreDBInstanceFromDBSnapshotMessage$Engine": null, - "RestoreDBInstanceFromDBSnapshotMessage$OptionGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$SourceDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$TargetDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$DBInstanceClass": null, - "RestoreDBInstanceToPointInTimeMessage$AvailabilityZone": null, - "RestoreDBInstanceToPointInTimeMessage$DBSubnetGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$LicenseModel": null, - "RestoreDBInstanceToPointInTimeMessage$DBName": null, - "RestoreDBInstanceToPointInTimeMessage$Engine": null, - "RestoreDBInstanceToPointInTimeMessage$OptionGroupName": null, - "RevokeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$CIDRIP": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "SourceIdsList$member": null, - "Subnet$SubnetIdentifier": null, - "Subnet$SubnetStatus": null, - "SubnetIdentifierList$member": null, - "Tag$Key": null, - "Tag$Value": null, - "VpcSecurityGroupIdList$member": null, - "VpcSecurityGroupMembership$VpcSecurityGroupId": null, - "VpcSecurityGroupMembership$Status": null - } - }, - "Subnet": { - "base": null, - "refs": { - "SubnetList$member": null - } - }, - "SubnetAlreadyInUse": { - "base": "

The DB subnet is already in use in the Availability Zone.

", - "refs": { - } - }, - "SubnetIdentifierList": { - "base": null, - "refs": { - "CreateDBSubnetGroupMessage$SubnetIds": null, - "ModifyDBSubnetGroupMessage$SubnetIds": null - } - }, - "SubnetList": { - "base": null, - "refs": { - "DBSubnetGroup$Subnets": null - } - }, - "SubscriptionAlreadyExistFault": { - "base": "

The supplied subscription name already exists.

", - "refs": { - } - }, - "SubscriptionCategoryNotFoundFault": { - "base": "

The supplied category does not exist.

", - "refs": { - } - }, - "SubscriptionNotFoundFault": { - "base": "

The subscription name does not exist.

", - "refs": { - } - }, - "SupportedCharacterSetsList": { - "base": null, - "refs": { - "DBEngineVersion$SupportedCharacterSets": null - } - }, - "TStamp": { - "base": null, - "refs": { - "DBInstance$InstanceCreateTime": null, - "DBInstance$LatestRestorableTime": null, - "DBSnapshot$SnapshotCreateTime": null, - "DBSnapshot$InstanceCreateTime": null, - "DescribeEventsMessage$StartTime": null, - "DescribeEventsMessage$EndTime": null, - "Event$Date": null, - "ReservedDBInstance$StartTime": null, - "RestoreDBInstanceToPointInTimeMessage$RestoreTime": null - } - }, - "Tag": { - "base": null, - "refs": { - "TagList$member": null - } - }, - "TagList": { - "base": null, - "refs": { - "AddTagsToResourceMessage$Tags": null, - "CopyDBSnapshotMessage$Tags": null, - "CreateDBInstanceMessage$Tags": null, - "CreateDBInstanceReadReplicaMessage$Tags": null, - "CreateDBParameterGroupMessage$Tags": null, - "CreateDBSecurityGroupMessage$Tags": null, - "CreateDBSnapshotMessage$Tags": null, - "CreateDBSubnetGroupMessage$Tags": null, - "CreateEventSubscriptionMessage$Tags": null, - "CreateOptionGroupMessage$Tags": null, - "PurchaseReservedDBInstancesOfferingMessage$Tags": null, - "RestoreDBInstanceFromDBSnapshotMessage$Tags": null, - "RestoreDBInstanceToPointInTimeMessage$Tags": null, - "TagListMessage$TagList": null - } - }, - "TagListMessage": { - "base": null, - "refs": { - } - }, - "VpcSecurityGroupIdList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$VpcSecurityGroupIds": null, - "ModifyDBInstanceMessage$VpcSecurityGroupIds": null, - "OptionConfiguration$VpcSecurityGroupMemberships": null - } - }, - "VpcSecurityGroupMembership": { - "base": null, - "refs": { - "VpcSecurityGroupMembershipList$member": null - } - }, - "VpcSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$VpcSecurityGroups": null, - "Option$VpcSecurityGroupMemberships": null - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/paginators-1.json deleted file mode 100644 index c51d8d15b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/paginators-1.json +++ /dev/null @@ -1,110 +0,0 @@ -{ - "pagination": { - "DescribeDBEngineVersions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBEngineVersions" - }, - "DescribeDBInstances": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBInstances" - }, - "DescribeDBLogFiles": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DescribeDBLogFiles" - }, - "DescribeDBParameterGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBParameterGroups" - }, - "DescribeDBParameters": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "Parameters" - }, - "DescribeDBSecurityGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSecurityGroups" - }, - "DescribeDBSnapshots": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSnapshots" - }, - "DescribeDBSubnetGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "DBSubnetGroups" - }, - "DescribeEngineDefaultParameters": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "EngineDefaults.Marker", - "result_key": "EngineDefaults.Parameters" - }, - "DescribeEventSubscriptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "EventSubscriptionsList" - }, - "DescribeEvents": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "Events" - }, - "DescribeOptionGroupOptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OptionGroupOptions" - }, - "DescribeOptionGroups": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OptionGroupsList" - }, - "DescribeOrderableDBInstanceOptions": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "OrderableDBInstanceOptions" - }, - "DescribeReservedDBInstances": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "ReservedDBInstances" - }, - "DescribeReservedDBInstancesOfferings": { - "input_token": "Marker", - "limit_key": "MaxRecords", - "output_token": "Marker", - "result_key": "ReservedDBInstancesOfferings" - }, - "DownloadDBLogFilePortion": { - "input_token": "Marker", - "limit_key": "NumberOfLines", - "more_results": "AdditionalDataPending", - "output_token": "Marker", - "result_key": "LogFileData" - }, - "ListTagsForResource": { - "result_key": "TagList" - } - } -} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/docs-2.json deleted file mode 100644 index 0cce34ce2..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/docs-2.json +++ /dev/null @@ -1,1932 +0,0 @@ -{ - "version": "2.0", - "service": null, - "operations": { - "AddSourceIdentifierToSubscription": null, - "AddTagsToResource": null, - "AuthorizeDBSecurityGroupIngress": null, - "CopyDBParameterGroup": null, - "CopyDBSnapshot": null, - "CopyOptionGroup": null, - "CreateDBInstance": null, - "CreateDBInstanceReadReplica": null, - "CreateDBParameterGroup": null, - "CreateDBSecurityGroup": null, - "CreateDBSnapshot": null, - "CreateDBSubnetGroup": null, - "CreateEventSubscription": null, - "CreateOptionGroup": null, - "DeleteDBInstance": null, - "DeleteDBParameterGroup": null, - "DeleteDBSecurityGroup": null, - "DeleteDBSnapshot": null, - "DeleteDBSubnetGroup": null, - "DeleteEventSubscription": null, - "DeleteOptionGroup": null, - "DescribeDBEngineVersions": null, - "DescribeDBInstances": null, - "DescribeDBLogFiles": null, - "DescribeDBParameterGroups": null, - "DescribeDBParameters": null, - "DescribeDBSecurityGroups": null, - "DescribeDBSnapshots": null, - "DescribeDBSubnetGroups": null, - "DescribeEngineDefaultParameters": null, - "DescribeEventCategories": null, - "DescribeEventSubscriptions": null, - "DescribeEvents": null, - "DescribeOptionGroupOptions": null, - "DescribeOptionGroups": null, - "DescribeOrderableDBInstanceOptions": null, - "DescribeReservedDBInstances": null, - "DescribeReservedDBInstancesOfferings": null, - "DownloadDBLogFilePortion": null, - "ListTagsForResource": null, - "ModifyDBInstance": null, - "ModifyDBParameterGroup": null, - "ModifyDBSubnetGroup": null, - "ModifyEventSubscription": null, - "ModifyOptionGroup": null, - "PromoteReadReplica": null, - "PurchaseReservedDBInstancesOffering": null, - "RebootDBInstance": null, - "RemoveSourceIdentifierFromSubscription": null, - "RemoveTagsFromResource": null, - "ResetDBParameterGroup": null, - "RestoreDBInstanceFromDBSnapshot": null, - "RestoreDBInstanceToPointInTime": null, - "RevokeDBSecurityGroupIngress": null - }, - "shapes": { - "AddSourceIdentifierToSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "AddSourceIdentifierToSubscriptionResult": { - "base": null, - "refs": { - } - }, - "AddTagsToResourceMessage": { - "base": null, - "refs": { - } - }, - "ApplyMethod": { - "base": null, - "refs": { - "Parameter$ApplyMethod": null - } - }, - "AuthorizationAlreadyExistsFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group is already authorized for the specified DB security group.

", - "refs": { - } - }, - "AuthorizationNotFoundFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group isn't authorized for the specified DB security group.

RDS also may not be authorized by using IAM to perform necessary actions on your behalf.

", - "refs": { - } - }, - "AuthorizationQuotaExceededFault": { - "base": "

The DB security group authorization quota has been reached.

", - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "AvailabilityZone": { - "base": null, - "refs": { - "AvailabilityZoneList$member": null, - "Subnet$SubnetAvailabilityZone": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "OrderableDBInstanceOption$AvailabilityZones": null - } - }, - "Boolean": { - "base": null, - "refs": { - "DBInstance$MultiAZ": null, - "DBInstance$AutoMinorVersionUpgrade": null, - "DBInstance$PubliclyAccessible": null, - "DBInstanceStatusInfo$Normal": null, - "DeleteDBInstanceMessage$SkipFinalSnapshot": null, - "DescribeDBEngineVersionsMessage$DefaultOnly": null, - "DownloadDBLogFilePortionDetails$AdditionalDataPending": null, - "EventSubscription$Enabled": null, - "ModifyDBInstanceMessage$ApplyImmediately": null, - "ModifyDBInstanceMessage$AllowMajorVersionUpgrade": null, - "ModifyOptionGroupMessage$ApplyImmediately": null, - "Option$Persistent": null, - "Option$Permanent": null, - "OptionGroup$AllowsVpcAndNonVpcInstanceMemberships": null, - "OptionGroupOption$PortRequired": null, - "OptionGroupOption$Persistent": null, - "OptionGroupOption$Permanent": null, - "OptionGroupOptionSetting$IsModifiable": null, - "OptionSetting$IsModifiable": null, - "OptionSetting$IsCollection": null, - "OrderableDBInstanceOption$MultiAZCapable": null, - "OrderableDBInstanceOption$ReadReplicaCapable": null, - "OrderableDBInstanceOption$Vpc": null, - "OrderableDBInstanceOption$SupportsIops": null, - "Parameter$IsModifiable": null, - "ReservedDBInstance$MultiAZ": null, - "ReservedDBInstancesOffering$MultiAZ": null, - "ResetDBParameterGroupMessage$ResetAllParameters": null, - "RestoreDBInstanceToPointInTimeMessage$UseLatestRestorableTime": null - } - }, - "BooleanOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$MultiAZ": null, - "CreateDBInstanceMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceMessage$PubliclyAccessible": null, - "CreateDBInstanceReadReplicaMessage$AutoMinorVersionUpgrade": null, - "CreateDBInstanceReadReplicaMessage$PubliclyAccessible": null, - "CreateEventSubscriptionMessage$Enabled": null, - "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets": null, - "DescribeOrderableDBInstanceOptionsMessage$Vpc": null, - "DescribeReservedDBInstancesMessage$MultiAZ": null, - "DescribeReservedDBInstancesOfferingsMessage$MultiAZ": null, - "ModifyDBInstanceMessage$MultiAZ": null, - "ModifyDBInstanceMessage$AutoMinorVersionUpgrade": null, - "ModifyEventSubscriptionMessage$Enabled": null, - "PendingModifiedValues$MultiAZ": null, - "RebootDBInstanceMessage$ForceFailover": null, - "RestoreDBInstanceFromDBSnapshotMessage$MultiAZ": null, - "RestoreDBInstanceFromDBSnapshotMessage$PubliclyAccessible": null, - "RestoreDBInstanceFromDBSnapshotMessage$AutoMinorVersionUpgrade": null, - "RestoreDBInstanceToPointInTimeMessage$MultiAZ": null, - "RestoreDBInstanceToPointInTimeMessage$PubliclyAccessible": null, - "RestoreDBInstanceToPointInTimeMessage$AutoMinorVersionUpgrade": null - } - }, - "CharacterSet": { - "base": null, - "refs": { - "DBEngineVersion$DefaultCharacterSet": null, - "SupportedCharacterSetsList$member": null - } - }, - "CopyDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "CopyDBParameterGroupResult": { - "base": null, - "refs": { - } - }, - "CopyDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CopyDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CopyOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "CopyOptionGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceReadReplicaResult": { - "base": null, - "refs": { - } - }, - "CreateDBInstanceResult": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBParameterGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSecurityGroupResult": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "CreateEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "CreateOptionGroupResult": { - "base": null, - "refs": { - } - }, - "DBEngineVersion": { - "base": null, - "refs": { - "DBEngineVersionList$member": null - } - }, - "DBEngineVersionList": { - "base": null, - "refs": { - "DBEngineVersionMessage$DBEngineVersions": null - } - }, - "DBEngineVersionMessage": { - "base": null, - "refs": { - } - }, - "DBInstance": { - "base": null, - "refs": { - "CreateDBInstanceReadReplicaResult$DBInstance": null, - "CreateDBInstanceResult$DBInstance": null, - "DBInstanceList$member": null, - "DeleteDBInstanceResult$DBInstance": null, - "ModifyDBInstanceResult$DBInstance": null, - "PromoteReadReplicaResult$DBInstance": null, - "RebootDBInstanceResult$DBInstance": null, - "RestoreDBInstanceFromDBSnapshotResult$DBInstance": null, - "RestoreDBInstanceToPointInTimeResult$DBInstance": null - } - }, - "DBInstanceAlreadyExistsFault": { - "base": "

The user already has a DB instance with the given identifier.

", - "refs": { - } - }, - "DBInstanceList": { - "base": null, - "refs": { - "DBInstanceMessage$DBInstances": null - } - }, - "DBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DBInstanceNotFoundFault": { - "base": "

DBInstanceIdentifier doesn't refer to an existing DB instance.

", - "refs": { - } - }, - "DBInstanceStatusInfo": { - "base": null, - "refs": { - "DBInstanceStatusInfoList$member": null - } - }, - "DBInstanceStatusInfoList": { - "base": null, - "refs": { - "DBInstance$StatusInfos": null - } - }, - "DBLogFileNotFoundFault": { - "base": "

LogFileName doesn't refer to an existing DB log file.

", - "refs": { - } - }, - "DBParameterGroup": { - "base": null, - "refs": { - "CopyDBParameterGroupResult$DBParameterGroup": null, - "CreateDBParameterGroupResult$DBParameterGroup": null, - "DBParameterGroupList$member": null - } - }, - "DBParameterGroupAlreadyExistsFault": { - "base": "

A DB parameter group with the same name exists.

", - "refs": { - } - }, - "DBParameterGroupDetails": { - "base": null, - "refs": { - } - }, - "DBParameterGroupList": { - "base": null, - "refs": { - "DBParameterGroupsMessage$DBParameterGroups": null - } - }, - "DBParameterGroupNameMessage": { - "base": null, - "refs": { - } - }, - "DBParameterGroupNotFoundFault": { - "base": "

DBParameterGroupName doesn't refer to an existing DB parameter group.

", - "refs": { - } - }, - "DBParameterGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB parameter groups.

", - "refs": { - } - }, - "DBParameterGroupStatus": { - "base": null, - "refs": { - "DBParameterGroupStatusList$member": null - } - }, - "DBParameterGroupStatusList": { - "base": null, - "refs": { - "DBInstance$DBParameterGroups": null - } - }, - "DBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroup": { - "base": null, - "refs": { - "AuthorizeDBSecurityGroupIngressResult$DBSecurityGroup": null, - "CreateDBSecurityGroupResult$DBSecurityGroup": null, - "DBSecurityGroups$member": null, - "RevokeDBSecurityGroupIngressResult$DBSecurityGroup": null - } - }, - "DBSecurityGroupAlreadyExistsFault": { - "base": "

A DB security group with the name specified in DBSecurityGroupName already exists.

", - "refs": { - } - }, - "DBSecurityGroupMembership": { - "base": null, - "refs": { - "DBSecurityGroupMembershipList$member": null - } - }, - "DBSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$DBSecurityGroups": null, - "Option$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSecurityGroupNameList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$DBSecurityGroups": null, - "ModifyDBInstanceMessage$DBSecurityGroups": null, - "OptionConfiguration$DBSecurityGroupMemberships": null - } - }, - "DBSecurityGroupNotFoundFault": { - "base": "

DBSecurityGroupName doesn't refer to an existing DB security group.

", - "refs": { - } - }, - "DBSecurityGroupNotSupportedFault": { - "base": "

A DB security group isn't allowed for this action.

", - "refs": { - } - }, - "DBSecurityGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB security groups.

", - "refs": { - } - }, - "DBSecurityGroups": { - "base": null, - "refs": { - "DBSecurityGroupMessage$DBSecurityGroups": null - } - }, - "DBSnapshot": { - "base": null, - "refs": { - "CopyDBSnapshotResult$DBSnapshot": null, - "CreateDBSnapshotResult$DBSnapshot": null, - "DBSnapshotList$member": null, - "DeleteDBSnapshotResult$DBSnapshot": null - } - }, - "DBSnapshotAlreadyExistsFault": { - "base": "

DBSnapshotIdentifier is already used by an existing snapshot.

", - "refs": { - } - }, - "DBSnapshotList": { - "base": null, - "refs": { - "DBSnapshotMessage$DBSnapshots": null - } - }, - "DBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DBSnapshotNotFoundFault": { - "base": "

DBSnapshotIdentifier doesn't refer to an existing DB snapshot.

", - "refs": { - } - }, - "DBSubnetGroup": { - "base": null, - "refs": { - "CreateDBSubnetGroupResult$DBSubnetGroup": null, - "DBInstance$DBSubnetGroup": null, - "DBSubnetGroups$member": null, - "ModifyDBSubnetGroupResult$DBSubnetGroup": null - } - }, - "DBSubnetGroupAlreadyExistsFault": { - "base": "

DBSubnetGroupName is already used by an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupDoesNotCoverEnoughAZs": { - "base": "

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", - "refs": { - } - }, - "DBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DBSubnetGroupNotAllowedFault": { - "base": "

The DBSubnetGroup shouldn't be specified while creating read replicas that lie in the same region as the source instance.

", - "refs": { - } - }, - "DBSubnetGroupNotFoundFault": { - "base": "

DBSubnetGroupName doesn't refer to an existing DB subnet group.

", - "refs": { - } - }, - "DBSubnetGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB subnet groups.

", - "refs": { - } - }, - "DBSubnetGroups": { - "base": null, - "refs": { - "DBSubnetGroupMessage$DBSubnetGroups": null - } - }, - "DBSubnetQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of subnets in a DB subnet groups.

", - "refs": { - } - }, - "DBUpgradeDependencyFailureFault": { - "base": "

The DB upgrade failed because a resource the DB depends on can't be modified.

", - "refs": { - } - }, - "DeleteDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBInstanceResult": { - "base": null, - "refs": { - } - }, - "DeleteDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSecurityGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "DeleteDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "DeleteDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "DeleteEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "DeleteOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBEngineVersionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBLogFilesDetails": { - "base": null, - "refs": { - "DescribeDBLogFilesList$member": null - } - }, - "DescribeDBLogFilesList": { - "base": null, - "refs": { - "DescribeDBLogFilesResponse$DescribeDBLogFiles": null - } - }, - "DescribeDBLogFilesMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBLogFilesResponse": { - "base": null, - "refs": { - } - }, - "DescribeDBParameterGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSecurityGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSnapshotsMessage": { - "base": null, - "refs": { - } - }, - "DescribeDBSubnetGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersMessage": { - "base": null, - "refs": { - } - }, - "DescribeEngineDefaultParametersResult": { - "base": null, - "refs": { - } - }, - "DescribeEventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeEventsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOptionGroupsMessage": { - "base": null, - "refs": { - } - }, - "DescribeOrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesMessage": { - "base": null, - "refs": { - } - }, - "DescribeReservedDBInstancesOfferingsMessage": { - "base": null, - "refs": { - } - }, - "Double": { - "base": null, - "refs": { - "RecurringCharge$RecurringChargeAmount": null, - "ReservedDBInstance$FixedPrice": null, - "ReservedDBInstance$UsagePrice": null, - "ReservedDBInstancesOffering$FixedPrice": null, - "ReservedDBInstancesOffering$UsagePrice": null - } - }, - "DownloadDBLogFilePortionDetails": { - "base": null, - "refs": { - } - }, - "DownloadDBLogFilePortionMessage": { - "base": null, - "refs": { - } - }, - "EC2SecurityGroup": { - "base": null, - "refs": { - "EC2SecurityGroupList$member": null - } - }, - "EC2SecurityGroupList": { - "base": null, - "refs": { - "DBSecurityGroup$EC2SecurityGroups": null - } - }, - "Endpoint": { - "base": null, - "refs": { - "DBInstance$Endpoint": null - } - }, - "EngineDefaults": { - "base": null, - "refs": { - "DescribeEngineDefaultParametersResult$EngineDefaults": null - } - }, - "Event": { - "base": null, - "refs": { - "EventList$member": null - } - }, - "EventCategoriesList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$EventCategories": null, - "DescribeEventsMessage$EventCategories": null, - "Event$EventCategories": null, - "EventCategoriesMap$EventCategories": null, - "EventSubscription$EventCategoriesList": null, - "ModifyEventSubscriptionMessage$EventCategories": null - } - }, - "EventCategoriesMap": { - "base": null, - "refs": { - "EventCategoriesMapList$member": null - } - }, - "EventCategoriesMapList": { - "base": null, - "refs": { - "EventCategoriesMessage$EventCategoriesMapList": null - } - }, - "EventCategoriesMessage": { - "base": null, - "refs": { - } - }, - "EventList": { - "base": null, - "refs": { - "EventsMessage$Events": null - } - }, - "EventSubscription": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionResult$EventSubscription": null, - "CreateEventSubscriptionResult$EventSubscription": null, - "DeleteEventSubscriptionResult$EventSubscription": null, - "EventSubscriptionsList$member": null, - "ModifyEventSubscriptionResult$EventSubscription": null, - "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription": null - } - }, - "EventSubscriptionQuotaExceededFault": { - "base": "

You have reached the maximum number of event subscriptions.

", - "refs": { - } - }, - "EventSubscriptionsList": { - "base": null, - "refs": { - "EventSubscriptionsMessage$EventSubscriptionsList": null - } - }, - "EventSubscriptionsMessage": { - "base": null, - "refs": { - } - }, - "EventsMessage": { - "base": null, - "refs": { - } - }, - "Filter": { - "base": null, - "refs": { - "FilterList$member": null - } - }, - "FilterList": { - "base": null, - "refs": { - "DescribeDBEngineVersionsMessage$Filters": null, - "DescribeDBInstancesMessage$Filters": null, - "DescribeDBLogFilesMessage$Filters": null, - "DescribeDBParameterGroupsMessage$Filters": null, - "DescribeDBParametersMessage$Filters": null, - "DescribeDBSecurityGroupsMessage$Filters": null, - "DescribeDBSnapshotsMessage$Filters": null, - "DescribeDBSubnetGroupsMessage$Filters": null, - "DescribeEngineDefaultParametersMessage$Filters": null, - "DescribeEventCategoriesMessage$Filters": null, - "DescribeEventSubscriptionsMessage$Filters": null, - "DescribeEventsMessage$Filters": null, - "DescribeOptionGroupOptionsMessage$Filters": null, - "DescribeOptionGroupsMessage$Filters": null, - "DescribeOrderableDBInstanceOptionsMessage$Filters": null, - "DescribeReservedDBInstancesMessage$Filters": null, - "DescribeReservedDBInstancesOfferingsMessage$Filters": null, - "ListTagsForResourceMessage$Filters": null - } - }, - "FilterValueList": { - "base": null, - "refs": { - "Filter$Values": null - } - }, - "IPRange": { - "base": null, - "refs": { - "IPRangeList$member": null - } - }, - "IPRangeList": { - "base": null, - "refs": { - "DBSecurityGroup$IPRanges": null - } - }, - "InstanceQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB instances.

", - "refs": { - } - }, - "InsufficientDBInstanceCapacityFault": { - "base": "

The specified DB instance class isn't available in the specified Availability Zone.

", - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "DBInstance$AllocatedStorage": null, - "DBInstance$BackupRetentionPeriod": null, - "DBSnapshot$AllocatedStorage": null, - "DBSnapshot$Port": null, - "DBSnapshot$PercentProgress": null, - "DownloadDBLogFilePortionMessage$NumberOfLines": null, - "Endpoint$Port": null, - "ReservedDBInstance$Duration": null, - "ReservedDBInstance$DBInstanceCount": null, - "ReservedDBInstancesOffering$Duration": null - } - }, - "IntegerOptional": { - "base": null, - "refs": { - "CreateDBInstanceMessage$AllocatedStorage": null, - "CreateDBInstanceMessage$BackupRetentionPeriod": null, - "CreateDBInstanceMessage$Port": null, - "CreateDBInstanceMessage$Iops": null, - "CreateDBInstanceReadReplicaMessage$Port": null, - "CreateDBInstanceReadReplicaMessage$Iops": null, - "DBInstance$Iops": null, - "DBSnapshot$Iops": null, - "DescribeDBEngineVersionsMessage$MaxRecords": null, - "DescribeDBInstancesMessage$MaxRecords": null, - "DescribeDBLogFilesMessage$MaxRecords": null, - "DescribeDBParameterGroupsMessage$MaxRecords": null, - "DescribeDBParametersMessage$MaxRecords": null, - "DescribeDBSecurityGroupsMessage$MaxRecords": null, - "DescribeDBSnapshotsMessage$MaxRecords": null, - "DescribeDBSubnetGroupsMessage$MaxRecords": null, - "DescribeEngineDefaultParametersMessage$MaxRecords": null, - "DescribeEventSubscriptionsMessage$MaxRecords": null, - "DescribeEventsMessage$Duration": null, - "DescribeEventsMessage$MaxRecords": null, - "DescribeOptionGroupOptionsMessage$MaxRecords": null, - "DescribeOptionGroupsMessage$MaxRecords": null, - "DescribeOrderableDBInstanceOptionsMessage$MaxRecords": null, - "DescribeReservedDBInstancesMessage$MaxRecords": null, - "DescribeReservedDBInstancesOfferingsMessage$MaxRecords": null, - "ModifyDBInstanceMessage$AllocatedStorage": null, - "ModifyDBInstanceMessage$BackupRetentionPeriod": null, - "ModifyDBInstanceMessage$Iops": null, - "Option$Port": null, - "OptionConfiguration$Port": null, - "OptionGroupOption$DefaultPort": null, - "PendingModifiedValues$AllocatedStorage": null, - "PendingModifiedValues$Port": null, - "PendingModifiedValues$BackupRetentionPeriod": null, - "PendingModifiedValues$Iops": null, - "PromoteReadReplicaMessage$BackupRetentionPeriod": null, - "PurchaseReservedDBInstancesOfferingMessage$DBInstanceCount": null, - "RestoreDBInstanceFromDBSnapshotMessage$Port": null, - "RestoreDBInstanceFromDBSnapshotMessage$Iops": null, - "RestoreDBInstanceToPointInTimeMessage$Port": null, - "RestoreDBInstanceToPointInTimeMessage$Iops": null - } - }, - "InvalidDBInstanceStateFault": { - "base": "

The specified DB instance isn't in the available state.

", - "refs": { - } - }, - "InvalidDBParameterGroupStateFault": { - "base": "

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you can't delete it when the parameter group is in this state.

", - "refs": { - } - }, - "InvalidDBSecurityGroupStateFault": { - "base": "

The state of the DB security group doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSnapshotStateFault": { - "base": "

The state of the DB snapshot doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSubnetGroupFault": { - "base": "

The DBSubnetGroup doesn't belong to the same VPC as that of an existing cross-region read replica of the same source instance.

", - "refs": { - } - }, - "InvalidDBSubnetGroupStateFault": { - "base": "

The DB subnet group cannot be deleted because it's in use.

", - "refs": { - } - }, - "InvalidDBSubnetStateFault": { - "base": "

The DB subnet isn't in the available state.

", - "refs": { - } - }, - "InvalidEventSubscriptionStateFault": { - "base": "

This error can occur if someone else is modifying a subscription. You should retry the action.

", - "refs": { - } - }, - "InvalidOptionGroupStateFault": { - "base": "

The option group isn't in the available state.

", - "refs": { - } - }, - "InvalidRestoreFault": { - "base": "

Cannot restore from VPC backup to non-VPC DB instance.

", - "refs": { - } - }, - "InvalidSubnet": { - "base": "

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", - "refs": { - } - }, - "InvalidVPCNetworkStateFault": { - "base": "

The DB subnet group doesn't cover all Availability Zones after it's created because of users' change.

", - "refs": { - } - }, - "KeyList": { - "base": null, - "refs": { - "RemoveTagsFromResourceMessage$TagKeys": null - } - }, - "ListTagsForResourceMessage": { - "base": null, - "refs": { - } - }, - "Long": { - "base": null, - "refs": { - "DescribeDBLogFilesDetails$LastWritten": null, - "DescribeDBLogFilesDetails$Size": null, - "DescribeDBLogFilesMessage$FileLastWritten": null, - "DescribeDBLogFilesMessage$FileSize": null - } - }, - "ModifyDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBInstanceResult": { - "base": null, - "refs": { - } - }, - "ModifyDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyDBSubnetGroupResult": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "ModifyEventSubscriptionResult": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupMessage": { - "base": null, - "refs": { - } - }, - "ModifyOptionGroupResult": { - "base": null, - "refs": { - } - }, - "Option": { - "base": null, - "refs": { - "OptionsList$member": null - } - }, - "OptionConfiguration": { - "base": null, - "refs": { - "OptionConfigurationList$member": null - } - }, - "OptionConfigurationList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToInclude": null - } - }, - "OptionGroup": { - "base": null, - "refs": { - "CopyOptionGroupResult$OptionGroup": null, - "CreateOptionGroupResult$OptionGroup": null, - "ModifyOptionGroupResult$OptionGroup": null, - "OptionGroupsList$member": null - } - }, - "OptionGroupAlreadyExistsFault": { - "base": "

The option group you are trying to create already exists.

", - "refs": { - } - }, - "OptionGroupMembership": { - "base": null, - "refs": { - "OptionGroupMembershipList$member": null - } - }, - "OptionGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$OptionGroupMemberships": null - } - }, - "OptionGroupNotFoundFault": { - "base": "

The specified option group could not be found.

", - "refs": { - } - }, - "OptionGroupOption": { - "base": null, - "refs": { - "OptionGroupOptionsList$member": null - } - }, - "OptionGroupOptionSetting": { - "base": null, - "refs": { - "OptionGroupOptionSettingsList$member": null - } - }, - "OptionGroupOptionSettingsList": { - "base": null, - "refs": { - "OptionGroupOption$OptionGroupOptionSettings": null - } - }, - "OptionGroupOptionsList": { - "base": null, - "refs": { - "OptionGroupOptionsMessage$OptionGroupOptions": null - } - }, - "OptionGroupOptionsMessage": { - "base": null, - "refs": { - } - }, - "OptionGroupQuotaExceededFault": { - "base": "

The quota of 20 option groups was exceeded for this AWS account.

", - "refs": { - } - }, - "OptionGroups": { - "base": null, - "refs": { - } - }, - "OptionGroupsList": { - "base": null, - "refs": { - "OptionGroups$OptionGroupsList": null - } - }, - "OptionNamesList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToRemove": null - } - }, - "OptionSetting": { - "base": null, - "refs": { - "OptionSettingConfigurationList$member": null, - "OptionSettingsList$member": null - } - }, - "OptionSettingConfigurationList": { - "base": null, - "refs": { - "Option$OptionSettings": null - } - }, - "OptionSettingsList": { - "base": null, - "refs": { - "OptionConfiguration$OptionSettings": null - } - }, - "OptionsDependedOn": { - "base": null, - "refs": { - "OptionGroupOption$OptionsDependedOn": null - } - }, - "OptionsList": { - "base": null, - "refs": { - "OptionGroup$Options": null - } - }, - "OrderableDBInstanceOption": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsList$member": null - } - }, - "OrderableDBInstanceOptionsList": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions": null - } - }, - "OrderableDBInstanceOptionsMessage": { - "base": null, - "refs": { - } - }, - "Parameter": { - "base": null, - "refs": { - "ParametersList$member": null - } - }, - "ParametersList": { - "base": null, - "refs": { - "DBParameterGroupDetails$Parameters": null, - "EngineDefaults$Parameters": null, - "ModifyDBParameterGroupMessage$Parameters": null, - "ResetDBParameterGroupMessage$Parameters": null - } - }, - "PendingModifiedValues": { - "base": null, - "refs": { - "DBInstance$PendingModifiedValues": null - } - }, - "PointInTimeRestoreNotEnabledFault": { - "base": "

SourceDBInstanceIdentifier refers to a DB instance with BackupRetentionPeriod equal to 0.

", - "refs": { - } - }, - "PromoteReadReplicaMessage": { - "base": null, - "refs": { - } - }, - "PromoteReadReplicaResult": { - "base": null, - "refs": { - } - }, - "ProvisionedIopsNotAvailableInAZFault": { - "base": "

Provisioned IOPS not available in the specified Availability Zone.

", - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "PurchaseReservedDBInstancesOfferingResult": { - "base": null, - "refs": { - } - }, - "ReadReplicaDBInstanceIdentifierList": { - "base": null, - "refs": { - "DBInstance$ReadReplicaDBInstanceIdentifiers": null - } - }, - "RebootDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "RebootDBInstanceResult": { - "base": null, - "refs": { - } - }, - "RecurringCharge": { - "base": null, - "refs": { - "RecurringChargeList$member": null - } - }, - "RecurringChargeList": { - "base": null, - "refs": { - "ReservedDBInstance$RecurringCharges": null, - "ReservedDBInstancesOffering$RecurringCharges": null - } - }, - "RemoveSourceIdentifierFromSubscriptionMessage": { - "base": null, - "refs": { - } - }, - "RemoveSourceIdentifierFromSubscriptionResult": { - "base": null, - "refs": { - } - }, - "RemoveTagsFromResourceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstance": { - "base": null, - "refs": { - "PurchaseReservedDBInstancesOfferingResult$ReservedDBInstance": null, - "ReservedDBInstanceList$member": null - } - }, - "ReservedDBInstanceAlreadyExistsFault": { - "base": "

User already has a reservation with the given identifier.

", - "refs": { - } - }, - "ReservedDBInstanceList": { - "base": null, - "refs": { - "ReservedDBInstanceMessage$ReservedDBInstances": null - } - }, - "ReservedDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstanceNotFoundFault": { - "base": "

The specified reserved DB Instance not found.

", - "refs": { - } - }, - "ReservedDBInstanceQuotaExceededFault": { - "base": "

Request would exceed the user's DB Instance quota.

", - "refs": { - } - }, - "ReservedDBInstancesOffering": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingList$member": null - } - }, - "ReservedDBInstancesOfferingList": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferings": null - } - }, - "ReservedDBInstancesOfferingMessage": { - "base": null, - "refs": { - } - }, - "ReservedDBInstancesOfferingNotFoundFault": { - "base": "

Specified offering does not exist.

", - "refs": { - } - }, - "ResetDBParameterGroupMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceFromDBSnapshotResult": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeMessage": { - "base": null, - "refs": { - } - }, - "RestoreDBInstanceToPointInTimeResult": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressMessage": { - "base": null, - "refs": { - } - }, - "RevokeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "SNSInvalidTopicFault": { - "base": "

SNS has responded that there is a problem with the SND topic specified.

", - "refs": { - } - }, - "SNSNoAuthorizationFault": { - "base": "

You do not have permission to publish to the SNS topic ARN.

", - "refs": { - } - }, - "SNSTopicArnNotFoundFault": { - "base": "

The SNS topic ARN does not exist.

", - "refs": { - } - }, - "SnapshotQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB snapshots.

", - "refs": { - } - }, - "SourceIdsList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$SourceIds": null, - "EventSubscription$SourceIdsList": null - } - }, - "SourceNotFoundFault": { - "base": "

The requested source could not be found.

", - "refs": { - } - }, - "SourceType": { - "base": null, - "refs": { - "DescribeEventsMessage$SourceType": null, - "Event$SourceType": null - } - }, - "StorageQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed amount of storage available across all DB instances.

", - "refs": { - } - }, - "StorageTypeNotSupportedFault": { - "base": "

Storage of the StorageType specified can't be associated with the DB instance.

", - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "AddSourceIdentifierToSubscriptionMessage$SubscriptionName": null, - "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier": null, - "AddTagsToResourceMessage$ResourceName": null, - "AuthorizeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$CIDRIP": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "AvailabilityZone$Name": null, - "CharacterSet$CharacterSetName": null, - "CharacterSet$CharacterSetDescription": null, - "CopyDBParameterGroupMessage$SourceDBParameterGroupIdentifier": null, - "CopyDBParameterGroupMessage$TargetDBParameterGroupIdentifier": null, - "CopyDBParameterGroupMessage$TargetDBParameterGroupDescription": null, - "CopyDBSnapshotMessage$SourceDBSnapshotIdentifier": null, - "CopyDBSnapshotMessage$TargetDBSnapshotIdentifier": null, - "CopyOptionGroupMessage$SourceOptionGroupIdentifier": null, - "CopyOptionGroupMessage$TargetOptionGroupIdentifier": null, - "CopyOptionGroupMessage$TargetOptionGroupDescription": null, - "CreateDBInstanceMessage$DBName": null, - "CreateDBInstanceMessage$DBInstanceIdentifier": null, - "CreateDBInstanceMessage$DBInstanceClass": null, - "CreateDBInstanceMessage$Engine": null, - "CreateDBInstanceMessage$MasterUsername": null, - "CreateDBInstanceMessage$MasterUserPassword": null, - "CreateDBInstanceMessage$AvailabilityZone": null, - "CreateDBInstanceMessage$DBSubnetGroupName": null, - "CreateDBInstanceMessage$PreferredMaintenanceWindow": null, - "CreateDBInstanceMessage$DBParameterGroupName": null, - "CreateDBInstanceMessage$PreferredBackupWindow": null, - "CreateDBInstanceMessage$EngineVersion": null, - "CreateDBInstanceMessage$LicenseModel": null, - "CreateDBInstanceMessage$OptionGroupName": null, - "CreateDBInstanceMessage$CharacterSetName": null, - "CreateDBInstanceMessage$StorageType": null, - "CreateDBInstanceMessage$TdeCredentialArn": null, - "CreateDBInstanceMessage$TdeCredentialPassword": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$SourceDBInstanceIdentifier": null, - "CreateDBInstanceReadReplicaMessage$DBInstanceClass": null, - "CreateDBInstanceReadReplicaMessage$AvailabilityZone": null, - "CreateDBInstanceReadReplicaMessage$OptionGroupName": null, - "CreateDBInstanceReadReplicaMessage$DBSubnetGroupName": null, - "CreateDBInstanceReadReplicaMessage$StorageType": null, - "CreateDBParameterGroupMessage$DBParameterGroupName": null, - "CreateDBParameterGroupMessage$DBParameterGroupFamily": null, - "CreateDBParameterGroupMessage$Description": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupName": null, - "CreateDBSecurityGroupMessage$DBSecurityGroupDescription": null, - "CreateDBSnapshotMessage$DBSnapshotIdentifier": null, - "CreateDBSnapshotMessage$DBInstanceIdentifier": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupName": null, - "CreateDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "CreateEventSubscriptionMessage$SubscriptionName": null, - "CreateEventSubscriptionMessage$SnsTopicArn": null, - "CreateEventSubscriptionMessage$SourceType": null, - "CreateOptionGroupMessage$OptionGroupName": null, - "CreateOptionGroupMessage$EngineName": null, - "CreateOptionGroupMessage$MajorEngineVersion": null, - "CreateOptionGroupMessage$OptionGroupDescription": null, - "DBEngineVersion$Engine": null, - "DBEngineVersion$EngineVersion": null, - "DBEngineVersion$DBParameterGroupFamily": null, - "DBEngineVersion$DBEngineDescription": null, - "DBEngineVersion$DBEngineVersionDescription": null, - "DBEngineVersionMessage$Marker": null, - "DBInstance$DBInstanceIdentifier": null, - "DBInstance$DBInstanceClass": null, - "DBInstance$Engine": null, - "DBInstance$DBInstanceStatus": null, - "DBInstance$MasterUsername": null, - "DBInstance$DBName": null, - "DBInstance$PreferredBackupWindow": null, - "DBInstance$AvailabilityZone": null, - "DBInstance$PreferredMaintenanceWindow": null, - "DBInstance$EngineVersion": null, - "DBInstance$ReadReplicaSourceDBInstanceIdentifier": null, - "DBInstance$LicenseModel": null, - "DBInstance$CharacterSetName": null, - "DBInstance$SecondaryAvailabilityZone": null, - "DBInstance$StorageType": null, - "DBInstance$TdeCredentialArn": null, - "DBInstanceMessage$Marker": null, - "DBInstanceStatusInfo$StatusType": null, - "DBInstanceStatusInfo$Status": null, - "DBInstanceStatusInfo$Message": null, - "DBParameterGroup$DBParameterGroupName": null, - "DBParameterGroup$DBParameterGroupFamily": null, - "DBParameterGroup$Description": null, - "DBParameterGroupDetails$Marker": null, - "DBParameterGroupNameMessage$DBParameterGroupName": null, - "DBParameterGroupStatus$DBParameterGroupName": null, - "DBParameterGroupStatus$ParameterApplyStatus": null, - "DBParameterGroupsMessage$Marker": null, - "DBSecurityGroup$OwnerId": null, - "DBSecurityGroup$DBSecurityGroupName": null, - "DBSecurityGroup$DBSecurityGroupDescription": null, - "DBSecurityGroup$VpcId": null, - "DBSecurityGroupMembership$DBSecurityGroupName": null, - "DBSecurityGroupMembership$Status": null, - "DBSecurityGroupMessage$Marker": null, - "DBSecurityGroupNameList$member": null, - "DBSnapshot$DBSnapshotIdentifier": null, - "DBSnapshot$DBInstanceIdentifier": null, - "DBSnapshot$Engine": null, - "DBSnapshot$Status": null, - "DBSnapshot$AvailabilityZone": null, - "DBSnapshot$VpcId": null, - "DBSnapshot$MasterUsername": null, - "DBSnapshot$EngineVersion": null, - "DBSnapshot$LicenseModel": null, - "DBSnapshot$SnapshotType": null, - "DBSnapshot$OptionGroupName": null, - "DBSnapshot$SourceRegion": null, - "DBSnapshot$StorageType": null, - "DBSnapshot$TdeCredentialArn": null, - "DBSnapshotMessage$Marker": null, - "DBSubnetGroup$DBSubnetGroupName": null, - "DBSubnetGroup$DBSubnetGroupDescription": null, - "DBSubnetGroup$VpcId": null, - "DBSubnetGroup$SubnetGroupStatus": null, - "DBSubnetGroupMessage$Marker": null, - "DeleteDBInstanceMessage$DBInstanceIdentifier": null, - "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier": null, - "DeleteDBParameterGroupMessage$DBParameterGroupName": null, - "DeleteDBSecurityGroupMessage$DBSecurityGroupName": null, - "DeleteDBSnapshotMessage$DBSnapshotIdentifier": null, - "DeleteDBSubnetGroupMessage$DBSubnetGroupName": null, - "DeleteEventSubscriptionMessage$SubscriptionName": null, - "DeleteOptionGroupMessage$OptionGroupName": null, - "DescribeDBEngineVersionsMessage$Engine": null, - "DescribeDBEngineVersionsMessage$EngineVersion": null, - "DescribeDBEngineVersionsMessage$DBParameterGroupFamily": null, - "DescribeDBEngineVersionsMessage$Marker": null, - "DescribeDBInstancesMessage$DBInstanceIdentifier": null, - "DescribeDBInstancesMessage$Marker": null, - "DescribeDBLogFilesDetails$LogFileName": null, - "DescribeDBLogFilesMessage$DBInstanceIdentifier": null, - "DescribeDBLogFilesMessage$FilenameContains": null, - "DescribeDBLogFilesMessage$Marker": null, - "DescribeDBLogFilesResponse$Marker": null, - "DescribeDBParameterGroupsMessage$DBParameterGroupName": null, - "DescribeDBParameterGroupsMessage$Marker": null, - "DescribeDBParametersMessage$DBParameterGroupName": null, - "DescribeDBParametersMessage$Source": null, - "DescribeDBParametersMessage$Marker": null, - "DescribeDBSecurityGroupsMessage$DBSecurityGroupName": null, - "DescribeDBSecurityGroupsMessage$Marker": null, - "DescribeDBSnapshotsMessage$DBInstanceIdentifier": null, - "DescribeDBSnapshotsMessage$DBSnapshotIdentifier": null, - "DescribeDBSnapshotsMessage$SnapshotType": null, - "DescribeDBSnapshotsMessage$Marker": null, - "DescribeDBSubnetGroupsMessage$DBSubnetGroupName": null, - "DescribeDBSubnetGroupsMessage$Marker": null, - "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily": null, - "DescribeEngineDefaultParametersMessage$Marker": null, - "DescribeEventCategoriesMessage$SourceType": null, - "DescribeEventSubscriptionsMessage$SubscriptionName": null, - "DescribeEventSubscriptionsMessage$Marker": null, - "DescribeEventsMessage$SourceIdentifier": null, - "DescribeEventsMessage$Marker": null, - "DescribeOptionGroupOptionsMessage$EngineName": null, - "DescribeOptionGroupOptionsMessage$MajorEngineVersion": null, - "DescribeOptionGroupOptionsMessage$Marker": null, - "DescribeOptionGroupsMessage$OptionGroupName": null, - "DescribeOptionGroupsMessage$Marker": null, - "DescribeOptionGroupsMessage$EngineName": null, - "DescribeOptionGroupsMessage$MajorEngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$Engine": null, - "DescribeOrderableDBInstanceOptionsMessage$EngineVersion": null, - "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass": null, - "DescribeOrderableDBInstanceOptionsMessage$LicenseModel": null, - "DescribeOrderableDBInstanceOptionsMessage$Marker": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstanceId": null, - "DescribeReservedDBInstancesMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesMessage$Duration": null, - "DescribeReservedDBInstancesMessage$ProductDescription": null, - "DescribeReservedDBInstancesMessage$OfferingType": null, - "DescribeReservedDBInstancesMessage$Marker": null, - "DescribeReservedDBInstancesOfferingsMessage$ReservedDBInstancesOfferingId": null, - "DescribeReservedDBInstancesOfferingsMessage$DBInstanceClass": null, - "DescribeReservedDBInstancesOfferingsMessage$Duration": null, - "DescribeReservedDBInstancesOfferingsMessage$ProductDescription": null, - "DescribeReservedDBInstancesOfferingsMessage$OfferingType": null, - "DescribeReservedDBInstancesOfferingsMessage$Marker": null, - "DownloadDBLogFilePortionDetails$LogFileData": null, - "DownloadDBLogFilePortionDetails$Marker": null, - "DownloadDBLogFilePortionMessage$DBInstanceIdentifier": null, - "DownloadDBLogFilePortionMessage$LogFileName": null, - "DownloadDBLogFilePortionMessage$Marker": null, - "EC2SecurityGroup$Status": null, - "EC2SecurityGroup$EC2SecurityGroupName": null, - "EC2SecurityGroup$EC2SecurityGroupId": null, - "EC2SecurityGroup$EC2SecurityGroupOwnerId": null, - "Endpoint$Address": null, - "EngineDefaults$DBParameterGroupFamily": null, - "EngineDefaults$Marker": null, - "Event$SourceIdentifier": null, - "Event$Message": null, - "EventCategoriesList$member": null, - "EventCategoriesMap$SourceType": null, - "EventSubscription$CustomerAwsId": null, - "EventSubscription$CustSubscriptionId": null, - "EventSubscription$SnsTopicArn": null, - "EventSubscription$Status": null, - "EventSubscription$SubscriptionCreationTime": null, - "EventSubscription$SourceType": null, - "EventSubscriptionsMessage$Marker": null, - "EventsMessage$Marker": null, - "Filter$Name": null, - "FilterValueList$member": null, - "IPRange$Status": null, - "IPRange$CIDRIP": null, - "KeyList$member": null, - "ListTagsForResourceMessage$ResourceName": null, - "ModifyDBInstanceMessage$DBInstanceIdentifier": null, - "ModifyDBInstanceMessage$DBInstanceClass": null, - "ModifyDBInstanceMessage$MasterUserPassword": null, - "ModifyDBInstanceMessage$DBParameterGroupName": null, - "ModifyDBInstanceMessage$PreferredBackupWindow": null, - "ModifyDBInstanceMessage$PreferredMaintenanceWindow": null, - "ModifyDBInstanceMessage$EngineVersion": null, - "ModifyDBInstanceMessage$OptionGroupName": null, - "ModifyDBInstanceMessage$NewDBInstanceIdentifier": null, - "ModifyDBInstanceMessage$StorageType": null, - "ModifyDBInstanceMessage$TdeCredentialArn": null, - "ModifyDBInstanceMessage$TdeCredentialPassword": null, - "ModifyDBParameterGroupMessage$DBParameterGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupName": null, - "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription": null, - "ModifyEventSubscriptionMessage$SubscriptionName": null, - "ModifyEventSubscriptionMessage$SnsTopicArn": null, - "ModifyEventSubscriptionMessage$SourceType": null, - "ModifyOptionGroupMessage$OptionGroupName": null, - "Option$OptionName": null, - "Option$OptionDescription": null, - "OptionConfiguration$OptionName": null, - "OptionGroup$OptionGroupName": null, - "OptionGroup$OptionGroupDescription": null, - "OptionGroup$EngineName": null, - "OptionGroup$MajorEngineVersion": null, - "OptionGroup$VpcId": null, - "OptionGroupMembership$OptionGroupName": null, - "OptionGroupMembership$Status": null, - "OptionGroupOption$Name": null, - "OptionGroupOption$Description": null, - "OptionGroupOption$EngineName": null, - "OptionGroupOption$MajorEngineVersion": null, - "OptionGroupOption$MinimumRequiredMinorEngineVersion": null, - "OptionGroupOptionSetting$SettingName": null, - "OptionGroupOptionSetting$SettingDescription": null, - "OptionGroupOptionSetting$DefaultValue": null, - "OptionGroupOptionSetting$ApplyType": null, - "OptionGroupOptionSetting$AllowedValues": null, - "OptionGroupOptionsMessage$Marker": null, - "OptionGroups$Marker": null, - "OptionNamesList$member": null, - "OptionSetting$Name": null, - "OptionSetting$Value": null, - "OptionSetting$DefaultValue": null, - "OptionSetting$Description": null, - "OptionSetting$ApplyType": null, - "OptionSetting$DataType": null, - "OptionSetting$AllowedValues": null, - "OptionsDependedOn$member": null, - "OrderableDBInstanceOption$Engine": null, - "OrderableDBInstanceOption$EngineVersion": null, - "OrderableDBInstanceOption$DBInstanceClass": null, - "OrderableDBInstanceOption$LicenseModel": null, - "OrderableDBInstanceOption$StorageType": null, - "OrderableDBInstanceOptionsMessage$Marker": null, - "Parameter$ParameterName": null, - "Parameter$ParameterValue": null, - "Parameter$Description": null, - "Parameter$Source": null, - "Parameter$ApplyType": null, - "Parameter$DataType": null, - "Parameter$AllowedValues": null, - "Parameter$MinimumEngineVersion": null, - "PendingModifiedValues$DBInstanceClass": null, - "PendingModifiedValues$MasterUserPassword": null, - "PendingModifiedValues$EngineVersion": null, - "PendingModifiedValues$DBInstanceIdentifier": null, - "PendingModifiedValues$StorageType": null, - "PromoteReadReplicaMessage$DBInstanceIdentifier": null, - "PromoteReadReplicaMessage$PreferredBackupWindow": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferingId": null, - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstanceId": null, - "ReadReplicaDBInstanceIdentifierList$member": null, - "RebootDBInstanceMessage$DBInstanceIdentifier": null, - "RecurringCharge$RecurringChargeFrequency": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName": null, - "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier": null, - "RemoveTagsFromResourceMessage$ResourceName": null, - "ReservedDBInstance$ReservedDBInstanceId": null, - "ReservedDBInstance$ReservedDBInstancesOfferingId": null, - "ReservedDBInstance$DBInstanceClass": null, - "ReservedDBInstance$CurrencyCode": null, - "ReservedDBInstance$ProductDescription": null, - "ReservedDBInstance$OfferingType": null, - "ReservedDBInstance$State": null, - "ReservedDBInstanceMessage$Marker": null, - "ReservedDBInstancesOffering$ReservedDBInstancesOfferingId": null, - "ReservedDBInstancesOffering$DBInstanceClass": null, - "ReservedDBInstancesOffering$CurrencyCode": null, - "ReservedDBInstancesOffering$ProductDescription": null, - "ReservedDBInstancesOffering$OfferingType": null, - "ReservedDBInstancesOfferingMessage$Marker": null, - "ResetDBParameterGroupMessage$DBParameterGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSnapshotIdentifier": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceClass": null, - "RestoreDBInstanceFromDBSnapshotMessage$AvailabilityZone": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBSubnetGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$LicenseModel": null, - "RestoreDBInstanceFromDBSnapshotMessage$DBName": null, - "RestoreDBInstanceFromDBSnapshotMessage$Engine": null, - "RestoreDBInstanceFromDBSnapshotMessage$OptionGroupName": null, - "RestoreDBInstanceFromDBSnapshotMessage$StorageType": null, - "RestoreDBInstanceFromDBSnapshotMessage$TdeCredentialArn": null, - "RestoreDBInstanceFromDBSnapshotMessage$TdeCredentialPassword": null, - "RestoreDBInstanceToPointInTimeMessage$SourceDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$TargetDBInstanceIdentifier": null, - "RestoreDBInstanceToPointInTimeMessage$DBInstanceClass": null, - "RestoreDBInstanceToPointInTimeMessage$AvailabilityZone": null, - "RestoreDBInstanceToPointInTimeMessage$DBSubnetGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$LicenseModel": null, - "RestoreDBInstanceToPointInTimeMessage$DBName": null, - "RestoreDBInstanceToPointInTimeMessage$Engine": null, - "RestoreDBInstanceToPointInTimeMessage$OptionGroupName": null, - "RestoreDBInstanceToPointInTimeMessage$StorageType": null, - "RestoreDBInstanceToPointInTimeMessage$TdeCredentialArn": null, - "RestoreDBInstanceToPointInTimeMessage$TdeCredentialPassword": null, - "RevokeDBSecurityGroupIngressMessage$DBSecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$CIDRIP": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupName": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupId": null, - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": null, - "SourceIdsList$member": null, - "Subnet$SubnetIdentifier": null, - "Subnet$SubnetStatus": null, - "SubnetIdentifierList$member": null, - "Tag$Key": null, - "Tag$Value": null, - "VpcSecurityGroupIdList$member": null, - "VpcSecurityGroupMembership$VpcSecurityGroupId": null, - "VpcSecurityGroupMembership$Status": null - } - }, - "Subnet": { - "base": null, - "refs": { - "SubnetList$member": null - } - }, - "SubnetAlreadyInUse": { - "base": "

The DB subnet is already in use in the Availability Zone.

", - "refs": { - } - }, - "SubnetIdentifierList": { - "base": null, - "refs": { - "CreateDBSubnetGroupMessage$SubnetIds": null, - "ModifyDBSubnetGroupMessage$SubnetIds": null - } - }, - "SubnetList": { - "base": null, - "refs": { - "DBSubnetGroup$Subnets": null - } - }, - "SubscriptionAlreadyExistFault": { - "base": "

The supplied subscription name already exists.

", - "refs": { - } - }, - "SubscriptionCategoryNotFoundFault": { - "base": "

The supplied category does not exist.

", - "refs": { - } - }, - "SubscriptionNotFoundFault": { - "base": "

The subscription name does not exist.

", - "refs": { - } - }, - "SupportedCharacterSetsList": { - "base": null, - "refs": { - "DBEngineVersion$SupportedCharacterSets": null - } - }, - "TStamp": { - "base": null, - "refs": { - "DBInstance$InstanceCreateTime": null, - "DBInstance$LatestRestorableTime": null, - "DBSnapshot$SnapshotCreateTime": null, - "DBSnapshot$InstanceCreateTime": null, - "DescribeEventsMessage$StartTime": null, - "DescribeEventsMessage$EndTime": null, - "Event$Date": null, - "ReservedDBInstance$StartTime": null, - "RestoreDBInstanceToPointInTimeMessage$RestoreTime": null - } - }, - "Tag": { - "base": null, - "refs": { - "TagList$member": null - } - }, - "TagList": { - "base": null, - "refs": { - "AddTagsToResourceMessage$Tags": null, - "CopyDBParameterGroupMessage$Tags": null, - "CopyDBSnapshotMessage$Tags": null, - "CopyOptionGroupMessage$Tags": null, - "CreateDBInstanceMessage$Tags": null, - "CreateDBInstanceReadReplicaMessage$Tags": null, - "CreateDBParameterGroupMessage$Tags": null, - "CreateDBSecurityGroupMessage$Tags": null, - "CreateDBSnapshotMessage$Tags": null, - "CreateDBSubnetGroupMessage$Tags": null, - "CreateEventSubscriptionMessage$Tags": null, - "CreateOptionGroupMessage$Tags": null, - "PurchaseReservedDBInstancesOfferingMessage$Tags": null, - "RestoreDBInstanceFromDBSnapshotMessage$Tags": null, - "RestoreDBInstanceToPointInTimeMessage$Tags": null, - "TagListMessage$TagList": null - } - }, - "TagListMessage": { - "base": null, - "refs": { - } - }, - "VpcSecurityGroupIdList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$VpcSecurityGroupIds": null, - "ModifyDBInstanceMessage$VpcSecurityGroupIds": null, - "OptionConfiguration$VpcSecurityGroupMemberships": null - } - }, - "VpcSecurityGroupMembership": { - "base": null, - "refs": { - "VpcSecurityGroupMembershipList$member": null - } - }, - "VpcSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$VpcSecurityGroups": null, - "Option$VpcSecurityGroupMemberships": null - } - } - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/examples-1.json deleted file mode 100644 index 0ea7e3b0b..000000000 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/examples-1.json +++ /dev/null @@ -1,5 +0,0 @@ -{ - "version": "1.0", - "examples": { - } -} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/api-2.json index e8557445c..e3ee0284e 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/api-2.json @@ -1436,7 +1436,8 @@ {"shape":"DBSubnetGroupNotFoundFault"}, {"shape":"InvalidSubnet"}, {"shape":"OptionGroupNotFoundFault"}, - {"shape":"KMSKeyNotAccessibleFault"} + {"shape":"KMSKeyNotAccessibleFault"}, + {"shape":"DBClusterParameterGroupNotFoundFault"} ] }, "RestoreDBClusterToPointInTime":{ @@ -1466,7 +1467,8 @@ {"shape":"InvalidVPCNetworkStateFault"}, {"shape":"KMSKeyNotAccessibleFault"}, {"shape":"OptionGroupNotFoundFault"}, - {"shape":"StorageQuotaExceededFault"} + {"shape":"StorageQuotaExceededFault"}, + {"shape":"DBClusterParameterGroupNotFoundFault"} ] }, "RestoreDBInstanceFromDBSnapshot":{ @@ -1498,7 +1500,8 @@ {"shape":"AuthorizationNotFoundFault"}, {"shape":"KMSKeyNotAccessibleFault"}, {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"DomainNotFoundFault"} + {"shape":"DomainNotFoundFault"}, + {"shape":"DBParameterGroupNotFoundFault"} ] }, "RestoreDBInstanceFromS3":{ @@ -1561,7 +1564,8 @@ {"shape":"AuthorizationNotFoundFault"}, {"shape":"KMSKeyNotAccessibleFault"}, {"shape":"DBSecurityGroupNotFoundFault"}, - {"shape":"DomainNotFoundFault"} + {"shape":"DomainNotFoundFault"}, + {"shape":"DBParameterGroupNotFoundFault"} ] }, "RevokeDBSecurityGroupIngress":{ @@ -1792,6 +1796,21 @@ "locationName":"AvailabilityZone" } }, + "AvailableProcessorFeature":{ + "type":"structure", + "members":{ + "Name":{"shape":"String"}, + "DefaultValue":{"shape":"String"}, + "AllowedValues":{"shape":"String"} + } + }, + "AvailableProcessorFeatureList":{ + "type":"list", + "member":{ + "shape":"AvailableProcessorFeature", + "locationName":"AvailableProcessorFeature" + } + }, "BacktrackDBClusterMessage":{ "type":"structure", "required":[ @@ -1990,7 +2009,8 @@ "KmsKeyId":{"shape":"String"}, "PreSignedUrl":{"shape":"String"}, "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, - "BacktrackWindow":{"shape":"LongOptional"} + "BacktrackWindow":{"shape":"LongOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} } }, "CreateDBClusterParameterGroupMessage":{ @@ -2086,7 +2106,9 @@ "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, "EnablePerformanceInsights":{"shape":"BooleanOptional"}, "PerformanceInsightsKMSKeyId":{"shape":"String"}, - "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} + "PerformanceInsightsRetentionPeriod":{"shape":"IntegerOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"} } }, "CreateDBInstanceReadReplicaMessage":{ @@ -2117,7 +2139,10 @@ "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, "EnablePerformanceInsights":{"shape":"BooleanOptional"}, "PerformanceInsightsKMSKeyId":{"shape":"String"}, - "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} + "PerformanceInsightsRetentionPeriod":{"shape":"IntegerOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"}, + "UseDefaultProcessorFeatures":{"shape":"BooleanOptional"} } }, "CreateDBInstanceReadReplicaResult":{ @@ -2292,7 +2317,8 @@ "ClusterCreateTime":{"shape":"TStamp"}, "EarliestBacktrackTime":{"shape":"TStamp"}, "BacktrackWindow":{"shape":"LongOptional"}, - "BacktrackConsumedChangeRecords":{"shape":"LongOptional"} + "BacktrackConsumedChangeRecords":{"shape":"LongOptional"}, + "EnabledCloudwatchLogsExports":{"shape":"LogTypeList"} }, "wrapper":true }, @@ -2674,7 +2700,9 @@ "IAMDatabaseAuthenticationEnabled":{"shape":"Boolean"}, "PerformanceInsightsEnabled":{"shape":"BooleanOptional"}, "PerformanceInsightsKMSKeyId":{"shape":"String"}, - "EnabledCloudwatchLogsExports":{"shape":"LogTypeList"} + "PerformanceInsightsRetentionPeriod":{"shape":"IntegerOptional"}, + "EnabledCloudwatchLogsExports":{"shape":"LogTypeList"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"} }, "wrapper":true }, @@ -2945,7 +2973,8 @@ "KmsKeyId":{"shape":"String"}, "DBSnapshotArn":{"shape":"String"}, "Timezone":{"shape":"String"}, - "IAMDatabaseAuthenticationEnabled":{"shape":"Boolean"} + "IAMDatabaseAuthenticationEnabled":{"shape":"Boolean"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"} }, "wrapper":true }, @@ -4075,6 +4104,7 @@ "PreferredMaintenanceWindow":{"shape":"String"}, "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, "BacktrackWindow":{"shape":"LongOptional"}, + "CloudwatchLogsExportConfiguration":{"shape":"CloudwatchLogsExportConfiguration"}, "EngineVersion":{"shape":"String"} } }, @@ -4153,7 +4183,10 @@ "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, "EnablePerformanceInsights":{"shape":"BooleanOptional"}, "PerformanceInsightsKMSKeyId":{"shape":"String"}, - "CloudwatchLogsExportConfiguration":{"shape":"CloudwatchLogsExportConfiguration"} + "PerformanceInsightsRetentionPeriod":{"shape":"IntegerOptional"}, + "CloudwatchLogsExportConfiguration":{"shape":"CloudwatchLogsExportConfiguration"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"}, + "UseDefaultProcessorFeatures":{"shape":"BooleanOptional"} } }, "ModifyDBInstanceResult":{ @@ -4508,7 +4541,8 @@ "MinIopsPerDbInstance":{"shape":"IntegerOptional"}, "MaxIopsPerDbInstance":{"shape":"IntegerOptional"}, "MinIopsPerGib":{"shape":"DoubleOptional"}, - "MaxIopsPerGib":{"shape":"DoubleOptional"} + "MaxIopsPerGib":{"shape":"DoubleOptional"}, + "AvailableProcessorFeatures":{"shape":"AvailableProcessorFeatureList"} }, "wrapper":true }, @@ -4603,7 +4637,8 @@ "StorageType":{"shape":"String"}, "CACertificateIdentifier":{"shape":"String"}, "DBSubnetGroupName":{"shape":"String"}, - "PendingCloudwatchLogsExports":{"shape":"PendingCloudwatchLogsExports"} + "PendingCloudwatchLogsExports":{"shape":"PendingCloudwatchLogsExports"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"} } }, "PointInTimeRestoreNotEnabledFault":{ @@ -4617,6 +4652,20 @@ }, "exception":true }, + "ProcessorFeature":{ + "type":"structure", + "members":{ + "Name":{"shape":"String"}, + "Value":{"shape":"String"} + } + }, + "ProcessorFeatureList":{ + "type":"list", + "member":{ + "shape":"ProcessorFeature", + "locationName":"ProcessorFeature" + } + }, "PromoteReadReplicaDBClusterMessage":{ "type":"structure", "required":["DBClusterIdentifier"], @@ -4960,7 +5009,8 @@ "S3BucketName":{"shape":"String"}, "S3Prefix":{"shape":"String"}, "S3IngestionRoleArn":{"shape":"String"}, - "BacktrackWindow":{"shape":"LongOptional"} + "BacktrackWindow":{"shape":"LongOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} } }, "RestoreDBClusterFromS3Result":{ @@ -4990,7 +5040,8 @@ "Tags":{"shape":"TagList"}, "KmsKeyId":{"shape":"String"}, "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, - "BacktrackWindow":{"shape":"LongOptional"} + "BacktrackWindow":{"shape":"LongOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} } }, "RestoreDBClusterFromSnapshotResult":{ @@ -5018,7 +5069,8 @@ "Tags":{"shape":"TagList"}, "KmsKeyId":{"shape":"String"}, "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, - "BacktrackWindow":{"shape":"LongOptional"} + "BacktrackWindow":{"shape":"LongOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} } }, "RestoreDBClusterToPointInTimeResult":{ @@ -5056,7 +5108,9 @@ "CopyTagsToSnapshot":{"shape":"BooleanOptional"}, "DomainIAMRoleName":{"shape":"String"}, "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, - "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"}, + "UseDefaultProcessorFeatures":{"shape":"BooleanOptional"} } }, "RestoreDBInstanceFromDBSnapshotResult":{ @@ -5115,7 +5169,10 @@ "S3IngestionRoleArn":{"shape":"String"}, "EnablePerformanceInsights":{"shape":"BooleanOptional"}, "PerformanceInsightsKMSKeyId":{"shape":"String"}, - "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} + "PerformanceInsightsRetentionPeriod":{"shape":"IntegerOptional"}, + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"}, + "UseDefaultProcessorFeatures":{"shape":"BooleanOptional"} } }, "RestoreDBInstanceFromS3Result":{ @@ -5155,7 +5212,9 @@ "Domain":{"shape":"String"}, "DomainIAMRoleName":{"shape":"String"}, "EnableIAMDatabaseAuthentication":{"shape":"BooleanOptional"}, - "EnableCloudwatchLogsExports":{"shape":"LogTypeList"} + "EnableCloudwatchLogsExports":{"shape":"LogTypeList"}, + "ProcessorFeatures":{"shape":"ProcessorFeatureList"}, + "UseDefaultProcessorFeatures":{"shape":"BooleanOptional"} } }, "RestoreDBInstanceToPointInTimeResult":{ @@ -5457,7 +5516,8 @@ "ValidDBInstanceModificationsMessage":{ "type":"structure", "members":{ - "Storage":{"shape":"ValidStorageOptionsList"} + "Storage":{"shape":"ValidStorageOptionsList"}, + "ValidProcessorFeatures":{"shape":"AvailableProcessorFeatureList"} }, "wrapper":true }, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/docs-2.json index f9c261047..01cb108fe 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/docs-2.json @@ -1,3253 +1,3307 @@ { - "version": "2.0", - "service": "Amazon Relational Database Service

Amazon Relational Database Service (Amazon RDS) is a web service that makes it easier to set up, operate, and scale a relational database in the cloud. It provides cost-efficient, resizable capacity for an industry-standard relational database and manages common database administration tasks, freeing up developers to focus on what makes their applications and businesses unique.

Amazon RDS gives you access to the capabilities of a MySQL, MariaDB, PostgreSQL, Microsoft SQL Server, Oracle, or Amazon Aurora database server. These capabilities mean that the code, applications, and tools you already use today with your existing databases work with Amazon RDS without modification. Amazon RDS automatically backs up your database and maintains the database software that powers your DB instance. Amazon RDS is flexible: you can scale your DB instance's compute resources and storage capacity to meet your application's demand. As with all Amazon Web Services, there are no up-front investments, and you pay only for the resources you use.

This interface reference for Amazon RDS contains documentation for a programming or command line interface you can use to manage Amazon RDS. Note that Amazon RDS is asynchronous, which means that some interfaces might require techniques such as polling or callback functions to determine when a command has been applied. In this reference, the parameter descriptions indicate whether a command is applied immediately, on the next instance reboot, or during the maintenance window. The reference structure is as follows, and we list following some related topics from the user guide.

Amazon RDS API Reference

Amazon RDS User Guide

", - "operations": { - "AddRoleToDBCluster": "

Associates an Identity and Access Management (IAM) role from an Aurora DB cluster. For more information, see Authorizing Amazon Aurora to Access Other AWS Services On Your Behalf.

", - "AddSourceIdentifierToSubscription": "

Adds a source identifier to an existing RDS event notification subscription.

", - "AddTagsToResource": "

Adds metadata tags to an Amazon RDS resource. These tags can also be used with cost allocation reporting to track cost associated with Amazon RDS resources, or used in a Condition statement in an IAM policy for Amazon RDS.

For an overview on tagging Amazon RDS resources, see Tagging Amazon RDS Resources.

", - "ApplyPendingMaintenanceAction": "

Applies a pending maintenance action to a resource (for example, to a DB instance).

", - "AuthorizeDBSecurityGroupIngress": "

Enables ingress to a DBSecurityGroup using one of two forms of authorization. First, EC2 or VPC security groups can be added to the DBSecurityGroup if the application using the database is running on EC2 or VPC instances. Second, IP ranges are available if the application accessing your database is running on the Internet. Required parameters for this API are one of CIDR range, EC2SecurityGroupId for VPC, or (EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId for non-VPC).

You can't authorize ingress from an EC2 security group in one AWS Region to an Amazon RDS DB instance in another. You can't authorize ingress from a VPC security group in one VPC to an Amazon RDS DB instance in another.

For an overview of CIDR ranges, go to the Wikipedia Tutorial.

", - "BacktrackDBCluster": "

Backtracks a DB cluster to a specific time, without creating a new DB cluster.

For more information on backtracking, see Backtracking an Aurora DB Cluster in the Amazon RDS User Guide.

", - "CopyDBClusterParameterGroup": "

Copies the specified DB cluster parameter group.

", - "CopyDBClusterSnapshot": "

Copies a snapshot of a DB cluster.

To copy a DB cluster snapshot from a shared manual DB cluster snapshot, SourceDBClusterSnapshotIdentifier must be the Amazon Resource Name (ARN) of the shared DB cluster snapshot.

You can copy an encrypted DB cluster snapshot from another AWS Region. In that case, the AWS Region where you call the CopyDBClusterSnapshot action is the destination AWS Region for the encrypted DB cluster snapshot to be copied to. To copy an encrypted DB cluster snapshot from another AWS Region, you must provide the following values:

  • KmsKeyId - The AWS Key Management System (AWS KMS) key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region.

  • PreSignedUrl - A URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot action to be called in the source AWS Region where the DB cluster snapshot is copied from. The pre-signed URL must be a valid request for the CopyDBClusterSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB cluster snapshot to be copied.

    The pre-signed URL request must contain the following parameter values:

    • KmsKeyId - The KMS key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. This is the same identifier for both the CopyDBClusterSnapshot action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

    • DestinationRegion - The name of the AWS Region that the DB cluster snapshot will be created in.

    • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster snapshot from the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:aurora-cluster1-snapshot-20161115.

    To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

  • TargetDBClusterSnapshotIdentifier - The identifier for the new copy of the DB cluster snapshot in the destination AWS Region.

  • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the ARN format for the source AWS Region and is the same value as the SourceDBClusterSnapshotIdentifier in the pre-signed URL.

To cancel the copy operation once it is in progress, delete the target DB cluster snapshot identified by TargetDBClusterSnapshotIdentifier while that DB cluster snapshot is in \"copying\" status.

For more information on copying encrypted DB cluster snapshots from one AWS Region to another, see Copying a DB Cluster Snapshot in the Same Account, Either in the Same Region or Across Regions in the Amazon RDS User Guide.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "CopyDBParameterGroup": "

Copies the specified DB parameter group.

", - "CopyDBSnapshot": "

Copies the specified DB snapshot. The source DB snapshot must be in the \"available\" state.

You can copy a snapshot from one AWS Region to another. In that case, the AWS Region where you call the CopyDBSnapshot action is the destination AWS Region for the DB snapshot copy.

For more information about copying snapshots, see Copying a DB Snapshot in the Amazon RDS User Guide.

", - "CopyOptionGroup": "

Copies the specified option group.

", - "CreateDBCluster": "

Creates a new Amazon Aurora DB cluster.

You can use the ReplicationSourceIdentifier parameter to create the DB cluster as a Read Replica of another DB cluster or Amazon RDS MySQL DB instance. For cross-region replication where the DB cluster identified by ReplicationSourceIdentifier is encrypted, you must also specify the PreSignedUrl parameter.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "CreateDBClusterParameterGroup": "

Creates a new DB cluster parameter group.

Parameters in a DB cluster parameter group apply to all of the instances in a DB cluster.

A DB cluster parameter group is initially created with the default parameters for the database engine used by instances in the DB cluster. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBClusterParameterGroup. Once you've created a DB cluster parameter group, you need to associate it with your DB cluster using ModifyDBCluster. When you associate a new DB cluster parameter group with a running DB cluster, you need to reboot the DB instances in the DB cluster without failover for the new DB cluster parameter group and associated settings to take effect.

After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the DB cluster parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBClusterParameters command to verify that your DB cluster parameter group has been created or modified.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "CreateDBClusterSnapshot": "

Creates a snapshot of a DB cluster. For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "CreateDBInstance": "

Creates a new DB instance.

", - "CreateDBInstanceReadReplica": "

Creates a new DB instance that acts as a Read Replica for an existing source DB instance. You can create a Read Replica for a DB instance running MySQL, MariaDB, or PostgreSQL. For more information, see Working with PostgreSQL, MySQL, and MariaDB Read Replicas.

Amazon Aurora doesn't support this action. You must call the CreateDBInstance action to create a DB instance for an Aurora DB cluster.

All Read Replica DB instances are created with backups disabled. All other DB instance attributes (including DB security groups and DB parameter groups) are inherited from the source DB instance, except as specified following.

Your source DB instance must have backup retention enabled.

", - "CreateDBParameterGroup": "

Creates a new DB parameter group.

A DB parameter group is initially created with the default parameters for the database engine used by the DB instance. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBParameterGroup. Once you've created a DB parameter group, you need to associate it with your DB instance using ModifyDBInstance. When you associate a new DB parameter group with a running DB instance, you need to reboot the DB instance without failover for the new DB parameter group and associated settings to take effect.

After you create a DB parameter group, you should wait at least 5 minutes before creating your first DB instance that uses that DB parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the parameter group is used as the default for a new DB instance. This is especially important for parameters that are critical when creating the default database for a DB instance, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBParameters command to verify that your DB parameter group has been created or modified.

", - "CreateDBSecurityGroup": "

Creates a new DB security group. DB security groups control access to a DB instance.

", - "CreateDBSnapshot": "

Creates a DBSnapshot. The source DBInstance must be in \"available\" state.

", - "CreateDBSubnetGroup": "

Creates a new DB subnet group. DB subnet groups must contain at least one subnet in at least two AZs in the AWS Region.

", - "CreateEventSubscription": "

Creates an RDS event notification subscription. This action requires a topic ARN (Amazon Resource Name) created by either the RDS console, the SNS console, or the SNS API. To obtain an ARN with SNS, you must create a topic in Amazon SNS and subscribe to the topic. The ARN is displayed in the SNS console.

You can specify the type of source (SourceType) you want to be notified of, provide a list of RDS sources (SourceIds) that triggers the events, and provide a list of event categories (EventCategories) for events you want to be notified of. For example, you can specify SourceType = db-instance, SourceIds = mydbinstance1, mydbinstance2 and EventCategories = Availability, Backup.

If you specify both the SourceType and SourceIds, such as SourceType = db-instance and SourceIdentifier = myDBInstance1, you are notified of all the db-instance events for the specified source. If you specify a SourceType but do not specify a SourceIdentifier, you receive notice of the events for that source type for all your RDS sources. If you do not specify either the SourceType nor the SourceIdentifier, you are notified of events generated from all RDS sources belonging to your customer account.

", - "CreateOptionGroup": "

Creates a new option group. You can create up to 20 option groups.

", - "DeleteDBCluster": "

The DeleteDBCluster action deletes a previously provisioned DB cluster. When you delete a DB cluster, all automated backups for that DB cluster are deleted and can't be recovered. Manual DB cluster snapshots of the specified DB cluster are not deleted.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DeleteDBClusterParameterGroup": "

Deletes a specified DB cluster parameter group. The DB cluster parameter group to be deleted can't be associated with any DB clusters.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DeleteDBClusterSnapshot": "

Deletes a DB cluster snapshot. If the snapshot is being copied, the copy operation is terminated.

The DB cluster snapshot must be in the available state to be deleted.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DeleteDBInstance": "

The DeleteDBInstance action deletes a previously provisioned DB instance. When you delete a DB instance, all automated backups for that instance are deleted and can't be recovered. Manual DB snapshots of the DB instance to be deleted by DeleteDBInstance are not deleted.

If you request a final DB snapshot the status of the Amazon RDS DB instance is deleting until the DB snapshot is created. The API action DescribeDBInstance is used to monitor the status of this operation. The action can't be canceled or reverted once submitted.

Note that when a DB instance is in a failure state and has a status of failed, incompatible-restore, or incompatible-network, you can only delete it when the SkipFinalSnapshot parameter is set to true.

If the specified DB instance is part of an Amazon Aurora DB cluster, you can't delete the DB instance if both of the following conditions are true:

  • The DB cluster is a Read Replica of another Amazon Aurora DB cluster.

  • The DB instance is the only instance in the DB cluster.

To delete a DB instance in this case, first call the PromoteReadReplicaDBCluster API action to promote the DB cluster so it's no longer a Read Replica. After the promotion completes, then call the DeleteDBInstance API action to delete the final instance in the DB cluster.

", - "DeleteDBParameterGroup": "

Deletes a specified DBParameterGroup. The DBParameterGroup to be deleted can't be associated with any DB instances.

", - "DeleteDBSecurityGroup": "

Deletes a DB security group.

The specified DB security group must not be associated with any DB instances.

", - "DeleteDBSnapshot": "

Deletes a DBSnapshot. If the snapshot is being copied, the copy operation is terminated.

The DBSnapshot must be in the available state to be deleted.

", - "DeleteDBSubnetGroup": "

Deletes a DB subnet group.

The specified database subnet group must not be associated with any DB instances.

", - "DeleteEventSubscription": "

Deletes an RDS event notification subscription.

", - "DeleteOptionGroup": "

Deletes an existing option group.

", - "DescribeAccountAttributes": "

Lists all of the attributes for a customer account. The attributes include Amazon RDS quotas for the account, such as the number of DB instances allowed. The description for a quota includes the quota name, current usage toward that quota, and the quota's maximum value.

This command doesn't take any parameters.

", - "DescribeCertificates": "

Lists the set of CA certificates provided by Amazon RDS for this AWS account.

", - "DescribeDBClusterBacktracks": "

Returns information about backtracks for a DB cluster.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DescribeDBClusterParameterGroups": "

Returns a list of DBClusterParameterGroup descriptions. If a DBClusterParameterGroupName parameter is specified, the list will contain only the description of the specified DB cluster parameter group.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DescribeDBClusterParameters": "

Returns the detailed parameter list for a particular DB cluster parameter group.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DescribeDBClusterSnapshotAttributes": "

Returns a list of DB cluster snapshot attribute names and values for a manual DB cluster snapshot.

When sharing snapshots with other AWS accounts, DescribeDBClusterSnapshotAttributes returns the restore attribute and a list of IDs for the AWS accounts that are authorized to copy or restore the manual DB cluster snapshot. If all is included in the list of values for the restore attribute, then the manual DB cluster snapshot is public and can be copied or restored by all AWS accounts.

To add or remove access for an AWS account to copy or restore a manual DB cluster snapshot, or to make the manual DB cluster snapshot public or private, use the ModifyDBClusterSnapshotAttribute API action.

", - "DescribeDBClusterSnapshots": "

Returns information about DB cluster snapshots. This API action supports pagination.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DescribeDBClusters": "

Returns information about provisioned Aurora DB clusters. This API supports pagination.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DescribeDBEngineVersions": "

Returns a list of the available DB engines.

", - "DescribeDBInstances": "

Returns information about provisioned RDS instances. This API supports pagination.

", - "DescribeDBLogFiles": "

Returns a list of DB log files for the DB instance.

", - "DescribeDBParameterGroups": "

Returns a list of DBParameterGroup descriptions. If a DBParameterGroupName is specified, the list will contain only the description of the specified DB parameter group.

", - "DescribeDBParameters": "

Returns the detailed parameter list for a particular DB parameter group.

", - "DescribeDBSecurityGroups": "

Returns a list of DBSecurityGroup descriptions. If a DBSecurityGroupName is specified, the list will contain only the descriptions of the specified DB security group.

", - "DescribeDBSnapshotAttributes": "

Returns a list of DB snapshot attribute names and values for a manual DB snapshot.

When sharing snapshots with other AWS accounts, DescribeDBSnapshotAttributes returns the restore attribute and a list of IDs for the AWS accounts that are authorized to copy or restore the manual DB snapshot. If all is included in the list of values for the restore attribute, then the manual DB snapshot is public and can be copied or restored by all AWS accounts.

To add or remove access for an AWS account to copy or restore a manual DB snapshot, or to make the manual DB snapshot public or private, use the ModifyDBSnapshotAttribute API action.

", - "DescribeDBSnapshots": "

Returns information about DB snapshots. This API action supports pagination.

", - "DescribeDBSubnetGroups": "

Returns a list of DBSubnetGroup descriptions. If a DBSubnetGroupName is specified, the list will contain only the descriptions of the specified DBSubnetGroup.

For an overview of CIDR ranges, go to the Wikipedia Tutorial.

", - "DescribeEngineDefaultClusterParameters": "

Returns the default engine and system parameter information for the cluster database engine.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "DescribeEngineDefaultParameters": "

Returns the default engine and system parameter information for the specified database engine.

", - "DescribeEventCategories": "

Displays a list of categories for all event source types, or, if specified, for a specified source type. You can see a list of the event categories and source types in the Events topic in the Amazon RDS User Guide.

", - "DescribeEventSubscriptions": "

Lists all the subscription descriptions for a customer account. The description for a subscription includes SubscriptionName, SNSTopicARN, CustomerID, SourceType, SourceID, CreationTime, and Status.

If you specify a SubscriptionName, lists the description for that subscription.

", - "DescribeEvents": "

Returns events related to DB instances, DB security groups, DB snapshots, and DB parameter groups for the past 14 days. Events specific to a particular DB instance, DB security group, database snapshot, or DB parameter group can be obtained by providing the name as a parameter. By default, the past hour of events are returned.

", - "DescribeOptionGroupOptions": "

Describes all available options.

", - "DescribeOptionGroups": "

Describes the available option groups.

", - "DescribeOrderableDBInstanceOptions": "

Returns a list of orderable DB instance options for the specified engine.

", - "DescribePendingMaintenanceActions": "

Returns a list of resources (for example, DB instances) that have at least one pending maintenance action.

", - "DescribeReservedDBInstances": "

Returns information about reserved DB instances for this account, or about a specified reserved DB instance.

", - "DescribeReservedDBInstancesOfferings": "

Lists available reserved DB instance offerings.

", - "DescribeSourceRegions": "

Returns a list of the source AWS Regions where the current AWS Region can create a Read Replica or copy a DB snapshot from. This API action supports pagination.

", - "DescribeValidDBInstanceModifications": "

You can call DescribeValidDBInstanceModifications to learn what modifications you can make to your DB instance. You can use this information when you call ModifyDBInstance.

", - "DownloadDBLogFilePortion": "

Downloads all or a portion of the specified log file, up to 1 MB in size.

", - "FailoverDBCluster": "

Forces a failover for a DB cluster.

A failover for a DB cluster promotes one of the Aurora Replicas (read-only instances) in the DB cluster to be the primary instance (the cluster writer).

Amazon Aurora will automatically fail over to an Aurora Replica, if one exists, when the primary instance fails. You can force a failover when you want to simulate a failure of a primary instance for testing. Because each instance in a DB cluster has its own endpoint address, you will need to clean up and re-establish any existing connections that use those endpoint addresses when the failover is complete.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "ListTagsForResource": "

Lists all tags on an Amazon RDS resource.

For an overview on tagging an Amazon RDS resource, see Tagging Amazon RDS Resources.

", - "ModifyDBCluster": "

Modify a setting for an Amazon Aurora DB cluster. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "ModifyDBClusterParameterGroup": "

Modifies the parameters of a DB cluster parameter group. To modify more than one parameter, submit a list of the following: ParameterName, ParameterValue, and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

Changes to dynamic parameters are applied immediately. Changes to static parameters require a reboot without failover to the DB cluster associated with the parameter group before the change can take effect.

After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBClusterParameters command to verify that your DB cluster parameter group has been created or modified.

", - "ModifyDBClusterSnapshotAttribute": "

Adds an attribute and values to, or removes an attribute and values from, a manual DB cluster snapshot.

To share a manual DB cluster snapshot with other AWS accounts, specify restore as the AttributeName and use the ValuesToAdd parameter to add a list of IDs of the AWS accounts that are authorized to restore the manual DB cluster snapshot. Use the value all to make the manual DB cluster snapshot public, which means that it can be copied or restored by all AWS accounts. Do not add the all value for any manual DB cluster snapshots that contain private information that you don't want available to all AWS accounts. If a manual DB cluster snapshot is encrypted, it can be shared, but only by specifying a list of authorized AWS account IDs for the ValuesToAdd parameter. You can't use all as a value for that parameter in this case.

To view which AWS accounts have access to copy or restore a manual DB cluster snapshot, or whether a manual DB cluster snapshot public or private, use the DescribeDBClusterSnapshotAttributes API action.

", - "ModifyDBInstance": "

Modifies settings for a DB instance. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. To learn what modifications you can make to your DB instance, call DescribeValidDBInstanceModifications before you call ModifyDBInstance.

", - "ModifyDBParameterGroup": "

Modifies the parameters of a DB parameter group. To modify more than one parameter, submit a list of the following: ParameterName, ParameterValue, and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

Changes to dynamic parameters are applied immediately. Changes to static parameters require a reboot without failover to the DB instance associated with the parameter group before the change can take effect.

After you modify a DB parameter group, you should wait at least 5 minutes before creating your first DB instance that uses that DB parameter group as the default parameter group. This allows Amazon RDS to fully complete the modify action before the parameter group is used as the default for a new DB instance. This is especially important for parameters that are critical when creating the default database for a DB instance, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBParameters command to verify that your DB parameter group has been created or modified.

", - "ModifyDBSnapshot": "

Updates a manual DB snapshot, which can be encrypted or not encrypted, with a new engine version.

Amazon RDS supports upgrading DB snapshots for MySQL and Oracle.

", - "ModifyDBSnapshotAttribute": "

Adds an attribute and values to, or removes an attribute and values from, a manual DB snapshot.

To share a manual DB snapshot with other AWS accounts, specify restore as the AttributeName and use the ValuesToAdd parameter to add a list of IDs of the AWS accounts that are authorized to restore the manual DB snapshot. Uses the value all to make the manual DB snapshot public, which means it can be copied or restored by all AWS accounts. Do not add the all value for any manual DB snapshots that contain private information that you don't want available to all AWS accounts. If the manual DB snapshot is encrypted, it can be shared, but only by specifying a list of authorized AWS account IDs for the ValuesToAdd parameter. You can't use all as a value for that parameter in this case.

To view which AWS accounts have access to copy or restore a manual DB snapshot, or whether a manual DB snapshot public or private, use the DescribeDBSnapshotAttributes API action.

", - "ModifyDBSubnetGroup": "

Modifies an existing DB subnet group. DB subnet groups must contain at least one subnet in at least two AZs in the AWS Region.

", - "ModifyEventSubscription": "

Modifies an existing RDS event notification subscription. Note that you can't modify the source identifiers using this call; to change source identifiers for a subscription, use the AddSourceIdentifierToSubscription and RemoveSourceIdentifierFromSubscription calls.

You can see a list of the event categories for a given SourceType in the Events topic in the Amazon RDS User Guide or by using the DescribeEventCategories action.

", - "ModifyOptionGroup": "

Modifies an existing option group.

", - "PromoteReadReplica": "

Promotes a Read Replica DB instance to a standalone DB instance.

  • Backup duration is a function of the amount of changes to the database since the previous backup. If you plan to promote a Read Replica to a standalone instance, we recommend that you enable backups and complete at least one backup prior to promotion. In addition, a Read Replica cannot be promoted to a standalone instance when it is in the backing-up status. If you have enabled backups on your Read Replica, configure the automated backup window so that daily backups do not interfere with Read Replica promotion.

  • This command doesn't apply to Aurora MySQL and Aurora PostgreSQL.

", - "PromoteReadReplicaDBCluster": "

Promotes a Read Replica DB cluster to a standalone DB cluster.

", - "PurchaseReservedDBInstancesOffering": "

Purchases a reserved DB instance offering.

", - "RebootDBInstance": "

You might need to reboot your DB instance, usually for maintenance reasons. For example, if you make certain modifications, or if you change the DB parameter group associated with the DB instance, you must reboot the instance for the changes to take effect.

Rebooting a DB instance restarts the database engine service. Rebooting a DB instance results in a momentary outage, during which the DB instance status is set to rebooting.

For more information about rebooting, see Rebooting a DB Instance.

", - "RemoveRoleFromDBCluster": "

Disassociates an Identity and Access Management (IAM) role from an Aurora DB cluster. For more information, see Authorizing Amazon Aurora to Access Other AWS Services On Your Behalf.

", - "RemoveSourceIdentifierFromSubscription": "

Removes a source identifier from an existing RDS event notification subscription.

", - "RemoveTagsFromResource": "

Removes metadata tags from an Amazon RDS resource.

For an overview on tagging an Amazon RDS resource, see Tagging Amazon RDS Resources.

", - "ResetDBClusterParameterGroup": "

Modifies the parameters of a DB cluster parameter group to the default value. To reset specific parameters submit a list of the following: ParameterName and ApplyMethod. To reset the entire DB cluster parameter group, specify the DBClusterParameterGroupName and ResetAllParameters parameters.

When resetting the entire group, dynamic parameters are updated immediately and static parameters are set to pending-reboot to take effect on the next DB instance restart or RebootDBInstance request. You must call RebootDBInstance for every DB instance in your DB cluster that you want the updated static parameter to apply to.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "ResetDBParameterGroup": "

Modifies the parameters of a DB parameter group to the engine/system default value. To reset specific parameters, provide a list of the following: ParameterName and ApplyMethod. To reset the entire DB parameter group, specify the DBParameterGroup name and ResetAllParameters parameters. When resetting the entire group, dynamic parameters are updated immediately and static parameters are set to pending-reboot to take effect on the next DB instance restart or RebootDBInstance request.

", - "RestoreDBClusterFromS3": "

Creates an Amazon Aurora DB cluster from data stored in an Amazon S3 bucket. Amazon RDS must be authorized to access the Amazon S3 bucket and the data must be created using the Percona XtraBackup utility as described in Migrating Data from MySQL by Using an Amazon S3 Bucket.

", - "RestoreDBClusterFromSnapshot": "

Creates a new DB cluster from a DB snapshot or DB cluster snapshot.

If a DB snapshot is specified, the target DB cluster is created from the source DB snapshot with a default configuration and default security group.

If a DB cluster snapshot is specified, the target DB cluster is created from the source DB cluster restore point with the same configuration as the original source DB cluster, except that the new DB cluster is created with the default security group.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "RestoreDBClusterToPointInTime": "

Restores a DB cluster to an arbitrary point in time. Users can restore to any point in time before LatestRestorableTime for up to BackupRetentionPeriod days. The target DB cluster is created from the source DB cluster with the same configuration as the original DB cluster, except that the new DB cluster is created with the default DB security group.

This action only restores the DB cluster, not the DB instances for that DB cluster. You must invoke the CreateDBInstance action to create DB instances for the restored DB cluster, specifying the identifier of the restored DB cluster in DBClusterIdentifier. You can create DB instances only after the RestoreDBClusterToPointInTime action has completed and the DB cluster is available.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", - "RestoreDBInstanceFromDBSnapshot": "

Creates a new DB instance from a DB snapshot. The target database is created from the source database restore point with the most of original configuration with the default security group and the default DB parameter group. By default, the new DB instance is created as a single-AZ deployment except when the instance is a SQL Server instance that has an option group that is associated with mirroring; in this case, the instance becomes a mirrored AZ deployment and not a single-AZ deployment.

If your intent is to replace your original DB instance with the new, restored DB instance, then rename your original DB instance before you call the RestoreDBInstanceFromDBSnapshot action. RDS doesn't allow two DB instances with the same name. Once you have renamed your original DB instance with a different identifier, then you can pass the original name of the DB instance as the DBInstanceIdentifier in the call to the RestoreDBInstanceFromDBSnapshot action. The result is that you will replace the original DB instance with the DB instance created from the snapshot.

If you are restoring from a shared manual DB snapshot, the DBSnapshotIdentifier must be the ARN of the shared DB snapshot.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL. For Aurora, use RestoreDBClusterFromSnapshot.

", - "RestoreDBInstanceFromS3": "

Amazon Relational Database Service (Amazon RDS) supports importing MySQL databases by using backup files. You can create a backup of your on-premises database, store it on Amazon Simple Storage Service (Amazon S3), and then restore the backup file onto a new Amazon RDS DB instance running MySQL. For more information, see Importing Data into an Amazon RDS MySQL DB Instance.

", - "RestoreDBInstanceToPointInTime": "

Restores a DB instance to an arbitrary point in time. You can restore to any point in time before the time identified by the LatestRestorableTime property. You can restore to a point up to the number of days specified by the BackupRetentionPeriod property.

The target database is created with most of the original configuration, but in a system-selected availability zone, with the default security group, the default subnet group, and the default DB parameter group. By default, the new DB instance is created as a single-AZ deployment except when the instance is a SQL Server instance that has an option group that is associated with mirroring; in this case, the instance becomes a mirrored deployment and not a single-AZ deployment.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL. For Aurora, use RestoreDBClusterToPointInTime.

", - "RevokeDBSecurityGroupIngress": "

Revokes ingress from a DBSecurityGroup for previously authorized IP ranges or EC2 or VPC Security Groups. Required parameters for this API are one of CIDRIP, EC2SecurityGroupId for VPC, or (EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId).

", - "StartDBInstance": "

Starts a DB instance that was stopped using the AWS console, the stop-db-instance AWS CLI command, or the StopDBInstance action. For more information, see Stopping and Starting a DB instance in the AWS RDS user guide.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL.

", - "StopDBInstance": "

Stops a DB instance. When you stop a DB instance, Amazon RDS retains the DB instance's metadata, including its endpoint, DB parameter group, and option group membership. Amazon RDS also retains the transaction logs so you can do a point-in-time restore if necessary. For more information, see Stopping and Starting a DB instance in the AWS RDS user guide.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL.

" + "version":"2.0", + "service":"Amazon Relational Database Service

Amazon Relational Database Service (Amazon RDS) is a web service that makes it easier to set up, operate, and scale a relational database in the cloud. It provides cost-efficient, resizable capacity for an industry-standard relational database and manages common database administration tasks, freeing up developers to focus on what makes their applications and businesses unique.

Amazon RDS gives you access to the capabilities of a MySQL, MariaDB, PostgreSQL, Microsoft SQL Server, Oracle, or Amazon Aurora database server. These capabilities mean that the code, applications, and tools you already use today with your existing databases work with Amazon RDS without modification. Amazon RDS automatically backs up your database and maintains the database software that powers your DB instance. Amazon RDS is flexible: you can scale your DB instance's compute resources and storage capacity to meet your application's demand. As with all Amazon Web Services, there are no up-front investments, and you pay only for the resources you use.

This interface reference for Amazon RDS contains documentation for a programming or command line interface you can use to manage Amazon RDS. Note that Amazon RDS is asynchronous, which means that some interfaces might require techniques such as polling or callback functions to determine when a command has been applied. In this reference, the parameter descriptions indicate whether a command is applied immediately, on the next instance reboot, or during the maintenance window. The reference structure is as follows, and we list following some related topics from the user guide.

Amazon RDS API Reference

Amazon RDS User Guide

", + "operations":{ + "AddRoleToDBCluster":"

Associates an Identity and Access Management (IAM) role from an Aurora DB cluster. For more information, see Authorizing Amazon Aurora to Access Other AWS Services On Your Behalf.

", + "AddSourceIdentifierToSubscription":"

Adds a source identifier to an existing RDS event notification subscription.

", + "AddTagsToResource":"

Adds metadata tags to an Amazon RDS resource. These tags can also be used with cost allocation reporting to track cost associated with Amazon RDS resources, or used in a Condition statement in an IAM policy for Amazon RDS.

For an overview on tagging Amazon RDS resources, see Tagging Amazon RDS Resources.

", + "ApplyPendingMaintenanceAction":"

Applies a pending maintenance action to a resource (for example, to a DB instance).

", + "AuthorizeDBSecurityGroupIngress":"

Enables ingress to a DBSecurityGroup using one of two forms of authorization. First, EC2 or VPC security groups can be added to the DBSecurityGroup if the application using the database is running on EC2 or VPC instances. Second, IP ranges are available if the application accessing your database is running on the Internet. Required parameters for this API are one of CIDR range, EC2SecurityGroupId for VPC, or (EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId for non-VPC).

You can't authorize ingress from an EC2 security group in one AWS Region to an Amazon RDS DB instance in another. You can't authorize ingress from a VPC security group in one VPC to an Amazon RDS DB instance in another.

For an overview of CIDR ranges, go to the Wikipedia Tutorial.

", + "BacktrackDBCluster":"

Backtracks a DB cluster to a specific time, without creating a new DB cluster.

For more information on backtracking, see Backtracking an Aurora DB Cluster in the Amazon RDS User Guide.

", + "CopyDBClusterParameterGroup":"

Copies the specified DB cluster parameter group.

", + "CopyDBClusterSnapshot":"

Copies a snapshot of a DB cluster.

To copy a DB cluster snapshot from a shared manual DB cluster snapshot, SourceDBClusterSnapshotIdentifier must be the Amazon Resource Name (ARN) of the shared DB cluster snapshot.

You can copy an encrypted DB cluster snapshot from another AWS Region. In that case, the AWS Region where you call the CopyDBClusterSnapshot action is the destination AWS Region for the encrypted DB cluster snapshot to be copied to. To copy an encrypted DB cluster snapshot from another AWS Region, you must provide the following values:

  • KmsKeyId - The AWS Key Management System (AWS KMS) key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region.

  • PreSignedUrl - A URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot action to be called in the source AWS Region where the DB cluster snapshot is copied from. The pre-signed URL must be a valid request for the CopyDBClusterSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB cluster snapshot to be copied.

    The pre-signed URL request must contain the following parameter values:

    • KmsKeyId - The KMS key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. This is the same identifier for both the CopyDBClusterSnapshot action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

    • DestinationRegion - The name of the AWS Region that the DB cluster snapshot will be created in.

    • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster snapshot from the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:aurora-cluster1-snapshot-20161115.

    To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

  • TargetDBClusterSnapshotIdentifier - The identifier for the new copy of the DB cluster snapshot in the destination AWS Region.

  • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the ARN format for the source AWS Region and is the same value as the SourceDBClusterSnapshotIdentifier in the pre-signed URL.

To cancel the copy operation once it is in progress, delete the target DB cluster snapshot identified by TargetDBClusterSnapshotIdentifier while that DB cluster snapshot is in \"copying\" status.

For more information on copying encrypted DB cluster snapshots from one AWS Region to another, see Copying a DB Cluster Snapshot in the Same Account, Either in the Same Region or Across Regions in the Amazon RDS User Guide.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "CopyDBParameterGroup":"

Copies the specified DB parameter group.

", + "CopyDBSnapshot":"

Copies the specified DB snapshot. The source DB snapshot must be in the \"available\" state.

You can copy a snapshot from one AWS Region to another. In that case, the AWS Region where you call the CopyDBSnapshot action is the destination AWS Region for the DB snapshot copy.

For more information about copying snapshots, see Copying a DB Snapshot in the Amazon RDS User Guide.

", + "CopyOptionGroup":"

Copies the specified option group.

", + "CreateDBCluster":"

Creates a new Amazon Aurora DB cluster.

You can use the ReplicationSourceIdentifier parameter to create the DB cluster as a Read Replica of another DB cluster or Amazon RDS MySQL DB instance. For cross-region replication where the DB cluster identified by ReplicationSourceIdentifier is encrypted, you must also specify the PreSignedUrl parameter.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "CreateDBClusterParameterGroup":"

Creates a new DB cluster parameter group.

Parameters in a DB cluster parameter group apply to all of the instances in a DB cluster.

A DB cluster parameter group is initially created with the default parameters for the database engine used by instances in the DB cluster. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBClusterParameterGroup. Once you've created a DB cluster parameter group, you need to associate it with your DB cluster using ModifyDBCluster. When you associate a new DB cluster parameter group with a running DB cluster, you need to reboot the DB instances in the DB cluster without failover for the new DB cluster parameter group and associated settings to take effect.

After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the DB cluster parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBClusterParameters command to verify that your DB cluster parameter group has been created or modified.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "CreateDBClusterSnapshot":"

Creates a snapshot of a DB cluster. For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "CreateDBInstance":"

Creates a new DB instance.

", + "CreateDBInstanceReadReplica":"

Creates a new DB instance that acts as a Read Replica for an existing source DB instance. You can create a Read Replica for a DB instance running MySQL, MariaDB, or PostgreSQL. For more information, see Working with PostgreSQL, MySQL, and MariaDB Read Replicas.

Amazon Aurora doesn't support this action. You must call the CreateDBInstance action to create a DB instance for an Aurora DB cluster.

All Read Replica DB instances are created with backups disabled. All other DB instance attributes (including DB security groups and DB parameter groups) are inherited from the source DB instance, except as specified following.

Your source DB instance must have backup retention enabled.

", + "CreateDBParameterGroup":"

Creates a new DB parameter group.

A DB parameter group is initially created with the default parameters for the database engine used by the DB instance. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBParameterGroup. Once you've created a DB parameter group, you need to associate it with your DB instance using ModifyDBInstance. When you associate a new DB parameter group with a running DB instance, you need to reboot the DB instance without failover for the new DB parameter group and associated settings to take effect.

After you create a DB parameter group, you should wait at least 5 minutes before creating your first DB instance that uses that DB parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the parameter group is used as the default for a new DB instance. This is especially important for parameters that are critical when creating the default database for a DB instance, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBParameters command to verify that your DB parameter group has been created or modified.

", + "CreateDBSecurityGroup":"

Creates a new DB security group. DB security groups control access to a DB instance.

A DB security group controls access to EC2-Classic DB instances that are not in a VPC.

", + "CreateDBSnapshot":"

Creates a DBSnapshot. The source DBInstance must be in \"available\" state.

", + "CreateDBSubnetGroup":"

Creates a new DB subnet group. DB subnet groups must contain at least one subnet in at least two AZs in the AWS Region.

", + "CreateEventSubscription":"

Creates an RDS event notification subscription. This action requires a topic ARN (Amazon Resource Name) created by either the RDS console, the SNS console, or the SNS API. To obtain an ARN with SNS, you must create a topic in Amazon SNS and subscribe to the topic. The ARN is displayed in the SNS console.

You can specify the type of source (SourceType) you want to be notified of, provide a list of RDS sources (SourceIds) that triggers the events, and provide a list of event categories (EventCategories) for events you want to be notified of. For example, you can specify SourceType = db-instance, SourceIds = mydbinstance1, mydbinstance2 and EventCategories = Availability, Backup.

If you specify both the SourceType and SourceIds, such as SourceType = db-instance and SourceIdentifier = myDBInstance1, you are notified of all the db-instance events for the specified source. If you specify a SourceType but do not specify a SourceIdentifier, you receive notice of the events for that source type for all your RDS sources. If you do not specify either the SourceType nor the SourceIdentifier, you are notified of events generated from all RDS sources belonging to your customer account.

", + "CreateOptionGroup":"

Creates a new option group. You can create up to 20 option groups.

", + "DeleteDBCluster":"

The DeleteDBCluster action deletes a previously provisioned DB cluster. When you delete a DB cluster, all automated backups for that DB cluster are deleted and can't be recovered. Manual DB cluster snapshots of the specified DB cluster are not deleted.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DeleteDBClusterParameterGroup":"

Deletes a specified DB cluster parameter group. The DB cluster parameter group to be deleted can't be associated with any DB clusters.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DeleteDBClusterSnapshot":"

Deletes a DB cluster snapshot. If the snapshot is being copied, the copy operation is terminated.

The DB cluster snapshot must be in the available state to be deleted.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DeleteDBInstance":"

The DeleteDBInstance action deletes a previously provisioned DB instance. When you delete a DB instance, all automated backups for that instance are deleted and can't be recovered. Manual DB snapshots of the DB instance to be deleted by DeleteDBInstance are not deleted.

If you request a final DB snapshot the status of the Amazon RDS DB instance is deleting until the DB snapshot is created. The API action DescribeDBInstance is used to monitor the status of this operation. The action can't be canceled or reverted once submitted.

Note that when a DB instance is in a failure state and has a status of failed, incompatible-restore, or incompatible-network, you can only delete it when the SkipFinalSnapshot parameter is set to true.

If the specified DB instance is part of an Amazon Aurora DB cluster, you can't delete the DB instance if both of the following conditions are true:

  • The DB cluster is a Read Replica of another Amazon Aurora DB cluster.

  • The DB instance is the only instance in the DB cluster.

To delete a DB instance in this case, first call the PromoteReadReplicaDBCluster API action to promote the DB cluster so it's no longer a Read Replica. After the promotion completes, then call the DeleteDBInstance API action to delete the final instance in the DB cluster.

", + "DeleteDBParameterGroup":"

Deletes a specified DBParameterGroup. The DBParameterGroup to be deleted can't be associated with any DB instances.

", + "DeleteDBSecurityGroup":"

Deletes a DB security group.

The specified DB security group must not be associated with any DB instances.

", + "DeleteDBSnapshot":"

Deletes a DBSnapshot. If the snapshot is being copied, the copy operation is terminated.

The DBSnapshot must be in the available state to be deleted.

", + "DeleteDBSubnetGroup":"

Deletes a DB subnet group.

The specified database subnet group must not be associated with any DB instances.

", + "DeleteEventSubscription":"

Deletes an RDS event notification subscription.

", + "DeleteOptionGroup":"

Deletes an existing option group.

", + "DescribeAccountAttributes":"

Lists all of the attributes for a customer account. The attributes include Amazon RDS quotas for the account, such as the number of DB instances allowed. The description for a quota includes the quota name, current usage toward that quota, and the quota's maximum value.

This command doesn't take any parameters.

", + "DescribeCertificates":"

Lists the set of CA certificates provided by Amazon RDS for this AWS account.

", + "DescribeDBClusterBacktracks":"

Returns information about backtracks for a DB cluster.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DescribeDBClusterParameterGroups":"

Returns a list of DBClusterParameterGroup descriptions. If a DBClusterParameterGroupName parameter is specified, the list will contain only the description of the specified DB cluster parameter group.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DescribeDBClusterParameters":"

Returns the detailed parameter list for a particular DB cluster parameter group.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DescribeDBClusterSnapshotAttributes":"

Returns a list of DB cluster snapshot attribute names and values for a manual DB cluster snapshot.

When sharing snapshots with other AWS accounts, DescribeDBClusterSnapshotAttributes returns the restore attribute and a list of IDs for the AWS accounts that are authorized to copy or restore the manual DB cluster snapshot. If all is included in the list of values for the restore attribute, then the manual DB cluster snapshot is public and can be copied or restored by all AWS accounts.

To add or remove access for an AWS account to copy or restore a manual DB cluster snapshot, or to make the manual DB cluster snapshot public or private, use the ModifyDBClusterSnapshotAttribute API action.

", + "DescribeDBClusterSnapshots":"

Returns information about DB cluster snapshots. This API action supports pagination.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DescribeDBClusters":"

Returns information about provisioned Aurora DB clusters. This API supports pagination.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DescribeDBEngineVersions":"

Returns a list of the available DB engines.

", + "DescribeDBInstances":"

Returns information about provisioned RDS instances. This API supports pagination.

", + "DescribeDBLogFiles":"

Returns a list of DB log files for the DB instance.

", + "DescribeDBParameterGroups":"

Returns a list of DBParameterGroup descriptions. If a DBParameterGroupName is specified, the list will contain only the description of the specified DB parameter group.

", + "DescribeDBParameters":"

Returns the detailed parameter list for a particular DB parameter group.

", + "DescribeDBSecurityGroups":"

Returns a list of DBSecurityGroup descriptions. If a DBSecurityGroupName is specified, the list will contain only the descriptions of the specified DB security group.

", + "DescribeDBSnapshotAttributes":"

Returns a list of DB snapshot attribute names and values for a manual DB snapshot.

When sharing snapshots with other AWS accounts, DescribeDBSnapshotAttributes returns the restore attribute and a list of IDs for the AWS accounts that are authorized to copy or restore the manual DB snapshot. If all is included in the list of values for the restore attribute, then the manual DB snapshot is public and can be copied or restored by all AWS accounts.

To add or remove access for an AWS account to copy or restore a manual DB snapshot, or to make the manual DB snapshot public or private, use the ModifyDBSnapshotAttribute API action.

", + "DescribeDBSnapshots":"

Returns information about DB snapshots. This API action supports pagination.

", + "DescribeDBSubnetGroups":"

Returns a list of DBSubnetGroup descriptions. If a DBSubnetGroupName is specified, the list will contain only the descriptions of the specified DBSubnetGroup.

For an overview of CIDR ranges, go to the Wikipedia Tutorial.

", + "DescribeEngineDefaultClusterParameters":"

Returns the default engine and system parameter information for the cluster database engine.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "DescribeEngineDefaultParameters":"

Returns the default engine and system parameter information for the specified database engine.

", + "DescribeEventCategories":"

Displays a list of categories for all event source types, or, if specified, for a specified source type. You can see a list of the event categories and source types in the Events topic in the Amazon RDS User Guide.

", + "DescribeEventSubscriptions":"

Lists all the subscription descriptions for a customer account. The description for a subscription includes SubscriptionName, SNSTopicARN, CustomerID, SourceType, SourceID, CreationTime, and Status.

If you specify a SubscriptionName, lists the description for that subscription.

", + "DescribeEvents":"

Returns events related to DB instances, DB security groups, DB snapshots, and DB parameter groups for the past 14 days. Events specific to a particular DB instance, DB security group, database snapshot, or DB parameter group can be obtained by providing the name as a parameter. By default, the past hour of events are returned.

", + "DescribeOptionGroupOptions":"

Describes all available options.

", + "DescribeOptionGroups":"

Describes the available option groups.

", + "DescribeOrderableDBInstanceOptions":"

Returns a list of orderable DB instance options for the specified engine.

", + "DescribePendingMaintenanceActions":"

Returns a list of resources (for example, DB instances) that have at least one pending maintenance action.

", + "DescribeReservedDBInstances":"

Returns information about reserved DB instances for this account, or about a specified reserved DB instance.

", + "DescribeReservedDBInstancesOfferings":"

Lists available reserved DB instance offerings.

", + "DescribeSourceRegions":"

Returns a list of the source AWS Regions where the current AWS Region can create a Read Replica or copy a DB snapshot from. This API action supports pagination.

", + "DescribeValidDBInstanceModifications":"

You can call DescribeValidDBInstanceModifications to learn what modifications you can make to your DB instance. You can use this information when you call ModifyDBInstance.

", + "DownloadDBLogFilePortion":"

Downloads all or a portion of the specified log file, up to 1 MB in size.

", + "FailoverDBCluster":"

Forces a failover for a DB cluster.

A failover for a DB cluster promotes one of the Aurora Replicas (read-only instances) in the DB cluster to be the primary instance (the cluster writer).

Amazon Aurora will automatically fail over to an Aurora Replica, if one exists, when the primary instance fails. You can force a failover when you want to simulate a failure of a primary instance for testing. Because each instance in a DB cluster has its own endpoint address, you will need to clean up and re-establish any existing connections that use those endpoint addresses when the failover is complete.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "ListTagsForResource":"

Lists all tags on an Amazon RDS resource.

For an overview on tagging an Amazon RDS resource, see Tagging Amazon RDS Resources.

", + "ModifyDBCluster":"

Modify a setting for an Amazon Aurora DB cluster. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "ModifyDBClusterParameterGroup":"

Modifies the parameters of a DB cluster parameter group. To modify more than one parameter, submit a list of the following: ParameterName, ParameterValue, and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

Changes to dynamic parameters are applied immediately. Changes to static parameters require a reboot without failover to the DB cluster associated with the parameter group before the change can take effect.

After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBClusterParameters command to verify that your DB cluster parameter group has been created or modified.

", + "ModifyDBClusterSnapshotAttribute":"

Adds an attribute and values to, or removes an attribute and values from, a manual DB cluster snapshot.

To share a manual DB cluster snapshot with other AWS accounts, specify restore as the AttributeName and use the ValuesToAdd parameter to add a list of IDs of the AWS accounts that are authorized to restore the manual DB cluster snapshot. Use the value all to make the manual DB cluster snapshot public, which means that it can be copied or restored by all AWS accounts. Do not add the all value for any manual DB cluster snapshots that contain private information that you don't want available to all AWS accounts. If a manual DB cluster snapshot is encrypted, it can be shared, but only by specifying a list of authorized AWS account IDs for the ValuesToAdd parameter. You can't use all as a value for that parameter in this case.

To view which AWS accounts have access to copy or restore a manual DB cluster snapshot, or whether a manual DB cluster snapshot public or private, use the DescribeDBClusterSnapshotAttributes API action.

", + "ModifyDBInstance":"

Modifies settings for a DB instance. You can change one or more database configuration parameters by specifying these parameters and the new values in the request. To learn what modifications you can make to your DB instance, call DescribeValidDBInstanceModifications before you call ModifyDBInstance.

", + "ModifyDBParameterGroup":"

Modifies the parameters of a DB parameter group. To modify more than one parameter, submit a list of the following: ParameterName, ParameterValue, and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

Changes to dynamic parameters are applied immediately. Changes to static parameters require a reboot without failover to the DB instance associated with the parameter group before the change can take effect.

After you modify a DB parameter group, you should wait at least 5 minutes before creating your first DB instance that uses that DB parameter group as the default parameter group. This allows Amazon RDS to fully complete the modify action before the parameter group is used as the default for a new DB instance. This is especially important for parameters that are critical when creating the default database for a DB instance, such as the character set for the default database defined by the character_set_database parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBParameters command to verify that your DB parameter group has been created or modified.

", + "ModifyDBSnapshot":"

Updates a manual DB snapshot, which can be encrypted or not encrypted, with a new engine version.

Amazon RDS supports upgrading DB snapshots for MySQL and Oracle.

", + "ModifyDBSnapshotAttribute":"

Adds an attribute and values to, or removes an attribute and values from, a manual DB snapshot.

To share a manual DB snapshot with other AWS accounts, specify restore as the AttributeName and use the ValuesToAdd parameter to add a list of IDs of the AWS accounts that are authorized to restore the manual DB snapshot. Uses the value all to make the manual DB snapshot public, which means it can be copied or restored by all AWS accounts. Do not add the all value for any manual DB snapshots that contain private information that you don't want available to all AWS accounts. If the manual DB snapshot is encrypted, it can be shared, but only by specifying a list of authorized AWS account IDs for the ValuesToAdd parameter. You can't use all as a value for that parameter in this case.

To view which AWS accounts have access to copy or restore a manual DB snapshot, or whether a manual DB snapshot public or private, use the DescribeDBSnapshotAttributes API action.

", + "ModifyDBSubnetGroup":"

Modifies an existing DB subnet group. DB subnet groups must contain at least one subnet in at least two AZs in the AWS Region.

", + "ModifyEventSubscription":"

Modifies an existing RDS event notification subscription. Note that you can't modify the source identifiers using this call; to change source identifiers for a subscription, use the AddSourceIdentifierToSubscription and RemoveSourceIdentifierFromSubscription calls.

You can see a list of the event categories for a given SourceType in the Events topic in the Amazon RDS User Guide or by using the DescribeEventCategories action.

", + "ModifyOptionGroup":"

Modifies an existing option group.

", + "PromoteReadReplica":"

Promotes a Read Replica DB instance to a standalone DB instance.

  • Backup duration is a function of the amount of changes to the database since the previous backup. If you plan to promote a Read Replica to a standalone instance, we recommend that you enable backups and complete at least one backup prior to promotion. In addition, a Read Replica cannot be promoted to a standalone instance when it is in the backing-up status. If you have enabled backups on your Read Replica, configure the automated backup window so that daily backups do not interfere with Read Replica promotion.

  • This command doesn't apply to Aurora MySQL and Aurora PostgreSQL.

", + "PromoteReadReplicaDBCluster":"

Promotes a Read Replica DB cluster to a standalone DB cluster.

", + "PurchaseReservedDBInstancesOffering":"

Purchases a reserved DB instance offering.

", + "RebootDBInstance":"

You might need to reboot your DB instance, usually for maintenance reasons. For example, if you make certain modifications, or if you change the DB parameter group associated with the DB instance, you must reboot the instance for the changes to take effect.

Rebooting a DB instance restarts the database engine service. Rebooting a DB instance results in a momentary outage, during which the DB instance status is set to rebooting.

For more information about rebooting, see Rebooting a DB Instance.

", + "RemoveRoleFromDBCluster":"

Disassociates an Identity and Access Management (IAM) role from an Aurora DB cluster. For more information, see Authorizing Amazon Aurora to Access Other AWS Services On Your Behalf.

", + "RemoveSourceIdentifierFromSubscription":"

Removes a source identifier from an existing RDS event notification subscription.

", + "RemoveTagsFromResource":"

Removes metadata tags from an Amazon RDS resource.

For an overview on tagging an Amazon RDS resource, see Tagging Amazon RDS Resources.

", + "ResetDBClusterParameterGroup":"

Modifies the parameters of a DB cluster parameter group to the default value. To reset specific parameters submit a list of the following: ParameterName and ApplyMethod. To reset the entire DB cluster parameter group, specify the DBClusterParameterGroupName and ResetAllParameters parameters.

When resetting the entire group, dynamic parameters are updated immediately and static parameters are set to pending-reboot to take effect on the next DB instance restart or RebootDBInstance request. You must call RebootDBInstance for every DB instance in your DB cluster that you want the updated static parameter to apply to.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "ResetDBParameterGroup":"

Modifies the parameters of a DB parameter group to the engine/system default value. To reset specific parameters, provide a list of the following: ParameterName and ApplyMethod. To reset the entire DB parameter group, specify the DBParameterGroup name and ResetAllParameters parameters. When resetting the entire group, dynamic parameters are updated immediately and static parameters are set to pending-reboot to take effect on the next DB instance restart or RebootDBInstance request.

", + "RestoreDBClusterFromS3":"

Creates an Amazon Aurora DB cluster from data stored in an Amazon S3 bucket. Amazon RDS must be authorized to access the Amazon S3 bucket and the data must be created using the Percona XtraBackup utility as described in Migrating Data from MySQL by Using an Amazon S3 Bucket.

", + "RestoreDBClusterFromSnapshot":"

Creates a new DB cluster from a DB snapshot or DB cluster snapshot.

If a DB snapshot is specified, the target DB cluster is created from the source DB snapshot with a default configuration and default security group.

If a DB cluster snapshot is specified, the target DB cluster is created from the source DB cluster restore point with the same configuration as the original source DB cluster, except that the new DB cluster is created with the default security group.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "RestoreDBClusterToPointInTime":"

Restores a DB cluster to an arbitrary point in time. Users can restore to any point in time before LatestRestorableTime for up to BackupRetentionPeriod days. The target DB cluster is created from the source DB cluster with the same configuration as the original DB cluster, except that the new DB cluster is created with the default DB security group.

This action only restores the DB cluster, not the DB instances for that DB cluster. You must invoke the CreateDBInstance action to create DB instances for the restored DB cluster, specifying the identifier of the restored DB cluster in DBClusterIdentifier. You can create DB instances only after the RestoreDBClusterToPointInTime action has completed and the DB cluster is available.

For more information on Amazon Aurora, see Aurora on Amazon RDS in the Amazon RDS User Guide.

", + "RestoreDBInstanceFromDBSnapshot":"

Creates a new DB instance from a DB snapshot. The target database is created from the source database restore point with the most of original configuration with the default security group and the default DB parameter group. By default, the new DB instance is created as a single-AZ deployment except when the instance is a SQL Server instance that has an option group that is associated with mirroring; in this case, the instance becomes a mirrored AZ deployment and not a single-AZ deployment.

If your intent is to replace your original DB instance with the new, restored DB instance, then rename your original DB instance before you call the RestoreDBInstanceFromDBSnapshot action. RDS doesn't allow two DB instances with the same name. Once you have renamed your original DB instance with a different identifier, then you can pass the original name of the DB instance as the DBInstanceIdentifier in the call to the RestoreDBInstanceFromDBSnapshot action. The result is that you will replace the original DB instance with the DB instance created from the snapshot.

If you are restoring from a shared manual DB snapshot, the DBSnapshotIdentifier must be the ARN of the shared DB snapshot.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL. For Aurora, use RestoreDBClusterFromSnapshot.

", + "RestoreDBInstanceFromS3":"

Amazon Relational Database Service (Amazon RDS) supports importing MySQL databases by using backup files. You can create a backup of your on-premises database, store it on Amazon Simple Storage Service (Amazon S3), and then restore the backup file onto a new Amazon RDS DB instance running MySQL. For more information, see Importing Data into an Amazon RDS MySQL DB Instance.

", + "RestoreDBInstanceToPointInTime":"

Restores a DB instance to an arbitrary point in time. You can restore to any point in time before the time identified by the LatestRestorableTime property. You can restore to a point up to the number of days specified by the BackupRetentionPeriod property.

The target database is created with most of the original configuration, but in a system-selected Availability Zone, with the default security group, the default subnet group, and the default DB parameter group. By default, the new DB instance is created as a single-AZ deployment except when the instance is a SQL Server instance that has an option group that is associated with mirroring; in this case, the instance becomes a mirrored deployment and not a single-AZ deployment.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL. For Aurora, use RestoreDBClusterToPointInTime.

", + "RevokeDBSecurityGroupIngress":"

Revokes ingress from a DBSecurityGroup for previously authorized IP ranges or EC2 or VPC Security Groups. Required parameters for this API are one of CIDRIP, EC2SecurityGroupId for VPC, or (EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId).

", + "StartDBInstance":"

Starts a DB instance that was stopped using the AWS console, the stop-db-instance AWS CLI command, or the StopDBInstance action. For more information, see Stopping and Starting a DB instance in the AWS RDS user guide.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL.

", + "StopDBInstance":"

Stops a DB instance. When you stop a DB instance, Amazon RDS retains the DB instance's metadata, including its endpoint, DB parameter group, and option group membership. Amazon RDS also retains the transaction logs so you can do a point-in-time restore if necessary. For more information, see Stopping and Starting a DB instance in the AWS RDS user guide.

This command doesn't apply to Aurora MySQL and Aurora PostgreSQL.

" }, - "shapes": { - "AccountAttributesMessage": { - "base": "

Data returned by the DescribeAccountAttributes action.

", - "refs": { + "shapes":{ + "AccountAttributesMessage":{ + "base":"

Data returned by the DescribeAccountAttributes action.

", + "refs":{ } }, - "AccountQuota": { - "base": "

Describes a quota for an AWS account, for example, the number of DB instances allowed.

", - "refs": { - "AccountQuotaList$member": null + "AccountQuota":{ + "base":"

Describes a quota for an AWS account, for example, the number of DB instances allowed.

", + "refs":{ + "AccountQuotaList$member":null } }, - "AccountQuotaList": { - "base": null, - "refs": { - "AccountAttributesMessage$AccountQuotas": "

A list of AccountQuota objects. Within this list, each quota has a name, a count of usage toward the quota maximum, and a maximum value for the quota.

" + "AccountQuotaList":{ + "base":null, + "refs":{ + "AccountAttributesMessage$AccountQuotas":"

A list of AccountQuota objects. Within this list, each quota has a name, a count of usage toward the quota maximum, and a maximum value for the quota.

" } }, - "AddRoleToDBClusterMessage": { - "base": null, - "refs": { + "AddRoleToDBClusterMessage":{ + "base":null, + "refs":{ } }, - "AddSourceIdentifierToSubscriptionMessage": { - "base": "

", - "refs": { + "AddSourceIdentifierToSubscriptionMessage":{ + "base":"

", + "refs":{ } }, - "AddSourceIdentifierToSubscriptionResult": { - "base": null, - "refs": { + "AddSourceIdentifierToSubscriptionResult":{ + "base":null, + "refs":{ } }, - "AddTagsToResourceMessage": { - "base": "

", - "refs": { + "AddTagsToResourceMessage":{ + "base":"

", + "refs":{ } }, - "ApplyMethod": { - "base": null, - "refs": { - "Parameter$ApplyMethod": "

Indicates when to apply parameter updates.

" + "ApplyMethod":{ + "base":null, + "refs":{ + "Parameter$ApplyMethod":"

Indicates when to apply parameter updates.

" } }, - "ApplyPendingMaintenanceActionMessage": { - "base": "

", - "refs": { + "ApplyPendingMaintenanceActionMessage":{ + "base":"

", + "refs":{ } }, - "ApplyPendingMaintenanceActionResult": { - "base": null, - "refs": { + "ApplyPendingMaintenanceActionResult":{ + "base":null, + "refs":{ } }, - "AttributeValueList": { - "base": null, - "refs": { - "DBClusterSnapshotAttribute$AttributeValues": "

The value(s) for the manual DB cluster snapshot attribute.

If the AttributeName field is set to restore, then this element returns a list of IDs of the AWS accounts that are authorized to copy or restore the manual DB cluster snapshot. If a value of all is in the list, then the manual DB cluster snapshot is public and available for any AWS account to copy or restore.

", - "DBSnapshotAttribute$AttributeValues": "

The value or values for the manual DB snapshot attribute.

If the AttributeName field is set to restore, then this element returns a list of IDs of the AWS accounts that are authorized to copy or restore the manual DB snapshot. If a value of all is in the list, then the manual DB snapshot is public and available for any AWS account to copy or restore.

", - "ModifyDBClusterSnapshotAttributeMessage$ValuesToAdd": "

A list of DB cluster snapshot attributes to add to the attribute specified by AttributeName.

To authorize other AWS accounts to copy or restore a manual DB cluster snapshot, set this list to include one or more AWS account IDs, or all to make the manual DB cluster snapshot restorable by any AWS account. Do not add the all value for any manual DB cluster snapshots that contain private information that you don't want available to all AWS accounts.

", - "ModifyDBClusterSnapshotAttributeMessage$ValuesToRemove": "

A list of DB cluster snapshot attributes to remove from the attribute specified by AttributeName.

To remove authorization for other AWS accounts to copy or restore a manual DB cluster snapshot, set this list to include one or more AWS account identifiers, or all to remove authorization for any AWS account to copy or restore the DB cluster snapshot. If you specify all, an AWS account whose account ID is explicitly added to the restore attribute can still copy or restore a manual DB cluster snapshot.

", - "ModifyDBSnapshotAttributeMessage$ValuesToAdd": "

A list of DB snapshot attributes to add to the attribute specified by AttributeName.

To authorize other AWS accounts to copy or restore a manual snapshot, set this list to include one or more AWS account IDs, or all to make the manual DB snapshot restorable by any AWS account. Do not add the all value for any manual DB snapshots that contain private information that you don't want available to all AWS accounts.

", - "ModifyDBSnapshotAttributeMessage$ValuesToRemove": "

A list of DB snapshot attributes to remove from the attribute specified by AttributeName.

To remove authorization for other AWS accounts to copy or restore a manual snapshot, set this list to include one or more AWS account identifiers, or all to remove authorization for any AWS account to copy or restore the DB snapshot. If you specify all, an AWS account whose account ID is explicitly added to the restore attribute can still copy or restore the manual DB snapshot.

" - } - }, - "AuthorizationAlreadyExistsFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group is already authorized for the specified DB security group.

", - "refs": { - } - }, - "AuthorizationNotFoundFault": { - "base": "

The specified CIDRIP or Amazon EC2 security group isn't authorized for the specified DB security group.

RDS also may not be authorized by using IAM to perform necessary actions on your behalf.

", - "refs": { - } - }, - "AuthorizationQuotaExceededFault": { - "base": "

The DB security group authorization quota has been reached.

", - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressMessage": { - "base": "

", - "refs": { - } - }, - "AuthorizeDBSecurityGroupIngressResult": { - "base": null, - "refs": { - } - }, - "AvailabilityZone": { - "base": "

Contains Availability Zone information.

This data type is used as an element in the following data type:

", - "refs": { - "AvailabilityZoneList$member": null, - "Subnet$SubnetAvailabilityZone": null - } - }, - "AvailabilityZoneList": { - "base": null, - "refs": { - "OrderableDBInstanceOption$AvailabilityZones": "

A list of Availability Zones for a DB instance.

" - } - }, - "AvailabilityZones": { - "base": null, - "refs": { - "CreateDBClusterMessage$AvailabilityZones": "

A list of EC2 Availability Zones that instances in the DB cluster can be created in. For information on AWS Regions and Availability Zones, see Regions and Availability Zones.

", - "DBCluster$AvailabilityZones": "

Provides the list of EC2 Availability Zones that instances in the DB cluster can be created in.

", - "DBClusterSnapshot$AvailabilityZones": "

Provides the list of EC2 Availability Zones that instances in the DB cluster snapshot can be restored in.

", - "RestoreDBClusterFromS3Message$AvailabilityZones": "

A list of EC2 Availability Zones that instances in the restored DB cluster can be created in.

", - "RestoreDBClusterFromSnapshotMessage$AvailabilityZones": "

Provides the list of EC2 Availability Zones that instances in the restored DB cluster can be created in.

" - } - }, - "BacktrackDBClusterMessage": { - "base": "

", - "refs": { - } - }, - "Boolean": { - "base": null, - "refs": { - "DBCluster$MultiAZ": "

Specifies whether the DB cluster has instances in multiple Availability Zones.

", - "DBCluster$StorageEncrypted": "

Specifies whether the DB cluster is encrypted.

", - "DBCluster$IAMDatabaseAuthenticationEnabled": "

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", - "DBClusterMember$IsClusterWriter": "

Value that is true if the cluster member is the primary instance for the DB cluster and false otherwise.

", - "DBClusterSnapshot$StorageEncrypted": "

Specifies whether the DB cluster snapshot is encrypted.

", - "DBClusterSnapshot$IAMDatabaseAuthenticationEnabled": "

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", - "DBEngineVersion$SupportsLogExportsToCloudwatchLogs": "

A value that indicates whether the engine version supports exporting the log types specified by ExportableLogTypes to CloudWatch Logs.

", - "DBEngineVersion$SupportsReadReplica": "

Indicates whether the database engine version supports read replicas.

", - "DBInstance$MultiAZ": "

Specifies if the DB instance is a Multi-AZ deployment.

", - "DBInstance$AutoMinorVersionUpgrade": "

Indicates that minor version patches are applied automatically.

", - "DBInstance$PubliclyAccessible": "

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC:true

  • VPC:false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", - "DBInstance$StorageEncrypted": "

Specifies whether the DB instance is encrypted.

", - "DBInstance$CopyTagsToSnapshot": "

Specifies whether tags are copied from the DB instance to snapshots of the DB instance.

", - "DBInstance$IAMDatabaseAuthenticationEnabled": "

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

IAM database authentication can be enabled for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

  • Aurora 5.6 or higher. To enable IAM database authentication for Aurora, see DBCluster Type.

", - "DBInstanceStatusInfo$Normal": "

Boolean value that is true if the instance is operating normally, or false if the instance is in an error state.

", - "DBSnapshot$Encrypted": "

Specifies whether the DB snapshot is encrypted.

", - "DBSnapshot$IAMDatabaseAuthenticationEnabled": "

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", - "DeleteDBClusterMessage$SkipFinalSnapshot": "

Determines whether a final DB cluster snapshot is created before the DB cluster is deleted. If true is specified, no DB cluster snapshot is created. If false is specified, a DB cluster snapshot is created before the DB cluster is deleted.

You must specify a FinalDBSnapshotIdentifier parameter if SkipFinalSnapshot is false.

Default: false

", - "DeleteDBInstanceMessage$SkipFinalSnapshot": "

Determines whether a final DB snapshot is created before the DB instance is deleted. If true is specified, no DBSnapshot is created. If false is specified, a DB snapshot is created before the DB instance is deleted.

Note that when a DB instance is in a failure state and has a status of 'failed', 'incompatible-restore', or 'incompatible-network', it can only be deleted when the SkipFinalSnapshot parameter is set to \"true\".

Specify true when deleting a Read Replica.

The FinalDBSnapshotIdentifier parameter must be specified if SkipFinalSnapshot is false.

Default: false

", - "DescribeDBClusterSnapshotsMessage$IncludeShared": "

True to include shared manual DB cluster snapshots from other AWS accounts that this AWS account has been given permission to copy or restore, and otherwise false. The default is false.

You can give an AWS account permission to restore a manual DB cluster snapshot from another AWS account by the ModifyDBClusterSnapshotAttribute API action.

", - "DescribeDBClusterSnapshotsMessage$IncludePublic": "

True to include manual DB cluster snapshots that are public and can be copied or restored by any AWS account, and otherwise false. The default is false. The default is false.

You can share a manual DB cluster snapshot as public by using the ModifyDBClusterSnapshotAttribute API action.

", - "DescribeDBEngineVersionsMessage$DefaultOnly": "

Indicates that only the default version of the specified engine or engine and major version combination is returned.

", - "DescribeDBSnapshotsMessage$IncludeShared": "

True to include shared manual DB snapshots from other AWS accounts that this AWS account has been given permission to copy or restore, and otherwise false. The default is false.

You can give an AWS account permission to restore a manual DB snapshot from another AWS account by using the ModifyDBSnapshotAttribute API action.

", - "DescribeDBSnapshotsMessage$IncludePublic": "

True to include manual DB snapshots that are public and can be copied or restored by any AWS account, and otherwise false. The default is false.

You can share a manual DB snapshot as public by using the ModifyDBSnapshotAttribute API.

", - "DownloadDBLogFilePortionDetails$AdditionalDataPending": "

Boolean value that if true, indicates there is more data to be downloaded.

", - "EventSubscription$Enabled": "

A Boolean value indicating if the subscription is enabled. True indicates the subscription is enabled.

", - "ModifyDBClusterMessage$ApplyImmediately": "

A value that specifies whether the modifications in this request and any pending modifications are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow setting for the DB cluster. If this parameter is set to false, changes to the DB cluster are applied during the next maintenance window.

The ApplyImmediately parameter only affects the NewDBClusterIdentifier and MasterUserPassword values. If you set the ApplyImmediately parameter value to false, then changes to the NewDBClusterIdentifier and MasterUserPassword values are applied during the next maintenance window. All other changes are applied immediately, regardless of the value of the ApplyImmediately parameter.

Default: false

", - "ModifyDBInstanceMessage$ApplyImmediately": "

Specifies whether the modifications in this request and any pending modifications are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow setting for the DB instance.

If this parameter is set to false, changes to the DB instance are applied during the next maintenance window. Some parameter changes can cause an outage and are applied on the next call to RebootDBInstance, or the next failure reboot. Review the table of parameters in Modifying a DB Instance and Using the Apply Immediately Parameter to see the impact that setting ApplyImmediately to true or false has for each modified parameter and to determine when the changes are applied.

Default: false

", - "ModifyDBInstanceMessage$AllowMajorVersionUpgrade": "

Indicates that major version upgrades are allowed. Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible.

Constraints: This parameter must be set to true when specifying a value for the EngineVersion parameter that is a different major version than the DB instance's current version.

", - "ModifyOptionGroupMessage$ApplyImmediately": "

Indicates whether the changes should be applied immediately, or during the next maintenance window for each instance associated with the option group.

", - "Option$Persistent": "

Indicate if this option is persistent.

", - "Option$Permanent": "

Indicate if this option is permanent.

", - "OptionGroup$AllowsVpcAndNonVpcInstanceMemberships": "

Indicates whether this option group can be applied to both VPC and non-VPC instances. The value true indicates the option group can be applied to both VPC and non-VPC instances.

", - "OptionGroupOption$PortRequired": "

Specifies whether the option requires a port.

", - "OptionGroupOption$Persistent": "

Persistent options can't be removed from an option group while DB instances are associated with the option group. If you disassociate all DB instances from the option group, your can remove the persistent option from the option group.

", - "OptionGroupOption$Permanent": "

Permanent options can never be removed from an option group. An option group containing a permanent option can't be removed from a DB instance.

", - "OptionGroupOption$RequiresAutoMinorEngineVersionUpgrade": "

If true, you must enable the Auto Minor Version Upgrade setting for your DB instance before you can use this option. You can enable Auto Minor Version Upgrade when you first create your DB instance, or by modifying your DB instance later.

", - "OptionGroupOption$VpcOnly": "

If true, you can only use this option with a DB instance that is in a VPC.

", - "OptionGroupOptionSetting$IsModifiable": "

Boolean value where true indicates that this option group option can be changed from the default value.

", - "OptionSetting$IsModifiable": "

A Boolean value that, when true, indicates the option setting can be modified from the default.

", - "OptionSetting$IsCollection": "

Indicates if the option setting is part of a collection.

", - "OptionVersion$IsDefault": "

True if the version is the default version of the option, and otherwise false.

", - "OrderableDBInstanceOption$MultiAZCapable": "

Indicates whether a DB instance is Multi-AZ capable.

", - "OrderableDBInstanceOption$ReadReplicaCapable": "

Indicates whether a DB instance can have a Read Replica.

", - "OrderableDBInstanceOption$Vpc": "

Indicates whether a DB instance is in a VPC.

", - "OrderableDBInstanceOption$SupportsStorageEncryption": "

Indicates whether a DB instance supports encrypted storage.

", - "OrderableDBInstanceOption$SupportsIops": "

Indicates whether a DB instance supports provisioned IOPS.

", - "OrderableDBInstanceOption$SupportsEnhancedMonitoring": "

Indicates whether a DB instance supports Enhanced Monitoring at intervals from 1 to 60 seconds.

", - "OrderableDBInstanceOption$SupportsIAMDatabaseAuthentication": "

Indicates whether a DB instance supports IAM database authentication.

", - "OrderableDBInstanceOption$SupportsPerformanceInsights": "

True if a DB instance supports Performance Insights, otherwise false.

", - "Parameter$IsModifiable": "

Indicates whether (true) or not (false) the parameter can be modified. Some parameters have security or operational implications that prevent them from being changed.

", - "ReservedDBInstance$MultiAZ": "

Indicates if the reservation applies to Multi-AZ deployments.

", - "ReservedDBInstancesOffering$MultiAZ": "

Indicates if the offering applies to Multi-AZ deployments.

", - "ResetDBClusterParameterGroupMessage$ResetAllParameters": "

A value that is set to true to reset all parameters in the DB cluster parameter group to their default values, and false otherwise. You can't use this parameter if there is a list of parameter names specified for the Parameters parameter.

", - "ResetDBParameterGroupMessage$ResetAllParameters": "

Specifies whether (true) or not (false) to reset all parameters in the DB parameter group to default values.

Default: true

", - "RestoreDBClusterToPointInTimeMessage$UseLatestRestorableTime": "

A value that is set to true to restore the DB cluster to the latest restorable backup time, and false otherwise.

Default: false

Constraints: Cannot be specified if RestoreToTime parameter is provided.

", - "RestoreDBInstanceToPointInTimeMessage$UseLatestRestorableTime": "

Specifies whether (true) or not (false) the DB instance is restored from the latest backup time.

Default: false

Constraints: Cannot be specified if RestoreTime parameter is provided.

", - "UpgradeTarget$AutoUpgrade": "

A value that indicates whether the target version is applied to any source DB instances that have AutoMinorVersionUpgrade set to true.

", - "UpgradeTarget$IsMajorVersionUpgrade": "

A value that indicates whether a database engine is upgraded to a major version.

" - } - }, - "BooleanOptional": { - "base": null, - "refs": { - "BacktrackDBClusterMessage$Force": "

A value that, if specified, forces the DB cluster to backtrack when binary logging is enabled. Otherwise, an error occurs when binary logging is enabled.

", - "BacktrackDBClusterMessage$UseEarliestTimeOnPointInTimeUnavailable": "

If BacktrackTo is set to a timestamp earlier than the earliest backtrack time, this value backtracks the DB cluster to the earliest possible backtrack time. Otherwise, an error occurs.

", - "CopyDBClusterSnapshotMessage$CopyTags": "

True to copy all tags from the source DB cluster snapshot to the target DB cluster snapshot, and otherwise false. The default is false.

", - "CopyDBSnapshotMessage$CopyTags": "

True to copy all tags from the source DB snapshot to the target DB snapshot, and otherwise false. The default is false.

", - "CreateDBClusterMessage$StorageEncrypted": "

Specifies whether the DB cluster is encrypted.

", - "CreateDBClusterMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", - "CreateDBInstanceMessage$MultiAZ": "

Specifies if the DB instance is a Multi-AZ deployment. You can't set the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", - "CreateDBInstanceMessage$AutoMinorVersionUpgrade": "

Indicates that minor engine upgrades are applied automatically to the DB instance during the maintenance window.

Default: true

", - "CreateDBInstanceMessage$PubliclyAccessible": "

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC: true

  • VPC: false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", - "CreateDBInstanceMessage$StorageEncrypted": "

Specifies whether the DB instance is encrypted.

Amazon Aurora

Not applicable. The encryption for DB instances is managed by the DB cluster. For more information, see CreateDBCluster.

Default: false

", - "CreateDBInstanceMessage$CopyTagsToSnapshot": "

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", - "CreateDBInstanceMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines:

Amazon Aurora

Not applicable. Mapping AWS IAM accounts to database accounts is managed by the DB cluster. For more information, see CreateDBCluster.

MySQL

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", - "CreateDBInstanceMessage$EnablePerformanceInsights": "

True to enable Performance Insights for the DB instance, and otherwise false.

", - "CreateDBInstanceReadReplicaMessage$MultiAZ": "

Specifies whether the Read Replica is in a Multi-AZ deployment.

You can create a Read Replica as a Multi-AZ DB instance. RDS creates a standby of your replica in another Availability Zone for failover support for the replica. Creating your Read Replica as a Multi-AZ DB instance is independent of whether the source database is a Multi-AZ DB instance.

", - "CreateDBInstanceReadReplicaMessage$AutoMinorVersionUpgrade": "

Indicates that minor engine upgrades are applied automatically to the Read Replica during the maintenance window.

Default: Inherits from the source DB instance

", - "CreateDBInstanceReadReplicaMessage$PubliclyAccessible": "

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC:true

  • VPC:false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", - "CreateDBInstanceReadReplicaMessage$CopyTagsToSnapshot": "

True to copy all tags from the Read Replica to snapshots of the Read Replica, and otherwise false. The default is false.

", - "CreateDBInstanceReadReplicaMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

  • Aurora 5.6 or higher.

Default: false

", - "CreateDBInstanceReadReplicaMessage$EnablePerformanceInsights": "

True to enable Performance Insights for the read replica, and otherwise false.

", - "CreateEventSubscriptionMessage$Enabled": "

A Boolean value; set to true to activate the subscription, set to false to create the subscription but not active it.

", - "DBInstance$PerformanceInsightsEnabled": "

True if Performance Insights is enabled for the DB instance, and otherwise false.

", - "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets": "

If this parameter is specified and the requested engine supports the CharacterSetName parameter for CreateDBInstance, the response includes a list of supported character sets for each engine version.

", - "DescribeDBEngineVersionsMessage$ListSupportedTimezones": "

If this parameter is specified and the requested engine supports the TimeZone parameter for CreateDBInstance, the response includes a list of supported time zones for each engine version.

", - "DescribeOrderableDBInstanceOptionsMessage$Vpc": "

The VPC filter value. Specify this parameter to show only the available VPC or non-VPC offerings.

", - "DescribeReservedDBInstancesMessage$MultiAZ": "

The Multi-AZ filter value. Specify this parameter to show only those reservations matching the specified Multi-AZ parameter.

", - "DescribeReservedDBInstancesOfferingsMessage$MultiAZ": "

The Multi-AZ filter value. Specify this parameter to show only the available offerings matching the specified Multi-AZ parameter.

", - "ModifyDBClusterMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", - "ModifyDBInstanceMessage$MultiAZ": "

Specifies if the DB instance is a Multi-AZ deployment. Changing this parameter doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

", - "ModifyDBInstanceMessage$AutoMinorVersionUpgrade": "

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window. Changing this parameter doesn't result in an outage except in the following case and the change is asynchronously applied as soon as possible. An outage will result if this parameter is set to true during the maintenance window, and a newer minor version is available, and RDS has enabled auto patching for that engine version.

", - "ModifyDBInstanceMessage$CopyTagsToSnapshot": "

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", - "ModifyDBInstanceMessage$PubliclyAccessible": "

Boolean value that indicates if the DB instance has a publicly resolvable DNS name. Set to True to make the DB instance Internet-facing with a publicly resolvable DNS name, which resolves to a public IP address. Set to False to make the DB instance internal with a DNS name that resolves to a private IP address.

PubliclyAccessible only applies to DB instances in a VPC. The DB instance must be part of a public subnet and PubliclyAccessible must be true in order for it to be publicly accessible.

Changes to the PubliclyAccessible parameter are applied immediately regardless of the value of the ApplyImmediately parameter.

Default: false

", - "ModifyDBInstanceMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

Amazon Aurora

Not applicable. Mapping AWS IAM accounts to database accounts is managed by the DB cluster. For more information, see ModifyDBCluster.

MySQL

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", - "ModifyDBInstanceMessage$EnablePerformanceInsights": "

True to enable Performance Insights for the DB instance, and otherwise false.

", - "ModifyEventSubscriptionMessage$Enabled": "

A Boolean value; set to true to activate the subscription.

", - "OptionGroupOption$SupportsOptionVersionDowngrade": "

If true, you can change the option to an earlier version of the option. This only applies to options that have different versions available.

", - "PendingModifiedValues$MultiAZ": "

Indicates that the Single-AZ DB instance is to change to a Multi-AZ deployment.

", - "RebootDBInstanceMessage$ForceFailover": "

When true, the reboot is conducted through a MultiAZ failover.

Constraint: You can't specify true if the instance is not configured for MultiAZ.

", - "RestoreDBClusterFromS3Message$StorageEncrypted": "

Specifies whether the restored DB cluster is encrypted.

", - "RestoreDBClusterFromS3Message$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", - "RestoreDBClusterFromSnapshotMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", - "RestoreDBClusterToPointInTimeMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", - "RestoreDBInstanceFromDBSnapshotMessage$MultiAZ": "

Specifies if the DB instance is a Multi-AZ deployment.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", - "RestoreDBInstanceFromDBSnapshotMessage$PubliclyAccessible": "

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC: true

  • VPC: false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", - "RestoreDBInstanceFromDBSnapshotMessage$AutoMinorVersionUpgrade": "

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window.

", - "RestoreDBInstanceFromDBSnapshotMessage$CopyTagsToSnapshot": "

True to copy all tags from the restored DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", - "RestoreDBInstanceFromDBSnapshotMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", - "RestoreDBInstanceFromS3Message$MultiAZ": "

Specifies whether the DB instance is a Multi-AZ deployment. If MultiAZ is set to true, you can't set the AvailabilityZone parameter.

", - "RestoreDBInstanceFromS3Message$AutoMinorVersionUpgrade": "

True to indicate that minor engine upgrades are applied automatically to the DB instance during the maintenance window, and otherwise false.

Default: true

", - "RestoreDBInstanceFromS3Message$PubliclyAccessible": "

Specifies whether the DB instance is publicly accessible or not. For more information, see CreateDBInstance.

", - "RestoreDBInstanceFromS3Message$StorageEncrypted": "

Specifies whether the new DB instance is encrypted or not.

", - "RestoreDBInstanceFromS3Message$CopyTagsToSnapshot": "

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false.

Default: false.

", - "RestoreDBInstanceFromS3Message$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", - "RestoreDBInstanceFromS3Message$EnablePerformanceInsights": "

True to enable Performance Insights for the DB instance, and otherwise false.

", - "RestoreDBInstanceToPointInTimeMessage$MultiAZ": "

Specifies if the DB instance is a Multi-AZ deployment.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", - "RestoreDBInstanceToPointInTimeMessage$PubliclyAccessible": "

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC:true

  • VPC:false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", - "RestoreDBInstanceToPointInTimeMessage$AutoMinorVersionUpgrade": "

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window.

", - "RestoreDBInstanceToPointInTimeMessage$CopyTagsToSnapshot": "

True to copy all tags from the restored DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", - "RestoreDBInstanceToPointInTimeMessage$EnableIAMDatabaseAuthentication": "

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

" - } - }, - "Certificate": { - "base": "

A CA certificate for an AWS account.

", - "refs": { - "CertificateList$member": null - } - }, - "CertificateList": { - "base": null, - "refs": { - "CertificateMessage$Certificates": "

The list of Certificate objects for the AWS account.

" - } - }, - "CertificateMessage": { - "base": "

Data returned by the DescribeCertificates action.

", - "refs": { - } - }, - "CertificateNotFoundFault": { - "base": "

CertificateIdentifier doesn't refer to an existing certificate.

", - "refs": { + "AttributeValueList":{ + "base":null, + "refs":{ + "DBClusterSnapshotAttribute$AttributeValues":"

The value(s) for the manual DB cluster snapshot attribute.

If the AttributeName field is set to restore, then this element returns a list of IDs of the AWS accounts that are authorized to copy or restore the manual DB cluster snapshot. If a value of all is in the list, then the manual DB cluster snapshot is public and available for any AWS account to copy or restore.

", + "DBSnapshotAttribute$AttributeValues":"

The value or values for the manual DB snapshot attribute.

If the AttributeName field is set to restore, then this element returns a list of IDs of the AWS accounts that are authorized to copy or restore the manual DB snapshot. If a value of all is in the list, then the manual DB snapshot is public and available for any AWS account to copy or restore.

", + "ModifyDBClusterSnapshotAttributeMessage$ValuesToAdd":"

A list of DB cluster snapshot attributes to add to the attribute specified by AttributeName.

To authorize other AWS accounts to copy or restore a manual DB cluster snapshot, set this list to include one or more AWS account IDs, or all to make the manual DB cluster snapshot restorable by any AWS account. Do not add the all value for any manual DB cluster snapshots that contain private information that you don't want available to all AWS accounts.

", + "ModifyDBClusterSnapshotAttributeMessage$ValuesToRemove":"

A list of DB cluster snapshot attributes to remove from the attribute specified by AttributeName.

To remove authorization for other AWS accounts to copy or restore a manual DB cluster snapshot, set this list to include one or more AWS account identifiers, or all to remove authorization for any AWS account to copy or restore the DB cluster snapshot. If you specify all, an AWS account whose account ID is explicitly added to the restore attribute can still copy or restore a manual DB cluster snapshot.

", + "ModifyDBSnapshotAttributeMessage$ValuesToAdd":"

A list of DB snapshot attributes to add to the attribute specified by AttributeName.

To authorize other AWS accounts to copy or restore a manual snapshot, set this list to include one or more AWS account IDs, or all to make the manual DB snapshot restorable by any AWS account. Do not add the all value for any manual DB snapshots that contain private information that you don't want available to all AWS accounts.

", + "ModifyDBSnapshotAttributeMessage$ValuesToRemove":"

A list of DB snapshot attributes to remove from the attribute specified by AttributeName.

To remove authorization for other AWS accounts to copy or restore a manual snapshot, set this list to include one or more AWS account identifiers, or all to remove authorization for any AWS account to copy or restore the DB snapshot. If you specify all, an AWS account whose account ID is explicitly added to the restore attribute can still copy or restore the manual DB snapshot.

" } }, - "CharacterSet": { - "base": "

This data type is used as a response element in the action DescribeDBEngineVersions.

", - "refs": { - "DBEngineVersion$DefaultCharacterSet": "

The default character set for new instances of this engine version, if the CharacterSetName parameter of the CreateDBInstance API is not specified.

", - "SupportedCharacterSetsList$member": null + "AuthorizationAlreadyExistsFault":{ + "base":"

The specified CIDRIP or Amazon EC2 security group is already authorized for the specified DB security group.

", + "refs":{ } }, - "CloudwatchLogsExportConfiguration": { - "base": "

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB instance or DB cluster.

", - "refs": { - "ModifyDBInstanceMessage$CloudwatchLogsExportConfiguration": "

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB instance or DB cluster.

" + "AuthorizationNotFoundFault":{ + "base":"

The specified CIDRIP or Amazon EC2 security group isn't authorized for the specified DB security group.

RDS also may not be authorized by using IAM to perform necessary actions on your behalf.

", + "refs":{ } }, - "CopyDBClusterParameterGroupMessage": { - "base": null, - "refs": { + "AuthorizationQuotaExceededFault":{ + "base":"

The DB security group authorization quota has been reached.

", + "refs":{ + } + }, + "AuthorizeDBSecurityGroupIngressMessage":{ + "base":"

", + "refs":{ + } + }, + "AuthorizeDBSecurityGroupIngressResult":{ + "base":null, + "refs":{ + } + }, + "AvailabilityZone":{ + "base":"

Contains Availability Zone information.

This data type is used as an element in the following data type:

", + "refs":{ + "AvailabilityZoneList$member":null, + "Subnet$SubnetAvailabilityZone":null + } + }, + "AvailabilityZoneList":{ + "base":null, + "refs":{ + "OrderableDBInstanceOption$AvailabilityZones":"

A list of Availability Zones for a DB instance.

" + } + }, + "AvailabilityZones":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$AvailabilityZones":"

A list of EC2 Availability Zones that instances in the DB cluster can be created in. For information on AWS Regions and Availability Zones, see Regions and Availability Zones.

", + "DBCluster$AvailabilityZones":"

Provides the list of EC2 Availability Zones that instances in the DB cluster can be created in.

", + "DBClusterSnapshot$AvailabilityZones":"

Provides the list of EC2 Availability Zones that instances in the DB cluster snapshot can be restored in.

", + "RestoreDBClusterFromS3Message$AvailabilityZones":"

A list of EC2 Availability Zones that instances in the restored DB cluster can be created in.

", + "RestoreDBClusterFromSnapshotMessage$AvailabilityZones":"

Provides the list of EC2 Availability Zones that instances in the restored DB cluster can be created in.

" + } + }, + "AvailableProcessorFeature":{ + "base":"

Contains the available processor feature information for the DB instance class of a DB instance.

For more information, see Configuring the Processor of the DB Instance Class in the Amazon RDS User Guide.

", + "refs":{ + "AvailableProcessorFeatureList$member":null + } + }, + "AvailableProcessorFeatureList":{ + "base":null, + "refs":{ + "OrderableDBInstanceOption$AvailableProcessorFeatures":"

A list of the available processor features for the DB instance class of a DB instance.

", + "ValidDBInstanceModificationsMessage$ValidProcessorFeatures":"

Valid processor features for your DB instance.

" + } + }, + "BacktrackDBClusterMessage":{ + "base":"

", + "refs":{ + } + }, + "Boolean":{ + "base":null, + "refs":{ + "DBCluster$MultiAZ":"

Specifies whether the DB cluster has instances in multiple Availability Zones.

", + "DBCluster$StorageEncrypted":"

Specifies whether the DB cluster is encrypted.

", + "DBCluster$IAMDatabaseAuthenticationEnabled":"

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", + "DBClusterMember$IsClusterWriter":"

Value that is true if the cluster member is the primary instance for the DB cluster and false otherwise.

", + "DBClusterSnapshot$StorageEncrypted":"

Specifies whether the DB cluster snapshot is encrypted.

", + "DBClusterSnapshot$IAMDatabaseAuthenticationEnabled":"

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", + "DBEngineVersion$SupportsLogExportsToCloudwatchLogs":"

A value that indicates whether the engine version supports exporting the log types specified by ExportableLogTypes to CloudWatch Logs.

", + "DBEngineVersion$SupportsReadReplica":"

Indicates whether the database engine version supports read replicas.

", + "DBInstance$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment.

", + "DBInstance$AutoMinorVersionUpgrade":"

Indicates that minor version patches are applied automatically.

", + "DBInstance$PubliclyAccessible":"

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC:true

  • VPC:false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", + "DBInstance$StorageEncrypted":"

Specifies whether the DB instance is encrypted.

", + "DBInstance$CopyTagsToSnapshot":"

Specifies whether tags are copied from the DB instance to snapshots of the DB instance.

", + "DBInstance$IAMDatabaseAuthenticationEnabled":"

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

IAM database authentication can be enabled for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

  • Aurora 5.6 or higher. To enable IAM database authentication for Aurora, see DBCluster Type.

", + "DBInstanceStatusInfo$Normal":"

Boolean value that is true if the instance is operating normally, or false if the instance is in an error state.

", + "DBSnapshot$Encrypted":"

Specifies whether the DB snapshot is encrypted.

", + "DBSnapshot$IAMDatabaseAuthenticationEnabled":"

True if mapping of AWS Identity and Access Management (IAM) accounts to database accounts is enabled, and otherwise false.

", + "DeleteDBClusterMessage$SkipFinalSnapshot":"

Determines whether a final DB cluster snapshot is created before the DB cluster is deleted. If true is specified, no DB cluster snapshot is created. If false is specified, a DB cluster snapshot is created before the DB cluster is deleted.

You must specify a FinalDBSnapshotIdentifier parameter if SkipFinalSnapshot is false.

Default: false

", + "DeleteDBInstanceMessage$SkipFinalSnapshot":"

Determines whether a final DB snapshot is created before the DB instance is deleted. If true is specified, no DBSnapshot is created. If false is specified, a DB snapshot is created before the DB instance is deleted.

Note that when a DB instance is in a failure state and has a status of 'failed', 'incompatible-restore', or 'incompatible-network', it can only be deleted when the SkipFinalSnapshot parameter is set to \"true\".

Specify true when deleting a Read Replica.

The FinalDBSnapshotIdentifier parameter must be specified if SkipFinalSnapshot is false.

Default: false

", + "DescribeDBClusterSnapshotsMessage$IncludeShared":"

True to include shared manual DB cluster snapshots from other AWS accounts that this AWS account has been given permission to copy or restore, and otherwise false. The default is false.

You can give an AWS account permission to restore a manual DB cluster snapshot from another AWS account by the ModifyDBClusterSnapshotAttribute API action.

", + "DescribeDBClusterSnapshotsMessage$IncludePublic":"

True to include manual DB cluster snapshots that are public and can be copied or restored by any AWS account, and otherwise false. The default is false. The default is false.

You can share a manual DB cluster snapshot as public by using the ModifyDBClusterSnapshotAttribute API action.

", + "DescribeDBEngineVersionsMessage$DefaultOnly":"

Indicates that only the default version of the specified engine or engine and major version combination is returned.

", + "DescribeDBSnapshotsMessage$IncludeShared":"

True to include shared manual DB snapshots from other AWS accounts that this AWS account has been given permission to copy or restore, and otherwise false. The default is false.

You can give an AWS account permission to restore a manual DB snapshot from another AWS account by using the ModifyDBSnapshotAttribute API action.

", + "DescribeDBSnapshotsMessage$IncludePublic":"

True to include manual DB snapshots that are public and can be copied or restored by any AWS account, and otherwise false. The default is false.

You can share a manual DB snapshot as public by using the ModifyDBSnapshotAttribute API.

", + "DownloadDBLogFilePortionDetails$AdditionalDataPending":"

Boolean value that if true, indicates there is more data to be downloaded.

", + "EventSubscription$Enabled":"

A Boolean value indicating if the subscription is enabled. True indicates the subscription is enabled.

", + "ModifyDBClusterMessage$ApplyImmediately":"

A value that specifies whether the modifications in this request and any pending modifications are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow setting for the DB cluster. If this parameter is set to false, changes to the DB cluster are applied during the next maintenance window.

The ApplyImmediately parameter only affects the NewDBClusterIdentifier and MasterUserPassword values. If you set the ApplyImmediately parameter value to false, then changes to the NewDBClusterIdentifier and MasterUserPassword values are applied during the next maintenance window. All other changes are applied immediately, regardless of the value of the ApplyImmediately parameter.

Default: false

", + "ModifyDBInstanceMessage$ApplyImmediately":"

Specifies whether the modifications in this request and any pending modifications are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow setting for the DB instance.

If this parameter is set to false, changes to the DB instance are applied during the next maintenance window. Some parameter changes can cause an outage and are applied on the next call to RebootDBInstance, or the next failure reboot. Review the table of parameters in Modifying a DB Instance and Using the Apply Immediately Parameter to see the impact that setting ApplyImmediately to true or false has for each modified parameter and to determine when the changes are applied.

Default: false

", + "ModifyDBInstanceMessage$AllowMajorVersionUpgrade":"

Indicates that major version upgrades are allowed. Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible.

Constraints: This parameter must be set to true when specifying a value for the EngineVersion parameter that is a different major version than the DB instance's current version.

", + "ModifyOptionGroupMessage$ApplyImmediately":"

Indicates whether the changes should be applied immediately, or during the next maintenance window for each instance associated with the option group.

", + "Option$Persistent":"

Indicate if this option is persistent.

", + "Option$Permanent":"

Indicate if this option is permanent.

", + "OptionGroup$AllowsVpcAndNonVpcInstanceMemberships":"

Indicates whether this option group can be applied to both VPC and non-VPC instances. The value true indicates the option group can be applied to both VPC and non-VPC instances.

", + "OptionGroupOption$PortRequired":"

Specifies whether the option requires a port.

", + "OptionGroupOption$Persistent":"

Persistent options can't be removed from an option group while DB instances are associated with the option group. If you disassociate all DB instances from the option group, your can remove the persistent option from the option group.

", + "OptionGroupOption$Permanent":"

Permanent options can never be removed from an option group. An option group containing a permanent option can't be removed from a DB instance.

", + "OptionGroupOption$RequiresAutoMinorEngineVersionUpgrade":"

If true, you must enable the Auto Minor Version Upgrade setting for your DB instance before you can use this option. You can enable Auto Minor Version Upgrade when you first create your DB instance, or by modifying your DB instance later.

", + "OptionGroupOption$VpcOnly":"

If true, you can only use this option with a DB instance that is in a VPC.

", + "OptionGroupOptionSetting$IsModifiable":"

Boolean value where true indicates that this option group option can be changed from the default value.

", + "OptionSetting$IsModifiable":"

A Boolean value that, when true, indicates the option setting can be modified from the default.

", + "OptionSetting$IsCollection":"

Indicates if the option setting is part of a collection.

", + "OptionVersion$IsDefault":"

True if the version is the default version of the option, and otherwise false.

", + "OrderableDBInstanceOption$MultiAZCapable":"

Indicates whether a DB instance is Multi-AZ capable.

", + "OrderableDBInstanceOption$ReadReplicaCapable":"

Indicates whether a DB instance can have a Read Replica.

", + "OrderableDBInstanceOption$Vpc":"

Indicates whether a DB instance is in a VPC.

", + "OrderableDBInstanceOption$SupportsStorageEncryption":"

Indicates whether a DB instance supports encrypted storage.

", + "OrderableDBInstanceOption$SupportsIops":"

Indicates whether a DB instance supports provisioned IOPS.

", + "OrderableDBInstanceOption$SupportsEnhancedMonitoring":"

Indicates whether a DB instance supports Enhanced Monitoring at intervals from 1 to 60 seconds.

", + "OrderableDBInstanceOption$SupportsIAMDatabaseAuthentication":"

Indicates whether a DB instance supports IAM database authentication.

", + "OrderableDBInstanceOption$SupportsPerformanceInsights":"

True if a DB instance supports Performance Insights, otherwise false.

", + "Parameter$IsModifiable":"

Indicates whether (true) or not (false) the parameter can be modified. Some parameters have security or operational implications that prevent them from being changed.

", + "ReservedDBInstance$MultiAZ":"

Indicates if the reservation applies to Multi-AZ deployments.

", + "ReservedDBInstancesOffering$MultiAZ":"

Indicates if the offering applies to Multi-AZ deployments.

", + "ResetDBClusterParameterGroupMessage$ResetAllParameters":"

A value that is set to true to reset all parameters in the DB cluster parameter group to their default values, and false otherwise. You can't use this parameter if there is a list of parameter names specified for the Parameters parameter.

", + "ResetDBParameterGroupMessage$ResetAllParameters":"

Specifies whether (true) or not (false) to reset all parameters in the DB parameter group to default values.

Default: true

", + "RestoreDBClusterToPointInTimeMessage$UseLatestRestorableTime":"

A value that is set to true to restore the DB cluster to the latest restorable backup time, and false otherwise.

Default: false

Constraints: Cannot be specified if RestoreToTime parameter is provided.

", + "RestoreDBInstanceToPointInTimeMessage$UseLatestRestorableTime":"

Specifies whether (true) or not (false) the DB instance is restored from the latest backup time.

Default: false

Constraints: Cannot be specified if RestoreTime parameter is provided.

", + "UpgradeTarget$AutoUpgrade":"

A value that indicates whether the target version is applied to any source DB instances that have AutoMinorVersionUpgrade set to true.

", + "UpgradeTarget$IsMajorVersionUpgrade":"

A value that indicates whether a database engine is upgraded to a major version.

" + } + }, + "BooleanOptional":{ + "base":null, + "refs":{ + "BacktrackDBClusterMessage$Force":"

A value that, if specified, forces the DB cluster to backtrack when binary logging is enabled. Otherwise, an error occurs when binary logging is enabled.

", + "BacktrackDBClusterMessage$UseEarliestTimeOnPointInTimeUnavailable":"

If BacktrackTo is set to a timestamp earlier than the earliest backtrack time, this value backtracks the DB cluster to the earliest possible backtrack time. Otherwise, an error occurs.

", + "CopyDBClusterSnapshotMessage$CopyTags":"

True to copy all tags from the source DB cluster snapshot to the target DB cluster snapshot, and otherwise false. The default is false.

", + "CopyDBSnapshotMessage$CopyTags":"

True to copy all tags from the source DB snapshot to the target DB snapshot, and otherwise false. The default is false.

", + "CreateDBClusterMessage$StorageEncrypted":"

Specifies whether the DB cluster is encrypted.

", + "CreateDBClusterMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "CreateDBInstanceMessage$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment. You can't set the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", + "CreateDBInstanceMessage$AutoMinorVersionUpgrade":"

Indicates that minor engine upgrades are applied automatically to the DB instance during the maintenance window.

Default: true

", + "CreateDBInstanceMessage$PubliclyAccessible":"

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC: true

  • VPC: false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", + "CreateDBInstanceMessage$StorageEncrypted":"

Specifies whether the DB instance is encrypted.

Amazon Aurora

Not applicable. The encryption for DB instances is managed by the DB cluster. For more information, see CreateDBCluster.

Default: false

", + "CreateDBInstanceMessage$CopyTagsToSnapshot":"

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", + "CreateDBInstanceMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines:

Amazon Aurora

Not applicable. Mapping AWS IAM accounts to database accounts is managed by the DB cluster. For more information, see CreateDBCluster.

MySQL

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", + "CreateDBInstanceMessage$EnablePerformanceInsights":"

True to enable Performance Insights for the DB instance, and otherwise false.

For more information, see Using Amazon Performance Insights in the Amazon Relational Database Service User Guide.

", + "CreateDBInstanceReadReplicaMessage$MultiAZ":"

Specifies whether the Read Replica is in a Multi-AZ deployment.

You can create a Read Replica as a Multi-AZ DB instance. RDS creates a standby of your replica in another Availability Zone for failover support for the replica. Creating your Read Replica as a Multi-AZ DB instance is independent of whether the source database is a Multi-AZ DB instance.

", + "CreateDBInstanceReadReplicaMessage$AutoMinorVersionUpgrade":"

Indicates that minor engine upgrades are applied automatically to the Read Replica during the maintenance window.

Default: Inherits from the source DB instance

", + "CreateDBInstanceReadReplicaMessage$PubliclyAccessible":"

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC:true

  • VPC:false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", + "CreateDBInstanceReadReplicaMessage$CopyTagsToSnapshot":"

True to copy all tags from the Read Replica to snapshots of the Read Replica, and otherwise false. The default is false.

", + "CreateDBInstanceReadReplicaMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

  • Aurora 5.6 or higher.

Default: false

", + "CreateDBInstanceReadReplicaMessage$EnablePerformanceInsights":"

True to enable Performance Insights for the read replica, and otherwise false.

For more information, see Using Amazon Performance Insights in the Amazon Relational Database Service User Guide.

", + "CreateDBInstanceReadReplicaMessage$UseDefaultProcessorFeatures":"

A value that specifies that the DB instance class of the DB instance uses its default processor features.

", + "CreateEventSubscriptionMessage$Enabled":"

A Boolean value; set to true to activate the subscription, set to false to create the subscription but not active it.

", + "DBInstance$PerformanceInsightsEnabled":"

True if Performance Insights is enabled for the DB instance, and otherwise false.

", + "DescribeDBEngineVersionsMessage$ListSupportedCharacterSets":"

If this parameter is specified and the requested engine supports the CharacterSetName parameter for CreateDBInstance, the response includes a list of supported character sets for each engine version.

", + "DescribeDBEngineVersionsMessage$ListSupportedTimezones":"

If this parameter is specified and the requested engine supports the TimeZone parameter for CreateDBInstance, the response includes a list of supported time zones for each engine version.

", + "DescribeOrderableDBInstanceOptionsMessage$Vpc":"

The VPC filter value. Specify this parameter to show only the available VPC or non-VPC offerings.

", + "DescribeReservedDBInstancesMessage$MultiAZ":"

The Multi-AZ filter value. Specify this parameter to show only those reservations matching the specified Multi-AZ parameter.

", + "DescribeReservedDBInstancesOfferingsMessage$MultiAZ":"

The Multi-AZ filter value. Specify this parameter to show only the available offerings matching the specified Multi-AZ parameter.

", + "ModifyDBClusterMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "ModifyDBInstanceMessage$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment. Changing this parameter doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

", + "ModifyDBInstanceMessage$AutoMinorVersionUpgrade":"

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window. Changing this parameter doesn't result in an outage except in the following case and the change is asynchronously applied as soon as possible. An outage will result if this parameter is set to true during the maintenance window, and a newer minor version is available, and RDS has enabled auto patching for that engine version.

", + "ModifyDBInstanceMessage$CopyTagsToSnapshot":"

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", + "ModifyDBInstanceMessage$PubliclyAccessible":"

Boolean value that indicates if the DB instance has a publicly resolvable DNS name. Set to True to make the DB instance Internet-facing with a publicly resolvable DNS name, which resolves to a public IP address. Set to False to make the DB instance internal with a DNS name that resolves to a private IP address.

PubliclyAccessible only applies to DB instances in a VPC. The DB instance must be part of a public subnet and PubliclyAccessible must be true in order for it to be publicly accessible.

Changes to the PubliclyAccessible parameter are applied immediately regardless of the value of the ApplyImmediately parameter.

Default: false

", + "ModifyDBInstanceMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

Amazon Aurora

Not applicable. Mapping AWS IAM accounts to database accounts is managed by the DB cluster. For more information, see ModifyDBCluster.

MySQL

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", + "ModifyDBInstanceMessage$EnablePerformanceInsights":"

True to enable Performance Insights for the DB instance, and otherwise false.

For more information, see Using Amazon Performance Insights in the Amazon Relational Database Service User Guide.

", + "ModifyDBInstanceMessage$UseDefaultProcessorFeatures":"

A value that specifies that the DB instance class of the DB instance uses its default processor features.

", + "ModifyEventSubscriptionMessage$Enabled":"

A Boolean value; set to true to activate the subscription.

", + "OptionGroupOption$SupportsOptionVersionDowngrade":"

If true, you can change the option to an earlier version of the option. This only applies to options that have different versions available.

", + "PendingModifiedValues$MultiAZ":"

Indicates that the Single-AZ DB instance is to change to a Multi-AZ deployment.

", + "RebootDBInstanceMessage$ForceFailover":"

When true, the reboot is conducted through a MultiAZ failover.

Constraint: You can't specify true if the instance is not configured for MultiAZ.

", + "RestoreDBClusterFromS3Message$StorageEncrypted":"

Specifies whether the restored DB cluster is encrypted.

", + "RestoreDBClusterFromS3Message$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "RestoreDBClusterFromSnapshotMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "RestoreDBClusterToPointInTimeMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "RestoreDBInstanceFromDBSnapshotMessage$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", + "RestoreDBInstanceFromDBSnapshotMessage$PubliclyAccessible":"

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC: true

  • VPC: false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", + "RestoreDBInstanceFromDBSnapshotMessage$AutoMinorVersionUpgrade":"

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window.

", + "RestoreDBInstanceFromDBSnapshotMessage$CopyTagsToSnapshot":"

True to copy all tags from the restored DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", + "RestoreDBInstanceFromDBSnapshotMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", + "RestoreDBInstanceFromDBSnapshotMessage$UseDefaultProcessorFeatures":"

A value that specifies that the DB instance class of the DB instance uses its default processor features.

", + "RestoreDBInstanceFromS3Message$MultiAZ":"

Specifies whether the DB instance is a Multi-AZ deployment. If MultiAZ is set to true, you can't set the AvailabilityZone parameter.

", + "RestoreDBInstanceFromS3Message$AutoMinorVersionUpgrade":"

True to indicate that minor engine upgrades are applied automatically to the DB instance during the maintenance window, and otherwise false.

Default: true

", + "RestoreDBInstanceFromS3Message$PubliclyAccessible":"

Specifies whether the DB instance is publicly accessible or not. For more information, see CreateDBInstance.

", + "RestoreDBInstanceFromS3Message$StorageEncrypted":"

Specifies whether the new DB instance is encrypted or not.

", + "RestoreDBInstanceFromS3Message$CopyTagsToSnapshot":"

True to copy all tags from the DB instance to snapshots of the DB instance, and otherwise false.

Default: false.

", + "RestoreDBInstanceFromS3Message$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

Default: false

", + "RestoreDBInstanceFromS3Message$EnablePerformanceInsights":"

True to enable Performance Insights for the DB instance, and otherwise false.

For more information, see Using Amazon Performance Insights in the Amazon Relational Database Service User Guide.

", + "RestoreDBInstanceFromS3Message$UseDefaultProcessorFeatures":"

A value that specifies that the DB instance class of the DB instance uses its default processor features.

", + "RestoreDBInstanceToPointInTimeMessage$MultiAZ":"

Specifies if the DB instance is a Multi-AZ deployment.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

", + "RestoreDBInstanceToPointInTimeMessage$PubliclyAccessible":"

Specifies the accessibility options for the DB instance. A value of true specifies an Internet-facing instance with a publicly resolvable DNS name, which resolves to a public IP address. A value of false specifies an internal instance with a DNS name that resolves to a private IP address.

Default: The default behavior varies depending on whether a VPC has been requested or not. The following list shows the default behavior in each case.

  • Default VPC:true

  • VPC:false

If no DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is publicly accessible. If a specific DB subnet group has been specified as part of the request and the PubliclyAccessible value has not been set, the DB instance is private.

", + "RestoreDBInstanceToPointInTimeMessage$AutoMinorVersionUpgrade":"

Indicates that minor version upgrades are applied automatically to the DB instance during the maintenance window.

", + "RestoreDBInstanceToPointInTimeMessage$CopyTagsToSnapshot":"

True to copy all tags from the restored DB instance to snapshots of the DB instance, and otherwise false. The default is false.

", + "RestoreDBInstanceToPointInTimeMessage$EnableIAMDatabaseAuthentication":"

True to enable mapping of AWS Identity and Access Management (IAM) accounts to database accounts, and otherwise false.

You can enable IAM database authentication for the following database engines

  • For MySQL 5.6, minor version 5.6.34 or higher

  • For MySQL 5.7, minor version 5.7.16 or higher

Default: false

", + "RestoreDBInstanceToPointInTimeMessage$UseDefaultProcessorFeatures":"

A value that specifies that the DB instance class of the DB instance uses its default processor features.

" + } + }, + "Certificate":{ + "base":"

A CA certificate for an AWS account.

", + "refs":{ + "CertificateList$member":null + } + }, + "CertificateList":{ + "base":null, + "refs":{ + "CertificateMessage$Certificates":"

The list of Certificate objects for the AWS account.

" + } + }, + "CertificateMessage":{ + "base":"

Data returned by the DescribeCertificates action.

", + "refs":{ + } + }, + "CertificateNotFoundFault":{ + "base":"

CertificateIdentifier doesn't refer to an existing certificate.

", + "refs":{ + } + }, + "CharacterSet":{ + "base":"

This data type is used as a response element in the action DescribeDBEngineVersions.

", + "refs":{ + "DBEngineVersion$DefaultCharacterSet":"

The default character set for new instances of this engine version, if the CharacterSetName parameter of the CreateDBInstance API is not specified.

", + "SupportedCharacterSetsList$member":null + } + }, + "CloudwatchLogsExportConfiguration":{ + "base":"

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB instance or DB cluster.

", + "refs":{ + "ModifyDBClusterMessage$CloudwatchLogsExportConfiguration":"

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB cluster.

", + "ModifyDBInstanceMessage$CloudwatchLogsExportConfiguration":"

The configuration setting for the log types to be enabled for export to CloudWatch Logs for a specific DB instance.

" } }, - "CopyDBClusterParameterGroupResult": { - "base": null, - "refs": { + "CopyDBClusterParameterGroupMessage":{ + "base":null, + "refs":{ } }, - "CopyDBClusterSnapshotMessage": { - "base": "

", - "refs": { + "CopyDBClusterParameterGroupResult":{ + "base":null, + "refs":{ } }, - "CopyDBClusterSnapshotResult": { - "base": null, - "refs": { + "CopyDBClusterSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "CopyDBParameterGroupMessage": { - "base": "

", - "refs": { + "CopyDBClusterSnapshotResult":{ + "base":null, + "refs":{ } }, - "CopyDBParameterGroupResult": { - "base": null, - "refs": { + "CopyDBParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "CopyDBSnapshotMessage": { - "base": "

", - "refs": { + "CopyDBParameterGroupResult":{ + "base":null, + "refs":{ } }, - "CopyDBSnapshotResult": { - "base": null, - "refs": { + "CopyDBSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "CopyOptionGroupMessage": { - "base": "

", - "refs": { + "CopyDBSnapshotResult":{ + "base":null, + "refs":{ } }, - "CopyOptionGroupResult": { - "base": null, - "refs": { + "CopyOptionGroupMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBClusterMessage": { - "base": "

", - "refs": { + "CopyOptionGroupResult":{ + "base":null, + "refs":{ } }, - "CreateDBClusterParameterGroupMessage": { - "base": "

", - "refs": { + "CreateDBClusterMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBClusterParameterGroupResult": { - "base": null, - "refs": { + "CreateDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBClusterResult": { - "base": null, - "refs": { + "CreateDBClusterParameterGroupResult":{ + "base":null, + "refs":{ } }, - "CreateDBClusterSnapshotMessage": { - "base": "

", - "refs": { + "CreateDBClusterResult":{ + "base":null, + "refs":{ } }, - "CreateDBClusterSnapshotResult": { - "base": null, - "refs": { + "CreateDBClusterSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBInstanceMessage": { - "base": "

", - "refs": { + "CreateDBClusterSnapshotResult":{ + "base":null, + "refs":{ } }, - "CreateDBInstanceReadReplicaMessage": { - "base": null, - "refs": { + "CreateDBInstanceMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBInstanceReadReplicaResult": { - "base": null, - "refs": { + "CreateDBInstanceReadReplicaMessage":{ + "base":null, + "refs":{ } }, - "CreateDBInstanceResult": { - "base": null, - "refs": { + "CreateDBInstanceReadReplicaResult":{ + "base":null, + "refs":{ } }, - "CreateDBParameterGroupMessage": { - "base": "

", - "refs": { + "CreateDBInstanceResult":{ + "base":null, + "refs":{ } }, - "CreateDBParameterGroupResult": { - "base": null, - "refs": { + "CreateDBParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBSecurityGroupMessage": { - "base": "

", - "refs": { + "CreateDBParameterGroupResult":{ + "base":null, + "refs":{ } }, - "CreateDBSecurityGroupResult": { - "base": null, - "refs": { + "CreateDBSecurityGroupMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBSnapshotMessage": { - "base": "

", - "refs": { + "CreateDBSecurityGroupResult":{ + "base":null, + "refs":{ } }, - "CreateDBSnapshotResult": { - "base": null, - "refs": { + "CreateDBSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "CreateDBSubnetGroupMessage": { - "base": "

", - "refs": { + "CreateDBSnapshotResult":{ + "base":null, + "refs":{ } }, - "CreateDBSubnetGroupResult": { - "base": null, - "refs": { + "CreateDBSubnetGroupMessage":{ + "base":"

", + "refs":{ } }, - "CreateEventSubscriptionMessage": { - "base": "

", - "refs": { + "CreateDBSubnetGroupResult":{ + "base":null, + "refs":{ } }, - "CreateEventSubscriptionResult": { - "base": null, - "refs": { + "CreateEventSubscriptionMessage":{ + "base":"

", + "refs":{ } }, - "CreateOptionGroupMessage": { - "base": "

", - "refs": { + "CreateEventSubscriptionResult":{ + "base":null, + "refs":{ } }, - "CreateOptionGroupResult": { - "base": null, - "refs": { + "CreateOptionGroupMessage":{ + "base":"

", + "refs":{ } }, - "DBCluster": { - "base": "

Contains the details of an Amazon RDS DB cluster.

This data type is used as a response element in the DescribeDBClusters action.

", - "refs": { - "CreateDBClusterResult$DBCluster": null, - "DBClusterList$member": null, - "DeleteDBClusterResult$DBCluster": null, - "FailoverDBClusterResult$DBCluster": null, - "ModifyDBClusterResult$DBCluster": null, - "PromoteReadReplicaDBClusterResult$DBCluster": null, - "RestoreDBClusterFromS3Result$DBCluster": null, - "RestoreDBClusterFromSnapshotResult$DBCluster": null, - "RestoreDBClusterToPointInTimeResult$DBCluster": null + "CreateOptionGroupResult":{ + "base":null, + "refs":{ } }, - "DBClusterAlreadyExistsFault": { - "base": "

The user already has a DB cluster with the given identifier.

", - "refs": { + "DBCluster":{ + "base":"

Contains the details of an Amazon RDS DB cluster.

This data type is used as a response element in the DescribeDBClusters action.

", + "refs":{ + "CreateDBClusterResult$DBCluster":null, + "DBClusterList$member":null, + "DeleteDBClusterResult$DBCluster":null, + "FailoverDBClusterResult$DBCluster":null, + "ModifyDBClusterResult$DBCluster":null, + "PromoteReadReplicaDBClusterResult$DBCluster":null, + "RestoreDBClusterFromS3Result$DBCluster":null, + "RestoreDBClusterFromSnapshotResult$DBCluster":null, + "RestoreDBClusterToPointInTimeResult$DBCluster":null } }, - "DBClusterBacktrack": { - "base": "

This data type is used as a response element in the DescribeDBClusterBacktracks action.

", - "refs": { - "DBClusterBacktrackList$member": null + "DBClusterAlreadyExistsFault":{ + "base":"

The user already has a DB cluster with the given identifier.

", + "refs":{ } }, - "DBClusterBacktrackList": { - "base": null, - "refs": { - "DBClusterBacktrackMessage$DBClusterBacktracks": "

Contains a list of backtracks for the user.

" + "DBClusterBacktrack":{ + "base":"

This data type is used as a response element in the DescribeDBClusterBacktracks action.

", + "refs":{ + "DBClusterBacktrackList$member":null } }, - "DBClusterBacktrackMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBClusterBacktracks action.

", - "refs": { + "DBClusterBacktrackList":{ + "base":null, + "refs":{ + "DBClusterBacktrackMessage$DBClusterBacktracks":"

Contains a list of backtracks for the user.

" } }, - "DBClusterBacktrackNotFoundFault": { - "base": "

BacktrackIdentifier doesn't refer to an existing backtrack.

", - "refs": { + "DBClusterBacktrackMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBClusterBacktracks action.

", + "refs":{ } }, - "DBClusterList": { - "base": null, - "refs": { - "DBClusterMessage$DBClusters": "

Contains a list of DB clusters for the user.

" + "DBClusterBacktrackNotFoundFault":{ + "base":"

BacktrackIdentifier doesn't refer to an existing backtrack.

", + "refs":{ } }, - "DBClusterMember": { - "base": "

Contains information about an instance that is part of a DB cluster.

", - "refs": { - "DBClusterMemberList$member": null + "DBClusterList":{ + "base":null, + "refs":{ + "DBClusterMessage$DBClusters":"

Contains a list of DB clusters for the user.

" } }, - "DBClusterMemberList": { - "base": null, - "refs": { - "DBCluster$DBClusterMembers": "

Provides the list of instances that make up the DB cluster.

" + "DBClusterMember":{ + "base":"

Contains information about an instance that is part of a DB cluster.

", + "refs":{ + "DBClusterMemberList$member":null } }, - "DBClusterMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBClusters action.

", - "refs": { + "DBClusterMemberList":{ + "base":null, + "refs":{ + "DBCluster$DBClusterMembers":"

Provides the list of instances that make up the DB cluster.

" } }, - "DBClusterNotFoundFault": { - "base": "

DBClusterIdentifier doesn't refer to an existing DB cluster.

", - "refs": { + "DBClusterMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBClusters action.

", + "refs":{ } }, - "DBClusterOptionGroupMemberships": { - "base": null, - "refs": { - "DBCluster$DBClusterOptionGroupMemberships": "

Provides the list of option group memberships for this DB cluster.

" + "DBClusterNotFoundFault":{ + "base":"

DBClusterIdentifier doesn't refer to an existing DB cluster.

", + "refs":{ } }, - "DBClusterOptionGroupStatus": { - "base": "

Contains status information for a DB cluster option group.

", - "refs": { - "DBClusterOptionGroupMemberships$member": null + "DBClusterOptionGroupMemberships":{ + "base":null, + "refs":{ + "DBCluster$DBClusterOptionGroupMemberships":"

Provides the list of option group memberships for this DB cluster.

" } }, - "DBClusterParameterGroup": { - "base": "

Contains the details of an Amazon RDS DB cluster parameter group.

This data type is used as a response element in the DescribeDBClusterParameterGroups action.

", - "refs": { - "CopyDBClusterParameterGroupResult$DBClusterParameterGroup": null, - "CreateDBClusterParameterGroupResult$DBClusterParameterGroup": null, - "DBClusterParameterGroupList$member": null + "DBClusterOptionGroupStatus":{ + "base":"

Contains status information for a DB cluster option group.

", + "refs":{ + "DBClusterOptionGroupMemberships$member":null } }, - "DBClusterParameterGroupDetails": { - "base": "

Provides details about a DB cluster parameter group including the parameters in the DB cluster parameter group.

", - "refs": { + "DBClusterParameterGroup":{ + "base":"

Contains the details of an Amazon RDS DB cluster parameter group.

This data type is used as a response element in the DescribeDBClusterParameterGroups action.

", + "refs":{ + "CopyDBClusterParameterGroupResult$DBClusterParameterGroup":null, + "CreateDBClusterParameterGroupResult$DBClusterParameterGroup":null, + "DBClusterParameterGroupList$member":null } }, - "DBClusterParameterGroupList": { - "base": null, - "refs": { - "DBClusterParameterGroupsMessage$DBClusterParameterGroups": "

A list of DB cluster parameter groups.

" + "DBClusterParameterGroupDetails":{ + "base":"

Provides details about a DB cluster parameter group including the parameters in the DB cluster parameter group.

", + "refs":{ } }, - "DBClusterParameterGroupNameMessage": { - "base": "

", - "refs": { + "DBClusterParameterGroupList":{ + "base":null, + "refs":{ + "DBClusterParameterGroupsMessage$DBClusterParameterGroups":"

A list of DB cluster parameter groups.

" } }, - "DBClusterParameterGroupNotFoundFault": { - "base": "

DBClusterParameterGroupName doesn't refer to an existing DB cluster parameter group.

", - "refs": { + "DBClusterParameterGroupNameMessage":{ + "base":"

", + "refs":{ } }, - "DBClusterParameterGroupsMessage": { - "base": "

", - "refs": { + "DBClusterParameterGroupNotFoundFault":{ + "base":"

DBClusterParameterGroupName doesn't refer to an existing DB cluster parameter group.

", + "refs":{ } }, - "DBClusterQuotaExceededFault": { - "base": "

The user attempted to create a new DB cluster and the user has already reached the maximum allowed DB cluster quota.

", - "refs": { + "DBClusterParameterGroupsMessage":{ + "base":"

", + "refs":{ } }, - "DBClusterRole": { - "base": "

Describes an AWS Identity and Access Management (IAM) role that is associated with a DB cluster.

", - "refs": { - "DBClusterRoles$member": null + "DBClusterQuotaExceededFault":{ + "base":"

The user attempted to create a new DB cluster and the user has already reached the maximum allowed DB cluster quota.

", + "refs":{ } }, - "DBClusterRoleAlreadyExistsFault": { - "base": "

The specified IAM role Amazon Resource Name (ARN) is already associated with the specified DB cluster.

", - "refs": { + "DBClusterRole":{ + "base":"

Describes an AWS Identity and Access Management (IAM) role that is associated with a DB cluster.

", + "refs":{ + "DBClusterRoles$member":null } }, - "DBClusterRoleNotFoundFault": { - "base": "

The specified IAM role Amazon Resource Name (ARN) isn't associated with the specified DB cluster.

", - "refs": { + "DBClusterRoleAlreadyExistsFault":{ + "base":"

The specified IAM role Amazon Resource Name (ARN) is already associated with the specified DB cluster.

", + "refs":{ } }, - "DBClusterRoleQuotaExceededFault": { - "base": "

You have exceeded the maximum number of IAM roles that can be associated with the specified DB cluster.

", - "refs": { + "DBClusterRoleNotFoundFault":{ + "base":"

The specified IAM role Amazon Resource Name (ARN) isn't associated with the specified DB cluster.

", + "refs":{ } }, - "DBClusterRoles": { - "base": null, - "refs": { - "DBCluster$AssociatedRoles": "

Provides a list of the AWS Identity and Access Management (IAM) roles that are associated with the DB cluster. IAM roles that are associated with a DB cluster grant permission for the DB cluster to access other AWS services on your behalf.

" + "DBClusterRoleQuotaExceededFault":{ + "base":"

You have exceeded the maximum number of IAM roles that can be associated with the specified DB cluster.

", + "refs":{ } }, - "DBClusterSnapshot": { - "base": "

Contains the details for an Amazon RDS DB cluster snapshot

This data type is used as a response element in the DescribeDBClusterSnapshots action.

", - "refs": { - "CopyDBClusterSnapshotResult$DBClusterSnapshot": null, - "CreateDBClusterSnapshotResult$DBClusterSnapshot": null, - "DBClusterSnapshotList$member": null, - "DeleteDBClusterSnapshotResult$DBClusterSnapshot": null + "DBClusterRoles":{ + "base":null, + "refs":{ + "DBCluster$AssociatedRoles":"

Provides a list of the AWS Identity and Access Management (IAM) roles that are associated with the DB cluster. IAM roles that are associated with a DB cluster grant permission for the DB cluster to access other AWS services on your behalf.

" } }, - "DBClusterSnapshotAlreadyExistsFault": { - "base": "

The user already has a DB cluster snapshot with the given identifier.

", - "refs": { + "DBClusterSnapshot":{ + "base":"

Contains the details for an Amazon RDS DB cluster snapshot

This data type is used as a response element in the DescribeDBClusterSnapshots action.

", + "refs":{ + "CopyDBClusterSnapshotResult$DBClusterSnapshot":null, + "CreateDBClusterSnapshotResult$DBClusterSnapshot":null, + "DBClusterSnapshotList$member":null, + "DeleteDBClusterSnapshotResult$DBClusterSnapshot":null } }, - "DBClusterSnapshotAttribute": { - "base": "

Contains the name and values of a manual DB cluster snapshot attribute.

Manual DB cluster snapshot attributes are used to authorize other AWS accounts to restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", - "refs": { - "DBClusterSnapshotAttributeList$member": null + "DBClusterSnapshotAlreadyExistsFault":{ + "base":"

The user already has a DB cluster snapshot with the given identifier.

", + "refs":{ } }, - "DBClusterSnapshotAttributeList": { - "base": null, - "refs": { - "DBClusterSnapshotAttributesResult$DBClusterSnapshotAttributes": "

The list of attributes and values for the manual DB cluster snapshot.

" + "DBClusterSnapshotAttribute":{ + "base":"

Contains the name and values of a manual DB cluster snapshot attribute.

Manual DB cluster snapshot attributes are used to authorize other AWS accounts to restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", + "refs":{ + "DBClusterSnapshotAttributeList$member":null } }, - "DBClusterSnapshotAttributesResult": { - "base": "

Contains the results of a successful call to the DescribeDBClusterSnapshotAttributes API action.

Manual DB cluster snapshot attributes are used to authorize other AWS accounts to copy or restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", - "refs": { - "DescribeDBClusterSnapshotAttributesResult$DBClusterSnapshotAttributesResult": null, - "ModifyDBClusterSnapshotAttributeResult$DBClusterSnapshotAttributesResult": null + "DBClusterSnapshotAttributeList":{ + "base":null, + "refs":{ + "DBClusterSnapshotAttributesResult$DBClusterSnapshotAttributes":"

The list of attributes and values for the manual DB cluster snapshot.

" } }, - "DBClusterSnapshotList": { - "base": null, - "refs": { - "DBClusterSnapshotMessage$DBClusterSnapshots": "

Provides a list of DB cluster snapshots for the user.

" + "DBClusterSnapshotAttributesResult":{ + "base":"

Contains the results of a successful call to the DescribeDBClusterSnapshotAttributes API action.

Manual DB cluster snapshot attributes are used to authorize other AWS accounts to copy or restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", + "refs":{ + "DescribeDBClusterSnapshotAttributesResult$DBClusterSnapshotAttributesResult":null, + "ModifyDBClusterSnapshotAttributeResult$DBClusterSnapshotAttributesResult":null } }, - "DBClusterSnapshotMessage": { - "base": "

Provides a list of DB cluster snapshots for the user as the result of a call to the DescribeDBClusterSnapshots action.

", - "refs": { + "DBClusterSnapshotList":{ + "base":null, + "refs":{ + "DBClusterSnapshotMessage$DBClusterSnapshots":"

Provides a list of DB cluster snapshots for the user.

" } }, - "DBClusterSnapshotNotFoundFault": { - "base": "

DBClusterSnapshotIdentifier doesn't refer to an existing DB cluster snapshot.

", - "refs": { + "DBClusterSnapshotMessage":{ + "base":"

Provides a list of DB cluster snapshots for the user as the result of a call to the DescribeDBClusterSnapshots action.

", + "refs":{ } }, - "DBEngineVersion": { - "base": "

This data type is used as a response element in the action DescribeDBEngineVersions.

", - "refs": { - "DBEngineVersionList$member": null + "DBClusterSnapshotNotFoundFault":{ + "base":"

DBClusterSnapshotIdentifier doesn't refer to an existing DB cluster snapshot.

", + "refs":{ } }, - "DBEngineVersionList": { - "base": null, - "refs": { - "DBEngineVersionMessage$DBEngineVersions": "

A list of DBEngineVersion elements.

" + "DBEngineVersion":{ + "base":"

This data type is used as a response element in the action DescribeDBEngineVersions.

", + "refs":{ + "DBEngineVersionList$member":null } }, - "DBEngineVersionMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBEngineVersions action.

", - "refs": { + "DBEngineVersionList":{ + "base":null, + "refs":{ + "DBEngineVersionMessage$DBEngineVersions":"

A list of DBEngineVersion elements.

" } }, - "DBInstance": { - "base": "

Contains the details of an Amazon RDS DB instance.

This data type is used as a response element in the DescribeDBInstances action.

", - "refs": { - "CreateDBInstanceReadReplicaResult$DBInstance": null, - "CreateDBInstanceResult$DBInstance": null, - "DBInstanceList$member": null, - "DeleteDBInstanceResult$DBInstance": null, - "ModifyDBInstanceResult$DBInstance": null, - "PromoteReadReplicaResult$DBInstance": null, - "RebootDBInstanceResult$DBInstance": null, - "RestoreDBInstanceFromDBSnapshotResult$DBInstance": null, - "RestoreDBInstanceFromS3Result$DBInstance": null, - "RestoreDBInstanceToPointInTimeResult$DBInstance": null, - "StartDBInstanceResult$DBInstance": null, - "StopDBInstanceResult$DBInstance": null + "DBEngineVersionMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBEngineVersions action.

", + "refs":{ } }, - "DBInstanceAlreadyExistsFault": { - "base": "

The user already has a DB instance with the given identifier.

", - "refs": { + "DBInstance":{ + "base":"

Contains the details of an Amazon RDS DB instance.

This data type is used as a response element in the DescribeDBInstances action.

", + "refs":{ + "CreateDBInstanceReadReplicaResult$DBInstance":null, + "CreateDBInstanceResult$DBInstance":null, + "DBInstanceList$member":null, + "DeleteDBInstanceResult$DBInstance":null, + "ModifyDBInstanceResult$DBInstance":null, + "PromoteReadReplicaResult$DBInstance":null, + "RebootDBInstanceResult$DBInstance":null, + "RestoreDBInstanceFromDBSnapshotResult$DBInstance":null, + "RestoreDBInstanceFromS3Result$DBInstance":null, + "RestoreDBInstanceToPointInTimeResult$DBInstance":null, + "StartDBInstanceResult$DBInstance":null, + "StopDBInstanceResult$DBInstance":null } }, - "DBInstanceList": { - "base": null, - "refs": { - "DBInstanceMessage$DBInstances": "

A list of DBInstance instances.

" + "DBInstanceAlreadyExistsFault":{ + "base":"

The user already has a DB instance with the given identifier.

", + "refs":{ } }, - "DBInstanceMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBInstances action.

", - "refs": { + "DBInstanceList":{ + "base":null, + "refs":{ + "DBInstanceMessage$DBInstances":"

A list of DBInstance instances.

" } }, - "DBInstanceNotFoundFault": { - "base": "

DBInstanceIdentifier doesn't refer to an existing DB instance.

", - "refs": { + "DBInstanceMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBInstances action.

", + "refs":{ } }, - "DBInstanceStatusInfo": { - "base": "

Provides a list of status information for a DB instance.

", - "refs": { - "DBInstanceStatusInfoList$member": null + "DBInstanceNotFoundFault":{ + "base":"

DBInstanceIdentifier doesn't refer to an existing DB instance.

", + "refs":{ } }, - "DBInstanceStatusInfoList": { - "base": null, - "refs": { - "DBInstance$StatusInfos": "

The status of a Read Replica. If the instance is not a Read Replica, this is blank.

" + "DBInstanceStatusInfo":{ + "base":"

Provides a list of status information for a DB instance.

", + "refs":{ + "DBInstanceStatusInfoList$member":null } }, - "DBLogFileNotFoundFault": { - "base": "

LogFileName doesn't refer to an existing DB log file.

", - "refs": { + "DBInstanceStatusInfoList":{ + "base":null, + "refs":{ + "DBInstance$StatusInfos":"

The status of a Read Replica. If the instance is not a Read Replica, this is blank.

" } }, - "DBParameterGroup": { - "base": "

Contains the details of an Amazon RDS DB parameter group.

This data type is used as a response element in the DescribeDBParameterGroups action.

", - "refs": { - "CopyDBParameterGroupResult$DBParameterGroup": null, - "CreateDBParameterGroupResult$DBParameterGroup": null, - "DBParameterGroupList$member": null + "DBLogFileNotFoundFault":{ + "base":"

LogFileName doesn't refer to an existing DB log file.

", + "refs":{ } }, - "DBParameterGroupAlreadyExistsFault": { - "base": "

A DB parameter group with the same name exists.

", - "refs": { + "DBParameterGroup":{ + "base":"

Contains the details of an Amazon RDS DB parameter group.

This data type is used as a response element in the DescribeDBParameterGroups action.

", + "refs":{ + "CopyDBParameterGroupResult$DBParameterGroup":null, + "CreateDBParameterGroupResult$DBParameterGroup":null, + "DBParameterGroupList$member":null } }, - "DBParameterGroupDetails": { - "base": "

Contains the result of a successful invocation of the DescribeDBParameters action.

", - "refs": { + "DBParameterGroupAlreadyExistsFault":{ + "base":"

A DB parameter group with the same name exists.

", + "refs":{ } }, - "DBParameterGroupList": { - "base": null, - "refs": { - "DBParameterGroupsMessage$DBParameterGroups": "

A list of DBParameterGroup instances.

" + "DBParameterGroupDetails":{ + "base":"

Contains the result of a successful invocation of the DescribeDBParameters action.

", + "refs":{ } }, - "DBParameterGroupNameMessage": { - "base": "

Contains the result of a successful invocation of the ModifyDBParameterGroup or ResetDBParameterGroup action.

", - "refs": { + "DBParameterGroupList":{ + "base":null, + "refs":{ + "DBParameterGroupsMessage$DBParameterGroups":"

A list of DBParameterGroup instances.

" } }, - "DBParameterGroupNotFoundFault": { - "base": "

DBParameterGroupName doesn't refer to an existing DB parameter group.

", - "refs": { + "DBParameterGroupNameMessage":{ + "base":"

Contains the result of a successful invocation of the ModifyDBParameterGroup or ResetDBParameterGroup action.

", + "refs":{ } }, - "DBParameterGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB parameter groups.

", - "refs": { + "DBParameterGroupNotFoundFault":{ + "base":"

DBParameterGroupName doesn't refer to an existing DB parameter group.

", + "refs":{ } }, - "DBParameterGroupStatus": { - "base": "

The status of the DB parameter group.

This data type is used as a response element in the following actions:

", - "refs": { - "DBParameterGroupStatusList$member": null + "DBParameterGroupQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed number of DB parameter groups.

", + "refs":{ } }, - "DBParameterGroupStatusList": { - "base": null, - "refs": { - "DBInstance$DBParameterGroups": "

Provides the list of DB parameter groups applied to this DB instance.

" + "DBParameterGroupStatus":{ + "base":"

The status of the DB parameter group.

This data type is used as a response element in the following actions:

", + "refs":{ + "DBParameterGroupStatusList$member":null } }, - "DBParameterGroupsMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBParameterGroups action.

", - "refs": { + "DBParameterGroupStatusList":{ + "base":null, + "refs":{ + "DBInstance$DBParameterGroups":"

Provides the list of DB parameter groups applied to this DB instance.

" } }, - "DBSecurityGroup": { - "base": "

Contains the details for an Amazon RDS DB security group.

This data type is used as a response element in the DescribeDBSecurityGroups action.

", - "refs": { - "AuthorizeDBSecurityGroupIngressResult$DBSecurityGroup": null, - "CreateDBSecurityGroupResult$DBSecurityGroup": null, - "DBSecurityGroups$member": null, - "RevokeDBSecurityGroupIngressResult$DBSecurityGroup": null + "DBParameterGroupsMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBParameterGroups action.

", + "refs":{ } }, - "DBSecurityGroupAlreadyExistsFault": { - "base": "

A DB security group with the name specified in DBSecurityGroupName already exists.

", - "refs": { + "DBSecurityGroup":{ + "base":"

Contains the details for an Amazon RDS DB security group.

This data type is used as a response element in the DescribeDBSecurityGroups action.

", + "refs":{ + "AuthorizeDBSecurityGroupIngressResult$DBSecurityGroup":null, + "CreateDBSecurityGroupResult$DBSecurityGroup":null, + "DBSecurityGroups$member":null, + "RevokeDBSecurityGroupIngressResult$DBSecurityGroup":null } }, - "DBSecurityGroupMembership": { - "base": "

This data type is used as a response element in the following actions:

", - "refs": { - "DBSecurityGroupMembershipList$member": null + "DBSecurityGroupAlreadyExistsFault":{ + "base":"

A DB security group with the name specified in DBSecurityGroupName already exists.

", + "refs":{ } }, - "DBSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$DBSecurityGroups": "

Provides List of DB security group elements containing only DBSecurityGroup.Name and DBSecurityGroup.Status subelements.

", - "Option$DBSecurityGroupMemberships": "

If the option requires access to a port, then this DB security group allows access to the port.

" + "DBSecurityGroupMembership":{ + "base":"

This data type is used as a response element in the following actions:

", + "refs":{ + "DBSecurityGroupMembershipList$member":null } }, - "DBSecurityGroupMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBSecurityGroups action.

", - "refs": { + "DBSecurityGroupMembershipList":{ + "base":null, + "refs":{ + "DBInstance$DBSecurityGroups":"

Provides List of DB security group elements containing only DBSecurityGroup.Name and DBSecurityGroup.Status subelements.

", + "Option$DBSecurityGroupMemberships":"

If the option requires access to a port, then this DB security group allows access to the port.

" } }, - "DBSecurityGroupNameList": { - "base": null, - "refs": { - "CreateDBInstanceMessage$DBSecurityGroups": "

A list of DB security groups to associate with this DB instance.

Default: The default DB security group for the database engine.

", - "ModifyDBInstanceMessage$DBSecurityGroups": "

A list of DB security groups to authorize on this DB instance. Changing this setting doesn't result in an outage and the change is asynchronously applied as soon as possible.

Constraints:

  • If supplied, must match existing DBSecurityGroups.

", - "OptionConfiguration$DBSecurityGroupMemberships": "

A list of DBSecurityGroupMemebrship name strings used for this option.

", - "RestoreDBInstanceFromS3Message$DBSecurityGroups": "

A list of DB security groups to associate with this DB instance.

Default: The default DB security group for the database engine.

" + "DBSecurityGroupMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBSecurityGroups action.

", + "refs":{ } }, - "DBSecurityGroupNotFoundFault": { - "base": "

DBSecurityGroupName doesn't refer to an existing DB security group.

", - "refs": { + "DBSecurityGroupNameList":{ + "base":null, + "refs":{ + "CreateDBInstanceMessage$DBSecurityGroups":"

A list of DB security groups to associate with this DB instance.

Default: The default DB security group for the database engine.

", + "ModifyDBInstanceMessage$DBSecurityGroups":"

A list of DB security groups to authorize on this DB instance. Changing this setting doesn't result in an outage and the change is asynchronously applied as soon as possible.

Constraints:

  • If supplied, must match existing DBSecurityGroups.

", + "OptionConfiguration$DBSecurityGroupMemberships":"

A list of DBSecurityGroupMemebrship name strings used for this option.

", + "RestoreDBInstanceFromS3Message$DBSecurityGroups":"

A list of DB security groups to associate with this DB instance.

Default: The default DB security group for the database engine.

" } }, - "DBSecurityGroupNotSupportedFault": { - "base": "

A DB security group isn't allowed for this action.

", - "refs": { + "DBSecurityGroupNotFoundFault":{ + "base":"

DBSecurityGroupName doesn't refer to an existing DB security group.

", + "refs":{ } }, - "DBSecurityGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB security groups.

", - "refs": { + "DBSecurityGroupNotSupportedFault":{ + "base":"

A DB security group isn't allowed for this action.

", + "refs":{ } }, - "DBSecurityGroups": { - "base": null, - "refs": { - "DBSecurityGroupMessage$DBSecurityGroups": "

A list of DBSecurityGroup instances.

" + "DBSecurityGroupQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed number of DB security groups.

", + "refs":{ } }, - "DBSnapshot": { - "base": "

Contains the details of an Amazon RDS DB snapshot.

This data type is used as a response element in the DescribeDBSnapshots action.

", - "refs": { - "CopyDBSnapshotResult$DBSnapshot": null, - "CreateDBSnapshotResult$DBSnapshot": null, - "DBSnapshotList$member": null, - "DeleteDBSnapshotResult$DBSnapshot": null, - "ModifyDBSnapshotResult$DBSnapshot": null + "DBSecurityGroups":{ + "base":null, + "refs":{ + "DBSecurityGroupMessage$DBSecurityGroups":"

A list of DBSecurityGroup instances.

" } }, - "DBSnapshotAlreadyExistsFault": { - "base": "

DBSnapshotIdentifier is already used by an existing snapshot.

", - "refs": { + "DBSnapshot":{ + "base":"

Contains the details of an Amazon RDS DB snapshot.

This data type is used as a response element in the DescribeDBSnapshots action.

", + "refs":{ + "CopyDBSnapshotResult$DBSnapshot":null, + "CreateDBSnapshotResult$DBSnapshot":null, + "DBSnapshotList$member":null, + "DeleteDBSnapshotResult$DBSnapshot":null, + "ModifyDBSnapshotResult$DBSnapshot":null } }, - "DBSnapshotAttribute": { - "base": "

Contains the name and values of a manual DB snapshot attribute

Manual DB snapshot attributes are used to authorize other AWS accounts to restore a manual DB snapshot. For more information, see the ModifyDBSnapshotAttribute API.

", - "refs": { - "DBSnapshotAttributeList$member": null + "DBSnapshotAlreadyExistsFault":{ + "base":"

DBSnapshotIdentifier is already used by an existing snapshot.

", + "refs":{ } }, - "DBSnapshotAttributeList": { - "base": null, - "refs": { - "DBSnapshotAttributesResult$DBSnapshotAttributes": "

The list of attributes and values for the manual DB snapshot.

" + "DBSnapshotAttribute":{ + "base":"

Contains the name and values of a manual DB snapshot attribute

Manual DB snapshot attributes are used to authorize other AWS accounts to restore a manual DB snapshot. For more information, see the ModifyDBSnapshotAttribute API.

", + "refs":{ + "DBSnapshotAttributeList$member":null } }, - "DBSnapshotAttributesResult": { - "base": "

Contains the results of a successful call to the DescribeDBSnapshotAttributes API action.

Manual DB snapshot attributes are used to authorize other AWS accounts to copy or restore a manual DB snapshot. For more information, see the ModifyDBSnapshotAttribute API action.

", - "refs": { - "DescribeDBSnapshotAttributesResult$DBSnapshotAttributesResult": null, - "ModifyDBSnapshotAttributeResult$DBSnapshotAttributesResult": null + "DBSnapshotAttributeList":{ + "base":null, + "refs":{ + "DBSnapshotAttributesResult$DBSnapshotAttributes":"

The list of attributes and values for the manual DB snapshot.

" } }, - "DBSnapshotList": { - "base": null, - "refs": { - "DBSnapshotMessage$DBSnapshots": "

A list of DBSnapshot instances.

" + "DBSnapshotAttributesResult":{ + "base":"

Contains the results of a successful call to the DescribeDBSnapshotAttributes API action.

Manual DB snapshot attributes are used to authorize other AWS accounts to copy or restore a manual DB snapshot. For more information, see the ModifyDBSnapshotAttribute API action.

", + "refs":{ + "DescribeDBSnapshotAttributesResult$DBSnapshotAttributesResult":null, + "ModifyDBSnapshotAttributeResult$DBSnapshotAttributesResult":null } }, - "DBSnapshotMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBSnapshots action.

", - "refs": { + "DBSnapshotList":{ + "base":null, + "refs":{ + "DBSnapshotMessage$DBSnapshots":"

A list of DBSnapshot instances.

" } }, - "DBSnapshotNotFoundFault": { - "base": "

DBSnapshotIdentifier doesn't refer to an existing DB snapshot.

", - "refs": { + "DBSnapshotMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBSnapshots action.

", + "refs":{ } }, - "DBSubnetGroup": { - "base": "

Contains the details of an Amazon RDS DB subnet group.

This data type is used as a response element in the DescribeDBSubnetGroups action.

", - "refs": { - "CreateDBSubnetGroupResult$DBSubnetGroup": null, - "DBInstance$DBSubnetGroup": "

Specifies information on the subnet group associated with the DB instance, including the name, description, and subnets in the subnet group.

", - "DBSubnetGroups$member": null, - "ModifyDBSubnetGroupResult$DBSubnetGroup": null + "DBSnapshotNotFoundFault":{ + "base":"

DBSnapshotIdentifier doesn't refer to an existing DB snapshot.

", + "refs":{ } }, - "DBSubnetGroupAlreadyExistsFault": { - "base": "

DBSubnetGroupName is already used by an existing DB subnet group.

", - "refs": { + "DBSubnetGroup":{ + "base":"

Contains the details of an Amazon RDS DB subnet group.

This data type is used as a response element in the DescribeDBSubnetGroups action.

", + "refs":{ + "CreateDBSubnetGroupResult$DBSubnetGroup":null, + "DBInstance$DBSubnetGroup":"

Specifies information on the subnet group associated with the DB instance, including the name, description, and subnets in the subnet group.

", + "DBSubnetGroups$member":null, + "ModifyDBSubnetGroupResult$DBSubnetGroup":null } }, - "DBSubnetGroupDoesNotCoverEnoughAZs": { - "base": "

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", - "refs": { + "DBSubnetGroupAlreadyExistsFault":{ + "base":"

DBSubnetGroupName is already used by an existing DB subnet group.

", + "refs":{ } }, - "DBSubnetGroupMessage": { - "base": "

Contains the result of a successful invocation of the DescribeDBSubnetGroups action.

", - "refs": { + "DBSubnetGroupDoesNotCoverEnoughAZs":{ + "base":"

Subnets in the DB subnet group should cover at least two Availability Zones unless there is only one Availability Zone.

", + "refs":{ } }, - "DBSubnetGroupNotAllowedFault": { - "base": "

The DBSubnetGroup shouldn't be specified while creating read replicas that lie in the same region as the source instance.

", - "refs": { + "DBSubnetGroupMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeDBSubnetGroups action.

", + "refs":{ } }, - "DBSubnetGroupNotFoundFault": { - "base": "

DBSubnetGroupName doesn't refer to an existing DB subnet group.

", - "refs": { + "DBSubnetGroupNotAllowedFault":{ + "base":"

The DBSubnetGroup shouldn't be specified while creating read replicas that lie in the same region as the source instance.

", + "refs":{ } }, - "DBSubnetGroupQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB subnet groups.

", - "refs": { + "DBSubnetGroupNotFoundFault":{ + "base":"

DBSubnetGroupName doesn't refer to an existing DB subnet group.

", + "refs":{ } }, - "DBSubnetGroups": { - "base": null, - "refs": { - "DBSubnetGroupMessage$DBSubnetGroups": "

A list of DBSubnetGroup instances.

" + "DBSubnetGroupQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed number of DB subnet groups.

", + "refs":{ } }, - "DBSubnetQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of subnets in a DB subnet groups.

", - "refs": { + "DBSubnetGroups":{ + "base":null, + "refs":{ + "DBSubnetGroupMessage$DBSubnetGroups":"

A list of DBSubnetGroup instances.

" } }, - "DBUpgradeDependencyFailureFault": { - "base": "

The DB upgrade failed because a resource the DB depends on can't be modified.

", - "refs": { + "DBSubnetQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed number of subnets in a DB subnet groups.

", + "refs":{ } }, - "DeleteDBClusterMessage": { - "base": "

", - "refs": { + "DBUpgradeDependencyFailureFault":{ + "base":"

The DB upgrade failed because a resource the DB depends on can't be modified.

", + "refs":{ } }, - "DeleteDBClusterParameterGroupMessage": { - "base": "

", - "refs": { + "DeleteDBClusterMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBClusterResult": { - "base": null, - "refs": { + "DeleteDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBClusterSnapshotMessage": { - "base": "

", - "refs": { + "DeleteDBClusterResult":{ + "base":null, + "refs":{ } }, - "DeleteDBClusterSnapshotResult": { - "base": null, - "refs": { + "DeleteDBClusterSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBInstanceMessage": { - "base": "

", - "refs": { + "DeleteDBClusterSnapshotResult":{ + "base":null, + "refs":{ } }, - "DeleteDBInstanceResult": { - "base": null, - "refs": { + "DeleteDBInstanceMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBParameterGroupMessage": { - "base": "

", - "refs": { + "DeleteDBInstanceResult":{ + "base":null, + "refs":{ } }, - "DeleteDBSecurityGroupMessage": { - "base": "

", - "refs": { + "DeleteDBParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBSnapshotMessage": { - "base": "

", - "refs": { + "DeleteDBSecurityGroupMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBSnapshotResult": { - "base": null, - "refs": { + "DeleteDBSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "DeleteDBSubnetGroupMessage": { - "base": "

", - "refs": { + "DeleteDBSnapshotResult":{ + "base":null, + "refs":{ } }, - "DeleteEventSubscriptionMessage": { - "base": "

", - "refs": { + "DeleteDBSubnetGroupMessage":{ + "base":"

", + "refs":{ } }, - "DeleteEventSubscriptionResult": { - "base": null, - "refs": { + "DeleteEventSubscriptionMessage":{ + "base":"

", + "refs":{ } }, - "DeleteOptionGroupMessage": { - "base": "

", - "refs": { + "DeleteEventSubscriptionResult":{ + "base":null, + "refs":{ } }, - "DescribeAccountAttributesMessage": { - "base": "

", - "refs": { + "DeleteOptionGroupMessage":{ + "base":"

", + "refs":{ } }, - "DescribeCertificatesMessage": { - "base": "

", - "refs": { + "DescribeAccountAttributesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBClusterBacktracksMessage": { - "base": "

", - "refs": { + "DescribeCertificatesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBClusterParameterGroupsMessage": { - "base": "

", - "refs": { + "DescribeDBClusterBacktracksMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBClusterParametersMessage": { - "base": "

", - "refs": { + "DescribeDBClusterParameterGroupsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBClusterSnapshotAttributesMessage": { - "base": "

", - "refs": { + "DescribeDBClusterParametersMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBClusterSnapshotAttributesResult": { - "base": null, - "refs": { + "DescribeDBClusterSnapshotAttributesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBClusterSnapshotsMessage": { - "base": "

", - "refs": { + "DescribeDBClusterSnapshotAttributesResult":{ + "base":null, + "refs":{ } }, - "DescribeDBClustersMessage": { - "base": "

", - "refs": { + "DescribeDBClusterSnapshotsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBEngineVersionsMessage": { - "base": null, - "refs": { + "DescribeDBClustersMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBInstancesMessage": { - "base": "

", - "refs": { + "DescribeDBEngineVersionsMessage":{ + "base":null, + "refs":{ } }, - "DescribeDBLogFilesDetails": { - "base": "

This data type is used as a response element to DescribeDBLogFiles.

", - "refs": { - "DescribeDBLogFilesList$member": null + "DescribeDBInstancesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBLogFilesList": { - "base": null, - "refs": { - "DescribeDBLogFilesResponse$DescribeDBLogFiles": "

The DB log files returned.

" + "DescribeDBLogFilesDetails":{ + "base":"

This data type is used as a response element to DescribeDBLogFiles.

", + "refs":{ + "DescribeDBLogFilesList$member":null } }, - "DescribeDBLogFilesMessage": { - "base": "

", - "refs": { + "DescribeDBLogFilesList":{ + "base":null, + "refs":{ + "DescribeDBLogFilesResponse$DescribeDBLogFiles":"

The DB log files returned.

" } }, - "DescribeDBLogFilesResponse": { - "base": "

The response from a call to DescribeDBLogFiles.

", - "refs": { + "DescribeDBLogFilesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBParameterGroupsMessage": { - "base": "

", - "refs": { + "DescribeDBLogFilesResponse":{ + "base":"

The response from a call to DescribeDBLogFiles.

", + "refs":{ } }, - "DescribeDBParametersMessage": { - "base": null, - "refs": { + "DescribeDBParameterGroupsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBSecurityGroupsMessage": { - "base": "

", - "refs": { + "DescribeDBParametersMessage":{ + "base":null, + "refs":{ } }, - "DescribeDBSnapshotAttributesMessage": { - "base": "

", - "refs": { + "DescribeDBSecurityGroupsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBSnapshotAttributesResult": { - "base": null, - "refs": { + "DescribeDBSnapshotAttributesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeDBSnapshotsMessage": { - "base": "

", - "refs": { + "DescribeDBSnapshotAttributesResult":{ + "base":null, + "refs":{ } }, - "DescribeDBSubnetGroupsMessage": { - "base": "

", - "refs": { + "DescribeDBSnapshotsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeEngineDefaultClusterParametersMessage": { - "base": "

", - "refs": { + "DescribeDBSubnetGroupsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeEngineDefaultClusterParametersResult": { - "base": null, - "refs": { + "DescribeEngineDefaultClusterParametersMessage":{ + "base":"

", + "refs":{ } }, - "DescribeEngineDefaultParametersMessage": { - "base": "

", - "refs": { + "DescribeEngineDefaultClusterParametersResult":{ + "base":null, + "refs":{ } }, - "DescribeEngineDefaultParametersResult": { - "base": null, - "refs": { + "DescribeEngineDefaultParametersMessage":{ + "base":"

", + "refs":{ } }, - "DescribeEventCategoriesMessage": { - "base": "

", - "refs": { + "DescribeEngineDefaultParametersResult":{ + "base":null, + "refs":{ } }, - "DescribeEventSubscriptionsMessage": { - "base": "

", - "refs": { + "DescribeEventCategoriesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeEventsMessage": { - "base": "

", - "refs": { + "DescribeEventSubscriptionsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeOptionGroupOptionsMessage": { - "base": "

", - "refs": { + "DescribeEventsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeOptionGroupsMessage": { - "base": "

", - "refs": { + "DescribeOptionGroupOptionsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeOrderableDBInstanceOptionsMessage": { - "base": "

", - "refs": { + "DescribeOptionGroupsMessage":{ + "base":"

", + "refs":{ } }, - "DescribePendingMaintenanceActionsMessage": { - "base": "

", - "refs": { + "DescribeOrderableDBInstanceOptionsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeReservedDBInstancesMessage": { - "base": "

", - "refs": { + "DescribePendingMaintenanceActionsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeReservedDBInstancesOfferingsMessage": { - "base": "

", - "refs": { + "DescribeReservedDBInstancesMessage":{ + "base":"

", + "refs":{ } }, - "DescribeSourceRegionsMessage": { - "base": "

", - "refs": { + "DescribeReservedDBInstancesOfferingsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeValidDBInstanceModificationsMessage": { - "base": "

", - "refs": { + "DescribeSourceRegionsMessage":{ + "base":"

", + "refs":{ } }, - "DescribeValidDBInstanceModificationsResult": { - "base": null, - "refs": { + "DescribeValidDBInstanceModificationsMessage":{ + "base":"

", + "refs":{ } }, - "DomainMembership": { - "base": "

An Active Directory Domain membership record associated with the DB instance.

", - "refs": { - "DomainMembershipList$member": null + "DescribeValidDBInstanceModificationsResult":{ + "base":null, + "refs":{ } }, - "DomainMembershipList": { - "base": "

List of Active Directory Domain membership records associated with a DB instance.

", - "refs": { - "DBInstance$DomainMemberships": "

The Active Directory Domain membership records associated with the DB instance.

" + "DomainMembership":{ + "base":"

An Active Directory Domain membership record associated with the DB instance.

", + "refs":{ + "DomainMembershipList$member":null } }, - "DomainNotFoundFault": { - "base": "

Domain doesn't refer to an existing Active Directory domain.

", - "refs": { + "DomainMembershipList":{ + "base":"

List of Active Directory Domain membership records associated with a DB instance.

", + "refs":{ + "DBInstance$DomainMemberships":"

The Active Directory Domain membership records associated with the DB instance.

" } }, - "Double": { - "base": null, - "refs": { - "DoubleRange$From": "

The minimum value in the range.

", - "DoubleRange$To": "

The maximum value in the range.

", - "RecurringCharge$RecurringChargeAmount": "

The amount of the recurring charge.

", - "ReservedDBInstance$FixedPrice": "

The fixed price charged for this reserved DB instance.

", - "ReservedDBInstance$UsagePrice": "

The hourly price charged for this reserved DB instance.

", - "ReservedDBInstancesOffering$FixedPrice": "

The fixed price charged for this offering.

", - "ReservedDBInstancesOffering$UsagePrice": "

The hourly price charged for this offering.

" + "DomainNotFoundFault":{ + "base":"

Domain doesn't refer to an existing Active Directory domain.

", + "refs":{ } }, - "DoubleOptional": { - "base": null, - "refs": { - "OrderableDBInstanceOption$MinIopsPerGib": "

Minimum provisioned IOPS per GiB for a DB instance.

", - "OrderableDBInstanceOption$MaxIopsPerGib": "

Maximum provisioned IOPS per GiB for a DB instance.

" + "Double":{ + "base":null, + "refs":{ + "DoubleRange$From":"

The minimum value in the range.

", + "DoubleRange$To":"

The maximum value in the range.

", + "RecurringCharge$RecurringChargeAmount":"

The amount of the recurring charge.

", + "ReservedDBInstance$FixedPrice":"

The fixed price charged for this reserved DB instance.

", + "ReservedDBInstance$UsagePrice":"

The hourly price charged for this reserved DB instance.

", + "ReservedDBInstancesOffering$FixedPrice":"

The fixed price charged for this offering.

", + "ReservedDBInstancesOffering$UsagePrice":"

The hourly price charged for this offering.

" } }, - "DoubleRange": { - "base": "

A range of double values.

", - "refs": { - "DoubleRangeList$member": null + "DoubleOptional":{ + "base":null, + "refs":{ + "OrderableDBInstanceOption$MinIopsPerGib":"

Minimum provisioned IOPS per GiB for a DB instance.

", + "OrderableDBInstanceOption$MaxIopsPerGib":"

Maximum provisioned IOPS per GiB for a DB instance.

" } }, - "DoubleRangeList": { - "base": null, - "refs": { - "ValidStorageOptions$IopsToStorageRatio": "

The valid range of Provisioned IOPS to gibibytes of storage multiplier. For example, 3-10, which means that provisioned IOPS can be between 3 and 10 times storage.

" + "DoubleRange":{ + "base":"

A range of double values.

", + "refs":{ + "DoubleRangeList$member":null } }, - "DownloadDBLogFilePortionDetails": { - "base": "

This data type is used as a response element to DownloadDBLogFilePortion.

", - "refs": { + "DoubleRangeList":{ + "base":null, + "refs":{ + "ValidStorageOptions$IopsToStorageRatio":"

The valid range of Provisioned IOPS to gibibytes of storage multiplier. For example, 3-10, which means that provisioned IOPS can be between 3 and 10 times storage.

" } }, - "DownloadDBLogFilePortionMessage": { - "base": "

", - "refs": { + "DownloadDBLogFilePortionDetails":{ + "base":"

This data type is used as a response element to DownloadDBLogFilePortion.

", + "refs":{ } }, - "EC2SecurityGroup": { - "base": "

This data type is used as a response element in the following actions:

", - "refs": { - "EC2SecurityGroupList$member": null + "DownloadDBLogFilePortionMessage":{ + "base":"

", + "refs":{ } }, - "EC2SecurityGroupList": { - "base": null, - "refs": { - "DBSecurityGroup$EC2SecurityGroups": "

Contains a list of EC2SecurityGroup elements.

" + "EC2SecurityGroup":{ + "base":"

This data type is used as a response element in the following actions:

", + "refs":{ + "EC2SecurityGroupList$member":null } }, - "Endpoint": { - "base": "

This data type is used as a response element in the following actions:

", - "refs": { - "DBInstance$Endpoint": "

Specifies the connection endpoint.

" + "EC2SecurityGroupList":{ + "base":null, + "refs":{ + "DBSecurityGroup$EC2SecurityGroups":"

Contains a list of EC2SecurityGroup elements.

" } }, - "EngineDefaults": { - "base": "

Contains the result of a successful invocation of the DescribeEngineDefaultParameters action.

", - "refs": { - "DescribeEngineDefaultClusterParametersResult$EngineDefaults": null, - "DescribeEngineDefaultParametersResult$EngineDefaults": null + "Endpoint":{ + "base":"

This data type is used as a response element in the following actions:

", + "refs":{ + "DBInstance$Endpoint":"

Specifies the connection endpoint.

" } }, - "Event": { - "base": "

This data type is used as a response element in the DescribeEvents action.

", - "refs": { - "EventList$member": null + "EngineDefaults":{ + "base":"

Contains the result of a successful invocation of the DescribeEngineDefaultParameters action.

", + "refs":{ + "DescribeEngineDefaultClusterParametersResult$EngineDefaults":null, + "DescribeEngineDefaultParametersResult$EngineDefaults":null } }, - "EventCategoriesList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$EventCategories": "

A list of event categories for a SourceType that you want to subscribe to. You can see a list of the categories for a given SourceType in the Events topic in the Amazon RDS User Guide or by using the DescribeEventCategories action.

", - "DescribeEventsMessage$EventCategories": "

A list of event categories that trigger notifications for a event notification subscription.

", - "Event$EventCategories": "

Specifies the category for the event.

", - "EventCategoriesMap$EventCategories": "

The event categories for the specified source type

", - "EventSubscription$EventCategoriesList": "

A list of event categories for the RDS event notification subscription.

", - "ModifyEventSubscriptionMessage$EventCategories": "

A list of event categories for a SourceType that you want to subscribe to. You can see a list of the categories for a given SourceType in the Events topic in the Amazon RDS User Guide or by using the DescribeEventCategories action.

" + "Event":{ + "base":"

This data type is used as a response element in the DescribeEvents action.

", + "refs":{ + "EventList$member":null } }, - "EventCategoriesMap": { - "base": "

Contains the results of a successful invocation of the DescribeEventCategories action.

", - "refs": { - "EventCategoriesMapList$member": null + "EventCategoriesList":{ + "base":null, + "refs":{ + "CreateEventSubscriptionMessage$EventCategories":"

A list of event categories for a SourceType that you want to subscribe to. You can see a list of the categories for a given SourceType in the Events topic in the Amazon RDS User Guide or by using the DescribeEventCategories action.

", + "DescribeEventsMessage$EventCategories":"

A list of event categories that trigger notifications for a event notification subscription.

", + "Event$EventCategories":"

Specifies the category for the event.

", + "EventCategoriesMap$EventCategories":"

The event categories for the specified source type

", + "EventSubscription$EventCategoriesList":"

A list of event categories for the RDS event notification subscription.

", + "ModifyEventSubscriptionMessage$EventCategories":"

A list of event categories for a SourceType that you want to subscribe to. You can see a list of the categories for a given SourceType in the Events topic in the Amazon RDS User Guide or by using the DescribeEventCategories action.

" } }, - "EventCategoriesMapList": { - "base": null, - "refs": { - "EventCategoriesMessage$EventCategoriesMapList": "

A list of EventCategoriesMap data types.

" + "EventCategoriesMap":{ + "base":"

Contains the results of a successful invocation of the DescribeEventCategories action.

", + "refs":{ + "EventCategoriesMapList$member":null } }, - "EventCategoriesMessage": { - "base": "

Data returned from the DescribeEventCategories action.

", - "refs": { + "EventCategoriesMapList":{ + "base":null, + "refs":{ + "EventCategoriesMessage$EventCategoriesMapList":"

A list of EventCategoriesMap data types.

" } }, - "EventList": { - "base": null, - "refs": { - "EventsMessage$Events": "

A list of Event instances.

" + "EventCategoriesMessage":{ + "base":"

Data returned from the DescribeEventCategories action.

", + "refs":{ } }, - "EventSubscription": { - "base": "

Contains the results of a successful invocation of the DescribeEventSubscriptions action.

", - "refs": { - "AddSourceIdentifierToSubscriptionResult$EventSubscription": null, - "CreateEventSubscriptionResult$EventSubscription": null, - "DeleteEventSubscriptionResult$EventSubscription": null, - "EventSubscriptionsList$member": null, - "ModifyEventSubscriptionResult$EventSubscription": null, - "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription": null + "EventList":{ + "base":null, + "refs":{ + "EventsMessage$Events":"

A list of Event instances.

" } }, - "EventSubscriptionQuotaExceededFault": { - "base": "

You have reached the maximum number of event subscriptions.

", - "refs": { + "EventSubscription":{ + "base":"

Contains the results of a successful invocation of the DescribeEventSubscriptions action.

", + "refs":{ + "AddSourceIdentifierToSubscriptionResult$EventSubscription":null, + "CreateEventSubscriptionResult$EventSubscription":null, + "DeleteEventSubscriptionResult$EventSubscription":null, + "EventSubscriptionsList$member":null, + "ModifyEventSubscriptionResult$EventSubscription":null, + "RemoveSourceIdentifierFromSubscriptionResult$EventSubscription":null } }, - "EventSubscriptionsList": { - "base": null, - "refs": { - "EventSubscriptionsMessage$EventSubscriptionsList": "

A list of EventSubscriptions data types.

" + "EventSubscriptionQuotaExceededFault":{ + "base":"

You have reached the maximum number of event subscriptions.

", + "refs":{ } }, - "EventSubscriptionsMessage": { - "base": "

Data returned by the DescribeEventSubscriptions action.

", - "refs": { + "EventSubscriptionsList":{ + "base":null, + "refs":{ + "EventSubscriptionsMessage$EventSubscriptionsList":"

A list of EventSubscriptions data types.

" } }, - "EventsMessage": { - "base": "

Contains the result of a successful invocation of the DescribeEvents action.

", - "refs": { + "EventSubscriptionsMessage":{ + "base":"

Data returned by the DescribeEventSubscriptions action.

", + "refs":{ } }, - "FailoverDBClusterMessage": { - "base": "

", - "refs": { + "EventsMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeEvents action.

", + "refs":{ } }, - "FailoverDBClusterResult": { - "base": null, - "refs": { + "FailoverDBClusterMessage":{ + "base":"

", + "refs":{ } }, - "Filter": { - "base": "

A filter name and value pair that is used to return a more specific list of results from a describe operation. Filters can be used to match a set of resources by specific criteria, such as IDs. The filters supported by a describe operation are documented with the describe operation.

Currently, wildcards are not supported in filters.

The following actions can be filtered:

", - "refs": { - "FilterList$member": null + "FailoverDBClusterResult":{ + "base":null, + "refs":{ } }, - "FilterList": { - "base": null, - "refs": { - "DescribeCertificatesMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBClusterBacktracksMessage$Filters": "

A filter that specifies one or more DB clusters to describe. Supported filters include the following:

  • db-cluster-backtrack-id - Accepts backtrack identifiers. The results list includes information about only the backtracks identified by these identifiers.

  • db-cluster-backtrack-status - Accepts any of the following backtrack status values:

    • applying

    • completed

    • failed

    • pending

    The results list includes information about only the backtracks identified by these values. For more information about backtrack status values, see DBClusterBacktrack.

", - "DescribeDBClusterParameterGroupsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBClusterParametersMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBClusterSnapshotsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBClustersMessage$Filters": "

A filter that specifies one or more DB clusters to describe.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include information about the DB clusters identified by these ARNs.

", - "DescribeDBEngineVersionsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBInstancesMessage$Filters": "

A filter that specifies one or more DB instances to describe.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include information about the DB instances associated with the DB clusters identified by these ARNs.

  • db-instance-id - Accepts DB instance identifiers and DB instance Amazon Resource Names (ARNs). The results list will only include information about the DB instances identified by these ARNs.

", - "DescribeDBLogFilesMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBParameterGroupsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBParametersMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBSecurityGroupsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBSnapshotsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeDBSubnetGroupsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeEngineDefaultClusterParametersMessage$Filters": "

This parameter is not currently supported.

", - "DescribeEngineDefaultParametersMessage$Filters": "

This parameter is not currently supported.

", - "DescribeEventCategoriesMessage$Filters": "

This parameter is not currently supported.

", - "DescribeEventSubscriptionsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeEventsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeOptionGroupOptionsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeOptionGroupsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeOrderableDBInstanceOptionsMessage$Filters": "

This parameter is not currently supported.

", - "DescribePendingMaintenanceActionsMessage$Filters": "

A filter that specifies one or more resources to return pending maintenance actions for.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include pending maintenance actions for the DB clusters identified by these ARNs.

  • db-instance-id - Accepts DB instance identifiers and DB instance ARNs. The results list will only include pending maintenance actions for the DB instances identified by these ARNs.

", - "DescribeReservedDBInstancesMessage$Filters": "

This parameter is not currently supported.

", - "DescribeReservedDBInstancesOfferingsMessage$Filters": "

This parameter is not currently supported.

", - "DescribeSourceRegionsMessage$Filters": "

This parameter is not currently supported.

", - "ListTagsForResourceMessage$Filters": "

This parameter is not currently supported.

" - } - }, - "FilterValueList": { - "base": null, - "refs": { - "Filter$Values": "

One or more filter values. Filter values are case-sensitive.

" - } - }, - "IPRange": { - "base": "

This data type is used as a response element in the DescribeDBSecurityGroups action.

", - "refs": { - "IPRangeList$member": null - } - }, - "IPRangeList": { - "base": null, - "refs": { - "DBSecurityGroup$IPRanges": "

Contains a list of IPRange elements.

" - } - }, - "InstanceQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB instances.

", - "refs": { - } - }, - "InsufficientDBClusterCapacityFault": { - "base": "

The DB cluster doesn't have enough capacity for the current operation.

", - "refs": { - } - }, - "InsufficientDBInstanceCapacityFault": { - "base": "

The specified DB instance class isn't available in the specified Availability Zone.

", - "refs": { - } - }, - "InsufficientStorageClusterCapacityFault": { - "base": "

There is insufficient storage available for the current action. You might be able to resolve this error by updating your subnet group to use different Availability Zones that have more storage available.

", - "refs": { - } - }, - "Integer": { - "base": null, - "refs": { - "DBClusterSnapshot$AllocatedStorage": "

Specifies the allocated storage size in gibibytes (GiB).

", - "DBClusterSnapshot$Port": "

Specifies the port that the DB cluster was listening on at the time of the snapshot.

", - "DBClusterSnapshot$PercentProgress": "

Specifies the percentage of the estimated data that has been transferred.

", - "DBInstance$AllocatedStorage": "

Specifies the allocated storage size specified in gibibytes.

", - "DBInstance$BackupRetentionPeriod": "

Specifies the number of days for which automatic DB snapshots are retained.

", - "DBInstance$DbInstancePort": "

Specifies the port that the DB instance listens on. If the DB instance is part of a DB cluster, this can be a different port than the DB cluster port.

", - "DBSnapshot$AllocatedStorage": "

Specifies the allocated storage size in gibibytes (GiB).

", - "DBSnapshot$Port": "

Specifies the port that the database engine was listening on at the time of the snapshot.

", - "DBSnapshot$PercentProgress": "

The percentage of the estimated data that has been transferred.

", - "DownloadDBLogFilePortionMessage$NumberOfLines": "

The number of lines to download. If the number of lines specified results in a file over 1 MB in size, the file is truncated at 1 MB in size.

If the NumberOfLines parameter is specified, then the block of lines returned can be from the beginning or the end of the log file, depending on the value of the Marker parameter.

  • If neither Marker or NumberOfLines are specified, the entire log file is returned up to a maximum of 10000 lines, starting with the most recent log entries first.

  • If NumberOfLines is specified and Marker is not specified, then the most recent lines from the end of the log file are returned.

  • If Marker is specified as \"0\", then the specified number of lines from the beginning of the log file are returned.

  • You can download the log file in blocks of lines by specifying the size of the block using the NumberOfLines parameter, and by specifying a value of \"0\" for the Marker parameter in your first request. Include the Marker value returned in the response as the Marker value for the next request, continuing until the AdditionalDataPending response element returns false.

", - "Endpoint$Port": "

Specifies the port that the database engine is listening on.

", - "Range$From": "

The minimum value in the range.

", - "Range$To": "

The maximum value in the range.

", - "ReservedDBInstance$Duration": "

The duration of the reservation in seconds.

", - "ReservedDBInstance$DBInstanceCount": "

The number of reserved DB instances.

", - "ReservedDBInstancesOffering$Duration": "

The duration of the offering in seconds.

" - } - }, - "IntegerOptional": { - "base": null, - "refs": { - "CreateDBClusterMessage$BackupRetentionPeriod": "

The number of days for which automated backups are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", - "CreateDBClusterMessage$Port": "

The port number on which the instances in the DB cluster accept connections.

Default: 3306 if engine is set as aurora or 5432 if set to aurora-postgresql.

", - "CreateDBInstanceMessage$AllocatedStorage": "

The amount of storage (in gibibytes) to allocate for the DB instance.

Type: Integer

Amazon Aurora

Not applicable. Aurora cluster volumes automatically grow as the amount of data in your database increases, though you are only charged for the space that you use in an Aurora cluster volume.

MySQL

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 5 to 3072.

MariaDB

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 5 to 3072.

PostgreSQL

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 5 to 3072.

Oracle

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 10 to 3072.

SQL Server

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2):

    • Enterprise and Standard editions: Must be an integer from 200 to 16384.

    • Web and Express editions: Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1):

    • Enterprise and Standard editions: Must be an integer from 200 to 16384.

    • Web and Express editions: Must be an integer from 100 to 16384.

  • Magnetic storage (standard):

    • Enterprise and Standard editions: Must be an integer from 200 to 1024.

    • Web and Express editions: Must be an integer from 20 to 1024.

", - "CreateDBInstanceMessage$BackupRetentionPeriod": "

The number of days for which automated backups are retained. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Amazon Aurora

Not applicable. The retention period for automated backups is managed by the DB cluster. For more information, see CreateDBCluster.

Default: 1

Constraints:

  • Must be a value from 0 to 35

  • Cannot be set to 0 if the DB instance is a source to Read Replicas

", - "CreateDBInstanceMessage$Port": "

The port number on which the database accepts connections.

MySQL

Default: 3306

Valid Values: 1150-65535

Type: Integer

MariaDB

Default: 3306

Valid Values: 1150-65535

Type: Integer

PostgreSQL

Default: 5432

Valid Values: 1150-65535

Type: Integer

Oracle

Default: 1521

Valid Values: 1150-65535

SQL Server

Default: 1433

Valid Values: 1150-65535 except for 1434, 3389, 47001, 49152, and 49152 through 49156.

Amazon Aurora

Default: 3306

Valid Values: 1150-65535

Type: Integer

", - "CreateDBInstanceMessage$Iops": "

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance. For information about valid Iops values, see see Amazon RDS Provisioned IOPS Storage to Improve Performance.

Constraints: Must be a multiple between 1 and 50 of the storage amount for the DB instance. Must also be an integer multiple of 1000. For example, if the size of your DB instance is 500 GiB, then your Iops value can be 2000, 3000, 4000, or 5000.

", - "CreateDBInstanceMessage$MonitoringInterval": "

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", - "CreateDBInstanceMessage$PromotionTier": "

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

Default: 1

Valid Values: 0 - 15

", - "CreateDBInstanceReadReplicaMessage$Port": "

The port number that the DB instance uses for connections.

Default: Inherits from the source DB instance

Valid Values: 1150-65535

", - "CreateDBInstanceReadReplicaMessage$Iops": "

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance.

", - "CreateDBInstanceReadReplicaMessage$MonitoringInterval": "

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the Read Replica. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", - "DBCluster$AllocatedStorage": "

For all database engines except Amazon Aurora, AllocatedStorage specifies the allocated storage size in gibibytes (GiB). For Aurora, AllocatedStorage always returns 1, because Aurora DB cluster storage size is not fixed, but instead automatically adjusts as needed.

", - "DBCluster$BackupRetentionPeriod": "

Specifies the number of days for which automatic DB snapshots are retained.

", - "DBCluster$Port": "

Specifies the port that the database engine is listening on.

", - "DBClusterMember$PromotionTier": "

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

", - "DBInstance$Iops": "

Specifies the Provisioned IOPS (I/O operations per second) value.

", - "DBInstance$MonitoringInterval": "

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance.

", - "DBInstance$PromotionTier": "

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

", - "DBSnapshot$Iops": "

Specifies the Provisioned IOPS (I/O operations per second) value of the DB instance at the time of the snapshot.

", - "DescribeCertificatesMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBClusterBacktracksMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBClusterParameterGroupsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBClusterParametersMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBClusterSnapshotsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBClustersMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBEngineVersionsMessage$MaxRecords": "

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBInstancesMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBLogFilesMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

", - "DescribeDBParameterGroupsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBParametersMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBSecurityGroupsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBSnapshotsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeDBSubnetGroupsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeEngineDefaultClusterParametersMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeEngineDefaultParametersMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeEventSubscriptionsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeEventsMessage$Duration": "

The number of minutes to retrieve events for.

Default: 60

", - "DescribeEventsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeOptionGroupOptionsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeOptionGroupsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeOrderableDBInstanceOptionsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribePendingMaintenanceActionsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeReservedDBInstancesMessage$MaxRecords": "

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeReservedDBInstancesOfferingsMessage$MaxRecords": "

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "DescribeSourceRegionsMessage$MaxRecords": "

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", - "ModifyDBClusterMessage$BackupRetentionPeriod": "

The number of days for which automated backups are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", - "ModifyDBClusterMessage$Port": "

The port number on which the DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", - "ModifyDBInstanceMessage$AllocatedStorage": "

The new amount of storage (in gibibytes) to allocate for the DB instance.

For MariaDB, MySQL, Oracle, and PostgreSQL, the value supplied must be at least 10% greater than the current value. Values that are not at least 10% greater than the existing value are rounded up so that they are 10% greater than the current value.

For the valid values for allocated storage for each engine, see CreateDBInstance.

", - "ModifyDBInstanceMessage$BackupRetentionPeriod": "

The number of days to retain automated backups. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Changing this parameter can result in an outage if you change from 0 to a non-zero value or from a non-zero value to 0. These changes are applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If you change the parameter from one non-zero value to another non-zero value, the change is asynchronously applied as soon as possible.

Amazon Aurora

Not applicable. The retention period for automated backups is managed by the DB cluster. For more information, see ModifyDBCluster.

Default: Uses existing setting

Constraints:

  • Must be a value from 0 to 35

  • Can be specified for a MySQL Read Replica only if the source is running MySQL 5.6

  • Can be specified for a PostgreSQL Read Replica only if the source is running PostgreSQL 9.3.5

  • Cannot be set to 0 if the DB instance is a source to Read Replicas

", - "ModifyDBInstanceMessage$Iops": "

The new Provisioned IOPS (I/O operations per second) value for the RDS instance.

Changing this setting doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If you are migrating from Provisioned IOPS to standard storage, set this value to 0. The DB instance will require a reboot for the change in storage type to take effect.

If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon RDS operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a Read Replica for the instance, and creating a DB snapshot of the instance.

Constraints: For MariaDB, MySQL, Oracle, and PostgreSQL, the value supplied must be at least 10% greater than the current value. Values that are not at least 10% greater than the existing value are rounded up so that they are 10% greater than the current value.

Default: Uses existing setting

", - "ModifyDBInstanceMessage$MonitoringInterval": "

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", - "ModifyDBInstanceMessage$DBPortNumber": "

The port number on which the database accepts connections.

The value of the DBPortNumber parameter must not match any of the port values specified for options in the option group for the DB instance.

Your database will restart when you change the DBPortNumber value regardless of the value of the ApplyImmediately parameter.

MySQL

Default: 3306

Valid Values: 1150-65535

MariaDB

Default: 3306

Valid Values: 1150-65535

PostgreSQL

Default: 5432

Valid Values: 1150-65535

Type: Integer

Oracle

Default: 1521

Valid Values: 1150-65535

SQL Server

Default: 1433

Valid Values: 1150-65535 except for 1434, 3389, 47001, 49152, and 49152 through 49156.

Amazon Aurora

Default: 3306

Valid Values: 1150-65535

", - "ModifyDBInstanceMessage$PromotionTier": "

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

Default: 1

Valid Values: 0 - 15

", - "Option$Port": "

If required, the port configured for this option to use.

", - "OptionConfiguration$Port": "

The optional port for the option.

", - "OptionGroupOption$DefaultPort": "

If the option requires a port, specifies the default port for the option.

", - "OrderableDBInstanceOption$MinStorageSize": "

Minimum storage size for a DB instance.

", - "OrderableDBInstanceOption$MaxStorageSize": "

Maximum storage size for a DB instance.

", - "OrderableDBInstanceOption$MinIopsPerDbInstance": "

Minimum total provisioned IOPS for a DB instance.

", - "OrderableDBInstanceOption$MaxIopsPerDbInstance": "

Maximum total provisioned IOPS for a DB instance.

", - "PendingModifiedValues$AllocatedStorage": "

Contains the new AllocatedStorage size for the DB instance that will be applied or is currently being applied.

", - "PendingModifiedValues$Port": "

Specifies the pending port for the DB instance.

", - "PendingModifiedValues$BackupRetentionPeriod": "

Specifies the pending number of days for which automated backups are retained.

", - "PendingModifiedValues$Iops": "

Specifies the new Provisioned IOPS value for the DB instance that will be applied or is currently being applied.

", - "PromoteReadReplicaMessage$BackupRetentionPeriod": "

The number of days to retain automated backups. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Default: 1

Constraints:

  • Must be a value from 0 to 8

", - "PurchaseReservedDBInstancesOfferingMessage$DBInstanceCount": "

The number of instances to reserve.

Default: 1

", - "Range$Step": "

The step value for the range. For example, if you have a range of 5,000 to 10,000, with a step value of 1,000, the valid values start at 5,000 and step up by 1,000. Even though 7,500 is within the range, it isn't a valid value for the range. The valid values are 5,000, 6,000, 7,000, 8,000...

", - "RestoreDBClusterFromS3Message$BackupRetentionPeriod": "

The number of days for which automated backups of the restored DB cluster are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", - "RestoreDBClusterFromS3Message$Port": "

The port number on which the instances in the restored DB cluster accept connections.

Default: 3306

", - "RestoreDBClusterFromSnapshotMessage$Port": "

The port number on which the new DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", - "RestoreDBClusterToPointInTimeMessage$Port": "

The port number on which the new DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", - "RestoreDBInstanceFromDBSnapshotMessage$Port": "

The port number on which the database accepts connections.

Default: The same port as the original DB instance

Constraints: Value must be 1150-65535

", - "RestoreDBInstanceFromDBSnapshotMessage$Iops": "

Specifies the amount of provisioned IOPS for the DB instance, expressed in I/O operations per second. If this parameter is not specified, the IOPS value is taken from the backup. If this parameter is set to 0, the new instance is converted to a non-PIOPS instance. The conversion takes additional time, though your DB instance is available for connections before the conversion starts.

The provisioned IOPS value must follow the requirements for your database engine. For more information, see Amazon RDS Provisioned IOPS Storage to Improve Performance.

Constraints: Must be an integer greater than 1000.

", - "RestoreDBInstanceFromS3Message$AllocatedStorage": "

The amount of storage (in gigabytes) to allocate initially for the DB instance. Follow the allocation rules specified in CreateDBInstance.

Be sure to allocate enough memory for your new DB instance so that the restore operation can succeed. You can also allocate additional memory for future growth.

", - "RestoreDBInstanceFromS3Message$BackupRetentionPeriod": "

The number of days for which automated backups are retained. Setting this parameter to a positive number enables backups. For more information, see CreateDBInstance.

", - "RestoreDBInstanceFromS3Message$Port": "

The port number on which the database accepts connections.

Type: Integer

Valid Values: 1150-65535

Default: 3306

", - "RestoreDBInstanceFromS3Message$Iops": "

The amount of Provisioned IOPS (input/output operations per second) to allocate initially for the DB instance. For information about valid Iops values, see see Amazon RDS Provisioned IOPS Storage to Improve Performance.

", - "RestoreDBInstanceFromS3Message$MonitoringInterval": "

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

Default: 0

", - "RestoreDBInstanceToPointInTimeMessage$Port": "

The port number on which the database accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB instance.

", - "RestoreDBInstanceToPointInTimeMessage$Iops": "

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance.

Constraints: Must be an integer greater than 1000.

SQL Server

Setting the IOPS value for the SQL Server database engine is not supported.

" - } - }, - "InvalidDBClusterSnapshotStateFault": { - "base": "

The supplied value isn't a valid DB cluster snapshot state.

", - "refs": { - } - }, - "InvalidDBClusterStateFault": { - "base": "

The DB cluster isn't in a valid state.

", - "refs": { - } - }, - "InvalidDBInstanceStateFault": { - "base": "

The specified DB instance isn't in the available state.

", - "refs": { - } - }, - "InvalidDBParameterGroupStateFault": { - "base": "

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you can't delete it when the parameter group is in this state.

", - "refs": { - } - }, - "InvalidDBSecurityGroupStateFault": { - "base": "

The state of the DB security group doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSnapshotStateFault": { - "base": "

The state of the DB snapshot doesn't allow deletion.

", - "refs": { - } - }, - "InvalidDBSubnetGroupFault": { - "base": "

The DBSubnetGroup doesn't belong to the same VPC as that of an existing cross-region read replica of the same source instance.

", - "refs": { - } - }, - "InvalidDBSubnetGroupStateFault": { - "base": "

The DB subnet group cannot be deleted because it's in use.

", - "refs": { - } - }, - "InvalidDBSubnetStateFault": { - "base": "

The DB subnet isn't in the available state.

", - "refs": { - } + "Filter":{ + "base":"

A filter name and value pair that is used to return a more specific list of results from a describe operation. Filters can be used to match a set of resources by specific criteria, such as IDs. The filters supported by a describe operation are documented with the describe operation.

Currently, wildcards are not supported in filters.

The following actions can be filtered:

", + "refs":{ + "FilterList$member":null + } + }, + "FilterList":{ + "base":null, + "refs":{ + "DescribeCertificatesMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClusterBacktracksMessage$Filters":"

A filter that specifies one or more DB clusters to describe. Supported filters include the following:

  • db-cluster-backtrack-id - Accepts backtrack identifiers. The results list includes information about only the backtracks identified by these identifiers.

  • db-cluster-backtrack-status - Accepts any of the following backtrack status values:

    • applying

    • completed

    • failed

    • pending

    The results list includes information about only the backtracks identified by these values. For more information about backtrack status values, see DBClusterBacktrack.

", + "DescribeDBClusterParameterGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClusterParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClusterSnapshotsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBClustersMessage$Filters":"

A filter that specifies one or more DB clusters to describe.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include information about the DB clusters identified by these ARNs.

", + "DescribeDBEngineVersionsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBInstancesMessage$Filters":"

A filter that specifies one or more DB instances to describe.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include information about the DB instances associated with the DB clusters identified by these ARNs.

  • db-instance-id - Accepts DB instance identifiers and DB instance Amazon Resource Names (ARNs). The results list will only include information about the DB instances identified by these ARNs.

", + "DescribeDBLogFilesMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBParameterGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBSecurityGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBSnapshotsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeDBSubnetGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEngineDefaultClusterParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEngineDefaultParametersMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEventCategoriesMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEventSubscriptionsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeEventsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeOptionGroupOptionsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeOptionGroupsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeOrderableDBInstanceOptionsMessage$Filters":"

This parameter is not currently supported.

", + "DescribePendingMaintenanceActionsMessage$Filters":"

A filter that specifies one or more resources to return pending maintenance actions for.

Supported filters:

  • db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon Resource Names (ARNs). The results list will only include pending maintenance actions for the DB clusters identified by these ARNs.

  • db-instance-id - Accepts DB instance identifiers and DB instance ARNs. The results list will only include pending maintenance actions for the DB instances identified by these ARNs.

", + "DescribeReservedDBInstancesMessage$Filters":"

This parameter is not currently supported.

", + "DescribeReservedDBInstancesOfferingsMessage$Filters":"

This parameter is not currently supported.

", + "DescribeSourceRegionsMessage$Filters":"

This parameter is not currently supported.

", + "ListTagsForResourceMessage$Filters":"

This parameter is not currently supported.

" + } + }, + "FilterValueList":{ + "base":null, + "refs":{ + "Filter$Values":"

One or more filter values. Filter values are case-sensitive.

" + } + }, + "IPRange":{ + "base":"

This data type is used as a response element in the DescribeDBSecurityGroups action.

", + "refs":{ + "IPRangeList$member":null + } + }, + "IPRangeList":{ + "base":null, + "refs":{ + "DBSecurityGroup$IPRanges":"

Contains a list of IPRange elements.

" + } + }, + "InstanceQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed number of DB instances.

", + "refs":{ + } + }, + "InsufficientDBClusterCapacityFault":{ + "base":"

The DB cluster doesn't have enough capacity for the current operation.

", + "refs":{ + } + }, + "InsufficientDBInstanceCapacityFault":{ + "base":"

The specified DB instance class isn't available in the specified Availability Zone.

", + "refs":{ + } + }, + "InsufficientStorageClusterCapacityFault":{ + "base":"

There is insufficient storage available for the current action. You might be able to resolve this error by updating your subnet group to use different Availability Zones that have more storage available.

", + "refs":{ + } + }, + "Integer":{ + "base":null, + "refs":{ + "DBClusterSnapshot$AllocatedStorage":"

Specifies the allocated storage size in gibibytes (GiB).

", + "DBClusterSnapshot$Port":"

Specifies the port that the DB cluster was listening on at the time of the snapshot.

", + "DBClusterSnapshot$PercentProgress":"

Specifies the percentage of the estimated data that has been transferred.

", + "DBInstance$AllocatedStorage":"

Specifies the allocated storage size specified in gibibytes.

", + "DBInstance$BackupRetentionPeriod":"

Specifies the number of days for which automatic DB snapshots are retained.

", + "DBInstance$DbInstancePort":"

Specifies the port that the DB instance listens on. If the DB instance is part of a DB cluster, this can be a different port than the DB cluster port.

", + "DBSnapshot$AllocatedStorage":"

Specifies the allocated storage size in gibibytes (GiB).

", + "DBSnapshot$Port":"

Specifies the port that the database engine was listening on at the time of the snapshot.

", + "DBSnapshot$PercentProgress":"

The percentage of the estimated data that has been transferred.

", + "DownloadDBLogFilePortionMessage$NumberOfLines":"

The number of lines to download. If the number of lines specified results in a file over 1 MB in size, the file is truncated at 1 MB in size.

If the NumberOfLines parameter is specified, then the block of lines returned can be from the beginning or the end of the log file, depending on the value of the Marker parameter.

  • If neither Marker or NumberOfLines are specified, the entire log file is returned up to a maximum of 10000 lines, starting with the most recent log entries first.

  • If NumberOfLines is specified and Marker is not specified, then the most recent lines from the end of the log file are returned.

  • If Marker is specified as \"0\", then the specified number of lines from the beginning of the log file are returned.

  • You can download the log file in blocks of lines by specifying the size of the block using the NumberOfLines parameter, and by specifying a value of \"0\" for the Marker parameter in your first request. Include the Marker value returned in the response as the Marker value for the next request, continuing until the AdditionalDataPending response element returns false.

", + "Endpoint$Port":"

Specifies the port that the database engine is listening on.

", + "Range$From":"

The minimum value in the range.

", + "Range$To":"

The maximum value in the range.

", + "ReservedDBInstance$Duration":"

The duration of the reservation in seconds.

", + "ReservedDBInstance$DBInstanceCount":"

The number of reserved DB instances.

", + "ReservedDBInstancesOffering$Duration":"

The duration of the offering in seconds.

" + } + }, + "IntegerOptional":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$BackupRetentionPeriod":"

The number of days for which automated backups are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", + "CreateDBClusterMessage$Port":"

The port number on which the instances in the DB cluster accept connections.

Default: 3306 if engine is set as aurora or 5432 if set to aurora-postgresql.

", + "CreateDBInstanceMessage$AllocatedStorage":"

The amount of storage (in gibibytes) to allocate for the DB instance.

Type: Integer

Amazon Aurora

Not applicable. Aurora cluster volumes automatically grow as the amount of data in your database increases, though you are only charged for the space that you use in an Aurora cluster volume.

MySQL

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 5 to 3072.

MariaDB

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 5 to 3072.

PostgreSQL

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 5 to 3072.

Oracle

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2): Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1): Must be an integer from 100 to 16384.

  • Magnetic storage (standard): Must be an integer from 10 to 3072.

SQL Server

Constraints to the amount of storage for each storage type are the following:

  • General Purpose (SSD) storage (gp2):

    • Enterprise and Standard editions: Must be an integer from 200 to 16384.

    • Web and Express editions: Must be an integer from 20 to 16384.

  • Provisioned IOPS storage (io1):

    • Enterprise and Standard editions: Must be an integer from 200 to 16384.

    • Web and Express editions: Must be an integer from 100 to 16384.

  • Magnetic storage (standard):

    • Enterprise and Standard editions: Must be an integer from 200 to 1024.

    • Web and Express editions: Must be an integer from 20 to 1024.

", + "CreateDBInstanceMessage$BackupRetentionPeriod":"

The number of days for which automated backups are retained. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Amazon Aurora

Not applicable. The retention period for automated backups is managed by the DB cluster. For more information, see CreateDBCluster.

Default: 1

Constraints:

  • Must be a value from 0 to 35

  • Cannot be set to 0 if the DB instance is a source to Read Replicas

", + "CreateDBInstanceMessage$Port":"

The port number on which the database accepts connections.

MySQL

Default: 3306

Valid Values: 1150-65535

Type: Integer

MariaDB

Default: 3306

Valid Values: 1150-65535

Type: Integer

PostgreSQL

Default: 5432

Valid Values: 1150-65535

Type: Integer

Oracle

Default: 1521

Valid Values: 1150-65535

SQL Server

Default: 1433

Valid Values: 1150-65535 except for 1434, 3389, 47001, 49152, and 49152 through 49156.

Amazon Aurora

Default: 3306

Valid Values: 1150-65535

Type: Integer

", + "CreateDBInstanceMessage$Iops":"

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance. For information about valid Iops values, see see Amazon RDS Provisioned IOPS Storage to Improve Performance.

Constraints: Must be a multiple between 1 and 50 of the storage amount for the DB instance. Must also be an integer multiple of 1000. For example, if the size of your DB instance is 500 GiB, then your Iops value can be 2000, 3000, 4000, or 5000.

", + "CreateDBInstanceMessage$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", + "CreateDBInstanceMessage$PromotionTier":"

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

Default: 1

Valid Values: 0 - 15

", + "CreateDBInstanceMessage$PerformanceInsightsRetentionPeriod":"

The amount of time, in days, to retain Performance Insights data. Valid values are 7 or 731 (2 years).

", + "CreateDBInstanceReadReplicaMessage$Port":"

The port number that the DB instance uses for connections.

Default: Inherits from the source DB instance

Valid Values: 1150-65535

", + "CreateDBInstanceReadReplicaMessage$Iops":"

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance.

", + "CreateDBInstanceReadReplicaMessage$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the Read Replica. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", + "CreateDBInstanceReadReplicaMessage$PerformanceInsightsRetentionPeriod":"

The amount of time, in days, to retain Performance Insights data. Valid values are 7 or 731 (2 years).

", + "DBCluster$AllocatedStorage":"

For all database engines except Amazon Aurora, AllocatedStorage specifies the allocated storage size in gibibytes (GiB). For Aurora, AllocatedStorage always returns 1, because Aurora DB cluster storage size is not fixed, but instead automatically adjusts as needed.

", + "DBCluster$BackupRetentionPeriod":"

Specifies the number of days for which automatic DB snapshots are retained.

", + "DBCluster$Port":"

Specifies the port that the database engine is listening on.

", + "DBClusterMember$PromotionTier":"

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

", + "DBInstance$Iops":"

Specifies the Provisioned IOPS (I/O operations per second) value.

", + "DBInstance$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance.

", + "DBInstance$PromotionTier":"

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

", + "DBInstance$PerformanceInsightsRetentionPeriod":"

The amount of time, in days, to retain Performance Insights data. Valid values are 7 or 731 (2 years).

", + "DBSnapshot$Iops":"

Specifies the Provisioned IOPS (I/O operations per second) value of the DB instance at the time of the snapshot.

", + "DescribeCertificatesMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClusterBacktracksMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClusterParameterGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClusterParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClusterSnapshotsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBClustersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBEngineVersionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBInstancesMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBLogFilesMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

", + "DescribeDBParameterGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBSecurityGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBSnapshotsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeDBSubnetGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEngineDefaultClusterParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEngineDefaultParametersMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEventSubscriptionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeEventsMessage$Duration":"

The number of minutes to retrieve events for.

Default: 60

", + "DescribeEventsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeOptionGroupOptionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeOptionGroupsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeOrderableDBInstanceOptionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribePendingMaintenanceActionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeReservedDBInstancesMessage$MaxRecords":"

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeReservedDBInstancesOfferingsMessage$MaxRecords":"

The maximum number of records to include in the response. If more than the MaxRecords value is available, a pagination token called a marker is included in the response so that the following results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "DescribeSourceRegionsMessage$MaxRecords":"

The maximum number of records to include in the response. If more records exist than the specified MaxRecords value, a pagination token called a marker is included in the response so that the remaining results can be retrieved.

Default: 100

Constraints: Minimum 20, maximum 100.

", + "ModifyDBClusterMessage$BackupRetentionPeriod":"

The number of days for which automated backups are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", + "ModifyDBClusterMessage$Port":"

The port number on which the DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", + "ModifyDBInstanceMessage$AllocatedStorage":"

The new amount of storage (in gibibytes) to allocate for the DB instance.

For MariaDB, MySQL, Oracle, and PostgreSQL, the value supplied must be at least 10% greater than the current value. Values that are not at least 10% greater than the existing value are rounded up so that they are 10% greater than the current value.

For the valid values for allocated storage for each engine, see CreateDBInstance.

", + "ModifyDBInstanceMessage$BackupRetentionPeriod":"

The number of days to retain automated backups. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Changing this parameter can result in an outage if you change from 0 to a non-zero value or from a non-zero value to 0. These changes are applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If you change the parameter from one non-zero value to another non-zero value, the change is asynchronously applied as soon as possible.

Amazon Aurora

Not applicable. The retention period for automated backups is managed by the DB cluster. For more information, see ModifyDBCluster.

Default: Uses existing setting

Constraints:

  • Must be a value from 0 to 35

  • Can be specified for a MySQL Read Replica only if the source is running MySQL 5.6

  • Can be specified for a PostgreSQL Read Replica only if the source is running PostgreSQL 9.3.5

  • Cannot be set to 0 if the DB instance is a source to Read Replicas

", + "ModifyDBInstanceMessage$Iops":"

The new Provisioned IOPS (I/O operations per second) value for the RDS instance.

Changing this setting doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If you are migrating from Provisioned IOPS to standard storage, set this value to 0. The DB instance will require a reboot for the change in storage type to take effect.

If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon RDS operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a Read Replica for the instance, and creating a DB snapshot of the instance.

Constraints: For MariaDB, MySQL, Oracle, and PostgreSQL, the value supplied must be at least 10% greater than the current value. Values that are not at least 10% greater than the existing value are rounded up so that they are 10% greater than the current value.

Default: Uses existing setting

", + "ModifyDBInstanceMessage$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0. The default is 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

", + "ModifyDBInstanceMessage$DBPortNumber":"

The port number on which the database accepts connections.

The value of the DBPortNumber parameter must not match any of the port values specified for options in the option group for the DB instance.

Your database will restart when you change the DBPortNumber value regardless of the value of the ApplyImmediately parameter.

MySQL

Default: 3306

Valid Values: 1150-65535

MariaDB

Default: 3306

Valid Values: 1150-65535

PostgreSQL

Default: 5432

Valid Values: 1150-65535

Type: Integer

Oracle

Default: 1521

Valid Values: 1150-65535

SQL Server

Default: 1433

Valid Values: 1150-65535 except for 1434, 3389, 47001, 49152, and 49152 through 49156.

Amazon Aurora

Default: 3306

Valid Values: 1150-65535

", + "ModifyDBInstanceMessage$PromotionTier":"

A value that specifies the order in which an Aurora Replica is promoted to the primary instance after a failure of the existing primary instance. For more information, see Fault Tolerance for an Aurora DB Cluster.

Default: 1

Valid Values: 0 - 15

", + "ModifyDBInstanceMessage$PerformanceInsightsRetentionPeriod":"

The amount of time, in days, to retain Performance Insights data. Valid values are 7 or 731 (2 years).

", + "Option$Port":"

If required, the port configured for this option to use.

", + "OptionConfiguration$Port":"

The optional port for the option.

", + "OptionGroupOption$DefaultPort":"

If the option requires a port, specifies the default port for the option.

", + "OrderableDBInstanceOption$MinStorageSize":"

Minimum storage size for a DB instance.

", + "OrderableDBInstanceOption$MaxStorageSize":"

Maximum storage size for a DB instance.

", + "OrderableDBInstanceOption$MinIopsPerDbInstance":"

Minimum total provisioned IOPS for a DB instance.

", + "OrderableDBInstanceOption$MaxIopsPerDbInstance":"

Maximum total provisioned IOPS for a DB instance.

", + "PendingModifiedValues$AllocatedStorage":"

Contains the new AllocatedStorage size for the DB instance that will be applied or is currently being applied.

", + "PendingModifiedValues$Port":"

Specifies the pending port for the DB instance.

", + "PendingModifiedValues$BackupRetentionPeriod":"

Specifies the pending number of days for which automated backups are retained.

", + "PendingModifiedValues$Iops":"

Specifies the new Provisioned IOPS value for the DB instance that will be applied or is currently being applied.

", + "PromoteReadReplicaMessage$BackupRetentionPeriod":"

The number of days to retain automated backups. Setting this parameter to a positive number enables backups. Setting this parameter to 0 disables automated backups.

Default: 1

Constraints:

  • Must be a value from 0 to 8

", + "PurchaseReservedDBInstancesOfferingMessage$DBInstanceCount":"

The number of instances to reserve.

Default: 1

", + "Range$Step":"

The step value for the range. For example, if you have a range of 5,000 to 10,000, with a step value of 1,000, the valid values start at 5,000 and step up by 1,000. Even though 7,500 is within the range, it isn't a valid value for the range. The valid values are 5,000, 6,000, 7,000, 8,000...

", + "RestoreDBClusterFromS3Message$BackupRetentionPeriod":"

The number of days for which automated backups of the restored DB cluster are retained. You must specify a minimum value of 1.

Default: 1

Constraints:

  • Must be a value from 1 to 35

", + "RestoreDBClusterFromS3Message$Port":"

The port number on which the instances in the restored DB cluster accept connections.

Default: 3306

", + "RestoreDBClusterFromSnapshotMessage$Port":"

The port number on which the new DB cluster accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$Port":"

The port number on which the new DB cluster accepts connections.

Constraints: A value from 1150-65535.

Default: The default port for the engine.

", + "RestoreDBInstanceFromDBSnapshotMessage$Port":"

The port number on which the database accepts connections.

Default: The same port as the original DB instance

Constraints: Value must be 1150-65535

", + "RestoreDBInstanceFromDBSnapshotMessage$Iops":"

Specifies the amount of provisioned IOPS for the DB instance, expressed in I/O operations per second. If this parameter is not specified, the IOPS value is taken from the backup. If this parameter is set to 0, the new instance is converted to a non-PIOPS instance. The conversion takes additional time, though your DB instance is available for connections before the conversion starts.

The provisioned IOPS value must follow the requirements for your database engine. For more information, see Amazon RDS Provisioned IOPS Storage to Improve Performance.

Constraints: Must be an integer greater than 1000.

", + "RestoreDBInstanceFromS3Message$AllocatedStorage":"

The amount of storage (in gigabytes) to allocate initially for the DB instance. Follow the allocation rules specified in CreateDBInstance.

Be sure to allocate enough memory for your new DB instance so that the restore operation can succeed. You can also allocate additional memory for future growth.

", + "RestoreDBInstanceFromS3Message$BackupRetentionPeriod":"

The number of days for which automated backups are retained. Setting this parameter to a positive number enables backups. For more information, see CreateDBInstance.

", + "RestoreDBInstanceFromS3Message$Port":"

The port number on which the database accepts connections.

Type: Integer

Valid Values: 1150-65535

Default: 3306

", + "RestoreDBInstanceFromS3Message$Iops":"

The amount of Provisioned IOPS (input/output operations per second) to allocate initially for the DB instance. For information about valid Iops values, see see Amazon RDS Provisioned IOPS Storage to Improve Performance.

", + "RestoreDBInstanceFromS3Message$MonitoringInterval":"

The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB instance. To disable collecting Enhanced Monitoring metrics, specify 0.

If MonitoringRoleArn is specified, then you must also set MonitoringInterval to a value other than 0.

Valid Values: 0, 1, 5, 10, 15, 30, 60

Default: 0

", + "RestoreDBInstanceFromS3Message$PerformanceInsightsRetentionPeriod":"

The amount of time, in days, to retain Performance Insights data. Valid values are 7 or 731 (2 years).

", + "RestoreDBInstanceToPointInTimeMessage$Port":"

The port number on which the database accepts connections.

Constraints: Value must be 1150-65535

Default: The same port as the original DB instance.

", + "RestoreDBInstanceToPointInTimeMessage$Iops":"

The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for the DB instance.

Constraints: Must be an integer greater than 1000.

SQL Server

Setting the IOPS value for the SQL Server database engine is not supported.

" + } + }, + "InvalidDBClusterSnapshotStateFault":{ + "base":"

The supplied value isn't a valid DB cluster snapshot state.

", + "refs":{ + } + }, + "InvalidDBClusterStateFault":{ + "base":"

The DB cluster isn't in a valid state.

", + "refs":{ + } + }, + "InvalidDBInstanceStateFault":{ + "base":"

The specified DB instance isn't in the available state.

", + "refs":{ + } + }, + "InvalidDBParameterGroupStateFault":{ + "base":"

The DB parameter group is in use or is in an invalid state. If you are attempting to delete the parameter group, you can't delete it when the parameter group is in this state.

", + "refs":{ + } + }, + "InvalidDBSecurityGroupStateFault":{ + "base":"

The state of the DB security group doesn't allow deletion.

", + "refs":{ + } + }, + "InvalidDBSnapshotStateFault":{ + "base":"

The state of the DB snapshot doesn't allow deletion.

", + "refs":{ + } + }, + "InvalidDBSubnetGroupFault":{ + "base":"

The DBSubnetGroup doesn't belong to the same VPC as that of an existing cross-region read replica of the same source instance.

", + "refs":{ + } + }, + "InvalidDBSubnetGroupStateFault":{ + "base":"

The DB subnet group cannot be deleted because it's in use.

", + "refs":{ + } + }, + "InvalidDBSubnetStateFault":{ + "base":"

The DB subnet isn't in the available state.

", + "refs":{ + } + }, + "InvalidEventSubscriptionStateFault":{ + "base":"

This error can occur if someone else is modifying a subscription. You should retry the action.

", + "refs":{ + } + }, + "InvalidOptionGroupStateFault":{ + "base":"

The option group isn't in the available state.

", + "refs":{ + } }, - "InvalidEventSubscriptionStateFault": { - "base": "

This error can occur if someone else is modifying a subscription. You should retry the action.

", - "refs": { + "InvalidRestoreFault":{ + "base":"

Cannot restore from VPC backup to non-VPC DB instance.

", + "refs":{ } }, - "InvalidOptionGroupStateFault": { - "base": "

The option group isn't in the available state.

", - "refs": { + "InvalidS3BucketFault":{ + "base":"

The specified Amazon S3 bucket name can't be found or Amazon RDS isn't authorized to access the specified Amazon S3 bucket. Verify the SourceS3BucketName and S3IngestionRoleArn values and try again.

", + "refs":{ } }, - "InvalidRestoreFault": { - "base": "

Cannot restore from VPC backup to non-VPC DB instance.

", - "refs": { + "InvalidSubnet":{ + "base":"

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", + "refs":{ } }, - "InvalidS3BucketFault": { - "base": "

The specified Amazon S3 bucket name can't be found or Amazon RDS isn't authorized to access the specified Amazon S3 bucket. Verify the SourceS3BucketName and S3IngestionRoleArn values and try again.

", - "refs": { + "InvalidVPCNetworkStateFault":{ + "base":"

The DB subnet group doesn't cover all Availability Zones after it's created because of users' change.

", + "refs":{ } }, - "InvalidSubnet": { - "base": "

The requested subnet is invalid, or multiple subnets were requested that are not all in a common VPC.

", - "refs": { + "KMSKeyNotAccessibleFault":{ + "base":"

An error occurred accessing an AWS KMS key.

", + "refs":{ } }, - "InvalidVPCNetworkStateFault": { - "base": "

The DB subnet group doesn't cover all Availability Zones after it's created because of users' change.

", - "refs": { + "KeyList":{ + "base":null, + "refs":{ + "RemoveTagsFromResourceMessage$TagKeys":"

The tag key (name) of the tag to be removed.

" } }, - "KMSKeyNotAccessibleFault": { - "base": "

An error occurred accessing an AWS KMS key.

", - "refs": { + "ListTagsForResourceMessage":{ + "base":"

", + "refs":{ } }, - "KeyList": { - "base": null, - "refs": { - "RemoveTagsFromResourceMessage$TagKeys": "

The tag key (name) of the tag to be removed.

" + "LogTypeList":{ + "base":null, + "refs":{ + "CloudwatchLogsExportConfiguration$EnableLogTypes":"

The list of log types to enable.

", + "CloudwatchLogsExportConfiguration$DisableLogTypes":"

The list of log types to disable.

", + "CreateDBClusterMessage$EnableCloudwatchLogsExports":"

The list of log types that need to be enabled for exporting to CloudWatch Logs.

", + "CreateDBInstanceMessage$EnableCloudwatchLogsExports":"

The list of log types that need to be enabled for exporting to CloudWatch Logs.

", + "CreateDBInstanceReadReplicaMessage$EnableCloudwatchLogsExports":"

The list of logs that the new DB instance is to export to CloudWatch Logs.

", + "DBCluster$EnabledCloudwatchLogsExports":"

A list of log types that this DB cluster is configured to export to CloudWatch Logs.

", + "DBEngineVersion$ExportableLogTypes":"

The types of logs that the database engine has available for export to CloudWatch Logs.

", + "DBInstance$EnabledCloudwatchLogsExports":"

A list of log types that this DB instance is configured to export to CloudWatch Logs.

", + "PendingCloudwatchLogsExports$LogTypesToEnable":"

Log types that are in the process of being deactivated. After they are deactivated, these log types aren't exported to CloudWatch Logs.

", + "PendingCloudwatchLogsExports$LogTypesToDisable":"

Log types that are in the process of being enabled. After they are enabled, these log types are exported to CloudWatch Logs.

", + "RestoreDBClusterFromS3Message$EnableCloudwatchLogsExports":"

The list of logs that the restored DB cluster is to export to CloudWatch Logs.

", + "RestoreDBClusterFromSnapshotMessage$EnableCloudwatchLogsExports":"

The list of logs that the restored DB cluster is to export to CloudWatch Logs.

", + "RestoreDBClusterToPointInTimeMessage$EnableCloudwatchLogsExports":"

The list of logs that the restored DB cluster is to export to CloudWatch Logs.

", + "RestoreDBInstanceFromDBSnapshotMessage$EnableCloudwatchLogsExports":"

The list of logs that the restored DB instance is to export to CloudWatch Logs.

", + "RestoreDBInstanceFromS3Message$EnableCloudwatchLogsExports":"

The list of logs that the restored DB instance is to export to CloudWatch Logs.

", + "RestoreDBInstanceToPointInTimeMessage$EnableCloudwatchLogsExports":"

The list of logs that the restored DB instance is to export to CloudWatch Logs.

" } }, - "ListTagsForResourceMessage": { - "base": "

", - "refs": { + "Long":{ + "base":null, + "refs":{ + "AccountQuota$Used":"

The amount currently used toward the quota maximum.

", + "AccountQuota$Max":"

The maximum allowed value for the quota.

", + "DescribeDBLogFilesDetails$LastWritten":"

A POSIX timestamp when the last log entry was written.

", + "DescribeDBLogFilesDetails$Size":"

The size, in bytes, of the log file for the specified DB instance.

", + "DescribeDBLogFilesMessage$FileLastWritten":"

Filters the available log files for files written since the specified date, in POSIX timestamp format with milliseconds.

", + "DescribeDBLogFilesMessage$FileSize":"

Filters the available log files for files larger than the specified size.

" } }, - "LogTypeList": { - "base": null, - "refs": { - "CloudwatchLogsExportConfiguration$EnableLogTypes": "

The list of log types to enable.

", - "CloudwatchLogsExportConfiguration$DisableLogTypes": "

The list of log types to disable.

", - "CreateDBInstanceMessage$EnableCloudwatchLogsExports": "

The list of log types that need to be enabled for exporting to CloudWatch Logs.

", - "CreateDBInstanceReadReplicaMessage$EnableCloudwatchLogsExports": "

The list of logs that the new DB instance is to export to CloudWatch Logs.

", - "DBEngineVersion$ExportableLogTypes": "

The types of logs that the database engine has available for export to CloudWatch Logs.

", - "DBInstance$EnabledCloudwatchLogsExports": "

A list of log types that this DB instance is configured to export to CloudWatch Logs.

", - "PendingCloudwatchLogsExports$LogTypesToEnable": "

Log types that are in the process of being deactivated. After they are deactivated, these log types aren't exported to CloudWatch Logs.

", - "PendingCloudwatchLogsExports$LogTypesToDisable": "

Log types that are in the process of being enabled. After they are enabled, these log types are exported to CloudWatch Logs.

", - "RestoreDBInstanceFromDBSnapshotMessage$EnableCloudwatchLogsExports": "

The list of logs that the restored DB instance is to export to CloudWatch Logs.

", - "RestoreDBInstanceFromS3Message$EnableCloudwatchLogsExports": "

The list of logs that the restored DB instance is to export to CloudWatch Logs.

", - "RestoreDBInstanceToPointInTimeMessage$EnableCloudwatchLogsExports": "

The list of logs that the restored DB instance is to export to CloudWatch Logs.

" + "LongOptional":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$BacktrackWindow":"

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", + "DBCluster$BacktrackWindow":"

The target backtrack window, in seconds. If this value is set to 0, backtracking is disabled for the DB cluster. Otherwise, backtracking is enabled.

", + "DBCluster$BacktrackConsumedChangeRecords":"

The number of change records stored for Backtrack.

", + "ModifyDBClusterMessage$BacktrackWindow":"

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", + "RestoreDBClusterFromS3Message$BacktrackWindow":"

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", + "RestoreDBClusterFromSnapshotMessage$BacktrackWindow":"

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", + "RestoreDBClusterToPointInTimeMessage$BacktrackWindow":"

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

" } }, - "Long": { - "base": null, - "refs": { - "AccountQuota$Used": "

The amount currently used toward the quota maximum.

", - "AccountQuota$Max": "

The maximum allowed value for the quota.

", - "DescribeDBLogFilesDetails$LastWritten": "

A POSIX timestamp when the last log entry was written.

", - "DescribeDBLogFilesDetails$Size": "

The size, in bytes, of the log file for the specified DB instance.

", - "DescribeDBLogFilesMessage$FileLastWritten": "

Filters the available log files for files written since the specified date, in POSIX timestamp format with milliseconds.

", - "DescribeDBLogFilesMessage$FileSize": "

Filters the available log files for files larger than the specified size.

" + "ModifyDBClusterMessage":{ + "base":"

", + "refs":{ } }, - "LongOptional": { - "base": null, - "refs": { - "CreateDBClusterMessage$BacktrackWindow": "

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", - "DBCluster$BacktrackWindow": "

The target backtrack window, in seconds. If this value is set to 0, backtracking is disabled for the DB cluster. Otherwise, backtracking is enabled.

", - "DBCluster$BacktrackConsumedChangeRecords": "

The number of change records stored for Backtrack.

", - "ModifyDBClusterMessage$BacktrackWindow": "

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", - "RestoreDBClusterFromS3Message$BacktrackWindow": "

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", - "RestoreDBClusterFromSnapshotMessage$BacktrackWindow": "

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

", - "RestoreDBClusterToPointInTimeMessage$BacktrackWindow": "

The target backtrack window, in seconds. To disable backtracking, set this value to 0.

Default: 0

Constraints:

  • If specified, this value must be set to a number from 0 to 259,200 (72 hours).

" + "ModifyDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBClusterMessage": { - "base": "

", - "refs": { + "ModifyDBClusterResult":{ + "base":null, + "refs":{ } }, - "ModifyDBClusterParameterGroupMessage": { - "base": "

", - "refs": { + "ModifyDBClusterSnapshotAttributeMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBClusterResult": { - "base": null, - "refs": { + "ModifyDBClusterSnapshotAttributeResult":{ + "base":null, + "refs":{ } }, - "ModifyDBClusterSnapshotAttributeMessage": { - "base": "

", - "refs": { + "ModifyDBInstanceMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBClusterSnapshotAttributeResult": { - "base": null, - "refs": { + "ModifyDBInstanceResult":{ + "base":null, + "refs":{ } }, - "ModifyDBInstanceMessage": { - "base": "

", - "refs": { + "ModifyDBParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBInstanceResult": { - "base": null, - "refs": { + "ModifyDBSnapshotAttributeMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBParameterGroupMessage": { - "base": "

", - "refs": { + "ModifyDBSnapshotAttributeResult":{ + "base":null, + "refs":{ } }, - "ModifyDBSnapshotAttributeMessage": { - "base": "

", - "refs": { + "ModifyDBSnapshotMessage":{ + "base":null, + "refs":{ } }, - "ModifyDBSnapshotAttributeResult": { - "base": null, - "refs": { + "ModifyDBSnapshotResult":{ + "base":null, + "refs":{ } }, - "ModifyDBSnapshotMessage": { - "base": null, - "refs": { + "ModifyDBSubnetGroupMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBSnapshotResult": { - "base": null, - "refs": { + "ModifyDBSubnetGroupResult":{ + "base":null, + "refs":{ } }, - "ModifyDBSubnetGroupMessage": { - "base": "

", - "refs": { + "ModifyEventSubscriptionMessage":{ + "base":"

", + "refs":{ } }, - "ModifyDBSubnetGroupResult": { - "base": null, - "refs": { + "ModifyEventSubscriptionResult":{ + "base":null, + "refs":{ } }, - "ModifyEventSubscriptionMessage": { - "base": "

", - "refs": { + "ModifyOptionGroupMessage":{ + "base":"

", + "refs":{ } }, - "ModifyEventSubscriptionResult": { - "base": null, - "refs": { + "ModifyOptionGroupResult":{ + "base":null, + "refs":{ } }, - "ModifyOptionGroupMessage": { - "base": "

", - "refs": { + "Option":{ + "base":"

Option details.

", + "refs":{ + "OptionsList$member":null } }, - "ModifyOptionGroupResult": { - "base": null, - "refs": { + "OptionConfiguration":{ + "base":"

A list of all available options

", + "refs":{ + "OptionConfigurationList$member":null } }, - "Option": { - "base": "

Option details.

", - "refs": { - "OptionsList$member": null + "OptionConfigurationList":{ + "base":null, + "refs":{ + "ModifyOptionGroupMessage$OptionsToInclude":"

Options in this list are added to the option group or, if already present, the specified configuration is used to update the existing configuration.

" } }, - "OptionConfiguration": { - "base": "

A list of all available options

", - "refs": { - "OptionConfigurationList$member": null + "OptionGroup":{ + "base":"

", + "refs":{ + "CopyOptionGroupResult$OptionGroup":null, + "CreateOptionGroupResult$OptionGroup":null, + "ModifyOptionGroupResult$OptionGroup":null, + "OptionGroupsList$member":null } }, - "OptionConfigurationList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToInclude": "

Options in this list are added to the option group or, if already present, the specified configuration is used to update the existing configuration.

" + "OptionGroupAlreadyExistsFault":{ + "base":"

The option group you are trying to create already exists.

", + "refs":{ } }, - "OptionGroup": { - "base": "

", - "refs": { - "CopyOptionGroupResult$OptionGroup": null, - "CreateOptionGroupResult$OptionGroup": null, - "ModifyOptionGroupResult$OptionGroup": null, - "OptionGroupsList$member": null + "OptionGroupMembership":{ + "base":"

Provides information on the option groups the DB instance is a member of.

", + "refs":{ + "OptionGroupMembershipList$member":null } }, - "OptionGroupAlreadyExistsFault": { - "base": "

The option group you are trying to create already exists.

", - "refs": { + "OptionGroupMembershipList":{ + "base":null, + "refs":{ + "DBInstance$OptionGroupMemberships":"

Provides the list of option group memberships for this DB instance.

" } }, - "OptionGroupMembership": { - "base": "

Provides information on the option groups the DB instance is a member of.

", - "refs": { - "OptionGroupMembershipList$member": null + "OptionGroupNotFoundFault":{ + "base":"

The specified option group could not be found.

", + "refs":{ } }, - "OptionGroupMembershipList": { - "base": null, - "refs": { - "DBInstance$OptionGroupMemberships": "

Provides the list of option group memberships for this DB instance.

" + "OptionGroupOption":{ + "base":"

Available option.

", + "refs":{ + "OptionGroupOptionsList$member":null } }, - "OptionGroupNotFoundFault": { - "base": "

The specified option group could not be found.

", - "refs": { + "OptionGroupOptionSetting":{ + "base":"

Option group option settings are used to display settings available for each option with their default values and other information. These values are used with the DescribeOptionGroupOptions action.

", + "refs":{ + "OptionGroupOptionSettingsList$member":null } }, - "OptionGroupOption": { - "base": "

Available option.

", - "refs": { - "OptionGroupOptionsList$member": null + "OptionGroupOptionSettingsList":{ + "base":null, + "refs":{ + "OptionGroupOption$OptionGroupOptionSettings":"

The option settings that are available (and the default value) for each option in an option group.

" } }, - "OptionGroupOptionSetting": { - "base": "

Option group option settings are used to display settings available for each option with their default values and other information. These values are used with the DescribeOptionGroupOptions action.

", - "refs": { - "OptionGroupOptionSettingsList$member": null + "OptionGroupOptionVersionsList":{ + "base":null, + "refs":{ + "OptionGroupOption$OptionGroupOptionVersions":"

The versions that are available for the option.

" } }, - "OptionGroupOptionSettingsList": { - "base": null, - "refs": { - "OptionGroupOption$OptionGroupOptionSettings": "

The option settings that are available (and the default value) for each option in an option group.

" + "OptionGroupOptionsList":{ + "base":"

List of available option group options.

", + "refs":{ + "OptionGroupOptionsMessage$OptionGroupOptions":null } }, - "OptionGroupOptionVersionsList": { - "base": null, - "refs": { - "OptionGroupOption$OptionGroupOptionVersions": "

The versions that are available for the option.

" + "OptionGroupOptionsMessage":{ + "base":"

", + "refs":{ } }, - "OptionGroupOptionsList": { - "base": "

List of available option group options.

", - "refs": { - "OptionGroupOptionsMessage$OptionGroupOptions": null + "OptionGroupQuotaExceededFault":{ + "base":"

The quota of 20 option groups was exceeded for this AWS account.

", + "refs":{ } }, - "OptionGroupOptionsMessage": { - "base": "

", - "refs": { + "OptionGroups":{ + "base":"

List of option groups.

", + "refs":{ } }, - "OptionGroupQuotaExceededFault": { - "base": "

The quota of 20 option groups was exceeded for this AWS account.

", - "refs": { + "OptionGroupsList":{ + "base":null, + "refs":{ + "OptionGroups$OptionGroupsList":"

List of option groups.

" } }, - "OptionGroups": { - "base": "

List of option groups.

", - "refs": { + "OptionNamesList":{ + "base":null, + "refs":{ + "ModifyOptionGroupMessage$OptionsToRemove":"

Options in this list are removed from the option group.

" } }, - "OptionGroupsList": { - "base": null, - "refs": { - "OptionGroups$OptionGroupsList": "

List of option groups.

" + "OptionSetting":{ + "base":"

Option settings are the actual settings being applied or configured for that option. It is used when you modify an option group or describe option groups. For example, the NATIVE_NETWORK_ENCRYPTION option has a setting called SQLNET.ENCRYPTION_SERVER that can have several different values.

", + "refs":{ + "OptionSettingConfigurationList$member":null, + "OptionSettingsList$member":null } }, - "OptionNamesList": { - "base": null, - "refs": { - "ModifyOptionGroupMessage$OptionsToRemove": "

Options in this list are removed from the option group.

" + "OptionSettingConfigurationList":{ + "base":null, + "refs":{ + "Option$OptionSettings":"

The option settings for this option.

" } }, - "OptionSetting": { - "base": "

Option settings are the actual settings being applied or configured for that option. It is used when you modify an option group or describe option groups. For example, the NATIVE_NETWORK_ENCRYPTION option has a setting called SQLNET.ENCRYPTION_SERVER that can have several different values.

", - "refs": { - "OptionSettingConfigurationList$member": null, - "OptionSettingsList$member": null + "OptionSettingsList":{ + "base":null, + "refs":{ + "OptionConfiguration$OptionSettings":"

The option settings to include in an option group.

" } }, - "OptionSettingConfigurationList": { - "base": null, - "refs": { - "Option$OptionSettings": "

The option settings for this option.

" + "OptionVersion":{ + "base":"

The version for an option. Option group option versions are returned by the DescribeOptionGroupOptions action.

", + "refs":{ + "OptionGroupOptionVersionsList$member":null } }, - "OptionSettingsList": { - "base": null, - "refs": { - "OptionConfiguration$OptionSettings": "

The option settings to include in an option group.

" + "OptionsConflictsWith":{ + "base":null, + "refs":{ + "OptionGroupOption$OptionsConflictsWith":"

The options that conflict with this option.

" } }, - "OptionVersion": { - "base": "

The version for an option. Option group option versions are returned by the DescribeOptionGroupOptions action.

", - "refs": { - "OptionGroupOptionVersionsList$member": null + "OptionsDependedOn":{ + "base":null, + "refs":{ + "OptionGroupOption$OptionsDependedOn":"

The options that are prerequisites for this option.

" } }, - "OptionsConflictsWith": { - "base": null, - "refs": { - "OptionGroupOption$OptionsConflictsWith": "

The options that conflict with this option.

" + "OptionsList":{ + "base":null, + "refs":{ + "OptionGroup$Options":"

Indicates what options are available in the option group.

" } }, - "OptionsDependedOn": { - "base": null, - "refs": { - "OptionGroupOption$OptionsDependedOn": "

The options that are prerequisites for this option.

" + "OrderableDBInstanceOption":{ + "base":"

Contains a list of available options for a DB instance.

This data type is used as a response element in the DescribeOrderableDBInstanceOptions action.

", + "refs":{ + "OrderableDBInstanceOptionsList$member":null } }, - "OptionsList": { - "base": null, - "refs": { - "OptionGroup$Options": "

Indicates what options are available in the option group.

" + "OrderableDBInstanceOptionsList":{ + "base":null, + "refs":{ + "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions":"

An OrderableDBInstanceOption structure containing information about orderable options for the DB instance.

" } }, - "OrderableDBInstanceOption": { - "base": "

Contains a list of available options for a DB instance.

This data type is used as a response element in the DescribeOrderableDBInstanceOptions action.

", - "refs": { - "OrderableDBInstanceOptionsList$member": null + "OrderableDBInstanceOptionsMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeOrderableDBInstanceOptions action.

", + "refs":{ } }, - "OrderableDBInstanceOptionsList": { - "base": null, - "refs": { - "OrderableDBInstanceOptionsMessage$OrderableDBInstanceOptions": "

An OrderableDBInstanceOption structure containing information about orderable options for the DB instance.

" + "Parameter":{ + "base":"

This data type is used as a request parameter in the ModifyDBParameterGroup and ResetDBParameterGroup actions.

This data type is used as a response element in the DescribeEngineDefaultParameters and DescribeDBParameters actions.

", + "refs":{ + "ParametersList$member":null } }, - "OrderableDBInstanceOptionsMessage": { - "base": "

Contains the result of a successful invocation of the DescribeOrderableDBInstanceOptions action.

", - "refs": { + "ParametersList":{ + "base":null, + "refs":{ + "DBClusterParameterGroupDetails$Parameters":"

Provides a list of parameters for the DB cluster parameter group.

", + "DBParameterGroupDetails$Parameters":"

A list of Parameter values.

", + "EngineDefaults$Parameters":"

Contains a list of engine default parameters.

", + "ModifyDBClusterParameterGroupMessage$Parameters":"

A list of parameters in the DB cluster parameter group to modify.

", + "ModifyDBParameterGroupMessage$Parameters":"

An array of parameter names, values, and the apply method for the parameter update. At least one parameter name, value, and apply method must be supplied; subsequent arguments are optional. A maximum of 20 parameters can be modified in a single request.

Valid Values (for the application method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when you reboot the DB instance without failover.

", + "ResetDBClusterParameterGroupMessage$Parameters":"

A list of parameter names in the DB cluster parameter group to reset to the default values. You can't use this parameter if the ResetAllParameters parameter is set to true.

", + "ResetDBParameterGroupMessage$Parameters":"

To reset the entire DB parameter group, specify the DBParameterGroup name and ResetAllParameters parameters. To reset specific parameters, provide a list of the following: ParameterName and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

MySQL

Valid Values (for Apply method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when DB instance reboots.

MariaDB

Valid Values (for Apply method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when DB instance reboots.

Oracle

Valid Values (for Apply method): pending-reboot

" } }, - "Parameter": { - "base": "

This data type is used as a request parameter in the ModifyDBParameterGroup and ResetDBParameterGroup actions.

This data type is used as a response element in the DescribeEngineDefaultParameters and DescribeDBParameters actions.

", - "refs": { - "ParametersList$member": null + "PendingCloudwatchLogsExports":{ + "base":"

A list of the log types whose configuration is still pending. In other words, these log types are in the process of being activated or deactivated.

", + "refs":{ + "PendingModifiedValues$PendingCloudwatchLogsExports":null } }, - "ParametersList": { - "base": null, - "refs": { - "DBClusterParameterGroupDetails$Parameters": "

Provides a list of parameters for the DB cluster parameter group.

", - "DBParameterGroupDetails$Parameters": "

A list of Parameter values.

", - "EngineDefaults$Parameters": "

Contains a list of engine default parameters.

", - "ModifyDBClusterParameterGroupMessage$Parameters": "

A list of parameters in the DB cluster parameter group to modify.

", - "ModifyDBParameterGroupMessage$Parameters": "

An array of parameter names, values, and the apply method for the parameter update. At least one parameter name, value, and apply method must be supplied; subsequent arguments are optional. A maximum of 20 parameters can be modified in a single request.

Valid Values (for the application method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when you reboot the DB instance without failover.

", - "ResetDBClusterParameterGroupMessage$Parameters": "

A list of parameter names in the DB cluster parameter group to reset to the default values. You can't use this parameter if the ResetAllParameters parameter is set to true.

", - "ResetDBParameterGroupMessage$Parameters": "

To reset the entire DB parameter group, specify the DBParameterGroup name and ResetAllParameters parameters. To reset specific parameters, provide a list of the following: ParameterName and ApplyMethod. A maximum of 20 parameters can be modified in a single request.

MySQL

Valid Values (for Apply method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when DB instance reboots.

MariaDB

Valid Values (for Apply method): immediate | pending-reboot

You can use the immediate value with dynamic parameters only. You can use the pending-reboot value for both dynamic and static parameters, and changes are applied when DB instance reboots.

Oracle

Valid Values (for Apply method): pending-reboot

" + "PendingMaintenanceAction":{ + "base":"

Provides information about a pending maintenance action for a resource.

", + "refs":{ + "PendingMaintenanceActionDetails$member":null } }, - "PendingCloudwatchLogsExports": { - "base": "

A list of the log types whose configuration is still pending. In other words, these log types are in the process of being activated or deactivated.

", - "refs": { - "PendingModifiedValues$PendingCloudwatchLogsExports": null + "PendingMaintenanceActionDetails":{ + "base":null, + "refs":{ + "ResourcePendingMaintenanceActions$PendingMaintenanceActionDetails":"

A list that provides details about the pending maintenance actions for the resource.

" } }, - "PendingMaintenanceAction": { - "base": "

Provides information about a pending maintenance action for a resource.

", - "refs": { - "PendingMaintenanceActionDetails$member": null + "PendingMaintenanceActions":{ + "base":null, + "refs":{ + "PendingMaintenanceActionsMessage$PendingMaintenanceActions":"

A list of the pending maintenance actions for the resource.

" } }, - "PendingMaintenanceActionDetails": { - "base": null, - "refs": { - "ResourcePendingMaintenanceActions$PendingMaintenanceActionDetails": "

A list that provides details about the pending maintenance actions for the resource.

" + "PendingMaintenanceActionsMessage":{ + "base":"

Data returned from the DescribePendingMaintenanceActions action.

", + "refs":{ } }, - "PendingMaintenanceActions": { - "base": null, - "refs": { - "PendingMaintenanceActionsMessage$PendingMaintenanceActions": "

A list of the pending maintenance actions for the resource.

" + "PendingModifiedValues":{ + "base":"

This data type is used as a response element in the ModifyDBInstance action.

", + "refs":{ + "DBInstance$PendingModifiedValues":"

Specifies that changes to the DB instance are pending. This element is only included when changes are pending. Specific changes are identified by subelements.

" } }, - "PendingMaintenanceActionsMessage": { - "base": "

Data returned from the DescribePendingMaintenanceActions action.

", - "refs": { + "PointInTimeRestoreNotEnabledFault":{ + "base":"

SourceDBInstanceIdentifier refers to a DB instance with BackupRetentionPeriod equal to 0.

", + "refs":{ } }, - "PendingModifiedValues": { - "base": "

This data type is used as a response element in the ModifyDBInstance action.

", - "refs": { - "DBInstance$PendingModifiedValues": "

Specifies that changes to the DB instance are pending. This element is only included when changes are pending. Specific changes are identified by subelements.

" + "ProcessorFeature":{ + "base":"

Contains the processor features of a DB instance class.

To specify the number of CPU cores, use the coreCount feature name for the Name parameter. To specify the number of threads per core, use the threadsPerCore feature name for the Name parameter.

You can set the processor features of the DB instance class for a DB instance when you call one of the following actions:

You can view the valid processor values for a particular instance class by calling the DescribeOrderableDBInstanceOptions action and specifying the instance class for the DBInstanceClass parameter.

In addition, you can use the following actions for DB instance class processor information:

For more information, see Configuring the Processor of the DB Instance Class in the Amazon RDS User Guide.

", + "refs":{ + "ProcessorFeatureList$member":null } }, - "PointInTimeRestoreNotEnabledFault": { - "base": "

SourceDBInstanceIdentifier refers to a DB instance with BackupRetentionPeriod equal to 0.

", - "refs": { + "ProcessorFeatureList":{ + "base":null, + "refs":{ + "CreateDBInstanceMessage$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "CreateDBInstanceReadReplicaMessage$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "DBInstance$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "DBSnapshot$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance when the DB snapshot was created.

", + "ModifyDBInstanceMessage$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "PendingModifiedValues$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "RestoreDBInstanceFromDBSnapshotMessage$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "RestoreDBInstanceFromS3Message$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

", + "RestoreDBInstanceToPointInTimeMessage$ProcessorFeatures":"

The number of CPU cores and the number of threads per core for the DB instance class of the DB instance.

" } }, - "PromoteReadReplicaDBClusterMessage": { - "base": "

", - "refs": { + "PromoteReadReplicaDBClusterMessage":{ + "base":"

", + "refs":{ } }, - "PromoteReadReplicaDBClusterResult": { - "base": null, - "refs": { + "PromoteReadReplicaDBClusterResult":{ + "base":null, + "refs":{ } }, - "PromoteReadReplicaMessage": { - "base": "

", - "refs": { + "PromoteReadReplicaMessage":{ + "base":"

", + "refs":{ } }, - "PromoteReadReplicaResult": { - "base": null, - "refs": { + "PromoteReadReplicaResult":{ + "base":null, + "refs":{ } }, - "ProvisionedIopsNotAvailableInAZFault": { - "base": "

Provisioned IOPS not available in the specified Availability Zone.

", - "refs": { + "ProvisionedIopsNotAvailableInAZFault":{ + "base":"

Provisioned IOPS not available in the specified Availability Zone.

", + "refs":{ } }, - "PurchaseReservedDBInstancesOfferingMessage": { - "base": "

", - "refs": { + "PurchaseReservedDBInstancesOfferingMessage":{ + "base":"

", + "refs":{ } }, - "PurchaseReservedDBInstancesOfferingResult": { - "base": null, - "refs": { + "PurchaseReservedDBInstancesOfferingResult":{ + "base":null, + "refs":{ } }, - "Range": { - "base": "

A range of integer values.

", - "refs": { - "RangeList$member": null + "Range":{ + "base":"

A range of integer values.

", + "refs":{ + "RangeList$member":null } }, - "RangeList": { - "base": null, - "refs": { - "ValidStorageOptions$StorageSize": "

The valid range of storage in gibibytes. For example, 100 to 16384.

", - "ValidStorageOptions$ProvisionedIops": "

The valid range of provisioned IOPS. For example, 1000-20000.

" + "RangeList":{ + "base":null, + "refs":{ + "ValidStorageOptions$StorageSize":"

The valid range of storage in gibibytes. For example, 100 to 16384.

", + "ValidStorageOptions$ProvisionedIops":"

The valid range of provisioned IOPS. For example, 1000-20000.

" } }, - "ReadReplicaDBClusterIdentifierList": { - "base": null, - "refs": { - "DBInstance$ReadReplicaDBClusterIdentifiers": "

Contains one or more identifiers of Aurora DB clusters that are Read Replicas of this DB instance.

" + "ReadReplicaDBClusterIdentifierList":{ + "base":null, + "refs":{ + "DBInstance$ReadReplicaDBClusterIdentifiers":"

Contains one or more identifiers of Aurora DB clusters that are Read Replicas of this DB instance.

" } }, - "ReadReplicaDBInstanceIdentifierList": { - "base": null, - "refs": { - "DBInstance$ReadReplicaDBInstanceIdentifiers": "

Contains one or more identifiers of the Read Replicas associated with this DB instance.

" + "ReadReplicaDBInstanceIdentifierList":{ + "base":null, + "refs":{ + "DBInstance$ReadReplicaDBInstanceIdentifiers":"

Contains one or more identifiers of the Read Replicas associated with this DB instance.

" } }, - "ReadReplicaIdentifierList": { - "base": null, - "refs": { - "DBCluster$ReadReplicaIdentifiers": "

Contains one or more identifiers of the Read Replicas associated with this DB cluster.

" + "ReadReplicaIdentifierList":{ + "base":null, + "refs":{ + "DBCluster$ReadReplicaIdentifiers":"

Contains one or more identifiers of the Read Replicas associated with this DB cluster.

" } }, - "RebootDBInstanceMessage": { - "base": "

", - "refs": { + "RebootDBInstanceMessage":{ + "base":"

", + "refs":{ } }, - "RebootDBInstanceResult": { - "base": null, - "refs": { + "RebootDBInstanceResult":{ + "base":null, + "refs":{ } }, - "RecurringCharge": { - "base": "

This data type is used as a response element in the DescribeReservedDBInstances and DescribeReservedDBInstancesOfferings actions.

", - "refs": { - "RecurringChargeList$member": null + "RecurringCharge":{ + "base":"

This data type is used as a response element in the DescribeReservedDBInstances and DescribeReservedDBInstancesOfferings actions.

", + "refs":{ + "RecurringChargeList$member":null } }, - "RecurringChargeList": { - "base": null, - "refs": { - "ReservedDBInstance$RecurringCharges": "

The recurring price charged to run this reserved DB instance.

", - "ReservedDBInstancesOffering$RecurringCharges": "

The recurring price charged to run this reserved DB instance.

" + "RecurringChargeList":{ + "base":null, + "refs":{ + "ReservedDBInstance$RecurringCharges":"

The recurring price charged to run this reserved DB instance.

", + "ReservedDBInstancesOffering$RecurringCharges":"

The recurring price charged to run this reserved DB instance.

" } }, - "RemoveRoleFromDBClusterMessage": { - "base": null, - "refs": { + "RemoveRoleFromDBClusterMessage":{ + "base":null, + "refs":{ } }, - "RemoveSourceIdentifierFromSubscriptionMessage": { - "base": "

", - "refs": { + "RemoveSourceIdentifierFromSubscriptionMessage":{ + "base":"

", + "refs":{ } }, - "RemoveSourceIdentifierFromSubscriptionResult": { - "base": null, - "refs": { + "RemoveSourceIdentifierFromSubscriptionResult":{ + "base":null, + "refs":{ } }, - "RemoveTagsFromResourceMessage": { - "base": "

", - "refs": { + "RemoveTagsFromResourceMessage":{ + "base":"

", + "refs":{ } }, - "ReservedDBInstance": { - "base": "

This data type is used as a response element in the DescribeReservedDBInstances and PurchaseReservedDBInstancesOffering actions.

", - "refs": { - "PurchaseReservedDBInstancesOfferingResult$ReservedDBInstance": null, - "ReservedDBInstanceList$member": null + "ReservedDBInstance":{ + "base":"

This data type is used as a response element in the DescribeReservedDBInstances and PurchaseReservedDBInstancesOffering actions.

", + "refs":{ + "PurchaseReservedDBInstancesOfferingResult$ReservedDBInstance":null, + "ReservedDBInstanceList$member":null } }, - "ReservedDBInstanceAlreadyExistsFault": { - "base": "

User already has a reservation with the given identifier.

", - "refs": { + "ReservedDBInstanceAlreadyExistsFault":{ + "base":"

User already has a reservation with the given identifier.

", + "refs":{ } }, - "ReservedDBInstanceList": { - "base": null, - "refs": { - "ReservedDBInstanceMessage$ReservedDBInstances": "

A list of reserved DB instances.

" + "ReservedDBInstanceList":{ + "base":null, + "refs":{ + "ReservedDBInstanceMessage$ReservedDBInstances":"

A list of reserved DB instances.

" } }, - "ReservedDBInstanceMessage": { - "base": "

Contains the result of a successful invocation of the DescribeReservedDBInstances action.

", - "refs": { + "ReservedDBInstanceMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeReservedDBInstances action.

", + "refs":{ } }, - "ReservedDBInstanceNotFoundFault": { - "base": "

The specified reserved DB Instance not found.

", - "refs": { + "ReservedDBInstanceNotFoundFault":{ + "base":"

The specified reserved DB Instance not found.

", + "refs":{ } }, - "ReservedDBInstanceQuotaExceededFault": { - "base": "

Request would exceed the user's DB Instance quota.

", - "refs": { + "ReservedDBInstanceQuotaExceededFault":{ + "base":"

Request would exceed the user's DB Instance quota.

", + "refs":{ } }, - "ReservedDBInstancesOffering": { - "base": "

This data type is used as a response element in the DescribeReservedDBInstancesOfferings action.

", - "refs": { - "ReservedDBInstancesOfferingList$member": null + "ReservedDBInstancesOffering":{ + "base":"

This data type is used as a response element in the DescribeReservedDBInstancesOfferings action.

", + "refs":{ + "ReservedDBInstancesOfferingList$member":null } }, - "ReservedDBInstancesOfferingList": { - "base": null, - "refs": { - "ReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferings": "

A list of reserved DB instance offerings.

" + "ReservedDBInstancesOfferingList":{ + "base":null, + "refs":{ + "ReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferings":"

A list of reserved DB instance offerings.

" } }, - "ReservedDBInstancesOfferingMessage": { - "base": "

Contains the result of a successful invocation of the DescribeReservedDBInstancesOfferings action.

", - "refs": { + "ReservedDBInstancesOfferingMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeReservedDBInstancesOfferings action.

", + "refs":{ } }, - "ReservedDBInstancesOfferingNotFoundFault": { - "base": "

Specified offering does not exist.

", - "refs": { + "ReservedDBInstancesOfferingNotFoundFault":{ + "base":"

Specified offering does not exist.

", + "refs":{ } }, - "ResetDBClusterParameterGroupMessage": { - "base": "

", - "refs": { + "ResetDBClusterParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "ResetDBParameterGroupMessage": { - "base": "

", - "refs": { + "ResetDBParameterGroupMessage":{ + "base":"

", + "refs":{ } }, - "ResourceNotFoundFault": { - "base": "

The specified resource ID was not found.

", - "refs": { + "ResourceNotFoundFault":{ + "base":"

The specified resource ID was not found.

", + "refs":{ } }, - "ResourcePendingMaintenanceActions": { - "base": "

Describes the pending maintenance actions for a resource.

", - "refs": { - "ApplyPendingMaintenanceActionResult$ResourcePendingMaintenanceActions": null, - "PendingMaintenanceActions$member": null + "ResourcePendingMaintenanceActions":{ + "base":"

Describes the pending maintenance actions for a resource.

", + "refs":{ + "ApplyPendingMaintenanceActionResult$ResourcePendingMaintenanceActions":null, + "PendingMaintenanceActions$member":null } }, - "RestoreDBClusterFromS3Message": { - "base": null, - "refs": { + "RestoreDBClusterFromS3Message":{ + "base":null, + "refs":{ } }, - "RestoreDBClusterFromS3Result": { - "base": null, - "refs": { + "RestoreDBClusterFromS3Result":{ + "base":null, + "refs":{ } }, - "RestoreDBClusterFromSnapshotMessage": { - "base": "

", - "refs": { + "RestoreDBClusterFromSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "RestoreDBClusterFromSnapshotResult": { - "base": null, - "refs": { + "RestoreDBClusterFromSnapshotResult":{ + "base":null, + "refs":{ } }, - "RestoreDBClusterToPointInTimeMessage": { - "base": "

", - "refs": { + "RestoreDBClusterToPointInTimeMessage":{ + "base":"

", + "refs":{ } }, - "RestoreDBClusterToPointInTimeResult": { - "base": null, - "refs": { + "RestoreDBClusterToPointInTimeResult":{ + "base":null, + "refs":{ } }, - "RestoreDBInstanceFromDBSnapshotMessage": { - "base": "

", - "refs": { + "RestoreDBInstanceFromDBSnapshotMessage":{ + "base":"

", + "refs":{ } }, - "RestoreDBInstanceFromDBSnapshotResult": { - "base": null, - "refs": { + "RestoreDBInstanceFromDBSnapshotResult":{ + "base":null, + "refs":{ } }, - "RestoreDBInstanceFromS3Message": { - "base": null, - "refs": { + "RestoreDBInstanceFromS3Message":{ + "base":null, + "refs":{ } }, - "RestoreDBInstanceFromS3Result": { - "base": null, - "refs": { + "RestoreDBInstanceFromS3Result":{ + "base":null, + "refs":{ } }, - "RestoreDBInstanceToPointInTimeMessage": { - "base": "

", - "refs": { + "RestoreDBInstanceToPointInTimeMessage":{ + "base":"

", + "refs":{ } }, - "RestoreDBInstanceToPointInTimeResult": { - "base": null, - "refs": { + "RestoreDBInstanceToPointInTimeResult":{ + "base":null, + "refs":{ } }, - "RevokeDBSecurityGroupIngressMessage": { - "base": "

", - "refs": { + "RevokeDBSecurityGroupIngressMessage":{ + "base":"

", + "refs":{ } }, - "RevokeDBSecurityGroupIngressResult": { - "base": null, - "refs": { + "RevokeDBSecurityGroupIngressResult":{ + "base":null, + "refs":{ } }, - "SNSInvalidTopicFault": { - "base": "

SNS has responded that there is a problem with the SND topic specified.

", - "refs": { + "SNSInvalidTopicFault":{ + "base":"

SNS has responded that there is a problem with the SND topic specified.

", + "refs":{ } }, - "SNSNoAuthorizationFault": { - "base": "

You do not have permission to publish to the SNS topic ARN.

", - "refs": { + "SNSNoAuthorizationFault":{ + "base":"

You do not have permission to publish to the SNS topic ARN.

", + "refs":{ } }, - "SNSTopicArnNotFoundFault": { - "base": "

The SNS topic ARN does not exist.

", - "refs": { + "SNSTopicArnNotFoundFault":{ + "base":"

The SNS topic ARN does not exist.

", + "refs":{ } }, - "SharedSnapshotQuotaExceededFault": { - "base": "

You have exceeded the maximum number of accounts that you can share a manual DB snapshot with.

", - "refs": { + "SharedSnapshotQuotaExceededFault":{ + "base":"

You have exceeded the maximum number of accounts that you can share a manual DB snapshot with.

", + "refs":{ } }, - "SnapshotQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed number of DB snapshots.

", - "refs": { + "SnapshotQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed number of DB snapshots.

", + "refs":{ } }, - "SourceIdsList": { - "base": null, - "refs": { - "CreateEventSubscriptionMessage$SourceIds": "

The list of identifiers of the event sources for which events are returned. If not specified, then all sources are included in the response. An identifier must begin with a letter and must contain only ASCII letters, digits, and hyphens; it can't end with a hyphen or contain two consecutive hyphens.

Constraints:

  • If SourceIds are supplied, SourceType must also be provided.

  • If the source type is a DB instance, then a DBInstanceIdentifier must be supplied.

  • If the source type is a DB security group, a DBSecurityGroupName must be supplied.

  • If the source type is a DB parameter group, a DBParameterGroupName must be supplied.

  • If the source type is a DB snapshot, a DBSnapshotIdentifier must be supplied.

", - "EventSubscription$SourceIdsList": "

A list of source IDs for the RDS event notification subscription.

" + "SourceIdsList":{ + "base":null, + "refs":{ + "CreateEventSubscriptionMessage$SourceIds":"

The list of identifiers of the event sources for which events are returned. If not specified, then all sources are included in the response. An identifier must begin with a letter and must contain only ASCII letters, digits, and hyphens; it can't end with a hyphen or contain two consecutive hyphens.

Constraints:

  • If SourceIds are supplied, SourceType must also be provided.

  • If the source type is a DB instance, then a DBInstanceIdentifier must be supplied.

  • If the source type is a DB security group, a DBSecurityGroupName must be supplied.

  • If the source type is a DB parameter group, a DBParameterGroupName must be supplied.

  • If the source type is a DB snapshot, a DBSnapshotIdentifier must be supplied.

", + "EventSubscription$SourceIdsList":"

A list of source IDs for the RDS event notification subscription.

" } }, - "SourceNotFoundFault": { - "base": "

The requested source could not be found.

", - "refs": { + "SourceNotFoundFault":{ + "base":"

The requested source could not be found.

", + "refs":{ } }, - "SourceRegion": { - "base": "

Contains an AWS Region name as the result of a successful call to the DescribeSourceRegions action.

", - "refs": { - "SourceRegionList$member": null + "SourceRegion":{ + "base":"

Contains an AWS Region name as the result of a successful call to the DescribeSourceRegions action.

", + "refs":{ + "SourceRegionList$member":null } }, - "SourceRegionList": { - "base": null, - "refs": { - "SourceRegionMessage$SourceRegions": "

A list of SourceRegion instances that contains each source AWS Region that the current AWS Region can get a Read Replica or a DB snapshot from.

" + "SourceRegionList":{ + "base":null, + "refs":{ + "SourceRegionMessage$SourceRegions":"

A list of SourceRegion instances that contains each source AWS Region that the current AWS Region can get a Read Replica or a DB snapshot from.

" } }, - "SourceRegionMessage": { - "base": "

Contains the result of a successful invocation of the DescribeSourceRegions action.

", - "refs": { + "SourceRegionMessage":{ + "base":"

Contains the result of a successful invocation of the DescribeSourceRegions action.

", + "refs":{ } }, - "SourceType": { - "base": null, - "refs": { - "DescribeEventsMessage$SourceType": "

The event source to retrieve events for. If no value is specified, all events are returned.

", - "Event$SourceType": "

Specifies the source type for this event.

" + "SourceType":{ + "base":null, + "refs":{ + "DescribeEventsMessage$SourceType":"

The event source to retrieve events for. If no value is specified, all events are returned.

", + "Event$SourceType":"

Specifies the source type for this event.

" } }, - "StartDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "StartDBInstanceResult": { - "base": null, - "refs": { - } - }, - "StopDBInstanceMessage": { - "base": null, - "refs": { - } - }, - "StopDBInstanceResult": { - "base": null, - "refs": { - } - }, - "StorageQuotaExceededFault": { - "base": "

The request would result in the user exceeding the allowed amount of storage available across all DB instances.

", - "refs": { - } - }, - "StorageTypeNotSupportedFault": { - "base": "

Storage of the StorageType specified can't be associated with the DB instance.

", - "refs": { - } - }, - "String": { - "base": null, - "refs": { - "AccountQuota$AccountQuotaName": "

The name of the Amazon RDS quota for this AWS account.

", - "AddRoleToDBClusterMessage$DBClusterIdentifier": "

The name of the DB cluster to associate the IAM role with.

", - "AddRoleToDBClusterMessage$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role to associate with the Aurora DB cluster, for example arn:aws:iam::123456789012:role/AuroraAccessRole.

", - "AddSourceIdentifierToSubscriptionMessage$SubscriptionName": "

The name of the RDS event notification subscription you want to add a source identifier to.

", - "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier": "

The identifier of the event source to be added.

Constraints:

  • If the source type is a DB instance, then a DBInstanceIdentifier must be supplied.

  • If the source type is a DB security group, a DBSecurityGroupName must be supplied.

  • If the source type is a DB parameter group, a DBParameterGroupName must be supplied.

  • If the source type is a DB snapshot, a DBSnapshotIdentifier must be supplied.

", - "AddTagsToResourceMessage$ResourceName": "

The Amazon RDS resource that the tags are added to. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", - "ApplyPendingMaintenanceActionMessage$ResourceIdentifier": "

The RDS Amazon Resource Name (ARN) of the resource that the pending maintenance action applies to. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", - "ApplyPendingMaintenanceActionMessage$ApplyAction": "

The pending maintenance action to apply to this resource.

Valid values: system-update, db-upgrade

", - "ApplyPendingMaintenanceActionMessage$OptInType": "

A value that specifies the type of opt-in request, or undoes an opt-in request. An opt-in request of type immediate can't be undone.

Valid values:

  • immediate - Apply the maintenance action immediately.

  • next-maintenance - Apply the maintenance action during the next maintenance window for the resource.

  • undo-opt-in - Cancel any existing next-maintenance opt-in requests.

", - "AttributeValueList$member": null, - "AuthorizeDBSecurityGroupIngressMessage$DBSecurityGroupName": "

The name of the DB security group to add authorization to.

", - "AuthorizeDBSecurityGroupIngressMessage$CIDRIP": "

The IP range to authorize.

", - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupName": "

Name of the EC2 security group to authorize. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupId": "

Id of the EC2 security group to authorize. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", - "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": "

AWS account number of the owner of the EC2 security group specified in the EC2SecurityGroupName parameter. The AWS Access Key ID is not an acceptable value. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", - "AvailabilityZone$Name": "

The name of the availability zone.

", - "AvailabilityZones$member": null, - "BacktrackDBClusterMessage$DBClusterIdentifier": "

The DB cluster identifier of the DB cluster to be backtracked. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 alphanumeric characters or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", - "Certificate$CertificateIdentifier": "

The unique key that identifies a certificate.

", - "Certificate$CertificateType": "

The type of the certificate.

", - "Certificate$Thumbprint": "

The thumbprint of the certificate.

", - "Certificate$CertificateArn": "

The Amazon Resource Name (ARN) for the certificate.

", - "CertificateMessage$Marker": "

An optional pagination token provided by a previous DescribeCertificates request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "CharacterSet$CharacterSetName": "

The name of the character set.

", - "CharacterSet$CharacterSetDescription": "

The description of the character set.

", - "CopyDBClusterParameterGroupMessage$SourceDBClusterParameterGroupIdentifier": "

The identifier or Amazon Resource Name (ARN) for the source DB cluster parameter group. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid DB cluster parameter group.

  • If the source DB cluster parameter group is in the same AWS Region as the copy, specify a valid DB parameter group identifier, for example my-db-cluster-param-group, or a valid ARN.

  • If the source DB parameter group is in a different AWS Region than the copy, specify a valid DB cluster parameter group ARN, for example arn:aws:rds:us-east-1:123456789012:cluster-pg:custom-cluster-group1.

", - "CopyDBClusterParameterGroupMessage$TargetDBClusterParameterGroupIdentifier": "

The identifier for the copied DB cluster parameter group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-cluster-param-group1

", - "CopyDBClusterParameterGroupMessage$TargetDBClusterParameterGroupDescription": "

A description for the copied DB cluster parameter group.

", - "CopyDBClusterSnapshotMessage$SourceDBClusterSnapshotIdentifier": "

The identifier of the DB cluster snapshot to copy. This parameter is not case-sensitive.

You can't copy an encrypted, shared DB cluster snapshot from one AWS Region to another.

Constraints:

  • Must specify a valid system snapshot in the \"available\" state.

  • If the source snapshot is in the same AWS Region as the copy, specify a valid DB snapshot identifier.

  • If the source snapshot is in a different AWS Region than the copy, specify a valid DB cluster snapshot ARN. For more information, go to Copying a DB Snapshot or DB Cluster Snapshot.

Example: my-cluster-snapshot1

", - "CopyDBClusterSnapshotMessage$TargetDBClusterSnapshotIdentifier": "

The identifier of the new DB cluster snapshot to create from the source DB cluster snapshot. This parameter is not case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster-snapshot2

", - "CopyDBClusterSnapshotMessage$KmsKeyId": "

The AWS AWS KMS key ID for an encrypted DB cluster snapshot. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you copy an encrypted DB cluster snapshot from your AWS account, you can specify a value for KmsKeyId to encrypt the copy with a new KMS encryption key. If you don't specify a value for KmsKeyId, then the copy of the DB cluster snapshot is encrypted with the same KMS key as the source DB cluster snapshot.

If you copy an encrypted DB cluster snapshot that is shared from another AWS account, then you must specify a value for KmsKeyId.

To copy an encrypted DB cluster snapshot to another AWS Region, you must set KmsKeyId to the KMS key ID you want to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

If you copy an unencrypted DB cluster snapshot and specify a value for the KmsKeyId parameter, an error is returned.

", - "CopyDBClusterSnapshotMessage$PreSignedUrl": "

The URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot API action in the AWS Region that contains the source DB cluster snapshot to copy. The PreSignedUrl parameter must be used when copying an encrypted DB cluster snapshot from another AWS Region.

The pre-signed URL must be a valid request for the CopyDBSClusterSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB cluster snapshot to be copied. The pre-signed URL request must contain the following parameter values:

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. This is the same identifier for both the CopyDBClusterSnapshot action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

  • DestinationRegion - The name of the AWS Region that the DB cluster snapshot will be created in.

  • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster snapshot from the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:aurora-cluster1-snapshot-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", - "CopyDBParameterGroupMessage$SourceDBParameterGroupIdentifier": "

The identifier or ARN for the source DB parameter group. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid DB parameter group.

  • Must specify a valid DB parameter group identifier, for example my-db-param-group, or a valid ARN.

", - "CopyDBParameterGroupMessage$TargetDBParameterGroupIdentifier": "

The identifier for the copied DB parameter group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-db-parameter-group

", - "CopyDBParameterGroupMessage$TargetDBParameterGroupDescription": "

A description for the copied DB parameter group.

", - "CopyDBSnapshotMessage$SourceDBSnapshotIdentifier": "

The identifier for the source DB snapshot.

If the source snapshot is in the same AWS Region as the copy, specify a valid DB snapshot identifier. For example, you might specify rds:mysql-instance1-snapshot-20130805.

If the source snapshot is in a different AWS Region than the copy, specify a valid DB snapshot ARN. For example, you might specify arn:aws:rds:us-west-2:123456789012:snapshot:mysql-instance1-snapshot-20130805.

If you are copying from a shared manual DB snapshot, this parameter must be the Amazon Resource Name (ARN) of the shared DB snapshot.

If you are copying an encrypted snapshot this parameter must be in the ARN format for the source AWS Region, and must match the SourceDBSnapshotIdentifier in the PreSignedUrl parameter.

Constraints:

  • Must specify a valid system snapshot in the \"available\" state.

Example: rds:mydb-2012-04-02-00-01

Example: arn:aws:rds:us-west-2:123456789012:snapshot:mysql-instance1-snapshot-20130805

", - "CopyDBSnapshotMessage$TargetDBSnapshotIdentifier": "

The identifier for the copy of the snapshot.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-db-snapshot

", - "CopyDBSnapshotMessage$KmsKeyId": "

The AWS KMS key ID for an encrypted DB snapshot. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you copy an encrypted DB snapshot from your AWS account, you can specify a value for this parameter to encrypt the copy with a new KMS encryption key. If you don't specify a value for this parameter, then the copy of the DB snapshot is encrypted with the same KMS key as the source DB snapshot.

If you copy an encrypted DB snapshot that is shared from another AWS account, then you must specify a value for this parameter.

If you specify this parameter when you copy an unencrypted snapshot, the copy is encrypted.

If you copy an encrypted snapshot to a different AWS Region, then you must specify a KMS key for the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

", - "CopyDBSnapshotMessage$PreSignedUrl": "

The URL that contains a Signature Version 4 signed request for the CopyDBSnapshot API action in the source AWS Region that contains the source DB snapshot to copy.

You must specify this parameter when you copy an encrypted DB snapshot from another AWS Region by using the Amazon RDS API. You can specify the --source-region option instead of this parameter when you copy an encrypted DB snapshot from another AWS Region by using the AWS CLI.

The presigned URL must be a valid request for the CopyDBSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB snapshot to be copied. The presigned URL request must contain the following parameter values:

  • DestinationRegion - The AWS Region that the encrypted DB snapshot is copied to. This AWS Region is the same one where the CopyDBSnapshot action is called that contains this presigned URL.

    For example, if you copy an encrypted DB snapshot from the us-west-2 AWS Region to the us-east-1 AWS Region, then you call the CopyDBSnapshot action in the us-east-1 AWS Region and provide a presigned URL that contains a call to the CopyDBSnapshot action in the us-west-2 AWS Region. For this example, the DestinationRegion in the presigned URL must be set to the us-east-1 AWS Region.

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB snapshot in the destination AWS Region. This is the same identifier for both the CopyDBSnapshot action that is called in the destination AWS Region, and the action contained in the presigned URL.

  • SourceDBSnapshotIdentifier - The DB snapshot identifier for the encrypted snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB snapshot from the us-west-2 AWS Region, then your SourceDBSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:snapshot:mysql-instance1-snapshot-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", - "CopyDBSnapshotMessage$OptionGroupName": "

The name of an option group to associate with the copy of the snapshot.

Specify this option if you are copying a snapshot from one AWS Region to another, and your DB instance uses a nondefault option group. If your source DB instance uses Transparent Data Encryption for Oracle or Microsoft SQL Server, you must specify this option when copying across AWS Regions. For more information, see Option Group Considerations.

", - "CopyOptionGroupMessage$SourceOptionGroupIdentifier": "

The identifier or ARN for the source option group. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid option group.

  • If the source option group is in the same AWS Region as the copy, specify a valid option group identifier, for example my-option-group, or a valid ARN.

  • If the source option group is in a different AWS Region than the copy, specify a valid option group ARN, for example arn:aws:rds:us-west-2:123456789012:og:special-options.

", - "CopyOptionGroupMessage$TargetOptionGroupIdentifier": "

The identifier for the copied option group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-option-group

", - "CopyOptionGroupMessage$TargetOptionGroupDescription": "

The description for the copied option group.

", - "CreateDBClusterMessage$CharacterSetName": "

A value that indicates that the DB cluster should be associated with the specified CharacterSet.

", - "CreateDBClusterMessage$DatabaseName": "

The name for your database of up to 64 alpha-numeric characters. If you do not provide a name, Amazon RDS will not create a database in the DB cluster you are creating.

", - "CreateDBClusterMessage$DBClusterIdentifier": "

The DB cluster identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", - "CreateDBClusterMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group to associate with this DB cluster. If this argument is omitted, default.aurora5.6 is used.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", - "CreateDBClusterMessage$DBSubnetGroupName": "

A DB subnet group to associate with this DB cluster.

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", - "CreateDBClusterMessage$Engine": "

The name of the database engine to be used for this DB cluster.

Valid Values: aurora (for MySQL 5.6-compatible Aurora), aurora-mysql (for MySQL 5.7-compatible Aurora), and aurora-postgresql

", - "CreateDBClusterMessage$EngineVersion": "

The version number of the database engine to use.

Aurora MySQL

Example: 5.6.10a, 5.7.12

Aurora PostgreSQL

Example: 9.6.3

", - "CreateDBClusterMessage$MasterUsername": "

The name of the master user for the DB cluster.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", - "CreateDBClusterMessage$MasterUserPassword": "

The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", - "CreateDBClusterMessage$OptionGroupName": "

A value that indicates that the DB cluster should be associated with the specified option group.

Permanent options can't be removed from an option group. The option group can't be removed from a DB cluster once it is associated with a DB cluster.

", - "CreateDBClusterMessage$PreferredBackupWindow": "

The daily time range during which automated backups are created if automated backups are enabled using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", - "CreateDBClusterMessage$PreferredMaintenanceWindow": "

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", - "CreateDBClusterMessage$ReplicationSourceIdentifier": "

The Amazon Resource Name (ARN) of the source DB instance or DB cluster if this DB cluster is created as a Read Replica.

", - "CreateDBClusterMessage$KmsKeyId": "

The AWS KMS key identifier for an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

If an encryption key is not specified in KmsKeyId:

  • If ReplicationSourceIdentifier identifies an encrypted source, then Amazon RDS will use the encryption key used to encrypt the source. Otherwise, Amazon RDS will use your default encryption key.

  • If the StorageEncrypted parameter is true and ReplicationSourceIdentifier is not specified, then Amazon RDS will use your default encryption key.

AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

If you create a Read Replica of an encrypted DB cluster in another AWS Region, you must set KmsKeyId to a KMS key ID that is valid in the destination AWS Region. This key is used to encrypt the Read Replica in that AWS Region.

", - "CreateDBClusterMessage$PreSignedUrl": "

A URL that contains a Signature Version 4 signed request for the CreateDBCluster action to be called in the source AWS Region where the DB cluster is replicated from. You only need to specify PreSignedUrl when you are performing cross-region replication from an encrypted DB cluster.

The pre-signed URL must be a valid request for the CreateDBCluster API action that can be executed in the source AWS Region that contains the encrypted DB cluster to be copied.

The pre-signed URL request must contain the following parameter values:

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB cluster in the destination AWS Region. This should refer to the same KMS key for both the CreateDBCluster action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

  • DestinationRegion - The name of the AWS Region that Aurora Read Replica will be created in.

  • ReplicationSourceIdentifier - The DB cluster identifier for the encrypted DB cluster to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster from the us-west-2 AWS Region, then your ReplicationSourceIdentifier would look like Example: arn:aws:rds:us-west-2:123456789012:cluster:aurora-cluster1.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", - "CreateDBClusterParameterGroupMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group.

Constraints:

  • Must match the name of an existing DBClusterParameterGroup.

This value is stored as a lowercase string.

", - "CreateDBClusterParameterGroupMessage$DBParameterGroupFamily": "

The DB cluster parameter group family name. A DB cluster parameter group can be associated with one and only one DB cluster parameter group family, and can be applied only to a DB cluster running a database engine and engine version compatible with that DB cluster parameter group family.

Aurora MySQL

Example: aurora5.6, aurora-mysql5.7

Aurora PostgreSQL

Example: aurora-postgresql9.6

", - "CreateDBClusterParameterGroupMessage$Description": "

The description for the DB cluster parameter group.

", - "CreateDBClusterSnapshotMessage$DBClusterSnapshotIdentifier": "

The identifier of the DB cluster snapshot. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1-snapshot1

", - "CreateDBClusterSnapshotMessage$DBClusterIdentifier": "

The identifier of the DB cluster to create a snapshot for. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

Example: my-cluster1

", - "CreateDBInstanceMessage$DBName": "

The meaning of this parameter differs according to the database engine you use.

Type: String

MySQL

The name of the database to create when the DB instance is created. If this parameter is not specified, no database is created in the DB instance.

Constraints:

  • Must contain 1 to 64 letters or numbers.

  • Cannot be a word reserved by the specified database engine

MariaDB

The name of the database to create when the DB instance is created. If this parameter is not specified, no database is created in the DB instance.

Constraints:

  • Must contain 1 to 64 letters or numbers.

  • Cannot be a word reserved by the specified database engine

PostgreSQL

The name of the database to create when the DB instance is created. If this parameter is not specified, the default \"postgres\" database is created in the DB instance.

Constraints:

  • Must contain 1 to 63 letters, numbers, or underscores.

  • Must begin with a letter or an underscore. Subsequent characters can be letters, underscores, or digits (0-9).

  • Cannot be a word reserved by the specified database engine

Oracle

The Oracle System ID (SID) of the created DB instance. If you specify null, the default value ORCL is used. You can't specify the string NULL, or any other reserved word, for DBName.

Default: ORCL

Constraints:

  • Cannot be longer than 8 characters

SQL Server

Not applicable. Must be null.

Amazon Aurora

The name of the database to create when the primary instance of the DB cluster is created. If this parameter is not specified, no database is created in the DB instance.

Constraints:

  • Must contain 1 to 64 letters or numbers.

  • Cannot be a word reserved by the specified database engine

", - "CreateDBInstanceMessage$DBInstanceIdentifier": "

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", - "CreateDBInstanceMessage$DBInstanceClass": "

The compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

", - "CreateDBInstanceMessage$Engine": "

The name of the database engine to be used for this instance.

Not every database engine is available for every AWS Region.

Valid Values:

  • aurora (for MySQL 5.6-compatible Aurora)

  • aurora-mysql (for MySQL 5.7-compatible Aurora)

  • aurora-postgresql

  • mariadb

  • mysql

  • oracle-ee

  • oracle-se2

  • oracle-se1

  • oracle-se

  • postgres

  • sqlserver-ee

  • sqlserver-se

  • sqlserver-ex

  • sqlserver-web

", - "CreateDBInstanceMessage$MasterUsername": "

The name for the master user.

Amazon Aurora

Not applicable. The name for the master user is managed by the DB cluster. For more information, see CreateDBCluster.

MariaDB

Constraints:

  • Required for MariaDB.

  • Must be 1 to 16 letters or numbers.

  • Cannot be a reserved word for the chosen database engine.

Microsoft SQL Server

Constraints:

  • Required for SQL Server.

  • Must be 1 to 128 letters or numbers.

  • The first character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

MySQL

Constraints:

  • Required for MySQL.

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

Oracle

Constraints:

  • Required for Oracle.

  • Must be 1 to 30 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

PostgreSQL

Constraints:

  • Required for PostgreSQL.

  • Must be 1 to 63 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", - "CreateDBInstanceMessage$MasterUserPassword": "

The password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Amazon Aurora

Not applicable. The password for the master user is managed by the DB cluster. For more information, see CreateDBCluster.

MariaDB

Constraints: Must contain from 8 to 41 characters.

Microsoft SQL Server

Constraints: Must contain from 8 to 128 characters.

MySQL

Constraints: Must contain from 8 to 41 characters.

Oracle

Constraints: Must contain from 8 to 30 characters.

PostgreSQL

Constraints: Must contain from 8 to 128 characters.

", - "CreateDBInstanceMessage$AvailabilityZone": "

The EC2 Availability Zone that the DB instance is created in. For information on AWS Regions and Availability Zones, see Regions and Availability Zones.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

Constraint: The AvailabilityZone parameter can't be specified if the MultiAZ parameter is set to true. The specified Availability Zone must be in the same AWS Region as the current endpoint.

", - "CreateDBInstanceMessage$DBSubnetGroupName": "

A DB subnet group to associate with this DB instance.

If there is no DB subnet group, then it is a non-VPC DB instance.

", - "CreateDBInstanceMessage$PreferredMaintenanceWindow": "

The time range each week during which system maintenance can occur, in Universal Coordinated Time (UTC). For more information, see Amazon RDS Maintenance Window.

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", - "CreateDBInstanceMessage$DBParameterGroupName": "

The name of the DB parameter group to associate with this DB instance. If this argument is omitted, the default DBParameterGroup for the specified engine is used.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", - "CreateDBInstanceMessage$PreferredBackupWindow": "

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter. For more information, see The Backup Window.

Amazon Aurora

Not applicable. The daily time range for creating automated backups is managed by the DB cluster. For more information, see CreateDBCluster.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred DB Instance Maintenance Window.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", - "CreateDBInstanceMessage$EngineVersion": "

The version number of the database engine to use.

The following are the database engines and major and minor versions that are available with Amazon RDS. Not every database engine is available for every AWS Region.

Amazon Aurora

Not applicable. The version number of the database engine to be used by the DB instance is managed by the DB cluster. For more information, see CreateDBCluster.

MariaDB

  • 10.2.12 (supported in all AWS Regions)

  • 10.2.11 (supported in all AWS Regions)

  • 10.1.31 (supported in all AWS Regions)

  • 10.1.26 (supported in all AWS Regions)

  • 10.1.23 (supported in all AWS Regions)

  • 10.1.19 (supported in all AWS Regions)

  • 10.1.14 (supported in all AWS Regions except us-east-2)

  • 10.0.34 (supported in all AWS Regions)

  • 10.0.32 (supported in all AWS Regions)

  • 10.0.31 (supported in all AWS Regions)

  • 10.0.28 (supported in all AWS Regions)

  • 10.0.24 (supported in all AWS Regions)

  • 10.0.17 (supported in all AWS Regions except us-east-2, ca-central-1, eu-west-2)

Microsoft SQL Server 2017

  • 14.00.1000.169.v1 (supported for all editions, and all AWS Regions)

Microsoft SQL Server 2016

  • 13.00.4451.0.v1 (supported for all editions, and all AWS Regions)

  • 13.00.4422.0.v1 (supported for all editions, and all AWS Regions)

  • 13.00.2164.0.v1 (supported for all editions, and all AWS Regions)

Microsoft SQL Server 2014

  • 12.00.5546.0.v1 (supported for all editions, and all AWS Regions)

  • 12.00.5000.0.v1 (supported for all editions, and all AWS Regions)

  • 12.00.4422.0.v1 (supported for all editions except Enterprise Edition, and all AWS Regions except ca-central-1 and eu-west-2)

Microsoft SQL Server 2012

  • 11.00.6594.0.v1 (supported for all editions, and all AWS Regions)

  • 11.00.6020.0.v1 (supported for all editions, and all AWS Regions)

  • 11.00.5058.0.v1 (supported for all editions, and all AWS Regions except us-east-2, ca-central-1, and eu-west-2)

  • 11.00.2100.60.v1 (supported for all editions, and all AWS Regions except us-east-2, ca-central-1, and eu-west-2)

Microsoft SQL Server 2008 R2

  • 10.50.6529.0.v1 (supported for all editions, and all AWS Regions except us-east-2, ca-central-1, and eu-west-2)

  • 10.50.6000.34.v1 (supported for all editions, and all AWS Regions except us-east-2, ca-central-1, and eu-west-2)

  • 10.50.2789.0.v1 (supported for all editions, and all AWS Regions except us-east-2, ca-central-1, and eu-west-2)

MySQL

  • 5.7.21 (supported in all AWS regions)

  • 5.7.19 (supported in all AWS regions)

  • 5.7.17 (supported in all AWS regions)

  • 5.7.16 (supported in all AWS regions)

  • 5.6.39 (supported in all AWS Regions)

  • 5.6.37 (supported in all AWS Regions)

  • 5.6.35 (supported in all AWS Regions)

  • 5.6.34 (supported in all AWS Regions)

  • 5.6.29 (supported in all AWS Regions)

  • 5.6.27 (supported in all AWS Regions except us-east-2, ca-central-1, eu-west-2)

  • 5.5.59 (supported in all AWS Regions)

  • 5.5.57 (supported in all AWS Regions)

  • 5.5.54 (supported in all AWS Regions)

  • 5.5.53 (supported in all AWS Regions)

  • 5.5.46 (supported in all AWS Regions)

Oracle 12c

  • 12.1.0.2.v9 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v8 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v7 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v6 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v5 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v4 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v3 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v2 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

  • 12.1.0.2.v1 (supported for EE in all AWS regions, and SE2 in all AWS regions except us-gov-west-1)

Oracle 11g

  • 11.2.0.4.v13 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v12 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v11 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v10 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v9 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v8 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v7 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v6 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v5 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v4 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v3 (supported for EE, SE1, and SE, in all AWS regions)

  • 11.2.0.4.v1 (supported for EE, SE1, and SE, in all AWS regions)

PostgreSQL

  • Version 10.1

  • Version 9.6.x: 9.6.6 | 9.6.5 | 9.6.3 | 9.6.2 | 9.6.1

  • Version 9.5.x: 9.5.9 | 9.5.7 | 9.5.6 | 9.5.4 | 9.5.2

  • Version 9.4.x: 9.4.14 | 9.4.12 | 9.4.11 | 9.4.9 | 9.4.7

  • Version 9.3.x: 9.3.19 | 9.3.17 | 9.3.16 | 9.3.14 | 9.3.12

", - "CreateDBInstanceMessage$LicenseModel": "

License model information for this DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", - "CreateDBInstanceMessage$OptionGroupName": "

Indicates that the DB instance should be associated with the specified option group.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", - "CreateDBInstanceMessage$CharacterSetName": "

For supported engines, indicates that the DB instance should be associated with the specified CharacterSet.

Amazon Aurora

Not applicable. The character set is managed by the DB cluster. For more information, see CreateDBCluster.

", - "CreateDBInstanceMessage$DBClusterIdentifier": "

The identifier of the DB cluster that the instance will belong to.

For information on creating a DB cluster, see CreateDBCluster.

Type: String

", - "CreateDBInstanceMessage$StorageType": "

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", - "CreateDBInstanceMessage$TdeCredentialArn": "

The ARN from the key store with which to associate the instance for TDE encryption.

", - "CreateDBInstanceMessage$TdeCredentialPassword": "

The password for the given ARN from the key store in order to access the device.

", - "CreateDBInstanceMessage$KmsKeyId": "

The AWS KMS key identifier for an encrypted DB instance.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB instance with the same AWS account that owns the KMS encryption key used to encrypt the new DB instance, then you can use the KMS key alias instead of the ARN for the KM encryption key.

Amazon Aurora

Not applicable. The KMS key identifier is managed by the DB cluster. For more information, see CreateDBCluster.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon RDS will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", - "CreateDBInstanceMessage$Domain": "

Specify the Active Directory Domain to create the instance in.

", - "CreateDBInstanceMessage$MonitoringRoleArn": "

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, go to Setting Up and Enabling Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", - "CreateDBInstanceMessage$DomainIAMRoleName": "

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", - "CreateDBInstanceMessage$Timezone": "

The time zone of the DB instance. The time zone parameter is currently supported only by Microsoft SQL Server.

", - "CreateDBInstanceMessage$PerformanceInsightsKMSKeyId": "

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", - "CreateDBInstanceReadReplicaMessage$DBInstanceIdentifier": "

The DB instance identifier of the Read Replica. This identifier is the unique key that identifies a DB instance. This parameter is stored as a lowercase string.

", - "CreateDBInstanceReadReplicaMessage$SourceDBInstanceIdentifier": "

The identifier of the DB instance that will act as the source for the Read Replica. Each DB instance can have up to five Read Replicas.

Constraints:

  • Must be the identifier of an existing MySQL, MariaDB, or PostgreSQL DB instance.

  • Can specify a DB instance that is a MySQL Read Replica only if the source is running MySQL 5.6.

  • Can specify a DB instance that is a PostgreSQL DB instance only if the source is running PostgreSQL 9.3.5 or later (9.4.7 and higher for cross-region replication).

  • The specified DB instance must have automatic backups enabled, its backup retention period must be greater than 0.

  • If the source DB instance is in the same AWS Region as the Read Replica, specify a valid DB instance identifier.

  • If the source DB instance is in a different AWS Region than the Read Replica, specify a valid DB instance ARN. For more information, go to Constructing a Amazon RDS Amazon Resource Name (ARN).

", - "CreateDBInstanceReadReplicaMessage$DBInstanceClass": "

The compute and memory capacity of the Read Replica, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Default: Inherits from the source DB instance.

", - "CreateDBInstanceReadReplicaMessage$AvailabilityZone": "

The Amazon EC2 Availability Zone that the Read Replica is created in.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

", - "CreateDBInstanceReadReplicaMessage$OptionGroupName": "

The option group the DB instance is associated with. If omitted, the default option group for the engine specified is used.

", - "CreateDBInstanceReadReplicaMessage$DBSubnetGroupName": "

Specifies a DB subnet group for the DB instance. The new DB instance is created in the VPC associated with the DB subnet group. If no DB subnet group is specified, then the new DB instance is not created in a VPC.

Constraints:

  • Can only be specified if the source DB instance identifier specifies a DB instance in another AWS Region.

  • If supplied, must match the name of an existing DBSubnetGroup.

  • The specified DB subnet group must be in the same AWS Region in which the operation is running.

  • All Read Replicas in one AWS Region that are created from the same source DB instance must either:>

    • Specify DB subnet groups from the same VPC. All these Read Replicas are created in the same VPC.

    • Not specify a DB subnet group. All these Read Replicas are created outside of any VPC.

Example: mySubnetgroup

", - "CreateDBInstanceReadReplicaMessage$StorageType": "

Specifies the storage type to be associated with the Read Replica.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", - "CreateDBInstanceReadReplicaMessage$MonitoringRoleArn": "

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, go to To create an IAM role for Amazon RDS Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", - "CreateDBInstanceReadReplicaMessage$KmsKeyId": "

The AWS KMS key ID for an encrypted Read Replica. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you specify this parameter when you create a Read Replica from an unencrypted DB instance, the Read Replica is encrypted.

If you create an encrypted Read Replica in the same AWS Region as the source DB instance, then you do not have to specify a value for this parameter. The Read Replica is encrypted with the same KMS key as the source DB instance.

If you create an encrypted Read Replica in a different AWS Region, then you must specify a KMS key for the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

", - "CreateDBInstanceReadReplicaMessage$PreSignedUrl": "

The URL that contains a Signature Version 4 signed request for the CreateDBInstanceReadReplica API action in the source AWS Region that contains the source DB instance.

You must specify this parameter when you create an encrypted Read Replica from another AWS Region by using the Amazon RDS API. You can specify the --source-region option instead of this parameter when you create an encrypted Read Replica from another AWS Region by using the AWS CLI.

The presigned URL must be a valid request for the CreateDBInstanceReadReplica API action that can be executed in the source AWS Region that contains the encrypted source DB instance. The presigned URL request must contain the following parameter values:

  • DestinationRegion - The AWS Region that the encrypted Read Replica is created in. This AWS Region is the same one where the CreateDBInstanceReadReplica action is called that contains this presigned URL.

    For example, if you create an encrypted DB instance in the us-west-1 AWS Region, from a source DB instance in the us-east-2 AWS Region, then you call the CreateDBInstanceReadReplica action in the us-east-1 AWS Region and provide a presigned URL that contains a call to the CreateDBInstanceReadReplica action in the us-west-2 AWS Region. For this example, the DestinationRegion in the presigned URL must be set to the us-east-1 AWS Region.

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the Read Replica in the destination AWS Region. This is the same identifier for both the CreateDBInstanceReadReplica action that is called in the destination AWS Region, and the action contained in the presigned URL.

  • SourceDBInstanceIdentifier - The DB instance identifier for the encrypted DB instance to be replicated. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are creating an encrypted Read Replica from a DB instance in the us-west-2 AWS Region, then your SourceDBInstanceIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:instance:mysql-instance1-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", - "CreateDBInstanceReadReplicaMessage$PerformanceInsightsKMSKeyId": "

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", - "CreateDBParameterGroupMessage$DBParameterGroupName": "

The name of the DB parameter group.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

This value is stored as a lowercase string.

", - "CreateDBParameterGroupMessage$DBParameterGroupFamily": "

The DB parameter group family name. A DB parameter group can be associated with one and only one DB parameter group family, and can be applied only to a DB instance running a database engine and engine version compatible with that DB parameter group family.

", - "CreateDBParameterGroupMessage$Description": "

The description for the DB parameter group.

", - "CreateDBSecurityGroupMessage$DBSecurityGroupName": "

The name for the DB security group. This value is stored as a lowercase string.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Must not be \"Default\"

Example: mysecuritygroup

", - "CreateDBSecurityGroupMessage$DBSecurityGroupDescription": "

The description for the DB security group.

", - "CreateDBSnapshotMessage$DBSnapshotIdentifier": "

The identifier for the DB snapshot.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", - "CreateDBSnapshotMessage$DBInstanceIdentifier": "

The identifier of the DB instance that you want to create the snapshot of.

Constraints:

  • Must match the identifier of an existing DBInstance.

", - "CreateDBSubnetGroupMessage$DBSubnetGroupName": "

The name for the DB subnet group. This value is stored as a lowercase string.

Constraints: Must contain no more than 255 letters, numbers, periods, underscores, spaces, or hyphens. Must not be default.

Example: mySubnetgroup

", - "CreateDBSubnetGroupMessage$DBSubnetGroupDescription": "

The description for the DB subnet group.

", - "CreateEventSubscriptionMessage$SubscriptionName": "

The name of the subscription.

Constraints: The name must be less than 255 characters.

", - "CreateEventSubscriptionMessage$SnsTopicArn": "

The Amazon Resource Name (ARN) of the SNS topic created for event notification. The ARN is created by Amazon SNS when you create a topic and subscribe to it.

", - "CreateEventSubscriptionMessage$SourceType": "

The type of source that is generating the events. For example, if you want to be notified of events generated by a DB instance, you would set this parameter to db-instance. if this value is not specified, all events are returned.

Valid values: db-instance | db-cluster | db-parameter-group | db-security-group | db-snapshot | db-cluster-snapshot

", - "CreateOptionGroupMessage$OptionGroupName": "

Specifies the name of the option group to be created.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: myoptiongroup

", - "CreateOptionGroupMessage$EngineName": "

Specifies the name of the engine that this option group should be associated with.

", - "CreateOptionGroupMessage$MajorEngineVersion": "

Specifies the major version of the engine that this option group should be associated with.

", - "CreateOptionGroupMessage$OptionGroupDescription": "

The description of the option group.

", - "DBCluster$CharacterSetName": "

If present, specifies the name of the character set that this cluster is associated with.

", - "DBCluster$DatabaseName": "

Contains the name of the initial database of this DB cluster that was provided at create time, if one was specified when the DB cluster was created. This same name is returned for the life of the DB cluster.

", - "DBCluster$DBClusterIdentifier": "

Contains a user-supplied DB cluster identifier. This identifier is the unique key that identifies a DB cluster.

", - "DBCluster$DBClusterParameterGroup": "

Specifies the name of the DB cluster parameter group for the DB cluster.

", - "DBCluster$DBSubnetGroup": "

Specifies information on the subnet group associated with the DB cluster, including the name, description, and subnets in the subnet group.

", - "DBCluster$Status": "

Specifies the current state of this DB cluster.

", - "DBCluster$PercentProgress": "

Specifies the progress of the operation as a percentage.

", - "DBCluster$Endpoint": "

Specifies the connection endpoint for the primary instance of the DB cluster.

", - "DBCluster$ReaderEndpoint": "

The reader endpoint for the DB cluster. The reader endpoint for a DB cluster load-balances connections across the Aurora Replicas that are available in a DB cluster. As clients request new connections to the reader endpoint, Aurora distributes the connection requests among the Aurora Replicas in the DB cluster. This functionality can help balance your read workload across multiple Aurora Replicas in your DB cluster.

If a failover occurs, and the Aurora Replica that you are connected to is promoted to be the primary instance, your connection is dropped. To continue sending your read workload to other Aurora Replicas in the cluster, you can then reconnect to the reader endpoint.

", - "DBCluster$Engine": "

Provides the name of the database engine to be used for this DB cluster.

", - "DBCluster$EngineVersion": "

Indicates the database engine version.

", - "DBCluster$MasterUsername": "

Contains the master username for the DB cluster.

", - "DBCluster$PreferredBackupWindow": "

Specifies the daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod.

", - "DBCluster$PreferredMaintenanceWindow": "

Specifies the weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

", - "DBCluster$ReplicationSourceIdentifier": "

Contains the identifier of the source DB cluster if this DB cluster is a Read Replica.

", - "DBCluster$HostedZoneId": "

Specifies the ID that Amazon Route 53 assigns when you create a hosted zone.

", - "DBCluster$KmsKeyId": "

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB cluster.

", - "DBCluster$DbClusterResourceId": "

The AWS Region-unique, immutable identifier for the DB cluster. This identifier is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB cluster is accessed.

", - "DBCluster$DBClusterArn": "

The Amazon Resource Name (ARN) for the DB cluster.

", - "DBCluster$CloneGroupId": "

Identifies the clone group to which the DB cluster is associated.

", - "DBClusterBacktrack$DBClusterIdentifier": "

Contains a user-supplied DB cluster identifier. This identifier is the unique key that identifies a DB cluster.

", - "DBClusterBacktrack$BacktrackIdentifier": "

Contains the backtrack identifier.

", - "DBClusterBacktrack$Status": "

The status of the backtrack. This property returns one of the following values:

  • applying - The backtrack is currently being applied to or rolled back from the DB cluster.

  • completed - The backtrack has successfully been applied to or rolled back from the DB cluster.

  • failed - An error occurred while the backtrack was applied to or rolled back from the DB cluster.

  • pending - The backtrack is currently pending application to or rollback from the DB cluster.

", - "DBClusterBacktrackMessage$Marker": "

A pagination token that can be used in a subsequent DescribeDBClusterBacktracks request.

", - "DBClusterMember$DBInstanceIdentifier": "

Specifies the instance identifier for this member of the DB cluster.

", - "DBClusterMember$DBClusterParameterGroupStatus": "

Specifies the status of the DB cluster parameter group for this member of the DB cluster.

", - "DBClusterMessage$Marker": "

A pagination token that can be used in a subsequent DescribeDBClusters request.

", - "DBClusterOptionGroupStatus$DBClusterOptionGroupName": "

Specifies the name of the DB cluster option group.

", - "DBClusterOptionGroupStatus$Status": "

Specifies the status of the DB cluster option group.

", - "DBClusterParameterGroup$DBClusterParameterGroupName": "

Provides the name of the DB cluster parameter group.

", - "DBClusterParameterGroup$DBParameterGroupFamily": "

Provides the name of the DB parameter group family that this DB cluster parameter group is compatible with.

", - "DBClusterParameterGroup$Description": "

Provides the customer-specified description for this DB cluster parameter group.

", - "DBClusterParameterGroup$DBClusterParameterGroupArn": "

The Amazon Resource Name (ARN) for the DB cluster parameter group.

", - "DBClusterParameterGroupDetails$Marker": "

An optional pagination token provided by a previous DescribeDBClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "DBClusterParameterGroupNameMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group.

Constraints:

  • Must be 1 to 255 letters or numbers.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

This value is stored as a lowercase string.

", - "DBClusterParameterGroupsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusterParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBClusterRole$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role that is associated with the DB cluster.

", - "DBClusterRole$Status": "

Describes the state of association between the IAM role and the DB cluster. The Status property returns one of the following values:

  • ACTIVE - the IAM role ARN is associated with the DB cluster and can be used to access other AWS services on your behalf.

  • PENDING - the IAM role ARN is being associated with the DB cluster.

  • INVALID - the IAM role ARN is associated with the DB cluster, but the DB cluster is unable to assume the IAM role in order to access other AWS services on your behalf.

", - "DBClusterSnapshot$DBClusterSnapshotIdentifier": "

Specifies the identifier for the DB cluster snapshot.

", - "DBClusterSnapshot$DBClusterIdentifier": "

Specifies the DB cluster identifier of the DB cluster that this DB cluster snapshot was created from.

", - "DBClusterSnapshot$Engine": "

Specifies the name of the database engine.

", - "DBClusterSnapshot$Status": "

Specifies the status of this DB cluster snapshot.

", - "DBClusterSnapshot$VpcId": "

Provides the VPC ID associated with the DB cluster snapshot.

", - "DBClusterSnapshot$MasterUsername": "

Provides the master username for the DB cluster snapshot.

", - "DBClusterSnapshot$EngineVersion": "

Provides the version of the database engine for this DB cluster snapshot.

", - "DBClusterSnapshot$LicenseModel": "

Provides the license model information for this DB cluster snapshot.

", - "DBClusterSnapshot$SnapshotType": "

Provides the type of the DB cluster snapshot.

", - "DBClusterSnapshot$KmsKeyId": "

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB cluster snapshot.

", - "DBClusterSnapshot$DBClusterSnapshotArn": "

The Amazon Resource Name (ARN) for the DB cluster snapshot.

", - "DBClusterSnapshot$SourceDBClusterSnapshotArn": "

If the DB cluster snapshot was copied from a source DB cluster snapshot, the Amazon Resource Name (ARN) for the source DB cluster snapshot, otherwise, a null value.

", - "DBClusterSnapshotAttribute$AttributeName": "

The name of the manual DB cluster snapshot attribute.

The attribute named restore refers to the list of AWS accounts that have permission to copy or restore the manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", - "DBClusterSnapshotAttributesResult$DBClusterSnapshotIdentifier": "

The identifier of the manual DB cluster snapshot that the attributes apply to.

", - "DBClusterSnapshotMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusterSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBEngineVersion$Engine": "

The name of the database engine.

", - "DBEngineVersion$EngineVersion": "

The version number of the database engine.

", - "DBEngineVersion$DBParameterGroupFamily": "

The name of the DB parameter group family for the database engine.

", - "DBEngineVersion$DBEngineDescription": "

The description of the database engine.

", - "DBEngineVersion$DBEngineVersionDescription": "

The description of the database engine version.

", - "DBEngineVersionMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBInstance$DBInstanceIdentifier": "

Contains a user-supplied database identifier. This identifier is the unique key that identifies a DB instance.

", - "DBInstance$DBInstanceClass": "

Contains the name of the compute and memory capacity class of the DB instance.

", - "DBInstance$Engine": "

Provides the name of the database engine to be used for this DB instance.

", - "DBInstance$DBInstanceStatus": "

Specifies the current state of this database.

", - "DBInstance$MasterUsername": "

Contains the master username for the DB instance.

", - "DBInstance$DBName": "

The meaning of this parameter differs according to the database engine you use. For example, this value returns MySQL, MariaDB, or PostgreSQL information when returning values from CreateDBInstanceReadReplica since Read Replicas are only supported for these engines.

MySQL, MariaDB, SQL Server, PostgreSQL

Contains the name of the initial database of this instance that was provided at create time, if one was specified when the DB instance was created. This same name is returned for the life of the DB instance.

Type: String

Oracle

Contains the Oracle System ID (SID) of the created DB instance. Not shown when the returned parameters do not apply to an Oracle DB instance.

", - "DBInstance$PreferredBackupWindow": "

Specifies the daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod.

", - "DBInstance$AvailabilityZone": "

Specifies the name of the Availability Zone the DB instance is located in.

", - "DBInstance$PreferredMaintenanceWindow": "

Specifies the weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

", - "DBInstance$EngineVersion": "

Indicates the database engine version.

", - "DBInstance$ReadReplicaSourceDBInstanceIdentifier": "

Contains the identifier of the source DB instance if this DB instance is a Read Replica.

", - "DBInstance$LicenseModel": "

License model information for this DB instance.

", - "DBInstance$CharacterSetName": "

If present, specifies the name of the character set that this instance is associated with.

", - "DBInstance$SecondaryAvailabilityZone": "

If present, specifies the name of the secondary Availability Zone for a DB instance with multi-AZ support.

", - "DBInstance$StorageType": "

Specifies the storage type associated with DB instance.

", - "DBInstance$TdeCredentialArn": "

The ARN from the key store with which the instance is associated for TDE encryption.

", - "DBInstance$DBClusterIdentifier": "

If the DB instance is a member of a DB cluster, contains the name of the DB cluster that the DB instance is a member of.

", - "DBInstance$KmsKeyId": "

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB instance.

", - "DBInstance$DbiResourceId": "

The AWS Region-unique, immutable identifier for the DB instance. This identifier is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB instance is accessed.

", - "DBInstance$CACertificateIdentifier": "

The identifier of the CA certificate for this DB instance.

", - "DBInstance$EnhancedMonitoringResourceArn": "

The Amazon Resource Name (ARN) of the Amazon CloudWatch Logs log stream that receives the Enhanced Monitoring metrics data for the DB instance.

", - "DBInstance$MonitoringRoleArn": "

The ARN for the IAM role that permits RDS to send Enhanced Monitoring metrics to Amazon CloudWatch Logs.

", - "DBInstance$DBInstanceArn": "

The Amazon Resource Name (ARN) for the DB instance.

", - "DBInstance$Timezone": "

The time zone of the DB instance. In most cases, the Timezone element is empty. Timezone content appears only for Microsoft SQL Server DB instances that were created with a time zone specified.

", - "DBInstance$PerformanceInsightsKMSKeyId": "

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", - "DBInstanceMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "DBInstanceStatusInfo$StatusType": "

This value is currently \"read replication.\"

", - "DBInstanceStatusInfo$Status": "

Status of the DB instance. For a StatusType of read replica, the values can be replicating, error, stopped, or terminated.

", - "DBInstanceStatusInfo$Message": "

Details of the error if there is an error for the instance. If the instance is not in an error state, this value is blank.

", - "DBParameterGroup$DBParameterGroupName": "

Provides the name of the DB parameter group.

", - "DBParameterGroup$DBParameterGroupFamily": "

Provides the name of the DB parameter group family that this DB parameter group is compatible with.

", - "DBParameterGroup$Description": "

Provides the customer-specified description for this DB parameter group.

", - "DBParameterGroup$DBParameterGroupArn": "

The Amazon Resource Name (ARN) for the DB parameter group.

", - "DBParameterGroupDetails$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBParameterGroupNameMessage$DBParameterGroupName": "

Provides the name of the DB parameter group.

", - "DBParameterGroupStatus$DBParameterGroupName": "

The name of the DP parameter group.

", - "DBParameterGroupStatus$ParameterApplyStatus": "

The status of parameter updates.

", - "DBParameterGroupsMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBSecurityGroup$OwnerId": "

Provides the AWS ID of the owner of a specific DB security group.

", - "DBSecurityGroup$DBSecurityGroupName": "

Specifies the name of the DB security group.

", - "DBSecurityGroup$DBSecurityGroupDescription": "

Provides the description of the DB security group.

", - "DBSecurityGroup$VpcId": "

Provides the VpcId of the DB security group.

", - "DBSecurityGroup$DBSecurityGroupArn": "

The Amazon Resource Name (ARN) for the DB security group.

", - "DBSecurityGroupMembership$DBSecurityGroupName": "

The name of the DB security group.

", - "DBSecurityGroupMembership$Status": "

The status of the DB security group.

", - "DBSecurityGroupMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBSecurityGroupNameList$member": null, - "DBSnapshot$DBSnapshotIdentifier": "

Specifies the identifier for the DB snapshot.

", - "DBSnapshot$DBInstanceIdentifier": "

Specifies the DB instance identifier of the DB instance this DB snapshot was created from.

", - "DBSnapshot$Engine": "

Specifies the name of the database engine.

", - "DBSnapshot$Status": "

Specifies the status of this DB snapshot.

", - "DBSnapshot$AvailabilityZone": "

Specifies the name of the Availability Zone the DB instance was located in at the time of the DB snapshot.

", - "DBSnapshot$VpcId": "

Provides the VPC ID associated with the DB snapshot.

", - "DBSnapshot$MasterUsername": "

Provides the master username for the DB snapshot.

", - "DBSnapshot$EngineVersion": "

Specifies the version of the database engine.

", - "DBSnapshot$LicenseModel": "

License model information for the restored DB instance.

", - "DBSnapshot$SnapshotType": "

Provides the type of the DB snapshot.

", - "DBSnapshot$OptionGroupName": "

Provides the option group name for the DB snapshot.

", - "DBSnapshot$SourceRegion": "

The AWS Region that the DB snapshot was created in or copied from.

", - "DBSnapshot$SourceDBSnapshotIdentifier": "

The DB snapshot Amazon Resource Name (ARN) that the DB snapshot was copied from. It only has value in case of cross-customer or cross-region copy.

", - "DBSnapshot$StorageType": "

Specifies the storage type associated with DB snapshot.

", - "DBSnapshot$TdeCredentialArn": "

The ARN from the key store with which to associate the instance for TDE encryption.

", - "DBSnapshot$KmsKeyId": "

If Encrypted is true, the AWS KMS key identifier for the encrypted DB snapshot.

", - "DBSnapshot$DBSnapshotArn": "

The Amazon Resource Name (ARN) for the DB snapshot.

", - "DBSnapshot$Timezone": "

The time zone of the DB snapshot. In most cases, the Timezone element is empty. Timezone content appears only for snapshots taken from Microsoft SQL Server DB instances that were created with a time zone specified.

", - "DBSnapshotAttribute$AttributeName": "

The name of the manual DB snapshot attribute.

The attribute named restore refers to the list of AWS accounts that have permission to copy or restore the manual DB cluster snapshot. For more information, see the ModifyDBSnapshotAttribute API action.

", - "DBSnapshotAttributesResult$DBSnapshotIdentifier": "

The identifier of the manual DB snapshot that the attributes apply to.

", - "DBSnapshotMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DBSubnetGroup$DBSubnetGroupName": "

The name of the DB subnet group.

", - "DBSubnetGroup$DBSubnetGroupDescription": "

Provides the description of the DB subnet group.

", - "DBSubnetGroup$VpcId": "

Provides the VpcId of the DB subnet group.

", - "DBSubnetGroup$SubnetGroupStatus": "

Provides the status of the DB subnet group.

", - "DBSubnetGroup$DBSubnetGroupArn": "

The Amazon Resource Name (ARN) for the DB subnet group.

", - "DBSubnetGroupMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DeleteDBClusterMessage$DBClusterIdentifier": "

The DB cluster identifier for the DB cluster to be deleted. This parameter isn't case-sensitive.

Constraints:

  • Must match an existing DBClusterIdentifier.

", - "DeleteDBClusterMessage$FinalDBSnapshotIdentifier": "

The DB cluster snapshot identifier of the new DB cluster snapshot created when SkipFinalSnapshot is set to false.

Specifying this parameter and also setting the SkipFinalShapshot parameter to true results in an error.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", - "DeleteDBClusterParameterGroupMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group.

Constraints:

  • Must be the name of an existing DB cluster parameter group.

  • You can't delete a default DB cluster parameter group.

  • Cannot be associated with any DB clusters.

", - "DeleteDBClusterSnapshotMessage$DBClusterSnapshotIdentifier": "

The identifier of the DB cluster snapshot to delete.

Constraints: Must be the name of an existing DB cluster snapshot in the available state.

", - "DeleteDBInstanceMessage$DBInstanceIdentifier": "

The DB instance identifier for the DB instance to be deleted. This parameter isn't case-sensitive.

Constraints:

  • Must match the name of an existing DB instance.

", - "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier": "

The DBSnapshotIdentifier of the new DBSnapshot created when SkipFinalSnapshot is set to false.

Specifying this parameter and also setting the SkipFinalShapshot parameter to true results in an error.

Constraints:

  • Must be 1 to 255 letters or numbers.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Cannot be specified when deleting a Read Replica.

", - "DeleteDBParameterGroupMessage$DBParameterGroupName": "

The name of the DB parameter group.

Constraints:

  • Must be the name of an existing DB parameter group

  • You can't delete a default DB parameter group

  • Cannot be associated with any DB instances

", - "DeleteDBSecurityGroupMessage$DBSecurityGroupName": "

The name of the DB security group to delete.

You can't delete the default DB security group.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Must not be \"Default\"

", - "DeleteDBSnapshotMessage$DBSnapshotIdentifier": "

The DBSnapshot identifier.

Constraints: Must be the name of an existing DB snapshot in the available state.

", - "DeleteDBSubnetGroupMessage$DBSubnetGroupName": "

The name of the database subnet group to delete.

You can't delete the default subnet group.

Constraints:

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", - "DeleteEventSubscriptionMessage$SubscriptionName": "

The name of the RDS event notification subscription you want to delete.

", - "DeleteOptionGroupMessage$OptionGroupName": "

The name of the option group to be deleted.

You can't delete default option groups.

", - "DescribeCertificatesMessage$CertificateIdentifier": "

The user-supplied certificate identifier. If this parameter is specified, information for only the identified certificate is returned. This parameter isn't case-sensitive.

Constraints:

  • Must match an existing CertificateIdentifier.

", - "DescribeCertificatesMessage$Marker": "

An optional pagination token provided by a previous DescribeCertificates request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBClusterBacktracksMessage$DBClusterIdentifier": "

The DB cluster identifier of the DB cluster to be described. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 alphanumeric characters or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", - "DescribeDBClusterBacktracksMessage$BacktrackIdentifier": "

If specified, this value is the backtrack identifier of the backtrack to be described.

Constraints:

Example: 123e4567-e89b-12d3-a456-426655440000

", - "DescribeDBClusterBacktracksMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusterBacktracks request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBClusterParameterGroupsMessage$DBClusterParameterGroupName": "

The name of a specific DB cluster parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", - "DescribeDBClusterParameterGroupsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusterParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBClusterParametersMessage$DBClusterParameterGroupName": "

The name of a specific DB cluster parameter group to return parameter details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", - "DescribeDBClusterParametersMessage$Source": "

A value that indicates to return only parameters for a specific source. Parameter sources can be engine, service, or customer.

", - "DescribeDBClusterParametersMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBClusterSnapshotAttributesMessage$DBClusterSnapshotIdentifier": "

The identifier for the DB cluster snapshot to describe the attributes for.

", - "DescribeDBClusterSnapshotsMessage$DBClusterIdentifier": "

The ID of the DB cluster to retrieve the list of DB cluster snapshots for. This parameter can't be used in conjunction with the DBClusterSnapshotIdentifier parameter. This parameter is not case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBCluster.

", - "DescribeDBClusterSnapshotsMessage$DBClusterSnapshotIdentifier": "

A specific DB cluster snapshot identifier to describe. This parameter can't be used in conjunction with the DBClusterIdentifier parameter. This value is stored as a lowercase string.

Constraints:

  • If supplied, must match the identifier of an existing DBClusterSnapshot.

  • If this identifier is for an automated snapshot, the SnapshotType parameter must also be specified.

", - "DescribeDBClusterSnapshotsMessage$SnapshotType": "

The type of DB cluster snapshots to be returned. You can specify one of the following values:

  • automated - Return all DB cluster snapshots that have been automatically taken by Amazon RDS for my AWS account.

  • manual - Return all DB cluster snapshots that have been taken by my AWS account.

  • shared - Return all manual DB cluster snapshots that have been shared to my AWS account.

  • public - Return all DB cluster snapshots that have been marked as public.

If you don't specify a SnapshotType value, then both automated and manual DB cluster snapshots are returned. You can include shared DB cluster snapshots with these results by setting the IncludeShared parameter to true. You can include public DB cluster snapshots with these results by setting the IncludePublic parameter to true.

The IncludeShared and IncludePublic parameters don't apply for SnapshotType values of manual or automated. The IncludePublic parameter doesn't apply when SnapshotType is set to shared. The IncludeShared parameter doesn't apply when SnapshotType is set to public.

", - "DescribeDBClusterSnapshotsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusterSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBClustersMessage$DBClusterIdentifier": "

The user-supplied DB cluster identifier. If this parameter is specified, information from only the specific DB cluster is returned. This parameter isn't case-sensitive.

Constraints:

  • If supplied, must match an existing DBClusterIdentifier.

", - "DescribeDBClustersMessage$Marker": "

An optional pagination token provided by a previous DescribeDBClusters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBEngineVersionsMessage$Engine": "

The database engine to return.

", - "DescribeDBEngineVersionsMessage$EngineVersion": "

The database engine version to return.

Example: 5.1.49

", - "DescribeDBEngineVersionsMessage$DBParameterGroupFamily": "

The name of a specific DB parameter group family to return details for.

Constraints:

  • If supplied, must match an existing DBParameterGroupFamily.

", - "DescribeDBEngineVersionsMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBInstancesMessage$DBInstanceIdentifier": "

The user-supplied instance identifier. If this parameter is specified, information from only the specific DB instance is returned. This parameter isn't case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBInstance.

", - "DescribeDBInstancesMessage$Marker": "

An optional pagination token provided by a previous DescribeDBInstances request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBLogFilesDetails$LogFileName": "

The name of the log file for the specified DB instance.

", - "DescribeDBLogFilesMessage$DBInstanceIdentifier": "

The customer-assigned name of the DB instance that contains the log files you want to list.

Constraints:

  • Must match the identifier of an existing DBInstance.

", - "DescribeDBLogFilesMessage$FilenameContains": "

Filters the available log files for log file names that contain the specified string.

", - "DescribeDBLogFilesMessage$Marker": "

The pagination token provided in the previous request. If this parameter is specified the response includes only records beyond the marker, up to MaxRecords.

", - "DescribeDBLogFilesResponse$Marker": "

A pagination token that can be used in a subsequent DescribeDBLogFiles request.

", - "DescribeDBParameterGroupsMessage$DBParameterGroupName": "

The name of a specific DB parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", - "DescribeDBParameterGroupsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBParametersMessage$DBParameterGroupName": "

The name of a specific DB parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBParameterGroup.

", - "DescribeDBParametersMessage$Source": "

The parameter types to return.

Default: All parameter types returned

Valid Values: user | system | engine-default

", - "DescribeDBParametersMessage$Marker": "

An optional pagination token provided by a previous DescribeDBParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBSecurityGroupsMessage$DBSecurityGroupName": "

The name of the DB security group to return details for.

", - "DescribeDBSecurityGroupsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBSecurityGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBSnapshotAttributesMessage$DBSnapshotIdentifier": "

The identifier for the DB snapshot to describe the attributes for.

", - "DescribeDBSnapshotsMessage$DBInstanceIdentifier": "

The ID of the DB instance to retrieve the list of DB snapshots for. This parameter can't be used in conjunction with DBSnapshotIdentifier. This parameter is not case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBInstance.

", - "DescribeDBSnapshotsMessage$DBSnapshotIdentifier": "

A specific DB snapshot identifier to describe. This parameter can't be used in conjunction with DBInstanceIdentifier. This value is stored as a lowercase string.

Constraints:

  • If supplied, must match the identifier of an existing DBSnapshot.

  • If this identifier is for an automated snapshot, the SnapshotType parameter must also be specified.

", - "DescribeDBSnapshotsMessage$SnapshotType": "

The type of snapshots to be returned. You can specify one of the following values:

  • automated - Return all DB snapshots that have been automatically taken by Amazon RDS for my AWS account.

  • manual - Return all DB snapshots that have been taken by my AWS account.

  • shared - Return all manual DB snapshots that have been shared to my AWS account.

  • public - Return all DB snapshots that have been marked as public.

If you don't specify a SnapshotType value, then both automated and manual snapshots are returned. Shared and public DB snapshots are not included in the returned results by default. You can include shared snapshots with these results by setting the IncludeShared parameter to true. You can include public snapshots with these results by setting the IncludePublic parameter to true.

The IncludeShared and IncludePublic parameters don't apply for SnapshotType values of manual or automated. The IncludePublic parameter doesn't apply when SnapshotType is set to shared. The IncludeShared parameter doesn't apply when SnapshotType is set to public.

", - "DescribeDBSnapshotsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeDBSubnetGroupsMessage$DBSubnetGroupName": "

The name of the DB subnet group to return details for.

", - "DescribeDBSubnetGroupsMessage$Marker": "

An optional pagination token provided by a previous DescribeDBSubnetGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeEngineDefaultClusterParametersMessage$DBParameterGroupFamily": "

The name of the DB cluster parameter group family to return engine parameter information for.

", - "DescribeEngineDefaultClusterParametersMessage$Marker": "

An optional pagination token provided by a previous DescribeEngineDefaultClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily": "

The name of the DB parameter group family.

", - "DescribeEngineDefaultParametersMessage$Marker": "

An optional pagination token provided by a previous DescribeEngineDefaultParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeEventCategoriesMessage$SourceType": "

The type of source that is generating the events.

Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot

", - "DescribeEventSubscriptionsMessage$SubscriptionName": "

The name of the RDS event notification subscription you want to describe.

", - "DescribeEventSubscriptionsMessage$Marker": "

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "DescribeEventsMessage$SourceIdentifier": "

The identifier of the event source for which events are returned. If not specified, then all sources are included in the response.

Constraints:

  • If SourceIdentifier is supplied, SourceType must also be provided.

  • If the source type is DBInstance, then a DBInstanceIdentifier must be supplied.

  • If the source type is DBSecurityGroup, a DBSecurityGroupName must be supplied.

  • If the source type is DBParameterGroup, a DBParameterGroupName must be supplied.

  • If the source type is DBSnapshot, a DBSnapshotIdentifier must be supplied.

  • Cannot end with a hyphen or contain two consecutive hyphens.

", - "DescribeEventsMessage$Marker": "

An optional pagination token provided by a previous DescribeEvents request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeOptionGroupOptionsMessage$EngineName": "

A required parameter. Options available for the given engine name are described.

", - "DescribeOptionGroupOptionsMessage$MajorEngineVersion": "

If specified, filters the results to include only options for the specified major engine version.

", - "DescribeOptionGroupOptionsMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeOptionGroupsMessage$OptionGroupName": "

The name of the option group to describe. Cannot be supplied together with EngineName or MajorEngineVersion.

", - "DescribeOptionGroupsMessage$Marker": "

An optional pagination token provided by a previous DescribeOptionGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeOptionGroupsMessage$EngineName": "

Filters the list of option groups to only include groups associated with a specific database engine.

", - "DescribeOptionGroupsMessage$MajorEngineVersion": "

Filters the list of option groups to only include groups associated with a specific database engine version. If specified, then EngineName must also be specified.

", - "DescribeOrderableDBInstanceOptionsMessage$Engine": "

The name of the engine to retrieve DB instance options for.

", - "DescribeOrderableDBInstanceOptionsMessage$EngineVersion": "

The engine version filter value. Specify this parameter to show only the available offerings matching the specified engine version.

", - "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass": "

The DB instance class filter value. Specify this parameter to show only the available offerings matching the specified DB instance class.

", - "DescribeOrderableDBInstanceOptionsMessage$LicenseModel": "

The license model filter value. Specify this parameter to show only the available offerings matching the specified license model.

", - "DescribeOrderableDBInstanceOptionsMessage$Marker": "

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "DescribePendingMaintenanceActionsMessage$ResourceIdentifier": "

The ARN of a resource to return pending maintenance actions for.

", - "DescribePendingMaintenanceActionsMessage$Marker": "

An optional pagination token provided by a previous DescribePendingMaintenanceActions request. If this parameter is specified, the response includes only records beyond the marker, up to a number of records specified by MaxRecords.

", - "DescribeReservedDBInstancesMessage$ReservedDBInstanceId": "

The reserved DB instance identifier filter value. Specify this parameter to show only the reservation that matches the specified reservation ID.

", - "DescribeReservedDBInstancesMessage$ReservedDBInstancesOfferingId": "

The offering identifier filter value. Specify this parameter to show only purchased reservations matching the specified offering identifier.

", - "DescribeReservedDBInstancesMessage$DBInstanceClass": "

The DB instance class filter value. Specify this parameter to show only those reservations matching the specified DB instances class.

", - "DescribeReservedDBInstancesMessage$Duration": "

The duration filter value, specified in years or seconds. Specify this parameter to show only reservations for this duration.

Valid Values: 1 | 3 | 31536000 | 94608000

", - "DescribeReservedDBInstancesMessage$ProductDescription": "

The product description filter value. Specify this parameter to show only those reservations matching the specified product description.

", - "DescribeReservedDBInstancesMessage$OfferingType": "

The offering type filter value. Specify this parameter to show only the available offerings matching the specified offering type.

Valid Values: \"Partial Upfront\" | \"All Upfront\" | \"No Upfront\"

", - "DescribeReservedDBInstancesMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeReservedDBInstancesOfferingsMessage$ReservedDBInstancesOfferingId": "

The offering identifier filter value. Specify this parameter to show only the available offering that matches the specified reservation identifier.

Example: 438012d3-4052-4cc7-b2e3-8d3372e0e706

", - "DescribeReservedDBInstancesOfferingsMessage$DBInstanceClass": "

The DB instance class filter value. Specify this parameter to show only the available offerings matching the specified DB instance class.

", - "DescribeReservedDBInstancesOfferingsMessage$Duration": "

Duration filter value, specified in years or seconds. Specify this parameter to show only reservations for this duration.

Valid Values: 1 | 3 | 31536000 | 94608000

", - "DescribeReservedDBInstancesOfferingsMessage$ProductDescription": "

Product description filter value. Specify this parameter to show only the available offerings matching the specified product description.

", - "DescribeReservedDBInstancesOfferingsMessage$OfferingType": "

The offering type filter value. Specify this parameter to show only the available offerings matching the specified offering type.

Valid Values: \"Partial Upfront\" | \"All Upfront\" | \"No Upfront\"

", - "DescribeReservedDBInstancesOfferingsMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeSourceRegionsMessage$RegionName": "

The source AWS Region name. For example, us-east-1.

Constraints:

  • Must specify a valid AWS Region name.

", - "DescribeSourceRegionsMessage$Marker": "

An optional pagination token provided by a previous DescribeSourceRegions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "DescribeValidDBInstanceModificationsMessage$DBInstanceIdentifier": "

The customer identifier or the ARN of your DB instance.

", - "DomainMembership$Domain": "

The identifier of the Active Directory Domain.

", - "DomainMembership$Status": "

The status of the DB instance's Active Directory Domain membership, such as joined, pending-join, failed etc).

", - "DomainMembership$FQDN": "

The fully qualified domain name of the Active Directory Domain.

", - "DomainMembership$IAMRoleName": "

The name of the IAM role to be used when making API calls to the Directory Service.

", - "DownloadDBLogFilePortionDetails$LogFileData": "

Entries from the specified log file.

", - "DownloadDBLogFilePortionDetails$Marker": "

A pagination token that can be used in a subsequent DownloadDBLogFilePortion request.

", - "DownloadDBLogFilePortionMessage$DBInstanceIdentifier": "

The customer-assigned name of the DB instance that contains the log files you want to list.

Constraints:

  • Must match the identifier of an existing DBInstance.

", - "DownloadDBLogFilePortionMessage$LogFileName": "

The name of the log file to be downloaded.

", - "DownloadDBLogFilePortionMessage$Marker": "

The pagination token provided in the previous request or \"0\". If the Marker parameter is specified the response includes only records beyond the marker until the end of the file or up to NumberOfLines.

", - "EC2SecurityGroup$Status": "

Provides the status of the EC2 security group. Status can be \"authorizing\", \"authorized\", \"revoking\", and \"revoked\".

", - "EC2SecurityGroup$EC2SecurityGroupName": "

Specifies the name of the EC2 security group.

", - "EC2SecurityGroup$EC2SecurityGroupId": "

Specifies the id of the EC2 security group.

", - "EC2SecurityGroup$EC2SecurityGroupOwnerId": "

Specifies the AWS ID of the owner of the EC2 security group specified in the EC2SecurityGroupName field.

", - "Endpoint$Address": "

Specifies the DNS address of the DB instance.

", - "Endpoint$HostedZoneId": "

Specifies the ID that Amazon Route 53 assigns when you create a hosted zone.

", - "EngineDefaults$DBParameterGroupFamily": "

Specifies the name of the DB parameter group family that the engine default parameters apply to.

", - "EngineDefaults$Marker": "

An optional pagination token provided by a previous EngineDefaults request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "Event$SourceIdentifier": "

Provides the identifier for the source of the event.

", - "Event$Message": "

Provides the text of this event.

", - "Event$SourceArn": "

The Amazon Resource Name (ARN) for the event.

", - "EventCategoriesList$member": null, - "EventCategoriesMap$SourceType": "

The source type that the returned categories belong to

", - "EventSubscription$CustomerAwsId": "

The AWS customer account associated with the RDS event notification subscription.

", - "EventSubscription$CustSubscriptionId": "

The RDS event notification subscription Id.

", - "EventSubscription$SnsTopicArn": "

The topic ARN of the RDS event notification subscription.

", - "EventSubscription$Status": "

The status of the RDS event notification subscription.

Constraints:

Can be one of the following: creating | modifying | deleting | active | no-permission | topic-not-exist

The status \"no-permission\" indicates that RDS no longer has permission to post to the SNS topic. The status \"topic-not-exist\" indicates that the topic was deleted after the subscription was created.

", - "EventSubscription$SubscriptionCreationTime": "

The time the RDS event notification subscription was created.

", - "EventSubscription$SourceType": "

The source type for the RDS event notification subscription.

", - "EventSubscription$EventSubscriptionArn": "

The Amazon Resource Name (ARN) for the event subscription.

", - "EventSubscriptionsMessage$Marker": "

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "EventsMessage$Marker": "

An optional pagination token provided by a previous Events request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "FailoverDBClusterMessage$DBClusterIdentifier": "

A DB cluster identifier to force a failover for. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

", - "FailoverDBClusterMessage$TargetDBInstanceIdentifier": "

The name of the instance to promote to the primary instance.

You must specify the instance identifier for an Aurora Replica in the DB cluster. For example, mydbcluster-replica1.

", - "Filter$Name": "

The name of the filter. Filter names are case-sensitive.

", - "FilterValueList$member": null, - "IPRange$Status": "

Specifies the status of the IP range. Status can be \"authorizing\", \"authorized\", \"revoking\", and \"revoked\".

", - "IPRange$CIDRIP": "

Specifies the IP range.

", - "KeyList$member": null, - "ListTagsForResourceMessage$ResourceName": "

The Amazon RDS resource with tags to be listed. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", - "LogTypeList$member": null, - "ModifyDBClusterMessage$DBClusterIdentifier": "

The DB cluster identifier for the cluster being modified. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

", - "ModifyDBClusterMessage$NewDBClusterIdentifier": "

The new DB cluster identifier for the DB cluster when renaming a DB cluster. This value is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • The first character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-cluster2

", - "ModifyDBClusterMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group to use for the DB cluster.

", - "ModifyDBClusterMessage$MasterUserPassword": "

The new password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", - "ModifyDBClusterMessage$OptionGroupName": "

A value that indicates that the DB cluster should be associated with the specified option group. Changing this parameter doesn't result in an outage except in the following case, and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If the parameter change results in an option group that enables OEM, this change can cause a brief (sub-second) period during which new connections are rejected but existing connections are not interrupted.

Permanent options can't be removed from an option group. The option group can't be removed from a DB cluster once it is associated with a DB cluster.

", - "ModifyDBClusterMessage$PreferredBackupWindow": "

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", - "ModifyDBClusterMessage$PreferredMaintenanceWindow": "

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", - "ModifyDBClusterMessage$EngineVersion": "

The version number of the database engine to which you want to upgrade. Changing this parameter results in an outage. The change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true.

For a list of valid engine versions, see CreateDBInstance, or call DescribeDBEngineVersions.

", - "ModifyDBClusterParameterGroupMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group to modify.

", - "ModifyDBClusterSnapshotAttributeMessage$DBClusterSnapshotIdentifier": "

The identifier for the DB cluster snapshot to modify the attributes for.

", - "ModifyDBClusterSnapshotAttributeMessage$AttributeName": "

The name of the DB cluster snapshot attribute to modify.

To manage authorization for other AWS accounts to copy or restore a manual DB cluster snapshot, set this value to restore.

", - "ModifyDBInstanceMessage$DBInstanceIdentifier": "

The DB instance identifier. This value is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing DBInstance.

", - "ModifyDBInstanceMessage$DBInstanceClass": "

The new compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

If you modify the DB instance class, an outage occurs during the change. The change is applied during the next maintenance window, unless ApplyImmediately is specified as true for this request.

Default: Uses existing setting

", - "ModifyDBInstanceMessage$DBSubnetGroupName": "

The new DB subnet group for the DB instance. You can use this parameter to move your DB instance to a different VPC. If your DB instance is not in a VPC, you can also use this parameter to move your DB instance into a VPC. For more information, see Updating the VPC for a DB Instance.

Changing the subnet group causes an outage during the change. The change is applied during the next maintenance window, unless you specify true for the ApplyImmediately parameter.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetGroup

", - "ModifyDBInstanceMessage$MasterUserPassword": "

The new password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible. Between the time of the request and the completion of the request, the MasterUserPassword element exists in the PendingModifiedValues element of the operation response.

Amazon Aurora

Not applicable. The password for the master user is managed by the DB cluster. For more information, see ModifyDBCluster.

Default: Uses existing setting

MariaDB

Constraints: Must contain from 8 to 41 characters.

Microsoft SQL Server

Constraints: Must contain from 8 to 128 characters.

MySQL

Constraints: Must contain from 8 to 41 characters.

Oracle

Constraints: Must contain from 8 to 30 characters.

PostgreSQL

Constraints: Must contain from 8 to 128 characters.

Amazon RDS API actions never return the password, so this action provides a way to regain access to a primary instance user if the password is lost. This includes restoring privileges that might have been accidentally revoked.

", - "ModifyDBInstanceMessage$DBParameterGroupName": "

The name of the DB parameter group to apply to the DB instance. Changing this setting doesn't result in an outage. The parameter group name itself is changed immediately, but the actual parameter changes are not applied until you reboot the instance without failover. The db instance will NOT be rebooted automatically and the parameter changes will NOT be applied during the next maintenance window.

Default: Uses existing setting

Constraints: The DB parameter group must be in the same DB parameter group family as this DB instance.

", - "ModifyDBInstanceMessage$PreferredBackupWindow": "

The daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod parameter. Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible.

Amazon Aurora

Not applicable. The daily time range for creating automated backups is managed by the DB cluster. For more information, see ModifyDBCluster.

Constraints:

  • Must be in the format hh24:mi-hh24:mi

  • Must be in Universal Time Coordinated (UTC)

  • Must not conflict with the preferred maintenance window

  • Must be at least 30 minutes

", - "ModifyDBInstanceMessage$PreferredMaintenanceWindow": "

The weekly time range (in UTC) during which system maintenance can occur, which might result in an outage. Changing this parameter doesn't result in an outage, except in the following situation, and the change is asynchronously applied as soon as possible. If there are pending actions that cause a reboot, and the maintenance window is changed to include the current time, then changing this parameter will cause a reboot of the DB instance. If moving this window to the current time, there must be at least 30 minutes between the current time and end of the window to ensure pending changes are applied.

Default: Uses existing setting

Format: ddd:hh24:mi-ddd:hh24:mi

Valid Days: Mon | Tue | Wed | Thu | Fri | Sat | Sun

Constraints: Must be at least 30 minutes

", - "ModifyDBInstanceMessage$EngineVersion": "

The version number of the database engine to upgrade to. Changing this parameter results in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

For major version upgrades, if a nondefault DB parameter group is currently in use, a new DB parameter group in the DB parameter group family for the new engine version must be specified. The new DB parameter group can be the default for that DB parameter group family.

For a list of valid engine versions, see CreateDBInstance.

", - "ModifyDBInstanceMessage$LicenseModel": "

The license model for the DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", - "ModifyDBInstanceMessage$OptionGroupName": "

Indicates that the DB instance should be associated with the specified option group. Changing this parameter doesn't result in an outage except in the following case and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If the parameter change results in an option group that enables OEM, this change can cause a brief (sub-second) period during which new connections are rejected but existing connections are not interrupted.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", - "ModifyDBInstanceMessage$NewDBInstanceIdentifier": "

The new DB instance identifier for the DB instance when renaming a DB instance. When you change the DB instance identifier, an instance reboot will occur immediately if you set Apply Immediately to true, or will occur during the next maintenance window if Apply Immediately to false. This value is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • The first character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", - "ModifyDBInstanceMessage$StorageType": "

Specifies the storage type to be associated with the DB instance.

If you specify Provisioned IOPS (io1), you must also include a value for the Iops parameter.

If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon RDS operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a Read Replica for the instance, and creating a DB snapshot of the instance.

Valid values: standard | gp2 | io1

Default: io1 if the Iops parameter is specified, otherwise standard

", - "ModifyDBInstanceMessage$TdeCredentialArn": "

The ARN from the key store with which to associate the instance for TDE encryption.

", - "ModifyDBInstanceMessage$TdeCredentialPassword": "

The password for the given ARN from the key store in order to access the device.

", - "ModifyDBInstanceMessage$CACertificateIdentifier": "

Indicates the certificate that needs to be associated with the instance.

", - "ModifyDBInstanceMessage$Domain": "

The Active Directory Domain to move the instance to. Specify none to remove the instance from its current domain. The domain must be created prior to this operation. Currently only a Microsoft SQL Server instance can be created in a Active Directory Domain.

", - "ModifyDBInstanceMessage$MonitoringRoleArn": "

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, go to To create an IAM role for Amazon RDS Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", - "ModifyDBInstanceMessage$DomainIAMRoleName": "

The name of the IAM role to use when making API calls to the Directory Service.

", - "ModifyDBInstanceMessage$PerformanceInsightsKMSKeyId": "

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", - "ModifyDBParameterGroupMessage$DBParameterGroupName": "

The name of the DB parameter group.

Constraints:

  • If supplied, must match the name of an existing DBParameterGroup.

", - "ModifyDBSnapshotAttributeMessage$DBSnapshotIdentifier": "

The identifier for the DB snapshot to modify the attributes for.

", - "ModifyDBSnapshotAttributeMessage$AttributeName": "

The name of the DB snapshot attribute to modify.

To manage authorization for other AWS accounts to copy or restore a manual DB snapshot, set this value to restore.

", - "ModifyDBSnapshotMessage$DBSnapshotIdentifier": "

The identifier of the DB snapshot to modify.

", - "ModifyDBSnapshotMessage$EngineVersion": "

The engine version to upgrade the DB snapshot to.

The following are the database engines and engine versions that are available when you upgrade a DB snapshot.

MySQL

  • 5.5.46 (supported for 5.1 DB snapshots)

Oracle

  • 12.1.0.2.v8 (supported for 12.1.0.1 DB snapshots)

  • 11.2.0.4.v12 (supported for 11.2.0.2 DB snapshots)

  • 11.2.0.4.v11 (supported for 11.2.0.3 DB snapshots)

", - "ModifyDBSnapshotMessage$OptionGroupName": "

The option group to identify with the upgraded DB snapshot.

You can specify this parameter when you upgrade an Oracle DB snapshot. The same option group considerations apply when upgrading a DB snapshot as when upgrading a DB instance. For more information, see Option Group Considerations.

", - "ModifyDBSubnetGroupMessage$DBSubnetGroupName": "

The name for the DB subnet group. This value is stored as a lowercase string. You can't modify the default subnet group.

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", - "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription": "

The description for the DB subnet group.

", - "ModifyEventSubscriptionMessage$SubscriptionName": "

The name of the RDS event notification subscription.

", - "ModifyEventSubscriptionMessage$SnsTopicArn": "

The Amazon Resource Name (ARN) of the SNS topic created for event notification. The ARN is created by Amazon SNS when you create a topic and subscribe to it.

", - "ModifyEventSubscriptionMessage$SourceType": "

The type of source that is generating the events. For example, if you want to be notified of events generated by a DB instance, you would set this parameter to db-instance. if this value is not specified, all events are returned.

Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot

", - "ModifyOptionGroupMessage$OptionGroupName": "

The name of the option group to be modified.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", - "Option$OptionName": "

The name of the option.

", - "Option$OptionDescription": "

The description of the option.

", - "Option$OptionVersion": "

The version of the option.

", - "OptionConfiguration$OptionName": "

The configuration of options to include in a group.

", - "OptionConfiguration$OptionVersion": "

The version for the option.

", - "OptionGroup$OptionGroupName": "

Specifies the name of the option group.

", - "OptionGroup$OptionGroupDescription": "

Provides a description of the option group.

", - "OptionGroup$EngineName": "

Indicates the name of the engine that this option group can be applied to.

", - "OptionGroup$MajorEngineVersion": "

Indicates the major engine version associated with this option group.

", - "OptionGroup$VpcId": "

If AllowsVpcAndNonVpcInstanceMemberships is false, this field is blank. If AllowsVpcAndNonVpcInstanceMemberships is true and this field is blank, then this option group can be applied to both VPC and non-VPC instances. If this field contains a value, then this option group can only be applied to instances that are in the VPC indicated by this field.

", - "OptionGroup$OptionGroupArn": "

The Amazon Resource Name (ARN) for the option group.

", - "OptionGroupMembership$OptionGroupName": "

The name of the option group that the instance belongs to.

", - "OptionGroupMembership$Status": "

The status of the DB instance's option group membership. Valid values are: in-sync, pending-apply, pending-removal, pending-maintenance-apply, pending-maintenance-removal, applying, removing, and failed.

", - "OptionGroupOption$Name": "

The name of the option.

", - "OptionGroupOption$Description": "

The description of the option.

", - "OptionGroupOption$EngineName": "

The name of the engine that this option can be applied to.

", - "OptionGroupOption$MajorEngineVersion": "

Indicates the major engine version that the option is available for.

", - "OptionGroupOption$MinimumRequiredMinorEngineVersion": "

The minimum required engine version for the option to be applied.

", - "OptionGroupOptionSetting$SettingName": "

The name of the option group option.

", - "OptionGroupOptionSetting$SettingDescription": "

The description of the option group option.

", - "OptionGroupOptionSetting$DefaultValue": "

The default value for the option group option.

", - "OptionGroupOptionSetting$ApplyType": "

The DB engine specific parameter type for the option group option.

", - "OptionGroupOptionSetting$AllowedValues": "

Indicates the acceptable values for the option group option.

", - "OptionGroupOptionsMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "OptionGroups$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "OptionNamesList$member": null, - "OptionSetting$Name": "

The name of the option that has settings that you can set.

", - "OptionSetting$Value": "

The current value of the option setting.

", - "OptionSetting$DefaultValue": "

The default value of the option setting.

", - "OptionSetting$Description": "

The description of the option setting.

", - "OptionSetting$ApplyType": "

The DB engine specific parameter type.

", - "OptionSetting$DataType": "

The data type of the option setting.

", - "OptionSetting$AllowedValues": "

The allowed values of the option setting.

", - "OptionVersion$Version": "

The version of the option.

", - "OptionsConflictsWith$member": null, - "OptionsDependedOn$member": null, - "OrderableDBInstanceOption$Engine": "

The engine type of a DB instance.

", - "OrderableDBInstanceOption$EngineVersion": "

The engine version of a DB instance.

", - "OrderableDBInstanceOption$DBInstanceClass": "

The DB instance class for a DB instance.

", - "OrderableDBInstanceOption$LicenseModel": "

The license model for a DB instance.

", - "OrderableDBInstanceOption$StorageType": "

Indicates the storage type for a DB instance.

", - "OrderableDBInstanceOptionsMessage$Marker": "

An optional pagination token provided by a previous OrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", - "Parameter$ParameterName": "

Specifies the name of the parameter.

", - "Parameter$ParameterValue": "

Specifies the value of the parameter.

", - "Parameter$Description": "

Provides a description of the parameter.

", - "Parameter$Source": "

Indicates the source of the parameter value.

", - "Parameter$ApplyType": "

Specifies the engine specific parameters type.

", - "Parameter$DataType": "

Specifies the valid data type for the parameter.

", - "Parameter$AllowedValues": "

Specifies the valid range of values for the parameter.

", - "Parameter$MinimumEngineVersion": "

The earliest engine version to which the parameter can apply.

", - "PendingMaintenanceAction$Action": "

The type of pending maintenance action that is available for the resource.

", - "PendingMaintenanceAction$OptInStatus": "

Indicates the type of opt-in request that has been received for the resource.

", - "PendingMaintenanceAction$Description": "

A description providing more detail about the maintenance action.

", - "PendingMaintenanceActionsMessage$Marker": "

An optional pagination token provided by a previous DescribePendingMaintenanceActions request. If this parameter is specified, the response includes only records beyond the marker, up to a number of records specified by MaxRecords.

", - "PendingModifiedValues$DBInstanceClass": "

Contains the new DBInstanceClass for the DB instance that will be applied or is currently being applied.

", - "PendingModifiedValues$MasterUserPassword": "

Contains the pending or currently-in-progress change of the master credentials for the DB instance.

", - "PendingModifiedValues$EngineVersion": "

Indicates the database engine version.

", - "PendingModifiedValues$LicenseModel": "

The license model for the DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", - "PendingModifiedValues$DBInstanceIdentifier": "

Contains the new DBInstanceIdentifier for the DB instance that will be applied or is currently being applied.

", - "PendingModifiedValues$StorageType": "

Specifies the storage type to be associated with the DB instance.

", - "PendingModifiedValues$CACertificateIdentifier": "

Specifies the identifier of the CA certificate for the DB instance.

", - "PendingModifiedValues$DBSubnetGroupName": "

The new DB subnet group for the DB instance.

", - "PromoteReadReplicaDBClusterMessage$DBClusterIdentifier": "

The identifier of the DB cluster Read Replica to promote. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster Read Replica.

Example: my-cluster-replica1

", - "PromoteReadReplicaMessage$DBInstanceIdentifier": "

The DB instance identifier. This value is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing Read Replica DB instance.

Example: mydbinstance

", - "PromoteReadReplicaMessage$PreferredBackupWindow": "

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferingId": "

The ID of the Reserved DB instance offering to purchase.

Example: 438012d3-4052-4cc7-b2e3-8d3372e0e706

", - "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstanceId": "

Customer-specified identifier to track this reservation.

Example: myreservationID

", - "ReadReplicaDBClusterIdentifierList$member": null, - "ReadReplicaDBInstanceIdentifierList$member": null, - "ReadReplicaIdentifierList$member": null, - "RebootDBInstanceMessage$DBInstanceIdentifier": "

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing DBInstance.

", - "RecurringCharge$RecurringChargeFrequency": "

The frequency of the recurring charge.

", - "RemoveRoleFromDBClusterMessage$DBClusterIdentifier": "

The name of the DB cluster to disassociate the IAM role from.

", - "RemoveRoleFromDBClusterMessage$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role to disassociate from the Aurora DB cluster, for example arn:aws:iam::123456789012:role/AuroraAccessRole.

", - "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName": "

The name of the RDS event notification subscription you want to remove a source identifier from.

", - "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier": "

The source identifier to be removed from the subscription, such as the DB instance identifier for a DB instance or the name of a security group.

", - "RemoveTagsFromResourceMessage$ResourceName": "

The Amazon RDS resource that the tags are removed from. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", - "ReservedDBInstance$ReservedDBInstanceId": "

The unique identifier for the reservation.

", - "ReservedDBInstance$ReservedDBInstancesOfferingId": "

The offering identifier.

", - "ReservedDBInstance$DBInstanceClass": "

The DB instance class for the reserved DB instance.

", - "ReservedDBInstance$CurrencyCode": "

The currency code for the reserved DB instance.

", - "ReservedDBInstance$ProductDescription": "

The description of the reserved DB instance.

", - "ReservedDBInstance$OfferingType": "

The offering type of this reserved DB instance.

", - "ReservedDBInstance$State": "

The state of the reserved DB instance.

", - "ReservedDBInstance$ReservedDBInstanceArn": "

The Amazon Resource Name (ARN) for the reserved DB instance.

", - "ReservedDBInstanceMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "ReservedDBInstancesOffering$ReservedDBInstancesOfferingId": "

The offering identifier.

", - "ReservedDBInstancesOffering$DBInstanceClass": "

The DB instance class for the reserved DB instance.

", - "ReservedDBInstancesOffering$CurrencyCode": "

The currency code for the reserved DB instance offering.

", - "ReservedDBInstancesOffering$ProductDescription": "

The database engine used by the offering.

", - "ReservedDBInstancesOffering$OfferingType": "

The offering type.

", - "ReservedDBInstancesOfferingMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "ResetDBClusterParameterGroupMessage$DBClusterParameterGroupName": "

The name of the DB cluster parameter group to reset.

", - "ResetDBParameterGroupMessage$DBParameterGroupName": "

The name of the DB parameter group.

Constraints:

  • Must match the name of an existing DBParameterGroup.

", - "ResourcePendingMaintenanceActions$ResourceIdentifier": "

The ARN of the resource that has pending maintenance actions.

", - "RestoreDBClusterFromS3Message$CharacterSetName": "

A value that indicates that the restored DB cluster should be associated with the specified CharacterSet.

", - "RestoreDBClusterFromS3Message$DatabaseName": "

The database name for the restored DB cluster.

", - "RestoreDBClusterFromS3Message$DBClusterIdentifier": "

The name of the DB cluster to create from the source data in the Amazon S3 bucket. This parameter is isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", - "RestoreDBClusterFromS3Message$DBClusterParameterGroupName": "

The name of the DB cluster parameter group to associate with the restored DB cluster. If this argument is omitted, default.aurora5.6 is used.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", - "RestoreDBClusterFromS3Message$DBSubnetGroupName": "

A DB subnet group to associate with the restored DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", - "RestoreDBClusterFromS3Message$Engine": "

The name of the database engine to be used for the restored DB cluster.

Valid Values: aurora, aurora-postgresql

", - "RestoreDBClusterFromS3Message$EngineVersion": "

The version number of the database engine to use.

Aurora MySQL

Example: 5.6.10a

Aurora PostgreSQL

Example: 9.6.3

", - "RestoreDBClusterFromS3Message$MasterUsername": "

The name of the master user for the restored DB cluster.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", - "RestoreDBClusterFromS3Message$MasterUserPassword": "

The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", - "RestoreDBClusterFromS3Message$OptionGroupName": "

A value that indicates that the restored DB cluster should be associated with the specified option group.

Permanent options can't be removed from an option group. An option group can't be removed from a DB cluster once it is associated with a DB cluster.

", - "RestoreDBClusterFromS3Message$PreferredBackupWindow": "

The daily time range during which automated backups are created if automated backups are enabled using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", - "RestoreDBClusterFromS3Message$PreferredMaintenanceWindow": "

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", - "RestoreDBClusterFromS3Message$KmsKeyId": "

The AWS KMS key identifier for an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KM encryption key.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon RDS will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", - "RestoreDBClusterFromS3Message$SourceEngine": "

The identifier for the database engine that was backed up to create the files stored in the Amazon S3 bucket.

Valid values: mysql

", - "RestoreDBClusterFromS3Message$SourceEngineVersion": "

The version of the database that the backup files were created from.

MySQL version 5.5 and 5.6 are supported.

Example: 5.6.22

", - "RestoreDBClusterFromS3Message$S3BucketName": "

The name of the Amazon S3 bucket that contains the data used to create the Amazon Aurora DB cluster.

", - "RestoreDBClusterFromS3Message$S3Prefix": "

The prefix for all of the file names that contain the data used to create the Amazon Aurora DB cluster. If you do not specify a SourceS3Prefix value, then the Amazon Aurora DB cluster is created by using all of the files in the Amazon S3 bucket.

", - "RestoreDBClusterFromS3Message$S3IngestionRoleArn": "

The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that authorizes Amazon RDS to access the Amazon S3 bucket on your behalf.

", - "RestoreDBClusterFromSnapshotMessage$DBClusterIdentifier": "

The name of the DB cluster to create from the DB snapshot or DB cluster snapshot. This parameter isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", - "RestoreDBClusterFromSnapshotMessage$SnapshotIdentifier": "

The identifier for the DB snapshot or DB cluster snapshot to restore from.

You can use either the name or the Amazon Resource Name (ARN) to specify a DB cluster snapshot. However, you can use only the ARN to specify a DB snapshot.

Constraints:

  • Must match the identifier of an existing Snapshot.

", - "RestoreDBClusterFromSnapshotMessage$Engine": "

The database engine to use for the new DB cluster.

Default: The same as source

Constraint: Must be compatible with the engine of the source

", - "RestoreDBClusterFromSnapshotMessage$EngineVersion": "

The version of the database engine to use for the new DB cluster.

", - "RestoreDBClusterFromSnapshotMessage$DBSubnetGroupName": "

The name of the DB subnet group to use for the new DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", - "RestoreDBClusterFromSnapshotMessage$DatabaseName": "

The database name for the restored DB cluster.

", - "RestoreDBClusterFromSnapshotMessage$OptionGroupName": "

The name of the option group to use for the restored DB cluster.

", - "RestoreDBClusterFromSnapshotMessage$KmsKeyId": "

The AWS KMS key identifier to use when restoring an encrypted DB cluster from a DB snapshot or DB cluster snapshot.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are restoring a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

If you do not specify a value for the KmsKeyId parameter, then the following will occur:

  • If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is encrypted, then the restored DB cluster is encrypted using the KMS key that was used to encrypt the DB snapshot or DB cluster snapshot.

  • If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is not encrypted, then the restored DB cluster is not encrypted.

", - "RestoreDBClusterToPointInTimeMessage$DBClusterIdentifier": "

The name of the new DB cluster to be created.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", - "RestoreDBClusterToPointInTimeMessage$RestoreType": "

The type of restore to be performed. You can specify one of the following values:

  • full-copy - The new DB cluster is restored as a full copy of the source DB cluster.

  • copy-on-write - The new DB cluster is restored as a clone of the source DB cluster.

Constraints: You can't specify copy-on-write if the engine version of the source DB cluster is earlier than 1.11.

If you don't specify a RestoreType value, then the new DB cluster is restored as a full copy of the source DB cluster.

", - "RestoreDBClusterToPointInTimeMessage$SourceDBClusterIdentifier": "

The identifier of the source DB cluster from which to restore.

Constraints:

  • Must match the identifier of an existing DBCluster.

", - "RestoreDBClusterToPointInTimeMessage$DBSubnetGroupName": "

The DB subnet group name to use for the new DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", - "RestoreDBClusterToPointInTimeMessage$OptionGroupName": "

The name of the option group for the new DB cluster.

", - "RestoreDBClusterToPointInTimeMessage$KmsKeyId": "

The AWS KMS key identifier to use when restoring an encrypted DB cluster from an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are restoring a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

You can restore to a new DB cluster and encrypt the new DB cluster with a KMS key that is different than the KMS key used to encrypt the source DB cluster. The new DB cluster is encrypted with the KMS key identified by the KmsKeyId parameter.

If you do not specify a value for the KmsKeyId parameter, then the following will occur:

  • If the DB cluster is encrypted, then the restored DB cluster is encrypted using the KMS key that was used to encrypt the source DB cluster.

  • If the DB cluster is not encrypted, then the restored DB cluster is not encrypted.

If DBClusterIdentifier refers to a DB cluster that is not encrypted, then the restore request is rejected.

", - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceIdentifier": "

Name of the DB instance to create from the DB snapshot. This parameter isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 numbers, letters, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", - "RestoreDBInstanceFromDBSnapshotMessage$DBSnapshotIdentifier": "

The identifier for the DB snapshot to restore from.

Constraints:

  • Must match the identifier of an existing DBSnapshot.

  • If you are restoring from a shared manual DB snapshot, the DBSnapshotIdentifier must be the ARN of the shared DB snapshot.

", - "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceClass": "

The compute and memory capacity of the Amazon RDS DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Default: The same DBInstanceClass as the original DB instance.

", - "RestoreDBInstanceFromDBSnapshotMessage$AvailabilityZone": "

The EC2 Availability Zone that the DB instance is created in.

Default: A random, system-chosen Availability Zone.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

Example: us-east-1a

", - "RestoreDBInstanceFromDBSnapshotMessage$DBSubnetGroupName": "

The DB subnet group name to use for the new instance.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", - "RestoreDBInstanceFromDBSnapshotMessage$LicenseModel": "

License model information for the restored DB instance.

Default: Same as source.

Valid values: license-included | bring-your-own-license | general-public-license

", - "RestoreDBInstanceFromDBSnapshotMessage$DBName": "

The database name for the restored DB instance.

This parameter doesn't apply to the MySQL, PostgreSQL, or MariaDB engines.

", - "RestoreDBInstanceFromDBSnapshotMessage$Engine": "

The database engine to use for the new instance.

Default: The same as source

Constraint: Must be compatible with the engine of the source. For example, you can restore a MariaDB 10.1 DB instance from a MySQL 5.6 snapshot.

Valid Values:

  • mariadb

  • mysql

  • oracle-ee

  • oracle-se2

  • oracle-se1

  • oracle-se

  • postgres

  • sqlserver-ee

  • sqlserver-se

  • sqlserver-ex

  • sqlserver-web

", - "RestoreDBInstanceFromDBSnapshotMessage$OptionGroupName": "

The name of the option group to be used for the restored DB instance.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", - "RestoreDBInstanceFromDBSnapshotMessage$StorageType": "

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", - "RestoreDBInstanceFromDBSnapshotMessage$TdeCredentialArn": "

The ARN from the key store with which to associate the instance for TDE encryption.

", - "RestoreDBInstanceFromDBSnapshotMessage$TdeCredentialPassword": "

The password for the given ARN from the key store in order to access the device.

", - "RestoreDBInstanceFromDBSnapshotMessage$Domain": "

Specify the Active Directory Domain to restore the instance in.

", - "RestoreDBInstanceFromDBSnapshotMessage$DomainIAMRoleName": "

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", - "RestoreDBInstanceFromS3Message$DBName": "

The name of the database to create when the DB instance is created. Follow the naming rules specified in CreateDBInstance.

", - "RestoreDBInstanceFromS3Message$DBInstanceIdentifier": "

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", - "RestoreDBInstanceFromS3Message$DBInstanceClass": "

The compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Importing from Amazon S3 is not supported on the db.t2.micro DB instance class.

", - "RestoreDBInstanceFromS3Message$Engine": "

The name of the database engine to be used for this instance.

Valid Values: mysql

", - "RestoreDBInstanceFromS3Message$MasterUsername": "

The name for the master user.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", - "RestoreDBInstanceFromS3Message$MasterUserPassword": "

The password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", - "RestoreDBInstanceFromS3Message$AvailabilityZone": "

The Availability Zone that the DB instance is created in. For information about AWS Regions and Availability Zones, see Regions and Availability Zones.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

Constraint: The AvailabilityZone parameter can't be specified if the MultiAZ parameter is set to true. The specified Availability Zone must be in the same AWS Region as the current endpoint.

", - "RestoreDBInstanceFromS3Message$DBSubnetGroupName": "

A DB subnet group to associate with this DB instance.

", - "RestoreDBInstanceFromS3Message$PreferredMaintenanceWindow": "

The time range each week during which system maintenance can occur, in Universal Coordinated Time (UTC). For more information, see Amazon RDS Maintenance Window.

Constraints:

  • Must be in the format ddd:hh24:mi-ddd:hh24:mi.

  • Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred backup window.

  • Must be at least 30 minutes.

", - "RestoreDBInstanceFromS3Message$DBParameterGroupName": "

The name of the DB parameter group to associate with this DB instance. If this argument is omitted, the default parameter group for the specified engine is used.

", - "RestoreDBInstanceFromS3Message$PreferredBackupWindow": "

The time range each day during which automated backups are created if automated backups are enabled. For more information, see The Backup Window.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", - "RestoreDBInstanceFromS3Message$EngineVersion": "

The version number of the database engine to use. Choose the latest minor version of your database engine as specified in CreateDBInstance.

", - "RestoreDBInstanceFromS3Message$LicenseModel": "

The license model for this DB instance. Use general-public-license.

", - "RestoreDBInstanceFromS3Message$OptionGroupName": "

The name of the option group to associate with this DB instance. If this argument is omitted, the default option group for the specified engine is used.

", - "RestoreDBInstanceFromS3Message$StorageType": "

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified; otherwise standard

", - "RestoreDBInstanceFromS3Message$KmsKeyId": "

The AWS KMS key identifier for an encrypted DB instance.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB instance with the same AWS account that owns the KMS encryption key used to encrypt the new DB instance, then you can use the KMS key alias instead of the ARN for the KM encryption key.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon RDS will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", - "RestoreDBInstanceFromS3Message$MonitoringRoleArn": "

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, see Setting Up and Enabling Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", - "RestoreDBInstanceFromS3Message$SourceEngine": "

The name of the engine of your source database.

Valid Values: mysql

", - "RestoreDBInstanceFromS3Message$SourceEngineVersion": "

The engine version of your source database.

Valid Values: 5.6

", - "RestoreDBInstanceFromS3Message$S3BucketName": "

The name of your Amazon S3 bucket that contains your database backup file.

", - "RestoreDBInstanceFromS3Message$S3Prefix": "

The prefix of your Amazon S3 bucket.

", - "RestoreDBInstanceFromS3Message$S3IngestionRoleArn": "

An AWS Identity and Access Management (IAM) role to allow Amazon RDS to access your Amazon S3 bucket.

", - "RestoreDBInstanceFromS3Message$PerformanceInsightsKMSKeyId": "

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), the KMS key identifier, or the KMS key alias for the KMS encryption key.

", - "RestoreDBInstanceToPointInTimeMessage$SourceDBInstanceIdentifier": "

The identifier of the source DB instance from which to restore.

Constraints:

  • Must match the identifier of an existing DB instance.

", - "RestoreDBInstanceToPointInTimeMessage$TargetDBInstanceIdentifier": "

The name of the new DB instance to be created.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", - "RestoreDBInstanceToPointInTimeMessage$DBInstanceClass": "

The compute and memory capacity of the Amazon RDS DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Default: The same DBInstanceClass as the original DB instance.

", - "RestoreDBInstanceToPointInTimeMessage$AvailabilityZone": "

The EC2 Availability Zone that the DB instance is created in.

Default: A random, system-chosen Availability Zone.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

Example: us-east-1a

", - "RestoreDBInstanceToPointInTimeMessage$DBSubnetGroupName": "

The DB subnet group name to use for the new instance.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", - "RestoreDBInstanceToPointInTimeMessage$LicenseModel": "

License model information for the restored DB instance.

Default: Same as source.

Valid values: license-included | bring-your-own-license | general-public-license

", - "RestoreDBInstanceToPointInTimeMessage$DBName": "

The database name for the restored DB instance.

This parameter is not used for the MySQL or MariaDB engines.

", - "RestoreDBInstanceToPointInTimeMessage$Engine": "

The database engine to use for the new instance.

Default: The same as source

Constraint: Must be compatible with the engine of the source

Valid Values:

  • mariadb

  • mysql

  • oracle-ee

  • oracle-se2

  • oracle-se1

  • oracle-se

  • postgres

  • sqlserver-ee

  • sqlserver-se

  • sqlserver-ex

  • sqlserver-web

", - "RestoreDBInstanceToPointInTimeMessage$OptionGroupName": "

The name of the option group to be used for the restored DB instance.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", - "RestoreDBInstanceToPointInTimeMessage$StorageType": "

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", - "RestoreDBInstanceToPointInTimeMessage$TdeCredentialArn": "

The ARN from the key store with which to associate the instance for TDE encryption.

", - "RestoreDBInstanceToPointInTimeMessage$TdeCredentialPassword": "

The password for the given ARN from the key store in order to access the device.

", - "RestoreDBInstanceToPointInTimeMessage$Domain": "

Specify the Active Directory Domain to restore the instance in.

", - "RestoreDBInstanceToPointInTimeMessage$DomainIAMRoleName": "

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", - "RevokeDBSecurityGroupIngressMessage$DBSecurityGroupName": "

The name of the DB security group to revoke ingress from.

", - "RevokeDBSecurityGroupIngressMessage$CIDRIP": "

The IP range to revoke access from. Must be a valid CIDR range. If CIDRIP is specified, EC2SecurityGroupName, EC2SecurityGroupId and EC2SecurityGroupOwnerId can't be provided.

", - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupName": "

The name of the EC2 security group to revoke access from. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupId": "

The id of the EC2 security group to revoke access from. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", - "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId": "

The AWS Account Number of the owner of the EC2 security group specified in the EC2SecurityGroupName parameter. The AWS Access Key ID is not an acceptable value. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", - "SourceIdsList$member": null, - "SourceRegion$RegionName": "

The name of the source AWS Region.

", - "SourceRegion$Endpoint": "

The endpoint for the source AWS Region endpoint.

", - "SourceRegion$Status": "

The status of the source AWS Region.

", - "SourceRegionMessage$Marker": "

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", - "StartDBInstanceMessage$DBInstanceIdentifier": "

The user-supplied instance identifier.

", - "StopDBInstanceMessage$DBInstanceIdentifier": "

The user-supplied instance identifier.

", - "StopDBInstanceMessage$DBSnapshotIdentifier": "

The user-supplied instance identifier of the DB Snapshot created immediately before the DB instance is stopped.

", - "Subnet$SubnetIdentifier": "

Specifies the identifier of the subnet.

", - "Subnet$SubnetStatus": "

Specifies the status of the subnet.

", - "SubnetIdentifierList$member": null, - "Tag$Key": "

A key is the required name of the tag. The string value can be from 1 to 128 Unicode characters in length and can't be prefixed with \"aws:\" or \"rds:\". The string can only contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-' (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").

", - "Tag$Value": "

A value is the optional value of the tag. The string value can be from 1 to 256 Unicode characters in length and can't be prefixed with \"aws:\" or \"rds:\". The string can only contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-' (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").

", - "Timezone$TimezoneName": "

The name of the time zone.

", - "UpgradeTarget$Engine": "

The name of the upgrade target database engine.

", - "UpgradeTarget$EngineVersion": "

The version number of the upgrade target database engine.

", - "UpgradeTarget$Description": "

The version of the database engine that a DB instance can be upgraded to.

", - "ValidStorageOptions$StorageType": "

The valid storage types for your DB instance. For example, gp2, io1.

", - "VpcSecurityGroupIdList$member": null, - "VpcSecurityGroupMembership$VpcSecurityGroupId": "

The name of the VPC security group.

", - "VpcSecurityGroupMembership$Status": "

The status of the VPC security group.

" - } - }, - "Subnet": { - "base": "

This data type is used as a response element in the DescribeDBSubnetGroups action.

", - "refs": { - "SubnetList$member": null - } - }, - "SubnetAlreadyInUse": { - "base": "

The DB subnet is already in use in the Availability Zone.

", - "refs": { - } - }, - "SubnetIdentifierList": { - "base": null, - "refs": { - "CreateDBSubnetGroupMessage$SubnetIds": "

The EC2 Subnet IDs for the DB subnet group.

", - "ModifyDBSubnetGroupMessage$SubnetIds": "

The EC2 subnet IDs for the DB subnet group.

" - } - }, - "SubnetList": { - "base": null, - "refs": { - "DBSubnetGroup$Subnets": "

Contains a list of Subnet elements.

" - } - }, - "SubscriptionAlreadyExistFault": { - "base": "

The supplied subscription name already exists.

", - "refs": { - } - }, - "SubscriptionCategoryNotFoundFault": { - "base": "

The supplied category does not exist.

", - "refs": { - } - }, - "SubscriptionNotFoundFault": { - "base": "

The subscription name does not exist.

", - "refs": { - } - }, - "SupportedCharacterSetsList": { - "base": null, - "refs": { - "DBEngineVersion$SupportedCharacterSets": "

A list of the character sets supported by this engine for the CharacterSetName parameter of the CreateDBInstance action.

" - } - }, - "SupportedTimezonesList": { - "base": null, - "refs": { - "DBEngineVersion$SupportedTimezones": "

A list of the time zones supported by this engine for the Timezone parameter of the CreateDBInstance action.

" - } - }, - "TStamp": { - "base": null, - "refs": { - "BacktrackDBClusterMessage$BacktrackTo": "

The timestamp of the time to backtrack the DB cluster to, specified in ISO 8601 format. For more information about ISO 8601, see the ISO8601 Wikipedia page.

If the specified time is not a consistent time for the DB cluster, Aurora automatically chooses the nearest possible consistent time for the DB cluster.

Constraints:

  • Must contain a valid ISO 8601 timestamp.

  • Cannot contain a timestamp set in the future.

Example: 2017-07-08T18:00Z

", - "Certificate$ValidFrom": "

The starting date from which the certificate is valid.

", - "Certificate$ValidTill": "

The final date that the certificate continues to be valid.

", - "DBCluster$EarliestRestorableTime": "

The earliest time to which a database can be restored with point-in-time restore.

", - "DBCluster$LatestRestorableTime": "

Specifies the latest time to which a database can be restored with point-in-time restore.

", - "DBCluster$ClusterCreateTime": "

Specifies the time when the DB cluster was created, in Universal Coordinated Time (UTC).

", - "DBCluster$EarliestBacktrackTime": "

The earliest time to which a DB cluster can be backtracked.

", - "DBClusterBacktrack$BacktrackTo": "

The timestamp of the time to which the DB cluster was backtracked.

", - "DBClusterBacktrack$BacktrackedFrom": "

The timestamp of the time from which the DB cluster was backtracked.

", - "DBClusterBacktrack$BacktrackRequestCreationTime": "

The timestamp of the time at which the backtrack was requested.

", - "DBClusterSnapshot$SnapshotCreateTime": "

Provides the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", - "DBClusterSnapshot$ClusterCreateTime": "

Specifies the time when the DB cluster was created, in Universal Coordinated Time (UTC).

", - "DBInstance$InstanceCreateTime": "

Provides the date and time the DB instance was created.

", - "DBInstance$LatestRestorableTime": "

Specifies the latest time to which a database can be restored with point-in-time restore.

", - "DBSnapshot$SnapshotCreateTime": "

Provides the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", - "DBSnapshot$InstanceCreateTime": "

Specifies the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", - "DescribeEventsMessage$StartTime": "

The beginning of the time interval to retrieve events for, specified in ISO 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia page.

Example: 2009-07-08T18:00Z

", - "DescribeEventsMessage$EndTime": "

The end of the time interval for which to retrieve events, specified in ISO 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia page.

Example: 2009-07-08T18:00Z

", - "Event$Date": "

Specifies the date and time of the event.

", - "PendingMaintenanceAction$AutoAppliedAfterDate": "

The date of the maintenance window when the action is applied. The maintenance action is applied to the resource during its first maintenance window after this date. If this date is specified, any next-maintenance opt-in requests are ignored.

", - "PendingMaintenanceAction$ForcedApplyDate": "

The date when the maintenance action is automatically applied. The maintenance action is applied to the resource on this date regardless of the maintenance window for the resource. If this date is specified, any immediate opt-in requests are ignored.

", - "PendingMaintenanceAction$CurrentApplyDate": "

The effective date when the pending maintenance action is applied to the resource. This date takes into account opt-in requests received from the ApplyPendingMaintenanceAction API, the AutoAppliedAfterDate, and the ForcedApplyDate. This value is blank if an opt-in request has not been received and nothing has been specified as AutoAppliedAfterDate or ForcedApplyDate.

", - "ReservedDBInstance$StartTime": "

The time the reservation started.

", - "RestoreDBClusterToPointInTimeMessage$RestoreToTime": "

The date and time to restore the DB cluster to.

Valid Values: Value must be a time in Universal Coordinated Time (UTC) format

Constraints:

  • Must be before the latest restorable time for the DB instance

  • Must be specified if UseLatestRestorableTime parameter is not provided

  • Cannot be specified if UseLatestRestorableTime parameter is true

  • Cannot be specified if RestoreType parameter is copy-on-write

Example: 2015-03-07T23:45:00Z

", - "RestoreDBInstanceToPointInTimeMessage$RestoreTime": "

The date and time to restore from.

Valid Values: Value must be a time in Universal Coordinated Time (UTC) format

Constraints:

  • Must be before the latest restorable time for the DB instance

  • Cannot be specified if UseLatestRestorableTime parameter is true

Example: 2009-09-07T23:45:00Z

" - } - }, - "Tag": { - "base": "

Metadata assigned to an Amazon RDS resource consisting of a key-value pair.

", - "refs": { - "TagList$member": null - } - }, - "TagList": { - "base": "

A list of tags. For more information, see Tagging Amazon RDS Resources.

", - "refs": { - "AddTagsToResourceMessage$Tags": "

The tags to be assigned to the Amazon RDS resource.

", - "CopyDBClusterParameterGroupMessage$Tags": null, - "CopyDBClusterSnapshotMessage$Tags": null, - "CopyDBParameterGroupMessage$Tags": null, - "CopyDBSnapshotMessage$Tags": null, - "CopyOptionGroupMessage$Tags": null, - "CreateDBClusterMessage$Tags": null, - "CreateDBClusterParameterGroupMessage$Tags": null, - "CreateDBClusterSnapshotMessage$Tags": "

The tags to be assigned to the DB cluster snapshot.

", - "CreateDBInstanceMessage$Tags": null, - "CreateDBInstanceReadReplicaMessage$Tags": null, - "CreateDBParameterGroupMessage$Tags": null, - "CreateDBSecurityGroupMessage$Tags": null, - "CreateDBSnapshotMessage$Tags": null, - "CreateDBSubnetGroupMessage$Tags": null, - "CreateEventSubscriptionMessage$Tags": null, - "CreateOptionGroupMessage$Tags": null, - "PurchaseReservedDBInstancesOfferingMessage$Tags": null, - "RestoreDBClusterFromS3Message$Tags": null, - "RestoreDBClusterFromSnapshotMessage$Tags": "

The tags to be assigned to the restored DB cluster.

", - "RestoreDBClusterToPointInTimeMessage$Tags": null, - "RestoreDBInstanceFromDBSnapshotMessage$Tags": null, - "RestoreDBInstanceFromS3Message$Tags": "

A list of tags to associate with this DB instance. For more information, see Tagging Amazon RDS Resources.

", - "RestoreDBInstanceToPointInTimeMessage$Tags": null, - "TagListMessage$TagList": "

List of tags returned by the ListTagsForResource operation.

" - } - }, - "TagListMessage": { - "base": "

", - "refs": { - } - }, - "Timezone": { - "base": "

A time zone associated with a DBInstance or a DBSnapshot. This data type is an element in the response to the DescribeDBInstances, the DescribeDBSnapshots, and the DescribeDBEngineVersions actions.

", - "refs": { - "SupportedTimezonesList$member": null - } - }, - "UpgradeTarget": { - "base": "

The version of the database engine that a DB instance can be upgraded to.

", - "refs": { - "ValidUpgradeTargetList$member": null - } - }, - "ValidDBInstanceModificationsMessage": { - "base": "

Information about valid modifications that you can make to your DB instance. Contains the result of a successful call to the DescribeValidDBInstanceModifications action. You can use this information when you call ModifyDBInstance.

", - "refs": { - "DescribeValidDBInstanceModificationsResult$ValidDBInstanceModificationsMessage": null - } - }, - "ValidStorageOptions": { - "base": "

Information about valid modifications that you can make to your DB instance. Contains the result of a successful call to the DescribeValidDBInstanceModifications action.

", - "refs": { - "ValidStorageOptionsList$member": null - } - }, - "ValidStorageOptionsList": { - "base": null, - "refs": { - "ValidDBInstanceModificationsMessage$Storage": "

Valid storage options for your DB instance.

" - } - }, - "ValidUpgradeTargetList": { - "base": null, - "refs": { - "DBEngineVersion$ValidUpgradeTarget": "

A list of engine versions that this database engine version can be upgraded to.

" - } - }, - "VpcSecurityGroupIdList": { - "base": null, - "refs": { - "CreateDBClusterMessage$VpcSecurityGroupIds": "

A list of EC2 VPC security groups to associate with this DB cluster.

", - "CreateDBInstanceMessage$VpcSecurityGroupIds": "

A list of EC2 VPC security groups to associate with this DB instance.

Amazon Aurora

Not applicable. The associated list of EC2 VPC security groups is managed by the DB cluster. For more information, see CreateDBCluster.

Default: The default EC2 VPC security group for the DB subnet group's VPC.

", - "ModifyDBClusterMessage$VpcSecurityGroupIds": "

A list of VPC security groups that the DB cluster will belong to.

", - "ModifyDBInstanceMessage$VpcSecurityGroupIds": "

A list of EC2 VPC security groups to authorize on this DB instance. This change is asynchronously applied as soon as possible.

Amazon Aurora

Not applicable. The associated list of EC2 VPC security groups is managed by the DB cluster. For more information, see ModifyDBCluster.

Constraints:

  • If supplied, must match existing VpcSecurityGroupIds.

", - "OptionConfiguration$VpcSecurityGroupMemberships": "

A list of VpcSecurityGroupMemebrship name strings used for this option.

", - "RestoreDBClusterFromS3Message$VpcSecurityGroupIds": "

A list of EC2 VPC security groups to associate with the restored DB cluster.

", - "RestoreDBClusterFromSnapshotMessage$VpcSecurityGroupIds": "

A list of VPC security groups that the new DB cluster will belong to.

", - "RestoreDBClusterToPointInTimeMessage$VpcSecurityGroupIds": "

A list of VPC security groups that the new DB cluster belongs to.

", - "RestoreDBInstanceFromS3Message$VpcSecurityGroupIds": "

A list of VPC security groups to associate with this DB instance.

" - } - }, - "VpcSecurityGroupMembership": { - "base": "

This data type is used as a response element for queries on VPC security group membership.

", - "refs": { - "VpcSecurityGroupMembershipList$member": null - } - }, - "VpcSecurityGroupMembershipList": { - "base": null, - "refs": { - "DBCluster$VpcSecurityGroups": "

Provides a list of VPC security groups that the DB cluster belongs to.

", - "DBInstance$VpcSecurityGroups": "

Provides a list of VPC security group elements that the DB instance belongs to.

", - "Option$VpcSecurityGroupMemberships": "

If the option requires access to a port, then this VPC security group allows access to the port.

" + "StartDBInstanceMessage":{ + "base":null, + "refs":{ + } + }, + "StartDBInstanceResult":{ + "base":null, + "refs":{ + } + }, + "StopDBInstanceMessage":{ + "base":null, + "refs":{ + } + }, + "StopDBInstanceResult":{ + "base":null, + "refs":{ + } + }, + "StorageQuotaExceededFault":{ + "base":"

The request would result in the user exceeding the allowed amount of storage available across all DB instances.

", + "refs":{ + } + }, + "StorageTypeNotSupportedFault":{ + "base":"

Storage of the StorageType specified can't be associated with the DB instance.

", + "refs":{ + } + }, + "String":{ + "base":null, + "refs":{ + "AccountQuota$AccountQuotaName":"

The name of the Amazon RDS quota for this AWS account.

", + "AddRoleToDBClusterMessage$DBClusterIdentifier":"

The name of the DB cluster to associate the IAM role with.

", + "AddRoleToDBClusterMessage$RoleArn":"

The Amazon Resource Name (ARN) of the IAM role to associate with the Aurora DB cluster, for example arn:aws:iam::123456789012:role/AuroraAccessRole.

", + "AddSourceIdentifierToSubscriptionMessage$SubscriptionName":"

The name of the RDS event notification subscription you want to add a source identifier to.

", + "AddSourceIdentifierToSubscriptionMessage$SourceIdentifier":"

The identifier of the event source to be added.

Constraints:

  • If the source type is a DB instance, then a DBInstanceIdentifier must be supplied.

  • If the source type is a DB security group, a DBSecurityGroupName must be supplied.

  • If the source type is a DB parameter group, a DBParameterGroupName must be supplied.

  • If the source type is a DB snapshot, a DBSnapshotIdentifier must be supplied.

", + "AddTagsToResourceMessage$ResourceName":"

The Amazon RDS resource that the tags are added to. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", + "ApplyPendingMaintenanceActionMessage$ResourceIdentifier":"

The RDS Amazon Resource Name (ARN) of the resource that the pending maintenance action applies to. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", + "ApplyPendingMaintenanceActionMessage$ApplyAction":"

The pending maintenance action to apply to this resource.

Valid values: system-update, db-upgrade

", + "ApplyPendingMaintenanceActionMessage$OptInType":"

A value that specifies the type of opt-in request, or undoes an opt-in request. An opt-in request of type immediate can't be undone.

Valid values:

  • immediate - Apply the maintenance action immediately.

  • next-maintenance - Apply the maintenance action during the next maintenance window for the resource.

  • undo-opt-in - Cancel any existing next-maintenance opt-in requests.

", + "AttributeValueList$member":null, + "AuthorizeDBSecurityGroupIngressMessage$DBSecurityGroupName":"

The name of the DB security group to add authorization to.

", + "AuthorizeDBSecurityGroupIngressMessage$CIDRIP":"

The IP range to authorize.

", + "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupName":"

Name of the EC2 security group to authorize. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", + "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupId":"

Id of the EC2 security group to authorize. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", + "AuthorizeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId":"

AWS account number of the owner of the EC2 security group specified in the EC2SecurityGroupName parameter. The AWS Access Key ID is not an acceptable value. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", + "AvailabilityZone$Name":"

The name of the Availability Zone.

", + "AvailabilityZones$member":null, + "AvailableProcessorFeature$Name":"

The name of the processor feature. Valid names are coreCount and threadsPerCore.

", + "AvailableProcessorFeature$DefaultValue":"

The default value for the processor feature of the DB instance class.

", + "AvailableProcessorFeature$AllowedValues":"

The allowed values for the processor feature of the DB instance class.

", + "BacktrackDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier of the DB cluster to be backtracked. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 alphanumeric characters or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", + "Certificate$CertificateIdentifier":"

The unique key that identifies a certificate.

", + "Certificate$CertificateType":"

The type of the certificate.

", + "Certificate$Thumbprint":"

The thumbprint of the certificate.

", + "Certificate$CertificateArn":"

The Amazon Resource Name (ARN) for the certificate.

", + "CertificateMessage$Marker":"

An optional pagination token provided by a previous DescribeCertificates request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "CharacterSet$CharacterSetName":"

The name of the character set.

", + "CharacterSet$CharacterSetDescription":"

The description of the character set.

", + "CopyDBClusterParameterGroupMessage$SourceDBClusterParameterGroupIdentifier":"

The identifier or Amazon Resource Name (ARN) for the source DB cluster parameter group. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid DB cluster parameter group.

  • If the source DB cluster parameter group is in the same AWS Region as the copy, specify a valid DB parameter group identifier, for example my-db-cluster-param-group, or a valid ARN.

  • If the source DB parameter group is in a different AWS Region than the copy, specify a valid DB cluster parameter group ARN, for example arn:aws:rds:us-east-1:123456789012:cluster-pg:custom-cluster-group1.

", + "CopyDBClusterParameterGroupMessage$TargetDBClusterParameterGroupIdentifier":"

The identifier for the copied DB cluster parameter group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-cluster-param-group1

", + "CopyDBClusterParameterGroupMessage$TargetDBClusterParameterGroupDescription":"

A description for the copied DB cluster parameter group.

", + "CopyDBClusterSnapshotMessage$SourceDBClusterSnapshotIdentifier":"

The identifier of the DB cluster snapshot to copy. This parameter is not case-sensitive.

You can't copy an encrypted, shared DB cluster snapshot from one AWS Region to another.

Constraints:

  • Must specify a valid system snapshot in the \"available\" state.

  • If the source snapshot is in the same AWS Region as the copy, specify a valid DB snapshot identifier.

  • If the source snapshot is in a different AWS Region than the copy, specify a valid DB cluster snapshot ARN. For more information, go to Copying a DB Snapshot or DB Cluster Snapshot.

Example: my-cluster-snapshot1

", + "CopyDBClusterSnapshotMessage$TargetDBClusterSnapshotIdentifier":"

The identifier of the new DB cluster snapshot to create from the source DB cluster snapshot. This parameter is not case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster-snapshot2

", + "CopyDBClusterSnapshotMessage$KmsKeyId":"

The AWS AWS KMS key ID for an encrypted DB cluster snapshot. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you copy an encrypted DB cluster snapshot from your AWS account, you can specify a value for KmsKeyId to encrypt the copy with a new KMS encryption key. If you don't specify a value for KmsKeyId, then the copy of the DB cluster snapshot is encrypted with the same KMS key as the source DB cluster snapshot.

If you copy an encrypted DB cluster snapshot that is shared from another AWS account, then you must specify a value for KmsKeyId.

To copy an encrypted DB cluster snapshot to another AWS Region, you must set KmsKeyId to the KMS key ID you want to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

If you copy an unencrypted DB cluster snapshot and specify a value for the KmsKeyId parameter, an error is returned.

", + "CopyDBClusterSnapshotMessage$PreSignedUrl":"

The URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot API action in the AWS Region that contains the source DB cluster snapshot to copy. The PreSignedUrl parameter must be used when copying an encrypted DB cluster snapshot from another AWS Region.

The pre-signed URL must be a valid request for the CopyDBSClusterSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB cluster snapshot to be copied. The pre-signed URL request must contain the following parameter values:

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB cluster snapshot in the destination AWS Region. This is the same identifier for both the CopyDBClusterSnapshot action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

  • DestinationRegion - The name of the AWS Region that the DB cluster snapshot will be created in.

  • SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for the encrypted DB cluster snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster snapshot from the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:aurora-cluster1-snapshot-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", + "CopyDBParameterGroupMessage$SourceDBParameterGroupIdentifier":"

The identifier or ARN for the source DB parameter group. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid DB parameter group.

  • Must specify a valid DB parameter group identifier, for example my-db-param-group, or a valid ARN.

", + "CopyDBParameterGroupMessage$TargetDBParameterGroupIdentifier":"

The identifier for the copied DB parameter group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-db-parameter-group

", + "CopyDBParameterGroupMessage$TargetDBParameterGroupDescription":"

A description for the copied DB parameter group.

", + "CopyDBSnapshotMessage$SourceDBSnapshotIdentifier":"

The identifier for the source DB snapshot.

If the source snapshot is in the same AWS Region as the copy, specify a valid DB snapshot identifier. For example, you might specify rds:mysql-instance1-snapshot-20130805.

If the source snapshot is in a different AWS Region than the copy, specify a valid DB snapshot ARN. For example, you might specify arn:aws:rds:us-west-2:123456789012:snapshot:mysql-instance1-snapshot-20130805.

If you are copying from a shared manual DB snapshot, this parameter must be the Amazon Resource Name (ARN) of the shared DB snapshot.

If you are copying an encrypted snapshot this parameter must be in the ARN format for the source AWS Region, and must match the SourceDBSnapshotIdentifier in the PreSignedUrl parameter.

Constraints:

  • Must specify a valid system snapshot in the \"available\" state.

Example: rds:mydb-2012-04-02-00-01

Example: arn:aws:rds:us-west-2:123456789012:snapshot:mysql-instance1-snapshot-20130805

", + "CopyDBSnapshotMessage$TargetDBSnapshotIdentifier":"

The identifier for the copy of the snapshot.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-db-snapshot

", + "CopyDBSnapshotMessage$KmsKeyId":"

The AWS KMS key ID for an encrypted DB snapshot. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you copy an encrypted DB snapshot from your AWS account, you can specify a value for this parameter to encrypt the copy with a new KMS encryption key. If you don't specify a value for this parameter, then the copy of the DB snapshot is encrypted with the same KMS key as the source DB snapshot.

If you copy an encrypted DB snapshot that is shared from another AWS account, then you must specify a value for this parameter.

If you specify this parameter when you copy an unencrypted snapshot, the copy is encrypted.

If you copy an encrypted snapshot to a different AWS Region, then you must specify a KMS key for the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

", + "CopyDBSnapshotMessage$PreSignedUrl":"

The URL that contains a Signature Version 4 signed request for the CopyDBSnapshot API action in the source AWS Region that contains the source DB snapshot to copy.

You must specify this parameter when you copy an encrypted DB snapshot from another AWS Region by using the Amazon RDS API. You can specify the --source-region option instead of this parameter when you copy an encrypted DB snapshot from another AWS Region by using the AWS CLI.

The presigned URL must be a valid request for the CopyDBSnapshot API action that can be executed in the source AWS Region that contains the encrypted DB snapshot to be copied. The presigned URL request must contain the following parameter values:

  • DestinationRegion - The AWS Region that the encrypted DB snapshot is copied to. This AWS Region is the same one where the CopyDBSnapshot action is called that contains this presigned URL.

    For example, if you copy an encrypted DB snapshot from the us-west-2 AWS Region to the us-east-1 AWS Region, then you call the CopyDBSnapshot action in the us-east-1 AWS Region and provide a presigned URL that contains a call to the CopyDBSnapshot action in the us-west-2 AWS Region. For this example, the DestinationRegion in the presigned URL must be set to the us-east-1 AWS Region.

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB snapshot in the destination AWS Region. This is the same identifier for both the CopyDBSnapshot action that is called in the destination AWS Region, and the action contained in the presigned URL.

  • SourceDBSnapshotIdentifier - The DB snapshot identifier for the encrypted snapshot to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB snapshot from the us-west-2 AWS Region, then your SourceDBSnapshotIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:snapshot:mysql-instance1-snapshot-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", + "CopyDBSnapshotMessage$OptionGroupName":"

The name of an option group to associate with the copy of the snapshot.

Specify this option if you are copying a snapshot from one AWS Region to another, and your DB instance uses a nondefault option group. If your source DB instance uses Transparent Data Encryption for Oracle or Microsoft SQL Server, you must specify this option when copying across AWS Regions. For more information, see Option Group Considerations.

", + "CopyOptionGroupMessage$SourceOptionGroupIdentifier":"

The identifier or ARN for the source option group. For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

Constraints:

  • Must specify a valid option group.

  • If the source option group is in the same AWS Region as the copy, specify a valid option group identifier, for example my-option-group, or a valid ARN.

  • If the source option group is in a different AWS Region than the copy, specify a valid option group ARN, for example arn:aws:rds:us-west-2:123456789012:og:special-options.

", + "CopyOptionGroupMessage$TargetOptionGroupIdentifier":"

The identifier for the copied option group.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-option-group

", + "CopyOptionGroupMessage$TargetOptionGroupDescription":"

The description for the copied option group.

", + "CreateDBClusterMessage$CharacterSetName":"

A value that indicates that the DB cluster should be associated with the specified CharacterSet.

", + "CreateDBClusterMessage$DatabaseName":"

The name for your database of up to 64 alpha-numeric characters. If you do not provide a name, Amazon RDS will not create a database in the DB cluster you are creating.

", + "CreateDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", + "CreateDBClusterMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to associate with this DB cluster. If this argument is omitted, default.aurora5.6 is used.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "CreateDBClusterMessage$DBSubnetGroupName":"

A DB subnet group to associate with this DB cluster.

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", + "CreateDBClusterMessage$Engine":"

The name of the database engine to be used for this DB cluster.

Valid Values: aurora (for MySQL 5.6-compatible Aurora), aurora-mysql (for MySQL 5.7-compatible Aurora), and aurora-postgresql

", + "CreateDBClusterMessage$EngineVersion":"

The version number of the database engine to use.

Aurora MySQL

Example: 5.6.10a, 5.7.12

Aurora PostgreSQL

Example: 9.6.3

", + "CreateDBClusterMessage$MasterUsername":"

The name of the master user for the DB cluster.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", + "CreateDBClusterMessage$MasterUserPassword":"

The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", + "CreateDBClusterMessage$OptionGroupName":"

A value that indicates that the DB cluster should be associated with the specified option group.

Permanent options can't be removed from an option group. The option group can't be removed from a DB cluster once it is associated with a DB cluster.

", + "CreateDBClusterMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "CreateDBClusterMessage$PreferredMaintenanceWindow":"

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "CreateDBClusterMessage$ReplicationSourceIdentifier":"

The Amazon Resource Name (ARN) of the source DB instance or DB cluster if this DB cluster is created as a Read Replica.

", + "CreateDBClusterMessage$KmsKeyId":"

The AWS KMS key identifier for an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

If an encryption key is not specified in KmsKeyId:

  • If ReplicationSourceIdentifier identifies an encrypted source, then Amazon RDS will use the encryption key used to encrypt the source. Otherwise, Amazon RDS will use your default encryption key.

  • If the StorageEncrypted parameter is true and ReplicationSourceIdentifier is not specified, then Amazon RDS will use your default encryption key.

AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

If you create a Read Replica of an encrypted DB cluster in another AWS Region, you must set KmsKeyId to a KMS key ID that is valid in the destination AWS Region. This key is used to encrypt the Read Replica in that AWS Region.

", + "CreateDBClusterMessage$PreSignedUrl":"

A URL that contains a Signature Version 4 signed request for the CreateDBCluster action to be called in the source AWS Region where the DB cluster is replicated from. You only need to specify PreSignedUrl when you are performing cross-region replication from an encrypted DB cluster.

The pre-signed URL must be a valid request for the CreateDBCluster API action that can be executed in the source AWS Region that contains the encrypted DB cluster to be copied.

The pre-signed URL request must contain the following parameter values:

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the copy of the DB cluster in the destination AWS Region. This should refer to the same KMS key for both the CreateDBCluster action that is called in the destination AWS Region, and the action contained in the pre-signed URL.

  • DestinationRegion - The name of the AWS Region that Aurora Read Replica will be created in.

  • ReplicationSourceIdentifier - The DB cluster identifier for the encrypted DB cluster to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are copying an encrypted DB cluster from the us-west-2 AWS Region, then your ReplicationSourceIdentifier would look like Example: arn:aws:rds:us-west-2:123456789012:cluster:aurora-cluster1.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", + "CreateDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group.

Constraints:

  • Must match the name of an existing DBClusterParameterGroup.

This value is stored as a lowercase string.

", + "CreateDBClusterParameterGroupMessage$DBParameterGroupFamily":"

The DB cluster parameter group family name. A DB cluster parameter group can be associated with one and only one DB cluster parameter group family, and can be applied only to a DB cluster running a database engine and engine version compatible with that DB cluster parameter group family.

Aurora MySQL

Example: aurora5.6, aurora-mysql5.7

Aurora PostgreSQL

Example: aurora-postgresql9.6

", + "CreateDBClusterParameterGroupMessage$Description":"

The description for the DB cluster parameter group.

", + "CreateDBClusterSnapshotMessage$DBClusterSnapshotIdentifier":"

The identifier of the DB cluster snapshot. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1-snapshot1

", + "CreateDBClusterSnapshotMessage$DBClusterIdentifier":"

The identifier of the DB cluster to create a snapshot for. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

Example: my-cluster1

", + "CreateDBInstanceMessage$DBName":"

The meaning of this parameter differs according to the database engine you use.

Type: String

MySQL

The name of the database to create when the DB instance is created. If this parameter is not specified, no database is created in the DB instance.

Constraints:

  • Must contain 1 to 64 letters or numbers.

  • Cannot be a word reserved by the specified database engine

MariaDB

The name of the database to create when the DB instance is created. If this parameter is not specified, no database is created in the DB instance.

Constraints:

  • Must contain 1 to 64 letters or numbers.

  • Cannot be a word reserved by the specified database engine

PostgreSQL

The name of the database to create when the DB instance is created. If this parameter is not specified, the default \"postgres\" database is created in the DB instance.

Constraints:

  • Must contain 1 to 63 letters, numbers, or underscores.

  • Must begin with a letter or an underscore. Subsequent characters can be letters, underscores, or digits (0-9).

  • Cannot be a word reserved by the specified database engine

Oracle

The Oracle System ID (SID) of the created DB instance. If you specify null, the default value ORCL is used. You can't specify the string NULL, or any other reserved word, for DBName.

Default: ORCL

Constraints:

  • Cannot be longer than 8 characters

SQL Server

Not applicable. Must be null.

Amazon Aurora

The name of the database to create when the primary instance of the DB cluster is created. If this parameter is not specified, no database is created in the DB instance.

Constraints:

  • Must contain 1 to 64 letters or numbers.

  • Cannot be a word reserved by the specified database engine

", + "CreateDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", + "CreateDBInstanceMessage$DBInstanceClass":"

The compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

", + "CreateDBInstanceMessage$Engine":"

The name of the database engine to be used for this instance.

Not every database engine is available for every AWS Region.

Valid Values:

  • aurora (for MySQL 5.6-compatible Aurora)

  • aurora-mysql (for MySQL 5.7-compatible Aurora)

  • aurora-postgresql

  • mariadb

  • mysql

  • oracle-ee

  • oracle-se2

  • oracle-se1

  • oracle-se

  • postgres

  • sqlserver-ee

  • sqlserver-se

  • sqlserver-ex

  • sqlserver-web

", + "CreateDBInstanceMessage$MasterUsername":"

The name for the master user.

Amazon Aurora

Not applicable. The name for the master user is managed by the DB cluster. For more information, see CreateDBCluster.

MariaDB

Constraints:

  • Required for MariaDB.

  • Must be 1 to 16 letters or numbers.

  • Cannot be a reserved word for the chosen database engine.

Microsoft SQL Server

Constraints:

  • Required for SQL Server.

  • Must be 1 to 128 letters or numbers.

  • The first character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

MySQL

Constraints:

  • Required for MySQL.

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

Oracle

Constraints:

  • Required for Oracle.

  • Must be 1 to 30 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

PostgreSQL

Constraints:

  • Required for PostgreSQL.

  • Must be 1 to 63 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", + "CreateDBInstanceMessage$MasterUserPassword":"

The password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Amazon Aurora

Not applicable. The password for the master user is managed by the DB cluster. For more information, see CreateDBCluster.

MariaDB

Constraints: Must contain from 8 to 41 characters.

Microsoft SQL Server

Constraints: Must contain from 8 to 128 characters.

MySQL

Constraints: Must contain from 8 to 41 characters.

Oracle

Constraints: Must contain from 8 to 30 characters.

PostgreSQL

Constraints: Must contain from 8 to 128 characters.

", + "CreateDBInstanceMessage$AvailabilityZone":"

The EC2 Availability Zone that the DB instance is created in. For information on AWS Regions and Availability Zones, see Regions and Availability Zones.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

Constraint: The AvailabilityZone parameter can't be specified if the MultiAZ parameter is set to true. The specified Availability Zone must be in the same AWS Region as the current endpoint.

", + "CreateDBInstanceMessage$DBSubnetGroupName":"

A DB subnet group to associate with this DB instance.

If there is no DB subnet group, then it is a non-VPC DB instance.

", + "CreateDBInstanceMessage$PreferredMaintenanceWindow":"

The time range each week during which system maintenance can occur, in Universal Coordinated Time (UTC). For more information, see Amazon RDS Maintenance Window.

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "CreateDBInstanceMessage$DBParameterGroupName":"

The name of the DB parameter group to associate with this DB instance. If this argument is omitted, the default DBParameterGroup for the specified engine is used.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "CreateDBInstanceMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter. For more information, see The Backup Window.

Amazon Aurora

Not applicable. The daily time range for creating automated backups is managed by the DB cluster. For more information, see CreateDBCluster.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred DB Instance Maintenance Window.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "CreateDBInstanceMessage$EngineVersion":"

The version number of the database engine to use.

For a list of valid engine versions, call DescribeDBEngineVersions.

The following are the database engines and links to information about the major and minor versions that are available with Amazon RDS. Not every database engine is available for every AWS Region.

Amazon Aurora

Not applicable. The version number of the database engine to be used by the DB instance is managed by the DB cluster. For more information, see CreateDBCluster.

MariaDB

See MariaDB on Amazon RDS Versions in the Amazon RDS User Guide.

Microsoft SQL Server

See Version and Feature Support on Amazon RDS in the Amazon RDS User Guide.

MySQL

See MySQL on Amazon RDS Versions in the Amazon RDS User Guide.

Oracle

See Oracle Database Engine Release Notes in the Amazon RDS User Guide.

PostgreSQL

See Supported PostgreSQL Database Versions in the Amazon RDS User Guide.

", + "CreateDBInstanceMessage$LicenseModel":"

License model information for this DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", + "CreateDBInstanceMessage$OptionGroupName":"

Indicates that the DB instance should be associated with the specified option group.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "CreateDBInstanceMessage$CharacterSetName":"

For supported engines, indicates that the DB instance should be associated with the specified CharacterSet.

Amazon Aurora

Not applicable. The character set is managed by the DB cluster. For more information, see CreateDBCluster.

", + "CreateDBInstanceMessage$DBClusterIdentifier":"

The identifier of the DB cluster that the instance will belong to.

For information on creating a DB cluster, see CreateDBCluster.

Type: String

", + "CreateDBInstanceMessage$StorageType":"

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", + "CreateDBInstanceMessage$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "CreateDBInstanceMessage$TdeCredentialPassword":"

The password for the given ARN from the key store in order to access the device.

", + "CreateDBInstanceMessage$KmsKeyId":"

The AWS KMS key identifier for an encrypted DB instance.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB instance with the same AWS account that owns the KMS encryption key used to encrypt the new DB instance, then you can use the KMS key alias instead of the ARN for the KM encryption key.

Amazon Aurora

Not applicable. The KMS key identifier is managed by the DB cluster. For more information, see CreateDBCluster.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon RDS will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", + "CreateDBInstanceMessage$Domain":"

Specify the Active Directory Domain to create the instance in.

", + "CreateDBInstanceMessage$MonitoringRoleArn":"

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, go to Setting Up and Enabling Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", + "CreateDBInstanceMessage$DomainIAMRoleName":"

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", + "CreateDBInstanceMessage$Timezone":"

The time zone of the DB instance. The time zone parameter is currently supported only by Microsoft SQL Server.

", + "CreateDBInstanceMessage$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "CreateDBInstanceReadReplicaMessage$DBInstanceIdentifier":"

The DB instance identifier of the Read Replica. This identifier is the unique key that identifies a DB instance. This parameter is stored as a lowercase string.

", + "CreateDBInstanceReadReplicaMessage$SourceDBInstanceIdentifier":"

The identifier of the DB instance that will act as the source for the Read Replica. Each DB instance can have up to five Read Replicas.

Constraints:

  • Must be the identifier of an existing MySQL, MariaDB, or PostgreSQL DB instance.

  • Can specify a DB instance that is a MySQL Read Replica only if the source is running MySQL 5.6.

  • Can specify a DB instance that is a PostgreSQL DB instance only if the source is running PostgreSQL 9.3.5 or later (9.4.7 and higher for cross-region replication).

  • The specified DB instance must have automatic backups enabled, its backup retention period must be greater than 0.

  • If the source DB instance is in the same AWS Region as the Read Replica, specify a valid DB instance identifier.

  • If the source DB instance is in a different AWS Region than the Read Replica, specify a valid DB instance ARN. For more information, go to Constructing a Amazon RDS Amazon Resource Name (ARN).

", + "CreateDBInstanceReadReplicaMessage$DBInstanceClass":"

The compute and memory capacity of the Read Replica, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Default: Inherits from the source DB instance.

", + "CreateDBInstanceReadReplicaMessage$AvailabilityZone":"

The Amazon EC2 Availability Zone that the Read Replica is created in.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

", + "CreateDBInstanceReadReplicaMessage$OptionGroupName":"

The option group the DB instance is associated with. If omitted, the default option group for the engine specified is used.

", + "CreateDBInstanceReadReplicaMessage$DBSubnetGroupName":"

Specifies a DB subnet group for the DB instance. The new DB instance is created in the VPC associated with the DB subnet group. If no DB subnet group is specified, then the new DB instance is not created in a VPC.

Constraints:

  • Can only be specified if the source DB instance identifier specifies a DB instance in another AWS Region.

  • If supplied, must match the name of an existing DBSubnetGroup.

  • The specified DB subnet group must be in the same AWS Region in which the operation is running.

  • All Read Replicas in one AWS Region that are created from the same source DB instance must either:>

    • Specify DB subnet groups from the same VPC. All these Read Replicas are created in the same VPC.

    • Not specify a DB subnet group. All these Read Replicas are created outside of any VPC.

Example: mySubnetgroup

", + "CreateDBInstanceReadReplicaMessage$StorageType":"

Specifies the storage type to be associated with the Read Replica.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", + "CreateDBInstanceReadReplicaMessage$MonitoringRoleArn":"

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, go to To create an IAM role for Amazon RDS Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", + "CreateDBInstanceReadReplicaMessage$KmsKeyId":"

The AWS KMS key ID for an encrypted Read Replica. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

If you specify this parameter when you create a Read Replica from an unencrypted DB instance, the Read Replica is encrypted.

If you create an encrypted Read Replica in the same AWS Region as the source DB instance, then you do not have to specify a value for this parameter. The Read Replica is encrypted with the same KMS key as the source DB instance.

If you create an encrypted Read Replica in a different AWS Region, then you must specify a KMS key for the destination AWS Region. KMS encryption keys are specific to the AWS Region that they are created in, and you can't use encryption keys from one AWS Region in another AWS Region.

", + "CreateDBInstanceReadReplicaMessage$PreSignedUrl":"

The URL that contains a Signature Version 4 signed request for the CreateDBInstanceReadReplica API action in the source AWS Region that contains the source DB instance.

You must specify this parameter when you create an encrypted Read Replica from another AWS Region by using the Amazon RDS API. You can specify the --source-region option instead of this parameter when you create an encrypted Read Replica from another AWS Region by using the AWS CLI.

The presigned URL must be a valid request for the CreateDBInstanceReadReplica API action that can be executed in the source AWS Region that contains the encrypted source DB instance. The presigned URL request must contain the following parameter values:

  • DestinationRegion - The AWS Region that the encrypted Read Replica is created in. This AWS Region is the same one where the CreateDBInstanceReadReplica action is called that contains this presigned URL.

    For example, if you create an encrypted DB instance in the us-west-1 AWS Region, from a source DB instance in the us-east-2 AWS Region, then you call the CreateDBInstanceReadReplica action in the us-east-1 AWS Region and provide a presigned URL that contains a call to the CreateDBInstanceReadReplica action in the us-west-2 AWS Region. For this example, the DestinationRegion in the presigned URL must be set to the us-east-1 AWS Region.

  • KmsKeyId - The AWS KMS key identifier for the key to use to encrypt the Read Replica in the destination AWS Region. This is the same identifier for both the CreateDBInstanceReadReplica action that is called in the destination AWS Region, and the action contained in the presigned URL.

  • SourceDBInstanceIdentifier - The DB instance identifier for the encrypted DB instance to be replicated. This identifier must be in the Amazon Resource Name (ARN) format for the source AWS Region. For example, if you are creating an encrypted Read Replica from a DB instance in the us-west-2 AWS Region, then your SourceDBInstanceIdentifier looks like the following example: arn:aws:rds:us-west-2:123456789012:instance:mysql-instance1-20161115.

To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (AWS Signature Version 4) and Signature Version 4 Signing Process.

", + "CreateDBInstanceReadReplicaMessage$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "CreateDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

This value is stored as a lowercase string.

", + "CreateDBParameterGroupMessage$DBParameterGroupFamily":"

The DB parameter group family name. A DB parameter group can be associated with one and only one DB parameter group family, and can be applied only to a DB instance running a database engine and engine version compatible with that DB parameter group family.

To list all of the available parameter group families, use the following command:

aws rds describe-db-engine-versions --query \"DBEngineVersions[].DBParameterGroupFamily\"

The output contains duplicates.

", + "CreateDBParameterGroupMessage$Description":"

The description for the DB parameter group.

", + "CreateDBSecurityGroupMessage$DBSecurityGroupName":"

The name for the DB security group. This value is stored as a lowercase string.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Must not be \"Default\"

Example: mysecuritygroup

", + "CreateDBSecurityGroupMessage$DBSecurityGroupDescription":"

The description for the DB security group.

", + "CreateDBSnapshotMessage$DBSnapshotIdentifier":"

The identifier for the DB snapshot.

Constraints:

  • Cannot be null, empty, or blank

  • Must contain from 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", + "CreateDBSnapshotMessage$DBInstanceIdentifier":"

The identifier of the DB instance that you want to create the snapshot of.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "CreateDBSubnetGroupMessage$DBSubnetGroupName":"

The name for the DB subnet group. This value is stored as a lowercase string.

Constraints: Must contain no more than 255 letters, numbers, periods, underscores, spaces, or hyphens. Must not be default.

Example: mySubnetgroup

", + "CreateDBSubnetGroupMessage$DBSubnetGroupDescription":"

The description for the DB subnet group.

", + "CreateEventSubscriptionMessage$SubscriptionName":"

The name of the subscription.

Constraints: The name must be less than 255 characters.

", + "CreateEventSubscriptionMessage$SnsTopicArn":"

The Amazon Resource Name (ARN) of the SNS topic created for event notification. The ARN is created by Amazon SNS when you create a topic and subscribe to it.

", + "CreateEventSubscriptionMessage$SourceType":"

The type of source that is generating the events. For example, if you want to be notified of events generated by a DB instance, you would set this parameter to db-instance. if this value is not specified, all events are returned.

Valid values: db-instance | db-cluster | db-parameter-group | db-security-group | db-snapshot | db-cluster-snapshot

", + "CreateOptionGroupMessage$OptionGroupName":"

Specifies the name of the option group to be created.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: myoptiongroup

", + "CreateOptionGroupMessage$EngineName":"

Specifies the name of the engine that this option group should be associated with.

", + "CreateOptionGroupMessage$MajorEngineVersion":"

Specifies the major version of the engine that this option group should be associated with.

", + "CreateOptionGroupMessage$OptionGroupDescription":"

The description of the option group.

", + "DBCluster$CharacterSetName":"

If present, specifies the name of the character set that this cluster is associated with.

", + "DBCluster$DatabaseName":"

Contains the name of the initial database of this DB cluster that was provided at create time, if one was specified when the DB cluster was created. This same name is returned for the life of the DB cluster.

", + "DBCluster$DBClusterIdentifier":"

Contains a user-supplied DB cluster identifier. This identifier is the unique key that identifies a DB cluster.

", + "DBCluster$DBClusterParameterGroup":"

Specifies the name of the DB cluster parameter group for the DB cluster.

", + "DBCluster$DBSubnetGroup":"

Specifies information on the subnet group associated with the DB cluster, including the name, description, and subnets in the subnet group.

", + "DBCluster$Status":"

Specifies the current state of this DB cluster.

", + "DBCluster$PercentProgress":"

Specifies the progress of the operation as a percentage.

", + "DBCluster$Endpoint":"

Specifies the connection endpoint for the primary instance of the DB cluster.

", + "DBCluster$ReaderEndpoint":"

The reader endpoint for the DB cluster. The reader endpoint for a DB cluster load-balances connections across the Aurora Replicas that are available in a DB cluster. As clients request new connections to the reader endpoint, Aurora distributes the connection requests among the Aurora Replicas in the DB cluster. This functionality can help balance your read workload across multiple Aurora Replicas in your DB cluster.

If a failover occurs, and the Aurora Replica that you are connected to is promoted to be the primary instance, your connection is dropped. To continue sending your read workload to other Aurora Replicas in the cluster, you can then reconnect to the reader endpoint.

", + "DBCluster$Engine":"

Provides the name of the database engine to be used for this DB cluster.

", + "DBCluster$EngineVersion":"

Indicates the database engine version.

", + "DBCluster$MasterUsername":"

Contains the master username for the DB cluster.

", + "DBCluster$PreferredBackupWindow":"

Specifies the daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod.

", + "DBCluster$PreferredMaintenanceWindow":"

Specifies the weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

", + "DBCluster$ReplicationSourceIdentifier":"

Contains the identifier of the source DB cluster if this DB cluster is a Read Replica.

", + "DBCluster$HostedZoneId":"

Specifies the ID that Amazon Route 53 assigns when you create a hosted zone.

", + "DBCluster$KmsKeyId":"

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB cluster.

", + "DBCluster$DbClusterResourceId":"

The AWS Region-unique, immutable identifier for the DB cluster. This identifier is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB cluster is accessed.

", + "DBCluster$DBClusterArn":"

The Amazon Resource Name (ARN) for the DB cluster.

", + "DBCluster$CloneGroupId":"

Identifies the clone group to which the DB cluster is associated.

", + "DBClusterBacktrack$DBClusterIdentifier":"

Contains a user-supplied DB cluster identifier. This identifier is the unique key that identifies a DB cluster.

", + "DBClusterBacktrack$BacktrackIdentifier":"

Contains the backtrack identifier.

", + "DBClusterBacktrack$Status":"

The status of the backtrack. This property returns one of the following values:

  • applying - The backtrack is currently being applied to or rolled back from the DB cluster.

  • completed - The backtrack has successfully been applied to or rolled back from the DB cluster.

  • failed - An error occurred while the backtrack was applied to or rolled back from the DB cluster.

  • pending - The backtrack is currently pending application to or rollback from the DB cluster.

", + "DBClusterBacktrackMessage$Marker":"

A pagination token that can be used in a subsequent DescribeDBClusterBacktracks request.

", + "DBClusterMember$DBInstanceIdentifier":"

Specifies the instance identifier for this member of the DB cluster.

", + "DBClusterMember$DBClusterParameterGroupStatus":"

Specifies the status of the DB cluster parameter group for this member of the DB cluster.

", + "DBClusterMessage$Marker":"

A pagination token that can be used in a subsequent DescribeDBClusters request.

", + "DBClusterOptionGroupStatus$DBClusterOptionGroupName":"

Specifies the name of the DB cluster option group.

", + "DBClusterOptionGroupStatus$Status":"

Specifies the status of the DB cluster option group.

", + "DBClusterParameterGroup$DBClusterParameterGroupName":"

Provides the name of the DB cluster parameter group.

", + "DBClusterParameterGroup$DBParameterGroupFamily":"

Provides the name of the DB parameter group family that this DB cluster parameter group is compatible with.

", + "DBClusterParameterGroup$Description":"

Provides the customer-specified description for this DB cluster parameter group.

", + "DBClusterParameterGroup$DBClusterParameterGroupArn":"

The Amazon Resource Name (ARN) for the DB cluster parameter group.

", + "DBClusterParameterGroupDetails$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DBClusterParameterGroupNameMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group.

Constraints:

  • Must be 1 to 255 letters or numbers.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

This value is stored as a lowercase string.

", + "DBClusterParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBClusterRole$RoleArn":"

The Amazon Resource Name (ARN) of the IAM role that is associated with the DB cluster.

", + "DBClusterRole$Status":"

Describes the state of association between the IAM role and the DB cluster. The Status property returns one of the following values:

  • ACTIVE - the IAM role ARN is associated with the DB cluster and can be used to access other AWS services on your behalf.

  • PENDING - the IAM role ARN is being associated with the DB cluster.

  • INVALID - the IAM role ARN is associated with the DB cluster, but the DB cluster is unable to assume the IAM role in order to access other AWS services on your behalf.

", + "DBClusterSnapshot$DBClusterSnapshotIdentifier":"

Specifies the identifier for the DB cluster snapshot.

", + "DBClusterSnapshot$DBClusterIdentifier":"

Specifies the DB cluster identifier of the DB cluster that this DB cluster snapshot was created from.

", + "DBClusterSnapshot$Engine":"

Specifies the name of the database engine.

", + "DBClusterSnapshot$Status":"

Specifies the status of this DB cluster snapshot.

", + "DBClusterSnapshot$VpcId":"

Provides the VPC ID associated with the DB cluster snapshot.

", + "DBClusterSnapshot$MasterUsername":"

Provides the master username for the DB cluster snapshot.

", + "DBClusterSnapshot$EngineVersion":"

Provides the version of the database engine for this DB cluster snapshot.

", + "DBClusterSnapshot$LicenseModel":"

Provides the license model information for this DB cluster snapshot.

", + "DBClusterSnapshot$SnapshotType":"

Provides the type of the DB cluster snapshot.

", + "DBClusterSnapshot$KmsKeyId":"

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB cluster snapshot.

", + "DBClusterSnapshot$DBClusterSnapshotArn":"

The Amazon Resource Name (ARN) for the DB cluster snapshot.

", + "DBClusterSnapshot$SourceDBClusterSnapshotArn":"

If the DB cluster snapshot was copied from a source DB cluster snapshot, the Amazon Resource Name (ARN) for the source DB cluster snapshot, otherwise, a null value.

", + "DBClusterSnapshotAttribute$AttributeName":"

The name of the manual DB cluster snapshot attribute.

The attribute named restore refers to the list of AWS accounts that have permission to copy or restore the manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute API action.

", + "DBClusterSnapshotAttributesResult$DBClusterSnapshotIdentifier":"

The identifier of the manual DB cluster snapshot that the attributes apply to.

", + "DBClusterSnapshotMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBEngineVersion$Engine":"

The name of the database engine.

", + "DBEngineVersion$EngineVersion":"

The version number of the database engine.

", + "DBEngineVersion$DBParameterGroupFamily":"

The name of the DB parameter group family for the database engine.

", + "DBEngineVersion$DBEngineDescription":"

The description of the database engine.

", + "DBEngineVersion$DBEngineVersionDescription":"

The description of the database engine version.

", + "DBEngineVersionMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBInstance$DBInstanceIdentifier":"

Contains a user-supplied database identifier. This identifier is the unique key that identifies a DB instance.

", + "DBInstance$DBInstanceClass":"

Contains the name of the compute and memory capacity class of the DB instance.

", + "DBInstance$Engine":"

Provides the name of the database engine to be used for this DB instance.

", + "DBInstance$DBInstanceStatus":"

Specifies the current state of this database.

", + "DBInstance$MasterUsername":"

Contains the master username for the DB instance.

", + "DBInstance$DBName":"

The meaning of this parameter differs according to the database engine you use. For example, this value returns MySQL, MariaDB, or PostgreSQL information when returning values from CreateDBInstanceReadReplica since Read Replicas are only supported for these engines.

MySQL, MariaDB, SQL Server, PostgreSQL

Contains the name of the initial database of this instance that was provided at create time, if one was specified when the DB instance was created. This same name is returned for the life of the DB instance.

Type: String

Oracle

Contains the Oracle System ID (SID) of the created DB instance. Not shown when the returned parameters do not apply to an Oracle DB instance.

", + "DBInstance$PreferredBackupWindow":"

Specifies the daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod.

", + "DBInstance$AvailabilityZone":"

Specifies the name of the Availability Zone the DB instance is located in.

", + "DBInstance$PreferredMaintenanceWindow":"

Specifies the weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

", + "DBInstance$EngineVersion":"

Indicates the database engine version.

", + "DBInstance$ReadReplicaSourceDBInstanceIdentifier":"

Contains the identifier of the source DB instance if this DB instance is a Read Replica.

", + "DBInstance$LicenseModel":"

License model information for this DB instance.

", + "DBInstance$CharacterSetName":"

If present, specifies the name of the character set that this instance is associated with.

", + "DBInstance$SecondaryAvailabilityZone":"

If present, specifies the name of the secondary Availability Zone for a DB instance with multi-AZ support.

", + "DBInstance$StorageType":"

Specifies the storage type associated with DB instance.

", + "DBInstance$TdeCredentialArn":"

The ARN from the key store with which the instance is associated for TDE encryption.

", + "DBInstance$DBClusterIdentifier":"

If the DB instance is a member of a DB cluster, contains the name of the DB cluster that the DB instance is a member of.

", + "DBInstance$KmsKeyId":"

If StorageEncrypted is true, the AWS KMS key identifier for the encrypted DB instance.

", + "DBInstance$DbiResourceId":"

The AWS Region-unique, immutable identifier for the DB instance. This identifier is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB instance is accessed.

", + "DBInstance$CACertificateIdentifier":"

The identifier of the CA certificate for this DB instance.

", + "DBInstance$EnhancedMonitoringResourceArn":"

The Amazon Resource Name (ARN) of the Amazon CloudWatch Logs log stream that receives the Enhanced Monitoring metrics data for the DB instance.

", + "DBInstance$MonitoringRoleArn":"

The ARN for the IAM role that permits RDS to send Enhanced Monitoring metrics to Amazon CloudWatch Logs.

", + "DBInstance$DBInstanceArn":"

The Amazon Resource Name (ARN) for the DB instance.

", + "DBInstance$Timezone":"

The time zone of the DB instance. In most cases, the Timezone element is empty. Timezone content appears only for Microsoft SQL Server DB instances that were created with a time zone specified.

", + "DBInstance$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "DBInstanceMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DBInstanceStatusInfo$StatusType":"

This value is currently \"read replication.\"

", + "DBInstanceStatusInfo$Status":"

Status of the DB instance. For a StatusType of read replica, the values can be replicating, error, stopped, or terminated.

", + "DBInstanceStatusInfo$Message":"

Details of the error if there is an error for the instance. If the instance is not in an error state, this value is blank.

", + "DBParameterGroup$DBParameterGroupName":"

Provides the name of the DB parameter group.

", + "DBParameterGroup$DBParameterGroupFamily":"

Provides the name of the DB parameter group family that this DB parameter group is compatible with.

", + "DBParameterGroup$Description":"

Provides the customer-specified description for this DB parameter group.

", + "DBParameterGroup$DBParameterGroupArn":"

The Amazon Resource Name (ARN) for the DB parameter group.

", + "DBParameterGroupDetails$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBParameterGroupNameMessage$DBParameterGroupName":"

Provides the name of the DB parameter group.

", + "DBParameterGroupStatus$DBParameterGroupName":"

The name of the DP parameter group.

", + "DBParameterGroupStatus$ParameterApplyStatus":"

The status of parameter updates.

", + "DBParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBSecurityGroup$OwnerId":"

Provides the AWS ID of the owner of a specific DB security group.

", + "DBSecurityGroup$DBSecurityGroupName":"

Specifies the name of the DB security group.

", + "DBSecurityGroup$DBSecurityGroupDescription":"

Provides the description of the DB security group.

", + "DBSecurityGroup$VpcId":"

Provides the VpcId of the DB security group.

", + "DBSecurityGroup$DBSecurityGroupArn":"

The Amazon Resource Name (ARN) for the DB security group.

", + "DBSecurityGroupMembership$DBSecurityGroupName":"

The name of the DB security group.

", + "DBSecurityGroupMembership$Status":"

The status of the DB security group.

", + "DBSecurityGroupMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBSecurityGroupNameList$member":null, + "DBSnapshot$DBSnapshotIdentifier":"

Specifies the identifier for the DB snapshot.

", + "DBSnapshot$DBInstanceIdentifier":"

Specifies the DB instance identifier of the DB instance this DB snapshot was created from.

", + "DBSnapshot$Engine":"

Specifies the name of the database engine.

", + "DBSnapshot$Status":"

Specifies the status of this DB snapshot.

", + "DBSnapshot$AvailabilityZone":"

Specifies the name of the Availability Zone the DB instance was located in at the time of the DB snapshot.

", + "DBSnapshot$VpcId":"

Provides the VPC ID associated with the DB snapshot.

", + "DBSnapshot$MasterUsername":"

Provides the master username for the DB snapshot.

", + "DBSnapshot$EngineVersion":"

Specifies the version of the database engine.

", + "DBSnapshot$LicenseModel":"

License model information for the restored DB instance.

", + "DBSnapshot$SnapshotType":"

Provides the type of the DB snapshot.

", + "DBSnapshot$OptionGroupName":"

Provides the option group name for the DB snapshot.

", + "DBSnapshot$SourceRegion":"

The AWS Region that the DB snapshot was created in or copied from.

", + "DBSnapshot$SourceDBSnapshotIdentifier":"

The DB snapshot Amazon Resource Name (ARN) that the DB snapshot was copied from. It only has value in case of cross-customer or cross-region copy.

", + "DBSnapshot$StorageType":"

Specifies the storage type associated with DB snapshot.

", + "DBSnapshot$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "DBSnapshot$KmsKeyId":"

If Encrypted is true, the AWS KMS key identifier for the encrypted DB snapshot.

", + "DBSnapshot$DBSnapshotArn":"

The Amazon Resource Name (ARN) for the DB snapshot.

", + "DBSnapshot$Timezone":"

The time zone of the DB snapshot. In most cases, the Timezone element is empty. Timezone content appears only for snapshots taken from Microsoft SQL Server DB instances that were created with a time zone specified.

", + "DBSnapshotAttribute$AttributeName":"

The name of the manual DB snapshot attribute.

The attribute named restore refers to the list of AWS accounts that have permission to copy or restore the manual DB cluster snapshot. For more information, see the ModifyDBSnapshotAttribute API action.

", + "DBSnapshotAttributesResult$DBSnapshotIdentifier":"

The identifier of the manual DB snapshot that the attributes apply to.

", + "DBSnapshotMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DBSubnetGroup$DBSubnetGroupName":"

The name of the DB subnet group.

", + "DBSubnetGroup$DBSubnetGroupDescription":"

Provides the description of the DB subnet group.

", + "DBSubnetGroup$VpcId":"

Provides the VpcId of the DB subnet group.

", + "DBSubnetGroup$SubnetGroupStatus":"

Provides the status of the DB subnet group.

", + "DBSubnetGroup$DBSubnetGroupArn":"

The Amazon Resource Name (ARN) for the DB subnet group.

", + "DBSubnetGroupMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DeleteDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier for the DB cluster to be deleted. This parameter isn't case-sensitive.

Constraints:

  • Must match an existing DBClusterIdentifier.

", + "DeleteDBClusterMessage$FinalDBSnapshotIdentifier":"

The DB cluster snapshot identifier of the new DB cluster snapshot created when SkipFinalSnapshot is set to false.

Specifying this parameter and also setting the SkipFinalShapshot parameter to true results in an error.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "DeleteDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group.

Constraints:

  • Must be the name of an existing DB cluster parameter group.

  • You can't delete a default DB cluster parameter group.

  • Cannot be associated with any DB clusters.

", + "DeleteDBClusterSnapshotMessage$DBClusterSnapshotIdentifier":"

The identifier of the DB cluster snapshot to delete.

Constraints: Must be the name of an existing DB cluster snapshot in the available state.

", + "DeleteDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier for the DB instance to be deleted. This parameter isn't case-sensitive.

Constraints:

  • Must match the name of an existing DB instance.

", + "DeleteDBInstanceMessage$FinalDBSnapshotIdentifier":"

The DBSnapshotIdentifier of the new DBSnapshot created when SkipFinalSnapshot is set to false.

Specifying this parameter and also setting the SkipFinalShapshot parameter to true results in an error.

Constraints:

  • Must be 1 to 255 letters or numbers.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Cannot be specified when deleting a Read Replica.

", + "DeleteDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • Must be the name of an existing DB parameter group

  • You can't delete a default DB parameter group

  • Cannot be associated with any DB instances

", + "DeleteDBSecurityGroupMessage$DBSecurityGroupName":"

The name of the DB security group to delete.

You can't delete the default DB security group.

Constraints:

  • Must be 1 to 255 letters, numbers, or hyphens.

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

  • Must not be \"Default\"

", + "DeleteDBSnapshotMessage$DBSnapshotIdentifier":"

The DBSnapshot identifier.

Constraints: Must be the name of an existing DB snapshot in the available state.

", + "DeleteDBSubnetGroupMessage$DBSubnetGroupName":"

The name of the database subnet group to delete.

You can't delete the default subnet group.

Constraints:

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", + "DeleteEventSubscriptionMessage$SubscriptionName":"

The name of the RDS event notification subscription you want to delete.

", + "DeleteOptionGroupMessage$OptionGroupName":"

The name of the option group to be deleted.

You can't delete default option groups.

", + "DescribeCertificatesMessage$CertificateIdentifier":"

The user-supplied certificate identifier. If this parameter is specified, information for only the identified certificate is returned. This parameter isn't case-sensitive.

Constraints:

  • Must match an existing CertificateIdentifier.

", + "DescribeCertificatesMessage$Marker":"

An optional pagination token provided by a previous DescribeCertificates request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClusterBacktracksMessage$DBClusterIdentifier":"

The DB cluster identifier of the DB cluster to be described. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 alphanumeric characters or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", + "DescribeDBClusterBacktracksMessage$BacktrackIdentifier":"

If specified, this value is the backtrack identifier of the backtrack to be described.

Constraints:

Example: 123e4567-e89b-12d3-a456-426655440000

", + "DescribeDBClusterBacktracksMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterBacktracks request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClusterParameterGroupsMessage$DBClusterParameterGroupName":"

The name of a specific DB cluster parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "DescribeDBClusterParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClusterParametersMessage$DBClusterParameterGroupName":"

The name of a specific DB cluster parameter group to return parameter details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "DescribeDBClusterParametersMessage$Source":"

A value that indicates to return only parameters for a specific source. Parameter sources can be engine, service, or customer.

", + "DescribeDBClusterParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClusterSnapshotAttributesMessage$DBClusterSnapshotIdentifier":"

The identifier for the DB cluster snapshot to describe the attributes for.

", + "DescribeDBClusterSnapshotsMessage$DBClusterIdentifier":"

The ID of the DB cluster to retrieve the list of DB cluster snapshots for. This parameter can't be used in conjunction with the DBClusterSnapshotIdentifier parameter. This parameter is not case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBCluster.

", + "DescribeDBClusterSnapshotsMessage$DBClusterSnapshotIdentifier":"

A specific DB cluster snapshot identifier to describe. This parameter can't be used in conjunction with the DBClusterIdentifier parameter. This value is stored as a lowercase string.

Constraints:

  • If supplied, must match the identifier of an existing DBClusterSnapshot.

  • If this identifier is for an automated snapshot, the SnapshotType parameter must also be specified.

", + "DescribeDBClusterSnapshotsMessage$SnapshotType":"

The type of DB cluster snapshots to be returned. You can specify one of the following values:

  • automated - Return all DB cluster snapshots that have been automatically taken by Amazon RDS for my AWS account.

  • manual - Return all DB cluster snapshots that have been taken by my AWS account.

  • shared - Return all manual DB cluster snapshots that have been shared to my AWS account.

  • public - Return all DB cluster snapshots that have been marked as public.

If you don't specify a SnapshotType value, then both automated and manual DB cluster snapshots are returned. You can include shared DB cluster snapshots with these results by setting the IncludeShared parameter to true. You can include public DB cluster snapshots with these results by setting the IncludePublic parameter to true.

The IncludeShared and IncludePublic parameters don't apply for SnapshotType values of manual or automated. The IncludePublic parameter doesn't apply when SnapshotType is set to shared. The IncludeShared parameter doesn't apply when SnapshotType is set to public.

", + "DescribeDBClusterSnapshotsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusterSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBClustersMessage$DBClusterIdentifier":"

The user-supplied DB cluster identifier. If this parameter is specified, information from only the specific DB cluster is returned. This parameter isn't case-sensitive.

Constraints:

  • If supplied, must match an existing DBClusterIdentifier.

", + "DescribeDBClustersMessage$Marker":"

An optional pagination token provided by a previous DescribeDBClusters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBEngineVersionsMessage$Engine":"

The database engine to return.

", + "DescribeDBEngineVersionsMessage$EngineVersion":"

The database engine version to return.

Example: 5.1.49

", + "DescribeDBEngineVersionsMessage$DBParameterGroupFamily":"

The name of a specific DB parameter group family to return details for.

Constraints:

  • If supplied, must match an existing DBParameterGroupFamily.

", + "DescribeDBEngineVersionsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBInstancesMessage$DBInstanceIdentifier":"

The user-supplied instance identifier. If this parameter is specified, information from only the specific DB instance is returned. This parameter isn't case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBInstance.

", + "DescribeDBInstancesMessage$Marker":"

An optional pagination token provided by a previous DescribeDBInstances request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBLogFilesDetails$LogFileName":"

The name of the log file for the specified DB instance.

", + "DescribeDBLogFilesMessage$DBInstanceIdentifier":"

The customer-assigned name of the DB instance that contains the log files you want to list.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "DescribeDBLogFilesMessage$FilenameContains":"

Filters the available log files for log file names that contain the specified string.

", + "DescribeDBLogFilesMessage$Marker":"

The pagination token provided in the previous request. If this parameter is specified the response includes only records beyond the marker, up to MaxRecords.

", + "DescribeDBLogFilesResponse$Marker":"

A pagination token that can be used in a subsequent DescribeDBLogFiles request.

", + "DescribeDBParameterGroupsMessage$DBParameterGroupName":"

The name of a specific DB parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "DescribeDBParameterGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBParameterGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBParametersMessage$DBParameterGroupName":"

The name of a specific DB parameter group to return details for.

Constraints:

  • If supplied, must match the name of an existing DBParameterGroup.

", + "DescribeDBParametersMessage$Source":"

The parameter types to return.

Default: All parameter types returned

Valid Values: user | system | engine-default

", + "DescribeDBParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeDBParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBSecurityGroupsMessage$DBSecurityGroupName":"

The name of the DB security group to return details for.

", + "DescribeDBSecurityGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBSecurityGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBSnapshotAttributesMessage$DBSnapshotIdentifier":"

The identifier for the DB snapshot to describe the attributes for.

", + "DescribeDBSnapshotsMessage$DBInstanceIdentifier":"

The ID of the DB instance to retrieve the list of DB snapshots for. This parameter can't be used in conjunction with DBSnapshotIdentifier. This parameter is not case-sensitive.

Constraints:

  • If supplied, must match the identifier of an existing DBInstance.

", + "DescribeDBSnapshotsMessage$DBSnapshotIdentifier":"

A specific DB snapshot identifier to describe. This parameter can't be used in conjunction with DBInstanceIdentifier. This value is stored as a lowercase string.

Constraints:

  • If supplied, must match the identifier of an existing DBSnapshot.

  • If this identifier is for an automated snapshot, the SnapshotType parameter must also be specified.

", + "DescribeDBSnapshotsMessage$SnapshotType":"

The type of snapshots to be returned. You can specify one of the following values:

  • automated - Return all DB snapshots that have been automatically taken by Amazon RDS for my AWS account.

  • manual - Return all DB snapshots that have been taken by my AWS account.

  • shared - Return all manual DB snapshots that have been shared to my AWS account.

  • public - Return all DB snapshots that have been marked as public.

If you don't specify a SnapshotType value, then both automated and manual snapshots are returned. Shared and public DB snapshots are not included in the returned results by default. You can include shared snapshots with these results by setting the IncludeShared parameter to true. You can include public snapshots with these results by setting the IncludePublic parameter to true.

The IncludeShared and IncludePublic parameters don't apply for SnapshotType values of manual or automated. The IncludePublic parameter doesn't apply when SnapshotType is set to shared. The IncludeShared parameter doesn't apply when SnapshotType is set to public.

", + "DescribeDBSnapshotsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBSnapshots request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeDBSubnetGroupsMessage$DBSubnetGroupName":"

The name of the DB subnet group to return details for.

", + "DescribeDBSubnetGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeDBSubnetGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeEngineDefaultClusterParametersMessage$DBParameterGroupFamily":"

The name of the DB cluster parameter group family to return engine parameter information for.

", + "DescribeEngineDefaultClusterParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeEngineDefaultClusterParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeEngineDefaultParametersMessage$DBParameterGroupFamily":"

The name of the DB parameter group family.

", + "DescribeEngineDefaultParametersMessage$Marker":"

An optional pagination token provided by a previous DescribeEngineDefaultParameters request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeEventCategoriesMessage$SourceType":"

The type of source that is generating the events.

Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot

", + "DescribeEventSubscriptionsMessage$SubscriptionName":"

The name of the RDS event notification subscription you want to describe.

", + "DescribeEventSubscriptionsMessage$Marker":"

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DescribeEventsMessage$SourceIdentifier":"

The identifier of the event source for which events are returned. If not specified, then all sources are included in the response.

Constraints:

  • If SourceIdentifier is supplied, SourceType must also be provided.

  • If the source type is DBInstance, then a DBInstanceIdentifier must be supplied.

  • If the source type is DBSecurityGroup, a DBSecurityGroupName must be supplied.

  • If the source type is DBParameterGroup, a DBParameterGroupName must be supplied.

  • If the source type is DBSnapshot, a DBSnapshotIdentifier must be supplied.

  • Cannot end with a hyphen or contain two consecutive hyphens.

", + "DescribeEventsMessage$Marker":"

An optional pagination token provided by a previous DescribeEvents request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeOptionGroupOptionsMessage$EngineName":"

A required parameter. Options available for the given engine name are described.

", + "DescribeOptionGroupOptionsMessage$MajorEngineVersion":"

If specified, filters the results to include only options for the specified major engine version.

", + "DescribeOptionGroupOptionsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeOptionGroupsMessage$OptionGroupName":"

The name of the option group to describe. Cannot be supplied together with EngineName or MajorEngineVersion.

", + "DescribeOptionGroupsMessage$Marker":"

An optional pagination token provided by a previous DescribeOptionGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeOptionGroupsMessage$EngineName":"

Filters the list of option groups to only include groups associated with a specific database engine.

", + "DescribeOptionGroupsMessage$MajorEngineVersion":"

Filters the list of option groups to only include groups associated with a specific database engine version. If specified, then EngineName must also be specified.

", + "DescribeOrderableDBInstanceOptionsMessage$Engine":"

The name of the engine to retrieve DB instance options for.

", + "DescribeOrderableDBInstanceOptionsMessage$EngineVersion":"

The engine version filter value. Specify this parameter to show only the available offerings matching the specified engine version.

", + "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass":"

The DB instance class filter value. Specify this parameter to show only the available offerings matching the specified DB instance class.

", + "DescribeOrderableDBInstanceOptionsMessage$LicenseModel":"

The license model filter value. Specify this parameter to show only the available offerings matching the specified license model.

", + "DescribeOrderableDBInstanceOptionsMessage$Marker":"

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "DescribePendingMaintenanceActionsMessage$ResourceIdentifier":"

The ARN of a resource to return pending maintenance actions for.

", + "DescribePendingMaintenanceActionsMessage$Marker":"

An optional pagination token provided by a previous DescribePendingMaintenanceActions request. If this parameter is specified, the response includes only records beyond the marker, up to a number of records specified by MaxRecords.

", + "DescribeReservedDBInstancesMessage$ReservedDBInstanceId":"

The reserved DB instance identifier filter value. Specify this parameter to show only the reservation that matches the specified reservation ID.

", + "DescribeReservedDBInstancesMessage$ReservedDBInstancesOfferingId":"

The offering identifier filter value. Specify this parameter to show only purchased reservations matching the specified offering identifier.

", + "DescribeReservedDBInstancesMessage$DBInstanceClass":"

The DB instance class filter value. Specify this parameter to show only those reservations matching the specified DB instances class.

", + "DescribeReservedDBInstancesMessage$Duration":"

The duration filter value, specified in years or seconds. Specify this parameter to show only reservations for this duration.

Valid Values: 1 | 3 | 31536000 | 94608000

", + "DescribeReservedDBInstancesMessage$ProductDescription":"

The product description filter value. Specify this parameter to show only those reservations matching the specified product description.

", + "DescribeReservedDBInstancesMessage$OfferingType":"

The offering type filter value. Specify this parameter to show only the available offerings matching the specified offering type.

Valid Values: \"Partial Upfront\" | \"All Upfront\" | \"No Upfront\"

", + "DescribeReservedDBInstancesMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeReservedDBInstancesOfferingsMessage$ReservedDBInstancesOfferingId":"

The offering identifier filter value. Specify this parameter to show only the available offering that matches the specified reservation identifier.

Example: 438012d3-4052-4cc7-b2e3-8d3372e0e706

", + "DescribeReservedDBInstancesOfferingsMessage$DBInstanceClass":"

The DB instance class filter value. Specify this parameter to show only the available offerings matching the specified DB instance class.

", + "DescribeReservedDBInstancesOfferingsMessage$Duration":"

Duration filter value, specified in years or seconds. Specify this parameter to show only reservations for this duration.

Valid Values: 1 | 3 | 31536000 | 94608000

", + "DescribeReservedDBInstancesOfferingsMessage$ProductDescription":"

Product description filter value. Specify this parameter to show only the available offerings that contain the specified product description.

The results show offerings that partially match the filter value.

", + "DescribeReservedDBInstancesOfferingsMessage$OfferingType":"

The offering type filter value. Specify this parameter to show only the available offerings matching the specified offering type.

Valid Values: \"Partial Upfront\" | \"All Upfront\" | \"No Upfront\"

", + "DescribeReservedDBInstancesOfferingsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeSourceRegionsMessage$RegionName":"

The source AWS Region name. For example, us-east-1.

Constraints:

  • Must specify a valid AWS Region name.

", + "DescribeSourceRegionsMessage$Marker":"

An optional pagination token provided by a previous DescribeSourceRegions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "DescribeValidDBInstanceModificationsMessage$DBInstanceIdentifier":"

The customer identifier or the ARN of your DB instance.

", + "DomainMembership$Domain":"

The identifier of the Active Directory Domain.

", + "DomainMembership$Status":"

The status of the DB instance's Active Directory Domain membership, such as joined, pending-join, failed etc).

", + "DomainMembership$FQDN":"

The fully qualified domain name of the Active Directory Domain.

", + "DomainMembership$IAMRoleName":"

The name of the IAM role to be used when making API calls to the Directory Service.

", + "DownloadDBLogFilePortionDetails$LogFileData":"

Entries from the specified log file.

", + "DownloadDBLogFilePortionDetails$Marker":"

A pagination token that can be used in a subsequent DownloadDBLogFilePortion request.

", + "DownloadDBLogFilePortionMessage$DBInstanceIdentifier":"

The customer-assigned name of the DB instance that contains the log files you want to list.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "DownloadDBLogFilePortionMessage$LogFileName":"

The name of the log file to be downloaded.

", + "DownloadDBLogFilePortionMessage$Marker":"

The pagination token provided in the previous request or \"0\". If the Marker parameter is specified the response includes only records beyond the marker until the end of the file or up to NumberOfLines.

", + "EC2SecurityGroup$Status":"

Provides the status of the EC2 security group. Status can be \"authorizing\", \"authorized\", \"revoking\", and \"revoked\".

", + "EC2SecurityGroup$EC2SecurityGroupName":"

Specifies the name of the EC2 security group.

", + "EC2SecurityGroup$EC2SecurityGroupId":"

Specifies the id of the EC2 security group.

", + "EC2SecurityGroup$EC2SecurityGroupOwnerId":"

Specifies the AWS ID of the owner of the EC2 security group specified in the EC2SecurityGroupName field.

", + "Endpoint$Address":"

Specifies the DNS address of the DB instance.

", + "Endpoint$HostedZoneId":"

Specifies the ID that Amazon Route 53 assigns when you create a hosted zone.

", + "EngineDefaults$DBParameterGroupFamily":"

Specifies the name of the DB parameter group family that the engine default parameters apply to.

", + "EngineDefaults$Marker":"

An optional pagination token provided by a previous EngineDefaults request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "Event$SourceIdentifier":"

Provides the identifier for the source of the event.

", + "Event$Message":"

Provides the text of this event.

", + "Event$SourceArn":"

The Amazon Resource Name (ARN) for the event.

", + "EventCategoriesList$member":null, + "EventCategoriesMap$SourceType":"

The source type that the returned categories belong to

", + "EventSubscription$CustomerAwsId":"

The AWS customer account associated with the RDS event notification subscription.

", + "EventSubscription$CustSubscriptionId":"

The RDS event notification subscription Id.

", + "EventSubscription$SnsTopicArn":"

The topic ARN of the RDS event notification subscription.

", + "EventSubscription$Status":"

The status of the RDS event notification subscription.

Constraints:

Can be one of the following: creating | modifying | deleting | active | no-permission | topic-not-exist

The status \"no-permission\" indicates that RDS no longer has permission to post to the SNS topic. The status \"topic-not-exist\" indicates that the topic was deleted after the subscription was created.

", + "EventSubscription$SubscriptionCreationTime":"

The time the RDS event notification subscription was created.

", + "EventSubscription$SourceType":"

The source type for the RDS event notification subscription.

", + "EventSubscription$EventSubscriptionArn":"

The Amazon Resource Name (ARN) for the event subscription.

", + "EventSubscriptionsMessage$Marker":"

An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "EventsMessage$Marker":"

An optional pagination token provided by a previous Events request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "FailoverDBClusterMessage$DBClusterIdentifier":"

A DB cluster identifier to force a failover for. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

", + "FailoverDBClusterMessage$TargetDBInstanceIdentifier":"

The name of the instance to promote to the primary instance.

You must specify the instance identifier for an Aurora Replica in the DB cluster. For example, mydbcluster-replica1.

", + "Filter$Name":"

The name of the filter. Filter names are case-sensitive.

", + "FilterValueList$member":null, + "IPRange$Status":"

Specifies the status of the IP range. Status can be \"authorizing\", \"authorized\", \"revoking\", and \"revoked\".

", + "IPRange$CIDRIP":"

Specifies the IP range.

", + "KeyList$member":null, + "ListTagsForResourceMessage$ResourceName":"

The Amazon RDS resource with tags to be listed. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", + "LogTypeList$member":null, + "ModifyDBClusterMessage$DBClusterIdentifier":"

The DB cluster identifier for the cluster being modified. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster.

", + "ModifyDBClusterMessage$NewDBClusterIdentifier":"

The new DB cluster identifier for the DB cluster when renaming a DB cluster. This value is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • The first character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-cluster2

", + "ModifyDBClusterMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to use for the DB cluster.

", + "ModifyDBClusterMessage$MasterUserPassword":"

The new password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", + "ModifyDBClusterMessage$OptionGroupName":"

A value that indicates that the DB cluster should be associated with the specified option group. Changing this parameter doesn't result in an outage except in the following case, and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If the parameter change results in an option group that enables OEM, this change can cause a brief (sub-second) period during which new connections are rejected but existing connections are not interrupted.

Permanent options can't be removed from an option group. The option group can't be removed from a DB cluster once it is associated with a DB cluster.

", + "ModifyDBClusterMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "ModifyDBClusterMessage$PreferredMaintenanceWindow":"

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "ModifyDBClusterMessage$EngineVersion":"

The version number of the database engine to which you want to upgrade. Changing this parameter results in an outage. The change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true.

For a list of valid engine versions, see CreateDBCluster, or call DescribeDBEngineVersions.

", + "ModifyDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to modify.

", + "ModifyDBClusterSnapshotAttributeMessage$DBClusterSnapshotIdentifier":"

The identifier for the DB cluster snapshot to modify the attributes for.

", + "ModifyDBClusterSnapshotAttributeMessage$AttributeName":"

The name of the DB cluster snapshot attribute to modify.

To manage authorization for other AWS accounts to copy or restore a manual DB cluster snapshot, set this value to restore.

", + "ModifyDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier. This value is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "ModifyDBInstanceMessage$DBInstanceClass":"

The new compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

If you modify the DB instance class, an outage occurs during the change. The change is applied during the next maintenance window, unless ApplyImmediately is specified as true for this request.

Default: Uses existing setting

", + "ModifyDBInstanceMessage$DBSubnetGroupName":"

The new DB subnet group for the DB instance. You can use this parameter to move your DB instance to a different VPC. If your DB instance is not in a VPC, you can also use this parameter to move your DB instance into a VPC. For more information, see Updating the VPC for a DB Instance.

Changing the subnet group causes an outage during the change. The change is applied during the next maintenance window, unless you specify true for the ApplyImmediately parameter.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetGroup

", + "ModifyDBInstanceMessage$MasterUserPassword":"

The new password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible. Between the time of the request and the completion of the request, the MasterUserPassword element exists in the PendingModifiedValues element of the operation response.

Amazon Aurora

Not applicable. The password for the master user is managed by the DB cluster. For more information, see ModifyDBCluster.

Default: Uses existing setting

MariaDB

Constraints: Must contain from 8 to 41 characters.

Microsoft SQL Server

Constraints: Must contain from 8 to 128 characters.

MySQL

Constraints: Must contain from 8 to 41 characters.

Oracle

Constraints: Must contain from 8 to 30 characters.

PostgreSQL

Constraints: Must contain from 8 to 128 characters.

Amazon RDS API actions never return the password, so this action provides a way to regain access to a primary instance user if the password is lost. This includes restoring privileges that might have been accidentally revoked.

", + "ModifyDBInstanceMessage$DBParameterGroupName":"

The name of the DB parameter group to apply to the DB instance. Changing this setting doesn't result in an outage. The parameter group name itself is changed immediately, but the actual parameter changes are not applied until you reboot the instance without failover. The db instance will NOT be rebooted automatically and the parameter changes will NOT be applied during the next maintenance window.

Default: Uses existing setting

Constraints: The DB parameter group must be in the same DB parameter group family as this DB instance.

", + "ModifyDBInstanceMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled, as determined by the BackupRetentionPeriod parameter. Changing this parameter doesn't result in an outage and the change is asynchronously applied as soon as possible.

Amazon Aurora

Not applicable. The daily time range for creating automated backups is managed by the DB cluster. For more information, see ModifyDBCluster.

Constraints:

  • Must be in the format hh24:mi-hh24:mi

  • Must be in Universal Time Coordinated (UTC)

  • Must not conflict with the preferred maintenance window

  • Must be at least 30 minutes

", + "ModifyDBInstanceMessage$PreferredMaintenanceWindow":"

The weekly time range (in UTC) during which system maintenance can occur, which might result in an outage. Changing this parameter doesn't result in an outage, except in the following situation, and the change is asynchronously applied as soon as possible. If there are pending actions that cause a reboot, and the maintenance window is changed to include the current time, then changing this parameter will cause a reboot of the DB instance. If moving this window to the current time, there must be at least 30 minutes between the current time and end of the window to ensure pending changes are applied.

Default: Uses existing setting

Format: ddd:hh24:mi-ddd:hh24:mi

Valid Days: Mon | Tue | Wed | Thu | Fri | Sat | Sun

Constraints: Must be at least 30 minutes

", + "ModifyDBInstanceMessage$EngineVersion":"

The version number of the database engine to upgrade to. Changing this parameter results in an outage and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request.

For major version upgrades, if a nondefault DB parameter group is currently in use, a new DB parameter group in the DB parameter group family for the new engine version must be specified. The new DB parameter group can be the default for that DB parameter group family.

For information about valid engine versions, see CreateDBInstance, or call DescribeDBEngineVersions.

", + "ModifyDBInstanceMessage$LicenseModel":"

The license model for the DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", + "ModifyDBInstanceMessage$OptionGroupName":"

Indicates that the DB instance should be associated with the specified option group. Changing this parameter doesn't result in an outage except in the following case and the change is applied during the next maintenance window unless the ApplyImmediately parameter is set to true for this request. If the parameter change results in an option group that enables OEM, this change can cause a brief (sub-second) period during which new connections are rejected but existing connections are not interrupted.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "ModifyDBInstanceMessage$NewDBInstanceIdentifier":"

The new DB instance identifier for the DB instance when renaming a DB instance. When you change the DB instance identifier, an instance reboot will occur immediately if you set Apply Immediately to true, or will occur during the next maintenance window if Apply Immediately to false. This value is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • The first character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", + "ModifyDBInstanceMessage$StorageType":"

Specifies the storage type to be associated with the DB instance.

If you specify Provisioned IOPS (io1), you must also include a value for the Iops parameter.

If you choose to migrate your DB instance from using standard storage to using Provisioned IOPS, or from using Provisioned IOPS to using standard storage, the process can take time. The duration of the migration depends on several factors such as database load, storage size, storage type (standard or Provisioned IOPS), amount of IOPS provisioned (if any), and the number of prior scale storage operations. Typical migration times are under 24 hours, but the process can take up to several days in some cases. During the migration, the DB instance is available for use, but might experience performance degradation. While the migration takes place, nightly backups for the instance are suspended. No other Amazon RDS operations can take place for the instance, including modifying the instance, rebooting the instance, deleting the instance, creating a Read Replica for the instance, and creating a DB snapshot of the instance.

Valid values: standard | gp2 | io1

Default: io1 if the Iops parameter is specified, otherwise standard

", + "ModifyDBInstanceMessage$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "ModifyDBInstanceMessage$TdeCredentialPassword":"

The password for the given ARN from the key store in order to access the device.

", + "ModifyDBInstanceMessage$CACertificateIdentifier":"

Indicates the certificate that needs to be associated with the instance.

", + "ModifyDBInstanceMessage$Domain":"

The Active Directory Domain to move the instance to. Specify none to remove the instance from its current domain. The domain must be created prior to this operation. Currently only a Microsoft SQL Server instance can be created in a Active Directory Domain.

", + "ModifyDBInstanceMessage$MonitoringRoleArn":"

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, go to To create an IAM role for Amazon RDS Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", + "ModifyDBInstanceMessage$DomainIAMRoleName":"

The name of the IAM role to use when making API calls to the Directory Service.

", + "ModifyDBInstanceMessage$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "ModifyDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • If supplied, must match the name of an existing DBParameterGroup.

", + "ModifyDBSnapshotAttributeMessage$DBSnapshotIdentifier":"

The identifier for the DB snapshot to modify the attributes for.

", + "ModifyDBSnapshotAttributeMessage$AttributeName":"

The name of the DB snapshot attribute to modify.

To manage authorization for other AWS accounts to copy or restore a manual DB snapshot, set this value to restore.

", + "ModifyDBSnapshotMessage$DBSnapshotIdentifier":"

The identifier of the DB snapshot to modify.

", + "ModifyDBSnapshotMessage$EngineVersion":"

The engine version to upgrade the DB snapshot to.

The following are the database engines and engine versions that are available when you upgrade a DB snapshot.

MySQL

  • 5.5.46 (supported for 5.1 DB snapshots)

Oracle

  • 12.1.0.2.v8 (supported for 12.1.0.1 DB snapshots)

  • 11.2.0.4.v12 (supported for 11.2.0.2 DB snapshots)

  • 11.2.0.4.v11 (supported for 11.2.0.3 DB snapshots)

", + "ModifyDBSnapshotMessage$OptionGroupName":"

The option group to identify with the upgraded DB snapshot.

You can specify this parameter when you upgrade an Oracle DB snapshot. The same option group considerations apply when upgrading a DB snapshot as when upgrading a DB instance. For more information, see Option Group Considerations.

", + "ModifyDBSubnetGroupMessage$DBSubnetGroupName":"

The name for the DB subnet group. This value is stored as a lowercase string. You can't modify the default subnet group.

Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.

Example: mySubnetgroup

", + "ModifyDBSubnetGroupMessage$DBSubnetGroupDescription":"

The description for the DB subnet group.

", + "ModifyEventSubscriptionMessage$SubscriptionName":"

The name of the RDS event notification subscription.

", + "ModifyEventSubscriptionMessage$SnsTopicArn":"

The Amazon Resource Name (ARN) of the SNS topic created for event notification. The ARN is created by Amazon SNS when you create a topic and subscribe to it.

", + "ModifyEventSubscriptionMessage$SourceType":"

The type of source that is generating the events. For example, if you want to be notified of events generated by a DB instance, you would set this parameter to db-instance. if this value is not specified, all events are returned.

Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot

", + "ModifyOptionGroupMessage$OptionGroupName":"

The name of the option group to be modified.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "Option$OptionName":"

The name of the option.

", + "Option$OptionDescription":"

The description of the option.

", + "Option$OptionVersion":"

The version of the option.

", + "OptionConfiguration$OptionName":"

The configuration of options to include in a group.

", + "OptionConfiguration$OptionVersion":"

The version for the option.

", + "OptionGroup$OptionGroupName":"

Specifies the name of the option group.

", + "OptionGroup$OptionGroupDescription":"

Provides a description of the option group.

", + "OptionGroup$EngineName":"

Indicates the name of the engine that this option group can be applied to.

", + "OptionGroup$MajorEngineVersion":"

Indicates the major engine version associated with this option group.

", + "OptionGroup$VpcId":"

If AllowsVpcAndNonVpcInstanceMemberships is false, this field is blank. If AllowsVpcAndNonVpcInstanceMemberships is true and this field is blank, then this option group can be applied to both VPC and non-VPC instances. If this field contains a value, then this option group can only be applied to instances that are in the VPC indicated by this field.

", + "OptionGroup$OptionGroupArn":"

The Amazon Resource Name (ARN) for the option group.

", + "OptionGroupMembership$OptionGroupName":"

The name of the option group that the instance belongs to.

", + "OptionGroupMembership$Status":"

The status of the DB instance's option group membership. Valid values are: in-sync, pending-apply, pending-removal, pending-maintenance-apply, pending-maintenance-removal, applying, removing, and failed.

", + "OptionGroupOption$Name":"

The name of the option.

", + "OptionGroupOption$Description":"

The description of the option.

", + "OptionGroupOption$EngineName":"

The name of the engine that this option can be applied to.

", + "OptionGroupOption$MajorEngineVersion":"

Indicates the major engine version that the option is available for.

", + "OptionGroupOption$MinimumRequiredMinorEngineVersion":"

The minimum required engine version for the option to be applied.

", + "OptionGroupOptionSetting$SettingName":"

The name of the option group option.

", + "OptionGroupOptionSetting$SettingDescription":"

The description of the option group option.

", + "OptionGroupOptionSetting$DefaultValue":"

The default value for the option group option.

", + "OptionGroupOptionSetting$ApplyType":"

The DB engine specific parameter type for the option group option.

", + "OptionGroupOptionSetting$AllowedValues":"

Indicates the acceptable values for the option group option.

", + "OptionGroupOptionsMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "OptionGroups$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "OptionNamesList$member":null, + "OptionSetting$Name":"

The name of the option that has settings that you can set.

", + "OptionSetting$Value":"

The current value of the option setting.

", + "OptionSetting$DefaultValue":"

The default value of the option setting.

", + "OptionSetting$Description":"

The description of the option setting.

", + "OptionSetting$ApplyType":"

The DB engine specific parameter type.

", + "OptionSetting$DataType":"

The data type of the option setting.

", + "OptionSetting$AllowedValues":"

The allowed values of the option setting.

", + "OptionVersion$Version":"

The version of the option.

", + "OptionsConflictsWith$member":null, + "OptionsDependedOn$member":null, + "OrderableDBInstanceOption$Engine":"

The engine type of a DB instance.

", + "OrderableDBInstanceOption$EngineVersion":"

The engine version of a DB instance.

", + "OrderableDBInstanceOption$DBInstanceClass":"

The DB instance class for a DB instance.

", + "OrderableDBInstanceOption$LicenseModel":"

The license model for a DB instance.

", + "OrderableDBInstanceOption$StorageType":"

Indicates the storage type for a DB instance.

", + "OrderableDBInstanceOptionsMessage$Marker":"

An optional pagination token provided by a previous OrderableDBInstanceOptions request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords .

", + "Parameter$ParameterName":"

Specifies the name of the parameter.

", + "Parameter$ParameterValue":"

Specifies the value of the parameter.

", + "Parameter$Description":"

Provides a description of the parameter.

", + "Parameter$Source":"

Indicates the source of the parameter value.

", + "Parameter$ApplyType":"

Specifies the engine specific parameters type.

", + "Parameter$DataType":"

Specifies the valid data type for the parameter.

", + "Parameter$AllowedValues":"

Specifies the valid range of values for the parameter.

", + "Parameter$MinimumEngineVersion":"

The earliest engine version to which the parameter can apply.

", + "PendingMaintenanceAction$Action":"

The type of pending maintenance action that is available for the resource.

", + "PendingMaintenanceAction$OptInStatus":"

Indicates the type of opt-in request that has been received for the resource.

", + "PendingMaintenanceAction$Description":"

A description providing more detail about the maintenance action.

", + "PendingMaintenanceActionsMessage$Marker":"

An optional pagination token provided by a previous DescribePendingMaintenanceActions request. If this parameter is specified, the response includes only records beyond the marker, up to a number of records specified by MaxRecords.

", + "PendingModifiedValues$DBInstanceClass":"

Contains the new DBInstanceClass for the DB instance that will be applied or is currently being applied.

", + "PendingModifiedValues$MasterUserPassword":"

Contains the pending or currently-in-progress change of the master credentials for the DB instance.

", + "PendingModifiedValues$EngineVersion":"

Indicates the database engine version.

", + "PendingModifiedValues$LicenseModel":"

The license model for the DB instance.

Valid values: license-included | bring-your-own-license | general-public-license

", + "PendingModifiedValues$DBInstanceIdentifier":"

Contains the new DBInstanceIdentifier for the DB instance that will be applied or is currently being applied.

", + "PendingModifiedValues$StorageType":"

Specifies the storage type to be associated with the DB instance.

", + "PendingModifiedValues$CACertificateIdentifier":"

Specifies the identifier of the CA certificate for the DB instance.

", + "PendingModifiedValues$DBSubnetGroupName":"

The new DB subnet group for the DB instance.

", + "ProcessorFeature$Name":"

The name of the processor feature. Valid names are coreCount and threadsPerCore.

", + "ProcessorFeature$Value":"

The value of a processor feature name.

", + "PromoteReadReplicaDBClusterMessage$DBClusterIdentifier":"

The identifier of the DB cluster Read Replica to promote. This parameter is not case-sensitive.

Constraints:

  • Must match the identifier of an existing DBCluster Read Replica.

Example: my-cluster-replica1

", + "PromoteReadReplicaMessage$DBInstanceIdentifier":"

The DB instance identifier. This value is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing Read Replica DB instance.

Example: mydbinstance

", + "PromoteReadReplicaMessage$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstancesOfferingId":"

The ID of the Reserved DB instance offering to purchase.

Example: 438012d3-4052-4cc7-b2e3-8d3372e0e706

", + "PurchaseReservedDBInstancesOfferingMessage$ReservedDBInstanceId":"

Customer-specified identifier to track this reservation.

Example: myreservationID

", + "ReadReplicaDBClusterIdentifierList$member":null, + "ReadReplicaDBInstanceIdentifierList$member":null, + "ReadReplicaIdentifierList$member":null, + "RebootDBInstanceMessage$DBInstanceIdentifier":"

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must match the identifier of an existing DBInstance.

", + "RecurringCharge$RecurringChargeFrequency":"

The frequency of the recurring charge.

", + "RemoveRoleFromDBClusterMessage$DBClusterIdentifier":"

The name of the DB cluster to disassociate the IAM role from.

", + "RemoveRoleFromDBClusterMessage$RoleArn":"

The Amazon Resource Name (ARN) of the IAM role to disassociate from the Aurora DB cluster, for example arn:aws:iam::123456789012:role/AuroraAccessRole.

", + "RemoveSourceIdentifierFromSubscriptionMessage$SubscriptionName":"

The name of the RDS event notification subscription you want to remove a source identifier from.

", + "RemoveSourceIdentifierFromSubscriptionMessage$SourceIdentifier":"

The source identifier to be removed from the subscription, such as the DB instance identifier for a DB instance or the name of a security group.

", + "RemoveTagsFromResourceMessage$ResourceName":"

The Amazon RDS resource that the tags are removed from. This value is an Amazon Resource Name (ARN). For information about creating an ARN, see Constructing an RDS Amazon Resource Name (ARN).

", + "ReservedDBInstance$ReservedDBInstanceId":"

The unique identifier for the reservation.

", + "ReservedDBInstance$ReservedDBInstancesOfferingId":"

The offering identifier.

", + "ReservedDBInstance$DBInstanceClass":"

The DB instance class for the reserved DB instance.

", + "ReservedDBInstance$CurrencyCode":"

The currency code for the reserved DB instance.

", + "ReservedDBInstance$ProductDescription":"

The description of the reserved DB instance.

", + "ReservedDBInstance$OfferingType":"

The offering type of this reserved DB instance.

", + "ReservedDBInstance$State":"

The state of the reserved DB instance.

", + "ReservedDBInstance$ReservedDBInstanceArn":"

The Amazon Resource Name (ARN) for the reserved DB instance.

", + "ReservedDBInstanceMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "ReservedDBInstancesOffering$ReservedDBInstancesOfferingId":"

The offering identifier.

", + "ReservedDBInstancesOffering$DBInstanceClass":"

The DB instance class for the reserved DB instance.

", + "ReservedDBInstancesOffering$CurrencyCode":"

The currency code for the reserved DB instance offering.

", + "ReservedDBInstancesOffering$ProductDescription":"

The database engine used by the offering.

", + "ReservedDBInstancesOffering$OfferingType":"

The offering type.

", + "ReservedDBInstancesOfferingMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "ResetDBClusterParameterGroupMessage$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to reset.

", + "ResetDBParameterGroupMessage$DBParameterGroupName":"

The name of the DB parameter group.

Constraints:

  • Must match the name of an existing DBParameterGroup.

", + "ResourcePendingMaintenanceActions$ResourceIdentifier":"

The ARN of the resource that has pending maintenance actions.

", + "RestoreDBClusterFromS3Message$CharacterSetName":"

A value that indicates that the restored DB cluster should be associated with the specified CharacterSet.

", + "RestoreDBClusterFromS3Message$DatabaseName":"

The database name for the restored DB cluster.

", + "RestoreDBClusterFromS3Message$DBClusterIdentifier":"

The name of the DB cluster to create from the source data in the Amazon S3 bucket. This parameter is isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: my-cluster1

", + "RestoreDBClusterFromS3Message$DBClusterParameterGroupName":"

The name of the DB cluster parameter group to associate with the restored DB cluster. If this argument is omitted, default.aurora5.6 is used.

Constraints:

  • If supplied, must match the name of an existing DBClusterParameterGroup.

", + "RestoreDBClusterFromS3Message$DBSubnetGroupName":"

A DB subnet group to associate with the restored DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBClusterFromS3Message$Engine":"

The name of the database engine to be used for the restored DB cluster.

Valid Values: aurora, aurora-postgresql

", + "RestoreDBClusterFromS3Message$EngineVersion":"

The version number of the database engine to use.

Aurora MySQL

Example: 5.6.10a

Aurora PostgreSQL

Example: 9.6.3

", + "RestoreDBClusterFromS3Message$MasterUsername":"

The name of the master user for the restored DB cluster.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", + "RestoreDBClusterFromS3Message$MasterUserPassword":"

The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", + "RestoreDBClusterFromS3Message$OptionGroupName":"

A value that indicates that the restored DB cluster should be associated with the specified option group.

Permanent options can't be removed from an option group. An option group can't be removed from a DB cluster once it is associated with a DB cluster.

", + "RestoreDBClusterFromS3Message$PreferredBackupWindow":"

The daily time range during which automated backups are created if automated backups are enabled using the BackupRetentionPeriod parameter.

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "RestoreDBClusterFromS3Message$PreferredMaintenanceWindow":"

The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).

Format: ddd:hh24:mi-ddd:hh24:mi

The default is a 30-minute window selected at random from an 8-hour block of time for each AWS Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred Maintenance Window in the Amazon RDS User Guide.

Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

Constraints: Minimum 30-minute window.

", + "RestoreDBClusterFromS3Message$KmsKeyId":"

The AWS KMS key identifier for an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KM encryption key.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon RDS will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", + "RestoreDBClusterFromS3Message$SourceEngine":"

The identifier for the database engine that was backed up to create the files stored in the Amazon S3 bucket.

Valid values: mysql

", + "RestoreDBClusterFromS3Message$SourceEngineVersion":"

The version of the database that the backup files were created from.

MySQL version 5.5 and 5.6 are supported.

Example: 5.6.22

", + "RestoreDBClusterFromS3Message$S3BucketName":"

The name of the Amazon S3 bucket that contains the data used to create the Amazon Aurora DB cluster.

", + "RestoreDBClusterFromS3Message$S3Prefix":"

The prefix for all of the file names that contain the data used to create the Amazon Aurora DB cluster. If you do not specify a SourceS3Prefix value, then the Amazon Aurora DB cluster is created by using all of the files in the Amazon S3 bucket.

", + "RestoreDBClusterFromS3Message$S3IngestionRoleArn":"

The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that authorizes Amazon RDS to access the Amazon S3 bucket on your behalf.

", + "RestoreDBClusterFromSnapshotMessage$DBClusterIdentifier":"

The name of the DB cluster to create from the DB snapshot or DB cluster snapshot. This parameter isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", + "RestoreDBClusterFromSnapshotMessage$SnapshotIdentifier":"

The identifier for the DB snapshot or DB cluster snapshot to restore from.

You can use either the name or the Amazon Resource Name (ARN) to specify a DB cluster snapshot. However, you can use only the ARN to specify a DB snapshot.

Constraints:

  • Must match the identifier of an existing Snapshot.

", + "RestoreDBClusterFromSnapshotMessage$Engine":"

The database engine to use for the new DB cluster.

Default: The same as source

Constraint: Must be compatible with the engine of the source

", + "RestoreDBClusterFromSnapshotMessage$EngineVersion":"

The version of the database engine to use for the new DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$DBSubnetGroupName":"

The name of the DB subnet group to use for the new DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBClusterFromSnapshotMessage$DatabaseName":"

The database name for the restored DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$OptionGroupName":"

The name of the option group to use for the restored DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$KmsKeyId":"

The AWS KMS key identifier to use when restoring an encrypted DB cluster from a DB snapshot or DB cluster snapshot.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are restoring a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

If you do not specify a value for the KmsKeyId parameter, then the following will occur:

  • If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is encrypted, then the restored DB cluster is encrypted using the KMS key that was used to encrypt the DB snapshot or DB cluster snapshot.

  • If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is not encrypted, then the restored DB cluster is not encrypted.

", + "RestoreDBClusterToPointInTimeMessage$DBClusterIdentifier":"

The name of the new DB cluster to be created.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "RestoreDBClusterToPointInTimeMessage$RestoreType":"

The type of restore to be performed. You can specify one of the following values:

  • full-copy - The new DB cluster is restored as a full copy of the source DB cluster.

  • copy-on-write - The new DB cluster is restored as a clone of the source DB cluster.

Constraints: You can't specify copy-on-write if the engine version of the source DB cluster is earlier than 1.11.

If you don't specify a RestoreType value, then the new DB cluster is restored as a full copy of the source DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$SourceDBClusterIdentifier":"

The identifier of the source DB cluster from which to restore.

Constraints:

  • Must match the identifier of an existing DBCluster.

", + "RestoreDBClusterToPointInTimeMessage$DBSubnetGroupName":"

The DB subnet group name to use for the new DB cluster.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBClusterToPointInTimeMessage$OptionGroupName":"

The name of the option group for the new DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$KmsKeyId":"

The AWS KMS key identifier to use when restoring an encrypted DB cluster from an encrypted DB cluster.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are restoring a DB cluster with the same AWS account that owns the KMS encryption key used to encrypt the new DB cluster, then you can use the KMS key alias instead of the ARN for the KMS encryption key.

You can restore to a new DB cluster and encrypt the new DB cluster with a KMS key that is different than the KMS key used to encrypt the source DB cluster. The new DB cluster is encrypted with the KMS key identified by the KmsKeyId parameter.

If you do not specify a value for the KmsKeyId parameter, then the following will occur:

  • If the DB cluster is encrypted, then the restored DB cluster is encrypted using the KMS key that was used to encrypt the source DB cluster.

  • If the DB cluster is not encrypted, then the restored DB cluster is not encrypted.

If DBClusterIdentifier refers to a DB cluster that is not encrypted, then the restore request is rejected.

", + "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceIdentifier":"

Name of the DB instance to create from the DB snapshot. This parameter isn't case-sensitive.

Constraints:

  • Must contain from 1 to 63 numbers, letters, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

Example: my-snapshot-id

", + "RestoreDBInstanceFromDBSnapshotMessage$DBSnapshotIdentifier":"

The identifier for the DB snapshot to restore from.

Constraints:

  • Must match the identifier of an existing DBSnapshot.

  • If you are restoring from a shared manual DB snapshot, the DBSnapshotIdentifier must be the ARN of the shared DB snapshot.

", + "RestoreDBInstanceFromDBSnapshotMessage$DBInstanceClass":"

The compute and memory capacity of the Amazon RDS DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Default: The same DBInstanceClass as the original DB instance.

", + "RestoreDBInstanceFromDBSnapshotMessage$AvailabilityZone":"

The EC2 Availability Zone that the DB instance is created in.

Default: A random, system-chosen Availability Zone.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

Example: us-east-1a

", + "RestoreDBInstanceFromDBSnapshotMessage$DBSubnetGroupName":"

The DB subnet group name to use for the new instance.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBInstanceFromDBSnapshotMessage$LicenseModel":"

License model information for the restored DB instance.

Default: Same as source.

Valid values: license-included | bring-your-own-license | general-public-license

", + "RestoreDBInstanceFromDBSnapshotMessage$DBName":"

The database name for the restored DB instance.

This parameter doesn't apply to the MySQL, PostgreSQL, or MariaDB engines.

", + "RestoreDBInstanceFromDBSnapshotMessage$Engine":"

The database engine to use for the new instance.

Default: The same as source

Constraint: Must be compatible with the engine of the source. For example, you can restore a MariaDB 10.1 DB instance from a MySQL 5.6 snapshot.

Valid Values:

  • mariadb

  • mysql

  • oracle-ee

  • oracle-se2

  • oracle-se1

  • oracle-se

  • postgres

  • sqlserver-ee

  • sqlserver-se

  • sqlserver-ex

  • sqlserver-web

", + "RestoreDBInstanceFromDBSnapshotMessage$OptionGroupName":"

The name of the option group to be used for the restored DB instance.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "RestoreDBInstanceFromDBSnapshotMessage$StorageType":"

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", + "RestoreDBInstanceFromDBSnapshotMessage$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "RestoreDBInstanceFromDBSnapshotMessage$TdeCredentialPassword":"

The password for the given ARN from the key store in order to access the device.

", + "RestoreDBInstanceFromDBSnapshotMessage$Domain":"

Specify the Active Directory Domain to restore the instance in.

", + "RestoreDBInstanceFromDBSnapshotMessage$DomainIAMRoleName":"

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", + "RestoreDBInstanceFromS3Message$DBName":"

The name of the database to create when the DB instance is created. Follow the naming rules specified in CreateDBInstance.

", + "RestoreDBInstanceFromS3Message$DBInstanceIdentifier":"

The DB instance identifier. This parameter is stored as a lowercase string.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens.

  • First character must be a letter.

  • Cannot end with a hyphen or contain two consecutive hyphens.

Example: mydbinstance

", + "RestoreDBInstanceFromS3Message$DBInstanceClass":"

The compute and memory capacity of the DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Importing from Amazon S3 is not supported on the db.t2.micro DB instance class.

", + "RestoreDBInstanceFromS3Message$Engine":"

The name of the database engine to be used for this instance.

Valid Values: mysql

", + "RestoreDBInstanceFromS3Message$MasterUsername":"

The name for the master user.

Constraints:

  • Must be 1 to 16 letters or numbers.

  • First character must be a letter.

  • Cannot be a reserved word for the chosen database engine.

", + "RestoreDBInstanceFromS3Message$MasterUserPassword":"

The password for the master user. The password can include any printable ASCII character except \"/\", \"\"\", or \"@\".

Constraints: Must contain from 8 to 41 characters.

", + "RestoreDBInstanceFromS3Message$AvailabilityZone":"

The Availability Zone that the DB instance is created in. For information about AWS Regions and Availability Zones, see Regions and Availability Zones.

Default: A random, system-chosen Availability Zone in the endpoint's AWS Region.

Example: us-east-1d

Constraint: The AvailabilityZone parameter can't be specified if the MultiAZ parameter is set to true. The specified Availability Zone must be in the same AWS Region as the current endpoint.

", + "RestoreDBInstanceFromS3Message$DBSubnetGroupName":"

A DB subnet group to associate with this DB instance.

", + "RestoreDBInstanceFromS3Message$PreferredMaintenanceWindow":"

The time range each week during which system maintenance can occur, in Universal Coordinated Time (UTC). For more information, see Amazon RDS Maintenance Window.

Constraints:

  • Must be in the format ddd:hh24:mi-ddd:hh24:mi.

  • Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred backup window.

  • Must be at least 30 minutes.

", + "RestoreDBInstanceFromS3Message$DBParameterGroupName":"

The name of the DB parameter group to associate with this DB instance. If this argument is omitted, the default parameter group for the specified engine is used.

", + "RestoreDBInstanceFromS3Message$PreferredBackupWindow":"

The time range each day during which automated backups are created if automated backups are enabled. For more information, see The Backup Window.

Constraints:

  • Must be in the format hh24:mi-hh24:mi.

  • Must be in Universal Coordinated Time (UTC).

  • Must not conflict with the preferred maintenance window.

  • Must be at least 30 minutes.

", + "RestoreDBInstanceFromS3Message$EngineVersion":"

The version number of the database engine to use. Choose the latest minor version of your database engine. For information about engine versions, see CreateDBInstance, or call DescribeDBEngineVersions.

", + "RestoreDBInstanceFromS3Message$LicenseModel":"

The license model for this DB instance. Use general-public-license.

", + "RestoreDBInstanceFromS3Message$OptionGroupName":"

The name of the option group to associate with this DB instance. If this argument is omitted, the default option group for the specified engine is used.

", + "RestoreDBInstanceFromS3Message$StorageType":"

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified; otherwise standard

", + "RestoreDBInstanceFromS3Message$KmsKeyId":"

The AWS KMS key identifier for an encrypted DB instance.

The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption key. If you are creating a DB instance with the same AWS account that owns the KMS encryption key used to encrypt the new DB instance, then you can use the KMS key alias instead of the ARN for the KM encryption key.

If the StorageEncrypted parameter is true, and you do not specify a value for the KmsKeyId parameter, then Amazon RDS will use your default encryption key. AWS KMS creates the default encryption key for your AWS account. Your AWS account has a different default encryption key for each AWS Region.

", + "RestoreDBInstanceFromS3Message$MonitoringRoleArn":"

The ARN for the IAM role that permits RDS to send enhanced monitoring metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. For information on creating a monitoring role, see Setting Up and Enabling Enhanced Monitoring.

If MonitoringInterval is set to a value other than 0, then you must supply a MonitoringRoleArn value.

", + "RestoreDBInstanceFromS3Message$SourceEngine":"

The name of the engine of your source database.

Valid Values: mysql

", + "RestoreDBInstanceFromS3Message$SourceEngineVersion":"

The engine version of your source database.

Valid Values: 5.6

", + "RestoreDBInstanceFromS3Message$S3BucketName":"

The name of your Amazon S3 bucket that contains your database backup file.

", + "RestoreDBInstanceFromS3Message$S3Prefix":"

The prefix of your Amazon S3 bucket.

", + "RestoreDBInstanceFromS3Message$S3IngestionRoleArn":"

An AWS Identity and Access Management (IAM) role to allow Amazon RDS to access your Amazon S3 bucket.

", + "RestoreDBInstanceFromS3Message$PerformanceInsightsKMSKeyId":"

The AWS KMS key identifier for encryption of Performance Insights data. The KMS key ID is the Amazon Resource Name (ARN), the KMS key identifier, or the KMS key alias for the KMS encryption key.

", + "RestoreDBInstanceToPointInTimeMessage$SourceDBInstanceIdentifier":"

The identifier of the source DB instance from which to restore.

Constraints:

  • Must match the identifier of an existing DB instance.

", + "RestoreDBInstanceToPointInTimeMessage$TargetDBInstanceIdentifier":"

The name of the new DB instance to be created.

Constraints:

  • Must contain from 1 to 63 letters, numbers, or hyphens

  • First character must be a letter

  • Cannot end with a hyphen or contain two consecutive hyphens

", + "RestoreDBInstanceToPointInTimeMessage$DBInstanceClass":"

The compute and memory capacity of the Amazon RDS DB instance, for example, db.m4.large. Not all DB instance classes are available in all AWS Regions, or for all database engines. For the full list of DB instance classes, and availability for your engine, see DB Instance Class in the Amazon RDS User Guide.

Default: The same DBInstanceClass as the original DB instance.

", + "RestoreDBInstanceToPointInTimeMessage$AvailabilityZone":"

The EC2 Availability Zone that the DB instance is created in.

Default: A random, system-chosen Availability Zone.

Constraint: You can't specify the AvailabilityZone parameter if the MultiAZ parameter is set to true.

Example: us-east-1a

", + "RestoreDBInstanceToPointInTimeMessage$DBSubnetGroupName":"

The DB subnet group name to use for the new instance.

Constraints: If supplied, must match the name of an existing DBSubnetGroup.

Example: mySubnetgroup

", + "RestoreDBInstanceToPointInTimeMessage$LicenseModel":"

License model information for the restored DB instance.

Default: Same as source.

Valid values: license-included | bring-your-own-license | general-public-license

", + "RestoreDBInstanceToPointInTimeMessage$DBName":"

The database name for the restored DB instance.

This parameter is not used for the MySQL or MariaDB engines.

", + "RestoreDBInstanceToPointInTimeMessage$Engine":"

The database engine to use for the new instance.

Default: The same as source

Constraint: Must be compatible with the engine of the source

Valid Values:

  • mariadb

  • mysql

  • oracle-ee

  • oracle-se2

  • oracle-se1

  • oracle-se

  • postgres

  • sqlserver-ee

  • sqlserver-se

  • sqlserver-ex

  • sqlserver-web

", + "RestoreDBInstanceToPointInTimeMessage$OptionGroupName":"

The name of the option group to be used for the restored DB instance.

Permanent options, such as the TDE option for Oracle Advanced Security TDE, can't be removed from an option group, and that option group can't be removed from a DB instance once it is associated with a DB instance

", + "RestoreDBInstanceToPointInTimeMessage$StorageType":"

Specifies the storage type to be associated with the DB instance.

Valid values: standard | gp2 | io1

If you specify io1, you must also include a value for the Iops parameter.

Default: io1 if the Iops parameter is specified, otherwise standard

", + "RestoreDBInstanceToPointInTimeMessage$TdeCredentialArn":"

The ARN from the key store with which to associate the instance for TDE encryption.

", + "RestoreDBInstanceToPointInTimeMessage$TdeCredentialPassword":"

The password for the given ARN from the key store in order to access the device.

", + "RestoreDBInstanceToPointInTimeMessage$Domain":"

Specify the Active Directory Domain to restore the instance in.

", + "RestoreDBInstanceToPointInTimeMessage$DomainIAMRoleName":"

Specify the name of the IAM role to be used when making API calls to the Directory Service.

", + "RevokeDBSecurityGroupIngressMessage$DBSecurityGroupName":"

The name of the DB security group to revoke ingress from.

", + "RevokeDBSecurityGroupIngressMessage$CIDRIP":"

The IP range to revoke access from. Must be a valid CIDR range. If CIDRIP is specified, EC2SecurityGroupName, EC2SecurityGroupId and EC2SecurityGroupOwnerId can't be provided.

", + "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupName":"

The name of the EC2 security group to revoke access from. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", + "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupId":"

The id of the EC2 security group to revoke access from. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", + "RevokeDBSecurityGroupIngressMessage$EC2SecurityGroupOwnerId":"

The AWS Account Number of the owner of the EC2 security group specified in the EC2SecurityGroupName parameter. The AWS Access Key ID is not an acceptable value. For VPC DB security groups, EC2SecurityGroupId must be provided. Otherwise, EC2SecurityGroupOwnerId and either EC2SecurityGroupName or EC2SecurityGroupId must be provided.

", + "SourceIdsList$member":null, + "SourceRegion$RegionName":"

The name of the source AWS Region.

", + "SourceRegion$Endpoint":"

The endpoint for the source AWS Region endpoint.

", + "SourceRegion$Status":"

The status of the source AWS Region.

", + "SourceRegionMessage$Marker":"

An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords.

", + "StartDBInstanceMessage$DBInstanceIdentifier":"

The user-supplied instance identifier.

", + "StopDBInstanceMessage$DBInstanceIdentifier":"

The user-supplied instance identifier.

", + "StopDBInstanceMessage$DBSnapshotIdentifier":"

The user-supplied instance identifier of the DB Snapshot created immediately before the DB instance is stopped.

", + "Subnet$SubnetIdentifier":"

Specifies the identifier of the subnet.

", + "Subnet$SubnetStatus":"

Specifies the status of the subnet.

", + "SubnetIdentifierList$member":null, + "Tag$Key":"

A key is the required name of the tag. The string value can be from 1 to 128 Unicode characters in length and can't be prefixed with \"aws:\" or \"rds:\". The string can only contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-' (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").

", + "Tag$Value":"

A value is the optional value of the tag. The string value can be from 1 to 256 Unicode characters in length and can't be prefixed with \"aws:\" or \"rds:\". The string can only contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-' (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").

", + "Timezone$TimezoneName":"

The name of the time zone.

", + "UpgradeTarget$Engine":"

The name of the upgrade target database engine.

", + "UpgradeTarget$EngineVersion":"

The version number of the upgrade target database engine.

", + "UpgradeTarget$Description":"

The version of the database engine that a DB instance can be upgraded to.

", + "ValidStorageOptions$StorageType":"

The valid storage types for your DB instance. For example, gp2, io1.

", + "VpcSecurityGroupIdList$member":null, + "VpcSecurityGroupMembership$VpcSecurityGroupId":"

The name of the VPC security group.

", + "VpcSecurityGroupMembership$Status":"

The status of the VPC security group.

" + } + }, + "Subnet":{ + "base":"

This data type is used as a response element in the DescribeDBSubnetGroups action.

", + "refs":{ + "SubnetList$member":null + } + }, + "SubnetAlreadyInUse":{ + "base":"

The DB subnet is already in use in the Availability Zone.

", + "refs":{ + } + }, + "SubnetIdentifierList":{ + "base":null, + "refs":{ + "CreateDBSubnetGroupMessage$SubnetIds":"

The EC2 Subnet IDs for the DB subnet group.

", + "ModifyDBSubnetGroupMessage$SubnetIds":"

The EC2 subnet IDs for the DB subnet group.

" + } + }, + "SubnetList":{ + "base":null, + "refs":{ + "DBSubnetGroup$Subnets":"

Contains a list of Subnet elements.

" + } + }, + "SubscriptionAlreadyExistFault":{ + "base":"

The supplied subscription name already exists.

", + "refs":{ + } + }, + "SubscriptionCategoryNotFoundFault":{ + "base":"

The supplied category does not exist.

", + "refs":{ + } + }, + "SubscriptionNotFoundFault":{ + "base":"

The subscription name does not exist.

", + "refs":{ + } + }, + "SupportedCharacterSetsList":{ + "base":null, + "refs":{ + "DBEngineVersion$SupportedCharacterSets":"

A list of the character sets supported by this engine for the CharacterSetName parameter of the CreateDBInstance action.

" + } + }, + "SupportedTimezonesList":{ + "base":null, + "refs":{ + "DBEngineVersion$SupportedTimezones":"

A list of the time zones supported by this engine for the Timezone parameter of the CreateDBInstance action.

" + } + }, + "TStamp":{ + "base":null, + "refs":{ + "BacktrackDBClusterMessage$BacktrackTo":"

The timestamp of the time to backtrack the DB cluster to, specified in ISO 8601 format. For more information about ISO 8601, see the ISO8601 Wikipedia page.

If the specified time is not a consistent time for the DB cluster, Aurora automatically chooses the nearest possible consistent time for the DB cluster.

Constraints:

  • Must contain a valid ISO 8601 timestamp.

  • Cannot contain a timestamp set in the future.

Example: 2017-07-08T18:00Z

", + "Certificate$ValidFrom":"

The starting date from which the certificate is valid.

", + "Certificate$ValidTill":"

The final date that the certificate continues to be valid.

", + "DBCluster$EarliestRestorableTime":"

The earliest time to which a database can be restored with point-in-time restore.

", + "DBCluster$LatestRestorableTime":"

Specifies the latest time to which a database can be restored with point-in-time restore.

", + "DBCluster$ClusterCreateTime":"

Specifies the time when the DB cluster was created, in Universal Coordinated Time (UTC).

", + "DBCluster$EarliestBacktrackTime":"

The earliest time to which a DB cluster can be backtracked.

", + "DBClusterBacktrack$BacktrackTo":"

The timestamp of the time to which the DB cluster was backtracked.

", + "DBClusterBacktrack$BacktrackedFrom":"

The timestamp of the time from which the DB cluster was backtracked.

", + "DBClusterBacktrack$BacktrackRequestCreationTime":"

The timestamp of the time at which the backtrack was requested.

", + "DBClusterSnapshot$SnapshotCreateTime":"

Provides the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", + "DBClusterSnapshot$ClusterCreateTime":"

Specifies the time when the DB cluster was created, in Universal Coordinated Time (UTC).

", + "DBInstance$InstanceCreateTime":"

Provides the date and time the DB instance was created.

", + "DBInstance$LatestRestorableTime":"

Specifies the latest time to which a database can be restored with point-in-time restore.

", + "DBSnapshot$SnapshotCreateTime":"

Provides the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", + "DBSnapshot$InstanceCreateTime":"

Specifies the time when the snapshot was taken, in Universal Coordinated Time (UTC).

", + "DescribeEventsMessage$StartTime":"

The beginning of the time interval to retrieve events for, specified in ISO 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia page.

Example: 2009-07-08T18:00Z

", + "DescribeEventsMessage$EndTime":"

The end of the time interval for which to retrieve events, specified in ISO 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia page.

Example: 2009-07-08T18:00Z

", + "Event$Date":"

Specifies the date and time of the event.

", + "PendingMaintenanceAction$AutoAppliedAfterDate":"

The date of the maintenance window when the action is applied. The maintenance action is applied to the resource during its first maintenance window after this date. If this date is specified, any next-maintenance opt-in requests are ignored.

", + "PendingMaintenanceAction$ForcedApplyDate":"

The date when the maintenance action is automatically applied. The maintenance action is applied to the resource on this date regardless of the maintenance window for the resource. If this date is specified, any immediate opt-in requests are ignored.

", + "PendingMaintenanceAction$CurrentApplyDate":"

The effective date when the pending maintenance action is applied to the resource. This date takes into account opt-in requests received from the ApplyPendingMaintenanceAction API, the AutoAppliedAfterDate, and the ForcedApplyDate. This value is blank if an opt-in request has not been received and nothing has been specified as AutoAppliedAfterDate or ForcedApplyDate.

", + "ReservedDBInstance$StartTime":"

The time the reservation started.

", + "RestoreDBClusterToPointInTimeMessage$RestoreToTime":"

The date and time to restore the DB cluster to.

Valid Values: Value must be a time in Universal Coordinated Time (UTC) format

Constraints:

  • Must be before the latest restorable time for the DB instance

  • Must be specified if UseLatestRestorableTime parameter is not provided

  • Cannot be specified if UseLatestRestorableTime parameter is true

  • Cannot be specified if RestoreType parameter is copy-on-write

Example: 2015-03-07T23:45:00Z

", + "RestoreDBInstanceToPointInTimeMessage$RestoreTime":"

The date and time to restore from.

Valid Values: Value must be a time in Universal Coordinated Time (UTC) format

Constraints:

  • Must be before the latest restorable time for the DB instance

  • Cannot be specified if UseLatestRestorableTime parameter is true

Example: 2009-09-07T23:45:00Z

" + } + }, + "Tag":{ + "base":"

Metadata assigned to an Amazon RDS resource consisting of a key-value pair.

", + "refs":{ + "TagList$member":null + } + }, + "TagList":{ + "base":"

A list of tags. For more information, see Tagging Amazon RDS Resources.

", + "refs":{ + "AddTagsToResourceMessage$Tags":"

The tags to be assigned to the Amazon RDS resource.

", + "CopyDBClusterParameterGroupMessage$Tags":null, + "CopyDBClusterSnapshotMessage$Tags":null, + "CopyDBParameterGroupMessage$Tags":null, + "CopyDBSnapshotMessage$Tags":null, + "CopyOptionGroupMessage$Tags":null, + "CreateDBClusterMessage$Tags":null, + "CreateDBClusterParameterGroupMessage$Tags":null, + "CreateDBClusterSnapshotMessage$Tags":"

The tags to be assigned to the DB cluster snapshot.

", + "CreateDBInstanceMessage$Tags":null, + "CreateDBInstanceReadReplicaMessage$Tags":null, + "CreateDBParameterGroupMessage$Tags":null, + "CreateDBSecurityGroupMessage$Tags":null, + "CreateDBSnapshotMessage$Tags":null, + "CreateDBSubnetGroupMessage$Tags":null, + "CreateEventSubscriptionMessage$Tags":null, + "CreateOptionGroupMessage$Tags":null, + "PurchaseReservedDBInstancesOfferingMessage$Tags":null, + "RestoreDBClusterFromS3Message$Tags":null, + "RestoreDBClusterFromSnapshotMessage$Tags":"

The tags to be assigned to the restored DB cluster.

", + "RestoreDBClusterToPointInTimeMessage$Tags":null, + "RestoreDBInstanceFromDBSnapshotMessage$Tags":null, + "RestoreDBInstanceFromS3Message$Tags":"

A list of tags to associate with this DB instance. For more information, see Tagging Amazon RDS Resources.

", + "RestoreDBInstanceToPointInTimeMessage$Tags":null, + "TagListMessage$TagList":"

List of tags returned by the ListTagsForResource operation.

" + } + }, + "TagListMessage":{ + "base":"

", + "refs":{ + } + }, + "Timezone":{ + "base":"

A time zone associated with a DBInstance or a DBSnapshot. This data type is an element in the response to the DescribeDBInstances, the DescribeDBSnapshots, and the DescribeDBEngineVersions actions.

", + "refs":{ + "SupportedTimezonesList$member":null + } + }, + "UpgradeTarget":{ + "base":"

The version of the database engine that a DB instance can be upgraded to.

", + "refs":{ + "ValidUpgradeTargetList$member":null + } + }, + "ValidDBInstanceModificationsMessage":{ + "base":"

Information about valid modifications that you can make to your DB instance. Contains the result of a successful call to the DescribeValidDBInstanceModifications action. You can use this information when you call ModifyDBInstance.

", + "refs":{ + "DescribeValidDBInstanceModificationsResult$ValidDBInstanceModificationsMessage":null + } + }, + "ValidStorageOptions":{ + "base":"

Information about valid modifications that you can make to your DB instance. Contains the result of a successful call to the DescribeValidDBInstanceModifications action.

", + "refs":{ + "ValidStorageOptionsList$member":null + } + }, + "ValidStorageOptionsList":{ + "base":null, + "refs":{ + "ValidDBInstanceModificationsMessage$Storage":"

Valid storage options for your DB instance.

" + } + }, + "ValidUpgradeTargetList":{ + "base":null, + "refs":{ + "DBEngineVersion$ValidUpgradeTarget":"

A list of engine versions that this database engine version can be upgraded to.

" + } + }, + "VpcSecurityGroupIdList":{ + "base":null, + "refs":{ + "CreateDBClusterMessage$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to associate with this DB cluster.

", + "CreateDBInstanceMessage$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to associate with this DB instance.

Amazon Aurora

Not applicable. The associated list of EC2 VPC security groups is managed by the DB cluster. For more information, see CreateDBCluster.

Default: The default EC2 VPC security group for the DB subnet group's VPC.

", + "ModifyDBClusterMessage$VpcSecurityGroupIds":"

A list of VPC security groups that the DB cluster will belong to.

", + "ModifyDBInstanceMessage$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to authorize on this DB instance. This change is asynchronously applied as soon as possible.

Amazon Aurora

Not applicable. The associated list of EC2 VPC security groups is managed by the DB cluster. For more information, see ModifyDBCluster.

Constraints:

  • If supplied, must match existing VpcSecurityGroupIds.

", + "OptionConfiguration$VpcSecurityGroupMemberships":"

A list of VpcSecurityGroupMemebrship name strings used for this option.

", + "RestoreDBClusterFromS3Message$VpcSecurityGroupIds":"

A list of EC2 VPC security groups to associate with the restored DB cluster.

", + "RestoreDBClusterFromSnapshotMessage$VpcSecurityGroupIds":"

A list of VPC security groups that the new DB cluster will belong to.

", + "RestoreDBClusterToPointInTimeMessage$VpcSecurityGroupIds":"

A list of VPC security groups that the new DB cluster belongs to.

", + "RestoreDBInstanceFromS3Message$VpcSecurityGroupIds":"

A list of VPC security groups to associate with this DB instance.

" + } + }, + "VpcSecurityGroupMembership":{ + "base":"

This data type is used as a response element for queries on VPC security group membership.

", + "refs":{ + "VpcSecurityGroupMembershipList$member":null + } + }, + "VpcSecurityGroupMembershipList":{ + "base":null, + "refs":{ + "DBCluster$VpcSecurityGroups":"

Provides a list of VPC security groups that the DB cluster belongs to.

", + "DBInstance$VpcSecurityGroups":"

Provides a list of VPC security group elements that the DB instance belongs to.

", + "Option$VpcSecurityGroupMemberships":"

If the option requires access to a port, then this VPC security group allows access to the port.

" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/examples-1.json index e72a328e8..b365c209c 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-10-31/examples-1.json @@ -1,1950 +1,1950 @@ { - "version": "1.0", - "examples": { - "AddSourceIdentifierToSubscription": [ + "version":"1.0", + "examples":{ + "AddSourceIdentifierToSubscription":[ { - "input": { - "SourceIdentifier": "mymysqlinstance", - "SubscriptionName": "mymysqleventsubscription" + "input":{ + "SourceIdentifier":"mymysqlinstance", + "SubscriptionName":"mymysqleventsubscription" }, - "output": { - "EventSubscription": { + "output":{ + "EventSubscription":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example add a source identifier to an event notification subscription.", - "id": "add-source-identifier-to-subscription-93fb6a15-0a59-4577-a7b5-e12db9752c14", - "title": "To add a source identifier to an event notification subscription" + "description":"This example add a source identifier to an event notification subscription.", + "id":"add-source-identifier-to-subscription-93fb6a15-0a59-4577-a7b5-e12db9752c14", + "title":"To add a source identifier to an event notification subscription" } ], - "AddTagsToResource": [ + "AddTagsToResource":[ { - "input": { - "ResourceName": "arn:aws:rds:us-east-1:992648334831:og:mymysqloptiongroup", - "Tags": [ + "input":{ + "ResourceName":"arn:aws:rds:us-east-1:992648334831:og:mymysqloptiongroup", + "Tags":[ { - "Key": "Staging", - "Value": "LocationDB" + "Key":"Staging", + "Value":"LocationDB" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example adds a tag to an option group.", - "id": "add-tags-to-resource-fa99ef50-228b-449d-b893-ca4d4e9768ab", - "title": "To add tags to a resource" + "description":"This example adds a tag to an option group.", + "id":"add-tags-to-resource-fa99ef50-228b-449d-b893-ca4d4e9768ab", + "title":"To add tags to a resource" } ], - "ApplyPendingMaintenanceAction": [ + "ApplyPendingMaintenanceAction":[ { - "input": { - "ApplyAction": "system-update", - "OptInType": "immediate", - "ResourceIdentifier": "arn:aws:rds:us-east-1:992648334831:db:mymysqlinstance" + "input":{ + "ApplyAction":"system-update", + "OptInType":"immediate", + "ResourceIdentifier":"arn:aws:rds:us-east-1:992648334831:db:mymysqlinstance" }, - "output": { - "ResourcePendingMaintenanceActions": { + "output":{ + "ResourcePendingMaintenanceActions":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example immediately applies a pending system update to a DB instance.", - "id": "apply-pending-maintenance-action-2a026047-8bbb-47fc-b695-abad9f308c24", - "title": "To apply a pending maintenance action" + "description":"This example immediately applies a pending system update to a DB instance.", + "id":"apply-pending-maintenance-action-2a026047-8bbb-47fc-b695-abad9f308c24", + "title":"To apply a pending maintenance action" } ], - "AuthorizeDBSecurityGroupIngress": [ + "AuthorizeDBSecurityGroupIngress":[ { - "input": { - "CIDRIP": "203.0.113.5/32", - "DBSecurityGroupName": "mydbsecuritygroup" + "input":{ + "CIDRIP":"203.0.113.5/32", + "DBSecurityGroupName":"mydbsecuritygroup" }, - "output": { - "DBSecurityGroup": { + "output":{ + "DBSecurityGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example authorizes access to the specified security group by the specified CIDR block.", - "id": "authorize-db-security-group-ingress-ebf9ab91-8912-4b07-a32e-ca150668164f", - "title": "To authorize DB security group integress" + "description":"This example authorizes access to the specified security group by the specified CIDR block.", + "id":"authorize-db-security-group-ingress-ebf9ab91-8912-4b07-a32e-ca150668164f", + "title":"To authorize DB security group integress" } ], - "CopyDBClusterParameterGroup": [ + "CopyDBClusterParameterGroup":[ { - "input": { - "SourceDBClusterParameterGroupIdentifier": "mydbclusterparametergroup", - "TargetDBClusterParameterGroupDescription": "My DB cluster parameter group copy", - "TargetDBClusterParameterGroupIdentifier": "mydbclusterparametergroup-copy" + "input":{ + "SourceDBClusterParameterGroupIdentifier":"mydbclusterparametergroup", + "TargetDBClusterParameterGroupDescription":"My DB cluster parameter group copy", + "TargetDBClusterParameterGroupIdentifier":"mydbclusterparametergroup-copy" }, - "output": { - "DBClusterParameterGroup": { + "output":{ + "DBClusterParameterGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example copies a DB cluster parameter group.", - "id": "copy-db-cluster-parameter-group-6fefaffe-cde9-4dba-9f0b-d3f593572fe4", - "title": "To copy a DB cluster parameter group" + "description":"This example copies a DB cluster parameter group.", + "id":"copy-db-cluster-parameter-group-6fefaffe-cde9-4dba-9f0b-d3f593572fe4", + "title":"To copy a DB cluster parameter group" } ], - "CopyDBClusterSnapshot": [ + "CopyDBClusterSnapshot":[ { - "input": { - "SourceDBClusterSnapshotIdentifier": "rds:sample-cluster-2016-09-14-10-38", - "TargetDBClusterSnapshotIdentifier": "cluster-snapshot-copy-1" + "input":{ + "SourceDBClusterSnapshotIdentifier":"rds:sample-cluster-2016-09-14-10-38", + "TargetDBClusterSnapshotIdentifier":"cluster-snapshot-copy-1" }, - "output": { - "DBClusterSnapshot": { + "output":{ + "DBClusterSnapshot":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example copies an automated snapshot of a DB cluster to a new DB cluster snapshot.", - "id": "to-copy-a-db-cluster-snapshot-1473879770564", - "title": "To copy a DB cluster snapshot" + "description":"The following example copies an automated snapshot of a DB cluster to a new DB cluster snapshot.", + "id":"to-copy-a-db-cluster-snapshot-1473879770564", + "title":"To copy a DB cluster snapshot" } ], - "CopyDBParameterGroup": [ + "CopyDBParameterGroup":[ { - "input": { - "SourceDBParameterGroupIdentifier": "mymysqlparametergroup", - "TargetDBParameterGroupDescription": "My MySQL parameter group copy", - "TargetDBParameterGroupIdentifier": "mymysqlparametergroup-copy" + "input":{ + "SourceDBParameterGroupIdentifier":"mymysqlparametergroup", + "TargetDBParameterGroupDescription":"My MySQL parameter group copy", + "TargetDBParameterGroupIdentifier":"mymysqlparametergroup-copy" }, - "output": { - "DBParameterGroup": { + "output":{ + "DBParameterGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example copies a DB parameter group.", - "id": "copy-db-parameter-group-610d4dba-2c87-467f-ae5d-edd7f8e47349", - "title": "To copy a DB parameter group" + "description":"This example copies a DB parameter group.", + "id":"copy-db-parameter-group-610d4dba-2c87-467f-ae5d-edd7f8e47349", + "title":"To copy a DB parameter group" } ], - "CopyDBSnapshot": [ + "CopyDBSnapshot":[ { - "input": { - "SourceDBSnapshotIdentifier": "mydbsnapshot", - "TargetDBSnapshotIdentifier": "mydbsnapshot-copy" + "input":{ + "SourceDBSnapshotIdentifier":"mydbsnapshot", + "TargetDBSnapshotIdentifier":"mydbsnapshot-copy" }, - "output": { - "DBSnapshot": { + "output":{ + "DBSnapshot":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example copies a DB snapshot.", - "id": "copy-db-snapshot-1b2f0210-bc67-415d-9822-6eecf447dc86", - "title": "To copy a DB snapshot" + "description":"This example copies a DB snapshot.", + "id":"copy-db-snapshot-1b2f0210-bc67-415d-9822-6eecf447dc86", + "title":"To copy a DB snapshot" } ], - "CopyOptionGroup": [ + "CopyOptionGroup":[ { - "input": { - "SourceOptionGroupIdentifier": "mymysqloptiongroup", - "TargetOptionGroupDescription": "My MySQL option group copy", - "TargetOptionGroupIdentifier": "mymysqloptiongroup-copy" + "input":{ + "SourceOptionGroupIdentifier":"mymysqloptiongroup", + "TargetOptionGroupDescription":"My MySQL option group copy", + "TargetOptionGroupIdentifier":"mymysqloptiongroup-copy" }, - "output": { - "OptionGroup": { + "output":{ + "OptionGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example copies an option group.", - "id": "copy-option-group-8d5c01c3-8846-4e9c-a4b0-1b7237f7d0ec", - "title": "To copy an option group" + "description":"This example copies an option group.", + "id":"copy-option-group-8d5c01c3-8846-4e9c-a4b0-1b7237f7d0ec", + "title":"To copy an option group" } ], - "CreateDBCluster": [ + "CreateDBCluster":[ { - "input": { - "AvailabilityZones": [ + "input":{ + "AvailabilityZones":[ "us-east-1a" ], - "BackupRetentionPeriod": 1, - "DBClusterIdentifier": "mydbcluster", - "DBClusterParameterGroupName": "mydbclusterparametergroup", - "DatabaseName": "myauroradb", - "Engine": "aurora", - "EngineVersion": "5.6.10a", - "MasterUserPassword": "mypassword", - "MasterUsername": "myuser", - "Port": 3306, - "StorageEncrypted": true + "BackupRetentionPeriod":1, + "DBClusterIdentifier":"mydbcluster", + "DBClusterParameterGroupName":"mydbclusterparametergroup", + "DatabaseName":"myauroradb", + "Engine":"aurora", + "EngineVersion":"5.6.10a", + "MasterUserPassword":"mypassword", + "MasterUsername":"myuser", + "Port":3306, + "StorageEncrypted":true }, - "output": { - "DBCluster": { + "output":{ + "DBCluster":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB cluster.", - "id": "create-db-cluster-423b998d-eba9-40dd-8e19-96c5b6e5f31d", - "title": "To create a DB cluster" + "description":"This example creates a DB cluster.", + "id":"create-db-cluster-423b998d-eba9-40dd-8e19-96c5b6e5f31d", + "title":"To create a DB cluster" } ], - "CreateDBClusterParameterGroup": [ + "CreateDBClusterParameterGroup":[ { - "input": { - "DBClusterParameterGroupName": "mydbclusterparametergroup", - "DBParameterGroupFamily": "aurora5.6", - "Description": "My DB cluster parameter group" + "input":{ + "DBClusterParameterGroupName":"mydbclusterparametergroup", + "DBParameterGroupFamily":"aurora5.6", + "Description":"My DB cluster parameter group" }, - "output": { - "DBClusterParameterGroup": { + "output":{ + "DBClusterParameterGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB cluster parameter group.", - "id": "create-db-cluster-parameter-group-8eb1c3ae-1965-4262-afe3-ee134c4430b1", - "title": "To create a DB cluster parameter group" + "description":"This example creates a DB cluster parameter group.", + "id":"create-db-cluster-parameter-group-8eb1c3ae-1965-4262-afe3-ee134c4430b1", + "title":"To create a DB cluster parameter group" } ], - "CreateDBClusterSnapshot": [ + "CreateDBClusterSnapshot":[ { - "input": { - "DBClusterIdentifier": "mydbcluster", - "DBClusterSnapshotIdentifier": "mydbclustersnapshot" + "input":{ + "DBClusterIdentifier":"mydbcluster", + "DBClusterSnapshotIdentifier":"mydbclustersnapshot" }, - "output": { - "DBClusterSnapshot": { + "output":{ + "DBClusterSnapshot":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB cluster snapshot.", - "id": "create-db-cluster-snapshot-", - "title": "To create a DB cluster snapshot" + "description":"This example creates a DB cluster snapshot.", + "id":"create-db-cluster-snapshot-", + "title":"To create a DB cluster snapshot" } ], - "CreateDBInstance": [ + "CreateDBInstance":[ { - "input": { - "AllocatedStorage": 5, - "DBInstanceClass": "db.t2.micro", - "DBInstanceIdentifier": "mymysqlinstance", - "Engine": "MySQL", - "MasterUserPassword": "MyPassword", - "MasterUsername": "MyUser" + "input":{ + "AllocatedStorage":5, + "DBInstanceClass":"db.t2.micro", + "DBInstanceIdentifier":"mymysqlinstance", + "Engine":"MySQL", + "MasterUserPassword":"MyPassword", + "MasterUsername":"MyUser" }, - "output": { - "DBInstance": { + "output":{ + "DBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB instance.", - "id": "create-db-instance-57eb5d16-8bf8-4c84-9709-1700322b37b9", - "title": "To create a DB instance." + "description":"This example creates a DB instance.", + "id":"create-db-instance-57eb5d16-8bf8-4c84-9709-1700322b37b9", + "title":"To create a DB instance." } ], - "CreateDBInstanceReadReplica": [ + "CreateDBInstanceReadReplica":[ { - "input": { - "AvailabilityZone": "us-east-1a", - "CopyTagsToSnapshot": true, - "DBInstanceClass": "db.t2.micro", - "DBInstanceIdentifier": "mydbreadreplica", - "PubliclyAccessible": true, - "SourceDBInstanceIdentifier": "mymysqlinstance", - "StorageType": "gp2", - "Tags": [ + "input":{ + "AvailabilityZone":"us-east-1a", + "CopyTagsToSnapshot":true, + "DBInstanceClass":"db.t2.micro", + "DBInstanceIdentifier":"mydbreadreplica", + "PubliclyAccessible":true, + "SourceDBInstanceIdentifier":"mymysqlinstance", + "StorageType":"gp2", + "Tags":[ { - "Key": "mydbreadreplicakey", - "Value": "mydbreadreplicavalue" + "Key":"mydbreadreplicakey", + "Value":"mydbreadreplicavalue" } ] }, - "output": { - "DBInstance": { + "output":{ + "DBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB instance read replica.", - "id": "create-db-instance-read-replica-81b41cd5-2871-4dae-bc59-3e264449d5fe", - "title": "To create a DB instance read replica." + "description":"This example creates a DB instance read replica.", + "id":"create-db-instance-read-replica-81b41cd5-2871-4dae-bc59-3e264449d5fe", + "title":"To create a DB instance read replica." } ], - "CreateDBParameterGroup": [ + "CreateDBParameterGroup":[ { - "input": { - "DBParameterGroupFamily": "mysql5.6", - "DBParameterGroupName": "mymysqlparametergroup", - "Description": "My MySQL parameter group" + "input":{ + "DBParameterGroupFamily":"mysql5.6", + "DBParameterGroupName":"mymysqlparametergroup", + "Description":"My MySQL parameter group" }, - "output": { - "DBParameterGroup": { + "output":{ + "DBParameterGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB parameter group.", - "id": "create-db-parameter-group-42afcc37-12e9-4b6a-a55c-b8a141246e87", - "title": "To create a DB parameter group." + "description":"This example creates a DB parameter group.", + "id":"create-db-parameter-group-42afcc37-12e9-4b6a-a55c-b8a141246e87", + "title":"To create a DB parameter group." } ], - "CreateDBSecurityGroup": [ + "CreateDBSecurityGroup":[ { - "input": { - "DBSecurityGroupDescription": "My DB security group", - "DBSecurityGroupName": "mydbsecuritygroup" + "input":{ + "DBSecurityGroupDescription":"My DB security group", + "DBSecurityGroupName":"mydbsecuritygroup" }, - "output": { - "DBSecurityGroup": { + "output":{ + "DBSecurityGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB security group.", - "id": "create-db-security-group-41b6786a-539e-42a5-a645-a8bc3cf99353", - "title": "To create a DB security group." + "description":"This example creates a DB security group.", + "id":"create-db-security-group-41b6786a-539e-42a5-a645-a8bc3cf99353", + "title":"To create a DB security group." } ], - "CreateDBSnapshot": [ + "CreateDBSnapshot":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance", - "DBSnapshotIdentifier": "mydbsnapshot" + "input":{ + "DBInstanceIdentifier":"mymysqlinstance", + "DBSnapshotIdentifier":"mydbsnapshot" }, - "output": { - "DBSnapshot": { + "output":{ + "DBSnapshot":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB snapshot.", - "id": "create-db-snapshot-e10e0e2c-9ac4-426d-9b17-6b6a3e382ce2", - "title": "To create a DB snapshot." + "description":"This example creates a DB snapshot.", + "id":"create-db-snapshot-e10e0e2c-9ac4-426d-9b17-6b6a3e382ce2", + "title":"To create a DB snapshot." } ], - "CreateDBSubnetGroup": [ + "CreateDBSubnetGroup":[ { - "input": { - "DBSubnetGroupDescription": "My DB subnet group", - "DBSubnetGroupName": "mydbsubnetgroup", - "SubnetIds": [ + "input":{ + "DBSubnetGroupDescription":"My DB subnet group", + "DBSubnetGroupName":"mydbsubnetgroup", + "SubnetIds":[ "subnet-1fab8a69", "subnet-d43a468c" ] }, - "output": { - "DBSubnetGroup": { + "output":{ + "DBSubnetGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates a DB subnet group.", - "id": "create-db-subnet-group-c3d162c2-0ec4-4955-ba89-18967615fdb8", - "title": "To create a DB subnet group." + "description":"This example creates a DB subnet group.", + "id":"create-db-subnet-group-c3d162c2-0ec4-4955-ba89-18967615fdb8", + "title":"To create a DB subnet group." } ], - "CreateEventSubscription": [ + "CreateEventSubscription":[ { - "input": { - "Enabled": true, - "EventCategories": [ + "input":{ + "Enabled":true, + "EventCategories":[ "availability" ], - "SnsTopicArn": "arn:aws:sns:us-east-1:992648334831:MyDemoSNSTopic", - "SourceIds": [ + "SnsTopicArn":"arn:aws:sns:us-east-1:992648334831:MyDemoSNSTopic", + "SourceIds":[ "mymysqlinstance" ], - "SourceType": "db-instance", - "SubscriptionName": "mymysqleventsubscription" + "SourceType":"db-instance", + "SubscriptionName":"mymysqleventsubscription" }, - "output": { - "EventSubscription": { + "output":{ + "EventSubscription":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates an event notification subscription.", - "id": "create-event-subscription-00dd0ee6-0e0f-4a38-ae83-e5f2ded5f69a", - "title": "To create an event notification subscription" + "description":"This example creates an event notification subscription.", + "id":"create-event-subscription-00dd0ee6-0e0f-4a38-ae83-e5f2ded5f69a", + "title":"To create an event notification subscription" } ], - "CreateOptionGroup": [ + "CreateOptionGroup":[ { - "input": { - "EngineName": "MySQL", - "MajorEngineVersion": "5.6", - "OptionGroupDescription": "My MySQL 5.6 option group", - "OptionGroupName": "mymysqloptiongroup" + "input":{ + "EngineName":"MySQL", + "MajorEngineVersion":"5.6", + "OptionGroupDescription":"My MySQL 5.6 option group", + "OptionGroupName":"mymysqloptiongroup" }, - "output": { - "OptionGroup": { + "output":{ + "OptionGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example creates an option group.", - "id": "create-option-group-a7708c87-1b79-4a5e-a762-21cf8fc62b78", - "title": "To create an option group" + "description":"This example creates an option group.", + "id":"create-option-group-a7708c87-1b79-4a5e-a762-21cf8fc62b78", + "title":"To create an option group" } ], - "DeleteDBCluster": [ + "DeleteDBCluster":[ { - "input": { - "DBClusterIdentifier": "mydbcluster", - "SkipFinalSnapshot": true + "input":{ + "DBClusterIdentifier":"mydbcluster", + "SkipFinalSnapshot":true }, - "output": { - "DBCluster": { + "output":{ + "DBCluster":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB cluster.", - "id": "delete-db-cluster-927fc2c8-6c67-4075-b1ba-75490be0f7d6", - "title": "To delete a DB cluster." + "description":"This example deletes the specified DB cluster.", + "id":"delete-db-cluster-927fc2c8-6c67-4075-b1ba-75490be0f7d6", + "title":"To delete a DB cluster." } ], - "DeleteDBClusterParameterGroup": [ + "DeleteDBClusterParameterGroup":[ { - "input": { - "DBClusterParameterGroupName": "mydbclusterparametergroup" + "input":{ + "DBClusterParameterGroupName":"mydbclusterparametergroup" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB cluster parameter group.", - "id": "delete-db-cluster-parameter-group-364f5555-ba0a-4cc8-979c-e769098924fc", - "title": "To delete a DB cluster parameter group." + "description":"This example deletes the specified DB cluster parameter group.", + "id":"delete-db-cluster-parameter-group-364f5555-ba0a-4cc8-979c-e769098924fc", + "title":"To delete a DB cluster parameter group." } ], - "DeleteDBClusterSnapshot": [ + "DeleteDBClusterSnapshot":[ { - "input": { - "DBClusterSnapshotIdentifier": "mydbclustersnapshot" + "input":{ + "DBClusterSnapshotIdentifier":"mydbclustersnapshot" }, - "output": { - "DBClusterSnapshot": { + "output":{ + "DBClusterSnapshot":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB cluster snapshot.", - "id": "delete-db-cluster-snapshot-c67e0d95-670e-4fb5-af90-6d9a70a91b07", - "title": "To delete a DB cluster snapshot." + "description":"This example deletes the specified DB cluster snapshot.", + "id":"delete-db-cluster-snapshot-c67e0d95-670e-4fb5-af90-6d9a70a91b07", + "title":"To delete a DB cluster snapshot." } ], - "DeleteDBInstance": [ + "DeleteDBInstance":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance", - "SkipFinalSnapshot": true + "input":{ + "DBInstanceIdentifier":"mymysqlinstance", + "SkipFinalSnapshot":true }, - "output": { - "DBInstance": { + "output":{ + "DBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB instance.", - "id": "delete-db-instance-4412e650-949c-488a-b32a-7d3038ebccc4", - "title": "To delete a DB instance." + "description":"This example deletes the specified DB instance.", + "id":"delete-db-instance-4412e650-949c-488a-b32a-7d3038ebccc4", + "title":"To delete a DB instance." } ], - "DeleteDBParameterGroup": [ + "DeleteDBParameterGroup":[ { - "input": { - "DBParameterGroupName": "mydbparamgroup3" + "input":{ + "DBParameterGroupName":"mydbparamgroup3" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example deletes a DB parameter group.", - "id": "to-delete-a-db-parameter-group-1473888796509", - "title": "To delete a DB parameter group" + "description":"The following example deletes a DB parameter group.", + "id":"to-delete-a-db-parameter-group-1473888796509", + "title":"To delete a DB parameter group" } ], - "DeleteDBSecurityGroup": [ + "DeleteDBSecurityGroup":[ { - "input": { - "DBSecurityGroupName": "mysecgroup" + "input":{ + "DBSecurityGroupName":"mysecgroup" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example deletes a DB security group.", - "id": "to-delete-a-db-security-group-1473960141889", - "title": "To delete a DB security group" + "description":"The following example deletes a DB security group.", + "id":"to-delete-a-db-security-group-1473960141889", + "title":"To delete a DB security group" } ], - "DeleteDBSnapshot": [ + "DeleteDBSnapshot":[ { - "input": { - "DBSnapshotIdentifier": "mydbsnapshot" + "input":{ + "DBSnapshotIdentifier":"mydbsnapshot" }, - "output": { - "DBSnapshot": { + "output":{ + "DBSnapshot":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB snapshot.", - "id": "delete-db-snapshot-505d6b4e-8ced-479c-856a-c460a33fe07b", - "title": "To delete a DB cluster snapshot." + "description":"This example deletes the specified DB snapshot.", + "id":"delete-db-snapshot-505d6b4e-8ced-479c-856a-c460a33fe07b", + "title":"To delete a DB cluster snapshot." } ], - "DeleteDBSubnetGroup": [ + "DeleteDBSubnetGroup":[ { - "input": { - "DBSubnetGroupName": "mydbsubnetgroup" + "input":{ + "DBSubnetGroupName":"mydbsubnetgroup" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB subnetgroup.", - "id": "delete-db-subnet-group-4ae00375-511e-443d-a01d-4b9f552244aa", - "title": "To delete a DB subnet group." + "description":"This example deletes the specified DB subnetgroup.", + "id":"delete-db-subnet-group-4ae00375-511e-443d-a01d-4b9f552244aa", + "title":"To delete a DB subnet group." } ], - "DeleteEventSubscription": [ + "DeleteEventSubscription":[ { - "input": { - "SubscriptionName": "myeventsubscription" + "input":{ + "SubscriptionName":"myeventsubscription" }, - "output": { - "EventSubscription": { + "output":{ + "EventSubscription":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified DB event subscription.", - "id": "delete-db-event-subscription-d33567e3-1d5d-48ff-873f-0270453f4a75", - "title": "To delete a DB event subscription." + "description":"This example deletes the specified DB event subscription.", + "id":"delete-db-event-subscription-d33567e3-1d5d-48ff-873f-0270453f4a75", + "title":"To delete a DB event subscription." } ], - "DeleteOptionGroup": [ + "DeleteOptionGroup":[ { - "input": { - "OptionGroupName": "mydboptiongroup" + "input":{ + "OptionGroupName":"mydboptiongroup" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified option group.", - "id": "delete-db-option-group-578be2be-3095-431a-9ea4-9a3c3b0daef4", - "title": "To delete an option group." + "description":"This example deletes the specified option group.", + "id":"delete-db-option-group-578be2be-3095-431a-9ea4-9a3c3b0daef4", + "title":"To delete an option group." } ], - "DescribeAccountAttributes": [ + "DescribeAccountAttributes":[ { - "input": { + "input":{ }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists account attributes.", - "id": "describe-account-attributes-683d3ff7-5524-421a-8da5-e88f1ea2222b", - "title": "To list account attributes" + "description":"This example lists account attributes.", + "id":"describe-account-attributes-683d3ff7-5524-421a-8da5-e88f1ea2222b", + "title":"To list account attributes" } ], - "DescribeCertificates": [ + "DescribeCertificates":[ { - "input": { - "CertificateIdentifier": "rds-ca-2015", - "MaxRecords": 20 + "input":{ + "CertificateIdentifier":"rds-ca-2015", + "MaxRecords":20 }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists up to 20 certificates for the specified certificate identifier.", - "id": "describe-certificates-9d71a70d-7908-4444-b43f-321d842c62dc", - "title": "To list certificates" + "description":"This example lists up to 20 certificates for the specified certificate identifier.", + "id":"describe-certificates-9d71a70d-7908-4444-b43f-321d842c62dc", + "title":"To list certificates" } ], - "DescribeDBClusterParameterGroups": [ + "DescribeDBClusterParameterGroups":[ { - "input": { - "DBClusterParameterGroupName": "mydbclusterparametergroup" + "input":{ + "DBClusterParameterGroupName":"mydbclusterparametergroup" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists settings for the specified DB cluster parameter group.", - "id": "describe-db-cluster-parameter-groups-cf9c6e66-664e-4f57-8e29-a9080abfc013", - "title": "To list DB cluster parameter group settings" + "description":"This example lists settings for the specified DB cluster parameter group.", + "id":"describe-db-cluster-parameter-groups-cf9c6e66-664e-4f57-8e29-a9080abfc013", + "title":"To list DB cluster parameter group settings" } ], - "DescribeDBClusterParameters": [ + "DescribeDBClusterParameters":[ { - "input": { - "DBClusterParameterGroupName": "mydbclusterparametergroup", - "Source": "system" + "input":{ + "DBClusterParameterGroupName":"mydbclusterparametergroup", + "Source":"system" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists system parameters for the specified DB cluster parameter group.", - "id": "describe-db-cluster-parameters-98043c28-e489-41a7-b118-bfd96dc779a1", - "title": "To list DB cluster parameters" + "description":"This example lists system parameters for the specified DB cluster parameter group.", + "id":"describe-db-cluster-parameters-98043c28-e489-41a7-b118-bfd96dc779a1", + "title":"To list DB cluster parameters" } ], - "DescribeDBClusterSnapshotAttributes": [ + "DescribeDBClusterSnapshotAttributes":[ { - "input": { - "DBClusterSnapshotIdentifier": "mydbclustersnapshot" + "input":{ + "DBClusterSnapshotIdentifier":"mydbclustersnapshot" }, - "output": { - "DBClusterSnapshotAttributesResult": { + "output":{ + "DBClusterSnapshotAttributesResult":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists attributes for the specified DB cluster snapshot.", - "id": "describe-db-cluster-snapshot-attributes-6752ade3-0c7b-4b06-a8e4-b76bf4e2d3571", - "title": "To list DB cluster snapshot attributes" + "description":"This example lists attributes for the specified DB cluster snapshot.", + "id":"describe-db-cluster-snapshot-attributes-6752ade3-0c7b-4b06-a8e4-b76bf4e2d3571", + "title":"To list DB cluster snapshot attributes" } ], - "DescribeDBClusterSnapshots": [ + "DescribeDBClusterSnapshots":[ { - "input": { - "DBClusterSnapshotIdentifier": "mydbclustersnapshot", - "SnapshotType": "manual" + "input":{ + "DBClusterSnapshotIdentifier":"mydbclustersnapshot", + "SnapshotType":"manual" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists settings for the specified, manually-created cluster snapshot.", - "id": "describe-db-cluster-snapshots-52f38af1-3431-4a51-9a6a-e6bb8c961b32", - "title": "To list DB cluster snapshots" + "description":"This example lists settings for the specified, manually-created cluster snapshot.", + "id":"describe-db-cluster-snapshots-52f38af1-3431-4a51-9a6a-e6bb8c961b32", + "title":"To list DB cluster snapshots" } ], - "DescribeDBClusters": [ + "DescribeDBClusters":[ { - "input": { - "DBClusterIdentifier": "mynewdbcluster" + "input":{ + "DBClusterIdentifier":"mynewdbcluster" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists settings for the specified DB cluster.", - "id": "describe-db-clusters-7aae8861-cb95-4b3b-9042-f62df7698635", - "title": "To list DB clusters" + "description":"This example lists settings for the specified DB cluster.", + "id":"describe-db-clusters-7aae8861-cb95-4b3b-9042-f62df7698635", + "title":"To list DB clusters" } ], - "DescribeDBEngineVersions": [ + "DescribeDBEngineVersions":[ { - "input": { - "DBParameterGroupFamily": "mysql5.6", - "DefaultOnly": true, - "Engine": "mysql", - "EngineVersion": "5.6", - "ListSupportedCharacterSets": true + "input":{ + "DBParameterGroupFamily":"mysql5.6", + "DefaultOnly":true, + "Engine":"mysql", + "EngineVersion":"5.6", + "ListSupportedCharacterSets":true }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists settings for the specified DB engine version.", - "id": "describe-db-engine-versions-8e698cf2-2162-425a-a854-111cdaceb52b", - "title": "To list DB engine version settings" + "description":"This example lists settings for the specified DB engine version.", + "id":"describe-db-engine-versions-8e698cf2-2162-425a-a854-111cdaceb52b", + "title":"To list DB engine version settings" } ], - "DescribeDBInstances": [ + "DescribeDBInstances":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance" + "input":{ + "DBInstanceIdentifier":"mymysqlinstance" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists settings for the specified DB instance.", - "id": "describe-db-instances-0e11a8c5-4ec3-4463-8cbf-f7254d04c4fc", - "title": "To list DB instance settings" + "description":"This example lists settings for the specified DB instance.", + "id":"describe-db-instances-0e11a8c5-4ec3-4463-8cbf-f7254d04c4fc", + "title":"To list DB instance settings" } ], - "DescribeDBLogFiles": [ + "DescribeDBLogFiles":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance", - "FileLastWritten": 1470873600000, - "FileSize": 0, - "FilenameContains": "error" + "input":{ + "DBInstanceIdentifier":"mymysqlinstance", + "FileLastWritten":1470873600000, + "FileSize":0, + "FilenameContains":"error" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists matching log file names for the specified DB instance, file name pattern, last write date in POSIX time with milleseconds, and minimum file size.", - "id": "describe-db-log-files-5f002d8d-5c1d-44c2-b5f4-bd284c0f1285", - "title": "To list DB log file names" + "description":"This example lists matching log file names for the specified DB instance, file name pattern, last write date in POSIX time with milleseconds, and minimum file size.", + "id":"describe-db-log-files-5f002d8d-5c1d-44c2-b5f4-bd284c0f1285", + "title":"To list DB log file names" } ], - "DescribeDBParameterGroups": [ + "DescribeDBParameterGroups":[ { - "input": { - "DBParameterGroupName": "mymysqlparametergroup" + "input":{ + "DBParameterGroupName":"mymysqlparametergroup" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information about the specified DB parameter group.", - "id": "describe-db-parameter-groups-", - "title": "To list information about DB parameter groups" + "description":"This example lists information about the specified DB parameter group.", + "id":"describe-db-parameter-groups-", + "title":"To list information about DB parameter groups" } ], - "DescribeDBParameters": [ + "DescribeDBParameters":[ { - "input": { - "DBParameterGroupName": "mymysqlparametergroup", - "MaxRecords": 20, - "Source": "system" + "input":{ + "DBParameterGroupName":"mymysqlparametergroup", + "MaxRecords":20, + "Source":"system" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for up to the first 20 system parameters for the specified DB parameter group.", - "id": "describe-db-parameters-09db4201-ef4f-4d97-a4b5-d71c0715b901", - "title": "To list information about DB parameters" + "description":"This example lists information for up to the first 20 system parameters for the specified DB parameter group.", + "id":"describe-db-parameters-09db4201-ef4f-4d97-a4b5-d71c0715b901", + "title":"To list information about DB parameters" } ], - "DescribeDBSecurityGroups": [ + "DescribeDBSecurityGroups":[ { - "input": { - "DBSecurityGroupName": "mydbsecuritygroup" + "input":{ + "DBSecurityGroupName":"mydbsecuritygroup" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists settings for the specified security group.", - "id": "describe-db-security-groups-66fe9ea1-17dd-4275-b82e-f771cee0c849", - "title": "To list DB security group settings" + "description":"This example lists settings for the specified security group.", + "id":"describe-db-security-groups-66fe9ea1-17dd-4275-b82e-f771cee0c849", + "title":"To list DB security group settings" } ], - "DescribeDBSnapshotAttributes": [ + "DescribeDBSnapshotAttributes":[ { - "input": { - "DBSnapshotIdentifier": "mydbsnapshot" + "input":{ + "DBSnapshotIdentifier":"mydbsnapshot" }, - "output": { - "DBSnapshotAttributesResult": { + "output":{ + "DBSnapshotAttributesResult":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists attributes for the specified DB snapshot.", - "id": "describe-db-snapshot-attributes-1d4fb750-34f6-4e43-8b3d-b2751d796a95", - "title": "To list DB snapshot attributes" + "description":"This example lists attributes for the specified DB snapshot.", + "id":"describe-db-snapshot-attributes-1d4fb750-34f6-4e43-8b3d-b2751d796a95", + "title":"To list DB snapshot attributes" } ], - "DescribeDBSnapshots": [ + "DescribeDBSnapshots":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance", - "IncludePublic": false, - "IncludeShared": true, - "SnapshotType": "manual" + "input":{ + "DBInstanceIdentifier":"mymysqlinstance", + "IncludePublic":false, + "IncludeShared":true, + "SnapshotType":"manual" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists all manually-created, shared snapshots for the specified DB instance.", - "id": "describe-db-snapshots-2c935989-a1ef-4c85-aea4-1d0f45f17f26", - "title": "To list DB snapshot attributes" + "description":"This example lists all manually-created, shared snapshots for the specified DB instance.", + "id":"describe-db-snapshots-2c935989-a1ef-4c85-aea4-1d0f45f17f26", + "title":"To list DB snapshot attributes" } ], - "DescribeDBSubnetGroups": [ + "DescribeDBSubnetGroups":[ { - "input": { - "DBSubnetGroupName": "mydbsubnetgroup" + "input":{ + "DBSubnetGroupName":"mydbsubnetgroup" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information about the specified DB subnet group.", - "id": "describe-db-subnet-groups-1d97b340-682f-4dd6-9653-8ed72a8d1221", - "title": "To list information about DB subnet groups" + "description":"This example lists information about the specified DB subnet group.", + "id":"describe-db-subnet-groups-1d97b340-682f-4dd6-9653-8ed72a8d1221", + "title":"To list information about DB subnet groups" } ], - "DescribeEngineDefaultClusterParameters": [ + "DescribeEngineDefaultClusterParameters":[ { - "input": { - "DBParameterGroupFamily": "aurora5.6" + "input":{ + "DBParameterGroupFamily":"aurora5.6" }, - "output": { - "EngineDefaults": { + "output":{ + "EngineDefaults":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists default parameters for the specified DB cluster engine.", - "id": "describe-engine-default-cluster-parameters-f130374a-7bee-434b-b51d-da20b6e000e0", - "title": "To list default parameters for a DB cluster engine" + "description":"This example lists default parameters for the specified DB cluster engine.", + "id":"describe-engine-default-cluster-parameters-f130374a-7bee-434b-b51d-da20b6e000e0", + "title":"To list default parameters for a DB cluster engine" } ], - "DescribeEngineDefaultParameters": [ + "DescribeEngineDefaultParameters":[ { - "input": { - "DBParameterGroupFamily": "mysql5.6" + "input":{ + "DBParameterGroupFamily":"mysql5.6" }, - "output": { - "EngineDefaults": { + "output":{ + "EngineDefaults":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists default parameters for the specified DB engine.", - "id": "describe-engine-default-parameters-35d5108e-1d44-4fac-8aeb-04b8fdfface1", - "title": "To list default parameters for a DB engine" + "description":"This example lists default parameters for the specified DB engine.", + "id":"describe-engine-default-parameters-35d5108e-1d44-4fac-8aeb-04b8fdfface1", + "title":"To list default parameters for a DB engine" } ], - "DescribeEventCategories": [ + "DescribeEventCategories":[ { - "input": { - "SourceType": "db-instance" + "input":{ + "SourceType":"db-instance" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists all DB instance event categories.", - "id": "describe-event-categories-97bd4c77-12da-4be6-b42f-edf77771428b", - "title": "To list event categories." + "description":"This example lists all DB instance event categories.", + "id":"describe-event-categories-97bd4c77-12da-4be6-b42f-edf77771428b", + "title":"To list event categories." } ], - "DescribeEventSubscriptions": [ + "DescribeEventSubscriptions":[ { - "input": { - "SubscriptionName": "mymysqleventsubscription" + "input":{ + "SubscriptionName":"mymysqleventsubscription" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for the specified DB event notification subscription.", - "id": "describe-event-subscriptions-11184a82-e58a-4d0c-b558-f3a7489e0850", - "title": "To list information about DB event notification subscriptions" + "description":"This example lists information for the specified DB event notification subscription.", + "id":"describe-event-subscriptions-11184a82-e58a-4d0c-b558-f3a7489e0850", + "title":"To list information about DB event notification subscriptions" } ], - "DescribeEvents": [ + "DescribeEvents":[ { - "input": { - "Duration": 10080, - "EventCategories": [ + "input":{ + "Duration":10080, + "EventCategories":[ "backup" ], - "SourceIdentifier": "mymysqlinstance", - "SourceType": "db-instance" + "SourceIdentifier":"mymysqlinstance", + "SourceType":"db-instance" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all backup-related events for the specified DB instance for the past 7 days (7 days * 24 hours * 60 minutes = 10,080 minutes).", - "id": "describe-events-3836e5ed-3913-4f76-8452-c77fcad5016b", - "title": "To list information about events" + "description":"This example lists information for all backup-related events for the specified DB instance for the past 7 days (7 days * 24 hours * 60 minutes = 10,080 minutes).", + "id":"describe-events-3836e5ed-3913-4f76-8452-c77fcad5016b", + "title":"To list information about events" } ], - "DescribeOptionGroupOptions": [ + "DescribeOptionGroupOptions":[ { - "input": { - "EngineName": "mysql", - "MajorEngineVersion": "5.6" + "input":{ + "EngineName":"mysql", + "MajorEngineVersion":"5.6" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all option group options for the specified DB engine.", - "id": "describe-option-group-options-30d735a4-81f1-49e4-b3f2-5dc45d50c8ed", - "title": "To list information about DB option group options" + "description":"This example lists information for all option group options for the specified DB engine.", + "id":"describe-option-group-options-30d735a4-81f1-49e4-b3f2-5dc45d50c8ed", + "title":"To list information about DB option group options" } ], - "DescribeOptionGroups": [ + "DescribeOptionGroups":[ { - "input": { - "EngineName": "mysql", - "MajorEngineVersion": "5.6" + "input":{ + "EngineName":"mysql", + "MajorEngineVersion":"5.6" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all option groups for the specified DB engine.", - "id": "describe-option-groups-4ef478a1-66d5-45f2-bec3-e608720418a4", - "title": "To list information about DB option groups" + "description":"This example lists information for all option groups for the specified DB engine.", + "id":"describe-option-groups-4ef478a1-66d5-45f2-bec3-e608720418a4", + "title":"To list information about DB option groups" } ], - "DescribeOrderableDBInstanceOptions": [ + "DescribeOrderableDBInstanceOptions":[ { - "input": { - "DBInstanceClass": "db.t2.micro", - "Engine": "mysql", - "EngineVersion": "5.6.27", - "LicenseModel": "general-public-license", - "Vpc": true + "input":{ + "DBInstanceClass":"db.t2.micro", + "Engine":"mysql", + "EngineVersion":"5.6.27", + "LicenseModel":"general-public-license", + "Vpc":true }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all orderable DB instance options for the specified DB engine, engine version, DB instance class, license model, and VPC settings.", - "id": "describe-orderable-db-instance-options-7444d3ed-82eb-42b9-9ed9-896b8c27a782", - "title": "To list information about orderable DB instance options" + "description":"This example lists information for all orderable DB instance options for the specified DB engine, engine version, DB instance class, license model, and VPC settings.", + "id":"describe-orderable-db-instance-options-7444d3ed-82eb-42b9-9ed9-896b8c27a782", + "title":"To list information about orderable DB instance options" } ], - "DescribePendingMaintenanceActions": [ + "DescribePendingMaintenanceActions":[ { - "input": { - "ResourceIdentifier": "arn:aws:rds:us-east-1:992648334831:db:mymysqlinstance" + "input":{ + "ResourceIdentifier":"arn:aws:rds:us-east-1:992648334831:db:mymysqlinstance" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all pending maintenance actions for the specified DB instance.", - "id": "describe-pending-maintenance-actions-e6021f7e-58ae-49cc-b874-11996176835c", - "title": "To list information about pending maintenance actions" + "description":"This example lists information for all pending maintenance actions for the specified DB instance.", + "id":"describe-pending-maintenance-actions-e6021f7e-58ae-49cc-b874-11996176835c", + "title":"To list information about pending maintenance actions" } ], - "DescribeReservedDBInstances": [ + "DescribeReservedDBInstances":[ { - "input": { - "DBInstanceClass": "db.t2.micro", - "Duration": "1y", - "MultiAZ": false, - "OfferingType": "No Upfront", - "ProductDescription": "mysql" + "input":{ + "DBInstanceClass":"db.t2.micro", + "Duration":"1y", + "MultiAZ":false, + "OfferingType":"No Upfront", + "ProductDescription":"mysql" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all reserved DB instances for the specified DB instance class, duration, product, offering type, and availability zone settings.", - "id": "describe-reserved-db-instances-d45adaca-2e30-407c-a0f3-aa7b98bea17f", - "title": "To list information about reserved DB instances" + "description":"This example lists information for all reserved DB instances for the specified DB instance class, duration, product, offering type, and availability zone settings.", + "id":"describe-reserved-db-instances-d45adaca-2e30-407c-a0f3-aa7b98bea17f", + "title":"To list information about reserved DB instances" } ], - "DescribeReservedDBInstancesOfferings": [ + "DescribeReservedDBInstancesOfferings":[ { - "input": { - "DBInstanceClass": "db.t2.micro", - "Duration": "1y", - "MultiAZ": false, - "OfferingType": "No Upfront", - "ProductDescription": "mysql" + "input":{ + "DBInstanceClass":"db.t2.micro", + "Duration":"1y", + "MultiAZ":false, + "OfferingType":"No Upfront", + "ProductDescription":"mysql" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for all reserved DB instance offerings for the specified DB instance class, duration, product, offering type, and availability zone settings.", - "id": "describe-reserved-db-instances-offerings-9de7d1fd-d6a6-4a72-84ae-b2ef58d47d8d", - "title": "To list information about reserved DB instance offerings" + "description":"This example lists information for all reserved DB instance offerings for the specified DB instance class, duration, product, offering type, and availability zone settings.", + "id":"describe-reserved-db-instances-offerings-9de7d1fd-d6a6-4a72-84ae-b2ef58d47d8d", + "title":"To list information about reserved DB instance offerings" } ], - "DescribeSourceRegions": [ + "DescribeSourceRegions":[ { - "input": { + "input":{ }, - "output": { - "SourceRegions": [ + "output":{ + "SourceRegions":[ { - "Endpoint": "https://rds.ap-northeast-1.amazonaws.com", - "RegionName": "ap-northeast-1", - "Status": "available" + "Endpoint":"https://rds.ap-northeast-1.amazonaws.com", + "RegionName":"ap-northeast-1", + "Status":"available" }, { - "Endpoint": "https://rds.ap-northeast-2.amazonaws.com", - "RegionName": "ap-northeast-2", - "Status": "available" + "Endpoint":"https://rds.ap-northeast-2.amazonaws.com", + "RegionName":"ap-northeast-2", + "Status":"available" }, { - "Endpoint": "https://rds.ap-south-1.amazonaws.com", - "RegionName": "ap-south-1", - "Status": "available" + "Endpoint":"https://rds.ap-south-1.amazonaws.com", + "RegionName":"ap-south-1", + "Status":"available" }, { - "Endpoint": "https://rds.ap-southeast-1.amazonaws.com", - "RegionName": "ap-southeast-1", - "Status": "available" + "Endpoint":"https://rds.ap-southeast-1.amazonaws.com", + "RegionName":"ap-southeast-1", + "Status":"available" }, { - "Endpoint": "https://rds.ap-southeast-2.amazonaws.com", - "RegionName": "ap-southeast-2", - "Status": "available" + "Endpoint":"https://rds.ap-southeast-2.amazonaws.com", + "RegionName":"ap-southeast-2", + "Status":"available" }, { - "Endpoint": "https://rds.eu-central-1.amazonaws.com", - "RegionName": "eu-central-1", - "Status": "available" + "Endpoint":"https://rds.eu-central-1.amazonaws.com", + "RegionName":"eu-central-1", + "Status":"available" }, { - "Endpoint": "https://rds.eu-west-1.amazonaws.com", - "RegionName": "eu-west-1", - "Status": "available" + "Endpoint":"https://rds.eu-west-1.amazonaws.com", + "RegionName":"eu-west-1", + "Status":"available" }, { - "Endpoint": "https://rds.sa-east-1.amazonaws.com", - "RegionName": "sa-east-1", - "Status": "available" + "Endpoint":"https://rds.sa-east-1.amazonaws.com", + "RegionName":"sa-east-1", + "Status":"available" }, { - "Endpoint": "https://rds.us-west-1.amazonaws.com", - "RegionName": "us-west-1", - "Status": "available" + "Endpoint":"https://rds.us-west-1.amazonaws.com", + "RegionName":"us-west-1", + "Status":"available" }, { - "Endpoint": "https://rds.us-west-2.amazonaws.com", - "RegionName": "us-west-2", - "Status": "available" + "Endpoint":"https://rds.us-west-2.amazonaws.com", + "RegionName":"us-west-2", + "Status":"available" } ] }, - "comments": { + "comments":{ }, - "description": "To list the AWS regions where a Read Replica can be created.", - "id": "to-describe-source-regions-1473457722410", - "title": "To describe source regions" + "description":"To list the AWS regions where a Read Replica can be created.", + "id":"to-describe-source-regions-1473457722410", + "title":"To describe source regions" } ], - "DownloadDBLogFilePortion": [ + "DownloadDBLogFilePortion":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance", - "LogFileName": "mysqlUpgrade" + "input":{ + "DBInstanceIdentifier":"mymysqlinstance", + "LogFileName":"mysqlUpgrade" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information for the specified log file for the specified DB instance.", - "id": "download-db-log-file-portion-54a82731-a441-4fc7-a010-8eccae6fa202", - "title": "To list information about DB log files" + "description":"This example lists information for the specified log file for the specified DB instance.", + "id":"download-db-log-file-portion-54a82731-a441-4fc7-a010-8eccae6fa202", + "title":"To list information about DB log files" } ], - "FailoverDBCluster": [ + "FailoverDBCluster":[ { - "input": { - "DBClusterIdentifier": "myaurorainstance-cluster", - "TargetDBInstanceIdentifier": "myaurorareplica" + "input":{ + "DBClusterIdentifier":"myaurorainstance-cluster", + "TargetDBInstanceIdentifier":"myaurorareplica" }, - "output": { - "DBCluster": { + "output":{ + "DBCluster":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example performs a failover for the specified DB cluster to the specified DB instance.", - "id": "failover-db-cluster-9e7f2f93-d98c-42c7-bb0e-d6c485c096d6", - "title": "To perform a failover for a DB cluster" + "description":"This example performs a failover for the specified DB cluster to the specified DB instance.", + "id":"failover-db-cluster-9e7f2f93-d98c-42c7-bb0e-d6c485c096d6", + "title":"To perform a failover for a DB cluster" } ], - "ListTagsForResource": [ + "ListTagsForResource":[ { - "input": { - "ResourceName": "arn:aws:rds:us-east-1:992648334831:og:mymysqloptiongroup" + "input":{ + "ResourceName":"arn:aws:rds:us-east-1:992648334831:og:mymysqloptiongroup" }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example lists information about all tags associated with the specified DB option group.", - "id": "list-tags-for-resource-8401f3c2-77cd-4f90-bfd5-b523f0adcc2f", - "title": "To list information about tags associated with a resource" + "description":"This example lists information about all tags associated with the specified DB option group.", + "id":"list-tags-for-resource-8401f3c2-77cd-4f90-bfd5-b523f0adcc2f", + "title":"To list information about tags associated with a resource" } ], - "ModifyDBCluster": [ + "ModifyDBCluster":[ { - "input": { - "ApplyImmediately": true, - "DBClusterIdentifier": "mydbcluster", - "MasterUserPassword": "mynewpassword", - "NewDBClusterIdentifier": "mynewdbcluster", - "PreferredBackupWindow": "04:00-04:30", - "PreferredMaintenanceWindow": "Tue:05:00-Tue:05:30" + "input":{ + "ApplyImmediately":true, + "DBClusterIdentifier":"mydbcluster", + "MasterUserPassword":"mynewpassword", + "NewDBClusterIdentifier":"mynewdbcluster", + "PreferredBackupWindow":"04:00-04:30", + "PreferredMaintenanceWindow":"Tue:05:00-Tue:05:30" }, - "output": { - "DBCluster": { + "output":{ + "DBCluster":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example changes the specified settings for the specified DB cluster.", - "id": "modify-db-cluster-a370ee1b-768d-450a-853b-707cb1ab663d", - "title": "To change DB cluster settings" + "description":"This example changes the specified settings for the specified DB cluster.", + "id":"modify-db-cluster-a370ee1b-768d-450a-853b-707cb1ab663d", + "title":"To change DB cluster settings" } ], - "ModifyDBClusterParameterGroup": [ + "ModifyDBClusterParameterGroup":[ { - "input": { - "DBClusterParameterGroupName": "mydbclusterparametergroup", - "Parameters": [ + "input":{ + "DBClusterParameterGroupName":"mydbclusterparametergroup", + "Parameters":[ { - "ApplyMethod": "immediate", - "ParameterName": "time_zone", - "ParameterValue": "America/Phoenix" + "ApplyMethod":"immediate", + "ParameterName":"time_zone", + "ParameterValue":"America/Phoenix" } ] }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example immediately changes the specified setting for the specified DB cluster parameter group.", - "id": "modify-db-cluster-parameter-group-f9156bc9-082a-442e-8d12-239542c1a113", - "title": "To change DB cluster parameter group settings" + "description":"This example immediately changes the specified setting for the specified DB cluster parameter group.", + "id":"modify-db-cluster-parameter-group-f9156bc9-082a-442e-8d12-239542c1a113", + "title":"To change DB cluster parameter group settings" } ], - "ModifyDBClusterSnapshotAttribute": [ + "ModifyDBClusterSnapshotAttribute":[ { - "input": { - "AttributeName": "restore", - "DBClusterSnapshotIdentifier": "manual-cluster-snapshot1", - "ValuesToAdd": [ + "input":{ + "AttributeName":"restore", + "DBClusterSnapshotIdentifier":"manual-cluster-snapshot1", + "ValuesToAdd":[ "123451234512", "123456789012" ], - "ValuesToRemove": [ + "ValuesToRemove":[ "all" ] }, - "output": { - "DBClusterSnapshotAttributesResult": { + "output":{ + "DBClusterSnapshotAttributesResult":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example gives two AWS accounts access to a manual DB cluster snapshot and ensures that the DB cluster snapshot is private by removing the value \"all\".", - "id": "to-add-or-remove-access-to-a-manual-db-cluster-snapshot-1473889426431", - "title": "To add or remove access to a manual DB cluster snapshot" + "description":"The following example gives two AWS accounts access to a manual DB cluster snapshot and ensures that the DB cluster snapshot is private by removing the value \"all\".", + "id":"to-add-or-remove-access-to-a-manual-db-cluster-snapshot-1473889426431", + "title":"To add or remove access to a manual DB cluster snapshot" } ], - "ModifyDBInstance": [ + "ModifyDBInstance":[ { - "input": { - "AllocatedStorage": 10, - "ApplyImmediately": true, - "BackupRetentionPeriod": 1, - "DBInstanceClass": "db.t2.small", - "DBInstanceIdentifier": "mymysqlinstance", - "MasterUserPassword": "mynewpassword", - "PreferredBackupWindow": "04:00-04:30", - "PreferredMaintenanceWindow": "Tue:05:00-Tue:05:30" + "input":{ + "AllocatedStorage":10, + "ApplyImmediately":true, + "BackupRetentionPeriod":1, + "DBInstanceClass":"db.t2.small", + "DBInstanceIdentifier":"mymysqlinstance", + "MasterUserPassword":"mynewpassword", + "PreferredBackupWindow":"04:00-04:30", + "PreferredMaintenanceWindow":"Tue:05:00-Tue:05:30" }, - "output": { - "DBInstance": { + "output":{ + "DBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example immediately changes the specified settings for the specified DB instance.", - "id": "modify-db-instance-6979a368-6254-467b-8a8d-61103f4fcde9", - "title": "To change DB instance settings" + "description":"This example immediately changes the specified settings for the specified DB instance.", + "id":"modify-db-instance-6979a368-6254-467b-8a8d-61103f4fcde9", + "title":"To change DB instance settings" } ], - "ModifyDBParameterGroup": [ + "ModifyDBParameterGroup":[ { - "input": { - "DBParameterGroupName": "mymysqlparametergroup", - "Parameters": [ + "input":{ + "DBParameterGroupName":"mymysqlparametergroup", + "Parameters":[ { - "ApplyMethod": "immediate", - "ParameterName": "time_zone", - "ParameterValue": "America/Phoenix" + "ApplyMethod":"immediate", + "ParameterName":"time_zone", + "ParameterValue":"America/Phoenix" } ] }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example immediately changes the specified setting for the specified DB parameter group.", - "id": "modify-db-parameter-group-f3a4e52a-68e4-4b88-b559-f912d34c457a", - "title": "To change DB parameter group settings" + "description":"This example immediately changes the specified setting for the specified DB parameter group.", + "id":"modify-db-parameter-group-f3a4e52a-68e4-4b88-b559-f912d34c457a", + "title":"To change DB parameter group settings" } ], - "ModifyDBSnapshotAttribute": [ + "ModifyDBSnapshotAttribute":[ { - "input": { - "AttributeName": "restore", - "DBSnapshotIdentifier": "mydbsnapshot", - "ValuesToAdd": [ + "input":{ + "AttributeName":"restore", + "DBSnapshotIdentifier":"mydbsnapshot", + "ValuesToAdd":[ "all" ] }, - "output": { - "DBSnapshotAttributesResult": { + "output":{ + "DBSnapshotAttributesResult":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example adds the specified attribute for the specified DB snapshot.", - "id": "modify-db-snapshot-attribute-2e66f120-2b21-4a7c-890b-4474da88bde6", - "title": "To change DB snapshot attributes" + "description":"This example adds the specified attribute for the specified DB snapshot.", + "id":"modify-db-snapshot-attribute-2e66f120-2b21-4a7c-890b-4474da88bde6", + "title":"To change DB snapshot attributes" } ], - "ModifyDBSubnetGroup": [ + "ModifyDBSubnetGroup":[ { - "input": { - "DBSubnetGroupName": "mydbsubnetgroup", - "SubnetIds": [ + "input":{ + "DBSubnetGroupName":"mydbsubnetgroup", + "SubnetIds":[ "subnet-70e1975a", "subnet-747a5c49" ] }, - "output": { - "DBSubnetGroup": { + "output":{ + "DBSubnetGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example changes the specified setting for the specified DB subnet group.", - "id": "modify-db-subnet-group-e34a97d9-8fe6-4239-a4ed-ad6e73a956b0", - "title": "To change DB subnet group settings" + "description":"This example changes the specified setting for the specified DB subnet group.", + "id":"modify-db-subnet-group-e34a97d9-8fe6-4239-a4ed-ad6e73a956b0", + "title":"To change DB subnet group settings" } ], - "ModifyEventSubscription": [ + "ModifyEventSubscription":[ { - "input": { - "Enabled": true, - "EventCategories": [ + "input":{ + "Enabled":true, + "EventCategories":[ "deletion", "low storage" ], - "SourceType": "db-instance", - "SubscriptionName": "mymysqleventsubscription" + "SourceType":"db-instance", + "SubscriptionName":"mymysqleventsubscription" }, - "output": { - "EventSubscription": { + "output":{ + "EventSubscription":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example changes the specified setting for the specified event notification subscription.", - "id": "modify-event-subscription-405ac869-1f02-42cd-b8f4-6950a435f30e", - "title": "To change event notification subscription settings" + "description":"This example changes the specified setting for the specified event notification subscription.", + "id":"modify-event-subscription-405ac869-1f02-42cd-b8f4-6950a435f30e", + "title":"To change event notification subscription settings" } ], - "ModifyOptionGroup": [ + "ModifyOptionGroup":[ { - "input": { - "ApplyImmediately": true, - "OptionGroupName": "myawsuser-og02", - "OptionsToInclude": [ + "input":{ + "ApplyImmediately":true, + "OptionGroupName":"myawsuser-og02", + "OptionsToInclude":[ { - "DBSecurityGroupMemberships": [ + "DBSecurityGroupMemberships":[ "default" ], - "OptionName": "MEMCACHED" + "OptionName":"MEMCACHED" } ] }, - "output": { - "OptionGroup": { + "output":{ + "OptionGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example adds an option to an option group.", - "id": "to-modify-an-option-group-1473890247875", - "title": "To modify an option group" + "description":"The following example adds an option to an option group.", + "id":"to-modify-an-option-group-1473890247875", + "title":"To modify an option group" } ], - "PromoteReadReplica": [ + "PromoteReadReplica":[ { - "input": { - "BackupRetentionPeriod": 1, - "DBInstanceIdentifier": "mydbreadreplica", - "PreferredBackupWindow": "03:30-04:00" + "input":{ + "BackupRetentionPeriod":1, + "DBInstanceIdentifier":"mydbreadreplica", + "PreferredBackupWindow":"03:30-04:00" }, - "output": { - "DBInstance": { + "output":{ + "DBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example promotes the specified read replica and sets its backup retention period and preferred backup window.", - "id": "promote-read-replica-cc580039-c55d-4035-838a-def4a1ae4181", - "title": "To promote a read replica" + "description":"This example promotes the specified read replica and sets its backup retention period and preferred backup window.", + "id":"promote-read-replica-cc580039-c55d-4035-838a-def4a1ae4181", + "title":"To promote a read replica" } ], - "PurchaseReservedDBInstancesOffering": [ + "PurchaseReservedDBInstancesOffering":[ { - "input": { - "ReservedDBInstanceId": "myreservationid", - "ReservedDBInstancesOfferingId": "fb29428a-646d-4390-850e-5fe89926e727" + "input":{ + "ReservedDBInstanceId":"myreservationid", + "ReservedDBInstancesOfferingId":"fb29428a-646d-4390-850e-5fe89926e727" }, - "output": { - "ReservedDBInstance": { + "output":{ + "ReservedDBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example purchases a reserved DB instance offering that matches the specified settings.", - "id": "purchase-reserved-db-instances-offfering-f423c736-8413-429b-ba13-850fd4fa4dcd", - "title": "To purchase a reserved DB instance offering" + "description":"This example purchases a reserved DB instance offering that matches the specified settings.", + "id":"purchase-reserved-db-instances-offfering-f423c736-8413-429b-ba13-850fd4fa4dcd", + "title":"To purchase a reserved DB instance offering" } ], - "RebootDBInstance": [ + "RebootDBInstance":[ { - "input": { - "DBInstanceIdentifier": "mymysqlinstance", - "ForceFailover": false + "input":{ + "DBInstanceIdentifier":"mymysqlinstance", + "ForceFailover":false }, - "output": { - "DBInstance": { + "output":{ + "DBInstance":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example reboots the specified DB instance without forcing a failover.", - "id": "reboot-db-instance-b9ce8a0a-2920-451d-a1f3-01d288aa7366", - "title": "To reboot a DB instance" + "description":"This example reboots the specified DB instance without forcing a failover.", + "id":"reboot-db-instance-b9ce8a0a-2920-451d-a1f3-01d288aa7366", + "title":"To reboot a DB instance" } ], - "RemoveSourceIdentifierFromSubscription": [ + "RemoveSourceIdentifierFromSubscription":[ { - "input": { - "SourceIdentifier": "mymysqlinstance", - "SubscriptionName": "myeventsubscription" + "input":{ + "SourceIdentifier":"mymysqlinstance", + "SubscriptionName":"myeventsubscription" }, - "output": { - "EventSubscription": { + "output":{ + "EventSubscription":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example removes the specified source identifier from the specified DB event subscription.", - "id": "remove-source-identifier-from-subscription-30d25493-c19d-4cf7-b4e5-68371d0d8770", - "title": "To remove a source identifier from a DB event subscription" + "description":"This example removes the specified source identifier from the specified DB event subscription.", + "id":"remove-source-identifier-from-subscription-30d25493-c19d-4cf7-b4e5-68371d0d8770", + "title":"To remove a source identifier from a DB event subscription" } ], - "RemoveTagsFromResource": [ + "RemoveTagsFromResource":[ { - "input": { - "ResourceName": "arn:aws:rds:us-east-1:992648334831:og:mydboptiongroup", - "TagKeys": [ + "input":{ + "ResourceName":"arn:aws:rds:us-east-1:992648334831:og:mydboptiongroup", + "TagKeys":[ "MyKey" ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example removes the specified tag associated with the specified DB option group.", - "id": "remove-tags-from-resource-49f00574-38f6-4d01-ac89-d3c668449ce3", - "title": "To remove tags from a resource" + "description":"This example removes the specified tag associated with the specified DB option group.", + "id":"remove-tags-from-resource-49f00574-38f6-4d01-ac89-d3c668449ce3", + "title":"To remove tags from a resource" } ], - "ResetDBClusterParameterGroup": [ + "ResetDBClusterParameterGroup":[ { - "input": { - "DBClusterParameterGroupName": "mydbclusterparametergroup", - "ResetAllParameters": true + "input":{ + "DBClusterParameterGroupName":"mydbclusterparametergroup", + "ResetAllParameters":true }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example resets all parameters for the specified DB cluster parameter group to their default values.", - "id": "reset-db-cluster-parameter-group-b04aeaf7-7f73-49e1-9bb4-857573ea3ee4", - "title": "To reset the values of a DB cluster parameter group" + "description":"This example resets all parameters for the specified DB cluster parameter group to their default values.", + "id":"reset-db-cluster-parameter-group-b04aeaf7-7f73-49e1-9bb4-857573ea3ee4", + "title":"To reset the values of a DB cluster parameter group" } ], - "ResetDBParameterGroup": [ + "ResetDBParameterGroup":[ { - "input": { - "DBParameterGroupName": "mydbparametergroup", - "ResetAllParameters": true + "input":{ + "DBParameterGroupName":"mydbparametergroup", + "ResetAllParameters":true }, - "output": { + "output":{ }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example resets all parameters for the specified DB parameter group to their default values.", - "id": "reset-db-parameter-group-ed2ed723-de0d-4824-8af5-3c65fa130abf", - "title": "To reset the values of a DB parameter group" + "description":"This example resets all parameters for the specified DB parameter group to their default values.", + "id":"reset-db-parameter-group-ed2ed723-de0d-4824-8af5-3c65fa130abf", + "title":"To reset the values of a DB parameter group" } ], - "RestoreDBClusterFromSnapshot": [ + "RestoreDBClusterFromSnapshot":[ { - "input": { - "DBClusterIdentifier": "restored-cluster1", - "Engine": "aurora", - "SnapshotIdentifier": "sample-cluster-snapshot1" + "input":{ + "DBClusterIdentifier":"restored-cluster1", + "Engine":"aurora", + "SnapshotIdentifier":"sample-cluster-snapshot1" }, - "output": { - "DBCluster": { + "output":{ + "DBCluster":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example restores an Amazon Aurora DB cluster from a DB cluster snapshot.", - "id": "to-restore-an-amazon-aurora-db-cluster-from-a-db-cluster-snapshot-1473958144325", - "title": "To restore an Amazon Aurora DB cluster from a DB cluster snapshot" + "description":"The following example restores an Amazon Aurora DB cluster from a DB cluster snapshot.", + "id":"to-restore-an-amazon-aurora-db-cluster-from-a-db-cluster-snapshot-1473958144325", + "title":"To restore an Amazon Aurora DB cluster from a DB cluster snapshot" } ], - "RestoreDBClusterToPointInTime": [ + "RestoreDBClusterToPointInTime":[ { - "input": { - "DBClusterIdentifier": "sample-restored-cluster1", - "RestoreToTime": "2016-09-13T18:45:00Z", - "SourceDBClusterIdentifier": "sample-cluster1" + "input":{ + "DBClusterIdentifier":"sample-restored-cluster1", + "RestoreToTime":"2016-09-13T18:45:00Z", + "SourceDBClusterIdentifier":"sample-cluster1" }, - "output": { - "DBCluster": { + "output":{ + "DBCluster":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example restores a DB cluster to a new DB cluster at a point in time from the source DB cluster.", - "id": "to-restore-a-db-cluster-to-a-point-in-time-1473962082214", - "title": "To restore a DB cluster to a point in time." + "description":"The following example restores a DB cluster to a new DB cluster at a point in time from the source DB cluster.", + "id":"to-restore-a-db-cluster-to-a-point-in-time-1473962082214", + "title":"To restore a DB cluster to a point in time." } ], - "RestoreDBInstanceFromDBSnapshot": [ + "RestoreDBInstanceFromDBSnapshot":[ { - "input": { - "DBInstanceIdentifier": "mysqldb-restored", - "DBSnapshotIdentifier": "rds:mysqldb-2014-04-22-08-15" + "input":{ + "DBInstanceIdentifier":"mysqldb-restored", + "DBSnapshotIdentifier":"rds:mysqldb-2014-04-22-08-15" }, - "output": { - "DBInstance": { - "AllocatedStorage": 200, - "AutoMinorVersionUpgrade": true, - "AvailabilityZone": "us-west-2b", - "BackupRetentionPeriod": 7, - "CACertificateIdentifier": "rds-ca-2015", - "CopyTagsToSnapshot": false, - "DBInstanceArn": "arn:aws:rds:us-west-2:123456789012:db:mysqldb-restored", - "DBInstanceClass": "db.t2.small", - "DBInstanceIdentifier": "mysqldb-restored", - "DBInstanceStatus": "available", - "DBName": "sample", - "DBParameterGroups": [ + "output":{ + "DBInstance":{ + "AllocatedStorage":200, + "AutoMinorVersionUpgrade":true, + "AvailabilityZone":"us-west-2b", + "BackupRetentionPeriod":7, + "CACertificateIdentifier":"rds-ca-2015", + "CopyTagsToSnapshot":false, + "DBInstanceArn":"arn:aws:rds:us-west-2:123456789012:db:mysqldb-restored", + "DBInstanceClass":"db.t2.small", + "DBInstanceIdentifier":"mysqldb-restored", + "DBInstanceStatus":"available", + "DBName":"sample", + "DBParameterGroups":[ { - "DBParameterGroupName": "default.mysql5.6", - "ParameterApplyStatus": "in-sync" + "DBParameterGroupName":"default.mysql5.6", + "ParameterApplyStatus":"in-sync" } ], - "DBSecurityGroups": [ + "DBSecurityGroups":[ ], - "DBSubnetGroup": { - "DBSubnetGroupDescription": "default", - "DBSubnetGroupName": "default", - "SubnetGroupStatus": "Complete", - "Subnets": [ + "DBSubnetGroup":{ + "DBSubnetGroupDescription":"default", + "DBSubnetGroupName":"default", + "SubnetGroupStatus":"Complete", + "Subnets":[ { - "SubnetAvailabilityZone": { - "Name": "us-west-2a" + "SubnetAvailabilityZone":{ + "Name":"us-west-2a" }, - "SubnetIdentifier": "subnet-77e8db03", - "SubnetStatus": "Active" + "SubnetIdentifier":"subnet-77e8db03", + "SubnetStatus":"Active" }, { - "SubnetAvailabilityZone": { - "Name": "us-west-2b" + "SubnetAvailabilityZone":{ + "Name":"us-west-2b" }, - "SubnetIdentifier": "subnet-c39989a1", - "SubnetStatus": "Active" + "SubnetIdentifier":"subnet-c39989a1", + "SubnetStatus":"Active" }, { - "SubnetAvailabilityZone": { - "Name": "us-west-2c" + "SubnetAvailabilityZone":{ + "Name":"us-west-2c" }, - "SubnetIdentifier": "subnet-4b267b0d", - "SubnetStatus": "Active" + "SubnetIdentifier":"subnet-4b267b0d", + "SubnetStatus":"Active" } ], - "VpcId": "vpc-c1c5b3a3" + "VpcId":"vpc-c1c5b3a3" }, - "DbInstancePort": 0, - "DbiResourceId": "db-VNZUCCBTEDC4WR7THXNJO72HVQ", - "DomainMemberships": [ + "DbInstancePort":0, + "DbiResourceId":"db-VNZUCCBTEDC4WR7THXNJO72HVQ", + "DomainMemberships":[ ], - "Engine": "mysql", - "EngineVersion": "5.6.27", - "LicenseModel": "general-public-license", - "MasterUsername": "mymasteruser", - "MonitoringInterval": 0, - "MultiAZ": false, - "OptionGroupMemberships": [ + "Engine":"mysql", + "EngineVersion":"5.6.27", + "LicenseModel":"general-public-license", + "MasterUsername":"mymasteruser", + "MonitoringInterval":0, + "MultiAZ":false, + "OptionGroupMemberships":[ { - "OptionGroupName": "default:mysql-5-6", - "Status": "in-sync" + "OptionGroupName":"default:mysql-5-6", + "Status":"in-sync" } ], - "PendingModifiedValues": { + "PendingModifiedValues":{ }, - "PreferredBackupWindow": "12:58-13:28", - "PreferredMaintenanceWindow": "tue:10:16-tue:10:46", - "PubliclyAccessible": true, - "ReadReplicaDBInstanceIdentifiers": [ + "PreferredBackupWindow":"12:58-13:28", + "PreferredMaintenanceWindow":"tue:10:16-tue:10:46", + "PubliclyAccessible":true, + "ReadReplicaDBInstanceIdentifiers":[ ], - "StorageEncrypted": false, - "StorageType": "gp2", - "VpcSecurityGroups": [ + "StorageEncrypted":false, + "StorageType":"gp2", + "VpcSecurityGroups":[ { - "Status": "active", - "VpcSecurityGroupId": "sg-e5e5b0d2" + "Status":"active", + "VpcSecurityGroupId":"sg-e5e5b0d2" } ] } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example restores a DB instance from a DB snapshot.", - "id": "to-restore-a-db-instance-from-a-db-snapshot-1473961657311", - "title": "To restore a DB instance from a DB snapshot." + "description":"The following example restores a DB instance from a DB snapshot.", + "id":"to-restore-a-db-instance-from-a-db-snapshot-1473961657311", + "title":"To restore a DB instance from a DB snapshot." } ], - "RestoreDBInstanceToPointInTime": [ + "RestoreDBInstanceToPointInTime":[ { - "input": { - "RestoreTime": "2016-09-13T18:45:00Z", - "SourceDBInstanceIdentifier": "mysql-sample", - "TargetDBInstanceIdentifier": "mysql-sample-restored" + "input":{ + "RestoreTime":"2016-09-13T18:45:00Z", + "SourceDBInstanceIdentifier":"mysql-sample", + "TargetDBInstanceIdentifier":"mysql-sample-restored" }, - "output": { - "DBInstance": { - "AllocatedStorage": 200, - "AutoMinorVersionUpgrade": true, - "AvailabilityZone": "us-west-2b", - "BackupRetentionPeriod": 7, - "CACertificateIdentifier": "rds-ca-2015", - "CopyTagsToSnapshot": false, - "DBInstanceArn": "arn:aws:rds:us-west-2:123456789012:db:mysql-sample-restored", - "DBInstanceClass": "db.t2.small", - "DBInstanceIdentifier": "mysql-sample-restored", - "DBInstanceStatus": "available", - "DBName": "sample", - "DBParameterGroups": [ + "output":{ + "DBInstance":{ + "AllocatedStorage":200, + "AutoMinorVersionUpgrade":true, + "AvailabilityZone":"us-west-2b", + "BackupRetentionPeriod":7, + "CACertificateIdentifier":"rds-ca-2015", + "CopyTagsToSnapshot":false, + "DBInstanceArn":"arn:aws:rds:us-west-2:123456789012:db:mysql-sample-restored", + "DBInstanceClass":"db.t2.small", + "DBInstanceIdentifier":"mysql-sample-restored", + "DBInstanceStatus":"available", + "DBName":"sample", + "DBParameterGroups":[ { - "DBParameterGroupName": "default.mysql5.6", - "ParameterApplyStatus": "in-sync" + "DBParameterGroupName":"default.mysql5.6", + "ParameterApplyStatus":"in-sync" } ], - "DBSecurityGroups": [ + "DBSecurityGroups":[ ], - "DBSubnetGroup": { - "DBSubnetGroupDescription": "default", - "DBSubnetGroupName": "default", - "SubnetGroupStatus": "Complete", - "Subnets": [ + "DBSubnetGroup":{ + "DBSubnetGroupDescription":"default", + "DBSubnetGroupName":"default", + "SubnetGroupStatus":"Complete", + "Subnets":[ { - "SubnetAvailabilityZone": { - "Name": "us-west-2a" + "SubnetAvailabilityZone":{ + "Name":"us-west-2a" }, - "SubnetIdentifier": "subnet-77e8db03", - "SubnetStatus": "Active" + "SubnetIdentifier":"subnet-77e8db03", + "SubnetStatus":"Active" }, { - "SubnetAvailabilityZone": { - "Name": "us-west-2b" + "SubnetAvailabilityZone":{ + "Name":"us-west-2b" }, - "SubnetIdentifier": "subnet-c39989a1", - "SubnetStatus": "Active" + "SubnetIdentifier":"subnet-c39989a1", + "SubnetStatus":"Active" }, { - "SubnetAvailabilityZone": { - "Name": "us-west-2c" + "SubnetAvailabilityZone":{ + "Name":"us-west-2c" }, - "SubnetIdentifier": "subnet-4b267b0d", - "SubnetStatus": "Active" + "SubnetIdentifier":"subnet-4b267b0d", + "SubnetStatus":"Active" } ], - "VpcId": "vpc-c1c5b3a3" + "VpcId":"vpc-c1c5b3a3" }, - "DbInstancePort": 0, - "DbiResourceId": "db-VNZUCCBTEDC4WR7THXNJO72HVQ", - "DomainMemberships": [ + "DbInstancePort":0, + "DbiResourceId":"db-VNZUCCBTEDC4WR7THXNJO72HVQ", + "DomainMemberships":[ ], - "Engine": "mysql", - "EngineVersion": "5.6.27", - "LicenseModel": "general-public-license", - "MasterUsername": "mymasteruser", - "MonitoringInterval": 0, - "MultiAZ": false, - "OptionGroupMemberships": [ + "Engine":"mysql", + "EngineVersion":"5.6.27", + "LicenseModel":"general-public-license", + "MasterUsername":"mymasteruser", + "MonitoringInterval":0, + "MultiAZ":false, + "OptionGroupMemberships":[ { - "OptionGroupName": "default:mysql-5-6", - "Status": "in-sync" + "OptionGroupName":"default:mysql-5-6", + "Status":"in-sync" } ], - "PendingModifiedValues": { + "PendingModifiedValues":{ }, - "PreferredBackupWindow": "12:58-13:28", - "PreferredMaintenanceWindow": "tue:10:16-tue:10:46", - "PubliclyAccessible": true, - "ReadReplicaDBInstanceIdentifiers": [ + "PreferredBackupWindow":"12:58-13:28", + "PreferredMaintenanceWindow":"tue:10:16-tue:10:46", + "PubliclyAccessible":true, + "ReadReplicaDBInstanceIdentifiers":[ ], - "StorageEncrypted": false, - "StorageType": "gp2", - "VpcSecurityGroups": [ + "StorageEncrypted":false, + "StorageType":"gp2", + "VpcSecurityGroups":[ { - "Status": "active", - "VpcSecurityGroupId": "sg-e5e5b0d2" + "Status":"active", + "VpcSecurityGroupId":"sg-e5e5b0d2" } ] } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example restores a DB instance to a new DB instance at a point in time from the source DB instance.", - "id": "to-restore-a-db-instance-to-a-point-in-time-1473962652154", - "title": "To restore a DB instance to a point in time." + "description":"The following example restores a DB instance to a new DB instance at a point in time from the source DB instance.", + "id":"to-restore-a-db-instance-to-a-point-in-time-1473962652154", + "title":"To restore a DB instance to a point in time." } ], - "RevokeDBSecurityGroupIngress": [ + "RevokeDBSecurityGroupIngress":[ { - "input": { - "CIDRIP": "203.0.113.5/32", - "DBSecurityGroupName": "mydbsecuritygroup" + "input":{ + "CIDRIP":"203.0.113.5/32", + "DBSecurityGroupName":"mydbsecuritygroup" }, - "output": { - "DBSecurityGroup": { + "output":{ + "DBSecurityGroup":{ } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example revokes ingress for the specified CIDR block associated with the specified DB security group.", - "id": "revoke-db-security-group-ingress-ce5b2c1c-bd4e-4809-b04a-6d78ec448813", - "title": "To revoke ingress for a DB security group" + "description":"This example revokes ingress for the specified CIDR block associated with the specified DB security group.", + "id":"revoke-db-security-group-ingress-ce5b2c1c-bd4e-4809-b04a-6d78ec448813", + "title":"To revoke ingress for a DB security group" } ] } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/redshift/2012-12-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/redshift/2012-12-01/docs-2.json index d19a3b58e..119427482 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/redshift/2012-12-01/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/redshift/2012-12-01/docs-2.json @@ -14,7 +14,7 @@ "CreateHsmClientCertificate": "

Creates an HSM client certificate that an Amazon Redshift cluster will use to connect to the client's HSM in order to store and retrieve the keys used to encrypt the cluster databases.

The command returns a public key, which you must store in the HSM. In addition to creating the HSM certificate, you must create an Amazon Redshift HSM configuration that provides a cluster the information needed to store and use encryption keys in the HSM. For more information, go to Hardware Security Modules in the Amazon Redshift Cluster Management Guide.

", "CreateHsmConfiguration": "

Creates an HSM configuration that contains the information required by an Amazon Redshift cluster to store and use database encryption keys in a Hardware Security Module (HSM). After creating the HSM configuration, you can specify it as a parameter when creating a cluster. The cluster will then store its encryption keys in the HSM.

In addition to creating an HSM configuration, you must also create an HSM client certificate. For more information, go to Hardware Security Modules in the Amazon Redshift Cluster Management Guide.

", "CreateSnapshotCopyGrant": "

Creates a snapshot copy grant that permits Amazon Redshift to use a customer master key (CMK) from AWS Key Management Service (AWS KMS) to encrypt copied snapshots in a destination region.

For more information about managing snapshot copy grants, go to Amazon Redshift Database Encryption in the Amazon Redshift Cluster Management Guide.

", - "CreateTags": "

Adds one or more tags to a specified resource.

A resource can have up to 10 tags. If you try to create more than 10 tags for a resource, you will receive an error and the attempt will fail.

If you specify a key that already exists for the resource, the value for that key will be updated with the new value.

", + "CreateTags": "

Adds one or more tags to a specified resource.

A resource can have up to 50 tags. If you try to create more than 50 tags for a resource, you will receive an error and the attempt will fail.

If you specify a key that already exists for the resource, the value for that key will be updated with the new value.

", "DeleteCluster": "

Deletes a previously provisioned cluster. A successful response from the web service indicates that the request was received correctly. Use DescribeClusters to monitor the status of the deletion. The delete operation cannot be canceled or reverted once submitted. For more information about managing clusters, go to Amazon Redshift Clusters in the Amazon Redshift Cluster Management Guide.

If you want to shut down the cluster and retain it for future use, set SkipFinalClusterSnapshot to false and specify a name for FinalClusterSnapshotIdentifier. You can later restore this snapshot to resume using the cluster. If a final cluster snapshot is requested, the status of the cluster will be \"final-snapshot\" while the snapshot is being taken, then it's \"deleting\" once Amazon Redshift begins deleting the cluster.

For more information about managing clusters, go to Amazon Redshift Clusters in the Amazon Redshift Cluster Management Guide.

", "DeleteClusterParameterGroup": "

Deletes a specified Amazon Redshift parameter group.

You cannot delete a parameter group if it is associated with a cluster.

", "DeleteClusterSecurityGroup": "

Deletes an Amazon Redshift security group.

You cannot delete a security group that is associated with any clusters. You cannot delete the default security group.

For information about managing security groups, go to Amazon Redshift Cluster Security Groups in the Amazon Redshift Cluster Management Guide.

", @@ -2181,7 +2181,7 @@ "base": null, "refs": { "CreateClusterMessage$VpcSecurityGroupIds": "

A list of Virtual Private Cloud (VPC) security groups to be associated with the cluster.

Default: The default VPC security group is associated with the cluster.

", - "ModifyClusterMessage$VpcSecurityGroupIds": "

A list of virtual private cloud (VPC) security groups to be associated with the cluster.

", + "ModifyClusterMessage$VpcSecurityGroupIds": "

A list of virtual private cloud (VPC) security groups to be associated with the cluster. This change is asynchronously applied as soon as possible.

", "RestoreFromClusterSnapshotMessage$VpcSecurityGroupIds": "

A list of Virtual Private Cloud (VPC) security groups to be associated with the cluster.

Default: The default VPC security group is associated with the cluster.

VPC security groups only apply to clusters in VPCs.

" } }, diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/docs-2.json index cbefd23a2..2ca436b96 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/docs-2.json @@ -1,1389 +1,1389 @@ { - "version": "2.0", - "service": "

This is the Amazon Rekognition API reference.

", - "operations": { - "CompareFaces": "

Compares a face in the source input image with each of the 100 largest faces detected in the target input image.

If the source image contains multiple faces, the service detects the largest face and compares it with each face detected in the target image.

You pass the input and target images either as base64-encoded image bytes or as a references to images in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

In response, the operation returns an array of face matches ordered by similarity score in descending order. For each face match, the response provides a bounding box of the face, facial landmarks, pose details (pitch, role, and yaw), quality (brightness and sharpness), and confidence value (indicating the level of confidence that the bounding box contains a face). The response also provides a similarity score, which indicates how closely the faces match.

By default, only faces with a similarity score of greater than or equal to 80% are returned in the response. You can change this value by specifying the SimilarityThreshold parameter.

CompareFaces also returns an array of faces that don't match the source image. For each face, it returns a bounding box, confidence value, landmarks, pose details, and quality. The response also returns information about the face in the source image, including the bounding box of the face and confidence value.

If the image doesn't contain Exif metadata, CompareFaces returns orientation information for the source and target images. Use these values to display the images with the correct image orientation.

If no faces are detected in the source or target images, CompareFaces returns an InvalidParameterException error.

This is a stateless API operation. That is, data returned by this operation doesn't persist.

For an example, see faces-compare-images.

This operation requires permissions to perform the rekognition:CompareFaces action.

", - "CreateCollection": "

Creates a collection in an AWS Region. You can add faces to the collection using the operation.

For example, you might create collections, one for each of your application users. A user can then index faces using the IndexFaces operation and persist results in a specific collection. Then, a user can search the collection for faces in the user-specific container.

Collection names are case-sensitive.

This operation requires permissions to perform the rekognition:CreateCollection action.

", - "CreateStreamProcessor": "

Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces in a streaming video.

Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. Rekognition Video sends analysis results to Amazon Kinesis Data Streams.

You provide as input a Kinesis video stream (Input) and a Kinesis data stream (Output) stream. You also specify the face recognition criteria in Settings. For example, the collection containing faces that you want to recognize. Use Name to assign an identifier for the stream processor. You use Name to manage the stream processor. For example, you can start processing the source video by calling with the Name field.

After you have finished analyzing a streaming video, use to stop processing. You can delete the stream processor by calling .

", - "DeleteCollection": "

Deletes the specified collection. Note that this operation removes all faces in the collection. For an example, see delete-collection-procedure.

This operation requires permissions to perform the rekognition:DeleteCollection action.

", - "DeleteFaces": "

Deletes faces from a collection. You specify a collection ID and an array of face IDs to remove from the collection.

This operation requires permissions to perform the rekognition:DeleteFaces action.

", - "DeleteStreamProcessor": "

Deletes the stream processor identified by Name. You assign the value for Name when you create the stream processor with . You might not be able to use the same name for a stream processor for a few seconds after calling DeleteStreamProcessor.

", - "DescribeStreamProcessor": "

Provides information about a stream processor created by . You can get information about the input and output streams, the input parameters for the face recognition being performed, and the current status of the stream processor.

", - "DetectFaces": "

Detects faces within an image that is provided as input.

DetectFaces detects the 100 largest faces in the image. For each face detected, the operation returns face details including a bounding box of the face, a confidence value (that the bounding box contains a face), and a fixed set of attributes such as facial landmarks (for example, coordinates of eye and mouth), gender, presence of beard, sunglasses, etc.

The face-detection algorithm is most effective on frontal faces. For non-frontal or obscured faces, the algorithm may not detect the faces or might detect faces with lower confidence.

You pass the input image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

This is a stateless API operation. That is, the operation does not persist any data.

For an example, see procedure-detecting-faces-in-images.

This operation requires permissions to perform the rekognition:DetectFaces action.

", - "DetectLabels": "

Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature. For an example, see images-s3.

DetectLabels does not support the detection of activities. However, activity detection is supported for label detection in videos. For more information, see .

You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

For each object, scene, and concept the API returns one or more labels. Each label provides the object name, and the level of confidence that the image contains the object. For example, suppose the input image has a lighthouse, the sea, and a rock. The response will include all three labels, one for each object.

{Name: lighthouse, Confidence: 98.4629}

{Name: rock,Confidence: 79.2097}

{Name: sea,Confidence: 75.061}

In the preceding example, the operation returns one label for each of the three objects. The operation can also return multiple labels for the same object in the image. For example, if the input image shows a flower (for example, a tulip), the operation might return the following three labels.

{Name: flower,Confidence: 99.0562}

{Name: plant,Confidence: 99.0562}

{Name: tulip,Confidence: 99.0562}

In this example, the detection algorithm more precisely identifies the flower as a tulip.

In response, the API returns an array of labels. In addition, the response also includes the orientation correction. Optionally, you can specify MinConfidence to control the confidence threshold for the labels returned. The default is 50%. You can also add the MaxLabels parameter to limit the number of labels returned.

If the object detected is a person, the operation doesn't provide the same facial details that the DetectFaces operation provides.

This is a stateless API operation. That is, the operation does not persist any data.

This operation requires permissions to perform the rekognition:DetectLabels action.

", - "DetectModerationLabels": "

Detects explicit or suggestive adult content in a specified JPEG or PNG format image. Use DetectModerationLabels to moderate images depending on your requirements. For example, you might want to filter images that contain nudity, but not images containing suggestive content.

To filter images, use the labels returned by DetectModerationLabels to determine which types of content are appropriate. For information about moderation labels, see moderation.

You pass the input image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

", - "DetectText": "

Detects text in the input image and converts it into machine-readable text.

Pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, you must pass it as a reference to an image in an Amazon S3 bucket. For the AWS CLI, passing image bytes is not supported. The image must be either a .png or .jpeg formatted file.

The DetectText operation returns text in an array of elements, TextDetections. Each TextDetection element provides information about a single word or line of text that was detected in the image.

A word is one or more ISO basic latin script characters that are not separated by spaces. DetectText can detect up to 50 words in an image.

A line is a string of equally spaced words. A line isn't necessarily a complete sentence. For example, a driver's license number is detected as a line. A line ends when there is no aligned text after it. Also, a line ends when there is a large gap between words, relative to the length of the words. This means, depending on the gap between words, Amazon Rekognition may detect multiple lines in text aligned in the same direction. Periods don't represent the end of a line. If a sentence spans multiple lines, the DetectText operation returns multiple lines.

To determine whether a TextDetection element is a line of text or a word, use the TextDetection object Type field.

To be detected, text must be within +/- 30 degrees orientation of the horizontal axis.

For more information, see text-detection.

", - "GetCelebrityInfo": "

Gets the name and additional information about a celebrity based on his or her Rekognition ID. The additional information is returned as an array of URLs. If there is no additional information about the celebrity, this list is empty. For more information, see get-celebrity-info-procedure.

This operation requires permissions to perform the rekognition:GetCelebrityInfo action.

", - "GetCelebrityRecognition": "

Gets the celebrity recognition results for a Rekognition Video analysis started by .

Celebrity recognition in a video is an asynchronous operation. Analysis is started by a call to which returns a job identifier (JobId). When the celebrity recognition operation finishes, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartCelebrityRecognition. To get the results of the celebrity recognition analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetCelebrityDetection and pass the job identifier (JobId) from the initial call to StartCelebrityDetection. For more information, see video.

GetCelebrityRecognition returns detected celebrities and the time(s) they are detected in an array (Celebrities) of objects. Each CelebrityRecognition contains information about the celebrity in a object and the time, Timestamp, the celebrity was detected.

GetCelebrityRecognition only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see .

By default, the Celebrities array is sorted by time (milliseconds from the start of the video). You can also sort the array by celebrity by specifying the value ID in the SortBy input parameter.

The CelebrityDetail object includes the celebrity identifer and additional information urls. If you don't store the additional information urls, you can get them later by calling with the celebrity identifer.

No information is returned for faces not recognized as celebrities.

Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetCelebrityDetection and populate the NextToken request parameter with the token value returned from the previous call to GetCelebrityRecognition.

", - "GetContentModeration": "

Gets the content moderation analysis results for a Rekognition Video analysis started by .

Content moderation analysis of a video is an asynchronous operation. You start analysis by calling . which returns a job identifier (JobId). When analysis finishes, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartContentModeration. To get the results of the content moderation analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetCelebrityDetection and pass the job identifier (JobId) from the initial call to StartCelebrityDetection. For more information, see video.

GetContentModeration returns detected content moderation labels, and the time they are detected, in an array, ModerationLabels, of objects.

By default, the moderated labels are returned sorted by time, in milliseconds from the start of the video. You can also sort them by moderated label by specifying NAME for the SortBy input parameter.

Since video analysis can return a large number of results, use the MaxResults parameter to limit the number of labels returned in a single call to GetContentModeration. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetContentModeration and populate the NextToken request parameter with the value of NextToken returned from the previous call to GetContentModeration.

For more information, see moderation.

", - "GetFaceDetection": "

Gets face detection results for a Rekognition Video analysis started by .

Face detection with Rekognition Video is an asynchronous operation. You start face detection by calling which returns a job identifier (JobId). When the face detection operation finishes, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceDetection. To get the results of the face detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartFaceDetection.

GetFaceDetection returns an array of detected faces (Faces) sorted by the time the faces were detected.

Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetFaceDetection and populate the NextToken request parameter with the token value returned from the previous call to GetFaceDetection.

", - "GetFaceSearch": "

Gets the face search results for Rekognition Video face search started by . The search returns faces in a collection that match the faces of persons detected in a video. It also includes the time(s) that faces are matched in the video.

Face search in a video is an asynchronous operation. You start face search by calling to which returns a job identifier (JobId). When the search operation finishes, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceSearch. To get the search results, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetFaceSearch and pass the job identifier (JobId) from the initial call to StartFaceSearch. For more information, see collections.

The search results are retured in an array, Persons, of objects. EachPersonMatch element contains details about the matching faces in the input collection, person information (facial attributes, bounding boxes, and person identifer) for the matched person, and the time the person was matched in the video.

GetFaceSearch only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see .

By default, the Persons array is sorted by the time, in milliseconds from the start of the video, persons are matched. You can also sort by persons by specifying INDEX for the SORTBY input parameter.

", - "GetLabelDetection": "

Gets the label detection results of a Rekognition Video analysis started by .

The label detection operation is started by a call to which returns a job identifier (JobId). When the label detection operation finishes, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartlabelDetection. To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartLabelDetection.

GetLabelDetection returns an array of detected labels (Labels) sorted by the time the labels were detected. You can also sort by the label name by specifying NAME for the SortBy input parameter.

The labels returned include the label name, the percentage confidence in the accuracy of the detected label, and the time the label was detected in the video.

Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetlabelDetection and populate the NextToken request parameter with the token value returned from the previous call to GetLabelDetection.

", - "GetPersonTracking": "

Gets the person tracking results of a Rekognition Video analysis started by .

The person detection operation is started by a call to StartPersonTracking which returns a job identifier (JobId). When the person detection operation finishes, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartPersonTracking.

To get the results of the person tracking operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartPersonTracking.

GetPersonTracking returns an array, Persons, of tracked persons and the time(s) they were tracked in the video.

GetPersonTracking only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see .

By default, the array is sorted by the time(s) a person is tracked in the video. You can sort by tracked persons by specifying INDEX for the SortBy input parameter.

Use the MaxResults parameter to limit the number of items returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetPersonTracking and populate the NextToken request parameter with the token value returned from the previous call to GetPersonTracking.

", - "IndexFaces": "

Detects faces in the input image and adds them to the specified collection.

Amazon Rekognition does not save the actual faces detected. Instead, the underlying detection algorithm first detects the faces in the input image, and for each face extracts facial features into a feature vector, and stores it in the back-end database. Amazon Rekognition uses feature vectors when performing face match and search operations using the and operations.

If you are using version 1.0 of the face detection model, IndexFaces indexes the 15 largest faces in the input image. Later versions of the face detection model index the 100 largest faces in the input image. To determine which version of the model you are using, check the the value of FaceModelVersion in the response from IndexFaces. For more information, see face-detection-model.

If you provide the optional ExternalImageID for the input image you provided, Amazon Rekognition associates this ID with all faces that it detects. When you call the operation, the response returns the external ID. You can use this external image ID to create a client-side index to associate the faces with each image. You can then use the index to find all faces in an image.

In response, the operation returns an array of metadata for all detected faces. This includes, the bounding box of the detected face, confidence value (indicating the bounding box contains a face), a face ID assigned by the service for each face that is detected and stored, and an image ID assigned by the service for the input image. If you request all facial attributes (using the detectionAttributes parameter, Amazon Rekognition returns detailed facial attributes such as facial landmarks (for example, location of eye and mount) and other facial attributes such gender. If you provide the same image, specify the same collection, and use the same external ID in the IndexFaces operation, Amazon Rekognition doesn't save duplicate face metadata.

The input image is passed either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

This operation requires permissions to perform the rekognition:IndexFaces action.

", - "ListCollections": "

Returns list of collection IDs in your account. If the result is truncated, the response also provides a NextToken that you can use in the subsequent request to fetch the next set of collection IDs.

For an example, see list-collection-procedure.

This operation requires permissions to perform the rekognition:ListCollections action.

", - "ListFaces": "

Returns metadata for faces in the specified collection. This metadata includes information such as the bounding box coordinates, the confidence (that the bounding box contains a face), and face ID. For an example, see list-faces-in-collection-procedure.

This operation requires permissions to perform the rekognition:ListFaces action.

", - "ListStreamProcessors": "

Gets a list of stream processors that you have created with .

", - "RecognizeCelebrities": "

Returns an array of celebrities recognized in the input image. For more information, see celebrities.

RecognizeCelebrities returns the 100 largest faces in the image. It lists recognized celebrities in the CelebrityFaces array and unrecognized faces in the UnrecognizedFaces array. RecognizeCelebrities doesn't return celebrities whose faces are not amongst the largest 100 faces in the image.

For each celebrity recognized, the RecognizeCelebrities returns a Celebrity object. The Celebrity object contains the celebrity name, ID, URL links to additional information, match confidence, and a ComparedFace object that you can use to locate the celebrity's face on the image.

Rekognition does not retain information about which images a celebrity has been recognized in. Your application must store this information and use the Celebrity ID property as a unique identifier for the celebrity. If you don't store the celebrity name or additional information URLs returned by RecognizeCelebrities, you will need the ID to identify the celebrity in a call to the operation.

You pass the imput image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

For an example, see celebrities-procedure-image.

This operation requires permissions to perform the rekognition:RecognizeCelebrities operation.

", - "SearchFaces": "

For a given input face ID, searches for matching faces in the collection the face belongs to. You get a face ID when you add a face to the collection using the IndexFaces operation. The operation compares the features of the input face with faces in the specified collection.

You can also search faces without indexing faces by using the SearchFacesByImage operation.

The operation response returns an array of faces that match, ordered by similarity score with the highest similarity first. More specifically, it is an array of metadata for each face match that is found. Along with the metadata, the response also includes a confidence value for each face match, indicating the confidence that the specific face matches the input face.

For an example, see search-face-with-id-procedure.

This operation requires permissions to perform the rekognition:SearchFaces action.

", - "SearchFacesByImage": "

For a given input image, first detects the largest face in the image, and then searches the specified collection for matching faces. The operation compares the features of the input face with faces in the specified collection.

To search for all faces in an input image, you might first call the operation, and then use the face IDs returned in subsequent calls to the operation.

You can also call the DetectFaces operation and use the bounding boxes in the response to make face crops, which then you can pass in to the SearchFacesByImage operation.

You pass the input image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

The response returns an array of faces that match, ordered by similarity score with the highest similarity first. More specifically, it is an array of metadata for each face match found. Along with the metadata, the response also includes a similarity indicating how similar the face is to the input face. In the response, the operation also returns the bounding box (and a confidence level that the bounding box contains a face) of the face that Amazon Rekognition used for the input image.

For an example, see search-face-with-image-procedure.

This operation requires permissions to perform the rekognition:SearchFacesByImage action.

", - "StartCelebrityRecognition": "

Starts asynchronous recognition of celebrities in a stored video.

Rekognition Video can detect celebrities in a video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartCelebrityRecognition returns a job identifier (JobId) which you use to get the results of the analysis. When celebrity recognition analysis is finished, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the results of the celebrity recognition analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartCelebrityRecognition. For more information, see celebrities.

", - "StartContentModeration": "

Starts asynchronous detection of explicit or suggestive adult content in a stored video.

Rekognition Video can moderate content in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartContentModeration returns a job identifier (JobId) which you use to get the results of the analysis. When content moderation analysis is finished, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel.

To get the results of the content moderation analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartContentModeration. For more information, see moderation.

", - "StartFaceDetection": "

Starts asynchronous detection of faces in a stored video.

Rekognition Video can detect faces in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartFaceDetection returns a job identifier (JobId) that you use to get the results of the operation. When face detection is finished, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartFaceDetection. For more information, see faces-video.

", - "StartFaceSearch": "

Starts the asynchronous search for faces in a collection that match the faces of persons detected in a stored video.

The video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartFaceSearch returns a job identifier (JobId) which you use to get the search results once the search has completed. When searching is finished, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the search results, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartFaceSearch. For more information, see collections-search-person.

", - "StartLabelDetection": "

Starts asynchronous detection of labels in a stored video.

Rekognition Video can detect labels in a video. Labels are instances of real-world entities. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; concepts like landscape, evening, and nature; and activities like a person getting out of a car or a person skiing.

The video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartLabelDetection returns a job identifier (JobId) which you use to get the results of the operation. When label detection is finished, Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel.

To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartLabelDetection.

", - "StartPersonTracking": "

Starts the asynchronous tracking of persons in a stored video.

Rekognition Video can track persons in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartPersonTracking returns a job identifier (JobId) which you use to get the results of the operation. When label detection is finished, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel.

To get the results of the person detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartPersonTracking.

", - "StartStreamProcessor": "

Starts processing a stream processor. You create a stream processor by calling . To tell StartStreamProcessor which stream processor to start, use the value of the Name field specified in the call to CreateStreamProcessor.

", - "StopStreamProcessor": "

Stops a running stream processor that was created by .

" + "version":"2.0", + "service":"

This is the Amazon Rekognition API reference.

", + "operations":{ + "CompareFaces":"

Compares a face in the source input image with each of the 100 largest faces detected in the target input image.

If the source image contains multiple faces, the service detects the largest face and compares it with each face detected in the target image.

You pass the input and target images either as base64-encoded image bytes or as a references to images in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

In response, the operation returns an array of face matches ordered by similarity score in descending order. For each face match, the response provides a bounding box of the face, facial landmarks, pose details (pitch, role, and yaw), quality (brightness and sharpness), and confidence value (indicating the level of confidence that the bounding box contains a face). The response also provides a similarity score, which indicates how closely the faces match.

By default, only faces with a similarity score of greater than or equal to 80% are returned in the response. You can change this value by specifying the SimilarityThreshold parameter.

CompareFaces also returns an array of faces that don't match the source image. For each face, it returns a bounding box, confidence value, landmarks, pose details, and quality. The response also returns information about the face in the source image, including the bounding box of the face and confidence value.

If the image doesn't contain Exif metadata, CompareFaces returns orientation information for the source and target images. Use these values to display the images with the correct image orientation.

If no faces are detected in the source or target images, CompareFaces returns an InvalidParameterException error.

This is a stateless API operation. That is, data returned by this operation doesn't persist.

For an example, see Comparing Faces in Images in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:CompareFaces action.

", + "CreateCollection":"

Creates a collection in an AWS Region. You can add faces to the collection using the operation.

For example, you might create collections, one for each of your application users. A user can then index faces using the IndexFaces operation and persist results in a specific collection. Then, a user can search the collection for faces in the user-specific container.

Collection names are case-sensitive.

This operation requires permissions to perform the rekognition:CreateCollection action.

", + "CreateStreamProcessor":"

Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces in a streaming video.

Amazon Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. Amazon Rekognition Video sends analysis results to Amazon Kinesis Data Streams.

You provide as input a Kinesis video stream (Input) and a Kinesis data stream (Output) stream. You also specify the face recognition criteria in Settings. For example, the collection containing faces that you want to recognize. Use Name to assign an identifier for the stream processor. You use Name to manage the stream processor. For example, you can start processing the source video by calling with the Name field.

After you have finished analyzing a streaming video, use to stop processing. You can delete the stream processor by calling .

", + "DeleteCollection":"

Deletes the specified collection. Note that this operation removes all faces in the collection. For an example, see delete-collection-procedure.

This operation requires permissions to perform the rekognition:DeleteCollection action.

", + "DeleteFaces":"

Deletes faces from a collection. You specify a collection ID and an array of face IDs to remove from the collection.

This operation requires permissions to perform the rekognition:DeleteFaces action.

", + "DeleteStreamProcessor":"

Deletes the stream processor identified by Name. You assign the value for Name when you create the stream processor with . You might not be able to use the same name for a stream processor for a few seconds after calling DeleteStreamProcessor.

", + "DescribeStreamProcessor":"

Provides information about a stream processor created by . You can get information about the input and output streams, the input parameters for the face recognition being performed, and the current status of the stream processor.

", + "DetectFaces":"

Detects faces within an image that is provided as input.

DetectFaces detects the 100 largest faces in the image. For each face detected, the operation returns face details including a bounding box of the face, a confidence value (that the bounding box contains a face), and a fixed set of attributes such as facial landmarks (for example, coordinates of eye and mouth), gender, presence of beard, sunglasses, etc.

The face-detection algorithm is most effective on frontal faces. For non-frontal or obscured faces, the algorithm may not detect the faces or might detect faces with lower confidence.

You pass the input image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

This is a stateless API operation. That is, the operation does not persist any data.

This operation requires permissions to perform the rekognition:DetectFaces action.

", + "DetectLabels":"

Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature.

For an example, see Analyzing Images Stored in an Amazon S3 Bucket in the Amazon Rekognition Developer Guide.

DetectLabels does not support the detection of activities. However, activity detection is supported for label detection in videos. For more information, see StartLabelDetection in the Amazon Rekognition Developer Guide.

You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

For each object, scene, and concept the API returns one or more labels. Each label provides the object name, and the level of confidence that the image contains the object. For example, suppose the input image has a lighthouse, the sea, and a rock. The response will include all three labels, one for each object.

{Name: lighthouse, Confidence: 98.4629}

{Name: rock,Confidence: 79.2097}

{Name: sea,Confidence: 75.061}

In the preceding example, the operation returns one label for each of the three objects. The operation can also return multiple labels for the same object in the image. For example, if the input image shows a flower (for example, a tulip), the operation might return the following three labels.

{Name: flower,Confidence: 99.0562}

{Name: plant,Confidence: 99.0562}

{Name: tulip,Confidence: 99.0562}

In this example, the detection algorithm more precisely identifies the flower as a tulip.

In response, the API returns an array of labels. In addition, the response also includes the orientation correction. Optionally, you can specify MinConfidence to control the confidence threshold for the labels returned. The default is 50%. You can also add the MaxLabels parameter to limit the number of labels returned.

If the object detected is a person, the operation doesn't provide the same facial details that the DetectFaces operation provides.

This is a stateless API operation. That is, the operation does not persist any data.

This operation requires permissions to perform the rekognition:DetectLabels action.

", + "DetectModerationLabels":"

Detects explicit or suggestive adult content in a specified JPEG or PNG format image. Use DetectModerationLabels to moderate images depending on your requirements. For example, you might want to filter images that contain nudity, but not images containing suggestive content.

To filter images, use the labels returned by DetectModerationLabels to determine which types of content are appropriate.

For information about moderation labels, see Detecting Unsafe Content in the Amazon Rekognition Developer Guide.

You pass the input image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

", + "DetectText":"

Detects text in the input image and converts it into machine-readable text.

Pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, you must pass it as a reference to an image in an Amazon S3 bucket. For the AWS CLI, passing image bytes is not supported. The image must be either a .png or .jpeg formatted file.

The DetectText operation returns text in an array of elements, TextDetections. Each TextDetection element provides information about a single word or line of text that was detected in the image.

A word is one or more ISO basic latin script characters that are not separated by spaces. DetectText can detect up to 50 words in an image.

A line is a string of equally spaced words. A line isn't necessarily a complete sentence. For example, a driver's license number is detected as a line. A line ends when there is no aligned text after it. Also, a line ends when there is a large gap between words, relative to the length of the words. This means, depending on the gap between words, Amazon Rekognition may detect multiple lines in text aligned in the same direction. Periods don't represent the end of a line. If a sentence spans multiple lines, the DetectText operation returns multiple lines.

To determine whether a TextDetection element is a line of text or a word, use the TextDetection object Type field.

To be detected, text must be within +/- 30 degrees orientation of the horizontal axis.

For more information, see DetectText in the Amazon Rekognition Developer Guide.

", + "GetCelebrityInfo":"

Gets the name and additional information about a celebrity based on his or her Rekognition ID. The additional information is returned as an array of URLs. If there is no additional information about the celebrity, this list is empty.

For more information, see Recognizing Celebrities in an Image in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:GetCelebrityInfo action.

", + "GetCelebrityRecognition":"

Gets the celebrity recognition results for a Amazon Rekognition Video analysis started by .

Celebrity recognition in a video is an asynchronous operation. Analysis is started by a call to which returns a job identifier (JobId). When the celebrity recognition operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartCelebrityRecognition. To get the results of the celebrity recognition analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetCelebrityDetection and pass the job identifier (JobId) from the initial call to StartCelebrityDetection.

For more information, see Working With Stored Videos in the Amazon Rekognition Developer Guide.

GetCelebrityRecognition returns detected celebrities and the time(s) they are detected in an array (Celebrities) of objects. Each CelebrityRecognition contains information about the celebrity in a object and the time, Timestamp, the celebrity was detected.

GetCelebrityRecognition only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see FaceDetail in the Amazon Rekognition Developer Guide.

By default, the Celebrities array is sorted by time (milliseconds from the start of the video). You can also sort the array by celebrity by specifying the value ID in the SortBy input parameter.

The CelebrityDetail object includes the celebrity identifer and additional information urls. If you don't store the additional information urls, you can get them later by calling with the celebrity identifer.

No information is returned for faces not recognized as celebrities.

Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetCelebrityDetection and populate the NextToken request parameter with the token value returned from the previous call to GetCelebrityRecognition.

", + "GetContentModeration":"

Gets the content moderation analysis results for a Amazon Rekognition Video analysis started by .

Content moderation analysis of a video is an asynchronous operation. You start analysis by calling . which returns a job identifier (JobId). When analysis finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartContentModeration. To get the results of the content moderation analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetCelebrityDetection and pass the job identifier (JobId) from the initial call to StartCelebrityDetection.

For more information, see Working with Stored Videos in the Amazon Rekognition Devlopers Guide.

GetContentModeration returns detected content moderation labels, and the time they are detected, in an array, ModerationLabels, of objects.

By default, the moderated labels are returned sorted by time, in milliseconds from the start of the video. You can also sort them by moderated label by specifying NAME for the SortBy input parameter.

Since video analysis can return a large number of results, use the MaxResults parameter to limit the number of labels returned in a single call to GetContentModeration. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetContentModeration and populate the NextToken request parameter with the value of NextToken returned from the previous call to GetContentModeration.

For more information, see Detecting Unsafe Content in the Amazon Rekognition Developer Guide.

", + "GetFaceDetection":"

Gets face detection results for a Amazon Rekognition Video analysis started by .

Face detection with Amazon Rekognition Video is an asynchronous operation. You start face detection by calling which returns a job identifier (JobId). When the face detection operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceDetection. To get the results of the face detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartFaceDetection.

GetFaceDetection returns an array of detected faces (Faces) sorted by the time the faces were detected.

Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetFaceDetection and populate the NextToken request parameter with the token value returned from the previous call to GetFaceDetection.

", + "GetFaceSearch":"

Gets the face search results for Amazon Rekognition Video face search started by . The search returns faces in a collection that match the faces of persons detected in a video. It also includes the time(s) that faces are matched in the video.

Face search in a video is an asynchronous operation. You start face search by calling to which returns a job identifier (JobId). When the search operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceSearch. To get the search results, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetFaceSearch and pass the job identifier (JobId) from the initial call to StartFaceSearch.

For more information, see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.

The search results are retured in an array, Persons, of objects. EachPersonMatch element contains details about the matching faces in the input collection, person information (facial attributes, bounding boxes, and person identifer) for the matched person, and the time the person was matched in the video.

GetFaceSearch only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see FaceDetail in the Amazon Rekognition Developer Guide.

By default, the Persons array is sorted by the time, in milliseconds from the start of the video, persons are matched. You can also sort by persons by specifying INDEX for the SORTBY input parameter.

", + "GetLabelDetection":"

Gets the label detection results of a Amazon Rekognition Video analysis started by .

The label detection operation is started by a call to which returns a job identifier (JobId). When the label detection operation finishes, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartlabelDetection. To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartLabelDetection.

GetLabelDetection returns an array of detected labels (Labels) sorted by the time the labels were detected. You can also sort by the label name by specifying NAME for the SortBy input parameter.

The labels returned include the label name, the percentage confidence in the accuracy of the detected label, and the time the label was detected in the video.

Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetlabelDetection and populate the NextToken request parameter with the token value returned from the previous call to GetLabelDetection.

", + "GetPersonTracking":"

Gets the person tracking results of a Amazon Rekognition Video analysis started by .

The person detection operation is started by a call to StartPersonTracking which returns a job identifier (JobId). When the person detection operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartPersonTracking.

To get the results of the person tracking operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartPersonTracking.

GetPersonTracking returns an array, Persons, of tracked persons and the time(s) they were tracked in the video.

GetPersonTracking only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned.

For more information, see FaceDetail in the Amazon Rekognition Developer Guide.

By default, the array is sorted by the time(s) a person is tracked in the video. You can sort by tracked persons by specifying INDEX for the SortBy input parameter.

Use the MaxResults parameter to limit the number of items returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetPersonTracking and populate the NextToken request parameter with the token value returned from the previous call to GetPersonTracking.

", + "IndexFaces":"

Detects faces in the input image and adds them to the specified collection.

Amazon Rekognition does not save the actual faces detected. Instead, the underlying detection algorithm first detects the faces in the input image, and for each face extracts facial features into a feature vector, and stores it in the back-end database. Amazon Rekognition uses feature vectors when performing face match and search operations using the and operations.

If you are using version 1.0 of the face detection model, IndexFaces indexes the 15 largest faces in the input image. Later versions of the face detection model index the 100 largest faces in the input image. To determine which version of the model you are using, check the the value of FaceModelVersion in the response from IndexFaces.

For more information, see Model Versioning in the Amazon Rekognition Developer Guide.

If you provide the optional ExternalImageID for the input image you provided, Amazon Rekognition associates this ID with all faces that it detects. When you call the operation, the response returns the external ID. You can use this external image ID to create a client-side index to associate the faces with each image. You can then use the index to find all faces in an image.

In response, the operation returns an array of metadata for all detected faces. This includes, the bounding box of the detected face, confidence value (indicating the bounding box contains a face), a face ID assigned by the service for each face that is detected and stored, and an image ID assigned by the service for the input image. If you request all facial attributes (using the detectionAttributes parameter, Amazon Rekognition returns detailed facial attributes such as facial landmarks (for example, location of eye and mount) and other facial attributes such gender. If you provide the same image, specify the same collection, and use the same external ID in the IndexFaces operation, Amazon Rekognition doesn't save duplicate face metadata.

For more information, see Adding Faces to a Collection in the Amazon Rekognition Developer Guide.

The input image is passed either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

This operation requires permissions to perform the rekognition:IndexFaces action.

", + "ListCollections":"

Returns list of collection IDs in your account. If the result is truncated, the response also provides a NextToken that you can use in the subsequent request to fetch the next set of collection IDs.

For an example, see Listing Collections in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:ListCollections action.

", + "ListFaces":"

Returns metadata for faces in the specified collection. This metadata includes information such as the bounding box coordinates, the confidence (that the bounding box contains a face), and face ID. For an example, see Listing Faces in a Collection in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:ListFaces action.

", + "ListStreamProcessors":"

Gets a list of stream processors that you have created with .

", + "RecognizeCelebrities":"

Returns an array of celebrities recognized in the input image. For more information, see Recognizing Celebrities in the Amazon Rekognition Developer Guide.

RecognizeCelebrities returns the 100 largest faces in the image. It lists recognized celebrities in the CelebrityFaces array and unrecognized faces in the UnrecognizedFaces array. RecognizeCelebrities doesn't return celebrities whose faces are not amongst the largest 100 faces in the image.

For each celebrity recognized, the RecognizeCelebrities returns a Celebrity object. The Celebrity object contains the celebrity name, ID, URL links to additional information, match confidence, and a ComparedFace object that you can use to locate the celebrity's face on the image.

Rekognition does not retain information about which images a celebrity has been recognized in. Your application must store this information and use the Celebrity ID property as a unique identifier for the celebrity. If you don't store the celebrity name or additional information URLs returned by RecognizeCelebrities, you will need the ID to identify the celebrity in a call to the operation.

You pass the imput image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

For an example, see Recognizing Celebrities in an Image in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:RecognizeCelebrities operation.

", + "SearchFaces":"

For a given input face ID, searches for matching faces in the collection the face belongs to. You get a face ID when you add a face to the collection using the IndexFaces operation. The operation compares the features of the input face with faces in the specified collection.

You can also search faces without indexing faces by using the SearchFacesByImage operation.

The operation response returns an array of faces that match, ordered by similarity score with the highest similarity first. More specifically, it is an array of metadata for each face match that is found. Along with the metadata, the response also includes a confidence value for each face match, indicating the confidence that the specific face matches the input face.

For an example, see Searching for a Face Using Its Face ID in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:SearchFaces action.

", + "SearchFacesByImage":"

For a given input image, first detects the largest face in the image, and then searches the specified collection for matching faces. The operation compares the features of the input face with faces in the specified collection.

To search for all faces in an input image, you might first call the operation, and then use the face IDs returned in subsequent calls to the operation.

You can also call the DetectFaces operation and use the bounding boxes in the response to make face crops, which then you can pass in to the SearchFacesByImage operation.

You pass the input image either as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.

The response returns an array of faces that match, ordered by similarity score with the highest similarity first. More specifically, it is an array of metadata for each face match found. Along with the metadata, the response also includes a similarity indicating how similar the face is to the input face. In the response, the operation also returns the bounding box (and a confidence level that the bounding box contains a face) of the face that Amazon Rekognition used for the input image.

For an example, Searching for a Face Using an Image in the Amazon Rekognition Developer Guide.

This operation requires permissions to perform the rekognition:SearchFacesByImage action.

", + "StartCelebrityRecognition":"

Starts asynchronous recognition of celebrities in a stored video.

Amazon Rekognition Video can detect celebrities in a video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartCelebrityRecognition returns a job identifier (JobId) which you use to get the results of the analysis. When celebrity recognition analysis is finished, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the results of the celebrity recognition analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartCelebrityRecognition.

For more information, see Recognizing Celebrities in the Amazon Rekognition Developer Guide.

", + "StartContentModeration":"

Starts asynchronous detection of explicit or suggestive adult content in a stored video.

Amazon Rekognition Video can moderate content in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartContentModeration returns a job identifier (JobId) which you use to get the results of the analysis. When content moderation analysis is finished, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel.

To get the results of the content moderation analysis, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartContentModeration.

For more information, see Detecting Unsafe Content in the Amazon Rekognition Developer Guide.

", + "StartFaceDetection":"

Starts asynchronous detection of faces in a stored video.

Amazon Rekognition Video can detect faces in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartFaceDetection returns a job identifier (JobId) that you use to get the results of the operation. When face detection is finished, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartFaceDetection.

For more information, see Detecting Faces in a Stored Video in the Amazon Rekognition Developer Guide.

", + "StartFaceSearch":"

Starts the asynchronous search for faces in a collection that match the faces of persons detected in a stored video.

The video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartFaceSearch returns a job identifier (JobId) which you use to get the search results once the search has completed. When searching is finished, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the search results, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartFaceSearch. For more information, see collections-search-person.

", + "StartLabelDetection":"

Starts asynchronous detection of labels in a stored video.

Amazon Rekognition Video can detect labels in a video. Labels are instances of real-world entities. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; concepts like landscape, evening, and nature; and activities like a person getting out of a car or a person skiing.

The video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartLabelDetection returns a job identifier (JobId) which you use to get the results of the operation. When label detection is finished, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel.

To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartLabelDetection.

", + "StartPersonTracking":"

Starts the asynchronous tracking of persons in a stored video.

Amazon Rekognition Video can track persons in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartPersonTracking returns a job identifier (JobId) which you use to get the results of the operation. When label detection is finished, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel.

To get the results of the person detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from the initial call to StartPersonTracking.

", + "StartStreamProcessor":"

Starts processing a stream processor. You create a stream processor by calling . To tell StartStreamProcessor which stream processor to start, use the value of the Name field specified in the call to CreateStreamProcessor.

", + "StopStreamProcessor":"

Stops a running stream processor that was created by .

" }, - "shapes": { - "AccessDeniedException": { - "base": "

You are not authorized to perform the action.

", - "refs": { + "shapes":{ + "AccessDeniedException":{ + "base":"

You are not authorized to perform the action.

", + "refs":{ } }, - "AgeRange": { - "base": "

Structure containing the estimated age range, in years, for a face.

Rekognition estimates an age-range for faces detected in the input image. Estimated age ranges can overlap; a face of a 5 year old may have an estimated range of 4-6 whilst the face of a 6 year old may have an estimated range of 4-8.

", - "refs": { - "FaceDetail$AgeRange": "

The estimated age range, in years, for the face. Low represents the lowest estimated age and High represents the highest estimated age.

" + "AgeRange":{ + "base":"

Structure containing the estimated age range, in years, for a face.

Rekognition estimates an age-range for faces detected in the input image. Estimated age ranges can overlap; a face of a 5 year old may have an estimated range of 4-6 whilst the face of a 6 year old may have an estimated range of 4-8.

", + "refs":{ + "FaceDetail$AgeRange":"

The estimated age range, in years, for the face. Low represents the lowest estimated age and High represents the highest estimated age.

" } }, - "Attribute": { - "base": null, - "refs": { - "Attributes$member": null + "Attribute":{ + "base":null, + "refs":{ + "Attributes$member":null } }, - "Attributes": { - "base": null, - "refs": { - "DetectFacesRequest$Attributes": "

An array of facial attributes you want to be returned. This can be the default list of attributes or all attributes. If you don't specify a value for Attributes or if you specify [\"DEFAULT\"], the API returns the following subset of facial attributes: BoundingBox, Confidence, Pose, Quality and Landmarks. If you provide [\"ALL\"], all facial attributes are returned but the operation will take longer to complete.

If you provide both, [\"ALL\", \"DEFAULT\"], the service uses a logical AND operator to determine which attributes to return (in this case, all attributes).

", - "IndexFacesRequest$DetectionAttributes": "

An array of facial attributes that you want to be returned. This can be the default list of attributes or all attributes. If you don't specify a value for Attributes or if you specify [\"DEFAULT\"], the API returns the following subset of facial attributes: BoundingBox, Confidence, Pose, Quality and Landmarks. If you provide [\"ALL\"], all facial attributes are returned but the operation will take longer to complete.

If you provide both, [\"ALL\", \"DEFAULT\"], the service uses a logical AND operator to determine which attributes to return (in this case, all attributes).

" + "Attributes":{ + "base":null, + "refs":{ + "DetectFacesRequest$Attributes":"

An array of facial attributes you want to be returned. This can be the default list of attributes or all attributes. If you don't specify a value for Attributes or if you specify [\"DEFAULT\"], the API returns the following subset of facial attributes: BoundingBox, Confidence, Pose, Quality and Landmarks. If you provide [\"ALL\"], all facial attributes are returned but the operation will take longer to complete.

If you provide both, [\"ALL\", \"DEFAULT\"], the service uses a logical AND operator to determine which attributes to return (in this case, all attributes).

", + "IndexFacesRequest$DetectionAttributes":"

An array of facial attributes that you want to be returned. This can be the default list of attributes or all attributes. If you don't specify a value for Attributes or if you specify [\"DEFAULT\"], the API returns the following subset of facial attributes: BoundingBox, Confidence, Pose, Quality and Landmarks. If you provide [\"ALL\"], all facial attributes are returned but the operation will take longer to complete.

If you provide both, [\"ALL\", \"DEFAULT\"], the service uses a logical AND operator to determine which attributes to return (in this case, all attributes).

" } }, - "Beard": { - "base": "

Indicates whether or not the face has a beard, and the confidence level in the determination.

", - "refs": { - "FaceDetail$Beard": "

Indicates whether or not the face has a beard, and the confidence level in the determination.

" + "Beard":{ + "base":"

Indicates whether or not the face has a beard, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$Beard":"

Indicates whether or not the face has a beard, and the confidence level in the determination.

" } }, - "Boolean": { - "base": null, - "refs": { - "Beard$Value": "

Boolean value that indicates whether the face has beard or not.

", - "EyeOpen$Value": "

Boolean value that indicates whether the eyes on the face are open.

", - "Eyeglasses$Value": "

Boolean value that indicates whether the face is wearing eye glasses or not.

", - "MouthOpen$Value": "

Boolean value that indicates whether the mouth on the face is open or not.

", - "Mustache$Value": "

Boolean value that indicates whether the face has mustache or not.

", - "Smile$Value": "

Boolean value that indicates whether the face is smiling or not.

", - "Sunglasses$Value": "

Boolean value that indicates whether the face is wearing sunglasses or not.

" + "Boolean":{ + "base":null, + "refs":{ + "Beard$Value":"

Boolean value that indicates whether the face has beard or not.

", + "EyeOpen$Value":"

Boolean value that indicates whether the eyes on the face are open.

", + "Eyeglasses$Value":"

Boolean value that indicates whether the face is wearing eye glasses or not.

", + "MouthOpen$Value":"

Boolean value that indicates whether the mouth on the face is open or not.

", + "Mustache$Value":"

Boolean value that indicates whether the face has mustache or not.

", + "Smile$Value":"

Boolean value that indicates whether the face is smiling or not.

", + "Sunglasses$Value":"

Boolean value that indicates whether the face is wearing sunglasses or not.

" } }, - "BoundingBox": { - "base": "

Identifies the bounding box around the object, face or text. The left (x-coordinate) and top (y-coordinate) are coordinates representing the top and left sides of the bounding box. Note that the upper-left corner of the image is the origin (0,0).

The top and left values returned are ratios of the overall image size. For example, if the input image is 700x200 pixels, and the top-left coordinate of the bounding box is 350x50 pixels, the API returns a left value of 0.5 (350/700) and a top value of 0.25 (50/200).

The width and height values represent the dimensions of the bounding box as a ratio of the overall image dimension. For example, if the input image is 700x200 pixels, and the bounding box width is 70 pixels, the width returned is 0.1.

The bounding box coordinates can have negative values. For example, if Amazon Rekognition is able to detect a face that is at the image edge and is only partially visible, the service can return coordinates that are outside the image bounds and, depending on the image edge, you might get negative values or values greater than 1 for the left or top values.

", - "refs": { - "CelebrityDetail$BoundingBox": "

Bounding box around the body of a celebrity.

", - "ComparedFace$BoundingBox": "

Bounding box of the face.

", - "ComparedSourceImageFace$BoundingBox": "

Bounding box of the face.

", - "Face$BoundingBox": "

Bounding box of the face.

", - "FaceDetail$BoundingBox": "

Bounding box of the face. Default attribute.

", - "Geometry$BoundingBox": "

An axis-aligned coarse representation of the detected text's location on the image.

", - "PersonDetail$BoundingBox": "

Bounding box around the detected person.

", - "SearchFacesByImageResponse$SearchedFaceBoundingBox": "

The bounding box around the face in the input image that Amazon Rekognition used for the search.

" + "BoundingBox":{ + "base":"

Identifies the bounding box around the object, face or text. The left (x-coordinate) and top (y-coordinate) are coordinates representing the top and left sides of the bounding box. Note that the upper-left corner of the image is the origin (0,0).

The top and left values returned are ratios of the overall image size. For example, if the input image is 700x200 pixels, and the top-left coordinate of the bounding box is 350x50 pixels, the API returns a left value of 0.5 (350/700) and a top value of 0.25 (50/200).

The width and height values represent the dimensions of the bounding box as a ratio of the overall image dimension. For example, if the input image is 700x200 pixels, and the bounding box width is 70 pixels, the width returned is 0.1.

The bounding box coordinates can have negative values. For example, if Amazon Rekognition is able to detect a face that is at the image edge and is only partially visible, the service can return coordinates that are outside the image bounds and, depending on the image edge, you might get negative values or values greater than 1 for the left or top values.

", + "refs":{ + "CelebrityDetail$BoundingBox":"

Bounding box around the body of a celebrity.

", + "ComparedFace$BoundingBox":"

Bounding box of the face.

", + "ComparedSourceImageFace$BoundingBox":"

Bounding box of the face.

", + "Face$BoundingBox":"

Bounding box of the face.

", + "FaceDetail$BoundingBox":"

Bounding box of the face. Default attribute.

", + "Geometry$BoundingBox":"

An axis-aligned coarse representation of the detected text's location on the image.

", + "PersonDetail$BoundingBox":"

Bounding box around the detected person.

", + "SearchFacesByImageResponse$SearchedFaceBoundingBox":"

The bounding box around the face in the input image that Amazon Rekognition used for the search.

" } }, - "Celebrity": { - "base": "

Provides information about a celebrity recognized by the operation.

", - "refs": { - "CelebrityList$member": null + "Celebrity":{ + "base":"

Provides information about a celebrity recognized by the operation.

", + "refs":{ + "CelebrityList$member":null } }, - "CelebrityDetail": { - "base": "

Information about a recognized celebrity.

", - "refs": { - "CelebrityRecognition$Celebrity": "

Information about a recognized celebrity.

" + "CelebrityDetail":{ + "base":"

Information about a recognized celebrity.

", + "refs":{ + "CelebrityRecognition$Celebrity":"

Information about a recognized celebrity.

" } }, - "CelebrityList": { - "base": null, - "refs": { - "RecognizeCelebritiesResponse$CelebrityFaces": "

Details about each celebrity found in the image. Amazon Rekognition can detect a maximum of 15 celebrities in an image.

" + "CelebrityList":{ + "base":null, + "refs":{ + "RecognizeCelebritiesResponse$CelebrityFaces":"

Details about each celebrity found in the image. Amazon Rekognition can detect a maximum of 15 celebrities in an image.

" } }, - "CelebrityRecognition": { - "base": "

Information about a detected celebrity and the time the celebrity was detected in a stored video. For more information, see .

", - "refs": { - "CelebrityRecognitions$member": null + "CelebrityRecognition":{ + "base":"

Information about a detected celebrity and the time the celebrity was detected in a stored video. For more information, see GetCelebrityRecognition in the Amazon Rekognition Developer Guide.

", + "refs":{ + "CelebrityRecognitions$member":null } }, - "CelebrityRecognitionSortBy": { - "base": null, - "refs": { - "GetCelebrityRecognitionRequest$SortBy": "

Sort to use for celebrities returned in Celebrities field. Specify ID to sort by the celebrity identifier, specify TIMESTAMP to sort by the time the celebrity was recognized.

" + "CelebrityRecognitionSortBy":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionRequest$SortBy":"

Sort to use for celebrities returned in Celebrities field. Specify ID to sort by the celebrity identifier, specify TIMESTAMP to sort by the time the celebrity was recognized.

" } }, - "CelebrityRecognitions": { - "base": null, - "refs": { - "GetCelebrityRecognitionResponse$Celebrities": "

Array of celebrities recognized in the video.

" + "CelebrityRecognitions":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionResponse$Celebrities":"

Array of celebrities recognized in the video.

" } }, - "ClientRequestToken": { - "base": null, - "refs": { - "StartCelebrityRecognitionRequest$ClientRequestToken": "

Idempotent token used to identify the start request. If you use the same token with multiple StartCelebrityRecognition requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", - "StartContentModerationRequest$ClientRequestToken": "

Idempotent token used to identify the start request. If you use the same token with multiple StartContentModeration requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", - "StartFaceDetectionRequest$ClientRequestToken": "

Idempotent token used to identify the start request. If you use the same token with multiple StartFaceDetection requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", - "StartFaceSearchRequest$ClientRequestToken": "

Idempotent token used to identify the start request. If you use the same token with multiple StartFaceSearch requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", - "StartLabelDetectionRequest$ClientRequestToken": "

Idempotent token used to identify the start request. If you use the same token with multiple StartLabelDetection requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", - "StartPersonTrackingRequest$ClientRequestToken": "

Idempotent token used to identify the start request. If you use the same token with multiple StartPersonTracking requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

" + "ClientRequestToken":{ + "base":null, + "refs":{ + "StartCelebrityRecognitionRequest$ClientRequestToken":"

Idempotent token used to identify the start request. If you use the same token with multiple StartCelebrityRecognition requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", + "StartContentModerationRequest$ClientRequestToken":"

Idempotent token used to identify the start request. If you use the same token with multiple StartContentModeration requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", + "StartFaceDetectionRequest$ClientRequestToken":"

Idempotent token used to identify the start request. If you use the same token with multiple StartFaceDetection requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", + "StartFaceSearchRequest$ClientRequestToken":"

Idempotent token used to identify the start request. If you use the same token with multiple StartFaceSearch requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", + "StartLabelDetectionRequest$ClientRequestToken":"

Idempotent token used to identify the start request. If you use the same token with multiple StartLabelDetection requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

", + "StartPersonTrackingRequest$ClientRequestToken":"

Idempotent token used to identify the start request. If you use the same token with multiple StartPersonTracking requests, the same JobId is returned. Use ClientRequestToken to prevent the same job from being accidently started more than once.

" } }, - "CollectionId": { - "base": null, - "refs": { - "CollectionIdList$member": null, - "CreateCollectionRequest$CollectionId": "

ID for the collection that you are creating.

", - "DeleteCollectionRequest$CollectionId": "

ID of the collection to delete.

", - "DeleteFacesRequest$CollectionId": "

Collection from which to remove the specific faces.

", - "FaceSearchSettings$CollectionId": "

The ID of a collection that contains faces that you want to search for.

", - "IndexFacesRequest$CollectionId": "

The ID of an existing collection to which you want to add the faces that are detected in the input images.

", - "ListFacesRequest$CollectionId": "

ID of the collection from which to list the faces.

", - "SearchFacesByImageRequest$CollectionId": "

ID of the collection to search.

", - "SearchFacesRequest$CollectionId": "

ID of the collection the face belongs to.

", - "StartFaceSearchRequest$CollectionId": "

ID of the collection that contains the faces you want to search for.

" + "CollectionId":{ + "base":null, + "refs":{ + "CollectionIdList$member":null, + "CreateCollectionRequest$CollectionId":"

ID for the collection that you are creating.

", + "DeleteCollectionRequest$CollectionId":"

ID of the collection to delete.

", + "DeleteFacesRequest$CollectionId":"

Collection from which to remove the specific faces.

", + "FaceSearchSettings$CollectionId":"

The ID of a collection that contains faces that you want to search for.

", + "IndexFacesRequest$CollectionId":"

The ID of an existing collection to which you want to add the faces that are detected in the input images.

", + "ListFacesRequest$CollectionId":"

ID of the collection from which to list the faces.

", + "SearchFacesByImageRequest$CollectionId":"

ID of the collection to search.

", + "SearchFacesRequest$CollectionId":"

ID of the collection the face belongs to.

", + "StartFaceSearchRequest$CollectionId":"

ID of the collection that contains the faces you want to search for.

" } }, - "CollectionIdList": { - "base": null, - "refs": { - "ListCollectionsResponse$CollectionIds": "

An array of collection IDs.

" + "CollectionIdList":{ + "base":null, + "refs":{ + "ListCollectionsResponse$CollectionIds":"

An array of collection IDs.

" } }, - "CompareFacesMatch": { - "base": "

Provides information about a face in a target image that matches the source image face analysed by CompareFaces. The Face property contains the bounding box of the face in the target image. The Similarity property is the confidence that the source image face matches the face in the bounding box.

", - "refs": { - "CompareFacesMatchList$member": null + "CompareFacesMatch":{ + "base":"

Provides information about a face in a target image that matches the source image face analysed by CompareFaces. The Face property contains the bounding box of the face in the target image. The Similarity property is the confidence that the source image face matches the face in the bounding box.

", + "refs":{ + "CompareFacesMatchList$member":null } }, - "CompareFacesMatchList": { - "base": null, - "refs": { - "CompareFacesResponse$FaceMatches": "

An array of faces in the target image that match the source image face. Each CompareFacesMatch object provides the bounding box, the confidence level that the bounding box contains a face, and the similarity score for the face in the bounding box and the face in the source image.

" + "CompareFacesMatchList":{ + "base":null, + "refs":{ + "CompareFacesResponse$FaceMatches":"

An array of faces in the target image that match the source image face. Each CompareFacesMatch object provides the bounding box, the confidence level that the bounding box contains a face, and the similarity score for the face in the bounding box and the face in the source image.

" } }, - "CompareFacesRequest": { - "base": null, - "refs": { + "CompareFacesRequest":{ + "base":null, + "refs":{ } }, - "CompareFacesResponse": { - "base": null, - "refs": { + "CompareFacesResponse":{ + "base":null, + "refs":{ } }, - "CompareFacesUnmatchList": { - "base": null, - "refs": { - "CompareFacesResponse$UnmatchedFaces": "

An array of faces in the target image that did not match the source image face.

" + "CompareFacesUnmatchList":{ + "base":null, + "refs":{ + "CompareFacesResponse$UnmatchedFaces":"

An array of faces in the target image that did not match the source image face.

" } }, - "ComparedFace": { - "base": "

Provides face metadata for target image faces that are analysed by CompareFaces and RecognizeCelebrities.

", - "refs": { - "Celebrity$Face": "

Provides information about the celebrity's face, such as its location on the image.

", - "CompareFacesMatch$Face": "

Provides face metadata (bounding box and confidence that the bounding box actually contains a face).

", - "CompareFacesUnmatchList$member": null, - "ComparedFaceList$member": null + "ComparedFace":{ + "base":"

Provides face metadata for target image faces that are analysed by CompareFaces and RecognizeCelebrities.

", + "refs":{ + "Celebrity$Face":"

Provides information about the celebrity's face, such as its location on the image.

", + "CompareFacesMatch$Face":"

Provides face metadata (bounding box and confidence that the bounding box actually contains a face).

", + "CompareFacesUnmatchList$member":null, + "ComparedFaceList$member":null } }, - "ComparedFaceList": { - "base": null, - "refs": { - "RecognizeCelebritiesResponse$UnrecognizedFaces": "

Details about each unrecognized face in the image.

" + "ComparedFaceList":{ + "base":null, + "refs":{ + "RecognizeCelebritiesResponse$UnrecognizedFaces":"

Details about each unrecognized face in the image.

" } }, - "ComparedSourceImageFace": { - "base": "

Type that describes the face Amazon Rekognition chose to compare with the faces in the target. This contains a bounding box for the selected face and confidence level that the bounding box contains a face. Note that Amazon Rekognition selects the largest face in the source image for this comparison.

", - "refs": { - "CompareFacesResponse$SourceImageFace": "

The face in the source image that was used for comparison.

" + "ComparedSourceImageFace":{ + "base":"

Type that describes the face Amazon Rekognition chose to compare with the faces in the target. This contains a bounding box for the selected face and confidence level that the bounding box contains a face. Note that Amazon Rekognition selects the largest face in the source image for this comparison.

", + "refs":{ + "CompareFacesResponse$SourceImageFace":"

The face in the source image that was used for comparison.

" } }, - "ContentModerationDetection": { - "base": "

Information about a moderation label detection in a stored video.

", - "refs": { - "ContentModerationDetections$member": null + "ContentModerationDetection":{ + "base":"

Information about a moderation label detection in a stored video.

", + "refs":{ + "ContentModerationDetections$member":null } }, - "ContentModerationDetections": { - "base": null, - "refs": { - "GetContentModerationResponse$ModerationLabels": "

The detected moderation labels and the time(s) they were detected.

" + "ContentModerationDetections":{ + "base":null, + "refs":{ + "GetContentModerationResponse$ModerationLabels":"

The detected moderation labels and the time(s) they were detected.

" } }, - "ContentModerationSortBy": { - "base": null, - "refs": { - "GetContentModerationRequest$SortBy": "

Sort to use for elements in the ModerationLabelDetections array. Use TIMESTAMP to sort array elements by the time labels are detected. Use NAME to alphabetically group elements for a label together. Within each label group, the array element are sorted by detection confidence. The default sort is by TIMESTAMP.

" + "ContentModerationSortBy":{ + "base":null, + "refs":{ + "GetContentModerationRequest$SortBy":"

Sort to use for elements in the ModerationLabelDetections array. Use TIMESTAMP to sort array elements by the time labels are detected. Use NAME to alphabetically group elements for a label together. Within each label group, the array element are sorted by detection confidence. The default sort is by TIMESTAMP.

" } }, - "CreateCollectionRequest": { - "base": null, - "refs": { + "CreateCollectionRequest":{ + "base":null, + "refs":{ } }, - "CreateCollectionResponse": { - "base": null, - "refs": { + "CreateCollectionResponse":{ + "base":null, + "refs":{ } }, - "CreateStreamProcessorRequest": { - "base": null, - "refs": { + "CreateStreamProcessorRequest":{ + "base":null, + "refs":{ } }, - "CreateStreamProcessorResponse": { - "base": null, - "refs": { + "CreateStreamProcessorResponse":{ + "base":null, + "refs":{ } }, - "DateTime": { - "base": null, - "refs": { - "DescribeStreamProcessorResponse$CreationTimestamp": "

Date and time the stream processor was created

", - "DescribeStreamProcessorResponse$LastUpdateTimestamp": "

The time, in Unix format, the stream processor was last updated. For example, when the stream processor moves from a running state to a failed state, or when the user starts or stops the stream processor.

" + "DateTime":{ + "base":null, + "refs":{ + "DescribeStreamProcessorResponse$CreationTimestamp":"

Date and time the stream processor was created

", + "DescribeStreamProcessorResponse$LastUpdateTimestamp":"

The time, in Unix format, the stream processor was last updated. For example, when the stream processor moves from a running state to a failed state, or when the user starts or stops the stream processor.

" } }, - "Degree": { - "base": null, - "refs": { - "Pose$Roll": "

Value representing the face rotation on the roll axis.

", - "Pose$Yaw": "

Value representing the face rotation on the yaw axis.

", - "Pose$Pitch": "

Value representing the face rotation on the pitch axis.

" + "Degree":{ + "base":null, + "refs":{ + "Pose$Roll":"

Value representing the face rotation on the roll axis.

", + "Pose$Yaw":"

Value representing the face rotation on the yaw axis.

", + "Pose$Pitch":"

Value representing the face rotation on the pitch axis.

" } }, - "DeleteCollectionRequest": { - "base": null, - "refs": { + "DeleteCollectionRequest":{ + "base":null, + "refs":{ } }, - "DeleteCollectionResponse": { - "base": null, - "refs": { + "DeleteCollectionResponse":{ + "base":null, + "refs":{ } }, - "DeleteFacesRequest": { - "base": null, - "refs": { + "DeleteFacesRequest":{ + "base":null, + "refs":{ } }, - "DeleteFacesResponse": { - "base": null, - "refs": { + "DeleteFacesResponse":{ + "base":null, + "refs":{ } }, - "DeleteStreamProcessorRequest": { - "base": null, - "refs": { + "DeleteStreamProcessorRequest":{ + "base":null, + "refs":{ } }, - "DeleteStreamProcessorResponse": { - "base": null, - "refs": { + "DeleteStreamProcessorResponse":{ + "base":null, + "refs":{ } }, - "DescribeStreamProcessorRequest": { - "base": null, - "refs": { + "DescribeStreamProcessorRequest":{ + "base":null, + "refs":{ } }, - "DescribeStreamProcessorResponse": { - "base": null, - "refs": { + "DescribeStreamProcessorResponse":{ + "base":null, + "refs":{ } }, - "DetectFacesRequest": { - "base": null, - "refs": { + "DetectFacesRequest":{ + "base":null, + "refs":{ } }, - "DetectFacesResponse": { - "base": null, - "refs": { + "DetectFacesResponse":{ + "base":null, + "refs":{ } }, - "DetectLabelsRequest": { - "base": null, - "refs": { + "DetectLabelsRequest":{ + "base":null, + "refs":{ } }, - "DetectLabelsResponse": { - "base": null, - "refs": { + "DetectLabelsResponse":{ + "base":null, + "refs":{ } }, - "DetectModerationLabelsRequest": { - "base": null, - "refs": { + "DetectModerationLabelsRequest":{ + "base":null, + "refs":{ } }, - "DetectModerationLabelsResponse": { - "base": null, - "refs": { + "DetectModerationLabelsResponse":{ + "base":null, + "refs":{ } }, - "DetectTextRequest": { - "base": null, - "refs": { + "DetectTextRequest":{ + "base":null, + "refs":{ } }, - "DetectTextResponse": { - "base": null, - "refs": { + "DetectTextResponse":{ + "base":null, + "refs":{ } }, - "Emotion": { - "base": "

The emotions detected on the face, and the confidence level in the determination. For example, HAPPY, SAD, and ANGRY.

", - "refs": { - "Emotions$member": null + "Emotion":{ + "base":"

The emotions detected on the face, and the confidence level in the determination. For example, HAPPY, SAD, and ANGRY.

", + "refs":{ + "Emotions$member":null } }, - "EmotionName": { - "base": null, - "refs": { - "Emotion$Type": "

Type of emotion detected.

" + "EmotionName":{ + "base":null, + "refs":{ + "Emotion$Type":"

Type of emotion detected.

" } }, - "Emotions": { - "base": null, - "refs": { - "FaceDetail$Emotions": "

The emotions detected on the face, and the confidence level in the determination. For example, HAPPY, SAD, and ANGRY.

" + "Emotions":{ + "base":null, + "refs":{ + "FaceDetail$Emotions":"

The emotions detected on the face, and the confidence level in the determination. For example, HAPPY, SAD, and ANGRY.

" } }, - "ExternalImageId": { - "base": null, - "refs": { - "Face$ExternalImageId": "

Identifier that you assign to all the faces in the input image.

", - "IndexFacesRequest$ExternalImageId": "

ID you want to assign to all the faces detected in the image.

" + "ExternalImageId":{ + "base":null, + "refs":{ + "Face$ExternalImageId":"

Identifier that you assign to all the faces in the input image.

", + "IndexFacesRequest$ExternalImageId":"

ID you want to assign to all the faces detected in the image.

" } }, - "EyeOpen": { - "base": "

Indicates whether or not the eyes on the face are open, and the confidence level in the determination.

", - "refs": { - "FaceDetail$EyesOpen": "

Indicates whether or not the eyes on the face are open, and the confidence level in the determination.

" + "EyeOpen":{ + "base":"

Indicates whether or not the eyes on the face are open, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$EyesOpen":"

Indicates whether or not the eyes on the face are open, and the confidence level in the determination.

" } }, - "Eyeglasses": { - "base": "

Indicates whether or not the face is wearing eye glasses, and the confidence level in the determination.

", - "refs": { - "FaceDetail$Eyeglasses": "

Indicates whether or not the face is wearing eye glasses, and the confidence level in the determination.

" + "Eyeglasses":{ + "base":"

Indicates whether or not the face is wearing eye glasses, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$Eyeglasses":"

Indicates whether or not the face is wearing eye glasses, and the confidence level in the determination.

" } }, - "Face": { - "base": "

Describes the face properties such as the bounding box, face ID, image ID of the input image, and external image ID that you assigned.

", - "refs": { - "FaceList$member": null, - "FaceMatch$Face": "

Describes the face properties such as the bounding box, face ID, image ID of the source image, and external image ID that you assigned.

", - "FaceRecord$Face": "

Describes the face properties such as the bounding box, face ID, image ID of the input image, and external image ID that you assigned.

" + "Face":{ + "base":"

Describes the face properties such as the bounding box, face ID, image ID of the input image, and external image ID that you assigned.

", + "refs":{ + "FaceList$member":null, + "FaceMatch$Face":"

Describes the face properties such as the bounding box, face ID, image ID of the source image, and external image ID that you assigned.

", + "FaceRecord$Face":"

Describes the face properties such as the bounding box, face ID, image ID of the input image, and external image ID that you assigned.

" } }, - "FaceAttributes": { - "base": null, - "refs": { - "StartFaceDetectionRequest$FaceAttributes": "

The face attributes you want returned.

DEFAULT - The following subset of facial attributes are returned: BoundingBox, Confidence, Pose, Quality and Landmarks.

ALL - All facial attributes are returned.

" + "FaceAttributes":{ + "base":null, + "refs":{ + "StartFaceDetectionRequest$FaceAttributes":"

The face attributes you want returned.

DEFAULT - The following subset of facial attributes are returned: BoundingBox, Confidence, Pose, Quality and Landmarks.

ALL - All facial attributes are returned.

" } }, - "FaceDetail": { - "base": "

Structure containing attributes of the face that the algorithm detected.

A FaceDetail object contains either the default facial attributes or all facial attributes. The default attributes are BoundingBox, Confidence, Landmarks, Pose, and Quality.

is the only Rekognition Video stored video operation that can return a FaceDetail object with all attributes. To specify which attributes to return, use the FaceAttributes input parameter for . The following Rekognition Video operations return only the default attributes. The corresponding Start operations don't have a FaceAttributes input parameter.

  • GetCelebrityRecognition

  • GetPersonTracking

  • GetFaceSearch

The Rekognition Image and operations can return all facial attributes. To specify which attributes to return, use the Attributes input parameter for DetectFaces. For IndexFaces, use the DetectAttributes input parameter.

", - "refs": { - "CelebrityDetail$Face": "

Face details for the recognized celebrity.

", - "FaceDetailList$member": null, - "FaceDetection$Face": "

The face properties for the detected face.

", - "FaceRecord$FaceDetail": "

Structure containing attributes of the face that the algorithm detected.

", - "PersonDetail$Face": "

Face details for the detected person.

" + "FaceDetail":{ + "base":"

Structure containing attributes of the face that the algorithm detected.

A FaceDetail object contains either the default facial attributes or all facial attributes. The default attributes are BoundingBox, Confidence, Landmarks, Pose, and Quality.

is the only Amazon Rekognition Video stored video operation that can return a FaceDetail object with all attributes. To specify which attributes to return, use the FaceAttributes input parameter for . The following Amazon Rekognition Video operations return only the default attributes. The corresponding Start operations don't have a FaceAttributes input parameter.

  • GetCelebrityRecognition

  • GetPersonTracking

  • GetFaceSearch

The Amazon Rekognition Image and operations can return all facial attributes. To specify which attributes to return, use the Attributes input parameter for DetectFaces. For IndexFaces, use the DetectAttributes input parameter.

", + "refs":{ + "CelebrityDetail$Face":"

Face details for the recognized celebrity.

", + "FaceDetailList$member":null, + "FaceDetection$Face":"

The face properties for the detected face.

", + "FaceRecord$FaceDetail":"

Structure containing attributes of the face that the algorithm detected.

", + "PersonDetail$Face":"

Face details for the detected person.

" } }, - "FaceDetailList": { - "base": null, - "refs": { - "DetectFacesResponse$FaceDetails": "

Details of each face found in the image.

" + "FaceDetailList":{ + "base":null, + "refs":{ + "DetectFacesResponse$FaceDetails":"

Details of each face found in the image.

" } }, - "FaceDetection": { - "base": "

Information about a face detected in a video analysis request and the time the face was detected in the video.

", - "refs": { - "FaceDetections$member": null + "FaceDetection":{ + "base":"

Information about a face detected in a video analysis request and the time the face was detected in the video.

", + "refs":{ + "FaceDetections$member":null } }, - "FaceDetections": { - "base": null, - "refs": { - "GetFaceDetectionResponse$Faces": "

An array of faces detected in the video. Each element contains a detected face's details and the time, in milliseconds from the start of the video, the face was detected.

" + "FaceDetections":{ + "base":null, + "refs":{ + "GetFaceDetectionResponse$Faces":"

An array of faces detected in the video. Each element contains a detected face's details and the time, in milliseconds from the start of the video, the face was detected.

" } }, - "FaceId": { - "base": null, - "refs": { - "Face$FaceId": "

Unique identifier that Amazon Rekognition assigns to the face.

", - "FaceIdList$member": null, - "SearchFacesRequest$FaceId": "

ID of a face to find matches for in the collection.

", - "SearchFacesResponse$SearchedFaceId": "

ID of the face that was searched for matches in a collection.

" + "FaceId":{ + "base":null, + "refs":{ + "Face$FaceId":"

Unique identifier that Amazon Rekognition assigns to the face.

", + "FaceIdList$member":null, + "SearchFacesRequest$FaceId":"

ID of a face to find matches for in the collection.

", + "SearchFacesResponse$SearchedFaceId":"

ID of the face that was searched for matches in a collection.

" } }, - "FaceIdList": { - "base": null, - "refs": { - "DeleteFacesRequest$FaceIds": "

An array of face IDs to delete.

", - "DeleteFacesResponse$DeletedFaces": "

An array of strings (face IDs) of the faces that were deleted.

" + "FaceIdList":{ + "base":null, + "refs":{ + "DeleteFacesRequest$FaceIds":"

An array of face IDs to delete.

", + "DeleteFacesResponse$DeletedFaces":"

An array of strings (face IDs) of the faces that were deleted.

" } }, - "FaceList": { - "base": null, - "refs": { - "ListFacesResponse$Faces": "

An array of Face objects.

" + "FaceList":{ + "base":null, + "refs":{ + "ListFacesResponse$Faces":"

An array of Face objects.

" } }, - "FaceMatch": { - "base": "

Provides face metadata. In addition, it also provides the confidence in the match of this face with the input face.

", - "refs": { - "FaceMatchList$member": null + "FaceMatch":{ + "base":"

Provides face metadata. In addition, it also provides the confidence in the match of this face with the input face.

", + "refs":{ + "FaceMatchList$member":null } }, - "FaceMatchList": { - "base": null, - "refs": { - "PersonMatch$FaceMatches": "

Information about the faces in the input collection that match the face of a person in the video.

", - "SearchFacesByImageResponse$FaceMatches": "

An array of faces that match the input face, along with the confidence in the match.

", - "SearchFacesResponse$FaceMatches": "

An array of faces that matched the input face, along with the confidence in the match.

" + "FaceMatchList":{ + "base":null, + "refs":{ + "PersonMatch$FaceMatches":"

Information about the faces in the input collection that match the face of a person in the video.

", + "SearchFacesByImageResponse$FaceMatches":"

An array of faces that match the input face, along with the confidence in the match.

", + "SearchFacesResponse$FaceMatches":"

An array of faces that matched the input face, along with the confidence in the match.

" } }, - "FaceModelVersionList": { - "base": null, - "refs": { - "ListCollectionsResponse$FaceModelVersions": "

Version numbers of the face detection models associated with the collections in the array CollectionIds. For example, the value of FaceModelVersions[2] is the version number for the face detection model used by the collection in CollectionId[2].

" + "FaceModelVersionList":{ + "base":null, + "refs":{ + "ListCollectionsResponse$FaceModelVersions":"

Version numbers of the face detection models associated with the collections in the array CollectionIds. For example, the value of FaceModelVersions[2] is the version number for the face detection model used by the collection in CollectionId[2].

" } }, - "FaceRecord": { - "base": "

Object containing both the face metadata (stored in the back-end database) and facial attributes that are detected but aren't stored in the database.

", - "refs": { - "FaceRecordList$member": null + "FaceRecord":{ + "base":"

Object containing both the face metadata (stored in the back-end database) and facial attributes that are detected but aren't stored in the database.

", + "refs":{ + "FaceRecordList$member":null } }, - "FaceRecordList": { - "base": null, - "refs": { - "IndexFacesResponse$FaceRecords": "

An array of faces detected and added to the collection. For more information, see collections-index-faces.

" + "FaceRecordList":{ + "base":null, + "refs":{ + "IndexFacesResponse$FaceRecords":"

An array of faces detected and added to the collection. For more information, see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.

" } }, - "FaceSearchSettings": { - "base": "

Input face recognition parameters for an Amazon Rekognition stream processor. FaceRecognitionSettings is a request parameter for .

", - "refs": { - "StreamProcessorSettings$FaceSearch": "

Face search settings to use on a streaming video.

" + "FaceSearchSettings":{ + "base":"

Input face recognition parameters for an Amazon Rekognition stream processor. FaceRecognitionSettings is a request parameter for .

", + "refs":{ + "StreamProcessorSettings$FaceSearch":"

Face search settings to use on a streaming video.

" } }, - "FaceSearchSortBy": { - "base": null, - "refs": { - "GetFaceSearchRequest$SortBy": "

Sort to use for grouping faces in the response. Use TIMESTAMP to group faces by the time that they are recognized. Use INDEX to sort by recognized faces.

" + "FaceSearchSortBy":{ + "base":null, + "refs":{ + "GetFaceSearchRequest$SortBy":"

Sort to use for grouping faces in the response. Use TIMESTAMP to group faces by the time that they are recognized. Use INDEX to sort by recognized faces.

" } }, - "Float": { - "base": null, - "refs": { - "BoundingBox$Width": "

Width of the bounding box as a ratio of the overall image width.

", - "BoundingBox$Height": "

Height of the bounding box as a ratio of the overall image height.

", - "BoundingBox$Left": "

Left coordinate of the bounding box as a ratio of overall image width.

", - "BoundingBox$Top": "

Top coordinate of the bounding box as a ratio of overall image height.

", - "ImageQuality$Brightness": "

Value representing brightness of the face. The service returns a value between 0 and 100 (inclusive). A higher value indicates a brighter face image.

", - "ImageQuality$Sharpness": "

Value representing sharpness of the face. The service returns a value between 0 and 100 (inclusive). A higher value indicates a sharper face image.

", - "Landmark$X": "

x-coordinate from the top left of the landmark expressed as the ratio of the width of the image. For example, if the images is 700x200 and the x-coordinate of the landmark is at 350 pixels, this value is 0.5.

", - "Landmark$Y": "

y-coordinate from the top left of the landmark expressed as the ratio of the height of the image. For example, if the images is 700x200 and the y-coordinate of the landmark is at 100 pixels, this value is 0.5.

", - "Point$X": "

The value of the X coordinate for a point on a Polygon.

", - "Point$Y": "

The value of the Y coordinate for a point on a Polygon.

", - "VideoMetadata$FrameRate": "

Number of frames per second in the video.

" + "Float":{ + "base":null, + "refs":{ + "BoundingBox$Width":"

Width of the bounding box as a ratio of the overall image width.

", + "BoundingBox$Height":"

Height of the bounding box as a ratio of the overall image height.

", + "BoundingBox$Left":"

Left coordinate of the bounding box as a ratio of overall image width.

", + "BoundingBox$Top":"

Top coordinate of the bounding box as a ratio of overall image height.

", + "ImageQuality$Brightness":"

Value representing brightness of the face. The service returns a value between 0 and 100 (inclusive). A higher value indicates a brighter face image.

", + "ImageQuality$Sharpness":"

Value representing sharpness of the face. The service returns a value between 0 and 100 (inclusive). A higher value indicates a sharper face image.

", + "Landmark$X":"

x-coordinate from the top left of the landmark expressed as the ratio of the width of the image. For example, if the images is 700x200 and the x-coordinate of the landmark is at 350 pixels, this value is 0.5.

", + "Landmark$Y":"

y-coordinate from the top left of the landmark expressed as the ratio of the height of the image. For example, if the images is 700x200 and the y-coordinate of the landmark is at 100 pixels, this value is 0.5.

", + "Point$X":"

The value of the X coordinate for a point on a Polygon.

", + "Point$Y":"

The value of the Y coordinate for a point on a Polygon.

", + "VideoMetadata$FrameRate":"

Number of frames per second in the video.

" } }, - "Gender": { - "base": "

Gender of the face and the confidence level in the determination.

", - "refs": { - "FaceDetail$Gender": "

Gender of the face and the confidence level in the determination.

" + "Gender":{ + "base":"

Gender of the face and the confidence level in the determination.

", + "refs":{ + "FaceDetail$Gender":"

Gender of the face and the confidence level in the determination.

" } }, - "GenderType": { - "base": null, - "refs": { - "Gender$Value": "

Gender of the face.

" + "GenderType":{ + "base":null, + "refs":{ + "Gender$Value":"

Gender of the face.

" } }, - "Geometry": { - "base": "

Information about where text detected by is located on an image.

", - "refs": { - "TextDetection$Geometry": "

The location of the detected text on the image. Includes an axis aligned coarse bounding box surrounding the text and a finer grain polygon for more accurate spatial information.

" + "Geometry":{ + "base":"

Information about where text detected by is located on an image.

", + "refs":{ + "TextDetection$Geometry":"

The location of the detected text on the image. Includes an axis aligned coarse bounding box surrounding the text and a finer grain polygon for more accurate spatial information.

" } }, - "GetCelebrityInfoRequest": { - "base": null, - "refs": { + "GetCelebrityInfoRequest":{ + "base":null, + "refs":{ } }, - "GetCelebrityInfoResponse": { - "base": null, - "refs": { + "GetCelebrityInfoResponse":{ + "base":null, + "refs":{ } }, - "GetCelebrityRecognitionRequest": { - "base": null, - "refs": { + "GetCelebrityRecognitionRequest":{ + "base":null, + "refs":{ } }, - "GetCelebrityRecognitionResponse": { - "base": null, - "refs": { + "GetCelebrityRecognitionResponse":{ + "base":null, + "refs":{ } }, - "GetContentModerationRequest": { - "base": null, - "refs": { + "GetContentModerationRequest":{ + "base":null, + "refs":{ } }, - "GetContentModerationResponse": { - "base": null, - "refs": { + "GetContentModerationResponse":{ + "base":null, + "refs":{ } }, - "GetFaceDetectionRequest": { - "base": null, - "refs": { + "GetFaceDetectionRequest":{ + "base":null, + "refs":{ } }, - "GetFaceDetectionResponse": { - "base": null, - "refs": { + "GetFaceDetectionResponse":{ + "base":null, + "refs":{ } }, - "GetFaceSearchRequest": { - "base": null, - "refs": { + "GetFaceSearchRequest":{ + "base":null, + "refs":{ } }, - "GetFaceSearchResponse": { - "base": null, - "refs": { + "GetFaceSearchResponse":{ + "base":null, + "refs":{ } }, - "GetLabelDetectionRequest": { - "base": null, - "refs": { + "GetLabelDetectionRequest":{ + "base":null, + "refs":{ } }, - "GetLabelDetectionResponse": { - "base": null, - "refs": { + "GetLabelDetectionResponse":{ + "base":null, + "refs":{ } }, - "GetPersonTrackingRequest": { - "base": null, - "refs": { + "GetPersonTrackingRequest":{ + "base":null, + "refs":{ } }, - "GetPersonTrackingResponse": { - "base": null, - "refs": { + "GetPersonTrackingResponse":{ + "base":null, + "refs":{ } }, - "IdempotentParameterMismatchException": { - "base": "

A ClientRequestToken input parameter was reused with an operation, but at least one of the other input parameters is different from the previous call to the operation.

", - "refs": { + "IdempotentParameterMismatchException":{ + "base":"

A ClientRequestToken input parameter was reused with an operation, but at least one of the other input parameters is different from the previous call to the operation.

", + "refs":{ } }, - "Image": { - "base": "

Provides the input image either as bytes or an S3 object.

You pass image bytes to a Rekognition API operation by using the Bytes property. For example, you would use the Bytes property to pass an image loaded from a local file system. Image bytes passed by using the Bytes property must be base64-encoded. Your code may not need to encode image bytes if you are using an AWS SDK to call Rekognition API operations. For more information, see images-bytes.

You pass images stored in an S3 bucket to a Rekognition API operation by using the S3Object property. Images stored in an S3 bucket do not need to be base64-encoded.

The region for the S3 bucket containing the S3 object must match the region you use for Amazon Rekognition operations.

If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes using the Bytes property is not supported. You must first upload the image to an Amazon S3 bucket and then call the operation using the S3Object property.

For Amazon Rekognition to process an S3 object, the user must have permission to access the S3 object. For more information, see manage-access-resource-policies.

", - "refs": { - "CompareFacesRequest$SourceImage": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "CompareFacesRequest$TargetImage": "

The target image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "DetectFacesRequest$Image": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "DetectLabelsRequest$Image": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "DetectModerationLabelsRequest$Image": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "DetectTextRequest$Image": "

The input image as base64-encoded bytes or an Amazon S3 object. If you use the AWS CLI to call Amazon Rekognition operations, you can't pass image bytes.

", - "IndexFacesRequest$Image": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "RecognizeCelebritiesRequest$Image": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", - "SearchFacesByImageRequest$Image": "

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

" + "Image":{ + "base":"

Provides the input image either as bytes or an S3 object.

You pass image bytes to a Rekognition API operation by using the Bytes property. For example, you would use the Bytes property to pass an image loaded from a local file system. Image bytes passed by using the Bytes property must be base64-encoded. Your code may not need to encode image bytes if you are using an AWS SDK to call Rekognition API operations.

For more information, see Analyzing an Image Loaded from a Local File System in the Amazon Rekognition Developer Guide.

You pass images stored in an S3 bucket to a Rekognition API operation by using the S3Object property. Images stored in an S3 bucket do not need to be base64-encoded.

The region for the S3 bucket containing the S3 object must match the region you use for Amazon Rekognition operations.

If you use the Amazon CLI to call Amazon Rekognition operations, passing image bytes using the Bytes property is not supported. You must first upload the image to an Amazon S3 bucket and then call the operation using the S3Object property.

For Amazon Rekognition to process an S3 object, the user must have permission to access the S3 object. For more information, see Resource Based Policies in the Amazon Rekognition Developer Guide.

", + "refs":{ + "CompareFacesRequest$SourceImage":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "CompareFacesRequest$TargetImage":"

The target image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "DetectFacesRequest$Image":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "DetectLabelsRequest$Image":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "DetectModerationLabelsRequest$Image":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "DetectTextRequest$Image":"

The input image as base64-encoded bytes or an Amazon S3 object. If you use the AWS CLI to call Amazon Rekognition operations, you can't pass image bytes.

", + "IndexFacesRequest$Image":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "RecognizeCelebritiesRequest$Image":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

", + "SearchFacesByImageRequest$Image":"

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes is not supported.

" } }, - "ImageBlob": { - "base": null, - "refs": { - "Image$Bytes": "

Blob of image bytes up to 5 MBs.

" + "ImageBlob":{ + "base":null, + "refs":{ + "Image$Bytes":"

Blob of image bytes up to 5 MBs.

" } }, - "ImageId": { - "base": null, - "refs": { - "Face$ImageId": "

Unique identifier that Amazon Rekognition assigns to the input image.

" + "ImageId":{ + "base":null, + "refs":{ + "Face$ImageId":"

Unique identifier that Amazon Rekognition assigns to the input image.

" } }, - "ImageQuality": { - "base": "

Identifies face image brightness and sharpness.

", - "refs": { - "ComparedFace$Quality": "

Identifies face image brightness and sharpness.

", - "FaceDetail$Quality": "

Identifies image brightness and sharpness. Default attribute.

" + "ImageQuality":{ + "base":"

Identifies face image brightness and sharpness.

", + "refs":{ + "ComparedFace$Quality":"

Identifies face image brightness and sharpness.

", + "FaceDetail$Quality":"

Identifies image brightness and sharpness. Default attribute.

" } }, - "ImageTooLargeException": { - "base": "

The input image size exceeds the allowed limit. For more information, see limits.

", - "refs": { + "ImageTooLargeException":{ + "base":"

The input image size exceeds the allowed limit. For more information, see Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide.

", + "refs":{ } }, - "IndexFacesRequest": { - "base": null, - "refs": { + "IndexFacesRequest":{ + "base":null, + "refs":{ } }, - "IndexFacesResponse": { - "base": null, - "refs": { + "IndexFacesResponse":{ + "base":null, + "refs":{ } }, - "InternalServerError": { - "base": "

Amazon Rekognition experienced a service issue. Try your call again.

", - "refs": { + "InternalServerError":{ + "base":"

Amazon Rekognition experienced a service issue. Try your call again.

", + "refs":{ } }, - "InvalidImageFormatException": { - "base": "

The provided image format is not supported.

", - "refs": { + "InvalidImageFormatException":{ + "base":"

The provided image format is not supported.

", + "refs":{ } }, - "InvalidPaginationTokenException": { - "base": "

Pagination token in the request is not valid.

", - "refs": { + "InvalidPaginationTokenException":{ + "base":"

Pagination token in the request is not valid.

", + "refs":{ } }, - "InvalidParameterException": { - "base": "

Input parameter violated a constraint. Validate your parameter before calling the API operation again.

", - "refs": { + "InvalidParameterException":{ + "base":"

Input parameter violated a constraint. Validate your parameter before calling the API operation again.

", + "refs":{ } }, - "InvalidS3ObjectException": { - "base": "

Amazon Rekognition is unable to access the S3 object specified in the request.

", - "refs": { + "InvalidS3ObjectException":{ + "base":"

Amazon Rekognition is unable to access the S3 object specified in the request.

", + "refs":{ } }, - "JobId": { - "base": null, - "refs": { - "GetCelebrityRecognitionRequest$JobId": "

Job identifier for the required celebrity recognition analysis. You can get the job identifer from a call to StartCelebrityRecognition.

", - "GetContentModerationRequest$JobId": "

The identifier for the content moderation job. Use JobId to identify the job in a subsequent call to GetContentModeration.

", - "GetFaceDetectionRequest$JobId": "

Unique identifier for the face detection job. The JobId is returned from StartFaceDetection.

", - "GetFaceSearchRequest$JobId": "

The job identifer for the search request. You get the job identifier from an initial call to StartFaceSearch.

", - "GetLabelDetectionRequest$JobId": "

Job identifier for the label detection operation for which you want results returned. You get the job identifer from an initial call to StartlabelDetection.

", - "GetPersonTrackingRequest$JobId": "

The identifier for a job that tracks persons in a video. You get the JobId from a call to StartPersonTracking.

", - "StartCelebrityRecognitionResponse$JobId": "

The identifier for the celebrity recognition analysis job. Use JobId to identify the job in a subsequent call to GetCelebrityRecognition.

", - "StartContentModerationResponse$JobId": "

The identifier for the content moderation analysis job. Use JobId to identify the job in a subsequent call to GetContentModeration.

", - "StartFaceDetectionResponse$JobId": "

The identifier for the face detection job. Use JobId to identify the job in a subsequent call to GetFaceDetection.

", - "StartFaceSearchResponse$JobId": "

The identifier for the search job. Use JobId to identify the job in a subsequent call to GetFaceSearch.

", - "StartLabelDetectionResponse$JobId": "

The identifier for the label detection job. Use JobId to identify the job in a subsequent call to GetLabelDetection.

", - "StartPersonTrackingResponse$JobId": "

The identifier for the person detection job. Use JobId to identify the job in a subsequent call to GetPersonTracking.

" + "JobId":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionRequest$JobId":"

Job identifier for the required celebrity recognition analysis. You can get the job identifer from a call to StartCelebrityRecognition.

", + "GetContentModerationRequest$JobId":"

The identifier for the content moderation job. Use JobId to identify the job in a subsequent call to GetContentModeration.

", + "GetFaceDetectionRequest$JobId":"

Unique identifier for the face detection job. The JobId is returned from StartFaceDetection.

", + "GetFaceSearchRequest$JobId":"

The job identifer for the search request. You get the job identifier from an initial call to StartFaceSearch.

", + "GetLabelDetectionRequest$JobId":"

Job identifier for the label detection operation for which you want results returned. You get the job identifer from an initial call to StartlabelDetection.

", + "GetPersonTrackingRequest$JobId":"

The identifier for a job that tracks persons in a video. You get the JobId from a call to StartPersonTracking.

", + "StartCelebrityRecognitionResponse$JobId":"

The identifier for the celebrity recognition analysis job. Use JobId to identify the job in a subsequent call to GetCelebrityRecognition.

", + "StartContentModerationResponse$JobId":"

The identifier for the content moderation analysis job. Use JobId to identify the job in a subsequent call to GetContentModeration.

", + "StartFaceDetectionResponse$JobId":"

The identifier for the face detection job. Use JobId to identify the job in a subsequent call to GetFaceDetection.

", + "StartFaceSearchResponse$JobId":"

The identifier for the search job. Use JobId to identify the job in a subsequent call to GetFaceSearch.

", + "StartLabelDetectionResponse$JobId":"

The identifier for the label detection job. Use JobId to identify the job in a subsequent call to GetLabelDetection.

", + "StartPersonTrackingResponse$JobId":"

The identifier for the person detection job. Use JobId to identify the job in a subsequent call to GetPersonTracking.

" } }, - "JobTag": { - "base": null, - "refs": { - "StartCelebrityRecognitionRequest$JobTag": "

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", - "StartContentModerationRequest$JobTag": "

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", - "StartFaceDetectionRequest$JobTag": "

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", - "StartFaceSearchRequest$JobTag": "

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", - "StartLabelDetectionRequest$JobTag": "

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", - "StartPersonTrackingRequest$JobTag": "

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

" + "JobTag":{ + "base":null, + "refs":{ + "StartCelebrityRecognitionRequest$JobTag":"

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", + "StartContentModerationRequest$JobTag":"

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", + "StartFaceDetectionRequest$JobTag":"

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", + "StartFaceSearchRequest$JobTag":"

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", + "StartLabelDetectionRequest$JobTag":"

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

", + "StartPersonTrackingRequest$JobTag":"

Unique identifier you specify to identify the job in the completion status published to the Amazon Simple Notification Service topic.

" } }, - "KinesisDataArn": { - "base": null, - "refs": { - "KinesisDataStream$Arn": "

ARN of the output Amazon Kinesis Data Streams stream.

" + "KinesisDataArn":{ + "base":null, + "refs":{ + "KinesisDataStream$Arn":"

ARN of the output Amazon Kinesis Data Streams stream.

" } }, - "KinesisDataStream": { - "base": "

The Kinesis data stream Amazon Rekognition to which the analysis results of a Amazon Rekognition stream processor are streamed. For more information, see .

", - "refs": { - "StreamProcessorOutput$KinesisDataStream": "

The Amazon Kinesis Data Streams stream to which the Amazon Rekognition stream processor streams the analysis results.

" + "KinesisDataStream":{ + "base":"

The Kinesis data stream Amazon Rekognition to which the analysis results of a Amazon Rekognition stream processor are streamed. For more information, see CreateStreamProcessor in the Amazon Rekognition Developer Guide.

", + "refs":{ + "StreamProcessorOutput$KinesisDataStream":"

The Amazon Kinesis Data Streams stream to which the Amazon Rekognition stream processor streams the analysis results.

" } }, - "KinesisVideoArn": { - "base": null, - "refs": { - "KinesisVideoStream$Arn": "

ARN of the Kinesis video stream stream that streams the source video.

" + "KinesisVideoArn":{ + "base":null, + "refs":{ + "KinesisVideoStream$Arn":"

ARN of the Kinesis video stream stream that streams the source video.

" } }, - "KinesisVideoStream": { - "base": "

Kinesis video stream stream that provides the source streaming video for a Rekognition Video stream processor. For more information, see .

", - "refs": { - "StreamProcessorInput$KinesisVideoStream": "

The Kinesis video stream input stream for the source streaming video.

" + "KinesisVideoStream":{ + "base":"

Kinesis video stream stream that provides the source streaming video for a Amazon Rekognition Video stream processor. For more information, see CreateStreamProcessor in the Amazon Rekognition Developer Guide.

", + "refs":{ + "StreamProcessorInput$KinesisVideoStream":"

The Kinesis video stream input stream for the source streaming video.

" } }, - "Label": { - "base": "

Structure containing details about the detected label, including name, and level of confidence.

", - "refs": { - "LabelDetection$Label": "

Details about the detected label.

", - "Labels$member": null + "Label":{ + "base":"

Structure containing details about the detected label, including name, and level of confidence.

", + "refs":{ + "LabelDetection$Label":"

Details about the detected label.

", + "Labels$member":null } }, - "LabelDetection": { - "base": "

Information about a label detected in a video analysis request and the time the label was detected in the video.

", - "refs": { - "LabelDetections$member": null + "LabelDetection":{ + "base":"

Information about a label detected in a video analysis request and the time the label was detected in the video.

", + "refs":{ + "LabelDetections$member":null } }, - "LabelDetectionSortBy": { - "base": null, - "refs": { - "GetLabelDetectionRequest$SortBy": "

Sort to use for elements in the Labels array. Use TIMESTAMP to sort array elements by the time labels are detected. Use NAME to alphabetically group elements for a label together. Within each label group, the array element are sorted by detection confidence. The default sort is by TIMESTAMP.

" + "LabelDetectionSortBy":{ + "base":null, + "refs":{ + "GetLabelDetectionRequest$SortBy":"

Sort to use for elements in the Labels array. Use TIMESTAMP to sort array elements by the time labels are detected. Use NAME to alphabetically group elements for a label together. Within each label group, the array element are sorted by detection confidence. The default sort is by TIMESTAMP.

" } }, - "LabelDetections": { - "base": null, - "refs": { - "GetLabelDetectionResponse$Labels": "

An array of labels detected in the video. Each element contains the detected label and the time, in milliseconds from the start of the video, that the label was detected.

" + "LabelDetections":{ + "base":null, + "refs":{ + "GetLabelDetectionResponse$Labels":"

An array of labels detected in the video. Each element contains the detected label and the time, in milliseconds from the start of the video, that the label was detected.

" } }, - "Labels": { - "base": null, - "refs": { - "DetectLabelsResponse$Labels": "

An array of labels for the real-world objects detected.

" + "Labels":{ + "base":null, + "refs":{ + "DetectLabelsResponse$Labels":"

An array of labels for the real-world objects detected.

" } }, - "Landmark": { - "base": "

Indicates the location of the landmark on the face.

", - "refs": { - "Landmarks$member": null + "Landmark":{ + "base":"

Indicates the location of the landmark on the face.

", + "refs":{ + "Landmarks$member":null } }, - "LandmarkType": { - "base": null, - "refs": { - "Landmark$Type": "

Type of the landmark.

" + "LandmarkType":{ + "base":null, + "refs":{ + "Landmark$Type":"

Type of the landmark.

" } }, - "Landmarks": { - "base": null, - "refs": { - "ComparedFace$Landmarks": "

An array of facial landmarks.

", - "FaceDetail$Landmarks": "

Indicates the location of landmarks on the face. Default attribute.

" + "Landmarks":{ + "base":null, + "refs":{ + "ComparedFace$Landmarks":"

An array of facial landmarks.

", + "FaceDetail$Landmarks":"

Indicates the location of landmarks on the face. Default attribute.

" } }, - "LimitExceededException": { - "base": "

An Amazon Rekognition service limit was exceeded. For example, if you start too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, for example) will raise a LimitExceededException exception (HTTP status code: 400) until the number of concurrently running jobs is below the Amazon Rekognition service limit.

", - "refs": { + "LimitExceededException":{ + "base":"

An Amazon Rekognition service limit was exceeded. For example, if you start too many Amazon Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, for example) will raise a LimitExceededException exception (HTTP status code: 400) until the number of concurrently running jobs is below the Amazon Rekognition service limit.

", + "refs":{ } }, - "ListCollectionsRequest": { - "base": null, - "refs": { + "ListCollectionsRequest":{ + "base":null, + "refs":{ } }, - "ListCollectionsResponse": { - "base": null, - "refs": { + "ListCollectionsResponse":{ + "base":null, + "refs":{ } }, - "ListFacesRequest": { - "base": null, - "refs": { + "ListFacesRequest":{ + "base":null, + "refs":{ } }, - "ListFacesResponse": { - "base": null, - "refs": { + "ListFacesResponse":{ + "base":null, + "refs":{ } }, - "ListStreamProcessorsRequest": { - "base": null, - "refs": { + "ListStreamProcessorsRequest":{ + "base":null, + "refs":{ } }, - "ListStreamProcessorsResponse": { - "base": null, - "refs": { + "ListStreamProcessorsResponse":{ + "base":null, + "refs":{ } }, - "MaxFaces": { - "base": null, - "refs": { - "SearchFacesByImageRequest$MaxFaces": "

Maximum number of faces to return. The operation returns the maximum number of faces with the highest confidence in the match.

", - "SearchFacesRequest$MaxFaces": "

Maximum number of faces to return. The operation returns the maximum number of faces with the highest confidence in the match.

" + "MaxFaces":{ + "base":null, + "refs":{ + "SearchFacesByImageRequest$MaxFaces":"

Maximum number of faces to return. The operation returns the maximum number of faces with the highest confidence in the match.

", + "SearchFacesRequest$MaxFaces":"

Maximum number of faces to return. The operation returns the maximum number of faces with the highest confidence in the match.

" } }, - "MaxResults": { - "base": null, - "refs": { - "GetCelebrityRecognitionRequest$MaxResults": "

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", - "GetContentModerationRequest$MaxResults": "

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", - "GetFaceDetectionRequest$MaxResults": "

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", - "GetFaceSearchRequest$MaxResults": "

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", - "GetLabelDetectionRequest$MaxResults": "

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", - "GetPersonTrackingRequest$MaxResults": "

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", - "ListStreamProcessorsRequest$MaxResults": "

Maximum number of stream processors you want Rekognition Video to return in the response. The default is 1000.

" + "MaxResults":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionRequest$MaxResults":"

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", + "GetContentModerationRequest$MaxResults":"

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", + "GetFaceDetectionRequest$MaxResults":"

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", + "GetFaceSearchRequest$MaxResults":"

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", + "GetLabelDetectionRequest$MaxResults":"

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", + "GetPersonTrackingRequest$MaxResults":"

Maximum number of results to return per paginated call. The largest value you can specify is 1000. If you specify a value greater than 1000, a maximum of 1000 results is returned. The default value is 1000.

", + "ListStreamProcessorsRequest$MaxResults":"

Maximum number of stream processors you want Amazon Rekognition Video to return in the response. The default is 1000.

" } }, - "ModerationLabel": { - "base": "

Provides information about a single type of moderated content found in an image or video. Each type of moderated content has a label within a hierarchical taxonomy. For more information, see moderation.

", - "refs": { - "ContentModerationDetection$ModerationLabel": "

The moderation label detected by in the stored video.

", - "ModerationLabels$member": null - } - }, - "ModerationLabels": { - "base": null, - "refs": { - "DetectModerationLabelsResponse$ModerationLabels": "

Array of detected Moderation labels and the time, in millseconds from the start of the video, they were detected.

" - } - }, - "MouthOpen": { - "base": "

Indicates whether or not the mouth on the face is open, and the confidence level in the determination.

", - "refs": { - "FaceDetail$MouthOpen": "

Indicates whether or not the mouth on the face is open, and the confidence level in the determination.

" - } - }, - "Mustache": { - "base": "

Indicates whether or not the face has a mustache, and the confidence level in the determination.

", - "refs": { - "FaceDetail$Mustache": "

Indicates whether or not the face has a mustache, and the confidence level in the determination.

" - } - }, - "NotificationChannel": { - "base": "

The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the completion status of a video analysis operation. For more information, see api-video.

", - "refs": { - "StartCelebrityRecognitionRequest$NotificationChannel": "

The Amazon SNS topic ARN that you want Rekognition Video to publish the completion status of the celebrity recognition analysis to.

", - "StartContentModerationRequest$NotificationChannel": "

The Amazon SNS topic ARN that you want Rekognition Video to publish the completion status of the content moderation analysis to.

", - "StartFaceDetectionRequest$NotificationChannel": "

The ARN of the Amazon SNS topic to which you want Rekognition Video to publish the completion status of the face detection operation.

", - "StartFaceSearchRequest$NotificationChannel": "

The ARN of the Amazon SNS topic to which you want Rekognition Video to publish the completion status of the search.

", - "StartLabelDetectionRequest$NotificationChannel": "

The Amazon SNS topic ARN you want Rekognition Video to publish the completion status of the label detection operation to.

", - "StartPersonTrackingRequest$NotificationChannel": "

The Amazon SNS topic ARN you want Rekognition Video to publish the completion status of the people detection operation to.

" - } - }, - "OrientationCorrection": { - "base": null, - "refs": { - "CompareFacesResponse$SourceImageOrientationCorrection": "

The orientation of the source image (counterclockwise direction). If your application displays the source image, you can use this value to correct image orientation. The bounding box coordinates returned in SourceImageFace represent the location of the face before the image orientation is corrected.

If the source image is in .jpeg format, it might contain exchangeable image (Exif) metadata that includes the image's orientation. If the Exif metadata for the source image populates the orientation field, the value of OrientationCorrection is null and the SourceImageFace bounding box coordinates represent the location of the face after Exif metadata is used to correct the orientation. Images in .png format don't contain Exif metadata.

", - "CompareFacesResponse$TargetImageOrientationCorrection": "

The orientation of the target image (in counterclockwise direction). If your application displays the target image, you can use this value to correct the orientation of the image. The bounding box coordinates returned in FaceMatches and UnmatchedFaces represent face locations before the image orientation is corrected.

If the target image is in .jpg format, it might contain Exif metadata that includes the orientation of the image. If the Exif metadata for the target image populates the orientation field, the value of OrientationCorrection is null and the bounding box coordinates in FaceMatches and UnmatchedFaces represent the location of the face after Exif metadata is used to correct the orientation. Images in .png format don't contain Exif metadata.

", - "DetectFacesResponse$OrientationCorrection": "

The orientation of the input image (counter-clockwise direction). If your application displays the image, you can use this value to correct image orientation. The bounding box coordinates returned in FaceDetails represent face locations before the image orientation is corrected.

If the input image is in .jpeg format, it might contain exchangeable image (Exif) metadata that includes the image's orientation. If so, and the Exif metadata for the input image populates the orientation field, the value of OrientationCorrection is null and the FaceDetails bounding box coordinates represent face locations after Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata.

", - "DetectLabelsResponse$OrientationCorrection": "

The orientation of the input image (counter-clockwise direction). If your application displays the image, you can use this value to correct the orientation. If Amazon Rekognition detects that the input image was rotated (for example, by 90 degrees), it first corrects the orientation before detecting the labels.

If the input image Exif metadata populates the orientation field, Amazon Rekognition does not perform orientation correction and the value of OrientationCorrection will be null.

", - "IndexFacesResponse$OrientationCorrection": "

The orientation of the input image (counterclockwise direction). If your application displays the image, you can use this value to correct image orientation. The bounding box coordinates returned in FaceRecords represent face locations before the image orientation is corrected.

If the input image is in jpeg format, it might contain exchangeable image (Exif) metadata. If so, and the Exif metadata populates the orientation field, the value of OrientationCorrection is null and the bounding box coordinates in FaceRecords represent face locations after Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata.

", - "RecognizeCelebritiesResponse$OrientationCorrection": "

The orientation of the input image (counterclockwise direction). If your application displays the image, you can use this value to correct the orientation. The bounding box coordinates returned in CelebrityFaces and UnrecognizedFaces represent face locations before the image orientation is corrected.

If the input image is in .jpeg format, it might contain exchangeable image (Exif) metadata that includes the image's orientation. If so, and the Exif metadata for the input image populates the orientation field, the value of OrientationCorrection is null and the CelebrityFaces and UnrecognizedFaces bounding box coordinates represent face locations after Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata.

" - } - }, - "PageSize": { - "base": null, - "refs": { - "ListCollectionsRequest$MaxResults": "

Maximum number of collection IDs to return.

", - "ListFacesRequest$MaxResults": "

Maximum number of faces to return.

" - } - }, - "PaginationToken": { - "base": null, - "refs": { - "GetCelebrityRecognitionRequest$NextToken": "

If the previous response was incomplete (because there is more recognized celebrities to retrieve), Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of celebrities.

", - "GetCelebrityRecognitionResponse$NextToken": "

If the response is truncated, Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of celebrities.

", - "GetContentModerationRequest$NextToken": "

If the previous response was incomplete (because there is more data to retrieve), Amazon Rekognition returns a pagination token in the response. You can use this pagination token to retrieve the next set of content moderation labels.

", - "GetContentModerationResponse$NextToken": "

If the response is truncated, Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of moderation labels.

", - "GetFaceDetectionRequest$NextToken": "

If the previous response was incomplete (because there are more faces to retrieve), Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of faces.

", - "GetFaceDetectionResponse$NextToken": "

If the response is truncated, Amazon Rekognition returns this token that you can use in the subsequent request to retrieve the next set of faces.

", - "GetFaceSearchRequest$NextToken": "

If the previous response was incomplete (because there is more search results to retrieve), Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of search results.

", - "GetFaceSearchResponse$NextToken": "

If the response is truncated, Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of search results.

", - "GetLabelDetectionRequest$NextToken": "

If the previous response was incomplete (because there are more labels to retrieve), Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of labels.

", - "GetLabelDetectionResponse$NextToken": "

If the response is truncated, Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of labels.

", - "GetPersonTrackingRequest$NextToken": "

If the previous response was incomplete (because there are more persons to retrieve), Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of persons.

", - "GetPersonTrackingResponse$NextToken": "

If the response is truncated, Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of persons.

", - "ListCollectionsRequest$NextToken": "

Pagination token from the previous response.

", - "ListCollectionsResponse$NextToken": "

If the result is truncated, the response provides a NextToken that you can use in the subsequent request to fetch the next set of collection IDs.

", - "ListFacesRequest$NextToken": "

If the previous response was incomplete (because there is more data to retrieve), Amazon Rekognition returns a pagination token in the response. You can use this pagination token to retrieve the next set of faces.

", - "ListStreamProcessorsRequest$NextToken": "

If the previous response was incomplete (because there are more stream processors to retrieve), Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of stream processors.

", - "ListStreamProcessorsResponse$NextToken": "

If the response is truncated, Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of stream processors.

" - } - }, - "Percent": { - "base": null, - "refs": { - "Beard$Confidence": "

Level of confidence in the determination.

", - "Celebrity$MatchConfidence": "

The confidence, in percentage, that Rekognition has that the recognized face is the celebrity.

", - "CelebrityDetail$Confidence": "

The confidence, in percentage, that Amazon Rekognition has that the recognized face is the celebrity.

", - "CompareFacesMatch$Similarity": "

Level of confidence that the faces match.

", - "CompareFacesRequest$SimilarityThreshold": "

The minimum level of confidence in the face matches that a match must meet to be included in the FaceMatches array.

", - "ComparedFace$Confidence": "

Level of confidence that what the bounding box contains is a face.

", - "ComparedSourceImageFace$Confidence": "

Confidence level that the selected bounding box contains a face.

", - "DetectLabelsRequest$MinConfidence": "

Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn't return any labels with confidence lower than this specified value.

If MinConfidence is not specified, the operation returns labels with a confidence values greater than or equal to 50 percent.

", - "DetectModerationLabelsRequest$MinConfidence": "

Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn't return any labels with a confidence level lower than this specified value.

If you don't specify MinConfidence, the operation returns labels with confidence values greater than or equal to 50 percent.

", - "Emotion$Confidence": "

Level of confidence in the determination.

", - "EyeOpen$Confidence": "

Level of confidence in the determination.

", - "Eyeglasses$Confidence": "

Level of confidence in the determination.

", - "Face$Confidence": "

Confidence level that the bounding box contains a face (and not a different object such as a tree).

", - "FaceDetail$Confidence": "

Confidence level that the bounding box contains a face (and not a different object such as a tree). Default attribute.

", - "FaceMatch$Similarity": "

Confidence in the match of this face with the input face.

", - "FaceSearchSettings$FaceMatchThreshold": "

Minimum face match confidence score that must be met to return a result for a recognized face. Default is 70. 0 is the lowest confidence. 100 is the highest confidence.

", - "Gender$Confidence": "

Level of confidence in the determination.

", - "Label$Confidence": "

Level of confidence.

", - "ModerationLabel$Confidence": "

Specifies the confidence that Amazon Rekognition has that the label has been correctly identified.

If you don't specify the MinConfidence parameter in the call to DetectModerationLabels, the operation returns labels with a confidence value greater than or equal to 50 percent.

", - "MouthOpen$Confidence": "

Level of confidence in the determination.

", - "Mustache$Confidence": "

Level of confidence in the determination.

", - "SearchFacesByImageRequest$FaceMatchThreshold": "

(Optional) Specifies the minimum confidence in the face match to return. For example, don't return any matches where confidence in matches is less than 70%.

", - "SearchFacesByImageResponse$SearchedFaceConfidence": "

The level of confidence that the searchedFaceBoundingBox, contains a face.

", - "SearchFacesRequest$FaceMatchThreshold": "

Optional value specifying the minimum confidence in the face match to return. For example, don't return any matches where confidence in matches is less than 70%.

", - "Smile$Confidence": "

Level of confidence in the determination.

", - "StartContentModerationRequest$MinConfidence": "

Specifies the minimum confidence that Amazon Rekognition must have in order to return a moderated content label. Confidence represents how certain Amazon Rekognition is that the moderated content is correctly identified. 0 is the lowest confidence. 100 is the highest confidence. Amazon Rekognition doesn't return any moderated content labels with a confidence level lower than this specified value.

", - "StartFaceSearchRequest$FaceMatchThreshold": "

The minimum confidence in the person match to return. For example, don't return any matches where confidence in matches is less than 70%.

", - "StartLabelDetectionRequest$MinConfidence": "

Specifies the minimum confidence that Rekognition Video must have in order to return a detected label. Confidence represents how certain Amazon Rekognition is that a label is correctly identified.0 is the lowest confidence. 100 is the highest confidence. Rekognition Video doesn't return any labels with a confidence level lower than this specified value.

If you don't specify MinConfidence, the operation returns labels with confidence values greater than or equal to 50 percent.

", - "Sunglasses$Confidence": "

Level of confidence in the determination.

", - "TextDetection$Confidence": "

The confidence that Amazon Rekognition has in the accuracy of the detected text and the accuracy of the geometry points around the detected text.

" - } + "ModerationLabel":{ + "base":"

Provides information about a single type of moderated content found in an image or video. Each type of moderated content has a label within a hierarchical taxonomy. For more information, see Detecting Unsafe Content in the Amazon Rekognition Developer Guide.

", + "refs":{ + "ContentModerationDetection$ModerationLabel":"

The moderation label detected by in the stored video.

", + "ModerationLabels$member":null + } + }, + "ModerationLabels":{ + "base":null, + "refs":{ + "DetectModerationLabelsResponse$ModerationLabels":"

Array of detected Moderation labels and the time, in millseconds from the start of the video, they were detected.

" + } + }, + "MouthOpen":{ + "base":"

Indicates whether or not the mouth on the face is open, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$MouthOpen":"

Indicates whether or not the mouth on the face is open, and the confidence level in the determination.

" + } + }, + "Mustache":{ + "base":"

Indicates whether or not the face has a mustache, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$Mustache":"

Indicates whether or not the face has a mustache, and the confidence level in the determination.

" + } + }, + "NotificationChannel":{ + "base":"

The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the completion status of a video analysis operation. For more information, see api-video.

", + "refs":{ + "StartCelebrityRecognitionRequest$NotificationChannel":"

The Amazon SNS topic ARN that you want Amazon Rekognition Video to publish the completion status of the celebrity recognition analysis to.

", + "StartContentModerationRequest$NotificationChannel":"

The Amazon SNS topic ARN that you want Amazon Rekognition Video to publish the completion status of the content moderation analysis to.

", + "StartFaceDetectionRequest$NotificationChannel":"

The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video to publish the completion status of the face detection operation.

", + "StartFaceSearchRequest$NotificationChannel":"

The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video to publish the completion status of the search.

", + "StartLabelDetectionRequest$NotificationChannel":"

The Amazon SNS topic ARN you want Amazon Rekognition Video to publish the completion status of the label detection operation to.

", + "StartPersonTrackingRequest$NotificationChannel":"

The Amazon SNS topic ARN you want Amazon Rekognition Video to publish the completion status of the people detection operation to.

" + } + }, + "OrientationCorrection":{ + "base":null, + "refs":{ + "CompareFacesResponse$SourceImageOrientationCorrection":"

The orientation of the source image (counterclockwise direction). If your application displays the source image, you can use this value to correct image orientation. The bounding box coordinates returned in SourceImageFace represent the location of the face before the image orientation is corrected.

If the source image is in .jpeg format, it might contain exchangeable image (Exif) metadata that includes the image's orientation. If the Exif metadata for the source image populates the orientation field, the value of OrientationCorrection is null and the SourceImageFace bounding box coordinates represent the location of the face after Exif metadata is used to correct the orientation. Images in .png format don't contain Exif metadata.

", + "CompareFacesResponse$TargetImageOrientationCorrection":"

The orientation of the target image (in counterclockwise direction). If your application displays the target image, you can use this value to correct the orientation of the image. The bounding box coordinates returned in FaceMatches and UnmatchedFaces represent face locations before the image orientation is corrected.

If the target image is in .jpg format, it might contain Exif metadata that includes the orientation of the image. If the Exif metadata for the target image populates the orientation field, the value of OrientationCorrection is null and the bounding box coordinates in FaceMatches and UnmatchedFaces represent the location of the face after Exif metadata is used to correct the orientation. Images in .png format don't contain Exif metadata.

", + "DetectFacesResponse$OrientationCorrection":"

The orientation of the input image (counter-clockwise direction). If your application displays the image, you can use this value to correct image orientation. The bounding box coordinates returned in FaceDetails represent face locations before the image orientation is corrected.

If the input image is in .jpeg format, it might contain exchangeable image (Exif) metadata that includes the image's orientation. If so, and the Exif metadata for the input image populates the orientation field, the value of OrientationCorrection is null and the FaceDetails bounding box coordinates represent face locations after Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata.

", + "DetectLabelsResponse$OrientationCorrection":"

The orientation of the input image (counter-clockwise direction). If your application displays the image, you can use this value to correct the orientation. If Amazon Rekognition detects that the input image was rotated (for example, by 90 degrees), it first corrects the orientation before detecting the labels.

If the input image Exif metadata populates the orientation field, Amazon Rekognition does not perform orientation correction and the value of OrientationCorrection will be null.

", + "IndexFacesResponse$OrientationCorrection":"

The orientation of the input image (counterclockwise direction). If your application displays the image, you can use this value to correct image orientation. The bounding box coordinates returned in FaceRecords represent face locations before the image orientation is corrected.

If the input image is in jpeg format, it might contain exchangeable image (Exif) metadata. If so, and the Exif metadata populates the orientation field, the value of OrientationCorrection is null and the bounding box coordinates in FaceRecords represent face locations after Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata.

", + "RecognizeCelebritiesResponse$OrientationCorrection":"

The orientation of the input image (counterclockwise direction). If your application displays the image, you can use this value to correct the orientation. The bounding box coordinates returned in CelebrityFaces and UnrecognizedFaces represent face locations before the image orientation is corrected.

If the input image is in .jpeg format, it might contain exchangeable image (Exif) metadata that includes the image's orientation. If so, and the Exif metadata for the input image populates the orientation field, the value of OrientationCorrection is null and the CelebrityFaces and UnrecognizedFaces bounding box coordinates represent face locations after Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata.

" + } + }, + "PageSize":{ + "base":null, + "refs":{ + "ListCollectionsRequest$MaxResults":"

Maximum number of collection IDs to return.

", + "ListFacesRequest$MaxResults":"

Maximum number of faces to return.

" + } + }, + "PaginationToken":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionRequest$NextToken":"

If the previous response was incomplete (because there is more recognized celebrities to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of celebrities.

", + "GetCelebrityRecognitionResponse$NextToken":"

If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of celebrities.

", + "GetContentModerationRequest$NextToken":"

If the previous response was incomplete (because there is more data to retrieve), Amazon Rekognition returns a pagination token in the response. You can use this pagination token to retrieve the next set of content moderation labels.

", + "GetContentModerationResponse$NextToken":"

If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of moderation labels.

", + "GetFaceDetectionRequest$NextToken":"

If the previous response was incomplete (because there are more faces to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of faces.

", + "GetFaceDetectionResponse$NextToken":"

If the response is truncated, Amazon Rekognition returns this token that you can use in the subsequent request to retrieve the next set of faces.

", + "GetFaceSearchRequest$NextToken":"

If the previous response was incomplete (because there is more search results to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of search results.

", + "GetFaceSearchResponse$NextToken":"

If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of search results.

", + "GetLabelDetectionRequest$NextToken":"

If the previous response was incomplete (because there are more labels to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of labels.

", + "GetLabelDetectionResponse$NextToken":"

If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of labels.

", + "GetPersonTrackingRequest$NextToken":"

If the previous response was incomplete (because there are more persons to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of persons.

", + "GetPersonTrackingResponse$NextToken":"

If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of persons.

", + "ListCollectionsRequest$NextToken":"

Pagination token from the previous response.

", + "ListCollectionsResponse$NextToken":"

If the result is truncated, the response provides a NextToken that you can use in the subsequent request to fetch the next set of collection IDs.

", + "ListFacesRequest$NextToken":"

If the previous response was incomplete (because there is more data to retrieve), Amazon Rekognition returns a pagination token in the response. You can use this pagination token to retrieve the next set of faces.

", + "ListStreamProcessorsRequest$NextToken":"

If the previous response was incomplete (because there are more stream processors to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of stream processors.

", + "ListStreamProcessorsResponse$NextToken":"

If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of stream processors.

" + } + }, + "Percent":{ + "base":null, + "refs":{ + "Beard$Confidence":"

Level of confidence in the determination.

", + "Celebrity$MatchConfidence":"

The confidence, in percentage, that Rekognition has that the recognized face is the celebrity.

", + "CelebrityDetail$Confidence":"

The confidence, in percentage, that Amazon Rekognition has that the recognized face is the celebrity.

", + "CompareFacesMatch$Similarity":"

Level of confidence that the faces match.

", + "CompareFacesRequest$SimilarityThreshold":"

The minimum level of confidence in the face matches that a match must meet to be included in the FaceMatches array.

", + "ComparedFace$Confidence":"

Level of confidence that what the bounding box contains is a face.

", + "ComparedSourceImageFace$Confidence":"

Confidence level that the selected bounding box contains a face.

", + "DetectLabelsRequest$MinConfidence":"

Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn't return any labels with confidence lower than this specified value.

If MinConfidence is not specified, the operation returns labels with a confidence values greater than or equal to 50 percent.

", + "DetectModerationLabelsRequest$MinConfidence":"

Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn't return any labels with a confidence level lower than this specified value.

If you don't specify MinConfidence, the operation returns labels with confidence values greater than or equal to 50 percent.

", + "Emotion$Confidence":"

Level of confidence in the determination.

", + "EyeOpen$Confidence":"

Level of confidence in the determination.

", + "Eyeglasses$Confidence":"

Level of confidence in the determination.

", + "Face$Confidence":"

Confidence level that the bounding box contains a face (and not a different object such as a tree).

", + "FaceDetail$Confidence":"

Confidence level that the bounding box contains a face (and not a different object such as a tree). Default attribute.

", + "FaceMatch$Similarity":"

Confidence in the match of this face with the input face.

", + "FaceSearchSettings$FaceMatchThreshold":"

Minimum face match confidence score that must be met to return a result for a recognized face. Default is 70. 0 is the lowest confidence. 100 is the highest confidence.

", + "Gender$Confidence":"

Level of confidence in the determination.

", + "Label$Confidence":"

Level of confidence.

", + "ModerationLabel$Confidence":"

Specifies the confidence that Amazon Rekognition has that the label has been correctly identified.

If you don't specify the MinConfidence parameter in the call to DetectModerationLabels, the operation returns labels with a confidence value greater than or equal to 50 percent.

", + "MouthOpen$Confidence":"

Level of confidence in the determination.

", + "Mustache$Confidence":"

Level of confidence in the determination.

", + "SearchFacesByImageRequest$FaceMatchThreshold":"

(Optional) Specifies the minimum confidence in the face match to return. For example, don't return any matches where confidence in matches is less than 70%.

", + "SearchFacesByImageResponse$SearchedFaceConfidence":"

The level of confidence that the searchedFaceBoundingBox, contains a face.

", + "SearchFacesRequest$FaceMatchThreshold":"

Optional value specifying the minimum confidence in the face match to return. For example, don't return any matches where confidence in matches is less than 70%.

", + "Smile$Confidence":"

Level of confidence in the determination.

", + "StartContentModerationRequest$MinConfidence":"

Specifies the minimum confidence that Amazon Rekognition must have in order to return a moderated content label. Confidence represents how certain Amazon Rekognition is that the moderated content is correctly identified. 0 is the lowest confidence. 100 is the highest confidence. Amazon Rekognition doesn't return any moderated content labels with a confidence level lower than this specified value.

", + "StartFaceSearchRequest$FaceMatchThreshold":"

The minimum confidence in the person match to return. For example, don't return any matches where confidence in matches is less than 70%.

", + "StartLabelDetectionRequest$MinConfidence":"

Specifies the minimum confidence that Amazon Rekognition Video must have in order to return a detected label. Confidence represents how certain Amazon Rekognition is that a label is correctly identified.0 is the lowest confidence. 100 is the highest confidence. Amazon Rekognition Video doesn't return any labels with a confidence level lower than this specified value.

If you don't specify MinConfidence, the operation returns labels with confidence values greater than or equal to 50 percent.

", + "Sunglasses$Confidence":"

Level of confidence in the determination.

", + "TextDetection$Confidence":"

The confidence that Amazon Rekognition has in the accuracy of the detected text and the accuracy of the geometry points around the detected text.

" + } }, - "PersonDetail": { - "base": "

Details about a person detected in a video analysis request.

", - "refs": { - "PersonDetection$Person": "

Details about a person tracked in a video.

", - "PersonMatch$Person": "

Information about the matched person.

" - } - }, - "PersonDetection": { - "base": "

Details and tracking information for a single time a person is tracked in a video. Amazon Rekognition operations that track persons return an array of PersonDetection objects with elements for each time a person is tracked in a video. For more information, see .

", - "refs": { - "PersonDetections$member": null - } - }, - "PersonDetections": { - "base": null, - "refs": { - "GetPersonTrackingResponse$Persons": "

An array of the persons detected in the video and the times they are tracked throughout the video. An array element will exist for each time the person is tracked.

" - } - }, - "PersonIndex": { - "base": null, - "refs": { - "PersonDetail$Index": "

Identifier for the person detected person within a video. Use to keep track of the person throughout the video. The identifier is not stored by Amazon Rekognition.

" + "PersonDetail":{ + "base":"

Details about a person detected in a video analysis request.

", + "refs":{ + "PersonDetection$Person":"

Details about a person tracked in a video.

", + "PersonMatch$Person":"

Information about the matched person.

" + } + }, + "PersonDetection":{ + "base":"

Details and tracking information for a single time a person is tracked in a video. Amazon Rekognition operations that track persons return an array of PersonDetection objects with elements for each time a person is tracked in a video.

For more information, see API_GetPersonTracking in the Amazon Rekognition Developer Guide.

", + "refs":{ + "PersonDetections$member":null + } + }, + "PersonDetections":{ + "base":null, + "refs":{ + "GetPersonTrackingResponse$Persons":"

An array of the persons detected in the video and the times they are tracked throughout the video. An array element will exist for each time the person is tracked.

" + } + }, + "PersonIndex":{ + "base":null, + "refs":{ + "PersonDetail$Index":"

Identifier for the person detected person within a video. Use to keep track of the person throughout the video. The identifier is not stored by Amazon Rekognition.

" } - }, - "PersonMatch": { - "base": "

Information about a person whose face matches a face(s) in a Amazon Rekognition collection. Includes information about the faces in the Amazon Rekognition collection (, information about the person (PersonDetail) and the timestamp for when the person was detected in a video. An array of PersonMatch objects is returned by .

", - "refs": { - "PersonMatches$member": null + }, + "PersonMatch":{ + "base":"

Information about a person whose face matches a face(s) in a Amazon Rekognition collection. Includes information about the faces in the Amazon Rekognition collection (, information about the person (PersonDetail) and the timestamp for when the person was detected in a video. An array of PersonMatch objects is returned by .

", + "refs":{ + "PersonMatches$member":null } }, - "PersonMatches": { - "base": null, - "refs": { - "GetFaceSearchResponse$Persons": "

An array of persons, , in the video whose face(s) match the face(s) in an Amazon Rekognition collection. It also includes time information for when persons are matched in the video. You specify the input collection in an initial call to StartFaceSearch. Each Persons element includes a time the person was matched, face match details (FaceMatches) for matching faces in the collection, and person information (Person) for the matched person.

" + "PersonMatches":{ + "base":null, + "refs":{ + "GetFaceSearchResponse$Persons":"

An array of persons, , in the video whose face(s) match the face(s) in an Amazon Rekognition collection. It also includes time information for when persons are matched in the video. You specify the input collection in an initial call to StartFaceSearch. Each Persons element includes a time the person was matched, face match details (FaceMatches) for matching faces in the collection, and person information (Person) for the matched person.

" } }, - "PersonTrackingSortBy": { - "base": null, - "refs": { - "GetPersonTrackingRequest$SortBy": "

Sort to use for elements in the Persons array. Use TIMESTAMP to sort array elements by the time persons are detected. Use INDEX to sort by the tracked persons. If you sort by INDEX, the array elements for each person are sorted by detection confidence. The default sort is by TIMESTAMP.

" + "PersonTrackingSortBy":{ + "base":null, + "refs":{ + "GetPersonTrackingRequest$SortBy":"

Sort to use for elements in the Persons array. Use TIMESTAMP to sort array elements by the time persons are detected. Use INDEX to sort by the tracked persons. If you sort by INDEX, the array elements for each person are sorted by detection confidence. The default sort is by TIMESTAMP.

" } }, - "Point": { - "base": "

The X and Y coordinates of a point on an image. The X and Y values returned are ratios of the overall image size. For example, if the input image is 700x200 and the operation returns X=0.5 and Y=0.25, then the point is at the (350,50) pixel coordinate on the image.

An array of Point objects, Polygon, is returned by . Polygon represents a fine-grained polygon around detected text. For more information, see .

", - "refs": { - "Polygon$member": null + "Point":{ + "base":"

The X and Y coordinates of a point on an image. The X and Y values returned are ratios of the overall image size. For example, if the input image is 700x200 and the operation returns X=0.5 and Y=0.25, then the point is at the (350,50) pixel coordinate on the image.

An array of Point objects, Polygon, is returned by . Polygon represents a fine-grained polygon around detected text. For more information, see Geometry in the Amazon Rekognition Developer Guide.

", + "refs":{ + "Polygon$member":null } }, - "Polygon": { - "base": null, - "refs": { - "Geometry$Polygon": "

Within the bounding box, a fine-grained polygon around the detected text.

" + "Polygon":{ + "base":null, + "refs":{ + "Geometry$Polygon":"

Within the bounding box, a fine-grained polygon around the detected text.

" } }, - "Pose": { - "base": "

Indicates the pose of the face as determined by its pitch, roll, and yaw.

", - "refs": { - "ComparedFace$Pose": "

Indicates the pose of the face as determined by its pitch, roll, and yaw.

", - "FaceDetail$Pose": "

Indicates the pose of the face as determined by its pitch, roll, and yaw. Default attribute.

" + "Pose":{ + "base":"

Indicates the pose of the face as determined by its pitch, roll, and yaw.

", + "refs":{ + "ComparedFace$Pose":"

Indicates the pose of the face as determined by its pitch, roll, and yaw.

", + "FaceDetail$Pose":"

Indicates the pose of the face as determined by its pitch, roll, and yaw. Default attribute.

" } }, - "ProvisionedThroughputExceededException": { - "base": "

The number of requests exceeded your throughput limit. If you want to increase this limit, contact Amazon Rekognition.

", - "refs": { + "ProvisionedThroughputExceededException":{ + "base":"

The number of requests exceeded your throughput limit. If you want to increase this limit, contact Amazon Rekognition.

", + "refs":{ } }, - "RecognizeCelebritiesRequest": { - "base": null, - "refs": { + "RecognizeCelebritiesRequest":{ + "base":null, + "refs":{ } }, - "RecognizeCelebritiesResponse": { - "base": null, - "refs": { + "RecognizeCelebritiesResponse":{ + "base":null, + "refs":{ } }, - "RekognitionUniqueId": { - "base": null, - "refs": { - "Celebrity$Id": "

A unique identifier for the celebrity.

", - "CelebrityDetail$Id": "

The unique identifier for the celebrity.

", - "GetCelebrityInfoRequest$Id": "

The ID for the celebrity. You get the celebrity ID from a call to the operation, which recognizes celebrities in an image.

" + "RekognitionUniqueId":{ + "base":null, + "refs":{ + "Celebrity$Id":"

A unique identifier for the celebrity.

", + "CelebrityDetail$Id":"

The unique identifier for the celebrity.

", + "GetCelebrityInfoRequest$Id":"

The ID for the celebrity. You get the celebrity ID from a call to the operation, which recognizes celebrities in an image.

" } }, - "ResourceAlreadyExistsException": { - "base": "

A collection with the specified ID already exists.

", - "refs": { + "ResourceAlreadyExistsException":{ + "base":"

A collection with the specified ID already exists.

", + "refs":{ } }, - "ResourceInUseException": { - "base": "

", - "refs": { + "ResourceInUseException":{ + "base":"

", + "refs":{ } }, - "ResourceNotFoundException": { - "base": "

The collection specified in the request cannot be found.

", - "refs": { + "ResourceNotFoundException":{ + "base":"

The collection specified in the request cannot be found.

", + "refs":{ } }, - "RoleArn": { - "base": null, - "refs": { - "CreateStreamProcessorRequest$RoleArn": "

ARN of the IAM role that allows access to the stream processor.

", - "DescribeStreamProcessorResponse$RoleArn": "

ARN of the IAM role that allows access to the stream processor.

", - "NotificationChannel$RoleArn": "

The ARN of an IAM role that gives Amazon Rekognition publishing permissions to the Amazon SNS topic.

" + "RoleArn":{ + "base":null, + "refs":{ + "CreateStreamProcessorRequest$RoleArn":"

ARN of the IAM role that allows access to the stream processor.

", + "DescribeStreamProcessorResponse$RoleArn":"

ARN of the IAM role that allows access to the stream processor.

", + "NotificationChannel$RoleArn":"

The ARN of an IAM role that gives Amazon Rekognition publishing permissions to the Amazon SNS topic.

" } }, - "S3Bucket": { - "base": null, - "refs": { - "S3Object$Bucket": "

Name of the S3 bucket.

" + "S3Bucket":{ + "base":null, + "refs":{ + "S3Object$Bucket":"

Name of the S3 bucket.

" } }, - "S3Object": { - "base": "

Provides the S3 bucket name and object name.

The region for the S3 bucket containing the S3 object must match the region you use for Amazon Rekognition operations.

For Amazon Rekognition to process an S3 object, the user must have permission to access the S3 object. For more information, see manage-access-resource-policies.

", - "refs": { - "Image$S3Object": "

Identifies an S3 object as the image source.

", - "Video$S3Object": "

The Amazon S3 bucket name and file name for the video.

" + "S3Object":{ + "base":"

Provides the S3 bucket name and object name.

The region for the S3 bucket containing the S3 object must match the region you use for Amazon Rekognition operations.

For Amazon Rekognition to process an S3 object, the user must have permission to access the S3 object. For more information, see Resource Based Policies in the Amazon Rekognition Developer Guide.

", + "refs":{ + "Image$S3Object":"

Identifies an S3 object as the image source.

", + "Video$S3Object":"

The Amazon S3 bucket name and file name for the video.

" } }, - "S3ObjectName": { - "base": null, - "refs": { - "S3Object$Name": "

S3 object key name.

" + "S3ObjectName":{ + "base":null, + "refs":{ + "S3Object$Name":"

S3 object key name.

" } }, - "S3ObjectVersion": { - "base": null, - "refs": { - "S3Object$Version": "

If the bucket is versioning enabled, you can specify the object version.

" + "S3ObjectVersion":{ + "base":null, + "refs":{ + "S3Object$Version":"

If the bucket is versioning enabled, you can specify the object version.

" } }, - "SNSTopicArn": { - "base": null, - "refs": { - "NotificationChannel$SNSTopicArn": "

The Amazon SNS topic to which Amazon Rekognition to posts the completion status.

" + "SNSTopicArn":{ + "base":null, + "refs":{ + "NotificationChannel$SNSTopicArn":"

The Amazon SNS topic to which Amazon Rekognition to posts the completion status.

" } }, - "SearchFacesByImageRequest": { - "base": null, - "refs": { + "SearchFacesByImageRequest":{ + "base":null, + "refs":{ } }, - "SearchFacesByImageResponse": { - "base": null, - "refs": { + "SearchFacesByImageResponse":{ + "base":null, + "refs":{ } }, - "SearchFacesRequest": { - "base": null, - "refs": { + "SearchFacesRequest":{ + "base":null, + "refs":{ } }, - "SearchFacesResponse": { - "base": null, - "refs": { + "SearchFacesResponse":{ + "base":null, + "refs":{ } }, - "Smile": { - "base": "

Indicates whether or not the face is smiling, and the confidence level in the determination.

", - "refs": { - "FaceDetail$Smile": "

Indicates whether or not the face is smiling, and the confidence level in the determination.

" + "Smile":{ + "base":"

Indicates whether or not the face is smiling, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$Smile":"

Indicates whether or not the face is smiling, and the confidence level in the determination.

" } }, - "StartCelebrityRecognitionRequest": { - "base": null, - "refs": { + "StartCelebrityRecognitionRequest":{ + "base":null, + "refs":{ } }, - "StartCelebrityRecognitionResponse": { - "base": null, - "refs": { + "StartCelebrityRecognitionResponse":{ + "base":null, + "refs":{ } }, - "StartContentModerationRequest": { - "base": null, - "refs": { + "StartContentModerationRequest":{ + "base":null, + "refs":{ } }, - "StartContentModerationResponse": { - "base": null, - "refs": { + "StartContentModerationResponse":{ + "base":null, + "refs":{ } }, - "StartFaceDetectionRequest": { - "base": null, - "refs": { + "StartFaceDetectionRequest":{ + "base":null, + "refs":{ } }, - "StartFaceDetectionResponse": { - "base": null, - "refs": { + "StartFaceDetectionResponse":{ + "base":null, + "refs":{ } }, - "StartFaceSearchRequest": { - "base": null, - "refs": { + "StartFaceSearchRequest":{ + "base":null, + "refs":{ } }, - "StartFaceSearchResponse": { - "base": null, - "refs": { + "StartFaceSearchResponse":{ + "base":null, + "refs":{ } }, - "StartLabelDetectionRequest": { - "base": null, - "refs": { + "StartLabelDetectionRequest":{ + "base":null, + "refs":{ } }, - "StartLabelDetectionResponse": { - "base": null, - "refs": { + "StartLabelDetectionResponse":{ + "base":null, + "refs":{ } }, - "StartPersonTrackingRequest": { - "base": null, - "refs": { + "StartPersonTrackingRequest":{ + "base":null, + "refs":{ } }, - "StartPersonTrackingResponse": { - "base": null, - "refs": { + "StartPersonTrackingResponse":{ + "base":null, + "refs":{ } }, - "StartStreamProcessorRequest": { - "base": null, - "refs": { + "StartStreamProcessorRequest":{ + "base":null, + "refs":{ } }, - "StartStreamProcessorResponse": { - "base": null, - "refs": { + "StartStreamProcessorResponse":{ + "base":null, + "refs":{ } }, - "StatusMessage": { - "base": null, - "refs": { - "GetCelebrityRecognitionResponse$StatusMessage": "

If the job fails, StatusMessage provides a descriptive error message.

", - "GetContentModerationResponse$StatusMessage": "

If the job fails, StatusMessage provides a descriptive error message.

", - "GetFaceDetectionResponse$StatusMessage": "

If the job fails, StatusMessage provides a descriptive error message.

", - "GetFaceSearchResponse$StatusMessage": "

If the job fails, StatusMessage provides a descriptive error message.

", - "GetLabelDetectionResponse$StatusMessage": "

If the job fails, StatusMessage provides a descriptive error message.

", - "GetPersonTrackingResponse$StatusMessage": "

If the job fails, StatusMessage provides a descriptive error message.

" + "StatusMessage":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionResponse$StatusMessage":"

If the job fails, StatusMessage provides a descriptive error message.

", + "GetContentModerationResponse$StatusMessage":"

If the job fails, StatusMessage provides a descriptive error message.

", + "GetFaceDetectionResponse$StatusMessage":"

If the job fails, StatusMessage provides a descriptive error message.

", + "GetFaceSearchResponse$StatusMessage":"

If the job fails, StatusMessage provides a descriptive error message.

", + "GetLabelDetectionResponse$StatusMessage":"

If the job fails, StatusMessage provides a descriptive error message.

", + "GetPersonTrackingResponse$StatusMessage":"

If the job fails, StatusMessage provides a descriptive error message.

" } }, - "StopStreamProcessorRequest": { - "base": null, - "refs": { + "StopStreamProcessorRequest":{ + "base":null, + "refs":{ } }, - "StopStreamProcessorResponse": { - "base": null, - "refs": { + "StopStreamProcessorResponse":{ + "base":null, + "refs":{ } }, - "StreamProcessor": { - "base": "

An object that recognizes faces in a streaming video. An Amazon Rekognition stream processor is created by a call to . The request parameters for CreateStreamProcessor describe the Kinesis video stream source for the streaming video, face recognition parameters, and where to stream the analysis resullts.

", - "refs": { - "StreamProcessorList$member": null + "StreamProcessor":{ + "base":"

An object that recognizes faces in a streaming video. An Amazon Rekognition stream processor is created by a call to . The request parameters for CreateStreamProcessor describe the Kinesis video stream source for the streaming video, face recognition parameters, and where to stream the analysis resullts.

", + "refs":{ + "StreamProcessorList$member":null } }, - "StreamProcessorArn": { - "base": null, - "refs": { - "CreateStreamProcessorResponse$StreamProcessorArn": "

ARN for the newly create stream processor.

", - "DescribeStreamProcessorResponse$StreamProcessorArn": "

ARN of the stream processor.

" + "StreamProcessorArn":{ + "base":null, + "refs":{ + "CreateStreamProcessorResponse$StreamProcessorArn":"

ARN for the newly create stream processor.

", + "DescribeStreamProcessorResponse$StreamProcessorArn":"

ARN of the stream processor.

" } }, - "StreamProcessorInput": { - "base": "

Information about the source streaming video.

", - "refs": { - "CreateStreamProcessorRequest$Input": "

Kinesis video stream stream that provides the source streaming video. If you are using the AWS CLI, the parameter name is StreamProcessorInput.

", - "DescribeStreamProcessorResponse$Input": "

Kinesis video stream that provides the source streaming video.

" + "StreamProcessorInput":{ + "base":"

Information about the source streaming video.

", + "refs":{ + "CreateStreamProcessorRequest$Input":"

Kinesis video stream stream that provides the source streaming video. If you are using the AWS CLI, the parameter name is StreamProcessorInput.

", + "DescribeStreamProcessorResponse$Input":"

Kinesis video stream that provides the source streaming video.

" } }, - "StreamProcessorList": { - "base": null, - "refs": { - "ListStreamProcessorsResponse$StreamProcessors": "

List of stream processors that you have created.

" + "StreamProcessorList":{ + "base":null, + "refs":{ + "ListStreamProcessorsResponse$StreamProcessors":"

List of stream processors that you have created.

" } }, - "StreamProcessorName": { - "base": null, - "refs": { - "CreateStreamProcessorRequest$Name": "

An identifier you assign to the stream processor. You can use Name to manage the stream processor. For example, you can get the current status of the stream processor by calling . Name is idempotent.

", - "DeleteStreamProcessorRequest$Name": "

The name of the stream processor you want to delete.

", - "DescribeStreamProcessorRequest$Name": "

Name of the stream processor for which you want information.

", - "DescribeStreamProcessorResponse$Name": "

Name of the stream processor.

", - "StartStreamProcessorRequest$Name": "

The name of the stream processor to start processing.

", - "StopStreamProcessorRequest$Name": "

The name of a stream processor created by .

", - "StreamProcessor$Name": "

Name of the Amazon Rekognition stream processor.

" + "StreamProcessorName":{ + "base":null, + "refs":{ + "CreateStreamProcessorRequest$Name":"

An identifier you assign to the stream processor. You can use Name to manage the stream processor. For example, you can get the current status of the stream processor by calling . Name is idempotent.

", + "DeleteStreamProcessorRequest$Name":"

The name of the stream processor you want to delete.

", + "DescribeStreamProcessorRequest$Name":"

Name of the stream processor for which you want information.

", + "DescribeStreamProcessorResponse$Name":"

Name of the stream processor.

", + "StartStreamProcessorRequest$Name":"

The name of the stream processor to start processing.

", + "StopStreamProcessorRequest$Name":"

The name of a stream processor created by .

", + "StreamProcessor$Name":"

Name of the Amazon Rekognition stream processor.

" } }, - "StreamProcessorOutput": { - "base": "

Information about the Amazon Kinesis Data Streams stream to which a Rekognition Video stream processor streams the results of a video analysis. For more information, see .

", - "refs": { - "CreateStreamProcessorRequest$Output": "

Kinesis data stream stream to which Rekognition Video puts the analysis results. If you are using the AWS CLI, the parameter name is StreamProcessorOutput.

", - "DescribeStreamProcessorResponse$Output": "

Kinesis data stream to which Rekognition Video puts the analysis results.

" + "StreamProcessorOutput":{ + "base":"

Information about the Amazon Kinesis Data Streams stream to which a Amazon Rekognition Video stream processor streams the results of a video analysis. For more information, see CreateStreamProcessor in the Amazon Rekognition Developer Guide.

", + "refs":{ + "CreateStreamProcessorRequest$Output":"

Kinesis data stream stream to which Amazon Rekognition Video puts the analysis results. If you are using the AWS CLI, the parameter name is StreamProcessorOutput.

", + "DescribeStreamProcessorResponse$Output":"

Kinesis data stream to which Amazon Rekognition Video puts the analysis results.

" } }, - "StreamProcessorSettings": { - "base": "

Input parameters used to recognize faces in a streaming video analyzed by a Amazon Rekognition stream processor.

", - "refs": { - "CreateStreamProcessorRequest$Settings": "

Face recognition input parameters to be used by the stream processor. Includes the collection to use for face recognition and the face attributes to detect.

", - "DescribeStreamProcessorResponse$Settings": "

Face recognition input parameters that are being used by the stream processor. Includes the collection to use for face recognition and the face attributes to detect.

" + "StreamProcessorSettings":{ + "base":"

Input parameters used to recognize faces in a streaming video analyzed by a Amazon Rekognition stream processor.

", + "refs":{ + "CreateStreamProcessorRequest$Settings":"

Face recognition input parameters to be used by the stream processor. Includes the collection to use for face recognition and the face attributes to detect.

", + "DescribeStreamProcessorResponse$Settings":"

Face recognition input parameters that are being used by the stream processor. Includes the collection to use for face recognition and the face attributes to detect.

" } }, - "StreamProcessorStatus": { - "base": null, - "refs": { - "DescribeStreamProcessorResponse$Status": "

Current status of the stream processor.

", - "StreamProcessor$Status": "

Current status of the Amazon Rekognition stream processor.

" + "StreamProcessorStatus":{ + "base":null, + "refs":{ + "DescribeStreamProcessorResponse$Status":"

Current status of the stream processor.

", + "StreamProcessor$Status":"

Current status of the Amazon Rekognition stream processor.

" } }, - "String": { - "base": null, - "refs": { - "Celebrity$Name": "

The name of the celebrity.

", - "CelebrityDetail$Name": "

The name of the celebrity.

", - "CreateCollectionResponse$CollectionArn": "

Amazon Resource Name (ARN) of the collection. You can use this to manage permissions on your resources.

", - "CreateCollectionResponse$FaceModelVersion": "

Version number of the face detection model associated with the collection you are creating.

", - "DescribeStreamProcessorResponse$StatusMessage": "

Detailed status message about the stream processor.

", - "FaceModelVersionList$member": null, - "GetCelebrityInfoResponse$Name": "

The name of the celebrity.

", - "IndexFacesResponse$FaceModelVersion": "

Version number of the face detection model associated with the input collection (CollectionId).

", - "Label$Name": "

The name (label) of the object.

", - "ListFacesResponse$NextToken": "

If the response is truncated, Amazon Rekognition returns this token that you can use in the subsequent request to retrieve the next set of faces.

", - "ListFacesResponse$FaceModelVersion": "

Version number of the face detection model associated with the input collection (CollectionId).

", - "ModerationLabel$Name": "

The label name for the type of content detected in the image.

", - "ModerationLabel$ParentName": "

The name for the parent label. Labels at the top-level of the hierarchy have the parent label \"\".

", - "SearchFacesByImageResponse$FaceModelVersion": "

Version number of the face detection model associated with the input collection (CollectionId).

", - "SearchFacesResponse$FaceModelVersion": "

Version number of the face detection model associated with the input collection (CollectionId).

", - "TextDetection$DetectedText": "

The word or line of text recognized by Amazon Rekognition.

", - "VideoMetadata$Codec": "

Type of compression used in the analyzed video.

", - "VideoMetadata$Format": "

Format of the analyzed video. Possible values are MP4, MOV and AVI.

" - } - }, - "Sunglasses": { - "base": "

Indicates whether or not the face is wearing sunglasses, and the confidence level in the determination.

", - "refs": { - "FaceDetail$Sunglasses": "

Indicates whether or not the face is wearing sunglasses, and the confidence level in the determination.

" - } - }, - "TextDetection": { - "base": "

Information about a word or line of text detected by .

The DetectedText field contains the text that Amazon Rekognition detected in the image.

Every word and line has an identifier (Id). Each word belongs to a line and has a parent identifier (ParentId) that identifies the line of text in which the word appears. The word Id is also an index for the word within a line of words.

For more information, see text-detection.

", - "refs": { - "TextDetectionList$member": null - } - }, - "TextDetectionList": { - "base": null, - "refs": { - "DetectTextResponse$TextDetections": "

An array of text that was detected in the input image.

" - } - }, - "TextTypes": { - "base": null, - "refs": { - "TextDetection$Type": "

The type of text that was detected.

" - } - }, - "ThrottlingException": { - "base": "

Amazon Rekognition is temporarily unable to process the request. Try your call again.

", - "refs": { - } - }, - "Timestamp": { - "base": null, - "refs": { - "CelebrityRecognition$Timestamp": "

The time, in milliseconds from the start of the video, that the celebrity was recognized.

", - "ContentModerationDetection$Timestamp": "

Time, in milliseconds from the beginning of the video, that the moderation label was detected.

", - "FaceDetection$Timestamp": "

Time, in milliseconds from the start of the video, that the face was detected.

", - "LabelDetection$Timestamp": "

Time, in milliseconds from the start of the video, that the label was detected.

", - "PersonDetection$Timestamp": "

The time, in milliseconds from the start of the video, that the person was tracked.

", - "PersonMatch$Timestamp": "

The time, in milliseconds from the beginning of the video, that the person was matched in the video.

" - } - }, - "UInteger": { - "base": null, - "refs": { - "AgeRange$Low": "

The lowest estimated age.

", - "AgeRange$High": "

The highest estimated age.

", - "CreateCollectionResponse$StatusCode": "

HTTP status code indicating the result of the operation.

", - "DeleteCollectionResponse$StatusCode": "

HTTP status code that indicates the result of the operation.

", - "DetectLabelsRequest$MaxLabels": "

Maximum number of labels you want the service to return in the response. The service returns the specified number of highest confidence labels.

", - "TextDetection$Id": "

The identifier for the detected text. The identifier is only unique for a single call to DetectText.

", - "TextDetection$ParentId": "

The Parent identifier for the detected text identified by the value of ID. If the type of detected text is LINE, the value of ParentId is Null.

" - } - }, - "ULong": { - "base": null, - "refs": { - "VideoMetadata$DurationMillis": "

Length of the video in milliseconds.

", - "VideoMetadata$FrameHeight": "

Vertical pixel dimension of the video.

", - "VideoMetadata$FrameWidth": "

Horizontal pixel dimension of the video.

" - } - }, - "Url": { - "base": null, - "refs": { - "Urls$member": null - } - }, - "Urls": { - "base": null, - "refs": { - "Celebrity$Urls": "

An array of URLs pointing to additional information about the celebrity. If there is no additional information about the celebrity, this list is empty.

", - "CelebrityDetail$Urls": "

An array of URLs pointing to additional celebrity information.

", - "GetCelebrityInfoResponse$Urls": "

An array of URLs pointing to additional celebrity information.

" - } - }, - "Video": { - "base": "

Video file stored in an Amazon S3 bucket. Amazon Rekognition video start operations such as use Video to specify a video for analysis. The supported file formats are .mp4, .mov and .avi.

", - "refs": { - "StartCelebrityRecognitionRequest$Video": "

The video in which you want to recognize celebrities. The video must be stored in an Amazon S3 bucket.

", - "StartContentModerationRequest$Video": "

The video in which you want to moderate content. The video must be stored in an Amazon S3 bucket.

", - "StartFaceDetectionRequest$Video": "

The video in which you want to detect faces. The video must be stored in an Amazon S3 bucket.

", - "StartFaceSearchRequest$Video": "

The video you want to search. The video must be stored in an Amazon S3 bucket.

", - "StartLabelDetectionRequest$Video": "

The video in which you want to detect labels. The video must be stored in an Amazon S3 bucket.

", - "StartPersonTrackingRequest$Video": "

The video in which you want to detect people. The video must be stored in an Amazon S3 bucket.

" - } - }, - "VideoJobStatus": { - "base": null, - "refs": { - "GetCelebrityRecognitionResponse$JobStatus": "

The current status of the celebrity recognition job.

", - "GetContentModerationResponse$JobStatus": "

The current status of the content moderation job.

", - "GetFaceDetectionResponse$JobStatus": "

The current status of the face detection job.

", - "GetFaceSearchResponse$JobStatus": "

The current status of the face search job.

", - "GetLabelDetectionResponse$JobStatus": "

The current status of the label detection job.

", - "GetPersonTrackingResponse$JobStatus": "

The current status of the person tracking job.

" - } + "String":{ + "base":null, + "refs":{ + "Celebrity$Name":"

The name of the celebrity.

", + "CelebrityDetail$Name":"

The name of the celebrity.

", + "CreateCollectionResponse$CollectionArn":"

Amazon Resource Name (ARN) of the collection. You can use this to manage permissions on your resources.

", + "CreateCollectionResponse$FaceModelVersion":"

Version number of the face detection model associated with the collection you are creating.

", + "DescribeStreamProcessorResponse$StatusMessage":"

Detailed status message about the stream processor.

", + "FaceModelVersionList$member":null, + "GetCelebrityInfoResponse$Name":"

The name of the celebrity.

", + "IndexFacesResponse$FaceModelVersion":"

Version number of the face detection model associated with the input collection (CollectionId).

", + "Label$Name":"

The name (label) of the object.

", + "ListFacesResponse$NextToken":"

If the response is truncated, Amazon Rekognition returns this token that you can use in the subsequent request to retrieve the next set of faces.

", + "ListFacesResponse$FaceModelVersion":"

Version number of the face detection model associated with the input collection (CollectionId).

", + "ModerationLabel$Name":"

The label name for the type of content detected in the image.

", + "ModerationLabel$ParentName":"

The name for the parent label. Labels at the top-level of the hierarchy have the parent label \"\".

", + "SearchFacesByImageResponse$FaceModelVersion":"

Version number of the face detection model associated with the input collection (CollectionId).

", + "SearchFacesResponse$FaceModelVersion":"

Version number of the face detection model associated with the input collection (CollectionId).

", + "TextDetection$DetectedText":"

The word or line of text recognized by Amazon Rekognition.

", + "VideoMetadata$Codec":"

Type of compression used in the analyzed video.

", + "VideoMetadata$Format":"

Format of the analyzed video. Possible values are MP4, MOV and AVI.

" + } + }, + "Sunglasses":{ + "base":"

Indicates whether or not the face is wearing sunglasses, and the confidence level in the determination.

", + "refs":{ + "FaceDetail$Sunglasses":"

Indicates whether or not the face is wearing sunglasses, and the confidence level in the determination.

" + } + }, + "TextDetection":{ + "base":"

Information about a word or line of text detected by .

The DetectedText field contains the text that Amazon Rekognition detected in the image.

Every word and line has an identifier (Id). Each word belongs to a line and has a parent identifier (ParentId) that identifies the line of text in which the word appears. The word Id is also an index for the word within a line of words.

For more information, see Detecting Text in the Amazon Rekognition Developer Guide.

", + "refs":{ + "TextDetectionList$member":null + } + }, + "TextDetectionList":{ + "base":null, + "refs":{ + "DetectTextResponse$TextDetections":"

An array of text that was detected in the input image.

" + } + }, + "TextTypes":{ + "base":null, + "refs":{ + "TextDetection$Type":"

The type of text that was detected.

" + } + }, + "ThrottlingException":{ + "base":"

Amazon Rekognition is temporarily unable to process the request. Try your call again.

", + "refs":{ + } + }, + "Timestamp":{ + "base":null, + "refs":{ + "CelebrityRecognition$Timestamp":"

The time, in milliseconds from the start of the video, that the celebrity was recognized.

", + "ContentModerationDetection$Timestamp":"

Time, in milliseconds from the beginning of the video, that the moderation label was detected.

", + "FaceDetection$Timestamp":"

Time, in milliseconds from the start of the video, that the face was detected.

", + "LabelDetection$Timestamp":"

Time, in milliseconds from the start of the video, that the label was detected.

", + "PersonDetection$Timestamp":"

The time, in milliseconds from the start of the video, that the person was tracked.

", + "PersonMatch$Timestamp":"

The time, in milliseconds from the beginning of the video, that the person was matched in the video.

" + } + }, + "UInteger":{ + "base":null, + "refs":{ + "AgeRange$Low":"

The lowest estimated age.

", + "AgeRange$High":"

The highest estimated age.

", + "CreateCollectionResponse$StatusCode":"

HTTP status code indicating the result of the operation.

", + "DeleteCollectionResponse$StatusCode":"

HTTP status code that indicates the result of the operation.

", + "DetectLabelsRequest$MaxLabels":"

Maximum number of labels you want the service to return in the response. The service returns the specified number of highest confidence labels.

", + "TextDetection$Id":"

The identifier for the detected text. The identifier is only unique for a single call to DetectText.

", + "TextDetection$ParentId":"

The Parent identifier for the detected text identified by the value of ID. If the type of detected text is LINE, the value of ParentId is Null.

" + } + }, + "ULong":{ + "base":null, + "refs":{ + "VideoMetadata$DurationMillis":"

Length of the video in milliseconds.

", + "VideoMetadata$FrameHeight":"

Vertical pixel dimension of the video.

", + "VideoMetadata$FrameWidth":"

Horizontal pixel dimension of the video.

" + } + }, + "Url":{ + "base":null, + "refs":{ + "Urls$member":null + } + }, + "Urls":{ + "base":null, + "refs":{ + "Celebrity$Urls":"

An array of URLs pointing to additional information about the celebrity. If there is no additional information about the celebrity, this list is empty.

", + "CelebrityDetail$Urls":"

An array of URLs pointing to additional celebrity information.

", + "GetCelebrityInfoResponse$Urls":"

An array of URLs pointing to additional celebrity information.

" + } + }, + "Video":{ + "base":"

Video file stored in an Amazon S3 bucket. Amazon Rekognition video start operations such as use Video to specify a video for analysis. The supported file formats are .mp4, .mov and .avi.

", + "refs":{ + "StartCelebrityRecognitionRequest$Video":"

The video in which you want to recognize celebrities. The video must be stored in an Amazon S3 bucket.

", + "StartContentModerationRequest$Video":"

The video in which you want to moderate content. The video must be stored in an Amazon S3 bucket.

", + "StartFaceDetectionRequest$Video":"

The video in which you want to detect faces. The video must be stored in an Amazon S3 bucket.

", + "StartFaceSearchRequest$Video":"

The video you want to search. The video must be stored in an Amazon S3 bucket.

", + "StartLabelDetectionRequest$Video":"

The video in which you want to detect labels. The video must be stored in an Amazon S3 bucket.

", + "StartPersonTrackingRequest$Video":"

The video in which you want to detect people. The video must be stored in an Amazon S3 bucket.

" + } + }, + "VideoJobStatus":{ + "base":null, + "refs":{ + "GetCelebrityRecognitionResponse$JobStatus":"

The current status of the celebrity recognition job.

", + "GetContentModerationResponse$JobStatus":"

The current status of the content moderation job.

", + "GetFaceDetectionResponse$JobStatus":"

The current status of the face detection job.

", + "GetFaceSearchResponse$JobStatus":"

The current status of the face search job.

", + "GetLabelDetectionResponse$JobStatus":"

The current status of the label detection job.

", + "GetPersonTrackingResponse$JobStatus":"

The current status of the person tracking job.

" + } }, - "VideoMetadata": { - "base": "

Information about a video that Amazon Rekognition analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition video operation.

", - "refs": { - "GetCelebrityRecognitionResponse$VideoMetadata": "

Information about a video that Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Rekognition Video operation.

", - "GetContentModerationResponse$VideoMetadata": "

Information about a video that Amazon Rekognition analyzed. Videometadata is returned in every page of paginated responses from GetContentModeration.

", - "GetFaceDetectionResponse$VideoMetadata": "

Information about a video that Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition video operation.

", - "GetFaceSearchResponse$VideoMetadata": "

Information about a video that Amazon Rekognition analyzed. Videometadata is returned in every page of paginated responses from a Rekognition Video operation.

", - "GetLabelDetectionResponse$VideoMetadata": "

Information about a video that Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition video operation.

", - "GetPersonTrackingResponse$VideoMetadata": "

Information about a video that Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Rekognition Video operation.

" - } - }, - "VideoTooLargeException": { - "base": "

The file size or duration of the supplied media is too large. The maximum file size is 8GB. The maximum duration is 2 hours.

", - "refs": { + "VideoMetadata":{ + "base":"

Information about a video that Amazon Rekognition analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition video operation.

", + "refs":{ + "GetCelebrityRecognitionResponse$VideoMetadata":"

Information about a video that Amazon Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition Video operation.

", + "GetContentModerationResponse$VideoMetadata":"

Information about a video that Amazon Rekognition analyzed. Videometadata is returned in every page of paginated responses from GetContentModeration.

", + "GetFaceDetectionResponse$VideoMetadata":"

Information about a video that Amazon Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition video operation.

", + "GetFaceSearchResponse$VideoMetadata":"

Information about a video that Amazon Rekognition analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition Video operation.

", + "GetLabelDetectionResponse$VideoMetadata":"

Information about a video that Amazon Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition video operation.

", + "GetPersonTrackingResponse$VideoMetadata":"

Information about a video that Amazon Rekognition Video analyzed. Videometadata is returned in every page of paginated responses from a Amazon Rekognition Video operation.

" + } + }, + "VideoTooLargeException":{ + "base":"

The file size or duration of the supplied media is too large. The maximum file size is 8GB. The maximum duration is 2 hours.

", + "refs":{ } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/examples-1.json index 039e04d60..2b6a2e9fb 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/rekognition/2016-06-27/examples-1.json @@ -1,650 +1,650 @@ { - "version": "1.0", - "examples": { - "CompareFaces": [ + "version":"1.0", + "examples":{ + "CompareFaces":[ { - "input": { - "SimilarityThreshold": 90, - "SourceImage": { - "S3Object": { - "Bucket": "mybucket", - "Name": "mysourceimage" + "input":{ + "SimilarityThreshold":90, + "SourceImage":{ + "S3Object":{ + "Bucket":"mybucket", + "Name":"mysourceimage" } }, - "TargetImage": { - "S3Object": { - "Bucket": "mybucket", - "Name": "mytargetimage" + "TargetImage":{ + "S3Object":{ + "Bucket":"mybucket", + "Name":"mytargetimage" } } }, - "output": { - "FaceMatches": [ + "output":{ + "FaceMatches":[ { - "Face": { - "BoundingBox": { - "Height": 0.33481481671333313, - "Left": 0.31888890266418457, - "Top": 0.4933333396911621, - "Width": 0.25 + "Face":{ + "BoundingBox":{ + "Height":0.33481481671333313, + "Left":0.31888890266418457, + "Top":0.4933333396911621, + "Width":0.25 }, - "Confidence": 99.9991226196289 + "Confidence":99.9991226196289 }, - "Similarity": 100 + "Similarity":100 } ], - "SourceImageFace": { - "BoundingBox": { - "Height": 0.33481481671333313, - "Left": 0.31888890266418457, - "Top": 0.4933333396911621, - "Width": 0.25 + "SourceImageFace":{ + "BoundingBox":{ + "Height":0.33481481671333313, + "Left":0.31888890266418457, + "Top":0.4933333396911621, + "Width":0.25 }, - "Confidence": 99.9991226196289 + "Confidence":99.9991226196289 } }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation compares the largest face detected in the source image with each face detected in the target image.", - "id": "to-compare-two-images-1482181985581", - "title": "To compare two images" + "description":"This operation compares the largest face detected in the source image with each face detected in the target image.", + "id":"to-compare-two-images-1482181985581", + "title":"To compare two images" } ], - "CreateCollection": [ + "CreateCollection":[ { - "input": { - "CollectionId": "myphotos" + "input":{ + "CollectionId":"myphotos" }, - "output": { - "CollectionArn": "aws:rekognition:us-west-2:123456789012:collection/myphotos", - "StatusCode": 200 + "output":{ + "CollectionArn":"aws:rekognition:us-west-2:123456789012:collection/myphotos", + "StatusCode":200 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation creates a Rekognition collection for storing image data.", - "id": "to-create-a-collection-1481833313674", - "title": "To create a collection" + "description":"This operation creates a Rekognition collection for storing image data.", + "id":"to-create-a-collection-1481833313674", + "title":"To create a collection" } ], - "DeleteCollection": [ + "DeleteCollection":[ { - "input": { - "CollectionId": "myphotos" + "input":{ + "CollectionId":"myphotos" }, - "output": { - "StatusCode": 200 + "output":{ + "StatusCode":200 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation deletes a Rekognition collection.", - "id": "to-delete-a-collection-1481838179973", - "title": "To delete a collection" + "description":"This operation deletes a Rekognition collection.", + "id":"to-delete-a-collection-1481838179973", + "title":"To delete a collection" } ], - "DeleteFaces": [ + "DeleteFaces":[ { - "input": { - "CollectionId": "myphotos", - "FaceIds": [ + "input":{ + "CollectionId":"myphotos", + "FaceIds":[ "ff43d742-0c13-5d16-a3e8-03d3f58e980b" ] }, - "output": { - "DeletedFaces": [ + "output":{ + "DeletedFaces":[ "ff43d742-0c13-5d16-a3e8-03d3f58e980b" ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation deletes one or more faces from a Rekognition collection.", - "id": "to-delete-a-face-1482182799377", - "title": "To delete a face" + "description":"This operation deletes one or more faces from a Rekognition collection.", + "id":"to-delete-a-face-1482182799377", + "title":"To delete a face" } ], - "DetectFaces": [ + "DetectFaces":[ { - "input": { - "Image": { - "S3Object": { - "Bucket": "mybucket", - "Name": "myphoto" + "input":{ + "Image":{ + "S3Object":{ + "Bucket":"mybucket", + "Name":"myphoto" } } }, - "output": { - "FaceDetails": [ + "output":{ + "FaceDetails":[ { - "BoundingBox": { - "Height": 0.18000000715255737, - "Left": 0.5555555820465088, - "Top": 0.33666667342185974, - "Width": 0.23999999463558197 + "BoundingBox":{ + "Height":0.18000000715255737, + "Left":0.5555555820465088, + "Top":0.33666667342185974, + "Width":0.23999999463558197 }, - "Confidence": 100, - "Landmarks": [ + "Confidence":100, + "Landmarks":[ { - "Type": "eyeLeft", - "X": 0.6394737362861633, - "Y": 0.40819624066352844 + "Type":"eyeLeft", + "X":0.6394737362861633, + "Y":0.40819624066352844 }, { - "Type": "eyeRight", - "X": 0.7266660928726196, - "Y": 0.41039225459098816 + "Type":"eyeRight", + "X":0.7266660928726196, + "Y":0.41039225459098816 }, { - "Type": "eyeRight", - "X": 0.6912462115287781, - "Y": 0.44240960478782654 + "Type":"eyeRight", + "X":0.6912462115287781, + "Y":0.44240960478782654 }, { - "Type": "mouthDown", - "X": 0.6306198239326477, - "Y": 0.46700039505958557 + "Type":"mouthDown", + "X":0.6306198239326477, + "Y":0.46700039505958557 }, { - "Type": "mouthUp", - "X": 0.7215608954429626, - "Y": 0.47114261984825134 + "Type":"mouthUp", + "X":0.7215608954429626, + "Y":0.47114261984825134 } ], - "Pose": { - "Pitch": 4.050806522369385, - "Roll": 0.9950747489929199, - "Yaw": 13.693790435791016 + "Pose":{ + "Pitch":4.050806522369385, + "Roll":0.9950747489929199, + "Yaw":13.693790435791016 }, - "Quality": { - "Brightness": 37.60169982910156, - "Sharpness": 80 + "Quality":{ + "Brightness":37.60169982910156, + "Sharpness":80 } } ], - "OrientationCorrection": "ROTATE_0" + "OrientationCorrection":"ROTATE_0" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation detects faces in an image stored in an AWS S3 bucket.", - "id": "to-detect-faces-in-an-image-1481841782793", - "title": "To detect faces in an image" + "description":"This operation detects faces in an image stored in an AWS S3 bucket.", + "id":"to-detect-faces-in-an-image-1481841782793", + "title":"To detect faces in an image" } ], - "DetectLabels": [ + "DetectLabels":[ { - "input": { - "Image": { - "S3Object": { - "Bucket": "mybucket", - "Name": "myphoto" + "input":{ + "Image":{ + "S3Object":{ + "Bucket":"mybucket", + "Name":"myphoto" } }, - "MaxLabels": 123, - "MinConfidence": 70 + "MaxLabels":123, + "MinConfidence":70 }, - "output": { - "Labels": [ + "output":{ + "Labels":[ { - "Confidence": 99.25072479248047, - "Name": "People" + "Confidence":99.25072479248047, + "Name":"People" }, { - "Confidence": 99.25074005126953, - "Name": "Person" + "Confidence":99.25074005126953, + "Name":"Person" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation detects labels in the supplied image", - "id": "to-detect-labels-1481834255770", - "title": "To detect labels" + "description":"This operation detects labels in the supplied image", + "id":"to-detect-labels-1481834255770", + "title":"To detect labels" } ], - "IndexFaces": [ + "IndexFaces":[ { - "input": { - "CollectionId": "myphotos", - "DetectionAttributes": [ + "input":{ + "CollectionId":"myphotos", + "DetectionAttributes":[ ], - "ExternalImageId": "myphotoid", - "Image": { - "S3Object": { - "Bucket": "mybucket", - "Name": "myphoto" + "ExternalImageId":"myphotoid", + "Image":{ + "S3Object":{ + "Bucket":"mybucket", + "Name":"myphoto" } } }, - "output": { - "FaceRecords": [ + "output":{ + "FaceRecords":[ { - "Face": { - "BoundingBox": { - "Height": 0.33481481671333313, - "Left": 0.31888890266418457, - "Top": 0.4933333396911621, - "Width": 0.25 + "Face":{ + "BoundingBox":{ + "Height":0.33481481671333313, + "Left":0.31888890266418457, + "Top":0.4933333396911621, + "Width":0.25 }, - "Confidence": 99.9991226196289, - "FaceId": "ff43d742-0c13-5d16-a3e8-03d3f58e980b", - "ImageId": "465f4e93-763e-51d0-b030-b9667a2d94b1" + "Confidence":99.9991226196289, + "FaceId":"ff43d742-0c13-5d16-a3e8-03d3f58e980b", + "ImageId":"465f4e93-763e-51d0-b030-b9667a2d94b1" }, - "FaceDetail": { - "BoundingBox": { - "Height": 0.33481481671333313, - "Left": 0.31888890266418457, - "Top": 0.4933333396911621, - "Width": 0.25 + "FaceDetail":{ + "BoundingBox":{ + "Height":0.33481481671333313, + "Left":0.31888890266418457, + "Top":0.4933333396911621, + "Width":0.25 }, - "Confidence": 99.9991226196289, - "Landmarks": [ + "Confidence":99.9991226196289, + "Landmarks":[ { - "Type": "eyeLeft", - "X": 0.3976764678955078, - "Y": 0.6248345971107483 + "Type":"eyeLeft", + "X":0.3976764678955078, + "Y":0.6248345971107483 }, { - "Type": "eyeRight", - "X": 0.4810936450958252, - "Y": 0.6317117214202881 + "Type":"eyeRight", + "X":0.4810936450958252, + "Y":0.6317117214202881 }, { - "Type": "noseLeft", - "X": 0.41986238956451416, - "Y": 0.7111940383911133 + "Type":"noseLeft", + "X":0.41986238956451416, + "Y":0.7111940383911133 }, { - "Type": "mouthDown", - "X": 0.40525302290916443, - "Y": 0.7497701048851013 + "Type":"mouthDown", + "X":0.40525302290916443, + "Y":0.7497701048851013 }, { - "Type": "mouthUp", - "X": 0.4753248989582062, - "Y": 0.7558549642562866 + "Type":"mouthUp", + "X":0.4753248989582062, + "Y":0.7558549642562866 } ], - "Pose": { - "Pitch": -9.713645935058594, - "Roll": 4.707281112670898, - "Yaw": -24.438663482666016 + "Pose":{ + "Pitch":-9.713645935058594, + "Roll":4.707281112670898, + "Yaw":-24.438663482666016 }, - "Quality": { - "Brightness": 29.23358917236328, - "Sharpness": 80 + "Quality":{ + "Brightness":29.23358917236328, + "Sharpness":80 } } }, { - "Face": { - "BoundingBox": { - "Height": 0.32592591643333435, - "Left": 0.5144444704055786, - "Top": 0.15111111104488373, - "Width": 0.24444444477558136 + "Face":{ + "BoundingBox":{ + "Height":0.32592591643333435, + "Left":0.5144444704055786, + "Top":0.15111111104488373, + "Width":0.24444444477558136 }, - "Confidence": 99.99950408935547, - "FaceId": "8be04dba-4e58-520d-850e-9eae4af70eb2", - "ImageId": "465f4e93-763e-51d0-b030-b9667a2d94b1" + "Confidence":99.99950408935547, + "FaceId":"8be04dba-4e58-520d-850e-9eae4af70eb2", + "ImageId":"465f4e93-763e-51d0-b030-b9667a2d94b1" }, - "FaceDetail": { - "BoundingBox": { - "Height": 0.32592591643333435, - "Left": 0.5144444704055786, - "Top": 0.15111111104488373, - "Width": 0.24444444477558136 + "FaceDetail":{ + "BoundingBox":{ + "Height":0.32592591643333435, + "Left":0.5144444704055786, + "Top":0.15111111104488373, + "Width":0.24444444477558136 }, - "Confidence": 99.99950408935547, - "Landmarks": [ + "Confidence":99.99950408935547, + "Landmarks":[ { - "Type": "eyeLeft", - "X": 0.6006892323493958, - "Y": 0.290842205286026 + "Type":"eyeLeft", + "X":0.6006892323493958, + "Y":0.290842205286026 }, { - "Type": "eyeRight", - "X": 0.6808141469955444, - "Y": 0.29609042406082153 + "Type":"eyeRight", + "X":0.6808141469955444, + "Y":0.29609042406082153 }, { - "Type": "noseLeft", - "X": 0.6395332217216492, - "Y": 0.3522595763206482 + "Type":"noseLeft", + "X":0.6395332217216492, + "Y":0.3522595763206482 }, { - "Type": "mouthDown", - "X": 0.5892083048820496, - "Y": 0.38689887523651123 + "Type":"mouthDown", + "X":0.5892083048820496, + "Y":0.38689887523651123 }, { - "Type": "mouthUp", - "X": 0.674560010433197, - "Y": 0.394125759601593 + "Type":"mouthUp", + "X":0.674560010433197, + "Y":0.394125759601593 } ], - "Pose": { - "Pitch": -4.683138370513916, - "Roll": 2.1029529571533203, - "Yaw": 6.716655254364014 + "Pose":{ + "Pitch":-4.683138370513916, + "Roll":2.1029529571533203, + "Yaw":6.716655254364014 }, - "Quality": { - "Brightness": 34.951698303222656, - "Sharpness": 160 + "Quality":{ + "Brightness":34.951698303222656, + "Sharpness":160 } } } ], - "OrientationCorrection": "ROTATE_0" + "OrientationCorrection":"ROTATE_0" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation detects faces in an image and adds them to the specified Rekognition collection.", - "id": "to-add-a-face-to-a-collection-1482179542923", - "title": "To add a face to a collection" + "description":"This operation detects faces in an image and adds them to the specified Rekognition collection.", + "id":"to-add-a-face-to-a-collection-1482179542923", + "title":"To add a face to a collection" } ], - "ListCollections": [ + "ListCollections":[ { - "input": { + "input":{ }, - "output": { - "CollectionIds": [ + "output":{ + "CollectionIds":[ "myphotos" ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation returns a list of Rekognition collections.", - "id": "to-list-the-collections-1482179199088", - "title": "To list the collections" + "description":"This operation returns a list of Rekognition collections.", + "id":"to-list-the-collections-1482179199088", + "title":"To list the collections" } ], - "ListFaces": [ + "ListFaces":[ { - "input": { - "CollectionId": "myphotos", - "MaxResults": 20 + "input":{ + "CollectionId":"myphotos", + "MaxResults":20 }, - "output": { - "Faces": [ + "output":{ + "Faces":[ { - "BoundingBox": { - "Height": 0.18000000715255737, - "Left": 0.5555559992790222, - "Top": 0.336667001247406, - "Width": 0.23999999463558197 + "BoundingBox":{ + "Height":0.18000000715255737, + "Left":0.5555559992790222, + "Top":0.336667001247406, + "Width":0.23999999463558197 }, - "Confidence": 100, - "FaceId": "1c62e8b5-69a7-5b7d-b3cd-db4338a8a7e7", - "ImageId": "147fdf82-7a71-52cf-819b-e786c7b9746e" + "Confidence":100, + "FaceId":"1c62e8b5-69a7-5b7d-b3cd-db4338a8a7e7", + "ImageId":"147fdf82-7a71-52cf-819b-e786c7b9746e" }, { - "BoundingBox": { - "Height": 0.16555599868297577, - "Left": 0.30963000655174255, - "Top": 0.7066670060157776, - "Width": 0.22074100375175476 + "BoundingBox":{ + "Height":0.16555599868297577, + "Left":0.30963000655174255, + "Top":0.7066670060157776, + "Width":0.22074100375175476 }, - "Confidence": 100, - "FaceId": "29a75abe-397b-5101-ba4f-706783b2246c", - "ImageId": "147fdf82-7a71-52cf-819b-e786c7b9746e" + "Confidence":100, + "FaceId":"29a75abe-397b-5101-ba4f-706783b2246c", + "ImageId":"147fdf82-7a71-52cf-819b-e786c7b9746e" }, { - "BoundingBox": { - "Height": 0.3234420120716095, - "Left": 0.3233329951763153, - "Top": 0.5, - "Width": 0.24222199618816376 + "BoundingBox":{ + "Height":0.3234420120716095, + "Left":0.3233329951763153, + "Top":0.5, + "Width":0.24222199618816376 }, - "Confidence": 99.99829864501953, - "FaceId": "38271d79-7bc2-5efb-b752-398a8d575b85", - "ImageId": "d5631190-d039-54e4-b267-abd22c8647c5" + "Confidence":99.99829864501953, + "FaceId":"38271d79-7bc2-5efb-b752-398a8d575b85", + "ImageId":"d5631190-d039-54e4-b267-abd22c8647c5" }, { - "BoundingBox": { - "Height": 0.03555560111999512, - "Left": 0.37388700246810913, - "Top": 0.2477779984474182, - "Width": 0.04747769981622696 + "BoundingBox":{ + "Height":0.03555560111999512, + "Left":0.37388700246810913, + "Top":0.2477779984474182, + "Width":0.04747769981622696 }, - "Confidence": 99.99210357666016, - "FaceId": "3b01bef0-c883-5654-ba42-d5ad28b720b3", - "ImageId": "812d9f04-86f9-54fc-9275-8d0dcbcb6784" + "Confidence":99.99210357666016, + "FaceId":"3b01bef0-c883-5654-ba42-d5ad28b720b3", + "ImageId":"812d9f04-86f9-54fc-9275-8d0dcbcb6784" }, { - "BoundingBox": { - "Height": 0.05333330109715462, - "Left": 0.2937690019607544, - "Top": 0.35666701197624207, - "Width": 0.07121659815311432 + "BoundingBox":{ + "Height":0.05333330109715462, + "Left":0.2937690019607544, + "Top":0.35666701197624207, + "Width":0.07121659815311432 }, - "Confidence": 99.99919891357422, - "FaceId": "4839a608-49d0-566c-8301-509d71b534d1", - "ImageId": "812d9f04-86f9-54fc-9275-8d0dcbcb6784" + "Confidence":99.99919891357422, + "FaceId":"4839a608-49d0-566c-8301-509d71b534d1", + "ImageId":"812d9f04-86f9-54fc-9275-8d0dcbcb6784" }, { - "BoundingBox": { - "Height": 0.3249259889125824, - "Left": 0.5155559778213501, - "Top": 0.1513350009918213, - "Width": 0.24333299696445465 + "BoundingBox":{ + "Height":0.3249259889125824, + "Left":0.5155559778213501, + "Top":0.1513350009918213, + "Width":0.24333299696445465 }, - "Confidence": 99.99949645996094, - "FaceId": "70008e50-75e4-55d0-8e80-363fb73b3a14", - "ImageId": "d5631190-d039-54e4-b267-abd22c8647c5" + "Confidence":99.99949645996094, + "FaceId":"70008e50-75e4-55d0-8e80-363fb73b3a14", + "ImageId":"d5631190-d039-54e4-b267-abd22c8647c5" }, { - "BoundingBox": { - "Height": 0.03777780011296272, - "Left": 0.7002969980239868, - "Top": 0.18777799606323242, - "Width": 0.05044509842991829 + "BoundingBox":{ + "Height":0.03777780011296272, + "Left":0.7002969980239868, + "Top":0.18777799606323242, + "Width":0.05044509842991829 }, - "Confidence": 99.92639923095703, - "FaceId": "7f5f88ed-d684-5a88-b0df-01e4a521552b", - "ImageId": "812d9f04-86f9-54fc-9275-8d0dcbcb6784" + "Confidence":99.92639923095703, + "FaceId":"7f5f88ed-d684-5a88-b0df-01e4a521552b", + "ImageId":"812d9f04-86f9-54fc-9275-8d0dcbcb6784" }, { - "BoundingBox": { - "Height": 0.05555560067296028, - "Left": 0.13946600258350372, - "Top": 0.46333301067352295, - "Width": 0.07270029932260513 + "BoundingBox":{ + "Height":0.05555560067296028, + "Left":0.13946600258350372, + "Top":0.46333301067352295, + "Width":0.07270029932260513 }, - "Confidence": 99.99469757080078, - "FaceId": "895b4e2c-81de-5902-a4bd-d1792bda00b2", - "ImageId": "812d9f04-86f9-54fc-9275-8d0dcbcb6784" + "Confidence":99.99469757080078, + "FaceId":"895b4e2c-81de-5902-a4bd-d1792bda00b2", + "ImageId":"812d9f04-86f9-54fc-9275-8d0dcbcb6784" }, { - "BoundingBox": { - "Height": 0.3259260058403015, - "Left": 0.5144439935684204, - "Top": 0.15111100673675537, - "Width": 0.24444399774074554 + "BoundingBox":{ + "Height":0.3259260058403015, + "Left":0.5144439935684204, + "Top":0.15111100673675537, + "Width":0.24444399774074554 }, - "Confidence": 99.99949645996094, - "FaceId": "8be04dba-4e58-520d-850e-9eae4af70eb2", - "ImageId": "465f4e93-763e-51d0-b030-b9667a2d94b1" + "Confidence":99.99949645996094, + "FaceId":"8be04dba-4e58-520d-850e-9eae4af70eb2", + "ImageId":"465f4e93-763e-51d0-b030-b9667a2d94b1" }, { - "BoundingBox": { - "Height": 0.18888899683952332, - "Left": 0.3783380091190338, - "Top": 0.2355560064315796, - "Width": 0.25222599506378174 + "BoundingBox":{ + "Height":0.18888899683952332, + "Left":0.3783380091190338, + "Top":0.2355560064315796, + "Width":0.25222599506378174 }, - "Confidence": 99.9999008178711, - "FaceId": "908544ad-edc3-59df-8faf-6a87cc256cf5", - "ImageId": "3c731605-d772-541a-a5e7-0375dbc68a07" + "Confidence":99.9999008178711, + "FaceId":"908544ad-edc3-59df-8faf-6a87cc256cf5", + "ImageId":"3c731605-d772-541a-a5e7-0375dbc68a07" }, { - "BoundingBox": { - "Height": 0.33481499552726746, - "Left": 0.31888899207115173, - "Top": 0.49333301186561584, - "Width": 0.25 + "BoundingBox":{ + "Height":0.33481499552726746, + "Left":0.31888899207115173, + "Top":0.49333301186561584, + "Width":0.25 }, - "Confidence": 99.99909973144531, - "FaceId": "ff43d742-0c13-5d16-a3e8-03d3f58e980b", - "ImageId": "465f4e93-763e-51d0-b030-b9667a2d94b1" + "Confidence":99.99909973144531, + "FaceId":"ff43d742-0c13-5d16-a3e8-03d3f58e980b", + "ImageId":"465f4e93-763e-51d0-b030-b9667a2d94b1" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation lists the faces in a Rekognition collection.", - "id": "to-list-the-faces-in-a-collection-1482181416530", - "title": "To list the faces in a collection" + "description":"This operation lists the faces in a Rekognition collection.", + "id":"to-list-the-faces-in-a-collection-1482181416530", + "title":"To list the faces in a collection" } ], - "SearchFaces": [ + "SearchFaces":[ { - "input": { - "CollectionId": "myphotos", - "FaceId": "70008e50-75e4-55d0-8e80-363fb73b3a14", - "FaceMatchThreshold": 90, - "MaxFaces": 10 + "input":{ + "CollectionId":"myphotos", + "FaceId":"70008e50-75e4-55d0-8e80-363fb73b3a14", + "FaceMatchThreshold":90, + "MaxFaces":10 }, - "output": { - "FaceMatches": [ + "output":{ + "FaceMatches":[ { - "Face": { - "BoundingBox": { - "Height": 0.3259260058403015, - "Left": 0.5144439935684204, - "Top": 0.15111100673675537, - "Width": 0.24444399774074554 + "Face":{ + "BoundingBox":{ + "Height":0.3259260058403015, + "Left":0.5144439935684204, + "Top":0.15111100673675537, + "Width":0.24444399774074554 }, - "Confidence": 99.99949645996094, - "FaceId": "8be04dba-4e58-520d-850e-9eae4af70eb2", - "ImageId": "465f4e93-763e-51d0-b030-b9667a2d94b1" + "Confidence":99.99949645996094, + "FaceId":"8be04dba-4e58-520d-850e-9eae4af70eb2", + "ImageId":"465f4e93-763e-51d0-b030-b9667a2d94b1" }, - "Similarity": 99.97222137451172 + "Similarity":99.97222137451172 }, { - "Face": { - "BoundingBox": { - "Height": 0.16555599868297577, - "Left": 0.30963000655174255, - "Top": 0.7066670060157776, - "Width": 0.22074100375175476 + "Face":{ + "BoundingBox":{ + "Height":0.16555599868297577, + "Left":0.30963000655174255, + "Top":0.7066670060157776, + "Width":0.22074100375175476 }, - "Confidence": 100, - "FaceId": "29a75abe-397b-5101-ba4f-706783b2246c", - "ImageId": "147fdf82-7a71-52cf-819b-e786c7b9746e" + "Confidence":100, + "FaceId":"29a75abe-397b-5101-ba4f-706783b2246c", + "ImageId":"147fdf82-7a71-52cf-819b-e786c7b9746e" }, - "Similarity": 97.04154968261719 + "Similarity":97.04154968261719 }, { - "Face": { - "BoundingBox": { - "Height": 0.18888899683952332, - "Left": 0.3783380091190338, - "Top": 0.2355560064315796, - "Width": 0.25222599506378174 + "Face":{ + "BoundingBox":{ + "Height":0.18888899683952332, + "Left":0.3783380091190338, + "Top":0.2355560064315796, + "Width":0.25222599506378174 }, - "Confidence": 99.9999008178711, - "FaceId": "908544ad-edc3-59df-8faf-6a87cc256cf5", - "ImageId": "3c731605-d772-541a-a5e7-0375dbc68a07" + "Confidence":99.9999008178711, + "FaceId":"908544ad-edc3-59df-8faf-6a87cc256cf5", + "ImageId":"3c731605-d772-541a-a5e7-0375dbc68a07" }, - "Similarity": 95.94520568847656 + "Similarity":95.94520568847656 } ], - "SearchedFaceId": "70008e50-75e4-55d0-8e80-363fb73b3a14" + "SearchedFaceId":"70008e50-75e4-55d0-8e80-363fb73b3a14" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation searches for matching faces in the collection the supplied face belongs to.", - "id": "to-delete-a-face-1482182799377", - "title": "To delete a face" + "description":"This operation searches for matching faces in the collection the supplied face belongs to.", + "id":"to-delete-a-face-1482182799377", + "title":"To delete a face" } ], - "SearchFacesByImage": [ + "SearchFacesByImage":[ { - "input": { - "CollectionId": "myphotos", - "FaceMatchThreshold": 95, - "Image": { - "S3Object": { - "Bucket": "mybucket", - "Name": "myphoto" + "input":{ + "CollectionId":"myphotos", + "FaceMatchThreshold":95, + "Image":{ + "S3Object":{ + "Bucket":"mybucket", + "Name":"myphoto" } }, - "MaxFaces": 5 + "MaxFaces":5 }, - "output": { - "FaceMatches": [ + "output":{ + "FaceMatches":[ { - "Face": { - "BoundingBox": { - "Height": 0.3234420120716095, - "Left": 0.3233329951763153, - "Top": 0.5, - "Width": 0.24222199618816376 + "Face":{ + "BoundingBox":{ + "Height":0.3234420120716095, + "Left":0.3233329951763153, + "Top":0.5, + "Width":0.24222199618816376 }, - "Confidence": 99.99829864501953, - "FaceId": "38271d79-7bc2-5efb-b752-398a8d575b85", - "ImageId": "d5631190-d039-54e4-b267-abd22c8647c5" + "Confidence":99.99829864501953, + "FaceId":"38271d79-7bc2-5efb-b752-398a8d575b85", + "ImageId":"d5631190-d039-54e4-b267-abd22c8647c5" }, - "Similarity": 99.97036743164062 + "Similarity":99.97036743164062 } ], - "SearchedFaceBoundingBox": { - "Height": 0.33481481671333313, - "Left": 0.31888890266418457, - "Top": 0.4933333396911621, - "Width": 0.25 + "SearchedFaceBoundingBox":{ + "Height":0.33481481671333313, + "Left":0.31888890266418457, + "Top":0.4933333396911621, + "Width":0.25 }, - "SearchedFaceConfidence": 99.9991226196289 + "SearchedFaceConfidence":99.9991226196289 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation searches for faces in a Rekognition collection that match the largest face in an S3 bucket stored image.", - "id": "to-search-for-faces-matching-a-supplied-image-1482175994491", - "title": "To search for faces matching a supplied image" + "description":"This operation searches for faces in a Rekognition collection that match the largest face in an S3 bucket stored image.", + "id":"to-search-for-faces-matching-a-supplied-image-1482175994491", + "title":"To search for faces matching a supplied image" } ] } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/api-2.json index 8730fba30..42eb371f5 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/api-2.json @@ -859,6 +859,7 @@ } }, "AccountId":{"type":"string"}, + "AllowQuotedRecordDelimiter":{"type":"boolean"}, "AllowedHeader":{"type":"string"}, "AllowedHeaders":{ "type":"list", @@ -1081,7 +1082,8 @@ "QuoteEscapeCharacter":{"shape":"QuoteEscapeCharacter"}, "RecordDelimiter":{"shape":"RecordDelimiter"}, "FieldDelimiter":{"shape":"FieldDelimiter"}, - "QuoteCharacter":{"shape":"QuoteCharacter"} + "QuoteCharacter":{"shape":"QuoteCharacter"}, + "AllowQuotedRecordDelimiter":{"shape":"AllowQuotedRecordDelimiter"} } }, "CSVOutput":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/docs-2.json index 56ba774b8..afb0d78bd 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/s3/2006-03-01/docs-2.json @@ -147,6 +147,12 @@ "InventoryS3BucketDestination$AccountId": "The ID of the account that owns the destination bucket." } }, + "AllowQuotedRecordDelimiter": { + "base": null, + "refs": { + "CSVInput$AllowQuotedRecordDelimiter": "Specifies that CSV field values may contain quoted record delimiters and such records should be allowed. Default value is FALSE. Setting this value to TRUE may lower performance." + } + }, "AllowedHeader": { "base": null, "refs": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/api-2.json index 1b64834b3..37faea423 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/api-2.json @@ -46,6 +46,19 @@ {"shape":"ResourceLimitExceeded"} ] }, + "CreateHyperParameterTuningJob":{ + "name":"CreateHyperParameterTuningJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"CreateHyperParameterTuningJobRequest"}, + "output":{"shape":"CreateHyperParameterTuningJobResponse"}, + "errors":[ + {"shape":"ResourceInUse"}, + {"shape":"ResourceLimitExceeded"} + ] + }, "CreateModel":{ "name":"CreateModel", "http":{ @@ -171,6 +184,18 @@ "input":{"shape":"DescribeEndpointConfigInput"}, "output":{"shape":"DescribeEndpointConfigOutput"} }, + "DescribeHyperParameterTuningJob":{ + "name":"DescribeHyperParameterTuningJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeHyperParameterTuningJobRequest"}, + "output":{"shape":"DescribeHyperParameterTuningJobResponse"}, + "errors":[ + {"shape":"ResourceNotFound"} + ] + }, "DescribeModel":{ "name":"DescribeModel", "http":{ @@ -228,6 +253,15 @@ "input":{"shape":"ListEndpointsInput"}, "output":{"shape":"ListEndpointsOutput"} }, + "ListHyperParameterTuningJobs":{ + "name":"ListHyperParameterTuningJobs", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListHyperParameterTuningJobsRequest"}, + "output":{"shape":"ListHyperParameterTuningJobsResponse"} + }, "ListModels":{ "name":"ListModels", "http":{ @@ -273,6 +307,18 @@ "input":{"shape":"ListTrainingJobsRequest"}, "output":{"shape":"ListTrainingJobsResponse"} }, + "ListTrainingJobsForHyperParameterTuningJob":{ + "name":"ListTrainingJobsForHyperParameterTuningJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"ListTrainingJobsForHyperParameterTuningJobRequest"}, + "output":{"shape":"ListTrainingJobsForHyperParameterTuningJobResponse"}, + "errors":[ + {"shape":"ResourceNotFound"} + ] + }, "StartNotebookInstance":{ "name":"StartNotebookInstance", "http":{ @@ -284,6 +330,17 @@ {"shape":"ResourceLimitExceeded"} ] }, + "StopHyperParameterTuningJob":{ + "name":"StopHyperParameterTuningJob", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"StopHyperParameterTuningJobRequest"}, + "errors":[ + {"shape":"ResourceNotFound"} + ] + }, "StopNotebookInstance":{ "name":"StopNotebookInstance", "http":{ @@ -385,6 +442,23 @@ "TrainingInputMode":{"shape":"TrainingInputMode"} } }, + "CategoricalParameterRange":{ + "type":"structure", + "required":[ + "Name", + "Values" + ], + "members":{ + "Name":{"shape":"ParameterKey"}, + "Values":{"shape":"ParameterValues"} + } + }, + "CategoricalParameterRanges":{ + "type":"list", + "member":{"shape":"CategoricalParameterRange"}, + "max":20, + "min":0 + }, "Channel":{ "type":"structure", "required":[ @@ -431,6 +505,25 @@ "type":"string", "max":256 }, + "ContinuousParameterRange":{ + "type":"structure", + "required":[ + "Name", + "MinValue", + "MaxValue" + ], + "members":{ + "Name":{"shape":"ParameterKey"}, + "MinValue":{"shape":"ParameterValue"}, + "MaxValue":{"shape":"ParameterValue"} + } + }, + "ContinuousParameterRanges":{ + "type":"list", + "member":{"shape":"ContinuousParameterRange"}, + "max":20, + "min":0 + }, "CreateEndpointConfigInput":{ "type":"structure", "required":[ @@ -470,6 +563,27 @@ "EndpointArn":{"shape":"EndpointArn"} } }, + "CreateHyperParameterTuningJobRequest":{ + "type":"structure", + "required":[ + "HyperParameterTuningJobName", + "HyperParameterTuningJobConfig", + "TrainingJobDefinition" + ], + "members":{ + "HyperParameterTuningJobName":{"shape":"HyperParameterTuningJobName"}, + "HyperParameterTuningJobConfig":{"shape":"HyperParameterTuningJobConfig"}, + "TrainingJobDefinition":{"shape":"HyperParameterTrainingJobDefinition"}, + "Tags":{"shape":"TagList"} + } + }, + "CreateHyperParameterTuningJobResponse":{ + "type":"structure", + "required":["HyperParameterTuningJobArn"], + "members":{ + "HyperParameterTuningJobArn":{"shape":"HyperParameterTuningJobArn"} + } + }, "CreateModelInput":{ "type":"structure", "required":[ @@ -687,6 +801,40 @@ "LastModifiedTime":{"shape":"Timestamp"} } }, + "DescribeHyperParameterTuningJobRequest":{ + "type":"structure", + "required":["HyperParameterTuningJobName"], + "members":{ + "HyperParameterTuningJobName":{"shape":"HyperParameterTuningJobName"} + } + }, + "DescribeHyperParameterTuningJobResponse":{ + "type":"structure", + "required":[ + "HyperParameterTuningJobName", + "HyperParameterTuningJobArn", + "HyperParameterTuningJobConfig", + "TrainingJobDefinition", + "HyperParameterTuningJobStatus", + "CreationTime", + "TrainingJobStatusCounters", + "ObjectiveStatusCounters" + ], + "members":{ + "HyperParameterTuningJobName":{"shape":"HyperParameterTuningJobName"}, + "HyperParameterTuningJobArn":{"shape":"HyperParameterTuningJobArn"}, + "HyperParameterTuningJobConfig":{"shape":"HyperParameterTuningJobConfig"}, + "TrainingJobDefinition":{"shape":"HyperParameterTrainingJobDefinition"}, + "HyperParameterTuningJobStatus":{"shape":"HyperParameterTuningJobStatus"}, + "CreationTime":{"shape":"Timestamp"}, + "HyperParameterTuningEndTime":{"shape":"Timestamp"}, + "LastModifiedTime":{"shape":"Timestamp"}, + "TrainingJobStatusCounters":{"shape":"TrainingJobStatusCounters"}, + "ObjectiveStatusCounters":{"shape":"ObjectiveStatusCounters"}, + "BestTrainingJob":{"shape":"HyperParameterTrainingJobSummary"}, + "FailureReason":{"shape":"FailureReason"} + } + }, "DescribeModelInput":{ "type":"structure", "required":["ModelName"], @@ -781,6 +929,7 @@ "members":{ "TrainingJobName":{"shape":"TrainingJobName"}, "TrainingJobArn":{"shape":"TrainingJobArn"}, + "TuningJobArn":{"shape":"HyperParameterTuningJobArn"}, "ModelArtifacts":{"shape":"ModelArtifacts"}, "TrainingJobStatus":{"shape":"TrainingJobStatus"}, "SecondaryStatus":{"shape":"SecondaryStatus"}, @@ -932,6 +1081,171 @@ "type":"string", "max":1024 }, + "FinalHyperParameterTuningJobObjectiveMetric":{ + "type":"structure", + "required":[ + "MetricName", + "Value" + ], + "members":{ + "Type":{"shape":"HyperParameterTuningJobObjectiveType"}, + "MetricName":{"shape":"MetricName"}, + "Value":{"shape":"MetricValue"} + } + }, + "HyperParameterAlgorithmSpecification":{ + "type":"structure", + "required":[ + "TrainingImage", + "TrainingInputMode" + ], + "members":{ + "TrainingImage":{"shape":"AlgorithmImage"}, + "TrainingInputMode":{"shape":"TrainingInputMode"}, + "MetricDefinitions":{"shape":"MetricDefinitionList"} + } + }, + "HyperParameterTrainingJobDefinition":{ + "type":"structure", + "required":[ + "AlgorithmSpecification", + "RoleArn", + "InputDataConfig", + "OutputDataConfig", + "ResourceConfig", + "StoppingCondition" + ], + "members":{ + "StaticHyperParameters":{"shape":"HyperParameters"}, + "AlgorithmSpecification":{"shape":"HyperParameterAlgorithmSpecification"}, + "RoleArn":{"shape":"RoleArn"}, + "InputDataConfig":{"shape":"InputDataConfig"}, + "VpcConfig":{"shape":"VpcConfig"}, + "OutputDataConfig":{"shape":"OutputDataConfig"}, + "ResourceConfig":{"shape":"ResourceConfig"}, + "StoppingCondition":{"shape":"StoppingCondition"} + } + }, + "HyperParameterTrainingJobSummaries":{ + "type":"list", + "member":{"shape":"HyperParameterTrainingJobSummary"} + }, + "HyperParameterTrainingJobSummary":{ + "type":"structure", + "required":[ + "TrainingJobName", + "TrainingJobArn", + "CreationTime", + "TrainingJobStatus", + "TunedHyperParameters" + ], + "members":{ + "TrainingJobName":{"shape":"TrainingJobName"}, + "TrainingJobArn":{"shape":"TrainingJobArn"}, + "CreationTime":{"shape":"Timestamp"}, + "TrainingStartTime":{"shape":"Timestamp"}, + "TrainingEndTime":{"shape":"Timestamp"}, + "TrainingJobStatus":{"shape":"TrainingJobStatus"}, + "TunedHyperParameters":{"shape":"HyperParameters"}, + "FailureReason":{"shape":"FailureReason"}, + "FinalHyperParameterTuningJobObjectiveMetric":{"shape":"FinalHyperParameterTuningJobObjectiveMetric"}, + "ObjectiveStatus":{"shape":"ObjectiveStatus"} + } + }, + "HyperParameterTuningJobArn":{ + "type":"string", + "max":256, + "pattern":"arn:aws[a-z\\-]*:sagemaker:[a-z0-9\\-]*:[0-9]{12}:hyper-parameter-tuning-job/.*" + }, + "HyperParameterTuningJobConfig":{ + "type":"structure", + "required":[ + "Strategy", + "HyperParameterTuningJobObjective", + "ResourceLimits", + "ParameterRanges" + ], + "members":{ + "Strategy":{"shape":"HyperParameterTuningJobStrategyType"}, + "HyperParameterTuningJobObjective":{"shape":"HyperParameterTuningJobObjective"}, + "ResourceLimits":{"shape":"ResourceLimits"}, + "ParameterRanges":{"shape":"ParameterRanges"} + } + }, + "HyperParameterTuningJobName":{ + "type":"string", + "max":32, + "min":1, + "pattern":"^[a-zA-Z0-9](-*[a-zA-Z0-9])*" + }, + "HyperParameterTuningJobObjective":{ + "type":"structure", + "required":[ + "Type", + "MetricName" + ], + "members":{ + "Type":{"shape":"HyperParameterTuningJobObjectiveType"}, + "MetricName":{"shape":"MetricName"} + } + }, + "HyperParameterTuningJobObjectiveType":{ + "type":"string", + "enum":[ + "Maximize", + "Minimize" + ] + }, + "HyperParameterTuningJobSortByOptions":{ + "type":"string", + "enum":[ + "Name", + "Status", + "CreationTime" + ] + }, + "HyperParameterTuningJobStatus":{ + "type":"string", + "enum":[ + "Completed", + "InProgress", + "Failed", + "Stopped", + "Stopping" + ] + }, + "HyperParameterTuningJobStrategyType":{ + "type":"string", + "enum":["Bayesian"] + }, + "HyperParameterTuningJobSummaries":{ + "type":"list", + "member":{"shape":"HyperParameterTuningJobSummary"} + }, + "HyperParameterTuningJobSummary":{ + "type":"structure", + "required":[ + "HyperParameterTuningJobName", + "HyperParameterTuningJobArn", + "HyperParameterTuningJobStatus", + "Strategy", + "CreationTime", + "TrainingJobStatusCounters", + "ObjectiveStatusCounters" + ], + "members":{ + "HyperParameterTuningJobName":{"shape":"HyperParameterTuningJobName"}, + "HyperParameterTuningJobArn":{"shape":"HyperParameterTuningJobArn"}, + "HyperParameterTuningJobStatus":{"shape":"HyperParameterTuningJobStatus"}, + "Strategy":{"shape":"HyperParameterTuningJobStrategyType"}, + "CreationTime":{"shape":"Timestamp"}, + "HyperParameterTuningEndTime":{"shape":"Timestamp"}, + "LastModifiedTime":{"shape":"Timestamp"}, + "TrainingJobStatusCounters":{"shape":"TrainingJobStatusCounters"}, + "ObjectiveStatusCounters":{"shape":"ObjectiveStatusCounters"}, + "ResourceLimits":{"shape":"ResourceLimits"} + } + }, "HyperParameters":{ "type":"map", "key":{"shape":"ParameterKey"}, @@ -970,6 +1284,25 @@ "ml.p3.16xlarge" ] }, + "IntegerParameterRange":{ + "type":"structure", + "required":[ + "Name", + "MinValue", + "MaxValue" + ], + "members":{ + "Name":{"shape":"ParameterKey"}, + "MinValue":{"shape":"ParameterValue"}, + "MaxValue":{"shape":"ParameterValue"} + } + }, + "IntegerParameterRanges":{ + "type":"list", + "member":{"shape":"IntegerParameterRange"}, + "max":20, + "min":0 + }, "KmsKeyId":{ "type":"string", "max":2048 @@ -1018,6 +1351,32 @@ "NextToken":{"shape":"PaginationToken"} } }, + "ListHyperParameterTuningJobsRequest":{ + "type":"structure", + "members":{ + "NextToken":{"shape":"NextToken"}, + "MaxResults":{ + "shape":"MaxResults", + "box":true + }, + "SortBy":{"shape":"HyperParameterTuningJobSortByOptions"}, + "SortOrder":{"shape":"SortOrder"}, + "NameContains":{"shape":"NameContains"}, + "CreationTimeAfter":{"shape":"Timestamp"}, + "CreationTimeBefore":{"shape":"Timestamp"}, + "LastModifiedTimeAfter":{"shape":"Timestamp"}, + "LastModifiedTimeBefore":{"shape":"Timestamp"}, + "StatusEquals":{"shape":"HyperParameterTuningJobStatus"} + } + }, + "ListHyperParameterTuningJobsResponse":{ + "type":"structure", + "required":["HyperParameterTuningJobSummaries"], + "members":{ + "HyperParameterTuningJobSummaries":{"shape":"HyperParameterTuningJobSummaries"}, + "NextToken":{"shape":"NextToken"} + } + }, "ListModelsInput":{ "type":"structure", "members":{ @@ -1102,6 +1461,26 @@ "NextToken":{"shape":"NextToken"} } }, + "ListTrainingJobsForHyperParameterTuningJobRequest":{ + "type":"structure", + "required":["HyperParameterTuningJobName"], + "members":{ + "HyperParameterTuningJobName":{"shape":"HyperParameterTuningJobName"}, + "NextToken":{"shape":"NextToken"}, + "MaxResults":{"shape":"MaxResults"}, + "StatusEquals":{"shape":"TrainingJobStatus"}, + "SortBy":{"shape":"TrainingJobSortByOptions"}, + "SortOrder":{"shape":"SortOrder"} + } + }, + "ListTrainingJobsForHyperParameterTuningJobResponse":{ + "type":"structure", + "required":["TrainingJobSummaries"], + "members":{ + "TrainingJobSummaries":{"shape":"HyperParameterTrainingJobSummaries"}, + "NextToken":{"shape":"NextToken"} + } + }, "ListTrainingJobsRequest":{ "type":"structure", "members":{ @@ -1128,6 +1507,14 @@ "NextToken":{"shape":"NextToken"} } }, + "MaxNumberOfTrainingJobs":{ + "type":"integer", + "min":1 + }, + "MaxParallelTrainingJobs":{ + "type":"integer", + "min":1 + }, "MaxResults":{ "type":"integer", "max":100, @@ -1137,6 +1524,34 @@ "type":"integer", "min":1 }, + "MetricDefinition":{ + "type":"structure", + "required":[ + "Name", + "Regex" + ], + "members":{ + "Name":{"shape":"MetricName"}, + "Regex":{"shape":"MetricRegex"} + } + }, + "MetricDefinitionList":{ + "type":"list", + "member":{"shape":"MetricDefinition"}, + "max":20, + "min":0 + }, + "MetricName":{ + "type":"string", + "max":255, + "min":1 + }, + "MetricRegex":{ + "type":"string", + "max":500, + "min":1 + }, + "MetricValue":{"type":"float"}, "ModelArn":{ "type":"string", "max":2048, @@ -1314,6 +1729,26 @@ "member":{"shape":"NotebookInstanceSummary"} }, "NotebookInstanceUrl":{"type":"string"}, + "ObjectiveStatus":{ + "type":"string", + "enum":[ + "Succeeded", + "Pending", + "Failed" + ] + }, + "ObjectiveStatusCounter":{ + "type":"integer", + "min":0 + }, + "ObjectiveStatusCounters":{ + "type":"structure", + "members":{ + "Succeeded":{"shape":"ObjectiveStatusCounter"}, + "Pending":{"shape":"ObjectiveStatusCounter"}, + "Failed":{"shape":"ObjectiveStatusCounter"} + } + }, "OrderKey":{ "type":"string", "enum":[ @@ -1337,10 +1772,24 @@ "type":"string", "max":256 }, + "ParameterRanges":{ + "type":"structure", + "members":{ + "IntegerParameterRanges":{"shape":"IntegerParameterRanges"}, + "ContinuousParameterRanges":{"shape":"ContinuousParameterRanges"}, + "CategoricalParameterRanges":{"shape":"CategoricalParameterRanges"} + } + }, "ParameterValue":{ "type":"string", "max":256 }, + "ParameterValues":{ + "type":"list", + "member":{"shape":"ParameterValue"}, + "max":20, + "min":1 + }, "ProductionVariant":{ "type":"structure", "required":[ @@ -1454,6 +1903,17 @@ }, "exception":true }, + "ResourceLimits":{ + "type":"structure", + "required":[ + "MaxNumberOfTrainingJobs", + "MaxParallelTrainingJobs" + ], + "members":{ + "MaxNumberOfTrainingJobs":{"shape":"MaxNumberOfTrainingJobs"}, + "MaxParallelTrainingJobs":{"shape":"MaxParallelTrainingJobs"} + } + }, "ResourceNotFound":{ "type":"structure", "members":{ @@ -1548,6 +2008,13 @@ "NotebookInstanceName":{"shape":"NotebookInstanceName"} } }, + "StopHyperParameterTuningJobRequest":{ + "type":"structure", + "required":["HyperParameterTuningJobName"], + "members":{ + "HyperParameterTuningJobName":{"shape":"HyperParameterTuningJobName"} + } + }, "StopNotebookInstanceInput":{ "type":"structure", "required":["NotebookInstanceName"], @@ -1663,7 +2130,7 @@ "TrainingJobArn":{ "type":"string", "max":256, - "pattern":"arn:aws[a-z\\-]*:sagemaker:[\\p{Alnum}\\-]*:[0-9]{12}:training-job/.*" + "pattern":"arn:aws[a-z\\-]*:sagemaker:[a-z0-9\\-]*:[0-9]{12}:training-job/.*" }, "TrainingJobName":{ "type":"string", @@ -1671,6 +2138,15 @@ "min":1, "pattern":"^[a-zA-Z0-9](-*[a-zA-Z0-9])*" }, + "TrainingJobSortByOptions":{ + "type":"string", + "enum":[ + "Name", + "CreationTime", + "Status", + "FinalObjectiveMetricValue" + ] + }, "TrainingJobStatus":{ "type":"string", "enum":[ @@ -1681,6 +2157,20 @@ "Stopped" ] }, + "TrainingJobStatusCounter":{ + "type":"integer", + "min":0 + }, + "TrainingJobStatusCounters":{ + "type":"structure", + "members":{ + "Completed":{"shape":"TrainingJobStatusCounter"}, + "InProgress":{"shape":"TrainingJobStatusCounter"}, + "RetryableError":{"shape":"TrainingJobStatusCounter"}, + "NonRetryableError":{"shape":"TrainingJobStatusCounter"}, + "Stopped":{"shape":"TrainingJobStatusCounter"} + } + }, "TrainingJobSummaries":{ "type":"list", "member":{"shape":"TrainingJobSummary"} diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/docs-2.json index e2077edc8..f4dd9ddb4 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/docs-2.json @@ -5,31 +5,36 @@ "AddTags": "

Adds or overwrites one or more tags for the specified Amazon SageMaker resource. You can add tags to notebook instances, training jobs, models, endpoint configurations, and endpoints.

Each tag consists of a key and an optional value. Tag keys must be unique per resource. For more information about tags, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.

", "CreateEndpoint": "

Creates an endpoint using the endpoint configuration specified in the request. Amazon SageMaker uses the endpoint to provision resources and deploy models. You create the endpoint configuration with the CreateEndpointConfig API.

Use this API only for hosting models using Amazon SageMaker hosting services.

The endpoint name must be unique within an AWS Region in your AWS account.

When it receives the request, Amazon SageMaker creates the endpoint, launches the resources (ML compute instances), and deploys the model(s) on them.

When Amazon SageMaker receives the request, it sets the endpoint status to Creating. After it creates the endpoint, it sets the status to InService. Amazon SageMaker can then process incoming requests for inferences. To check the status of an endpoint, use the DescribeEndpoint API.

For an example, see Exercise 1: Using the K-Means Algorithm Provided by Amazon SageMaker.

", "CreateEndpointConfig": "

Creates an endpoint configuration that Amazon SageMaker hosting services uses to deploy models. In the configuration, you identify one or more models, created using the CreateModel API, to deploy and the resources that you want Amazon SageMaker to provision. Then you call the CreateEndpoint API.

Use this API only if you want to use Amazon SageMaker hosting services to deploy models into production.

In the request, you define one or more ProductionVariants, each of which identifies a model. Each ProductionVariant parameter also describes the resources that you want Amazon SageMaker to provision. This includes the number and type of ML compute instances to deploy.

If you are hosting multiple models, you also assign a VariantWeight to specify how much traffic you want to allocate to each model. For example, suppose that you want to host two models, A and B, and you assign traffic weight 2 for model A and 1 for model B. Amazon SageMaker distributes two-thirds of the traffic to Model A, and one-third to model B.

", + "CreateHyperParameterTuningJob": "

Starts a hyperparameter tuning job.

", "CreateModel": "

Creates a model in Amazon SageMaker. In the request, you name the model and describe one or more containers. For each container, you specify the docker image containing inference code, artifacts (from prior training), and custom environment map that the inference code uses when you deploy the model into production.

Use this API to create a model only if you want to use Amazon SageMaker hosting services. To host your model, you create an endpoint configuration with the CreateEndpointConfig API, and then create an endpoint with the CreateEndpoint API.

Amazon SageMaker then deploys all of the containers that you defined for the model in the hosting environment.

In the CreateModel request, you must define a container with the PrimaryContainer parameter.

In the request, you also provide an IAM role that Amazon SageMaker can assume to access model artifacts and docker image for deployment on ML compute hosting instances. In addition, you also use the IAM role to manage permissions the inference code needs. For example, if the inference code access any other AWS resources, you grant necessary permissions via this role.

", "CreateNotebookInstance": "

Creates an Amazon SageMaker notebook instance. A notebook instance is a machine learning (ML) compute instance running on a Jupyter notebook.

In a CreateNotebookInstance request, specify the type of ML compute instance that you want to run. Amazon SageMaker launches the instance, installs common libraries that you can use to explore datasets for model training, and attaches an ML storage volume to the notebook instance.

Amazon SageMaker also provides a set of example notebooks. Each notebook demonstrates how to use Amazon SageMaker with a specific algorithm or with a machine learning framework.

After receiving the request, Amazon SageMaker does the following:

  1. Creates a network interface in the Amazon SageMaker VPC.

  2. (Option) If you specified SubnetId, Amazon SageMaker creates a network interface in your own VPC, which is inferred from the subnet ID that you provide in the input. When creating this network interface, Amazon SageMaker attaches the security group that you specified in the request to the network interface that it creates in your VPC.

  3. Launches an EC2 instance of the type specified in the request in the Amazon SageMaker VPC. If you specified SubnetId of your VPC, Amazon SageMaker specifies both network interfaces when launching this instance. This enables inbound traffic from your own VPC to the notebook instance, assuming that the security groups allow it.

After creating the notebook instance, Amazon SageMaker returns its Amazon Resource Name (ARN).

After Amazon SageMaker creates the notebook instance, you can connect to the Jupyter server and work in Jupyter notebooks. For example, you can write code to explore a dataset that you can use for model training, train a model, host models by creating Amazon SageMaker endpoints, and validate hosted models.

For more information, see How It Works.

", "CreateNotebookInstanceLifecycleConfig": "

Creates a lifecycle configuration that you can associate with a notebook instance. A lifecycle configuration is a collection of shell scripts that run when you create or start a notebook instance.

Each lifecycle configuration script has a limit of 16384 characters.

The value of the $PATH environment variable that is available to both scripts is /sbin:bin:/usr/sbin:/usr/bin.

View CloudWatch Logs for notebook instance lifecycle configurations in log group /aws/sagemaker/NotebookInstances in log stream [notebook-instance-name]/[LifecycleConfigHook].

Lifecycle configuration scripts cannot run for longer than 5 minutes. If a script runs for longer than 5 minutes, it fails and the notebook instance is not created or started.

For information about notebook instance lifestyle configurations, see notebook-lifecycle-config.

", - "CreatePresignedNotebookInstanceUrl": "

Returns a URL that you can use to connect to the Juypter server from a notebook instance. In the Amazon SageMaker console, when you choose Open next to a notebook instance, Amazon SageMaker opens a new tab showing the Jupyter server home page from the notebook instance. The console uses this API to get the URL and show the page.

", - "CreateTrainingJob": "

Starts a model training job. After training completes, Amazon SageMaker saves the resulting model artifacts to an Amazon S3 location that you specify.

If you choose to host your model using Amazon SageMaker hosting services, you can use the resulting model artifacts as part of the model. You can also use the artifacts in a deep learning service other than Amazon SageMaker, provided that you know how to use them for inferences.

In the request body, you provide the following:

  • AlgorithmSpecification - Identifies the training algorithm to use.

  • HyperParameters - Specify these algorithm-specific parameters to influence the quality of the final model. For a list of hyperparameters for each training algorithm provided by Amazon SageMaker, see Algorithms.

  • InputDataConfig - Describes the training dataset and the Amazon S3 location where it is stored.

  • OutputDataConfig - Identifies the Amazon S3 location where you want Amazon SageMaker to save the results of model training.

  • ResourceConfig - Identifies the resources, ML compute instances, and ML storage volumes to deploy for model training. In distributed training, you specify more than one instance.

  • RoleARN - The Amazon Resource Number (ARN) that Amazon SageMaker assumes to perform tasks on your behalf during model training. You must grant this role the necessary permissions so that Amazon SageMaker can successfully complete model training.

  • StoppingCondition - Sets a duration for training. Use this parameter to cap model training costs.

For more information about Amazon SageMaker, see How It Works.

", + "CreatePresignedNotebookInstanceUrl": "

Returns a URL that you can use to connect to the Jupyter server from a notebook instance. In the Amazon SageMaker console, when you choose Open next to a notebook instance, Amazon SageMaker opens a new tab showing the Jupyter server home page from the notebook instance. The console uses this API to get the URL and show the page.

", + "CreateTrainingJob": "

Starts a model training job. After training completes, Amazon SageMaker saves the resulting model artifacts to an Amazon S3 location that you specify.

If you choose to host your model using Amazon SageMaker hosting services, you can use the resulting model artifacts as part of the model. You can also use the artifacts in a deep learning service other than Amazon SageMaker, provided that you know how to use them for inferences.

In the request body, you provide the following:

  • AlgorithmSpecification - Identifies the training algorithm to use.

  • HyperParameters - Specify these algorithm-specific parameters to influence the quality of the final model. For a list of hyperparameters for each training algorithm provided by Amazon SageMaker, see Algorithms.

  • InputDataConfig - Describes the training dataset and the Amazon S3 location where it is stored.

  • OutputDataConfig - Identifies the Amazon S3 location where you want Amazon SageMaker to save the results of model training.

  • ResourceConfig - Identifies the resources, ML compute instances, and ML storage volumes to deploy for model training. In distributed training, you specify more than one instance.

  • RoleARN - The Amazon Resource Number (ARN) that Amazon SageMaker assumes to perform tasks on your behalf during model training. You must grant this role the necessary permissions so that Amazon SageMaker can successfully complete model training.

  • StoppingCondition - Sets a duration for training. Use this parameter to cap model training costs.

For more information about Amazon SageMaker, see How It Works.

", "DeleteEndpoint": "

Deletes an endpoint. Amazon SageMaker frees up all of the resources that were deployed when the endpoint was created.

", - "DeleteEndpointConfig": "

Deletes an endpoint configuration. The DeleteEndpoingConfig API deletes only the specified configuration. It does not delete endpoints created using the configuration.

", + "DeleteEndpointConfig": "

Deletes an endpoint configuration. The DeleteEndpointConfig API deletes only the specified configuration. It does not delete endpoints created using the configuration.

", "DeleteModel": "

Deletes a model. The DeleteModel API deletes only the model entry that was created in Amazon SageMaker when you called the CreateModel API. It does not delete model artifacts, inference code, or the IAM role that you specified when creating the model.

", "DeleteNotebookInstance": "

Deletes an Amazon SageMaker notebook instance. Before you can delete a notebook instance, you must call the StopNotebookInstance API.

When you delete a notebook instance, you lose all of your data. Amazon SageMaker removes the ML compute instance, and deletes the ML storage volume and the network interface associated with the notebook instance.

", "DeleteNotebookInstanceLifecycleConfig": "

Deletes a notebook instance lifecycle configuration.

", "DeleteTags": "

Deletes the specified tags from an Amazon SageMaker resource.

To list a resource's tags, use the ListTags API.

", "DescribeEndpoint": "

Returns the description of an endpoint.

", "DescribeEndpointConfig": "

Returns the description of an endpoint configuration created using the CreateEndpointConfig API.

", + "DescribeHyperParameterTuningJob": "

Gets a description of a hyperparameter tuning job.

", "DescribeModel": "

Describes a model that you created using the CreateModel API.

", "DescribeNotebookInstance": "

Returns information about a notebook instance.

", "DescribeNotebookInstanceLifecycleConfig": "

Returns a description of a notebook instance lifecycle configuration.

For information about notebook instance lifestyle configurations, see notebook-lifecycle-config.

", "DescribeTrainingJob": "

Returns information about a training job.

", "ListEndpointConfigs": "

Lists endpoint configurations.

", "ListEndpoints": "

Lists endpoints.

", + "ListHyperParameterTuningJobs": "

Gets a list of objects that describe the hyperparameter tuning jobs launched in your account.

", "ListModels": "

Lists models created with the CreateModel API.

", "ListNotebookInstanceLifecycleConfigs": "

Lists notebook instance lifestyle configurations created with the API.

", "ListNotebookInstances": "

Returns a list of the Amazon SageMaker notebook instances in the requester's account in an AWS Region.

", "ListTags": "

Returns the tags for the specified Amazon SageMaker resource.

", "ListTrainingJobs": "

Lists training jobs.

", + "ListTrainingJobsForHyperParameterTuningJob": "

Gets a list of objects that describe the training jobs that a hyperparameter tuning job launched.

", "StartNotebookInstance": "

Launches an ML compute instance with the latest version of the libraries and attaches your ML storage volume. After configuring the notebook instance, Amazon SageMaker sets the notebook instance status to InService. A notebook instance's status must be InService before you can connect to your Jupyter notebook.

", + "StopHyperParameterTuningJob": "

Stops a running hyperparameter tuning job and all running training jobs that the tuning job launched.

All model artifacts output from the training jobs are stored in Amazon Simple Storage Service (Amazon S3). All data that the training jobs write toAmazon CloudWatch Logs are still available in CloudWatch. After the tuning job moves to the Stopped state, it releases all reserved resources for the tuning job.

", "StopNotebookInstance": "

Terminates the ML compute instance. Before terminating the instance, Amazon SageMaker disconnects the ML storage volume from it. Amazon SageMaker preserves the ML storage volume.

To access data on the ML storage volume for a notebook instance that has been terminated, call the StartNotebookInstance API. StartNotebookInstance launches another ML compute instance, configures it, and attaches the preserved ML storage volume so you can continue your work.

", "StopTrainingJob": "

Stops a training job. To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for 120 seconds. Algorithms might use this 120-second window to save the model artifacts, so the results of the training is not lost.

Training algorithms provided by Amazon SageMaker save the intermediate results of a model training job. This intermediate data is a valid model artifact. You can use the model artifacts that are saved when Amazon SageMaker stops a training job to create a model.

When it receives a StopTrainingJob request, Amazon SageMaker changes the status of the job to Stopping. After Amazon SageMaker stops the job, it sets the status to Stopped.

", "UpdateEndpoint": "

Deploys the new EndpointConfig specified in the request, switches to using newly created endpoint, and then deletes resources provisioned for the endpoint using the previous EndpointConfig (there is no availability loss).

When Amazon SageMaker receives the request, it sets the endpoint status to Updating. After updating the endpoint, it sets the status to InService. To check the status of an endpoint, use the DescribeEndpoint API.

", @@ -51,16 +56,29 @@ "AlgorithmImage": { "base": null, "refs": { - "AlgorithmSpecification$TrainingImage": "

The registry path of the Docker image that contains the training algorithm. For information about docker registry paths for built-in algorithms, see sagemaker-algo-docker-registry-paths.

" + "AlgorithmSpecification$TrainingImage": "

The registry path of the Docker image that contains the training algorithm. For information about docker registry paths for built-in algorithms, see sagemaker-algo-docker-registry-paths.

", + "HyperParameterAlgorithmSpecification$TrainingImage": "

The registry path of the Docker image that contains the training algorithm. For information about Docker registry paths for built-in algorithms, see sagemaker-algo-docker-registry-paths.

" } }, "AlgorithmSpecification": { - "base": "

Specifies the training algorithm to use in a CreateTrainingJob request.

For more information about algorithms provided by Amazon SageMaker, see Algorithms. For information about using your own algorithms, see your-algorithms.

", + "base": "

Specifies the training algorithm to use in a CreateTrainingJob request.

For more information about algorithms provided by Amazon SageMaker, see Algorithms. For information about using your own algorithms, see your-algorithms.

", "refs": { "CreateTrainingJobRequest$AlgorithmSpecification": "

The registry path of the Docker image that contains the training algorithm and algorithm-specific metadata, including the input mode. For more information about algorithms provided by Amazon SageMaker, see Algorithms. For information about providing your own algorithms, see your-algorithms.

", "DescribeTrainingJobResponse$AlgorithmSpecification": "

Information about the algorithm used for training, and algorithm metadata.

" } }, + "CategoricalParameterRange": { + "base": "

A list of categorical hyperparameters to tune.

", + "refs": { + "CategoricalParameterRanges$member": null + } + }, + "CategoricalParameterRanges": { + "base": null, + "refs": { + "ParameterRanges$CategoricalParameterRanges": "

The array of CategoricalParameterRange objects that specify ranges of categorical hyperparameters that a hyperparameter tuning job searches.

" + } + }, "Channel": { "base": "

A channel is a named input source that training algorithms can consume.

", "refs": { @@ -76,7 +94,7 @@ "CompressionType": { "base": null, "refs": { - "Channel$CompressionType": "

If training data is compressed, the compression type. The default value is None. CompressionType is used only in PIPE input mode. In FILE mode, leave this field unset or set it to None.

" + "Channel$CompressionType": "

If training data is compressed, the compression type. The default value is None. CompressionType is used only in Pipe input mode. In File mode, leave this field unset or set it to None.

" } }, "ContainerDefinition": { @@ -98,6 +116,18 @@ "Channel$ContentType": "

The MIME type of the data.

" } }, + "ContinuousParameterRange": { + "base": "

A list of continuous hyperparameters to tune.

", + "refs": { + "ContinuousParameterRanges$member": null + } + }, + "ContinuousParameterRanges": { + "base": null, + "refs": { + "ParameterRanges$ContinuousParameterRanges": "

The array of ContinuousParameterRange objects that specify ranges of continuous hyperparameters that a hyperparameter tuning job searches.

" + } + }, "CreateEndpointConfigInput": { "base": null, "refs": { @@ -118,6 +148,16 @@ "refs": { } }, + "CreateHyperParameterTuningJobRequest": { + "base": null, + "refs": { + } + }, + "CreateHyperParameterTuningJobResponse": { + "base": null, + "refs": { + } + }, "CreateModelInput": { "base": null, "refs": { @@ -242,6 +282,16 @@ "refs": { } }, + "DescribeHyperParameterTuningJobRequest": { + "base": null, + "refs": { + } + }, + "DescribeHyperParameterTuningJobResponse": { + "base": null, + "refs": { + } + }, "DescribeModelInput": { "base": null, "refs": { @@ -422,18 +472,127 @@ "base": null, "refs": { "DescribeEndpointOutput$FailureReason": "

If the status of the endpoint is Failed, the reason why it failed.

", + "DescribeHyperParameterTuningJobResponse$FailureReason": "

If the tuning job failed, the reason it failed.

", "DescribeNotebookInstanceOutput$FailureReason": "

If status is failed, the reason it failed.

", "DescribeTrainingJobResponse$FailureReason": "

If the training job failed, the reason it failed.

", + "HyperParameterTrainingJobSummary$FailureReason": "

The reason that the

", "ResourceInUse$Message": null, "ResourceLimitExceeded$Message": null, "ResourceNotFound$Message": null } }, + "FinalHyperParameterTuningJobObjectiveMetric": { + "base": "

Shows the final value for the objective metric for a training job that was launched by a hyperparameter tuning job. You define the objective metric in the HyperParameterTuningJobObjective parameter of HyperParameterTuningJobConfig.

", + "refs": { + "HyperParameterTrainingJobSummary$FinalHyperParameterTuningJobObjectiveMetric": "

The object that specifies the value of the objective metric of the tuning job that launched this training job.

" + } + }, + "HyperParameterAlgorithmSpecification": { + "base": "

Specifies which training algorithm to use for training jobs that a hyperparameter tuning job launches and the metrics to monitor.

", + "refs": { + "HyperParameterTrainingJobDefinition$AlgorithmSpecification": "

The object that specifies the algorithm to use for the training jobs that the tuning job launches.

" + } + }, + "HyperParameterTrainingJobDefinition": { + "base": "

Defines the training jobs launched by a hyperparameter tuning job.

", + "refs": { + "CreateHyperParameterTuningJobRequest$TrainingJobDefinition": "

The object that describes the training jobs that this tuning job launches, including static hyperparameters, input data configuration, output data configuration, resource configuration, and stopping condition.

", + "DescribeHyperParameterTuningJobResponse$TrainingJobDefinition": "

The object that specifies the definition of the training jobs that this tuning job launches.

" + } + }, + "HyperParameterTrainingJobSummaries": { + "base": null, + "refs": { + "ListTrainingJobsForHyperParameterTuningJobResponse$TrainingJobSummaries": "

A list of objects that describe the training jobs that the ListTrainingJobsForHyperParameterTuningJob request returned.

" + } + }, + "HyperParameterTrainingJobSummary": { + "base": "

Specifies summary information about a training job.

", + "refs": { + "DescribeHyperParameterTuningJobResponse$BestTrainingJob": "

A object that describes the training job that completed with the best current .

", + "HyperParameterTrainingJobSummaries$member": null + } + }, + "HyperParameterTuningJobArn": { + "base": null, + "refs": { + "CreateHyperParameterTuningJobResponse$HyperParameterTuningJobArn": "

The Amazon Resource Name (ARN) of the tuning job.

", + "DescribeHyperParameterTuningJobResponse$HyperParameterTuningJobArn": "

The Amazon Resource Name (ARN) of the tuning job.

", + "DescribeTrainingJobResponse$TuningJobArn": "

The Amazon Resource Name (ARN) of the associated hyperparameter tuning job if the training job was launched by a hyperparameter tuning job.

", + "HyperParameterTuningJobSummary$HyperParameterTuningJobArn": "

The Amazon Resource Name (ARN) of the tuning job.

" + } + }, + "HyperParameterTuningJobConfig": { + "base": "

Configures a hyperparameter tuning job.

", + "refs": { + "CreateHyperParameterTuningJobRequest$HyperParameterTuningJobConfig": "

The object that describes the tuning job, including the search strategy, metric used to evaluate training jobs, ranges of parameters to search, and resource limits for the tuning job.

", + "DescribeHyperParameterTuningJobResponse$HyperParameterTuningJobConfig": "

The object that specifies the configuration of the tuning job.

" + } + }, + "HyperParameterTuningJobName": { + "base": null, + "refs": { + "CreateHyperParameterTuningJobRequest$HyperParameterTuningJobName": "

The name of the tuning job. This name is the prefix for the names of all training jobs that this tuning job launches. The name must be unique within the same AWS account and AWS Region. Names are not case sensitive, and must be between 1-32 characters.

", + "DescribeHyperParameterTuningJobRequest$HyperParameterTuningJobName": "

The name of the tuning job to describe.

", + "DescribeHyperParameterTuningJobResponse$HyperParameterTuningJobName": "

The name of the tuning job.

", + "HyperParameterTuningJobSummary$HyperParameterTuningJobName": "

The name of the tuning job.

", + "ListTrainingJobsForHyperParameterTuningJobRequest$HyperParameterTuningJobName": "

The name of the tuning job whose training jobs you want to list.

", + "StopHyperParameterTuningJobRequest$HyperParameterTuningJobName": "

The name of the tuning job to stop.

" + } + }, + "HyperParameterTuningJobObjective": { + "base": "

Defines the objective metric for a hyperparameter tuning job. Hyperparameter tuning uses the value of this metric to evaluate the training jobs it launches, and returns the training job that results in either the highest or lowest value for this metric, depending on the value you specify for the Type parameter.

", + "refs": { + "HyperParameterTuningJobConfig$HyperParameterTuningJobObjective": "

The object that specifies the objective metric for this tuning job.

" + } + }, + "HyperParameterTuningJobObjectiveType": { + "base": null, + "refs": { + "FinalHyperParameterTuningJobObjectiveMetric$Type": "

Whether to minimize or maximize the objective metric. Valid values are Minimize and Maximize.

", + "HyperParameterTuningJobObjective$Type": "

Whether to minimize or maximize the objective metric.

" + } + }, + "HyperParameterTuningJobSortByOptions": { + "base": null, + "refs": { + "ListHyperParameterTuningJobsRequest$SortBy": "

The field to sort results by. The default is Name.

" + } + }, + "HyperParameterTuningJobStatus": { + "base": null, + "refs": { + "DescribeHyperParameterTuningJobResponse$HyperParameterTuningJobStatus": "

The status of the tuning job: InProgress, Completed, Failed, Stopping, or Stopped.

", + "HyperParameterTuningJobSummary$HyperParameterTuningJobStatus": "

The status of the tuning job.

", + "ListHyperParameterTuningJobsRequest$StatusEquals": "

A filter that returns only tuning jobs with the specified status.

" + } + }, + "HyperParameterTuningJobStrategyType": { + "base": "

The strategy hyperparameter tuning uses to find the best combination of hyperparameters for your model. Currently, the only supported value is Bayesian.

", + "refs": { + "HyperParameterTuningJobConfig$Strategy": "

Specifies the search strategy for hyperparameters. Currently, the only valid value is Bayesian.

", + "HyperParameterTuningJobSummary$Strategy": "

Specifies the search strategy hyperparameter tuning uses to choose which hyperparameters to use for each iteration. Currently, the only valid value is Bayesian.

" + } + }, + "HyperParameterTuningJobSummaries": { + "base": null, + "refs": { + "ListHyperParameterTuningJobsResponse$HyperParameterTuningJobSummaries": "

A list of objects that describe the tuning jobs that the ListHyperParameterTuningJobs request returned.

" + } + }, + "HyperParameterTuningJobSummary": { + "base": "

Provides summary information about a hyperparameter tuning job.

", + "refs": { + "HyperParameterTuningJobSummaries$member": null + } + }, "HyperParameters": { "base": null, "refs": { - "CreateTrainingJobRequest$HyperParameters": "

Algorithm-specific parameters. You set hyperparameters before you start the learning process. Hyperparameters influence the quality of the model. For a list of hyperparameters for each training algorithm provided by Amazon SageMaker, see Algorithms.

You can specify a maximum of 100 hyperparameters. Each hyperparameter is a key-value pair. Each key and value is limited to 256 characters, as specified by the Length Constraint.

", - "DescribeTrainingJobResponse$HyperParameters": "

Algorithm-specific parameters.

" + "CreateTrainingJobRequest$HyperParameters": "

Algorithm-specific parameters that influence the quality of the model. You set hyperparameters before you start the learning process. For a list of hyperparameters for each training algorithm provided by Amazon SageMaker, see Algorithms.

You can specify a maximum of 100 hyperparameters. Each hyperparameter is a key-value pair. Each key and value is limited to 256 characters, as specified by the Length Constraint.

", + "DescribeTrainingJobResponse$HyperParameters": "

Algorithm-specific parameters.

", + "HyperParameterTrainingJobDefinition$StaticHyperParameters": "

Specifies the values of hyperparameters that do not change for the tuning job.

", + "HyperParameterTrainingJobSummary$TunedHyperParameters": "

A list of the hyperparameters for which you specified ranges to search.

" } }, "Image": { @@ -446,7 +605,8 @@ "base": null, "refs": { "CreateTrainingJobRequest$InputDataConfig": "

An array of Channel objects. Each channel is a named input source. InputDataConfig describes the input data and its location.

Algorithms can accept input data from one or more channels. For example, an algorithm might have two channels of input data, training_data and validation_data. The configuration for each channel provides the S3 location where the input data is stored. It also provides information about the stored data: the MIME type, compression method, and whether the data is wrapped in RecordIO format.

Depending on the input mode that the algorithm supports, Amazon SageMaker either copies input data files from an S3 bucket to a local directory in the Docker container, or makes it available as input streams.

", - "DescribeTrainingJobResponse$InputDataConfig": "

An array of Channel objects that describes each data input channel.

" + "DescribeTrainingJobResponse$InputDataConfig": "

An array of Channel objects that describes each data input channel.

", + "HyperParameterTrainingJobDefinition$InputDataConfig": "

An array of objects that specify the input for the training jobs that the tuning job launches.

" } }, "InstanceType": { @@ -458,6 +618,18 @@ "UpdateNotebookInstanceInput$InstanceType": "

The Amazon ML compute instance type.

" } }, + "IntegerParameterRange": { + "base": "

For a hyperparameter of the integer type, specifies the range that a hyperparameter tuning job searches.

", + "refs": { + "IntegerParameterRanges$member": null + } + }, + "IntegerParameterRanges": { + "base": null, + "refs": { + "ParameterRanges$IntegerParameterRanges": "

The array of IntegerParameterRange objects that specify ranges of integer hyperparameters that a hyperparameter tuning job searches.

" + } + }, "KmsKeyId": { "base": null, "refs": { @@ -465,7 +637,7 @@ "CreateNotebookInstanceInput$KmsKeyId": "

If you provide a AWS KMS key ID, Amazon SageMaker uses it to encrypt data at rest on the ML storage volume that is attached to your notebook instance.

", "DescribeEndpointConfigOutput$KmsKeyId": "

AWS KMS key ID Amazon SageMaker uses to encrypt data when storing it on the ML storage volume attached to the instance.

", "DescribeNotebookInstanceOutput$KmsKeyId": "

AWS KMS key ID Amazon SageMaker uses to encrypt data when storing it on the ML storage volume attached to the instance.

", - "OutputDataConfig$KmsKeyId": "

The AWS Key Management Service (AWS KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption.

If the configuration of the output S3 bucket requires server-side encryption for objects, and you don't provide the KMS key ID, Amazon SageMaker uses the default service key. For more information, see KMS-Managed Encryption Keys in Amazon Simple Storage Service developer guide.

The KMS key policy must grant permission to the IAM role you specify in your CreateTrainingJob request. Using Key Policies in AWS KMS in the AWS Key Management Service Developer Guide.

", + "OutputDataConfig$KmsKeyId": "

The AWS Key Management Service (AWS KMS) key that Amazon SageMaker uses to encrypt the model artifacts at rest using Amazon S3 server-side encryption.

If you don't provide the KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's account. For more information, see KMS-Managed Encryption Keys in Amazon Simple Storage Service developer guide.

The KMS key policy must grant permission to the IAM role you specify in your CreateTrainingJob request. Using Key Policies in AWS KMS in the AWS Key Management Service Developer Guide.

", "ResourceConfig$VolumeKmsKeyId": "

The Amazon Resource Name (ARN) of a AWS Key Management Service key that Amazon SageMaker uses to encrypt data on the storage volume attached to the ML compute instance(s) that run the training job.

" } }, @@ -502,6 +674,16 @@ "refs": { } }, + "ListHyperParameterTuningJobsRequest": { + "base": null, + "refs": { + } + }, + "ListHyperParameterTuningJobsResponse": { + "base": null, + "refs": { + } + }, "ListModelsInput": { "base": null, "refs": { @@ -548,6 +730,16 @@ "refs": { } }, + "ListTrainingJobsForHyperParameterTuningJobRequest": { + "base": null, + "refs": { + } + }, + "ListTrainingJobsForHyperParameterTuningJobResponse": { + "base": null, + "refs": { + } + }, "ListTrainingJobsRequest": { "base": null, "refs": { @@ -558,14 +750,28 @@ "refs": { } }, + "MaxNumberOfTrainingJobs": { + "base": null, + "refs": { + "ResourceLimits$MaxNumberOfTrainingJobs": "

The maximum number of training jobs that a hyperparameter tuning job can launch.

" + } + }, + "MaxParallelTrainingJobs": { + "base": null, + "refs": { + "ResourceLimits$MaxParallelTrainingJobs": "

The maximum number of concurrent training jobs that a hyperparameter tuning job can launch.

" + } + }, "MaxResults": { "base": null, "refs": { "ListEndpointConfigsInput$MaxResults": "

The maximum number of training jobs to return in the response.

", "ListEndpointsInput$MaxResults": "

The maximum number of endpoints to return in the response.

", + "ListHyperParameterTuningJobsRequest$MaxResults": "

The maximum number of tuning jobs to return.

", "ListModelsInput$MaxResults": "

The maximum number of models to return in the response.

", "ListNotebookInstanceLifecycleConfigsInput$MaxResults": "

The maximum number of lifecycle configurations to return in the response.

", "ListNotebookInstancesInput$MaxResults": "

The maximum number of notebook instances to return.

", + "ListTrainingJobsForHyperParameterTuningJobRequest$MaxResults": "

The maximum number of training jobs to return.

", "ListTrainingJobsRequest$MaxResults": "

The maximum number of training jobs to return in the response.

" } }, @@ -575,6 +781,38 @@ "StoppingCondition$MaxRuntimeInSeconds": "

The maximum length of time, in seconds, that the training job can run. If model training does not complete during this time, Amazon SageMaker ends the job. If value is not specified, default value is 1 day. Maximum value is 5 days.

" } }, + "MetricDefinition": { + "base": "

Specifies a metric that the training algorithm writes to stderr or stdout. Amazon SageMakerHyperparamter tuning captures all defined metrics. You specify one metric that a hyperparameter tuning job uses as its objective metric to choose the best training job.

", + "refs": { + "MetricDefinitionList$member": null + } + }, + "MetricDefinitionList": { + "base": null, + "refs": { + "HyperParameterAlgorithmSpecification$MetricDefinitions": "

An array of objects that specify the metrics that the algorithm emits.

" + } + }, + "MetricName": { + "base": null, + "refs": { + "FinalHyperParameterTuningJobObjectiveMetric$MetricName": "

The name of the objective metric.

", + "HyperParameterTuningJobObjective$MetricName": "

The name of the metric to use for the objective metric.

", + "MetricDefinition$Name": "

The name of the metric.

" + } + }, + "MetricRegex": { + "base": null, + "refs": { + "MetricDefinition$Regex": "

A regular expression that searches the output of a training job and gets the value of the metric. For more information about using regular expressions to define metrics, see hpo-define-metrics.

" + } + }, + "MetricValue": { + "base": null, + "refs": { + "FinalHyperParameterTuningJobObjectiveMetric$Value": "

The value of the objective metric.

" + } + }, "ModelArn": { "base": null, "refs": { @@ -627,7 +865,8 @@ "NameContains": { "base": null, "refs": { - "ListTrainingJobsRequest$NameContains": "

A string in the training job name. This filter returns only models whose name contains the specified string.

" + "ListHyperParameterTuningJobsRequest$NameContains": "

A string in the tuning job name. This filter returns only tuning jobs whose name contains the specified string.

", + "ListTrainingJobsRequest$NameContains": "

A string in the training job name. This filter returns only training jobs whose name contains the specified string.

" } }, "NetworkInterfaceId": { @@ -639,12 +878,16 @@ "NextToken": { "base": null, "refs": { + "ListHyperParameterTuningJobsRequest$NextToken": "

If the result of the previous ListHyperParameterTuningJobs request was truncated, the response includes a NextToken. To retrieve the next set of tuning jobs, use the token in the next request.

", + "ListHyperParameterTuningJobsResponse$NextToken": "

If the result of this ListHyperParameterTuningJobs request was truncated, the response includes a NextToken. To retrieve the next set of tuning jobs, use the token in the next request.

", "ListNotebookInstanceLifecycleConfigsInput$NextToken": "

If the result of a ListNotebookInstanceLifecycleConfigs request was truncated, the response includes a NextToken. To get the next set of lifecycle configurations, use the token in the next request.

", "ListNotebookInstanceLifecycleConfigsOutput$NextToken": "

If the response is truncated, Amazon SageMaker returns this token. To get the next set of lifecycle configurations, use it in the next request.

", "ListNotebookInstancesInput$NextToken": "

If the previous call to the ListNotebookInstances is truncated, the response includes a NextToken. You can use this token in your subsequent ListNotebookInstances request to fetch the next set of notebook instances.

You might specify a filter or a sort order in your request. When response is truncated, you must use the same values for the filer and sort order in the next request.

", "ListNotebookInstancesOutput$NextToken": "

If the response to the previous ListNotebookInstances request was truncated, Amazon SageMaker returns this token. To retrieve the next set of notebook instances, use the token in the next request.

", "ListTagsInput$NextToken": "

If the response to the previous ListTags request is truncated, Amazon SageMaker returns this token. To retrieve the next set of tags, use it in the subsequent request.

", "ListTagsOutput$NextToken": "

If response is truncated, Amazon SageMaker includes a token in the response. You can use this token in your subsequent request to fetch next set of tokens.

", + "ListTrainingJobsForHyperParameterTuningJobRequest$NextToken": "

If the result of the previous ListTrainingJobsForHyperParameterTuningJob request was truncated, the response includes a NextToken. To retrieve the next set of training jobs, use the token in the next request.

", + "ListTrainingJobsForHyperParameterTuningJobResponse$NextToken": "

If the result of this ListTrainingJobsForHyperParameterTuningJob request was truncated, the response includes a NextToken. To retrieve the next set of training jobs, use the token in the next request.

", "ListTrainingJobsRequest$NextToken": "

If the result of the previous ListTrainingJobs request was truncated, the response includes a NextToken. To retrieve the next set of training jobs, use the token in the next request.

", "ListTrainingJobsResponse$NextToken": "

If the response is truncated, Amazon SageMaker returns this token. To retrieve the next set of training jobs, use it in the subsequent request.

" } @@ -793,6 +1036,27 @@ "NotebookInstanceSummary$Url": "

The URL that you use to connect to the Jupyter instance running in your notebook instance.

" } }, + "ObjectiveStatus": { + "base": null, + "refs": { + "HyperParameterTrainingJobSummary$ObjectiveStatus": "

The status of the objective metric for the training job:

  • Succeeded: The final objective metric for the training job was evaluated by the hyperparameter tuning job and used in the hyperparameter tuning process.

  • Pending: The training job is in progress and evaluation of its final objective metric is pending.

  • Failed: The final objective metric for the training job was not evaluated, and was not used in the hyperparameter tuning process. This typically occurs when the training job failed or did not emit an objective metric.

" + } + }, + "ObjectiveStatusCounter": { + "base": null, + "refs": { + "ObjectiveStatusCounters$Succeeded": "

The number of training jobs whose final objective metric was evaluated by the hyperparameter tuning job and used in the hyperparameter tuning process.

", + "ObjectiveStatusCounters$Pending": "

The number of training jobs that are in progress and pending evaluation of their final objective metric.

", + "ObjectiveStatusCounters$Failed": "

The number of training jobs whose final objective metric was not evaluated and used in the hyperparameter tuning process. This typically occurs when the training job failed or did not emit an objective metric.

" + } + }, + "ObjectiveStatusCounters": { + "base": "

Specifies the number of training jobs that this hyperparameter tuning job launched, categorized by the status of their objective metric. The objective metric status shows whether the final objective metric for the training job has been evaluated by the tuning job and used in the hyperparameter tuning process.

", + "refs": { + "DescribeHyperParameterTuningJobResponse$ObjectiveStatusCounters": "

The object that specifies the number of training jobs, categorized by the status of their final objective metric, that this tuning job launched.

", + "HyperParameterTuningJobSummary$ObjectiveStatusCounters": "

The object that specifies the numbers of training jobs, categorized by objective metric status, that this tuning job launched.

" + } + }, "OrderKey": { "base": null, "refs": { @@ -805,7 +1069,8 @@ "base": "

Provides information about how to store model training results (model artifacts).

", "refs": { "CreateTrainingJobRequest$OutputDataConfig": "

Specifies the path to the S3 bucket where you want to store model artifacts. Amazon SageMaker creates subfolders for the artifacts.

", - "DescribeTrainingJobResponse$OutputDataConfig": "

The S3 path where model artifacts that you configured when creating the job are stored. Amazon SageMaker creates subfolders for model artifacts.

" + "DescribeTrainingJobResponse$OutputDataConfig": "

The S3 path where model artifacts that you configured when creating the job are stored. Amazon SageMaker creates subfolders for model artifacts.

", + "HyperParameterTrainingJobDefinition$OutputDataConfig": "

Specifies the path to the Amazon S3 bucket where you store model artifacts from the training jobs that the tuning job launches.

" } }, "PaginationToken": { @@ -822,13 +1087,33 @@ "ParameterKey": { "base": null, "refs": { - "HyperParameters$key": null + "CategoricalParameterRange$Name": "

The name of the categorical hyperparameter to tune.

", + "ContinuousParameterRange$Name": "

The name of the continuous hyperparameter to tune.

", + "HyperParameters$key": null, + "IntegerParameterRange$Name": "

The name of the hyperparameter to search.

" + } + }, + "ParameterRanges": { + "base": "

Specifies ranges of integer, continuous, and categorical hyperparameters that a hyperparameter tuning job searches.

", + "refs": { + "HyperParameterTuningJobConfig$ParameterRanges": "

The object that specifies the ranges of hyperparameters that this tuning job searches.

" } }, "ParameterValue": { "base": null, "refs": { - "HyperParameters$value": null + "ContinuousParameterRange$MinValue": "

The minimum value for the hyperparameter. The tuning job uses floating-point values between this value and MaxValuefor tuning.

", + "ContinuousParameterRange$MaxValue": "

The maximum value for the hyperparameter. The tuning job uses floating-point values between MinValue value and this value for tuning.

", + "HyperParameters$value": null, + "IntegerParameterRange$MinValue": "

The minimum value of the hyperparameter to search.

", + "IntegerParameterRange$MaxValue": "

The maximum value of the hyperparameter to search.

", + "ParameterValues$member": null + } + }, + "ParameterValues": { + "base": null, + "refs": { + "CategoricalParameterRange$Values": "

A list of the categories for the hyperparameter.

" } }, "ProductionVariant": { @@ -865,7 +1150,7 @@ "RecordWrapper": { "base": null, "refs": { - "Channel$RecordWrapperType": "

Specify RecordIO as the value when input data is in raw format but the training algorithm requires the RecordIO format, in which caseAmazon SageMaker wraps each individual S3 object in a RecordIO record. If the input data is already in RecordIO format, you don't need to set this attribute. For more information, see Create a Dataset Using RecordIO.

In FILE mode, leave this field unset or set it to None.

" + "Channel$RecordWrapperType": "

Specify RecordIO as the value when input data is in raw format but the training algorithm requires the RecordIO format, in which case, Amazon SageMaker wraps each individual S3 object in a RecordIO record. If the input data is already in RecordIO format, you don't need to set this attribute. For more information, see Create a Dataset Using RecordIO.

In FILE mode, leave this field unset or set it to None.

" } }, "ResourceArn": { @@ -880,7 +1165,8 @@ "base": "

Describes the resources, including ML compute instances and ML storage volumes, to use for model training.

", "refs": { "CreateTrainingJobRequest$ResourceConfig": "

The resources, including the ML compute instances and ML storage volumes, to use for model training.

ML storage volumes store model artifacts and incremental states. Training algorithms might also use ML storage volumes for scratch space. If you want Amazon SageMaker to use the ML storage volume to store the training data, choose File as the TrainingInputMode in the algorithm specification. For distributed training algorithms, specify an instance count greater than 1.

", - "DescribeTrainingJobResponse$ResourceConfig": "

Resources, including ML compute instances and ML storage volumes, that are configured for model training.

" + "DescribeTrainingJobResponse$ResourceConfig": "

Resources, including ML compute instances and ML storage volumes, that are configured for model training.

", + "HyperParameterTrainingJobDefinition$ResourceConfig": "

The resources, including the compute instances and storage volumes, to use for the training jobs that the tuning job launches.

Storage volumes store model artifacts and incremental states. Training algorithms might also use storage volumes for scratch space. If you want Amazon SageMaker to use the storage volume to store the training data, choose File as the TrainingInputMode in the algorithm specification. For distributed training algorithms, specify an instance count greater than 1.

" } }, "ResourceInUse": { @@ -893,6 +1179,13 @@ "refs": { } }, + "ResourceLimits": { + "base": "

Specifies the maximum number of training jobs and parallel training jobs that a hyperparameter tuning job can launch.

", + "refs": { + "HyperParameterTuningJobConfig$ResourceLimits": "

The object that specifies the maximum number of training jobs and parallel training jobs for this tuning job.

", + "HyperParameterTuningJobSummary$ResourceLimits": "

The object that specifies the maximum number of training jobs and parallel training jobs allowed for this tuning job.

" + } + }, "ResourceNotFound": { "base": "

Resource being access is not found.

", "refs": { @@ -901,13 +1194,14 @@ "RoleArn": { "base": null, "refs": { - "CreateModelInput$ExecutionRoleArn": "

The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker can assume to access model artifacts and docker image for deployment on ML compute instances. Deploying on ML compute instances is part of model hosting. For more information, see Amazon SageMaker Roles.

", - "CreateNotebookInstanceInput$RoleArn": "

When you send any requests to AWS resources from the notebook instance, Amazon SageMaker assumes this role to perform tasks on your behalf. You must grant this role necessary permissions so Amazon SageMaker can perform these tasks. The policy must allow the Amazon SageMaker service principal (sagemaker.amazonaws.com) permissions to assume this role. For more information, see Amazon SageMaker Roles.

", - "CreateTrainingJobRequest$RoleArn": "

The Amazon Resource Name (ARN) of an IAM role that Amazon SageMaker can assume to perform tasks on your behalf.

During model training, Amazon SageMaker needs your permission to read input data from an S3 bucket, download a Docker image that contains training code, write model artifacts to an S3 bucket, write logs to Amazon CloudWatch Logs, and publish metrics to Amazon CloudWatch. You grant permissions for all of these tasks to an IAM role. For more information, see Amazon SageMaker Roles.

", + "CreateModelInput$ExecutionRoleArn": "

The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker can assume to access model artifacts and docker image for deployment on ML compute instances. Deploying on ML compute instances is part of model hosting. For more information, see Amazon SageMaker Roles.

To be able to pass this role to Amazon SageMaker, the caller of this API must have the iam:PassRole permission.

", + "CreateNotebookInstanceInput$RoleArn": "

When you send any requests to AWS resources from the notebook instance, Amazon SageMaker assumes this role to perform tasks on your behalf. You must grant this role necessary permissions so Amazon SageMaker can perform these tasks. The policy must allow the Amazon SageMaker service principal (sagemaker.amazonaws.com) permissions to assume this role. For more information, see Amazon SageMaker Roles.

To be able to pass this role to Amazon SageMaker, the caller of this API must have the iam:PassRole permission.

", + "CreateTrainingJobRequest$RoleArn": "

The Amazon Resource Name (ARN) of an IAM role that Amazon SageMaker can assume to perform tasks on your behalf.

During model training, Amazon SageMaker needs your permission to read input data from an S3 bucket, download a Docker image that contains training code, write model artifacts to an S3 bucket, write logs to Amazon CloudWatch Logs, and publish metrics to Amazon CloudWatch. You grant permissions for all of these tasks to an IAM role. For more information, see Amazon SageMaker Roles.

To be able to pass this role to Amazon SageMaker, the caller of this API must have the iam:PassRole permission.

", "DescribeModelOutput$ExecutionRoleArn": "

The Amazon Resource Name (ARN) of the IAM role that you specified for the model.

", "DescribeNotebookInstanceOutput$RoleArn": "

Amazon Resource Name (ARN) of the IAM role associated with the instance.

", "DescribeTrainingJobResponse$RoleArn": "

The AWS Identity and Access Management (IAM) role configured for the training job.

", - "UpdateNotebookInstanceInput$RoleArn": "

Amazon Resource Name (ARN) of the IAM role to associate with the instance.

" + "HyperParameterTrainingJobDefinition$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role associated with the training jobs that the tuning job launches.

", + "UpdateNotebookInstanceInput$RoleArn": "

The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker can assume to access the notebook instance. For more information, see Amazon SageMaker Roles.

To be able to pass this role to Amazon SageMaker, the caller of this API must have the iam:PassRole permission.

" } }, "S3DataDistribution": { @@ -971,6 +1265,8 @@ "SortOrder": { "base": null, "refs": { + "ListHyperParameterTuningJobsRequest$SortOrder": "

The sort order for results. The default is Ascending.

", + "ListTrainingJobsForHyperParameterTuningJobRequest$SortOrder": "

The sort order for results. The default is Ascending.

", "ListTrainingJobsRequest$SortOrder": "

The sort order for results. The default is Ascending.

" } }, @@ -979,6 +1275,11 @@ "refs": { } }, + "StopHyperParameterTuningJobRequest": { + "base": null, + "refs": { + } + }, "StopNotebookInstanceInput": { "base": null, "refs": { @@ -993,7 +1294,8 @@ "base": "

Specifies how long model training can run. When model training reaches the limit, Amazon SageMaker ends the training job. Use this API to cap model training cost.

To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for120 seconds. Algorithms might use this 120-second window to save the model artifacts, so the results of training is not lost.

Training algorithms provided by Amazon SageMaker automatically saves the intermediate results of a model training job (it is best effort case, as model might not be ready to save as some stages, for example training just started). This intermediate data is a valid model artifact. You can use it to create a model (CreateModel).

", "refs": { "CreateTrainingJobRequest$StoppingCondition": "

Sets a duration for training. Use this parameter to cap model training costs. To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for 120 seconds. Algorithms might use this 120-second window to save the model artifacts.

When Amazon SageMaker terminates a job because the stopping condition has been met, training algorithms provided by Amazon SageMaker save the intermediate results of the job. This intermediate data is a valid model artifact. You can use it to create a model using the CreateModel API.

", - "DescribeTrainingJobResponse$StoppingCondition": "

The condition under which to stop the training job.

" + "DescribeTrainingJobResponse$StoppingCondition": "

The condition under which to stop the training job.

", + "HyperParameterTrainingJobDefinition$StoppingCondition": "

Sets a maximum duration for the training jobs that the tuning job launches. Use this parameter to limit model training costs.

To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal. This delays job termination for 120 seconds. Algorithms might use this 120-second window to save the model artifacts.

When Amazon SageMaker terminates a job because the stopping condition has been met, training algorithms provided by Amazon SageMaker save the intermediate results of the job.

" } }, "SubnetId": { @@ -1036,6 +1338,7 @@ "AddTagsOutput$Tags": "

A list of tags associated with the Amazon SageMaker resource.

", "CreateEndpointConfigInput$Tags": "

An array of key-value pairs. For more information, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.

", "CreateEndpointInput$Tags": "

An array of key-value pairs. For more information, see Using Cost Allocation Tagsin the AWS Billing and Cost Management User Guide.

", + "CreateHyperParameterTuningJobRequest$Tags": "

An array of key-value pairs. You can use tags to categorize your AWS resources in different ways, for example, by purpose, owner, or environment. For more information, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.

", "CreateModelInput$Tags": "

An array of key-value pairs. For more information, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.

", "CreateNotebookInstanceInput$Tags": "

A list of tags to associate with the notebook instance. You can add tags later by using the CreateTags API.

", "CreateTrainingJobRequest$Tags": "

An array of key-value pairs. For more information, see Using Cost Allocation Tags in the AWS Billing and Cost Management User Guide.

", @@ -1063,20 +1366,33 @@ "DescribeEndpointConfigOutput$CreationTime": "

A timestamp that shows when the endpoint configuration was created.

", "DescribeEndpointOutput$CreationTime": "

A timestamp that shows when the endpoint was created.

", "DescribeEndpointOutput$LastModifiedTime": "

A timestamp that shows when the endpoint was last modified.

", + "DescribeHyperParameterTuningJobResponse$CreationTime": "

The date and time that the tuning job started.

", + "DescribeHyperParameterTuningJobResponse$HyperParameterTuningEndTime": "

The date and time that the tuning job ended.

", + "DescribeHyperParameterTuningJobResponse$LastModifiedTime": "

The date and time that the status of the tuning job was modified.

", "DescribeModelOutput$CreationTime": "

A timestamp that shows when the model was created.

", "DescribeTrainingJobResponse$CreationTime": "

A timestamp that indicates when the training job was created.

", - "DescribeTrainingJobResponse$TrainingStartTime": "

A timestamp that indicates when training started.

", - "DescribeTrainingJobResponse$TrainingEndTime": "

A timestamp that indicates when model training ended.

", + "DescribeTrainingJobResponse$TrainingStartTime": "

Indicates the time when the training job starts on training instances. You are billed for the time interval between this time and the value of TrainingEndTime. The start time in CloudWatch Logs might be later than this time. The difference is due to the time it takes to download the training data and to the size of the training container.

", + "DescribeTrainingJobResponse$TrainingEndTime": "

Indicates the time when the training job ends on training instances. You are billed for the time interval between the value of TrainingStartTime and this time. For successful jobs and stopped jobs, this is the time after model artifacts are uploaded. For failed jobs, this is the time when Amazon SageMaker detects a job failure.

", "DescribeTrainingJobResponse$LastModifiedTime": "

A timestamp that indicates when the status of the training job was last modified.

", "EndpointConfigSummary$CreationTime": "

A timestamp that shows when the endpoint configuration was created.

", "EndpointSummary$CreationTime": "

A timestamp that shows when the endpoint was created.

", "EndpointSummary$LastModifiedTime": "

A timestamp that shows when the endpoint was last modified.

", + "HyperParameterTrainingJobSummary$CreationTime": "

The date and time that the training job was created.

", + "HyperParameterTrainingJobSummary$TrainingStartTime": "

The date and time that the training job started.

", + "HyperParameterTrainingJobSummary$TrainingEndTime": "

The date and time that the training job ended.

", + "HyperParameterTuningJobSummary$CreationTime": "

The date and time that the tuning job was created.

", + "HyperParameterTuningJobSummary$HyperParameterTuningEndTime": "

The date and time that the tuning job ended.

", + "HyperParameterTuningJobSummary$LastModifiedTime": "

The date and time that the tuning job was modified.

", "ListEndpointConfigsInput$CreationTimeBefore": "

A filter that returns only endpoint configurations created before the specified time (timestamp).

", "ListEndpointConfigsInput$CreationTimeAfter": "

A filter that returns only endpoint configurations created after the specified time (timestamp).

", "ListEndpointsInput$CreationTimeBefore": "

A filter that returns only endpoints that were created before the specified time (timestamp).

", "ListEndpointsInput$CreationTimeAfter": "

A filter that returns only endpoints that were created after the specified time (timestamp).

", "ListEndpointsInput$LastModifiedTimeBefore": "

A filter that returns only endpoints that were modified before the specified timestamp.

", "ListEndpointsInput$LastModifiedTimeAfter": "

A filter that returns only endpoints that were modified after the specified timestamp.

", + "ListHyperParameterTuningJobsRequest$CreationTimeAfter": "

A filter that returns only tuning jobs that were created after the specified time.

", + "ListHyperParameterTuningJobsRequest$CreationTimeBefore": "

A filter that returns only tuning jobs that were created before the specified time.

", + "ListHyperParameterTuningJobsRequest$LastModifiedTimeAfter": "

A filter that returns only tuning jobs that were modified after the specified time.

", + "ListHyperParameterTuningJobsRequest$LastModifiedTimeBefore": "

A filter that returns only tuning jobs that were modified before the specified time.

", "ListModelsInput$CreationTimeBefore": "

A filter that returns only models created before the specified time (timestamp).

", "ListModelsInput$CreationTimeAfter": "

A filter that returns only models created after the specified time (timestamp).

", "ListTrainingJobsRequest$CreationTimeAfter": "

A filter that only training jobs created after the specified time (timestamp).

", @@ -1092,7 +1408,8 @@ "TrainingInputMode": { "base": null, "refs": { - "AlgorithmSpecification$TrainingInputMode": "

The input mode that the algorithm supports. For the input modes that Amazon SageMaker algorithms support, see Algorithms. If an algorithm supports the File input mode, Amazon SageMaker downloads the training data from S3 to the provisioned ML storage Volume, and mounts the directory to docker volume for training container. If an algorithm supports the Pipe input mode, Amazon SageMaker streams data directly from S3 to the container.

In File mode, make sure you provision ML storage volume with sufficient capacity to accommodate the data download from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container use ML storage volume to also store intermediate information, if any.

For distributed algorithms using File mode, training data is distributed uniformly, and your training duration is predictable if the input data objects size is approximately same. Amazon SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed where one host in a training cluster is overloaded, thus becoming bottleneck in training.

" + "AlgorithmSpecification$TrainingInputMode": "

The input mode that the algorithm supports. For the input modes that Amazon SageMaker algorithms support, see Algorithms. If an algorithm supports the File input mode, Amazon SageMaker downloads the training data from S3 to the provisioned ML storage Volume, and mounts the directory to docker volume for training container. If an algorithm supports the Pipe input mode, Amazon SageMaker streams data directly from S3 to the container.

In File mode, make sure you provision ML storage volume with sufficient capacity to accommodate the data download from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container use ML storage volume to also store intermediate information, if any.

For distributed algorithms using File mode, training data is distributed uniformly, and your training duration is predictable if the input data objects size is approximately same. Amazon SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed where one host in a training cluster is overloaded, thus becoming bottleneck in training.

", + "HyperParameterAlgorithmSpecification$TrainingInputMode": "

The input mode that the algorithm supports: File or Pipe. In File input mode, Amazon SageMaker downloads the training data from Amazon S3 to the storage volume that is attached to the training instance and mounts the directory to the Docker volume for the training container. In Pipe input mode, Amazon SageMaker streams data directly from Amazon S3 to the container.

If you specify File mode, make sure that you provision the storage volume that is attached to the training instance with enough capacity to accommodate the training data downloaded from Amazon S3, the model artifacts, and intermediate information.

For more information about input modes, see Algorithms.

" } }, "TrainingInstanceCount": { @@ -1112,6 +1429,7 @@ "refs": { "CreateTrainingJobResponse$TrainingJobArn": "

The Amazon Resource Name (ARN) of the training job.

", "DescribeTrainingJobResponse$TrainingJobArn": "

The Amazon Resource Name (ARN) of the training job.

", + "HyperParameterTrainingJobSummary$TrainingJobArn": "

The Amazon Resource Name (ARN) of the training job.

", "TrainingJobSummary$TrainingJobArn": "

The Amazon Resource Name (ARN) of the training job.

" } }, @@ -1121,18 +1439,44 @@ "CreateTrainingJobRequest$TrainingJobName": "

The name of the training job. The name must be unique within an AWS Region in an AWS account. It appears in the Amazon SageMaker console.

", "DescribeTrainingJobRequest$TrainingJobName": "

The name of the training job.

", "DescribeTrainingJobResponse$TrainingJobName": "

Name of the model training job.

", + "HyperParameterTrainingJobSummary$TrainingJobName": "

The name of the training job.

", "StopTrainingJobRequest$TrainingJobName": "

The name of the training job to stop.

", "TrainingJobSummary$TrainingJobName": "

The name of the training job that you want a summary for.

" } }, + "TrainingJobSortByOptions": { + "base": null, + "refs": { + "ListTrainingJobsForHyperParameterTuningJobRequest$SortBy": "

The field to sort results by. The default is Name.

" + } + }, "TrainingJobStatus": { "base": null, "refs": { "DescribeTrainingJobResponse$TrainingJobStatus": "

The status of the training job.

For the InProgress status, Amazon SageMaker can return these secondary statuses:

  • Starting - Preparing for training.

  • Downloading - Optional stage for algorithms that support File training input mode. It indicates data is being downloaded to ML storage volumes.

  • Training - Training is in progress.

  • Uploading - Training is complete and model upload is in progress.

For the Stopped training status, Amazon SageMaker can return these secondary statuses:

  • MaxRuntimeExceeded - Job stopped as a result of maximum allowed runtime exceeded.

", + "HyperParameterTrainingJobSummary$TrainingJobStatus": "

The status of the training job.

", + "ListTrainingJobsForHyperParameterTuningJobRequest$StatusEquals": "

A filter that returns only training jobs with the specified status.

", "ListTrainingJobsRequest$StatusEquals": "

A filter that retrieves only training jobs with a specific status.

", "TrainingJobSummary$TrainingJobStatus": "

The status of the training job.

" } }, + "TrainingJobStatusCounter": { + "base": null, + "refs": { + "TrainingJobStatusCounters$Completed": "

The number of completed training jobs launched by a hyperparameter tuning job.

", + "TrainingJobStatusCounters$InProgress": "

The number of in-progress training jobs launched by a hyperparameter tuning job.

", + "TrainingJobStatusCounters$RetryableError": "

The number of training jobs that failed, but can be retried. A failed training job can be retried only if it failed because an internal service error occurred.

", + "TrainingJobStatusCounters$NonRetryableError": "

The number of training jobs that failed and can't be retried. A failed training job can't be retried if it failed because a client error occurred.

", + "TrainingJobStatusCounters$Stopped": "

The number of training jobs launched by a hyperparameter tuning job that were manually stopped.

" + } + }, + "TrainingJobStatusCounters": { + "base": "

The numbers of training jobs launched by a hyperparameter tuning job, categorized by status.

", + "refs": { + "DescribeHyperParameterTuningJobResponse$TrainingJobStatusCounters": "

The object that specifies the number of training jobs, categorized by status, that this tuning job launched.

", + "HyperParameterTuningJobSummary$TrainingJobStatusCounters": "

The object that specifies the numbers of training jobs, categorized by status, that this tuning job launched.

" + } + }, "TrainingJobSummaries": { "base": null, "refs": { @@ -1217,10 +1561,11 @@ "VpcConfig": { "base": "

Specifies a VPC that your training jobs and hosted models have access to. Control access to and from your training and model containers by configuring the VPC. For more information, see host-vpc and train-vpc.

", "refs": { - "CreateModelInput$VpcConfig": "

A object that specifies the VPC that you want your model to connect to. Control access to and from your training container by configuring the VPC. For more information, see host-vpc.

", + "CreateModelInput$VpcConfig": "

A object that specifies the VPC that you want your model to connect to. Control access to and from your model container by configuring the VPC. For more information, see host-vpc.

", "CreateTrainingJobRequest$VpcConfig": "

A object that specifies the VPC that you want your training job to connect to. Control access to and from your training container by configuring the VPC. For more information, see train-vpc

", "DescribeModelOutput$VpcConfig": "

A object that specifies the VPC that this model has access to. For more information, see host-vpc

", - "DescribeTrainingJobResponse$VpcConfig": "

A object that specifies the VPC that this training job has access to. For more information, see train-vpc.

" + "DescribeTrainingJobResponse$VpcConfig": "

A object that specifies the VPC that this training job has access to. For more information, see train-vpc.

", + "HyperParameterTrainingJobDefinition$VpcConfig": "

The object that specifies the VPC that you want the training jobs that this hyperparameter tuning job launches to connect to. Control access to and from your training container by configuring the VPC. For more information, see train-vpc.

" } }, "VpcSecurityGroupIds": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/paginators-1.json index ae761302e..ddfccd19d 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sagemaker/2017-07-24/paginators-1.json @@ -10,6 +10,11 @@ "output_token": "NextToken", "limit_key": "MaxResults" }, + "ListHyperParameterTuningJobs": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" + }, "ListModels": { "input_token": "NextToken", "output_token": "NextToken", @@ -34,6 +39,11 @@ "input_token": "NextToken", "output_token": "NextToken", "limit_key": "MaxResults" + }, + "ListTrainingJobsForHyperParameterTuningJob": { + "input_token": "NextToken", + "output_token": "NextToken", + "limit_key": "MaxResults" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/api-2.json index 8bbd98ac0..8d706823b 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/api-2.json @@ -44,7 +44,22 @@ {"shape":"ResourceExistsException"}, {"shape":"ResourceNotFoundException"}, {"shape":"MalformedPolicyDocumentException"}, - {"shape":"InternalServiceError"} + {"shape":"InternalServiceError"}, + {"shape":"PreconditionNotMetException"} + ] + }, + "DeleteResourcePolicy":{ + "name":"DeleteResourcePolicy", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DeleteResourcePolicyRequest"}, + "output":{"shape":"DeleteResourcePolicyResponse"}, + "errors":[ + {"shape":"ResourceNotFoundException"}, + {"shape":"InternalServiceError"}, + {"shape":"InvalidRequestException"} ] }, "DeleteSecret":{ @@ -89,6 +104,20 @@ {"shape":"InternalServiceError"} ] }, + "GetResourcePolicy":{ + "name":"GetResourcePolicy", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"GetResourcePolicyRequest"}, + "output":{"shape":"GetResourcePolicyResponse"}, + "errors":[ + {"shape":"ResourceNotFoundException"}, + {"shape":"InternalServiceError"}, + {"shape":"InvalidRequestException"} + ] + }, "GetSecretValue":{ "name":"GetSecretValue", "http":{ @@ -133,6 +162,22 @@ {"shape":"InternalServiceError"} ] }, + "PutResourcePolicy":{ + "name":"PutResourcePolicy", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"PutResourcePolicyRequest"}, + "output":{"shape":"PutResourcePolicyResponse"}, + "errors":[ + {"shape":"MalformedPolicyDocumentException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidParameterException"}, + {"shape":"InternalServiceError"}, + {"shape":"InvalidRequestException"} + ] + }, "PutSecretValue":{ "name":"PutSecretValue", "http":{ @@ -190,6 +235,7 @@ "input":{"shape":"TagResourceRequest"}, "errors":[ {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidRequestException"}, {"shape":"InvalidParameterException"}, {"shape":"InternalServiceError"} ] @@ -203,6 +249,7 @@ "input":{"shape":"UntagResourceRequest"}, "errors":[ {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidRequestException"}, {"shape":"InvalidParameterException"}, {"shape":"InternalServiceError"} ] @@ -223,7 +270,8 @@ {"shape":"ResourceExistsException"}, {"shape":"ResourceNotFoundException"}, {"shape":"MalformedPolicyDocumentException"}, - {"shape":"InternalServiceError"} + {"shape":"InternalServiceError"}, + {"shape":"PreconditionNotMetException"} ] }, "UpdateSecretVersionStage":{ @@ -302,6 +350,20 @@ }, "exception":true }, + "DeleteResourcePolicyRequest":{ + "type":"structure", + "required":["SecretId"], + "members":{ + "SecretId":{"shape":"SecretIdType"} + } + }, + "DeleteResourcePolicyResponse":{ + "type":"structure", + "members":{ + "ARN":{"shape":"SecretARNType"}, + "Name":{"shape":"NameType"} + } + }, "DeleteSecretRequest":{ "type":"structure", "required":["SecretId"], @@ -427,6 +489,21 @@ "RandomPassword":{"shape":"RandomPasswordType"} } }, + "GetResourcePolicyRequest":{ + "type":"structure", + "required":["SecretId"], + "members":{ + "SecretId":{"shape":"SecretIdType"} + } + }, + "GetResourcePolicyResponse":{ + "type":"structure", + "members":{ + "ARN":{"shape":"SecretARNType"}, + "Name":{"shape":"NameType"}, + "ResourcePolicy":{"shape":"NonEmptyResourcePolicyType"} + } + }, "GetSecretValueRequest":{ "type":"structure", "required":["SecretId"], @@ -560,11 +637,41 @@ "max":4096, "min":1 }, + "NonEmptyResourcePolicyType":{ + "type":"string", + "max":4096, + "min":1 + }, "PasswordLengthType":{ "type":"long", "max":4096, "min":1 }, + "PreconditionNotMetException":{ + "type":"structure", + "members":{ + "Message":{"shape":"ErrorMessage"} + }, + "exception":true + }, + "PutResourcePolicyRequest":{ + "type":"structure", + "required":[ + "SecretId", + "ResourcePolicy" + ], + "members":{ + "SecretId":{"shape":"SecretIdType"}, + "ResourcePolicy":{"shape":"NonEmptyResourcePolicyType"} + } + }, + "PutResourcePolicyResponse":{ + "type":"structure", + "members":{ + "ARN":{"shape":"SecretARNType"}, + "Name":{"shape":"NameType"} + } + }, "PutSecretValueRequest":{ "type":"structure", "required":["SecretId"], diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/docs-2.json index 4e2d921f1..13c0ba548 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/docs-2.json @@ -3,20 +3,23 @@ "service": "AWS Secrets Manager API Reference

AWS Secrets Manager is a web service that enables you to store, manage, and retrieve, secrets.

This guide provides descriptions of the Secrets Manager API. For more information about using this service, see the AWS Secrets Manager User Guide.

API Version

This version of the Secrets Manager API Reference documents the Secrets Manager API version 2017-10-17.

As an alternative to using the API directly, you can use one of the AWS SDKs, which consist of libraries and sample code for various programming languages and platforms (such as Java, Ruby, .NET, iOS, and Android). The SDKs provide a convenient way to create programmatic access to AWS Secrets Manager. For example, the SDKs take care of cryptographically signing requests, managing errors, and retrying requests automatically. For more information about the AWS SDKs, including how to download and install them, see Tools for Amazon Web Services.

We recommend that you use the AWS SDKs to make programmatic API calls to Secrets Manager. However, you also can use the Secrets Manager HTTP Query API to make direct calls to the Secrets Manager web service. To learn more about the Secrets Manager HTTP Query API, see Making Query Requests in the AWS Secrets Manager User Guide.

Secrets Manager supports GET and POST requests for all actions. That is, the API doesn't require you to use GET for some actions and POST for others. However, GET requests are subject to the limitation size of a URL. Therefore, for operations that require larger sizes, use a POST request.

Support and Feedback for AWS Secrets Manager

We welcome your feedback. Send your comments to awssecretsmanager-feedback@amazon.com, or post your feedback and questions in the AWS Secrets Manager Discussion Forum. For more information about the AWS Discussion Forums, see Forums Help.

How examples are presented

The JSON that AWS Secrets Manager expects as your request parameters and that the service returns as a response to HTTP query requests are single, long strings without line breaks or white space formatting. The JSON shown in the examples is formatted with both line breaks and white space to improve readability. When example input parameters would also result in long strings that extend beyond the screen, we insert line breaks to enhance readability. You should always submit the input as a single JSON text string.

Logging API Requests

AWS Secrets Manager supports AWS CloudTrail, a service that records AWS API calls for your AWS account and delivers log files to an Amazon S3 bucket. By using information that's collected by AWS CloudTrail, you can determine which requests were successfully made to Secrets Manager, who made the request, when it was made, and so on. For more about AWS Secrets Manager and its support for AWS CloudTrail, see Logging AWS Secrets Manager Events with AWS CloudTrail in the AWS Secrets Manager User Guide. To learn more about CloudTrail, including how to turn it on and find your log files, see the AWS CloudTrail User Guide.

", "operations": { "CancelRotateSecret": "

Disables automatic scheduled rotation and cancels the rotation of a secret if one is currently in progress.

To re-enable scheduled rotation, call RotateSecret with AutomaticallyRotateAfterDays set to a value greater than 0. This will immediately rotate your secret and then enable the automatic schedule.

If you cancel a rotation that is in progress, it can leave the VersionStage labels in an unexpected state. Depending on what step of the rotation was in progress, you might need to remove the staging label AWSPENDING from the partially created version, specified by the SecretVersionId response value. You should also evaluate the partially rotated new version to see if it should be deleted, which you can do by removing all staging labels from the new version's VersionStage field.

To successfully start a rotation, the staging label AWSPENDING must be in one of the following states:

  • Not be attached to any version at all

  • Attached to the same version as the staging label AWSCURRENT

If the staging label AWSPENDING is attached to a different version than the version with AWSCURRENT then the attempt to rotate fails.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:CancelRotateSecret

Related operations

  • To configure rotation for a secret or to manually trigger a rotation, use RotateSecret.

  • To get the rotation configuration details for a secret, use DescribeSecret.

  • To list all of the currently available secrets, use ListSecrets.

  • To list all of the versions currently associated with a secret, use ListSecretVersionIds.

", - "CreateSecret": "

Creates a new secret. A secret in Secrets Manager consists of both the protected secret data and the important information needed to manage the secret.

Secrets Manager stores the encrypted secret data in one of a collection of \"versions\" associated with the secret. Each version contains a copy of the encrypted secret data. Each version is associated with one or more \"staging labels\" that identify where the version is in the rotation cycle. The SecretVersionsToStages field of the secret contains the mapping of staging labels to the active versions of the secret. Versions without a staging label are considered deprecated and are not included in the list.

You provide the secret data to be encrypted by putting text in either the SecretString parameter or binary data in the SecretBinary parameter, but not both. If you include SecretString or SecretBinary then Secrets Manager also creates an initial secret version and automatically attaches the staging label AWSCURRENT to the new version.

  • If you call an operation that needs to encrypt or decrypt the SecretString or SecretBinary for a secret in the same account as the calling user and that secret doesn't specify a KMS encryption key, Secrets Manager uses the account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. If this key doesn't already exist in your account then Secrets Manager creates it for you automatically. All users in the same AWS account automatically have access to use the default CMK. Note that if an Secrets Manager API call results in AWS having to create the account's AWS-managed CMK, it can result in a one-time significant delay in returning the result.

  • If the secret is in a different AWS account from the credentials calling an API that requires encryption or decryption of the secret value then you must create and use a custom KMS CMK because you can't access the default CMK for the account using credentials from a different AWS account. Store the ARN of the CMK in the secret when you create the secret or when you update it by including it in the KMSKeyId. If you call an API that must encrypt or decrypt SecretString or SecretBinary using credentials from a different account then the KMS key policy must grant cross-account access to that other account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:CreateSecret

  • kms:GenerateDataKey - needed only if you use a customer-created KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

  • kms:Decrypt - needed only if you use a customer-created KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

Related operations

  • To delete a secret, use DeleteSecret.

  • To modify an existing secret, use UpdateSecret.

  • To create a new version of a secret, use PutSecretValue.

  • To retrieve the encrypted secure string and secure binary values, use GetSecretValue.

  • To retrieve all other details for a secret, use DescribeSecret. This does not include the encrypted secure string and secure binary values.

  • To retrieve the list of secret versions associated with the current secret, use DescribeSecret and examine the SecretVersionsToStages response value.

", + "CreateSecret": "

Creates a new secret. A secret in Secrets Manager consists of both the protected secret data and the important information needed to manage the secret.

Secrets Manager stores the encrypted secret data in one of a collection of \"versions\" associated with the secret. Each version contains a copy of the encrypted secret data. Each version is associated with one or more \"staging labels\" that identify where the version is in the rotation cycle. The SecretVersionsToStages field of the secret contains the mapping of staging labels to the active versions of the secret. Versions without a staging label are considered deprecated and are not included in the list.

You provide the secret data to be encrypted by putting text in either the SecretString parameter or binary data in the SecretBinary parameter, but not both. If you include SecretString or SecretBinary then Secrets Manager also creates an initial secret version and automatically attaches the staging label AWSCURRENT to the new version.

  • If you call an operation that needs to encrypt or decrypt the SecretString or SecretBinary for a secret in the same account as the calling user and that secret doesn't specify a AWS KMS encryption key, Secrets Manager uses the account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. If this key doesn't already exist in your account then Secrets Manager creates it for you automatically. All users in the same AWS account automatically have access to use the default CMK. Note that if an Secrets Manager API call results in AWS having to create the account's AWS-managed CMK, it can result in a one-time significant delay in returning the result.

  • If the secret is in a different AWS account from the credentials calling an API that requires encryption or decryption of the secret value then you must create and use a custom AWS KMS CMK because you can't access the default CMK for the account using credentials from a different AWS account. Store the ARN of the CMK in the secret when you create the secret or when you update it by including it in the KMSKeyId. If you call an API that must encrypt or decrypt SecretString or SecretBinary using credentials from a different account then the AWS KMS key policy must grant cross-account access to that other account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:CreateSecret

  • kms:GenerateDataKey - needed only if you use a customer-managed AWS KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

  • kms:Decrypt - needed only if you use a customer-managed AWS KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

Related operations

  • To delete a secret, use DeleteSecret.

  • To modify an existing secret, use UpdateSecret.

  • To create a new version of a secret, use PutSecretValue.

  • To retrieve the encrypted secure string and secure binary values, use GetSecretValue.

  • To retrieve all other details for a secret, use DescribeSecret. This does not include the encrypted secure string and secure binary values.

  • To retrieve the list of secret versions associated with the current secret, use DescribeSecret and examine the SecretVersionsToStages response value.

", + "DeleteResourcePolicy": "

Deletes the resource-based permission policy that's attached to the secret.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:DeleteResourcePolicy

Related operations

  • To attach a resource policy to a secret, use PutResourcePolicy.

  • To retrieve the current resource-based policy that's attached to a secret, use GetResourcePolicy.

  • To list all of the currently available secrets, use ListSecrets.

", "DeleteSecret": "

Deletes an entire secret and all of its versions. You can optionally include a recovery window during which you can restore the secret. If you don't specify a recovery window value, the operation defaults to 30 days. Secrets Manager attaches a DeletionDate stamp to the secret that specifies the end of the recovery window. At the end of the recovery window, Secrets Manager deletes the secret permanently.

At any time before recovery window ends, you can use RestoreSecret to remove the DeletionDate and cancel the deletion of the secret.

You cannot access the encrypted secret information in any secret that is scheduled for deletion. If you need to access that information, you must cancel the deletion with RestoreSecret and then retrieve the information.

  • There is no explicit operation to delete a version of a secret. Instead, remove all staging labels from the VersionStage field of a version. That marks the version as deprecated and allows Secrets Manager to delete it as needed. Versions that do not have any staging labels do not show up in ListSecretVersionIds unless you specify IncludeDeprecated.

  • The permanent secret deletion at the end of the waiting period is performed as a background task with low priority. There is no guarantee of a specific time after the recovery window for the actual delete operation to occur.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:DeleteSecret

Related operations

  • To create a secret, use CreateSecret.

  • To cancel deletion of a version of a secret before the recovery window has expired, use RestoreSecret.

", "DescribeSecret": "

Retrieves the details of a secret. It does not include the encrypted fields. Only those fields that are populated with a value are returned in the response.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:DescribeSecret

Related operations

", "GetRandomPassword": "

Generates a random password of the specified complexity. This operation is intended for use in the Lambda rotation function. Per best practice, we recommend that you specify the maximum length and include every character type that the system you are generating a password for can support.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:GetRandomPassword

", - "GetSecretValue": "

Retrieves the contents of the encrypted fields SecretString or SecretBinary from the specified version of a secret, whichever contains content.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:GetSecretValue

  • kms:Decrypt - required only if you use a customer-created KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

Related operations

  • To create a new version of the secret with different encrypted information, use PutSecretValue.

  • To retrieve the non-encrypted details for the secret, use DescribeSecret.

", + "GetResourcePolicy": "

Retrieves the JSON text of the resource-based policy document that's attached to the specified secret. The JSON request string input and response output are shown formatted with white space and line breaks for better readability. Submit your input as a single line JSON string.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:GetResourcePolicy

Related operations

", + "GetSecretValue": "

Retrieves the contents of the encrypted fields SecretString or SecretBinary from the specified version of a secret, whichever contains content.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:GetSecretValue

  • kms:Decrypt - required only if you use a customer-managed AWS KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

Related operations

  • To create a new version of the secret with different encrypted information, use PutSecretValue.

  • To retrieve the non-encrypted details for the secret, use DescribeSecret.

", "ListSecretVersionIds": "

Lists all of the versions attached to the specified secret. The output does not include the SecretString or SecretBinary fields. By default, the list includes only versions that have at least one staging label in VersionStage attached.

Always check the NextToken response parameter when calling any of the List* operations. These operations can occasionally return an empty or shorter than expected list of results even when there are more results available. When this happens, the NextToken response parameter contains a value to pass to the next call to the same API to request the next part of the list.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:ListSecretVersionIds

Related operations

", "ListSecrets": "

Lists all of the secrets that are stored by Secrets Manager in the AWS account. To list the versions currently stored for a specific secret, use ListSecretVersionIds. The encrypted fields SecretString and SecretBinary are not included in the output. To get that information, call the GetSecretValue operation.

Always check the NextToken response parameter when calling any of the List* operations. These operations can occasionally return an empty or shorter than expected list of results even when there are more results available. When this happens, the NextToken response parameter contains a value to pass to the next call to the same API to request the next part of the list.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:ListSecrets

Related operations

", - "PutSecretValue": "

Stores a new encrypted secret value in the specified secret. To do this, the operation creates a new version and attaches it to the secret. The version can contain a new SecretString value or a new SecretBinary value. You can also specify the staging labels that are initially attached to the new version.

The Secrets Manager console uses only the SecretString field. To add binary data to a secret with the SecretBinary field you must use the AWS CLI or one of the AWS SDKs.

  • If this operation creates the first version for the secret then Secrets Manager automatically attaches the staging label AWSCURRENT to the new version.

  • If another version of this secret already exists, then this operation does not automatically move any staging labels other than those that you explicitly specify in the VersionStages parameter.

  • If this operation moves the staging label AWSCURRENT from another version to this version (because you included it in the StagingLabels parameter) then Secrets Manager also automatically moves the staging label AWSPREVIOUS to the version that AWSCURRENT was removed from.

  • This operation is idempotent. If a version with a SecretVersionId with the same value as the ClientRequestToken parameter already exists and you specify the same secret data, the operation succeeds but does nothing. However, if the secret data is different, then the operation fails because you cannot modify an existing version; you can only create new ones.

  • If you call an operation that needs to encrypt or decrypt the SecretString or SecretBinary for a secret in the same account as the calling user and that secret doesn't specify a KMS encryption key, Secrets Manager uses the account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. If this key doesn't already exist in your account then Secrets Manager creates it for you automatically. All users in the same AWS account automatically have access to use the default CMK. Note that if an Secrets Manager API call results in AWS having to create the account's AWS-managed CMK, it can result in a one-time significant delay in returning the result.

  • If the secret is in a different AWS account from the credentials calling an API that requires encryption or decryption of the secret value then you must create and use a custom KMS CMK because you can't access the default CMK for the account using credentials from a different AWS account. Store the ARN of the CMK in the secret when you create the secret or when you update it by including it in the KMSKeyId. If you call an API that must encrypt or decrypt SecretString or SecretBinary using credentials from a different account then the KMS key policy must grant cross-account access to that other account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:PutSecretValue

  • kms:GenerateDataKey - needed only if you use a customer-created KMS key to encrypt the secret. You do not need this permission to use the account's AWS managed CMK for Secrets Manager.

  • kms:Encrypt - needed only if you use a customer-created KMS key to encrypt the secret. You do not need this permission to use the account's AWS managed CMK for Secrets Manager.

Related operations

", + "PutResourcePolicy": "

Attaches the contents of the specified resource-based permission policy to a secret. A resource-based policy is optional. Alternatively, you can use IAM identity-based policies that specify the secret's Amazon Resource Name (ARN) in the policy statement's Resources element. You can also use a combination of both identity-based and resource-based policies. The affected users and roles receive the permissions that are permitted by all of the relevant policies. For more information, see Using Resource-Based Policies for AWS Secrets Manager. For the complete description of the AWS policy syntax and grammar, see IAM JSON Policy Reference in the IAM User Guide.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:PutResourcePolicy

Related operations

  • To retrieve the resource policy that's attached to a secret, use GetResourcePolicy.

  • To delete the resource-based policy that's attached to a secret, use DeleteResourcePolicy.

  • To list all of the currently available secrets, use ListSecrets.

", + "PutSecretValue": "

Stores a new encrypted secret value in the specified secret. To do this, the operation creates a new version and attaches it to the secret. The version can contain a new SecretString value or a new SecretBinary value. You can also specify the staging labels that are initially attached to the new version.

The Secrets Manager console uses only the SecretString field. To add binary data to a secret with the SecretBinary field you must use the AWS CLI or one of the AWS SDKs.

  • If this operation creates the first version for the secret then Secrets Manager automatically attaches the staging label AWSCURRENT to the new version.

  • If another version of this secret already exists, then this operation does not automatically move any staging labels other than those that you explicitly specify in the VersionStages parameter.

  • If this operation moves the staging label AWSCURRENT from another version to this version (because you included it in the StagingLabels parameter) then Secrets Manager also automatically moves the staging label AWSPREVIOUS to the version that AWSCURRENT was removed from.

  • This operation is idempotent. If a version with a SecretVersionId with the same value as the ClientRequestToken parameter already exists and you specify the same secret data, the operation succeeds but does nothing. However, if the secret data is different, then the operation fails because you cannot modify an existing version; you can only create new ones.

  • If you call an operation that needs to encrypt or decrypt the SecretString or SecretBinary for a secret in the same account as the calling user and that secret doesn't specify a AWS KMS encryption key, Secrets Manager uses the account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. If this key doesn't already exist in your account then Secrets Manager creates it for you automatically. All users in the same AWS account automatically have access to use the default CMK. Note that if an Secrets Manager API call results in AWS having to create the account's AWS-managed CMK, it can result in a one-time significant delay in returning the result.

  • If the secret is in a different AWS account from the credentials calling an API that requires encryption or decryption of the secret value then you must create and use a custom AWS KMS CMK because you can't access the default CMK for the account using credentials from a different AWS account. Store the ARN of the CMK in the secret when you create the secret or when you update it by including it in the KMSKeyId. If you call an API that must encrypt or decrypt SecretString or SecretBinary using credentials from a different account then the AWS KMS key policy must grant cross-account access to that other account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:PutSecretValue

  • kms:GenerateDataKey - needed only if you use a customer-managed AWS KMS key to encrypt the secret. You do not need this permission to use the account's default AWS managed CMK for Secrets Manager.

Related operations

", "RestoreSecret": "

Cancels the scheduled deletion of a secret by removing the DeletedDate time stamp. This makes the secret accessible to query once again.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:RestoreSecret

Related operations

", "RotateSecret": "

Configures and starts the asynchronous process of rotating this secret. If you include the configuration parameters, the operation sets those values for the secret and then immediately starts a rotation. If you do not include the configuration parameters, the operation starts a rotation with the values already stored in the secret. After the rotation completes, the protected service and its clients all use the new version of the secret.

This required configuration information includes the ARN of an AWS Lambda function and the time between scheduled rotations. The Lambda rotation function creates a new version of the secret and creates or updates the credentials on the protected service to match. After testing the new credentials, the function marks the new secret with the staging label AWSCURRENT so that your clients all immediately begin to use the new version. For more information about rotating secrets and how to configure a Lambda function to rotate the secrets for your protected service, see Rotating Secrets in AWS Secrets Manager in the AWS Secrets Manager User Guide.

The rotation function must end with the versions of the secret in one of two states:

  • The AWSPENDING and AWSCURRENT staging labels are attached to the same version of the secret, or

  • The AWSPENDING staging label is not attached to any version of the secret.

If instead the AWSPENDING staging label is present but is not attached to the same version as AWSCURRENT then any later invocation of RotateSecret assumes that a previous rotation request is still in progress and returns an error.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:RotateSecret

  • lambda:InvokeFunction (on the function specified in the secret's metadata)

Related operations

", "TagResource": "

Attaches one or more tags, each consisting of a key name and a value, to the specified secret. Tags are part of the secret's overall metadata, and are not associated with any specific version of the secret. This operation only appends tags to the existing list of tags. To remove tags, you must use UntagResource.

The following basic restrictions apply to tags:

  • Maximum number of tags per secret—50

  • Maximum key length—127 Unicode characters in UTF-8

  • Maximum value length—255 Unicode characters in UTF-8

  • Tag keys and values are case sensitive.

  • Do not use the aws: prefix in your tag names or values because it is reserved for AWS use. You can't edit or delete tag names or values with this prefix. Tags with this prefix do not count against your tags per secret limit.

  • If your tagging schema will be used across multiple services and resources, remember that other services might have restrictions on allowed characters. Generally allowed characters are: letters, spaces, and numbers representable in UTF-8, plus the following special characters: + - = . _ : / @.

If you use tags as part of your security strategy, then adding or removing a tag can change permissions. If successfully completing this operation would result in you losing your permissions for this secret, then the operation is blocked and returns an Access Denied error.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:TagResource

Related operations

  • To remove one or more tags from the collection attached to a secret, use UntagResource.

  • To view the list of tags attached to a secret, use DescribeSecret.

", "UntagResource": "

Removes one or more tags from the specified secret.

This operation is idempotent. If a requested tag is not attached to the secret, no error is returned and the secret metadata is unchanged.

If you use tags as part of your security strategy, then removing a tag can change permissions. If successfully completing this operation would result in you losing your permissions for this secret, then the operation is blocked and returns an Access Denied error.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:UntagResource

Related operations

  • To add one or more tags to the collection attached to a secret, use TagResource.

  • To view the list of tags attached to a secret, use DescribeSecret.

", - "UpdateSecret": "

Modifies many of the details of a secret. If you include a ClientRequestToken and either SecretString or SecretBinary then it also creates a new version attached to the secret.

To modify the rotation configuration of a secret, use RotateSecret instead.

The Secrets Manager console uses only the SecretString parameter and therefore limits you to encrypting and storing only a text string. To encrypt and store binary data as part of the version of a secret, you must use either the AWS CLI or one of the AWS SDKs.

  • If a version with a SecretVersionId with the same value as the ClientRequestToken parameter already exists, the operation generates an error. You cannot modify an existing version, you can only create new ones.

  • If you include SecretString or SecretBinary to create a new secret version, Secrets Manager automatically attaches the staging label AWSCURRENT to the new version.

  • If you call an operation that needs to encrypt or decrypt the SecretString or SecretBinary for a secret in the same account as the calling user and that secret doesn't specify a KMS encryption key, Secrets Manager uses the account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. If this key doesn't already exist in your account then Secrets Manager creates it for you automatically. All users in the same AWS account automatically have access to use the default CMK. Note that if an Secrets Manager API call results in AWS having to create the account's AWS-managed CMK, it can result in a one-time significant delay in returning the result.

  • If the secret is in a different AWS account from the credentials calling an API that requires encryption or decryption of the secret value then you must create and use a custom KMS CMK because you can't access the default CMK for the account using credentials from a different AWS account. Store the ARN of the CMK in the secret when you create the secret or when you update it by including it in the KMSKeyId. If you call an API that must encrypt or decrypt SecretString or SecretBinary using credentials from a different account then the KMS key policy must grant cross-account access to that other account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:UpdateSecret

  • kms:GenerateDataKey - needed only if you use a custom KMS key to encrypt the secret. You do not need this permission to use the account's AWS managed CMK for Secrets Manager.

  • kms:Decrypt - needed only if you use a custom KMS key to encrypt the secret. You do not need this permission to use the account's AWS managed CMK for Secrets Manager.

Related operations

", - "UpdateSecretVersionStage": "

Modifies the staging labels attached to a version of a secret. Staging labels are used to track a version as it progresses through the secret rotation process. You can attach a staging label to only one version of a secret at a time. If a staging label to be added is already attached to another version, then it is moved--removed from the other version first and then attached to this one. For more information about staging labels, see Staging Labels in the AWS Secrets Manager User Guide.

The staging labels that you specify in the VersionStage parameter are added to the existing list of staging labels--they don't replace it.

You can move the AWSCURRENT staging label to this version by including it in this call.

Whenever you move AWSCURRENT, Secrets Manager automatically moves the label AWSPREVIOUS to the version that AWSCURRENT was removed from.

If this action results in the last label being removed from a version, then the version is considered to be 'deprecated' and can be deleted by Secrets Manager.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:UpdateSecretVersionStage

Related operations

  • To get the list of staging labels that are currently associated with a version of a secret, use DescribeSecret and examine the SecretVersionsToStages response value.

" + "UpdateSecret": "

Modifies many of the details of a secret. If you include a ClientRequestToken and either SecretString or SecretBinary then it also creates a new version attached to the secret.

To modify the rotation configuration of a secret, use RotateSecret instead.

The Secrets Manager console uses only the SecretString parameter and therefore limits you to encrypting and storing only a text string. To encrypt and store binary data as part of the version of a secret, you must use either the AWS CLI or one of the AWS SDKs.

  • If a version with a SecretVersionId with the same value as the ClientRequestToken parameter already exists, the operation generates an error. You cannot modify an existing version, you can only create new ones.

  • If you include SecretString or SecretBinary to create a new secret version, Secrets Manager automatically attaches the staging label AWSCURRENT to the new version.

  • If you call an operation that needs to encrypt or decrypt the SecretString or SecretBinary for a secret in the same account as the calling user and that secret doesn't specify a AWS KMS encryption key, Secrets Manager uses the account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. If this key doesn't already exist in your account then Secrets Manager creates it for you automatically. All users in the same AWS account automatically have access to use the default CMK. Note that if an Secrets Manager API call results in AWS having to create the account's AWS-managed CMK, it can result in a one-time significant delay in returning the result.

  • If the secret is in a different AWS account from the credentials calling an API that requires encryption or decryption of the secret value then you must create and use a custom AWS KMS CMK because you can't access the default CMK for the account using credentials from a different AWS account. Store the ARN of the CMK in the secret when you create the secret or when you update it by including it in the KMSKeyId. If you call an API that must encrypt or decrypt SecretString or SecretBinary using credentials from a different account then the AWS KMS key policy must grant cross-account access to that other account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:UpdateSecret

  • kms:GenerateDataKey - needed only if you use a custom AWS KMS key to encrypt the secret. You do not need this permission to use the account's AWS managed CMK for Secrets Manager.

  • kms:Decrypt - needed only if you use a custom AWS KMS key to encrypt the secret. You do not need this permission to use the account's AWS managed CMK for Secrets Manager.

Related operations

", + "UpdateSecretVersionStage": "

Modifies the staging labels attached to a version of a secret. Staging labels are used to track a version as it progresses through the secret rotation process. You can attach a staging label to only one version of a secret at a time. If a staging label to be added is already attached to another version, then it is moved--removed from the other version first and then attached to this one. For more information about staging labels, see Staging Labels in the AWS Secrets Manager User Guide.

The staging labels that you specify in the VersionStage parameter are added to the existing list of staging labels--they don't replace it.

You can move the AWSCURRENT staging label to this version by including it in this call.

Whenever you move AWSCURRENT, Secrets Manager automatically moves the label AWSPREVIOUS to the version that AWSCURRENT was removed from.

If this action results in the last label being removed from a version, then the version is considered to be 'deprecated' and can be deleted by Secrets Manager.

Minimum permissions

To run this command, you must have the following permissions:

  • secretsmanager:UpdateSecretVersionStage

Related operations

  • To get the list of staging labels that are currently associated with a version of a secret, use DescribeSecret and examine the SecretVersionsToStages response value.

" }, "shapes": { "AutomaticallyRotateAfterDaysType": { @@ -44,7 +47,7 @@ "ClientRequestTokenType": { "base": null, "refs": { - "CreateSecretRequest$ClientRequestToken": "

(Optional) If you include SecretString or SecretBinary, then an initial version is created as part of the secret, and this parameter specifies a unique identifier for the new version.

If you use the AWS CLI or one of the AWS SDK to call this operation, then you can leave this parameter empty. The CLI or SDK generates a random UUID for you and includes as the value for this parameter in the request. If you don't use the SDK and instead generate a raw HTTP request to the Secrets Manager service endpoint, then you must generate a ClientRequestToken yourself for the new version and include that value in the request.

This value helps ensure idempotency. Secrets Manager uses this value to prevent the accidental creation of duplicate versions if there are failures and retries during a rotation. We recommend that you generate a UUID-type value to ensure uniqueness of your versions within the specified secret.

  • If the ClientRequestToken value isn't already associated with a version of the secret then a new version of the secret is created.

  • If a version with this value already exists and that version's SecretString and SecretBinary values are the same as those in the request, then the request is ignored (the operation is idempotent).

  • If a version with this value already exists and that version's SecretString and SecretBinary values are different from those in the request then the request fails because you cannot modify an existing version. Instead, use PutSecretValue to create a new version.

This value becomes the SecretVersionId of the new version.

", + "CreateSecretRequest$ClientRequestToken": "

(Optional) If you include SecretString or SecretBinary, then an initial version is created as part of the secret, and this parameter specifies a unique identifier for the new version.

If you use the AWS CLI or one of the AWS SDK to call this operation, then you can leave this parameter empty. The CLI or SDK generates a random UUID for you and includes it as the value for this parameter in the request. If you don't use the SDK and instead generate a raw HTTP request to the Secrets Manager service endpoint, then you must generate a ClientRequestToken yourself for the new version and include that value in the request.

This value helps ensure idempotency. Secrets Manager uses this value to prevent the accidental creation of duplicate versions if there are failures and retries during a rotation. We recommend that you generate a UUID-type value to ensure uniqueness of your versions within the specified secret.

  • If the ClientRequestToken value isn't already associated with a version of the secret then a new version of the secret is created.

  • If a version with this value already exists and that version's SecretString and SecretBinary values are the same as those in the request, then the request is ignored (the operation is idempotent).

  • If a version with this value already exists and that version's SecretString and SecretBinary values are different from those in the request then the request fails because you cannot modify an existing version. Instead, use PutSecretValue to create a new version.

This value becomes the SecretVersionId of the new version.

", "PutSecretValueRequest$ClientRequestToken": "

(Optional) Specifies a unique identifier for the new version of the secret.

If you use the AWS CLI or one of the AWS SDK to call this operation, then you can leave this parameter empty. The CLI or SDK generates a random UUID for you and includes that in the request. If you don't use the SDK and instead generate a raw HTTP request to the Secrets Manager service endpoint, then you must generate a ClientRequestToken yourself for new versions and include that value in the request.

This value helps ensure idempotency. Secrets Manager uses this value to prevent the accidental creation of duplicate versions if there are failures and retries during the Lambda rotation function's processing. We recommend that you generate a UUID-type value to ensure uniqueness within the specified secret.

  • If the ClientRequestToken value isn't already associated with a version of the secret then a new version of the secret is created.

  • If a version with this value already exists and that version's SecretString or SecretBinary values are the same as those in the request then the request is ignored (the operation is idempotent).

  • If a version with this value already exists and that version's SecretString and SecretBinary values are different from those in the request then the request fails because you cannot modify an existing secret version. You can only create new versions to store new secret values.

This value becomes the SecretVersionId of the new version.

", "RotateSecretRequest$ClientRequestToken": "

(Optional) Specifies a unique identifier for the new version of the secret that helps ensure idempotency.

If you use the AWS CLI or one of the AWS SDK to call this operation, then you can leave this parameter empty. The CLI or SDK generates a random UUID for you and includes that in the request for this parameter. If you don't use the SDK and instead generate a raw HTTP request to the Secrets Manager service endpoint, then you must generate a ClientRequestToken yourself for new versions and include that value in the request.

You only need to specify your own value if you are implementing your own retry logic and want to ensure that a given secret is not created twice. We recommend that you generate a UUID-type value to ensure uniqueness within the specified secret.

Secrets Manager uses this value to prevent the accidental creation of duplicate versions if there are failures and retries during the function's processing.

  • If the ClientRequestToken value isn't already associated with a version of the secret then a new version of the secret is created.

  • If a version with this value already exists and that version's SecretString and SecretBinary values are the same as the request, then the request is ignored (the operation is idempotent).

  • If a version with this value already exists and that version's SecretString and SecretBinary values are different from the request then an error occurs because you cannot modify an existing secret value.

This value becomes the SecretVersionId of the new version.

", "UpdateSecretRequest$ClientRequestToken": "

(Optional) If you want to add a new version to the secret, this parameter specifies a unique identifier for the new version that helps ensure idempotency.

If you use the AWS CLI or one of the AWS SDK to call this operation, then you can leave this parameter empty. The CLI or SDK generates a random UUID for you and includes that in the request. If you don't use the SDK and instead generate a raw HTTP request to the Secrets Manager service endpoint, then you must generate a ClientRequestToken yourself for new versions and include that value in the request.

You typically only need to interact with this value if you implement your own retry logic and want to ensure that a given secret is not created twice. We recommend that you generate a UUID-type value to ensure uniqueness within the specified secret.

Secrets Manager uses this value to prevent the accidental creation of duplicate versions if there are failures and retries during the Lambda rotation function's processing.

  • If the ClientRequestToken value isn't already associated with a version of the secret then a new version of the secret is created.

  • If a version with this value already exists and that version's SecretString and SecretBinary values are the same as those in the request then the request is ignored (the operation is idempotent).

  • If a version with this value already exists and that version's SecretString and SecretBinary values are different from the request then an error occurs because you cannot modify an existing secret value.

This value becomes the SecretVersionId of the new version.

" @@ -72,6 +75,16 @@ "refs": { } }, + "DeleteResourcePolicyRequest": { + "base": null, + "refs": { + } + }, + "DeleteResourcePolicyResponse": { + "base": null, + "refs": { + } + }, "DeleteSecretRequest": { "base": null, "refs": { @@ -130,6 +143,7 @@ "InvalidRequestException$Message": null, "LimitExceededException$Message": null, "MalformedPolicyDocumentException$Message": null, + "PreconditionNotMetException$Message": null, "ResourceExistsException$Message": null, "ResourceNotFoundException$Message": null } @@ -174,6 +188,16 @@ "refs": { } }, + "GetResourcePolicyRequest": { + "base": null, + "refs": { + } + }, + "GetResourcePolicyResponse": { + "base": null, + "refs": { + } + }, "GetSecretValueRequest": { "base": null, "refs": { @@ -206,17 +230,17 @@ } }, "InvalidRequestException": { - "base": "

You provided a parameter value that is not valid for the current state of the resource. For example, if you try to enable rotation on a secret, you must already have a Lambda function ARN configured or included as a parameter in this call.

", + "base": "

You provided a parameter value that is not valid for the current state of the resource.

Possible causes:

  • You tried to perform the operation on a secret that's currently marked deleted.

  • You tried to enable rotation on a secret that doesn't already have a Lambda function ARN configured and you didn't include such an ARN as a parameter in this call.

", "refs": { } }, "KmsKeyIdType": { "base": null, "refs": { - "CreateSecretRequest$KmsKeyId": "

(Optional) Specifies the ARN or alias of the AWS KMS customer master key (CMK) to be used to encrypt the SecretString or SecretBinary values in the versions stored in this secret.

If you don't specify this value, then Secrets Manager defaults to using the AWS account's default CMK (the one named aws/secretsmanager). If a KMS CMK with that name doesn't yet exist, then Secrets Manager creates it for you automatically the first time it needs to encrypt a version's SecretString or SecretBinary fields.

You can use the account's default CMK to encrypt and decrypt only if you call this operation using credentials from the same account that owns the secret. If the secret is in a different account, then you must create a custom CMK and specify the ARN in this field.

", - "DescribeSecretResponse$KmsKeyId": "

The ARN or alias of the AWS KMS customer master key (CMK) that's used to encrypt the SecretString or SecretBinary fields in each version of the secret. If you don't provide a key, then Secrets Manager defaults to encrypting the secret fields with the default KMS CMK (the one named awssecretsmanager) for this account.

", + "CreateSecretRequest$KmsKeyId": "

(Optional) Specifies the ARN, Key ID, or alias of the AWS KMS customer master key (CMK) to be used to encrypt the SecretString or SecretBinary values in the versions stored in this secret.

You can specify any of the supported ways to identify a AWS KMS key ID. If you need to reference a CMK in a different account, you can use only the key ARN or the alias ARN.

If you don't specify this value, then Secrets Manager defaults to using the AWS account's default CMK (the one named aws/secretsmanager). If a AWS KMS CMK with that name doesn't yet exist, then Secrets Manager creates it for you automatically the first time it needs to encrypt a version's SecretString or SecretBinary fields.

You can use the account's default CMK to encrypt and decrypt only if you call this operation using credentials from the same account that owns the secret. If the secret is in a different account, then you must create a custom CMK and specify the ARN in this field.

", + "DescribeSecretResponse$KmsKeyId": "

The ARN or alias of the AWS KMS customer master key (CMK) that's used to encrypt the SecretString or SecretBinary fields in each version of the secret. If you don't provide a key, then Secrets Manager defaults to encrypting the secret fields with the default AWS KMS CMK (the one named awssecretsmanager) for this account.

", "SecretListEntry$KmsKeyId": "

The ARN or alias of the AWS KMS customer master key (CMK) that's used to encrypt the SecretString and SecretBinary fields in each version of the secret. If you don't provide a key, then Secrets Manager defaults to encrypting the secret fields with the default KMS CMK (the one named awssecretsmanager) for this account.

", - "UpdateSecretRequest$KmsKeyId": "

(Optional) Specifies the ARN or alias of the KMS customer master key (CMK) to be used to encrypt the protected text in the versions of this secret.

If you don't specify this value, then Secrets Manager defaults to using the default CMK in the account (the one named aws/secretsmanager). If a KMS CMK with that name doesn't exist, then Secrets Manager creates it for you automatically the first time it needs to encrypt a version's Plaintext or PlaintextString fields.

You can only use the account's default CMK to encrypt and decrypt if you call this operation using credentials from the same account that owns the secret. If the secret is in a different account, then you must create a custom CMK and provide the ARN in this field.

" + "UpdateSecretRequest$KmsKeyId": "

(Optional) Specifies the ARN or alias of the AWS KMS customer master key (CMK) to be used to encrypt the protected text in the versions of this secret.

If you don't specify this value, then Secrets Manager defaults to using the default CMK in the account (the one named aws/secretsmanager). If a AWS KMS CMK with that name doesn't exist, then Secrets Manager creates it for you automatically the first time it needs to encrypt a version's Plaintext or PlaintextString fields.

You can only use the account's default CMK to encrypt and decrypt if you call this operation using credentials from the same account that owns the secret. If the secret is in a different account, then you must create a custom CMK and provide the ARN in this field.

" } }, "LastAccessedDateType": { @@ -281,7 +305,10 @@ "NameType": { "base": null, "refs": { - "CreateSecretRequest$Name": "

Specifies the friendly name of the new secret.

" + "CreateSecretRequest$Name": "

Specifies the friendly name of the new secret.

The secret name must be ASCII letters, digits, or the following characters : /_+=.@-

", + "DeleteResourcePolicyResponse$Name": "

The friendly name of the secret that the resource-based policy was deleted for.

", + "GetResourcePolicyResponse$Name": "

The friendly name of the secret that the resource-based policy was retrieved for.

", + "PutResourcePolicyResponse$Name": "

The friendly name of the secret that the resource-based policy was retrieved for.

" } }, "NextTokenType": { @@ -293,12 +320,34 @@ "ListSecretsResponse$NextToken": "

If present in the response, this value indicates that there's more output available than what's included in the current response. This can occur even when the response includes no values at all, such as when you ask for a filtered view of a very long list. Use this value in the NextToken request parameter in a subsequent call to the operation to continue processing and get the next part of the output. You should repeat this until the NextToken response element comes back empty (as null).

" } }, + "NonEmptyResourcePolicyType": { + "base": null, + "refs": { + "GetResourcePolicyResponse$ResourcePolicy": "

A JSON-formatted string that describes the permissions that are associated with the attached secret. These permissions are combined with any permissions that are associated with the user or role that attempts to access this secret. The combined permissions specify who can access the secret and what actions they can perform. For more information, see Authentication and Access Control for AWS Secrets Manager in the AWS Secrets Manager User Guide.

", + "PutResourcePolicyRequest$ResourcePolicy": "

A JSON-formatted string that's constructed according to the grammar and syntax for an AWS resource-based policy. The policy in the string identifies who can access or manage this secret and its versions. For information on how to format a JSON parameter for the various command line tool environments, see Using JSON for Parameters in the AWS CLI User Guide.

" + } + }, "PasswordLengthType": { "base": null, "refs": { "GetRandomPasswordRequest$PasswordLength": "

The desired length of the generated password. The default value if you do not include this parameter is 32 characters.

" } }, + "PreconditionNotMetException": { + "base": "

The request failed because you did not complete all the prerequisite steps.

", + "refs": { + } + }, + "PutResourcePolicyRequest": { + "base": null, + "refs": { + } + }, + "PutResourcePolicyResponse": { + "base": null, + "refs": { + } + }, "PutSecretValueRequest": { "base": null, "refs": { @@ -385,14 +434,17 @@ "refs": { "CancelRotateSecretResponse$ARN": "

The ARN of the secret for which rotation was canceled.

", "CreateSecretResponse$ARN": "

The Amazon Resource Name (ARN) of the secret that you just created.

Secrets Manager automatically adds several random characters to the name at the end of the ARN when you initially create a secret. This affects only the ARN and not the actual friendly name. This ensures that if you create a new secret with the same name as an old secret that you previously deleted, then users with access to the old secret don't automatically get access to the new secret because the ARNs are different.

", + "DeleteResourcePolicyResponse$ARN": "

The ARN of the secret that the resource-based policy was deleted for.

", "DeleteSecretResponse$ARN": "

The ARN of the secret that is now scheduled for deletion.

", "DescribeSecretResponse$ARN": "

The ARN of the secret.

", + "GetResourcePolicyResponse$ARN": "

The ARN of the secret that the resource-based policy was retrieved for.

", "GetSecretValueResponse$ARN": "

The ARN of the secret.

", "ListSecretVersionIdsResponse$ARN": "

The Amazon Resource Name (ARN) for the secret.

Secrets Manager automatically adds several random characters to the name at the end of the ARN when you initially create a secret. This affects only the ARN and not the actual friendly name. This ensures that if you create a new secret with the same name as an old secret that you previously deleted, then users with access to the old secret don't automatically get access to the new secret because the ARNs are different.

", + "PutResourcePolicyResponse$ARN": "

The ARN of the secret that the resource-based policy was retrieved for.

", "PutSecretValueResponse$ARN": "

The Amazon Resource Name (ARN) for the secret for which you just created a version.

", "RestoreSecretResponse$ARN": "

The ARN of the secret that was restored.

", "RotateSecretResponse$ARN": "

The ARN of the secret.

", - "SecretListEntry$ARN": "

The Amazon Resource Name (ARN) of the secret.

For more information about ARNs in Secrets Manager, see Policy Resources in the AWS Secrets Manager User Guide.

", + "SecretListEntry$ARN": "

The Amazon Resource Name (ARN) of the secret.

For more information about ARNs in Secrets Manager, see Policy Resources in the AWS Secrets Manager User Guide.

", "UpdateSecretResponse$ARN": "

The ARN of this secret.

Secrets Manager automatically adds several random characters to the name at the end of the ARN when you initially create a secret. This affects only the ARN and not the actual friendly name. This ensures that if you create a new secret with the same name as an old secret that you previously deleted, then users with access to the old secret don't automatically get access to the new secret because the ARNs are different.

", "UpdateSecretVersionStageResponse$ARN": "

The ARN of the secret with the staging labels that were modified.

" } @@ -410,10 +462,13 @@ "base": null, "refs": { "CancelRotateSecretRequest$SecretId": "

Specifies the secret for which you want to cancel a rotation request. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", + "DeleteResourcePolicyRequest$SecretId": "

Specifies the secret that you want to delete the attached resource-based policy for. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", "DeleteSecretRequest$SecretId": "

Specifies the secret that you want to delete. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", "DescribeSecretRequest$SecretId": "

The identifier of the secret whose details you want to retrieve. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", + "GetResourcePolicyRequest$SecretId": "

Specifies the secret that you want to retrieve the attached resource-based policy for. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", "GetSecretValueRequest$SecretId": "

Specifies the secret containing the version that you want to retrieve. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", "ListSecretVersionIdsRequest$SecretId": "

The identifier for the secret containing the versions you want to list. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", + "PutResourcePolicyRequest$SecretId": "

Specifies the secret that you want to attach the resource-based policy to. You can specify either the ARN or the friendly name of the secret.

", "PutSecretValueRequest$SecretId": "

Specifies the secret to which you want to add a new version. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret. The secret must already exist.

", "RestoreSecretRequest$SecretId": "

Specifies the secret that you want to restore from a previously scheduled deletion. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", "RotateSecretRequest$SecretId": "

Specifies the secret that you want to rotate. You can specify either the Amazon Resource Name (ARN) or the friendly name of the secret.

", diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/examples-1.json index 6bae5627f..8433121eb 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/examples-1.json @@ -45,6 +45,26 @@ "title": "To create a basic secret" } ], + "DeleteResourcePolicy": [ + { + "input": { + "SecretId": "MyTestDatabaseSecret" + }, + "output": { + "ARN": "arn:aws:secretsmanager:us-west-2:123456789012:secret:MyTestDatabaseMasterSecret-a1b2c3", + "Name": "MyTestDatabaseSecret" + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "The following example shows how to delete the resource-based policy that is attached to a secret.", + "id": "to-delete-the-resource-based-policy-attached-to-a-secret-1530209419204", + "title": "To delete the resource-based policy attached to a secret" + } + ], "DeleteSecret": [ { "input": { @@ -136,6 +156,27 @@ "title": "To generate a random password" } ], + "GetResourcePolicy": [ + { + "input": { + "SecretId": "MyTestDatabaseSecret" + }, + "output": { + "ARN": "arn:aws:secretsmanager:us-west-2:123456789012:secret:MyTestDatabaseSecret-a1b2c3", + "Name": "MyTestDatabaseSecret", + "ResourcePolicy": "{\n\"Version\":\"2012-10-17\",\n\"Statement\":[{\n\"Effect\":\"Allow\",\n\"Principal\":{\n\"AWS\":\"arn:aws:iam::123456789012:root\"\n},\n\"Action\":\"secretsmanager:GetSecretValue\",\n\"Resource\":\"*\"\n}]\n}" + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "The following example shows how to retrieve the resource-based policy that is attached to a secret.", + "id": "to-retrieve-the-resource-based-policy-attached-to-a-secret-1530209677536", + "title": "To retrieve the resource-based policy attached to a secret" + } + ], "GetSecretValue": [ { "input": { @@ -245,6 +286,27 @@ "title": "To list the secrets in your account" } ], + "PutResourcePolicy": [ + { + "input": { + "ResourcePolicy": "{\n\"Version\":\"2012-10-17\",\n\"Statement\":[{\n\"Effect\":\"Allow\",\n\"Principal\":{\n\"AWS\":\"arn:aws:iam::123456789012:root\"\n},\n\"Action\":\"secretsmanager:GetSecretValue\",\n\"Resource\":\"*\"\n}]\n}", + "SecretId": "MyTestDatabaseSecret" + }, + "output": { + "ARN": "arn:aws:secretsmanager:us-west-2:123456789012:secret:MyTestDatabaseSecret-a1b2c3", + "Name": "MyTestDatabaseSecret" + }, + "comments": { + "input": { + }, + "output": { + } + }, + "description": "The following example shows how to add a resource-based policy to a secret.", + "id": "to-add-a-resource-based-policy-to-a-secret-1530209881839", + "title": "To add a resource-based policy to a secret" + } + ], "PutSecretValue": [ { "input": { diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/smoke.json b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/smoke.json similarity index 63% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/smoke.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/smoke.json index 068b23492..51e9dc240 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2014-09-01/smoke.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/secretsmanager/2017-10-17/smoke.json @@ -3,14 +3,14 @@ "defaultRegion": "us-west-2", "testCases": [ { - "operationName": "DescribeDBEngineVersions", + "operationName": "ListSecrets", "input": {}, "errorExpectedFromService": false }, { - "operationName": "DescribeDBInstances", + "operationName": "DescribeSecret", "input": { - "DBInstanceIdentifier": "fake-id" + "SecretId": "fake-secret-id" }, "errorExpectedFromService": true } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/servicecatalog/2015-12-10/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/servicecatalog/2015-12-10/api-2.json index 24cd74110..425faeb6d 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/servicecatalog/2015-12-10/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/servicecatalog/2015-12-10/api-2.json @@ -1955,7 +1955,8 @@ "enum":[ "CLOUD_FORMATION_TEMPLATE", "MARKETPLACE" - ] + ], + "max":8191 }, "ProductViewAggregationType":{"type":"string"}, "ProductViewAggregationValue":{ @@ -2007,9 +2008,18 @@ "key":{"shape":"ProductViewFilterBy"}, "value":{"shape":"ProductViewFilterValues"} }, - "ProductViewName":{"type":"string"}, - "ProductViewOwner":{"type":"string"}, - "ProductViewShortDescription":{"type":"string"}, + "ProductViewName":{ + "type":"string", + "max":8191 + }, + "ProductViewOwner":{ + "type":"string", + "max":8191 + }, + "ProductViewShortDescription":{ + "type":"string", + "max":8191 + }, "ProductViewSortBy":{ "type":"string", "enum":[ @@ -2619,9 +2629,18 @@ "type":"string", "pattern":"[\\u0009\\u000a\\u000d\\u0020-\\uD7FF\\uE000-\\uFFFD]*" }, - "SupportDescription":{"type":"string"}, - "SupportEmail":{"type":"string"}, - "SupportUrl":{"type":"string"}, + "SupportDescription":{ + "type":"string", + "max":8191 + }, + "SupportEmail":{ + "type":"string", + "max":254 + }, + "SupportUrl":{ + "type":"string", + "max":2083 + }, "Tag":{ "type":"structure", "required":[ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/api-2.json index d19f6ad50..2696b1d99 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/api-2.json @@ -12,6 +12,42 @@ "uid":"shield-2016-06-02" }, "operations":{ + "AssociateDRTLogBucket":{ + "name":"AssociateDRTLogBucket", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"AssociateDRTLogBucketRequest"}, + "output":{"shape":"AssociateDRTLogBucketResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"InvalidOperationException"}, + {"shape":"NoAssociatedRoleException"}, + {"shape":"LimitsExceededException"}, + {"shape":"InvalidParameterException"}, + {"shape":"AccessDeniedForDependencyException"}, + {"shape":"OptimisticLockException"}, + {"shape":"ResourceNotFoundException"} + ] + }, + "AssociateDRTRole":{ + "name":"AssociateDRTRole", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"AssociateDRTRoleRequest"}, + "output":{"shape":"AssociateDRTRoleResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"InvalidOperationException"}, + {"shape":"InvalidParameterException"}, + {"shape":"AccessDeniedForDependencyException"}, + {"shape":"OptimisticLockException"}, + {"shape":"ResourceNotFoundException"} + ] + }, "CreateProtection":{ "name":"CreateProtection", "http":{ @@ -69,7 +105,8 @@ {"shape":"InternalErrorException"}, {"shape":"LockedSubscriptionException"}, {"shape":"ResourceNotFoundException"} - ] + ], + "deprecated":true }, "DescribeAttack":{ "name":"DescribeAttack", @@ -84,6 +121,32 @@ {"shape":"InvalidParameterException"} ] }, + "DescribeDRTAccess":{ + "name":"DescribeDRTAccess", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeDRTAccessRequest"}, + "output":{"shape":"DescribeDRTAccessResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"ResourceNotFoundException"} + ] + }, + "DescribeEmergencyContactSettings":{ + "name":"DescribeEmergencyContactSettings", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeEmergencyContactSettingsRequest"}, + "output":{"shape":"DescribeEmergencyContactSettingsResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"ResourceNotFoundException"} + ] + }, "DescribeProtection":{ "name":"DescribeProtection", "http":{ @@ -110,6 +173,38 @@ {"shape":"ResourceNotFoundException"} ] }, + "DisassociateDRTLogBucket":{ + "name":"DisassociateDRTLogBucket", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DisassociateDRTLogBucketRequest"}, + "output":{"shape":"DisassociateDRTLogBucketResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"InvalidOperationException"}, + {"shape":"NoAssociatedRoleException"}, + {"shape":"AccessDeniedForDependencyException"}, + {"shape":"OptimisticLockException"}, + {"shape":"ResourceNotFoundException"} + ] + }, + "DisassociateDRTRole":{ + "name":"DisassociateDRTRole", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DisassociateDRTRoleRequest"}, + "output":{"shape":"DisassociateDRTRoleResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"InvalidOperationException"}, + {"shape":"OptimisticLockException"}, + {"shape":"ResourceNotFoundException"} + ] + }, "GetSubscriptionState":{ "name":"GetSubscriptionState", "http":{ @@ -146,11 +241,74 @@ "output":{"shape":"ListProtectionsResponse"}, "errors":[ {"shape":"InternalErrorException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidPaginationTokenException"} + ] + }, + "UpdateEmergencyContactSettings":{ + "name":"UpdateEmergencyContactSettings", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"UpdateEmergencyContactSettingsRequest"}, + "output":{"shape":"UpdateEmergencyContactSettingsResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"InvalidParameterException"}, + {"shape":"OptimisticLockException"}, {"shape":"ResourceNotFoundException"} ] + }, + "UpdateSubscription":{ + "name":"UpdateSubscription", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"UpdateSubscriptionRequest"}, + "output":{"shape":"UpdateSubscriptionResponse"}, + "errors":[ + {"shape":"InternalErrorException"}, + {"shape":"LockedSubscriptionException"}, + {"shape":"ResourceNotFoundException"}, + {"shape":"InvalidParameterException"}, + {"shape":"OptimisticLockException"} + ] } }, "shapes":{ + "AccessDeniedForDependencyException":{ + "type":"structure", + "members":{ + "message":{"shape":"errorMessage"} + }, + "exception":true + }, + "AssociateDRTLogBucketRequest":{ + "type":"structure", + "required":["LogBucket"], + "members":{ + "LogBucket":{"shape":"LogBucket"} + } + }, + "AssociateDRTLogBucketResponse":{ + "type":"structure", + "members":{ + } + }, + "AssociateDRTRoleRequest":{ + "type":"structure", + "required":["RoleArn"], + "members":{ + "RoleArn":{"shape":"RoleArn"} + } + }, + "AssociateDRTRoleResponse":{ + "type":"structure", + "members":{ + } + }, "AttackDetail":{ "type":"structure", "members":{ @@ -228,6 +386,13 @@ "type":"list", "member":{"shape":"AttackVectorDescription"} }, + "AutoRenew":{ + "type":"string", + "enum":[ + "ENABLED", + "DISABLED" + ] + }, "Contributor":{ "type":"structure", "members":{ @@ -277,12 +442,14 @@ "DeleteSubscriptionRequest":{ "type":"structure", "members":{ - } + }, + "deprecated":true }, "DeleteSubscriptionResponse":{ "type":"structure", "members":{ - } + }, + "deprecated":true }, "DescribeAttackRequest":{ "type":"structure", @@ -297,6 +464,29 @@ "Attack":{"shape":"AttackDetail"} } }, + "DescribeDRTAccessRequest":{ + "type":"structure", + "members":{ + } + }, + "DescribeDRTAccessResponse":{ + "type":"structure", + "members":{ + "RoleArn":{"shape":"RoleArn"}, + "LogBucketList":{"shape":"LogBucketList"} + } + }, + "DescribeEmergencyContactSettingsRequest":{ + "type":"structure", + "members":{ + } + }, + "DescribeEmergencyContactSettingsResponse":{ + "type":"structure", + "members":{ + "EmergencyContactList":{"shape":"EmergencyContactList"} + } + }, "DescribeProtectionRequest":{ "type":"structure", "required":["ProtectionId"], @@ -321,11 +511,50 @@ "Subscription":{"shape":"Subscription"} } }, + "DisassociateDRTLogBucketRequest":{ + "type":"structure", + "required":["LogBucket"], + "members":{ + "LogBucket":{"shape":"LogBucket"} + } + }, + "DisassociateDRTLogBucketResponse":{ + "type":"structure", + "members":{ + } + }, + "DisassociateDRTRoleRequest":{ + "type":"structure", + "members":{ + } + }, + "DisassociateDRTRoleResponse":{ + "type":"structure", + "members":{ + } + }, "Double":{"type":"double"}, "DurationInSeconds":{ "type":"long", "min":0 }, + "EmailAddress":{ + "type":"string", + "pattern":"^\\S+@\\S+\\.\\S+$" + }, + "EmergencyContact":{ + "type":"structure", + "required":["EmailAddress"], + "members":{ + "EmailAddress":{"shape":"EmailAddress"} + } + }, + "EmergencyContactList":{ + "type":"list", + "member":{"shape":"EmergencyContact"}, + "max":10, + "min":0 + }, "GetSubscriptionStateRequest":{ "type":"structure", "members":{ @@ -354,6 +583,13 @@ }, "exception":true }, + "InvalidPaginationTokenException":{ + "type":"structure", + "members":{ + "message":{"shape":"errorMessage"} + }, + "exception":true + }, "InvalidParameterException":{ "type":"structure", "members":{ @@ -368,8 +604,19 @@ }, "exception":true }, + "Limit":{ + "type":"structure", + "members":{ + "Type":{"shape":"String"}, + "Max":{"shape":"Long"} + } + }, "LimitNumber":{"type":"long"}, "LimitType":{"type":"string"}, + "Limits":{ + "type":"list", + "member":{"shape":"Limit"} + }, "LimitsExceededException":{ "type":"structure", "members":{ @@ -417,6 +664,18 @@ }, "exception":true }, + "LogBucket":{ + "type":"string", + "max":63, + "min":3, + "pattern":"^([a-z]|(\\d(?!\\d{0,2}\\.\\d{1,3}\\.\\d{1,3}\\.\\d{1,3})))([a-z\\d]|(\\.(?!(\\.|-)))|(-(?!\\.))){1,61}[a-z\\d]$" + }, + "LogBucketList":{ + "type":"list", + "member":{"shape":"LogBucket"}, + "max":10, + "min":0 + }, "Long":{"type":"long"}, "MaxResults":{ "type":"integer", @@ -434,6 +693,13 @@ "type":"list", "member":{"shape":"Mitigation"} }, + "NoAssociatedRoleException":{ + "type":"structure", + "members":{ + "message":{"shape":"errorMessage"} + }, + "exception":true + }, "OptimisticLockException":{ "type":"structure", "members":{ @@ -487,6 +753,11 @@ }, "exception":true }, + "RoleArn":{ + "type":"string", + "max":96, + "pattern":"^arn:aws:iam::\\d{12}:role/?[a-zA-Z_0-9+=,.@\\-_/]+" + }, "String":{"type":"string"}, "SubResourceSummary":{ "type":"structure", @@ -512,7 +783,10 @@ "type":"structure", "members":{ "StartTime":{"shape":"Timestamp"}, - "TimeCommitmentInSeconds":{"shape":"DurationInSeconds"} + "EndTime":{"shape":"Timestamp"}, + "TimeCommitmentInSeconds":{"shape":"DurationInSeconds"}, + "AutoRenew":{"shape":"AutoRenew"}, + "Limits":{"shape":"Limits"} } }, "SubscriptionState":{ @@ -574,6 +848,28 @@ "REQUESTS" ] }, + "UpdateEmergencyContactSettingsRequest":{ + "type":"structure", + "members":{ + "EmergencyContactList":{"shape":"EmergencyContactList"} + } + }, + "UpdateEmergencyContactSettingsResponse":{ + "type":"structure", + "members":{ + } + }, + "UpdateSubscriptionRequest":{ + "type":"structure", + "members":{ + "AutoRenew":{"shape":"AutoRenew"} + } + }, + "UpdateSubscriptionResponse":{ + "type":"structure", + "members":{ + } + }, "errorMessage":{"type":"string"} } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/docs-2.json index b32a177f8..56d56a1ec 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/shield/2016-06-02/docs-2.json @@ -2,18 +2,51 @@ "version": "2.0", "service": "AWS Shield Advanced

This is the AWS Shield Advanced API Reference. This guide is for developers who need detailed information about the AWS Shield Advanced API actions, data types, and errors. For detailed information about AWS WAF and AWS Shield Advanced features and an overview of how to use the AWS WAF and AWS Shield Advanced APIs, see the AWS WAF and AWS Shield Developer Guide.

", "operations": { - "CreateProtection": "

Enables AWS Shield Advanced for a specific AWS resource. The resource can be an Amazon CloudFront distribution, Elastic Load Balancing load balancer, Elastic IP Address, or an Amazon Route 53 hosted zone.

", - "CreateSubscription": "

Activates AWS Shield Advanced for an account.

", + "AssociateDRTLogBucket": "

Authorizes the DDoS Response team (DRT) to access the specified Amazon S3 bucket containing your flow logs. You can associate up to 10 Amazon S3 buckets with your subscription.

To use the services of the DRT and make an AssociateDRTLogBucket request, you must be subscribed to the Business Support plan or the Enterprise Support plan.

", + "AssociateDRTRole": "

Authorizes the DDoS Response team (DRT), using the specified role, to access your AWS account to assist with DDoS attack mitigation during potential attacks. This enables the DRT to inspect your AWS WAF configuration and create or update AWS WAF rules and web ACLs.

You can associate only one RoleArn with your subscription. If you submit an AssociateDRTRole request for an account that already has an associated role, the new RoleArn will replace the existing RoleArn.

Prior to making the AssociateDRTRole request, you must attach the AWSShieldDRTAccessPolicy managed policy to the role you will specify in the request. For more information see Attaching and Detaching IAM Policies. The role must also trust the service principal drt.shield.amazonaws.com. For more information, see IAM JSON Policy Elements: Principal.

The DRT will have access only to your AWS WAF and Shield resources. By submitting this request, you authorize the DRT to inspect your AWS WAF and Shield configuration and create and update AWS WAF rules and web ACLs on your behalf. The DRT takes these actions only if explicitly authorized by you.

You must have the iam:PassRole permission to make an AssociateDRTRole request. For more information, see Granting a User Permissions to Pass a Role to an AWS Service.

To use the services of the DRT and make an AssociateDRTRole request, you must be subscribed to the Business Support plan or the Enterprise Support plan.

", + "CreateProtection": "

Enables AWS Shield Advanced for a specific AWS resource. The resource can be an Amazon CloudFront distribution, Elastic Load Balancing load balancer, Elastic IP Address, or an Amazon Route 53 hosted zone.

You can add protection to only a single resource with each CreateProtection request. If you want to add protection to multiple resources at once, use the AWS WAF console. For more information see Getting Started with AWS Shield Advanced and Add AWS Shield Advanced Protection to more AWS Resources.

", + "CreateSubscription": "

Activates AWS Shield Advanced for an account.

As part of this request you can specify EmergencySettings that automaticaly grant the DDoS response team (DRT) needed permissions to assist you during a suspected DDoS attack. For more information see Authorize the DDoS Response Team to Create Rules and Web ACLs on Your Behalf.

When you initally create a subscription, your subscription is set to be automatically renewed at the end of the existing subscription period. You can change this by submitting an UpdateSubscription request.

", "DeleteProtection": "

Deletes an AWS Shield Advanced Protection.

", "DeleteSubscription": "

Removes AWS Shield Advanced from an account. AWS Shield Advanced requires a 1-year subscription commitment. You cannot delete a subscription prior to the completion of that commitment.

", "DescribeAttack": "

Describes the details of a DDoS attack.

", + "DescribeDRTAccess": "

Returns the current role and list of Amazon S3 log buckets used by the DDoS Response team (DRT) to access your AWS account while assisting with attack mitigation.

", + "DescribeEmergencyContactSettings": "

Lists the email addresses that the DRT can use to contact you during a suspected attack.

", "DescribeProtection": "

Lists the details of a Protection object.

", "DescribeSubscription": "

Provides details about the AWS Shield Advanced subscription for an account.

", + "DisassociateDRTLogBucket": "

Removes the DDoS Response team's (DRT) access to the specified Amazon S3 bucket containing your flow logs.

To make a DisassociateDRTLogBucket request, you must be subscribed to the Business Support plan or the Enterprise Support plan. However, if you are not subscribed to one of these support plans, but had been previously and had granted the DRT access to your account, you can submit a DisassociateDRTLogBucket request to remove this access.

", + "DisassociateDRTRole": "

Removes the DDoS Response team's (DRT) access to your AWS account.

To make a DisassociateDRTRole request, you must be subscribed to the Business Support plan or the Enterprise Support plan. However, if you are not subscribed to one of these support plans, but had been previously and had granted the DRT access to your account, you can submit a DisassociateDRTRole request to remove this access.

", "GetSubscriptionState": "

Returns the SubscriptionState, either Active or Inactive.

", "ListAttacks": "

Returns all ongoing DDoS attacks or all DDoS attacks during a specified time period.

", - "ListProtections": "

Lists all Protection objects for the account.

" + "ListProtections": "

Lists all Protection objects for the account.

", + "UpdateEmergencyContactSettings": "

Updates the details of the list of email addresses that the DRT can use to contact you during a suspected attack.

", + "UpdateSubscription": "

Updates the details of an existing subscription. Only enter values for parameters you want to change. Empty parameters are not updated.

" }, "shapes": { + "AccessDeniedForDependencyException": { + "base": "

In order to grant the necessary access to the DDoS Response Team, the user submitting AssociateDRTRole must have the iam:PassRole permission. This error indicates the user did not have the appropriate permissions. For more information, see Granting a User Permissions to Pass a Role to an AWS Service.

", + "refs": { + } + }, + "AssociateDRTLogBucketRequest": { + "base": null, + "refs": { + } + }, + "AssociateDRTLogBucketResponse": { + "base": null, + "refs": { + } + }, + "AssociateDRTRoleRequest": { + "base": null, + "refs": { + } + }, + "AssociateDRTRoleResponse": { + "base": null, + "refs": { + } + }, "AttackDetail": { "base": "

The details of a DDoS attack.

", "refs": { @@ -86,6 +119,13 @@ "AttackSummary$AttackVectors": "

The list of attacks for a specified time period.

" } }, + "AutoRenew": { + "base": null, + "refs": { + "Subscription$AutoRenew": "

If ENABLED, the subscription will be automatically renewed at the end of the existing subscription period.

When you initally create a subscription, AutoRenew is set to ENABLED. You can change this by submitting an UpdateSubscription request. If the UpdateSubscription request does not included a value for AutoRenew, the existing value for AutoRenew remains unchanged.

", + "UpdateSubscriptionRequest$AutoRenew": "

When you initally create a subscription, AutoRenew is set to ENABLED. If ENABLED, the subscription will be automatically renewed at the end of the existing subscription period. You can change this by submitting an UpdateSubscription request. If the UpdateSubscription request does not included a value for AutoRenew, the existing value for AutoRenew remains unchanged.

" + } + }, "Contributor": { "base": "

A contributor to the attack and their contribution.

", "refs": { @@ -142,6 +182,26 @@ "refs": { } }, + "DescribeDRTAccessRequest": { + "base": null, + "refs": { + } + }, + "DescribeDRTAccessResponse": { + "base": null, + "refs": { + } + }, + "DescribeEmergencyContactSettingsRequest": { + "base": null, + "refs": { + } + }, + "DescribeEmergencyContactSettingsResponse": { + "base": null, + "refs": { + } + }, "DescribeProtectionRequest": { "base": null, "refs": { @@ -162,6 +222,26 @@ "refs": { } }, + "DisassociateDRTLogBucketRequest": { + "base": null, + "refs": { + } + }, + "DisassociateDRTLogBucketResponse": { + "base": null, + "refs": { + } + }, + "DisassociateDRTRoleRequest": { + "base": null, + "refs": { + } + }, + "DisassociateDRTRoleResponse": { + "base": null, + "refs": { + } + }, "Double": { "base": null, "refs": { @@ -176,6 +256,25 @@ "Subscription$TimeCommitmentInSeconds": "

The length, in seconds, of the AWS Shield Advanced subscription for the account.

" } }, + "EmailAddress": { + "base": null, + "refs": { + "EmergencyContact$EmailAddress": "

An email address that the DRT can use to contact you during a suspected attack.

" + } + }, + "EmergencyContact": { + "base": "

Contact information that the DRT can use to contact you during a suspected attack.

", + "refs": { + "EmergencyContactList$member": null + } + }, + "EmergencyContactList": { + "base": null, + "refs": { + "DescribeEmergencyContactSettingsResponse$EmergencyContactList": "

A list of email addresses that the DRT can use to contact you during a suspected attack.

", + "UpdateEmergencyContactSettingsRequest$EmergencyContactList": "

A list of email addresses that the DRT can use to contact you during a suspected attack.

" + } + }, "GetSubscriptionStateRequest": { "base": null, "refs": { @@ -202,6 +301,11 @@ "refs": { } }, + "InvalidPaginationTokenException": { + "base": "

Exception that indicates that the NextToken specified in the request is invalid. Submit the request using the NextToken value that was returned in the response.

", + "refs": { + } + }, "InvalidParameterException": { "base": "

Exception that indicates that the parameters passed to the API are invalid.

", "refs": { @@ -212,6 +316,12 @@ "refs": { } }, + "Limit": { + "base": "

Specifies how many protections of a given type you can create.

", + "refs": { + "Limits$member": null + } + }, "LimitNumber": { "base": null, "refs": { @@ -224,6 +334,12 @@ "LimitsExceededException$Type": null } }, + "Limits": { + "base": null, + "refs": { + "Subscription$Limits": "

Specifies how many protections of a given type you can create.

" + } + }, "LimitsExceededException": { "base": "

Exception that indicates that the operation would exceed a limit.

Type is the type of limit that would be exceeded.

Limit is the threshold that would be exceeded.

", "refs": { @@ -250,15 +366,30 @@ } }, "LockedSubscriptionException": { - "base": "

Exception that indicates that the subscription you are trying to delete has not yet completed the 1-year commitment. You cannot delete this subscription.

", + "base": "

You are trying to update a subscription that has not yet completed the 1-year commitment. You can change the AutoRenew parameter during the last 30 days of your subscription. This exception indicates that you are attempting to change AutoRenew prior to that period.

", "refs": { } }, + "LogBucket": { + "base": null, + "refs": { + "AssociateDRTLogBucketRequest$LogBucket": "

The Amazon S3 bucket that contains your flow logs.

", + "DisassociateDRTLogBucketRequest$LogBucket": "

The Amazon S3 bucket that contains your flow logs.

", + "LogBucketList$member": null + } + }, + "LogBucketList": { + "base": null, + "refs": { + "DescribeDRTAccessResponse$LogBucketList": "

The list of Amazon S3 buckets accessed by the DRT.

" + } + }, "Long": { "base": null, "refs": { "AttackProperty$Total": "

The total contributions made to this attack by all contributors, not just the five listed in the TopContributors list.

", - "Contributor$Value": "

The contribution of this contributor expressed in Protection units. For example 10,000.

" + "Contributor$Value": "

The contribution of this contributor expressed in Protection units. For example 10,000.

", + "Limit$Max": "

The maximum number of protections that can be created for the specified Type.

" } }, "MaxResults": { @@ -280,6 +411,11 @@ "AttackDetail$Mitigations": "

List of mitigation actions taken for the attack.

" } }, + "NoAssociatedRoleException": { + "base": "

The ARN of the role that you specifed does not exist.

", + "refs": { + } + }, "OptimisticLockException": { "base": "

Exception that indicates that the protection state has been modified by another client. You can retry the request.

", "refs": { @@ -339,6 +475,13 @@ "refs": { } }, + "RoleArn": { + "base": null, + "refs": { + "AssociateDRTRoleRequest$RoleArn": "

The Amazon Resource Name (ARN) of the role the DRT will use to access your AWS account.

Prior to making the AssociateDRTRole request, you must attach the AWSShieldDRTAccessPolicy managed policy to this role. For more information see Attaching and Detaching IAM Policies.

", + "DescribeDRTAccessResponse$RoleArn": "

The Amazon Resource Name (ARN) of the role the DRT used to access your AWS account.

" + } + }, "String": { "base": null, "refs": { @@ -346,6 +489,7 @@ "AttackSummary$ResourceArn": "

The ARN (Amazon Resource Name) of the resource that was attacked.

", "AttackVectorDescription$VectorType": "

The attack type. Valid values:

  • UDP_TRAFFIC

  • UDP_FRAGMENT

  • GENERIC_UDP_REFLECTION

  • DNS_REFLECTION

  • NTP_REFLECTION

  • CHARGEN_REFLECTION

  • SSDP_REFLECTION

  • PORT_MAPPER

  • RIP_REFLECTION

  • SNMP_REFLECTION

  • MSSQL_REFLECTION

  • NET_BIOS_REFLECTION

  • SYN_FLOOD

  • ACK_FLOOD

  • REQUEST_FLOOD

", "Contributor$Name": "

The name of the contributor. This is dependent on the AttackPropertyIdentifier. For example, if the AttackPropertyIdentifier is SOURCE_COUNTRY, the Name could be United States.

", + "Limit$Type": "

The type of protection.

", "Mitigation$MitigationName": "

The name of the mitigation taken for this attack.

", "SubResourceSummary$Id": "

The unique identifier (ID) of the SubResource.

", "SummarizedAttackVector$VectorType": "

The attack type, for example, SNMP reflection or SYN flood.

", @@ -419,7 +563,8 @@ "Timestamp": { "base": null, "refs": { - "Subscription$StartTime": "

The start time of the subscription, in Unix time in seconds. For more information see timestamp.

" + "Subscription$StartTime": "

The start time of the subscription, in Unix time in seconds. For more information see timestamp.

", + "Subscription$EndTime": "

The date and time your subscription will end.

" } }, "Token": { @@ -443,15 +588,38 @@ "AttackProperty$Unit": "

The unit of the Value of the contributions.

" } }, + "UpdateEmergencyContactSettingsRequest": { + "base": null, + "refs": { + } + }, + "UpdateEmergencyContactSettingsResponse": { + "base": null, + "refs": { + } + }, + "UpdateSubscriptionRequest": { + "base": null, + "refs": { + } + }, + "UpdateSubscriptionResponse": { + "base": null, + "refs": { + } + }, "errorMessage": { "base": null, "refs": { + "AccessDeniedForDependencyException$message": null, "InternalErrorException$message": null, "InvalidOperationException$message": null, + "InvalidPaginationTokenException$message": null, "InvalidParameterException$message": null, "InvalidResourceException$message": null, "LimitsExceededException$message": null, "LockedSubscriptionException$message": null, + "NoAssociatedRoleException$message": null, "OptimisticLockException$message": null, "ResourceAlreadyExistsException$message": null, "ResourceNotFoundException$message": null diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/api-2.json index de483026a..944854b9b 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/api-2.json @@ -1,13 +1,14 @@ { "version":"2.0", "metadata":{ - "uid":"sns-2010-03-31", "apiVersion":"2010-03-31", "endpointPrefix":"sns", "protocol":"query", "serviceAbbreviation":"Amazon SNS", "serviceFullName":"Amazon Simple Notification Service", + "serviceId":"SNS", "signatureVersion":"v4", + "uid":"sns-2010-03-31", "xmlNamespace":"http://sns.amazonaws.com/doc/2010-03-31/" }, "operations":{ @@ -458,6 +459,7 @@ "input":{"shape":"SetSubscriptionAttributesInput"}, "errors":[ {"shape":"InvalidParameterException"}, + {"shape":"FilterPolicyLimitExceededException"}, {"shape":"InternalErrorException"}, {"shape":"NotFoundException"}, {"shape":"AuthorizationErrorException"} @@ -490,6 +492,7 @@ }, "errors":[ {"shape":"SubscriptionLimitExceededException"}, + {"shape":"FilterPolicyLimitExceededException"}, {"shape":"InvalidParameterException"}, {"shape":"InternalErrorException"}, {"shape":"NotFoundException"}, @@ -670,6 +673,18 @@ }, "exception":true }, + "FilterPolicyLimitExceededException":{ + "type":"structure", + "members":{ + "message":{"shape":"string"} + }, + "error":{ + "code":"FilterPolicyLimitExceeded", + "httpStatusCode":403, + "senderFault":true + }, + "exception":true + }, "GetEndpointAttributesInput":{ "type":"structure", "required":["EndpointArn"], @@ -1035,7 +1050,9 @@ "members":{ "TopicArn":{"shape":"topicARN"}, "Protocol":{"shape":"protocol"}, - "Endpoint":{"shape":"endpoint"} + "Endpoint":{"shape":"endpoint"}, + "Attributes":{"shape":"SubscriptionAttributesMap"}, + "ReturnSubscriptionArn":{"shape":"boolean"} } }, "SubscribeResponse":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/docs-2.json index 6073ae181..750def982 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/docs-2.json @@ -16,22 +16,22 @@ "GetSMSAttributes": "

Returns the settings for sending SMS messages from your account.

These settings are set with the SetSMSAttributes action.

", "GetSubscriptionAttributes": "

Returns all of the properties of a subscription.

", "GetTopicAttributes": "

Returns all of the properties of a topic. Topic properties returned might differ based on the authorization of the user.

", - "ListEndpointsByPlatformApplication": "

Lists the endpoints and endpoint attributes for devices in a supported push notification service, such as GCM and APNS. The results for ListEndpointsByPlatformApplication are paginated and return a limited list of endpoints, up to 100. If additional records are available after the first page results, then a NextToken string will be returned. To receive the next page, you call ListEndpointsByPlatformApplication again using the NextToken string received from the previous call. When there are no more records to return, NextToken will be null. For more information, see Using Amazon SNS Mobile Push Notifications.

", + "ListEndpointsByPlatformApplication": "

Lists the endpoints and endpoint attributes for devices in a supported push notification service, such as GCM and APNS. The results for ListEndpointsByPlatformApplication are paginated and return a limited list of endpoints, up to 100. If additional records are available after the first page results, then a NextToken string will be returned. To receive the next page, you call ListEndpointsByPlatformApplication again using the NextToken string received from the previous call. When there are no more records to return, NextToken will be null. For more information, see Using Amazon SNS Mobile Push Notifications.

This action is throttled at 30 transactions per second (TPS).

", "ListPhoneNumbersOptedOut": "

Returns a list of phone numbers that are opted out, meaning you cannot send SMS messages to them.

The results for ListPhoneNumbersOptedOut are paginated, and each page returns up to 100 phone numbers. If additional phone numbers are available after the first page of results, then a NextToken string will be returned. To receive the next page, you call ListPhoneNumbersOptedOut again using the NextToken string received from the previous call. When there are no more records to return, NextToken will be null.

", - "ListPlatformApplications": "

Lists the platform application objects for the supported push notification services, such as APNS and GCM. The results for ListPlatformApplications are paginated and return a limited list of applications, up to 100. If additional records are available after the first page results, then a NextToken string will be returned. To receive the next page, you call ListPlatformApplications using the NextToken string received from the previous call. When there are no more records to return, NextToken will be null. For more information, see Using Amazon SNS Mobile Push Notifications.

", - "ListSubscriptions": "

Returns a list of the requester's subscriptions. Each call returns a limited list of subscriptions, up to 100. If there are more subscriptions, a NextToken is also returned. Use the NextToken parameter in a new ListSubscriptions call to get further results.

", - "ListSubscriptionsByTopic": "

Returns a list of the subscriptions to a specific topic. Each call returns a limited list of subscriptions, up to 100. If there are more subscriptions, a NextToken is also returned. Use the NextToken parameter in a new ListSubscriptionsByTopic call to get further results.

", - "ListTopics": "

Returns a list of the requester's topics. Each call returns a limited list of topics, up to 100. If there are more topics, a NextToken is also returned. Use the NextToken parameter in a new ListTopics call to get further results.

", + "ListPlatformApplications": "

Lists the platform application objects for the supported push notification services, such as APNS and GCM. The results for ListPlatformApplications are paginated and return a limited list of applications, up to 100. If additional records are available after the first page results, then a NextToken string will be returned. To receive the next page, you call ListPlatformApplications using the NextToken string received from the previous call. When there are no more records to return, NextToken will be null. For more information, see Using Amazon SNS Mobile Push Notifications.

This action is throttled at 15 transactions per second (TPS).

", + "ListSubscriptions": "

Returns a list of the requester's subscriptions. Each call returns a limited list of subscriptions, up to 100. If there are more subscriptions, a NextToken is also returned. Use the NextToken parameter in a new ListSubscriptions call to get further results.

This action is throttled at 30 transactions per second (TPS).

", + "ListSubscriptionsByTopic": "

Returns a list of the subscriptions to a specific topic. Each call returns a limited list of subscriptions, up to 100. If there are more subscriptions, a NextToken is also returned. Use the NextToken parameter in a new ListSubscriptionsByTopic call to get further results.

This action is throttled at 30 transactions per second (TPS).

", + "ListTopics": "

Returns a list of the requester's topics. Each call returns a limited list of topics, up to 100. If there are more topics, a NextToken is also returned. Use the NextToken parameter in a new ListTopics call to get further results.

This action is throttled at 30 transactions per second (TPS).

", "OptInPhoneNumber": "

Use this request to opt in a phone number that is opted out, which enables you to resume sending SMS messages to the number.

You can opt in a phone number only once every 30 days.

", - "Publish": "

Sends a message to all of a topic's subscribed endpoints. When a messageId is returned, the message has been saved and Amazon SNS will attempt to deliver it to the topic's subscribers shortly. The format of the outgoing message to each subscribed endpoint depends on the notification protocol.

To use the Publish action for sending a message to a mobile endpoint, such as an app on a Kindle device or mobile phone, you must specify the EndpointArn for the TargetArn parameter. The EndpointArn is returned when making a call with the CreatePlatformEndpoint action.

For more information about formatting messages, see Send Custom Platform-Specific Payloads in Messages to Mobile Devices.

", + "Publish": "

Sends a message to an Amazon SNS topic or sends a text message (SMS message) directly to a phone number.

If you send a message to a topic, Amazon SNS delivers the message to each endpoint that is subscribed to the topic. The format of the message depends on the notification protocol for each subscribed endpoint.

When a messageId is returned, the message has been saved and Amazon SNS will attempt to deliver it shortly.

To use the Publish action for sending a message to a mobile endpoint, such as an app on a Kindle device or mobile phone, you must specify the EndpointArn for the TargetArn parameter. The EndpointArn is returned when making a call with the CreatePlatformEndpoint action.

For more information about formatting messages, see Send Custom Platform-Specific Payloads in Messages to Mobile Devices.

", "RemovePermission": "

Removes a statement from a topic's access control policy.

", "SetEndpointAttributes": "

Sets the attributes for an endpoint for a device on one of the supported push notification services, such as GCM and APNS. For more information, see Using Amazon SNS Mobile Push Notifications.

", "SetPlatformApplicationAttributes": "

Sets the attributes of the platform application object for the supported push notification services, such as APNS and GCM. For more information, see Using Amazon SNS Mobile Push Notifications. For information on configuring attributes for message delivery status, see Using Amazon SNS Application Attributes for Message Delivery Status.

", "SetSMSAttributes": "

Use this request to set the default settings for sending SMS messages and receiving daily SMS usage reports.

You can override some of these settings for a single message when you use the Publish action with the MessageAttributes.entry.N parameter. For more information, see Sending an SMS Message in the Amazon SNS Developer Guide.

", - "SetSubscriptionAttributes": "

Allows a subscription owner to set an attribute of the topic to a new value.

", + "SetSubscriptionAttributes": "

Allows a subscription owner to set an attribute of the subscription to a new value.

", "SetTopicAttributes": "

Allows a topic owner to set an attribute of the topic to a new value.

", - "Subscribe": "

Prepares to subscribe an endpoint by sending the endpoint a confirmation message. To actually create a subscription, the endpoint owner must call the ConfirmSubscription action with the token from the confirmation message. Confirmation tokens are valid for three days.

", - "Unsubscribe": "

Deletes a subscription. If the subscription requires authentication for deletion, only the owner of the subscription or the topic's owner can unsubscribe, and an AWS signature is required. If the Unsubscribe call does not require authentication and the requester is not the subscription owner, a final cancellation message is delivered to the endpoint, so that the endpoint owner can easily resubscribe to the topic if the Unsubscribe request was unintended.

" + "Subscribe": "

Prepares to subscribe an endpoint by sending the endpoint a confirmation message. To actually create a subscription, the endpoint owner must call the ConfirmSubscription action with the token from the confirmation message. Confirmation tokens are valid for three days.

This action is throttled at 100 transactions per second (TPS).

", + "Unsubscribe": "

Deletes a subscription. If the subscription requires authentication for deletion, only the owner of the subscription or the topic's owner can unsubscribe, and an AWS signature is required. If the Unsubscribe call does not require authentication and the requester is not the subscription owner, a final cancellation message is delivered to the endpoint, so that the endpoint owner can easily resubscribe to the topic if the Unsubscribe request was unintended.

This action is throttled at 100 transactions per second (TPS).

" }, "shapes": { "ActionsList": { @@ -138,6 +138,11 @@ "refs": { } }, + "FilterPolicyLimitExceededException": { + "base": "

Indicates that the number of filter polices in your AWS account exceeds the limit. To add more filter polices, submit an SNS Limit Increase case in the AWS Support Center.

", + "refs": { + } + }, "GetEndpointAttributesInput": { "base": "

Input for GetEndpointAttributes action.

", "refs": { @@ -293,7 +298,7 @@ "PlatformApplication$Attributes": "

Attributes for platform application object.

", "SetEndpointAttributesInput$Attributes": "

A map of the endpoint attributes. Attributes in this map include the following:

  • CustomUserData -- arbitrary user data to associate with the endpoint. Amazon SNS does not use this data. The data must be in UTF-8 format and less than 2KB.

  • Enabled -- flag that enables/disables delivery to the endpoint. Amazon SNS will set this to false when a notification service indicates to Amazon SNS that the endpoint is invalid. Users can set it back to true, typically after updating Token.

  • Token -- device token, also referred to as a registration id, for an app and mobile device. This is returned from the notification service when an app and mobile device are registered with the notification service.

", "SetPlatformApplicationAttributesInput$Attributes": "

A map of the platform application attributes. Attributes in this map include the following:

  • PlatformCredential -- The credential received from the notification service. For APNS/APNS_SANDBOX, PlatformCredential is private key. For GCM, PlatformCredential is \"API key\". For ADM, PlatformCredential is \"client secret\".

  • PlatformPrincipal -- The principal received from the notification service. For APNS/APNS_SANDBOX, PlatformPrincipal is SSL certificate. For GCM, PlatformPrincipal is not applicable. For ADM, PlatformPrincipal is \"client id\".

  • EventEndpointCreated -- Topic ARN to which EndpointCreated event notifications should be sent.

  • EventEndpointDeleted -- Topic ARN to which EndpointDeleted event notifications should be sent.

  • EventEndpointUpdated -- Topic ARN to which EndpointUpdate event notifications should be sent.

  • EventDeliveryFailure -- Topic ARN to which DeliveryFailure event notifications should be sent upon Direct Publish delivery failure (permanent) to one of the application's endpoints.

  • SuccessFeedbackRoleArn -- IAM role ARN used to give Amazon SNS write access to use CloudWatch Logs on your behalf.

  • FailureFeedbackRoleArn -- IAM role ARN used to give Amazon SNS write access to use CloudWatch Logs on your behalf.

  • SuccessFeedbackSampleRate -- Sample rate percentage (0-100) of successfully delivered messages.

", - "SetSMSAttributesInput$attributes": "

The default settings for sending SMS messages from your account. You can set values for the following attribute names:

MonthlySpendLimit – The maximum amount in USD that you are willing to spend each month to send SMS messages. When Amazon SNS determines that sending an SMS message would incur a cost that exceeds this limit, it stops sending SMS messages within minutes.

Amazon SNS stops sending SMS messages within minutes of the limit being crossed. During that interval, if you continue to send SMS messages, you will incur costs that exceed your limit.

By default, the spend limit is set to the maximum allowed by Amazon SNS. If you want to exceed the maximum, contact AWS Support or your AWS sales representative for a service limit increase.

DeliveryStatusIAMRole – The ARN of the IAM role that allows Amazon SNS to write logs about SMS deliveries in CloudWatch Logs. For each SMS message that you send, Amazon SNS writes a log that includes the message price, the success or failure status, the reason for failure (if the message failed), the message dwell time, and other information.

DeliveryStatusSuccessSamplingRate – The percentage of successful SMS deliveries for which Amazon SNS will write logs in CloudWatch Logs. The value can be an integer from 0 - 100. For example, to write logs only for failed deliveries, set this value to 0. To write logs for 10% of your successful deliveries, set it to 10.

DefaultSenderID – A string, such as your business brand, that is displayed as the sender on the receiving device. Support for sender IDs varies by country. The sender ID can be 1 - 11 alphanumeric characters, and it must contain at least one letter.

DefaultSMSType – The type of SMS message that you will send by default. You can assign the following values:

  • Promotional – (Default) Noncritical messages, such as marketing messages. Amazon SNS optimizes the message delivery to incur the lowest cost.

  • Transactional – Critical messages that support customer transactions, such as one-time passcodes for multi-factor authentication. Amazon SNS optimizes the message delivery to achieve the highest reliability.

UsageReportS3Bucket – The name of the Amazon S3 bucket to receive daily SMS usage reports from Amazon SNS. Each day, Amazon SNS will deliver a usage report as a CSV file to the bucket. The report includes the following information for each SMS message that was successfully delivered by your account:

  • Time that the message was published (in UTC)

  • Message ID

  • Destination phone number

  • Message type

  • Delivery status

  • Message price (in USD)

  • Part number (a message is split into multiple parts if it is too long for a single message)

  • Total number of parts

To receive the report, the bucket must have a policy that allows the Amazon SNS service principle to perform the s3:PutObject and s3:GetBucketLocation actions.

For an example bucket policy and usage report, see Monitoring SMS Activity in the Amazon SNS Developer Guide.

" + "SetSMSAttributesInput$attributes": "

The default settings for sending SMS messages from your account. You can set values for the following attribute names:

MonthlySpendLimit – The maximum amount in USD that you are willing to spend each month to send SMS messages. When Amazon SNS determines that sending an SMS message would incur a cost that exceeds this limit, it stops sending SMS messages within minutes.

Amazon SNS stops sending SMS messages within minutes of the limit being crossed. During that interval, if you continue to send SMS messages, you will incur costs that exceed your limit.

By default, the spend limit is set to the maximum allowed by Amazon SNS. If you want to raise the limit, submit an SNS Limit Increase case. For New limit value, enter your desired monthly spend limit. In the Use Case Description field, explain that you are requesting an SMS monthly spend limit increase.

DeliveryStatusIAMRole – The ARN of the IAM role that allows Amazon SNS to write logs about SMS deliveries in CloudWatch Logs. For each SMS message that you send, Amazon SNS writes a log that includes the message price, the success or failure status, the reason for failure (if the message failed), the message dwell time, and other information.

DeliveryStatusSuccessSamplingRate – The percentage of successful SMS deliveries for which Amazon SNS will write logs in CloudWatch Logs. The value can be an integer from 0 - 100. For example, to write logs only for failed deliveries, set this value to 0. To write logs for 10% of your successful deliveries, set it to 10.

DefaultSenderID – A string, such as your business brand, that is displayed as the sender on the receiving device. Support for sender IDs varies by country. The sender ID can be 1 - 11 alphanumeric characters, and it must contain at least one letter.

DefaultSMSType – The type of SMS message that you will send by default. You can assign the following values:

  • Promotional – (Default) Noncritical messages, such as marketing messages. Amazon SNS optimizes the message delivery to incur the lowest cost.

  • Transactional – Critical messages that support customer transactions, such as one-time passcodes for multi-factor authentication. Amazon SNS optimizes the message delivery to achieve the highest reliability.

UsageReportS3Bucket – The name of the Amazon S3 bucket to receive daily SMS usage reports from Amazon SNS. Each day, Amazon SNS will deliver a usage report as a CSV file to the bucket. The report includes the following information for each SMS message that was successfully delivered by your account:

  • Time that the message was published (in UTC)

  • Message ID

  • Destination phone number

  • Message type

  • Delivery status

  • Message price (in USD)

  • Part number (a message is split into multiple parts if it is too long for a single message)

  • Total number of parts

To receive the report, the bucket must have a policy that allows the Amazon SNS service principle to perform the s3:PutObject and s3:GetBucketLocation actions.

For an example bucket policy and usage report, see Monitoring SMS Activity in the Amazon SNS Developer Guide.

" } }, "MessageAttributeMap": { @@ -417,7 +422,7 @@ "MapStringToString$key": null, "MapStringToString$value": null, "MessageAttributeMap$key": null, - "MessageAttributeValue$DataType": "

Amazon SNS supports the following logical data types: String, Number, and Binary. For more information, see Message Attribute Data Types.

", + "MessageAttributeValue$DataType": "

Amazon SNS supports the following logical data types: String, String.Array, Number, and Binary. For more information, see Message Attribute Data Types.

", "MessageAttributeValue$StringValue": "

Strings are Unicode with UTF8 binary encoding. For a list of code values, see http://en.wikipedia.org/wiki/ASCII#ASCII_printable_characters.

", "PlatformApplication$PlatformApplicationArn": "

PlatformApplicationArn for platform application object.

", "PublishInput$TargetArn": "

Either TopicArn or EndpointArn, but not both.

If you don't specify a value for the TargetArn parameter, you must specify a value for the PhoneNumber or TopicArn parameters.

", @@ -445,7 +450,8 @@ "SubscriptionAttributesMap": { "base": null, "refs": { - "GetSubscriptionAttributesResponse$Attributes": "

A map of the subscription's attributes. Attributes in this map include the following:

  • SubscriptionArn -- the subscription's ARN

  • TopicArn -- the topic ARN that the subscription is associated with

  • Owner -- the AWS account ID of the subscription's owner

  • ConfirmationWasAuthenticated -- true if the subscription confirmation request was authenticated

  • DeliveryPolicy -- the JSON serialization of the subscription's delivery policy

  • EffectiveDeliveryPolicy -- the JSON serialization of the effective delivery policy that takes into account the topic delivery policy and account system defaults

" + "GetSubscriptionAttributesResponse$Attributes": "

A map of the subscription's attributes. Attributes in this map include the following:

  • ConfirmationWasAuthenticated -- true if the subscription confirmation request was authenticated.

  • DeliveryPolicy -- The JSON serialization of the subscription's delivery policy.

  • EffectiveDeliveryPolicy -- The JSON serialization of the effective delivery policy that takes into account the topic delivery policy and account system defaults.

  • FilterPolicy -- The filter policy JSON that is assigned to the subscription.

  • Owner -- The AWS account ID of the subscription's owner.

  • PendingConfirmation -- true if the subscription hasn't been confirmed. To confirm a pending subscription, call the ConfirmSubscription action with a confirmation token.

  • RawMessageDelivery -- true if raw message delivery is enabled for the subscription. Raw messages are free of JSON formatting and can be sent to HTTP/S and Amazon SQS endpoints.

  • SubscriptionArn -- The subscription's ARN.

  • TopicArn -- The topic ARN that the subscription is associated with.

", + "SubscribeInput$Attributes": "

Assigns attributes to the subscription as a map of key-value pairs. You can assign any attribute that is supported by the SetSubscriptionAttributes action.

" } }, "SubscriptionLimitExceededException": { @@ -508,7 +514,7 @@ "attributeName": { "base": null, "refs": { - "SetSubscriptionAttributesInput$AttributeName": "

The name of the attribute you want to set. Only a subset of the subscriptions attributes are mutable.

Valid values: DeliveryPolicy | RawMessageDelivery

", + "SetSubscriptionAttributesInput$AttributeName": "

The name of the attribute you want to set. Only a subset of the subscriptions attributes are mutable.

Valid values: DeliveryPolicy | FilterPolicy | RawMessageDelivery

", "SetTopicAttributesInput$AttributeName": "

The name of the attribute you want to set. Only a subset of the topic's attributes are mutable.

Valid values: Policy | DisplayName | DeliveryPolicy

", "SubscriptionAttributesMap$key": null, "TopicAttributesMap$key": null @@ -532,7 +538,8 @@ "boolean": { "base": null, "refs": { - "CheckIfPhoneNumberIsOptedOutResponse$isOptedOut": "

Indicates whether the phone number is opted out:

  • true – The phone number is opted out, meaning you cannot publish SMS messages to it.

  • false – The phone number is opted in, meaning you can publish SMS messages to it.

" + "CheckIfPhoneNumberIsOptedOutResponse$isOptedOut": "

Indicates whether the phone number is opted out:

  • true – The phone number is opted out, meaning you cannot publish SMS messages to it.

  • false – The phone number is opted in, meaning you can publish SMS messages to it.

", + "SubscribeInput$ReturnSubscriptionArn": "

Sets whether the response from the Subscribe request includes the subscription ARN, even if the subscription is not yet confirmed.

If you set this parameter to false, the response includes the ARN for confirmed subscriptions, but it includes an ARN value of \"pending subscription\" for subscriptions that are not yet confirmed. A subscription becomes confirmed when the subscriber calls the ConfirmSubscription action with a confirmation token.

If you set this parameter to true, the response includes the ARN in all cases, even if the subscription is not yet confirmed.

The default value is false.

" } }, "delegate": { @@ -558,7 +565,7 @@ "message": { "base": null, "refs": { - "PublishInput$Message": "

The message you want to send to the topic.

If you want to send the same message to all transport protocols, include the text of the message as a String value.

If you want to send different messages for each transport protocol, set the value of the MessageStructure parameter to json and use a JSON object for the Message parameter.

Constraints: Messages must be UTF-8 encoded strings at most 256 KB in size (262144 bytes, not 262144 characters).

JSON-specific constraints:

  • Keys in the JSON object that correspond to supported transport protocols must have simple JSON string values.

  • The values will be parsed (unescaped) before they are used in outgoing messages.

  • Outbound notifications are JSON encoded (meaning that the characters will be reescaped for sending).

  • Values have a minimum length of 0 (the empty string, \"\", is allowed).

  • Values have a maximum length bounded by the overall message size (so, including multiple protocols may limit message sizes).

  • Non-string values will cause the key to be ignored.

  • Keys that do not correspond to supported transport protocols are ignored.

  • Duplicate keys are not allowed.

  • Failure to parse or validate any key or value in the message will cause the Publish call to return an error (no partial delivery).

" + "PublishInput$Message": "

The message you want to send.

If you are publishing to a topic and you want to send the same message to all transport protocols, include the text of the message as a String value. If you want to send different messages for each transport protocol, set the value of the MessageStructure parameter to json and use a JSON object for the Message parameter.

Constraints:

  • With the exception of SMS, messages must be UTF-8 encoded strings and at most 256 KB in size (262144 bytes, not 262144 characters).

  • For SMS, each message can contain up to 140 bytes, and the character limit depends on the encoding scheme. For example, an SMS message can contain 160 GSM characters, 140 ASCII characters, or 70 UCS-2 characters. If you publish a message that exceeds the size limit, Amazon SNS sends it as multiple messages, each fitting within the size limit. Messages are not cut off in the middle of a word but on whole-word boundaries. The total size limit for a single SMS publish action is 1600 bytes.

JSON-specific constraints:

  • Keys in the JSON object that correspond to supported transport protocols must have simple JSON string values.

  • The values will be parsed (unescaped) before they are used in outgoing messages.

  • Outbound notifications are JSON encoded (meaning that the characters will be reescaped for sending).

  • Values have a minimum length of 0 (the empty string, \"\", is allowed).

  • Values have a maximum length bounded by the overall message size (so, including multiple protocols may limit message sizes).

  • Non-string values will cause the key to be ignored.

  • Keys that do not correspond to supported transport protocols are ignored.

  • Duplicate keys are not allowed.

  • Failure to parse or validate any key or value in the message will cause the Publish call to return an error (no partial delivery).

" } }, "messageId": { @@ -596,6 +603,7 @@ "refs": { "AuthorizationErrorException$message": null, "EndpointDisabledException$message": "

Message for endpoint disabled.

", + "FilterPolicyLimitExceededException$message": null, "InternalErrorException$message": null, "InvalidParameterException$message": null, "InvalidParameterValueException$message": "

The parameter value is invalid.

", @@ -620,7 +628,7 @@ "ConfirmSubscriptionResponse$SubscriptionArn": "

The ARN of the created subscription.

", "GetSubscriptionAttributesInput$SubscriptionArn": "

The ARN of the subscription whose properties you want to get.

", "SetSubscriptionAttributesInput$SubscriptionArn": "

The ARN of the subscription to modify.

", - "SubscribeResponse$SubscriptionArn": "

The ARN of the subscription, if the service was able to create a subscription immediately (without requiring endpoint owner confirmation).

", + "SubscribeResponse$SubscriptionArn": "

The ARN of the subscription if it is confirmed, or the string \"pending confirmation\" if the subscription requires confirmation. However, if the API request parameter ReturnSubscriptionArn is true, then the value is always the subscription ARN, even if the subscription requires confirmation.

", "Subscription$SubscriptionArn": "

The subscription's ARN.

", "UnsubscribeInput$SubscriptionArn": "

The ARN of the subscription to be deleted.

" } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/paginators-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/paginators-1.json index 455e4708e..df5bc6bdd 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/paginators-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/paginators-1.json @@ -26,4 +26,4 @@ "result_key": "Topics" } } -} +} \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/smoke.json b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/smoke.json similarity index 63% rename from vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/smoke.json rename to vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/smoke.json index 068b23492..0dcf07ba9 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/rds/2013-09-09/smoke.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/sns/2010-03-31/smoke.json @@ -3,14 +3,15 @@ "defaultRegion": "us-west-2", "testCases": [ { - "operationName": "DescribeDBEngineVersions", + "operationName": "ListTopics", "input": {}, "errorExpectedFromService": false }, { - "operationName": "DescribeDBInstances", + "operationName": "Publish", "input": { - "DBInstanceIdentifier": "fake-id" + "Message": "hello", + "TopicArn": "fake_topic" }, "errorExpectedFromService": true } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/api-2.json index 4e872229c..ae7e0062f 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/api-2.json @@ -1950,6 +1950,20 @@ "max":64, "min":1 }, + "CloudWatchLogGroupName":{ + "type":"string", + "max":512, + "min":1, + "pattern":"[\\.\\-_/#A-Za-z0-9]+" + }, + "CloudWatchOutputConfig":{ + "type":"structure", + "members":{ + "CloudWatchLogGroupName":{"shape":"CloudWatchLogGroupName"}, + "CloudWatchOutputEnabled":{"shape":"CloudWatchOutputEnabled"} + } + }, + "CloudWatchOutputEnabled":{"type":"boolean"}, "Command":{ "type":"structure", "members":{ @@ -1972,8 +1986,10 @@ "TargetCount":{"shape":"TargetCount"}, "CompletedCount":{"shape":"CompletedCount"}, "ErrorCount":{"shape":"ErrorCount"}, + "DeliveryTimedOutCount":{"shape":"DeliveryTimedOutCount"}, "ServiceRole":{"shape":"ServiceRole"}, - "NotificationConfig":{"shape":"NotificationConfig"} + "NotificationConfig":{"shape":"NotificationConfig"}, + "CloudWatchOutputConfig":{"shape":"CloudWatchOutputConfig"} } }, "CommandFilter":{ @@ -2027,7 +2043,8 @@ "StandardErrorUrl":{"shape":"Url"}, "CommandPlugins":{"shape":"CommandPluginList"}, "ServiceRole":{"shape":"ServiceRole"}, - "NotificationConfig":{"shape":"NotificationConfig"} + "NotificationConfig":{"shape":"NotificationConfig"}, + "CloudWatchOutputConfig":{"shape":"CloudWatchOutputConfig"} } }, "CommandInvocationList":{ @@ -2601,6 +2618,7 @@ "members":{ } }, + "DeliveryTimedOutCount":{"type":"integer"}, "DeregisterManagedInstanceRequest":{ "type":"structure", "required":["InstanceId"], @@ -3522,7 +3540,8 @@ "StandardOutputContent":{"shape":"StandardOutputContent"}, "StandardOutputUrl":{"shape":"Url"}, "StandardErrorContent":{"shape":"StandardErrorContent"}, - "StandardErrorUrl":{"shape":"Url"} + "StandardErrorUrl":{"shape":"Url"}, + "CloudWatchOutputConfig":{"shape":"CloudWatchOutputConfig"} } }, "GetDefaultPatchBaselineRequest":{ @@ -5470,6 +5489,7 @@ "enum":[ "WINDOWS", "AMAZON_LINUX", + "AMAZON_LINUX_2", "UBUNTU", "REDHAT_ENTERPRISE_LINUX", "SUSE", @@ -6444,7 +6464,8 @@ "MaxConcurrency":{"shape":"MaxConcurrency"}, "MaxErrors":{"shape":"MaxErrors"}, "ServiceRoleArn":{"shape":"ServiceRole"}, - "NotificationConfig":{"shape":"NotificationConfig"} + "NotificationConfig":{"shape":"NotificationConfig"}, + "CloudWatchOutputConfig":{"shape":"CloudWatchOutputConfig"} } }, "SendCommandResult":{ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/docs-2.json index edbe2359f..68b5d27dc 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/docs-2.json @@ -1,5304 +1,5331 @@ { - "version": "2.0", - "service": "AWS Systems Manager

AWS Systems Manager is a collection of capabilities that helps you automate management tasks such as collecting system inventory, applying operating system (OS) patches, automating the creation of Amazon Machine Images (AMIs), and configuring operating systems (OSs) and applications at scale. Systems Manager lets you remotely and securely manage the configuration of your managed instances. A managed instance is any Amazon EC2 instance or on-premises machine in your hybrid environment that has been configured for Systems Manager.

This reference is intended to be used with the AWS Systems Manager User Guide.

To get started, verify prerequisites and configure managed instances. For more information, see Systems Manager Prerequisites.

For information about other API actions you can perform on Amazon EC2 instances, see the Amazon EC2 API Reference. For information about how to use a Query API, see Making API Requests.

", - "operations": { - "AddTagsToResource": "

Adds or overwrites one or more tags for the specified resource. Tags are metadata that you can assign to your documents, managed instances, Maintenance Windows, Parameter Store parameters, and patch baselines. Tags enable you to categorize your resources in different ways, for example, by purpose, owner, or environment. Each tag consists of a key and an optional value, both of which you define. For example, you could define a set of tags for your account's managed instances that helps you track each instance's owner and stack level. For example: Key=Owner and Value=DbAdmin, SysAdmin, or Dev. Or Key=Stack and Value=Production, Pre-Production, or Test.

Each resource can have a maximum of 50 tags.

We recommend that you devise a set of tag keys that meets your needs for each resource type. Using a consistent set of tag keys makes it easier for you to manage your resources. You can search and filter the resources based on the tags you add. Tags don't have any semantic meaning to Amazon EC2 and are interpreted strictly as a string of characters.

For more information about tags, see Tagging Your Amazon EC2 Resources in the Amazon EC2 User Guide.

", - "CancelCommand": "

Attempts to cancel the command specified by the Command ID. There is no guarantee that the command will be terminated and the underlying process stopped.

", - "CreateActivation": "

Registers your on-premises server or virtual machine with Amazon EC2 so that you can manage these resources using Run Command. An on-premises server or virtual machine that has been registered with EC2 is called a managed instance. For more information about activations, see Setting Up Systems Manager in Hybrid Environments.

", - "CreateAssociation": "

Associates the specified Systems Manager document with the specified instances or targets.

When you associate a document with one or more instances using instance IDs or tags, the SSM Agent running on the instance processes the document and configures the instance as specified.

If you associate a document with an instance that already has an associated document, the system throws the AssociationAlreadyExists exception.

", - "CreateAssociationBatch": "

Associates the specified Systems Manager document with the specified instances or targets.

When you associate a document with one or more instances using instance IDs or tags, the SSM Agent running on the instance processes the document and configures the instance as specified.

If you associate a document with an instance that already has an associated document, the system throws the AssociationAlreadyExists exception.

", - "CreateDocument": "

Creates a Systems Manager document.

After you create a document, you can use CreateAssociation to associate it with one or more running instances.

", - "CreateMaintenanceWindow": "

Creates a new Maintenance Window.

", - "CreatePatchBaseline": "

Creates a patch baseline.

For information about valid key and value pairs in PatchFilters for each supported operating system type, see PatchFilter.

", - "CreateResourceDataSync": "

Creates a resource data sync configuration to a single bucket in Amazon S3. This is an asynchronous operation that returns immediately. After a successful initial sync is completed, the system continuously syncs data to the Amazon S3 bucket. To check the status of the sync, use the ListResourceDataSync.

By default, data is not encrypted in Amazon S3. We strongly recommend that you enable encryption in Amazon S3 to ensure secure data storage. We also recommend that you secure access to the Amazon S3 bucket by creating a restrictive bucket policy. To view an example of a restrictive Amazon S3 bucket policy for Resource Data Sync, see Configuring Resource Data Sync for Inventory.

", - "DeleteActivation": "

Deletes an activation. You are not required to delete an activation. If you delete an activation, you can no longer use it to register additional managed instances. Deleting an activation does not de-register managed instances. You must manually de-register managed instances.

", - "DeleteAssociation": "

Disassociates the specified Systems Manager document from the specified instance.

When you disassociate a document from an instance, it does not change the configuration of the instance. To change the configuration state of an instance after you disassociate a document, you must create a new document with the desired configuration and associate it with the instance.

", - "DeleteDocument": "

Deletes the Systems Manager document and all instance associations to the document.

Before you delete the document, we recommend that you use DeleteAssociation to disassociate all instances that are associated with the document.

", - "DeleteInventory": "

Delete a custom inventory type, or the data associated with a custom Inventory type. Deleting a custom inventory type is also referred to as deleting a custom inventory schema.

", - "DeleteMaintenanceWindow": "

Deletes a Maintenance Window.

", - "DeleteParameter": "

Delete a parameter from the system.

", - "DeleteParameters": "

Delete a list of parameters. This API is used to delete parameters by using the Amazon EC2 console.

", - "DeletePatchBaseline": "

Deletes a patch baseline.

", - "DeleteResourceDataSync": "

Deletes a Resource Data Sync configuration. After the configuration is deleted, changes to inventory data on managed instances are no longer synced with the target Amazon S3 bucket. Deleting a sync configuration does not delete data in the target Amazon S3 bucket.

", - "DeregisterManagedInstance": "

Removes the server or virtual machine from the list of registered servers. You can reregister the instance again at any time. If you don't plan to use Run Command on the server, we suggest uninstalling the SSM Agent first.

", - "DeregisterPatchBaselineForPatchGroup": "

Removes a patch group from a patch baseline.

", - "DeregisterTargetFromMaintenanceWindow": "

Removes a target from a Maintenance Window.

", - "DeregisterTaskFromMaintenanceWindow": "

Removes a task from a Maintenance Window.

", - "DescribeActivations": "

Details about the activation, including: the date and time the activation was created, the expiration date, the IAM role assigned to the instances in the activation, and the number of instances activated by this registration.

", - "DescribeAssociation": "

Describes the association for the specified target or instance. If you created the association by using the Targets parameter, then you must retrieve the association by using the association ID. If you created the association by specifying an instance ID and a Systems Manager document, then you retrieve the association by specifying the document name and the instance ID.

", - "DescribeAutomationExecutions": "

Provides details about all active and terminated Automation executions.

", - "DescribeAutomationStepExecutions": "

Information about all active and terminated step executions in an Automation workflow.

", - "DescribeAvailablePatches": "

Lists all patches that could possibly be included in a patch baseline.

", - "DescribeDocument": "

Describes the specified Systems Manager document.

", - "DescribeDocumentPermission": "

Describes the permissions for a Systems Manager document. If you created the document, you are the owner. If a document is shared, it can either be shared privately (by specifying a user's AWS account ID) or publicly (All).

", - "DescribeEffectiveInstanceAssociations": "

All associations for the instance(s).

", - "DescribeEffectivePatchesForPatchBaseline": "

Retrieves the current effective patches (the patch and the approval state) for the specified patch baseline. Note that this API applies only to Windows patch baselines.

", - "DescribeInstanceAssociationsStatus": "

The status of the associations for the instance(s).

", - "DescribeInstanceInformation": "

Describes one or more of your instances. You can use this to get information about instances like the operating system platform, the SSM Agent version (Linux), status etc. If you specify one or more instance IDs, it returns information for those instances. If you do not specify instance IDs, it returns information for all your instances. If you specify an instance ID that is not valid or an instance that you do not own, you receive an error.

", - "DescribeInstancePatchStates": "

Retrieves the high-level patch state of one or more instances.

", - "DescribeInstancePatchStatesForPatchGroup": "

Retrieves the high-level patch state for the instances in the specified patch group.

", - "DescribeInstancePatches": "

Retrieves information about the patches on the specified instance and their state relative to the patch baseline being used for the instance.

", - "DescribeInventoryDeletions": "

Describes a specific delete inventory operation.

", - "DescribeMaintenanceWindowExecutionTaskInvocations": "

Retrieves the individual task executions (one per target) for a particular task executed as part of a Maintenance Window execution.

", - "DescribeMaintenanceWindowExecutionTasks": "

For a given Maintenance Window execution, lists the tasks that were executed.

", - "DescribeMaintenanceWindowExecutions": "

Lists the executions of a Maintenance Window. This includes information about when the Maintenance Window was scheduled to be active, and information about tasks registered and run with the Maintenance Window.

", - "DescribeMaintenanceWindowTargets": "

Lists the targets registered with the Maintenance Window.

", - "DescribeMaintenanceWindowTasks": "

Lists the tasks in a Maintenance Window.

", - "DescribeMaintenanceWindows": "

Retrieves the Maintenance Windows in an AWS account.

", - "DescribeParameters": "

Get information about a parameter.

Request results are returned on a best-effort basis. If you specify MaxResults in the request, the response includes information up to the limit specified. The number of items returned, however, can be between zero and the value of MaxResults. If the service reaches an internal limit while processing the results, it stops the operation and returns the matching values up to that point and a NextToken. You can specify the NextToken in a subsequent call to get the next set of results.

", - "DescribePatchBaselines": "

Lists the patch baselines in your AWS account.

", - "DescribePatchGroupState": "

Returns high-level aggregated patch compliance state for a patch group.

", - "DescribePatchGroups": "

Lists all patch groups that have been registered with patch baselines.

", - "GetAutomationExecution": "

Get detailed information about a particular Automation execution.

", - "GetCommandInvocation": "

Returns detailed information about command execution for an invocation or plugin.

", - "GetDefaultPatchBaseline": "

Retrieves the default patch baseline. Note that Systems Manager supports creating multiple default patch baselines. For example, you can create a default patch baseline for each operating system.

If you do not specify an operating system value, the default patch baseline for Windows is returned.

", - "GetDeployablePatchSnapshotForInstance": "

Retrieves the current snapshot for the patch baseline the instance uses. This API is primarily used by the AWS-RunPatchBaseline Systems Manager document.

", - "GetDocument": "

Gets the contents of the specified Systems Manager document.

", - "GetInventory": "

Query inventory information.

", - "GetInventorySchema": "

Return a list of inventory type names for the account, or return a list of attribute names for a specific Inventory item type.

", - "GetMaintenanceWindow": "

Retrieves a Maintenance Window.

", - "GetMaintenanceWindowExecution": "

Retrieves details about a specific task executed as part of a Maintenance Window execution.

", - "GetMaintenanceWindowExecutionTask": "

Retrieves the details about a specific task executed as part of a Maintenance Window execution.

", - "GetMaintenanceWindowExecutionTaskInvocation": "

Retrieves a task invocation. A task invocation is a specific task executing on a specific target. Maintenance Windows report status for all invocations.

", - "GetMaintenanceWindowTask": "

Lists the tasks in a Maintenance Window.

", - "GetParameter": "

Get information about a parameter by using the parameter name.

", - "GetParameterHistory": "

Query a list of all parameters used by the AWS account.

", - "GetParameters": "

Get details of a parameter.

", - "GetParametersByPath": "

Retrieve parameters in a specific hierarchy. For more information, see Working with Systems Manager Parameters.

Request results are returned on a best-effort basis. If you specify MaxResults in the request, the response includes information up to the limit specified. The number of items returned, however, can be between zero and the value of MaxResults. If the service reaches an internal limit while processing the results, it stops the operation and returns the matching values up to that point and a NextToken. You can specify the NextToken in a subsequent call to get the next set of results.

This API action doesn't support filtering by tags.

", - "GetPatchBaseline": "

Retrieves information about a patch baseline.

", - "GetPatchBaselineForPatchGroup": "

Retrieves the patch baseline that should be used for the specified patch group.

", - "ListAssociationVersions": "

Retrieves all versions of an association for a specific association ID.

", - "ListAssociations": "

Lists the associations for the specified Systems Manager document or instance.

", - "ListCommandInvocations": "

An invocation is copy of a command sent to a specific instance. A command can apply to one or more instances. A command invocation applies to one instance. For example, if a user executes SendCommand against three instances, then a command invocation is created for each requested instance ID. ListCommandInvocations provide status about command execution.

", - "ListCommands": "

Lists the commands requested by users of the AWS account.

", - "ListComplianceItems": "

For a specified resource ID, this API action returns a list of compliance statuses for different resource types. Currently, you can only specify one resource ID per call. List results depend on the criteria specified in the filter.

", - "ListComplianceSummaries": "

Returns a summary count of compliant and non-compliant resources for a compliance type. For example, this call can return State Manager associations, patches, or custom compliance types according to the filter criteria that you specify.

", - "ListDocumentVersions": "

List all versions for a document.

", - "ListDocuments": "

Describes one or more of your Systems Manager documents.

", - "ListInventoryEntries": "

A list of inventory items returned by the request.

", - "ListResourceComplianceSummaries": "

Returns a resource-level summary count. The summary includes information about compliant and non-compliant statuses and detailed compliance-item severity counts, according to the filter criteria you specify.

", - "ListResourceDataSync": "

Lists your resource data sync configurations. Includes information about the last time a sync attempted to start, the last sync status, and the last time a sync successfully completed.

The number of sync configurations might be too large to return using a single call to ListResourceDataSync. You can limit the number of sync configurations returned by using the MaxResults parameter. To determine whether there are more sync configurations to list, check the value of NextToken in the output. If there are more sync configurations to list, you can request them by specifying the NextToken returned in the call to the parameter of a subsequent call.

", - "ListTagsForResource": "

Returns a list of the tags assigned to the specified resource.

", - "ModifyDocumentPermission": "

Shares a Systems Manager document publicly or privately. If you share a document privately, you must specify the AWS user account IDs for those people who can use the document. If you share a document publicly, you must specify All as the account ID.

", - "PutComplianceItems": "

Registers a compliance type and other compliance details on a designated resource. This action lets you register custom compliance details with a resource. This call overwrites existing compliance information on the resource, so you must provide a full list of compliance items each time that you send the request.

ComplianceType can be one of the following:

  • ExecutionId: The execution ID when the patch, association, or custom compliance item was applied.

  • ExecutionType: Specify patch, association, or Custom:string.

  • ExecutionTime. The time the patch, association, or custom compliance item was applied to the instance.

  • Id: The patch, association, or custom compliance ID.

  • Title: A title.

  • Status: The status of the compliance item. For example, approved for patches, or Failed for associations.

  • Severity: A patch severity. For example, critical.

  • DocumentName: A SSM document name. For example, AWS-RunPatchBaseline.

  • DocumentVersion: An SSM document version number. For example, 4.

  • Classification: A patch classification. For example, security updates.

  • PatchBaselineId: A patch baseline ID.

  • PatchSeverity: A patch severity. For example, Critical.

  • PatchState: A patch state. For example, InstancesWithFailedPatches.

  • PatchGroup: The name of a patch group.

  • InstalledTime: The time the association, patch, or custom compliance item was applied to the resource. Specify the time by using the following format: yyyy-MM-dd'T'HH:mm:ss'Z'

", - "PutInventory": "

Bulk update custom inventory items on one more instance. The request adds an inventory item, if it doesn't already exist, or updates an inventory item, if it does exist.

", - "PutParameter": "

Add a parameter to the system.

", - "RegisterDefaultPatchBaseline": "

Defines the default patch baseline.

", - "RegisterPatchBaselineForPatchGroup": "

Registers a patch baseline for a patch group.

", - "RegisterTargetWithMaintenanceWindow": "

Registers a target with a Maintenance Window.

", - "RegisterTaskWithMaintenanceWindow": "

Adds a new task to a Maintenance Window.

", - "RemoveTagsFromResource": "

Removes all tags from the specified resource.

", - "SendAutomationSignal": "

Sends a signal to an Automation execution to change the current behavior or status of the execution.

", - "SendCommand": "

Executes commands on one or more managed instances.

", - "StartAutomationExecution": "

Initiates execution of an Automation document.

", - "StopAutomationExecution": "

Stop an Automation that is currently executing.

", - "UpdateAssociation": "

Updates an association. You can update the association name and version, the document version, schedule, parameters, and Amazon S3 output.

", - "UpdateAssociationStatus": "

Updates the status of the Systems Manager document associated with the specified instance.

", - "UpdateDocument": "

The document you want to update.

", - "UpdateDocumentDefaultVersion": "

Set the default version of a document.

", - "UpdateMaintenanceWindow": "

Updates an existing Maintenance Window. Only specified parameters are modified.

", - "UpdateMaintenanceWindowTarget": "

Modifies the target of an existing Maintenance Window. You can't change the target type, but you can change the following:

The target from being an ID target to a Tag target, or a Tag target to an ID target.

IDs for an ID target.

Tags for a Tag target.

Owner.

Name.

Description.

If a parameter is null, then the corresponding field is not modified.

", - "UpdateMaintenanceWindowTask": "

Modifies a task assigned to a Maintenance Window. You can't change the task type, but you can change the following values:

  • TaskARN. For example, you can change a RUN_COMMAND task from AWS-RunPowerShellScript to AWS-RunShellScript.

  • ServiceRoleArn

  • TaskInvocationParameters

  • Priority

  • MaxConcurrency

  • MaxErrors

If a parameter is null, then the corresponding field is not modified. Also, if you set Replace to true, then all fields required by the RegisterTaskWithMaintenanceWindow action are required for this request. Optional fields that aren't specified are set to null.

", - "UpdateManagedInstanceRole": "

Assigns or changes an Amazon Identity and Access Management (IAM) role to the managed instance.

", - "UpdatePatchBaseline": "

Modifies an existing patch baseline. Fields not specified in the request are left unchanged.

For information about valid key and value pairs in PatchFilters for each supported operating system type, see PatchFilter.

" + "version":"2.0", + "service":"AWS Systems Manager

AWS Systems Manager is a collection of capabilities that helps you automate management tasks such as collecting system inventory, applying operating system (OS) patches, automating the creation of Amazon Machine Images (AMIs), and configuring operating systems (OSs) and applications at scale. Systems Manager lets you remotely and securely manage the configuration of your managed instances. A managed instance is any Amazon EC2 instance or on-premises machine in your hybrid environment that has been configured for Systems Manager.

This reference is intended to be used with the AWS Systems Manager User Guide.

To get started, verify prerequisites and configure managed instances. For more information, see Systems Manager Prerequisites in the AWS Systems Manager User Guide.

For information about other API actions you can perform on Amazon EC2 instances, see the Amazon EC2 API Reference. For information about how to use a Query API, see Making API Requests.

", + "operations":{ + "AddTagsToResource":"

Adds or overwrites one or more tags for the specified resource. Tags are metadata that you can assign to your documents, managed instances, Maintenance Windows, Parameter Store parameters, and patch baselines. Tags enable you to categorize your resources in different ways, for example, by purpose, owner, or environment. Each tag consists of a key and an optional value, both of which you define. For example, you could define a set of tags for your account's managed instances that helps you track each instance's owner and stack level. For example: Key=Owner and Value=DbAdmin, SysAdmin, or Dev. Or Key=Stack and Value=Production, Pre-Production, or Test.

Each resource can have a maximum of 50 tags.

We recommend that you devise a set of tag keys that meets your needs for each resource type. Using a consistent set of tag keys makes it easier for you to manage your resources. You can search and filter the resources based on the tags you add. Tags don't have any semantic meaning to Amazon EC2 and are interpreted strictly as a string of characters.

For more information about tags, see Tagging Your Amazon EC2 Resources in the Amazon EC2 User Guide.

", + "CancelCommand":"

Attempts to cancel the command specified by the Command ID. There is no guarantee that the command will be terminated and the underlying process stopped.

", + "CreateActivation":"

Registers your on-premises server or virtual machine with Amazon EC2 so that you can manage these resources using Run Command. An on-premises server or virtual machine that has been registered with EC2 is called a managed instance. For more information about activations, see Setting Up Systems Manager in Hybrid Environments.

", + "CreateAssociation":"

Associates the specified Systems Manager document with the specified instances or targets.

When you associate a document with one or more instances using instance IDs or tags, the SSM Agent running on the instance processes the document and configures the instance as specified.

If you associate a document with an instance that already has an associated document, the system throws the AssociationAlreadyExists exception.

", + "CreateAssociationBatch":"

Associates the specified Systems Manager document with the specified instances or targets.

When you associate a document with one or more instances using instance IDs or tags, the SSM Agent running on the instance processes the document and configures the instance as specified.

If you associate a document with an instance that already has an associated document, the system throws the AssociationAlreadyExists exception.

", + "CreateDocument":"

Creates a Systems Manager document.

After you create a document, you can use CreateAssociation to associate it with one or more running instances.

", + "CreateMaintenanceWindow":"

Creates a new Maintenance Window.

", + "CreatePatchBaseline":"

Creates a patch baseline.

For information about valid key and value pairs in PatchFilters for each supported operating system type, see PatchFilter.

", + "CreateResourceDataSync":"

Creates a resource data sync configuration to a single bucket in Amazon S3. This is an asynchronous operation that returns immediately. After a successful initial sync is completed, the system continuously syncs data to the Amazon S3 bucket. To check the status of the sync, use the ListResourceDataSync.

By default, data is not encrypted in Amazon S3. We strongly recommend that you enable encryption in Amazon S3 to ensure secure data storage. We also recommend that you secure access to the Amazon S3 bucket by creating a restrictive bucket policy. To view an example of a restrictive Amazon S3 bucket policy for Resource Data Sync, see Create a Resource Data Sync for Inventory in the AWS Systems Manager User Guide.

", + "DeleteActivation":"

Deletes an activation. You are not required to delete an activation. If you delete an activation, you can no longer use it to register additional managed instances. Deleting an activation does not de-register managed instances. You must manually de-register managed instances.

", + "DeleteAssociation":"

Disassociates the specified Systems Manager document from the specified instance.

When you disassociate a document from an instance, it does not change the configuration of the instance. To change the configuration state of an instance after you disassociate a document, you must create a new document with the desired configuration and associate it with the instance.

", + "DeleteDocument":"

Deletes the Systems Manager document and all instance associations to the document.

Before you delete the document, we recommend that you use DeleteAssociation to disassociate all instances that are associated with the document.

", + "DeleteInventory":"

Delete a custom inventory type, or the data associated with a custom Inventory type. Deleting a custom inventory type is also referred to as deleting a custom inventory schema.

", + "DeleteMaintenanceWindow":"

Deletes a Maintenance Window.

", + "DeleteParameter":"

Delete a parameter from the system.

", + "DeleteParameters":"

Delete a list of parameters. This API is used to delete parameters by using the Amazon EC2 console.

", + "DeletePatchBaseline":"

Deletes a patch baseline.

", + "DeleteResourceDataSync":"

Deletes a Resource Data Sync configuration. After the configuration is deleted, changes to inventory data on managed instances are no longer synced with the target Amazon S3 bucket. Deleting a sync configuration does not delete data in the target Amazon S3 bucket.

", + "DeregisterManagedInstance":"

Removes the server or virtual machine from the list of registered servers. You can reregister the instance again at any time. If you don't plan to use Run Command on the server, we suggest uninstalling SSM Agent first.

", + "DeregisterPatchBaselineForPatchGroup":"

Removes a patch group from a patch baseline.

", + "DeregisterTargetFromMaintenanceWindow":"

Removes a target from a Maintenance Window.

", + "DeregisterTaskFromMaintenanceWindow":"

Removes a task from a Maintenance Window.

", + "DescribeActivations":"

Details about the activation, including: the date and time the activation was created, the expiration date, the IAM role assigned to the instances in the activation, and the number of instances activated by this registration.

", + "DescribeAssociation":"

Describes the association for the specified target or instance. If you created the association by using the Targets parameter, then you must retrieve the association by using the association ID. If you created the association by specifying an instance ID and a Systems Manager document, then you retrieve the association by specifying the document name and the instance ID.

", + "DescribeAutomationExecutions":"

Provides details about all active and terminated Automation executions.

", + "DescribeAutomationStepExecutions":"

Information about all active and terminated step executions in an Automation workflow.

", + "DescribeAvailablePatches":"

Lists all patches that could possibly be included in a patch baseline.

", + "DescribeDocument":"

Describes the specified Systems Manager document.

", + "DescribeDocumentPermission":"

Describes the permissions for a Systems Manager document. If you created the document, you are the owner. If a document is shared, it can either be shared privately (by specifying a user's AWS account ID) or publicly (All).

", + "DescribeEffectiveInstanceAssociations":"

All associations for the instance(s).

", + "DescribeEffectivePatchesForPatchBaseline":"

Retrieves the current effective patches (the patch and the approval state) for the specified patch baseline. Note that this API applies only to Windows patch baselines.

", + "DescribeInstanceAssociationsStatus":"

The status of the associations for the instance(s).

", + "DescribeInstanceInformation":"

Describes one or more of your instances. You can use this to get information about instances like the operating system platform, the SSM Agent version (Linux), status etc. If you specify one or more instance IDs, it returns information for those instances. If you do not specify instance IDs, it returns information for all your instances. If you specify an instance ID that is not valid or an instance that you do not own, you receive an error.

The IamRole field for this API action is the Amazon Identity and Access Management (IAM) role assigned to on-premises instances. This call does not return the IAM role for Amazon EC2 instances.

", + "DescribeInstancePatchStates":"

Retrieves the high-level patch state of one or more instances.

", + "DescribeInstancePatchStatesForPatchGroup":"

Retrieves the high-level patch state for the instances in the specified patch group.

", + "DescribeInstancePatches":"

Retrieves information about the patches on the specified instance and their state relative to the patch baseline being used for the instance.

", + "DescribeInventoryDeletions":"

Describes a specific delete inventory operation.

", + "DescribeMaintenanceWindowExecutionTaskInvocations":"

Retrieves the individual task executions (one per target) for a particular task executed as part of a Maintenance Window execution.

", + "DescribeMaintenanceWindowExecutionTasks":"

For a given Maintenance Window execution, lists the tasks that were executed.

", + "DescribeMaintenanceWindowExecutions":"

Lists the executions of a Maintenance Window. This includes information about when the Maintenance Window was scheduled to be active, and information about tasks registered and run with the Maintenance Window.

", + "DescribeMaintenanceWindowTargets":"

Lists the targets registered with the Maintenance Window.

", + "DescribeMaintenanceWindowTasks":"

Lists the tasks in a Maintenance Window.

", + "DescribeMaintenanceWindows":"

Retrieves the Maintenance Windows in an AWS account.

", + "DescribeParameters":"

Get information about a parameter.

Request results are returned on a best-effort basis. If you specify MaxResults in the request, the response includes information up to the limit specified. The number of items returned, however, can be between zero and the value of MaxResults. If the service reaches an internal limit while processing the results, it stops the operation and returns the matching values up to that point and a NextToken. You can specify the NextToken in a subsequent call to get the next set of results.

", + "DescribePatchBaselines":"

Lists the patch baselines in your AWS account.

", + "DescribePatchGroupState":"

Returns high-level aggregated patch compliance state for a patch group.

", + "DescribePatchGroups":"

Lists all patch groups that have been registered with patch baselines.

", + "GetAutomationExecution":"

Get detailed information about a particular Automation execution.

", + "GetCommandInvocation":"

Returns detailed information about command execution for an invocation or plugin.

", + "GetDefaultPatchBaseline":"

Retrieves the default patch baseline. Note that Systems Manager supports creating multiple default patch baselines. For example, you can create a default patch baseline for each operating system.

If you do not specify an operating system value, the default patch baseline for Windows is returned.

", + "GetDeployablePatchSnapshotForInstance":"

Retrieves the current snapshot for the patch baseline the instance uses. This API is primarily used by the AWS-RunPatchBaseline Systems Manager document.

", + "GetDocument":"

Gets the contents of the specified Systems Manager document.

", + "GetInventory":"

Query inventory information.

", + "GetInventorySchema":"

Return a list of inventory type names for the account, or return a list of attribute names for a specific Inventory item type.

", + "GetMaintenanceWindow":"

Retrieves a Maintenance Window.

", + "GetMaintenanceWindowExecution":"

Retrieves details about a specific task executed as part of a Maintenance Window execution.

", + "GetMaintenanceWindowExecutionTask":"

Retrieves the details about a specific task executed as part of a Maintenance Window execution.

", + "GetMaintenanceWindowExecutionTaskInvocation":"

Retrieves a task invocation. A task invocation is a specific task executing on a specific target. Maintenance Windows report status for all invocations.

", + "GetMaintenanceWindowTask":"

Lists the tasks in a Maintenance Window.

", + "GetParameter":"

Get information about a parameter by using the parameter name. Don't confuse this API action with the GetParameters API action.

", + "GetParameterHistory":"

Query a list of all parameters used by the AWS account.

", + "GetParameters":"

Get details of a parameter. Don't confuse this API action with the GetParameter API action.

", + "GetParametersByPath":"

Retrieve parameters in a specific hierarchy. For more information, see Working with Systems Manager Parameters in the AWS Systems Manager User Guide.

Request results are returned on a best-effort basis. If you specify MaxResults in the request, the response includes information up to the limit specified. The number of items returned, however, can be between zero and the value of MaxResults. If the service reaches an internal limit while processing the results, it stops the operation and returns the matching values up to that point and a NextToken. You can specify the NextToken in a subsequent call to get the next set of results.

This API action doesn't support filtering by tags.

", + "GetPatchBaseline":"

Retrieves information about a patch baseline.

", + "GetPatchBaselineForPatchGroup":"

Retrieves the patch baseline that should be used for the specified patch group.

", + "ListAssociationVersions":"

Retrieves all versions of an association for a specific association ID.

", + "ListAssociations":"

Lists the associations for the specified Systems Manager document or instance.

", + "ListCommandInvocations":"

An invocation is copy of a command sent to a specific instance. A command can apply to one or more instances. A command invocation applies to one instance. For example, if a user executes SendCommand against three instances, then a command invocation is created for each requested instance ID. ListCommandInvocations provide status about command execution.

", + "ListCommands":"

Lists the commands requested by users of the AWS account.

", + "ListComplianceItems":"

For a specified resource ID, this API action returns a list of compliance statuses for different resource types. Currently, you can only specify one resource ID per call. List results depend on the criteria specified in the filter.

", + "ListComplianceSummaries":"

Returns a summary count of compliant and non-compliant resources for a compliance type. For example, this call can return State Manager associations, patches, or custom compliance types according to the filter criteria that you specify.

", + "ListDocumentVersions":"

List all versions for a document.

", + "ListDocuments":"

Describes one or more of your Systems Manager documents.

", + "ListInventoryEntries":"

A list of inventory items returned by the request.

", + "ListResourceComplianceSummaries":"

Returns a resource-level summary count. The summary includes information about compliant and non-compliant statuses and detailed compliance-item severity counts, according to the filter criteria you specify.

", + "ListResourceDataSync":"

Lists your resource data sync configurations. Includes information about the last time a sync attempted to start, the last sync status, and the last time a sync successfully completed.

The number of sync configurations might be too large to return using a single call to ListResourceDataSync. You can limit the number of sync configurations returned by using the MaxResults parameter. To determine whether there are more sync configurations to list, check the value of NextToken in the output. If there are more sync configurations to list, you can request them by specifying the NextToken returned in the call to the parameter of a subsequent call.

", + "ListTagsForResource":"

Returns a list of the tags assigned to the specified resource.

", + "ModifyDocumentPermission":"

Shares a Systems Manager document publicly or privately. If you share a document privately, you must specify the AWS user account IDs for those people who can use the document. If you share a document publicly, you must specify All as the account ID.

", + "PutComplianceItems":"

Registers a compliance type and other compliance details on a designated resource. This action lets you register custom compliance details with a resource. This call overwrites existing compliance information on the resource, so you must provide a full list of compliance items each time that you send the request.

ComplianceType can be one of the following:

  • ExecutionId: The execution ID when the patch, association, or custom compliance item was applied.

  • ExecutionType: Specify patch, association, or Custom:string.

  • ExecutionTime. The time the patch, association, or custom compliance item was applied to the instance.

  • Id: The patch, association, or custom compliance ID.

  • Title: A title.

  • Status: The status of the compliance item. For example, approved for patches, or Failed for associations.

  • Severity: A patch severity. For example, critical.

  • DocumentName: A SSM document name. For example, AWS-RunPatchBaseline.

  • DocumentVersion: An SSM document version number. For example, 4.

  • Classification: A patch classification. For example, security updates.

  • PatchBaselineId: A patch baseline ID.

  • PatchSeverity: A patch severity. For example, Critical.

  • PatchState: A patch state. For example, InstancesWithFailedPatches.

  • PatchGroup: The name of a patch group.

  • InstalledTime: The time the association, patch, or custom compliance item was applied to the resource. Specify the time by using the following format: yyyy-MM-dd'T'HH:mm:ss'Z'

", + "PutInventory":"

Bulk update custom inventory items on one more instance. The request adds an inventory item, if it doesn't already exist, or updates an inventory item, if it does exist.

", + "PutParameter":"

Add a parameter to the system.

", + "RegisterDefaultPatchBaseline":"

Defines the default patch baseline.

", + "RegisterPatchBaselineForPatchGroup":"

Registers a patch baseline for a patch group.

", + "RegisterTargetWithMaintenanceWindow":"

Registers a target with a Maintenance Window.

", + "RegisterTaskWithMaintenanceWindow":"

Adds a new task to a Maintenance Window.

", + "RemoveTagsFromResource":"

Removes all tags from the specified resource.

", + "SendAutomationSignal":"

Sends a signal to an Automation execution to change the current behavior or status of the execution.

", + "SendCommand":"

Executes commands on one or more managed instances.

", + "StartAutomationExecution":"

Initiates execution of an Automation document.

", + "StopAutomationExecution":"

Stop an Automation that is currently executing.

", + "UpdateAssociation":"

Updates an association. You can update the association name and version, the document version, schedule, parameters, and Amazon S3 output.

", + "UpdateAssociationStatus":"

Updates the status of the Systems Manager document associated with the specified instance.

", + "UpdateDocument":"

The document you want to update.

", + "UpdateDocumentDefaultVersion":"

Set the default version of a document.

", + "UpdateMaintenanceWindow":"

Updates an existing Maintenance Window. Only specified parameters are modified.

", + "UpdateMaintenanceWindowTarget":"

Modifies the target of an existing Maintenance Window. You can't change the target type, but you can change the following:

The target from being an ID target to a Tag target, or a Tag target to an ID target.

IDs for an ID target.

Tags for a Tag target.

Owner.

Name.

Description.

If a parameter is null, then the corresponding field is not modified.

", + "UpdateMaintenanceWindowTask":"

Modifies a task assigned to a Maintenance Window. You can't change the task type, but you can change the following values:

  • TaskARN. For example, you can change a RUN_COMMAND task from AWS-RunPowerShellScript to AWS-RunShellScript.

  • ServiceRoleArn

  • TaskInvocationParameters

  • Priority

  • MaxConcurrency

  • MaxErrors

If a parameter is null, then the corresponding field is not modified. Also, if you set Replace to true, then all fields required by the RegisterTaskWithMaintenanceWindow action are required for this request. Optional fields that aren't specified are set to null.

", + "UpdateManagedInstanceRole":"

Assigns or changes an Amazon Identity and Access Management (IAM) role to the managed instance.

", + "UpdatePatchBaseline":"

Modifies an existing patch baseline. Fields not specified in the request are left unchanged.

For information about valid key and value pairs in PatchFilters for each supported operating system type, see PatchFilter.

" }, - "shapes": { - "AccountId": { - "base": null, - "refs": { - "AccountIdList$member": null + "shapes":{ + "AccountId":{ + "base":null, + "refs":{ + "AccountIdList$member":null } }, - "AccountIdList": { - "base": null, - "refs": { - "DescribeDocumentPermissionResponse$AccountIds": "

The account IDs that have permission to use this document. The ID can be either an AWS account or All.

", - "ModifyDocumentPermissionRequest$AccountIdsToAdd": "

The AWS user accounts that should have access to the document. The account IDs can either be a group of account IDs or All.

", - "ModifyDocumentPermissionRequest$AccountIdsToRemove": "

The AWS user accounts that should no longer have access to the document. The AWS user account can either be a group of account IDs or All. This action has a higher priority than AccountIdsToAdd. If you specify an account ID to add and the same ID to remove, the system removes access to the document.

" + "AccountIdList":{ + "base":null, + "refs":{ + "DescribeDocumentPermissionResponse$AccountIds":"

The account IDs that have permission to use this document. The ID can be either an AWS account or All.

", + "ModifyDocumentPermissionRequest$AccountIdsToAdd":"

The AWS user accounts that should have access to the document. The account IDs can either be a group of account IDs or All.

", + "ModifyDocumentPermissionRequest$AccountIdsToRemove":"

The AWS user accounts that should no longer have access to the document. The AWS user account can either be a group of account IDs or All. This action has a higher priority than AccountIdsToAdd. If you specify an account ID to add and the same ID to remove, the system removes access to the document.

" } }, - "Activation": { - "base": "

An activation registers one or more on-premises servers or virtual machines (VMs) with AWS so that you can configure those servers or VMs using Run Command. A server or VM that has been registered with AWS is called a managed instance.

", - "refs": { - "ActivationList$member": null + "Activation":{ + "base":"

An activation registers one or more on-premises servers or virtual machines (VMs) with AWS so that you can configure those servers or VMs using Run Command. A server or VM that has been registered with AWS is called a managed instance.

", + "refs":{ + "ActivationList$member":null } }, - "ActivationCode": { - "base": null, - "refs": { - "CreateActivationResult$ActivationCode": "

The code the system generates when it processes the activation. The activation code functions like a password to validate the activation ID.

" + "ActivationCode":{ + "base":null, + "refs":{ + "CreateActivationResult$ActivationCode":"

The code the system generates when it processes the activation. The activation code functions like a password to validate the activation ID.

" } }, - "ActivationDescription": { - "base": null, - "refs": { - "Activation$Description": "

A user defined description of the activation.

", - "CreateActivationRequest$Description": "

A user-defined description of the resource that you want to register with Amazon EC2.

Do not enter personally identifiable information in this field.

" + "ActivationDescription":{ + "base":null, + "refs":{ + "Activation$Description":"

A user defined description of the activation.

", + "CreateActivationRequest$Description":"

A user-defined description of the resource that you want to register with Amazon EC2.

Do not enter personally identifiable information in this field.

" } }, - "ActivationId": { - "base": null, - "refs": { - "Activation$ActivationId": "

The ID created by Systems Manager when you submitted the activation.

", - "CreateActivationResult$ActivationId": "

The ID number generated by the system when it processed the activation. The activation ID functions like a user name.

", - "DeleteActivationRequest$ActivationId": "

The ID of the activation that you want to delete.

", - "InstanceInformation$ActivationId": "

The activation ID created by Systems Manager when the server or VM was registered.

" + "ActivationId":{ + "base":null, + "refs":{ + "Activation$ActivationId":"

The ID created by Systems Manager when you submitted the activation.

", + "CreateActivationResult$ActivationId":"

The ID number generated by the system when it processed the activation. The activation ID functions like a user name.

", + "DeleteActivationRequest$ActivationId":"

The ID of the activation that you want to delete.

", + "InstanceInformation$ActivationId":"

The activation ID created by Systems Manager when the server or VM was registered.

" } }, - "ActivationList": { - "base": null, - "refs": { - "DescribeActivationsResult$ActivationList": "

A list of activations for your AWS account.

" + "ActivationList":{ + "base":null, + "refs":{ + "DescribeActivationsResult$ActivationList":"

A list of activations for your AWS account.

" } }, - "AddTagsToResourceRequest": { - "base": null, - "refs": { + "AddTagsToResourceRequest":{ + "base":null, + "refs":{ } }, - "AddTagsToResourceResult": { - "base": null, - "refs": { + "AddTagsToResourceResult":{ + "base":null, + "refs":{ } }, - "AgentErrorCode": { - "base": null, - "refs": { - "InstanceAssociationStatusInfo$ErrorCode": "

An error code returned by the request to create the association.

" + "AgentErrorCode":{ + "base":null, + "refs":{ + "InstanceAssociationStatusInfo$ErrorCode":"

An error code returned by the request to create the association.

" } }, - "AggregatorSchemaOnly": { - "base": null, - "refs": { - "GetInventorySchemaRequest$Aggregator": "

Returns inventory schemas that support aggregation. For example, this call returns the AWS:InstanceInformation type, because it supports aggregation based on the PlatformName, PlatformType, and PlatformVersion attributes.

" + "AggregatorSchemaOnly":{ + "base":null, + "refs":{ + "GetInventorySchemaRequest$Aggregator":"

Returns inventory schemas that support aggregation. For example, this call returns the AWS:InstanceInformation type, because it supports aggregation based on the PlatformName, PlatformType, and PlatformVersion attributes.

" } }, - "AllowedPattern": { - "base": null, - "refs": { - "ParameterHistory$AllowedPattern": "

Parameter names can include the following letters and symbols.

a-zA-Z0-9_.-

", - "ParameterMetadata$AllowedPattern": "

A parameter name can include only the following letters and symbols.

a-zA-Z0-9_.-

", - "PutParameterRequest$AllowedPattern": "

A regular expression used to validate the parameter value. For example, for String types with values restricted to numbers, you can specify the following: AllowedPattern=^\\d+$

" + "AllowedPattern":{ + "base":null, + "refs":{ + "ParameterHistory$AllowedPattern":"

Parameter names can include the following letters and symbols.

a-zA-Z0-9_.-

", + "ParameterMetadata$AllowedPattern":"

A parameter name can include only the following letters and symbols.

a-zA-Z0-9_.-

", + "PutParameterRequest$AllowedPattern":"

A regular expression used to validate the parameter value. For example, for String types with values restricted to numbers, you can specify the following: AllowedPattern=^\\d+$

" } }, - "AlreadyExistsException": { - "base": "

Error returned if an attempt is made to register a patch group with a patch baseline that is already registered with a different patch baseline.

", - "refs": { + "AlreadyExistsException":{ + "base":"

Error returned if an attempt is made to register a patch group with a patch baseline that is already registered with a different patch baseline.

", + "refs":{ } }, - "ApproveAfterDays": { - "base": null, - "refs": { - "PatchRule$ApproveAfterDays": "

The number of days after the release date of each patch matched by the rule the patch is marked as approved in the patch baseline.

" + "ApproveAfterDays":{ + "base":null, + "refs":{ + "PatchRule$ApproveAfterDays":"

The number of days after the release date of each patch matched by the rule that the patch is marked as approved in the patch baseline. For example, a value of 7 means that patches are approved seven days after they are released.

" } }, - "AssociatedInstances": { - "base": "

You must disassociate a document from all instances before you can delete it.

", - "refs": { + "AssociatedInstances":{ + "base":"

You must disassociate a document from all instances before you can delete it.

", + "refs":{ } }, - "Association": { - "base": "

Describes an association of a Systems Manager document and an instance.

", - "refs": { - "AssociationList$member": null + "Association":{ + "base":"

Describes an association of a Systems Manager document and an instance.

", + "refs":{ + "AssociationList$member":null } }, - "AssociationAlreadyExists": { - "base": "

The specified association already exists.

", - "refs": { + "AssociationAlreadyExists":{ + "base":"

The specified association already exists.

", + "refs":{ } }, - "AssociationDescription": { - "base": "

Describes the parameters for a document.

", - "refs": { - "AssociationDescriptionList$member": null, - "CreateAssociationResult$AssociationDescription": "

Information about the association.

", - "DescribeAssociationResult$AssociationDescription": "

Information about the association.

", - "UpdateAssociationResult$AssociationDescription": "

The description of the association that was updated.

", - "UpdateAssociationStatusResult$AssociationDescription": "

Information about the association.

" + "AssociationDescription":{ + "base":"

Describes the parameters for a document.

", + "refs":{ + "AssociationDescriptionList$member":null, + "CreateAssociationResult$AssociationDescription":"

Information about the association.

", + "DescribeAssociationResult$AssociationDescription":"

Information about the association.

", + "UpdateAssociationResult$AssociationDescription":"

The description of the association that was updated.

", + "UpdateAssociationStatusResult$AssociationDescription":"

Information about the association.

" } }, - "AssociationDescriptionList": { - "base": null, - "refs": { - "CreateAssociationBatchResult$Successful": "

Information about the associations that succeeded.

" + "AssociationDescriptionList":{ + "base":null, + "refs":{ + "CreateAssociationBatchResult$Successful":"

Information about the associations that succeeded.

" } }, - "AssociationDoesNotExist": { - "base": "

The specified association does not exist.

", - "refs": { + "AssociationDoesNotExist":{ + "base":"

The specified association does not exist.

", + "refs":{ } }, - "AssociationFilter": { - "base": "

Describes a filter.

", - "refs": { - "AssociationFilterList$member": null + "AssociationFilter":{ + "base":"

Describes a filter.

", + "refs":{ + "AssociationFilterList$member":null } }, - "AssociationFilterKey": { - "base": null, - "refs": { - "AssociationFilter$key": "

The name of the filter.

" + "AssociationFilterKey":{ + "base":null, + "refs":{ + "AssociationFilter$key":"

The name of the filter.

" } }, - "AssociationFilterList": { - "base": null, - "refs": { - "ListAssociationsRequest$AssociationFilterList": "

One or more filters. Use a filter to return a more specific list of results.

" + "AssociationFilterList":{ + "base":null, + "refs":{ + "ListAssociationsRequest$AssociationFilterList":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "AssociationFilterValue": { - "base": null, - "refs": { - "AssociationFilter$value": "

The filter value.

" + "AssociationFilterValue":{ + "base":null, + "refs":{ + "AssociationFilter$value":"

The filter value.

" } }, - "AssociationId": { - "base": null, - "refs": { - "Association$AssociationId": "

The ID created by the system when you create an association. An association is a binding between a document and a set of targets with a schedule.

", - "AssociationDescription$AssociationId": "

The association ID.

", - "AssociationVersionInfo$AssociationId": "

The ID created by the system when the association was created.

", - "DeleteAssociationRequest$AssociationId": "

The association ID that you want to delete.

", - "DescribeAssociationRequest$AssociationId": "

The association ID for which you want information.

", - "InstanceAssociation$AssociationId": "

The association ID.

", - "InstanceAssociationStatusInfo$AssociationId": "

The association ID.

", - "ListAssociationVersionsRequest$AssociationId": "

The association ID for which you want to view all versions.

", - "UpdateAssociationRequest$AssociationId": "

The ID of the association you want to update.

" + "AssociationId":{ + "base":null, + "refs":{ + "Association$AssociationId":"

The ID created by the system when you create an association. An association is a binding between a document and a set of targets with a schedule.

", + "AssociationDescription$AssociationId":"

The association ID.

", + "AssociationVersionInfo$AssociationId":"

The ID created by the system when the association was created.

", + "DeleteAssociationRequest$AssociationId":"

The association ID that you want to delete.

", + "DescribeAssociationRequest$AssociationId":"

The association ID for which you want information.

", + "InstanceAssociation$AssociationId":"

The association ID.

", + "InstanceAssociationStatusInfo$AssociationId":"

The association ID.

", + "ListAssociationVersionsRequest$AssociationId":"

The association ID for which you want to view all versions.

", + "UpdateAssociationRequest$AssociationId":"

The ID of the association you want to update.

" } }, - "AssociationLimitExceeded": { - "base": "

You can have at most 2,000 active associations.

", - "refs": { + "AssociationLimitExceeded":{ + "base":"

You can have at most 2,000 active associations.

", + "refs":{ } }, - "AssociationList": { - "base": null, - "refs": { - "ListAssociationsResult$Associations": "

The associations.

" + "AssociationList":{ + "base":null, + "refs":{ + "ListAssociationsResult$Associations":"

The associations.

" } }, - "AssociationName": { - "base": null, - "refs": { - "Association$AssociationName": "

The association name.

", - "AssociationDescription$AssociationName": "

The association name.

", - "AssociationVersionInfo$AssociationName": "

The name specified for the association version when the association version was created.

", - "CreateAssociationBatchRequestEntry$AssociationName": "

Specify a descriptive name for the association.

", - "CreateAssociationRequest$AssociationName": "

Specify a descriptive name for the association.

", - "InstanceAssociationStatusInfo$AssociationName": "

The name of the association applied to the instance.

", - "UpdateAssociationRequest$AssociationName": "

The name of the association that you want to update.

" + "AssociationName":{ + "base":null, + "refs":{ + "Association$AssociationName":"

The association name.

", + "AssociationDescription$AssociationName":"

The association name.

", + "AssociationVersionInfo$AssociationName":"

The name specified for the association version when the association version was created.

", + "CreateAssociationBatchRequestEntry$AssociationName":"

Specify a descriptive name for the association.

", + "CreateAssociationRequest$AssociationName":"

Specify a descriptive name for the association.

", + "InstanceAssociationStatusInfo$AssociationName":"

The name of the association applied to the instance.

", + "UpdateAssociationRequest$AssociationName":"

The name of the association that you want to update.

" } }, - "AssociationOverview": { - "base": "

Information about the association.

", - "refs": { - "Association$Overview": "

Information about the association.

", - "AssociationDescription$Overview": "

Information about the association.

" + "AssociationOverview":{ + "base":"

Information about the association.

", + "refs":{ + "Association$Overview":"

Information about the association.

", + "AssociationDescription$Overview":"

Information about the association.

" } }, - "AssociationStatus": { - "base": "

Describes an association status.

", - "refs": { - "AssociationDescription$Status": "

The association status.

", - "UpdateAssociationStatusRequest$AssociationStatus": "

The association status.

" + "AssociationStatus":{ + "base":"

Describes an association status.

", + "refs":{ + "AssociationDescription$Status":"

The association status.

", + "UpdateAssociationStatusRequest$AssociationStatus":"

The association status.

" } }, - "AssociationStatusAggregatedCount": { - "base": null, - "refs": { - "AssociationOverview$AssociationStatusAggregatedCount": "

Returns the number of targets for the association status. For example, if you created an association with two instances, and one of them was successful, this would return the count of instances by status.

" + "AssociationStatusAggregatedCount":{ + "base":null, + "refs":{ + "AssociationOverview$AssociationStatusAggregatedCount":"

Returns the number of targets for the association status. For example, if you created an association with two instances, and one of them was successful, this would return the count of instances by status.

" } }, - "AssociationStatusName": { - "base": null, - "refs": { - "AssociationStatus$Name": "

The status.

" + "AssociationStatusName":{ + "base":null, + "refs":{ + "AssociationStatus$Name":"

The status.

" } }, - "AssociationVersion": { - "base": null, - "refs": { - "Association$AssociationVersion": "

The association version.

", - "AssociationDescription$AssociationVersion": "

The association version.

", - "AssociationVersionInfo$AssociationVersion": "

The association version.

", - "DescribeAssociationRequest$AssociationVersion": "

Specify the association version to retrieve. To view the latest version, either specify $LATEST for this parameter, or omit this parameter. To view a list of all associations for an instance, use ListInstanceAssociations. To get a list of versions for a specific association, use ListAssociationVersions.

", - "InstanceAssociation$AssociationVersion": "

Version information for the association on the instance.

", - "InstanceAssociationStatusInfo$AssociationVersion": "

The version of the association applied to the instance.

", - "UpdateAssociationRequest$AssociationVersion": "

This parameter is provided for concurrency control purposes. You must specify the latest association version in the service. If you want to ensure that this request succeeds, either specify $LATEST, or omit this parameter.

" + "AssociationVersion":{ + "base":null, + "refs":{ + "Association$AssociationVersion":"

The association version.

", + "AssociationDescription$AssociationVersion":"

The association version.

", + "AssociationVersionInfo$AssociationVersion":"

The association version.

", + "DescribeAssociationRequest$AssociationVersion":"

Specify the association version to retrieve. To view the latest version, either specify $LATEST for this parameter, or omit this parameter. To view a list of all associations for an instance, use ListInstanceAssociations. To get a list of versions for a specific association, use ListAssociationVersions.

", + "InstanceAssociation$AssociationVersion":"

Version information for the association on the instance.

", + "InstanceAssociationStatusInfo$AssociationVersion":"

The version of the association applied to the instance.

", + "UpdateAssociationRequest$AssociationVersion":"

This parameter is provided for concurrency control purposes. You must specify the latest association version in the service. If you want to ensure that this request succeeds, either specify $LATEST, or omit this parameter.

" } }, - "AssociationVersionInfo": { - "base": "

Information about the association version.

", - "refs": { - "AssociationVersionList$member": null + "AssociationVersionInfo":{ + "base":"

Information about the association version.

", + "refs":{ + "AssociationVersionList$member":null } }, - "AssociationVersionLimitExceeded": { - "base": "

You have reached the maximum number versions allowed for an association. Each association has a limit of 1,000 versions.

", - "refs": { + "AssociationVersionLimitExceeded":{ + "base":"

You have reached the maximum number versions allowed for an association. Each association has a limit of 1,000 versions.

", + "refs":{ } }, - "AssociationVersionList": { - "base": null, - "refs": { - "ListAssociationVersionsResult$AssociationVersions": "

Information about all versions of the association for the specified association ID.

" + "AssociationVersionList":{ + "base":null, + "refs":{ + "ListAssociationVersionsResult$AssociationVersions":"

Information about all versions of the association for the specified association ID.

" } }, - "AttributeName": { - "base": null, - "refs": { - "ComplianceItemDetails$key": null, - "InventoryItemContentContext$key": null, - "InventoryItemEntry$key": null + "AttributeName":{ + "base":null, + "refs":{ + "ComplianceItemDetails$key":null, + "InventoryItemContentContext$key":null, + "InventoryItemEntry$key":null } }, - "AttributeValue": { - "base": null, - "refs": { - "ComplianceItemDetails$value": null, - "InventoryItemContentContext$value": null, - "InventoryItemEntry$value": null + "AttributeValue":{ + "base":null, + "refs":{ + "ComplianceItemDetails$value":null, + "InventoryItemContentContext$value":null, + "InventoryItemEntry$value":null } }, - "AutomationActionName": { - "base": null, - "refs": { - "StepExecution$Action": "

The action this step performs. The action determines the behavior of the step.

" + "AutomationActionName":{ + "base":null, + "refs":{ + "StepExecution$Action":"

The action this step performs. The action determines the behavior of the step.

" } }, - "AutomationDefinitionNotFoundException": { - "base": "

An Automation document with the specified name could not be found.

", - "refs": { + "AutomationDefinitionNotFoundException":{ + "base":"

An Automation document with the specified name could not be found.

", + "refs":{ } }, - "AutomationDefinitionVersionNotFoundException": { - "base": "

An Automation document with the specified name and version could not be found.

", - "refs": { + "AutomationDefinitionVersionNotFoundException":{ + "base":"

An Automation document with the specified name and version could not be found.

", + "refs":{ } }, - "AutomationExecution": { - "base": "

Detailed information about the current state of an individual Automation execution.

", - "refs": { - "GetAutomationExecutionResult$AutomationExecution": "

Detailed information about the current state of an automation execution.

" + "AutomationExecution":{ + "base":"

Detailed information about the current state of an individual Automation execution.

", + "refs":{ + "GetAutomationExecutionResult$AutomationExecution":"

Detailed information about the current state of an automation execution.

" } }, - "AutomationExecutionFilter": { - "base": "

A filter used to match specific automation executions. This is used to limit the scope of Automation execution information returned.

", - "refs": { - "AutomationExecutionFilterList$member": null + "AutomationExecutionFilter":{ + "base":"

A filter used to match specific automation executions. This is used to limit the scope of Automation execution information returned.

", + "refs":{ + "AutomationExecutionFilterList$member":null } }, - "AutomationExecutionFilterKey": { - "base": null, - "refs": { - "AutomationExecutionFilter$Key": "

One or more keys to limit the results. Valid filter keys include the following: DocumentNamePrefix, ExecutionStatus, ExecutionId, ParentExecutionId, CurrentAction, StartTimeBefore, StartTimeAfter.

" + "AutomationExecutionFilterKey":{ + "base":null, + "refs":{ + "AutomationExecutionFilter$Key":"

One or more keys to limit the results. Valid filter keys include the following: DocumentNamePrefix, ExecutionStatus, ExecutionId, ParentExecutionId, CurrentAction, StartTimeBefore, StartTimeAfter.

" } }, - "AutomationExecutionFilterList": { - "base": null, - "refs": { - "DescribeAutomationExecutionsRequest$Filters": "

Filters used to limit the scope of executions that are requested.

" + "AutomationExecutionFilterList":{ + "base":null, + "refs":{ + "DescribeAutomationExecutionsRequest$Filters":"

Filters used to limit the scope of executions that are requested.

" } }, - "AutomationExecutionFilterValue": { - "base": null, - "refs": { - "AutomationExecutionFilterValueList$member": null + "AutomationExecutionFilterValue":{ + "base":null, + "refs":{ + "AutomationExecutionFilterValueList$member":null } }, - "AutomationExecutionFilterValueList": { - "base": null, - "refs": { - "AutomationExecutionFilter$Values": "

The values used to limit the execution information associated with the filter's key.

" + "AutomationExecutionFilterValueList":{ + "base":null, + "refs":{ + "AutomationExecutionFilter$Values":"

The values used to limit the execution information associated with the filter's key.

" } }, - "AutomationExecutionId": { - "base": null, - "refs": { - "AutomationExecution$AutomationExecutionId": "

The execution ID.

", - "AutomationExecution$ParentAutomationExecutionId": "

The AutomationExecutionId of the parent automation.

", - "AutomationExecutionMetadata$AutomationExecutionId": "

The execution ID.

", - "AutomationExecutionMetadata$ParentAutomationExecutionId": "

The ExecutionId of the parent Automation.

", - "DescribeAutomationStepExecutionsRequest$AutomationExecutionId": "

The Automation execution ID for which you want step execution descriptions.

", - "GetAutomationExecutionRequest$AutomationExecutionId": "

The unique identifier for an existing automation execution to examine. The execution ID is returned by StartAutomationExecution when the execution of an Automation document is initiated.

", - "SendAutomationSignalRequest$AutomationExecutionId": "

The unique identifier for an existing Automation execution that you want to send the signal to.

", - "StartAutomationExecutionResult$AutomationExecutionId": "

The unique ID of a newly scheduled automation execution.

", - "StopAutomationExecutionRequest$AutomationExecutionId": "

The execution ID of the Automation to stop.

" + "AutomationExecutionId":{ + "base":null, + "refs":{ + "AutomationExecution$AutomationExecutionId":"

The execution ID.

", + "AutomationExecution$ParentAutomationExecutionId":"

The AutomationExecutionId of the parent automation.

", + "AutomationExecutionMetadata$AutomationExecutionId":"

The execution ID.

", + "AutomationExecutionMetadata$ParentAutomationExecutionId":"

The ExecutionId of the parent Automation.

", + "DescribeAutomationStepExecutionsRequest$AutomationExecutionId":"

The Automation execution ID for which you want step execution descriptions.

", + "GetAutomationExecutionRequest$AutomationExecutionId":"

The unique identifier for an existing automation execution to examine. The execution ID is returned by StartAutomationExecution when the execution of an Automation document is initiated.

", + "SendAutomationSignalRequest$AutomationExecutionId":"

The unique identifier for an existing Automation execution that you want to send the signal to.

", + "StartAutomationExecutionResult$AutomationExecutionId":"

The unique ID of a newly scheduled automation execution.

", + "StopAutomationExecutionRequest$AutomationExecutionId":"

The execution ID of the Automation to stop.

" } }, - "AutomationExecutionLimitExceededException": { - "base": "

The number of simultaneously running Automation executions exceeded the allowable limit.

", - "refs": { + "AutomationExecutionLimitExceededException":{ + "base":"

The number of simultaneously running Automation executions exceeded the allowable limit.

", + "refs":{ } }, - "AutomationExecutionMetadata": { - "base": "

Details about a specific Automation execution.

", - "refs": { - "AutomationExecutionMetadataList$member": null + "AutomationExecutionMetadata":{ + "base":"

Details about a specific Automation execution.

", + "refs":{ + "AutomationExecutionMetadataList$member":null } }, - "AutomationExecutionMetadataList": { - "base": null, - "refs": { - "DescribeAutomationExecutionsResult$AutomationExecutionMetadataList": "

The list of details about each automation execution which has occurred which matches the filter specification, if any.

" + "AutomationExecutionMetadataList":{ + "base":null, + "refs":{ + "DescribeAutomationExecutionsResult$AutomationExecutionMetadataList":"

The list of details about each automation execution which has occurred which matches the filter specification, if any.

" } }, - "AutomationExecutionNotFoundException": { - "base": "

There is no automation execution information for the requested automation execution ID.

", - "refs": { + "AutomationExecutionNotFoundException":{ + "base":"

There is no automation execution information for the requested automation execution ID.

", + "refs":{ } }, - "AutomationExecutionStatus": { - "base": null, - "refs": { - "AutomationExecution$AutomationExecutionStatus": "

The execution status of the Automation.

", - "AutomationExecutionMetadata$AutomationExecutionStatus": "

The status of the execution. Valid values include: Running, Succeeded, Failed, Timed out, or Cancelled.

", - "StepExecution$StepStatus": "

The execution status for this step. Valid values include: Pending, InProgress, Success, Cancelled, Failed, and TimedOut.

" + "AutomationExecutionStatus":{ + "base":null, + "refs":{ + "AutomationExecution$AutomationExecutionStatus":"

The execution status of the Automation.

", + "AutomationExecutionMetadata$AutomationExecutionStatus":"

The status of the execution. Valid values include: Running, Succeeded, Failed, Timed out, or Cancelled.

", + "StepExecution$StepStatus":"

The execution status for this step. Valid values include: Pending, InProgress, Success, Cancelled, Failed, and TimedOut.

" } }, - "AutomationParameterKey": { - "base": null, - "refs": { - "AutomationExecution$TargetParameterName": "

The parameter name.

", - "AutomationExecutionMetadata$TargetParameterName": "

The list of execution outputs as defined in the Automation document.

", - "AutomationParameterMap$key": null, - "StartAutomationExecutionRequest$TargetParameterName": "

The name of the parameter used as the target resource for the rate-controlled execution. Required if you specify Targets.

" - } - }, - "AutomationParameterMap": { - "base": null, - "refs": { - "AutomationExecution$Parameters": "

The key-value map of execution parameters, which were supplied when calling StartAutomationExecution.

", - "AutomationExecution$Outputs": "

The list of execution outputs as defined in the automation document.

", - "AutomationExecutionMetadata$Outputs": "

The list of execution outputs as defined in the Automation document.

", - "FailureDetails$Details": "

Detailed information about the Automation step failure.

", - "MaintenanceWindowAutomationParameters$Parameters": "

The parameters for the AUTOMATION task.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For AUTOMATION task types, Systems Manager ignores any values specified for these parameters.

", - "SendAutomationSignalRequest$Payload": "

The data sent with the signal. The data schema depends on the type of signal used in the request.

", - "StartAutomationExecutionRequest$Parameters": "

A key-value map of execution parameters, which match the declared parameters in the Automation document.

", - "StepExecution$Outputs": "

Returned values from the execution of the step.

", - "StepExecution$OverriddenParameters": "

A user-specified list of parameters to override when executing a step.

" - } - }, - "AutomationParameterValue": { - "base": null, - "refs": { - "AutomationParameterValueList$member": null - } - }, - "AutomationParameterValueList": { - "base": null, - "refs": { - "AutomationParameterMap$value": null - } - }, - "AutomationStepNotFoundException": { - "base": "

The specified step name and execution ID don't exist. Verify the information and try again.

", - "refs": { - } - }, - "BaselineDescription": { - "base": null, - "refs": { - "CreatePatchBaselineRequest$Description": "

A description of the patch baseline.

", - "GetPatchBaselineResult$Description": "

A description of the patch baseline.

", - "PatchBaselineIdentity$BaselineDescription": "

The description of the patch baseline.

", - "UpdatePatchBaselineRequest$Description": "

A description of the patch baseline.

", - "UpdatePatchBaselineResult$Description": "

A description of the Patch Baseline.

" - } - }, - "BaselineId": { - "base": null, - "refs": { - "CreatePatchBaselineResult$BaselineId": "

The ID of the created patch baseline.

", - "DeletePatchBaselineRequest$BaselineId": "

The ID of the patch baseline to delete.

", - "DeletePatchBaselineResult$BaselineId": "

The ID of the deleted patch baseline.

", - "DeregisterPatchBaselineForPatchGroupRequest$BaselineId": "

The ID of the patch baseline to deregister the patch group from.

", - "DeregisterPatchBaselineForPatchGroupResult$BaselineId": "

The ID of the patch baseline the patch group was deregistered from.

", - "DescribeEffectivePatchesForPatchBaselineRequest$BaselineId": "

The ID of the patch baseline to retrieve the effective patches for.

", - "GetDefaultPatchBaselineResult$BaselineId": "

The ID of the default patch baseline.

", - "GetPatchBaselineForPatchGroupResult$BaselineId": "

The ID of the patch baseline that should be used for the patch group.

", - "GetPatchBaselineRequest$BaselineId": "

The ID of the patch baseline to retrieve.

", - "GetPatchBaselineResult$BaselineId": "

The ID of the retrieved patch baseline.

", - "InstancePatchState$BaselineId": "

The ID of the patch baseline used to patch the instance.

", - "PatchBaselineIdentity$BaselineId": "

The ID of the patch baseline.

", - "RegisterDefaultPatchBaselineRequest$BaselineId": "

The ID of the patch baseline that should be the default patch baseline.

", - "RegisterDefaultPatchBaselineResult$BaselineId": "

The ID of the default patch baseline.

", - "RegisterPatchBaselineForPatchGroupRequest$BaselineId": "

The ID of the patch baseline to register the patch group with.

", - "RegisterPatchBaselineForPatchGroupResult$BaselineId": "

The ID of the patch baseline the patch group was registered with.

", - "UpdatePatchBaselineRequest$BaselineId": "

The ID of the patch baseline to update.

", - "UpdatePatchBaselineResult$BaselineId": "

The ID of the deleted patch baseline.

" - } - }, - "BaselineName": { - "base": null, - "refs": { - "CreatePatchBaselineRequest$Name": "

The name of the patch baseline.

", - "GetPatchBaselineResult$Name": "

The name of the patch baseline.

", - "PatchBaselineIdentity$BaselineName": "

The name of the patch baseline.

", - "UpdatePatchBaselineRequest$Name": "

The name of the patch baseline.

", - "UpdatePatchBaselineResult$Name": "

The name of the patch baseline.

" - } - }, - "BatchErrorMessage": { - "base": null, - "refs": { - "FailedCreateAssociation$Message": "

A description of the failure.

" - } - }, - "Boolean": { - "base": null, - "refs": { - "Activation$Expired": "

Whether or not the activation is expired.

", - "AutomationExecution$StepExecutionsTruncated": "

A boolean value that indicates if the response contains the full list of the Automation step executions. If true, use the DescribeAutomationStepExecutions API action to get the full list of step executions.

", - "CreatePatchBaselineRequest$ApprovedPatchesEnableNonSecurity": "

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

", - "DeregisterTargetFromMaintenanceWindowRequest$Safe": "

The system checks if the target is being referenced by a task. If the target is being referenced, the system returns an error and does not deregister the target from the Maintenance Window.

", - "DescribeAutomationStepExecutionsRequest$ReverseOrder": "

A boolean that indicates whether to list step executions in reverse order by start time. The default value is false.

", - "DocumentVersionInfo$IsDefaultVersion": "

An identifier for the default version of the document.

", - "GetParameterHistoryRequest$WithDecryption": "

Return decrypted values for secure string parameters. This flag is ignored for String and StringList parameter types.

", - "GetParameterRequest$WithDecryption": "

Return decrypted values for secure string parameters. This flag is ignored for String and StringList parameter types.

", - "GetParametersByPathRequest$Recursive": "

Retrieve all parameters within a hierarchy.

If a user has access to a path, then the user can access all levels of that path. For example, if a user has permission to access path /a, then the user can also access /a/b. Even if a user has explicitly been denied access in IAM for parameter /a, they can still call the GetParametersByPath API action recursively and view /a/b.

", - "GetParametersByPathRequest$WithDecryption": "

Retrieve all parameters in a hierarchy with their value decrypted.

", - "GetParametersRequest$WithDecryption": "

Return decrypted secure string value. Return decrypted values for secure string parameters. This flag is ignored for String and StringList parameter types.

", - "GetPatchBaselineResult$ApprovedPatchesEnableNonSecurity": "

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

", - "InstanceInformation$IsLatestVersion": "

Indicates whether latest version of the SSM Agent is running on your instance. Some older versions of Windows Server use the EC2Config service to process SSM requests. For this reason, this field does not indicate whether or not the latest version is installed on Windows managed instances.

", - "ListCommandInvocationsRequest$Details": "

(Optional) If set this returns the response of the command executions and any command output. By default this is set to False.

", - "PatchRule$EnableNonSecurity": "

For instances identified by the approval rule filters, enables a patch baseline to apply non-security updates available in the specified repository. The default value is 'false'. Applies to Linux instances only.

", - "PutParameterRequest$Overwrite": "

Overwrite an existing parameter. If not specified, will default to \"false\".

", - "ResolvedTargets$Truncated": "

A boolean value indicating whether the resolved target list is truncated.

", - "UpdateMaintenanceWindowRequest$Replace": "

If True, then all fields that are required by the CreateMaintenanceWindow action are also required for this API request. Optional fields that are not specified are set to null.

", - "UpdateMaintenanceWindowTargetRequest$Replace": "

If True, then all fields that are required by the RegisterTargetWithMaintenanceWindow action are also required for this API request. Optional fields that are not specified are set to null.

", - "UpdateMaintenanceWindowTaskRequest$Replace": "

If True, then all fields that are required by the RegisterTaskWithMaintenanceWndow action are also required for this API request. Optional fields that are not specified are set to null.

", - "UpdatePatchBaselineRequest$ApprovedPatchesEnableNonSecurity": "

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

", - "UpdatePatchBaselineRequest$Replace": "

If True, then all fields that are required by the CreatePatchBaseline action are also required for this API request. Optional fields that are not specified are set to null.

", - "UpdatePatchBaselineResult$ApprovedPatchesEnableNonSecurity": "

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

" - } - }, - "CancelCommandRequest": { - "base": "

", - "refs": { - } - }, - "CancelCommandResult": { - "base": "

Whether or not the command was successfully canceled. There is no guarantee that a request can be canceled.

", - "refs": { - } - }, - "ClientToken": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$ClientToken": "

User-provided idempotency token.

", - "CreatePatchBaselineRequest$ClientToken": "

User-provided idempotency token.

", - "DeleteInventoryRequest$ClientToken": "

User-provided idempotency token.

", - "RegisterTargetWithMaintenanceWindowRequest$ClientToken": "

User-provided idempotency token.

", - "RegisterTaskWithMaintenanceWindowRequest$ClientToken": "

User-provided idempotency token.

" - } - }, - "Command": { - "base": "

Describes a command request.

", - "refs": { - "CommandList$member": null, - "SendCommandResult$Command": "

The request as it was received by Systems Manager. Also provides the command ID which can be used future references to this request.

" - } - }, - "CommandFilter": { - "base": "

Describes a command filter.

", - "refs": { - "CommandFilterList$member": null - } - }, - "CommandFilterKey": { - "base": null, - "refs": { - "CommandFilter$key": "

The name of the filter.

" - } - }, - "CommandFilterList": { - "base": null, - "refs": { - "ListCommandInvocationsRequest$Filters": "

(Optional) One or more filters. Use a filter to return a more specific list of results.

", - "ListCommandsRequest$Filters": "

(Optional) One or more filters. Use a filter to return a more specific list of results.

" + "AutomationParameterKey":{ + "base":null, + "refs":{ + "AutomationExecution$TargetParameterName":"

The parameter name.

", + "AutomationExecutionMetadata$TargetParameterName":"

The list of execution outputs as defined in the Automation document.

", + "AutomationParameterMap$key":null, + "StartAutomationExecutionRequest$TargetParameterName":"

The name of the parameter used as the target resource for the rate-controlled execution. Required if you specify Targets.

" + } + }, + "AutomationParameterMap":{ + "base":null, + "refs":{ + "AutomationExecution$Parameters":"

The key-value map of execution parameters, which were supplied when calling StartAutomationExecution.

", + "AutomationExecution$Outputs":"

The list of execution outputs as defined in the automation document.

", + "AutomationExecutionMetadata$Outputs":"

The list of execution outputs as defined in the Automation document.

", + "FailureDetails$Details":"

Detailed information about the Automation step failure.

", + "MaintenanceWindowAutomationParameters$Parameters":"

The parameters for the AUTOMATION task.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For AUTOMATION task types, Systems Manager ignores any values specified for these parameters.

", + "SendAutomationSignalRequest$Payload":"

The data sent with the signal. The data schema depends on the type of signal used in the request.

", + "StartAutomationExecutionRequest$Parameters":"

A key-value map of execution parameters, which match the declared parameters in the Automation document.

", + "StepExecution$Outputs":"

Returned values from the execution of the step.

", + "StepExecution$OverriddenParameters":"

A user-specified list of parameters to override when executing a step.

" + } + }, + "AutomationParameterValue":{ + "base":null, + "refs":{ + "AutomationParameterValueList$member":null + } + }, + "AutomationParameterValueList":{ + "base":null, + "refs":{ + "AutomationParameterMap$value":null + } + }, + "AutomationStepNotFoundException":{ + "base":"

The specified step name and execution ID don't exist. Verify the information and try again.

", + "refs":{ + } + }, + "BaselineDescription":{ + "base":null, + "refs":{ + "CreatePatchBaselineRequest$Description":"

A description of the patch baseline.

", + "GetPatchBaselineResult$Description":"

A description of the patch baseline.

", + "PatchBaselineIdentity$BaselineDescription":"

The description of the patch baseline.

", + "UpdatePatchBaselineRequest$Description":"

A description of the patch baseline.

", + "UpdatePatchBaselineResult$Description":"

A description of the Patch Baseline.

" + } + }, + "BaselineId":{ + "base":null, + "refs":{ + "CreatePatchBaselineResult$BaselineId":"

The ID of the created patch baseline.

", + "DeletePatchBaselineRequest$BaselineId":"

The ID of the patch baseline to delete.

", + "DeletePatchBaselineResult$BaselineId":"

The ID of the deleted patch baseline.

", + "DeregisterPatchBaselineForPatchGroupRequest$BaselineId":"

The ID of the patch baseline to deregister the patch group from.

", + "DeregisterPatchBaselineForPatchGroupResult$BaselineId":"

The ID of the patch baseline the patch group was deregistered from.

", + "DescribeEffectivePatchesForPatchBaselineRequest$BaselineId":"

The ID of the patch baseline to retrieve the effective patches for.

", + "GetDefaultPatchBaselineResult$BaselineId":"

The ID of the default patch baseline.

", + "GetPatchBaselineForPatchGroupResult$BaselineId":"

The ID of the patch baseline that should be used for the patch group.

", + "GetPatchBaselineRequest$BaselineId":"

The ID of the patch baseline to retrieve.

", + "GetPatchBaselineResult$BaselineId":"

The ID of the retrieved patch baseline.

", + "InstancePatchState$BaselineId":"

The ID of the patch baseline used to patch the instance.

", + "PatchBaselineIdentity$BaselineId":"

The ID of the patch baseline.

", + "RegisterDefaultPatchBaselineRequest$BaselineId":"

The ID of the patch baseline that should be the default patch baseline.

", + "RegisterDefaultPatchBaselineResult$BaselineId":"

The ID of the default patch baseline.

", + "RegisterPatchBaselineForPatchGroupRequest$BaselineId":"

The ID of the patch baseline to register the patch group with.

", + "RegisterPatchBaselineForPatchGroupResult$BaselineId":"

The ID of the patch baseline the patch group was registered with.

", + "UpdatePatchBaselineRequest$BaselineId":"

The ID of the patch baseline to update.

", + "UpdatePatchBaselineResult$BaselineId":"

The ID of the deleted patch baseline.

" + } + }, + "BaselineName":{ + "base":null, + "refs":{ + "CreatePatchBaselineRequest$Name":"

The name of the patch baseline.

", + "GetPatchBaselineResult$Name":"

The name of the patch baseline.

", + "PatchBaselineIdentity$BaselineName":"

The name of the patch baseline.

", + "UpdatePatchBaselineRequest$Name":"

The name of the patch baseline.

", + "UpdatePatchBaselineResult$Name":"

The name of the patch baseline.

" + } + }, + "BatchErrorMessage":{ + "base":null, + "refs":{ + "FailedCreateAssociation$Message":"

A description of the failure.

" + } + }, + "Boolean":{ + "base":null, + "refs":{ + "Activation$Expired":"

Whether or not the activation is expired.

", + "AutomationExecution$StepExecutionsTruncated":"

A boolean value that indicates if the response contains the full list of the Automation step executions. If true, use the DescribeAutomationStepExecutions API action to get the full list of step executions.

", + "CreatePatchBaselineRequest$ApprovedPatchesEnableNonSecurity":"

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

", + "DeregisterTargetFromMaintenanceWindowRequest$Safe":"

The system checks if the target is being referenced by a task. If the target is being referenced, the system returns an error and does not deregister the target from the Maintenance Window.

", + "DescribeAutomationStepExecutionsRequest$ReverseOrder":"

A boolean that indicates whether to list step executions in reverse order by start time. The default value is false.

", + "DocumentVersionInfo$IsDefaultVersion":"

An identifier for the default version of the document.

", + "GetParameterHistoryRequest$WithDecryption":"

Return decrypted values for secure string parameters. This flag is ignored for String and StringList parameter types.

", + "GetParameterRequest$WithDecryption":"

Return decrypted values for secure string parameters. This flag is ignored for String and StringList parameter types.

", + "GetParametersByPathRequest$Recursive":"

Retrieve all parameters within a hierarchy.

If a user has access to a path, then the user can access all levels of that path. For example, if a user has permission to access path /a, then the user can also access /a/b. Even if a user has explicitly been denied access in IAM for parameter /a, they can still call the GetParametersByPath API action recursively and view /a/b.

", + "GetParametersByPathRequest$WithDecryption":"

Retrieve all parameters in a hierarchy with their value decrypted.

", + "GetParametersRequest$WithDecryption":"

Return decrypted secure string value. Return decrypted values for secure string parameters. This flag is ignored for String and StringList parameter types.

", + "GetPatchBaselineResult$ApprovedPatchesEnableNonSecurity":"

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

", + "InstanceInformation$IsLatestVersion":"

Indicates whether latest version of SSM Agent is running on your instance. Some older versions of Windows Server use the EC2Config service to process SSM requests. For this reason, this field does not indicate whether or not the latest version is installed on Windows managed instances.

", + "ListCommandInvocationsRequest$Details":"

(Optional) If set this returns the response of the command executions and any command output. By default this is set to False.

", + "PatchRule$EnableNonSecurity":"

For instances identified by the approval rule filters, enables a patch baseline to apply non-security updates available in the specified repository. The default value is 'false'. Applies to Linux instances only.

", + "PutParameterRequest$Overwrite":"

Overwrite an existing parameter. If not specified, will default to \"false\".

", + "ResolvedTargets$Truncated":"

A boolean value indicating whether the resolved target list is truncated.

", + "UpdateMaintenanceWindowRequest$Replace":"

If True, then all fields that are required by the CreateMaintenanceWindow action are also required for this API request. Optional fields that are not specified are set to null.

", + "UpdateMaintenanceWindowTargetRequest$Replace":"

If True, then all fields that are required by the RegisterTargetWithMaintenanceWindow action are also required for this API request. Optional fields that are not specified are set to null.

", + "UpdateMaintenanceWindowTaskRequest$Replace":"

If True, then all fields that are required by the RegisterTaskWithMaintenanceWndow action are also required for this API request. Optional fields that are not specified are set to null.

", + "UpdatePatchBaselineRequest$ApprovedPatchesEnableNonSecurity":"

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

", + "UpdatePatchBaselineRequest$Replace":"

If True, then all fields that are required by the CreatePatchBaseline action are also required for this API request. Optional fields that are not specified are set to null.

", + "UpdatePatchBaselineResult$ApprovedPatchesEnableNonSecurity":"

Indicates whether the list of approved patches includes non-security updates that should be applied to the instances. The default value is 'false'. Applies to Linux instances only.

" + } + }, + "CancelCommandRequest":{ + "base":"

", + "refs":{ + } + }, + "CancelCommandResult":{ + "base":"

Whether or not the command was successfully canceled. There is no guarantee that a request can be canceled.

", + "refs":{ + } + }, + "ClientToken":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$ClientToken":"

User-provided idempotency token.

", + "CreatePatchBaselineRequest$ClientToken":"

User-provided idempotency token.

", + "DeleteInventoryRequest$ClientToken":"

User-provided idempotency token.

", + "RegisterTargetWithMaintenanceWindowRequest$ClientToken":"

User-provided idempotency token.

", + "RegisterTaskWithMaintenanceWindowRequest$ClientToken":"

User-provided idempotency token.

" + } + }, + "CloudWatchLogGroupName":{ + "base":null, + "refs":{ + "CloudWatchOutputConfig$CloudWatchLogGroupName":"

The name of the CloudWatch log group where you want to send command output. If you don't specify a group name, Systems Manager automatically creates a log group for you. The log group uses the following naming format: aws/ssm/SystemsManagerDocumentName.

" + } + }, + "CloudWatchOutputConfig":{ + "base":"

Configuration options for sending command output to CloudWatch Logs.

", + "refs":{ + "Command$CloudWatchOutputConfig":"

CloudWatch Logs information where you want Systems Manager to send the command output.

", + "CommandInvocation$CloudWatchOutputConfig":"

CloudWatch Logs information where you want Systems Manager to send the command output.

", + "GetCommandInvocationResult$CloudWatchOutputConfig":"

CloudWatch Logs information where Systems Manager sent the command output.

", + "SendCommandRequest$CloudWatchOutputConfig":"

Enables Systems Manager to send Run Command output to Amazon CloudWatch Logs.

" + } + }, + "CloudWatchOutputEnabled":{ + "base":null, + "refs":{ + "CloudWatchOutputConfig$CloudWatchOutputEnabled":"

Enables Systems Manager to send command output to CloudWatch Logs.

" + } + }, + "Command":{ + "base":"

Describes a command request.

", + "refs":{ + "CommandList$member":null, + "SendCommandResult$Command":"

The request as it was received by Systems Manager. Also provides the command ID which can be used future references to this request.

" } }, - "CommandFilterValue": { - "base": null, - "refs": { - "CommandFilter$value": "

The filter value.

" + "CommandFilter":{ + "base":"

Describes a command filter.

", + "refs":{ + "CommandFilterList$member":null } }, - "CommandId": { - "base": null, - "refs": { - "CancelCommandRequest$CommandId": "

The ID of the command you want to cancel.

", - "Command$CommandId": "

A unique identifier for this command.

", - "CommandInvocation$CommandId": "

The command against which this invocation was requested.

", - "GetCommandInvocationRequest$CommandId": "

(Required) The parent command ID of the invocation plugin.

", - "GetCommandInvocationResult$CommandId": "

The parent command ID of the invocation plugin.

", - "ListCommandInvocationsRequest$CommandId": "

(Optional) The invocations for a specific command ID.

", - "ListCommandsRequest$CommandId": "

(Optional) If provided, lists only the specified command.

" + "CommandFilterKey":{ + "base":null, + "refs":{ + "CommandFilter$key":"

The name of the filter.

" } }, - "CommandInvocation": { - "base": "

An invocation is copy of a command sent to a specific instance. A command can apply to one or more instances. A command invocation applies to one instance. For example, if a user executes SendCommand against three instances, then a command invocation is created for each requested instance ID. A command invocation returns status and detail information about a command you executed.

", - "refs": { - "CommandInvocationList$member": null + "CommandFilterList":{ + "base":null, + "refs":{ + "ListCommandInvocationsRequest$Filters":"

(Optional) One or more filters. Use a filter to return a more specific list of results.

", + "ListCommandsRequest$Filters":"

(Optional) One or more filters. Use a filter to return a more specific list of results.

" } }, - "CommandInvocationList": { - "base": null, - "refs": { - "ListCommandInvocationsResult$CommandInvocations": "

(Optional) A list of all invocations.

" + "CommandFilterValue":{ + "base":null, + "refs":{ + "CommandFilter$value":"

The filter value.

" } }, - "CommandInvocationStatus": { - "base": null, - "refs": { - "CommandInvocation$Status": "

Whether or not the invocation succeeded, failed, or is pending.

", - "GetCommandInvocationResult$Status": "

The status of this invocation plugin. This status can be different than StatusDetails.

" + "CommandId":{ + "base":null, + "refs":{ + "CancelCommandRequest$CommandId":"

The ID of the command you want to cancel.

", + "Command$CommandId":"

A unique identifier for this command.

", + "CommandInvocation$CommandId":"

The command against which this invocation was requested.

", + "GetCommandInvocationRequest$CommandId":"

(Required) The parent command ID of the invocation plugin.

", + "GetCommandInvocationResult$CommandId":"

The parent command ID of the invocation plugin.

", + "ListCommandInvocationsRequest$CommandId":"

(Optional) The invocations for a specific command ID.

", + "ListCommandsRequest$CommandId":"

(Optional) If provided, lists only the specified command.

" } }, - "CommandList": { - "base": null, - "refs": { - "ListCommandsResult$Commands": "

(Optional) The list of commands requested by the user.

" + "CommandInvocation":{ + "base":"

An invocation is copy of a command sent to a specific instance. A command can apply to one or more instances. A command invocation applies to one instance. For example, if a user executes SendCommand against three instances, then a command invocation is created for each requested instance ID. A command invocation returns status and detail information about a command you executed.

", + "refs":{ + "CommandInvocationList$member":null } }, - "CommandMaxResults": { - "base": null, - "refs": { - "ListCommandInvocationsRequest$MaxResults": "

(Optional) The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListCommandsRequest$MaxResults": "

(Optional) The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + "CommandInvocationList":{ + "base":null, + "refs":{ + "ListCommandInvocationsResult$CommandInvocations":"

(Optional) A list of all invocations.

" } }, - "CommandPlugin": { - "base": "

Describes plugin details.

", - "refs": { - "CommandPluginList$member": null + "CommandInvocationStatus":{ + "base":null, + "refs":{ + "CommandInvocation$Status":"

Whether or not the invocation succeeded, failed, or is pending.

", + "GetCommandInvocationResult$Status":"

The status of this invocation plugin. This status can be different than StatusDetails.

" } }, - "CommandPluginList": { - "base": null, - "refs": { - "CommandInvocation$CommandPlugins": null + "CommandList":{ + "base":null, + "refs":{ + "ListCommandsResult$Commands":"

(Optional) The list of commands requested by the user.

" } }, - "CommandPluginName": { - "base": null, - "refs": { - "CommandPlugin$Name": "

The name of the plugin. Must be one of the following: aws:updateAgent, aws:domainjoin, aws:applications, aws:runPowerShellScript, aws:psmodule, aws:cloudWatch, aws:runShellScript, or aws:updateSSMAgent.

", - "GetCommandInvocationRequest$PluginName": "

(Optional) The name of the plugin for which you want detailed results. If the document contains only one plugin, the name can be omitted and the details will be returned.

", - "GetCommandInvocationResult$PluginName": "

The name of the plugin for which you want detailed results. For example, aws:RunShellScript is a plugin.

" + "CommandMaxResults":{ + "base":null, + "refs":{ + "ListCommandInvocationsRequest$MaxResults":"

(Optional) The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListCommandsRequest$MaxResults":"

(Optional) The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" } }, - "CommandPluginOutput": { - "base": null, - "refs": { - "CommandPlugin$Output": "

Output of the plugin execution.

" + "CommandPlugin":{ + "base":"

Describes plugin details.

", + "refs":{ + "CommandPluginList$member":null } }, - "CommandPluginStatus": { - "base": null, - "refs": { - "CommandPlugin$Status": "

The status of this plugin. You can execute a document with multiple plugins.

" + "CommandPluginList":{ + "base":null, + "refs":{ + "CommandInvocation$CommandPlugins":null } }, - "CommandStatus": { - "base": null, - "refs": { - "Command$Status": "

The status of the command.

" + "CommandPluginName":{ + "base":null, + "refs":{ + "CommandPlugin$Name":"

The name of the plugin. Must be one of the following: aws:updateAgent, aws:domainjoin, aws:applications, aws:runPowerShellScript, aws:psmodule, aws:cloudWatch, aws:runShellScript, or aws:updateSSMAgent.

", + "GetCommandInvocationRequest$PluginName":"

(Optional) The name of the plugin for which you want detailed results. If the document contains only one plugin, the name can be omitted and the details will be returned.

", + "GetCommandInvocationResult$PluginName":"

The name of the plugin for which you want detailed results. For example, aws:RunShellScript is a plugin.

" } }, - "Comment": { - "base": null, - "refs": { - "Command$Comment": "

User-specified information about the command, such as a brief description of what the command should do.

", - "CommandInvocation$Comment": "

User-specified information about the command, such as a brief description of what the command should do.

", - "GetCommandInvocationResult$Comment": "

The comment text for the command.

", - "MaintenanceWindowRunCommandParameters$Comment": "

Information about the command(s) to execute.

", - "SendCommandRequest$Comment": "

User-specified information about the command, such as a brief description of what the command should do.

" + "CommandPluginOutput":{ + "base":null, + "refs":{ + "CommandPlugin$Output":"

Output of the plugin execution.

" } }, - "CompletedCount": { - "base": null, - "refs": { - "Command$CompletedCount": "

The number of targets for which the command invocation reached a terminal state. Terminal states include the following: Success, Failed, Execution Timed Out, Delivery Timed Out, Canceled, Terminated, or Undeliverable.

" + "CommandPluginStatus":{ + "base":null, + "refs":{ + "CommandPlugin$Status":"

The status of this plugin. You can execute a document with multiple plugins.

" } }, - "ComplianceExecutionId": { - "base": null, - "refs": { - "ComplianceExecutionSummary$ExecutionId": "

An ID created by the system when PutComplianceItems was called. For example, CommandID is a valid execution ID. You can use this ID in subsequent calls.

" + "CommandStatus":{ + "base":null, + "refs":{ + "Command$Status":"

The status of the command.

" } }, - "ComplianceExecutionSummary": { - "base": "

A summary of the call execution that includes an execution ID, the type of execution (for example, Command), and the date/time of the execution using a datetime object that is saved in the following format: yyyy-MM-dd'T'HH:mm:ss'Z'.

", - "refs": { - "ComplianceItem$ExecutionSummary": "

A summary for the compliance item. The summary includes an execution ID, the execution type (for example, command), and the execution time.

", - "PutComplianceItemsRequest$ExecutionSummary": "

A summary of the call execution that includes an execution ID, the type of execution (for example, Command), and the date/time of the execution using a datetime object that is saved in the following format: yyyy-MM-dd'T'HH:mm:ss'Z'.

", - "ResourceComplianceSummaryItem$ExecutionSummary": "

Information about the execution.

" + "Comment":{ + "base":null, + "refs":{ + "Command$Comment":"

User-specified information about the command, such as a brief description of what the command should do.

", + "CommandInvocation$Comment":"

User-specified information about the command, such as a brief description of what the command should do.

", + "GetCommandInvocationResult$Comment":"

The comment text for the command.

", + "MaintenanceWindowRunCommandParameters$Comment":"

Information about the command(s) to execute.

", + "SendCommandRequest$Comment":"

User-specified information about the command, such as a brief description of what the command should do.

" } }, - "ComplianceExecutionType": { - "base": null, - "refs": { - "ComplianceExecutionSummary$ExecutionType": "

The type of execution. For example, Command is a valid execution type.

" + "CompletedCount":{ + "base":null, + "refs":{ + "Command$CompletedCount":"

The number of targets for which the command invocation reached a terminal state. Terminal states include the following: Success, Failed, Execution Timed Out, Delivery Timed Out, Canceled, Terminated, or Undeliverable.

" } }, - "ComplianceFilterValue": { - "base": null, - "refs": { - "ComplianceStringFilterValueList$member": null + "ComplianceExecutionId":{ + "base":null, + "refs":{ + "ComplianceExecutionSummary$ExecutionId":"

An ID created by the system when PutComplianceItems was called. For example, CommandID is a valid execution ID. You can use this ID in subsequent calls.

" } }, - "ComplianceItem": { - "base": "

Information about the compliance as defined by the resource type. For example, for a patch resource type, Items includes information about the PatchSeverity, Classification, etc.

", - "refs": { - "ComplianceItemList$member": null + "ComplianceExecutionSummary":{ + "base":"

A summary of the call execution that includes an execution ID, the type of execution (for example, Command), and the date/time of the execution using a datetime object that is saved in the following format: yyyy-MM-dd'T'HH:mm:ss'Z'.

", + "refs":{ + "ComplianceItem$ExecutionSummary":"

A summary for the compliance item. The summary includes an execution ID, the execution type (for example, command), and the execution time.

", + "PutComplianceItemsRequest$ExecutionSummary":"

A summary of the call execution that includes an execution ID, the type of execution (for example, Command), and the date/time of the execution using a datetime object that is saved in the following format: yyyy-MM-dd'T'HH:mm:ss'Z'.

", + "ResourceComplianceSummaryItem$ExecutionSummary":"

Information about the execution.

" } }, - "ComplianceItemContentHash": { - "base": null, - "refs": { - "PutComplianceItemsRequest$ItemContentHash": "

MD5 or SHA-256 content hash. The content hash is used to determine if existing information should be overwritten or ignored. If the content hashes match, the request to put compliance information is ignored.

" + "ComplianceExecutionType":{ + "base":null, + "refs":{ + "ComplianceExecutionSummary$ExecutionType":"

The type of execution. For example, Command is a valid execution type.

" } }, - "ComplianceItemDetails": { - "base": null, - "refs": { - "ComplianceItem$Details": "

A \"Key\": \"Value\" tag combination for the compliance item.

", - "ComplianceItemEntry$Details": "

A \"Key\": \"Value\" tag combination for the compliance item.

" + "ComplianceFilterValue":{ + "base":null, + "refs":{ + "ComplianceStringFilterValueList$member":null } }, - "ComplianceItemEntry": { - "base": "

Information about a compliance item.

", - "refs": { - "ComplianceItemEntryList$member": null + "ComplianceItem":{ + "base":"

Information about the compliance as defined by the resource type. For example, for a patch resource type, Items includes information about the PatchSeverity, Classification, etc.

", + "refs":{ + "ComplianceItemList$member":null } }, - "ComplianceItemEntryList": { - "base": null, - "refs": { - "PutComplianceItemsRequest$Items": "

Information about the compliance as defined by the resource type. For example, for a patch compliance type, Items includes information about the PatchSeverity, Classification, etc.

" + "ComplianceItemContentHash":{ + "base":null, + "refs":{ + "PutComplianceItemsRequest$ItemContentHash":"

MD5 or SHA-256 content hash. The content hash is used to determine if existing information should be overwritten or ignored. If the content hashes match, the request to put compliance information is ignored.

" } }, - "ComplianceItemId": { - "base": null, - "refs": { - "ComplianceItem$Id": "

An ID for the compliance item. For example, if the compliance item is a Windows patch, the ID could be the number of the KB article; for example: KB4010320.

", - "ComplianceItemEntry$Id": "

The compliance item ID. For example, if the compliance item is a Windows patch, the ID could be the number of the KB article.

" + "ComplianceItemDetails":{ + "base":null, + "refs":{ + "ComplianceItem$Details":"

A \"Key\": \"Value\" tag combination for the compliance item.

", + "ComplianceItemEntry$Details":"

A \"Key\": \"Value\" tag combination for the compliance item.

" } }, - "ComplianceItemList": { - "base": null, - "refs": { - "ListComplianceItemsResult$ComplianceItems": "

A list of compliance information for the specified resource ID.

" + "ComplianceItemEntry":{ + "base":"

Information about a compliance item.

", + "refs":{ + "ComplianceItemEntryList$member":null } }, - "ComplianceItemTitle": { - "base": null, - "refs": { - "ComplianceItem$Title": "

A title for the compliance item. For example, if the compliance item is a Windows patch, the title could be the title of the KB article for the patch; for example: Security Update for Active Directory Federation Services.

", - "ComplianceItemEntry$Title": "

The title of the compliance item. For example, if the compliance item is a Windows patch, the title could be the title of the KB article for the patch; for example: Security Update for Active Directory Federation Services.

" + "ComplianceItemEntryList":{ + "base":null, + "refs":{ + "PutComplianceItemsRequest$Items":"

Information about the compliance as defined by the resource type. For example, for a patch compliance type, Items includes information about the PatchSeverity, Classification, etc.

" } }, - "ComplianceQueryOperatorType": { - "base": null, - "refs": { - "ComplianceStringFilter$Type": "

The type of comparison that should be performed for the value: Equal, NotEqual, BeginWith, LessThan, or GreaterThan.

" + "ComplianceItemId":{ + "base":null, + "refs":{ + "ComplianceItem$Id":"

An ID for the compliance item. For example, if the compliance item is a Windows patch, the ID could be the number of the KB article; for example: KB4010320.

", + "ComplianceItemEntry$Id":"

The compliance item ID. For example, if the compliance item is a Windows patch, the ID could be the number of the KB article.

" } }, - "ComplianceResourceId": { - "base": null, - "refs": { - "ComplianceItem$ResourceId": "

An ID for the resource. For a managed instance, this is the instance ID.

", - "ComplianceResourceIdList$member": null, - "PutComplianceItemsRequest$ResourceId": "

Specify an ID for this resource. For a managed instance, this is the instance ID.

", - "ResourceComplianceSummaryItem$ResourceId": "

The resource ID.

" + "ComplianceItemList":{ + "base":null, + "refs":{ + "ListComplianceItemsResult$ComplianceItems":"

A list of compliance information for the specified resource ID.

" } }, - "ComplianceResourceIdList": { - "base": null, - "refs": { - "ListComplianceItemsRequest$ResourceIds": "

The ID for the resources from which to get compliance information. Currently, you can only specify one resource ID.

" + "ComplianceItemTitle":{ + "base":null, + "refs":{ + "ComplianceItem$Title":"

A title for the compliance item. For example, if the compliance item is a Windows patch, the title could be the title of the KB article for the patch; for example: Security Update for Active Directory Federation Services.

", + "ComplianceItemEntry$Title":"

The title of the compliance item. For example, if the compliance item is a Windows patch, the title could be the title of the KB article for the patch; for example: Security Update for Active Directory Federation Services.

" } }, - "ComplianceResourceType": { - "base": null, - "refs": { - "ComplianceItem$ResourceType": "

The type of resource. ManagedInstance is currently the only supported resource type.

", - "ComplianceResourceTypeList$member": null, - "PutComplianceItemsRequest$ResourceType": "

Specify the type of resource. ManagedInstance is currently the only supported resource type.

", - "ResourceComplianceSummaryItem$ResourceType": "

The resource type.

" + "ComplianceQueryOperatorType":{ + "base":null, + "refs":{ + "ComplianceStringFilter$Type":"

The type of comparison that should be performed for the value: Equal, NotEqual, BeginWith, LessThan, or GreaterThan.

" } }, - "ComplianceResourceTypeList": { - "base": null, - "refs": { - "ListComplianceItemsRequest$ResourceTypes": "

The type of resource from which to get compliance information. Currently, the only supported resource type is ManagedInstance.

" + "ComplianceResourceId":{ + "base":null, + "refs":{ + "ComplianceItem$ResourceId":"

An ID for the resource. For a managed instance, this is the instance ID.

", + "ComplianceResourceIdList$member":null, + "PutComplianceItemsRequest$ResourceId":"

Specify an ID for this resource. For a managed instance, this is the instance ID.

", + "ResourceComplianceSummaryItem$ResourceId":"

The resource ID.

" } }, - "ComplianceSeverity": { - "base": null, - "refs": { - "ComplianceItem$Severity": "

The severity of the compliance status. Severity can be one of the following: Critical, High, Medium, Low, Informational, Unspecified.

", - "ComplianceItemEntry$Severity": "

The severity of the compliance status. Severity can be one of the following: Critical, High, Medium, Low, Informational, Unspecified.

", - "ResourceComplianceSummaryItem$OverallSeverity": "

The highest severity item found for the resource. The resource is compliant for this item.

" + "ComplianceResourceIdList":{ + "base":null, + "refs":{ + "ListComplianceItemsRequest$ResourceIds":"

The ID for the resources from which to get compliance information. Currently, you can only specify one resource ID.

" } }, - "ComplianceStatus": { - "base": null, - "refs": { - "ComplianceItem$Status": "

The status of the compliance item. An item is either COMPLIANT or NON_COMPLIANT.

", - "ComplianceItemEntry$Status": "

The status of the compliance item. An item is either COMPLIANT or NON_COMPLIANT.

", - "ResourceComplianceSummaryItem$Status": "

The compliance status for the resource.

" + "ComplianceResourceType":{ + "base":null, + "refs":{ + "ComplianceItem$ResourceType":"

The type of resource. ManagedInstance is currently the only supported resource type.

", + "ComplianceResourceTypeList$member":null, + "PutComplianceItemsRequest$ResourceType":"

Specify the type of resource. ManagedInstance is currently the only supported resource type.

", + "ResourceComplianceSummaryItem$ResourceType":"

The resource type.

" } }, - "ComplianceStringFilter": { - "base": "

One or more filters. Use a filter to return a more specific list of results.

", - "refs": { - "ComplianceStringFilterList$member": null + "ComplianceResourceTypeList":{ + "base":null, + "refs":{ + "ListComplianceItemsRequest$ResourceTypes":"

The type of resource from which to get compliance information. Currently, the only supported resource type is ManagedInstance.

" } }, - "ComplianceStringFilterKey": { - "base": null, - "refs": { - "ComplianceStringFilter$Key": "

The name of the filter.

" + "ComplianceSeverity":{ + "base":null, + "refs":{ + "ComplianceItem$Severity":"

The severity of the compliance status. Severity can be one of the following: Critical, High, Medium, Low, Informational, Unspecified.

", + "ComplianceItemEntry$Severity":"

The severity of the compliance status. Severity can be one of the following: Critical, High, Medium, Low, Informational, Unspecified.

", + "ResourceComplianceSummaryItem$OverallSeverity":"

The highest severity item found for the resource. The resource is compliant for this item.

" } }, - "ComplianceStringFilterList": { - "base": null, - "refs": { - "ListComplianceItemsRequest$Filters": "

One or more compliance filters. Use a filter to return a more specific list of results.

", - "ListComplianceSummariesRequest$Filters": "

One or more compliance or inventory filters. Use a filter to return a more specific list of results.

", - "ListResourceComplianceSummariesRequest$Filters": "

One or more filters. Use a filter to return a more specific list of results.

" + "ComplianceStatus":{ + "base":null, + "refs":{ + "ComplianceItem$Status":"

The status of the compliance item. An item is either COMPLIANT or NON_COMPLIANT.

", + "ComplianceItemEntry$Status":"

The status of the compliance item. An item is either COMPLIANT or NON_COMPLIANT.

", + "ResourceComplianceSummaryItem$Status":"

The compliance status for the resource.

" } }, - "ComplianceStringFilterValueList": { - "base": null, - "refs": { - "ComplianceStringFilter$Values": "

The value for which to search.

" + "ComplianceStringFilter":{ + "base":"

One or more filters. Use a filter to return a more specific list of results.

", + "refs":{ + "ComplianceStringFilterList$member":null } }, - "ComplianceSummaryCount": { - "base": null, - "refs": { - "CompliantSummary$CompliantCount": "

The total number of resources that are compliant.

", - "NonCompliantSummary$NonCompliantCount": "

The total number of compliance items that are not compliant.

", - "SeveritySummary$CriticalCount": "

The total number of resources or compliance items that have a severity level of critical. Critical severity is determined by the organization that published the compliance items.

", - "SeveritySummary$HighCount": "

The total number of resources or compliance items that have a severity level of high. High severity is determined by the organization that published the compliance items.

", - "SeveritySummary$MediumCount": "

The total number of resources or compliance items that have a severity level of medium. Medium severity is determined by the organization that published the compliance items.

", - "SeveritySummary$LowCount": "

The total number of resources or compliance items that have a severity level of low. Low severity is determined by the organization that published the compliance items.

", - "SeveritySummary$InformationalCount": "

The total number of resources or compliance items that have a severity level of informational. Informational severity is determined by the organization that published the compliance items.

", - "SeveritySummary$UnspecifiedCount": "

The total number of resources or compliance items that have a severity level of unspecified. Unspecified severity is determined by the organization that published the compliance items.

" + "ComplianceStringFilterKey":{ + "base":null, + "refs":{ + "ComplianceStringFilter$Key":"

The name of the filter.

" } }, - "ComplianceSummaryItem": { - "base": "

A summary of compliance information by compliance type.

", - "refs": { - "ComplianceSummaryItemList$member": null + "ComplianceStringFilterList":{ + "base":null, + "refs":{ + "ListComplianceItemsRequest$Filters":"

One or more compliance filters. Use a filter to return a more specific list of results.

", + "ListComplianceSummariesRequest$Filters":"

One or more compliance or inventory filters. Use a filter to return a more specific list of results.

", + "ListResourceComplianceSummariesRequest$Filters":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "ComplianceSummaryItemList": { - "base": null, - "refs": { - "ListComplianceSummariesResult$ComplianceSummaryItems": "

A list of compliant and non-compliant summary counts based on compliance types. For example, this call returns State Manager associations, patches, or custom compliance types according to the filter criteria that you specified.

" + "ComplianceStringFilterValueList":{ + "base":null, + "refs":{ + "ComplianceStringFilter$Values":"

The value for which to search.

" } }, - "ComplianceTypeCountLimitExceededException": { - "base": "

You specified too many custom compliance types. You can specify a maximum of 10 different types.

", - "refs": { + "ComplianceSummaryCount":{ + "base":null, + "refs":{ + "CompliantSummary$CompliantCount":"

The total number of resources that are compliant.

", + "NonCompliantSummary$NonCompliantCount":"

The total number of compliance items that are not compliant.

", + "SeveritySummary$CriticalCount":"

The total number of resources or compliance items that have a severity level of critical. Critical severity is determined by the organization that published the compliance items.

", + "SeveritySummary$HighCount":"

The total number of resources or compliance items that have a severity level of high. High severity is determined by the organization that published the compliance items.

", + "SeveritySummary$MediumCount":"

The total number of resources or compliance items that have a severity level of medium. Medium severity is determined by the organization that published the compliance items.

", + "SeveritySummary$LowCount":"

The total number of resources or compliance items that have a severity level of low. Low severity is determined by the organization that published the compliance items.

", + "SeveritySummary$InformationalCount":"

The total number of resources or compliance items that have a severity level of informational. Informational severity is determined by the organization that published the compliance items.

", + "SeveritySummary$UnspecifiedCount":"

The total number of resources or compliance items that have a severity level of unspecified. Unspecified severity is determined by the organization that published the compliance items.

" } }, - "ComplianceTypeName": { - "base": null, - "refs": { - "ComplianceItem$ComplianceType": "

The compliance type. For example, Association (for a State Manager association), Patch, or Custom:string are all valid compliance types.

", - "ComplianceSummaryItem$ComplianceType": "

The type of compliance item. For example, the compliance type can be Association, Patch, or Custom:string.

", - "PutComplianceItemsRequest$ComplianceType": "

Specify the compliance type. For example, specify Association (for a State Manager association), Patch, or Custom:string.

", - "ResourceComplianceSummaryItem$ComplianceType": "

The compliance type.

" + "ComplianceSummaryItem":{ + "base":"

A summary of compliance information by compliance type.

", + "refs":{ + "ComplianceSummaryItemList$member":null } }, - "CompliantSummary": { - "base": "

A summary of resources that are compliant. The summary is organized according to the resource count for each compliance type.

", - "refs": { - "ComplianceSummaryItem$CompliantSummary": "

A list of COMPLIANT items for the specified compliance type.

", - "ResourceComplianceSummaryItem$CompliantSummary": "

A list of items that are compliant for the resource.

" + "ComplianceSummaryItemList":{ + "base":null, + "refs":{ + "ListComplianceSummariesResult$ComplianceSummaryItems":"

A list of compliant and non-compliant summary counts based on compliance types. For example, this call returns State Manager associations, patches, or custom compliance types according to the filter criteria that you specified.

" } }, - "ComputerName": { - "base": null, - "refs": { - "InstanceInformation$ComputerName": "

The fully qualified host name of the managed instance.

" + "ComplianceTypeCountLimitExceededException":{ + "base":"

You specified too many custom compliance types. You can specify a maximum of 10 different types.

", + "refs":{ } }, - "CreateActivationRequest": { - "base": null, - "refs": { + "ComplianceTypeName":{ + "base":null, + "refs":{ + "ComplianceItem$ComplianceType":"

The compliance type. For example, Association (for a State Manager association), Patch, or Custom:string are all valid compliance types.

", + "ComplianceSummaryItem$ComplianceType":"

The type of compliance item. For example, the compliance type can be Association, Patch, or Custom:string.

", + "PutComplianceItemsRequest$ComplianceType":"

Specify the compliance type. For example, specify Association (for a State Manager association), Patch, or Custom:string.

", + "ResourceComplianceSummaryItem$ComplianceType":"

The compliance type.

" } }, - "CreateActivationResult": { - "base": null, - "refs": { + "CompliantSummary":{ + "base":"

A summary of resources that are compliant. The summary is organized according to the resource count for each compliance type.

", + "refs":{ + "ComplianceSummaryItem$CompliantSummary":"

A list of COMPLIANT items for the specified compliance type.

", + "ResourceComplianceSummaryItem$CompliantSummary":"

A list of items that are compliant for the resource.

" } }, - "CreateAssociationBatchRequest": { - "base": null, - "refs": { + "ComputerName":{ + "base":null, + "refs":{ + "InstanceInformation$ComputerName":"

The fully qualified host name of the managed instance.

" } }, - "CreateAssociationBatchRequestEntries": { - "base": null, - "refs": { - "CreateAssociationBatchRequest$Entries": "

One or more associations.

" + "CreateActivationRequest":{ + "base":null, + "refs":{ } }, - "CreateAssociationBatchRequestEntry": { - "base": "

Describes the association of a Systems Manager document and an instance.

", - "refs": { - "CreateAssociationBatchRequestEntries$member": null, - "FailedCreateAssociation$Entry": "

The association.

" + "CreateActivationResult":{ + "base":null, + "refs":{ } }, - "CreateAssociationBatchResult": { - "base": null, - "refs": { + "CreateAssociationBatchRequest":{ + "base":null, + "refs":{ } }, - "CreateAssociationRequest": { - "base": null, - "refs": { + "CreateAssociationBatchRequestEntries":{ + "base":null, + "refs":{ + "CreateAssociationBatchRequest$Entries":"

One or more associations.

" } }, - "CreateAssociationResult": { - "base": null, - "refs": { + "CreateAssociationBatchRequestEntry":{ + "base":"

Describes the association of a Systems Manager document and an instance.

", + "refs":{ + "CreateAssociationBatchRequestEntries$member":null, + "FailedCreateAssociation$Entry":"

The association.

" } }, - "CreateDocumentRequest": { - "base": null, - "refs": { + "CreateAssociationBatchResult":{ + "base":null, + "refs":{ } }, - "CreateDocumentResult": { - "base": null, - "refs": { + "CreateAssociationRequest":{ + "base":null, + "refs":{ } }, - "CreateMaintenanceWindowRequest": { - "base": null, - "refs": { + "CreateAssociationResult":{ + "base":null, + "refs":{ } }, - "CreateMaintenanceWindowResult": { - "base": null, - "refs": { + "CreateDocumentRequest":{ + "base":null, + "refs":{ } }, - "CreatePatchBaselineRequest": { - "base": null, - "refs": { + "CreateDocumentResult":{ + "base":null, + "refs":{ } }, - "CreatePatchBaselineResult": { - "base": null, - "refs": { + "CreateMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "CreateResourceDataSyncRequest": { - "base": null, - "refs": { + "CreateMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "CreateResourceDataSyncResult": { - "base": null, - "refs": { + "CreatePatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "CreatedDate": { - "base": null, - "refs": { - "Activation$CreatedDate": "

The date the activation was created.

" + "CreatePatchBaselineResult":{ + "base":null, + "refs":{ } }, - "CustomSchemaCountLimitExceededException": { - "base": "

You have exceeded the limit for custom schemas. Delete one or more custom schemas and try again.

", - "refs": { + "CreateResourceDataSyncRequest":{ + "base":null, + "refs":{ } }, - "DateTime": { - "base": null, - "refs": { - "Association$LastExecutionDate": "

The date on which the association was last run.

", - "AssociationDescription$Date": "

The date when the association was made.

", - "AssociationDescription$LastUpdateAssociationDate": "

The date when the association was last updated.

", - "AssociationDescription$LastExecutionDate": "

The date on which the association was last run.

", - "AssociationDescription$LastSuccessfulExecutionDate": "

The last date on which the association was successfully run.

", - "AssociationStatus$Date": "

The date when the status changed.

", - "AssociationVersionInfo$CreatedDate": "

The date the association version was created.

", - "AutomationExecution$ExecutionStartTime": "

The time the execution started.

", - "AutomationExecution$ExecutionEndTime": "

The time the execution finished.

", - "AutomationExecutionMetadata$ExecutionStartTime": "

The time the execution started.>

", - "AutomationExecutionMetadata$ExecutionEndTime": "

The time the execution finished. This is not populated if the execution is still in progress.

", - "Command$ExpiresAfter": "

If this time is reached and the command has not already started executing, it will not run. Calculated based on the ExpiresAfter user input provided as part of the SendCommand API.

", - "Command$RequestedDateTime": "

The date and time the command was requested.

", - "CommandInvocation$RequestedDateTime": "

The time and date the request was sent to this instance.

", - "CommandPlugin$ResponseStartDateTime": "

The time the plugin started executing.

", - "CommandPlugin$ResponseFinishDateTime": "

The time the plugin stopped executing. Could stop prematurely if, for example, a cancel command was sent.

", - "ComplianceExecutionSummary$ExecutionTime": "

The time the execution ran as a datetime object that is saved in the following format: yyyy-MM-dd'T'HH:mm:ss'Z'.

", - "DocumentDescription$CreatedDate": "

The date when the document was created.

", - "DocumentVersionInfo$CreatedDate": "

The date the document was created.

", - "GetMaintenanceWindowExecutionResult$StartTime": "

The time the Maintenance Window started executing.

", - "GetMaintenanceWindowExecutionResult$EndTime": "

The time the Maintenance Window finished executing.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$StartTime": "

The time that the task started executing on the target.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$EndTime": "

The time that the task finished executing on the target.

", - "GetMaintenanceWindowExecutionTaskResult$StartTime": "

The time the task execution started.

", - "GetMaintenanceWindowExecutionTaskResult$EndTime": "

The time the task execution completed.

", - "GetMaintenanceWindowResult$CreatedDate": "

The date the Maintenance Window was created.

", - "GetMaintenanceWindowResult$ModifiedDate": "

The date the Maintenance Window was last modified.

", - "GetPatchBaselineResult$CreatedDate": "

The date the patch baseline was created.

", - "GetPatchBaselineResult$ModifiedDate": "

The date the patch baseline was last modified.

", - "InstanceAssociationStatusInfo$ExecutionDate": "

The date the instance association executed.

", - "InstanceInformation$LastPingDateTime": "

The date and time when agent last pinged Systems Manager service.

", - "InstanceInformation$RegistrationDate": "

The date the server or VM was registered with AWS as a managed instance.

", - "InstanceInformation$LastAssociationExecutionDate": "

The date the association was last executed.

", - "InstanceInformation$LastSuccessfulAssociationExecutionDate": "

The last date the association was successfully run.

", - "InstancePatchState$OperationStartTime": "

The time the most recent patching operation was started on the instance.

", - "InstancePatchState$OperationEndTime": "

The time the most recent patching operation completed on the instance.

", - "MaintenanceWindowExecution$StartTime": "

The time the execution started.

", - "MaintenanceWindowExecution$EndTime": "

The time the execution finished.

", - "MaintenanceWindowExecutionTaskIdentity$StartTime": "

The time the task execution started.

", - "MaintenanceWindowExecutionTaskIdentity$EndTime": "

The time the task execution finished.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$StartTime": "

The time the invocation started.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$EndTime": "

The time the invocation finished.

", - "ParameterHistory$LastModifiedDate": "

Date the parameter was last changed or updated.

", - "ParameterMetadata$LastModifiedDate": "

Date the parameter was last changed or updated.

", - "Patch$ReleaseDate": "

The date the patch was released.

", - "PatchComplianceData$InstalledTime": "

The date/time the patch was installed on the instance. Note that not all operating systems provide this level of information.

", - "PatchStatus$ApprovalDate": "

The date the patch was approved (or will be approved if the status is PENDING_APPROVAL).

", - "StepExecution$ExecutionStartTime": "

If a step has begun execution, this contains the time the step started. If the step is in Pending status, this field is not populated.

", - "StepExecution$ExecutionEndTime": "

If a step has finished execution, this contains the time the execution ended. If the step has not yet concluded, this field is not populated.

", - "UpdatePatchBaselineResult$CreatedDate": "

The date when the patch baseline was created.

", - "UpdatePatchBaselineResult$ModifiedDate": "

The date when the patch baseline was last modified.

" + "CreateResourceDataSyncResult":{ + "base":null, + "refs":{ } }, - "DefaultBaseline": { - "base": null, - "refs": { - "PatchBaselineIdentity$DefaultBaseline": "

Whether this is the default baseline. Note that Systems Manager supports creating multiple default patch baselines. For example, you can create a default patch baseline for each operating system.

" + "CreatedDate":{ + "base":null, + "refs":{ + "Activation$CreatedDate":"

The date the activation was created.

" } }, - "DefaultInstanceName": { - "base": null, - "refs": { - "Activation$DefaultInstanceName": "

A name for the managed instance when it is created.

", - "CreateActivationRequest$DefaultInstanceName": "

The name of the registered, managed instance as it will appear in the Amazon EC2 console or when you use the AWS command line tools to list EC2 resources.

Do not enter personally identifiable information in this field.

" + "CustomSchemaCountLimitExceededException":{ + "base":"

You have exceeded the limit for custom schemas. Delete one or more custom schemas and try again.

", + "refs":{ } }, - "DeleteActivationRequest": { - "base": null, - "refs": { + "DateTime":{ + "base":null, + "refs":{ + "Association$LastExecutionDate":"

The date on which the association was last run.

", + "AssociationDescription$Date":"

The date when the association was made.

", + "AssociationDescription$LastUpdateAssociationDate":"

The date when the association was last updated.

", + "AssociationDescription$LastExecutionDate":"

The date on which the association was last run.

", + "AssociationDescription$LastSuccessfulExecutionDate":"

The last date on which the association was successfully run.

", + "AssociationStatus$Date":"

The date when the status changed.

", + "AssociationVersionInfo$CreatedDate":"

The date the association version was created.

", + "AutomationExecution$ExecutionStartTime":"

The time the execution started.

", + "AutomationExecution$ExecutionEndTime":"

The time the execution finished.

", + "AutomationExecutionMetadata$ExecutionStartTime":"

The time the execution started.>

", + "AutomationExecutionMetadata$ExecutionEndTime":"

The time the execution finished. This is not populated if the execution is still in progress.

", + "Command$ExpiresAfter":"

If this time is reached and the command has not already started executing, it will not run. Calculated based on the ExpiresAfter user input provided as part of the SendCommand API.

", + "Command$RequestedDateTime":"

The date and time the command was requested.

", + "CommandInvocation$RequestedDateTime":"

The time and date the request was sent to this instance.

", + "CommandPlugin$ResponseStartDateTime":"

The time the plugin started executing.

", + "CommandPlugin$ResponseFinishDateTime":"

The time the plugin stopped executing. Could stop prematurely if, for example, a cancel command was sent.

", + "ComplianceExecutionSummary$ExecutionTime":"

The time the execution ran as a datetime object that is saved in the following format: yyyy-MM-dd'T'HH:mm:ss'Z'.

", + "DocumentDescription$CreatedDate":"

The date when the document was created.

", + "DocumentVersionInfo$CreatedDate":"

The date the document was created.

", + "GetMaintenanceWindowExecutionResult$StartTime":"

The time the Maintenance Window started executing.

", + "GetMaintenanceWindowExecutionResult$EndTime":"

The time the Maintenance Window finished executing.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$StartTime":"

The time that the task started executing on the target.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$EndTime":"

The time that the task finished executing on the target.

", + "GetMaintenanceWindowExecutionTaskResult$StartTime":"

The time the task execution started.

", + "GetMaintenanceWindowExecutionTaskResult$EndTime":"

The time the task execution completed.

", + "GetMaintenanceWindowResult$CreatedDate":"

The date the Maintenance Window was created.

", + "GetMaintenanceWindowResult$ModifiedDate":"

The date the Maintenance Window was last modified.

", + "GetPatchBaselineResult$CreatedDate":"

The date the patch baseline was created.

", + "GetPatchBaselineResult$ModifiedDate":"

The date the patch baseline was last modified.

", + "InstanceAssociationStatusInfo$ExecutionDate":"

The date the instance association executed.

", + "InstanceInformation$LastPingDateTime":"

The date and time when agent last pinged Systems Manager service.

", + "InstanceInformation$RegistrationDate":"

The date the server or VM was registered with AWS as a managed instance.

", + "InstanceInformation$LastAssociationExecutionDate":"

The date the association was last executed.

", + "InstanceInformation$LastSuccessfulAssociationExecutionDate":"

The last date the association was successfully run.

", + "InstancePatchState$OperationStartTime":"

The time the most recent patching operation was started on the instance.

", + "InstancePatchState$OperationEndTime":"

The time the most recent patching operation completed on the instance.

", + "MaintenanceWindowExecution$StartTime":"

The time the execution started.

", + "MaintenanceWindowExecution$EndTime":"

The time the execution finished.

", + "MaintenanceWindowExecutionTaskIdentity$StartTime":"

The time the task execution started.

", + "MaintenanceWindowExecutionTaskIdentity$EndTime":"

The time the task execution finished.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$StartTime":"

The time the invocation started.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$EndTime":"

The time the invocation finished.

", + "ParameterHistory$LastModifiedDate":"

Date the parameter was last changed or updated.

", + "ParameterMetadata$LastModifiedDate":"

Date the parameter was last changed or updated.

", + "Patch$ReleaseDate":"

The date the patch was released.

", + "PatchComplianceData$InstalledTime":"

The date/time the patch was installed on the instance. Note that not all operating systems provide this level of information.

", + "PatchStatus$ApprovalDate":"

The date the patch was approved (or will be approved if the status is PENDING_APPROVAL).

", + "StepExecution$ExecutionStartTime":"

If a step has begun execution, this contains the time the step started. If the step is in Pending status, this field is not populated.

", + "StepExecution$ExecutionEndTime":"

If a step has finished execution, this contains the time the execution ended. If the step has not yet concluded, this field is not populated.

", + "UpdatePatchBaselineResult$CreatedDate":"

The date when the patch baseline was created.

", + "UpdatePatchBaselineResult$ModifiedDate":"

The date when the patch baseline was last modified.

" } }, - "DeleteActivationResult": { - "base": null, - "refs": { + "DefaultBaseline":{ + "base":null, + "refs":{ + "PatchBaselineIdentity$DefaultBaseline":"

Whether this is the default baseline. Note that Systems Manager supports creating multiple default patch baselines. For example, you can create a default patch baseline for each operating system.

" } }, - "DeleteAssociationRequest": { - "base": null, - "refs": { + "DefaultInstanceName":{ + "base":null, + "refs":{ + "Activation$DefaultInstanceName":"

A name for the managed instance when it is created.

", + "CreateActivationRequest$DefaultInstanceName":"

The name of the registered, managed instance as it will appear in the Amazon EC2 console or when you use the AWS command line tools to list EC2 resources.

Do not enter personally identifiable information in this field.

" } }, - "DeleteAssociationResult": { - "base": null, - "refs": { + "DeleteActivationRequest":{ + "base":null, + "refs":{ } }, - "DeleteDocumentRequest": { - "base": null, - "refs": { + "DeleteActivationResult":{ + "base":null, + "refs":{ } }, - "DeleteDocumentResult": { - "base": null, - "refs": { + "DeleteAssociationRequest":{ + "base":null, + "refs":{ } }, - "DeleteInventoryRequest": { - "base": null, - "refs": { + "DeleteAssociationResult":{ + "base":null, + "refs":{ } }, - "DeleteInventoryResult": { - "base": null, - "refs": { + "DeleteDocumentRequest":{ + "base":null, + "refs":{ } }, - "DeleteMaintenanceWindowRequest": { - "base": null, - "refs": { + "DeleteDocumentResult":{ + "base":null, + "refs":{ } }, - "DeleteMaintenanceWindowResult": { - "base": null, - "refs": { + "DeleteInventoryRequest":{ + "base":null, + "refs":{ } }, - "DeleteParameterRequest": { - "base": null, - "refs": { + "DeleteInventoryResult":{ + "base":null, + "refs":{ } }, - "DeleteParameterResult": { - "base": null, - "refs": { + "DeleteMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "DeleteParametersRequest": { - "base": null, - "refs": { + "DeleteMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "DeleteParametersResult": { - "base": null, - "refs": { + "DeleteParameterRequest":{ + "base":null, + "refs":{ } }, - "DeletePatchBaselineRequest": { - "base": null, - "refs": { + "DeleteParameterResult":{ + "base":null, + "refs":{ } }, - "DeletePatchBaselineResult": { - "base": null, - "refs": { + "DeleteParametersRequest":{ + "base":null, + "refs":{ } }, - "DeleteResourceDataSyncRequest": { - "base": null, - "refs": { + "DeleteParametersResult":{ + "base":null, + "refs":{ } }, - "DeleteResourceDataSyncResult": { - "base": null, - "refs": { + "DeletePatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "DeregisterManagedInstanceRequest": { - "base": null, - "refs": { + "DeletePatchBaselineResult":{ + "base":null, + "refs":{ } }, - "DeregisterManagedInstanceResult": { - "base": null, - "refs": { + "DeleteResourceDataSyncRequest":{ + "base":null, + "refs":{ } }, - "DeregisterPatchBaselineForPatchGroupRequest": { - "base": null, - "refs": { + "DeleteResourceDataSyncResult":{ + "base":null, + "refs":{ } }, - "DeregisterPatchBaselineForPatchGroupResult": { - "base": null, - "refs": { + "DeliveryTimedOutCount":{ + "base":null, + "refs":{ + "Command$DeliveryTimedOutCount":"

The number of targets for which the status is Delivery Timed Out.

" } }, - "DeregisterTargetFromMaintenanceWindowRequest": { - "base": null, - "refs": { + "DeregisterManagedInstanceRequest":{ + "base":null, + "refs":{ } }, - "DeregisterTargetFromMaintenanceWindowResult": { - "base": null, - "refs": { + "DeregisterManagedInstanceResult":{ + "base":null, + "refs":{ } }, - "DeregisterTaskFromMaintenanceWindowRequest": { - "base": null, - "refs": { + "DeregisterPatchBaselineForPatchGroupRequest":{ + "base":null, + "refs":{ } }, - "DeregisterTaskFromMaintenanceWindowResult": { - "base": null, - "refs": { + "DeregisterPatchBaselineForPatchGroupResult":{ + "base":null, + "refs":{ } }, - "DescribeActivationsFilter": { - "base": "

Filter for the DescribeActivation API.

", - "refs": { - "DescribeActivationsFilterList$member": null + "DeregisterTargetFromMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "DescribeActivationsFilterKeys": { - "base": null, - "refs": { - "DescribeActivationsFilter$FilterKey": "

The name of the filter.

" + "DeregisterTargetFromMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "DescribeActivationsFilterList": { - "base": null, - "refs": { - "DescribeActivationsRequest$Filters": "

A filter to view information about your activations.

" + "DeregisterTaskFromMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "DescribeActivationsRequest": { - "base": null, - "refs": { + "DeregisterTaskFromMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "DescribeActivationsResult": { - "base": null, - "refs": { + "DescribeActivationsFilter":{ + "base":"

Filter for the DescribeActivation API.

", + "refs":{ + "DescribeActivationsFilterList$member":null } }, - "DescribeAssociationRequest": { - "base": null, - "refs": { + "DescribeActivationsFilterKeys":{ + "base":null, + "refs":{ + "DescribeActivationsFilter$FilterKey":"

The name of the filter.

" } }, - "DescribeAssociationResult": { - "base": null, - "refs": { + "DescribeActivationsFilterList":{ + "base":null, + "refs":{ + "DescribeActivationsRequest$Filters":"

A filter to view information about your activations.

" } }, - "DescribeAutomationExecutionsRequest": { - "base": null, - "refs": { + "DescribeActivationsRequest":{ + "base":null, + "refs":{ } }, - "DescribeAutomationExecutionsResult": { - "base": null, - "refs": { + "DescribeActivationsResult":{ + "base":null, + "refs":{ } }, - "DescribeAutomationStepExecutionsRequest": { - "base": null, - "refs": { + "DescribeAssociationRequest":{ + "base":null, + "refs":{ } }, - "DescribeAutomationStepExecutionsResult": { - "base": null, - "refs": { + "DescribeAssociationResult":{ + "base":null, + "refs":{ } }, - "DescribeAvailablePatchesRequest": { - "base": null, - "refs": { + "DescribeAutomationExecutionsRequest":{ + "base":null, + "refs":{ } }, - "DescribeAvailablePatchesResult": { - "base": null, - "refs": { + "DescribeAutomationExecutionsResult":{ + "base":null, + "refs":{ } }, - "DescribeDocumentPermissionRequest": { - "base": null, - "refs": { + "DescribeAutomationStepExecutionsRequest":{ + "base":null, + "refs":{ } }, - "DescribeDocumentPermissionResponse": { - "base": null, - "refs": { + "DescribeAutomationStepExecutionsResult":{ + "base":null, + "refs":{ } }, - "DescribeDocumentRequest": { - "base": null, - "refs": { + "DescribeAvailablePatchesRequest":{ + "base":null, + "refs":{ } }, - "DescribeDocumentResult": { - "base": null, - "refs": { + "DescribeAvailablePatchesResult":{ + "base":null, + "refs":{ } }, - "DescribeEffectiveInstanceAssociationsRequest": { - "base": null, - "refs": { + "DescribeDocumentPermissionRequest":{ + "base":null, + "refs":{ } }, - "DescribeEffectiveInstanceAssociationsResult": { - "base": null, - "refs": { + "DescribeDocumentPermissionResponse":{ + "base":null, + "refs":{ } }, - "DescribeEffectivePatchesForPatchBaselineRequest": { - "base": null, - "refs": { + "DescribeDocumentRequest":{ + "base":null, + "refs":{ } }, - "DescribeEffectivePatchesForPatchBaselineResult": { - "base": null, - "refs": { + "DescribeDocumentResult":{ + "base":null, + "refs":{ } }, - "DescribeInstanceAssociationsStatusRequest": { - "base": null, - "refs": { + "DescribeEffectiveInstanceAssociationsRequest":{ + "base":null, + "refs":{ } }, - "DescribeInstanceAssociationsStatusResult": { - "base": null, - "refs": { + "DescribeEffectiveInstanceAssociationsResult":{ + "base":null, + "refs":{ } }, - "DescribeInstanceInformationRequest": { - "base": null, - "refs": { + "DescribeEffectivePatchesForPatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "DescribeInstanceInformationResult": { - "base": null, - "refs": { + "DescribeEffectivePatchesForPatchBaselineResult":{ + "base":null, + "refs":{ } }, - "DescribeInstancePatchStatesForPatchGroupRequest": { - "base": null, - "refs": { + "DescribeInstanceAssociationsStatusRequest":{ + "base":null, + "refs":{ } }, - "DescribeInstancePatchStatesForPatchGroupResult": { - "base": null, - "refs": { + "DescribeInstanceAssociationsStatusResult":{ + "base":null, + "refs":{ } }, - "DescribeInstancePatchStatesRequest": { - "base": null, - "refs": { + "DescribeInstanceInformationRequest":{ + "base":null, + "refs":{ } }, - "DescribeInstancePatchStatesResult": { - "base": null, - "refs": { + "DescribeInstanceInformationResult":{ + "base":null, + "refs":{ } }, - "DescribeInstancePatchesRequest": { - "base": null, - "refs": { + "DescribeInstancePatchStatesForPatchGroupRequest":{ + "base":null, + "refs":{ } }, - "DescribeInstancePatchesResult": { - "base": null, - "refs": { + "DescribeInstancePatchStatesForPatchGroupResult":{ + "base":null, + "refs":{ } }, - "DescribeInventoryDeletionsRequest": { - "base": null, - "refs": { + "DescribeInstancePatchStatesRequest":{ + "base":null, + "refs":{ } }, - "DescribeInventoryDeletionsResult": { - "base": null, - "refs": { + "DescribeInstancePatchStatesResult":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowExecutionTaskInvocationsRequest": { - "base": null, - "refs": { + "DescribeInstancePatchesRequest":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowExecutionTaskInvocationsResult": { - "base": null, - "refs": { + "DescribeInstancePatchesResult":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowExecutionTasksRequest": { - "base": null, - "refs": { + "DescribeInventoryDeletionsRequest":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowExecutionTasksResult": { - "base": null, - "refs": { + "DescribeInventoryDeletionsResult":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowExecutionsRequest": { - "base": null, - "refs": { + "DescribeMaintenanceWindowExecutionTaskInvocationsRequest":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowExecutionsResult": { - "base": null, - "refs": { + "DescribeMaintenanceWindowExecutionTaskInvocationsResult":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowTargetsRequest": { - "base": null, - "refs": { + "DescribeMaintenanceWindowExecutionTasksRequest":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowTargetsResult": { - "base": null, - "refs": { + "DescribeMaintenanceWindowExecutionTasksResult":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowTasksRequest": { - "base": null, - "refs": { + "DescribeMaintenanceWindowExecutionsRequest":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowTasksResult": { - "base": null, - "refs": { + "DescribeMaintenanceWindowExecutionsResult":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowsRequest": { - "base": null, - "refs": { + "DescribeMaintenanceWindowTargetsRequest":{ + "base":null, + "refs":{ } }, - "DescribeMaintenanceWindowsResult": { - "base": null, - "refs": { + "DescribeMaintenanceWindowTargetsResult":{ + "base":null, + "refs":{ } }, - "DescribeParametersRequest": { - "base": null, - "refs": { + "DescribeMaintenanceWindowTasksRequest":{ + "base":null, + "refs":{ } }, - "DescribeParametersResult": { - "base": null, - "refs": { + "DescribeMaintenanceWindowTasksResult":{ + "base":null, + "refs":{ } }, - "DescribePatchBaselinesRequest": { - "base": null, - "refs": { + "DescribeMaintenanceWindowsRequest":{ + "base":null, + "refs":{ } }, - "DescribePatchBaselinesResult": { - "base": null, - "refs": { + "DescribeMaintenanceWindowsResult":{ + "base":null, + "refs":{ } }, - "DescribePatchGroupStateRequest": { - "base": null, - "refs": { + "DescribeParametersRequest":{ + "base":null, + "refs":{ } }, - "DescribePatchGroupStateResult": { - "base": null, - "refs": { + "DescribeParametersResult":{ + "base":null, + "refs":{ } }, - "DescribePatchGroupsRequest": { - "base": null, - "refs": { + "DescribePatchBaselinesRequest":{ + "base":null, + "refs":{ } }, - "DescribePatchGroupsResult": { - "base": null, - "refs": { + "DescribePatchBaselinesResult":{ + "base":null, + "refs":{ } }, - "DescriptionInDocument": { - "base": null, - "refs": { - "DocumentDescription$Description": "

A description of the document.

" + "DescribePatchGroupStateRequest":{ + "base":null, + "refs":{ } }, - "DocumentARN": { - "base": null, - "refs": { - "DescribeDocumentRequest$Name": "

The name of the Systems Manager document.

", - "DocumentDescription$Name": "

The name of the Systems Manager document.

", - "DocumentIdentifier$Name": "

The name of the Systems Manager document.

", - "GetDocumentRequest$Name": "

The name of the Systems Manager document.

", - "GetDocumentResult$Name": "

The name of the Systems Manager document.

", - "SendCommandRequest$DocumentName": "

Required. The name of the Systems Manager document to execute. This can be a public document or a custom document.

", - "StartAutomationExecutionRequest$DocumentName": "

The name of the Automation document to use for this execution.

" + "DescribePatchGroupStateResult":{ + "base":null, + "refs":{ } }, - "DocumentAlreadyExists": { - "base": "

The specified document already exists.

", - "refs": { + "DescribePatchGroupsRequest":{ + "base":null, + "refs":{ } }, - "DocumentContent": { - "base": null, - "refs": { - "CreateDocumentRequest$Content": "

A valid JSON or YAML string.

", - "GetDocumentResult$Content": "

The contents of the Systems Manager document.

", - "InstanceAssociation$Content": "

The content of the association document for the instance(s).

", - "UpdateDocumentRequest$Content": "

The content in a document that you want to update.

" + "DescribePatchGroupsResult":{ + "base":null, + "refs":{ } }, - "DocumentDefaultVersionDescription": { - "base": "

A default version of a document.

", - "refs": { - "UpdateDocumentDefaultVersionResult$Description": "

The description of a custom document that you want to set as the default version.

" + "DescriptionInDocument":{ + "base":null, + "refs":{ + "DocumentDescription$Description":"

A description of the document.

" } }, - "DocumentDescription": { - "base": "

Describes a Systems Manager document.

", - "refs": { - "CreateDocumentResult$DocumentDescription": "

Information about the Systems Manager document.

", - "DescribeDocumentResult$Document": "

Information about the Systems Manager document.

", - "UpdateDocumentResult$DocumentDescription": "

A description of the document that was updated.

" + "DocumentARN":{ + "base":null, + "refs":{ + "DescribeDocumentRequest$Name":"

The name of the Systems Manager document.

", + "DocumentDescription$Name":"

The name of the Systems Manager document.

", + "DocumentIdentifier$Name":"

The name of the Systems Manager document.

", + "GetDocumentRequest$Name":"

The name of the Systems Manager document.

", + "GetDocumentResult$Name":"

The name of the Systems Manager document.

", + "SendCommandRequest$DocumentName":"

Required. The name of the Systems Manager document to execute. This can be a public document or a custom document.

", + "StartAutomationExecutionRequest$DocumentName":"

The name of the Automation document to use for this execution.

" } }, - "DocumentFilter": { - "base": "

Describes a filter.

", - "refs": { - "DocumentFilterList$member": null + "DocumentAlreadyExists":{ + "base":"

The specified document already exists.

", + "refs":{ } }, - "DocumentFilterKey": { - "base": null, - "refs": { - "DocumentFilter$key": "

The name of the filter.

" + "DocumentContent":{ + "base":null, + "refs":{ + "CreateDocumentRequest$Content":"

A valid JSON or YAML string.

", + "GetDocumentResult$Content":"

The contents of the Systems Manager document.

", + "InstanceAssociation$Content":"

The content of the association document for the instance(s).

", + "UpdateDocumentRequest$Content":"

The content in a document that you want to update.

" } }, - "DocumentFilterList": { - "base": null, - "refs": { - "ListDocumentsRequest$DocumentFilterList": "

One or more filters. Use a filter to return a more specific list of results.

" + "DocumentDefaultVersionDescription":{ + "base":"

A default version of a document.

", + "refs":{ + "UpdateDocumentDefaultVersionResult$Description":"

The description of a custom document that you want to set as the default version.

" } }, - "DocumentFilterValue": { - "base": null, - "refs": { - "DocumentFilter$value": "

The value of the filter.

" + "DocumentDescription":{ + "base":"

Describes a Systems Manager document.

", + "refs":{ + "CreateDocumentResult$DocumentDescription":"

Information about the Systems Manager document.

", + "DescribeDocumentResult$Document":"

Information about the Systems Manager document.

", + "UpdateDocumentResult$DocumentDescription":"

A description of the document that was updated.

" } }, - "DocumentFormat": { - "base": null, - "refs": { - "CreateDocumentRequest$DocumentFormat": "

Specify the document format for the request. The document format can be either JSON or YAML. JSON is the default format.

", - "DocumentDescription$DocumentFormat": "

The document format, either JSON or YAML.

", - "DocumentIdentifier$DocumentFormat": "

The document format, either JSON or YAML.

", - "DocumentVersionInfo$DocumentFormat": "

The document format, either JSON or YAML.

", - "GetDocumentRequest$DocumentFormat": "

Returns the document in the specified format. The document format can be either JSON or YAML. JSON is the default format.

", - "GetDocumentResult$DocumentFormat": "

The document format, either JSON or YAML.

", - "UpdateDocumentRequest$DocumentFormat": "

Specify the document format for the new document version. Systems Manager supports JSON and YAML documents. JSON is the default format.

" + "DocumentFilter":{ + "base":"

Describes a filter.

", + "refs":{ + "DocumentFilterList$member":null } }, - "DocumentHash": { - "base": null, - "refs": { - "DocumentDescription$Hash": "

The Sha256 or Sha1 hash created by the system when the document was created.

Sha1 hashes have been deprecated.

", - "MaintenanceWindowRunCommandParameters$DocumentHash": "

The SHA-256 or SHA-1 hash created by the system when the document was created. SHA-1 hashes have been deprecated.

", - "SendCommandRequest$DocumentHash": "

The Sha256 or Sha1 hash created by the system when the document was created.

Sha1 hashes have been deprecated.

" + "DocumentFilterKey":{ + "base":null, + "refs":{ + "DocumentFilter$key":"

The name of the filter.

" } }, - "DocumentHashType": { - "base": null, - "refs": { - "DocumentDescription$HashType": "

Sha256 or Sha1.

Sha1 hashes have been deprecated.

", - "MaintenanceWindowRunCommandParameters$DocumentHashType": "

SHA-256 or SHA-1. SHA-1 hashes have been deprecated.

", - "SendCommandRequest$DocumentHashType": "

Sha256 or Sha1.

Sha1 hashes have been deprecated.

" + "DocumentFilterList":{ + "base":null, + "refs":{ + "ListDocumentsRequest$DocumentFilterList":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "DocumentIdentifier": { - "base": "

Describes the name of a Systems Manager document.

", - "refs": { - "DocumentIdentifierList$member": null + "DocumentFilterValue":{ + "base":null, + "refs":{ + "DocumentFilter$value":"

The value of the filter.

" } }, - "DocumentIdentifierList": { - "base": null, - "refs": { - "ListDocumentsResult$DocumentIdentifiers": "

The names of the Systems Manager documents.

" + "DocumentFormat":{ + "base":null, + "refs":{ + "CreateDocumentRequest$DocumentFormat":"

Specify the document format for the request. The document format can be either JSON or YAML. JSON is the default format.

", + "DocumentDescription$DocumentFormat":"

The document format, either JSON or YAML.

", + "DocumentIdentifier$DocumentFormat":"

The document format, either JSON or YAML.

", + "DocumentVersionInfo$DocumentFormat":"

The document format, either JSON or YAML.

", + "GetDocumentRequest$DocumentFormat":"

Returns the document in the specified format. The document format can be either JSON or YAML. JSON is the default format.

", + "GetDocumentResult$DocumentFormat":"

The document format, either JSON or YAML.

", + "UpdateDocumentRequest$DocumentFormat":"

Specify the document format for the new document version. Systems Manager supports JSON and YAML documents. JSON is the default format.

" } }, - "DocumentKeyValuesFilter": { - "base": "

One or more filters. Use a filter to return a more specific list of documents.

For keys, you can specify one or more tags that have been applied to a document.

Other valid values include Owner, Name, PlatformTypes, and DocumentType.

Note that only one Owner can be specified in a request. For example: Key=Owner,Values=Self.

If you use Name as a key, you can use a name prefix to return a list of documents. For example, in the AWS CLI, to return a list of all documents that begin with Te, run the following command:

aws ssm list-documents --filters Key=Name,Values=Te

If you specify more than two keys, only documents that are identified by all the tags are returned in the results. If you specify more than two values for a key, documents that are identified by any of the values are returned in the results.

To specify a custom key and value pair, use the format Key=tag:[tagName],Values=[valueName].

For example, if you created a Key called region and are using the AWS CLI to call the list-documents command:

aws ssm list-documents --filters Key=tag:region,Values=east,west Key=Owner,Values=Self

", - "refs": { - "DocumentKeyValuesFilterList$member": null + "DocumentHash":{ + "base":null, + "refs":{ + "DocumentDescription$Hash":"

The Sha256 or Sha1 hash created by the system when the document was created.

Sha1 hashes have been deprecated.

", + "MaintenanceWindowRunCommandParameters$DocumentHash":"

The SHA-256 or SHA-1 hash created by the system when the document was created. SHA-1 hashes have been deprecated.

", + "SendCommandRequest$DocumentHash":"

The Sha256 or Sha1 hash created by the system when the document was created.

Sha1 hashes have been deprecated.

" } }, - "DocumentKeyValuesFilterKey": { - "base": null, - "refs": { - "DocumentKeyValuesFilter$Key": "

The name of the filter key.

" + "DocumentHashType":{ + "base":null, + "refs":{ + "DocumentDescription$HashType":"

Sha256 or Sha1.

Sha1 hashes have been deprecated.

", + "MaintenanceWindowRunCommandParameters$DocumentHashType":"

SHA-256 or SHA-1. SHA-1 hashes have been deprecated.

", + "SendCommandRequest$DocumentHashType":"

Sha256 or Sha1.

Sha1 hashes have been deprecated.

" } }, - "DocumentKeyValuesFilterList": { - "base": null, - "refs": { - "ListDocumentsRequest$Filters": "

One or more filters. Use a filter to return a more specific list of results.

" + "DocumentIdentifier":{ + "base":"

Describes the name of a Systems Manager document.

", + "refs":{ + "DocumentIdentifierList$member":null } }, - "DocumentKeyValuesFilterValue": { - "base": null, - "refs": { - "DocumentKeyValuesFilterValues$member": null + "DocumentIdentifierList":{ + "base":null, + "refs":{ + "ListDocumentsResult$DocumentIdentifiers":"

The names of the Systems Manager documents.

" } }, - "DocumentKeyValuesFilterValues": { - "base": null, - "refs": { - "DocumentKeyValuesFilter$Values": "

The value for the filter key.

" + "DocumentKeyValuesFilter":{ + "base":"

One or more filters. Use a filter to return a more specific list of documents.

For keys, you can specify one or more tags that have been applied to a document.

Other valid values include Owner, Name, PlatformTypes, and DocumentType.

Note that only one Owner can be specified in a request. For example: Key=Owner,Values=Self.

If you use Name as a key, you can use a name prefix to return a list of documents. For example, in the AWS CLI, to return a list of all documents that begin with Te, run the following command:

aws ssm list-documents --filters Key=Name,Values=Te

If you specify more than two keys, only documents that are identified by all the tags are returned in the results. If you specify more than two values for a key, documents that are identified by any of the values are returned in the results.

To specify a custom key and value pair, use the format Key=tag:[tagName],Values=[valueName].

For example, if you created a Key called region and are using the AWS CLI to call the list-documents command:

aws ssm list-documents --filters Key=tag:region,Values=east,west Key=Owner,Values=Self

", + "refs":{ + "DocumentKeyValuesFilterList$member":null } }, - "DocumentLimitExceeded": { - "base": "

You can have at most 200 active Systems Manager documents.

", - "refs": { + "DocumentKeyValuesFilterKey":{ + "base":null, + "refs":{ + "DocumentKeyValuesFilter$Key":"

The name of the filter key.

" } }, - "DocumentName": { - "base": null, - "refs": { - "Association$Name": "

The name of the Systems Manager document.

", - "AssociationDescription$Name": "

The name of the Systems Manager document.

", - "AssociationVersionInfo$Name": "

The name specified when the association was created.

", - "AutomationExecution$DocumentName": "

The name of the Automation document used during the execution.

", - "AutomationExecutionMetadata$DocumentName": "

The name of the Automation document used during execution.

", - "Command$DocumentName": "

The name of the document requested for execution.

", - "CommandInvocation$DocumentName": "

The document name that was requested for execution.

", - "CreateAssociationBatchRequestEntry$Name": "

The name of the configuration document.

", - "CreateAssociationRequest$Name": "

The name of the Systems Manager document.

", - "CreateDocumentRequest$Name": "

A name for the Systems Manager document.

Do not use the following to begin the names of documents you create. They are reserved by AWS for use as document prefixes:

  • aws

  • amazon

  • amzn

", - "DeleteAssociationRequest$Name": "

The name of the Systems Manager document.

", - "DeleteDocumentRequest$Name": "

The name of the document.

", - "DescribeAssociationRequest$Name": "

The name of the Systems Manager document.

", - "DescribeDocumentPermissionRequest$Name": "

The name of the document for which you are the owner.

", - "DocumentDefaultVersionDescription$Name": "

The name of the document.

", - "DocumentVersionInfo$Name": "

The document name.

", - "GetCommandInvocationResult$DocumentName": "

The name of the document that was executed. For example, AWS-RunShellScript.

", - "InstanceAssociationStatusInfo$Name": "

The name of the association.

", - "ListDocumentVersionsRequest$Name": "

The name of the document about which you want version information.

", - "ModifyDocumentPermissionRequest$Name": "

The name of the document that you want to share.

", - "UpdateAssociationRequest$Name": "

The name of the association document.

", - "UpdateAssociationStatusRequest$Name": "

The name of the Systems Manager document.

", - "UpdateDocumentDefaultVersionRequest$Name": "

The name of a custom document that you want to set as the default version.

", - "UpdateDocumentRequest$Name": "

The name of the document that you want to update.

" + "DocumentKeyValuesFilterList":{ + "base":null, + "refs":{ + "ListDocumentsRequest$Filters":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "DocumentOwner": { - "base": null, - "refs": { - "DocumentDescription$Owner": "

The AWS user account that created the document.

", - "DocumentIdentifier$Owner": "

The AWS user account that created the document.

" + "DocumentKeyValuesFilterValue":{ + "base":null, + "refs":{ + "DocumentKeyValuesFilterValues$member":null } }, - "DocumentParameter": { - "base": "

Parameters specified in a System Manager document that execute on the server when the command is run.

", - "refs": { - "DocumentParameterList$member": null - } + "DocumentKeyValuesFilterValues":{ + "base":null, + "refs":{ + "DocumentKeyValuesFilter$Values":"

The value for the filter key.

" + } + }, + "DocumentLimitExceeded":{ + "base":"

You can have at most 200 active Systems Manager documents.

", + "refs":{ + } + }, + "DocumentName":{ + "base":null, + "refs":{ + "Association$Name":"

The name of the Systems Manager document.

", + "AssociationDescription$Name":"

The name of the Systems Manager document.

", + "AssociationVersionInfo$Name":"

The name specified when the association was created.

", + "AutomationExecution$DocumentName":"

The name of the Automation document used during the execution.

", + "AutomationExecutionMetadata$DocumentName":"

The name of the Automation document used during execution.

", + "Command$DocumentName":"

The name of the document requested for execution.

", + "CommandInvocation$DocumentName":"

The document name that was requested for execution.

", + "CreateAssociationBatchRequestEntry$Name":"

The name of the configuration document.

", + "CreateAssociationRequest$Name":"

The name of the Systems Manager document.

", + "CreateDocumentRequest$Name":"

A name for the Systems Manager document.

Do not use the following to begin the names of documents you create. They are reserved by AWS for use as document prefixes:

  • aws

  • amazon

  • amzn

", + "DeleteAssociationRequest$Name":"

The name of the Systems Manager document.

", + "DeleteDocumentRequest$Name":"

The name of the document.

", + "DescribeAssociationRequest$Name":"

The name of the Systems Manager document.

", + "DescribeDocumentPermissionRequest$Name":"

The name of the document for which you are the owner.

", + "DocumentDefaultVersionDescription$Name":"

The name of the document.

", + "DocumentVersionInfo$Name":"

The document name.

", + "GetCommandInvocationResult$DocumentName":"

The name of the document that was executed. For example, AWS-RunShellScript.

", + "InstanceAssociationStatusInfo$Name":"

The name of the association.

", + "ListDocumentVersionsRequest$Name":"

The name of the document about which you want version information.

", + "ModifyDocumentPermissionRequest$Name":"

The name of the document that you want to share.

", + "UpdateAssociationRequest$Name":"

The name of the association document.

", + "UpdateAssociationStatusRequest$Name":"

The name of the Systems Manager document.

", + "UpdateDocumentDefaultVersionRequest$Name":"

The name of a custom document that you want to set as the default version.

", + "UpdateDocumentRequest$Name":"

The name of the document that you want to update.

" + } + }, + "DocumentOwner":{ + "base":null, + "refs":{ + "DocumentDescription$Owner":"

The AWS user account that created the document.

", + "DocumentIdentifier$Owner":"

The AWS user account that created the document.

" + } + }, + "DocumentParameter":{ + "base":"

Parameters specified in a System Manager document that execute on the server when the command is run.

", + "refs":{ + "DocumentParameterList$member":null + } }, - "DocumentParameterDefaultValue": { - "base": null, - "refs": { - "DocumentParameter$DefaultValue": "

If specified, the default values for the parameters. Parameters without a default value are required. Parameters with a default value are optional.

" - } + "DocumentParameterDefaultValue":{ + "base":null, + "refs":{ + "DocumentParameter$DefaultValue":"

If specified, the default values for the parameters. Parameters without a default value are required. Parameters with a default value are optional.

" + } }, - "DocumentParameterDescrption": { - "base": null, - "refs": { - "DocumentParameter$Description": "

A description of what the parameter does, how to use it, the default value, and whether or not the parameter is optional.

" + "DocumentParameterDescrption":{ + "base":null, + "refs":{ + "DocumentParameter$Description":"

A description of what the parameter does, how to use it, the default value, and whether or not the parameter is optional.

" } }, - "DocumentParameterList": { - "base": null, - "refs": { - "DocumentDescription$Parameters": "

A description of the parameters for a document.

" + "DocumentParameterList":{ + "base":null, + "refs":{ + "DocumentDescription$Parameters":"

A description of the parameters for a document.

" } }, - "DocumentParameterName": { - "base": null, - "refs": { - "DocumentParameter$Name": "

The name of the parameter.

" + "DocumentParameterName":{ + "base":null, + "refs":{ + "DocumentParameter$Name":"

The name of the parameter.

" } }, - "DocumentParameterType": { - "base": null, - "refs": { - "DocumentParameter$Type": "

The type of parameter. The type can be either String or StringList.

" + "DocumentParameterType":{ + "base":null, + "refs":{ + "DocumentParameter$Type":"

The type of parameter. The type can be either String or StringList.

" } }, - "DocumentPermissionLimit": { - "base": "

The document cannot be shared with more AWS user accounts. You can share a document with a maximum of 20 accounts. You can publicly share up to five documents. If you need to increase this limit, contact AWS Support.

", - "refs": { + "DocumentPermissionLimit":{ + "base":"

The document cannot be shared with more AWS user accounts. You can share a document with a maximum of 20 accounts. You can publicly share up to five documents. If you need to increase this limit, contact AWS Support.

", + "refs":{ } }, - "DocumentPermissionType": { - "base": null, - "refs": { - "DescribeDocumentPermissionRequest$PermissionType": "

The permission type for the document. The permission type can be Share.

", - "ModifyDocumentPermissionRequest$PermissionType": "

The permission type for the document. The permission type can be Share.

" + "DocumentPermissionType":{ + "base":null, + "refs":{ + "DescribeDocumentPermissionRequest$PermissionType":"

The permission type for the document. The permission type can be Share.

", + "ModifyDocumentPermissionRequest$PermissionType":"

The permission type for the document. The permission type can be Share.

" } }, - "DocumentSchemaVersion": { - "base": null, - "refs": { - "DocumentDescription$SchemaVersion": "

The schema version.

", - "DocumentIdentifier$SchemaVersion": "

The schema version.

" + "DocumentSchemaVersion":{ + "base":null, + "refs":{ + "DocumentDescription$SchemaVersion":"

The schema version.

", + "DocumentIdentifier$SchemaVersion":"

The schema version.

" } }, - "DocumentSha1": { - "base": null, - "refs": { - "DocumentDescription$Sha1": "

The SHA1 hash of the document, which you can use for verification.

" + "DocumentSha1":{ + "base":null, + "refs":{ + "DocumentDescription$Sha1":"

The SHA1 hash of the document, which you can use for verification.

" } }, - "DocumentStatus": { - "base": null, - "refs": { - "DocumentDescription$Status": "

The status of the Systems Manager document.

" + "DocumentStatus":{ + "base":null, + "refs":{ + "DocumentDescription$Status":"

The status of the Systems Manager document.

" } }, - "DocumentType": { - "base": null, - "refs": { - "CreateDocumentRequest$DocumentType": "

The type of document to create. Valid document types include: Policy, Automation, and Command.

", - "DocumentDescription$DocumentType": "

The type of document.

", - "DocumentIdentifier$DocumentType": "

The document type.

", - "GetDocumentResult$DocumentType": "

The document type.

" + "DocumentType":{ + "base":null, + "refs":{ + "CreateDocumentRequest$DocumentType":"

The type of document to create. Valid document types include: Policy, Automation, and Command.

", + "DocumentDescription$DocumentType":"

The type of document.

", + "DocumentIdentifier$DocumentType":"

The document type.

", + "GetDocumentResult$DocumentType":"

The document type.

" } }, - "DocumentVersion": { - "base": null, - "refs": { - "Association$DocumentVersion": "

The version of the document used in the association.

", - "AssociationDescription$DocumentVersion": "

The document version.

", - "AssociationVersionInfo$DocumentVersion": "

The version of a Systems Manager document used when the association version was created.

", - "AutomationExecution$DocumentVersion": "

The version of the document to use during execution.

", - "AutomationExecutionMetadata$DocumentVersion": "

The document version used during the execution.

", - "Command$DocumentVersion": "

The SSM document version.

", - "CommandInvocation$DocumentVersion": "

The SSM document version.

", - "CreateAssociationBatchRequestEntry$DocumentVersion": "

The document version.

", - "CreateAssociationRequest$DocumentVersion": "

The document version you want to associate with the target(s). Can be a specific version or the default version.

", - "DescribeDocumentRequest$DocumentVersion": "

The document version for which you want information. Can be a specific version or the default version.

", - "DocumentDefaultVersionDescription$DefaultVersion": "

The default version of the document.

", - "DocumentDescription$DocumentVersion": "

The document version.

", - "DocumentDescription$LatestVersion": "

The latest version of the document.

", - "DocumentDescription$DefaultVersion": "

The default version.

", - "DocumentIdentifier$DocumentVersion": "

The document version.

", - "DocumentVersionInfo$DocumentVersion": "

The document version.

", - "GetCommandInvocationResult$DocumentVersion": "

The SSM document version used in the request.

", - "GetDocumentRequest$DocumentVersion": "

The document version for which you want information.

", - "GetDocumentResult$DocumentVersion": "

The document version.

", - "InstanceAssociationStatusInfo$DocumentVersion": "

The association document verions.

", - "MaintenanceWindowAutomationParameters$DocumentVersion": "

The version of an Automation document to use during task execution.

", - "SendCommandRequest$DocumentVersion": "

The SSM document version to use in the request. You can specify Default, Latest, or a specific version number.

", - "StartAutomationExecutionRequest$DocumentVersion": "

The version of the Automation document to use for this execution.

", - "UpdateAssociationRequest$DocumentVersion": "

The document version you want update for the association.

", - "UpdateDocumentRequest$DocumentVersion": "

The version of the document that you want to update.

" + "DocumentVersion":{ + "base":null, + "refs":{ + "Association$DocumentVersion":"

The version of the document used in the association.

", + "AssociationDescription$DocumentVersion":"

The document version.

", + "AssociationVersionInfo$DocumentVersion":"

The version of a Systems Manager document used when the association version was created.

", + "AutomationExecution$DocumentVersion":"

The version of the document to use during execution.

", + "AutomationExecutionMetadata$DocumentVersion":"

The document version used during the execution.

", + "Command$DocumentVersion":"

The SSM document version.

", + "CommandInvocation$DocumentVersion":"

The SSM document version.

", + "CreateAssociationBatchRequestEntry$DocumentVersion":"

The document version.

", + "CreateAssociationRequest$DocumentVersion":"

The document version you want to associate with the target(s). Can be a specific version or the default version.

", + "DescribeDocumentRequest$DocumentVersion":"

The document version for which you want information. Can be a specific version or the default version.

", + "DocumentDefaultVersionDescription$DefaultVersion":"

The default version of the document.

", + "DocumentDescription$DocumentVersion":"

The document version.

", + "DocumentDescription$LatestVersion":"

The latest version of the document.

", + "DocumentDescription$DefaultVersion":"

The default version.

", + "DocumentIdentifier$DocumentVersion":"

The document version.

", + "DocumentVersionInfo$DocumentVersion":"

The document version.

", + "GetCommandInvocationResult$DocumentVersion":"

The SSM document version used in the request.

", + "GetDocumentRequest$DocumentVersion":"

The document version for which you want information.

", + "GetDocumentResult$DocumentVersion":"

The document version.

", + "InstanceAssociationStatusInfo$DocumentVersion":"

The association document verions.

", + "MaintenanceWindowAutomationParameters$DocumentVersion":"

The version of an Automation document to use during task execution.

", + "SendCommandRequest$DocumentVersion":"

The SSM document version to use in the request. You can specify $DEFAULT, $LATEST, or a specific version number. If you execute commands by using the AWS CLI, then you must escape the first two options by using a backslash. If you specify a version number, then you don't need to use the backslash. For example:

--document-version \"\\$DEFAULT\"

--document-version \"\\$LATEST\"

--document-version \"3\"

", + "StartAutomationExecutionRequest$DocumentVersion":"

The version of the Automation document to use for this execution.

", + "UpdateAssociationRequest$DocumentVersion":"

The document version you want update for the association.

", + "UpdateDocumentRequest$DocumentVersion":"

The version of the document that you want to update.

" } }, - "DocumentVersionInfo": { - "base": "

Version information about the document.

", - "refs": { - "DocumentVersionList$member": null + "DocumentVersionInfo":{ + "base":"

Version information about the document.

", + "refs":{ + "DocumentVersionList$member":null } }, - "DocumentVersionLimitExceeded": { - "base": "

The document has too many versions. Delete one or more document versions and try again.

", - "refs": { + "DocumentVersionLimitExceeded":{ + "base":"

The document has too many versions. Delete one or more document versions and try again.

", + "refs":{ } }, - "DocumentVersionList": { - "base": null, - "refs": { - "ListDocumentVersionsResult$DocumentVersions": "

The document versions.

" + "DocumentVersionList":{ + "base":null, + "refs":{ + "ListDocumentVersionsResult$DocumentVersions":"

The document versions.

" } }, - "DocumentVersionNumber": { - "base": null, - "refs": { - "UpdateDocumentDefaultVersionRequest$DocumentVersion": "

The version of a custom document that you want to set as the default version.

" + "DocumentVersionNumber":{ + "base":null, + "refs":{ + "UpdateDocumentDefaultVersionRequest$DocumentVersion":"

The version of a custom document that you want to set as the default version.

" } }, - "DoesNotExistException": { - "base": "

Error returned when the ID specified for a resource, such as a Maintenance Window or Patch baseline, doesn't exist.

For information about resource limits in Systems Manager, see AWS Systems Manager Limits.

", - "refs": { + "DoesNotExistException":{ + "base":"

Error returned when the ID specified for a resource, such as a Maintenance Window or Patch baseline, doesn't exist.

For information about resource limits in Systems Manager, see AWS Systems Manager Limits.

", + "refs":{ } }, - "DryRun": { - "base": null, - "refs": { - "DeleteInventoryRequest$DryRun": "

Use this option to view a summary of the deletion request without deleting any data or the data type. This option is useful when you only want to understand what will be deleted. Once you validate that the data to be deleted is what you intend to delete, you can run the same command without specifying the DryRun option.

" + "DryRun":{ + "base":null, + "refs":{ + "DeleteInventoryRequest$DryRun":"

Use this option to view a summary of the deletion request without deleting any data or the data type. This option is useful when you only want to understand what will be deleted. Once you validate that the data to be deleted is what you intend to delete, you can run the same command without specifying the DryRun option.

" } }, - "DuplicateDocumentContent": { - "base": "

The content of the association document matches another document. Change the content of the document and try again.

", - "refs": { + "DuplicateDocumentContent":{ + "base":"

The content of the association document matches another document. Change the content of the document and try again.

", + "refs":{ } }, - "DuplicateInstanceId": { - "base": "

You cannot specify an instance ID in more than one association.

", - "refs": { + "DuplicateInstanceId":{ + "base":"

You cannot specify an instance ID in more than one association.

", + "refs":{ } }, - "EffectiveInstanceAssociationMaxResults": { - "base": null, - "refs": { - "DescribeEffectiveInstanceAssociationsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + "EffectiveInstanceAssociationMaxResults":{ + "base":null, + "refs":{ + "DescribeEffectiveInstanceAssociationsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" } }, - "EffectivePatch": { - "base": "

The EffectivePatch structure defines metadata about a patch along with the approval state of the patch in a particular patch baseline. The approval state includes information about whether the patch is currently approved, due to be approved by a rule, explicitly approved, or explicitly rejected and the date the patch was or will be approved.

", - "refs": { - "EffectivePatchList$member": null + "EffectivePatch":{ + "base":"

The EffectivePatch structure defines metadata about a patch along with the approval state of the patch in a particular patch baseline. The approval state includes information about whether the patch is currently approved, due to be approved by a rule, explicitly approved, or explicitly rejected and the date the patch was or will be approved.

", + "refs":{ + "EffectivePatchList$member":null } }, - "EffectivePatchList": { - "base": null, - "refs": { - "DescribeEffectivePatchesForPatchBaselineResult$EffectivePatches": "

An array of patches and patch status.

" + "EffectivePatchList":{ + "base":null, + "refs":{ + "DescribeEffectivePatchesForPatchBaselineResult$EffectivePatches":"

An array of patches and patch status.

" } }, - "ErrorCount": { - "base": null, - "refs": { - "Command$ErrorCount": "

The number of targets for which the status is Failed or Execution Timed Out.

" + "ErrorCount":{ + "base":null, + "refs":{ + "Command$ErrorCount":"

The number of targets for which the status is Failed or Execution Timed Out.

" } }, - "ExecutionMode": { - "base": null, - "refs": { - "AutomationExecution$Mode": "

The automation execution mode.

", - "AutomationExecutionMetadata$Mode": "

The Automation execution mode.

", - "StartAutomationExecutionRequest$Mode": "

The execution mode of the automation. Valid modes include the following: Auto and Interactive. The default mode is Auto.

" + "ExecutionMode":{ + "base":null, + "refs":{ + "AutomationExecution$Mode":"

The automation execution mode.

", + "AutomationExecutionMetadata$Mode":"

The Automation execution mode.

", + "StartAutomationExecutionRequest$Mode":"

The execution mode of the automation. Valid modes include the following: Auto and Interactive. The default mode is Auto.

" } }, - "ExpirationDate": { - "base": null, - "refs": { - "Activation$ExpirationDate": "

The date when this activation can no longer be used to register managed instances.

", - "CreateActivationRequest$ExpirationDate": "

The date by which this activation request should expire. The default value is 24 hours.

" + "ExpirationDate":{ + "base":null, + "refs":{ + "Activation$ExpirationDate":"

The date when this activation can no longer be used to register managed instances.

", + "CreateActivationRequest$ExpirationDate":"

The date by which this activation request should expire. The default value is 24 hours.

" } }, - "FailedCreateAssociation": { - "base": "

Describes a failed association.

", - "refs": { - "FailedCreateAssociationList$member": null + "FailedCreateAssociation":{ + "base":"

Describes a failed association.

", + "refs":{ + "FailedCreateAssociationList$member":null } }, - "FailedCreateAssociationList": { - "base": null, - "refs": { - "CreateAssociationBatchResult$Failed": "

Information about the associations that failed.

" + "FailedCreateAssociationList":{ + "base":null, + "refs":{ + "CreateAssociationBatchResult$Failed":"

Information about the associations that failed.

" } }, - "FailureDetails": { - "base": "

Information about an Automation failure.

", - "refs": { - "StepExecution$FailureDetails": "

Information about the Automation failure.

" + "FailureDetails":{ + "base":"

Information about an Automation failure.

", + "refs":{ + "StepExecution$FailureDetails":"

Information about the Automation failure.

" } }, - "Fault": { - "base": null, - "refs": { - "FailedCreateAssociation$Fault": "

The source of the failure.

" + "Fault":{ + "base":null, + "refs":{ + "FailedCreateAssociation$Fault":"

The source of the failure.

" } }, - "FeatureNotAvailableException": { - "base": "

You attempted to register a LAMBDA or STEP_FUNCTION task in a region where the corresponding service is not available.

", - "refs": { + "FeatureNotAvailableException":{ + "base":"

You attempted to register a LAMBDA or STEP_FUNCTION task in a region where the corresponding service is not available.

", + "refs":{ } }, - "GetAutomationExecutionRequest": { - "base": null, - "refs": { + "GetAutomationExecutionRequest":{ + "base":null, + "refs":{ } }, - "GetAutomationExecutionResult": { - "base": null, - "refs": { + "GetAutomationExecutionResult":{ + "base":null, + "refs":{ } }, - "GetCommandInvocationRequest": { - "base": null, - "refs": { + "GetCommandInvocationRequest":{ + "base":null, + "refs":{ } }, - "GetCommandInvocationResult": { - "base": null, - "refs": { + "GetCommandInvocationResult":{ + "base":null, + "refs":{ } }, - "GetDefaultPatchBaselineRequest": { - "base": null, - "refs": { + "GetDefaultPatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "GetDefaultPatchBaselineResult": { - "base": null, - "refs": { + "GetDefaultPatchBaselineResult":{ + "base":null, + "refs":{ } }, - "GetDeployablePatchSnapshotForInstanceRequest": { - "base": null, - "refs": { + "GetDeployablePatchSnapshotForInstanceRequest":{ + "base":null, + "refs":{ } }, - "GetDeployablePatchSnapshotForInstanceResult": { - "base": null, - "refs": { + "GetDeployablePatchSnapshotForInstanceResult":{ + "base":null, + "refs":{ } }, - "GetDocumentRequest": { - "base": null, - "refs": { + "GetDocumentRequest":{ + "base":null, + "refs":{ } }, - "GetDocumentResult": { - "base": null, - "refs": { + "GetDocumentResult":{ + "base":null, + "refs":{ } }, - "GetInventoryRequest": { - "base": null, - "refs": { + "GetInventoryRequest":{ + "base":null, + "refs":{ } }, - "GetInventoryResult": { - "base": null, - "refs": { + "GetInventoryResult":{ + "base":null, + "refs":{ } }, - "GetInventorySchemaMaxResults": { - "base": null, - "refs": { - "GetInventorySchemaRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + "GetInventorySchemaMaxResults":{ + "base":null, + "refs":{ + "GetInventorySchemaRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" } }, - "GetInventorySchemaRequest": { - "base": null, - "refs": { + "GetInventorySchemaRequest":{ + "base":null, + "refs":{ } }, - "GetInventorySchemaResult": { - "base": null, - "refs": { + "GetInventorySchemaResult":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowExecutionRequest": { - "base": null, - "refs": { + "GetMaintenanceWindowExecutionRequest":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowExecutionResult": { - "base": null, - "refs": { + "GetMaintenanceWindowExecutionResult":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowExecutionTaskInvocationRequest": { - "base": null, - "refs": { + "GetMaintenanceWindowExecutionTaskInvocationRequest":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowExecutionTaskInvocationResult": { - "base": null, - "refs": { + "GetMaintenanceWindowExecutionTaskInvocationResult":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowExecutionTaskRequest": { - "base": null, - "refs": { + "GetMaintenanceWindowExecutionTaskRequest":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowExecutionTaskResult": { - "base": null, - "refs": { + "GetMaintenanceWindowExecutionTaskResult":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowRequest": { - "base": null, - "refs": { + "GetMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowResult": { - "base": null, - "refs": { + "GetMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowTaskRequest": { - "base": null, - "refs": { + "GetMaintenanceWindowTaskRequest":{ + "base":null, + "refs":{ } }, - "GetMaintenanceWindowTaskResult": { - "base": null, - "refs": { + "GetMaintenanceWindowTaskResult":{ + "base":null, + "refs":{ } }, - "GetParameterHistoryRequest": { - "base": null, - "refs": { + "GetParameterHistoryRequest":{ + "base":null, + "refs":{ } }, - "GetParameterHistoryResult": { - "base": null, - "refs": { + "GetParameterHistoryResult":{ + "base":null, + "refs":{ } }, - "GetParameterRequest": { - "base": null, - "refs": { + "GetParameterRequest":{ + "base":null, + "refs":{ } }, - "GetParameterResult": { - "base": null, - "refs": { + "GetParameterResult":{ + "base":null, + "refs":{ } }, - "GetParametersByPathMaxResults": { - "base": null, - "refs": { - "GetParametersByPathRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + "GetParametersByPathMaxResults":{ + "base":null, + "refs":{ + "GetParametersByPathRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" } }, - "GetParametersByPathRequest": { - "base": null, - "refs": { + "GetParametersByPathRequest":{ + "base":null, + "refs":{ } }, - "GetParametersByPathResult": { - "base": null, - "refs": { + "GetParametersByPathResult":{ + "base":null, + "refs":{ } }, - "GetParametersRequest": { - "base": null, - "refs": { + "GetParametersRequest":{ + "base":null, + "refs":{ } }, - "GetParametersResult": { - "base": null, - "refs": { + "GetParametersResult":{ + "base":null, + "refs":{ } }, - "GetPatchBaselineForPatchGroupRequest": { - "base": null, - "refs": { + "GetPatchBaselineForPatchGroupRequest":{ + "base":null, + "refs":{ } }, - "GetPatchBaselineForPatchGroupResult": { - "base": null, - "refs": { + "GetPatchBaselineForPatchGroupResult":{ + "base":null, + "refs":{ } }, - "GetPatchBaselineRequest": { - "base": null, - "refs": { + "GetPatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "GetPatchBaselineResult": { - "base": null, - "refs": { + "GetPatchBaselineResult":{ + "base":null, + "refs":{ } }, - "HierarchyLevelLimitExceededException": { - "base": "

A hierarchy can have a maximum of 15 levels. For more information, see Working with Systems Manager Parameters.

", - "refs": { + "HierarchyLevelLimitExceededException":{ + "base":"

A hierarchy can have a maximum of 15 levels. For more information, see Working with Systems Manager Parameters in the AWS Systems Manager User Guide.

", + "refs":{ } }, - "HierarchyTypeMismatchException": { - "base": "

Parameter Store does not support changing a parameter type in a hierarchy. For example, you can't change a parameter from a String type to a SecureString type. You must create a new, unique parameter.

", - "refs": { + "HierarchyTypeMismatchException":{ + "base":"

Parameter Store does not support changing a parameter type in a hierarchy. For example, you can't change a parameter from a String type to a SecureString type. You must create a new, unique parameter.

", + "refs":{ } }, - "IPAddress": { - "base": null, - "refs": { - "InstanceInformation$IPAddress": "

The IP address of the managed instance.

" + "IPAddress":{ + "base":null, + "refs":{ + "InstanceInformation$IPAddress":"

The IP address of the managed instance.

" } }, - "IamRole": { - "base": null, - "refs": { - "Activation$IamRole": "

The Amazon Identity and Access Management (IAM) role to assign to the managed instance.

", - "CreateActivationRequest$IamRole": "

The Amazon Identity and Access Management (IAM) role that you want to assign to the managed instance.

", - "InstanceInformation$IamRole": "

The Amazon Identity and Access Management (IAM) role assigned to EC2 instances or managed instances.

", - "UpdateManagedInstanceRoleRequest$IamRole": "

The IAM role you want to assign or change.

" + "IamRole":{ + "base":null, + "refs":{ + "Activation$IamRole":"

The Amazon Identity and Access Management (IAM) role to assign to the managed instance.

", + "CreateActivationRequest$IamRole":"

The Amazon Identity and Access Management (IAM) role that you want to assign to the managed instance.

", + "InstanceInformation$IamRole":"

The Amazon Identity and Access Management (IAM) role assigned to the on-premises Systems Manager managed instances. This call does not return the IAM role for Amazon EC2 instances.

", + "UpdateManagedInstanceRoleRequest$IamRole":"

The IAM role you want to assign or change.

" } }, - "IdempotencyToken": { - "base": null, - "refs": { - "StartAutomationExecutionRequest$ClientToken": "

User-provided idempotency token. The token must be unique, is case insensitive, enforces the UUID format, and can't be reused.

" + "IdempotencyToken":{ + "base":null, + "refs":{ + "StartAutomationExecutionRequest$ClientToken":"

User-provided idempotency token. The token must be unique, is case insensitive, enforces the UUID format, and can't be reused.

" } }, - "IdempotentParameterMismatch": { - "base": "

Error returned when an idempotent operation is retried and the parameters don't match the original call to the API with the same idempotency token.

", - "refs": { + "IdempotentParameterMismatch":{ + "base":"

Error returned when an idempotent operation is retried and the parameters don't match the original call to the API with the same idempotency token.

", + "refs":{ } }, - "InstanceAggregatedAssociationOverview": { - "base": "

Status information about the aggregated associations.

", - "refs": { - "InstanceInformation$AssociationOverview": "

Information about the association.

" + "InstanceAggregatedAssociationOverview":{ + "base":"

Status information about the aggregated associations.

", + "refs":{ + "InstanceInformation$AssociationOverview":"

Information about the association.

" } }, - "InstanceAssociation": { - "base": "

One or more association documents on the instance.

", - "refs": { - "InstanceAssociationList$member": null + "InstanceAssociation":{ + "base":"

One or more association documents on the instance.

", + "refs":{ + "InstanceAssociationList$member":null } }, - "InstanceAssociationExecutionSummary": { - "base": null, - "refs": { - "InstanceAssociationStatusInfo$ExecutionSummary": "

Summary information about association execution.

" + "InstanceAssociationExecutionSummary":{ + "base":null, + "refs":{ + "InstanceAssociationStatusInfo$ExecutionSummary":"

Summary information about association execution.

" } }, - "InstanceAssociationList": { - "base": null, - "refs": { - "DescribeEffectiveInstanceAssociationsResult$Associations": "

The associations for the requested instance.

" + "InstanceAssociationList":{ + "base":null, + "refs":{ + "DescribeEffectiveInstanceAssociationsResult$Associations":"

The associations for the requested instance.

" } }, - "InstanceAssociationOutputLocation": { - "base": "

An Amazon S3 bucket where you want to store the results of this request.

", - "refs": { - "AssociationDescription$OutputLocation": "

An Amazon S3 bucket where you want to store the output details of the request.

", - "AssociationVersionInfo$OutputLocation": "

The location in Amazon S3 specified for the association when the association version was created.

", - "CreateAssociationBatchRequestEntry$OutputLocation": "

An Amazon S3 bucket where you want to store the results of this request.

", - "CreateAssociationRequest$OutputLocation": "

An Amazon S3 bucket where you want to store the output details of the request.

", - "UpdateAssociationRequest$OutputLocation": "

An Amazon S3 bucket where you want to store the results of this request.

" + "InstanceAssociationOutputLocation":{ + "base":"

An Amazon S3 bucket where you want to store the results of this request.

", + "refs":{ + "AssociationDescription$OutputLocation":"

An Amazon S3 bucket where you want to store the output details of the request.

", + "AssociationVersionInfo$OutputLocation":"

The location in Amazon S3 specified for the association when the association version was created.

", + "CreateAssociationBatchRequestEntry$OutputLocation":"

An Amazon S3 bucket where you want to store the results of this request.

", + "CreateAssociationRequest$OutputLocation":"

An Amazon S3 bucket where you want to store the output details of the request.

", + "UpdateAssociationRequest$OutputLocation":"

An Amazon S3 bucket where you want to store the results of this request.

" } }, - "InstanceAssociationOutputUrl": { - "base": "

The URL of Amazon S3 bucket where you want to store the results of this request.

", - "refs": { - "InstanceAssociationStatusInfo$OutputUrl": "

A URL for an Amazon S3 bucket where you want to store the results of this request.

" + "InstanceAssociationOutputUrl":{ + "base":"

The URL of Amazon S3 bucket where you want to store the results of this request.

", + "refs":{ + "InstanceAssociationStatusInfo$OutputUrl":"

A URL for an Amazon S3 bucket where you want to store the results of this request.

" } }, - "InstanceAssociationStatusAggregatedCount": { - "base": null, - "refs": { - "InstanceAggregatedAssociationOverview$InstanceAssociationStatusAggregatedCount": "

The number of associations for the instance(s).

" + "InstanceAssociationStatusAggregatedCount":{ + "base":null, + "refs":{ + "InstanceAggregatedAssociationOverview$InstanceAssociationStatusAggregatedCount":"

The number of associations for the instance(s).

" } }, - "InstanceAssociationStatusInfo": { - "base": "

Status information about the instance association.

", - "refs": { - "InstanceAssociationStatusInfos$member": null + "InstanceAssociationStatusInfo":{ + "base":"

Status information about the instance association.

", + "refs":{ + "InstanceAssociationStatusInfos$member":null } }, - "InstanceAssociationStatusInfos": { - "base": null, - "refs": { - "DescribeInstanceAssociationsStatusResult$InstanceAssociationStatusInfos": "

Status information about the association.

" + "InstanceAssociationStatusInfos":{ + "base":null, + "refs":{ + "DescribeInstanceAssociationsStatusResult$InstanceAssociationStatusInfos":"

Status information about the association.

" } }, - "InstanceCount": { - "base": null, - "refs": { - "AssociationStatusAggregatedCount$value": null, - "InstanceAssociationStatusAggregatedCount$value": null + "InstanceCount":{ + "base":null, + "refs":{ + "AssociationStatusAggregatedCount$value":null, + "InstanceAssociationStatusAggregatedCount$value":null } }, - "InstanceId": { - "base": null, - "refs": { - "Association$InstanceId": "

The ID of the instance.

", - "AssociationDescription$InstanceId": "

The ID of the instance.

", - "CommandInvocation$InstanceId": "

The instance ID in which this invocation was requested.

", - "CreateAssociationBatchRequestEntry$InstanceId": "

The ID of the instance.

", - "CreateAssociationRequest$InstanceId": "

The instance ID.

", - "DeleteAssociationRequest$InstanceId": "

The ID of the instance.

", - "DescribeAssociationRequest$InstanceId": "

The instance ID.

", - "DescribeEffectiveInstanceAssociationsRequest$InstanceId": "

The instance ID for which you want to view all associations.

", - "DescribeInstanceAssociationsStatusRequest$InstanceId": "

The instance IDs for which you want association status information.

", - "DescribeInstancePatchesRequest$InstanceId": "

The ID of the instance whose patch state information should be retrieved.

", - "GetCommandInvocationRequest$InstanceId": "

(Required) The ID of the managed instance targeted by the command. A managed instance can be an Amazon EC2 instance or an instance in your hybrid environment that is configured for Systems Manager.

", - "GetCommandInvocationResult$InstanceId": "

The ID of the managed instance targeted by the command. A managed instance can be an Amazon EC2 instance or an instance in your hybrid environment that is configured for Systems Manager.

", - "GetDeployablePatchSnapshotForInstanceRequest$InstanceId": "

The ID of the instance for which the appropriate patch snapshot should be retrieved.

", - "GetDeployablePatchSnapshotForInstanceResult$InstanceId": "

The ID of the instance.

", - "InstanceAssociation$InstanceId": "

The instance ID.

", - "InstanceAssociationStatusInfo$InstanceId": "

The instance ID where the association was created.

", - "InstanceIdList$member": null, - "InstanceInformation$InstanceId": "

The instance ID.

", - "InstancePatchState$InstanceId": "

The ID of the managed instance the high-level patch compliance information was collected for.

", - "ListCommandInvocationsRequest$InstanceId": "

(Optional) The command execution details for a specific instance ID.

", - "ListCommandsRequest$InstanceId": "

(Optional) Lists commands issued against this instance ID.

", - "ListInventoryEntriesRequest$InstanceId": "

The instance ID for which you want inventory information.

", - "ListInventoryEntriesResult$InstanceId": "

The instance ID targeted by the request to query inventory information.

", - "PutInventoryRequest$InstanceId": "

One or more instance IDs where you want to add or update inventory items.

", - "UpdateAssociationStatusRequest$InstanceId": "

The ID of the instance.

" + "InstanceId":{ + "base":null, + "refs":{ + "Association$InstanceId":"

The ID of the instance.

", + "AssociationDescription$InstanceId":"

The ID of the instance.

", + "CommandInvocation$InstanceId":"

The instance ID in which this invocation was requested.

", + "CreateAssociationBatchRequestEntry$InstanceId":"

The ID of the instance.

", + "CreateAssociationRequest$InstanceId":"

The instance ID.

", + "DeleteAssociationRequest$InstanceId":"

The ID of the instance.

", + "DescribeAssociationRequest$InstanceId":"

The instance ID.

", + "DescribeEffectiveInstanceAssociationsRequest$InstanceId":"

The instance ID for which you want to view all associations.

", + "DescribeInstanceAssociationsStatusRequest$InstanceId":"

The instance IDs for which you want association status information.

", + "DescribeInstancePatchesRequest$InstanceId":"

The ID of the instance whose patch state information should be retrieved.

", + "GetCommandInvocationRequest$InstanceId":"

(Required) The ID of the managed instance targeted by the command. A managed instance can be an Amazon EC2 instance or an instance in your hybrid environment that is configured for Systems Manager.

", + "GetCommandInvocationResult$InstanceId":"

The ID of the managed instance targeted by the command. A managed instance can be an Amazon EC2 instance or an instance in your hybrid environment that is configured for Systems Manager.

", + "GetDeployablePatchSnapshotForInstanceRequest$InstanceId":"

The ID of the instance for which the appropriate patch snapshot should be retrieved.

", + "GetDeployablePatchSnapshotForInstanceResult$InstanceId":"

The ID of the instance.

", + "InstanceAssociation$InstanceId":"

The instance ID.

", + "InstanceAssociationStatusInfo$InstanceId":"

The instance ID where the association was created.

", + "InstanceIdList$member":null, + "InstanceInformation$InstanceId":"

The instance ID.

", + "InstancePatchState$InstanceId":"

The ID of the managed instance the high-level patch compliance information was collected for.

", + "ListCommandInvocationsRequest$InstanceId":"

(Optional) The command execution details for a specific instance ID.

", + "ListCommandsRequest$InstanceId":"

(Optional) Lists commands issued against this instance ID.

", + "ListInventoryEntriesRequest$InstanceId":"

The instance ID for which you want inventory information.

", + "ListInventoryEntriesResult$InstanceId":"

The instance ID targeted by the request to query inventory information.

", + "PutInventoryRequest$InstanceId":"

One or more instance IDs where you want to add or update inventory items.

", + "UpdateAssociationStatusRequest$InstanceId":"

The ID of the instance.

" } }, - "InstanceIdList": { - "base": null, - "refs": { - "CancelCommandRequest$InstanceIds": "

(Optional) A list of instance IDs on which you want to cancel the command. If not provided, the command is canceled on every instance on which it was requested.

", - "Command$InstanceIds": "

The instance IDs against which this command was requested.

", - "DescribeInstancePatchStatesRequest$InstanceIds": "

The ID of the instance whose patch state information should be retrieved.

", - "SendCommandRequest$InstanceIds": "

The instance IDs where the command should execute. You can specify a maximum of 50 IDs. If you prefer not to list individual instance IDs, you can instead send commands to a fleet of instances using the Targets parameter, which accepts EC2 tags. For more information about how to use Targets, see Sending Commands to a Fleet.

" + "InstanceIdList":{ + "base":null, + "refs":{ + "CancelCommandRequest$InstanceIds":"

(Optional) A list of instance IDs on which you want to cancel the command. If not provided, the command is canceled on every instance on which it was requested.

", + "Command$InstanceIds":"

The instance IDs against which this command was requested.

", + "DescribeInstancePatchStatesRequest$InstanceIds":"

The ID of the instance whose patch state information should be retrieved.

", + "SendCommandRequest$InstanceIds":"

The instance IDs where the command should execute. You can specify a maximum of 50 IDs. If you prefer not to list individual instance IDs, you can instead send commands to a fleet of instances using the Targets parameter, which accepts EC2 tags. For more information about how to use Targets, see Sending Commands to a Fleet in the AWS Systems Manager User Guide.

" } }, - "InstanceInformation": { - "base": "

Describes a filter for a specific list of instances.

", - "refs": { - "InstanceInformationList$member": null + "InstanceInformation":{ + "base":"

Describes a filter for a specific list of instances.

", + "refs":{ + "InstanceInformationList$member":null } }, - "InstanceInformationFilter": { - "base": "

Describes a filter for a specific list of instances.

", - "refs": { - "InstanceInformationFilterList$member": null + "InstanceInformationFilter":{ + "base":"

Describes a filter for a specific list of instances.

", + "refs":{ + "InstanceInformationFilterList$member":null } }, - "InstanceInformationFilterKey": { - "base": null, - "refs": { - "InstanceInformationFilter$key": "

The name of the filter.

" + "InstanceInformationFilterKey":{ + "base":null, + "refs":{ + "InstanceInformationFilter$key":"

The name of the filter.

" } }, - "InstanceInformationFilterList": { - "base": null, - "refs": { - "DescribeInstanceInformationRequest$InstanceInformationFilterList": "

One or more filters. Use a filter to return a more specific list of instances.

" + "InstanceInformationFilterList":{ + "base":null, + "refs":{ + "DescribeInstanceInformationRequest$InstanceInformationFilterList":"

One or more filters. Use a filter to return a more specific list of instances.

" } }, - "InstanceInformationFilterValue": { - "base": null, - "refs": { - "InstanceInformationFilterValueSet$member": null + "InstanceInformationFilterValue":{ + "base":null, + "refs":{ + "InstanceInformationFilterValueSet$member":null } }, - "InstanceInformationFilterValueSet": { - "base": null, - "refs": { - "InstanceInformationFilter$valueSet": "

The filter values.

", - "InstanceInformationStringFilter$Values": "

The filter values.

" + "InstanceInformationFilterValueSet":{ + "base":null, + "refs":{ + "InstanceInformationFilter$valueSet":"

The filter values.

", + "InstanceInformationStringFilter$Values":"

The filter values.

" } }, - "InstanceInformationList": { - "base": null, - "refs": { - "DescribeInstanceInformationResult$InstanceInformationList": "

The instance information list.

" + "InstanceInformationList":{ + "base":null, + "refs":{ + "DescribeInstanceInformationResult$InstanceInformationList":"

The instance information list.

" } }, - "InstanceInformationStringFilter": { - "base": "

The filters to describe or get information about your managed instances.

", - "refs": { - "InstanceInformationStringFilterList$member": null + "InstanceInformationStringFilter":{ + "base":"

The filters to describe or get information about your managed instances.

", + "refs":{ + "InstanceInformationStringFilterList$member":null } }, - "InstanceInformationStringFilterKey": { - "base": null, - "refs": { - "InstanceInformationStringFilter$Key": "

The filter key name to describe your instances. For example:

\"InstanceIds\"|\"AgentVersion\"|\"PingStatus\"|\"PlatformTypes\"|\"ActivationIds\"|\"IamRole\"|\"ResourceType\"|\"AssociationStatus\"|\"Tag Key\"

" + "InstanceInformationStringFilterKey":{ + "base":null, + "refs":{ + "InstanceInformationStringFilter$Key":"

The filter key name to describe your instances. For example:

\"InstanceIds\"|\"AgentVersion\"|\"PingStatus\"|\"PlatformTypes\"|\"ActivationIds\"|\"IamRole\"|\"ResourceType\"|\"AssociationStatus\"|\"Tag Key\"

" } }, - "InstanceInformationStringFilterList": { - "base": null, - "refs": { - "DescribeInstanceInformationRequest$Filters": "

One or more filters. Use a filter to return a more specific list of instances.

" + "InstanceInformationStringFilterList":{ + "base":null, + "refs":{ + "DescribeInstanceInformationRequest$Filters":"

One or more filters. Use a filter to return a more specific list of instances.

" } }, - "InstancePatchState": { - "base": "

Defines the high-level patch compliance state for a managed instance, providing information about the number of installed, missing, not applicable, and failed patches along with metadata about the operation when this information was gathered for the instance.

", - "refs": { - "InstancePatchStateList$member": null, - "InstancePatchStatesList$member": null + "InstancePatchState":{ + "base":"

Defines the high-level patch compliance state for a managed instance, providing information about the number of installed, missing, not applicable, and failed patches along with metadata about the operation when this information was gathered for the instance.

", + "refs":{ + "InstancePatchStateList$member":null, + "InstancePatchStatesList$member":null } }, - "InstancePatchStateFilter": { - "base": "

Defines a filter used in DescribeInstancePatchStatesForPatchGroup used to scope down the information returned by the API.

", - "refs": { - "InstancePatchStateFilterList$member": null + "InstancePatchStateFilter":{ + "base":"

Defines a filter used in DescribeInstancePatchStatesForPatchGroup used to scope down the information returned by the API.

", + "refs":{ + "InstancePatchStateFilterList$member":null } }, - "InstancePatchStateFilterKey": { - "base": null, - "refs": { - "InstancePatchStateFilter$Key": "

The key for the filter. Supported values are FailedCount, InstalledCount, InstalledOtherCount, MissingCount and NotApplicableCount.

" + "InstancePatchStateFilterKey":{ + "base":null, + "refs":{ + "InstancePatchStateFilter$Key":"

The key for the filter. Supported values are FailedCount, InstalledCount, InstalledOtherCount, MissingCount and NotApplicableCount.

" } }, - "InstancePatchStateFilterList": { - "base": null, - "refs": { - "DescribeInstancePatchStatesForPatchGroupRequest$Filters": "

Each entry in the array is a structure containing:

Key (string between 1 and 200 characters)

Values (array containing a single string)

Type (string \"Equal\", \"NotEqual\", \"LessThan\", \"GreaterThan\")

" + "InstancePatchStateFilterList":{ + "base":null, + "refs":{ + "DescribeInstancePatchStatesForPatchGroupRequest$Filters":"

Each entry in the array is a structure containing:

Key (string between 1 and 200 characters)

Values (array containing a single string)

Type (string \"Equal\", \"NotEqual\", \"LessThan\", \"GreaterThan\")

" } }, - "InstancePatchStateFilterValue": { - "base": null, - "refs": { - "InstancePatchStateFilterValues$member": null + "InstancePatchStateFilterValue":{ + "base":null, + "refs":{ + "InstancePatchStateFilterValues$member":null } }, - "InstancePatchStateFilterValues": { - "base": null, - "refs": { - "InstancePatchStateFilter$Values": "

The value for the filter, must be an integer greater than or equal to 0.

" + "InstancePatchStateFilterValues":{ + "base":null, + "refs":{ + "InstancePatchStateFilter$Values":"

The value for the filter, must be an integer greater than or equal to 0.

" } }, - "InstancePatchStateList": { - "base": null, - "refs": { - "DescribeInstancePatchStatesResult$InstancePatchStates": "

The high-level patch state for the requested instances.

" + "InstancePatchStateList":{ + "base":null, + "refs":{ + "DescribeInstancePatchStatesResult$InstancePatchStates":"

The high-level patch state for the requested instances.

" } }, - "InstancePatchStateOperatorType": { - "base": null, - "refs": { - "InstancePatchStateFilter$Type": "

The type of comparison that should be performed for the value: Equal, NotEqual, LessThan or GreaterThan.

" + "InstancePatchStateOperatorType":{ + "base":null, + "refs":{ + "InstancePatchStateFilter$Type":"

The type of comparison that should be performed for the value: Equal, NotEqual, LessThan or GreaterThan.

" } }, - "InstancePatchStatesList": { - "base": null, - "refs": { - "DescribeInstancePatchStatesForPatchGroupResult$InstancePatchStates": "

The high-level patch state for the requested instances.

" + "InstancePatchStatesList":{ + "base":null, + "refs":{ + "DescribeInstancePatchStatesForPatchGroupResult$InstancePatchStates":"

The high-level patch state for the requested instances.

" } }, - "InstanceTagName": { - "base": null, - "refs": { - "CommandInvocation$InstanceName": "

The name of the invocation target. For Amazon EC2 instances this is the value for the aws:Name tag. For on-premises instances, this is the name of the instance.

" + "InstanceTagName":{ + "base":null, + "refs":{ + "CommandInvocation$InstanceName":"

The name of the invocation target. For Amazon EC2 instances this is the value for the aws:Name tag. For on-premises instances, this is the name of the instance.

" } }, - "Integer": { - "base": null, - "refs": { - "DescribePatchGroupStateResult$Instances": "

The number of instances in the patch group.

", - "DescribePatchGroupStateResult$InstancesWithInstalledPatches": "

The number of instances with installed patches.

", - "DescribePatchGroupStateResult$InstancesWithInstalledOtherPatches": "

The number of instances with patches installed that aren't defined in the patch baseline.

", - "DescribePatchGroupStateResult$InstancesWithMissingPatches": "

The number of instances with missing patches from the patch baseline.

", - "DescribePatchGroupStateResult$InstancesWithFailedPatches": "

The number of instances with patches from the patch baseline that failed to install.

", - "DescribePatchGroupStateResult$InstancesWithNotApplicablePatches": "

The number of instances with patches that aren't applicable.

", - "StepExecution$MaxAttempts": "

The maximum number of tries to run the action of the step. The default value is 1.

" + "Integer":{ + "base":null, + "refs":{ + "DescribePatchGroupStateResult$Instances":"

The number of instances in the patch group.

", + "DescribePatchGroupStateResult$InstancesWithInstalledPatches":"

The number of instances with installed patches.

", + "DescribePatchGroupStateResult$InstancesWithInstalledOtherPatches":"

The number of instances with patches installed that aren't defined in the patch baseline.

", + "DescribePatchGroupStateResult$InstancesWithMissingPatches":"

The number of instances with missing patches from the patch baseline.

", + "DescribePatchGroupStateResult$InstancesWithFailedPatches":"

The number of instances with patches from the patch baseline that failed to install.

", + "DescribePatchGroupStateResult$InstancesWithNotApplicablePatches":"

The number of instances with patches that aren't applicable.

", + "StepExecution$MaxAttempts":"

The maximum number of tries to run the action of the step. The default value is 1.

" } }, - "InternalServerError": { - "base": "

An error occurred on the server side.

", - "refs": { + "InternalServerError":{ + "base":"

An error occurred on the server side.

", + "refs":{ } }, - "InvalidActivation": { - "base": "

The activation is not valid. The activation might have been deleted, or the ActivationId and the ActivationCode do not match.

", - "refs": { + "InvalidActivation":{ + "base":"

The activation is not valid. The activation might have been deleted, or the ActivationId and the ActivationCode do not match.

", + "refs":{ } }, - "InvalidActivationId": { - "base": "

The activation ID is not valid. Verify the you entered the correct ActivationId or ActivationCode and try again.

", - "refs": { + "InvalidActivationId":{ + "base":"

The activation ID is not valid. Verify the you entered the correct ActivationId or ActivationCode and try again.

", + "refs":{ } }, - "InvalidAllowedPatternException": { - "base": "

The request does not meet the regular expression requirement.

", - "refs": { + "InvalidAllowedPatternException":{ + "base":"

The request does not meet the regular expression requirement.

", + "refs":{ } }, - "InvalidAssociationVersion": { - "base": "

The version you specified is not valid. Use ListAssociationVersions to view all versions of an association according to the association ID. Or, use the $LATEST parameter to view the latest version of the association.

", - "refs": { + "InvalidAssociationVersion":{ + "base":"

The version you specified is not valid. Use ListAssociationVersions to view all versions of an association according to the association ID. Or, use the $LATEST parameter to view the latest version of the association.

", + "refs":{ } }, - "InvalidAutomationExecutionParametersException": { - "base": "

The supplied parameters for invoking the specified Automation document are incorrect. For example, they may not match the set of parameters permitted for the specified Automation document.

", - "refs": { + "InvalidAutomationExecutionParametersException":{ + "base":"

The supplied parameters for invoking the specified Automation document are incorrect. For example, they may not match the set of parameters permitted for the specified Automation document.

", + "refs":{ } }, - "InvalidAutomationSignalException": { - "base": "

The signal is not valid for the current Automation execution.

", - "refs": { + "InvalidAutomationSignalException":{ + "base":"

The signal is not valid for the current Automation execution.

", + "refs":{ } }, - "InvalidAutomationStatusUpdateException": { - "base": "

The specified update status operation is not valid.

", - "refs": { + "InvalidAutomationStatusUpdateException":{ + "base":"

The specified update status operation is not valid.

", + "refs":{ } }, - "InvalidCommandId": { - "base": null, - "refs": { + "InvalidCommandId":{ + "base":null, + "refs":{ } }, - "InvalidDeleteInventoryParametersException": { - "base": "

One or more of the parameters specified for the delete operation is not valid. Verify all parameters and try again.

", - "refs": { + "InvalidDeleteInventoryParametersException":{ + "base":"

One or more of the parameters specified for the delete operation is not valid. Verify all parameters and try again.

", + "refs":{ } }, - "InvalidDeletionIdException": { - "base": "

The ID specified for the delete operation does not exist or is not valide. Verify the ID and try again.

", - "refs": { + "InvalidDeletionIdException":{ + "base":"

The ID specified for the delete operation does not exist or is not valide. Verify the ID and try again.

", + "refs":{ } }, - "InvalidDocument": { - "base": "

The specified document does not exist.

", - "refs": { + "InvalidDocument":{ + "base":"

The specified document does not exist.

", + "refs":{ } }, - "InvalidDocumentContent": { - "base": "

The content for the document is not valid.

", - "refs": { + "InvalidDocumentContent":{ + "base":"

The content for the document is not valid.

", + "refs":{ } }, - "InvalidDocumentOperation": { - "base": "

You attempted to delete a document while it is still shared. You must stop sharing the document before you can delete it.

", - "refs": { + "InvalidDocumentOperation":{ + "base":"

You attempted to delete a document while it is still shared. You must stop sharing the document before you can delete it.

", + "refs":{ } }, - "InvalidDocumentSchemaVersion": { - "base": "

The version of the document schema is not supported.

", - "refs": { + "InvalidDocumentSchemaVersion":{ + "base":"

The version of the document schema is not supported.

", + "refs":{ } }, - "InvalidDocumentVersion": { - "base": "

The document version is not valid or does not exist.

", - "refs": { + "InvalidDocumentVersion":{ + "base":"

The document version is not valid or does not exist.

", + "refs":{ } }, - "InvalidFilter": { - "base": "

The filter name is not valid. Verify the you entered the correct name and try again.

", - "refs": { + "InvalidFilter":{ + "base":"

The filter name is not valid. Verify the you entered the correct name and try again.

", + "refs":{ } }, - "InvalidFilterKey": { - "base": "

The specified key is not valid.

", - "refs": { + "InvalidFilterKey":{ + "base":"

The specified key is not valid.

", + "refs":{ } }, - "InvalidFilterOption": { - "base": "

The specified filter option is not valid. Valid options are Equals and BeginsWith. For Path filter, valid options are Recursive and OneLevel.

", - "refs": { + "InvalidFilterOption":{ + "base":"

The specified filter option is not valid. Valid options are Equals and BeginsWith. For Path filter, valid options are Recursive and OneLevel.

", + "refs":{ } }, - "InvalidFilterValue": { - "base": "

The filter value is not valid. Verify the value and try again.

", - "refs": { + "InvalidFilterValue":{ + "base":"

The filter value is not valid. Verify the value and try again.

", + "refs":{ } }, - "InvalidInstanceId": { - "base": "

The following problems can cause this exception:

You do not have permission to access the instance.

The SSM Agent is not running. On managed instances and Linux instances, verify that the SSM Agent is running. On EC2 Windows instances, verify that the EC2Config service is running.

The SSM Agent or EC2Config service is not registered to the SSM endpoint. Try reinstalling the SSM Agent or EC2Config service.

The instance is not in valid state. Valid states are: Running, Pending, Stopped, Stopping. Invalid states are: Shutting-down and Terminated.

", - "refs": { + "InvalidInstanceId":{ + "base":"

The following problems can cause this exception:

You do not have permission to access the instance.

SSM Agent is not running. On managed instances and Linux instances, verify that the SSM Agent is running. On EC2 Windows instances, verify that the EC2Config service is running.

SSM Agent or EC2Config service is not registered to the SSM endpoint. Try reinstalling SSM Agent or EC2Config service.

The instance is not in valid state. Valid states are: Running, Pending, Stopped, Stopping. Invalid states are: Shutting-down and Terminated.

", + "refs":{ } }, - "InvalidInstanceInformationFilterValue": { - "base": "

The specified filter value is not valid.

", - "refs": { + "InvalidInstanceInformationFilterValue":{ + "base":"

The specified filter value is not valid.

", + "refs":{ } }, - "InvalidInventoryItemContextException": { - "base": "

You specified invalid keys or values in the Context attribute for InventoryItem. Verify the keys and values, and try again.

", - "refs": { + "InvalidInventoryItemContextException":{ + "base":"

You specified invalid keys or values in the Context attribute for InventoryItem. Verify the keys and values, and try again.

", + "refs":{ } }, - "InvalidInventoryRequestException": { - "base": "

The request is not valid.

", - "refs": { + "InvalidInventoryRequestException":{ + "base":"

The request is not valid.

", + "refs":{ } }, - "InvalidItemContentException": { - "base": "

One or more content items is not valid.

", - "refs": { + "InvalidItemContentException":{ + "base":"

One or more content items is not valid.

", + "refs":{ } }, - "InvalidKeyId": { - "base": "

The query key ID is not valid.

", - "refs": { + "InvalidKeyId":{ + "base":"

The query key ID is not valid.

", + "refs":{ } }, - "InvalidNextToken": { - "base": "

The specified token is not valid.

", - "refs": { + "InvalidNextToken":{ + "base":"

The specified token is not valid.

", + "refs":{ } }, - "InvalidNotificationConfig": { - "base": "

One or more configuration items is not valid. Verify that a valid Amazon Resource Name (ARN) was provided for an Amazon SNS topic.

", - "refs": { + "InvalidNotificationConfig":{ + "base":"

One or more configuration items is not valid. Verify that a valid Amazon Resource Name (ARN) was provided for an Amazon SNS topic.

", + "refs":{ } }, - "InvalidOptionException": { - "base": "

The delete inventory option specified is not valid. Verify the option and try again.

", - "refs": { + "InvalidOptionException":{ + "base":"

The delete inventory option specified is not valid. Verify the option and try again.

", + "refs":{ } }, - "InvalidOutputFolder": { - "base": "

The S3 bucket does not exist.

", - "refs": { + "InvalidOutputFolder":{ + "base":"

The S3 bucket does not exist.

", + "refs":{ } }, - "InvalidOutputLocation": { - "base": "

The output location is not valid or does not exist.

", - "refs": { + "InvalidOutputLocation":{ + "base":"

The output location is not valid or does not exist.

", + "refs":{ } }, - "InvalidParameters": { - "base": "

You must specify values for all required parameters in the Systems Manager document. You can only supply values to parameters defined in the Systems Manager document.

", - "refs": { + "InvalidParameters":{ + "base":"

You must specify values for all required parameters in the Systems Manager document. You can only supply values to parameters defined in the Systems Manager document.

", + "refs":{ } }, - "InvalidPermissionType": { - "base": "

The permission type is not supported. Share is the only supported permission type.

", - "refs": { + "InvalidPermissionType":{ + "base":"

The permission type is not supported. Share is the only supported permission type.

", + "refs":{ } }, - "InvalidPluginName": { - "base": "

The plugin name is not valid.

", - "refs": { + "InvalidPluginName":{ + "base":"

The plugin name is not valid.

", + "refs":{ } }, - "InvalidResourceId": { - "base": "

The resource ID is not valid. Verify that you entered the correct ID and try again.

", - "refs": { + "InvalidResourceId":{ + "base":"

The resource ID is not valid. Verify that you entered the correct ID and try again.

", + "refs":{ } }, - "InvalidResourceType": { - "base": "

The resource type is not valid. For example, if you are attempting to tag an instance, the instance must be a registered, managed instance.

", - "refs": { + "InvalidResourceType":{ + "base":"

The resource type is not valid. For example, if you are attempting to tag an instance, the instance must be a registered, managed instance.

", + "refs":{ } }, - "InvalidResultAttributeException": { - "base": "

The specified inventory item result attribute is not valid.

", - "refs": { + "InvalidResultAttributeException":{ + "base":"

The specified inventory item result attribute is not valid.

", + "refs":{ } }, - "InvalidRole": { - "base": "

The role name can't contain invalid characters. Also verify that you specified an IAM role for notifications that includes the required trust policy. For information about configuring the IAM role for Run Command notifications, see Configuring Amazon SNS Notifications for Run Command in the AWS Systems Manager User Guide.

", - "refs": { + "InvalidRole":{ + "base":"

The role name can't contain invalid characters. Also verify that you specified an IAM role for notifications that includes the required trust policy. For information about configuring the IAM role for Run Command notifications, see Configuring Amazon SNS Notifications for Run Command in the AWS Systems Manager User Guide.

", + "refs":{ } }, - "InvalidSchedule": { - "base": "

The schedule is invalid. Verify your cron or rate expression and try again.

", - "refs": { + "InvalidSchedule":{ + "base":"

The schedule is invalid. Verify your cron or rate expression and try again.

", + "refs":{ } }, - "InvalidTarget": { - "base": "

The target is not valid or does not exist. It might not be configured for EC2 Systems Manager or you might not have permission to perform the operation.

", - "refs": { + "InvalidTarget":{ + "base":"

The target is not valid or does not exist. It might not be configured for EC2 Systems Manager or you might not have permission to perform the operation.

", + "refs":{ } }, - "InvalidTypeNameException": { - "base": "

The parameter type name is not valid.

", - "refs": { + "InvalidTypeNameException":{ + "base":"

The parameter type name is not valid.

", + "refs":{ } }, - "InvalidUpdate": { - "base": "

The update is not valid.

", - "refs": { + "InvalidUpdate":{ + "base":"

The update is not valid.

", + "refs":{ } }, - "InventoryAggregator": { - "base": "

Specifies the inventory type and attribute for the aggregation execution.

", - "refs": { - "InventoryAggregatorList$member": null + "InventoryAggregator":{ + "base":"

Specifies the inventory type and attribute for the aggregation execution.

", + "refs":{ + "InventoryAggregatorList$member":null } }, - "InventoryAggregatorExpression": { - "base": null, - "refs": { - "InventoryAggregator$Expression": "

The inventory type and attribute name for aggregation.

" + "InventoryAggregatorExpression":{ + "base":null, + "refs":{ + "InventoryAggregator$Expression":"

The inventory type and attribute name for aggregation.

" } }, - "InventoryAggregatorList": { - "base": null, - "refs": { - "GetInventoryRequest$Aggregators": "

Returns counts of inventory types based on one or more expressions. For example, if you aggregate by using an expression that uses the AWS:InstanceInformation.PlatformType type, you can see a count of how many Windows and Linux instances exist in your inventoried fleet.

", - "InventoryAggregator$Aggregators": "

Nested aggregators to further refine aggregation for an inventory type.

" + "InventoryAggregatorList":{ + "base":null, + "refs":{ + "GetInventoryRequest$Aggregators":"

Returns counts of inventory types based on one or more expressions. For example, if you aggregate by using an expression that uses the AWS:InstanceInformation.PlatformType type, you can see a count of how many Windows and Linux instances exist in your inventoried fleet.

", + "InventoryAggregator$Aggregators":"

Nested aggregators to further refine aggregation for an inventory type.

" } }, - "InventoryAttributeDataType": { - "base": null, - "refs": { - "InventoryItemAttribute$DataType": "

The data type of the inventory item attribute.

" + "InventoryAttributeDataType":{ + "base":null, + "refs":{ + "InventoryItemAttribute$DataType":"

The data type of the inventory item attribute.

" } }, - "InventoryDeletionId": { - "base": null, - "refs": { - "DeleteInventoryResult$DeletionId": "

Every DeleteInventory action is assigned a unique ID. This option returns a unique ID. You can use this ID to query the status of a delete operation. This option is useful for ensuring that a delete operation has completed before you begin other actions.

", - "DescribeInventoryDeletionsRequest$DeletionId": "

Specify the delete inventory ID for which you want information. This ID was returned by the DeleteInventory action.

", - "InventoryDeletionStatusItem$DeletionId": "

The deletion ID returned by the DeleteInventory action.

" + "InventoryDeletionId":{ + "base":null, + "refs":{ + "DeleteInventoryResult$DeletionId":"

Every DeleteInventory action is assigned a unique ID. This option returns a unique ID. You can use this ID to query the status of a delete operation. This option is useful for ensuring that a delete operation has completed before you begin other actions.

", + "DescribeInventoryDeletionsRequest$DeletionId":"

Specify the delete inventory ID for which you want information. This ID was returned by the DeleteInventory action.

", + "InventoryDeletionStatusItem$DeletionId":"

The deletion ID returned by the DeleteInventory action.

" } }, - "InventoryDeletionLastStatusMessage": { - "base": null, - "refs": { - "InventoryDeletionStatusItem$LastStatusMessage": "

Information about the status.

" + "InventoryDeletionLastStatusMessage":{ + "base":null, + "refs":{ + "InventoryDeletionStatusItem$LastStatusMessage":"

Information about the status.

" } }, - "InventoryDeletionLastStatusUpdateTime": { - "base": null, - "refs": { - "InventoryDeletionStatusItem$LastStatusUpdateTime": "

The UTC timestamp of when the last status report.

" + "InventoryDeletionLastStatusUpdateTime":{ + "base":null, + "refs":{ + "InventoryDeletionStatusItem$LastStatusUpdateTime":"

The UTC timestamp of when the last status report.

" } }, - "InventoryDeletionStartTime": { - "base": null, - "refs": { - "InventoryDeletionStatusItem$DeletionStartTime": "

The UTC timestamp when the delete operation started.

" + "InventoryDeletionStartTime":{ + "base":null, + "refs":{ + "InventoryDeletionStatusItem$DeletionStartTime":"

The UTC timestamp when the delete operation started.

" } }, - "InventoryDeletionStatus": { - "base": null, - "refs": { - "InventoryDeletionStatusItem$LastStatus": "

The status of the operation. Possible values are InProgress and Complete.

" + "InventoryDeletionStatus":{ + "base":null, + "refs":{ + "InventoryDeletionStatusItem$LastStatus":"

The status of the operation. Possible values are InProgress and Complete.

" } }, - "InventoryDeletionStatusItem": { - "base": "

Status information returned by the DeleteInventory action.

", - "refs": { - "InventoryDeletionsList$member": null + "InventoryDeletionStatusItem":{ + "base":"

Status information returned by the DeleteInventory action.

", + "refs":{ + "InventoryDeletionsList$member":null } }, - "InventoryDeletionSummary": { - "base": "

Information about the delete operation.

", - "refs": { - "DeleteInventoryResult$DeletionSummary": "

A summary of the delete operation. For more information about this summary, see Understanding the Delete Inventory Summary.

", - "InventoryDeletionStatusItem$DeletionSummary": "

Information about the delete operation. For more information about this summary, see Understanding the Delete Inventory Summary.

" + "InventoryDeletionSummary":{ + "base":"

Information about the delete operation.

", + "refs":{ + "DeleteInventoryResult$DeletionSummary":"

A summary of the delete operation. For more information about this summary, see Understanding the Delete Inventory Summary in the AWS Systems Manager User Guide.

", + "InventoryDeletionStatusItem$DeletionSummary":"

Information about the delete operation. For more information about this summary, see Understanding the Delete Inventory Summary in the AWS Systems Manager User Guide.

" } }, - "InventoryDeletionSummaryItem": { - "base": "

Either a count, remaining count, or a version number in a delete inventory summary.

", - "refs": { - "InventoryDeletionSummaryItems$member": null + "InventoryDeletionSummaryItem":{ + "base":"

Either a count, remaining count, or a version number in a delete inventory summary.

", + "refs":{ + "InventoryDeletionSummaryItems$member":null } }, - "InventoryDeletionSummaryItems": { - "base": null, - "refs": { - "InventoryDeletionSummary$SummaryItems": "

A list of counts and versions for deleted items.

" + "InventoryDeletionSummaryItems":{ + "base":null, + "refs":{ + "InventoryDeletionSummary$SummaryItems":"

A list of counts and versions for deleted items.

" } }, - "InventoryDeletionsList": { - "base": null, - "refs": { - "DescribeInventoryDeletionsResult$InventoryDeletions": "

A list of status items for deleted inventory.

" + "InventoryDeletionsList":{ + "base":null, + "refs":{ + "DescribeInventoryDeletionsResult$InventoryDeletions":"

A list of status items for deleted inventory.

" } }, - "InventoryFilter": { - "base": "

One or more filters. Use a filter to return a more specific list of results.

", - "refs": { - "InventoryFilterList$member": null + "InventoryFilter":{ + "base":"

One or more filters. Use a filter to return a more specific list of results.

", + "refs":{ + "InventoryFilterList$member":null } }, - "InventoryFilterKey": { - "base": null, - "refs": { - "InventoryFilter$Key": "

The name of the filter key.

" + "InventoryFilterKey":{ + "base":null, + "refs":{ + "InventoryFilter$Key":"

The name of the filter key.

" } }, - "InventoryFilterList": { - "base": null, - "refs": { - "GetInventoryRequest$Filters": "

One or more filters. Use a filter to return a more specific list of results.

", - "ListInventoryEntriesRequest$Filters": "

One or more filters. Use a filter to return a more specific list of results.

" + "InventoryFilterList":{ + "base":null, + "refs":{ + "GetInventoryRequest$Filters":"

One or more filters. Use a filter to return a more specific list of results.

", + "ListInventoryEntriesRequest$Filters":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "InventoryFilterValue": { - "base": null, - "refs": { - "InventoryFilterValueList$member": null + "InventoryFilterValue":{ + "base":null, + "refs":{ + "InventoryFilterValueList$member":null } }, - "InventoryFilterValueList": { - "base": null, - "refs": { - "InventoryFilter$Values": "

Inventory filter values. Example: inventory filter where instance IDs are specified as values Key=AWS:InstanceInformation.InstanceId,Values= i-a12b3c4d5e6g, i-1a2b3c4d5e6,Type=Equal

" + "InventoryFilterValueList":{ + "base":null, + "refs":{ + "InventoryFilter$Values":"

Inventory filter values. Example: inventory filter where instance IDs are specified as values Key=AWS:InstanceInformation.InstanceId,Values= i-a12b3c4d5e6g, i-1a2b3c4d5e6,Type=Equal

" } }, - "InventoryItem": { - "base": "

Information collected from managed instances based on your inventory policy document

", - "refs": { - "InventoryItemList$member": null + "InventoryItem":{ + "base":"

Information collected from managed instances based on your inventory policy document

", + "refs":{ + "InventoryItemList$member":null } }, - "InventoryItemAttribute": { - "base": "

Attributes are the entries within the inventory item content. It contains name and value.

", - "refs": { - "InventoryItemAttributeList$member": null + "InventoryItemAttribute":{ + "base":"

Attributes are the entries within the inventory item content. It contains name and value.

", + "refs":{ + "InventoryItemAttributeList$member":null } }, - "InventoryItemAttributeList": { - "base": null, - "refs": { - "InventoryItemSchema$Attributes": "

The schema attributes for inventory. This contains data type and attribute name.

" + "InventoryItemAttributeList":{ + "base":null, + "refs":{ + "InventoryItemSchema$Attributes":"

The schema attributes for inventory. This contains data type and attribute name.

" } }, - "InventoryItemAttributeName": { - "base": null, - "refs": { - "InventoryItemAttribute$Name": "

Name of the inventory item attribute.

" + "InventoryItemAttributeName":{ + "base":null, + "refs":{ + "InventoryItemAttribute$Name":"

Name of the inventory item attribute.

" } }, - "InventoryItemCaptureTime": { - "base": null, - "refs": { - "InventoryItem$CaptureTime": "

The time the inventory information was collected.

", - "InventoryResultItem$CaptureTime": "

The time inventory item data was captured.

", - "ListInventoryEntriesResult$CaptureTime": "

The time that inventory information was collected for the instance(s).

" + "InventoryItemCaptureTime":{ + "base":null, + "refs":{ + "InventoryItem$CaptureTime":"

The time the inventory information was collected.

", + "InventoryResultItem$CaptureTime":"

The time inventory item data was captured.

", + "ListInventoryEntriesResult$CaptureTime":"

The time that inventory information was collected for the instance(s).

" } }, - "InventoryItemContentContext": { - "base": null, - "refs": { - "InventoryItem$Context": "

A map of associated properties for a specified inventory type. For example, with this attribute, you can specify the ExecutionId, ExecutionType, ComplianceType properties of the AWS:ComplianceItem type.

" + "InventoryItemContentContext":{ + "base":null, + "refs":{ + "InventoryItem$Context":"

A map of associated properties for a specified inventory type. For example, with this attribute, you can specify the ExecutionId, ExecutionType, ComplianceType properties of the AWS:ComplianceItem type.

" } }, - "InventoryItemContentHash": { - "base": null, - "refs": { - "InventoryItem$ContentHash": "

MD5 hash of the inventory item type contents. The content hash is used to determine whether to update inventory information. The PutInventory API does not update the inventory item type contents if the MD5 hash has not changed since last update.

", - "InventoryResultItem$ContentHash": "

MD5 hash of the inventory item type contents. The content hash is used to determine whether to update inventory information. The PutInventory API does not update the inventory item type contents if the MD5 hash has not changed since last update.

" + "InventoryItemContentHash":{ + "base":null, + "refs":{ + "InventoryItem$ContentHash":"

MD5 hash of the inventory item type contents. The content hash is used to determine whether to update inventory information. The PutInventory API does not update the inventory item type contents if the MD5 hash has not changed since last update.

", + "InventoryResultItem$ContentHash":"

MD5 hash of the inventory item type contents. The content hash is used to determine whether to update inventory information. The PutInventory API does not update the inventory item type contents if the MD5 hash has not changed since last update.

" } }, - "InventoryItemEntry": { - "base": null, - "refs": { - "InventoryItemEntryList$member": null + "InventoryItemEntry":{ + "base":null, + "refs":{ + "InventoryItemEntryList$member":null } }, - "InventoryItemEntryList": { - "base": null, - "refs": { - "InventoryItem$Content": "

The inventory data of the inventory type.

", - "InventoryResultItem$Content": "

Contains all the inventory data of the item type. Results include attribute names and values.

", - "ListInventoryEntriesResult$Entries": "

A list of inventory items on the instance(s).

" + "InventoryItemEntryList":{ + "base":null, + "refs":{ + "InventoryItem$Content":"

The inventory data of the inventory type.

", + "InventoryResultItem$Content":"

Contains all the inventory data of the item type. Results include attribute names and values.

", + "ListInventoryEntriesResult$Entries":"

A list of inventory items on the instance(s).

" } }, - "InventoryItemList": { - "base": null, - "refs": { - "PutInventoryRequest$Items": "

The inventory items that you want to add or update on instances.

" + "InventoryItemList":{ + "base":null, + "refs":{ + "PutInventoryRequest$Items":"

The inventory items that you want to add or update on instances.

" } }, - "InventoryItemSchema": { - "base": "

The inventory item schema definition. Users can use this to compose inventory query filters.

", - "refs": { - "InventoryItemSchemaResultList$member": null + "InventoryItemSchema":{ + "base":"

The inventory item schema definition. Users can use this to compose inventory query filters.

", + "refs":{ + "InventoryItemSchemaResultList$member":null } }, - "InventoryItemSchemaResultList": { - "base": null, - "refs": { - "GetInventorySchemaResult$Schemas": "

Inventory schemas returned by the request.

" + "InventoryItemSchemaResultList":{ + "base":null, + "refs":{ + "GetInventorySchemaResult$Schemas":"

Inventory schemas returned by the request.

" } }, - "InventoryItemSchemaVersion": { - "base": null, - "refs": { - "InventoryDeletionSummaryItem$Version": "

The inventory type version.

", - "InventoryItem$SchemaVersion": "

The schema version for the inventory item.

", - "InventoryItemSchema$Version": "

The schema version for the inventory item.

", - "InventoryResultItem$SchemaVersion": "

The schema version for the inventory result item/

", - "ListInventoryEntriesResult$SchemaVersion": "

The inventory schema version used by the instance(s).

" + "InventoryItemSchemaVersion":{ + "base":null, + "refs":{ + "InventoryDeletionSummaryItem$Version":"

The inventory type version.

", + "InventoryItem$SchemaVersion":"

The schema version for the inventory item.

", + "InventoryItemSchema$Version":"

The schema version for the inventory item.

", + "InventoryResultItem$SchemaVersion":"

The schema version for the inventory result item/

", + "ListInventoryEntriesResult$SchemaVersion":"

The inventory schema version used by the instance(s).

" } }, - "InventoryItemTypeName": { - "base": null, - "refs": { - "DeleteInventoryRequest$TypeName": "

The name of the custom inventory type for which you want to delete either all previously collected data, or the inventory type itself.

", - "DeleteInventoryResult$TypeName": "

The name of the inventory data type specified in the request.

", - "InvalidItemContentException$TypeName": null, - "InventoryDeletionStatusItem$TypeName": "

The name of the inventory data type.

", - "InventoryItem$TypeName": "

The name of the inventory type. Default inventory item type names start with AWS. Custom inventory type names will start with Custom. Default inventory item types include the following: AWS:AWSComponent, AWS:Application, AWS:InstanceInformation, AWS:Network, and AWS:WindowsUpdate.

", - "InventoryItemSchema$TypeName": "

The name of the inventory type. Default inventory item type names start with AWS. Custom inventory type names will start with Custom. Default inventory item types include the following: AWS:AWSComponent, AWS:Application, AWS:InstanceInformation, AWS:Network, and AWS:WindowsUpdate.

", - "InventoryResultItem$TypeName": "

The name of the inventory result item type.

", - "ItemContentMismatchException$TypeName": null, - "ItemSizeLimitExceededException$TypeName": null, - "ListInventoryEntriesRequest$TypeName": "

The type of inventory item for which you want information.

", - "ListInventoryEntriesResult$TypeName": "

The type of inventory item returned by the request.

", - "ResultAttribute$TypeName": "

Name of the inventory item type. Valid value: AWS:InstanceInformation. Default Value: AWS:InstanceInformation.

", - "UnsupportedInventoryItemContextException$TypeName": null + "InventoryItemTypeName":{ + "base":null, + "refs":{ + "DeleteInventoryRequest$TypeName":"

The name of the custom inventory type for which you want to delete either all previously collected data, or the inventory type itself.

", + "DeleteInventoryResult$TypeName":"

The name of the inventory data type specified in the request.

", + "InvalidItemContentException$TypeName":null, + "InventoryDeletionStatusItem$TypeName":"

The name of the inventory data type.

", + "InventoryItem$TypeName":"

The name of the inventory type. Default inventory item type names start with AWS. Custom inventory type names will start with Custom. Default inventory item types include the following: AWS:AWSComponent, AWS:Application, AWS:InstanceInformation, AWS:Network, and AWS:WindowsUpdate.

", + "InventoryItemSchema$TypeName":"

The name of the inventory type. Default inventory item type names start with AWS. Custom inventory type names will start with Custom. Default inventory item types include the following: AWS:AWSComponent, AWS:Application, AWS:InstanceInformation, AWS:Network, and AWS:WindowsUpdate.

", + "InventoryResultItem$TypeName":"

The name of the inventory result item type.

", + "ItemContentMismatchException$TypeName":null, + "ItemSizeLimitExceededException$TypeName":null, + "ListInventoryEntriesRequest$TypeName":"

The type of inventory item for which you want information.

", + "ListInventoryEntriesResult$TypeName":"

The type of inventory item returned by the request.

", + "ResultAttribute$TypeName":"

Name of the inventory item type. Valid value: AWS:InstanceInformation. Default Value: AWS:InstanceInformation.

", + "UnsupportedInventoryItemContextException$TypeName":null } }, - "InventoryItemTypeNameFilter": { - "base": null, - "refs": { - "GetInventorySchemaRequest$TypeName": "

The type of inventory item to return.

" + "InventoryItemTypeNameFilter":{ + "base":null, + "refs":{ + "GetInventorySchemaRequest$TypeName":"

The type of inventory item to return.

" } }, - "InventoryQueryOperatorType": { - "base": null, - "refs": { - "InventoryFilter$Type": "

The type of filter. Valid values include the following: \"Equal\"|\"NotEqual\"|\"BeginWith\"|\"LessThan\"|\"GreaterThan\"

" + "InventoryQueryOperatorType":{ + "base":null, + "refs":{ + "InventoryFilter$Type":"

The type of filter. Valid values include the following: \"Equal\"|\"NotEqual\"|\"BeginWith\"|\"LessThan\"|\"GreaterThan\"

" } }, - "InventoryResultEntity": { - "base": "

Inventory query results.

", - "refs": { - "InventoryResultEntityList$member": null + "InventoryResultEntity":{ + "base":"

Inventory query results.

", + "refs":{ + "InventoryResultEntityList$member":null } }, - "InventoryResultEntityId": { - "base": null, - "refs": { - "InventoryResultEntity$Id": "

ID of the inventory result entity. For example, for managed instance inventory the result will be the managed instance ID. For EC2 instance inventory, the result will be the instance ID.

" + "InventoryResultEntityId":{ + "base":null, + "refs":{ + "InventoryResultEntity$Id":"

ID of the inventory result entity. For example, for managed instance inventory the result will be the managed instance ID. For EC2 instance inventory, the result will be the instance ID.

" } }, - "InventoryResultEntityList": { - "base": null, - "refs": { - "GetInventoryResult$Entities": "

Collection of inventory entities such as a collection of instance inventory.

" + "InventoryResultEntityList":{ + "base":null, + "refs":{ + "GetInventoryResult$Entities":"

Collection of inventory entities such as a collection of instance inventory.

" } }, - "InventoryResultItem": { - "base": "

The inventory result item.

", - "refs": { - "InventoryResultItemMap$value": null + "InventoryResultItem":{ + "base":"

The inventory result item.

", + "refs":{ + "InventoryResultItemMap$value":null } }, - "InventoryResultItemKey": { - "base": null, - "refs": { - "InventoryResultItemMap$key": null + "InventoryResultItemKey":{ + "base":null, + "refs":{ + "InventoryResultItemMap$key":null } }, - "InventoryResultItemMap": { - "base": null, - "refs": { - "InventoryResultEntity$Data": "

The data section in the inventory result entity JSON.

" + "InventoryResultItemMap":{ + "base":null, + "refs":{ + "InventoryResultEntity$Data":"

The data section in the inventory result entity JSON.

" } }, - "InventorySchemaDeleteOption": { - "base": null, - "refs": { - "DeleteInventoryRequest$SchemaDeleteOption": "

Use the SchemaDeleteOption to delete a custom inventory type (schema). If you don't choose this option, the system only deletes existing inventory data associated with the custom inventory type. Choose one of the following options:

DisableSchema: If you choose this option, the system ignores all inventory data for the specified version, and any earlier versions. To enable this schema again, you must call the PutInventory action for a version greater than the disbled version.

DeleteSchema: This option deletes the specified custom type from the Inventory service. You can recreate the schema later, if you want.

" + "InventorySchemaDeleteOption":{ + "base":null, + "refs":{ + "DeleteInventoryRequest$SchemaDeleteOption":"

Use the SchemaDeleteOption to delete a custom inventory type (schema). If you don't choose this option, the system only deletes existing inventory data associated with the custom inventory type. Choose one of the following options:

DisableSchema: If you choose this option, the system ignores all inventory data for the specified version, and any earlier versions. To enable this schema again, you must call the PutInventory action for a version greater than the disbled version.

DeleteSchema: This option deletes the specified custom type from the Inventory service. You can recreate the schema later, if you want.

" } }, - "InventoryTypeDisplayName": { - "base": null, - "refs": { - "InventoryItemSchema$DisplayName": "

The alias name of the inventory type. The alias name is used for display purposes.

" + "InventoryTypeDisplayName":{ + "base":null, + "refs":{ + "InventoryItemSchema$DisplayName":"

The alias name of the inventory type. The alias name is used for display purposes.

" } }, - "InvocationDoesNotExist": { - "base": "

The command ID and instance ID you specified did not match any invocations. Verify the command ID adn the instance ID and try again.

", - "refs": { + "InvocationDoesNotExist":{ + "base":"

The command ID and instance ID you specified did not match any invocations. Verify the command ID adn the instance ID and try again.

", + "refs":{ } }, - "InvocationTraceOutput": { - "base": null, - "refs": { - "CommandInvocation$TraceOutput": "

Gets the trace output sent by the agent.

" + "InvocationTraceOutput":{ + "base":null, + "refs":{ + "CommandInvocation$TraceOutput":"

Gets the trace output sent by the agent.

" } }, - "IsSubTypeSchema": { - "base": null, - "refs": { - "GetInventorySchemaRequest$SubType": "

Returns the sub-type schema for a specified inventory type.

" + "IsSubTypeSchema":{ + "base":null, + "refs":{ + "GetInventorySchemaRequest$SubType":"

Returns the sub-type schema for a specified inventory type.

" } }, - "ItemContentMismatchException": { - "base": "

The inventory item has invalid content.

", - "refs": { + "ItemContentMismatchException":{ + "base":"

The inventory item has invalid content.

", + "refs":{ } }, - "ItemSizeLimitExceededException": { - "base": "

The inventory item size has exceeded the size limit.

", - "refs": { + "ItemSizeLimitExceededException":{ + "base":"

The inventory item size has exceeded the size limit.

", + "refs":{ } }, - "KeyList": { - "base": null, - "refs": { - "RemoveTagsFromResourceRequest$TagKeys": "

Tag keys that you want to remove from the specified resource.

" + "KeyList":{ + "base":null, + "refs":{ + "RemoveTagsFromResourceRequest$TagKeys":"

Tag keys that you want to remove from the specified resource.

" } }, - "LastResourceDataSyncMessage": { - "base": null, - "refs": { - "ResourceDataSyncItem$LastSyncStatusMessage": "

The status message details reported by the last sync.

" + "LastResourceDataSyncMessage":{ + "base":null, + "refs":{ + "ResourceDataSyncItem$LastSyncStatusMessage":"

The status message details reported by the last sync.

" } }, - "LastResourceDataSyncStatus": { - "base": null, - "refs": { - "ResourceDataSyncItem$LastStatus": "

The status reported by the last sync.

" + "LastResourceDataSyncStatus":{ + "base":null, + "refs":{ + "ResourceDataSyncItem$LastStatus":"

The status reported by the last sync.

" } }, - "LastResourceDataSyncTime": { - "base": null, - "refs": { - "ResourceDataSyncItem$LastSyncTime": "

The last time the configuration attempted to sync (UTC).

" + "LastResourceDataSyncTime":{ + "base":null, + "refs":{ + "ResourceDataSyncItem$LastSyncTime":"

The last time the configuration attempted to sync (UTC).

" } }, - "LastSuccessfulResourceDataSyncTime": { - "base": null, - "refs": { - "ResourceDataSyncItem$LastSuccessfulSyncTime": "

The last time the sync operations returned a status of SUCCESSFUL (UTC).

" + "LastSuccessfulResourceDataSyncTime":{ + "base":null, + "refs":{ + "ResourceDataSyncItem$LastSuccessfulSyncTime":"

The last time the sync operations returned a status of SUCCESSFUL (UTC).

" } }, - "ListAssociationVersionsRequest": { - "base": null, - "refs": { + "ListAssociationVersionsRequest":{ + "base":null, + "refs":{ } }, - "ListAssociationVersionsResult": { - "base": null, - "refs": { + "ListAssociationVersionsResult":{ + "base":null, + "refs":{ } }, - "ListAssociationsRequest": { - "base": null, - "refs": { + "ListAssociationsRequest":{ + "base":null, + "refs":{ } }, - "ListAssociationsResult": { - "base": null, - "refs": { + "ListAssociationsResult":{ + "base":null, + "refs":{ } }, - "ListCommandInvocationsRequest": { - "base": null, - "refs": { + "ListCommandInvocationsRequest":{ + "base":null, + "refs":{ } }, - "ListCommandInvocationsResult": { - "base": null, - "refs": { + "ListCommandInvocationsResult":{ + "base":null, + "refs":{ } }, - "ListCommandsRequest": { - "base": null, - "refs": { + "ListCommandsRequest":{ + "base":null, + "refs":{ } }, - "ListCommandsResult": { - "base": null, - "refs": { + "ListCommandsResult":{ + "base":null, + "refs":{ } }, - "ListComplianceItemsRequest": { - "base": null, - "refs": { + "ListComplianceItemsRequest":{ + "base":null, + "refs":{ } }, - "ListComplianceItemsResult": { - "base": null, - "refs": { + "ListComplianceItemsResult":{ + "base":null, + "refs":{ } }, - "ListComplianceSummariesRequest": { - "base": null, - "refs": { + "ListComplianceSummariesRequest":{ + "base":null, + "refs":{ } }, - "ListComplianceSummariesResult": { - "base": null, - "refs": { + "ListComplianceSummariesResult":{ + "base":null, + "refs":{ } }, - "ListDocumentVersionsRequest": { - "base": null, - "refs": { + "ListDocumentVersionsRequest":{ + "base":null, + "refs":{ } }, - "ListDocumentVersionsResult": { - "base": null, - "refs": { + "ListDocumentVersionsResult":{ + "base":null, + "refs":{ } }, - "ListDocumentsRequest": { - "base": null, - "refs": { + "ListDocumentsRequest":{ + "base":null, + "refs":{ } }, - "ListDocumentsResult": { - "base": null, - "refs": { + "ListDocumentsResult":{ + "base":null, + "refs":{ } }, - "ListInventoryEntriesRequest": { - "base": null, - "refs": { + "ListInventoryEntriesRequest":{ + "base":null, + "refs":{ } }, - "ListInventoryEntriesResult": { - "base": null, - "refs": { + "ListInventoryEntriesResult":{ + "base":null, + "refs":{ } }, - "ListResourceComplianceSummariesRequest": { - "base": null, - "refs": { + "ListResourceComplianceSummariesRequest":{ + "base":null, + "refs":{ } }, - "ListResourceComplianceSummariesResult": { - "base": null, - "refs": { + "ListResourceComplianceSummariesResult":{ + "base":null, + "refs":{ } }, - "ListResourceDataSyncRequest": { - "base": null, - "refs": { + "ListResourceDataSyncRequest":{ + "base":null, + "refs":{ } }, - "ListResourceDataSyncResult": { - "base": null, - "refs": { + "ListResourceDataSyncResult":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceRequest": { - "base": null, - "refs": { + "ListTagsForResourceRequest":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceResult": { - "base": null, - "refs": { + "ListTagsForResourceResult":{ + "base":null, + "refs":{ } }, - "LoggingInfo": { - "base": "

Information about an Amazon S3 bucket to write instance-level logs to.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "refs": { - "GetMaintenanceWindowTaskResult$LoggingInfo": "

The location in Amazon S3 where the task results are logged.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "MaintenanceWindowTask$LoggingInfo": "

Information about an Amazon S3 bucket to write task-level logs to.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "RegisterTaskWithMaintenanceWindowRequest$LoggingInfo": "

A structure containing information about an Amazon S3 bucket to write instance-level logs to.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "UpdateMaintenanceWindowTaskRequest$LoggingInfo": "

The new logging location in Amazon S3 to specify.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "UpdateMaintenanceWindowTaskResult$LoggingInfo": "

The updated logging information in Amazon S3.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

" + "LoggingInfo":{ + "base":"

Information about an Amazon S3 bucket to write instance-level logs to.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "refs":{ + "GetMaintenanceWindowTaskResult$LoggingInfo":"

The location in Amazon S3 where the task results are logged.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "MaintenanceWindowTask$LoggingInfo":"

Information about an Amazon S3 bucket to write task-level logs to.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "RegisterTaskWithMaintenanceWindowRequest$LoggingInfo":"

A structure containing information about an Amazon S3 bucket to write instance-level logs to.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "UpdateMaintenanceWindowTaskRequest$LoggingInfo":"

The new logging location in Amazon S3 to specify.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "UpdateMaintenanceWindowTaskResult$LoggingInfo":"

The updated logging information in Amazon S3.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

" } }, - "Long": { - "base": null, - "refs": { - "StepExecution$TimeoutSeconds": "

The timeout seconds of the step.

" + "Long":{ + "base":null, + "refs":{ + "StepExecution$TimeoutSeconds":"

The timeout seconds of the step.

" } }, - "MaintenanceWindowAllowUnassociatedTargets": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$AllowUnassociatedTargets": "

Enables a Maintenance Window task to execute on managed instances, even if you have not registered those instances as targets. If enabled, then you must specify the unregistered instances (by instance ID) when you register a task with the Maintenance Window

If you don't enable this option, then you must specify previously-registered targets when you register a task with the Maintenance Window.

", - "GetMaintenanceWindowResult$AllowUnassociatedTargets": "

Whether targets must be registered with the Maintenance Window before tasks can be defined for those targets.

", - "UpdateMaintenanceWindowRequest$AllowUnassociatedTargets": "

Whether targets must be registered with the Maintenance Window before tasks can be defined for those targets.

", - "UpdateMaintenanceWindowResult$AllowUnassociatedTargets": "

Whether targets must be registered with the Maintenance Window before tasks can be defined for those targets.

" + "MaintenanceWindowAllowUnassociatedTargets":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$AllowUnassociatedTargets":"

Enables a Maintenance Window task to execute on managed instances, even if you have not registered those instances as targets. If enabled, then you must specify the unregistered instances (by instance ID) when you register a task with the Maintenance Window

If you don't enable this option, then you must specify previously-registered targets when you register a task with the Maintenance Window.

", + "GetMaintenanceWindowResult$AllowUnassociatedTargets":"

Whether targets must be registered with the Maintenance Window before tasks can be defined for those targets.

", + "UpdateMaintenanceWindowRequest$AllowUnassociatedTargets":"

Whether targets must be registered with the Maintenance Window before tasks can be defined for those targets.

", + "UpdateMaintenanceWindowResult$AllowUnassociatedTargets":"

Whether targets must be registered with the Maintenance Window before tasks can be defined for those targets.

" } }, - "MaintenanceWindowAutomationParameters": { - "base": "

The parameters for an AUTOMATION task type.

", - "refs": { - "MaintenanceWindowTaskInvocationParameters$Automation": "

The parameters for an AUTOMATION task type.

" + "MaintenanceWindowAutomationParameters":{ + "base":"

The parameters for an AUTOMATION task type.

", + "refs":{ + "MaintenanceWindowTaskInvocationParameters$Automation":"

The parameters for an AUTOMATION task type.

" } }, - "MaintenanceWindowCutoff": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$Cutoff": "

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", - "GetMaintenanceWindowResult$Cutoff": "

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", - "MaintenanceWindowIdentity$Cutoff": "

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", - "UpdateMaintenanceWindowRequest$Cutoff": "

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", - "UpdateMaintenanceWindowResult$Cutoff": "

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

" + "MaintenanceWindowCutoff":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$Cutoff":"

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", + "GetMaintenanceWindowResult$Cutoff":"

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", + "MaintenanceWindowIdentity$Cutoff":"

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", + "UpdateMaintenanceWindowRequest$Cutoff":"

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

", + "UpdateMaintenanceWindowResult$Cutoff":"

The number of hours before the end of the Maintenance Window that Systems Manager stops scheduling new tasks for execution.

" } }, - "MaintenanceWindowDescription": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$Description": "

An optional description for the Maintenance Window. We recommend specifying a description to help you organize your Maintenance Windows.

", - "GetMaintenanceWindowResult$Description": "

The description of the Maintenance Window.

", - "GetMaintenanceWindowTaskResult$Description": "

The retrieved task description.

", - "MaintenanceWindowIdentity$Description": "

A description of the Maintenance Window.

", - "MaintenanceWindowTarget$Description": "

A description of the target.

", - "MaintenanceWindowTask$Description": "

A description of the task.

", - "RegisterTargetWithMaintenanceWindowRequest$Description": "

An optional description for the target.

", - "RegisterTaskWithMaintenanceWindowRequest$Description": "

An optional description for the task.

", - "UpdateMaintenanceWindowRequest$Description": "

An optional description for the update request.

", - "UpdateMaintenanceWindowResult$Description": "

An optional description of the update.

", - "UpdateMaintenanceWindowTargetRequest$Description": "

An optional description for the update.

", - "UpdateMaintenanceWindowTargetResult$Description": "

The updated description.

", - "UpdateMaintenanceWindowTaskRequest$Description": "

The new task description to specify.

", - "UpdateMaintenanceWindowTaskResult$Description": "

The updated task description.

" + "MaintenanceWindowDescription":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$Description":"

An optional description for the Maintenance Window. We recommend specifying a description to help you organize your Maintenance Windows.

", + "GetMaintenanceWindowResult$Description":"

The description of the Maintenance Window.

", + "GetMaintenanceWindowTaskResult$Description":"

The retrieved task description.

", + "MaintenanceWindowIdentity$Description":"

A description of the Maintenance Window.

", + "MaintenanceWindowTarget$Description":"

A description of the target.

", + "MaintenanceWindowTask$Description":"

A description of the task.

", + "RegisterTargetWithMaintenanceWindowRequest$Description":"

An optional description for the target.

", + "RegisterTaskWithMaintenanceWindowRequest$Description":"

An optional description for the task.

", + "UpdateMaintenanceWindowRequest$Description":"

An optional description for the update request.

", + "UpdateMaintenanceWindowResult$Description":"

An optional description of the update.

", + "UpdateMaintenanceWindowTargetRequest$Description":"

An optional description for the update.

", + "UpdateMaintenanceWindowTargetResult$Description":"

The updated description.

", + "UpdateMaintenanceWindowTaskRequest$Description":"

The new task description to specify.

", + "UpdateMaintenanceWindowTaskResult$Description":"

The updated task description.

" } }, - "MaintenanceWindowDurationHours": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$Duration": "

The duration of the Maintenance Window in hours.

", - "GetMaintenanceWindowResult$Duration": "

The duration of the Maintenance Window in hours.

", - "MaintenanceWindowIdentity$Duration": "

The duration of the Maintenance Window in hours.

", - "UpdateMaintenanceWindowRequest$Duration": "

The duration of the Maintenance Window in hours.

", - "UpdateMaintenanceWindowResult$Duration": "

The duration of the Maintenance Window in hours.

" + "MaintenanceWindowDurationHours":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$Duration":"

The duration of the Maintenance Window in hours.

", + "GetMaintenanceWindowResult$Duration":"

The duration of the Maintenance Window in hours.

", + "MaintenanceWindowIdentity$Duration":"

The duration of the Maintenance Window in hours.

", + "UpdateMaintenanceWindowRequest$Duration":"

The duration of the Maintenance Window in hours.

", + "UpdateMaintenanceWindowResult$Duration":"

The duration of the Maintenance Window in hours.

" } - }, - "MaintenanceWindowEnabled": { - "base": null, - "refs": { - "GetMaintenanceWindowResult$Enabled": "

Whether the Maintenance Windows is enabled.

", - "MaintenanceWindowIdentity$Enabled": "

Whether the Maintenance Window is enabled.

", - "UpdateMaintenanceWindowRequest$Enabled": "

Whether the Maintenance Window is enabled.

", - "UpdateMaintenanceWindowResult$Enabled": "

Whether the Maintenance Window is enabled.

" - } - }, - "MaintenanceWindowExecution": { - "base": "

Describes the information about an execution of a Maintenance Window.

", - "refs": { - "MaintenanceWindowExecutionList$member": null - } - }, - "MaintenanceWindowExecutionId": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$WindowExecutionId": "

The ID of the Maintenance Window execution the task is part of.

", - "DescribeMaintenanceWindowExecutionTasksRequest$WindowExecutionId": "

The ID of the Maintenance Window execution whose task executions should be retrieved.

", - "GetMaintenanceWindowExecutionRequest$WindowExecutionId": "

The ID of the Maintenance Window execution that includes the task.

", - "GetMaintenanceWindowExecutionResult$WindowExecutionId": "

The ID of the Maintenance Window execution.

", - "GetMaintenanceWindowExecutionTaskInvocationRequest$WindowExecutionId": "

The ID of the Maintenance Window execution for which the task is a part.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$WindowExecutionId": "

The Maintenance Window execution ID.

", - "GetMaintenanceWindowExecutionTaskRequest$WindowExecutionId": "

The ID of the Maintenance Window execution that includes the task.

", - "GetMaintenanceWindowExecutionTaskResult$WindowExecutionId": "

The ID of the Maintenance Window execution that includes the task.

", - "MaintenanceWindowExecution$WindowExecutionId": "

The ID of the Maintenance Window execution.

", - "MaintenanceWindowExecutionTaskIdentity$WindowExecutionId": "

The ID of the Maintenance Window execution that ran the task.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$WindowExecutionId": "

The ID of the Maintenance Window execution that ran the task.

" - } - }, - "MaintenanceWindowExecutionList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionsResult$WindowExecutions": "

Information about the Maintenance Windows execution.

" - } - }, - "MaintenanceWindowExecutionStatus": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionResult$Status": "

The status of the Maintenance Window execution.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$Status": "

The task status for an invocation.

", - "GetMaintenanceWindowExecutionTaskResult$Status": "

The status of the task.

", - "MaintenanceWindowExecution$Status": "

The status of the execution.

", - "MaintenanceWindowExecutionTaskIdentity$Status": "

The status of the task execution.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$Status": "

The status of the task invocation.

" - } - }, - "MaintenanceWindowExecutionStatusDetails": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionResult$StatusDetails": "

The details explaining the Status. Only available for certain status values.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$StatusDetails": "

The details explaining the status. Details are only available for certain status values.

", - "GetMaintenanceWindowExecutionTaskResult$StatusDetails": "

The details explaining the Status. Only available for certain status values.

", - "MaintenanceWindowExecution$StatusDetails": "

The details explaining the Status. Only available for certain status values.

", - "MaintenanceWindowExecutionTaskIdentity$StatusDetails": "

The details explaining the status of the task execution. Only available for certain status values.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$StatusDetails": "

The details explaining the status of the task invocation. Only available for certain Status values.

" - } - }, - "MaintenanceWindowExecutionTaskExecutionId": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskInvocationResult$ExecutionId": "

The execution ID.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$ExecutionId": "

The ID of the action performed in the service that actually handled the task invocation. If the task type is RUN_COMMAND, this value is the command ID.

" - } - }, - "MaintenanceWindowExecutionTaskId": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$TaskId": "

The ID of the specific task in the Maintenance Window task that should be retrieved.

", - "GetMaintenanceWindowExecutionTaskInvocationRequest$TaskId": "

The ID of the specific task in the Maintenance Window task that should be retrieved.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$TaskExecutionId": "

The task execution ID.

", - "GetMaintenanceWindowExecutionTaskRequest$TaskId": "

The ID of the specific task execution in the Maintenance Window task that should be retrieved.

", - "GetMaintenanceWindowExecutionTaskResult$TaskExecutionId": "

The ID of the specific task execution in the Maintenance Window task that was retrieved.

", - "MaintenanceWindowExecutionTaskIdList$member": null, - "MaintenanceWindowExecutionTaskIdentity$TaskExecutionId": "

The ID of the specific task execution in the Maintenance Window execution.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$TaskExecutionId": "

The ID of the specific task execution in the Maintenance Window execution.

" - } - }, - "MaintenanceWindowExecutionTaskIdList": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionResult$TaskIds": "

The ID of the task executions from the Maintenance Window execution.

" - } - }, - "MaintenanceWindowExecutionTaskIdentity": { - "base": "

Information about a task execution performed as part of a Maintenance Window execution.

", - "refs": { - "MaintenanceWindowExecutionTaskIdentityList$member": null - } - }, - "MaintenanceWindowExecutionTaskIdentityList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionTasksResult$WindowExecutionTaskIdentities": "

Information about the task executions.

" - } - }, - "MaintenanceWindowExecutionTaskInvocationId": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskInvocationRequest$InvocationId": "

The invocation ID to retrieve.

", - "GetMaintenanceWindowExecutionTaskInvocationResult$InvocationId": "

The invocation ID.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$InvocationId": "

The ID of the task invocation.

" - } - }, - "MaintenanceWindowExecutionTaskInvocationIdentity": { - "base": "

Describes the information about a task invocation for a particular target as part of a task execution performed as part of a Maintenance Window execution.

", - "refs": { - "MaintenanceWindowExecutionTaskInvocationIdentityList$member": null - } - }, - "MaintenanceWindowExecutionTaskInvocationIdentityList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionTaskInvocationsResult$WindowExecutionTaskInvocationIdentities": "

Information about the task invocation results per invocation.

" - } - }, - "MaintenanceWindowExecutionTaskInvocationParameters": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskInvocationResult$Parameters": "

The parameters used at the time that the task executed.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$Parameters": "

The parameters that were provided for the invocation when it was executed.

" - } - }, - "MaintenanceWindowFilter": { - "base": "

Filter used in the request.

", - "refs": { - "MaintenanceWindowFilterList$member": null + }, + "MaintenanceWindowEnabled":{ + "base":null, + "refs":{ + "GetMaintenanceWindowResult$Enabled":"

Whether the Maintenance Windows is enabled.

", + "MaintenanceWindowIdentity$Enabled":"

Whether the Maintenance Window is enabled.

", + "UpdateMaintenanceWindowRequest$Enabled":"

Whether the Maintenance Window is enabled.

", + "UpdateMaintenanceWindowResult$Enabled":"

Whether the Maintenance Window is enabled.

" + } + }, + "MaintenanceWindowExecution":{ + "base":"

Describes the information about an execution of a Maintenance Window.

", + "refs":{ + "MaintenanceWindowExecutionList$member":null + } + }, + "MaintenanceWindowExecutionId":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$WindowExecutionId":"

The ID of the Maintenance Window execution the task is part of.

", + "DescribeMaintenanceWindowExecutionTasksRequest$WindowExecutionId":"

The ID of the Maintenance Window execution whose task executions should be retrieved.

", + "GetMaintenanceWindowExecutionRequest$WindowExecutionId":"

The ID of the Maintenance Window execution that includes the task.

", + "GetMaintenanceWindowExecutionResult$WindowExecutionId":"

The ID of the Maintenance Window execution.

", + "GetMaintenanceWindowExecutionTaskInvocationRequest$WindowExecutionId":"

The ID of the Maintenance Window execution for which the task is a part.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$WindowExecutionId":"

The Maintenance Window execution ID.

", + "GetMaintenanceWindowExecutionTaskRequest$WindowExecutionId":"

The ID of the Maintenance Window execution that includes the task.

", + "GetMaintenanceWindowExecutionTaskResult$WindowExecutionId":"

The ID of the Maintenance Window execution that includes the task.

", + "MaintenanceWindowExecution$WindowExecutionId":"

The ID of the Maintenance Window execution.

", + "MaintenanceWindowExecutionTaskIdentity$WindowExecutionId":"

The ID of the Maintenance Window execution that ran the task.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$WindowExecutionId":"

The ID of the Maintenance Window execution that ran the task.

" + } + }, + "MaintenanceWindowExecutionList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionsResult$WindowExecutions":"

Information about the Maintenance Windows execution.

" + } + }, + "MaintenanceWindowExecutionStatus":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionResult$Status":"

The status of the Maintenance Window execution.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$Status":"

The task status for an invocation.

", + "GetMaintenanceWindowExecutionTaskResult$Status":"

The status of the task.

", + "MaintenanceWindowExecution$Status":"

The status of the execution.

", + "MaintenanceWindowExecutionTaskIdentity$Status":"

The status of the task execution.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$Status":"

The status of the task invocation.

" + } + }, + "MaintenanceWindowExecutionStatusDetails":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionResult$StatusDetails":"

The details explaining the Status. Only available for certain status values.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$StatusDetails":"

The details explaining the status. Details are only available for certain status values.

", + "GetMaintenanceWindowExecutionTaskResult$StatusDetails":"

The details explaining the Status. Only available for certain status values.

", + "MaintenanceWindowExecution$StatusDetails":"

The details explaining the Status. Only available for certain status values.

", + "MaintenanceWindowExecutionTaskIdentity$StatusDetails":"

The details explaining the status of the task execution. Only available for certain status values.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$StatusDetails":"

The details explaining the status of the task invocation. Only available for certain Status values.

" + } + }, + "MaintenanceWindowExecutionTaskExecutionId":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskInvocationResult$ExecutionId":"

The execution ID.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$ExecutionId":"

The ID of the action performed in the service that actually handled the task invocation. If the task type is RUN_COMMAND, this value is the command ID.

" + } + }, + "MaintenanceWindowExecutionTaskId":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$TaskId":"

The ID of the specific task in the Maintenance Window task that should be retrieved.

", + "GetMaintenanceWindowExecutionTaskInvocationRequest$TaskId":"

The ID of the specific task in the Maintenance Window task that should be retrieved.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$TaskExecutionId":"

The task execution ID.

", + "GetMaintenanceWindowExecutionTaskRequest$TaskId":"

The ID of the specific task execution in the Maintenance Window task that should be retrieved.

", + "GetMaintenanceWindowExecutionTaskResult$TaskExecutionId":"

The ID of the specific task execution in the Maintenance Window task that was retrieved.

", + "MaintenanceWindowExecutionTaskIdList$member":null, + "MaintenanceWindowExecutionTaskIdentity$TaskExecutionId":"

The ID of the specific task execution in the Maintenance Window execution.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$TaskExecutionId":"

The ID of the specific task execution in the Maintenance Window execution.

" + } + }, + "MaintenanceWindowExecutionTaskIdList":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionResult$TaskIds":"

The ID of the task executions from the Maintenance Window execution.

" + } + }, + "MaintenanceWindowExecutionTaskIdentity":{ + "base":"

Information about a task execution performed as part of a Maintenance Window execution.

", + "refs":{ + "MaintenanceWindowExecutionTaskIdentityList$member":null + } + }, + "MaintenanceWindowExecutionTaskIdentityList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionTasksResult$WindowExecutionTaskIdentities":"

Information about the task executions.

" + } + }, + "MaintenanceWindowExecutionTaskInvocationId":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskInvocationRequest$InvocationId":"

The invocation ID to retrieve.

", + "GetMaintenanceWindowExecutionTaskInvocationResult$InvocationId":"

The invocation ID.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$InvocationId":"

The ID of the task invocation.

" + } + }, + "MaintenanceWindowExecutionTaskInvocationIdentity":{ + "base":"

Describes the information about a task invocation for a particular target as part of a task execution performed as part of a Maintenance Window execution.

", + "refs":{ + "MaintenanceWindowExecutionTaskInvocationIdentityList$member":null + } + }, + "MaintenanceWindowExecutionTaskInvocationIdentityList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionTaskInvocationsResult$WindowExecutionTaskInvocationIdentities":"

Information about the task invocation results per invocation.

" + } + }, + "MaintenanceWindowExecutionTaskInvocationParameters":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskInvocationResult$Parameters":"

The parameters used at the time that the task executed.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$Parameters":"

The parameters that were provided for the invocation when it was executed.

" + } + }, + "MaintenanceWindowFilter":{ + "base":"

Filter used in the request.

", + "refs":{ + "MaintenanceWindowFilterList$member":null } }, - "MaintenanceWindowFilterKey": { - "base": null, - "refs": { - "MaintenanceWindowFilter$Key": "

The name of the filter.

" - } - }, - "MaintenanceWindowFilterList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$Filters": "

Optional filters used to scope down the returned task invocations. The supported filter key is STATUS with the corresponding values PENDING, IN_PROGRESS, SUCCESS, FAILED, TIMED_OUT, CANCELLING, and CANCELLED.

", - "DescribeMaintenanceWindowExecutionTasksRequest$Filters": "

Optional filters used to scope down the returned tasks. The supported filter key is STATUS with the corresponding values PENDING, IN_PROGRESS, SUCCESS, FAILED, TIMED_OUT, CANCELLING, and CANCELLED.

", - "DescribeMaintenanceWindowExecutionsRequest$Filters": "

Each entry in the array is a structure containing:

Key (string, between 1 and 128 characters)

Values (array of strings, each string is between 1 and 256 characters)

The supported Keys are ExecutedBefore and ExecutedAfter with the value being a date/time string such as 2016-11-04T05:00:00Z.

", - "DescribeMaintenanceWindowTargetsRequest$Filters": "

Optional filters that can be used to narrow down the scope of the returned window targets. The supported filter keys are Type, WindowTargetId and OwnerInformation.

", - "DescribeMaintenanceWindowTasksRequest$Filters": "

Optional filters used to narrow down the scope of the returned tasks. The supported filter keys are WindowTaskId, TaskArn, Priority, and TaskType.

", - "DescribeMaintenanceWindowsRequest$Filters": "

Optional filters used to narrow down the scope of the returned Maintenance Windows. Supported filter keys are Name and Enabled.

" - } - }, - "MaintenanceWindowFilterValue": { - "base": null, - "refs": { - "MaintenanceWindowFilterValues$member": null - } - }, - "MaintenanceWindowFilterValues": { - "base": null, - "refs": { - "MaintenanceWindowFilter$Values": "

The filter values.

" - } - }, - "MaintenanceWindowId": { - "base": null, - "refs": { - "CreateMaintenanceWindowResult$WindowId": "

The ID of the created Maintenance Window.

", - "DeleteMaintenanceWindowRequest$WindowId": "

The ID of the Maintenance Window to delete.

", - "DeleteMaintenanceWindowResult$WindowId": "

The ID of the deleted Maintenance Window.

", - "DeregisterTargetFromMaintenanceWindowRequest$WindowId": "

The ID of the Maintenance Window the target should be removed from.

", - "DeregisterTargetFromMaintenanceWindowResult$WindowId": "

The ID of the Maintenance Window the target was removed from.

", - "DeregisterTaskFromMaintenanceWindowRequest$WindowId": "

The ID of the Maintenance Window the task should be removed from.

", - "DeregisterTaskFromMaintenanceWindowResult$WindowId": "

The ID of the Maintenance Window the task was removed from.

", - "DescribeMaintenanceWindowExecutionsRequest$WindowId": "

The ID of the Maintenance Window whose executions should be retrieved.

", - "DescribeMaintenanceWindowTargetsRequest$WindowId": "

The ID of the Maintenance Window whose targets should be retrieved.

", - "DescribeMaintenanceWindowTasksRequest$WindowId": "

The ID of the Maintenance Window whose tasks should be retrieved.

", - "GetMaintenanceWindowRequest$WindowId": "

The ID of the desired Maintenance Window.

", - "GetMaintenanceWindowResult$WindowId": "

The ID of the created Maintenance Window.

", - "GetMaintenanceWindowTaskRequest$WindowId": "

The Maintenance Window ID that includes the task to retrieve.

", - "GetMaintenanceWindowTaskResult$WindowId": "

The retrieved Maintenance Window ID.

", - "MaintenanceWindowExecution$WindowId": "

The ID of the Maintenance Window.

", - "MaintenanceWindowIdentity$WindowId": "

The ID of the Maintenance Window.

", - "MaintenanceWindowTarget$WindowId": "

The Maintenance Window ID where the target is registered.

", - "MaintenanceWindowTask$WindowId": "

The Maintenance Window ID where the task is registered.

", - "RegisterTargetWithMaintenanceWindowRequest$WindowId": "

The ID of the Maintenance Window the target should be registered with.

", - "RegisterTaskWithMaintenanceWindowRequest$WindowId": "

The ID of the Maintenance Window the task should be added to.

", - "UpdateMaintenanceWindowRequest$WindowId": "

The ID of the Maintenance Window to update.

", - "UpdateMaintenanceWindowResult$WindowId": "

The ID of the created Maintenance Window.

", - "UpdateMaintenanceWindowTargetRequest$WindowId": "

The Maintenance Window ID with which to modify the target.

", - "UpdateMaintenanceWindowTargetResult$WindowId": "

The Maintenance Window ID specified in the update request.

", - "UpdateMaintenanceWindowTaskRequest$WindowId": "

The Maintenance Window ID that contains the task to modify.

", - "UpdateMaintenanceWindowTaskResult$WindowId": "

The ID of the Maintenance Window that was updated.

" - } + "MaintenanceWindowFilterKey":{ + "base":null, + "refs":{ + "MaintenanceWindowFilter$Key":"

The name of the filter.

" + } + }, + "MaintenanceWindowFilterList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$Filters":"

Optional filters used to scope down the returned task invocations. The supported filter key is STATUS with the corresponding values PENDING, IN_PROGRESS, SUCCESS, FAILED, TIMED_OUT, CANCELLING, and CANCELLED.

", + "DescribeMaintenanceWindowExecutionTasksRequest$Filters":"

Optional filters used to scope down the returned tasks. The supported filter key is STATUS with the corresponding values PENDING, IN_PROGRESS, SUCCESS, FAILED, TIMED_OUT, CANCELLING, and CANCELLED.

", + "DescribeMaintenanceWindowExecutionsRequest$Filters":"

Each entry in the array is a structure containing:

Key (string, between 1 and 128 characters)

Values (array of strings, each string is between 1 and 256 characters)

The supported Keys are ExecutedBefore and ExecutedAfter with the value being a date/time string such as 2016-11-04T05:00:00Z.

", + "DescribeMaintenanceWindowTargetsRequest$Filters":"

Optional filters that can be used to narrow down the scope of the returned window targets. The supported filter keys are Type, WindowTargetId and OwnerInformation.

", + "DescribeMaintenanceWindowTasksRequest$Filters":"

Optional filters used to narrow down the scope of the returned tasks. The supported filter keys are WindowTaskId, TaskArn, Priority, and TaskType.

", + "DescribeMaintenanceWindowsRequest$Filters":"

Optional filters used to narrow down the scope of the returned Maintenance Windows. Supported filter keys are Name and Enabled.

" + } + }, + "MaintenanceWindowFilterValue":{ + "base":null, + "refs":{ + "MaintenanceWindowFilterValues$member":null + } + }, + "MaintenanceWindowFilterValues":{ + "base":null, + "refs":{ + "MaintenanceWindowFilter$Values":"

The filter values.

" + } + }, + "MaintenanceWindowId":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowResult$WindowId":"

The ID of the created Maintenance Window.

", + "DeleteMaintenanceWindowRequest$WindowId":"

The ID of the Maintenance Window to delete.

", + "DeleteMaintenanceWindowResult$WindowId":"

The ID of the deleted Maintenance Window.

", + "DeregisterTargetFromMaintenanceWindowRequest$WindowId":"

The ID of the Maintenance Window the target should be removed from.

", + "DeregisterTargetFromMaintenanceWindowResult$WindowId":"

The ID of the Maintenance Window the target was removed from.

", + "DeregisterTaskFromMaintenanceWindowRequest$WindowId":"

The ID of the Maintenance Window the task should be removed from.

", + "DeregisterTaskFromMaintenanceWindowResult$WindowId":"

The ID of the Maintenance Window the task was removed from.

", + "DescribeMaintenanceWindowExecutionsRequest$WindowId":"

The ID of the Maintenance Window whose executions should be retrieved.

", + "DescribeMaintenanceWindowTargetsRequest$WindowId":"

The ID of the Maintenance Window whose targets should be retrieved.

", + "DescribeMaintenanceWindowTasksRequest$WindowId":"

The ID of the Maintenance Window whose tasks should be retrieved.

", + "GetMaintenanceWindowRequest$WindowId":"

The ID of the desired Maintenance Window.

", + "GetMaintenanceWindowResult$WindowId":"

The ID of the created Maintenance Window.

", + "GetMaintenanceWindowTaskRequest$WindowId":"

The Maintenance Window ID that includes the task to retrieve.

", + "GetMaintenanceWindowTaskResult$WindowId":"

The retrieved Maintenance Window ID.

", + "MaintenanceWindowExecution$WindowId":"

The ID of the Maintenance Window.

", + "MaintenanceWindowIdentity$WindowId":"

The ID of the Maintenance Window.

", + "MaintenanceWindowTarget$WindowId":"

The Maintenance Window ID where the target is registered.

", + "MaintenanceWindowTask$WindowId":"

The Maintenance Window ID where the task is registered.

", + "RegisterTargetWithMaintenanceWindowRequest$WindowId":"

The ID of the Maintenance Window the target should be registered with.

", + "RegisterTaskWithMaintenanceWindowRequest$WindowId":"

The ID of the Maintenance Window the task should be added to.

", + "UpdateMaintenanceWindowRequest$WindowId":"

The ID of the Maintenance Window to update.

", + "UpdateMaintenanceWindowResult$WindowId":"

The ID of the created Maintenance Window.

", + "UpdateMaintenanceWindowTargetRequest$WindowId":"

The Maintenance Window ID with which to modify the target.

", + "UpdateMaintenanceWindowTargetResult$WindowId":"

The Maintenance Window ID specified in the update request.

", + "UpdateMaintenanceWindowTaskRequest$WindowId":"

The Maintenance Window ID that contains the task to modify.

", + "UpdateMaintenanceWindowTaskResult$WindowId":"

The ID of the Maintenance Window that was updated.

" + } }, - "MaintenanceWindowIdentity": { - "base": "

Information about the Maintenance Window.

", - "refs": { - "MaintenanceWindowIdentityList$member": null - } + "MaintenanceWindowIdentity":{ + "base":"

Information about the Maintenance Window.

", + "refs":{ + "MaintenanceWindowIdentityList$member":null + } }, - "MaintenanceWindowIdentityList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowsResult$WindowIdentities": "

Information about the Maintenance Windows.

" - } + "MaintenanceWindowIdentityList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowsResult$WindowIdentities":"

Information about the Maintenance Windows.

" + } }, - "MaintenanceWindowLambdaClientContext": { - "base": null, - "refs": { - "MaintenanceWindowLambdaParameters$ClientContext": "

Pass client-specific information to the Lambda function that you are invoking. You can then process the client information in your Lambda function as you choose through the context variable.

" - } - }, - "MaintenanceWindowLambdaParameters": { - "base": "

The parameters for a LAMBDA task type.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For Lambda tasks, Systems Manager ignores any values specified for TaskParameters and LoggingInfo.

", - "refs": { - "MaintenanceWindowTaskInvocationParameters$Lambda": "

The parameters for a LAMBDA task type.

" - } - }, - "MaintenanceWindowLambdaPayload": { - "base": null, - "refs": { - "MaintenanceWindowLambdaParameters$Payload": "

JSON to provide to your Lambda function as input.

" - } - }, - "MaintenanceWindowLambdaQualifier": { - "base": null, - "refs": { - "MaintenanceWindowLambdaParameters$Qualifier": "

(Optional) Specify a Lambda function version or alias name. If you specify a function version, the action uses the qualified function ARN to invoke a specific Lambda function. If you specify an alias name, the action uses the alias ARN to invoke the Lambda function version to which the alias points.

" - } - }, - "MaintenanceWindowMaxResults": { - "base": null, - "refs": { - "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeMaintenanceWindowExecutionTasksRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeMaintenanceWindowExecutionsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeMaintenanceWindowTargetsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeMaintenanceWindowTasksRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeMaintenanceWindowsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" - } - }, - "MaintenanceWindowName": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$Name": "

The name of the Maintenance Window.

", - "GetMaintenanceWindowResult$Name": "

The name of the Maintenance Window.

", - "GetMaintenanceWindowTaskResult$Name": "

The retrieved task name.

", - "MaintenanceWindowIdentity$Name": "

The name of the Maintenance Window.

", - "MaintenanceWindowTarget$Name": "

The target name.

", - "MaintenanceWindowTask$Name": "

The task name.

", - "RegisterTargetWithMaintenanceWindowRequest$Name": "

An optional name for the target.

", - "RegisterTaskWithMaintenanceWindowRequest$Name": "

An optional name for the task.

", - "UpdateMaintenanceWindowRequest$Name": "

The name of the Maintenance Window.

", - "UpdateMaintenanceWindowResult$Name": "

The name of the Maintenance Window.

", - "UpdateMaintenanceWindowTargetRequest$Name": "

A name for the update.

", - "UpdateMaintenanceWindowTargetResult$Name": "

The updated name.

", - "UpdateMaintenanceWindowTaskRequest$Name": "

The new task name to specify.

", - "UpdateMaintenanceWindowTaskResult$Name": "

The updated task name.

" - } - }, - "MaintenanceWindowResourceType": { - "base": null, - "refs": { - "MaintenanceWindowTarget$ResourceType": "

The type of target.

", - "RegisterTargetWithMaintenanceWindowRequest$ResourceType": "

The type of target being registered with the Maintenance Window.

" - } - }, - "MaintenanceWindowRunCommandParameters": { - "base": "

The parameters for a RUN_COMMAND task type.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For Run Command tasks, Systems Manager uses specified values for TaskParameters and LoggingInfo only if no values are specified for TaskInvocationParameters.

", - "refs": { - "MaintenanceWindowTaskInvocationParameters$RunCommand": "

The parameters for a RUN_COMMAND task type.

" - } - }, - "MaintenanceWindowSchedule": { - "base": null, - "refs": { - "CreateMaintenanceWindowRequest$Schedule": "

The schedule of the Maintenance Window in the form of a cron or rate expression.

", - "GetMaintenanceWindowResult$Schedule": "

The schedule of the Maintenance Window in the form of a cron or rate expression.

", - "UpdateMaintenanceWindowRequest$Schedule": "

The schedule of the Maintenance Window in the form of a cron or rate expression.

", - "UpdateMaintenanceWindowResult$Schedule": "

The schedule of the Maintenance Window in the form of a cron or rate expression.

" + "MaintenanceWindowLambdaClientContext":{ + "base":null, + "refs":{ + "MaintenanceWindowLambdaParameters$ClientContext":"

Pass client-specific information to the Lambda function that you are invoking. You can then process the client information in your Lambda function as you choose through the context variable.

" + } + }, + "MaintenanceWindowLambdaParameters":{ + "base":"

The parameters for a LAMBDA task type.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For Lambda tasks, Systems Manager ignores any values specified for TaskParameters and LoggingInfo.

", + "refs":{ + "MaintenanceWindowTaskInvocationParameters$Lambda":"

The parameters for a LAMBDA task type.

" + } + }, + "MaintenanceWindowLambdaPayload":{ + "base":null, + "refs":{ + "MaintenanceWindowLambdaParameters$Payload":"

JSON to provide to your Lambda function as input.

" + } + }, + "MaintenanceWindowLambdaQualifier":{ + "base":null, + "refs":{ + "MaintenanceWindowLambdaParameters$Qualifier":"

(Optional) Specify a Lambda function version or alias name. If you specify a function version, the action uses the qualified function ARN to invoke a specific Lambda function. If you specify an alias name, the action uses the alias ARN to invoke the Lambda function version to which the alias points.

" + } + }, + "MaintenanceWindowMaxResults":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeMaintenanceWindowExecutionTasksRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeMaintenanceWindowExecutionsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeMaintenanceWindowTargetsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeMaintenanceWindowTasksRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeMaintenanceWindowsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + } + }, + "MaintenanceWindowName":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$Name":"

The name of the Maintenance Window.

", + "GetMaintenanceWindowResult$Name":"

The name of the Maintenance Window.

", + "GetMaintenanceWindowTaskResult$Name":"

The retrieved task name.

", + "MaintenanceWindowIdentity$Name":"

The name of the Maintenance Window.

", + "MaintenanceWindowTarget$Name":"

The target name.

", + "MaintenanceWindowTask$Name":"

The task name.

", + "RegisterTargetWithMaintenanceWindowRequest$Name":"

An optional name for the target.

", + "RegisterTaskWithMaintenanceWindowRequest$Name":"

An optional name for the task.

", + "UpdateMaintenanceWindowRequest$Name":"

The name of the Maintenance Window.

", + "UpdateMaintenanceWindowResult$Name":"

The name of the Maintenance Window.

", + "UpdateMaintenanceWindowTargetRequest$Name":"

A name for the update.

", + "UpdateMaintenanceWindowTargetResult$Name":"

The updated name.

", + "UpdateMaintenanceWindowTaskRequest$Name":"

The new task name to specify.

", + "UpdateMaintenanceWindowTaskResult$Name":"

The updated task name.

" + } + }, + "MaintenanceWindowResourceType":{ + "base":null, + "refs":{ + "MaintenanceWindowTarget$ResourceType":"

The type of target.

", + "RegisterTargetWithMaintenanceWindowRequest$ResourceType":"

The type of target being registered with the Maintenance Window.

" + } + }, + "MaintenanceWindowRunCommandParameters":{ + "base":"

The parameters for a RUN_COMMAND task type.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For Run Command tasks, Systems Manager uses specified values for TaskParameters and LoggingInfo only if no values are specified for TaskInvocationParameters.

", + "refs":{ + "MaintenanceWindowTaskInvocationParameters$RunCommand":"

The parameters for a RUN_COMMAND task type.

" + } + }, + "MaintenanceWindowSchedule":{ + "base":null, + "refs":{ + "CreateMaintenanceWindowRequest$Schedule":"

The schedule of the Maintenance Window in the form of a cron or rate expression.

", + "GetMaintenanceWindowResult$Schedule":"

The schedule of the Maintenance Window in the form of a cron or rate expression.

", + "UpdateMaintenanceWindowRequest$Schedule":"

The schedule of the Maintenance Window in the form of a cron or rate expression.

", + "UpdateMaintenanceWindowResult$Schedule":"

The schedule of the Maintenance Window in the form of a cron or rate expression.

" } }, - "MaintenanceWindowStepFunctionsInput": { - "base": null, - "refs": { - "MaintenanceWindowStepFunctionsParameters$Input": "

The inputs for the STEP_FUNCTION task.

" + "MaintenanceWindowStepFunctionsInput":{ + "base":null, + "refs":{ + "MaintenanceWindowStepFunctionsParameters$Input":"

The inputs for the STEP_FUNCTION task.

" } }, - "MaintenanceWindowStepFunctionsName": { - "base": null, - "refs": { - "MaintenanceWindowStepFunctionsParameters$Name": "

The name of the STEP_FUNCTION task.

" + "MaintenanceWindowStepFunctionsName":{ + "base":null, + "refs":{ + "MaintenanceWindowStepFunctionsParameters$Name":"

The name of the STEP_FUNCTION task.

" } }, - "MaintenanceWindowStepFunctionsParameters": { - "base": "

The parameters for a STEP_FUNCTION task.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For Step Functions tasks, Systems Manager ignores any values specified for TaskParameters and LoggingInfo.

", - "refs": { - "MaintenanceWindowTaskInvocationParameters$StepFunctions": "

The parameters for a STEP_FUNCTION task type.

" + "MaintenanceWindowStepFunctionsParameters":{ + "base":"

The parameters for a STEP_FUNCTION task.

For information about specifying and updating task parameters, see RegisterTaskWithMaintenanceWindow and UpdateMaintenanceWindowTask.

LoggingInfo has been deprecated. To specify an S3 bucket to contain logs, instead use the OutputS3BucketName and OutputS3KeyPrefix options in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

For Step Functions tasks, Systems Manager ignores any values specified for TaskParameters and LoggingInfo.

", + "refs":{ + "MaintenanceWindowTaskInvocationParameters$StepFunctions":"

The parameters for a STEP_FUNCTION task type.

" } }, - "MaintenanceWindowTarget": { - "base": "

The target registered with the Maintenance Window.

", - "refs": { - "MaintenanceWindowTargetList$member": null + "MaintenanceWindowTarget":{ + "base":"

The target registered with the Maintenance Window.

", + "refs":{ + "MaintenanceWindowTargetList$member":null } }, - "MaintenanceWindowTargetId": { - "base": null, - "refs": { - "DeregisterTargetFromMaintenanceWindowRequest$WindowTargetId": "

The ID of the target definition to remove.

", - "DeregisterTargetFromMaintenanceWindowResult$WindowTargetId": "

The ID of the removed target definition.

", - "MaintenanceWindowTarget$WindowTargetId": "

The ID of the target.

", - "RegisterTargetWithMaintenanceWindowResult$WindowTargetId": "

The ID of the target definition in this Maintenance Window.

", - "UpdateMaintenanceWindowTargetRequest$WindowTargetId": "

The target ID to modify.

", - "UpdateMaintenanceWindowTargetResult$WindowTargetId": "

The target ID specified in the update request.

" + "MaintenanceWindowTargetId":{ + "base":null, + "refs":{ + "DeregisterTargetFromMaintenanceWindowRequest$WindowTargetId":"

The ID of the target definition to remove.

", + "DeregisterTargetFromMaintenanceWindowResult$WindowTargetId":"

The ID of the removed target definition.

", + "MaintenanceWindowTarget$WindowTargetId":"

The ID of the target.

", + "RegisterTargetWithMaintenanceWindowResult$WindowTargetId":"

The ID of the target definition in this Maintenance Window.

", + "UpdateMaintenanceWindowTargetRequest$WindowTargetId":"

The target ID to modify.

", + "UpdateMaintenanceWindowTargetResult$WindowTargetId":"

The target ID specified in the update request.

" } }, - "MaintenanceWindowTargetList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowTargetsResult$Targets": "

Information about the targets in the Maintenance Window.

" + "MaintenanceWindowTargetList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowTargetsResult$Targets":"

Information about the targets in the Maintenance Window.

" } }, - "MaintenanceWindowTask": { - "base": "

Information about a task defined for a Maintenance Window.

", - "refs": { - "MaintenanceWindowTaskList$member": null + "MaintenanceWindowTask":{ + "base":"

Information about a task defined for a Maintenance Window.

", + "refs":{ + "MaintenanceWindowTaskList$member":null } }, - "MaintenanceWindowTaskArn": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskResult$TaskArn": "

The ARN of the executed task.

", - "GetMaintenanceWindowTaskResult$TaskArn": "

The resource that the task used during execution. For RUN_COMMAND and AUTOMATION task types, the TaskArn is the Systems Manager Document name/ARN. For LAMBDA tasks, the value is the function name/ARN. For STEP_FUNCTION tasks, the value is the state machine ARN.

", - "MaintenanceWindowExecutionTaskIdentity$TaskArn": "

The ARN of the executed task.

", - "MaintenanceWindowTask$TaskArn": "

The resource that the task uses during execution. For RUN_COMMAND and AUTOMATION task types, TaskArn is the Systems Manager document name or ARN. For LAMBDA tasks, it's the function name or ARN. For STEP_FUNCTION tasks, it's the state machine ARN.

", - "RegisterTaskWithMaintenanceWindowRequest$TaskArn": "

The ARN of the task to execute

", - "UpdateMaintenanceWindowTaskRequest$TaskArn": "

The task ARN to modify.

", - "UpdateMaintenanceWindowTaskResult$TaskArn": "

The updated task ARN value.

" + "MaintenanceWindowTaskArn":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskResult$TaskArn":"

The ARN of the executed task.

", + "GetMaintenanceWindowTaskResult$TaskArn":"

The resource that the task used during execution. For RUN_COMMAND and AUTOMATION task types, the TaskArn is the Systems Manager Document name/ARN. For LAMBDA tasks, the value is the function name/ARN. For STEP_FUNCTION tasks, the value is the state machine ARN.

", + "MaintenanceWindowExecutionTaskIdentity$TaskArn":"

The ARN of the executed task.

", + "MaintenanceWindowTask$TaskArn":"

The resource that the task uses during execution. For RUN_COMMAND and AUTOMATION task types, TaskArn is the Systems Manager document name or ARN. For LAMBDA tasks, it's the function name or ARN. For STEP_FUNCTION tasks, it's the state machine ARN.

", + "RegisterTaskWithMaintenanceWindowRequest$TaskArn":"

The ARN of the task to execute

", + "UpdateMaintenanceWindowTaskRequest$TaskArn":"

The task ARN to modify.

", + "UpdateMaintenanceWindowTaskResult$TaskArn":"

The updated task ARN value.

" } - }, - "MaintenanceWindowTaskId": { - "base": null, - "refs": { - "DeregisterTaskFromMaintenanceWindowRequest$WindowTaskId": "

The ID of the task to remove from the Maintenance Window.

", - "DeregisterTaskFromMaintenanceWindowResult$WindowTaskId": "

The ID of the task removed from the Maintenance Window.

", - "GetMaintenanceWindowTaskRequest$WindowTaskId": "

The Maintenance Window task ID to retrieve.

", - "GetMaintenanceWindowTaskResult$WindowTaskId": "

The retrieved Maintenance Window task ID.

", - "MaintenanceWindowTask$WindowTaskId": "

The task ID.

", - "RegisterTaskWithMaintenanceWindowResult$WindowTaskId": "

The id of the task in the Maintenance Window.

", - "UpdateMaintenanceWindowTaskRequest$WindowTaskId": "

The task ID to modify.

", - "UpdateMaintenanceWindowTaskResult$WindowTaskId": "

The task ID of the Maintenance Window that was updated.

" + }, + "MaintenanceWindowTaskId":{ + "base":null, + "refs":{ + "DeregisterTaskFromMaintenanceWindowRequest$WindowTaskId":"

The ID of the task to remove from the Maintenance Window.

", + "DeregisterTaskFromMaintenanceWindowResult$WindowTaskId":"

The ID of the task removed from the Maintenance Window.

", + "GetMaintenanceWindowTaskRequest$WindowTaskId":"

The Maintenance Window task ID to retrieve.

", + "GetMaintenanceWindowTaskResult$WindowTaskId":"

The retrieved Maintenance Window task ID.

", + "MaintenanceWindowTask$WindowTaskId":"

The task ID.

", + "RegisterTaskWithMaintenanceWindowResult$WindowTaskId":"

The id of the task in the Maintenance Window.

", + "UpdateMaintenanceWindowTaskRequest$WindowTaskId":"

The task ID to modify.

", + "UpdateMaintenanceWindowTaskResult$WindowTaskId":"

The task ID of the Maintenance Window that was updated.

" } }, - "MaintenanceWindowTaskInvocationParameters": { - "base": "

The parameters for task execution.

", - "refs": { - "GetMaintenanceWindowTaskResult$TaskInvocationParameters": "

The parameters to pass to the task when it executes.

", - "RegisterTaskWithMaintenanceWindowRequest$TaskInvocationParameters": "

The parameters that the task should use during execution. Populate only the fields that match the task type. All other fields should be empty.

", - "UpdateMaintenanceWindowTaskRequest$TaskInvocationParameters": "

The parameters that the task should use during execution. Populate only the fields that match the task type. All other fields should be empty.

", - "UpdateMaintenanceWindowTaskResult$TaskInvocationParameters": "

The updated parameter values.

" - } + "MaintenanceWindowTaskInvocationParameters":{ + "base":"

The parameters for task execution.

", + "refs":{ + "GetMaintenanceWindowTaskResult$TaskInvocationParameters":"

The parameters to pass to the task when it executes.

", + "RegisterTaskWithMaintenanceWindowRequest$TaskInvocationParameters":"

The parameters that the task should use during execution. Populate only the fields that match the task type. All other fields should be empty.

", + "UpdateMaintenanceWindowTaskRequest$TaskInvocationParameters":"

The parameters that the task should use during execution. Populate only the fields that match the task type. All other fields should be empty.

", + "UpdateMaintenanceWindowTaskResult$TaskInvocationParameters":"

The updated parameter values.

" + } }, - "MaintenanceWindowTaskList": { - "base": null, - "refs": { - "DescribeMaintenanceWindowTasksResult$Tasks": "

Information about the tasks in the Maintenance Window.

" + "MaintenanceWindowTaskList":{ + "base":null, + "refs":{ + "DescribeMaintenanceWindowTasksResult$Tasks":"

Information about the tasks in the Maintenance Window.

" } }, - "MaintenanceWindowTaskParameterName": { - "base": null, - "refs": { - "MaintenanceWindowTaskParameters$key": null - } - }, - "MaintenanceWindowTaskParameterValue": { - "base": null, - "refs": { - "MaintenanceWindowTaskParameterValueList$member": null - } - }, - "MaintenanceWindowTaskParameterValueExpression": { - "base": "

Defines the values for a task parameter.

", - "refs": { - "MaintenanceWindowTaskParameters$value": null - } - }, - "MaintenanceWindowTaskParameterValueList": { - "base": null, - "refs": { - "MaintenanceWindowTaskParameterValueExpression$Values": "

This field contains an array of 0 or more strings, each 1 to 255 characters in length.

" - } - }, - "MaintenanceWindowTaskParameters": { - "base": null, - "refs": { - "GetMaintenanceWindowTaskResult$TaskParameters": "

The parameters to pass to the task when it executes.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "MaintenanceWindowTask$TaskParameters": "

The parameters that should be passed to the task when it is executed.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "MaintenanceWindowTaskParametersList$member": null, - "RegisterTaskWithMaintenanceWindowRequest$TaskParameters": "

The parameters that should be passed to the task when it is executed.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", - "UpdateMaintenanceWindowTaskRequest$TaskParameters": "

The parameters to modify.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

The map has the following format:

Key: string, between 1 and 255 characters

Value: an array of strings, each string is between 1 and 255 characters

", - "UpdateMaintenanceWindowTaskResult$TaskParameters": "

The updated parameter values.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

" - } - }, - "MaintenanceWindowTaskParametersList": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskResult$TaskParameters": "

The parameters passed to the task when it was executed.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

The map has the following format:

Key: string, between 1 and 255 characters

Value: an array of strings, each string is between 1 and 255 characters

" - } - }, - "MaintenanceWindowTaskPriority": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskResult$Priority": "

The priority of the task.

", - "GetMaintenanceWindowTaskResult$Priority": "

The priority of the task when it executes. The lower the number, the higher the priority. Tasks that have the same priority are scheduled in parallel.

", - "MaintenanceWindowTask$Priority": "

The priority of the task in the Maintenance Window. The lower the number, the higher the priority. Tasks that have the same priority are scheduled in parallel.

", - "RegisterTaskWithMaintenanceWindowRequest$Priority": "

The priority of the task in the Maintenance Window, the lower the number the higher the priority. Tasks in a Maintenance Window are scheduled in priority order with tasks that have the same priority scheduled in parallel.

", - "UpdateMaintenanceWindowTaskRequest$Priority": "

The new task priority to specify. The lower the number, the higher the priority. Tasks that have the same priority are scheduled in parallel.

", - "UpdateMaintenanceWindowTaskResult$Priority": "

The updated priority value.

" - } - }, - "MaintenanceWindowTaskTargetId": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskInvocationResult$WindowTargetId": "

The Maintenance Window target ID.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$WindowTargetId": "

The ID of the target definition in this Maintenance Window the invocation was performed for.

" - } - }, - "MaintenanceWindowTaskType": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskInvocationResult$TaskType": "

Retrieves the task type for a Maintenance Window. Task types include the following: LAMBDA, STEP_FUNCTION, AUTOMATION, RUN_COMMAND.

", - "GetMaintenanceWindowExecutionTaskResult$Type": "

The type of task executed.

", - "GetMaintenanceWindowTaskResult$TaskType": "

The type of task to execute.

", - "MaintenanceWindowExecutionTaskIdentity$TaskType": "

The type of executed task.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$TaskType": "

The task type.

", - "MaintenanceWindowTask$Type": "

The type of task. The type can be one of the following: RUN_COMMAND, AUTOMATION, LAMBDA, or STEP_FUNCTION.

", - "RegisterTaskWithMaintenanceWindowRequest$TaskType": "

The type of task being registered.

" - } - }, - "ManagedInstanceId": { - "base": null, - "refs": { - "DeregisterManagedInstanceRequest$InstanceId": "

The ID assigned to the managed instance when you registered it using the activation process.

", - "UpdateManagedInstanceRoleRequest$InstanceId": "

The ID of the managed instance where you want to update the role.

" - } - }, - "MaxConcurrency": { - "base": null, - "refs": { - "AutomationExecution$MaxConcurrency": "

The MaxConcurrency value specified by the user when the execution started.

", - "AutomationExecutionMetadata$MaxConcurrency": "

The MaxConcurrency value specified by the user when starting the Automation.

", - "Command$MaxConcurrency": "

The maximum number of instances that are allowed to execute the command at the same time. You can specify a number of instances, such as 10, or a percentage of instances, such as 10%. The default value is 50. For more information about how to use MaxConcurrency, see Executing a Command Using Systems Manager Run Command.

", - "GetMaintenanceWindowExecutionTaskResult$MaxConcurrency": "

The defined maximum number of task executions that could be run in parallel.

", - "GetMaintenanceWindowTaskResult$MaxConcurrency": "

The maximum number of targets allowed to run this task in parallel.

", - "MaintenanceWindowTask$MaxConcurrency": "

The maximum number of targets this task can be run for in parallel.

", - "RegisterTaskWithMaintenanceWindowRequest$MaxConcurrency": "

The maximum number of targets this task can be run for in parallel.

", - "SendCommandRequest$MaxConcurrency": "

(Optional) The maximum number of instances that are allowed to execute the command at the same time. You can specify a number such as 10 or a percentage such as 10%. The default value is 50. For more information about how to use MaxConcurrency, see Using Concurrency Controls.

", - "StartAutomationExecutionRequest$MaxConcurrency": "

The maximum number of targets allowed to run this task in parallel. You can specify a number, such as 10, or a percentage, such as 10%. The default value is 10.

", - "UpdateMaintenanceWindowTaskRequest$MaxConcurrency": "

The new MaxConcurrency value you want to specify. MaxConcurrency is the number of targets that are allowed to run this task in parallel.

", - "UpdateMaintenanceWindowTaskResult$MaxConcurrency": "

The updated MaxConcurrency value.

" - } - }, - "MaxDocumentSizeExceeded": { - "base": "

The size limit of a document is 64 KB.

", - "refs": { - } - }, - "MaxErrors": { - "base": null, - "refs": { - "AutomationExecution$MaxErrors": "

The MaxErrors value specified by the user when the execution started.

", - "AutomationExecutionMetadata$MaxErrors": "

The MaxErrors value specified by the user when starting the Automation.

", - "Command$MaxErrors": "

The maximum number of errors allowed before the system stops sending the command to additional targets. You can specify a number of errors, such as 10, or a percentage or errors, such as 10%. The default value is 0. For more information about how to use MaxErrors, see Executing a Command Using Systems Manager Run Command.

", - "GetMaintenanceWindowExecutionTaskResult$MaxErrors": "

The defined maximum number of task execution errors allowed before scheduling of the task execution would have been stopped.

", - "GetMaintenanceWindowTaskResult$MaxErrors": "

The maximum number of errors allowed before the task stops being scheduled.

", - "MaintenanceWindowTask$MaxErrors": "

The maximum number of errors allowed before this task stops being scheduled.

", - "RegisterTaskWithMaintenanceWindowRequest$MaxErrors": "

The maximum number of errors allowed before this task stops being scheduled.

", - "SendCommandRequest$MaxErrors": "

The maximum number of errors allowed without the command failing. When the command fails one more time beyond the value of MaxErrors, the systems stops sending the command to additional targets. You can specify a number like 10 or a percentage like 10%. The default value is 0. For more information about how to use MaxErrors, see Using Error Controls.

", - "StartAutomationExecutionRequest$MaxErrors": "

The number of errors that are allowed before the system stops running the automation on additional targets. You can specify either an absolute number of errors, for example 10, or a percentage of the target set, for example 10%. If you specify 3, for example, the system stops running the automation when the fourth error is received. If you specify 0, then the system stops running the automation on additional targets after the first error result is returned. If you run an automation on 50 resources and set max-errors to 10%, then the system stops running the automation on additional targets when the sixth error is received.

Executions that are already running an automation when max-errors is reached are allowed to complete, but some of these executions may fail as well. If you need to ensure that there won't be more than max-errors failed executions, set max-concurrency to 1 so the executions proceed one at a time.

", - "UpdateMaintenanceWindowTaskRequest$MaxErrors": "

The new MaxErrors value to specify. MaxErrors is the maximum number of errors that are allowed before the task stops being scheduled.

", - "UpdateMaintenanceWindowTaskResult$MaxErrors": "

The updated MaxErrors value.

" - } - }, - "MaxResults": { - "base": null, - "refs": { - "DescribeActivationsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeAutomationExecutionsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeAutomationStepExecutionsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeInstanceAssociationsStatusRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeInventoryDeletionsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "DescribeParametersRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "GetInventoryRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "GetParameterHistoryRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListAssociationVersionsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListAssociationsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListComplianceItemsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListComplianceSummariesRequest$MaxResults": "

The maximum number of items to return for this call. Currently, you can specify null or 50. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListDocumentVersionsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListDocumentsRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListInventoryEntriesRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListResourceComplianceSummariesRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", - "ListResourceDataSyncRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" - } - }, - "MaxResultsEC2Compatible": { - "base": null, - "refs": { - "DescribeInstanceInformationRequest$MaxResults": "

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" - } - }, - "ModifyDocumentPermissionRequest": { - "base": null, - "refs": { - } - }, - "ModifyDocumentPermissionResponse": { - "base": null, - "refs": { - } - }, - "NextToken": { - "base": null, - "refs": { - "DescribeActivationsRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "DescribeActivationsResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "DescribeAutomationExecutionsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeAutomationExecutionsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeAutomationStepExecutionsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeAutomationStepExecutionsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeAvailablePatchesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeAvailablePatchesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeEffectiveInstanceAssociationsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeEffectiveInstanceAssociationsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeEffectivePatchesForPatchBaselineRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeEffectivePatchesForPatchBaselineResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeInstanceAssociationsStatusRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeInstanceAssociationsStatusResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeInstanceInformationRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeInstanceInformationResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeInstancePatchStatesForPatchGroupRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeInstancePatchStatesForPatchGroupResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeInstancePatchStatesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeInstancePatchStatesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeInstancePatchesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeInstancePatchesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeInventoryDeletionsRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "DescribeInventoryDeletionsResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeMaintenanceWindowExecutionTaskInvocationsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeMaintenanceWindowExecutionTasksRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeMaintenanceWindowExecutionTasksResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeMaintenanceWindowExecutionsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeMaintenanceWindowExecutionsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeMaintenanceWindowTargetsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeMaintenanceWindowTargetsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeMaintenanceWindowTasksRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeMaintenanceWindowTasksResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeMaintenanceWindowsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeMaintenanceWindowsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribeParametersRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribeParametersResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribePatchBaselinesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribePatchBaselinesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "DescribePatchGroupsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "DescribePatchGroupsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "GetInventoryRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "GetInventoryResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "GetInventorySchemaRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "GetInventorySchemaResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "GetParameterHistoryRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "GetParameterHistoryResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "GetParametersByPathRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "GetParametersByPathResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "ListAssociationVersionsRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "ListAssociationVersionsResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "ListAssociationsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "ListAssociationsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "ListCommandInvocationsRequest$NextToken": "

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", - "ListCommandInvocationsResult$NextToken": "

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", - "ListCommandsRequest$NextToken": "

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", - "ListCommandsResult$NextToken": "

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", - "ListComplianceItemsRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "ListComplianceItemsResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "ListComplianceSummariesRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "ListComplianceSummariesResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "ListDocumentVersionsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "ListDocumentVersionsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "ListDocumentsRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "ListDocumentsResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "ListInventoryEntriesRequest$NextToken": "

The token for the next set of items to return. (You received this token from a previous call.)

", - "ListInventoryEntriesResult$NextToken": "

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", - "ListResourceComplianceSummariesRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "ListResourceComplianceSummariesResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

", - "ListResourceDataSyncRequest$NextToken": "

A token to start the list. Use this token to get the next set of results.

", - "ListResourceDataSyncResult$NextToken": "

The token for the next set of items to return. Use this token to get the next set of results.

" - } - }, - "NonCompliantSummary": { - "base": "

A summary of resources that are not compliant. The summary is organized according to resource type.

", - "refs": { - "ComplianceSummaryItem$NonCompliantSummary": "

A list of NON_COMPLIANT items for the specified compliance type.

", - "ResourceComplianceSummaryItem$NonCompliantSummary": "

A list of items that aren't compliant for the resource.

" - } - }, - "NormalStringMap": { - "base": null, - "refs": { - "StepExecution$Inputs": "

Fully-resolved values passed into the step before execution.

" - } - }, - "NotificationArn": { - "base": null, - "refs": { - "NotificationConfig$NotificationArn": "

An Amazon Resource Name (ARN) for a Simple Notification Service (SNS) topic. Run Command pushes notifications about command status changes to this topic.

" - } - }, - "NotificationConfig": { - "base": "

Configurations for sending notifications.

", - "refs": { - "Command$NotificationConfig": "

Configurations for sending notifications about command status changes.

", - "CommandInvocation$NotificationConfig": "

Configurations for sending notifications about command status changes on a per instance basis.

", - "MaintenanceWindowRunCommandParameters$NotificationConfig": "

Configurations for sending notifications about command status changes on a per-instance basis.

", - "SendCommandRequest$NotificationConfig": "

Configurations for sending notifications.

" - } - }, - "NotificationEvent": { - "base": null, - "refs": { - "NotificationEventList$member": null - } - }, - "NotificationEventList": { - "base": null, - "refs": { - "NotificationConfig$NotificationEvents": "

The different events for which you can receive notifications. These events include the following: All (events), InProgress, Success, TimedOut, Cancelled, Failed. To learn more about these events, see Setting Up Events and Notifications in the AWS Systems Manager User Guide.

" - } - }, - "NotificationType": { - "base": null, - "refs": { - "NotificationConfig$NotificationType": "

Command: Receive notification when the status of a command changes. Invocation: For commands sent to multiple instances, receive notification on a per-instance basis when the status of a command changes.

" - } - }, - "OperatingSystem": { - "base": null, - "refs": { - "CreatePatchBaselineRequest$OperatingSystem": "

Defines the operating system the patch baseline applies to. The Default value is WINDOWS.

", - "GetDefaultPatchBaselineRequest$OperatingSystem": "

Returns the default patch baseline for the specified operating system.

", - "GetDefaultPatchBaselineResult$OperatingSystem": "

The operating system for the returned patch baseline.

", - "GetPatchBaselineForPatchGroupRequest$OperatingSystem": "

Returns he operating system rule specified for patch groups using the patch baseline.

", - "GetPatchBaselineForPatchGroupResult$OperatingSystem": "

The operating system rule specified for patch groups using the patch baseline.

", - "GetPatchBaselineResult$OperatingSystem": "

Returns the operating system specified for the patch baseline.

", - "PatchBaselineIdentity$OperatingSystem": "

Defines the operating system the patch baseline applies to. The Default value is WINDOWS.

", - "UpdatePatchBaselineResult$OperatingSystem": "

The operating system rule used by the updated patch baseline.

" - } - }, - "OwnerInformation": { - "base": null, - "refs": { - "GetMaintenanceWindowExecutionTaskInvocationResult$OwnerInformation": "

User-provided value to be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", - "InstancePatchState$OwnerInformation": "

Placeholder information. This field will always be empty in the current release of the service.

", - "MaintenanceWindowExecutionTaskInvocationIdentity$OwnerInformation": "

User-provided value that was specified when the target was registered with the Maintenance Window. This was also included in any CloudWatch events raised during the task invocation.

", - "MaintenanceWindowTarget$OwnerInformation": "

User-provided value that will be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", - "RegisterTargetWithMaintenanceWindowRequest$OwnerInformation": "

User-provided value that will be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", - "UpdateMaintenanceWindowTargetRequest$OwnerInformation": "

User-provided value that will be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", - "UpdateMaintenanceWindowTargetResult$OwnerInformation": "

The updated owner.

" - } - }, - "PSParameterName": { - "base": null, - "refs": { - "DeleteParameterRequest$Name": "

The name of the parameter to delete.

", - "GetParameterHistoryRequest$Name": "

The name of a parameter you want to query.

", - "GetParameterRequest$Name": "

The name of the parameter you want to query.

", - "GetParametersByPathRequest$Path": "

The hierarchy for the parameter. Hierarchies start with a forward slash (/) and end with the parameter name. A hierarchy can have a maximum of 15 levels. Here is an example of a hierarchy: /Finance/Prod/IAD/WinServ2016/license33

", - "Parameter$Name": "

The name of the parameter.

", - "ParameterHistory$Name": "

The name of the parameter.

", - "ParameterMetadata$Name": "

The parameter name.

", - "ParameterNameList$member": null, - "PutParameterRequest$Name": "

The fully qualified name of the parameter that you want to add to the system. The fully qualified name includes the complete hierarchy of the parameter path and name. For example: /Dev/DBServer/MySQL/db-string13

For information about parameter name requirements and restrictions, see About Creating Systems Manager Parameters in the AWS Systems Manager User Guide.

The maximum length constraint listed below includes capacity for additional system attributes that are not part of the name. The maximum length for the fully qualified parameter name is 1011 characters.

" + "MaintenanceWindowTaskParameterName":{ + "base":null, + "refs":{ + "MaintenanceWindowTaskParameters$key":null + } + }, + "MaintenanceWindowTaskParameterValue":{ + "base":null, + "refs":{ + "MaintenanceWindowTaskParameterValueList$member":null + } + }, + "MaintenanceWindowTaskParameterValueExpression":{ + "base":"

Defines the values for a task parameter.

", + "refs":{ + "MaintenanceWindowTaskParameters$value":null + } + }, + "MaintenanceWindowTaskParameterValueList":{ + "base":null, + "refs":{ + "MaintenanceWindowTaskParameterValueExpression$Values":"

This field contains an array of 0 or more strings, each 1 to 255 characters in length.

" + } + }, + "MaintenanceWindowTaskParameters":{ + "base":null, + "refs":{ + "GetMaintenanceWindowTaskResult$TaskParameters":"

The parameters to pass to the task when it executes.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "MaintenanceWindowTask$TaskParameters":"

The parameters that should be passed to the task when it is executed.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "MaintenanceWindowTaskParametersList$member":null, + "RegisterTaskWithMaintenanceWindowRequest$TaskParameters":"

The parameters that should be passed to the task when it is executed.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

", + "UpdateMaintenanceWindowTaskRequest$TaskParameters":"

The parameters to modify.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

The map has the following format:

Key: string, between 1 and 255 characters

Value: an array of strings, each string is between 1 and 255 characters

", + "UpdateMaintenanceWindowTaskResult$TaskParameters":"

The updated parameter values.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

" + } + }, + "MaintenanceWindowTaskParametersList":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskResult$TaskParameters":"

The parameters passed to the task when it was executed.

TaskParameters has been deprecated. To specify parameters to pass to a task when it runs, instead use the Parameters option in the TaskInvocationParameters structure. For information about how Systems Manager handles these options for the supported Maintenance Window task types, see MaintenanceWindowTaskInvocationParameters.

The map has the following format:

Key: string, between 1 and 255 characters

Value: an array of strings, each string is between 1 and 255 characters

" + } + }, + "MaintenanceWindowTaskPriority":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskResult$Priority":"

The priority of the task.

", + "GetMaintenanceWindowTaskResult$Priority":"

The priority of the task when it executes. The lower the number, the higher the priority. Tasks that have the same priority are scheduled in parallel.

", + "MaintenanceWindowTask$Priority":"

The priority of the task in the Maintenance Window. The lower the number, the higher the priority. Tasks that have the same priority are scheduled in parallel.

", + "RegisterTaskWithMaintenanceWindowRequest$Priority":"

The priority of the task in the Maintenance Window, the lower the number the higher the priority. Tasks in a Maintenance Window are scheduled in priority order with tasks that have the same priority scheduled in parallel.

", + "UpdateMaintenanceWindowTaskRequest$Priority":"

The new task priority to specify. The lower the number, the higher the priority. Tasks that have the same priority are scheduled in parallel.

", + "UpdateMaintenanceWindowTaskResult$Priority":"

The updated priority value.

" + } + }, + "MaintenanceWindowTaskTargetId":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskInvocationResult$WindowTargetId":"

The Maintenance Window target ID.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$WindowTargetId":"

The ID of the target definition in this Maintenance Window the invocation was performed for.

" + } + }, + "MaintenanceWindowTaskType":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskInvocationResult$TaskType":"

Retrieves the task type for a Maintenance Window. Task types include the following: LAMBDA, STEP_FUNCTION, AUTOMATION, RUN_COMMAND.

", + "GetMaintenanceWindowExecutionTaskResult$Type":"

The type of task executed.

", + "GetMaintenanceWindowTaskResult$TaskType":"

The type of task to execute.

", + "MaintenanceWindowExecutionTaskIdentity$TaskType":"

The type of executed task.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$TaskType":"

The task type.

", + "MaintenanceWindowTask$Type":"

The type of task. The type can be one of the following: RUN_COMMAND, AUTOMATION, LAMBDA, or STEP_FUNCTION.

", + "RegisterTaskWithMaintenanceWindowRequest$TaskType":"

The type of task being registered.

" + } + }, + "ManagedInstanceId":{ + "base":null, + "refs":{ + "DeregisterManagedInstanceRequest$InstanceId":"

The ID assigned to the managed instance when you registered it using the activation process.

", + "UpdateManagedInstanceRoleRequest$InstanceId":"

The ID of the managed instance where you want to update the role.

" + } + }, + "MaxConcurrency":{ + "base":null, + "refs":{ + "AutomationExecution$MaxConcurrency":"

The MaxConcurrency value specified by the user when the execution started.

", + "AutomationExecutionMetadata$MaxConcurrency":"

The MaxConcurrency value specified by the user when starting the Automation.

", + "Command$MaxConcurrency":"

The maximum number of instances that are allowed to execute the command at the same time. You can specify a number of instances, such as 10, or a percentage of instances, such as 10%. The default value is 50. For more information about how to use MaxConcurrency, see Executing Commands Using Systems Manager Run Command in the AWS Systems Manager User Guide.

", + "GetMaintenanceWindowExecutionTaskResult$MaxConcurrency":"

The defined maximum number of task executions that could be run in parallel.

", + "GetMaintenanceWindowTaskResult$MaxConcurrency":"

The maximum number of targets allowed to run this task in parallel.

", + "MaintenanceWindowTask$MaxConcurrency":"

The maximum number of targets this task can be run for in parallel.

", + "RegisterTaskWithMaintenanceWindowRequest$MaxConcurrency":"

The maximum number of targets this task can be run for in parallel.

", + "SendCommandRequest$MaxConcurrency":"

(Optional) The maximum number of instances that are allowed to execute the command at the same time. You can specify a number such as 10 or a percentage such as 10%. The default value is 50. For more information about how to use MaxConcurrency, see Using Concurrency Controls in the AWS Systems Manager User Guide.

", + "StartAutomationExecutionRequest$MaxConcurrency":"

The maximum number of targets allowed to run this task in parallel. You can specify a number, such as 10, or a percentage, such as 10%. The default value is 10.

", + "UpdateMaintenanceWindowTaskRequest$MaxConcurrency":"

The new MaxConcurrency value you want to specify. MaxConcurrency is the number of targets that are allowed to run this task in parallel.

", + "UpdateMaintenanceWindowTaskResult$MaxConcurrency":"

The updated MaxConcurrency value.

" + } + }, + "MaxDocumentSizeExceeded":{ + "base":"

The size limit of a document is 64 KB.

", + "refs":{ + } + }, + "MaxErrors":{ + "base":null, + "refs":{ + "AutomationExecution$MaxErrors":"

The MaxErrors value specified by the user when the execution started.

", + "AutomationExecutionMetadata$MaxErrors":"

The MaxErrors value specified by the user when starting the Automation.

", + "Command$MaxErrors":"

The maximum number of errors allowed before the system stops sending the command to additional targets. You can specify a number of errors, such as 10, or a percentage or errors, such as 10%. The default value is 0. For more information about how to use MaxErrors, see Executing Commands Using Systems Manager Run Command in the AWS Systems Manager User Guide.

", + "GetMaintenanceWindowExecutionTaskResult$MaxErrors":"

The defined maximum number of task execution errors allowed before scheduling of the task execution would have been stopped.

", + "GetMaintenanceWindowTaskResult$MaxErrors":"

The maximum number of errors allowed before the task stops being scheduled.

", + "MaintenanceWindowTask$MaxErrors":"

The maximum number of errors allowed before this task stops being scheduled.

", + "RegisterTaskWithMaintenanceWindowRequest$MaxErrors":"

The maximum number of errors allowed before this task stops being scheduled.

", + "SendCommandRequest$MaxErrors":"

The maximum number of errors allowed without the command failing. When the command fails one more time beyond the value of MaxErrors, the systems stops sending the command to additional targets. You can specify a number like 10 or a percentage like 10%. The default value is 0. For more information about how to use MaxErrors, see Using Error Controls in the AWS Systems Manager User Guide.

", + "StartAutomationExecutionRequest$MaxErrors":"

The number of errors that are allowed before the system stops running the automation on additional targets. You can specify either an absolute number of errors, for example 10, or a percentage of the target set, for example 10%. If you specify 3, for example, the system stops running the automation when the fourth error is received. If you specify 0, then the system stops running the automation on additional targets after the first error result is returned. If you run an automation on 50 resources and set max-errors to 10%, then the system stops running the automation on additional targets when the sixth error is received.

Executions that are already running an automation when max-errors is reached are allowed to complete, but some of these executions may fail as well. If you need to ensure that there won't be more than max-errors failed executions, set max-concurrency to 1 so the executions proceed one at a time.

", + "UpdateMaintenanceWindowTaskRequest$MaxErrors":"

The new MaxErrors value to specify. MaxErrors is the maximum number of errors that are allowed before the task stops being scheduled.

", + "UpdateMaintenanceWindowTaskResult$MaxErrors":"

The updated MaxErrors value.

" + } + }, + "MaxResults":{ + "base":null, + "refs":{ + "DescribeActivationsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeAutomationExecutionsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeAutomationStepExecutionsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeInstanceAssociationsStatusRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeInventoryDeletionsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "DescribeParametersRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "GetInventoryRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "GetParameterHistoryRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListAssociationVersionsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListAssociationsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListComplianceItemsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListComplianceSummariesRequest$MaxResults":"

The maximum number of items to return for this call. Currently, you can specify null or 50. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListDocumentVersionsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListDocumentsRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListInventoryEntriesRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListResourceComplianceSummariesRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

", + "ListResourceDataSyncRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + } + }, + "MaxResultsEC2Compatible":{ + "base":null, + "refs":{ + "DescribeInstanceInformationRequest$MaxResults":"

The maximum number of items to return for this call. The call also returns a token that you can specify in a subsequent call to get the next set of results.

" + } + }, + "ModifyDocumentPermissionRequest":{ + "base":null, + "refs":{ + } + }, + "ModifyDocumentPermissionResponse":{ + "base":null, + "refs":{ + } + }, + "NextToken":{ + "base":null, + "refs":{ + "DescribeActivationsRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "DescribeActivationsResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "DescribeAutomationExecutionsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeAutomationExecutionsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeAutomationStepExecutionsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeAutomationStepExecutionsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeAvailablePatchesRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeAvailablePatchesResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeEffectiveInstanceAssociationsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeEffectiveInstanceAssociationsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeEffectivePatchesForPatchBaselineRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeEffectivePatchesForPatchBaselineResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeInstanceAssociationsStatusRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeInstanceAssociationsStatusResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeInstanceInformationRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeInstanceInformationResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeInstancePatchStatesForPatchGroupRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeInstancePatchStatesForPatchGroupResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeInstancePatchStatesRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeInstancePatchStatesResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeInstancePatchesRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeInstancePatchesResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeInventoryDeletionsRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "DescribeInventoryDeletionsResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "DescribeMaintenanceWindowExecutionTaskInvocationsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeMaintenanceWindowExecutionTaskInvocationsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeMaintenanceWindowExecutionTasksRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeMaintenanceWindowExecutionTasksResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeMaintenanceWindowExecutionsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeMaintenanceWindowExecutionsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeMaintenanceWindowTargetsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeMaintenanceWindowTargetsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeMaintenanceWindowTasksRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeMaintenanceWindowTasksResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeMaintenanceWindowsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeMaintenanceWindowsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribeParametersRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribeParametersResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribePatchBaselinesRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribePatchBaselinesResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "DescribePatchGroupsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "DescribePatchGroupsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "GetInventoryRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "GetInventoryResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "GetInventorySchemaRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "GetInventorySchemaResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "GetParameterHistoryRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "GetParameterHistoryResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "GetParametersByPathRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "GetParametersByPathResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "ListAssociationVersionsRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "ListAssociationVersionsResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "ListAssociationsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "ListAssociationsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "ListCommandInvocationsRequest$NextToken":"

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", + "ListCommandInvocationsResult$NextToken":"

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", + "ListCommandsRequest$NextToken":"

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", + "ListCommandsResult$NextToken":"

(Optional) The token for the next set of items to return. (You received this token from a previous call.)

", + "ListComplianceItemsRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "ListComplianceItemsResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "ListComplianceSummariesRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "ListComplianceSummariesResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "ListDocumentVersionsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "ListDocumentVersionsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "ListDocumentsRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "ListDocumentsResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "ListInventoryEntriesRequest$NextToken":"

The token for the next set of items to return. (You received this token from a previous call.)

", + "ListInventoryEntriesResult$NextToken":"

The token to use when requesting the next set of items. If there are no additional items to return, the string is empty.

", + "ListResourceComplianceSummariesRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "ListResourceComplianceSummariesResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

", + "ListResourceDataSyncRequest$NextToken":"

A token to start the list. Use this token to get the next set of results.

", + "ListResourceDataSyncResult$NextToken":"

The token for the next set of items to return. Use this token to get the next set of results.

" + } + }, + "NonCompliantSummary":{ + "base":"

A summary of resources that are not compliant. The summary is organized according to resource type.

", + "refs":{ + "ComplianceSummaryItem$NonCompliantSummary":"

A list of NON_COMPLIANT items for the specified compliance type.

", + "ResourceComplianceSummaryItem$NonCompliantSummary":"

A list of items that aren't compliant for the resource.

" + } + }, + "NormalStringMap":{ + "base":null, + "refs":{ + "StepExecution$Inputs":"

Fully-resolved values passed into the step before execution.

" + } + }, + "NotificationArn":{ + "base":null, + "refs":{ + "NotificationConfig$NotificationArn":"

An Amazon Resource Name (ARN) for a Simple Notification Service (SNS) topic. Run Command pushes notifications about command status changes to this topic.

" + } + }, + "NotificationConfig":{ + "base":"

Configurations for sending notifications.

", + "refs":{ + "Command$NotificationConfig":"

Configurations for sending notifications about command status changes.

", + "CommandInvocation$NotificationConfig":"

Configurations for sending notifications about command status changes on a per instance basis.

", + "MaintenanceWindowRunCommandParameters$NotificationConfig":"

Configurations for sending notifications about command status changes on a per-instance basis.

", + "SendCommandRequest$NotificationConfig":"

Configurations for sending notifications.

" + } + }, + "NotificationEvent":{ + "base":null, + "refs":{ + "NotificationEventList$member":null + } + }, + "NotificationEventList":{ + "base":null, + "refs":{ + "NotificationConfig$NotificationEvents":"

The different events for which you can receive notifications. These events include the following: All (events), InProgress, Success, TimedOut, Cancelled, Failed. To learn more about these events, see Configuring Amazon SNS Notifications for Run Command in the AWS Systems Manager User Guide.

" + } + }, + "NotificationType":{ + "base":null, + "refs":{ + "NotificationConfig$NotificationType":"

Command: Receive notification when the status of a command changes. Invocation: For commands sent to multiple instances, receive notification on a per-instance basis when the status of a command changes.

" + } + }, + "OperatingSystem":{ + "base":null, + "refs":{ + "CreatePatchBaselineRequest$OperatingSystem":"

Defines the operating system the patch baseline applies to. The Default value is WINDOWS.

", + "GetDefaultPatchBaselineRequest$OperatingSystem":"

Returns the default patch baseline for the specified operating system.

", + "GetDefaultPatchBaselineResult$OperatingSystem":"

The operating system for the returned patch baseline.

", + "GetPatchBaselineForPatchGroupRequest$OperatingSystem":"

Returns he operating system rule specified for patch groups using the patch baseline.

", + "GetPatchBaselineForPatchGroupResult$OperatingSystem":"

The operating system rule specified for patch groups using the patch baseline.

", + "GetPatchBaselineResult$OperatingSystem":"

Returns the operating system specified for the patch baseline.

", + "PatchBaselineIdentity$OperatingSystem":"

Defines the operating system the patch baseline applies to. The Default value is WINDOWS.

", + "UpdatePatchBaselineResult$OperatingSystem":"

The operating system rule used by the updated patch baseline.

" + } + }, + "OwnerInformation":{ + "base":null, + "refs":{ + "GetMaintenanceWindowExecutionTaskInvocationResult$OwnerInformation":"

User-provided value to be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", + "InstancePatchState$OwnerInformation":"

Placeholder information. This field will always be empty in the current release of the service.

", + "MaintenanceWindowExecutionTaskInvocationIdentity$OwnerInformation":"

User-provided value that was specified when the target was registered with the Maintenance Window. This was also included in any CloudWatch events raised during the task invocation.

", + "MaintenanceWindowTarget$OwnerInformation":"

User-provided value that will be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", + "RegisterTargetWithMaintenanceWindowRequest$OwnerInformation":"

User-provided value that will be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", + "UpdateMaintenanceWindowTargetRequest$OwnerInformation":"

User-provided value that will be included in any CloudWatch events raised while running tasks for these targets in this Maintenance Window.

", + "UpdateMaintenanceWindowTargetResult$OwnerInformation":"

The updated owner.

" + } + }, + "PSParameterName":{ + "base":null, + "refs":{ + "DeleteParameterRequest$Name":"

The name of the parameter to delete.

", + "GetParameterHistoryRequest$Name":"

The name of a parameter you want to query.

", + "GetParameterRequest$Name":"

The name of the parameter you want to query.

", + "GetParametersByPathRequest$Path":"

The hierarchy for the parameter. Hierarchies start with a forward slash (/) and end with the parameter name. A hierarchy can have a maximum of 15 levels. Here is an example of a hierarchy: /Finance/Prod/IAD/WinServ2016/license33

", + "Parameter$Name":"

The name of the parameter.

", + "ParameterHistory$Name":"

The name of the parameter.

", + "ParameterMetadata$Name":"

The parameter name.

", + "ParameterNameList$member":null, + "PutParameterRequest$Name":"

The fully qualified name of the parameter that you want to add to the system. The fully qualified name includes the complete hierarchy of the parameter path and name. For example: /Dev/DBServer/MySQL/db-string13

For information about parameter name requirements and restrictions, see Creating Systems Manager Parameters in the AWS Systems Manager User Guide.

The maximum length constraint listed below includes capacity for additional system attributes that are not part of the name. The maximum length for the fully qualified parameter name is 1011 characters.

" } }, - "PSParameterValue": { - "base": null, - "refs": { - "Parameter$Value": "

The parameter value.

", - "ParameterHistory$Value": "

The parameter value.

", - "PutParameterRequest$Value": "

The parameter value that you want to add to the system.

" + "PSParameterValue":{ + "base":null, + "refs":{ + "Parameter$Value":"

The parameter value.

", + "ParameterHistory$Value":"

The parameter value.

", + "PutParameterRequest$Value":"

The parameter value that you want to add to the system.

" } }, - "PSParameterVersion": { - "base": null, - "refs": { - "Parameter$Version": "

The parameter version.

", - "ParameterHistory$Version": "

The parameter version.

", - "ParameterMetadata$Version": "

The parameter version.

", - "PutParameterResult$Version": "

The new version number of a parameter. If you edit a parameter value, Parameter Store automatically creates a new version and assigns this new version a unique ID. You can reference a parameter version ID in API actions or in Systems Manager documents (SSM documents). By default, if you don't specify a specific version, the system returns the latest parameter value when a parameter is called.

" + "PSParameterVersion":{ + "base":null, + "refs":{ + "Parameter$Version":"

The parameter version.

", + "ParameterHistory$Version":"

The parameter version.

", + "ParameterMetadata$Version":"

The parameter version.

", + "PutParameterResult$Version":"

The new version number of a parameter. If you edit a parameter value, Parameter Store automatically creates a new version and assigns this new version a unique ID. You can reference a parameter version ID in API actions or in Systems Manager documents (SSM documents). By default, if you don't specify a specific version, the system returns the latest parameter value when a parameter is called.

" } }, - "Parameter": { - "base": "

An Amazon EC2 Systems Manager parameter in Parameter Store.

", - "refs": { - "GetParameterResult$Parameter": "

Information about a parameter.

", - "ParameterList$member": null + "Parameter":{ + "base":"

An Amazon EC2 Systems Manager parameter in Parameter Store.

", + "refs":{ + "GetParameterResult$Parameter":"

Information about a parameter.

", + "ParameterList$member":null } }, - "ParameterAlreadyExists": { - "base": "

The parameter already exists. You can't create duplicate parameters.

", - "refs": { + "ParameterAlreadyExists":{ + "base":"

The parameter already exists. You can't create duplicate parameters.

", + "refs":{ } }, - "ParameterDescription": { - "base": null, - "refs": { - "ParameterHistory$Description": "

Information about the parameter.

", - "ParameterMetadata$Description": "

Description of the parameter actions.

", - "PutParameterRequest$Description": "

Information about the parameter that you want to add to the system.

Do not enter personally identifiable information in this field.

" + "ParameterDescription":{ + "base":null, + "refs":{ + "ParameterHistory$Description":"

Information about the parameter.

", + "ParameterMetadata$Description":"

Description of the parameter actions.

", + "PutParameterRequest$Description":"

Information about the parameter that you want to add to the system.

Do not enter personally identifiable information in this field.

" } }, - "ParameterHistory": { - "base": "

Information about parameter usage.

", - "refs": { - "ParameterHistoryList$member": null + "ParameterHistory":{ + "base":"

Information about parameter usage.

", + "refs":{ + "ParameterHistoryList$member":null } }, - "ParameterHistoryList": { - "base": null, - "refs": { - "GetParameterHistoryResult$Parameters": "

A list of parameters returned by the request.

" + "ParameterHistoryList":{ + "base":null, + "refs":{ + "GetParameterHistoryResult$Parameters":"

A list of parameters returned by the request.

" } }, - "ParameterKeyId": { - "base": null, - "refs": { - "ParameterHistory$KeyId": "

The ID of the query key used for this parameter.

", - "ParameterMetadata$KeyId": "

The ID of the query key used for this parameter.

", - "PutParameterRequest$KeyId": "

The KMS Key ID that you want to use to encrypt a parameter when you choose the SecureString data type. If you don't specify a key ID, the system uses the default key associated with your AWS account.

" + "ParameterKeyId":{ + "base":null, + "refs":{ + "ParameterHistory$KeyId":"

The ID of the query key used for this parameter.

", + "ParameterMetadata$KeyId":"

The ID of the query key used for this parameter.

", + "PutParameterRequest$KeyId":"

The KMS Key ID that you want to use to encrypt a parameter when you choose the SecureString data type. If you don't specify a key ID, the system uses the default key associated with your AWS account.

" } }, - "ParameterLimitExceeded": { - "base": "

You have exceeded the number of parameters for this AWS account. Delete one or more parameters and try again.

", - "refs": { + "ParameterLimitExceeded":{ + "base":"

You have exceeded the number of parameters for this AWS account. Delete one or more parameters and try again.

", + "refs":{ } }, - "ParameterList": { - "base": null, - "refs": { - "GetParametersByPathResult$Parameters": "

A list of parameters found in the specified hierarchy.

", - "GetParametersResult$Parameters": "

A list of details for a parameter.

" + "ParameterList":{ + "base":null, + "refs":{ + "GetParametersByPathResult$Parameters":"

A list of parameters found in the specified hierarchy.

", + "GetParametersResult$Parameters":"

A list of details for a parameter.

" } }, - "ParameterMaxVersionLimitExceeded": { - "base": "

The parameter exceeded the maximum number of allowed versions.

", - "refs": { + "ParameterMaxVersionLimitExceeded":{ + "base":"

The parameter exceeded the maximum number of allowed versions.

", + "refs":{ } }, - "ParameterMetadata": { - "base": "

Metada includes information like the ARN of the last user and the date/time the parameter was last used.

", - "refs": { - "ParameterMetadataList$member": null + "ParameterMetadata":{ + "base":"

Metada includes information like the ARN of the last user and the date/time the parameter was last used.

", + "refs":{ + "ParameterMetadataList$member":null } }, - "ParameterMetadataList": { - "base": null, - "refs": { - "DescribeParametersResult$Parameters": "

Parameters returned by the request.

" + "ParameterMetadataList":{ + "base":null, + "refs":{ + "DescribeParametersResult$Parameters":"

Parameters returned by the request.

" } }, - "ParameterName": { - "base": null, - "refs": { - "Parameters$key": null + "ParameterName":{ + "base":null, + "refs":{ + "Parameters$key":null } }, - "ParameterNameList": { - "base": null, - "refs": { - "DeleteParametersRequest$Names": "

The names of the parameters to delete.

", - "DeleteParametersResult$DeletedParameters": "

The names of the deleted parameters.

", - "DeleteParametersResult$InvalidParameters": "

The names of parameters that weren't deleted because the parameters are not valid.

", - "GetParametersRequest$Names": "

Names of the parameters for which you want to query information.

", - "GetParametersResult$InvalidParameters": "

A list of parameters that are not formatted correctly or do not run when executed.

" + "ParameterNameList":{ + "base":null, + "refs":{ + "DeleteParametersRequest$Names":"

The names of the parameters to delete.

", + "DeleteParametersResult$DeletedParameters":"

The names of the deleted parameters.

", + "DeleteParametersResult$InvalidParameters":"

The names of parameters that weren't deleted because the parameters are not valid.

", + "GetParametersRequest$Names":"

Names of the parameters for which you want to query information.

", + "GetParametersResult$InvalidParameters":"

A list of parameters that are not formatted correctly or do not run when executed.

" } }, - "ParameterNotFound": { - "base": "

The parameter could not be found. Verify the name and try again.

", - "refs": { + "ParameterNotFound":{ + "base":"

The parameter could not be found. Verify the name and try again.

", + "refs":{ } }, - "ParameterPatternMismatchException": { - "base": "

The parameter name is not valid.

", - "refs": { + "ParameterPatternMismatchException":{ + "base":"

The parameter name is not valid.

", + "refs":{ } }, - "ParameterStringFilter": { - "base": "

One or more filters. Use a filter to return a more specific list of results.

", - "refs": { - "ParameterStringFilterList$member": null + "ParameterStringFilter":{ + "base":"

One or more filters. Use a filter to return a more specific list of results.

The Name field can't be used with the GetParametersByPath API action.

", + "refs":{ + "ParameterStringFilterList$member":null } }, - "ParameterStringFilterKey": { - "base": null, - "refs": { - "ParameterStringFilter$Key": "

The name of the filter.

" + "ParameterStringFilterKey":{ + "base":null, + "refs":{ + "ParameterStringFilter$Key":"

The name of the filter.

" } }, - "ParameterStringFilterList": { - "base": null, - "refs": { - "DescribeParametersRequest$ParameterFilters": "

Filters to limit the request results.

", - "GetParametersByPathRequest$ParameterFilters": "

Filters to limit the request results.

You can't filter using the parameter name.

" + "ParameterStringFilterList":{ + "base":null, + "refs":{ + "DescribeParametersRequest$ParameterFilters":"

Filters to limit the request results.

", + "GetParametersByPathRequest$ParameterFilters":"

Filters to limit the request results.

You can't filter using the parameter name.

" } }, - "ParameterStringFilterValue": { - "base": null, - "refs": { - "ParameterStringFilterValueList$member": null + "ParameterStringFilterValue":{ + "base":null, + "refs":{ + "ParameterStringFilterValueList$member":null } }, - "ParameterStringFilterValueList": { - "base": null, - "refs": { - "ParameterStringFilter$Values": "

The value you want to search for.

" + "ParameterStringFilterValueList":{ + "base":null, + "refs":{ + "ParameterStringFilter$Values":"

The value you want to search for.

" } }, - "ParameterStringQueryOption": { - "base": null, - "refs": { - "ParameterStringFilter$Option": "

Valid options are Equals and BeginsWith. For Path filter, valid options are Recursive and OneLevel.

" + "ParameterStringQueryOption":{ + "base":null, + "refs":{ + "ParameterStringFilter$Option":"

Valid options are Equals and BeginsWith. For Path filter, valid options are Recursive and OneLevel.

" } }, - "ParameterType": { - "base": null, - "refs": { - "Parameter$Type": "

The type of parameter. Valid values include the following: String, String list, Secure string.

", - "ParameterHistory$Type": "

The type of parameter used.

", - "ParameterMetadata$Type": "

The type of parameter. Valid parameter types include the following: String, String list, Secure string.

", - "PutParameterRequest$Type": "

The type of parameter that you want to add to the system.

" + "ParameterType":{ + "base":null, + "refs":{ + "Parameter$Type":"

The type of parameter. Valid values include the following: String, String list, Secure string.

", + "ParameterHistory$Type":"

The type of parameter used.

", + "ParameterMetadata$Type":"

The type of parameter. Valid parameter types include the following: String, String list, Secure string.

", + "PutParameterRequest$Type":"

The type of parameter that you want to add to the system.

" } }, - "ParameterValue": { - "base": null, - "refs": { - "ParameterValueList$member": null, - "TargetParameterList$member": null + "ParameterValue":{ + "base":null, + "refs":{ + "ParameterValueList$member":null, + "TargetParameterList$member":null } }, - "ParameterValueList": { - "base": null, - "refs": { - "Parameters$value": null + "ParameterValueList":{ + "base":null, + "refs":{ + "Parameters$value":null } }, - "ParameterVersionNotFound": { - "base": "

The specified parameter version was not found. Verify the parameter name and version, and try again.

", - "refs": { + "ParameterVersionNotFound":{ + "base":"

The specified parameter version was not found. Verify the parameter name and version, and try again.

", + "refs":{ } }, - "Parameters": { - "base": null, - "refs": { - "AssociationDescription$Parameters": "

A description of the parameters for a document.

", - "AssociationVersionInfo$Parameters": "

Parameters specified when the association version was created.

", - "Command$Parameters": "

The parameter values to be inserted in the document when executing the command.

", - "CreateAssociationBatchRequestEntry$Parameters": "

A description of the parameters for a document.

", - "CreateAssociationRequest$Parameters": "

The parameters for the documents runtime configuration.

", - "MaintenanceWindowRunCommandParameters$Parameters": "

The parameters for the RUN_COMMAND task execution.

", - "SendCommandRequest$Parameters": "

The required and optional parameters specified in the document being executed.

", - "UpdateAssociationRequest$Parameters": "

The parameters you want to update for the association. If you create a parameter using Parameter Store, you can reference the parameter using {{ssm:parameter-name}}

" + "Parameters":{ + "base":null, + "refs":{ + "AssociationDescription$Parameters":"

A description of the parameters for a document.

", + "AssociationVersionInfo$Parameters":"

Parameters specified when the association version was created.

", + "Command$Parameters":"

The parameter values to be inserted in the document when executing the command.

", + "CreateAssociationBatchRequestEntry$Parameters":"

A description of the parameters for a document.

", + "CreateAssociationRequest$Parameters":"

The parameters for the documents runtime configuration.

", + "MaintenanceWindowRunCommandParameters$Parameters":"

The parameters for the RUN_COMMAND task execution.

", + "SendCommandRequest$Parameters":"

The required and optional parameters specified in the document being executed.

", + "UpdateAssociationRequest$Parameters":"

The parameters you want to update for the association. If you create a parameter using Parameter Store, you can reference the parameter using {{ssm:parameter-name}}

" } }, - "ParametersFilter": { - "base": "

This data type is deprecated. Instead, use ParameterStringFilter.

", - "refs": { - "ParametersFilterList$member": null + "ParametersFilter":{ + "base":"

This data type is deprecated. Instead, use ParameterStringFilter.

", + "refs":{ + "ParametersFilterList$member":null } }, - "ParametersFilterKey": { - "base": null, - "refs": { - "ParametersFilter$Key": "

The name of the filter.

" + "ParametersFilterKey":{ + "base":null, + "refs":{ + "ParametersFilter$Key":"

The name of the filter.

" } }, - "ParametersFilterList": { - "base": null, - "refs": { - "DescribeParametersRequest$Filters": "

One or more filters. Use a filter to return a more specific list of results.

" + "ParametersFilterList":{ + "base":null, + "refs":{ + "DescribeParametersRequest$Filters":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "ParametersFilterValue": { - "base": null, - "refs": { - "ParametersFilterValueList$member": null + "ParametersFilterValue":{ + "base":null, + "refs":{ + "ParametersFilterValueList$member":null } }, - "ParametersFilterValueList": { - "base": null, - "refs": { - "ParametersFilter$Values": "

The filter values.

" + "ParametersFilterValueList":{ + "base":null, + "refs":{ + "ParametersFilter$Values":"

The filter values.

" } }, - "Patch": { - "base": "

Represents metadata about a patch.

", - "refs": { - "EffectivePatch$Patch": "

Provides metadata for a patch, including information such as the KB ID, severity, classification and a URL for where more information can be obtained about the patch.

", - "PatchList$member": null + "Patch":{ + "base":"

Represents metadata about a patch.

", + "refs":{ + "EffectivePatch$Patch":"

Provides metadata for a patch, including information such as the KB ID, severity, classification and a URL for where more information can be obtained about the patch.

", + "PatchList$member":null } }, - "PatchBaselineIdentity": { - "base": "

Defines the basic information about a patch baseline.

", - "refs": { - "PatchBaselineIdentityList$member": null, - "PatchGroupPatchBaselineMapping$BaselineIdentity": "

The patch baseline the patch group is registered with.

" + "PatchBaselineIdentity":{ + "base":"

Defines the basic information about a patch baseline.

", + "refs":{ + "PatchBaselineIdentityList$member":null, + "PatchGroupPatchBaselineMapping$BaselineIdentity":"

The patch baseline the patch group is registered with.

" } }, - "PatchBaselineIdentityList": { - "base": null, - "refs": { - "DescribePatchBaselinesResult$BaselineIdentities": "

An array of PatchBaselineIdentity elements.

" + "PatchBaselineIdentityList":{ + "base":null, + "refs":{ + "DescribePatchBaselinesResult$BaselineIdentities":"

An array of PatchBaselineIdentity elements.

" } }, - "PatchBaselineMaxResults": { - "base": null, - "refs": { - "DescribeAvailablePatchesRequest$MaxResults": "

The maximum number of patches to return (per page).

", - "DescribeEffectivePatchesForPatchBaselineRequest$MaxResults": "

The maximum number of patches to return (per page).

", - "DescribePatchBaselinesRequest$MaxResults": "

The maximum number of patch baselines to return (per page).

", - "DescribePatchGroupsRequest$MaxResults": "

The maximum number of patch groups to return (per page).

" + "PatchBaselineMaxResults":{ + "base":null, + "refs":{ + "DescribeAvailablePatchesRequest$MaxResults":"

The maximum number of patches to return (per page).

", + "DescribeEffectivePatchesForPatchBaselineRequest$MaxResults":"

The maximum number of patches to return (per page).

", + "DescribePatchBaselinesRequest$MaxResults":"

The maximum number of patch baselines to return (per page).

", + "DescribePatchGroupsRequest$MaxResults":"

The maximum number of patch groups to return (per page).

" } }, - "PatchClassification": { - "base": null, - "refs": { - "Patch$Classification": "

The classification of the patch (for example, SecurityUpdates, Updates, CriticalUpdates).

", - "PatchComplianceData$Classification": "

The classification of the patch (for example, SecurityUpdates, Updates, CriticalUpdates).

" + "PatchClassification":{ + "base":null, + "refs":{ + "Patch$Classification":"

The classification of the patch (for example, SecurityUpdates, Updates, CriticalUpdates).

", + "PatchComplianceData$Classification":"

The classification of the patch (for example, SecurityUpdates, Updates, CriticalUpdates).

" } }, - "PatchComplianceData": { - "base": "

Information about the state of a patch on a particular instance as it relates to the patch baseline used to patch the instance.

", - "refs": { - "PatchComplianceDataList$member": null + "PatchComplianceData":{ + "base":"

Information about the state of a patch on a particular instance as it relates to the patch baseline used to patch the instance.

", + "refs":{ + "PatchComplianceDataList$member":null } }, - "PatchComplianceDataList": { - "base": null, - "refs": { - "DescribeInstancePatchesResult$Patches": "

Each entry in the array is a structure containing:

Title (string)

KBId (string)

Classification (string)

Severity (string)

State (string: \"INSTALLED\", \"INSTALLED OTHER\", \"MISSING\", \"NOT APPLICABLE\", \"FAILED\")

InstalledTime (DateTime)

InstalledBy (string)

" + "PatchComplianceDataList":{ + "base":null, + "refs":{ + "DescribeInstancePatchesResult$Patches":"

Each entry in the array is a structure containing:

Title (string)

KBId (string)

Classification (string)

Severity (string)

State (string: \"INSTALLED\", \"INSTALLED OTHER\", \"MISSING\", \"NOT APPLICABLE\", \"FAILED\")

InstalledTime (DateTime)

InstalledBy (string)

" } }, - "PatchComplianceDataState": { - "base": null, - "refs": { - "PatchComplianceData$State": "

The state of the patch on the instance (INSTALLED, INSTALLED_OTHER, MISSING, NOT_APPLICABLE or FAILED).

" + "PatchComplianceDataState":{ + "base":null, + "refs":{ + "PatchComplianceData$State":"

The state of the patch on the instance (INSTALLED, INSTALLED_OTHER, MISSING, NOT_APPLICABLE or FAILED).

" } }, - "PatchComplianceLevel": { - "base": null, - "refs": { - "CreatePatchBaselineRequest$ApprovedPatchesComplianceLevel": "

Defines the compliance level for approved patches. This means that if an approved patch is reported as missing, this is the severity of the compliance violation. The default value is UNSPECIFIED.

", - "GetPatchBaselineResult$ApprovedPatchesComplianceLevel": "

Returns the specified compliance severity level for approved patches in the patch baseline.

", - "PatchRule$ComplianceLevel": "

A compliance severity level for all approved patches in a patch baseline. Valid compliance severity levels include the following: Unspecified, Critical, High, Medium, Low, and Informational.

", - "PatchStatus$ComplianceLevel": "

The compliance severity level for a patch.

", - "UpdatePatchBaselineRequest$ApprovedPatchesComplianceLevel": "

Assigns a new compliance severity level to an existing patch baseline.

", - "UpdatePatchBaselineResult$ApprovedPatchesComplianceLevel": "

The compliance severity level assigned to the patch baseline after the update completed.

" + "PatchComplianceLevel":{ + "base":null, + "refs":{ + "CreatePatchBaselineRequest$ApprovedPatchesComplianceLevel":"

Defines the compliance level for approved patches. This means that if an approved patch is reported as missing, this is the severity of the compliance violation. The default value is UNSPECIFIED.

", + "GetPatchBaselineResult$ApprovedPatchesComplianceLevel":"

Returns the specified compliance severity level for approved patches in the patch baseline.

", + "PatchRule$ComplianceLevel":"

A compliance severity level for all approved patches in a patch baseline. Valid compliance severity levels include the following: Unspecified, Critical, High, Medium, Low, and Informational.

", + "PatchStatus$ComplianceLevel":"

The compliance severity level for a patch.

", + "UpdatePatchBaselineRequest$ApprovedPatchesComplianceLevel":"

Assigns a new compliance severity level to an existing patch baseline.

", + "UpdatePatchBaselineResult$ApprovedPatchesComplianceLevel":"

The compliance severity level assigned to the patch baseline after the update completed.

" } }, - "PatchComplianceMaxResults": { - "base": null, - "refs": { - "DescribeInstancePatchStatesForPatchGroupRequest$MaxResults": "

The maximum number of patches to return (per page).

", - "DescribeInstancePatchStatesRequest$MaxResults": "

The maximum number of instances to return (per page).

", - "DescribeInstancePatchesRequest$MaxResults": "

The maximum number of patches to return (per page).

" + "PatchComplianceMaxResults":{ + "base":null, + "refs":{ + "DescribeInstancePatchStatesForPatchGroupRequest$MaxResults":"

The maximum number of patches to return (per page).

", + "DescribeInstancePatchStatesRequest$MaxResults":"

The maximum number of instances to return (per page).

", + "DescribeInstancePatchesRequest$MaxResults":"

The maximum number of patches to return (per page).

" } }, - "PatchContentUrl": { - "base": null, - "refs": { - "Patch$ContentUrl": "

The URL where more information can be obtained about the patch.

" + "PatchContentUrl":{ + "base":null, + "refs":{ + "Patch$ContentUrl":"

The URL where more information can be obtained about the patch.

" } }, - "PatchDeploymentStatus": { - "base": null, - "refs": { - "PatchStatus$DeploymentStatus": "

The approval status of a patch (APPROVED, PENDING_APPROVAL, EXPLICIT_APPROVED, EXPLICIT_REJECTED).

" + "PatchDeploymentStatus":{ + "base":null, + "refs":{ + "PatchStatus$DeploymentStatus":"

The approval status of a patch (APPROVED, PENDING_APPROVAL, EXPLICIT_APPROVED, EXPLICIT_REJECTED).

" } }, - "PatchDescription": { - "base": null, - "refs": { - "Patch$Description": "

The description of the patch.

" + "PatchDescription":{ + "base":null, + "refs":{ + "Patch$Description":"

The description of the patch.

" } }, - "PatchFailedCount": { - "base": null, - "refs": { - "InstancePatchState$FailedCount": "

The number of patches from the patch baseline that were attempted to be installed during the last patching operation, but failed to install.

" + "PatchFailedCount":{ + "base":null, + "refs":{ + "InstancePatchState$FailedCount":"

The number of patches from the patch baseline that were attempted to be installed during the last patching operation, but failed to install.

" } }, - "PatchFilter": { - "base": "

Defines a patch filter.

A patch filter consists of key/value pairs, but not all keys are valid for all operating system types. For example, the key PRODUCT is valid for all supported operating system types. The key MSRC_SEVERITY, however, is valid only for Windows operating systems, and the key SECTION is valid only for Ubuntu operating systems.

Refer to the following sections for information about which keys may be used with each major operating system, and which values are valid for each key.

Windows Operating Systems

The supported keys for Windows operating systems are PRODUCT, CLASSIFICATION, and MSRC_SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • Windows7

  • Windows8

  • Windows8.1

  • Windows8Embedded

  • Windows10

  • Windows10LTSB

  • WindowsServer2008

  • WindowsServer2008R2

  • WindowsServer2012

  • WindowsServer2012R2

  • WindowsServer2016

Supported key: CLASSIFICATION

Supported values:

  • CriticalUpdates

  • DefinitionUpdates

  • Drivers

  • FeaturePacks

  • SecurityUpdates

  • ServicePacks

  • Tools

  • UpdateRollups

  • Updates

  • Upgrades

Supported key: MSRC_SEVERITY

Supported values:

  • Critical

  • Important

  • Moderate

  • Low

  • Unspecified

Ubuntu Operating Systems

The supported keys for Ubuntu operating systems are PRODUCT, PRIORITY, and SECTION. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • Ubuntu14.04

  • Ubuntu16.04

Supported key: PRIORITY

Supported values:

  • Required

  • Important

  • Standard

  • Optional

  • Extra

Supported key: SECTION

Only the length of the key value is validated. Minimum length is 1. Maximum length is 64.

Amazon Linux Operating Systems

The supported keys for Amazon Linux operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • AmazonLinux2012.03

  • AmazonLinux2012.09

  • AmazonLinux2013.03

  • AmazonLinux2013.09

  • AmazonLinux2014.03

  • AmazonLinux2014.09

  • AmazonLinux2015.03

  • AmazonLinux2015.09

  • AmazonLinux2016.03

  • AmazonLinux2016.09

  • AmazonLinux2017.03

  • AmazonLinux2017.09

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

RedHat Enterprise Linux (RHEL) Operating Systems

The supported keys for RedHat Enterprise Linux operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • RedhatEnterpriseLinux6.5

  • RedhatEnterpriseLinux6.6

  • RedhatEnterpriseLinux6.7

  • RedhatEnterpriseLinux6.8

  • RedhatEnterpriseLinux6.9

  • RedhatEnterpriseLinux7.0

  • RedhatEnterpriseLinux7.1

  • RedhatEnterpriseLinux7.2

  • RedhatEnterpriseLinux7.3

  • RedhatEnterpriseLinux7.4

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

SUSE Linux Enterprise Server (SLES) Operating Systems

The supported keys for SLES operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • Suse12.0

  • Suse12.1

  • Suse12.2

  • Suse12.3

  • Suse12.4

  • Suse12.5

  • Suse12.6

  • Suse12.7

  • Suse12.8

  • Suse12.9

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Recommended

  • Optional

  • Feature

  • Document

  • Yast

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Moderate

  • Low

CentOS Operating Systems

The supported keys for CentOS operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • CentOS6.5

  • CentOS6.6

  • CentOS6.7

  • CentOS6.8

  • CentOS6.9

  • CentOS7.0

  • CentOS7.1

  • CentOS7.2

  • CentOS7.3

  • CentOS7.4

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

", - "refs": { - "PatchFilterList$member": null + "PatchFilter":{ + "base":"

Defines a patch filter.

A patch filter consists of key/value pairs, but not all keys are valid for all operating system types. For example, the key PRODUCT is valid for all supported operating system types. The key MSRC_SEVERITY, however, is valid only for Windows operating systems, and the key SECTION is valid only for Ubuntu operating systems.

Refer to the following sections for information about which keys may be used with each major operating system, and which values are valid for each key.

Windows Operating Systems

The supported keys for Windows operating systems are PRODUCT, CLASSIFICATION, and MSRC_SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • Windows7

  • Windows8

  • Windows8.1

  • Windows8Embedded

  • Windows10

  • Windows10LTSB

  • WindowsServer2008

  • WindowsServer2008R2

  • WindowsServer2012

  • WindowsServer2012R2

  • WindowsServer2016

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: CLASSIFICATION

Supported values:

  • CriticalUpdates

  • DefinitionUpdates

  • Drivers

  • FeaturePacks

  • SecurityUpdates

  • ServicePacks

  • Tools

  • UpdateRollups

  • Updates

  • Upgrades

Supported key: MSRC_SEVERITY

Supported values:

  • Critical

  • Important

  • Moderate

  • Low

  • Unspecified

Ubuntu Operating Systems

The supported keys for Ubuntu operating systems are PRODUCT, PRIORITY, and SECTION. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • Ubuntu14.04

  • Ubuntu16.04

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: PRIORITY

Supported values:

  • Required

  • Important

  • Standard

  • Optional

  • Extra

Supported key: SECTION

Only the length of the key value is validated. Minimum length is 1. Maximum length is 64.

Amazon Linux Operating Systems

The supported keys for Amazon Linux operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • AmazonLinux2012.03

  • AmazonLinux2012.09

  • AmazonLinux2013.03

  • AmazonLinux2013.09

  • AmazonLinux2014.03

  • AmazonLinux2014.09

  • AmazonLinux2015.03

  • AmazonLinux2015.09

  • AmazonLinux2016.03

  • AmazonLinux2016.09

  • AmazonLinux2017.03

  • AmazonLinux2017.09

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

Amazon Linux 2 Operating Systems

The supported keys for Amazon Linux 2 operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • AmazonLinux2

  • AmazonLinux2.0

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

RedHat Enterprise Linux (RHEL) Operating Systems

The supported keys for RedHat Enterprise Linux operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • RedhatEnterpriseLinux6.5

  • RedhatEnterpriseLinux6.6

  • RedhatEnterpriseLinux6.7

  • RedhatEnterpriseLinux6.8

  • RedhatEnterpriseLinux6.9

  • RedhatEnterpriseLinux7.0

  • RedhatEnterpriseLinux7.1

  • RedhatEnterpriseLinux7.2

  • RedhatEnterpriseLinux7.3

  • RedhatEnterpriseLinux7.4

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

SUSE Linux Enterprise Server (SLES) Operating Systems

The supported keys for SLES operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • Suse12.0

  • Suse12.1

  • Suse12.2

  • Suse12.3

  • Suse12.4

  • Suse12.5

  • Suse12.6

  • Suse12.7

  • Suse12.8

  • Suse12.9

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Recommended

  • Optional

  • Feature

  • Document

  • Yast

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Moderate

  • Low

CentOS Operating Systems

The supported keys for CentOS operating systems are PRODUCT, CLASSIFICATION, and SEVERITY. See the following lists for valid values for each of these keys.

Supported key: PRODUCT

Supported values:

  • CentOS6.5

  • CentOS6.6

  • CentOS6.7

  • CentOS6.8

  • CentOS6.9

  • CentOS7.0

  • CentOS7.1

  • CentOS7.2

  • CentOS7.3

  • CentOS7.4

  • *

    Use a wildcard character (*) to target all supported operating system versions.

Supported key: CLASSIFICATION

Supported values:

  • Security

  • Bugfix

  • Enhancement

  • Recommended

  • Newpackage

Supported key: SEVERITY

Supported values:

  • Critical

  • Important

  • Medium

  • Low

", + "refs":{ + "PatchFilterList$member":null } }, - "PatchFilterGroup": { - "base": "

A set of patch filters, typically used for approval rules.

", - "refs": { - "CreatePatchBaselineRequest$GlobalFilters": "

A set of global filters used to exclude patches from the baseline.

", - "GetPatchBaselineResult$GlobalFilters": "

A set of global filters used to exclude patches from the baseline.

", - "PatchRule$PatchFilterGroup": "

The patch filter group that defines the criteria for the rule.

", - "UpdatePatchBaselineRequest$GlobalFilters": "

A set of global filters used to exclude patches from the baseline.

", - "UpdatePatchBaselineResult$GlobalFilters": "

A set of global filters used to exclude patches from the baseline.

" + "PatchFilterGroup":{ + "base":"

A set of patch filters, typically used for approval rules.

", + "refs":{ + "CreatePatchBaselineRequest$GlobalFilters":"

A set of global filters used to exclude patches from the baseline.

", + "GetPatchBaselineResult$GlobalFilters":"

A set of global filters used to exclude patches from the baseline.

", + "PatchRule$PatchFilterGroup":"

The patch filter group that defines the criteria for the rule.

", + "UpdatePatchBaselineRequest$GlobalFilters":"

A set of global filters used to exclude patches from the baseline.

", + "UpdatePatchBaselineResult$GlobalFilters":"

A set of global filters used to exclude patches from the baseline.

" } }, - "PatchFilterKey": { - "base": null, - "refs": { - "PatchFilter$Key": "

The key for the filter.

See PatchFilter for lists of valid keys for each operating system type.

" + "PatchFilterKey":{ + "base":null, + "refs":{ + "PatchFilter$Key":"

The key for the filter.

See PatchFilter for lists of valid keys for each operating system type.

" } }, - "PatchFilterList": { - "base": null, - "refs": { - "PatchFilterGroup$PatchFilters": "

The set of patch filters that make up the group.

" + "PatchFilterList":{ + "base":null, + "refs":{ + "PatchFilterGroup$PatchFilters":"

The set of patch filters that make up the group.

" } }, - "PatchFilterValue": { - "base": null, - "refs": { - "PatchFilterValueList$member": null + "PatchFilterValue":{ + "base":null, + "refs":{ + "PatchFilterValueList$member":null } }, - "PatchFilterValueList": { - "base": null, - "refs": { - "PatchFilter$Values": "

The value for the filter key.

See PatchFilter for lists of valid values for each key based on operating system type.

" + "PatchFilterValueList":{ + "base":null, + "refs":{ + "PatchFilter$Values":"

The value for the filter key.

See PatchFilter for lists of valid values for each key based on operating system type.

" } }, - "PatchGroup": { - "base": null, - "refs": { - "DeregisterPatchBaselineForPatchGroupRequest$PatchGroup": "

The name of the patch group that should be deregistered from the patch baseline.

", - "DeregisterPatchBaselineForPatchGroupResult$PatchGroup": "

The name of the patch group deregistered from the patch baseline.

", - "DescribeInstancePatchStatesForPatchGroupRequest$PatchGroup": "

The name of the patch group for which the patch state information should be retrieved.

", - "DescribePatchGroupStateRequest$PatchGroup": "

The name of the patch group whose patch snapshot should be retrieved.

", - "GetPatchBaselineForPatchGroupRequest$PatchGroup": "

The name of the patch group whose patch baseline should be retrieved.

", - "GetPatchBaselineForPatchGroupResult$PatchGroup": "

The name of the patch group.

", - "InstancePatchState$PatchGroup": "

The name of the patch group the managed instance belongs to.

", - "PatchGroupList$member": null, - "PatchGroupPatchBaselineMapping$PatchGroup": "

The name of the patch group registered with the patch baseline.

", - "RegisterPatchBaselineForPatchGroupRequest$PatchGroup": "

The name of the patch group that should be registered with the patch baseline.

", - "RegisterPatchBaselineForPatchGroupResult$PatchGroup": "

The name of the patch group registered with the patch baseline.

" + "PatchGroup":{ + "base":null, + "refs":{ + "DeregisterPatchBaselineForPatchGroupRequest$PatchGroup":"

The name of the patch group that should be deregistered from the patch baseline.

", + "DeregisterPatchBaselineForPatchGroupResult$PatchGroup":"

The name of the patch group deregistered from the patch baseline.

", + "DescribeInstancePatchStatesForPatchGroupRequest$PatchGroup":"

The name of the patch group for which the patch state information should be retrieved.

", + "DescribePatchGroupStateRequest$PatchGroup":"

The name of the patch group whose patch snapshot should be retrieved.

", + "GetPatchBaselineForPatchGroupRequest$PatchGroup":"

The name of the patch group whose patch baseline should be retrieved.

", + "GetPatchBaselineForPatchGroupResult$PatchGroup":"

The name of the patch group.

", + "InstancePatchState$PatchGroup":"

The name of the patch group the managed instance belongs to.

", + "PatchGroupList$member":null, + "PatchGroupPatchBaselineMapping$PatchGroup":"

The name of the patch group registered with the patch baseline.

", + "RegisterPatchBaselineForPatchGroupRequest$PatchGroup":"

The name of the patch group that should be registered with the patch baseline.

", + "RegisterPatchBaselineForPatchGroupResult$PatchGroup":"

The name of the patch group registered with the patch baseline.

" } }, - "PatchGroupList": { - "base": null, - "refs": { - "GetPatchBaselineResult$PatchGroups": "

Patch groups included in the patch baseline.

" + "PatchGroupList":{ + "base":null, + "refs":{ + "GetPatchBaselineResult$PatchGroups":"

Patch groups included in the patch baseline.

" } }, - "PatchGroupPatchBaselineMapping": { - "base": "

The mapping between a patch group and the patch baseline the patch group is registered with.

", - "refs": { - "PatchGroupPatchBaselineMappingList$member": null + "PatchGroupPatchBaselineMapping":{ + "base":"

The mapping between a patch group and the patch baseline the patch group is registered with.

", + "refs":{ + "PatchGroupPatchBaselineMappingList$member":null } }, - "PatchGroupPatchBaselineMappingList": { - "base": null, - "refs": { - "DescribePatchGroupsResult$Mappings": "

Each entry in the array contains:

PatchGroup: string (between 1 and 256 characters, Regex: ^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$)

PatchBaselineIdentity: A PatchBaselineIdentity element.

" + "PatchGroupPatchBaselineMappingList":{ + "base":null, + "refs":{ + "DescribePatchGroupsResult$Mappings":"

Each entry in the array contains:

PatchGroup: string (between 1 and 256 characters, Regex: ^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$)

PatchBaselineIdentity: A PatchBaselineIdentity element.

" } }, - "PatchId": { - "base": null, - "refs": { - "Patch$Id": "

The ID of the patch (this is different than the Microsoft Knowledge Base ID).

", - "PatchIdList$member": null + "PatchId":{ + "base":null, + "refs":{ + "Patch$Id":"

The ID of the patch (this is different than the Microsoft Knowledge Base ID).

", + "PatchIdList$member":null } }, - "PatchIdList": { - "base": null, - "refs": { - "CreatePatchBaselineRequest$ApprovedPatches": "

A list of explicitly approved patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", - "CreatePatchBaselineRequest$RejectedPatches": "

A list of explicitly rejected patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", - "GetPatchBaselineResult$ApprovedPatches": "

A list of explicitly approved patches for the baseline.

", - "GetPatchBaselineResult$RejectedPatches": "

A list of explicitly rejected patches for the baseline.

", - "UpdatePatchBaselineRequest$ApprovedPatches": "

A list of explicitly approved patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", - "UpdatePatchBaselineRequest$RejectedPatches": "

A list of explicitly rejected patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", - "UpdatePatchBaselineResult$ApprovedPatches": "

A list of explicitly approved patches for the baseline.

", - "UpdatePatchBaselineResult$RejectedPatches": "

A list of explicitly rejected patches for the baseline.

" + "PatchIdList":{ + "base":null, + "refs":{ + "CreatePatchBaselineRequest$ApprovedPatches":"

A list of explicitly approved patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", + "CreatePatchBaselineRequest$RejectedPatches":"

A list of explicitly rejected patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", + "GetPatchBaselineResult$ApprovedPatches":"

A list of explicitly approved patches for the baseline.

", + "GetPatchBaselineResult$RejectedPatches":"

A list of explicitly rejected patches for the baseline.

", + "UpdatePatchBaselineRequest$ApprovedPatches":"

A list of explicitly approved patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", + "UpdatePatchBaselineRequest$RejectedPatches":"

A list of explicitly rejected patches for the baseline.

For information about accepted formats for lists of approved patches and rejected patches, see Package Name Formats for Approved and Rejected Patch Lists in the AWS Systems Manager User Guide.

", + "UpdatePatchBaselineResult$ApprovedPatches":"

A list of explicitly approved patches for the baseline.

", + "UpdatePatchBaselineResult$RejectedPatches":"

A list of explicitly rejected patches for the baseline.

" } }, - "PatchInstalledCount": { - "base": null, - "refs": { - "InstancePatchState$InstalledCount": "

The number of patches from the patch baseline that are installed on the instance.

" + "PatchInstalledCount":{ + "base":null, + "refs":{ + "InstancePatchState$InstalledCount":"

The number of patches from the patch baseline that are installed on the instance.

" } }, - "PatchInstalledOtherCount": { - "base": null, - "refs": { - "InstancePatchState$InstalledOtherCount": "

The number of patches not specified in the patch baseline that are installed on the instance.

" + "PatchInstalledOtherCount":{ + "base":null, + "refs":{ + "InstancePatchState$InstalledOtherCount":"

The number of patches not specified in the patch baseline that are installed on the instance.

" } }, - "PatchKbNumber": { - "base": null, - "refs": { - "Patch$KbNumber": "

The Microsoft Knowledge Base ID of the patch.

", - "PatchComplianceData$KBId": "

The operating system-specific ID of the patch.

" + "PatchKbNumber":{ + "base":null, + "refs":{ + "Patch$KbNumber":"

The Microsoft Knowledge Base ID of the patch.

", + "PatchComplianceData$KBId":"

The operating system-specific ID of the patch.

" } }, - "PatchLanguage": { - "base": null, - "refs": { - "Patch$Language": "

The language of the patch if it's language-specific.

" + "PatchLanguage":{ + "base":null, + "refs":{ + "Patch$Language":"

The language of the patch if it's language-specific.

" } }, - "PatchList": { - "base": null, - "refs": { - "DescribeAvailablePatchesResult$Patches": "

An array of patches. Each entry in the array is a patch structure.

" + "PatchList":{ + "base":null, + "refs":{ + "DescribeAvailablePatchesResult$Patches":"

An array of patches. Each entry in the array is a patch structure.

" } }, - "PatchMissingCount": { - "base": null, - "refs": { - "InstancePatchState$MissingCount": "

The number of patches from the patch baseline that are applicable for the instance but aren't currently installed.

" + "PatchMissingCount":{ + "base":null, + "refs":{ + "InstancePatchState$MissingCount":"

The number of patches from the patch baseline that are applicable for the instance but aren't currently installed.

" } }, - "PatchMsrcNumber": { - "base": null, - "refs": { - "Patch$MsrcNumber": "

The ID of the MSRC bulletin the patch is related to.

" + "PatchMsrcNumber":{ + "base":null, + "refs":{ + "Patch$MsrcNumber":"

The ID of the MSRC bulletin the patch is related to.

" } }, - "PatchMsrcSeverity": { - "base": null, - "refs": { - "Patch$MsrcSeverity": "

The severity of the patch (for example Critical, Important, Moderate).

" + "PatchMsrcSeverity":{ + "base":null, + "refs":{ + "Patch$MsrcSeverity":"

The severity of the patch (for example Critical, Important, Moderate).

" } }, - "PatchNotApplicableCount": { - "base": null, - "refs": { - "InstancePatchState$NotApplicableCount": "

The number of patches from the patch baseline that aren't applicable for the instance and hence aren't installed on the instance.

" + "PatchNotApplicableCount":{ + "base":null, + "refs":{ + "InstancePatchState$NotApplicableCount":"

The number of patches from the patch baseline that aren't applicable for the instance and hence aren't installed on the instance.

" } }, - "PatchOperationType": { - "base": null, - "refs": { - "InstancePatchState$Operation": "

The type of patching operation that was performed: SCAN (assess patch compliance state) or INSTALL (install missing patches).

" + "PatchOperationType":{ + "base":null, + "refs":{ + "InstancePatchState$Operation":"

The type of patching operation that was performed: SCAN (assess patch compliance state) or INSTALL (install missing patches).

" } }, - "PatchOrchestratorFilter": { - "base": "

Defines a filter used in Patch Manager APIs.

", - "refs": { - "PatchOrchestratorFilterList$member": null + "PatchOrchestratorFilter":{ + "base":"

Defines a filter used in Patch Manager APIs.

", + "refs":{ + "PatchOrchestratorFilterList$member":null } }, - "PatchOrchestratorFilterKey": { - "base": null, - "refs": { - "PatchOrchestratorFilter$Key": "

The key for the filter.

" + "PatchOrchestratorFilterKey":{ + "base":null, + "refs":{ + "PatchOrchestratorFilter$Key":"

The key for the filter.

" } }, - "PatchOrchestratorFilterList": { - "base": null, - "refs": { - "DescribeAvailablePatchesRequest$Filters": "

Filters used to scope down the returned patches.

", - "DescribeInstancePatchesRequest$Filters": "

Each entry in the array is a structure containing:

Key (string, between 1 and 128 characters)

Values (array of strings, each string between 1 and 256 characters)

", - "DescribePatchBaselinesRequest$Filters": "

Each element in the array is a structure containing:

Key: (string, \"NAME_PREFIX\" or \"OWNER\")

Value: (array of strings, exactly 1 entry, between 1 and 255 characters)

", - "DescribePatchGroupsRequest$Filters": "

One or more filters. Use a filter to return a more specific list of results.

" + "PatchOrchestratorFilterList":{ + "base":null, + "refs":{ + "DescribeAvailablePatchesRequest$Filters":"

Filters used to scope down the returned patches.

", + "DescribeInstancePatchesRequest$Filters":"

Each entry in the array is a structure containing:

Key (string, between 1 and 128 characters)

Values (array of strings, each string between 1 and 256 characters)

", + "DescribePatchBaselinesRequest$Filters":"

Each element in the array is a structure containing:

Key: (string, \"NAME_PREFIX\" or \"OWNER\")

Value: (array of strings, exactly 1 entry, between 1 and 255 characters)

", + "DescribePatchGroupsRequest$Filters":"

One or more filters. Use a filter to return a more specific list of results.

" } }, - "PatchOrchestratorFilterValue": { - "base": null, - "refs": { - "PatchOrchestratorFilterValues$member": null + "PatchOrchestratorFilterValue":{ + "base":null, + "refs":{ + "PatchOrchestratorFilterValues$member":null } }, - "PatchOrchestratorFilterValues": { - "base": null, - "refs": { - "PatchOrchestratorFilter$Values": "

The value for the filter.

" + "PatchOrchestratorFilterValues":{ + "base":null, + "refs":{ + "PatchOrchestratorFilter$Values":"

The value for the filter.

" } }, - "PatchProduct": { - "base": null, - "refs": { - "Patch$Product": "

The specific product the patch is applicable for (for example, WindowsServer2016).

" + "PatchProduct":{ + "base":null, + "refs":{ + "Patch$Product":"

The specific product the patch is applicable for (for example, WindowsServer2016).

" } }, - "PatchProductFamily": { - "base": null, - "refs": { - "Patch$ProductFamily": "

The product family the patch is applicable for (for example, Windows).

" + "PatchProductFamily":{ + "base":null, + "refs":{ + "Patch$ProductFamily":"

The product family the patch is applicable for (for example, Windows).

" } }, - "PatchRule": { - "base": "

Defines an approval rule for a patch baseline.

", - "refs": { - "PatchRuleList$member": null + "PatchRule":{ + "base":"

Defines an approval rule for a patch baseline.

", + "refs":{ + "PatchRuleList$member":null } }, - "PatchRuleGroup": { - "base": "

A set of rules defining the approval rules for a patch baseline.

", - "refs": { - "CreatePatchBaselineRequest$ApprovalRules": "

A set of rules used to include patches in the baseline.

", - "GetPatchBaselineResult$ApprovalRules": "

A set of rules used to include patches in the baseline.

", - "UpdatePatchBaselineRequest$ApprovalRules": "

A set of rules used to include patches in the baseline.

", - "UpdatePatchBaselineResult$ApprovalRules": "

A set of rules used to include patches in the baseline.

" + "PatchRuleGroup":{ + "base":"

A set of rules defining the approval rules for a patch baseline.

", + "refs":{ + "CreatePatchBaselineRequest$ApprovalRules":"

A set of rules used to include patches in the baseline.

", + "GetPatchBaselineResult$ApprovalRules":"

A set of rules used to include patches in the baseline.

", + "UpdatePatchBaselineRequest$ApprovalRules":"

A set of rules used to include patches in the baseline.

", + "UpdatePatchBaselineResult$ApprovalRules":"

A set of rules used to include patches in the baseline.

" } }, - "PatchRuleList": { - "base": null, - "refs": { - "PatchRuleGroup$PatchRules": "

The rules that make up the rule group.

" + "PatchRuleList":{ + "base":null, + "refs":{ + "PatchRuleGroup$PatchRules":"

The rules that make up the rule group.

" } }, - "PatchSeverity": { - "base": null, - "refs": { - "PatchComplianceData$Severity": "

The severity of the patch (for example, Critical, Important, Moderate).

" + "PatchSeverity":{ + "base":null, + "refs":{ + "PatchComplianceData$Severity":"

The severity of the patch (for example, Critical, Important, Moderate).

" } }, - "PatchSource": { - "base": "

Information about the patches to use to update the instances, including target operating systems and source repository. Applies to Linux instances only.

", - "refs": { - "PatchSourceList$member": null + "PatchSource":{ + "base":"

Information about the patches to use to update the instances, including target operating systems and source repository. Applies to Linux instances only.

", + "refs":{ + "PatchSourceList$member":null } }, - "PatchSourceConfiguration": { - "base": null, - "refs": { - "PatchSource$Configuration": "

The value of the yum repo configuration. For example:

cachedir=/var/cache/yum/$basesearch

$releasever

keepcache=0

debualevel=2

" + "PatchSourceConfiguration":{ + "base":null, + "refs":{ + "PatchSource$Configuration":"

The value of the yum repo configuration. For example:

cachedir=/var/cache/yum/$basesearch

$releasever

keepcache=0

debualevel=2

" } }, - "PatchSourceList": { - "base": null, - "refs": { - "CreatePatchBaselineRequest$Sources": "

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

", - "GetPatchBaselineResult$Sources": "

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

", - "UpdatePatchBaselineRequest$Sources": "

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

", - "UpdatePatchBaselineResult$Sources": "

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

" + "PatchSourceList":{ + "base":null, + "refs":{ + "CreatePatchBaselineRequest$Sources":"

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

", + "GetPatchBaselineResult$Sources":"

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

", + "UpdatePatchBaselineRequest$Sources":"

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

", + "UpdatePatchBaselineResult$Sources":"

Information about the patches to use to update the instances, including target operating systems and source repositories. Applies to Linux instances only.

" } }, - "PatchSourceName": { - "base": null, - "refs": { - "PatchSource$Name": "

The name specified to identify the patch source.

" + "PatchSourceName":{ + "base":null, + "refs":{ + "PatchSource$Name":"

The name specified to identify the patch source.

" } }, - "PatchSourceProduct": { - "base": null, - "refs": { - "PatchSourceProductList$member": null + "PatchSourceProduct":{ + "base":null, + "refs":{ + "PatchSourceProductList$member":null } }, - "PatchSourceProductList": { - "base": null, - "refs": { - "PatchSource$Products": "

The specific operating system versions a patch repository applies to, such as \"Ubuntu16.04\", \"AmazonLinux2016.09\", \"RedhatEnterpriseLinux7.2\" or \"Suse12.7\". For lists of supported product values, see PatchFilter.

" + "PatchSourceProductList":{ + "base":null, + "refs":{ + "PatchSource$Products":"

The specific operating system versions a patch repository applies to, such as \"Ubuntu16.04\", \"AmazonLinux2016.09\", \"RedhatEnterpriseLinux7.2\" or \"Suse12.7\". For lists of supported product values, see PatchFilter.

" } }, - "PatchStatus": { - "base": "

Information about the approval status of a patch.

", - "refs": { - "EffectivePatch$PatchStatus": "

The status of the patch in a patch baseline. This includes information about whether the patch is currently approved, due to be approved by a rule, explicitly approved, or explicitly rejected and the date the patch was or will be approved.

" + "PatchStatus":{ + "base":"

Information about the approval status of a patch.

", + "refs":{ + "EffectivePatch$PatchStatus":"

The status of the patch in a patch baseline. This includes information about whether the patch is currently approved, due to be approved by a rule, explicitly approved, or explicitly rejected and the date the patch was or will be approved.

" } }, - "PatchTitle": { - "base": null, - "refs": { - "Patch$Title": "

The title of the patch.

", - "PatchComplianceData$Title": "

The title of the patch.

" + "PatchTitle":{ + "base":null, + "refs":{ + "Patch$Title":"

The title of the patch.

", + "PatchComplianceData$Title":"

The title of the patch.

" } }, - "PatchVendor": { - "base": null, - "refs": { - "Patch$Vendor": "

The name of the vendor providing the patch.

" + "PatchVendor":{ + "base":null, + "refs":{ + "Patch$Vendor":"

The name of the vendor providing the patch.

" } }, - "PingStatus": { - "base": null, - "refs": { - "InstanceInformation$PingStatus": "

Connection status of the SSM Agent.

" + "PingStatus":{ + "base":null, + "refs":{ + "InstanceInformation$PingStatus":"

Connection status of SSM Agent.

" } }, - "PlatformType": { - "base": null, - "refs": { - "InstanceInformation$PlatformType": "

The operating system platform type.

", - "PlatformTypeList$member": null + "PlatformType":{ + "base":null, + "refs":{ + "InstanceInformation$PlatformType":"

The operating system platform type.

", + "PlatformTypeList$member":null } }, - "PlatformTypeList": { - "base": null, - "refs": { - "DocumentDescription$PlatformTypes": "

The list of OS platforms compatible with this Systems Manager document.

", - "DocumentIdentifier$PlatformTypes": "

The operating system platform.

" + "PlatformTypeList":{ + "base":null, + "refs":{ + "DocumentDescription$PlatformTypes":"

The list of OS platforms compatible with this Systems Manager document.

", + "DocumentIdentifier$PlatformTypes":"

The operating system platform.

" } }, - "Product": { - "base": null, - "refs": { - "GetDeployablePatchSnapshotForInstanceResult$Product": "

Returns the specific operating system (for example Windows Server 2012 or Amazon Linux 2015.09) on the instance for the specified patch snapshot.

" + "Product":{ + "base":null, + "refs":{ + "GetDeployablePatchSnapshotForInstanceResult$Product":"

Returns the specific operating system (for example Windows Server 2012 or Amazon Linux 2015.09) on the instance for the specified patch snapshot.

" } }, - "PutComplianceItemsRequest": { - "base": null, - "refs": { + "PutComplianceItemsRequest":{ + "base":null, + "refs":{ } }, - "PutComplianceItemsResult": { - "base": null, - "refs": { + "PutComplianceItemsResult":{ + "base":null, + "refs":{ } }, - "PutInventoryMessage": { - "base": null, - "refs": { - "PutInventoryResult$Message": "

Information about the request.

" + "PutInventoryMessage":{ + "base":null, + "refs":{ + "PutInventoryResult$Message":"

Information about the request.

" } }, - "PutInventoryRequest": { - "base": null, - "refs": { + "PutInventoryRequest":{ + "base":null, + "refs":{ } }, - "PutInventoryResult": { - "base": null, - "refs": { + "PutInventoryResult":{ + "base":null, + "refs":{ } }, - "PutParameterRequest": { - "base": null, - "refs": { + "PutParameterRequest":{ + "base":null, + "refs":{ } }, - "PutParameterResult": { - "base": null, - "refs": { + "PutParameterResult":{ + "base":null, + "refs":{ } }, - "RegisterDefaultPatchBaselineRequest": { - "base": null, - "refs": { + "RegisterDefaultPatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "RegisterDefaultPatchBaselineResult": { - "base": null, - "refs": { + "RegisterDefaultPatchBaselineResult":{ + "base":null, + "refs":{ } }, - "RegisterPatchBaselineForPatchGroupRequest": { - "base": null, - "refs": { + "RegisterPatchBaselineForPatchGroupRequest":{ + "base":null, + "refs":{ } }, - "RegisterPatchBaselineForPatchGroupResult": { - "base": null, - "refs": { + "RegisterPatchBaselineForPatchGroupResult":{ + "base":null, + "refs":{ } }, - "RegisterTargetWithMaintenanceWindowRequest": { - "base": null, - "refs": { + "RegisterTargetWithMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "RegisterTargetWithMaintenanceWindowResult": { - "base": null, - "refs": { + "RegisterTargetWithMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "RegisterTaskWithMaintenanceWindowRequest": { - "base": null, - "refs": { + "RegisterTaskWithMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "RegisterTaskWithMaintenanceWindowResult": { - "base": null, - "refs": { + "RegisterTaskWithMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "RegistrationLimit": { - "base": null, - "refs": { - "Activation$RegistrationLimit": "

The maximum number of managed instances that can be registered using this activation.

", - "CreateActivationRequest$RegistrationLimit": "

Specify the maximum number of managed instances you want to register. The default value is 1 instance.

" + "RegistrationLimit":{ + "base":null, + "refs":{ + "Activation$RegistrationLimit":"

The maximum number of managed instances that can be registered using this activation.

", + "CreateActivationRequest$RegistrationLimit":"

Specify the maximum number of managed instances you want to register. The default value is 1 instance.

" } }, - "RegistrationsCount": { - "base": null, - "refs": { - "Activation$RegistrationsCount": "

The number of managed instances already registered with this activation.

" + "RegistrationsCount":{ + "base":null, + "refs":{ + "Activation$RegistrationsCount":"

The number of managed instances already registered with this activation.

" } }, - "RemainingCount": { - "base": null, - "refs": { - "InventoryDeletionSummary$RemainingCount": "

Remaining number of items to delete.

", - "InventoryDeletionSummaryItem$RemainingCount": "

The remaining number of items to delete.

" + "RemainingCount":{ + "base":null, + "refs":{ + "InventoryDeletionSummary$RemainingCount":"

Remaining number of items to delete.

", + "InventoryDeletionSummaryItem$RemainingCount":"

The remaining number of items to delete.

" } }, - "RemoveTagsFromResourceRequest": { - "base": null, - "refs": { + "RemoveTagsFromResourceRequest":{ + "base":null, + "refs":{ } }, - "RemoveTagsFromResourceResult": { - "base": null, - "refs": { + "RemoveTagsFromResourceResult":{ + "base":null, + "refs":{ } }, - "ResolvedTargets": { - "base": "

Information about targets that resolved during the Automation execution.

", - "refs": { - "AutomationExecution$ResolvedTargets": "

A list of resolved targets in the rate control execution.

", - "AutomationExecutionMetadata$ResolvedTargets": "

A list of targets that resolved during the execution.

" + "ResolvedTargets":{ + "base":"

Information about targets that resolved during the Automation execution.

", + "refs":{ + "AutomationExecution$ResolvedTargets":"

A list of resolved targets in the rate control execution.

", + "AutomationExecutionMetadata$ResolvedTargets":"

A list of targets that resolved during the execution.

" } }, - "ResourceComplianceSummaryItem": { - "base": "

Compliance summary information for a specific resource.

", - "refs": { - "ResourceComplianceSummaryItemList$member": null + "ResourceComplianceSummaryItem":{ + "base":"

Compliance summary information for a specific resource.

", + "refs":{ + "ResourceComplianceSummaryItemList$member":null } }, - "ResourceComplianceSummaryItemList": { - "base": null, - "refs": { - "ListResourceComplianceSummariesResult$ResourceComplianceSummaryItems": "

A summary count for specified or targeted managed instances. Summary count includes information about compliant and non-compliant State Manager associations, patch status, or custom items according to the filter criteria that you specify.

" + "ResourceComplianceSummaryItemList":{ + "base":null, + "refs":{ + "ListResourceComplianceSummariesResult$ResourceComplianceSummaryItems":"

A summary count for specified or targeted managed instances. Summary count includes information about compliant and non-compliant State Manager associations, patch status, or custom items according to the filter criteria that you specify.

" } }, - "ResourceCount": { - "base": null, - "refs": { - "InventoryDeletionSummaryItem$Count": "

A count of the number of deleted items.

" + "ResourceCount":{ + "base":null, + "refs":{ + "InventoryDeletionSummaryItem$Count":"

A count of the number of deleted items.

" } }, - "ResourceDataSyncAWSKMSKeyARN": { - "base": null, - "refs": { - "ResourceDataSyncS3Destination$AWSKMSKeyARN": "

The ARN of an encryption key for a destination in Amazon S3. Must belong to the same region as the destination Amazon S3 bucket.

" + "ResourceDataSyncAWSKMSKeyARN":{ + "base":null, + "refs":{ + "ResourceDataSyncS3Destination$AWSKMSKeyARN":"

The ARN of an encryption key for a destination in Amazon S3. Must belong to the same region as the destination Amazon S3 bucket.

" } }, - "ResourceDataSyncAlreadyExistsException": { - "base": "

A sync configuration with the same name already exists.

", - "refs": { + "ResourceDataSyncAlreadyExistsException":{ + "base":"

A sync configuration with the same name already exists.

", + "refs":{ } }, - "ResourceDataSyncCountExceededException": { - "base": "

You have exceeded the allowed maximum sync configurations.

", - "refs": { + "ResourceDataSyncCountExceededException":{ + "base":"

You have exceeded the allowed maximum sync configurations.

", + "refs":{ } }, - "ResourceDataSyncCreatedTime": { - "base": null, - "refs": { - "ResourceDataSyncItem$SyncCreatedTime": "

The date and time the configuration was created (UTC).

" + "ResourceDataSyncCreatedTime":{ + "base":null, + "refs":{ + "ResourceDataSyncItem$SyncCreatedTime":"

The date and time the configuration was created (UTC).

" } }, - "ResourceDataSyncInvalidConfigurationException": { - "base": "

The specified sync configuration is invalid.

", - "refs": { + "ResourceDataSyncInvalidConfigurationException":{ + "base":"

The specified sync configuration is invalid.

", + "refs":{ } }, - "ResourceDataSyncItem": { - "base": "

Information about a Resource Data Sync configuration, including its current status and last successful sync.

", - "refs": { - "ResourceDataSyncItemList$member": null + "ResourceDataSyncItem":{ + "base":"

Information about a Resource Data Sync configuration, including its current status and last successful sync.

", + "refs":{ + "ResourceDataSyncItemList$member":null } }, - "ResourceDataSyncItemList": { - "base": null, - "refs": { - "ListResourceDataSyncResult$ResourceDataSyncItems": "

A list of your current Resource Data Sync configurations and their statuses.

" + "ResourceDataSyncItemList":{ + "base":null, + "refs":{ + "ListResourceDataSyncResult$ResourceDataSyncItems":"

A list of your current Resource Data Sync configurations and their statuses.

" } }, - "ResourceDataSyncName": { - "base": null, - "refs": { - "CreateResourceDataSyncRequest$SyncName": "

A name for the configuration.

", - "DeleteResourceDataSyncRequest$SyncName": "

The name of the configuration to delete.

", - "ResourceDataSyncAlreadyExistsException$SyncName": null, - "ResourceDataSyncItem$SyncName": "

The name of the Resource Data Sync.

", - "ResourceDataSyncNotFoundException$SyncName": null + "ResourceDataSyncName":{ + "base":null, + "refs":{ + "CreateResourceDataSyncRequest$SyncName":"

A name for the configuration.

", + "DeleteResourceDataSyncRequest$SyncName":"

The name of the configuration to delete.

", + "ResourceDataSyncAlreadyExistsException$SyncName":null, + "ResourceDataSyncItem$SyncName":"

The name of the Resource Data Sync.

", + "ResourceDataSyncNotFoundException$SyncName":null } }, - "ResourceDataSyncNotFoundException": { - "base": "

The specified sync name was not found.

", - "refs": { + "ResourceDataSyncNotFoundException":{ + "base":"

The specified sync name was not found.

", + "refs":{ } }, - "ResourceDataSyncS3BucketName": { - "base": null, - "refs": { - "ResourceDataSyncS3Destination$BucketName": "

The name of the Amazon S3 bucket where the aggregated data is stored.

" + "ResourceDataSyncS3BucketName":{ + "base":null, + "refs":{ + "ResourceDataSyncS3Destination$BucketName":"

The name of the Amazon S3 bucket where the aggregated data is stored.

" } }, - "ResourceDataSyncS3Destination": { - "base": "

Information about the target Amazon S3 bucket for the Resource Data Sync.

", - "refs": { - "CreateResourceDataSyncRequest$S3Destination": "

Amazon S3 configuration details for the sync.

", - "ResourceDataSyncItem$S3Destination": "

Configuration information for the target Amazon S3 bucket.

" + "ResourceDataSyncS3Destination":{ + "base":"

Information about the target Amazon S3 bucket for the Resource Data Sync.

", + "refs":{ + "CreateResourceDataSyncRequest$S3Destination":"

Amazon S3 configuration details for the sync.

", + "ResourceDataSyncItem$S3Destination":"

Configuration information for the target Amazon S3 bucket.

" } }, - "ResourceDataSyncS3Format": { - "base": null, - "refs": { - "ResourceDataSyncS3Destination$SyncFormat": "

A supported sync format. The following format is currently supported: JsonSerDe

" + "ResourceDataSyncS3Format":{ + "base":null, + "refs":{ + "ResourceDataSyncS3Destination$SyncFormat":"

A supported sync format. The following format is currently supported: JsonSerDe

" } }, - "ResourceDataSyncS3Prefix": { - "base": null, - "refs": { - "ResourceDataSyncS3Destination$Prefix": "

An Amazon S3 prefix for the bucket.

" + "ResourceDataSyncS3Prefix":{ + "base":null, + "refs":{ + "ResourceDataSyncS3Destination$Prefix":"

An Amazon S3 prefix for the bucket.

" } }, - "ResourceDataSyncS3Region": { - "base": null, - "refs": { - "ResourceDataSyncS3Destination$Region": "

The AWS Region with the Amazon S3 bucket targeted by the Resource Data Sync.

" + "ResourceDataSyncS3Region":{ + "base":null, + "refs":{ + "ResourceDataSyncS3Destination$Region":"

The AWS Region with the Amazon S3 bucket targeted by the Resource Data Sync.

" } }, - "ResourceId": { - "base": null, - "refs": { - "AddTagsToResourceRequest$ResourceId": "

The resource ID you want to tag.

Use the ID of the resource. Here are some examples:

ManagedInstance: mi-012345abcde

MaintenanceWindow: mw-012345abcde

PatchBaseline: pb-012345abcde

For the Document and Parameter values, use the name of the resource.

The ManagedInstance type for this API action is only for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

", - "ListTagsForResourceRequest$ResourceId": "

The resource ID for which you want to see a list of tags.

", - "RemoveTagsFromResourceRequest$ResourceId": "

The resource ID for which you want to remove tags. Use the ID of the resource. Here are some examples:

ManagedInstance: mi-012345abcde

MaintenanceWindow: mw-012345abcde

PatchBaseline: pb-012345abcde

For the Document and Parameter values, use the name of the resource.

The ManagedInstance type for this API action is only for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

" + "ResourceId":{ + "base":null, + "refs":{ + "AddTagsToResourceRequest$ResourceId":"

The resource ID you want to tag.

Use the ID of the resource. Here are some examples:

ManagedInstance: mi-012345abcde

MaintenanceWindow: mw-012345abcde

PatchBaseline: pb-012345abcde

For the Document and Parameter values, use the name of the resource.

The ManagedInstance type for this API action is only for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

", + "ListTagsForResourceRequest$ResourceId":"

The resource ID for which you want to see a list of tags.

", + "RemoveTagsFromResourceRequest$ResourceId":"

The resource ID for which you want to remove tags. Use the ID of the resource. Here are some examples:

ManagedInstance: mi-012345abcde

MaintenanceWindow: mw-012345abcde

PatchBaseline: pb-012345abcde

For the Document and Parameter values, use the name of the resource.

The ManagedInstance type for this API action is only for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

" } }, - "ResourceInUseException": { - "base": "

Error returned if an attempt is made to delete a patch baseline that is registered for a patch group.

", - "refs": { + "ResourceInUseException":{ + "base":"

Error returned if an attempt is made to delete a patch baseline that is registered for a patch group.

", + "refs":{ } }, - "ResourceLimitExceededException": { - "base": "

Error returned when the caller has exceeded the default resource limits. For example, too many Maintenance Windows or Patch baselines have been created.

For information about resource limits in Systems Manager, see AWS Systems Manager Limits.

", - "refs": { + "ResourceLimitExceededException":{ + "base":"

Error returned when the caller has exceeded the default resource limits. For example, too many Maintenance Windows or Patch baselines have been created.

For information about resource limits in Systems Manager, see AWS Systems Manager Limits.

", + "refs":{ } }, - "ResourceType": { - "base": null, - "refs": { - "InstanceInformation$ResourceType": "

The type of instance. Instances are either EC2 instances or managed instances.

" + "ResourceType":{ + "base":null, + "refs":{ + "InstanceInformation$ResourceType":"

The type of instance. Instances are either EC2 instances or managed instances.

" } }, - "ResourceTypeForTagging": { - "base": null, - "refs": { - "AddTagsToResourceRequest$ResourceType": "

Specifies the type of resource you are tagging.

The ManagedInstance type for this API action is for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

", - "ListTagsForResourceRequest$ResourceType": "

Returns a list of tags for a specific resource type.

", - "RemoveTagsFromResourceRequest$ResourceType": "

The type of resource of which you want to remove a tag.

The ManagedInstance type for this API action is only for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

" + "ResourceTypeForTagging":{ + "base":null, + "refs":{ + "AddTagsToResourceRequest$ResourceType":"

Specifies the type of resource you are tagging.

The ManagedInstance type for this API action is for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

", + "ListTagsForResourceRequest$ResourceType":"

Returns a list of tags for a specific resource type.

", + "RemoveTagsFromResourceRequest$ResourceType":"

The type of resource of which you want to remove a tag.

The ManagedInstance type for this API action is only for on-premises managed instances. You must specify the the name of the managed instance in the following format: mi-ID_number. For example, mi-1a2b3c4d5e6f.

" } }, - "ResponseCode": { - "base": null, - "refs": { - "CommandPlugin$ResponseCode": "

A numeric response code generated after executing the plugin.

", - "GetCommandInvocationResult$ResponseCode": "

The error level response code for the plugin script. If the response code is -1, then the command has not started executing on the instance, or it was not received by the instance.

" + "ResponseCode":{ + "base":null, + "refs":{ + "CommandPlugin$ResponseCode":"

A numeric response code generated after executing the plugin.

", + "GetCommandInvocationResult$ResponseCode":"

The error level response code for the plugin script. If the response code is -1, then the command has not started executing on the instance, or it was not received by the instance.

" } }, - "ResultAttribute": { - "base": "

The inventory item result attribute.

", - "refs": { - "ResultAttributeList$member": null + "ResultAttribute":{ + "base":"

The inventory item result attribute.

", + "refs":{ + "ResultAttributeList$member":null } }, - "ResultAttributeList": { - "base": null, - "refs": { - "GetInventoryRequest$ResultAttributes": "

The list of inventory item types to return.

" + "ResultAttributeList":{ + "base":null, + "refs":{ + "GetInventoryRequest$ResultAttributes":"

The list of inventory item types to return.

" } }, - "S3BucketName": { - "base": null, - "refs": { - "Command$OutputS3BucketName": "

The S3 bucket where the responses to the command executions should be stored. This was requested when issuing the command.

", - "CommandPlugin$OutputS3BucketName": "

The S3 bucket where the responses to the command executions should be stored. This was requested when issuing the command. For example, in the following response:

test_folder/ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix/i-1234567876543/awsrunShellScript

test_folder is the name of the Amazon S3 bucket;

ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix is the name of the S3 prefix;

i-1234567876543 is the instance ID;

awsrunShellScript is the name of the plugin.

", - "LoggingInfo$S3BucketName": "

The name of an Amazon S3 bucket where execution logs are stored .

", - "MaintenanceWindowRunCommandParameters$OutputS3BucketName": "

The name of the Amazon S3 bucket.

", - "S3OutputLocation$OutputS3BucketName": "

The name of the Amazon S3 bucket.

", - "SendCommandRequest$OutputS3BucketName": "

The name of the S3 bucket where command execution responses should be stored.

" + "S3BucketName":{ + "base":null, + "refs":{ + "Command$OutputS3BucketName":"

The S3 bucket where the responses to the command executions should be stored. This was requested when issuing the command.

", + "CommandPlugin$OutputS3BucketName":"

The S3 bucket where the responses to the command executions should be stored. This was requested when issuing the command. For example, in the following response:

test_folder/ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix/i-1234567876543/awsrunShellScript

test_folder is the name of the Amazon S3 bucket;

ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix is the name of the S3 prefix;

i-1234567876543 is the instance ID;

awsrunShellScript is the name of the plugin.

", + "LoggingInfo$S3BucketName":"

The name of an Amazon S3 bucket where execution logs are stored .

", + "MaintenanceWindowRunCommandParameters$OutputS3BucketName":"

The name of the Amazon S3 bucket.

", + "S3OutputLocation$OutputS3BucketName":"

The name of the Amazon S3 bucket.

", + "SendCommandRequest$OutputS3BucketName":"

The name of the S3 bucket where command execution responses should be stored.

" } }, - "S3KeyPrefix": { - "base": null, - "refs": { - "Command$OutputS3KeyPrefix": "

The S3 directory path inside the bucket where the responses to the command executions should be stored. This was requested when issuing the command.

", - "CommandPlugin$OutputS3KeyPrefix": "

The S3 directory path inside the bucket where the responses to the command executions should be stored. This was requested when issuing the command. For example, in the following response:

test_folder/ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix/i-1234567876543/awsrunShellScript

test_folder is the name of the Amazon S3 bucket;

ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix is the name of the S3 prefix;

i-1234567876543 is the instance ID;

awsrunShellScript is the name of the plugin.

", - "LoggingInfo$S3KeyPrefix": "

(Optional) The Amazon S3 bucket subfolder.

", - "MaintenanceWindowRunCommandParameters$OutputS3KeyPrefix": "

The Amazon S3 bucket subfolder.

", - "S3OutputLocation$OutputS3KeyPrefix": "

The Amazon S3 bucket subfolder.

", - "SendCommandRequest$OutputS3KeyPrefix": "

The directory structure within the S3 bucket where the responses should be stored.

" + "S3KeyPrefix":{ + "base":null, + "refs":{ + "Command$OutputS3KeyPrefix":"

The S3 directory path inside the bucket where the responses to the command executions should be stored. This was requested when issuing the command.

", + "CommandPlugin$OutputS3KeyPrefix":"

The S3 directory path inside the bucket where the responses to the command executions should be stored. This was requested when issuing the command. For example, in the following response:

test_folder/ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix/i-1234567876543/awsrunShellScript

test_folder is the name of the Amazon S3 bucket;

ab19cb99-a030-46dd-9dfc-8eSAMPLEPre-Fix is the name of the S3 prefix;

i-1234567876543 is the instance ID;

awsrunShellScript is the name of the plugin.

", + "LoggingInfo$S3KeyPrefix":"

(Optional) The Amazon S3 bucket subfolder.

", + "MaintenanceWindowRunCommandParameters$OutputS3KeyPrefix":"

The Amazon S3 bucket subfolder.

", + "S3OutputLocation$OutputS3KeyPrefix":"

The Amazon S3 bucket subfolder.

", + "SendCommandRequest$OutputS3KeyPrefix":"

The directory structure within the S3 bucket where the responses should be stored.

" } }, - "S3OutputLocation": { - "base": "

An Amazon S3 bucket where you want to store the results of this request.

", - "refs": { - "InstanceAssociationOutputLocation$S3Location": "

An Amazon S3 bucket where you want to store the results of this request.

" + "S3OutputLocation":{ + "base":"

An Amazon S3 bucket where you want to store the results of this request.

", + "refs":{ + "InstanceAssociationOutputLocation$S3Location":"

An Amazon S3 bucket where you want to store the results of this request.

" } }, - "S3OutputUrl": { - "base": "

A URL for the Amazon S3 bucket where you want to store the results of this request.

", - "refs": { - "InstanceAssociationOutputUrl$S3OutputUrl": "

The URL of Amazon S3 bucket where you want to store the results of this request.

" + "S3OutputUrl":{ + "base":"

A URL for the Amazon S3 bucket where you want to store the results of this request.

", + "refs":{ + "InstanceAssociationOutputUrl$S3OutputUrl":"

The URL of Amazon S3 bucket where you want to store the results of this request.

" } }, - "S3Region": { - "base": null, - "refs": { - "Command$OutputS3Region": "

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

", - "CommandPlugin$OutputS3Region": "

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

", - "LoggingInfo$S3Region": "

The region where the Amazon S3 bucket is located.

", - "S3OutputLocation$OutputS3Region": "

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

", - "SendCommandRequest$OutputS3Region": "

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

" + "S3Region":{ + "base":null, + "refs":{ + "Command$OutputS3Region":"

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

", + "CommandPlugin$OutputS3Region":"

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

", + "LoggingInfo$S3Region":"

The region where the Amazon S3 bucket is located.

", + "S3OutputLocation$OutputS3Region":"

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

", + "SendCommandRequest$OutputS3Region":"

(Deprecated) You can no longer specify this parameter. The system ignores it. Instead, Systems Manager automatically determines the Amazon S3 bucket region.

" } }, - "ScheduleExpression": { - "base": null, - "refs": { - "Association$ScheduleExpression": "

A cron expression that specifies a schedule when the association runs.

", - "AssociationDescription$ScheduleExpression": "

A cron expression that specifies a schedule when the association runs.

", - "AssociationVersionInfo$ScheduleExpression": "

The cron or rate schedule specified for the association when the association version was created.

", - "CreateAssociationBatchRequestEntry$ScheduleExpression": "

A cron expression that specifies a schedule when the association runs.

", - "CreateAssociationRequest$ScheduleExpression": "

A cron expression when the association will be applied to the target(s).

", - "UpdateAssociationRequest$ScheduleExpression": "

The cron expression used to schedule the association that you want to update.

" + "ScheduleExpression":{ + "base":null, + "refs":{ + "Association$ScheduleExpression":"

A cron expression that specifies a schedule when the association runs.

", + "AssociationDescription$ScheduleExpression":"

A cron expression that specifies a schedule when the association runs.

", + "AssociationVersionInfo$ScheduleExpression":"

The cron or rate schedule specified for the association when the association version was created.

", + "CreateAssociationBatchRequestEntry$ScheduleExpression":"

A cron expression that specifies a schedule when the association runs.

", + "CreateAssociationRequest$ScheduleExpression":"

A cron expression when the association will be applied to the target(s).

", + "UpdateAssociationRequest$ScheduleExpression":"

The cron expression used to schedule the association that you want to update.

" } }, - "SendAutomationSignalRequest": { - "base": null, - "refs": { + "SendAutomationSignalRequest":{ + "base":null, + "refs":{ } }, - "SendAutomationSignalResult": { - "base": null, - "refs": { + "SendAutomationSignalResult":{ + "base":null, + "refs":{ } }, - "SendCommandRequest": { - "base": null, - "refs": { + "SendCommandRequest":{ + "base":null, + "refs":{ } }, - "SendCommandResult": { - "base": null, - "refs": { + "SendCommandResult":{ + "base":null, + "refs":{ } }, - "ServiceRole": { - "base": null, - "refs": { - "Command$ServiceRole": "

The IAM service role that Run Command uses to act on your behalf when sending notifications about command status changes.

", - "CommandInvocation$ServiceRole": "

The IAM service role that Run Command uses to act on your behalf when sending notifications about command status changes on a per instance basis.

", - "GetMaintenanceWindowExecutionTaskResult$ServiceRole": "

The role that was assumed when executing the task.

", - "GetMaintenanceWindowTaskResult$ServiceRoleArn": "

The IAM service role to assume during task execution.

", - "MaintenanceWindowRunCommandParameters$ServiceRoleArn": "

The IAM service role to assume during task execution.

", - "MaintenanceWindowTask$ServiceRoleArn": "

The role that should be assumed when executing the task

", - "RegisterTaskWithMaintenanceWindowRequest$ServiceRoleArn": "

The role that should be assumed when executing the task.

", - "SendCommandRequest$ServiceRoleArn": "

The IAM role that Systems Manager uses to send notifications.

", - "UpdateMaintenanceWindowTaskRequest$ServiceRoleArn": "

The IAM service role ARN to modify. The system assumes this role during task execution.

", - "UpdateMaintenanceWindowTaskResult$ServiceRoleArn": "

The updated service role ARN value.

" + "ServiceRole":{ + "base":null, + "refs":{ + "Command$ServiceRole":"

The IAM service role that Run Command uses to act on your behalf when sending notifications about command status changes.

", + "CommandInvocation$ServiceRole":"

The IAM service role that Run Command uses to act on your behalf when sending notifications about command status changes on a per instance basis.

", + "GetMaintenanceWindowExecutionTaskResult$ServiceRole":"

The role that was assumed when executing the task.

", + "GetMaintenanceWindowTaskResult$ServiceRoleArn":"

The IAM service role to assume during task execution.

", + "MaintenanceWindowRunCommandParameters$ServiceRoleArn":"

The IAM service role to assume during task execution.

", + "MaintenanceWindowTask$ServiceRoleArn":"

The role that should be assumed when executing the task

", + "RegisterTaskWithMaintenanceWindowRequest$ServiceRoleArn":"

The role that should be assumed when executing the task.

", + "SendCommandRequest$ServiceRoleArn":"

The IAM role that Systems Manager uses to send notifications.

", + "UpdateMaintenanceWindowTaskRequest$ServiceRoleArn":"

The IAM service role ARN to modify. The system assumes this role during task execution.

", + "UpdateMaintenanceWindowTaskResult$ServiceRoleArn":"

The updated service role ARN value.

" } }, - "SeveritySummary": { - "base": "

The number of managed instances found for each patch severity level defined in the request filter.

", - "refs": { - "CompliantSummary$SeveritySummary": "

A summary of the compliance severity by compliance type.

", - "NonCompliantSummary$SeveritySummary": "

A summary of the non-compliance severity by compliance type

" + "SeveritySummary":{ + "base":"

The number of managed instances found for each patch severity level defined in the request filter.

", + "refs":{ + "CompliantSummary$SeveritySummary":"

A summary of the compliance severity by compliance type.

", + "NonCompliantSummary$SeveritySummary":"

A summary of the non-compliance severity by compliance type

" } }, - "SignalType": { - "base": null, - "refs": { - "SendAutomationSignalRequest$SignalType": "

The type of signal. Valid signal types include the following: Approve and Reject

" + "SignalType":{ + "base":null, + "refs":{ + "SendAutomationSignalRequest$SignalType":"

The type of signal. Valid signal types include the following: Approve and Reject

" } }, - "SnapshotDownloadUrl": { - "base": null, - "refs": { - "GetDeployablePatchSnapshotForInstanceResult$SnapshotDownloadUrl": "

A pre-signed Amazon S3 URL that can be used to download the patch snapshot.

" + "SnapshotDownloadUrl":{ + "base":null, + "refs":{ + "GetDeployablePatchSnapshotForInstanceResult$SnapshotDownloadUrl":"

A pre-signed Amazon S3 URL that can be used to download the patch snapshot.

" } }, - "SnapshotId": { - "base": null, - "refs": { - "GetDeployablePatchSnapshotForInstanceRequest$SnapshotId": "

The user-defined snapshot ID.

", - "GetDeployablePatchSnapshotForInstanceResult$SnapshotId": "

The user-defined snapshot ID.

", - "InstancePatchState$SnapshotId": "

The ID of the patch baseline snapshot used during the patching operation when this compliance data was collected.

" + "SnapshotId":{ + "base":null, + "refs":{ + "GetDeployablePatchSnapshotForInstanceRequest$SnapshotId":"

The user-defined snapshot ID.

", + "GetDeployablePatchSnapshotForInstanceResult$SnapshotId":"

The user-defined snapshot ID.

", + "InstancePatchState$SnapshotId":"

The ID of the patch baseline snapshot used during the patching operation when this compliance data was collected.

" } }, - "StandardErrorContent": { - "base": null, - "refs": { - "GetCommandInvocationResult$StandardErrorContent": "

The first 8,000 characters written by the plugin to stderr. If the command has not finished executing, then this string is empty.

" + "StandardErrorContent":{ + "base":null, + "refs":{ + "GetCommandInvocationResult$StandardErrorContent":"

The first 8,000 characters written by the plugin to stderr. If the command has not finished executing, then this string is empty.

" } }, - "StandardOutputContent": { - "base": null, - "refs": { - "GetCommandInvocationResult$StandardOutputContent": "

The first 24,000 characters written by the plugin to stdout. If the command has not finished executing, if ExecutionStatus is neither Succeeded nor Failed, then this string is empty.

" + "StandardOutputContent":{ + "base":null, + "refs":{ + "GetCommandInvocationResult$StandardOutputContent":"

The first 24,000 characters written by the plugin to stdout. If the command has not finished executing, if ExecutionStatus is neither Succeeded nor Failed, then this string is empty.

" } }, - "StartAutomationExecutionRequest": { - "base": null, - "refs": { + "StartAutomationExecutionRequest":{ + "base":null, + "refs":{ } }, - "StartAutomationExecutionResult": { - "base": null, - "refs": { + "StartAutomationExecutionResult":{ + "base":null, + "refs":{ } }, - "StatusAdditionalInfo": { - "base": null, - "refs": { - "AssociationStatus$AdditionalInfo": "

A user-defined string.

" + "StatusAdditionalInfo":{ + "base":null, + "refs":{ + "AssociationStatus$AdditionalInfo":"

A user-defined string.

" } }, - "StatusDetails": { - "base": null, - "refs": { - "Command$StatusDetails": "

A detailed status of the command execution. StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Run Command Status. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to any instances.

  • In Progress: The command has been sent to at least one instance but has not reached a final state on all instances.

  • Success: The command successfully executed on all invocations. This is a terminal state.

  • Delivery Timed Out: The value of MaxErrors or more command invocations shows a status of Delivery Timed Out. This is a terminal state.

  • Execution Timed Out: The value of MaxErrors or more command invocations shows a status of Execution Timed Out. This is a terminal state.

  • Failed: The value of MaxErrors or more command invocations shows a status of Failed. This is a terminal state.

  • Incomplete: The command was attempted on all instances and one or more invocations does not have a value of Success but not enough invocations failed for the status to be Failed. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Rate Exceeded: The number of instances targeted by the command exceeded the account limit for pending invocations. The system has canceled the command before executing it on any instance. This is a terminal state.

", - "CommandInvocation$StatusDetails": "

A detailed status of the command execution for each invocation (each instance targeted by the command). StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Run Command Status. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to the instance.

  • In Progress: The command has been sent to the instance but has not reached a terminal state.

  • Success: The execution of the command or plugin was successfully completed. This is a terminal state.

  • Delivery Timed Out: The command was not delivered to the instance before the delivery timeout expired. Delivery timeouts do not count against the parent command's MaxErrors limit, but they do contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Execution Timed Out: Command execution started on the instance, but the execution was not complete before the execution timeout expired. Execution timeouts count against the MaxErrors limit of the parent command. This is a terminal state.

  • Failed: The command was not successful on the instance. For a plugin, this indicates that the result code was not zero. For a command invocation, this indicates that the result code for one or more plugins was not zero. Invocation failures count against the MaxErrors limit of the parent command. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Undeliverable: The command can't be delivered to the instance. The instance might not exist or might not be responding. Undeliverable invocations don't count against the parent command's MaxErrors limit and don't contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Terminated: The parent command exceeded its MaxErrors limit and subsequent command invocations were canceled by the system. This is a terminal state.

", - "CommandPlugin$StatusDetails": "

A detailed status of the plugin execution. StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Run Command Status. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to the instance.

  • In Progress: The command has been sent to the instance but has not reached a terminal state.

  • Success: The execution of the command or plugin was successfully completed. This is a terminal state.

  • Delivery Timed Out: The command was not delivered to the instance before the delivery timeout expired. Delivery timeouts do not count against the parent command's MaxErrors limit, but they do contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Execution Timed Out: Command execution started on the instance, but the execution was not complete before the execution timeout expired. Execution timeouts count against the MaxErrors limit of the parent command. This is a terminal state.

  • Failed: The command was not successful on the instance. For a plugin, this indicates that the result code was not zero. For a command invocation, this indicates that the result code for one or more plugins was not zero. Invocation failures count against the MaxErrors limit of the parent command. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Undeliverable: The command can't be delivered to the instance. The instance might not exist, or it might not be responding. Undeliverable invocations don't count against the parent command's MaxErrors limit, and they don't contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Terminated: The parent command exceeded its MaxErrors limit and subsequent command invocations were canceled by the system. This is a terminal state.

", - "GetCommandInvocationResult$StatusDetails": "

A detailed status of the command execution for an invocation. StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Run Command Status. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to the instance.

  • In Progress: The command has been sent to the instance but has not reached a terminal state.

  • Delayed: The system attempted to send the command to the target, but the target was not available. The instance might not be available because of network issues, the instance was stopped, etc. The system will try to deliver the command again.

  • Success: The command or plugin was executed successfully. This is a terminal state.

  • Delivery Timed Out: The command was not delivered to the instance before the delivery timeout expired. Delivery timeouts do not count against the parent command's MaxErrors limit, but they do contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Execution Timed Out: The command started to execute on the instance, but the execution was not complete before the timeout expired. Execution timeouts count against the MaxErrors limit of the parent command. This is a terminal state.

  • Failed: The command wasn't executed successfully on the instance. For a plugin, this indicates that the result code was not zero. For a command invocation, this indicates that the result code for one or more plugins was not zero. Invocation failures count against the MaxErrors limit of the parent command. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Undeliverable: The command can't be delivered to the instance. The instance might not exist or might not be responding. Undeliverable invocations don't count against the parent command's MaxErrors limit and don't contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Terminated: The parent command exceeded its MaxErrors limit and subsequent command invocations were canceled by the system. This is a terminal state.

" + "StatusDetails":{ + "base":null, + "refs":{ + "Command$StatusDetails":"

A detailed status of the command execution. StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Understanding Command Statuses in the AWS Systems Manager User Guide. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to any instances.

  • In Progress: The command has been sent to at least one instance but has not reached a final state on all instances.

  • Success: The command successfully executed on all invocations. This is a terminal state.

  • Delivery Timed Out: The value of MaxErrors or more command invocations shows a status of Delivery Timed Out. This is a terminal state.

  • Execution Timed Out: The value of MaxErrors or more command invocations shows a status of Execution Timed Out. This is a terminal state.

  • Failed: The value of MaxErrors or more command invocations shows a status of Failed. This is a terminal state.

  • Incomplete: The command was attempted on all instances and one or more invocations does not have a value of Success but not enough invocations failed for the status to be Failed. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Rate Exceeded: The number of instances targeted by the command exceeded the account limit for pending invocations. The system has canceled the command before executing it on any instance. This is a terminal state.

", + "CommandInvocation$StatusDetails":"

A detailed status of the command execution for each invocation (each instance targeted by the command). StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Understanding Command Statuses in the AWS Systems Manager User Guide. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to the instance.

  • In Progress: The command has been sent to the instance but has not reached a terminal state.

  • Success: The execution of the command or plugin was successfully completed. This is a terminal state.

  • Delivery Timed Out: The command was not delivered to the instance before the delivery timeout expired. Delivery timeouts do not count against the parent command's MaxErrors limit, but they do contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Execution Timed Out: Command execution started on the instance, but the execution was not complete before the execution timeout expired. Execution timeouts count against the MaxErrors limit of the parent command. This is a terminal state.

  • Failed: The command was not successful on the instance. For a plugin, this indicates that the result code was not zero. For a command invocation, this indicates that the result code for one or more plugins was not zero. Invocation failures count against the MaxErrors limit of the parent command. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Undeliverable: The command can't be delivered to the instance. The instance might not exist or might not be responding. Undeliverable invocations don't count against the parent command's MaxErrors limit and don't contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Terminated: The parent command exceeded its MaxErrors limit and subsequent command invocations were canceled by the system. This is a terminal state.

", + "CommandPlugin$StatusDetails":"

A detailed status of the plugin execution. StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Understanding Command Statuses in the AWS Systems Manager User Guide. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to the instance.

  • In Progress: The command has been sent to the instance but has not reached a terminal state.

  • Success: The execution of the command or plugin was successfully completed. This is a terminal state.

  • Delivery Timed Out: The command was not delivered to the instance before the delivery timeout expired. Delivery timeouts do not count against the parent command's MaxErrors limit, but they do contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Execution Timed Out: Command execution started on the instance, but the execution was not complete before the execution timeout expired. Execution timeouts count against the MaxErrors limit of the parent command. This is a terminal state.

  • Failed: The command was not successful on the instance. For a plugin, this indicates that the result code was not zero. For a command invocation, this indicates that the result code for one or more plugins was not zero. Invocation failures count against the MaxErrors limit of the parent command. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Undeliverable: The command can't be delivered to the instance. The instance might not exist, or it might not be responding. Undeliverable invocations don't count against the parent command's MaxErrors limit, and they don't contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Terminated: The parent command exceeded its MaxErrors limit and subsequent command invocations were canceled by the system. This is a terminal state.

", + "GetCommandInvocationResult$StatusDetails":"

A detailed status of the command execution for an invocation. StatusDetails includes more information than Status because it includes states resulting from error and concurrency control parameters. StatusDetails can show different results than Status. For more information about these statuses, see Understanding Command Statuses in the AWS Systems Manager User Guide. StatusDetails can be one of the following values:

  • Pending: The command has not been sent to the instance.

  • In Progress: The command has been sent to the instance but has not reached a terminal state.

  • Delayed: The system attempted to send the command to the target, but the target was not available. The instance might not be available because of network issues, the instance was stopped, etc. The system will try to deliver the command again.

  • Success: The command or plugin was executed successfully. This is a terminal state.

  • Delivery Timed Out: The command was not delivered to the instance before the delivery timeout expired. Delivery timeouts do not count against the parent command's MaxErrors limit, but they do contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Execution Timed Out: The command started to execute on the instance, but the execution was not complete before the timeout expired. Execution timeouts count against the MaxErrors limit of the parent command. This is a terminal state.

  • Failed: The command wasn't executed successfully on the instance. For a plugin, this indicates that the result code was not zero. For a command invocation, this indicates that the result code for one or more plugins was not zero. Invocation failures count against the MaxErrors limit of the parent command. This is a terminal state.

  • Canceled: The command was terminated before it was completed. This is a terminal state.

  • Undeliverable: The command can't be delivered to the instance. The instance might not exist or might not be responding. Undeliverable invocations don't count against the parent command's MaxErrors limit and don't contribute to whether the parent command status is Success or Incomplete. This is a terminal state.

  • Terminated: The parent command exceeded its MaxErrors limit and subsequent command invocations were canceled by the system. This is a terminal state.

" } }, - "StatusMessage": { - "base": null, - "refs": { - "AssociationStatus$Message": "

The reason for the status.

" + "StatusMessage":{ + "base":null, + "refs":{ + "AssociationStatus$Message":"

The reason for the status.

" } }, - "StatusName": { - "base": null, - "refs": { - "AssociationOverview$Status": "

The status of the association. Status can be: Pending, Success, or Failed.

", - "AssociationOverview$DetailedStatus": "

A detailed status of the association.

", - "AssociationStatusAggregatedCount$key": null, - "InstanceAggregatedAssociationOverview$DetailedStatus": "

Detailed status information about the aggregated associations.

", - "InstanceAssociationStatusAggregatedCount$key": null, - "InstanceAssociationStatusInfo$Status": "

Status information about the instance association.

", - "InstanceAssociationStatusInfo$DetailedStatus": "

Detailed status information about the instance association.

", - "InstanceInformation$AssociationStatus": "

The status of the association.

" + "StatusName":{ + "base":null, + "refs":{ + "AssociationOverview$Status":"

The status of the association. Status can be: Pending, Success, or Failed.

", + "AssociationOverview$DetailedStatus":"

A detailed status of the association.

", + "AssociationStatusAggregatedCount$key":null, + "InstanceAggregatedAssociationOverview$DetailedStatus":"

Detailed status information about the aggregated associations.

", + "InstanceAssociationStatusAggregatedCount$key":null, + "InstanceAssociationStatusInfo$Status":"

Status information about the instance association.

", + "InstanceAssociationStatusInfo$DetailedStatus":"

Detailed status information about the instance association.

", + "InstanceInformation$AssociationStatus":"

The status of the association.

" } }, - "StatusUnchanged": { - "base": "

The updated status is the same as the current status.

", - "refs": { - } - }, - "StepExecution": { - "base": "

Detailed information about an the execution state of an Automation step.

", - "refs": { - "StepExecutionList$member": null - } - }, - "StepExecutionFilter": { - "base": "

A filter to limit the amount of step execution information returned by the call.

", - "refs": { - "StepExecutionFilterList$member": null - } - }, - "StepExecutionFilterKey": { - "base": null, - "refs": { - "StepExecutionFilter$Key": "

One or more keys to limit the results. Valid filter keys include the following: StepName, Action, StepExecutionId, StepExecutionStatus, StartTimeBefore, StartTimeAfter.

" - } - }, - "StepExecutionFilterList": { - "base": null, - "refs": { - "DescribeAutomationStepExecutionsRequest$Filters": "

One or more filters to limit the number of step executions returned by the request.

" - } - }, - "StepExecutionFilterValue": { - "base": null, - "refs": { - "StepExecutionFilterValueList$member": null - } - }, - "StepExecutionFilterValueList": { - "base": null, - "refs": { - "StepExecutionFilter$Values": "

The values of the filter key.

" - } - }, - "StepExecutionList": { - "base": null, - "refs": { - "AutomationExecution$StepExecutions": "

A list of details about the current state of all steps that comprise an execution. An Automation document contains a list of steps that are executed in order.

", - "DescribeAutomationStepExecutionsResult$StepExecutions": "

A list of details about the current state of all steps that make up an execution.

" - } - }, - "StopAutomationExecutionRequest": { - "base": null, - "refs": { - } - }, - "StopAutomationExecutionResult": { - "base": null, - "refs": { - } - }, - "StopType": { - "base": null, - "refs": { - "StopAutomationExecutionRequest$Type": "

The stop request type. Valid types include the following: Cancel and Complete. The default type is Cancel.

" - } - }, - "String": { - "base": null, - "refs": { - "AlreadyExistsException$Message": null, - "AssociationDoesNotExist$Message": null, - "AssociationVersionLimitExceeded$Message": null, - "AutomationDefinitionNotFoundException$Message": null, - "AutomationDefinitionVersionNotFoundException$Message": null, - "AutomationExecution$FailureMessage": "

A message describing why an execution has failed, if the status is set to Failed.

", - "AutomationExecution$ExecutedBy": "

The Amazon Resource Name (ARN) of the user who executed the automation.

", - "AutomationExecution$CurrentStepName": "

The name of the currently executing step.

", - "AutomationExecution$CurrentAction": "

The action of the currently executing step.

", - "AutomationExecution$Target": "

The target of the execution.

", - "AutomationExecutionLimitExceededException$Message": null, - "AutomationExecutionMetadata$ExecutedBy": "

The IAM role ARN of the user who executed the Automation.

", - "AutomationExecutionMetadata$LogFile": "

An Amazon S3 bucket where execution information is stored.

", - "AutomationExecutionMetadata$CurrentStepName": "

The name of the currently executing step.

", - "AutomationExecutionMetadata$CurrentAction": "

The action of the currently executing step.

", - "AutomationExecutionMetadata$FailureMessage": "

The list of execution outputs as defined in the Automation document.

", - "AutomationExecutionMetadata$Target": "

The list of execution outputs as defined in the Automation document.

", - "AutomationExecutionNotFoundException$Message": null, - "AutomationStepNotFoundException$Message": null, - "ComplianceTypeCountLimitExceededException$Message": null, - "CustomSchemaCountLimitExceededException$Message": null, - "DocumentAlreadyExists$Message": null, - "DocumentLimitExceeded$Message": null, - "DocumentPermissionLimit$Message": null, - "DocumentVersionLimitExceeded$Message": null, - "DoesNotExistException$Message": null, - "DuplicateDocumentContent$Message": null, - "FailureDetails$FailureStage": "

The stage of the Automation execution when the failure occurred. The stages include the following: InputValidation, PreVerification, Invocation, PostVerification.

", - "FailureDetails$FailureType": "

The type of Automation failure. Failure types include the following: Action, Permission, Throttling, Verification, Internal.

", - "FeatureNotAvailableException$Message": null, - "HierarchyLevelLimitExceededException$message": "

A hierarchy can have a maximum of 15 levels. For more information, see Working with Systems Manager Parameters.

", - "HierarchyTypeMismatchException$message": "

Parameter Store does not support changing a parameter type in a hierarchy. For example, you can't change a parameter from a String type to a SecureString type. You must create a new, unique parameter.

", - "IdempotentParameterMismatch$Message": null, - "InstanceInformation$PlatformName": "

The name of the operating system platform running on your instance.

", - "InstanceInformation$PlatformVersion": "

The version of the OS platform running on your instance.

", - "InstanceInformation$Name": "

The name of the managed instance.

", - "InternalServerError$Message": null, - "InvalidActivation$Message": null, - "InvalidActivationId$Message": null, - "InvalidAllowedPatternException$message": "

The request does not meet the regular expression requirement.

", - "InvalidAssociationVersion$Message": null, - "InvalidAutomationExecutionParametersException$Message": null, - "InvalidAutomationSignalException$Message": null, - "InvalidAutomationStatusUpdateException$Message": null, - "InvalidDeleteInventoryParametersException$Message": null, - "InvalidDeletionIdException$Message": null, - "InvalidDocument$Message": "

The document does not exist or the document is not available to the user. This exception can be issued by CreateAssociation, CreateAssociationBatch, DeleteAssociation, DeleteDocument, DescribeAssociation, DescribeDocument, GetDocument, SendCommand, or UpdateAssociationStatus.

", - "InvalidDocumentContent$Message": "

A description of the validation error.

", - "InvalidDocumentOperation$Message": null, - "InvalidDocumentSchemaVersion$Message": null, - "InvalidDocumentVersion$Message": null, - "InvalidFilter$Message": null, - "InvalidFilterOption$message": "

The specified filter option is not valid. Valid options are Equals and BeginsWith. For Path filter, valid options are Recursive and OneLevel.

", - "InvalidFilterValue$Message": null, - "InvalidInstanceId$Message": null, - "InvalidInstanceInformationFilterValue$message": null, - "InvalidInventoryItemContextException$Message": null, - "InvalidInventoryRequestException$Message": null, - "InvalidItemContentException$Message": null, - "InvalidKeyId$message": null, - "InvalidNextToken$Message": null, - "InvalidNotificationConfig$Message": null, - "InvalidOptionException$Message": null, - "InvalidParameters$Message": null, - "InvalidPermissionType$Message": null, - "InvalidResultAttributeException$Message": null, - "InvalidRole$Message": null, - "InvalidSchedule$Message": null, - "InvalidTarget$Message": null, - "InvalidTypeNameException$Message": null, - "InvalidUpdate$Message": null, - "ItemContentMismatchException$Message": null, - "ItemSizeLimitExceededException$Message": null, - "MaxDocumentSizeExceeded$Message": null, - "NormalStringMap$key": null, - "NormalStringMap$value": null, - "ParameterAlreadyExists$message": null, - "ParameterHistory$LastModifiedUser": "

Amazon Resource Name (ARN) of the AWS user who last changed the parameter.

", - "ParameterLimitExceeded$message": null, - "ParameterMaxVersionLimitExceeded$message": null, - "ParameterMetadata$LastModifiedUser": "

Amazon Resource Name (ARN) of the AWS user who last changed the parameter.

", - "ParameterNotFound$message": null, - "ParameterPatternMismatchException$message": "

The parameter name is not valid.

", - "ParameterVersionNotFound$message": null, - "ResourceDataSyncCountExceededException$Message": null, - "ResourceDataSyncInvalidConfigurationException$Message": null, - "ResourceInUseException$Message": null, - "ResourceLimitExceededException$Message": null, - "StepExecution$StepName": "

The name of this execution step.

", - "StepExecution$OnFailure": "

The action to take if the step fails. The default value is Abort.

", - "StepExecution$ResponseCode": "

The response code returned by the execution of the step.

", - "StepExecution$Response": "

A message associated with the response code for an execution.

", - "StepExecution$FailureMessage": "

If a step failed, this message explains why the execution failed.

", - "StepExecution$StepExecutionId": "

The unique ID of a step execution.

", - "StringList$member": null, - "SubTypeCountLimitExceededException$Message": null, - "TargetInUseException$Message": null, - "TooManyUpdates$Message": null, - "TotalSizeLimitExceededException$Message": null, - "UnsupportedInventoryItemContextException$Message": null, - "UnsupportedInventorySchemaVersionException$Message": null, - "UnsupportedOperatingSystem$Message": null, - "UnsupportedParameterType$message": null, - "UnsupportedPlatformType$Message": null - } - }, - "StringDateTime": { - "base": null, - "refs": { - "GetCommandInvocationResult$ExecutionStartDateTime": "

The date and time the plugin started executing. Date and time are written in ISO 8601 format. For example, June 7, 2017 is represented as 2017-06-7. The following sample AWS CLI command uses the InvokedBefore filter.

aws ssm list-commands --filters key=InvokedBefore,value=2017-06-07T00:00:00Z

If the plugin has not started to execute, the string is empty.

", - "GetCommandInvocationResult$ExecutionElapsedTime": "

Duration since ExecutionStartDateTime.

", - "GetCommandInvocationResult$ExecutionEndDateTime": "

The date and time the plugin was finished executing. Date and time are written in ISO 8601 format. For example, June 7, 2017 is represented as 2017-06-7. The following sample AWS CLI command uses the InvokedAfter filter.

aws ssm list-commands --filters key=InvokedAfter,value=2017-06-07T00:00:00Z

If the plugin has not started to execute, the string is empty.

" - } - }, - "StringList": { - "base": null, - "refs": { - "DescribeActivationsFilter$FilterValues": "

The filter values.

" - } - }, - "SubTypeCountLimitExceededException": { - "base": "

The sub-type count exceeded the limit for the inventory type.

", - "refs": { - } - }, - "Tag": { - "base": "

Metadata that you assign to your AWS resources. Tags enable you to categorize your resources in different ways, for example, by purpose, owner, or environment. In Systems Manager, you can apply tags to documents, managed instances, Maintenance Windows, Parameter Store parameters, and patch baselines.

", - "refs": { - "TagList$member": null - } - }, - "TagKey": { - "base": null, - "refs": { - "KeyList$member": null, - "Tag$Key": "

The name of the tag.

" - } - }, - "TagList": { - "base": null, - "refs": { - "AddTagsToResourceRequest$Tags": "

One or more tags. The value parameter is required, but if you don't want the tag to have a value, specify the parameter with no value, and we set the value to an empty string.

Do not enter personally identifiable information in this field.

", - "DocumentDescription$Tags": "

The tags, or metadata, that have been applied to the document.

", - "DocumentIdentifier$Tags": "

The tags, or metadata, that have been applied to the document.

", - "ListTagsForResourceResult$TagList": "

A list of tags.

" - } - }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": "

The value of the tag.

" - } - }, - "Target": { - "base": "

An array of search criteria that targets instances using a Key,Value combination that you specify. Targets is required if you don't provide one or more instance IDs in the call.

", - "refs": { - "Targets$member": null - } - }, - "TargetCount": { - "base": null, - "refs": { - "Command$TargetCount": "

The number of targets for the command.

" - } - }, - "TargetInUseException": { - "base": "

You specified the Safe option for the DeregisterTargetFromMaintenanceWindow operation, but the target is still referenced in a task.

", - "refs": { - } - }, - "TargetKey": { - "base": null, - "refs": { - "Target$Key": "

User-defined criteria for sending commands that target instances that meet the criteria. Key can be tag:<Amazon EC2 tag> or InstanceIds. For more information about how to send commands that target instances using Key,Value parameters, see Executing a Command Using Systems Manager Run Command.

" + "StatusUnchanged":{ + "base":"

The updated status is the same as the current status.

", + "refs":{ + } + }, + "StepExecution":{ + "base":"

Detailed information about an the execution state of an Automation step.

", + "refs":{ + "StepExecutionList$member":null + } + }, + "StepExecutionFilter":{ + "base":"

A filter to limit the amount of step execution information returned by the call.

", + "refs":{ + "StepExecutionFilterList$member":null + } + }, + "StepExecutionFilterKey":{ + "base":null, + "refs":{ + "StepExecutionFilter$Key":"

One or more keys to limit the results. Valid filter keys include the following: StepName, Action, StepExecutionId, StepExecutionStatus, StartTimeBefore, StartTimeAfter.

" + } + }, + "StepExecutionFilterList":{ + "base":null, + "refs":{ + "DescribeAutomationStepExecutionsRequest$Filters":"

One or more filters to limit the number of step executions returned by the request.

" + } + }, + "StepExecutionFilterValue":{ + "base":null, + "refs":{ + "StepExecutionFilterValueList$member":null + } + }, + "StepExecutionFilterValueList":{ + "base":null, + "refs":{ + "StepExecutionFilter$Values":"

The values of the filter key.

" + } + }, + "StepExecutionList":{ + "base":null, + "refs":{ + "AutomationExecution$StepExecutions":"

A list of details about the current state of all steps that comprise an execution. An Automation document contains a list of steps that are executed in order.

", + "DescribeAutomationStepExecutionsResult$StepExecutions":"

A list of details about the current state of all steps that make up an execution.

" + } + }, + "StopAutomationExecutionRequest":{ + "base":null, + "refs":{ + } + }, + "StopAutomationExecutionResult":{ + "base":null, + "refs":{ + } + }, + "StopType":{ + "base":null, + "refs":{ + "StopAutomationExecutionRequest$Type":"

The stop request type. Valid types include the following: Cancel and Complete. The default type is Cancel.

" + } + }, + "String":{ + "base":null, + "refs":{ + "AlreadyExistsException$Message":null, + "AssociationDoesNotExist$Message":null, + "AssociationVersionLimitExceeded$Message":null, + "AutomationDefinitionNotFoundException$Message":null, + "AutomationDefinitionVersionNotFoundException$Message":null, + "AutomationExecution$FailureMessage":"

A message describing why an execution has failed, if the status is set to Failed.

", + "AutomationExecution$ExecutedBy":"

The Amazon Resource Name (ARN) of the user who executed the automation.

", + "AutomationExecution$CurrentStepName":"

The name of the currently executing step.

", + "AutomationExecution$CurrentAction":"

The action of the currently executing step.

", + "AutomationExecution$Target":"

The target of the execution.

", + "AutomationExecutionLimitExceededException$Message":null, + "AutomationExecutionMetadata$ExecutedBy":"

The IAM role ARN of the user who executed the Automation.

", + "AutomationExecutionMetadata$LogFile":"

An Amazon S3 bucket where execution information is stored.

", + "AutomationExecutionMetadata$CurrentStepName":"

The name of the currently executing step.

", + "AutomationExecutionMetadata$CurrentAction":"

The action of the currently executing step.

", + "AutomationExecutionMetadata$FailureMessage":"

The list of execution outputs as defined in the Automation document.

", + "AutomationExecutionMetadata$Target":"

The list of execution outputs as defined in the Automation document.

", + "AutomationExecutionNotFoundException$Message":null, + "AutomationStepNotFoundException$Message":null, + "ComplianceTypeCountLimitExceededException$Message":null, + "CustomSchemaCountLimitExceededException$Message":null, + "DocumentAlreadyExists$Message":null, + "DocumentLimitExceeded$Message":null, + "DocumentPermissionLimit$Message":null, + "DocumentVersionLimitExceeded$Message":null, + "DoesNotExistException$Message":null, + "DuplicateDocumentContent$Message":null, + "FailureDetails$FailureStage":"

The stage of the Automation execution when the failure occurred. The stages include the following: InputValidation, PreVerification, Invocation, PostVerification.

", + "FailureDetails$FailureType":"

The type of Automation failure. Failure types include the following: Action, Permission, Throttling, Verification, Internal.

", + "FeatureNotAvailableException$Message":null, + "HierarchyLevelLimitExceededException$message":"

A hierarchy can have a maximum of 15 levels. For more information, see Working with Systems Manager Parameters in the AWS Systems Manager User Guide.

", + "HierarchyTypeMismatchException$message":"

Parameter Store does not support changing a parameter type in a hierarchy. For example, you can't change a parameter from a String type to a SecureString type. You must create a new, unique parameter.

", + "IdempotentParameterMismatch$Message":null, + "InstanceInformation$PlatformName":"

The name of the operating system platform running on your instance.

", + "InstanceInformation$PlatformVersion":"

The version of the OS platform running on your instance.

", + "InstanceInformation$Name":"

The name of the managed instance.

", + "InternalServerError$Message":null, + "InvalidActivation$Message":null, + "InvalidActivationId$Message":null, + "InvalidAllowedPatternException$message":"

The request does not meet the regular expression requirement.

", + "InvalidAssociationVersion$Message":null, + "InvalidAutomationExecutionParametersException$Message":null, + "InvalidAutomationSignalException$Message":null, + "InvalidAutomationStatusUpdateException$Message":null, + "InvalidDeleteInventoryParametersException$Message":null, + "InvalidDeletionIdException$Message":null, + "InvalidDocument$Message":"

The document does not exist or the document is not available to the user. This exception can be issued by CreateAssociation, CreateAssociationBatch, DeleteAssociation, DeleteDocument, DescribeAssociation, DescribeDocument, GetDocument, SendCommand, or UpdateAssociationStatus.

", + "InvalidDocumentContent$Message":"

A description of the validation error.

", + "InvalidDocumentOperation$Message":null, + "InvalidDocumentSchemaVersion$Message":null, + "InvalidDocumentVersion$Message":null, + "InvalidFilter$Message":null, + "InvalidFilterOption$message":"

The specified filter option is not valid. Valid options are Equals and BeginsWith. For Path filter, valid options are Recursive and OneLevel.

", + "InvalidFilterValue$Message":null, + "InvalidInstanceId$Message":null, + "InvalidInstanceInformationFilterValue$message":null, + "InvalidInventoryItemContextException$Message":null, + "InvalidInventoryRequestException$Message":null, + "InvalidItemContentException$Message":null, + "InvalidKeyId$message":null, + "InvalidNextToken$Message":null, + "InvalidNotificationConfig$Message":null, + "InvalidOptionException$Message":null, + "InvalidParameters$Message":null, + "InvalidPermissionType$Message":null, + "InvalidResultAttributeException$Message":null, + "InvalidRole$Message":null, + "InvalidSchedule$Message":null, + "InvalidTarget$Message":null, + "InvalidTypeNameException$Message":null, + "InvalidUpdate$Message":null, + "ItemContentMismatchException$Message":null, + "ItemSizeLimitExceededException$Message":null, + "MaxDocumentSizeExceeded$Message":null, + "NormalStringMap$key":null, + "NormalStringMap$value":null, + "ParameterAlreadyExists$message":null, + "ParameterHistory$LastModifiedUser":"

Amazon Resource Name (ARN) of the AWS user who last changed the parameter.

", + "ParameterLimitExceeded$message":null, + "ParameterMaxVersionLimitExceeded$message":null, + "ParameterMetadata$LastModifiedUser":"

Amazon Resource Name (ARN) of the AWS user who last changed the parameter.

", + "ParameterNotFound$message":null, + "ParameterPatternMismatchException$message":"

The parameter name is not valid.

", + "ParameterVersionNotFound$message":null, + "ResourceDataSyncCountExceededException$Message":null, + "ResourceDataSyncInvalidConfigurationException$Message":null, + "ResourceInUseException$Message":null, + "ResourceLimitExceededException$Message":null, + "StepExecution$StepName":"

The name of this execution step.

", + "StepExecution$OnFailure":"

The action to take if the step fails. The default value is Abort.

", + "StepExecution$ResponseCode":"

The response code returned by the execution of the step.

", + "StepExecution$Response":"

A message associated with the response code for an execution.

", + "StepExecution$FailureMessage":"

If a step failed, this message explains why the execution failed.

", + "StepExecution$StepExecutionId":"

The unique ID of a step execution.

", + "StringList$member":null, + "SubTypeCountLimitExceededException$Message":null, + "TargetInUseException$Message":null, + "TooManyUpdates$Message":null, + "TotalSizeLimitExceededException$Message":null, + "UnsupportedInventoryItemContextException$Message":null, + "UnsupportedInventorySchemaVersionException$Message":null, + "UnsupportedOperatingSystem$Message":null, + "UnsupportedParameterType$message":null, + "UnsupportedPlatformType$Message":null + } + }, + "StringDateTime":{ + "base":null, + "refs":{ + "GetCommandInvocationResult$ExecutionStartDateTime":"

The date and time the plugin started executing. Date and time are written in ISO 8601 format. For example, June 7, 2017 is represented as 2017-06-7. The following sample AWS CLI command uses the InvokedBefore filter.

aws ssm list-commands --filters key=InvokedBefore,value=2017-06-07T00:00:00Z

If the plugin has not started to execute, the string is empty.

", + "GetCommandInvocationResult$ExecutionElapsedTime":"

Duration since ExecutionStartDateTime.

", + "GetCommandInvocationResult$ExecutionEndDateTime":"

The date and time the plugin was finished executing. Date and time are written in ISO 8601 format. For example, June 7, 2017 is represented as 2017-06-7. The following sample AWS CLI command uses the InvokedAfter filter.

aws ssm list-commands --filters key=InvokedAfter,value=2017-06-07T00:00:00Z

If the plugin has not started to execute, the string is empty.

" + } + }, + "StringList":{ + "base":null, + "refs":{ + "DescribeActivationsFilter$FilterValues":"

The filter values.

" + } + }, + "SubTypeCountLimitExceededException":{ + "base":"

The sub-type count exceeded the limit for the inventory type.

", + "refs":{ + } + }, + "Tag":{ + "base":"

Metadata that you assign to your AWS resources. Tags enable you to categorize your resources in different ways, for example, by purpose, owner, or environment. In Systems Manager, you can apply tags to documents, managed instances, Maintenance Windows, Parameter Store parameters, and patch baselines.

", + "refs":{ + "TagList$member":null + } + }, + "TagKey":{ + "base":null, + "refs":{ + "KeyList$member":null, + "Tag$Key":"

The name of the tag.

" + } + }, + "TagList":{ + "base":null, + "refs":{ + "AddTagsToResourceRequest$Tags":"

One or more tags. The value parameter is required, but if you don't want the tag to have a value, specify the parameter with no value, and we set the value to an empty string.

Do not enter personally identifiable information in this field.

", + "DocumentDescription$Tags":"

The tags, or metadata, that have been applied to the document.

", + "DocumentIdentifier$Tags":"

The tags, or metadata, that have been applied to the document.

", + "ListTagsForResourceResult$TagList":"

A list of tags.

" + } + }, + "TagValue":{ + "base":null, + "refs":{ + "Tag$Value":"

The value of the tag.

" + } + }, + "Target":{ + "base":"

An array of search criteria that targets instances using a Key,Value combination that you specify. Targets is required if you don't provide one or more instance IDs in the call.

", + "refs":{ + "Targets$member":null + } + }, + "TargetCount":{ + "base":null, + "refs":{ + "Command$TargetCount":"

The number of targets for the command.

" + } + }, + "TargetInUseException":{ + "base":"

You specified the Safe option for the DeregisterTargetFromMaintenanceWindow operation, but the target is still referenced in a task.

", + "refs":{ + } + }, + "TargetKey":{ + "base":null, + "refs":{ + "Target$Key":"

User-defined criteria for sending commands that target instances that meet the criteria. Key can be tag:<Amazon EC2 tag> or InstanceIds. For more information about how to send commands that target instances using Key,Value parameters, see Targeting Multiple Instances in the AWS Systems Manager User Guide.

" } }, - "TargetParameterList": { - "base": null, - "refs": { - "ResolvedTargets$ParameterValues": "

A list of parameter values sent to targets that resolved during the Automation execution.

" + "TargetParameterList":{ + "base":null, + "refs":{ + "ResolvedTargets$ParameterValues":"

A list of parameter values sent to targets that resolved during the Automation execution.

" } }, - "TargetType": { - "base": null, - "refs": { - "CreateDocumentRequest$TargetType": "

Specify a target type to define the kinds of resources the document can run on. For example, to run a document on EC2 instances, specify the following value: /AWS::EC2::Instance. If you specify a value of '/' the document can run on all types of resources. If you don't specify a value, the document can't run on any resources. For a list of valid resource types, see AWS Resource Types Reference in the AWS CloudFormation User Guide.

", - "DocumentDescription$TargetType": "

The target type which defines the kinds of resources the document can run on. For example, /AWS::EC2::Instance. For a list of valid resource types, see AWS Resource Types Reference in the AWS CloudFormation User Guide.

", - "DocumentIdentifier$TargetType": "

The target type which defines the kinds of resources the document can run on. For example, /AWS::EC2::Instance. For a list of valid resource types, see AWS Resource Types Reference in the AWS CloudFormation User Guide.

", - "UpdateDocumentRequest$TargetType": "

Specify a new target type for the document.

" + "TargetType":{ + "base":null, + "refs":{ + "CreateDocumentRequest$TargetType":"

Specify a target type to define the kinds of resources the document can run on. For example, to run a document on EC2 instances, specify the following value: /AWS::EC2::Instance. If you specify a value of '/' the document can run on all types of resources. If you don't specify a value, the document can't run on any resources. For a list of valid resource types, see AWS Resource Types Reference in the AWS CloudFormation User Guide.

", + "DocumentDescription$TargetType":"

The target type which defines the kinds of resources the document can run on. For example, /AWS::EC2::Instance. For a list of valid resource types, see AWS Resource Types Reference in the AWS CloudFormation User Guide.

", + "DocumentIdentifier$TargetType":"

The target type which defines the kinds of resources the document can run on. For example, /AWS::EC2::Instance. For a list of valid resource types, see AWS Resource Types Reference in the AWS CloudFormation User Guide.

", + "UpdateDocumentRequest$TargetType":"

Specify a new target type for the document.

" } }, - "TargetValue": { - "base": null, - "refs": { - "TargetValues$member": null + "TargetValue":{ + "base":null, + "refs":{ + "TargetValues$member":null } }, - "TargetValues": { - "base": null, - "refs": { - "Target$Values": "

User-defined criteria that maps to Key. For example, if you specified tag:ServerRole, you could specify value:WebServer to execute a command on instances that include Amazon EC2 tags of ServerRole,WebServer. For more information about how to send commands that target instances using Key,Value parameters, see Executing a Command Using Systems Manager Run Command.

" + "TargetValues":{ + "base":null, + "refs":{ + "Target$Values":"

User-defined criteria that maps to Key. For example, if you specified tag:ServerRole, you could specify value:WebServer to execute a command on instances that include Amazon EC2 tags of ServerRole,WebServer. For more information about how to send commands that target instances using Key,Value parameters, see Sending Commands to a Fleet in the AWS Systems Manager User Guide.

" } }, - "Targets": { - "base": null, - "refs": { - "Association$Targets": "

The instances targeted by the request to create an association.

", - "AssociationDescription$Targets": "

The instances targeted by the request.

", - "AssociationVersionInfo$Targets": "

The targets specified for the association when the association version was created.

", - "AutomationExecution$Targets": "

The specified targets.

", - "AutomationExecutionMetadata$Targets": "

The targets defined by the user when starting the Automation.

", - "Command$Targets": "

An array of search criteria that targets instances using a Key,Value combination that you specify. Targets is required if you don't provide one or more instance IDs in the call.

", - "CreateAssociationBatchRequestEntry$Targets": "

The instances targeted by the request.

", - "CreateAssociationRequest$Targets": "

The targets (either instances or tags) for the association.

", - "GetMaintenanceWindowTaskResult$Targets": "

The targets where the task should execute.

", - "MaintenanceWindowTarget$Targets": "

The targets (either instances or tags). Instances are specified using Key=instanceids,Values=<instanceid1>,<instanceid2>. Tags are specified using Key=<tag name>,Values=<tag value>.

", - "MaintenanceWindowTask$Targets": "

The targets (either instances or tags). Instances are specified using Key=instanceids,Values=<instanceid1>,<instanceid2>. Tags are specified using Key=<tag name>,Values=<tag value>.

", - "RegisterTargetWithMaintenanceWindowRequest$Targets": "

The targets (either instances or tags).

Specify instances using the following format:

Key=InstanceIds,Values=<instance-id-1>,<instance-id-2>

Specify tags using either of the following formats:

Key=tag:<tag-key>,Values=<tag-value-1>,<tag-value-2>

Key=tag-key,Values=<tag-key-1>,<tag-key-2>

", - "RegisterTaskWithMaintenanceWindowRequest$Targets": "

The targets (either instances or Maintenance Window targets).

Specify instances using the following format:

Key=InstanceIds,Values=<instance-id-1>,<instance-id-2>

Specify Maintenance Window targets using the following format:

Key=<WindowTargetIds>,Values=<window-target-id-1>,<window-target-id-2>

", - "SendCommandRequest$Targets": "

(Optional) An array of search criteria that targets instances using a Key,Value combination that you specify. Targets is required if you don't provide one or more instance IDs in the call. For more information about how to use Targets, see Sending Commands to a Fleet.

", - "StartAutomationExecutionRequest$Targets": "

A key-value mapping to target resources. Required if you specify TargetParameterName.

", - "UpdateAssociationRequest$Targets": "

The targets of the association.

", - "UpdateMaintenanceWindowTargetRequest$Targets": "

The targets to add or replace.

", - "UpdateMaintenanceWindowTargetResult$Targets": "

The updated targets.

", - "UpdateMaintenanceWindowTaskRequest$Targets": "

The targets (either instances or tags) to modify. Instances are specified using Key=instanceids,Values=instanceID_1,instanceID_2. Tags are specified using Key=tag_name,Values=tag_value.

", - "UpdateMaintenanceWindowTaskResult$Targets": "

The updated target values.

" + "Targets":{ + "base":null, + "refs":{ + "Association$Targets":"

The instances targeted by the request to create an association.

", + "AssociationDescription$Targets":"

The instances targeted by the request.

", + "AssociationVersionInfo$Targets":"

The targets specified for the association when the association version was created.

", + "AutomationExecution$Targets":"

The specified targets.

", + "AutomationExecutionMetadata$Targets":"

The targets defined by the user when starting the Automation.

", + "Command$Targets":"

An array of search criteria that targets instances using a Key,Value combination that you specify. Targets is required if you don't provide one or more instance IDs in the call.

", + "CreateAssociationBatchRequestEntry$Targets":"

The instances targeted by the request.

", + "CreateAssociationRequest$Targets":"

The targets (either instances or tags) for the association.

", + "GetMaintenanceWindowTaskResult$Targets":"

The targets where the task should execute.

", + "MaintenanceWindowTarget$Targets":"

The targets (either instances or tags). Instances are specified using Key=instanceids,Values=<instanceid1>,<instanceid2>. Tags are specified using Key=<tag name>,Values=<tag value>.

", + "MaintenanceWindowTask$Targets":"

The targets (either instances or tags). Instances are specified using Key=instanceids,Values=<instanceid1>,<instanceid2>. Tags are specified using Key=<tag name>,Values=<tag value>.

", + "RegisterTargetWithMaintenanceWindowRequest$Targets":"

The targets (either instances or tags).

Specify instances using the following format:

Key=InstanceIds,Values=<instance-id-1>,<instance-id-2>

Specify tags using either of the following formats:

Key=tag:<tag-key>,Values=<tag-value-1>,<tag-value-2>

Key=tag-key,Values=<tag-key-1>,<tag-key-2>

", + "RegisterTaskWithMaintenanceWindowRequest$Targets":"

The targets (either instances or Maintenance Window targets).

Specify instances using the following format:

Key=InstanceIds,Values=<instance-id-1>,<instance-id-2>

Specify Maintenance Window targets using the following format:

Key=<WindowTargetIds>,Values=<window-target-id-1>,<window-target-id-2>

", + "SendCommandRequest$Targets":"

(Optional) An array of search criteria that targets instances using a Key,Value combination that you specify. Targets is required if you don't provide one or more instance IDs in the call. For more information about how to use Targets, see Sending Commands to a Fleet in the AWS Systems Manager User Guide.

", + "StartAutomationExecutionRequest$Targets":"

A key-value mapping to target resources. Required if you specify TargetParameterName.

", + "UpdateAssociationRequest$Targets":"

The targets of the association.

", + "UpdateMaintenanceWindowTargetRequest$Targets":"

The targets to add or replace.

", + "UpdateMaintenanceWindowTargetResult$Targets":"

The updated targets.

", + "UpdateMaintenanceWindowTaskRequest$Targets":"

The targets (either instances or tags) to modify. Instances are specified using Key=instanceids,Values=instanceID_1,instanceID_2. Tags are specified using Key=tag_name,Values=tag_value.

", + "UpdateMaintenanceWindowTaskResult$Targets":"

The updated target values.

" } }, - "TimeoutSeconds": { - "base": null, - "refs": { - "MaintenanceWindowRunCommandParameters$TimeoutSeconds": "

If this time is reached and the command has not already started executing, it doesn not execute.

", - "SendCommandRequest$TimeoutSeconds": "

If this time is reached and the command has not already started executing, it will not run.

" + "TimeoutSeconds":{ + "base":null, + "refs":{ + "MaintenanceWindowRunCommandParameters$TimeoutSeconds":"

If this time is reached and the command has not already started executing, it doesn not execute.

", + "SendCommandRequest$TimeoutSeconds":"

If this time is reached and the command has not already started executing, it will not run.

" } }, - "TooManyTagsError": { - "base": "

The Targets parameter includes too many tags. Remove one or more tags and try the command again.

", - "refs": { + "TooManyTagsError":{ + "base":"

The Targets parameter includes too many tags. Remove one or more tags and try the command again.

", + "refs":{ } }, - "TooManyUpdates": { - "base": "

There are concurrent updates for a resource that supports one update at a time.

", - "refs": { + "TooManyUpdates":{ + "base":"

There are concurrent updates for a resource that supports one update at a time.

", + "refs":{ } }, - "TotalCount": { - "base": null, - "refs": { - "InventoryDeletionSummary$TotalCount": "

The total number of items to delete. This count does not change during the delete operation.

" + "TotalCount":{ + "base":null, + "refs":{ + "InventoryDeletionSummary$TotalCount":"

The total number of items to delete. This count does not change during the delete operation.

" } }, - "TotalSizeLimitExceededException": { - "base": "

The size of inventory data has exceeded the total size limit for the resource.

", - "refs": { + "TotalSizeLimitExceededException":{ + "base":"

The size of inventory data has exceeded the total size limit for the resource.

", + "refs":{ } }, - "UnsupportedInventoryItemContextException": { - "base": "

The Context attribute that you specified for the InventoryItem is not allowed for this inventory type. You can only use the Context attribute with inventory types like AWS:ComplianceItem.

", - "refs": { + "UnsupportedInventoryItemContextException":{ + "base":"

The Context attribute that you specified for the InventoryItem is not allowed for this inventory type. You can only use the Context attribute with inventory types like AWS:ComplianceItem.

", + "refs":{ } }, - "UnsupportedInventorySchemaVersionException": { - "base": "

Inventory item type schema version has to match supported versions in the service. Check output of GetInventorySchema to see the available schema version for each type.

", - "refs": { + "UnsupportedInventorySchemaVersionException":{ + "base":"

Inventory item type schema version has to match supported versions in the service. Check output of GetInventorySchema to see the available schema version for each type.

", + "refs":{ } }, - "UnsupportedOperatingSystem": { - "base": "

The operating systems you specified is not supported, or the operation is not supported for the operating system. Valid operating systems include: Windows, AmazonLinux, RedhatEnterpriseLinux, and Ubuntu.

", - "refs": { + "UnsupportedOperatingSystem":{ + "base":"

The operating systems you specified is not supported, or the operation is not supported for the operating system. Valid operating systems include: Windows, AmazonLinux, RedhatEnterpriseLinux, and Ubuntu.

", + "refs":{ } }, - "UnsupportedParameterType": { - "base": "

The parameter type is not supported.

", - "refs": { + "UnsupportedParameterType":{ + "base":"

The parameter type is not supported.

", + "refs":{ } }, - "UnsupportedPlatformType": { - "base": "

The document does not support the platform type of the given instance ID(s). For example, you sent an document for a Windows instance to a Linux instance.

", - "refs": { + "UnsupportedPlatformType":{ + "base":"

The document does not support the platform type of the given instance ID(s). For example, you sent an document for a Windows instance to a Linux instance.

", + "refs":{ } }, - "UpdateAssociationRequest": { - "base": null, - "refs": { + "UpdateAssociationRequest":{ + "base":null, + "refs":{ } }, - "UpdateAssociationResult": { - "base": null, - "refs": { + "UpdateAssociationResult":{ + "base":null, + "refs":{ } }, - "UpdateAssociationStatusRequest": { - "base": null, - "refs": { + "UpdateAssociationStatusRequest":{ + "base":null, + "refs":{ } }, - "UpdateAssociationStatusResult": { - "base": null, - "refs": { + "UpdateAssociationStatusResult":{ + "base":null, + "refs":{ } }, - "UpdateDocumentDefaultVersionRequest": { - "base": null, - "refs": { + "UpdateDocumentDefaultVersionRequest":{ + "base":null, + "refs":{ } }, - "UpdateDocumentDefaultVersionResult": { - "base": null, - "refs": { + "UpdateDocumentDefaultVersionResult":{ + "base":null, + "refs":{ } }, - "UpdateDocumentRequest": { - "base": null, - "refs": { + "UpdateDocumentRequest":{ + "base":null, + "refs":{ } }, - "UpdateDocumentResult": { - "base": null, - "refs": { + "UpdateDocumentResult":{ + "base":null, + "refs":{ } }, - "UpdateMaintenanceWindowRequest": { - "base": null, - "refs": { + "UpdateMaintenanceWindowRequest":{ + "base":null, + "refs":{ } }, - "UpdateMaintenanceWindowResult": { - "base": null, - "refs": { + "UpdateMaintenanceWindowResult":{ + "base":null, + "refs":{ } }, - "UpdateMaintenanceWindowTargetRequest": { - "base": null, - "refs": { + "UpdateMaintenanceWindowTargetRequest":{ + "base":null, + "refs":{ } }, - "UpdateMaintenanceWindowTargetResult": { - "base": null, - "refs": { + "UpdateMaintenanceWindowTargetResult":{ + "base":null, + "refs":{ } }, - "UpdateMaintenanceWindowTaskRequest": { - "base": null, - "refs": { + "UpdateMaintenanceWindowTaskRequest":{ + "base":null, + "refs":{ } }, - "UpdateMaintenanceWindowTaskResult": { - "base": null, - "refs": { + "UpdateMaintenanceWindowTaskResult":{ + "base":null, + "refs":{ } }, - "UpdateManagedInstanceRoleRequest": { - "base": null, - "refs": { + "UpdateManagedInstanceRoleRequest":{ + "base":null, + "refs":{ } }, - "UpdateManagedInstanceRoleResult": { - "base": null, - "refs": { + "UpdateManagedInstanceRoleResult":{ + "base":null, + "refs":{ } }, - "UpdatePatchBaselineRequest": { - "base": null, - "refs": { + "UpdatePatchBaselineRequest":{ + "base":null, + "refs":{ } }, - "UpdatePatchBaselineResult": { - "base": null, - "refs": { + "UpdatePatchBaselineResult":{ + "base":null, + "refs":{ } }, - "Url": { - "base": null, - "refs": { - "CommandInvocation$StandardOutputUrl": "

The URL to the plugin's StdOut file in Amazon S3, if the Amazon S3 bucket was defined for the parent command. For an invocation, StandardOutputUrl is populated if there is just one plugin defined for the command, and the Amazon S3 bucket was defined for the command.

", - "CommandInvocation$StandardErrorUrl": "

The URL to the plugin's StdErr file in Amazon S3, if the Amazon S3 bucket was defined for the parent command. For an invocation, StandardErrorUrl is populated if there is just one plugin defined for the command, and the Amazon S3 bucket was defined for the command.

", - "CommandPlugin$StandardOutputUrl": "

The URL for the complete text written by the plugin to stdout in Amazon S3. If the Amazon S3 bucket for the command was not specified, then this string is empty.

", - "CommandPlugin$StandardErrorUrl": "

The URL for the complete text written by the plugin to stderr. If execution is not yet complete, then this string is empty.

", - "GetCommandInvocationResult$StandardOutputUrl": "

The URL for the complete text written by the plugin to stdout in Amazon S3. If an Amazon S3 bucket was not specified, then this string is empty.

", - "GetCommandInvocationResult$StandardErrorUrl": "

The URL for the complete text written by the plugin to stderr. If the command has not finished executing, then this string is empty.

", - "S3OutputUrl$OutputUrl": "

A URL for an Amazon S3 bucket where you want to store the results of this request.

" + "Url":{ + "base":null, + "refs":{ + "CommandInvocation$StandardOutputUrl":"

The URL to the plugin's StdOut file in Amazon S3, if the Amazon S3 bucket was defined for the parent command. For an invocation, StandardOutputUrl is populated if there is just one plugin defined for the command, and the Amazon S3 bucket was defined for the command.

", + "CommandInvocation$StandardErrorUrl":"

The URL to the plugin's StdErr file in Amazon S3, if the Amazon S3 bucket was defined for the parent command. For an invocation, StandardErrorUrl is populated if there is just one plugin defined for the command, and the Amazon S3 bucket was defined for the command.

", + "CommandPlugin$StandardOutputUrl":"

The URL for the complete text written by the plugin to stdout in Amazon S3. If the Amazon S3 bucket for the command was not specified, then this string is empty.

", + "CommandPlugin$StandardErrorUrl":"

The URL for the complete text written by the plugin to stderr. If execution is not yet complete, then this string is empty.

", + "GetCommandInvocationResult$StandardOutputUrl":"

The URL for the complete text written by the plugin to stdout in Amazon S3. If an Amazon S3 bucket was not specified, then this string is empty.

", + "GetCommandInvocationResult$StandardErrorUrl":"

The URL for the complete text written by the plugin to stderr. If the command has not finished executing, then this string is empty.

", + "S3OutputUrl$OutputUrl":"

A URL for an Amazon S3 bucket where you want to store the results of this request.

" } }, - "Version": { - "base": null, - "refs": { - "InstanceInformation$AgentVersion": "

The version of the SSM Agent running on your Linux instance.

" + "Version":{ + "base":null, + "refs":{ + "InstanceInformation$AgentVersion":"

The version of SSM Agent running on your Linux instance.

" } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/examples-1.json index 0ea7e3b0b..faff76894 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/ssm/2014-11-06/examples-1.json @@ -1,5 +1,5 @@ { - "version": "1.0", - "examples": { + "version":"1.0", + "examples":{ } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/api-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/api-2.json index b5b4617d0..3d6957721 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/api-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/api-2.json @@ -129,6 +129,19 @@ {"shape":"InternalServerError"} ] }, + "CreateSMBFileShare":{ + "name":"CreateSMBFileShare", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"CreateSMBFileShareInput"}, + "output":{"shape":"CreateSMBFileShareOutput"}, + "errors":[ + {"shape":"InvalidGatewayRequestException"}, + {"shape":"InternalServerError"} + ] + }, "CreateSnapshot":{ "name":"CreateSnapshot", "http":{ @@ -391,6 +404,32 @@ {"shape":"InternalServerError"} ] }, + "DescribeSMBFileShares":{ + "name":"DescribeSMBFileShares", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeSMBFileSharesInput"}, + "output":{"shape":"DescribeSMBFileSharesOutput"}, + "errors":[ + {"shape":"InvalidGatewayRequestException"}, + {"shape":"InternalServerError"} + ] + }, + "DescribeSMBSettings":{ + "name":"DescribeSMBSettings", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"DescribeSMBSettingsInput"}, + "output":{"shape":"DescribeSMBSettingsOutput"}, + "errors":[ + {"shape":"InvalidGatewayRequestException"}, + {"shape":"InternalServerError"} + ] + }, "DescribeSnapshotSchedule":{ "name":"DescribeSnapshotSchedule", "http":{ @@ -508,6 +547,19 @@ {"shape":"InternalServerError"} ] }, + "JoinDomain":{ + "name":"JoinDomain", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"JoinDomainInput"}, + "output":{"shape":"JoinDomainOutput"}, + "errors":[ + {"shape":"InvalidGatewayRequestException"}, + {"shape":"InternalServerError"} + ] + }, "ListFileShares":{ "name":"ListFileShares", "http":{ @@ -703,6 +755,19 @@ {"shape":"InternalServerError"} ] }, + "SetSMBGuestPassword":{ + "name":"SetSMBGuestPassword", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"SetSMBGuestPasswordInput"}, + "output":{"shape":"SetSMBGuestPasswordOutput"}, + "errors":[ + {"shape":"InvalidGatewayRequestException"}, + {"shape":"InternalServerError"} + ] + }, "ShutdownGateway":{ "name":"ShutdownGateway", "http":{ @@ -807,6 +872,19 @@ {"shape":"InternalServerError"} ] }, + "UpdateSMBFileShare":{ + "name":"UpdateSMBFileShare", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"UpdateSMBFileShareInput"}, + "output":{"shape":"UpdateSMBFileShareOutput"}, + "errors":[ + {"shape":"InvalidGatewayRequestException"}, + {"shape":"InternalServerError"} + ] + }, "UpdateSnapshotSchedule":{ "name":"UpdateSnapshotSchedule", "http":{ @@ -932,6 +1010,11 @@ "GatewayARN":{"shape":"GatewayARN"} } }, + "Authentication":{ + "type":"string", + "max":15, + "min":5 + }, "BandwidthDownloadRateLimit":{ "type":"long", "min":102400 @@ -958,7 +1041,8 @@ "SourceSnapshotId":{"shape":"SnapshotId"}, "VolumeiSCSIAttributes":{"shape":"VolumeiSCSIAttributes"}, "CreatedDate":{"shape":"CreatedDate"}, - "VolumeUsedInBytes":{"shape":"VolumeUsedInBytes"} + "VolumeUsedInBytes":{"shape":"VolumeUsedInBytes"}, + "KMSKey":{"shape":"KMSKey"} } }, "CachediSCSIVolumes":{ @@ -1038,7 +1122,9 @@ "TargetName":{"shape":"TargetName"}, "SourceVolumeARN":{"shape":"VolumeARN"}, "NetworkInterfaceId":{"shape":"NetworkInterfaceId"}, - "ClientToken":{"shape":"ClientToken"} + "ClientToken":{"shape":"ClientToken"}, + "KMSEncrypted":{"shape":"Boolean"}, + "KMSKey":{"shape":"KMSKey"} } }, "CreateCachediSCSIVolumeOutput":{ @@ -1079,6 +1165,37 @@ "FileShareARN":{"shape":"FileShareARN"} } }, + "CreateSMBFileShareInput":{ + "type":"structure", + "required":[ + "ClientToken", + "GatewayARN", + "Role", + "LocationARN" + ], + "members":{ + "ClientToken":{"shape":"ClientToken"}, + "GatewayARN":{"shape":"GatewayARN"}, + "KMSEncrypted":{"shape":"Boolean"}, + "KMSKey":{"shape":"KMSKey"}, + "Role":{"shape":"Role"}, + "LocationARN":{"shape":"LocationARN"}, + "DefaultStorageClass":{"shape":"StorageClass"}, + "ObjectACL":{"shape":"ObjectACL"}, + "ReadOnly":{"shape":"Boolean"}, + "GuessMIMETypeEnabled":{"shape":"Boolean"}, + "RequesterPays":{"shape":"Boolean"}, + "ValidUserList":{"shape":"FileShareUserList"}, + "InvalidUserList":{"shape":"FileShareUserList"}, + "Authentication":{"shape":"Authentication"} + } + }, + "CreateSMBFileShareOutput":{ + "type":"structure", + "members":{ + "FileShareARN":{"shape":"FileShareARN"} + } + }, "CreateSnapshotFromVolumeRecoveryPointInput":{ "type":"structure", "required":[ @@ -1152,7 +1269,9 @@ "members":{ "GatewayARN":{"shape":"GatewayARN"}, "TapeSizeInBytes":{"shape":"TapeSize"}, - "TapeBarcode":{"shape":"TapeBarcode"} + "TapeBarcode":{"shape":"TapeBarcode"}, + "KMSEncrypted":{"shape":"Boolean"}, + "KMSKey":{"shape":"KMSKey"} } }, "CreateTapeWithBarcodeOutput":{ @@ -1175,7 +1294,9 @@ "TapeSizeInBytes":{"shape":"TapeSize"}, "ClientToken":{"shape":"ClientToken"}, "NumTapesToCreate":{"shape":"NumTapesToCreate"}, - "TapeBarcodePrefix":{"shape":"TapeBarcodePrefix"} + "TapeBarcodePrefix":{"shape":"TapeBarcodePrefix"}, + "KMSEncrypted":{"shape":"Boolean"}, + "KMSKey":{"shape":"KMSKey"} } }, "CreateTapesOutput":{ @@ -1419,6 +1540,34 @@ "NFSFileShareInfoList":{"shape":"NFSFileShareInfoList"} } }, + "DescribeSMBFileSharesInput":{ + "type":"structure", + "required":["FileShareARNList"], + "members":{ + "FileShareARNList":{"shape":"FileShareARNList"} + } + }, + "DescribeSMBFileSharesOutput":{ + "type":"structure", + "members":{ + "SMBFileShareInfoList":{"shape":"SMBFileShareInfoList"} + } + }, + "DescribeSMBSettingsInput":{ + "type":"structure", + "required":["GatewayARN"], + "members":{ + "GatewayARN":{"shape":"GatewayARN"} + } + }, + "DescribeSMBSettingsOutput":{ + "type":"structure", + "members":{ + "GatewayARN":{"shape":"GatewayARN"}, + "DomainName":{"shape":"DomainName"}, + "SMBGuestPasswordSet":{"shape":"Boolean"} + } + }, "DescribeSnapshotScheduleInput":{ "type":"structure", "required":["VolumeARN"], @@ -1610,6 +1759,19 @@ "type":"list", "member":{"shape":"Disk"} }, + "DomainName":{ + "type":"string", + "pattern":"^([a-z0-9]+(-[a-z0-9]+)*\\.)+[a-z]{2,}$" + }, + "DomainUserName":{ + "type":"string", + "pattern":"^\\w[\\w\\.\\- ]*$" + }, + "DomainUserPassword":{ + "type":"string", + "pattern":"^[ -~]+$", + "sensitive":true + }, "DoubleObject":{"type":"double"}, "ErrorCode":{ "type":"string", @@ -1702,6 +1864,7 @@ "FileShareInfo":{ "type":"structure", "members":{ + "FileShareType":{"shape":"FileShareType"}, "FileShareARN":{"shape":"FileShareARN"}, "FileShareId":{"shape":"FileShareId"}, "FileShareStatus":{"shape":"FileShareStatus"}, @@ -1717,6 +1880,24 @@ "max":50, "min":3 }, + "FileShareType":{ + "type":"string", + "enum":[ + "NFS", + "SMB" + ] + }, + "FileShareUser":{ + "type":"string", + "max":64, + "min":1 + }, + "FileShareUserList":{ + "type":"list", + "member":{"shape":"FileShareUser"}, + "max":100, + "min":0 + }, "GatewayARN":{ "type":"string", "max":500, @@ -1811,6 +1992,27 @@ "min":1, "pattern":"[0-9a-z:.-]+" }, + "JoinDomainInput":{ + "type":"structure", + "required":[ + "GatewayARN", + "DomainName", + "UserName", + "Password" + ], + "members":{ + "GatewayARN":{"shape":"GatewayARN"}, + "DomainName":{"shape":"DomainName"}, + "UserName":{"shape":"DomainUserName"}, + "Password":{"shape":"DomainUserPassword"} + } + }, + "JoinDomainOutput":{ + "type":"structure", + "members":{ + "GatewayARN":{"shape":"GatewayARN"} + } + }, "KMSKey":{ "type":"string", "max":2048, @@ -2168,6 +2370,39 @@ "max":2048, "min":20 }, + "SMBFileShareInfo":{ + "type":"structure", + "members":{ + "FileShareARN":{"shape":"FileShareARN"}, + "FileShareId":{"shape":"FileShareId"}, + "FileShareStatus":{"shape":"FileShareStatus"}, + "GatewayARN":{"shape":"GatewayARN"}, + "KMSEncrypted":{"shape":"boolean"}, + "KMSKey":{"shape":"KMSKey"}, + "Path":{"shape":"Path"}, + "Role":{"shape":"Role"}, + "LocationARN":{"shape":"LocationARN"}, + "DefaultStorageClass":{"shape":"StorageClass"}, + "ObjectACL":{"shape":"ObjectACL"}, + "ReadOnly":{"shape":"Boolean"}, + "GuessMIMETypeEnabled":{"shape":"Boolean"}, + "RequesterPays":{"shape":"Boolean"}, + "ValidUserList":{"shape":"FileShareUserList"}, + "InvalidUserList":{"shape":"FileShareUserList"}, + "Authentication":{"shape":"Authentication"} + } + }, + "SMBFileShareInfoList":{ + "type":"list", + "member":{"shape":"SMBFileShareInfo"} + }, + "SMBGuestPassword":{ + "type":"string", + "max":512, + "min":6, + "pattern":"^[ -~]+$", + "sensitive":true + }, "ServiceUnavailableError":{ "type":"structure", "members":{ @@ -2193,6 +2428,23 @@ "GatewayARN":{"shape":"GatewayARN"} } }, + "SetSMBGuestPasswordInput":{ + "type":"structure", + "required":[ + "GatewayARN", + "Password" + ], + "members":{ + "GatewayARN":{"shape":"GatewayARN"}, + "Password":{"shape":"SMBGuestPassword"} + } + }, + "SetSMBGuestPasswordOutput":{ + "type":"structure", + "members":{ + "GatewayARN":{"shape":"GatewayARN"} + } + }, "ShutdownGatewayInput":{ "type":"structure", "required":["GatewayARN"], @@ -2281,7 +2533,7 @@ "type":"string", "max":128, "min":1, - "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-%@]*)$" + "pattern":"^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-@]*)$" }, "TagKeys":{ "type":"list", @@ -2305,14 +2557,15 @@ "TapeStatus":{"shape":"TapeStatus"}, "VTLDevice":{"shape":"VTLDeviceARN"}, "Progress":{"shape":"DoubleObject"}, - "TapeUsedInBytes":{"shape":"TapeUsage"} + "TapeUsedInBytes":{"shape":"TapeUsage"}, + "KMSKey":{"shape":"KMSKey"} } }, "TapeARN":{ "type":"string", "max":500, "min":50, - "pattern":"^arn:(aws|aws-cn):storagegateway:[a-z\\-0-9]+:[0-9]+:tape\\/[0-9A-Z]{7,16}$" + "pattern":"^arn:(aws|aws-cn|aws-us-gov):storagegateway:[a-z\\-0-9]+:[0-9]+:tape\\/[0-9A-Z]{7,16}$" }, "TapeARNs":{ "type":"list", @@ -2328,7 +2581,8 @@ "CompletionTime":{"shape":"Time"}, "RetrievedTo":{"shape":"GatewayARN"}, "TapeStatus":{"shape":"TapeArchiveStatus"}, - "TapeUsedInBytes":{"shape":"TapeUsage"} + "TapeUsedInBytes":{"shape":"TapeUsage"}, + "KMSKey":{"shape":"KMSKey"} } }, "TapeArchiveStatus":{"type":"string"}, @@ -2509,6 +2763,28 @@ "FileShareARN":{"shape":"FileShareARN"} } }, + "UpdateSMBFileShareInput":{ + "type":"structure", + "required":["FileShareARN"], + "members":{ + "FileShareARN":{"shape":"FileShareARN"}, + "KMSEncrypted":{"shape":"Boolean"}, + "KMSKey":{"shape":"KMSKey"}, + "DefaultStorageClass":{"shape":"StorageClass"}, + "ObjectACL":{"shape":"ObjectACL"}, + "ReadOnly":{"shape":"Boolean"}, + "GuessMIMETypeEnabled":{"shape":"Boolean"}, + "RequesterPays":{"shape":"Boolean"}, + "ValidUserList":{"shape":"FileShareUserList"}, + "InvalidUserList":{"shape":"FileShareUserList"} + } + }, + "UpdateSMBFileShareOutput":{ + "type":"structure", + "members":{ + "FileShareARN":{"shape":"FileShareARN"} + } + }, "UpdateSnapshotScheduleInput":{ "type":"structure", "required":[ diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/docs-2.json b/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/docs-2.json index f7ad414e7..68314253b 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/docs-2.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/docs-2.json @@ -1,1773 +1,1958 @@ { - "version": "2.0", - "service": "AWS Storage Gateway Service

AWS Storage Gateway is the service that connects an on-premises software appliance with cloud-based storage to provide seamless and secure integration between an organization's on-premises IT environment and AWS's storage infrastructure. The service enables you to securely upload data to the AWS cloud for cost effective backup and rapid disaster recovery.

Use the following links to get started using the AWS Storage Gateway Service API Reference:

AWS Storage Gateway resource IDs are in uppercase. When you use these resource IDs with the Amazon EC2 API, EC2 expects resource IDs in lowercase. You must change your resource ID to lowercase to use it with the EC2 API. For example, in Storage Gateway the ID for a volume might be vol-AA22BB012345DAF670. When you use this ID with the EC2 API, you must change it to vol-aa22bb012345daf670. Otherwise, the EC2 API might not behave as expected.

IDs for Storage Gateway volumes and Amazon EBS snapshots created from gateway volumes are changing to a longer format. Starting in December 2016, all new volumes and snapshots will be created with a 17-character string. Starting in April 2016, you will be able to use these longer IDs so you can test your systems with the new format. For more information, see Longer EC2 and EBS Resource IDs.

For example, a volume Amazon Resource Name (ARN) with the longer volume ID format looks like the following:

arn:aws:storagegateway:us-west-2:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABBCCDDEEFFG.

A snapshot ID with the longer ID format looks like the following: snap-78e226633445566ee.

For more information, see Announcement: Heads-up – Longer AWS Storage Gateway volume and snapshot IDs coming in 2016.

", - "operations": { - "ActivateGateway": "

Activates the gateway you previously deployed on your host. In the activation process, you specify information such as the region you want to use for storing snapshots or tapes, the time zone for scheduled snapshots the gateway snapshot schedule window, an activation key, and a name for your gateway. The activation process also associates your gateway with your account; for more information, see UpdateGatewayInformation.

You must turn on the gateway VM before you can activate your gateway.

", - "AddCache": "

Configures one or more gateway local disks as cache for a gateway. This operation is only supported in the cached volume, tape and file gateway type (see Storage Gateway Concepts).

In the request, you specify the gateway Amazon Resource Name (ARN) to which you want to add cache, and one or more disk IDs that you want to configure as cache.

", - "AddTagsToResource": "

Adds one or more tags to the specified resource. You use tags to add metadata to resources, which you can use to categorize these resources. For example, you can categorize resources by purpose, owner, environment, or team. Each tag consists of a key and a value, which you define. You can add tags to the following AWS Storage Gateway resources:

  • Storage gateways of all types

  • Storage Volumes

  • Virtual Tapes

You can create a maximum of 10 tags for each resource. Virtual tapes and storage volumes that are recovered to a new gateway maintain their tags.

", - "AddUploadBuffer": "

Configures one or more gateway local disks as upload buffer for a specified gateway. This operation is supported for the stored volume, cached volume and tape gateway types.

In the request, you specify the gateway Amazon Resource Name (ARN) to which you want to add upload buffer, and one or more disk IDs that you want to configure as upload buffer.

", - "AddWorkingStorage": "

Configures one or more gateway local disks as working storage for a gateway. This operation is only supported in the stored volume gateway type. This operation is deprecated in cached volume API version 20120630. Use AddUploadBuffer instead.

Working storage is also referred to as upload buffer. You can also use the AddUploadBuffer operation to add upload buffer to a stored volume gateway.

In the request, you specify the gateway Amazon Resource Name (ARN) to which you want to add working storage, and one or more disk IDs that you want to configure as working storage.

", - "CancelArchival": "

Cancels archiving of a virtual tape to the virtual tape shelf (VTS) after the archiving process is initiated. This operation is only supported in the tape gateway type.

", - "CancelRetrieval": "

Cancels retrieval of a virtual tape from the virtual tape shelf (VTS) to a gateway after the retrieval process is initiated. The virtual tape is returned to the VTS. This operation is only supported in the tape gateway type.

", - "CreateCachediSCSIVolume": "

Creates a cached volume on a specified cached volume gateway. This operation is only supported in the cached volume gateway type.

Cache storage must be allocated to the gateway before you can create a cached volume. Use the AddCache operation to add cache storage to a gateway.

In the request, you must specify the gateway, size of the volume in bytes, the iSCSI target name, an IP address on which to expose the target, and a unique client token. In response, the gateway creates the volume and returns information about it. This information includes the volume Amazon Resource Name (ARN), its size, and the iSCSI target ARN that initiators can use to connect to the volume target.

Optionally, you can provide the ARN for an existing volume as the SourceVolumeARN for this cached volume, which creates an exact copy of the existing volume’s latest recovery point. The VolumeSizeInBytes value must be equal to or larger than the size of the copied volume, in bytes.

", - "CreateNFSFileShare": "

Creates a file share on an existing file gateway. In Storage Gateway, a file share is a file system mount point backed by Amazon S3 cloud storage. Storage Gateway exposes file shares using a Network File System (NFS) interface. This operation is only supported in the file gateway type.

File gateway requires AWS Security Token Service (AWS STS) to be activated to enable you create a file share. Make sure AWS STS is activated in the region you are creating your file gateway in. If AWS STS is not activated in the region, activate it. For information about how to activate AWS STS, see Activating and Deactivating AWS STS in an AWS Region in the AWS Identity and Access Management User Guide.

File gateway does not support creating hard or symbolic links on a file share.

", - "CreateSnapshot": "

Initiates a snapshot of a volume.

AWS Storage Gateway provides the ability to back up point-in-time snapshots of your data to Amazon Simple Storage (S3) for durable off-site recovery, as well as import the data to an Amazon Elastic Block Store (EBS) volume in Amazon Elastic Compute Cloud (EC2). You can take snapshots of your gateway volume on a scheduled or ad-hoc basis. This API enables you to take ad-hoc snapshot. For more information, see Editing a Snapshot Schedule.

In the CreateSnapshot request you identify the volume by providing its Amazon Resource Name (ARN). You must also provide description for the snapshot. When AWS Storage Gateway takes the snapshot of specified volume, the snapshot and description appears in the AWS Storage Gateway Console. In response, AWS Storage Gateway returns you a snapshot ID. You can use this snapshot ID to check the snapshot progress or later use it when you want to create a volume from a snapshot. This operation is only supported in stored and cached volume gateway type.

To list or delete a snapshot, you must use the Amazon EC2 API. For more information, see DescribeSnapshots or DeleteSnapshot in the EC2 API reference.

Volume and snapshot IDs are changing to a longer length ID format. For more information, see the important note on the Welcome page.

", - "CreateSnapshotFromVolumeRecoveryPoint": "

Initiates a snapshot of a gateway from a volume recovery point. This operation is only supported in the cached volume gateway type.

A volume recovery point is a point in time at which all data of the volume is consistent and from which you can create a snapshot. To get a list of volume recovery point for cached volume gateway, use ListVolumeRecoveryPoints.

In the CreateSnapshotFromVolumeRecoveryPoint request, you identify the volume by providing its Amazon Resource Name (ARN). You must also provide a description for the snapshot. When the gateway takes a snapshot of the specified volume, the snapshot and its description appear in the AWS Storage Gateway console. In response, the gateway returns you a snapshot ID. You can use this snapshot ID to check the snapshot progress or later use it when you want to create a volume from a snapshot.

To list or delete a snapshot, you must use the Amazon EC2 API. For more information, in Amazon Elastic Compute Cloud API Reference.

", - "CreateStorediSCSIVolume": "

Creates a volume on a specified gateway. This operation is only supported in the stored volume gateway type.

The size of the volume to create is inferred from the disk size. You can choose to preserve existing data on the disk, create volume from an existing snapshot, or create an empty volume. If you choose to create an empty gateway volume, then any existing data on the disk is erased.

In the request you must specify the gateway and the disk information on which you are creating the volume. In response, the gateway creates the volume and returns volume information such as the volume Amazon Resource Name (ARN), its size, and the iSCSI target ARN that initiators can use to connect to the volume target.

", - "CreateTapeWithBarcode": "

Creates a virtual tape by using your own barcode. You write data to the virtual tape and then archive the tape. A barcode is unique and can not be reused if it has already been used on a tape . This applies to barcodes used on deleted tapes. This operation is only supported in the tape gateway type.

Cache storage must be allocated to the gateway before you can create a virtual tape. Use the AddCache operation to add cache storage to a gateway.

", - "CreateTapes": "

Creates one or more virtual tapes. You write data to the virtual tapes and then archive the tapes. This operation is only supported in the tape gateway type.

Cache storage must be allocated to the gateway before you can create virtual tapes. Use the AddCache operation to add cache storage to a gateway.

", - "DeleteBandwidthRateLimit": "

Deletes the bandwidth rate limits of a gateway. You can delete either the upload and download bandwidth rate limit, or you can delete both. If you delete only one of the limits, the other limit remains unchanged. To specify which gateway to work with, use the Amazon Resource Name (ARN) of the gateway in your request.

", - "DeleteChapCredentials": "

Deletes Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target and initiator pair.

", - "DeleteFileShare": "

Deletes a file share from a file gateway. This operation is only supported in the file gateway type.

", - "DeleteGateway": "

Deletes a gateway. To specify which gateway to delete, use the Amazon Resource Name (ARN) of the gateway in your request. The operation deletes the gateway; however, it does not delete the gateway virtual machine (VM) from your host computer.

After you delete a gateway, you cannot reactivate it. Completed snapshots of the gateway volumes are not deleted upon deleting the gateway, however, pending snapshots will not complete. After you delete a gateway, your next step is to remove it from your environment.

You no longer pay software charges after the gateway is deleted; however, your existing Amazon EBS snapshots persist and you will continue to be billed for these snapshots. You can choose to remove all remaining Amazon EBS snapshots by canceling your Amazon EC2 subscription.  If you prefer not to cancel your Amazon EC2 subscription, you can delete your snapshots using the Amazon EC2 console. For more information, see the AWS Storage Gateway Detail Page.

", - "DeleteSnapshotSchedule": "

Deletes a snapshot of a volume.

You can take snapshots of your gateway volumes on a scheduled or ad hoc basis. This API action enables you to delete a snapshot schedule for a volume. For more information, see Working with Snapshots. In the DeleteSnapshotSchedule request, you identify the volume by providing its Amazon Resource Name (ARN). This operation is only supported in stored and cached volume gateway types.

To list or delete a snapshot, you must use the Amazon EC2 API. in Amazon Elastic Compute Cloud API Reference.

", - "DeleteTape": "

Deletes the specified virtual tape. This operation is only supported in the tape gateway type.

", - "DeleteTapeArchive": "

Deletes the specified virtual tape from the virtual tape shelf (VTS). This operation is only supported in the tape gateway type.

", - "DeleteVolume": "

Deletes the specified storage volume that you previously created using the CreateCachediSCSIVolume or CreateStorediSCSIVolume API. This operation is only supported in the cached volume and stored volume types. For stored volume gateways, the local disk that was configured as the storage volume is not deleted. You can reuse the local disk to create another storage volume.

Before you delete a volume, make sure there are no iSCSI connections to the volume you are deleting. You should also make sure there is no snapshot in progress. You can use the Amazon Elastic Compute Cloud (Amazon EC2) API to query snapshots on the volume you are deleting and check the snapshot status. For more information, go to DescribeSnapshots in the Amazon Elastic Compute Cloud API Reference.

In the request, you must provide the Amazon Resource Name (ARN) of the storage volume you want to delete.

", - "DescribeBandwidthRateLimit": "

Returns the bandwidth rate limits of a gateway. By default, these limits are not set, which means no bandwidth rate limiting is in effect.

This operation only returns a value for a bandwidth rate limit only if the limit is set. If no limits are set for the gateway, then this operation returns only the gateway ARN in the response body. To specify which gateway to describe, use the Amazon Resource Name (ARN) of the gateway in your request.

", - "DescribeCache": "

Returns information about the cache of a gateway. This operation is only supported in the cached volume, tape and file gateway types.

The response includes disk IDs that are configured as cache, and it includes the amount of cache allocated and used.

", - "DescribeCachediSCSIVolumes": "

Returns a description of the gateway volumes specified in the request. This operation is only supported in the cached volume gateway types.

The list of gateway volumes in the request must be from one gateway. In the response Amazon Storage Gateway returns volume information sorted by volume Amazon Resource Name (ARN).

", - "DescribeChapCredentials": "

Returns an array of Challenge-Handshake Authentication Protocol (CHAP) credentials information for a specified iSCSI target, one for each target-initiator pair.

", - "DescribeGatewayInformation": "

Returns metadata about a gateway such as its name, network interfaces, configured time zone, and the state (whether the gateway is running or not). To specify which gateway to describe, use the Amazon Resource Name (ARN) of the gateway in your request.

", - "DescribeMaintenanceStartTime": "

Returns your gateway's weekly maintenance start time including the day and time of the week. Note that values are in terms of the gateway's time zone.

", - "DescribeNFSFileShares": "

Gets a description for one or more file shares from a file gateway. This operation is only supported in the file gateway type.

", - "DescribeSnapshotSchedule": "

Describes the snapshot schedule for the specified gateway volume. The snapshot schedule information includes intervals at which snapshots are automatically initiated on the volume. This operation is only supported in the cached volume and stored volume types.

", - "DescribeStorediSCSIVolumes": "

Returns the description of the gateway volumes specified in the request. The list of gateway volumes in the request must be from one gateway. In the response Amazon Storage Gateway returns volume information sorted by volume ARNs. This operation is only supported in stored volume gateway type.

", - "DescribeTapeArchives": "

Returns a description of specified virtual tapes in the virtual tape shelf (VTS). This operation is only supported in the tape gateway type.

If a specific TapeARN is not specified, AWS Storage Gateway returns a description of all virtual tapes found in the VTS associated with your account.

", - "DescribeTapeRecoveryPoints": "

Returns a list of virtual tape recovery points that are available for the specified tape gateway.

A recovery point is a point-in-time view of a virtual tape at which all the data on the virtual tape is consistent. If your gateway crashes, virtual tapes that have recovery points can be recovered to a new gateway. This operation is only supported in the tape gateway type.

", - "DescribeTapes": "

Returns a description of the specified Amazon Resource Name (ARN) of virtual tapes. If a TapeARN is not specified, returns a description of all virtual tapes associated with the specified gateway. This operation is only supported in the tape gateway type.

", - "DescribeUploadBuffer": "

Returns information about the upload buffer of a gateway. This operation is supported for the stored volume, cached volume and tape gateway types.

The response includes disk IDs that are configured as upload buffer space, and it includes the amount of upload buffer space allocated and used.

", - "DescribeVTLDevices": "

Returns a description of virtual tape library (VTL) devices for the specified tape gateway. In the response, AWS Storage Gateway returns VTL device information.

This operation is only supported in the tape gateway type.

", - "DescribeWorkingStorage": "

Returns information about the working storage of a gateway. This operation is only supported in the stored volumes gateway type. This operation is deprecated in cached volumes API version (20120630). Use DescribeUploadBuffer instead.

Working storage is also referred to as upload buffer. You can also use the DescribeUploadBuffer operation to add upload buffer to a stored volume gateway.

The response includes disk IDs that are configured as working storage, and it includes the amount of working storage allocated and used.

", - "DisableGateway": "

Disables a tape gateway when the gateway is no longer functioning. For example, if your gateway VM is damaged, you can disable the gateway so you can recover virtual tapes.

Use this operation for a tape gateway that is not reachable or not functioning. This operation is only supported in the tape gateway type.

Once a gateway is disabled it cannot be enabled.

", - "ListFileShares": "

Gets a list of the file shares for a specific file gateway, or the list of file shares that belong to the calling user account. This operation is only supported in the file gateway type.

", - "ListGateways": "

Lists gateways owned by an AWS account in a region specified in the request. The returned list is ordered by gateway Amazon Resource Name (ARN).

By default, the operation returns a maximum of 100 gateways. This operation supports pagination that allows you to optionally reduce the number of gateways returned in a response.

If you have more gateways than are returned in a response (that is, the response returns only a truncated list of your gateways), the response contains a marker that you can specify in your next request to fetch the next page of gateways.

", - "ListLocalDisks": "

Returns a list of the gateway's local disks. To specify which gateway to describe, you use the Amazon Resource Name (ARN) of the gateway in the body of the request.

The request returns a list of all disks, specifying which are configured as working storage, cache storage, or stored volume or not configured at all. The response includes a DiskStatus field. This field can have a value of present (the disk is available to use), missing (the disk is no longer connected to the gateway), or mismatch (the disk node is occupied by a disk that has incorrect metadata or the disk content is corrupted).

", - "ListTagsForResource": "

Lists the tags that have been added to the specified resource. This operation is only supported in the cached volume, stored volume and tape gateway type.

", - "ListTapes": "

Lists virtual tapes in your virtual tape library (VTL) and your virtual tape shelf (VTS). You specify the tapes to list by specifying one or more tape Amazon Resource Names (ARNs). If you don't specify a tape ARN, the operation lists all virtual tapes in both your VTL and VTS.

This operation supports pagination. By default, the operation returns a maximum of up to 100 tapes. You can optionally specify the Limit parameter in the body to limit the number of tapes in the response. If the number of tapes returned in the response is truncated, the response includes a Marker element that you can use in your subsequent request to retrieve the next set of tapes. This operation is only supported in the tape gateway type.

", - "ListVolumeInitiators": "

Lists iSCSI initiators that are connected to a volume. You can use this operation to determine whether a volume is being used or not. This operation is only supported in the cached volume and stored volume gateway types.

", - "ListVolumeRecoveryPoints": "

Lists the recovery points for a specified gateway. This operation is only supported in the cached volume gateway type.

Each cache volume has one recovery point. A volume recovery point is a point in time at which all data of the volume is consistent and from which you can create a snapshot or clone a new cached volume from a source volume. To create a snapshot from a volume recovery point use the CreateSnapshotFromVolumeRecoveryPoint operation.

", - "ListVolumes": "

Lists the iSCSI stored volumes of a gateway. Results are sorted by volume ARN. The response includes only the volume ARNs. If you want additional volume information, use the DescribeStorediSCSIVolumes or the DescribeCachediSCSIVolumes API.

The operation supports pagination. By default, the operation returns a maximum of up to 100 volumes. You can optionally specify the Limit field in the body to limit the number of volumes in the response. If the number of volumes returned in the response is truncated, the response includes a Marker field. You can use this Marker value in your subsequent request to retrieve the next set of volumes. This operation is only supported in the cached volume and stored volume gateway types.

", - "NotifyWhenUploaded": "

Sends you notification through CloudWatch Events when all files written to your NFS file share have been uploaded to Amazon S3.

AWS Storage Gateway can send a notification through Amazon CloudWatch Events when all files written to your file share up to that point in time have been uploaded to Amazon S3. These files include files written to the NFS file share up to the time that you make a request for notification. When the upload is done, Storage Gateway sends you notification through an Amazon CloudWatch Event. You can configure CloudWatch Events to send the notification through event targets such as Amazon SNS or AWS Lambda function. This operation is only supported in the file gateway type.

For more information, see Getting File Upload Notification in the Storage Gateway User Guide (https://docs.aws.amazon.com/storagegateway/latest/userguide/monitoring-file-gateway.html#get-upload-notification).

", - "RefreshCache": "

Refreshes the cache for the specified file share. This operation finds objects in the Amazon S3 bucket that were added, removed or replaced since the gateway last listed the bucket's contents and cached the results. This operation is only supported in the file gateway type.

", - "RemoveTagsFromResource": "

Removes one or more tags from the specified resource. This operation is only supported in the cached volume, stored volume and tape gateway types.

", - "ResetCache": "

Resets all cache disks that have encountered a error and makes the disks available for reconfiguration as cache storage. If your cache disk encounters a error, the gateway prevents read and write operations on virtual tapes in the gateway. For example, an error can occur when a disk is corrupted or removed from the gateway. When a cache is reset, the gateway loses its cache storage. At this point you can reconfigure the disks as cache disks. This operation is only supported in the cached volume and tape types.

If the cache disk you are resetting contains data that has not been uploaded to Amazon S3 yet, that data can be lost. After you reset cache disks, there will be no configured cache disks left in the gateway, so you must configure at least one new cache disk for your gateway to function properly.

", - "RetrieveTapeArchive": "

Retrieves an archived virtual tape from the virtual tape shelf (VTS) to a tape gateway. Virtual tapes archived in the VTS are not associated with any gateway. However after a tape is retrieved, it is associated with a gateway, even though it is also listed in the VTS, that is, archive. This operation is only supported in the tape gateway type.

Once a tape is successfully retrieved to a gateway, it cannot be retrieved again to another gateway. You must archive the tape again before you can retrieve it to another gateway. This operation is only supported in the tape gateway type.

", - "RetrieveTapeRecoveryPoint": "

Retrieves the recovery point for the specified virtual tape. This operation is only supported in the tape gateway type.

A recovery point is a point in time view of a virtual tape at which all the data on the tape is consistent. If your gateway crashes, virtual tapes that have recovery points can be recovered to a new gateway.

The virtual tape can be retrieved to only one gateway. The retrieved tape is read-only. The virtual tape can be retrieved to only a tape gateway. There is no charge for retrieving recovery points.

", - "SetLocalConsolePassword": "

Sets the password for your VM local console. When you log in to the local console for the first time, you log in to the VM with the default credentials. We recommend that you set a new password. You don't need to know the default password to set a new password.

", - "ShutdownGateway": "

Shuts down a gateway. To specify which gateway to shut down, use the Amazon Resource Name (ARN) of the gateway in the body of your request.

The operation shuts down the gateway service component running in the gateway's virtual machine (VM) and not the host VM.

If you want to shut down the VM, it is recommended that you first shut down the gateway component in the VM to avoid unpredictable conditions.

After the gateway is shutdown, you cannot call any other API except StartGateway, DescribeGatewayInformation, and ListGateways. For more information, see ActivateGateway. Your applications cannot read from or write to the gateway's storage volumes, and there are no snapshots taken.

When you make a shutdown request, you will get a 200 OK success response immediately. However, it might take some time for the gateway to shut down. You can call the DescribeGatewayInformation API to check the status. For more information, see ActivateGateway.

If do not intend to use the gateway again, you must delete the gateway (using DeleteGateway) to no longer pay software charges associated with the gateway.

", - "StartGateway": "

Starts a gateway that you previously shut down (see ShutdownGateway). After the gateway starts, you can then make other API calls, your applications can read from or write to the gateway's storage volumes and you will be able to take snapshot backups.

When you make a request, you will get a 200 OK success response immediately. However, it might take some time for the gateway to be ready. You should call DescribeGatewayInformation and check the status before making any additional API calls. For more information, see ActivateGateway.

To specify which gateway to start, use the Amazon Resource Name (ARN) of the gateway in your request.

", - "UpdateBandwidthRateLimit": "

Updates the bandwidth rate limits of a gateway. You can update both the upload and download bandwidth rate limit or specify only one of the two. If you don't set a bandwidth rate limit, the existing rate limit remains.

By default, a gateway's bandwidth rate limits are not set. If you don't set any limit, the gateway does not have any limitations on its bandwidth usage and could potentially use the maximum available bandwidth.

To specify which gateway to update, use the Amazon Resource Name (ARN) of the gateway in your request.

", - "UpdateChapCredentials": "

Updates the Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target. By default, a gateway does not have CHAP enabled; however, for added security, you might use it.

When you update CHAP credentials, all existing connections on the target are closed and initiators must reconnect with the new credentials.

", - "UpdateGatewayInformation": "

Updates a gateway's metadata, which includes the gateway's name and time zone. To specify which gateway to update, use the Amazon Resource Name (ARN) of the gateway in your request.

For Gateways activated after September 2, 2015, the gateway's ARN contains the gateway ID rather than the gateway name. However, changing the name of the gateway has no effect on the gateway's ARN.

", - "UpdateGatewaySoftwareNow": "

Updates the gateway virtual machine (VM) software. The request immediately triggers the software update.

When you make this request, you get a 200 OK success response immediately. However, it might take some time for the update to complete. You can call DescribeGatewayInformation to verify the gateway is in the STATE_RUNNING state.

A software update forces a system restart of your gateway. You can minimize the chance of any disruption to your applications by increasing your iSCSI Initiators' timeouts. For more information about increasing iSCSI Initiator timeouts for Windows and Linux, see Customizing Your Windows iSCSI Settings and Customizing Your Linux iSCSI Settings, respectively.

", - "UpdateMaintenanceStartTime": "

Updates a gateway's weekly maintenance start time information, including day and time of the week. The maintenance time is the time in your gateway's time zone.

", - "UpdateNFSFileShare": "

Updates a file share. This operation is only supported in the file gateway type.

To leave a file share field unchanged, set the corresponding input field to null.

Updates the following file share setting:

  • Default storage class for your S3 bucket

  • Metadata defaults for your S3 bucket

  • Allowed NFS clients for your file share

  • Squash settings

  • Write status of your file share

To leave a file share field unchanged, set the corresponding input field to null. This operation is only supported in file gateways.

", - "UpdateSnapshotSchedule": "

Updates a snapshot schedule configured for a gateway volume. This operation is only supported in the cached volume and stored volume gateway types.

The default snapshot schedule for volume is once every 24 hours, starting at the creation time of the volume. You can use this API to change the snapshot schedule configured for the volume.

In the request you must identify the gateway volume whose snapshot schedule you want to update, and the schedule information, including when you want the snapshot to begin on a day and the frequency (in hours) of snapshots.

", - "UpdateVTLDeviceType": "

Updates the type of medium changer in a tape gateway. When you activate a tape gateway, you select a medium changer type for the tape gateway. This operation enables you to select a different type of medium changer after a tape gateway is activated. This operation is only supported in the tape gateway type.

" + "version":"2.0", + "service":"AWS Storage Gateway Service

AWS Storage Gateway is the service that connects an on-premises software appliance with cloud-based storage to provide seamless and secure integration between an organization's on-premises IT environment and AWS's storage infrastructure. The service enables you to securely upload data to the AWS cloud for cost effective backup and rapid disaster recovery.

Use the following links to get started using the AWS Storage Gateway Service API Reference:

AWS Storage Gateway resource IDs are in uppercase. When you use these resource IDs with the Amazon EC2 API, EC2 expects resource IDs in lowercase. You must change your resource ID to lowercase to use it with the EC2 API. For example, in Storage Gateway the ID for a volume might be vol-AA22BB012345DAF670. When you use this ID with the EC2 API, you must change it to vol-aa22bb012345daf670. Otherwise, the EC2 API might not behave as expected.

IDs for Storage Gateway volumes and Amazon EBS snapshots created from gateway volumes are changing to a longer format. Starting in December 2016, all new volumes and snapshots will be created with a 17-character string. Starting in April 2016, you will be able to use these longer IDs so you can test your systems with the new format. For more information, see Longer EC2 and EBS Resource IDs.

For example, a volume Amazon Resource Name (ARN) with the longer volume ID format looks like the following:

arn:aws:storagegateway:us-west-2:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABBCCDDEEFFG.

A snapshot ID with the longer ID format looks like the following: snap-78e226633445566ee.

For more information, see Announcement: Heads-up – Longer AWS Storage Gateway volume and snapshot IDs coming in 2016.

", + "operations":{ + "ActivateGateway":"

Activates the gateway you previously deployed on your host. In the activation process, you specify information such as the region you want to use for storing snapshots or tapes, the time zone for scheduled snapshots the gateway snapshot schedule window, an activation key, and a name for your gateway. The activation process also associates your gateway with your account; for more information, see UpdateGatewayInformation.

You must turn on the gateway VM before you can activate your gateway.

", + "AddCache":"

Configures one or more gateway local disks as cache for a gateway. This operation is only supported in the cached volume, tape and file gateway type (see Storage Gateway Concepts).

In the request, you specify the gateway Amazon Resource Name (ARN) to which you want to add cache, and one or more disk IDs that you want to configure as cache.

", + "AddTagsToResource":"

Adds one or more tags to the specified resource. You use tags to add metadata to resources, which you can use to categorize these resources. For example, you can categorize resources by purpose, owner, environment, or team. Each tag consists of a key and a value, which you define. You can add tags to the following AWS Storage Gateway resources:

  • Storage gateways of all types

  • Storage Volumes

  • Virtual Tapes

You can create a maximum of 10 tags for each resource. Virtual tapes and storage volumes that are recovered to a new gateway maintain their tags.

", + "AddUploadBuffer":"

Configures one or more gateway local disks as upload buffer for a specified gateway. This operation is supported for the stored volume, cached volume and tape gateway types.

In the request, you specify the gateway Amazon Resource Name (ARN) to which you want to add upload buffer, and one or more disk IDs that you want to configure as upload buffer.

", + "AddWorkingStorage":"

Configures one or more gateway local disks as working storage for a gateway. This operation is only supported in the stored volume gateway type. This operation is deprecated in cached volume API version 20120630. Use AddUploadBuffer instead.

Working storage is also referred to as upload buffer. You can also use the AddUploadBuffer operation to add upload buffer to a stored volume gateway.

In the request, you specify the gateway Amazon Resource Name (ARN) to which you want to add working storage, and one or more disk IDs that you want to configure as working storage.

", + "CancelArchival":"

Cancels archiving of a virtual tape to the virtual tape shelf (VTS) after the archiving process is initiated. This operation is only supported in the tape gateway type.

", + "CancelRetrieval":"

Cancels retrieval of a virtual tape from the virtual tape shelf (VTS) to a gateway after the retrieval process is initiated. The virtual tape is returned to the VTS. This operation is only supported in the tape gateway type.

", + "CreateCachediSCSIVolume":"

Creates a cached volume on a specified cached volume gateway. This operation is only supported in the cached volume gateway type.

Cache storage must be allocated to the gateway before you can create a cached volume. Use the AddCache operation to add cache storage to a gateway.

In the request, you must specify the gateway, size of the volume in bytes, the iSCSI target name, an IP address on which to expose the target, and a unique client token. In response, the gateway creates the volume and returns information about it. This information includes the volume Amazon Resource Name (ARN), its size, and the iSCSI target ARN that initiators can use to connect to the volume target.

Optionally, you can provide the ARN for an existing volume as the SourceVolumeARN for this cached volume, which creates an exact copy of the existing volume’s latest recovery point. The VolumeSizeInBytes value must be equal to or larger than the size of the copied volume, in bytes.

", + "CreateNFSFileShare":"

Creates a Network File System (NFS) file share on an existing file gateway. In Storage Gateway, a file share is a file system mount point backed by Amazon S3 cloud storage. Storage Gateway exposes file shares using a NFS interface. This operation is only supported in the file gateway type.

File gateway requires AWS Security Token Service (AWS STS) to be activated to enable you create a file share. Make sure AWS STS is activated in the region you are creating your file gateway in. If AWS STS is not activated in the region, activate it. For information about how to activate AWS STS, see Activating and Deactivating AWS STS in an AWS Region in the AWS Identity and Access Management User Guide.

File gateway does not support creating hard or symbolic links on a file share.

", + "CreateSMBFileShare":"

Creates a Server Message Block (SMB) file share on an existing file gateway. In Storage Gateway, a file share is a file system mount point backed by Amazon S3 cloud storage. Storage Gateway expose file shares using a SMB interface. This operation is only supported in the file gateway type.

File gateway requires AWS Security Token Service (AWS STS) to be activated to enable you create a file share. Make sure AWS STS is activated in the region you are creating your file gateway in. If AWS STS is not activated in the region, activate it. For information about how to activate AWS STS, see Activating and Deactivating AWS STS in an AWS Region in the AWS Identity and Access Management User Guide.

File gateway does not support creating hard or symbolic links on a file share.

", + "CreateSnapshot":"

Initiates a snapshot of a volume.

AWS Storage Gateway provides the ability to back up point-in-time snapshots of your data to Amazon Simple Storage (S3) for durable off-site recovery, as well as import the data to an Amazon Elastic Block Store (EBS) volume in Amazon Elastic Compute Cloud (EC2). You can take snapshots of your gateway volume on a scheduled or ad-hoc basis. This API enables you to take ad-hoc snapshot. For more information, see Editing a Snapshot Schedule.

In the CreateSnapshot request you identify the volume by providing its Amazon Resource Name (ARN). You must also provide description for the snapshot. When AWS Storage Gateway takes the snapshot of specified volume, the snapshot and description appears in the AWS Storage Gateway Console. In response, AWS Storage Gateway returns you a snapshot ID. You can use this snapshot ID to check the snapshot progress or later use it when you want to create a volume from a snapshot. This operation is only supported in stored and cached volume gateway type.

To list or delete a snapshot, you must use the Amazon EC2 API. For more information, see DescribeSnapshots or DeleteSnapshot in the EC2 API reference.

Volume and snapshot IDs are changing to a longer length ID format. For more information, see the important note on the Welcome page.

", + "CreateSnapshotFromVolumeRecoveryPoint":"

Initiates a snapshot of a gateway from a volume recovery point. This operation is only supported in the cached volume gateway type.

A volume recovery point is a point in time at which all data of the volume is consistent and from which you can create a snapshot. To get a list of volume recovery point for cached volume gateway, use ListVolumeRecoveryPoints.

In the CreateSnapshotFromVolumeRecoveryPoint request, you identify the volume by providing its Amazon Resource Name (ARN). You must also provide a description for the snapshot. When the gateway takes a snapshot of the specified volume, the snapshot and its description appear in the AWS Storage Gateway console. In response, the gateway returns you a snapshot ID. You can use this snapshot ID to check the snapshot progress or later use it when you want to create a volume from a snapshot.

To list or delete a snapshot, you must use the Amazon EC2 API. For more information, in Amazon Elastic Compute Cloud API Reference.

", + "CreateStorediSCSIVolume":"

Creates a volume on a specified gateway. This operation is only supported in the stored volume gateway type.

The size of the volume to create is inferred from the disk size. You can choose to preserve existing data on the disk, create volume from an existing snapshot, or create an empty volume. If you choose to create an empty gateway volume, then any existing data on the disk is erased.

In the request you must specify the gateway and the disk information on which you are creating the volume. In response, the gateway creates the volume and returns volume information such as the volume Amazon Resource Name (ARN), its size, and the iSCSI target ARN that initiators can use to connect to the volume target.

", + "CreateTapeWithBarcode":"

Creates a virtual tape by using your own barcode. You write data to the virtual tape and then archive the tape. A barcode is unique and can not be reused if it has already been used on a tape . This applies to barcodes used on deleted tapes. This operation is only supported in the tape gateway type.

Cache storage must be allocated to the gateway before you can create a virtual tape. Use the AddCache operation to add cache storage to a gateway.

", + "CreateTapes":"

Creates one or more virtual tapes. You write data to the virtual tapes and then archive the tapes. This operation is only supported in the tape gateway type.

Cache storage must be allocated to the gateway before you can create virtual tapes. Use the AddCache operation to add cache storage to a gateway.

", + "DeleteBandwidthRateLimit":"

Deletes the bandwidth rate limits of a gateway. You can delete either the upload and download bandwidth rate limit, or you can delete both. If you delete only one of the limits, the other limit remains unchanged. To specify which gateway to work with, use the Amazon Resource Name (ARN) of the gateway in your request.

", + "DeleteChapCredentials":"

Deletes Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target and initiator pair.

", + "DeleteFileShare":"

Deletes a file share from a file gateway. This operation is only supported in the file gateway type.

", + "DeleteGateway":"

Deletes a gateway. To specify which gateway to delete, use the Amazon Resource Name (ARN) of the gateway in your request. The operation deletes the gateway; however, it does not delete the gateway virtual machine (VM) from your host computer.

After you delete a gateway, you cannot reactivate it. Completed snapshots of the gateway volumes are not deleted upon deleting the gateway, however, pending snapshots will not complete. After you delete a gateway, your next step is to remove it from your environment.

You no longer pay software charges after the gateway is deleted; however, your existing Amazon EBS snapshots persist and you will continue to be billed for these snapshots. You can choose to remove all remaining Amazon EBS snapshots by canceling your Amazon EC2 subscription.  If you prefer not to cancel your Amazon EC2 subscription, you can delete your snapshots using the Amazon EC2 console. For more information, see the AWS Storage Gateway Detail Page.

", + "DeleteSnapshotSchedule":"

Deletes a snapshot of a volume.

You can take snapshots of your gateway volumes on a scheduled or ad hoc basis. This API action enables you to delete a snapshot schedule for a volume. For more information, see Working with Snapshots. In the DeleteSnapshotSchedule request, you identify the volume by providing its Amazon Resource Name (ARN). This operation is only supported in stored and cached volume gateway types.

To list or delete a snapshot, you must use the Amazon EC2 API. in Amazon Elastic Compute Cloud API Reference.

", + "DeleteTape":"

Deletes the specified virtual tape. This operation is only supported in the tape gateway type.

", + "DeleteTapeArchive":"

Deletes the specified virtual tape from the virtual tape shelf (VTS). This operation is only supported in the tape gateway type.

", + "DeleteVolume":"

Deletes the specified storage volume that you previously created using the CreateCachediSCSIVolume or CreateStorediSCSIVolume API. This operation is only supported in the cached volume and stored volume types. For stored volume gateways, the local disk that was configured as the storage volume is not deleted. You can reuse the local disk to create another storage volume.

Before you delete a volume, make sure there are no iSCSI connections to the volume you are deleting. You should also make sure there is no snapshot in progress. You can use the Amazon Elastic Compute Cloud (Amazon EC2) API to query snapshots on the volume you are deleting and check the snapshot status. For more information, go to DescribeSnapshots in the Amazon Elastic Compute Cloud API Reference.

In the request, you must provide the Amazon Resource Name (ARN) of the storage volume you want to delete.

", + "DescribeBandwidthRateLimit":"

Returns the bandwidth rate limits of a gateway. By default, these limits are not set, which means no bandwidth rate limiting is in effect.

This operation only returns a value for a bandwidth rate limit only if the limit is set. If no limits are set for the gateway, then this operation returns only the gateway ARN in the response body. To specify which gateway to describe, use the Amazon Resource Name (ARN) of the gateway in your request.

", + "DescribeCache":"

Returns information about the cache of a gateway. This operation is only supported in the cached volume, tape and file gateway types.

The response includes disk IDs that are configured as cache, and it includes the amount of cache allocated and used.

", + "DescribeCachediSCSIVolumes":"

Returns a description of the gateway volumes specified in the request. This operation is only supported in the cached volume gateway types.

The list of gateway volumes in the request must be from one gateway. In the response Amazon Storage Gateway returns volume information sorted by volume Amazon Resource Name (ARN).

", + "DescribeChapCredentials":"

Returns an array of Challenge-Handshake Authentication Protocol (CHAP) credentials information for a specified iSCSI target, one for each target-initiator pair.

", + "DescribeGatewayInformation":"

Returns metadata about a gateway such as its name, network interfaces, configured time zone, and the state (whether the gateway is running or not). To specify which gateway to describe, use the Amazon Resource Name (ARN) of the gateway in your request.

", + "DescribeMaintenanceStartTime":"

Returns your gateway's weekly maintenance start time including the day and time of the week. Note that values are in terms of the gateway's time zone.

", + "DescribeNFSFileShares":"

Gets a description for one or more Network File System (NFS) file shares from a file gateway. This operation is only supported in the file gateway type.

", + "DescribeSMBFileShares":"

Gets a description for one or more Server Message Block (SMB) file shares from a file gateway. This operation is only supported in the file gateway type.

", + "DescribeSMBSettings":"

Gets a description of a Server Message Block (SMB) file share settings from a file gateway. This operation is only supported in the file gateway type.

", + "DescribeSnapshotSchedule":"

Describes the snapshot schedule for the specified gateway volume. The snapshot schedule information includes intervals at which snapshots are automatically initiated on the volume. This operation is only supported in the cached volume and stored volume types.

", + "DescribeStorediSCSIVolumes":"

Returns the description of the gateway volumes specified in the request. The list of gateway volumes in the request must be from one gateway. In the response Amazon Storage Gateway returns volume information sorted by volume ARNs. This operation is only supported in stored volume gateway type.

", + "DescribeTapeArchives":"

Returns a description of specified virtual tapes in the virtual tape shelf (VTS). This operation is only supported in the tape gateway type.

If a specific TapeARN is not specified, AWS Storage Gateway returns a description of all virtual tapes found in the VTS associated with your account.

", + "DescribeTapeRecoveryPoints":"

Returns a list of virtual tape recovery points that are available for the specified tape gateway.

A recovery point is a point-in-time view of a virtual tape at which all the data on the virtual tape is consistent. If your gateway crashes, virtual tapes that have recovery points can be recovered to a new gateway. This operation is only supported in the tape gateway type.

", + "DescribeTapes":"

Returns a description of the specified Amazon Resource Name (ARN) of virtual tapes. If a TapeARN is not specified, returns a description of all virtual tapes associated with the specified gateway. This operation is only supported in the tape gateway type.

", + "DescribeUploadBuffer":"

Returns information about the upload buffer of a gateway. This operation is supported for the stored volume, cached volume and tape gateway types.

The response includes disk IDs that are configured as upload buffer space, and it includes the amount of upload buffer space allocated and used.

", + "DescribeVTLDevices":"

Returns a description of virtual tape library (VTL) devices for the specified tape gateway. In the response, AWS Storage Gateway returns VTL device information.

This operation is only supported in the tape gateway type.

", + "DescribeWorkingStorage":"

Returns information about the working storage of a gateway. This operation is only supported in the stored volumes gateway type. This operation is deprecated in cached volumes API version (20120630). Use DescribeUploadBuffer instead.

Working storage is also referred to as upload buffer. You can also use the DescribeUploadBuffer operation to add upload buffer to a stored volume gateway.

The response includes disk IDs that are configured as working storage, and it includes the amount of working storage allocated and used.

", + "DisableGateway":"

Disables a tape gateway when the gateway is no longer functioning. For example, if your gateway VM is damaged, you can disable the gateway so you can recover virtual tapes.

Use this operation for a tape gateway that is not reachable or not functioning. This operation is only supported in the tape gateway type.

Once a gateway is disabled it cannot be enabled.

", + "JoinDomain":"

Adds a file gateway to an Active Directory domain. This operation is only supported in the file gateway type that supports the SMB file protocol.

", + "ListFileShares":"

Gets a list of the file shares for a specific file gateway, or the list of file shares that belong to the calling user account. This operation is only supported in the file gateway type.

", + "ListGateways":"

Lists gateways owned by an AWS account in a region specified in the request. The returned list is ordered by gateway Amazon Resource Name (ARN).

By default, the operation returns a maximum of 100 gateways. This operation supports pagination that allows you to optionally reduce the number of gateways returned in a response.

If you have more gateways than are returned in a response (that is, the response returns only a truncated list of your gateways), the response contains a marker that you can specify in your next request to fetch the next page of gateways.

", + "ListLocalDisks":"

Returns a list of the gateway's local disks. To specify which gateway to describe, you use the Amazon Resource Name (ARN) of the gateway in the body of the request.

The request returns a list of all disks, specifying which are configured as working storage, cache storage, or stored volume or not configured at all. The response includes a DiskStatus field. This field can have a value of present (the disk is available to use), missing (the disk is no longer connected to the gateway), or mismatch (the disk node is occupied by a disk that has incorrect metadata or the disk content is corrupted).

", + "ListTagsForResource":"

Lists the tags that have been added to the specified resource. This operation is only supported in the cached volume, stored volume and tape gateway type.

", + "ListTapes":"

Lists virtual tapes in your virtual tape library (VTL) and your virtual tape shelf (VTS). You specify the tapes to list by specifying one or more tape Amazon Resource Names (ARNs). If you don't specify a tape ARN, the operation lists all virtual tapes in both your VTL and VTS.

This operation supports pagination. By default, the operation returns a maximum of up to 100 tapes. You can optionally specify the Limit parameter in the body to limit the number of tapes in the response. If the number of tapes returned in the response is truncated, the response includes a Marker element that you can use in your subsequent request to retrieve the next set of tapes. This operation is only supported in the tape gateway type.

", + "ListVolumeInitiators":"

Lists iSCSI initiators that are connected to a volume. You can use this operation to determine whether a volume is being used or not. This operation is only supported in the cached volume and stored volume gateway types.

", + "ListVolumeRecoveryPoints":"

Lists the recovery points for a specified gateway. This operation is only supported in the cached volume gateway type.

Each cache volume has one recovery point. A volume recovery point is a point in time at which all data of the volume is consistent and from which you can create a snapshot or clone a new cached volume from a source volume. To create a snapshot from a volume recovery point use the CreateSnapshotFromVolumeRecoveryPoint operation.

", + "ListVolumes":"

Lists the iSCSI stored volumes of a gateway. Results are sorted by volume ARN. The response includes only the volume ARNs. If you want additional volume information, use the DescribeStorediSCSIVolumes or the DescribeCachediSCSIVolumes API.

The operation supports pagination. By default, the operation returns a maximum of up to 100 volumes. You can optionally specify the Limit field in the body to limit the number of volumes in the response. If the number of volumes returned in the response is truncated, the response includes a Marker field. You can use this Marker value in your subsequent request to retrieve the next set of volumes. This operation is only supported in the cached volume and stored volume gateway types.

", + "NotifyWhenUploaded":"

Sends you notification through CloudWatch Events when all files written to your NFS file share have been uploaded to Amazon S3.

AWS Storage Gateway can send a notification through Amazon CloudWatch Events when all files written to your file share up to that point in time have been uploaded to Amazon S3. These files include files written to the NFS file share up to the time that you make a request for notification. When the upload is done, Storage Gateway sends you notification through an Amazon CloudWatch Event. You can configure CloudWatch Events to send the notification through event targets such as Amazon SNS or AWS Lambda function. This operation is only supported in the file gateway type.

For more information, see Getting File Upload Notification in the Storage Gateway User Guide (https://docs.aws.amazon.com/storagegateway/latest/userguide/monitoring-file-gateway.html#get-upload-notification).

", + "RefreshCache":"

Refreshes the cache for the specified file share. This operation finds objects in the Amazon S3 bucket that were added, removed or replaced since the gateway last listed the bucket's contents and cached the results. This operation is only supported in the file gateway type.

", + "RemoveTagsFromResource":"

Removes one or more tags from the specified resource. This operation is only supported in the cached volume, stored volume and tape gateway types.

", + "ResetCache":"

Resets all cache disks that have encountered a error and makes the disks available for reconfiguration as cache storage. If your cache disk encounters a error, the gateway prevents read and write operations on virtual tapes in the gateway. For example, an error can occur when a disk is corrupted or removed from the gateway. When a cache is reset, the gateway loses its cache storage. At this point you can reconfigure the disks as cache disks. This operation is only supported in the cached volume and tape types.

If the cache disk you are resetting contains data that has not been uploaded to Amazon S3 yet, that data can be lost. After you reset cache disks, there will be no configured cache disks left in the gateway, so you must configure at least one new cache disk for your gateway to function properly.

", + "RetrieveTapeArchive":"

Retrieves an archived virtual tape from the virtual tape shelf (VTS) to a tape gateway. Virtual tapes archived in the VTS are not associated with any gateway. However after a tape is retrieved, it is associated with a gateway, even though it is also listed in the VTS, that is, archive. This operation is only supported in the tape gateway type.

Once a tape is successfully retrieved to a gateway, it cannot be retrieved again to another gateway. You must archive the tape again before you can retrieve it to another gateway. This operation is only supported in the tape gateway type.

", + "RetrieveTapeRecoveryPoint":"

Retrieves the recovery point for the specified virtual tape. This operation is only supported in the tape gateway type.

A recovery point is a point in time view of a virtual tape at which all the data on the tape is consistent. If your gateway crashes, virtual tapes that have recovery points can be recovered to a new gateway.

The virtual tape can be retrieved to only one gateway. The retrieved tape is read-only. The virtual tape can be retrieved to only a tape gateway. There is no charge for retrieving recovery points.

", + "SetLocalConsolePassword":"

Sets the password for your VM local console. When you log in to the local console for the first time, you log in to the VM with the default credentials. We recommend that you set a new password. You don't need to know the default password to set a new password.

", + "SetSMBGuestPassword":"

Sets the password for the guest user “smbguest”. \"smbguest\" is the user when the Authentication method for the file share is “GuestAccess”.

", + "ShutdownGateway":"

Shuts down a gateway. To specify which gateway to shut down, use the Amazon Resource Name (ARN) of the gateway in the body of your request.

The operation shuts down the gateway service component running in the gateway's virtual machine (VM) and not the host VM.

If you want to shut down the VM, it is recommended that you first shut down the gateway component in the VM to avoid unpredictable conditions.

After the gateway is shutdown, you cannot call any other API except StartGateway, DescribeGatewayInformation, and ListGateways. For more information, see ActivateGateway. Your applications cannot read from or write to the gateway's storage volumes, and there are no snapshots taken.

When you make a shutdown request, you will get a 200 OK success response immediately. However, it might take some time for the gateway to shut down. You can call the DescribeGatewayInformation API to check the status. For more information, see ActivateGateway.

If do not intend to use the gateway again, you must delete the gateway (using DeleteGateway) to no longer pay software charges associated with the gateway.

", + "StartGateway":"

Starts a gateway that you previously shut down (see ShutdownGateway). After the gateway starts, you can then make other API calls, your applications can read from or write to the gateway's storage volumes and you will be able to take snapshot backups.

When you make a request, you will get a 200 OK success response immediately. However, it might take some time for the gateway to be ready. You should call DescribeGatewayInformation and check the status before making any additional API calls. For more information, see ActivateGateway.

To specify which gateway to start, use the Amazon Resource Name (ARN) of the gateway in your request.

", + "UpdateBandwidthRateLimit":"

Updates the bandwidth rate limits of a gateway. You can update both the upload and download bandwidth rate limit or specify only one of the two. If you don't set a bandwidth rate limit, the existing rate limit remains.

By default, a gateway's bandwidth rate limits are not set. If you don't set any limit, the gateway does not have any limitations on its bandwidth usage and could potentially use the maximum available bandwidth.

To specify which gateway to update, use the Amazon Resource Name (ARN) of the gateway in your request.

", + "UpdateChapCredentials":"

Updates the Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target. By default, a gateway does not have CHAP enabled; however, for added security, you might use it.

When you update CHAP credentials, all existing connections on the target are closed and initiators must reconnect with the new credentials.

", + "UpdateGatewayInformation":"

Updates a gateway's metadata, which includes the gateway's name and time zone. To specify which gateway to update, use the Amazon Resource Name (ARN) of the gateway in your request.

For Gateways activated after September 2, 2015, the gateway's ARN contains the gateway ID rather than the gateway name. However, changing the name of the gateway has no effect on the gateway's ARN.

", + "UpdateGatewaySoftwareNow":"

Updates the gateway virtual machine (VM) software. The request immediately triggers the software update.

When you make this request, you get a 200 OK success response immediately. However, it might take some time for the update to complete. You can call DescribeGatewayInformation to verify the gateway is in the STATE_RUNNING state.

A software update forces a system restart of your gateway. You can minimize the chance of any disruption to your applications by increasing your iSCSI Initiators' timeouts. For more information about increasing iSCSI Initiator timeouts for Windows and Linux, see Customizing Your Windows iSCSI Settings and Customizing Your Linux iSCSI Settings, respectively.

", + "UpdateMaintenanceStartTime":"

Updates a gateway's weekly maintenance start time information, including day and time of the week. The maintenance time is the time in your gateway's time zone.

", + "UpdateNFSFileShare":"

Updates a Network File System (NFS) file share. This operation is only supported in the file gateway type.

To leave a file share field unchanged, set the corresponding input field to null.

Updates the following file share setting:

  • Default storage class for your S3 bucket

  • Metadata defaults for your S3 bucket

  • Allowed NFS clients for your file share

  • Squash settings

  • Write status of your file share

To leave a file share field unchanged, set the corresponding input field to null. This operation is only supported in file gateways.

", + "UpdateSMBFileShare":"

Updates a Server Message Block (SMB) file share. This operation is only supported in the file gateway type.

To leave a file share field unchanged, set the corresponding input field to null. This operation is only supported in the file gateway type.

File gateway requires AWS Security Token Service (AWS STS) to be activated to enable you create a file share. Make sure AWS STS is activated in the region you are creating your file gateway in. If AWS STS is not activated in the region, activate it. For information about how to activate AWS STS, see Activating and Deactivating AWS STS in an AWS Region in the AWS Identity and Access Management User Guide.

File gateway does not support creating hard or symbolic links on a file share.

", + "UpdateSnapshotSchedule":"

Updates a snapshot schedule configured for a gateway volume. This operation is only supported in the cached volume and stored volume gateway types.

The default snapshot schedule for volume is once every 24 hours, starting at the creation time of the volume. You can use this API to change the snapshot schedule configured for the volume.

In the request you must identify the gateway volume whose snapshot schedule you want to update, and the schedule information, including when you want the snapshot to begin on a day and the frequency (in hours) of snapshots.

", + "UpdateVTLDeviceType":"

Updates the type of medium changer in a tape gateway. When you activate a tape gateway, you select a medium changer type for the tape gateway. This operation enables you to select a different type of medium changer after a tape gateway is activated. This operation is only supported in the tape gateway type.

" }, - "shapes": { - "ActivateGatewayInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "shapes":{ + "ActivateGatewayInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "ActivateGatewayOutput": { - "base": "

AWS Storage Gateway returns the Amazon Resource Name (ARN) of the activated gateway. It is a string made of information such as your account, gateway name, and region. This ARN is used to reference the gateway in other API operations as well as resource-based authorization.

For gateways activated prior to September 02, 2015, the gateway ARN contains the gateway name rather than the gateway ID. Changing the name of the gateway has no effect on the gateway ARN.

", - "refs": { + "ActivateGatewayOutput":{ + "base":"

AWS Storage Gateway returns the Amazon Resource Name (ARN) of the activated gateway. It is a string made of information such as your account, gateway name, and region. This ARN is used to reference the gateway in other API operations as well as resource-based authorization.

For gateways activated prior to September 02, 2015, the gateway ARN contains the gateway name rather than the gateway ID. Changing the name of the gateway has no effect on the gateway ARN.

", + "refs":{ } }, - "ActivationKey": { - "base": null, - "refs": { - "ActivateGatewayInput$ActivationKey": "

Your gateway activation key. You can obtain the activation key by sending an HTTP GET request with redirects enabled to the gateway IP address (port 80). The redirect URL returned in the response provides you the activation key for your gateway in the query string parameter activationKey. It may also include other activation-related parameters, however, these are merely defaults -- the arguments you pass to the ActivateGateway API call determine the actual configuration of your gateway.

For more information, see https://docs.aws.amazon.com/storagegateway/latest/userguide/get-activation-key.html in the Storage Gateway User Guide.

" + "ActivationKey":{ + "base":null, + "refs":{ + "ActivateGatewayInput$ActivationKey":"

Your gateway activation key. You can obtain the activation key by sending an HTTP GET request with redirects enabled to the gateway IP address (port 80). The redirect URL returned in the response provides you the activation key for your gateway in the query string parameter activationKey. It may also include other activation-related parameters, however, these are merely defaults -- the arguments you pass to the ActivateGateway API call determine the actual configuration of your gateway.

For more information, see https://docs.aws.amazon.com/storagegateway/latest/userguide/get-activation-key.html in the Storage Gateway User Guide.

" } }, - "AddCacheInput": { - "base": null, - "refs": { + "AddCacheInput":{ + "base":null, + "refs":{ } }, - "AddCacheOutput": { - "base": null, - "refs": { + "AddCacheOutput":{ + "base":null, + "refs":{ } }, - "AddTagsToResourceInput": { - "base": "

AddTagsToResourceInput

", - "refs": { + "AddTagsToResourceInput":{ + "base":"

AddTagsToResourceInput

", + "refs":{ } }, - "AddTagsToResourceOutput": { - "base": "

AddTagsToResourceOutput

", - "refs": { + "AddTagsToResourceOutput":{ + "base":"

AddTagsToResourceOutput

", + "refs":{ } }, - "AddUploadBufferInput": { - "base": null, - "refs": { + "AddUploadBufferInput":{ + "base":null, + "refs":{ } }, - "AddUploadBufferOutput": { - "base": null, - "refs": { + "AddUploadBufferOutput":{ + "base":null, + "refs":{ } }, - "AddWorkingStorageInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "AddWorkingStorageInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "AddWorkingStorageOutput": { - "base": "

A JSON object containing the of the gateway for which working storage was configured.

", - "refs": { + "AddWorkingStorageOutput":{ + "base":"

A JSON object containing the of the gateway for which working storage was configured.

", + "refs":{ } }, - "BandwidthDownloadRateLimit": { - "base": null, - "refs": { - "DescribeBandwidthRateLimitOutput$AverageDownloadRateLimitInBitsPerSec": "

The average download bandwidth rate limit in bits per second. This field does not appear in the response if the download rate limit is not set.

", - "UpdateBandwidthRateLimitInput$AverageDownloadRateLimitInBitsPerSec": "

The average download bandwidth rate limit in bits per second.

" + "Authentication":{ + "base":"

The authentication method of the file share. Valid values: \"ActiveDirectory\" or \"GuestAccess\". The default is \"ActiveDirectory\".

", + "refs":{ + "CreateSMBFileShareInput$Authentication":"

The authentication method that users use to access the file share.

Valid values: \"ActiveDirectory\" or \"GuestAccess\". The default is \"ActiveDirectory\".

", + "SMBFileShareInfo$Authentication":null } }, - "BandwidthType": { - "base": null, - "refs": { - "DeleteBandwidthRateLimitInput$BandwidthType": "

One of the BandwidthType values that indicates the gateway bandwidth rate limit to delete.

Valid Values: Upload, Download, All.

" + "BandwidthDownloadRateLimit":{ + "base":null, + "refs":{ + "DescribeBandwidthRateLimitOutput$AverageDownloadRateLimitInBitsPerSec":"

The average download bandwidth rate limit in bits per second. This field does not appear in the response if the download rate limit is not set.

", + "UpdateBandwidthRateLimitInput$AverageDownloadRateLimitInBitsPerSec":"

The average download bandwidth rate limit in bits per second.

" } }, - "BandwidthUploadRateLimit": { - "base": null, - "refs": { - "DescribeBandwidthRateLimitOutput$AverageUploadRateLimitInBitsPerSec": "

The average upload bandwidth rate limit in bits per second. This field does not appear in the response if the upload rate limit is not set.

", - "UpdateBandwidthRateLimitInput$AverageUploadRateLimitInBitsPerSec": "

The average upload bandwidth rate limit in bits per second.

" + "BandwidthType":{ + "base":null, + "refs":{ + "DeleteBandwidthRateLimitInput$BandwidthType":"

One of the BandwidthType values that indicates the gateway bandwidth rate limit to delete.

Valid Values: Upload, Download, All.

" } }, - "Boolean": { - "base": null, - "refs": { - "CreateNFSFileShareInput$KMSEncrypted": "

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", - "CreateNFSFileShareInput$ReadOnly": "

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", - "CreateNFSFileShareInput$GuessMIMETypeEnabled": "

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", - "CreateNFSFileShareInput$RequesterPays": "

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", - "NFSFileShareInfo$ReadOnly": "

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", - "NFSFileShareInfo$GuessMIMETypeEnabled": "

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", - "NFSFileShareInfo$RequesterPays": "

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", - "UpdateNFSFileShareInput$KMSEncrypted": "

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", - "UpdateNFSFileShareInput$ReadOnly": "

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", - "UpdateNFSFileShareInput$GuessMIMETypeEnabled": "

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", - "UpdateNFSFileShareInput$RequesterPays": "

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

" + "BandwidthUploadRateLimit":{ + "base":null, + "refs":{ + "DescribeBandwidthRateLimitOutput$AverageUploadRateLimitInBitsPerSec":"

The average upload bandwidth rate limit in bits per second. This field does not appear in the response if the upload rate limit is not set.

", + "UpdateBandwidthRateLimitInput$AverageUploadRateLimitInBitsPerSec":"

The average upload bandwidth rate limit in bits per second.

" } }, - "CachediSCSIVolume": { - "base": "

Describes an iSCSI cached volume.

", - "refs": { - "CachediSCSIVolumes$member": null + "Boolean":{ + "base":null, + "refs":{ + "CreateCachediSCSIVolumeInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "CreateNFSFileShareInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "CreateNFSFileShareInput$ReadOnly":"

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", + "CreateNFSFileShareInput$GuessMIMETypeEnabled":"

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", + "CreateNFSFileShareInput$RequesterPays":"

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", + "CreateSMBFileShareInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "CreateSMBFileShareInput$ReadOnly":"

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", + "CreateSMBFileShareInput$GuessMIMETypeEnabled":"

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", + "CreateSMBFileShareInput$RequesterPays":"

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", + "CreateTapeWithBarcodeInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "CreateTapesInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "DescribeSMBSettingsOutput$SMBGuestPasswordSet":"

This value is true if a password for the guest user “smbguest” is set, and otherwise false.

", + "NFSFileShareInfo$ReadOnly":"

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", + "NFSFileShareInfo$GuessMIMETypeEnabled":"

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", + "NFSFileShareInfo$RequesterPays":"

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", + "SMBFileShareInfo$ReadOnly":"

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", + "SMBFileShareInfo$GuessMIMETypeEnabled":"

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", + "SMBFileShareInfo$RequesterPays":"

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", + "UpdateNFSFileShareInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "UpdateNFSFileShareInput$ReadOnly":"

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", + "UpdateNFSFileShareInput$GuessMIMETypeEnabled":"

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", + "UpdateNFSFileShareInput$RequesterPays":"

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

", + "UpdateSMBFileShareInput$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own AWS KMS key, or false to use a key managed by Amazon S3. Optional.

", + "UpdateSMBFileShareInput$ReadOnly":"

Sets the write status of a file share. This value is true if the write status is read-only, and otherwise false.

", + "UpdateSMBFileShareInput$GuessMIMETypeEnabled":"

Enables guessing of the MIME type for uploaded objects based on file extensions. Set this value to true to enable MIME type guessing, and otherwise to false. The default value is true.

", + "UpdateSMBFileShareInput$RequesterPays":"

Sets who pays the cost of the request and the data download from the Amazon S3 bucket. Set this value to true if you want the requester to pay instead of the bucket owner, and otherwise to false.

" } }, - "CachediSCSIVolumes": { - "base": null, - "refs": { - "DescribeCachediSCSIVolumesOutput$CachediSCSIVolumes": "

An array of objects where each object contains metadata about one cached volume.

" + "CachediSCSIVolume":{ + "base":"

Describes an iSCSI cached volume.

", + "refs":{ + "CachediSCSIVolumes$member":null } }, - "CancelArchivalInput": { - "base": "

CancelArchivalInput

", - "refs": { + "CachediSCSIVolumes":{ + "base":null, + "refs":{ + "DescribeCachediSCSIVolumesOutput$CachediSCSIVolumes":"

An array of objects where each object contains metadata about one cached volume.

" } }, - "CancelArchivalOutput": { - "base": "

CancelArchivalOutput

", - "refs": { + "CancelArchivalInput":{ + "base":"

CancelArchivalInput

", + "refs":{ } }, - "CancelRetrievalInput": { - "base": "

CancelRetrievalInput

", - "refs": { + "CancelArchivalOutput":{ + "base":"

CancelArchivalOutput

", + "refs":{ } }, - "CancelRetrievalOutput": { - "base": "

CancelRetrievalOutput

", - "refs": { + "CancelRetrievalInput":{ + "base":"

CancelRetrievalInput

", + "refs":{ } }, - "ChapCredentials": { - "base": null, - "refs": { - "DescribeChapCredentialsOutput$ChapCredentials": "

An array of ChapInfo objects that represent CHAP credentials. Each object in the array contains CHAP credential information for one target-initiator pair. If no CHAP credentials are set, an empty array is returned. CHAP credential information is provided in a JSON object with the following fields:

  • InitiatorName: The iSCSI initiator that connects to the target.

  • SecretToAuthenticateInitiator: The secret key that the initiator (for example, the Windows client) must provide to participate in mutual CHAP with the target.

  • SecretToAuthenticateTarget: The secret key that the target must provide to participate in mutual CHAP with the initiator (e.g. Windows client).

  • TargetARN: The Amazon Resource Name (ARN) of the storage volume.

" + "CancelRetrievalOutput":{ + "base":"

CancelRetrievalOutput

", + "refs":{ } }, - "ChapInfo": { - "base": "

Describes Challenge-Handshake Authentication Protocol (CHAP) information that supports authentication between your gateway and iSCSI initiators.

", - "refs": { - "ChapCredentials$member": null + "ChapCredentials":{ + "base":null, + "refs":{ + "DescribeChapCredentialsOutput$ChapCredentials":"

An array of ChapInfo objects that represent CHAP credentials. Each object in the array contains CHAP credential information for one target-initiator pair. If no CHAP credentials are set, an empty array is returned. CHAP credential information is provided in a JSON object with the following fields:

  • InitiatorName: The iSCSI initiator that connects to the target.

  • SecretToAuthenticateInitiator: The secret key that the initiator (for example, the Windows client) must provide to participate in mutual CHAP with the target.

  • SecretToAuthenticateTarget: The secret key that the target must provide to participate in mutual CHAP with the initiator (e.g. Windows client).

  • TargetARN: The Amazon Resource Name (ARN) of the storage volume.

" } }, - "ChapSecret": { - "base": null, - "refs": { - "ChapInfo$SecretToAuthenticateInitiator": "

The secret key that the initiator (for example, the Windows client) must provide to participate in mutual CHAP with the target.

", - "ChapInfo$SecretToAuthenticateTarget": "

The secret key that the target must provide to participate in mutual CHAP with the initiator (e.g. Windows client).

", - "UpdateChapCredentialsInput$SecretToAuthenticateInitiator": "

The secret key that the initiator (for example, the Windows client) must provide to participate in mutual CHAP with the target.

The secret key must be between 12 and 16 bytes when encoded in UTF-8.

", - "UpdateChapCredentialsInput$SecretToAuthenticateTarget": "

The secret key that the target must provide to participate in mutual CHAP with the initiator (e.g. Windows client).

Byte constraints: Minimum bytes of 12. Maximum bytes of 16.

The secret key must be between 12 and 16 bytes when encoded in UTF-8.

" + "ChapInfo":{ + "base":"

Describes Challenge-Handshake Authentication Protocol (CHAP) information that supports authentication between your gateway and iSCSI initiators.

", + "refs":{ + "ChapCredentials$member":null } }, - "ClientToken": { - "base": null, - "refs": { - "CreateCachediSCSIVolumeInput$ClientToken": null, - "CreateNFSFileShareInput$ClientToken": "

A unique string value that you supply that is used by file gateway to ensure idempotent file share creation.

", - "CreateTapesInput$ClientToken": "

A unique identifier that you use to retry a request. If you retry a request, use the same ClientToken you specified in the initial request.

Using the same ClientToken prevents creating the tape multiple times.

" + "ChapSecret":{ + "base":null, + "refs":{ + "ChapInfo$SecretToAuthenticateInitiator":"

The secret key that the initiator (for example, the Windows client) must provide to participate in mutual CHAP with the target.

", + "ChapInfo$SecretToAuthenticateTarget":"

The secret key that the target must provide to participate in mutual CHAP with the initiator (e.g. Windows client).

", + "UpdateChapCredentialsInput$SecretToAuthenticateInitiator":"

The secret key that the initiator (for example, the Windows client) must provide to participate in mutual CHAP with the target.

The secret key must be between 12 and 16 bytes when encoded in UTF-8.

", + "UpdateChapCredentialsInput$SecretToAuthenticateTarget":"

The secret key that the target must provide to participate in mutual CHAP with the initiator (e.g. Windows client).

Byte constraints: Minimum bytes of 12. Maximum bytes of 16.

The secret key must be between 12 and 16 bytes when encoded in UTF-8.

" } }, - "CreateCachediSCSIVolumeInput": { - "base": null, - "refs": { + "ClientToken":{ + "base":null, + "refs":{ + "CreateCachediSCSIVolumeInput$ClientToken":"

A unique identifier that you use to retry a request. If you retry a request, use the same ClientToken you specified in the initial request.

", + "CreateNFSFileShareInput$ClientToken":"

A unique string value that you supply that is used by file gateway to ensure idempotent file share creation.

", + "CreateSMBFileShareInput$ClientToken":"

A unique string value that you supply that is used by file gateway to ensure idempotent file share creation.

", + "CreateTapesInput$ClientToken":"

A unique identifier that you use to retry a request. If you retry a request, use the same ClientToken you specified in the initial request.

Using the same ClientToken prevents creating the tape multiple times.

" } }, - "CreateCachediSCSIVolumeOutput": { - "base": null, - "refs": { + "CreateCachediSCSIVolumeInput":{ + "base":null, + "refs":{ } }, - "CreateNFSFileShareInput": { - "base": "

CreateNFSFileShareInput

", - "refs": { + "CreateCachediSCSIVolumeOutput":{ + "base":null, + "refs":{ } }, - "CreateNFSFileShareOutput": { - "base": "

CreateNFSFileShareOutput

", - "refs": { + "CreateNFSFileShareInput":{ + "base":"

CreateNFSFileShareInput

", + "refs":{ } }, - "CreateSnapshotFromVolumeRecoveryPointInput": { - "base": null, - "refs": { + "CreateNFSFileShareOutput":{ + "base":"

CreateNFSFileShareOutput

", + "refs":{ } }, - "CreateSnapshotFromVolumeRecoveryPointOutput": { - "base": null, - "refs": { + "CreateSMBFileShareInput":{ + "base":"

CreateSMBFileShareInput

", + "refs":{ } }, - "CreateSnapshotInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "CreateSMBFileShareOutput":{ + "base":"

CreateSMBFileShareOutput

", + "refs":{ } }, - "CreateSnapshotOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "CreateSnapshotFromVolumeRecoveryPointInput":{ + "base":null, + "refs":{ } }, - "CreateStorediSCSIVolumeInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "CreateSnapshotFromVolumeRecoveryPointOutput":{ + "base":null, + "refs":{ } }, - "CreateStorediSCSIVolumeOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "CreateSnapshotInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "CreateTapeWithBarcodeInput": { - "base": "

CreateTapeWithBarcodeInput

", - "refs": { + "CreateSnapshotOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "CreateTapeWithBarcodeOutput": { - "base": "

CreateTapeOutput

", - "refs": { + "CreateStorediSCSIVolumeInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "CreateTapesInput": { - "base": "

CreateTapesInput

", - "refs": { + "CreateStorediSCSIVolumeOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "CreateTapesOutput": { - "base": "

CreateTapeOutput

", - "refs": { + "CreateTapeWithBarcodeInput":{ + "base":"

CreateTapeWithBarcodeInput

", + "refs":{ } }, - "CreatedDate": { - "base": null, - "refs": { - "CachediSCSIVolume$CreatedDate": "

The date the volume was created. Volumes created prior to March 28, 2017 don’t have this time stamp.

", - "StorediSCSIVolume$CreatedDate": "

The date the volume was created. Volumes created prior to March 28, 2017 don’t have this time stamp.

" + "CreateTapeWithBarcodeOutput":{ + "base":"

CreateTapeOutput

", + "refs":{ } }, - "DayOfWeek": { - "base": null, - "refs": { - "DescribeMaintenanceStartTimeOutput$DayOfWeek": "

An ordinal number between 0 and 6 that represents the day of the week, where 0 represents Sunday and 6 represents Saturday. The day of week is in the time zone of the gateway.

", - "UpdateMaintenanceStartTimeInput$DayOfWeek": "

The maintenance start time day of the week represented as an ordinal number from 0 to 6, where 0 represents Sunday and 6 Saturday.

" + "CreateTapesInput":{ + "base":"

CreateTapesInput

", + "refs":{ } }, - "DeleteBandwidthRateLimitInput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "CreateTapesOutput":{ + "base":"

CreateTapeOutput

", + "refs":{ } }, - "DeleteBandwidthRateLimitOutput": { - "base": "

A JSON object containing the of the gateway whose bandwidth rate information was deleted.

", - "refs": { + "CreatedDate":{ + "base":null, + "refs":{ + "CachediSCSIVolume$CreatedDate":"

The date the volume was created. Volumes created prior to March 28, 2017 don’t have this time stamp.

", + "StorediSCSIVolume$CreatedDate":"

The date the volume was created. Volumes created prior to March 28, 2017 don’t have this time stamp.

" } }, - "DeleteChapCredentialsInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "DayOfWeek":{ + "base":null, + "refs":{ + "DescribeMaintenanceStartTimeOutput$DayOfWeek":"

An ordinal number between 0 and 6 that represents the day of the week, where 0 represents Sunday and 6 represents Saturday. The day of week is in the time zone of the gateway.

", + "UpdateMaintenanceStartTimeInput$DayOfWeek":"

The maintenance start time day of the week represented as an ordinal number from 0 to 6, where 0 represents Sunday and 6 Saturday.

" } }, - "DeleteChapCredentialsOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "DeleteBandwidthRateLimitInput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DeleteFileShareInput": { - "base": "

DeleteFileShareInput

", - "refs": { + "DeleteBandwidthRateLimitOutput":{ + "base":"

A JSON object containing the of the gateway whose bandwidth rate information was deleted.

", + "refs":{ } }, - "DeleteFileShareOutput": { - "base": "

DeleteFileShareOutput

", - "refs": { + "DeleteChapCredentialsInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "DeleteGatewayInput": { - "base": "

A JSON object containing the ID of the gateway to delete.

", - "refs": { + "DeleteChapCredentialsOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DeleteGatewayOutput": { - "base": "

A JSON object containing the ID of the deleted gateway.

", - "refs": { + "DeleteFileShareInput":{ + "base":"

DeleteFileShareInput

", + "refs":{ } }, - "DeleteSnapshotScheduleInput": { - "base": null, - "refs": { + "DeleteFileShareOutput":{ + "base":"

DeleteFileShareOutput

", + "refs":{ } }, - "DeleteSnapshotScheduleOutput": { - "base": null, - "refs": { + "DeleteGatewayInput":{ + "base":"

A JSON object containing the ID of the gateway to delete.

", + "refs":{ } }, - "DeleteTapeArchiveInput": { - "base": "

DeleteTapeArchiveInput

", - "refs": { + "DeleteGatewayOutput":{ + "base":"

A JSON object containing the ID of the deleted gateway.

", + "refs":{ } }, - "DeleteTapeArchiveOutput": { - "base": "

DeleteTapeArchiveOutput

", - "refs": { + "DeleteSnapshotScheduleInput":{ + "base":null, + "refs":{ } }, - "DeleteTapeInput": { - "base": "

DeleteTapeInput

", - "refs": { + "DeleteSnapshotScheduleOutput":{ + "base":null, + "refs":{ } }, - "DeleteTapeOutput": { - "base": "

DeleteTapeOutput

", - "refs": { + "DeleteTapeArchiveInput":{ + "base":"

DeleteTapeArchiveInput

", + "refs":{ } }, - "DeleteVolumeInput": { - "base": "

A JSON object containing the DeleteVolumeInput$VolumeARN to delete.

", - "refs": { + "DeleteTapeArchiveOutput":{ + "base":"

DeleteTapeArchiveOutput

", + "refs":{ } }, - "DeleteVolumeOutput": { - "base": "

A JSON object containing the of the storage volume that was deleted

", - "refs": { + "DeleteTapeInput":{ + "base":"

DeleteTapeInput

", + "refs":{ } }, - "DescribeBandwidthRateLimitInput": { - "base": "

A JSON object containing the of the gateway.

", - "refs": { + "DeleteTapeOutput":{ + "base":"

DeleteTapeOutput

", + "refs":{ } }, - "DescribeBandwidthRateLimitOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "DeleteVolumeInput":{ + "base":"

A JSON object containing the DeleteVolumeInput$VolumeARN to delete.

", + "refs":{ } }, - "DescribeCacheInput": { - "base": null, - "refs": { + "DeleteVolumeOutput":{ + "base":"

A JSON object containing the of the storage volume that was deleted

", + "refs":{ } }, - "DescribeCacheOutput": { - "base": null, - "refs": { + "DescribeBandwidthRateLimitInput":{ + "base":"

A JSON object containing the of the gateway.

", + "refs":{ } }, - "DescribeCachediSCSIVolumesInput": { - "base": null, - "refs": { + "DescribeBandwidthRateLimitOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DescribeCachediSCSIVolumesOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "DescribeCacheInput":{ + "base":null, + "refs":{ } }, - "DescribeChapCredentialsInput": { - "base": "

A JSON object containing the Amazon Resource Name (ARN) of the iSCSI volume target.

", - "refs": { + "DescribeCacheOutput":{ + "base":null, + "refs":{ } }, - "DescribeChapCredentialsOutput": { - "base": "

A JSON object containing a .

", - "refs": { + "DescribeCachediSCSIVolumesInput":{ + "base":null, + "refs":{ } }, - "DescribeGatewayInformationInput": { - "base": "

A JSON object containing the ID of the gateway.

", - "refs": { + "DescribeCachediSCSIVolumesOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DescribeGatewayInformationOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "DescribeChapCredentialsInput":{ + "base":"

A JSON object containing the Amazon Resource Name (ARN) of the iSCSI volume target.

", + "refs":{ } }, - "DescribeMaintenanceStartTimeInput": { - "base": "

A JSON object containing the of the gateway.

", - "refs": { + "DescribeChapCredentialsOutput":{ + "base":"

A JSON object containing a .

", + "refs":{ } }, - "DescribeMaintenanceStartTimeOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "DescribeGatewayInformationInput":{ + "base":"

A JSON object containing the ID of the gateway.

", + "refs":{ } }, - "DescribeNFSFileSharesInput": { - "base": "

DescribeNFSFileSharesInput

", - "refs": { + "DescribeGatewayInformationOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DescribeNFSFileSharesOutput": { - "base": "

DescribeNFSFileSharesOutput

", - "refs": { + "DescribeMaintenanceStartTimeInput":{ + "base":"

A JSON object containing the of the gateway.

", + "refs":{ } }, - "DescribeSnapshotScheduleInput": { - "base": "

A JSON object containing the DescribeSnapshotScheduleInput$VolumeARN of the volume.

", - "refs": { + "DescribeMaintenanceStartTimeOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DescribeSnapshotScheduleOutput": { - "base": null, - "refs": { + "DescribeNFSFileSharesInput":{ + "base":"

DescribeNFSFileSharesInput

", + "refs":{ } }, - "DescribeStorediSCSIVolumesInput": { - "base": "

A JSON object containing a list of DescribeStorediSCSIVolumesInput$VolumeARNs.

", - "refs": { + "DescribeNFSFileSharesOutput":{ + "base":"

DescribeNFSFileSharesOutput

", + "refs":{ } }, - "DescribeStorediSCSIVolumesOutput": { - "base": null, - "refs": { + "DescribeSMBFileSharesInput":{ + "base":"

DescribeSMBFileSharesInput

", + "refs":{ } }, - "DescribeTapeArchivesInput": { - "base": "

DescribeTapeArchivesInput

", - "refs": { + "DescribeSMBFileSharesOutput":{ + "base":"

DescribeSMBFileSharesOutput

", + "refs":{ } }, - "DescribeTapeArchivesOutput": { - "base": "

DescribeTapeArchivesOutput

", - "refs": { + "DescribeSMBSettingsInput":{ + "base":null, + "refs":{ } }, - "DescribeTapeRecoveryPointsInput": { - "base": "

DescribeTapeRecoveryPointsInput

", - "refs": { + "DescribeSMBSettingsOutput":{ + "base":null, + "refs":{ } }, - "DescribeTapeRecoveryPointsOutput": { - "base": "

DescribeTapeRecoveryPointsOutput

", - "refs": { + "DescribeSnapshotScheduleInput":{ + "base":"

A JSON object containing the DescribeSnapshotScheduleInput$VolumeARN of the volume.

", + "refs":{ } }, - "DescribeTapesInput": { - "base": "

DescribeTapesInput

", - "refs": { + "DescribeSnapshotScheduleOutput":{ + "base":null, + "refs":{ } }, - "DescribeTapesOutput": { - "base": "

DescribeTapesOutput

", - "refs": { + "DescribeStorediSCSIVolumesInput":{ + "base":"

A JSON object containing a list of DescribeStorediSCSIVolumesInput$VolumeARNs.

", + "refs":{ } }, - "DescribeUploadBufferInput": { - "base": null, - "refs": { + "DescribeStorediSCSIVolumesOutput":{ + "base":null, + "refs":{ } }, - "DescribeUploadBufferOutput": { - "base": null, - "refs": { + "DescribeTapeArchivesInput":{ + "base":"

DescribeTapeArchivesInput

", + "refs":{ } }, - "DescribeVTLDevicesInput": { - "base": "

DescribeVTLDevicesInput

", - "refs": { + "DescribeTapeArchivesOutput":{ + "base":"

DescribeTapeArchivesOutput

", + "refs":{ } }, - "DescribeVTLDevicesOutput": { - "base": "

DescribeVTLDevicesOutput

", - "refs": { + "DescribeTapeRecoveryPointsInput":{ + "base":"

DescribeTapeRecoveryPointsInput

", + "refs":{ } }, - "DescribeWorkingStorageInput": { - "base": "

A JSON object containing the of the gateway.

", - "refs": { + "DescribeTapeRecoveryPointsOutput":{ + "base":"

DescribeTapeRecoveryPointsOutput

", + "refs":{ } }, - "DescribeWorkingStorageOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "DescribeTapesInput":{ + "base":"

DescribeTapesInput

", + "refs":{ } }, - "Description": { - "base": null, - "refs": { - "DescribeSnapshotScheduleOutput$Description": null, - "UpdateSnapshotScheduleInput$Description": "

Optional description of the snapshot that overwrites the existing description.

" + "DescribeTapesOutput":{ + "base":"

DescribeTapesOutput

", + "refs":{ } }, - "DeviceType": { - "base": null, - "refs": { - "UpdateVTLDeviceTypeInput$DeviceType": "

The type of medium changer you want to select.

Valid Values: \"STK-L700\", \"AWS-Gateway-VTL\"

" + "DescribeUploadBufferInput":{ + "base":null, + "refs":{ } }, - "DeviceiSCSIAttributes": { - "base": "

Lists iSCSI information about a VTL device.

", - "refs": { - "VTLDevice$DeviceiSCSIAttributes": "

A list of iSCSI information about a VTL device.

" + "DescribeUploadBufferOutput":{ + "base":null, + "refs":{ } }, - "DisableGatewayInput": { - "base": "

DisableGatewayInput

", - "refs": { + "DescribeVTLDevicesInput":{ + "base":"

DescribeVTLDevicesInput

", + "refs":{ } }, - "DisableGatewayOutput": { - "base": "

DisableGatewayOutput

", - "refs": { + "DescribeVTLDevicesOutput":{ + "base":"

DescribeVTLDevicesOutput

", + "refs":{ } }, - "Disk": { - "base": null, - "refs": { - "Disks$member": null + "DescribeWorkingStorageInput":{ + "base":"

A JSON object containing the of the gateway.

", + "refs":{ } }, - "DiskAllocationType": { - "base": null, - "refs": { - "Disk$DiskAllocationType": null + "DescribeWorkingStorageOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "DiskId": { - "base": null, - "refs": { - "CreateStorediSCSIVolumeInput$DiskId": "

The unique identifier for the gateway local disk that is configured as a stored volume. Use ListLocalDisks to list disk IDs for a gateway.

", - "Disk$DiskId": null, - "DiskIds$member": null, - "StorediSCSIVolume$VolumeDiskId": "

The ID of the local disk that was specified in the CreateStorediSCSIVolume operation.

" + "Description":{ + "base":null, + "refs":{ + "DescribeSnapshotScheduleOutput$Description":null, + "UpdateSnapshotScheduleInput$Description":"

Optional description of the snapshot that overwrites the existing description.

" } }, - "DiskIds": { - "base": null, - "refs": { - "AddCacheInput$DiskIds": null, - "AddUploadBufferInput$DiskIds": null, - "AddWorkingStorageInput$DiskIds": "

An array of strings that identify disks that are to be configured as working storage. Each string have a minimum length of 1 and maximum length of 300. You can get the disk IDs from the ListLocalDisks API.

", - "DescribeCacheOutput$DiskIds": null, - "DescribeUploadBufferOutput$DiskIds": null, - "DescribeWorkingStorageOutput$DiskIds": "

An array of the gateway's local disk IDs that are configured as working storage. Each local disk ID is specified as a string (minimum length of 1 and maximum length of 300). If no local disks are configured as working storage, then the DiskIds array is empty.

" + "DeviceType":{ + "base":null, + "refs":{ + "UpdateVTLDeviceTypeInput$DeviceType":"

The type of medium changer you want to select.

Valid Values: \"STK-L700\", \"AWS-Gateway-VTL\"

" } }, - "Disks": { - "base": null, - "refs": { - "ListLocalDisksOutput$Disks": null + "DeviceiSCSIAttributes":{ + "base":"

Lists iSCSI information about a VTL device.

", + "refs":{ + "VTLDevice$DeviceiSCSIAttributes":"

A list of iSCSI information about a VTL device.

" } }, - "DoubleObject": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeProgress": "

Represents the percentage complete if the volume is restoring or bootstrapping that represents the percent of data transferred. This field does not appear in the response if the cached volume is not restoring or bootstrapping.

", - "StorediSCSIVolume$VolumeProgress": "

Represents the percentage complete if the volume is restoring or bootstrapping that represents the percent of data transferred. This field does not appear in the response if the stored volume is not restoring or bootstrapping.

", - "Tape$Progress": "

For archiving virtual tapes, indicates how much data remains to be uploaded before archiving is complete.

Range: 0 (not started) to 100 (complete).

" + "DisableGatewayInput":{ + "base":"

DisableGatewayInput

", + "refs":{ } - }, - "ErrorCode": { - "base": null, - "refs": { - "StorageGatewayError$errorCode": "

Additional information about the error.

" - } - }, - "FileShareARN": { - "base": "

The Amazon Resource Name (ARN) of the file share.

", - "refs": { - "CreateNFSFileShareOutput$FileShareARN": "

The Amazon Resource Name (ARN) of the newly created file share.

", - "DeleteFileShareInput$FileShareARN": "

The Amazon Resource Name (ARN) of the file share to be deleted.

", - "DeleteFileShareOutput$FileShareARN": "

The Amazon Resource Name (ARN) of the deleted file share.

", - "FileShareARNList$member": null, - "FileShareInfo$FileShareARN": null, - "NFSFileShareInfo$FileShareARN": null, - "NotifyWhenUploadedInput$FileShareARN": null, - "NotifyWhenUploadedOutput$FileShareARN": null, - "RefreshCacheInput$FileShareARN": null, - "RefreshCacheOutput$FileShareARN": null, - "UpdateNFSFileShareInput$FileShareARN": "

The Amazon Resource Name (ARN) of the file share to be updated.

", - "UpdateNFSFileShareOutput$FileShareARN": "

The Amazon Resource Name (ARN) of the updated file share.

" - } - }, - "FileShareARNList": { - "base": null, - "refs": { - "DescribeNFSFileSharesInput$FileShareARNList": "

An array containing the Amazon Resource Name (ARN) of each file share to be described.

" - } - }, - "FileShareClientList": { - "base": "

The list of clients that are allowed to access the file gateway. The list must contain either valid IP addresses or valid CIDR blocks.

", - "refs": { - "CreateNFSFileShareInput$ClientList": "

The list of clients that are allowed to access the file gateway. The list must contain either valid IP addresses or valid CIDR blocks.

", - "NFSFileShareInfo$ClientList": null, - "UpdateNFSFileShareInput$ClientList": "

The list of clients that are allowed to access the file gateway. The list must contain either valid IP addresses or valid CIDR blocks.

" - } - }, - "FileShareId": { - "base": "

The ID of the file share.

", - "refs": { - "FileShareInfo$FileShareId": null, - "NFSFileShareInfo$FileShareId": null - } - }, - "FileShareInfo": { - "base": "

Describes a file share.

", - "refs": { - "FileShareInfoList$member": null - } - }, - "FileShareInfoList": { - "base": null, - "refs": { - "ListFileSharesOutput$FileShareInfoList": "

An array of information about the file gateway's file shares.

" - } - }, - "FileShareStatus": { - "base": "

The status of the file share. Possible values are CREATING, UPDATING, AVAILABLE and DELETING.

", - "refs": { - "FileShareInfo$FileShareStatus": null, - "NFSFileShareInfo$FileShareStatus": null - } - }, - "GatewayARN": { - "base": "

The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation to return a list of gateways for your account and region.

", - "refs": { - "ActivateGatewayOutput$GatewayARN": null, - "AddCacheInput$GatewayARN": null, - "AddCacheOutput$GatewayARN": null, - "AddUploadBufferInput$GatewayARN": null, - "AddUploadBufferOutput$GatewayARN": null, - "AddWorkingStorageInput$GatewayARN": null, - "AddWorkingStorageOutput$GatewayARN": null, - "CancelArchivalInput$GatewayARN": null, - "CancelRetrievalInput$GatewayARN": null, - "CreateCachediSCSIVolumeInput$GatewayARN": null, - "CreateNFSFileShareInput$GatewayARN": "

The Amazon Resource Name (ARN) of the file gateway on which you want to create a file share.

", - "CreateStorediSCSIVolumeInput$GatewayARN": null, - "CreateTapeWithBarcodeInput$GatewayARN": "

The unique Amazon Resource Name (ARN) that represents the gateway to associate the virtual tape with. Use the ListGateways operation to return a list of gateways for your account and region.

", - "CreateTapesInput$GatewayARN": "

The unique Amazon Resource Name (ARN) that represents the gateway to associate the virtual tapes with. Use the ListGateways operation to return a list of gateways for your account and region.

", - "DeleteBandwidthRateLimitInput$GatewayARN": null, - "DeleteBandwidthRateLimitOutput$GatewayARN": null, - "DeleteGatewayInput$GatewayARN": null, - "DeleteGatewayOutput$GatewayARN": null, - "DeleteTapeInput$GatewayARN": "

The unique Amazon Resource Name (ARN) of the gateway that the virtual tape to delete is associated with. Use the ListGateways operation to return a list of gateways for your account and region.

", - "DescribeBandwidthRateLimitInput$GatewayARN": null, - "DescribeBandwidthRateLimitOutput$GatewayARN": null, - "DescribeCacheInput$GatewayARN": null, - "DescribeCacheOutput$GatewayARN": null, - "DescribeGatewayInformationInput$GatewayARN": null, - "DescribeGatewayInformationOutput$GatewayARN": null, - "DescribeMaintenanceStartTimeInput$GatewayARN": null, - "DescribeMaintenanceStartTimeOutput$GatewayARN": null, - "DescribeTapeRecoveryPointsInput$GatewayARN": null, - "DescribeTapeRecoveryPointsOutput$GatewayARN": null, - "DescribeTapesInput$GatewayARN": null, - "DescribeUploadBufferInput$GatewayARN": null, - "DescribeUploadBufferOutput$GatewayARN": null, - "DescribeVTLDevicesInput$GatewayARN": null, - "DescribeVTLDevicesOutput$GatewayARN": null, - "DescribeWorkingStorageInput$GatewayARN": null, - "DescribeWorkingStorageOutput$GatewayARN": null, - "DisableGatewayInput$GatewayARN": null, - "DisableGatewayOutput$GatewayARN": "

The unique Amazon Resource Name of the disabled gateway.

", - "FileShareInfo$GatewayARN": null, - "GatewayInfo$GatewayARN": "

The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation to return a list of gateways for your account and region.

", - "ListFileSharesInput$GatewayARN": "

The Amazon resource Name (ARN) of the gateway whose file shares you want to list. If this field is not present, all file shares under your account are listed.

", - "ListLocalDisksInput$GatewayARN": null, - "ListLocalDisksOutput$GatewayARN": null, - "ListVolumeRecoveryPointsInput$GatewayARN": null, - "ListVolumeRecoveryPointsOutput$GatewayARN": null, - "ListVolumesInput$GatewayARN": null, - "ListVolumesOutput$GatewayARN": null, - "NFSFileShareInfo$GatewayARN": null, - "ResetCacheInput$GatewayARN": null, - "ResetCacheOutput$GatewayARN": null, - "RetrieveTapeArchiveInput$GatewayARN": "

The Amazon Resource Name (ARN) of the gateway you want to retrieve the virtual tape to. Use the ListGateways operation to return a list of gateways for your account and region.

You retrieve archived virtual tapes to only one gateway and the gateway must be a tape gateway.

", - "RetrieveTapeRecoveryPointInput$GatewayARN": null, - "SetLocalConsolePasswordInput$GatewayARN": null, - "SetLocalConsolePasswordOutput$GatewayARN": null, - "ShutdownGatewayInput$GatewayARN": null, - "ShutdownGatewayOutput$GatewayARN": null, - "StartGatewayInput$GatewayARN": null, - "StartGatewayOutput$GatewayARN": null, - "TapeArchive$RetrievedTo": "

The Amazon Resource Name (ARN) of the tape gateway that the virtual tape is being retrieved to.

The virtual tape is retrieved from the virtual tape shelf (VTS).

", - "TapeInfo$GatewayARN": "

The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation to return a list of gateways for your account and region.

", - "UpdateBandwidthRateLimitInput$GatewayARN": null, - "UpdateBandwidthRateLimitOutput$GatewayARN": null, - "UpdateGatewayInformationInput$GatewayARN": null, - "UpdateGatewayInformationOutput$GatewayARN": null, - "UpdateGatewaySoftwareNowInput$GatewayARN": null, - "UpdateGatewaySoftwareNowOutput$GatewayARN": null, - "UpdateMaintenanceStartTimeInput$GatewayARN": null, - "UpdateMaintenanceStartTimeOutput$GatewayARN": null, - "VolumeInfo$GatewayARN": null - } - }, - "GatewayId": { - "base": null, - "refs": { - "DescribeGatewayInformationOutput$GatewayId": "

The unique identifier assigned to your gateway during activation. This ID becomes part of the gateway Amazon Resource Name (ARN), which you use as input for other operations.

", - "GatewayInfo$GatewayId": "

The unique identifier assigned to your gateway during activation. This ID becomes part of the gateway Amazon Resource Name (ARN), which you use as input for other operations.

", - "VolumeInfo$GatewayId": "

The unique identifier assigned to your gateway during activation. This ID becomes part of the gateway Amazon Resource Name (ARN), which you use as input for other operations.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

" - } - }, - "GatewayInfo": { - "base": "

Describes a gateway object.

", - "refs": { - "Gateways$member": null - } - }, - "GatewayName": { - "base": "

The name you configured for your gateway.

", - "refs": { - "ActivateGatewayInput$GatewayName": "

The name you configured for your gateway.

", - "UpdateGatewayInformationInput$GatewayName": null - } - }, - "GatewayNetworkInterfaces": { - "base": null, - "refs": { - "DescribeGatewayInformationOutput$GatewayNetworkInterfaces": "

A NetworkInterface array that contains descriptions of the gateway network interfaces.

" - } - }, - "GatewayOperationalState": { - "base": null, - "refs": { - "GatewayInfo$GatewayOperationalState": "

The state of the gateway.

Valid Values: DISABLED or ACTIVE

" - } - }, - "GatewayState": { - "base": null, - "refs": { - "DescribeGatewayInformationOutput$GatewayState": "

A value that indicates the operating state of the gateway.

" - } - }, - "GatewayTimezone": { - "base": null, - "refs": { - "ActivateGatewayInput$GatewayTimezone": "

A value that indicates the time zone you want to set for the gateway. The time zone is of the format \"GMT-hr:mm\" or \"GMT+hr:mm\". For example, GMT-4:00 indicates the time is 4 hours behind GMT. GMT+2:00 indicates the time is 2 hours ahead of GMT. The time zone is used, for example, for scheduling snapshots and your gateway's maintenance schedule.

", - "DescribeGatewayInformationOutput$GatewayTimezone": "

A value that indicates the time zone configured for the gateway.

", - "DescribeMaintenanceStartTimeOutput$Timezone": null, - "DescribeSnapshotScheduleOutput$Timezone": null, - "UpdateGatewayInformationInput$GatewayTimezone": null - } - }, - "GatewayType": { - "base": null, - "refs": { - "ActivateGatewayInput$GatewayType": "

A value that defines the type of gateway to activate. The type specified is critical to all later functions of the gateway and cannot be changed after activation. The default value is STORED.

Valid Values: \"STORED\", \"CACHED\", \"VTL\", \"FILE_S3\"

", - "DescribeGatewayInformationOutput$GatewayType": "

The type of the gateway.

", - "GatewayInfo$GatewayType": "

The type of the gateway.

" + }, + "DisableGatewayOutput":{ + "base":"

DisableGatewayOutput

", + "refs":{ + } + }, + "Disk":{ + "base":null, + "refs":{ + "Disks$member":null + } + }, + "DiskAllocationType":{ + "base":null, + "refs":{ + "Disk$DiskAllocationType":null + } + }, + "DiskId":{ + "base":null, + "refs":{ + "CreateStorediSCSIVolumeInput$DiskId":"

The unique identifier for the gateway local disk that is configured as a stored volume. Use ListLocalDisks to list disk IDs for a gateway.

", + "Disk$DiskId":null, + "DiskIds$member":null, + "StorediSCSIVolume$VolumeDiskId":"

The ID of the local disk that was specified in the CreateStorediSCSIVolume operation.

" + } + }, + "DiskIds":{ + "base":null, + "refs":{ + "AddCacheInput$DiskIds":null, + "AddUploadBufferInput$DiskIds":null, + "AddWorkingStorageInput$DiskIds":"

An array of strings that identify disks that are to be configured as working storage. Each string have a minimum length of 1 and maximum length of 300. You can get the disk IDs from the ListLocalDisks API.

", + "DescribeCacheOutput$DiskIds":null, + "DescribeUploadBufferOutput$DiskIds":null, + "DescribeWorkingStorageOutput$DiskIds":"

An array of the gateway's local disk IDs that are configured as working storage. Each local disk ID is specified as a string (minimum length of 1 and maximum length of 300). If no local disks are configured as working storage, then the DiskIds array is empty.

" + } + }, + "Disks":{ + "base":null, + "refs":{ + "ListLocalDisksOutput$Disks":null + } + }, + "DomainName":{ + "base":null, + "refs":{ + "DescribeSMBSettingsOutput$DomainName":"

The name of the domain that the gateway is joined to.

", + "JoinDomainInput$DomainName":"

The name of the domain that you want the gateway to join.

" + } + }, + "DomainUserName":{ + "base":null, + "refs":{ + "JoinDomainInput$UserName":"

Sets the user name of user who has permission to add the gateway to the Active Directory domain.

" + } + }, + "DomainUserPassword":{ + "base":null, + "refs":{ + "JoinDomainInput$Password":"

Sets the password of the user who has permission to add the gateway to the Active Directory domain.

" + } + }, + "DoubleObject":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeProgress":"

Represents the percentage complete if the volume is restoring or bootstrapping that represents the percent of data transferred. This field does not appear in the response if the cached volume is not restoring or bootstrapping.

", + "StorediSCSIVolume$VolumeProgress":"

Represents the percentage complete if the volume is restoring or bootstrapping that represents the percent of data transferred. This field does not appear in the response if the stored volume is not restoring or bootstrapping.

", + "Tape$Progress":"

For archiving virtual tapes, indicates how much data remains to be uploaded before archiving is complete.

Range: 0 (not started) to 100 (complete).

" + } + }, + "ErrorCode":{ + "base":null, + "refs":{ + "StorageGatewayError$errorCode":"

Additional information about the error.

" + } + }, + "FileShareARN":{ + "base":"

The Amazon Resource Name (ARN) of the file share.

", + "refs":{ + "CreateNFSFileShareOutput$FileShareARN":"

The Amazon Resource Name (ARN) of the newly created file share.

", + "CreateSMBFileShareOutput$FileShareARN":"

The Amazon Resource Name (ARN) of the newly created file share.

", + "DeleteFileShareInput$FileShareARN":"

The Amazon Resource Name (ARN) of the file share to be deleted.

", + "DeleteFileShareOutput$FileShareARN":"

The Amazon Resource Name (ARN) of the deleted file share.

", + "FileShareARNList$member":null, + "FileShareInfo$FileShareARN":null, + "NFSFileShareInfo$FileShareARN":null, + "NotifyWhenUploadedInput$FileShareARN":null, + "NotifyWhenUploadedOutput$FileShareARN":null, + "RefreshCacheInput$FileShareARN":null, + "RefreshCacheOutput$FileShareARN":null, + "SMBFileShareInfo$FileShareARN":null, + "UpdateNFSFileShareInput$FileShareARN":"

The Amazon Resource Name (ARN) of the file share to be updated.

", + "UpdateNFSFileShareOutput$FileShareARN":"

The Amazon Resource Name (ARN) of the updated file share.

", + "UpdateSMBFileShareInput$FileShareARN":"

The Amazon Resource Name (ARN) of the SMB file share you want to update.

", + "UpdateSMBFileShareOutput$FileShareARN":"

The Amazon Resource Name (ARN) of the updated SMB file share.

" + } + }, + "FileShareARNList":{ + "base":null, + "refs":{ + "DescribeNFSFileSharesInput$FileShareARNList":"

An array containing the Amazon Resource Name (ARN) of each file share to be described.

", + "DescribeSMBFileSharesInput$FileShareARNList":"

An array containing the Amazon Resource Name (ARN) of each file share to be described.

" + } + }, + "FileShareClientList":{ + "base":"

The list of clients that are allowed to access the file gateway. The list must contain either valid IP addresses or valid CIDR blocks.

", + "refs":{ + "CreateNFSFileShareInput$ClientList":"

The list of clients that are allowed to access the file gateway. The list must contain either valid IP addresses or valid CIDR blocks.

", + "NFSFileShareInfo$ClientList":null, + "UpdateNFSFileShareInput$ClientList":"

The list of clients that are allowed to access the file gateway. The list must contain either valid IP addresses or valid CIDR blocks.

" + } + }, + "FileShareId":{ + "base":"

The ID of the file share.

", + "refs":{ + "FileShareInfo$FileShareId":null, + "NFSFileShareInfo$FileShareId":null, + "SMBFileShareInfo$FileShareId":null + } + }, + "FileShareInfo":{ + "base":"

Describes a file share.

", + "refs":{ + "FileShareInfoList$member":null + } + }, + "FileShareInfoList":{ + "base":null, + "refs":{ + "ListFileSharesOutput$FileShareInfoList":"

An array of information about the file gateway's file shares.

" + } + }, + "FileShareStatus":{ + "base":"

The status of the file share. Possible values are CREATING, UPDATING, AVAILABLE and DELETING.

", + "refs":{ + "FileShareInfo$FileShareStatus":null, + "NFSFileShareInfo$FileShareStatus":null, + "SMBFileShareInfo$FileShareStatus":null + } + }, + "FileShareType":{ + "base":"

The type of the file share.

", + "refs":{ + "FileShareInfo$FileShareType":null + } + }, + "FileShareUser":{ + "base":null, + "refs":{ + "FileShareUserList$member":null + } + }, + "FileShareUserList":{ + "base":null, + "refs":{ + "CreateSMBFileShareInput$ValidUserList":"

A list of users in the Active Directory that are allowed to access the file share. Can only be set if Authentication is set to \"ActiveDirectory\".

", + "CreateSMBFileShareInput$InvalidUserList":"

A list of users in the Active Directory that are not allowed to access the file share. Can only be set if Authentication is set to \"ActiveDirectory\".

", + "SMBFileShareInfo$ValidUserList":"

A list of users in the Active Directory that are allowed to access the file share. Can only be set if Authentication is set to \"ActiveDirectory\".

", + "SMBFileShareInfo$InvalidUserList":"

A list of users in the Active Directory that are not allowed to access the file share. Can only be set if Authentication is set to \"ActiveDirectory\".

", + "UpdateSMBFileShareInput$ValidUserList":"

A list of users in the Active Directory that are allowed to access the file share. Can only be set if Authentication is set to \"ActiveDirectory\".

", + "UpdateSMBFileShareInput$InvalidUserList":"

A list of users in the Active Directory that are not allowed to access the file share. Can only be set if Authentication is set to \"ActiveDirectory\".

" + } + }, + "GatewayARN":{ + "base":"

The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation to return a list of gateways for your account and region.

", + "refs":{ + "ActivateGatewayOutput$GatewayARN":null, + "AddCacheInput$GatewayARN":null, + "AddCacheOutput$GatewayARN":null, + "AddUploadBufferInput$GatewayARN":null, + "AddUploadBufferOutput$GatewayARN":null, + "AddWorkingStorageInput$GatewayARN":null, + "AddWorkingStorageOutput$GatewayARN":null, + "CancelArchivalInput$GatewayARN":null, + "CancelRetrievalInput$GatewayARN":null, + "CreateCachediSCSIVolumeInput$GatewayARN":null, + "CreateNFSFileShareInput$GatewayARN":"

The Amazon Resource Name (ARN) of the file gateway on which you want to create a file share.

", + "CreateSMBFileShareInput$GatewayARN":"

The Amazon Resource Name (ARN) of the file gateway on which you want to create a file share.

", + "CreateStorediSCSIVolumeInput$GatewayARN":null, + "CreateTapeWithBarcodeInput$GatewayARN":"

The unique Amazon Resource Name (ARN) that represents the gateway to associate the virtual tape with. Use the ListGateways operation to return a list of gateways for your account and region.

", + "CreateTapesInput$GatewayARN":"

The unique Amazon Resource Name (ARN) that represents the gateway to associate the virtual tapes with. Use the ListGateways operation to return a list of gateways for your account and region.

", + "DeleteBandwidthRateLimitInput$GatewayARN":null, + "DeleteBandwidthRateLimitOutput$GatewayARN":null, + "DeleteGatewayInput$GatewayARN":null, + "DeleteGatewayOutput$GatewayARN":null, + "DeleteTapeInput$GatewayARN":"

The unique Amazon Resource Name (ARN) of the gateway that the virtual tape to delete is associated with. Use the ListGateways operation to return a list of gateways for your account and region.

", + "DescribeBandwidthRateLimitInput$GatewayARN":null, + "DescribeBandwidthRateLimitOutput$GatewayARN":null, + "DescribeCacheInput$GatewayARN":null, + "DescribeCacheOutput$GatewayARN":null, + "DescribeGatewayInformationInput$GatewayARN":null, + "DescribeGatewayInformationOutput$GatewayARN":null, + "DescribeMaintenanceStartTimeInput$GatewayARN":null, + "DescribeMaintenanceStartTimeOutput$GatewayARN":null, + "DescribeSMBSettingsInput$GatewayARN":null, + "DescribeSMBSettingsOutput$GatewayARN":null, + "DescribeTapeRecoveryPointsInput$GatewayARN":null, + "DescribeTapeRecoveryPointsOutput$GatewayARN":null, + "DescribeTapesInput$GatewayARN":null, + "DescribeUploadBufferInput$GatewayARN":null, + "DescribeUploadBufferOutput$GatewayARN":null, + "DescribeVTLDevicesInput$GatewayARN":null, + "DescribeVTLDevicesOutput$GatewayARN":null, + "DescribeWorkingStorageInput$GatewayARN":null, + "DescribeWorkingStorageOutput$GatewayARN":null, + "DisableGatewayInput$GatewayARN":null, + "DisableGatewayOutput$GatewayARN":"

The unique Amazon Resource Name of the disabled gateway.

", + "FileShareInfo$GatewayARN":null, + "GatewayInfo$GatewayARN":"

The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation to return a list of gateways for your account and region.

", + "JoinDomainInput$GatewayARN":"

The unique Amazon Resource Name of the file gateway you want to add to the Active Directory domain.

", + "JoinDomainOutput$GatewayARN":"

The unique Amazon Resource Name of the gateway that joined the domain.

", + "ListFileSharesInput$GatewayARN":"

The Amazon resource Name (ARN) of the gateway whose file shares you want to list. If this field is not present, all file shares under your account are listed.

", + "ListLocalDisksInput$GatewayARN":null, + "ListLocalDisksOutput$GatewayARN":null, + "ListVolumeRecoveryPointsInput$GatewayARN":null, + "ListVolumeRecoveryPointsOutput$GatewayARN":null, + "ListVolumesInput$GatewayARN":null, + "ListVolumesOutput$GatewayARN":null, + "NFSFileShareInfo$GatewayARN":null, + "ResetCacheInput$GatewayARN":null, + "ResetCacheOutput$GatewayARN":null, + "RetrieveTapeArchiveInput$GatewayARN":"

The Amazon Resource Name (ARN) of the gateway you want to retrieve the virtual tape to. Use the ListGateways operation to return a list of gateways for your account and region.

You retrieve archived virtual tapes to only one gateway and the gateway must be a tape gateway.

", + "RetrieveTapeRecoveryPointInput$GatewayARN":null, + "SMBFileShareInfo$GatewayARN":null, + "SetLocalConsolePasswordInput$GatewayARN":null, + "SetLocalConsolePasswordOutput$GatewayARN":null, + "SetSMBGuestPasswordInput$GatewayARN":"

The Amazon Resource Name (ARN) of the file gateway the SMB file share is associated with.

", + "SetSMBGuestPasswordOutput$GatewayARN":null, + "ShutdownGatewayInput$GatewayARN":null, + "ShutdownGatewayOutput$GatewayARN":null, + "StartGatewayInput$GatewayARN":null, + "StartGatewayOutput$GatewayARN":null, + "TapeArchive$RetrievedTo":"

The Amazon Resource Name (ARN) of the tape gateway that the virtual tape is being retrieved to.

The virtual tape is retrieved from the virtual tape shelf (VTS).

", + "TapeInfo$GatewayARN":"

The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation to return a list of gateways for your account and region.

", + "UpdateBandwidthRateLimitInput$GatewayARN":null, + "UpdateBandwidthRateLimitOutput$GatewayARN":null, + "UpdateGatewayInformationInput$GatewayARN":null, + "UpdateGatewayInformationOutput$GatewayARN":null, + "UpdateGatewaySoftwareNowInput$GatewayARN":null, + "UpdateGatewaySoftwareNowOutput$GatewayARN":null, + "UpdateMaintenanceStartTimeInput$GatewayARN":null, + "UpdateMaintenanceStartTimeOutput$GatewayARN":null, + "VolumeInfo$GatewayARN":null + } + }, + "GatewayId":{ + "base":null, + "refs":{ + "DescribeGatewayInformationOutput$GatewayId":"

The unique identifier assigned to your gateway during activation. This ID becomes part of the gateway Amazon Resource Name (ARN), which you use as input for other operations.

", + "GatewayInfo$GatewayId":"

The unique identifier assigned to your gateway during activation. This ID becomes part of the gateway Amazon Resource Name (ARN), which you use as input for other operations.

", + "VolumeInfo$GatewayId":"

The unique identifier assigned to your gateway during activation. This ID becomes part of the gateway Amazon Resource Name (ARN), which you use as input for other operations.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

" + } + }, + "GatewayInfo":{ + "base":"

Describes a gateway object.

", + "refs":{ + "Gateways$member":null + } + }, + "GatewayName":{ + "base":"

The name you configured for your gateway.

", + "refs":{ + "ActivateGatewayInput$GatewayName":"

The name you configured for your gateway.

", + "UpdateGatewayInformationInput$GatewayName":null + } + }, + "GatewayNetworkInterfaces":{ + "base":null, + "refs":{ + "DescribeGatewayInformationOutput$GatewayNetworkInterfaces":"

A NetworkInterface array that contains descriptions of the gateway network interfaces.

" + } + }, + "GatewayOperationalState":{ + "base":null, + "refs":{ + "GatewayInfo$GatewayOperationalState":"

The state of the gateway.

Valid Values: DISABLED or ACTIVE

" + } + }, + "GatewayState":{ + "base":null, + "refs":{ + "DescribeGatewayInformationOutput$GatewayState":"

A value that indicates the operating state of the gateway.

" + } + }, + "GatewayTimezone":{ + "base":null, + "refs":{ + "ActivateGatewayInput$GatewayTimezone":"

A value that indicates the time zone you want to set for the gateway. The time zone is of the format \"GMT-hr:mm\" or \"GMT+hr:mm\". For example, GMT-4:00 indicates the time is 4 hours behind GMT. GMT+2:00 indicates the time is 2 hours ahead of GMT. The time zone is used, for example, for scheduling snapshots and your gateway's maintenance schedule.

", + "DescribeGatewayInformationOutput$GatewayTimezone":"

A value that indicates the time zone configured for the gateway.

", + "DescribeMaintenanceStartTimeOutput$Timezone":null, + "DescribeSnapshotScheduleOutput$Timezone":null, + "UpdateGatewayInformationInput$GatewayTimezone":null + } + }, + "GatewayType":{ + "base":null, + "refs":{ + "ActivateGatewayInput$GatewayType":"

A value that defines the type of gateway to activate. The type specified is critical to all later functions of the gateway and cannot be changed after activation. The default value is STORED.

Valid Values: \"STORED\", \"CACHED\", \"VTL\", \"FILE_S3\"

", + "DescribeGatewayInformationOutput$GatewayType":"

The type of the gateway.

", + "GatewayInfo$GatewayType":"

The type of the gateway.

" + } + }, + "Gateways":{ + "base":null, + "refs":{ + "ListGatewaysOutput$Gateways":null + } + }, + "HourOfDay":{ + "base":null, + "refs":{ + "DescribeMaintenanceStartTimeOutput$HourOfDay":"

The hour component of the maintenance start time represented as hh, where hh is the hour (0 to 23). The hour of the day is in the time zone of the gateway.

", + "DescribeSnapshotScheduleOutput$StartAt":null, + "UpdateMaintenanceStartTimeInput$HourOfDay":"

The hour component of the maintenance start time represented as hh, where hh is the hour (00 to 23). The hour of the day is in the time zone of the gateway.

", + "UpdateSnapshotScheduleInput$StartAt":"

The hour of the day at which the snapshot schedule begins represented as hh, where hh is the hour (0 to 23). The hour of the day is in the time zone of the gateway.

" + } + }, + "IPV4AddressCIDR":{ + "base":null, + "refs":{ + "FileShareClientList$member":null + } + }, + "Initiator":{ + "base":null, + "refs":{ + "Initiators$member":null + } + }, + "Initiators":{ + "base":null, + "refs":{ + "ListVolumeInitiatorsOutput$Initiators":"

The host names and port numbers of all iSCSI initiators that are connected to the gateway.

" + } + }, + "InternalServerError":{ + "base":"

An internal server error has occurred during the request. For more information, see the error and message fields.

", + "refs":{ + } + }, + "InvalidGatewayRequestException":{ + "base":"

An exception occurred because an invalid gateway request was issued to the service. For more information, see the error and message fields.

", + "refs":{ + } + }, + "IqnName":{ + "base":null, + "refs":{ + "ChapInfo$InitiatorName":"

The iSCSI initiator that connects to the target.

", + "DeleteChapCredentialsInput$InitiatorName":"

The iSCSI initiator that connects to the target.

", + "DeleteChapCredentialsOutput$InitiatorName":"

The iSCSI initiator that connects to the target.

", + "UpdateChapCredentialsInput$InitiatorName":"

The iSCSI initiator that connects to the target.

", + "UpdateChapCredentialsOutput$InitiatorName":"

The iSCSI initiator that connects to the target. This is the same initiator name specified in the request.

" + } + }, + "JoinDomainInput":{ + "base":"

JoinDomainInput

", + "refs":{ } }, - "Gateways": { - "base": null, - "refs": { - "ListGatewaysOutput$Gateways": null + "JoinDomainOutput":{ + "base":"

JoinDomainOutput

", + "refs":{ } }, - "HourOfDay": { - "base": null, - "refs": { - "DescribeMaintenanceStartTimeOutput$HourOfDay": "

The hour component of the maintenance start time represented as hh, where hh is the hour (0 to 23). The hour of the day is in the time zone of the gateway.

", - "DescribeSnapshotScheduleOutput$StartAt": null, - "UpdateMaintenanceStartTimeInput$HourOfDay": "

The hour component of the maintenance start time represented as hh, where hh is the hour (00 to 23). The hour of the day is in the time zone of the gateway.

", - "UpdateSnapshotScheduleInput$StartAt": "

The hour of the day at which the snapshot schedule begins represented as hh, where hh is the hour (0 to 23). The hour of the day is in the time zone of the gateway.

" + "KMSKey":{ + "base":"

The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "refs":{ + "CachediSCSIVolume$KMSKey":null, + "CreateCachediSCSIVolumeInput$KMSKey":"

The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "CreateNFSFileShareInput$KMSKey":"

The Amazon Resource Name (ARN) KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "CreateSMBFileShareInput$KMSKey":"

The Amazon Resource Name (ARN) KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "CreateTapeWithBarcodeInput$KMSKey":"

The Amazon Resource Name (ARN) of the KMS Key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "CreateTapesInput$KMSKey":"

The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "NFSFileShareInfo$KMSKey":null, + "SMBFileShareInfo$KMSKey":null, + "Tape$KMSKey":null, + "TapeArchive$KMSKey":null, + "UpdateNFSFileShareInput$KMSKey":"

The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

", + "UpdateSMBFileShareInput$KMSKey":"

The Amazon Resource Name (ARN) KMS key used for Amazon S3 server side encryption. This value can only be set when KMSEncrypted is true. Optional.

" } }, - "IPV4AddressCIDR": { - "base": null, - "refs": { - "FileShareClientList$member": null + "LastSoftwareUpdate":{ + "base":null, + "refs":{ + "DescribeGatewayInformationOutput$LastSoftwareUpdate":"

The date on which the last software update was applied to the gateway. If the gateway has never been updated, this field does not return a value in the response.

" } }, - "Initiator": { - "base": null, - "refs": { - "Initiators$member": null + "ListFileSharesInput":{ + "base":"

ListFileShareInput

", + "refs":{ } }, - "Initiators": { - "base": null, - "refs": { - "ListVolumeInitiatorsOutput$Initiators": "

The host names and port numbers of all iSCSI initiators that are connected to the gateway.

" + "ListFileSharesOutput":{ + "base":"

ListFileShareOutput

", + "refs":{ } }, - "InternalServerError": { - "base": "

An internal server error has occurred during the request. For more information, see the error and message fields.

", - "refs": { + "ListGatewaysInput":{ + "base":"

A JSON object containing zero or more of the following fields:

", + "refs":{ } }, - "InvalidGatewayRequestException": { - "base": "

An exception occurred because an invalid gateway request was issued to the service. For more information, see the error and message fields.

", - "refs": { + "ListGatewaysOutput":{ + "base":null, + "refs":{ } }, - "IqnName": { - "base": null, - "refs": { - "ChapInfo$InitiatorName": "

The iSCSI initiator that connects to the target.

", - "DeleteChapCredentialsInput$InitiatorName": "

The iSCSI initiator that connects to the target.

", - "DeleteChapCredentialsOutput$InitiatorName": "

The iSCSI initiator that connects to the target.

", - "UpdateChapCredentialsInput$InitiatorName": "

The iSCSI initiator that connects to the target.

", - "UpdateChapCredentialsOutput$InitiatorName": "

The iSCSI initiator that connects to the target. This is the same initiator name specified in the request.

" + "ListLocalDisksInput":{ + "base":"

A JSON object containing the of the gateway.

", + "refs":{ } }, - "KMSKey": { - "base": "

The ARN of the KMS key used for Amazon S3 server side encryption.

", - "refs": { - "CreateNFSFileShareInput$KMSKey": "

The KMS key used for Amazon S3 server side encryption. This value can only be set when KmsEncrypted is true. Optional.

", - "NFSFileShareInfo$KMSKey": null, - "UpdateNFSFileShareInput$KMSKey": "

The KMS key used for Amazon S3 server side encryption. This value can only be set when KmsEncrypted is true. Optional.

" + "ListLocalDisksOutput":{ + "base":null, + "refs":{ } }, - "LastSoftwareUpdate": { - "base": null, - "refs": { - "DescribeGatewayInformationOutput$LastSoftwareUpdate": "

The date on which the last software update was applied to the gateway. If the gateway has never been updated, this field does not return a value in the response.

" + "ListTagsForResourceInput":{ + "base":"

ListTagsForResourceInput

", + "refs":{ } }, - "ListFileSharesInput": { - "base": "

ListFileShareInput

", - "refs": { + "ListTagsForResourceOutput":{ + "base":"

ListTagsForResourceOutput

", + "refs":{ } }, - "ListFileSharesOutput": { - "base": "

ListFileShareOutput

", - "refs": { + "ListTapesInput":{ + "base":"

A JSON object that contains one or more of the following fields:

", + "refs":{ } }, - "ListGatewaysInput": { - "base": "

A JSON object containing zero or more of the following fields:

", - "refs": { + "ListTapesOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "ListGatewaysOutput": { - "base": null, - "refs": { + "ListVolumeInitiatorsInput":{ + "base":"

ListVolumeInitiatorsInput

", + "refs":{ } }, - "ListLocalDisksInput": { - "base": "

A JSON object containing the of the gateway.

", - "refs": { + "ListVolumeInitiatorsOutput":{ + "base":"

ListVolumeInitiatorsOutput

", + "refs":{ } }, - "ListLocalDisksOutput": { - "base": null, - "refs": { + "ListVolumeRecoveryPointsInput":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceInput": { - "base": "

ListTagsForResourceInput

", - "refs": { + "ListVolumeRecoveryPointsOutput":{ + "base":null, + "refs":{ } }, - "ListTagsForResourceOutput": { - "base": "

ListTagsForResourceOutput

", - "refs": { + "ListVolumesInput":{ + "base":"

A JSON object that contains one or more of the following fields:

", + "refs":{ } }, - "ListTapesInput": { - "base": "

A JSON object that contains one or more of the following fields:

", - "refs": { + "ListVolumesOutput":{ + "base":null, + "refs":{ } }, - "ListTapesOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "LocalConsolePassword":{ + "base":null, + "refs":{ + "SetLocalConsolePasswordInput$LocalConsolePassword":"

The password you want to set for your VM local console.

" } }, - "ListVolumeInitiatorsInput": { - "base": "

ListVolumeInitiatorsInput

", - "refs": { + "LocationARN":{ + "base":"

The ARN of the backend storage used for storing file data.

", + "refs":{ + "CreateNFSFileShareInput$LocationARN":"

The ARN of the backed storage used for storing file data.

", + "CreateSMBFileShareInput$LocationARN":"

The ARN of the backed storage used for storing file data.

", + "NFSFileShareInfo$LocationARN":null, + "SMBFileShareInfo$LocationARN":null } }, - "ListVolumeInitiatorsOutput": { - "base": "

ListVolumeInitiatorsOutput

", - "refs": { + "Marker":{ + "base":null, + "refs":{ + "DescribeTapeArchivesInput$Marker":"

An opaque string that indicates the position at which to begin describing virtual tapes.

", + "DescribeTapeArchivesOutput$Marker":"

An opaque string that indicates the position at which the virtual tapes that were fetched for description ended. Use this marker in your next request to fetch the next set of virtual tapes in the virtual tape shelf (VTS). If there are no more virtual tapes to describe, this field does not appear in the response.

", + "DescribeTapeRecoveryPointsInput$Marker":"

An opaque string that indicates the position at which to begin describing the virtual tape recovery points.

", + "DescribeTapeRecoveryPointsOutput$Marker":"

An opaque string that indicates the position at which the virtual tape recovery points that were listed for description ended.

Use this marker in your next request to list the next set of virtual tape recovery points in the list. If there are no more recovery points to describe, this field does not appear in the response.

", + "DescribeTapesInput$Marker":"

A marker value, obtained in a previous call to DescribeTapes. This marker indicates which page of results to retrieve.

If not specified, the first page of results is retrieved.

", + "DescribeTapesOutput$Marker":"

An opaque string which can be used as part of a subsequent DescribeTapes call to retrieve the next page of results.

If a response does not contain a marker, then there are no more results to be retrieved.

", + "DescribeVTLDevicesInput$Marker":"

An opaque string that indicates the position at which to begin describing the VTL devices.

", + "DescribeVTLDevicesOutput$Marker":"

An opaque string that indicates the position at which the VTL devices that were fetched for description ended. Use the marker in your next request to fetch the next set of VTL devices in the list. If there are no more VTL devices to describe, this field does not appear in the response.

", + "ListFileSharesInput$Marker":"

Opaque pagination token returned from a previous ListFileShares operation. If present, Marker specifies where to continue the list from after a previous call to ListFileShares. Optional.

", + "ListFileSharesOutput$Marker":"

If the request includes Marker, the response returns that value in this field.

", + "ListFileSharesOutput$NextMarker":"

If a value is present, there are more file shares to return. In a subsequent request, use NextMarker as the value for Marker to retrieve the next set of file shares.

", + "ListGatewaysInput$Marker":"

An opaque string that indicates the position at which to begin the returned list of gateways.

", + "ListGatewaysOutput$Marker":null, + "ListTagsForResourceInput$Marker":"

An opaque string that indicates the position at which to begin returning the list of tags.

", + "ListTagsForResourceOutput$Marker":"

An opaque string that indicates the position at which to stop returning the list of tags.

", + "ListTapesInput$Marker":"

A string that indicates the position at which to begin the returned list of tapes.

", + "ListTapesOutput$Marker":"

A string that indicates the position at which to begin returning the next list of tapes. Use the marker in your next request to continue pagination of tapes. If there are no more tapes to list, this element does not appear in the response body.

", + "ListVolumesInput$Marker":"

A string that indicates the position at which to begin the returned list of volumes. Obtain the marker from the response of a previous List iSCSI Volumes request.

", + "ListVolumesOutput$Marker":null } }, - "ListVolumeRecoveryPointsInput": { - "base": null, - "refs": { + "MediumChangerType":{ + "base":null, + "refs":{ + "ActivateGatewayInput$MediumChangerType":"

The value that indicates the type of medium changer to use for tape gateway. This field is optional.

Valid Values: \"STK-L700\", \"AWS-Gateway-VTL\"

" } }, - "ListVolumeRecoveryPointsOutput": { - "base": null, - "refs": { + "MinuteOfHour":{ + "base":null, + "refs":{ + "DescribeMaintenanceStartTimeOutput$MinuteOfHour":"

The minute component of the maintenance start time represented as mm, where mm is the minute (0 to 59). The minute of the hour is in the time zone of the gateway.

", + "UpdateMaintenanceStartTimeInput$MinuteOfHour":"

The minute component of the maintenance start time represented as mm, where mm is the minute (00 to 59). The minute of the hour is in the time zone of the gateway.

" } }, - "ListVolumesInput": { - "base": "

A JSON object that contains one or more of the following fields:

", - "refs": { + "NFSFileShareDefaults":{ + "base":"

Describes Network File System (NFS) file share default values. Files and folders stored as Amazon S3 objects in S3 buckets don't, by default, have Unix file permissions assigned to them. Upon discovery in an S3 bucket by Storage Gateway, the S3 objects that represent files and folders are assigned these default Unix permissions. This operation is only supported in the file gateway type.

", + "refs":{ + "CreateNFSFileShareInput$NFSFileShareDefaults":"

File share default values. Optional.

", + "NFSFileShareInfo$NFSFileShareDefaults":null, + "UpdateNFSFileShareInput$NFSFileShareDefaults":"

The default values for the file share. Optional.

" } }, - "ListVolumesOutput": { - "base": null, - "refs": { + "NFSFileShareInfo":{ + "base":"

The Unix file permissions and ownership information assigned, by default, to native S3 objects when file gateway discovers them in S3 buckets. This operation is only supported in file gateways.

", + "refs":{ + "NFSFileShareInfoList$member":null } }, - "LocalConsolePassword": { - "base": null, - "refs": { - "SetLocalConsolePasswordInput$LocalConsolePassword": "

The password you want to set for your VM local console.

" + "NFSFileShareInfoList":{ + "base":null, + "refs":{ + "DescribeNFSFileSharesOutput$NFSFileShareInfoList":"

An array containing a description for each requested file share.

" } }, - "LocationARN": { - "base": "

The ARN of the backend storage used for storing file data.

", - "refs": { - "CreateNFSFileShareInput$LocationARN": "

The ARN of the backed storage used for storing file data.

", - "NFSFileShareInfo$LocationARN": null + "NetworkInterface":{ + "base":"

Describes a gateway's network interface.

", + "refs":{ + "GatewayNetworkInterfaces$member":null } }, - "Marker": { - "base": null, - "refs": { - "DescribeTapeArchivesInput$Marker": "

An opaque string that indicates the position at which to begin describing virtual tapes.

", - "DescribeTapeArchivesOutput$Marker": "

An opaque string that indicates the position at which the virtual tapes that were fetched for description ended. Use this marker in your next request to fetch the next set of virtual tapes in the virtual tape shelf (VTS). If there are no more virtual tapes to describe, this field does not appear in the response.

", - "DescribeTapeRecoveryPointsInput$Marker": "

An opaque string that indicates the position at which to begin describing the virtual tape recovery points.

", - "DescribeTapeRecoveryPointsOutput$Marker": "

An opaque string that indicates the position at which the virtual tape recovery points that were listed for description ended.

Use this marker in your next request to list the next set of virtual tape recovery points in the list. If there are no more recovery points to describe, this field does not appear in the response.

", - "DescribeTapesInput$Marker": "

A marker value, obtained in a previous call to DescribeTapes. This marker indicates which page of results to retrieve.

If not specified, the first page of results is retrieved.

", - "DescribeTapesOutput$Marker": "

An opaque string which can be used as part of a subsequent DescribeTapes call to retrieve the next page of results.

If a response does not contain a marker, then there are no more results to be retrieved.

", - "DescribeVTLDevicesInput$Marker": "

An opaque string that indicates the position at which to begin describing the VTL devices.

", - "DescribeVTLDevicesOutput$Marker": "

An opaque string that indicates the position at which the VTL devices that were fetched for description ended. Use the marker in your next request to fetch the next set of VTL devices in the list. If there are no more VTL devices to describe, this field does not appear in the response.

", - "ListFileSharesInput$Marker": "

Opaque pagination token returned from a previous ListFileShares operation. If present, Marker specifies where to continue the list from after a previous call to ListFileShares. Optional.

", - "ListFileSharesOutput$Marker": "

If the request includes Marker, the response returns that value in this field.

", - "ListFileSharesOutput$NextMarker": "

If a value is present, there are more file shares to return. In a subsequent request, use NextMarker as the value for Marker to retrieve the next set of file shares.

", - "ListGatewaysInput$Marker": "

An opaque string that indicates the position at which to begin the returned list of gateways.

", - "ListGatewaysOutput$Marker": null, - "ListTagsForResourceInput$Marker": "

An opaque string that indicates the position at which to begin returning the list of tags.

", - "ListTagsForResourceOutput$Marker": "

An opaque string that indicates the position at which to stop returning the list of tags.

", - "ListTapesInput$Marker": "

A string that indicates the position at which to begin the returned list of tapes.

", - "ListTapesOutput$Marker": "

A string that indicates the position at which to begin returning the next list of tapes. Use the marker in your next request to continue pagination of tapes. If there are no more tapes to list, this element does not appear in the response body.

", - "ListVolumesInput$Marker": "

A string that indicates the position at which to begin the returned list of volumes. Obtain the marker from the response of a previous List iSCSI Volumes request.

", - "ListVolumesOutput$Marker": null + "NetworkInterfaceId":{ + "base":null, + "refs":{ + "CreateCachediSCSIVolumeInput$NetworkInterfaceId":"

The network interface of the gateway on which to expose the iSCSI target. Only IPv4 addresses are accepted. Use DescribeGatewayInformation to get a list of the network interfaces available on a gateway.

Valid Values: A valid IP address.

", + "CreateStorediSCSIVolumeInput$NetworkInterfaceId":"

The network interface of the gateway on which to expose the iSCSI target. Only IPv4 addresses are accepted. Use DescribeGatewayInformation to get a list of the network interfaces available on a gateway.

Valid Values: A valid IP address.

", + "DeviceiSCSIAttributes$NetworkInterfaceId":"

The network interface identifier of the VTL device.

", + "VolumeiSCSIAttributes$NetworkInterfaceId":"

The network interface identifier.

" } }, - "MediumChangerType": { - "base": null, - "refs": { - "ActivateGatewayInput$MediumChangerType": "

The value that indicates the type of medium changer to use for tape gateway. This field is optional.

Valid Values: \"STK-L700\", \"AWS-Gateway-VTL\"

" + "NextUpdateAvailabilityDate":{ + "base":null, + "refs":{ + "DescribeGatewayInformationOutput$NextUpdateAvailabilityDate":"

The date on which an update to the gateway is available. This date is in the time zone of the gateway. If the gateway is not available for an update this field is not returned in the response.

" } }, - "MinuteOfHour": { - "base": null, - "refs": { - "DescribeMaintenanceStartTimeOutput$MinuteOfHour": "

The minute component of the maintenance start time represented as mm, where mm is the minute (0 to 59). The minute of the hour is in the time zone of the gateway.

", - "UpdateMaintenanceStartTimeInput$MinuteOfHour": "

The minute component of the maintenance start time represented as mm, where mm is the minute (00 to 59). The minute of the hour is in the time zone of the gateway.

" + "NotificationId":{ + "base":"

The randomly generated ID of the notification that was sent. This ID is in UUID format.

", + "refs":{ + "NotifyWhenUploadedOutput$NotificationId":null } }, - "NFSFileShareDefaults": { - "base": "

Describes file share default values. Files and folders stored as Amazon S3 objects in S3 buckets don't, by default, have Unix file permissions assigned to them. Upon discovery in an S3 bucket by Storage Gateway, the S3 objects that represent files and folders are assigned these default Unix permissions. This operation is only supported in the file gateway type.

", - "refs": { - "CreateNFSFileShareInput$NFSFileShareDefaults": "

File share default values. Optional.

", - "NFSFileShareInfo$NFSFileShareDefaults": null, - "UpdateNFSFileShareInput$NFSFileShareDefaults": "

The default values for the file share. Optional.

" + "NotifyWhenUploadedInput":{ + "base":null, + "refs":{ } }, - "NFSFileShareInfo": { - "base": "

The Unix file permissions and ownership information assigned, by default, to native S3 objects when file gateway discovers them in S3 buckets. This operation is only supported in file gateways.

", - "refs": { - "NFSFileShareInfoList$member": null + "NotifyWhenUploadedOutput":{ + "base":null, + "refs":{ } }, - "NFSFileShareInfoList": { - "base": null, - "refs": { - "DescribeNFSFileSharesOutput$NFSFileShareInfoList": "

An array containing a description for each requested file share.

" + "NumTapesToCreate":{ + "base":null, + "refs":{ + "CreateTapesInput$NumTapesToCreate":"

The number of virtual tapes that you want to create.

" } }, - "NetworkInterface": { - "base": "

Describes a gateway's network interface.

", - "refs": { - "GatewayNetworkInterfaces$member": null + "ObjectACL":{ + "base":"

Sets the access control list permission for objects in the S3 bucket that a file gateway puts objects into. The default value is \"private\".

", + "refs":{ + "CreateNFSFileShareInput$ObjectACL":"

Sets the access control list permission for objects in the Amazon S3 bucket that a file gateway puts objects into. The default value is \"private\".

", + "CreateSMBFileShareInput$ObjectACL":"

Sets the access control list permission for objects in the Amazon S3 bucket that a file gateway puts objects into. The default value is \"private\".

", + "NFSFileShareInfo$ObjectACL":null, + "SMBFileShareInfo$ObjectACL":null, + "UpdateNFSFileShareInput$ObjectACL":"

Sets the access control list permission for objects in the S3 bucket that a file gateway puts objects into. The default value is \"private\".

", + "UpdateSMBFileShareInput$ObjectACL":"

Sets the access control list permission for objects in the Amazon S3 bucket that a file gateway puts objects into. The default value is \"private\".

" } }, - "NetworkInterfaceId": { - "base": null, - "refs": { - "CreateCachediSCSIVolumeInput$NetworkInterfaceId": null, - "CreateStorediSCSIVolumeInput$NetworkInterfaceId": "

The network interface of the gateway on which to expose the iSCSI target. Only IPv4 addresses are accepted. Use DescribeGatewayInformation to get a list of the network interfaces available on a gateway.

Valid Values: A valid IP address.

", - "DeviceiSCSIAttributes$NetworkInterfaceId": "

The network interface identifier of the VTL device.

", - "VolumeiSCSIAttributes$NetworkInterfaceId": "

The network interface identifier.

" + "Path":{ + "base":"

The file share path used by the NFS client to identify the mount point.

", + "refs":{ + "NFSFileShareInfo$Path":null, + "SMBFileShareInfo$Path":"

The file share path used by the SMB client to identify the mount point.

" } }, - "NextUpdateAvailabilityDate": { - "base": null, - "refs": { - "DescribeGatewayInformationOutput$NextUpdateAvailabilityDate": "

The date on which an update to the gateway is available. This date is in the time zone of the gateway. If the gateway is not available for an update this field is not returned in the response.

" + "PermissionId":{ + "base":null, + "refs":{ + "NFSFileShareDefaults$GroupId":"

The default group ID for the file share (unless the files have another group ID specified). The default value is nfsnobody.

", + "NFSFileShareDefaults$OwnerId":"

The default owner ID for files in the file share (unless the files have another owner ID specified). The default value is nfsnobody.

" } }, - "NotificationId": { - "base": "

The randomly generated ID of the notification that was sent. This ID is in UUID format.

", - "refs": { - "NotifyWhenUploadedOutput$NotificationId": null + "PermissionMode":{ + "base":null, + "refs":{ + "NFSFileShareDefaults$FileMode":"

The Unix file mode in the form \"nnnn\". For example, \"0666\" represents the default file mode inside the file share. The default value is 0666.

", + "NFSFileShareDefaults$DirectoryMode":"

The Unix directory mode in the form \"nnnn\". For example, \"0666\" represents the default access mode for all directories inside the file share. The default value is 0777.

" } }, - "NotifyWhenUploadedInput": { - "base": null, - "refs": { + "PositiveIntObject":{ + "base":null, + "refs":{ + "DescribeTapeArchivesInput$Limit":"

Specifies that the number of virtual tapes descried be limited to the specified number.

", + "DescribeTapeRecoveryPointsInput$Limit":"

Specifies that the number of virtual tape recovery points that are described be limited to the specified number.

", + "DescribeTapesInput$Limit":"

Specifies that the number of virtual tapes described be limited to the specified number.

Amazon Web Services may impose its own limit, if this field is not set.

", + "DescribeVTLDevicesInput$Limit":"

Specifies that the number of VTL devices described be limited to the specified number.

", + "ListFileSharesInput$Limit":"

The maximum number of file shares to return in the response. The value must be an integer with a value greater than zero. Optional.

", + "ListGatewaysInput$Limit":"

Specifies that the list of gateways returned be limited to the specified number of items.

", + "ListTagsForResourceInput$Limit":"

Specifies that the list of tags returned be limited to the specified number of items.

", + "ListTapesInput$Limit":"

An optional number limit for the tapes in the list returned by this call.

", + "ListVolumesInput$Limit":"

Specifies that the list of volumes returned be limited to the specified number of items.

", + "VolumeiSCSIAttributes$LunNumber":"

The logical disk number.

" } }, - "NotifyWhenUploadedOutput": { - "base": null, - "refs": { + "RecurrenceInHours":{ + "base":null, + "refs":{ + "DescribeSnapshotScheduleOutput$RecurrenceInHours":null, + "UpdateSnapshotScheduleInput$RecurrenceInHours":"

Frequency of snapshots. Specify the number of hours between snapshots.

" } }, - "NumTapesToCreate": { - "base": null, - "refs": { - "CreateTapesInput$NumTapesToCreate": "

The number of virtual tapes that you want to create.

" + "RefreshCacheInput":{ + "base":null, + "refs":{ } }, - "ObjectACL": { - "base": "

Sets the access control list permission for objects in the S3 bucket that a file gateway puts objects into. The default value is \"private\".

", - "refs": { - "CreateNFSFileShareInput$ObjectACL": "

Sets the access control list permission for objects in the Amazon S3 bucket that a file gateway puts objects into. The default value is \"private\".

", - "NFSFileShareInfo$ObjectACL": null, - "UpdateNFSFileShareInput$ObjectACL": "

Sets the access control list permission for objects in the S3 bucket that a file gateway puts objects into. The default value is \"private\".

" + "RefreshCacheOutput":{ + "base":null, + "refs":{ } }, - "Path": { - "base": "

The file share path used by the NFS client to identify the mount point.

", - "refs": { - "NFSFileShareInfo$Path": null + "RegionId":{ + "base":null, + "refs":{ + "ActivateGatewayInput$GatewayRegion":"

A value that indicates the region where you want to store your data. The gateway region specified must be the same region as the region in your Host header in the request. For more information about available regions and endpoints for AWS Storage Gateway, see Regions and Endpoints in the Amazon Web Services Glossary.

Valid Values: \"us-east-1\", \"us-east-2\", \"us-west-1\", \"us-west-2\", \"ca-central-1\", \"eu-west-1\", \"eu-central-1\", \"eu-west-2\", \"eu-west-3\", \"ap-northeast-1\", \"ap-northeast-2\", \"ap-southeast-1\", \"ap-southeast-2\", \"ap-south-1\", \"sa-east-1\"

" } }, - "PermissionId": { - "base": null, - "refs": { - "NFSFileShareDefaults$GroupId": "

The default group ID for the file share (unless the files have another group ID specified). The default value is nfsnobody.

", - "NFSFileShareDefaults$OwnerId": "

The default owner ID for files in the file share (unless the files have another owner ID specified). The default value is nfsnobody.

" + "RemoveTagsFromResourceInput":{ + "base":"

RemoveTagsFromResourceInput

", + "refs":{ } }, - "PermissionMode": { - "base": null, - "refs": { - "NFSFileShareDefaults$FileMode": "

The Unix file mode in the form \"nnnn\". For example, \"0666\" represents the default file mode inside the file share. The default value is 0666.

", - "NFSFileShareDefaults$DirectoryMode": "

The Unix directory mode in the form \"nnnn\". For example, \"0666\" represents the default access mode for all directories inside the file share. The default value is 0777.

" + "RemoveTagsFromResourceOutput":{ + "base":"

RemoveTagsFromResourceOutput

", + "refs":{ } }, - "PositiveIntObject": { - "base": null, - "refs": { - "DescribeTapeArchivesInput$Limit": "

Specifies that the number of virtual tapes descried be limited to the specified number.

", - "DescribeTapeRecoveryPointsInput$Limit": "

Specifies that the number of virtual tape recovery points that are described be limited to the specified number.

", - "DescribeTapesInput$Limit": "

Specifies that the number of virtual tapes described be limited to the specified number.

Amazon Web Services may impose its own limit, if this field is not set.

", - "DescribeVTLDevicesInput$Limit": "

Specifies that the number of VTL devices described be limited to the specified number.

", - "ListFileSharesInput$Limit": "

The maximum number of file shares to return in the response. The value must be an integer with a value greater than zero. Optional.

", - "ListGatewaysInput$Limit": "

Specifies that the list of gateways returned be limited to the specified number of items.

", - "ListTagsForResourceInput$Limit": "

Specifies that the list of tags returned be limited to the specified number of items.

", - "ListTapesInput$Limit": "

An optional number limit for the tapes in the list returned by this call.

", - "ListVolumesInput$Limit": "

Specifies that the list of volumes returned be limited to the specified number of items.

", - "VolumeiSCSIAttributes$LunNumber": "

The logical disk number.

" + "ResetCacheInput":{ + "base":null, + "refs":{ } }, - "RecurrenceInHours": { - "base": null, - "refs": { - "DescribeSnapshotScheduleOutput$RecurrenceInHours": null, - "UpdateSnapshotScheduleInput$RecurrenceInHours": "

Frequency of snapshots. Specify the number of hours between snapshots.

" + "ResetCacheOutput":{ + "base":null, + "refs":{ } }, - "RefreshCacheInput": { - "base": null, - "refs": { + "ResourceARN":{ + "base":null, + "refs":{ + "AddTagsToResourceInput$ResourceARN":"

The Amazon Resource Name (ARN) of the resource you want to add tags to.

", + "AddTagsToResourceOutput$ResourceARN":"

The Amazon Resource Name (ARN) of the resource you want to add tags to.

", + "ListTagsForResourceInput$ResourceARN":"

The Amazon Resource Name (ARN) of the resource for which you want to list tags.

", + "ListTagsForResourceOutput$ResourceARN":"

he Amazon Resource Name (ARN) of the resource for which you want to list tags.

", + "RemoveTagsFromResourceInput$ResourceARN":"

The Amazon Resource Name (ARN) of the resource you want to remove the tags from.

", + "RemoveTagsFromResourceOutput$ResourceARN":"

The Amazon Resource Name (ARN) of the resource that the tags were removed from.

" } }, - "RefreshCacheOutput": { - "base": null, - "refs": { + "RetrieveTapeArchiveInput":{ + "base":"

RetrieveTapeArchiveInput

", + "refs":{ } }, - "RegionId": { - "base": null, - "refs": { - "ActivateGatewayInput$GatewayRegion": "

A value that indicates the region where you want to store your data. The gateway region specified must be the same region as the region in your Host header in the request. For more information about available regions and endpoints for AWS Storage Gateway, see Regions and Endpoints in the Amazon Web Services Glossary.

Valid Values: \"us-east-1\", \"us-east-2\", \"us-west-1\", \"us-west-2\", \"ca-central-1\", \"eu-west-1\", \"eu-central-1\", \"eu-west-2\", \"eu-west-3\", \"ap-northeast-1\", \"ap-northeast-2\", \"ap-southeast-1\", \"ap-southeast-2\", \"ap-south-1\", \"sa-east-1\"

" + "RetrieveTapeArchiveOutput":{ + "base":"

RetrieveTapeArchiveOutput

", + "refs":{ } }, - "RemoveTagsFromResourceInput": { - "base": "

RemoveTagsFromResourceInput

", - "refs": { + "RetrieveTapeRecoveryPointInput":{ + "base":"

RetrieveTapeRecoveryPointInput

", + "refs":{ } }, - "RemoveTagsFromResourceOutput": { - "base": "

RemoveTagsFromResourceOutput

", - "refs": { + "RetrieveTapeRecoveryPointOutput":{ + "base":"

RetrieveTapeRecoveryPointOutput

", + "refs":{ } }, - "ResetCacheInput": { - "base": null, - "refs": { + "Role":{ + "base":"

The ARN of the IAM role that file gateway assumes when it accesses the underlying storage.

", + "refs":{ + "CreateNFSFileShareInput$Role":"

The ARN of the AWS Identity and Access Management (IAM) role that a file gateway assumes when it accesses the underlying storage.

", + "CreateSMBFileShareInput$Role":"

The ARN of the AWS Identity and Access Management (IAM) role that a file gateway assumes when it accesses the underlying storage.

", + "NFSFileShareInfo$Role":null, + "SMBFileShareInfo$Role":null } }, - "ResetCacheOutput": { - "base": null, - "refs": { + "SMBFileShareInfo":{ + "base":"

The Windows file permissions and ownership information assigned, by default, to native S3 objects when file gateway discovers them in S3 buckets. This operation is only supported in file gateways.

", + "refs":{ + "SMBFileShareInfoList$member":null } }, - "ResourceARN": { - "base": null, - "refs": { - "AddTagsToResourceInput$ResourceARN": "

The Amazon Resource Name (ARN) of the resource you want to add tags to.

", - "AddTagsToResourceOutput$ResourceARN": "

The Amazon Resource Name (ARN) of the resource you want to add tags to.

", - "ListTagsForResourceInput$ResourceARN": "

The Amazon Resource Name (ARN) of the resource for which you want to list tags.

", - "ListTagsForResourceOutput$ResourceARN": "

he Amazon Resource Name (ARN) of the resource for which you want to list tags.

", - "RemoveTagsFromResourceInput$ResourceARN": "

The Amazon Resource Name (ARN) of the resource you want to remove the tags from.

", - "RemoveTagsFromResourceOutput$ResourceARN": "

The Amazon Resource Name (ARN) of the resource that the tags were removed from.

" + "SMBFileShareInfoList":{ + "base":null, + "refs":{ + "DescribeSMBFileSharesOutput$SMBFileShareInfoList":"

An array containing a description for each requested file share.

" } }, - "RetrieveTapeArchiveInput": { - "base": "

RetrieveTapeArchiveInput

", - "refs": { + "SMBGuestPassword":{ + "base":null, + "refs":{ + "SetSMBGuestPasswordInput$Password":"

The password you want to set for your SMB Server.

" } }, - "RetrieveTapeArchiveOutput": { - "base": "

RetrieveTapeArchiveOutput

", - "refs": { + "ServiceUnavailableError":{ + "base":"

An internal server error has occurred because the service is unavailable. For more information, see the error and message fields.

", + "refs":{ } }, - "RetrieveTapeRecoveryPointInput": { - "base": "

RetrieveTapeRecoveryPointInput

", - "refs": { + "SetLocalConsolePasswordInput":{ + "base":"

SetLocalConsolePasswordInput

", + "refs":{ } }, - "RetrieveTapeRecoveryPointOutput": { - "base": "

RetrieveTapeRecoveryPointOutput

", - "refs": { + "SetLocalConsolePasswordOutput":{ + "base":null, + "refs":{ } }, - "Role": { - "base": "

The ARN of the IAM role that file gateway assumes when it accesses the underlying storage.

", - "refs": { - "CreateNFSFileShareInput$Role": "

The ARN of the AWS Identity and Access Management (IAM) role that a file gateway assumes when it accesses the underlying storage.

", - "NFSFileShareInfo$Role": null + "SetSMBGuestPasswordInput":{ + "base":"

SetSMBGuestPasswordInput

", + "refs":{ } }, - "ServiceUnavailableError": { - "base": "

An internal server error has occurred because the service is unavailable. For more information, see the error and message fields.

", - "refs": { + "SetSMBGuestPasswordOutput":{ + "base":null, + "refs":{ } }, - "SetLocalConsolePasswordInput": { - "base": "

SetLocalConsolePasswordInput

", - "refs": { + "ShutdownGatewayInput":{ + "base":"

A JSON object containing the of the gateway to shut down.

", + "refs":{ } }, - "SetLocalConsolePasswordOutput": { - "base": null, - "refs": { + "ShutdownGatewayOutput":{ + "base":"

A JSON object containing the of the gateway that was shut down.

", + "refs":{ } }, - "ShutdownGatewayInput": { - "base": "

A JSON object containing the of the gateway to shut down.

", - "refs": { + "SnapshotDescription":{ + "base":null, + "refs":{ + "CreateSnapshotFromVolumeRecoveryPointInput$SnapshotDescription":null, + "CreateSnapshotInput$SnapshotDescription":"

Textual description of the snapshot that appears in the Amazon EC2 console, Elastic Block Store snapshots panel in the Description field, and in the AWS Storage Gateway snapshot Details pane, Description field

" } }, - "ShutdownGatewayOutput": { - "base": "

A JSON object containing the of the gateway that was shut down.

", - "refs": { + "SnapshotId":{ + "base":null, + "refs":{ + "CachediSCSIVolume$SourceSnapshotId":"

If the cached volume was created from a snapshot, this field contains the snapshot ID used, e.g. snap-78e22663. Otherwise, this field is not included.

", + "CreateCachediSCSIVolumeInput$SnapshotId":"

The snapshot ID (e.g. \"snap-1122aabb\") of the snapshot to restore as the new cached volume. Specify this field if you want to create the iSCSI storage volume from a snapshot otherwise do not include this field. To list snapshots for your account use DescribeSnapshots in the Amazon Elastic Compute Cloud API Reference.

", + "CreateSnapshotFromVolumeRecoveryPointOutput$SnapshotId":null, + "CreateSnapshotOutput$SnapshotId":"

The snapshot ID that is used to refer to the snapshot in future operations such as describing snapshots (Amazon Elastic Compute Cloud API DescribeSnapshots) or creating a volume from a snapshot (CreateStorediSCSIVolume).

", + "CreateStorediSCSIVolumeInput$SnapshotId":"

The snapshot ID (e.g. \"snap-1122aabb\") of the snapshot to restore as the new stored volume. Specify this field if you want to create the iSCSI storage volume from a snapshot otherwise do not include this field. To list snapshots for your account use DescribeSnapshots in the Amazon Elastic Compute Cloud API Reference.

", + "StorediSCSIVolume$SourceSnapshotId":"

If the stored volume was created from a snapshot, this field contains the snapshot ID used, e.g. snap-78e22663. Otherwise, this field is not included.

" } }, - "SnapshotDescription": { - "base": null, - "refs": { - "CreateSnapshotFromVolumeRecoveryPointInput$SnapshotDescription": null, - "CreateSnapshotInput$SnapshotDescription": "

Textual description of the snapshot that appears in the Amazon EC2 console, Elastic Block Store snapshots panel in the Description field, and in the AWS Storage Gateway snapshot Details pane, Description field

" + "Squash":{ + "base":"

The user mapped to anonymous user. Valid options are the following:

  • \"RootSquash\" - Only root is mapped to anonymous user.

  • \"NoSquash\" - No one is mapped to anonymous user

  • \"AllSquash\" - Everyone is mapped to anonymous user.

", + "refs":{ + "CreateNFSFileShareInput$Squash":"

Maps a user to anonymous user. Valid options are the following:

  • \"RootSquash\" - Only root is mapped to anonymous user.

  • \"NoSquash\" - No one is mapped to anonymous user.

  • \"AllSquash\" - Everyone is mapped to anonymous user.

", + "NFSFileShareInfo$Squash":null, + "UpdateNFSFileShareInput$Squash":"

The user mapped to anonymous user. Valid options are the following:

  • \"RootSquash\" - Only root is mapped to anonymous user.

  • \"NoSquash\" - No one is mapped to anonymous user

  • \"AllSquash\" - Everyone is mapped to anonymous user.

" } }, - "SnapshotId": { - "base": null, - "refs": { - "CachediSCSIVolume$SourceSnapshotId": "

If the cached volume was created from a snapshot, this field contains the snapshot ID used, e.g. snap-78e22663. Otherwise, this field is not included.

", - "CreateCachediSCSIVolumeInput$SnapshotId": null, - "CreateSnapshotFromVolumeRecoveryPointOutput$SnapshotId": null, - "CreateSnapshotOutput$SnapshotId": "

The snapshot ID that is used to refer to the snapshot in future operations such as describing snapshots (Amazon Elastic Compute Cloud API DescribeSnapshots) or creating a volume from a snapshot (CreateStorediSCSIVolume).

", - "CreateStorediSCSIVolumeInput$SnapshotId": "

The snapshot ID (e.g. \"snap-1122aabb\") of the snapshot to restore as the new stored volume. Specify this field if you want to create the iSCSI storage volume from a snapshot otherwise do not include this field. To list snapshots for your account use DescribeSnapshots in the Amazon Elastic Compute Cloud API Reference.

", - "StorediSCSIVolume$SourceSnapshotId": "

If the stored volume was created from a snapshot, this field contains the snapshot ID used, e.g. snap-78e22663. Otherwise, this field is not included.

" + "StartGatewayInput":{ + "base":"

A JSON object containing the of the gateway to start.

", + "refs":{ } }, - "Squash": { - "base": "

The user mapped to anonymous user. Valid options are the following:

  • \"RootSquash\" - Only root is mapped to anonymous user.

  • \"NoSquash\" - No one is mapped to anonymous user

  • \"AllSquash\" - Everyone is mapped to anonymous user.

", - "refs": { - "CreateNFSFileShareInput$Squash": "

Maps a user to anonymous user. Valid options are the following:

  • \"RootSquash\" - Only root is mapped to anonymous user.

  • \"NoSquash\" - No one is mapped to anonymous user.

  • \"AllSquash\" - Everyone is mapped to anonymous user.

", - "NFSFileShareInfo$Squash": null, - "UpdateNFSFileShareInput$Squash": "

The user mapped to anonymous user. Valid options are the following:

  • \"RootSquash\" - Only root is mapped to anonymous user.

  • \"NoSquash\" - No one is mapped to anonymous user

  • \"AllSquash\" - Everyone is mapped to anonymous user.

" + "StartGatewayOutput":{ + "base":"

A JSON object containing the of the gateway that was restarted.

", + "refs":{ } }, - "StartGatewayInput": { - "base": "

A JSON object containing the of the gateway to start.

", - "refs": { + "StorageClass":{ + "base":"

", + "refs":{ + "CreateNFSFileShareInput$DefaultStorageClass":"

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", + "CreateSMBFileShareInput$DefaultStorageClass":"

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", + "NFSFileShareInfo$DefaultStorageClass":"

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", + "SMBFileShareInfo$DefaultStorageClass":"

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", + "UpdateNFSFileShareInput$DefaultStorageClass":"

The default storage class for objects put into an Amazon S3 bucket by a file gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", + "UpdateSMBFileShareInput$DefaultStorageClass":"

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

" } }, - "StartGatewayOutput": { - "base": "

A JSON object containing the of the gateway that was restarted.

", - "refs": { + "StorageGatewayError":{ + "base":"

Provides additional information about an error that was returned by the service as an or. See the errorCode and errorDetails members for more information about the error.

", + "refs":{ + "InternalServerError$error":"

A StorageGatewayError that provides more information about the cause of the error.

", + "InvalidGatewayRequestException$error":"

A StorageGatewayError that provides more detail about the cause of the error.

", + "ServiceUnavailableError$error":"

A StorageGatewayError that provides more information about the cause of the error.

" } }, - "StorageClass": { - "base": "

", - "refs": { - "CreateNFSFileShareInput$DefaultStorageClass": "

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD or S3_STANDARD_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", - "NFSFileShareInfo$DefaultStorageClass": "

The default storage class for objects put into an Amazon S3 bucket by file gateway. Possible values are S3_STANDARD or S3_STANDARD_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

", - "UpdateNFSFileShareInput$DefaultStorageClass": "

The default storage class for objects put into an Amazon S3 bucket by a file gateway. Possible values are S3_STANDARD or S3_STANDARD_IA. If this field is not populated, the default value S3_STANDARD is used. Optional.

" + "StorediSCSIVolume":{ + "base":"

Describes an iSCSI stored volume.

", + "refs":{ + "StorediSCSIVolumes$member":null } }, - "StorageGatewayError": { - "base": "

Provides additional information about an error that was returned by the service as an or. See the errorCode and errorDetails members for more information about the error.

", - "refs": { - "InternalServerError$error": "

A StorageGatewayError that provides more information about the cause of the error.

", - "InvalidGatewayRequestException$error": "

A StorageGatewayError that provides more detail about the cause of the error.

", - "ServiceUnavailableError$error": "

A StorageGatewayError that provides more information about the cause of the error.

" + "StorediSCSIVolumes":{ + "base":null, + "refs":{ + "DescribeStorediSCSIVolumesOutput$StorediSCSIVolumes":null } }, - "StorediSCSIVolume": { - "base": "

Describes an iSCSI stored volume.

", - "refs": { - "StorediSCSIVolumes$member": null + "Tag":{ + "base":null, + "refs":{ + "Tags$member":null } }, - "StorediSCSIVolumes": { - "base": null, - "refs": { - "DescribeStorediSCSIVolumesOutput$StorediSCSIVolumes": null + "TagKey":{ + "base":null, + "refs":{ + "Tag$Key":null, + "TagKeys$member":null } }, - "Tag": { - "base": null, - "refs": { - "Tags$member": null + "TagKeys":{ + "base":null, + "refs":{ + "RemoveTagsFromResourceInput$TagKeys":"

The keys of the tags you want to remove from the specified resource. A tag is composed of a key/value pair.

" } }, - "TagKey": { - "base": null, - "refs": { - "Tag$Key": null, - "TagKeys$member": null + "TagValue":{ + "base":null, + "refs":{ + "Tag$Value":null } }, - "TagKeys": { - "base": null, - "refs": { - "RemoveTagsFromResourceInput$TagKeys": "

The keys of the tags you want to remove from the specified resource. A tag is composed of a key/value pair.

" + "Tags":{ + "base":null, + "refs":{ + "AddTagsToResourceInput$Tags":"

The key-value pair that represents the tag you want to add to the resource. The value can be an empty string.

Valid characters for key and value are letters, spaces, and numbers representable in UTF-8 format, and the following special characters: + - = . _ : / @.

", + "ListTagsForResourceOutput$Tags":"

An array that contains the tags for the specified resource.

" } }, - "TagValue": { - "base": null, - "refs": { - "Tag$Value": null + "Tape":{ + "base":"

Describes a virtual tape object.

", + "refs":{ + "Tapes$member":null } }, - "Tags": { - "base": null, - "refs": { - "AddTagsToResourceInput$Tags": "

The key-value pair that represents the tag you want to add to the resource. The value can be an empty string.

Valid characters for key and value are letters, spaces, and numbers representable in UTF-8 format, and the following special characters: + - = . _ : / @.

", - "ListTagsForResourceOutput$Tags": "

An array that contains the tags for the specified resource.

" + "TapeARN":{ + "base":null, + "refs":{ + "CancelArchivalInput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape you want to cancel archiving for.

", + "CancelArchivalOutput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape for which archiving was canceled.

", + "CancelRetrievalInput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape you want to cancel retrieval for.

", + "CancelRetrievalOutput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape for which retrieval was canceled.

", + "CreateTapeWithBarcodeOutput$TapeARN":"

A unique Amazon Resource Name (ARN) that represents the virtual tape that was created.

", + "DeleteTapeArchiveInput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape to delete from the virtual tape shelf (VTS).

", + "DeleteTapeArchiveOutput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape that was deleted from the virtual tape shelf (VTS).

", + "DeleteTapeInput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape to delete.

", + "DeleteTapeOutput$TapeARN":"

The Amazon Resource Name (ARN) of the deleted virtual tape.

", + "RetrieveTapeArchiveInput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape you want to retrieve from the virtual tape shelf (VTS).

", + "RetrieveTapeArchiveOutput$TapeARN":"

The Amazon Resource Name (ARN) of the retrieved virtual tape.

", + "RetrieveTapeRecoveryPointInput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape for which you want to retrieve the recovery point.

", + "RetrieveTapeRecoveryPointOutput$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape for which the recovery point was retrieved.

", + "Tape$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape.

", + "TapeARNs$member":null, + "TapeArchive$TapeARN":"

The Amazon Resource Name (ARN) of an archived virtual tape.

", + "TapeInfo$TapeARN":"

The Amazon Resource Name (ARN) of a virtual tape.

", + "TapeRecoveryPointInfo$TapeARN":"

The Amazon Resource Name (ARN) of the virtual tape.

" } }, - "Tape": { - "base": "

Describes a virtual tape object.

", - "refs": { - "Tapes$member": null + "TapeARNs":{ + "base":"

The Amazon Resource Name (ARN) of each of the tapes you want to list. If you don't specify a tape ARN, the response lists all tapes in both your VTL and VTS.

", + "refs":{ + "CreateTapesOutput$TapeARNs":"

A list of unique Amazon Resource Names (ARNs) that represents the virtual tapes that were created.

", + "DescribeTapeArchivesInput$TapeARNs":"

Specifies one or more unique Amazon Resource Names (ARNs) that represent the virtual tapes you want to describe.

", + "DescribeTapesInput$TapeARNs":"

Specifies one or more unique Amazon Resource Names (ARNs) that represent the virtual tapes you want to describe. If this parameter is not specified, Tape gateway returns a description of all virtual tapes associated with the specified gateway.

", + "ListTapesInput$TapeARNs":null } }, - "TapeARN": { - "base": null, - "refs": { - "CancelArchivalInput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape you want to cancel archiving for.

", - "CancelArchivalOutput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape for which archiving was canceled.

", - "CancelRetrievalInput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape you want to cancel retrieval for.

", - "CancelRetrievalOutput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape for which retrieval was canceled.

", - "CreateTapeWithBarcodeOutput$TapeARN": "

A unique Amazon Resource Name (ARN) that represents the virtual tape that was created.

", - "DeleteTapeArchiveInput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape to delete from the virtual tape shelf (VTS).

", - "DeleteTapeArchiveOutput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape that was deleted from the virtual tape shelf (VTS).

", - "DeleteTapeInput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape to delete.

", - "DeleteTapeOutput$TapeARN": "

The Amazon Resource Name (ARN) of the deleted virtual tape.

", - "RetrieveTapeArchiveInput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape you want to retrieve from the virtual tape shelf (VTS).

", - "RetrieveTapeArchiveOutput$TapeARN": "

The Amazon Resource Name (ARN) of the retrieved virtual tape.

", - "RetrieveTapeRecoveryPointInput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape for which you want to retrieve the recovery point.

", - "RetrieveTapeRecoveryPointOutput$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape for which the recovery point was retrieved.

", - "Tape$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape.

", - "TapeARNs$member": null, - "TapeArchive$TapeARN": "

The Amazon Resource Name (ARN) of an archived virtual tape.

", - "TapeInfo$TapeARN": "

The Amazon Resource Name (ARN) of a virtual tape.

", - "TapeRecoveryPointInfo$TapeARN": "

The Amazon Resource Name (ARN) of the virtual tape.

" + "TapeArchive":{ + "base":"

Represents a virtual tape that is archived in the virtual tape shelf (VTS).

", + "refs":{ + "TapeArchives$member":null } }, - "TapeARNs": { - "base": "

The Amazon Resource Name (ARN) of each of the tapes you want to list. If you don't specify a tape ARN, the response lists all tapes in both your VTL and VTS.

", - "refs": { - "CreateTapesOutput$TapeARNs": "

A list of unique Amazon Resource Names (ARNs) that represents the virtual tapes that were created.

", - "DescribeTapeArchivesInput$TapeARNs": "

Specifies one or more unique Amazon Resource Names (ARNs) that represent the virtual tapes you want to describe.

", - "DescribeTapesInput$TapeARNs": "

Specifies one or more unique Amazon Resource Names (ARNs) that represent the virtual tapes you want to describe. If this parameter is not specified, Tape gateway returns a description of all virtual tapes associated with the specified gateway.

", - "ListTapesInput$TapeARNs": null + "TapeArchiveStatus":{ + "base":null, + "refs":{ + "TapeArchive$TapeStatus":"

The current state of the archived virtual tape.

" } }, - "TapeArchive": { - "base": "

Represents a virtual tape that is archived in the virtual tape shelf (VTS).

", - "refs": { - "TapeArchives$member": null + "TapeArchives":{ + "base":null, + "refs":{ + "DescribeTapeArchivesOutput$TapeArchives":"

An array of virtual tape objects in the virtual tape shelf (VTS). The description includes of the Amazon Resource Name(ARN) of the virtual tapes. The information returned includes the Amazon Resource Names (ARNs) of the tapes, size of the tapes, status of the tapes, progress of the description and tape barcode.

" } }, - "TapeArchiveStatus": { - "base": null, - "refs": { - "TapeArchive$TapeStatus": "

The current state of the archived virtual tape.

" + "TapeBarcode":{ + "base":null, + "refs":{ + "CreateTapeWithBarcodeInput$TapeBarcode":"

The barcode that you want to assign to the tape.

Barcodes cannot be reused. This includes barcodes used for tapes that have been deleted.

", + "Tape$TapeBarcode":"

The barcode that identifies a specific virtual tape.

", + "TapeArchive$TapeBarcode":"

The barcode that identifies the archived virtual tape.

", + "TapeInfo$TapeBarcode":"

The barcode that identifies a specific virtual tape.

" } }, - "TapeArchives": { - "base": null, - "refs": { - "DescribeTapeArchivesOutput$TapeArchives": "

An array of virtual tape objects in the virtual tape shelf (VTS). The description includes of the Amazon Resource Name(ARN) of the virtual tapes. The information returned includes the Amazon Resource Names (ARNs) of the tapes, size of the tapes, status of the tapes, progress of the description and tape barcode.

" + "TapeBarcodePrefix":{ + "base":null, + "refs":{ + "CreateTapesInput$TapeBarcodePrefix":"

A prefix that you append to the barcode of the virtual tape you are creating. This prefix makes the barcode unique.

The prefix must be 1 to 4 characters in length and must be one of the uppercase letters from A to Z.

" } }, - "TapeBarcode": { - "base": null, - "refs": { - "CreateTapeWithBarcodeInput$TapeBarcode": "

The barcode that you want to assign to the tape.

Barcodes cannot be reused. This includes barcodes used for tapes that have been deleted.

", - "Tape$TapeBarcode": "

The barcode that identifies a specific virtual tape.

", - "TapeArchive$TapeBarcode": "

The barcode that identifies the archived virtual tape.

", - "TapeInfo$TapeBarcode": "

The barcode that identifies a specific virtual tape.

" + "TapeDriveType":{ + "base":null, + "refs":{ + "ActivateGatewayInput$TapeDriveType":"

The value that indicates the type of tape drive to use for tape gateway. This field is optional.

Valid Values: \"IBM-ULT3580-TD5\"

" } }, - "TapeBarcodePrefix": { - "base": null, - "refs": { - "CreateTapesInput$TapeBarcodePrefix": "

A prefix that you append to the barcode of the virtual tape you are creating. This prefix makes the barcode unique.

The prefix must be 1 to 4 characters in length and must be one of the uppercase letters from A to Z.

" + "TapeInfo":{ + "base":"

Describes a virtual tape.

", + "refs":{ + "TapeInfos$member":null } }, - "TapeDriveType": { - "base": null, - "refs": { - "ActivateGatewayInput$TapeDriveType": "

The value that indicates the type of tape drive to use for tape gateway. This field is optional.

Valid Values: \"IBM-ULT3580-TD5\"

" + "TapeInfos":{ + "base":"

An array of TapeInfo objects, where each object describes an a single tape. If there not tapes in the tape library or VTS, then the TapeInfos is an empty array.

", + "refs":{ + "ListTapesOutput$TapeInfos":null } }, - "TapeInfo": { - "base": "

Describes a virtual tape.

", - "refs": { - "TapeInfos$member": null + "TapeRecoveryPointInfo":{ + "base":"

Describes a recovery point.

", + "refs":{ + "TapeRecoveryPointInfos$member":null } }, - "TapeInfos": { - "base": "

An array of TapeInfo objects, where each object describes an a single tape. If there not tapes in the tape library or VTS, then the TapeInfos is an empty array.

", - "refs": { - "ListTapesOutput$TapeInfos": null + "TapeRecoveryPointInfos":{ + "base":null, + "refs":{ + "DescribeTapeRecoveryPointsOutput$TapeRecoveryPointInfos":"

An array of TapeRecoveryPointInfos that are available for the specified gateway.

" } }, - "TapeRecoveryPointInfo": { - "base": "

Describes a recovery point.

", - "refs": { - "TapeRecoveryPointInfos$member": null + "TapeRecoveryPointStatus":{ + "base":null, + "refs":{ + "TapeRecoveryPointInfo$TapeStatus":null } }, - "TapeRecoveryPointInfos": { - "base": null, - "refs": { - "DescribeTapeRecoveryPointsOutput$TapeRecoveryPointInfos": "

An array of TapeRecoveryPointInfos that are available for the specified gateway.

" + "TapeSize":{ + "base":null, + "refs":{ + "CreateTapeWithBarcodeInput$TapeSizeInBytes":"

The size, in bytes, of the virtual tape that you want to create.

The size must be aligned by gigabyte (1024*1024*1024 byte).

", + "CreateTapesInput$TapeSizeInBytes":"

The size, in bytes, of the virtual tapes that you want to create.

The size must be aligned by gigabyte (1024*1024*1024 byte).

", + "Tape$TapeSizeInBytes":"

The size, in bytes, of the virtual tape capacity.

", + "TapeArchive$TapeSizeInBytes":"

The size, in bytes, of the archived virtual tape.

", + "TapeInfo$TapeSizeInBytes":"

The size, in bytes, of a virtual tape.

", + "TapeRecoveryPointInfo$TapeSizeInBytes":"

The size, in bytes, of the virtual tapes to recover.

" } }, - "TapeRecoveryPointStatus": { - "base": null, - "refs": { - "TapeRecoveryPointInfo$TapeStatus": null + "TapeStatus":{ + "base":null, + "refs":{ + "Tape$TapeStatus":"

The current state of the virtual tape.

", + "TapeInfo$TapeStatus":"

The status of the tape.

" } }, - "TapeSize": { - "base": null, - "refs": { - "CreateTapeWithBarcodeInput$TapeSizeInBytes": "

The size, in bytes, of the virtual tape that you want to create.

The size must be aligned by gigabyte (1024*1024*1024 byte).

", - "CreateTapesInput$TapeSizeInBytes": "

The size, in bytes, of the virtual tapes that you want to create.

The size must be aligned by gigabyte (1024*1024*1024 byte).

", - "Tape$TapeSizeInBytes": "

The size, in bytes, of the virtual tape capacity.

", - "TapeArchive$TapeSizeInBytes": "

The size, in bytes, of the archived virtual tape.

", - "TapeInfo$TapeSizeInBytes": "

The size, in bytes, of a virtual tape.

", - "TapeRecoveryPointInfo$TapeSizeInBytes": "

The size, in bytes, of the virtual tapes to recover.

" + "TapeUsage":{ + "base":null, + "refs":{ + "Tape$TapeUsedInBytes":"

The size, in bytes, of data stored on the virtual tape.

This value is not available for tapes created prior to May 13, 2015.

", + "TapeArchive$TapeUsedInBytes":"

The size, in bytes, of data stored on the virtual tape.

This value is not available for tapes created prior to May 13, 2015.

" } }, - "TapeStatus": { - "base": null, - "refs": { - "Tape$TapeStatus": "

The current state of the virtual tape.

", - "TapeInfo$TapeStatus": "

The status of the tape.

" + "Tapes":{ + "base":null, + "refs":{ + "DescribeTapesOutput$Tapes":"

An array of virtual tape descriptions.

" } }, - "TapeUsage": { - "base": null, - "refs": { - "Tape$TapeUsedInBytes": "

The size, in bytes, of data stored on the virtual tape.

This value is not available for tapes created prior to May 13, 2015.

", - "TapeArchive$TapeUsedInBytes": "

The size, in bytes, of data stored on the virtual tape.

This value is not available for tapes created prior to May 13, 2015.

" + "TargetARN":{ + "base":null, + "refs":{ + "ChapInfo$TargetARN":"

The Amazon Resource Name (ARN) of the volume.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

", + "CreateCachediSCSIVolumeOutput$TargetARN":"

he Amazon Resource Name (ARN) of the volume target that includes the iSCSI name that initiators can use to connect to the target.

", + "CreateStorediSCSIVolumeOutput$TargetARN":"

he Amazon Resource Name (ARN) of the volume target that includes the iSCSI name that initiators can use to connect to the target.

", + "DeleteChapCredentialsInput$TargetARN":"

The Amazon Resource Name (ARN) of the iSCSI volume target. Use the DescribeStorediSCSIVolumes operation to return to retrieve the TargetARN for specified VolumeARN.

", + "DeleteChapCredentialsOutput$TargetARN":"

The Amazon Resource Name (ARN) of the target.

", + "DescribeChapCredentialsInput$TargetARN":"

The Amazon Resource Name (ARN) of the iSCSI volume target. Use the DescribeStorediSCSIVolumes operation to return to retrieve the TargetARN for specified VolumeARN.

", + "DeviceiSCSIAttributes$TargetARN":"

Specifies the unique Amazon Resource Name(ARN) that encodes the iSCSI qualified name(iqn) of a tape drive or media changer target.

", + "UpdateChapCredentialsInput$TargetARN":"

The Amazon Resource Name (ARN) of the iSCSI volume target. Use the DescribeStorediSCSIVolumes operation to return the TargetARN for specified VolumeARN.

", + "UpdateChapCredentialsOutput$TargetARN":"

The Amazon Resource Name (ARN) of the target. This is the same target specified in the request.

", + "VolumeiSCSIAttributes$TargetARN":"

The Amazon Resource Name (ARN) of the volume target.

" } }, - "Tapes": { - "base": null, - "refs": { - "DescribeTapesOutput$Tapes": "

An array of virtual tape descriptions.

" + "TargetName":{ + "base":null, + "refs":{ + "CreateCachediSCSIVolumeInput$TargetName":"

The name of the iSCSI target used by initiators to connect to the target and as a suffix for the target ARN. For example, specifying TargetName as myvolume results in the target ARN of arn:aws:storagegateway:us-east-2:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume. The target name must be unique across all volumes of a gateway.

", + "CreateStorediSCSIVolumeInput$TargetName":"

The name of the iSCSI target used by initiators to connect to the target and as a suffix for the target ARN. For example, specifying TargetName as myvolume results in the target ARN of arn:aws:storagegateway:us-east-2:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume. The target name must be unique across all volumes of a gateway.

" } }, - "TargetARN": { - "base": null, - "refs": { - "ChapInfo$TargetARN": "

The Amazon Resource Name (ARN) of the volume.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

", - "CreateCachediSCSIVolumeOutput$TargetARN": null, - "CreateStorediSCSIVolumeOutput$TargetARN": "

he Amazon Resource Name (ARN) of the volume target that includes the iSCSI name that initiators can use to connect to the target.

", - "DeleteChapCredentialsInput$TargetARN": "

The Amazon Resource Name (ARN) of the iSCSI volume target. Use the DescribeStorediSCSIVolumes operation to return to retrieve the TargetARN for specified VolumeARN.

", - "DeleteChapCredentialsOutput$TargetARN": "

The Amazon Resource Name (ARN) of the target.

", - "DescribeChapCredentialsInput$TargetARN": "

The Amazon Resource Name (ARN) of the iSCSI volume target. Use the DescribeStorediSCSIVolumes operation to return to retrieve the TargetARN for specified VolumeARN.

", - "DeviceiSCSIAttributes$TargetARN": "

Specifies the unique Amazon Resource Name(ARN) that encodes the iSCSI qualified name(iqn) of a tape drive or media changer target.

", - "UpdateChapCredentialsInput$TargetARN": "

The Amazon Resource Name (ARN) of the iSCSI volume target. Use the DescribeStorediSCSIVolumes operation to return the TargetARN for specified VolumeARN.

", - "UpdateChapCredentialsOutput$TargetARN": "

The Amazon Resource Name (ARN) of the target. This is the same target specified in the request.

", - "VolumeiSCSIAttributes$TargetARN": "

The Amazon Resource Name (ARN) of the volume target.

" + "Time":{ + "base":null, + "refs":{ + "Tape$TapeCreatedDate":"

The date the virtual tape was created.

", + "TapeArchive$TapeCreatedDate":"

The date the virtual tape was created.

", + "TapeArchive$CompletionTime":"

The time that the archiving of the virtual tape was completed.

The default time stamp format is in the ISO8601 extended YYYY-MM-DD'T'HH:MM:SS'Z' format.

", + "TapeRecoveryPointInfo$TapeRecoveryPointTime":"

The time when the point-in-time view of the virtual tape was replicated for later recovery.

The default time stamp format of the tape recovery point time is in the ISO8601 extended YYYY-MM-DD'T'HH:MM:SS'Z' format.

" } }, - "TargetName": { - "base": null, - "refs": { - "CreateCachediSCSIVolumeInput$TargetName": null, - "CreateStorediSCSIVolumeInput$TargetName": "

The name of the iSCSI target used by initiators to connect to the target and as a suffix for the target ARN. For example, specifying TargetName as myvolume results in the target ARN of arn:aws:storagegateway:us-east-2:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume. The target name must be unique across all volumes of a gateway.

" + "UpdateBandwidthRateLimitInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "Time": { - "base": null, - "refs": { - "Tape$TapeCreatedDate": "

The date the virtual tape was created.

", - "TapeArchive$TapeCreatedDate": null, - "TapeArchive$CompletionTime": "

The time that the archiving of the virtual tape was completed.

The string format of the completion time is in the ISO8601 extended YYYY-MM-DD'T'HH:MM:SS'Z' format.

", - "TapeRecoveryPointInfo$TapeRecoveryPointTime": "

The time when the point-in-time view of the virtual tape was replicated for later recovery.

The string format of the tape recovery point time is in the ISO8601 extended YYYY-MM-DD'T'HH:MM:SS'Z' format.

" + "UpdateBandwidthRateLimitOutput":{ + "base":"

A JSON object containing the of the gateway whose throttle information was updated.

", + "refs":{ } }, - "UpdateBandwidthRateLimitInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "UpdateChapCredentialsInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "UpdateBandwidthRateLimitOutput": { - "base": "

A JSON object containing the of the gateway whose throttle information was updated.

", - "refs": { + "UpdateChapCredentialsOutput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "UpdateChapCredentialsInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "UpdateGatewayInformationInput":{ + "base":null, + "refs":{ } }, - "UpdateChapCredentialsOutput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "UpdateGatewayInformationOutput":{ + "base":"

A JSON object containing the ARN of the gateway that was updated.

", + "refs":{ } }, - "UpdateGatewayInformationInput": { - "base": null, - "refs": { + "UpdateGatewaySoftwareNowInput":{ + "base":"

A JSON object containing the of the gateway to update.

", + "refs":{ } }, - "UpdateGatewayInformationOutput": { - "base": "

A JSON object containing the ARN of the gateway that was updated.

", - "refs": { + "UpdateGatewaySoftwareNowOutput":{ + "base":"

A JSON object containing the of the gateway that was updated.

", + "refs":{ } }, - "UpdateGatewaySoftwareNowInput": { - "base": "

A JSON object containing the of the gateway to update.

", - "refs": { + "UpdateMaintenanceStartTimeInput":{ + "base":"

A JSON object containing the following fields:

", + "refs":{ } }, - "UpdateGatewaySoftwareNowOutput": { - "base": "

A JSON object containing the of the gateway that was updated.

", - "refs": { + "UpdateMaintenanceStartTimeOutput":{ + "base":"

A JSON object containing the of the gateway whose maintenance start time is updated.

", + "refs":{ } }, - "UpdateMaintenanceStartTimeInput": { - "base": "

A JSON object containing the following fields:

", - "refs": { + "UpdateNFSFileShareInput":{ + "base":"

UpdateNFSFileShareInput

", + "refs":{ } }, - "UpdateMaintenanceStartTimeOutput": { - "base": "

A JSON object containing the of the gateway whose maintenance start time is updated.

", - "refs": { + "UpdateNFSFileShareOutput":{ + "base":"

UpdateNFSFileShareOutput

", + "refs":{ } }, - "UpdateNFSFileShareInput": { - "base": "

UpdateNFSFileShareInput

", - "refs": { + "UpdateSMBFileShareInput":{ + "base":"

UpdateSMBFileShareInput

", + "refs":{ } }, - "UpdateNFSFileShareOutput": { - "base": "

UpdateNFSFileShareOutput

", - "refs": { + "UpdateSMBFileShareOutput":{ + "base":"

UpdateSMBFileShareOutput

", + "refs":{ } }, - "UpdateSnapshotScheduleInput": { - "base": "

A JSON object containing one or more of the following fields:

", - "refs": { + "UpdateSnapshotScheduleInput":{ + "base":"

A JSON object containing one or more of the following fields:

", + "refs":{ } }, - "UpdateSnapshotScheduleOutput": { - "base": "

A JSON object containing the of the updated storage volume.

", - "refs": { + "UpdateSnapshotScheduleOutput":{ + "base":"

A JSON object containing the of the updated storage volume.

", + "refs":{ } }, - "UpdateVTLDeviceTypeInput": { - "base": null, - "refs": { + "UpdateVTLDeviceTypeInput":{ + "base":null, + "refs":{ } }, - "UpdateVTLDeviceTypeOutput": { - "base": "

UpdateVTLDeviceTypeOutput

", - "refs": { + "UpdateVTLDeviceTypeOutput":{ + "base":"

UpdateVTLDeviceTypeOutput

", + "refs":{ } }, - "VTLDevice": { - "base": "

Represents a device object associated with a tape gateway.

", - "refs": { - "VTLDevices$member": null + "VTLDevice":{ + "base":"

Represents a device object associated with a tape gateway.

", + "refs":{ + "VTLDevices$member":null } }, - "VTLDeviceARN": { - "base": null, - "refs": { - "Tape$VTLDevice": "

The virtual tape library (VTL) device that the virtual tape is associated with.

", - "UpdateVTLDeviceTypeInput$VTLDeviceARN": "

The Amazon Resource Name (ARN) of the medium changer you want to select.

", - "UpdateVTLDeviceTypeOutput$VTLDeviceARN": "

The Amazon Resource Name (ARN) of the medium changer you have selected.

", - "VTLDevice$VTLDeviceARN": "

Specifies the unique Amazon Resource Name (ARN) of the device (tape drive or media changer).

", - "VTLDeviceARNs$member": null + "VTLDeviceARN":{ + "base":null, + "refs":{ + "Tape$VTLDevice":"

The virtual tape library (VTL) device that the virtual tape is associated with.

", + "UpdateVTLDeviceTypeInput$VTLDeviceARN":"

The Amazon Resource Name (ARN) of the medium changer you want to select.

", + "UpdateVTLDeviceTypeOutput$VTLDeviceARN":"

The Amazon Resource Name (ARN) of the medium changer you have selected.

", + "VTLDevice$VTLDeviceARN":"

Specifies the unique Amazon Resource Name (ARN) of the device (tape drive or media changer).

", + "VTLDeviceARNs$member":null } }, - "VTLDeviceARNs": { - "base": null, - "refs": { - "DescribeVTLDevicesInput$VTLDeviceARNs": "

An array of strings, where each string represents the Amazon Resource Name (ARN) of a VTL device.

All of the specified VTL devices must be from the same gateway. If no VTL devices are specified, the result will contain all devices on the specified gateway.

" + "VTLDeviceARNs":{ + "base":null, + "refs":{ + "DescribeVTLDevicesInput$VTLDeviceARNs":"

An array of strings, where each string represents the Amazon Resource Name (ARN) of a VTL device.

All of the specified VTL devices must be from the same gateway. If no VTL devices are specified, the result will contain all devices on the specified gateway.

" } }, - "VTLDeviceProductIdentifier": { - "base": null, - "refs": { - "VTLDevice$VTLDeviceProductIdentifier": null + "VTLDeviceProductIdentifier":{ + "base":null, + "refs":{ + "VTLDevice$VTLDeviceProductIdentifier":null } }, - "VTLDeviceType": { - "base": null, - "refs": { - "VTLDevice$VTLDeviceType": null + "VTLDeviceType":{ + "base":null, + "refs":{ + "VTLDevice$VTLDeviceType":null } }, - "VTLDeviceVendor": { - "base": null, - "refs": { - "VTLDevice$VTLDeviceVendor": null + "VTLDeviceVendor":{ + "base":null, + "refs":{ + "VTLDevice$VTLDeviceVendor":null } }, - "VTLDevices": { - "base": null, - "refs": { - "DescribeVTLDevicesOutput$VTLDevices": "

An array of VTL device objects composed of the Amazon Resource Name(ARN) of the VTL devices.

" + "VTLDevices":{ + "base":null, + "refs":{ + "DescribeVTLDevicesOutput$VTLDevices":"

An array of VTL device objects composed of the Amazon Resource Name(ARN) of the VTL devices.

" } }, - "VolumeARN": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeARN": "

The Amazon Resource Name (ARN) of the storage volume.

", - "CreateCachediSCSIVolumeInput$SourceVolumeARN": "

The ARN for an existing volume. Specifying this ARN makes the new volume into an exact copy of the specified existing volume's latest recovery point. The VolumeSizeInBytes value for this new volume must be equal to or larger than the size of the existing volume, in bytes.

", - "CreateCachediSCSIVolumeOutput$VolumeARN": null, - "CreateSnapshotFromVolumeRecoveryPointInput$VolumeARN": null, - "CreateSnapshotFromVolumeRecoveryPointOutput$VolumeARN": null, - "CreateSnapshotInput$VolumeARN": "

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", - "CreateSnapshotOutput$VolumeARN": "

The Amazon Resource Name (ARN) of the volume of which the snapshot was taken.

", - "CreateStorediSCSIVolumeOutput$VolumeARN": "

The Amazon Resource Name (ARN) of the configured volume.

", - "DeleteSnapshotScheduleInput$VolumeARN": null, - "DeleteSnapshotScheduleOutput$VolumeARN": null, - "DeleteVolumeInput$VolumeARN": "

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", - "DeleteVolumeOutput$VolumeARN": "

The Amazon Resource Name (ARN) of the storage volume that was deleted. It is the same ARN you provided in the request.

", - "DescribeSnapshotScheduleInput$VolumeARN": "

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", - "DescribeSnapshotScheduleOutput$VolumeARN": null, - "ListVolumeInitiatorsInput$VolumeARN": "

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes for the gateway.

", - "StorediSCSIVolume$VolumeARN": "

The Amazon Resource Name (ARN) of the storage volume.

", - "UpdateSnapshotScheduleInput$VolumeARN": "

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", - "UpdateSnapshotScheduleOutput$VolumeARN": "

", - "VolumeARNs$member": null, - "VolumeInfo$VolumeARN": "

The Amazon Resource Name (ARN) for the storage volume. For example, the following is a valid ARN:

arn:aws:storagegateway:us-east-2:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

", - "VolumeRecoveryPointInfo$VolumeARN": null + "VolumeARN":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeARN":"

The Amazon Resource Name (ARN) of the storage volume.

", + "CreateCachediSCSIVolumeInput$SourceVolumeARN":"

The ARN for an existing volume. Specifying this ARN makes the new volume into an exact copy of the specified existing volume's latest recovery point. The VolumeSizeInBytes value for this new volume must be equal to or larger than the size of the existing volume, in bytes.

", + "CreateCachediSCSIVolumeOutput$VolumeARN":"

The Amazon Resource Name (ARN) of the configured volume.

", + "CreateSnapshotFromVolumeRecoveryPointInput$VolumeARN":null, + "CreateSnapshotFromVolumeRecoveryPointOutput$VolumeARN":null, + "CreateSnapshotInput$VolumeARN":"

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", + "CreateSnapshotOutput$VolumeARN":"

The Amazon Resource Name (ARN) of the volume of which the snapshot was taken.

", + "CreateStorediSCSIVolumeOutput$VolumeARN":"

The Amazon Resource Name (ARN) of the configured volume.

", + "DeleteSnapshotScheduleInput$VolumeARN":null, + "DeleteSnapshotScheduleOutput$VolumeARN":null, + "DeleteVolumeInput$VolumeARN":"

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", + "DeleteVolumeOutput$VolumeARN":"

The Amazon Resource Name (ARN) of the storage volume that was deleted. It is the same ARN you provided in the request.

", + "DescribeSnapshotScheduleInput$VolumeARN":"

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", + "DescribeSnapshotScheduleOutput$VolumeARN":null, + "ListVolumeInitiatorsInput$VolumeARN":"

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes for the gateway.

", + "StorediSCSIVolume$VolumeARN":"

The Amazon Resource Name (ARN) of the storage volume.

", + "UpdateSnapshotScheduleInput$VolumeARN":"

The Amazon Resource Name (ARN) of the volume. Use the ListVolumes operation to return a list of gateway volumes.

", + "UpdateSnapshotScheduleOutput$VolumeARN":"

", + "VolumeARNs$member":null, + "VolumeInfo$VolumeARN":"

The Amazon Resource Name (ARN) for the storage volume. For example, the following is a valid ARN:

arn:aws:storagegateway:us-east-2:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

", + "VolumeRecoveryPointInfo$VolumeARN":null } }, - "VolumeARNs": { - "base": null, - "refs": { - "DescribeCachediSCSIVolumesInput$VolumeARNs": null, - "DescribeStorediSCSIVolumesInput$VolumeARNs": "

An array of strings where each string represents the Amazon Resource Name (ARN) of a stored volume. All of the specified stored volumes must from the same gateway. Use ListVolumes to get volume ARNs for a gateway.

" + "VolumeARNs":{ + "base":null, + "refs":{ + "DescribeCachediSCSIVolumesInput$VolumeARNs":null, + "DescribeStorediSCSIVolumesInput$VolumeARNs":"

An array of strings where each string represents the Amazon Resource Name (ARN) of a stored volume. All of the specified stored volumes must from the same gateway. Use ListVolumes to get volume ARNs for a gateway.

" } }, - "VolumeId": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeId": "

The unique identifier of the volume, e.g. vol-AE4B946D.

", - "StorediSCSIVolume$VolumeId": "

The unique identifier of the volume, e.g. vol-AE4B946D.

", - "VolumeInfo$VolumeId": "

The unique identifier assigned to the volume. This ID becomes part of the volume Amazon Resource Name (ARN), which you use as input for other operations.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

" + "VolumeId":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeId":"

The unique identifier of the volume, e.g. vol-AE4B946D.

", + "StorediSCSIVolume$VolumeId":"

The unique identifier of the volume, e.g. vol-AE4B946D.

", + "VolumeInfo$VolumeId":"

The unique identifier assigned to the volume. This ID becomes part of the volume Amazon Resource Name (ARN), which you use as input for other operations.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

" } }, - "VolumeInfo": { - "base": "

Describes a storage volume object.

", - "refs": { - "VolumeInfos$member": null + "VolumeInfo":{ + "base":"

Describes a storage volume object.

", + "refs":{ + "VolumeInfos$member":null } }, - "VolumeInfos": { - "base": null, - "refs": { - "ListVolumesOutput$VolumeInfos": null - } - }, - "VolumeRecoveryPointInfo": { - "base": null, - "refs": { - "VolumeRecoveryPointInfos$member": null - } - }, - "VolumeRecoveryPointInfos": { - "base": null, - "refs": { - "ListVolumeRecoveryPointsOutput$VolumeRecoveryPointInfos": null - } - }, - "VolumeStatus": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeStatus": "

One of the VolumeStatus values that indicates the state of the storage volume.

", - "StorediSCSIVolume$VolumeStatus": "

One of the VolumeStatus values that indicates the state of the storage volume.

" - } - }, - "VolumeType": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeType": "

One of the VolumeType enumeration values that describes the type of the volume.

", - "StorediSCSIVolume$VolumeType": "

One of the VolumeType enumeration values describing the type of the volume.

", - "VolumeInfo$VolumeType": null - } - }, - "VolumeUsedInBytes": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeUsedInBytes": "

The size of the data stored on the volume in bytes.

This value is not available for volumes created prior to May 13, 2015, until you store data on the volume.

", - "StorediSCSIVolume$VolumeUsedInBytes": "

The size of the data stored on the volume in bytes.

This value is not available for volumes created prior to May 13, 2015, until you store data on the volume.

" - } - }, - "VolumeiSCSIAttributes": { - "base": "

Lists iSCSI information about a volume.

", - "refs": { - "CachediSCSIVolume$VolumeiSCSIAttributes": "

An VolumeiSCSIAttributes object that represents a collection of iSCSI attributes for one stored volume.

", - "StorediSCSIVolume$VolumeiSCSIAttributes": "

An VolumeiSCSIAttributes object that represents a collection of iSCSI attributes for one stored volume.

" - } - }, - "boolean": { - "base": null, - "refs": { - "CreateStorediSCSIVolumeInput$PreserveExistingData": "

Specify this field as true if you want to preserve the data on the local disk. Otherwise, specifying this field as false creates an empty volume.

Valid Values: true, false

", - "DeleteFileShareInput$ForceDelete": "

If this value is set to true, the operation deletes a file share immediately and aborts all data uploads to AWS. Otherwise, the file share is not deleted until all data is uploaded to AWS. This process aborts the data upload process, and the file share enters the FORCE_DELETING status.

", - "DeviceiSCSIAttributes$ChapEnabled": "

Indicates whether mutual CHAP is enabled for the iSCSI target.

", - "NFSFileShareInfo$KMSEncrypted": "

True to use Amazon S3 server side encryption with your own KMS key, or false to use a key managed by Amazon S3. Optional.

", - "StorediSCSIVolume$PreservedExistingData": "

Indicates if when the stored volume was created, existing data on the underlying local disk was preserved.

Valid Values: true, false

", - "VolumeiSCSIAttributes$ChapEnabled": "

Indicates whether mutual CHAP is enabled for the iSCSI target.

" - } - }, - "double": { - "base": null, - "refs": { - "DescribeCacheOutput$CacheUsedPercentage": null, - "DescribeCacheOutput$CacheDirtyPercentage": null, - "DescribeCacheOutput$CacheHitPercentage": null, - "DescribeCacheOutput$CacheMissPercentage": null - } - }, - "errorDetails": { - "base": null, - "refs": { - "StorageGatewayError$errorDetails": "

Human-readable text that provides detail about the error that occurred.

" - } - }, - "integer": { - "base": null, - "refs": { - "DeviceiSCSIAttributes$NetworkInterfacePort": "

The port used to communicate with iSCSI VTL device targets.

", - "VolumeiSCSIAttributes$NetworkInterfacePort": "

The port used to communicate with iSCSI targets.

" - } - }, - "long": { - "base": null, - "refs": { - "CachediSCSIVolume$VolumeSizeInBytes": "

The size, in bytes, of the volume capacity.

", - "CreateCachediSCSIVolumeInput$VolumeSizeInBytes": null, - "CreateStorediSCSIVolumeOutput$VolumeSizeInBytes": "

The size of the volume in bytes.

", - "DescribeCacheOutput$CacheAllocatedInBytes": null, - "DescribeUploadBufferOutput$UploadBufferUsedInBytes": null, - "DescribeUploadBufferOutput$UploadBufferAllocatedInBytes": null, - "DescribeWorkingStorageOutput$WorkingStorageUsedInBytes": "

The total working storage in bytes in use by the gateway. If no working storage is configured for the gateway, this field returns 0.

", - "DescribeWorkingStorageOutput$WorkingStorageAllocatedInBytes": "

The total working storage in bytes allocated for the gateway. If no working storage is configured for the gateway, this field returns 0.

", - "Disk$DiskSizeInBytes": null, - "StorediSCSIVolume$VolumeSizeInBytes": "

The size of the volume in bytes.

", - "VolumeInfo$VolumeSizeInBytes": "

The size of the volume in bytes.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

", - "VolumeRecoveryPointInfo$VolumeSizeInBytes": null, - "VolumeRecoveryPointInfo$VolumeUsageInBytes": null - } - }, - "string": { - "base": null, - "refs": { - "CreateSnapshotFromVolumeRecoveryPointOutput$VolumeRecoveryPointTime": null, - "DescribeGatewayInformationOutput$GatewayName": "

The name you configured for your gateway.

", - "Disk$DiskPath": null, - "Disk$DiskNode": null, - "Disk$DiskStatus": null, - "Disk$DiskAllocationResource": null, - "GatewayInfo$GatewayName": "

The name of the gateway.

", - "InternalServerError$message": "

A human-readable message describing the error that occurred.

", - "InvalidGatewayRequestException$message": "

A human-readable message describing the error that occurred.

", - "NetworkInterface$Ipv4Address": "

The Internet Protocol version 4 (IPv4) address of the interface.

", - "NetworkInterface$MacAddress": "

The Media Access Control (MAC) address of the interface.

This is currently unsupported and will not be returned in output.

", - "NetworkInterface$Ipv6Address": "

The Internet Protocol version 6 (IPv6) address of the interface. Currently not supported.

", - "ServiceUnavailableError$message": "

A human-readable message describing the error that occurred.

", - "UpdateGatewayInformationOutput$GatewayName": null, - "VolumeRecoveryPointInfo$VolumeRecoveryPointTime": null, - "errorDetails$key": null, - "errorDetails$value": null + "VolumeInfos":{ + "base":null, + "refs":{ + "ListVolumesOutput$VolumeInfos":null + } + }, + "VolumeRecoveryPointInfo":{ + "base":null, + "refs":{ + "VolumeRecoveryPointInfos$member":null + } + }, + "VolumeRecoveryPointInfos":{ + "base":null, + "refs":{ + "ListVolumeRecoveryPointsOutput$VolumeRecoveryPointInfos":null + } + }, + "VolumeStatus":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeStatus":"

One of the VolumeStatus values that indicates the state of the storage volume.

", + "StorediSCSIVolume$VolumeStatus":"

One of the VolumeStatus values that indicates the state of the storage volume.

" + } + }, + "VolumeType":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeType":"

One of the VolumeType enumeration values that describes the type of the volume.

", + "StorediSCSIVolume$VolumeType":"

One of the VolumeType enumeration values describing the type of the volume.

", + "VolumeInfo$VolumeType":null + } + }, + "VolumeUsedInBytes":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeUsedInBytes":"

The size of the data stored on the volume in bytes.

This value is not available for volumes created prior to May 13, 2015, until you store data on the volume.

", + "StorediSCSIVolume$VolumeUsedInBytes":"

The size of the data stored on the volume in bytes.

This value is not available for volumes created prior to May 13, 2015, until you store data on the volume.

" + } + }, + "VolumeiSCSIAttributes":{ + "base":"

Lists iSCSI information about a volume.

", + "refs":{ + "CachediSCSIVolume$VolumeiSCSIAttributes":"

An VolumeiSCSIAttributes object that represents a collection of iSCSI attributes for one stored volume.

", + "StorediSCSIVolume$VolumeiSCSIAttributes":"

An VolumeiSCSIAttributes object that represents a collection of iSCSI attributes for one stored volume.

" + } + }, + "boolean":{ + "base":null, + "refs":{ + "CreateStorediSCSIVolumeInput$PreserveExistingData":"

Specify this field as true if you want to preserve the data on the local disk. Otherwise, specifying this field as false creates an empty volume.

Valid Values: true, false

", + "DeleteFileShareInput$ForceDelete":"

If this value is set to true, the operation deletes a file share immediately and aborts all data uploads to AWS. Otherwise, the file share is not deleted until all data is uploaded to AWS. This process aborts the data upload process, and the file share enters the FORCE_DELETING status.

", + "DeviceiSCSIAttributes$ChapEnabled":"

Indicates whether mutual CHAP is enabled for the iSCSI target.

", + "NFSFileShareInfo$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own KMS key, or false to use a key managed by Amazon S3. Optional.

", + "SMBFileShareInfo$KMSEncrypted":"

True to use Amazon S3 server side encryption with your own KMS key, or false to use a key managed by Amazon S3. Optional.

", + "StorediSCSIVolume$PreservedExistingData":"

Indicates if when the stored volume was created, existing data on the underlying local disk was preserved.

Valid Values: true, false

", + "VolumeiSCSIAttributes$ChapEnabled":"

Indicates whether mutual CHAP is enabled for the iSCSI target.

" + } + }, + "double":{ + "base":null, + "refs":{ + "DescribeCacheOutput$CacheUsedPercentage":null, + "DescribeCacheOutput$CacheDirtyPercentage":null, + "DescribeCacheOutput$CacheHitPercentage":null, + "DescribeCacheOutput$CacheMissPercentage":null + } + }, + "errorDetails":{ + "base":null, + "refs":{ + "StorageGatewayError$errorDetails":"

Human-readable text that provides detail about the error that occurred.

" + } + }, + "integer":{ + "base":null, + "refs":{ + "DeviceiSCSIAttributes$NetworkInterfacePort":"

The port used to communicate with iSCSI VTL device targets.

", + "VolumeiSCSIAttributes$NetworkInterfacePort":"

The port used to communicate with iSCSI targets.

" + } + }, + "long":{ + "base":null, + "refs":{ + "CachediSCSIVolume$VolumeSizeInBytes":"

The size, in bytes, of the volume capacity.

", + "CreateCachediSCSIVolumeInput$VolumeSizeInBytes":"

The size of the volume in bytes.

", + "CreateStorediSCSIVolumeOutput$VolumeSizeInBytes":"

The size of the volume in bytes.

", + "DescribeCacheOutput$CacheAllocatedInBytes":null, + "DescribeUploadBufferOutput$UploadBufferUsedInBytes":null, + "DescribeUploadBufferOutput$UploadBufferAllocatedInBytes":null, + "DescribeWorkingStorageOutput$WorkingStorageUsedInBytes":"

The total working storage in bytes in use by the gateway. If no working storage is configured for the gateway, this field returns 0.

", + "DescribeWorkingStorageOutput$WorkingStorageAllocatedInBytes":"

The total working storage in bytes allocated for the gateway. If no working storage is configured for the gateway, this field returns 0.

", + "Disk$DiskSizeInBytes":null, + "StorediSCSIVolume$VolumeSizeInBytes":"

The size of the volume in bytes.

", + "VolumeInfo$VolumeSizeInBytes":"

The size of the volume in bytes.

Valid Values: 50 to 500 lowercase letters, numbers, periods (.), and hyphens (-).

", + "VolumeRecoveryPointInfo$VolumeSizeInBytes":null, + "VolumeRecoveryPointInfo$VolumeUsageInBytes":null + } + }, + "string":{ + "base":null, + "refs":{ + "CreateSnapshotFromVolumeRecoveryPointOutput$VolumeRecoveryPointTime":null, + "DescribeGatewayInformationOutput$GatewayName":"

The name you configured for your gateway.

", + "Disk$DiskPath":null, + "Disk$DiskNode":null, + "Disk$DiskStatus":null, + "Disk$DiskAllocationResource":null, + "GatewayInfo$GatewayName":"

The name of the gateway.

", + "InternalServerError$message":"

A human-readable message describing the error that occurred.

", + "InvalidGatewayRequestException$message":"

A human-readable message describing the error that occurred.

", + "NetworkInterface$Ipv4Address":"

The Internet Protocol version 4 (IPv4) address of the interface.

", + "NetworkInterface$MacAddress":"

The Media Access Control (MAC) address of the interface.

This is currently unsupported and will not be returned in output.

", + "NetworkInterface$Ipv6Address":"

The Internet Protocol version 6 (IPv6) address of the interface. Currently not supported.

", + "ServiceUnavailableError$message":"

A human-readable message describing the error that occurred.

", + "UpdateGatewayInformationOutput$GatewayName":null, + "VolumeRecoveryPointInfo$VolumeRecoveryPointTime":null, + "errorDetails$key":null, + "errorDetails$value":null } } } diff --git a/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/examples-1.json b/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/examples-1.json index 7cc0d7d41..6bdb137cc 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/examples-1.json +++ b/vendor/github.com/aws/aws-sdk-go/models/apis/storagegateway/2013-06-30/examples-1.json @@ -1,1380 +1,1380 @@ { - "version": "1.0", - "examples": { - "ActivateGateway": [ + "version":"1.0", + "examples":{ + "ActivateGateway":[ { - "input": { - "ActivationKey": "29AV1-3OFV9-VVIUB-NKT0I-LRO6V", - "GatewayName": "My_Gateway", - "GatewayRegion": "us-east-1", - "GatewayTimezone": "GMT-12:00", - "GatewayType": "STORED", - "MediumChangerType": "AWS-Gateway-VTL", - "TapeDriveType": "IBM-ULT3580-TD5" + "input":{ + "ActivationKey":"29AV1-3OFV9-VVIUB-NKT0I-LRO6V", + "GatewayName":"My_Gateway", + "GatewayRegion":"us-east-1", + "GatewayTimezone":"GMT-12:00", + "GatewayType":"STORED", + "MediumChangerType":"AWS-Gateway-VTL", + "TapeDriveType":"IBM-ULT3580-TD5" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Activates the gateway you previously deployed on your host.", - "id": "to-activate-the-gateway-1471281611207", - "title": "To activate the gateway" + "description":"Activates the gateway you previously deployed on your host.", + "id":"to-activate-the-gateway-1471281611207", + "title":"To activate the gateway" } ], - "AddCache": [ + "AddCache":[ { - "input": { - "DiskIds": [ + "input":{ + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:03:00.0-scsi-0:0:1:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The following example shows a request that activates a gateway-stored volume.", - "id": "to-add-a-cache-1471043606854", - "title": "To add a cache" + "description":"The following example shows a request that activates a gateway-stored volume.", + "id":"to-add-a-cache-1471043606854", + "title":"To add a cache" } ], - "AddTagsToResource": [ + "AddTagsToResource":[ { - "input": { - "ResourceARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B", - "Tags": [ + "input":{ + "ResourceARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B", + "Tags":[ { - "Key": "Dev Gatgeway Region", - "Value": "East Coast" + "Key":"Dev Gatgeway Region", + "Value":"East Coast" } ] }, - "output": { - "ResourceARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" + "output":{ + "ResourceARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Adds one or more tags to the specified resource.", - "id": "to-add-tags-to-resource-1471283689460", - "title": "To add tags to resource" + "description":"Adds one or more tags to the specified resource.", + "id":"to-add-tags-to-resource-1471283689460", + "title":"To add tags to resource" } ], - "AddUploadBuffer": [ + "AddUploadBuffer":[ { - "input": { - "DiskIds": [ + "input":{ + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:03:00.0-scsi-0:0:1:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Configures one or more gateway local disks as upload buffer for a specified gateway.", - "id": "to-add-upload-buffer-on-local-disk-1471293902847", - "title": "To add upload buffer on local disk" + "description":"Configures one or more gateway local disks as upload buffer for a specified gateway.", + "id":"to-add-upload-buffer-on-local-disk-1471293902847", + "title":"To add upload buffer on local disk" } ], - "AddWorkingStorage": [ + "AddWorkingStorage":[ { - "input": { - "DiskIds": [ + "input":{ + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:03:00.0-scsi-0:0:1:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Configures one or more gateway local disks as working storage for a gateway. (Working storage is also referred to as upload buffer.)", - "id": "to-add-storage-on-local-disk-1471294305401", - "title": "To add storage on local disk" + "description":"Configures one or more gateway local disks as working storage for a gateway. (Working storage is also referred to as upload buffer.)", + "id":"to-add-storage-on-local-disk-1471294305401", + "title":"To add storage on local disk" } ], - "CancelArchival": [ + "CancelArchival":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Cancels archiving of a virtual tape to the virtual tape shelf (VTS) after the archiving process is initiated.", - "id": "to-cancel-virtual-tape-archiving-1471294865203", - "title": "To cancel virtual tape archiving" + "description":"Cancels archiving of a virtual tape to the virtual tape shelf (VTS) after the archiving process is initiated.", + "id":"to-cancel-virtual-tape-archiving-1471294865203", + "title":"To cancel virtual tape archiving" } ], - "CancelRetrieval": [ + "CancelRetrieval":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Cancels retrieval of a virtual tape from the virtual tape shelf (VTS) to a gateway after the retrieval process is initiated.", - "id": "to-cancel-virtual-tape-retrieval-1471295704491", - "title": "To cancel virtual tape retrieval" + "description":"Cancels retrieval of a virtual tape from the virtual tape shelf (VTS) to a gateway after the retrieval process is initiated.", + "id":"to-cancel-virtual-tape-retrieval-1471295704491", + "title":"To cancel virtual tape retrieval" } ], - "CreateCachediSCSIVolume": [ + "CreateCachediSCSIVolume":[ { - "input": { - "ClientToken": "cachedvol112233", - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "NetworkInterfaceId": "10.1.1.1", - "SnapshotId": "snap-f47b7b94", - "TargetName": "my-volume", - "VolumeSizeInBytes": 536870912000 + "input":{ + "ClientToken":"cachedvol112233", + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "NetworkInterfaceId":"10.1.1.1", + "SnapshotId":"snap-f47b7b94", + "TargetName":"my-volume", + "VolumeSizeInBytes":536870912000 }, - "output": { - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "output":{ + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Creates a cached volume on a specified cached gateway.", - "id": "to-create-a-cached-iscsi-volume-1471296661787", - "title": "To create a cached iSCSI volume" + "description":"Creates a cached volume on a specified cached gateway.", + "id":"to-create-a-cached-iscsi-volume-1471296661787", + "title":"To create a cached iSCSI volume" } ], - "CreateSnapshot": [ + "CreateSnapshot":[ { - "input": { - "SnapshotDescription": "My root volume snapshot as of 10/03/2017", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "input":{ + "SnapshotDescription":"My root volume snapshot as of 10/03/2017", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "output": { - "SnapshotId": "snap-78e22663", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "output":{ + "SnapshotId":"snap-78e22663", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Initiates an ad-hoc snapshot of a gateway volume.", - "id": "to-create-a-snapshot-of-a-gateway-volume-1471301469561", - "title": "To create a snapshot of a gateway volume" + "description":"Initiates an ad-hoc snapshot of a gateway volume.", + "id":"to-create-a-snapshot-of-a-gateway-volume-1471301469561", + "title":"To create a snapshot of a gateway volume" } ], - "CreateSnapshotFromVolumeRecoveryPoint": [ + "CreateSnapshotFromVolumeRecoveryPoint":[ { - "input": { - "SnapshotDescription": "My root volume snapshot as of 2017-06-30T10:10:10.000Z", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "input":{ + "SnapshotDescription":"My root volume snapshot as of 2017-06-30T10:10:10.000Z", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "output": { - "SnapshotId": "snap-78e22663", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", - "VolumeRecoveryPointTime": "2017-06-30T10:10:10.000Z" + "output":{ + "SnapshotId":"snap-78e22663", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", + "VolumeRecoveryPointTime":"2017-06-30T10:10:10.000Z" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Initiates a snapshot of a gateway from a volume recovery point.", - "id": "to-create-a-snapshot-of-a-gateway-volume-1471301469561", - "title": "To create a snapshot of a gateway volume" + "description":"Initiates a snapshot of a gateway from a volume recovery point.", + "id":"to-create-a-snapshot-of-a-gateway-volume-1471301469561", + "title":"To create a snapshot of a gateway volume" } ], - "CreateStorediSCSIVolume": [ + "CreateStorediSCSIVolume":[ { - "input": { - "DiskId": "pci-0000:03:00.0-scsi-0:0:0:0", - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "NetworkInterfaceId": "10.1.1.1", - "PreserveExistingData": true, - "SnapshotId": "snap-f47b7b94", - "TargetName": "my-volume" + "input":{ + "DiskId":"pci-0000:03:00.0-scsi-0:0:0:0", + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "NetworkInterfaceId":"10.1.1.1", + "PreserveExistingData":true, + "SnapshotId":"snap-f47b7b94", + "TargetName":"my-volume" }, - "output": { - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", - "VolumeSizeInBytes": 1099511627776 + "output":{ + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", + "VolumeSizeInBytes":1099511627776 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Creates a stored volume on a specified stored gateway.", - "id": "to-create-a-stored-iscsi-volume-1471367662813", - "title": "To create a stored iSCSI volume" + "description":"Creates a stored volume on a specified stored gateway.", + "id":"to-create-a-stored-iscsi-volume-1471367662813", + "title":"To create a stored iSCSI volume" } ], - "CreateTapeWithBarcode": [ + "CreateTapeWithBarcode":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "TapeBarcode": "TEST12345", - "TapeSizeInBytes": 107374182400 + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "TapeBarcode":"TEST12345", + "TapeSizeInBytes":107374182400 }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST12345" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST12345" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Creates a virtual tape by using your own barcode.", - "id": "to-create-a-virtual-tape-using-a-barcode-1471371842452", - "title": "To create a virtual tape using a barcode" + "description":"Creates a virtual tape by using your own barcode.", + "id":"to-create-a-virtual-tape-using-a-barcode-1471371842452", + "title":"To create a virtual tape using a barcode" } ], - "CreateTapes": [ + "CreateTapes":[ { - "input": { - "ClientToken": "77777", - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "NumTapesToCreate": 3, - "TapeBarcodePrefix": "TEST", - "TapeSizeInBytes": 107374182400 + "input":{ + "ClientToken":"77777", + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "NumTapesToCreate":3, + "TapeBarcodePrefix":"TEST", + "TapeSizeInBytes":107374182400 }, - "output": { - "TapeARNs": [ + "output":{ + "TapeARNs":[ "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST38A29D", "arn:aws:storagegateway:us-east-1:204469490176:tape/TEST3AA29F", "arn:aws:storagegateway:us-east-1:204469490176:tape/TEST3BA29E" ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Creates one or more virtual tapes.", - "id": "to-create-a-virtual-tape-1471372061659", - "title": "To create a virtual tape" + "description":"Creates one or more virtual tapes.", + "id":"to-create-a-virtual-tape-1471372061659", + "title":"To create a virtual tape" } ], - "DeleteBandwidthRateLimit": [ + "DeleteBandwidthRateLimit":[ { - "input": { - "BandwidthType": "All", - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "BandwidthType":"All", + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Deletes the bandwidth rate limits of a gateway; either the upload or download limit, or both.", - "id": "to-delete-bandwidth-rate-limits-of-gateway-1471373225520", - "title": "To delete bandwidth rate limits of gateway" + "description":"Deletes the bandwidth rate limits of a gateway; either the upload or download limit, or both.", + "id":"to-delete-bandwidth-rate-limits-of-gateway-1471373225520", + "title":"To delete bandwidth rate limits of gateway" } ], - "DeleteChapCredentials": [ + "DeleteChapCredentials":[ { - "input": { - "InitiatorName": "iqn.1991-05.com.microsoft:computername.domain.example.com", - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "input":{ + "InitiatorName":"iqn.1991-05.com.microsoft:computername.domain.example.com", + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" }, - "output": { - "InitiatorName": "iqn.1991-05.com.microsoft:computername.domain.example.com", - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "output":{ + "InitiatorName":"iqn.1991-05.com.microsoft:computername.domain.example.com", + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Deletes Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target and initiator pair.", - "id": "to-delete-chap-credentials-1471375025612", - "title": "To delete CHAP credentials" + "description":"Deletes Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target and initiator pair.", + "id":"to-delete-chap-credentials-1471375025612", + "title":"To delete CHAP credentials" } ], - "DeleteGateway": [ + "DeleteGateway":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation deletes the gateway, but not the gateway's VM from the host computer.", - "id": "to-delete-a-gatgeway-1471381697333", - "title": "To delete a gatgeway" + "description":"This operation deletes the gateway, but not the gateway's VM from the host computer.", + "id":"to-delete-a-gatgeway-1471381697333", + "title":"To delete a gatgeway" } ], - "DeleteSnapshotSchedule": [ + "DeleteSnapshotSchedule":[ { - "input": { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "input":{ + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "output": { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "output":{ + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This action enables you to delete a snapshot schedule for a volume.", - "id": "to-delete-a-snapshot-of-a-volume-1471382234377", - "title": "To delete a snapshot of a volume" + "description":"This action enables you to delete a snapshot schedule for a volume.", + "id":"to-delete-a-snapshot-of-a-volume-1471382234377", + "title":"To delete a snapshot of a volume" } ], - "DeleteTape": [ + "DeleteTape":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:204469490176:gateway/sgw-12A3456B", - "TapeARN": "arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:204469490176:gateway/sgw-12A3456B", + "TapeARN":"arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This example deletes the specified virtual tape.", - "id": "to-delete-a-virtual-tape-1471382444157", - "title": "To delete a virtual tape" + "description":"This example deletes the specified virtual tape.", + "id":"to-delete-a-virtual-tape-1471382444157", + "title":"To delete a virtual tape" } ], - "DeleteTapeArchive": [ + "DeleteTapeArchive":[ { - "input": { - "TapeARN": "arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" + "input":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:204469490176:tape/TEST05A2A0" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Deletes the specified virtual tape from the virtual tape shelf (VTS).", - "id": "to-delete-a-virtual-tape-from-the-shelf-vts-1471383964329", - "title": "To delete a virtual tape from the shelf (VTS)" + "description":"Deletes the specified virtual tape from the virtual tape shelf (VTS).", + "id":"to-delete-a-virtual-tape-from-the-shelf-vts-1471383964329", + "title":"To delete a virtual tape from the shelf (VTS)" } ], - "DeleteVolume": [ + "DeleteVolume":[ { - "input": { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "input":{ + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "output": { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "output":{ + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Deletes the specified gateway volume that you previously created using the CreateCachediSCSIVolume or CreateStorediSCSIVolume API.", - "id": "to-delete-a-gateway-volume-1471384418416", - "title": "To delete a gateway volume" + "description":"Deletes the specified gateway volume that you previously created using the CreateCachediSCSIVolume or CreateStorediSCSIVolume API.", + "id":"to-delete-a-gateway-volume-1471384418416", + "title":"To delete a gateway volume" } ], - "DescribeBandwidthRateLimit": [ + "DescribeBandwidthRateLimit":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "AverageDownloadRateLimitInBitsPerSec": 204800, - "AverageUploadRateLimitInBitsPerSec": 102400, - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "AverageDownloadRateLimitInBitsPerSec":204800, + "AverageUploadRateLimitInBitsPerSec":102400, + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns a value for a bandwidth rate limit if set. If not set, then only the gateway ARN is returned.", - "id": "to-describe-the-bandwidth-rate-limits-of-a-gateway-1471384826404", - "title": "To describe the bandwidth rate limits of a gateway" + "description":"Returns a value for a bandwidth rate limit if set. If not set, then only the gateway ARN is returned.", + "id":"to-describe-the-bandwidth-rate-limits-of-a-gateway-1471384826404", + "title":"To describe the bandwidth rate limits of a gateway" } ], - "DescribeCache": [ + "DescribeCache":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "CacheAllocatedInBytes": 2199023255552, - "CacheDirtyPercentage": 0.07, - "CacheHitPercentage": 99.68, - "CacheMissPercentage": 0.32, - "CacheUsedPercentage": 0.07, - "DiskIds": [ + "output":{ + "CacheAllocatedInBytes":2199023255552, + "CacheDirtyPercentage":0.07, + "CacheHitPercentage":99.68, + "CacheMissPercentage":0.32, + "CacheUsedPercentage":0.07, + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:04:00.0-scsi-0:1:0:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns information about the cache of a gateway.", - "id": "to-describe-cache-information-1471385756036", - "title": "To describe cache information" + "description":"Returns information about the cache of a gateway.", + "id":"to-describe-cache-information-1471385756036", + "title":"To describe cache information" } ], - "DescribeCachediSCSIVolumes": [ + "DescribeCachediSCSIVolumes":[ { - "input": { - "VolumeARNs": [ + "input":{ + "VolumeARNs":[ "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" ] }, - "output": { - "CachediSCSIVolumes": [ + "output":{ + "CachediSCSIVolumes":[ { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", - "VolumeId": "vol-1122AABB", - "VolumeSizeInBytes": 1099511627776, - "VolumeStatus": "AVAILABLE", - "VolumeType": "CACHED iSCSI", - "VolumeiSCSIAttributes": { - "ChapEnabled": true, - "LunNumber": 1, - "NetworkInterfaceId": "10.243.43.207", - "NetworkInterfacePort": 3260, - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", + "VolumeId":"vol-1122AABB", + "VolumeSizeInBytes":1099511627776, + "VolumeStatus":"AVAILABLE", + "VolumeType":"CACHED iSCSI", + "VolumeiSCSIAttributes":{ + "ChapEnabled":true, + "LunNumber":1, + "NetworkInterfaceId":"10.243.43.207", + "NetworkInterfacePort":3260, + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" } } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns a description of the gateway cached iSCSI volumes specified in the request.", - "id": "to-describe-gateway-cached-iscsi-volumes-1471458094649", - "title": "To describe gateway cached iSCSI volumes" + "description":"Returns a description of the gateway cached iSCSI volumes specified in the request.", + "id":"to-describe-gateway-cached-iscsi-volumes-1471458094649", + "title":"To describe gateway cached iSCSI volumes" } ], - "DescribeChapCredentials": [ + "DescribeChapCredentials":[ { - "input": { - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "input":{ + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" }, - "output": { - "ChapCredentials": [ + "output":{ + "ChapCredentials":[ { - "InitiatorName": "iqn.1991-05.com.microsoft:computername.domain.example.com", - "SecretToAuthenticateInitiator": "111111111111", - "SecretToAuthenticateTarget": "222222222222", - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "InitiatorName":"iqn.1991-05.com.microsoft:computername.domain.example.com", + "SecretToAuthenticateInitiator":"111111111111", + "SecretToAuthenticateTarget":"222222222222", + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns an array of Challenge-Handshake Authentication Protocol (CHAP) credentials information for a specified iSCSI target, one for each target-initiator pair.", - "id": "to-describe-chap-credetnitals-for-an-iscsi-1471467462967", - "title": "To describe CHAP credetnitals for an iSCSI" + "description":"Returns an array of Challenge-Handshake Authentication Protocol (CHAP) credentials information for a specified iSCSI target, one for each target-initiator pair.", + "id":"to-describe-chap-credetnitals-for-an-iscsi-1471467462967", + "title":"To describe CHAP credetnitals for an iSCSI" } ], - "DescribeGatewayInformation": [ + "DescribeGatewayInformation":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "GatewayId": "sgw-AABB1122", - "GatewayName": "My_Gateway", - "GatewayNetworkInterfaces": [ + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "GatewayId":"sgw-AABB1122", + "GatewayName":"My_Gateway", + "GatewayNetworkInterfaces":[ { - "Ipv4Address": "10.35.69.216" + "Ipv4Address":"10.35.69.216" } ], - "GatewayState": "STATE_RUNNING", - "GatewayTimezone": "GMT-8:00", - "GatewayType": "STORED", - "LastSoftwareUpdate": "2016-01-02T16:00:00", - "NextUpdateAvailabilityDate": "2017-01-02T16:00:00" + "GatewayState":"STATE_RUNNING", + "GatewayTimezone":"GMT-8:00", + "GatewayType":"STORED", + "LastSoftwareUpdate":"2016-01-02T16:00:00", + "NextUpdateAvailabilityDate":"2017-01-02T16:00:00" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns metadata about a gateway such as its name, network interfaces, configured time zone, and the state (whether the gateway is running or not).", - "id": "to-describe-metadata-about-the-gateway-1471467849079", - "title": "To describe metadata about the gateway" + "description":"Returns metadata about a gateway such as its name, network interfaces, configured time zone, and the state (whether the gateway is running or not).", + "id":"to-describe-metadata-about-the-gateway-1471467849079", + "title":"To describe metadata about the gateway" } ], - "DescribeMaintenanceStartTime": [ + "DescribeMaintenanceStartTime":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "DayOfWeek": 2, - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "HourOfDay": 15, - "MinuteOfHour": 35, - "Timezone": "GMT+7:00" + "output":{ + "DayOfWeek":2, + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "HourOfDay":15, + "MinuteOfHour":35, + "Timezone":"GMT+7:00" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns your gateway's weekly maintenance start time including the day and time of the week.", - "id": "to-describe-gateways-maintenance-start-time-1471470727387", - "title": "To describe gateway's maintenance start time" + "description":"Returns your gateway's weekly maintenance start time including the day and time of the week.", + "id":"to-describe-gateways-maintenance-start-time-1471470727387", + "title":"To describe gateway's maintenance start time" } ], - "DescribeSnapshotSchedule": [ + "DescribeSnapshotSchedule":[ { - "input": { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "input":{ + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "output": { - "Description": "sgw-AABB1122:vol-AABB1122:Schedule", - "RecurrenceInHours": 24, - "StartAt": 6, - "Timezone": "GMT+7:00", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "output":{ + "Description":"sgw-AABB1122:vol-AABB1122:Schedule", + "RecurrenceInHours":24, + "StartAt":6, + "Timezone":"GMT+7:00", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Describes the snapshot schedule for the specified gateway volume including intervals at which snapshots are automatically initiated.", - "id": "to-describe-snapshot-schedule-for-gateway-volume-1471471139538", - "title": "To describe snapshot schedule for gateway volume" + "description":"Describes the snapshot schedule for the specified gateway volume including intervals at which snapshots are automatically initiated.", + "id":"to-describe-snapshot-schedule-for-gateway-volume-1471471139538", + "title":"To describe snapshot schedule for gateway volume" } ], - "DescribeStorediSCSIVolumes": [ + "DescribeStorediSCSIVolumes":[ { - "input": { - "VolumeARNs": [ + "input":{ + "VolumeARNs":[ "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" ] }, - "output": { - "StorediSCSIVolumes": [ + "output":{ + "StorediSCSIVolumes":[ { - "PreservedExistingData": false, - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", - "VolumeDiskId": "pci-0000:03:00.0-scsi-0:0:0:0", - "VolumeId": "vol-1122AABB", - "VolumeProgress": 23.7, - "VolumeSizeInBytes": 1099511627776, - "VolumeStatus": "BOOTSTRAPPING", - "VolumeiSCSIAttributes": { - "ChapEnabled": true, - "NetworkInterfaceId": "10.243.43.207", - "NetworkInterfacePort": 3260, - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "PreservedExistingData":false, + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", + "VolumeDiskId":"pci-0000:03:00.0-scsi-0:0:0:0", + "VolumeId":"vol-1122AABB", + "VolumeProgress":23.7, + "VolumeSizeInBytes":1099511627776, + "VolumeStatus":"BOOTSTRAPPING", + "VolumeiSCSIAttributes":{ + "ChapEnabled":true, + "NetworkInterfaceId":"10.243.43.207", + "NetworkInterfacePort":3260, + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" } } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns the description of the gateway volumes specified in the request belonging to the same gateway.", - "id": "to-describe-the-volumes-of-a-gateway-1471472640660", - "title": "To describe the volumes of a gateway" + "description":"Returns the description of the gateway volumes specified in the request belonging to the same gateway.", + "id":"to-describe-the-volumes-of-a-gateway-1471472640660", + "title":"To describe the volumes of a gateway" } ], - "DescribeTapeArchives": [ + "DescribeTapeArchives":[ { - "input": { - "Limit": 123, - "Marker": "1", - "TapeARNs": [ + "input":{ + "Limit":123, + "Marker":"1", + "TapeARNs":[ "arn:aws:storagegateway:us-east-1:999999999999:tape/AM08A1AD", "arn:aws:storagegateway:us-east-1:999999999999:tape/AMZN01A2A4" ] }, - "output": { - "Marker": "1", - "TapeArchives": [ + "output":{ + "Marker":"1", + "TapeArchives":[ { - "CompletionTime": "2016-12-16T13:50Z", - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999:tape/AM08A1AD", - "TapeBarcode": "AM08A1AD", - "TapeSizeInBytes": 107374182400, - "TapeStatus": "ARCHIVED" + "CompletionTime":"2016-12-16T13:50Z", + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999:tape/AM08A1AD", + "TapeBarcode":"AM08A1AD", + "TapeSizeInBytes":107374182400, + "TapeStatus":"ARCHIVED" }, { - "CompletionTime": "2016-12-16T13:59Z", - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999:tape/AMZN01A2A4", - "TapeBarcode": "AMZN01A2A4", - "TapeSizeInBytes": 429496729600, - "TapeStatus": "ARCHIVED" + "CompletionTime":"2016-12-16T13:59Z", + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999:tape/AMZN01A2A4", + "TapeBarcode":"AMZN01A2A4", + "TapeSizeInBytes":429496729600, + "TapeStatus":"ARCHIVED" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns a description of specified virtual tapes in the virtual tape shelf (VTS).", - "id": "to-describe-virtual-tapes-in-the-vts-1471473188198", - "title": "To describe virtual tapes in the VTS" + "description":"Returns a description of specified virtual tapes in the virtual tape shelf (VTS).", + "id":"to-describe-virtual-tapes-in-the-vts-1471473188198", + "title":"To describe virtual tapes in the VTS" } ], - "DescribeTapeRecoveryPoints": [ + "DescribeTapeRecoveryPoints":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "Limit": 1, - "Marker": "1" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "Limit":1, + "Marker":"1" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "Marker": "1", - "TapeRecoveryPointInfos": [ + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "Marker":"1", + "TapeRecoveryPointInfos":[ { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999:tape/AMZN01A2A4", - "TapeRecoveryPointTime": "2016-12-16T13:50Z", - "TapeSizeInBytes": 1471550497, - "TapeStatus": "AVAILABLE" + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999:tape/AMZN01A2A4", + "TapeRecoveryPointTime":"2016-12-16T13:50Z", + "TapeSizeInBytes":1471550497, + "TapeStatus":"AVAILABLE" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns a list of virtual tape recovery points that are available for the specified gateway-VTL.", - "id": "to-describe-virtual-tape-recovery-points-1471542042026", - "title": "To describe virtual tape recovery points" + "description":"Returns a list of virtual tape recovery points that are available for the specified gateway-VTL.", + "id":"to-describe-virtual-tape-recovery-points-1471542042026", + "title":"To describe virtual tape recovery points" } ], - "DescribeTapes": [ + "DescribeTapes":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "Limit": 2, - "Marker": "1", - "TapeARNs": [ + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "Limit":2, + "Marker":"1", + "TapeARNs":[ "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST04A2A1", "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST05A2A0" ] }, - "output": { - "Marker": "1", - "Tapes": [ + "output":{ + "Marker":"1", + "Tapes":[ { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST04A2A1", - "TapeBarcode": "TEST04A2A1", - "TapeSizeInBytes": 107374182400, - "TapeStatus": "AVAILABLE" + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST04A2A1", + "TapeBarcode":"TEST04A2A1", + "TapeSizeInBytes":107374182400, + "TapeStatus":"AVAILABLE" }, { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST05A2A0", - "TapeBarcode": "TEST05A2A0", - "TapeSizeInBytes": 107374182400, - "TapeStatus": "AVAILABLE" + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST05A2A0", + "TapeBarcode":"TEST05A2A0", + "TapeSizeInBytes":107374182400, + "TapeStatus":"AVAILABLE" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns a description of the specified Amazon Resource Name (ARN) of virtual tapes. If a TapeARN is not specified, returns a description of all virtual tapes.", - "id": "to-describe-virtual-tapes-associated-with-gateway-1471629287727", - "title": "To describe virtual tape(s) associated with gateway" + "description":"Returns a description of the specified Amazon Resource Name (ARN) of virtual tapes. If a TapeARN is not specified, returns a description of all virtual tapes.", + "id":"to-describe-virtual-tapes-associated-with-gateway-1471629287727", + "title":"To describe virtual tape(s) associated with gateway" } ], - "DescribeUploadBuffer": [ + "DescribeUploadBuffer":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "DiskIds": [ + "output":{ + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:04:00.0-scsi-0:1:0:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "UploadBufferAllocatedInBytes": 0, - "UploadBufferUsedInBytes": 161061273600 + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "UploadBufferAllocatedInBytes":0, + "UploadBufferUsedInBytes":161061273600 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns information about the upload buffer of a gateway including disk IDs and the amount of upload buffer space allocated/used.", - "id": "to-describe-upload-buffer-of-gateway-1471631099003", - "title": "To describe upload buffer of gateway" + "description":"Returns information about the upload buffer of a gateway including disk IDs and the amount of upload buffer space allocated/used.", + "id":"to-describe-upload-buffer-of-gateway-1471631099003", + "title":"To describe upload buffer of gateway" }, { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "DiskIds": [ + "output":{ + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:04:00.0-scsi-0:1:0:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "UploadBufferAllocatedInBytes": 161061273600, - "UploadBufferUsedInBytes": 0 + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "UploadBufferAllocatedInBytes":161061273600, + "UploadBufferUsedInBytes":0 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns information about the upload buffer of a gateway including disk IDs and the amount of upload buffer space allocated and used.", - "id": "to-describe-upload-buffer-of-a-gateway--1471904566370", - "title": "To describe upload buffer of a gateway" + "description":"Returns information about the upload buffer of a gateway including disk IDs and the amount of upload buffer space allocated and used.", + "id":"to-describe-upload-buffer-of-a-gateway--1471904566370", + "title":"To describe upload buffer of a gateway" } ], - "DescribeVTLDevices": [ + "DescribeVTLDevices":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "Limit": 123, - "Marker": "1", - "VTLDeviceARNs": [ + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "Limit":123, + "Marker":"1", + "VTLDeviceARNs":[ ] }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "Marker": "1", - "VTLDevices": [ + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "Marker":"1", + "VTLDevices":[ { - "DeviceiSCSIAttributes": { - "ChapEnabled": false, - "NetworkInterfaceId": "10.243.43.207", - "NetworkInterfacePort": 3260, - "TargetARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:sgw-1fad4876-mediachanger" + "DeviceiSCSIAttributes":{ + "ChapEnabled":false, + "NetworkInterfaceId":"10.243.43.207", + "NetworkInterfacePort":3260, + "TargetARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:sgw-1fad4876-mediachanger" }, - "VTLDeviceARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_MEDIACHANGER_00001", - "VTLDeviceProductIdentifier": "L700", - "VTLDeviceType": "Medium Changer", - "VTLDeviceVendor": "STK" + "VTLDeviceARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_MEDIACHANGER_00001", + "VTLDeviceProductIdentifier":"L700", + "VTLDeviceType":"Medium Changer", + "VTLDeviceVendor":"STK" }, { - "DeviceiSCSIAttributes": { - "ChapEnabled": false, - "NetworkInterfaceId": "10.243.43.209", - "NetworkInterfacePort": 3260, - "TargetARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:sgw-1fad4876-tapedrive-01" + "DeviceiSCSIAttributes":{ + "ChapEnabled":false, + "NetworkInterfaceId":"10.243.43.209", + "NetworkInterfacePort":3260, + "TargetARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:sgw-1fad4876-tapedrive-01" }, - "VTLDeviceARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_TAPEDRIVE_00001", - "VTLDeviceProductIdentifier": "ULT3580-TD5", - "VTLDeviceType": "Tape Drive", - "VTLDeviceVendor": "IBM" + "VTLDeviceARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_TAPEDRIVE_00001", + "VTLDeviceProductIdentifier":"ULT3580-TD5", + "VTLDeviceType":"Tape Drive", + "VTLDeviceVendor":"IBM" }, { - "DeviceiSCSIAttributes": { - "ChapEnabled": false, - "NetworkInterfaceId": "10.243.43.209", - "NetworkInterfacePort": 3260, - "TargetARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:sgw-1fad4876-tapedrive-02" + "DeviceiSCSIAttributes":{ + "ChapEnabled":false, + "NetworkInterfaceId":"10.243.43.209", + "NetworkInterfacePort":3260, + "TargetARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:sgw-1fad4876-tapedrive-02" }, - "VTLDeviceARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_TAPEDRIVE_00002", - "VTLDeviceProductIdentifier": "ULT3580-TD5", - "VTLDeviceType": "Tape Drive", - "VTLDeviceVendor": "IBM" + "VTLDeviceARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_TAPEDRIVE_00002", + "VTLDeviceProductIdentifier":"ULT3580-TD5", + "VTLDeviceType":"Tape Drive", + "VTLDeviceVendor":"IBM" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Returns a description of virtual tape library (VTL) devices for the specified gateway.", - "id": "to-describe-virtual-tape-library-vtl-devices-of-a-single-gateway-1471906071410", - "title": "To describe virtual tape library (VTL) devices of a single gateway" + "description":"Returns a description of virtual tape library (VTL) devices for the specified gateway.", + "id":"to-describe-virtual-tape-library-vtl-devices-of-a-single-gateway-1471906071410", + "title":"To describe virtual tape library (VTL) devices of a single gateway" } ], - "DescribeWorkingStorage": [ + "DescribeWorkingStorage":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "DiskIds": [ + "output":{ + "DiskIds":[ "pci-0000:03:00.0-scsi-0:0:0:0", "pci-0000:03:00.0-scsi-0:0:1:0" ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "WorkingStorageAllocatedInBytes": 2199023255552, - "WorkingStorageUsedInBytes": 789207040 + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "WorkingStorageAllocatedInBytes":2199023255552, + "WorkingStorageUsedInBytes":789207040 }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation is supported only for the gateway-stored volume architecture. This operation is deprecated in cached-volumes API version (20120630). Use DescribeUploadBuffer instead.", - "id": "to-describe-the-working-storage-of-a-gateway-depreciated-1472070842332", - "title": "To describe the working storage of a gateway [Depreciated]" + "description":"This operation is supported only for the gateway-stored volume architecture. This operation is deprecated in cached-volumes API version (20120630). Use DescribeUploadBuffer instead.", + "id":"to-describe-the-working-storage-of-a-gateway-depreciated-1472070842332", + "title":"To describe the working storage of a gateway [Depreciated]" } ], - "DisableGateway": [ + "DisableGateway":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Disables a gateway when the gateway is no longer functioning. Use this operation for a gateway-VTL that is not reachable or not functioning.", - "id": "to-disable-a-gateway-when-it-is-no-longer-functioning-1472076046936", - "title": "To disable a gateway when it is no longer functioning" + "description":"Disables a gateway when the gateway is no longer functioning. Use this operation for a gateway-VTL that is not reachable or not functioning.", + "id":"to-disable-a-gateway-when-it-is-no-longer-functioning-1472076046936", + "title":"To disable a gateway when it is no longer functioning" } ], - "ListGateways": [ + "ListGateways":[ { - "input": { - "Limit": 2, - "Marker": "1" + "input":{ + "Limit":2, + "Marker":"1" }, - "output": { - "Gateways": [ + "output":{ + "Gateways":[ { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-23A4567C" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-23A4567C" } ], - "Marker": "1" + "Marker":"1" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Lists gateways owned by an AWS account in a specified region as requested. Results are sorted by gateway ARN up to a maximum of 100 gateways.", - "id": "to-lists-region-specific-gateways-per-aws-account-1472077860657", - "title": "To lists region specific gateways per AWS account" + "description":"Lists gateways owned by an AWS account in a specified region as requested. Results are sorted by gateway ARN up to a maximum of 100 gateways.", + "id":"to-lists-region-specific-gateways-per-aws-account-1472077860657", + "title":"To lists region specific gateways per AWS account" } ], - "ListLocalDisks": [ + "ListLocalDisks":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "Disks": [ + "output":{ + "Disks":[ { - "DiskAllocationType": "CACHE_STORAGE", - "DiskId": "pci-0000:03:00.0-scsi-0:0:0:0", - "DiskNode": "SCSI(0:0)", - "DiskPath": "/dev/sda", - "DiskSizeInBytes": 1099511627776, - "DiskStatus": "missing" + "DiskAllocationType":"CACHE_STORAGE", + "DiskId":"pci-0000:03:00.0-scsi-0:0:0:0", + "DiskNode":"SCSI(0:0)", + "DiskPath":"/dev/sda", + "DiskSizeInBytes":1099511627776, + "DiskStatus":"missing" }, { - "DiskAllocationResource": "", - "DiskAllocationType": "UPLOAD_BUFFER", - "DiskId": "pci-0000:03:00.0-scsi-0:0:1:0", - "DiskNode": "SCSI(0:1)", - "DiskPath": "/dev/sdb", - "DiskSizeInBytes": 1099511627776, - "DiskStatus": "present" + "DiskAllocationResource":"", + "DiskAllocationType":"UPLOAD_BUFFER", + "DiskId":"pci-0000:03:00.0-scsi-0:0:1:0", + "DiskNode":"SCSI(0:1)", + "DiskPath":"/dev/sdb", + "DiskSizeInBytes":1099511627776, + "DiskStatus":"present" } ], - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "The request returns a list of all disks, specifying which are configured as working storage, cache storage, or stored volume or not configured at all.", - "id": "to-list-the-gateways-local-disks-1472079564618", - "title": "To list the gateway's local disks" + "description":"The request returns a list of all disks, specifying which are configured as working storage, cache storage, or stored volume or not configured at all.", + "id":"to-list-the-gateways-local-disks-1472079564618", + "title":"To list the gateway's local disks" } ], - "ListTagsForResource": [ + "ListTagsForResource":[ { - "input": { - "Limit": 1, - "Marker": "1", - "ResourceARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" + "input":{ + "Limit":1, + "Marker":"1", + "ResourceARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" }, - "output": { - "Marker": "1", - "ResourceARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B", - "Tags": [ + "output":{ + "Marker":"1", + "ResourceARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B", + "Tags":[ { - "Key": "Dev Gatgeway Region", - "Value": "East Coast" + "Key":"Dev Gatgeway Region", + "Value":"East Coast" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Lists the tags that have been added to the specified resource.", - "id": "to-list-tags-that-have-been-added-to-a-resource-1472080268972", - "title": "To list tags that have been added to a resource" + "description":"Lists the tags that have been added to the specified resource.", + "id":"to-list-tags-that-have-been-added-to-a-resource-1472080268972", + "title":"To list tags that have been added to a resource" } ], - "ListVolumeRecoveryPoints": [ + "ListVolumeRecoveryPoints":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "VolumeRecoveryPointInfos": [ + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "VolumeRecoveryPointInfos":[ { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", - "VolumeRecoveryPointTime": "2012-09-04T21:08:44.627Z", - "VolumeSizeInBytes": 536870912000 + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", + "VolumeRecoveryPointTime":"2012-09-04T21:08:44.627Z", + "VolumeSizeInBytes":536870912000 } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Lists the recovery points for a specified gateway in which all data of the volume is consistent and can be used to create a snapshot.", - "id": "to-list-recovery-points-for-a-gateway-1472143015088", - "title": "To list recovery points for a gateway" + "description":"Lists the recovery points for a specified gateway in which all data of the volume is consistent and can be used to create a snapshot.", + "id":"to-list-recovery-points-for-a-gateway-1472143015088", + "title":"To list recovery points for a gateway" } ], - "ListVolumes": [ + "ListVolumes":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "Limit": 2, - "Marker": "1" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "Limit":2, + "Marker":"1" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "Marker": "1", - "VolumeInfos": [ + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "Marker":"1", + "VolumeInfos":[ { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "GatewayId": "sgw-12A3456B", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", - "VolumeId": "vol-1122AABB", - "VolumeSizeInBytes": 107374182400, - "VolumeType": "STORED" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "GatewayId":"sgw-12A3456B", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB", + "VolumeId":"vol-1122AABB", + "VolumeSizeInBytes":107374182400, + "VolumeType":"STORED" }, { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C", - "GatewayId": "sgw-gw-13B4567C", - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C/volume/vol-3344CCDD", - "VolumeId": "vol-1122AABB", - "VolumeSizeInBytes": 107374182400, - "VolumeType": "STORED" + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C", + "GatewayId":"sgw-gw-13B4567C", + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C/volume/vol-3344CCDD", + "VolumeId":"vol-1122AABB", + "VolumeSizeInBytes":107374182400, + "VolumeType":"STORED" } ] }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Lists the iSCSI stored volumes of a gateway. Results are sorted by volume ARN up to a maximum of 100 volumes.", - "id": "to-list-the-iscsi-stored-volumes-of-a-gateway-1472145723653", - "title": "To list the iSCSI stored volumes of a gateway" + "description":"Lists the iSCSI stored volumes of a gateway. Results are sorted by volume ARN up to a maximum of 100 volumes.", + "id":"to-list-the-iscsi-stored-volumes-of-a-gateway-1472145723653", + "title":"To list the iSCSI stored volumes of a gateway" } ], - "RemoveTagsFromResource": [ + "RemoveTagsFromResource":[ { - "input": { - "ResourceARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B", - "TagKeys": [ + "input":{ + "ResourceARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B", + "TagKeys":[ "Dev Gatgeway Region", "East Coast" ] }, - "output": { - "ResourceARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" + "output":{ + "ResourceARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-11A2222B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Lists the iSCSI stored volumes of a gateway. Removes one or more tags from the specified resource.", - "id": "to-remove-tags-from-a-resource-1472147210553", - "title": "To remove tags from a resource" + "description":"Lists the iSCSI stored volumes of a gateway. Removes one or more tags from the specified resource.", + "id":"to-remove-tags-from-a-resource-1472147210553", + "title":"To remove tags from a resource" } ], - "ResetCache": [ + "ResetCache":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-13B4567C" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Resets all cache disks that have encountered a error and makes the disks available for reconfiguration as cache storage.", - "id": "to-reset-cache-disks-in-error-status-1472148909807", - "title": "To reset cache disks in error status" + "description":"Resets all cache disks that have encountered a error and makes the disks available for reconfiguration as cache storage.", + "id":"to-reset-cache-disks-in-error-status-1472148909807", + "title":"To reset cache disks in error status" } ], - "RetrieveTapeArchive": [ + "RetrieveTapeArchive":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Retrieves an archived virtual tape from the virtual tape shelf (VTS) to a gateway-VTL. Virtual tapes archived in the VTS are not associated with any gateway.", - "id": "to-retrieve-an-archived-tape-from-the-vts-1472149812358", - "title": "To retrieve an archived tape from the VTS" + "description":"Retrieves an archived virtual tape from the virtual tape shelf (VTS) to a gateway-VTL. Virtual tapes archived in the VTS are not associated with any gateway.", + "id":"to-retrieve-an-archived-tape-from-the-vts-1472149812358", + "title":"To retrieve an archived tape from the VTS" } ], - "RetrieveTapeRecoveryPoint": [ + "RetrieveTapeRecoveryPoint":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" }, - "output": { - "TapeARN": "arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" + "output":{ + "TapeARN":"arn:aws:storagegateway:us-east-1:999999999999:tape/TEST0AA2AF" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Retrieves the recovery point for the specified virtual tape.", - "id": "to-retrieve-the-recovery-point-of-a-virtual-tape-1472150014805", - "title": "To retrieve the recovery point of a virtual tape" + "description":"Retrieves the recovery point for the specified virtual tape.", + "id":"to-retrieve-the-recovery-point-of-a-virtual-tape-1472150014805", + "title":"To retrieve the recovery point of a virtual tape" } ], - "SetLocalConsolePassword": [ + "SetLocalConsolePassword":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", - "LocalConsolePassword": "PassWordMustBeAtLeast6Chars." + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B", + "LocalConsolePassword":"PassWordMustBeAtLeast6Chars." }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Sets the password for your VM local console.", - "id": "to-set-a-password-for-your-vm-1472150202632", - "title": "To set a password for your VM" + "description":"Sets the password for your VM local console.", + "id":"to-set-a-password-for-your-vm-1472150202632", + "title":"To set a password for your VM" } ], - "ShutdownGateway": [ + "ShutdownGateway":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "This operation shuts down the gateway service component running in the storage gateway's virtual machine (VM) and not the VM.", - "id": "to-shut-down-a-gateway-service-1472150508835", - "title": "To shut down a gateway service" + "description":"This operation shuts down the gateway service component running in the storage gateway's virtual machine (VM) and not the VM.", + "id":"to-shut-down-a-gateway-service-1472150508835", + "title":"To shut down a gateway service" } ], - "StartGateway": [ + "StartGateway":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Starts a gateway service that was previously shut down.", - "id": "to-start-a-gateway-service-1472150722315", - "title": "To start a gateway service" + "description":"Starts a gateway service that was previously shut down.", + "id":"to-start-a-gateway-service-1472150722315", + "title":"To start a gateway service" } ], - "UpdateBandwidthRateLimit": [ + "UpdateBandwidthRateLimit":[ { - "input": { - "AverageDownloadRateLimitInBitsPerSec": 102400, - "AverageUploadRateLimitInBitsPerSec": 51200, - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "AverageDownloadRateLimitInBitsPerSec":102400, + "AverageUploadRateLimitInBitsPerSec":51200, + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates the bandwidth rate limits of a gateway. Both the upload and download bandwidth rate limit can be set, or either one of the two. If a new limit is not set, the existing rate limit remains.", - "id": "to-update-the-bandwidth-rate-limits-of-a-gateway-1472151016202", - "title": "To update the bandwidth rate limits of a gateway" + "description":"Updates the bandwidth rate limits of a gateway. Both the upload and download bandwidth rate limit can be set, or either one of the two. If a new limit is not set, the existing rate limit remains.", + "id":"to-update-the-bandwidth-rate-limits-of-a-gateway-1472151016202", + "title":"To update the bandwidth rate limits of a gateway" } ], - "UpdateChapCredentials": [ + "UpdateChapCredentials":[ { - "input": { - "InitiatorName": "iqn.1991-05.com.microsoft:computername.domain.example.com", - "SecretToAuthenticateInitiator": "111111111111", - "SecretToAuthenticateTarget": "222222222222", - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "input":{ + "InitiatorName":"iqn.1991-05.com.microsoft:computername.domain.example.com", + "SecretToAuthenticateInitiator":"111111111111", + "SecretToAuthenticateTarget":"222222222222", + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" }, - "output": { - "InitiatorName": "iqn.1991-05.com.microsoft:computername.domain.example.com", - "TargetARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" + "output":{ + "InitiatorName":"iqn.1991-05.com.microsoft:computername.domain.example.com", + "TargetARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates the Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target.", - "id": "to-update-chap-credentials-for-an-iscsi-target-1472151325795", - "title": "To update CHAP credentials for an iSCSI target" + "description":"Updates the Challenge-Handshake Authentication Protocol (CHAP) credentials for a specified iSCSI target.", + "id":"to-update-chap-credentials-for-an-iscsi-target-1472151325795", + "title":"To update CHAP credentials for an iSCSI target" } ], - "UpdateGatewayInformation": [ + "UpdateGatewayInformation":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "GatewayName": "MyGateway2", - "GatewayTimezone": "GMT-12:00" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "GatewayName":"MyGateway2", + "GatewayTimezone":"GMT-12:00" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "GatewayName": "" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "GatewayName":"" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates a gateway's metadata, which includes the gateway's name and time zone.", - "id": "to-update-a-gateways-metadata-1472151688693", - "title": "To update a gateway's metadata" + "description":"Updates a gateway's metadata, which includes the gateway's name and time zone.", + "id":"to-update-a-gateways-metadata-1472151688693", + "title":"To update a gateway's metadata" } ], - "UpdateGatewaySoftwareNow": [ + "UpdateGatewaySoftwareNow":[ { - "input": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "input":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates the gateway virtual machine (VM) software. The request immediately triggers the software update.", - "id": "to-update-a-gateways-vm-software-1472152020929", - "title": "To update a gateway's VM software" + "description":"Updates the gateway virtual machine (VM) software. The request immediately triggers the software update.", + "id":"to-update-a-gateways-vm-software-1472152020929", + "title":"To update a gateway's VM software" } ], - "UpdateMaintenanceStartTime": [ + "UpdateMaintenanceStartTime":[ { - "input": { - "DayOfWeek": 2, - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", - "HourOfDay": 0, - "MinuteOfHour": 30 + "input":{ + "DayOfWeek":2, + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B", + "HourOfDay":0, + "MinuteOfHour":30 }, - "output": { - "GatewayARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" + "output":{ + "GatewayARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates a gateway's weekly maintenance start time information, including day and time of the week. The maintenance time is in your gateway's time zone.", - "id": "to-update-a-gateways-maintenance-start-time-1472152552031", - "title": "To update a gateway's maintenance start time" + "description":"Updates a gateway's weekly maintenance start time information, including day and time of the week. The maintenance time is in your gateway's time zone.", + "id":"to-update-a-gateways-maintenance-start-time-1472152552031", + "title":"To update a gateway's maintenance start time" } ], - "UpdateSnapshotSchedule": [ + "UpdateSnapshotSchedule":[ { - "input": { - "Description": "Hourly snapshot", - "RecurrenceInHours": 1, - "StartAt": 0, - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "input":{ + "Description":"Hourly snapshot", + "RecurrenceInHours":1, + "StartAt":0, + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "output": { - "VolumeARN": "arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" + "output":{ + "VolumeARN":"arn:aws:storagegateway:us-east-1:111122223333:gateway/sgw-12A3456B/volume/vol-1122AABB" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates a snapshot schedule configured for a gateway volume.", - "id": "to-update-a-volume-snapshot-schedule-1472152757068", - "title": "To update a volume snapshot schedule" + "description":"Updates a snapshot schedule configured for a gateway volume.", + "id":"to-update-a-volume-snapshot-schedule-1472152757068", + "title":"To update a volume snapshot schedule" } ], - "UpdateVTLDeviceType": [ + "UpdateVTLDeviceType":[ { - "input": { - "DeviceType": "Medium Changer", - "VTLDeviceARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_MEDIACHANGER_00001" + "input":{ + "DeviceType":"Medium Changer", + "VTLDeviceARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_MEDIACHANGER_00001" }, - "output": { - "VTLDeviceARN": "arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_MEDIACHANGER_00001" + "output":{ + "VTLDeviceARN":"arn:aws:storagegateway:us-east-1:999999999999:gateway/sgw-12A3456B/device/AMZN_SGW-1FAD4876_MEDIACHANGER_00001" }, - "comments": { - "input": { + "comments":{ + "input":{ }, - "output": { + "output":{ } }, - "description": "Updates the type of medium changer in a gateway-VTL after a gateway-VTL is activated.", - "id": "to-update-a-vtl-device-type-1472153012967", - "title": "To update a VTL device type" + "description":"Updates the type of medium changer in a gateway-VTL after a gateway-VTL is activated.", + "id":"to-update-a-vtl-device-type-1472153012967", + "title":"To update a VTL device type" } ] } diff --git a/vendor/github.com/aws/aws-sdk-go/models/endpoints/endpoints.json b/vendor/github.com/aws/aws-sdk-go/models/endpoints/endpoints.json index 350321704..a69bec48b 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/endpoints/endpoints.json +++ b/vendor/github.com/aws/aws-sdk-go/models/endpoints/endpoints.json @@ -97,6 +97,15 @@ "us-west-2" : { } } }, + "api.mediatailor" : { + "endpoints" : { + "ap-northeast-1" : { }, + "ap-southeast-1" : { }, + "ap-southeast-2" : { }, + "eu-west-1" : { }, + "us-east-1" : { } + } + }, "api.pricing" : { "defaults" : { "credentialScope" : { @@ -221,6 +230,7 @@ "endpoints" : { "ap-northeast-1" : { }, "ap-northeast-2" : { }, + "ap-south-1" : { }, "ap-southeast-1" : { }, "ap-southeast-2" : { }, "ca-central-1" : { }, @@ -338,6 +348,7 @@ "ca-central-1" : { }, "eu-central-1" : { }, "eu-west-1" : { }, + "eu-west-2" : { }, "us-east-1" : { }, "us-east-2" : { }, "us-west-1" : { }, @@ -812,6 +823,7 @@ }, "elasticfilesystem" : { "endpoints" : { + "ap-northeast-2" : { }, "ap-southeast-2" : { }, "eu-central-1" : { }, "eu-west-1" : { }, @@ -940,8 +952,10 @@ "firehose" : { "endpoints" : { "ap-northeast-1" : { }, + "ap-northeast-2" : { }, "ap-southeast-1" : { }, "ap-southeast-2" : { }, + "ca-central-1" : { }, "eu-central-1" : { }, "eu-west-1" : { }, "us-east-1" : { }, @@ -955,6 +969,7 @@ "protocols" : [ "https" ] }, "endpoints" : { + "eu-west-1" : { }, "us-east-1" : { }, "us-west-2" : { } } @@ -1001,11 +1016,13 @@ "glue" : { "endpoints" : { "ap-northeast-1" : { }, + "ap-northeast-2" : { }, "ap-south-1" : { }, "ap-southeast-1" : { }, "ap-southeast-2" : { }, "eu-central-1" : { }, "eu-west-1" : { }, + "eu-west-2" : { }, "us-east-1" : { }, "us-east-2" : { }, "us-west-2" : { } @@ -1253,8 +1270,10 @@ "medialive" : { "endpoints" : { "ap-northeast-1" : { }, + "ap-northeast-2" : { }, "ap-southeast-1" : { }, "ap-southeast-2" : { }, + "eu-central-1" : { }, "eu-west-1" : { }, "us-east-1" : { }, "us-west-2" : { } @@ -1361,6 +1380,34 @@ }, "isRegionalized" : false }, + "neptune" : { + "endpoints" : { + "eu-west-1" : { + "credentialScope" : { + "region" : "eu-west-1" + }, + "hostname" : "rds.eu-west-1.amazonaws.com" + }, + "us-east-1" : { + "credentialScope" : { + "region" : "us-east-1" + }, + "hostname" : "rds.us-east-1.amazonaws.com" + }, + "us-east-2" : { + "credentialScope" : { + "region" : "us-east-2" + }, + "hostname" : "rds.us-east-2.amazonaws.com" + }, + "us-west-2" : { + "credentialScope" : { + "region" : "us-west-2" + }, + "hostname" : "rds.us-west-2.amazonaws.com" + } + } + }, "opsworks" : { "endpoints" : { "ap-northeast-1" : { }, @@ -1533,6 +1580,7 @@ }, "runtime.sagemaker" : { "endpoints" : { + "ap-northeast-1" : { }, "eu-west-1" : { }, "us-east-1" : { }, "us-east-2" : { }, @@ -1597,6 +1645,7 @@ }, "sagemaker" : { "endpoints" : { + "ap-northeast-1" : { }, "eu-west-1" : { }, "us-east-1" : { }, "us-east-2" : { }, @@ -1712,6 +1761,7 @@ "eu-west-1" : { }, "us-east-1" : { }, "us-east-2" : { }, + "us-west-1" : { }, "us-west-2" : { } } }, @@ -1800,6 +1850,10 @@ "eu-west-1" : { }, "eu-west-2" : { }, "eu-west-3" : { }, + "fips-us-east-1" : { }, + "fips-us-east-2" : { }, + "fips-us-west-1" : { }, + "fips-us-west-2" : { }, "sa-east-1" : { }, "us-east-1" : { "sslCommonName" : "queue.{dnsSuffix}" @@ -2195,12 +2249,14 @@ }, "ecr" : { "endpoints" : { - "cn-north-1" : { } + "cn-north-1" : { }, + "cn-northwest-1" : { } } }, "ecs" : { "endpoints" : { - "cn-north-1" : { } + "cn-north-1" : { }, + "cn-northwest-1" : { } } }, "elasticache" : { @@ -2552,6 +2608,11 @@ "isRegionalized" : false, "partitionEndpoint" : "aws-us-gov-global" }, + "inspector" : { + "endpoints" : { + "us-gov-west-1" : { } + } + }, "kinesis" : { "endpoints" : { "us-gov-west-1" : { } diff --git a/vendor/github.com/aws/aws-sdk-go/models/protocol_tests/generate.go b/vendor/github.com/aws/aws-sdk-go/models/protocol_tests/generate.go index 05a399d19..b173376bd 100644 --- a/vendor/github.com/aws/aws-sdk-go/models/protocol_tests/generate.go +++ b/vendor/github.com/aws/aws-sdk-go/models/protocol_tests/generate.go @@ -394,6 +394,7 @@ func generateTestSuite(filename string) string { suite.API.NoConstServiceNames = true // don't generate service names suite.API.Setup() suite.API.Metadata.EndpointPrefix = suite.API.PackageName() + suite.API.Metadata.EndpointsID = suite.API.Metadata.EndpointPrefix // Sort in order for deterministic test generation names := make([]string, 0, len(suite.API.Shapes)) @@ -481,7 +482,7 @@ func GenerateAssertions(out interface{}, shape *api.Shape, prefix string) string case "timestamp": return fmtAssertEqual( fmt.Sprintf("time.Unix(%#v, 0).UTC().String()", out), - fmt.Sprintf("%s.String()", prefix), + fmt.Sprintf("%s.UTC().String()", prefix), ) case "blob": return fmtAssertEqual( diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/api.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/api.go index 10573f075..81086edc6 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/api.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/api.go @@ -10,6 +10,7 @@ import ( "io/ioutil" "path" "path/filepath" + "regexp" "sort" "strings" "text/template" @@ -54,6 +55,8 @@ type API struct { path string BaseCrosslinkURL string + + HasEventStream bool `json:"-"` } // A Metadata is the metadata about an API's definition. @@ -69,6 +72,7 @@ type Metadata struct { Protocol string UID string EndpointsID string + ServiceID string NoResolveEndpoint bool } @@ -416,30 +420,62 @@ var tplServiceDoc = template.Must(template.New("service docs").Funcs(template.Fu // https://docs.aws.amazon.com/sdk-for-go/api/service/{{ .PackageName }}/#New `)) +var serviceIDRegex = regexp.MustCompile("[^a-zA-Z0-9 ]+") +var prefixDigitRegex = regexp.MustCompile("^[0-9]+") + +// ServiceID will return a unique identifier specific to a service. +func ServiceID(a *API) string { + if len(a.Metadata.ServiceID) > 0 { + return a.Metadata.ServiceID + } + + name := a.Metadata.ServiceAbbreviation + if len(name) == 0 { + name = a.Metadata.ServiceFullName + } + + name = strings.Replace(name, "Amazon", "", -1) + name = strings.Replace(name, "AWS", "", -1) + name = serviceIDRegex.ReplaceAllString(name, "") + name = prefixDigitRegex.ReplaceAllString(name, "") + name = strings.TrimSpace(name) + return name +} + // A tplService defines the template for the service generated code. var tplService = template.Must(template.New("service").Funcs(template.FuncMap{ + "ServiceNameConstValue": ServiceName, "ServiceNameValue": func(a *API) string { - if a.NoConstServiceNames { - return fmt.Sprintf("%q", a.Metadata.EndpointPrefix) + if !a.NoConstServiceNames { + return "ServiceName" } - return "ServiceName" + return fmt.Sprintf("%q", ServiceName(a)) }, "EndpointsIDConstValue": func(a *API) string { if a.NoConstServiceNames { - return fmt.Sprintf("%q", a.Metadata.EndpointPrefix) + return fmt.Sprintf("%q", a.Metadata.EndpointsID) } - if a.Metadata.EndpointPrefix == a.Metadata.EndpointsID { + if a.Metadata.EndpointsID == ServiceName(a) { return "ServiceName" } + return fmt.Sprintf("%q", a.Metadata.EndpointsID) }, "EndpointsIDValue": func(a *API) string { if a.NoConstServiceNames { - return fmt.Sprintf("%q", a.Metadata.EndpointPrefix) + return fmt.Sprintf("%q", a.Metadata.EndpointsID) } return "EndpointsID" }, + "ServiceIDVar": func(a *API) string { + if a.NoConstServiceNames { + return fmt.Sprintf("%q", ServiceID(a)) + } + + return "ServiceID" + }, + "ServiceID": ServiceID, }).Parse(` // {{ .StructName }} provides the API operation methods for making requests to // {{ .Metadata.ServiceFullName }}. See this package's package overview docs @@ -462,8 +498,9 @@ var initRequest func(*request.Request) {{ if not .NoConstServiceNames -}} // Service information constants const ( - ServiceName = "{{ .Metadata.EndpointPrefix }}" // Service endpoint prefix API calls made to. - EndpointsID = {{ EndpointsIDConstValue . }} // Service ID for Regions and Endpoints metadata. + ServiceName = "{{ ServiceNameConstValue . }}" // Name of service. + EndpointsID = {{ EndpointsIDConstValue . }} // ID to lookup a service endpoint with. + ServiceID = "{{ ServiceID . }}" // ServiceID is a unique identifer of a specific service. ) {{- end }} @@ -504,6 +541,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: {{ ServiceNameValue . }}, + ServiceID : {{ ServiceIDVar . }}, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -528,6 +566,9 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio svc.Handlers.Unmarshal.PushBackNamed({{ .ProtocolPackage }}.UnmarshalHandler) svc.Handlers.UnmarshalMeta.PushBackNamed({{ .ProtocolPackage }}.UnmarshalMetaHandler) svc.Handlers.UnmarshalError.PushBackNamed({{ .ProtocolPackage }}.UnmarshalErrorHandler) + {{ if .HasEventStream }} + svc.Handlers.UnmarshalStream.PushBackNamed({{ .ProtocolPackage }}.UnmarshalHandler) + {{ end }} {{ if .UseInitMethods }}// Run custom client initialization if present if initClient != nil { diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream.go index e6437e69f..005ef9698 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream.go @@ -2,27 +2,536 @@ package api -func (a *API) suppressEventStreams() { +import ( + "bytes" + "fmt" + "io" + "text/template" +) + +// EventStreamAPI provides details about the event stream async API and +// associated EventStream shapes. +type EventStreamAPI struct { + Name string + Operation *Operation + Shape *Shape + Inbound *EventStream + Outbound *EventStream +} + +// EventStream represents a single eventstream group (input/output) and the +// modeled events that are known for the stream. +type EventStream struct { + Name string + Shape *Shape + Events []*Event +} + +// Event is a single EventStream event that can be sent or received in an +// EventStream. +type Event struct { + Name string + Shape *Shape + For *EventStream +} + +// ShapeDoc returns the docstring for the EventStream API. +func (esAPI *EventStreamAPI) ShapeDoc() string { + tmpl := template.Must(template.New("eventStreamShapeDoc").Parse(` +{{- $.Name }} provides handling of EventStreams for +the {{ $.Operation.ExportedName }} API. +{{- if $.Inbound }} + +Use this type to receive {{ $.Inbound.Name }} events. The events +can be read from the Events channel member. + +The events that can be received are: +{{ range $_, $event := $.Inbound.Events }} + * {{ $event.Shape.ShapeName }} +{{- end }} + +{{- end }} + +{{- if $.Outbound }} + +Use this type to send {{ $.Outbound.Name }} events. The events +can be sent with the Send method. + +The events that can be sent are: +{{ range $_, $event := $.Outbound.Events -}} + * {{ $event.Shape.ShapeName }} +{{- end }} + +{{- end }}`)) + + var w bytes.Buffer + if err := tmpl.Execute(&w, esAPI); err != nil { + panic(fmt.Sprintf("failed to generate eventstream shape template for %v, %v", esAPI.Name, err)) + } + + return commentify(w.String()) +} + +func eventStreamAPIShapeRefDoc(refName string) string { + return commentify(fmt.Sprintf("Use %s to use the API's stream.", refName)) +} + +func (a *API) setupEventStreams() { const eventStreamMemberName = "EventStream" - for name, op := range a.Operations { - outbound := hasEventStream(op.InputRef.Shape) - inbound := hasEventStream(op.OutputRef.Shape) + for _, op := range a.Operations { + outbound := setupEventStream(op.InputRef.Shape) + inbound := setupEventStream(op.OutputRef.Shape) - if !(outbound || inbound) { + if outbound == nil && inbound == nil { continue } - a.removeOperation(name) + if outbound != nil { + panic(fmt.Sprintf("Outbound stream support not implemented, %s, %s", + outbound.Name, outbound.Shape.ShapeName)) + } + + switch a.Metadata.Protocol { + case `rest-json`, `rest-xml`, `json`: + default: + panic(fmt.Sprintf("EventStream not supported for protocol %v", + a.Metadata.Protocol)) + } + + eventStreamAPI := &EventStreamAPI{ + Name: op.ExportedName + eventStreamMemberName, + Operation: op, + Outbound: outbound, + Inbound: inbound, + } + + streamShape := &Shape{ + API: a, + ShapeName: eventStreamAPI.Name, + Documentation: eventStreamAPI.ShapeDoc(), + Type: "structure", + EventStreamAPI: eventStreamAPI, + } + streamShapeRef := &ShapeRef{ + API: a, + ShapeName: streamShape.ShapeName, + Shape: streamShape, + Documentation: eventStreamAPIShapeRefDoc(eventStreamMemberName), + } + streamShape.refs = []*ShapeRef{streamShapeRef} + eventStreamAPI.Shape = streamShape + + if _, ok := op.OutputRef.Shape.MemberRefs[eventStreamMemberName]; ok { + panic(fmt.Sprintf("shape ref already exists, %s.%s", + op.OutputRef.Shape.ShapeName, eventStreamMemberName)) + } + op.OutputRef.Shape.MemberRefs[eventStreamMemberName] = streamShapeRef + op.OutputRef.Shape.EventStreamsMemberName = eventStreamMemberName + if _, ok := a.Shapes[streamShape.ShapeName]; ok { + panic("shape already exists, " + streamShape.ShapeName) + } + a.Shapes[streamShape.ShapeName] = streamShape + + a.HasEventStream = true } } -func hasEventStream(topShape *Shape) bool { - for _, ref := range topShape.MemberRefs { - if ref.Shape.IsEventStream { - return true +func setupEventStream(topShape *Shape) *EventStream { + var eventStream *EventStream + for refName, ref := range topShape.MemberRefs { + if !ref.Shape.IsEventStream { + continue + } + if eventStream != nil { + panic(fmt.Sprintf("multiple shape ref eventstreams, %s, prev: %s", + refName, eventStream.Name)) + } + + eventStream = &EventStream{ + Name: ref.Shape.ShapeName, + Shape: ref.Shape, + } + for _, eventRefName := range ref.Shape.MemberNames() { + eventRef := ref.Shape.MemberRefs[eventRefName] + if !eventRef.Shape.IsEvent { + panic(fmt.Sprintf("unexpected non-event member reference %s.%s", + ref.Shape.ShapeName, eventRefName)) + } + + updateEventPayloadRef(eventRef.Shape) + + eventRef.Shape.EventFor = append(eventRef.Shape.EventFor, eventStream) + eventStream.Events = append(eventStream.Events, &Event{ + Name: eventRefName, + Shape: eventRef.Shape, + For: eventStream, + }) + } + + // Remove the eventstream references as they will be added elsewhere. + ref.Shape.removeRef(ref) + delete(topShape.MemberRefs, refName) + delete(topShape.API.Shapes, ref.Shape.ShapeName) + } + + return eventStream +} + +func updateEventPayloadRef(parent *Shape) { + refName := parent.PayloadRefName() + if len(refName) == 0 { + return + } + + payloadRef := parent.MemberRefs[refName] + + if payloadRef.Shape.Type == "blob" { + return + } + + if len(payloadRef.LocationName) != 0 { + return + } + + payloadRef.LocationName = refName +} + +func renderEventStreamAPIShape(w io.Writer, s *Shape) error { + // Imports needed by the EventStream APIs. + s.API.imports["bytes"] = true + s.API.imports["io"] = true + s.API.imports["sync"] = true + s.API.imports["sync/atomic"] = true + s.API.imports["github.com/aws/aws-sdk-go/private/protocol/eventstream"] = true + s.API.imports["github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi"] = true + + return eventStreamAPIShapeTmpl.Execute(w, s) +} + +// EventStreamReaderInterfaceName returns the interface name for the +// EventStream's reader interface. +func EventStreamReaderInterfaceName(s *Shape) string { + return s.ShapeName + "Reader" +} + +// Template for an EventStream API Shape that will provide read/writing events +// across the EventStream. This is a special shape that's only public members +// are the Events channel and a Close and Err method. +// +// Executed in the context of a Shape. +var eventStreamAPIShapeTmpl = func() *template.Template { + t := template.Must( + template.New("eventStreamAPIShapeTmpl"). + Funcs(template.FuncMap{}). + Parse(eventStreamAPITmplDef), + ) + + template.Must( + t.AddParseTree( + "eventStreamAPIReaderTmpl", eventStreamAPIReaderTmpl.Tree), + ) + + return t +}() + +const eventStreamAPITmplDef = ` +{{ $.Documentation }} +type {{ $.ShapeName }} struct { + {{- if $.EventStreamAPI.Inbound }} + // Reader is the EventStream reader for the {{ $.EventStreamAPI.Inbound.Name }} + // events. This value is automatically set by the SDK when the API call is made + // Use this member when unit testing your code with the SDK to mock out the + // EventStream Reader. + // + // Must not be nil. + Reader {{ $.ShapeName }}Reader + + {{ end -}} + + {{- if $.EventStreamAPI.Outbound }} + // Writer is the EventStream reader for the {{ $.EventStreamAPI.Inbound.Name }} + // events. This value is automatically set by the SDK when the API call is made + // Use this member when unit testing your code with the SDK to mock out the + // EventStream Writer. + // + // Must not be nil. + Writer *{{ $.ShapeName }}Writer + + {{ end -}} + + // StreamCloser is the io.Closer for the EventStream connection. For HTTP + // EventStream this is the response Body. The stream will be closed when + // the Close method of the EventStream is called. + StreamCloser io.Closer +} + +// Close closes the EventStream. This will also cause the Events channel to be +// closed. You can use the closing of the Events channel to terminate your +// application's read from the API's EventStream. +{{- if $.EventStreamAPI.Inbound }} +// +// Will close the underlying EventStream reader. For EventStream over HTTP +// connection this will also close the HTTP connection. +{{ end -}} +// +// Close must be called when done using the EventStream API. Not calling Close +// may result in resource leaks. +func (es *{{ $.ShapeName }}) Close() (err error) { + {{- if $.EventStreamAPI.Inbound }} + es.Reader.Close() + {{ end -}} + {{- if $.EventStreamAPI.Outbound }} + es.Writer.Close() + {{ end -}} + + return es.Err() +} + +// Err returns any error that occurred while reading EventStream Events from +// the service API's response. Returns nil if there were no errors. +func (es *{{ $.ShapeName }}) Err() error { + {{- if $.EventStreamAPI.Outbound }} + if err := es.Writer.Err(); err != nil { + return err + } + {{ end -}} + + {{- if $.EventStreamAPI.Inbound }} + if err := es.Reader.Err(); err != nil { + return err + } + {{ end -}} + + es.StreamCloser.Close() + + return nil +} + +{{ if $.EventStreamAPI.Inbound }} + // Events returns a channel to read EventStream Events from the + // {{ $.EventStreamAPI.Operation.ExportedName }} API. + // + // These events are: + // {{ range $_, $event := $.EventStreamAPI.Inbound.Events }} + // * {{ $event.Shape.ShapeName }} + {{- end }} + func (es *{{ $.ShapeName }}) Events() <-chan {{ $.EventStreamAPI.Inbound.Name }}Event { + return es.Reader.Events() + } + + {{ template "eventStreamAPIReaderTmpl" $ }} +{{ end }} + +{{ if $.EventStreamAPI.Outbound }} + // TODO writer helper method. +{{ end }} + +` + +var eventStreamAPIReaderTmpl = template.Must(template.New("eventStreamAPIReaderTmpl"). + Funcs(template.FuncMap{}). + Parse(` +// {{ $.EventStreamAPI.Inbound.Name }}Event groups together all EventStream +// events read from the {{ $.EventStreamAPI.Operation.ExportedName }} API. +// +// These events are: +// {{ range $_, $event := $.EventStreamAPI.Inbound.Events }} +// * {{ $event.Shape.ShapeName }} +{{- end }} +type {{ $.EventStreamAPI.Inbound.Name }}Event interface { + event{{ $.EventStreamAPI.Name }}() +} + +// {{ $.ShapeName }}Reader provides the interface for reading EventStream +// Events from the {{ $.EventStreamAPI.Operation.ExportedName }} API. The +// default implementation for this interface will be {{ $.ShapeName }}. +// +// The reader's Close method must allow multiple concurrent calls. +// +// These events are: +// {{ range $_, $event := $.EventStreamAPI.Inbound.Events }} +// * {{ $event.Shape.ShapeName }} +{{- end }} +type {{ $.ShapeName }}Reader interface { + // Returns a channel of events as they are read from the event stream. + Events() <-chan {{ $.EventStreamAPI.Inbound.Name }}Event + + // Close will close the underlying event stream reader. For event stream over + // HTTP this will also close the HTTP connection. + Close() error + + // Returns any error that has occured while reading from the event stream. + Err() error +} + +type read{{ $.ShapeName }} struct { + eventReader *eventstreamapi.EventReader + stream chan {{ $.EventStreamAPI.Inbound.Name }}Event + errVal atomic.Value + + done chan struct{} + closeOnce sync.Once +} + +func newRead{{ $.ShapeName }}( + reader io.ReadCloser, + unmarshalers request.HandlerList, + logger aws.Logger, + logLevel aws.LogLevelType, +) *read{{ $.ShapeName }} { + r := &read{{ $.ShapeName }}{ + stream: make(chan {{ $.EventStreamAPI.Inbound.Name }}Event), + done: make(chan struct{}), + } + + r.eventReader = eventstreamapi.NewEventReader( + reader, + protocol.HandlerPayloadUnmarshal{ + Unmarshalers: unmarshalers, + }, + r.unmarshalerForEventType, + ) + r.eventReader.UseLogger(logger, logLevel) + + return r +} + +// Close will close the underlying event stream reader. For EventStream over +// HTTP this will also close the HTTP connection. +func (r *read{{ $.ShapeName }}) Close() error { + r.closeOnce.Do(r.safeClose) + + return r.Err() +} + +func (r *read{{ $.ShapeName }}) safeClose() { + close(r.done) + err := r.eventReader.Close() + if err != nil { + r.errVal.Store(err) + } +} + +func (r *read{{ $.ShapeName }}) Err() error { + if v := r.errVal.Load(); v != nil { + return v.(error) + } + + return nil +} + +func (r *read{{ $.ShapeName }}) Events() <-chan {{ $.EventStreamAPI.Inbound.Name }}Event { + return r.stream +} + +func (r *read{{ $.ShapeName }}) readEventStream() { + defer close(r.stream) + + for { + event, err := r.eventReader.ReadEvent() + if err != nil { + if err == io.EOF { + return + } + select { + case <-r.done: + // If closed already ignore the error + return + default: + } + r.errVal.Store(err) + return + } + + select { + case r.stream <- event.({{ $.EventStreamAPI.Inbound.Name }}Event): + case <-r.done: + return } } - - return false } + +func (r *read{{ $.ShapeName }}) unmarshalerForEventType( + eventType string, +) (eventstreamapi.Unmarshaler, error) { + switch eventType { + {{- range $_, $event := $.EventStreamAPI.Inbound.Events }} + case {{ printf "%q" $event.Name }}: + return &{{ $event.Shape.ShapeName }}{}, nil + {{ end -}} + default: + return nil, fmt.Errorf( + "unknown event type name, %s, for {{ $.ShapeName }}", eventType) + } +} +`)) + +// Template for the EventStream API Output shape that contains the EventStream +// member. +// +// Executed in the context of a Shape. +var eventStreamAPILoopMethodTmpl = template.Must( + template.New("eventStreamAPILoopMethodTmpl").Parse(` +func (s *{{ $.ShapeName }}) runEventStreamLoop(r *request.Request) { + if r.Error != nil { + return + } + + {{- $esMemberRef := index $.MemberRefs $.EventStreamsMemberName }} + {{- if $esMemberRef.Shape.EventStreamAPI.Inbound }} + reader := newRead{{ $esMemberRef.ShapeName }}( + r.HTTPResponse.Body, + r.Handlers.UnmarshalStream, + r.Config.Logger, + r.Config.LogLevel.Value(), + ) + go reader.readEventStream() + + eventStream := &{{ $esMemberRef.ShapeName }} { + StreamCloser: r.HTTPResponse.Body, + Reader: reader, + } + {{ end -}} + + s.{{ $.EventStreamsMemberName }} = eventStream +} +`)) + +// Template for an EventStream Event shape. This is a normal API shape that is +// decorated as an EventStream Event. +// +// Executed in the context of a Shape. +var eventStreamEventShapeTmpl = template.Must(template.New("eventStreamEventShapeTmpl").Parse(` +{{ range $_, $eventstream := $.EventFor }} + // The {{ $.ShapeName }} is and event in the {{ $eventstream.Name }} group of events. + func (s *{{ $.ShapeName }}) event{{ $eventstream.Name }}() {} +{{ end }} + +// UnmarshalEvent unmarshals the EventStream Message into the {{ $.ShapeName }} value. +// This method is only used internally within the SDK's EventStream handling. +func (s *{{ $.ShapeName }}) UnmarshalEvent( + payloadUnmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + {{- range $fieldIdx, $fieldName := $.MemberNames }} + {{- $fieldRef := index $.MemberRefs $fieldName -}} + {{ if $fieldRef.IsEventHeader }} + // TODO handle event header, {{ $fieldName }} + {{- else if (and ($fieldRef.IsEventPayload) (eq $fieldRef.Shape.Type "blob")) }} + s.{{ $fieldName }} = make([]byte, len(msg.Payload)) + copy(s.{{ $fieldName }}, msg.Payload) + {{- else }} + if err := payloadUnmarshaler.UnmarshalPayload( + bytes.NewReader(msg.Payload), s, + ); err != nil { + return fmt.Errorf("failed to unmarshal payload, %v", err) + } + {{- end }} + {{- end }} + return nil +} +`)) + +var eventStreamTestTmpl = template.Must(template.New("eventStreamTestTmpl").Parse(` +`)) diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream_test.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream_test.go deleted file mode 100644 index 41804a181..000000000 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/eventstream_test.go +++ /dev/null @@ -1,79 +0,0 @@ -// +build go1.6,codegen - -package api - -import ( - "testing" -) - -func TestSuppressEventStream(t *testing.T) { - cases := []struct { - API *API - Ops []string - Shapes []string - }{ - { - API: &API{ - Operations: map[string]*Operation{ - "abc": { - InputRef: ShapeRef{ - ShapeName: "abcRequest", - }, - OutputRef: ShapeRef{ - ShapeName: "abcResponse", - }, - }, - "eventStreamOp": { - InputRef: ShapeRef{ - ShapeName: "eventStreamOpRequest", - }, - OutputRef: ShapeRef{ - ShapeName: "eventStreamOpResponse", - }, - }, - }, - Shapes: map[string]*Shape{ - "abcRequest": {}, - "abcResponse": {}, - "eventStreamOpRequest": {}, - "eventStreamOpResponse": { - MemberRefs: map[string]*ShapeRef{ - "eventStreamShape": { - ShapeName: "eventStreamShape", - }, - }, - }, - "eventStreamShape": { - IsEventStream: true, - }, - }, - }, - Ops: []string{"Abc"}, - Shapes: []string{"AbcInput", "AbcOutput"}, - }, - } - - for _, c := range cases { - c.API.Setup() - if e, a := c.Ops, c.API.OperationNames(); !stringsEqual(e, a) { - t.Errorf("expect %v ops, got %v", e, a) - } - - if e, a := c.Shapes, c.API.ShapeNames(); !stringsEqual(e, a) { - t.Errorf("expect %v ops, got %v", e, a) - } - } -} - -func stringsEqual(a, b []string) bool { - if len(a) != len(b) { - return false - } - - for i := 0; i < len(a); i++ { - if a[i] != b[i] { - return false - } - } - return true -} diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/load.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/load.go index 8fb52d2ae..a920e350a 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/load.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/load.go @@ -47,7 +47,7 @@ func (a *API) Setup() { a.renameCollidingFields() a.updateTopLevelShapeReferences() a.createInputOutputShapes() - a.suppressEventStreams() + a.setupEventStreams() a.customizationPasses() if !a.NoRemoveUnusedShapes { diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/operation.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/operation.go index 59a716ed9..ccab396a7 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/operation.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/operation.go @@ -44,6 +44,7 @@ func (o *Operation) HasOutput() bool { return o.OutputRef.ShapeName != "" } +// GetSigner returns the signer that should be used for a API request. func (o *Operation) GetSigner() string { if o.AuthType == "v4-unsigned-body" { o.API.imports["github.com/aws/aws-sdk-go/aws/signer/v4"] = true @@ -66,7 +67,8 @@ func (o *Operation) GetSigner() string { // tplOperation defines a template for rendering an API Operation var tplOperation = template.Must(template.New("operation").Funcs(template.FuncMap{ - "GetCrosslinkURL": GetCrosslinkURL, + "GetCrosslinkURL": GetCrosslinkURL, + "EnableStopOnSameToken": enableStopOnSameToken, }).Parse(` const op{{ .ExportedName }} = "{{ .Name }}" @@ -120,10 +122,17 @@ func (c *{{ .API.StructName }}) {{ .ExportedName }}Request(` + } output = &{{ .OutputRef.GoTypeElem }}{} - req = c.newRequest(op, input, output){{ if eq .OutputRef.Shape.Placeholder true }} - req.Handlers.Unmarshal.Remove({{ .API.ProtocolPackage }}.UnmarshalHandler) - req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler){{ end }} + req = c.newRequest(op, input, output) + {{ if eq .OutputRef.Shape.Placeholder true -}} + req.Handlers.Unmarshal.Remove({{ .API.ProtocolPackage }}.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + {{ end -}} {{ if ne .AuthType "" }}{{ .GetSigner }}{{ end -}} + {{ if .OutputRef.Shape.EventStreamsMemberName -}} + req.Handlers.Send.Swap(client.LogHTTPResponseHandler.Name, client.LogHTTPResponseHeaderHandler) + req.Handlers.Unmarshal.Swap({{ .API.ProtocolPackage }}.UnmarshalHandler.Name, rest.UnmarshalHandler) + req.Handlers.Unmarshal.PushBack(output.runEventStreamLoop) + {{ end -}} return } @@ -214,6 +223,8 @@ func (c *{{ .API.StructName }}) {{ .ExportedName }}PagesWithContext(` + `fn func({{ .OutputRef.GoType }}, bool) bool, ` + `opts ...request.Option) error { p := request.Pagination { + {{ if EnableStopOnSameToken .API.PackageName -}}EndPageOnSameToken: true, + {{ end -}} NewRequest: func() (*request.Request, error) { var inCpy {{ .InputRef.GoType }} if input != nil { @@ -239,6 +250,13 @@ func (c *{{ .API.StructName }}) {{ .ExportedName }}PagesWithContext(` + // GoCode returns a string of rendered GoCode for this Operation func (o *Operation) GoCode() string { var buf bytes.Buffer + + if len(o.OutputRef.Shape.EventStreamsMemberName) != 0 { + // TODO need better was of updating protocol unmarshalers + o.API.imports["github.com/aws/aws-sdk-go/aws/client"] = true + o.API.imports["github.com/aws/aws-sdk-go/private/protocol/rest"] = true + } + err := tplOperation.Execute(&buf, o) if err != nil { panic(err) diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/pagination.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/pagination.go index 441a47529..f9ab64409 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/pagination.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/pagination.go @@ -89,3 +89,12 @@ func (p *paginationDefinitions) setup() { } } } + +func enableStopOnSameToken(service string) bool { + switch service { + case "cloudwatchlogs": + return true + default: + return false + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/passes.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/passes.go index 450da1ca8..0773f06c5 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/passes.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/passes.go @@ -59,6 +59,7 @@ func (a *API) resolveReferences() { for i := range o.ErrorRefs { resolver.resolveReference(&o.ErrorRefs[i]) o.ErrorRefs[i].Shape.IsError = true + o.ErrorRefs[i].Shape.ErrorInfo.Type = o.ErrorRefs[i].Shape.ShapeName } } } diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/service_name.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/service_name.go new file mode 100644 index 000000000..737a5c5b9 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/service_name.go @@ -0,0 +1,159 @@ +// +build codegen + +package api + +// ServiceName returns the SDK's naming of the service. Has +// backwards compatability built in for services that were +// incorrectly named with the service's endpoint prefix. +func ServiceName(a *API) string { + if oldName, ok := oldServiceNames[a.PackageName()]; ok { + return oldName + } + + return ServiceID(a) +} + +var oldServiceNames = map[string]string{ + "migrationhub": "mgh", + "acmpca": "acm-pca", + "acm": "acm", + "alexaforbusiness": "a4b", + "apigateway": "apigateway", + "applicationautoscaling": "autoscaling", + "appstream": "appstream2", + "appsync": "appsync", + "athena": "athena", + "autoscalingplans": "autoscaling", + "autoscaling": "autoscaling", + "batch": "batch", + "budgets": "budgets", + "costexplorer": "ce", + "cloud9": "cloud9", + "clouddirectory": "clouddirectory", + "cloudformation": "cloudformation", + "cloudfront": "cloudfront", + "cloudhsm": "cloudhsm", + "cloudhsmv2": "cloudhsmv2", + "cloudsearch": "cloudsearch", + "cloudsearchdomain": "cloudsearchdomain", + "cloudtrail": "cloudtrail", + "codebuild": "codebuild", + "codecommit": "codecommit", + "codedeploy": "codedeploy", + "codepipeline": "codepipeline", + "codestar": "codestar", + "cognitoidentity": "cognito-identity", + "cognitoidentityprovider": "cognito-idp", + "cognitosync": "cognito-sync", + "comprehend": "comprehend", + "configservice": "config", + "connect": "connect", + "costandusagereportservice": "cur", + "datapipeline": "datapipeline", + "dax": "dax", + "devicefarm": "devicefarm", + "directconnect": "directconnect", + "applicationdiscoveryservice": "discovery", + "databasemigrationservice": "dms", + "directoryservice": "ds", + "dynamodb": "dynamodb", + "ec2": "ec2", + "ecr": "ecr", + "ecs": "ecs", + "eks": "eks", + "elasticache": "elasticache", + "elasticbeanstalk": "elasticbeanstalk", + "efs": "elasticfilesystem", + "elb": "elasticloadbalancing", + "elbv2": "elasticloadbalancing", + "emr": "elasticmapreduce", + "elastictranscoder": "elastictranscoder", + "ses": "email", + "marketplaceentitlementservice": "entitlement.marketplace", + "elasticsearchservice": "es", + "cloudwatchevents": "events", + "firehose": "firehose", + "fms": "fms", + "gamelift": "gamelift", + "glacier": "glacier", + "glue": "glue", + "greengrass": "greengrass", + "guardduty": "guardduty", + "health": "health", + "iam": "iam", + "inspector": "inspector", + "iotdataplane": "data.iot", + "iotjobsdataplane": "data.jobs.iot", + "iot": "iot", + "iot1clickdevicesservice": "devices.iot1click", + "iot1clickprojects": "projects.iot1click", + "iotanalytics": "iotanalytics", + "kinesisvideoarchivedmedia": "kinesisvideo", + "kinesisvideomedia": "kinesisvideo", + "kinesis": "kinesis", + "kinesisanalytics": "kinesisanalytics", + "kinesisvideo": "kinesisvideo", + "kms": "kms", + "lambda": "lambda", + "lexmodelbuildingservice": "models.lex", + "lightsail": "lightsail", + "cloudwatchlogs": "logs", + "machinelearning": "machinelearning", + "marketplacecommerceanalytics": "marketplacecommerceanalytics", + "mediaconvert": "mediaconvert", + "medialive": "medialive", + "mediapackage": "mediapackage", + "mediastoredata": "data.mediastore", + "mediastore": "mediastore", + "mediatailor": "api.mediatailor", + "marketplacemetering": "metering.marketplace", + "mobile": "mobile", + "mobileanalytics": "mobileanalytics", + "cloudwatch": "monitoring", + "mq": "mq", + "mturk": "mturk-requester", + "neptune": "rds", + "opsworks": "opsworks", + "opsworkscm": "opsworks-cm", + "organizations": "organizations", + "pi": "pi", + "pinpoint": "pinpoint", + "polly": "polly", + "pricing": "api.pricing", + "rds": "rds", + "redshift": "redshift", + "rekognition": "rekognition", + "resourcegroups": "resource-groups", + "resourcegroupstaggingapi": "tagging", + "route53": "route53", + "route53domains": "route53domains", + "lexruntimeservice": "runtime.lex", + "sagemakerruntime": "runtime.sagemaker", + "s3": "s3", + "sagemaker": "sagemaker", + "simpledb": "sdb", + "secretsmanager": "secretsmanager", + "serverlessapplicationrepository": "serverlessrepo", + "servicecatalog": "servicecatalog", + "servicediscovery": "servicediscovery", + "shield": "shield", + "sms": "sms", + "snowball": "snowball", + "sns": "sns", + "sqs": "sqs", + "ssm": "ssm", + "sfn": "states", + "storagegateway": "storagegateway", + "dynamodbstreams": "streams.dynamodb", + "sts": "sts", + "support": "support", + "swf": "swf", + "transcribeservice": "transcribe", + "translate": "translate", + "wafregional": "waf-regional", + "waf": "waf", + "workdocs": "workdocs", + "workmail": "workmail", + "workspaces": "workspaces", + "xray": "xray", +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/model/api/shape.go b/vendor/github.com/aws/aws-sdk-go/private/model/api/shape.go index 86d3453fe..09cd21f85 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/model/api/shape.go +++ b/vendor/github.com/aws/aws-sdk-go/private/model/api/shape.go @@ -12,6 +12,19 @@ import ( "text/template" ) +// ErrorInfo represents the error block of a shape's structure +type ErrorInfo struct { + Type string + Code string + HTTPStatusCode int +} + +// A XMLInfo defines URL and prefix for Shapes when rendered as XML +type XMLInfo struct { + Prefix string + URI string +} + // A ShapeRef defines the usage of a shape within the API. type ShapeRef struct { API *API `json:"-"` @@ -35,18 +48,9 @@ type ShapeRef struct { OrigShapeName string `json:"-"` GenerateGetter bool -} -// ErrorInfo represents the error block of a shape's structure -type ErrorInfo struct { - Code string - HTTPStatusCode int -} - -// A XMLInfo defines URL and prefix for Shapes when rendered as XML -type XMLInfo struct { - Prefix string - URI string + IsEventPayload bool `json:"eventpayload"` + IsEventHeader bool `json:"eventheader"` } // A Shape defines the definition of a shape type @@ -73,7 +77,12 @@ type Shape struct { Min float64 // optional Minimum length (string, list) or value (number) Max float64 // optional Maximum length (string, list) or value (number) + EventStreamsMemberName string `json:"-"` + EventStreamAPI *EventStreamAPI `json:"-"` + EventFor []*EventStream `json:"-"` + IsEventStream bool `json:"eventstream"` + IsEvent bool `json:"event"` refs []*ShapeRef // References to this shape resolvePkg string // use this package in the goType() if present @@ -101,7 +110,7 @@ func (s *Shape) ErrorCodeName() string { // ErrorName will return the shape's name or error code if available based // on the API's protocol. This is the error code string returned by the service. func (s *Shape) ErrorName() string { - name := s.ShapeName + name := s.ErrorInfo.Type switch s.API.Metadata.Protocol { case "query", "ec2query", "rest-xml": if len(s.ErrorInfo.Code) > 0 { @@ -112,6 +121,23 @@ func (s *Shape) ErrorName() string { return name } +// PayloadRefName returns the payload member of the shape if there is one +// modeled. If no payload is modeled, empty string will be returned. +func (s *Shape) PayloadRefName() string { + if name := s.Payload; len(name) != 0 { + // Root shape + return name + } + + for name, ref := range s.MemberRefs { + if ref.IsEventPayload { + return name + } + } + + return "" +} + // GoTags returns the struct tags for a shape. func (s *Shape) GoTags(root, required bool) string { ref := &ShapeRef{ShapeName: s.ShapeName, API: s.API, Shape: s} @@ -149,6 +175,8 @@ func (s *Shape) GoTypeWithPkgName() string { return goType(s, true) } +// GoTypeWithPkgNameElem returns the shapes type as a string with the "*" +// removed if there was one preset. func (s *Shape) GoTypeWithPkgNameElem() string { t := goType(s, true) if strings.HasPrefix(t, "*") { @@ -157,7 +185,7 @@ func (s *Shape) GoTypeWithPkgNameElem() string { return t } -// GenAccessors returns if the shape's reference should have setters generated. +// UseIndirection returns if the shape's reference should use indirection or not. func (s *ShapeRef) UseIndirection() bool { switch s.Shape.Type { case "map", "list", "blob", "structure", "jsonvalue": @@ -401,8 +429,8 @@ func (ref *ShapeRef) GoTags(toplevel bool, isRequired bool) string { } if toplevel { - if ref.Shape.Payload != "" { - tags = append(tags, ShapeTag{"payload", ref.Shape.Payload}) + if name := ref.Shape.PayloadRefName(); len(name) > 0 { + tags = append(tags, ShapeTag{"payload", name}) } } @@ -514,9 +542,29 @@ func (s *Shape) NestedShape() *Shape { return nestedShape } -var structShapeTmpl = template.Must(template.New("StructShape").Funcs(template.FuncMap{ - "GetCrosslinkURL": GetCrosslinkURL, -}).Parse(` +var structShapeTmpl = func() *template.Template { + shapeTmpl := template.Must( + template.New("structShapeTmpl"). + Funcs(template.FuncMap{ + "GetCrosslinkURL": GetCrosslinkURL, + }). + Parse(structShapeTmplDef), + ) + + template.Must( + shapeTmpl.AddParseTree( + "eventStreamAPILoopMethodTmpl", eventStreamAPILoopMethodTmpl.Tree), + ) + + template.Must( + shapeTmpl.AddParseTree( + "eventStreamEventShapeTmpl", eventStreamEventShapeTmpl.Tree), + ) + + return shapeTmpl +}() + +const structShapeTmplDef = ` {{ .Docstring }} {{ $context := . -}} type {{ .ShapeName }} struct { @@ -557,38 +605,43 @@ type {{ .ShapeName }} struct { {{ end }} {{ if not .API.NoGenStructFieldAccessors }} + {{ $builderShapeName := print .ShapeName -}} + {{ range $_, $name := $context.MemberNames -}} + {{ $elem := index $context.MemberRefs $name -}} -{{ $builderShapeName := print .ShapeName -}} - -{{ range $_, $name := $context.MemberNames -}} - {{ $elem := index $context.MemberRefs $name -}} - -// Set{{ $name }} sets the {{ $name }} field's value. -func (s *{{ $builderShapeName }}) Set{{ $name }}(v {{ $context.GoStructValueType $name $elem }}) *{{ $builderShapeName }} { - {{ if $elem.UseIndirection -}} - s.{{ $name }} = &v - {{ else -}} - s.{{ $name }} = v - {{ end -}} - return s -} - -{{ if $elem.GenerateGetter -}} -func (s *{{ $builderShapeName }}) get{{ $name }}() (v {{ $context.GoStructValueType $name $elem }}) { - {{ if $elem.UseIndirection -}} - if s.{{ $name }} == nil { - return v + // Set{{ $name }} sets the {{ $name }} field's value. + func (s *{{ $builderShapeName }}) Set{{ $name }}(v {{ $context.GoStructValueType $name $elem }}) *{{ $builderShapeName }} { + {{ if $elem.UseIndirection -}} + s.{{ $name }} = &v + {{ else -}} + s.{{ $name }} = v + {{ end -}} + return s } - return *s.{{ $name }} - {{ else -}} - return s.{{ $name }} - {{ end -}} -} -{{- end }} + {{ if $elem.GenerateGetter -}} + func (s *{{ $builderShapeName }}) get{{ $name }}() (v {{ $context.GoStructValueType $name $elem }}) { + {{ if $elem.UseIndirection -}} + if s.{{ $name }} == nil { + return v + } + return *s.{{ $name }} + {{ else -}} + return s.{{ $name }} + {{ end -}} + } + {{- end }} + {{ end }} {{ end }} + +{{ if $.EventStreamsMemberName }} + {{ template "eventStreamAPILoopMethodTmpl" $ }} {{ end }} -`)) + +{{ if $.IsEvent }} + {{ template "eventStreamEventShapeTmpl" $ }} +{{ end }} +` var enumShapeTmpl = template.Must(template.New("EnumShape").Parse(` {{ .Docstring }} @@ -605,22 +658,38 @@ const ( // GoCode returns the rendered Go code for the Shape. func (s *Shape) GoCode() string { - b := &bytes.Buffer{} + w := &bytes.Buffer{} switch { + case s.EventStreamAPI != nil: + if err := renderEventStreamAPIShape(w, s); err != nil { + panic( + fmt.Sprintf( + "failed to generate eventstream API shape, %s, %v", + s.ShapeName, err), + ) + } case s.Type == "structure": - if err := structShapeTmpl.Execute(b, s); err != nil { - panic(fmt.Sprintf("Failed to generate struct shape %s, %v\n", s.ShapeName, err)) + if err := structShapeTmpl.Execute(w, s); err != nil { + panic( + fmt.Sprintf( + "Failed to generate struct shape %s, %v", + s.ShapeName, err), + ) } case s.IsEnum(): - if err := enumShapeTmpl.Execute(b, s); err != nil { - panic(fmt.Sprintf("Failed to generate enum shape %s, %v\n", s.ShapeName, err)) + if err := enumShapeTmpl.Execute(w, s); err != nil { + panic( + fmt.Sprintf( + "Failed to generate enum shape %s, %v", + s.ShapeName, err), + ) } default: panic(fmt.Sprintln("Cannot generate toplevel shape for", s.Type)) } - return b.String() + return w.String() } // IsEnum returns whether this shape is an enum list diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/build_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/build_test.go index 28f81b60b..8654f3698 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/build_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/build_test.go @@ -75,7 +75,8 @@ func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice1protocoltest", + ServiceName: "InputService1ProtocolTest", + ServiceID: "InputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -228,7 +229,8 @@ func newInputService2ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice2protocoltest", + ServiceName: "InputService2ProtocolTest", + ServiceID: "InputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -389,7 +391,8 @@ func newInputService3ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice3protocoltest", + ServiceName: "InputService3ProtocolTest", + ServiceID: "InputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -546,7 +549,8 @@ func newInputService4ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice4protocoltest", + ServiceName: "InputService4ProtocolTest", + ServiceID: "InputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -715,7 +719,8 @@ func newInputService5ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice5protocoltest", + ServiceName: "InputService5ProtocolTest", + ServiceID: "InputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -860,7 +865,8 @@ func newInputService6ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice6protocoltest", + ServiceName: "InputService6ProtocolTest", + ServiceID: "InputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1005,7 +1011,8 @@ func newInputService7ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice7protocoltest", + ServiceName: "InputService7ProtocolTest", + ServiceID: "InputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1151,7 +1158,8 @@ func newInputService8ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice8protocoltest", + ServiceName: "InputService8ProtocolTest", + ServiceID: "InputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1296,7 +1304,8 @@ func newInputService9ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice9protocoltest", + ServiceName: "InputService9ProtocolTest", + ServiceID: "InputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1515,7 +1524,8 @@ func newInputService10ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice10protocoltest", + ServiceName: "InputService10ProtocolTest", + ServiceID: "InputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/unmarshal_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/unmarshal_test.go index 0e2377c36..408d269c8 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/unmarshal_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/ec2query/unmarshal_test.go @@ -75,7 +75,8 @@ func newOutputService1ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice1protocoltest", + ServiceName: "OutputService1ProtocolTest", + ServiceID: "OutputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -274,7 +275,8 @@ func newOutputService2ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice2protocoltest", + ServiceName: "OutputService2ProtocolTest", + ServiceID: "OutputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -418,7 +420,8 @@ func newOutputService3ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice3protocoltest", + ServiceName: "OutputService3ProtocolTest", + ServiceID: "OutputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -561,7 +564,8 @@ func newOutputService4ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice4protocoltest", + ServiceName: "OutputService4ProtocolTest", + ServiceID: "OutputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -704,7 +708,8 @@ func newOutputService5ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice5protocoltest", + ServiceName: "OutputService5ProtocolTest", + ServiceID: "OutputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -847,7 +852,8 @@ func newOutputService6ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice6protocoltest", + ServiceName: "OutputService6ProtocolTest", + ServiceID: "OutputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1002,7 +1008,8 @@ func newOutputService7ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice7protocoltest", + ServiceName: "OutputService7ProtocolTest", + ServiceID: "OutputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1145,7 +1152,8 @@ func newOutputService8ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice8protocoltest", + ServiceName: "OutputService8ProtocolTest", + ServiceID: "OutputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1288,7 +1296,8 @@ func newOutputService9ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice9protocoltest", + ServiceName: "OutputService9ProtocolTest", + ServiceID: "OutputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1431,7 +1440,8 @@ func newOutputService10ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice10protocoltest", + ServiceName: "OutputService10ProtocolTest", + ServiceID: "OutputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/debug.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/debug.go new file mode 100644 index 000000000..ecc7bf82f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/debug.go @@ -0,0 +1,144 @@ +package eventstream + +import ( + "bytes" + "encoding/base64" + "encoding/json" + "fmt" + "strconv" +) + +type decodedMessage struct { + rawMessage + Headers decodedHeaders `json:"headers"` +} +type jsonMessage struct { + Length json.Number `json:"total_length"` + HeadersLen json.Number `json:"headers_length"` + PreludeCRC json.Number `json:"prelude_crc"` + Headers decodedHeaders `json:"headers"` + Payload []byte `json:"payload"` + CRC json.Number `json:"message_crc"` +} + +func (d *decodedMessage) UnmarshalJSON(b []byte) (err error) { + var jsonMsg jsonMessage + if err = json.Unmarshal(b, &jsonMsg); err != nil { + return err + } + + d.Length, err = numAsUint32(jsonMsg.Length) + if err != nil { + return err + } + d.HeadersLen, err = numAsUint32(jsonMsg.HeadersLen) + if err != nil { + return err + } + d.PreludeCRC, err = numAsUint32(jsonMsg.PreludeCRC) + if err != nil { + return err + } + d.Headers = jsonMsg.Headers + d.Payload = jsonMsg.Payload + d.CRC, err = numAsUint32(jsonMsg.CRC) + if err != nil { + return err + } + + return nil +} + +func (d *decodedMessage) MarshalJSON() ([]byte, error) { + jsonMsg := jsonMessage{ + Length: json.Number(strconv.Itoa(int(d.Length))), + HeadersLen: json.Number(strconv.Itoa(int(d.HeadersLen))), + PreludeCRC: json.Number(strconv.Itoa(int(d.PreludeCRC))), + Headers: d.Headers, + Payload: d.Payload, + CRC: json.Number(strconv.Itoa(int(d.CRC))), + } + + return json.Marshal(jsonMsg) +} + +func numAsUint32(n json.Number) (uint32, error) { + v, err := n.Int64() + if err != nil { + return 0, fmt.Errorf("failed to get int64 json number, %v", err) + } + + return uint32(v), nil +} + +func (d decodedMessage) Message() Message { + return Message{ + Headers: Headers(d.Headers), + Payload: d.Payload, + } +} + +type decodedHeaders Headers + +func (hs *decodedHeaders) UnmarshalJSON(b []byte) error { + var jsonHeaders []struct { + Name string `json:"name"` + Type valueType `json:"type"` + Value interface{} `json:"value"` + } + + decoder := json.NewDecoder(bytes.NewReader(b)) + decoder.UseNumber() + if err := decoder.Decode(&jsonHeaders); err != nil { + return err + } + + var headers Headers + for _, h := range jsonHeaders { + value, err := valueFromType(h.Type, h.Value) + if err != nil { + return err + } + headers.Set(h.Name, value) + } + (*hs) = decodedHeaders(headers) + + return nil +} + +func valueFromType(typ valueType, val interface{}) (Value, error) { + switch typ { + case trueValueType: + return BoolValue(true), nil + case falseValueType: + return BoolValue(false), nil + case int8ValueType: + v, err := val.(json.Number).Int64() + return Int8Value(int8(v)), err + case int16ValueType: + v, err := val.(json.Number).Int64() + return Int16Value(int16(v)), err + case int32ValueType: + v, err := val.(json.Number).Int64() + return Int32Value(int32(v)), err + case int64ValueType: + v, err := val.(json.Number).Int64() + return Int64Value(v), err + case bytesValueType: + v, err := base64.StdEncoding.DecodeString(val.(string)) + return BytesValue(v), err + case stringValueType: + v, err := base64.StdEncoding.DecodeString(val.(string)) + return StringValue(string(v)), err + case timestampValueType: + v, err := val.(json.Number).Int64() + return TimestampValue(timeFromEpochMilli(v)), err + case uuidValueType: + v, err := base64.StdEncoding.DecodeString(val.(string)) + var tv UUIDValue + copy(tv[:], v) + return tv, err + default: + panic(fmt.Sprintf("unknown type, %s, %T", typ.String(), val)) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/decode.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/decode.go new file mode 100644 index 000000000..4b972b2d6 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/decode.go @@ -0,0 +1,199 @@ +package eventstream + +import ( + "bytes" + "encoding/binary" + "encoding/hex" + "encoding/json" + "fmt" + "hash" + "hash/crc32" + "io" + + "github.com/aws/aws-sdk-go/aws" +) + +// Decoder provides decoding of an Event Stream messages. +type Decoder struct { + r io.Reader + logger aws.Logger +} + +// NewDecoder initializes and returns a Decoder for decoding event +// stream messages from the reader provided. +func NewDecoder(r io.Reader) *Decoder { + return &Decoder{ + r: r, + } +} + +// Decode attempts to decode a single message from the event stream reader. +// Will return the event stream message, or error if Decode fails to read +// the message from the stream. +func (d *Decoder) Decode(payloadBuf []byte) (m Message, err error) { + reader := d.r + if d.logger != nil { + debugMsgBuf := bytes.NewBuffer(nil) + reader = io.TeeReader(reader, debugMsgBuf) + defer func() { + logMessageDecode(d.logger, debugMsgBuf, m, err) + }() + } + + crc := crc32.New(crc32IEEETable) + hashReader := io.TeeReader(reader, crc) + + prelude, err := decodePrelude(hashReader, crc) + if err != nil { + return Message{}, err + } + + if prelude.HeadersLen > 0 { + lr := io.LimitReader(hashReader, int64(prelude.HeadersLen)) + m.Headers, err = decodeHeaders(lr) + if err != nil { + return Message{}, err + } + } + + if payloadLen := prelude.PayloadLen(); payloadLen > 0 { + buf, err := decodePayload(payloadBuf, io.LimitReader(hashReader, int64(payloadLen))) + if err != nil { + return Message{}, err + } + m.Payload = buf + } + + msgCRC := crc.Sum32() + if err := validateCRC(reader, msgCRC); err != nil { + return Message{}, err + } + + return m, nil +} + +// UseLogger specifies the Logger that that the decoder should use to log the +// message decode to. +func (d *Decoder) UseLogger(logger aws.Logger) { + d.logger = logger +} + +func logMessageDecode(logger aws.Logger, msgBuf *bytes.Buffer, msg Message, decodeErr error) { + w := bytes.NewBuffer(nil) + defer func() { logger.Log(w.String()) }() + + fmt.Fprintf(w, "Raw message:\n%s\n", + hex.Dump(msgBuf.Bytes())) + + if decodeErr != nil { + fmt.Fprintf(w, "Decode error: %v\n", decodeErr) + return + } + + rawMsg, err := msg.rawMessage() + if err != nil { + fmt.Fprintf(w, "failed to create raw message, %v\n", err) + return + } + + decodedMsg := decodedMessage{ + rawMessage: rawMsg, + Headers: decodedHeaders(msg.Headers), + } + + fmt.Fprintf(w, "Decoded message:\n") + encoder := json.NewEncoder(w) + if err := encoder.Encode(decodedMsg); err != nil { + fmt.Fprintf(w, "failed to generate decoded message, %v\n", err) + } +} + +func decodePrelude(r io.Reader, crc hash.Hash32) (messagePrelude, error) { + var p messagePrelude + + var err error + p.Length, err = decodeUint32(r) + if err != nil { + return messagePrelude{}, err + } + + p.HeadersLen, err = decodeUint32(r) + if err != nil { + return messagePrelude{}, err + } + + if err := p.ValidateLens(); err != nil { + return messagePrelude{}, err + } + + preludeCRC := crc.Sum32() + if err := validateCRC(r, preludeCRC); err != nil { + return messagePrelude{}, err + } + + p.PreludeCRC = preludeCRC + + return p, nil +} + +func decodePayload(buf []byte, r io.Reader) ([]byte, error) { + w := bytes.NewBuffer(buf[0:0]) + + _, err := io.Copy(w, r) + return w.Bytes(), err +} + +func decodeUint8(r io.Reader) (uint8, error) { + type byteReader interface { + ReadByte() (byte, error) + } + + if br, ok := r.(byteReader); ok { + v, err := br.ReadByte() + return uint8(v), err + } + + var b [1]byte + _, err := io.ReadFull(r, b[:]) + return uint8(b[0]), err +} +func decodeUint16(r io.Reader) (uint16, error) { + var b [2]byte + bs := b[:] + _, err := io.ReadFull(r, bs) + if err != nil { + return 0, err + } + return binary.BigEndian.Uint16(bs), nil +} +func decodeUint32(r io.Reader) (uint32, error) { + var b [4]byte + bs := b[:] + _, err := io.ReadFull(r, bs) + if err != nil { + return 0, err + } + return binary.BigEndian.Uint32(bs), nil +} +func decodeUint64(r io.Reader) (uint64, error) { + var b [8]byte + bs := b[:] + _, err := io.ReadFull(r, bs) + if err != nil { + return 0, err + } + return binary.BigEndian.Uint64(bs), nil +} + +func validateCRC(r io.Reader, expect uint32) error { + msgCRC, err := decodeUint32(r) + if err != nil { + return err + } + + if msgCRC != expect { + return ChecksumError{} + } + + return nil +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/decode_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/decode_test.go new file mode 100644 index 000000000..2dcdbbfab --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/decode_test.go @@ -0,0 +1,168 @@ +package eventstream + +import ( + "bytes" + "encoding/hex" + "os" + "path/filepath" + "reflect" + "testing" +) + +func TestWriteEncodedFromDecoded(t *testing.T) { + cases, err := readPositiveTests("testdata") + if err != nil { + t.Fatalf("failed to load positive tests, %v", err) + } + + for _, c := range cases { + f, err := os.Create(filepath.Join("testdata", "encoded", "positive", c.Name)) + if err != nil { + t.Fatalf("failed to open %q, %v", c.Name, err) + } + defer f.Close() + + encoder := NewEncoder(f) + + msg := c.Decoded.Message() + if err := encoder.Encode(msg); err != nil { + t.Errorf("failed to encode %q, %v", c.Name, err) + } + } +} + +func TestDecoder_Decode(t *testing.T) { + cases, err := readPositiveTests("testdata") + if err != nil { + t.Fatalf("failed to load positive tests, %v", err) + } + + for _, c := range cases { + decoder := NewDecoder(bytes.NewBuffer(c.Encoded)) + + msg, err := decoder.Decode(nil) + if err != nil { + t.Fatalf("%s, expect no decode error, got %v", c.Name, err) + } + + raw, err := msg.rawMessage() // rawMessage will fail if payload read CRC fails + if err != nil { + t.Fatalf("%s, failed to get raw decoded message %v", c.Name, err) + } + + if e, a := c.Decoded.Length, raw.Length; e != a { + t.Errorf("%s, expect %v length, got %v", c.Name, e, a) + } + if e, a := c.Decoded.HeadersLen, raw.HeadersLen; e != a { + t.Errorf("%s, expect %v HeadersLen, got %v", c.Name, e, a) + } + if e, a := c.Decoded.PreludeCRC, raw.PreludeCRC; e != a { + t.Errorf("%s, expect %v PreludeCRC, got %v", c.Name, e, a) + } + if e, a := Headers(c.Decoded.Headers), msg.Headers; !reflect.DeepEqual(e, a) { + t.Errorf("%s, expect %v headers, got %v", c.Name, e, a) + } + if e, a := c.Decoded.Payload, raw.Payload; !bytes.Equal(e, a) { + t.Errorf("%s, expect %v payload, got %v", c.Name, e, a) + } + if e, a := c.Decoded.CRC, raw.CRC; e != a { + t.Errorf("%s, expect %v CRC, got %v", c.Name, e, a) + } + } +} + +func TestDecoder_Decode_Negative(t *testing.T) { + cases, err := readNegativeTests("testdata") + if err != nil { + t.Fatalf("failed to load negative tests, %v", err) + } + + for _, c := range cases { + decoder := NewDecoder(bytes.NewBuffer(c.Encoded)) + + msg, err := decoder.Decode(nil) + if err == nil { + rawMsg, rawMsgErr := msg.rawMessage() + t.Fatalf("%s, expect error, got none, %s,\n%s\n%#v, %v\n", c.Name, + c.Err, hex.Dump(c.Encoded), rawMsg, rawMsgErr) + } + } +} + +var testEncodedMsg = []byte{0, 0, 0, 61, 0, 0, 0, 32, 7, 253, 131, 150, 12, 99, 111, 110, 116, 101, 110, 116, 45, 116, 121, 112, 101, 7, 0, 16, 97, 112, 112, 108, 105, 99, 97, 116, 105, 111, 110, 47, 106, 115, 111, 110, 123, 39, 102, 111, 111, 39, 58, 39, 98, 97, 114, 39, 125, 141, 156, 8, 177} + +func TestDecoder_DecodeMultipleMessages(t *testing.T) { + const ( + expectMsgCount = 10 + expectPayloadLen = 13 + ) + + r := bytes.NewBuffer(nil) + for i := 0; i < expectMsgCount; i++ { + r.Write(testEncodedMsg) + } + + decoder := NewDecoder(r) + + var err error + var msg Message + var count int + for { + msg, err = decoder.Decode(nil) + if err != nil { + break + } + count++ + + if e, a := expectPayloadLen, len(msg.Payload); e != a { + t.Errorf("expect %v payload len, got %v", e, a) + } + + if e, a := []byte(`{'foo':'bar'}`), msg.Payload; !bytes.Equal(e, a) { + t.Errorf("expect %v payload, got %v", e, a) + } + } + + type causer interface { + Cause() error + } + if err != nil && count != expectMsgCount { + t.Fatalf("expect, no error, got %v", err) + } + + if e, a := expectMsgCount, count; e != a { + t.Errorf("expect %v messages read, got %v", e, a) + } +} + +func BenchmarkDecode(b *testing.B) { + r := bytes.NewReader(testEncodedMsg) + decoder := NewDecoder(r) + payloadBuf := make([]byte, 0, 5*1024) + b.ResetTimer() + + for i := 0; i < b.N; i++ { + msg, err := decoder.Decode(payloadBuf) + if err != nil { + b.Fatal(err) + } + + // Release the payload buffer + payloadBuf = msg.Payload[0:0] + r.Seek(0, 0) + } +} + +func BenchmarkDecode_NoPayloadBuf(b *testing.B) { + r := bytes.NewReader(testEncodedMsg) + decoder := NewDecoder(r) + b.ResetTimer() + + for i := 0; i < b.N; i++ { + _, err := decoder.Decode(nil) + if err != nil { + b.Fatal(err) + } + r.Seek(0, 0) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/encode.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/encode.go new file mode 100644 index 000000000..150a60981 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/encode.go @@ -0,0 +1,114 @@ +package eventstream + +import ( + "bytes" + "encoding/binary" + "hash" + "hash/crc32" + "io" +) + +// Encoder provides EventStream message encoding. +type Encoder struct { + w io.Writer + + headersBuf *bytes.Buffer +} + +// NewEncoder initializes and returns an Encoder to encode Event Stream +// messages to an io.Writer. +func NewEncoder(w io.Writer) *Encoder { + return &Encoder{ + w: w, + headersBuf: bytes.NewBuffer(nil), + } +} + +// Encode encodes a single EventStream message to the io.Writer the Encoder +// was created with. An error is returned if writing the message fails. +func (e *Encoder) Encode(msg Message) error { + e.headersBuf.Reset() + + err := encodeHeaders(e.headersBuf, msg.Headers) + if err != nil { + return err + } + + crc := crc32.New(crc32IEEETable) + hashWriter := io.MultiWriter(e.w, crc) + + headersLen := uint32(e.headersBuf.Len()) + payloadLen := uint32(len(msg.Payload)) + + if err := encodePrelude(hashWriter, crc, headersLen, payloadLen); err != nil { + return err + } + + if headersLen > 0 { + if _, err := io.Copy(hashWriter, e.headersBuf); err != nil { + return err + } + } + + if payloadLen > 0 { + if _, err := hashWriter.Write(msg.Payload); err != nil { + return err + } + } + + msgCRC := crc.Sum32() + return binary.Write(e.w, binary.BigEndian, msgCRC) +} + +func encodePrelude(w io.Writer, crc hash.Hash32, headersLen, payloadLen uint32) error { + p := messagePrelude{ + Length: minMsgLen + headersLen + payloadLen, + HeadersLen: headersLen, + } + if err := p.ValidateLens(); err != nil { + return err + } + + err := binaryWriteFields(w, binary.BigEndian, + p.Length, + p.HeadersLen, + ) + if err != nil { + return err + } + + p.PreludeCRC = crc.Sum32() + err = binary.Write(w, binary.BigEndian, p.PreludeCRC) + if err != nil { + return err + } + + return nil +} + +func encodeHeaders(w io.Writer, headers Headers) error { + for _, h := range headers { + hn := headerName{ + Len: uint8(len(h.Name)), + } + copy(hn.Name[:hn.Len], h.Name) + if err := hn.encode(w); err != nil { + return err + } + + if err := h.Value.encode(w); err != nil { + return err + } + } + + return nil +} + +func binaryWriteFields(w io.Writer, order binary.ByteOrder, vs ...interface{}) error { + for _, v := range vs { + if err := binary.Write(w, order, v); err != nil { + return err + } + } + return nil +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/encode_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/encode_test.go new file mode 100644 index 000000000..0cbfe4a7d --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/encode_test.go @@ -0,0 +1,50 @@ +package eventstream + +import ( + "bytes" + "encoding/hex" + "reflect" + "testing" +) + +func TestEncoder_Encode(t *testing.T) { + cases, err := readPositiveTests("testdata") + if err != nil { + t.Fatalf("failed to load positive tests, %v", err) + } + + for _, c := range cases { + var w bytes.Buffer + encoder := NewEncoder(&w) + + err = encoder.Encode(c.Decoded.Message()) + if err != nil { + t.Fatalf("%s, failed to encode message, %v", c.Name, err) + } + + if e, a := c.Encoded, w.Bytes(); !reflect.DeepEqual(e, a) { + t.Errorf("%s, expect:\n%v\nactual:\n%v\n", c.Name, + hex.Dump(e), hex.Dump(a)) + } + } +} + +func BenchmarkEncode(b *testing.B) { + var w bytes.Buffer + encoder := NewEncoder(&w) + msg := Message{ + Headers: Headers{ + {Name: "event-id", Value: Int16Value(123)}, + }, + Payload: []byte(`{"abc":123}`), + } + + b.ResetTimer() + + for i := 0; i < b.N; i++ { + err := encoder.Encode(msg) + if err != nil { + b.Fatal(err) + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/error.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/error.go new file mode 100644 index 000000000..5481ef307 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/error.go @@ -0,0 +1,23 @@ +package eventstream + +import "fmt" + +// LengthError provides the error for items being larger than a maximum length. +type LengthError struct { + Part string + Want int + Have int + Value interface{} +} + +func (e LengthError) Error() string { + return fmt.Sprintf("%s length invalid, %d/%d, %v", + e.Part, e.Want, e.Have, e.Value) +} + +// ChecksumError provides the error for message checksum invalidation errors. +type ChecksumError struct{} + +func (e ChecksumError) Error() string { + return "message checksum mismatch" +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/api.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/api.go new file mode 100644 index 000000000..4a4e64c71 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/api.go @@ -0,0 +1,160 @@ +package eventstreamapi + +import ( + "fmt" + "io" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/eventstream" +) + +// Unmarshaler provides the interface for unmarshaling a EventStream +// message into a SDK type. +type Unmarshaler interface { + UnmarshalEvent(protocol.PayloadUnmarshaler, eventstream.Message) error +} + +// EventStream headers with specific meaning to async API functionality. +const ( + MessageTypeHeader = `:message-type` // Identifies type of message. + EventMessageType = `event` + ErrorMessageType = `error` + ExceptionMessageType = `exception` + + // Message Events + EventTypeHeader = `:event-type` // Identifies message event type e.g. "Stats". + + // Message Error + ErrorCodeHeader = `:error-code` + ErrorMessageHeader = `:error-message` + + // Message Exception + ExceptionTypeHeader = `:exception-type` +) + +// EventReader provides reading from the EventStream of an reader. +type EventReader struct { + reader io.ReadCloser + decoder *eventstream.Decoder + + unmarshalerForEventType func(string) (Unmarshaler, error) + payloadUnmarshaler protocol.PayloadUnmarshaler + + payloadBuf []byte +} + +// NewEventReader returns a EventReader built from the reader and unmarshaler +// provided. Use ReadStream method to start reading from the EventStream. +func NewEventReader( + reader io.ReadCloser, + payloadUnmarshaler protocol.PayloadUnmarshaler, + unmarshalerForEventType func(string) (Unmarshaler, error), +) *EventReader { + return &EventReader{ + reader: reader, + decoder: eventstream.NewDecoder(reader), + payloadUnmarshaler: payloadUnmarshaler, + unmarshalerForEventType: unmarshalerForEventType, + payloadBuf: make([]byte, 10*1024), + } +} + +// UseLogger instructs the EventReader to use the logger and log level +// specified. +func (r *EventReader) UseLogger(logger aws.Logger, logLevel aws.LogLevelType) { + if logger != nil && logLevel.Matches(aws.LogDebugWithEventStreamBody) { + r.decoder.UseLogger(logger) + } +} + +// ReadEvent attempts to read a message from the EventStream and return the +// unmarshaled event value that the message is for. +// +// For EventStream API errors check if the returned error satisfies the +// awserr.Error interface to get the error's Code and Message components. +// +// EventUnmarshalers called with EventStream messages must take copies of the +// message's Payload. The payload will is reused between events read. +func (r *EventReader) ReadEvent() (event interface{}, err error) { + msg, err := r.decoder.Decode(r.payloadBuf) + if err != nil { + return nil, err + } + defer func() { + // Reclaim payload buffer for next message read. + r.payloadBuf = msg.Payload[0:0] + }() + + typ, err := GetHeaderString(msg, MessageTypeHeader) + if err != nil { + return nil, err + } + + switch typ { + case EventMessageType: + return r.unmarshalEventMessage(msg) + case ErrorMessageType: + return nil, r.unmarshalErrorMessage(msg) + default: + return nil, fmt.Errorf("unknown eventstream message type, %v", typ) + } +} + +func (r *EventReader) unmarshalEventMessage( + msg eventstream.Message, +) (event interface{}, err error) { + eventType, err := GetHeaderString(msg, EventTypeHeader) + if err != nil { + return nil, err + } + + ev, err := r.unmarshalerForEventType(eventType) + if err != nil { + return nil, err + } + + err = ev.UnmarshalEvent(r.payloadUnmarshaler, msg) + if err != nil { + return nil, err + } + + return ev, nil +} + +func (r *EventReader) unmarshalErrorMessage(msg eventstream.Message) (err error) { + var msgErr messageError + + msgErr.code, err = GetHeaderString(msg, ErrorCodeHeader) + if err != nil { + return err + } + + msgErr.msg, err = GetHeaderString(msg, ErrorMessageHeader) + if err != nil { + return err + } + + return msgErr +} + +// Close closes the EventReader's EventStream reader. +func (r *EventReader) Close() error { + return r.reader.Close() +} + +// GetHeaderString returns the value of the header as a string. If the header +// is not set or the value is not a string an error will be returned. +func GetHeaderString(msg eventstream.Message, headerName string) (string, error) { + headerVal := msg.Headers.Get(headerName) + if headerVal == nil { + return "", fmt.Errorf("error header %s not present", headerName) + } + + v, ok := headerVal.Get().(string) + if !ok { + return "", fmt.Errorf("error header value is not a string, %T", headerVal) + } + + return v, nil +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/api_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/api_test.go new file mode 100644 index 000000000..565d9277d --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/api_test.go @@ -0,0 +1,197 @@ +package eventstreamapi + +import ( + "bytes" + "fmt" + "io" + "io/ioutil" + "testing" + + "github.com/aws/aws-sdk-go/aws/awserr" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/eventstream" + "github.com/aws/aws-sdk-go/private/protocol/restjson" +) + +var eventMessageTypeHeader = eventstream.Header{ + Name: MessageTypeHeader, + Value: eventstream.StringValue(EventMessageType), +} + +func TestEventReader(t *testing.T) { + stream := createStream( + eventstream.Message{ + Headers: eventstream.Headers{ + eventMessageTypeHeader, + eventstream.Header{ + Name: EventTypeHeader, + Value: eventstream.StringValue("eventABC"), + }, + }, + }, + eventstream.Message{ + Headers: eventstream.Headers{ + eventMessageTypeHeader, + eventstream.Header{ + Name: EventTypeHeader, + Value: eventstream.StringValue("eventEFG"), + }, + }, + }, + ) + + var unmarshalers request.HandlerList + unmarshalers.PushBackNamed(restjson.UnmarshalHandler) + + eventReader := NewEventReader(stream, + protocol.HandlerPayloadUnmarshal{ + Unmarshalers: unmarshalers, + }, + unmarshalerForEventType, + ) + + event, err := eventReader.ReadEvent() + if err != nil { + t.Fatalf("expect no error, got %v", err) + } + + if event == nil { + t.Fatalf("expect event got none") + } + + event, err = eventReader.ReadEvent() + if err == nil { + t.Fatalf("expect error for unknown event, got none") + } + + if event != nil { + t.Fatalf("expect no event, got %T, %v", event, event) + } +} + +func TestEventReader_Error(t *testing.T) { + stream := createStream( + eventstream.Message{ + Headers: eventstream.Headers{ + eventstream.Header{ + Name: MessageTypeHeader, + Value: eventstream.StringValue(ErrorMessageType), + }, + eventstream.Header{ + Name: ErrorCodeHeader, + Value: eventstream.StringValue("errorCode"), + }, + eventstream.Header{ + Name: ErrorMessageHeader, + Value: eventstream.StringValue("error message occur"), + }, + }, + }, + ) + + var unmarshalers request.HandlerList + unmarshalers.PushBackNamed(restjson.UnmarshalHandler) + + eventReader := NewEventReader(stream, + protocol.HandlerPayloadUnmarshal{ + Unmarshalers: unmarshalers, + }, + unmarshalerForEventType, + ) + + event, err := eventReader.ReadEvent() + if err == nil { + t.Fatalf("expect error got none") + } + + if event != nil { + t.Fatalf("expect no event, got %v", event) + } + + aerr := err.(awserr.Error) + if e, a := "errorCode", aerr.Code(); e != a { + t.Errorf("expect %v code, got %v", e, a) + } + if e, a := "error message occur", aerr.Message(); e != a { + t.Errorf("expect %v message, got %v", e, a) + } +} + +func BenchmarkEventReader(b *testing.B) { + var buf bytes.Buffer + encoder := eventstream.NewEncoder(&buf) + msg := eventstream.Message{ + Headers: eventstream.Headers{ + eventMessageTypeHeader, + eventstream.Header{ + Name: EventTypeHeader, + Value: eventstream.StringValue("eventABC"), + }, + }, + } + if err := encoder.Encode(msg); err != nil { + b.Fatalf("failed to encode message, %v", err) + } + stream := bytes.NewReader(buf.Bytes()) + + var unmarshalers request.HandlerList + unmarshalers.PushBackNamed(restjson.UnmarshalHandler) + + eventReader := NewEventReader(ioutil.NopCloser(stream), + protocol.HandlerPayloadUnmarshal{ + Unmarshalers: unmarshalers, + }, + unmarshalerForEventType, + ) + b.ResetTimer() + + for i := 0; i < b.N; i++ { + stream.Seek(0, 0) + + event, err := eventReader.ReadEvent() + if err != nil { + b.Fatalf("expect no error, got %v", err) + } + if event == nil { + b.Fatalf("expect event got none") + } + } +} + +func unmarshalerForEventType(eventType string) (Unmarshaler, error) { + switch eventType { + case "eventABC": + return &eventABC{}, nil + default: + return nil, fmt.Errorf("unknown event type, %v", eventType) + } +} + +type eventABC struct { + _ struct{} + + HeaderField string + Payload []byte +} + +func (e *eventABC) UnmarshalEvent( + unmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + return nil +} + +func createStream(msgs ...eventstream.Message) io.ReadCloser { + w := bytes.NewBuffer(nil) + + encoder := eventstream.NewEncoder(w) + + for _, msg := range msgs { + if err := encoder.Encode(msg); err != nil { + panic("createStream failed, " + err.Error()) + } + } + + return ioutil.NopCloser(w) +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/error.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/error.go new file mode 100644 index 000000000..5ea5a988b --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi/error.go @@ -0,0 +1,24 @@ +package eventstreamapi + +import "fmt" + +type messageError struct { + code string + msg string +} + +func (e messageError) Code() string { + return e.code +} + +func (e messageError) Message() string { + return e.msg +} + +func (e messageError) Error() string { + return fmt.Sprintf("%s: %s", e.code, e.msg) +} + +func (e messageError) OrigErr() error { + return nil +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamtest/testing.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamtest/testing.go new file mode 100644 index 000000000..c566f285f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamtest/testing.go @@ -0,0 +1,116 @@ +package eventstreamtest + +import ( + "bytes" + "crypto/tls" + "fmt" + "io" + "net/http" + "net/http/httptest" + "testing" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/awstesting/unit" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/eventstream" + "github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi" + "golang.org/x/net/http2" +) + +// ServeEventStream provides serving EventStream messages from a HTTP server to +// the client. The events are sent sequentially to the client without delay. +type ServeEventStream struct { + T *testing.T + Events []eventstream.Message +} + +func (s ServeEventStream) ServeHTTP(w http.ResponseWriter, r *http.Request) { + w.WriteHeader(http.StatusOK) + encoder := eventstream.NewEncoder(flushWriter{w}) + + for _, event := range s.Events { + encoder.Encode(event) + } +} + +// SetupEventStreamSession creates a HTTP server SDK session for communicating +// with that server to be used for EventStream APIs. If HTTP/2 is enabled the +// server/client will only attempt to use HTTP/2. +func SetupEventStreamSession( + t *testing.T, handler http.Handler, h2 bool, +) (sess *session.Session, cleanupFn func(), err error) { + server := httptest.NewUnstartedServer(handler) + server.Config.TLSConfig = &tls.Config{ + InsecureSkipVerify: true, + } + + clientTrans := &http.Transport{ + TLSClientConfig: &tls.Config{ + InsecureSkipVerify: true, + }, + } + + if h2 { + http2.ConfigureServer(server.Config, nil) + http2.ConfigureTransport(clientTrans) + server.Config.TLSConfig.NextProtos = []string{http2.NextProtoTLS} + clientTrans.TLSClientConfig.NextProtos = []string{http2.NextProtoTLS} + } + server.TLS = server.Config.TLSConfig + + server.StartTLS() + cleanupFn = func() { + server.Close() + } + + client := &http.Client{ + Transport: clientTrans, + } + + sess, err = session.NewSession(unit.Session.Config, &aws.Config{ + Endpoint: &server.URL, + DisableParamValidation: aws.Bool(true), + HTTPClient: client, + // LogLevel: aws.LogLevel(aws.LogDebugWithEventStreamBody), + }) + if err != nil { + return nil, nil, err + } + + return sess, cleanupFn, nil +} + +type flushWriter struct { + w io.Writer +} + +func (fw flushWriter) Write(p []byte) (n int, err error) { + n, err = fw.w.Write(p) + if f, ok := fw.w.(http.Flusher); ok { + f.Flush() + } + return +} + +// MarshalEventPayload marshals a SDK API shape into its associated wire +// protocol payload. +func MarshalEventPayload( + payloadMarshaler protocol.PayloadMarshaler, + v interface{}, +) []byte { + var w bytes.Buffer + err := payloadMarshaler.MarshalPayload(&w, v) + if err != nil { + panic(fmt.Sprintf("failed to marshal event %T, %v", v, v)) + } + + return w.Bytes() +} + +// EventMessageTypeHeader is a event message type header for marking an +// message as being the event type. +var EventMessageTypeHeader = eventstream.Header{ + Name: eventstreamapi.MessageTypeHeader, + Value: eventstream.StringValue(eventstreamapi.EventMessageType), +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header.go new file mode 100644 index 000000000..3b44dde2f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header.go @@ -0,0 +1,166 @@ +package eventstream + +import ( + "encoding/binary" + "fmt" + "io" +) + +// Headers are a collection of EventStream header values. +type Headers []Header + +// Header is a single EventStream Key Value header pair. +type Header struct { + Name string + Value Value +} + +// Set associates the name with a value. If the header name already exists in +// the Headers the value will be replaced with the new one. +func (hs *Headers) Set(name string, value Value) { + var i int + for ; i < len(*hs); i++ { + if (*hs)[i].Name == name { + (*hs)[i].Value = value + return + } + } + + *hs = append(*hs, Header{ + Name: name, Value: value, + }) +} + +// Get returns the Value associated with the header. Nil is returned if the +// value does not exist. +func (hs Headers) Get(name string) Value { + for i := 0; i < len(hs); i++ { + if h := hs[i]; h.Name == name { + return h.Value + } + } + return nil +} + +// Del deletes the value in the Headers if it exists. +func (hs *Headers) Del(name string) { + for i := 0; i < len(*hs); i++ { + if (*hs)[i].Name == name { + copy((*hs)[i:], (*hs)[i+1:]) + (*hs) = (*hs)[:len(*hs)-1] + } + } +} + +func decodeHeaders(r io.Reader) (Headers, error) { + hs := Headers{} + + for { + name, err := decodeHeaderName(r) + if err != nil { + if err == io.EOF { + // EOF while getting header name means no more headers + break + } + return nil, err + } + + value, err := decodeHeaderValue(r) + if err != nil { + return nil, err + } + + hs.Set(name, value) + } + + return hs, nil +} + +func decodeHeaderName(r io.Reader) (string, error) { + var n headerName + + var err error + n.Len, err = decodeUint8(r) + if err != nil { + return "", err + } + + name := n.Name[:n.Len] + if _, err := io.ReadFull(r, name); err != nil { + return "", err + } + + return string(name), nil +} + +func decodeHeaderValue(r io.Reader) (Value, error) { + var raw rawValue + + typ, err := decodeUint8(r) + if err != nil { + return nil, err + } + raw.Type = valueType(typ) + + var v Value + + switch raw.Type { + case trueValueType: + v = BoolValue(true) + case falseValueType: + v = BoolValue(false) + case int8ValueType: + var tv Int8Value + err = tv.decode(r) + v = tv + case int16ValueType: + var tv Int16Value + err = tv.decode(r) + v = tv + case int32ValueType: + var tv Int32Value + err = tv.decode(r) + v = tv + case int64ValueType: + var tv Int64Value + err = tv.decode(r) + v = tv + case bytesValueType: + var tv BytesValue + err = tv.decode(r) + v = tv + case stringValueType: + var tv StringValue + err = tv.decode(r) + v = tv + case timestampValueType: + var tv TimestampValue + err = tv.decode(r) + v = tv + case uuidValueType: + var tv UUIDValue + err = tv.decode(r) + v = tv + default: + panic(fmt.Sprintf("unknown value type %d", raw.Type)) + } + + // Error could be EOF, let caller deal with it + return v, err +} + +const maxHeaderNameLen = 255 + +type headerName struct { + Len uint8 + Name [maxHeaderNameLen]byte +} + +func (v headerName) encode(w io.Writer) error { + if err := binary.Write(w, binary.BigEndian, v.Len); err != nil { + return err + } + + _, err := w.Write(v.Name[:v.Len]) + return err +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_test.go new file mode 100644 index 000000000..67a0cbcfc --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_test.go @@ -0,0 +1,66 @@ +package eventstream + +import ( + "reflect" + "testing" + "time" +) + +func TestHeaders_Set(t *testing.T) { + expect := Headers{ + {Name: "ABC", Value: StringValue("123")}, + {Name: "EFG", Value: TimestampValue(time.Time{})}, + } + + var actual Headers + actual.Set("ABC", Int32Value(123)) + actual.Set("ABC", StringValue("123")) // replase case + actual.Set("EFG", TimestampValue(time.Time{})) + + if e, a := expect, actual; !reflect.DeepEqual(e, a) { + t.Errorf("expect %v headers, got %v", e, a) + } +} + +func TestHeaders_Get(t *testing.T) { + headers := Headers{ + {Name: "ABC", Value: StringValue("123")}, + {Name: "EFG", Value: TimestampValue(time.Time{})}, + } + + cases := []struct { + Name string + Value Value + }{ + {Name: "ABC", Value: StringValue("123")}, + {Name: "EFG", Value: TimestampValue(time.Time{})}, + {Name: "NotFound"}, + } + + for i, c := range cases { + actual := headers.Get(c.Name) + if e, a := c.Value, actual; !reflect.DeepEqual(e, a) { + t.Errorf("%d, expect %v value, got %v", i, e, a) + } + } +} + +func TestHeaders_Del(t *testing.T) { + headers := Headers{ + {Name: "ABC", Value: StringValue("123")}, + {Name: "EFG", Value: TimestampValue(time.Time{})}, + {Name: "HIJ", Value: StringValue("123")}, + {Name: "KML", Value: TimestampValue(time.Time{})}, + } + expectAfterDel := Headers{ + {Name: "EFG", Value: TimestampValue(time.Time{})}, + } + + headers.Del("HIJ") + headers.Del("ABC") + headers.Del("KML") + + if e, a := expectAfterDel, headers; !reflect.DeepEqual(e, a) { + t.Errorf("expect %v headers, got %v", e, a) + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_value.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_value.go new file mode 100644 index 000000000..d7786f92c --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_value.go @@ -0,0 +1,501 @@ +package eventstream + +import ( + "encoding/base64" + "encoding/binary" + "fmt" + "io" + "strconv" + "time" +) + +const maxHeaderValueLen = 1<<15 - 1 // 2^15-1 or 32KB - 1 + +// valueType is the EventStream header value type. +type valueType uint8 + +// Header value types +const ( + trueValueType valueType = iota + falseValueType + int8ValueType // Byte + int16ValueType // Short + int32ValueType // Integer + int64ValueType // Long + bytesValueType + stringValueType + timestampValueType + uuidValueType +) + +func (t valueType) String() string { + switch t { + case trueValueType: + return "bool" + case falseValueType: + return "bool" + case int8ValueType: + return "int8" + case int16ValueType: + return "int16" + case int32ValueType: + return "int32" + case int64ValueType: + return "int64" + case bytesValueType: + return "byte_array" + case stringValueType: + return "string" + case timestampValueType: + return "timestamp" + case uuidValueType: + return "uuid" + default: + return fmt.Sprintf("unknown value type %d", uint8(t)) + } +} + +type rawValue struct { + Type valueType + Len uint16 // Only set for variable length slices + Value []byte // byte representation of value, BigEndian encoding. +} + +func (r rawValue) encodeScalar(w io.Writer, v interface{}) error { + return binaryWriteFields(w, binary.BigEndian, + r.Type, + v, + ) +} + +func (r rawValue) encodeFixedSlice(w io.Writer, v []byte) error { + binary.Write(w, binary.BigEndian, r.Type) + + _, err := w.Write(v) + return err +} + +func (r rawValue) encodeBytes(w io.Writer, v []byte) error { + if len(v) > maxHeaderValueLen { + return LengthError{ + Part: "header value", + Want: maxHeaderValueLen, Have: len(v), + Value: v, + } + } + r.Len = uint16(len(v)) + + err := binaryWriteFields(w, binary.BigEndian, + r.Type, + r.Len, + ) + if err != nil { + return err + } + + _, err = w.Write(v) + return err +} + +func (r rawValue) encodeString(w io.Writer, v string) error { + if len(v) > maxHeaderValueLen { + return LengthError{ + Part: "header value", + Want: maxHeaderValueLen, Have: len(v), + Value: v, + } + } + r.Len = uint16(len(v)) + + type stringWriter interface { + WriteString(string) (int, error) + } + + err := binaryWriteFields(w, binary.BigEndian, + r.Type, + r.Len, + ) + if err != nil { + return err + } + + if sw, ok := w.(stringWriter); ok { + _, err = sw.WriteString(v) + } else { + _, err = w.Write([]byte(v)) + } + + return err +} + +func decodeFixedBytesValue(r io.Reader, buf []byte) error { + _, err := io.ReadFull(r, buf) + return err +} + +func decodeBytesValue(r io.Reader) ([]byte, error) { + var raw rawValue + var err error + raw.Len, err = decodeUint16(r) + if err != nil { + return nil, err + } + + buf := make([]byte, raw.Len) + _, err = io.ReadFull(r, buf) + if err != nil { + return nil, err + } + + return buf, nil +} + +func decodeStringValue(r io.Reader) (string, error) { + v, err := decodeBytesValue(r) + return string(v), err +} + +// Value represents the abstract header value. +type Value interface { + Get() interface{} + String() string + valueType() valueType + encode(io.Writer) error +} + +// An BoolValue provides eventstream encoding, and representation +// of a Go bool value. +type BoolValue bool + +// Get returns the underlying type +func (v BoolValue) Get() interface{} { + return bool(v) +} + +// valueType returns the EventStream header value type value. +func (v BoolValue) valueType() valueType { + if v { + return trueValueType + } + return falseValueType +} + +func (v BoolValue) String() string { + return strconv.FormatBool(bool(v)) +} + +// encode encodes the BoolValue into an eventstream binary value +// representation. +func (v BoolValue) encode(w io.Writer) error { + return binary.Write(w, binary.BigEndian, v.valueType()) +} + +// An Int8Value provides eventstream encoding, and representation of a Go +// int8 value. +type Int8Value int8 + +// Get returns the underlying value. +func (v Int8Value) Get() interface{} { + return int8(v) +} + +// valueType returns the EventStream header value type value. +func (Int8Value) valueType() valueType { + return int8ValueType +} + +func (v Int8Value) String() string { + return fmt.Sprintf("0x%02x", int8(v)) +} + +// encode encodes the Int8Value into an eventstream binary value +// representation. +func (v Int8Value) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + + return raw.encodeScalar(w, v) +} + +func (v *Int8Value) decode(r io.Reader) error { + n, err := decodeUint8(r) + if err != nil { + return err + } + + *v = Int8Value(n) + return nil +} + +// An Int16Value provides eventstream encoding, and representation of a Go +// int16 value. +type Int16Value int16 + +// Get returns the underlying value. +func (v Int16Value) Get() interface{} { + return int16(v) +} + +// valueType returns the EventStream header value type value. +func (Int16Value) valueType() valueType { + return int16ValueType +} + +func (v Int16Value) String() string { + return fmt.Sprintf("0x%04x", int16(v)) +} + +// encode encodes the Int16Value into an eventstream binary value +// representation. +func (v Int16Value) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + return raw.encodeScalar(w, v) +} + +func (v *Int16Value) decode(r io.Reader) error { + n, err := decodeUint16(r) + if err != nil { + return err + } + + *v = Int16Value(n) + return nil +} + +// An Int32Value provides eventstream encoding, and representation of a Go +// int32 value. +type Int32Value int32 + +// Get returns the underlying value. +func (v Int32Value) Get() interface{} { + return int32(v) +} + +// valueType returns the EventStream header value type value. +func (Int32Value) valueType() valueType { + return int32ValueType +} + +func (v Int32Value) String() string { + return fmt.Sprintf("0x%08x", int32(v)) +} + +// encode encodes the Int32Value into an eventstream binary value +// representation. +func (v Int32Value) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + return raw.encodeScalar(w, v) +} + +func (v *Int32Value) decode(r io.Reader) error { + n, err := decodeUint32(r) + if err != nil { + return err + } + + *v = Int32Value(n) + return nil +} + +// An Int64Value provides eventstream encoding, and representation of a Go +// int64 value. +type Int64Value int64 + +// Get returns the underlying value. +func (v Int64Value) Get() interface{} { + return int64(v) +} + +// valueType returns the EventStream header value type value. +func (Int64Value) valueType() valueType { + return int64ValueType +} + +func (v Int64Value) String() string { + return fmt.Sprintf("0x%016x", int64(v)) +} + +// encode encodes the Int64Value into an eventstream binary value +// representation. +func (v Int64Value) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + return raw.encodeScalar(w, v) +} + +func (v *Int64Value) decode(r io.Reader) error { + n, err := decodeUint64(r) + if err != nil { + return err + } + + *v = Int64Value(n) + return nil +} + +// An BytesValue provides eventstream encoding, and representation of a Go +// byte slice. +type BytesValue []byte + +// Get returns the underlying value. +func (v BytesValue) Get() interface{} { + return []byte(v) +} + +// valueType returns the EventStream header value type value. +func (BytesValue) valueType() valueType { + return bytesValueType +} + +func (v BytesValue) String() string { + return base64.StdEncoding.EncodeToString([]byte(v)) +} + +// encode encodes the BytesValue into an eventstream binary value +// representation. +func (v BytesValue) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + + return raw.encodeBytes(w, []byte(v)) +} + +func (v *BytesValue) decode(r io.Reader) error { + buf, err := decodeBytesValue(r) + if err != nil { + return err + } + + *v = BytesValue(buf) + return nil +} + +// An StringValue provides eventstream encoding, and representation of a Go +// string. +type StringValue string + +// Get returns the underlying value. +func (v StringValue) Get() interface{} { + return string(v) +} + +// valueType returns the EventStream header value type value. +func (StringValue) valueType() valueType { + return stringValueType +} + +func (v StringValue) String() string { + return string(v) +} + +// encode encodes the StringValue into an eventstream binary value +// representation. +func (v StringValue) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + + return raw.encodeString(w, string(v)) +} + +func (v *StringValue) decode(r io.Reader) error { + s, err := decodeStringValue(r) + if err != nil { + return err + } + + *v = StringValue(s) + return nil +} + +// An TimestampValue provides eventstream encoding, and representation of a Go +// timestamp. +type TimestampValue time.Time + +// Get returns the underlying value. +func (v TimestampValue) Get() interface{} { + return time.Time(v) +} + +// valueType returns the EventStream header value type value. +func (TimestampValue) valueType() valueType { + return timestampValueType +} + +func (v TimestampValue) epochMilli() int64 { + nano := time.Time(v).UnixNano() + msec := nano / int64(time.Millisecond) + return msec +} + +func (v TimestampValue) String() string { + msec := v.epochMilli() + return strconv.FormatInt(msec, 10) +} + +// encode encodes the TimestampValue into an eventstream binary value +// representation. +func (v TimestampValue) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + + msec := v.epochMilli() + return raw.encodeScalar(w, msec) +} + +func (v *TimestampValue) decode(r io.Reader) error { + n, err := decodeUint64(r) + if err != nil { + return err + } + + *v = TimestampValue(timeFromEpochMilli(int64(n))) + return nil +} + +func timeFromEpochMilli(t int64) time.Time { + secs := t / 1e3 + msec := t % 1e3 + return time.Unix(secs, msec*int64(time.Millisecond)) +} + +// An UUIDValue provides eventstream encoding, and representation of a UUID +// value. +type UUIDValue [16]byte + +// Get returns the underlying value. +func (v UUIDValue) Get() interface{} { + return v[:] +} + +// valueType returns the EventStream header value type value. +func (UUIDValue) valueType() valueType { + return uuidValueType +} + +func (v UUIDValue) String() string { + return fmt.Sprintf(`%X-%X-%X-%X-%X`, v[0:4], v[4:6], v[6:8], v[8:10], v[10:]) +} + +// encode encodes the UUIDValue into an eventstream binary value +// representation. +func (v UUIDValue) encode(w io.Writer) error { + raw := rawValue{ + Type: v.valueType(), + } + + return raw.encodeFixedSlice(w, v[:]) +} + +func (v *UUIDValue) decode(r io.Reader) error { + tv := (*v)[:] + return decodeFixedBytesValue(r, tv) +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_value_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_value_test.go new file mode 100644 index 000000000..f2e67f2fc --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/header_value_test.go @@ -0,0 +1,203 @@ +package eventstream + +import ( + "bytes" + "encoding/binary" + "io" + "reflect" + "testing" + "time" +) + +func binWrite(v interface{}) []byte { + var w bytes.Buffer + binary.Write(&w, binary.BigEndian, v) + return w.Bytes() +} + +var testValueEncodingCases = []struct { + Val Value + Expect []byte + Decode func(io.Reader) (Value, error) +}{ + { + BoolValue(true), + []byte{byte(trueValueType)}, + nil, + }, + { + BoolValue(false), + []byte{byte(falseValueType)}, + nil, + }, + { + Int8Value(0x0f), + []byte{byte(int8ValueType), 0x0f}, + func(r io.Reader) (Value, error) { + var v Int8Value + err := v.decode(r) + return v, err + }, + }, + { + Int16Value(0x0f), + append([]byte{byte(int16ValueType)}, binWrite(int16(0x0f))...), + func(r io.Reader) (Value, error) { + var v Int16Value + err := v.decode(r) + return v, err + }, + }, + { + Int32Value(0x0f), + append([]byte{byte(int32ValueType)}, binWrite(int32(0x0f))...), + func(r io.Reader) (Value, error) { + var v Int32Value + err := v.decode(r) + return v, err + }, + }, + { + Int64Value(0x0f), + append([]byte{byte(int64ValueType)}, binWrite(int64(0x0f))...), + func(r io.Reader) (Value, error) { + var v Int64Value + err := v.decode(r) + return v, err + }, + }, + { + BytesValue([]byte{0, 1, 2, 3}), + []byte{byte(bytesValueType), 0x00, 0x04, 0, 1, 2, 3}, + func(r io.Reader) (Value, error) { + var v BytesValue + err := v.decode(r) + return v, err + }, + }, + { + StringValue("abc123"), + append([]byte{byte(stringValueType), 0, 6}, []byte("abc123")...), + func(r io.Reader) (Value, error) { + var v StringValue + err := v.decode(r) + return v, err + }, + }, + { + TimestampValue( + time.Date(2014, 04, 04, 0, 1, 0, 0, time.FixedZone("PDT", -7)), + ), + append([]byte{byte(timestampValueType)}, binWrite(int64(1396569667000))...), + func(r io.Reader) (Value, error) { + var v TimestampValue + err := v.decode(r) + return v, err + }, + }, + { + UUIDValue( + [16]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0xa, 0xb, 0xc, 0xd, 0xe, 0xf}, + ), + []byte{byte(uuidValueType), 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0xa, 0xb, 0xc, 0xd, 0xe, 0xf}, + func(r io.Reader) (Value, error) { + var v UUIDValue + err := v.decode(r) + return v, err + }, + }, +} + +func TestValue_MarshalValue(t *testing.T) { + for i, c := range testValueEncodingCases { + var w bytes.Buffer + + if err := c.Val.encode(&w); err != nil { + t.Fatalf("%d, expect no error, got %v", i, err) + } + + if e, a := c.Expect, w.Bytes(); !reflect.DeepEqual(e, a) { + t.Errorf("%d, expect %v, got %v", i, e, a) + } + } +} + +func TestHeader_DecodeValues(t *testing.T) { + for i, c := range testValueEncodingCases { + r := bytes.NewBuffer(c.Expect) + v, err := decodeHeaderValue(r) + if err != nil { + t.Fatalf("%d, expect no error, got %v", i, err) + } + + switch tv := v.(type) { + case TimestampValue: + exp := time.Time(c.Val.(TimestampValue)) + if e, a := exp, time.Time(tv); !e.Equal(a) { + t.Errorf("%d, expect %v, got %v", i, e, a) + } + default: + if e, a := c.Val, v; !reflect.DeepEqual(e, a) { + t.Errorf("%d, expect %v, got %v", i, e, a) + } + } + } +} + +func TestValue_Decode(t *testing.T) { + for i, c := range testValueEncodingCases { + if c.Decode == nil { + continue + } + + r := bytes.NewBuffer(c.Expect) + r.ReadByte() // strip off Type field + + v, err := c.Decode(r) + if err != nil { + t.Fatalf("%d, expect no error, got %v", i, err) + } + + switch tv := v.(type) { + case TimestampValue: + exp := time.Time(c.Val.(TimestampValue)) + if e, a := exp, time.Time(tv); !e.Equal(a) { + t.Errorf("%d, expect %v, got %v", i, e, a) + } + default: + if e, a := c.Val, v; !reflect.DeepEqual(e, a) { + t.Errorf("%d, expect %v, got %v", i, e, a) + } + } + } +} + +func TestValue_String(t *testing.T) { + cases := []struct { + Val Value + Expect string + }{ + {BoolValue(true), "true"}, + {BoolValue(false), "false"}, + {Int8Value(0x0f), "0x0f"}, + {Int16Value(0x0f), "0x000f"}, + {Int32Value(0x0f), "0x0000000f"}, + {Int64Value(0x0f), "0x000000000000000f"}, + {BytesValue([]byte{0, 1, 2, 3}), "AAECAw=="}, + {StringValue("abc123"), "abc123"}, + {TimestampValue( + time.Date(2014, 04, 04, 0, 1, 0, 0, time.FixedZone("PDT", -7)), + ), + "1396569667000", + }, + {UUIDValue([16]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0xa, 0xb, 0xc, 0xd, 0xe, 0xf}), + "00010203-0405-0607-0809-0A0B0C0D0E0F", + }, + } + + for i, c := range cases { + if e, a := c.Expect, c.Val.String(); e != a { + t.Errorf("%d, expect %v, got %v", i, e, a) + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/message.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/message.go new file mode 100644 index 000000000..2dc012a66 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/message.go @@ -0,0 +1,103 @@ +package eventstream + +import ( + "bytes" + "encoding/binary" + "hash/crc32" +) + +const preludeLen = 8 +const preludeCRCLen = 4 +const msgCRCLen = 4 +const minMsgLen = preludeLen + preludeCRCLen + msgCRCLen +const maxPayloadLen = 1024 * 1024 * 16 // 16MB +const maxHeadersLen = 1024 * 128 // 128KB +const maxMsgLen = minMsgLen + maxHeadersLen + maxPayloadLen + +var crc32IEEETable = crc32.MakeTable(crc32.IEEE) + +// A Message provides the eventstream message representation. +type Message struct { + Headers Headers + Payload []byte +} + +func (m *Message) rawMessage() (rawMessage, error) { + var raw rawMessage + + if len(m.Headers) > 0 { + var headers bytes.Buffer + if err := encodeHeaders(&headers, m.Headers); err != nil { + return rawMessage{}, err + } + raw.Headers = headers.Bytes() + raw.HeadersLen = uint32(len(raw.Headers)) + } + + raw.Length = raw.HeadersLen + uint32(len(m.Payload)) + minMsgLen + + hash := crc32.New(crc32IEEETable) + binaryWriteFields(hash, binary.BigEndian, raw.Length, raw.HeadersLen) + raw.PreludeCRC = hash.Sum32() + + binaryWriteFields(hash, binary.BigEndian, raw.PreludeCRC) + + if raw.HeadersLen > 0 { + hash.Write(raw.Headers) + } + + // Read payload bytes and update hash for it as well. + if len(m.Payload) > 0 { + raw.Payload = m.Payload + hash.Write(raw.Payload) + } + + raw.CRC = hash.Sum32() + + return raw, nil +} + +type messagePrelude struct { + Length uint32 + HeadersLen uint32 + PreludeCRC uint32 +} + +func (p messagePrelude) PayloadLen() uint32 { + return p.Length - p.HeadersLen - minMsgLen +} + +func (p messagePrelude) ValidateLens() error { + if p.Length == 0 || p.Length > maxMsgLen { + return LengthError{ + Part: "message prelude", + Want: maxMsgLen, + Have: int(p.Length), + } + } + if p.HeadersLen > maxHeadersLen { + return LengthError{ + Part: "message headers", + Want: maxHeadersLen, + Have: int(p.HeadersLen), + } + } + if payloadLen := p.PayloadLen(); payloadLen > maxPayloadLen { + return LengthError{ + Part: "message payload", + Want: maxPayloadLen, + Have: int(payloadLen), + } + } + + return nil +} + +type rawMessage struct { + messagePrelude + + Headers []byte + Payload []byte + + CRC uint32 +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/shared_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/shared_test.go new file mode 100644 index 000000000..09e0f963b --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/shared_test.go @@ -0,0 +1,152 @@ +package eventstream + +import ( + "bufio" + "bytes" + "encoding/json" + "fmt" + "io/ioutil" + "path/filepath" + "testing" +) + +type testCase struct { + Name string + Encoded []byte + Decoded decodedMessage +} + +type testErrorCase struct { + Name string + Encoded []byte + Err string +} + +type rawTestCase struct { + Name string + Encoded, Decoded []byte +} + +func readRawTestCases(root, class string) (map[string]rawTestCase, error) { + encoded, err := readTests(filepath.Join(root, "encoded", class)) + if err != nil { + return nil, err + } + + decoded, err := readTests(filepath.Join(root, "decoded", class)) + if err != nil { + return nil, err + } + + if len(encoded) == 0 { + return nil, fmt.Errorf("expect encoded cases, found none") + } + + if len(encoded) != len(decoded) { + return nil, fmt.Errorf("encoded and decoded sets different") + } + + rawCases := map[string]rawTestCase{} + for name, encData := range encoded { + decData, ok := decoded[name] + if !ok { + return nil, fmt.Errorf("encoded %q case not found in decoded set", name) + } + + rawCases[name] = rawTestCase{ + Name: name, + Encoded: encData, + Decoded: decData, + } + } + + return rawCases, nil +} + +func readNegativeTests(root string) ([]testErrorCase, error) { + rawCases, err := readRawTestCases(root, "negative") + if err != nil { + return nil, err + } + + cases := make([]testErrorCase, 0, len(rawCases)) + for name, rawCase := range rawCases { + cases = append(cases, testErrorCase{ + Name: name, + Encoded: rawCase.Encoded, + Err: string(rawCase.Decoded), + }) + } + + return cases, nil +} + +func readPositiveTests(root string) ([]testCase, error) { + rawCases, err := readRawTestCases(root, "positive") + if err != nil { + return nil, err + } + + cases := make([]testCase, 0, len(rawCases)) + for name, rawCase := range rawCases { + + var dec decodedMessage + if err := json.Unmarshal(rawCase.Decoded, &dec); err != nil { + return nil, fmt.Errorf("failed to decode %q, %v", name, err) + } + + cases = append(cases, testCase{ + Name: name, + Encoded: rawCase.Encoded, + Decoded: dec, + }) + } + + return cases, nil +} + +func readTests(root string) (map[string][]byte, error) { + items, err := ioutil.ReadDir(root) + if err != nil { + return nil, fmt.Errorf("failed to read test suite %q dirs, %v", root, err) + } + + cases := map[string][]byte{} + for _, item := range items { + if item.IsDir() { + continue + } + + filename := filepath.Join(root, item.Name()) + data, err := ioutil.ReadFile(filename) + if err != nil { + return nil, fmt.Errorf("failed to read test_data file %q, %v", filename, err) + } + + cases[item.Name()] = data + } + + return cases, nil +} + +func compareLines(t *testing.T, a, b []byte) bool { + as := bufio.NewScanner(bytes.NewBuffer(a)) + bs := bufio.NewScanner(bytes.NewBuffer(b)) + + var failed bool + for { + if ab, bb := as.Scan(), bs.Scan(); ab != bb { + t.Errorf("expect a & b to have same number of lines") + return false + } else if !ab { + break + } + + if v1, v2 := as.Text(), bs.Text(); v1 != v2 { + t.Errorf("expect %q to be %q", v1, v2) + failed = true + } + } + + return !failed +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_header_len b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_header_len new file mode 100644 index 000000000..73e4d6f1f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_header_len @@ -0,0 +1 @@ +Prelude checksum mismatch \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_headers b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_headers new file mode 100644 index 000000000..f56e5c887 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_headers @@ -0,0 +1 @@ +Message checksum mismatch \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_length b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_length new file mode 100644 index 000000000..73e4d6f1f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_length @@ -0,0 +1 @@ +Prelude checksum mismatch \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_payload b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_payload new file mode 100644 index 000000000..f56e5c887 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/negative/corrupted_payload @@ -0,0 +1 @@ +Message checksum mismatch \ No newline at end of file diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/all_headers b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/all_headers new file mode 100644 index 000000000..fd8f96b88 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/all_headers @@ -0,0 +1,58 @@ +{ + "total_length": 204, + "headers_length": 175, + "prelude_crc": 263087306, + "headers": [ { + "name": "event-type", + "type": 4, + "value": 40972 + }, + { + "name": "content-type", + "type": 7, + "value": "YXBwbGljYXRpb24vanNvbg==" + }, + { + "name": "bool false", + "type": 1, + "value": false + }, + { + "name": "bool true", + "type": 0, + "value": true + }, + { + "name": "byte", + "type": 2, + "value": -49 + }, + { + "name": "byte buf", + "type": 6, + "value": "SSdtIGEgbGl0dGxlIHRlYXBvdCE=" + }, + { + "name": "timestamp", + "type": 8, + "value": 8675309 + }, + { + "name": "int16", + "type": 3, + "value": 42 + }, + { + "name": "int64", + "type": 5, + "value": 42424242 + }, + { + "name": "uuid", + "type": 9, + "value": "AQIDBAUGBwgJCgsMDQ4PEA==" + } + ], + "payload": "eydmb28nOidiYXInfQ==", + "message_crc": -1415188212 +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/empty_message b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/empty_message new file mode 100644 index 000000000..1d35df8e6 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/empty_message @@ -0,0 +1,8 @@ +{ + "total_length": 16, + "headers_length": 0, + "prelude_crc": 96618731, + "headers": [ ], + "payload": "", + "message_crc": 2107164927 +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/int32_header b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/int32_header new file mode 100644 index 000000000..852a0db6e --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/int32_header @@ -0,0 +1,13 @@ +{ + "total_length": 45, + "headers_length": 16, + "prelude_crc": 1103373496, + "headers": [ { + "name": "event-type", + "type": 4, + "value": 40972 + } + ], + "payload": "eydmb28nOidiYXInfQ==", + "message_crc": 921993376 +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/payload_no_headers b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/payload_no_headers new file mode 100644 index 000000000..1c96631ca --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/payload_no_headers @@ -0,0 +1,8 @@ +{ + "total_length": 29, + "headers_length": 0, + "prelude_crc": -44921766, + "headers": [ ], + "payload": "eydmb28nOidiYXInfQ==", + "message_crc": -1016776394 +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/payload_one_str_header b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/payload_one_str_header new file mode 100644 index 000000000..e3bfd312f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/decoded/positive/payload_one_str_header @@ -0,0 +1,13 @@ +{ + "total_length": 61, + "headers_length": 32, + "prelude_crc": 134054806, + "headers": [ { + "name": "content-type", + "type": 7, + "value": "YXBwbGljYXRpb24vanNvbg==" + } + ], + "payload": "eydmb28nOidiYXInfQ==", + "message_crc": -1919153999 +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_header_len b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_header_len new file mode 100644 index 000000000..474929c83 Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_header_len differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_headers b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_headers new file mode 100644 index 000000000..802a2276c Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_headers differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_length b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_length new file mode 100644 index 000000000..4e55a3196 Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_length differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_payload b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_payload new file mode 100644 index 000000000..2ee9ef3ce Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/negative/corrupted_payload differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/all_headers b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/all_headers new file mode 100644 index 000000000..b986af148 Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/all_headers differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/empty_message b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/empty_message new file mode 100644 index 000000000..663632857 Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/empty_message differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/int32_header b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/int32_header new file mode 100644 index 000000000..4e13b503d Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/int32_header differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/payload_no_headers b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/payload_no_headers new file mode 100644 index 000000000..47733a111 Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/payload_no_headers differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/payload_one_str_header b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/payload_one_str_header new file mode 100644 index 000000000..d4abaa7f8 Binary files /dev/null and b/vendor/github.com/aws/aws-sdk-go/private/protocol/eventstream/testdata/encoded/positive/payload_one_str_header differ diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/build_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/build_test.go index 4bd8238a0..f9433a224 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/build_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/build_test.go @@ -75,7 +75,8 @@ func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice1protocoltest", + ServiceName: "InputService1ProtocolTest", + ServiceID: "InputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -223,7 +224,8 @@ func newInputService2ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice2protocoltest", + ServiceName: "InputService2ProtocolTest", + ServiceID: "InputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -370,7 +372,8 @@ func newInputService3ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice3protocoltest", + ServiceName: "InputService3ProtocolTest", + ServiceID: "InputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -600,7 +603,8 @@ func newInputService4ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice4protocoltest", + ServiceName: "InputService4ProtocolTest", + ServiceID: "InputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -748,7 +752,8 @@ func newInputService5ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice5protocoltest", + ServiceName: "InputService5ProtocolTest", + ServiceID: "InputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1301,7 +1306,8 @@ func newInputService6ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice6protocoltest", + ServiceName: "InputService6ProtocolTest", + ServiceID: "InputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1449,7 +1455,8 @@ func newInputService7ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice7protocoltest", + ServiceName: "InputService7ProtocolTest", + ServiceID: "InputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1670,7 +1677,8 @@ func newInputService8ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice8protocoltest", + ServiceName: "InputService8ProtocolTest", + ServiceID: "InputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/unmarshal_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/unmarshal_test.go index a749cf1cc..61e57810e 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/unmarshal_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc/unmarshal_test.go @@ -75,7 +75,8 @@ func newOutputService1ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice1protocoltest", + ServiceName: "OutputService1ProtocolTest", + ServiceID: "OutputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -274,7 +275,8 @@ func newOutputService2ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice2protocoltest", + ServiceName: "OutputService2ProtocolTest", + ServiceID: "OutputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -439,7 +441,8 @@ func newOutputService3ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice3protocoltest", + ServiceName: "OutputService3ProtocolTest", + ServiceID: "OutputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -602,7 +605,8 @@ func newOutputService4ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice4protocoltest", + ServiceName: "OutputService4ProtocolTest", + ServiceID: "OutputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -837,7 +841,8 @@ func newOutputService5ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice5protocoltest", + ServiceName: "OutputService5ProtocolTest", + ServiceID: "OutputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -980,7 +985,8 @@ func newOutputService6ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice6protocoltest", + ServiceName: "OutputService6ProtocolTest", + ServiceID: "OutputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1123,7 +1129,8 @@ func newOutputService7ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice7protocoltest", + ServiceName: "OutputService7ProtocolTest", + ServiceID: "OutputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1351,10 +1358,10 @@ func TestOutputService3ProtocolTestTimestampMembersCase1(t *testing.T) { if out == nil { t.Errorf("expect not to be nil") } - if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.StructMember.Foo.String(); e != a { + if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.StructMember.Foo.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.TimeMember.String(); e != a { + if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.TimeMember.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/payload.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/payload.go new file mode 100644 index 000000000..e21614a12 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/payload.go @@ -0,0 +1,81 @@ +package protocol + +import ( + "io" + "io/ioutil" + "net/http" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/request" +) + +// PayloadUnmarshaler provides the interface for unmarshaling a payload's +// reader into a SDK shape. +type PayloadUnmarshaler interface { + UnmarshalPayload(io.Reader, interface{}) error +} + +// HandlerPayloadUnmarshal implements the PayloadUnmarshaler from a +// HandlerList. This provides the support for unmarshaling a payload reader to +// a shape without needing a SDK request first. +type HandlerPayloadUnmarshal struct { + Unmarshalers request.HandlerList +} + +// UnmarshalPayload unmarshals the io.Reader payload into the SDK shape using +// the Unmarshalers HandlerList provided. Returns an error if unable +// unmarshaling fails. +func (h HandlerPayloadUnmarshal) UnmarshalPayload(r io.Reader, v interface{}) error { + req := &request.Request{ + HTTPRequest: &http.Request{}, + HTTPResponse: &http.Response{ + StatusCode: 200, + Header: http.Header{}, + Body: ioutil.NopCloser(r), + }, + Data: v, + } + + h.Unmarshalers.Run(req) + + return req.Error +} + +// PayloadMarshaler provides the interface for marshaling a SDK shape into and +// io.Writer. +type PayloadMarshaler interface { + MarshalPayload(io.Writer, interface{}) error +} + +// HandlerPayloadMarshal implements the PayloadMarshaler from a HandlerList. +// This provides support for marshaling a SDK shape into an io.Writer without +// needing a SDK request first. +type HandlerPayloadMarshal struct { + Marshalers request.HandlerList +} + +// MarshalPayload marshals the SDK shape into the io.Writer using the +// Marshalers HandlerList provided. Returns an error if unable if marshal +// fails. +func (h HandlerPayloadMarshal) MarshalPayload(w io.Writer, v interface{}) error { + req := request.New( + aws.Config{}, + metadata.ClientInfo{}, + request.Handlers{}, + nil, + &request.Operation{HTTPMethod: "GET"}, + v, + nil, + ) + + h.Marshalers.Run(req) + + if req.Error != nil { + return req.Error + } + + io.Copy(w, req.GetBody()) + + return nil +} diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/query/build_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/query/build_test.go index 0cef7b41a..50dcea52d 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/query/build_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/query/build_test.go @@ -75,7 +75,8 @@ func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice1protocoltest", + ServiceName: "InputService1ProtocolTest", + ServiceID: "InputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -384,7 +385,8 @@ func newInputService2ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice2protocoltest", + ServiceName: "InputService2ProtocolTest", + ServiceID: "InputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -541,7 +543,8 @@ func newInputService3ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice3protocoltest", + ServiceName: "InputService3ProtocolTest", + ServiceID: "InputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -760,7 +763,8 @@ func newInputService4ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice4protocoltest", + ServiceName: "InputService4ProtocolTest", + ServiceID: "InputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -995,7 +999,8 @@ func newInputService5ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice5protocoltest", + ServiceName: "InputService5ProtocolTest", + ServiceID: "InputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1140,7 +1145,8 @@ func newInputService6ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice6protocoltest", + ServiceName: "InputService6ProtocolTest", + ServiceID: "InputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1285,7 +1291,8 @@ func newInputService7ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice7protocoltest", + ServiceName: "InputService7ProtocolTest", + ServiceID: "InputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1438,7 +1445,8 @@ func newInputService8ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice8protocoltest", + ServiceName: "InputService8ProtocolTest", + ServiceID: "InputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1583,7 +1591,8 @@ func newInputService9ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice9protocoltest", + ServiceName: "InputService9ProtocolTest", + ServiceID: "InputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1728,7 +1737,8 @@ func newInputService10ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice10protocoltest", + ServiceName: "InputService10ProtocolTest", + ServiceID: "InputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1874,7 +1884,8 @@ func newInputService11ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice11protocoltest", + ServiceName: "InputService11ProtocolTest", + ServiceID: "InputService11ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2019,7 +2030,8 @@ func newInputService12ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice12protocoltest", + ServiceName: "InputService12ProtocolTest", + ServiceID: "InputService12ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2164,7 +2176,8 @@ func newInputService13ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice13protocoltest", + ServiceName: "InputService13ProtocolTest", + ServiceID: "InputService13ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2715,7 +2728,8 @@ func newInputService14ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice14protocoltest", + ServiceName: "InputService14ProtocolTest", + ServiceID: "InputService14ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2936,7 +2950,8 @@ func newInputService15ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice15protocoltest", + ServiceName: "InputService15ProtocolTest", + ServiceID: "InputService15ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/query/unmarshal_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/query/unmarshal_test.go index 17cdd3594..a1872cc74 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/query/unmarshal_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/query/unmarshal_test.go @@ -75,7 +75,8 @@ func newOutputService1ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice1protocoltest", + ServiceName: "OutputService1ProtocolTest", + ServiceID: "OutputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -282,7 +283,8 @@ func newOutputService2ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice2protocoltest", + ServiceName: "OutputService2ProtocolTest", + ServiceID: "OutputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -433,7 +435,8 @@ func newOutputService3ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice3protocoltest", + ServiceName: "OutputService3ProtocolTest", + ServiceID: "OutputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -577,7 +580,8 @@ func newOutputService4ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice4protocoltest", + ServiceName: "OutputService4ProtocolTest", + ServiceID: "OutputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -720,7 +724,8 @@ func newOutputService5ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice5protocoltest", + ServiceName: "OutputService5ProtocolTest", + ServiceID: "OutputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -863,7 +868,8 @@ func newOutputService6ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice6protocoltest", + ServiceName: "OutputService6ProtocolTest", + ServiceID: "OutputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1006,7 +1012,8 @@ func newOutputService7ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice7protocoltest", + ServiceName: "OutputService7ProtocolTest", + ServiceID: "OutputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1149,7 +1156,8 @@ func newOutputService8ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice8protocoltest", + ServiceName: "OutputService8ProtocolTest", + ServiceID: "OutputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1320,7 +1328,8 @@ func newOutputService9ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice9protocoltest", + ServiceName: "OutputService9ProtocolTest", + ServiceID: "OutputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1491,7 +1500,8 @@ func newOutputService10ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice10protocoltest", + ServiceName: "OutputService10ProtocolTest", + ServiceID: "OutputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1634,7 +1644,8 @@ func newOutputService11ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice11protocoltest", + ServiceName: "OutputService11ProtocolTest", + ServiceID: "OutputService11ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1789,7 +1800,8 @@ func newOutputService12ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice12protocoltest", + ServiceName: "OutputService12ProtocolTest", + ServiceID: "OutputService12ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1932,7 +1944,8 @@ func newOutputService13ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice13protocoltest", + ServiceName: "OutputService13ProtocolTest", + ServiceID: "OutputService13ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2075,7 +2088,8 @@ func newOutputService14ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice14protocoltest", + ServiceName: "OutputService14ProtocolTest", + ServiceID: "OutputService14ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2218,7 +2232,8 @@ func newOutputService15ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice15protocoltest", + ServiceName: "OutputService15ProtocolTest", + ServiceID: "OutputService15ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2361,7 +2376,8 @@ func newOutputService16ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice16protocoltest", + ServiceName: "OutputService16ProtocolTest", + ServiceID: "OutputService16ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2534,7 +2550,7 @@ func TestOutputService1ProtocolTestScalarMembersCase1(t *testing.T) { if e, a := "myname", *out.Str; e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.String(); e != a { + if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } if e, a := true, *out.TrueBool; e != a { diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/build.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/build.go index c405288d7..f761e0b3a 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/build.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/build.go @@ -20,8 +20,10 @@ import ( "github.com/aws/aws-sdk-go/private/protocol" ) -// RFC822 returns an RFC822 formatted timestamp for AWS protocols -const RFC822 = "Mon, 2 Jan 2006 15:04:05 GMT" +// RFC1123GMT is a RFC1123 (RFC822) formated timestame. This format is not +// using the standard library's time.RFC1123 due to the desire to always use +// GMT as the timezone. +const RFC1123GMT = "Mon, 2 Jan 2006 15:04:05 GMT" // Whether the byte value can be sent without escaping in AWS URLs var noEscape [256]bool @@ -270,7 +272,7 @@ func convertType(v reflect.Value, tag reflect.StructTag) (str string, err error) case float64: str = strconv.FormatFloat(value, 'f', -1, 64) case time.Time: - str = value.UTC().Format(RFC822) + str = value.UTC().Format(RFC1123GMT) case aws.JSONValue: if len(value) == 0 { return "", errValueNotSet diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/unmarshal.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/unmarshal.go index 823f045ee..9d4e76267 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/unmarshal.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/rest/unmarshal.go @@ -198,7 +198,7 @@ func unmarshalHeader(v reflect.Value, header string, tag reflect.StructTag) erro } v.Set(reflect.ValueOf(&f)) case *time.Time: - t, err := time.Parse(RFC822, header) + t, err := time.Parse(time.RFC1123, header) if err != nil { return err } diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/build_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/build_test.go index c927fbc98..85e0b0c3b 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/build_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/build_test.go @@ -75,7 +75,8 @@ func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice1protocoltest", + ServiceName: "InputService1ProtocolTest", + ServiceID: "InputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -213,7 +214,8 @@ func newInputService2ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice2protocoltest", + ServiceName: "InputService2ProtocolTest", + ServiceID: "InputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -359,7 +361,8 @@ func newInputService3ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice3protocoltest", + ServiceName: "InputService3ProtocolTest", + ServiceID: "InputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -505,7 +508,8 @@ func newInputService4ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice4protocoltest", + ServiceName: "InputService4ProtocolTest", + ServiceID: "InputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -651,7 +655,8 @@ func newInputService5ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice5protocoltest", + ServiceName: "InputService5ProtocolTest", + ServiceID: "InputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -805,7 +810,8 @@ func newInputService6ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice6protocoltest", + ServiceName: "InputService6ProtocolTest", + ServiceID: "InputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -959,7 +965,8 @@ func newInputService7ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice7protocoltest", + ServiceName: "InputService7ProtocolTest", + ServiceID: "InputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1180,7 +1187,8 @@ func newInputService8ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice8protocoltest", + ServiceName: "InputService8ProtocolTest", + ServiceID: "InputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1342,7 +1350,8 @@ func newInputService9ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice9protocoltest", + ServiceName: "InputService9ProtocolTest", + ServiceID: "InputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1532,7 +1541,8 @@ func newInputService10ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice10protocoltest", + ServiceName: "InputService10ProtocolTest", + ServiceID: "InputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1730,7 +1740,8 @@ func newInputService11ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice11protocoltest", + ServiceName: "InputService11ProtocolTest", + ServiceID: "InputService11ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1906,7 +1917,8 @@ func newInputService12ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice12protocoltest", + ServiceName: "InputService12ProtocolTest", + ServiceID: "InputService12ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2075,7 +2087,8 @@ func newInputService13ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice13protocoltest", + ServiceName: "InputService13ProtocolTest", + ServiceID: "InputService13ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2296,7 +2309,8 @@ func newInputService14ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice14protocoltest", + ServiceName: "InputService14ProtocolTest", + ServiceID: "InputService14ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2529,7 +2543,8 @@ func newInputService15ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice15protocoltest", + ServiceName: "InputService15ProtocolTest", + ServiceID: "InputService15ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2750,7 +2765,8 @@ func newInputService16ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice16protocoltest", + ServiceName: "InputService16ProtocolTest", + ServiceID: "InputService16ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3307,7 +3323,8 @@ func newInputService17ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice17protocoltest", + ServiceName: "InputService17ProtocolTest", + ServiceID: "InputService17ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3536,7 +3553,8 @@ func newInputService18ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice18protocoltest", + ServiceName: "InputService18ProtocolTest", + ServiceID: "InputService18ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3682,7 +3700,8 @@ func newInputService19ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice19protocoltest", + ServiceName: "InputService19ProtocolTest", + ServiceID: "InputService19ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3828,7 +3847,8 @@ func newInputService20ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice20protocoltest", + ServiceName: "InputService20ProtocolTest", + ServiceID: "InputService20ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -4049,7 +4069,8 @@ func newInputService21ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice21protocoltest", + ServiceName: "InputService21ProtocolTest", + ServiceID: "InputService21ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -4381,7 +4402,8 @@ func newInputService22ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice22protocoltest", + ServiceName: "InputService22ProtocolTest", + ServiceID: "InputService22ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/unmarshal_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/unmarshal_test.go index 492ebe635..32c13e882 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/unmarshal_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/restjson/unmarshal_test.go @@ -75,7 +75,8 @@ func newOutputService1ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice1protocoltest", + ServiceName: "OutputService1ProtocolTest", + ServiceID: "OutputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -298,7 +299,8 @@ func newOutputService2ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice2protocoltest", + ServiceName: "OutputService2ProtocolTest", + ServiceID: "OutputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -463,7 +465,8 @@ func newOutputService3ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice3protocoltest", + ServiceName: "OutputService3ProtocolTest", + ServiceID: "OutputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -626,7 +629,8 @@ func newOutputService4ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice4protocoltest", + ServiceName: "OutputService4ProtocolTest", + ServiceID: "OutputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -769,7 +773,8 @@ func newOutputService5ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice5protocoltest", + ServiceName: "OutputService5ProtocolTest", + ServiceID: "OutputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -924,7 +929,8 @@ func newOutputService6ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice6protocoltest", + ServiceName: "OutputService6ProtocolTest", + ServiceID: "OutputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1067,7 +1073,8 @@ func newOutputService7ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice7protocoltest", + ServiceName: "OutputService7ProtocolTest", + ServiceID: "OutputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1210,7 +1217,8 @@ func newOutputService8ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice8protocoltest", + ServiceName: "OutputService8ProtocolTest", + ServiceID: "OutputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1353,7 +1361,8 @@ func newOutputService9ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice9protocoltest", + ServiceName: "OutputService9ProtocolTest", + ServiceID: "OutputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1504,7 +1513,8 @@ func newOutputService10ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice10protocoltest", + ServiceName: "OutputService10ProtocolTest", + ServiceID: "OutputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1667,7 +1677,8 @@ func newOutputService11ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice11protocoltest", + ServiceName: "OutputService11ProtocolTest", + ServiceID: "OutputService11ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1810,7 +1821,8 @@ func newOutputService12ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice12protocoltest", + ServiceName: "OutputService12ProtocolTest", + ServiceID: "OutputService12ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2113,7 +2125,8 @@ func newOutputService13ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice13protocoltest", + ServiceName: "OutputService13ProtocolTest", + ServiceID: "OutputService13ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2442,10 +2455,10 @@ func TestOutputService3ProtocolTestTimestampMembersCase1(t *testing.T) { if out == nil { t.Errorf("expect not to be nil") } - if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.StructMember.Foo.String(); e != a { + if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.StructMember.Foo.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.TimeMember.String(); e != a { + if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.TimeMember.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } @@ -2564,10 +2577,10 @@ func TestOutputService7ProtocolTestComplexMapValuesCase1(t *testing.T) { if out == nil { t.Errorf("expect not to be nil") } - if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.MapMember["a"].String(); e != a { + if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.MapMember["a"].UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.MapMember["b"].String(); e != a { + if e, a := time.Unix(1.398796238e+09, 0).UTC().String(), out.MapMember["b"].UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/build_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/build_test.go index e599edf11..a13c421a7 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/build_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/build_test.go @@ -75,7 +75,8 @@ func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice1protocoltest", + ServiceName: "InputService1ProtocolTest", + ServiceID: "InputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -383,7 +384,8 @@ func newInputService2ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice2protocoltest", + ServiceName: "InputService2ProtocolTest", + ServiceID: "InputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -553,7 +555,8 @@ func newInputService3ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice3protocoltest", + ServiceName: "InputService3ProtocolTest", + ServiceID: "InputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -802,7 +805,8 @@ func newInputService4ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice4protocoltest", + ServiceName: "InputService4ProtocolTest", + ServiceID: "InputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -976,7 +980,8 @@ func newInputService5ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice5protocoltest", + ServiceName: "InputService5ProtocolTest", + ServiceID: "InputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1122,7 +1127,8 @@ func newInputService6ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice6protocoltest", + ServiceName: "InputService6ProtocolTest", + ServiceID: "InputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1268,7 +1274,8 @@ func newInputService7ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice7protocoltest", + ServiceName: "InputService7ProtocolTest", + ServiceID: "InputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1414,7 +1421,8 @@ func newInputService8ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice8protocoltest", + ServiceName: "InputService8ProtocolTest", + ServiceID: "InputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1560,7 +1568,8 @@ func newInputService9ProtocolTestClient(cfg aws.Config, handlers request.Handler Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice9protocoltest", + ServiceName: "InputService9ProtocolTest", + ServiceID: "InputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1718,7 +1727,8 @@ func newInputService10ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice10protocoltest", + ServiceName: "InputService10ProtocolTest", + ServiceID: "InputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1885,7 +1895,8 @@ func newInputService11ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice11protocoltest", + ServiceName: "InputService11ProtocolTest", + ServiceID: "InputService11ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2031,7 +2042,8 @@ func newInputService12ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice12protocoltest", + ServiceName: "InputService12ProtocolTest", + ServiceID: "InputService12ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2177,7 +2189,8 @@ func newInputService13ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice13protocoltest", + ServiceName: "InputService13ProtocolTest", + ServiceID: "InputService13ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2331,7 +2344,8 @@ func newInputService14ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice14protocoltest", + ServiceName: "InputService14ProtocolTest", + ServiceID: "InputService14ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2485,7 +2499,8 @@ func newInputService15ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice15protocoltest", + ServiceName: "InputService15ProtocolTest", + ServiceID: "InputService15ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2706,7 +2721,8 @@ func newInputService16ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice16protocoltest", + ServiceName: "InputService16ProtocolTest", + ServiceID: "InputService16ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2852,7 +2868,8 @@ func newInputService17ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice17protocoltest", + ServiceName: "InputService17ProtocolTest", + ServiceID: "InputService17ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3073,7 +3090,8 @@ func newInputService18ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice18protocoltest", + ServiceName: "InputService18ProtocolTest", + ServiceID: "InputService18ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3456,7 +3474,8 @@ func newInputService19ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice19protocoltest", + ServiceName: "InputService19ProtocolTest", + ServiceID: "InputService19ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3634,7 +3653,8 @@ func newInputService20ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice20protocoltest", + ServiceName: "InputService20ProtocolTest", + ServiceID: "InputService20ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -3788,7 +3808,8 @@ func newInputService21ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice21protocoltest", + ServiceName: "InputService21ProtocolTest", + ServiceID: "InputService21ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -4009,7 +4030,8 @@ func newInputService22ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice22protocoltest", + ServiceName: "InputService22ProtocolTest", + ServiceID: "InputService22ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -4566,7 +4588,8 @@ func newInputService23ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice23protocoltest", + ServiceName: "InputService23ProtocolTest", + ServiceID: "InputService23ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -4712,7 +4735,8 @@ func newInputService24ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice24protocoltest", + ServiceName: "InputService24ProtocolTest", + ServiceID: "InputService24ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -4933,7 +4957,8 @@ func newInputService25ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "inputservice25protocoltest", + ServiceName: "InputService25ProtocolTest", + ServiceID: "InputService25ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/unmarshal_test.go b/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/unmarshal_test.go index 05db75612..1b9f9c5bb 100644 --- a/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/unmarshal_test.go +++ b/vendor/github.com/aws/aws-sdk-go/private/protocol/restxml/unmarshal_test.go @@ -75,7 +75,8 @@ func newOutputService1ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice1protocoltest", + ServiceName: "OutputService1ProtocolTest", + ServiceID: "OutputService1ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -458,7 +459,8 @@ func newOutputService2ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice2protocoltest", + ServiceName: "OutputService2ProtocolTest", + ServiceID: "OutputService2ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -602,7 +604,8 @@ func newOutputService3ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice3protocoltest", + ServiceName: "OutputService3ProtocolTest", + ServiceID: "OutputService3ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -745,7 +748,8 @@ func newOutputService4ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice4protocoltest", + ServiceName: "OutputService4ProtocolTest", + ServiceID: "OutputService4ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -888,7 +892,8 @@ func newOutputService5ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice5protocoltest", + ServiceName: "OutputService5ProtocolTest", + ServiceID: "OutputService5ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1031,7 +1036,8 @@ func newOutputService6ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice6protocoltest", + ServiceName: "OutputService6ProtocolTest", + ServiceID: "OutputService6ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1186,7 +1192,8 @@ func newOutputService7ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice7protocoltest", + ServiceName: "OutputService7ProtocolTest", + ServiceID: "OutputService7ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1329,7 +1336,8 @@ func newOutputService8ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice8protocoltest", + ServiceName: "OutputService8ProtocolTest", + ServiceID: "OutputService8ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1472,7 +1480,8 @@ func newOutputService9ProtocolTestClient(cfg aws.Config, handlers request.Handle Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice9protocoltest", + ServiceName: "OutputService9ProtocolTest", + ServiceID: "OutputService9ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1635,7 +1644,8 @@ func newOutputService10ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice10protocoltest", + ServiceName: "OutputService10ProtocolTest", + ServiceID: "OutputService10ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1778,7 +1788,8 @@ func newOutputService11ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice11protocoltest", + ServiceName: "OutputService11ProtocolTest", + ServiceID: "OutputService11ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -1985,7 +1996,8 @@ func newOutputService12ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice12protocoltest", + ServiceName: "OutputService12ProtocolTest", + ServiceID: "OutputService12ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2128,7 +2140,8 @@ func newOutputService13ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice13protocoltest", + ServiceName: "OutputService13ProtocolTest", + ServiceID: "OutputService13ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2377,7 +2390,8 @@ func newOutputService14ProtocolTestClient(cfg aws.Config, handlers request.Handl Client: client.New( cfg, metadata.ClientInfo{ - ServiceName: "outputservice14protocoltest", + ServiceName: "OutputService14ProtocolTest", + ServiceID: "OutputService14ProtocolTest", SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -2587,7 +2601,7 @@ func TestOutputService1ProtocolTestScalarMembersCase1(t *testing.T) { if e, a := "myname", *out.Str; e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.String(); e != a { + if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } if e, a := true, *out.TrueBool; e != a { @@ -2645,7 +2659,7 @@ func TestOutputService1ProtocolTestScalarMembersCase2(t *testing.T) { if e, a := "", *out.Str; e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.String(); e != a { + if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } if e, a := true, *out.TrueBool; e != a { @@ -2680,10 +2694,10 @@ func TestOutputService1ProtocolTestScalarMembersCase3(t *testing.T) { if e, a := "value2", string(out.Blobs[1]); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamps[0].String(); e != a { + if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamps[0].UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.422172801e+09, 0).UTC().String(), out.Timestamps[1].String(); e != a { + if e, a := time.Unix(1.422172801e+09, 0).UTC().String(), out.Timestamps[1].UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } @@ -2995,7 +3009,7 @@ func TestOutputService11ProtocolTestScalarMembersInHeadersCase1(t *testing.T) { if e, a := "string", *out.Str; e != a { t.Errorf("expect %v, got %v", e, a) } - if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.String(); e != a { + if e, a := time.Unix(1.4221728e+09, 0).UTC().String(), out.Timestamp.UTC().String(); e != a { t.Errorf("expect %v, got %v", e, a) } if e, a := true, *out.TrueBool; e != a { diff --git a/vendor/github.com/aws/aws-sdk-go/service/acm/service.go b/vendor/github.com/aws/aws-sdk-go/service/acm/service.go index b083c37d6..9817d0c0a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/acm/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/acm/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "acm" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "acm" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ACM" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ACM client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/acmpca/acmpcaiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/acmpca/acmpcaiface/interface.go index bdd8fe19b..2bd092d02 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/acmpca/acmpcaiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/acmpca/acmpcaiface/interface.go @@ -108,6 +108,10 @@ type ACMPCAAPI interface { ListTagsWithContext(aws.Context, *acmpca.ListTagsInput, ...request.Option) (*acmpca.ListTagsOutput, error) ListTagsRequest(*acmpca.ListTagsInput) (*request.Request, *acmpca.ListTagsOutput) + RestoreCertificateAuthority(*acmpca.RestoreCertificateAuthorityInput) (*acmpca.RestoreCertificateAuthorityOutput, error) + RestoreCertificateAuthorityWithContext(aws.Context, *acmpca.RestoreCertificateAuthorityInput, ...request.Option) (*acmpca.RestoreCertificateAuthorityOutput, error) + RestoreCertificateAuthorityRequest(*acmpca.RestoreCertificateAuthorityInput) (*request.Request, *acmpca.RestoreCertificateAuthorityOutput) + RevokeCertificate(*acmpca.RevokeCertificateInput) (*acmpca.RevokeCertificateOutput, error) RevokeCertificateWithContext(aws.Context, *acmpca.RevokeCertificateInput, ...request.Option) (*acmpca.RevokeCertificateOutput, error) RevokeCertificateRequest(*acmpca.RevokeCertificateInput) (*request.Request, *acmpca.RevokeCertificateOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/acmpca/api.go b/vendor/github.com/aws/aws-sdk-go/service/acmpca/api.go index c6a2d319b..1987336d5 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/acmpca/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/acmpca/api.go @@ -65,7 +65,7 @@ func (c *ACMPCA) CreateCertificateAuthorityRequest(input *CreateCertificateAutho // (certificate revocation list) configuration specifies the CRL expiration // period in days (the validity period of the CRL), the Amazon S3 bucket that // will contain the CRL, and a CNAME alias for the S3 bucket that is included -// in certificates issued by the CA. If successful, this function returns the +// in certificates issued by the CA. If successful, this operation returns the // Amazon Resource Name (ARN) of the CA. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -155,7 +155,7 @@ func (c *ACMPCA) CreateCertificateAuthorityAuditReportRequest(input *CreateCerti // // Creates an audit report that lists every time that the your CA private key // is used. The report is saved in the Amazon S3 bucket that you specify on -// input. The IssueCertificate and RevokeCertificate functions use the private +// input. The IssueCertificate and RevokeCertificate operations use the private // key. You can generate a new report every 30 minutes. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -253,17 +253,25 @@ func (c *ACMPCA) DeleteCertificateAuthorityRequest(input *DeleteCertificateAutho // DeleteCertificateAuthority API operation for AWS Certificate Manager Private Certificate Authority. // -// Deletes the private certificate authority (CA) that you created or started -// to create by calling the CreateCertificateAuthority function. This action -// requires that you enter an ARN (Amazon Resource Name) for the private CA -// that you want to delete. You can find the ARN by calling the ListCertificateAuthorities -// function. You can delete the CA if you are waiting for it to be created (the -// Status field of the CertificateAuthority is CREATING) or if the CA has been -// created but you haven't yet imported the signed certificate (the Status is -// PENDING_CERTIFICATE) into ACM PCA. If you've already imported the certificate, -// you cannot delete the CA unless it has been disabled for more than 30 days. -// To disable a CA, call the UpdateCertificateAuthority function and set the -// CertificateAuthorityStatus argument to DISABLED. +// Deletes a private certificate authority (CA). You must provide the ARN (Amazon +// Resource Name) of the private CA that you want to delete. You can find the +// ARN by calling the ListCertificateAuthorities operation. Before you can delete +// a CA, you must disable it. Call the UpdateCertificateAuthority operation +// and set the CertificateAuthorityStatus parameter to DISABLED. +// +// Additionally, you can delete a CA if you are waiting for it to be created +// (the Status field of the CertificateAuthority is CREATING). You can also +// delete it if the CA has been created but you haven't yet imported the signed +// certificate (the Status is PENDING_CERTIFICATE) into ACM PCA. +// +// If the CA is in one of the aforementioned states and you call DeleteCertificateAuthority, +// the CA's status changes to DELETED. However, the CA won't be permentantly +// deleted until the restoration period has passed. By default, if you do not +// set the PermanentDeletionTimeInDays parameter, the CA remains restorable +// for 30 days. You can set the parameter from 7 to 30 days. The DescribeCertificateAuthority +// operation returns the time remaining in the restoration window of a Private +// CA in the DELETED state. To restore an eligable CA, call the RestoreCertificateAuthority +// operation. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -356,21 +364,25 @@ func (c *ACMPCA) DescribeCertificateAuthorityRequest(input *DescribeCertificateA // the private CA on input by its ARN (Amazon Resource Name). The output contains // the status of your CA. This can be any of the following: // -// * CREATING: ACM PCA is creating your private certificate authority. +// * CREATING - ACM PCA is creating your private certificate authority. // -// * PENDING_CERTIFICATE: The certificate is pending. You must use your on-premises -// root or subordinate CA to sign your private CA CSR and then import it -// into PCA. +// * PENDING_CERTIFICATE - The certificate is pending. You must use your +// on-premises root or subordinate CA to sign your private CA CSR and then +// import it into PCA. // -// * ACTIVE: Your private CA is active. +// * ACTIVE - Your private CA is active. // -// * DISABLED: Your private CA has been disabled. +// * DISABLED - Your private CA has been disabled. // -// * EXPIRED: Your private CA certificate has expired. +// * EXPIRED - Your private CA certificate has expired. // -// * FAILED: Your private CA has failed. Your CA can fail for problems such -// a network outage or backend AWS failure or other errors. A failed CA can -// never return to the pending state. You must create a new CA. +// * FAILED - Your private CA has failed. Your CA can fail because of problems +// such a network outage or backend AWS failure or other errors. A failed +// CA can never return to the pending state. You must create a new CA. +// +// * DELETED - Your private CA is within the restoration period, after which +// it will be permanently deleted. The length of time remaining in the CA's +// restoration period will also be included in this operation's output. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -454,9 +466,9 @@ func (c *ACMPCA) DescribeCertificateAuthorityAuditReportRequest(input *DescribeC // DescribeCertificateAuthorityAuditReport API operation for AWS Certificate Manager Private Certificate Authority. // // Lists information about a specific audit report created by calling the CreateCertificateAuthorityAuditReport -// function. Audit information is created every time the certificate authority +// operation. Audit information is created every time the certificate authority // (CA) private key is used. The private key is used when you call the IssueCertificate -// function or the RevokeCertificate function. +// operation or the RevokeCertificate operation. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -470,6 +482,9 @@ func (c *ACMPCA) DescribeCertificateAuthorityAuditReportRequest(input *DescribeC // A resource such as a private CA, S3 bucket, certificate, or audit report // cannot be found. // +// * ErrCodeInvalidArnException "InvalidArnException" +// The requested Amazon Resource Name (ARN) does not refer to an existing resource. +// // * ErrCodeInvalidArgsException "InvalidArgsException" // One or more of the specified arguments was not valid. // @@ -540,11 +555,11 @@ func (c *ACMPCA) GetCertificateRequest(input *GetCertificateInput) (req *request // GetCertificate API operation for AWS Certificate Manager Private Certificate Authority. // // Retrieves a certificate from your private CA. The ARN of the certificate -// is returned when you call the IssueCertificate function. You must specify +// is returned when you call the IssueCertificate operation. You must specify // both the ARN of your private CA and the ARN of the issued certificate when -// calling the GetCertificate function. You can retrieve the certificate if +// calling the GetCertificate operation. You can retrieve the certificate if // it is in the ISSUED state. You can call the CreateCertificateAuthorityAuditReport -// function to create a report that contains information about all of the certificates +// operation to create a report that contains information about all of the certificates // issued and revoked by your private CA. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -728,9 +743,9 @@ func (c *ACMPCA) GetCertificateAuthorityCsrRequest(input *GetCertificateAuthorit // // Retrieves the certificate signing request (CSR) for your private certificate // authority (CA). The CSR is created when you call the CreateCertificateAuthority -// function. Take the CSR to your on-premises X.509 infrastructure and sign +// operation. Take the CSR to your on-premises X.509 infrastructure and sign // it by using your root or a subordinate CA. Then import the signed certificate -// back into ACM PCA by calling the ImportCertificateAuthorityCertificate function. +// back into ACM PCA by calling the ImportCertificateAuthorityCertificate operation. // The CSR is returned as a base64 PEM-encoded string. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -754,6 +769,9 @@ func (c *ACMPCA) GetCertificateAuthorityCsrRequest(input *GetCertificateAuthorit // * ErrCodeInvalidArnException "InvalidArnException" // The requested Amazon Resource Name (ARN) does not refer to an existing resource. // +// * ErrCodeInvalidStateException "InvalidStateException" +// The private CA is in a state during which a report cannot be generated. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/acm-pca-2017-08-22/GetCertificateAuthorityCsr func (c *ACMPCA) GetCertificateAuthorityCsr(input *GetCertificateAuthorityCsrInput) (*GetCertificateAuthorityCsrOutput, error) { req, out := c.GetCertificateAuthorityCsrRequest(input) @@ -823,9 +841,9 @@ func (c *ACMPCA) ImportCertificateAuthorityCertificateRequest(input *ImportCerti // ImportCertificateAuthorityCertificate API operation for AWS Certificate Manager Private Certificate Authority. // // Imports your signed private CA certificate into ACM PCA. Before you can call -// this function, you must create the private certificate authority by calling -// the CreateCertificateAuthority function. You must then generate a certificate -// signing request (CSR) by calling the GetCertificateAuthorityCsr function. +// this operation, you must create the private certificate authority by calling +// the CreateCertificateAuthority operation. You must then generate a certificate +// signing request (CSR) by calling the GetCertificateAuthorityCsr operation. // Take the CSR to your on-premises CA and use the root certificate or a subordinate // certificate to sign it. Create a certificate chain and copy the signed certificate // and the certificate chain to your working directory. @@ -864,6 +882,9 @@ func (c *ACMPCA) ImportCertificateAuthorityCertificateRequest(input *ImportCerti // * ErrCodeInvalidArnException "InvalidArnException" // The requested Amazon Resource Name (ARN) does not refer to an existing resource. // +// * ErrCodeInvalidStateException "InvalidStateException" +// The private CA is in a state during which a report cannot be generated. +// // * ErrCodeMalformedCertificateException "MalformedCertificateException" // One or more fields in the certificate are invalid. // @@ -938,11 +959,11 @@ func (c *ACMPCA) IssueCertificateRequest(input *IssueCertificateInput) (req *req // IssueCertificate API operation for AWS Certificate Manager Private Certificate Authority. // // Uses your private certificate authority (CA) to issue a client certificate. -// This function returns the Amazon Resource Name (ARN) of the certificate. -// You can retrieve the certificate by calling the GetCertificate function and -// specifying the ARN. +// This operation returns the Amazon Resource Name (ARN) of the certificate. +// You can retrieve the certificate by calling the GetCertificate operation +// and specifying the ARN. // -// You cannot use the ACM ListCertificateAuthorities function to retrieve the +// You cannot use the ACM ListCertificateAuthorities operation to retrieve the // ARNs of the certificates that you issue by using ACM PCA. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -1040,7 +1061,7 @@ func (c *ACMPCA) ListCertificateAuthoritiesRequest(input *ListCertificateAuthori // ListCertificateAuthorities API operation for AWS Certificate Manager Private Certificate Authority. // // Lists the private certificate authorities that you created by using the CreateCertificateAuthority -// function. +// operation. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1122,8 +1143,8 @@ func (c *ACMPCA) ListTagsRequest(input *ListTagsInput) (req *request.Request, ou // // Lists the tags, if any, that are associated with your private CA. Tags are // labels that you can use to identify and organize your CAs. Each tag consists -// of a key and an optional value. Call the TagCertificateAuthority function -// to add one or more tags to your CA. Call the UntagCertificateAuthority function +// of a key and an optional value. Call the TagCertificateAuthority operation +// to add one or more tags to your CA. Call the UntagCertificateAuthority operation // to remove tags. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -1163,6 +1184,107 @@ func (c *ACMPCA) ListTagsWithContext(ctx aws.Context, input *ListTagsInput, opts return out, req.Send() } +const opRestoreCertificateAuthority = "RestoreCertificateAuthority" + +// RestoreCertificateAuthorityRequest generates a "aws/request.Request" representing the +// client's request for the RestoreCertificateAuthority operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RestoreCertificateAuthority for more information on using the RestoreCertificateAuthority +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RestoreCertificateAuthorityRequest method. +// req, resp := client.RestoreCertificateAuthorityRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/acm-pca-2017-08-22/RestoreCertificateAuthority +func (c *ACMPCA) RestoreCertificateAuthorityRequest(input *RestoreCertificateAuthorityInput) (req *request.Request, output *RestoreCertificateAuthorityOutput) { + op := &request.Operation{ + Name: opRestoreCertificateAuthority, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RestoreCertificateAuthorityInput{} + } + + output = &RestoreCertificateAuthorityOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(jsonrpc.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// RestoreCertificateAuthority API operation for AWS Certificate Manager Private Certificate Authority. +// +// Restores a certificate authority (CA) that is in the DELETED state. You can +// restore a CA during the period that you defined in the PermanentDeletionTimeInDays +// parameter of the DeleteCertificateAuthority operation. Currently, you can +// specify 7 to 30 days. If you did not specify a PermanentDeletionTimeInDays +// value, by default you can restore the CA at any time in a 30 day period. +// You can check the time remaining in the restoration period of a private CA +// in the DELETED state by calling the DescribeCertificateAuthority or ListCertificateAuthorities +// operations. The status of a restored CA is set to its pre-deletion status +// when the RestoreCertificateAuthority operation returns. To change its status +// to ACTIVE, call the UpdateCertificateAuthority operation. If the private +// CA was in the PENDING_CERTIFICATE state at deletion, you must use the ImportCertificateAuthorityCertificate +// operation to import a certificate authority into the private CA before it +// can be activated. You cannot restore a CA after the restoration period has +// ended. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Certificate Manager Private Certificate Authority's +// API operation RestoreCertificateAuthority for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// A resource such as a private CA, S3 bucket, certificate, or audit report +// cannot be found. +// +// * ErrCodeInvalidStateException "InvalidStateException" +// The private CA is in a state during which a report cannot be generated. +// +// * ErrCodeInvalidArnException "InvalidArnException" +// The requested Amazon Resource Name (ARN) does not refer to an existing resource. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/acm-pca-2017-08-22/RestoreCertificateAuthority +func (c *ACMPCA) RestoreCertificateAuthority(input *RestoreCertificateAuthorityInput) (*RestoreCertificateAuthorityOutput, error) { + req, out := c.RestoreCertificateAuthorityRequest(input) + return out, req.Send() +} + +// RestoreCertificateAuthorityWithContext is the same as RestoreCertificateAuthority with the addition of +// the ability to pass a context and additional request options. +// +// See RestoreCertificateAuthority for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *ACMPCA) RestoreCertificateAuthorityWithContext(ctx aws.Context, input *RestoreCertificateAuthorityInput, opts ...request.Option) (*RestoreCertificateAuthorityOutput, error) { + req, out := c.RestoreCertificateAuthorityRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opRevokeCertificate = "RevokeCertificate" // RevokeCertificateRequest generates a "aws/request.Request" representing the @@ -1209,7 +1331,7 @@ func (c *ACMPCA) RevokeCertificateRequest(input *RevokeCertificateInput) (req *r // RevokeCertificate API operation for AWS Certificate Manager Private Certificate Authority. // -// Revokes a certificate that you issued by calling the IssueCertificate function. +// Revokes a certificate that you issued by calling the IssueCertificate operation. // If you enable a certificate revocation list (CRL) when you create or update // your private CA, information about the revoked certificates will be included // in the CRL. ACM PCA writes the CRL to an S3 bucket that you specify. For @@ -1322,8 +1444,8 @@ func (c *ACMPCA) TagCertificateAuthorityRequest(input *TagCertificateAuthorityIn // a tag to just one private CA if you want to identify a specific characteristic // of that CA, or you can apply the same tag to multiple private CAs if you // want to filter for a common relationship among those CAs. To remove one or -// more tags, use the UntagCertificateAuthority function. Call the ListTags -// function to see what tags are associated with your CA. +// more tags, use the UntagCertificateAuthority operation. Call the ListTags +// operation to see what tags are associated with your CA. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1340,6 +1462,9 @@ func (c *ACMPCA) TagCertificateAuthorityRequest(input *TagCertificateAuthorityIn // * ErrCodeInvalidArnException "InvalidArnException" // The requested Amazon Resource Name (ARN) does not refer to an existing resource. // +// * ErrCodeInvalidStateException "InvalidStateException" +// The private CA is in a state during which a report cannot be generated. +// // * ErrCodeInvalidTagException "InvalidTagException" // The tag associated with the CA is not valid. The invalid argument is contained // in the message field. @@ -1418,10 +1543,10 @@ func (c *ACMPCA) UntagCertificateAuthorityRequest(input *UntagCertificateAuthori // // Remove one or more tags from your private CA. A tag consists of a key-value // pair. If you do not specify the value portion of the tag when calling this -// function, the tag will be removed regardless of value. If you specify a value, -// the tag is removed only if it is associated with the specified value. To -// add tags to a private CA, use the TagCertificateAuthority. Call the ListTags -// function to see what tags are associated with your CA. +// operation, the tag will be removed regardless of value. If you specify a +// value, the tag is removed only if it is associated with the specified value. +// To add tags to a private CA, use the TagCertificateAuthority. Call the ListTags +// operation to see what tags are associated with your CA. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1438,6 +1563,9 @@ func (c *ACMPCA) UntagCertificateAuthorityRequest(input *UntagCertificateAuthori // * ErrCodeInvalidArnException "InvalidArnException" // The requested Amazon Resource Name (ARN) does not refer to an existing resource. // +// * ErrCodeInvalidStateException "InvalidStateException" +// The private CA is in a state during which a report cannot be generated. +// // * ErrCodeInvalidTagException "InvalidTagException" // The tag associated with the CA is not valid. The invalid argument is contained // in the message field. @@ -1572,15 +1700,15 @@ func (c *ACMPCA) UpdateCertificateAuthorityWithContext(ctx aws.Context, input *U // a user, computer, device, or service. The Subject must contain an X.500 distinguished // name (DN). A DN is a sequence of relative distinguished names (RDNs). The // RDNs are separated by commas in the certificate. The DN must be unique for -// each for each entity, but your private CA can issue more than one certificate -// with the same DN to the same entity. +// each entity, but your private CA can issue more than one certificate with +// the same DN to the same entity. type ASN1Subject struct { _ struct{} `type:"structure"` // Fully qualified domain name (FQDN) associated with the certificate subject. CommonName *string `type:"string"` - // Two digit code that specifies the country in which the certificate subject + // Two-digit code that specifies the country in which the certificate subject // located. Country *string `type:"string"` @@ -1621,11 +1749,11 @@ type ASN1Subject struct { // State in which the subject of the certificate is located. State *string `type:"string"` - // Family name. In the US and the UK for example, the surname of an individual + // Family name. In the US and the UK, for example, the surname of an individual // is ordered last. In Asian cultures the surname is typically ordered first. Surname *string `type:"string"` - // A title such as Mr. or Ms. which is pre-pended to the name to refer formally + // A title such as Mr. or Ms., which is pre-pended to the name to refer formally // to the certificate subject. Title *string `type:"string"` } @@ -1728,11 +1856,11 @@ func (s *ASN1Subject) SetTitle(v string) *ASN1Subject { // private CA can issue and revoke X.509 digital certificates. Digital certificates // verify that the entity named in the certificate Subject field owns or controls // the public key contained in the Subject Public Key Info field. Call the CreateCertificateAuthority -// function to create your private CA. You must then call the GetCertificateAuthorityCertificate -// function to retrieve a private CA certificate signing request (CSR). Take +// operation to create your private CA. You must then call the GetCertificateAuthorityCertificate +// operation to retrieve a private CA certificate signing request (CSR). Take // the CSR to your on-premises CA and sign it with the root CA certificate or // a subordinate certificate. Call the ImportCertificateAuthorityCertificate -// function to import the signed certificate into AWS Certificate Manager (ACM). +// operation to import the signed certificate into AWS Certificate Manager (ACM). type CertificateAuthority struct { _ struct{} `type:"structure"` @@ -1758,6 +1886,11 @@ type CertificateAuthority struct { // Date and time before which your private CA certificate is not valid. NotBefore *time.Time `type:"timestamp" timestampFormat:"unix"` + // The period during which a deleted CA can be restored. For more information, + // see the PermanentDeletionTimeInDays parameter of the DeleteCertificateAuthorityRequest + // operation. + RestorableUntil *time.Time `type:"timestamp" timestampFormat:"unix"` + // Information about the certificate revocation list (CRL) created and maintained // by your private CA. RevocationConfiguration *RevocationConfiguration `type:"structure"` @@ -1824,6 +1957,12 @@ func (s *CertificateAuthority) SetNotBefore(v time.Time) *CertificateAuthority { return s } +// SetRestorableUntil sets the RestorableUntil field's value. +func (s *CertificateAuthority) SetRestorableUntil(v time.Time) *CertificateAuthority { + s.RestorableUntil = &v + return s +} + // SetRevocationConfiguration sets the RevocationConfiguration field's value. func (s *CertificateAuthority) SetRevocationConfiguration(v *RevocationConfiguration) *CertificateAuthority { s.RevocationConfiguration = v @@ -1853,7 +1992,7 @@ func (s *CertificateAuthority) SetType(v string) *CertificateAuthority { // the key pair that your private CA creates when it issues a certificate, the // signature algorithm it uses used when issuing certificates, and its X.500 // distinguished name. You must specify this information when you call the CreateCertificateAuthority -// function. +// operation. type CertificateAuthorityConfiguration struct { _ struct{} `type:"structure"` @@ -1932,7 +2071,7 @@ type CreateCertificateAuthorityAuditReportInput struct { // Amazon Resource Name (ARN) of the CA to be audited. This is of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2042,9 +2181,9 @@ type CreateCertificateAuthorityInput struct { // Alphanumeric string that can be used to distinguish between calls to CreateCertificateAuthority. // Idempotency tokens time out after five minutes. Therefore, if you call CreateCertificateAuthority // multiple times with the same idempotency token within a five minute period, - // ACM PCA recognizes that you are requesting only one certificate and will - // issue only one. If you change the idempotency token for each call, however, - // ACM PCA recognizes that you are requesting multiple certificates. + // ACM PCA recognizes that you are requesting only one certificate. As a result, + // ACM PCA issues only one. If you change the idempotency token for each call, + // however, ACM PCA recognizes that you are requesting multiple certificates. IdempotencyToken *string `min:"1" type:"string"` // Contains a Boolean value that you can use to enable a certification revocation @@ -2124,7 +2263,7 @@ type CreateCertificateAuthorityOutput struct { // If successful, the Amazon Resource Name (ARN) of the certificate authority // (CA). This is of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. CertificateAuthorityArn *string `min:"5" type:"string"` } @@ -2211,8 +2350,8 @@ type CrlConfiguration struct { // Boolean value that specifies whether certificate revocation lists (CRLs) // are enabled. You can use this value to enable certificate revocation for - // a new CA when you call the CreateCertificateAuthority function or for an - // existing CA when you call the UpdateCertificateAuthority function. + // a new CA when you call the CreateCertificateAuthority operation or for an + // existing CA when you call the UpdateCertificateAuthority operation. // // Enabled is a required field Enabled *bool `type:"boolean" required:"true"` @@ -2223,7 +2362,7 @@ type CrlConfiguration struct { // Name of the S3 bucket that contains the CRL. If you do not provide a value // for the CustomCname argument, the name of your S3 bucket is placed into the // CRL Distribution Points extension of the issued certificate. You can change - // the name of your bucket by calling the UpdateCertificateAuthority function. + // the name of your bucket by calling the UpdateCertificateAuthority operation. // You must specify a bucket policy that allows ACM PCA to write the CRL to // your bucket. S3BucketName *string `min:"3" type:"string"` @@ -2286,12 +2425,16 @@ type DeleteCertificateAuthorityInput struct { _ struct{} `type:"structure"` // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. - // This must be of the form: + // This must have the following form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` + + // The number of days to make a CA restorable after it has been deleted. This + // can be anywhere from 7 to 30 days, with 30 being the default. + PermanentDeletionTimeInDays *int64 `min:"7" type:"integer"` } // String returns the string representation @@ -2313,6 +2456,9 @@ func (s *DeleteCertificateAuthorityInput) Validate() error { if s.CertificateAuthorityArn != nil && len(*s.CertificateAuthorityArn) < 5 { invalidParams.Add(request.NewErrParamMinLen("CertificateAuthorityArn", 5)) } + if s.PermanentDeletionTimeInDays != nil && *s.PermanentDeletionTimeInDays < 7 { + invalidParams.Add(request.NewErrParamMinValue("PermanentDeletionTimeInDays", 7)) + } if invalidParams.Len() > 0 { return invalidParams @@ -2326,6 +2472,12 @@ func (s *DeleteCertificateAuthorityInput) SetCertificateAuthorityArn(v string) * return s } +// SetPermanentDeletionTimeInDays sets the PermanentDeletionTimeInDays field's value. +func (s *DeleteCertificateAuthorityInput) SetPermanentDeletionTimeInDays(v int64) *DeleteCertificateAuthorityInput { + s.PermanentDeletionTimeInDays = &v + return s +} + type DeleteCertificateAuthorityOutput struct { _ struct{} `type:"structure"` } @@ -2344,14 +2496,14 @@ type DescribeCertificateAuthorityAuditReportInput struct { _ struct{} `type:"structure"` // The report ID returned by calling the CreateCertificateAuthorityAuditReport - // function. + // operation. // // AuditReportId is a required field AuditReportId *string `min:"36" type:"string" required:"true"` // The Amazon Resource Name (ARN) of the private CA. This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2457,7 +2609,7 @@ type DescribeCertificateAuthorityInput struct { // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. // This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2524,7 +2676,7 @@ type GetCertificateAuthorityCertificateInput struct { // The Amazon Resource Name (ARN) of your private CA. This is of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2600,9 +2752,9 @@ type GetCertificateAuthorityCsrInput struct { _ struct{} `type:"structure"` // The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority - // function. This must be of the form: + // operation. This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2670,7 +2822,7 @@ type GetCertificateInput struct { // The ARN of the issued certificate. The ARN contains the certificate serial // number and must be in the following form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245 // // CertificateArn is a required field CertificateArn *string `min:"5" type:"string" required:"true"` @@ -2678,7 +2830,7 @@ type GetCertificateInput struct { // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. // This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012. // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2775,7 +2927,7 @@ type ImportCertificateAuthorityCertificateInput struct { // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. // This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2864,7 +3016,7 @@ type IssueCertificateInput struct { // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. // This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -2889,9 +3041,9 @@ type IssueCertificateInput struct { Csr []byte `min:"1" type:"blob" required:"true"` // Custom string that can be used to distinguish between calls to the IssueCertificate - // function. Idempotency tokens time out after one hour. Therefore, if you call - // IssueCertificate multiple times with the same idempotency token within 5 - // minutes, ACM PCA recognizes that you are requesting only one certificate + // operation. Idempotency tokens time out after one hour. Therefore, if you + // call IssueCertificate multiple times with the same idempotency token within + // 5 minutes, ACM PCA recognizes that you are requesting only one certificate // and will issue only one. If you change the idempotency token for each call, // PCA recognizes that you are requesting multiple certificates. IdempotencyToken *string `min:"1" type:"string"` @@ -2990,7 +3142,7 @@ type IssueCertificateOutput struct { // The Amazon Resource Name (ARN) of the issued certificate and the certificate // serial number. This is of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012/certificate/286535153982981100925020015808220737245 CertificateArn *string `min:"5" type:"string"` } @@ -3100,9 +3252,9 @@ type ListTagsInput struct { _ struct{} `type:"structure"` // The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority - // function. This must be of the form: + // operation. This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -3202,8 +3354,66 @@ func (s *ListTagsOutput) SetTags(v []*Tag) *ListTagsOutput { return s } +type RestoreCertificateAuthorityInput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) that was returned when you called the CreateCertificateAuthority + // operation. This must be of the form: + // + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // + // CertificateAuthorityArn is a required field + CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` +} + +// String returns the string representation +func (s RestoreCertificateAuthorityInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RestoreCertificateAuthorityInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RestoreCertificateAuthorityInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RestoreCertificateAuthorityInput"} + if s.CertificateAuthorityArn == nil { + invalidParams.Add(request.NewErrParamRequired("CertificateAuthorityArn")) + } + if s.CertificateAuthorityArn != nil && len(*s.CertificateAuthorityArn) < 5 { + invalidParams.Add(request.NewErrParamMinLen("CertificateAuthorityArn", 5)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCertificateAuthorityArn sets the CertificateAuthorityArn field's value. +func (s *RestoreCertificateAuthorityInput) SetCertificateAuthorityArn(v string) *RestoreCertificateAuthorityInput { + s.CertificateAuthorityArn = &v + return s +} + +type RestoreCertificateAuthorityOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s RestoreCertificateAuthorityOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RestoreCertificateAuthorityOutput) GoString() string { + return s.String() +} + // Certificate revocation information used by the CreateCertificateAuthority -// and UpdateCertificateAuthority functions. Your private certificate authority +// and UpdateCertificateAuthority operations. Your private certificate authority // (CA) can create and maintain a certificate revocation list (CRL). A CRL contains // information about certificates revoked by your CA. For more information, // see RevokeCertificate. @@ -3252,7 +3462,7 @@ type RevokeCertificateInput struct { // Amazon Resource Name (ARN) of the private CA that issued the certificate // to be revoked. This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -3260,15 +3470,15 @@ type RevokeCertificateInput struct { // Serial number of the certificate to be revoked. This must be in hexadecimal // format. You can retrieve the serial number by calling GetCertificate with // the Amazon Resource Name (ARN) of the certificate you want and the ARN of - // your private CA. The GetCertificate function retrieves the certificate in + // your private CA. The GetCertificate operation retrieves the certificate in // the PEM format. You can use the following OpenSSL command to list the certificate // in text format and copy the hexadecimal serial number. // // openssl x509 -in file_path -text -noout // // You can also copy the serial number from the console or use the DescribeCertificate - // (http://docs.aws.amazon.comacm/latest/APIReferenceAPI_DescribeCertificate.html) - // function in the AWS Certificate Manager API Reference. + // (https://docs.aws.amazon.com/acm/latest/APIReference/API_DescribeCertificate.html) + // operation in the AWS Certificate Manager API Reference. // // CertificateSerial is a required field CertificateSerial *string `type:"string" required:"true"` @@ -3346,8 +3556,8 @@ func (s RevokeCertificateOutput) GoString() string { // Tags are labels that you can use to identify and organize your private CAs. // Each tag consists of a key and an optional value. You can associate up to // 50 tags with a private CA. To add one or more tags to a private CA, call -// the TagCertificateAuthority function. To remove a tag, call the UntagCertificateAuthority -// function. +// the TagCertificateAuthority operation. To remove a tag, call the UntagCertificateAuthority +// operation. type Tag struct { _ struct{} `type:"structure"` @@ -3404,7 +3614,7 @@ type TagCertificateAuthorityInput struct { // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. // This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -3489,7 +3699,7 @@ type UntagCertificateAuthorityInput struct { // The Amazon Resource Name (ARN) that was returned when you called CreateCertificateAuthority. // This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -3574,7 +3784,7 @@ type UpdateCertificateAuthorityInput struct { // Amazon Resource Name (ARN) of the private CA that issued the certificate // to be revoked. This must be of the form: // - // arn:aws:acm:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 + // arn:aws:acm-pca:region:account:certificate-authority/12345678-1234-1234-1234-123456789012 // // CertificateAuthorityArn is a required field CertificateAuthorityArn *string `min:"5" type:"string" required:"true"` @@ -3651,7 +3861,7 @@ func (s UpdateCertificateAuthorityOutput) GoString() string { // Length of time for which the certificate issued by your private certificate // authority (CA), or by the private CA itself, is valid in days, months, or -// years. You can issue a certificate by calling the IssueCertificate function. +// years. You can issue a certificate by calling the IssueCertificate operation. type Validity struct { _ struct{} `type:"structure"` @@ -3736,6 +3946,9 @@ const ( // CertificateAuthorityStatusActive is a CertificateAuthorityStatus enum value CertificateAuthorityStatusActive = "ACTIVE" + // CertificateAuthorityStatusDeleted is a CertificateAuthorityStatus enum value + CertificateAuthorityStatusDeleted = "DELETED" + // CertificateAuthorityStatusDisabled is a CertificateAuthorityStatus enum value CertificateAuthorityStatusDisabled = "DISABLED" diff --git a/vendor/github.com/aws/aws-sdk-go/service/acmpca/doc.go b/vendor/github.com/aws/aws-sdk-go/service/acmpca/doc.go index f20c964ff..3ca437764 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/acmpca/doc.go +++ b/vendor/github.com/aws/aws-sdk-go/service/acmpca/doc.go @@ -4,9 +4,9 @@ // requests to AWS Certificate Manager Private Certificate Authority. // // You can use the ACM PCA API to create a private certificate authority (CA). -// You must first call the CreateCertificateAuthority function. If successful, -// the function returns an Amazon Resource Name (ARN) for your private CA. Use -// this ARN as input to the GetCertificateAuthorityCsr function to retrieve +// You must first call the CreateCertificateAuthority operation. If successful, +// the operation returns an Amazon Resource Name (ARN) for your private CA. +// Use this ARN as input to the GetCertificateAuthorityCsr operation to retrieve // the certificate signing request (CSR) for your private CA certificate. Sign // the CSR using the root or an intermediate CA in your on-premises PKI hierarchy, // and call the ImportCertificateAuthorityCertificate to import your signed @@ -15,8 +15,8 @@ // Use your private CA to issue and revoke certificates. These are private certificates // that identify and secure client computers, servers, applications, services, // devices, and users over SSLS/TLS connections within your organization. Call -// the IssueCertificate function to issue a certificate. Call the RevokeCertificate -// function to revoke a certificate. +// the IssueCertificate operation to issue a certificate. Call the RevokeCertificate +// operation to revoke a certificate. // // Certificates issued by your private CA can be trusted only within your organization, // not publicly. @@ -24,13 +24,13 @@ // Your private CA can optionally create a certificate revocation list (CRL) // to track the certificates you revoke. To create a CRL, you must specify a // RevocationConfiguration object when you call the CreateCertificateAuthority -// function. ACM PCA writes the CRL to an S3 bucket that you specify. You must +// operation. ACM PCA writes the CRL to an S3 bucket that you specify. You must // specify a bucket policy that grants ACM PCA write permission. // // You can also call the CreateCertificateAuthorityAuditReport to create an // optional audit report that lists every time the CA private key is used. The // private key is used for signing when the IssueCertificate or RevokeCertificate -// function is called. +// operation is called. // // See https://docs.aws.amazon.com/goto/WebAPI/acm-pca-2017-08-22 for more information on this service. // diff --git a/vendor/github.com/aws/aws-sdk-go/service/acmpca/service.go b/vendor/github.com/aws/aws-sdk-go/service/acmpca/service.go index 2040a19f6..6c231c1d7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/acmpca/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/acmpca/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "acm-pca" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "acm-pca" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ACM PCA" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ACMPCA client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/api.go b/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/api.go index 215077f1e..4241c640a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/api.go @@ -152,6 +152,8 @@ func (c *AlexaForBusiness) AssociateDeviceWithRoomRequest(input *AssociateDevice // You are performing an action that would put you beyond your account's limits. // HTTP Status Code: 400 // +// * ErrCodeDeviceNotRegisteredException "DeviceNotRegisteredException" +// // See also, https://docs.aws.amazon.com/goto/WebAPI/alexaforbusiness-2017-11-09/AssociateDeviceWithRoom func (c *AlexaForBusiness) AssociateDeviceWithRoom(input *AssociateDeviceWithRoomInput) (*AssociateDeviceWithRoomOutput, error) { req, out := c.AssociateDeviceWithRoomRequest(input) @@ -1423,6 +1425,10 @@ func (c *AlexaForBusiness) DisassociateDeviceFromRoomRequest(input *Disassociate // // See the AWS API reference guide for Alexa For Business's // API operation DisassociateDeviceFromRoom for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDeviceNotRegisteredException "DeviceNotRegisteredException" +// // See also, https://docs.aws.amazon.com/goto/WebAPI/alexaforbusiness-2017-11-09/DisassociateDeviceFromRoom func (c *AlexaForBusiness) DisassociateDeviceFromRoom(input *DisassociateDeviceFromRoomInput) (*DisassociateDeviceFromRoomOutput, error) { req, out := c.DisassociateDeviceFromRoomRequest(input) @@ -3759,6 +3765,10 @@ func (c *AlexaForBusiness) StartDeviceSyncRequest(input *StartDeviceSyncInput) ( // // See the AWS API reference guide for Alexa For Business's // API operation StartDeviceSync for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDeviceNotRegisteredException "DeviceNotRegisteredException" +// // See also, https://docs.aws.amazon.com/goto/WebAPI/alexaforbusiness-2017-11-09/StartDeviceSync func (c *AlexaForBusiness) StartDeviceSync(input *StartDeviceSyncInput) (*StartDeviceSyncOutput, error) { req, out := c.StartDeviceSyncRequest(input) @@ -4157,6 +4167,8 @@ func (c *AlexaForBusiness) UpdateDeviceRequest(input *UpdateDeviceInput) (req *r // * ErrCodeNotFoundException "NotFoundException" // The resource is not found. HTTP Status Code: 400 // +// * ErrCodeDeviceNotRegisteredException "DeviceNotRegisteredException" +// // See also, https://docs.aws.amazon.com/goto/WebAPI/alexaforbusiness-2017-11-09/UpdateDevice func (c *AlexaForBusiness) UpdateDevice(input *UpdateDeviceInput) (*UpdateDeviceOutput, error) { req, out := c.UpdateDeviceRequest(input) diff --git a/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/errors.go b/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/errors.go index c2ad81011..618a64a5d 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/errors.go @@ -10,6 +10,10 @@ const ( // The resource being created already exists. HTTP Status Code: 400 ErrCodeAlreadyExistsException = "AlreadyExistsException" + // ErrCodeDeviceNotRegisteredException for service response error code + // "DeviceNotRegisteredException". + ErrCodeDeviceNotRegisteredException = "DeviceNotRegisteredException" + // ErrCodeInvalidUserStatusException for service response error code // "InvalidUserStatusException". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/service.go b/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/service.go index f2aaf6887..1368edcfa 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/alexaforbusiness/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "a4b" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "a4b" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Alexa For Business" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the AlexaForBusiness client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/apigateway/api.go b/vendor/github.com/aws/aws-sdk-go/service/apigateway/api.go index 95cf464ad..d1804357c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/apigateway/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/apigateway/api.go @@ -15203,7 +15203,7 @@ type EndpointConfiguration struct { // A list of endpoint types of an API (RestApi) or its custom domain name (DomainName). // For an edge-optimized API and its custom domain name, the endpoint type is // "EDGE". For a regional API and its custom domain name, the endpoint type - // is REGIONAL. + // is REGIONAL. For a private API, the endpoint type is PRIVATE. Types []*string `locationName:"types" type:"list"` } @@ -16577,11 +16577,11 @@ type GetExportInput struct { // A key-value map of query string parameters that specify properties of the // export, depending on the requested exportType. For exportTypeswagger, any - // combination of the following parameters are supported: integrations will - // export the API with x-amazon-apigateway-integration extensions. authorizers - // will export the API with x-amazon-apigateway-authorizer extensions. postman - // will export the API with Postman extensions, allowing for import to the Postman - // tool + // combination of the following parameters are supported: extensions='integrations' + // or extensions='apigateway' will export the API with x-amazon-apigateway-integration + // extensions. extensions='authorizers' will export the API with x-amazon-apigateway-authorizer + // extensions. postman will export the API with Postman extensions, allowing + // for import to the Postman tool Parameters map[string]*string `location:"querystring" locationName:"parameters" type:"map"` // [Required] The string identifier of the associated RestApi. @@ -19141,8 +19141,9 @@ type ImportRestApiInput struct { // // To exclude DocumentationParts from the import, set parameters as ignore=documentation. // - // To configure the endpoint type, set parameters as endpointConfigurationTypes=EDGE - // orendpointConfigurationTypes=REGIONAL. The default endpoint type is EDGE. + // To configure the endpoint type, set parameters as endpointConfigurationTypes=EDGE, + // endpointConfigurationTypes=REGIONAL, or endpointConfigurationTypes=PRIVATE. + // The default endpoint type is EDGE. // // To handle imported basePath, set parameters as basePath=ignore, basePath=prepend // or basePath=split. @@ -19151,11 +19152,11 @@ type ImportRestApiInput struct { // API is: // // aws apigateway import-rest-api --parameters ignore=documentation --body - // 'file:///path/to/imported-api-body.json + // 'file:///path/to/imported-api-body.json' // The AWS CLI command to set the regional endpoint on the imported API is: // // aws apigateway import-rest-api --parameters endpointConfigurationTypes=REGIONAL - // --body 'file:///path/to/imported-api-body.json + // --body 'file:///path/to/imported-api-body.json' Parameters map[string]*string `location:"querystring" locationName:"parameters" type:"map"` } @@ -20996,7 +20997,7 @@ type PutRestApiInput struct { // Custom header parameters as part of the request. For example, to exclude // DocumentationParts from an imported API, set ignore=documentation as a parameters // value, as in the AWS CLI command of aws apigateway import-rest-api --parameters - // ignore=documentation --body 'file:///path/to/imported-api-body.json. + // ignore=documentation --body 'file:///path/to/imported-api-body.json'. Parameters map[string]*string `location:"querystring" locationName:"parameters" type:"map"` // [Required] The string identifier of the associated RestApi. @@ -24162,15 +24163,19 @@ const ( DocumentationPartTypeResponseBody = "RESPONSE_BODY" ) -// The endpoint type. The valid value is EDGE for edge-optimized API setup, -// most suitable for mobile applications, REGIONAL for regional API endpoint -// setup, most suitable for calling from AWS Region +// The endpoint type. The valid values are EDGE for edge-optimized API setup, +// most suitable for mobile applications; REGIONAL for regional API endpoint +// setup, most suitable for calling from AWS Region; and PRIVATE for private +// APIs. const ( // EndpointTypeRegional is a EndpointType enum value EndpointTypeRegional = "REGIONAL" // EndpointTypeEdge is a EndpointType enum value EndpointTypeEdge = "EDGE" + + // EndpointTypePrivate is a EndpointType enum value + EndpointTypePrivate = "PRIVATE" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/apigateway/service.go b/vendor/github.com/aws/aws-sdk-go/service/apigateway/service.go index 690b6b2f3..8064d24fc 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/apigateway/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/apigateway/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "apigateway" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "apigateway" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "API Gateway" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the APIGateway client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/applicationautoscaling/service.go b/vendor/github.com/aws/aws-sdk-go/service/applicationautoscaling/service.go index 909c1b80e..902d81d42 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/applicationautoscaling/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/applicationautoscaling/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "autoscaling" // Service endpoint prefix API calls made to. - EndpointsID = "application-autoscaling" // Service ID for Regions and Endpoints metadata. + ServiceName = "autoscaling" // Name of service. + EndpointsID = "application-autoscaling" // ID to lookup a service endpoint with. + ServiceID = "Application Auto Scaling" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ApplicationAutoScaling client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/applicationdiscoveryservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/applicationdiscoveryservice/service.go index 00f66dd77..11c2aa643 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/applicationdiscoveryservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/applicationdiscoveryservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "discovery" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "discovery" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Application Discovery Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ApplicationDiscoveryService client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/appstream/api.go b/vendor/github.com/aws/aws-sdk-go/service/appstream/api.go index 0c913939b..baad3f76e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/appstream/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/appstream/api.go @@ -246,7 +246,9 @@ func (c *AppStream) CreateDirectoryConfigRequest(input *CreateDirectoryConfigInp // CreateDirectoryConfig API operation for Amazon AppStream. // -// Creates a directory configuration. +// Creates a Directory Config object in AppStream 2.0. This object includes +// the information required to join streaming instances to an Active Directory +// domain. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -332,7 +334,8 @@ func (c *AppStream) CreateFleetRequest(input *CreateFleetInput) (req *request.Re // CreateFleet API operation for Amazon AppStream. // -// Creates a fleet. +// Creates a fleet. A fleet consists of streaming instances that run a specified +// image. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -436,7 +439,8 @@ func (c *AppStream) CreateImageBuilderRequest(input *CreateImageBuilderInput) (r // CreateImageBuilder API operation for Amazon AppStream. // -// Creates an image builder. +// Creates an image builder. An image builder is a virtual machine that is used +// to create an image. // // The initial state of the builder is PENDING. When it is ready, the state // is RUNNING. @@ -625,7 +629,8 @@ func (c *AppStream) CreateStackRequest(input *CreateStackInput) (req *request.Re // CreateStack API operation for Amazon AppStream. // -// Creates a stack. +// Creates a stack to start streaming applications to users. A stack consists +// of an associated fleet, user access policies, and storage configurations. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -723,7 +728,9 @@ func (c *AppStream) CreateStreamingURLRequest(input *CreateStreamingURLInput) (r // CreateStreamingURL API operation for Amazon AppStream. // -// Creates a URL to start a streaming session for the specified user. +// Creates a temporary URL to start an AppStream 2.0 streaming session for the +// specified user. A streaming URL enables application streaming to be tested +// without user setup. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -811,7 +818,9 @@ func (c *AppStream) DeleteDirectoryConfigRequest(input *DeleteDirectoryConfigInp // DeleteDirectoryConfig API operation for Amazon AppStream. // -// Deletes the specified directory configuration. +// Deletes the specified Directory Config object from AppStream 2.0. This object +// includes the information required to join streaming instances to an Active +// Directory domain. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -978,9 +987,8 @@ func (c *AppStream) DeleteImageRequest(input *DeleteImageInput) (req *request.Re // DeleteImage API operation for Amazon AppStream. // -// Deletes the specified image. You cannot delete an image that is currently -// in use. After you delete an image, you cannot provision new capacity using -// the image. +// Deletes the specified image. You cannot delete an image when it is in use. +// After you delete an image, you cannot provision new capacity using the image. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1153,8 +1161,10 @@ func (c *AppStream) DeleteStackRequest(input *DeleteStackInput) (req *request.Re // DeleteStack API operation for Amazon AppStream. // -// Deletes the specified stack. After this operation completes, the environment -// can no longer be activated and any reservations made for the stack are released. +// Deletes the specified stack. After the stack is deleted, the application +// streaming environment provided by the stack is no longer available to users. +// Also, any reservations made for application streaming sessions for the stack +// are released. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1239,9 +1249,14 @@ func (c *AppStream) DescribeDirectoryConfigsRequest(input *DescribeDirectoryConf // DescribeDirectoryConfigs API operation for Amazon AppStream. // -// Describes the specified directory configurations. Note that although the -// response syntax in this topic includes the account password, this password -// is not returned in the actual response. +// Retrieves a list that describes one or more specified Directory Config objects +// for AppStream 2.0, if the names for these objects are provided. Otherwise, +// all Directory Config objects in the account are described. These objects +// include the information required to join streaming instances to an Active +// Directory domain. +// +// Although the response syntax in this topic includes the account password, +// this password is not returned in the actual response. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1320,7 +1335,8 @@ func (c *AppStream) DescribeFleetsRequest(input *DescribeFleetsInput) (req *requ // DescribeFleets API operation for Amazon AppStream. // -// Describes the specified fleets or all fleets in the account. +// Retrieves a list that describes one or more specified fleets, if the fleet +// names are provided. Otherwise, all fleets in the account are described. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1399,7 +1415,9 @@ func (c *AppStream) DescribeImageBuildersRequest(input *DescribeImageBuildersInp // DescribeImageBuilders API operation for Amazon AppStream. // -// Describes the specified image builders or all image builders in the account. +// Retrieves a list that describes one or more specified image builders, if +// the image builder names are provided. Otherwise, all image builders in the +// account are described. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1478,7 +1496,8 @@ func (c *AppStream) DescribeImagesRequest(input *DescribeImagesInput) (req *requ // DescribeImages API operation for Amazon AppStream. // -// Describes the specified images or all images in the account. +// Retrieves a list that describes one or more specified images, if the image +// names are provided. Otherwise, all images in the account are described. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1557,10 +1576,10 @@ func (c *AppStream) DescribeSessionsRequest(input *DescribeSessionsInput) (req * // DescribeSessions API operation for Amazon AppStream. // -// Describes the streaming sessions for the specified stack and fleet. If a -// user ID is provided, only the streaming sessions for only that user are returned. -// If an authentication type is not provided, the default is to authenticate -// users using a streaming URL. +// Retrieves a list that describes the streaming sessions for a specified stack +// and fleet. If a user ID is provided for the stack and fleet, only streaming +// sessions for that user are described. If an authentication type is not provided, +// the default is to authenticate users using a streaming URL. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1639,7 +1658,8 @@ func (c *AppStream) DescribeStacksRequest(input *DescribeStacksInput) (req *requ // DescribeStacks API operation for Amazon AppStream. // -// Describes the specified stacks or all stacks in the account. +// Retrieves a list that describes one or more specified stacks, if the stack +// names are provided. Otherwise, all stacks in the account are described. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1803,7 +1823,7 @@ func (c *AppStream) ExpireSessionRequest(input *ExpireSessionInput) (req *reques // ExpireSession API operation for Amazon AppStream. // -// Stops the specified streaming session. +// Immediately stops the specified streaming session. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1877,7 +1897,7 @@ func (c *AppStream) ListAssociatedFleetsRequest(input *ListAssociatedFleetsInput // ListAssociatedFleets API operation for Amazon AppStream. // -// Lists the fleets associated with the specified stack. +// Retrieves the name of the fleet that is associated with the specified stack. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1951,7 +1971,7 @@ func (c *AppStream) ListAssociatedStacksRequest(input *ListAssociatedStacksInput // ListAssociatedStacks API operation for Amazon AppStream. // -// Lists the stacks associated with the specified fleet. +// Retrieves the name of the stack with which the specified fleet is associated. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2025,8 +2045,8 @@ func (c *AppStream) ListTagsForResourceRequest(input *ListTagsForResourceInput) // ListTagsForResource API operation for Amazon AppStream. // -// Lists the tags for the specified AppStream 2.0 resource. You can tag AppStream -// 2.0 image builders, images, fleets, and stacks. +// Retrieves a list of all tags for the specified AppStream 2.0 resource. You +// can tag AppStream 2.0 image builders, images, fleets, and stacks. // // For more information about tags, see Tagging Your Resources (http://docs.aws.amazon.com/appstream2/latest/developerguide/tagging-basic.html) // in the Amazon AppStream 2.0 Developer Guide. @@ -2555,7 +2575,8 @@ func (c *AppStream) UntagResourceRequest(input *UntagResourceInput) (req *reques // UntagResource API operation for Amazon AppStream. // -// Disassociates the specified tags from the specified AppStream 2.0 resource. +// Disassociates one or more specified tags from the specified AppStream 2.0 +// resource. // // To list the current tags for your resources, use ListTagsForResource. // @@ -2639,7 +2660,9 @@ func (c *AppStream) UpdateDirectoryConfigRequest(input *UpdateDirectoryConfigInp // UpdateDirectoryConfig API operation for Amazon AppStream. // -// Updates the specified directory configuration. +// Updates the specified Directory Config object in AppStream 2.0. This object +// includes the information required to join streaming instances to an Active +// Directory domain. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2836,7 +2859,7 @@ func (c *AppStream) UpdateStackRequest(input *UpdateStackInput) (req *request.Re // UpdateStack API operation for Amazon AppStream. // -// Updates the specified stack. +// Updates the specified fields for the specified stack. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2868,6 +2891,9 @@ func (c *AppStream) UpdateStackRequest(input *UpdateStackInput) (req *request.Re // * ErrCodeIncompatibleImageException "IncompatibleImageException" // The image does not support storage connectors. // +// * ErrCodeOperationNotPermittedException "OperationNotPermittedException" +// The attempted operation is not permitted. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/appstream-2016-12-01/UpdateStack func (c *AppStream) UpdateStack(input *UpdateStackInput) (*UpdateStackOutput, error) { req, out := c.UpdateStackRequest(input) @@ -3820,13 +3846,17 @@ type CreateStackInput struct { // The name of the stack. // // Name is a required field - Name *string `min:"1" type:"string" required:"true"` + Name *string `type:"string" required:"true"` // The URL that users are redirected to after their streaming session ends. RedirectURL *string `type:"string"` // The storage connectors to enable. StorageConnectors []*StorageConnector `type:"list"` + + // The actions that are enabled or disabled for users during their streaming + // sessions. By default, these actions are enabled. + UserSettings []*UserSetting `min:"1" type:"list"` } // String returns the string representation @@ -3845,8 +3875,8 @@ func (s *CreateStackInput) Validate() error { if s.Name == nil { invalidParams.Add(request.NewErrParamRequired("Name")) } - if s.Name != nil && len(*s.Name) < 1 { - invalidParams.Add(request.NewErrParamMinLen("Name", 1)) + if s.UserSettings != nil && len(s.UserSettings) < 1 { + invalidParams.Add(request.NewErrParamMinLen("UserSettings", 1)) } if s.StorageConnectors != nil { for i, v := range s.StorageConnectors { @@ -3858,6 +3888,16 @@ func (s *CreateStackInput) Validate() error { } } } + if s.UserSettings != nil { + for i, v := range s.UserSettings { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "UserSettings", i), err.(request.ErrInvalidParams)) + } + } + } if invalidParams.Len() > 0 { return invalidParams @@ -3901,6 +3941,12 @@ func (s *CreateStackInput) SetStorageConnectors(v []*StorageConnector) *CreateSt return s } +// SetUserSettings sets the UserSettings field's value. +func (s *CreateStackInput) SetUserSettings(v []*UserSetting) *CreateStackInput { + s.UserSettings = v + return s +} + type CreateStackOutput struct { _ struct{} `type:"structure"` @@ -5723,7 +5769,7 @@ func (s *ListAssociatedFleetsInput) SetStackName(v string) *ListAssociatedFleets type ListAssociatedFleetsOutput struct { _ struct{} `type:"structure"` - // The names of the fleets. + // The name of the fleet. Names []*string `type:"list"` // The pagination token to use to retrieve the next page of results for this @@ -5810,7 +5856,7 @@ func (s *ListAssociatedStacksInput) SetNextToken(v string) *ListAssociatedStacks type ListAssociatedStacksOutput struct { _ struct{} `type:"structure"` - // The names of the stacks. + // The name of the stack. Names []*string `type:"list"` // The pagination token to use to retrieve the next page of results for this @@ -5901,6 +5947,42 @@ func (s *ListTagsForResourceOutput) SetTags(v map[string]*string) *ListTagsForRe return s } +// The network details of the fleet instance for the streaming session. +type NetworkAccessConfiguration struct { + _ struct{} `type:"structure"` + + // The resource identifier of the elastic network interface that is attached + // to instances in your VPC. All network interfaces have the eni-xxxxxxxx resource + // identifier. + EniId *string `min:"1" type:"string"` + + // The private IP address of the elastic network interface that is attached + // to instances in your VPC. + EniPrivateIpAddress *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s NetworkAccessConfiguration) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s NetworkAccessConfiguration) GoString() string { + return s.String() +} + +// SetEniId sets the EniId field's value. +func (s *NetworkAccessConfiguration) SetEniId(v string) *NetworkAccessConfiguration { + s.EniId = &v + return s +} + +// SetEniPrivateIpAddress sets the EniPrivateIpAddress field's value. +func (s *NetworkAccessConfiguration) SetEniPrivateIpAddress(v string) *NetworkAccessConfiguration { + s.EniPrivateIpAddress = &v + return s +} + // Describes a resource error. type ResourceError struct { _ struct{} `type:"structure"` @@ -6023,6 +6105,9 @@ type Session struct { // Id is a required field Id *string `min:"1" type:"string" required:"true"` + // The network details for the streaming session. + NetworkAccessConfiguration *NetworkAccessConfiguration `type:"structure"` + // The name of the stack for the streaming session. // // StackName is a required field @@ -6067,6 +6152,12 @@ func (s *Session) SetId(v string) *Session { return s } +// SetNetworkAccessConfiguration sets the NetworkAccessConfiguration field's value. +func (s *Session) SetNetworkAccessConfiguration(v *NetworkAccessConfiguration) *Session { + s.NetworkAccessConfiguration = v + return s +} + // SetStackName sets the StackName field's value. func (s *Session) SetStackName(v string) *Session { s.StackName = &v @@ -6118,6 +6209,10 @@ type Stack struct { // The storage connectors to enable. StorageConnectors []*StorageConnector `type:"list"` + + // The actions that are enabled or disabled for users during their streaming + // sessions. By default these actions are enabled. + UserSettings []*UserSetting `min:"1" type:"list"` } // String returns the string representation @@ -6184,6 +6279,12 @@ func (s *Stack) SetStorageConnectors(v []*StorageConnector) *Stack { return s } +// SetUserSettings sets the UserSettings field's value. +func (s *Stack) SetUserSettings(v []*UserSetting) *Stack { + s.UserSettings = v + return s +} + // Describes a stack error. type StackError struct { _ struct{} `type:"structure"` @@ -6468,7 +6569,7 @@ func (s *StopImageBuilderOutput) SetImageBuilder(v *ImageBuilder) *StopImageBuil return s } -// Describes a storage connector. +// Describes a connector to enable persistent storage for users. type StorageConnector struct { _ struct{} `type:"structure"` @@ -6477,6 +6578,9 @@ type StorageConnector struct { // ConnectorType is a required field ConnectorType *string `type:"string" required:"true" enum:"StorageConnectorType"` + // The names of the domains for the G Suite account. + Domains []*string `type:"list"` + // The ARN of the storage connector. ResourceIdentifier *string `min:"1" type:"string"` } @@ -6513,6 +6617,12 @@ func (s *StorageConnector) SetConnectorType(v string) *StorageConnector { return s } +// SetDomains sets the Domains field's value. +func (s *StorageConnector) SetDomains(v []*string) *StorageConnector { + s.Domains = v + return s +} + // SetResourceIdentifier sets the ResourceIdentifier field's value. func (s *StorageConnector) SetResourceIdentifier(v string) *StorageConnector { s.ResourceIdentifier = &v @@ -6663,7 +6773,7 @@ func (s UntagResourceOutput) GoString() string { type UpdateDirectoryConfigInput struct { _ struct{} `type:"structure"` - // The name of the directory configuration. + // The name of the Directory Config object. // // DirectoryName is a required field DirectoryName *string `type:"string" required:"true"` @@ -6725,7 +6835,7 @@ func (s *UpdateDirectoryConfigInput) SetServiceAccountCredentials(v *ServiceAcco type UpdateDirectoryConfigOutput struct { _ struct{} `type:"structure"` - // Information about the directory configuration. + // Information about the Directory Config object. DirectoryConfig *DirectoryConfig `type:"structure"` } @@ -7002,6 +7112,10 @@ type UpdateStackInput struct { // The storage connectors to enable. StorageConnectors []*StorageConnector `type:"list"` + + // The actions that are enabled or disabled for users during their streaming + // sessions. By default, these actions are enabled. + UserSettings []*UserSetting `min:"1" type:"list"` } // String returns the string representation @@ -7023,6 +7137,9 @@ func (s *UpdateStackInput) Validate() error { if s.Name != nil && len(*s.Name) < 1 { invalidParams.Add(request.NewErrParamMinLen("Name", 1)) } + if s.UserSettings != nil && len(s.UserSettings) < 1 { + invalidParams.Add(request.NewErrParamMinLen("UserSettings", 1)) + } if s.StorageConnectors != nil { for i, v := range s.StorageConnectors { if v == nil { @@ -7033,6 +7150,16 @@ func (s *UpdateStackInput) Validate() error { } } } + if s.UserSettings != nil { + for i, v := range s.UserSettings { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "UserSettings", i), err.(request.ErrInvalidParams)) + } + } + } if invalidParams.Len() > 0 { return invalidParams @@ -7088,6 +7215,12 @@ func (s *UpdateStackInput) SetStorageConnectors(v []*StorageConnector) *UpdateSt return s } +// SetUserSettings sets the UserSettings field's value. +func (s *UpdateStackInput) SetUserSettings(v []*UserSetting) *UpdateStackInput { + s.UserSettings = v + return s +} + type UpdateStackOutput struct { _ struct{} `type:"structure"` @@ -7111,6 +7244,60 @@ func (s *UpdateStackOutput) SetStack(v *Stack) *UpdateStackOutput { return s } +// Describes an action and whether the action is enabled or disabled for users +// during their streaming sessions. +type UserSetting struct { + _ struct{} `type:"structure"` + + // The action that is enabled or disabled. + // + // Action is a required field + Action *string `type:"string" required:"true" enum:"Action"` + + // Indicates whether the action is enabled or disabled. + // + // Permission is a required field + Permission *string `type:"string" required:"true" enum:"Permission"` +} + +// String returns the string representation +func (s UserSetting) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UserSetting) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *UserSetting) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "UserSetting"} + if s.Action == nil { + invalidParams.Add(request.NewErrParamRequired("Action")) + } + if s.Permission == nil { + invalidParams.Add(request.NewErrParamRequired("Permission")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAction sets the Action field's value. +func (s *UserSetting) SetAction(v string) *UserSetting { + s.Action = &v + return s +} + +// SetPermission sets the Permission field's value. +func (s *UserSetting) SetPermission(v string) *UserSetting { + s.Permission = &v + return s +} + // Describes VPC configuration information. type VpcConfig struct { _ struct{} `type:"structure"` @@ -7144,6 +7331,23 @@ func (s *VpcConfig) SetSubnetIds(v []*string) *VpcConfig { return s } +const ( + // ActionClipboardCopyFromLocalDevice is a Action enum value + ActionClipboardCopyFromLocalDevice = "CLIPBOARD_COPY_FROM_LOCAL_DEVICE" + + // ActionClipboardCopyToLocalDevice is a Action enum value + ActionClipboardCopyToLocalDevice = "CLIPBOARD_COPY_TO_LOCAL_DEVICE" + + // ActionFileUpload is a Action enum value + ActionFileUpload = "FILE_UPLOAD" + + // ActionFileDownload is a Action enum value + ActionFileDownload = "FILE_DOWNLOAD" + + // ActionPrintingToLocalDevice is a Action enum value + ActionPrintingToLocalDevice = "PRINTING_TO_LOCAL_DEVICE" +) + const ( // AuthenticationTypeApi is a AuthenticationType enum value AuthenticationTypeApi = "API" @@ -7334,6 +7538,14 @@ const ( ImageStateChangeReasonCodeImageCopyFailure = "IMAGE_COPY_FAILURE" ) +const ( + // PermissionEnabled is a Permission enum value + PermissionEnabled = "ENABLED" + + // PermissionDisabled is a Permission enum value + PermissionDisabled = "DISABLED" +) + const ( // PlatformTypeWindows is a PlatformType enum value PlatformTypeWindows = "WINDOWS" @@ -7355,6 +7567,12 @@ const ( // StackAttributeStorageConnectors is a StackAttribute enum value StackAttributeStorageConnectors = "STORAGE_CONNECTORS" + // StackAttributeStorageConnectorHomefolders is a StackAttribute enum value + StackAttributeStorageConnectorHomefolders = "STORAGE_CONNECTOR_HOMEFOLDERS" + + // StackAttributeStorageConnectorGoogleDrive is a StackAttribute enum value + StackAttributeStorageConnectorGoogleDrive = "STORAGE_CONNECTOR_GOOGLE_DRIVE" + // StackAttributeRedirectUrl is a StackAttribute enum value StackAttributeRedirectUrl = "REDIRECT_URL" @@ -7363,6 +7581,9 @@ const ( // StackAttributeThemeName is a StackAttribute enum value StackAttributeThemeName = "THEME_NAME" + + // StackAttributeUserSettings is a StackAttribute enum value + StackAttributeUserSettings = "USER_SETTINGS" ) const ( @@ -7377,6 +7598,9 @@ const ( const ( // StorageConnectorTypeHomefolders is a StorageConnectorType enum value StorageConnectorTypeHomefolders = "HOMEFOLDERS" + + // StorageConnectorTypeGoogleDrive is a StorageConnectorType enum value + StorageConnectorTypeGoogleDrive = "GOOGLE_DRIVE" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/appstream/service.go b/vendor/github.com/aws/aws-sdk-go/service/appstream/service.go index acf6963fa..168cf44e0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/appstream/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/appstream/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "appstream2" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "appstream2" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "AppStream" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the AppStream client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/appsync/service.go b/vendor/github.com/aws/aws-sdk-go/service/appsync/service.go index 6af2be6d9..e53f6e71b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/appsync/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/appsync/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "appsync" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "appsync" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "AppSync" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the AppSync client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/athena/service.go b/vendor/github.com/aws/aws-sdk-go/service/athena/service.go index c1fd29b0f..6806a62ec 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/athena/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/athena/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "athena" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "athena" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Athena" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Athena client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/autoscaling/service.go b/vendor/github.com/aws/aws-sdk-go/service/autoscaling/service.go index 5e63d1c03..e1da9fd75 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/autoscaling/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/autoscaling/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "autoscaling" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "autoscaling" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Auto Scaling" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the AutoScaling client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/autoscalingplans/service.go b/vendor/github.com/aws/aws-sdk-go/service/autoscalingplans/service.go index a9a9495d4..dd6a1ece4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/autoscalingplans/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/autoscalingplans/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "autoscaling" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "autoscaling" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Auto Scaling Plans" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the AutoScalingPlans client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/batch/service.go b/vendor/github.com/aws/aws-sdk-go/service/batch/service.go index 11539325a..0d04e7497 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/batch/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/batch/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "batch" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "batch" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Batch" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Batch client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/budgets/service.go b/vendor/github.com/aws/aws-sdk-go/service/budgets/service.go index d74d9c96a..dc10c2aaa 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/budgets/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/budgets/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "budgets" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "budgets" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Budgets" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Budgets client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloud9/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloud9/service.go index 1a01d2d92..8b9d30d54 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloud9/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloud9/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloud9" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloud9" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Cloud9" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Cloud9 client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/api.go b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/api.go index e5d0dd891..a0563192f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/api.go @@ -38,7 +38,7 @@ const opAddFacetToObject = "AddFacetToObject" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AddFacetToObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AddFacetToObject func (c *CloudDirectory) AddFacetToObjectRequest(input *AddFacetToObjectInput) (req *request.Request, output *AddFacetToObjectOutput) { op := &request.Operation{ Name: opAddFacetToObject, @@ -106,7 +106,7 @@ func (c *CloudDirectory) AddFacetToObjectRequest(input *AddFacetToObjectInput) ( // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AddFacetToObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AddFacetToObject func (c *CloudDirectory) AddFacetToObject(input *AddFacetToObjectInput) (*AddFacetToObjectOutput, error) { req, out := c.AddFacetToObjectRequest(input) return out, req.Send() @@ -153,7 +153,7 @@ const opApplySchema = "ApplySchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ApplySchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ApplySchema func (c *CloudDirectory) ApplySchemaRequest(input *ApplySchemaInput) (req *request.Request, output *ApplySchemaOutput) { op := &request.Operation{ Name: opApplySchema, @@ -211,6 +211,10 @@ func (c *CloudDirectory) ApplySchemaRequest(input *ApplySchemaInput) (req *reque // * ErrCodeAccessDeniedException "AccessDeniedException" // Access denied. Check your permissions. // +// * ErrCodeSchemaAlreadyExistsException "SchemaAlreadyExistsException" +// Indicates that a schema could not be created due to a naming conflict. Please +// select a different name and then try again. +// // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // @@ -219,7 +223,7 @@ func (c *CloudDirectory) ApplySchemaRequest(input *ApplySchemaInput) (req *reque // to apply a schema with the same name has occurred. Rename the link or the // schema and then try again. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ApplySchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ApplySchema func (c *CloudDirectory) ApplySchema(input *ApplySchemaInput) (*ApplySchemaOutput, error) { req, out := c.ApplySchemaRequest(input) return out, req.Send() @@ -266,7 +270,7 @@ const opAttachObject = "AttachObject" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachObject func (c *CloudDirectory) AttachObjectRequest(input *AttachObjectInput) (req *request.Request, output *AttachObjectOutput) { op := &request.Operation{ Name: opAttachObject, @@ -351,7 +355,7 @@ func (c *CloudDirectory) AttachObjectRequest(input *AttachObjectInput) (req *req // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachObject func (c *CloudDirectory) AttachObject(input *AttachObjectInput) (*AttachObjectOutput, error) { req, out := c.AttachObjectRequest(input) return out, req.Send() @@ -398,7 +402,7 @@ const opAttachPolicy = "AttachPolicy" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachPolicy +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachPolicy func (c *CloudDirectory) AttachPolicyRequest(input *AttachPolicyInput) (req *request.Request, output *AttachPolicyOutput) { op := &request.Operation{ Name: opAttachPolicy, @@ -465,7 +469,7 @@ func (c *CloudDirectory) AttachPolicyRequest(input *AttachPolicyInput) (req *req // * ErrCodeNotPolicyException "NotPolicyException" // Indicates that the requested operation can only operate on policy objects. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachPolicy +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachPolicy func (c *CloudDirectory) AttachPolicy(input *AttachPolicyInput) (*AttachPolicyOutput, error) { req, out := c.AttachPolicyRequest(input) return out, req.Send() @@ -512,7 +516,7 @@ const opAttachToIndex = "AttachToIndex" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachToIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachToIndex func (c *CloudDirectory) AttachToIndexRequest(input *AttachToIndexInput) (req *request.Request, output *AttachToIndexOutput) { op := &request.Operation{ Name: opAttachToIndex, @@ -591,7 +595,7 @@ func (c *CloudDirectory) AttachToIndexRequest(input *AttachToIndexInput) (req *r // * ErrCodeNotIndexException "NotIndexException" // Indicates that the requested operation can only operate on index objects. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachToIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachToIndex func (c *CloudDirectory) AttachToIndex(input *AttachToIndexInput) (*AttachToIndexOutput, error) { req, out := c.AttachToIndexRequest(input) return out, req.Send() @@ -638,7 +642,7 @@ const opAttachTypedLink = "AttachTypedLink" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachTypedLink +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachTypedLink func (c *CloudDirectory) AttachTypedLinkRequest(input *AttachTypedLinkInput) (req *request.Request, output *AttachTypedLinkOutput) { op := &request.Operation{ Name: opAttachTypedLink, @@ -715,7 +719,7 @@ func (c *CloudDirectory) AttachTypedLinkRequest(input *AttachTypedLinkInput) (re // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/AttachTypedLink +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/AttachTypedLink func (c *CloudDirectory) AttachTypedLink(input *AttachTypedLinkInput) (*AttachTypedLinkOutput, error) { req, out := c.AttachTypedLinkRequest(input) return out, req.Send() @@ -762,7 +766,7 @@ const opBatchRead = "BatchRead" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/BatchRead +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/BatchRead func (c *CloudDirectory) BatchReadRequest(input *BatchReadInput) (req *request.Request, output *BatchReadOutput) { op := &request.Operation{ Name: opBatchRead, @@ -822,7 +826,7 @@ func (c *CloudDirectory) BatchReadRequest(input *BatchReadInput) (req *request.R // * ErrCodeDirectoryNotEnabledException "DirectoryNotEnabledException" // Operations are only permitted on enabled directories. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/BatchRead +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/BatchRead func (c *CloudDirectory) BatchRead(input *BatchReadInput) (*BatchReadOutput, error) { req, out := c.BatchReadRequest(input) return out, req.Send() @@ -869,7 +873,7 @@ const opBatchWrite = "BatchWrite" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/BatchWrite +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/BatchWrite func (c *CloudDirectory) BatchWriteRequest(input *BatchWriteInput) (req *request.Request, output *BatchWriteOutput) { op := &request.Operation{ Name: opBatchWrite, @@ -933,7 +937,7 @@ func (c *CloudDirectory) BatchWriteRequest(input *BatchWriteInput) (req *request // * ErrCodeBatchWriteException "BatchWriteException" // A BatchWrite exception has occurred. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/BatchWrite +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/BatchWrite func (c *CloudDirectory) BatchWrite(input *BatchWriteInput) (*BatchWriteOutput, error) { req, out := c.BatchWriteRequest(input) return out, req.Send() @@ -980,7 +984,7 @@ const opCreateDirectory = "CreateDirectory" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateDirectory func (c *CloudDirectory) CreateDirectoryRequest(input *CreateDirectoryInput) (req *request.Request, output *CreateDirectoryOutput) { op := &request.Operation{ Name: opCreateDirectory, @@ -1045,7 +1049,7 @@ func (c *CloudDirectory) CreateDirectoryRequest(input *CreateDirectoryInput) (re // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateDirectory func (c *CloudDirectory) CreateDirectory(input *CreateDirectoryInput) (*CreateDirectoryOutput, error) { req, out := c.CreateDirectoryRequest(input) return out, req.Send() @@ -1092,7 +1096,7 @@ const opCreateFacet = "CreateFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateFacet func (c *CloudDirectory) CreateFacetRequest(input *CreateFacetInput) (req *request.Request, output *CreateFacetOutput) { op := &request.Operation{ Name: opCreateFacet, @@ -1163,7 +1167,7 @@ func (c *CloudDirectory) CreateFacetRequest(input *CreateFacetInput) (req *reque // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateFacet func (c *CloudDirectory) CreateFacet(input *CreateFacetInput) (*CreateFacetOutput, error) { req, out := c.CreateFacetRequest(input) return out, req.Send() @@ -1210,7 +1214,7 @@ const opCreateIndex = "CreateIndex" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateIndex func (c *CloudDirectory) CreateIndexRequest(input *CreateIndexInput) (req *request.Request, output *CreateIndexOutput) { op := &request.Operation{ Name: opCreateIndex, @@ -1285,7 +1289,7 @@ func (c *CloudDirectory) CreateIndexRequest(input *CreateIndexInput) (req *reque // * ErrCodeUnsupportedIndexTypeException "UnsupportedIndexTypeException" // Indicates that the requested index type is not supported. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateIndex func (c *CloudDirectory) CreateIndex(input *CreateIndexInput) (*CreateIndexOutput, error) { req, out := c.CreateIndexRequest(input) return out, req.Send() @@ -1332,7 +1336,7 @@ const opCreateObject = "CreateObject" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateObject func (c *CloudDirectory) CreateObjectRequest(input *CreateObjectInput) (req *request.Request, output *CreateObjectOutput) { op := &request.Operation{ Name: opCreateObject, @@ -1409,7 +1413,7 @@ func (c *CloudDirectory) CreateObjectRequest(input *CreateObjectInput) (req *req // * ErrCodeUnsupportedIndexTypeException "UnsupportedIndexTypeException" // Indicates that the requested index type is not supported. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateObject func (c *CloudDirectory) CreateObject(input *CreateObjectInput) (*CreateObjectOutput, error) { req, out := c.CreateObjectRequest(input) return out, req.Send() @@ -1456,7 +1460,7 @@ const opCreateSchema = "CreateSchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateSchema func (c *CloudDirectory) CreateSchemaRequest(input *CreateSchemaInput) (req *request.Request, output *CreateSchemaOutput) { op := &request.Operation{ Name: opCreateSchema, @@ -1532,7 +1536,7 @@ func (c *CloudDirectory) CreateSchemaRequest(input *CreateSchemaInput) (req *req // * ErrCodeAccessDeniedException "AccessDeniedException" // Access denied. Check your permissions. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateSchema func (c *CloudDirectory) CreateSchema(input *CreateSchemaInput) (*CreateSchemaOutput, error) { req, out := c.CreateSchemaRequest(input) return out, req.Send() @@ -1579,7 +1583,7 @@ const opCreateTypedLinkFacet = "CreateTypedLinkFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateTypedLinkFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateTypedLinkFacet func (c *CloudDirectory) CreateTypedLinkFacetRequest(input *CreateTypedLinkFacetInput) (req *request.Request, output *CreateTypedLinkFacetOutput) { op := &request.Operation{ Name: opCreateTypedLinkFacet, @@ -1649,7 +1653,7 @@ func (c *CloudDirectory) CreateTypedLinkFacetRequest(input *CreateTypedLinkFacet // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/CreateTypedLinkFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/CreateTypedLinkFacet func (c *CloudDirectory) CreateTypedLinkFacet(input *CreateTypedLinkFacetInput) (*CreateTypedLinkFacetOutput, error) { req, out := c.CreateTypedLinkFacetRequest(input) return out, req.Send() @@ -1696,7 +1700,7 @@ const opDeleteDirectory = "DeleteDirectory" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteDirectory func (c *CloudDirectory) DeleteDirectoryRequest(input *DeleteDirectoryInput) (req *request.Request, output *DeleteDirectoryOutput) { op := &request.Operation{ Name: opDeleteDirectory, @@ -1764,7 +1768,7 @@ func (c *CloudDirectory) DeleteDirectoryRequest(input *DeleteDirectoryInput) (re // * ErrCodeInvalidArnException "InvalidArnException" // Indicates that the provided ARN value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteDirectory func (c *CloudDirectory) DeleteDirectory(input *DeleteDirectoryInput) (*DeleteDirectoryOutput, error) { req, out := c.DeleteDirectoryRequest(input) return out, req.Send() @@ -1811,7 +1815,7 @@ const opDeleteFacet = "DeleteFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteFacet func (c *CloudDirectory) DeleteFacetRequest(input *DeleteFacetInput) (req *request.Request, output *DeleteFacetOutput) { op := &request.Operation{ Name: opDeleteFacet, @@ -1879,7 +1883,7 @@ func (c *CloudDirectory) DeleteFacetRequest(input *DeleteFacetInput) (req *reque // Occurs when deleting a facet that contains an attribute that is a target // to an attribute reference in a different facet. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteFacet func (c *CloudDirectory) DeleteFacet(input *DeleteFacetInput) (*DeleteFacetOutput, error) { req, out := c.DeleteFacetRequest(input) return out, req.Send() @@ -1926,7 +1930,7 @@ const opDeleteObject = "DeleteObject" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteObject func (c *CloudDirectory) DeleteObjectRequest(input *DeleteObjectInput) (req *request.Request, output *DeleteObjectOutput) { op := &request.Operation{ Name: opDeleteObject, @@ -1994,7 +1998,7 @@ func (c *CloudDirectory) DeleteObjectRequest(input *DeleteObjectInput) (req *req // Indicates that the requested operation cannot be completed because the object // has not been detached from the tree. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteObject func (c *CloudDirectory) DeleteObject(input *DeleteObjectInput) (*DeleteObjectOutput, error) { req, out := c.DeleteObjectRequest(input) return out, req.Send() @@ -2041,7 +2045,7 @@ const opDeleteSchema = "DeleteSchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteSchema func (c *CloudDirectory) DeleteSchemaRequest(input *DeleteSchemaInput) (req *request.Request, output *DeleteSchemaOutput) { op := &request.Operation{ Name: opDeleteSchema, @@ -2106,7 +2110,7 @@ func (c *CloudDirectory) DeleteSchemaRequest(input *DeleteSchemaInput) (req *req // The object could not be deleted because links still exist. Remove the links // and then try the operation again. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteSchema func (c *CloudDirectory) DeleteSchema(input *DeleteSchemaInput) (*DeleteSchemaOutput, error) { req, out := c.DeleteSchemaRequest(input) return out, req.Send() @@ -2153,7 +2157,7 @@ const opDeleteTypedLinkFacet = "DeleteTypedLinkFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteTypedLinkFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteTypedLinkFacet func (c *CloudDirectory) DeleteTypedLinkFacetRequest(input *DeleteTypedLinkFacetInput) (req *request.Request, output *DeleteTypedLinkFacetOutput) { op := &request.Operation{ Name: opDeleteTypedLinkFacet, @@ -2216,7 +2220,7 @@ func (c *CloudDirectory) DeleteTypedLinkFacetRequest(input *DeleteTypedLinkFacet // * ErrCodeFacetNotFoundException "FacetNotFoundException" // The specified Facet could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DeleteTypedLinkFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DeleteTypedLinkFacet func (c *CloudDirectory) DeleteTypedLinkFacet(input *DeleteTypedLinkFacetInput) (*DeleteTypedLinkFacetOutput, error) { req, out := c.DeleteTypedLinkFacetRequest(input) return out, req.Send() @@ -2263,7 +2267,7 @@ const opDetachFromIndex = "DetachFromIndex" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachFromIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachFromIndex func (c *CloudDirectory) DetachFromIndexRequest(input *DetachFromIndexInput) (req *request.Request, output *DetachFromIndexOutput) { op := &request.Operation{ Name: opDetachFromIndex, @@ -2332,7 +2336,7 @@ func (c *CloudDirectory) DetachFromIndexRequest(input *DetachFromIndexInput) (re // * ErrCodeNotIndexException "NotIndexException" // Indicates that the requested operation can only operate on index objects. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachFromIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachFromIndex func (c *CloudDirectory) DetachFromIndex(input *DetachFromIndexInput) (*DetachFromIndexOutput, error) { req, out := c.DetachFromIndexRequest(input) return out, req.Send() @@ -2379,7 +2383,7 @@ const opDetachObject = "DetachObject" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachObject func (c *CloudDirectory) DetachObjectRequest(input *DetachObjectInput) (req *request.Request, output *DetachObjectOutput) { op := &request.Operation{ Name: opDetachObject, @@ -2447,7 +2451,7 @@ func (c *CloudDirectory) DetachObjectRequest(input *DetachObjectInput) (req *req // Occurs when any invalid operations are performed on an object that is not // a node, such as calling ListObjectChildren for a leaf node object. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachObject func (c *CloudDirectory) DetachObject(input *DetachObjectInput) (*DetachObjectOutput, error) { req, out := c.DetachObjectRequest(input) return out, req.Send() @@ -2494,7 +2498,7 @@ const opDetachPolicy = "DetachPolicy" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachPolicy +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachPolicy func (c *CloudDirectory) DetachPolicyRequest(input *DetachPolicyInput) (req *request.Request, output *DetachPolicyOutput) { op := &request.Operation{ Name: opDetachPolicy, @@ -2560,7 +2564,7 @@ func (c *CloudDirectory) DetachPolicyRequest(input *DetachPolicyInput) (req *req // * ErrCodeNotPolicyException "NotPolicyException" // Indicates that the requested operation can only operate on policy objects. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachPolicy +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachPolicy func (c *CloudDirectory) DetachPolicy(input *DetachPolicyInput) (*DetachPolicyOutput, error) { req, out := c.DetachPolicyRequest(input) return out, req.Send() @@ -2607,7 +2611,7 @@ const opDetachTypedLink = "DetachTypedLink" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachTypedLink +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachTypedLink func (c *CloudDirectory) DetachTypedLinkRequest(input *DetachTypedLinkInput) (req *request.Request, output *DetachTypedLinkOutput) { op := &request.Operation{ Name: opDetachTypedLink, @@ -2677,7 +2681,7 @@ func (c *CloudDirectory) DetachTypedLinkRequest(input *DetachTypedLinkInput) (re // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DetachTypedLink +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DetachTypedLink func (c *CloudDirectory) DetachTypedLink(input *DetachTypedLinkInput) (*DetachTypedLinkOutput, error) { req, out := c.DetachTypedLinkRequest(input) return out, req.Send() @@ -2724,7 +2728,7 @@ const opDisableDirectory = "DisableDirectory" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DisableDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DisableDirectory func (c *CloudDirectory) DisableDirectoryRequest(input *DisableDirectoryInput) (req *request.Request, output *DisableDirectoryOutput) { op := &request.Operation{ Name: opDisableDirectory, @@ -2790,7 +2794,7 @@ func (c *CloudDirectory) DisableDirectoryRequest(input *DisableDirectoryInput) ( // * ErrCodeInvalidArnException "InvalidArnException" // Indicates that the provided ARN value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/DisableDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/DisableDirectory func (c *CloudDirectory) DisableDirectory(input *DisableDirectoryInput) (*DisableDirectoryOutput, error) { req, out := c.DisableDirectoryRequest(input) return out, req.Send() @@ -2837,7 +2841,7 @@ const opEnableDirectory = "EnableDirectory" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/EnableDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/EnableDirectory func (c *CloudDirectory) EnableDirectoryRequest(input *EnableDirectoryInput) (req *request.Request, output *EnableDirectoryOutput) { op := &request.Operation{ Name: opEnableDirectory, @@ -2902,7 +2906,7 @@ func (c *CloudDirectory) EnableDirectoryRequest(input *EnableDirectoryInput) (re // * ErrCodeInvalidArnException "InvalidArnException" // Indicates that the provided ARN value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/EnableDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/EnableDirectory func (c *CloudDirectory) EnableDirectory(input *EnableDirectoryInput) (*EnableDirectoryOutput, error) { req, out := c.EnableDirectoryRequest(input) return out, req.Send() @@ -2949,7 +2953,7 @@ const opGetAppliedSchemaVersion = "GetAppliedSchemaVersion" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetAppliedSchemaVersion +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetAppliedSchemaVersion func (c *CloudDirectory) GetAppliedSchemaVersionRequest(input *GetAppliedSchemaVersionInput) (req *request.Request, output *GetAppliedSchemaVersionOutput) { op := &request.Operation{ Name: opGetAppliedSchemaVersion, @@ -3010,7 +3014,7 @@ func (c *CloudDirectory) GetAppliedSchemaVersionRequest(input *GetAppliedSchemaV // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetAppliedSchemaVersion +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetAppliedSchemaVersion func (c *CloudDirectory) GetAppliedSchemaVersion(input *GetAppliedSchemaVersionInput) (*GetAppliedSchemaVersionOutput, error) { req, out := c.GetAppliedSchemaVersionRequest(input) return out, req.Send() @@ -3057,7 +3061,7 @@ const opGetDirectory = "GetDirectory" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetDirectory func (c *CloudDirectory) GetDirectoryRequest(input *GetDirectoryInput) (req *request.Request, output *GetDirectoryOutput) { op := &request.Operation{ Name: opGetDirectory, @@ -3114,7 +3118,7 @@ func (c *CloudDirectory) GetDirectoryRequest(input *GetDirectoryInput) (req *req // * ErrCodeAccessDeniedException "AccessDeniedException" // Access denied. Check your permissions. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetDirectory +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetDirectory func (c *CloudDirectory) GetDirectory(input *GetDirectoryInput) (*GetDirectoryOutput, error) { req, out := c.GetDirectoryRequest(input) return out, req.Send() @@ -3161,7 +3165,7 @@ const opGetFacet = "GetFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetFacet func (c *CloudDirectory) GetFacetRequest(input *GetFacetInput) (req *request.Request, output *GetFacetOutput) { op := &request.Operation{ Name: opGetFacet, @@ -3226,7 +3230,7 @@ func (c *CloudDirectory) GetFacetRequest(input *GetFacetInput) (req *request.Req // * ErrCodeFacetNotFoundException "FacetNotFoundException" // The specified Facet could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetFacet func (c *CloudDirectory) GetFacet(input *GetFacetInput) (*GetFacetOutput, error) { req, out := c.GetFacetRequest(input) return out, req.Send() @@ -3248,6 +3252,120 @@ func (c *CloudDirectory) GetFacetWithContext(ctx aws.Context, input *GetFacetInp return out, req.Send() } +const opGetLinkAttributes = "GetLinkAttributes" + +// GetLinkAttributesRequest generates a "aws/request.Request" representing the +// client's request for the GetLinkAttributes operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See GetLinkAttributes for more information on using the GetLinkAttributes +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the GetLinkAttributesRequest method. +// req, resp := client.GetLinkAttributesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetLinkAttributes +func (c *CloudDirectory) GetLinkAttributesRequest(input *GetLinkAttributesInput) (req *request.Request, output *GetLinkAttributesOutput) { + op := &request.Operation{ + Name: opGetLinkAttributes, + HTTPMethod: "POST", + HTTPPath: "/amazonclouddirectory/2017-01-11/typedlink/attributes/get", + } + + if input == nil { + input = &GetLinkAttributesInput{} + } + + output = &GetLinkAttributesOutput{} + req = c.newRequest(op, input, output) + return +} + +// GetLinkAttributes API operation for Amazon CloudDirectory. +// +// Retrieves attributes that are associated with a typed link. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon CloudDirectory's +// API operation GetLinkAttributes for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalServiceException "InternalServiceException" +// Indicates a problem that must be resolved by Amazon Web Services. This might +// be a transient error in which case you can retry your request until it succeeds. +// Otherwise, go to the AWS Service Health Dashboard (http://status.aws.amazon.com/) +// site to see if there are any operational issues with the service. +// +// * ErrCodeInvalidArnException "InvalidArnException" +// Indicates that the provided ARN value is not valid. +// +// * ErrCodeRetryableConflictException "RetryableConflictException" +// Occurs when a conflict with a previous successful write is detected. For +// example, if a write operation occurs on an object and then an attempt is +// made to read the object using “SERIALIZABLE” consistency, this exception +// may result. This generally occurs when the previous write did not have time +// to propagate to the host serving the current request. A retry (with appropriate +// backoff logic) is the recommended response to this exception. +// +// * ErrCodeValidationException "ValidationException" +// Indicates that your request is malformed in some manner. See the exception +// message. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// Indicates that limits are exceeded. See Limits (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/limits.html) +// for more information. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// Access denied. Check your permissions. +// +// * ErrCodeDirectoryNotEnabledException "DirectoryNotEnabledException" +// Operations are only permitted on enabled directories. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource could not be found. +// +// * ErrCodeFacetValidationException "FacetValidationException" +// The Facet that you provided was not well formed or could not be validated +// with the schema. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetLinkAttributes +func (c *CloudDirectory) GetLinkAttributes(input *GetLinkAttributesInput) (*GetLinkAttributesOutput, error) { + req, out := c.GetLinkAttributesRequest(input) + return out, req.Send() +} + +// GetLinkAttributesWithContext is the same as GetLinkAttributes with the addition of +// the ability to pass a context and additional request options. +// +// See GetLinkAttributes for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *CloudDirectory) GetLinkAttributesWithContext(ctx aws.Context, input *GetLinkAttributesInput, opts ...request.Option) (*GetLinkAttributesOutput, error) { + req, out := c.GetLinkAttributesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opGetObjectAttributes = "GetObjectAttributes" // GetObjectAttributesRequest generates a "aws/request.Request" representing the @@ -3273,7 +3391,7 @@ const opGetObjectAttributes = "GetObjectAttributes" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetObjectAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetObjectAttributes func (c *CloudDirectory) GetObjectAttributesRequest(input *GetObjectAttributesInput) (req *request.Request, output *GetObjectAttributesOutput) { op := &request.Operation{ Name: opGetObjectAttributes, @@ -3340,7 +3458,7 @@ func (c *CloudDirectory) GetObjectAttributesRequest(input *GetObjectAttributesIn // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetObjectAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetObjectAttributes func (c *CloudDirectory) GetObjectAttributes(input *GetObjectAttributesInput) (*GetObjectAttributesOutput, error) { req, out := c.GetObjectAttributesRequest(input) return out, req.Send() @@ -3387,7 +3505,7 @@ const opGetObjectInformation = "GetObjectInformation" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetObjectInformation +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetObjectInformation func (c *CloudDirectory) GetObjectInformationRequest(input *GetObjectInformationInput) (req *request.Request, output *GetObjectInformationOutput) { op := &request.Operation{ Name: opGetObjectInformation, @@ -3450,7 +3568,7 @@ func (c *CloudDirectory) GetObjectInformationRequest(input *GetObjectInformation // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetObjectInformation +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetObjectInformation func (c *CloudDirectory) GetObjectInformation(input *GetObjectInformationInput) (*GetObjectInformationOutput, error) { req, out := c.GetObjectInformationRequest(input) return out, req.Send() @@ -3497,7 +3615,7 @@ const opGetSchemaAsJson = "GetSchemaAsJson" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetSchemaAsJson +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetSchemaAsJson func (c *CloudDirectory) GetSchemaAsJsonRequest(input *GetSchemaAsJsonInput) (req *request.Request, output *GetSchemaAsJsonOutput) { op := &request.Operation{ Name: opGetSchemaAsJson, @@ -3562,7 +3680,7 @@ func (c *CloudDirectory) GetSchemaAsJsonRequest(input *GetSchemaAsJsonInput) (re // Indicates that your request is malformed in some manner. See the exception // message. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetSchemaAsJson +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetSchemaAsJson func (c *CloudDirectory) GetSchemaAsJson(input *GetSchemaAsJsonInput) (*GetSchemaAsJsonOutput, error) { req, out := c.GetSchemaAsJsonRequest(input) return out, req.Send() @@ -3609,7 +3727,7 @@ const opGetTypedLinkFacetInformation = "GetTypedLinkFacetInformation" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetTypedLinkFacetInformation +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetTypedLinkFacetInformation func (c *CloudDirectory) GetTypedLinkFacetInformationRequest(input *GetTypedLinkFacetInformationInput) (req *request.Request, output *GetTypedLinkFacetInformationOutput) { op := &request.Operation{ Name: opGetTypedLinkFacetInformation, @@ -3676,7 +3794,7 @@ func (c *CloudDirectory) GetTypedLinkFacetInformationRequest(input *GetTypedLink // * ErrCodeFacetNotFoundException "FacetNotFoundException" // The specified Facet could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/GetTypedLinkFacetInformation +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/GetTypedLinkFacetInformation func (c *CloudDirectory) GetTypedLinkFacetInformation(input *GetTypedLinkFacetInformationInput) (*GetTypedLinkFacetInformationOutput, error) { req, out := c.GetTypedLinkFacetInformationRequest(input) return out, req.Send() @@ -3723,7 +3841,7 @@ const opListAppliedSchemaArns = "ListAppliedSchemaArns" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListAppliedSchemaArns +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListAppliedSchemaArns func (c *CloudDirectory) ListAppliedSchemaArnsRequest(input *ListAppliedSchemaArnsInput) (req *request.Request, output *ListAppliedSchemaArnsOutput) { op := &request.Operation{ Name: opListAppliedSchemaArns, @@ -3793,7 +3911,7 @@ func (c *CloudDirectory) ListAppliedSchemaArnsRequest(input *ListAppliedSchemaAr // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListAppliedSchemaArns +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListAppliedSchemaArns func (c *CloudDirectory) ListAppliedSchemaArns(input *ListAppliedSchemaArnsInput) (*ListAppliedSchemaArnsOutput, error) { req, out := c.ListAppliedSchemaArnsRequest(input) return out, req.Send() @@ -3890,7 +4008,7 @@ const opListAttachedIndices = "ListAttachedIndices" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListAttachedIndices +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListAttachedIndices func (c *CloudDirectory) ListAttachedIndicesRequest(input *ListAttachedIndicesInput) (req *request.Request, output *ListAttachedIndicesOutput) { op := &request.Operation{ Name: opListAttachedIndices, @@ -3959,7 +4077,7 @@ func (c *CloudDirectory) ListAttachedIndicesRequest(input *ListAttachedIndicesIn // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListAttachedIndices +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListAttachedIndices func (c *CloudDirectory) ListAttachedIndices(input *ListAttachedIndicesInput) (*ListAttachedIndicesOutput, error) { req, out := c.ListAttachedIndicesRequest(input) return out, req.Send() @@ -4056,7 +4174,7 @@ const opListDevelopmentSchemaArns = "ListDevelopmentSchemaArns" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListDevelopmentSchemaArns +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListDevelopmentSchemaArns func (c *CloudDirectory) ListDevelopmentSchemaArnsRequest(input *ListDevelopmentSchemaArnsInput) (req *request.Request, output *ListDevelopmentSchemaArnsOutput) { op := &request.Operation{ Name: opListDevelopmentSchemaArns, @@ -4125,7 +4243,7 @@ func (c *CloudDirectory) ListDevelopmentSchemaArnsRequest(input *ListDevelopment // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListDevelopmentSchemaArns +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListDevelopmentSchemaArns func (c *CloudDirectory) ListDevelopmentSchemaArns(input *ListDevelopmentSchemaArnsInput) (*ListDevelopmentSchemaArnsOutput, error) { req, out := c.ListDevelopmentSchemaArnsRequest(input) return out, req.Send() @@ -4222,7 +4340,7 @@ const opListDirectories = "ListDirectories" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListDirectories +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListDirectories func (c *CloudDirectory) ListDirectoriesRequest(input *ListDirectoriesInput) (req *request.Request, output *ListDirectoriesOutput) { op := &request.Operation{ Name: opListDirectories, @@ -4288,7 +4406,7 @@ func (c *CloudDirectory) ListDirectoriesRequest(input *ListDirectoriesInput) (re // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListDirectories +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListDirectories func (c *CloudDirectory) ListDirectories(input *ListDirectoriesInput) (*ListDirectoriesOutput, error) { req, out := c.ListDirectoriesRequest(input) return out, req.Send() @@ -4385,7 +4503,7 @@ const opListFacetAttributes = "ListFacetAttributes" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListFacetAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListFacetAttributes func (c *CloudDirectory) ListFacetAttributesRequest(input *ListFacetAttributesInput) (req *request.Request, output *ListFacetAttributesOutput) { op := &request.Operation{ Name: opListFacetAttributes, @@ -4457,7 +4575,7 @@ func (c *CloudDirectory) ListFacetAttributesRequest(input *ListFacetAttributesIn // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListFacetAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListFacetAttributes func (c *CloudDirectory) ListFacetAttributes(input *ListFacetAttributesInput) (*ListFacetAttributesOutput, error) { req, out := c.ListFacetAttributesRequest(input) return out, req.Send() @@ -4554,7 +4672,7 @@ const opListFacetNames = "ListFacetNames" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListFacetNames +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListFacetNames func (c *CloudDirectory) ListFacetNamesRequest(input *ListFacetNamesInput) (req *request.Request, output *ListFacetNamesOutput) { op := &request.Operation{ Name: opListFacetNames, @@ -4623,7 +4741,7 @@ func (c *CloudDirectory) ListFacetNamesRequest(input *ListFacetNamesInput) (req // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListFacetNames +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListFacetNames func (c *CloudDirectory) ListFacetNames(input *ListFacetNamesInput) (*ListFacetNamesOutput, error) { req, out := c.ListFacetNamesRequest(input) return out, req.Send() @@ -4720,7 +4838,7 @@ const opListIncomingTypedLinks = "ListIncomingTypedLinks" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListIncomingTypedLinks +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListIncomingTypedLinks func (c *CloudDirectory) ListIncomingTypedLinksRequest(input *ListIncomingTypedLinksInput) (req *request.Request, output *ListIncomingTypedLinksOutput) { op := &request.Operation{ Name: opListIncomingTypedLinks, @@ -4792,7 +4910,7 @@ func (c *CloudDirectory) ListIncomingTypedLinksRequest(input *ListIncomingTypedL // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListIncomingTypedLinks +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListIncomingTypedLinks func (c *CloudDirectory) ListIncomingTypedLinks(input *ListIncomingTypedLinksInput) (*ListIncomingTypedLinksOutput, error) { req, out := c.ListIncomingTypedLinksRequest(input) return out, req.Send() @@ -4839,7 +4957,7 @@ const opListIndex = "ListIndex" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListIndex func (c *CloudDirectory) ListIndexRequest(input *ListIndexInput) (req *request.Request, output *ListIndexOutput) { op := &request.Operation{ Name: opListIndex, @@ -4918,7 +5036,7 @@ func (c *CloudDirectory) ListIndexRequest(input *ListIndexInput) (req *request.R // * ErrCodeNotIndexException "NotIndexException" // Indicates that the requested operation can only operate on index objects. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListIndex +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListIndex func (c *CloudDirectory) ListIndex(input *ListIndexInput) (*ListIndexOutput, error) { req, out := c.ListIndexRequest(input) return out, req.Send() @@ -4990,6 +5108,162 @@ func (c *CloudDirectory) ListIndexPagesWithContext(ctx aws.Context, input *ListI return p.Err() } +const opListManagedSchemaArns = "ListManagedSchemaArns" + +// ListManagedSchemaArnsRequest generates a "aws/request.Request" representing the +// client's request for the ListManagedSchemaArns operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListManagedSchemaArns for more information on using the ListManagedSchemaArns +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListManagedSchemaArnsRequest method. +// req, resp := client.ListManagedSchemaArnsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListManagedSchemaArns +func (c *CloudDirectory) ListManagedSchemaArnsRequest(input *ListManagedSchemaArnsInput) (req *request.Request, output *ListManagedSchemaArnsOutput) { + op := &request.Operation{ + Name: opListManagedSchemaArns, + HTTPMethod: "POST", + HTTPPath: "/amazonclouddirectory/2017-01-11/schema/managed", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListManagedSchemaArnsInput{} + } + + output = &ListManagedSchemaArnsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListManagedSchemaArns API operation for Amazon CloudDirectory. +// +// Lists the major version families of each managed schema. If a major version +// ARN is provided as SchemaArn, the minor version revisions in that family +// are listed instead. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon CloudDirectory's +// API operation ListManagedSchemaArns for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalServiceException "InternalServiceException" +// Indicates a problem that must be resolved by Amazon Web Services. This might +// be a transient error in which case you can retry your request until it succeeds. +// Otherwise, go to the AWS Service Health Dashboard (http://status.aws.amazon.com/) +// site to see if there are any operational issues with the service. +// +// * ErrCodeInvalidArnException "InvalidArnException" +// Indicates that the provided ARN value is not valid. +// +// * ErrCodeValidationException "ValidationException" +// Indicates that your request is malformed in some manner. See the exception +// message. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// Access denied. Check your permissions. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource could not be found. +// +// * ErrCodeInvalidNextTokenException "InvalidNextTokenException" +// Indicates that the NextToken value is not valid. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListManagedSchemaArns +func (c *CloudDirectory) ListManagedSchemaArns(input *ListManagedSchemaArnsInput) (*ListManagedSchemaArnsOutput, error) { + req, out := c.ListManagedSchemaArnsRequest(input) + return out, req.Send() +} + +// ListManagedSchemaArnsWithContext is the same as ListManagedSchemaArns with the addition of +// the ability to pass a context and additional request options. +// +// See ListManagedSchemaArns for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *CloudDirectory) ListManagedSchemaArnsWithContext(ctx aws.Context, input *ListManagedSchemaArnsInput, opts ...request.Option) (*ListManagedSchemaArnsOutput, error) { + req, out := c.ListManagedSchemaArnsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListManagedSchemaArnsPages iterates over the pages of a ListManagedSchemaArns operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListManagedSchemaArns method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListManagedSchemaArns operation. +// pageNum := 0 +// err := client.ListManagedSchemaArnsPages(params, +// func(page *ListManagedSchemaArnsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *CloudDirectory) ListManagedSchemaArnsPages(input *ListManagedSchemaArnsInput, fn func(*ListManagedSchemaArnsOutput, bool) bool) error { + return c.ListManagedSchemaArnsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListManagedSchemaArnsPagesWithContext same as ListManagedSchemaArnsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *CloudDirectory) ListManagedSchemaArnsPagesWithContext(ctx aws.Context, input *ListManagedSchemaArnsInput, fn func(*ListManagedSchemaArnsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListManagedSchemaArnsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListManagedSchemaArnsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListManagedSchemaArnsOutput), !p.HasNextPage()) + } + return p.Err() +} + const opListObjectAttributes = "ListObjectAttributes" // ListObjectAttributesRequest generates a "aws/request.Request" representing the @@ -5015,7 +5289,7 @@ const opListObjectAttributes = "ListObjectAttributes" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectAttributes func (c *CloudDirectory) ListObjectAttributesRequest(input *ListObjectAttributesInput) (req *request.Request, output *ListObjectAttributesOutput) { op := &request.Operation{ Name: opListObjectAttributes, @@ -5091,7 +5365,7 @@ func (c *CloudDirectory) ListObjectAttributesRequest(input *ListObjectAttributes // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectAttributes func (c *CloudDirectory) ListObjectAttributes(input *ListObjectAttributesInput) (*ListObjectAttributesOutput, error) { req, out := c.ListObjectAttributesRequest(input) return out, req.Send() @@ -5188,7 +5462,7 @@ const opListObjectChildren = "ListObjectChildren" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectChildren +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectChildren func (c *CloudDirectory) ListObjectChildrenRequest(input *ListObjectChildrenInput) (req *request.Request, output *ListObjectChildrenOutput) { op := &request.Operation{ Name: opListObjectChildren, @@ -5265,7 +5539,7 @@ func (c *CloudDirectory) ListObjectChildrenRequest(input *ListObjectChildrenInpu // Occurs when any invalid operations are performed on an object that is not // a node, such as calling ListObjectChildren for a leaf node object. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectChildren +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectChildren func (c *CloudDirectory) ListObjectChildren(input *ListObjectChildrenInput) (*ListObjectChildrenOutput, error) { req, out := c.ListObjectChildrenRequest(input) return out, req.Send() @@ -5362,7 +5636,7 @@ const opListObjectParentPaths = "ListObjectParentPaths" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectParentPaths +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectParentPaths func (c *CloudDirectory) ListObjectParentPathsRequest(input *ListObjectParentPathsInput) (req *request.Request, output *ListObjectParentPathsOutput) { op := &request.Operation{ Name: opListObjectParentPaths, @@ -5444,7 +5718,7 @@ func (c *CloudDirectory) ListObjectParentPathsRequest(input *ListObjectParentPat // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectParentPaths +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectParentPaths func (c *CloudDirectory) ListObjectParentPaths(input *ListObjectParentPathsInput) (*ListObjectParentPathsOutput, error) { req, out := c.ListObjectParentPathsRequest(input) return out, req.Send() @@ -5541,7 +5815,7 @@ const opListObjectParents = "ListObjectParents" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectParents +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectParents func (c *CloudDirectory) ListObjectParentsRequest(input *ListObjectParentsInput) (req *request.Request, output *ListObjectParentsOutput) { op := &request.Operation{ Name: opListObjectParents, @@ -5617,7 +5891,7 @@ func (c *CloudDirectory) ListObjectParentsRequest(input *ListObjectParentsInput) // * ErrCodeCannotListParentOfRootException "CannotListParentOfRootException" // Cannot list the parents of a Directory root. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectParents +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectParents func (c *CloudDirectory) ListObjectParents(input *ListObjectParentsInput) (*ListObjectParentsOutput, error) { req, out := c.ListObjectParentsRequest(input) return out, req.Send() @@ -5714,7 +5988,7 @@ const opListObjectPolicies = "ListObjectPolicies" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectPolicies +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectPolicies func (c *CloudDirectory) ListObjectPoliciesRequest(input *ListObjectPoliciesInput) (req *request.Request, output *ListObjectPoliciesOutput) { op := &request.Operation{ Name: opListObjectPolicies, @@ -5786,7 +6060,7 @@ func (c *CloudDirectory) ListObjectPoliciesRequest(input *ListObjectPoliciesInpu // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListObjectPolicies +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListObjectPolicies func (c *CloudDirectory) ListObjectPolicies(input *ListObjectPoliciesInput) (*ListObjectPoliciesOutput, error) { req, out := c.ListObjectPoliciesRequest(input) return out, req.Send() @@ -5883,7 +6157,7 @@ const opListOutgoingTypedLinks = "ListOutgoingTypedLinks" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListOutgoingTypedLinks +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListOutgoingTypedLinks func (c *CloudDirectory) ListOutgoingTypedLinksRequest(input *ListOutgoingTypedLinksInput) (req *request.Request, output *ListOutgoingTypedLinksOutput) { op := &request.Operation{ Name: opListOutgoingTypedLinks, @@ -5955,7 +6229,7 @@ func (c *CloudDirectory) ListOutgoingTypedLinksRequest(input *ListOutgoingTypedL // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListOutgoingTypedLinks +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListOutgoingTypedLinks func (c *CloudDirectory) ListOutgoingTypedLinks(input *ListOutgoingTypedLinksInput) (*ListOutgoingTypedLinksOutput, error) { req, out := c.ListOutgoingTypedLinksRequest(input) return out, req.Send() @@ -6002,7 +6276,7 @@ const opListPolicyAttachments = "ListPolicyAttachments" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListPolicyAttachments +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListPolicyAttachments func (c *CloudDirectory) ListPolicyAttachmentsRequest(input *ListPolicyAttachmentsInput) (req *request.Request, output *ListPolicyAttachmentsOutput) { op := &request.Operation{ Name: opListPolicyAttachments, @@ -6077,7 +6351,7 @@ func (c *CloudDirectory) ListPolicyAttachmentsRequest(input *ListPolicyAttachmen // * ErrCodeNotPolicyException "NotPolicyException" // Indicates that the requested operation can only operate on policy objects. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListPolicyAttachments +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListPolicyAttachments func (c *CloudDirectory) ListPolicyAttachments(input *ListPolicyAttachmentsInput) (*ListPolicyAttachmentsOutput, error) { req, out := c.ListPolicyAttachmentsRequest(input) return out, req.Send() @@ -6174,7 +6448,7 @@ const opListPublishedSchemaArns = "ListPublishedSchemaArns" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListPublishedSchemaArns +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListPublishedSchemaArns func (c *CloudDirectory) ListPublishedSchemaArnsRequest(input *ListPublishedSchemaArnsInput) (req *request.Request, output *ListPublishedSchemaArnsOutput) { op := &request.Operation{ Name: opListPublishedSchemaArns, @@ -6245,7 +6519,7 @@ func (c *CloudDirectory) ListPublishedSchemaArnsRequest(input *ListPublishedSche // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListPublishedSchemaArns +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListPublishedSchemaArns func (c *CloudDirectory) ListPublishedSchemaArns(input *ListPublishedSchemaArnsInput) (*ListPublishedSchemaArnsOutput, error) { req, out := c.ListPublishedSchemaArnsRequest(input) return out, req.Send() @@ -6342,7 +6616,7 @@ const opListTagsForResource = "ListTagsForResource" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListTagsForResource +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListTagsForResource func (c *CloudDirectory) ListTagsForResourceRequest(input *ListTagsForResourceInput) (req *request.Request, output *ListTagsForResourceOutput) { op := &request.Operation{ Name: opListTagsForResource, @@ -6415,7 +6689,7 @@ func (c *CloudDirectory) ListTagsForResourceRequest(input *ListTagsForResourceIn // exist or if you specify a higher number of tags for a resource than the allowed // limit. Allowed limit is 50 tags per resource. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListTagsForResource +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListTagsForResource func (c *CloudDirectory) ListTagsForResource(input *ListTagsForResourceInput) (*ListTagsForResourceOutput, error) { req, out := c.ListTagsForResourceRequest(input) return out, req.Send() @@ -6512,7 +6786,7 @@ const opListTypedLinkFacetAttributes = "ListTypedLinkFacetAttributes" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListTypedLinkFacetAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListTypedLinkFacetAttributes func (c *CloudDirectory) ListTypedLinkFacetAttributesRequest(input *ListTypedLinkFacetAttributesInput) (req *request.Request, output *ListTypedLinkFacetAttributesOutput) { op := &request.Operation{ Name: opListTypedLinkFacetAttributes, @@ -6585,7 +6859,7 @@ func (c *CloudDirectory) ListTypedLinkFacetAttributesRequest(input *ListTypedLin // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListTypedLinkFacetAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListTypedLinkFacetAttributes func (c *CloudDirectory) ListTypedLinkFacetAttributes(input *ListTypedLinkFacetAttributesInput) (*ListTypedLinkFacetAttributesOutput, error) { req, out := c.ListTypedLinkFacetAttributesRequest(input) return out, req.Send() @@ -6682,7 +6956,7 @@ const opListTypedLinkFacetNames = "ListTypedLinkFacetNames" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListTypedLinkFacetNames +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListTypedLinkFacetNames func (c *CloudDirectory) ListTypedLinkFacetNamesRequest(input *ListTypedLinkFacetNamesInput) (req *request.Request, output *ListTypedLinkFacetNamesOutput) { op := &request.Operation{ Name: opListTypedLinkFacetNames, @@ -6752,7 +7026,7 @@ func (c *CloudDirectory) ListTypedLinkFacetNamesRequest(input *ListTypedLinkFace // * ErrCodeInvalidNextTokenException "InvalidNextTokenException" // Indicates that the NextToken value is not valid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/ListTypedLinkFacetNames +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/ListTypedLinkFacetNames func (c *CloudDirectory) ListTypedLinkFacetNames(input *ListTypedLinkFacetNamesInput) (*ListTypedLinkFacetNamesOutput, error) { req, out := c.ListTypedLinkFacetNamesRequest(input) return out, req.Send() @@ -6849,7 +7123,7 @@ const opLookupPolicy = "LookupPolicy" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/LookupPolicy +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/LookupPolicy func (c *CloudDirectory) LookupPolicyRequest(input *LookupPolicyInput) (req *request.Request, output *LookupPolicyOutput) { op := &request.Operation{ Name: opLookupPolicy, @@ -6927,7 +7201,7 @@ func (c *CloudDirectory) LookupPolicyRequest(input *LookupPolicyInput) (req *req // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/LookupPolicy +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/LookupPolicy func (c *CloudDirectory) LookupPolicy(input *LookupPolicyInput) (*LookupPolicyOutput, error) { req, out := c.LookupPolicyRequest(input) return out, req.Send() @@ -7024,7 +7298,7 @@ const opPublishSchema = "PublishSchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/PublishSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/PublishSchema func (c *CloudDirectory) PublishSchemaRequest(input *PublishSchemaInput) (req *request.Request, output *PublishSchemaOutput) { op := &request.Operation{ Name: opPublishSchema, @@ -7088,7 +7362,7 @@ func (c *CloudDirectory) PublishSchemaRequest(input *PublishSchemaInput) (req *r // * ErrCodeSchemaAlreadyPublishedException "SchemaAlreadyPublishedException" // Indicates that a schema is already published. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/PublishSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/PublishSchema func (c *CloudDirectory) PublishSchema(input *PublishSchemaInput) (*PublishSchemaOutput, error) { req, out := c.PublishSchemaRequest(input) return out, req.Send() @@ -7135,7 +7409,7 @@ const opPutSchemaFromJson = "PutSchemaFromJson" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/PutSchemaFromJson +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/PutSchemaFromJson func (c *CloudDirectory) PutSchemaFromJsonRequest(input *PutSchemaFromJsonInput) (req *request.Request, output *PutSchemaFromJsonOutput) { op := &request.Operation{ Name: opPutSchemaFromJson, @@ -7200,7 +7474,7 @@ func (c *CloudDirectory) PutSchemaFromJsonRequest(input *PutSchemaFromJsonInput) // * ErrCodeInvalidRuleException "InvalidRuleException" // Occurs when any of the rule parameter keys or values are invalid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/PutSchemaFromJson +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/PutSchemaFromJson func (c *CloudDirectory) PutSchemaFromJson(input *PutSchemaFromJsonInput) (*PutSchemaFromJsonOutput, error) { req, out := c.PutSchemaFromJsonRequest(input) return out, req.Send() @@ -7247,7 +7521,7 @@ const opRemoveFacetFromObject = "RemoveFacetFromObject" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/RemoveFacetFromObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/RemoveFacetFromObject func (c *CloudDirectory) RemoveFacetFromObjectRequest(input *RemoveFacetFromObjectInput) (req *request.Request, output *RemoveFacetFromObjectOutput) { op := &request.Operation{ Name: opRemoveFacetFromObject, @@ -7314,7 +7588,7 @@ func (c *CloudDirectory) RemoveFacetFromObjectRequest(input *RemoveFacetFromObje // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/RemoveFacetFromObject +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/RemoveFacetFromObject func (c *CloudDirectory) RemoveFacetFromObject(input *RemoveFacetFromObjectInput) (*RemoveFacetFromObjectOutput, error) { req, out := c.RemoveFacetFromObjectRequest(input) return out, req.Send() @@ -7361,7 +7635,7 @@ const opTagResource = "TagResource" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/TagResource +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/TagResource func (c *CloudDirectory) TagResourceRequest(input *TagResourceInput) (req *request.Request, output *TagResourceOutput) { op := &request.Operation{ Name: opTagResource, @@ -7426,7 +7700,7 @@ func (c *CloudDirectory) TagResourceRequest(input *TagResourceInput) (req *reque // exist or if you specify a higher number of tags for a resource than the allowed // limit. Allowed limit is 50 tags per resource. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/TagResource +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/TagResource func (c *CloudDirectory) TagResource(input *TagResourceInput) (*TagResourceOutput, error) { req, out := c.TagResourceRequest(input) return out, req.Send() @@ -7473,7 +7747,7 @@ const opUntagResource = "UntagResource" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UntagResource +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UntagResource func (c *CloudDirectory) UntagResourceRequest(input *UntagResourceInput) (req *request.Request, output *UntagResourceOutput) { op := &request.Operation{ Name: opUntagResource, @@ -7538,7 +7812,7 @@ func (c *CloudDirectory) UntagResourceRequest(input *UntagResourceInput) (req *r // exist or if you specify a higher number of tags for a resource than the allowed // limit. Allowed limit is 50 tags per resource. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UntagResource +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UntagResource func (c *CloudDirectory) UntagResource(input *UntagResourceInput) (*UntagResourceOutput, error) { req, out := c.UntagResourceRequest(input) return out, req.Send() @@ -7585,7 +7859,7 @@ const opUpdateFacet = "UpdateFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateFacet func (c *CloudDirectory) UpdateFacetRequest(input *UpdateFacetInput) (req *request.Request, output *UpdateFacetOutput) { op := &request.Operation{ Name: opUpdateFacet, @@ -7651,6 +7925,10 @@ func (c *CloudDirectory) UpdateFacetRequest(input *UpdateFacetInput) (req *reque // * ErrCodeInvalidFacetUpdateException "InvalidFacetUpdateException" // An attempt to modify a Facet resulted in an invalid schema exception. // +// * ErrCodeFacetValidationException "FacetValidationException" +// The Facet that you provided was not well formed or could not be validated +// with the schema. +// // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // @@ -7660,7 +7938,7 @@ func (c *CloudDirectory) UpdateFacetRequest(input *UpdateFacetInput) (req *reque // * ErrCodeInvalidRuleException "InvalidRuleException" // Occurs when any of the rule parameter keys or values are invalid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateFacet func (c *CloudDirectory) UpdateFacet(input *UpdateFacetInput) (*UpdateFacetOutput, error) { req, out := c.UpdateFacetRequest(input) return out, req.Send() @@ -7682,6 +7960,121 @@ func (c *CloudDirectory) UpdateFacetWithContext(ctx aws.Context, input *UpdateFa return out, req.Send() } +const opUpdateLinkAttributes = "UpdateLinkAttributes" + +// UpdateLinkAttributesRequest generates a "aws/request.Request" representing the +// client's request for the UpdateLinkAttributes operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See UpdateLinkAttributes for more information on using the UpdateLinkAttributes +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the UpdateLinkAttributesRequest method. +// req, resp := client.UpdateLinkAttributesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateLinkAttributes +func (c *CloudDirectory) UpdateLinkAttributesRequest(input *UpdateLinkAttributesInput) (req *request.Request, output *UpdateLinkAttributesOutput) { + op := &request.Operation{ + Name: opUpdateLinkAttributes, + HTTPMethod: "POST", + HTTPPath: "/amazonclouddirectory/2017-01-11/typedlink/attributes/update", + } + + if input == nil { + input = &UpdateLinkAttributesInput{} + } + + output = &UpdateLinkAttributesOutput{} + req = c.newRequest(op, input, output) + return +} + +// UpdateLinkAttributes API operation for Amazon CloudDirectory. +// +// Updates a given typed link’s attributes. Attributes to be updated must not +// contribute to the typed link’s identity, as defined by its IdentityAttributeOrder. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon CloudDirectory's +// API operation UpdateLinkAttributes for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalServiceException "InternalServiceException" +// Indicates a problem that must be resolved by Amazon Web Services. This might +// be a transient error in which case you can retry your request until it succeeds. +// Otherwise, go to the AWS Service Health Dashboard (http://status.aws.amazon.com/) +// site to see if there are any operational issues with the service. +// +// * ErrCodeInvalidArnException "InvalidArnException" +// Indicates that the provided ARN value is not valid. +// +// * ErrCodeRetryableConflictException "RetryableConflictException" +// Occurs when a conflict with a previous successful write is detected. For +// example, if a write operation occurs on an object and then an attempt is +// made to read the object using “SERIALIZABLE” consistency, this exception +// may result. This generally occurs when the previous write did not have time +// to propagate to the host serving the current request. A retry (with appropriate +// backoff logic) is the recommended response to this exception. +// +// * ErrCodeValidationException "ValidationException" +// Indicates that your request is malformed in some manner. See the exception +// message. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// Indicates that limits are exceeded. See Limits (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/limits.html) +// for more information. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// Access denied. Check your permissions. +// +// * ErrCodeDirectoryNotEnabledException "DirectoryNotEnabledException" +// Operations are only permitted on enabled directories. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource could not be found. +// +// * ErrCodeFacetValidationException "FacetValidationException" +// The Facet that you provided was not well formed or could not be validated +// with the schema. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateLinkAttributes +func (c *CloudDirectory) UpdateLinkAttributes(input *UpdateLinkAttributesInput) (*UpdateLinkAttributesOutput, error) { + req, out := c.UpdateLinkAttributesRequest(input) + return out, req.Send() +} + +// UpdateLinkAttributesWithContext is the same as UpdateLinkAttributes with the addition of +// the ability to pass a context and additional request options. +// +// See UpdateLinkAttributes for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *CloudDirectory) UpdateLinkAttributesWithContext(ctx aws.Context, input *UpdateLinkAttributesInput, opts ...request.Option) (*UpdateLinkAttributesOutput, error) { + req, out := c.UpdateLinkAttributesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opUpdateObjectAttributes = "UpdateObjectAttributes" // UpdateObjectAttributesRequest generates a "aws/request.Request" representing the @@ -7707,7 +8100,7 @@ const opUpdateObjectAttributes = "UpdateObjectAttributes" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateObjectAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateObjectAttributes func (c *CloudDirectory) UpdateObjectAttributesRequest(input *UpdateObjectAttributesInput) (req *request.Request, output *UpdateObjectAttributesOutput) { op := &request.Operation{ Name: opUpdateObjectAttributes, @@ -7770,11 +8163,15 @@ func (c *CloudDirectory) UpdateObjectAttributesRequest(input *UpdateObjectAttrib // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // +// * ErrCodeLinkNameAlreadyInUseException "LinkNameAlreadyInUseException" +// Indicates that a link could not be created due to a naming conflict. Choose +// a different name and then try again. +// // * ErrCodeFacetValidationException "FacetValidationException" // The Facet that you provided was not well formed or could not be validated // with the schema. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateObjectAttributes +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateObjectAttributes func (c *CloudDirectory) UpdateObjectAttributes(input *UpdateObjectAttributesInput) (*UpdateObjectAttributesOutput, error) { req, out := c.UpdateObjectAttributesRequest(input) return out, req.Send() @@ -7821,7 +8218,7 @@ const opUpdateSchema = "UpdateSchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateSchema func (c *CloudDirectory) UpdateSchemaRequest(input *UpdateSchemaInput) (req *request.Request, output *UpdateSchemaOutput) { op := &request.Operation{ Name: opUpdateSchema, @@ -7882,7 +8279,7 @@ func (c *CloudDirectory) UpdateSchemaRequest(input *UpdateSchemaInput) (req *req // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The specified resource could not be found. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateSchema func (c *CloudDirectory) UpdateSchema(input *UpdateSchemaInput) (*UpdateSchemaOutput, error) { req, out := c.UpdateSchemaRequest(input) return out, req.Send() @@ -7929,7 +8326,7 @@ const opUpdateTypedLinkFacet = "UpdateTypedLinkFacet" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateTypedLinkFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateTypedLinkFacet func (c *CloudDirectory) UpdateTypedLinkFacetRequest(input *UpdateTypedLinkFacetInput) (req *request.Request, output *UpdateTypedLinkFacetOutput) { op := &request.Operation{ Name: opUpdateTypedLinkFacet, @@ -8002,7 +8399,7 @@ func (c *CloudDirectory) UpdateTypedLinkFacetRequest(input *UpdateTypedLinkFacet // * ErrCodeInvalidRuleException "InvalidRuleException" // Occurs when any of the rule parameter keys or values are invalid. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpdateTypedLinkFacet +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpdateTypedLinkFacet func (c *CloudDirectory) UpdateTypedLinkFacet(input *UpdateTypedLinkFacetInput) (*UpdateTypedLinkFacetOutput, error) { req, out := c.UpdateTypedLinkFacetRequest(input) return out, req.Send() @@ -8049,7 +8446,7 @@ const opUpgradeAppliedSchema = "UpgradeAppliedSchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpgradeAppliedSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpgradeAppliedSchema func (c *CloudDirectory) UpgradeAppliedSchemaRequest(input *UpgradeAppliedSchemaInput) (req *request.Request, output *UpgradeAppliedSchemaOutput) { op := &request.Operation{ Name: opUpgradeAppliedSchema, @@ -8120,7 +8517,11 @@ func (c *CloudDirectory) UpgradeAppliedSchemaRequest(input *UpgradeAppliedSchema // to apply a schema with the same name has occurred. Rename the link or the // schema and then try again. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpgradeAppliedSchema +// * ErrCodeSchemaAlreadyExistsException "SchemaAlreadyExistsException" +// Indicates that a schema could not be created due to a naming conflict. Please +// select a different name and then try again. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpgradeAppliedSchema func (c *CloudDirectory) UpgradeAppliedSchema(input *UpgradeAppliedSchemaInput) (*UpgradeAppliedSchemaOutput, error) { req, out := c.UpgradeAppliedSchemaRequest(input) return out, req.Send() @@ -8167,7 +8568,7 @@ const opUpgradePublishedSchema = "UpgradePublishedSchema" // fmt.Println(resp) // } // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpgradePublishedSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpgradePublishedSchema func (c *CloudDirectory) UpgradePublishedSchemaRequest(input *UpgradePublishedSchemaInput) (req *request.Request, output *UpgradePublishedSchemaOutput) { op := &request.Operation{ Name: opUpgradePublishedSchema, @@ -8238,7 +8639,7 @@ func (c *CloudDirectory) UpgradePublishedSchemaRequest(input *UpgradePublishedSc // Indicates that limits are exceeded. See Limits (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/limits.html) // for more information. // -// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10/UpgradePublishedSchema +// See also, https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11/UpgradePublishedSchema func (c *CloudDirectory) UpgradePublishedSchema(input *UpgradePublishedSchemaInput) (*UpgradePublishedSchemaOutput, error) { req, out := c.UpgradePublishedSchemaRequest(input) return out, req.Send() @@ -10115,6 +10516,89 @@ func (s BatchDetachTypedLinkResponse) GoString() string { return s.String() } +// Retrieves attributes that are associated with a typed link inside a BatchRead +// operation. For more information, see GetLinkAttributes and BatchReadRequest$Operations. +type BatchGetLinkAttributes struct { + _ struct{} `type:"structure"` + + // A list of attribute names whose values will be retrieved. + // + // AttributeNames is a required field + AttributeNames []*string `type:"list" required:"true"` + + // Allows a typed link specifier to be accepted as input. + // + // TypedLinkSpecifier is a required field + TypedLinkSpecifier *TypedLinkSpecifier `type:"structure" required:"true"` +} + +// String returns the string representation +func (s BatchGetLinkAttributes) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s BatchGetLinkAttributes) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *BatchGetLinkAttributes) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "BatchGetLinkAttributes"} + if s.AttributeNames == nil { + invalidParams.Add(request.NewErrParamRequired("AttributeNames")) + } + if s.TypedLinkSpecifier == nil { + invalidParams.Add(request.NewErrParamRequired("TypedLinkSpecifier")) + } + if s.TypedLinkSpecifier != nil { + if err := s.TypedLinkSpecifier.Validate(); err != nil { + invalidParams.AddNested("TypedLinkSpecifier", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAttributeNames sets the AttributeNames field's value. +func (s *BatchGetLinkAttributes) SetAttributeNames(v []*string) *BatchGetLinkAttributes { + s.AttributeNames = v + return s +} + +// SetTypedLinkSpecifier sets the TypedLinkSpecifier field's value. +func (s *BatchGetLinkAttributes) SetTypedLinkSpecifier(v *TypedLinkSpecifier) *BatchGetLinkAttributes { + s.TypedLinkSpecifier = v + return s +} + +// Represents the output of a GetLinkAttributes response operation. +type BatchGetLinkAttributesResponse struct { + _ struct{} `type:"structure"` + + // The attributes that are associated with the typed link. + Attributes []*AttributeKeyAndValue `type:"list"` +} + +// String returns the string representation +func (s BatchGetLinkAttributesResponse) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s BatchGetLinkAttributesResponse) GoString() string { + return s.String() +} + +// SetAttributes sets the Attributes field's value. +func (s *BatchGetLinkAttributesResponse) SetAttributes(v []*AttributeKeyAndValue) *BatchGetLinkAttributesResponse { + s.Attributes = v + return s +} + // Retrieves attributes within a facet that are associated with an object inside // an BatchRead operation. For more information, see GetObjectAttributes and // BatchReadRequest$Operations. @@ -11454,6 +11938,9 @@ func (s *BatchReadInput) SetOperations(v []*BatchReadOperation) *BatchReadInput type BatchReadOperation struct { _ struct{} `type:"structure"` + // Retrieves attributes that are associated with a typed link. + GetLinkAttributes *BatchGetLinkAttributes `type:"structure"` + // Retrieves attributes within a facet that are associated with an object. GetObjectAttributes *BatchGetObjectAttributes `type:"structure"` @@ -11517,6 +12004,11 @@ func (s BatchReadOperation) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *BatchReadOperation) Validate() error { invalidParams := request.ErrInvalidParams{Context: "BatchReadOperation"} + if s.GetLinkAttributes != nil { + if err := s.GetLinkAttributes.Validate(); err != nil { + invalidParams.AddNested("GetLinkAttributes", err.(request.ErrInvalidParams)) + } + } if s.GetObjectAttributes != nil { if err := s.GetObjectAttributes.Validate(); err != nil { invalidParams.AddNested("GetObjectAttributes", err.(request.ErrInvalidParams)) @@ -11584,6 +12076,12 @@ func (s *BatchReadOperation) Validate() error { return nil } +// SetGetLinkAttributes sets the GetLinkAttributes field's value. +func (s *BatchReadOperation) SetGetLinkAttributes(v *BatchGetLinkAttributes) *BatchReadOperation { + s.GetLinkAttributes = v + return s +} + // SetGetObjectAttributes sets the GetObjectAttributes field's value. func (s *BatchReadOperation) SetGetObjectAttributes(v *BatchGetObjectAttributes) *BatchReadOperation { s.GetObjectAttributes = v @@ -11716,6 +12214,9 @@ func (s *BatchReadOutput) SetResponses(v []*BatchReadOperationResponse) *BatchRe type BatchReadSuccessfulResponse struct { _ struct{} `type:"structure"` + // The list of attributes to retrieve from the typed link. + GetLinkAttributes *BatchGetLinkAttributesResponse `type:"structure"` + // Retrieves attributes within a facet that are associated with an object. GetObjectAttributes *BatchGetObjectAttributesResponse `type:"structure"` @@ -11776,6 +12277,12 @@ func (s BatchReadSuccessfulResponse) GoString() string { return s.String() } +// SetGetLinkAttributes sets the GetLinkAttributes field's value. +func (s *BatchReadSuccessfulResponse) SetGetLinkAttributes(v *BatchGetLinkAttributesResponse) *BatchReadSuccessfulResponse { + s.GetLinkAttributes = v + return s +} + // SetGetObjectAttributes sets the GetObjectAttributes field's value. func (s *BatchReadSuccessfulResponse) SetGetObjectAttributes(v *BatchGetObjectAttributesResponse) *BatchReadSuccessfulResponse { s.GetObjectAttributes = v @@ -11921,6 +12428,92 @@ func (s BatchRemoveFacetFromObjectResponse) GoString() string { return s.String() } +// Updates a given typed link’s attributes inside a BatchRead operation. Attributes +// to be updated must not contribute to the typed link’s identity, as defined +// by its IdentityAttributeOrder. For more information, see UpdateLinkAttributes +// and BatchReadRequest$Operations. +type BatchUpdateLinkAttributes struct { + _ struct{} `type:"structure"` + + // The attributes update structure. + // + // AttributeUpdates is a required field + AttributeUpdates []*LinkAttributeUpdate `type:"list" required:"true"` + + // Allows a typed link specifier to be accepted as input. + // + // TypedLinkSpecifier is a required field + TypedLinkSpecifier *TypedLinkSpecifier `type:"structure" required:"true"` +} + +// String returns the string representation +func (s BatchUpdateLinkAttributes) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s BatchUpdateLinkAttributes) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *BatchUpdateLinkAttributes) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "BatchUpdateLinkAttributes"} + if s.AttributeUpdates == nil { + invalidParams.Add(request.NewErrParamRequired("AttributeUpdates")) + } + if s.TypedLinkSpecifier == nil { + invalidParams.Add(request.NewErrParamRequired("TypedLinkSpecifier")) + } + if s.AttributeUpdates != nil { + for i, v := range s.AttributeUpdates { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AttributeUpdates", i), err.(request.ErrInvalidParams)) + } + } + } + if s.TypedLinkSpecifier != nil { + if err := s.TypedLinkSpecifier.Validate(); err != nil { + invalidParams.AddNested("TypedLinkSpecifier", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAttributeUpdates sets the AttributeUpdates field's value. +func (s *BatchUpdateLinkAttributes) SetAttributeUpdates(v []*LinkAttributeUpdate) *BatchUpdateLinkAttributes { + s.AttributeUpdates = v + return s +} + +// SetTypedLinkSpecifier sets the TypedLinkSpecifier field's value. +func (s *BatchUpdateLinkAttributes) SetTypedLinkSpecifier(v *TypedLinkSpecifier) *BatchUpdateLinkAttributes { + s.TypedLinkSpecifier = v + return s +} + +// Represents the output of a UpdateLinkAttributes response operation. +type BatchUpdateLinkAttributesResponse struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s BatchUpdateLinkAttributesResponse) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s BatchUpdateLinkAttributesResponse) GoString() string { + return s.String() +} + // Represents the output of a BatchUpdate operation. type BatchUpdateObjectAttributes struct { _ struct{} `type:"structure"` @@ -12118,6 +12711,9 @@ type BatchWriteOperation struct { // A batch operation that removes a facet from an object. RemoveFacetFromObject *BatchRemoveFacetFromObject `type:"structure"` + // Updates a given object's attributes. + UpdateLinkAttributes *BatchUpdateLinkAttributes `type:"structure"` + // Updates a given object's attributes. UpdateObjectAttributes *BatchUpdateObjectAttributes `type:"structure"` } @@ -12200,6 +12796,11 @@ func (s *BatchWriteOperation) Validate() error { invalidParams.AddNested("RemoveFacetFromObject", err.(request.ErrInvalidParams)) } } + if s.UpdateLinkAttributes != nil { + if err := s.UpdateLinkAttributes.Validate(); err != nil { + invalidParams.AddNested("UpdateLinkAttributes", err.(request.ErrInvalidParams)) + } + } if s.UpdateObjectAttributes != nil { if err := s.UpdateObjectAttributes.Validate(); err != nil { invalidParams.AddNested("UpdateObjectAttributes", err.(request.ErrInvalidParams)) @@ -12290,6 +12891,12 @@ func (s *BatchWriteOperation) SetRemoveFacetFromObject(v *BatchRemoveFacetFromOb return s } +// SetUpdateLinkAttributes sets the UpdateLinkAttributes field's value. +func (s *BatchWriteOperation) SetUpdateLinkAttributes(v *BatchUpdateLinkAttributes) *BatchWriteOperation { + s.UpdateLinkAttributes = v + return s +} + // SetUpdateObjectAttributes sets the UpdateObjectAttributes field's value. func (s *BatchWriteOperation) SetUpdateObjectAttributes(v *BatchUpdateObjectAttributes) *BatchWriteOperation { s.UpdateObjectAttributes = v @@ -12343,6 +12950,9 @@ type BatchWriteOperationResponse struct { // The result of a batch remove facet from object operation. RemoveFacetFromObject *BatchRemoveFacetFromObjectResponse `type:"structure"` + // Represents the output of a BatchWrite response operation. + UpdateLinkAttributes *BatchUpdateLinkAttributesResponse `type:"structure"` + // Updates a given object’s attributes. UpdateObjectAttributes *BatchUpdateObjectAttributesResponse `type:"structure"` } @@ -12435,6 +13045,12 @@ func (s *BatchWriteOperationResponse) SetRemoveFacetFromObject(v *BatchRemoveFac return s } +// SetUpdateLinkAttributes sets the UpdateLinkAttributes field's value. +func (s *BatchWriteOperationResponse) SetUpdateLinkAttributes(v *BatchUpdateLinkAttributesResponse) *BatchWriteOperationResponse { + s.UpdateLinkAttributes = v + return s +} + // SetUpdateObjectAttributes sets the UpdateObjectAttributes field's value. func (s *BatchWriteOperationResponse) SetUpdateObjectAttributes(v *BatchUpdateObjectAttributesResponse) *BatchWriteOperationResponse { s.UpdateObjectAttributes = v @@ -12587,6 +13203,11 @@ type CreateFacetInput struct { // The attributes that are associated with the Facet. Attributes []*FacetAttribute `type:"list"` + // There are two different styles that you can define on any given facet, Static + // and Dynamic. For static facets, all attributes must be defined in the schema. + // For dynamic facets, attributes can be defined during data plane operations. + FacetStyle *string `type:"string" enum:"FacetStyle"` + // The name of the Facet, which is unique for a given schema. // // Name is a required field @@ -12603,9 +13224,7 @@ type CreateFacetInput struct { // information, see Policies (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/cd_key_concepts.html#policies). // // * Index: Can be created with the Index API. - // - // ObjectType is a required field - ObjectType *string `type:"string" required:"true" enum:"ObjectType"` + ObjectType *string `type:"string" enum:"ObjectType"` // The schema ARN in which the new Facet will be created. For more information, // see arns. @@ -12633,9 +13252,6 @@ func (s *CreateFacetInput) Validate() error { if s.Name != nil && len(*s.Name) < 1 { invalidParams.Add(request.NewErrParamMinLen("Name", 1)) } - if s.ObjectType == nil { - invalidParams.Add(request.NewErrParamRequired("ObjectType")) - } if s.SchemaArn == nil { invalidParams.Add(request.NewErrParamRequired("SchemaArn")) } @@ -12662,6 +13278,12 @@ func (s *CreateFacetInput) SetAttributes(v []*FacetAttribute) *CreateFacetInput return s } +// SetFacetStyle sets the FacetStyle field's value. +func (s *CreateFacetInput) SetFacetStyle(v string) *CreateFacetInput { + s.FacetStyle = &v + return s +} + // SetName sets the Name field's value. func (s *CreateFacetInput) SetName(v string) *CreateFacetInput { s.Name = &v @@ -13933,6 +14555,11 @@ func (s *EnableDirectoryOutput) SetDirectoryArn(v string) *EnableDirectoryOutput type Facet struct { _ struct{} `type:"structure"` + // There are two different styles that you can define on any given facet, Static + // and Dynamic. For static facets, all attributes must be defined in the schema. + // For dynamic facets, attributes can be defined during data plane operations. + FacetStyle *string `type:"string" enum:"FacetStyle"` + // The name of the Facet. Name *string `min:"1" type:"string"` @@ -13951,6 +14578,12 @@ func (s Facet) GoString() string { return s.String() } +// SetFacetStyle sets the FacetStyle field's value. +func (s *Facet) SetFacetStyle(v string) *Facet { + s.FacetStyle = &v + return s +} + // SetName sets the Name field's value. func (s *Facet) SetName(v string) *Facet { s.Name = &v @@ -14429,6 +15062,110 @@ func (s *GetFacetOutput) SetFacet(v *Facet) *GetFacetOutput { return s } +type GetLinkAttributesInput struct { + _ struct{} `type:"structure"` + + // A list of attribute names whose values will be retrieved. + // + // AttributeNames is a required field + AttributeNames []*string `type:"list" required:"true"` + + // The consistency level at which to retrieve the attributes on a typed link. + ConsistencyLevel *string `type:"string" enum:"ConsistencyLevel"` + + // The Amazon Resource Name (ARN) that is associated with the Directory where + // the typed link resides. For more information, see arns or Typed link (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/objectsandlinks.html#typedlink). + // + // DirectoryArn is a required field + DirectoryArn *string `location:"header" locationName:"x-amz-data-partition" type:"string" required:"true"` + + // Allows a typed link specifier to be accepted as input. + // + // TypedLinkSpecifier is a required field + TypedLinkSpecifier *TypedLinkSpecifier `type:"structure" required:"true"` +} + +// String returns the string representation +func (s GetLinkAttributesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetLinkAttributesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *GetLinkAttributesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "GetLinkAttributesInput"} + if s.AttributeNames == nil { + invalidParams.Add(request.NewErrParamRequired("AttributeNames")) + } + if s.DirectoryArn == nil { + invalidParams.Add(request.NewErrParamRequired("DirectoryArn")) + } + if s.TypedLinkSpecifier == nil { + invalidParams.Add(request.NewErrParamRequired("TypedLinkSpecifier")) + } + if s.TypedLinkSpecifier != nil { + if err := s.TypedLinkSpecifier.Validate(); err != nil { + invalidParams.AddNested("TypedLinkSpecifier", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAttributeNames sets the AttributeNames field's value. +func (s *GetLinkAttributesInput) SetAttributeNames(v []*string) *GetLinkAttributesInput { + s.AttributeNames = v + return s +} + +// SetConsistencyLevel sets the ConsistencyLevel field's value. +func (s *GetLinkAttributesInput) SetConsistencyLevel(v string) *GetLinkAttributesInput { + s.ConsistencyLevel = &v + return s +} + +// SetDirectoryArn sets the DirectoryArn field's value. +func (s *GetLinkAttributesInput) SetDirectoryArn(v string) *GetLinkAttributesInput { + s.DirectoryArn = &v + return s +} + +// SetTypedLinkSpecifier sets the TypedLinkSpecifier field's value. +func (s *GetLinkAttributesInput) SetTypedLinkSpecifier(v *TypedLinkSpecifier) *GetLinkAttributesInput { + s.TypedLinkSpecifier = v + return s +} + +type GetLinkAttributesOutput struct { + _ struct{} `type:"structure"` + + // The attributes that are associated with the typed link. + Attributes []*AttributeKeyAndValue `type:"list"` +} + +// String returns the string representation +func (s GetLinkAttributesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetLinkAttributesOutput) GoString() string { + return s.String() +} + +// SetAttributes sets the Attributes field's value. +func (s *GetLinkAttributesOutput) SetAttributes(v []*AttributeKeyAndValue) *GetLinkAttributesOutput { + s.Attributes = v + return s +} + type GetObjectAttributesInput struct { _ struct{} `type:"structure"` @@ -14832,6 +15569,88 @@ func (s *IndexAttachment) SetObjectIdentifier(v string) *IndexAttachment { return s } +// The action to take on a typed link attribute value. Updates are only supported +// for attributes which don’t contribute to link identity. +type LinkAttributeAction struct { + _ struct{} `type:"structure"` + + // A type that can be either UPDATE_OR_CREATE or DELETE. + AttributeActionType *string `type:"string" enum:"UpdateActionType"` + + // The value that you want to update to. + AttributeUpdateValue *TypedAttributeValue `type:"structure"` +} + +// String returns the string representation +func (s LinkAttributeAction) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s LinkAttributeAction) GoString() string { + return s.String() +} + +// SetAttributeActionType sets the AttributeActionType field's value. +func (s *LinkAttributeAction) SetAttributeActionType(v string) *LinkAttributeAction { + s.AttributeActionType = &v + return s +} + +// SetAttributeUpdateValue sets the AttributeUpdateValue field's value. +func (s *LinkAttributeAction) SetAttributeUpdateValue(v *TypedAttributeValue) *LinkAttributeAction { + s.AttributeUpdateValue = v + return s +} + +// Structure that contains attribute update information. +type LinkAttributeUpdate struct { + _ struct{} `type:"structure"` + + // The action to perform as part of the attribute update. + AttributeAction *LinkAttributeAction `type:"structure"` + + // The key of the attribute being updated. + AttributeKey *AttributeKey `type:"structure"` +} + +// String returns the string representation +func (s LinkAttributeUpdate) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s LinkAttributeUpdate) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *LinkAttributeUpdate) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "LinkAttributeUpdate"} + if s.AttributeKey != nil { + if err := s.AttributeKey.Validate(); err != nil { + invalidParams.AddNested("AttributeKey", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAttributeAction sets the AttributeAction field's value. +func (s *LinkAttributeUpdate) SetAttributeAction(v *LinkAttributeAction) *LinkAttributeUpdate { + s.AttributeAction = v + return s +} + +// SetAttributeKey sets the AttributeKey field's value. +func (s *LinkAttributeUpdate) SetAttributeKey(v *AttributeKey) *LinkAttributeUpdate { + s.AttributeKey = v + return s +} + type ListAppliedSchemaArnsInput struct { _ struct{} `type:"structure"` @@ -15700,6 +16519,93 @@ func (s *ListIndexOutput) SetNextToken(v string) *ListIndexOutput { return s } +type ListManagedSchemaArnsInput struct { + _ struct{} `type:"structure"` + + // The maximum number of results to retrieve. + MaxResults *int64 `min:"1" type:"integer"` + + // The pagination token. + NextToken *string `type:"string"` + + // The response for ListManagedSchemaArns. When this parameter is used, all + // minor version ARNs for a major version are listed. + SchemaArn *string `type:"string"` +} + +// String returns the string representation +func (s ListManagedSchemaArnsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListManagedSchemaArnsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListManagedSchemaArnsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListManagedSchemaArnsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListManagedSchemaArnsInput) SetMaxResults(v int64) *ListManagedSchemaArnsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListManagedSchemaArnsInput) SetNextToken(v string) *ListManagedSchemaArnsInput { + s.NextToken = &v + return s +} + +// SetSchemaArn sets the SchemaArn field's value. +func (s *ListManagedSchemaArnsInput) SetSchemaArn(v string) *ListManagedSchemaArnsInput { + s.SchemaArn = &v + return s +} + +type ListManagedSchemaArnsOutput struct { + _ struct{} `type:"structure"` + + // The pagination token. + NextToken *string `type:"string"` + + // The ARNs for all AWS managed schemas. + SchemaArns []*string `type:"list"` +} + +// String returns the string representation +func (s ListManagedSchemaArnsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListManagedSchemaArnsOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *ListManagedSchemaArnsOutput) SetNextToken(v string) *ListManagedSchemaArnsOutput { + s.NextToken = &v + return s +} + +// SetSchemaArns sets the SchemaArns field's value. +func (s *ListManagedSchemaArnsOutput) SetSchemaArns(v []*string) *ListManagedSchemaArnsOutput { + s.SchemaArns = v + return s +} + type ListObjectAttributesInput struct { _ struct{} `type:"structure"` @@ -18539,6 +19445,103 @@ func (s UpdateFacetOutput) GoString() string { return s.String() } +type UpdateLinkAttributesInput struct { + _ struct{} `type:"structure"` + + // The attributes update structure. + // + // AttributeUpdates is a required field + AttributeUpdates []*LinkAttributeUpdate `type:"list" required:"true"` + + // The Amazon Resource Name (ARN) that is associated with the Directory where + // the updated typed link resides. For more information, see arns or Typed link + // (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/objectsandlinks.html#typedlink). + // + // DirectoryArn is a required field + DirectoryArn *string `location:"header" locationName:"x-amz-data-partition" type:"string" required:"true"` + + // Allows a typed link specifier to be accepted as input. + // + // TypedLinkSpecifier is a required field + TypedLinkSpecifier *TypedLinkSpecifier `type:"structure" required:"true"` +} + +// String returns the string representation +func (s UpdateLinkAttributesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateLinkAttributesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *UpdateLinkAttributesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "UpdateLinkAttributesInput"} + if s.AttributeUpdates == nil { + invalidParams.Add(request.NewErrParamRequired("AttributeUpdates")) + } + if s.DirectoryArn == nil { + invalidParams.Add(request.NewErrParamRequired("DirectoryArn")) + } + if s.TypedLinkSpecifier == nil { + invalidParams.Add(request.NewErrParamRequired("TypedLinkSpecifier")) + } + if s.AttributeUpdates != nil { + for i, v := range s.AttributeUpdates { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AttributeUpdates", i), err.(request.ErrInvalidParams)) + } + } + } + if s.TypedLinkSpecifier != nil { + if err := s.TypedLinkSpecifier.Validate(); err != nil { + invalidParams.AddNested("TypedLinkSpecifier", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAttributeUpdates sets the AttributeUpdates field's value. +func (s *UpdateLinkAttributesInput) SetAttributeUpdates(v []*LinkAttributeUpdate) *UpdateLinkAttributesInput { + s.AttributeUpdates = v + return s +} + +// SetDirectoryArn sets the DirectoryArn field's value. +func (s *UpdateLinkAttributesInput) SetDirectoryArn(v string) *UpdateLinkAttributesInput { + s.DirectoryArn = &v + return s +} + +// SetTypedLinkSpecifier sets the TypedLinkSpecifier field's value. +func (s *UpdateLinkAttributesInput) SetTypedLinkSpecifier(v *TypedLinkSpecifier) *UpdateLinkAttributesInput { + s.TypedLinkSpecifier = v + return s +} + +type UpdateLinkAttributesOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s UpdateLinkAttributesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateLinkAttributesOutput) GoString() string { + return s.String() +} + type UpdateObjectAttributesInput struct { _ struct{} `type:"structure"` @@ -19162,6 +20165,17 @@ const ( // FacetAttributeTypeDatetime is a FacetAttributeType enum value FacetAttributeTypeDatetime = "DATETIME" + + // FacetAttributeTypeVariant is a FacetAttributeType enum value + FacetAttributeTypeVariant = "VARIANT" +) + +const ( + // FacetStyleStatic is a FacetStyle enum value + FacetStyleStatic = "STATIC" + + // FacetStyleDynamic is a FacetStyle enum value + FacetStyleDynamic = "DYNAMIC" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/clouddirectoryiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/clouddirectoryiface/interface.go index 94f80f98f..266cf63de 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/clouddirectoryiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/clouddirectoryiface/interface.go @@ -172,6 +172,10 @@ type CloudDirectoryAPI interface { GetFacetWithContext(aws.Context, *clouddirectory.GetFacetInput, ...request.Option) (*clouddirectory.GetFacetOutput, error) GetFacetRequest(*clouddirectory.GetFacetInput) (*request.Request, *clouddirectory.GetFacetOutput) + GetLinkAttributes(*clouddirectory.GetLinkAttributesInput) (*clouddirectory.GetLinkAttributesOutput, error) + GetLinkAttributesWithContext(aws.Context, *clouddirectory.GetLinkAttributesInput, ...request.Option) (*clouddirectory.GetLinkAttributesOutput, error) + GetLinkAttributesRequest(*clouddirectory.GetLinkAttributesInput) (*request.Request, *clouddirectory.GetLinkAttributesOutput) + GetObjectAttributes(*clouddirectory.GetObjectAttributesInput) (*clouddirectory.GetObjectAttributesOutput, error) GetObjectAttributesWithContext(aws.Context, *clouddirectory.GetObjectAttributesInput, ...request.Option) (*clouddirectory.GetObjectAttributesOutput, error) GetObjectAttributesRequest(*clouddirectory.GetObjectAttributesInput) (*request.Request, *clouddirectory.GetObjectAttributesOutput) @@ -241,6 +245,13 @@ type CloudDirectoryAPI interface { ListIndexPages(*clouddirectory.ListIndexInput, func(*clouddirectory.ListIndexOutput, bool) bool) error ListIndexPagesWithContext(aws.Context, *clouddirectory.ListIndexInput, func(*clouddirectory.ListIndexOutput, bool) bool, ...request.Option) error + ListManagedSchemaArns(*clouddirectory.ListManagedSchemaArnsInput) (*clouddirectory.ListManagedSchemaArnsOutput, error) + ListManagedSchemaArnsWithContext(aws.Context, *clouddirectory.ListManagedSchemaArnsInput, ...request.Option) (*clouddirectory.ListManagedSchemaArnsOutput, error) + ListManagedSchemaArnsRequest(*clouddirectory.ListManagedSchemaArnsInput) (*request.Request, *clouddirectory.ListManagedSchemaArnsOutput) + + ListManagedSchemaArnsPages(*clouddirectory.ListManagedSchemaArnsInput, func(*clouddirectory.ListManagedSchemaArnsOutput, bool) bool) error + ListManagedSchemaArnsPagesWithContext(aws.Context, *clouddirectory.ListManagedSchemaArnsInput, func(*clouddirectory.ListManagedSchemaArnsOutput, bool) bool, ...request.Option) error + ListObjectAttributes(*clouddirectory.ListObjectAttributesInput) (*clouddirectory.ListObjectAttributesOutput, error) ListObjectAttributesWithContext(aws.Context, *clouddirectory.ListObjectAttributesInput, ...request.Option) (*clouddirectory.ListObjectAttributesOutput, error) ListObjectAttributesRequest(*clouddirectory.ListObjectAttributesInput) (*request.Request, *clouddirectory.ListObjectAttributesOutput) @@ -346,6 +357,10 @@ type CloudDirectoryAPI interface { UpdateFacetWithContext(aws.Context, *clouddirectory.UpdateFacetInput, ...request.Option) (*clouddirectory.UpdateFacetOutput, error) UpdateFacetRequest(*clouddirectory.UpdateFacetInput) (*request.Request, *clouddirectory.UpdateFacetOutput) + UpdateLinkAttributes(*clouddirectory.UpdateLinkAttributesInput) (*clouddirectory.UpdateLinkAttributesOutput, error) + UpdateLinkAttributesWithContext(aws.Context, *clouddirectory.UpdateLinkAttributesInput, ...request.Option) (*clouddirectory.UpdateLinkAttributesOutput, error) + UpdateLinkAttributesRequest(*clouddirectory.UpdateLinkAttributesInput) (*request.Request, *clouddirectory.UpdateLinkAttributesOutput) + UpdateObjectAttributes(*clouddirectory.UpdateObjectAttributesInput) (*clouddirectory.UpdateObjectAttributesOutput, error) UpdateObjectAttributesWithContext(aws.Context, *clouddirectory.UpdateObjectAttributesInput, ...request.Option) (*clouddirectory.UpdateObjectAttributesOutput, error) UpdateObjectAttributesRequest(*clouddirectory.UpdateObjectAttributesInput) (*request.Request, *clouddirectory.UpdateObjectAttributesOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/doc.go b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/doc.go index b5fa52d51..387682de0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/doc.go +++ b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/doc.go @@ -10,7 +10,7 @@ // about AWS Directory Services features, see AWS Directory Service (https://aws.amazon.com/directoryservice/) // and the AWS Directory Service Administration Guide (http://docs.aws.amazon.com/directoryservice/latest/admin-guide/what_is.html). // -// See https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2016-05-10 for more information on this service. +// See https://docs.aws.amazon.com/goto/WebAPI/clouddirectory-2017-01-11 for more information on this service. // // See clouddirectory package documentation for more information. // https://docs.aws.amazon.com/sdk-for-go/api/service/clouddirectory/ diff --git a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/service.go b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/service.go index e0ce4ba6c..9748f97a7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/clouddirectory/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "clouddirectory" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "clouddirectory" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudDirectory" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudDirectory client with a session. @@ -58,10 +59,11 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, - APIVersion: "2016-05-10", + APIVersion: "2017-01-11", }, handlers, ), diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudformation/api.go b/vendor/github.com/aws/aws-sdk-go/service/cloudformation/api.go index 673b3636c..3cd918390 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudformation/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudformation/api.go @@ -3618,7 +3618,8 @@ func (c *CloudFormation) UpdateStackSetRequest(input *UpdateStackSetInput) (req // UpdateStackSet API operation for AWS CloudFormation. // -// Updates the stack set and all associated stack instances. +// Updates the stack set, and associated stack instances in the specified accounts +// and regions. // // Even if the stack set operation created by updating the stack set fails (completely // or partially, below or above a specified failure tolerance), the stack set @@ -3650,6 +3651,9 @@ func (c *CloudFormation) UpdateStackSetRequest(input *UpdateStackSetInput) (req // * ErrCodeInvalidOperationException "InvalidOperationException" // The specified operation isn't valid. // +// * ErrCodeStackInstanceNotFoundException "StackInstanceNotFoundException" +// The specified stack instance doesn't exist. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/cloudformation-2010-05-15/UpdateStackSet func (c *CloudFormation) UpdateStackSet(input *UpdateStackSetInput) (*UpdateStackSetOutput, error) { req, out := c.UpdateStackSetRequest(input) @@ -5130,8 +5134,8 @@ type CreateStackSetInput struct { // // Specify an IAM role only if you are using customized administrator roles // to control which users or groups can manage specific stack sets within the - // same administrator account. For more information, see Define Permissions - // for Multiple Administrators (http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/stacksets-prereqs.html) + // same administrator account. For more information, see Prerequisites: Granting + // Permissions for Stack Set Operations (http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/stacksets-prereqs.html) // in the AWS CloudFormation User Guide. AdministrationRoleARN *string `min:"20" type:"string"` @@ -5184,6 +5188,14 @@ type CreateStackSetInput struct { // stack set's purpose or other important information. Description *string `min:"1" type:"string"` + // The name of the IAM execution role to use to create the stack set. If you + // do not specify an execution role, AWS CloudFormation uses the AWSCloudFormationStackSetExecutionRole + // role for the stack set operation. + // + // Specify an IAM role only if you are using customized execution roles to control + // which stack resources users and groups can include in their stack sets. + ExecutionRoleName *string `min:"1" type:"string"` + // The input parameters for the stack set template. Parameters []*Parameter `type:"list"` @@ -5248,6 +5260,9 @@ func (s *CreateStackSetInput) Validate() error { if s.Description != nil && len(*s.Description) < 1 { invalidParams.Add(request.NewErrParamMinLen("Description", 1)) } + if s.ExecutionRoleName != nil && len(*s.ExecutionRoleName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ExecutionRoleName", 1)) + } if s.StackSetName == nil { invalidParams.Add(request.NewErrParamRequired("StackSetName")) } @@ -5298,6 +5313,12 @@ func (s *CreateStackSetInput) SetDescription(v string) *CreateStackSetInput { return s } +// SetExecutionRoleName sets the ExecutionRoleName field's value. +func (s *CreateStackSetInput) SetExecutionRoleName(v string) *CreateStackSetInput { + s.ExecutionRoleName = &v + return s +} + // SetParameters sets the Parameters field's value. func (s *CreateStackSetInput) SetParameters(v []*Parameter) *CreateStackSetInput { s.Parameters = v @@ -9765,7 +9786,7 @@ type StackSet struct { // // Use customized administrator roles to control which users or groups can manage // specific stack sets within the same administrator account. For more information, - // see Define Permissions for Multiple Administrators (http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/stacksets-prereqs.html) + // see Prerequisites: Granting Permissions for Stack Set Operations (http://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/stacksets-prereqs.html) // in the AWS CloudFormation User Guide. AdministrationRoleARN *string `min:"20" type:"string"` @@ -9780,6 +9801,12 @@ type StackSet struct { // or updated. Description *string `min:"1" type:"string"` + // The name of the IAM execution role used to create or update the stack set. + // + // Use customized execution roles to control which stack resources users and + // groups can include in their stack sets. + ExecutionRoleName *string `min:"1" type:"string"` + // A list of input parameters for a stack set. Parameters []*Parameter `type:"list"` @@ -9832,6 +9859,12 @@ func (s *StackSet) SetDescription(v string) *StackSet { return s } +// SetExecutionRoleName sets the ExecutionRoleName field's value. +func (s *StackSet) SetExecutionRoleName(v string) *StackSet { + s.ExecutionRoleName = &v + return s +} + // SetParameters sets the Parameters field's value. func (s *StackSet) SetParameters(v []*Parameter) *StackSet { s.Parameters = v @@ -9905,6 +9938,12 @@ type StackSetOperation struct { // set operation was successful, or even attempted, in each account or region. EndTimestamp *time.Time `type:"timestamp" timestampFormat:"iso8601"` + // The name of the IAM execution role used to create or update the stack set. + // + // Use customized execution roles to control which stack resources users and + // groups can include in their stack sets. + ExecutionRoleName *string `min:"1" type:"string"` + // The unique ID of a stack set operation. OperationId *string `min:"1" type:"string"` @@ -9976,6 +10015,12 @@ func (s *StackSetOperation) SetEndTimestamp(v time.Time) *StackSetOperation { return s } +// SetExecutionRoleName sets the ExecutionRoleName field's value. +func (s *StackSetOperation) SetExecutionRoleName(v string) *StackSetOperation { + s.ExecutionRoleName = &v + return s +} + // SetOperationId sets the OperationId field's value. func (s *StackSetOperation) SetOperationId(v string) *StackSetOperation { s.OperationId = &v @@ -11168,6 +11213,23 @@ func (s *UpdateStackOutput) SetStackId(v string) *UpdateStackOutput { type UpdateStackSetInput struct { _ struct{} `type:"structure"` + // The accounts in which to update associated stack instances. If you specify + // accounts, you must also specify the regions in which to update stack set + // instances. + // + // To update all the stack instances associated with this stack set, do not + // specify the Accounts or Regions properties. + // + // If the stack set update includes changes to the template (that is, if the + // TemplateBody or TemplateURL properties are specified), or the Parameters + // property, AWS CloudFormation marks all stack instances with a status of OUTDATED + // prior to updating the stack instances in the specified accounts and regions. + // If the stack set update does not include changes to the template or parameters, + // AWS CloudFormation updates the stack instances in the specified accounts + // and regions, while leaving all other stack instances with their existing + // stack instance status. + Accounts []*string `type:"list"` + // The Amazon Resource Number (ARN) of the IAM role to use to update this stack // set. // @@ -11223,6 +11285,20 @@ type UpdateStackSetInput struct { // A brief description of updates that you are making. Description *string `min:"1" type:"string"` + // The name of the IAM execution role to use to update the stack set. If you + // do not specify an execution role, AWS CloudFormation uses the AWSCloudFormationStackSetExecutionRole + // role for the stack set operation. + // + // Specify an IAM role only if you are using customized execution roles to control + // which stack resources users and groups can include in their stack sets. + // + // If you specify a customized execution role, AWS CloudFormation uses that + // role to update the stack. If you do not specify a customized execution role, + // AWS CloudFormation performs the update using the role previously associated + // with the stack set, so long as you have permissions to perform operations + // on the stack set. + ExecutionRoleName *string `min:"1" type:"string"` + // The unique ID for this stack set operation. // // The operation ID also functions as an idempotency token, to ensure that AWS @@ -11242,6 +11318,22 @@ type UpdateStackSetInput struct { // A list of input parameters for the stack set template. Parameters []*Parameter `type:"list"` + // The regions in which to update associated stack instances. If you specify + // regions, you must also specify accounts in which to update stack set instances. + // + // To update all the stack instances associated with this stack set, do not + // specify the Accounts or Regions properties. + // + // If the stack set update includes changes to the template (that is, if the + // TemplateBody or TemplateURL properties are specified), or the Parameters + // property, AWS CloudFormation marks all stack instances with a status of OUTDATED + // prior to updating the stack instances in the specified accounts and regions. + // If the stack set update does not include changes to the template or parameters, + // AWS CloudFormation updates the stack instances in the specified accounts + // and regions, while leaving all other stack instances with their existing + // stack instance status. + Regions []*string `type:"list"` + // The name or unique ID of the stack set that you want to update. // // StackSetName is a required field @@ -11321,6 +11413,9 @@ func (s *UpdateStackSetInput) Validate() error { if s.Description != nil && len(*s.Description) < 1 { invalidParams.Add(request.NewErrParamMinLen("Description", 1)) } + if s.ExecutionRoleName != nil && len(*s.ExecutionRoleName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ExecutionRoleName", 1)) + } if s.OperationId != nil && len(*s.OperationId) < 1 { invalidParams.Add(request.NewErrParamMinLen("OperationId", 1)) } @@ -11355,6 +11450,12 @@ func (s *UpdateStackSetInput) Validate() error { return nil } +// SetAccounts sets the Accounts field's value. +func (s *UpdateStackSetInput) SetAccounts(v []*string) *UpdateStackSetInput { + s.Accounts = v + return s +} + // SetAdministrationRoleARN sets the AdministrationRoleARN field's value. func (s *UpdateStackSetInput) SetAdministrationRoleARN(v string) *UpdateStackSetInput { s.AdministrationRoleARN = &v @@ -11373,6 +11474,12 @@ func (s *UpdateStackSetInput) SetDescription(v string) *UpdateStackSetInput { return s } +// SetExecutionRoleName sets the ExecutionRoleName field's value. +func (s *UpdateStackSetInput) SetExecutionRoleName(v string) *UpdateStackSetInput { + s.ExecutionRoleName = &v + return s +} + // SetOperationId sets the OperationId field's value. func (s *UpdateStackSetInput) SetOperationId(v string) *UpdateStackSetInput { s.OperationId = &v @@ -11391,6 +11498,12 @@ func (s *UpdateStackSetInput) SetParameters(v []*Parameter) *UpdateStackSetInput return s } +// SetRegions sets the Regions field's value. +func (s *UpdateStackSetInput) SetRegions(v []*string) *UpdateStackSetInput { + s.Regions = v + return s +} + // SetStackSetName sets the StackSetName field's value. func (s *UpdateStackSetInput) SetStackSetName(v string) *UpdateStackSetInput { s.StackSetName = &v diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudformation/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudformation/service.go index 0115c5bb0..65df49a0c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudformation/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudformation/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudformation" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudformation" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudFormation" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudFormation client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/api.go b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/api.go index afc3cf2ea..3c98c555e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/api.go @@ -72,7 +72,7 @@ func (c *CloudFront) CreateCloudFrontOriginAccessIdentityRequest(input *CreateCl // API operation CreateCloudFrontOriginAccessIdentity for usage and error information. // // Returned Error Codes: -// * ErrCodeOriginAccessIdentityAlreadyExists "OriginAccessIdentityAlreadyExists" +// * ErrCodeOriginAccessIdentityAlreadyExists "CloudFrontOriginAccessIdentityAlreadyExists" // If the CallerReference is a value you already sent in a previous request // to create an identity but the content of the CloudFrontOriginAccessIdentityConfig // is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists @@ -1239,7 +1239,7 @@ func (c *CloudFront) DeleteCloudFrontOriginAccessIdentityRequest(input *DeleteCl // The precondition given in one or more of the request-header fields evaluated // to false. // -// * ErrCodeOriginAccessIdentityInUse "OriginAccessIdentityInUse" +// * ErrCodeOriginAccessIdentityInUse "CloudFrontOriginAccessIdentityInUse" // // See also, https://docs.aws.amazon.com/goto/WebAPI/cloudfront-2017-10-30/DeleteCloudFrontOriginAccessIdentity func (c *CloudFront) DeleteCloudFrontOriginAccessIdentity(input *DeleteCloudFrontOriginAccessIdentityInput) (*DeleteCloudFrontOriginAccessIdentityOutput, error) { @@ -1638,92 +1638,6 @@ func (c *CloudFront) DeletePublicKeyWithContext(ctx aws.Context, input *DeletePu return out, req.Send() } -const opDeleteServiceLinkedRole = "DeleteServiceLinkedRole2017_10_30" - -// DeleteServiceLinkedRoleRequest generates a "aws/request.Request" representing the -// client's request for the DeleteServiceLinkedRole operation. The "output" return -// value will be populated with the request's response once the request completes -// successfuly. -// -// Use "Send" method on the returned Request to send the API call to the service. -// the "output" return value is not valid until after Send returns without error. -// -// See DeleteServiceLinkedRole for more information on using the DeleteServiceLinkedRole -// API call, and error handling. -// -// This method is useful when you want to inject custom logic or configuration -// into the SDK's request lifecycle. Such as custom headers, or retry logic. -// -// -// // Example sending a request using the DeleteServiceLinkedRoleRequest method. -// req, resp := client.DeleteServiceLinkedRoleRequest(params) -// -// err := req.Send() -// if err == nil { // resp is now filled -// fmt.Println(resp) -// } -// -// See also, https://docs.aws.amazon.com/goto/WebAPI/cloudfront-2017-10-30/DeleteServiceLinkedRole -func (c *CloudFront) DeleteServiceLinkedRoleRequest(input *DeleteServiceLinkedRoleInput) (req *request.Request, output *DeleteServiceLinkedRoleOutput) { - op := &request.Operation{ - Name: opDeleteServiceLinkedRole, - HTTPMethod: "DELETE", - HTTPPath: "/2017-10-30/service-linked-role/{RoleName}", - } - - if input == nil { - input = &DeleteServiceLinkedRoleInput{} - } - - output = &DeleteServiceLinkedRoleOutput{} - req = c.newRequest(op, input, output) - req.Handlers.Unmarshal.Remove(restxml.UnmarshalHandler) - req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) - return -} - -// DeleteServiceLinkedRole API operation for Amazon CloudFront. -// -// Returns awserr.Error for service API and SDK errors. Use runtime type assertions -// with awserr.Error's Code and Message methods to get detailed information about -// the error. -// -// See the AWS API reference guide for Amazon CloudFront's -// API operation DeleteServiceLinkedRole for usage and error information. -// -// Returned Error Codes: -// * ErrCodeInvalidArgument "InvalidArgument" -// The argument is invalid. -// -// * ErrCodeAccessDenied "AccessDenied" -// Access denied. -// -// * ErrCodeResourceInUse "ResourceInUse" -// -// * ErrCodeNoSuchResource "NoSuchResource" -// -// See also, https://docs.aws.amazon.com/goto/WebAPI/cloudfront-2017-10-30/DeleteServiceLinkedRole -func (c *CloudFront) DeleteServiceLinkedRole(input *DeleteServiceLinkedRoleInput) (*DeleteServiceLinkedRoleOutput, error) { - req, out := c.DeleteServiceLinkedRoleRequest(input) - return out, req.Send() -} - -// DeleteServiceLinkedRoleWithContext is the same as DeleteServiceLinkedRole with the addition of -// the ability to pass a context and additional request options. -// -// See DeleteServiceLinkedRole for details on how to use this API operation. -// -// The context must be non-nil and will be used for request cancellation. If -// the context is nil a panic will occur. In the future the SDK may create -// sub-contexts for http.Requests. See https://golang.org/pkg/context/ -// for more information on using Contexts. -func (c *CloudFront) DeleteServiceLinkedRoleWithContext(ctx aws.Context, input *DeleteServiceLinkedRoleInput, opts ...request.Option) (*DeleteServiceLinkedRoleOutput, error) { - req, out := c.DeleteServiceLinkedRoleRequest(input) - req.SetContext(ctx) - req.ApplyOptions(opts...) - return out, req.Send() -} - const opDeleteStreamingDistribution = "DeleteStreamingDistribution2017_10_30" // DeleteStreamingDistributionRequest generates a "aws/request.Request" representing the @@ -7551,56 +7465,6 @@ func (s DeletePublicKeyOutput) GoString() string { return s.String() } -type DeleteServiceLinkedRoleInput struct { - _ struct{} `type:"structure"` - - // RoleName is a required field - RoleName *string `location:"uri" locationName:"RoleName" type:"string" required:"true"` -} - -// String returns the string representation -func (s DeleteServiceLinkedRoleInput) String() string { - return awsutil.Prettify(s) -} - -// GoString returns the string representation -func (s DeleteServiceLinkedRoleInput) GoString() string { - return s.String() -} - -// Validate inspects the fields of the type to determine if they are valid. -func (s *DeleteServiceLinkedRoleInput) Validate() error { - invalidParams := request.ErrInvalidParams{Context: "DeleteServiceLinkedRoleInput"} - if s.RoleName == nil { - invalidParams.Add(request.NewErrParamRequired("RoleName")) - } - - if invalidParams.Len() > 0 { - return invalidParams - } - return nil -} - -// SetRoleName sets the RoleName field's value. -func (s *DeleteServiceLinkedRoleInput) SetRoleName(v string) *DeleteServiceLinkedRoleInput { - s.RoleName = &v - return s -} - -type DeleteServiceLinkedRoleOutput struct { - _ struct{} `type:"structure"` -} - -// String returns the string representation -func (s DeleteServiceLinkedRoleOutput) String() string { - return awsutil.Prettify(s) -} - -// GoString returns the string representation -func (s DeleteServiceLinkedRoleOutput) GoString() string { - return s.String() -} - // The request to delete a streaming distribution. type DeleteStreamingDistributionInput struct { _ struct{} `type:"structure"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/cloudfrontiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/cloudfrontiface/interface.go index 472ebe5ad..a3c1c0036 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/cloudfrontiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/cloudfrontiface/interface.go @@ -116,10 +116,6 @@ type CloudFrontAPI interface { DeletePublicKeyWithContext(aws.Context, *cloudfront.DeletePublicKeyInput, ...request.Option) (*cloudfront.DeletePublicKeyOutput, error) DeletePublicKeyRequest(*cloudfront.DeletePublicKeyInput) (*request.Request, *cloudfront.DeletePublicKeyOutput) - DeleteServiceLinkedRole(*cloudfront.DeleteServiceLinkedRoleInput) (*cloudfront.DeleteServiceLinkedRoleOutput, error) - DeleteServiceLinkedRoleWithContext(aws.Context, *cloudfront.DeleteServiceLinkedRoleInput, ...request.Option) (*cloudfront.DeleteServiceLinkedRoleOutput, error) - DeleteServiceLinkedRoleRequest(*cloudfront.DeleteServiceLinkedRoleInput) (*request.Request, *cloudfront.DeleteServiceLinkedRoleOutput) - DeleteStreamingDistribution(*cloudfront.DeleteStreamingDistributionInput) (*cloudfront.DeleteStreamingDistributionOutput, error) DeleteStreamingDistributionWithContext(aws.Context, *cloudfront.DeleteStreamingDistributionInput, ...request.Option) (*cloudfront.DeleteStreamingDistributionOutput, error) DeleteStreamingDistributionRequest(*cloudfront.DeleteStreamingDistributionInput) (*request.Request, *cloudfront.DeleteStreamingDistributionOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/errors.go b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/errors.go index 1f5ce88f1..ede41494f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/errors.go @@ -264,17 +264,17 @@ const ( ErrCodeNoSuchStreamingDistribution = "NoSuchStreamingDistribution" // ErrCodeOriginAccessIdentityAlreadyExists for service response error code - // "OriginAccessIdentityAlreadyExists". + // "CloudFrontOriginAccessIdentityAlreadyExists". // // If the CallerReference is a value you already sent in a previous request // to create an identity but the content of the CloudFrontOriginAccessIdentityConfig // is different from the original request, CloudFront returns a CloudFrontOriginAccessIdentityAlreadyExists // error. - ErrCodeOriginAccessIdentityAlreadyExists = "OriginAccessIdentityAlreadyExists" + ErrCodeOriginAccessIdentityAlreadyExists = "CloudFrontOriginAccessIdentityAlreadyExists" // ErrCodeOriginAccessIdentityInUse for service response error code - // "OriginAccessIdentityInUse". - ErrCodeOriginAccessIdentityInUse = "OriginAccessIdentityInUse" + // "CloudFrontOriginAccessIdentityInUse". + ErrCodeOriginAccessIdentityInUse = "CloudFrontOriginAccessIdentityInUse" // ErrCodePreconditionFailed for service response error code // "PreconditionFailed". @@ -301,10 +301,6 @@ const ( // No profile specified for the field-level encryption query argument. ErrCodeQueryArgProfileEmpty = "QueryArgProfileEmpty" - // ErrCodeResourceInUse for service response error code - // "ResourceInUse". - ErrCodeResourceInUse = "ResourceInUse" - // ErrCodeStreamingDistributionAlreadyExists for service response error code // "StreamingDistributionAlreadyExists". ErrCodeStreamingDistributionAlreadyExists = "StreamingDistributionAlreadyExists" diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/service.go index a07f02113..103bab6a2 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudfront/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudfront/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudfront" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudfront" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudFront" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudFront client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudhsm/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudhsm/service.go index 6e499a568..4c948a31f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudhsm/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudhsm/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudhsm" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudhsm" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudHSM" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudHSM client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudhsmv2/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudhsmv2/service.go index 62ed752a3..c86db6ae7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudhsmv2/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudhsmv2/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudhsmv2" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudhsmv2" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudHSM V2" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudHSMV2 client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudsearch/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudsearch/service.go index 6ef7a62d4..850bc1370 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudsearch/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudsearch/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudsearch" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudsearch" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudSearch" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudSearch client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudsearchdomain/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudsearchdomain/service.go index af75b1eaf..65200bffc 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudsearchdomain/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudsearchdomain/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudsearchdomain" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudsearchdomain" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudSearch Domain" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudSearchDomain client with a session. @@ -63,6 +64,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/api.go b/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/api.go index c6ad246ec..c815f8a8b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/api.go @@ -73,7 +73,7 @@ func (c *CloudTrail) AddTagsRequest(input *AddTagsInput) (req *request.Request, // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // This exception is thrown when the specified resource is not found. // -// * ErrCodeARNInvalidException "ARNInvalidException" +// * ErrCodeARNInvalidException "CloudTrailARNInvalidException" // This exception is thrown when an operation is called with an invalid trail // ARN. The format of a trail ARN is: // @@ -842,7 +842,7 @@ func (c *CloudTrail) ListTagsRequest(input *ListTagsInput) (req *request.Request // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // This exception is thrown when the specified resource is not found. // -// * ErrCodeARNInvalidException "ARNInvalidException" +// * ErrCodeARNInvalidException "CloudTrailARNInvalidException" // This exception is thrown when an operation is called with an invalid trail // ARN. The format of a trail ARN is: // @@ -1271,7 +1271,7 @@ func (c *CloudTrail) RemoveTagsRequest(input *RemoveTagsInput) (req *request.Req // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // This exception is thrown when the specified resource is not found. // -// * ErrCodeARNInvalidException "ARNInvalidException" +// * ErrCodeARNInvalidException "CloudTrailARNInvalidException" // This exception is thrown when an operation is called with an invalid trail // ARN. The format of a trail ARN is: // diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/errors.go b/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/errors.go index 0da999b33..0ee616cbe 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/errors.go @@ -5,13 +5,13 @@ package cloudtrail const ( // ErrCodeARNInvalidException for service response error code - // "ARNInvalidException". + // "CloudTrailARNInvalidException". // // This exception is thrown when an operation is called with an invalid trail // ARN. The format of a trail ARN is: // // arn:aws:cloudtrail:us-east-1:123456789012:trail/MyTrail - ErrCodeARNInvalidException = "ARNInvalidException" + ErrCodeARNInvalidException = "CloudTrailARNInvalidException" // ErrCodeCloudWatchLogsDeliveryUnavailableException for service response error code // "CloudWatchLogsDeliveryUnavailableException". diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/service.go index 49198fc63..5f1d4ddc4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudtrail/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cloudtrail" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cloudtrail" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudTrail" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudTrail client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudwatch/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudwatch/service.go index 4b0aa76ed..0d4786622 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudwatch/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudwatch/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "monitoring" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "monitoring" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudWatch" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudWatch client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudwatchevents/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudwatchevents/service.go index 1fb24f16b..2a7f6969c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudwatchevents/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudwatchevents/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "events" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "events" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudWatch Events" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudWatchEvents client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/api.go b/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/api.go index 43a77742f..7d05f0826 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/api.go @@ -1281,6 +1281,7 @@ func (c *CloudWatchLogs) DescribeDestinationsPages(input *DescribeDestinationsIn // for more information on using Contexts. func (c *CloudWatchLogs) DescribeDestinationsPagesWithContext(ctx aws.Context, input *DescribeDestinationsInput, fn func(*DescribeDestinationsOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *DescribeDestinationsInput if input != nil { @@ -1503,6 +1504,7 @@ func (c *CloudWatchLogs) DescribeLogGroupsPages(input *DescribeLogGroupsInput, f // for more information on using Contexts. func (c *CloudWatchLogs) DescribeLogGroupsPagesWithContext(ctx aws.Context, input *DescribeLogGroupsInput, fn func(*DescribeLogGroupsOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *DescribeLogGroupsInput if input != nil { @@ -1649,6 +1651,7 @@ func (c *CloudWatchLogs) DescribeLogStreamsPages(input *DescribeLogStreamsInput, // for more information on using Contexts. func (c *CloudWatchLogs) DescribeLogStreamsPagesWithContext(ctx aws.Context, input *DescribeLogStreamsInput, fn func(*DescribeLogStreamsOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *DescribeLogStreamsInput if input != nil { @@ -1792,6 +1795,7 @@ func (c *CloudWatchLogs) DescribeMetricFiltersPages(input *DescribeMetricFilters // for more information on using Contexts. func (c *CloudWatchLogs) DescribeMetricFiltersPagesWithContext(ctx aws.Context, input *DescribeMetricFiltersInput, fn func(*DescribeMetricFiltersOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *DescribeMetricFiltersInput if input != nil { @@ -2017,6 +2021,7 @@ func (c *CloudWatchLogs) DescribeSubscriptionFiltersPages(input *DescribeSubscri // for more information on using Contexts. func (c *CloudWatchLogs) DescribeSubscriptionFiltersPagesWithContext(ctx aws.Context, input *DescribeSubscriptionFiltersInput, fn func(*DescribeSubscriptionFiltersOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *DescribeSubscriptionFiltersInput if input != nil { @@ -2264,6 +2269,7 @@ func (c *CloudWatchLogs) FilterLogEventsPages(input *FilterLogEventsInput, fn fu // for more information on using Contexts. func (c *CloudWatchLogs) FilterLogEventsPagesWithContext(ctx aws.Context, input *FilterLogEventsInput, fn func(*FilterLogEventsOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *FilterLogEventsInput if input != nil { @@ -2410,6 +2416,7 @@ func (c *CloudWatchLogs) GetLogEventsPages(input *GetLogEventsInput, fn func(*Ge // for more information on using Contexts. func (c *CloudWatchLogs) GetLogEventsPagesWithContext(ctx aws.Context, input *GetLogEventsInput, fn func(*GetLogEventsOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ + EndPageOnSameToken: true, NewRequest: func() (*request.Request, error) { var inCpy *GetLogEventsInput if input != nil { diff --git a/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/service.go b/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/service.go index 8e6094d58..8d5f929df 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cloudwatchlogs/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "logs" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "logs" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CloudWatch Logs" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CloudWatchLogs client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/codebuild/api.go b/vendor/github.com/aws/aws-sdk-go/service/codebuild/api.go index 644ff0df7..816aff014 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codebuild/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codebuild/api.go @@ -4757,6 +4757,9 @@ const ( ) const ( + // EnvironmentTypeWindowsContainer is a EnvironmentType enum value + EnvironmentTypeWindowsContainer = "WINDOWS_CONTAINER" + // EnvironmentTypeLinuxContainer is a EnvironmentType enum value EnvironmentTypeLinuxContainer = "LINUX_CONTAINER" ) @@ -4807,6 +4810,9 @@ const ( // PlatformTypeUbuntu is a PlatformType enum value PlatformTypeUbuntu = "UBUNTU" + + // PlatformTypeWindowsServer is a PlatformType enum value + PlatformTypeWindowsServer = "WINDOWS_SERVER" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/codebuild/service.go b/vendor/github.com/aws/aws-sdk-go/service/codebuild/service.go index d45aed633..b9ff2b6f7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codebuild/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codebuild/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "codebuild" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "codebuild" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CodeBuild" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CodeBuild client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/codecommit/service.go b/vendor/github.com/aws/aws-sdk-go/service/codecommit/service.go index cb25eaba7..c8cad394a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codecommit/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codecommit/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "codecommit" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "codecommit" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CodeCommit" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CodeCommit client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/codedeploy/api.go b/vendor/github.com/aws/aws-sdk-go/service/codedeploy/api.go index d5169b081..4c3a3c8d7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codedeploy/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codedeploy/api.go @@ -5105,7 +5105,8 @@ type BlueInstanceTerminationOption struct { Action *string `locationName:"action" type:"string" enum:"InstanceAction"` // The number of minutes to wait after a successful blue/green deployment before - // terminating instances from the original environment. + // terminating instances from the original environment. The maximum setting + // is 2880 minutes (2 days). TerminationWaitTimeInMinutes *int64 `locationName:"terminationWaitTimeInMinutes" type:"integer"` } @@ -8760,11 +8761,15 @@ type LoadBalancerInfo struct { // An array containing information about the load balancer to use for load balancing // in a deployment. In Elastic Load Balancing, load balancers are used with // Classic Load Balancers. + // + // Adding more than one load balancer to the array is not supported. ElbInfoList []*ELBInfo `locationName:"elbInfoList" type:"list"` // An array containing information about the target group to use for load balancing // in a deployment. In Elastic Load Balancing, target groups are used with Application // Load Balancers. + // + // Adding more than one target group to the array is not supported. TargetGroupInfoList []*TargetGroupInfo `locationName:"targetGroupInfoList" type:"list"` } diff --git a/vendor/github.com/aws/aws-sdk-go/service/codedeploy/service.go b/vendor/github.com/aws/aws-sdk-go/service/codedeploy/service.go index e6524d04d..dd1eaf413 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codedeploy/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codedeploy/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "codedeploy" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "codedeploy" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CodeDeploy" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CodeDeploy client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/codepipeline/api.go b/vendor/github.com/aws/aws-sdk-go/service/codepipeline/api.go index d308cfb94..466db4a05 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codepipeline/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codepipeline/api.go @@ -2809,6 +2809,10 @@ func (c *CodePipeline) UpdatePipelineRequest(input *UpdatePipelineInput) (req *r // * ErrCodeInvalidStructureException "InvalidStructureException" // The specified structure was specified in an invalid format. // +// * ErrCodeLimitExceededException "LimitExceededException" +// The number of pipelines associated with the AWS account has exceeded the +// limit allowed for the account. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/codepipeline-2015-07-09/UpdatePipeline func (c *CodePipeline) UpdatePipeline(input *UpdatePipelineInput) (*UpdatePipelineOutput, error) { req, out := c.UpdatePipelineRequest(input) diff --git a/vendor/github.com/aws/aws-sdk-go/service/codepipeline/service.go b/vendor/github.com/aws/aws-sdk-go/service/codepipeline/service.go index 8b7737abb..397a00f9e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codepipeline/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codepipeline/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "codepipeline" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "codepipeline" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CodePipeline" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CodePipeline client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/codestar/service.go b/vendor/github.com/aws/aws-sdk-go/service/codestar/service.go index e2a7bc762..5c0a3fb51 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/codestar/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/codestar/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "codestar" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "codestar" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "CodeStar" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CodeStar client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cognitoidentity/service.go b/vendor/github.com/aws/aws-sdk-go/service/cognitoidentity/service.go index ee82b18ce..9ebae103f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cognitoidentity/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cognitoidentity/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cognito-identity" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cognito-identity" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Cognito Identity" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CognitoIdentity client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/api.go b/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/api.go index 739f5f15a..0a86e0c2e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/api.go @@ -9911,7 +9911,8 @@ func (c *CognitoIdentityProvider) VerifySoftwareTokenRequest(input *VerifySoftwa // VerifySoftwareToken API operation for Amazon Cognito Identity Provider. // // Use this API to register a user's entered TOTP code and mark the user's software -// token MFA status as "verified" if successful, +// token MFA status as "verified" if successful. The request takes an access +// token or a session string, but not both. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -11751,6 +11752,13 @@ type AdminInitiateAuthOutput struct { // is returned to you in the AdminInitiateAuth response if you need to pass // another challenge. // + // * MFA_SETUP: If MFA is required, users who do not have at least one of + // the MFA methods set up are presented with an MFA_SETUP challenge. The + // user must set up at least one MFA type to continue to authenticate. + // + // * SELECT_MFA_TYPE: Selects the MFA type. Valid MFA options are SMS_MFA + // for text SMS MFA, and SOFTWARE_TOKEN_MFA for TOTP software token MFA. + // // * SMS_MFA: Next challenge is to supply an SMS_MFA_CODE, delivered via // SMS. // @@ -13555,7 +13563,7 @@ type AuthenticationResultType struct { // The access token. AccessToken *string `type:"string"` - // The expiration period of the authentication result. + // The expiration period of the authentication result in seconds. ExpiresIn *int64 `type:"integer"` // The ID token. @@ -14851,7 +14859,19 @@ type CreateUserPoolClientInput struct { // user pool. AnalyticsConfiguration *AnalyticsConfigurationType `type:"structure"` - // A list of allowed callback URLs for the identity providers. + // A list of allowed redirect (callback) URLs for the identity providers. + // + // A redirect URI must: + // + // * Be an absolute URI. + // + // * Be registered with the authorization server. + // + // * Not use HTTP without TLS (i.e. use HTTPS instead of HTTP). + // + // * Not include a fragment component. + // + // See OAuth 2.0 - Redirection Endpoint (https://tools.ietf.org/html/rfc6749#section-3.1.2). CallbackURLs []*string `type:"list"` // The client name for the user pool client you would like to create. @@ -14860,6 +14880,18 @@ type CreateUserPoolClientInput struct { ClientName *string `min:"1" type:"string" required:"true"` // The default redirect URI. Must be in the CallbackURLs list. + // + // A redirect URI must: + // + // * Be an absolute URI. + // + // * Be registered with the authorization server. + // + // * Not use HTTP without TLS (i.e. use HTTPS instead of HTTP). + // + // * Not include a fragment component. + // + // See OAuth 2.0 - Redirection Endpoint (https://tools.ietf.org/html/rfc6749#section-3.1.2). DefaultRedirectURI *string `min:"1" type:"string"` // The explicit authentication flows. @@ -20488,7 +20520,7 @@ type SchemaAttributeType struct { // Specifies whether the attribute type is developer only. DeveloperOnlyAttribute *bool `type:"boolean"` - // Specifies whether the attribute can be changed once it has been created. + // Specifies whether the value of the attribute can be changed. Mutable *bool `type:"boolean"` // A schema attribute of the name type. @@ -22344,7 +22376,19 @@ type UpdateUserPoolClientInput struct { // user pool. AnalyticsConfiguration *AnalyticsConfigurationType `type:"structure"` - // A list of allowed callback URLs for the identity providers. + // A list of allowed redirect (callback) URLs for the identity providers. + // + // A redirect URI must: + // + // * Be an absolute URI. + // + // * Be registered with the authorization server. + // + // * Not use HTTP without TLS (i.e. use HTTPS instead of HTTP). + // + // * Not include a fragment component. + // + // See OAuth 2.0 - Redirection Endpoint (https://tools.ietf.org/html/rfc6749#section-3.1.2). CallbackURLs []*string `type:"list"` // The ID of the client associated with the user pool. @@ -22356,6 +22400,18 @@ type UpdateUserPoolClientInput struct { ClientName *string `min:"1" type:"string"` // The default redirect URI. Must be in the CallbackURLs list. + // + // A redirect URI must: + // + // * Be an absolute URI. + // + // * Be registered with the authorization server. + // + // * Not use HTTP without TLS (i.e. use HTTPS instead of HTTP). + // + // * Not include a fragment component. + // + // See OAuth 2.0 - Redirection Endpoint (https://tools.ietf.org/html/rfc6749#section-3.1.2). DefaultRedirectURI *string `min:"1" type:"string"` // Explicit authentication flows. @@ -23084,7 +23140,19 @@ type UserPoolClientType struct { // The Amazon Pinpoint analytics configuration for the user pool client. AnalyticsConfiguration *AnalyticsConfigurationType `type:"structure"` - // A list of allowed callback URLs for the identity providers. + // A list of allowed redirect (callback) URLs for the identity providers. + // + // A redirect URI must: + // + // * Be an absolute URI. + // + // * Be registered with the authorization server. + // + // * Not use HTTP without TLS (i.e. use HTTPS instead of HTTP). + // + // * Not include a fragment component. + // + // See OAuth 2.0 - Redirection Endpoint (https://tools.ietf.org/html/rfc6749#section-3.1.2). CallbackURLs []*string `type:"list"` // The ID of the client associated with the user pool. @@ -23100,6 +23168,18 @@ type UserPoolClientType struct { CreationDate *time.Time `type:"timestamp" timestampFormat:"unix"` // The default redirect URI. Must be in the CallbackURLs list. + // + // A redirect URI must: + // + // * Be an absolute URI. + // + // * Be registered with the authorization server. + // + // * Not use HTTP without TLS (i.e. use HTTPS instead of HTTP). + // + // * Not include a fragment component. + // + // See OAuth 2.0 - Redirection Endpoint (https://tools.ietf.org/html/rfc6749#section-3.1.2). DefaultRedirectURI *string `min:"1" type:"string"` // The explicit authentication flows. @@ -23365,6 +23445,9 @@ type UserPoolType struct { // Specifies the attributes that are aliased in a user pool. AliasAttributes []*string `type:"list"` + // The Amazon Resource Name (ARN) for the user pool. + Arn *string `min:"20" type:"string"` + // Specifies the attributes that are auto-verified in a user pool. AutoVerifiedAttributes []*string `type:"list"` @@ -23395,7 +23478,7 @@ type UserPoolType struct { // The ID of the user pool. Id *string `min:"1" type:"string"` - // The AWS Lambda triggers associated with tue user pool. + // The AWS Lambda triggers associated with the user pool. LambdaConfig *LambdaConfigType `type:"structure"` // The date the user pool was last modified. @@ -23473,6 +23556,12 @@ func (s *UserPoolType) SetAliasAttributes(v []*string) *UserPoolType { return s } +// SetArn sets the Arn field's value. +func (s *UserPoolType) SetArn(v string) *UserPoolType { + s.Arn = &v + return s +} + // SetAutoVerifiedAttributes sets the AutoVerifiedAttributes field's value. func (s *UserPoolType) SetAutoVerifiedAttributes(v []*string) *UserPoolType { s.AutoVerifiedAttributes = v @@ -24228,6 +24317,9 @@ const ( // IdentityProviderTypeTypeLoginWithAmazon is a IdentityProviderTypeType enum value IdentityProviderTypeTypeLoginWithAmazon = "LoginWithAmazon" + + // IdentityProviderTypeTypeOidc is a IdentityProviderTypeType enum value + IdentityProviderTypeTypeOidc = "OIDC" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/service.go b/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/service.go index 190d20711..68efbd80b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cognitoidentityprovider/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cognito-idp" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cognito-idp" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Cognito Identity Provider" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CognitoIdentityProvider client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/cognitosync/service.go b/vendor/github.com/aws/aws-sdk-go/service/cognitosync/service.go index 1094e5416..a35cbbd5a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/cognitosync/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/cognitosync/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cognito-sync" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cognito-sync" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Cognito Sync" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CognitoSync client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/comprehend/api.go b/vendor/github.com/aws/aws-sdk-go/service/comprehend/api.go index 3d331d350..7924f4991 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/comprehend/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/comprehend/api.go @@ -145,8 +145,8 @@ func (c *Comprehend) BatchDetectEntitiesRequest(input *BatchDetectEntitiesInput) // BatchDetectEntities API operation for Amazon Comprehend. // -// Inspects the text of a batch of documents and returns information about them. -// For more information about entities, see how-entities +// Inspects the text of a batch of documents for named entities and returns +// information about them. For more information about named entities, see how-entities // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -388,6 +388,362 @@ func (c *Comprehend) BatchDetectSentimentWithContext(ctx aws.Context, input *Bat return out, req.Send() } +const opDescribeDominantLanguageDetectionJob = "DescribeDominantLanguageDetectionJob" + +// DescribeDominantLanguageDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDominantLanguageDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDominantLanguageDetectionJob for more information on using the DescribeDominantLanguageDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDominantLanguageDetectionJobRequest method. +// req, resp := client.DescribeDominantLanguageDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeDominantLanguageDetectionJob +func (c *Comprehend) DescribeDominantLanguageDetectionJobRequest(input *DescribeDominantLanguageDetectionJobInput) (req *request.Request, output *DescribeDominantLanguageDetectionJobOutput) { + op := &request.Operation{ + Name: opDescribeDominantLanguageDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDominantLanguageDetectionJobInput{} + } + + output = &DescribeDominantLanguageDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDominantLanguageDetectionJob API operation for Amazon Comprehend. +// +// Gets the properties associated with a dominant language detection job. Use +// this operation to get the status of a detection job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation DescribeDominantLanguageDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeDominantLanguageDetectionJob +func (c *Comprehend) DescribeDominantLanguageDetectionJob(input *DescribeDominantLanguageDetectionJobInput) (*DescribeDominantLanguageDetectionJobOutput, error) { + req, out := c.DescribeDominantLanguageDetectionJobRequest(input) + return out, req.Send() +} + +// DescribeDominantLanguageDetectionJobWithContext is the same as DescribeDominantLanguageDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDominantLanguageDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) DescribeDominantLanguageDetectionJobWithContext(ctx aws.Context, input *DescribeDominantLanguageDetectionJobInput, opts ...request.Option) (*DescribeDominantLanguageDetectionJobOutput, error) { + req, out := c.DescribeDominantLanguageDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeEntitiesDetectionJob = "DescribeEntitiesDetectionJob" + +// DescribeEntitiesDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEntitiesDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEntitiesDetectionJob for more information on using the DescribeEntitiesDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEntitiesDetectionJobRequest method. +// req, resp := client.DescribeEntitiesDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeEntitiesDetectionJob +func (c *Comprehend) DescribeEntitiesDetectionJobRequest(input *DescribeEntitiesDetectionJobInput) (req *request.Request, output *DescribeEntitiesDetectionJobOutput) { + op := &request.Operation{ + Name: opDescribeEntitiesDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeEntitiesDetectionJobInput{} + } + + output = &DescribeEntitiesDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEntitiesDetectionJob API operation for Amazon Comprehend. +// +// Gets the properties associated with an entities detection job. Use this operation +// to get the status of a detection job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation DescribeEntitiesDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeEntitiesDetectionJob +func (c *Comprehend) DescribeEntitiesDetectionJob(input *DescribeEntitiesDetectionJobInput) (*DescribeEntitiesDetectionJobOutput, error) { + req, out := c.DescribeEntitiesDetectionJobRequest(input) + return out, req.Send() +} + +// DescribeEntitiesDetectionJobWithContext is the same as DescribeEntitiesDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEntitiesDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) DescribeEntitiesDetectionJobWithContext(ctx aws.Context, input *DescribeEntitiesDetectionJobInput, opts ...request.Option) (*DescribeEntitiesDetectionJobOutput, error) { + req, out := c.DescribeEntitiesDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeKeyPhrasesDetectionJob = "DescribeKeyPhrasesDetectionJob" + +// DescribeKeyPhrasesDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the DescribeKeyPhrasesDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeKeyPhrasesDetectionJob for more information on using the DescribeKeyPhrasesDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeKeyPhrasesDetectionJobRequest method. +// req, resp := client.DescribeKeyPhrasesDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeKeyPhrasesDetectionJob +func (c *Comprehend) DescribeKeyPhrasesDetectionJobRequest(input *DescribeKeyPhrasesDetectionJobInput) (req *request.Request, output *DescribeKeyPhrasesDetectionJobOutput) { + op := &request.Operation{ + Name: opDescribeKeyPhrasesDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeKeyPhrasesDetectionJobInput{} + } + + output = &DescribeKeyPhrasesDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeKeyPhrasesDetectionJob API operation for Amazon Comprehend. +// +// Gets the properties associated with a key phrases detection job. Use this +// operation to get the status of a detection job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation DescribeKeyPhrasesDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeKeyPhrasesDetectionJob +func (c *Comprehend) DescribeKeyPhrasesDetectionJob(input *DescribeKeyPhrasesDetectionJobInput) (*DescribeKeyPhrasesDetectionJobOutput, error) { + req, out := c.DescribeKeyPhrasesDetectionJobRequest(input) + return out, req.Send() +} + +// DescribeKeyPhrasesDetectionJobWithContext is the same as DescribeKeyPhrasesDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeKeyPhrasesDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) DescribeKeyPhrasesDetectionJobWithContext(ctx aws.Context, input *DescribeKeyPhrasesDetectionJobInput, opts ...request.Option) (*DescribeKeyPhrasesDetectionJobOutput, error) { + req, out := c.DescribeKeyPhrasesDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeSentimentDetectionJob = "DescribeSentimentDetectionJob" + +// DescribeSentimentDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the DescribeSentimentDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeSentimentDetectionJob for more information on using the DescribeSentimentDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeSentimentDetectionJobRequest method. +// req, resp := client.DescribeSentimentDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeSentimentDetectionJob +func (c *Comprehend) DescribeSentimentDetectionJobRequest(input *DescribeSentimentDetectionJobInput) (req *request.Request, output *DescribeSentimentDetectionJobOutput) { + op := &request.Operation{ + Name: opDescribeSentimentDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeSentimentDetectionJobInput{} + } + + output = &DescribeSentimentDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeSentimentDetectionJob API operation for Amazon Comprehend. +// +// Gets the properties associated with a sentiment detection job. Use this operation +// to get the status of a detection job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation DescribeSentimentDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/DescribeSentimentDetectionJob +func (c *Comprehend) DescribeSentimentDetectionJob(input *DescribeSentimentDetectionJobInput) (*DescribeSentimentDetectionJobOutput, error) { + req, out := c.DescribeSentimentDetectionJobRequest(input) + return out, req.Send() +} + +// DescribeSentimentDetectionJobWithContext is the same as DescribeSentimentDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeSentimentDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) DescribeSentimentDetectionJobWithContext(ctx aws.Context, input *DescribeSentimentDetectionJobInput, opts ...request.Option) (*DescribeSentimentDetectionJobOutput, error) { + req, out := c.DescribeSentimentDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDescribeTopicsDetectionJob = "DescribeTopicsDetectionJob" // DescribeTopicsDetectionJobRequest generates a "aws/request.Request" representing the @@ -608,8 +964,8 @@ func (c *Comprehend) DetectEntitiesRequest(input *DetectEntitiesInput) (req *req // DetectEntities API operation for Amazon Comprehend. // -// Inspects text for entities, and returns information about them. For more -// information, about entities, see how-entities. +// Inspects text for named entities, and returns information about them. For +// more information, about named entities, see how-entities. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -839,6 +1195,586 @@ func (c *Comprehend) DetectSentimentWithContext(ctx aws.Context, input *DetectSe return out, req.Send() } +const opListDominantLanguageDetectionJobs = "ListDominantLanguageDetectionJobs" + +// ListDominantLanguageDetectionJobsRequest generates a "aws/request.Request" representing the +// client's request for the ListDominantLanguageDetectionJobs operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListDominantLanguageDetectionJobs for more information on using the ListDominantLanguageDetectionJobs +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListDominantLanguageDetectionJobsRequest method. +// req, resp := client.ListDominantLanguageDetectionJobsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListDominantLanguageDetectionJobs +func (c *Comprehend) ListDominantLanguageDetectionJobsRequest(input *ListDominantLanguageDetectionJobsInput) (req *request.Request, output *ListDominantLanguageDetectionJobsOutput) { + op := &request.Operation{ + Name: opListDominantLanguageDetectionJobs, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListDominantLanguageDetectionJobsInput{} + } + + output = &ListDominantLanguageDetectionJobsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListDominantLanguageDetectionJobs API operation for Amazon Comprehend. +// +// Gets a list of the dominant language detection jobs that you have submitted. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation ListDominantLanguageDetectionJobs for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInvalidFilterException "InvalidFilterException" +// The filter specified for the ListTopicDetectionJobs operation is invalid. +// Specify a different filter. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListDominantLanguageDetectionJobs +func (c *Comprehend) ListDominantLanguageDetectionJobs(input *ListDominantLanguageDetectionJobsInput) (*ListDominantLanguageDetectionJobsOutput, error) { + req, out := c.ListDominantLanguageDetectionJobsRequest(input) + return out, req.Send() +} + +// ListDominantLanguageDetectionJobsWithContext is the same as ListDominantLanguageDetectionJobs with the addition of +// the ability to pass a context and additional request options. +// +// See ListDominantLanguageDetectionJobs for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListDominantLanguageDetectionJobsWithContext(ctx aws.Context, input *ListDominantLanguageDetectionJobsInput, opts ...request.Option) (*ListDominantLanguageDetectionJobsOutput, error) { + req, out := c.ListDominantLanguageDetectionJobsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListDominantLanguageDetectionJobsPages iterates over the pages of a ListDominantLanguageDetectionJobs operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListDominantLanguageDetectionJobs method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListDominantLanguageDetectionJobs operation. +// pageNum := 0 +// err := client.ListDominantLanguageDetectionJobsPages(params, +// func(page *ListDominantLanguageDetectionJobsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Comprehend) ListDominantLanguageDetectionJobsPages(input *ListDominantLanguageDetectionJobsInput, fn func(*ListDominantLanguageDetectionJobsOutput, bool) bool) error { + return c.ListDominantLanguageDetectionJobsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListDominantLanguageDetectionJobsPagesWithContext same as ListDominantLanguageDetectionJobsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListDominantLanguageDetectionJobsPagesWithContext(ctx aws.Context, input *ListDominantLanguageDetectionJobsInput, fn func(*ListDominantLanguageDetectionJobsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListDominantLanguageDetectionJobsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListDominantLanguageDetectionJobsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListDominantLanguageDetectionJobsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opListEntitiesDetectionJobs = "ListEntitiesDetectionJobs" + +// ListEntitiesDetectionJobsRequest generates a "aws/request.Request" representing the +// client's request for the ListEntitiesDetectionJobs operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListEntitiesDetectionJobs for more information on using the ListEntitiesDetectionJobs +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListEntitiesDetectionJobsRequest method. +// req, resp := client.ListEntitiesDetectionJobsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListEntitiesDetectionJobs +func (c *Comprehend) ListEntitiesDetectionJobsRequest(input *ListEntitiesDetectionJobsInput) (req *request.Request, output *ListEntitiesDetectionJobsOutput) { + op := &request.Operation{ + Name: opListEntitiesDetectionJobs, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListEntitiesDetectionJobsInput{} + } + + output = &ListEntitiesDetectionJobsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListEntitiesDetectionJobs API operation for Amazon Comprehend. +// +// Gets a list of the entity detection jobs that you have submitted. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation ListEntitiesDetectionJobs for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInvalidFilterException "InvalidFilterException" +// The filter specified for the ListTopicDetectionJobs operation is invalid. +// Specify a different filter. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListEntitiesDetectionJobs +func (c *Comprehend) ListEntitiesDetectionJobs(input *ListEntitiesDetectionJobsInput) (*ListEntitiesDetectionJobsOutput, error) { + req, out := c.ListEntitiesDetectionJobsRequest(input) + return out, req.Send() +} + +// ListEntitiesDetectionJobsWithContext is the same as ListEntitiesDetectionJobs with the addition of +// the ability to pass a context and additional request options. +// +// See ListEntitiesDetectionJobs for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListEntitiesDetectionJobsWithContext(ctx aws.Context, input *ListEntitiesDetectionJobsInput, opts ...request.Option) (*ListEntitiesDetectionJobsOutput, error) { + req, out := c.ListEntitiesDetectionJobsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListEntitiesDetectionJobsPages iterates over the pages of a ListEntitiesDetectionJobs operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListEntitiesDetectionJobs method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListEntitiesDetectionJobs operation. +// pageNum := 0 +// err := client.ListEntitiesDetectionJobsPages(params, +// func(page *ListEntitiesDetectionJobsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Comprehend) ListEntitiesDetectionJobsPages(input *ListEntitiesDetectionJobsInput, fn func(*ListEntitiesDetectionJobsOutput, bool) bool) error { + return c.ListEntitiesDetectionJobsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListEntitiesDetectionJobsPagesWithContext same as ListEntitiesDetectionJobsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListEntitiesDetectionJobsPagesWithContext(ctx aws.Context, input *ListEntitiesDetectionJobsInput, fn func(*ListEntitiesDetectionJobsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListEntitiesDetectionJobsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListEntitiesDetectionJobsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListEntitiesDetectionJobsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opListKeyPhrasesDetectionJobs = "ListKeyPhrasesDetectionJobs" + +// ListKeyPhrasesDetectionJobsRequest generates a "aws/request.Request" representing the +// client's request for the ListKeyPhrasesDetectionJobs operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListKeyPhrasesDetectionJobs for more information on using the ListKeyPhrasesDetectionJobs +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListKeyPhrasesDetectionJobsRequest method. +// req, resp := client.ListKeyPhrasesDetectionJobsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListKeyPhrasesDetectionJobs +func (c *Comprehend) ListKeyPhrasesDetectionJobsRequest(input *ListKeyPhrasesDetectionJobsInput) (req *request.Request, output *ListKeyPhrasesDetectionJobsOutput) { + op := &request.Operation{ + Name: opListKeyPhrasesDetectionJobs, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListKeyPhrasesDetectionJobsInput{} + } + + output = &ListKeyPhrasesDetectionJobsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListKeyPhrasesDetectionJobs API operation for Amazon Comprehend. +// +// Get a list of key phrase detection jobs that you have submitted. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation ListKeyPhrasesDetectionJobs for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInvalidFilterException "InvalidFilterException" +// The filter specified for the ListTopicDetectionJobs operation is invalid. +// Specify a different filter. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListKeyPhrasesDetectionJobs +func (c *Comprehend) ListKeyPhrasesDetectionJobs(input *ListKeyPhrasesDetectionJobsInput) (*ListKeyPhrasesDetectionJobsOutput, error) { + req, out := c.ListKeyPhrasesDetectionJobsRequest(input) + return out, req.Send() +} + +// ListKeyPhrasesDetectionJobsWithContext is the same as ListKeyPhrasesDetectionJobs with the addition of +// the ability to pass a context and additional request options. +// +// See ListKeyPhrasesDetectionJobs for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListKeyPhrasesDetectionJobsWithContext(ctx aws.Context, input *ListKeyPhrasesDetectionJobsInput, opts ...request.Option) (*ListKeyPhrasesDetectionJobsOutput, error) { + req, out := c.ListKeyPhrasesDetectionJobsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListKeyPhrasesDetectionJobsPages iterates over the pages of a ListKeyPhrasesDetectionJobs operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListKeyPhrasesDetectionJobs method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListKeyPhrasesDetectionJobs operation. +// pageNum := 0 +// err := client.ListKeyPhrasesDetectionJobsPages(params, +// func(page *ListKeyPhrasesDetectionJobsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Comprehend) ListKeyPhrasesDetectionJobsPages(input *ListKeyPhrasesDetectionJobsInput, fn func(*ListKeyPhrasesDetectionJobsOutput, bool) bool) error { + return c.ListKeyPhrasesDetectionJobsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListKeyPhrasesDetectionJobsPagesWithContext same as ListKeyPhrasesDetectionJobsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListKeyPhrasesDetectionJobsPagesWithContext(ctx aws.Context, input *ListKeyPhrasesDetectionJobsInput, fn func(*ListKeyPhrasesDetectionJobsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListKeyPhrasesDetectionJobsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListKeyPhrasesDetectionJobsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListKeyPhrasesDetectionJobsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opListSentimentDetectionJobs = "ListSentimentDetectionJobs" + +// ListSentimentDetectionJobsRequest generates a "aws/request.Request" representing the +// client's request for the ListSentimentDetectionJobs operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListSentimentDetectionJobs for more information on using the ListSentimentDetectionJobs +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListSentimentDetectionJobsRequest method. +// req, resp := client.ListSentimentDetectionJobsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListSentimentDetectionJobs +func (c *Comprehend) ListSentimentDetectionJobsRequest(input *ListSentimentDetectionJobsInput) (req *request.Request, output *ListSentimentDetectionJobsOutput) { + op := &request.Operation{ + Name: opListSentimentDetectionJobs, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListSentimentDetectionJobsInput{} + } + + output = &ListSentimentDetectionJobsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListSentimentDetectionJobs API operation for Amazon Comprehend. +// +// Gets a list of sentiment detection jobs that you have submitted. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation ListSentimentDetectionJobs for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInvalidFilterException "InvalidFilterException" +// The filter specified for the ListTopicDetectionJobs operation is invalid. +// Specify a different filter. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/ListSentimentDetectionJobs +func (c *Comprehend) ListSentimentDetectionJobs(input *ListSentimentDetectionJobsInput) (*ListSentimentDetectionJobsOutput, error) { + req, out := c.ListSentimentDetectionJobsRequest(input) + return out, req.Send() +} + +// ListSentimentDetectionJobsWithContext is the same as ListSentimentDetectionJobs with the addition of +// the ability to pass a context and additional request options. +// +// See ListSentimentDetectionJobs for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListSentimentDetectionJobsWithContext(ctx aws.Context, input *ListSentimentDetectionJobsInput, opts ...request.Option) (*ListSentimentDetectionJobsOutput, error) { + req, out := c.ListSentimentDetectionJobsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListSentimentDetectionJobsPages iterates over the pages of a ListSentimentDetectionJobs operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListSentimentDetectionJobs method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListSentimentDetectionJobs operation. +// pageNum := 0 +// err := client.ListSentimentDetectionJobsPages(params, +// func(page *ListSentimentDetectionJobsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Comprehend) ListSentimentDetectionJobsPages(input *ListSentimentDetectionJobsInput, fn func(*ListSentimentDetectionJobsOutput, bool) bool) error { + return c.ListSentimentDetectionJobsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListSentimentDetectionJobsPagesWithContext same as ListSentimentDetectionJobsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) ListSentimentDetectionJobsPagesWithContext(ctx aws.Context, input *ListSentimentDetectionJobsInput, fn func(*ListSentimentDetectionJobsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListSentimentDetectionJobsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListSentimentDetectionJobsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListSentimentDetectionJobsOutput), !p.HasNextPage()) + } + return p.Err() +} + const opListTopicsDetectionJobs = "ListTopicsDetectionJobs" // ListTopicsDetectionJobsRequest generates a "aws/request.Request" representing the @@ -984,6 +1920,350 @@ func (c *Comprehend) ListTopicsDetectionJobsPagesWithContext(ctx aws.Context, in return p.Err() } +const opStartDominantLanguageDetectionJob = "StartDominantLanguageDetectionJob" + +// StartDominantLanguageDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StartDominantLanguageDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StartDominantLanguageDetectionJob for more information on using the StartDominantLanguageDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StartDominantLanguageDetectionJobRequest method. +// req, resp := client.StartDominantLanguageDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartDominantLanguageDetectionJob +func (c *Comprehend) StartDominantLanguageDetectionJobRequest(input *StartDominantLanguageDetectionJobInput) (req *request.Request, output *StartDominantLanguageDetectionJobOutput) { + op := &request.Operation{ + Name: opStartDominantLanguageDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StartDominantLanguageDetectionJobInput{} + } + + output = &StartDominantLanguageDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StartDominantLanguageDetectionJob API operation for Amazon Comprehend. +// +// Starts an asynchronous dominant language detection job for a collection of +// documents. Use the operation to track the status of a job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StartDominantLanguageDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartDominantLanguageDetectionJob +func (c *Comprehend) StartDominantLanguageDetectionJob(input *StartDominantLanguageDetectionJobInput) (*StartDominantLanguageDetectionJobOutput, error) { + req, out := c.StartDominantLanguageDetectionJobRequest(input) + return out, req.Send() +} + +// StartDominantLanguageDetectionJobWithContext is the same as StartDominantLanguageDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StartDominantLanguageDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StartDominantLanguageDetectionJobWithContext(ctx aws.Context, input *StartDominantLanguageDetectionJobInput, opts ...request.Option) (*StartDominantLanguageDetectionJobOutput, error) { + req, out := c.StartDominantLanguageDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opStartEntitiesDetectionJob = "StartEntitiesDetectionJob" + +// StartEntitiesDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StartEntitiesDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StartEntitiesDetectionJob for more information on using the StartEntitiesDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StartEntitiesDetectionJobRequest method. +// req, resp := client.StartEntitiesDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartEntitiesDetectionJob +func (c *Comprehend) StartEntitiesDetectionJobRequest(input *StartEntitiesDetectionJobInput) (req *request.Request, output *StartEntitiesDetectionJobOutput) { + op := &request.Operation{ + Name: opStartEntitiesDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StartEntitiesDetectionJobInput{} + } + + output = &StartEntitiesDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StartEntitiesDetectionJob API operation for Amazon Comprehend. +// +// Starts an asynchronous entity detection job for a collection of documents. +// Use the operation to track the status of a job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StartEntitiesDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartEntitiesDetectionJob +func (c *Comprehend) StartEntitiesDetectionJob(input *StartEntitiesDetectionJobInput) (*StartEntitiesDetectionJobOutput, error) { + req, out := c.StartEntitiesDetectionJobRequest(input) + return out, req.Send() +} + +// StartEntitiesDetectionJobWithContext is the same as StartEntitiesDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StartEntitiesDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StartEntitiesDetectionJobWithContext(ctx aws.Context, input *StartEntitiesDetectionJobInput, opts ...request.Option) (*StartEntitiesDetectionJobOutput, error) { + req, out := c.StartEntitiesDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opStartKeyPhrasesDetectionJob = "StartKeyPhrasesDetectionJob" + +// StartKeyPhrasesDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StartKeyPhrasesDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StartKeyPhrasesDetectionJob for more information on using the StartKeyPhrasesDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StartKeyPhrasesDetectionJobRequest method. +// req, resp := client.StartKeyPhrasesDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartKeyPhrasesDetectionJob +func (c *Comprehend) StartKeyPhrasesDetectionJobRequest(input *StartKeyPhrasesDetectionJobInput) (req *request.Request, output *StartKeyPhrasesDetectionJobOutput) { + op := &request.Operation{ + Name: opStartKeyPhrasesDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StartKeyPhrasesDetectionJobInput{} + } + + output = &StartKeyPhrasesDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StartKeyPhrasesDetectionJob API operation for Amazon Comprehend. +// +// Starts an asynchronous key phrase detection job for a collection of documents. +// Use the operation to track the status of a job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StartKeyPhrasesDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartKeyPhrasesDetectionJob +func (c *Comprehend) StartKeyPhrasesDetectionJob(input *StartKeyPhrasesDetectionJobInput) (*StartKeyPhrasesDetectionJobOutput, error) { + req, out := c.StartKeyPhrasesDetectionJobRequest(input) + return out, req.Send() +} + +// StartKeyPhrasesDetectionJobWithContext is the same as StartKeyPhrasesDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StartKeyPhrasesDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StartKeyPhrasesDetectionJobWithContext(ctx aws.Context, input *StartKeyPhrasesDetectionJobInput, opts ...request.Option) (*StartKeyPhrasesDetectionJobOutput, error) { + req, out := c.StartKeyPhrasesDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opStartSentimentDetectionJob = "StartSentimentDetectionJob" + +// StartSentimentDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StartSentimentDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StartSentimentDetectionJob for more information on using the StartSentimentDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StartSentimentDetectionJobRequest method. +// req, resp := client.StartSentimentDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartSentimentDetectionJob +func (c *Comprehend) StartSentimentDetectionJobRequest(input *StartSentimentDetectionJobInput) (req *request.Request, output *StartSentimentDetectionJobOutput) { + op := &request.Operation{ + Name: opStartSentimentDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StartSentimentDetectionJobInput{} + } + + output = &StartSentimentDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StartSentimentDetectionJob API operation for Amazon Comprehend. +// +// Starts an asynchronous sentiment detection job for a collection of documents. +// use the operation to track the status of a job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StartSentimentDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// The number of requests exceeds the limit. Resubmit your request later. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StartSentimentDetectionJob +func (c *Comprehend) StartSentimentDetectionJob(input *StartSentimentDetectionJobInput) (*StartSentimentDetectionJobOutput, error) { + req, out := c.StartSentimentDetectionJobRequest(input) + return out, req.Send() +} + +// StartSentimentDetectionJobWithContext is the same as StartSentimentDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StartSentimentDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StartSentimentDetectionJobWithContext(ctx aws.Context, input *StartSentimentDetectionJobInput, opts ...request.Option) (*StartSentimentDetectionJobOutput, error) { + req, out := c.StartSentimentDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opStartTopicsDetectionJob = "StartTopicsDetectionJob" // StartTopicsDetectionJobRequest generates a "aws/request.Request" representing the @@ -1070,6 +2350,382 @@ func (c *Comprehend) StartTopicsDetectionJobWithContext(ctx aws.Context, input * return out, req.Send() } +const opStopDominantLanguageDetectionJob = "StopDominantLanguageDetectionJob" + +// StopDominantLanguageDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StopDominantLanguageDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StopDominantLanguageDetectionJob for more information on using the StopDominantLanguageDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StopDominantLanguageDetectionJobRequest method. +// req, resp := client.StopDominantLanguageDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopDominantLanguageDetectionJob +func (c *Comprehend) StopDominantLanguageDetectionJobRequest(input *StopDominantLanguageDetectionJobInput) (req *request.Request, output *StopDominantLanguageDetectionJobOutput) { + op := &request.Operation{ + Name: opStopDominantLanguageDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StopDominantLanguageDetectionJobInput{} + } + + output = &StopDominantLanguageDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StopDominantLanguageDetectionJob API operation for Amazon Comprehend. +// +// Stops a dominant language detection job in progress. +// +// If the job state is IN_PROGRESS the job will be marked for termination and +// put into the STOPPING state. +// +// If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob +// operation, the operation will return a 400 Internal Request Exception. +// +// When a job is stopped, any document that has already been processed will +// be written to the output location. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StopDominantLanguageDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopDominantLanguageDetectionJob +func (c *Comprehend) StopDominantLanguageDetectionJob(input *StopDominantLanguageDetectionJobInput) (*StopDominantLanguageDetectionJobOutput, error) { + req, out := c.StopDominantLanguageDetectionJobRequest(input) + return out, req.Send() +} + +// StopDominantLanguageDetectionJobWithContext is the same as StopDominantLanguageDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StopDominantLanguageDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StopDominantLanguageDetectionJobWithContext(ctx aws.Context, input *StopDominantLanguageDetectionJobInput, opts ...request.Option) (*StopDominantLanguageDetectionJobOutput, error) { + req, out := c.StopDominantLanguageDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opStopEntitiesDetectionJob = "StopEntitiesDetectionJob" + +// StopEntitiesDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StopEntitiesDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StopEntitiesDetectionJob for more information on using the StopEntitiesDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StopEntitiesDetectionJobRequest method. +// req, resp := client.StopEntitiesDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopEntitiesDetectionJob +func (c *Comprehend) StopEntitiesDetectionJobRequest(input *StopEntitiesDetectionJobInput) (req *request.Request, output *StopEntitiesDetectionJobOutput) { + op := &request.Operation{ + Name: opStopEntitiesDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StopEntitiesDetectionJobInput{} + } + + output = &StopEntitiesDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StopEntitiesDetectionJob API operation for Amazon Comprehend. +// +// Stops an entities detection job in progress. +// +// If the job state is IN_PROGRESS the job will be marked for termination and +// put into the STOPPING state. +// +// If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob +// operation, the operation will return a 400 Internal Request Exception. +// +// When a job is stopped, any document that has already been processed will +// be written to the output location. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StopEntitiesDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopEntitiesDetectionJob +func (c *Comprehend) StopEntitiesDetectionJob(input *StopEntitiesDetectionJobInput) (*StopEntitiesDetectionJobOutput, error) { + req, out := c.StopEntitiesDetectionJobRequest(input) + return out, req.Send() +} + +// StopEntitiesDetectionJobWithContext is the same as StopEntitiesDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StopEntitiesDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StopEntitiesDetectionJobWithContext(ctx aws.Context, input *StopEntitiesDetectionJobInput, opts ...request.Option) (*StopEntitiesDetectionJobOutput, error) { + req, out := c.StopEntitiesDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opStopKeyPhrasesDetectionJob = "StopKeyPhrasesDetectionJob" + +// StopKeyPhrasesDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StopKeyPhrasesDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StopKeyPhrasesDetectionJob for more information on using the StopKeyPhrasesDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StopKeyPhrasesDetectionJobRequest method. +// req, resp := client.StopKeyPhrasesDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopKeyPhrasesDetectionJob +func (c *Comprehend) StopKeyPhrasesDetectionJobRequest(input *StopKeyPhrasesDetectionJobInput) (req *request.Request, output *StopKeyPhrasesDetectionJobOutput) { + op := &request.Operation{ + Name: opStopKeyPhrasesDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StopKeyPhrasesDetectionJobInput{} + } + + output = &StopKeyPhrasesDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StopKeyPhrasesDetectionJob API operation for Amazon Comprehend. +// +// Stops a key phrases detection job in progress. +// +// If the job state is IN_PROGRESS the job will be marked for termination and +// put into the STOPPING state. +// +// If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob +// operation, the operation will return a 400 Internal Request Exception. +// +// When a job is stopped, any document that has already been processed will +// be written to the output location. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StopKeyPhrasesDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopKeyPhrasesDetectionJob +func (c *Comprehend) StopKeyPhrasesDetectionJob(input *StopKeyPhrasesDetectionJobInput) (*StopKeyPhrasesDetectionJobOutput, error) { + req, out := c.StopKeyPhrasesDetectionJobRequest(input) + return out, req.Send() +} + +// StopKeyPhrasesDetectionJobWithContext is the same as StopKeyPhrasesDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StopKeyPhrasesDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StopKeyPhrasesDetectionJobWithContext(ctx aws.Context, input *StopKeyPhrasesDetectionJobInput, opts ...request.Option) (*StopKeyPhrasesDetectionJobOutput, error) { + req, out := c.StopKeyPhrasesDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opStopSentimentDetectionJob = "StopSentimentDetectionJob" + +// StopSentimentDetectionJobRequest generates a "aws/request.Request" representing the +// client's request for the StopSentimentDetectionJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StopSentimentDetectionJob for more information on using the StopSentimentDetectionJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StopSentimentDetectionJobRequest method. +// req, resp := client.StopSentimentDetectionJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopSentimentDetectionJob +func (c *Comprehend) StopSentimentDetectionJobRequest(input *StopSentimentDetectionJobInput) (req *request.Request, output *StopSentimentDetectionJobOutput) { + op := &request.Operation{ + Name: opStopSentimentDetectionJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StopSentimentDetectionJobInput{} + } + + output = &StopSentimentDetectionJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// StopSentimentDetectionJob API operation for Amazon Comprehend. +// +// Stops a sentiment detection job in progress. +// +// If the job state is IN_PROGRESS the job will be marked for termination and +// put into the STOPPING state. +// +// If the job is in the COMPLETED or FAILED state when you call the StopDominantLanguageDetectionJob +// operation, the operation will return a 400 Internal Request Exception. +// +// When a job is stopped, any document that has already been processed will +// be written to the output location. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Comprehend's +// API operation StopSentimentDetectionJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is invalid. +// +// * ErrCodeJobNotFoundException "JobNotFoundException" +// The specified job was not found. Check the job ID and try again. +// +// * ErrCodeInternalServerException "InternalServerException" +// An internal server error occurred. Retry your request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/comprehend-2017-11-27/StopSentimentDetectionJob +func (c *Comprehend) StopSentimentDetectionJob(input *StopSentimentDetectionJobInput) (*StopSentimentDetectionJobOutput, error) { + req, out := c.StopSentimentDetectionJobRequest(input) + return out, req.Send() +} + +// StopSentimentDetectionJobWithContext is the same as StopSentimentDetectionJob with the addition of +// the ability to pass a context and additional request options. +// +// See StopSentimentDetectionJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Comprehend) StopSentimentDetectionJobWithContext(ctx aws.Context, input *StopSentimentDetectionJobInput, opts ...request.Option) (*StopSentimentDetectionJobOutput, error) { + req, out := c.StopSentimentDetectionJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + type BatchDetectDominantLanguageInput struct { _ struct{} `type:"structure"` @@ -1190,10 +2846,11 @@ func (s *BatchDetectDominantLanguageOutput) SetResultList(v []*BatchDetectDomina type BatchDetectEntitiesInput struct { _ struct{} `type:"structure"` - // The language of the input documents. All documents must be in the same language. + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. // // LanguageCode is a required field - LanguageCode *string `min:"1" type:"string" required:"true"` + LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` // A list containing the text of the input documents. The list can contain a // maximum of 25 documents. Each document must contain fewer than 5,000 bytes @@ -1219,9 +2876,6 @@ func (s *BatchDetectEntitiesInput) Validate() error { if s.LanguageCode == nil { invalidParams.Add(request.NewErrParamRequired("LanguageCode")) } - if s.LanguageCode != nil && len(*s.LanguageCode) < 1 { - invalidParams.Add(request.NewErrParamMinLen("LanguageCode", 1)) - } if s.TextList == nil { invalidParams.Add(request.NewErrParamRequired("TextList")) } @@ -1323,10 +2977,11 @@ func (s *BatchDetectEntitiesOutput) SetResultList(v []*BatchDetectEntitiesItemRe type BatchDetectKeyPhrasesInput struct { _ struct{} `type:"structure"` - // The language of the input documents. All documents must be in the same language. + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. // // LanguageCode is a required field - LanguageCode *string `min:"1" type:"string" required:"true"` + LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` // A list containing the text of the input documents. The list can contain a // maximum of 25 documents. Each document must contain fewer that 5,000 bytes @@ -1352,9 +3007,6 @@ func (s *BatchDetectKeyPhrasesInput) Validate() error { if s.LanguageCode == nil { invalidParams.Add(request.NewErrParamRequired("LanguageCode")) } - if s.LanguageCode != nil && len(*s.LanguageCode) < 1 { - invalidParams.Add(request.NewErrParamMinLen("LanguageCode", 1)) - } if s.TextList == nil { invalidParams.Add(request.NewErrParamRequired("TextList")) } @@ -1456,10 +3108,11 @@ func (s *BatchDetectKeyPhrasesOutput) SetResultList(v []*BatchDetectKeyPhrasesIt type BatchDetectSentimentInput struct { _ struct{} `type:"structure"` - // The language of the input documents. All documents must be in the same language. + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. // // LanguageCode is a required field - LanguageCode *string `min:"1" type:"string" required:"true"` + LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` // A list containing the text of the input documents. The list can contain a // maximum of 25 documents. Each document must contain fewer that 5,000 bytes @@ -1485,9 +3138,6 @@ func (s *BatchDetectSentimentInput) Validate() error { if s.LanguageCode == nil { invalidParams.Add(request.NewErrParamRequired("LanguageCode")) } - if s.LanguageCode != nil && len(*s.LanguageCode) < 1 { - invalidParams.Add(request.NewErrParamMinLen("LanguageCode", 1)) - } if s.TextList == nil { invalidParams.Add(request.NewErrParamRequired("TextList")) } @@ -1640,6 +3290,270 @@ func (s *BatchItemError) SetIndex(v int64) *BatchItemError { return s } +type DescribeDominantLanguageDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier that Amazon Comprehend generated for the job. The operation + // returns this identifier in its response. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeDominantLanguageDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDominantLanguageDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDominantLanguageDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDominantLanguageDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *DescribeDominantLanguageDetectionJobInput) SetJobId(v string) *DescribeDominantLanguageDetectionJobInput { + s.JobId = &v + return s +} + +type DescribeDominantLanguageDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // An object that contains the properties associated with a dominant language + // detection job. + DominantLanguageDetectionJobProperties *DominantLanguageDetectionJobProperties `type:"structure"` +} + +// String returns the string representation +func (s DescribeDominantLanguageDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDominantLanguageDetectionJobOutput) GoString() string { + return s.String() +} + +// SetDominantLanguageDetectionJobProperties sets the DominantLanguageDetectionJobProperties field's value. +func (s *DescribeDominantLanguageDetectionJobOutput) SetDominantLanguageDetectionJobProperties(v *DominantLanguageDetectionJobProperties) *DescribeDominantLanguageDetectionJobOutput { + s.DominantLanguageDetectionJobProperties = v + return s +} + +type DescribeEntitiesDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier that Amazon Comprehend generated for the job. The operation + // returns this identifier in its response. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeEntitiesDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEntitiesDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeEntitiesDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeEntitiesDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *DescribeEntitiesDetectionJobInput) SetJobId(v string) *DescribeEntitiesDetectionJobInput { + s.JobId = &v + return s +} + +type DescribeEntitiesDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // An object that contains the properties associated with an entities detection + // job. + EntitiesDetectionJobProperties *EntitiesDetectionJobProperties `type:"structure"` +} + +// String returns the string representation +func (s DescribeEntitiesDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEntitiesDetectionJobOutput) GoString() string { + return s.String() +} + +// SetEntitiesDetectionJobProperties sets the EntitiesDetectionJobProperties field's value. +func (s *DescribeEntitiesDetectionJobOutput) SetEntitiesDetectionJobProperties(v *EntitiesDetectionJobProperties) *DescribeEntitiesDetectionJobOutput { + s.EntitiesDetectionJobProperties = v + return s +} + +type DescribeKeyPhrasesDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier that Amazon Comprehend generated for the job. The operation + // returns this identifier in its response. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeKeyPhrasesDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeKeyPhrasesDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeKeyPhrasesDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeKeyPhrasesDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *DescribeKeyPhrasesDetectionJobInput) SetJobId(v string) *DescribeKeyPhrasesDetectionJobInput { + s.JobId = &v + return s +} + +type DescribeKeyPhrasesDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // An object that contains the properties associated with a key phrases detection + // job. + KeyPhrasesDetectionJobProperties *KeyPhrasesDetectionJobProperties `type:"structure"` +} + +// String returns the string representation +func (s DescribeKeyPhrasesDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeKeyPhrasesDetectionJobOutput) GoString() string { + return s.String() +} + +// SetKeyPhrasesDetectionJobProperties sets the KeyPhrasesDetectionJobProperties field's value. +func (s *DescribeKeyPhrasesDetectionJobOutput) SetKeyPhrasesDetectionJobProperties(v *KeyPhrasesDetectionJobProperties) *DescribeKeyPhrasesDetectionJobOutput { + s.KeyPhrasesDetectionJobProperties = v + return s +} + +type DescribeSentimentDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier that Amazon Comprehend generated for the job. The operation + // returns this identifier in its response. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeSentimentDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeSentimentDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeSentimentDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeSentimentDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *DescribeSentimentDetectionJobInput) SetJobId(v string) *DescribeSentimentDetectionJobInput { + s.JobId = &v + return s +} + +type DescribeSentimentDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // An object that contains the properties associated with a sentiment detection + // job. + SentimentDetectionJobProperties *SentimentDetectionJobProperties `type:"structure"` +} + +// String returns the string representation +func (s DescribeSentimentDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeSentimentDetectionJobOutput) GoString() string { + return s.String() +} + +// SetSentimentDetectionJobProperties sets the SentimentDetectionJobProperties field's value. +func (s *DescribeSentimentDetectionJobOutput) SetSentimentDetectionJobProperties(v *SentimentDetectionJobProperties) *DescribeSentimentDetectionJobOutput { + s.SentimentDetectionJobProperties = v + return s +} + type DescribeTopicsDetectionJobInput struct { _ struct{} `type:"structure"` @@ -1776,11 +3690,8 @@ func (s *DetectDominantLanguageOutput) SetLanguages(v []*DominantLanguage) *Dete type DetectEntitiesInput struct { _ struct{} `type:"structure"` - // The RFC 5646 language code of the input text. If the request does not specify - // the language code, the service detects the dominant language. If you specify - // a language code that the service does not support, it returns UnsupportedLanguageException - // exception. For more information about RFC 5646, see Tags for Identifying - // Languages (https://tools.ietf.org/html/rfc5646) on the IETF Tools web site. + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. // // LanguageCode is a required field LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` @@ -1862,12 +3773,8 @@ func (s *DetectEntitiesOutput) SetEntities(v []*Entity) *DetectEntitiesOutput { type DetectKeyPhrasesInput struct { _ struct{} `type:"structure"` - // The RFC 5646 language code for the input text. If you don't specify a language - // code, Amazon Comprehend detects the dominant language. If you specify the - // code for a language that Amazon Comprehend does not support, it returns and - // UnsupportedLanguageException. For more information about RFC 5646, see Tags - // for Identifying Languages (https://tools.ietf.org/html/rfc5646) on the IETF - // Tools web site. + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. // // LanguageCode is a required field LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` @@ -1949,12 +3856,8 @@ func (s *DetectKeyPhrasesOutput) SetKeyPhrases(v []*KeyPhrase) *DetectKeyPhrases type DetectSentimentInput struct { _ struct{} `type:"structure"` - // The RFC 5646 language code for the input text. If you don't specify a language - // code, Amazon Comprehend detects the dominant language. If you specify the - // code for a language that Amazon Comprehend does not support, it returns and - // UnsupportedLanguageException. For more information about RFC 5646, see Tags - // for Identifying Languages (https://tools.ietf.org/html/rfc5646) on the IETF - // Tools web site. + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. // // LanguageCode is a required field LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` @@ -2045,7 +3948,9 @@ func (s *DetectSentimentOutput) SetSentimentScore(v *SentimentScore) *DetectSent type DominantLanguage struct { _ struct{} `type:"structure"` - // The RFC 5646 language code for the dominant language. + // The RFC 5646 language code for the dominant language. For more information + // about RFC 5646, see Tags for Identifying Languages (https://tools.ietf.org/html/rfc5646) + // on the IETF Tools web site. LanguageCode *string `min:"1" type:"string"` // The level of confidence that Amazon Comprehend has in the accuracy of the @@ -2075,6 +3980,335 @@ func (s *DominantLanguage) SetScore(v float64) *DominantLanguage { return s } +// Provides information for filtering a list of dominant language detection +// jobs. For more information, see the operation. +type DominantLanguageDetectionJobFilter struct { + _ struct{} `type:"structure"` + + // Filters on the name of the job. + JobName *string `min:"1" type:"string"` + + // Filters the list of jobs based on job status. Returns only jobs with the + // specified status. + JobStatus *string `type:"string" enum:"JobStatus"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted after the specified time. Jobs are + // returned in descending order, newest to oldest. + SubmitTimeAfter *time.Time `type:"timestamp" timestampFormat:"unix"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted before the specified time. Jobs are + // returned in ascending order, oldest to newest. + SubmitTimeBefore *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s DominantLanguageDetectionJobFilter) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DominantLanguageDetectionJobFilter) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DominantLanguageDetectionJobFilter) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DominantLanguageDetectionJobFilter"} + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobName sets the JobName field's value. +func (s *DominantLanguageDetectionJobFilter) SetJobName(v string) *DominantLanguageDetectionJobFilter { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *DominantLanguageDetectionJobFilter) SetJobStatus(v string) *DominantLanguageDetectionJobFilter { + s.JobStatus = &v + return s +} + +// SetSubmitTimeAfter sets the SubmitTimeAfter field's value. +func (s *DominantLanguageDetectionJobFilter) SetSubmitTimeAfter(v time.Time) *DominantLanguageDetectionJobFilter { + s.SubmitTimeAfter = &v + return s +} + +// SetSubmitTimeBefore sets the SubmitTimeBefore field's value. +func (s *DominantLanguageDetectionJobFilter) SetSubmitTimeBefore(v time.Time) *DominantLanguageDetectionJobFilter { + s.SubmitTimeBefore = &v + return s +} + +// Provides information about a dominant language detection job. +type DominantLanguageDetectionJobProperties struct { + _ struct{} `type:"structure"` + + // The time that the dominant language detection job completed. + EndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The input data configuration that you supplied when you created the dominant + // language detection job. + InputDataConfig *InputDataConfig `type:"structure"` + + // The identifier assigned to the dominant language detection job. + JobId *string `min:"1" type:"string"` + + // The name that you assigned to the dominant language detection job. + JobName *string `min:"1" type:"string"` + + // The current status of the dominant language detection job. If the status + // is FAILED, the Message field shows the reason for the failure. + JobStatus *string `type:"string" enum:"JobStatus"` + + // A description for the status of a job. + Message *string `type:"string"` + + // The output data configuration that you supplied when you created the dominant + // language detection job. + OutputDataConfig *OutputDataConfig `type:"structure"` + + // The time that the dominant language detection job was submitted for processing. + SubmitTime *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s DominantLanguageDetectionJobProperties) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DominantLanguageDetectionJobProperties) GoString() string { + return s.String() +} + +// SetEndTime sets the EndTime field's value. +func (s *DominantLanguageDetectionJobProperties) SetEndTime(v time.Time) *DominantLanguageDetectionJobProperties { + s.EndTime = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *DominantLanguageDetectionJobProperties) SetInputDataConfig(v *InputDataConfig) *DominantLanguageDetectionJobProperties { + s.InputDataConfig = v + return s +} + +// SetJobId sets the JobId field's value. +func (s *DominantLanguageDetectionJobProperties) SetJobId(v string) *DominantLanguageDetectionJobProperties { + s.JobId = &v + return s +} + +// SetJobName sets the JobName field's value. +func (s *DominantLanguageDetectionJobProperties) SetJobName(v string) *DominantLanguageDetectionJobProperties { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *DominantLanguageDetectionJobProperties) SetJobStatus(v string) *DominantLanguageDetectionJobProperties { + s.JobStatus = &v + return s +} + +// SetMessage sets the Message field's value. +func (s *DominantLanguageDetectionJobProperties) SetMessage(v string) *DominantLanguageDetectionJobProperties { + s.Message = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *DominantLanguageDetectionJobProperties) SetOutputDataConfig(v *OutputDataConfig) *DominantLanguageDetectionJobProperties { + s.OutputDataConfig = v + return s +} + +// SetSubmitTime sets the SubmitTime field's value. +func (s *DominantLanguageDetectionJobProperties) SetSubmitTime(v time.Time) *DominantLanguageDetectionJobProperties { + s.SubmitTime = &v + return s +} + +// Provides information for filtering a list of dominant language detection +// jobs. For more information, see the operation. +type EntitiesDetectionJobFilter struct { + _ struct{} `type:"structure"` + + // Filters on the name of the job. + JobName *string `min:"1" type:"string"` + + // Filters the list of jobs based on job status. Returns only jobs with the + // specified status. + JobStatus *string `type:"string" enum:"JobStatus"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted after the specified time. Jobs are + // returned in descending order, newest to oldest. + SubmitTimeAfter *time.Time `type:"timestamp" timestampFormat:"unix"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted before the specified time. Jobs are + // returned in ascending order, oldest to newest. + SubmitTimeBefore *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s EntitiesDetectionJobFilter) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EntitiesDetectionJobFilter) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *EntitiesDetectionJobFilter) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "EntitiesDetectionJobFilter"} + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobName sets the JobName field's value. +func (s *EntitiesDetectionJobFilter) SetJobName(v string) *EntitiesDetectionJobFilter { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *EntitiesDetectionJobFilter) SetJobStatus(v string) *EntitiesDetectionJobFilter { + s.JobStatus = &v + return s +} + +// SetSubmitTimeAfter sets the SubmitTimeAfter field's value. +func (s *EntitiesDetectionJobFilter) SetSubmitTimeAfter(v time.Time) *EntitiesDetectionJobFilter { + s.SubmitTimeAfter = &v + return s +} + +// SetSubmitTimeBefore sets the SubmitTimeBefore field's value. +func (s *EntitiesDetectionJobFilter) SetSubmitTimeBefore(v time.Time) *EntitiesDetectionJobFilter { + s.SubmitTimeBefore = &v + return s +} + +// Provides information about an entities detection job. +type EntitiesDetectionJobProperties struct { + _ struct{} `type:"structure"` + + // The time that the entities detection job completed + EndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The input data configuration that you supplied when you created the entities + // detection job. + InputDataConfig *InputDataConfig `type:"structure"` + + // The identifier assigned to the entities detection job. + JobId *string `min:"1" type:"string"` + + // The name that you assigned the entities detection job. + JobName *string `min:"1" type:"string"` + + // The current status of the entities detection job. If the status is FAILED, + // the Message field shows the reason for the failure. + JobStatus *string `type:"string" enum:"JobStatus"` + + // The language code of the input documents. + LanguageCode *string `type:"string" enum:"LanguageCode"` + + // A description of the status of a job. + Message *string `type:"string"` + + // The output data configuration that you supplied when you created the entities + // detection job. + OutputDataConfig *OutputDataConfig `type:"structure"` + + // The time that the entities detection job was submitted for processing. + SubmitTime *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s EntitiesDetectionJobProperties) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EntitiesDetectionJobProperties) GoString() string { + return s.String() +} + +// SetEndTime sets the EndTime field's value. +func (s *EntitiesDetectionJobProperties) SetEndTime(v time.Time) *EntitiesDetectionJobProperties { + s.EndTime = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *EntitiesDetectionJobProperties) SetInputDataConfig(v *InputDataConfig) *EntitiesDetectionJobProperties { + s.InputDataConfig = v + return s +} + +// SetJobId sets the JobId field's value. +func (s *EntitiesDetectionJobProperties) SetJobId(v string) *EntitiesDetectionJobProperties { + s.JobId = &v + return s +} + +// SetJobName sets the JobName field's value. +func (s *EntitiesDetectionJobProperties) SetJobName(v string) *EntitiesDetectionJobProperties { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *EntitiesDetectionJobProperties) SetJobStatus(v string) *EntitiesDetectionJobProperties { + s.JobStatus = &v + return s +} + +// SetLanguageCode sets the LanguageCode field's value. +func (s *EntitiesDetectionJobProperties) SetLanguageCode(v string) *EntitiesDetectionJobProperties { + s.LanguageCode = &v + return s +} + +// SetMessage sets the Message field's value. +func (s *EntitiesDetectionJobProperties) SetMessage(v string) *EntitiesDetectionJobProperties { + s.Message = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *EntitiesDetectionJobProperties) SetOutputDataConfig(v *OutputDataConfig) *EntitiesDetectionJobProperties { + s.OutputDataConfig = v + return s +} + +// SetSubmitTime sets the SubmitTime field's value. +func (s *EntitiesDetectionJobProperties) SetSubmitTime(v time.Time) *EntitiesDetectionJobProperties { + s.SubmitTime = &v + return s +} + // Provides information about an entity. type Entity struct { _ struct{} `type:"structure"` @@ -2264,6 +4498,559 @@ func (s *KeyPhrase) SetText(v string) *KeyPhrase { return s } +// Provides information for filtering a list of dominant language detection +// jobs. For more information, see the operation. +type KeyPhrasesDetectionJobFilter struct { + _ struct{} `type:"structure"` + + // Filters on the name of the job. + JobName *string `min:"1" type:"string"` + + // Filters the list of jobs based on job status. Returns only jobs with the + // specified status. + JobStatus *string `type:"string" enum:"JobStatus"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted after the specified time. Jobs are + // returned in descending order, newest to oldest. + SubmitTimeAfter *time.Time `type:"timestamp" timestampFormat:"unix"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted before the specified time. Jobs are + // returned in ascending order, oldest to newest. + SubmitTimeBefore *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s KeyPhrasesDetectionJobFilter) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s KeyPhrasesDetectionJobFilter) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *KeyPhrasesDetectionJobFilter) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "KeyPhrasesDetectionJobFilter"} + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobName sets the JobName field's value. +func (s *KeyPhrasesDetectionJobFilter) SetJobName(v string) *KeyPhrasesDetectionJobFilter { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *KeyPhrasesDetectionJobFilter) SetJobStatus(v string) *KeyPhrasesDetectionJobFilter { + s.JobStatus = &v + return s +} + +// SetSubmitTimeAfter sets the SubmitTimeAfter field's value. +func (s *KeyPhrasesDetectionJobFilter) SetSubmitTimeAfter(v time.Time) *KeyPhrasesDetectionJobFilter { + s.SubmitTimeAfter = &v + return s +} + +// SetSubmitTimeBefore sets the SubmitTimeBefore field's value. +func (s *KeyPhrasesDetectionJobFilter) SetSubmitTimeBefore(v time.Time) *KeyPhrasesDetectionJobFilter { + s.SubmitTimeBefore = &v + return s +} + +// Provides information about a key phrases detection job. +type KeyPhrasesDetectionJobProperties struct { + _ struct{} `type:"structure"` + + // The time that the key phrases detection job completed. + EndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The input data configuration that you supplied when you created the key phrases + // detection job. + InputDataConfig *InputDataConfig `type:"structure"` + + // The identifier assigned to the key phrases detection job. + JobId *string `min:"1" type:"string"` + + // The name that you assigned the key phrases detection job. + JobName *string `min:"1" type:"string"` + + // The current status of the key phrases detection job. If the status is FAILED, + // the Message field shows the reason for the failure. + JobStatus *string `type:"string" enum:"JobStatus"` + + // The language code of the input documents. + LanguageCode *string `type:"string" enum:"LanguageCode"` + + // A description of the status of a job. + Message *string `type:"string"` + + // The output data configuration that you supplied when you created the key + // phrases detection job. + OutputDataConfig *OutputDataConfig `type:"structure"` + + // The time that the key phrases detection job was submitted for processing. + SubmitTime *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s KeyPhrasesDetectionJobProperties) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s KeyPhrasesDetectionJobProperties) GoString() string { + return s.String() +} + +// SetEndTime sets the EndTime field's value. +func (s *KeyPhrasesDetectionJobProperties) SetEndTime(v time.Time) *KeyPhrasesDetectionJobProperties { + s.EndTime = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *KeyPhrasesDetectionJobProperties) SetInputDataConfig(v *InputDataConfig) *KeyPhrasesDetectionJobProperties { + s.InputDataConfig = v + return s +} + +// SetJobId sets the JobId field's value. +func (s *KeyPhrasesDetectionJobProperties) SetJobId(v string) *KeyPhrasesDetectionJobProperties { + s.JobId = &v + return s +} + +// SetJobName sets the JobName field's value. +func (s *KeyPhrasesDetectionJobProperties) SetJobName(v string) *KeyPhrasesDetectionJobProperties { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *KeyPhrasesDetectionJobProperties) SetJobStatus(v string) *KeyPhrasesDetectionJobProperties { + s.JobStatus = &v + return s +} + +// SetLanguageCode sets the LanguageCode field's value. +func (s *KeyPhrasesDetectionJobProperties) SetLanguageCode(v string) *KeyPhrasesDetectionJobProperties { + s.LanguageCode = &v + return s +} + +// SetMessage sets the Message field's value. +func (s *KeyPhrasesDetectionJobProperties) SetMessage(v string) *KeyPhrasesDetectionJobProperties { + s.Message = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *KeyPhrasesDetectionJobProperties) SetOutputDataConfig(v *OutputDataConfig) *KeyPhrasesDetectionJobProperties { + s.OutputDataConfig = v + return s +} + +// SetSubmitTime sets the SubmitTime field's value. +func (s *KeyPhrasesDetectionJobProperties) SetSubmitTime(v time.Time) *KeyPhrasesDetectionJobProperties { + s.SubmitTime = &v + return s +} + +type ListDominantLanguageDetectionJobsInput struct { + _ struct{} `type:"structure"` + + // Filters that jobs that are returned. You can filter jobs on their name, status, + // or the date and time that they were submitted. You can only set one filter + // at a time. + Filter *DominantLanguageDetectionJobFilter `type:"structure"` + + // The maximum number of results to return in each page. The default is 100. + MaxResults *int64 `min:"1" type:"integer"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListDominantLanguageDetectionJobsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListDominantLanguageDetectionJobsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListDominantLanguageDetectionJobsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListDominantLanguageDetectionJobsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + if s.NextToken != nil && len(*s.NextToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NextToken", 1)) + } + if s.Filter != nil { + if err := s.Filter.Validate(); err != nil { + invalidParams.AddNested("Filter", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilter sets the Filter field's value. +func (s *ListDominantLanguageDetectionJobsInput) SetFilter(v *DominantLanguageDetectionJobFilter) *ListDominantLanguageDetectionJobsInput { + s.Filter = v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListDominantLanguageDetectionJobsInput) SetMaxResults(v int64) *ListDominantLanguageDetectionJobsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListDominantLanguageDetectionJobsInput) SetNextToken(v string) *ListDominantLanguageDetectionJobsInput { + s.NextToken = &v + return s +} + +type ListDominantLanguageDetectionJobsOutput struct { + _ struct{} `type:"structure"` + + // A list containing the properties of each job that is returned. + DominantLanguageDetectionJobPropertiesList []*DominantLanguageDetectionJobProperties `type:"list"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListDominantLanguageDetectionJobsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListDominantLanguageDetectionJobsOutput) GoString() string { + return s.String() +} + +// SetDominantLanguageDetectionJobPropertiesList sets the DominantLanguageDetectionJobPropertiesList field's value. +func (s *ListDominantLanguageDetectionJobsOutput) SetDominantLanguageDetectionJobPropertiesList(v []*DominantLanguageDetectionJobProperties) *ListDominantLanguageDetectionJobsOutput { + s.DominantLanguageDetectionJobPropertiesList = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListDominantLanguageDetectionJobsOutput) SetNextToken(v string) *ListDominantLanguageDetectionJobsOutput { + s.NextToken = &v + return s +} + +type ListEntitiesDetectionJobsInput struct { + _ struct{} `type:"structure"` + + // Filters the jobs that are returned. You can filter jobs on their name, status, + // or the date and time that they were submitted. You can only set one filter + // at a time. + Filter *EntitiesDetectionJobFilter `type:"structure"` + + // The maximum number of results to return in each page. The default is 100. + MaxResults *int64 `min:"1" type:"integer"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListEntitiesDetectionJobsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListEntitiesDetectionJobsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListEntitiesDetectionJobsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListEntitiesDetectionJobsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + if s.NextToken != nil && len(*s.NextToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NextToken", 1)) + } + if s.Filter != nil { + if err := s.Filter.Validate(); err != nil { + invalidParams.AddNested("Filter", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilter sets the Filter field's value. +func (s *ListEntitiesDetectionJobsInput) SetFilter(v *EntitiesDetectionJobFilter) *ListEntitiesDetectionJobsInput { + s.Filter = v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListEntitiesDetectionJobsInput) SetMaxResults(v int64) *ListEntitiesDetectionJobsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListEntitiesDetectionJobsInput) SetNextToken(v string) *ListEntitiesDetectionJobsInput { + s.NextToken = &v + return s +} + +type ListEntitiesDetectionJobsOutput struct { + _ struct{} `type:"structure"` + + // A list containing the properties of each job that is returned. + EntitiesDetectionJobPropertiesList []*EntitiesDetectionJobProperties `type:"list"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListEntitiesDetectionJobsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListEntitiesDetectionJobsOutput) GoString() string { + return s.String() +} + +// SetEntitiesDetectionJobPropertiesList sets the EntitiesDetectionJobPropertiesList field's value. +func (s *ListEntitiesDetectionJobsOutput) SetEntitiesDetectionJobPropertiesList(v []*EntitiesDetectionJobProperties) *ListEntitiesDetectionJobsOutput { + s.EntitiesDetectionJobPropertiesList = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListEntitiesDetectionJobsOutput) SetNextToken(v string) *ListEntitiesDetectionJobsOutput { + s.NextToken = &v + return s +} + +type ListKeyPhrasesDetectionJobsInput struct { + _ struct{} `type:"structure"` + + // Filters the jobs that are returned. You can filter jobs on their name, status, + // or the date and time that they were submitted. You can only set one filter + // at a time. + Filter *KeyPhrasesDetectionJobFilter `type:"structure"` + + // The maximum number of results to return in each page. The default is 100. + MaxResults *int64 `min:"1" type:"integer"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListKeyPhrasesDetectionJobsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListKeyPhrasesDetectionJobsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListKeyPhrasesDetectionJobsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListKeyPhrasesDetectionJobsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + if s.NextToken != nil && len(*s.NextToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NextToken", 1)) + } + if s.Filter != nil { + if err := s.Filter.Validate(); err != nil { + invalidParams.AddNested("Filter", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilter sets the Filter field's value. +func (s *ListKeyPhrasesDetectionJobsInput) SetFilter(v *KeyPhrasesDetectionJobFilter) *ListKeyPhrasesDetectionJobsInput { + s.Filter = v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListKeyPhrasesDetectionJobsInput) SetMaxResults(v int64) *ListKeyPhrasesDetectionJobsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListKeyPhrasesDetectionJobsInput) SetNextToken(v string) *ListKeyPhrasesDetectionJobsInput { + s.NextToken = &v + return s +} + +type ListKeyPhrasesDetectionJobsOutput struct { + _ struct{} `type:"structure"` + + // A list containing the properties of each job that is returned. + KeyPhrasesDetectionJobPropertiesList []*KeyPhrasesDetectionJobProperties `type:"list"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListKeyPhrasesDetectionJobsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListKeyPhrasesDetectionJobsOutput) GoString() string { + return s.String() +} + +// SetKeyPhrasesDetectionJobPropertiesList sets the KeyPhrasesDetectionJobPropertiesList field's value. +func (s *ListKeyPhrasesDetectionJobsOutput) SetKeyPhrasesDetectionJobPropertiesList(v []*KeyPhrasesDetectionJobProperties) *ListKeyPhrasesDetectionJobsOutput { + s.KeyPhrasesDetectionJobPropertiesList = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListKeyPhrasesDetectionJobsOutput) SetNextToken(v string) *ListKeyPhrasesDetectionJobsOutput { + s.NextToken = &v + return s +} + +type ListSentimentDetectionJobsInput struct { + _ struct{} `type:"structure"` + + // Filters the jobs that are returned. You can filter jobs on their name, status, + // or the date and time that they were submitted. You can only set one filter + // at a time. + Filter *SentimentDetectionJobFilter `type:"structure"` + + // The maximum number of results to return in each page. The default is 100. + MaxResults *int64 `min:"1" type:"integer"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s ListSentimentDetectionJobsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListSentimentDetectionJobsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListSentimentDetectionJobsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListSentimentDetectionJobsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + if s.NextToken != nil && len(*s.NextToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NextToken", 1)) + } + if s.Filter != nil { + if err := s.Filter.Validate(); err != nil { + invalidParams.AddNested("Filter", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilter sets the Filter field's value. +func (s *ListSentimentDetectionJobsInput) SetFilter(v *SentimentDetectionJobFilter) *ListSentimentDetectionJobsInput { + s.Filter = v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListSentimentDetectionJobsInput) SetMaxResults(v int64) *ListSentimentDetectionJobsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListSentimentDetectionJobsInput) SetNextToken(v string) *ListSentimentDetectionJobsInput { + s.NextToken = &v + return s +} + +type ListSentimentDetectionJobsOutput struct { + _ struct{} `type:"structure"` + + // Identifies the next page of results to return. + NextToken *string `min:"1" type:"string"` + + // A list containing the properties of each job that is returned. + SentimentDetectionJobPropertiesList []*SentimentDetectionJobProperties `type:"list"` +} + +// String returns the string representation +func (s ListSentimentDetectionJobsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListSentimentDetectionJobsOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *ListSentimentDetectionJobsOutput) SetNextToken(v string) *ListSentimentDetectionJobsOutput { + s.NextToken = &v + return s +} + +// SetSentimentDetectionJobPropertiesList sets the SentimentDetectionJobPropertiesList field's value. +func (s *ListSentimentDetectionJobsOutput) SetSentimentDetectionJobPropertiesList(v []*SentimentDetectionJobProperties) *ListSentimentDetectionJobsOutput { + s.SentimentDetectionJobPropertiesList = v + return s +} + type ListTopicsDetectionJobsInput struct { _ struct{} `type:"structure"` @@ -2272,7 +5059,7 @@ type ListTopicsDetectionJobsInput struct { // at a time. Filter *TopicsDetectionJobFilter `type:"structure"` - // The maximum number of results to return in each page. + // The maximum number of results to return in each page. The default is 100. MaxResults *int64 `min:"1" type:"integer"` // Identifies the next page of results to return. @@ -2364,13 +5151,16 @@ func (s *ListTopicsDetectionJobsOutput) SetTopicsDetectionJobPropertiesList(v [] type OutputDataConfig struct { _ struct{} `type:"structure"` - // The Amazon S3 URI where you want to write the output data. The URI must be - // in the same region as the API endpoint that you are calling. + // When you use the OutputDataConfig object with asynchronous operations, you + // specify the Amazon S3 location where you want to write the output data. The + // URI must be in the same region as the API endpoint that you are calling. + // The location is used as the prefix for the actual location of the output + // file. // - // The service creates an output file called output.tar.gz. It is a compressed - // archive that contains two files, topic-terms.csv that lists the terms associated - // with each topic, and doc-topics.csv that lists the documents associated with - // each topic. For more information, see topic-modeling. + // When the topic detection job is finished, the service creates an output file + // in a directory specific to the job. The S3Uri field contains the location + // of the output file, called output.tar.gz. It is a compressed archive that + // contains the ouput of the operation. // // S3Uri is a required field S3Uri *string `type:"string" required:"true"` @@ -2405,6 +5195,175 @@ func (s *OutputDataConfig) SetS3Uri(v string) *OutputDataConfig { return s } +// Provides information for filtering a list of dominant language detection +// jobs. For more information, see the operation. +type SentimentDetectionJobFilter struct { + _ struct{} `type:"structure"` + + // Filters on the name of the job. + JobName *string `min:"1" type:"string"` + + // Filters the list of jobs based on job status. Returns only jobs with the + // specified status. + JobStatus *string `type:"string" enum:"JobStatus"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted after the specified time. Jobs are + // returned in descending order, newest to oldest. + SubmitTimeAfter *time.Time `type:"timestamp" timestampFormat:"unix"` + + // Filters the list of jobs based on the time that the job was submitted for + // processing. Returns only jobs submitted before the specified time. Jobs are + // returned in ascending order, oldest to newest. + SubmitTimeBefore *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s SentimentDetectionJobFilter) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SentimentDetectionJobFilter) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *SentimentDetectionJobFilter) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "SentimentDetectionJobFilter"} + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobName sets the JobName field's value. +func (s *SentimentDetectionJobFilter) SetJobName(v string) *SentimentDetectionJobFilter { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *SentimentDetectionJobFilter) SetJobStatus(v string) *SentimentDetectionJobFilter { + s.JobStatus = &v + return s +} + +// SetSubmitTimeAfter sets the SubmitTimeAfter field's value. +func (s *SentimentDetectionJobFilter) SetSubmitTimeAfter(v time.Time) *SentimentDetectionJobFilter { + s.SubmitTimeAfter = &v + return s +} + +// SetSubmitTimeBefore sets the SubmitTimeBefore field's value. +func (s *SentimentDetectionJobFilter) SetSubmitTimeBefore(v time.Time) *SentimentDetectionJobFilter { + s.SubmitTimeBefore = &v + return s +} + +// Provides information about a sentiment detection job. +type SentimentDetectionJobProperties struct { + _ struct{} `type:"structure"` + + // The time that the sentiment detection job ended. + EndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The input data configuration that you supplied when you created the sentiment + // detection job. + InputDataConfig *InputDataConfig `type:"structure"` + + // The identifier assigned to the sentiment detection job. + JobId *string `min:"1" type:"string"` + + // The name that you assigned to the sentiment detection job + JobName *string `min:"1" type:"string"` + + // The current status of the sentiment detection job. If the status is FAILED, + // the Messages field shows the reason for the failure. + JobStatus *string `type:"string" enum:"JobStatus"` + + // The language code of the input documents. + LanguageCode *string `type:"string" enum:"LanguageCode"` + + // A description of the status of a job. + Message *string `type:"string"` + + // The output data configuration that you supplied when you created the sentiment + // detection job. + OutputDataConfig *OutputDataConfig `type:"structure"` + + // The time that the sentiment detection job was submitted for processing. + SubmitTime *time.Time `type:"timestamp" timestampFormat:"unix"` +} + +// String returns the string representation +func (s SentimentDetectionJobProperties) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SentimentDetectionJobProperties) GoString() string { + return s.String() +} + +// SetEndTime sets the EndTime field's value. +func (s *SentimentDetectionJobProperties) SetEndTime(v time.Time) *SentimentDetectionJobProperties { + s.EndTime = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *SentimentDetectionJobProperties) SetInputDataConfig(v *InputDataConfig) *SentimentDetectionJobProperties { + s.InputDataConfig = v + return s +} + +// SetJobId sets the JobId field's value. +func (s *SentimentDetectionJobProperties) SetJobId(v string) *SentimentDetectionJobProperties { + s.JobId = &v + return s +} + +// SetJobName sets the JobName field's value. +func (s *SentimentDetectionJobProperties) SetJobName(v string) *SentimentDetectionJobProperties { + s.JobName = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *SentimentDetectionJobProperties) SetJobStatus(v string) *SentimentDetectionJobProperties { + s.JobStatus = &v + return s +} + +// SetLanguageCode sets the LanguageCode field's value. +func (s *SentimentDetectionJobProperties) SetLanguageCode(v string) *SentimentDetectionJobProperties { + s.LanguageCode = &v + return s +} + +// SetMessage sets the Message field's value. +func (s *SentimentDetectionJobProperties) SetMessage(v string) *SentimentDetectionJobProperties { + s.Message = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *SentimentDetectionJobProperties) SetOutputDataConfig(v *OutputDataConfig) *SentimentDetectionJobProperties { + s.OutputDataConfig = v + return s +} + +// SetSubmitTime sets the SubmitTime field's value. +func (s *SentimentDetectionJobProperties) SetSubmitTime(v time.Time) *SentimentDetectionJobProperties { + s.SubmitTime = &v + return s +} + // Describes the level of confidence that Amazon Comprehend has in the accuracy // of its detection of sentiments. type SentimentScore struct { @@ -2461,6 +5420,623 @@ func (s *SentimentScore) SetPositive(v float64) *SentimentScore { return s } +type StartDominantLanguageDetectionJobInput struct { + _ struct{} `type:"structure"` + + // A unique identifier for the request. If you do not set the client request + // token, Amazon Comprehend generates one. + ClientRequestToken *string `min:"1" type:"string" idempotencyToken:"true"` + + // The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role + // that grants Amazon Comprehend read access to your input data. + // + // DataAccessRoleArn is a required field + DataAccessRoleArn *string `type:"string" required:"true"` + + // Specifies the format and location of the input data for the job. + // + // InputDataConfig is a required field + InputDataConfig *InputDataConfig `type:"structure" required:"true"` + + // An identifier for the job. + JobName *string `min:"1" type:"string"` + + // Specifies where to send the output files. + // + // OutputDataConfig is a required field + OutputDataConfig *OutputDataConfig `type:"structure" required:"true"` +} + +// String returns the string representation +func (s StartDominantLanguageDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartDominantLanguageDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StartDominantLanguageDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StartDominantLanguageDetectionJobInput"} + if s.ClientRequestToken != nil && len(*s.ClientRequestToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ClientRequestToken", 1)) + } + if s.DataAccessRoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("DataAccessRoleArn")) + } + if s.InputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("InputDataConfig")) + } + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + if s.OutputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("OutputDataConfig")) + } + if s.InputDataConfig != nil { + if err := s.InputDataConfig.Validate(); err != nil { + invalidParams.AddNested("InputDataConfig", err.(request.ErrInvalidParams)) + } + } + if s.OutputDataConfig != nil { + if err := s.OutputDataConfig.Validate(); err != nil { + invalidParams.AddNested("OutputDataConfig", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetClientRequestToken sets the ClientRequestToken field's value. +func (s *StartDominantLanguageDetectionJobInput) SetClientRequestToken(v string) *StartDominantLanguageDetectionJobInput { + s.ClientRequestToken = &v + return s +} + +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *StartDominantLanguageDetectionJobInput) SetDataAccessRoleArn(v string) *StartDominantLanguageDetectionJobInput { + s.DataAccessRoleArn = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *StartDominantLanguageDetectionJobInput) SetInputDataConfig(v *InputDataConfig) *StartDominantLanguageDetectionJobInput { + s.InputDataConfig = v + return s +} + +// SetJobName sets the JobName field's value. +func (s *StartDominantLanguageDetectionJobInput) SetJobName(v string) *StartDominantLanguageDetectionJobInput { + s.JobName = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *StartDominantLanguageDetectionJobInput) SetOutputDataConfig(v *OutputDataConfig) *StartDominantLanguageDetectionJobInput { + s.OutputDataConfig = v + return s +} + +type StartDominantLanguageDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier generated for the job. To get the status of a job, use this + // identifier with the operation. + JobId *string `min:"1" type:"string"` + + // The status of the job. + // + // * SUBMITTED - The job has been received and is queued for processing. + // + // * IN_PROGRESS - Amazon Comprehend is processing the job. + // + // * COMPLETED - The job was successfully completed and the output is available. + // + // * FAILED - The job did not complete. To get details, use the operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StartDominantLanguageDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartDominantLanguageDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StartDominantLanguageDetectionJobOutput) SetJobId(v string) *StartDominantLanguageDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StartDominantLanguageDetectionJobOutput) SetJobStatus(v string) *StartDominantLanguageDetectionJobOutput { + s.JobStatus = &v + return s +} + +type StartEntitiesDetectionJobInput struct { + _ struct{} `type:"structure"` + + // A unique identifier for the request. If you don't set the client request + // token, Amazon Comprehend generates one. + ClientRequestToken *string `min:"1" type:"string" idempotencyToken:"true"` + + // The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role + // that grants Amazon Comprehend read access to your input data. + // + // DataAccessRoleArn is a required field + DataAccessRoleArn *string `type:"string" required:"true"` + + // Specifies the format and location of the input data for the job. + // + // InputDataConfig is a required field + InputDataConfig *InputDataConfig `type:"structure" required:"true"` + + // The identifier of the job. + JobName *string `min:"1" type:"string"` + + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. + // + // LanguageCode is a required field + LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` + + // Specifies where to send the output files. + // + // OutputDataConfig is a required field + OutputDataConfig *OutputDataConfig `type:"structure" required:"true"` +} + +// String returns the string representation +func (s StartEntitiesDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartEntitiesDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StartEntitiesDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StartEntitiesDetectionJobInput"} + if s.ClientRequestToken != nil && len(*s.ClientRequestToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ClientRequestToken", 1)) + } + if s.DataAccessRoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("DataAccessRoleArn")) + } + if s.InputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("InputDataConfig")) + } + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + if s.LanguageCode == nil { + invalidParams.Add(request.NewErrParamRequired("LanguageCode")) + } + if s.OutputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("OutputDataConfig")) + } + if s.InputDataConfig != nil { + if err := s.InputDataConfig.Validate(); err != nil { + invalidParams.AddNested("InputDataConfig", err.(request.ErrInvalidParams)) + } + } + if s.OutputDataConfig != nil { + if err := s.OutputDataConfig.Validate(); err != nil { + invalidParams.AddNested("OutputDataConfig", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetClientRequestToken sets the ClientRequestToken field's value. +func (s *StartEntitiesDetectionJobInput) SetClientRequestToken(v string) *StartEntitiesDetectionJobInput { + s.ClientRequestToken = &v + return s +} + +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *StartEntitiesDetectionJobInput) SetDataAccessRoleArn(v string) *StartEntitiesDetectionJobInput { + s.DataAccessRoleArn = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *StartEntitiesDetectionJobInput) SetInputDataConfig(v *InputDataConfig) *StartEntitiesDetectionJobInput { + s.InputDataConfig = v + return s +} + +// SetJobName sets the JobName field's value. +func (s *StartEntitiesDetectionJobInput) SetJobName(v string) *StartEntitiesDetectionJobInput { + s.JobName = &v + return s +} + +// SetLanguageCode sets the LanguageCode field's value. +func (s *StartEntitiesDetectionJobInput) SetLanguageCode(v string) *StartEntitiesDetectionJobInput { + s.LanguageCode = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *StartEntitiesDetectionJobInput) SetOutputDataConfig(v *OutputDataConfig) *StartEntitiesDetectionJobInput { + s.OutputDataConfig = v + return s +} + +type StartEntitiesDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier generated for the job. To get the status of job, use this + // identifier with the operation. + JobId *string `min:"1" type:"string"` + + // The status of the job. + // + // * SUBMITTED - The job has been received and is queued for processing. + // + // * IN_PROGRESS - Amazon Comprehend is processing the job. + // + // * COMPLETED - The job was successfully completed and the output is available. + // + // * FAILED - The job did not complete. To get details, use the operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StartEntitiesDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartEntitiesDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StartEntitiesDetectionJobOutput) SetJobId(v string) *StartEntitiesDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StartEntitiesDetectionJobOutput) SetJobStatus(v string) *StartEntitiesDetectionJobOutput { + s.JobStatus = &v + return s +} + +type StartKeyPhrasesDetectionJobInput struct { + _ struct{} `type:"structure"` + + // A unique identifier for the request. If you don't set the client request + // token, Amazon Comprehend generates one. + ClientRequestToken *string `min:"1" type:"string" idempotencyToken:"true"` + + // The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role + // that grants Amazon Comprehend read access to your input data. + // + // DataAccessRoleArn is a required field + DataAccessRoleArn *string `type:"string" required:"true"` + + // Specifies the format and location of the input data for the job. + // + // InputDataConfig is a required field + InputDataConfig *InputDataConfig `type:"structure" required:"true"` + + // The identifier of the job. + JobName *string `min:"1" type:"string"` + + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. + // + // LanguageCode is a required field + LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` + + // Specifies where to send the output files. + // + // OutputDataConfig is a required field + OutputDataConfig *OutputDataConfig `type:"structure" required:"true"` +} + +// String returns the string representation +func (s StartKeyPhrasesDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartKeyPhrasesDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StartKeyPhrasesDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StartKeyPhrasesDetectionJobInput"} + if s.ClientRequestToken != nil && len(*s.ClientRequestToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ClientRequestToken", 1)) + } + if s.DataAccessRoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("DataAccessRoleArn")) + } + if s.InputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("InputDataConfig")) + } + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + if s.LanguageCode == nil { + invalidParams.Add(request.NewErrParamRequired("LanguageCode")) + } + if s.OutputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("OutputDataConfig")) + } + if s.InputDataConfig != nil { + if err := s.InputDataConfig.Validate(); err != nil { + invalidParams.AddNested("InputDataConfig", err.(request.ErrInvalidParams)) + } + } + if s.OutputDataConfig != nil { + if err := s.OutputDataConfig.Validate(); err != nil { + invalidParams.AddNested("OutputDataConfig", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetClientRequestToken sets the ClientRequestToken field's value. +func (s *StartKeyPhrasesDetectionJobInput) SetClientRequestToken(v string) *StartKeyPhrasesDetectionJobInput { + s.ClientRequestToken = &v + return s +} + +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *StartKeyPhrasesDetectionJobInput) SetDataAccessRoleArn(v string) *StartKeyPhrasesDetectionJobInput { + s.DataAccessRoleArn = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *StartKeyPhrasesDetectionJobInput) SetInputDataConfig(v *InputDataConfig) *StartKeyPhrasesDetectionJobInput { + s.InputDataConfig = v + return s +} + +// SetJobName sets the JobName field's value. +func (s *StartKeyPhrasesDetectionJobInput) SetJobName(v string) *StartKeyPhrasesDetectionJobInput { + s.JobName = &v + return s +} + +// SetLanguageCode sets the LanguageCode field's value. +func (s *StartKeyPhrasesDetectionJobInput) SetLanguageCode(v string) *StartKeyPhrasesDetectionJobInput { + s.LanguageCode = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *StartKeyPhrasesDetectionJobInput) SetOutputDataConfig(v *OutputDataConfig) *StartKeyPhrasesDetectionJobInput { + s.OutputDataConfig = v + return s +} + +type StartKeyPhrasesDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier generated for the job. To get the status of a job, use this + // identifier with the operation. + JobId *string `min:"1" type:"string"` + + // The status of the job. + // + // * SUBMITTED - The job has been received and is queued for processing. + // + // * IN_PROGRESS - Amazon Comprehend is processing the job. + // + // * COMPLETED - The job was successfully completed and the output is available. + // + // * FAILED - The job did not complete. To get details, use the operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StartKeyPhrasesDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartKeyPhrasesDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StartKeyPhrasesDetectionJobOutput) SetJobId(v string) *StartKeyPhrasesDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StartKeyPhrasesDetectionJobOutput) SetJobStatus(v string) *StartKeyPhrasesDetectionJobOutput { + s.JobStatus = &v + return s +} + +type StartSentimentDetectionJobInput struct { + _ struct{} `type:"structure"` + + // A unique identifier for the request. If you don't set the client request + // token, Amazon Comprehend generates one. + ClientRequestToken *string `min:"1" type:"string" idempotencyToken:"true"` + + // The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role + // that grants Amazon Comprehend read access to your input data. + // + // DataAccessRoleArn is a required field + DataAccessRoleArn *string `type:"string" required:"true"` + + // Specifies the format and location of the input data for the job. + // + // InputDataConfig is a required field + InputDataConfig *InputDataConfig `type:"structure" required:"true"` + + // The identifier of the job. + JobName *string `min:"1" type:"string"` + + // The language of the input documents. You can specify English ("en") or Spanish + // ("es"). All documents must be in the same language. + // + // LanguageCode is a required field + LanguageCode *string `type:"string" required:"true" enum:"LanguageCode"` + + // Specifies where to send the output files. + // + // OutputDataConfig is a required field + OutputDataConfig *OutputDataConfig `type:"structure" required:"true"` +} + +// String returns the string representation +func (s StartSentimentDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartSentimentDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StartSentimentDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StartSentimentDetectionJobInput"} + if s.ClientRequestToken != nil && len(*s.ClientRequestToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ClientRequestToken", 1)) + } + if s.DataAccessRoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("DataAccessRoleArn")) + } + if s.InputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("InputDataConfig")) + } + if s.JobName != nil && len(*s.JobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobName", 1)) + } + if s.LanguageCode == nil { + invalidParams.Add(request.NewErrParamRequired("LanguageCode")) + } + if s.OutputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("OutputDataConfig")) + } + if s.InputDataConfig != nil { + if err := s.InputDataConfig.Validate(); err != nil { + invalidParams.AddNested("InputDataConfig", err.(request.ErrInvalidParams)) + } + } + if s.OutputDataConfig != nil { + if err := s.OutputDataConfig.Validate(); err != nil { + invalidParams.AddNested("OutputDataConfig", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetClientRequestToken sets the ClientRequestToken field's value. +func (s *StartSentimentDetectionJobInput) SetClientRequestToken(v string) *StartSentimentDetectionJobInput { + s.ClientRequestToken = &v + return s +} + +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *StartSentimentDetectionJobInput) SetDataAccessRoleArn(v string) *StartSentimentDetectionJobInput { + s.DataAccessRoleArn = &v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *StartSentimentDetectionJobInput) SetInputDataConfig(v *InputDataConfig) *StartSentimentDetectionJobInput { + s.InputDataConfig = v + return s +} + +// SetJobName sets the JobName field's value. +func (s *StartSentimentDetectionJobInput) SetJobName(v string) *StartSentimentDetectionJobInput { + s.JobName = &v + return s +} + +// SetLanguageCode sets the LanguageCode field's value. +func (s *StartSentimentDetectionJobInput) SetLanguageCode(v string) *StartSentimentDetectionJobInput { + s.LanguageCode = &v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *StartSentimentDetectionJobInput) SetOutputDataConfig(v *OutputDataConfig) *StartSentimentDetectionJobInput { + s.OutputDataConfig = v + return s +} + +type StartSentimentDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier generated for the job. To get the status of a job, use this + // identifier with the operation. + JobId *string `min:"1" type:"string"` + + // The status of the job. + // + // * SUBMITTED - The job has been received and is queued for processing. + // + // * IN_PROGRESS - Amazon Comprehend is processing the job. + // + // * COMPLETED - The job was successfully completed and the output is available. + // + // * FAILED - The job did not complete. To get details, use the operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StartSentimentDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StartSentimentDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StartSentimentDetectionJobOutput) SetJobId(v string) *StartSentimentDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StartSentimentDetectionJobOutput) SetJobStatus(v string) *StartSentimentDetectionJobOutput { + s.JobStatus = &v + return s +} + type StartTopicsDetectionJobInput struct { _ struct{} `type:"structure"` @@ -2485,7 +6061,9 @@ type StartTopicsDetectionJobInput struct { // The number of topics to detect. NumberOfTopics *int64 `min:"1" type:"integer"` - // Specifies where to send the output files. + // Specifies where to send the output files. The output is a compressed archive + // with two files, topic-terms.csv that lists the terms associated with each + // topic, and doc-topics.csv that lists the documents associated with each topic // // OutputDataConfig is a required field OutputDataConfig *OutputDataConfig `type:"structure" required:"true"` @@ -2617,6 +6195,302 @@ func (s *StartTopicsDetectionJobOutput) SetJobStatus(v string) *StartTopicsDetec return s } +type StopDominantLanguageDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier of the dominant language detection job to stop. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s StopDominantLanguageDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopDominantLanguageDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StopDominantLanguageDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StopDominantLanguageDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *StopDominantLanguageDetectionJobInput) SetJobId(v string) *StopDominantLanguageDetectionJobInput { + s.JobId = &v + return s +} + +type StopDominantLanguageDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier of the dominant language detection job to stop. + JobId *string `min:"1" type:"string"` + + // Either STOPPING if the job is currently running, or STOPPED if the job was + // previously stopped with the StopDominantLanguageDetectionJob operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StopDominantLanguageDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopDominantLanguageDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StopDominantLanguageDetectionJobOutput) SetJobId(v string) *StopDominantLanguageDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StopDominantLanguageDetectionJobOutput) SetJobStatus(v string) *StopDominantLanguageDetectionJobOutput { + s.JobStatus = &v + return s +} + +type StopEntitiesDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier of the entities detection job to stop. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s StopEntitiesDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopEntitiesDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StopEntitiesDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StopEntitiesDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *StopEntitiesDetectionJobInput) SetJobId(v string) *StopEntitiesDetectionJobInput { + s.JobId = &v + return s +} + +type StopEntitiesDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier of the entities detection job to stop. + JobId *string `min:"1" type:"string"` + + // Either STOPPING if the job is currently running, or STOPPED if the job was + // previously stopped with the StopEntitiesDetectionJob operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StopEntitiesDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopEntitiesDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StopEntitiesDetectionJobOutput) SetJobId(v string) *StopEntitiesDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StopEntitiesDetectionJobOutput) SetJobStatus(v string) *StopEntitiesDetectionJobOutput { + s.JobStatus = &v + return s +} + +type StopKeyPhrasesDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier of the key phrases detection job to stop. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s StopKeyPhrasesDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopKeyPhrasesDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StopKeyPhrasesDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StopKeyPhrasesDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *StopKeyPhrasesDetectionJobInput) SetJobId(v string) *StopKeyPhrasesDetectionJobInput { + s.JobId = &v + return s +} + +type StopKeyPhrasesDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier of the key phrases detection job to stop. + JobId *string `min:"1" type:"string"` + + // Either STOPPING if the job is currently running, or STOPPED if the job was + // previously stopped with the StopKeyPhrasesDetectionJob operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StopKeyPhrasesDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopKeyPhrasesDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StopKeyPhrasesDetectionJobOutput) SetJobId(v string) *StopKeyPhrasesDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StopKeyPhrasesDetectionJobOutput) SetJobStatus(v string) *StopKeyPhrasesDetectionJobOutput { + s.JobStatus = &v + return s +} + +type StopSentimentDetectionJobInput struct { + _ struct{} `type:"structure"` + + // The identifier of the sentiment detection job to stop. + // + // JobId is a required field + JobId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s StopSentimentDetectionJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopSentimentDetectionJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StopSentimentDetectionJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StopSentimentDetectionJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetJobId sets the JobId field's value. +func (s *StopSentimentDetectionJobInput) SetJobId(v string) *StopSentimentDetectionJobInput { + s.JobId = &v + return s +} + +type StopSentimentDetectionJobOutput struct { + _ struct{} `type:"structure"` + + // The identifier of the sentiment detection job to stop. + JobId *string `min:"1" type:"string"` + + // Either STOPPING if the job is currently running, or STOPPED if the job was + // previously stopped with the StopSentimentDetectionJob operation. + JobStatus *string `type:"string" enum:"JobStatus"` +} + +// String returns the string representation +func (s StopSentimentDetectionJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopSentimentDetectionJobOutput) GoString() string { + return s.String() +} + +// SetJobId sets the JobId field's value. +func (s *StopSentimentDetectionJobOutput) SetJobId(v string) *StopSentimentDetectionJobOutput { + s.JobId = &v + return s +} + +// SetJobStatus sets the JobStatus field's value. +func (s *StopSentimentDetectionJobOutput) SetJobStatus(v string) *StopSentimentDetectionJobOutput { + s.JobStatus = &v + return s +} + // Provides information for filtering topic detection jobs. For more information, // see . type TopicsDetectionJobFilter struct { @@ -2835,6 +6709,12 @@ const ( // JobStatusFailed is a JobStatus enum value JobStatusFailed = "FAILED" + + // JobStatusStopRequested is a JobStatus enum value + JobStatusStopRequested = "STOP_REQUESTED" + + // JobStatusStopped is a JobStatus enum value + JobStatusStopped = "STOPPED" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/comprehend/comprehendiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/comprehend/comprehendiface/interface.go index 5a473127c..c20f44016 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/comprehend/comprehendiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/comprehend/comprehendiface/interface.go @@ -76,6 +76,22 @@ type ComprehendAPI interface { BatchDetectSentimentWithContext(aws.Context, *comprehend.BatchDetectSentimentInput, ...request.Option) (*comprehend.BatchDetectSentimentOutput, error) BatchDetectSentimentRequest(*comprehend.BatchDetectSentimentInput) (*request.Request, *comprehend.BatchDetectSentimentOutput) + DescribeDominantLanguageDetectionJob(*comprehend.DescribeDominantLanguageDetectionJobInput) (*comprehend.DescribeDominantLanguageDetectionJobOutput, error) + DescribeDominantLanguageDetectionJobWithContext(aws.Context, *comprehend.DescribeDominantLanguageDetectionJobInput, ...request.Option) (*comprehend.DescribeDominantLanguageDetectionJobOutput, error) + DescribeDominantLanguageDetectionJobRequest(*comprehend.DescribeDominantLanguageDetectionJobInput) (*request.Request, *comprehend.DescribeDominantLanguageDetectionJobOutput) + + DescribeEntitiesDetectionJob(*comprehend.DescribeEntitiesDetectionJobInput) (*comprehend.DescribeEntitiesDetectionJobOutput, error) + DescribeEntitiesDetectionJobWithContext(aws.Context, *comprehend.DescribeEntitiesDetectionJobInput, ...request.Option) (*comprehend.DescribeEntitiesDetectionJobOutput, error) + DescribeEntitiesDetectionJobRequest(*comprehend.DescribeEntitiesDetectionJobInput) (*request.Request, *comprehend.DescribeEntitiesDetectionJobOutput) + + DescribeKeyPhrasesDetectionJob(*comprehend.DescribeKeyPhrasesDetectionJobInput) (*comprehend.DescribeKeyPhrasesDetectionJobOutput, error) + DescribeKeyPhrasesDetectionJobWithContext(aws.Context, *comprehend.DescribeKeyPhrasesDetectionJobInput, ...request.Option) (*comprehend.DescribeKeyPhrasesDetectionJobOutput, error) + DescribeKeyPhrasesDetectionJobRequest(*comprehend.DescribeKeyPhrasesDetectionJobInput) (*request.Request, *comprehend.DescribeKeyPhrasesDetectionJobOutput) + + DescribeSentimentDetectionJob(*comprehend.DescribeSentimentDetectionJobInput) (*comprehend.DescribeSentimentDetectionJobOutput, error) + DescribeSentimentDetectionJobWithContext(aws.Context, *comprehend.DescribeSentimentDetectionJobInput, ...request.Option) (*comprehend.DescribeSentimentDetectionJobOutput, error) + DescribeSentimentDetectionJobRequest(*comprehend.DescribeSentimentDetectionJobInput) (*request.Request, *comprehend.DescribeSentimentDetectionJobOutput) + DescribeTopicsDetectionJob(*comprehend.DescribeTopicsDetectionJobInput) (*comprehend.DescribeTopicsDetectionJobOutput, error) DescribeTopicsDetectionJobWithContext(aws.Context, *comprehend.DescribeTopicsDetectionJobInput, ...request.Option) (*comprehend.DescribeTopicsDetectionJobOutput, error) DescribeTopicsDetectionJobRequest(*comprehend.DescribeTopicsDetectionJobInput) (*request.Request, *comprehend.DescribeTopicsDetectionJobOutput) @@ -96,6 +112,34 @@ type ComprehendAPI interface { DetectSentimentWithContext(aws.Context, *comprehend.DetectSentimentInput, ...request.Option) (*comprehend.DetectSentimentOutput, error) DetectSentimentRequest(*comprehend.DetectSentimentInput) (*request.Request, *comprehend.DetectSentimentOutput) + ListDominantLanguageDetectionJobs(*comprehend.ListDominantLanguageDetectionJobsInput) (*comprehend.ListDominantLanguageDetectionJobsOutput, error) + ListDominantLanguageDetectionJobsWithContext(aws.Context, *comprehend.ListDominantLanguageDetectionJobsInput, ...request.Option) (*comprehend.ListDominantLanguageDetectionJobsOutput, error) + ListDominantLanguageDetectionJobsRequest(*comprehend.ListDominantLanguageDetectionJobsInput) (*request.Request, *comprehend.ListDominantLanguageDetectionJobsOutput) + + ListDominantLanguageDetectionJobsPages(*comprehend.ListDominantLanguageDetectionJobsInput, func(*comprehend.ListDominantLanguageDetectionJobsOutput, bool) bool) error + ListDominantLanguageDetectionJobsPagesWithContext(aws.Context, *comprehend.ListDominantLanguageDetectionJobsInput, func(*comprehend.ListDominantLanguageDetectionJobsOutput, bool) bool, ...request.Option) error + + ListEntitiesDetectionJobs(*comprehend.ListEntitiesDetectionJobsInput) (*comprehend.ListEntitiesDetectionJobsOutput, error) + ListEntitiesDetectionJobsWithContext(aws.Context, *comprehend.ListEntitiesDetectionJobsInput, ...request.Option) (*comprehend.ListEntitiesDetectionJobsOutput, error) + ListEntitiesDetectionJobsRequest(*comprehend.ListEntitiesDetectionJobsInput) (*request.Request, *comprehend.ListEntitiesDetectionJobsOutput) + + ListEntitiesDetectionJobsPages(*comprehend.ListEntitiesDetectionJobsInput, func(*comprehend.ListEntitiesDetectionJobsOutput, bool) bool) error + ListEntitiesDetectionJobsPagesWithContext(aws.Context, *comprehend.ListEntitiesDetectionJobsInput, func(*comprehend.ListEntitiesDetectionJobsOutput, bool) bool, ...request.Option) error + + ListKeyPhrasesDetectionJobs(*comprehend.ListKeyPhrasesDetectionJobsInput) (*comprehend.ListKeyPhrasesDetectionJobsOutput, error) + ListKeyPhrasesDetectionJobsWithContext(aws.Context, *comprehend.ListKeyPhrasesDetectionJobsInput, ...request.Option) (*comprehend.ListKeyPhrasesDetectionJobsOutput, error) + ListKeyPhrasesDetectionJobsRequest(*comprehend.ListKeyPhrasesDetectionJobsInput) (*request.Request, *comprehend.ListKeyPhrasesDetectionJobsOutput) + + ListKeyPhrasesDetectionJobsPages(*comprehend.ListKeyPhrasesDetectionJobsInput, func(*comprehend.ListKeyPhrasesDetectionJobsOutput, bool) bool) error + ListKeyPhrasesDetectionJobsPagesWithContext(aws.Context, *comprehend.ListKeyPhrasesDetectionJobsInput, func(*comprehend.ListKeyPhrasesDetectionJobsOutput, bool) bool, ...request.Option) error + + ListSentimentDetectionJobs(*comprehend.ListSentimentDetectionJobsInput) (*comprehend.ListSentimentDetectionJobsOutput, error) + ListSentimentDetectionJobsWithContext(aws.Context, *comprehend.ListSentimentDetectionJobsInput, ...request.Option) (*comprehend.ListSentimentDetectionJobsOutput, error) + ListSentimentDetectionJobsRequest(*comprehend.ListSentimentDetectionJobsInput) (*request.Request, *comprehend.ListSentimentDetectionJobsOutput) + + ListSentimentDetectionJobsPages(*comprehend.ListSentimentDetectionJobsInput, func(*comprehend.ListSentimentDetectionJobsOutput, bool) bool) error + ListSentimentDetectionJobsPagesWithContext(aws.Context, *comprehend.ListSentimentDetectionJobsInput, func(*comprehend.ListSentimentDetectionJobsOutput, bool) bool, ...request.Option) error + ListTopicsDetectionJobs(*comprehend.ListTopicsDetectionJobsInput) (*comprehend.ListTopicsDetectionJobsOutput, error) ListTopicsDetectionJobsWithContext(aws.Context, *comprehend.ListTopicsDetectionJobsInput, ...request.Option) (*comprehend.ListTopicsDetectionJobsOutput, error) ListTopicsDetectionJobsRequest(*comprehend.ListTopicsDetectionJobsInput) (*request.Request, *comprehend.ListTopicsDetectionJobsOutput) @@ -103,9 +147,41 @@ type ComprehendAPI interface { ListTopicsDetectionJobsPages(*comprehend.ListTopicsDetectionJobsInput, func(*comprehend.ListTopicsDetectionJobsOutput, bool) bool) error ListTopicsDetectionJobsPagesWithContext(aws.Context, *comprehend.ListTopicsDetectionJobsInput, func(*comprehend.ListTopicsDetectionJobsOutput, bool) bool, ...request.Option) error + StartDominantLanguageDetectionJob(*comprehend.StartDominantLanguageDetectionJobInput) (*comprehend.StartDominantLanguageDetectionJobOutput, error) + StartDominantLanguageDetectionJobWithContext(aws.Context, *comprehend.StartDominantLanguageDetectionJobInput, ...request.Option) (*comprehend.StartDominantLanguageDetectionJobOutput, error) + StartDominantLanguageDetectionJobRequest(*comprehend.StartDominantLanguageDetectionJobInput) (*request.Request, *comprehend.StartDominantLanguageDetectionJobOutput) + + StartEntitiesDetectionJob(*comprehend.StartEntitiesDetectionJobInput) (*comprehend.StartEntitiesDetectionJobOutput, error) + StartEntitiesDetectionJobWithContext(aws.Context, *comprehend.StartEntitiesDetectionJobInput, ...request.Option) (*comprehend.StartEntitiesDetectionJobOutput, error) + StartEntitiesDetectionJobRequest(*comprehend.StartEntitiesDetectionJobInput) (*request.Request, *comprehend.StartEntitiesDetectionJobOutput) + + StartKeyPhrasesDetectionJob(*comprehend.StartKeyPhrasesDetectionJobInput) (*comprehend.StartKeyPhrasesDetectionJobOutput, error) + StartKeyPhrasesDetectionJobWithContext(aws.Context, *comprehend.StartKeyPhrasesDetectionJobInput, ...request.Option) (*comprehend.StartKeyPhrasesDetectionJobOutput, error) + StartKeyPhrasesDetectionJobRequest(*comprehend.StartKeyPhrasesDetectionJobInput) (*request.Request, *comprehend.StartKeyPhrasesDetectionJobOutput) + + StartSentimentDetectionJob(*comprehend.StartSentimentDetectionJobInput) (*comprehend.StartSentimentDetectionJobOutput, error) + StartSentimentDetectionJobWithContext(aws.Context, *comprehend.StartSentimentDetectionJobInput, ...request.Option) (*comprehend.StartSentimentDetectionJobOutput, error) + StartSentimentDetectionJobRequest(*comprehend.StartSentimentDetectionJobInput) (*request.Request, *comprehend.StartSentimentDetectionJobOutput) + StartTopicsDetectionJob(*comprehend.StartTopicsDetectionJobInput) (*comprehend.StartTopicsDetectionJobOutput, error) StartTopicsDetectionJobWithContext(aws.Context, *comprehend.StartTopicsDetectionJobInput, ...request.Option) (*comprehend.StartTopicsDetectionJobOutput, error) StartTopicsDetectionJobRequest(*comprehend.StartTopicsDetectionJobInput) (*request.Request, *comprehend.StartTopicsDetectionJobOutput) + + StopDominantLanguageDetectionJob(*comprehend.StopDominantLanguageDetectionJobInput) (*comprehend.StopDominantLanguageDetectionJobOutput, error) + StopDominantLanguageDetectionJobWithContext(aws.Context, *comprehend.StopDominantLanguageDetectionJobInput, ...request.Option) (*comprehend.StopDominantLanguageDetectionJobOutput, error) + StopDominantLanguageDetectionJobRequest(*comprehend.StopDominantLanguageDetectionJobInput) (*request.Request, *comprehend.StopDominantLanguageDetectionJobOutput) + + StopEntitiesDetectionJob(*comprehend.StopEntitiesDetectionJobInput) (*comprehend.StopEntitiesDetectionJobOutput, error) + StopEntitiesDetectionJobWithContext(aws.Context, *comprehend.StopEntitiesDetectionJobInput, ...request.Option) (*comprehend.StopEntitiesDetectionJobOutput, error) + StopEntitiesDetectionJobRequest(*comprehend.StopEntitiesDetectionJobInput) (*request.Request, *comprehend.StopEntitiesDetectionJobOutput) + + StopKeyPhrasesDetectionJob(*comprehend.StopKeyPhrasesDetectionJobInput) (*comprehend.StopKeyPhrasesDetectionJobOutput, error) + StopKeyPhrasesDetectionJobWithContext(aws.Context, *comprehend.StopKeyPhrasesDetectionJobInput, ...request.Option) (*comprehend.StopKeyPhrasesDetectionJobOutput, error) + StopKeyPhrasesDetectionJobRequest(*comprehend.StopKeyPhrasesDetectionJobInput) (*request.Request, *comprehend.StopKeyPhrasesDetectionJobOutput) + + StopSentimentDetectionJob(*comprehend.StopSentimentDetectionJobInput) (*comprehend.StopSentimentDetectionJobOutput, error) + StopSentimentDetectionJobWithContext(aws.Context, *comprehend.StopSentimentDetectionJobInput, ...request.Option) (*comprehend.StopSentimentDetectionJobOutput, error) + StopSentimentDetectionJobRequest(*comprehend.StopSentimentDetectionJobInput) (*request.Request, *comprehend.StopSentimentDetectionJobOutput) } var _ ComprehendAPI = (*comprehend.Comprehend)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/comprehend/service.go b/vendor/github.com/aws/aws-sdk-go/service/comprehend/service.go index 933eb819d..1cb8f88f6 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/comprehend/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/comprehend/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "comprehend" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "comprehend" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Comprehend" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Comprehend client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/configservice/api.go b/vendor/github.com/aws/aws-sdk-go/service/configservice/api.go index e4e77c934..2b9dfc548 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/configservice/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/configservice/api.go @@ -709,6 +709,91 @@ func (c *ConfigService) DeletePendingAggregationRequestWithContext(ctx aws.Conte return out, req.Send() } +const opDeleteRetentionConfiguration = "DeleteRetentionConfiguration" + +// DeleteRetentionConfigurationRequest generates a "aws/request.Request" representing the +// client's request for the DeleteRetentionConfiguration operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteRetentionConfiguration for more information on using the DeleteRetentionConfiguration +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteRetentionConfigurationRequest method. +// req, resp := client.DeleteRetentionConfigurationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/config-2014-11-12/DeleteRetentionConfiguration +func (c *ConfigService) DeleteRetentionConfigurationRequest(input *DeleteRetentionConfigurationInput) (req *request.Request, output *DeleteRetentionConfigurationOutput) { + op := &request.Operation{ + Name: opDeleteRetentionConfiguration, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteRetentionConfigurationInput{} + } + + output = &DeleteRetentionConfigurationOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(jsonrpc.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeleteRetentionConfiguration API operation for AWS Config. +// +// Deletes the retention configuration. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Config's +// API operation DeleteRetentionConfiguration for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidParameterValueException "InvalidParameterValueException" +// One or more of the specified parameters are invalid. Verify that your parameters +// are valid and try again. +// +// * ErrCodeNoSuchRetentionConfigurationException "NoSuchRetentionConfigurationException" +// You have specified a retention configuration that does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/config-2014-11-12/DeleteRetentionConfiguration +func (c *ConfigService) DeleteRetentionConfiguration(input *DeleteRetentionConfigurationInput) (*DeleteRetentionConfigurationOutput, error) { + req, out := c.DeleteRetentionConfigurationRequest(input) + return out, req.Send() +} + +// DeleteRetentionConfigurationWithContext is the same as DeleteRetentionConfiguration with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteRetentionConfiguration for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *ConfigService) DeleteRetentionConfigurationWithContext(ctx aws.Context, input *DeleteRetentionConfigurationInput, opts ...request.Option) (*DeleteRetentionConfigurationOutput, error) { + req, out := c.DeleteRetentionConfigurationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDeliverConfigSnapshot = "DeliverConfigSnapshot" // DeliverConfigSnapshotRequest generates a "aws/request.Request" representing the @@ -1985,6 +2070,98 @@ func (c *ConfigService) DescribePendingAggregationRequestsWithContext(ctx aws.Co return out, req.Send() } +const opDescribeRetentionConfigurations = "DescribeRetentionConfigurations" + +// DescribeRetentionConfigurationsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeRetentionConfigurations operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeRetentionConfigurations for more information on using the DescribeRetentionConfigurations +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeRetentionConfigurationsRequest method. +// req, resp := client.DescribeRetentionConfigurationsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/config-2014-11-12/DescribeRetentionConfigurations +func (c *ConfigService) DescribeRetentionConfigurationsRequest(input *DescribeRetentionConfigurationsInput) (req *request.Request, output *DescribeRetentionConfigurationsOutput) { + op := &request.Operation{ + Name: opDescribeRetentionConfigurations, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeRetentionConfigurationsInput{} + } + + output = &DescribeRetentionConfigurationsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeRetentionConfigurations API operation for AWS Config. +// +// Returns the details of one or more retention configurations. If the retention +// configuration name is not specified, this action returns the details for +// all the retention configurations for that account. +// +// Currently, AWS Config supports only one retention configuration per region +// in your account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Config's +// API operation DescribeRetentionConfigurations for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidParameterValueException "InvalidParameterValueException" +// One or more of the specified parameters are invalid. Verify that your parameters +// are valid and try again. +// +// * ErrCodeNoSuchRetentionConfigurationException "NoSuchRetentionConfigurationException" +// You have specified a retention configuration that does not exist. +// +// * ErrCodeInvalidNextTokenException "InvalidNextTokenException" +// The specified next token is invalid. Specify the nextToken string that was +// returned in the previous response to get the next page of results. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/config-2014-11-12/DescribeRetentionConfigurations +func (c *ConfigService) DescribeRetentionConfigurations(input *DescribeRetentionConfigurationsInput) (*DescribeRetentionConfigurationsOutput, error) { + req, out := c.DescribeRetentionConfigurationsRequest(input) + return out, req.Send() +} + +// DescribeRetentionConfigurationsWithContext is the same as DescribeRetentionConfigurations with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeRetentionConfigurations for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *ConfigService) DescribeRetentionConfigurationsWithContext(ctx aws.Context, input *DescribeRetentionConfigurationsInput, opts ...request.Option) (*DescribeRetentionConfigurationsOutput, error) { + req, out := c.DescribeRetentionConfigurationsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opGetAggregateComplianceDetailsByConfigRule = "GetAggregateComplianceDetailsByConfigRule" // GetAggregateComplianceDetailsByConfigRuleRequest generates a "aws/request.Request" representing the @@ -2673,7 +2850,9 @@ func (c *ConfigService) GetResourceConfigHistoryRequest(input *GetResourceConfig // // Returns a list of configuration items for the specified resource. The list // contains details about each state of the resource during the specified time -// interval. +// interval. If you specified a retention period to retain your ConfigurationItems +// between a minimum of 30 days and a maximum of 7 years (2557 days), AWS Config +// returns the ConfigurationItems for the specified retention period. // // The response is paginated. By default, AWS Config returns a limit of 10 configuration // items per page. You can customize this number with the limit parameter. The @@ -3517,6 +3696,97 @@ func (c *ConfigService) PutEvaluationsWithContext(ctx aws.Context, input *PutEva return out, req.Send() } +const opPutRetentionConfiguration = "PutRetentionConfiguration" + +// PutRetentionConfigurationRequest generates a "aws/request.Request" representing the +// client's request for the PutRetentionConfiguration operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See PutRetentionConfiguration for more information on using the PutRetentionConfiguration +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the PutRetentionConfigurationRequest method. +// req, resp := client.PutRetentionConfigurationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/config-2014-11-12/PutRetentionConfiguration +func (c *ConfigService) PutRetentionConfigurationRequest(input *PutRetentionConfigurationInput) (req *request.Request, output *PutRetentionConfigurationOutput) { + op := &request.Operation{ + Name: opPutRetentionConfiguration, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &PutRetentionConfigurationInput{} + } + + output = &PutRetentionConfigurationOutput{} + req = c.newRequest(op, input, output) + return +} + +// PutRetentionConfiguration API operation for AWS Config. +// +// Creates and updates the retention configuration with details about retention +// period (number of days) that AWS Config stores your historical information. +// The API creates the RetentionConfiguration object and names the object as +// default. When you have a RetentionConfiguration object named default, calling +// the API modifies the default object. +// +// Currently, AWS Config supports only one retention configuration per region +// in your account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Config's +// API operation PutRetentionConfiguration for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidParameterValueException "InvalidParameterValueException" +// One or more of the specified parameters are invalid. Verify that your parameters +// are valid and try again. +// +// * ErrCodeMaxNumberOfRetentionConfigurationsExceededException "MaxNumberOfRetentionConfigurationsExceededException" +// Failed to add the retention configuration because a retention configuration +// with that name already exists. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/config-2014-11-12/PutRetentionConfiguration +func (c *ConfigService) PutRetentionConfiguration(input *PutRetentionConfigurationInput) (*PutRetentionConfigurationOutput, error) { + req, out := c.PutRetentionConfigurationRequest(input) + return out, req.Send() +} + +// PutRetentionConfigurationWithContext is the same as PutRetentionConfiguration with the addition of +// the ability to pass a context and additional request options. +// +// See PutRetentionConfiguration for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *ConfigService) PutRetentionConfigurationWithContext(ctx aws.Context, input *PutRetentionConfigurationInput, opts ...request.Option) (*PutRetentionConfigurationOutput, error) { + req, out := c.PutRetentionConfigurationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opStartConfigRulesEvaluation = "StartConfigRulesEvaluation" // StartConfigRulesEvaluationRequest generates a "aws/request.Request" representing the @@ -3817,7 +4087,7 @@ type AccountAggregationSource struct { // AccountIds is a required field AccountIds []*string `min:"1" type:"list" required:"true"` - // If true, aggreagate existing AWS Config regions and future regions. + // If true, aggregate existing AWS Config regions and future regions. AllAwsRegions *bool `type:"boolean"` // The source regions being aggregated. @@ -6094,6 +6364,61 @@ func (s DeletePendingAggregationRequestOutput) GoString() string { return s.String() } +type DeleteRetentionConfigurationInput struct { + _ struct{} `type:"structure"` + + // The name of the retention configuration to delete. + // + // RetentionConfigurationName is a required field + RetentionConfigurationName *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteRetentionConfigurationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteRetentionConfigurationInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteRetentionConfigurationInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteRetentionConfigurationInput"} + if s.RetentionConfigurationName == nil { + invalidParams.Add(request.NewErrParamRequired("RetentionConfigurationName")) + } + if s.RetentionConfigurationName != nil && len(*s.RetentionConfigurationName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("RetentionConfigurationName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetRetentionConfigurationName sets the RetentionConfigurationName field's value. +func (s *DeleteRetentionConfigurationInput) SetRetentionConfigurationName(v string) *DeleteRetentionConfigurationInput { + s.RetentionConfigurationName = &v + return s +} + +type DeleteRetentionConfigurationOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeleteRetentionConfigurationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteRetentionConfigurationOutput) GoString() string { + return s.String() +} + // The input for the DeliverConfigSnapshot action. type DeliverConfigSnapshotInput struct { _ struct{} `type:"structure"` @@ -6911,7 +7236,7 @@ func (s *DescribeConfigurationAggregatorSourcesStatusInput) SetUpdateStatus(v [] type DescribeConfigurationAggregatorSourcesStatusOutput struct { _ struct{} `type:"structure"` - // Retuns an AggregatedSourceStatus object. + // Returns an AggregatedSourceStatus object. AggregatedSourceStatusList []*AggregatedSourceStatus `type:"list"` // The nextToken string returned on a previous page that you use to get the @@ -7278,6 +7603,77 @@ func (s *DescribePendingAggregationRequestsOutput) SetPendingAggregationRequests return s } +type DescribeRetentionConfigurationsInput struct { + _ struct{} `type:"structure"` + + // The nextToken string returned on a previous page that you use to get the + // next page of results in a paginated response. + NextToken *string `type:"string"` + + // A list of names of retention configurations for which you want details. If + // you do not specify a name, AWS Config returns details for all the retention + // configurations for that account. + // + // Currently, AWS Config supports only one retention configuration per region + // in your account. + RetentionConfigurationNames []*string `type:"list"` +} + +// String returns the string representation +func (s DescribeRetentionConfigurationsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeRetentionConfigurationsInput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *DescribeRetentionConfigurationsInput) SetNextToken(v string) *DescribeRetentionConfigurationsInput { + s.NextToken = &v + return s +} + +// SetRetentionConfigurationNames sets the RetentionConfigurationNames field's value. +func (s *DescribeRetentionConfigurationsInput) SetRetentionConfigurationNames(v []*string) *DescribeRetentionConfigurationsInput { + s.RetentionConfigurationNames = v + return s +} + +type DescribeRetentionConfigurationsOutput struct { + _ struct{} `type:"structure"` + + // The nextToken string returned on a previous page that you use to get the + // next page of results in a paginated response. + NextToken *string `type:"string"` + + // Returns a retention configuration object. + RetentionConfigurations []*RetentionConfiguration `type:"list"` +} + +// String returns the string representation +func (s DescribeRetentionConfigurationsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeRetentionConfigurationsOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *DescribeRetentionConfigurationsOutput) SetNextToken(v string) *DescribeRetentionConfigurationsOutput { + s.NextToken = &v + return s +} + +// SetRetentionConfigurations sets the RetentionConfigurations field's value. +func (s *DescribeRetentionConfigurationsOutput) SetRetentionConfigurations(v []*RetentionConfiguration) *DescribeRetentionConfigurationsOutput { + s.RetentionConfigurations = v + return s +} + // Identifies an AWS resource and indicates whether it complies with the AWS // Config rule that it was evaluated against. type Evaluation struct { @@ -8520,7 +8916,7 @@ func (s *ListDiscoveredResourcesOutput) SetResourceIdentifiers(v []*ResourceIden type OrganizationAggregationSource struct { _ struct{} `type:"structure"` - // If true, aggreagate existing AWS Config regions and future regions. + // If true, aggregate existing AWS Config regions and future regions. AllAwsRegions *bool `type:"boolean"` // The source regions being aggregated. @@ -9059,6 +9455,72 @@ func (s *PutEvaluationsOutput) SetFailedEvaluations(v []*Evaluation) *PutEvaluat return s } +type PutRetentionConfigurationInput struct { + _ struct{} `type:"structure"` + + // Number of days AWS Config stores your historical information. + // + // Currently, only applicable to the configuration item history. + // + // RetentionPeriodInDays is a required field + RetentionPeriodInDays *int64 `min:"30" type:"integer" required:"true"` +} + +// String returns the string representation +func (s PutRetentionConfigurationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PutRetentionConfigurationInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *PutRetentionConfigurationInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "PutRetentionConfigurationInput"} + if s.RetentionPeriodInDays == nil { + invalidParams.Add(request.NewErrParamRequired("RetentionPeriodInDays")) + } + if s.RetentionPeriodInDays != nil && *s.RetentionPeriodInDays < 30 { + invalidParams.Add(request.NewErrParamMinValue("RetentionPeriodInDays", 30)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetRetentionPeriodInDays sets the RetentionPeriodInDays field's value. +func (s *PutRetentionConfigurationInput) SetRetentionPeriodInDays(v int64) *PutRetentionConfigurationInput { + s.RetentionPeriodInDays = &v + return s +} + +type PutRetentionConfigurationOutput struct { + _ struct{} `type:"structure"` + + // Returns a retention configuration object. + RetentionConfiguration *RetentionConfiguration `type:"structure"` +} + +// String returns the string representation +func (s PutRetentionConfigurationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PutRetentionConfigurationOutput) GoString() string { + return s.String() +} + +// SetRetentionConfiguration sets the RetentionConfiguration field's value. +func (s *PutRetentionConfigurationOutput) SetRetentionConfiguration(v *RetentionConfiguration) *PutRetentionConfigurationOutput { + s.RetentionConfiguration = v + return s +} + // Specifies the types of AWS resource for which AWS Config records configuration // changes. // @@ -9349,6 +9811,47 @@ func (s *ResourceKey) SetResourceType(v string) *ResourceKey { return s } +// An object with the name of the retention configuration and the retention +// period in days. The object stores the configuration for data retention in +// AWS Config. +type RetentionConfiguration struct { + _ struct{} `type:"structure"` + + // The name of the retention configuration object. + // + // Name is a required field + Name *string `min:"1" type:"string" required:"true"` + + // Number of days AWS Config stores your historical information. + // + // Currently, only applicable to the configuration item history. + // + // RetentionPeriodInDays is a required field + RetentionPeriodInDays *int64 `min:"30" type:"integer" required:"true"` +} + +// String returns the string representation +func (s RetentionConfiguration) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RetentionConfiguration) GoString() string { + return s.String() +} + +// SetName sets the Name field's value. +func (s *RetentionConfiguration) SetName(v string) *RetentionConfiguration { + s.Name = &v + return s +} + +// SetRetentionPeriodInDays sets the RetentionPeriodInDays field's value. +func (s *RetentionConfiguration) SetRetentionPeriodInDays(v int64) *RetentionConfiguration { + s.RetentionPeriodInDays = &v + return s +} + // Defines which resources trigger an evaluation for an AWS Config rule. The // scope can include one or more resource types, a combination of a tag key // and value, or a combination of one resource type and one resource ID. Specify diff --git a/vendor/github.com/aws/aws-sdk-go/service/configservice/configserviceiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/configservice/configserviceiface/interface.go index 623501a3a..9941693f8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/configservice/configserviceiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/configservice/configserviceiface/interface.go @@ -92,6 +92,10 @@ type ConfigServiceAPI interface { DeletePendingAggregationRequestWithContext(aws.Context, *configservice.DeletePendingAggregationRequestInput, ...request.Option) (*configservice.DeletePendingAggregationRequestOutput, error) DeletePendingAggregationRequestRequest(*configservice.DeletePendingAggregationRequestInput) (*request.Request, *configservice.DeletePendingAggregationRequestOutput) + DeleteRetentionConfiguration(*configservice.DeleteRetentionConfigurationInput) (*configservice.DeleteRetentionConfigurationOutput, error) + DeleteRetentionConfigurationWithContext(aws.Context, *configservice.DeleteRetentionConfigurationInput, ...request.Option) (*configservice.DeleteRetentionConfigurationOutput, error) + DeleteRetentionConfigurationRequest(*configservice.DeleteRetentionConfigurationInput) (*request.Request, *configservice.DeleteRetentionConfigurationOutput) + DeliverConfigSnapshot(*configservice.DeliverConfigSnapshotInput) (*configservice.DeliverConfigSnapshotOutput, error) DeliverConfigSnapshotWithContext(aws.Context, *configservice.DeliverConfigSnapshotInput, ...request.Option) (*configservice.DeliverConfigSnapshotOutput, error) DeliverConfigSnapshotRequest(*configservice.DeliverConfigSnapshotInput) (*request.Request, *configservice.DeliverConfigSnapshotOutput) @@ -148,6 +152,10 @@ type ConfigServiceAPI interface { DescribePendingAggregationRequestsWithContext(aws.Context, *configservice.DescribePendingAggregationRequestsInput, ...request.Option) (*configservice.DescribePendingAggregationRequestsOutput, error) DescribePendingAggregationRequestsRequest(*configservice.DescribePendingAggregationRequestsInput) (*request.Request, *configservice.DescribePendingAggregationRequestsOutput) + DescribeRetentionConfigurations(*configservice.DescribeRetentionConfigurationsInput) (*configservice.DescribeRetentionConfigurationsOutput, error) + DescribeRetentionConfigurationsWithContext(aws.Context, *configservice.DescribeRetentionConfigurationsInput, ...request.Option) (*configservice.DescribeRetentionConfigurationsOutput, error) + DescribeRetentionConfigurationsRequest(*configservice.DescribeRetentionConfigurationsInput) (*request.Request, *configservice.DescribeRetentionConfigurationsOutput) + GetAggregateComplianceDetailsByConfigRule(*configservice.GetAggregateComplianceDetailsByConfigRuleInput) (*configservice.GetAggregateComplianceDetailsByConfigRuleOutput, error) GetAggregateComplianceDetailsByConfigRuleWithContext(aws.Context, *configservice.GetAggregateComplianceDetailsByConfigRuleInput, ...request.Option) (*configservice.GetAggregateComplianceDetailsByConfigRuleOutput, error) GetAggregateComplianceDetailsByConfigRuleRequest(*configservice.GetAggregateComplianceDetailsByConfigRuleInput) (*request.Request, *configservice.GetAggregateComplianceDetailsByConfigRuleOutput) @@ -211,6 +219,10 @@ type ConfigServiceAPI interface { PutEvaluationsWithContext(aws.Context, *configservice.PutEvaluationsInput, ...request.Option) (*configservice.PutEvaluationsOutput, error) PutEvaluationsRequest(*configservice.PutEvaluationsInput) (*request.Request, *configservice.PutEvaluationsOutput) + PutRetentionConfiguration(*configservice.PutRetentionConfigurationInput) (*configservice.PutRetentionConfigurationOutput, error) + PutRetentionConfigurationWithContext(aws.Context, *configservice.PutRetentionConfigurationInput, ...request.Option) (*configservice.PutRetentionConfigurationOutput, error) + PutRetentionConfigurationRequest(*configservice.PutRetentionConfigurationInput) (*request.Request, *configservice.PutRetentionConfigurationOutput) + StartConfigRulesEvaluation(*configservice.StartConfigRulesEvaluationInput) (*configservice.StartConfigRulesEvaluationOutput, error) StartConfigRulesEvaluationWithContext(aws.Context, *configservice.StartConfigRulesEvaluationInput, ...request.Option) (*configservice.StartConfigRulesEvaluationOutput, error) StartConfigRulesEvaluationRequest(*configservice.StartConfigRulesEvaluationInput) (*request.Request, *configservice.StartConfigRulesEvaluationOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/configservice/errors.go b/vendor/github.com/aws/aws-sdk-go/service/configservice/errors.go index 433f766a8..b57a5cec4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/configservice/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/configservice/errors.go @@ -126,6 +126,13 @@ const ( // You have reached the limit of the number of delivery channels you can create. ErrCodeMaxNumberOfDeliveryChannelsExceededException = "MaxNumberOfDeliveryChannelsExceededException" + // ErrCodeMaxNumberOfRetentionConfigurationsExceededException for service response error code + // "MaxNumberOfRetentionConfigurationsExceededException". + // + // Failed to add the retention configuration because a retention configuration + // with that name already exists. + ErrCodeMaxNumberOfRetentionConfigurationsExceededException = "MaxNumberOfRetentionConfigurationsExceededException" + // ErrCodeNoAvailableConfigurationRecorderException for service response error code // "NoAvailableConfigurationRecorderException". // @@ -182,6 +189,12 @@ const ( // You have specified a delivery channel that does not exist. ErrCodeNoSuchDeliveryChannelException = "NoSuchDeliveryChannelException" + // ErrCodeNoSuchRetentionConfigurationException for service response error code + // "NoSuchRetentionConfigurationException". + // + // You have specified a retention configuration that does not exist. + ErrCodeNoSuchRetentionConfigurationException = "NoSuchRetentionConfigurationException" + // ErrCodeOrganizationAccessDeniedException for service response error code // "OrganizationAccessDeniedException". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/configservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/configservice/service.go index 3593126ad..2fdea9556 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/configservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/configservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "config" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "config" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Config Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ConfigService client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/connect/service.go b/vendor/github.com/aws/aws-sdk-go/service/connect/service.go index 975c8cb28..2f1093c9a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/connect/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/connect/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "connect" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "connect" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Connect" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Connect client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/costandusagereportservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/costandusagereportservice/service.go index d862e2be5..39e3cedfe 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/costandusagereportservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/costandusagereportservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "cur" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "cur" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Cost and Usage Report Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CostandUsageReportService client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/costexplorer/api.go b/vendor/github.com/aws/aws-sdk-go/service/costexplorer/api.go index bd2059a21..a9ff105a7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/costexplorer/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/costexplorer/api.go @@ -977,9 +977,9 @@ func (s *EC2Specification) SetOfferingClass(v string) *EC2Specification { // DataTransfer). The Expression for that looks like this: // // { "And": [ {"Or": [ {"Dimensions": { "Key": "INSTANCE_TYPE", "Values": [ -// "m4.x.large", "c4.large" ] }}, {"Tag": { "Key": "TagName", "Values": ["Value1"] -// } } ]}, {"Not": {"dimensions": { "Key": "USAGE_TYPE", "Values": ["DataTransfer"] -// }}} ] } +// "m4.x.large", "c4.large" ] }}, {"Tags": { "Key": "TagName", "Values": +// ["Value1"] } } ]}, {"Not": {"Dimensions": { "Key": "USAGE_TYPE", "Values": +// ["DataTransfer"] }}} ] } // // Because each Expression can have only one operator, the service returns an // error if more than one is specified. The following example shows an Expression @@ -2358,12 +2358,31 @@ func (s *RDSInstanceDetails) SetSizeFlexEligible(v bool) *RDSInstanceDetails { type ReservationAggregates struct { _ struct{} `type:"structure"` + // The monthly cost of your RI, amortized over the RI period. + AmortizedRecurringFee *string `type:"string"` + + // The upfront cost of your RI, amortized over the RI period. + AmortizedUpfrontFee *string `type:"string"` + + // How much you saved due to purchasing and utilizing RIs. This is calculated + // by subtracting the TotalAmortizedFee from the OnDemandCostOfRIHoursUsed. + NetRISavings *string `type:"string"` + + // How much your RIs would cost if charged On-Demand rates. + OnDemandCostOfRIHoursUsed *string `type:"string"` + // How many RI hours that you purchased. PurchasedHours *string `type:"string"` // The total number of RI hours that you used. TotalActualHours *string `type:"string"` + // The total cost of your RI, amortized over the RI period. + TotalAmortizedFee *string `type:"string"` + + // How much you could save if you use your entire reservation. + TotalPotentialRISavings *string `type:"string"` + // The number of RI hours that you didn't use. UnusedHours *string `type:"string"` @@ -2381,6 +2400,30 @@ func (s ReservationAggregates) GoString() string { return s.String() } +// SetAmortizedRecurringFee sets the AmortizedRecurringFee field's value. +func (s *ReservationAggregates) SetAmortizedRecurringFee(v string) *ReservationAggregates { + s.AmortizedRecurringFee = &v + return s +} + +// SetAmortizedUpfrontFee sets the AmortizedUpfrontFee field's value. +func (s *ReservationAggregates) SetAmortizedUpfrontFee(v string) *ReservationAggregates { + s.AmortizedUpfrontFee = &v + return s +} + +// SetNetRISavings sets the NetRISavings field's value. +func (s *ReservationAggregates) SetNetRISavings(v string) *ReservationAggregates { + s.NetRISavings = &v + return s +} + +// SetOnDemandCostOfRIHoursUsed sets the OnDemandCostOfRIHoursUsed field's value. +func (s *ReservationAggregates) SetOnDemandCostOfRIHoursUsed(v string) *ReservationAggregates { + s.OnDemandCostOfRIHoursUsed = &v + return s +} + // SetPurchasedHours sets the PurchasedHours field's value. func (s *ReservationAggregates) SetPurchasedHours(v string) *ReservationAggregates { s.PurchasedHours = &v @@ -2393,6 +2436,18 @@ func (s *ReservationAggregates) SetTotalActualHours(v string) *ReservationAggreg return s } +// SetTotalAmortizedFee sets the TotalAmortizedFee field's value. +func (s *ReservationAggregates) SetTotalAmortizedFee(v string) *ReservationAggregates { + s.TotalAmortizedFee = &v + return s +} + +// SetTotalPotentialRISavings sets the TotalPotentialRISavings field's value. +func (s *ReservationAggregates) SetTotalPotentialRISavings(v string) *ReservationAggregates { + s.TotalPotentialRISavings = &v + return s +} + // SetUnusedHours sets the UnusedHours field's value. func (s *ReservationAggregates) SetUnusedHours(v string) *ReservationAggregates { s.UnusedHours = &v diff --git a/vendor/github.com/aws/aws-sdk-go/service/costexplorer/service.go b/vendor/github.com/aws/aws-sdk-go/service/costexplorer/service.go index e9e9e812c..f8fa582b3 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/costexplorer/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/costexplorer/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "ce" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "ce" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Cost Explorer" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the CostExplorer client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/databasemigrationservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/databasemigrationservice/service.go index bf5b47613..8ee775e03 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/databasemigrationservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/databasemigrationservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "dms" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "dms" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Database Migration Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DatabaseMigrationService client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/datapipeline/service.go b/vendor/github.com/aws/aws-sdk-go/service/datapipeline/service.go index 1c7122883..ebd5c29b2 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/datapipeline/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/datapipeline/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "datapipeline" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "datapipeline" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Data Pipeline" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DataPipeline client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/dax/service.go b/vendor/github.com/aws/aws-sdk-go/service/dax/service.go index a80ed1441..545ea0312 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dax/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dax/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "dax" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "dax" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "DAX" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DAX client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/devicefarm/api.go b/vendor/github.com/aws/aws-sdk-go/service/devicefarm/api.go index f84b3bd9f..f30075d6d 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/devicefarm/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/devicefarm/api.go @@ -7105,6 +7105,9 @@ type CreateRemoteAccessSessionConfiguration struct { // The billing method for the remote access session. BillingMethod *string `locationName:"billingMethod" type:"string" enum:"BillingMethod"` + + // An array of Amazon Resource Names (ARNs) included in the VPC endpoint configuration. + VpceConfigurationArns []*string `locationName:"vpceConfigurationArns" type:"list"` } // String returns the string representation @@ -7123,6 +7126,12 @@ func (s *CreateRemoteAccessSessionConfiguration) SetBillingMethod(v string) *Cre return s } +// SetVpceConfigurationArns sets the VpceConfigurationArns field's value. +func (s *CreateRemoteAccessSessionConfiguration) SetVpceConfigurationArns(v []*string) *CreateRemoteAccessSessionConfiguration { + s.VpceConfigurationArns = v + return s +} + // Creates and submits a request to start a remote access session. type CreateRemoteAccessSessionInput struct { _ struct{} `type:"structure"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/devicefarm/service.go b/vendor/github.com/aws/aws-sdk-go/service/devicefarm/service.go index 76cf7ca61..0f2354a4e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/devicefarm/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/devicefarm/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "devicefarm" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "devicefarm" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Device Farm" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DeviceFarm client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/directconnect/api.go b/vendor/github.com/aws/aws-sdk-go/service/directconnect/api.go index 6ad8c8e30..46b46b662 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directconnect/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directconnect/api.go @@ -75,11 +75,11 @@ func (c *DirectConnect) AllocateConnectionOnInterconnectRequest(input *AllocateC // API operation AllocateConnectionOnInterconnect for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -165,11 +165,11 @@ func (c *DirectConnect) AllocateHostedConnectionRequest(input *AllocateHostedCon // API operation AllocateHostedConnection for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -254,11 +254,11 @@ func (c *DirectConnect) AllocatePrivateVirtualInterfaceRequest(input *AllocatePr // API operation AllocatePrivateVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -350,11 +350,11 @@ func (c *DirectConnect) AllocatePublicVirtualInterfaceRequest(input *AllocatePub // API operation AllocatePublicVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -450,11 +450,11 @@ func (c *DirectConnect) AssociateConnectionWithLagRequest(input *AssociateConnec // API operation AssociateConnectionWithLag for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -540,11 +540,11 @@ func (c *DirectConnect) AssociateHostedConnectionRequest(input *AssociateHostedC // API operation AssociateHostedConnection for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -638,11 +638,11 @@ func (c *DirectConnect) AssociateVirtualInterfaceRequest(input *AssociateVirtual // API operation AssociateVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -726,11 +726,11 @@ func (c *DirectConnect) ConfirmConnectionRequest(input *ConfirmConnectionInput) // API operation ConfirmConnection for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -814,11 +814,11 @@ func (c *DirectConnect) ConfirmPrivateVirtualInterfaceRequest(input *ConfirmPriv // API operation ConfirmPrivateVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -901,11 +901,11 @@ func (c *DirectConnect) ConfirmPublicVirtualInterfaceRequest(input *ConfirmPubli // API operation ConfirmPublicVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -997,11 +997,11 @@ func (c *DirectConnect) CreateBGPPeerRequest(input *CreateBGPPeerInput) (req *re // API operation CreateBGPPeer for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1098,11 +1098,11 @@ func (c *DirectConnect) CreateConnectionRequest(input *CreateConnectionInput) (r // API operation CreateConnection for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1189,11 +1189,11 @@ func (c *DirectConnect) CreateDirectConnectGatewayRequest(input *CreateDirectCon // API operation CreateDirectConnectGateway for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1275,11 +1275,11 @@ func (c *DirectConnect) CreateDirectConnectGatewayAssociationRequest(input *Crea // API operation CreateDirectConnectGatewayAssociation for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1382,11 +1382,11 @@ func (c *DirectConnect) CreateInterconnectRequest(input *CreateInterconnectInput // API operation CreateInterconnect for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1490,11 +1490,11 @@ func (c *DirectConnect) CreateLagRequest(input *CreateLagInput) (req *request.Re // API operation CreateLag for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1576,11 +1576,11 @@ func (c *DirectConnect) CreatePrivateVirtualInterfaceRequest(input *CreatePrivat // API operation CreatePrivateVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1667,11 +1667,11 @@ func (c *DirectConnect) CreatePublicVirtualInterfaceRequest(input *CreatePublicV // API operation CreatePublicVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1753,11 +1753,11 @@ func (c *DirectConnect) DeleteBGPPeerRequest(input *DeleteBGPPeerInput) (req *re // API operation DeleteBGPPeer for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1842,11 +1842,11 @@ func (c *DirectConnect) DeleteConnectionRequest(input *DeleteConnectionInput) (r // API operation DeleteConnection for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -1928,11 +1928,11 @@ func (c *DirectConnect) DeleteDirectConnectGatewayRequest(input *DeleteDirectCon // API operation DeleteDirectConnectGateway for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2013,11 +2013,11 @@ func (c *DirectConnect) DeleteDirectConnectGatewayAssociationRequest(input *Dele // API operation DeleteDirectConnectGatewayAssociation for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2099,11 +2099,11 @@ func (c *DirectConnect) DeleteInterconnectRequest(input *DeleteInterconnectInput // API operation DeleteInterconnect for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2184,11 +2184,11 @@ func (c *DirectConnect) DeleteLagRequest(input *DeleteLagInput) (req *request.Re // API operation DeleteLag for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2268,11 +2268,11 @@ func (c *DirectConnect) DeleteVirtualInterfaceRequest(input *DeleteVirtualInterf // API operation DeleteVirtualInterface for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2363,11 +2363,11 @@ func (c *DirectConnect) DescribeConnectionLoaRequest(input *DescribeConnectionLo // API operation DescribeConnectionLoa for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2449,11 +2449,11 @@ func (c *DirectConnect) DescribeConnectionsRequest(input *DescribeConnectionsInp // API operation DescribeConnections for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2540,11 +2540,11 @@ func (c *DirectConnect) DescribeConnectionsOnInterconnectRequest(input *Describe // API operation DescribeConnectionsOnInterconnect for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2630,11 +2630,11 @@ func (c *DirectConnect) DescribeDirectConnectGatewayAssociationsRequest(input *D // API operation DescribeDirectConnectGatewayAssociations for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2720,11 +2720,11 @@ func (c *DirectConnect) DescribeDirectConnectGatewayAttachmentsRequest(input *De // API operation DescribeDirectConnectGatewayAttachments for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2808,11 +2808,11 @@ func (c *DirectConnect) DescribeDirectConnectGatewaysRequest(input *DescribeDire // API operation DescribeDirectConnectGateways for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2895,11 +2895,11 @@ func (c *DirectConnect) DescribeHostedConnectionsRequest(input *DescribeHostedCo // API operation DescribeHostedConnections for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -2990,11 +2990,11 @@ func (c *DirectConnect) DescribeInterconnectLoaRequest(input *DescribeInterconne // API operation DescribeInterconnectLoa for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3076,11 +3076,11 @@ func (c *DirectConnect) DescribeInterconnectsRequest(input *DescribeInterconnect // API operation DescribeInterconnects for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3162,11 +3162,11 @@ func (c *DirectConnect) DescribeLagsRequest(input *DescribeLagsInput) (req *requ // API operation DescribeLags for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3253,11 +3253,11 @@ func (c *DirectConnect) DescribeLoaRequest(input *DescribeLoaInput) (req *reques // API operation DescribeLoa for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3339,11 +3339,11 @@ func (c *DirectConnect) DescribeLocationsRequest(input *DescribeLocationsInput) // API operation DescribeLocations for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3423,11 +3423,11 @@ func (c *DirectConnect) DescribeTagsRequest(input *DescribeTagsInput) (req *requ // API operation DescribeTags for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3513,11 +3513,11 @@ func (c *DirectConnect) DescribeVirtualGatewaysRequest(input *DescribeVirtualGat // API operation DescribeVirtualGateways for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3604,11 +3604,11 @@ func (c *DirectConnect) DescribeVirtualInterfacesRequest(input *DescribeVirtualI // API operation DescribeVirtualInterfaces for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3698,11 +3698,11 @@ func (c *DirectConnect) DisassociateConnectionFromLagRequest(input *Disassociate // API operation DisassociateConnectionFromLag for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3794,11 +3794,11 @@ func (c *DirectConnect) TagResourceRequest(input *TagResourceInput) (req *reques // You have reached the limit on the number of tags that can be assigned to // a Direct Connect resource. // -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3878,11 +3878,11 @@ func (c *DirectConnect) UntagResourceRequest(input *UntagResourceInput) (req *re // API operation UntagResource for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // @@ -3976,11 +3976,11 @@ func (c *DirectConnect) UpdateLagRequest(input *UpdateLagInput) (req *request.Re // API operation UpdateLag for usage and error information. // // Returned Error Codes: -// * ErrCodeServerException "ServerException" +// * ErrCodeServerException "DirectConnectServerException" // A server-side error occurred during the API call. The error message will // contain additional details about the cause. // -// * ErrCodeClientException "ClientException" +// * ErrCodeClientException "DirectConnectClientException" // The API was called with invalid parameters. The error message will contain // additional details about the cause. // diff --git a/vendor/github.com/aws/aws-sdk-go/service/directconnect/errors.go b/vendor/github.com/aws/aws-sdk-go/service/directconnect/errors.go index cbf9bf2f5..454cfe5ae 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directconnect/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directconnect/errors.go @@ -5,11 +5,11 @@ package directconnect const ( // ErrCodeClientException for service response error code - // "ClientException". + // "DirectConnectClientException". // // The API was called with invalid parameters. The error message will contain // additional details about the cause. - ErrCodeClientException = "ClientException" + ErrCodeClientException = "DirectConnectClientException" // ErrCodeDuplicateTagKeysException for service response error code // "DuplicateTagKeysException". @@ -18,11 +18,11 @@ const ( ErrCodeDuplicateTagKeysException = "DuplicateTagKeysException" // ErrCodeServerException for service response error code - // "ServerException". + // "DirectConnectServerException". // // A server-side error occurred during the API call. The error message will // contain additional details about the cause. - ErrCodeServerException = "ServerException" + ErrCodeServerException = "DirectConnectServerException" // ErrCodeTooManyTagsException for service response error code // "TooManyTagsException". diff --git a/vendor/github.com/aws/aws-sdk-go/service/directconnect/service.go b/vendor/github.com/aws/aws-sdk-go/service/directconnect/service.go index 1798ae11a..bb182821c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directconnect/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directconnect/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "directconnect" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "directconnect" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Direct Connect" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DirectConnect client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/api.go b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/api.go index fc3674146..55a1a1200 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/api.go @@ -3239,6 +3239,105 @@ func (c *DirectoryService) RemoveTagsFromResourceWithContext(ctx aws.Context, in return out, req.Send() } +const opResetUserPassword = "ResetUserPassword" + +// ResetUserPasswordRequest generates a "aws/request.Request" representing the +// client's request for the ResetUserPassword operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ResetUserPassword for more information on using the ResetUserPassword +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ResetUserPasswordRequest method. +// req, resp := client.ResetUserPasswordRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/ds-2015-04-16/ResetUserPassword +func (c *DirectoryService) ResetUserPasswordRequest(input *ResetUserPasswordInput) (req *request.Request, output *ResetUserPasswordOutput) { + op := &request.Operation{ + Name: opResetUserPassword, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ResetUserPasswordInput{} + } + + output = &ResetUserPasswordOutput{} + req = c.newRequest(op, input, output) + return +} + +// ResetUserPassword API operation for AWS Directory Service. +// +// Resets the password for any user in your AWS Managed Microsoft AD or Simple +// AD directory. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Directory Service's +// API operation ResetUserPassword for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDirectoryUnavailableException "DirectoryUnavailableException" +// The specified directory is unavailable or could not be found. +// +// * ErrCodeUserDoesNotExistException "UserDoesNotExistException" +// The user provided a username that does not exist in your directory. +// +// * ErrCodeInvalidPasswordException "InvalidPasswordException" +// The new password provided by the user does not meet the password complexity +// requirements defined in your directory. +// +// * ErrCodeUnsupportedOperationException "UnsupportedOperationException" +// The operation is not supported. +// +// * ErrCodeEntityDoesNotExistException "EntityDoesNotExistException" +// The specified entity could not be found. +// +// * ErrCodeClientException "ClientException" +// A client exception has occurred. +// +// * ErrCodeServiceException "ServiceException" +// An exception has occurred in AWS Directory Service. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/ds-2015-04-16/ResetUserPassword +func (c *DirectoryService) ResetUserPassword(input *ResetUserPasswordInput) (*ResetUserPasswordOutput, error) { + req, out := c.ResetUserPasswordRequest(input) + return out, req.Send() +} + +// ResetUserPasswordWithContext is the same as ResetUserPassword with the addition of +// the ability to pass a context and additional request options. +// +// See ResetUserPassword for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *DirectoryService) ResetUserPasswordWithContext(ctx aws.Context, input *ResetUserPasswordInput, opts ...request.Option) (*ResetUserPasswordOutput, error) { + req, out := c.ResetUserPasswordRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opRestoreFromSnapshot = "RestoreFromSnapshot" // RestoreFromSnapshotRequest generates a "aws/request.Request" representing the @@ -7812,6 +7911,93 @@ func (s RemoveTagsFromResourceOutput) GoString() string { return s.String() } +type ResetUserPasswordInput struct { + _ struct{} `type:"structure"` + + // Identifier of the AWS Managed Microsoft AD or Simple AD directory in which + // the user resides. + // + // DirectoryId is a required field + DirectoryId *string `type:"string" required:"true"` + + // The new password that will be reset. + // + // NewPassword is a required field + NewPassword *string `min:"1" type:"string" required:"true"` + + // The username of the user whose password will be reset. + // + // UserName is a required field + UserName *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s ResetUserPasswordInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResetUserPasswordInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ResetUserPasswordInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ResetUserPasswordInput"} + if s.DirectoryId == nil { + invalidParams.Add(request.NewErrParamRequired("DirectoryId")) + } + if s.NewPassword == nil { + invalidParams.Add(request.NewErrParamRequired("NewPassword")) + } + if s.NewPassword != nil && len(*s.NewPassword) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NewPassword", 1)) + } + if s.UserName == nil { + invalidParams.Add(request.NewErrParamRequired("UserName")) + } + if s.UserName != nil && len(*s.UserName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("UserName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDirectoryId sets the DirectoryId field's value. +func (s *ResetUserPasswordInput) SetDirectoryId(v string) *ResetUserPasswordInput { + s.DirectoryId = &v + return s +} + +// SetNewPassword sets the NewPassword field's value. +func (s *ResetUserPasswordInput) SetNewPassword(v string) *ResetUserPasswordInput { + s.NewPassword = &v + return s +} + +// SetUserName sets the UserName field's value. +func (s *ResetUserPasswordInput) SetUserName(v string) *ResetUserPasswordInput { + s.UserName = &v + return s +} + +type ResetUserPasswordOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s ResetUserPasswordOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResetUserPasswordOutput) GoString() string { + return s.String() +} + // An object representing the inputs for the RestoreFromSnapshot operation. type RestoreFromSnapshotInput struct { _ struct{} `type:"structure"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/directoryserviceiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/directoryserviceiface/interface.go index bf6fc35bb..662594130 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/directoryserviceiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/directoryserviceiface/interface.go @@ -199,6 +199,10 @@ type DirectoryServiceAPI interface { RemoveTagsFromResourceWithContext(aws.Context, *directoryservice.RemoveTagsFromResourceInput, ...request.Option) (*directoryservice.RemoveTagsFromResourceOutput, error) RemoveTagsFromResourceRequest(*directoryservice.RemoveTagsFromResourceInput) (*request.Request, *directoryservice.RemoveTagsFromResourceOutput) + ResetUserPassword(*directoryservice.ResetUserPasswordInput) (*directoryservice.ResetUserPasswordOutput, error) + ResetUserPasswordWithContext(aws.Context, *directoryservice.ResetUserPasswordInput, ...request.Option) (*directoryservice.ResetUserPasswordOutput, error) + ResetUserPasswordRequest(*directoryservice.ResetUserPasswordInput) (*request.Request, *directoryservice.ResetUserPasswordOutput) + RestoreFromSnapshot(*directoryservice.RestoreFromSnapshotInput) (*directoryservice.RestoreFromSnapshotOutput, error) RestoreFromSnapshotWithContext(aws.Context, *directoryservice.RestoreFromSnapshotInput, ...request.Option) (*directoryservice.RestoreFromSnapshotOutput, error) RestoreFromSnapshotRequest(*directoryservice.RestoreFromSnapshotInput) (*request.Request, *directoryservice.RestoreFromSnapshotOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/errors.go b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/errors.go index f7560305b..85d831364 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/errors.go @@ -67,6 +67,13 @@ const ( // One or more parameters are not valid. ErrCodeInvalidParameterException = "InvalidParameterException" + // ErrCodeInvalidPasswordException for service response error code + // "InvalidPasswordException". + // + // The new password provided by the user does not meet the password complexity + // requirements defined in your directory. + ErrCodeInvalidPasswordException = "InvalidPasswordException" + // ErrCodeIpRouteLimitExceededException for service response error code // "IpRouteLimitExceededException". // @@ -99,4 +106,10 @@ const ( // // The operation is not supported. ErrCodeUnsupportedOperationException = "UnsupportedOperationException" + + // ErrCodeUserDoesNotExistException for service response error code + // "UserDoesNotExistException". + // + // The user provided a username that does not exist in your directory. + ErrCodeUserDoesNotExistException = "UserDoesNotExistException" ) diff --git a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/service.go index 59299ae5d..0743c9632 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/directoryservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/directoryservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "ds" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "ds" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Directory Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DirectoryService client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/api.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/api.go index bf9d7a4e1..5dc8c9021 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/api.go @@ -472,8 +472,7 @@ func (c *DynamoDB) CreateBackupRequest(input *CreateBackupInput) (req *request.R // table. The backups is either being created, deleted or restored to a table. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -558,20 +557,17 @@ func (c *DynamoDB) CreateGlobalTableRequest(input *CreateGlobalTableInput) (req // relationship between two or more DynamoDB tables with the same table name // in the provided regions. // -// Tables can only be added as the replicas of a global table group under the -// following conditions: +// If you want to add a new replica table to a global table, each of the following +// conditions must be true: // -// * The tables must have the same name. +// * The table must have the same primary key as all of the other replicas. // -// * The tables must contain no items. +// * The table must have the same name as all of the other replicas. // -// * The tables must have the same hash key and sort key (if present). -// -// * The tables must have DynamoDB Streams enabled (NEW_AND_OLD_IMAGES). -// -// -// * The tables must have same provisioned and maximum write capacity units. +// * The table must have DynamoDB Streams enabled, with the stream containing +// both the new and the old images of the item. // +// * None of the replica tables in the global table can contain any data. // // If global secondary indexes are specified, then the following conditions // must also be met: @@ -581,8 +577,15 @@ func (c *DynamoDB) CreateGlobalTableRequest(input *CreateGlobalTableInput) (req // * The global secondary indexes must have the same hash key and sort key // (if present). // -// * The global secondary indexes must have the same provisioned and maximum -// write capacity units. +// Write capacity settings should be set consistently across your replica tables +// and secondary indexes. DynamoDB strongly recommends enabling auto scaling +// to manage the write capacity settings for all of your global tables replicas +// and indexes. +// +// If you prefer to manage write capacity settings manually, you should provision +// equal replicated write capacity units to your replica tables. You should +// also provision equal replicated write capacity units to matching secondary +// indexes across your global table. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -593,8 +596,7 @@ func (c *DynamoDB) CreateGlobalTableRequest(input *CreateGlobalTableInput) (req // // Returned Error Codes: // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -712,8 +714,7 @@ func (c *DynamoDB) CreateTableRequest(input *CreateTableInput) (req *request.Req // in the CREATING state. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -814,8 +815,7 @@ func (c *DynamoDB) DeleteBackupRequest(input *DeleteBackupInput) (req *request.R // table. The backups is either being created, deleted or restored to a table. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -1044,8 +1044,7 @@ func (c *DynamoDB) DeleteTableRequest(input *DeleteTableInput) (req *request.Req // might not be specified correctly, or its status might not be ACTIVE. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -2642,8 +2641,7 @@ func (c *DynamoDB) RestoreTableFromBackupRequest(input *RestoreTableFromBackupIn // table. The backups is either being created, deleted or restored to a table. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -2782,8 +2780,7 @@ func (c *DynamoDB) RestoreTableToPointInTimeRequest(input *RestoreTableToPointIn // A target table with the specified name is either being created or deleted. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -3064,8 +3061,7 @@ func (c *DynamoDB) TagResourceRequest(input *TagResourceInput) (req *request.Req // // Returned Error Codes: // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -3172,8 +3168,7 @@ func (c *DynamoDB) UntagResourceRequest(input *UntagResourceInput) (req *request // // Returned Error Codes: // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -3490,8 +3485,7 @@ func (c *DynamoDB) UpdateGlobalTableSettingsRequest(input *UpdateGlobalTableSett // The operation tried to access a nonexistent index. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -3719,8 +3713,7 @@ func (c *DynamoDB) UpdateTableRequest(input *UpdateTableInput) (req *request.Req // might not be specified correctly, or its status might not be ACTIVE. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -3848,8 +3841,7 @@ func (c *DynamoDB) UpdateTimeToLiveRequest(input *UpdateTimeToLiveInput) (req *r // might not be specified correctly, or its status might not be ACTIVE. // // * ErrCodeLimitExceededException "LimitExceededException" -// Up to 50 CreateBackup operations are allowed per second, per account. There -// is no limit to the number of daily on-demand backups that can be taken. +// There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, @@ -6312,7 +6304,8 @@ func (s *DescribeContinuousBackupsInput) SetTableName(v string) *DescribeContinu type DescribeContinuousBackupsOutput struct { _ struct{} `type:"structure"` - // ContinuousBackupsDescription can be one of the following : ENABLED, DISABLED. + // Represents the continuous backups and point in time recovery settings on + // the table. ContinuousBackupsDescription *ContinuousBackupsDescription `type:"structure"` } @@ -7870,7 +7863,10 @@ func (s *KeysAndAttributes) SetProjectionExpression(v string) *KeysAndAttributes type ListBackupsInput struct { _ struct{} `type:"structure"` - // LastEvaluatedBackupARN returned by the previous ListBackups call. + // LastEvaluatedBackupArn is the ARN of the backup last evaluated when the current + // page of results was returned, inclusive of the current page of results. This + // value may be specified as the ExclusiveStartBackupArn of a new ListBackups + // operation in order to fetch the next page of results. ExclusiveStartBackupArn *string `min:"37" type:"string"` // Maximum number of backups to return at once. @@ -7952,7 +7948,17 @@ type ListBackupsOutput struct { // List of BackupSummary objects. BackupSummaries []*BackupSummary `type:"list"` - // Last evaluated BackupARN. + // The ARN of the backup last evaluated when the current page of results was + // returned, inclusive of the current page of results. This value may be specified + // as the ExclusiveStartBackupArn of a new ListBackups operation in order to + // fetch the next page of results. + // + // If LastEvaluatedBackupArn is empty, then the last page of results has been + // processed and there are no more results to be retrieved. + // + // If LastEvaluatedBackupArn is not empty, this may or may not indicate there + // is more data to be returned. All results are guaranteed to have been returned + // if and only if no value for LastEvaluatedBackupArn is returned. LastEvaluatedBackupArn *string `min:"37" type:"string"` } @@ -10291,6 +10297,16 @@ func (s *RestoreTableToPointInTimeOutput) SetTableDescription(v *TableDescriptio type SSEDescription struct { _ struct{} `type:"structure"` + // The KMS master key ARN used for the KMS encryption. + KMSMasterKeyArn *string `type:"string"` + + // Server-side encryption type: + // + // * AES256 - Server-side encryption which uses the AES256 algorithm. + // + // * KMS - Server-side encryption which uses AWS Key Management Service. + SSEType *string `type:"string" enum:"SSEType"` + // The current state of server-side encryption: // // * ENABLING - Server-side encryption is being enabled. @@ -10313,6 +10329,18 @@ func (s SSEDescription) GoString() string { return s.String() } +// SetKMSMasterKeyArn sets the KMSMasterKeyArn field's value. +func (s *SSEDescription) SetKMSMasterKeyArn(v string) *SSEDescription { + s.KMSMasterKeyArn = &v + return s +} + +// SetSSEType sets the SSEType field's value. +func (s *SSEDescription) SetSSEType(v string) *SSEDescription { + s.SSEType = &v + return s +} + // SetStatus sets the Status field's value. func (s *SSEDescription) SetStatus(v string) *SSEDescription { s.Status = &v @@ -12930,6 +12958,14 @@ const ( SSEStatusDisabled = "DISABLED" ) +const ( + // SSETypeAes256 is a SSEType enum value + SSETypeAes256 = "AES256" + + // SSETypeKms is a SSEType enum value + SSETypeKms = "KMS" +) + const ( // ScalarAttributeTypeS is a ScalarAttributeType enum value ScalarAttributeTypeS = "S" diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode.go index e02497568..41b67b3ae 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode.go @@ -1,6 +1,7 @@ package dynamodbattribute import ( + "encoding/base64" "fmt" "reflect" "strconv" @@ -538,6 +539,16 @@ func (d *Decoder) decodeString(s *string, v reflect.Value, fieldTag tag) error { switch v.Kind() { case reflect.String: v.SetString(*s) + case reflect.Slice: + // To maintain backwards compatibility with the ConvertFrom family of methods + // which converted []byte into base64-encoded strings if the input was typed + if v.Type() == byteSliceType { + decoded, err := base64.StdEncoding.DecodeString(*s) + if err != nil { + return &UnmarshalError{Err: err, Value: "string", Type: v.Type()} + } + v.SetBytes(decoded) + } case reflect.Interface: // Ensure type aliasing is handled properly v.Set(reflect.ValueOf(*s).Convert(v.Type())) diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode_test.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode_test.go index 1313064af..0f73a68b0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/decode_test.go @@ -234,6 +234,30 @@ func TestUnmarshalListError(t *testing.T) { } } +func TestUnmarshalConvertToData(t *testing.T) { + type T struct { + Int int + Str string + ByteSlice []byte + StrSlice []string + } + + exp := T{ + Int: 42, + Str: "foo", + ByteSlice: []byte{42, 97, 83}, + StrSlice: []string{"cat", "dog"}, + } + av, err := ConvertToMap(exp) + if err != nil { + t.Fatalf("expect no error, got %v", err) + } + + var act T + err = UnmarshalMap(av, &act) + assertConvertTest(t, 0, act, exp, err, nil) +} + func TestUnmarshalMapShared(t *testing.T) { for i, c := range sharedMapTestCases { err := UnmarshalMap(c.in, c.actual) diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/doc.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/doc.go index 7a51ac076..b83a29c95 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/doc.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/dynamodbattribute/doc.go @@ -81,7 +81,7 @@ // The ConvertTo, ConvertToList, ConvertToMap, ConvertFrom, ConvertFromMap // and ConvertFromList methods have been deprecated. The Marshal and Unmarshal // functions should be used instead. The ConvertTo|From marshallers do not -// support BinarySet, NumberSet, nor StringSets, and will incorrect marshal +// support BinarySet, NumberSet, nor StringSets, and will incorrectly marshal // binary data fields in structs as base64 strings. // // The Marshal and Unmarshal functions correct this behavior, and removes @@ -91,5 +91,11 @@ // replaced with have been replaced with dynamodbattribute.Marshaler and // dynamodbattribute.Unmarshaler interfaces. // +// The Unmarshal functions are backwards compatible with data marshalled by +// ConvertTo*, but the reverse is not true: objects marshalled using Marshal +// are not necessarily usable by ConvertFrom*. This backward compatibility is +// intended to assist with incremental upgrading of data following a switch +// away from the Convert* family of functions. +// // `time.Time` is marshaled as RFC3339 format. package dynamodbattribute diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/errors.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/errors.go index 5f6016521..4abbbe663 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/errors.go @@ -70,8 +70,7 @@ const ( // ErrCodeLimitExceededException for service response error code // "LimitExceededException". // - // Up to 50 CreateBackup operations are allowed per second, per account. There - // is no limit to the number of daily on-demand backups that can be taken. + // There is no limit to the number of daily on-demand backups that can be taken. // // Up to 10 simultaneous table operations are allowed per account. These operations // include CreateTable, UpdateTable, DeleteTable,UpdateTimeToLive, RestoreTableFromBackup, diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression.go index 75a33f06b..5efe0c9c9 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression.go @@ -488,7 +488,10 @@ func (e Expression) returnExpression(expressionType expressionType) *string { if e.expressionMap == nil { return nil } - return aws.String(e.expressionMap[expressionType]) + if s, exists := e.expressionMap[expressionType]; exists { + return &s + } + return nil } // exprNode are the generic nodes that represents both Operands and diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression_test.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression_test.go index 29af86206..1e20b7288 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/expression/expression_test.go @@ -978,6 +978,39 @@ func TestBuildExpressionString(t *testing.T) { } } +func TestReturnExpression(t *testing.T) { + cases := []struct { + name string + input Expression + expected *string + }{ + { + name: "projection exists", + input: Expression{ + expressionMap: map[expressionType]string{ + projection: "#0, #1, #2", + }, + }, + expected: aws.String("#0, #1, #2"), + }, + { + name: "projection not exists", + input: Expression{ + expressionMap: map[expressionType]string{}, + }, + expected: nil, + }, + } + for _, c := range cases { + t.Run(c.name, func(t *testing.T) { + actual := c.input.returnExpression(projection) + if e, a := c.expected, actual; !reflect.DeepEqual(a, e) { + t.Errorf("expect %v, got %v", e, a) + } + }) + } +} + func TestAliasValue(t *testing.T) { cases := []struct { name string diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/service.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/service.go index 80dcd19fd..346f52ce4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodb/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodb/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "dynamodb" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "dynamodb" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "DynamoDB" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DynamoDB client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/dynamodbstreams/service.go b/vendor/github.com/aws/aws-sdk-go/service/dynamodbstreams/service.go index b315870d5..700c52b5b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/dynamodbstreams/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/dynamodbstreams/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "streams.dynamodb" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "streams.dynamodb" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "DynamoDB Streams" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the DynamoDBStreams client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/ec2/api.go b/vendor/github.com/aws/aws-sdk-go/service/ec2/api.go index 99d12a66e..b48e40e20 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ec2/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ec2/api.go @@ -2268,11 +2268,7 @@ func (c *EC2) CancelSpotInstanceRequestsRequest(input *CancelSpotInstanceRequest // CancelSpotInstanceRequests API operation for Amazon Elastic Compute Cloud. // -// Cancels one or more Spot Instance requests. Spot Instances are instances -// that Amazon EC2 starts on your behalf when the maximum price that you specify -// exceeds the current Spot price. For more information, see Spot Instance Requests -// (http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/spot-requests.html) in -// the Amazon EC2 User Guide for Linux Instances. +// Cancels one or more Spot Instance requests. // // Canceling a Spot Instance request does not terminate running Spot Instances // associated with the request. @@ -4179,8 +4175,8 @@ func (c *EC2) CreateNetworkInterfacePermissionRequest(input *CreateNetworkInterf // CreateNetworkInterfacePermission API operation for Amazon Elastic Compute Cloud. // -// Grants an AWS authorized partner account permission to attach the specified -// network interface to an instance in their account. +// Grants an AWS-authorized account permission to attach the specified network +// interface to an instance in their account. // // You can grant permission to a single AWS account only, and only one account // at a time. @@ -13675,11 +13671,7 @@ func (c *EC2) DescribeSpotInstanceRequestsRequest(input *DescribeSpotInstanceReq // DescribeSpotInstanceRequests API operation for Amazon Elastic Compute Cloud. // -// Describes the Spot Instance requests that belong to your account. Spot Instances -// are instances that Amazon EC2 launches when the Spot price that you specify -// exceeds the current Spot price. For more information, see Spot Instance Requests -// (http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/spot-requests.html) in -// the Amazon EC2 User Guide for Linux Instances. +// Describes the specified Spot Instance requests. // // You can use DescribeSpotInstanceRequests to find a running Spot Instance // by examining the response. If the status of the Spot Instance is fulfilled, @@ -21367,9 +21359,9 @@ func (c *EC2) RequestSpotInstancesRequest(input *RequestSpotInstancesInput) (req // RequestSpotInstances API operation for Amazon Elastic Compute Cloud. // -// Creates a Spot Instance request. Spot Instances are instances that Amazon -// EC2 launches when the maximum price that you specify exceeds the current -// Spot price. For more information, see Spot Instance Requests (http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/spot-requests.html) +// Creates a Spot Instance request. +// +// For more information, see Spot Instance Requests (http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/spot-requests.html) // in the Amazon EC2 User Guide for Linux Instances. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -37615,7 +37607,7 @@ type DescribeInstancesInput struct { // The maximum number of results to return in a single call. To retrieve the // remaining results, make another call with the returned NextToken value. This // value can be between 5 and 1000. You cannot specify this parameter and the - // instance IDs parameter or tag filters in the same call. + // instance IDs parameter in the same call. MaxResults *int64 `locationName:"maxResults" type:"integer"` // The token to request the next page of results. @@ -66458,19 +66450,23 @@ type StateReason struct { // The message for the state change. // - // * Server.InsufficientInstanceCapacity: There was insufficient instance - // capacity to satisfy the launch request. + // * Server.InsufficientInstanceCapacity: There was insufficient capacity + // available to satisfy the launch request. // - // * Server.InternalError: An internal error occurred during instance launch, - // resulting in termination. + // * Server.InternalError: An internal error caused the instance to terminate + // during launch. // // * Server.ScheduledStop: The instance was stopped due to a scheduled retirement. // - // * Server.SpotInstanceTermination: A Spot Instance was terminated due to - // an increase in the Spot price. + // * Server.SpotInstanceShutdown: The instance was stopped because the number + // of Spot requests with a maximum price equal to or higher than the Spot + // price exceeded available capacity or because of an increase in the Spot + // price. // - // * Client.InternalError: A client error caused the instance to terminate - // on launch. + // * Server.SpotInstanceTermination: The instance was terminated because + // the number of Spot requests with a maximum price equal to or higher than + // the Spot price exceeded available capacity or because of an increase in + // the Spot price. // // * Client.InstanceInitiatedShutdown: The instance was shut down using the // shutdown -h command from the instance. @@ -66478,14 +66474,17 @@ type StateReason struct { // * Client.InstanceTerminated: The instance was terminated or rebooted during // AMI creation. // + // * Client.InternalError: A client error caused the instance to terminate + // during launch. + // + // * Client.InvalidSnapshot.NotFound: The specified snapshot was not found. + // // * Client.UserInitiatedShutdown: The instance was shut down using the Amazon // EC2 API. // // * Client.VolumeLimitExceeded: The limit on the number of EBS volumes or // total storage was exceeded. Decrease usage or request an increase in your - // limits. - // - // * Client.InvalidSnapshot.NotFound: The specified snapshot was not found. + // account limits. Message *string `locationName:"message" type:"string"` } @@ -66969,7 +66968,7 @@ type TagSpecification struct { _ struct{} `type:"structure"` // The type of resource to tag. Currently, the resource types that support tagging - // on creation are instance and volume. + // on creation are instance, snapshot, and volume. ResourceType *string `locationName:"resourceType" type:"string" enum:"ResourceType"` // The tags to apply to the resource. @@ -70694,6 +70693,9 @@ const ( // InstanceTypeI316xlarge is a InstanceType enum value InstanceTypeI316xlarge = "i3.16xlarge" + // InstanceTypeI3Metal is a InstanceType enum value + InstanceTypeI3Metal = "i3.metal" + // InstanceTypeHi14xlarge is a InstanceType enum value InstanceTypeHi14xlarge = "hi1.4xlarge" @@ -70754,6 +70756,24 @@ const ( // InstanceTypeC518xlarge is a InstanceType enum value InstanceTypeC518xlarge = "c5.18xlarge" + // InstanceTypeC5dLarge is a InstanceType enum value + InstanceTypeC5dLarge = "c5d.large" + + // InstanceTypeC5dXlarge is a InstanceType enum value + InstanceTypeC5dXlarge = "c5d.xlarge" + + // InstanceTypeC5d2xlarge is a InstanceType enum value + InstanceTypeC5d2xlarge = "c5d.2xlarge" + + // InstanceTypeC5d4xlarge is a InstanceType enum value + InstanceTypeC5d4xlarge = "c5d.4xlarge" + + // InstanceTypeC5d9xlarge is a InstanceType enum value + InstanceTypeC5d9xlarge = "c5d.9xlarge" + + // InstanceTypeC5d18xlarge is a InstanceType enum value + InstanceTypeC5d18xlarge = "c5d.18xlarge" + // InstanceTypeCc14xlarge is a InstanceType enum value InstanceTypeCc14xlarge = "cc1.4xlarge" @@ -70832,6 +70852,24 @@ const ( // InstanceTypeM524xlarge is a InstanceType enum value InstanceTypeM524xlarge = "m5.24xlarge" + // InstanceTypeM5dLarge is a InstanceType enum value + InstanceTypeM5dLarge = "m5d.large" + + // InstanceTypeM5dXlarge is a InstanceType enum value + InstanceTypeM5dXlarge = "m5d.xlarge" + + // InstanceTypeM5d2xlarge is a InstanceType enum value + InstanceTypeM5d2xlarge = "m5d.2xlarge" + + // InstanceTypeM5d4xlarge is a InstanceType enum value + InstanceTypeM5d4xlarge = "m5d.4xlarge" + + // InstanceTypeM5d12xlarge is a InstanceType enum value + InstanceTypeM5d12xlarge = "m5d.12xlarge" + + // InstanceTypeM5d24xlarge is a InstanceType enum value + InstanceTypeM5d24xlarge = "m5d.24xlarge" + // InstanceTypeH12xlarge is a InstanceType enum value InstanceTypeH12xlarge = "h1.2xlarge" diff --git a/vendor/github.com/aws/aws-sdk-go/service/ec2/service.go b/vendor/github.com/aws/aws-sdk-go/service/ec2/service.go index ba4433d38..6acbc43fe 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ec2/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ec2/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "ec2" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "ec2" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "EC2" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the EC2 client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/ecr/service.go b/vendor/github.com/aws/aws-sdk-go/service/ecr/service.go index 95de12e25..7bdf21370 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ecr/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ecr/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "ecr" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "ecr" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ECR" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ECR client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/ecs/api.go b/vendor/github.com/aws/aws-sdk-go/service/ecs/api.go index f6e48bfd7..ba586cfbb 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ecs/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ecs/api.go @@ -180,17 +180,21 @@ func (c *ECS) CreateServiceRequest(input *CreateServiceInput) (req *request.Requ // balancer are considered healthy if they are in the RUNNING state. Tasks for // services that do use a load balancer are considered healthy if they are in // the RUNNING state and the container instance they are hosted on is reported -// as healthy by the load balancer. The default value for minimumHealthyPercent -// is 50% in the console and 100% for the AWS CLI, the AWS SDKs, and the APIs. +// as healthy by the load balancer. The default value for a replica service +// for minimumHealthyPercent is 50% in the console and 100% for the AWS CLI, +// the AWS SDKs, and the APIs. The default value for a daemon service for minimumHealthyPercent +// is 0% for the AWS CLI, the AWS SDKs, and the APIs and 50% for the console. // // The maximumPercent parameter represents an upper limit on the number of your // service's tasks that are allowed in the RUNNING or PENDING state during a // deployment, as a percentage of the desiredCount (rounded down to the nearest // integer). This parameter enables you to define the deployment batch size. -// For example, if your service has a desiredCount of four tasks and a maximumPercent -// value of 200%, the scheduler can start four new tasks before stopping the -// four older tasks (provided that the cluster resources required to do this -// are available). The default value for maximumPercent is 200%. +// For example, if your replica service has a desiredCount of four tasks and +// a maximumPercent value of 200%, the scheduler can start four new tasks before +// stopping the four older tasks (provided that the cluster resources required +// to do this are available). The default value for a replica service for maximumPercent +// is 200%. If you are using a daemon service type, the maximumPercent should +// remain at 100%, which is the default value. // // When the service scheduler launches new tasks, it determines task placement // in your cluster using the following logic: @@ -203,11 +207,11 @@ func (c *ECS) CreateServiceRequest(input *CreateServiceInput) (req *request.Requ // Zones in this manner (although you can choose a different placement strategy) // with the placementStrategy parameter): // -// Sort the valid container instances by the fewest number of running tasks -// for this service in the same Availability Zone as the instance. For example, -// if zone A has one running service task and zones B and C each have zero, -// valid container instances in either zone B or C are considered optimal -// for placement. +// Sort the valid container instances, giving priority to instances that have +// the fewest number of running tasks for this service in their respective +// Availability Zone. For example, if zone A has one running service task +// and zones B and C each have zero, valid container instances in either +// zone B or C are considered optimal for placement. // // Place the new service task on a valid container instance in an optimal Availability // Zone (based on the previous steps), favoring container instances with @@ -4201,6 +4205,8 @@ type ContainerDefinition struct { // the Create a container (https://docs.docker.com/engine/reference/api/docker_remote_api_v1.27/#create-a-container) // section of the Docker Remote API (https://docs.docker.com/engine/reference/api/docker_remote_api_v1.27/) // and the --hostname option to docker run (https://docs.docker.com/engine/reference/run/). + // + // The hostname parameter is not supported if using the awsvpc networkMode. Hostname *string `locationName:"hostname" type:"string"` // The image used to start a container. This string is passed directly to the @@ -4707,12 +4713,12 @@ type ContainerInstance struct { RegisteredResources []*Resource `locationName:"registeredResources" type:"list"` // For CPU and memory resource types, this parameter describes the remaining - // CPU and memory on the that has not already been allocated to tasks (and is - // therefore available for new tasks). For port resource types, this parameter - // describes the ports that were reserved by the Amazon ECS container agent - // (at instance registration time) and any task containers that have reserved - // port mappings on the host (with the host or bridge network mode). Any port - // that is not specified here is available for new tasks. + // CPU and memory that has not already been allocated to tasks and is therefore + // available for new tasks. For port resource types, this parameter describes + // the ports that were reserved by the Amazon ECS container agent (at instance + // registration time) and any task containers that have reserved port mappings + // on the host (with the host or bridge network mode). Any port that is not + // specified here is available for new tasks. RemainingResources []*Resource `locationName:"remainingResources" type:"list"` // The number of tasks on the container instance that are in the RUNNING status. @@ -5027,8 +5033,8 @@ func (s *CreateClusterOutput) SetCluster(v *Cluster) *CreateClusterOutput { type CreateServiceInput struct { _ struct{} `type:"structure"` - // Unique, case-sensitive identifier you provide to ensure the idempotency of - // the request. Up to 32 ASCII characters are allowed. + // Unique, case-sensitive identifier that you provide to ensure the idempotency + // of the request. Up to 32 ASCII characters are allowed. ClientToken *string `locationName:"clientToken" type:"string"` // The short name or full Amazon Resource Name (ARN) of the cluster on which @@ -5042,9 +5048,7 @@ type CreateServiceInput struct { // The number of instantiations of the specified task definition to place and // keep running on your cluster. - // - // DesiredCount is a required field - DesiredCount *int64 `locationName:"desiredCount" type:"integer" required:"true"` + DesiredCount *int64 `locationName:"desiredCount" type:"integer"` // The period of time, in seconds, that the Amazon ECS service scheduler should // ignore unhealthy Elastic Load Balancing target health checks after a task @@ -5077,6 +5081,13 @@ type CreateServiceInput struct { // balancer. When a task from this service is placed on a container instance, // the container instance and port combination is registered as a target in // the target group specified here. + // + // Services with tasks that use the awsvpc network mode (for example, those + // with the Fargate launch type) only support Application Load Balancers and + // Network Load Balancers; Classic Load Balancers are not supported. Also, when + // you create any target groups for these services, you must choose ip as the + // target type, not instance, because tasks that use the awsvpc network mode + // are associated with an elastic network interface, not an Amazon EC2 instance. LoadBalancers []*LoadBalancer `locationName:"loadBalancers" type:"list"` // The network configuration for the service. This parameter is required for @@ -5121,6 +5132,25 @@ type CreateServiceInput struct { // in the IAM User Guide. Role *string `locationName:"role" type:"string"` + // The scheduling strategy to use for the service. For more information, see + // Services (http://docs.aws.amazon.com/AmazonECS/latest/developerguideecs_services.html). + // + // There are two service scheduler strategies available: + // + // * REPLICA-The replica scheduling strategy places and maintains the desired + // number of tasks across your cluster. By default, the service scheduler + // spreads tasks across Availability Zones. You can use task placement strategies + // and constraints to customize task placement decisions. + // + // * DAEMON-The daemon scheduling strategy deploys exactly one task on each + // active container instance that meets all of the task placement constraints + // that you specify in your cluster. When using this strategy, there is no + // need to specify a desired number of tasks, a task placement strategy, + // or use Service Auto Scaling policies. + // + // Fargate tasks do not support the DAEMON scheduling strategy. + SchedulingStrategy *string `locationName:"schedulingStrategy" type:"string" enum:"SchedulingStrategy"` + // The name of your service. Up to 255 letters (uppercase and lowercase), numbers, // hyphens, and underscores are allowed. Service names must be unique within // a cluster, but you can have similarly named services in multiple clusters @@ -5130,7 +5160,11 @@ type CreateServiceInput struct { ServiceName *string `locationName:"serviceName" type:"string" required:"true"` // The details of the service discovery registries you want to assign to this - // service. For more information, see Service Discovery (http://docs.aws.amazon.com/AmazonECS/latest/developerguideservice-discovery.html). + // service. For more information, see Service Discovery (http://docs.aws.amazon.com/AmazonECS/latest/developerguide/service-discovery.html). + // + // Service discovery is supported for Fargate tasks if using platform version + // v1.1.0 or later. For more information, see AWS Fargate Platform Versions + // (http://docs.aws.amazon.com/AmazonECS/latest/developerguide/platform_versions.html). ServiceRegistries []*ServiceRegistry `locationName:"serviceRegistries" type:"list"` // The family and revision (family:revision) or full ARN of the task definition @@ -5154,9 +5188,6 @@ func (s CreateServiceInput) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *CreateServiceInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "CreateServiceInput"} - if s.DesiredCount == nil { - invalidParams.Add(request.NewErrParamRequired("DesiredCount")) - } if s.ServiceName == nil { invalidParams.Add(request.NewErrParamRequired("ServiceName")) } @@ -5247,6 +5278,12 @@ func (s *CreateServiceInput) SetRole(v string) *CreateServiceInput { return s } +// SetSchedulingStrategy sets the SchedulingStrategy field's value. +func (s *CreateServiceInput) SetSchedulingStrategy(v string) *CreateServiceInput { + s.SchedulingStrategy = &v + return s +} + // SetServiceName sets the ServiceName field's value. func (s *CreateServiceInput) SetServiceName(v string) *CreateServiceInput { s.ServiceName = &v @@ -5442,6 +5479,11 @@ type DeleteServiceInput struct { // is assumed. Cluster *string `locationName:"cluster" type:"string"` + // If true, allows you to delete a service even if it has not been scaled down + // to zero tasks. It is only necessary to use this if the service is using the + // REPLICA scheduling strategy. + Force *bool `locationName:"force" type:"boolean"` + // The name of the service to delete. // // Service is a required field @@ -5477,6 +5519,12 @@ func (s *DeleteServiceInput) SetCluster(v string) *DeleteServiceInput { return s } +// SetForce sets the Force field's value. +func (s *DeleteServiceInput) SetForce(v bool) *DeleteServiceInput { + s.Force = &v + return s +} + // SetService sets the Service field's value. func (s *DeleteServiceInput) SetService(v string) *DeleteServiceInput { s.Service = &v @@ -5914,7 +5962,8 @@ type DescribeContainerInstancesInput struct { // default cluster is assumed. Cluster *string `locationName:"cluster" type:"string"` - // A list of container instance IDs or full ARN entries. + // A list of up to 100 container instance IDs or full Amazon Resource Name (ARN) + // entries. // // ContainerInstances is a required field ContainerInstances []*string `locationName:"containerInstances" type:"list" required:"true"` @@ -6689,12 +6738,18 @@ type LinuxParameters struct { // command: sudo docker version | grep "Server API version" InitProcessEnabled *bool `locationName:"initProcessEnabled" type:"boolean"` - // The value for the size of the /dev/shm volume. This parameter maps to the - // --shm-size option to docker run (https://docs.docker.com/engine/reference/run/). + // The value for the size (in MiB) of the /dev/shm volume. This parameter maps + // to the --shm-size option to docker run (https://docs.docker.com/engine/reference/run/). + // + // If you are using tasks that use the Fargate launch type, the sharedMemorySize + // parameter is not supported. SharedMemorySize *int64 `locationName:"sharedMemorySize" type:"integer"` - // The container path, mount options, and size of the tmpfs mount. This parameter - // maps to the --tmpfs option to docker run (https://docs.docker.com/engine/reference/run/). + // The container path, mount options, and size (in MiB) of the tmpfs mount. + // This parameter maps to the --tmpfs option to docker run (https://docs.docker.com/engine/reference/run/). + // + // If you are using tasks that use the Fargate launch type, the tmpfs parameter + // is not supported. Tmpfs []*Tmpfs `locationName:"tmpfs" type:"list"` } @@ -7125,6 +7180,9 @@ type ListServicesInput struct { // This token should be treated as an opaque identifier that is only used to // retrieve the next items in a list and not for other programmatic purposes. NextToken *string `locationName:"nextToken" type:"string"` + + // The scheduling strategy for services to list. + SchedulingStrategy *string `locationName:"schedulingStrategy" type:"string" enum:"SchedulingStrategy"` } // String returns the string representation @@ -7161,6 +7219,12 @@ func (s *ListServicesInput) SetNextToken(v string) *ListServicesInput { return s } +// SetSchedulingStrategy sets the SchedulingStrategy field's value. +func (s *ListServicesInput) SetSchedulingStrategy(v string) *ListServicesInput { + s.SchedulingStrategy = &v + return s +} + type ListServicesOutput struct { _ struct{} `type:"structure"` @@ -7581,6 +7645,13 @@ func (s *ListTasksOutput) SetTaskArns(v []*string) *ListTasksOutput { } // Details on a load balancer that is used with a service. +// +// Services with tasks that use the awsvpc network mode (for example, those +// with the Fargate launch type) only support Application Load Balancers and +// Network Load Balancers; Classic Load Balancers are not supported. Also, when +// you create any target groups for these services, you must choose ip as the +// target type, not instance, because tasks that use the awsvpc network mode +// are associated with an elastic network interface, not an Amazon EC2 instance. type LoadBalancer struct { _ struct{} `type:"structure"` @@ -7599,6 +7670,11 @@ type LoadBalancer struct { // The full Amazon Resource Name (ARN) of the Elastic Load Balancing target // group associated with a service. + // + // If your service's task definition uses the awsvpc network mode (which is + // required for the Fargate launch type), you must choose ip as the target type, + // not instance, because tasks that use the awsvpc network mode are associated + // with an elastic network interface, not an Amazon EC2 instance. TargetGroupArn *string `locationName:"targetGroupArn" type:"string"` } @@ -8542,7 +8618,8 @@ type Resource struct { // precision floating-point type. LongValue *int64 `locationName:"longValue" type:"long"` - // The name of the resource, such as cpu, memory, ports, or a user-defined resource. + // The name of the resource, such as CPU, MEMORY, PORTS, PORTS_UDP, or a user-defined + // resource. Name *string `locationName:"name" type:"string"` // When the stringSetValue type is set, the value of the resource must be a @@ -8832,6 +8909,13 @@ type Service struct { // A list of Elastic Load Balancing load balancer objects, containing the load // balancer name, the container name (as it appears in a container definition), // and the container port to access from the load balancer. + // + // Services with tasks that use the awsvpc network mode (for example, those + // with the Fargate launch type) only support Application Load Balancers and + // Network Load Balancers; Classic Load Balancers are not supported. Also, when + // you create any target groups for these services, you must choose ip as the + // target type, not instance, because tasks that use the awsvpc network mode + // are associated with an elastic network interface, not an Amazon EC2 instance. LoadBalancers []*LoadBalancer `locationName:"loadBalancers" type:"list"` // The VPC subnet and security group configuration for tasks that receive their @@ -8860,6 +8944,23 @@ type Service struct { // The number of tasks in the cluster that are in the RUNNING state. RunningCount *int64 `locationName:"runningCount" type:"integer"` + // The scheduling strategy to use for the service. For more information, see + // Services (http://docs.aws.amazon.com/AmazonECS/latest/developerguideecs_services.html). + // + // There are two service scheduler strategies available: + // + // * REPLICA-The replica scheduling strategy places and maintains the desired + // number of tasks across your cluster. By default, the service scheduler + // spreads tasks across Availability Zones. You can use task placement strategies + // and constraints to customize task placement decisions. + // + // * DAEMON-The daemon scheduling strategy deploys exactly one task on each + // container instance in your cluster. When using this strategy, do not specify + // a desired number of tasks or any task placement strategies. + // + // Fargate tasks do not support the DAEMON scheduling strategy. + SchedulingStrategy *string `locationName:"schedulingStrategy" type:"string" enum:"SchedulingStrategy"` + // The ARN that identifies the service. The ARN contains the arn:aws:ecs namespace, // followed by the region of the service, the AWS account ID of the service // owner, the service namespace, and then the service name. For example, arn:aws:ecs:region:012345678910:service/my-service. @@ -8988,6 +9089,12 @@ func (s *Service) SetRunningCount(v int64) *Service { return s } +// SetSchedulingStrategy sets the SchedulingStrategy field's value. +func (s *Service) SetSchedulingStrategy(v string) *Service { + s.SchedulingStrategy = &v + return s +} + // SetServiceArn sets the ServiceArn field's value. func (s *Service) SetServiceArn(v string) *Service { s.ServiceArn = &v @@ -9064,12 +9171,32 @@ func (s *ServiceEvent) SetMessage(v string) *ServiceEvent { type ServiceRegistry struct { _ struct{} `type:"structure"` - // The port value used if your Service Discovery service specified an SRV record. + // The container name value, already specified in the task definition, to be + // used for your service discovery service. If the task definition that your + // service task specifies uses the bridge or host network mode, you must specify + // a containerName and containerPort combination from the task definition. If + // the task definition that your service task specifies uses the awsvpc network + // mode and a type SRV DNS record is used, you must specify either a containerName + // and containerPort combination or a port value, but not both. + ContainerName *string `locationName:"containerName" type:"string"` + + // The port value, already specified in the task definition, to be used for + // your service discovery service. If the task definition your service task + // specifies uses the bridge or host network mode, you must specify a containerName + // and containerPort combination from the task definition. If the task definition + // your service task specifies uses the awsvpc network mode and a type SRV DNS + // record is used, you must specify either a containerName and containerPort + // combination or a port value, but not both. + ContainerPort *int64 `locationName:"containerPort" type:"integer"` + + // The port value used if your service discovery service specified an SRV record. + // This field is required if both the awsvpc network mode and SRV records are + // used. Port *int64 `locationName:"port" type:"integer"` - // The Amazon Resource Name (ARN) of the Service Registry. The currently supported - // service registry is Amazon Route 53 Auto Naming Service. For more information, - // see Service (https://docs.aws.amazon.com/Route53/latest/APIReference/API_autonaming_Service.html). + // The Amazon Resource Name (ARN) of the service registry. The currently supported + // service registry is Amazon Route 53 Auto Naming. For more information, see + // Service (https://docs.aws.amazon.com/Route53/latest/APIReference/API_autonaming_Service.html). RegistryArn *string `locationName:"registryArn" type:"string"` } @@ -9083,6 +9210,18 @@ func (s ServiceRegistry) GoString() string { return s.String() } +// SetContainerName sets the ContainerName field's value. +func (s *ServiceRegistry) SetContainerName(v string) *ServiceRegistry { + s.ContainerName = &v + return s +} + +// SetContainerPort sets the ContainerPort field's value. +func (s *ServiceRegistry) SetContainerPort(v int64) *ServiceRegistry { + s.ContainerPort = &v + return s +} + // SetPort sets the Port field's value. func (s *ServiceRegistry) SetPort(v int64) *ServiceRegistry { s.Port = &v @@ -10261,7 +10400,7 @@ type Tmpfs struct { // | "slave" | "rslave" | "relatime" | "norelatime" | "strictatime" | "nostrictatime" MountOptions []*string `locationName:"mountOptions" type:"list"` - // The size of the tmpfs volume. + // The size (in MiB) of the tmpfs volume. // // Size is a required field Size *int64 `locationName:"size" type:"integer" required:"true"` @@ -10991,6 +11130,14 @@ const ( PlacementStrategyTypeBinpack = "binpack" ) +const ( + // SchedulingStrategyReplica is a SchedulingStrategy enum value + SchedulingStrategyReplica = "REPLICA" + + // SchedulingStrategyDaemon is a SchedulingStrategy enum value + SchedulingStrategyDaemon = "DAEMON" +) + const ( // SortOrderAsc is a SortOrder enum value SortOrderAsc = "ASC" diff --git a/vendor/github.com/aws/aws-sdk-go/service/ecs/service.go b/vendor/github.com/aws/aws-sdk-go/service/ecs/service.go index 6082b9282..c268614ec 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ecs/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ecs/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "ecs" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "ecs" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ECS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ECS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/efs/service.go b/vendor/github.com/aws/aws-sdk-go/service/efs/service.go index 4d6bb2cdc..6b1a11c90 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/efs/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/efs/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elasticfilesystem" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elasticfilesystem" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "EFS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the EFS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/eks/api.go b/vendor/github.com/aws/aws-sdk-go/service/eks/api.go new file mode 100644 index 000000000..0c6d34213 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/eks/api.go @@ -0,0 +1,1012 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package eks + +import ( + "time" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awsutil" + "github.com/aws/aws-sdk-go/aws/request" +) + +const opCreateCluster = "CreateCluster" + +// CreateClusterRequest generates a "aws/request.Request" representing the +// client's request for the CreateCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateCluster for more information on using the CreateCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateClusterRequest method. +// req, resp := client.CreateClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/CreateCluster +func (c *EKS) CreateClusterRequest(input *CreateClusterInput) (req *request.Request, output *CreateClusterOutput) { + op := &request.Operation{ + Name: opCreateCluster, + HTTPMethod: "POST", + HTTPPath: "/clusters", + } + + if input == nil { + input = &CreateClusterInput{} + } + + output = &CreateClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateCluster API operation for Amazon Elastic Container Service for Kubernetes. +// +// Creates an Amazon EKS control plane. +// +// The Amazon EKS control plane consists of control plane instances that run +// the Kubernetes software, like etcd and the API server. The control plane +// runs in an account managed by AWS, and the Kubernetes API is exposed via +// the Amazon EKS API server endpoint. +// +// Amazon EKS worker nodes run in your AWS account and connect to your cluster's +// control plane via the Kubernetes API server endpoint and a certificate file +// that is created for your cluster. +// +// The cluster control plane is provisioned across multiple Availability Zones +// and fronted by an Elastic Load Balancing Network Load Balancer. Amazon EKS +// also provisions elastic network interfaces in your VPC subnets to provide +// connectivity from the control plane instances to the worker nodes (for example, +// to support kubectl exec, logs, and proxy data flows). +// +// After you create an Amazon EKS cluster, you must configure your Kubernetes +// tooling to communicate with the API server and launch worker nodes into your +// cluster. For more information, see Managing Cluster Authentication (http://docs.aws.amazon.com/eks/latest/userguide/managing-auth.html) +// and Launching Amazon EKS Worker Nodes (http://docs.aws.amazon.com/eks/latest/userguide/launch-workers.html)in +// the Amazon EKS User Guide. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Elastic Container Service for Kubernetes's +// API operation CreateCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceInUseException "ResourceInUseException" +// The specified resource is in use. +// +// * ErrCodeResourceLimitExceededException "ResourceLimitExceededException" +// You have encountered a service limit on the specified resource. +// +// * ErrCodeInvalidParameterException "InvalidParameterException" +// The specified parameter is invalid. Review the available parameters for the +// API request. +// +// * ErrCodeClientException "ClientException" +// These errors are usually caused by a client action, such as using an action +// or resource on behalf of a user that doesn't have permissions to use the +// action or resource, or specifying an identifier that is not valid. +// +// * ErrCodeServerException "ServerException" +// These errors are usually caused by a server-side issue. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is unavailable, back off and retry the operation. +// +// * ErrCodeUnsupportedAvailabilityZoneException "UnsupportedAvailabilityZoneException" +// At least one of your specified cluster subnets is in an Availability Zone +// that does not support Amazon EKS. The exception output will specify the supported +// Availability Zones for your account, from which you can choose subnets for +// your cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/CreateCluster +func (c *EKS) CreateCluster(input *CreateClusterInput) (*CreateClusterOutput, error) { + req, out := c.CreateClusterRequest(input) + return out, req.Send() +} + +// CreateClusterWithContext is the same as CreateCluster with the addition of +// the ability to pass a context and additional request options. +// +// See CreateCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *EKS) CreateClusterWithContext(ctx aws.Context, input *CreateClusterInput, opts ...request.Option) (*CreateClusterOutput, error) { + req, out := c.CreateClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteCluster = "DeleteCluster" + +// DeleteClusterRequest generates a "aws/request.Request" representing the +// client's request for the DeleteCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteCluster for more information on using the DeleteCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteClusterRequest method. +// req, resp := client.DeleteClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/DeleteCluster +func (c *EKS) DeleteClusterRequest(input *DeleteClusterInput) (req *request.Request, output *DeleteClusterOutput) { + op := &request.Operation{ + Name: opDeleteCluster, + HTTPMethod: "DELETE", + HTTPPath: "/clusters/{name}", + } + + if input == nil { + input = &DeleteClusterInput{} + } + + output = &DeleteClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteCluster API operation for Amazon Elastic Container Service for Kubernetes. +// +// Deletes the Amazon EKS cluster control plane. +// +// If you have active services in your cluster that are associated with a load +// balancer, you must delete those services before deleting the cluster so that +// the load balancers are deleted properly. Otherwise, you can have orphaned +// resources in your VPC that prevent you from being able to delete the VPC. +// For more information, see Deleting a Cluster (http://docs.aws.amazon.com/eks/latest/userguide/delete-cluster.html) +// in the Amazon EKS User Guide. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Elastic Container Service for Kubernetes's +// API operation DeleteCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceInUseException "ResourceInUseException" +// The specified resource is in use. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource could not be found. You can view your available clusters +// with ListClusters. Amazon EKS clusters are region-specific. +// +// * ErrCodeClientException "ClientException" +// These errors are usually caused by a client action, such as using an action +// or resource on behalf of a user that doesn't have permissions to use the +// action or resource, or specifying an identifier that is not valid. +// +// * ErrCodeServerException "ServerException" +// These errors are usually caused by a server-side issue. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is unavailable, back off and retry the operation. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/DeleteCluster +func (c *EKS) DeleteCluster(input *DeleteClusterInput) (*DeleteClusterOutput, error) { + req, out := c.DeleteClusterRequest(input) + return out, req.Send() +} + +// DeleteClusterWithContext is the same as DeleteCluster with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *EKS) DeleteClusterWithContext(ctx aws.Context, input *DeleteClusterInput, opts ...request.Option) (*DeleteClusterOutput, error) { + req, out := c.DeleteClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeCluster = "DescribeCluster" + +// DescribeClusterRequest generates a "aws/request.Request" representing the +// client's request for the DescribeCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeCluster for more information on using the DescribeCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeClusterRequest method. +// req, resp := client.DescribeClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/DescribeCluster +func (c *EKS) DescribeClusterRequest(input *DescribeClusterInput) (req *request.Request, output *DescribeClusterOutput) { + op := &request.Operation{ + Name: opDescribeCluster, + HTTPMethod: "GET", + HTTPPath: "/clusters/{name}", + } + + if input == nil { + input = &DescribeClusterInput{} + } + + output = &DescribeClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeCluster API operation for Amazon Elastic Container Service for Kubernetes. +// +// Returns descriptive information about an Amazon EKS cluster. +// +// The API server endpoint and certificate authority data returned by this operation +// are required for kubelet and kubectl to communicate with your Kubernetes +// API server. For more information, see Create a kubeconfig for Amazon EKS +// (http://docs.aws.amazon.com/eks/latest/userguide/create-kubeconfig.html). +// +// The API server endpoint and certificate authority data are not available +// until the cluster reaches the ACTIVE state. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Elastic Container Service for Kubernetes's +// API operation DescribeCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource could not be found. You can view your available clusters +// with ListClusters. Amazon EKS clusters are region-specific. +// +// * ErrCodeClientException "ClientException" +// These errors are usually caused by a client action, such as using an action +// or resource on behalf of a user that doesn't have permissions to use the +// action or resource, or specifying an identifier that is not valid. +// +// * ErrCodeServerException "ServerException" +// These errors are usually caused by a server-side issue. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is unavailable, back off and retry the operation. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/DescribeCluster +func (c *EKS) DescribeCluster(input *DescribeClusterInput) (*DescribeClusterOutput, error) { + req, out := c.DescribeClusterRequest(input) + return out, req.Send() +} + +// DescribeClusterWithContext is the same as DescribeCluster with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *EKS) DescribeClusterWithContext(ctx aws.Context, input *DescribeClusterInput, opts ...request.Option) (*DescribeClusterOutput, error) { + req, out := c.DescribeClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opListClusters = "ListClusters" + +// ListClustersRequest generates a "aws/request.Request" representing the +// client's request for the ListClusters operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListClusters for more information on using the ListClusters +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListClustersRequest method. +// req, resp := client.ListClustersRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/ListClusters +func (c *EKS) ListClustersRequest(input *ListClustersInput) (req *request.Request, output *ListClustersOutput) { + op := &request.Operation{ + Name: opListClusters, + HTTPMethod: "GET", + HTTPPath: "/clusters", + } + + if input == nil { + input = &ListClustersInput{} + } + + output = &ListClustersOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListClusters API operation for Amazon Elastic Container Service for Kubernetes. +// +// Lists the Amazon EKS clusters in your AWS account in the specified region. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Elastic Container Service for Kubernetes's +// API operation ListClusters for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidParameterException "InvalidParameterException" +// The specified parameter is invalid. Review the available parameters for the +// API request. +// +// * ErrCodeClientException "ClientException" +// These errors are usually caused by a client action, such as using an action +// or resource on behalf of a user that doesn't have permissions to use the +// action or resource, or specifying an identifier that is not valid. +// +// * ErrCodeServerException "ServerException" +// These errors are usually caused by a server-side issue. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is unavailable, back off and retry the operation. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01/ListClusters +func (c *EKS) ListClusters(input *ListClustersInput) (*ListClustersOutput, error) { + req, out := c.ListClustersRequest(input) + return out, req.Send() +} + +// ListClustersWithContext is the same as ListClusters with the addition of +// the ability to pass a context and additional request options. +// +// See ListClusters for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *EKS) ListClustersWithContext(ctx aws.Context, input *ListClustersInput, opts ...request.Option) (*ListClustersOutput, error) { + req, out := c.ListClustersRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// An object representing the certificate-authority-data for your cluster. +type Certificate struct { + _ struct{} `type:"structure"` + + // The base64 encoded certificate data required to communicate with your cluster. + // Add this to the certificate-authority-data section of the kubeconfig file + // for your cluster. + Data *string `locationName:"data" type:"string"` +} + +// String returns the string representation +func (s Certificate) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Certificate) GoString() string { + return s.String() +} + +// SetData sets the Data field's value. +func (s *Certificate) SetData(v string) *Certificate { + s.Data = &v + return s +} + +// An object representing an Amazon EKS cluster. +type Cluster struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the cluster. + Arn *string `locationName:"arn" type:"string"` + + // The certificate-authority-data for your cluster. + CertificateAuthority *Certificate `locationName:"certificateAuthority" type:"structure"` + + // Unique, case-sensitive identifier you provide to ensure the idempotency of + // the request. + ClientRequestToken *string `locationName:"clientRequestToken" type:"string"` + + // The Unix epoch time stamp in seconds for when the cluster was created. + CreatedAt *time.Time `locationName:"createdAt" type:"timestamp" timestampFormat:"unix"` + + // The endpoint for your Kubernetes API server. + Endpoint *string `locationName:"endpoint" type:"string"` + + // The name of the cluster. + Name *string `locationName:"name" type:"string"` + + // The VPC subnets and security groups used by the cluster control plane. Amazon + // EKS VPC resources have specific requirements to work properly with Kubernetes. + // For more information, see Cluster VPC Considerations (http://docs.aws.amazon.com/eks/latest/userguide/network_reqs.html) + // and Cluster Security Group Considerations (http://docs.aws.amazon.com/eks/latest/userguide/sec-group-reqs.html) + // in the Amazon EKS User Guide. + ResourcesVpcConfig *VpcConfigResponse `locationName:"resourcesVpcConfig" type:"structure"` + + // The Amazon Resource Name (ARN) of the IAM role that provides permissions + // for the Kubernetes control plane to make calls to AWS API operations on your + // behalf. + RoleArn *string `locationName:"roleArn" type:"string"` + + // The current status of the cluster. + Status *string `locationName:"status" type:"string" enum:"ClusterStatus"` + + // The Kubernetes server version for the cluster. + Version *string `locationName:"version" type:"string"` +} + +// String returns the string representation +func (s Cluster) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Cluster) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *Cluster) SetArn(v string) *Cluster { + s.Arn = &v + return s +} + +// SetCertificateAuthority sets the CertificateAuthority field's value. +func (s *Cluster) SetCertificateAuthority(v *Certificate) *Cluster { + s.CertificateAuthority = v + return s +} + +// SetClientRequestToken sets the ClientRequestToken field's value. +func (s *Cluster) SetClientRequestToken(v string) *Cluster { + s.ClientRequestToken = &v + return s +} + +// SetCreatedAt sets the CreatedAt field's value. +func (s *Cluster) SetCreatedAt(v time.Time) *Cluster { + s.CreatedAt = &v + return s +} + +// SetEndpoint sets the Endpoint field's value. +func (s *Cluster) SetEndpoint(v string) *Cluster { + s.Endpoint = &v + return s +} + +// SetName sets the Name field's value. +func (s *Cluster) SetName(v string) *Cluster { + s.Name = &v + return s +} + +// SetResourcesVpcConfig sets the ResourcesVpcConfig field's value. +func (s *Cluster) SetResourcesVpcConfig(v *VpcConfigResponse) *Cluster { + s.ResourcesVpcConfig = v + return s +} + +// SetRoleArn sets the RoleArn field's value. +func (s *Cluster) SetRoleArn(v string) *Cluster { + s.RoleArn = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *Cluster) SetStatus(v string) *Cluster { + s.Status = &v + return s +} + +// SetVersion sets the Version field's value. +func (s *Cluster) SetVersion(v string) *Cluster { + s.Version = &v + return s +} + +type CreateClusterInput struct { + _ struct{} `type:"structure"` + + // Unique, case-sensitive identifier you provide to ensure the idempotency of + // the request. + ClientRequestToken *string `locationName:"clientRequestToken" type:"string" idempotencyToken:"true"` + + // The unique name to give to your cluster. + // + // Name is a required field + Name *string `locationName:"name" min:"1" type:"string" required:"true"` + + // The VPC subnets and security groups used by the cluster control plane. Amazon + // EKS VPC resources have specific requirements to work properly with Kubernetes. + // For more information, see Cluster VPC Considerations (http://docs.aws.amazon.com/eks/latest/userguide/network_reqs.html) + // and Cluster Security Group Considerations (http://docs.aws.amazon.com/eks/latest/userguide/sec-group-reqs.html) + // in the Amazon EKS User Guide. + // + // ResourcesVpcConfig is a required field + ResourcesVpcConfig *VpcConfigRequest `locationName:"resourcesVpcConfig" type:"structure" required:"true"` + + // The Amazon Resource Name (ARN) of the IAM role that provides permissions + // for Amazon EKS to make calls to other AWS API operations on your behalf. + // For more information, see Amazon EKS Service IAM Role (http://docs.aws.amazon.com/eks/latest/userguide/service_IAM_role.html) + // in the Amazon EKS User Guide + // + // RoleArn is a required field + RoleArn *string `locationName:"roleArn" type:"string" required:"true"` + + // The desired Kubernetes version for your cluster. If you do not specify a + // value here, the latest version available in Amazon EKS is used. + Version *string `locationName:"version" type:"string"` +} + +// String returns the string representation +func (s CreateClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateClusterInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + if s.Name != nil && len(*s.Name) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Name", 1)) + } + if s.ResourcesVpcConfig == nil { + invalidParams.Add(request.NewErrParamRequired("ResourcesVpcConfig")) + } + if s.RoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("RoleArn")) + } + if s.ResourcesVpcConfig != nil { + if err := s.ResourcesVpcConfig.Validate(); err != nil { + invalidParams.AddNested("ResourcesVpcConfig", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetClientRequestToken sets the ClientRequestToken field's value. +func (s *CreateClusterInput) SetClientRequestToken(v string) *CreateClusterInput { + s.ClientRequestToken = &v + return s +} + +// SetName sets the Name field's value. +func (s *CreateClusterInput) SetName(v string) *CreateClusterInput { + s.Name = &v + return s +} + +// SetResourcesVpcConfig sets the ResourcesVpcConfig field's value. +func (s *CreateClusterInput) SetResourcesVpcConfig(v *VpcConfigRequest) *CreateClusterInput { + s.ResourcesVpcConfig = v + return s +} + +// SetRoleArn sets the RoleArn field's value. +func (s *CreateClusterInput) SetRoleArn(v string) *CreateClusterInput { + s.RoleArn = &v + return s +} + +// SetVersion sets the Version field's value. +func (s *CreateClusterInput) SetVersion(v string) *CreateClusterInput { + s.Version = &v + return s +} + +type CreateClusterOutput struct { + _ struct{} `type:"structure"` + + // The full description of your new cluster. + Cluster *Cluster `locationName:"cluster" type:"structure"` +} + +// String returns the string representation +func (s CreateClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateClusterOutput) GoString() string { + return s.String() +} + +// SetCluster sets the Cluster field's value. +func (s *CreateClusterOutput) SetCluster(v *Cluster) *CreateClusterOutput { + s.Cluster = v + return s +} + +type DeleteClusterInput struct { + _ struct{} `type:"structure"` + + // The name of the cluster to delete. + // + // Name is a required field + Name *string `location:"uri" locationName:"name" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteClusterInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *DeleteClusterInput) SetName(v string) *DeleteClusterInput { + s.Name = &v + return s +} + +type DeleteClusterOutput struct { + _ struct{} `type:"structure"` + + // The full description of the cluster to delete. + Cluster *Cluster `locationName:"cluster" type:"structure"` +} + +// String returns the string representation +func (s DeleteClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteClusterOutput) GoString() string { + return s.String() +} + +// SetCluster sets the Cluster field's value. +func (s *DeleteClusterOutput) SetCluster(v *Cluster) *DeleteClusterOutput { + s.Cluster = v + return s +} + +type DescribeClusterInput struct { + _ struct{} `type:"structure"` + + // The name of the cluster to describe. + // + // Name is a required field + Name *string `location:"uri" locationName:"name" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeClusterInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *DescribeClusterInput) SetName(v string) *DescribeClusterInput { + s.Name = &v + return s +} + +type DescribeClusterOutput struct { + _ struct{} `type:"structure"` + + // The full description of your specified cluster. + Cluster *Cluster `locationName:"cluster" type:"structure"` +} + +// String returns the string representation +func (s DescribeClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeClusterOutput) GoString() string { + return s.String() +} + +// SetCluster sets the Cluster field's value. +func (s *DescribeClusterOutput) SetCluster(v *Cluster) *DescribeClusterOutput { + s.Cluster = v + return s +} + +type ListClustersInput struct { + _ struct{} `type:"structure"` + + // The maximum number of cluster results returned by ListClusters in paginated + // output. When this parameter is used, ListClusters only returns maxResults + // results in a single page along with a nextToken response element. The remaining + // results of the initial request can be seen by sending another ListClusters + // request with the returned nextToken value. This value can be between 1 and + // 100. If this parameter is not used, then ListClusters returns up to 100 results + // and a nextToken value if applicable. + MaxResults *int64 `location:"querystring" locationName:"maxResults" min:"1" type:"integer"` + + // The nextToken value returned from a previous paginated ListClusters request + // where maxResults was used and the results exceeded the value of that parameter. + // Pagination continues from the end of the previous results that returned the + // nextToken value. + // + // This token should be treated as an opaque identifier that is only used to + // retrieve the next items in a list and not for other programmatic purposes. + NextToken *string `location:"querystring" locationName:"nextToken" type:"string"` +} + +// String returns the string representation +func (s ListClustersInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListClustersInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListClustersInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListClustersInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListClustersInput) SetMaxResults(v int64) *ListClustersInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListClustersInput) SetNextToken(v string) *ListClustersInput { + s.NextToken = &v + return s +} + +type ListClustersOutput struct { + _ struct{} `type:"structure"` + + // A list of all of the clusters for your account in the specified region. + Clusters []*string `locationName:"clusters" type:"list"` + + // The nextToken value to include in a future ListClusters request. When the + // results of a ListClusters request exceed maxResults, this value can be used + // to retrieve the next page of results. This value is null when there are no + // more results to return. + NextToken *string `locationName:"nextToken" type:"string"` +} + +// String returns the string representation +func (s ListClustersOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListClustersOutput) GoString() string { + return s.String() +} + +// SetClusters sets the Clusters field's value. +func (s *ListClustersOutput) SetClusters(v []*string) *ListClustersOutput { + s.Clusters = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListClustersOutput) SetNextToken(v string) *ListClustersOutput { + s.NextToken = &v + return s +} + +// An object representing an Amazon EKS cluster VPC configuration request. +type VpcConfigRequest struct { + _ struct{} `type:"structure"` + + // Specify one or more security groups for the cross-account elastic network + // interfaces that Amazon EKS creates to use to allow communication between + // your worker nodes and the Kubernetes control plane. + SecurityGroupIds []*string `locationName:"securityGroupIds" type:"list"` + + // Specify subnets for your Amazon EKS worker nodes. Amazon EKS creates cross-account + // elastic network interfaces in these subnets to allow communication between + // your worker nodes and the Kubernetes control plane. + // + // SubnetIds is a required field + SubnetIds []*string `locationName:"subnetIds" type:"list" required:"true"` +} + +// String returns the string representation +func (s VpcConfigRequest) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s VpcConfigRequest) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *VpcConfigRequest) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "VpcConfigRequest"} + if s.SubnetIds == nil { + invalidParams.Add(request.NewErrParamRequired("SubnetIds")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSecurityGroupIds sets the SecurityGroupIds field's value. +func (s *VpcConfigRequest) SetSecurityGroupIds(v []*string) *VpcConfigRequest { + s.SecurityGroupIds = v + return s +} + +// SetSubnetIds sets the SubnetIds field's value. +func (s *VpcConfigRequest) SetSubnetIds(v []*string) *VpcConfigRequest { + s.SubnetIds = v + return s +} + +// An object representing an Amazon EKS cluster VPC configuration response. +type VpcConfigResponse struct { + _ struct{} `type:"structure"` + + // The security groups associated with the cross-account elastic network interfaces + // that are used to allow communication between your worker nodes and the Kubernetes + // control plane. + SecurityGroupIds []*string `locationName:"securityGroupIds" type:"list"` + + // The subnets associated with your cluster. + SubnetIds []*string `locationName:"subnetIds" type:"list"` + + // The VPC associated with your cluster. + VpcId *string `locationName:"vpcId" type:"string"` +} + +// String returns the string representation +func (s VpcConfigResponse) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s VpcConfigResponse) GoString() string { + return s.String() +} + +// SetSecurityGroupIds sets the SecurityGroupIds field's value. +func (s *VpcConfigResponse) SetSecurityGroupIds(v []*string) *VpcConfigResponse { + s.SecurityGroupIds = v + return s +} + +// SetSubnetIds sets the SubnetIds field's value. +func (s *VpcConfigResponse) SetSubnetIds(v []*string) *VpcConfigResponse { + s.SubnetIds = v + return s +} + +// SetVpcId sets the VpcId field's value. +func (s *VpcConfigResponse) SetVpcId(v string) *VpcConfigResponse { + s.VpcId = &v + return s +} + +const ( + // ClusterStatusCreating is a ClusterStatus enum value + ClusterStatusCreating = "CREATING" + + // ClusterStatusActive is a ClusterStatus enum value + ClusterStatusActive = "ACTIVE" + + // ClusterStatusDeleting is a ClusterStatus enum value + ClusterStatusDeleting = "DELETING" + + // ClusterStatusFailed is a ClusterStatus enum value + ClusterStatusFailed = "FAILED" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/eks/doc.go b/vendor/github.com/aws/aws-sdk-go/service/eks/doc.go new file mode 100644 index 000000000..9f819afc6 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/eks/doc.go @@ -0,0 +1,54 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package eks provides the client and types for making API +// requests to Amazon Elastic Container Service for Kubernetes. +// +// Amazon Elastic Container Service for Kubernetes (Amazon EKS) is a managed +// service that makes it easy for you to run Kubernetes on AWS without needing +// to stand up or maintain your own Kubernetes control plane. Kubernetes is +// an open-source system for automating the deployment, scaling, and management +// of containerized applications. +// +// Amazon EKS runs three Kubernetes control plane instances across three Availability +// Zones to ensure high availability. Amazon EKS automatically detects and replaces +// unhealthy control plane instances, and it provides automated version upgrades +// and patching for them. +// +// Amazon EKS is also integrated with many AWS services to provide scalability +// and security for your applications, including the following: +// +// * Elastic Load Balancing for load distribution +// +// * IAM for authentication +// +// * Amazon VPC for isolation +// +// Amazon EKS runs up to date versions of the open-source Kubernetes software, +// so you can use all the existing plugins and tooling from the Kubernetes community. +// Applications running on Amazon EKS are fully compatible with applications +// running on any standard Kubernetes environment, whether running in on-premises +// data centers or public clouds. This means that you can easily migrate any +// standard Kubernetes application to Amazon EKS without any code modification +// required. +// +// See https://docs.aws.amazon.com/goto/WebAPI/eks-2017-11-01 for more information on this service. +// +// See eks package documentation for more information. +// https://docs.aws.amazon.com/sdk-for-go/api/service/eks/ +// +// Using the Client +// +// To contact Amazon Elastic Container Service for Kubernetes with the SDK use the New function to create +// a new service client. With that client you can make API requests to the service. +// These clients are safe to use concurrently. +// +// See the SDK's documentation for more information on how to use the SDK. +// https://docs.aws.amazon.com/sdk-for-go/api/ +// +// See aws.Config documentation for more information on configuring SDK clients. +// https://docs.aws.amazon.com/sdk-for-go/api/aws/#Config +// +// See the Amazon Elastic Container Service for Kubernetes client EKS for more +// information on creating client for this service. +// https://docs.aws.amazon.com/sdk-for-go/api/service/eks/#New +package eks diff --git a/vendor/github.com/aws/aws-sdk-go/service/eks/eksiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/eks/eksiface/interface.go new file mode 100644 index 000000000..970953be5 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/eks/eksiface/interface.go @@ -0,0 +1,80 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package eksiface provides an interface to enable mocking the Amazon Elastic Container Service for Kubernetes service client +// for testing your code. +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. +package eksiface + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/service/eks" +) + +// EKSAPI provides an interface to enable mocking the +// eks.EKS service client's API operation, +// paginators, and waiters. This make unit testing your code that calls out +// to the SDK's service client's calls easier. +// +// The best way to use this interface is so the SDK's service client's calls +// can be stubbed out for unit testing your code with the SDK without needing +// to inject custom request handlers into the SDK's request pipeline. +// +// // myFunc uses an SDK service client to make a request to +// // Amazon Elastic Container Service for Kubernetes. +// func myFunc(svc eksiface.EKSAPI) bool { +// // Make svc.CreateCluster request +// } +// +// func main() { +// sess := session.New() +// svc := eks.New(sess) +// +// myFunc(svc) +// } +// +// In your _test.go file: +// +// // Define a mock struct to be used in your unit tests of myFunc. +// type mockEKSClient struct { +// eksiface.EKSAPI +// } +// func (m *mockEKSClient) CreateCluster(input *eks.CreateClusterInput) (*eks.CreateClusterOutput, error) { +// // mock response/functionality +// } +// +// func TestMyFunc(t *testing.T) { +// // Setup Test +// mockSvc := &mockEKSClient{} +// +// myfunc(mockSvc) +// +// // Verify myFunc's functionality +// } +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. Its suggested to use the pattern above for testing, or using +// tooling to generate mocks to satisfy the interfaces. +type EKSAPI interface { + CreateCluster(*eks.CreateClusterInput) (*eks.CreateClusterOutput, error) + CreateClusterWithContext(aws.Context, *eks.CreateClusterInput, ...request.Option) (*eks.CreateClusterOutput, error) + CreateClusterRequest(*eks.CreateClusterInput) (*request.Request, *eks.CreateClusterOutput) + + DeleteCluster(*eks.DeleteClusterInput) (*eks.DeleteClusterOutput, error) + DeleteClusterWithContext(aws.Context, *eks.DeleteClusterInput, ...request.Option) (*eks.DeleteClusterOutput, error) + DeleteClusterRequest(*eks.DeleteClusterInput) (*request.Request, *eks.DeleteClusterOutput) + + DescribeCluster(*eks.DescribeClusterInput) (*eks.DescribeClusterOutput, error) + DescribeClusterWithContext(aws.Context, *eks.DescribeClusterInput, ...request.Option) (*eks.DescribeClusterOutput, error) + DescribeClusterRequest(*eks.DescribeClusterInput) (*request.Request, *eks.DescribeClusterOutput) + + ListClusters(*eks.ListClustersInput) (*eks.ListClustersOutput, error) + ListClustersWithContext(aws.Context, *eks.ListClustersInput, ...request.Option) (*eks.ListClustersOutput, error) + ListClustersRequest(*eks.ListClustersInput) (*request.Request, *eks.ListClustersOutput) +} + +var _ EKSAPI = (*eks.EKS)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/eks/errors.go b/vendor/github.com/aws/aws-sdk-go/service/eks/errors.go new file mode 100644 index 000000000..825e7d2db --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/eks/errors.go @@ -0,0 +1,61 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package eks + +const ( + + // ErrCodeClientException for service response error code + // "ClientException". + // + // These errors are usually caused by a client action, such as using an action + // or resource on behalf of a user that doesn't have permissions to use the + // action or resource, or specifying an identifier that is not valid. + ErrCodeClientException = "ClientException" + + // ErrCodeInvalidParameterException for service response error code + // "InvalidParameterException". + // + // The specified parameter is invalid. Review the available parameters for the + // API request. + ErrCodeInvalidParameterException = "InvalidParameterException" + + // ErrCodeResourceInUseException for service response error code + // "ResourceInUseException". + // + // The specified resource is in use. + ErrCodeResourceInUseException = "ResourceInUseException" + + // ErrCodeResourceLimitExceededException for service response error code + // "ResourceLimitExceededException". + // + // You have encountered a service limit on the specified resource. + ErrCodeResourceLimitExceededException = "ResourceLimitExceededException" + + // ErrCodeResourceNotFoundException for service response error code + // "ResourceNotFoundException". + // + // The specified resource could not be found. You can view your available clusters + // with ListClusters. Amazon EKS clusters are region-specific. + ErrCodeResourceNotFoundException = "ResourceNotFoundException" + + // ErrCodeServerException for service response error code + // "ServerException". + // + // These errors are usually caused by a server-side issue. + ErrCodeServerException = "ServerException" + + // ErrCodeServiceUnavailableException for service response error code + // "ServiceUnavailableException". + // + // The service is unavailable, back off and retry the operation. + ErrCodeServiceUnavailableException = "ServiceUnavailableException" + + // ErrCodeUnsupportedAvailabilityZoneException for service response error code + // "UnsupportedAvailabilityZoneException". + // + // At least one of your specified cluster subnets is in an Availability Zone + // that does not support Amazon EKS. The exception output will specify the supported + // Availability Zones for your account, from which you can choose subnets for + // your cluster. + ErrCodeUnsupportedAvailabilityZoneException = "UnsupportedAvailabilityZoneException" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/eks/examples_test.go b/vendor/github.com/aws/aws-sdk-go/service/eks/examples_test.go new file mode 100644 index 000000000..4ddb524ed --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/eks/examples_test.go @@ -0,0 +1,185 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package eks_test + +import ( + "fmt" + "strings" + "time" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awserr" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/service/eks" +) + +var _ time.Duration +var _ strings.Reader +var _ aws.Config + +func parseTime(layout, value string) *time.Time { + t, err := time.Parse(layout, value) + if err != nil { + panic(err) + } + return &t +} + +// To create a new cluster +// +// The following example creates an Amazon EKS cluster called prod. +func ExampleEKS_CreateCluster_shared00() { + svc := eks.New(session.New()) + input := &eks.CreateClusterInput{ + ClientRequestToken: aws.String("1d2129a1-3d38-460a-9756-e5b91fddb951"), + Name: aws.String("prod"), + ResourcesVpcConfig: &eks.VpcConfigRequest{ + SecurityGroupIds: []*string{ + aws.String("sg-6979fe18"), + }, + SubnetIds: []*string{ + aws.String("subnet-6782e71e"), + aws.String("subnet-e7e761ac"), + }, + }, + RoleArn: aws.String("arn:aws:iam::012345678910:role/eks-service-role-AWSServiceRoleForAmazonEKS-J7ONKE3BQ4PI"), + Version: aws.String("1.10"), + } + + result, err := svc.CreateCluster(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case eks.ErrCodeResourceInUseException: + fmt.Println(eks.ErrCodeResourceInUseException, aerr.Error()) + case eks.ErrCodeResourceLimitExceededException: + fmt.Println(eks.ErrCodeResourceLimitExceededException, aerr.Error()) + case eks.ErrCodeInvalidParameterException: + fmt.Println(eks.ErrCodeInvalidParameterException, aerr.Error()) + case eks.ErrCodeClientException: + fmt.Println(eks.ErrCodeClientException, aerr.Error()) + case eks.ErrCodeServerException: + fmt.Println(eks.ErrCodeServerException, aerr.Error()) + case eks.ErrCodeServiceUnavailableException: + fmt.Println(eks.ErrCodeServiceUnavailableException, aerr.Error()) + case eks.ErrCodeUnsupportedAvailabilityZoneException: + fmt.Println(eks.ErrCodeUnsupportedAvailabilityZoneException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} + +// To delete a cluster +// +// This example command deletes a cluster named `devel` in your default region. +func ExampleEKS_DeleteCluster_shared00() { + svc := eks.New(session.New()) + input := &eks.DeleteClusterInput{ + Name: aws.String("devel"), + } + + result, err := svc.DeleteCluster(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case eks.ErrCodeResourceInUseException: + fmt.Println(eks.ErrCodeResourceInUseException, aerr.Error()) + case eks.ErrCodeResourceNotFoundException: + fmt.Println(eks.ErrCodeResourceNotFoundException, aerr.Error()) + case eks.ErrCodeClientException: + fmt.Println(eks.ErrCodeClientException, aerr.Error()) + case eks.ErrCodeServerException: + fmt.Println(eks.ErrCodeServerException, aerr.Error()) + case eks.ErrCodeServiceUnavailableException: + fmt.Println(eks.ErrCodeServiceUnavailableException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} + +// To describe a cluster +// +// This example command provides a description of the specified cluster in your default +// region. +func ExampleEKS_DescribeCluster_shared00() { + svc := eks.New(session.New()) + input := &eks.DescribeClusterInput{ + Name: aws.String("devel"), + } + + result, err := svc.DescribeCluster(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case eks.ErrCodeResourceNotFoundException: + fmt.Println(eks.ErrCodeResourceNotFoundException, aerr.Error()) + case eks.ErrCodeClientException: + fmt.Println(eks.ErrCodeClientException, aerr.Error()) + case eks.ErrCodeServerException: + fmt.Println(eks.ErrCodeServerException, aerr.Error()) + case eks.ErrCodeServiceUnavailableException: + fmt.Println(eks.ErrCodeServiceUnavailableException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} + +// To list your available clusters +// +// This example command lists all of your available clusters in your default region. +func ExampleEKS_ListClusters_shared00() { + svc := eks.New(session.New()) + input := &eks.ListClustersInput{} + + result, err := svc.ListClusters(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case eks.ErrCodeInvalidParameterException: + fmt.Println(eks.ErrCodeInvalidParameterException, aerr.Error()) + case eks.ErrCodeClientException: + fmt.Println(eks.ErrCodeClientException, aerr.Error()) + case eks.ErrCodeServerException: + fmt.Println(eks.ErrCodeServerException, aerr.Error()) + case eks.ErrCodeServiceUnavailableException: + fmt.Println(eks.ErrCodeServiceUnavailableException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/eks/service.go b/vendor/github.com/aws/aws-sdk-go/service/eks/service.go new file mode 100644 index 000000000..e72b2c362 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/eks/service.go @@ -0,0 +1,99 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package eks + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/signer/v4" + "github.com/aws/aws-sdk-go/private/protocol/restjson" +) + +// EKS provides the API operation methods for making requests to +// Amazon Elastic Container Service for Kubernetes. See this package's package overview docs +// for details on the service. +// +// EKS methods are safe to use concurrently. It is not safe to +// modify mutate any of the struct's properties though. +type EKS struct { + *client.Client +} + +// Used for custom client initialization logic +var initClient func(*client.Client) + +// Used for custom request initialization logic +var initRequest func(*request.Request) + +// Service information constants +const ( + ServiceName = "eks" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "EKS" // ServiceID is a unique identifer of a specific service. +) + +// New creates a new instance of the EKS client with a session. +// If additional configuration is needed for the client instance use the optional +// aws.Config parameter to add your extra config. +// +// Example: +// // Create a EKS client from just a session. +// svc := eks.New(mySession) +// +// // Create a EKS client with additional configuration +// svc := eks.New(mySession, aws.NewConfig().WithRegion("us-west-2")) +func New(p client.ConfigProvider, cfgs ...*aws.Config) *EKS { + c := p.ClientConfig(EndpointsID, cfgs...) + if c.SigningNameDerived || len(c.SigningName) == 0 { + c.SigningName = "eks" + } + return newClient(*c.Config, c.Handlers, c.Endpoint, c.SigningRegion, c.SigningName) +} + +// newClient creates, initializes and returns a new service client instance. +func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion, signingName string) *EKS { + svc := &EKS{ + Client: client.New( + cfg, + metadata.ClientInfo{ + ServiceName: ServiceName, + ServiceID: ServiceID, + SigningName: signingName, + SigningRegion: signingRegion, + Endpoint: endpoint, + APIVersion: "2017-11-01", + JSONVersion: "1.1", + }, + handlers, + ), + } + + // Handlers + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(restjson.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(restjson.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(restjson.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(restjson.UnmarshalErrorHandler) + + // Run custom client initialization if present + if initClient != nil { + initClient(svc.Client) + } + + return svc +} + +// newRequest creates a new request for a EKS operation and runs any +// custom request initialization. +func (c *EKS) newRequest(op *request.Operation, params, data interface{}) *request.Request { + req := c.NewRequest(op, params, data) + + // Run custom request initialization if present + if initRequest != nil { + initRequest(req) + } + + return req +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/elasticache/service.go b/vendor/github.com/aws/aws-sdk-go/service/elasticache/service.go index 40bed298a..fd5f8c517 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elasticache/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elasticache/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elasticache" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elasticache" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ElastiCache" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ElastiCache client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/api.go b/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/api.go index 5849124ce..d3c3004c9 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/api.go @@ -152,7 +152,7 @@ func (c *ElasticBeanstalk) ApplyEnvironmentManagedActionRequest(input *ApplyEnvi // API operation ApplyEnvironmentManagedAction for usage and error information. // // Returned Error Codes: -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // * ErrCodeManagedActionInvalidStateException "ManagedActionInvalidStateException" @@ -776,7 +776,7 @@ func (c *ElasticBeanstalk) CreatePlatformVersionRequest(input *CreatePlatformVer // The specified account does not have sufficient privileges for one or more // AWS services. // -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // * ErrCodeTooManyPlatformsException "TooManyPlatformsException" @@ -1315,7 +1315,7 @@ func (c *ElasticBeanstalk) DeletePlatformVersionRequest(input *DeletePlatformVer // The specified account does not have sufficient privileges for one or more // AWS services. // -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // * ErrCodePlatformVersionStillReferencedException "PlatformVersionStillReferencedException" @@ -1809,7 +1809,7 @@ func (c *ElasticBeanstalk) DescribeEnvironmentHealthRequest(input *DescribeEnvir // One or more input parameters is not valid. Please correct the input parameters // and try the operation again. // -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticbeanstalk-2010-12-01/DescribeEnvironmentHealth @@ -1888,7 +1888,7 @@ func (c *ElasticBeanstalk) DescribeEnvironmentManagedActionHistoryRequest(input // API operation DescribeEnvironmentManagedActionHistory for usage and error information. // // Returned Error Codes: -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticbeanstalk-2010-12-01/DescribeEnvironmentManagedActionHistory @@ -1967,7 +1967,7 @@ func (c *ElasticBeanstalk) DescribeEnvironmentManagedActionsRequest(input *Descr // API operation DescribeEnvironmentManagedActions for usage and error information. // // Returned Error Codes: -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticbeanstalk-2010-12-01/DescribeEnvironmentManagedActions @@ -2322,8 +2322,8 @@ func (c *ElasticBeanstalk) DescribeInstancesHealthRequest(input *DescribeInstanc // DescribeInstancesHealth API operation for AWS Elastic Beanstalk. // -// Retrives detailed information about the health of instances in your AWS Elastic -// Beanstalk. This operation requires enhanced health reporting (http://docs.aws.amazon.com/elasticbeanstalk/latest/dg/health-enhanced.html). +// Retrieves detailed information about the health of instances in your AWS +// Elastic Beanstalk. This operation requires enhanced health reporting (http://docs.aws.amazon.com/elasticbeanstalk/latest/dg/health-enhanced.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2337,7 +2337,7 @@ func (c *ElasticBeanstalk) DescribeInstancesHealthRequest(input *DescribeInstanc // One or more input parameters is not valid. Please correct the input parameters // and try the operation again. // -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticbeanstalk-2010-12-01/DescribeInstancesHealth @@ -2420,7 +2420,7 @@ func (c *ElasticBeanstalk) DescribePlatformVersionRequest(input *DescribePlatfor // The specified account does not have sufficient privileges for one or more // AWS services. // -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticbeanstalk-2010-12-01/DescribePlatformVersion @@ -2578,7 +2578,7 @@ func (c *ElasticBeanstalk) ListPlatformVersionsRequest(input *ListPlatformVersio // The specified account does not have sufficient privileges for one or more // AWS services. // -// * ErrCodeServiceException "ServiceException" +// * ErrCodeServiceException "ElasticBeanstalkServiceException" // A generic service exception has occurred. // // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticbeanstalk-2010-12-01/ListPlatformVersions @@ -4039,6 +4039,14 @@ type ApplicationResourceLifecycleConfig struct { _ struct{} `type:"structure"` // The ARN of an IAM service role that Elastic Beanstalk has permission to assume. + // + // The ServiceRole property is required the first time that you provide a VersionLifecycleConfig + // for the application in one of the supporting calls (CreateApplication or + // UpdateApplicationResourceLifecycle). After you provide it once, in either + // one of the calls, Elastic Beanstalk persists the Service Role with the application, + // and you don't need to specify it again in subsequent UpdateApplicationResourceLifecycle + // calls. You can, however, specify it in subsequent calls to change the Service + // Role to another value. ServiceRole *string `type:"string"` // The application version lifecycle configuration. @@ -4112,7 +4120,25 @@ type ApplicationVersionDescription struct { // S3. SourceBundle *S3Location `type:"structure"` - // The processing status of the application version. + // The processing status of the application version. Reflects the state of the + // application version during its creation. Many of the values are only applicable + // if you specified True for the Process parameter of the CreateApplicationVersion + // action. The following list describes the possible values. + // + // * Unprocessed – Application version wasn't pre-processed or validated. + // Elastic Beanstalk will validate configuration files during deployment + // of the application version to an environment. + // + // * Processing – Elastic Beanstalk is currently processing the application + // version. + // + // * Building – Application version is currently undergoing an AWS CodeBuild + // build. + // + // * Processed – Elastic Beanstalk was successfully pre-processed and validated. + // + // * Failed – Either the AWS CodeBuild build failed or configuration files + // didn't pass validation. This application version isn't usable. Status *string `type:"string" enum:"ApplicationVersionStatus"` // A unique identifier for the application version. @@ -5180,11 +5206,15 @@ type CreateApplicationVersionInput struct { // Describes this version. Description *string `type:"string"` - // Preprocesses and validates the environment manifest (env.yaml) and configuration + // Pre-processes and validates the environment manifest (env.yaml) and configuration // files (*.config files in the .ebextensions folder) in the source bundle. // Validating configuration files can identify issues prior to deploying the // application version to an environment. // + // You must turn processing on for application versions that you create using + // AWS CodeBuild or AWS CodeCommit. For application versions built from a source + // bundle in Amazon S3, processing is optional. + // // The Process option validates Elastic Beanstalk configuration files. It doesn't // validate your application's configuration files, like proxy server or Docker // configuration. @@ -5518,6 +5548,9 @@ type CreateEnvironmentInput struct { // This is an alternative to specifying a template name. If specified, AWS Elastic // Beanstalk sets the configuration values to the default values associated // with the specified solution stack. + // + // For a list of current solution stacks, see Elastic Beanstalk Supported Platforms + // (http://docs.aws.amazon.com/elasticbeanstalk/latest/dg/concepts.platforms.html). SolutionStackName *string `type:"string"` // This specifies the tags applied to resources in the environment. @@ -8063,7 +8096,11 @@ type EnvironmentTier struct { // The type of this environment tier. Type *string `type:"string"` - // The version of this environment tier. + // The version of this environment tier. When you don't set a value to it, Elastic + // Beanstalk uses the latest compatible worker tier version. + // + // This member is deprecated. Any specific version that you set may become out + // of date. We recommend leaving it unspecified. Version *string `type:"string"` } @@ -11512,6 +11549,9 @@ const ( // EnvironmentHealthStatusSevere is a EnvironmentHealthStatus enum value EnvironmentHealthStatusSevere = "Severe" + + // EnvironmentHealthStatusSuspended is a EnvironmentHealthStatus enum value + EnvironmentHealthStatusSuspended = "Suspended" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/errors.go b/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/errors.go index adb0f27e3..d13041287 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/errors.go @@ -77,10 +77,10 @@ const ( ErrCodeS3SubscriptionRequiredException = "S3SubscriptionRequiredException" // ErrCodeServiceException for service response error code - // "ServiceException". + // "ElasticBeanstalkServiceException". // // A generic service exception has occurred. - ErrCodeServiceException = "ServiceException" + ErrCodeServiceException = "ElasticBeanstalkServiceException" // ErrCodeSourceBundleDeletionException for service response error code // "SourceBundleDeletionFailure". diff --git a/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/service.go b/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/service.go index e56509610..12e8b1c81 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elasticbeanstalk/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elasticbeanstalk" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elasticbeanstalk" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Elastic Beanstalk" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ElasticBeanstalk client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/elasticsearchservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/elasticsearchservice/service.go index c4388e79b..d2f8f3827 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elasticsearchservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elasticsearchservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "es" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "es" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Elasticsearch Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ElasticsearchService client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/elastictranscoder/service.go b/vendor/github.com/aws/aws-sdk-go/service/elastictranscoder/service.go index 29d15e2e7..30acb8d1b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elastictranscoder/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elastictranscoder/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elastictranscoder" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elastictranscoder" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Elastic Transcoder" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ElasticTranscoder client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/elb/service.go b/vendor/github.com/aws/aws-sdk-go/service/elb/service.go index 057530f6c..5dfdd322c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elb/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elb/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elasticloadbalancing" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elasticloadbalancing" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Elastic Load Balancing" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ELB client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/elbv2/api.go b/vendor/github.com/aws/aws-sdk-go/service/elbv2/api.go index 5a60ca656..228c6a56c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elbv2/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elbv2/api.go @@ -306,6 +306,12 @@ func (c *ELBV2) CreateListenerRequest(input *CreateListenerInput) (req *request. // * ErrCodeTooManyTargetsException "TooManyTargets" // You've reached the limit on the number of targets. // +// * ErrCodeTooManyActionsException "TooManyActions" +// You've reached the limit on the number of actions per rule. +// +// * ErrCodeInvalidLoadBalancerActionException "InvalidLoadBalancerAction" +// The requested action is not valid. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticloadbalancingv2-2015-12-01/CreateListener func (c *ELBV2) CreateListener(input *CreateListenerInput) (*CreateListenerOutput, error) { req, out := c.CreateListenerRequest(input) @@ -512,9 +518,9 @@ func (c *ELBV2) CreateRuleRequest(input *CreateRuleInput) (req *request.Request, // with an Application Load Balancer. // // Rules are evaluated in priority order, from the lowest value to the highest -// value. When the condition for a rule is met, the specified action is taken. -// If no conditions are met, the action for the default rule is taken. For more -// information, see Listener Rules (http://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-listeners.html#listener-rules) +// value. When the conditions for a rule are met, its actions are performed. +// If the conditions for no rules are met, the actions for the default rule +// are performed. For more information, see Listener Rules (http://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-listeners.html#listener-rules) // in the Application Load Balancers Guide. // // To view your current rules, use DescribeRules. To update a rule, use ModifyRule. @@ -560,6 +566,15 @@ func (c *ELBV2) CreateRuleRequest(input *CreateRuleInput) (req *request.Request, // * ErrCodeTooManyTargetsException "TooManyTargets" // You've reached the limit on the number of targets. // +// * ErrCodeUnsupportedProtocolException "UnsupportedProtocol" +// The specified protocol is not supported. +// +// * ErrCodeTooManyActionsException "TooManyActions" +// You've reached the limit on the number of actions per rule. +// +// * ErrCodeInvalidLoadBalancerActionException "InvalidLoadBalancerAction" +// The requested action is not valid. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticloadbalancingv2-2015-12-01/CreateRule func (c *ELBV2) CreateRule(input *CreateRuleInput) (*CreateRuleOutput, error) { req, out := c.CreateRuleRequest(input) @@ -1338,6 +1353,9 @@ func (c *ELBV2) DescribeListenersRequest(input *DescribeListenersInput) (req *re // * ErrCodeLoadBalancerNotFoundException "LoadBalancerNotFound" // The specified load balancer does not exist. // +// * ErrCodeUnsupportedProtocolException "UnsupportedProtocol" +// The specified protocol is not supported. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticloadbalancingv2-2015-12-01/DescribeListeners func (c *ELBV2) DescribeListeners(input *DescribeListenersInput) (*DescribeListenersOutput, error) { req, out := c.DescribeListenersRequest(input) @@ -1457,6 +1475,10 @@ func (c *ELBV2) DescribeLoadBalancerAttributesRequest(input *DescribeLoadBalance // Describes the attributes for the specified Application Load Balancer or Network // Load Balancer. // +// For more information, see Load Balancer Attributes (http://docs.aws.amazon.com/elasticloadbalancing/latest/application/application-load-balancers.html#load-balancer-attributes) +// in the Application Load Balancers Guide or Load Balancer Attributes (http://docs.aws.amazon.com/elasticloadbalancing/latest/network/network-load-balancers.html#load-balancer-attributes) +// in the Network Load Balancers Guide. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -1689,6 +1711,9 @@ func (c *ELBV2) DescribeRulesRequest(input *DescribeRulesInput) (req *request.Re // * ErrCodeRuleNotFoundException "RuleNotFound" // The specified rule does not exist. // +// * ErrCodeUnsupportedProtocolException "UnsupportedProtocol" +// The specified protocol is not supported. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticloadbalancingv2-2015-12-01/DescribeRules func (c *ELBV2) DescribeRules(input *DescribeRulesInput) (*DescribeRulesOutput, error) { req, out := c.DescribeRulesRequest(input) @@ -1929,6 +1954,10 @@ func (c *ELBV2) DescribeTargetGroupAttributesRequest(input *DescribeTargetGroupA // // Describes the attributes for the specified target group. // +// For more information, see Target Group Attributes (http://docs.aws.amazon.com/elasticloadbalancing/latest/application/load-balancer-target-groups.html#target-group-attributes) +// in the Application Load Balancers Guide or Target Group Attributes (http://docs.aws.amazon.com/elasticloadbalancing/latest/network/load-balancer-target-groups.html#target-group-attributes) +// in the Network Load Balancers Guide. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -2292,6 +2321,12 @@ func (c *ELBV2) ModifyListenerRequest(input *ModifyListenerInput) (req *request. // * ErrCodeTooManyTargetsException "TooManyTargets" // You've reached the limit on the number of targets. // +// * ErrCodeTooManyActionsException "TooManyActions" +// You've reached the limit on the number of actions per rule. +// +// * ErrCodeInvalidLoadBalancerActionException "InvalidLoadBalancerAction" +// The requested action is not valid. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticloadbalancingv2-2015-12-01/ModifyListener func (c *ELBV2) ModifyListener(input *ModifyListenerInput) (*ModifyListenerOutput, error) { req, out := c.ModifyListenerRequest(input) @@ -2449,7 +2484,7 @@ func (c *ELBV2) ModifyRuleRequest(input *ModifyRuleInput) (req *request.Request, // // Any existing properties that you do not modify retain their current values. // -// To modify the default action, use ModifyListener. +// To modify the actions for the default rule, use ModifyListener. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2481,6 +2516,15 @@ func (c *ELBV2) ModifyRuleRequest(input *ModifyRuleInput) (req *request.Request, // * ErrCodeTargetGroupNotFoundException "TargetGroupNotFound" // The specified target group does not exist. // +// * ErrCodeUnsupportedProtocolException "UnsupportedProtocol" +// The specified protocol is not supported. +// +// * ErrCodeTooManyActionsException "TooManyActions" +// You've reached the limit on the number of actions per rule. +// +// * ErrCodeInvalidLoadBalancerActionException "InvalidLoadBalancerAction" +// The requested action is not valid. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/elasticloadbalancingv2-2015-12-01/ModifyRule func (c *ELBV2) ModifyRule(input *ModifyRuleInput) (*ModifyRuleOutput, error) { req, out := c.ModifyRuleRequest(input) @@ -3323,12 +3367,27 @@ func (c *ELBV2) SetSubnetsWithContext(ctx aws.Context, input *SetSubnetsInput, o type Action struct { _ struct{} `type:"structure"` - // The Amazon Resource Name (ARN) of the target group. - // - // TargetGroupArn is a required field - TargetGroupArn *string `type:"string" required:"true"` + // [HTTPS listener] Information for using Amazon Cognito to authenticate users. + // Specify only when Type is authenticate-cognito. + AuthenticateCognitoConfig *AuthenticateCognitoActionConfig `type:"structure"` - // The type of action. + // [HTTPS listener] Information about an identity provider that is compliant + // with OpenID Connect (OIDC). Specify only when Type is authenticate-oidc. + AuthenticateOidcConfig *AuthenticateOidcActionConfig `type:"structure"` + + // The order for the action. This value is required for rules with multiple + // actions. The action with the lowest value for order is performed first. The + // forward action must be performed last. + Order *int64 `min:"1" type:"integer"` + + // The Amazon Resource Name (ARN) of the target group. Specify only when Type + // is forward. + // + // For a default rule, the protocol of the target group must be HTTP or HTTPS + // for an Application Load Balancer or TCP for a Network Load Balancer. + TargetGroupArn *string `type:"string"` + + // The type of action. Each rule must include one forward action. // // Type is a required field Type *string `type:"string" required:"true" enum:"ActionTypeEnum"` @@ -3347,12 +3406,22 @@ func (s Action) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *Action) Validate() error { invalidParams := request.ErrInvalidParams{Context: "Action"} - if s.TargetGroupArn == nil { - invalidParams.Add(request.NewErrParamRequired("TargetGroupArn")) + if s.Order != nil && *s.Order < 1 { + invalidParams.Add(request.NewErrParamMinValue("Order", 1)) } if s.Type == nil { invalidParams.Add(request.NewErrParamRequired("Type")) } + if s.AuthenticateCognitoConfig != nil { + if err := s.AuthenticateCognitoConfig.Validate(); err != nil { + invalidParams.AddNested("AuthenticateCognitoConfig", err.(request.ErrInvalidParams)) + } + } + if s.AuthenticateOidcConfig != nil { + if err := s.AuthenticateOidcConfig.Validate(); err != nil { + invalidParams.AddNested("AuthenticateOidcConfig", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -3360,6 +3429,24 @@ func (s *Action) Validate() error { return nil } +// SetAuthenticateCognitoConfig sets the AuthenticateCognitoConfig field's value. +func (s *Action) SetAuthenticateCognitoConfig(v *AuthenticateCognitoActionConfig) *Action { + s.AuthenticateCognitoConfig = v + return s +} + +// SetAuthenticateOidcConfig sets the AuthenticateOidcConfig field's value. +func (s *Action) SetAuthenticateOidcConfig(v *AuthenticateOidcActionConfig) *Action { + s.AuthenticateOidcConfig = v + return s +} + +// SetOrder sets the Order field's value. +func (s *Action) SetOrder(v int64) *Action { + s.Order = &v + return s +} + // SetTargetGroupArn sets the TargetGroupArn field's value. func (s *Action) SetTargetGroupArn(v string) *Action { s.TargetGroupArn = &v @@ -3526,6 +3613,305 @@ func (s AddTagsOutput) GoString() string { return s.String() } +// Request parameters to use when integrating with Amazon Cognito to authenticate +// users. +type AuthenticateCognitoActionConfig struct { + _ struct{} `type:"structure"` + + // The query parameters (up to 10) to include in the redirect request to the + // authorization endpoint. + AuthenticationRequestExtraParams map[string]*string `type:"map"` + + // The behavior if the user is not authenticated. The following are possible + // values: + // + // * deny - Return an HTTP 401 Unauthorized error. + // + // * allow - Allow the request to be forwarded to the target. + // + // authenticate + OnUnauthenticatedRequest *string `type:"string" enum:"AuthenticateCognitoActionConditionalBehaviorEnum"` + + // The set of user claims to be requested from the IdP. The default is openid. + // + // To verify which scope values your IdP supports and how to separate multiple + // values, see the documentation for your IdP. + Scope *string `type:"string"` + + // The name of the cookie used to maintain session information. The default + // is AWSELBAuthSessionCookie. + SessionCookieName *string `type:"string"` + + // The maximum duration of the authentication session, in seconds. The default + // is 604800 seconds (7 days). + SessionTimeout *int64 `type:"long"` + + // The Amazon Resource Name (ARN) of the Amazon Cognito user pool. + // + // UserPoolArn is a required field + UserPoolArn *string `type:"string" required:"true"` + + // The ID of the Amazon Cognito user pool client. + // + // UserPoolClientId is a required field + UserPoolClientId *string `type:"string" required:"true"` + + // The domain prefix or fully-qualified domain name of the Amazon Cognito user + // pool. + // + // UserPoolDomain is a required field + UserPoolDomain *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s AuthenticateCognitoActionConfig) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AuthenticateCognitoActionConfig) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AuthenticateCognitoActionConfig) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AuthenticateCognitoActionConfig"} + if s.UserPoolArn == nil { + invalidParams.Add(request.NewErrParamRequired("UserPoolArn")) + } + if s.UserPoolClientId == nil { + invalidParams.Add(request.NewErrParamRequired("UserPoolClientId")) + } + if s.UserPoolDomain == nil { + invalidParams.Add(request.NewErrParamRequired("UserPoolDomain")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAuthenticationRequestExtraParams sets the AuthenticationRequestExtraParams field's value. +func (s *AuthenticateCognitoActionConfig) SetAuthenticationRequestExtraParams(v map[string]*string) *AuthenticateCognitoActionConfig { + s.AuthenticationRequestExtraParams = v + return s +} + +// SetOnUnauthenticatedRequest sets the OnUnauthenticatedRequest field's value. +func (s *AuthenticateCognitoActionConfig) SetOnUnauthenticatedRequest(v string) *AuthenticateCognitoActionConfig { + s.OnUnauthenticatedRequest = &v + return s +} + +// SetScope sets the Scope field's value. +func (s *AuthenticateCognitoActionConfig) SetScope(v string) *AuthenticateCognitoActionConfig { + s.Scope = &v + return s +} + +// SetSessionCookieName sets the SessionCookieName field's value. +func (s *AuthenticateCognitoActionConfig) SetSessionCookieName(v string) *AuthenticateCognitoActionConfig { + s.SessionCookieName = &v + return s +} + +// SetSessionTimeout sets the SessionTimeout field's value. +func (s *AuthenticateCognitoActionConfig) SetSessionTimeout(v int64) *AuthenticateCognitoActionConfig { + s.SessionTimeout = &v + return s +} + +// SetUserPoolArn sets the UserPoolArn field's value. +func (s *AuthenticateCognitoActionConfig) SetUserPoolArn(v string) *AuthenticateCognitoActionConfig { + s.UserPoolArn = &v + return s +} + +// SetUserPoolClientId sets the UserPoolClientId field's value. +func (s *AuthenticateCognitoActionConfig) SetUserPoolClientId(v string) *AuthenticateCognitoActionConfig { + s.UserPoolClientId = &v + return s +} + +// SetUserPoolDomain sets the UserPoolDomain field's value. +func (s *AuthenticateCognitoActionConfig) SetUserPoolDomain(v string) *AuthenticateCognitoActionConfig { + s.UserPoolDomain = &v + return s +} + +// Request parameters when using an identity provider (IdP) that is compliant +// with OpenID Connect (OIDC) to authenticate users. +type AuthenticateOidcActionConfig struct { + _ struct{} `type:"structure"` + + // The query parameters (up to 10) to include in the redirect request to the + // authorization endpoint. + AuthenticationRequestExtraParams map[string]*string `type:"map"` + + // The authorization endpoint of the IdP. This must be a full URL, including + // the HTTPS protocol, the domain, and the path. + // + // AuthorizationEndpoint is a required field + AuthorizationEndpoint *string `type:"string" required:"true"` + + // The OAuth 2.0 client identifier. + // + // ClientId is a required field + ClientId *string `type:"string" required:"true"` + + // The OAuth 2.0 client secret. + // + // ClientSecret is a required field + ClientSecret *string `type:"string" required:"true"` + + // The OIDC issuer identifier of the IdP. This must be a full URL, including + // the HTTPS protocol, the domain, and the path. + // + // Issuer is a required field + Issuer *string `type:"string" required:"true"` + + // The behavior if the user is not authenticated. The following are possible + // values: + // + // * deny - Return an HTTP 401 Unauthorized error. + // + // * allow - Allow the request to be forwarded to the target. + // + // authenticate + OnUnauthenticatedRequest *string `type:"string" enum:"AuthenticateOidcActionConditionalBehaviorEnum"` + + // The set of user claims to be requested from the IdP. The default is openid. + // + // To verify which scope values your IdP supports and how to separate multiple + // values, see the documentation for your IdP. + Scope *string `type:"string"` + + // The name of the cookie used to maintain session information. The default + // is AWSELBAuthSessionCookie. + SessionCookieName *string `type:"string"` + + // The maximum duration of the authentication session, in seconds. The default + // is 604800 seconds (7 days). + SessionTimeout *int64 `type:"long"` + + // The token endpoint of the IdP. This must be a full URL, including the HTTPS + // protocol, the domain, and the path. + // + // TokenEndpoint is a required field + TokenEndpoint *string `type:"string" required:"true"` + + // The user info endpoint of the IdP. This must be a full URL, including the + // HTTPS protocol, the domain, and the path. + // + // UserInfoEndpoint is a required field + UserInfoEndpoint *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s AuthenticateOidcActionConfig) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AuthenticateOidcActionConfig) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AuthenticateOidcActionConfig) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AuthenticateOidcActionConfig"} + if s.AuthorizationEndpoint == nil { + invalidParams.Add(request.NewErrParamRequired("AuthorizationEndpoint")) + } + if s.ClientId == nil { + invalidParams.Add(request.NewErrParamRequired("ClientId")) + } + if s.ClientSecret == nil { + invalidParams.Add(request.NewErrParamRequired("ClientSecret")) + } + if s.Issuer == nil { + invalidParams.Add(request.NewErrParamRequired("Issuer")) + } + if s.TokenEndpoint == nil { + invalidParams.Add(request.NewErrParamRequired("TokenEndpoint")) + } + if s.UserInfoEndpoint == nil { + invalidParams.Add(request.NewErrParamRequired("UserInfoEndpoint")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAuthenticationRequestExtraParams sets the AuthenticationRequestExtraParams field's value. +func (s *AuthenticateOidcActionConfig) SetAuthenticationRequestExtraParams(v map[string]*string) *AuthenticateOidcActionConfig { + s.AuthenticationRequestExtraParams = v + return s +} + +// SetAuthorizationEndpoint sets the AuthorizationEndpoint field's value. +func (s *AuthenticateOidcActionConfig) SetAuthorizationEndpoint(v string) *AuthenticateOidcActionConfig { + s.AuthorizationEndpoint = &v + return s +} + +// SetClientId sets the ClientId field's value. +func (s *AuthenticateOidcActionConfig) SetClientId(v string) *AuthenticateOidcActionConfig { + s.ClientId = &v + return s +} + +// SetClientSecret sets the ClientSecret field's value. +func (s *AuthenticateOidcActionConfig) SetClientSecret(v string) *AuthenticateOidcActionConfig { + s.ClientSecret = &v + return s +} + +// SetIssuer sets the Issuer field's value. +func (s *AuthenticateOidcActionConfig) SetIssuer(v string) *AuthenticateOidcActionConfig { + s.Issuer = &v + return s +} + +// SetOnUnauthenticatedRequest sets the OnUnauthenticatedRequest field's value. +func (s *AuthenticateOidcActionConfig) SetOnUnauthenticatedRequest(v string) *AuthenticateOidcActionConfig { + s.OnUnauthenticatedRequest = &v + return s +} + +// SetScope sets the Scope field's value. +func (s *AuthenticateOidcActionConfig) SetScope(v string) *AuthenticateOidcActionConfig { + s.Scope = &v + return s +} + +// SetSessionCookieName sets the SessionCookieName field's value. +func (s *AuthenticateOidcActionConfig) SetSessionCookieName(v string) *AuthenticateOidcActionConfig { + s.SessionCookieName = &v + return s +} + +// SetSessionTimeout sets the SessionTimeout field's value. +func (s *AuthenticateOidcActionConfig) SetSessionTimeout(v int64) *AuthenticateOidcActionConfig { + s.SessionTimeout = &v + return s +} + +// SetTokenEndpoint sets the TokenEndpoint field's value. +func (s *AuthenticateOidcActionConfig) SetTokenEndpoint(v string) *AuthenticateOidcActionConfig { + s.TokenEndpoint = &v + return s +} + +// SetUserInfoEndpoint sets the UserInfoEndpoint field's value. +func (s *AuthenticateOidcActionConfig) SetUserInfoEndpoint(v string) *AuthenticateOidcActionConfig { + s.UserInfoEndpoint = &v + return s +} + // Information about an Availability Zone. type AvailabilityZone struct { _ struct{} `type:"structure"` @@ -3637,13 +4023,22 @@ func (s *Cipher) SetPriority(v int64) *Cipher { type CreateListenerInput struct { _ struct{} `type:"structure"` - // [HTTPS listeners] The SSL server certificate. You must provide exactly one - // certificate. + // [HTTPS listeners] The default SSL server certificate. You must provide exactly + // one certificate. To create a certificate list, use AddListenerCertificates. Certificates []*Certificate `type:"list"` - // The default action for the listener. For Application Load Balancers, the - // protocol of the specified target group must be HTTP or HTTPS. For Network - // Load Balancers, the protocol of the specified target group must be TCP. + // The actions for the default rule. The rule must include one forward action. + // + // If the action type is forward, you can specify a single target group. The + // protocol of the target group must be HTTP or HTTPS for an Application Load + // Balancer or TCP for a Network Load Balancer. + // + // If the action type is authenticate-oidc, you can use an identity provider + // that is OpenID Connect (OIDC) compliant to authenticate users as they access + // your application. + // + // If the action type is authenticate-cognito, you can use Amazon Cognito to + // authenticate users as they access your application. // // DefaultActions is a required field DefaultActions []*Action `type:"list" required:"true"` @@ -3786,8 +4181,8 @@ type CreateLoadBalancerInput struct { // The name of the load balancer. // // This name must be unique per region per account, can have a maximum of 32 - // characters, must contain only alphanumeric characters or hyphens, and must - // not begin or end with a hyphen. + // characters, must contain only alphanumeric characters or hyphens, must not + // begin or end with a hyphen, and must not begin with "internal-". // // Name is a required field Name *string `type:"string" required:"true"` @@ -3946,7 +4341,16 @@ func (s *CreateLoadBalancerOutput) SetLoadBalancers(v []*LoadBalancer) *CreateLo type CreateRuleInput struct { _ struct{} `type:"structure"` - // An action. Each action has the type forward and specifies a target group. + // The actions. Each rule must include one forward action. + // + // If the action type is forward, you can specify a single target group. + // + // If the action type is authenticate-oidc, you can use an identity provider + // that is OpenID Connect (OIDC) compliant to authenticate users as they access + // your application. + // + // If the action type is authenticate-cognito, you can use Amazon Cognito to + // authenticate users as they access your application. // // Actions is a required field Actions []*Action `type:"list" required:"true"` @@ -3989,8 +4393,7 @@ type CreateRuleInput struct { // ListenerArn is a required field ListenerArn *string `type:"string" required:"true"` - // The priority for the rule. A listener can't have multiple rules with the - // same priority. + // The rule priority. A listener can't have multiple rules with the same priority. // // Priority is a required field Priority *int64 `min:"1" type:"integer" required:"true"` @@ -5839,32 +6242,35 @@ type LoadBalancerAttribute struct { // The name of the attribute. // - // * access_logs.s3.enabled - [Application Load Balancers] Indicates whether - // access logs stored in Amazon S3 are enabled. The value is true or false. - // - // * access_logs.s3.bucket - [Application Load Balancers] The name of the - // S3 bucket for the access logs. This attribute is required if access logs - // in Amazon S3 are enabled. The bucket must exist in the same region as - // the load balancer and have a bucket policy that grants Elastic Load Balancing - // permission to write to the bucket. - // - // * access_logs.s3.prefix - [Application Load Balancers] The prefix for - // the location in the S3 bucket. If you don't specify a prefix, the access - // logs are stored in the root of the bucket. + // The following attributes are supported by both Application Load Balancers + // and Network Load Balancers: // // * deletion_protection.enabled - Indicates whether deletion protection - // is enabled. The value is true or false. + // is enabled. The value is true or false. The default is false. // - // * idle_timeout.timeout_seconds - [Application Load Balancers] The idle - // timeout value, in seconds. The valid range is 1-4000. The default is 60 - // seconds. + // The following attributes are supported by only Application Load Balancers: // - // * load_balancing.cross_zone.enabled - [Network Load Balancers] Indicates - // whether cross-zone load balancing is enabled. The value is true or false. - // The default is false. + // * access_logs.s3.enabled - Indicates whether access logs are enabled. + // The value is true or false. The default is false. // - // * routing.http2.enabled - [Application Load Balancers] Indicates whether - // HTTP/2 is enabled. The value is true or false. The default is true. + // * access_logs.s3.bucket - The name of the S3 bucket for the access logs. + // This attribute is required if access logs are enabled. The bucket must + // exist in the same region as the load balancer and have a bucket policy + // that grants Elastic Load Balancing permission to write to the bucket. + // + // * access_logs.s3.prefix - The prefix for the location in the S3 bucket + // for the access logs. + // + // * idle_timeout.timeout_seconds - The idle timeout value, in seconds. The + // valid range is 1-4000 seconds. The default is 60 seconds. + // + // * routing.http2.enabled - Indicates whether HTTP/2 is enabled. The value + // is true or false. The default is true. + // + // The following attributes are supported by only Network Load Balancers: + // + // * load_balancing.cross_zone.enabled - Indicates whether cross-zone load + // balancing is enabled. The value is true or false. The default is false. Key *string `type:"string"` // The value of the attribute. @@ -5976,12 +6382,22 @@ func (s *Matcher) SetHttpCode(v string) *Matcher { type ModifyListenerInput struct { _ struct{} `type:"structure"` - // The default SSL server certificate. + // [HTTPS listeners] The default SSL server certificate. You must provide exactly + // one certificate. To create a certificate list, use AddListenerCertificates. Certificates []*Certificate `type:"list"` - // The default action. For Application Load Balancers, the protocol of the specified - // target group must be HTTP or HTTPS. For Network Load Balancers, the protocol - // of the specified target group must be TCP. + // The actions for the default rule. The rule must include one forward action. + // + // If the action type is forward, you can specify a single target group. The + // protocol of the target group must be HTTP or HTTPS for an Application Load + // Balancer or TCP for a Network Load Balancer. + // + // If the action type is authenticate-oidc, you can use an identity provider + // that is OpenID Connect (OIDC) compliant to authenticate users as they access + // your application. + // + // If the action type is authenticate-cognito, you can use Amazon Cognito to + // authenticate users as they access your application. DefaultActions []*Action `type:"list"` // The Amazon Resource Name (ARN) of the listener. @@ -5997,8 +6413,8 @@ type ModifyListenerInput struct { // TCP. Protocol *string `type:"string" enum:"ProtocolEnum"` - // The security policy that defines which protocols and ciphers are supported. - // For more information, see Security Policies (http://docs.aws.amazon.com/elasticloadbalancing/latest/application/create-https-listener.html#describe-ssl-policies) + // [HTTPS listeners] The security policy that defines which protocols and ciphers + // are supported. For more information, see Security Policies (http://docs.aws.amazon.com/elasticloadbalancing/latest/application/create-https-listener.html#describe-ssl-policies) // in the Application Load Balancers Guide. SslPolicy *string `type:"string"` } @@ -6078,7 +6494,7 @@ func (s *ModifyListenerInput) SetSslPolicy(v string) *ModifyListenerInput { type ModifyListenerOutput struct { _ struct{} `type:"structure"` - // Information about the modified listeners. + // Information about the modified listener. Listeners []*Listener `type:"list"` } @@ -6176,10 +6592,47 @@ func (s *ModifyLoadBalancerAttributesOutput) SetAttributes(v []*LoadBalancerAttr type ModifyRuleInput struct { _ struct{} `type:"structure"` - // The actions. The target group must use the HTTP or HTTPS protocol. + // The actions. + // + // If the action type is forward, you can specify a single target group. + // + // If the action type is authenticate-oidc, you can use an identity provider + // that is OpenID Connect (OIDC) compliant to authenticate users as they access + // your application. + // + // If the action type is authenticate-cognito, you can use Amazon Cognito to + // authenticate users as they access your application. Actions []*Action `type:"list"` - // The conditions. + // The conditions. Each condition specifies a field name and a single value. + // + // If the field name is host-header, you can specify a single host name (for + // example, my.example.com). A host name is case insensitive, can be up to 128 + // characters in length, and can contain any of the following characters. Note + // that you can include up to three wildcard characters. + // + // * A-Z, a-z, 0-9 + // + // * - . + // + // * * (matches 0 or more characters) + // + // * ? (matches exactly 1 character) + // + // If the field name is path-pattern, you can specify a single path pattern. + // A path pattern is case sensitive, can be up to 128 characters in length, + // and can contain any of the following characters. Note that you can include + // up to three wildcard characters. + // + // * A-Z, a-z, 0-9 + // + // * _ - . $ / ~ " ' @ : + + // + // * & (using &) + // + // * * (matches 0 or more characters) + // + // * ? (matches exactly 1 character) Conditions []*RuleCondition `type:"list"` // The Amazon Resource Name (ARN) of the rule. @@ -6242,7 +6695,7 @@ func (s *ModifyRuleInput) SetRuleArn(v string) *ModifyRuleInput { type ModifyRuleOutput struct { _ struct{} `type:"structure"` - // Information about the rule. + // Information about the modified rule. Rules []*Rule `type:"list"` } @@ -6480,7 +6933,7 @@ func (s *ModifyTargetGroupInput) SetUnhealthyThresholdCount(v int64) *ModifyTarg type ModifyTargetGroupOutput struct { _ struct{} `type:"structure"` - // Information about the target group. + // Information about the modified target group. TargetGroups []*TargetGroup `type:"list"` } @@ -7116,9 +7569,7 @@ type SetSubnetsInput struct { // The IDs of the public subnets. You must specify subnets from at least two // Availability Zones. You can specify only one subnet per Availability Zone. // You must specify either subnets or subnet mappings. - // - // Subnets is a required field - Subnets []*string `type:"list" required:"true"` + Subnets []*string `type:"list"` } // String returns the string representation @@ -7137,9 +7588,6 @@ func (s *SetSubnetsInput) Validate() error { if s.LoadBalancerArn == nil { invalidParams.Add(request.NewErrParamRequired("LoadBalancerArn")) } - if s.Subnets == nil { - invalidParams.Add(request.NewErrParamRequired("Subnets")) - } if invalidParams.Len() > 0 { return invalidParams @@ -7581,25 +8029,38 @@ type TargetGroupAttribute struct { // The name of the attribute. // - // * deregistration_delay.timeout_seconds - The amount time for Elastic Load - // Balancing to wait before changing the state of a deregistering target - // from draining to unused. The range is 0-3600 seconds. The default value - // is 300 seconds. + // The following attributes are supported by both Application Load Balancers + // and Network Load Balancers: // - // * proxy_protocol_v2.enabled - [Network Load Balancers] Indicates whether - // Proxy Protocol version 2 is enabled. + // * deregistration_delay.timeout_seconds - The amount of time, in seconds, + // for Elastic Load Balancing to wait before changing the state of a deregistering + // target from draining to unused. The range is 0-3600 seconds. The default + // value is 300 seconds. // - // * stickiness.enabled - [Application Load Balancers] Indicates whether - // sticky sessions are enabled. The value is true or false. + // The following attributes are supported by only Application Load Balancers: // - // * stickiness.type - [Application Load Balancers] The type of sticky sessions. - // The possible value is lb_cookie. + // * slow_start.duration_seconds - The time period, in seconds, during which + // a newly registered target receives a linearly increasing share of the + // traffic to the target group. After this time period ends, the target receives + // its full share of traffic. The range is 30-900 seconds (15 minutes). Slow + // start mode is disabled by default. // - // * stickiness.lb_cookie.duration_seconds - [Application Load Balancers] - // The time period, in seconds, during which requests from a client should - // be routed to the same target. After this time period expires, the load - // balancer-generated cookie is considered stale. The range is 1 second to - // 1 week (604800 seconds). The default value is 1 day (86400 seconds). + // * stickiness.enabled - Indicates whether sticky sessions are enabled. + // The value is true or false. The default is false. + // + // * stickiness.type - The type of sticky sessions. The possible value is + // lb_cookie. + // + // * stickiness.lb_cookie.duration_seconds - The time period, in seconds, + // during which requests from a client should be routed to the same target. + // After this time period expires, the load balancer-generated cookie is + // considered stale. The range is 1 second to 1 week (604800 seconds). The + // default value is 1 day (86400 seconds). + // + // The following attributes are supported by only Network Load Balancers: + // + // * proxy_protocol_v2.enabled - Indicates whether Proxy Protocol version + // 2 is enabled. The value is true or false. The default is false. Key *string `type:"string"` // The value of the attribute. @@ -7759,6 +8220,34 @@ func (s *TargetHealthDescription) SetTargetHealth(v *TargetHealth) *TargetHealth const ( // ActionTypeEnumForward is a ActionTypeEnum enum value ActionTypeEnumForward = "forward" + + // ActionTypeEnumAuthenticateOidc is a ActionTypeEnum enum value + ActionTypeEnumAuthenticateOidc = "authenticate-oidc" + + // ActionTypeEnumAuthenticateCognito is a ActionTypeEnum enum value + ActionTypeEnumAuthenticateCognito = "authenticate-cognito" +) + +const ( + // AuthenticateCognitoActionConditionalBehaviorEnumDeny is a AuthenticateCognitoActionConditionalBehaviorEnum enum value + AuthenticateCognitoActionConditionalBehaviorEnumDeny = "deny" + + // AuthenticateCognitoActionConditionalBehaviorEnumAllow is a AuthenticateCognitoActionConditionalBehaviorEnum enum value + AuthenticateCognitoActionConditionalBehaviorEnumAllow = "allow" + + // AuthenticateCognitoActionConditionalBehaviorEnumAuthenticate is a AuthenticateCognitoActionConditionalBehaviorEnum enum value + AuthenticateCognitoActionConditionalBehaviorEnumAuthenticate = "authenticate" +) + +const ( + // AuthenticateOidcActionConditionalBehaviorEnumDeny is a AuthenticateOidcActionConditionalBehaviorEnum enum value + AuthenticateOidcActionConditionalBehaviorEnumDeny = "deny" + + // AuthenticateOidcActionConditionalBehaviorEnumAllow is a AuthenticateOidcActionConditionalBehaviorEnum enum value + AuthenticateOidcActionConditionalBehaviorEnumAllow = "allow" + + // AuthenticateOidcActionConditionalBehaviorEnumAuthenticate is a AuthenticateOidcActionConditionalBehaviorEnum enum value + AuthenticateOidcActionConditionalBehaviorEnumAuthenticate = "authenticate" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/elbv2/errors.go b/vendor/github.com/aws/aws-sdk-go/service/elbv2/errors.go index 88edc02e9..b813ebeff 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elbv2/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elbv2/errors.go @@ -65,6 +65,12 @@ const ( // The requested configuration is not valid. ErrCodeInvalidConfigurationRequestException = "InvalidConfigurationRequest" + // ErrCodeInvalidLoadBalancerActionException for service response error code + // "InvalidLoadBalancerAction". + // + // The requested action is not valid. + ErrCodeInvalidLoadBalancerActionException = "InvalidLoadBalancerAction" + // ErrCodeInvalidSchemeException for service response error code // "InvalidScheme". // @@ -150,6 +156,12 @@ const ( // The specified target group does not exist. ErrCodeTargetGroupNotFoundException = "TargetGroupNotFound" + // ErrCodeTooManyActionsException for service response error code + // "TooManyActions". + // + // You've reached the limit on the number of actions per rule. + ErrCodeTooManyActionsException = "TooManyActions" + // ErrCodeTooManyCertificatesException for service response error code // "TooManyCertificates". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/elbv2/examples_test.go b/vendor/github.com/aws/aws-sdk-go/service/elbv2/examples_test.go index 531022042..4810b427c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elbv2/examples_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elbv2/examples_test.go @@ -120,6 +120,10 @@ func ExampleELBV2_CreateListener_shared00() { fmt.Println(elbv2.ErrCodeTooManyRegistrationsForTargetIdException, aerr.Error()) case elbv2.ErrCodeTooManyTargetsException: fmt.Println(elbv2.ErrCodeTooManyTargetsException, aerr.Error()) + case elbv2.ErrCodeTooManyActionsException: + fmt.Println(elbv2.ErrCodeTooManyActionsException, aerr.Error()) + case elbv2.ErrCodeInvalidLoadBalancerActionException: + fmt.Println(elbv2.ErrCodeInvalidLoadBalancerActionException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -192,6 +196,10 @@ func ExampleELBV2_CreateListener_shared01() { fmt.Println(elbv2.ErrCodeTooManyRegistrationsForTargetIdException, aerr.Error()) case elbv2.ErrCodeTooManyTargetsException: fmt.Println(elbv2.ErrCodeTooManyTargetsException, aerr.Error()) + case elbv2.ErrCodeTooManyActionsException: + fmt.Println(elbv2.ErrCodeTooManyActionsException, aerr.Error()) + case elbv2.ErrCodeInvalidLoadBalancerActionException: + fmt.Println(elbv2.ErrCodeInvalidLoadBalancerActionException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -372,6 +380,12 @@ func ExampleELBV2_CreateRule_shared00() { fmt.Println(elbv2.ErrCodeTooManyRegistrationsForTargetIdException, aerr.Error()) case elbv2.ErrCodeTooManyTargetsException: fmt.Println(elbv2.ErrCodeTooManyTargetsException, aerr.Error()) + case elbv2.ErrCodeUnsupportedProtocolException: + fmt.Println(elbv2.ErrCodeUnsupportedProtocolException, aerr.Error()) + case elbv2.ErrCodeTooManyActionsException: + fmt.Println(elbv2.ErrCodeTooManyActionsException, aerr.Error()) + case elbv2.ErrCodeInvalidLoadBalancerActionException: + fmt.Println(elbv2.ErrCodeInvalidLoadBalancerActionException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -600,6 +614,8 @@ func ExampleELBV2_DescribeListeners_shared00() { fmt.Println(elbv2.ErrCodeListenerNotFoundException, aerr.Error()) case elbv2.ErrCodeLoadBalancerNotFoundException: fmt.Println(elbv2.ErrCodeLoadBalancerNotFoundException, aerr.Error()) + case elbv2.ErrCodeUnsupportedProtocolException: + fmt.Println(elbv2.ErrCodeUnsupportedProtocolException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -693,6 +709,8 @@ func ExampleELBV2_DescribeRules_shared00() { fmt.Println(elbv2.ErrCodeListenerNotFoundException, aerr.Error()) case elbv2.ErrCodeRuleNotFoundException: fmt.Println(elbv2.ErrCodeRuleNotFoundException, aerr.Error()) + case elbv2.ErrCodeUnsupportedProtocolException: + fmt.Println(elbv2.ErrCodeUnsupportedProtocolException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -956,6 +974,10 @@ func ExampleELBV2_ModifyListener_shared00() { fmt.Println(elbv2.ErrCodeTooManyRegistrationsForTargetIdException, aerr.Error()) case elbv2.ErrCodeTooManyTargetsException: fmt.Println(elbv2.ErrCodeTooManyTargetsException, aerr.Error()) + case elbv2.ErrCodeTooManyActionsException: + fmt.Println(elbv2.ErrCodeTooManyActionsException, aerr.Error()) + case elbv2.ErrCodeInvalidLoadBalancerActionException: + fmt.Println(elbv2.ErrCodeInvalidLoadBalancerActionException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -1014,6 +1036,10 @@ func ExampleELBV2_ModifyListener_shared01() { fmt.Println(elbv2.ErrCodeTooManyRegistrationsForTargetIdException, aerr.Error()) case elbv2.ErrCodeTooManyTargetsException: fmt.Println(elbv2.ErrCodeTooManyTargetsException, aerr.Error()) + case elbv2.ErrCodeTooManyActionsException: + fmt.Println(elbv2.ErrCodeTooManyActionsException, aerr.Error()) + case elbv2.ErrCodeInvalidLoadBalancerActionException: + fmt.Println(elbv2.ErrCodeInvalidLoadBalancerActionException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -1184,6 +1210,12 @@ func ExampleELBV2_ModifyRule_shared00() { fmt.Println(elbv2.ErrCodeTooManyTargetsException, aerr.Error()) case elbv2.ErrCodeTargetGroupNotFoundException: fmt.Println(elbv2.ErrCodeTargetGroupNotFoundException, aerr.Error()) + case elbv2.ErrCodeUnsupportedProtocolException: + fmt.Println(elbv2.ErrCodeUnsupportedProtocolException, aerr.Error()) + case elbv2.ErrCodeTooManyActionsException: + fmt.Println(elbv2.ErrCodeTooManyActionsException, aerr.Error()) + case elbv2.ErrCodeInvalidLoadBalancerActionException: + fmt.Println(elbv2.ErrCodeInvalidLoadBalancerActionException, aerr.Error()) default: fmt.Println(aerr.Error()) } diff --git a/vendor/github.com/aws/aws-sdk-go/service/elbv2/service.go b/vendor/github.com/aws/aws-sdk-go/service/elbv2/service.go index c3733846c..ad97e8df8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/elbv2/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/elbv2/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elasticloadbalancing" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elasticloadbalancing" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Elastic Load Balancing v2" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ELBV2 client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/emr/service.go b/vendor/github.com/aws/aws-sdk-go/service/emr/service.go index 61fc42581..92735a793 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/emr/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/emr/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "elasticmapreduce" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "elasticmapreduce" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "EMR" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the EMR client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/firehose/service.go b/vendor/github.com/aws/aws-sdk-go/service/firehose/service.go index 973386c05..bcdf23dff 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/firehose/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/firehose/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "firehose" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "firehose" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Firehose" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Firehose client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/fms/service.go b/vendor/github.com/aws/aws-sdk-go/service/fms/service.go index d0445c8c3..6103e57fd 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/fms/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/fms/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "fms" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "fms" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "FMS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the FMS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/gamelift/service.go b/vendor/github.com/aws/aws-sdk-go/service/gamelift/service.go index b79ac2047..a2361e476 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/gamelift/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/gamelift/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "gamelift" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "gamelift" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "GameLift" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the GameLift client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/glacier/service.go b/vendor/github.com/aws/aws-sdk-go/service/glacier/service.go index b875f0faf..85e6e367b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/glacier/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/glacier/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "glacier" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "glacier" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Glacier" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Glacier client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/glue/api.go b/vendor/github.com/aws/aws-sdk-go/service/glue/api.go index 553546ed3..fe60755ac 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/glue/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/glue/api.go @@ -7626,6 +7626,9 @@ type Action struct { // The name of a job to be executed. JobName *string `min:"1" type:"string"` + // Specifies configuration properties of a job run notification. + NotificationProperty *NotificationProperty `type:"structure"` + // The job run timeout in minutes. It overrides the timeout value of the job. Timeout *int64 `min:"1" type:"integer"` } @@ -7649,6 +7652,11 @@ func (s *Action) Validate() error { if s.Timeout != nil && *s.Timeout < 1 { invalidParams.Add(request.NewErrParamMinValue("Timeout", 1)) } + if s.NotificationProperty != nil { + if err := s.NotificationProperty.Validate(); err != nil { + invalidParams.AddNested("NotificationProperty", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -7668,6 +7676,12 @@ func (s *Action) SetJobName(v string) *Action { return s } +// SetNotificationProperty sets the NotificationProperty field's value. +func (s *Action) SetNotificationProperty(v *NotificationProperty) *Action { + s.NotificationProperty = v + return s +} + // SetTimeout sets the Timeout field's value. func (s *Action) SetTimeout(v int64) *Action { s.Timeout = &v @@ -10269,6 +10283,9 @@ type CreateJobInput struct { // Name is a required field Name *string `min:"1" type:"string" required:"true"` + // Specifies configuration properties of a job notification. + NotificationProperty *NotificationProperty `type:"structure"` + // The name or ARN of the IAM role associated with this job. // // Role is a required field @@ -10306,6 +10323,11 @@ func (s *CreateJobInput) Validate() error { if s.Timeout != nil && *s.Timeout < 1 { invalidParams.Add(request.NewErrParamMinValue("Timeout", 1)) } + if s.NotificationProperty != nil { + if err := s.NotificationProperty.Validate(); err != nil { + invalidParams.AddNested("NotificationProperty", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -10367,6 +10389,12 @@ func (s *CreateJobInput) SetName(v string) *CreateJobInput { return s } +// SetNotificationProperty sets the NotificationProperty field's value. +func (s *CreateJobInput) SetNotificationProperty(v *NotificationProperty) *CreateJobInput { + s.NotificationProperty = v + return s +} + // SetRole sets the Role field's value. func (s *CreateJobInput) SetRole(v string) *CreateJobInput { s.Role = &v @@ -15173,6 +15201,9 @@ type Job struct { // The name you assign to this job definition. Name *string `min:"1" type:"string"` + // Specifies configuration properties of a job notification. + NotificationProperty *NotificationProperty `type:"structure"` + // The name or ARN of the IAM role associated with this job. Role *string `type:"string"` @@ -15256,6 +15287,12 @@ func (s *Job) SetName(v string) *Job { return s } +// SetNotificationProperty sets the NotificationProperty field's value. +func (s *Job) SetNotificationProperty(v *NotificationProperty) *Job { + s.NotificationProperty = v + return s +} + // SetRole sets the Role field's value. func (s *Job) SetRole(v string) *Job { s.Role = &v @@ -15411,6 +15448,9 @@ type JobRun struct { // The last time this job run was modified. LastModifiedOn *time.Time `type:"timestamp" timestampFormat:"unix"` + // Specifies configuration properties of a job run notification. + NotificationProperty *NotificationProperty `type:"structure"` + // A list of predecessors to this job run. PredecessorRuns []*Predecessor `type:"list"` @@ -15498,6 +15538,12 @@ func (s *JobRun) SetLastModifiedOn(v time.Time) *JobRun { return s } +// SetNotificationProperty sets the NotificationProperty field's value. +func (s *JobRun) SetNotificationProperty(v *NotificationProperty) *JobRun { + s.NotificationProperty = v + return s +} + // SetPredecessorRuns sets the PredecessorRuns field's value. func (s *JobRun) SetPredecessorRuns(v []*Predecessor) *JobRun { s.PredecessorRuns = v @@ -15573,6 +15619,9 @@ type JobUpdate struct { // The maximum number of times to retry this job if it fails. MaxRetries *int64 `type:"integer"` + // Specifies configuration properties of a job notification. + NotificationProperty *NotificationProperty `type:"structure"` + // The name or ARN of the IAM role associated with this job (required). Role *string `type:"string"` @@ -15596,6 +15645,11 @@ func (s *JobUpdate) Validate() error { if s.Timeout != nil && *s.Timeout < 1 { invalidParams.Add(request.NewErrParamMinValue("Timeout", 1)) } + if s.NotificationProperty != nil { + if err := s.NotificationProperty.Validate(); err != nil { + invalidParams.AddNested("NotificationProperty", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -15651,6 +15705,12 @@ func (s *JobUpdate) SetMaxRetries(v int64) *JobUpdate { return s } +// SetNotificationProperty sets the NotificationProperty field's value. +func (s *JobUpdate) SetNotificationProperty(v *NotificationProperty) *JobUpdate { + s.NotificationProperty = v + return s +} + // SetRole sets the Role field's value. func (s *JobUpdate) SetRole(v string) *JobUpdate { s.Role = &v @@ -15930,6 +15990,44 @@ func (s *MappingEntry) SetTargetType(v string) *MappingEntry { return s } +// Specifies configuration properties of a notification. +type NotificationProperty struct { + _ struct{} `type:"structure"` + + // After a job run starts, the number of minutes to wait before sending a job + // run delay notification. + NotifyDelayAfter *int64 `min:"1" type:"integer"` +} + +// String returns the string representation +func (s NotificationProperty) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s NotificationProperty) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *NotificationProperty) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "NotificationProperty"} + if s.NotifyDelayAfter != nil && *s.NotifyDelayAfter < 1 { + invalidParams.Add(request.NewErrParamMinValue("NotifyDelayAfter", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetNotifyDelayAfter sets the NotifyDelayAfter field's value. +func (s *NotificationProperty) SetNotifyDelayAfter(v int64) *NotificationProperty { + s.NotifyDelayAfter = &v + return s +} + // Specifies the sort order of a sorted column. type Order struct { _ struct{} `type:"structure"` @@ -16881,6 +16979,9 @@ type StartJobRunInput struct { // The ID of a previous JobRun to retry. JobRunId *string `min:"1" type:"string"` + // Specifies configuration properties of a job run notification. + NotificationProperty *NotificationProperty `type:"structure"` + // The job run timeout in minutes. It overrides the timeout value of the job. Timeout *int64 `min:"1" type:"integer"` } @@ -16910,6 +17011,11 @@ func (s *StartJobRunInput) Validate() error { if s.Timeout != nil && *s.Timeout < 1 { invalidParams.Add(request.NewErrParamMinValue("Timeout", 1)) } + if s.NotificationProperty != nil { + if err := s.NotificationProperty.Validate(); err != nil { + invalidParams.AddNested("NotificationProperty", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -16941,6 +17047,12 @@ func (s *StartJobRunInput) SetJobRunId(v string) *StartJobRunInput { return s } +// SetNotificationProperty sets the NotificationProperty field's value. +func (s *StartJobRunInput) SetNotificationProperty(v *NotificationProperty) *StartJobRunInput { + s.NotificationProperty = v + return s +} + // SetTimeout sets the Timeout field's value. func (s *StartJobRunInput) SetTimeout(v int64) *StartJobRunInput { s.Timeout = &v diff --git a/vendor/github.com/aws/aws-sdk-go/service/glue/service.go b/vendor/github.com/aws/aws-sdk-go/service/glue/service.go index 5e017698a..075b0a1df 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/glue/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/glue/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "glue" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "glue" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Glue" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Glue client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/greengrass/service.go b/vendor/github.com/aws/aws-sdk-go/service/greengrass/service.go index 795892e32..5c449ebcc 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/greengrass/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/greengrass/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "greengrass" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "greengrass" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Greengrass" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Greengrass client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/guardduty/service.go b/vendor/github.com/aws/aws-sdk-go/service/guardduty/service.go index 54c24d4f7..0ca09946c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/guardduty/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/guardduty/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "guardduty" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "guardduty" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "GuardDuty" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the GuardDuty client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/health/service.go b/vendor/github.com/aws/aws-sdk-go/service/health/service.go index e85c6844b..d79b9cc49 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/health/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/health/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "health" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "health" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Health" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Health client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iam/service.go b/vendor/github.com/aws/aws-sdk-go/service/iam/service.go index 4f798c63d..940b4ce32 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iam/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iam/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "iam" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "iam" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IAM" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IAM client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/inspector/api.go b/vendor/github.com/aws/aws-sdk-go/service/inspector/api.go index b59beb031..9bda4e841 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/inspector/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/inspector/api.go @@ -182,6 +182,10 @@ func (c *Inspector) CreateAssessmentTargetRequest(input *CreateAssessmentTargetI // The request was rejected because it referenced an entity that does not exist. // The error code describes the entity. // +// * ErrCodeInvalidCrossAccountRoleException "InvalidCrossAccountRoleException" +// Amazon Inspector cannot assume the cross-account role that it needs to list +// your EC2 instances during the assessment run. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/CreateAssessmentTarget func (c *Inspector) CreateAssessmentTarget(input *CreateAssessmentTargetInput) (*CreateAssessmentTargetOutput, error) { req, out := c.CreateAssessmentTargetRequest(input) @@ -302,6 +306,102 @@ func (c *Inspector) CreateAssessmentTemplateWithContext(ctx aws.Context, input * return out, req.Send() } +const opCreateExclusionsPreview = "CreateExclusionsPreview" + +// CreateExclusionsPreviewRequest generates a "aws/request.Request" representing the +// client's request for the CreateExclusionsPreview operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateExclusionsPreview for more information on using the CreateExclusionsPreview +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateExclusionsPreviewRequest method. +// req, resp := client.CreateExclusionsPreviewRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/CreateExclusionsPreview +func (c *Inspector) CreateExclusionsPreviewRequest(input *CreateExclusionsPreviewInput) (req *request.Request, output *CreateExclusionsPreviewOutput) { + op := &request.Operation{ + Name: opCreateExclusionsPreview, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateExclusionsPreviewInput{} + } + + output = &CreateExclusionsPreviewOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateExclusionsPreview API operation for Amazon Inspector. +// +// Starts the generation of an exclusions preview for the specified assessment +// template. The exclusions preview lists the potential exclusions (ExclusionPreview) +// that Inspector can detect before it runs the assessment. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Inspector's +// API operation CreateExclusionsPreview for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodePreviewGenerationInProgressException "PreviewGenerationInProgressException" +// The request is rejected. The specified assessment template is currently generating +// an exclusions preview. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeNoSuchEntityException "NoSuchEntityException" +// The request was rejected because it referenced an entity that does not exist. +// The error code describes the entity. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/CreateExclusionsPreview +func (c *Inspector) CreateExclusionsPreview(input *CreateExclusionsPreviewInput) (*CreateExclusionsPreviewOutput, error) { + req, out := c.CreateExclusionsPreviewRequest(input) + return out, req.Send() +} + +// CreateExclusionsPreviewWithContext is the same as CreateExclusionsPreview with the addition of +// the ability to pass a context and additional request options. +// +// See CreateExclusionsPreview for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Inspector) CreateExclusionsPreviewWithContext(ctx aws.Context, input *CreateExclusionsPreviewInput, opts ...request.Option) (*CreateExclusionsPreviewOutput, error) { + req, out := c.CreateExclusionsPreviewRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opCreateResourceGroup = "CreateResourceGroup" // CreateResourceGroupRequest generates a "aws/request.Request" representing the @@ -1017,6 +1117,89 @@ func (c *Inspector) DescribeCrossAccountAccessRoleWithContext(ctx aws.Context, i return out, req.Send() } +const opDescribeExclusions = "DescribeExclusions" + +// DescribeExclusionsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeExclusions operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeExclusions for more information on using the DescribeExclusions +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeExclusionsRequest method. +// req, resp := client.DescribeExclusionsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/DescribeExclusions +func (c *Inspector) DescribeExclusionsRequest(input *DescribeExclusionsInput) (req *request.Request, output *DescribeExclusionsOutput) { + op := &request.Operation{ + Name: opDescribeExclusions, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeExclusionsInput{} + } + + output = &DescribeExclusionsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeExclusions API operation for Amazon Inspector. +// +// Describes the exclusions that are specified by the exclusions' ARNs. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Inspector's +// API operation DescribeExclusions for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/DescribeExclusions +func (c *Inspector) DescribeExclusions(input *DescribeExclusionsInput) (*DescribeExclusionsOutput, error) { + req, out := c.DescribeExclusionsRequest(input) + return out, req.Send() +} + +// DescribeExclusionsWithContext is the same as DescribeExclusions with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeExclusions for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Inspector) DescribeExclusionsWithContext(ctx aws.Context, input *DescribeExclusionsInput, opts ...request.Option) (*DescribeExclusionsOutput, error) { + req, out := c.DescribeExclusionsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDescribeFindings = "DescribeFindings" // DescribeFindingsRequest generates a "aws/request.Request" representing the @@ -1370,6 +1553,154 @@ func (c *Inspector) GetAssessmentReportWithContext(ctx aws.Context, input *GetAs return out, req.Send() } +const opGetExclusionsPreview = "GetExclusionsPreview" + +// GetExclusionsPreviewRequest generates a "aws/request.Request" representing the +// client's request for the GetExclusionsPreview operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See GetExclusionsPreview for more information on using the GetExclusionsPreview +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the GetExclusionsPreviewRequest method. +// req, resp := client.GetExclusionsPreviewRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/GetExclusionsPreview +func (c *Inspector) GetExclusionsPreviewRequest(input *GetExclusionsPreviewInput) (req *request.Request, output *GetExclusionsPreviewOutput) { + op := &request.Operation{ + Name: opGetExclusionsPreview, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"nextToken"}, + OutputTokens: []string{"nextToken"}, + LimitToken: "maxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &GetExclusionsPreviewInput{} + } + + output = &GetExclusionsPreviewOutput{} + req = c.newRequest(op, input, output) + return +} + +// GetExclusionsPreview API operation for Amazon Inspector. +// +// Retrieves the exclusions preview (a list of ExclusionPreview objects) specified +// by the preview token. You can obtain the preview token by running the CreateExclusionsPreview +// API. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Inspector's +// API operation GetExclusionsPreview for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeNoSuchEntityException "NoSuchEntityException" +// The request was rejected because it referenced an entity that does not exist. +// The error code describes the entity. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/GetExclusionsPreview +func (c *Inspector) GetExclusionsPreview(input *GetExclusionsPreviewInput) (*GetExclusionsPreviewOutput, error) { + req, out := c.GetExclusionsPreviewRequest(input) + return out, req.Send() +} + +// GetExclusionsPreviewWithContext is the same as GetExclusionsPreview with the addition of +// the ability to pass a context and additional request options. +// +// See GetExclusionsPreview for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Inspector) GetExclusionsPreviewWithContext(ctx aws.Context, input *GetExclusionsPreviewInput, opts ...request.Option) (*GetExclusionsPreviewOutput, error) { + req, out := c.GetExclusionsPreviewRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// GetExclusionsPreviewPages iterates over the pages of a GetExclusionsPreview operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See GetExclusionsPreview method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a GetExclusionsPreview operation. +// pageNum := 0 +// err := client.GetExclusionsPreviewPages(params, +// func(page *GetExclusionsPreviewOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Inspector) GetExclusionsPreviewPages(input *GetExclusionsPreviewInput, fn func(*GetExclusionsPreviewOutput, bool) bool) error { + return c.GetExclusionsPreviewPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// GetExclusionsPreviewPagesWithContext same as GetExclusionsPreviewPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Inspector) GetExclusionsPreviewPagesWithContext(ctx aws.Context, input *GetExclusionsPreviewInput, fn func(*GetExclusionsPreviewOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *GetExclusionsPreviewInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.GetExclusionsPreviewRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*GetExclusionsPreviewOutput), !p.HasNextPage()) + } + return p.Err() +} + const opGetTelemetryMetadata = "GetTelemetryMetadata" // GetTelemetryMetadataRequest generates a "aws/request.Request" representing the @@ -2194,6 +2525,152 @@ func (c *Inspector) ListEventSubscriptionsPagesWithContext(ctx aws.Context, inpu return p.Err() } +const opListExclusions = "ListExclusions" + +// ListExclusionsRequest generates a "aws/request.Request" representing the +// client's request for the ListExclusions operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListExclusions for more information on using the ListExclusions +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListExclusionsRequest method. +// req, resp := client.ListExclusionsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/ListExclusions +func (c *Inspector) ListExclusionsRequest(input *ListExclusionsInput) (req *request.Request, output *ListExclusionsOutput) { + op := &request.Operation{ + Name: opListExclusions, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"nextToken"}, + OutputTokens: []string{"nextToken"}, + LimitToken: "maxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListExclusionsInput{} + } + + output = &ListExclusionsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListExclusions API operation for Amazon Inspector. +// +// List exclusions that are generated by the assessment run. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Inspector's +// API operation ListExclusions for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeNoSuchEntityException "NoSuchEntityException" +// The request was rejected because it referenced an entity that does not exist. +// The error code describes the entity. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/inspector-2016-02-16/ListExclusions +func (c *Inspector) ListExclusions(input *ListExclusionsInput) (*ListExclusionsOutput, error) { + req, out := c.ListExclusionsRequest(input) + return out, req.Send() +} + +// ListExclusionsWithContext is the same as ListExclusions with the addition of +// the ability to pass a context and additional request options. +// +// See ListExclusions for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Inspector) ListExclusionsWithContext(ctx aws.Context, input *ListExclusionsInput, opts ...request.Option) (*ListExclusionsOutput, error) { + req, out := c.ListExclusionsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListExclusionsPages iterates over the pages of a ListExclusions operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListExclusions method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListExclusions operation. +// pageNum := 0 +// err := client.ListExclusionsPages(params, +// func(page *ListExclusionsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Inspector) ListExclusionsPages(input *ListExclusionsInput, fn func(*ListExclusionsOutput, bool) bool) error { + return c.ListExclusionsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListExclusionsPagesWithContext same as ListExclusionsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Inspector) ListExclusionsPagesWithContext(ctx aws.Context, input *ListExclusionsInput, fn func(*ListExclusionsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListExclusionsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListExclusionsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListExclusionsOutput), !p.HasNextPage()) + } + return p.Err() +} + const opListFindings = "ListFindings" // ListFindingsRequest generates a "aws/request.Request" representing the @@ -4278,9 +4755,7 @@ type AssessmentTarget struct { // The ARN that specifies the resource group that is associated with the assessment // target. - // - // ResourceGroupArn is a required field - ResourceGroupArn *string `locationName:"resourceGroupArn" min:"1" type:"string" required:"true"` + ResourceGroupArn *string `locationName:"resourceGroupArn" min:"1" type:"string"` // The time at which UpdateAssessmentTarget is called. // @@ -4394,7 +4869,7 @@ type AssessmentTemplate struct { // CreatedAt is a required field CreatedAt *time.Time `locationName:"createdAt" type:"timestamp" timestampFormat:"unix" required:"true"` - // The duration in seconds specified for this assessment tempate. The default + // The duration in seconds specified for this assessment template. The default // value is 3600 seconds (one hour). The maximum value is 86400 seconds (one // day). // @@ -4403,7 +4878,7 @@ type AssessmentTemplate struct { // The Amazon Resource Name (ARN) of the most recent assessment run associated // with this assessment template. This value exists only when the value of assessmentRunCount - // is greater than zero. + // is greaterpa than zero. LastAssessmentRunArn *string `locationName:"lastAssessmentRunArn" min:"1" type:"string"` // The name of the assessment template. @@ -4692,9 +5167,7 @@ type CreateAssessmentTargetInput struct { // The ARN that specifies the resource group that is used to create the assessment // target. - // - // ResourceGroupArn is a required field - ResourceGroupArn *string `locationName:"resourceGroupArn" min:"1" type:"string" required:"true"` + ResourceGroupArn *string `locationName:"resourceGroupArn" min:"1" type:"string"` } // String returns the string representation @@ -4716,9 +5189,6 @@ func (s *CreateAssessmentTargetInput) Validate() error { if s.AssessmentTargetName != nil && len(*s.AssessmentTargetName) < 1 { invalidParams.Add(request.NewErrParamMinLen("AssessmentTargetName", 1)) } - if s.ResourceGroupArn == nil { - invalidParams.Add(request.NewErrParamRequired("ResourceGroupArn")) - } if s.ResourceGroupArn != nil && len(*s.ResourceGroupArn) < 1 { invalidParams.Add(request.NewErrParamMinLen("ResourceGroupArn", 1)) } @@ -4908,6 +5378,75 @@ func (s *CreateAssessmentTemplateOutput) SetAssessmentTemplateArn(v string) *Cre return s } +type CreateExclusionsPreviewInput struct { + _ struct{} `type:"structure"` + + // The ARN that specifies the assessment template for which you want to create + // an exclusions preview. + // + // AssessmentTemplateArn is a required field + AssessmentTemplateArn *string `locationName:"assessmentTemplateArn" min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s CreateExclusionsPreviewInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateExclusionsPreviewInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateExclusionsPreviewInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateExclusionsPreviewInput"} + if s.AssessmentTemplateArn == nil { + invalidParams.Add(request.NewErrParamRequired("AssessmentTemplateArn")) + } + if s.AssessmentTemplateArn != nil && len(*s.AssessmentTemplateArn) < 1 { + invalidParams.Add(request.NewErrParamMinLen("AssessmentTemplateArn", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAssessmentTemplateArn sets the AssessmentTemplateArn field's value. +func (s *CreateExclusionsPreviewInput) SetAssessmentTemplateArn(v string) *CreateExclusionsPreviewInput { + s.AssessmentTemplateArn = &v + return s +} + +type CreateExclusionsPreviewOutput struct { + _ struct{} `type:"structure"` + + // Specifies the unique identifier of the requested exclusions preview. You + // can use the unique identifier to retrieve the exclusions preview when running + // the GetExclusionsPreview API. + // + // PreviewToken is a required field + PreviewToken *string `locationName:"previewToken" type:"string" required:"true"` +} + +// String returns the string representation +func (s CreateExclusionsPreviewOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateExclusionsPreviewOutput) GoString() string { + return s.String() +} + +// SetPreviewToken sets the PreviewToken field's value. +func (s *CreateExclusionsPreviewOutput) SetPreviewToken(v string) *CreateExclusionsPreviewOutput { + s.PreviewToken = &v + return s +} + type CreateResourceGroupInput struct { _ struct{} `type:"structure"` @@ -5446,6 +5985,94 @@ func (s *DescribeCrossAccountAccessRoleOutput) SetValid(v bool) *DescribeCrossAc return s } +type DescribeExclusionsInput struct { + _ struct{} `type:"structure"` + + // The list of ARNs that specify the exclusions that you want to describe. + // + // ExclusionArns is a required field + ExclusionArns []*string `locationName:"exclusionArns" min:"1" type:"list" required:"true"` + + // The locale into which you want to translate the exclusion's title, description, + // and recommendation. + Locale *string `locationName:"locale" type:"string" enum:"Locale"` +} + +// String returns the string representation +func (s DescribeExclusionsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeExclusionsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeExclusionsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeExclusionsInput"} + if s.ExclusionArns == nil { + invalidParams.Add(request.NewErrParamRequired("ExclusionArns")) + } + if s.ExclusionArns != nil && len(s.ExclusionArns) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ExclusionArns", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetExclusionArns sets the ExclusionArns field's value. +func (s *DescribeExclusionsInput) SetExclusionArns(v []*string) *DescribeExclusionsInput { + s.ExclusionArns = v + return s +} + +// SetLocale sets the Locale field's value. +func (s *DescribeExclusionsInput) SetLocale(v string) *DescribeExclusionsInput { + s.Locale = &v + return s +} + +type DescribeExclusionsOutput struct { + _ struct{} `type:"structure"` + + // Information about the exclusions. + // + // Exclusions is a required field + Exclusions map[string]*Exclusion `locationName:"exclusions" min:"1" type:"map" required:"true"` + + // Exclusion details that cannot be described. An error code is provided for + // each failed item. + // + // FailedItems is a required field + FailedItems map[string]*FailedItemDetails `locationName:"failedItems" type:"map" required:"true"` +} + +// String returns the string representation +func (s DescribeExclusionsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeExclusionsOutput) GoString() string { + return s.String() +} + +// SetExclusions sets the Exclusions field's value. +func (s *DescribeExclusionsOutput) SetExclusions(v map[string]*Exclusion) *DescribeExclusionsOutput { + s.Exclusions = v + return s +} + +// SetFailedItems sets the FailedItems field's value. +func (s *DescribeExclusionsOutput) SetFailedItems(v map[string]*FailedItemDetails) *DescribeExclusionsOutput { + s.FailedItems = v + return s +} + type DescribeFindingsInput struct { _ struct{} `type:"structure"` @@ -5787,6 +6414,154 @@ func (s *EventSubscription) SetSubscribedAt(v time.Time) *EventSubscription { return s } +// Contains information about what was excluded from an assessment run. +type Exclusion struct { + _ struct{} `type:"structure"` + + // The ARN that specifies the exclusion. + // + // Arn is a required field + Arn *string `locationName:"arn" min:"1" type:"string" required:"true"` + + // The system-defined attributes for the exclusion. + Attributes []*Attribute `locationName:"attributes" type:"list"` + + // The description of the exclusion. + // + // Description is a required field + Description *string `locationName:"description" type:"string" required:"true"` + + // The recommendation for the exclusion. + // + // Recommendation is a required field + Recommendation *string `locationName:"recommendation" type:"string" required:"true"` + + // The AWS resources for which the exclusion pertains. + // + // Scopes is a required field + Scopes []*Scope `locationName:"scopes" min:"1" type:"list" required:"true"` + + // The name of the exclusion. + // + // Title is a required field + Title *string `locationName:"title" type:"string" required:"true"` +} + +// String returns the string representation +func (s Exclusion) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Exclusion) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *Exclusion) SetArn(v string) *Exclusion { + s.Arn = &v + return s +} + +// SetAttributes sets the Attributes field's value. +func (s *Exclusion) SetAttributes(v []*Attribute) *Exclusion { + s.Attributes = v + return s +} + +// SetDescription sets the Description field's value. +func (s *Exclusion) SetDescription(v string) *Exclusion { + s.Description = &v + return s +} + +// SetRecommendation sets the Recommendation field's value. +func (s *Exclusion) SetRecommendation(v string) *Exclusion { + s.Recommendation = &v + return s +} + +// SetScopes sets the Scopes field's value. +func (s *Exclusion) SetScopes(v []*Scope) *Exclusion { + s.Scopes = v + return s +} + +// SetTitle sets the Title field's value. +func (s *Exclusion) SetTitle(v string) *Exclusion { + s.Title = &v + return s +} + +// Contains information about what is excluded from an assessment run given +// the current state of the assessment template. +type ExclusionPreview struct { + _ struct{} `type:"structure"` + + // The system-defined attributes for the exclusion preview. + Attributes []*Attribute `locationName:"attributes" type:"list"` + + // The description of the exclusion preview. + // + // Description is a required field + Description *string `locationName:"description" type:"string" required:"true"` + + // The recommendation for the exclusion preview. + // + // Recommendation is a required field + Recommendation *string `locationName:"recommendation" type:"string" required:"true"` + + // The AWS resources for which the exclusion preview pertains. + // + // Scopes is a required field + Scopes []*Scope `locationName:"scopes" min:"1" type:"list" required:"true"` + + // The name of the exclusion preview. + // + // Title is a required field + Title *string `locationName:"title" type:"string" required:"true"` +} + +// String returns the string representation +func (s ExclusionPreview) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ExclusionPreview) GoString() string { + return s.String() +} + +// SetAttributes sets the Attributes field's value. +func (s *ExclusionPreview) SetAttributes(v []*Attribute) *ExclusionPreview { + s.Attributes = v + return s +} + +// SetDescription sets the Description field's value. +func (s *ExclusionPreview) SetDescription(v string) *ExclusionPreview { + s.Description = &v + return s +} + +// SetRecommendation sets the Recommendation field's value. +func (s *ExclusionPreview) SetRecommendation(v string) *ExclusionPreview { + s.Recommendation = &v + return s +} + +// SetScopes sets the Scopes field's value. +func (s *ExclusionPreview) SetScopes(v []*Scope) *ExclusionPreview { + s.Scopes = v + return s +} + +// SetTitle sets the Title field's value. +func (s *ExclusionPreview) SetTitle(v string) *ExclusionPreview { + s.Title = &v + return s +} + // Includes details about the failed items. type FailedItemDetails struct { _ struct{} `type:"structure"` @@ -6252,6 +7027,143 @@ func (s *GetAssessmentReportOutput) SetUrl(v string) *GetAssessmentReportOutput return s } +type GetExclusionsPreviewInput struct { + _ struct{} `type:"structure"` + + // The ARN that specifies the assessment template for which the exclusions preview + // was requested. + // + // AssessmentTemplateArn is a required field + AssessmentTemplateArn *string `locationName:"assessmentTemplateArn" min:"1" type:"string" required:"true"` + + // The locale into which you want to translate the exclusion's title, description, + // and recommendation. + Locale *string `locationName:"locale" type:"string" enum:"Locale"` + + // You can use this parameter to indicate the maximum number of items you want + // in the response. The default value is 100. The maximum value is 500. + MaxResults *int64 `locationName:"maxResults" type:"integer"` + + // You can use this parameter when paginating results. Set the value of this + // parameter to null on your first call to the GetExclusionsPreviewRequest action. + // Subsequent calls to the action fill nextToken in the request with the value + // of nextToken from the previous response to continue listing data. + NextToken *string `locationName:"nextToken" min:"1" type:"string"` + + // The unique identifier associated of the exclusions preview. + // + // PreviewToken is a required field + PreviewToken *string `locationName:"previewToken" type:"string" required:"true"` +} + +// String returns the string representation +func (s GetExclusionsPreviewInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetExclusionsPreviewInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *GetExclusionsPreviewInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "GetExclusionsPreviewInput"} + if s.AssessmentTemplateArn == nil { + invalidParams.Add(request.NewErrParamRequired("AssessmentTemplateArn")) + } + if s.AssessmentTemplateArn != nil && len(*s.AssessmentTemplateArn) < 1 { + invalidParams.Add(request.NewErrParamMinLen("AssessmentTemplateArn", 1)) + } + if s.NextToken != nil && len(*s.NextToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NextToken", 1)) + } + if s.PreviewToken == nil { + invalidParams.Add(request.NewErrParamRequired("PreviewToken")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAssessmentTemplateArn sets the AssessmentTemplateArn field's value. +func (s *GetExclusionsPreviewInput) SetAssessmentTemplateArn(v string) *GetExclusionsPreviewInput { + s.AssessmentTemplateArn = &v + return s +} + +// SetLocale sets the Locale field's value. +func (s *GetExclusionsPreviewInput) SetLocale(v string) *GetExclusionsPreviewInput { + s.Locale = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *GetExclusionsPreviewInput) SetMaxResults(v int64) *GetExclusionsPreviewInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *GetExclusionsPreviewInput) SetNextToken(v string) *GetExclusionsPreviewInput { + s.NextToken = &v + return s +} + +// SetPreviewToken sets the PreviewToken field's value. +func (s *GetExclusionsPreviewInput) SetPreviewToken(v string) *GetExclusionsPreviewInput { + s.PreviewToken = &v + return s +} + +type GetExclusionsPreviewOutput struct { + _ struct{} `type:"structure"` + + // Information about the exclusions included in the preview. + ExclusionPreviews []*ExclusionPreview `locationName:"exclusionPreviews" type:"list"` + + // When a response is generated, if there is more data to be listed, this parameters + // is present in the response and contains the value to use for the nextToken + // parameter in a subsequent pagination request. If there is no more data to + // be listed, this parameter is set to null. + NextToken *string `locationName:"nextToken" min:"1" type:"string"` + + // Specifies the status of the request to generate an exclusions preview. + // + // PreviewStatus is a required field + PreviewStatus *string `locationName:"previewStatus" type:"string" required:"true" enum:"PreviewStatus"` +} + +// String returns the string representation +func (s GetExclusionsPreviewOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetExclusionsPreviewOutput) GoString() string { + return s.String() +} + +// SetExclusionPreviews sets the ExclusionPreviews field's value. +func (s *GetExclusionsPreviewOutput) SetExclusionPreviews(v []*ExclusionPreview) *GetExclusionsPreviewOutput { + s.ExclusionPreviews = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *GetExclusionsPreviewOutput) SetNextToken(v string) *GetExclusionsPreviewOutput { + s.NextToken = &v + return s +} + +// SetPreviewStatus sets the PreviewStatus field's value. +func (s *GetExclusionsPreviewOutput) SetPreviewStatus(v string) *GetExclusionsPreviewOutput { + s.PreviewStatus = &v + return s +} + type GetTelemetryMetadataInput struct { _ struct{} `type:"structure"` @@ -6877,6 +7789,110 @@ func (s *ListEventSubscriptionsOutput) SetSubscriptions(v []*Subscription) *List return s } +type ListExclusionsInput struct { + _ struct{} `type:"structure"` + + // The ARN of the assessment run that generated the exclusions that you want + // to list. + // + // AssessmentRunArn is a required field + AssessmentRunArn *string `locationName:"assessmentRunArn" min:"1" type:"string" required:"true"` + + // You can use this parameter to indicate the maximum number of items you want + // in the response. The default value is 100. The maximum value is 500. + MaxResults *int64 `locationName:"maxResults" type:"integer"` + + // You can use this parameter when paginating results. Set the value of this + // parameter to null on your first call to the ListExclusionsRequest action. + // Subsequent calls to the action fill nextToken in the request with the value + // of nextToken from the previous response to continue listing data. + NextToken *string `locationName:"nextToken" min:"1" type:"string"` +} + +// String returns the string representation +func (s ListExclusionsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListExclusionsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListExclusionsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListExclusionsInput"} + if s.AssessmentRunArn == nil { + invalidParams.Add(request.NewErrParamRequired("AssessmentRunArn")) + } + if s.AssessmentRunArn != nil && len(*s.AssessmentRunArn) < 1 { + invalidParams.Add(request.NewErrParamMinLen("AssessmentRunArn", 1)) + } + if s.NextToken != nil && len(*s.NextToken) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NextToken", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAssessmentRunArn sets the AssessmentRunArn field's value. +func (s *ListExclusionsInput) SetAssessmentRunArn(v string) *ListExclusionsInput { + s.AssessmentRunArn = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListExclusionsInput) SetMaxResults(v int64) *ListExclusionsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListExclusionsInput) SetNextToken(v string) *ListExclusionsInput { + s.NextToken = &v + return s +} + +type ListExclusionsOutput struct { + _ struct{} `type:"structure"` + + // A list of exclusions' ARNs returned by the action. + // + // ExclusionArns is a required field + ExclusionArns []*string `locationName:"exclusionArns" type:"list" required:"true"` + + // When a response is generated, if there is more data to be listed, this parameters + // is present in the response and contains the value to use for the nextToken + // parameter in a subsequent pagination request. If there is no more data to + // be listed, this parameter is set to null. + NextToken *string `locationName:"nextToken" min:"1" type:"string"` +} + +// String returns the string representation +func (s ListExclusionsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListExclusionsOutput) GoString() string { + return s.String() +} + +// SetExclusionArns sets the ExclusionArns field's value. +func (s *ListExclusionsOutput) SetExclusionArns(v []*string) *ListExclusionsOutput { + s.ExclusionArns = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListExclusionsOutput) SetNextToken(v string) *ListExclusionsOutput { + s.NextToken = &v + return s +} + type ListFindingsInput struct { _ struct{} `type:"structure"` @@ -7559,6 +8575,39 @@ func (s *RulesPackage) SetVersion(v string) *RulesPackage { return s } +// This data type contains key-value pairs that identify various Amazon resources. +type Scope struct { + _ struct{} `type:"structure"` + + // The type of the scope. + Key *string `locationName:"key" type:"string" enum:"ScopeType"` + + // The resource identifier for the specified scope type. + Value *string `locationName:"value" type:"string"` +} + +// String returns the string representation +func (s Scope) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Scope) GoString() string { + return s.String() +} + +// SetKey sets the Key field's value. +func (s *Scope) SetKey(v string) *Scope { + s.Key = &v + return s +} + +// SetValue sets the Value field's value. +func (s *Scope) SetValue(v string) *Scope { + s.Value = &v + return s +} + // This data type is used in the Finding data type. type ServiceAttributes struct { _ struct{} `type:"structure"` @@ -8203,9 +9252,7 @@ type UpdateAssessmentTargetInput struct { // The ARN of the resource group that is used to specify the new resource group // to associate with the assessment target. - // - // ResourceGroupArn is a required field - ResourceGroupArn *string `locationName:"resourceGroupArn" min:"1" type:"string" required:"true"` + ResourceGroupArn *string `locationName:"resourceGroupArn" min:"1" type:"string"` } // String returns the string representation @@ -8233,9 +9280,6 @@ func (s *UpdateAssessmentTargetInput) Validate() error { if s.AssessmentTargetName != nil && len(*s.AssessmentTargetName) < 1 { invalidParams.Add(request.NewErrParamMinLen("AssessmentTargetName", 1)) } - if s.ResourceGroupArn == nil { - invalidParams.Add(request.NewErrParamRequired("ResourceGroupArn")) - } if s.ResourceGroupArn != nil && len(*s.ResourceGroupArn) < 1 { invalidParams.Add(request.NewErrParamMinLen("ResourceGroupArn", 1)) } @@ -8652,6 +9696,14 @@ const ( NoSuchEntityErrorCodeIamRoleDoesNotExist = "IAM_ROLE_DOES_NOT_EXIST" ) +const ( + // PreviewStatusWorkInProgress is a PreviewStatus enum value + PreviewStatusWorkInProgress = "WORK_IN_PROGRESS" + + // PreviewStatusCompleted is a PreviewStatus enum value + PreviewStatusCompleted = "COMPLETED" +) + const ( // ReportFileFormatHtml is a ReportFileFormat enum value ReportFileFormatHtml = "HTML" @@ -8679,6 +9731,14 @@ const ( ReportTypeFull = "FULL" ) +const ( + // ScopeTypeInstanceId is a ScopeType enum value + ScopeTypeInstanceId = "INSTANCE_ID" + + // ScopeTypeRulesPackageArn is a ScopeType enum value + ScopeTypeRulesPackageArn = "RULES_PACKAGE_ARN" +) + const ( // SeverityLow is a Severity enum value SeverityLow = "Low" diff --git a/vendor/github.com/aws/aws-sdk-go/service/inspector/errors.go b/vendor/github.com/aws/aws-sdk-go/service/inspector/errors.go index abdadcccb..8d532b03e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/inspector/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/inspector/errors.go @@ -58,6 +58,13 @@ const ( // The error code describes the entity. ErrCodeNoSuchEntityException = "NoSuchEntityException" + // ErrCodePreviewGenerationInProgressException for service response error code + // "PreviewGenerationInProgressException". + // + // The request is rejected. The specified assessment template is currently generating + // an exclusions preview. + ErrCodePreviewGenerationInProgressException = "PreviewGenerationInProgressException" + // ErrCodeUnsupportedFeatureException for service response error code // "UnsupportedFeatureException". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/inspector/examples_test.go b/vendor/github.com/aws/aws-sdk-go/service/inspector/examples_test.go index 6f843c7a7..c3d21df5f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/inspector/examples_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/inspector/examples_test.go @@ -95,6 +95,8 @@ func ExampleInspector_CreateAssessmentTarget_shared00() { fmt.Println(inspector.ErrCodeAccessDeniedException, aerr.Error()) case inspector.ErrCodeNoSuchEntityException: fmt.Println(inspector.ErrCodeNoSuchEntityException, aerr.Error()) + case inspector.ErrCodeInvalidCrossAccountRoleException: + fmt.Println(inspector.ErrCodeInvalidCrossAccountRoleException, aerr.Error()) default: fmt.Println(aerr.Error()) } diff --git a/vendor/github.com/aws/aws-sdk-go/service/inspector/inspectoriface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/inspector/inspectoriface/interface.go index eaed8aa17..90ac08d76 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/inspector/inspectoriface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/inspector/inspectoriface/interface.go @@ -72,6 +72,10 @@ type InspectorAPI interface { CreateAssessmentTemplateWithContext(aws.Context, *inspector.CreateAssessmentTemplateInput, ...request.Option) (*inspector.CreateAssessmentTemplateOutput, error) CreateAssessmentTemplateRequest(*inspector.CreateAssessmentTemplateInput) (*request.Request, *inspector.CreateAssessmentTemplateOutput) + CreateExclusionsPreview(*inspector.CreateExclusionsPreviewInput) (*inspector.CreateExclusionsPreviewOutput, error) + CreateExclusionsPreviewWithContext(aws.Context, *inspector.CreateExclusionsPreviewInput, ...request.Option) (*inspector.CreateExclusionsPreviewOutput, error) + CreateExclusionsPreviewRequest(*inspector.CreateExclusionsPreviewInput) (*request.Request, *inspector.CreateExclusionsPreviewOutput) + CreateResourceGroup(*inspector.CreateResourceGroupInput) (*inspector.CreateResourceGroupOutput, error) CreateResourceGroupWithContext(aws.Context, *inspector.CreateResourceGroupInput, ...request.Option) (*inspector.CreateResourceGroupOutput, error) CreateResourceGroupRequest(*inspector.CreateResourceGroupInput) (*request.Request, *inspector.CreateResourceGroupOutput) @@ -104,6 +108,10 @@ type InspectorAPI interface { DescribeCrossAccountAccessRoleWithContext(aws.Context, *inspector.DescribeCrossAccountAccessRoleInput, ...request.Option) (*inspector.DescribeCrossAccountAccessRoleOutput, error) DescribeCrossAccountAccessRoleRequest(*inspector.DescribeCrossAccountAccessRoleInput) (*request.Request, *inspector.DescribeCrossAccountAccessRoleOutput) + DescribeExclusions(*inspector.DescribeExclusionsInput) (*inspector.DescribeExclusionsOutput, error) + DescribeExclusionsWithContext(aws.Context, *inspector.DescribeExclusionsInput, ...request.Option) (*inspector.DescribeExclusionsOutput, error) + DescribeExclusionsRequest(*inspector.DescribeExclusionsInput) (*request.Request, *inspector.DescribeExclusionsOutput) + DescribeFindings(*inspector.DescribeFindingsInput) (*inspector.DescribeFindingsOutput, error) DescribeFindingsWithContext(aws.Context, *inspector.DescribeFindingsInput, ...request.Option) (*inspector.DescribeFindingsOutput, error) DescribeFindingsRequest(*inspector.DescribeFindingsInput) (*request.Request, *inspector.DescribeFindingsOutput) @@ -120,6 +128,13 @@ type InspectorAPI interface { GetAssessmentReportWithContext(aws.Context, *inspector.GetAssessmentReportInput, ...request.Option) (*inspector.GetAssessmentReportOutput, error) GetAssessmentReportRequest(*inspector.GetAssessmentReportInput) (*request.Request, *inspector.GetAssessmentReportOutput) + GetExclusionsPreview(*inspector.GetExclusionsPreviewInput) (*inspector.GetExclusionsPreviewOutput, error) + GetExclusionsPreviewWithContext(aws.Context, *inspector.GetExclusionsPreviewInput, ...request.Option) (*inspector.GetExclusionsPreviewOutput, error) + GetExclusionsPreviewRequest(*inspector.GetExclusionsPreviewInput) (*request.Request, *inspector.GetExclusionsPreviewOutput) + + GetExclusionsPreviewPages(*inspector.GetExclusionsPreviewInput, func(*inspector.GetExclusionsPreviewOutput, bool) bool) error + GetExclusionsPreviewPagesWithContext(aws.Context, *inspector.GetExclusionsPreviewInput, func(*inspector.GetExclusionsPreviewOutput, bool) bool, ...request.Option) error + GetTelemetryMetadata(*inspector.GetTelemetryMetadataInput) (*inspector.GetTelemetryMetadataOutput, error) GetTelemetryMetadataWithContext(aws.Context, *inspector.GetTelemetryMetadataInput, ...request.Option) (*inspector.GetTelemetryMetadataOutput, error) GetTelemetryMetadataRequest(*inspector.GetTelemetryMetadataInput) (*request.Request, *inspector.GetTelemetryMetadataOutput) @@ -159,6 +174,13 @@ type InspectorAPI interface { ListEventSubscriptionsPages(*inspector.ListEventSubscriptionsInput, func(*inspector.ListEventSubscriptionsOutput, bool) bool) error ListEventSubscriptionsPagesWithContext(aws.Context, *inspector.ListEventSubscriptionsInput, func(*inspector.ListEventSubscriptionsOutput, bool) bool, ...request.Option) error + ListExclusions(*inspector.ListExclusionsInput) (*inspector.ListExclusionsOutput, error) + ListExclusionsWithContext(aws.Context, *inspector.ListExclusionsInput, ...request.Option) (*inspector.ListExclusionsOutput, error) + ListExclusionsRequest(*inspector.ListExclusionsInput) (*request.Request, *inspector.ListExclusionsOutput) + + ListExclusionsPages(*inspector.ListExclusionsInput, func(*inspector.ListExclusionsOutput, bool) bool) error + ListExclusionsPagesWithContext(aws.Context, *inspector.ListExclusionsInput, func(*inspector.ListExclusionsOutput, bool) bool, ...request.Option) error + ListFindings(*inspector.ListFindingsInput) (*inspector.ListFindingsOutput, error) ListFindingsWithContext(aws.Context, *inspector.ListFindingsInput, ...request.Option) (*inspector.ListFindingsOutput, error) ListFindingsRequest(*inspector.ListFindingsInput) (*request.Request, *inspector.ListFindingsOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/inspector/service.go b/vendor/github.com/aws/aws-sdk-go/service/inspector/service.go index 1d65f0707..2e68b4e4d 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/inspector/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/inspector/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "inspector" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "inspector" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Inspector" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Inspector client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iot/api.go b/vendor/github.com/aws/aws-sdk-go/service/iot/api.go index a86abf00e..4f24727d3 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iot/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iot/api.go @@ -266,7 +266,7 @@ func (c *IoT) AssociateTargetsWithJobRequest(input *AssociateTargetsWithJobInput // The specified resource does not exist. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -368,7 +368,7 @@ func (c *IoT) AttachPolicyRequest(input *AttachPolicyInput) (req *request.Reques // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) AttachPolicy(input *AttachPolicyInput) (*AttachPolicyOutput, error) { req, out := c.AttachPolicyRequest(input) @@ -470,7 +470,7 @@ func (c *IoT) AttachPrincipalPolicyRequest(input *AttachPrincipalPolicyInput) (r // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) AttachPrincipalPolicy(input *AttachPrincipalPolicyInput) (*AttachPrincipalPolicyOutput, error) { req, out := c.AttachPrincipalPolicyRequest(input) @@ -775,6 +775,102 @@ func (c *IoT) CancelJobWithContext(ctx aws.Context, input *CancelJobInput, opts return out, req.Send() } +const opCancelJobExecution = "CancelJobExecution" + +// CancelJobExecutionRequest generates a "aws/request.Request" representing the +// client's request for the CancelJobExecution operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CancelJobExecution for more information on using the CancelJobExecution +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CancelJobExecutionRequest method. +// req, resp := client.CancelJobExecutionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +func (c *IoT) CancelJobExecutionRequest(input *CancelJobExecutionInput) (req *request.Request, output *CancelJobExecutionOutput) { + op := &request.Operation{ + Name: opCancelJobExecution, + HTTPMethod: "PUT", + HTTPPath: "/things/{thingName}/jobs/{jobId}/cancel", + } + + if input == nil { + input = &CancelJobExecutionInput{} + } + + output = &CancelJobExecutionOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(restjson.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// CancelJobExecution API operation for AWS IoT. +// +// Cancels the execution of a job for a given thing. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS IoT's +// API operation CancelJobExecution for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is not valid. +// +// * ErrCodeInvalidStateTransitionException "InvalidStateTransitionException" +// An attempt was made to change to an invalid state, for example by deleting +// a job or a job execution which is "IN_PROGRESS" without setting the force +// parameter. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource does not exist. +// +// * ErrCodeThrottlingException "ThrottlingException" +// The rate exceeds the limit. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is temporarily unavailable. +// +// * ErrCodeVersionConflictException "VersionConflictException" +// An exception thrown when the version of an entity specified with the expectedVersion +// parameter does not match the latest version in the system. +// +func (c *IoT) CancelJobExecution(input *CancelJobExecutionInput) (*CancelJobExecutionOutput, error) { + req, out := c.CancelJobExecutionRequest(input) + return out, req.Send() +} + +// CancelJobExecutionWithContext is the same as CancelJobExecution with the addition of +// the ability to pass a context and additional request options. +// +// See CancelJobExecution for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *IoT) CancelJobExecutionWithContext(ctx aws.Context, input *CancelJobExecutionInput, opts ...request.Option) (*CancelJobExecutionOutput, error) { + req, out := c.CancelJobExecutionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opClearDefaultAuthorizer = "ClearDefaultAuthorizer" // ClearDefaultAuthorizerRequest generates a "aws/request.Request" representing the @@ -925,7 +1021,7 @@ func (c *IoT) CreateAuthorizerRequest(input *CreateAuthorizerInput) (req *reques // The request is not valid. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -1151,7 +1247,7 @@ func (c *IoT) CreateJobRequest(input *CreateJobInput) (req *request.Request, out // The resource already exists. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -1627,7 +1723,7 @@ func (c *IoT) CreateRoleAliasRequest(input *CreateRoleAliasInput) (req *request. // The request is not valid. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -2416,6 +2512,202 @@ func (c *IoT) DeleteCertificateWithContext(ctx aws.Context, input *DeleteCertifi return out, req.Send() } +const opDeleteJob = "DeleteJob" + +// DeleteJobRequest generates a "aws/request.Request" representing the +// client's request for the DeleteJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteJob for more information on using the DeleteJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteJobRequest method. +// req, resp := client.DeleteJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +func (c *IoT) DeleteJobRequest(input *DeleteJobInput) (req *request.Request, output *DeleteJobOutput) { + op := &request.Operation{ + Name: opDeleteJob, + HTTPMethod: "DELETE", + HTTPPath: "/jobs/{jobId}", + } + + if input == nil { + input = &DeleteJobInput{} + } + + output = &DeleteJobOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(restjson.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeleteJob API operation for AWS IoT. +// +// Deletes a job and its related job executions. +// +// Deleting a job may take time, depending on the number of job executions created +// for the job and various other factors. While the job is being deleted, the +// status of the job will be shown as "DELETION_IN_PROGRESS". Attempting to +// delete or cancel a job whose status is already "DELETION_IN_PROGRESS" will +// result in an error. +// +// Only 10 jobs may have status "DELETION_IN_PROGRESS" at the same time, or +// a LimitExceededException will occur. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS IoT's +// API operation DeleteJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is not valid. +// +// * ErrCodeInvalidStateTransitionException "InvalidStateTransitionException" +// An attempt was made to change to an invalid state, for example by deleting +// a job or a job execution which is "IN_PROGRESS" without setting the force +// parameter. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource does not exist. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// A limit has been exceeded. +// +// * ErrCodeThrottlingException "ThrottlingException" +// The rate exceeds the limit. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is temporarily unavailable. +// +func (c *IoT) DeleteJob(input *DeleteJobInput) (*DeleteJobOutput, error) { + req, out := c.DeleteJobRequest(input) + return out, req.Send() +} + +// DeleteJobWithContext is the same as DeleteJob with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *IoT) DeleteJobWithContext(ctx aws.Context, input *DeleteJobInput, opts ...request.Option) (*DeleteJobOutput, error) { + req, out := c.DeleteJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteJobExecution = "DeleteJobExecution" + +// DeleteJobExecutionRequest generates a "aws/request.Request" representing the +// client's request for the DeleteJobExecution operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteJobExecution for more information on using the DeleteJobExecution +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteJobExecutionRequest method. +// req, resp := client.DeleteJobExecutionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +func (c *IoT) DeleteJobExecutionRequest(input *DeleteJobExecutionInput) (req *request.Request, output *DeleteJobExecutionOutput) { + op := &request.Operation{ + Name: opDeleteJobExecution, + HTTPMethod: "DELETE", + HTTPPath: "/things/{thingName}/jobs/{jobId}/executionNumber/{executionNumber}", + } + + if input == nil { + input = &DeleteJobExecutionInput{} + } + + output = &DeleteJobExecutionOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(restjson.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeleteJobExecution API operation for AWS IoT. +// +// Deletes a job execution. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS IoT's +// API operation DeleteJobExecution for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request is not valid. +// +// * ErrCodeInvalidStateTransitionException "InvalidStateTransitionException" +// An attempt was made to change to an invalid state, for example by deleting +// a job or a job execution which is "IN_PROGRESS" without setting the force +// parameter. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// The specified resource does not exist. +// +// * ErrCodeThrottlingException "ThrottlingException" +// The rate exceeds the limit. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is temporarily unavailable. +// +func (c *IoT) DeleteJobExecution(input *DeleteJobExecutionInput) (*DeleteJobExecutionOutput, error) { + req, out := c.DeleteJobExecutionRequest(input) + return out, req.Send() +} + +// DeleteJobExecutionWithContext is the same as DeleteJobExecution with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteJobExecution for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *IoT) DeleteJobExecutionWithContext(ctx aws.Context, input *DeleteJobExecutionInput, opts ...request.Option) (*DeleteJobExecutionOutput, error) { + req, out := c.DeleteJobExecutionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDeleteOTAUpdate = "DeleteOTAUpdate" // DeleteOTAUpdateRequest generates a "aws/request.Request" representing the @@ -3044,8 +3336,8 @@ func (c *IoT) DeleteThingRequest(input *DeleteThingInput) (req *request.Request, // The specified resource does not exist. // // * ErrCodeVersionConflictException "VersionConflictException" -// An exception thrown when the version of a thing passed to a command is different -// than the version specified with the --version parameter. +// An exception thrown when the version of an entity specified with the expectedVersion +// parameter does not match the latest version in the system. // // * ErrCodeInvalidRequestException "InvalidRequestException" // The request is not valid. @@ -3139,8 +3431,8 @@ func (c *IoT) DeleteThingGroupRequest(input *DeleteThingGroupInput) (req *reques // The request is not valid. // // * ErrCodeVersionConflictException "VersionConflictException" -// An exception thrown when the version of a thing passed to a command is different -// than the version specified with the --version parameter. +// An exception thrown when the version of an entity specified with the expectedVersion +// parameter does not match the latest version in the system. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -4923,7 +5215,7 @@ func (c *IoT) DetachPolicyRequest(input *DetachPolicyInput) (req *request.Reques // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) DetachPolicy(input *DetachPolicyInput) (*DetachPolicyOutput, error) { req, out := c.DetachPolicyRequest(input) @@ -5351,7 +5643,8 @@ func (c *IoT) GetEffectivePoliciesRequest(input *GetEffectivePoliciesInput) (req // GetEffectivePolicies API operation for AWS IoT. // -// Gets effective policies. +// Gets a list of the policies that have an effect on the authorization behavior +// of the specified device when it connects to the AWS IoT device gateway. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -5380,7 +5673,7 @@ func (c *IoT) GetEffectivePoliciesRequest(input *GetEffectivePoliciesInput) (req // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) GetEffectivePolicies(input *GetEffectivePoliciesInput) (*GetEffectivePoliciesOutput, error) { req, out := c.GetEffectivePoliciesRequest(input) @@ -6258,7 +6551,7 @@ func (c *IoT) ListAttachedPoliciesRequest(input *ListAttachedPoliciesInput) (req // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) ListAttachedPolicies(input *ListAttachedPoliciesInput) (*ListAttachedPoliciesOutput, error) { req, out := c.ListAttachedPoliciesRequest(input) @@ -7868,7 +8161,7 @@ func (c *IoT) ListTargetsForPolicyRequest(input *ListTargetsForPolicyInput) (req // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) ListTargetsForPolicy(input *ListTargetsForPolicyInput) (*ListTargetsForPolicyOutput, error) { req, out := c.ListTargetsForPolicyRequest(input) @@ -8818,7 +9111,7 @@ func (c *IoT) RegisterCACertificateRequest(input *RegisterCACertificateInput) (r // The rate exceeds the limit. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // // * ErrCodeUnauthorizedException "UnauthorizedException" // You are not authorized to perform this operation. @@ -10089,7 +10382,9 @@ func (c *IoT) TestAuthorizationRequest(input *TestAuthorizationInput) (req *requ // TestAuthorization API operation for AWS IoT. // -// Test custom authorization. +// Tests if a specified principal is authorized to perform an AWS IoT action +// on a specified resource. Use this to test and debug the authorization behavior +// of devices that connect to the AWS IoT device gateway. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -10118,7 +10413,7 @@ func (c *IoT) TestAuthorizationRequest(input *TestAuthorizationInput) (req *requ // An unexpected error has occurred. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // func (c *IoT) TestAuthorization(input *TestAuthorizationInput) (*TestAuthorizationOutput, error) { req, out := c.TestAuthorizationRequest(input) @@ -10183,7 +10478,9 @@ func (c *IoT) TestInvokeAuthorizerRequest(input *TestInvokeAuthorizerInput) (req // TestInvokeAuthorizer API operation for AWS IoT. // -// Invoke the specified custom authorizer for testing purposes. +// Tests a custom authorization behavior by invoking a specified custom authorizer. +// Use this to test and debug the custom authorization behavior of devices that +// connect to the AWS IoT device gateway. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -10403,7 +10700,7 @@ func (c *IoT) UpdateAuthorizerRequest(input *UpdateAuthorizerInput) (req *reques // The request is not valid. // // * ErrCodeLimitExceededException "LimitExceededException" -// The number of attached entities exceeds the limit. +// A limit has been exceeded. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -11041,8 +11338,8 @@ func (c *IoT) UpdateThingRequest(input *UpdateThingInput) (req *request.Request, // The request is not valid. // // * ErrCodeVersionConflictException "VersionConflictException" -// An exception thrown when the version of a thing passed to a command is different -// than the version specified with the --version parameter. +// An exception thrown when the version of an entity specified with the expectedVersion +// parameter does not match the latest version in the system. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -11136,8 +11433,8 @@ func (c *IoT) UpdateThingGroupRequest(input *UpdateThingGroupInput) (req *reques // The request is not valid. // // * ErrCodeVersionConflictException "VersionConflictException" -// An exception thrown when the version of a thing passed to a command is different -// than the version specified with the --version parameter. +// An exception thrown when the version of an entity specified with the expectedVersion +// parameter does not match the latest version in the system. // // * ErrCodeThrottlingException "ThrottlingException" // The rate exceeds the limit. @@ -12436,12 +12733,137 @@ func (s CancelCertificateTransferOutput) GoString() string { return s.String() } +type CancelJobExecutionInput struct { + _ struct{} `type:"structure"` + + // (Optional) The expected current version of the job execution. Each time you + // update the job execution, its version is incremented. If the version of the + // job execution stored in Jobs does not match, the update is rejected with + // a VersionMismatch error, and an ErrorResponse that contains the current job + // execution status data is returned. (This makes it unnecessary to perform + // a separate DescribeJobExecution request in order to obtain the job execution + // status data.) + ExpectedVersion *int64 `locationName:"expectedVersion" type:"long"` + + // (Optional) If true the job execution will be canceled if it has status IN_PROGRESS + // or QUEUED, otherwise the job execution will be canceled only if it has status + // QUEUED. If you attempt to cancel a job execution that is IN_PROGRESS, and + // you do not set force to true, then an InvalidStateTransitionException will + // be thrown. The default is false. + // + // Canceling a job execution which is "IN_PROGRESS", will cause the device to + // be unable to update the job execution status. Use caution and ensure that + // the device is able to recover to a valid state. + Force *bool `location:"querystring" locationName:"force" type:"boolean"` + + // The ID of the job to be canceled. + // + // JobId is a required field + JobId *string `location:"uri" locationName:"jobId" min:"1" type:"string" required:"true"` + + // A collection of name/value pairs that describe the status of the job execution. + // If not specified, the statusDetails are unchanged. You can specify at most + // 10 name/value pairs. + StatusDetails map[string]*string `locationName:"statusDetails" type:"map"` + + // The name of the thing whose execution of the job will be canceled. + // + // ThingName is a required field + ThingName *string `location:"uri" locationName:"thingName" min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s CancelJobExecutionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CancelJobExecutionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CancelJobExecutionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CancelJobExecutionInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + if s.ThingName == nil { + invalidParams.Add(request.NewErrParamRequired("ThingName")) + } + if s.ThingName != nil && len(*s.ThingName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ThingName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetExpectedVersion sets the ExpectedVersion field's value. +func (s *CancelJobExecutionInput) SetExpectedVersion(v int64) *CancelJobExecutionInput { + s.ExpectedVersion = &v + return s +} + +// SetForce sets the Force field's value. +func (s *CancelJobExecutionInput) SetForce(v bool) *CancelJobExecutionInput { + s.Force = &v + return s +} + +// SetJobId sets the JobId field's value. +func (s *CancelJobExecutionInput) SetJobId(v string) *CancelJobExecutionInput { + s.JobId = &v + return s +} + +// SetStatusDetails sets the StatusDetails field's value. +func (s *CancelJobExecutionInput) SetStatusDetails(v map[string]*string) *CancelJobExecutionInput { + s.StatusDetails = v + return s +} + +// SetThingName sets the ThingName field's value. +func (s *CancelJobExecutionInput) SetThingName(v string) *CancelJobExecutionInput { + s.ThingName = &v + return s +} + +type CancelJobExecutionOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s CancelJobExecutionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CancelJobExecutionOutput) GoString() string { + return s.String() +} + type CancelJobInput struct { _ struct{} `type:"structure"` // An optional comment string describing why the job was canceled. Comment *string `locationName:"comment" type:"string"` + // (Optional) If true job executions with status "IN_PROGRESS" and "QUEUED" + // are canceled, otherwise only job executions with status "QUEUED" are canceled. + // The default is false. + // + // Canceling a job which is "IN_PROGRESS", will cause a device which is executing + // the job to be unable to update the job execution status. Use caution and + // ensure that each device executing a job which is canceled is able to recover + // to a valid state. + Force *bool `location:"querystring" locationName:"force" type:"boolean"` + // The unique identifier you assigned to this job when it was created. // // JobId is a required field @@ -12480,6 +12902,12 @@ func (s *CancelJobInput) SetComment(v string) *CancelJobInput { return s } +// SetForce sets the Force field's value. +func (s *CancelJobInput) SetForce(v bool) *CancelJobInput { + s.Force = &v + return s +} + // SetJobId sets the JobId field's value. func (s *CancelJobInput) SetJobId(v string) *CancelJobInput { s.JobId = &v @@ -14876,6 +15304,187 @@ func (s DeleteCertificateOutput) GoString() string { return s.String() } +type DeleteJobExecutionInput struct { + _ struct{} `type:"structure"` + + // The ID of the job execution to be deleted. The executionNumber refers to + // the execution of a particular job on a particular device. + // + // Note that once a job execution is deleted, the executionNumber may be reused + // by IoT, so be sure you get and use the correct value here. + // + // ExecutionNumber is a required field + ExecutionNumber *int64 `location:"uri" locationName:"executionNumber" type:"long" required:"true"` + + // (Optional) When true, you can delete a job execution which is "IN_PROGRESS". + // Otherwise, you can only delete a job execution which is in a terminal state + // ("SUCCEEDED", "FAILED", "REJECTED", "REMOVED" or "CANCELED") or an exception + // will occur. The default is false. + // + // Deleting a job execution which is "IN_PROGRESS", will cause the device to + // be unable to access job information or update the job execution status. Use + // caution and ensure that the device is able to recover to a valid state. + Force *bool `location:"querystring" locationName:"force" type:"boolean"` + + // The ID of the job whose execution on a particular device will be deleted. + // + // JobId is a required field + JobId *string `location:"uri" locationName:"jobId" min:"1" type:"string" required:"true"` + + // The name of the thing whose job execution will be deleted. + // + // ThingName is a required field + ThingName *string `location:"uri" locationName:"thingName" min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteJobExecutionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteJobExecutionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteJobExecutionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteJobExecutionInput"} + if s.ExecutionNumber == nil { + invalidParams.Add(request.NewErrParamRequired("ExecutionNumber")) + } + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + if s.ThingName == nil { + invalidParams.Add(request.NewErrParamRequired("ThingName")) + } + if s.ThingName != nil && len(*s.ThingName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ThingName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetExecutionNumber sets the ExecutionNumber field's value. +func (s *DeleteJobExecutionInput) SetExecutionNumber(v int64) *DeleteJobExecutionInput { + s.ExecutionNumber = &v + return s +} + +// SetForce sets the Force field's value. +func (s *DeleteJobExecutionInput) SetForce(v bool) *DeleteJobExecutionInput { + s.Force = &v + return s +} + +// SetJobId sets the JobId field's value. +func (s *DeleteJobExecutionInput) SetJobId(v string) *DeleteJobExecutionInput { + s.JobId = &v + return s +} + +// SetThingName sets the ThingName field's value. +func (s *DeleteJobExecutionInput) SetThingName(v string) *DeleteJobExecutionInput { + s.ThingName = &v + return s +} + +type DeleteJobExecutionOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeleteJobExecutionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteJobExecutionOutput) GoString() string { + return s.String() +} + +type DeleteJobInput struct { + _ struct{} `type:"structure"` + + // (Optional) When true, you can delete a job which is "IN_PROGRESS". Otherwise, + // you can only delete a job which is in a terminal state ("COMPLETED" or "CANCELED") + // or an exception will occur. The default is false. + // + // Deleting a job which is "IN_PROGRESS", will cause a device which is executing + // the job to be unable to access job information or update the job execution + // status. Use caution and ensure that each device executing a job which is + // deleted is able to recover to a valid state. + Force *bool `location:"querystring" locationName:"force" type:"boolean"` + + // The ID of the job to be deleted. + // + // After a job deletion is completed, you may reuse this jobId when you create + // a new job. However, this is not recommended, and you must ensure that your + // devices are not using the jobId to refer to the deleted job. + // + // JobId is a required field + JobId *string `location:"uri" locationName:"jobId" min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteJobInput"} + if s.JobId == nil { + invalidParams.Add(request.NewErrParamRequired("JobId")) + } + if s.JobId != nil && len(*s.JobId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("JobId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetForce sets the Force field's value. +func (s *DeleteJobInput) SetForce(v bool) *DeleteJobInput { + s.Force = &v + return s +} + +// SetJobId sets the JobId field's value. +func (s *DeleteJobInput) SetJobId(v string) *DeleteJobInput { + s.JobId = &v + return s +} + +type DeleteJobOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeleteJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteJobOutput) GoString() string { + return s.String() +} + type DeleteOTAUpdateInput struct { _ struct{} `type:"structure"` @@ -18410,7 +19019,7 @@ func (s *ImplicitDeny) SetPolicies(v []*Policy) *ImplicitDeny { return s } -// Sends message data to an AWS IoT Analytics channel. +// Sends messge data to an AWS IoT Analytics channel. type IotAnalyticsAction struct { _ struct{} `type:"structure"` @@ -18421,8 +19030,8 @@ type IotAnalyticsAction struct { // The name of the IoT Analytics channel to which message data will be sent. ChannelName *string `locationName:"channelName" type:"string"` - // The ARN of the role which has a policy that grants IoT permission to send - // message data via IoT Analytics (iotanalytics:BatchPutMessage). + // The ARN of the role which has a policy that grants IoT Analytics permission + // to send message data via IoT Analytics (iotanalytics:BatchPutMessage). RoleArn *string `locationName:"roleArn" type:"string"` } @@ -18473,6 +19082,10 @@ type Job struct { // The parameters specified for the job document. DocumentParameters map[string]*string `locationName:"documentParameters" type:"map"` + // Will be true if the job was canceled with the optional force parameter set + // to true. + ForceCanceled *bool `locationName:"forceCanceled" type:"boolean"` + // An ARN identifying the job with format "arn:aws:iot:region:account:job/jobId". JobArn *string `locationName:"jobArn" type:"string"` @@ -18546,6 +19159,12 @@ func (s *Job) SetDocumentParameters(v map[string]*string) *Job { return s } +// SetForceCanceled sets the ForceCanceled field's value. +func (s *Job) SetForceCanceled(v bool) *Job { + s.ForceCanceled = &v + return s +} + // SetJobArn sets the JobArn field's value. func (s *Job) SetJobArn(v string) *Job { s.JobArn = &v @@ -18610,6 +19229,10 @@ type JobExecution struct { // which return or update job execution information. ExecutionNumber *int64 `locationName:"executionNumber" type:"long"` + // Will be true if the job execution was canceled with the optional force parameter + // set to true. + ForceCanceled *bool `locationName:"forceCanceled" type:"boolean"` + // The unique identifier you assigned to the job when it was created. JobId *string `locationName:"jobId" min:"1" type:"string"` @@ -18632,6 +19255,10 @@ type JobExecution struct { // The ARN of the thing on which the job execution is running. ThingArn *string `locationName:"thingArn" type:"string"` + + // The version of the job execution. Job execution versions are incremented + // each time they are updated by a device. + VersionNumber *int64 `locationName:"versionNumber" type:"long"` } // String returns the string representation @@ -18650,6 +19277,12 @@ func (s *JobExecution) SetExecutionNumber(v int64) *JobExecution { return s } +// SetForceCanceled sets the ForceCanceled field's value. +func (s *JobExecution) SetForceCanceled(v bool) *JobExecution { + s.ForceCanceled = &v + return s +} + // SetJobId sets the JobId field's value. func (s *JobExecution) SetJobId(v string) *JobExecution { s.JobId = &v @@ -18692,6 +19325,12 @@ func (s *JobExecution) SetThingArn(v string) *JobExecution { return s } +// SetVersionNumber sets the VersionNumber field's value. +func (s *JobExecution) SetVersionNumber(v int64) *JobExecution { + s.VersionNumber = &v + return s +} + // Details of the job execution status. type JobExecutionStatusDetails struct { _ struct{} `type:"structure"` @@ -18910,7 +19549,9 @@ type JobProcessDetails struct { // The number of things which successfully completed the job. NumberOfSucceededThings *int64 `locationName:"numberOfSucceededThings" type:"integer"` - // The devices on which the job is executing. + // The target devices to which the job execution is being rolled out. This value + // will be null after the job execution has finished rolling out to all the + // target devices. ProcessingTargets []*string `locationName:"processingTargets" type:"list"` } @@ -23952,9 +24593,9 @@ func (s SetV2LoggingOptionsOutput) GoString() string { type SnsAction struct { _ struct{} `type:"structure"` - // The message format of the message to publish. Optional. Accepted values are - // "JSON" and "RAW". The default value of the attribute is "RAW". SNS uses this - // setting to determine if the payload should be parsed and relevant platform-specific + // (Optional) The message format of the message to publish. Accepted values + // are "JSON" and "RAW". The default value of the attribute is "RAW". SNS uses + // this setting to determine if the payload should be parsed and relevant platform-specific // bits of the payload should be extracted. To read more about SNS message formats, // see http://docs.aws.amazon.com/sns/latest/dg/json-formats.html (http://docs.aws.amazon.com/sns/latest/dg/json-formats.html) // refer to their official documentation. @@ -26522,6 +27163,9 @@ const ( // JobStatusCompleted is a JobStatus enum value JobStatusCompleted = "COMPLETED" + + // JobStatusDeletionInProgress is a JobStatus enum value + JobStatusDeletionInProgress = "DELETION_IN_PROGRESS" ) const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/iot/errors.go b/vendor/github.com/aws/aws-sdk-go/service/iot/errors.go index 6edb10706..8320d782b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iot/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iot/errors.go @@ -73,10 +73,18 @@ const ( // The response is invalid. ErrCodeInvalidResponseException = "InvalidResponseException" + // ErrCodeInvalidStateTransitionException for service response error code + // "InvalidStateTransitionException". + // + // An attempt was made to change to an invalid state, for example by deleting + // a job or a job execution which is "IN_PROGRESS" without setting the force + // parameter. + ErrCodeInvalidStateTransitionException = "InvalidStateTransitionException" + // ErrCodeLimitExceededException for service response error code // "LimitExceededException". // - // The number of attached entities exceeds the limit. + // A limit has been exceeded. ErrCodeLimitExceededException = "LimitExceededException" // ErrCodeMalformedPolicyException for service response error code @@ -156,8 +164,8 @@ const ( // ErrCodeVersionConflictException for service response error code // "VersionConflictException". // - // An exception thrown when the version of a thing passed to a command is different - // than the version specified with the --version parameter. + // An exception thrown when the version of an entity specified with the expectedVersion + // parameter does not match the latest version in the system. ErrCodeVersionConflictException = "VersionConflictException" // ErrCodeVersionsLimitExceededException for service response error code diff --git a/vendor/github.com/aws/aws-sdk-go/service/iot/iotiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/iot/iotiface/interface.go index 4251c9783..e4a98aaa6 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iot/iotiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iot/iotiface/interface.go @@ -92,6 +92,10 @@ type IoTAPI interface { CancelJobWithContext(aws.Context, *iot.CancelJobInput, ...request.Option) (*iot.CancelJobOutput, error) CancelJobRequest(*iot.CancelJobInput) (*request.Request, *iot.CancelJobOutput) + CancelJobExecution(*iot.CancelJobExecutionInput) (*iot.CancelJobExecutionOutput, error) + CancelJobExecutionWithContext(aws.Context, *iot.CancelJobExecutionInput, ...request.Option) (*iot.CancelJobExecutionOutput, error) + CancelJobExecutionRequest(*iot.CancelJobExecutionInput) (*request.Request, *iot.CancelJobExecutionOutput) + ClearDefaultAuthorizer(*iot.ClearDefaultAuthorizerInput) (*iot.ClearDefaultAuthorizerOutput, error) ClearDefaultAuthorizerWithContext(aws.Context, *iot.ClearDefaultAuthorizerInput, ...request.Option) (*iot.ClearDefaultAuthorizerOutput, error) ClearDefaultAuthorizerRequest(*iot.ClearDefaultAuthorizerInput) (*request.Request, *iot.ClearDefaultAuthorizerOutput) @@ -160,6 +164,14 @@ type IoTAPI interface { DeleteCertificateWithContext(aws.Context, *iot.DeleteCertificateInput, ...request.Option) (*iot.DeleteCertificateOutput, error) DeleteCertificateRequest(*iot.DeleteCertificateInput) (*request.Request, *iot.DeleteCertificateOutput) + DeleteJob(*iot.DeleteJobInput) (*iot.DeleteJobOutput, error) + DeleteJobWithContext(aws.Context, *iot.DeleteJobInput, ...request.Option) (*iot.DeleteJobOutput, error) + DeleteJobRequest(*iot.DeleteJobInput) (*request.Request, *iot.DeleteJobOutput) + + DeleteJobExecution(*iot.DeleteJobExecutionInput) (*iot.DeleteJobExecutionOutput, error) + DeleteJobExecutionWithContext(aws.Context, *iot.DeleteJobExecutionInput, ...request.Option) (*iot.DeleteJobExecutionOutput, error) + DeleteJobExecutionRequest(*iot.DeleteJobExecutionInput) (*request.Request, *iot.DeleteJobExecutionOutput) + DeleteOTAUpdate(*iot.DeleteOTAUpdateInput) (*iot.DeleteOTAUpdateOutput, error) DeleteOTAUpdateWithContext(aws.Context, *iot.DeleteOTAUpdateInput, ...request.Option) (*iot.DeleteOTAUpdateOutput, error) DeleteOTAUpdateRequest(*iot.DeleteOTAUpdateInput) (*request.Request, *iot.DeleteOTAUpdateOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/iot/service.go b/vendor/github.com/aws/aws-sdk-go/service/iot/service.go index 1f03372ac..10a95d560 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iot/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iot/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "iot" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "iot" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IoT" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IoT client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iot1clickdevicesservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/iot1clickdevicesservice/service.go index e9612dede..9614dd402 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iot1clickdevicesservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iot1clickdevicesservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "devices.iot1click" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "devices.iot1click" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IoT 1Click Devices Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IoT1ClickDevicesService client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iot1clickprojects/service.go b/vendor/github.com/aws/aws-sdk-go/service/iot1clickprojects/service.go index da0fa5eab..5aea2ea77 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iot1clickprojects/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iot1clickprojects/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "projects.iot1click" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "projects.iot1click" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IoT 1Click Projects" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IoT1ClickProjects client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/api.go b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/api.go index a2cd0b2e6..206b79621 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/api.go @@ -2260,6 +2260,100 @@ func (c *IoTAnalytics) ListPipelinesPagesWithContext(ctx aws.Context, input *Lis return p.Err() } +const opListTagsForResource = "ListTagsForResource" + +// ListTagsForResourceRequest generates a "aws/request.Request" representing the +// client's request for the ListTagsForResource operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListTagsForResource for more information on using the ListTagsForResource +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListTagsForResourceRequest method. +// req, resp := client.ListTagsForResourceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27/ListTagsForResource +func (c *IoTAnalytics) ListTagsForResourceRequest(input *ListTagsForResourceInput) (req *request.Request, output *ListTagsForResourceOutput) { + op := &request.Operation{ + Name: opListTagsForResource, + HTTPMethod: "GET", + HTTPPath: "/tags", + } + + if input == nil { + input = &ListTagsForResourceInput{} + } + + output = &ListTagsForResourceOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListTagsForResource API operation for AWS IoT Analytics. +// +// Lists the tags (metadata) which you have assigned to the resource. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS IoT Analytics's +// API operation ListTagsForResource for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request was not valid. +// +// * ErrCodeInternalFailureException "InternalFailureException" +// There was an internal failure. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is temporarily unavailable. +// +// * ErrCodeThrottlingException "ThrottlingException" +// The request was denied due to request throttling. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// The command caused an internal limit to be exceeded. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// A resource with the specified name could not be found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27/ListTagsForResource +func (c *IoTAnalytics) ListTagsForResource(input *ListTagsForResourceInput) (*ListTagsForResourceOutput, error) { + req, out := c.ListTagsForResourceRequest(input) + return out, req.Send() +} + +// ListTagsForResourceWithContext is the same as ListTagsForResource with the addition of +// the ability to pass a context and additional request options. +// +// See ListTagsForResource for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *IoTAnalytics) ListTagsForResourceWithContext(ctx aws.Context, input *ListTagsForResourceInput, opts ...request.Option) (*ListTagsForResourceOutput, error) { + req, out := c.ListTagsForResourceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opPutLoggingOptions = "PutLoggingOptions" // PutLoggingOptionsRequest generates a "aws/request.Request" representing the @@ -2308,6 +2402,12 @@ func (c *IoTAnalytics) PutLoggingOptionsRequest(input *PutLoggingOptionsInput) ( // // Sets or updates the AWS IoT Analytics logging options. // +// Note that if you update the value of any loggingOptions field, it takes up +// to one minute for the change to take effect. Also, if you change the policy +// attached to the role you specified in the roleArn field (for example, to +// correct an invalid policy) it takes up to 5 minutes for that change to take +// effect. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -2624,6 +2724,195 @@ func (c *IoTAnalytics) StartPipelineReprocessingWithContext(ctx aws.Context, inp return out, req.Send() } +const opTagResource = "TagResource" + +// TagResourceRequest generates a "aws/request.Request" representing the +// client's request for the TagResource operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See TagResource for more information on using the TagResource +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the TagResourceRequest method. +// req, resp := client.TagResourceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27/TagResource +func (c *IoTAnalytics) TagResourceRequest(input *TagResourceInput) (req *request.Request, output *TagResourceOutput) { + op := &request.Operation{ + Name: opTagResource, + HTTPMethod: "POST", + HTTPPath: "/tags", + } + + if input == nil { + input = &TagResourceInput{} + } + + output = &TagResourceOutput{} + req = c.newRequest(op, input, output) + return +} + +// TagResource API operation for AWS IoT Analytics. +// +// Adds to or modifies the tags of the given resource. Tags are metadata which +// can be used to manage a resource. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS IoT Analytics's +// API operation TagResource for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request was not valid. +// +// * ErrCodeInternalFailureException "InternalFailureException" +// There was an internal failure. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is temporarily unavailable. +// +// * ErrCodeThrottlingException "ThrottlingException" +// The request was denied due to request throttling. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// The command caused an internal limit to be exceeded. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// A resource with the specified name could not be found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27/TagResource +func (c *IoTAnalytics) TagResource(input *TagResourceInput) (*TagResourceOutput, error) { + req, out := c.TagResourceRequest(input) + return out, req.Send() +} + +// TagResourceWithContext is the same as TagResource with the addition of +// the ability to pass a context and additional request options. +// +// See TagResource for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *IoTAnalytics) TagResourceWithContext(ctx aws.Context, input *TagResourceInput, opts ...request.Option) (*TagResourceOutput, error) { + req, out := c.TagResourceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opUntagResource = "UntagResource" + +// UntagResourceRequest generates a "aws/request.Request" representing the +// client's request for the UntagResource operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See UntagResource for more information on using the UntagResource +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the UntagResourceRequest method. +// req, resp := client.UntagResourceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27/UntagResource +func (c *IoTAnalytics) UntagResourceRequest(input *UntagResourceInput) (req *request.Request, output *UntagResourceOutput) { + op := &request.Operation{ + Name: opUntagResource, + HTTPMethod: "DELETE", + HTTPPath: "/tags", + } + + if input == nil { + input = &UntagResourceInput{} + } + + output = &UntagResourceOutput{} + req = c.newRequest(op, input, output) + return +} + +// UntagResource API operation for AWS IoT Analytics. +// +// Removes the given tags (metadata) from the resource. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS IoT Analytics's +// API operation UntagResource for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidRequestException "InvalidRequestException" +// The request was not valid. +// +// * ErrCodeInternalFailureException "InternalFailureException" +// There was an internal failure. +// +// * ErrCodeServiceUnavailableException "ServiceUnavailableException" +// The service is temporarily unavailable. +// +// * ErrCodeThrottlingException "ThrottlingException" +// The request was denied due to request throttling. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// The command caused an internal limit to be exceeded. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// A resource with the specified name could not be found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27/UntagResource +func (c *IoTAnalytics) UntagResource(input *UntagResourceInput) (*UntagResourceOutput, error) { + req, out := c.UntagResourceRequest(input) + return out, req.Send() +} + +// UntagResourceWithContext is the same as UntagResource with the addition of +// the ability to pass a context and additional request options. +// +// See UntagResource for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *IoTAnalytics) UntagResourceWithContext(ctx aws.Context, input *UntagResourceInput, opts ...request.Option) (*UntagResourceOutput, error) { + req, out := c.UntagResourceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opUpdateChannel = "UpdateChannel" // UpdateChannelRequest generates a "aws/request.Request" representing the @@ -3478,6 +3767,9 @@ type CreateChannelInput struct { // How long, in days, message data is kept for the channel. RetentionPeriod *RetentionPeriod `locationName:"retentionPeriod" type:"structure"` + + // Metadata which can be used to manage the channel. + Tags []*Tag `locationName:"tags" min:"1" type:"list"` } // String returns the string representation @@ -3499,11 +3791,24 @@ func (s *CreateChannelInput) Validate() error { if s.ChannelName != nil && len(*s.ChannelName) < 1 { invalidParams.Add(request.NewErrParamMinLen("ChannelName", 1)) } + if s.Tags != nil && len(s.Tags) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Tags", 1)) + } if s.RetentionPeriod != nil { if err := s.RetentionPeriod.Validate(); err != nil { invalidParams.AddNested("RetentionPeriod", err.(request.ErrInvalidParams)) } } + if s.Tags != nil { + for i, v := range s.Tags { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Tags", i), err.(request.ErrInvalidParams)) + } + } + } if invalidParams.Len() > 0 { return invalidParams @@ -3523,6 +3828,12 @@ func (s *CreateChannelInput) SetRetentionPeriod(v *RetentionPeriod) *CreateChann return s } +// SetTags sets the Tags field's value. +func (s *CreateChannelInput) SetTags(v []*Tag) *CreateChannelInput { + s.Tags = v + return s +} + type CreateChannelOutput struct { _ struct{} `type:"structure"` @@ -3633,6 +3944,9 @@ type CreateDatasetInput struct { // DatasetName is a required field DatasetName *string `locationName:"datasetName" min:"1" type:"string" required:"true"` + // Metadata which can be used to manage the data set. + Tags []*Tag `locationName:"tags" min:"1" type:"list"` + // A list of triggers. A trigger causes data set content to be populated at // a specified time or time interval. The list of triggers can be empty or contain // up to five DataSetTrigger objects. @@ -3664,6 +3978,9 @@ func (s *CreateDatasetInput) Validate() error { if s.DatasetName != nil && len(*s.DatasetName) < 1 { invalidParams.Add(request.NewErrParamMinLen("DatasetName", 1)) } + if s.Tags != nil && len(s.Tags) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Tags", 1)) + } if s.Actions != nil { for i, v := range s.Actions { if v == nil { @@ -3674,6 +3991,16 @@ func (s *CreateDatasetInput) Validate() error { } } } + if s.Tags != nil { + for i, v := range s.Tags { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Tags", i), err.(request.ErrInvalidParams)) + } + } + } if invalidParams.Len() > 0 { return invalidParams @@ -3693,6 +4020,12 @@ func (s *CreateDatasetInput) SetDatasetName(v string) *CreateDatasetInput { return s } +// SetTags sets the Tags field's value. +func (s *CreateDatasetInput) SetTags(v []*Tag) *CreateDatasetInput { + s.Tags = v + return s +} + // SetTriggers sets the Triggers field's value. func (s *CreateDatasetInput) SetTriggers(v []*DatasetTrigger) *CreateDatasetInput { s.Triggers = v @@ -3741,6 +4074,9 @@ type CreateDatastoreInput struct { // How long, in days, message data is kept for the data store. RetentionPeriod *RetentionPeriod `locationName:"retentionPeriod" type:"structure"` + + // Metadata which can be used to manage the data store. + Tags []*Tag `locationName:"tags" min:"1" type:"list"` } // String returns the string representation @@ -3762,11 +4098,24 @@ func (s *CreateDatastoreInput) Validate() error { if s.DatastoreName != nil && len(*s.DatastoreName) < 1 { invalidParams.Add(request.NewErrParamMinLen("DatastoreName", 1)) } + if s.Tags != nil && len(s.Tags) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Tags", 1)) + } if s.RetentionPeriod != nil { if err := s.RetentionPeriod.Validate(); err != nil { invalidParams.AddNested("RetentionPeriod", err.(request.ErrInvalidParams)) } } + if s.Tags != nil { + for i, v := range s.Tags { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Tags", i), err.(request.ErrInvalidParams)) + } + } + } if invalidParams.Len() > 0 { return invalidParams @@ -3786,6 +4135,12 @@ func (s *CreateDatastoreInput) SetRetentionPeriod(v *RetentionPeriod) *CreateDat return s } +// SetTags sets the Tags field's value. +func (s *CreateDatastoreInput) SetTags(v []*Tag) *CreateDatastoreInput { + s.Tags = v + return s +} + type CreateDatastoreOutput struct { _ struct{} `type:"structure"` @@ -3845,6 +4200,9 @@ type CreatePipelineInput struct { // // PipelineName is a required field PipelineName *string `locationName:"pipelineName" min:"1" type:"string" required:"true"` + + // Metadata which can be used to manage the pipeline. + Tags []*Tag `locationName:"tags" min:"1" type:"list"` } // String returns the string representation @@ -3872,6 +4230,9 @@ func (s *CreatePipelineInput) Validate() error { if s.PipelineName != nil && len(*s.PipelineName) < 1 { invalidParams.Add(request.NewErrParamMinLen("PipelineName", 1)) } + if s.Tags != nil && len(s.Tags) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Tags", 1)) + } if s.PipelineActivities != nil { for i, v := range s.PipelineActivities { if v == nil { @@ -3882,6 +4243,16 @@ func (s *CreatePipelineInput) Validate() error { } } } + if s.Tags != nil { + for i, v := range s.Tags { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Tags", i), err.(request.ErrInvalidParams)) + } + } + } if invalidParams.Len() > 0 { return invalidParams @@ -3901,6 +4272,12 @@ func (s *CreatePipelineInput) SetPipelineName(v string) *CreatePipelineInput { return s } +// SetTags sets the Tags field's value. +func (s *CreatePipelineInput) SetTags(v []*Tag) *CreatePipelineInput { + s.Tags = v + return s +} + type CreatePipelineOutput struct { _ struct{} `type:"structure"` @@ -5757,6 +6134,70 @@ func (s *ListPipelinesOutput) SetPipelineSummaries(v []*PipelineSummary) *ListPi return s } +type ListTagsForResourceInput struct { + _ struct{} `type:"structure"` + + // The ARN of the resource whose tags you want to list. + // + // ResourceArn is a required field + ResourceArn *string `location:"querystring" locationName:"resourceArn" min:"20" type:"string" required:"true"` +} + +// String returns the string representation +func (s ListTagsForResourceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListTagsForResourceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListTagsForResourceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListTagsForResourceInput"} + if s.ResourceArn == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceArn")) + } + if s.ResourceArn != nil && len(*s.ResourceArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("ResourceArn", 20)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetResourceArn sets the ResourceArn field's value. +func (s *ListTagsForResourceInput) SetResourceArn(v string) *ListTagsForResourceInput { + s.ResourceArn = &v + return s +} + +type ListTagsForResourceOutput struct { + _ struct{} `type:"structure"` + + // The tags (metadata) which you have assigned to the resource. + Tags []*Tag `locationName:"tags" min:"1" type:"list"` +} + +// String returns the string representation +func (s ListTagsForResourceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListTagsForResourceOutput) GoString() string { + return s.String() +} + +// SetTags sets the Tags field's value. +func (s *ListTagsForResourceOutput) SetTags(v []*Tag) *ListTagsForResourceOutput { + s.Tags = v + return s +} + // Information about logging options. type LoggingOptions struct { _ struct{} `type:"structure"` @@ -5921,7 +6362,8 @@ func (s *MathActivity) SetNext(v string) *MathActivity { type Message struct { _ struct{} `type:"structure"` - // The ID you wish to assign to the message. + // The ID you wish to assign to the message. Each "messageId" must be unique + // within each batch sent. // // MessageId is a required field MessageId *string `locationName:"messageId" min:"1" type:"string" required:"true"` @@ -6892,6 +7334,219 @@ func (s *StartPipelineReprocessingOutput) SetReprocessingId(v string) *StartPipe return s } +// A set of key/value pairs which are used to manage the resource. +type Tag struct { + _ struct{} `type:"structure"` + + // The tag's key. + // + // Key is a required field + Key *string `locationName:"key" min:"1" type:"string" required:"true"` + + // The tag's value. + // + // Value is a required field + Value *string `locationName:"value" min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s Tag) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Tag) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *Tag) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Tag"} + if s.Key == nil { + invalidParams.Add(request.NewErrParamRequired("Key")) + } + if s.Key != nil && len(*s.Key) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Key", 1)) + } + if s.Value == nil { + invalidParams.Add(request.NewErrParamRequired("Value")) + } + if s.Value != nil && len(*s.Value) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Value", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetKey sets the Key field's value. +func (s *Tag) SetKey(v string) *Tag { + s.Key = &v + return s +} + +// SetValue sets the Value field's value. +func (s *Tag) SetValue(v string) *Tag { + s.Value = &v + return s +} + +type TagResourceInput struct { + _ struct{} `type:"structure"` + + // The ARN of the resource whose tags will be modified. + // + // ResourceArn is a required field + ResourceArn *string `location:"querystring" locationName:"resourceArn" min:"20" type:"string" required:"true"` + + // The new or modified tags for the resource. + // + // Tags is a required field + Tags []*Tag `locationName:"tags" min:"1" type:"list" required:"true"` +} + +// String returns the string representation +func (s TagResourceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s TagResourceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *TagResourceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "TagResourceInput"} + if s.ResourceArn == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceArn")) + } + if s.ResourceArn != nil && len(*s.ResourceArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("ResourceArn", 20)) + } + if s.Tags == nil { + invalidParams.Add(request.NewErrParamRequired("Tags")) + } + if s.Tags != nil && len(s.Tags) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Tags", 1)) + } + if s.Tags != nil { + for i, v := range s.Tags { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Tags", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetResourceArn sets the ResourceArn field's value. +func (s *TagResourceInput) SetResourceArn(v string) *TagResourceInput { + s.ResourceArn = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *TagResourceInput) SetTags(v []*Tag) *TagResourceInput { + s.Tags = v + return s +} + +type TagResourceOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s TagResourceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s TagResourceOutput) GoString() string { + return s.String() +} + +type UntagResourceInput struct { + _ struct{} `type:"structure"` + + // The ARN of the resource whose tags will be removed. + // + // ResourceArn is a required field + ResourceArn *string `location:"querystring" locationName:"resourceArn" min:"20" type:"string" required:"true"` + + // The keys of those tags which will be removed. + // + // TagKeys is a required field + TagKeys []*string `location:"querystring" locationName:"tagKeys" min:"1" type:"list" required:"true"` +} + +// String returns the string representation +func (s UntagResourceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UntagResourceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *UntagResourceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "UntagResourceInput"} + if s.ResourceArn == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceArn")) + } + if s.ResourceArn != nil && len(*s.ResourceArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("ResourceArn", 20)) + } + if s.TagKeys == nil { + invalidParams.Add(request.NewErrParamRequired("TagKeys")) + } + if s.TagKeys != nil && len(s.TagKeys) < 1 { + invalidParams.Add(request.NewErrParamMinLen("TagKeys", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetResourceArn sets the ResourceArn field's value. +func (s *UntagResourceInput) SetResourceArn(v string) *UntagResourceInput { + s.ResourceArn = &v + return s +} + +// SetTagKeys sets the TagKeys field's value. +func (s *UntagResourceInput) SetTagKeys(v []*string) *UntagResourceInput { + s.TagKeys = v + return s +} + +type UntagResourceOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s UntagResourceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UntagResourceOutput) GoString() string { + return s.String() +} + type UpdateChannelInput struct { _ struct{} `type:"structure"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/doc.go b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/doc.go index ab94d9b75..9ea1c7e8c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/doc.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/doc.go @@ -3,30 +3,29 @@ // Package iotanalytics provides the client and types for making API // requests to AWS IoT Analytics. // -// AWS IoT Analytics provides advanced data analysis for AWS IoT. It allows -// you to collect large amounts of device data, process messages, store them, -// and then query the data and run sophisticated analytics to make accurate -// decisions in your IoT applications and machine learning use cases. AWS IoT -// Analytics enables advanced data exploration through integration with Jupyter -// Notebooks and data visualization through integration with Amazon QuickSight. +// AWS IoT Analytics allows you to collect large amounts of device data, process +// messages, and store them. You can then query the data and run sophisticated +// analytics on it. AWS IoT Analytics enables advanced data exploration through +// integration with Jupyter Notebooks and data visualization through integration +// with Amazon QuickSight. // // Traditional analytics and business intelligence tools are designed to process // structured data. IoT data often comes from devices that record noisy processes -// (such as temperature, motion, or sound). As a result, the data from these +// (such as temperature, motion, or sound). As a result the data from these // devices can have significant gaps, corrupted messages, and false readings // that must be cleaned up before analysis can occur. Also, IoT data is often // only meaningful in the context of other data from external sources. // -// AWS IoT Analytics automates each of the steps required to analyze data from -// IoT devices. AWS IoT Analytics filters, transforms, and enriches IoT data -// before storing it in a time-series data store for analysis. You can set up -// the service to collect only the data you need from your devices, apply mathematical -// transforms to process the data, and enrich the data with device-specific -// metadata such as device type and location before storing it. Then, you can -// analyze your data by running queries using the built-in SQL query engine, -// or perform more complex analytics and machine learning inference. AWS IoT -// Analytics includes models for common IoT use cases so you can answer questions -// like which devices are about to fail or which customers are at risk of abandoning +// AWS IoT Analytics automates the steps required to analyze data from IoT devices. +// AWS IoT Analytics filters, transforms, and enriches IoT data before storing +// it in a time-series data store for analysis. You can set up the service to +// collect only the data you need from your devices, apply mathematical transforms +// to process the data, and enrich the data with device-specific metadata such +// as device type and location before storing it. Then, you can analyze your +// data by running queries using the built-in SQL query engine, or perform more +// complex analytics and machine learning inference. AWS IoT Analytics includes +// pre-built models for common IoT use cases so you can answer questions like +// which devices are about to fail or which customers are at risk of abandoning // their wearable devices. // // See https://docs.aws.amazon.com/goto/WebAPI/iotanalytics-2017-11-27 for more information on this service. diff --git a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/iotanalyticsiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/iotanalyticsiface/interface.go index 2f8665d4f..35c4b1dcb 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/iotanalyticsiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/iotanalyticsiface/interface.go @@ -160,6 +160,10 @@ type IoTAnalyticsAPI interface { ListPipelinesPages(*iotanalytics.ListPipelinesInput, func(*iotanalytics.ListPipelinesOutput, bool) bool) error ListPipelinesPagesWithContext(aws.Context, *iotanalytics.ListPipelinesInput, func(*iotanalytics.ListPipelinesOutput, bool) bool, ...request.Option) error + ListTagsForResource(*iotanalytics.ListTagsForResourceInput) (*iotanalytics.ListTagsForResourceOutput, error) + ListTagsForResourceWithContext(aws.Context, *iotanalytics.ListTagsForResourceInput, ...request.Option) (*iotanalytics.ListTagsForResourceOutput, error) + ListTagsForResourceRequest(*iotanalytics.ListTagsForResourceInput) (*request.Request, *iotanalytics.ListTagsForResourceOutput) + PutLoggingOptions(*iotanalytics.PutLoggingOptionsInput) (*iotanalytics.PutLoggingOptionsOutput, error) PutLoggingOptionsWithContext(aws.Context, *iotanalytics.PutLoggingOptionsInput, ...request.Option) (*iotanalytics.PutLoggingOptionsOutput, error) PutLoggingOptionsRequest(*iotanalytics.PutLoggingOptionsInput) (*request.Request, *iotanalytics.PutLoggingOptionsOutput) @@ -176,6 +180,14 @@ type IoTAnalyticsAPI interface { StartPipelineReprocessingWithContext(aws.Context, *iotanalytics.StartPipelineReprocessingInput, ...request.Option) (*iotanalytics.StartPipelineReprocessingOutput, error) StartPipelineReprocessingRequest(*iotanalytics.StartPipelineReprocessingInput) (*request.Request, *iotanalytics.StartPipelineReprocessingOutput) + TagResource(*iotanalytics.TagResourceInput) (*iotanalytics.TagResourceOutput, error) + TagResourceWithContext(aws.Context, *iotanalytics.TagResourceInput, ...request.Option) (*iotanalytics.TagResourceOutput, error) + TagResourceRequest(*iotanalytics.TagResourceInput) (*request.Request, *iotanalytics.TagResourceOutput) + + UntagResource(*iotanalytics.UntagResourceInput) (*iotanalytics.UntagResourceOutput, error) + UntagResourceWithContext(aws.Context, *iotanalytics.UntagResourceInput, ...request.Option) (*iotanalytics.UntagResourceOutput, error) + UntagResourceRequest(*iotanalytics.UntagResourceInput) (*request.Request, *iotanalytics.UntagResourceOutput) + UpdateChannel(*iotanalytics.UpdateChannelInput) (*iotanalytics.UpdateChannelOutput, error) UpdateChannelWithContext(aws.Context, *iotanalytics.UpdateChannelInput, ...request.Option) (*iotanalytics.UpdateChannelOutput, error) UpdateChannelRequest(*iotanalytics.UpdateChannelInput) (*request.Request, *iotanalytics.UpdateChannelOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/service.go b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/service.go index 7e49b0ea0..54edbd56f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iotanalytics/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "iotanalytics" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "iotanalytics" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IoTAnalytics" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IoTAnalytics client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iotdataplane/service.go b/vendor/github.com/aws/aws-sdk-go/service/iotdataplane/service.go index 684d96a0f..a7318b4d7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iotdataplane/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iotdataplane/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "data.iot" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "data.iot" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IoT Data Plane" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IoTDataPlane client with a session. @@ -63,6 +64,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/iotjobsdataplane/service.go b/vendor/github.com/aws/aws-sdk-go/service/iotjobsdataplane/service.go index 7dfba8c32..bd84865cc 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/iotjobsdataplane/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/iotjobsdataplane/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "data.jobs.iot" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "data.jobs.iot" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "IoT Jobs Data Plane" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the IoTJobsDataPlane client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/kinesis/service.go b/vendor/github.com/aws/aws-sdk-go/service/kinesis/service.go index 17a59119a..2d55ac957 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kinesis/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kinesis/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "kinesis" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "kinesis" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Kinesis" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Kinesis client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/kinesisanalytics/service.go b/vendor/github.com/aws/aws-sdk-go/service/kinesisanalytics/service.go index 1a8d574d4..153daad6e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kinesisanalytics/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kinesisanalytics/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "kinesisanalytics" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "kinesisanalytics" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Kinesis Analytics" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the KinesisAnalytics client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/kinesisvideo/service.go b/vendor/github.com/aws/aws-sdk-go/service/kinesisvideo/service.go index 7f16ac919..a723794cd 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kinesisvideo/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kinesisvideo/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "kinesisvideo" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "kinesisvideo" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Kinesis Video" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the KinesisVideo client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/kinesisvideoarchivedmedia/service.go b/vendor/github.com/aws/aws-sdk-go/service/kinesisvideoarchivedmedia/service.go index 2f967f7e2..2da92bd3b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kinesisvideoarchivedmedia/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kinesisvideoarchivedmedia/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "kinesisvideo" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "kinesisvideo" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Kinesis Video Archived Media" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the KinesisVideoArchivedMedia client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/kinesisvideomedia/service.go b/vendor/github.com/aws/aws-sdk-go/service/kinesisvideomedia/service.go index a06b6175c..ff9d2163e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kinesisvideomedia/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kinesisvideomedia/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "kinesisvideo" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "kinesisvideo" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Kinesis Video Media" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the KinesisVideoMedia client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/kms/api.go b/vendor/github.com/aws/aws-sdk-go/service/kms/api.go index eae393011..48efb9dd4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kms/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kms/api.go @@ -85,11 +85,11 @@ func (c *KMS) CancelKeyDeletionRequest(input *CancelKeyDeletionInput) (req *requ // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -212,7 +212,7 @@ func (c *KMS) CreateAliasRequest(input *CreateAliasInput) (req *request.Request, // * ErrCodeInvalidAliasNameException "InvalidAliasNameException" // The request was rejected because the specified alias name is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -221,7 +221,7 @@ func (c *KMS) CreateAliasRequest(input *CreateAliasInput) (req *request.Request, // see Limits (http://docs.aws.amazon.com/kms/latest/developerguide/limits.html) // in the AWS Key Management Service Developer Guide. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -326,7 +326,7 @@ func (c *KMS) CreateGrantRequest(input *CreateGrantInput) (req *request.Request, // * ErrCodeInvalidArnException "InvalidArnException" // The request was rejected because a specified ARN was not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -338,7 +338,7 @@ func (c *KMS) CreateGrantRequest(input *CreateGrantInput) (req *request.Request, // see Limits (http://docs.aws.amazon.com/kms/latest/developerguide/limits.html) // in the AWS Key Management Service Developer Guide. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -449,7 +449,7 @@ func (c *KMS) CreateKeyRequest(input *CreateKeyInput) (req *request.Request, out // The request was rejected because a specified parameter is not supported or // a specified resource is not valid for this operation. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -576,11 +576,11 @@ func (c *KMS) DecryptRequest(input *DecryptInput) (req *request.Request, output // * ErrCodeInvalidGrantTokenException "InvalidGrantTokenException" // The request was rejected because the specified grant token is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -684,11 +684,11 @@ func (c *KMS) DeleteAliasRequest(input *DeleteAliasInput) (req *request.Request, // The request was rejected because the specified entity or resource could not // be found. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -799,11 +799,11 @@ func (c *KMS) DeleteImportedKeyMaterialRequest(input *DeleteImportedKeyMaterialI // The request was rejected because the specified entity or resource could not // be found. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -901,7 +901,7 @@ func (c *KMS) DescribeKeyRequest(input *DescribeKeyInput) (req *request.Request, // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -1000,11 +1000,11 @@ func (c *KMS) DisableKeyRequest(input *DisableKeyInput) (req *request.Request, o // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1106,11 +1106,11 @@ func (c *KMS) DisableKeyRotationRequest(input *DisableKeyRotationInput) (req *re // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1213,7 +1213,7 @@ func (c *KMS) EnableKeyRequest(input *EnableKeyInput) (req *request.Request, out // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -1222,7 +1222,7 @@ func (c *KMS) EnableKeyRequest(input *EnableKeyInput) (req *request.Request, out // see Limits (http://docs.aws.amazon.com/kms/latest/developerguide/limits.html) // in the AWS Key Management Service Developer Guide. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1324,11 +1324,11 @@ func (c *KMS) EnableKeyRotationRequest(input *EnableKeyRotationInput) (req *requ // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1459,11 +1459,11 @@ func (c *KMS) EncryptRequest(input *EncryptInput) (req *request.Request, output // * ErrCodeInvalidGrantTokenException "InvalidGrantTokenException" // The request was rejected because the specified grant token is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1613,11 +1613,11 @@ func (c *KMS) GenerateDataKeyRequest(input *GenerateDataKeyInput) (req *request. // * ErrCodeInvalidGrantTokenException "InvalidGrantTokenException" // The request was rejected because the specified grant token is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1739,11 +1739,11 @@ func (c *KMS) GenerateDataKeyWithoutPlaintextRequest(input *GenerateDataKeyWitho // * ErrCodeInvalidGrantTokenException "InvalidGrantTokenException" // The request was rejected because the specified grant token is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -1835,7 +1835,7 @@ func (c *KMS) GenerateRandomRequest(input *GenerateRandomInput) (req *request.Re // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -1927,11 +1927,11 @@ func (c *KMS) GetKeyPolicyRequest(input *GetKeyPolicyInput) (req *request.Reques // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -2030,11 +2030,11 @@ func (c *KMS) GetKeyRotationStatusRequest(input *GetKeyRotationStatusInput) (req // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -2153,11 +2153,11 @@ func (c *KMS) GetParametersForImportRequest(input *GetParametersForImportInput) // The request was rejected because the specified entity or resource could not // be found. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -2291,11 +2291,11 @@ func (c *KMS) ImportKeyMaterialRequest(input *ImportKeyMaterialInput) (req *requ // The request was rejected because the specified entity or resource could not // be found. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -2419,7 +2419,7 @@ func (c *KMS) ListAliasesRequest(input *ListAliasesInput) (req *request.Request, // The request was rejected because the marker that specifies where pagination // should next begin is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -2573,11 +2573,11 @@ func (c *KMS) ListGrantsRequest(input *ListGrantsInput) (req *request.Request, o // * ErrCodeInvalidArnException "InvalidArnException" // The request was rejected because a specified ARN was not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -2731,11 +2731,11 @@ func (c *KMS) ListKeyPoliciesRequest(input *ListKeyPoliciesInput) (req *request. // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -2880,7 +2880,7 @@ func (c *KMS) ListKeysRequest(input *ListKeysInput) (req *request.Request, outpu // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -3016,7 +3016,7 @@ func (c *KMS) ListResourceTagsRequest(input *ListResourceTagsInput) (req *reques // API operation ListResourceTags for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -3126,7 +3126,7 @@ func (c *KMS) ListRetirableGrantsRequest(input *ListRetirableGrantsInput) (req * // The request was rejected because the specified entity or resource could not // be found. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -3231,7 +3231,7 @@ func (c *KMS) PutKeyPolicyRequest(input *PutKeyPolicyInput) (req *request.Reques // The request was rejected because a specified parameter is not supported or // a specified resource is not valid for this operation. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -3240,7 +3240,7 @@ func (c *KMS) PutKeyPolicyRequest(input *PutKeyPolicyInput) (req *request.Reques // see Limits (http://docs.aws.amazon.com/kms/latest/developerguide/limits.html) // in the AWS Key Management Service Developer Guide. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -3363,11 +3363,11 @@ func (c *KMS) ReEncryptRequest(input *ReEncryptInput) (req *request.Request, out // * ErrCodeInvalidGrantTokenException "InvalidGrantTokenException" // The request was rejected because the specified grant token is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -3485,11 +3485,11 @@ func (c *KMS) RetireGrantRequest(input *RetireGrantInput) (req *request.Request, // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -3593,11 +3593,11 @@ func (c *KMS) RevokeGrantRequest(input *RevokeGrantInput) (req *request.Request, // * ErrCodeInvalidGrantIdException "InvalidGrantIdException" // The request was rejected because the specified GrantId is not valid. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -3709,11 +3709,11 @@ func (c *KMS) ScheduleKeyDeletionRequest(input *ScheduleKeyDeletionInput) (req * // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -3813,7 +3813,7 @@ func (c *KMS) TagResourceRequest(input *TagResourceInput) (req *request.Request, // API operation TagResource for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -3824,7 +3824,7 @@ func (c *KMS) TagResourceRequest(input *TagResourceInput) (req *request.Request, // * ErrCodeInvalidArnException "InvalidArnException" // The request was rejected because a specified ARN was not valid. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -3923,7 +3923,7 @@ func (c *KMS) UntagResourceRequest(input *UntagResourceInput) (req *request.Requ // API operation UntagResource for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // @@ -3934,7 +3934,7 @@ func (c *KMS) UntagResourceRequest(input *UntagResourceInput) (req *request.Requ // * ErrCodeInvalidArnException "InvalidArnException" // The request was rejected because a specified ARN was not valid. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -4050,11 +4050,11 @@ func (c *KMS) UpdateAliasRequest(input *UpdateAliasInput) (req *request.Request, // The request was rejected because the specified entity or resource could not // be found. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // @@ -4154,11 +4154,11 @@ func (c *KMS) UpdateKeyDescriptionRequest(input *UpdateKeyDescriptionInput) (req // The system timed out while trying to fulfill the request. The request can // be retried. // -// * ErrCodeInternalException "InternalException" +// * ErrCodeInternalException "KMSInternalException" // The request was rejected because an internal exception occurred. The request // can be retried. // -// * ErrCodeInvalidStateException "InvalidStateException" +// * ErrCodeInvalidStateException "KMSInvalidStateException" // The request was rejected because the state of the specified resource is not // valid for this request. // diff --git a/vendor/github.com/aws/aws-sdk-go/service/kms/errors.go b/vendor/github.com/aws/aws-sdk-go/service/kms/errors.go index d79e4321b..2a6511da9 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kms/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kms/errors.go @@ -41,11 +41,11 @@ const ( ErrCodeIncorrectKeyMaterialException = "IncorrectKeyMaterialException" // ErrCodeInternalException for service response error code - // "InternalException". + // "KMSInternalException". // // The request was rejected because an internal exception occurred. The request // can be retried. - ErrCodeInternalException = "InternalException" + ErrCodeInternalException = "KMSInternalException" // ErrCodeInvalidAliasNameException for service response error code // "InvalidAliasNameException". @@ -100,7 +100,7 @@ const ( ErrCodeInvalidMarkerException = "InvalidMarkerException" // ErrCodeInvalidStateException for service response error code - // "InvalidStateException". + // "KMSInvalidStateException". // // The request was rejected because the state of the specified resource is not // valid for this request. @@ -108,7 +108,7 @@ const ( // For more information about how key state affects the use of a CMK, see How // Key State Affects Use of a Customer Master Key (http://docs.aws.amazon.com/kms/latest/developerguide/key-state.html) // in the AWS Key Management Service Developer Guide. - ErrCodeInvalidStateException = "InvalidStateException" + ErrCodeInvalidStateException = "KMSInvalidStateException" // ErrCodeKeyUnavailableException for service response error code // "KeyUnavailableException". diff --git a/vendor/github.com/aws/aws-sdk-go/service/kms/service.go b/vendor/github.com/aws/aws-sdk-go/service/kms/service.go index 3ff65de5e..6d062f32f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/kms/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/kms/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "kms" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "kms" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "KMS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the KMS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/lambda/api.go b/vendor/github.com/aws/aws-sdk-go/service/lambda/api.go index 8024623fe..8babcb686 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lambda/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lambda/api.go @@ -269,19 +269,21 @@ func (c *Lambda) CreateEventSourceMappingRequest(input *CreateEventSourceMapping // CreateEventSourceMapping API operation for AWS Lambda. // -// Identifies a stream as an event source for a Lambda function. It can be either -// an Amazon Kinesis stream or an Amazon DynamoDB stream. AWS Lambda invokes -// the specified function when records are posted to the stream. +// Identifies a poll-based event source for a Lambda function. It can be either +// an Amazon Kinesis or DynamoDB stream, or an Amazon SQS queue. AWS Lambda +// invokes the specified function when records are posted to the event source. // -// This association between a stream source and a Lambda function is called +// This association between a poll-based source and a Lambda function is called // the event source mapping. // -// You provide mapping information (for example, which stream to read from and -// which Lambda function to invoke) in the request body. +// You provide mapping information (for example, which stream or SQS queue to +// read from and which Lambda function to invoke) in the request body. // -// Each event source, such as an Amazon Kinesis or a DynamoDB stream, can be -// associated with multiple AWS Lambda functions. A given Lambda function can -// be associated with multiple AWS event sources. +// Amazon Kinesis or DynamoDB stream event sources can be associated with multiple +// AWS Lambda functions and a given Lambda function can be associated with multiple +// AWS event sources. For Amazon SQS, you can configure multiple queues as event +// sources for a single Lambda function, but an SQS queue can be mapped only +// to a single Lambda function. // // If you are using versioning, you can specify a specific function version // or an alias via the function name parameter. For more information about versioning, @@ -605,6 +607,11 @@ func (c *Lambda) DeleteEventSourceMappingRequest(input *DeleteEventSourceMapping // // * ErrCodeTooManyRequestsException "TooManyRequestsException" // +// * ErrCodeResourceInUseException "ResourceInUseException" +// The operation conflicts with the resource's availability. For example, you +// attempted to update an EventSoure Mapping in CREATING, or tried to delete +// a EventSoure mapping currently in the UPDATING state. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/lambda-2015-03-31/DeleteEventSourceMapping func (c *Lambda) DeleteEventSourceMapping(input *DeleteEventSourceMappingInput) (*EventSourceMappingConfiguration, error) { req, out := c.DeleteEventSourceMappingRequest(input) @@ -2935,6 +2942,11 @@ func (c *Lambda) UpdateEventSourceMappingRequest(input *UpdateEventSourceMapping // * ErrCodeResourceConflictException "ResourceConflictException" // The resource already exists. // +// * ErrCodeResourceInUseException "ResourceInUseException" +// The operation conflicts with the resource's availability. For example, you +// attempted to update an EventSoure Mapping in CREATING, or tried to delete +// a EventSoure mapping currently in the UPDATING state. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/lambda-2015-03-31/UpdateEventSourceMapping func (c *Lambda) UpdateEventSourceMapping(input *UpdateEventSourceMappingInput) (*EventSourceMappingConfiguration, error) { req, out := c.UpdateEventSourceMappingRequest(input) @@ -3684,18 +3696,18 @@ type CreateEventSourceMappingInput struct { // The largest number of records that AWS Lambda will retrieve from your event // source at the time of invoking your function. Your function receives an event - // with all the retrieved records. The default is 100 records. + // with all the retrieved records. The default for Amazon Kinesis and Amazon + // DynamoDB is 100 records. For SQS, the default is 1. BatchSize *int64 `min:"1" type:"integer"` // Indicates whether AWS Lambda should begin polling the event source. By default, // Enabled is true. Enabled *bool `type:"boolean"` - // The Amazon Resource Name (ARN) of the Amazon Kinesis or the Amazon DynamoDB - // stream that is the event source. Any record added to this stream could cause - // AWS Lambda to invoke your Lambda function, it depends on the BatchSize. AWS - // Lambda POSTs the Amazon Kinesis event, containing records, to your Lambda - // function as JSON. + // The Amazon Resource Name (ARN) of the event source. Any record added to this + // source could cause AWS Lambda to invoke your Lambda function, it depends + // on the BatchSize. AWS Lambda POSTs the event's records to your Lambda function + // as JSON. // // EventSourceArn is a required field EventSourceArn *string `type:"string" required:"true"` @@ -3724,9 +3736,7 @@ type CreateEventSourceMappingInput struct { // in the Amazon Kinesis API Reference Guide or GetShardIterator (http://docs.aws.amazon.com/amazondynamodb/latest/APIReference/API_streams_GetShardIterator.html) // in the Amazon DynamoDB API Reference Guide. The AT_TIMESTAMP value is supported // only for Kinesis streams (http://docs.aws.amazon.com/streams/latest/dev/amazon-kinesis-streams.html). - // - // StartingPosition is a required field - StartingPosition *string `type:"string" required:"true" enum:"EventSourcePosition"` + StartingPosition *string `type:"string" enum:"EventSourcePosition"` // The timestamp of the data record from which to start reading. Used with shard // iterator type (http://docs.aws.amazon.com/kinesis/latest/APIReference/API_GetShardIterator.html#Kinesis-GetShardIterator-request-ShardIteratorType) @@ -3762,9 +3772,6 @@ func (s *CreateEventSourceMappingInput) Validate() error { if s.FunctionName != nil && len(*s.FunctionName) < 1 { invalidParams.Add(request.NewErrParamMinLen("FunctionName", 1)) } - if s.StartingPosition == nil { - invalidParams.Add(request.NewErrParamRequired("StartingPosition")) - } if invalidParams.Len() > 0 { return invalidParams @@ -4422,7 +4429,8 @@ func (s *EnvironmentResponse) SetVariables(v map[string]*string) *EnvironmentRes return s } -// Describes mapping between an Amazon Kinesis stream and a Lambda function. +// Describes mapping between an Amazon Kinesis or DynamoDB stream or an Amazon +// SQS queue and a Lambda function. type EventSourceMappingConfiguration struct { _ struct{} `type:"structure"` @@ -4431,11 +4439,12 @@ type EventSourceMappingConfiguration struct { // with all the retrieved records. BatchSize *int64 `min:"1" type:"integer"` - // The Amazon Resource Name (ARN) of the Amazon Kinesis stream that is the source - // of events. + // The Amazon Resource Name (ARN) of the Amazon Kinesis or DynamoDB stream or + // the SQS queue that is the source of events. EventSourceArn *string `type:"string"` - // The Lambda function to invoke when AWS Lambda detects an event on the stream. + // The Lambda function to invoke when AWS Lambda detects an event on the poll-based + // source. FunctionArn *string `type:"string"` // The UTC time string indicating the last time the event mapping was updated. @@ -5648,8 +5657,8 @@ func (s *ListAliasesOutput) SetNextMarker(v string) *ListAliasesOutput { type ListEventSourceMappingsInput struct { _ struct{} `type:"structure"` - // The Amazon Resource Name (ARN) of the Amazon Kinesis stream. (This parameter - // is optional.) + // The Amazon Resource Name (ARN) of the Amazon Kinesis or DynamoDB stream, + // or an SQS queue. (This parameter is optional.) EventSourceArn *string `location:"querystring" locationName:"EventSourceArn" type:"string"` // The name of the Lambda function. diff --git a/vendor/github.com/aws/aws-sdk-go/service/lambda/errors.go b/vendor/github.com/aws/aws-sdk-go/service/lambda/errors.go index 57daa1c34..447b49ed1 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lambda/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lambda/errors.go @@ -130,6 +130,14 @@ const ( // The resource already exists. ErrCodeResourceConflictException = "ResourceConflictException" + // ErrCodeResourceInUseException for service response error code + // "ResourceInUseException". + // + // The operation conflicts with the resource's availability. For example, you + // attempted to update an EventSoure Mapping in CREATING, or tried to delete + // a EventSoure mapping currently in the UPDATING state. + ErrCodeResourceInUseException = "ResourceInUseException" + // ErrCodeResourceNotFoundException for service response error code // "ResourceNotFoundException". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/lambda/examples_test.go b/vendor/github.com/aws/aws-sdk-go/service/lambda/examples_test.go index dc2225b83..fc266087b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lambda/examples_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lambda/examples_test.go @@ -174,6 +174,8 @@ func ExampleLambda_DeleteEventSourceMapping_shared00() { fmt.Println(lambda.ErrCodeInvalidParameterValueException, aerr.Error()) case lambda.ErrCodeTooManyRequestsException: fmt.Println(lambda.ErrCodeTooManyRequestsException, aerr.Error()) + case lambda.ErrCodeResourceInUseException: + fmt.Println(lambda.ErrCodeResourceInUseException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -797,6 +799,8 @@ func ExampleLambda_UpdateEventSourceMapping_shared00() { fmt.Println(lambda.ErrCodeTooManyRequestsException, aerr.Error()) case lambda.ErrCodeResourceConflictException: fmt.Println(lambda.ErrCodeResourceConflictException, aerr.Error()) + case lambda.ErrCodeResourceInUseException: + fmt.Println(lambda.ErrCodeResourceInUseException, aerr.Error()) default: fmt.Println(aerr.Error()) } diff --git a/vendor/github.com/aws/aws-sdk-go/service/lambda/service.go b/vendor/github.com/aws/aws-sdk-go/service/lambda/service.go index 83c5e3094..1cccdda08 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lambda/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lambda/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "lambda" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "lambda" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Lambda" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Lambda client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/lexmodelbuildingservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/lexmodelbuildingservice/service.go index cc122d1a2..86d1a44d2 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lexmodelbuildingservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lexmodelbuildingservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "models.lex" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "models.lex" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Lex Model Building Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the LexModelBuildingService client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/lexruntimeservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/lexruntimeservice/service.go index a3e785e88..6f41b11a5 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lexruntimeservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lexruntimeservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "runtime.lex" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "runtime.lex" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Lex Runtime Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the LexRuntimeService client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/lightsail/service.go b/vendor/github.com/aws/aws-sdk-go/service/lightsail/service.go index a76cf79e0..b9f97faa8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/lightsail/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/lightsail/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "lightsail" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "lightsail" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Lightsail" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Lightsail client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/machinelearning/service.go b/vendor/github.com/aws/aws-sdk-go/service/machinelearning/service.go index c68da96a4..f4bfcd464 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/machinelearning/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/machinelearning/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "machinelearning" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "machinelearning" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Machine Learning" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MachineLearning client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/macie/api.go b/vendor/github.com/aws/aws-sdk-go/service/macie/api.go new file mode 100644 index 000000000..62a12b0c1 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/macie/api.go @@ -0,0 +1,1611 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package macie + +import ( + "fmt" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awsutil" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/jsonrpc" +) + +const opAssociateMemberAccount = "AssociateMemberAccount" + +// AssociateMemberAccountRequest generates a "aws/request.Request" representing the +// client's request for the AssociateMemberAccount operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AssociateMemberAccount for more information on using the AssociateMemberAccount +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AssociateMemberAccountRequest method. +// req, resp := client.AssociateMemberAccountRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/AssociateMemberAccount +func (c *Macie) AssociateMemberAccountRequest(input *AssociateMemberAccountInput) (req *request.Request, output *AssociateMemberAccountOutput) { + op := &request.Operation{ + Name: opAssociateMemberAccount, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AssociateMemberAccountInput{} + } + + output = &AssociateMemberAccountOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(jsonrpc.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// AssociateMemberAccount API operation for Amazon Macie. +// +// Associates a specified AWS account with Amazon Macie as a member account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation AssociateMemberAccount for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// The request was rejected because it attempted to create resources beyond +// the current AWS account limits. The error code describes the limit exceeded. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/AssociateMemberAccount +func (c *Macie) AssociateMemberAccount(input *AssociateMemberAccountInput) (*AssociateMemberAccountOutput, error) { + req, out := c.AssociateMemberAccountRequest(input) + return out, req.Send() +} + +// AssociateMemberAccountWithContext is the same as AssociateMemberAccount with the addition of +// the ability to pass a context and additional request options. +// +// See AssociateMemberAccount for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) AssociateMemberAccountWithContext(ctx aws.Context, input *AssociateMemberAccountInput, opts ...request.Option) (*AssociateMemberAccountOutput, error) { + req, out := c.AssociateMemberAccountRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opAssociateS3Resources = "AssociateS3Resources" + +// AssociateS3ResourcesRequest generates a "aws/request.Request" representing the +// client's request for the AssociateS3Resources operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AssociateS3Resources for more information on using the AssociateS3Resources +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AssociateS3ResourcesRequest method. +// req, resp := client.AssociateS3ResourcesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/AssociateS3Resources +func (c *Macie) AssociateS3ResourcesRequest(input *AssociateS3ResourcesInput) (req *request.Request, output *AssociateS3ResourcesOutput) { + op := &request.Operation{ + Name: opAssociateS3Resources, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AssociateS3ResourcesInput{} + } + + output = &AssociateS3ResourcesOutput{} + req = c.newRequest(op, input, output) + return +} + +// AssociateS3Resources API operation for Amazon Macie. +// +// Associates specified S3 resources with Amazon Macie for monitoring and data +// classification. If memberAccountId isn't specified, the action associates +// specified S3 resources with Macie for the current master account. If memberAccountId +// is specified, the action associates specified S3 resources with Macie for +// the specified member account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation AssociateS3Resources for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeLimitExceededException "LimitExceededException" +// The request was rejected because it attempted to create resources beyond +// the current AWS account limits. The error code describes the limit exceeded. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/AssociateS3Resources +func (c *Macie) AssociateS3Resources(input *AssociateS3ResourcesInput) (*AssociateS3ResourcesOutput, error) { + req, out := c.AssociateS3ResourcesRequest(input) + return out, req.Send() +} + +// AssociateS3ResourcesWithContext is the same as AssociateS3Resources with the addition of +// the ability to pass a context and additional request options. +// +// See AssociateS3Resources for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) AssociateS3ResourcesWithContext(ctx aws.Context, input *AssociateS3ResourcesInput, opts ...request.Option) (*AssociateS3ResourcesOutput, error) { + req, out := c.AssociateS3ResourcesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDisassociateMemberAccount = "DisassociateMemberAccount" + +// DisassociateMemberAccountRequest generates a "aws/request.Request" representing the +// client's request for the DisassociateMemberAccount operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DisassociateMemberAccount for more information on using the DisassociateMemberAccount +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DisassociateMemberAccountRequest method. +// req, resp := client.DisassociateMemberAccountRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/DisassociateMemberAccount +func (c *Macie) DisassociateMemberAccountRequest(input *DisassociateMemberAccountInput) (req *request.Request, output *DisassociateMemberAccountOutput) { + op := &request.Operation{ + Name: opDisassociateMemberAccount, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DisassociateMemberAccountInput{} + } + + output = &DisassociateMemberAccountOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(jsonrpc.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DisassociateMemberAccount API operation for Amazon Macie. +// +// Removes the specified member account from Amazon Macie. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation DisassociateMemberAccount for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/DisassociateMemberAccount +func (c *Macie) DisassociateMemberAccount(input *DisassociateMemberAccountInput) (*DisassociateMemberAccountOutput, error) { + req, out := c.DisassociateMemberAccountRequest(input) + return out, req.Send() +} + +// DisassociateMemberAccountWithContext is the same as DisassociateMemberAccount with the addition of +// the ability to pass a context and additional request options. +// +// See DisassociateMemberAccount for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) DisassociateMemberAccountWithContext(ctx aws.Context, input *DisassociateMemberAccountInput, opts ...request.Option) (*DisassociateMemberAccountOutput, error) { + req, out := c.DisassociateMemberAccountRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDisassociateS3Resources = "DisassociateS3Resources" + +// DisassociateS3ResourcesRequest generates a "aws/request.Request" representing the +// client's request for the DisassociateS3Resources operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DisassociateS3Resources for more information on using the DisassociateS3Resources +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DisassociateS3ResourcesRequest method. +// req, resp := client.DisassociateS3ResourcesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/DisassociateS3Resources +func (c *Macie) DisassociateS3ResourcesRequest(input *DisassociateS3ResourcesInput) (req *request.Request, output *DisassociateS3ResourcesOutput) { + op := &request.Operation{ + Name: opDisassociateS3Resources, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DisassociateS3ResourcesInput{} + } + + output = &DisassociateS3ResourcesOutput{} + req = c.newRequest(op, input, output) + return +} + +// DisassociateS3Resources API operation for Amazon Macie. +// +// Removes specified S3 resources from being monitored by Amazon Macie. If memberAccountId +// isn't specified, the action removes specified S3 resources from Macie for +// the current master account. If memberAccountId is specified, the action removes +// specified S3 resources from Macie for the specified member account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation DisassociateS3Resources for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/DisassociateS3Resources +func (c *Macie) DisassociateS3Resources(input *DisassociateS3ResourcesInput) (*DisassociateS3ResourcesOutput, error) { + req, out := c.DisassociateS3ResourcesRequest(input) + return out, req.Send() +} + +// DisassociateS3ResourcesWithContext is the same as DisassociateS3Resources with the addition of +// the ability to pass a context and additional request options. +// +// See DisassociateS3Resources for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) DisassociateS3ResourcesWithContext(ctx aws.Context, input *DisassociateS3ResourcesInput, opts ...request.Option) (*DisassociateS3ResourcesOutput, error) { + req, out := c.DisassociateS3ResourcesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opListMemberAccounts = "ListMemberAccounts" + +// ListMemberAccountsRequest generates a "aws/request.Request" representing the +// client's request for the ListMemberAccounts operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListMemberAccounts for more information on using the ListMemberAccounts +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListMemberAccountsRequest method. +// req, resp := client.ListMemberAccountsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/ListMemberAccounts +func (c *Macie) ListMemberAccountsRequest(input *ListMemberAccountsInput) (req *request.Request, output *ListMemberAccountsOutput) { + op := &request.Operation{ + Name: opListMemberAccounts, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"nextToken"}, + OutputTokens: []string{"nextToken"}, + LimitToken: "maxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListMemberAccountsInput{} + } + + output = &ListMemberAccountsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListMemberAccounts API operation for Amazon Macie. +// +// Lists all Amazon Macie member accounts for the current Amazon Macie master +// account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation ListMemberAccounts for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/ListMemberAccounts +func (c *Macie) ListMemberAccounts(input *ListMemberAccountsInput) (*ListMemberAccountsOutput, error) { + req, out := c.ListMemberAccountsRequest(input) + return out, req.Send() +} + +// ListMemberAccountsWithContext is the same as ListMemberAccounts with the addition of +// the ability to pass a context and additional request options. +// +// See ListMemberAccounts for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) ListMemberAccountsWithContext(ctx aws.Context, input *ListMemberAccountsInput, opts ...request.Option) (*ListMemberAccountsOutput, error) { + req, out := c.ListMemberAccountsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListMemberAccountsPages iterates over the pages of a ListMemberAccounts operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListMemberAccounts method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListMemberAccounts operation. +// pageNum := 0 +// err := client.ListMemberAccountsPages(params, +// func(page *ListMemberAccountsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Macie) ListMemberAccountsPages(input *ListMemberAccountsInput, fn func(*ListMemberAccountsOutput, bool) bool) error { + return c.ListMemberAccountsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListMemberAccountsPagesWithContext same as ListMemberAccountsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) ListMemberAccountsPagesWithContext(ctx aws.Context, input *ListMemberAccountsInput, fn func(*ListMemberAccountsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListMemberAccountsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListMemberAccountsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListMemberAccountsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opListS3Resources = "ListS3Resources" + +// ListS3ResourcesRequest generates a "aws/request.Request" representing the +// client's request for the ListS3Resources operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListS3Resources for more information on using the ListS3Resources +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListS3ResourcesRequest method. +// req, resp := client.ListS3ResourcesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/ListS3Resources +func (c *Macie) ListS3ResourcesRequest(input *ListS3ResourcesInput) (req *request.Request, output *ListS3ResourcesOutput) { + op := &request.Operation{ + Name: opListS3Resources, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"nextToken"}, + OutputTokens: []string{"nextToken"}, + LimitToken: "maxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListS3ResourcesInput{} + } + + output = &ListS3ResourcesOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListS3Resources API operation for Amazon Macie. +// +// Lists all the S3 resources associated with Amazon Macie. If memberAccountId +// isn't specified, the action lists the S3 resources associated with Amazon +// Macie for the current master account. If memberAccountId is specified, the +// action lists the S3 resources associated with Amazon Macie for the specified +// member account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation ListS3Resources for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/ListS3Resources +func (c *Macie) ListS3Resources(input *ListS3ResourcesInput) (*ListS3ResourcesOutput, error) { + req, out := c.ListS3ResourcesRequest(input) + return out, req.Send() +} + +// ListS3ResourcesWithContext is the same as ListS3Resources with the addition of +// the ability to pass a context and additional request options. +// +// See ListS3Resources for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) ListS3ResourcesWithContext(ctx aws.Context, input *ListS3ResourcesInput, opts ...request.Option) (*ListS3ResourcesOutput, error) { + req, out := c.ListS3ResourcesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListS3ResourcesPages iterates over the pages of a ListS3Resources operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListS3Resources method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListS3Resources operation. +// pageNum := 0 +// err := client.ListS3ResourcesPages(params, +// func(page *ListS3ResourcesOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Macie) ListS3ResourcesPages(input *ListS3ResourcesInput, fn func(*ListS3ResourcesOutput, bool) bool) error { + return c.ListS3ResourcesPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListS3ResourcesPagesWithContext same as ListS3ResourcesPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) ListS3ResourcesPagesWithContext(ctx aws.Context, input *ListS3ResourcesInput, fn func(*ListS3ResourcesOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListS3ResourcesInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListS3ResourcesRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListS3ResourcesOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opUpdateS3Resources = "UpdateS3Resources" + +// UpdateS3ResourcesRequest generates a "aws/request.Request" representing the +// client's request for the UpdateS3Resources operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See UpdateS3Resources for more information on using the UpdateS3Resources +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the UpdateS3ResourcesRequest method. +// req, resp := client.UpdateS3ResourcesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/UpdateS3Resources +func (c *Macie) UpdateS3ResourcesRequest(input *UpdateS3ResourcesInput) (req *request.Request, output *UpdateS3ResourcesOutput) { + op := &request.Operation{ + Name: opUpdateS3Resources, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &UpdateS3ResourcesInput{} + } + + output = &UpdateS3ResourcesOutput{} + req = c.newRequest(op, input, output) + return +} + +// UpdateS3Resources API operation for Amazon Macie. +// +// Updates the classification types for the specified S3 resources. If memberAccountId +// isn't specified, the action updates the classification types of the S3 resources +// associated with Amazon Macie for the current master account. If memberAccountId +// is specified, the action updates the classification types of the S3 resources +// associated with Amazon Macie for the specified member account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Macie's +// API operation UpdateS3Resources for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidInputException "InvalidInputException" +// The request was rejected because an invalid or out-of-range value was supplied +// for an input parameter. +// +// * ErrCodeAccessDeniedException "AccessDeniedException" +// You do not have required permissions to access the requested resource. +// +// * ErrCodeInternalException "InternalException" +// Internal server error. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19/UpdateS3Resources +func (c *Macie) UpdateS3Resources(input *UpdateS3ResourcesInput) (*UpdateS3ResourcesOutput, error) { + req, out := c.UpdateS3ResourcesRequest(input) + return out, req.Send() +} + +// UpdateS3ResourcesWithContext is the same as UpdateS3Resources with the addition of +// the ability to pass a context and additional request options. +// +// See UpdateS3Resources for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Macie) UpdateS3ResourcesWithContext(ctx aws.Context, input *UpdateS3ResourcesInput, opts ...request.Option) (*UpdateS3ResourcesOutput, error) { + req, out := c.UpdateS3ResourcesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +type AssociateMemberAccountInput struct { + _ struct{} `type:"structure"` + + // The ID of the AWS account that you want to associate with Amazon Macie as + // a member account. + // + // MemberAccountId is a required field + MemberAccountId *string `locationName:"memberAccountId" type:"string" required:"true"` +} + +// String returns the string representation +func (s AssociateMemberAccountInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateMemberAccountInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AssociateMemberAccountInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AssociateMemberAccountInput"} + if s.MemberAccountId == nil { + invalidParams.Add(request.NewErrParamRequired("MemberAccountId")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMemberAccountId sets the MemberAccountId field's value. +func (s *AssociateMemberAccountInput) SetMemberAccountId(v string) *AssociateMemberAccountInput { + s.MemberAccountId = &v + return s +} + +type AssociateMemberAccountOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s AssociateMemberAccountOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateMemberAccountOutput) GoString() string { + return s.String() +} + +type AssociateS3ResourcesInput struct { + _ struct{} `type:"structure"` + + // The ID of the Amazon Macie member account whose resources you want to associate + // with Macie. + MemberAccountId *string `locationName:"memberAccountId" type:"string"` + + // The S3 resources that you want to associate with Amazon Macie for monitoring + // and data classification. + // + // S3Resources is a required field + S3Resources []*S3ResourceClassification `locationName:"s3Resources" type:"list" required:"true"` +} + +// String returns the string representation +func (s AssociateS3ResourcesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateS3ResourcesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AssociateS3ResourcesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AssociateS3ResourcesInput"} + if s.S3Resources == nil { + invalidParams.Add(request.NewErrParamRequired("S3Resources")) + } + if s.S3Resources != nil { + for i, v := range s.S3Resources { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "S3Resources", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMemberAccountId sets the MemberAccountId field's value. +func (s *AssociateS3ResourcesInput) SetMemberAccountId(v string) *AssociateS3ResourcesInput { + s.MemberAccountId = &v + return s +} + +// SetS3Resources sets the S3Resources field's value. +func (s *AssociateS3ResourcesInput) SetS3Resources(v []*S3ResourceClassification) *AssociateS3ResourcesInput { + s.S3Resources = v + return s +} + +type AssociateS3ResourcesOutput struct { + _ struct{} `type:"structure"` + + // S3 resources that couldn't be associated with Amazon Macie. An error code + // and an error message are provided for each failed item. + FailedS3Resources []*FailedS3Resource `locationName:"failedS3Resources" type:"list"` +} + +// String returns the string representation +func (s AssociateS3ResourcesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateS3ResourcesOutput) GoString() string { + return s.String() +} + +// SetFailedS3Resources sets the FailedS3Resources field's value. +func (s *AssociateS3ResourcesOutput) SetFailedS3Resources(v []*FailedS3Resource) *AssociateS3ResourcesOutput { + s.FailedS3Resources = v + return s +} + +// The classification type that Amazon Macie applies to the associated S3 resources. +type ClassificationType struct { + _ struct{} `type:"structure"` + + // A continuous classification of the objects that are added to a specified + // S3 bucket. Amazon Macie begins performing continuous classification after + // a bucket is successfully associated with Amazon Macie. + // + // Continuous is a required field + Continuous *string `locationName:"continuous" type:"string" required:"true" enum:"S3ContinuousClassificationType"` + + // A one-time classification of all of the existing objects in a specified S3 + // bucket. + // + // OneTime is a required field + OneTime *string `locationName:"oneTime" type:"string" required:"true" enum:"S3OneTimeClassificationType"` +} + +// String returns the string representation +func (s ClassificationType) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ClassificationType) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ClassificationType) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ClassificationType"} + if s.Continuous == nil { + invalidParams.Add(request.NewErrParamRequired("Continuous")) + } + if s.OneTime == nil { + invalidParams.Add(request.NewErrParamRequired("OneTime")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetContinuous sets the Continuous field's value. +func (s *ClassificationType) SetContinuous(v string) *ClassificationType { + s.Continuous = &v + return s +} + +// SetOneTime sets the OneTime field's value. +func (s *ClassificationType) SetOneTime(v string) *ClassificationType { + s.OneTime = &v + return s +} + +// The classification type that Amazon Macie applies to the associated S3 resources. +// At least one of the classification types (oneTime or continuous) must be +// specified. +type ClassificationTypeUpdate struct { + _ struct{} `type:"structure"` + + // A continuous classification of the objects that are added to a specified + // S3 bucket. Amazon Macie begins performing continuous classification after + // a bucket is successfully associated with Amazon Macie. + Continuous *string `locationName:"continuous" type:"string" enum:"S3ContinuousClassificationType"` + + // A one-time classification of all of the existing objects in a specified S3 + // bucket. + OneTime *string `locationName:"oneTime" type:"string" enum:"S3OneTimeClassificationType"` +} + +// String returns the string representation +func (s ClassificationTypeUpdate) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ClassificationTypeUpdate) GoString() string { + return s.String() +} + +// SetContinuous sets the Continuous field's value. +func (s *ClassificationTypeUpdate) SetContinuous(v string) *ClassificationTypeUpdate { + s.Continuous = &v + return s +} + +// SetOneTime sets the OneTime field's value. +func (s *ClassificationTypeUpdate) SetOneTime(v string) *ClassificationTypeUpdate { + s.OneTime = &v + return s +} + +type DisassociateMemberAccountInput struct { + _ struct{} `type:"structure"` + + // The ID of the member account that you want to remove from Amazon Macie. + // + // MemberAccountId is a required field + MemberAccountId *string `locationName:"memberAccountId" type:"string" required:"true"` +} + +// String returns the string representation +func (s DisassociateMemberAccountInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateMemberAccountInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DisassociateMemberAccountInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DisassociateMemberAccountInput"} + if s.MemberAccountId == nil { + invalidParams.Add(request.NewErrParamRequired("MemberAccountId")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMemberAccountId sets the MemberAccountId field's value. +func (s *DisassociateMemberAccountInput) SetMemberAccountId(v string) *DisassociateMemberAccountInput { + s.MemberAccountId = &v + return s +} + +type DisassociateMemberAccountOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DisassociateMemberAccountOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateMemberAccountOutput) GoString() string { + return s.String() +} + +type DisassociateS3ResourcesInput struct { + _ struct{} `type:"structure"` + + // The S3 resources (buckets or prefixes) that you want to remove from being + // monitored and classified by Amazon Macie. + // + // AssociatedS3Resources is a required field + AssociatedS3Resources []*S3Resource `locationName:"associatedS3Resources" type:"list" required:"true"` + + // The ID of the Amazon Macie member account whose resources you want to remove + // from being monitored by Amazon Macie. + MemberAccountId *string `locationName:"memberAccountId" type:"string"` +} + +// String returns the string representation +func (s DisassociateS3ResourcesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateS3ResourcesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DisassociateS3ResourcesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DisassociateS3ResourcesInput"} + if s.AssociatedS3Resources == nil { + invalidParams.Add(request.NewErrParamRequired("AssociatedS3Resources")) + } + if s.AssociatedS3Resources != nil { + for i, v := range s.AssociatedS3Resources { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AssociatedS3Resources", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAssociatedS3Resources sets the AssociatedS3Resources field's value. +func (s *DisassociateS3ResourcesInput) SetAssociatedS3Resources(v []*S3Resource) *DisassociateS3ResourcesInput { + s.AssociatedS3Resources = v + return s +} + +// SetMemberAccountId sets the MemberAccountId field's value. +func (s *DisassociateS3ResourcesInput) SetMemberAccountId(v string) *DisassociateS3ResourcesInput { + s.MemberAccountId = &v + return s +} + +type DisassociateS3ResourcesOutput struct { + _ struct{} `type:"structure"` + + // S3 resources that couldn't be removed from being monitored and classified + // by Amazon Macie. An error code and an error message are provided for each + // failed item. + FailedS3Resources []*FailedS3Resource `locationName:"failedS3Resources" type:"list"` +} + +// String returns the string representation +func (s DisassociateS3ResourcesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateS3ResourcesOutput) GoString() string { + return s.String() +} + +// SetFailedS3Resources sets the FailedS3Resources field's value. +func (s *DisassociateS3ResourcesOutput) SetFailedS3Resources(v []*FailedS3Resource) *DisassociateS3ResourcesOutput { + s.FailedS3Resources = v + return s +} + +// Includes details about the failed S3 resources. +type FailedS3Resource struct { + _ struct{} `type:"structure"` + + // The status code of a failed item. + ErrorCode *string `locationName:"errorCode" type:"string"` + + // The error message of a failed item. + ErrorMessage *string `locationName:"errorMessage" type:"string"` + + // The failed S3 resources. + FailedItem *S3Resource `locationName:"failedItem" type:"structure"` +} + +// String returns the string representation +func (s FailedS3Resource) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FailedS3Resource) GoString() string { + return s.String() +} + +// SetErrorCode sets the ErrorCode field's value. +func (s *FailedS3Resource) SetErrorCode(v string) *FailedS3Resource { + s.ErrorCode = &v + return s +} + +// SetErrorMessage sets the ErrorMessage field's value. +func (s *FailedS3Resource) SetErrorMessage(v string) *FailedS3Resource { + s.ErrorMessage = &v + return s +} + +// SetFailedItem sets the FailedItem field's value. +func (s *FailedS3Resource) SetFailedItem(v *S3Resource) *FailedS3Resource { + s.FailedItem = v + return s +} + +type ListMemberAccountsInput struct { + _ struct{} `type:"structure"` + + // Use this parameter to indicate the maximum number of items that you want + // in the response. The default value is 250. + MaxResults *int64 `locationName:"maxResults" type:"integer"` + + // Use this parameter when paginating results. Set the value of this parameter + // to null on your first call to the ListMemberAccounts action. Subsequent calls + // to the action fill nextToken in the request with the value of nextToken from + // the previous response to continue listing data. + NextToken *string `locationName:"nextToken" type:"string"` +} + +// String returns the string representation +func (s ListMemberAccountsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListMemberAccountsInput) GoString() string { + return s.String() +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListMemberAccountsInput) SetMaxResults(v int64) *ListMemberAccountsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListMemberAccountsInput) SetNextToken(v string) *ListMemberAccountsInput { + s.NextToken = &v + return s +} + +type ListMemberAccountsOutput struct { + _ struct{} `type:"structure"` + + // A list of the Amazon Macie member accounts returned by the action. The current + // master account is also included in this list. + MemberAccounts []*MemberAccount `locationName:"memberAccounts" type:"list"` + + // When a response is generated, if there is more data to be listed, this parameter + // is present in the response and contains the value to use for the nextToken + // parameter in a subsequent pagination request. If there is no more data to + // be listed, this parameter is set to null. + NextToken *string `locationName:"nextToken" type:"string"` +} + +// String returns the string representation +func (s ListMemberAccountsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListMemberAccountsOutput) GoString() string { + return s.String() +} + +// SetMemberAccounts sets the MemberAccounts field's value. +func (s *ListMemberAccountsOutput) SetMemberAccounts(v []*MemberAccount) *ListMemberAccountsOutput { + s.MemberAccounts = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListMemberAccountsOutput) SetNextToken(v string) *ListMemberAccountsOutput { + s.NextToken = &v + return s +} + +type ListS3ResourcesInput struct { + _ struct{} `type:"structure"` + + // Use this parameter to indicate the maximum number of items that you want + // in the response. The default value is 250. + MaxResults *int64 `locationName:"maxResults" type:"integer"` + + // The Amazon Macie member account ID whose associated S3 resources you want + // to list. + MemberAccountId *string `locationName:"memberAccountId" type:"string"` + + // Use this parameter when paginating results. Set its value to null on your + // first call to the ListS3Resources action. Subsequent calls to the action + // fill nextToken in the request with the value of nextToken from the previous + // response to continue listing data. + NextToken *string `locationName:"nextToken" type:"string"` +} + +// String returns the string representation +func (s ListS3ResourcesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListS3ResourcesInput) GoString() string { + return s.String() +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListS3ResourcesInput) SetMaxResults(v int64) *ListS3ResourcesInput { + s.MaxResults = &v + return s +} + +// SetMemberAccountId sets the MemberAccountId field's value. +func (s *ListS3ResourcesInput) SetMemberAccountId(v string) *ListS3ResourcesInput { + s.MemberAccountId = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListS3ResourcesInput) SetNextToken(v string) *ListS3ResourcesInput { + s.NextToken = &v + return s +} + +type ListS3ResourcesOutput struct { + _ struct{} `type:"structure"` + + // When a response is generated, if there is more data to be listed, this parameter + // is present in the response and contains the value to use for the nextToken + // parameter in a subsequent pagination request. If there is no more data to + // be listed, this parameter is set to null. + NextToken *string `locationName:"nextToken" type:"string"` + + // A list of the associated S3 resources returned by the action. + S3Resources []*S3ResourceClassification `locationName:"s3Resources" type:"list"` +} + +// String returns the string representation +func (s ListS3ResourcesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListS3ResourcesOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *ListS3ResourcesOutput) SetNextToken(v string) *ListS3ResourcesOutput { + s.NextToken = &v + return s +} + +// SetS3Resources sets the S3Resources field's value. +func (s *ListS3ResourcesOutput) SetS3Resources(v []*S3ResourceClassification) *ListS3ResourcesOutput { + s.S3Resources = v + return s +} + +// Contains information about the Amazon Macie member account. +type MemberAccount struct { + _ struct{} `type:"structure"` + + // The AWS account ID of the Amazon Macie member account. + AccountId *string `locationName:"accountId" type:"string"` +} + +// String returns the string representation +func (s MemberAccount) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s MemberAccount) GoString() string { + return s.String() +} + +// SetAccountId sets the AccountId field's value. +func (s *MemberAccount) SetAccountId(v string) *MemberAccount { + s.AccountId = &v + return s +} + +// Contains information about the S3 resource. This data type is used as a request +// parameter in the DisassociateS3Resources action and can be used as a response +// parameter in the AssociateS3Resources and UpdateS3Resources actions. +type S3Resource struct { + _ struct{} `type:"structure"` + + // The name of the S3 bucket. + // + // BucketName is a required field + BucketName *string `locationName:"bucketName" type:"string" required:"true"` + + // The prefix of the S3 bucket. + Prefix *string `locationName:"prefix" type:"string"` +} + +// String returns the string representation +func (s S3Resource) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s S3Resource) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *S3Resource) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "S3Resource"} + if s.BucketName == nil { + invalidParams.Add(request.NewErrParamRequired("BucketName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetBucketName sets the BucketName field's value. +func (s *S3Resource) SetBucketName(v string) *S3Resource { + s.BucketName = &v + return s +} + +// SetPrefix sets the Prefix field's value. +func (s *S3Resource) SetPrefix(v string) *S3Resource { + s.Prefix = &v + return s +} + +// The S3 resources that you want to associate with Amazon Macie for monitoring +// and data classification. This data type is used as a request parameter in +// the AssociateS3Resources action and a response parameter in the ListS3Resources +// action. +type S3ResourceClassification struct { + _ struct{} `type:"structure"` + + // The name of the S3 bucket that you want to associate with Amazon Macie. + // + // BucketName is a required field + BucketName *string `locationName:"bucketName" type:"string" required:"true"` + + // The classification type that you want to specify for the resource associated + // with Amazon Macie. + // + // ClassificationType is a required field + ClassificationType *ClassificationType `locationName:"classificationType" type:"structure" required:"true"` + + // The prefix of the S3 bucket that you want to associate with Amazon Macie. + Prefix *string `locationName:"prefix" type:"string"` +} + +// String returns the string representation +func (s S3ResourceClassification) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s S3ResourceClassification) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *S3ResourceClassification) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "S3ResourceClassification"} + if s.BucketName == nil { + invalidParams.Add(request.NewErrParamRequired("BucketName")) + } + if s.ClassificationType == nil { + invalidParams.Add(request.NewErrParamRequired("ClassificationType")) + } + if s.ClassificationType != nil { + if err := s.ClassificationType.Validate(); err != nil { + invalidParams.AddNested("ClassificationType", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetBucketName sets the BucketName field's value. +func (s *S3ResourceClassification) SetBucketName(v string) *S3ResourceClassification { + s.BucketName = &v + return s +} + +// SetClassificationType sets the ClassificationType field's value. +func (s *S3ResourceClassification) SetClassificationType(v *ClassificationType) *S3ResourceClassification { + s.ClassificationType = v + return s +} + +// SetPrefix sets the Prefix field's value. +func (s *S3ResourceClassification) SetPrefix(v string) *S3ResourceClassification { + s.Prefix = &v + return s +} + +// The S3 resources whose classification types you want to update. This data +// type is used as a request parameter in the UpdateS3Resources action. +type S3ResourceClassificationUpdate struct { + _ struct{} `type:"structure"` + + // The name of the S3 bucket whose classification types you want to update. + // + // BucketName is a required field + BucketName *string `locationName:"bucketName" type:"string" required:"true"` + + // The classification type that you want to update for the resource associated + // with Amazon Macie. + // + // ClassificationTypeUpdate is a required field + ClassificationTypeUpdate *ClassificationTypeUpdate `locationName:"classificationTypeUpdate" type:"structure" required:"true"` + + // The prefix of the S3 bucket whose classification types you want to update. + Prefix *string `locationName:"prefix" type:"string"` +} + +// String returns the string representation +func (s S3ResourceClassificationUpdate) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s S3ResourceClassificationUpdate) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *S3ResourceClassificationUpdate) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "S3ResourceClassificationUpdate"} + if s.BucketName == nil { + invalidParams.Add(request.NewErrParamRequired("BucketName")) + } + if s.ClassificationTypeUpdate == nil { + invalidParams.Add(request.NewErrParamRequired("ClassificationTypeUpdate")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetBucketName sets the BucketName field's value. +func (s *S3ResourceClassificationUpdate) SetBucketName(v string) *S3ResourceClassificationUpdate { + s.BucketName = &v + return s +} + +// SetClassificationTypeUpdate sets the ClassificationTypeUpdate field's value. +func (s *S3ResourceClassificationUpdate) SetClassificationTypeUpdate(v *ClassificationTypeUpdate) *S3ResourceClassificationUpdate { + s.ClassificationTypeUpdate = v + return s +} + +// SetPrefix sets the Prefix field's value. +func (s *S3ResourceClassificationUpdate) SetPrefix(v string) *S3ResourceClassificationUpdate { + s.Prefix = &v + return s +} + +type UpdateS3ResourcesInput struct { + _ struct{} `type:"structure"` + + // The AWS ID of the Amazon Macie member account whose S3 resources' classification + // types you want to update. + MemberAccountId *string `locationName:"memberAccountId" type:"string"` + + // The S3 resources whose classification types you want to update. + // + // S3ResourcesUpdate is a required field + S3ResourcesUpdate []*S3ResourceClassificationUpdate `locationName:"s3ResourcesUpdate" type:"list" required:"true"` +} + +// String returns the string representation +func (s UpdateS3ResourcesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateS3ResourcesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *UpdateS3ResourcesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "UpdateS3ResourcesInput"} + if s.S3ResourcesUpdate == nil { + invalidParams.Add(request.NewErrParamRequired("S3ResourcesUpdate")) + } + if s.S3ResourcesUpdate != nil { + for i, v := range s.S3ResourcesUpdate { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "S3ResourcesUpdate", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMemberAccountId sets the MemberAccountId field's value. +func (s *UpdateS3ResourcesInput) SetMemberAccountId(v string) *UpdateS3ResourcesInput { + s.MemberAccountId = &v + return s +} + +// SetS3ResourcesUpdate sets the S3ResourcesUpdate field's value. +func (s *UpdateS3ResourcesInput) SetS3ResourcesUpdate(v []*S3ResourceClassificationUpdate) *UpdateS3ResourcesInput { + s.S3ResourcesUpdate = v + return s +} + +type UpdateS3ResourcesOutput struct { + _ struct{} `type:"structure"` + + // The S3 resources whose classification types can't be updated. An error code + // and an error message are provided for each failed item. + FailedS3Resources []*FailedS3Resource `locationName:"failedS3Resources" type:"list"` +} + +// String returns the string representation +func (s UpdateS3ResourcesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateS3ResourcesOutput) GoString() string { + return s.String() +} + +// SetFailedS3Resources sets the FailedS3Resources field's value. +func (s *UpdateS3ResourcesOutput) SetFailedS3Resources(v []*FailedS3Resource) *UpdateS3ResourcesOutput { + s.FailedS3Resources = v + return s +} + +const ( + // S3ContinuousClassificationTypeFull is a S3ContinuousClassificationType enum value + S3ContinuousClassificationTypeFull = "FULL" +) + +const ( + // S3OneTimeClassificationTypeFull is a S3OneTimeClassificationType enum value + S3OneTimeClassificationTypeFull = "FULL" + + // S3OneTimeClassificationTypeNone is a S3OneTimeClassificationType enum value + S3OneTimeClassificationTypeNone = "NONE" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/macie/doc.go b/vendor/github.com/aws/aws-sdk-go/service/macie/doc.go new file mode 100644 index 000000000..1b8f9632f --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/macie/doc.go @@ -0,0 +1,33 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package macie provides the client and types for making API +// requests to Amazon Macie. +// +// Amazon Macie is a security service that uses machine learning to automatically +// discover, classify, and protect sensitive data in AWS. Macie recognizes sensitive +// data such as personally identifiable information (PII) or intellectual property, +// and provides you with dashboards and alerts that give visibility into how +// this data is being accessed or moved. For more information, see the Macie +// User Guide (https://docs.aws.amazon.com/macie/latest/userguide/what-is-macie.html). +// +// See https://docs.aws.amazon.com/goto/WebAPI/macie-2017-12-19 for more information on this service. +// +// See macie package documentation for more information. +// https://docs.aws.amazon.com/sdk-for-go/api/service/macie/ +// +// Using the Client +// +// To contact Amazon Macie with the SDK use the New function to create +// a new service client. With that client you can make API requests to the service. +// These clients are safe to use concurrently. +// +// See the SDK's documentation for more information on how to use the SDK. +// https://docs.aws.amazon.com/sdk-for-go/api/ +// +// See aws.Config documentation for more information on configuring SDK clients. +// https://docs.aws.amazon.com/sdk-for-go/api/aws/#Config +// +// See the Amazon Macie client Macie for more +// information on creating client for this service. +// https://docs.aws.amazon.com/sdk-for-go/api/service/macie/#New +package macie diff --git a/vendor/github.com/aws/aws-sdk-go/service/macie/errors.go b/vendor/github.com/aws/aws-sdk-go/service/macie/errors.go new file mode 100644 index 000000000..77768d52e --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/macie/errors.go @@ -0,0 +1,32 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package macie + +const ( + + // ErrCodeAccessDeniedException for service response error code + // "AccessDeniedException". + // + // You do not have required permissions to access the requested resource. + ErrCodeAccessDeniedException = "AccessDeniedException" + + // ErrCodeInternalException for service response error code + // "InternalException". + // + // Internal server error. + ErrCodeInternalException = "InternalException" + + // ErrCodeInvalidInputException for service response error code + // "InvalidInputException". + // + // The request was rejected because an invalid or out-of-range value was supplied + // for an input parameter. + ErrCodeInvalidInputException = "InvalidInputException" + + // ErrCodeLimitExceededException for service response error code + // "LimitExceededException". + // + // The request was rejected because it attempted to create resources beyond + // the current AWS account limits. The error code describes the limit exceeded. + ErrCodeLimitExceededException = "LimitExceededException" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/macie/macieiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/macie/macieiface/interface.go new file mode 100644 index 000000000..423098cf5 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/macie/macieiface/interface.go @@ -0,0 +1,98 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package macieiface provides an interface to enable mocking the Amazon Macie service client +// for testing your code. +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. +package macieiface + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/service/macie" +) + +// MacieAPI provides an interface to enable mocking the +// macie.Macie service client's API operation, +// paginators, and waiters. This make unit testing your code that calls out +// to the SDK's service client's calls easier. +// +// The best way to use this interface is so the SDK's service client's calls +// can be stubbed out for unit testing your code with the SDK without needing +// to inject custom request handlers into the SDK's request pipeline. +// +// // myFunc uses an SDK service client to make a request to +// // Amazon Macie. +// func myFunc(svc macieiface.MacieAPI) bool { +// // Make svc.AssociateMemberAccount request +// } +// +// func main() { +// sess := session.New() +// svc := macie.New(sess) +// +// myFunc(svc) +// } +// +// In your _test.go file: +// +// // Define a mock struct to be used in your unit tests of myFunc. +// type mockMacieClient struct { +// macieiface.MacieAPI +// } +// func (m *mockMacieClient) AssociateMemberAccount(input *macie.AssociateMemberAccountInput) (*macie.AssociateMemberAccountOutput, error) { +// // mock response/functionality +// } +// +// func TestMyFunc(t *testing.T) { +// // Setup Test +// mockSvc := &mockMacieClient{} +// +// myfunc(mockSvc) +// +// // Verify myFunc's functionality +// } +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. Its suggested to use the pattern above for testing, or using +// tooling to generate mocks to satisfy the interfaces. +type MacieAPI interface { + AssociateMemberAccount(*macie.AssociateMemberAccountInput) (*macie.AssociateMemberAccountOutput, error) + AssociateMemberAccountWithContext(aws.Context, *macie.AssociateMemberAccountInput, ...request.Option) (*macie.AssociateMemberAccountOutput, error) + AssociateMemberAccountRequest(*macie.AssociateMemberAccountInput) (*request.Request, *macie.AssociateMemberAccountOutput) + + AssociateS3Resources(*macie.AssociateS3ResourcesInput) (*macie.AssociateS3ResourcesOutput, error) + AssociateS3ResourcesWithContext(aws.Context, *macie.AssociateS3ResourcesInput, ...request.Option) (*macie.AssociateS3ResourcesOutput, error) + AssociateS3ResourcesRequest(*macie.AssociateS3ResourcesInput) (*request.Request, *macie.AssociateS3ResourcesOutput) + + DisassociateMemberAccount(*macie.DisassociateMemberAccountInput) (*macie.DisassociateMemberAccountOutput, error) + DisassociateMemberAccountWithContext(aws.Context, *macie.DisassociateMemberAccountInput, ...request.Option) (*macie.DisassociateMemberAccountOutput, error) + DisassociateMemberAccountRequest(*macie.DisassociateMemberAccountInput) (*request.Request, *macie.DisassociateMemberAccountOutput) + + DisassociateS3Resources(*macie.DisassociateS3ResourcesInput) (*macie.DisassociateS3ResourcesOutput, error) + DisassociateS3ResourcesWithContext(aws.Context, *macie.DisassociateS3ResourcesInput, ...request.Option) (*macie.DisassociateS3ResourcesOutput, error) + DisassociateS3ResourcesRequest(*macie.DisassociateS3ResourcesInput) (*request.Request, *macie.DisassociateS3ResourcesOutput) + + ListMemberAccounts(*macie.ListMemberAccountsInput) (*macie.ListMemberAccountsOutput, error) + ListMemberAccountsWithContext(aws.Context, *macie.ListMemberAccountsInput, ...request.Option) (*macie.ListMemberAccountsOutput, error) + ListMemberAccountsRequest(*macie.ListMemberAccountsInput) (*request.Request, *macie.ListMemberAccountsOutput) + + ListMemberAccountsPages(*macie.ListMemberAccountsInput, func(*macie.ListMemberAccountsOutput, bool) bool) error + ListMemberAccountsPagesWithContext(aws.Context, *macie.ListMemberAccountsInput, func(*macie.ListMemberAccountsOutput, bool) bool, ...request.Option) error + + ListS3Resources(*macie.ListS3ResourcesInput) (*macie.ListS3ResourcesOutput, error) + ListS3ResourcesWithContext(aws.Context, *macie.ListS3ResourcesInput, ...request.Option) (*macie.ListS3ResourcesOutput, error) + ListS3ResourcesRequest(*macie.ListS3ResourcesInput) (*request.Request, *macie.ListS3ResourcesOutput) + + ListS3ResourcesPages(*macie.ListS3ResourcesInput, func(*macie.ListS3ResourcesOutput, bool) bool) error + ListS3ResourcesPagesWithContext(aws.Context, *macie.ListS3ResourcesInput, func(*macie.ListS3ResourcesOutput, bool) bool, ...request.Option) error + + UpdateS3Resources(*macie.UpdateS3ResourcesInput) (*macie.UpdateS3ResourcesOutput, error) + UpdateS3ResourcesWithContext(aws.Context, *macie.UpdateS3ResourcesInput, ...request.Option) (*macie.UpdateS3ResourcesOutput, error) + UpdateS3ResourcesRequest(*macie.UpdateS3ResourcesInput) (*request.Request, *macie.UpdateS3ResourcesOutput) +} + +var _ MacieAPI = (*macie.Macie)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/macie/service.go b/vendor/github.com/aws/aws-sdk-go/service/macie/service.go new file mode 100644 index 000000000..0b38598f0 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/macie/service.go @@ -0,0 +1,97 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package macie + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/signer/v4" + "github.com/aws/aws-sdk-go/private/protocol/jsonrpc" +) + +// Macie provides the API operation methods for making requests to +// Amazon Macie. See this package's package overview docs +// for details on the service. +// +// Macie methods are safe to use concurrently. It is not safe to +// modify mutate any of the struct's properties though. +type Macie struct { + *client.Client +} + +// Used for custom client initialization logic +var initClient func(*client.Client) + +// Used for custom request initialization logic +var initRequest func(*request.Request) + +// Service information constants +const ( + ServiceName = "Macie" // Name of service. + EndpointsID = "macie" // ID to lookup a service endpoint with. + ServiceID = "Macie" // ServiceID is a unique identifer of a specific service. +) + +// New creates a new instance of the Macie client with a session. +// If additional configuration is needed for the client instance use the optional +// aws.Config parameter to add your extra config. +// +// Example: +// // Create a Macie client from just a session. +// svc := macie.New(mySession) +// +// // Create a Macie client with additional configuration +// svc := macie.New(mySession, aws.NewConfig().WithRegion("us-west-2")) +func New(p client.ConfigProvider, cfgs ...*aws.Config) *Macie { + c := p.ClientConfig(EndpointsID, cfgs...) + return newClient(*c.Config, c.Handlers, c.Endpoint, c.SigningRegion, c.SigningName) +} + +// newClient creates, initializes and returns a new service client instance. +func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion, signingName string) *Macie { + svc := &Macie{ + Client: client.New( + cfg, + metadata.ClientInfo{ + ServiceName: ServiceName, + ServiceID: ServiceID, + SigningName: signingName, + SigningRegion: signingRegion, + Endpoint: endpoint, + APIVersion: "2017-12-19", + JSONVersion: "1.1", + TargetPrefix: "MacieService", + }, + handlers, + ), + } + + // Handlers + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(jsonrpc.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(jsonrpc.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(jsonrpc.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(jsonrpc.UnmarshalErrorHandler) + + // Run custom client initialization if present + if initClient != nil { + initClient(svc.Client) + } + + return svc +} + +// newRequest creates a new request for a Macie operation and runs any +// custom request initialization. +func (c *Macie) newRequest(op *request.Operation, params, data interface{}) *request.Request { + req := c.NewRequest(op, params, data) + + // Run custom request initialization if present + if initRequest != nil { + initRequest(req) + } + + return req +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/api.go b/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/api.go index b83510b5e..90b65b662 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/api.go @@ -73,7 +73,7 @@ func (c *MarketplaceCommerceAnalytics) GenerateDataSetRequest(input *GenerateDat // API operation GenerateDataSet for usage and error information. // // Returned Error Codes: -// * ErrCodeException "Exception" +// * ErrCodeException "MarketplaceCommerceAnalyticsException" // This exception is thrown when an internal service error occurs. // // See also, https://docs.aws.amazon.com/goto/WebAPI/marketplacecommerceanalytics-2015-07-01/GenerateDataSet @@ -162,7 +162,7 @@ func (c *MarketplaceCommerceAnalytics) StartSupportDataExportRequest(input *Star // API operation StartSupportDataExport for usage and error information. // // Returned Error Codes: -// * ErrCodeException "Exception" +// * ErrCodeException "MarketplaceCommerceAnalyticsException" // This exception is thrown when an internal service error occurs. // // See also, https://docs.aws.amazon.com/goto/WebAPI/marketplacecommerceanalytics-2015-07-01/StartSupportDataExport diff --git a/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/errors.go b/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/errors.go index 3d81dcbfc..ff07e02b5 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/errors.go @@ -5,8 +5,8 @@ package marketplacecommerceanalytics const ( // ErrCodeException for service response error code - // "Exception". + // "MarketplaceCommerceAnalyticsException". // // This exception is thrown when an internal service error occurs. - ErrCodeException = "Exception" + ErrCodeException = "MarketplaceCommerceAnalyticsException" ) diff --git a/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/service.go b/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/service.go index 721da2fcf..88543dd51 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/marketplacecommerceanalytics/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "marketplacecommerceanalytics" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "marketplacecommerceanalytics" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Marketplace Commerce Analytics" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MarketplaceCommerceAnalytics client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/marketplaceentitlementservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/marketplaceentitlementservice/service.go index ebb07d98f..69a7e11a8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/marketplaceentitlementservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/marketplaceentitlementservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "entitlement.marketplace" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "entitlement.marketplace" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Marketplace Entitlement Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MarketplaceEntitlementService client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/marketplacemetering/service.go b/vendor/github.com/aws/aws-sdk-go/service/marketplacemetering/service.go index c189a8ab0..aad4f1114 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/marketplacemetering/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/marketplacemetering/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "metering.marketplace" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "metering.marketplace" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Marketplace Metering" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MarketplaceMetering client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/api.go b/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/api.go index 27c8e2895..576c54310 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/api.go @@ -3,6 +3,7 @@ package mediaconvert import ( + "fmt" "time" "github.com/aws/aws-sdk-go/aws" @@ -1786,7 +1787,12 @@ func (c *MediaConvert) UpdateQueueWithContext(ctx aws.Context, input *UpdateQueu } // Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to -// the value AAC. +// the value AAC. The service accepts one of two mutually exclusive groups of +// AAC settings--VBR and CBR. To select one of these modes, set the value of +// Bitrate control mode (rateControlMode) to "VBR" or "CBR". In VBR mode, you +// control the audio quality with the setting VBR quality (vbrQuality). In CBR +// mode, you use the setting Bitrate (bitrate). Defaults and valid values depend +// on the rate control mode. type AacSettings struct { _ struct{} `type:"structure"` @@ -1801,9 +1807,9 @@ type AacSettings struct { // and FollowInputAudioType. AudioDescriptionBroadcasterMix *string `locationName:"audioDescriptionBroadcasterMix" type:"string" enum:"AacAudioDescriptionBroadcasterMix"` - // Average bitrate in bits/second. Valid values depend on rate control mode - // and profile. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + // Average bitrate in bits/second. Defaults and valid values depend on rate + // control mode and profile. + Bitrate *int64 `locationName:"bitrate" min:"6000" type:"integer"` // AAC Profile. CodecProfile *string `locationName:"codecProfile" type:"string" enum:"AacCodecProfile"` @@ -1813,7 +1819,9 @@ type AacSettings struct { // Mix)" setting receives a stereo description plus control track and emits // a mono AAC encode of the description track, with control data emitted in // the PES header as per ETSI TS 101 154 Annex E. - CodingMode *string `locationName:"codingMode" type:"string" enum:"AacCodingMode"` + // + // CodingMode is a required field + CodingMode *string `locationName:"codingMode" type:"string" required:"true" enum:"AacCodingMode"` // Rate Control Mode. RateControlMode *string `locationName:"rateControlMode" type:"string" enum:"AacRateControlMode"` @@ -1823,7 +1831,9 @@ type AacSettings struct { RawFormat *string `locationName:"rawFormat" type:"string" enum:"AacRawFormat"` // Sample rate in Hz. Valid values depend on rate control mode and profile. - SampleRate *int64 `locationName:"sampleRate" type:"integer"` + // + // SampleRate is a required field + SampleRate *int64 `locationName:"sampleRate" min:"8000" type:"integer" required:"true"` // Use MPEG-2 AAC instead of MPEG-4 AAC audio for raw or MPEG-2 Transport Stream // containers. @@ -1843,6 +1853,28 @@ func (s AacSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AacSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AacSettings"} + if s.Bitrate != nil && *s.Bitrate < 6000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 6000)) + } + if s.CodingMode == nil { + invalidParams.Add(request.NewErrParamRequired("CodingMode")) + } + if s.SampleRate == nil { + invalidParams.Add(request.NewErrParamRequired("SampleRate")) + } + if s.SampleRate != nil && *s.SampleRate < 8000 { + invalidParams.Add(request.NewErrParamMinValue("SampleRate", 8000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioDescriptionBroadcasterMix sets the AudioDescriptionBroadcasterMix field's value. func (s *AacSettings) SetAudioDescriptionBroadcasterMix(v string) *AacSettings { s.AudioDescriptionBroadcasterMix = &v @@ -1903,7 +1935,7 @@ type Ac3Settings struct { _ struct{} `type:"structure"` // Average bitrate in bits/second. Valid bitrates depend on the coding mode. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + Bitrate *int64 `locationName:"bitrate" min:"64000" type:"integer"` // Specifies the "Bitstream Mode" (bsmod) for the emitted AC-3 stream. See ATSC // A/52-2012 for background on these values. @@ -1914,7 +1946,7 @@ type Ac3Settings struct { // Sets the dialnorm for the output. If blank and input audio is Dolby Digital, // dialnorm will be passed through. - Dialnorm *int64 `locationName:"dialnorm" type:"integer"` + Dialnorm *int64 `locationName:"dialnorm" min:"1" type:"integer"` // If set to FILM_STANDARD, adds dynamic range compression signaling to the // output bitstream as defined in the Dolby Digital specification. @@ -1930,7 +1962,7 @@ type Ac3Settings struct { MetadataControl *string `locationName:"metadataControl" type:"string" enum:"Ac3MetadataControl"` // Sample rate in hz. Sample rate is always 48000. - SampleRate *int64 `locationName:"sampleRate" type:"integer"` + SampleRate *int64 `locationName:"sampleRate" min:"48000" type:"integer"` } // String returns the string representation @@ -1943,6 +1975,25 @@ func (s Ac3Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Ac3Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Ac3Settings"} + if s.Bitrate != nil && *s.Bitrate < 64000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 64000)) + } + if s.Dialnorm != nil && *s.Dialnorm < 1 { + invalidParams.Add(request.NewErrParamMinValue("Dialnorm", 1)) + } + if s.SampleRate != nil && *s.SampleRate < 48000 { + invalidParams.Add(request.NewErrParamMinValue("SampleRate", 48000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBitrate sets the Bitrate field's value. func (s *Ac3Settings) SetBitrate(v int64) *Ac3Settings { s.Bitrate = &v @@ -1998,15 +2049,15 @@ type AiffSettings struct { // Specify Bit depth (BitDepth), in bits per sample, to choose the encoding // quality for this audio track. - BitDepth *int64 `locationName:"bitDepth" type:"integer"` + BitDepth *int64 `locationName:"bitDepth" min:"16" type:"integer"` // Set Channels to specify the number of channels in this output audio track. // Choosing Mono in the console will give you 1 output channel; choosing Stereo // will give you 2. In the API, valid values are 1 and 2. - Channels *int64 `locationName:"channels" type:"integer"` + Channels *int64 `locationName:"channels" min:"1" type:"integer"` // Sample rate in hz. - SampleRate *int64 `locationName:"sampleRate" type:"integer"` + SampleRate *int64 `locationName:"sampleRate" min:"8000" type:"integer"` } // String returns the string representation @@ -2019,6 +2070,25 @@ func (s AiffSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AiffSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AiffSettings"} + if s.BitDepth != nil && *s.BitDepth < 16 { + invalidParams.Add(request.NewErrParamMinValue("BitDepth", 16)) + } + if s.Channels != nil && *s.Channels < 1 { + invalidParams.Add(request.NewErrParamMinValue("Channels", 1)) + } + if s.SampleRate != nil && *s.SampleRate < 8000 { + invalidParams.Add(request.NewErrParamMinValue("SampleRate", 8000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBitDepth sets the BitDepth field's value. func (s *AiffSettings) SetBitDepth(v int64) *AiffSettings { s.BitDepth = &v @@ -2043,7 +2113,7 @@ type AncillarySourceSettings struct { // Specifies the 608 channel number in the ancillary data track from which to // extract captions. Unused for passthrough. - SourceAncillaryChannelNumber *int64 `locationName:"sourceAncillaryChannelNumber" type:"integer"` + SourceAncillaryChannelNumber *int64 `locationName:"sourceAncillaryChannelNumber" min:"1" type:"integer"` } // String returns the string representation @@ -2056,6 +2126,19 @@ func (s AncillarySourceSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AncillarySourceSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AncillarySourceSettings"} + if s.SourceAncillaryChannelNumber != nil && *s.SourceAncillaryChannelNumber < 1 { + invalidParams.Add(request.NewErrParamMinValue("SourceAncillaryChannelNumber", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetSourceAncillaryChannelNumber sets the SourceAncillaryChannelNumber field's value. func (s *AncillarySourceSettings) SetSourceAncillaryChannelNumber(v int64) *AncillarySourceSettings { s.SourceAncillaryChannelNumber = &v @@ -2072,7 +2155,12 @@ type AudioCodecSettings struct { _ struct{} `type:"structure"` // Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to - // the value AAC. + // the value AAC. The service accepts one of two mutually exclusive groups of + // AAC settings--VBR and CBR. To select one of these modes, set the value of + // Bitrate control mode (rateControlMode) to "VBR" or "CBR". In VBR mode, you + // control the audio quality with the setting VBR quality (vbrQuality). In CBR + // mode, you use the setting Bitrate (bitrate). Defaults and valid values depend + // on the rate control mode. AacSettings *AacSettings `locationName:"aacSettings" type:"structure"` // Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to @@ -2084,7 +2172,9 @@ type AudioCodecSettings struct { AiffSettings *AiffSettings `locationName:"aiffSettings" type:"structure"` // Type of Audio codec. - Codec *string `locationName:"codec" type:"string" enum:"AudioCodec"` + // + // Codec is a required field + Codec *string `locationName:"codec" type:"string" required:"true" enum:"AudioCodec"` // Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to // the value EAC3. @@ -2109,6 +2199,49 @@ func (s AudioCodecSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AudioCodecSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AudioCodecSettings"} + if s.Codec == nil { + invalidParams.Add(request.NewErrParamRequired("Codec")) + } + if s.AacSettings != nil { + if err := s.AacSettings.Validate(); err != nil { + invalidParams.AddNested("AacSettings", err.(request.ErrInvalidParams)) + } + } + if s.Ac3Settings != nil { + if err := s.Ac3Settings.Validate(); err != nil { + invalidParams.AddNested("Ac3Settings", err.(request.ErrInvalidParams)) + } + } + if s.AiffSettings != nil { + if err := s.AiffSettings.Validate(); err != nil { + invalidParams.AddNested("AiffSettings", err.(request.ErrInvalidParams)) + } + } + if s.Eac3Settings != nil { + if err := s.Eac3Settings.Validate(); err != nil { + invalidParams.AddNested("Eac3Settings", err.(request.ErrInvalidParams)) + } + } + if s.Mp2Settings != nil { + if err := s.Mp2Settings.Validate(); err != nil { + invalidParams.AddNested("Mp2Settings", err.(request.ErrInvalidParams)) + } + } + if s.WavSettings != nil { + if err := s.WavSettings.Validate(); err != nil { + invalidParams.AddNested("WavSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAacSettings sets the AacSettings field's value. func (s *AudioCodecSettings) SetAacSettings(v *AacSettings) *AudioCodecSettings { s.AacSettings = v @@ -2189,7 +2322,16 @@ type AudioDescription struct { // enum you choose, define the corresponding settings object. The following // lists the codec enum, settings object pairs. * AAC, AacSettings * MP2, Mp2Settings // * WAV, WavSettings * AIFF, AiffSettings * AC3, Ac3Settings * EAC3, Eac3Settings - CodecSettings *AudioCodecSettings `locationName:"codecSettings" type:"structure"` + // + // CodecSettings is a required field + CodecSettings *AudioCodecSettings `locationName:"codecSettings" type:"structure" required:"true"` + + // Specify the language for this audio output track, using the ISO 639-2 or + // ISO 639-3 three-letter language code. The language specified will be used + // when 'Follow Input Language Code' is not selected or when 'Follow Input Language + // Code' is selected but there is no ISO 639 language code specified by the + // input. + CustomLanguageCode *string `locationName:"customLanguageCode" min:"3" type:"string"` // Indicates the language of the audio output track. The ISO 639 language specified // in the 'Language Code' drop down will be used when 'Follow Input Language @@ -2222,6 +2364,37 @@ func (s AudioDescription) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AudioDescription) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AudioDescription"} + if s.CodecSettings == nil { + invalidParams.Add(request.NewErrParamRequired("CodecSettings")) + } + if s.CustomLanguageCode != nil && len(*s.CustomLanguageCode) < 3 { + invalidParams.Add(request.NewErrParamMinLen("CustomLanguageCode", 3)) + } + if s.AudioNormalizationSettings != nil { + if err := s.AudioNormalizationSettings.Validate(); err != nil { + invalidParams.AddNested("AudioNormalizationSettings", err.(request.ErrInvalidParams)) + } + } + if s.CodecSettings != nil { + if err := s.CodecSettings.Validate(); err != nil { + invalidParams.AddNested("CodecSettings", err.(request.ErrInvalidParams)) + } + } + if s.RemixSettings != nil { + if err := s.RemixSettings.Validate(); err != nil { + invalidParams.AddNested("RemixSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioNormalizationSettings sets the AudioNormalizationSettings field's value. func (s *AudioDescription) SetAudioNormalizationSettings(v *AudioNormalizationSettings) *AudioDescription { s.AudioNormalizationSettings = v @@ -2252,6 +2425,12 @@ func (s *AudioDescription) SetCodecSettings(v *AudioCodecSettings) *AudioDescrip return s } +// SetCustomLanguageCode sets the CustomLanguageCode field's value. +func (s *AudioDescription) SetCustomLanguageCode(v string) *AudioDescription { + s.CustomLanguageCode = &v + return s +} + // SetLanguageCode sets the LanguageCode field's value. func (s *AudioDescription) SetLanguageCode(v string) *AudioDescription { s.LanguageCode = &v @@ -2317,6 +2496,19 @@ func (s AudioNormalizationSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AudioNormalizationSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AudioNormalizationSettings"} + if s.CorrectionGateLevel != nil && *s.CorrectionGateLevel < -70 { + invalidParams.Add(request.NewErrParamMinValue("CorrectionGateLevel", -70)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAlgorithm sets the Algorithm field's value. func (s *AudioNormalizationSettings) SetAlgorithm(v string) *AudioNormalizationSettings { s.Algorithm = &v @@ -2357,10 +2549,13 @@ func (s *AudioNormalizationSettings) SetTargetLkfs(v float64) *AudioNormalizatio type AudioSelector struct { _ struct{} `type:"structure"` - // When an "Audio Description":#audio_description specifies an AudioSelector - // or AudioSelectorGroup for which no matching source is found in the input, - // then the audio selector marked as DEFAULT will be used. If none are marked - // as default, silence will be inserted for the duration of the input. + // Selects a specific language code from within an audio source, using the ISO + // 639-2 or ISO 639-3 three-letter language code + CustomLanguageCode *string `locationName:"customLanguageCode" min:"3" type:"string"` + + // Enable this setting on one audio selector to set it as the default for the + // job. The service uses this default for outputs where it can't find the specified + // input audio. If you don't set a default, those outputs have no audio. DefaultSelection *string `locationName:"defaultSelection" type:"string" enum:"AudioDefaultSelection"` // Specifies audio data from an external file source. @@ -2377,23 +2572,31 @@ type AudioSelector struct { // 0x101). Pids []*int64 `locationName:"pids" type:"list"` - // Applies only when input streams contain Dolby E. Enter the program ID (according - // to the metadata in the audio) of the Dolby E program to extract from the - // specified track. One program extracted per audio selector. To select multiple - // programs, create multiple selectors with the same Track and different Program - // numbers. "All channels" means to ignore the program IDs and include all the - // channels in this selector; useful if metadata is known to be incorrect. + // Use this setting for input streams that contain Dolby E, to have the service + // extract specific program data from the track. To select multiple programs, + // create multiple selectors with the same Track and different Program numbers. + // In the console, this setting is visible when you set Selector type to Track. + // Choose the program number from the dropdown list. If you are sending a JSON + // file, provide the program ID, which is part of the audio metadata. If your + // input file has incorrect metadata, you can choose All channels instead of + // a program number to have the service ignore the program IDs and include all + // the programs in the track. ProgramSelection *int64 `locationName:"programSelection" type:"integer"` - // Advanced audio remixing settings. + // Use these settings to reorder the audio channels of one input to match those + // of another input. This allows you to combine the two files into a single + // output, one after the other. RemixSettings *RemixSettings `locationName:"remixSettings" type:"structure"` // Specifies the type of the audio selector. SelectorType *string `locationName:"selectorType" type:"string" enum:"AudioSelectorType"` - // Identify the channel to include in this selector by entering the 1-based - // track index. To combine several tracks, enter a comma-separated list, e.g. - // "1,2,3" for tracks 1-3. + // Identify a track from the input audio to include in this selector by entering + // the track index number. To include several tracks in a single audio selector, + // specify multiple tracks as follows. Using the console, enter a comma-separated + // list. For examle, type "1,2,3" to include tracks 1 through 3. Specifying + // directly in your JSON job file, provide the track numbers in an array. For + // example, "tracks": [1,2,3]. Tracks []*int64 `locationName:"tracks" type:"list"` } @@ -2407,6 +2610,33 @@ func (s AudioSelector) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AudioSelector) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AudioSelector"} + if s.CustomLanguageCode != nil && len(*s.CustomLanguageCode) < 3 { + invalidParams.Add(request.NewErrParamMinLen("CustomLanguageCode", 3)) + } + if s.Offset != nil && *s.Offset < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Offset", -2.147483648e+09)) + } + if s.RemixSettings != nil { + if err := s.RemixSettings.Validate(); err != nil { + invalidParams.AddNested("RemixSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCustomLanguageCode sets the CustomLanguageCode field's value. +func (s *AudioSelector) SetCustomLanguageCode(v string) *AudioSelector { + s.CustomLanguageCode = &v + return s +} + // SetDefaultSelection sets the DefaultSelection field's value. func (s *AudioSelector) SetDefaultSelection(v string) *AudioSelector { s.DefaultSelection = &v @@ -2465,11 +2695,13 @@ func (s *AudioSelector) SetTracks(v []*int64) *AudioSelector { type AudioSelectorGroup struct { _ struct{} `type:"structure"` - // Name of an "Audio Selector":#inputs-audio_selector within the same input - // to include in the group. Audio selector names are standardized, based on - // their order within the input (e.g. "Audio Selector 1"). The audio_selector_name - // parameter can be repeated to add any number of audio selectors to the group. - AudioSelectorNames []*string `locationName:"audioSelectorNames" type:"list"` + // Name of an Audio Selector within the same input to include in the group. + // Audio selector names are standardized, based on their order within the input + // (e.g., "Audio Selector 1"). The audio selector name parameter can be repeated + // to add any number of audio selectors to the group. + // + // AudioSelectorNames is a required field + AudioSelectorNames []*string `locationName:"audioSelectorNames" type:"list" required:"true"` } // String returns the string representation @@ -2482,6 +2714,19 @@ func (s AudioSelectorGroup) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AudioSelectorGroup) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AudioSelectorGroup"} + if s.AudioSelectorNames == nil { + invalidParams.Add(request.NewErrParamRequired("AudioSelectorNames")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioSelectorNames sets the AudioSelectorNames field's value. func (s *AudioSelectorGroup) SetAudioSelectorNames(v []*string) *AudioSelectorGroup { s.AudioSelectorNames = v @@ -2494,7 +2739,7 @@ type AvailBlanking struct { // Blanking image to be used. Leave empty for solid black. Only bmp and png // images are supported. - AvailBlankingImage *string `locationName:"availBlankingImage" type:"string"` + AvailBlankingImage *string `locationName:"availBlankingImage" min:"14" type:"string"` } // String returns the string representation @@ -2507,6 +2752,19 @@ func (s AvailBlanking) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *AvailBlanking) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AvailBlanking"} + if s.AvailBlankingImage != nil && len(*s.AvailBlankingImage) < 14 { + invalidParams.Add(request.NewErrParamMinLen("AvailBlankingImage", 14)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAvailBlankingImage sets the AvailBlankingImage field's value. func (s *AvailBlanking) SetAvailBlankingImage(v string) *AvailBlanking { s.AvailBlankingImage = &v @@ -2525,7 +2783,9 @@ type BurninDestinationSettings struct { // This option is not valid for source captions that are STL, 608/embedded or // teletext. These source settings are already pre-defined by the caption stream. // All burn-in and DVB-Sub font settings must match. - Alignment *string `locationName:"alignment" type:"string" enum:"BurninSubtitleAlignment"` + // + // Alignment is a required field + Alignment *string `locationName:"alignment" type:"string" required:"true" enum:"BurninSubtitleAlignment"` // Specifies the color of the rectangle behind the captions.All burn-in and // DVB-Sub font settings must match. @@ -2544,11 +2804,13 @@ type BurninDestinationSettings struct { // Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.All // burn-in and DVB-Sub font settings must match. - FontOpacity *int64 `locationName:"fontOpacity" type:"integer"` + // + // FontOpacity is a required field + FontOpacity *int64 `locationName:"fontOpacity" type:"integer" required:"true"` // Font resolution in DPI (dots per inch); default is 96 dpi.All burn-in and // DVB-Sub font settings must match. - FontResolution *int64 `locationName:"fontResolution" type:"integer"` + FontResolution *int64 `locationName:"fontResolution" min:"96" type:"integer"` // A positive integer indicates the exact font size in points. Set to 0 for // automatic font size selection. All burn-in and DVB-Sub font settings must @@ -2559,13 +2821,17 @@ type BurninDestinationSettings struct { // that are either 608/embedded or teletext. These source settings are already // pre-defined by the caption stream. All burn-in and DVB-Sub font settings // must match. - OutlineColor *string `locationName:"outlineColor" type:"string" enum:"BurninSubtitleOutlineColor"` + // + // OutlineColor is a required field + OutlineColor *string `locationName:"outlineColor" type:"string" required:"true" enum:"BurninSubtitleOutlineColor"` // Specifies font outline size in pixels. This option is not valid for source // captions that are either 608/embedded or teletext. These source settings // are already pre-defined by the caption stream. All burn-in and DVB-Sub font // settings must match. - OutlineSize *int64 `locationName:"outlineSize" type:"integer"` + // + // OutlineSize is a required field + OutlineSize *int64 `locationName:"outlineSize" type:"integer" required:"true"` // Specifies the color of the shadow cast by the captions.All burn-in and DVB-Sub // font settings must match. @@ -2586,9 +2852,11 @@ type BurninDestinationSettings struct { // burn-in and DVB-Sub font settings must match. ShadowYOffset *int64 `locationName:"shadowYOffset" type:"integer"` - // Controls whether a fixed grid size or proportional font spacing will be used - // to generate the output subtitles bitmap. Only applicable for Teletext inputs - // and DVB-Sub/Burn-in outputs. + // Only applies to jobs with input captions in Teletext or STL formats. Specify + // whether the spacing between letters in your captions is set by the captions + // grid or varies depending on letter width. Choose fixed grid to conform to + // the spacing specified in the captions file more accurately. Choose proportional + // to make the text easier to read if the captions are closed caption. TeletextSpacing *string `locationName:"teletextSpacing" type:"string" enum:"BurninSubtitleTeletextSpacing"` // Specifies the horizontal position of the caption relative to the left side @@ -2620,6 +2888,37 @@ func (s BurninDestinationSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *BurninDestinationSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "BurninDestinationSettings"} + if s.Alignment == nil { + invalidParams.Add(request.NewErrParamRequired("Alignment")) + } + if s.FontOpacity == nil { + invalidParams.Add(request.NewErrParamRequired("FontOpacity")) + } + if s.FontResolution != nil && *s.FontResolution < 96 { + invalidParams.Add(request.NewErrParamMinValue("FontResolution", 96)) + } + if s.OutlineColor == nil { + invalidParams.Add(request.NewErrParamRequired("OutlineColor")) + } + if s.OutlineSize == nil { + invalidParams.Add(request.NewErrParamRequired("OutlineSize")) + } + if s.ShadowXOffset != nil && *s.ShadowXOffset < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ShadowXOffset", -2.147483648e+09)) + } + if s.ShadowYOffset != nil && *s.ShadowYOffset < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ShadowYOffset", -2.147483648e+09)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAlignment sets the Alignment field's value. func (s *BurninDestinationSettings) SetAlignment(v string) *BurninDestinationSettings { s.Alignment = &v @@ -2778,11 +3077,19 @@ type CaptionDescription struct { // input when generating captions. The name should be of the format "Caption // Selector ", which denotes that the Nth Caption Selector will be used from // each input. - CaptionSelectorName *string `locationName:"captionSelectorName" type:"string"` + // + // CaptionSelectorName is a required field + CaptionSelectorName *string `locationName:"captionSelectorName" min:"1" type:"string" required:"true"` + + // Indicates the language of the caption output track, using the ISO 639-2 or + // ISO 639-3 three-letter language code + CustomLanguageCode *string `locationName:"customLanguageCode" min:"3" type:"string"` // Specific settings required by destination type. Note that burnin_destination_settings // are not available if the source of the caption data is Embedded or Teletext. - DestinationSettings *CaptionDestinationSettings `locationName:"destinationSettings" type:"structure"` + // + // DestinationSettings is a required field + DestinationSettings *CaptionDestinationSettings `locationName:"destinationSettings" type:"structure" required:"true"` // Indicates the language of the caption output track. LanguageCode *string `locationName:"languageCode" type:"string" enum:"LanguageCode"` @@ -2803,12 +3110,45 @@ func (s CaptionDescription) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CaptionDescription) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CaptionDescription"} + if s.CaptionSelectorName == nil { + invalidParams.Add(request.NewErrParamRequired("CaptionSelectorName")) + } + if s.CaptionSelectorName != nil && len(*s.CaptionSelectorName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("CaptionSelectorName", 1)) + } + if s.CustomLanguageCode != nil && len(*s.CustomLanguageCode) < 3 { + invalidParams.Add(request.NewErrParamMinLen("CustomLanguageCode", 3)) + } + if s.DestinationSettings == nil { + invalidParams.Add(request.NewErrParamRequired("DestinationSettings")) + } + if s.DestinationSettings != nil { + if err := s.DestinationSettings.Validate(); err != nil { + invalidParams.AddNested("DestinationSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCaptionSelectorName sets the CaptionSelectorName field's value. func (s *CaptionDescription) SetCaptionSelectorName(v string) *CaptionDescription { s.CaptionSelectorName = &v return s } +// SetCustomLanguageCode sets the CustomLanguageCode field's value. +func (s *CaptionDescription) SetCustomLanguageCode(v string) *CaptionDescription { + s.CustomLanguageCode = &v + return s +} + // SetDestinationSettings sets the DestinationSettings field's value. func (s *CaptionDescription) SetDestinationSettings(v *CaptionDestinationSettings) *CaptionDescription { s.DestinationSettings = v @@ -2831,9 +3171,15 @@ func (s *CaptionDescription) SetLanguageDescription(v string) *CaptionDescriptio type CaptionDescriptionPreset struct { _ struct{} `type:"structure"` + // Indicates the language of the caption output track, using the ISO 639-2 or + // ISO 639-3 three-letter language code + CustomLanguageCode *string `locationName:"customLanguageCode" min:"3" type:"string"` + // Specific settings required by destination type. Note that burnin_destination_settings // are not available if the source of the caption data is Embedded or Teletext. - DestinationSettings *CaptionDestinationSettings `locationName:"destinationSettings" type:"structure"` + // + // DestinationSettings is a required field + DestinationSettings *CaptionDestinationSettings `locationName:"destinationSettings" type:"structure" required:"true"` // Indicates the language of the caption output track. LanguageCode *string `locationName:"languageCode" type:"string" enum:"LanguageCode"` @@ -2854,6 +3200,33 @@ func (s CaptionDescriptionPreset) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CaptionDescriptionPreset) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CaptionDescriptionPreset"} + if s.CustomLanguageCode != nil && len(*s.CustomLanguageCode) < 3 { + invalidParams.Add(request.NewErrParamMinLen("CustomLanguageCode", 3)) + } + if s.DestinationSettings == nil { + invalidParams.Add(request.NewErrParamRequired("DestinationSettings")) + } + if s.DestinationSettings != nil { + if err := s.DestinationSettings.Validate(); err != nil { + invalidParams.AddNested("DestinationSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCustomLanguageCode sets the CustomLanguageCode field's value. +func (s *CaptionDescriptionPreset) SetCustomLanguageCode(v string) *CaptionDescriptionPreset { + s.CustomLanguageCode = &v + return s +} + // SetDestinationSettings sets the DestinationSettings field's value. func (s *CaptionDescriptionPreset) SetDestinationSettings(v *CaptionDestinationSettings) *CaptionDescriptionPreset { s.DestinationSettings = v @@ -2882,7 +3255,9 @@ type CaptionDestinationSettings struct { // Type of Caption output, including Burn-In, Embedded, SCC, SRT, TTML, WebVTT, // DVB-Sub, Teletext. - DestinationType *string `locationName:"destinationType" type:"string" enum:"CaptionDestinationType"` + // + // DestinationType is a required field + DestinationType *string `locationName:"destinationType" type:"string" required:"true" enum:"CaptionDestinationType"` // DVB-Sub Destination Settings DvbSubDestinationSettings *DvbSubDestinationSettings `locationName:"dvbSubDestinationSettings" type:"structure"` @@ -2908,6 +3283,34 @@ func (s CaptionDestinationSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CaptionDestinationSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CaptionDestinationSettings"} + if s.DestinationType == nil { + invalidParams.Add(request.NewErrParamRequired("DestinationType")) + } + if s.BurninDestinationSettings != nil { + if err := s.BurninDestinationSettings.Validate(); err != nil { + invalidParams.AddNested("BurninDestinationSettings", err.(request.ErrInvalidParams)) + } + } + if s.DvbSubDestinationSettings != nil { + if err := s.DvbSubDestinationSettings.Validate(); err != nil { + invalidParams.AddNested("DvbSubDestinationSettings", err.(request.ErrInvalidParams)) + } + } + if s.TeletextDestinationSettings != nil { + if err := s.TeletextDestinationSettings.Validate(); err != nil { + invalidParams.AddNested("TeletextDestinationSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBurninDestinationSettings sets the BurninDestinationSettings field's value. func (s *CaptionDestinationSettings) SetBurninDestinationSettings(v *BurninDestinationSettings) *CaptionDestinationSettings { s.BurninDestinationSettings = v @@ -2944,10 +3347,19 @@ func (s *CaptionDestinationSettings) SetTtmlDestinationSettings(v *TtmlDestinati return s } -// Caption inputs to be mapped to caption outputs. +// Set up captions in your outputs by first selecting them from your input here. type CaptionSelector struct { _ struct{} `type:"structure"` + // The specific language to extract from source, using the ISO 639-2 or ISO + // 639-3 three-letter language code. If input is SCTE-27, complete this field + // and/or PID to select the caption language to extract. If input is DVB-Sub + // and output is Burn-in or SMPTE-TT, complete this field and/or PID to select + // the caption language to extract. If input is DVB-Sub that is being passed + // through, omit this field (and PID field); there is no way to extract a specific + // language with pass-through captions. + CustomLanguageCode *string `locationName:"customLanguageCode" min:"3" type:"string"` + // The specific language to extract from source. If input is SCTE-27, complete // this field and/or PID to select the caption language to extract. If input // is DVB-Sub and output is Burn-in or SMPTE-TT, complete this field and/or @@ -2958,7 +3370,9 @@ type CaptionSelector struct { // Source settings (SourceSettings) contains the group of settings for captions // in the input. - SourceSettings *CaptionSourceSettings `locationName:"sourceSettings" type:"structure"` + // + // SourceSettings is a required field + SourceSettings *CaptionSourceSettings `locationName:"sourceSettings" type:"structure" required:"true"` } // String returns the string representation @@ -2971,6 +3385,33 @@ func (s CaptionSelector) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CaptionSelector) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CaptionSelector"} + if s.CustomLanguageCode != nil && len(*s.CustomLanguageCode) < 3 { + invalidParams.Add(request.NewErrParamMinLen("CustomLanguageCode", 3)) + } + if s.SourceSettings == nil { + invalidParams.Add(request.NewErrParamRequired("SourceSettings")) + } + if s.SourceSettings != nil { + if err := s.SourceSettings.Validate(); err != nil { + invalidParams.AddNested("SourceSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCustomLanguageCode sets the CustomLanguageCode field's value. +func (s *CaptionSelector) SetCustomLanguageCode(v string) *CaptionSelector { + s.CustomLanguageCode = &v + return s +} + // SetLanguageCode sets the LanguageCode field's value. func (s *CaptionSelector) SetLanguageCode(v string) *CaptionSelector { s.LanguageCode = &v @@ -3002,7 +3443,9 @@ type CaptionSourceSettings struct { // Use Source (SourceType) to identify the format of your input captions. The // service cannot auto-detect caption format. - SourceType *string `locationName:"sourceType" type:"string" enum:"CaptionSourceType"` + // + // SourceType is a required field + SourceType *string `locationName:"sourceType" type:"string" required:"true" enum:"CaptionSourceType"` // Settings specific to Teletext caption sources, including Page number. TeletextSourceSettings *TeletextSourceSettings `locationName:"teletextSourceSettings" type:"structure"` @@ -3018,6 +3461,44 @@ func (s CaptionSourceSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CaptionSourceSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CaptionSourceSettings"} + if s.SourceType == nil { + invalidParams.Add(request.NewErrParamRequired("SourceType")) + } + if s.AncillarySourceSettings != nil { + if err := s.AncillarySourceSettings.Validate(); err != nil { + invalidParams.AddNested("AncillarySourceSettings", err.(request.ErrInvalidParams)) + } + } + if s.DvbSubSourceSettings != nil { + if err := s.DvbSubSourceSettings.Validate(); err != nil { + invalidParams.AddNested("DvbSubSourceSettings", err.(request.ErrInvalidParams)) + } + } + if s.EmbeddedSourceSettings != nil { + if err := s.EmbeddedSourceSettings.Validate(); err != nil { + invalidParams.AddNested("EmbeddedSourceSettings", err.(request.ErrInvalidParams)) + } + } + if s.FileSourceSettings != nil { + if err := s.FileSourceSettings.Validate(); err != nil { + invalidParams.AddNested("FileSourceSettings", err.(request.ErrInvalidParams)) + } + } + if s.TeletextSourceSettings != nil { + if err := s.TeletextSourceSettings.Validate(); err != nil { + invalidParams.AddNested("TeletextSourceSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAncillarySourceSettings sets the AncillarySourceSettings field's value. func (s *CaptionSourceSettings) SetAncillarySourceSettings(v *AncillarySourceSettings) *CaptionSourceSettings { s.AncillarySourceSettings = v @@ -3062,7 +3543,9 @@ type ChannelMapping struct { _ struct{} `type:"structure"` // List of output channels - OutputChannels []*OutputChannelMapping `locationName:"outputChannels" type:"list"` + // + // OutputChannels is a required field + OutputChannels []*OutputChannelMapping `locationName:"outputChannels" type:"list" required:"true"` } // String returns the string representation @@ -3075,18 +3558,328 @@ func (s ChannelMapping) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *ChannelMapping) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ChannelMapping"} + if s.OutputChannels == nil { + invalidParams.Add(request.NewErrParamRequired("OutputChannels")) + } + if s.OutputChannels != nil { + for i, v := range s.OutputChannels { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "OutputChannels", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetOutputChannels sets the OutputChannels field's value. func (s *ChannelMapping) SetOutputChannels(v []*OutputChannelMapping) *ChannelMapping { s.OutputChannels = v return s } +// Settings for CMAF encryption +type CmafEncryptionSettings struct { + _ struct{} `type:"structure"` + + // This is a 128-bit, 16-byte hex value represented by a 32-character text string. + // If this parameter is not set then the Initialization Vector will follow the + // segment number by default. + ConstantInitializationVector *string `locationName:"constantInitializationVector" min:"32" type:"string"` + + // Encrypts the segments with the given encryption scheme. Leave blank to disable. + // Selecting 'Disabled' in the web interface also disables encryption. + EncryptionMethod *string `locationName:"encryptionMethod" type:"string" enum:"CmafEncryptionType"` + + // The Initialization Vector is a 128-bit number used in conjunction with the + // key for encrypting blocks. If set to INCLUDE, Initialization Vector is listed + // in the manifest. Otherwise Initialization Vector is not in the manifest. + InitializationVectorInManifest *string `locationName:"initializationVectorInManifest" type:"string" enum:"CmafInitializationVectorInManifest"` + + // Settings for use with a SPEKE key provider. + StaticKeyProvider *StaticKeyProvider `locationName:"staticKeyProvider" type:"structure"` + + // Indicates which type of key provider is used for encryption. + // + // Type is a required field + Type *string `locationName:"type" type:"string" required:"true" enum:"CmafKeyProviderType"` +} + +// String returns the string representation +func (s CmafEncryptionSettings) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CmafEncryptionSettings) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CmafEncryptionSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CmafEncryptionSettings"} + if s.ConstantInitializationVector != nil && len(*s.ConstantInitializationVector) < 32 { + invalidParams.Add(request.NewErrParamMinLen("ConstantInitializationVector", 32)) + } + if s.Type == nil { + invalidParams.Add(request.NewErrParamRequired("Type")) + } + if s.StaticKeyProvider != nil { + if err := s.StaticKeyProvider.Validate(); err != nil { + invalidParams.AddNested("StaticKeyProvider", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetConstantInitializationVector sets the ConstantInitializationVector field's value. +func (s *CmafEncryptionSettings) SetConstantInitializationVector(v string) *CmafEncryptionSettings { + s.ConstantInitializationVector = &v + return s +} + +// SetEncryptionMethod sets the EncryptionMethod field's value. +func (s *CmafEncryptionSettings) SetEncryptionMethod(v string) *CmafEncryptionSettings { + s.EncryptionMethod = &v + return s +} + +// SetInitializationVectorInManifest sets the InitializationVectorInManifest field's value. +func (s *CmafEncryptionSettings) SetInitializationVectorInManifest(v string) *CmafEncryptionSettings { + s.InitializationVectorInManifest = &v + return s +} + +// SetStaticKeyProvider sets the StaticKeyProvider field's value. +func (s *CmafEncryptionSettings) SetStaticKeyProvider(v *StaticKeyProvider) *CmafEncryptionSettings { + s.StaticKeyProvider = v + return s +} + +// SetType sets the Type field's value. +func (s *CmafEncryptionSettings) SetType(v string) *CmafEncryptionSettings { + s.Type = &v + return s +} + +// Required when you set (Type) under (OutputGroups)>(OutputGroupSettings) to +// CMAF_GROUP_SETTINGS. Each output in a CMAF Output Group may only contain +// a single video, audio, or caption output. +type CmafGroupSettings struct { + _ struct{} `type:"structure"` + + // A partial URI prefix that will be put in the manifest file at the top level + // BaseURL element. Can be used if streams are delivered from a different URL + // than the manifest file. + BaseUrl *string `locationName:"baseUrl" type:"string"` + + // When set to ENABLED, sets #EXT-X-ALLOW-CACHE:no tag, which prevents client + // from saving media segments for later replay. + ClientCache *string `locationName:"clientCache" type:"string" enum:"CmafClientCache"` + + // Specification to use (RFC-6381 or the default RFC-4281) during m3u8 playlist + // generation. + CodecSpecification *string `locationName:"codecSpecification" type:"string" enum:"CmafCodecSpecification"` + + // Use Destination (Destination) to specify the S3 output location and the output + // filename base. Destination accepts format identifiers. If you do not specify + // the base filename in the URI, the service will use the filename of the input + // file. If your job has multiple inputs, the service uses the filename of the + // first input file. + Destination *string `locationName:"destination" type:"string"` + + // DRM settings. + Encryption *CmafEncryptionSettings `locationName:"encryption" type:"structure"` + + // Length of fragments to generate (in seconds). Fragment length must be compatible + // with GOP size and Framerate. Note that fragments will end on the next keyframe + // after this number of seconds, so actual fragment length may be longer. When + // Emit Single File is checked, the fragmentation is internal to a single output + // file and it does not cause the creation of many output files as in other + // output types. + // + // FragmentLength is a required field + FragmentLength *int64 `locationName:"fragmentLength" min:"1" type:"integer" required:"true"` + + // When set to GZIP, compresses HLS playlist. + ManifestCompression *string `locationName:"manifestCompression" type:"string" enum:"CmafManifestCompression"` + + // Indicates whether the output manifest should use floating point values for + // segment duration. + ManifestDurationFormat *string `locationName:"manifestDurationFormat" type:"string" enum:"CmafManifestDurationFormat"` + + // Minimum time of initially buffered media that is needed to ensure smooth + // playout. + MinBufferTime *int64 `locationName:"minBufferTime" type:"integer"` + + // When set to SINGLE_FILE, a single output file is generated, which is internally + // segmented using the Fragment Length and Segment Length. When set to SEGMENTED_FILES, + // separate segment files will be created. + SegmentControl *string `locationName:"segmentControl" type:"string" enum:"CmafSegmentControl"` + + // Use this setting to specify the length, in seconds, of each individual CMAF + // segment. This value applies to the whole package; that is, to every output + // in the output group. Note that segments end on the first keyframe after this + // number of seconds, so the actual segment length might be slightly longer. + // If you set Segment control (CmafSegmentControl) to single file, the service + // puts the content of each output in a single file that has metadata that marks + // these segments. If you set it to segmented files, the service creates multiple + // files for each output, each with the content of one segment. + // + // SegmentLength is a required field + SegmentLength *int64 `locationName:"segmentLength" min:"1" type:"integer" required:"true"` + + // Include or exclude RESOLUTION attribute for video in EXT-X-STREAM-INF tag + // of variant manifest. + StreamInfResolution *string `locationName:"streamInfResolution" type:"string" enum:"CmafStreamInfResolution"` + + // When set to ENABLED, a DASH MPD manifest will be generated for this output. + WriteDashManifest *string `locationName:"writeDashManifest" type:"string" enum:"CmafWriteDASHManifest"` + + // When set to ENABLED, an Apple HLS manifest will be generated for this output. + WriteHlsManifest *string `locationName:"writeHlsManifest" type:"string" enum:"CmafWriteHLSManifest"` +} + +// String returns the string representation +func (s CmafGroupSettings) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CmafGroupSettings) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CmafGroupSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CmafGroupSettings"} + if s.FragmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("FragmentLength")) + } + if s.FragmentLength != nil && *s.FragmentLength < 1 { + invalidParams.Add(request.NewErrParamMinValue("FragmentLength", 1)) + } + if s.SegmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("SegmentLength")) + } + if s.SegmentLength != nil && *s.SegmentLength < 1 { + invalidParams.Add(request.NewErrParamMinValue("SegmentLength", 1)) + } + if s.Encryption != nil { + if err := s.Encryption.Validate(); err != nil { + invalidParams.AddNested("Encryption", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetBaseUrl sets the BaseUrl field's value. +func (s *CmafGroupSettings) SetBaseUrl(v string) *CmafGroupSettings { + s.BaseUrl = &v + return s +} + +// SetClientCache sets the ClientCache field's value. +func (s *CmafGroupSettings) SetClientCache(v string) *CmafGroupSettings { + s.ClientCache = &v + return s +} + +// SetCodecSpecification sets the CodecSpecification field's value. +func (s *CmafGroupSettings) SetCodecSpecification(v string) *CmafGroupSettings { + s.CodecSpecification = &v + return s +} + +// SetDestination sets the Destination field's value. +func (s *CmafGroupSettings) SetDestination(v string) *CmafGroupSettings { + s.Destination = &v + return s +} + +// SetEncryption sets the Encryption field's value. +func (s *CmafGroupSettings) SetEncryption(v *CmafEncryptionSettings) *CmafGroupSettings { + s.Encryption = v + return s +} + +// SetFragmentLength sets the FragmentLength field's value. +func (s *CmafGroupSettings) SetFragmentLength(v int64) *CmafGroupSettings { + s.FragmentLength = &v + return s +} + +// SetManifestCompression sets the ManifestCompression field's value. +func (s *CmafGroupSettings) SetManifestCompression(v string) *CmafGroupSettings { + s.ManifestCompression = &v + return s +} + +// SetManifestDurationFormat sets the ManifestDurationFormat field's value. +func (s *CmafGroupSettings) SetManifestDurationFormat(v string) *CmafGroupSettings { + s.ManifestDurationFormat = &v + return s +} + +// SetMinBufferTime sets the MinBufferTime field's value. +func (s *CmafGroupSettings) SetMinBufferTime(v int64) *CmafGroupSettings { + s.MinBufferTime = &v + return s +} + +// SetSegmentControl sets the SegmentControl field's value. +func (s *CmafGroupSettings) SetSegmentControl(v string) *CmafGroupSettings { + s.SegmentControl = &v + return s +} + +// SetSegmentLength sets the SegmentLength field's value. +func (s *CmafGroupSettings) SetSegmentLength(v int64) *CmafGroupSettings { + s.SegmentLength = &v + return s +} + +// SetStreamInfResolution sets the StreamInfResolution field's value. +func (s *CmafGroupSettings) SetStreamInfResolution(v string) *CmafGroupSettings { + s.StreamInfResolution = &v + return s +} + +// SetWriteDashManifest sets the WriteDashManifest field's value. +func (s *CmafGroupSettings) SetWriteDashManifest(v string) *CmafGroupSettings { + s.WriteDashManifest = &v + return s +} + +// SetWriteHlsManifest sets the WriteHlsManifest field's value. +func (s *CmafGroupSettings) SetWriteHlsManifest(v string) *CmafGroupSettings { + s.WriteHlsManifest = &v + return s +} + // Settings for color correction. type ColorCorrector struct { _ struct{} `type:"structure"` // Brightness level. - Brightness *int64 `locationName:"brightness" type:"integer"` + Brightness *int64 `locationName:"brightness" min:"1" type:"integer"` // Determines if colorspace conversion will be performed. If set to _None_, // no conversion will be performed. If _Force 601_ or _Force 709_ are selected, @@ -3096,19 +3889,22 @@ type ColorCorrector struct { ColorSpaceConversion *string `locationName:"colorSpaceConversion" type:"string" enum:"ColorSpaceConversion"` // Contrast level. - Contrast *int64 `locationName:"contrast" type:"integer"` + Contrast *int64 `locationName:"contrast" min:"1" type:"integer"` - // Use the HDR master display (Hdr10Metadata) settings to provide values for - // HDR color. These values vary depending on the input video and must be provided - // by a color grader. Range is 0 to 50,000, each increment represents 0.00002 - // in CIE1931 color coordinate. + // Use the HDR master display (Hdr10Metadata) settings to correct HDR metadata + // or to provide missing metadata. These values vary depending on the input + // video and must be provided by a color grader. Range is 0 to 50,000, each + // increment represents 0.00002 in CIE1931 color coordinate. Note that these + // settings are not color correction. Note that if you are creating HDR outputs + // inside of an HLS CMAF package, to comply with the Apple specification, you + // must use the HVC1 for H.265 setting. Hdr10Metadata *Hdr10Metadata `locationName:"hdr10Metadata" type:"structure"` // Hue in degrees. Hue *int64 `locationName:"hue" type:"integer"` // Saturation level. - Saturation *int64 `locationName:"saturation" type:"integer"` + Saturation *int64 `locationName:"saturation" min:"1" type:"integer"` } // String returns the string representation @@ -3121,6 +3917,33 @@ func (s ColorCorrector) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *ColorCorrector) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ColorCorrector"} + if s.Brightness != nil && *s.Brightness < 1 { + invalidParams.Add(request.NewErrParamMinValue("Brightness", 1)) + } + if s.Contrast != nil && *s.Contrast < 1 { + invalidParams.Add(request.NewErrParamMinValue("Contrast", 1)) + } + if s.Hue != nil && *s.Hue < -180 { + invalidParams.Add(request.NewErrParamMinValue("Hue", -180)) + } + if s.Saturation != nil && *s.Saturation < 1 { + invalidParams.Add(request.NewErrParamMinValue("Saturation", 1)) + } + if s.Hdr10Metadata != nil { + if err := s.Hdr10Metadata.Validate(); err != nil { + invalidParams.AddNested("Hdr10Metadata", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBrightness sets the Brightness field's value. func (s *ColorCorrector) SetBrightness(v int64) *ColorCorrector { s.Brightness = &v @@ -3163,7 +3986,9 @@ type ContainerSettings struct { // Container for this output. Some containers require a container settings object. // If not specified, the default object will be created. - Container *string `locationName:"container" type:"string" enum:"ContainerType"` + // + // Container is a required field + Container *string `locationName:"container" type:"string" required:"true" enum:"ContainerType"` // Settings for F4v container F4vSettings *F4vSettings `locationName:"f4vSettings" type:"structure"` @@ -3191,6 +4016,29 @@ func (s ContainerSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *ContainerSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ContainerSettings"} + if s.Container == nil { + invalidParams.Add(request.NewErrParamRequired("Container")) + } + if s.M2tsSettings != nil { + if err := s.M2tsSettings.Validate(); err != nil { + invalidParams.AddNested("M2tsSettings", err.(request.ErrInvalidParams)) + } + } + if s.M3u8Settings != nil { + if err := s.M3u8Settings.Validate(); err != nil { + invalidParams.AddNested("M3u8Settings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetContainer sets the Container field's value. func (s *ContainerSettings) SetContainer(v string) *ContainerSettings { s.Container = &v @@ -3246,10 +4094,14 @@ type CreateJobInput struct { // Required. The IAM role you use for creating this job. For details about permissions, // see the User Guide topic at the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html. - Role *string `locationName:"role" type:"string"` + // + // Role is a required field + Role *string `locationName:"role" type:"string" required:"true"` // JobSettings contains all the transcode settings for a job. - Settings *JobSettings `locationName:"settings" type:"structure"` + // + // Settings is a required field + Settings *JobSettings `locationName:"settings" type:"structure" required:"true"` // User-defined metadata that you want to associate with an MediaConvert job. // You specify metadata in key/value pairs. @@ -3266,6 +4118,27 @@ func (s CreateJobInput) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateJobInput"} + if s.Role == nil { + invalidParams.Add(request.NewErrParamRequired("Role")) + } + if s.Settings == nil { + invalidParams.Add(request.NewErrParamRequired("Settings")) + } + if s.Settings != nil { + if err := s.Settings.Validate(); err != nil { + invalidParams.AddNested("Settings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetClientRequestToken sets the ClientRequestToken field's value. func (s *CreateJobInput) SetClientRequestToken(v string) *CreateJobInput { s.ClientRequestToken = &v @@ -3340,7 +4213,9 @@ type CreateJobTemplateInput struct { Description *string `locationName:"description" type:"string"` // The name of the job template you are creating. - Name *string `locationName:"name" type:"string"` + // + // Name is a required field + Name *string `locationName:"name" type:"string" required:"true"` // Optional. The queue that jobs created from this template are assigned to. // If you don't specify this, jobs will go to the default queue. @@ -3348,7 +4223,9 @@ type CreateJobTemplateInput struct { // JobTemplateSettings contains all the transcode settings saved in the template // that will be applied to jobs created from it. - Settings *JobTemplateSettings `locationName:"settings" type:"structure"` + // + // Settings is a required field + Settings *JobTemplateSettings `locationName:"settings" type:"structure" required:"true"` } // String returns the string representation @@ -3361,6 +4238,27 @@ func (s CreateJobTemplateInput) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateJobTemplateInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateJobTemplateInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + if s.Settings == nil { + invalidParams.Add(request.NewErrParamRequired("Settings")) + } + if s.Settings != nil { + if err := s.Settings.Validate(); err != nil { + invalidParams.AddNested("Settings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCategory sets the Category field's value. func (s *CreateJobTemplateInput) SetCategory(v string) *CreateJobTemplateInput { s.Category = &v @@ -3428,10 +4326,14 @@ type CreatePresetInput struct { Description *string `locationName:"description" type:"string"` // The name of the preset you are creating. - Name *string `locationName:"name" type:"string"` + // + // Name is a required field + Name *string `locationName:"name" type:"string" required:"true"` // Settings for preset - Settings *PresetSettings `locationName:"settings" type:"structure"` + // + // Settings is a required field + Settings *PresetSettings `locationName:"settings" type:"structure" required:"true"` } // String returns the string representation @@ -3444,6 +4346,27 @@ func (s CreatePresetInput) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreatePresetInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreatePresetInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + if s.Settings == nil { + invalidParams.Add(request.NewErrParamRequired("Settings")) + } + if s.Settings != nil { + if err := s.Settings.Validate(); err != nil { + invalidParams.AddNested("Settings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCategory sets the Category field's value. func (s *CreatePresetInput) SetCategory(v string) *CreatePresetInput { s.Category = &v @@ -3501,7 +4424,9 @@ type CreateQueueInput struct { Description *string `locationName:"description" type:"string"` // The name of the queue you are creating. - Name *string `locationName:"name" type:"string"` + // + // Name is a required field + Name *string `locationName:"name" type:"string" required:"true"` } // String returns the string representation @@ -3514,6 +4439,19 @@ func (s CreateQueueInput) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateQueueInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateQueueInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetDescription sets the Description field's value. func (s *CreateQueueInput) SetDescription(v string) *CreateQueueInput { s.Description = &v @@ -3558,7 +4496,9 @@ type DashIsoEncryptionSettings struct { _ struct{} `type:"structure"` // Settings for use with a SPEKE key provider - SpekeKeyProvider *SpekeKeyProvider `locationName:"spekeKeyProvider" type:"structure"` + // + // SpekeKeyProvider is a required field + SpekeKeyProvider *SpekeKeyProvider `locationName:"spekeKeyProvider" type:"structure" required:"true"` } // String returns the string representation @@ -3571,6 +4511,24 @@ func (s DashIsoEncryptionSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DashIsoEncryptionSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DashIsoEncryptionSettings"} + if s.SpekeKeyProvider == nil { + invalidParams.Add(request.NewErrParamRequired("SpekeKeyProvider")) + } + if s.SpekeKeyProvider != nil { + if err := s.SpekeKeyProvider.Validate(); err != nil { + invalidParams.AddNested("SpekeKeyProvider", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetSpekeKeyProvider sets the SpekeKeyProvider field's value. func (s *DashIsoEncryptionSettings) SetSpekeKeyProvider(v *SpekeKeyProvider) *DashIsoEncryptionSettings { s.SpekeKeyProvider = v @@ -3603,7 +4561,9 @@ type DashIsoGroupSettings struct { // Emit Single File is checked, the fragmentation is internal to a single output // file and it does not cause the creation of many output files as in other // output types. - FragmentLength *int64 `locationName:"fragmentLength" type:"integer"` + // + // FragmentLength is a required field + FragmentLength *int64 `locationName:"fragmentLength" min:"1" type:"integer" required:"true"` // Supports HbbTV specification as indicated HbbtvCompliance *string `locationName:"hbbtvCompliance" type:"string" enum:"DashIsoHbbtvCompliance"` @@ -3622,7 +4582,9 @@ type DashIsoGroupSettings struct { // may be longer. When Emit Single File is checked, the segmentation is internal // to a single output file and it does not cause the creation of many output // files as in other output types. - SegmentLength *int64 `locationName:"segmentLength" type:"integer"` + // + // SegmentLength is a required field + SegmentLength *int64 `locationName:"segmentLength" min:"1" type:"integer" required:"true"` } // String returns the string representation @@ -3635,6 +4597,33 @@ func (s DashIsoGroupSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DashIsoGroupSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DashIsoGroupSettings"} + if s.FragmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("FragmentLength")) + } + if s.FragmentLength != nil && *s.FragmentLength < 1 { + invalidParams.Add(request.NewErrParamMinValue("FragmentLength", 1)) + } + if s.SegmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("SegmentLength")) + } + if s.SegmentLength != nil && *s.SegmentLength < 1 { + invalidParams.Add(request.NewErrParamMinValue("SegmentLength", 1)) + } + if s.Encryption != nil { + if err := s.Encryption.Validate(); err != nil { + invalidParams.AddNested("Encryption", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBaseUrl sets the BaseUrl field's value. func (s *DashIsoGroupSettings) SetBaseUrl(v string) *DashIsoGroupSettings { s.BaseUrl = &v @@ -3979,15 +4968,21 @@ type DvbNitSettings struct { _ struct{} `type:"structure"` // The numeric value placed in the Network Information Table (NIT). - NetworkId *int64 `locationName:"networkId" type:"integer"` + // + // NetworkId is a required field + NetworkId *int64 `locationName:"networkId" type:"integer" required:"true"` // The network name text placed in the network_name_descriptor inside the Network // Information Table. Maximum length is 256 characters. - NetworkName *string `locationName:"networkName" type:"string"` + // + // NetworkName is a required field + NetworkName *string `locationName:"networkName" min:"1" type:"string" required:"true"` // The number of milliseconds between instances of this table in the output // transport stream. - NitInterval *int64 `locationName:"nitInterval" type:"integer"` + // + // NitInterval is a required field + NitInterval *int64 `locationName:"nitInterval" min:"25" type:"integer" required:"true"` } // String returns the string representation @@ -4000,6 +4995,31 @@ func (s DvbNitSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DvbNitSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DvbNitSettings"} + if s.NetworkId == nil { + invalidParams.Add(request.NewErrParamRequired("NetworkId")) + } + if s.NetworkName == nil { + invalidParams.Add(request.NewErrParamRequired("NetworkName")) + } + if s.NetworkName != nil && len(*s.NetworkName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NetworkName", 1)) + } + if s.NitInterval == nil { + invalidParams.Add(request.NewErrParamRequired("NitInterval")) + } + if s.NitInterval != nil && *s.NitInterval < 25 { + invalidParams.Add(request.NewErrParamMinValue("NitInterval", 25)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetNetworkId sets the NetworkId field's value. func (s *DvbNitSettings) SetNetworkId(v int64) *DvbNitSettings { s.NetworkId = &v @@ -4033,15 +5053,15 @@ type DvbSdtSettings struct { // The number of milliseconds between instances of this table in the output // transport stream. - SdtInterval *int64 `locationName:"sdtInterval" type:"integer"` + SdtInterval *int64 `locationName:"sdtInterval" min:"25" type:"integer"` // The service name placed in the service_descriptor in the Service Description // Table. Maximum length is 256 characters. - ServiceName *string `locationName:"serviceName" type:"string"` + ServiceName *string `locationName:"serviceName" min:"1" type:"string"` // The service provider name placed in the service_descriptor in the Service // Description Table. Maximum length is 256 characters. - ServiceProviderName *string `locationName:"serviceProviderName" type:"string"` + ServiceProviderName *string `locationName:"serviceProviderName" min:"1" type:"string"` } // String returns the string representation @@ -4054,6 +5074,25 @@ func (s DvbSdtSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DvbSdtSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DvbSdtSettings"} + if s.SdtInterval != nil && *s.SdtInterval < 25 { + invalidParams.Add(request.NewErrParamMinValue("SdtInterval", 25)) + } + if s.ServiceName != nil && len(*s.ServiceName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ServiceName", 1)) + } + if s.ServiceProviderName != nil && len(*s.ServiceProviderName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ServiceProviderName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetOutputSdt sets the OutputSdt field's value. func (s *DvbSdtSettings) SetOutputSdt(v string) *DvbSdtSettings { s.OutputSdt = &v @@ -4090,7 +5129,9 @@ type DvbSubDestinationSettings struct { // This option is not valid for source captions that are STL, 608/embedded or // teletext. These source settings are already pre-defined by the caption stream. // All burn-in and DVB-Sub font settings must match. - Alignment *string `locationName:"alignment" type:"string" enum:"DvbSubtitleAlignment"` + // + // Alignment is a required field + Alignment *string `locationName:"alignment" type:"string" required:"true" enum:"DvbSubtitleAlignment"` // Specifies the color of the rectangle behind the captions.All burn-in and // DVB-Sub font settings must match. @@ -4109,11 +5150,13 @@ type DvbSubDestinationSettings struct { // Specifies the opacity of the burned-in captions. 255 is opaque; 0 is transparent.All // burn-in and DVB-Sub font settings must match. - FontOpacity *int64 `locationName:"fontOpacity" type:"integer"` + // + // FontOpacity is a required field + FontOpacity *int64 `locationName:"fontOpacity" type:"integer" required:"true"` // Font resolution in DPI (dots per inch); default is 96 dpi.All burn-in and // DVB-Sub font settings must match. - FontResolution *int64 `locationName:"fontResolution" type:"integer"` + FontResolution *int64 `locationName:"fontResolution" min:"96" type:"integer"` // A positive integer indicates the exact font size in points. Set to 0 for // automatic font size selection. All burn-in and DVB-Sub font settings must @@ -4124,13 +5167,17 @@ type DvbSubDestinationSettings struct { // that are either 608/embedded or teletext. These source settings are already // pre-defined by the caption stream. All burn-in and DVB-Sub font settings // must match. - OutlineColor *string `locationName:"outlineColor" type:"string" enum:"DvbSubtitleOutlineColor"` + // + // OutlineColor is a required field + OutlineColor *string `locationName:"outlineColor" type:"string" required:"true" enum:"DvbSubtitleOutlineColor"` // Specifies font outline size in pixels. This option is not valid for source // captions that are either 608/embedded or teletext. These source settings // are already pre-defined by the caption stream. All burn-in and DVB-Sub font // settings must match. - OutlineSize *int64 `locationName:"outlineSize" type:"integer"` + // + // OutlineSize is a required field + OutlineSize *int64 `locationName:"outlineSize" type:"integer" required:"true"` // Specifies the color of the shadow cast by the captions.All burn-in and DVB-Sub // font settings must match. @@ -4151,9 +5198,11 @@ type DvbSubDestinationSettings struct { // burn-in and DVB-Sub font settings must match. ShadowYOffset *int64 `locationName:"shadowYOffset" type:"integer"` - // Controls whether a fixed grid size or proportional font spacing will be used - // to generate the output subtitles bitmap. Only applicable for Teletext inputs - // and DVB-Sub/Burn-in outputs. + // Only applies to jobs with input captions in Teletext or STL formats. Specify + // whether the spacing between letters in your captions is set by the captions + // grid or varies depending on letter width. Choose fixed grid to conform to + // the spacing specified in the captions file more accurately. Choose proportional + // to make the text easier to read if the captions are closed caption. TeletextSpacing *string `locationName:"teletextSpacing" type:"string" enum:"DvbSubtitleTeletextSpacing"` // Specifies the horizontal position of the caption relative to the left side @@ -4185,6 +5234,37 @@ func (s DvbSubDestinationSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DvbSubDestinationSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DvbSubDestinationSettings"} + if s.Alignment == nil { + invalidParams.Add(request.NewErrParamRequired("Alignment")) + } + if s.FontOpacity == nil { + invalidParams.Add(request.NewErrParamRequired("FontOpacity")) + } + if s.FontResolution != nil && *s.FontResolution < 96 { + invalidParams.Add(request.NewErrParamMinValue("FontResolution", 96)) + } + if s.OutlineColor == nil { + invalidParams.Add(request.NewErrParamRequired("OutlineColor")) + } + if s.OutlineSize == nil { + invalidParams.Add(request.NewErrParamRequired("OutlineSize")) + } + if s.ShadowXOffset != nil && *s.ShadowXOffset < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ShadowXOffset", -2.147483648e+09)) + } + if s.ShadowYOffset != nil && *s.ShadowYOffset < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ShadowYOffset", -2.147483648e+09)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAlignment sets the Alignment field's value. func (s *DvbSubDestinationSettings) SetAlignment(v string) *DvbSubDestinationSettings { s.Alignment = &v @@ -4288,7 +5368,7 @@ type DvbSubSourceSettings struct { // When using DVB-Sub with Burn-In or SMPTE-TT, use this PID for the source // content. Unused for DVB-Sub passthrough. All DVB-Sub content is passed through, // regardless of selectors. - Pid *int64 `locationName:"pid" type:"integer"` + Pid *int64 `locationName:"pid" min:"1" type:"integer"` } // String returns the string representation @@ -4301,6 +5381,19 @@ func (s DvbSubSourceSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DvbSubSourceSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DvbSubSourceSettings"} + if s.Pid != nil && *s.Pid < 1 { + invalidParams.Add(request.NewErrParamMinValue("Pid", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetPid sets the Pid field's value. func (s *DvbSubSourceSettings) SetPid(v int64) *DvbSubSourceSettings { s.Pid = &v @@ -4313,7 +5406,9 @@ type DvbTdtSettings struct { // The number of milliseconds between instances of this table in the output // transport stream. - TdtInterval *int64 `locationName:"tdtInterval" type:"integer"` + // + // TdtInterval is a required field + TdtInterval *int64 `locationName:"tdtInterval" min:"1000" type:"integer" required:"true"` } // String returns the string representation @@ -4326,6 +5421,22 @@ func (s DvbTdtSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *DvbTdtSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DvbTdtSettings"} + if s.TdtInterval == nil { + invalidParams.Add(request.NewErrParamRequired("TdtInterval")) + } + if s.TdtInterval != nil && *s.TdtInterval < 1000 { + invalidParams.Add(request.NewErrParamMinValue("TdtInterval", 1000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetTdtInterval sets the TdtInterval field's value. func (s *DvbTdtSettings) SetTdtInterval(v int64) *DvbTdtSettings { s.TdtInterval = &v @@ -4342,7 +5453,7 @@ type Eac3Settings struct { AttenuationControl *string `locationName:"attenuationControl" type:"string" enum:"Eac3AttenuationControl"` // Average bitrate in bits/second. Valid bitrates depend on the coding mode. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + Bitrate *int64 `locationName:"bitrate" min:"64000" type:"integer"` // Specifies the "Bitstream Mode" (bsmod) for the emitted E-AC-3 stream. See // ATSC A/52-2012 (Annex E) for background on these values. @@ -4356,7 +5467,7 @@ type Eac3Settings struct { // Sets the dialnorm for the output. If blank and input audio is Dolby Digital // Plus, dialnorm will be passed through. - Dialnorm *int64 `locationName:"dialnorm" type:"integer"` + Dialnorm *int64 `locationName:"dialnorm" min:"1" type:"integer"` // Enables Dynamic Range Compression that restricts the absolute peak level // for a signal. @@ -4405,7 +5516,7 @@ type Eac3Settings struct { PhaseControl *string `locationName:"phaseControl" type:"string" enum:"Eac3PhaseControl"` // Sample rate in hz. Sample rate is always 48000. - SampleRate *int64 `locationName:"sampleRate" type:"integer"` + SampleRate *int64 `locationName:"sampleRate" min:"48000" type:"integer"` // Stereo downmix preference. Only used for 3/2 coding mode. StereoDownmix *string `locationName:"stereoDownmix" type:"string" enum:"Eac3StereoDownmix"` @@ -4429,6 +5540,25 @@ func (s Eac3Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Eac3Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Eac3Settings"} + if s.Bitrate != nil && *s.Bitrate < 64000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 64000)) + } + if s.Dialnorm != nil && *s.Dialnorm < 1 { + invalidParams.Add(request.NewErrParamMinValue("Dialnorm", 1)) + } + if s.SampleRate != nil && *s.SampleRate < 48000 { + invalidParams.Add(request.NewErrParamMinValue("SampleRate", 48000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAttenuationControl sets the AttenuationControl field's value. func (s *Eac3Settings) SetAttenuationControl(v string) *Eac3Settings { s.AttenuationControl = &v @@ -4566,11 +5696,11 @@ type EmbeddedSourceSettings struct { // Specifies the 608/708 channel number within the video track from which to // extract captions. Unused for passthrough. - Source608ChannelNumber *int64 `locationName:"source608ChannelNumber" type:"integer"` + Source608ChannelNumber *int64 `locationName:"source608ChannelNumber" min:"1" type:"integer"` // Specifies the video track index used for extracting captions. The system // only supports one input video track, so this should always be set to '1'. - Source608TrackNumber *int64 `locationName:"source608TrackNumber" type:"integer"` + Source608TrackNumber *int64 `locationName:"source608TrackNumber" min:"1" type:"integer"` } // String returns the string representation @@ -4583,6 +5713,22 @@ func (s EmbeddedSourceSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *EmbeddedSourceSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "EmbeddedSourceSettings"} + if s.Source608ChannelNumber != nil && *s.Source608ChannelNumber < 1 { + invalidParams.Add(request.NewErrParamMinValue("Source608ChannelNumber", 1)) + } + if s.Source608TrackNumber != nil && *s.Source608TrackNumber < 1 { + invalidParams.Add(request.NewErrParamMinValue("Source608TrackNumber", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetConvert608To708 sets the Convert608To708 field's value. func (s *EmbeddedSourceSettings) SetConvert608To708(v string) *EmbeddedSourceSettings { s.Convert608To708 = &v @@ -4691,7 +5837,9 @@ type FileSourceSettings struct { // External caption file used for loading captions. Accepted file extensions // are 'scc', 'ttml', 'dfxp', 'stl', 'srt', and 'smi'. - SourceFile *string `locationName:"sourceFile" type:"string"` + // + // SourceFile is a required field + SourceFile *string `locationName:"sourceFile" min:"14" type:"string" required:"true"` // Specifies a time delta in seconds to offset the captions from the source // file. @@ -4708,6 +5856,25 @@ func (s FileSourceSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *FileSourceSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "FileSourceSettings"} + if s.SourceFile == nil { + invalidParams.Add(request.NewErrParamRequired("SourceFile")) + } + if s.SourceFile != nil && len(*s.SourceFile) < 14 { + invalidParams.Add(request.NewErrParamMinLen("SourceFile", 14)) + } + if s.TimeDelta != nil && *s.TimeDelta < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("TimeDelta", -2.147483648e+09)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetConvert608To708 sets the Convert608To708 field's value. func (s *FileSourceSettings) SetConvert608To708(v string) *FileSourceSettings { s.Convert608To708 = &v @@ -4737,7 +5904,7 @@ type FrameCaptureSettings struct { // 1/3 frame per second) will capture the first frame, then 1 frame every 3s. // Files will be named as filename.n.jpg where n is the 0-based sequence number // of each Capture. - FramerateDenominator *int64 `locationName:"framerateDenominator" type:"integer"` + FramerateDenominator *int64 `locationName:"framerateDenominator" min:"1" type:"integer"` // Frame capture will encode the first frame of the output stream, then one // frame every framerateDenominator/framerateNumerator seconds. For example, @@ -4745,13 +5912,13 @@ type FrameCaptureSettings struct { // 1/3 frame per second) will capture the first frame, then 1 frame every 3s. // Files will be named as filename.NNNNNNN.jpg where N is the 0-based frame // sequence number zero padded to 7 decimal places. - FramerateNumerator *int64 `locationName:"framerateNumerator" type:"integer"` + FramerateNumerator *int64 `locationName:"framerateNumerator" min:"1" type:"integer"` // Maximum number of captures (encoded jpg output files). - MaxCaptures *int64 `locationName:"maxCaptures" type:"integer"` + MaxCaptures *int64 `locationName:"maxCaptures" min:"1" type:"integer"` // JPEG Quality - a higher value equals higher quality. - Quality *int64 `locationName:"quality" type:"integer"` + Quality *int64 `locationName:"quality" min:"1" type:"integer"` } // String returns the string representation @@ -4764,6 +5931,28 @@ func (s FrameCaptureSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *FrameCaptureSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "FrameCaptureSettings"} + if s.FramerateDenominator != nil && *s.FramerateDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateDenominator", 1)) + } + if s.FramerateNumerator != nil && *s.FramerateNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateNumerator", 1)) + } + if s.MaxCaptures != nil && *s.MaxCaptures < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxCaptures", 1)) + } + if s.Quality != nil && *s.Quality < 1 { + invalidParams.Add(request.NewErrParamMinValue("Quality", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetFramerateDenominator sets the FramerateDenominator field's value. func (s *FrameCaptureSettings) SetFramerateDenominator(v int64) *FrameCaptureSettings { s.FramerateDenominator = &v @@ -5055,11 +6244,9 @@ type H264Settings struct { // quality. AdaptiveQuantization *string `locationName:"adaptiveQuantization" type:"string" enum:"H264AdaptiveQuantization"` - // Average bitrate in bits/second. Required for VBR, CBR, and ABR. Five megabits - // can be entered as 5000000 or 5m. Five hundred kilobits can be entered as - // 500000 or 0.5m. For MS Smooth outputs, bitrates must be unique when rounded - // down to the nearest multiple of 1000. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + // Average bitrate in bits/second. Required for VBR and CBR. For MS Smooth outputs, + // bitrates must be unique when rounded down to the nearest multiple of 1000. + Bitrate *int64 `locationName:"bitrate" min:"1000" type:"integer"` // H.264 Level. CodecLevel *string `locationName:"codecLevel" type:"string" enum:"H264CodecLevel"` @@ -5077,9 +6264,17 @@ type H264Settings struct { // Adjust quantization within each frame to reduce flicker or 'pop' on I-frames. FlickerAdaptiveQuantization *string `locationName:"flickerAdaptiveQuantization" type:"string" enum:"H264FlickerAdaptiveQuantization"` - // Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want - // the service to use the framerate from the input. Using the console, do this - // by choosing INITIALIZE_FROM_SOURCE for Framerate. + // If you are using the console, use the Framerate setting to specify the framerate + // for this output. If you want to keep the same framerate as the input video, + // choose Follow source. If you want to do framerate conversion, choose a framerate + // from the dropdown list or choose Custom. The framerates shown in the dropdown + // list are decimal approximations of fractions. If you choose Custom, specify + // your framerate as a fraction. If you are creating your transcoding job specification + // as a JSON file without the console, use FramerateControl to specify which + // value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE + // if you want the service to use the framerate from the input. Choose SPECIFIED + // if you want the service to use the framerate you specify in the settings + // FramerateNumerator and FramerateDenominator. FramerateControl *string `locationName:"framerateControl" type:"string" enum:"H264FramerateControl"` // When set to INTERPOLATE, produces smoother motion during framerate conversion. @@ -5091,11 +6286,11 @@ type H264Settings struct { // example, use 1001 for the value of FramerateDenominator. When you use the // console for transcode jobs that use framerate conversion, provide the value // as a decimal number for Framerate. In this example, specify 23.976. - FramerateDenominator *int64 `locationName:"framerateDenominator" type:"integer"` + FramerateDenominator *int64 `locationName:"framerateDenominator" min:"1" type:"integer"` // Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 // fps. - FramerateNumerator *int64 `locationName:"framerateNumerator" type:"integer"` + FramerateNumerator *int64 `locationName:"framerateNumerator" min:"1" type:"integer"` // If enable, use reference B frames for GOP structures that have B frames > // 1. @@ -5117,27 +6312,26 @@ type H264Settings struct { // Percentage of the buffer that should initially be filled (HRD buffer model). HrdBufferInitialFillPercentage *int64 `locationName:"hrdBufferInitialFillPercentage" type:"integer"` - // Size of buffer (HRD buffer model). Five megabits can be entered as 5000000 - // or 5m. Five hundred kilobits can be entered as 500000 or 0.5m. + // Size of buffer (HRD buffer model) in bits. For example, enter five megabits + // as 5000000. HrdBufferSize *int64 `locationName:"hrdBufferSize" type:"integer"` // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top - // or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default + // or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, - // behavior depends on the input scan type. - If the source is interlaced, the - // output will be interlaced with the same polarity as the source (it will follow - // the source). The output could therefore be a mix of "top field first" and - // "bottom field first". - If the source is progressive, the output will be - // interlaced with "top field first" or "bottom field first" polarity, depending + // behavior depends on the input scan type, as follows. - If the source is interlaced, + // the output will be interlaced with the same polarity as the source (it will + // follow the source). The output could therefore be a mix of "top field first" + // and "bottom field first". - If the source is progressive, the output will + // be interlaced with "top field first" or "bottom field first" polarity, depending // on which of the Follow options you chose. InterlaceMode *string `locationName:"interlaceMode" type:"string" enum:"H264InterlaceMode"` - // Maximum bitrate in bits/second (for VBR mode only). Five megabits can be - // entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 - // or 0.5m. - MaxBitrate *int64 `locationName:"maxBitrate" type:"integer"` + // Maximum bitrate in bits/second. For example, enter five megabits per second + // as 5000000. + MaxBitrate *int64 `locationName:"maxBitrate" min:"1000" type:"integer"` // Enforces separation between repeated (cadence) I-frames and I-frames inserted // by Scene Change Detection. If a scene change I-frame is within I-interval @@ -5153,7 +6347,7 @@ type H264Settings struct { // Number of reference frames to use. The encoder may use more than requested // if using B-frames and/or interlaced encoding. - NumberReferenceFrames *int64 `locationName:"numberReferenceFrames" type:"integer"` + NumberReferenceFrames *int64 `locationName:"numberReferenceFrames" min:"1" type:"integer"` // Using the API, enable ParFollowSource if you want the service to use the // pixel aspect ratio from the input. Using the console, do this by choosing @@ -5161,18 +6355,18 @@ type H264Settings struct { ParControl *string `locationName:"parControl" type:"string" enum:"H264ParControl"` // Pixel Aspect Ratio denominator. - ParDenominator *int64 `locationName:"parDenominator" type:"integer"` + ParDenominator *int64 `locationName:"parDenominator" min:"1" type:"integer"` // Pixel Aspect Ratio numerator. - ParNumerator *int64 `locationName:"parNumerator" type:"integer"` + ParNumerator *int64 `locationName:"parNumerator" min:"1" type:"integer"` // Use Quality tuning level (H264QualityTuningLevel) to specifiy whether to // use fast single-pass, high-quality singlepass, or high-quality multipass // video encoding. QualityTuningLevel *string `locationName:"qualityTuningLevel" type:"string" enum:"H264QualityTuningLevel"` - // Rate control mode. CQ uses constant quantizer (qp), ABR (average bitrate) - // does not write HRD parameters. + // Use this setting to specify whether this output has a variable bitrate (VBR) + // or constant bitrate (CBR). RateControlMode *string `locationName:"rateControlMode" type:"string" enum:"H264RateControlMode"` // Places a PPS header on each encoded picture, even if repeated. @@ -5184,7 +6378,7 @@ type H264Settings struct { // Number of slices per picture. Must be less than or equal to the number of // macroblock rows for progressive pictures, and less than or equal to half // the number of macroblock rows for interlaced pictures. - Slices *int64 `locationName:"slices" type:"integer"` + Slices *int64 `locationName:"slices" min:"1" type:"integer"` // Enables Slow PAL rate conversion. 23.976fps and 24fps input is relabeled // as 25fps, and audio is sped up correspondingly. @@ -5228,6 +6422,40 @@ func (s H264Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *H264Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "H264Settings"} + if s.Bitrate != nil && *s.Bitrate < 1000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 1000)) + } + if s.FramerateDenominator != nil && *s.FramerateDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateDenominator", 1)) + } + if s.FramerateNumerator != nil && *s.FramerateNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateNumerator", 1)) + } + if s.MaxBitrate != nil && *s.MaxBitrate < 1000 { + invalidParams.Add(request.NewErrParamMinValue("MaxBitrate", 1000)) + } + if s.NumberReferenceFrames != nil && *s.NumberReferenceFrames < 1 { + invalidParams.Add(request.NewErrParamMinValue("NumberReferenceFrames", 1)) + } + if s.ParDenominator != nil && *s.ParDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParDenominator", 1)) + } + if s.ParNumerator != nil && *s.ParNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParNumerator", 1)) + } + if s.Slices != nil && *s.Slices < 1 { + invalidParams.Add(request.NewErrParamMinValue("Slices", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAdaptiveQuantization sets the AdaptiveQuantization field's value. func (s *H264Settings) SetAdaptiveQuantization(v string) *H264Settings { s.AdaptiveQuantization = &v @@ -5462,11 +6690,9 @@ type H265Settings struct { // Log Gamma (HLG) Electro-Optical Transfer Function (EOTF). AlternateTransferFunctionSei *string `locationName:"alternateTransferFunctionSei" type:"string" enum:"H265AlternateTransferFunctionSei"` - // Average bitrate in bits/second. Required for VBR, CBR, and ABR. Five megabits - // can be entered as 5000000 or 5m. Five hundred kilobits can be entered as - // 500000 or 0.5m. For MS Smooth outputs, bitrates must be unique when rounded - // down to the nearest multiple of 1000. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + // Average bitrate in bits/second. Required for VBR and CBR. For MS Smooth outputs, + // bitrates must be unique when rounded down to the nearest multiple of 1000. + Bitrate *int64 `locationName:"bitrate" min:"1000" type:"integer"` // H.265 Level. CodecLevel *string `locationName:"codecLevel" type:"string" enum:"H265CodecLevel"` @@ -5479,20 +6705,28 @@ type H265Settings struct { // Adjust quantization within each frame to reduce flicker or 'pop' on I-frames. FlickerAdaptiveQuantization *string `locationName:"flickerAdaptiveQuantization" type:"string" enum:"H265FlickerAdaptiveQuantization"` - // Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want - // the service to use the framerate from the input. Using the console, do this - // by choosing INITIALIZE_FROM_SOURCE for Framerate. + // If you are using the console, use the Framerate setting to specify the framerate + // for this output. If you want to keep the same framerate as the input video, + // choose Follow source. If you want to do framerate conversion, choose a framerate + // from the dropdown list or choose Custom. The framerates shown in the dropdown + // list are decimal approximations of fractions. If you choose Custom, specify + // your framerate as a fraction. If you are creating your transcoding job sepecification + // as a JSON file without the console, use FramerateControl to specify which + // value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE + // if you want the service to use the framerate from the input. Choose SPECIFIED + // if you want the service to use the framerate you specify in the settings + // FramerateNumerator and FramerateDenominator. FramerateControl *string `locationName:"framerateControl" type:"string" enum:"H265FramerateControl"` // When set to INTERPOLATE, produces smoother motion during framerate conversion. FramerateConversionAlgorithm *string `locationName:"framerateConversionAlgorithm" type:"string" enum:"H265FramerateConversionAlgorithm"` // Framerate denominator. - FramerateDenominator *int64 `locationName:"framerateDenominator" type:"integer"` + FramerateDenominator *int64 `locationName:"framerateDenominator" min:"1" type:"integer"` // Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 // fps. - FramerateNumerator *int64 `locationName:"framerateNumerator" type:"integer"` + FramerateNumerator *int64 `locationName:"framerateNumerator" min:"1" type:"integer"` // If enable, use reference B frames for GOP structures that have B frames > // 1. @@ -5514,14 +6748,14 @@ type H265Settings struct { // Percentage of the buffer that should initially be filled (HRD buffer model). HrdBufferInitialFillPercentage *int64 `locationName:"hrdBufferInitialFillPercentage" type:"integer"` - // Size of buffer (HRD buffer model). Five megabits can be entered as 5000000 - // or 5m. Five hundred kilobits can be entered as 500000 or 0.5m. + // Size of buffer (HRD buffer model) in bits. For example, enter five megabits + // as 5000000. HrdBufferSize *int64 `locationName:"hrdBufferSize" type:"integer"` // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top - // or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default + // or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, // behavior depends on the input scan type. - If the source is interlaced, the // output will be interlaced with the same polarity as the source (it will follow @@ -5531,10 +6765,8 @@ type H265Settings struct { // on which of the Follow options you chose. InterlaceMode *string `locationName:"interlaceMode" type:"string" enum:"H265InterlaceMode"` - // Maximum bitrate in bits/second (for VBR mode only). Five megabits can be - // entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 - // or 0.5m. - MaxBitrate *int64 `locationName:"maxBitrate" type:"integer"` + // Maximum bitrate in bits/second. + MaxBitrate *int64 `locationName:"maxBitrate" min:"1000" type:"integer"` // Enforces separation between repeated (cadence) I-frames and I-frames inserted // by Scene Change Detection. If a scene change I-frame is within I-interval @@ -5550,7 +6782,7 @@ type H265Settings struct { // Number of reference frames to use. The encoder may use more than requested // if using B-frames and/or interlaced encoding. - NumberReferenceFrames *int64 `locationName:"numberReferenceFrames" type:"integer"` + NumberReferenceFrames *int64 `locationName:"numberReferenceFrames" min:"1" type:"integer"` // Using the API, enable ParFollowSource if you want the service to use the // pixel aspect ratio from the input. Using the console, do this by choosing @@ -5558,18 +6790,18 @@ type H265Settings struct { ParControl *string `locationName:"parControl" type:"string" enum:"H265ParControl"` // Pixel Aspect Ratio denominator. - ParDenominator *int64 `locationName:"parDenominator" type:"integer"` + ParDenominator *int64 `locationName:"parDenominator" min:"1" type:"integer"` // Pixel Aspect Ratio numerator. - ParNumerator *int64 `locationName:"parNumerator" type:"integer"` + ParNumerator *int64 `locationName:"parNumerator" min:"1" type:"integer"` // Use Quality tuning level (H265QualityTuningLevel) to specifiy whether to // use fast single-pass, high-quality singlepass, or high-quality multipass // video encoding. QualityTuningLevel *string `locationName:"qualityTuningLevel" type:"string" enum:"H265QualityTuningLevel"` - // Rate control mode. CQ uses constant quantizer (qp), ABR (average bitrate) - // does not write HRD parameters. + // Use this setting to specify whether this output has a variable bitrate (VBR) + // or constant bitrate (CBR). RateControlMode *string `locationName:"rateControlMode" type:"string" enum:"H265RateControlMode"` // Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically @@ -5582,7 +6814,7 @@ type H265Settings struct { // Number of slices per picture. Must be less than or equal to the number of // macroblock rows for progressive pictures, and less than or equal to half // the number of macroblock rows for interlaced pictures. - Slices *int64 `locationName:"slices" type:"integer"` + Slices *int64 `locationName:"slices" min:"1" type:"integer"` // Enables Slow PAL rate conversion. 23.976fps and 24fps input is relabeled // as 25fps, and audio is sped up correspondingly. @@ -5621,6 +6853,12 @@ type H265Settings struct { // Inserts timecode for each frame as 4 bytes of an unregistered SEI message. UnregisteredSeiTimecode *string `locationName:"unregisteredSeiTimecode" type:"string" enum:"H265UnregisteredSeiTimecode"` + + // If HVC1, output that is H.265 will be marked as HVC1 and adhere to the ISO-IECJTC1-SC29_N13798_Text_ISOIEC_FDIS_14496-15_3rd_E + // spec which states that parameter set NAL units will be stored in the sample + // headers but not in the samples directly. If HEV1, then H.265 will be marked + // as HEV1 and parameter set NAL units will be written into the samples. + WriteMp4PackagingType *string `locationName:"writeMp4PackagingType" type:"string" enum:"H265WriteMp4PackagingType"` } // String returns the string representation @@ -5633,6 +6871,40 @@ func (s H265Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *H265Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "H265Settings"} + if s.Bitrate != nil && *s.Bitrate < 1000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 1000)) + } + if s.FramerateDenominator != nil && *s.FramerateDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateDenominator", 1)) + } + if s.FramerateNumerator != nil && *s.FramerateNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateNumerator", 1)) + } + if s.MaxBitrate != nil && *s.MaxBitrate < 1000 { + invalidParams.Add(request.NewErrParamMinValue("MaxBitrate", 1000)) + } + if s.NumberReferenceFrames != nil && *s.NumberReferenceFrames < 1 { + invalidParams.Add(request.NewErrParamMinValue("NumberReferenceFrames", 1)) + } + if s.ParDenominator != nil && *s.ParDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParDenominator", 1)) + } + if s.ParNumerator != nil && *s.ParNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParNumerator", 1)) + } + if s.Slices != nil && *s.Slices < 1 { + invalidParams.Add(request.NewErrParamMinValue("Slices", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAdaptiveQuantization sets the AdaptiveQuantization field's value. func (s *H265Settings) SetAdaptiveQuantization(v string) *H265Settings { s.AdaptiveQuantization = &v @@ -5849,40 +7121,53 @@ func (s *H265Settings) SetUnregisteredSeiTimecode(v string) *H265Settings { return s } -// Use the HDR master display (Hdr10Metadata) settings to provide values for -// HDR color. These values vary depending on the input video and must be provided -// by a color grader. Range is 0 to 50,000, each increment represents 0.00002 -// in CIE1931 color coordinate. +// SetWriteMp4PackagingType sets the WriteMp4PackagingType field's value. +func (s *H265Settings) SetWriteMp4PackagingType(v string) *H265Settings { + s.WriteMp4PackagingType = &v + return s +} + +// Use the HDR master display (Hdr10Metadata) settings to correct HDR metadata +// or to provide missing metadata. These values vary depending on the input +// video and must be provided by a color grader. Range is 0 to 50,000, each +// increment represents 0.00002 in CIE1931 color coordinate. Note that these +// settings are not color correction. Note that if you are creating HDR outputs +// inside of an HLS CMAF package, to comply with the Apple specification, you +// must use the HVC1 for H.265 setting. type Hdr10Metadata struct { _ struct{} `type:"structure"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. BluePrimaryX *int64 `locationName:"bluePrimaryX" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. BluePrimaryY *int64 `locationName:"bluePrimaryY" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. GreenPrimaryX *int64 `locationName:"greenPrimaryX" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. GreenPrimaryY *int64 `locationName:"greenPrimaryY" type:"integer"` // Maximum light level among all samples in the coded video sequence, in units // of candelas per square meter. - MaxContentLightLevel *int64 `locationName:"maxContentLightLevel" type:"integer"` + // + // MaxContentLightLevel is a required field + MaxContentLightLevel *int64 `locationName:"maxContentLightLevel" type:"integer" required:"true"` // Maximum average light level of any frame in the coded video sequence, in // units of candelas per square meter. - MaxFrameAverageLightLevel *int64 `locationName:"maxFrameAverageLightLevel" type:"integer"` + // + // MaxFrameAverageLightLevel is a required field + MaxFrameAverageLightLevel *int64 `locationName:"maxFrameAverageLightLevel" type:"integer" required:"true"` // Nominal maximum mastering display luminance in units of of 0.0001 candelas // per square meter. @@ -5892,24 +7177,24 @@ type Hdr10Metadata struct { // per square meter MinLuminance *int64 `locationName:"minLuminance" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. RedPrimaryX *int64 `locationName:"redPrimaryX" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. RedPrimaryY *int64 `locationName:"redPrimaryY" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. WhitePointX *int64 `locationName:"whitePointX" type:"integer"` - // HDR Master Display Information comes from the color grader and the color - // grading tools. Range is 0 to 50,000, each increment represents 0.00002 in - // CIE1931 color coordinate. + // HDR Master Display Information must be provided by a color grader, using + // color grading tools. Range is 0 to 50,000, each increment represents 0.00002 + // in CIE1931 color coordinate. Note that this setting is not for color correction. WhitePointY *int64 `locationName:"whitePointY" type:"integer"` } @@ -5923,6 +7208,22 @@ func (s Hdr10Metadata) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Hdr10Metadata) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Hdr10Metadata"} + if s.MaxContentLightLevel == nil { + invalidParams.Add(request.NewErrParamRequired("MaxContentLightLevel")) + } + if s.MaxFrameAverageLightLevel == nil { + invalidParams.Add(request.NewErrParamRequired("MaxFrameAverageLightLevel")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBluePrimaryX sets the BluePrimaryX field's value. func (s *Hdr10Metadata) SetBluePrimaryX(v int64) *Hdr10Metadata { s.BluePrimaryX = &v @@ -6002,8 +7303,11 @@ type HlsCaptionLanguageMapping struct { // Caption channel. CaptionChannel *int64 `locationName:"captionChannel" type:"integer"` - // Code to specify the language, following the specification "ISO 639-2 three-digit - // code":http://www.loc.gov/standards/iso639-2/ + // Specify the language for this caption channel, using the ISO 639-2 or ISO + // 639-3 three-letter language code + CustomLanguageCode *string `locationName:"customLanguageCode" min:"3" type:"string"` + + // Specify the language, using the ISO 639-2 three-letter code listed at https://www.loc.gov/standards/iso639-2/php/code_list.php. LanguageCode *string `locationName:"languageCode" type:"string" enum:"LanguageCode"` // Caption language description. @@ -6020,12 +7324,34 @@ func (s HlsCaptionLanguageMapping) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *HlsCaptionLanguageMapping) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HlsCaptionLanguageMapping"} + if s.CaptionChannel != nil && *s.CaptionChannel < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("CaptionChannel", -2.147483648e+09)) + } + if s.CustomLanguageCode != nil && len(*s.CustomLanguageCode) < 3 { + invalidParams.Add(request.NewErrParamMinLen("CustomLanguageCode", 3)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCaptionChannel sets the CaptionChannel field's value. func (s *HlsCaptionLanguageMapping) SetCaptionChannel(v int64) *HlsCaptionLanguageMapping { s.CaptionChannel = &v return s } +// SetCustomLanguageCode sets the CustomLanguageCode field's value. +func (s *HlsCaptionLanguageMapping) SetCustomLanguageCode(v string) *HlsCaptionLanguageMapping { + s.CustomLanguageCode = &v + return s +} + // SetLanguageCode sets the LanguageCode field's value. func (s *HlsCaptionLanguageMapping) SetLanguageCode(v string) *HlsCaptionLanguageMapping { s.LanguageCode = &v @@ -6045,7 +7371,7 @@ type HlsEncryptionSettings struct { // This is a 128-bit, 16-byte hex value represented by a 32-character text string. // If this parameter is not set then the Initialization Vector will follow the // segment number by default. - ConstantInitializationVector *string `locationName:"constantInitializationVector" type:"string"` + ConstantInitializationVector *string `locationName:"constantInitializationVector" min:"32" type:"string"` // Encrypts the segments with the given encryption scheme. Leave blank to disable. // Selecting 'Disabled' in the web interface also disables encryption. @@ -6063,7 +7389,9 @@ type HlsEncryptionSettings struct { StaticKeyProvider *StaticKeyProvider `locationName:"staticKeyProvider" type:"structure"` // Indicates which type of key provider is used for encryption. - Type *string `locationName:"type" type:"string" enum:"HlsKeyProviderType"` + // + // Type is a required field + Type *string `locationName:"type" type:"string" required:"true" enum:"HlsKeyProviderType"` } // String returns the string representation @@ -6076,6 +7404,32 @@ func (s HlsEncryptionSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *HlsEncryptionSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HlsEncryptionSettings"} + if s.ConstantInitializationVector != nil && len(*s.ConstantInitializationVector) < 32 { + invalidParams.Add(request.NewErrParamMinLen("ConstantInitializationVector", 32)) + } + if s.Type == nil { + invalidParams.Add(request.NewErrParamRequired("Type")) + } + if s.SpekeKeyProvider != nil { + if err := s.SpekeKeyProvider.Validate(); err != nil { + invalidParams.AddNested("SpekeKeyProvider", err.(request.ErrInvalidParams)) + } + } + if s.StaticKeyProvider != nil { + if err := s.StaticKeyProvider.Validate(); err != nil { + invalidParams.AddNested("StaticKeyProvider", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetConstantInitializationVector sets the ConstantInitializationVector field's value. func (s *HlsEncryptionSettings) SetConstantInitializationVector(v string) *HlsEncryptionSettings { s.ConstantInitializationVector = &v @@ -6171,7 +7525,9 @@ type HlsGroupSettings struct { // When set, Minimum Segment Size is enforced by looking ahead and back within // the specified range for a nearby avail and extending the segment size if // needed. - MinSegmentLength *int64 `locationName:"minSegmentLength" type:"integer"` + // + // MinSegmentLength is a required field + MinSegmentLength *int64 `locationName:"minSegmentLength" type:"integer" required:"true"` // Indicates whether the .m3u8 manifest file should be generated for this HLS // output group. @@ -6193,11 +7549,13 @@ type HlsGroupSettings struct { // Length of MPEG-2 Transport Stream segments to create (in seconds). Note that // segments will end on the next keyframe after this number of seconds, so actual // segment length may be longer. - SegmentLength *int64 `locationName:"segmentLength" type:"integer"` + // + // SegmentLength is a required field + SegmentLength *int64 `locationName:"segmentLength" min:"1" type:"integer" required:"true"` // Number of segments to write to a subdirectory before starting a new one. // directoryStructure must be SINGLE_DIRECTORY for this setting to have an effect. - SegmentsPerSubdirectory *int64 `locationName:"segmentsPerSubdirectory" type:"integer"` + SegmentsPerSubdirectory *int64 `locationName:"segmentsPerSubdirectory" min:"1" type:"integer"` // Include or exclude RESOLUTION attribute for video in EXT-X-STREAM-INF tag // of variant manifest. @@ -6223,6 +7581,49 @@ func (s HlsGroupSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *HlsGroupSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HlsGroupSettings"} + if s.MinSegmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("MinSegmentLength")) + } + if s.SegmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("SegmentLength")) + } + if s.SegmentLength != nil && *s.SegmentLength < 1 { + invalidParams.Add(request.NewErrParamMinValue("SegmentLength", 1)) + } + if s.SegmentsPerSubdirectory != nil && *s.SegmentsPerSubdirectory < 1 { + invalidParams.Add(request.NewErrParamMinValue("SegmentsPerSubdirectory", 1)) + } + if s.TimedMetadataId3Period != nil && *s.TimedMetadataId3Period < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("TimedMetadataId3Period", -2.147483648e+09)) + } + if s.TimestampDeltaMilliseconds != nil && *s.TimestampDeltaMilliseconds < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("TimestampDeltaMilliseconds", -2.147483648e+09)) + } + if s.CaptionLanguageMappings != nil { + for i, v := range s.CaptionLanguageMappings { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "CaptionLanguageMappings", i), err.(request.ErrInvalidParams)) + } + } + } + if s.Encryption != nil { + if err := s.Encryption.Validate(); err != nil { + invalidParams.AddNested("Encryption", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAdMarkers sets the AdMarkers field's value. func (s *HlsGroupSettings) SetAdMarkers(v []*string) *HlsGroupSettings { s.AdMarkers = v @@ -6430,15 +7831,19 @@ func (s *HlsSettings) SetSegmentModifier(v string) *HlsSettings { // To insert ID3 tags in your output, specify two values. Use ID3 tag (Id3) // to specify the base 64 encoded string and use Timecode (TimeCode) to specify // the time when the tag should be inserted. To insert multiple ID3 tags in -// your output, create mulitple instances of ID3 insertion (Id3Insertion). +// your output, create multiple instances of ID3 insertion (Id3Insertion). type Id3Insertion struct { _ struct{} `type:"structure"` // Use ID3 tag (Id3) to provide a tag value in base64-encode format. - Id3 *string `locationName:"id3" type:"string"` + // + // Id3 is a required field + Id3 *string `locationName:"id3" type:"string" required:"true"` // Provide a Timecode (TimeCode) in HH:MM:SS:FF or HH:MM:SS;FF format. - Timecode *string `locationName:"timecode" type:"string"` + // + // Timecode is a required field + Timecode *string `locationName:"timecode" type:"string" required:"true"` } // String returns the string representation @@ -6451,6 +7856,22 @@ func (s Id3Insertion) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Id3Insertion) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Id3Insertion"} + if s.Id3 == nil { + invalidParams.Add(request.NewErrParamRequired("Id3")) + } + if s.Timecode == nil { + invalidParams.Add(request.NewErrParamRequired("Timecode")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetId3 sets the Id3 field's value. func (s *Id3Insertion) SetId3(v string) *Id3Insertion { s.Id3 = &v @@ -6471,7 +7892,9 @@ type ImageInserter struct { // Image to insert. Must be 32 bit windows BMP, PNG, or TGA file. Must not be // larger than the output frames. - InsertableImages []*InsertableImage `locationName:"insertableImages" type:"list"` + // + // InsertableImages is a required field + InsertableImages []*InsertableImage `locationName:"insertableImages" type:"list" required:"true"` } // String returns the string representation @@ -6484,6 +7907,29 @@ func (s ImageInserter) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *ImageInserter) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ImageInserter"} + if s.InsertableImages == nil { + invalidParams.Add(request.NewErrParamRequired("InsertableImages")) + } + if s.InsertableImages != nil { + for i, v := range s.InsertableImages { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "InsertableImages", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetInsertableImages sets the InsertableImages field's value. func (s *ImageInserter) SetInsertableImages(v []*InsertableImage) *ImageInserter { s.InsertableImages = v @@ -6522,7 +7968,9 @@ type Input struct { // Use Input (fileInput) to define the source file used in the transcode job. // There can be multiple inputs in a job. These inputs are concantenated, in // the order they are specified in the job, to create the output. - FileInput *string `locationName:"fileInput" type:"string"` + // + // FileInput is a required field + FileInput *string `locationName:"fileInput" type:"string" required:"true"` // Use Filter enable (InputFilterEnable) to specify how the transcoding service // applies the denoise and deblock filters. You must also enable the filters @@ -6549,20 +7997,20 @@ type Input struct { // transport stream. Note that Quad 4K is not currently supported. Default is // the first program within the transport stream. If the program you specify // doesn't exist, the transcoding service will use this default. - ProgramNumber *int64 `locationName:"programNumber" type:"integer"` + ProgramNumber *int64 `locationName:"programNumber" min:"1" type:"integer"` // Set PSI control (InputPsiControl) for transport stream inputs to specify // which data the demux process to scans. * Ignore PSI - Scan all PIDs for audio // and video. * Use PSI - Scan only PSI data. PsiControl *string `locationName:"psiControl" type:"string" enum:"InputPsiControl"` - // Use Timecode source (InputTimecodeSource) to specify how timecode information - // from your input is adjusted and encoded in all outputs for the job. Default - // is embedded. Set to Embedded (EMBEDDED) to use the timecode that is in the - // input video. If no embedded timecode is in the source, will set the timecode - // for the first frame to 00:00:00:00. Set to Start at 0 (ZEROBASED) to set - // the timecode of the initial frame to 00:00:00:00. Set to Specified start - // (SPECIFIEDSTART) to provide the initial timecode yourself the setting (Start). + // Timecode source under input settings (InputTimecodeSource) only affects the + // behavior of features that apply to a single input at a time, such as input + // clipping and synchronizing some captions formats. Use this setting to specify + // whether the service counts frames by timecodes embedded in the video (EMBEDDED) + // or by starting the first frame at zero (ZEROBASED). In both cases, the timecode + // format is HH:MM:SS:FF or HH:MM:SS;FF, where FF is the frame number. Only + // set this to EMBEDDED if your source video has embedded timecodes. TimecodeSource *string `locationName:"timecodeSource" type:"string" enum:"InputTimecodeSource"` // Selector for video. @@ -6579,6 +8027,60 @@ func (s Input) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Input) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Input"} + if s.FileInput == nil { + invalidParams.Add(request.NewErrParamRequired("FileInput")) + } + if s.FilterStrength != nil && *s.FilterStrength < -5 { + invalidParams.Add(request.NewErrParamMinValue("FilterStrength", -5)) + } + if s.ProgramNumber != nil && *s.ProgramNumber < 1 { + invalidParams.Add(request.NewErrParamMinValue("ProgramNumber", 1)) + } + if s.AudioSelectorGroups != nil { + for i, v := range s.AudioSelectorGroups { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AudioSelectorGroups", i), err.(request.ErrInvalidParams)) + } + } + } + if s.AudioSelectors != nil { + for i, v := range s.AudioSelectors { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AudioSelectors", i), err.(request.ErrInvalidParams)) + } + } + } + if s.CaptionSelectors != nil { + for i, v := range s.CaptionSelectors { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "CaptionSelectors", i), err.(request.ErrInvalidParams)) + } + } + } + if s.VideoSelector != nil { + if err := s.VideoSelector.Validate(); err != nil { + invalidParams.AddNested("VideoSelector", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioSelectorGroups sets the AudioSelectorGroups field's value. func (s *Input) SetAudioSelectorGroups(v map[string]*AudioSelectorGroup) *Input { s.AudioSelectorGroups = v @@ -6657,25 +8159,31 @@ func (s *Input) SetVideoSelector(v *VideoSelector) *Input { return s } -// Include one instance of (InputClipping) for each input clip. +// To transcode only portions of your input (clips), include one Input clipping +// (one instance of InputClipping in the JSON job file) for each input clip. +// All input clips you specify will be included in every output of the job. type InputClipping struct { _ struct{} `type:"structure"` // Set End timecode (EndTimecode) to the end of the portion of the input you // are clipping. The frame corresponding to the End timecode value is included // in the clip. Start timecode or End timecode may be left blank, but not both. - // When choosing this value, take into account your setting for Input timecode - // source. For example, if you have embedded timecodes that start at 01:00:00:00 - // and you want your clip to begin five minutes into the video, use 01:00:05:00. + // Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the + // minute, SS is the second, and FF is the frame number. When choosing this + // value, take into account your setting for timecode source under input settings + // (InputTimecodeSource). For example, if you have embedded timecodes that start + // at 01:00:00:00 and you want your clip to end six minutes into the video, + // use 01:06:00:00. EndTimecode *string `locationName:"endTimecode" type:"string"` // Set Start timecode (StartTimecode) to the beginning of the portion of the // input you are clipping. The frame corresponding to the Start timecode value // is included in the clip. Start timecode or End timecode may be left blank, - // but not both. When choosing this value, take into account your setting for - // Input timecode source. For example, if you have embedded timecodes that start - // at 01:00:00:00 and you want your clip to begin five minutes into the video, - // use 01:00:05:00. + // but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the + // hour, MM is the minute, SS is the second, and FF is the frame number. When + // choosing this value, take into account your setting for Input timecode source. + // For example, if you have embedded timecodes that start at 01:00:00:00 and + // you want your clip to begin five minutes into the video, use 01:05:00:00. StartTimecode *string `locationName:"startTimecode" type:"string"` } @@ -6755,20 +8263,20 @@ type InputTemplate struct { // transport stream. Note that Quad 4K is not currently supported. Default is // the first program within the transport stream. If the program you specify // doesn't exist, the transcoding service will use this default. - ProgramNumber *int64 `locationName:"programNumber" type:"integer"` + ProgramNumber *int64 `locationName:"programNumber" min:"1" type:"integer"` // Set PSI control (InputPsiControl) for transport stream inputs to specify // which data the demux process to scans. * Ignore PSI - Scan all PIDs for audio // and video. * Use PSI - Scan only PSI data. PsiControl *string `locationName:"psiControl" type:"string" enum:"InputPsiControl"` - // Use Timecode source (InputTimecodeSource) to specify how timecode information - // from your input is adjusted and encoded in all outputs for the job. Default - // is embedded. Set to Embedded (EMBEDDED) to use the timecode that is in the - // input video. If no embedded timecode is in the source, will set the timecode - // for the first frame to 00:00:00:00. Set to Start at 0 (ZEROBASED) to set - // the timecode of the initial frame to 00:00:00:00. Set to Specified start - // (SPECIFIEDSTART) to provide the initial timecode yourself the setting (Start). + // Timecode source under input settings (InputTimecodeSource) only affects the + // behavior of features that apply to a single input at a time, such as input + // clipping and synchronizing some captions formats. Use this setting to specify + // whether the service counts frames by timecodes embedded in the video (EMBEDDED) + // or by starting the first frame at zero (ZEROBASED). In both cases, the timecode + // format is HH:MM:SS:FF or HH:MM:SS;FF, where FF is the frame number. Only + // set this to EMBEDDED if your source video has embedded timecodes. TimecodeSource *string `locationName:"timecodeSource" type:"string" enum:"InputTimecodeSource"` // Selector for video. @@ -6785,6 +8293,57 @@ func (s InputTemplate) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *InputTemplate) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "InputTemplate"} + if s.FilterStrength != nil && *s.FilterStrength < -5 { + invalidParams.Add(request.NewErrParamMinValue("FilterStrength", -5)) + } + if s.ProgramNumber != nil && *s.ProgramNumber < 1 { + invalidParams.Add(request.NewErrParamMinValue("ProgramNumber", 1)) + } + if s.AudioSelectorGroups != nil { + for i, v := range s.AudioSelectorGroups { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AudioSelectorGroups", i), err.(request.ErrInvalidParams)) + } + } + } + if s.AudioSelectors != nil { + for i, v := range s.AudioSelectors { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AudioSelectors", i), err.(request.ErrInvalidParams)) + } + } + } + if s.CaptionSelectors != nil { + for i, v := range s.CaptionSelectors { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "CaptionSelectors", i), err.(request.ErrInvalidParams)) + } + } + } + if s.VideoSelector != nil { + if err := s.VideoSelector.Validate(); err != nil { + invalidParams.AddNested("VideoSelector", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioSelectorGroups sets the AudioSelectorGroups field's value. func (s *InputTemplate) SetAudioSelectorGroups(v map[string]*AudioSelectorGroup) *InputTemplate { s.AudioSelectorGroups = v @@ -6883,28 +8442,39 @@ type InsertableImage struct { // Use Image location (imageInserterInput) to specify the Amazon S3 location // of the image to be inserted into the output. Use a 32 bit BMP, PNG, or TGA // file that fits inside the video frame. - ImageInserterInput *string `locationName:"imageInserterInput" type:"string"` + // + // ImageInserterInput is a required field + ImageInserterInput *string `locationName:"imageInserterInput" min:"14" type:"string" required:"true"` // Use Left (ImageX) to set the distance, in pixels, between the inserted image // and the left edge of the frame. Required for BMP, PNG and TGA input. - ImageX *int64 `locationName:"imageX" type:"integer"` + // + // ImageX is a required field + ImageX *int64 `locationName:"imageX" type:"integer" required:"true"` // Use Top (ImageY) to set the distance, in pixels, between the inserted image // and the top edge of the video frame. Required for BMP, PNG and TGA input. - ImageY *int64 `locationName:"imageY" type:"integer"` + // + // ImageY is a required field + ImageY *int64 `locationName:"imageY" type:"integer" required:"true"` // Use Layer (Layer) to specify how overlapping inserted images appear. Images // with higher values of layer appear on top of images with lower values of // layer. - Layer *int64 `locationName:"layer" type:"integer"` + // + // Layer is a required field + Layer *int64 `locationName:"layer" type:"integer" required:"true"` // Use Opacity (Opacity) to specify how much of the underlying video shows through // the inserted image. 0 is transparent and 100 is fully opaque. Default is // 50. - Opacity *int64 `locationName:"opacity" type:"integer"` + // + // Opacity is a required field + Opacity *int64 `locationName:"opacity" type:"integer" required:"true"` // Use Start time (StartTime) to specify the video timecode when the image is - // inserted in the output. This must be in timecode format (HH:MM:SS:FF) + // inserted in the output. This must be in timecode (HH:MM:SS:FF or HH:MM:SS;FF) + // format. StartTime *string `locationName:"startTime" type:"string"` // Specify the Width (Width) of the inserted image. Use a value that is less @@ -6923,6 +8493,55 @@ func (s InsertableImage) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *InsertableImage) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "InsertableImage"} + if s.Duration != nil && *s.Duration < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Duration", -2.147483648e+09)) + } + if s.FadeIn != nil && *s.FadeIn < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("FadeIn", -2.147483648e+09)) + } + if s.FadeOut != nil && *s.FadeOut < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("FadeOut", -2.147483648e+09)) + } + if s.Height != nil && *s.Height < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Height", -2.147483648e+09)) + } + if s.ImageInserterInput == nil { + invalidParams.Add(request.NewErrParamRequired("ImageInserterInput")) + } + if s.ImageInserterInput != nil && len(*s.ImageInserterInput) < 14 { + invalidParams.Add(request.NewErrParamMinLen("ImageInserterInput", 14)) + } + if s.ImageX == nil { + invalidParams.Add(request.NewErrParamRequired("ImageX")) + } + if s.ImageX != nil && *s.ImageX < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ImageX", -2.147483648e+09)) + } + if s.ImageY == nil { + invalidParams.Add(request.NewErrParamRequired("ImageY")) + } + if s.ImageY != nil && *s.ImageY < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ImageY", -2.147483648e+09)) + } + if s.Layer == nil { + invalidParams.Add(request.NewErrParamRequired("Layer")) + } + if s.Opacity == nil { + invalidParams.Add(request.NewErrParamRequired("Opacity")) + } + if s.Width != nil && *s.Width < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Width", -2.147483648e+09)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetDuration sets the Duration field's value. func (s *InsertableImage) SetDuration(v int64) *InsertableImage { s.Duration = &v @@ -7024,10 +8643,14 @@ type Job struct { // The IAM role you use for creating this job. For details about permissions, // see the User Guide topic at the User Guide at http://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html - Role *string `locationName:"role" type:"string"` + // + // Role is a required field + Role *string `locationName:"role" type:"string" required:"true"` // JobSettings contains all the transcode settings for a job. - Settings *JobSettings `locationName:"settings" type:"structure"` + // + // Settings is a required field + Settings *JobSettings `locationName:"settings" type:"structure" required:"true"` // A job's status can be SUBMITTED, PROGRESSING, COMPLETE, CANCELED, or ERROR. Status *string `locationName:"status" type:"string" enum:"JobStatus"` @@ -7144,21 +8767,25 @@ type JobSettings struct { // Use Inputs (inputs) to define source file used in the transcode job. There // can be multiple inputs add in a job. These inputs will be concantenated together // to create the output. - Inputs []*Input `locationName:"inputs" type:"list"` + // + // Inputs is a required field + Inputs []*Input `locationName:"inputs" type:"list" required:"true"` // Settings for Nielsen Configuration NielsenConfiguration *NielsenConfiguration `locationName:"nielsenConfiguration" type:"structure"` - // **!!**(OutputGroups) contains one group of settings for each set of outputs - // that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, - // Quicktime, MXF, and no container) are grouped in a single output group as - // well. Required in (OutputGroups) is a group of settings that apply to the - // whole group. This required object depends on the value you set for (Type) - // under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are - // as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, - // HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, - // MsSmoothGroupSettings - OutputGroups []*OutputGroup `locationName:"outputGroups" type:"list"` + // (OutputGroups) contains one group of settings for each set of outputs that + // share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, + // MXF, and no container) are grouped in a single output group as well. Required + // in (OutputGroups) is a group of settings that apply to the whole group. This + // required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). + // Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings + // * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings + // * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, + // CmafGroupSettings + // + // OutputGroups is a required field + OutputGroups []*OutputGroup `locationName:"outputGroups" type:"list" required:"true"` // Contains settings used to acquire and adjust timecode information from inputs. TimecodeConfig *TimecodeConfig `locationName:"timecodeConfig" type:"structure"` @@ -7180,6 +8807,55 @@ func (s JobSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *JobSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "JobSettings"} + if s.AdAvailOffset != nil && *s.AdAvailOffset < -1000 { + invalidParams.Add(request.NewErrParamMinValue("AdAvailOffset", -1000)) + } + if s.Inputs == nil { + invalidParams.Add(request.NewErrParamRequired("Inputs")) + } + if s.OutputGroups == nil { + invalidParams.Add(request.NewErrParamRequired("OutputGroups")) + } + if s.AvailBlanking != nil { + if err := s.AvailBlanking.Validate(); err != nil { + invalidParams.AddNested("AvailBlanking", err.(request.ErrInvalidParams)) + } + } + if s.Inputs != nil { + for i, v := range s.Inputs { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Inputs", i), err.(request.ErrInvalidParams)) + } + } + } + if s.OutputGroups != nil { + for i, v := range s.OutputGroups { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "OutputGroups", i), err.(request.ErrInvalidParams)) + } + } + } + if s.TimedMetadataInsertion != nil { + if err := s.TimedMetadataInsertion.Validate(); err != nil { + invalidParams.AddNested("TimedMetadataInsertion", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAdAvailOffset sets the AdAvailOffset field's value. func (s *JobSettings) SetAdAvailOffset(v int64) *JobSettings { s.AdAvailOffset = &v @@ -7244,7 +8920,9 @@ type JobTemplate struct { // A name you create for each job template. Each name must be unique within // your account. - Name *string `locationName:"name" type:"string"` + // + // Name is a required field + Name *string `locationName:"name" type:"string" required:"true"` // Optional. The queue that jobs created from this template are assigned to. // If you don't specify this, jobs will go to the default queue. @@ -7252,7 +8930,9 @@ type JobTemplate struct { // JobTemplateSettings contains all the transcode settings saved in the template // that will be applied to jobs created from it. - Settings *JobTemplateSettings `locationName:"settings" type:"structure"` + // + // Settings is a required field + Settings *JobTemplateSettings `locationName:"settings" type:"structure" required:"true"` // A job template can be of two types: system or custom. System or built-in // job templates can't be modified or deleted by the user. @@ -7344,16 +9024,18 @@ type JobTemplateSettings struct { // Settings for Nielsen Configuration NielsenConfiguration *NielsenConfiguration `locationName:"nielsenConfiguration" type:"structure"` - // **!!**(OutputGroups) contains one group of settings for each set of outputs - // that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, - // Quicktime, MXF, and no container) are grouped in a single output group as - // well. Required in (OutputGroups) is a group of settings that apply to the - // whole group. This required object depends on the value you set for (Type) - // under (OutputGroups)>(OutputGroupSettings). Type, settings object pairs are - // as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, - // HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, - // MsSmoothGroupSettings - OutputGroups []*OutputGroup `locationName:"outputGroups" type:"list"` + // (OutputGroups) contains one group of settings for each set of outputs that + // share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, + // MXF, and no container) are grouped in a single output group as well. Required + // in (OutputGroups) is a group of settings that apply to the whole group. This + // required object depends on the value you set for (Type) under (OutputGroups)>(OutputGroupSettings). + // Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings + // * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings + // * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, + // CmafGroupSettings + // + // OutputGroups is a required field + OutputGroups []*OutputGroup `locationName:"outputGroups" type:"list" required:"true"` // Contains settings used to acquire and adjust timecode information from inputs. TimecodeConfig *TimecodeConfig `locationName:"timecodeConfig" type:"structure"` @@ -7375,6 +9057,52 @@ func (s JobTemplateSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *JobTemplateSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "JobTemplateSettings"} + if s.AdAvailOffset != nil && *s.AdAvailOffset < -1000 { + invalidParams.Add(request.NewErrParamMinValue("AdAvailOffset", -1000)) + } + if s.OutputGroups == nil { + invalidParams.Add(request.NewErrParamRequired("OutputGroups")) + } + if s.AvailBlanking != nil { + if err := s.AvailBlanking.Validate(); err != nil { + invalidParams.AddNested("AvailBlanking", err.(request.ErrInvalidParams)) + } + } + if s.Inputs != nil { + for i, v := range s.Inputs { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Inputs", i), err.(request.ErrInvalidParams)) + } + } + } + if s.OutputGroups != nil { + for i, v := range s.OutputGroups { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "OutputGroups", i), err.(request.ErrInvalidParams)) + } + } + } + if s.TimedMetadataInsertion != nil { + if err := s.TimedMetadataInsertion.Validate(); err != nil { + invalidParams.AddNested("TimedMetadataInsertion", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAdAvailOffset sets the AdAvailOffset field's value. func (s *JobTemplateSettings) SetAdAvailOffset(v int64) *JobTemplateSettings { s.AdAvailOffset = &v @@ -7856,7 +9584,7 @@ type M2tsSettings struct { DvbTdtSettings *DvbTdtSettings `locationName:"dvbTdtSettings" type:"structure"` // Packet Identifier (PID) for input source DVB Teletext data to this output. - DvbTeletextPid *int64 `locationName:"dvbTeletextPid" type:"integer"` + DvbTeletextPid *int64 `locationName:"dvbTeletextPid" min:"32" type:"integer"` // When set to VIDEO_AND_FIXED_INTERVALS, audio EBP markers will be added to // partitions 3 and 4. The interval between these additional markers will be @@ -7911,7 +9639,7 @@ type M2tsSettings struct { // Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport // stream. When no value is given, the encoder will assign the same value as // the Video PID. - PcrPid *int64 `locationName:"pcrPid" type:"integer"` + PcrPid *int64 `locationName:"pcrPid" min:"32" type:"integer"` // The number of milliseconds between instances of this table in the output // transport stream. @@ -7919,10 +9647,10 @@ type M2tsSettings struct { // Packet Identifier (PID) for the Program Map Table (PMT) in the transport // stream. - PmtPid *int64 `locationName:"pmtPid" type:"integer"` + PmtPid *int64 `locationName:"pmtPid" min:"32" type:"integer"` // Packet Identifier (PID) of the private metadata stream in the transport stream. - PrivateMetadataPid *int64 `locationName:"privateMetadataPid" type:"integer"` + PrivateMetadataPid *int64 `locationName:"privateMetadataPid" min:"32" type:"integer"` // The value of the program number field in the Program Map Table. ProgramNumber *int64 `locationName:"programNumber" type:"integer"` @@ -7933,7 +9661,7 @@ type M2tsSettings struct { RateMode *string `locationName:"rateMode" type:"string" enum:"M2tsRateMode"` // Packet Identifier (PID) of the SCTE-35 stream in the transport stream. - Scte35Pid *int64 `locationName:"scte35Pid" type:"integer"` + Scte35Pid *int64 `locationName:"scte35Pid" min:"32" type:"integer"` // Enables SCTE-35 passthrough (scte35Source) to pass any SCTE-35 signals from // input to output. @@ -7966,13 +9694,13 @@ type M2tsSettings struct { SegmentationTime *float64 `locationName:"segmentationTime" type:"double"` // Packet Identifier (PID) of the timed metadata stream in the transport stream. - TimedMetadataPid *int64 `locationName:"timedMetadataPid" type:"integer"` + TimedMetadataPid *int64 `locationName:"timedMetadataPid" min:"32" type:"integer"` // The value of the transport stream ID field in the Program Map Table. TransportStreamId *int64 `locationName:"transportStreamId" type:"integer"` // Packet Identifier (PID) of the elementary video stream in the transport stream. - VideoPid *int64 `locationName:"videoPid" type:"integer"` + VideoPid *int64 `locationName:"videoPid" min:"32" type:"integer"` } // String returns the string representation @@ -7985,6 +9713,52 @@ func (s M2tsSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *M2tsSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "M2tsSettings"} + if s.DvbTeletextPid != nil && *s.DvbTeletextPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("DvbTeletextPid", 32)) + } + if s.PcrPid != nil && *s.PcrPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("PcrPid", 32)) + } + if s.PmtPid != nil && *s.PmtPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("PmtPid", 32)) + } + if s.PrivateMetadataPid != nil && *s.PrivateMetadataPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("PrivateMetadataPid", 32)) + } + if s.Scte35Pid != nil && *s.Scte35Pid < 32 { + invalidParams.Add(request.NewErrParamMinValue("Scte35Pid", 32)) + } + if s.TimedMetadataPid != nil && *s.TimedMetadataPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("TimedMetadataPid", 32)) + } + if s.VideoPid != nil && *s.VideoPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("VideoPid", 32)) + } + if s.DvbNitSettings != nil { + if err := s.DvbNitSettings.Validate(); err != nil { + invalidParams.AddNested("DvbNitSettings", err.(request.ErrInvalidParams)) + } + } + if s.DvbSdtSettings != nil { + if err := s.DvbSdtSettings.Validate(); err != nil { + invalidParams.AddNested("DvbSdtSettings", err.(request.ErrInvalidParams)) + } + } + if s.DvbTdtSettings != nil { + if err := s.DvbTdtSettings.Validate(); err != nil { + invalidParams.AddNested("DvbTdtSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioBufferModel sets the AudioBufferModel field's value. func (s *M2tsSettings) SetAudioBufferModel(v string) *M2tsSettings { s.AudioBufferModel = &v @@ -8217,7 +9991,7 @@ type M3u8Settings struct { // Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport // stream. When no value is given, the encoder will assign the same value as // the Video PID. - PcrPid *int64 `locationName:"pcrPid" type:"integer"` + PcrPid *int64 `locationName:"pcrPid" min:"32" type:"integer"` // The number of milliseconds between instances of this table in the output // transport stream. @@ -8225,33 +9999,33 @@ type M3u8Settings struct { // Packet Identifier (PID) for the Program Map Table (PMT) in the transport // stream. - PmtPid *int64 `locationName:"pmtPid" type:"integer"` + PmtPid *int64 `locationName:"pmtPid" min:"32" type:"integer"` // Packet Identifier (PID) of the private metadata stream in the transport stream. - PrivateMetadataPid *int64 `locationName:"privateMetadataPid" type:"integer"` + PrivateMetadataPid *int64 `locationName:"privateMetadataPid" min:"32" type:"integer"` // The value of the program number field in the Program Map Table. ProgramNumber *int64 `locationName:"programNumber" type:"integer"` // Packet Identifier (PID) of the SCTE-35 stream in the transport stream. - Scte35Pid *int64 `locationName:"scte35Pid" type:"integer"` + Scte35Pid *int64 `locationName:"scte35Pid" min:"32" type:"integer"` // Enables SCTE-35 passthrough (scte35Source) to pass any SCTE-35 signals from // input to output. Scte35Source *string `locationName:"scte35Source" type:"string" enum:"M3u8Scte35Source"` - // If PASSTHROUGH, inserts ID3 timed metadata from the timed_metadata REST command - // into this output. + // Applies only to HLS outputs. Use this setting to specify whether the service + // inserts the ID3 timed metadata from the input in this output. TimedMetadata *string `locationName:"timedMetadata" type:"string" enum:"TimedMetadata"` // Packet Identifier (PID) of the timed metadata stream in the transport stream. - TimedMetadataPid *int64 `locationName:"timedMetadataPid" type:"integer"` + TimedMetadataPid *int64 `locationName:"timedMetadataPid" min:"32" type:"integer"` // The value of the transport stream ID field in the Program Map Table. TransportStreamId *int64 `locationName:"transportStreamId" type:"integer"` // Packet Identifier (PID) of the elementary video stream in the transport stream. - VideoPid *int64 `locationName:"videoPid" type:"integer"` + VideoPid *int64 `locationName:"videoPid" min:"32" type:"integer"` } // String returns the string representation @@ -8264,6 +10038,34 @@ func (s M3u8Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *M3u8Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "M3u8Settings"} + if s.PcrPid != nil && *s.PcrPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("PcrPid", 32)) + } + if s.PmtPid != nil && *s.PmtPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("PmtPid", 32)) + } + if s.PrivateMetadataPid != nil && *s.PrivateMetadataPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("PrivateMetadataPid", 32)) + } + if s.Scte35Pid != nil && *s.Scte35Pid < 32 { + invalidParams.Add(request.NewErrParamMinValue("Scte35Pid", 32)) + } + if s.TimedMetadataPid != nil && *s.TimedMetadataPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("TimedMetadataPid", 32)) + } + if s.VideoPid != nil && *s.VideoPid < 32 { + invalidParams.Add(request.NewErrParamMinValue("VideoPid", 32)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioFramesPerPes sets the AudioFramesPerPes field's value. func (s *M3u8Settings) SetAudioFramesPerPes(v int64) *M3u8Settings { s.AudioFramesPerPes = &v @@ -8434,15 +10236,15 @@ type Mp2Settings struct { _ struct{} `type:"structure"` // Average bitrate in bits/second. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + Bitrate *int64 `locationName:"bitrate" min:"32000" type:"integer"` // Set Channels to specify the number of channels in this output audio track. // Choosing Mono in the console will give you 1 output channel; choosing Stereo // will give you 2. In the API, valid values are 1 and 2. - Channels *int64 `locationName:"channels" type:"integer"` + Channels *int64 `locationName:"channels" min:"1" type:"integer"` // Sample rate in hz. - SampleRate *int64 `locationName:"sampleRate" type:"integer"` + SampleRate *int64 `locationName:"sampleRate" min:"32000" type:"integer"` } // String returns the string representation @@ -8455,6 +10257,25 @@ func (s Mp2Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Mp2Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Mp2Settings"} + if s.Bitrate != nil && *s.Bitrate < 32000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 32000)) + } + if s.Channels != nil && *s.Channels < 1 { + invalidParams.Add(request.NewErrParamMinValue("Channels", 1)) + } + if s.SampleRate != nil && *s.SampleRate < 32000 { + invalidParams.Add(request.NewErrParamMinValue("SampleRate", 32000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBitrate sets the Bitrate field's value. func (s *Mp2Settings) SetBitrate(v int64) *Mp2Settings { s.Bitrate = &v @@ -8539,11 +10360,9 @@ type Mpeg2Settings struct { // quality. AdaptiveQuantization *string `locationName:"adaptiveQuantization" type:"string" enum:"Mpeg2AdaptiveQuantization"` - // Average bitrate in bits/second. Required for VBR, CBR, and ABR. Five megabits - // can be entered as 5000000 or 5m. Five hundred kilobits can be entered as - // 500000 or 0.5m. For MS Smooth outputs, bitrates must be unique when rounded - // down to the nearest multiple of 1000. - Bitrate *int64 `locationName:"bitrate" type:"integer"` + // Average bitrate in bits/second. Required for VBR and CBR. For MS Smooth outputs, + // bitrates must be unique when rounded down to the nearest multiple of 1000. + Bitrate *int64 `locationName:"bitrate" min:"1000" type:"integer"` // Use Level (Mpeg2CodecLevel) to set the MPEG-2 level for the video output. CodecLevel *string `locationName:"codecLevel" type:"string" enum:"Mpeg2CodecLevel"` @@ -8551,20 +10370,28 @@ type Mpeg2Settings struct { // Use Profile (Mpeg2CodecProfile) to set the MPEG-2 profile for the video output. CodecProfile *string `locationName:"codecProfile" type:"string" enum:"Mpeg2CodecProfile"` - // Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want - // the service to use the framerate from the input. Using the console, do this - // by choosing INITIALIZE_FROM_SOURCE for Framerate. + // If you are using the console, use the Framerate setting to specify the framerate + // for this output. If you want to keep the same framerate as the input video, + // choose Follow source. If you want to do framerate conversion, choose a framerate + // from the dropdown list or choose Custom. The framerates shown in the dropdown + // list are decimal approximations of fractions. If you choose Custom, specify + // your framerate as a fraction. If you are creating your transcoding job sepecification + // as a JSON file without the console, use FramerateControl to specify which + // value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE + // if you want the service to use the framerate from the input. Choose SPECIFIED + // if you want the service to use the framerate you specify in the settings + // FramerateNumerator and FramerateDenominator. FramerateControl *string `locationName:"framerateControl" type:"string" enum:"Mpeg2FramerateControl"` // When set to INTERPOLATE, produces smoother motion during framerate conversion. FramerateConversionAlgorithm *string `locationName:"framerateConversionAlgorithm" type:"string" enum:"Mpeg2FramerateConversionAlgorithm"` // Framerate denominator. - FramerateDenominator *int64 `locationName:"framerateDenominator" type:"integer"` + FramerateDenominator *int64 `locationName:"framerateDenominator" min:"1" type:"integer"` // Framerate numerator - framerate is a fraction, e.g. 24000 / 1001 = 23.976 // fps. - FramerateNumerator *int64 `locationName:"framerateNumerator" type:"integer"` + FramerateNumerator *int64 `locationName:"framerateNumerator" min:"24" type:"integer"` // Frequency of closed GOPs. In streaming applications, it is recommended that // this be set to 1 so a decoder joining mid-stream will receive an IDR frame @@ -8582,14 +10409,14 @@ type Mpeg2Settings struct { // Percentage of the buffer that should initially be filled (HRD buffer model). HrdBufferInitialFillPercentage *int64 `locationName:"hrdBufferInitialFillPercentage" type:"integer"` - // Size of buffer (HRD buffer model). Five megabits can be entered as 5000000 - // or 5m. Five hundred kilobits can be entered as 500000 or 0.5m. + // Size of buffer (HRD buffer model) in bits. For example, enter five megabits + // as 5000000. HrdBufferSize *int64 `locationName:"hrdBufferSize" type:"integer"` // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top - // or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default + // or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, // behavior depends on the input scan type. - If the source is interlaced, the // output will be interlaced with the same polarity as the source (it will follow @@ -8605,10 +10432,9 @@ type Mpeg2Settings struct { // ratio. IntraDcPrecision *string `locationName:"intraDcPrecision" type:"string" enum:"Mpeg2IntraDcPrecision"` - // Maximum bitrate in bits/second (for VBR mode only). Five megabits can be - // entered as 5000000 or 5m. Five hundred kilobits can be entered as 500000 - // or 0.5m. - MaxBitrate *int64 `locationName:"maxBitrate" type:"integer"` + // Maximum bitrate in bits/second. For example, enter five megabits per second + // as 5000000. + MaxBitrate *int64 `locationName:"maxBitrate" min:"1000" type:"integer"` // Enforces separation between repeated (cadence) I-frames and I-frames inserted // by Scene Change Detection. If a scene change I-frame is within I-interval @@ -8628,10 +10454,10 @@ type Mpeg2Settings struct { ParControl *string `locationName:"parControl" type:"string" enum:"Mpeg2ParControl"` // Pixel Aspect Ratio denominator. - ParDenominator *int64 `locationName:"parDenominator" type:"integer"` + ParDenominator *int64 `locationName:"parDenominator" min:"1" type:"integer"` // Pixel Aspect Ratio numerator. - ParNumerator *int64 `locationName:"parNumerator" type:"integer"` + ParNumerator *int64 `locationName:"parNumerator" min:"1" type:"integer"` // Use Quality tuning level (Mpeg2QualityTuningLevel) to specifiy whether to // use single-pass or multipass video encoding. @@ -8680,6 +10506,34 @@ func (s Mpeg2Settings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Mpeg2Settings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Mpeg2Settings"} + if s.Bitrate != nil && *s.Bitrate < 1000 { + invalidParams.Add(request.NewErrParamMinValue("Bitrate", 1000)) + } + if s.FramerateDenominator != nil && *s.FramerateDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateDenominator", 1)) + } + if s.FramerateNumerator != nil && *s.FramerateNumerator < 24 { + invalidParams.Add(request.NewErrParamMinValue("FramerateNumerator", 24)) + } + if s.MaxBitrate != nil && *s.MaxBitrate < 1000 { + invalidParams.Add(request.NewErrParamMinValue("MaxBitrate", 1000)) + } + if s.ParDenominator != nil && *s.ParDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParDenominator", 1)) + } + if s.ParNumerator != nil && *s.ParNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParNumerator", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAdaptiveQuantization sets the AdaptiveQuantization field's value. func (s *Mpeg2Settings) SetAdaptiveQuantization(v string) *Mpeg2Settings { s.AdaptiveQuantization = &v @@ -8866,7 +10720,9 @@ type MsSmoothEncryptionSettings struct { _ struct{} `type:"structure"` // Settings for use with a SPEKE key provider - SpekeKeyProvider *SpekeKeyProvider `locationName:"spekeKeyProvider" type:"structure"` + // + // SpekeKeyProvider is a required field + SpekeKeyProvider *SpekeKeyProvider `locationName:"spekeKeyProvider" type:"structure" required:"true"` } // String returns the string representation @@ -8879,6 +10735,24 @@ func (s MsSmoothEncryptionSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *MsSmoothEncryptionSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "MsSmoothEncryptionSettings"} + if s.SpekeKeyProvider == nil { + invalidParams.Add(request.NewErrParamRequired("SpekeKeyProvider")) + } + if s.SpekeKeyProvider != nil { + if err := s.SpekeKeyProvider.Validate(); err != nil { + invalidParams.AddNested("SpekeKeyProvider", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetSpekeKeyProvider sets the SpekeKeyProvider field's value. func (s *MsSmoothEncryptionSettings) SetSpekeKeyProvider(v *SpekeKeyProvider) *MsSmoothEncryptionSettings { s.SpekeKeyProvider = v @@ -8907,7 +10781,9 @@ type MsSmoothGroupSettings struct { // Use Fragment length (FragmentLength) to specify the mp4 fragment sizes in // seconds. Fragment length must be compatible with GOP size and framerate. - FragmentLength *int64 `locationName:"fragmentLength" type:"integer"` + // + // FragmentLength is a required field + FragmentLength *int64 `locationName:"fragmentLength" min:"1" type:"integer" required:"true"` // Use Manifest encoding (MsSmoothManifestEncoding) to specify the encoding // format for the server and client manifest. Valid options are utf8 and utf16. @@ -8924,6 +10800,27 @@ func (s MsSmoothGroupSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *MsSmoothGroupSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "MsSmoothGroupSettings"} + if s.FragmentLength == nil { + invalidParams.Add(request.NewErrParamRequired("FragmentLength")) + } + if s.FragmentLength != nil && *s.FragmentLength < 1 { + invalidParams.Add(request.NewErrParamMinValue("FragmentLength", 1)) + } + if s.Encryption != nil { + if err := s.Encryption.Validate(); err != nil { + invalidParams.AddNested("Encryption", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioDeduplication sets the AudioDeduplication field's value. func (s *MsSmoothGroupSettings) SetAudioDeduplication(v string) *MsSmoothGroupSettings { s.AudioDeduplication = &v @@ -9000,10 +10897,12 @@ type NoiseReducer struct { // Use Noise reducer filter (NoiseReducerFilter) to select one of the following // spatial image filtering functions. To use this setting, you must also enable // Noise reducer (NoiseReducer). * Bilateral is an edge preserving noise reduction - // filter * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) are convolution - // filters * Conserve is a min/max noise reduction filter * Spatial is frequency-domain + // filter. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) are convolution + // filters. * Conserve is a min/max noise reduction filter. * Spatial is a frequency-domain // filter based on JND principles. - Filter *string `locationName:"filter" type:"string" enum:"NoiseReducerFilter"` + // + // Filter is a required field + Filter *string `locationName:"filter" type:"string" required:"true" enum:"NoiseReducerFilter"` // Settings for a noise reducer filter FilterSettings *NoiseReducerFilterSettings `locationName:"filterSettings" type:"structure"` @@ -9022,6 +10921,24 @@ func (s NoiseReducer) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *NoiseReducer) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "NoiseReducer"} + if s.Filter == nil { + invalidParams.Add(request.NewErrParamRequired("Filter")) + } + if s.SpatialFilterSettings != nil { + if err := s.SpatialFilterSettings.Validate(); err != nil { + invalidParams.AddNested("SpatialFilterSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetFilter sets the Filter field's value. func (s *NoiseReducer) SetFilter(v string) *NoiseReducer { s.Filter = &v @@ -9092,6 +11009,19 @@ func (s NoiseReducerSpatialFilterSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *NoiseReducerSpatialFilterSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "NoiseReducerSpatialFilterSettings"} + if s.Speed != nil && *s.Speed < -2 { + invalidParams.Add(request.NewErrParamMinValue("Speed", -2)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetPostFilterSharpenStrength sets the PostFilterSharpenStrength field's value. func (s *NoiseReducerSpatialFilterSettings) SetPostFilterSharpenStrength(v int64) *NoiseReducerSpatialFilterSettings { s.PostFilterSharpenStrength = &v @@ -9142,7 +11072,7 @@ type Output struct { // identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ // or $Time$ in one output, you must use them in the same way in all outputs // of the output group. - NameModifier *string `locationName:"nameModifier" type:"string"` + NameModifier *string `locationName:"nameModifier" min:"1" type:"string"` // Specific settings for this type of output. OutputSettings *OutputSettings `locationName:"outputSettings" type:"structure"` @@ -9168,6 +11098,49 @@ func (s Output) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Output) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Output"} + if s.NameModifier != nil && len(*s.NameModifier) < 1 { + invalidParams.Add(request.NewErrParamMinLen("NameModifier", 1)) + } + if s.AudioDescriptions != nil { + for i, v := range s.AudioDescriptions { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AudioDescriptions", i), err.(request.ErrInvalidParams)) + } + } + } + if s.CaptionDescriptions != nil { + for i, v := range s.CaptionDescriptions { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "CaptionDescriptions", i), err.(request.ErrInvalidParams)) + } + } + } + if s.ContainerSettings != nil { + if err := s.ContainerSettings.Validate(); err != nil { + invalidParams.AddNested("ContainerSettings", err.(request.ErrInvalidParams)) + } + } + if s.VideoDescription != nil { + if err := s.VideoDescription.Validate(); err != nil { + invalidParams.AddNested("VideoDescription", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioDescriptions sets the AudioDescriptions field's value. func (s *Output) SetAudioDescriptions(v []*AudioDescription) *Output { s.AudioDescriptions = v @@ -9221,7 +11194,9 @@ type OutputChannelMapping struct { _ struct{} `type:"structure"` // List of input channels - InputChannels []*int64 `locationName:"inputChannels" type:"list"` + // + // InputChannels is a required field + InputChannels []*int64 `locationName:"inputChannels" type:"list" required:"true"` } // String returns the string representation @@ -9234,6 +11209,19 @@ func (s OutputChannelMapping) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *OutputChannelMapping) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "OutputChannelMapping"} + if s.InputChannels == nil { + invalidParams.Add(request.NewErrParamRequired("InputChannels")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetInputChannels sets the InputChannels field's value. func (s *OutputChannelMapping) SetInputChannels(v []*int64) *OutputChannelMapping { s.InputChannels = v @@ -9287,11 +11275,15 @@ type OutputGroup struct { Name *string `locationName:"name" type:"string"` // Output Group settings, including type - OutputGroupSettings *OutputGroupSettings `locationName:"outputGroupSettings" type:"structure"` + // + // OutputGroupSettings is a required field + OutputGroupSettings *OutputGroupSettings `locationName:"outputGroupSettings" type:"structure" required:"true"` // This object holds groups of encoding settings, one group of settings per // output. - Outputs []*Output `locationName:"outputs" type:"list"` + // + // Outputs is a required field + Outputs []*Output `locationName:"outputs" type:"list" required:"true"` } // String returns the string representation @@ -9304,6 +11296,37 @@ func (s OutputGroup) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *OutputGroup) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "OutputGroup"} + if s.OutputGroupSettings == nil { + invalidParams.Add(request.NewErrParamRequired("OutputGroupSettings")) + } + if s.Outputs == nil { + invalidParams.Add(request.NewErrParamRequired("Outputs")) + } + if s.OutputGroupSettings != nil { + if err := s.OutputGroupSettings.Validate(); err != nil { + invalidParams.AddNested("OutputGroupSettings", err.(request.ErrInvalidParams)) + } + } + if s.Outputs != nil { + for i, v := range s.Outputs { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Outputs", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCustomName sets the CustomName field's value. func (s *OutputGroup) SetCustomName(v string) *OutputGroup { s.CustomName = &v @@ -9356,6 +11379,11 @@ func (s *OutputGroupDetail) SetOutputDetails(v []*OutputDetail) *OutputGroupDeta type OutputGroupSettings struct { _ struct{} `type:"structure"` + // Required when you set (Type) under (OutputGroups)>(OutputGroupSettings) to + // CMAF_GROUP_SETTINGS. Each output in a CMAF Output Group may only contain + // a single video, audio, or caption output. + CmafGroupSettings *CmafGroupSettings `locationName:"cmafGroupSettings" type:"structure"` + // Required when you set (Type) under (OutputGroups)>(OutputGroupSettings) to // DASH_ISO_GROUP_SETTINGS. DashIsoGroupSettings *DashIsoGroupSettings `locationName:"dashIsoGroupSettings" type:"structure"` @@ -9372,8 +11400,11 @@ type OutputGroupSettings struct { // MS_SMOOTH_GROUP_SETTINGS. MsSmoothGroupSettings *MsSmoothGroupSettings `locationName:"msSmoothGroupSettings" type:"structure"` - // Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming) - Type *string `locationName:"type" type:"string" enum:"OutputGroupType"` + // Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming, + // CMAF) + // + // Type is a required field + Type *string `locationName:"type" type:"string" required:"true" enum:"OutputGroupType"` } // String returns the string representation @@ -9386,6 +11417,45 @@ func (s OutputGroupSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *OutputGroupSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "OutputGroupSettings"} + if s.Type == nil { + invalidParams.Add(request.NewErrParamRequired("Type")) + } + if s.CmafGroupSettings != nil { + if err := s.CmafGroupSettings.Validate(); err != nil { + invalidParams.AddNested("CmafGroupSettings", err.(request.ErrInvalidParams)) + } + } + if s.DashIsoGroupSettings != nil { + if err := s.DashIsoGroupSettings.Validate(); err != nil { + invalidParams.AddNested("DashIsoGroupSettings", err.(request.ErrInvalidParams)) + } + } + if s.HlsGroupSettings != nil { + if err := s.HlsGroupSettings.Validate(); err != nil { + invalidParams.AddNested("HlsGroupSettings", err.(request.ErrInvalidParams)) + } + } + if s.MsSmoothGroupSettings != nil { + if err := s.MsSmoothGroupSettings.Validate(); err != nil { + invalidParams.AddNested("MsSmoothGroupSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCmafGroupSettings sets the CmafGroupSettings field's value. +func (s *OutputGroupSettings) SetCmafGroupSettings(v *CmafGroupSettings) *OutputGroupSettings { + s.CmafGroupSettings = v + return s +} + // SetDashIsoGroupSettings sets the DashIsoGroupSettings field's value. func (s *OutputGroupSettings) SetDashIsoGroupSettings(v *DashIsoGroupSettings) *OutputGroupSettings { s.DashIsoGroupSettings = v @@ -9461,10 +11531,14 @@ type Preset struct { LastUpdated *time.Time `locationName:"lastUpdated" type:"timestamp" timestampFormat:"unix"` // A name you create for each preset. Each name must be unique within your account. - Name *string `locationName:"name" type:"string"` + // + // Name is a required field + Name *string `locationName:"name" type:"string" required:"true"` // Settings for preset - Settings *PresetSettings `locationName:"settings" type:"structure"` + // + // Settings is a required field + Settings *PresetSettings `locationName:"settings" type:"structure" required:"true"` // A preset can be of two types: system or custom. System or built-in preset // can't be modified or deleted by the user. @@ -9561,6 +11635,46 @@ func (s PresetSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *PresetSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "PresetSettings"} + if s.AudioDescriptions != nil { + for i, v := range s.AudioDescriptions { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "AudioDescriptions", i), err.(request.ErrInvalidParams)) + } + } + } + if s.CaptionDescriptions != nil { + for i, v := range s.CaptionDescriptions { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "CaptionDescriptions", i), err.(request.ErrInvalidParams)) + } + } + } + if s.ContainerSettings != nil { + if err := s.ContainerSettings.Validate(); err != nil { + invalidParams.AddNested("ContainerSettings", err.(request.ErrInvalidParams)) + } + } + if s.VideoDescription != nil { + if err := s.VideoDescription.Validate(); err != nil { + invalidParams.AddNested("VideoDescription", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAudioDescriptions sets the AudioDescriptions field's value. func (s *PresetSettings) SetAudioDescriptions(v []*AudioDescription) *PresetSettings { s.AudioDescriptions = v @@ -9594,27 +11708,35 @@ type ProresSettings struct { // to use for this output. CodecProfile *string `locationName:"codecProfile" type:"string" enum:"ProresCodecProfile"` - // Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want - // the service to use the framerate from the input. Using the console, do this - // by choosing INITIALIZE_FROM_SOURCE for Framerate. + // If you are using the console, use the Framerate setting to specify the framerate + // for this output. If you want to keep the same framerate as the input video, + // choose Follow source. If you want to do framerate conversion, choose a framerate + // from the dropdown list or choose Custom. The framerates shown in the dropdown + // list are decimal approximations of fractions. If you choose Custom, specify + // your framerate as a fraction. If you are creating your transcoding job sepecification + // as a JSON file without the console, use FramerateControl to specify which + // value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE + // if you want the service to use the framerate from the input. Choose SPECIFIED + // if you want the service to use the framerate you specify in the settings + // FramerateNumerator and FramerateDenominator. FramerateControl *string `locationName:"framerateControl" type:"string" enum:"ProresFramerateControl"` // When set to INTERPOLATE, produces smoother motion during framerate conversion. FramerateConversionAlgorithm *string `locationName:"framerateConversionAlgorithm" type:"string" enum:"ProresFramerateConversionAlgorithm"` // Framerate denominator. - FramerateDenominator *int64 `locationName:"framerateDenominator" type:"integer"` + FramerateDenominator *int64 `locationName:"framerateDenominator" min:"1" type:"integer"` // When you use the API for transcode jobs that use framerate conversion, specify // the framerate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use // FramerateNumerator to specify the numerator of this fraction. In this example, // use 24000 for the value of FramerateNumerator. - FramerateNumerator *int64 `locationName:"framerateNumerator" type:"integer"` + FramerateNumerator *int64 `locationName:"framerateNumerator" min:"1" type:"integer"` // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top - // or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default + // or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, // behavior depends on the input scan type. - If the source is interlaced, the // output will be interlaced with the same polarity as the source (it will follow @@ -9632,10 +11754,10 @@ type ProresSettings struct { ParControl *string `locationName:"parControl" type:"string" enum:"ProresParControl"` // Pixel Aspect Ratio denominator. - ParDenominator *int64 `locationName:"parDenominator" type:"integer"` + ParDenominator *int64 `locationName:"parDenominator" min:"1" type:"integer"` // Pixel Aspect Ratio numerator. - ParNumerator *int64 `locationName:"parNumerator" type:"integer"` + ParNumerator *int64 `locationName:"parNumerator" min:"1" type:"integer"` // Enables Slow PAL rate conversion. 23.976fps and 24fps input is relabeled // as 25fps, and audio is sped up correspondingly. @@ -9658,6 +11780,28 @@ func (s ProresSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *ProresSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ProresSettings"} + if s.FramerateDenominator != nil && *s.FramerateDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateDenominator", 1)) + } + if s.FramerateNumerator != nil && *s.FramerateNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("FramerateNumerator", 1)) + } + if s.ParDenominator != nil && *s.ParDenominator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParDenominator", 1)) + } + if s.ParNumerator != nil && *s.ParNumerator < 1 { + invalidParams.Add(request.NewErrParamMinValue("ParNumerator", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCodecProfile sets the CodecProfile field's value. func (s *ProresSettings) SetCodecProfile(v string) *ProresSettings { s.CodecProfile = &v @@ -9743,7 +11887,9 @@ type Queue struct { LastUpdated *time.Time `locationName:"lastUpdated" type:"timestamp" timestampFormat:"unix"` // A name you create for each queue. Each name must be unique within your account. - Name *string `locationName:"name" type:"string"` + // + // Name is a required field + Name *string `locationName:"name" type:"string" required:"true"` // Queues can be ACTIVE or PAUSED. If you pause a queue, jobs in that queue // will not begin. Jobs running when a queue is paused continue to run until @@ -9812,18 +11958,26 @@ type Rectangle struct { _ struct{} `type:"structure"` // Height of rectangle in pixels. - Height *int64 `locationName:"height" type:"integer"` + // + // Height is a required field + Height *int64 `locationName:"height" type:"integer" required:"true"` // Width of rectangle in pixels. - Width *int64 `locationName:"width" type:"integer"` + // + // Width is a required field + Width *int64 `locationName:"width" type:"integer" required:"true"` // The distance, in pixels, between the rectangle and the left edge of the video // frame. - X *int64 `locationName:"x" type:"integer"` + // + // X is a required field + X *int64 `locationName:"x" type:"integer" required:"true"` // The distance, in pixels, between the rectangle and the top edge of the video // frame. - Y *int64 `locationName:"y" type:"integer"` + // + // Y is a required field + Y *int64 `locationName:"y" type:"integer" required:"true"` } // String returns the string representation @@ -9836,6 +11990,40 @@ func (s Rectangle) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *Rectangle) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Rectangle"} + if s.Height == nil { + invalidParams.Add(request.NewErrParamRequired("Height")) + } + if s.Height != nil && *s.Height < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Height", -2.147483648e+09)) + } + if s.Width == nil { + invalidParams.Add(request.NewErrParamRequired("Width")) + } + if s.Width != nil && *s.Width < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Width", -2.147483648e+09)) + } + if s.X == nil { + invalidParams.Add(request.NewErrParamRequired("X")) + } + if s.X != nil && *s.X < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("X", -2.147483648e+09)) + } + if s.Y == nil { + invalidParams.Add(request.NewErrParamRequired("Y")) + } + if s.Y != nil && *s.Y < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("Y", -2.147483648e+09)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetHeight sets the Height field's value. func (s *Rectangle) SetHeight(v int64) *Rectangle { s.Height = &v @@ -9861,8 +12049,8 @@ func (s *Rectangle) SetY(v int64) *Rectangle { } // Use Manual audio remixing (RemixSettings) to adjust audio levels for each -// output channel. With audio remixing, you can output more or fewer audio channels -// than your input audio source provides. +// audio channel in each output of your job. With audio remixing, you can output +// more or fewer audio channels than your input audio source provides. type RemixSettings struct { _ struct{} `type:"structure"` @@ -9870,16 +12058,22 @@ type RemixSettings struct { // remixing value for each channel. Units are in dB. Acceptable values are within // the range from -60 (mute) through 6. A setting of 0 passes the input channel // unchanged to the output channel (no attenuation or amplification). - ChannelMapping *ChannelMapping `locationName:"channelMapping" type:"structure"` + // + // ChannelMapping is a required field + ChannelMapping *ChannelMapping `locationName:"channelMapping" type:"structure" required:"true"` // Specify the number of audio channels from your input that you want to use // in your output. With remixing, you might combine or split the data in these // channels, so the number of channels in your final output might be different. - ChannelsIn *int64 `locationName:"channelsIn" type:"integer"` + // + // ChannelsIn is a required field + ChannelsIn *int64 `locationName:"channelsIn" min:"1" type:"integer" required:"true"` // Specify the number of channels in this output after remixing. Valid values: // 1, 2, 4, 6, 8 - ChannelsOut *int64 `locationName:"channelsOut" type:"integer"` + // + // ChannelsOut is a required field + ChannelsOut *int64 `locationName:"channelsOut" min:"1" type:"integer" required:"true"` } // String returns the string representation @@ -9892,6 +12086,36 @@ func (s RemixSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *RemixSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RemixSettings"} + if s.ChannelMapping == nil { + invalidParams.Add(request.NewErrParamRequired("ChannelMapping")) + } + if s.ChannelsIn == nil { + invalidParams.Add(request.NewErrParamRequired("ChannelsIn")) + } + if s.ChannelsIn != nil && *s.ChannelsIn < 1 { + invalidParams.Add(request.NewErrParamMinValue("ChannelsIn", 1)) + } + if s.ChannelsOut == nil { + invalidParams.Add(request.NewErrParamRequired("ChannelsOut")) + } + if s.ChannelsOut != nil && *s.ChannelsOut < 1 { + invalidParams.Add(request.NewErrParamMinValue("ChannelsOut", 1)) + } + if s.ChannelMapping != nil { + if err := s.ChannelMapping.Validate(); err != nil { + invalidParams.AddNested("ChannelMapping", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetChannelMapping sets the ChannelMapping field's value. func (s *RemixSettings) SetChannelMapping(v *ChannelMapping) *RemixSettings { s.ChannelMapping = v @@ -9943,15 +12167,21 @@ type SpekeKeyProvider struct { _ struct{} `type:"structure"` // The SPEKE-compliant server uses Resource ID (ResourceId) to identify content. - ResourceId *string `locationName:"resourceId" type:"string"` + // + // ResourceId is a required field + ResourceId *string `locationName:"resourceId" type:"string" required:"true"` // Relates to SPEKE implementation. DRM system identifiers. DASH output groups // support a max of two system ids. Other group types support one system id. - SystemIds []*string `locationName:"systemIds" type:"list"` + // + // SystemIds is a required field + SystemIds []*string `locationName:"systemIds" type:"list" required:"true"` // Use URL (Url) to specify the SPEKE-compliant server that will provide keys // for content. - Url *string `locationName:"url" type:"string"` + // + // Url is a required field + Url *string `locationName:"url" type:"string" required:"true"` } // String returns the string representation @@ -9964,6 +12194,25 @@ func (s SpekeKeyProvider) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *SpekeKeyProvider) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "SpekeKeyProvider"} + if s.ResourceId == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceId")) + } + if s.SystemIds == nil { + invalidParams.Add(request.NewErrParamRequired("SystemIds")) + } + if s.Url == nil { + invalidParams.Add(request.NewErrParamRequired("Url")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetResourceId sets the ResourceId field's value. func (s *SpekeKeyProvider) SetResourceId(v string) *SpekeKeyProvider { s.ResourceId = &v @@ -9997,11 +12246,15 @@ type StaticKeyProvider struct { // Relates to DRM implementation. Use a 32-character hexidecimal string to specify // Key Value (StaticKeyValue). - StaticKeyValue *string `locationName:"staticKeyValue" type:"string"` + // + // StaticKeyValue is a required field + StaticKeyValue *string `locationName:"staticKeyValue" type:"string" required:"true"` // Relates to DRM implementation. The location of the license server used for // protecting content. - Url *string `locationName:"url" type:"string"` + // + // Url is a required field + Url *string `locationName:"url" type:"string" required:"true"` } // String returns the string representation @@ -10014,6 +12267,22 @@ func (s StaticKeyProvider) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *StaticKeyProvider) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StaticKeyProvider"} + if s.StaticKeyValue == nil { + invalidParams.Add(request.NewErrParamRequired("StaticKeyValue")) + } + if s.Url == nil { + invalidParams.Add(request.NewErrParamRequired("Url")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetKeyFormat sets the KeyFormat field's value. func (s *StaticKeyProvider) SetKeyFormat(v string) *StaticKeyProvider { s.KeyFormat = &v @@ -10046,7 +12315,7 @@ type TeletextDestinationSettings struct { // this output. This value must be a three-digit hexadecimal string; strings // ending in -FF are invalid. If you are passing through the entire set of Teletext // data, do not use this field. - PageNumber *string `locationName:"pageNumber" type:"string"` + PageNumber *string `locationName:"pageNumber" min:"3" type:"string"` } // String returns the string representation @@ -10059,6 +12328,19 @@ func (s TeletextDestinationSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *TeletextDestinationSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "TeletextDestinationSettings"} + if s.PageNumber != nil && len(*s.PageNumber) < 3 { + invalidParams.Add(request.NewErrParamMinLen("PageNumber", 3)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetPageNumber sets the PageNumber field's value. func (s *TeletextDestinationSettings) SetPageNumber(v string) *TeletextDestinationSettings { s.PageNumber = &v @@ -10072,7 +12354,7 @@ type TeletextSourceSettings struct { // Use Page Number (PageNumber) to specify the three-digit hexadecimal page // number that will be used for Teletext captions. Do not use this setting if // you are passing through teletext from the input source to output. - PageNumber *string `locationName:"pageNumber" type:"string"` + PageNumber *string `locationName:"pageNumber" min:"3" type:"string"` } // String returns the string representation @@ -10085,6 +12367,19 @@ func (s TeletextSourceSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *TeletextSourceSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "TeletextSourceSettings"} + if s.PageNumber != nil && len(*s.PageNumber) < 3 { + invalidParams.Add(request.NewErrParamMinLen("PageNumber", 3)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetPageNumber sets the PageNumber field's value. func (s *TeletextSourceSettings) SetPageNumber(v string) *TeletextSourceSettings { s.PageNumber = &v @@ -10098,7 +12393,7 @@ type TimecodeBurnin struct { // Use Font Size (FontSize) to set the font size of any burned-in timecode. // Valid values are 10, 16, 32, 48. - FontSize *int64 `locationName:"fontSize" type:"integer"` + FontSize *int64 `locationName:"fontSize" min:"10" type:"integer"` // Use Position (Position) under under Timecode burn-in (TimecodeBurnIn) to // specify the location the burned-in timecode on output video. @@ -10122,6 +12417,19 @@ func (s TimecodeBurnin) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *TimecodeBurnin) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "TimecodeBurnin"} + if s.FontSize != nil && *s.FontSize < 10 { + invalidParams.Add(request.NewErrParamMinValue("FontSize", 10)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetFontSize sets the FontSize field's value. func (s *TimecodeBurnin) SetFontSize(v int64) *TimecodeBurnin { s.FontSize = &v @@ -10140,7 +12448,8 @@ func (s *TimecodeBurnin) SetPrefix(v string) *TimecodeBurnin { return s } -// Contains settings used to acquire and adjust timecode information from inputs. +// These settings control how the service handles timecodes throughout the job. +// These settings don't affect input clipping. type TimecodeConfig struct { _ struct{} `type:"structure"` @@ -10148,40 +12457,40 @@ type TimecodeConfig struct { // Timecode (Anchor) to specify a timecode that will match the input video frame // to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) // or (HH:MM:SS;FF). This setting ignores framerate conversion. System behavior - // for Anchor Timecode varies depending on your setting for Timecode source - // (TimecodeSource). * If Timecode source (TimecodeSource) is set to Specified - // Start (specifiedstart), the first input frame is the specified value in Start - // Timecode (Start). Anchor Timecode (Anchor) and Start Timecode (Start) are - // used calculate output timecode. * If Timecode source (TimecodeSource) is - // set to Start at 0 (zerobased) the first frame is 00:00:00:00. * If Timecode - // source (TimecodeSource) is set to Embedded (embedded), the first frame is - // the timecode value on the first input frame of the input. + // for Anchor Timecode varies depending on your setting for Source (TimecodeSource). + // * If Source (TimecodeSource) is set to Specified Start (SPECIFIEDSTART), + // the first input frame is the specified value in Start Timecode (Start). Anchor + // Timecode (Anchor) and Start Timecode (Start) are used calculate output timecode. + // * If Source (TimecodeSource) is set to Start at 0 (ZEROBASED) the first frame + // is 00:00:00:00. * If Source (TimecodeSource) is set to Embedded (EMBEDDED), + // the first frame is the timecode value on the first input frame of the input. Anchor *string `locationName:"anchor" type:"string"` - // Use Timecode source (TimecodeSource) to set how timecodes are handled within - // this input. To make sure that your video, audio, captions, and markers are - // synchronized and that time-based features, such as image inserter, work correctly, - // choose the Timecode source option that matches your assets. All timecodes - // are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - // - Use the timecode that is in the input video. If no embedded timecode is - // in the source, the service will use Start at 0 (ZEROBASED) instead. * Start + // Use Source (TimecodeSource) to set how timecodes are handled within this + // job. To make sure that your video, audio, captions, and markers are synchronized + // and that time-based features, such as image inserter, work correctly, choose + // the Timecode source option that matches your assets. All timecodes are in + // a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - + // Use the timecode that is in the input video. If no embedded timecode is in + // the source, the service will use Start at 0 (ZEROBASED) instead. * Start // at 0 (ZEROBASED) - Set the timecode of the initial frame to 00:00:00:00. // * Specified Start (SPECIFIEDSTART) - Set the timecode of the initial frame // to a value other than zero. You use Start timecode (Start) to provide this // value. Source *string `locationName:"source" type:"string" enum:"TimecodeSource"` - // Only use when you set Timecode Source (TimecodeSource) to Specified Start - // (SPECIFIEDSTART). Use Start timecode (Start) to specify the timecode for - // the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or - // (HH:MM:SS;FF). + // Only use when you set Source (TimecodeSource) to Specified start (SPECIFIEDSTART). + // Use Start timecode (Start) to specify the timecode for the initial frame. + // Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). Start *string `locationName:"start" type:"string"` - // Only applies to outputs that support program-date-time stamp. Use Time stamp + // Only applies to outputs that support program-date-time stamp. Use Timestamp // offset (TimestampOffset) to overwrite the timecode date without affecting // the time and frame number. Provide the new date as a string in the format // "yyyy-mm-dd". To use Time stamp offset, you must also enable Insert program-date-time - // (InsertProgramDateTime) in the output settings. + // (InsertProgramDateTime) in the output settings. For example, if the date + // part of your timecodes is 2002-1-25 and you want to change it to one year + // later, set Timestamp offset (TimestampOffset) to 2003-1-25. TimestampOffset *string `locationName:"timestampOffset" type:"string"` } @@ -10227,7 +12536,9 @@ type TimedMetadataInsertion struct { _ struct{} `type:"structure"` // Id3Insertions contains the array of Id3Insertion instances. - Id3Insertions []*Id3Insertion `locationName:"id3Insertions" type:"list"` + // + // Id3Insertions is a required field + Id3Insertions []*Id3Insertion `locationName:"id3Insertions" type:"list" required:"true"` } // String returns the string representation @@ -10240,6 +12551,29 @@ func (s TimedMetadataInsertion) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *TimedMetadataInsertion) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "TimedMetadataInsertion"} + if s.Id3Insertions == nil { + invalidParams.Add(request.NewErrParamRequired("Id3Insertions")) + } + if s.Id3Insertions != nil { + for i, v := range s.Id3Insertions { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Id3Insertions", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetId3Insertions sets the Id3Insertions field's value. func (s *TimedMetadataInsertion) SetId3Insertions(v []*Id3Insertion) *TimedMetadataInsertion { s.Id3Insertions = v @@ -10356,6 +12690,11 @@ func (s *UpdateJobTemplateInput) Validate() error { if s.Name == nil { invalidParams.Add(request.NewErrParamRequired("Name")) } + if s.Settings != nil { + if err := s.Settings.Validate(); err != nil { + invalidParams.AddNested("Settings", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -10456,6 +12795,11 @@ func (s *UpdatePresetInput) Validate() error { if s.Name == nil { invalidParams.Add(request.NewErrParamRequired("Name")) } + if s.Settings != nil { + if err := s.Settings.Validate(); err != nil { + invalidParams.AddNested("Settings", err.(request.ErrInvalidParams)) + } + } if invalidParams.Len() > 0 { return invalidParams @@ -10610,7 +12954,9 @@ type VideoCodecSettings struct { _ struct{} `type:"structure"` // Type of video codec - Codec *string `locationName:"codec" type:"string" enum:"VideoCodec"` + // + // Codec is a required field + Codec *string `locationName:"codec" type:"string" required:"true" enum:"VideoCodec"` // Required when you set (Codec) under (VideoDescription)>(CodecSettings) to // the value FRAME_CAPTURE. @@ -10642,6 +12988,44 @@ func (s VideoCodecSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *VideoCodecSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "VideoCodecSettings"} + if s.Codec == nil { + invalidParams.Add(request.NewErrParamRequired("Codec")) + } + if s.FrameCaptureSettings != nil { + if err := s.FrameCaptureSettings.Validate(); err != nil { + invalidParams.AddNested("FrameCaptureSettings", err.(request.ErrInvalidParams)) + } + } + if s.H264Settings != nil { + if err := s.H264Settings.Validate(); err != nil { + invalidParams.AddNested("H264Settings", err.(request.ErrInvalidParams)) + } + } + if s.H265Settings != nil { + if err := s.H265Settings.Validate(); err != nil { + invalidParams.AddNested("H265Settings", err.(request.ErrInvalidParams)) + } + } + if s.Mpeg2Settings != nil { + if err := s.Mpeg2Settings.Validate(); err != nil { + invalidParams.AddNested("Mpeg2Settings", err.(request.ErrInvalidParams)) + } + } + if s.ProresSettings != nil { + if err := s.ProresSettings.Validate(); err != nil { + invalidParams.AddNested("ProresSettings", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetCodec sets the Codec field's value. func (s *VideoCodecSettings) SetCodec(v string) *VideoCodecSettings { s.Codec = &v @@ -10683,11 +13067,11 @@ type VideoDescription struct { _ struct{} `type:"structure"` // This setting only applies to H.264 and MPEG2 outputs. Use Insert AFD signaling - // (AfdSignaling) to whether there are AFD values in the output video data and - // what those values are. * Choose None to remove all AFD values from this output. - // * Choose Fixed to ignore input AFD values and instead encode the value specified - // in the job. * Choose Auto to calculate output AFD values based on the input - // AFD scaler data. + // (AfdSignaling) to specify whether the service includes AFD values in the + // output video data and what those values are. * Choose None to remove all + // AFD values from this output. * Choose Fixed to ignore input AFD values and + // instead encode the value specified in the job. * Choose Auto to calculate + // output AFD values based on the input AFD scaler data. AfdSignaling *string `locationName:"afdSignaling" type:"string" enum:"AfdSignaling"` // Enable Anti-alias (AntiAlias) to enhance sharp edges in video output when @@ -10702,7 +13086,9 @@ type VideoDescription struct { // lists the codec enum, settings object pairs. * H_264, H264Settings * H_265, // H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * FRAME_CAPTURE, // FrameCaptureSettings - CodecSettings *VideoCodecSettings `locationName:"codecSettings" type:"structure"` + // + // CodecSettings is a required field + CodecSettings *VideoCodecSettings `locationName:"codecSettings" type:"structure" required:"true"` // Enable Insert color metadata (ColorMetadata) to include color metadata in // this output. This setting is enabled by default. @@ -10729,7 +13115,7 @@ type VideoDescription struct { // Use the Height (Height) setting to define the video resolution height for // this output. Specify in pixels. If you don't provide a value here, the service // will use the input height. - Height *int64 `locationName:"height" type:"integer"` + Height *int64 `locationName:"height" min:"32" type:"integer"` // Use Position (Position) to point to a rectangle object to define your position. // This setting overrides any other aspect ratio. @@ -10759,12 +13145,18 @@ type VideoDescription struct { // setting, 100 the sharpest, and 50 recommended for most content. Sharpness *int64 `locationName:"sharpness" type:"integer"` - // Enable Timecode insertion to include timecode information in this output. - // Do this in the API by setting (VideoTimecodeInsertion) to (PIC_TIMING_SEI). - // To get timecodes to appear correctly in your output, also set up the timecode - // configuration for your job in the input settings. Only enable Timecode insertion - // when the input framerate is identical to output framerate. Disable this setting - // to remove the timecode from the output. Default is disabled. + // Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode + // insertion when the input framerate is identical to the output framerate. + // To include timecodes in this output, set Timecode insertion (VideoTimecodeInsertion) + // to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. + // When the service inserts timecodes in an output, by default, it uses any + // embedded timecodes from the input. If none are present, the service will + // set the timecode for the first output frame to zero. To change this default + // behavior, adjust the settings under Timecode configuration (TimecodeConfig). + // In the console, these settings are located under Job > Job settings > Timecode + // configuration. Note - Timecode source under input settings (InputTimecodeSource) + // does not affect the timecodes that are inserted in the output. Source under + // Job settings > Timecode configuration (TimecodeSource) does. TimecodeInsertion *string `locationName:"timecodeInsertion" type:"string" enum:"VideoTimecodeInsertion"` // Find additional transcoding features under Preprocessors (VideoPreprocessors). @@ -10775,7 +13167,7 @@ type VideoDescription struct { // Use Width (Width) to define the video resolution width, in pixels, for this // output. If you don't provide a value here, the service will use the input // width. - Width *int64 `locationName:"width" type:"integer"` + Width *int64 `locationName:"width" min:"32" type:"integer"` } // String returns the string representation @@ -10788,6 +13180,45 @@ func (s VideoDescription) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *VideoDescription) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "VideoDescription"} + if s.CodecSettings == nil { + invalidParams.Add(request.NewErrParamRequired("CodecSettings")) + } + if s.Height != nil && *s.Height < 32 { + invalidParams.Add(request.NewErrParamMinValue("Height", 32)) + } + if s.Width != nil && *s.Width < 32 { + invalidParams.Add(request.NewErrParamMinValue("Width", 32)) + } + if s.CodecSettings != nil { + if err := s.CodecSettings.Validate(); err != nil { + invalidParams.AddNested("CodecSettings", err.(request.ErrInvalidParams)) + } + } + if s.Crop != nil { + if err := s.Crop.Validate(); err != nil { + invalidParams.AddNested("Crop", err.(request.ErrInvalidParams)) + } + } + if s.Position != nil { + if err := s.Position.Validate(); err != nil { + invalidParams.AddNested("Position", err.(request.ErrInvalidParams)) + } + } + if s.VideoPreprocessors != nil { + if err := s.VideoPreprocessors.Validate(); err != nil { + invalidParams.AddNested("VideoPreprocessors", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetAfdSignaling sets the AfdSignaling field's value. func (s *VideoDescription) SetAfdSignaling(v string) *VideoDescription { s.AfdSignaling = &v @@ -10951,6 +13382,36 @@ func (s VideoPreprocessor) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *VideoPreprocessor) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "VideoPreprocessor"} + if s.ColorCorrector != nil { + if err := s.ColorCorrector.Validate(); err != nil { + invalidParams.AddNested("ColorCorrector", err.(request.ErrInvalidParams)) + } + } + if s.ImageInserter != nil { + if err := s.ImageInserter.Validate(); err != nil { + invalidParams.AddNested("ImageInserter", err.(request.ErrInvalidParams)) + } + } + if s.NoiseReducer != nil { + if err := s.NoiseReducer.Validate(); err != nil { + invalidParams.AddNested("NoiseReducer", err.(request.ErrInvalidParams)) + } + } + if s.TimecodeBurnin != nil { + if err := s.TimecodeBurnin.Validate(); err != nil { + invalidParams.AddNested("TimecodeBurnin", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetColorCorrector sets the ColorCorrector field's value. func (s *VideoPreprocessor) SetColorCorrector(v *ColorCorrector) *VideoPreprocessor { s.ColorCorrector = v @@ -10985,30 +13446,39 @@ func (s *VideoPreprocessor) SetTimecodeBurnin(v *TimecodeBurnin) *VideoPreproces type VideoSelector struct { _ struct{} `type:"structure"` - // Specifies the colorspace of an input. This setting works in tandem with "Color - // Corrector":#color_corrector > color_space_conversion to determine if any - // conversion will be performed. + // If your input video has accurate color space metadata, or if you don't know + // about color space, leave this set to the default value FOLLOW. The service + // will automatically detect your input color space. If your input video has + // metadata indicating the wrong color space, or if your input video is missing + // color space metadata that should be there, specify the accurate color space + // here. If you choose HDR10, you can also correct inaccurate color space coefficients, + // using the HDR master display information controls. You must also set Color + // space usage (ColorSpaceUsage) to FORCE for the service to use these values. ColorSpace *string `locationName:"colorSpace" type:"string" enum:"ColorSpace"` - // There are two sources for color metadata, the input file and the job configuration. - // This enum controls which takes precedence. FORCE: System will use color metadata - // supplied by user, if any. If the user does not supply color metadata the - // system will use data from the source. FALLBACK: System will use color metadata - // from the source. If source has no color metadata, the system will use user-supplied - // color metadata values if available. + // There are two sources for color metadata, the input file and the job configuration + // (in the Color space and HDR master display informaiton settings). The Color + // space usage setting controls which takes precedence. FORCE: The system will + // use color metadata supplied by user, if any. If the user does not supply + // color metadata, the system will use data from the source. FALLBACK: The system + // will use color metadata from the source. If source has no color metadata, + // the system will use user-supplied color metadata values if available. ColorSpaceUsage *string `locationName:"colorSpaceUsage" type:"string" enum:"ColorSpaceUsage"` - // Use the HDR master display (Hdr10Metadata) settings to provide values for - // HDR color. These values vary depending on the input video and must be provided - // by a color grader. Range is 0 to 50,000, each increment represents 0.00002 - // in CIE1931 color coordinate. + // Use the HDR master display (Hdr10Metadata) settings to correct HDR metadata + // or to provide missing metadata. These values vary depending on the input + // video and must be provided by a color grader. Range is 0 to 50,000, each + // increment represents 0.00002 in CIE1931 color coordinate. Note that these + // settings are not color correction. Note that if you are creating HDR outputs + // inside of an HLS CMAF package, to comply with the Apple specification, you + // must use the HVC1 for H.265 setting. Hdr10Metadata *Hdr10Metadata `locationName:"hdr10Metadata" type:"structure"` // Use PID (Pid) to select specific video data from an input file. Specify this // value as an integer; the system automatically converts it to the hexidecimal // value. For example, 257 selects PID 0x101. A PID, or packet identifier, is // an identifier for a set of data in an MPEG-2 transport stream container. - Pid *int64 `locationName:"pid" type:"integer"` + Pid *int64 `locationName:"pid" min:"1" type:"integer"` // Selects a specific program from within a multi-program transport stream. // Note that Quad 4K is not currently supported. @@ -11025,6 +13495,27 @@ func (s VideoSelector) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *VideoSelector) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "VideoSelector"} + if s.Pid != nil && *s.Pid < 1 { + invalidParams.Add(request.NewErrParamMinValue("Pid", 1)) + } + if s.ProgramNumber != nil && *s.ProgramNumber < -2.147483648e+09 { + invalidParams.Add(request.NewErrParamMinValue("ProgramNumber", -2.147483648e+09)) + } + if s.Hdr10Metadata != nil { + if err := s.Hdr10Metadata.Validate(); err != nil { + invalidParams.AddNested("Hdr10Metadata", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetColorSpace sets the ColorSpace field's value. func (s *VideoSelector) SetColorSpace(v string) *VideoSelector { s.ColorSpace = &v @@ -11062,15 +13553,20 @@ type WavSettings struct { // Specify Bit depth (BitDepth), in bits per sample, to choose the encoding // quality for this audio track. - BitDepth *int64 `locationName:"bitDepth" type:"integer"` + BitDepth *int64 `locationName:"bitDepth" min:"16" type:"integer"` // Set Channels to specify the number of channels in this output audio track. // With WAV, valid values 1, 2, 4, and 8. In the console, these values are Mono, // Stereo, 4-Channel, and 8-Channel, respectively. - Channels *int64 `locationName:"channels" type:"integer"` + Channels *int64 `locationName:"channels" min:"1" type:"integer"` + + // The service defaults to using RIFF for WAV outputs. If your output audio + // is likely to exceed 4 GB in file size, or if you otherwise need the extended + // support of the RF64 format, set your output WAV file format to RF64. + Format *string `locationName:"format" type:"string" enum:"WavFormat"` // Sample rate in Hz. - SampleRate *int64 `locationName:"sampleRate" type:"integer"` + SampleRate *int64 `locationName:"sampleRate" min:"8000" type:"integer"` } // String returns the string representation @@ -11083,6 +13579,25 @@ func (s WavSettings) GoString() string { return s.String() } +// Validate inspects the fields of the type to determine if they are valid. +func (s *WavSettings) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "WavSettings"} + if s.BitDepth != nil && *s.BitDepth < 16 { + invalidParams.Add(request.NewErrParamMinValue("BitDepth", 16)) + } + if s.Channels != nil && *s.Channels < 1 { + invalidParams.Add(request.NewErrParamMinValue("Channels", 1)) + } + if s.SampleRate != nil && *s.SampleRate < 8000 { + invalidParams.Add(request.NewErrParamMinValue("SampleRate", 8000)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + // SetBitDepth sets the BitDepth field's value. func (s *WavSettings) SetBitDepth(v int64) *WavSettings { s.BitDepth = &v @@ -11095,6 +13610,12 @@ func (s *WavSettings) SetChannels(v int64) *WavSettings { return s } +// SetFormat sets the Format field's value. +func (s *WavSettings) SetFormat(v string) *WavSettings { + s.Format = &v + return s +} + // SetSampleRate sets the SampleRate field's value. func (s *WavSettings) SetSampleRate(v int64) *WavSettings { s.SampleRate = &v @@ -11271,11 +13792,11 @@ const ( ) // This setting only applies to H.264 and MPEG2 outputs. Use Insert AFD signaling -// (AfdSignaling) to whether there are AFD values in the output video data and -// what those values are. * Choose None to remove all AFD values from this output. -// * Choose Fixed to ignore input AFD values and instead encode the value specified -// in the job. * Choose Auto to calculate output AFD values based on the input -// AFD scaler data. +// (AfdSignaling) to specify whether the service includes AFD values in the +// output video data and what those values are. * Choose None to remove all +// AFD values from this output. * Choose Fixed to ignore input AFD values and +// instead encode the value specified in the job. * Choose Auto to calculate +// output AFD values based on the input AFD scaler data. const ( // AfdSignalingNone is a AfdSignaling enum value AfdSignalingNone = "NONE" @@ -11322,10 +13843,9 @@ const ( AudioCodecPassthrough = "PASSTHROUGH" ) -// When an "Audio Description":#audio_description specifies an AudioSelector -// or AudioSelectorGroup for which no matching source is found in the input, -// then the audio selector marked as DEFAULT will be used. If none are marked -// as default, silence will be inserted for the duration of the input. +// Enable this setting on one audio selector to set it as the default for the +// job. The service uses this default for outputs where it can't find the specified +// input audio. If you don't set a default, those outputs have no audio. const ( // AudioDefaultSelectionDefault is a AudioDefaultSelection enum value AudioDefaultSelectionDefault = "DEFAULT" @@ -11500,9 +14020,11 @@ const ( BurninSubtitleShadowColorWhite = "WHITE" ) -// Controls whether a fixed grid size or proportional font spacing will be used -// to generate the output subtitles bitmap. Only applicable for Teletext inputs -// and DVB-Sub/Burn-in outputs. +// Only applies to jobs with input captions in Teletext or STL formats. Specify +// whether the spacing between letters in your captions is set by the captions +// grid or varies depending on letter width. Choose fixed grid to conform to +// the spacing specified in the captions file more accurately. Choose proportional +// to make the text easier to read if the captions are closed caption. const ( // BurninSubtitleTeletextSpacingFixedGrid is a BurninSubtitleTeletextSpacing enum value BurninSubtitleTeletextSpacingFixedGrid = "FIXED_GRID" @@ -11570,6 +14092,108 @@ const ( CaptionSourceTypeNullSource = "NULL_SOURCE" ) +// When set to ENABLED, sets #EXT-X-ALLOW-CACHE:no tag, which prevents client +// from saving media segments for later replay. +const ( + // CmafClientCacheDisabled is a CmafClientCache enum value + CmafClientCacheDisabled = "DISABLED" + + // CmafClientCacheEnabled is a CmafClientCache enum value + CmafClientCacheEnabled = "ENABLED" +) + +// Specification to use (RFC-6381 or the default RFC-4281) during m3u8 playlist +// generation. +const ( + // CmafCodecSpecificationRfc6381 is a CmafCodecSpecification enum value + CmafCodecSpecificationRfc6381 = "RFC_6381" + + // CmafCodecSpecificationRfc4281 is a CmafCodecSpecification enum value + CmafCodecSpecificationRfc4281 = "RFC_4281" +) + +// Encrypts the segments with the given encryption scheme. Leave blank to disable. +// Selecting 'Disabled' in the web interface also disables encryption. +const ( + // CmafEncryptionTypeSampleAes is a CmafEncryptionType enum value + CmafEncryptionTypeSampleAes = "SAMPLE_AES" +) + +// The Initialization Vector is a 128-bit number used in conjunction with the +// key for encrypting blocks. If set to INCLUDE, Initialization Vector is listed +// in the manifest. Otherwise Initialization Vector is not in the manifest. +const ( + // CmafInitializationVectorInManifestInclude is a CmafInitializationVectorInManifest enum value + CmafInitializationVectorInManifestInclude = "INCLUDE" + + // CmafInitializationVectorInManifestExclude is a CmafInitializationVectorInManifest enum value + CmafInitializationVectorInManifestExclude = "EXCLUDE" +) + +// Indicates which type of key provider is used for encryption. +const ( + // CmafKeyProviderTypeStaticKey is a CmafKeyProviderType enum value + CmafKeyProviderTypeStaticKey = "STATIC_KEY" +) + +// When set to GZIP, compresses HLS playlist. +const ( + // CmafManifestCompressionGzip is a CmafManifestCompression enum value + CmafManifestCompressionGzip = "GZIP" + + // CmafManifestCompressionNone is a CmafManifestCompression enum value + CmafManifestCompressionNone = "NONE" +) + +// Indicates whether the output manifest should use floating point values for +// segment duration. +const ( + // CmafManifestDurationFormatFloatingPoint is a CmafManifestDurationFormat enum value + CmafManifestDurationFormatFloatingPoint = "FLOATING_POINT" + + // CmafManifestDurationFormatInteger is a CmafManifestDurationFormat enum value + CmafManifestDurationFormatInteger = "INTEGER" +) + +// When set to SINGLE_FILE, a single output file is generated, which is internally +// segmented using the Fragment Length and Segment Length. When set to SEGMENTED_FILES, +// separate segment files will be created. +const ( + // CmafSegmentControlSingleFile is a CmafSegmentControl enum value + CmafSegmentControlSingleFile = "SINGLE_FILE" + + // CmafSegmentControlSegmentedFiles is a CmafSegmentControl enum value + CmafSegmentControlSegmentedFiles = "SEGMENTED_FILES" +) + +// Include or exclude RESOLUTION attribute for video in EXT-X-STREAM-INF tag +// of variant manifest. +const ( + // CmafStreamInfResolutionInclude is a CmafStreamInfResolution enum value + CmafStreamInfResolutionInclude = "INCLUDE" + + // CmafStreamInfResolutionExclude is a CmafStreamInfResolution enum value + CmafStreamInfResolutionExclude = "EXCLUDE" +) + +// When set to ENABLED, a DASH MPD manifest will be generated for this output. +const ( + // CmafWriteDASHManifestDisabled is a CmafWriteDASHManifest enum value + CmafWriteDASHManifestDisabled = "DISABLED" + + // CmafWriteDASHManifestEnabled is a CmafWriteDASHManifest enum value + CmafWriteDASHManifestEnabled = "ENABLED" +) + +// When set to ENABLED, an Apple HLS manifest will be generated for this output. +const ( + // CmafWriteHLSManifestDisabled is a CmafWriteHLSManifest enum value + CmafWriteHLSManifestDisabled = "DISABLED" + + // CmafWriteHLSManifestEnabled is a CmafWriteHLSManifest enum value + CmafWriteHLSManifestEnabled = "ENABLED" +) + // Enable Insert color metadata (ColorMetadata) to include color metadata in // this output. This setting is enabled by default. const ( @@ -11580,9 +14204,14 @@ const ( ColorMetadataInsert = "INSERT" ) -// Specifies the colorspace of an input. This setting works in tandem with "Color -// Corrector":#color_corrector > color_space_conversion to determine if any -// conversion will be performed. +// If your input video has accurate color space metadata, or if you don't know +// about color space, leave this set to the default value FOLLOW. The service +// will automatically detect your input color space. If your input video has +// metadata indicating the wrong color space, or if your input video is missing +// color space metadata that should be there, specify the accurate color space +// here. If you choose HDR10, you can also correct inaccurate color space coefficients, +// using the HDR master display information controls. You must also set Color +// space usage (ColorSpaceUsage) to FORCE for the service to use these values. const ( // ColorSpaceFollow is a ColorSpace enum value ColorSpaceFollow = "FOLLOW" @@ -11622,12 +14251,13 @@ const ( ColorSpaceConversionForceHlg2020 = "FORCE_HLG_2020" ) -// There are two sources for color metadata, the input file and the job configuration. -// This enum controls which takes precedence. FORCE: System will use color metadata -// supplied by user, if any. If the user does not supply color metadata the -// system will use data from the source. FALLBACK: System will use color metadata -// from the source. If source has no color metadata, the system will use user-supplied -// color metadata values if available. +// There are two sources for color metadata, the input file and the job configuration +// (in the Color space and HDR master display informaiton settings). The Color +// space usage setting controls which takes precedence. FORCE: The system will +// use color metadata supplied by user, if any. If the user does not supply +// color metadata, the system will use data from the source. FALLBACK: The system +// will use color metadata from the source. If source has no color metadata, +// the system will use user-supplied color metadata values if available. const ( // ColorSpaceUsageForce is a ColorSpaceUsage enum value ColorSpaceUsageForce = "FORCE" @@ -11651,6 +14281,9 @@ const ( // ContainerTypeM3u8 is a ContainerType enum value ContainerTypeM3u8 = "M3U8" + // ContainerTypeCmfc is a ContainerType enum value + ContainerTypeCmfc = "CMFC" + // ContainerTypeMov is a ContainerType enum value ContainerTypeMov = "MOV" @@ -11839,9 +14472,11 @@ const ( DvbSubtitleShadowColorWhite = "WHITE" ) -// Controls whether a fixed grid size or proportional font spacing will be used -// to generate the output subtitles bitmap. Only applicable for Teletext inputs -// and DVB-Sub/Burn-in outputs. +// Only applies to jobs with input captions in Teletext or STL formats. Specify +// whether the spacing between letters in your captions is set by the captions +// grid or varies depending on letter width. Choose fixed grid to conform to +// the spacing specified in the captions file more accurately. Choose proportional +// to make the text easier to read if the captions are closed caption. const ( // DvbSubtitleTeletextSpacingFixedGrid is a DvbSubtitleTeletextSpacing enum value DvbSubtitleTeletextSpacingFixedGrid = "FIXED_GRID" @@ -12195,9 +14830,17 @@ const ( H264FlickerAdaptiveQuantizationEnabled = "ENABLED" ) -// Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want -// the service to use the framerate from the input. Using the console, do this -// by choosing INITIALIZE_FROM_SOURCE for Framerate. +// If you are using the console, use the Framerate setting to specify the framerate +// for this output. If you want to keep the same framerate as the input video, +// choose Follow source. If you want to do framerate conversion, choose a framerate +// from the dropdown list or choose Custom. The framerates shown in the dropdown +// list are decimal approximations of fractions. If you choose Custom, specify +// your framerate as a fraction. If you are creating your transcoding job specification +// as a JSON file without the console, use FramerateControl to specify which +// value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE +// if you want the service to use the framerate from the input. Choose SPECIFIED +// if you want the service to use the framerate you specify in the settings +// FramerateNumerator and FramerateDenominator. const ( // H264FramerateControlInitializeFromSource is a H264FramerateControl enum value H264FramerateControlInitializeFromSource = "INITIALIZE_FROM_SOURCE" @@ -12238,13 +14881,13 @@ const ( // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top -// or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default +// or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, -// behavior depends on the input scan type. - If the source is interlaced, the -// output will be interlaced with the same polarity as the source (it will follow -// the source). The output could therefore be a mix of "top field first" and -// "bottom field first". - If the source is progressive, the output will be -// interlaced with "top field first" or "bottom field first" polarity, depending +// behavior depends on the input scan type, as follows. - If the source is interlaced, +// the output will be interlaced with the same polarity as the source (it will +// follow the source). The output could therefore be a mix of "top field first" +// and "bottom field first". - If the source is progressive, the output will +// be interlaced with "top field first" or "bottom field first" polarity, depending // on which of the Follow options you chose. const ( // H264InterlaceModeProgressive is a H264InterlaceMode enum value @@ -12288,8 +14931,8 @@ const ( H264QualityTuningLevelMultiPassHq = "MULTI_PASS_HQ" ) -// Rate control mode. CQ uses constant quantizer (qp), ABR (average bitrate) -// does not write HRD parameters. +// Use this setting to specify whether this output has a variable bitrate (VBR) +// or constant bitrate (CBR). const ( // H264RateControlModeVbr is a H264RateControlMode enum value H264RateControlModeVbr = "VBR" @@ -12497,9 +15140,17 @@ const ( H265FlickerAdaptiveQuantizationEnabled = "ENABLED" ) -// Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want -// the service to use the framerate from the input. Using the console, do this -// by choosing INITIALIZE_FROM_SOURCE for Framerate. +// If you are using the console, use the Framerate setting to specify the framerate +// for this output. If you want to keep the same framerate as the input video, +// choose Follow source. If you want to do framerate conversion, choose a framerate +// from the dropdown list or choose Custom. The framerates shown in the dropdown +// list are decimal approximations of fractions. If you choose Custom, specify +// your framerate as a fraction. If you are creating your transcoding job sepecification +// as a JSON file without the console, use FramerateControl to specify which +// value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE +// if you want the service to use the framerate from the input. Choose SPECIFIED +// if you want the service to use the framerate you specify in the settings +// FramerateNumerator and FramerateDenominator. const ( // H265FramerateControlInitializeFromSource is a H265FramerateControl enum value H265FramerateControlInitializeFromSource = "INITIALIZE_FROM_SOURCE" @@ -12540,7 +15191,7 @@ const ( // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top -// or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default +// or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, // behavior depends on the input scan type. - If the source is interlaced, the // output will be interlaced with the same polarity as the source (it will follow @@ -12590,8 +15241,8 @@ const ( H265QualityTuningLevelMultiPassHq = "MULTI_PASS_HQ" ) -// Rate control mode. CQ uses constant quantizer (qp), ABR (average bitrate) -// does not write HRD parameters. +// Use this setting to specify whether this output has a variable bitrate (VBR) +// or constant bitrate (CBR). const ( // H265RateControlModeVbr is a H265RateControlMode enum value H265RateControlModeVbr = "VBR" @@ -12705,6 +15356,18 @@ const ( H265UnregisteredSeiTimecodeEnabled = "ENABLED" ) +// If HVC1, output that is H.265 will be marked as HVC1 and adhere to the ISO-IECJTC1-SC29_N13798_Text_ISOIEC_FDIS_14496-15_3rd_E +// spec which states that parameter set NAL units will be stored in the sample +// headers but not in the samples directly. If HEV1, then H.265 will be marked +// as HEV1 and parameter set NAL units will be written into the samples. +const ( + // H265WriteMp4PackagingTypeHvc1 is a H265WriteMp4PackagingType enum value + H265WriteMp4PackagingTypeHvc1 = "HVC1" + + // H265WriteMp4PackagingTypeHev1 is a H265WriteMp4PackagingType enum value + H265WriteMp4PackagingTypeHev1 = "HEV1" +) + const ( // HlsAdMarkersElemental is a HlsAdMarkers enum value HlsAdMarkersElemental = "ELEMENTAL" @@ -12951,13 +15614,13 @@ const ( InputPsiControlUsePsi = "USE_PSI" ) -// Use Timecode source (InputTimecodeSource) to specify how timecode information -// from your input is adjusted and encoded in all outputs for the job. Default -// is embedded. Set to Embedded (EMBEDDED) to use the timecode that is in the -// input video. If no embedded timecode is in the source, will set the timecode -// for the first frame to 00:00:00:00. Set to Start at 0 (ZEROBASED) to set -// the timecode of the initial frame to 00:00:00:00. Set to Specified start -// (SPECIFIEDSTART) to provide the initial timecode yourself the setting (Start). +// Timecode source under input settings (InputTimecodeSource) only affects the +// behavior of features that apply to a single input at a time, such as input +// clipping and synchronizing some captions formats. Use this setting to specify +// whether the service counts frames by timecodes embedded in the video (EMBEDDED) +// or by starting the first frame at zero (ZEROBASED). In both cases, the timecode +// format is HH:MM:SS:FF or HH:MM:SS;FF, where FF is the frame number. Only +// set this to EMBEDDED if your source video has embedded timecodes. const ( // InputTimecodeSourceEmbedded is a InputTimecodeSource enum value InputTimecodeSourceEmbedded = "EMBEDDED" @@ -13001,8 +15664,7 @@ const ( JobTemplateListBySystem = "SYSTEM" ) -// Code to specify the language, following the specification "ISO 639-2 three-digit -// code":http://www.loc.gov/standards/iso639-2/ +// Specify the language, using the ISO 639-2 three-letter code listed at https://www.loc.gov/standards/iso639-2/php/code_list.php. const ( // LanguageCodeEng is a LanguageCode enum value LanguageCodeEng = "ENG" @@ -13880,9 +16542,17 @@ const ( Mpeg2CodecProfileProfile422 = "PROFILE_422" ) -// Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want -// the service to use the framerate from the input. Using the console, do this -// by choosing INITIALIZE_FROM_SOURCE for Framerate. +// If you are using the console, use the Framerate setting to specify the framerate +// for this output. If you want to keep the same framerate as the input video, +// choose Follow source. If you want to do framerate conversion, choose a framerate +// from the dropdown list or choose Custom. The framerates shown in the dropdown +// list are decimal approximations of fractions. If you choose Custom, specify +// your framerate as a fraction. If you are creating your transcoding job sepecification +// as a JSON file without the console, use FramerateControl to specify which +// value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE +// if you want the service to use the framerate from the input. Choose SPECIFIED +// if you want the service to use the framerate you specify in the settings +// FramerateNumerator and FramerateDenominator. const ( // Mpeg2FramerateControlInitializeFromSource is a Mpeg2FramerateControl enum value Mpeg2FramerateControlInitializeFromSource = "INITIALIZE_FROM_SOURCE" @@ -13913,7 +16583,7 @@ const ( // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top -// or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default +// or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, // behavior depends on the input scan type. - If the source is interlaced, the // output will be interlaced with the same polarity as the source (it will follow @@ -14076,8 +16746,8 @@ const ( // Use Noise reducer filter (NoiseReducerFilter) to select one of the following // spatial image filtering functions. To use this setting, you must also enable // Noise reducer (NoiseReducer). * Bilateral is an edge preserving noise reduction -// filter * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) are convolution -// filters * Conserve is a min/max noise reduction filter * Spatial is frequency-domain +// filter. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) are convolution +// filters. * Conserve is a min/max noise reduction filter. * Spatial is a frequency-domain // filter based on JND principles. const ( // NoiseReducerFilterBilateral is a NoiseReducerFilter enum value @@ -14112,7 +16782,8 @@ const ( OrderDescending = "DESCENDING" ) -// Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming) +// Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming, +// CMAF) const ( // OutputGroupTypeHlsGroupSettings is a OutputGroupType enum value OutputGroupTypeHlsGroupSettings = "HLS_GROUP_SETTINGS" @@ -14125,6 +16796,9 @@ const ( // OutputGroupTypeMsSmoothGroupSettings is a OutputGroupType enum value OutputGroupTypeMsSmoothGroupSettings = "MS_SMOOTH_GROUP_SETTINGS" + + // OutputGroupTypeCmafGroupSettings is a OutputGroupType enum value + OutputGroupTypeCmafGroupSettings = "CMAF_GROUP_SETTINGS" ) // Selects method of inserting SDT information into output stream. "Follow input @@ -14177,9 +16851,17 @@ const ( ProresCodecProfileAppleProres422Proxy = "APPLE_PRORES_422_PROXY" ) -// Using the API, set FramerateControl to INITIALIZE_FROM_SOURCE if you want -// the service to use the framerate from the input. Using the console, do this -// by choosing INITIALIZE_FROM_SOURCE for Framerate. +// If you are using the console, use the Framerate setting to specify the framerate +// for this output. If you want to keep the same framerate as the input video, +// choose Follow source. If you want to do framerate conversion, choose a framerate +// from the dropdown list or choose Custom. The framerates shown in the dropdown +// list are decimal approximations of fractions. If you choose Custom, specify +// your framerate as a fraction. If you are creating your transcoding job sepecification +// as a JSON file without the console, use FramerateControl to specify which +// value the service uses for the framerate for this output. Choose INITIALIZE_FROM_SOURCE +// if you want the service to use the framerate from the input. Choose SPECIFIED +// if you want the service to use the framerate you specify in the settings +// FramerateNumerator and FramerateDenominator. const ( // ProresFramerateControlInitializeFromSource is a ProresFramerateControl enum value ProresFramerateControlInitializeFromSource = "INITIALIZE_FROM_SOURCE" @@ -14200,7 +16882,7 @@ const ( // Use Interlace mode (InterlaceMode) to choose the scan line type for the output. // * Top Field First (TOP_FIELD) and Bottom Field First (BOTTOM_FIELD) produce // interlaced output with the entire output having the same field polarity (top -// or bottom first). * Follow, Default Top (FOLLOw_TOP_FIELD) and Follow, Default +// or bottom first). * Follow, Default Top (FOLLOW_TOP_FIELD) and Follow, Default // Bottom (FOLLOW_BOTTOM_FIELD) use the same field polarity as the source. Therefore, // behavior depends on the input scan type. - If the source is interlaced, the // output will be interlaced with the same polarity as the source (it will follow @@ -14364,13 +17046,13 @@ const ( TimecodeBurninPositionBottomRight = "BOTTOM_RIGHT" ) -// Use Timecode source (TimecodeSource) to set how timecodes are handled within -// this input. To make sure that your video, audio, captions, and markers are -// synchronized and that time-based features, such as image inserter, work correctly, -// choose the Timecode source option that matches your assets. All timecodes -// are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) -// - Use the timecode that is in the input video. If no embedded timecode is -// in the source, the service will use Start at 0 (ZEROBASED) instead. * Start +// Use Source (TimecodeSource) to set how timecodes are handled within this +// job. To make sure that your video, audio, captions, and markers are synchronized +// and that time-based features, such as image inserter, work correctly, choose +// the Timecode source option that matches your assets. All timecodes are in +// a 24-hour format with frame number (HH:MM:SS:FF). * Embedded (EMBEDDED) - +// Use the timecode that is in the input video. If no embedded timecode is in +// the source, the service will use Start at 0 (ZEROBASED) instead. * Start // at 0 (ZEROBASED) - Set the timecode of the initial frame to 00:00:00:00. // * Specified Start (SPECIFIEDSTART) - Set the timecode of the initial frame // to a value other than zero. You use Start timecode (Start) to provide this @@ -14386,8 +17068,8 @@ const ( TimecodeSourceSpecifiedstart = "SPECIFIEDSTART" ) -// If PASSTHROUGH, inserts ID3 timed metadata from the timed_metadata REST command -// into this output. +// Applies only to HLS outputs. Use this setting to specify whether the service +// inserts the ID3 timed metadata from the input in this output. const ( // TimedMetadataPassthrough is a TimedMetadata enum value TimedMetadataPassthrough = "PASSTHROUGH" @@ -14432,12 +17114,18 @@ const ( VideoCodecProres = "PRORES" ) -// Enable Timecode insertion to include timecode information in this output. -// Do this in the API by setting (VideoTimecodeInsertion) to (PIC_TIMING_SEI). -// To get timecodes to appear correctly in your output, also set up the timecode -// configuration for your job in the input settings. Only enable Timecode insertion -// when the input framerate is identical to output framerate. Disable this setting -// to remove the timecode from the output. Default is disabled. +// Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode +// insertion when the input framerate is identical to the output framerate. +// To include timecodes in this output, set Timecode insertion (VideoTimecodeInsertion) +// to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. +// When the service inserts timecodes in an output, by default, it uses any +// embedded timecodes from the input. If none are present, the service will +// set the timecode for the first output frame to zero. To change this default +// behavior, adjust the settings under Timecode configuration (TimecodeConfig). +// In the console, these settings are located under Job > Job settings > Timecode +// configuration. Note - Timecode source under input settings (InputTimecodeSource) +// does not affect the timecodes that are inserted in the output. Source under +// Job settings > Timecode configuration (TimecodeSource) does. const ( // VideoTimecodeInsertionDisabled is a VideoTimecodeInsertion enum value VideoTimecodeInsertionDisabled = "DISABLED" @@ -14445,3 +17133,14 @@ const ( // VideoTimecodeInsertionPicTimingSei is a VideoTimecodeInsertion enum value VideoTimecodeInsertionPicTimingSei = "PIC_TIMING_SEI" ) + +// The service defaults to using RIFF for WAV outputs. If your output audio +// is likely to exceed 4 GB in file size, or if you otherwise need the extended +// support of the RF64 format, set your output WAV file format to RF64. +const ( + // WavFormatRiff is a WavFormat enum value + WavFormatRiff = "RIFF" + + // WavFormatRf64 is a WavFormat enum value + WavFormatRf64 = "RF64" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/service.go b/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/service.go index b5610d358..441039b68 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mediaconvert/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mediaconvert" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mediaconvert" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MediaConvert" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MediaConvert client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/medialive/api.go b/vendor/github.com/aws/aws-sdk-go/service/medialive/api.go index f926d2e59..be658cc00 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/medialive/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/medialive/api.go @@ -64,10 +64,10 @@ func (c *MediaLive) CreateChannelRequest(input *CreateChannelInput) (req *reques // API operation CreateChannel for usage and error information. // // Returned Error Codes: -// * ErrCodeUnprocessableEntityException "UnprocessableEntityException" -// // * ErrCodeBadRequestException "BadRequestException" // +// * ErrCodeUnprocessableEntityException "UnprocessableEntityException" +// // * ErrCodeInternalServerErrorException "InternalServerErrorException" // // * ErrCodeForbiddenException "ForbiddenException" @@ -552,6 +552,98 @@ func (c *MediaLive) DeleteInputSecurityGroupWithContext(ctx aws.Context, input * return out, req.Send() } +const opDeleteReservation = "DeleteReservation" + +// DeleteReservationRequest generates a "aws/request.Request" representing the +// client's request for the DeleteReservation operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteReservation for more information on using the DeleteReservation +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteReservationRequest method. +// req, resp := client.DeleteReservationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/DeleteReservation +func (c *MediaLive) DeleteReservationRequest(input *DeleteReservationInput) (req *request.Request, output *DeleteReservationOutput) { + op := &request.Operation{ + Name: opDeleteReservation, + HTTPMethod: "DELETE", + HTTPPath: "/prod/reservations/{reservationId}", + } + + if input == nil { + input = &DeleteReservationInput{} + } + + output = &DeleteReservationOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteReservation API operation for AWS Elemental MediaLive. +// +// Delete an expired reservation. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Elemental MediaLive's +// API operation DeleteReservation for usage and error information. +// +// Returned Error Codes: +// * ErrCodeBadRequestException "BadRequestException" +// +// * ErrCodeInternalServerErrorException "InternalServerErrorException" +// +// * ErrCodeForbiddenException "ForbiddenException" +// +// * ErrCodeBadGatewayException "BadGatewayException" +// +// * ErrCodeNotFoundException "NotFoundException" +// +// * ErrCodeGatewayTimeoutException "GatewayTimeoutException" +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// +// * ErrCodeConflictException "ConflictException" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/DeleteReservation +func (c *MediaLive) DeleteReservation(input *DeleteReservationInput) (*DeleteReservationOutput, error) { + req, out := c.DeleteReservationRequest(input) + return out, req.Send() +} + +// DeleteReservationWithContext is the same as DeleteReservation with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteReservation for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) DeleteReservationWithContext(ctx aws.Context, input *DeleteReservationInput, opts ...request.Option) (*DeleteReservationOutput, error) { + req, out := c.DeleteReservationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDescribeChannel = "DescribeChannel" // DescribeChannelRequest generates a "aws/request.Request" representing the @@ -822,6 +914,186 @@ func (c *MediaLive) DescribeInputSecurityGroupWithContext(ctx aws.Context, input return out, req.Send() } +const opDescribeOffering = "DescribeOffering" + +// DescribeOfferingRequest generates a "aws/request.Request" representing the +// client's request for the DescribeOffering operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeOffering for more information on using the DescribeOffering +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeOfferingRequest method. +// req, resp := client.DescribeOfferingRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/DescribeOffering +func (c *MediaLive) DescribeOfferingRequest(input *DescribeOfferingInput) (req *request.Request, output *DescribeOfferingOutput) { + op := &request.Operation{ + Name: opDescribeOffering, + HTTPMethod: "GET", + HTTPPath: "/prod/offerings/{offeringId}", + } + + if input == nil { + input = &DescribeOfferingInput{} + } + + output = &DescribeOfferingOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeOffering API operation for AWS Elemental MediaLive. +// +// Get details for an offering. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Elemental MediaLive's +// API operation DescribeOffering for usage and error information. +// +// Returned Error Codes: +// * ErrCodeBadRequestException "BadRequestException" +// +// * ErrCodeInternalServerErrorException "InternalServerErrorException" +// +// * ErrCodeForbiddenException "ForbiddenException" +// +// * ErrCodeBadGatewayException "BadGatewayException" +// +// * ErrCodeNotFoundException "NotFoundException" +// +// * ErrCodeGatewayTimeoutException "GatewayTimeoutException" +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/DescribeOffering +func (c *MediaLive) DescribeOffering(input *DescribeOfferingInput) (*DescribeOfferingOutput, error) { + req, out := c.DescribeOfferingRequest(input) + return out, req.Send() +} + +// DescribeOfferingWithContext is the same as DescribeOffering with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeOffering for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) DescribeOfferingWithContext(ctx aws.Context, input *DescribeOfferingInput, opts ...request.Option) (*DescribeOfferingOutput, error) { + req, out := c.DescribeOfferingRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeReservation = "DescribeReservation" + +// DescribeReservationRequest generates a "aws/request.Request" representing the +// client's request for the DescribeReservation operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeReservation for more information on using the DescribeReservation +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeReservationRequest method. +// req, resp := client.DescribeReservationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/DescribeReservation +func (c *MediaLive) DescribeReservationRequest(input *DescribeReservationInput) (req *request.Request, output *DescribeReservationOutput) { + op := &request.Operation{ + Name: opDescribeReservation, + HTTPMethod: "GET", + HTTPPath: "/prod/reservations/{reservationId}", + } + + if input == nil { + input = &DescribeReservationInput{} + } + + output = &DescribeReservationOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeReservation API operation for AWS Elemental MediaLive. +// +// Get details for a reservation. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Elemental MediaLive's +// API operation DescribeReservation for usage and error information. +// +// Returned Error Codes: +// * ErrCodeBadRequestException "BadRequestException" +// +// * ErrCodeInternalServerErrorException "InternalServerErrorException" +// +// * ErrCodeForbiddenException "ForbiddenException" +// +// * ErrCodeBadGatewayException "BadGatewayException" +// +// * ErrCodeNotFoundException "NotFoundException" +// +// * ErrCodeGatewayTimeoutException "GatewayTimeoutException" +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/DescribeReservation +func (c *MediaLive) DescribeReservation(input *DescribeReservationInput) (*DescribeReservationOutput, error) { + req, out := c.DescribeReservationRequest(input) + return out, req.Send() +} + +// DescribeReservationWithContext is the same as DescribeReservation with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeReservation for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) DescribeReservationWithContext(ctx aws.Context, input *DescribeReservationInput, opts ...request.Option) (*DescribeReservationOutput, error) { + req, out := c.DescribeReservationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opListChannels = "ListChannels" // ListChannelsRequest generates a "aws/request.Request" representing the @@ -1254,6 +1526,386 @@ func (c *MediaLive) ListInputsPagesWithContext(ctx aws.Context, input *ListInput return p.Err() } +const opListOfferings = "ListOfferings" + +// ListOfferingsRequest generates a "aws/request.Request" representing the +// client's request for the ListOfferings operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListOfferings for more information on using the ListOfferings +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListOfferingsRequest method. +// req, resp := client.ListOfferingsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/ListOfferings +func (c *MediaLive) ListOfferingsRequest(input *ListOfferingsInput) (req *request.Request, output *ListOfferingsOutput) { + op := &request.Operation{ + Name: opListOfferings, + HTTPMethod: "GET", + HTTPPath: "/prod/offerings", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListOfferingsInput{} + } + + output = &ListOfferingsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListOfferings API operation for AWS Elemental MediaLive. +// +// List offerings available for purchase. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Elemental MediaLive's +// API operation ListOfferings for usage and error information. +// +// Returned Error Codes: +// * ErrCodeBadRequestException "BadRequestException" +// +// * ErrCodeInternalServerErrorException "InternalServerErrorException" +// +// * ErrCodeForbiddenException "ForbiddenException" +// +// * ErrCodeBadGatewayException "BadGatewayException" +// +// * ErrCodeGatewayTimeoutException "GatewayTimeoutException" +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/ListOfferings +func (c *MediaLive) ListOfferings(input *ListOfferingsInput) (*ListOfferingsOutput, error) { + req, out := c.ListOfferingsRequest(input) + return out, req.Send() +} + +// ListOfferingsWithContext is the same as ListOfferings with the addition of +// the ability to pass a context and additional request options. +// +// See ListOfferings for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) ListOfferingsWithContext(ctx aws.Context, input *ListOfferingsInput, opts ...request.Option) (*ListOfferingsOutput, error) { + req, out := c.ListOfferingsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListOfferingsPages iterates over the pages of a ListOfferings operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListOfferings method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListOfferings operation. +// pageNum := 0 +// err := client.ListOfferingsPages(params, +// func(page *ListOfferingsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *MediaLive) ListOfferingsPages(input *ListOfferingsInput, fn func(*ListOfferingsOutput, bool) bool) error { + return c.ListOfferingsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListOfferingsPagesWithContext same as ListOfferingsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) ListOfferingsPagesWithContext(ctx aws.Context, input *ListOfferingsInput, fn func(*ListOfferingsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListOfferingsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListOfferingsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListOfferingsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opListReservations = "ListReservations" + +// ListReservationsRequest generates a "aws/request.Request" representing the +// client's request for the ListReservations operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListReservations for more information on using the ListReservations +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListReservationsRequest method. +// req, resp := client.ListReservationsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/ListReservations +func (c *MediaLive) ListReservationsRequest(input *ListReservationsInput) (req *request.Request, output *ListReservationsOutput) { + op := &request.Operation{ + Name: opListReservations, + HTTPMethod: "GET", + HTTPPath: "/prod/reservations", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListReservationsInput{} + } + + output = &ListReservationsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListReservations API operation for AWS Elemental MediaLive. +// +// List purchased reservations. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Elemental MediaLive's +// API operation ListReservations for usage and error information. +// +// Returned Error Codes: +// * ErrCodeBadRequestException "BadRequestException" +// +// * ErrCodeInternalServerErrorException "InternalServerErrorException" +// +// * ErrCodeForbiddenException "ForbiddenException" +// +// * ErrCodeBadGatewayException "BadGatewayException" +// +// * ErrCodeGatewayTimeoutException "GatewayTimeoutException" +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/ListReservations +func (c *MediaLive) ListReservations(input *ListReservationsInput) (*ListReservationsOutput, error) { + req, out := c.ListReservationsRequest(input) + return out, req.Send() +} + +// ListReservationsWithContext is the same as ListReservations with the addition of +// the ability to pass a context and additional request options. +// +// See ListReservations for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) ListReservationsWithContext(ctx aws.Context, input *ListReservationsInput, opts ...request.Option) (*ListReservationsOutput, error) { + req, out := c.ListReservationsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListReservationsPages iterates over the pages of a ListReservations operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListReservations method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListReservations operation. +// pageNum := 0 +// err := client.ListReservationsPages(params, +// func(page *ListReservationsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *MediaLive) ListReservationsPages(input *ListReservationsInput, fn func(*ListReservationsOutput, bool) bool) error { + return c.ListReservationsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListReservationsPagesWithContext same as ListReservationsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) ListReservationsPagesWithContext(ctx aws.Context, input *ListReservationsInput, fn func(*ListReservationsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListReservationsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListReservationsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListReservationsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opPurchaseOffering = "PurchaseOffering" + +// PurchaseOfferingRequest generates a "aws/request.Request" representing the +// client's request for the PurchaseOffering operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See PurchaseOffering for more information on using the PurchaseOffering +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the PurchaseOfferingRequest method. +// req, resp := client.PurchaseOfferingRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/PurchaseOffering +func (c *MediaLive) PurchaseOfferingRequest(input *PurchaseOfferingInput) (req *request.Request, output *PurchaseOfferingOutput) { + op := &request.Operation{ + Name: opPurchaseOffering, + HTTPMethod: "POST", + HTTPPath: "/prod/offerings/{offeringId}/purchase", + } + + if input == nil { + input = &PurchaseOfferingInput{} + } + + output = &PurchaseOfferingOutput{} + req = c.newRequest(op, input, output) + return +} + +// PurchaseOffering API operation for AWS Elemental MediaLive. +// +// Purchase an offering and create a reservation. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Elemental MediaLive's +// API operation PurchaseOffering for usage and error information. +// +// Returned Error Codes: +// * ErrCodeBadRequestException "BadRequestException" +// +// * ErrCodeInternalServerErrorException "InternalServerErrorException" +// +// * ErrCodeForbiddenException "ForbiddenException" +// +// * ErrCodeBadGatewayException "BadGatewayException" +// +// * ErrCodeNotFoundException "NotFoundException" +// +// * ErrCodeGatewayTimeoutException "GatewayTimeoutException" +// +// * ErrCodeTooManyRequestsException "TooManyRequestsException" +// +// * ErrCodeConflictException "ConflictException" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/medialive-2017-10-14/PurchaseOffering +func (c *MediaLive) PurchaseOffering(input *PurchaseOfferingInput) (*PurchaseOfferingOutput, error) { + req, out := c.PurchaseOfferingRequest(input) + return out, req.Send() +} + +// PurchaseOfferingWithContext is the same as PurchaseOffering with the addition of +// the ability to pass a context and additional request options. +// +// See PurchaseOffering for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaLive) PurchaseOfferingWithContext(ctx aws.Context, input *PurchaseOfferingInput, opts ...request.Option) (*PurchaseOfferingOutput, error) { + req, out := c.PurchaseOfferingRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opStartChannel = "StartChannel" // StartChannelRequest generates a "aws/request.Request" representing the @@ -1492,10 +2144,10 @@ func (c *MediaLive) UpdateChannelRequest(input *UpdateChannelInput) (req *reques // API operation UpdateChannel for usage and error information. // // Returned Error Codes: -// * ErrCodeUnprocessableEntityException "UnprocessableEntityException" -// // * ErrCodeBadRequestException "BadRequestException" // +// * ErrCodeUnprocessableEntityException "UnprocessableEntityException" +// // * ErrCodeInternalServerErrorException "InternalServerErrorException" // // * ErrCodeForbiddenException "ForbiddenException" @@ -3625,6 +4277,9 @@ type Channel struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + // The log level being written to CloudWatch Logs. + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + // The name of the channel. (user-mutable) Name *string `locationName:"name" type:"string"` @@ -3689,6 +4344,12 @@ func (s *Channel) SetInputSpecification(v *InputSpecification) *Channel { return s } +// SetLogLevel sets the LogLevel field's value. +func (s *Channel) SetLogLevel(v string) *Channel { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *Channel) SetName(v string) *Channel { s.Name = &v @@ -3758,6 +4419,9 @@ type ChannelSummary struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + // The log level being written to CloudWatch Logs. + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + // The name of the channel. (user-mutable) Name *string `locationName:"name" type:"string"` @@ -3816,6 +4480,12 @@ func (s *ChannelSummary) SetInputSpecification(v *InputSpecification) *ChannelSu return s } +// SetLogLevel sets the LogLevel field's value. +func (s *ChannelSummary) SetLogLevel(v string) *ChannelSummary { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *ChannelSummary) SetName(v string) *ChannelSummary { s.Name = &v @@ -3851,6 +4521,8 @@ type CreateChannelInput struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + Name *string `locationName:"name" type:"string"` RequestId *string `locationName:"requestId" type:"string" idempotencyToken:"true"` @@ -3919,6 +4591,12 @@ func (s *CreateChannelInput) SetInputSpecification(v *InputSpecification) *Creat return s } +// SetLogLevel sets the LogLevel field's value. +func (s *CreateChannelInput) SetLogLevel(v string) *CreateChannelInput { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *CreateChannelInput) SetName(v string) *CreateChannelInput { s.Name = &v @@ -4147,6 +4825,8 @@ type DeleteChannelOutput struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + Name *string `locationName:"name" type:"string"` PipelinesRunningCount *int64 `locationName:"pipelinesRunningCount" type:"integer"` @@ -4208,6 +4888,12 @@ func (s *DeleteChannelOutput) SetInputSpecification(v *InputSpecification) *Dele return s } +// SetLogLevel sets the LogLevel field's value. +func (s *DeleteChannelOutput) SetLogLevel(v string) *DeleteChannelOutput { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *DeleteChannelOutput) SetName(v string) *DeleteChannelOutput { s.Name = &v @@ -4332,6 +5018,196 @@ func (s DeleteInputSecurityGroupOutput) GoString() string { return s.String() } +type DeleteReservationInput struct { + _ struct{} `type:"structure"` + + // ReservationId is a required field + ReservationId *string `location:"uri" locationName:"reservationId" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteReservationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteReservationInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteReservationInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteReservationInput"} + if s.ReservationId == nil { + invalidParams.Add(request.NewErrParamRequired("ReservationId")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetReservationId sets the ReservationId field's value. +func (s *DeleteReservationInput) SetReservationId(v string) *DeleteReservationInput { + s.ReservationId = &v + return s +} + +type DeleteReservationOutput struct { + _ struct{} `type:"structure"` + + Arn *string `locationName:"arn" type:"string"` + + Count *int64 `locationName:"count" type:"integer"` + + CurrencyCode *string `locationName:"currencyCode" type:"string"` + + Duration *int64 `locationName:"duration" type:"integer"` + + // Units for duration, e.g. 'MONTHS' + DurationUnits *string `locationName:"durationUnits" type:"string" enum:"OfferingDurationUnits"` + + End *string `locationName:"end" type:"string"` + + FixedPrice *float64 `locationName:"fixedPrice" type:"double"` + + Name *string `locationName:"name" type:"string"` + + OfferingDescription *string `locationName:"offeringDescription" type:"string"` + + OfferingId *string `locationName:"offeringId" type:"string"` + + // Offering type, e.g. 'NO_UPFRONT' + OfferingType *string `locationName:"offeringType" type:"string" enum:"OfferingType"` + + Region *string `locationName:"region" type:"string"` + + ReservationId *string `locationName:"reservationId" type:"string"` + + // Resource configuration (codec, resolution, bitrate, ...) + ResourceSpecification *ReservationResourceSpecification `locationName:"resourceSpecification" type:"structure"` + + Start *string `locationName:"start" type:"string"` + + // Current reservation state + State *string `locationName:"state" type:"string" enum:"ReservationState"` + + UsagePrice *float64 `locationName:"usagePrice" type:"double"` +} + +// String returns the string representation +func (s DeleteReservationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteReservationOutput) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *DeleteReservationOutput) SetArn(v string) *DeleteReservationOutput { + s.Arn = &v + return s +} + +// SetCount sets the Count field's value. +func (s *DeleteReservationOutput) SetCount(v int64) *DeleteReservationOutput { + s.Count = &v + return s +} + +// SetCurrencyCode sets the CurrencyCode field's value. +func (s *DeleteReservationOutput) SetCurrencyCode(v string) *DeleteReservationOutput { + s.CurrencyCode = &v + return s +} + +// SetDuration sets the Duration field's value. +func (s *DeleteReservationOutput) SetDuration(v int64) *DeleteReservationOutput { + s.Duration = &v + return s +} + +// SetDurationUnits sets the DurationUnits field's value. +func (s *DeleteReservationOutput) SetDurationUnits(v string) *DeleteReservationOutput { + s.DurationUnits = &v + return s +} + +// SetEnd sets the End field's value. +func (s *DeleteReservationOutput) SetEnd(v string) *DeleteReservationOutput { + s.End = &v + return s +} + +// SetFixedPrice sets the FixedPrice field's value. +func (s *DeleteReservationOutput) SetFixedPrice(v float64) *DeleteReservationOutput { + s.FixedPrice = &v + return s +} + +// SetName sets the Name field's value. +func (s *DeleteReservationOutput) SetName(v string) *DeleteReservationOutput { + s.Name = &v + return s +} + +// SetOfferingDescription sets the OfferingDescription field's value. +func (s *DeleteReservationOutput) SetOfferingDescription(v string) *DeleteReservationOutput { + s.OfferingDescription = &v + return s +} + +// SetOfferingId sets the OfferingId field's value. +func (s *DeleteReservationOutput) SetOfferingId(v string) *DeleteReservationOutput { + s.OfferingId = &v + return s +} + +// SetOfferingType sets the OfferingType field's value. +func (s *DeleteReservationOutput) SetOfferingType(v string) *DeleteReservationOutput { + s.OfferingType = &v + return s +} + +// SetRegion sets the Region field's value. +func (s *DeleteReservationOutput) SetRegion(v string) *DeleteReservationOutput { + s.Region = &v + return s +} + +// SetReservationId sets the ReservationId field's value. +func (s *DeleteReservationOutput) SetReservationId(v string) *DeleteReservationOutput { + s.ReservationId = &v + return s +} + +// SetResourceSpecification sets the ResourceSpecification field's value. +func (s *DeleteReservationOutput) SetResourceSpecification(v *ReservationResourceSpecification) *DeleteReservationOutput { + s.ResourceSpecification = v + return s +} + +// SetStart sets the Start field's value. +func (s *DeleteReservationOutput) SetStart(v string) *DeleteReservationOutput { + s.Start = &v + return s +} + +// SetState sets the State field's value. +func (s *DeleteReservationOutput) SetState(v string) *DeleteReservationOutput { + s.State = &v + return s +} + +// SetUsagePrice sets the UsagePrice field's value. +func (s *DeleteReservationOutput) SetUsagePrice(v float64) *DeleteReservationOutput { + s.UsagePrice = &v + return s +} + type DescribeChannelInput struct { _ struct{} `type:"structure"` @@ -4385,6 +5261,8 @@ type DescribeChannelOutput struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + Name *string `locationName:"name" type:"string"` PipelinesRunningCount *int64 `locationName:"pipelinesRunningCount" type:"integer"` @@ -4446,6 +5324,12 @@ func (s *DescribeChannelOutput) SetInputSpecification(v *InputSpecification) *De return s } +// SetLogLevel sets the LogLevel field's value. +func (s *DescribeChannelOutput) SetLogLevel(v string) *DescribeChannelOutput { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *DescribeChannelOutput) SetName(v string) *DescribeChannelOutput { s.Name = &v @@ -4682,6 +5566,337 @@ func (s *DescribeInputSecurityGroupOutput) SetWhitelistRules(v []*InputWhitelist return s } +type DescribeOfferingInput struct { + _ struct{} `type:"structure"` + + // OfferingId is a required field + OfferingId *string `location:"uri" locationName:"offeringId" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeOfferingInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeOfferingInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeOfferingInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeOfferingInput"} + if s.OfferingId == nil { + invalidParams.Add(request.NewErrParamRequired("OfferingId")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetOfferingId sets the OfferingId field's value. +func (s *DescribeOfferingInput) SetOfferingId(v string) *DescribeOfferingInput { + s.OfferingId = &v + return s +} + +type DescribeOfferingOutput struct { + _ struct{} `type:"structure"` + + Arn *string `locationName:"arn" type:"string"` + + CurrencyCode *string `locationName:"currencyCode" type:"string"` + + Duration *int64 `locationName:"duration" type:"integer"` + + // Units for duration, e.g. 'MONTHS' + DurationUnits *string `locationName:"durationUnits" type:"string" enum:"OfferingDurationUnits"` + + FixedPrice *float64 `locationName:"fixedPrice" type:"double"` + + OfferingDescription *string `locationName:"offeringDescription" type:"string"` + + OfferingId *string `locationName:"offeringId" type:"string"` + + // Offering type, e.g. 'NO_UPFRONT' + OfferingType *string `locationName:"offeringType" type:"string" enum:"OfferingType"` + + Region *string `locationName:"region" type:"string"` + + // Resource configuration (codec, resolution, bitrate, ...) + ResourceSpecification *ReservationResourceSpecification `locationName:"resourceSpecification" type:"structure"` + + UsagePrice *float64 `locationName:"usagePrice" type:"double"` +} + +// String returns the string representation +func (s DescribeOfferingOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeOfferingOutput) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *DescribeOfferingOutput) SetArn(v string) *DescribeOfferingOutput { + s.Arn = &v + return s +} + +// SetCurrencyCode sets the CurrencyCode field's value. +func (s *DescribeOfferingOutput) SetCurrencyCode(v string) *DescribeOfferingOutput { + s.CurrencyCode = &v + return s +} + +// SetDuration sets the Duration field's value. +func (s *DescribeOfferingOutput) SetDuration(v int64) *DescribeOfferingOutput { + s.Duration = &v + return s +} + +// SetDurationUnits sets the DurationUnits field's value. +func (s *DescribeOfferingOutput) SetDurationUnits(v string) *DescribeOfferingOutput { + s.DurationUnits = &v + return s +} + +// SetFixedPrice sets the FixedPrice field's value. +func (s *DescribeOfferingOutput) SetFixedPrice(v float64) *DescribeOfferingOutput { + s.FixedPrice = &v + return s +} + +// SetOfferingDescription sets the OfferingDescription field's value. +func (s *DescribeOfferingOutput) SetOfferingDescription(v string) *DescribeOfferingOutput { + s.OfferingDescription = &v + return s +} + +// SetOfferingId sets the OfferingId field's value. +func (s *DescribeOfferingOutput) SetOfferingId(v string) *DescribeOfferingOutput { + s.OfferingId = &v + return s +} + +// SetOfferingType sets the OfferingType field's value. +func (s *DescribeOfferingOutput) SetOfferingType(v string) *DescribeOfferingOutput { + s.OfferingType = &v + return s +} + +// SetRegion sets the Region field's value. +func (s *DescribeOfferingOutput) SetRegion(v string) *DescribeOfferingOutput { + s.Region = &v + return s +} + +// SetResourceSpecification sets the ResourceSpecification field's value. +func (s *DescribeOfferingOutput) SetResourceSpecification(v *ReservationResourceSpecification) *DescribeOfferingOutput { + s.ResourceSpecification = v + return s +} + +// SetUsagePrice sets the UsagePrice field's value. +func (s *DescribeOfferingOutput) SetUsagePrice(v float64) *DescribeOfferingOutput { + s.UsagePrice = &v + return s +} + +type DescribeReservationInput struct { + _ struct{} `type:"structure"` + + // ReservationId is a required field + ReservationId *string `location:"uri" locationName:"reservationId" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeReservationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeReservationInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeReservationInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeReservationInput"} + if s.ReservationId == nil { + invalidParams.Add(request.NewErrParamRequired("ReservationId")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetReservationId sets the ReservationId field's value. +func (s *DescribeReservationInput) SetReservationId(v string) *DescribeReservationInput { + s.ReservationId = &v + return s +} + +type DescribeReservationOutput struct { + _ struct{} `type:"structure"` + + Arn *string `locationName:"arn" type:"string"` + + Count *int64 `locationName:"count" type:"integer"` + + CurrencyCode *string `locationName:"currencyCode" type:"string"` + + Duration *int64 `locationName:"duration" type:"integer"` + + // Units for duration, e.g. 'MONTHS' + DurationUnits *string `locationName:"durationUnits" type:"string" enum:"OfferingDurationUnits"` + + End *string `locationName:"end" type:"string"` + + FixedPrice *float64 `locationName:"fixedPrice" type:"double"` + + Name *string `locationName:"name" type:"string"` + + OfferingDescription *string `locationName:"offeringDescription" type:"string"` + + OfferingId *string `locationName:"offeringId" type:"string"` + + // Offering type, e.g. 'NO_UPFRONT' + OfferingType *string `locationName:"offeringType" type:"string" enum:"OfferingType"` + + Region *string `locationName:"region" type:"string"` + + ReservationId *string `locationName:"reservationId" type:"string"` + + // Resource configuration (codec, resolution, bitrate, ...) + ResourceSpecification *ReservationResourceSpecification `locationName:"resourceSpecification" type:"structure"` + + Start *string `locationName:"start" type:"string"` + + // Current reservation state + State *string `locationName:"state" type:"string" enum:"ReservationState"` + + UsagePrice *float64 `locationName:"usagePrice" type:"double"` +} + +// String returns the string representation +func (s DescribeReservationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeReservationOutput) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *DescribeReservationOutput) SetArn(v string) *DescribeReservationOutput { + s.Arn = &v + return s +} + +// SetCount sets the Count field's value. +func (s *DescribeReservationOutput) SetCount(v int64) *DescribeReservationOutput { + s.Count = &v + return s +} + +// SetCurrencyCode sets the CurrencyCode field's value. +func (s *DescribeReservationOutput) SetCurrencyCode(v string) *DescribeReservationOutput { + s.CurrencyCode = &v + return s +} + +// SetDuration sets the Duration field's value. +func (s *DescribeReservationOutput) SetDuration(v int64) *DescribeReservationOutput { + s.Duration = &v + return s +} + +// SetDurationUnits sets the DurationUnits field's value. +func (s *DescribeReservationOutput) SetDurationUnits(v string) *DescribeReservationOutput { + s.DurationUnits = &v + return s +} + +// SetEnd sets the End field's value. +func (s *DescribeReservationOutput) SetEnd(v string) *DescribeReservationOutput { + s.End = &v + return s +} + +// SetFixedPrice sets the FixedPrice field's value. +func (s *DescribeReservationOutput) SetFixedPrice(v float64) *DescribeReservationOutput { + s.FixedPrice = &v + return s +} + +// SetName sets the Name field's value. +func (s *DescribeReservationOutput) SetName(v string) *DescribeReservationOutput { + s.Name = &v + return s +} + +// SetOfferingDescription sets the OfferingDescription field's value. +func (s *DescribeReservationOutput) SetOfferingDescription(v string) *DescribeReservationOutput { + s.OfferingDescription = &v + return s +} + +// SetOfferingId sets the OfferingId field's value. +func (s *DescribeReservationOutput) SetOfferingId(v string) *DescribeReservationOutput { + s.OfferingId = &v + return s +} + +// SetOfferingType sets the OfferingType field's value. +func (s *DescribeReservationOutput) SetOfferingType(v string) *DescribeReservationOutput { + s.OfferingType = &v + return s +} + +// SetRegion sets the Region field's value. +func (s *DescribeReservationOutput) SetRegion(v string) *DescribeReservationOutput { + s.Region = &v + return s +} + +// SetReservationId sets the ReservationId field's value. +func (s *DescribeReservationOutput) SetReservationId(v string) *DescribeReservationOutput { + s.ReservationId = &v + return s +} + +// SetResourceSpecification sets the ResourceSpecification field's value. +func (s *DescribeReservationOutput) SetResourceSpecification(v *ReservationResourceSpecification) *DescribeReservationOutput { + s.ResourceSpecification = v + return s +} + +// SetStart sets the Start field's value. +func (s *DescribeReservationOutput) SetStart(v string) *DescribeReservationOutput { + s.Start = &v + return s +} + +// SetState sets the State field's value. +func (s *DescribeReservationOutput) SetState(v string) *DescribeReservationOutput { + s.State = &v + return s +} + +// SetUsagePrice sets the UsagePrice field's value. +func (s *DescribeReservationOutput) SetUsagePrice(v float64) *DescribeReservationOutput { + s.UsagePrice = &v + return s +} + // DVB Network Information Table (NIT) type DvbNitSettings struct { _ struct{} `type:"structure"` @@ -8122,6 +9337,272 @@ func (s *ListInputsOutput) SetNextToken(v string) *ListInputsOutput { return s } +type ListOfferingsInput struct { + _ struct{} `type:"structure"` + + ChannelConfiguration *string `location:"querystring" locationName:"channelConfiguration" type:"string"` + + Codec *string `location:"querystring" locationName:"codec" type:"string"` + + MaxResults *int64 `location:"querystring" locationName:"maxResults" min:"1" type:"integer"` + + MaximumBitrate *string `location:"querystring" locationName:"maximumBitrate" type:"string"` + + MaximumFramerate *string `location:"querystring" locationName:"maximumFramerate" type:"string"` + + NextToken *string `location:"querystring" locationName:"nextToken" type:"string"` + + Resolution *string `location:"querystring" locationName:"resolution" type:"string"` + + ResourceType *string `location:"querystring" locationName:"resourceType" type:"string"` + + SpecialFeature *string `location:"querystring" locationName:"specialFeature" type:"string"` + + VideoQuality *string `location:"querystring" locationName:"videoQuality" type:"string"` +} + +// String returns the string representation +func (s ListOfferingsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListOfferingsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListOfferingsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListOfferingsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetChannelConfiguration sets the ChannelConfiguration field's value. +func (s *ListOfferingsInput) SetChannelConfiguration(v string) *ListOfferingsInput { + s.ChannelConfiguration = &v + return s +} + +// SetCodec sets the Codec field's value. +func (s *ListOfferingsInput) SetCodec(v string) *ListOfferingsInput { + s.Codec = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListOfferingsInput) SetMaxResults(v int64) *ListOfferingsInput { + s.MaxResults = &v + return s +} + +// SetMaximumBitrate sets the MaximumBitrate field's value. +func (s *ListOfferingsInput) SetMaximumBitrate(v string) *ListOfferingsInput { + s.MaximumBitrate = &v + return s +} + +// SetMaximumFramerate sets the MaximumFramerate field's value. +func (s *ListOfferingsInput) SetMaximumFramerate(v string) *ListOfferingsInput { + s.MaximumFramerate = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListOfferingsInput) SetNextToken(v string) *ListOfferingsInput { + s.NextToken = &v + return s +} + +// SetResolution sets the Resolution field's value. +func (s *ListOfferingsInput) SetResolution(v string) *ListOfferingsInput { + s.Resolution = &v + return s +} + +// SetResourceType sets the ResourceType field's value. +func (s *ListOfferingsInput) SetResourceType(v string) *ListOfferingsInput { + s.ResourceType = &v + return s +} + +// SetSpecialFeature sets the SpecialFeature field's value. +func (s *ListOfferingsInput) SetSpecialFeature(v string) *ListOfferingsInput { + s.SpecialFeature = &v + return s +} + +// SetVideoQuality sets the VideoQuality field's value. +func (s *ListOfferingsInput) SetVideoQuality(v string) *ListOfferingsInput { + s.VideoQuality = &v + return s +} + +type ListOfferingsOutput struct { + _ struct{} `type:"structure"` + + NextToken *string `locationName:"nextToken" type:"string"` + + Offerings []*Offering `locationName:"offerings" type:"list"` +} + +// String returns the string representation +func (s ListOfferingsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListOfferingsOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *ListOfferingsOutput) SetNextToken(v string) *ListOfferingsOutput { + s.NextToken = &v + return s +} + +// SetOfferings sets the Offerings field's value. +func (s *ListOfferingsOutput) SetOfferings(v []*Offering) *ListOfferingsOutput { + s.Offerings = v + return s +} + +type ListReservationsInput struct { + _ struct{} `type:"structure"` + + Codec *string `location:"querystring" locationName:"codec" type:"string"` + + MaxResults *int64 `location:"querystring" locationName:"maxResults" min:"1" type:"integer"` + + MaximumBitrate *string `location:"querystring" locationName:"maximumBitrate" type:"string"` + + MaximumFramerate *string `location:"querystring" locationName:"maximumFramerate" type:"string"` + + NextToken *string `location:"querystring" locationName:"nextToken" type:"string"` + + Resolution *string `location:"querystring" locationName:"resolution" type:"string"` + + ResourceType *string `location:"querystring" locationName:"resourceType" type:"string"` + + SpecialFeature *string `location:"querystring" locationName:"specialFeature" type:"string"` + + VideoQuality *string `location:"querystring" locationName:"videoQuality" type:"string"` +} + +// String returns the string representation +func (s ListReservationsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListReservationsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListReservationsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListReservationsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCodec sets the Codec field's value. +func (s *ListReservationsInput) SetCodec(v string) *ListReservationsInput { + s.Codec = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListReservationsInput) SetMaxResults(v int64) *ListReservationsInput { + s.MaxResults = &v + return s +} + +// SetMaximumBitrate sets the MaximumBitrate field's value. +func (s *ListReservationsInput) SetMaximumBitrate(v string) *ListReservationsInput { + s.MaximumBitrate = &v + return s +} + +// SetMaximumFramerate sets the MaximumFramerate field's value. +func (s *ListReservationsInput) SetMaximumFramerate(v string) *ListReservationsInput { + s.MaximumFramerate = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListReservationsInput) SetNextToken(v string) *ListReservationsInput { + s.NextToken = &v + return s +} + +// SetResolution sets the Resolution field's value. +func (s *ListReservationsInput) SetResolution(v string) *ListReservationsInput { + s.Resolution = &v + return s +} + +// SetResourceType sets the ResourceType field's value. +func (s *ListReservationsInput) SetResourceType(v string) *ListReservationsInput { + s.ResourceType = &v + return s +} + +// SetSpecialFeature sets the SpecialFeature field's value. +func (s *ListReservationsInput) SetSpecialFeature(v string) *ListReservationsInput { + s.SpecialFeature = &v + return s +} + +// SetVideoQuality sets the VideoQuality field's value. +func (s *ListReservationsInput) SetVideoQuality(v string) *ListReservationsInput { + s.VideoQuality = &v + return s +} + +type ListReservationsOutput struct { + _ struct{} `type:"structure"` + + NextToken *string `locationName:"nextToken" type:"string"` + + Reservations []*Reservation `locationName:"reservations" type:"list"` +} + +// String returns the string representation +func (s ListReservationsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListReservationsOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *ListReservationsOutput) SetNextToken(v string) *ListReservationsOutput { + s.NextToken = &v + return s +} + +// SetReservations sets the Reservations field's value. +func (s *ListReservationsOutput) SetReservations(v []*Reservation) *ListReservationsOutput { + s.Reservations = v + return s +} + type M2tsSettings struct { _ struct{} `type:"structure"` @@ -9170,6 +10651,121 @@ func (s *NetworkInputSettings) SetServerValidation(v string) *NetworkInputSettin return s } +// Reserved resources available for purchase +type Offering struct { + _ struct{} `type:"structure"` + + // Unique offering ARN, e.g. 'arn:aws:medialive:us-west-2:123456789012:offering:87654321' + Arn *string `locationName:"arn" type:"string"` + + // Currency code for usagePrice and fixedPrice in ISO-4217 format, e.g. 'USD' + CurrencyCode *string `locationName:"currencyCode" type:"string"` + + // Lease duration, e.g. '12' + Duration *int64 `locationName:"duration" type:"integer"` + + // Units for duration, e.g. 'MONTHS' + DurationUnits *string `locationName:"durationUnits" type:"string" enum:"OfferingDurationUnits"` + + // One-time charge for each reserved resource, e.g. '0.0' for a NO_UPFRONT offering + FixedPrice *float64 `locationName:"fixedPrice" type:"double"` + + // Offering description, e.g. 'HD AVC output at 10-20 Mbps, 30 fps, and standard + // VQ in US West (Oregon)' + OfferingDescription *string `locationName:"offeringDescription" type:"string"` + + // Unique offering ID, e.g. '87654321' + OfferingId *string `locationName:"offeringId" type:"string"` + + // Offering type, e.g. 'NO_UPFRONT' + OfferingType *string `locationName:"offeringType" type:"string" enum:"OfferingType"` + + // AWS region, e.g. 'us-west-2' + Region *string `locationName:"region" type:"string"` + + // Resource configuration details + ResourceSpecification *ReservationResourceSpecification `locationName:"resourceSpecification" type:"structure"` + + // Recurring usage charge for each reserved resource, e.g. '157.0' + UsagePrice *float64 `locationName:"usagePrice" type:"double"` +} + +// String returns the string representation +func (s Offering) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Offering) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *Offering) SetArn(v string) *Offering { + s.Arn = &v + return s +} + +// SetCurrencyCode sets the CurrencyCode field's value. +func (s *Offering) SetCurrencyCode(v string) *Offering { + s.CurrencyCode = &v + return s +} + +// SetDuration sets the Duration field's value. +func (s *Offering) SetDuration(v int64) *Offering { + s.Duration = &v + return s +} + +// SetDurationUnits sets the DurationUnits field's value. +func (s *Offering) SetDurationUnits(v string) *Offering { + s.DurationUnits = &v + return s +} + +// SetFixedPrice sets the FixedPrice field's value. +func (s *Offering) SetFixedPrice(v float64) *Offering { + s.FixedPrice = &v + return s +} + +// SetOfferingDescription sets the OfferingDescription field's value. +func (s *Offering) SetOfferingDescription(v string) *Offering { + s.OfferingDescription = &v + return s +} + +// SetOfferingId sets the OfferingId field's value. +func (s *Offering) SetOfferingId(v string) *Offering { + s.OfferingId = &v + return s +} + +// SetOfferingType sets the OfferingType field's value. +func (s *Offering) SetOfferingType(v string) *Offering { + s.OfferingType = &v + return s +} + +// SetRegion sets the Region field's value. +func (s *Offering) SetRegion(v string) *Offering { + s.Region = &v + return s +} + +// SetResourceSpecification sets the ResourceSpecification field's value. +func (s *Offering) SetResourceSpecification(v *ReservationResourceSpecification) *Offering { + s.ResourceSpecification = v + return s +} + +// SetUsagePrice sets the UsagePrice field's value. +func (s *Offering) SetUsagePrice(v float64) *Offering { + s.UsagePrice = &v + return s +} + // Output settings. There can be multiple outputs within a group. type Output struct { _ struct{} `type:"structure"` @@ -9617,6 +11213,92 @@ func (s PassThroughSettings) GoString() string { return s.String() } +type PurchaseOfferingInput struct { + _ struct{} `type:"structure"` + + Count *int64 `locationName:"count" min:"1" type:"integer"` + + Name *string `locationName:"name" type:"string"` + + // OfferingId is a required field + OfferingId *string `location:"uri" locationName:"offeringId" type:"string" required:"true"` + + RequestId *string `locationName:"requestId" type:"string" idempotencyToken:"true"` +} + +// String returns the string representation +func (s PurchaseOfferingInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PurchaseOfferingInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *PurchaseOfferingInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "PurchaseOfferingInput"} + if s.Count != nil && *s.Count < 1 { + invalidParams.Add(request.NewErrParamMinValue("Count", 1)) + } + if s.OfferingId == nil { + invalidParams.Add(request.NewErrParamRequired("OfferingId")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCount sets the Count field's value. +func (s *PurchaseOfferingInput) SetCount(v int64) *PurchaseOfferingInput { + s.Count = &v + return s +} + +// SetName sets the Name field's value. +func (s *PurchaseOfferingInput) SetName(v string) *PurchaseOfferingInput { + s.Name = &v + return s +} + +// SetOfferingId sets the OfferingId field's value. +func (s *PurchaseOfferingInput) SetOfferingId(v string) *PurchaseOfferingInput { + s.OfferingId = &v + return s +} + +// SetRequestId sets the RequestId field's value. +func (s *PurchaseOfferingInput) SetRequestId(v string) *PurchaseOfferingInput { + s.RequestId = &v + return s +} + +type PurchaseOfferingOutput struct { + _ struct{} `type:"structure"` + + // Reserved resources available to use + Reservation *Reservation `locationName:"reservation" type:"structure"` +} + +// String returns the string representation +func (s PurchaseOfferingOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PurchaseOfferingOutput) GoString() string { + return s.String() +} + +// SetReservation sets the Reservation field's value. +func (s *PurchaseOfferingOutput) SetReservation(v *Reservation) *PurchaseOfferingOutput { + s.Reservation = v + return s +} + type RemixSettings struct { _ struct{} `type:"structure"` @@ -9689,6 +11371,253 @@ func (s *RemixSettings) SetChannelsOut(v int64) *RemixSettings { return s } +// Reserved resources available to use +type Reservation struct { + _ struct{} `type:"structure"` + + // Unique reservation ARN, e.g. 'arn:aws:medialive:us-west-2:123456789012:reservation:1234567' + Arn *string `locationName:"arn" type:"string"` + + // Number of reserved resources + Count *int64 `locationName:"count" type:"integer"` + + // Currency code for usagePrice and fixedPrice in ISO-4217 format, e.g. 'USD' + CurrencyCode *string `locationName:"currencyCode" type:"string"` + + // Lease duration, e.g. '12' + Duration *int64 `locationName:"duration" type:"integer"` + + // Units for duration, e.g. 'MONTHS' + DurationUnits *string `locationName:"durationUnits" type:"string" enum:"OfferingDurationUnits"` + + // Reservation UTC end date and time in ISO-8601 format, e.g. '2019-03-01T00:00:00' + End *string `locationName:"end" type:"string"` + + // One-time charge for each reserved resource, e.g. '0.0' for a NO_UPFRONT offering + FixedPrice *float64 `locationName:"fixedPrice" type:"double"` + + // User specified reservation name + Name *string `locationName:"name" type:"string"` + + // Offering description, e.g. 'HD AVC output at 10-20 Mbps, 30 fps, and standard + // VQ in US West (Oregon)' + OfferingDescription *string `locationName:"offeringDescription" type:"string"` + + // Unique offering ID, e.g. '87654321' + OfferingId *string `locationName:"offeringId" type:"string"` + + // Offering type, e.g. 'NO_UPFRONT' + OfferingType *string `locationName:"offeringType" type:"string" enum:"OfferingType"` + + // AWS region, e.g. 'us-west-2' + Region *string `locationName:"region" type:"string"` + + // Unique reservation ID, e.g. '1234567' + ReservationId *string `locationName:"reservationId" type:"string"` + + // Resource configuration details + ResourceSpecification *ReservationResourceSpecification `locationName:"resourceSpecification" type:"structure"` + + // Reservation UTC start date and time in ISO-8601 format, e.g. '2018-03-01T00:00:00' + Start *string `locationName:"start" type:"string"` + + // Current state of reservation, e.g. 'ACTIVE' + State *string `locationName:"state" type:"string" enum:"ReservationState"` + + // Recurring usage charge for each reserved resource, e.g. '157.0' + UsagePrice *float64 `locationName:"usagePrice" type:"double"` +} + +// String returns the string representation +func (s Reservation) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Reservation) GoString() string { + return s.String() +} + +// SetArn sets the Arn field's value. +func (s *Reservation) SetArn(v string) *Reservation { + s.Arn = &v + return s +} + +// SetCount sets the Count field's value. +func (s *Reservation) SetCount(v int64) *Reservation { + s.Count = &v + return s +} + +// SetCurrencyCode sets the CurrencyCode field's value. +func (s *Reservation) SetCurrencyCode(v string) *Reservation { + s.CurrencyCode = &v + return s +} + +// SetDuration sets the Duration field's value. +func (s *Reservation) SetDuration(v int64) *Reservation { + s.Duration = &v + return s +} + +// SetDurationUnits sets the DurationUnits field's value. +func (s *Reservation) SetDurationUnits(v string) *Reservation { + s.DurationUnits = &v + return s +} + +// SetEnd sets the End field's value. +func (s *Reservation) SetEnd(v string) *Reservation { + s.End = &v + return s +} + +// SetFixedPrice sets the FixedPrice field's value. +func (s *Reservation) SetFixedPrice(v float64) *Reservation { + s.FixedPrice = &v + return s +} + +// SetName sets the Name field's value. +func (s *Reservation) SetName(v string) *Reservation { + s.Name = &v + return s +} + +// SetOfferingDescription sets the OfferingDescription field's value. +func (s *Reservation) SetOfferingDescription(v string) *Reservation { + s.OfferingDescription = &v + return s +} + +// SetOfferingId sets the OfferingId field's value. +func (s *Reservation) SetOfferingId(v string) *Reservation { + s.OfferingId = &v + return s +} + +// SetOfferingType sets the OfferingType field's value. +func (s *Reservation) SetOfferingType(v string) *Reservation { + s.OfferingType = &v + return s +} + +// SetRegion sets the Region field's value. +func (s *Reservation) SetRegion(v string) *Reservation { + s.Region = &v + return s +} + +// SetReservationId sets the ReservationId field's value. +func (s *Reservation) SetReservationId(v string) *Reservation { + s.ReservationId = &v + return s +} + +// SetResourceSpecification sets the ResourceSpecification field's value. +func (s *Reservation) SetResourceSpecification(v *ReservationResourceSpecification) *Reservation { + s.ResourceSpecification = v + return s +} + +// SetStart sets the Start field's value. +func (s *Reservation) SetStart(v string) *Reservation { + s.Start = &v + return s +} + +// SetState sets the State field's value. +func (s *Reservation) SetState(v string) *Reservation { + s.State = &v + return s +} + +// SetUsagePrice sets the UsagePrice field's value. +func (s *Reservation) SetUsagePrice(v float64) *Reservation { + s.UsagePrice = &v + return s +} + +// Resource configuration (codec, resolution, bitrate, ...) +type ReservationResourceSpecification struct { + _ struct{} `type:"structure"` + + // Codec, e.g. 'AVC' + Codec *string `locationName:"codec" type:"string" enum:"ReservationCodec"` + + // Maximum bitrate, e.g. 'MAX_20_MBPS' + MaximumBitrate *string `locationName:"maximumBitrate" type:"string" enum:"ReservationMaximumBitrate"` + + // Maximum framerate, e.g. 'MAX_30_FPS' (Outputs only) + MaximumFramerate *string `locationName:"maximumFramerate" type:"string" enum:"ReservationMaximumFramerate"` + + // Resolution, e.g. 'HD' + Resolution *string `locationName:"resolution" type:"string" enum:"ReservationResolution"` + + // Resource type, 'INPUT', 'OUTPUT', or 'CHANNEL' + ResourceType *string `locationName:"resourceType" type:"string" enum:"ReservationResourceType"` + + // Special feature, e.g. 'AUDIO_NORMALIZATION' (Channels only) + SpecialFeature *string `locationName:"specialFeature" type:"string" enum:"ReservationSpecialFeature"` + + // Video quality, e.g. 'STANDARD' (Outputs only) + VideoQuality *string `locationName:"videoQuality" type:"string" enum:"ReservationVideoQuality"` +} + +// String returns the string representation +func (s ReservationResourceSpecification) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ReservationResourceSpecification) GoString() string { + return s.String() +} + +// SetCodec sets the Codec field's value. +func (s *ReservationResourceSpecification) SetCodec(v string) *ReservationResourceSpecification { + s.Codec = &v + return s +} + +// SetMaximumBitrate sets the MaximumBitrate field's value. +func (s *ReservationResourceSpecification) SetMaximumBitrate(v string) *ReservationResourceSpecification { + s.MaximumBitrate = &v + return s +} + +// SetMaximumFramerate sets the MaximumFramerate field's value. +func (s *ReservationResourceSpecification) SetMaximumFramerate(v string) *ReservationResourceSpecification { + s.MaximumFramerate = &v + return s +} + +// SetResolution sets the Resolution field's value. +func (s *ReservationResourceSpecification) SetResolution(v string) *ReservationResourceSpecification { + s.Resolution = &v + return s +} + +// SetResourceType sets the ResourceType field's value. +func (s *ReservationResourceSpecification) SetResourceType(v string) *ReservationResourceSpecification { + s.ResourceType = &v + return s +} + +// SetSpecialFeature sets the SpecialFeature field's value. +func (s *ReservationResourceSpecification) SetSpecialFeature(v string) *ReservationResourceSpecification { + s.SpecialFeature = &v + return s +} + +// SetVideoQuality sets the VideoQuality field's value. +func (s *ReservationResourceSpecification) SetVideoQuality(v string) *ReservationResourceSpecification { + s.VideoQuality = &v + return s +} + type RtmpCaptionInfoDestinationSettings struct { _ struct{} `type:"structure"` } @@ -10208,6 +12137,8 @@ type StartChannelOutput struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + Name *string `locationName:"name" type:"string"` PipelinesRunningCount *int64 `locationName:"pipelinesRunningCount" type:"integer"` @@ -10269,6 +12200,12 @@ func (s *StartChannelOutput) SetInputSpecification(v *InputSpecification) *Start return s } +// SetLogLevel sets the LogLevel field's value. +func (s *StartChannelOutput) SetLogLevel(v string) *StartChannelOutput { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *StartChannelOutput) SetName(v string) *StartChannelOutput { s.Name = &v @@ -10297,7 +12234,9 @@ type StaticKeySettings struct { _ struct{} `type:"structure"` // The URL of the license server used for protecting content. - KeyProviderServer *InputLocation `locationName:"keyProviderServer" type:"structure"` + // + // KeyProviderServer is a required field + KeyProviderServer *InputLocation `locationName:"keyProviderServer" type:"structure" required:"true"` // Static key value as a 32 character hexadecimal string. // @@ -10318,6 +12257,9 @@ func (s StaticKeySettings) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *StaticKeySettings) Validate() error { invalidParams := request.ErrInvalidParams{Context: "StaticKeySettings"} + if s.KeyProviderServer == nil { + invalidParams.Add(request.NewErrParamRequired("KeyProviderServer")) + } if s.StaticKeyValue == nil { invalidParams.Add(request.NewErrParamRequired("StaticKeyValue")) } @@ -10401,6 +12343,8 @@ type StopChannelOutput struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + Name *string `locationName:"name" type:"string"` PipelinesRunningCount *int64 `locationName:"pipelinesRunningCount" type:"integer"` @@ -10462,6 +12406,12 @@ func (s *StopChannelOutput) SetInputSpecification(v *InputSpecification) *StopCh return s } +// SetLogLevel sets the LogLevel field's value. +func (s *StopChannelOutput) SetLogLevel(v string) *StopChannelOutput { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *StopChannelOutput) SetName(v string) *StopChannelOutput { s.Name = &v @@ -10788,6 +12738,8 @@ type UpdateChannelInput struct { InputSpecification *InputSpecification `locationName:"inputSpecification" type:"structure"` + LogLevel *string `locationName:"logLevel" type:"string" enum:"LogLevel"` + Name *string `locationName:"name" type:"string"` RoleArn *string `locationName:"roleArn" type:"string"` @@ -10861,6 +12813,12 @@ func (s *UpdateChannelInput) SetInputSpecification(v *InputSpecification) *Updat return s } +// SetLogLevel sets the LogLevel field's value. +func (s *UpdateChannelInput) SetLogLevel(v string) *UpdateChannelInput { + s.LogLevel = &v + return s +} + // SetName sets the Name field's value. func (s *UpdateChannelInput) SetName(v string) *UpdateChannelInput { s.Name = &v @@ -12603,6 +14561,23 @@ const ( InputTypeUrlPull = "URL_PULL" ) +const ( + // LogLevelError is a LogLevel enum value + LogLevelError = "ERROR" + + // LogLevelWarning is a LogLevel enum value + LogLevelWarning = "WARNING" + + // LogLevelInfo is a LogLevel enum value + LogLevelInfo = "INFO" + + // LogLevelDebug is a LogLevel enum value + LogLevelDebug = "DEBUG" + + // LogLevelDisabled is a LogLevel enum value + LogLevelDisabled = "DISABLED" +) + const ( // M2tsAbsentInputAudioBehaviorDrop is a M2tsAbsentInputAudioBehavior enum value M2tsAbsentInputAudioBehaviorDrop = "DROP" @@ -12799,6 +14774,115 @@ const ( NetworkInputServerValidationCheckCryptographyOnly = "CHECK_CRYPTOGRAPHY_ONLY" ) +// Units for duration, e.g. 'MONTHS' +const ( + // OfferingDurationUnitsMonths is a OfferingDurationUnits enum value + OfferingDurationUnitsMonths = "MONTHS" +) + +// Offering type, e.g. 'NO_UPFRONT' +const ( + // OfferingTypeNoUpfront is a OfferingType enum value + OfferingTypeNoUpfront = "NO_UPFRONT" +) + +// Codec, 'MPEG2', 'AVC', 'HEVC', or 'AUDIO' +const ( + // ReservationCodecMpeg2 is a ReservationCodec enum value + ReservationCodecMpeg2 = "MPEG2" + + // ReservationCodecAvc is a ReservationCodec enum value + ReservationCodecAvc = "AVC" + + // ReservationCodecHevc is a ReservationCodec enum value + ReservationCodecHevc = "HEVC" + + // ReservationCodecAudio is a ReservationCodec enum value + ReservationCodecAudio = "AUDIO" +) + +// Maximum bitrate in megabits per second +const ( + // ReservationMaximumBitrateMax10Mbps is a ReservationMaximumBitrate enum value + ReservationMaximumBitrateMax10Mbps = "MAX_10_MBPS" + + // ReservationMaximumBitrateMax20Mbps is a ReservationMaximumBitrate enum value + ReservationMaximumBitrateMax20Mbps = "MAX_20_MBPS" + + // ReservationMaximumBitrateMax50Mbps is a ReservationMaximumBitrate enum value + ReservationMaximumBitrateMax50Mbps = "MAX_50_MBPS" +) + +// Maximum framerate in frames per second (Outputs only) +const ( + // ReservationMaximumFramerateMax30Fps is a ReservationMaximumFramerate enum value + ReservationMaximumFramerateMax30Fps = "MAX_30_FPS" + + // ReservationMaximumFramerateMax60Fps is a ReservationMaximumFramerate enum value + ReservationMaximumFramerateMax60Fps = "MAX_60_FPS" +) + +// Resolution based on lines of vertical resolution; SD is less than 720 lines, +// HD is 720 to 1080 lines, UHD is greater than 1080 lines +const ( + // ReservationResolutionSd is a ReservationResolution enum value + ReservationResolutionSd = "SD" + + // ReservationResolutionHd is a ReservationResolution enum value + ReservationResolutionHd = "HD" + + // ReservationResolutionUhd is a ReservationResolution enum value + ReservationResolutionUhd = "UHD" +) + +// Resource type, 'INPUT', 'OUTPUT', or 'CHANNEL' +const ( + // ReservationResourceTypeInput is a ReservationResourceType enum value + ReservationResourceTypeInput = "INPUT" + + // ReservationResourceTypeOutput is a ReservationResourceType enum value + ReservationResourceTypeOutput = "OUTPUT" + + // ReservationResourceTypeChannel is a ReservationResourceType enum value + ReservationResourceTypeChannel = "CHANNEL" +) + +// Special features, 'ADVANCED_AUDIO' or 'AUDIO_NORMALIZATION' +const ( + // ReservationSpecialFeatureAdvancedAudio is a ReservationSpecialFeature enum value + ReservationSpecialFeatureAdvancedAudio = "ADVANCED_AUDIO" + + // ReservationSpecialFeatureAudioNormalization is a ReservationSpecialFeature enum value + ReservationSpecialFeatureAudioNormalization = "AUDIO_NORMALIZATION" +) + +// Current reservation state +const ( + // ReservationStateActive is a ReservationState enum value + ReservationStateActive = "ACTIVE" + + // ReservationStateExpired is a ReservationState enum value + ReservationStateExpired = "EXPIRED" + + // ReservationStateCanceled is a ReservationState enum value + ReservationStateCanceled = "CANCELED" + + // ReservationStateDeleted is a ReservationState enum value + ReservationStateDeleted = "DELETED" +) + +// Video quality, e.g. 'STANDARD' (Outputs only) +const ( + // ReservationVideoQualityStandard is a ReservationVideoQuality enum value + ReservationVideoQualityStandard = "STANDARD" + + // ReservationVideoQualityEnhanced is a ReservationVideoQuality enum value + ReservationVideoQualityEnhanced = "ENHANCED" + + // ReservationVideoQualityPremium is a ReservationVideoQuality enum value + ReservationVideoQualityPremium = "PREMIUM" +) + const ( // RtmpCacheFullBehaviorDisconnectImmediately is a RtmpCacheFullBehavior enum value RtmpCacheFullBehaviorDisconnectImmediately = "DISCONNECT_IMMEDIATELY" diff --git a/vendor/github.com/aws/aws-sdk-go/service/medialive/medialiveiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/medialive/medialiveiface/interface.go index 42c10a27b..a870998b8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/medialive/medialiveiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/medialive/medialiveiface/interface.go @@ -84,6 +84,10 @@ type MediaLiveAPI interface { DeleteInputSecurityGroupWithContext(aws.Context, *medialive.DeleteInputSecurityGroupInput, ...request.Option) (*medialive.DeleteInputSecurityGroupOutput, error) DeleteInputSecurityGroupRequest(*medialive.DeleteInputSecurityGroupInput) (*request.Request, *medialive.DeleteInputSecurityGroupOutput) + DeleteReservation(*medialive.DeleteReservationInput) (*medialive.DeleteReservationOutput, error) + DeleteReservationWithContext(aws.Context, *medialive.DeleteReservationInput, ...request.Option) (*medialive.DeleteReservationOutput, error) + DeleteReservationRequest(*medialive.DeleteReservationInput) (*request.Request, *medialive.DeleteReservationOutput) + DescribeChannel(*medialive.DescribeChannelInput) (*medialive.DescribeChannelOutput, error) DescribeChannelWithContext(aws.Context, *medialive.DescribeChannelInput, ...request.Option) (*medialive.DescribeChannelOutput, error) DescribeChannelRequest(*medialive.DescribeChannelInput) (*request.Request, *medialive.DescribeChannelOutput) @@ -96,6 +100,14 @@ type MediaLiveAPI interface { DescribeInputSecurityGroupWithContext(aws.Context, *medialive.DescribeInputSecurityGroupInput, ...request.Option) (*medialive.DescribeInputSecurityGroupOutput, error) DescribeInputSecurityGroupRequest(*medialive.DescribeInputSecurityGroupInput) (*request.Request, *medialive.DescribeInputSecurityGroupOutput) + DescribeOffering(*medialive.DescribeOfferingInput) (*medialive.DescribeOfferingOutput, error) + DescribeOfferingWithContext(aws.Context, *medialive.DescribeOfferingInput, ...request.Option) (*medialive.DescribeOfferingOutput, error) + DescribeOfferingRequest(*medialive.DescribeOfferingInput) (*request.Request, *medialive.DescribeOfferingOutput) + + DescribeReservation(*medialive.DescribeReservationInput) (*medialive.DescribeReservationOutput, error) + DescribeReservationWithContext(aws.Context, *medialive.DescribeReservationInput, ...request.Option) (*medialive.DescribeReservationOutput, error) + DescribeReservationRequest(*medialive.DescribeReservationInput) (*request.Request, *medialive.DescribeReservationOutput) + ListChannels(*medialive.ListChannelsInput) (*medialive.ListChannelsOutput, error) ListChannelsWithContext(aws.Context, *medialive.ListChannelsInput, ...request.Option) (*medialive.ListChannelsOutput, error) ListChannelsRequest(*medialive.ListChannelsInput) (*request.Request, *medialive.ListChannelsOutput) @@ -117,6 +129,24 @@ type MediaLiveAPI interface { ListInputsPages(*medialive.ListInputsInput, func(*medialive.ListInputsOutput, bool) bool) error ListInputsPagesWithContext(aws.Context, *medialive.ListInputsInput, func(*medialive.ListInputsOutput, bool) bool, ...request.Option) error + ListOfferings(*medialive.ListOfferingsInput) (*medialive.ListOfferingsOutput, error) + ListOfferingsWithContext(aws.Context, *medialive.ListOfferingsInput, ...request.Option) (*medialive.ListOfferingsOutput, error) + ListOfferingsRequest(*medialive.ListOfferingsInput) (*request.Request, *medialive.ListOfferingsOutput) + + ListOfferingsPages(*medialive.ListOfferingsInput, func(*medialive.ListOfferingsOutput, bool) bool) error + ListOfferingsPagesWithContext(aws.Context, *medialive.ListOfferingsInput, func(*medialive.ListOfferingsOutput, bool) bool, ...request.Option) error + + ListReservations(*medialive.ListReservationsInput) (*medialive.ListReservationsOutput, error) + ListReservationsWithContext(aws.Context, *medialive.ListReservationsInput, ...request.Option) (*medialive.ListReservationsOutput, error) + ListReservationsRequest(*medialive.ListReservationsInput) (*request.Request, *medialive.ListReservationsOutput) + + ListReservationsPages(*medialive.ListReservationsInput, func(*medialive.ListReservationsOutput, bool) bool) error + ListReservationsPagesWithContext(aws.Context, *medialive.ListReservationsInput, func(*medialive.ListReservationsOutput, bool) bool, ...request.Option) error + + PurchaseOffering(*medialive.PurchaseOfferingInput) (*medialive.PurchaseOfferingOutput, error) + PurchaseOfferingWithContext(aws.Context, *medialive.PurchaseOfferingInput, ...request.Option) (*medialive.PurchaseOfferingOutput, error) + PurchaseOfferingRequest(*medialive.PurchaseOfferingInput) (*request.Request, *medialive.PurchaseOfferingOutput) + StartChannel(*medialive.StartChannelInput) (*medialive.StartChannelOutput, error) StartChannelWithContext(aws.Context, *medialive.StartChannelInput, ...request.Option) (*medialive.StartChannelOutput, error) StartChannelRequest(*medialive.StartChannelInput) (*request.Request, *medialive.StartChannelOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/medialive/service.go b/vendor/github.com/aws/aws-sdk-go/service/medialive/service.go index af727d2a3..582d570f1 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/medialive/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/medialive/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "medialive" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "medialive" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MediaLive" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MediaLive client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediapackage/service.go b/vendor/github.com/aws/aws-sdk-go/service/mediapackage/service.go index d5aa3cca9..96dd74dcf 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mediapackage/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mediapackage/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mediapackage" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mediapackage" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MediaPackage" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MediaPackage client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediastore/service.go b/vendor/github.com/aws/aws-sdk-go/service/mediastore/service.go index 23ce93615..492b3bf04 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mediastore/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mediastore/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mediastore" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mediastore" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MediaStore" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MediaStore client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediastoredata/service.go b/vendor/github.com/aws/aws-sdk-go/service/mediastoredata/service.go index e4116d9cf..420314014 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mediastoredata/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mediastoredata/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "data.mediastore" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "data.mediastore" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MediaStore Data" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MediaStoreData client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediatailor/api.go b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/api.go new file mode 100644 index 000000000..7f0cc57f0 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/api.go @@ -0,0 +1,851 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package mediatailor + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awsutil" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/restjson" +) + +const opDeletePlaybackConfiguration = "DeletePlaybackConfiguration" + +// DeletePlaybackConfigurationRequest generates a "aws/request.Request" representing the +// client's request for the DeletePlaybackConfiguration operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeletePlaybackConfiguration for more information on using the DeletePlaybackConfiguration +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeletePlaybackConfigurationRequest method. +// req, resp := client.DeletePlaybackConfigurationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/DeletePlaybackConfiguration +func (c *MediaTailor) DeletePlaybackConfigurationRequest(input *DeletePlaybackConfigurationInput) (req *request.Request, output *DeletePlaybackConfigurationOutput) { + op := &request.Operation{ + Name: opDeletePlaybackConfiguration, + HTTPMethod: "DELETE", + HTTPPath: "/playbackConfiguration/{Name}", + } + + if input == nil { + input = &DeletePlaybackConfigurationInput{} + } + + output = &DeletePlaybackConfigurationOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(restjson.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeletePlaybackConfiguration API operation for AWS MediaTailor. +// +// Deletes the configuration for the specified name. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS MediaTailor's +// API operation DeletePlaybackConfiguration for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/DeletePlaybackConfiguration +func (c *MediaTailor) DeletePlaybackConfiguration(input *DeletePlaybackConfigurationInput) (*DeletePlaybackConfigurationOutput, error) { + req, out := c.DeletePlaybackConfigurationRequest(input) + return out, req.Send() +} + +// DeletePlaybackConfigurationWithContext is the same as DeletePlaybackConfiguration with the addition of +// the ability to pass a context and additional request options. +// +// See DeletePlaybackConfiguration for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaTailor) DeletePlaybackConfigurationWithContext(ctx aws.Context, input *DeletePlaybackConfigurationInput, opts ...request.Option) (*DeletePlaybackConfigurationOutput, error) { + req, out := c.DeletePlaybackConfigurationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opGetPlaybackConfiguration = "GetPlaybackConfiguration" + +// GetPlaybackConfigurationRequest generates a "aws/request.Request" representing the +// client's request for the GetPlaybackConfiguration operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See GetPlaybackConfiguration for more information on using the GetPlaybackConfiguration +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the GetPlaybackConfigurationRequest method. +// req, resp := client.GetPlaybackConfigurationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/GetPlaybackConfiguration +func (c *MediaTailor) GetPlaybackConfigurationRequest(input *GetPlaybackConfigurationInput) (req *request.Request, output *GetPlaybackConfigurationOutput) { + op := &request.Operation{ + Name: opGetPlaybackConfiguration, + HTTPMethod: "GET", + HTTPPath: "/playbackConfiguration/{Name}", + } + + if input == nil { + input = &GetPlaybackConfigurationInput{} + } + + output = &GetPlaybackConfigurationOutput{} + req = c.newRequest(op, input, output) + return +} + +// GetPlaybackConfiguration API operation for AWS MediaTailor. +// +// Returns the configuration for the specified name. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS MediaTailor's +// API operation GetPlaybackConfiguration for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/GetPlaybackConfiguration +func (c *MediaTailor) GetPlaybackConfiguration(input *GetPlaybackConfigurationInput) (*GetPlaybackConfigurationOutput, error) { + req, out := c.GetPlaybackConfigurationRequest(input) + return out, req.Send() +} + +// GetPlaybackConfigurationWithContext is the same as GetPlaybackConfiguration with the addition of +// the ability to pass a context and additional request options. +// +// See GetPlaybackConfiguration for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaTailor) GetPlaybackConfigurationWithContext(ctx aws.Context, input *GetPlaybackConfigurationInput, opts ...request.Option) (*GetPlaybackConfigurationOutput, error) { + req, out := c.GetPlaybackConfigurationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opListPlaybackConfigurations = "ListPlaybackConfigurations" + +// ListPlaybackConfigurationsRequest generates a "aws/request.Request" representing the +// client's request for the ListPlaybackConfigurations operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListPlaybackConfigurations for more information on using the ListPlaybackConfigurations +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListPlaybackConfigurationsRequest method. +// req, resp := client.ListPlaybackConfigurationsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/ListPlaybackConfigurations +func (c *MediaTailor) ListPlaybackConfigurationsRequest(input *ListPlaybackConfigurationsInput) (req *request.Request, output *ListPlaybackConfigurationsOutput) { + op := &request.Operation{ + Name: opListPlaybackConfigurations, + HTTPMethod: "GET", + HTTPPath: "/playbackConfigurations", + } + + if input == nil { + input = &ListPlaybackConfigurationsInput{} + } + + output = &ListPlaybackConfigurationsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListPlaybackConfigurations API operation for AWS MediaTailor. +// +// Returns a list of the configurations defined in AWS Elemental MediaTailor. +// You can specify a max number of configurations to return at a time. The default +// max is 50. Results are returned in pagefuls. If AWS Elemental MediaTailor +// has more configurations than the specified max, it provides parameters in +// the response that you can use to retrieve the next pageful. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS MediaTailor's +// API operation ListPlaybackConfigurations for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/ListPlaybackConfigurations +func (c *MediaTailor) ListPlaybackConfigurations(input *ListPlaybackConfigurationsInput) (*ListPlaybackConfigurationsOutput, error) { + req, out := c.ListPlaybackConfigurationsRequest(input) + return out, req.Send() +} + +// ListPlaybackConfigurationsWithContext is the same as ListPlaybackConfigurations with the addition of +// the ability to pass a context and additional request options. +// +// See ListPlaybackConfigurations for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaTailor) ListPlaybackConfigurationsWithContext(ctx aws.Context, input *ListPlaybackConfigurationsInput, opts ...request.Option) (*ListPlaybackConfigurationsOutput, error) { + req, out := c.ListPlaybackConfigurationsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opPutPlaybackConfiguration = "PutPlaybackConfiguration" + +// PutPlaybackConfigurationRequest generates a "aws/request.Request" representing the +// client's request for the PutPlaybackConfiguration operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See PutPlaybackConfiguration for more information on using the PutPlaybackConfiguration +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the PutPlaybackConfigurationRequest method. +// req, resp := client.PutPlaybackConfigurationRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/PutPlaybackConfiguration +func (c *MediaTailor) PutPlaybackConfigurationRequest(input *PutPlaybackConfigurationInput) (req *request.Request, output *PutPlaybackConfigurationOutput) { + op := &request.Operation{ + Name: opPutPlaybackConfiguration, + HTTPMethod: "PUT", + HTTPPath: "/playbackConfiguration", + } + + if input == nil { + input = &PutPlaybackConfigurationInput{} + } + + output = &PutPlaybackConfigurationOutput{} + req = c.newRequest(op, input, output) + return +} + +// PutPlaybackConfiguration API operation for AWS MediaTailor. +// +// Adds a new configuration to AWS Elemental MediaTailor. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS MediaTailor's +// API operation PutPlaybackConfiguration for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23/PutPlaybackConfiguration +func (c *MediaTailor) PutPlaybackConfiguration(input *PutPlaybackConfigurationInput) (*PutPlaybackConfigurationOutput, error) { + req, out := c.PutPlaybackConfigurationRequest(input) + return out, req.Send() +} + +// PutPlaybackConfigurationWithContext is the same as PutPlaybackConfiguration with the addition of +// the ability to pass a context and additional request options. +// +// See PutPlaybackConfiguration for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *MediaTailor) PutPlaybackConfigurationWithContext(ctx aws.Context, input *PutPlaybackConfigurationInput, opts ...request.Option) (*PutPlaybackConfigurationOutput, error) { + req, out := c.PutPlaybackConfigurationRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// The configuration for using a content delivery network (CDN), like Amazon +// CloudFront, for content and ad segment management. +type CdnConfiguration struct { + _ struct{} `type:"structure"` + + // A non-default content delivery network (CDN) to serve ad segments. By default, + // AWS Elemental MediaTailor uses Amazon CloudFront with default cache settings + // as its CDN for ad segments. To set up an alternate CDN, create a rule in + // your CDN for the following origin: ads.mediatailor..amazonaws.com. + // Then specify the rule's name in this AdSegmentUrlPrefix. When AWS Elemental + // MediaTailor serves a manifest, it reports your CDN as the source for ad segments. + AdSegmentUrlPrefix *string `type:"string"` + + // A content delivery network (CDN) to cache content segments, so that content + // requests don’t always have to go to the origin server. First, create a rule + // in your CDN for the content segment origin server. Then specify the rule's + // name in this ContentSegmentUrlPrefix. When AWS Elemental MediaTailor serves + // a manifest, it reports your CDN as the source for content segments. + ContentSegmentUrlPrefix *string `type:"string"` +} + +// String returns the string representation +func (s CdnConfiguration) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CdnConfiguration) GoString() string { + return s.String() +} + +// SetAdSegmentUrlPrefix sets the AdSegmentUrlPrefix field's value. +func (s *CdnConfiguration) SetAdSegmentUrlPrefix(v string) *CdnConfiguration { + s.AdSegmentUrlPrefix = &v + return s +} + +// SetContentSegmentUrlPrefix sets the ContentSegmentUrlPrefix field's value. +func (s *CdnConfiguration) SetContentSegmentUrlPrefix(v string) *CdnConfiguration { + s.ContentSegmentUrlPrefix = &v + return s +} + +type DeletePlaybackConfigurationInput struct { + _ struct{} `type:"structure"` + + // Name is a required field + Name *string `location:"uri" locationName:"Name" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeletePlaybackConfigurationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeletePlaybackConfigurationInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeletePlaybackConfigurationInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeletePlaybackConfigurationInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *DeletePlaybackConfigurationInput) SetName(v string) *DeletePlaybackConfigurationInput { + s.Name = &v + return s +} + +type DeletePlaybackConfigurationOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeletePlaybackConfigurationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeletePlaybackConfigurationOutput) GoString() string { + return s.String() +} + +type GetPlaybackConfigurationInput struct { + _ struct{} `type:"structure"` + + // Name is a required field + Name *string `location:"uri" locationName:"Name" type:"string" required:"true"` +} + +// String returns the string representation +func (s GetPlaybackConfigurationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetPlaybackConfigurationInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *GetPlaybackConfigurationInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "GetPlaybackConfigurationInput"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *GetPlaybackConfigurationInput) SetName(v string) *GetPlaybackConfigurationInput { + s.Name = &v + return s +} + +type GetPlaybackConfigurationOutput struct { + _ struct{} `type:"structure"` + + // The URL for the ad decision server (ADS). This includes the specification + // of static parameters and placeholders for dynamic parameters. AWS Elemental + // MediaTailor substitutes player-specific and session-specific parameters as + // needed when calling the ADS. Alternately, for testing, you can provide a + // static VAST URL. The maximum length is 25000 characters. + AdDecisionServerUrl *string `type:"string"` + + // The configuration for using a content delivery network (CDN), like Amazon + // CloudFront, for content and ad segment management. + CdnConfiguration *CdnConfiguration `type:"structure"` + + // The configuration for HLS content. + HlsConfiguration *HlsConfiguration `type:"structure"` + + // The identifier for the configuration. + Name *string `type:"string"` + + // The URL that the player accesses to get a manifest from AWS Elemental MediaTailor. + // This session will use server-side reporting. + PlaybackEndpointPrefix *string `type:"string"` + + // The URL that the player uses to initialize a session that uses client-side + // reporting. + SessionInitializationEndpointPrefix *string `type:"string"` + + // URL for a high-quality video asset to transcode and use to fill in time that's + // not used by ads. AWS Elemental MediaTailor shows the slate to fill in gaps + // in media content. Configuring the slate is optional for non-VPAID configurations. + // For VPAID, the slate is required because AWS Elemental MediaTailor provides + // it in the slots designated for dynamic ad content. The slate must be a high-quality + // asset that contains both audio and video. + SlateAdUrl *string `type:"string"` + + // The URL prefix for the master playlist for the stream, minus the asset ID. + // The maximum length is 512 characters. + VideoContentSourceUrl *string `type:"string"` +} + +// String returns the string representation +func (s GetPlaybackConfigurationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetPlaybackConfigurationOutput) GoString() string { + return s.String() +} + +// SetAdDecisionServerUrl sets the AdDecisionServerUrl field's value. +func (s *GetPlaybackConfigurationOutput) SetAdDecisionServerUrl(v string) *GetPlaybackConfigurationOutput { + s.AdDecisionServerUrl = &v + return s +} + +// SetCdnConfiguration sets the CdnConfiguration field's value. +func (s *GetPlaybackConfigurationOutput) SetCdnConfiguration(v *CdnConfiguration) *GetPlaybackConfigurationOutput { + s.CdnConfiguration = v + return s +} + +// SetHlsConfiguration sets the HlsConfiguration field's value. +func (s *GetPlaybackConfigurationOutput) SetHlsConfiguration(v *HlsConfiguration) *GetPlaybackConfigurationOutput { + s.HlsConfiguration = v + return s +} + +// SetName sets the Name field's value. +func (s *GetPlaybackConfigurationOutput) SetName(v string) *GetPlaybackConfigurationOutput { + s.Name = &v + return s +} + +// SetPlaybackEndpointPrefix sets the PlaybackEndpointPrefix field's value. +func (s *GetPlaybackConfigurationOutput) SetPlaybackEndpointPrefix(v string) *GetPlaybackConfigurationOutput { + s.PlaybackEndpointPrefix = &v + return s +} + +// SetSessionInitializationEndpointPrefix sets the SessionInitializationEndpointPrefix field's value. +func (s *GetPlaybackConfigurationOutput) SetSessionInitializationEndpointPrefix(v string) *GetPlaybackConfigurationOutput { + s.SessionInitializationEndpointPrefix = &v + return s +} + +// SetSlateAdUrl sets the SlateAdUrl field's value. +func (s *GetPlaybackConfigurationOutput) SetSlateAdUrl(v string) *GetPlaybackConfigurationOutput { + s.SlateAdUrl = &v + return s +} + +// SetVideoContentSourceUrl sets the VideoContentSourceUrl field's value. +func (s *GetPlaybackConfigurationOutput) SetVideoContentSourceUrl(v string) *GetPlaybackConfigurationOutput { + s.VideoContentSourceUrl = &v + return s +} + +// The configuration for HLS content. +type HlsConfiguration struct { + _ struct{} `type:"structure"` + + // The URL that is used to initiate a playback session for devices that support + // Apple HLS. The session uses server-side reporting. + ManifestEndpointPrefix *string `type:"string"` +} + +// String returns the string representation +func (s HlsConfiguration) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HlsConfiguration) GoString() string { + return s.String() +} + +// SetManifestEndpointPrefix sets the ManifestEndpointPrefix field's value. +func (s *HlsConfiguration) SetManifestEndpointPrefix(v string) *HlsConfiguration { + s.ManifestEndpointPrefix = &v + return s +} + +type ListPlaybackConfigurationsInput struct { + _ struct{} `type:"structure"` + + MaxResults *int64 `location:"querystring" locationName:"MaxResults" min:"1" type:"integer"` + + NextToken *string `location:"querystring" locationName:"NextToken" type:"string"` +} + +// String returns the string representation +func (s ListPlaybackConfigurationsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListPlaybackConfigurationsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListPlaybackConfigurationsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListPlaybackConfigurationsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListPlaybackConfigurationsInput) SetMaxResults(v int64) *ListPlaybackConfigurationsInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListPlaybackConfigurationsInput) SetNextToken(v string) *ListPlaybackConfigurationsInput { + s.NextToken = &v + return s +} + +type ListPlaybackConfigurationsOutput struct { + _ struct{} `type:"structure"` + + // Array of playback configurations. This may be all of the available configurations + // or a subset, depending on the settings you provide and on the total number + // of configurations stored. + Items []*PlaybackConfiguration `type:"list"` + + // Pagination token returned by the GET list request when results overrun the + // meximum allowed. Use the token to fetch the next page of results. + NextToken *string `type:"string"` +} + +// String returns the string representation +func (s ListPlaybackConfigurationsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListPlaybackConfigurationsOutput) GoString() string { + return s.String() +} + +// SetItems sets the Items field's value. +func (s *ListPlaybackConfigurationsOutput) SetItems(v []*PlaybackConfiguration) *ListPlaybackConfigurationsOutput { + s.Items = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListPlaybackConfigurationsOutput) SetNextToken(v string) *ListPlaybackConfigurationsOutput { + s.NextToken = &v + return s +} + +type PlaybackConfiguration struct { + _ struct{} `type:"structure"` + + AdDecisionServerUrl *string `type:"string"` + + // The configuration for using a content delivery network (CDN), like Amazon + // CloudFront, for content and ad segment management. + CdnConfiguration *CdnConfiguration `type:"structure"` + + Name *string `type:"string"` + + SlateAdUrl *string `type:"string"` + + VideoContentSourceUrl *string `type:"string"` +} + +// String returns the string representation +func (s PlaybackConfiguration) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PlaybackConfiguration) GoString() string { + return s.String() +} + +// SetAdDecisionServerUrl sets the AdDecisionServerUrl field's value. +func (s *PlaybackConfiguration) SetAdDecisionServerUrl(v string) *PlaybackConfiguration { + s.AdDecisionServerUrl = &v + return s +} + +// SetCdnConfiguration sets the CdnConfiguration field's value. +func (s *PlaybackConfiguration) SetCdnConfiguration(v *CdnConfiguration) *PlaybackConfiguration { + s.CdnConfiguration = v + return s +} + +// SetName sets the Name field's value. +func (s *PlaybackConfiguration) SetName(v string) *PlaybackConfiguration { + s.Name = &v + return s +} + +// SetSlateAdUrl sets the SlateAdUrl field's value. +func (s *PlaybackConfiguration) SetSlateAdUrl(v string) *PlaybackConfiguration { + s.SlateAdUrl = &v + return s +} + +// SetVideoContentSourceUrl sets the VideoContentSourceUrl field's value. +func (s *PlaybackConfiguration) SetVideoContentSourceUrl(v string) *PlaybackConfiguration { + s.VideoContentSourceUrl = &v + return s +} + +type PutPlaybackConfigurationInput struct { + _ struct{} `type:"structure"` + + // The URL for the ad decision server (ADS). This includes the specification + // of static parameters and placeholders for dynamic parameters. AWS Elemental + // MediaTailor substitutes player-specific and session-specific parameters as + // needed when calling the ADS. Alternately, for testing you can provide a static + // VAST URL. The maximum length is 25000 characters. + AdDecisionServerUrl *string `type:"string"` + + // The configuration for using a content delivery network (CDN), like Amazon + // CloudFront, for content and ad segment management. + CdnConfiguration *CdnConfiguration `type:"structure"` + + // The identifier for the configuration. + Name *string `type:"string"` + + // The URL for a high-quality video asset to transcode and use to fill in time + // that's not used by ads. AWS Elemental MediaTailor shows the slate to fill + // in gaps in media content. Configuring the slate is optional for non-VPAID + // configurations. For VPAID, the slate is required because AWS Elemental MediaTailor + // provides it in the slots that are designated for dynamic ad content. The + // slate must be a high-quality asset that contains both audio and video. + SlateAdUrl *string `type:"string"` + + // The URL prefix for the master playlist for the stream, minus the asset ID. + // The maximum length is 512 characters. + VideoContentSourceUrl *string `type:"string"` +} + +// String returns the string representation +func (s PutPlaybackConfigurationInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PutPlaybackConfigurationInput) GoString() string { + return s.String() +} + +// SetAdDecisionServerUrl sets the AdDecisionServerUrl field's value. +func (s *PutPlaybackConfigurationInput) SetAdDecisionServerUrl(v string) *PutPlaybackConfigurationInput { + s.AdDecisionServerUrl = &v + return s +} + +// SetCdnConfiguration sets the CdnConfiguration field's value. +func (s *PutPlaybackConfigurationInput) SetCdnConfiguration(v *CdnConfiguration) *PutPlaybackConfigurationInput { + s.CdnConfiguration = v + return s +} + +// SetName sets the Name field's value. +func (s *PutPlaybackConfigurationInput) SetName(v string) *PutPlaybackConfigurationInput { + s.Name = &v + return s +} + +// SetSlateAdUrl sets the SlateAdUrl field's value. +func (s *PutPlaybackConfigurationInput) SetSlateAdUrl(v string) *PutPlaybackConfigurationInput { + s.SlateAdUrl = &v + return s +} + +// SetVideoContentSourceUrl sets the VideoContentSourceUrl field's value. +func (s *PutPlaybackConfigurationInput) SetVideoContentSourceUrl(v string) *PutPlaybackConfigurationInput { + s.VideoContentSourceUrl = &v + return s +} + +type PutPlaybackConfigurationOutput struct { + _ struct{} `type:"structure"` + + AdDecisionServerUrl *string `type:"string"` + + // The configuration for using a content delivery network (CDN), like Amazon + // CloudFront, for content and ad segment management. + CdnConfiguration *CdnConfiguration `type:"structure"` + + // The configuration for HLS content. + HlsConfiguration *HlsConfiguration `type:"structure"` + + Name *string `type:"string"` + + PlaybackEndpointPrefix *string `type:"string"` + + SessionInitializationEndpointPrefix *string `type:"string"` + + SlateAdUrl *string `type:"string"` + + VideoContentSourceUrl *string `type:"string"` +} + +// String returns the string representation +func (s PutPlaybackConfigurationOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PutPlaybackConfigurationOutput) GoString() string { + return s.String() +} + +// SetAdDecisionServerUrl sets the AdDecisionServerUrl field's value. +func (s *PutPlaybackConfigurationOutput) SetAdDecisionServerUrl(v string) *PutPlaybackConfigurationOutput { + s.AdDecisionServerUrl = &v + return s +} + +// SetCdnConfiguration sets the CdnConfiguration field's value. +func (s *PutPlaybackConfigurationOutput) SetCdnConfiguration(v *CdnConfiguration) *PutPlaybackConfigurationOutput { + s.CdnConfiguration = v + return s +} + +// SetHlsConfiguration sets the HlsConfiguration field's value. +func (s *PutPlaybackConfigurationOutput) SetHlsConfiguration(v *HlsConfiguration) *PutPlaybackConfigurationOutput { + s.HlsConfiguration = v + return s +} + +// SetName sets the Name field's value. +func (s *PutPlaybackConfigurationOutput) SetName(v string) *PutPlaybackConfigurationOutput { + s.Name = &v + return s +} + +// SetPlaybackEndpointPrefix sets the PlaybackEndpointPrefix field's value. +func (s *PutPlaybackConfigurationOutput) SetPlaybackEndpointPrefix(v string) *PutPlaybackConfigurationOutput { + s.PlaybackEndpointPrefix = &v + return s +} + +// SetSessionInitializationEndpointPrefix sets the SessionInitializationEndpointPrefix field's value. +func (s *PutPlaybackConfigurationOutput) SetSessionInitializationEndpointPrefix(v string) *PutPlaybackConfigurationOutput { + s.SessionInitializationEndpointPrefix = &v + return s +} + +// SetSlateAdUrl sets the SlateAdUrl field's value. +func (s *PutPlaybackConfigurationOutput) SetSlateAdUrl(v string) *PutPlaybackConfigurationOutput { + s.SlateAdUrl = &v + return s +} + +// SetVideoContentSourceUrl sets the VideoContentSourceUrl field's value. +func (s *PutPlaybackConfigurationOutput) SetVideoContentSourceUrl(v string) *PutPlaybackConfigurationOutput { + s.VideoContentSourceUrl = &v + return s +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediatailor/doc.go b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/doc.go new file mode 100644 index 000000000..50eaf1e03 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/doc.go @@ -0,0 +1,36 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package mediatailor provides the client and types for making API +// requests to AWS MediaTailor. +// +// Use the AWS Elemental MediaTailor SDK to configure scalable ad insertion +// for your live and VOD content. With AWS Elemental MediaTailor, you can serve +// targeted ads to viewers while maintaining broadcast quality in over-the-top +// (OTT) video applications. For information about using the service, including +// detailed information about the settings covered in this guide, see the AWS +// Elemental MediaTailor User Guide.Through the SDK, you manage AWS Elemental +// MediaTailor configurations the same as you do through the console. For example, +// you specify ad insertion behavior and mapping information for the origin +// server and the ad decision server (ADS). +// +// See https://docs.aws.amazon.com/goto/WebAPI/mediatailor-2018-04-23 for more information on this service. +// +// See mediatailor package documentation for more information. +// https://docs.aws.amazon.com/sdk-for-go/api/service/mediatailor/ +// +// Using the Client +// +// To contact AWS MediaTailor with the SDK use the New function to create +// a new service client. With that client you can make API requests to the service. +// These clients are safe to use concurrently. +// +// See the SDK's documentation for more information on how to use the SDK. +// https://docs.aws.amazon.com/sdk-for-go/api/ +// +// See aws.Config documentation for more information on configuring SDK clients. +// https://docs.aws.amazon.com/sdk-for-go/api/aws/#Config +// +// See the AWS MediaTailor client MediaTailor for more +// information on creating client for this service. +// https://docs.aws.amazon.com/sdk-for-go/api/service/mediatailor/#New +package mediatailor diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediatailor/errors.go b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/errors.go new file mode 100644 index 000000000..49d22eb13 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/errors.go @@ -0,0 +1,3 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package mediatailor diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediatailor/mediatailoriface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/mediatailoriface/interface.go new file mode 100644 index 000000000..53fec2655 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/mediatailoriface/interface.go @@ -0,0 +1,80 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package mediatailoriface provides an interface to enable mocking the AWS MediaTailor service client +// for testing your code. +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. +package mediatailoriface + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/service/mediatailor" +) + +// MediaTailorAPI provides an interface to enable mocking the +// mediatailor.MediaTailor service client's API operation, +// paginators, and waiters. This make unit testing your code that calls out +// to the SDK's service client's calls easier. +// +// The best way to use this interface is so the SDK's service client's calls +// can be stubbed out for unit testing your code with the SDK without needing +// to inject custom request handlers into the SDK's request pipeline. +// +// // myFunc uses an SDK service client to make a request to +// // AWS MediaTailor. +// func myFunc(svc mediatailoriface.MediaTailorAPI) bool { +// // Make svc.DeletePlaybackConfiguration request +// } +// +// func main() { +// sess := session.New() +// svc := mediatailor.New(sess) +// +// myFunc(svc) +// } +// +// In your _test.go file: +// +// // Define a mock struct to be used in your unit tests of myFunc. +// type mockMediaTailorClient struct { +// mediatailoriface.MediaTailorAPI +// } +// func (m *mockMediaTailorClient) DeletePlaybackConfiguration(input *mediatailor.DeletePlaybackConfigurationInput) (*mediatailor.DeletePlaybackConfigurationOutput, error) { +// // mock response/functionality +// } +// +// func TestMyFunc(t *testing.T) { +// // Setup Test +// mockSvc := &mockMediaTailorClient{} +// +// myfunc(mockSvc) +// +// // Verify myFunc's functionality +// } +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. Its suggested to use the pattern above for testing, or using +// tooling to generate mocks to satisfy the interfaces. +type MediaTailorAPI interface { + DeletePlaybackConfiguration(*mediatailor.DeletePlaybackConfigurationInput) (*mediatailor.DeletePlaybackConfigurationOutput, error) + DeletePlaybackConfigurationWithContext(aws.Context, *mediatailor.DeletePlaybackConfigurationInput, ...request.Option) (*mediatailor.DeletePlaybackConfigurationOutput, error) + DeletePlaybackConfigurationRequest(*mediatailor.DeletePlaybackConfigurationInput) (*request.Request, *mediatailor.DeletePlaybackConfigurationOutput) + + GetPlaybackConfiguration(*mediatailor.GetPlaybackConfigurationInput) (*mediatailor.GetPlaybackConfigurationOutput, error) + GetPlaybackConfigurationWithContext(aws.Context, *mediatailor.GetPlaybackConfigurationInput, ...request.Option) (*mediatailor.GetPlaybackConfigurationOutput, error) + GetPlaybackConfigurationRequest(*mediatailor.GetPlaybackConfigurationInput) (*request.Request, *mediatailor.GetPlaybackConfigurationOutput) + + ListPlaybackConfigurations(*mediatailor.ListPlaybackConfigurationsInput) (*mediatailor.ListPlaybackConfigurationsOutput, error) + ListPlaybackConfigurationsWithContext(aws.Context, *mediatailor.ListPlaybackConfigurationsInput, ...request.Option) (*mediatailor.ListPlaybackConfigurationsOutput, error) + ListPlaybackConfigurationsRequest(*mediatailor.ListPlaybackConfigurationsInput) (*request.Request, *mediatailor.ListPlaybackConfigurationsOutput) + + PutPlaybackConfiguration(*mediatailor.PutPlaybackConfigurationInput) (*mediatailor.PutPlaybackConfigurationOutput, error) + PutPlaybackConfigurationWithContext(aws.Context, *mediatailor.PutPlaybackConfigurationInput, ...request.Option) (*mediatailor.PutPlaybackConfigurationOutput, error) + PutPlaybackConfigurationRequest(*mediatailor.PutPlaybackConfigurationInput) (*request.Request, *mediatailor.PutPlaybackConfigurationOutput) +} + +var _ MediaTailorAPI = (*mediatailor.MediaTailor)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/mediatailor/service.go b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/service.go new file mode 100644 index 000000000..a3471868a --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/mediatailor/service.go @@ -0,0 +1,99 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package mediatailor + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/signer/v4" + "github.com/aws/aws-sdk-go/private/protocol/restjson" +) + +// MediaTailor provides the API operation methods for making requests to +// AWS MediaTailor. See this package's package overview docs +// for details on the service. +// +// MediaTailor methods are safe to use concurrently. It is not safe to +// modify mutate any of the struct's properties though. +type MediaTailor struct { + *client.Client +} + +// Used for custom client initialization logic +var initClient func(*client.Client) + +// Used for custom request initialization logic +var initRequest func(*request.Request) + +// Service information constants +const ( + ServiceName = "api.mediatailor" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MediaTailor" // ServiceID is a unique identifer of a specific service. +) + +// New creates a new instance of the MediaTailor client with a session. +// If additional configuration is needed for the client instance use the optional +// aws.Config parameter to add your extra config. +// +// Example: +// // Create a MediaTailor client from just a session. +// svc := mediatailor.New(mySession) +// +// // Create a MediaTailor client with additional configuration +// svc := mediatailor.New(mySession, aws.NewConfig().WithRegion("us-west-2")) +func New(p client.ConfigProvider, cfgs ...*aws.Config) *MediaTailor { + c := p.ClientConfig(EndpointsID, cfgs...) + if c.SigningNameDerived || len(c.SigningName) == 0 { + c.SigningName = "mediatailor" + } + return newClient(*c.Config, c.Handlers, c.Endpoint, c.SigningRegion, c.SigningName) +} + +// newClient creates, initializes and returns a new service client instance. +func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion, signingName string) *MediaTailor { + svc := &MediaTailor{ + Client: client.New( + cfg, + metadata.ClientInfo{ + ServiceName: ServiceName, + ServiceID: ServiceID, + SigningName: signingName, + SigningRegion: signingRegion, + Endpoint: endpoint, + APIVersion: "2018-04-23", + JSONVersion: "1.1", + }, + handlers, + ), + } + + // Handlers + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(restjson.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(restjson.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(restjson.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(restjson.UnmarshalErrorHandler) + + // Run custom client initialization if present + if initClient != nil { + initClient(svc.Client) + } + + return svc +} + +// newRequest creates a new request for a MediaTailor operation and runs any +// custom request initialization. +func (c *MediaTailor) newRequest(op *request.Operation, params, data interface{}) *request.Request { + req := c.NewRequest(op, params, data) + + // Run custom request initialization if present + if initRequest != nil { + initRequest(req) + } + + return req +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/migrationhub/api.go b/vendor/github.com/aws/aws-sdk-go/service/migrationhub/api.go index 552987760..7c9f057c8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/migrationhub/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/migrationhub/api.go @@ -1673,6 +1673,7 @@ func (c *MigrationHub) PutResourceAttributesRequest(input *PutResourceAttributes // address. // // Note the instructions regarding the special use case of the ResourceAttributeList +// (https://docs.aws.amazon.com/migrationhub/latest/ug/API_PutResourceAttributes.html#migrationhub-PutResourceAttributes-request-ResourceAttributeList) // parameter when specifying any "VM" related value. // // Because this is an asynchronous call, it will always return 200, whether @@ -3399,17 +3400,22 @@ type PutResourceAttributesInput struct { // used to map the task to a resource in the Application Discovery Service (ADS)'s // repository. // - // In the ResourceAttribute object array, the Type field is reserved for the - // following values: IPV4_ADDRESS | IPV6_ADDRESS | MAC_ADDRESS | FQDN | VM_MANAGER_ID - // | VM_MANAGED_OBJECT_REFERENCE | VM_NAME | VM_PATH | BIOS_ID | MOTHERBOARD_SERIAL_NUMBER, - // and the identifying value can be a string up to 256 characters. + // Takes the object array of ResourceAttribute where the Type field is reserved + // for the following values: IPV4_ADDRESS | IPV6_ADDRESS | MAC_ADDRESS | FQDN + // | VM_MANAGER_ID | VM_MANAGED_OBJECT_REFERENCE | VM_NAME | VM_PATH | BIOS_ID + // | MOTHERBOARD_SERIAL_NUMBER where the identifying value can be a string up + // to 256 characters. // - // If any "VM" related value is used for a ResourceAttribute object, it is required - // that VM_MANAGER_ID, as a minimum, is always used. If it is not used, the - // server will not be associated in the Application Discovery Service (ADS)'s - // repository using any of the other "VM" related values, and you will experience - // data loss. See the Example section below for a use case of specifying "VM" - // related values. + // If any "VM" related value is set for a ResourceAttribute object, it is required + // that VM_MANAGER_ID, as a minimum, is always set. If VM_MANAGER_ID is not + // set, then all "VM" fields will be discarded and "VM" fields will not be used + // for matching the migration task to a server in Application Discovery Service + // (ADS)'s repository. See the Example (https://docs.aws.amazon.com/migrationhub/latest/ug/API_PutResourceAttributes.html#API_PutResourceAttributes_Examples) + // section below for a use case of specifying "VM" related values. + // + // If a server you are trying to match has multiple IP or MAC addresses, you + // should provide as many as you know in separate type/value pairs passed to + // the ResourceAttributeList parameter to maximize the chances of matching. // // ResourceAttributeList is a required field ResourceAttributeList []*ResourceAttribute `min:"1" type:"list" required:"true"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/migrationhub/service.go b/vendor/github.com/aws/aws-sdk-go/service/migrationhub/service.go index e7d767fbe..eb23ca0e8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/migrationhub/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/migrationhub/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mgh" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mgh" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Migration Hub" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MigrationHub client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mobile/service.go b/vendor/github.com/aws/aws-sdk-go/service/mobile/service.go index 20f342817..e41d7e337 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mobile/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mobile/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mobile" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mobile" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Mobile" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Mobile client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mobileanalytics/service.go b/vendor/github.com/aws/aws-sdk-go/service/mobileanalytics/service.go index bfdcea030..92d6f4c90 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mobileanalytics/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mobileanalytics/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mobileanalytics" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mobileanalytics" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Mobile Analytics" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MobileAnalytics client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mq/service.go b/vendor/github.com/aws/aws-sdk-go/service/mq/service.go index 455e3622f..1ebb5aeca 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mq/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mq/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mq" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mq" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "mq" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MQ client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/mturk/service.go b/vendor/github.com/aws/aws-sdk-go/service/mturk/service.go index 6823e2fb3..ddd789add 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/mturk/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/mturk/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "mturk-requester" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "mturk-requester" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "MTurk" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the MTurk client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/neptune/api.go b/vendor/github.com/aws/aws-sdk-go/service/neptune/api.go new file mode 100644 index 000000000..ecf6f2b28 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/neptune/api.go @@ -0,0 +1,16693 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package neptune + +import ( + "fmt" + "time" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awsutil" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/query" +) + +const opAddRoleToDBCluster = "AddRoleToDBCluster" + +// AddRoleToDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the AddRoleToDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AddRoleToDBCluster for more information on using the AddRoleToDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AddRoleToDBClusterRequest method. +// req, resp := client.AddRoleToDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/AddRoleToDBCluster +func (c *Neptune) AddRoleToDBClusterRequest(input *AddRoleToDBClusterInput) (req *request.Request, output *AddRoleToDBClusterOutput) { + op := &request.Operation{ + Name: opAddRoleToDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AddRoleToDBClusterInput{} + } + + output = &AddRoleToDBClusterOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// AddRoleToDBCluster API operation for Amazon Neptune. +// +// Associates an Identity and Access Management (IAM) role from an Neptune DB +// cluster. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation AddRoleToDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeDBClusterRoleAlreadyExistsFault "DBClusterRoleAlreadyExists" +// The specified IAM role Amazon Resource Name (ARN) is already associated with +// the specified DB cluster. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeDBClusterRoleQuotaExceededFault "DBClusterRoleQuotaExceeded" +// You have exceeded the maximum number of IAM roles that can be associated +// with the specified DB cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/AddRoleToDBCluster +func (c *Neptune) AddRoleToDBCluster(input *AddRoleToDBClusterInput) (*AddRoleToDBClusterOutput, error) { + req, out := c.AddRoleToDBClusterRequest(input) + return out, req.Send() +} + +// AddRoleToDBClusterWithContext is the same as AddRoleToDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See AddRoleToDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) AddRoleToDBClusterWithContext(ctx aws.Context, input *AddRoleToDBClusterInput, opts ...request.Option) (*AddRoleToDBClusterOutput, error) { + req, out := c.AddRoleToDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opAddSourceIdentifierToSubscription = "AddSourceIdentifierToSubscription" + +// AddSourceIdentifierToSubscriptionRequest generates a "aws/request.Request" representing the +// client's request for the AddSourceIdentifierToSubscription operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AddSourceIdentifierToSubscription for more information on using the AddSourceIdentifierToSubscription +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AddSourceIdentifierToSubscriptionRequest method. +// req, resp := client.AddSourceIdentifierToSubscriptionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/AddSourceIdentifierToSubscription +func (c *Neptune) AddSourceIdentifierToSubscriptionRequest(input *AddSourceIdentifierToSubscriptionInput) (req *request.Request, output *AddSourceIdentifierToSubscriptionOutput) { + op := &request.Operation{ + Name: opAddSourceIdentifierToSubscription, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AddSourceIdentifierToSubscriptionInput{} + } + + output = &AddSourceIdentifierToSubscriptionOutput{} + req = c.newRequest(op, input, output) + return +} + +// AddSourceIdentifierToSubscription API operation for Amazon Neptune. +// +// Adds a source identifier to an existing event notification subscription. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation AddSourceIdentifierToSubscription for usage and error information. +// +// Returned Error Codes: +// * ErrCodeSubscriptionNotFoundFault "SubscriptionNotFound" +// +// * ErrCodeSourceNotFoundFault "SourceNotFound" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/AddSourceIdentifierToSubscription +func (c *Neptune) AddSourceIdentifierToSubscription(input *AddSourceIdentifierToSubscriptionInput) (*AddSourceIdentifierToSubscriptionOutput, error) { + req, out := c.AddSourceIdentifierToSubscriptionRequest(input) + return out, req.Send() +} + +// AddSourceIdentifierToSubscriptionWithContext is the same as AddSourceIdentifierToSubscription with the addition of +// the ability to pass a context and additional request options. +// +// See AddSourceIdentifierToSubscription for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) AddSourceIdentifierToSubscriptionWithContext(ctx aws.Context, input *AddSourceIdentifierToSubscriptionInput, opts ...request.Option) (*AddSourceIdentifierToSubscriptionOutput, error) { + req, out := c.AddSourceIdentifierToSubscriptionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opAddTagsToResource = "AddTagsToResource" + +// AddTagsToResourceRequest generates a "aws/request.Request" representing the +// client's request for the AddTagsToResource operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AddTagsToResource for more information on using the AddTagsToResource +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AddTagsToResourceRequest method. +// req, resp := client.AddTagsToResourceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/AddTagsToResource +func (c *Neptune) AddTagsToResourceRequest(input *AddTagsToResourceInput) (req *request.Request, output *AddTagsToResourceOutput) { + op := &request.Operation{ + Name: opAddTagsToResource, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AddTagsToResourceInput{} + } + + output = &AddTagsToResourceOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// AddTagsToResource API operation for Amazon Neptune. +// +// Adds metadata tags to an Amazon Neptune resource. These tags can also be +// used with cost allocation reporting to track cost associated with Amazon +// Neptune resources, or used in a Condition statement in an IAM policy for +// Amazon Neptune. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation AddTagsToResource for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeDBSnapshotNotFoundFault "DBSnapshotNotFound" +// DBSnapshotIdentifier does not refer to an existing DB snapshot. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/AddTagsToResource +func (c *Neptune) AddTagsToResource(input *AddTagsToResourceInput) (*AddTagsToResourceOutput, error) { + req, out := c.AddTagsToResourceRequest(input) + return out, req.Send() +} + +// AddTagsToResourceWithContext is the same as AddTagsToResource with the addition of +// the ability to pass a context and additional request options. +// +// See AddTagsToResource for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) AddTagsToResourceWithContext(ctx aws.Context, input *AddTagsToResourceInput, opts ...request.Option) (*AddTagsToResourceOutput, error) { + req, out := c.AddTagsToResourceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opApplyPendingMaintenanceAction = "ApplyPendingMaintenanceAction" + +// ApplyPendingMaintenanceActionRequest generates a "aws/request.Request" representing the +// client's request for the ApplyPendingMaintenanceAction operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ApplyPendingMaintenanceAction for more information on using the ApplyPendingMaintenanceAction +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ApplyPendingMaintenanceActionRequest method. +// req, resp := client.ApplyPendingMaintenanceActionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ApplyPendingMaintenanceAction +func (c *Neptune) ApplyPendingMaintenanceActionRequest(input *ApplyPendingMaintenanceActionInput) (req *request.Request, output *ApplyPendingMaintenanceActionOutput) { + op := &request.Operation{ + Name: opApplyPendingMaintenanceAction, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ApplyPendingMaintenanceActionInput{} + } + + output = &ApplyPendingMaintenanceActionOutput{} + req = c.newRequest(op, input, output) + return +} + +// ApplyPendingMaintenanceAction API operation for Amazon Neptune. +// +// Applies a pending maintenance action to a resource (for example, to a DB +// instance). +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ApplyPendingMaintenanceAction for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFoundFault "ResourceNotFoundFault" +// The specified resource ID was not found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ApplyPendingMaintenanceAction +func (c *Neptune) ApplyPendingMaintenanceAction(input *ApplyPendingMaintenanceActionInput) (*ApplyPendingMaintenanceActionOutput, error) { + req, out := c.ApplyPendingMaintenanceActionRequest(input) + return out, req.Send() +} + +// ApplyPendingMaintenanceActionWithContext is the same as ApplyPendingMaintenanceAction with the addition of +// the ability to pass a context and additional request options. +// +// See ApplyPendingMaintenanceAction for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ApplyPendingMaintenanceActionWithContext(ctx aws.Context, input *ApplyPendingMaintenanceActionInput, opts ...request.Option) (*ApplyPendingMaintenanceActionOutput, error) { + req, out := c.ApplyPendingMaintenanceActionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCopyDBClusterParameterGroup = "CopyDBClusterParameterGroup" + +// CopyDBClusterParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the CopyDBClusterParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CopyDBClusterParameterGroup for more information on using the CopyDBClusterParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CopyDBClusterParameterGroupRequest method. +// req, resp := client.CopyDBClusterParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CopyDBClusterParameterGroup +func (c *Neptune) CopyDBClusterParameterGroupRequest(input *CopyDBClusterParameterGroupInput) (req *request.Request, output *CopyDBClusterParameterGroupOutput) { + op := &request.Operation{ + Name: opCopyDBClusterParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CopyDBClusterParameterGroupInput{} + } + + output = &CopyDBClusterParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// CopyDBClusterParameterGroup API operation for Amazon Neptune. +// +// Copies the specified DB cluster parameter group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CopyDBClusterParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// * ErrCodeDBParameterGroupQuotaExceededFault "DBParameterGroupQuotaExceeded" +// Request would result in user exceeding the allowed number of DB parameter +// groups. +// +// * ErrCodeDBParameterGroupAlreadyExistsFault "DBParameterGroupAlreadyExists" +// A DB parameter group with the same name exists. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CopyDBClusterParameterGroup +func (c *Neptune) CopyDBClusterParameterGroup(input *CopyDBClusterParameterGroupInput) (*CopyDBClusterParameterGroupOutput, error) { + req, out := c.CopyDBClusterParameterGroupRequest(input) + return out, req.Send() +} + +// CopyDBClusterParameterGroupWithContext is the same as CopyDBClusterParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See CopyDBClusterParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CopyDBClusterParameterGroupWithContext(ctx aws.Context, input *CopyDBClusterParameterGroupInput, opts ...request.Option) (*CopyDBClusterParameterGroupOutput, error) { + req, out := c.CopyDBClusterParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCopyDBClusterSnapshot = "CopyDBClusterSnapshot" + +// CopyDBClusterSnapshotRequest generates a "aws/request.Request" representing the +// client's request for the CopyDBClusterSnapshot operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CopyDBClusterSnapshot for more information on using the CopyDBClusterSnapshot +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CopyDBClusterSnapshotRequest method. +// req, resp := client.CopyDBClusterSnapshotRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CopyDBClusterSnapshot +func (c *Neptune) CopyDBClusterSnapshotRequest(input *CopyDBClusterSnapshotInput) (req *request.Request, output *CopyDBClusterSnapshotOutput) { + op := &request.Operation{ + Name: opCopyDBClusterSnapshot, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CopyDBClusterSnapshotInput{} + } + + output = &CopyDBClusterSnapshotOutput{} + req = c.newRequest(op, input, output) + return +} + +// CopyDBClusterSnapshot API operation for Amazon Neptune. +// +// Copies a snapshot of a DB cluster. +// +// To copy a DB cluster snapshot from a shared manual DB cluster snapshot, SourceDBClusterSnapshotIdentifier +// must be the Amazon Resource Name (ARN) of the shared DB cluster snapshot. +// +// You can copy an encrypted DB cluster snapshot from another AWS Region. In +// that case, the AWS Region where you call the CopyDBClusterSnapshot action +// is the destination AWS Region for the encrypted DB cluster snapshot to be +// copied to. To copy an encrypted DB cluster snapshot from another AWS Region, +// you must provide the following values: +// +// * KmsKeyId - The AWS Key Management System (AWS KMS) key identifier for +// the key to use to encrypt the copy of the DB cluster snapshot in the destination +// AWS Region. +// +// * PreSignedUrl - A URL that contains a Signature Version 4 signed request +// for the CopyDBClusterSnapshot action to be called in the source AWS Region +// where the DB cluster snapshot is copied from. The pre-signed URL must +// be a valid request for the CopyDBClusterSnapshot API action that can be +// executed in the source AWS Region that contains the encrypted DB cluster +// snapshot to be copied. +// +// The pre-signed URL request must contain the following parameter values: +// +// KmsKeyId - The KMS key identifier for the key to use to encrypt the copy +// of the DB cluster snapshot in the destination AWS Region. This is the +// same identifier for both the CopyDBClusterSnapshot action that is called +// in the destination AWS Region, and the action contained in the pre-signed +// URL. +// +// DestinationRegion - The name of the AWS Region that the DB cluster snapshot +// will be created in. +// +// SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier for +// the encrypted DB cluster snapshot to be copied. This identifier must be +// in the Amazon Resource Name (ARN) format for the source AWS Region. For +// example, if you are copying an encrypted DB cluster snapshot from the +// us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier looks +// like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:neptune-cluster1-snapshot-20161115. +// +// To learn how to generate a Signature Version 4 signed request, see Authenticating +// Requests: Using Query Parameters (AWS Signature Version 4) (http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html) +// and Signature Version 4 Signing Process (http://docs.aws.amazon.com/general/latest/gr/signature-version-4.html). +// +// * TargetDBClusterSnapshotIdentifier - The identifier for the new copy +// of the DB cluster snapshot in the destination AWS Region. +// +// * SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier +// for the encrypted DB cluster snapshot to be copied. This identifier must +// be in the ARN format for the source AWS Region and is the same value as +// the SourceDBClusterSnapshotIdentifier in the pre-signed URL. +// +// To cancel the copy operation once it is in progress, delete the target DB +// cluster snapshot identified by TargetDBClusterSnapshotIdentifier while that +// DB cluster snapshot is in "copying" status. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CopyDBClusterSnapshot for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterSnapshotAlreadyExistsFault "DBClusterSnapshotAlreadyExistsFault" +// User already has a DB cluster snapshot with the given identifier. +// +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// * ErrCodeSnapshotQuotaExceededFault "SnapshotQuotaExceeded" +// Request would result in user exceeding the allowed number of DB snapshots. +// +// * ErrCodeKMSKeyNotAccessibleFault "KMSKeyNotAccessibleFault" +// Error accessing KMS key. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CopyDBClusterSnapshot +func (c *Neptune) CopyDBClusterSnapshot(input *CopyDBClusterSnapshotInput) (*CopyDBClusterSnapshotOutput, error) { + req, out := c.CopyDBClusterSnapshotRequest(input) + return out, req.Send() +} + +// CopyDBClusterSnapshotWithContext is the same as CopyDBClusterSnapshot with the addition of +// the ability to pass a context and additional request options. +// +// See CopyDBClusterSnapshot for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CopyDBClusterSnapshotWithContext(ctx aws.Context, input *CopyDBClusterSnapshotInput, opts ...request.Option) (*CopyDBClusterSnapshotOutput, error) { + req, out := c.CopyDBClusterSnapshotRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCopyDBParameterGroup = "CopyDBParameterGroup" + +// CopyDBParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the CopyDBParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CopyDBParameterGroup for more information on using the CopyDBParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CopyDBParameterGroupRequest method. +// req, resp := client.CopyDBParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CopyDBParameterGroup +func (c *Neptune) CopyDBParameterGroupRequest(input *CopyDBParameterGroupInput) (req *request.Request, output *CopyDBParameterGroupOutput) { + op := &request.Operation{ + Name: opCopyDBParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CopyDBParameterGroupInput{} + } + + output = &CopyDBParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// CopyDBParameterGroup API operation for Amazon Neptune. +// +// Copies the specified DB parameter group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CopyDBParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// * ErrCodeDBParameterGroupAlreadyExistsFault "DBParameterGroupAlreadyExists" +// A DB parameter group with the same name exists. +// +// * ErrCodeDBParameterGroupQuotaExceededFault "DBParameterGroupQuotaExceeded" +// Request would result in user exceeding the allowed number of DB parameter +// groups. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CopyDBParameterGroup +func (c *Neptune) CopyDBParameterGroup(input *CopyDBParameterGroupInput) (*CopyDBParameterGroupOutput, error) { + req, out := c.CopyDBParameterGroupRequest(input) + return out, req.Send() +} + +// CopyDBParameterGroupWithContext is the same as CopyDBParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See CopyDBParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CopyDBParameterGroupWithContext(ctx aws.Context, input *CopyDBParameterGroupInput, opts ...request.Option) (*CopyDBParameterGroupOutput, error) { + req, out := c.CopyDBParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateDBCluster = "CreateDBCluster" + +// CreateDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the CreateDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateDBCluster for more information on using the CreateDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateDBClusterRequest method. +// req, resp := client.CreateDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBCluster +func (c *Neptune) CreateDBClusterRequest(input *CreateDBClusterInput) (req *request.Request, output *CreateDBClusterOutput) { + op := &request.Operation{ + Name: opCreateDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateDBClusterInput{} + } + + output = &CreateDBClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateDBCluster API operation for Amazon Neptune. +// +// Creates a new Amazon Neptune DB cluster. +// +// You can use the ReplicationSourceIdentifier parameter to create the DB cluster +// as a Read Replica of another DB cluster or Amazon Neptune DB instance. For +// cross-region replication where the DB cluster identified by ReplicationSourceIdentifier +// is encrypted, you must also specify the PreSignedUrl parameter. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterAlreadyExistsFault "DBClusterAlreadyExistsFault" +// User already has a DB cluster with the given identifier. +// +// * ErrCodeInsufficientStorageClusterCapacityFault "InsufficientStorageClusterCapacity" +// There is insufficient storage available for the current action. You may be +// able to resolve this error by updating your subnet group to use different +// Availability Zones that have more storage available. +// +// * ErrCodeDBClusterQuotaExceededFault "DBClusterQuotaExceededFault" +// User attempted to create a new DB cluster and the user has already reached +// the maximum allowed DB cluster quota. +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeInvalidVPCNetworkStateFault "InvalidVPCNetworkStateFault" +// DB subnet group does not cover all Availability Zones after it is created +// because users' change. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeInvalidDBSubnetGroupStateFault "InvalidDBSubnetGroupStateFault" +// The DB subnet group cannot be deleted because it is in use. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// * ErrCodeDBClusterParameterGroupNotFoundFault "DBClusterParameterGroupNotFound" +// DBClusterParameterGroupName does not refer to an existing DB Cluster parameter +// group. +// +// * ErrCodeKMSKeyNotAccessibleFault "KMSKeyNotAccessibleFault" +// Error accessing KMS key. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeDBSubnetGroupDoesNotCoverEnoughAZs "DBSubnetGroupDoesNotCoverEnoughAZs" +// Subnets in the DB subnet group should cover at least two Availability Zones +// unless there is only one Availability Zone. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBCluster +func (c *Neptune) CreateDBCluster(input *CreateDBClusterInput) (*CreateDBClusterOutput, error) { + req, out := c.CreateDBClusterRequest(input) + return out, req.Send() +} + +// CreateDBClusterWithContext is the same as CreateDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See CreateDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateDBClusterWithContext(ctx aws.Context, input *CreateDBClusterInput, opts ...request.Option) (*CreateDBClusterOutput, error) { + req, out := c.CreateDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateDBClusterParameterGroup = "CreateDBClusterParameterGroup" + +// CreateDBClusterParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the CreateDBClusterParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateDBClusterParameterGroup for more information on using the CreateDBClusterParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateDBClusterParameterGroupRequest method. +// req, resp := client.CreateDBClusterParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBClusterParameterGroup +func (c *Neptune) CreateDBClusterParameterGroupRequest(input *CreateDBClusterParameterGroupInput) (req *request.Request, output *CreateDBClusterParameterGroupOutput) { + op := &request.Operation{ + Name: opCreateDBClusterParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateDBClusterParameterGroupInput{} + } + + output = &CreateDBClusterParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateDBClusterParameterGroup API operation for Amazon Neptune. +// +// Creates a new DB cluster parameter group. +// +// Parameters in a DB cluster parameter group apply to all of the instances +// in a DB cluster. +// +// A DB cluster parameter group is initially created with the default parameters +// for the database engine used by instances in the DB cluster. To provide custom +// values for any of the parameters, you must modify the group after creating +// it using ModifyDBClusterParameterGroup. Once you've created a DB cluster +// parameter group, you need to associate it with your DB cluster using ModifyDBCluster. +// When you associate a new DB cluster parameter group with a running DB cluster, +// you need to reboot the DB instances in the DB cluster without failover for +// the new DB cluster parameter group and associated settings to take effect. +// +// After you create a DB cluster parameter group, you should wait at least 5 +// minutes before creating your first DB cluster that uses that DB cluster parameter +// group as the default parameter group. This allows Amazon Neptune to fully +// complete the create action before the DB cluster parameter group is used +// as the default for a new DB cluster. This is especially important for parameters +// that are critical when creating the default database for a DB cluster, such +// as the character set for the default database defined by the character_set_database +// parameter. You can use the Parameter Groups option of the Amazon Neptune +// console (https://console.aws.amazon.com/rds/) or the DescribeDBClusterParameters +// command to verify that your DB cluster parameter group has been created or +// modified. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateDBClusterParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupQuotaExceededFault "DBParameterGroupQuotaExceeded" +// Request would result in user exceeding the allowed number of DB parameter +// groups. +// +// * ErrCodeDBParameterGroupAlreadyExistsFault "DBParameterGroupAlreadyExists" +// A DB parameter group with the same name exists. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBClusterParameterGroup +func (c *Neptune) CreateDBClusterParameterGroup(input *CreateDBClusterParameterGroupInput) (*CreateDBClusterParameterGroupOutput, error) { + req, out := c.CreateDBClusterParameterGroupRequest(input) + return out, req.Send() +} + +// CreateDBClusterParameterGroupWithContext is the same as CreateDBClusterParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See CreateDBClusterParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateDBClusterParameterGroupWithContext(ctx aws.Context, input *CreateDBClusterParameterGroupInput, opts ...request.Option) (*CreateDBClusterParameterGroupOutput, error) { + req, out := c.CreateDBClusterParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateDBClusterSnapshot = "CreateDBClusterSnapshot" + +// CreateDBClusterSnapshotRequest generates a "aws/request.Request" representing the +// client's request for the CreateDBClusterSnapshot operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateDBClusterSnapshot for more information on using the CreateDBClusterSnapshot +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateDBClusterSnapshotRequest method. +// req, resp := client.CreateDBClusterSnapshotRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBClusterSnapshot +func (c *Neptune) CreateDBClusterSnapshotRequest(input *CreateDBClusterSnapshotInput) (req *request.Request, output *CreateDBClusterSnapshotOutput) { + op := &request.Operation{ + Name: opCreateDBClusterSnapshot, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateDBClusterSnapshotInput{} + } + + output = &CreateDBClusterSnapshotOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateDBClusterSnapshot API operation for Amazon Neptune. +// +// Creates a snapshot of a DB cluster. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateDBClusterSnapshot for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterSnapshotAlreadyExistsFault "DBClusterSnapshotAlreadyExistsFault" +// User already has a DB cluster snapshot with the given identifier. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeSnapshotQuotaExceededFault "SnapshotQuotaExceeded" +// Request would result in user exceeding the allowed number of DB snapshots. +// +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBClusterSnapshot +func (c *Neptune) CreateDBClusterSnapshot(input *CreateDBClusterSnapshotInput) (*CreateDBClusterSnapshotOutput, error) { + req, out := c.CreateDBClusterSnapshotRequest(input) + return out, req.Send() +} + +// CreateDBClusterSnapshotWithContext is the same as CreateDBClusterSnapshot with the addition of +// the ability to pass a context and additional request options. +// +// See CreateDBClusterSnapshot for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateDBClusterSnapshotWithContext(ctx aws.Context, input *CreateDBClusterSnapshotInput, opts ...request.Option) (*CreateDBClusterSnapshotOutput, error) { + req, out := c.CreateDBClusterSnapshotRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateDBInstance = "CreateDBInstance" + +// CreateDBInstanceRequest generates a "aws/request.Request" representing the +// client's request for the CreateDBInstance operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateDBInstance for more information on using the CreateDBInstance +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateDBInstanceRequest method. +// req, resp := client.CreateDBInstanceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBInstance +func (c *Neptune) CreateDBInstanceRequest(input *CreateDBInstanceInput) (req *request.Request, output *CreateDBInstanceOutput) { + op := &request.Operation{ + Name: opCreateDBInstance, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateDBInstanceInput{} + } + + output = &CreateDBInstanceOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateDBInstance API operation for Amazon Neptune. +// +// Creates a new DB instance. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateDBInstance for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceAlreadyExistsFault "DBInstanceAlreadyExists" +// User already has a DB instance with the given identifier. +// +// * ErrCodeInsufficientDBInstanceCapacityFault "InsufficientDBInstanceCapacity" +// Specified DB instance class is not available in the specified Availability +// Zone. +// +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// * ErrCodeDBSecurityGroupNotFoundFault "DBSecurityGroupNotFound" +// DBSecurityGroupName does not refer to an existing DB security group. +// +// * ErrCodeInstanceQuotaExceededFault "InstanceQuotaExceeded" +// Request would result in user exceeding the allowed number of DB instances. +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeDBSubnetGroupDoesNotCoverEnoughAZs "DBSubnetGroupDoesNotCoverEnoughAZs" +// Subnets in the DB subnet group should cover at least two Availability Zones +// unless there is only one Availability Zone. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// * ErrCodeInvalidVPCNetworkStateFault "InvalidVPCNetworkStateFault" +// DB subnet group does not cover all Availability Zones after it is created +// because users' change. +// +// * ErrCodeProvisionedIopsNotAvailableInAZFault "ProvisionedIopsNotAvailableInAZFault" +// Provisioned IOPS not available in the specified Availability Zone. +// +// * ErrCodeOptionGroupNotFoundFault "OptionGroupNotFoundFault" +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeStorageTypeNotSupportedFault "StorageTypeNotSupported" +// StorageType specified cannot be associated with the DB Instance. +// +// * ErrCodeAuthorizationNotFoundFault "AuthorizationNotFound" +// Specified CIDRIP or EC2 security group is not authorized for the specified +// DB security group. +// +// Neptune may not also be authorized via IAM to perform necessary actions on +// your behalf. +// +// * ErrCodeKMSKeyNotAccessibleFault "KMSKeyNotAccessibleFault" +// Error accessing KMS key. +// +// * ErrCodeDomainNotFoundFault "DomainNotFoundFault" +// Domain does not refer to an existing Active Directory Domain. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBInstance +func (c *Neptune) CreateDBInstance(input *CreateDBInstanceInput) (*CreateDBInstanceOutput, error) { + req, out := c.CreateDBInstanceRequest(input) + return out, req.Send() +} + +// CreateDBInstanceWithContext is the same as CreateDBInstance with the addition of +// the ability to pass a context and additional request options. +// +// See CreateDBInstance for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateDBInstanceWithContext(ctx aws.Context, input *CreateDBInstanceInput, opts ...request.Option) (*CreateDBInstanceOutput, error) { + req, out := c.CreateDBInstanceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateDBParameterGroup = "CreateDBParameterGroup" + +// CreateDBParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the CreateDBParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateDBParameterGroup for more information on using the CreateDBParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateDBParameterGroupRequest method. +// req, resp := client.CreateDBParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBParameterGroup +func (c *Neptune) CreateDBParameterGroupRequest(input *CreateDBParameterGroupInput) (req *request.Request, output *CreateDBParameterGroupOutput) { + op := &request.Operation{ + Name: opCreateDBParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateDBParameterGroupInput{} + } + + output = &CreateDBParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateDBParameterGroup API operation for Amazon Neptune. +// +// Creates a new DB parameter group. +// +// A DB parameter group is initially created with the default parameters for +// the database engine used by the DB instance. To provide custom values for +// any of the parameters, you must modify the group after creating it using +// ModifyDBParameterGroup. Once you've created a DB parameter group, you need +// to associate it with your DB instance using ModifyDBInstance. When you associate +// a new DB parameter group with a running DB instance, you need to reboot the +// DB instance without failover for the new DB parameter group and associated +// settings to take effect. +// +// After you create a DB parameter group, you should wait at least 5 minutes +// before creating your first DB instance that uses that DB parameter group +// as the default parameter group. This allows Amazon Neptune to fully complete +// the create action before the parameter group is used as the default for a +// new DB instance. This is especially important for parameters that are critical +// when creating the default database for a DB instance, such as the character +// set for the default database defined by the character_set_database parameter. +// You can use the Parameter Groups option of the Amazon Neptune console or +// the DescribeDBParameters command to verify that your DB parameter group has +// been created or modified. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateDBParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupQuotaExceededFault "DBParameterGroupQuotaExceeded" +// Request would result in user exceeding the allowed number of DB parameter +// groups. +// +// * ErrCodeDBParameterGroupAlreadyExistsFault "DBParameterGroupAlreadyExists" +// A DB parameter group with the same name exists. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBParameterGroup +func (c *Neptune) CreateDBParameterGroup(input *CreateDBParameterGroupInput) (*CreateDBParameterGroupOutput, error) { + req, out := c.CreateDBParameterGroupRequest(input) + return out, req.Send() +} + +// CreateDBParameterGroupWithContext is the same as CreateDBParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See CreateDBParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateDBParameterGroupWithContext(ctx aws.Context, input *CreateDBParameterGroupInput, opts ...request.Option) (*CreateDBParameterGroupOutput, error) { + req, out := c.CreateDBParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateDBSubnetGroup = "CreateDBSubnetGroup" + +// CreateDBSubnetGroupRequest generates a "aws/request.Request" representing the +// client's request for the CreateDBSubnetGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateDBSubnetGroup for more information on using the CreateDBSubnetGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateDBSubnetGroupRequest method. +// req, resp := client.CreateDBSubnetGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBSubnetGroup +func (c *Neptune) CreateDBSubnetGroupRequest(input *CreateDBSubnetGroupInput) (req *request.Request, output *CreateDBSubnetGroupOutput) { + op := &request.Operation{ + Name: opCreateDBSubnetGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateDBSubnetGroupInput{} + } + + output = &CreateDBSubnetGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateDBSubnetGroup API operation for Amazon Neptune. +// +// Creates a new DB subnet group. DB subnet groups must contain at least one +// subnet in at least two AZs in the AWS Region. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateDBSubnetGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBSubnetGroupAlreadyExistsFault "DBSubnetGroupAlreadyExists" +// DBSubnetGroupName is already used by an existing DB subnet group. +// +// * ErrCodeDBSubnetGroupQuotaExceededFault "DBSubnetGroupQuotaExceeded" +// Request would result in user exceeding the allowed number of DB subnet groups. +// +// * ErrCodeDBSubnetQuotaExceededFault "DBSubnetQuotaExceededFault" +// Request would result in user exceeding the allowed number of subnets in a +// DB subnet groups. +// +// * ErrCodeDBSubnetGroupDoesNotCoverEnoughAZs "DBSubnetGroupDoesNotCoverEnoughAZs" +// Subnets in the DB subnet group should cover at least two Availability Zones +// unless there is only one Availability Zone. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateDBSubnetGroup +func (c *Neptune) CreateDBSubnetGroup(input *CreateDBSubnetGroupInput) (*CreateDBSubnetGroupOutput, error) { + req, out := c.CreateDBSubnetGroupRequest(input) + return out, req.Send() +} + +// CreateDBSubnetGroupWithContext is the same as CreateDBSubnetGroup with the addition of +// the ability to pass a context and additional request options. +// +// See CreateDBSubnetGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateDBSubnetGroupWithContext(ctx aws.Context, input *CreateDBSubnetGroupInput, opts ...request.Option) (*CreateDBSubnetGroupOutput, error) { + req, out := c.CreateDBSubnetGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opCreateEventSubscription = "CreateEventSubscription" + +// CreateEventSubscriptionRequest generates a "aws/request.Request" representing the +// client's request for the CreateEventSubscription operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateEventSubscription for more information on using the CreateEventSubscription +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateEventSubscriptionRequest method. +// req, resp := client.CreateEventSubscriptionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateEventSubscription +func (c *Neptune) CreateEventSubscriptionRequest(input *CreateEventSubscriptionInput) (req *request.Request, output *CreateEventSubscriptionOutput) { + op := &request.Operation{ + Name: opCreateEventSubscription, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateEventSubscriptionInput{} + } + + output = &CreateEventSubscriptionOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateEventSubscription API operation for Amazon Neptune. +// +// Creates an event notification subscription. This action requires a topic +// ARN (Amazon Resource Name) created by either the Neptune console, the SNS +// console, or the SNS API. To obtain an ARN with SNS, you must create a topic +// in Amazon SNS and subscribe to the topic. The ARN is displayed in the SNS +// console. +// +// You can specify the type of source (SourceType) you want to be notified of, +// provide a list of Neptune sources (SourceIds) that triggers the events, and +// provide a list of event categories (EventCategories) for events you want +// to be notified of. For example, you can specify SourceType = db-instance, +// SourceIds = mydbinstance1, mydbinstance2 and EventCategories = Availability, +// Backup. +// +// If you specify both the SourceType and SourceIds, such as SourceType = db-instance +// and SourceIdentifier = myDBInstance1, you are notified of all the db-instance +// events for the specified source. If you specify a SourceType but do not specify +// a SourceIdentifier, you receive notice of the events for that source type +// for all your Neptune sources. If you do not specify either the SourceType +// nor the SourceIdentifier, you are notified of events generated from all Neptune +// sources belonging to your customer account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation CreateEventSubscription for usage and error information. +// +// Returned Error Codes: +// * ErrCodeEventSubscriptionQuotaExceededFault "EventSubscriptionQuotaExceeded" +// +// * ErrCodeSubscriptionAlreadyExistFault "SubscriptionAlreadyExist" +// +// * ErrCodeSNSInvalidTopicFault "SNSInvalidTopic" +// +// * ErrCodeSNSNoAuthorizationFault "SNSNoAuthorization" +// +// * ErrCodeSNSTopicArnNotFoundFault "SNSTopicArnNotFound" +// +// * ErrCodeSubscriptionCategoryNotFoundFault "SubscriptionCategoryNotFound" +// +// * ErrCodeSourceNotFoundFault "SourceNotFound" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/CreateEventSubscription +func (c *Neptune) CreateEventSubscription(input *CreateEventSubscriptionInput) (*CreateEventSubscriptionOutput, error) { + req, out := c.CreateEventSubscriptionRequest(input) + return out, req.Send() +} + +// CreateEventSubscriptionWithContext is the same as CreateEventSubscription with the addition of +// the ability to pass a context and additional request options. +// +// See CreateEventSubscription for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) CreateEventSubscriptionWithContext(ctx aws.Context, input *CreateEventSubscriptionInput, opts ...request.Option) (*CreateEventSubscriptionOutput, error) { + req, out := c.CreateEventSubscriptionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteDBCluster = "DeleteDBCluster" + +// DeleteDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the DeleteDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteDBCluster for more information on using the DeleteDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteDBClusterRequest method. +// req, resp := client.DeleteDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBCluster +func (c *Neptune) DeleteDBClusterRequest(input *DeleteDBClusterInput) (req *request.Request, output *DeleteDBClusterOutput) { + op := &request.Operation{ + Name: opDeleteDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteDBClusterInput{} + } + + output = &DeleteDBClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteDBCluster API operation for Amazon Neptune. +// +// The DeleteDBCluster action deletes a previously provisioned DB cluster. When +// you delete a DB cluster, all automated backups for that DB cluster are deleted +// and can't be recovered. Manual DB cluster snapshots of the specified DB cluster +// are not deleted. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeDBClusterSnapshotAlreadyExistsFault "DBClusterSnapshotAlreadyExistsFault" +// User already has a DB cluster snapshot with the given identifier. +// +// * ErrCodeSnapshotQuotaExceededFault "SnapshotQuotaExceeded" +// Request would result in user exceeding the allowed number of DB snapshots. +// +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBCluster +func (c *Neptune) DeleteDBCluster(input *DeleteDBClusterInput) (*DeleteDBClusterOutput, error) { + req, out := c.DeleteDBClusterRequest(input) + return out, req.Send() +} + +// DeleteDBClusterWithContext is the same as DeleteDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteDBClusterWithContext(ctx aws.Context, input *DeleteDBClusterInput, opts ...request.Option) (*DeleteDBClusterOutput, error) { + req, out := c.DeleteDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteDBClusterParameterGroup = "DeleteDBClusterParameterGroup" + +// DeleteDBClusterParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the DeleteDBClusterParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteDBClusterParameterGroup for more information on using the DeleteDBClusterParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteDBClusterParameterGroupRequest method. +// req, resp := client.DeleteDBClusterParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBClusterParameterGroup +func (c *Neptune) DeleteDBClusterParameterGroupRequest(input *DeleteDBClusterParameterGroupInput) (req *request.Request, output *DeleteDBClusterParameterGroupOutput) { + op := &request.Operation{ + Name: opDeleteDBClusterParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteDBClusterParameterGroupInput{} + } + + output = &DeleteDBClusterParameterGroupOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeleteDBClusterParameterGroup API operation for Amazon Neptune. +// +// Deletes a specified DB cluster parameter group. The DB cluster parameter +// group to be deleted can't be associated with any DB clusters. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteDBClusterParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBParameterGroupStateFault "InvalidDBParameterGroupState" +// The DB parameter group is in use or is in an invalid state. If you are attempting +// to delete the parameter group, you cannot delete it when the parameter group +// is in this state. +// +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBClusterParameterGroup +func (c *Neptune) DeleteDBClusterParameterGroup(input *DeleteDBClusterParameterGroupInput) (*DeleteDBClusterParameterGroupOutput, error) { + req, out := c.DeleteDBClusterParameterGroupRequest(input) + return out, req.Send() +} + +// DeleteDBClusterParameterGroupWithContext is the same as DeleteDBClusterParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteDBClusterParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteDBClusterParameterGroupWithContext(ctx aws.Context, input *DeleteDBClusterParameterGroupInput, opts ...request.Option) (*DeleteDBClusterParameterGroupOutput, error) { + req, out := c.DeleteDBClusterParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteDBClusterSnapshot = "DeleteDBClusterSnapshot" + +// DeleteDBClusterSnapshotRequest generates a "aws/request.Request" representing the +// client's request for the DeleteDBClusterSnapshot operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteDBClusterSnapshot for more information on using the DeleteDBClusterSnapshot +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteDBClusterSnapshotRequest method. +// req, resp := client.DeleteDBClusterSnapshotRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBClusterSnapshot +func (c *Neptune) DeleteDBClusterSnapshotRequest(input *DeleteDBClusterSnapshotInput) (req *request.Request, output *DeleteDBClusterSnapshotOutput) { + op := &request.Operation{ + Name: opDeleteDBClusterSnapshot, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteDBClusterSnapshotInput{} + } + + output = &DeleteDBClusterSnapshotOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteDBClusterSnapshot API operation for Amazon Neptune. +// +// Deletes a DB cluster snapshot. If the snapshot is being copied, the copy +// operation is terminated. +// +// The DB cluster snapshot must be in the available state to be deleted. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteDBClusterSnapshot for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBClusterSnapshot +func (c *Neptune) DeleteDBClusterSnapshot(input *DeleteDBClusterSnapshotInput) (*DeleteDBClusterSnapshotOutput, error) { + req, out := c.DeleteDBClusterSnapshotRequest(input) + return out, req.Send() +} + +// DeleteDBClusterSnapshotWithContext is the same as DeleteDBClusterSnapshot with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteDBClusterSnapshot for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteDBClusterSnapshotWithContext(ctx aws.Context, input *DeleteDBClusterSnapshotInput, opts ...request.Option) (*DeleteDBClusterSnapshotOutput, error) { + req, out := c.DeleteDBClusterSnapshotRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteDBInstance = "DeleteDBInstance" + +// DeleteDBInstanceRequest generates a "aws/request.Request" representing the +// client's request for the DeleteDBInstance operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteDBInstance for more information on using the DeleteDBInstance +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteDBInstanceRequest method. +// req, resp := client.DeleteDBInstanceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBInstance +func (c *Neptune) DeleteDBInstanceRequest(input *DeleteDBInstanceInput) (req *request.Request, output *DeleteDBInstanceOutput) { + op := &request.Operation{ + Name: opDeleteDBInstance, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteDBInstanceInput{} + } + + output = &DeleteDBInstanceOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteDBInstance API operation for Amazon Neptune. +// +// The DeleteDBInstance action deletes a previously provisioned DB instance. +// When you delete a DB instance, all automated backups for that instance are +// deleted and can't be recovered. Manual DB snapshots of the DB instance to +// be deleted by DeleteDBInstance are not deleted. +// +// If you request a final DB snapshot the status of the Amazon Neptune DB instance +// is deleting until the DB snapshot is created. The API action DescribeDBInstance +// is used to monitor the status of this operation. The action can't be canceled +// or reverted once submitted. +// +// Note that when a DB instance is in a failure state and has a status of failed, +// incompatible-restore, or incompatible-network, you can only delete it when +// the SkipFinalSnapshot parameter is set to true. +// +// If the specified DB instance is part of a DB cluster, you can't delete the +// DB instance if both of the following conditions are true: +// +// * The DB cluster is a Read Replica of another DB cluster. +// +// * The DB instance is the only instance in the DB cluster. +// +// To delete a DB instance in this case, first call the PromoteReadReplicaDBCluster +// API action to promote the DB cluster so it's no longer a Read Replica. After +// the promotion completes, then call the DeleteDBInstance API action to delete +// the final instance in the DB cluster. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteDBInstance for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// * ErrCodeDBSnapshotAlreadyExistsFault "DBSnapshotAlreadyExists" +// DBSnapshotIdentifier is already used by an existing snapshot. +// +// * ErrCodeSnapshotQuotaExceededFault "SnapshotQuotaExceeded" +// Request would result in user exceeding the allowed number of DB snapshots. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBInstance +func (c *Neptune) DeleteDBInstance(input *DeleteDBInstanceInput) (*DeleteDBInstanceOutput, error) { + req, out := c.DeleteDBInstanceRequest(input) + return out, req.Send() +} + +// DeleteDBInstanceWithContext is the same as DeleteDBInstance with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteDBInstance for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteDBInstanceWithContext(ctx aws.Context, input *DeleteDBInstanceInput, opts ...request.Option) (*DeleteDBInstanceOutput, error) { + req, out := c.DeleteDBInstanceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteDBParameterGroup = "DeleteDBParameterGroup" + +// DeleteDBParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the DeleteDBParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteDBParameterGroup for more information on using the DeleteDBParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteDBParameterGroupRequest method. +// req, resp := client.DeleteDBParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBParameterGroup +func (c *Neptune) DeleteDBParameterGroupRequest(input *DeleteDBParameterGroupInput) (req *request.Request, output *DeleteDBParameterGroupOutput) { + op := &request.Operation{ + Name: opDeleteDBParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteDBParameterGroupInput{} + } + + output = &DeleteDBParameterGroupOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeleteDBParameterGroup API operation for Amazon Neptune. +// +// Deletes a specified DBParameterGroup. The DBParameterGroup to be deleted +// can't be associated with any DB instances. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteDBParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBParameterGroupStateFault "InvalidDBParameterGroupState" +// The DB parameter group is in use or is in an invalid state. If you are attempting +// to delete the parameter group, you cannot delete it when the parameter group +// is in this state. +// +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBParameterGroup +func (c *Neptune) DeleteDBParameterGroup(input *DeleteDBParameterGroupInput) (*DeleteDBParameterGroupOutput, error) { + req, out := c.DeleteDBParameterGroupRequest(input) + return out, req.Send() +} + +// DeleteDBParameterGroupWithContext is the same as DeleteDBParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteDBParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteDBParameterGroupWithContext(ctx aws.Context, input *DeleteDBParameterGroupInput, opts ...request.Option) (*DeleteDBParameterGroupOutput, error) { + req, out := c.DeleteDBParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteDBSubnetGroup = "DeleteDBSubnetGroup" + +// DeleteDBSubnetGroupRequest generates a "aws/request.Request" representing the +// client's request for the DeleteDBSubnetGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteDBSubnetGroup for more information on using the DeleteDBSubnetGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteDBSubnetGroupRequest method. +// req, resp := client.DeleteDBSubnetGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBSubnetGroup +func (c *Neptune) DeleteDBSubnetGroupRequest(input *DeleteDBSubnetGroupInput) (req *request.Request, output *DeleteDBSubnetGroupOutput) { + op := &request.Operation{ + Name: opDeleteDBSubnetGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteDBSubnetGroupInput{} + } + + output = &DeleteDBSubnetGroupOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// DeleteDBSubnetGroup API operation for Amazon Neptune. +// +// Deletes a DB subnet group. +// +// The specified database subnet group must not be associated with any DB instances. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteDBSubnetGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBSubnetGroupStateFault "InvalidDBSubnetGroupStateFault" +// The DB subnet group cannot be deleted because it is in use. +// +// * ErrCodeInvalidDBSubnetStateFault "InvalidDBSubnetStateFault" +// The DB subnet is not in the available state. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteDBSubnetGroup +func (c *Neptune) DeleteDBSubnetGroup(input *DeleteDBSubnetGroupInput) (*DeleteDBSubnetGroupOutput, error) { + req, out := c.DeleteDBSubnetGroupRequest(input) + return out, req.Send() +} + +// DeleteDBSubnetGroupWithContext is the same as DeleteDBSubnetGroup with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteDBSubnetGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteDBSubnetGroupWithContext(ctx aws.Context, input *DeleteDBSubnetGroupInput, opts ...request.Option) (*DeleteDBSubnetGroupOutput, error) { + req, out := c.DeleteDBSubnetGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDeleteEventSubscription = "DeleteEventSubscription" + +// DeleteEventSubscriptionRequest generates a "aws/request.Request" representing the +// client's request for the DeleteEventSubscription operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteEventSubscription for more information on using the DeleteEventSubscription +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteEventSubscriptionRequest method. +// req, resp := client.DeleteEventSubscriptionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteEventSubscription +func (c *Neptune) DeleteEventSubscriptionRequest(input *DeleteEventSubscriptionInput) (req *request.Request, output *DeleteEventSubscriptionOutput) { + op := &request.Operation{ + Name: opDeleteEventSubscription, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteEventSubscriptionInput{} + } + + output = &DeleteEventSubscriptionOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteEventSubscription API operation for Amazon Neptune. +// +// Deletes an event notification subscription. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DeleteEventSubscription for usage and error information. +// +// Returned Error Codes: +// * ErrCodeSubscriptionNotFoundFault "SubscriptionNotFound" +// +// * ErrCodeInvalidEventSubscriptionStateFault "InvalidEventSubscriptionState" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DeleteEventSubscription +func (c *Neptune) DeleteEventSubscription(input *DeleteEventSubscriptionInput) (*DeleteEventSubscriptionOutput, error) { + req, out := c.DeleteEventSubscriptionRequest(input) + return out, req.Send() +} + +// DeleteEventSubscriptionWithContext is the same as DeleteEventSubscription with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteEventSubscription for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DeleteEventSubscriptionWithContext(ctx aws.Context, input *DeleteEventSubscriptionInput, opts ...request.Option) (*DeleteEventSubscriptionOutput, error) { + req, out := c.DeleteEventSubscriptionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeDBClusterParameterGroups = "DescribeDBClusterParameterGroups" + +// DescribeDBClusterParameterGroupsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBClusterParameterGroups operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBClusterParameterGroups for more information on using the DescribeDBClusterParameterGroups +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBClusterParameterGroupsRequest method. +// req, resp := client.DescribeDBClusterParameterGroupsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterParameterGroups +func (c *Neptune) DescribeDBClusterParameterGroupsRequest(input *DescribeDBClusterParameterGroupsInput) (req *request.Request, output *DescribeDBClusterParameterGroupsOutput) { + op := &request.Operation{ + Name: opDescribeDBClusterParameterGroups, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDBClusterParameterGroupsInput{} + } + + output = &DescribeDBClusterParameterGroupsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBClusterParameterGroups API operation for Amazon Neptune. +// +// Returns a list of DBClusterParameterGroup descriptions. If a DBClusterParameterGroupName +// parameter is specified, the list will contain only the description of the +// specified DB cluster parameter group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBClusterParameterGroups for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterParameterGroups +func (c *Neptune) DescribeDBClusterParameterGroups(input *DescribeDBClusterParameterGroupsInput) (*DescribeDBClusterParameterGroupsOutput, error) { + req, out := c.DescribeDBClusterParameterGroupsRequest(input) + return out, req.Send() +} + +// DescribeDBClusterParameterGroupsWithContext is the same as DescribeDBClusterParameterGroups with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBClusterParameterGroups for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBClusterParameterGroupsWithContext(ctx aws.Context, input *DescribeDBClusterParameterGroupsInput, opts ...request.Option) (*DescribeDBClusterParameterGroupsOutput, error) { + req, out := c.DescribeDBClusterParameterGroupsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeDBClusterParameters = "DescribeDBClusterParameters" + +// DescribeDBClusterParametersRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBClusterParameters operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBClusterParameters for more information on using the DescribeDBClusterParameters +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBClusterParametersRequest method. +// req, resp := client.DescribeDBClusterParametersRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterParameters +func (c *Neptune) DescribeDBClusterParametersRequest(input *DescribeDBClusterParametersInput) (req *request.Request, output *DescribeDBClusterParametersOutput) { + op := &request.Operation{ + Name: opDescribeDBClusterParameters, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDBClusterParametersInput{} + } + + output = &DescribeDBClusterParametersOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBClusterParameters API operation for Amazon Neptune. +// +// Returns the detailed parameter list for a particular DB cluster parameter +// group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBClusterParameters for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterParameters +func (c *Neptune) DescribeDBClusterParameters(input *DescribeDBClusterParametersInput) (*DescribeDBClusterParametersOutput, error) { + req, out := c.DescribeDBClusterParametersRequest(input) + return out, req.Send() +} + +// DescribeDBClusterParametersWithContext is the same as DescribeDBClusterParameters with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBClusterParameters for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBClusterParametersWithContext(ctx aws.Context, input *DescribeDBClusterParametersInput, opts ...request.Option) (*DescribeDBClusterParametersOutput, error) { + req, out := c.DescribeDBClusterParametersRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeDBClusterSnapshotAttributes = "DescribeDBClusterSnapshotAttributes" + +// DescribeDBClusterSnapshotAttributesRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBClusterSnapshotAttributes operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBClusterSnapshotAttributes for more information on using the DescribeDBClusterSnapshotAttributes +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBClusterSnapshotAttributesRequest method. +// req, resp := client.DescribeDBClusterSnapshotAttributesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterSnapshotAttributes +func (c *Neptune) DescribeDBClusterSnapshotAttributesRequest(input *DescribeDBClusterSnapshotAttributesInput) (req *request.Request, output *DescribeDBClusterSnapshotAttributesOutput) { + op := &request.Operation{ + Name: opDescribeDBClusterSnapshotAttributes, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDBClusterSnapshotAttributesInput{} + } + + output = &DescribeDBClusterSnapshotAttributesOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBClusterSnapshotAttributes API operation for Amazon Neptune. +// +// Returns a list of DB cluster snapshot attribute names and values for a manual +// DB cluster snapshot. +// +// When sharing snapshots with other AWS accounts, DescribeDBClusterSnapshotAttributes +// returns the restore attribute and a list of IDs for the AWS accounts that +// are authorized to copy or restore the manual DB cluster snapshot. If all +// is included in the list of values for the restore attribute, then the manual +// DB cluster snapshot is public and can be copied or restored by all AWS accounts. +// +// To add or remove access for an AWS account to copy or restore a manual DB +// cluster snapshot, or to make the manual DB cluster snapshot public or private, +// use the ModifyDBClusterSnapshotAttribute API action. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBClusterSnapshotAttributes for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterSnapshotAttributes +func (c *Neptune) DescribeDBClusterSnapshotAttributes(input *DescribeDBClusterSnapshotAttributesInput) (*DescribeDBClusterSnapshotAttributesOutput, error) { + req, out := c.DescribeDBClusterSnapshotAttributesRequest(input) + return out, req.Send() +} + +// DescribeDBClusterSnapshotAttributesWithContext is the same as DescribeDBClusterSnapshotAttributes with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBClusterSnapshotAttributes for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBClusterSnapshotAttributesWithContext(ctx aws.Context, input *DescribeDBClusterSnapshotAttributesInput, opts ...request.Option) (*DescribeDBClusterSnapshotAttributesOutput, error) { + req, out := c.DescribeDBClusterSnapshotAttributesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeDBClusterSnapshots = "DescribeDBClusterSnapshots" + +// DescribeDBClusterSnapshotsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBClusterSnapshots operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBClusterSnapshots for more information on using the DescribeDBClusterSnapshots +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBClusterSnapshotsRequest method. +// req, resp := client.DescribeDBClusterSnapshotsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterSnapshots +func (c *Neptune) DescribeDBClusterSnapshotsRequest(input *DescribeDBClusterSnapshotsInput) (req *request.Request, output *DescribeDBClusterSnapshotsOutput) { + op := &request.Operation{ + Name: opDescribeDBClusterSnapshots, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDBClusterSnapshotsInput{} + } + + output = &DescribeDBClusterSnapshotsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBClusterSnapshots API operation for Amazon Neptune. +// +// Returns information about DB cluster snapshots. This API action supports +// pagination. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBClusterSnapshots for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusterSnapshots +func (c *Neptune) DescribeDBClusterSnapshots(input *DescribeDBClusterSnapshotsInput) (*DescribeDBClusterSnapshotsOutput, error) { + req, out := c.DescribeDBClusterSnapshotsRequest(input) + return out, req.Send() +} + +// DescribeDBClusterSnapshotsWithContext is the same as DescribeDBClusterSnapshots with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBClusterSnapshots for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBClusterSnapshotsWithContext(ctx aws.Context, input *DescribeDBClusterSnapshotsInput, opts ...request.Option) (*DescribeDBClusterSnapshotsOutput, error) { + req, out := c.DescribeDBClusterSnapshotsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeDBClusters = "DescribeDBClusters" + +// DescribeDBClustersRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBClusters operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBClusters for more information on using the DescribeDBClusters +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBClustersRequest method. +// req, resp := client.DescribeDBClustersRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusters +func (c *Neptune) DescribeDBClustersRequest(input *DescribeDBClustersInput) (req *request.Request, output *DescribeDBClustersOutput) { + op := &request.Operation{ + Name: opDescribeDBClusters, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDBClustersInput{} + } + + output = &DescribeDBClustersOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBClusters API operation for Amazon Neptune. +// +// Returns information about provisioned DB clusters. This API supports pagination. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBClusters for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBClusters +func (c *Neptune) DescribeDBClusters(input *DescribeDBClustersInput) (*DescribeDBClustersOutput, error) { + req, out := c.DescribeDBClustersRequest(input) + return out, req.Send() +} + +// DescribeDBClustersWithContext is the same as DescribeDBClusters with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBClusters for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBClustersWithContext(ctx aws.Context, input *DescribeDBClustersInput, opts ...request.Option) (*DescribeDBClustersOutput, error) { + req, out := c.DescribeDBClustersRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeDBEngineVersions = "DescribeDBEngineVersions" + +// DescribeDBEngineVersionsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBEngineVersions operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBEngineVersions for more information on using the DescribeDBEngineVersions +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBEngineVersionsRequest method. +// req, resp := client.DescribeDBEngineVersionsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBEngineVersions +func (c *Neptune) DescribeDBEngineVersionsRequest(input *DescribeDBEngineVersionsInput) (req *request.Request, output *DescribeDBEngineVersionsOutput) { + op := &request.Operation{ + Name: opDescribeDBEngineVersions, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeDBEngineVersionsInput{} + } + + output = &DescribeDBEngineVersionsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBEngineVersions API operation for Amazon Neptune. +// +// Returns a list of the available DB engines. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBEngineVersions for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBEngineVersions +func (c *Neptune) DescribeDBEngineVersions(input *DescribeDBEngineVersionsInput) (*DescribeDBEngineVersionsOutput, error) { + req, out := c.DescribeDBEngineVersionsRequest(input) + return out, req.Send() +} + +// DescribeDBEngineVersionsWithContext is the same as DescribeDBEngineVersions with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBEngineVersions for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBEngineVersionsWithContext(ctx aws.Context, input *DescribeDBEngineVersionsInput, opts ...request.Option) (*DescribeDBEngineVersionsOutput, error) { + req, out := c.DescribeDBEngineVersionsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeDBEngineVersionsPages iterates over the pages of a DescribeDBEngineVersions operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeDBEngineVersions method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeDBEngineVersions operation. +// pageNum := 0 +// err := client.DescribeDBEngineVersionsPages(params, +// func(page *DescribeDBEngineVersionsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeDBEngineVersionsPages(input *DescribeDBEngineVersionsInput, fn func(*DescribeDBEngineVersionsOutput, bool) bool) error { + return c.DescribeDBEngineVersionsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeDBEngineVersionsPagesWithContext same as DescribeDBEngineVersionsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBEngineVersionsPagesWithContext(ctx aws.Context, input *DescribeDBEngineVersionsInput, fn func(*DescribeDBEngineVersionsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeDBEngineVersionsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBEngineVersionsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeDBEngineVersionsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeDBInstances = "DescribeDBInstances" + +// DescribeDBInstancesRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBInstances operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBInstances for more information on using the DescribeDBInstances +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBInstancesRequest method. +// req, resp := client.DescribeDBInstancesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBInstances +func (c *Neptune) DescribeDBInstancesRequest(input *DescribeDBInstancesInput) (req *request.Request, output *DescribeDBInstancesOutput) { + op := &request.Operation{ + Name: opDescribeDBInstances, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeDBInstancesInput{} + } + + output = &DescribeDBInstancesOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBInstances API operation for Amazon Neptune. +// +// Returns information about provisioned instances. This API supports pagination. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBInstances for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBInstances +func (c *Neptune) DescribeDBInstances(input *DescribeDBInstancesInput) (*DescribeDBInstancesOutput, error) { + req, out := c.DescribeDBInstancesRequest(input) + return out, req.Send() +} + +// DescribeDBInstancesWithContext is the same as DescribeDBInstances with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBInstances for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBInstancesWithContext(ctx aws.Context, input *DescribeDBInstancesInput, opts ...request.Option) (*DescribeDBInstancesOutput, error) { + req, out := c.DescribeDBInstancesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeDBInstancesPages iterates over the pages of a DescribeDBInstances operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeDBInstances method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeDBInstances operation. +// pageNum := 0 +// err := client.DescribeDBInstancesPages(params, +// func(page *DescribeDBInstancesOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeDBInstancesPages(input *DescribeDBInstancesInput, fn func(*DescribeDBInstancesOutput, bool) bool) error { + return c.DescribeDBInstancesPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeDBInstancesPagesWithContext same as DescribeDBInstancesPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBInstancesPagesWithContext(ctx aws.Context, input *DescribeDBInstancesInput, fn func(*DescribeDBInstancesOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeDBInstancesInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBInstancesRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeDBInstancesOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeDBParameterGroups = "DescribeDBParameterGroups" + +// DescribeDBParameterGroupsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBParameterGroups operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBParameterGroups for more information on using the DescribeDBParameterGroups +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBParameterGroupsRequest method. +// req, resp := client.DescribeDBParameterGroupsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBParameterGroups +func (c *Neptune) DescribeDBParameterGroupsRequest(input *DescribeDBParameterGroupsInput) (req *request.Request, output *DescribeDBParameterGroupsOutput) { + op := &request.Operation{ + Name: opDescribeDBParameterGroups, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeDBParameterGroupsInput{} + } + + output = &DescribeDBParameterGroupsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBParameterGroups API operation for Amazon Neptune. +// +// Returns a list of DBParameterGroup descriptions. If a DBParameterGroupName +// is specified, the list will contain only the description of the specified +// DB parameter group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBParameterGroups for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBParameterGroups +func (c *Neptune) DescribeDBParameterGroups(input *DescribeDBParameterGroupsInput) (*DescribeDBParameterGroupsOutput, error) { + req, out := c.DescribeDBParameterGroupsRequest(input) + return out, req.Send() +} + +// DescribeDBParameterGroupsWithContext is the same as DescribeDBParameterGroups with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBParameterGroups for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBParameterGroupsWithContext(ctx aws.Context, input *DescribeDBParameterGroupsInput, opts ...request.Option) (*DescribeDBParameterGroupsOutput, error) { + req, out := c.DescribeDBParameterGroupsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeDBParameterGroupsPages iterates over the pages of a DescribeDBParameterGroups operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeDBParameterGroups method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeDBParameterGroups operation. +// pageNum := 0 +// err := client.DescribeDBParameterGroupsPages(params, +// func(page *DescribeDBParameterGroupsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeDBParameterGroupsPages(input *DescribeDBParameterGroupsInput, fn func(*DescribeDBParameterGroupsOutput, bool) bool) error { + return c.DescribeDBParameterGroupsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeDBParameterGroupsPagesWithContext same as DescribeDBParameterGroupsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBParameterGroupsPagesWithContext(ctx aws.Context, input *DescribeDBParameterGroupsInput, fn func(*DescribeDBParameterGroupsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeDBParameterGroupsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBParameterGroupsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeDBParameterGroupsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeDBParameters = "DescribeDBParameters" + +// DescribeDBParametersRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBParameters operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBParameters for more information on using the DescribeDBParameters +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBParametersRequest method. +// req, resp := client.DescribeDBParametersRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBParameters +func (c *Neptune) DescribeDBParametersRequest(input *DescribeDBParametersInput) (req *request.Request, output *DescribeDBParametersOutput) { + op := &request.Operation{ + Name: opDescribeDBParameters, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeDBParametersInput{} + } + + output = &DescribeDBParametersOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBParameters API operation for Amazon Neptune. +// +// Returns the detailed parameter list for a particular DB parameter group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBParameters for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBParameters +func (c *Neptune) DescribeDBParameters(input *DescribeDBParametersInput) (*DescribeDBParametersOutput, error) { + req, out := c.DescribeDBParametersRequest(input) + return out, req.Send() +} + +// DescribeDBParametersWithContext is the same as DescribeDBParameters with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBParameters for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBParametersWithContext(ctx aws.Context, input *DescribeDBParametersInput, opts ...request.Option) (*DescribeDBParametersOutput, error) { + req, out := c.DescribeDBParametersRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeDBParametersPages iterates over the pages of a DescribeDBParameters operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeDBParameters method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeDBParameters operation. +// pageNum := 0 +// err := client.DescribeDBParametersPages(params, +// func(page *DescribeDBParametersOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeDBParametersPages(input *DescribeDBParametersInput, fn func(*DescribeDBParametersOutput, bool) bool) error { + return c.DescribeDBParametersPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeDBParametersPagesWithContext same as DescribeDBParametersPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBParametersPagesWithContext(ctx aws.Context, input *DescribeDBParametersInput, fn func(*DescribeDBParametersOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeDBParametersInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBParametersRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeDBParametersOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeDBSubnetGroups = "DescribeDBSubnetGroups" + +// DescribeDBSubnetGroupsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDBSubnetGroups operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDBSubnetGroups for more information on using the DescribeDBSubnetGroups +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDBSubnetGroupsRequest method. +// req, resp := client.DescribeDBSubnetGroupsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBSubnetGroups +func (c *Neptune) DescribeDBSubnetGroupsRequest(input *DescribeDBSubnetGroupsInput) (req *request.Request, output *DescribeDBSubnetGroupsOutput) { + op := &request.Operation{ + Name: opDescribeDBSubnetGroups, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeDBSubnetGroupsInput{} + } + + output = &DescribeDBSubnetGroupsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDBSubnetGroups API operation for Amazon Neptune. +// +// Returns a list of DBSubnetGroup descriptions. If a DBSubnetGroupName is specified, +// the list will contain only the descriptions of the specified DBSubnetGroup. +// +// For an overview of CIDR ranges, go to the Wikipedia Tutorial (http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing). +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeDBSubnetGroups for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeDBSubnetGroups +func (c *Neptune) DescribeDBSubnetGroups(input *DescribeDBSubnetGroupsInput) (*DescribeDBSubnetGroupsOutput, error) { + req, out := c.DescribeDBSubnetGroupsRequest(input) + return out, req.Send() +} + +// DescribeDBSubnetGroupsWithContext is the same as DescribeDBSubnetGroups with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDBSubnetGroups for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBSubnetGroupsWithContext(ctx aws.Context, input *DescribeDBSubnetGroupsInput, opts ...request.Option) (*DescribeDBSubnetGroupsOutput, error) { + req, out := c.DescribeDBSubnetGroupsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeDBSubnetGroupsPages iterates over the pages of a DescribeDBSubnetGroups operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeDBSubnetGroups method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeDBSubnetGroups operation. +// pageNum := 0 +// err := client.DescribeDBSubnetGroupsPages(params, +// func(page *DescribeDBSubnetGroupsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeDBSubnetGroupsPages(input *DescribeDBSubnetGroupsInput, fn func(*DescribeDBSubnetGroupsOutput, bool) bool) error { + return c.DescribeDBSubnetGroupsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeDBSubnetGroupsPagesWithContext same as DescribeDBSubnetGroupsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeDBSubnetGroupsPagesWithContext(ctx aws.Context, input *DescribeDBSubnetGroupsInput, fn func(*DescribeDBSubnetGroupsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeDBSubnetGroupsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBSubnetGroupsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeDBSubnetGroupsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeEngineDefaultClusterParameters = "DescribeEngineDefaultClusterParameters" + +// DescribeEngineDefaultClusterParametersRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEngineDefaultClusterParameters operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEngineDefaultClusterParameters for more information on using the DescribeEngineDefaultClusterParameters +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEngineDefaultClusterParametersRequest method. +// req, resp := client.DescribeEngineDefaultClusterParametersRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEngineDefaultClusterParameters +func (c *Neptune) DescribeEngineDefaultClusterParametersRequest(input *DescribeEngineDefaultClusterParametersInput) (req *request.Request, output *DescribeEngineDefaultClusterParametersOutput) { + op := &request.Operation{ + Name: opDescribeEngineDefaultClusterParameters, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeEngineDefaultClusterParametersInput{} + } + + output = &DescribeEngineDefaultClusterParametersOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEngineDefaultClusterParameters API operation for Amazon Neptune. +// +// Returns the default engine and system parameter information for the cluster +// database engine. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeEngineDefaultClusterParameters for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEngineDefaultClusterParameters +func (c *Neptune) DescribeEngineDefaultClusterParameters(input *DescribeEngineDefaultClusterParametersInput) (*DescribeEngineDefaultClusterParametersOutput, error) { + req, out := c.DescribeEngineDefaultClusterParametersRequest(input) + return out, req.Send() +} + +// DescribeEngineDefaultClusterParametersWithContext is the same as DescribeEngineDefaultClusterParameters with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEngineDefaultClusterParameters for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEngineDefaultClusterParametersWithContext(ctx aws.Context, input *DescribeEngineDefaultClusterParametersInput, opts ...request.Option) (*DescribeEngineDefaultClusterParametersOutput, error) { + req, out := c.DescribeEngineDefaultClusterParametersRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeEngineDefaultParameters = "DescribeEngineDefaultParameters" + +// DescribeEngineDefaultParametersRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEngineDefaultParameters operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEngineDefaultParameters for more information on using the DescribeEngineDefaultParameters +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEngineDefaultParametersRequest method. +// req, resp := client.DescribeEngineDefaultParametersRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEngineDefaultParameters +func (c *Neptune) DescribeEngineDefaultParametersRequest(input *DescribeEngineDefaultParametersInput) (req *request.Request, output *DescribeEngineDefaultParametersOutput) { + op := &request.Operation{ + Name: opDescribeEngineDefaultParameters, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"EngineDefaults.Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeEngineDefaultParametersInput{} + } + + output = &DescribeEngineDefaultParametersOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEngineDefaultParameters API operation for Amazon Neptune. +// +// Returns the default engine and system parameter information for the specified +// database engine. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeEngineDefaultParameters for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEngineDefaultParameters +func (c *Neptune) DescribeEngineDefaultParameters(input *DescribeEngineDefaultParametersInput) (*DescribeEngineDefaultParametersOutput, error) { + req, out := c.DescribeEngineDefaultParametersRequest(input) + return out, req.Send() +} + +// DescribeEngineDefaultParametersWithContext is the same as DescribeEngineDefaultParameters with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEngineDefaultParameters for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEngineDefaultParametersWithContext(ctx aws.Context, input *DescribeEngineDefaultParametersInput, opts ...request.Option) (*DescribeEngineDefaultParametersOutput, error) { + req, out := c.DescribeEngineDefaultParametersRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeEngineDefaultParametersPages iterates over the pages of a DescribeEngineDefaultParameters operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeEngineDefaultParameters method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeEngineDefaultParameters operation. +// pageNum := 0 +// err := client.DescribeEngineDefaultParametersPages(params, +// func(page *DescribeEngineDefaultParametersOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeEngineDefaultParametersPages(input *DescribeEngineDefaultParametersInput, fn func(*DescribeEngineDefaultParametersOutput, bool) bool) error { + return c.DescribeEngineDefaultParametersPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeEngineDefaultParametersPagesWithContext same as DescribeEngineDefaultParametersPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEngineDefaultParametersPagesWithContext(ctx aws.Context, input *DescribeEngineDefaultParametersInput, fn func(*DescribeEngineDefaultParametersOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeEngineDefaultParametersInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeEngineDefaultParametersRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeEngineDefaultParametersOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeEventCategories = "DescribeEventCategories" + +// DescribeEventCategoriesRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEventCategories operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEventCategories for more information on using the DescribeEventCategories +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEventCategoriesRequest method. +// req, resp := client.DescribeEventCategoriesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEventCategories +func (c *Neptune) DescribeEventCategoriesRequest(input *DescribeEventCategoriesInput) (req *request.Request, output *DescribeEventCategoriesOutput) { + op := &request.Operation{ + Name: opDescribeEventCategories, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeEventCategoriesInput{} + } + + output = &DescribeEventCategoriesOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEventCategories API operation for Amazon Neptune. +// +// Displays a list of categories for all event source types, or, if specified, +// for a specified source type. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeEventCategories for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEventCategories +func (c *Neptune) DescribeEventCategories(input *DescribeEventCategoriesInput) (*DescribeEventCategoriesOutput, error) { + req, out := c.DescribeEventCategoriesRequest(input) + return out, req.Send() +} + +// DescribeEventCategoriesWithContext is the same as DescribeEventCategories with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEventCategories for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEventCategoriesWithContext(ctx aws.Context, input *DescribeEventCategoriesInput, opts ...request.Option) (*DescribeEventCategoriesOutput, error) { + req, out := c.DescribeEventCategoriesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeEventSubscriptions = "DescribeEventSubscriptions" + +// DescribeEventSubscriptionsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEventSubscriptions operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEventSubscriptions for more information on using the DescribeEventSubscriptions +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEventSubscriptionsRequest method. +// req, resp := client.DescribeEventSubscriptionsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEventSubscriptions +func (c *Neptune) DescribeEventSubscriptionsRequest(input *DescribeEventSubscriptionsInput) (req *request.Request, output *DescribeEventSubscriptionsOutput) { + op := &request.Operation{ + Name: opDescribeEventSubscriptions, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeEventSubscriptionsInput{} + } + + output = &DescribeEventSubscriptionsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEventSubscriptions API operation for Amazon Neptune. +// +// Lists all the subscription descriptions for a customer account. The description +// for a subscription includes SubscriptionName, SNSTopicARN, CustomerID, SourceType, +// SourceID, CreationTime, and Status. +// +// If you specify a SubscriptionName, lists the description for that subscription. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeEventSubscriptions for usage and error information. +// +// Returned Error Codes: +// * ErrCodeSubscriptionNotFoundFault "SubscriptionNotFound" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEventSubscriptions +func (c *Neptune) DescribeEventSubscriptions(input *DescribeEventSubscriptionsInput) (*DescribeEventSubscriptionsOutput, error) { + req, out := c.DescribeEventSubscriptionsRequest(input) + return out, req.Send() +} + +// DescribeEventSubscriptionsWithContext is the same as DescribeEventSubscriptions with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEventSubscriptions for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEventSubscriptionsWithContext(ctx aws.Context, input *DescribeEventSubscriptionsInput, opts ...request.Option) (*DescribeEventSubscriptionsOutput, error) { + req, out := c.DescribeEventSubscriptionsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeEventSubscriptionsPages iterates over the pages of a DescribeEventSubscriptions operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeEventSubscriptions method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeEventSubscriptions operation. +// pageNum := 0 +// err := client.DescribeEventSubscriptionsPages(params, +// func(page *DescribeEventSubscriptionsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeEventSubscriptionsPages(input *DescribeEventSubscriptionsInput, fn func(*DescribeEventSubscriptionsOutput, bool) bool) error { + return c.DescribeEventSubscriptionsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeEventSubscriptionsPagesWithContext same as DescribeEventSubscriptionsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEventSubscriptionsPagesWithContext(ctx aws.Context, input *DescribeEventSubscriptionsInput, fn func(*DescribeEventSubscriptionsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeEventSubscriptionsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeEventSubscriptionsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeEventSubscriptionsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeEvents = "DescribeEvents" + +// DescribeEventsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEvents operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEvents for more information on using the DescribeEvents +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEventsRequest method. +// req, resp := client.DescribeEventsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEvents +func (c *Neptune) DescribeEventsRequest(input *DescribeEventsInput) (req *request.Request, output *DescribeEventsOutput) { + op := &request.Operation{ + Name: opDescribeEvents, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeEventsInput{} + } + + output = &DescribeEventsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEvents API operation for Amazon Neptune. +// +// Returns events related to DB instances, DB security groups, DB snapshots, +// and DB parameter groups for the past 14 days. Events specific to a particular +// DB instance, DB security group, database snapshot, or DB parameter group +// can be obtained by providing the name as a parameter. By default, the past +// hour of events are returned. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeEvents for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeEvents +func (c *Neptune) DescribeEvents(input *DescribeEventsInput) (*DescribeEventsOutput, error) { + req, out := c.DescribeEventsRequest(input) + return out, req.Send() +} + +// DescribeEventsWithContext is the same as DescribeEvents with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEvents for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEventsWithContext(ctx aws.Context, input *DescribeEventsInput, opts ...request.Option) (*DescribeEventsOutput, error) { + req, out := c.DescribeEventsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeEventsPages iterates over the pages of a DescribeEvents operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeEvents method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeEvents operation. +// pageNum := 0 +// err := client.DescribeEventsPages(params, +// func(page *DescribeEventsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeEventsPages(input *DescribeEventsInput, fn func(*DescribeEventsOutput, bool) bool) error { + return c.DescribeEventsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeEventsPagesWithContext same as DescribeEventsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeEventsPagesWithContext(ctx aws.Context, input *DescribeEventsInput, fn func(*DescribeEventsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeEventsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeEventsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeEventsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribeOrderableDBInstanceOptions = "DescribeOrderableDBInstanceOptions" + +// DescribeOrderableDBInstanceOptionsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeOrderableDBInstanceOptions operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeOrderableDBInstanceOptions for more information on using the DescribeOrderableDBInstanceOptions +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeOrderableDBInstanceOptionsRequest method. +// req, resp := client.DescribeOrderableDBInstanceOptionsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeOrderableDBInstanceOptions +func (c *Neptune) DescribeOrderableDBInstanceOptionsRequest(input *DescribeOrderableDBInstanceOptionsInput) (req *request.Request, output *DescribeOrderableDBInstanceOptionsOutput) { + op := &request.Operation{ + Name: opDescribeOrderableDBInstanceOptions, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"Marker"}, + OutputTokens: []string{"Marker"}, + LimitToken: "MaxRecords", + TruncationToken: "", + }, + } + + if input == nil { + input = &DescribeOrderableDBInstanceOptionsInput{} + } + + output = &DescribeOrderableDBInstanceOptionsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeOrderableDBInstanceOptions API operation for Amazon Neptune. +// +// Returns a list of orderable DB instance options for the specified engine. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeOrderableDBInstanceOptions for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeOrderableDBInstanceOptions +func (c *Neptune) DescribeOrderableDBInstanceOptions(input *DescribeOrderableDBInstanceOptionsInput) (*DescribeOrderableDBInstanceOptionsOutput, error) { + req, out := c.DescribeOrderableDBInstanceOptionsRequest(input) + return out, req.Send() +} + +// DescribeOrderableDBInstanceOptionsWithContext is the same as DescribeOrderableDBInstanceOptions with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeOrderableDBInstanceOptions for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeOrderableDBInstanceOptionsWithContext(ctx aws.Context, input *DescribeOrderableDBInstanceOptionsInput, opts ...request.Option) (*DescribeOrderableDBInstanceOptionsOutput, error) { + req, out := c.DescribeOrderableDBInstanceOptionsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// DescribeOrderableDBInstanceOptionsPages iterates over the pages of a DescribeOrderableDBInstanceOptions operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See DescribeOrderableDBInstanceOptions method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a DescribeOrderableDBInstanceOptions operation. +// pageNum := 0 +// err := client.DescribeOrderableDBInstanceOptionsPages(params, +// func(page *DescribeOrderableDBInstanceOptionsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *Neptune) DescribeOrderableDBInstanceOptionsPages(input *DescribeOrderableDBInstanceOptionsInput, fn func(*DescribeOrderableDBInstanceOptionsOutput, bool) bool) error { + return c.DescribeOrderableDBInstanceOptionsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// DescribeOrderableDBInstanceOptionsPagesWithContext same as DescribeOrderableDBInstanceOptionsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeOrderableDBInstanceOptionsPagesWithContext(ctx aws.Context, input *DescribeOrderableDBInstanceOptionsInput, fn func(*DescribeOrderableDBInstanceOptionsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *DescribeOrderableDBInstanceOptionsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeOrderableDBInstanceOptionsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*DescribeOrderableDBInstanceOptionsOutput), !p.HasNextPage()) + } + return p.Err() +} + +const opDescribePendingMaintenanceActions = "DescribePendingMaintenanceActions" + +// DescribePendingMaintenanceActionsRequest generates a "aws/request.Request" representing the +// client's request for the DescribePendingMaintenanceActions operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribePendingMaintenanceActions for more information on using the DescribePendingMaintenanceActions +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribePendingMaintenanceActionsRequest method. +// req, resp := client.DescribePendingMaintenanceActionsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribePendingMaintenanceActions +func (c *Neptune) DescribePendingMaintenanceActionsRequest(input *DescribePendingMaintenanceActionsInput) (req *request.Request, output *DescribePendingMaintenanceActionsOutput) { + op := &request.Operation{ + Name: opDescribePendingMaintenanceActions, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribePendingMaintenanceActionsInput{} + } + + output = &DescribePendingMaintenanceActionsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribePendingMaintenanceActions API operation for Amazon Neptune. +// +// Returns a list of resources (for example, DB instances) that have at least +// one pending maintenance action. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribePendingMaintenanceActions for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFoundFault "ResourceNotFoundFault" +// The specified resource ID was not found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribePendingMaintenanceActions +func (c *Neptune) DescribePendingMaintenanceActions(input *DescribePendingMaintenanceActionsInput) (*DescribePendingMaintenanceActionsOutput, error) { + req, out := c.DescribePendingMaintenanceActionsRequest(input) + return out, req.Send() +} + +// DescribePendingMaintenanceActionsWithContext is the same as DescribePendingMaintenanceActions with the addition of +// the ability to pass a context and additional request options. +// +// See DescribePendingMaintenanceActions for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribePendingMaintenanceActionsWithContext(ctx aws.Context, input *DescribePendingMaintenanceActionsInput, opts ...request.Option) (*DescribePendingMaintenanceActionsOutput, error) { + req, out := c.DescribePendingMaintenanceActionsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeValidDBInstanceModifications = "DescribeValidDBInstanceModifications" + +// DescribeValidDBInstanceModificationsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeValidDBInstanceModifications operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeValidDBInstanceModifications for more information on using the DescribeValidDBInstanceModifications +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeValidDBInstanceModificationsRequest method. +// req, resp := client.DescribeValidDBInstanceModificationsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeValidDBInstanceModifications +func (c *Neptune) DescribeValidDBInstanceModificationsRequest(input *DescribeValidDBInstanceModificationsInput) (req *request.Request, output *DescribeValidDBInstanceModificationsOutput) { + op := &request.Operation{ + Name: opDescribeValidDBInstanceModifications, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeValidDBInstanceModificationsInput{} + } + + output = &DescribeValidDBInstanceModificationsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeValidDBInstanceModifications API operation for Amazon Neptune. +// +// You can call DescribeValidDBInstanceModifications to learn what modifications +// you can make to your DB instance. You can use this information when you call +// ModifyDBInstance. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation DescribeValidDBInstanceModifications for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/DescribeValidDBInstanceModifications +func (c *Neptune) DescribeValidDBInstanceModifications(input *DescribeValidDBInstanceModificationsInput) (*DescribeValidDBInstanceModificationsOutput, error) { + req, out := c.DescribeValidDBInstanceModificationsRequest(input) + return out, req.Send() +} + +// DescribeValidDBInstanceModificationsWithContext is the same as DescribeValidDBInstanceModifications with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeValidDBInstanceModifications for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) DescribeValidDBInstanceModificationsWithContext(ctx aws.Context, input *DescribeValidDBInstanceModificationsInput, opts ...request.Option) (*DescribeValidDBInstanceModificationsOutput, error) { + req, out := c.DescribeValidDBInstanceModificationsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opFailoverDBCluster = "FailoverDBCluster" + +// FailoverDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the FailoverDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See FailoverDBCluster for more information on using the FailoverDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the FailoverDBClusterRequest method. +// req, resp := client.FailoverDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/FailoverDBCluster +func (c *Neptune) FailoverDBClusterRequest(input *FailoverDBClusterInput) (req *request.Request, output *FailoverDBClusterOutput) { + op := &request.Operation{ + Name: opFailoverDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &FailoverDBClusterInput{} + } + + output = &FailoverDBClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// FailoverDBCluster API operation for Amazon Neptune. +// +// Forces a failover for a DB cluster. +// +// A failover for a DB cluster promotes one of the Read Replicas (read-only +// instances) in the DB cluster to be the primary instance (the cluster writer). +// +// Amazon Neptune will automatically fail over to a Read Replica, if one exists, +// when the primary instance fails. You can force a failover when you want to +// simulate a failure of a primary instance for testing. Because each instance +// in a DB cluster has its own endpoint address, you will need to clean up and +// re-establish any existing connections that use those endpoint addresses when +// the failover is complete. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation FailoverDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/FailoverDBCluster +func (c *Neptune) FailoverDBCluster(input *FailoverDBClusterInput) (*FailoverDBClusterOutput, error) { + req, out := c.FailoverDBClusterRequest(input) + return out, req.Send() +} + +// FailoverDBClusterWithContext is the same as FailoverDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See FailoverDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) FailoverDBClusterWithContext(ctx aws.Context, input *FailoverDBClusterInput, opts ...request.Option) (*FailoverDBClusterOutput, error) { + req, out := c.FailoverDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opListTagsForResource = "ListTagsForResource" + +// ListTagsForResourceRequest generates a "aws/request.Request" representing the +// client's request for the ListTagsForResource operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListTagsForResource for more information on using the ListTagsForResource +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListTagsForResourceRequest method. +// req, resp := client.ListTagsForResourceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ListTagsForResource +func (c *Neptune) ListTagsForResourceRequest(input *ListTagsForResourceInput) (req *request.Request, output *ListTagsForResourceOutput) { + op := &request.Operation{ + Name: opListTagsForResource, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ListTagsForResourceInput{} + } + + output = &ListTagsForResourceOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListTagsForResource API operation for Amazon Neptune. +// +// Lists all tags on an Amazon Neptune resource. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ListTagsForResource for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeDBSnapshotNotFoundFault "DBSnapshotNotFound" +// DBSnapshotIdentifier does not refer to an existing DB snapshot. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ListTagsForResource +func (c *Neptune) ListTagsForResource(input *ListTagsForResourceInput) (*ListTagsForResourceOutput, error) { + req, out := c.ListTagsForResourceRequest(input) + return out, req.Send() +} + +// ListTagsForResourceWithContext is the same as ListTagsForResource with the addition of +// the ability to pass a context and additional request options. +// +// See ListTagsForResource for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ListTagsForResourceWithContext(ctx aws.Context, input *ListTagsForResourceInput, opts ...request.Option) (*ListTagsForResourceOutput, error) { + req, out := c.ListTagsForResourceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyDBCluster = "ModifyDBCluster" + +// ModifyDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the ModifyDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyDBCluster for more information on using the ModifyDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyDBClusterRequest method. +// req, resp := client.ModifyDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBCluster +func (c *Neptune) ModifyDBClusterRequest(input *ModifyDBClusterInput) (req *request.Request, output *ModifyDBClusterOutput) { + op := &request.Operation{ + Name: opModifyDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyDBClusterInput{} + } + + output = &ModifyDBClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyDBCluster API operation for Amazon Neptune. +// +// Modify a setting for a DB cluster. You can change one or more database configuration +// parameters by specifying these parameters and the new values in the request. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeInvalidVPCNetworkStateFault "InvalidVPCNetworkStateFault" +// DB subnet group does not cover all Availability Zones after it is created +// because users' change. +// +// * ErrCodeInvalidDBSubnetGroupStateFault "InvalidDBSubnetGroupStateFault" +// The DB subnet group cannot be deleted because it is in use. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// * ErrCodeDBClusterParameterGroupNotFoundFault "DBClusterParameterGroupNotFound" +// DBClusterParameterGroupName does not refer to an existing DB Cluster parameter +// group. +// +// * ErrCodeInvalidDBSecurityGroupStateFault "InvalidDBSecurityGroupState" +// The state of the DB security group does not allow deletion. +// +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// * ErrCodeDBClusterAlreadyExistsFault "DBClusterAlreadyExistsFault" +// User already has a DB cluster with the given identifier. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBCluster +func (c *Neptune) ModifyDBCluster(input *ModifyDBClusterInput) (*ModifyDBClusterOutput, error) { + req, out := c.ModifyDBClusterRequest(input) + return out, req.Send() +} + +// ModifyDBClusterWithContext is the same as ModifyDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyDBClusterWithContext(ctx aws.Context, input *ModifyDBClusterInput, opts ...request.Option) (*ModifyDBClusterOutput, error) { + req, out := c.ModifyDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyDBClusterParameterGroup = "ModifyDBClusterParameterGroup" + +// ModifyDBClusterParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the ModifyDBClusterParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyDBClusterParameterGroup for more information on using the ModifyDBClusterParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyDBClusterParameterGroupRequest method. +// req, resp := client.ModifyDBClusterParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBClusterParameterGroup +func (c *Neptune) ModifyDBClusterParameterGroupRequest(input *ModifyDBClusterParameterGroupInput) (req *request.Request, output *ResetDBClusterParameterGroupOutput) { + op := &request.Operation{ + Name: opModifyDBClusterParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyDBClusterParameterGroupInput{} + } + + output = &ResetDBClusterParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyDBClusterParameterGroup API operation for Amazon Neptune. +// +// Modifies the parameters of a DB cluster parameter group. To modify more than +// one parameter, submit a list of the following: ParameterName, ParameterValue, +// and ApplyMethod. A maximum of 20 parameters can be modified in a single request. +// +// Changes to dynamic parameters are applied immediately. Changes to static +// parameters require a reboot without failover to the DB cluster associated +// with the parameter group before the change can take effect. +// +// After you create a DB cluster parameter group, you should wait at least 5 +// minutes before creating your first DB cluster that uses that DB cluster parameter +// group as the default parameter group. This allows Amazon Neptune to fully +// complete the create action before the parameter group is used as the default +// for a new DB cluster. This is especially important for parameters that are +// critical when creating the default database for a DB cluster, such as the +// character set for the default database defined by the character_set_database +// parameter. You can use the Parameter Groups option of the Amazon Neptune +// console or the DescribeDBClusterParameters command to verify that your DB +// cluster parameter group has been created or modified. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyDBClusterParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// * ErrCodeInvalidDBParameterGroupStateFault "InvalidDBParameterGroupState" +// The DB parameter group is in use or is in an invalid state. If you are attempting +// to delete the parameter group, you cannot delete it when the parameter group +// is in this state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBClusterParameterGroup +func (c *Neptune) ModifyDBClusterParameterGroup(input *ModifyDBClusterParameterGroupInput) (*ResetDBClusterParameterGroupOutput, error) { + req, out := c.ModifyDBClusterParameterGroupRequest(input) + return out, req.Send() +} + +// ModifyDBClusterParameterGroupWithContext is the same as ModifyDBClusterParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyDBClusterParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyDBClusterParameterGroupWithContext(ctx aws.Context, input *ModifyDBClusterParameterGroupInput, opts ...request.Option) (*ResetDBClusterParameterGroupOutput, error) { + req, out := c.ModifyDBClusterParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyDBClusterSnapshotAttribute = "ModifyDBClusterSnapshotAttribute" + +// ModifyDBClusterSnapshotAttributeRequest generates a "aws/request.Request" representing the +// client's request for the ModifyDBClusterSnapshotAttribute operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyDBClusterSnapshotAttribute for more information on using the ModifyDBClusterSnapshotAttribute +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyDBClusterSnapshotAttributeRequest method. +// req, resp := client.ModifyDBClusterSnapshotAttributeRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBClusterSnapshotAttribute +func (c *Neptune) ModifyDBClusterSnapshotAttributeRequest(input *ModifyDBClusterSnapshotAttributeInput) (req *request.Request, output *ModifyDBClusterSnapshotAttributeOutput) { + op := &request.Operation{ + Name: opModifyDBClusterSnapshotAttribute, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyDBClusterSnapshotAttributeInput{} + } + + output = &ModifyDBClusterSnapshotAttributeOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyDBClusterSnapshotAttribute API operation for Amazon Neptune. +// +// Adds an attribute and values to, or removes an attribute and values from, +// a manual DB cluster snapshot. +// +// To share a manual DB cluster snapshot with other AWS accounts, specify restore +// as the AttributeName and use the ValuesToAdd parameter to add a list of IDs +// of the AWS accounts that are authorized to restore the manual DB cluster +// snapshot. Use the value all to make the manual DB cluster snapshot public, +// which means that it can be copied or restored by all AWS accounts. Do not +// add the all value for any manual DB cluster snapshots that contain private +// information that you don't want available to all AWS accounts. If a manual +// DB cluster snapshot is encrypted, it can be shared, but only by specifying +// a list of authorized AWS account IDs for the ValuesToAdd parameter. You can't +// use all as a value for that parameter in this case. +// +// To view which AWS accounts have access to copy or restore a manual DB cluster +// snapshot, or whether a manual DB cluster snapshot public or private, use +// the DescribeDBClusterSnapshotAttributes API action. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyDBClusterSnapshotAttribute for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// * ErrCodeSharedSnapshotQuotaExceededFault "SharedSnapshotQuotaExceeded" +// You have exceeded the maximum number of accounts that you can share a manual +// DB snapshot with. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBClusterSnapshotAttribute +func (c *Neptune) ModifyDBClusterSnapshotAttribute(input *ModifyDBClusterSnapshotAttributeInput) (*ModifyDBClusterSnapshotAttributeOutput, error) { + req, out := c.ModifyDBClusterSnapshotAttributeRequest(input) + return out, req.Send() +} + +// ModifyDBClusterSnapshotAttributeWithContext is the same as ModifyDBClusterSnapshotAttribute with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyDBClusterSnapshotAttribute for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyDBClusterSnapshotAttributeWithContext(ctx aws.Context, input *ModifyDBClusterSnapshotAttributeInput, opts ...request.Option) (*ModifyDBClusterSnapshotAttributeOutput, error) { + req, out := c.ModifyDBClusterSnapshotAttributeRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyDBInstance = "ModifyDBInstance" + +// ModifyDBInstanceRequest generates a "aws/request.Request" representing the +// client's request for the ModifyDBInstance operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyDBInstance for more information on using the ModifyDBInstance +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyDBInstanceRequest method. +// req, resp := client.ModifyDBInstanceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBInstance +func (c *Neptune) ModifyDBInstanceRequest(input *ModifyDBInstanceInput) (req *request.Request, output *ModifyDBInstanceOutput) { + op := &request.Operation{ + Name: opModifyDBInstance, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyDBInstanceInput{} + } + + output = &ModifyDBInstanceOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyDBInstance API operation for Amazon Neptune. +// +// Modifies settings for a DB instance. You can change one or more database +// configuration parameters by specifying these parameters and the new values +// in the request. To learn what modifications you can make to your DB instance, +// call DescribeValidDBInstanceModifications before you call ModifyDBInstance. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyDBInstance for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// * ErrCodeInvalidDBSecurityGroupStateFault "InvalidDBSecurityGroupState" +// The state of the DB security group does not allow deletion. +// +// * ErrCodeDBInstanceAlreadyExistsFault "DBInstanceAlreadyExists" +// User already has a DB instance with the given identifier. +// +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeDBSecurityGroupNotFoundFault "DBSecurityGroupNotFound" +// DBSecurityGroupName does not refer to an existing DB security group. +// +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// * ErrCodeInsufficientDBInstanceCapacityFault "InsufficientDBInstanceCapacity" +// Specified DB instance class is not available in the specified Availability +// Zone. +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// * ErrCodeInvalidVPCNetworkStateFault "InvalidVPCNetworkStateFault" +// DB subnet group does not cover all Availability Zones after it is created +// because users' change. +// +// * ErrCodeProvisionedIopsNotAvailableInAZFault "ProvisionedIopsNotAvailableInAZFault" +// Provisioned IOPS not available in the specified Availability Zone. +// +// * ErrCodeOptionGroupNotFoundFault "OptionGroupNotFoundFault" +// +// * ErrCodeDBUpgradeDependencyFailureFault "DBUpgradeDependencyFailure" +// The DB upgrade failed because a resource the DB depends on could not be modified. +// +// * ErrCodeStorageTypeNotSupportedFault "StorageTypeNotSupported" +// StorageType specified cannot be associated with the DB Instance. +// +// * ErrCodeAuthorizationNotFoundFault "AuthorizationNotFound" +// Specified CIDRIP or EC2 security group is not authorized for the specified +// DB security group. +// +// Neptune may not also be authorized via IAM to perform necessary actions on +// your behalf. +// +// * ErrCodeCertificateNotFoundFault "CertificateNotFound" +// CertificateIdentifier does not refer to an existing certificate. +// +// * ErrCodeDomainNotFoundFault "DomainNotFoundFault" +// Domain does not refer to an existing Active Directory Domain. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBInstance +func (c *Neptune) ModifyDBInstance(input *ModifyDBInstanceInput) (*ModifyDBInstanceOutput, error) { + req, out := c.ModifyDBInstanceRequest(input) + return out, req.Send() +} + +// ModifyDBInstanceWithContext is the same as ModifyDBInstance with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyDBInstance for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyDBInstanceWithContext(ctx aws.Context, input *ModifyDBInstanceInput, opts ...request.Option) (*ModifyDBInstanceOutput, error) { + req, out := c.ModifyDBInstanceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyDBParameterGroup = "ModifyDBParameterGroup" + +// ModifyDBParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the ModifyDBParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyDBParameterGroup for more information on using the ModifyDBParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyDBParameterGroupRequest method. +// req, resp := client.ModifyDBParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBParameterGroup +func (c *Neptune) ModifyDBParameterGroupRequest(input *ModifyDBParameterGroupInput) (req *request.Request, output *ResetDBParameterGroupOutput) { + op := &request.Operation{ + Name: opModifyDBParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyDBParameterGroupInput{} + } + + output = &ResetDBParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyDBParameterGroup API operation for Amazon Neptune. +// +// Modifies the parameters of a DB parameter group. To modify more than one +// parameter, submit a list of the following: ParameterName, ParameterValue, +// and ApplyMethod. A maximum of 20 parameters can be modified in a single request. +// +// Changes to dynamic parameters are applied immediately. Changes to static +// parameters require a reboot without failover to the DB instance associated +// with the parameter group before the change can take effect. +// +// After you modify a DB parameter group, you should wait at least 5 minutes +// before creating your first DB instance that uses that DB parameter group +// as the default parameter group. This allows Amazon Neptune to fully complete +// the modify action before the parameter group is used as the default for a +// new DB instance. This is especially important for parameters that are critical +// when creating the default database for a DB instance, such as the character +// set for the default database defined by the character_set_database parameter. +// You can use the Parameter Groups option of the Amazon Neptune console or +// the DescribeDBParameters command to verify that your DB parameter group has +// been created or modified. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyDBParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// * ErrCodeInvalidDBParameterGroupStateFault "InvalidDBParameterGroupState" +// The DB parameter group is in use or is in an invalid state. If you are attempting +// to delete the parameter group, you cannot delete it when the parameter group +// is in this state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBParameterGroup +func (c *Neptune) ModifyDBParameterGroup(input *ModifyDBParameterGroupInput) (*ResetDBParameterGroupOutput, error) { + req, out := c.ModifyDBParameterGroupRequest(input) + return out, req.Send() +} + +// ModifyDBParameterGroupWithContext is the same as ModifyDBParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyDBParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyDBParameterGroupWithContext(ctx aws.Context, input *ModifyDBParameterGroupInput, opts ...request.Option) (*ResetDBParameterGroupOutput, error) { + req, out := c.ModifyDBParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyDBSubnetGroup = "ModifyDBSubnetGroup" + +// ModifyDBSubnetGroupRequest generates a "aws/request.Request" representing the +// client's request for the ModifyDBSubnetGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyDBSubnetGroup for more information on using the ModifyDBSubnetGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyDBSubnetGroupRequest method. +// req, resp := client.ModifyDBSubnetGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBSubnetGroup +func (c *Neptune) ModifyDBSubnetGroupRequest(input *ModifyDBSubnetGroupInput) (req *request.Request, output *ModifyDBSubnetGroupOutput) { + op := &request.Operation{ + Name: opModifyDBSubnetGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyDBSubnetGroupInput{} + } + + output = &ModifyDBSubnetGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyDBSubnetGroup API operation for Amazon Neptune. +// +// Modifies an existing DB subnet group. DB subnet groups must contain at least +// one subnet in at least two AZs in the AWS Region. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyDBSubnetGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeDBSubnetQuotaExceededFault "DBSubnetQuotaExceededFault" +// Request would result in user exceeding the allowed number of subnets in a +// DB subnet groups. +// +// * ErrCodeSubnetAlreadyInUse "SubnetAlreadyInUse" +// The DB subnet is already in use in the Availability Zone. +// +// * ErrCodeDBSubnetGroupDoesNotCoverEnoughAZs "DBSubnetGroupDoesNotCoverEnoughAZs" +// Subnets in the DB subnet group should cover at least two Availability Zones +// unless there is only one Availability Zone. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyDBSubnetGroup +func (c *Neptune) ModifyDBSubnetGroup(input *ModifyDBSubnetGroupInput) (*ModifyDBSubnetGroupOutput, error) { + req, out := c.ModifyDBSubnetGroupRequest(input) + return out, req.Send() +} + +// ModifyDBSubnetGroupWithContext is the same as ModifyDBSubnetGroup with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyDBSubnetGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyDBSubnetGroupWithContext(ctx aws.Context, input *ModifyDBSubnetGroupInput, opts ...request.Option) (*ModifyDBSubnetGroupOutput, error) { + req, out := c.ModifyDBSubnetGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opModifyEventSubscription = "ModifyEventSubscription" + +// ModifyEventSubscriptionRequest generates a "aws/request.Request" representing the +// client's request for the ModifyEventSubscription operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ModifyEventSubscription for more information on using the ModifyEventSubscription +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ModifyEventSubscriptionRequest method. +// req, resp := client.ModifyEventSubscriptionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyEventSubscription +func (c *Neptune) ModifyEventSubscriptionRequest(input *ModifyEventSubscriptionInput) (req *request.Request, output *ModifyEventSubscriptionOutput) { + op := &request.Operation{ + Name: opModifyEventSubscription, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ModifyEventSubscriptionInput{} + } + + output = &ModifyEventSubscriptionOutput{} + req = c.newRequest(op, input, output) + return +} + +// ModifyEventSubscription API operation for Amazon Neptune. +// +// Modifies an existing event notification subscription. Note that you can't +// modify the source identifiers using this call; to change source identifiers +// for a subscription, use the AddSourceIdentifierToSubscription and RemoveSourceIdentifierFromSubscription +// calls. +// +// You can see a list of the event categories for a given SourceType by using +// the DescribeEventCategories action. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ModifyEventSubscription for usage and error information. +// +// Returned Error Codes: +// * ErrCodeEventSubscriptionQuotaExceededFault "EventSubscriptionQuotaExceeded" +// +// * ErrCodeSubscriptionNotFoundFault "SubscriptionNotFound" +// +// * ErrCodeSNSInvalidTopicFault "SNSInvalidTopic" +// +// * ErrCodeSNSNoAuthorizationFault "SNSNoAuthorization" +// +// * ErrCodeSNSTopicArnNotFoundFault "SNSTopicArnNotFound" +// +// * ErrCodeSubscriptionCategoryNotFoundFault "SubscriptionCategoryNotFound" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ModifyEventSubscription +func (c *Neptune) ModifyEventSubscription(input *ModifyEventSubscriptionInput) (*ModifyEventSubscriptionOutput, error) { + req, out := c.ModifyEventSubscriptionRequest(input) + return out, req.Send() +} + +// ModifyEventSubscriptionWithContext is the same as ModifyEventSubscription with the addition of +// the ability to pass a context and additional request options. +// +// See ModifyEventSubscription for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ModifyEventSubscriptionWithContext(ctx aws.Context, input *ModifyEventSubscriptionInput, opts ...request.Option) (*ModifyEventSubscriptionOutput, error) { + req, out := c.ModifyEventSubscriptionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opPromoteReadReplicaDBCluster = "PromoteReadReplicaDBCluster" + +// PromoteReadReplicaDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the PromoteReadReplicaDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See PromoteReadReplicaDBCluster for more information on using the PromoteReadReplicaDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the PromoteReadReplicaDBClusterRequest method. +// req, resp := client.PromoteReadReplicaDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/PromoteReadReplicaDBCluster +func (c *Neptune) PromoteReadReplicaDBClusterRequest(input *PromoteReadReplicaDBClusterInput) (req *request.Request, output *PromoteReadReplicaDBClusterOutput) { + op := &request.Operation{ + Name: opPromoteReadReplicaDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &PromoteReadReplicaDBClusterInput{} + } + + output = &PromoteReadReplicaDBClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// PromoteReadReplicaDBCluster API operation for Amazon Neptune. +// +// Promotes a Read Replica DB cluster to a standalone DB cluster. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation PromoteReadReplicaDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/PromoteReadReplicaDBCluster +func (c *Neptune) PromoteReadReplicaDBCluster(input *PromoteReadReplicaDBClusterInput) (*PromoteReadReplicaDBClusterOutput, error) { + req, out := c.PromoteReadReplicaDBClusterRequest(input) + return out, req.Send() +} + +// PromoteReadReplicaDBClusterWithContext is the same as PromoteReadReplicaDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See PromoteReadReplicaDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) PromoteReadReplicaDBClusterWithContext(ctx aws.Context, input *PromoteReadReplicaDBClusterInput, opts ...request.Option) (*PromoteReadReplicaDBClusterOutput, error) { + req, out := c.PromoteReadReplicaDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opRebootDBInstance = "RebootDBInstance" + +// RebootDBInstanceRequest generates a "aws/request.Request" representing the +// client's request for the RebootDBInstance operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RebootDBInstance for more information on using the RebootDBInstance +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RebootDBInstanceRequest method. +// req, resp := client.RebootDBInstanceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RebootDBInstance +func (c *Neptune) RebootDBInstanceRequest(input *RebootDBInstanceInput) (req *request.Request, output *RebootDBInstanceOutput) { + op := &request.Operation{ + Name: opRebootDBInstance, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RebootDBInstanceInput{} + } + + output = &RebootDBInstanceOutput{} + req = c.newRequest(op, input, output) + return +} + +// RebootDBInstance API operation for Amazon Neptune. +// +// You might need to reboot your DB instance, usually for maintenance reasons. +// For example, if you make certain modifications, or if you change the DB parameter +// group associated with the DB instance, you must reboot the instance for the +// changes to take effect. +// +// Rebooting a DB instance restarts the database engine service. Rebooting a +// DB instance results in a momentary outage, during which the DB instance status +// is set to rebooting. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation RebootDBInstance for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBInstanceStateFault "InvalidDBInstanceState" +// The specified DB instance is not in the available state. +// +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RebootDBInstance +func (c *Neptune) RebootDBInstance(input *RebootDBInstanceInput) (*RebootDBInstanceOutput, error) { + req, out := c.RebootDBInstanceRequest(input) + return out, req.Send() +} + +// RebootDBInstanceWithContext is the same as RebootDBInstance with the addition of +// the ability to pass a context and additional request options. +// +// See RebootDBInstance for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) RebootDBInstanceWithContext(ctx aws.Context, input *RebootDBInstanceInput, opts ...request.Option) (*RebootDBInstanceOutput, error) { + req, out := c.RebootDBInstanceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opRemoveRoleFromDBCluster = "RemoveRoleFromDBCluster" + +// RemoveRoleFromDBClusterRequest generates a "aws/request.Request" representing the +// client's request for the RemoveRoleFromDBCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RemoveRoleFromDBCluster for more information on using the RemoveRoleFromDBCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RemoveRoleFromDBClusterRequest method. +// req, resp := client.RemoveRoleFromDBClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RemoveRoleFromDBCluster +func (c *Neptune) RemoveRoleFromDBClusterRequest(input *RemoveRoleFromDBClusterInput) (req *request.Request, output *RemoveRoleFromDBClusterOutput) { + op := &request.Operation{ + Name: opRemoveRoleFromDBCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RemoveRoleFromDBClusterInput{} + } + + output = &RemoveRoleFromDBClusterOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// RemoveRoleFromDBCluster API operation for Amazon Neptune. +// +// Disassociates an Identity and Access Management (IAM) role from a DB cluster. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation RemoveRoleFromDBCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeDBClusterRoleNotFoundFault "DBClusterRoleNotFound" +// The specified IAM role Amazon Resource Name (ARN) is not associated with +// the specified DB cluster. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RemoveRoleFromDBCluster +func (c *Neptune) RemoveRoleFromDBCluster(input *RemoveRoleFromDBClusterInput) (*RemoveRoleFromDBClusterOutput, error) { + req, out := c.RemoveRoleFromDBClusterRequest(input) + return out, req.Send() +} + +// RemoveRoleFromDBClusterWithContext is the same as RemoveRoleFromDBCluster with the addition of +// the ability to pass a context and additional request options. +// +// See RemoveRoleFromDBCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) RemoveRoleFromDBClusterWithContext(ctx aws.Context, input *RemoveRoleFromDBClusterInput, opts ...request.Option) (*RemoveRoleFromDBClusterOutput, error) { + req, out := c.RemoveRoleFromDBClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opRemoveSourceIdentifierFromSubscription = "RemoveSourceIdentifierFromSubscription" + +// RemoveSourceIdentifierFromSubscriptionRequest generates a "aws/request.Request" representing the +// client's request for the RemoveSourceIdentifierFromSubscription operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RemoveSourceIdentifierFromSubscription for more information on using the RemoveSourceIdentifierFromSubscription +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RemoveSourceIdentifierFromSubscriptionRequest method. +// req, resp := client.RemoveSourceIdentifierFromSubscriptionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RemoveSourceIdentifierFromSubscription +func (c *Neptune) RemoveSourceIdentifierFromSubscriptionRequest(input *RemoveSourceIdentifierFromSubscriptionInput) (req *request.Request, output *RemoveSourceIdentifierFromSubscriptionOutput) { + op := &request.Operation{ + Name: opRemoveSourceIdentifierFromSubscription, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RemoveSourceIdentifierFromSubscriptionInput{} + } + + output = &RemoveSourceIdentifierFromSubscriptionOutput{} + req = c.newRequest(op, input, output) + return +} + +// RemoveSourceIdentifierFromSubscription API operation for Amazon Neptune. +// +// Removes a source identifier from an existing event notification subscription. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation RemoveSourceIdentifierFromSubscription for usage and error information. +// +// Returned Error Codes: +// * ErrCodeSubscriptionNotFoundFault "SubscriptionNotFound" +// +// * ErrCodeSourceNotFoundFault "SourceNotFound" +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RemoveSourceIdentifierFromSubscription +func (c *Neptune) RemoveSourceIdentifierFromSubscription(input *RemoveSourceIdentifierFromSubscriptionInput) (*RemoveSourceIdentifierFromSubscriptionOutput, error) { + req, out := c.RemoveSourceIdentifierFromSubscriptionRequest(input) + return out, req.Send() +} + +// RemoveSourceIdentifierFromSubscriptionWithContext is the same as RemoveSourceIdentifierFromSubscription with the addition of +// the ability to pass a context and additional request options. +// +// See RemoveSourceIdentifierFromSubscription for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) RemoveSourceIdentifierFromSubscriptionWithContext(ctx aws.Context, input *RemoveSourceIdentifierFromSubscriptionInput, opts ...request.Option) (*RemoveSourceIdentifierFromSubscriptionOutput, error) { + req, out := c.RemoveSourceIdentifierFromSubscriptionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opRemoveTagsFromResource = "RemoveTagsFromResource" + +// RemoveTagsFromResourceRequest generates a "aws/request.Request" representing the +// client's request for the RemoveTagsFromResource operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RemoveTagsFromResource for more information on using the RemoveTagsFromResource +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RemoveTagsFromResourceRequest method. +// req, resp := client.RemoveTagsFromResourceRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RemoveTagsFromResource +func (c *Neptune) RemoveTagsFromResourceRequest(input *RemoveTagsFromResourceInput) (req *request.Request, output *RemoveTagsFromResourceOutput) { + op := &request.Operation{ + Name: opRemoveTagsFromResource, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RemoveTagsFromResourceInput{} + } + + output = &RemoveTagsFromResourceOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(query.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// RemoveTagsFromResource API operation for Amazon Neptune. +// +// Removes metadata tags from an Amazon Neptune resource. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation RemoveTagsFromResource for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBInstanceNotFoundFault "DBInstanceNotFound" +// DBInstanceIdentifier does not refer to an existing DB instance. +// +// * ErrCodeDBSnapshotNotFoundFault "DBSnapshotNotFound" +// DBSnapshotIdentifier does not refer to an existing DB snapshot. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RemoveTagsFromResource +func (c *Neptune) RemoveTagsFromResource(input *RemoveTagsFromResourceInput) (*RemoveTagsFromResourceOutput, error) { + req, out := c.RemoveTagsFromResourceRequest(input) + return out, req.Send() +} + +// RemoveTagsFromResourceWithContext is the same as RemoveTagsFromResource with the addition of +// the ability to pass a context and additional request options. +// +// See RemoveTagsFromResource for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) RemoveTagsFromResourceWithContext(ctx aws.Context, input *RemoveTagsFromResourceInput, opts ...request.Option) (*RemoveTagsFromResourceOutput, error) { + req, out := c.RemoveTagsFromResourceRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opResetDBClusterParameterGroup = "ResetDBClusterParameterGroup" + +// ResetDBClusterParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the ResetDBClusterParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ResetDBClusterParameterGroup for more information on using the ResetDBClusterParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ResetDBClusterParameterGroupRequest method. +// req, resp := client.ResetDBClusterParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ResetDBClusterParameterGroup +func (c *Neptune) ResetDBClusterParameterGroupRequest(input *ResetDBClusterParameterGroupInput) (req *request.Request, output *ResetDBClusterParameterGroupOutput) { + op := &request.Operation{ + Name: opResetDBClusterParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ResetDBClusterParameterGroupInput{} + } + + output = &ResetDBClusterParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// ResetDBClusterParameterGroup API operation for Amazon Neptune. +// +// Modifies the parameters of a DB cluster parameter group to the default value. +// To reset specific parameters submit a list of the following: ParameterName +// and ApplyMethod. To reset the entire DB cluster parameter group, specify +// the DBClusterParameterGroupName and ResetAllParameters parameters. +// +// When resetting the entire group, dynamic parameters are updated immediately +// and static parameters are set to pending-reboot to take effect on the next +// DB instance restart or RebootDBInstance request. You must call RebootDBInstance +// for every DB instance in your DB cluster that you want the updated static +// parameter to apply to. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ResetDBClusterParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBParameterGroupStateFault "InvalidDBParameterGroupState" +// The DB parameter group is in use or is in an invalid state. If you are attempting +// to delete the parameter group, you cannot delete it when the parameter group +// is in this state. +// +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ResetDBClusterParameterGroup +func (c *Neptune) ResetDBClusterParameterGroup(input *ResetDBClusterParameterGroupInput) (*ResetDBClusterParameterGroupOutput, error) { + req, out := c.ResetDBClusterParameterGroupRequest(input) + return out, req.Send() +} + +// ResetDBClusterParameterGroupWithContext is the same as ResetDBClusterParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See ResetDBClusterParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ResetDBClusterParameterGroupWithContext(ctx aws.Context, input *ResetDBClusterParameterGroupInput, opts ...request.Option) (*ResetDBClusterParameterGroupOutput, error) { + req, out := c.ResetDBClusterParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opResetDBParameterGroup = "ResetDBParameterGroup" + +// ResetDBParameterGroupRequest generates a "aws/request.Request" representing the +// client's request for the ResetDBParameterGroup operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ResetDBParameterGroup for more information on using the ResetDBParameterGroup +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ResetDBParameterGroupRequest method. +// req, resp := client.ResetDBParameterGroupRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ResetDBParameterGroup +func (c *Neptune) ResetDBParameterGroupRequest(input *ResetDBParameterGroupInput) (req *request.Request, output *ResetDBParameterGroupOutput) { + op := &request.Operation{ + Name: opResetDBParameterGroup, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &ResetDBParameterGroupInput{} + } + + output = &ResetDBParameterGroupOutput{} + req = c.newRequest(op, input, output) + return +} + +// ResetDBParameterGroup API operation for Amazon Neptune. +// +// Modifies the parameters of a DB parameter group to the engine/system default +// value. To reset specific parameters, provide a list of the following: ParameterName +// and ApplyMethod. To reset the entire DB parameter group, specify the DBParameterGroup +// name and ResetAllParameters parameters. When resetting the entire group, +// dynamic parameters are updated immediately and static parameters are set +// to pending-reboot to take effect on the next DB instance restart or RebootDBInstance +// request. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation ResetDBParameterGroup for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidDBParameterGroupStateFault "InvalidDBParameterGroupState" +// The DB parameter group is in use or is in an invalid state. If you are attempting +// to delete the parameter group, you cannot delete it when the parameter group +// is in this state. +// +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName does not refer to an existing DB parameter group. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/ResetDBParameterGroup +func (c *Neptune) ResetDBParameterGroup(input *ResetDBParameterGroupInput) (*ResetDBParameterGroupOutput, error) { + req, out := c.ResetDBParameterGroupRequest(input) + return out, req.Send() +} + +// ResetDBParameterGroupWithContext is the same as ResetDBParameterGroup with the addition of +// the ability to pass a context and additional request options. +// +// See ResetDBParameterGroup for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) ResetDBParameterGroupWithContext(ctx aws.Context, input *ResetDBParameterGroupInput, opts ...request.Option) (*ResetDBParameterGroupOutput, error) { + req, out := c.ResetDBParameterGroupRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opRestoreDBClusterFromSnapshot = "RestoreDBClusterFromSnapshot" + +// RestoreDBClusterFromSnapshotRequest generates a "aws/request.Request" representing the +// client's request for the RestoreDBClusterFromSnapshot operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RestoreDBClusterFromSnapshot for more information on using the RestoreDBClusterFromSnapshot +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RestoreDBClusterFromSnapshotRequest method. +// req, resp := client.RestoreDBClusterFromSnapshotRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RestoreDBClusterFromSnapshot +func (c *Neptune) RestoreDBClusterFromSnapshotRequest(input *RestoreDBClusterFromSnapshotInput) (req *request.Request, output *RestoreDBClusterFromSnapshotOutput) { + op := &request.Operation{ + Name: opRestoreDBClusterFromSnapshot, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RestoreDBClusterFromSnapshotInput{} + } + + output = &RestoreDBClusterFromSnapshotOutput{} + req = c.newRequest(op, input, output) + return +} + +// RestoreDBClusterFromSnapshot API operation for Amazon Neptune. +// +// Creates a new DB cluster from a DB snapshot or DB cluster snapshot. +// +// If a DB snapshot is specified, the target DB cluster is created from the +// source DB snapshot with a default configuration and default security group. +// +// If a DB cluster snapshot is specified, the target DB cluster is created from +// the source DB cluster restore point with the same configuration as the original +// source DB cluster, except that the new DB cluster is created with the default +// security group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation RestoreDBClusterFromSnapshot for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterAlreadyExistsFault "DBClusterAlreadyExistsFault" +// User already has a DB cluster with the given identifier. +// +// * ErrCodeDBClusterQuotaExceededFault "DBClusterQuotaExceededFault" +// User attempted to create a new DB cluster and the user has already reached +// the maximum allowed DB cluster quota. +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeDBSnapshotNotFoundFault "DBSnapshotNotFound" +// DBSnapshotIdentifier does not refer to an existing DB snapshot. +// +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// * ErrCodeInsufficientDBClusterCapacityFault "InsufficientDBClusterCapacityFault" +// The DB cluster does not have enough capacity for the current operation. +// +// * ErrCodeInsufficientStorageClusterCapacityFault "InsufficientStorageClusterCapacity" +// There is insufficient storage available for the current action. You may be +// able to resolve this error by updating your subnet group to use different +// Availability Zones that have more storage available. +// +// * ErrCodeInvalidDBSnapshotStateFault "InvalidDBSnapshotState" +// The state of the DB snapshot does not allow deletion. +// +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// * ErrCodeInvalidVPCNetworkStateFault "InvalidVPCNetworkStateFault" +// DB subnet group does not cover all Availability Zones after it is created +// because users' change. +// +// * ErrCodeInvalidRestoreFault "InvalidRestoreFault" +// Cannot restore from vpc backup to non-vpc DB instance. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// * ErrCodeOptionGroupNotFoundFault "OptionGroupNotFoundFault" +// +// * ErrCodeKMSKeyNotAccessibleFault "KMSKeyNotAccessibleFault" +// Error accessing KMS key. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RestoreDBClusterFromSnapshot +func (c *Neptune) RestoreDBClusterFromSnapshot(input *RestoreDBClusterFromSnapshotInput) (*RestoreDBClusterFromSnapshotOutput, error) { + req, out := c.RestoreDBClusterFromSnapshotRequest(input) + return out, req.Send() +} + +// RestoreDBClusterFromSnapshotWithContext is the same as RestoreDBClusterFromSnapshot with the addition of +// the ability to pass a context and additional request options. +// +// See RestoreDBClusterFromSnapshot for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) RestoreDBClusterFromSnapshotWithContext(ctx aws.Context, input *RestoreDBClusterFromSnapshotInput, opts ...request.Option) (*RestoreDBClusterFromSnapshotOutput, error) { + req, out := c.RestoreDBClusterFromSnapshotRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opRestoreDBClusterToPointInTime = "RestoreDBClusterToPointInTime" + +// RestoreDBClusterToPointInTimeRequest generates a "aws/request.Request" representing the +// client's request for the RestoreDBClusterToPointInTime operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See RestoreDBClusterToPointInTime for more information on using the RestoreDBClusterToPointInTime +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the RestoreDBClusterToPointInTimeRequest method. +// req, resp := client.RestoreDBClusterToPointInTimeRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RestoreDBClusterToPointInTime +func (c *Neptune) RestoreDBClusterToPointInTimeRequest(input *RestoreDBClusterToPointInTimeInput) (req *request.Request, output *RestoreDBClusterToPointInTimeOutput) { + op := &request.Operation{ + Name: opRestoreDBClusterToPointInTime, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &RestoreDBClusterToPointInTimeInput{} + } + + output = &RestoreDBClusterToPointInTimeOutput{} + req = c.newRequest(op, input, output) + return +} + +// RestoreDBClusterToPointInTime API operation for Amazon Neptune. +// +// Restores a DB cluster to an arbitrary point in time. Users can restore to +// any point in time before LatestRestorableTime for up to BackupRetentionPeriod +// days. The target DB cluster is created from the source DB cluster with the +// same configuration as the original DB cluster, except that the new DB cluster +// is created with the default DB security group. +// +// This action only restores the DB cluster, not the DB instances for that DB +// cluster. You must invoke the CreateDBInstance action to create DB instances +// for the restored DB cluster, specifying the identifier of the restored DB +// cluster in DBClusterIdentifier. You can create DB instances only after the +// RestoreDBClusterToPointInTime action has completed and the DB cluster is +// available. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Neptune's +// API operation RestoreDBClusterToPointInTime for usage and error information. +// +// Returned Error Codes: +// * ErrCodeDBClusterAlreadyExistsFault "DBClusterAlreadyExistsFault" +// User already has a DB cluster with the given identifier. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier does not refer to an existing DB cluster. +// +// * ErrCodeDBClusterQuotaExceededFault "DBClusterQuotaExceededFault" +// User attempted to create a new DB cluster and the user has already reached +// the maximum allowed DB cluster quota. +// +// * ErrCodeDBClusterSnapshotNotFoundFault "DBClusterSnapshotNotFoundFault" +// DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. +// +// * ErrCodeDBSubnetGroupNotFoundFault "DBSubnetGroupNotFoundFault" +// DBSubnetGroupName does not refer to an existing DB subnet group. +// +// * ErrCodeInsufficientDBClusterCapacityFault "InsufficientDBClusterCapacityFault" +// The DB cluster does not have enough capacity for the current operation. +// +// * ErrCodeInsufficientStorageClusterCapacityFault "InsufficientStorageClusterCapacity" +// There is insufficient storage available for the current action. You may be +// able to resolve this error by updating your subnet group to use different +// Availability Zones that have more storage available. +// +// * ErrCodeInvalidDBClusterSnapshotStateFault "InvalidDBClusterSnapshotStateFault" +// The supplied value is not a valid DB cluster snapshot state. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The DB cluster is not in a valid state. +// +// * ErrCodeInvalidDBSnapshotStateFault "InvalidDBSnapshotState" +// The state of the DB snapshot does not allow deletion. +// +// * ErrCodeInvalidRestoreFault "InvalidRestoreFault" +// Cannot restore from vpc backup to non-vpc DB instance. +// +// * ErrCodeInvalidSubnet "InvalidSubnet" +// The requested subnet is invalid, or multiple subnets were requested that +// are not all in a common VPC. +// +// * ErrCodeInvalidVPCNetworkStateFault "InvalidVPCNetworkStateFault" +// DB subnet group does not cover all Availability Zones after it is created +// because users' change. +// +// * ErrCodeKMSKeyNotAccessibleFault "KMSKeyNotAccessibleFault" +// Error accessing KMS key. +// +// * ErrCodeOptionGroupNotFoundFault "OptionGroupNotFoundFault" +// +// * ErrCodeStorageQuotaExceededFault "StorageQuotaExceeded" +// Request would result in user exceeding the allowed amount of storage available +// across all DB instances. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31/RestoreDBClusterToPointInTime +func (c *Neptune) RestoreDBClusterToPointInTime(input *RestoreDBClusterToPointInTimeInput) (*RestoreDBClusterToPointInTimeOutput, error) { + req, out := c.RestoreDBClusterToPointInTimeRequest(input) + return out, req.Send() +} + +// RestoreDBClusterToPointInTimeWithContext is the same as RestoreDBClusterToPointInTime with the addition of +// the ability to pass a context and additional request options. +// +// See RestoreDBClusterToPointInTime for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) RestoreDBClusterToPointInTimeWithContext(ctx aws.Context, input *RestoreDBClusterToPointInTimeInput, opts ...request.Option) (*RestoreDBClusterToPointInTimeOutput, error) { + req, out := c.RestoreDBClusterToPointInTimeRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +type AddRoleToDBClusterInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster to associate the IAM role with. + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The Amazon Resource Name (ARN) of the IAM role to associate with the Neptune + // DB cluster, for example arn:aws:iam::123456789012:role/NeptuneAccessRole. + // + // RoleArn is a required field + RoleArn *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s AddRoleToDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AddRoleToDBClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AddRoleToDBClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AddRoleToDBClusterInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + if s.RoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("RoleArn")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *AddRoleToDBClusterInput) SetDBClusterIdentifier(v string) *AddRoleToDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +// SetRoleArn sets the RoleArn field's value. +func (s *AddRoleToDBClusterInput) SetRoleArn(v string) *AddRoleToDBClusterInput { + s.RoleArn = &v + return s +} + +type AddRoleToDBClusterOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s AddRoleToDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AddRoleToDBClusterOutput) GoString() string { + return s.String() +} + +type AddSourceIdentifierToSubscriptionInput struct { + _ struct{} `type:"structure"` + + // The identifier of the event source to be added. + // + // Constraints: + // + // * If the source type is a DB instance, then a DBInstanceIdentifier must + // be supplied. + // + // * If the source type is a DB security group, a DBSecurityGroupName must + // be supplied. + // + // * If the source type is a DB parameter group, a DBParameterGroupName must + // be supplied. + // + // * If the source type is a DB snapshot, a DBSnapshotIdentifier must be + // supplied. + // + // SourceIdentifier is a required field + SourceIdentifier *string `type:"string" required:"true"` + + // The name of the event notification subscription you want to add a source + // identifier to. + // + // SubscriptionName is a required field + SubscriptionName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s AddSourceIdentifierToSubscriptionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AddSourceIdentifierToSubscriptionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AddSourceIdentifierToSubscriptionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AddSourceIdentifierToSubscriptionInput"} + if s.SourceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SourceIdentifier")) + } + if s.SubscriptionName == nil { + invalidParams.Add(request.NewErrParamRequired("SubscriptionName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSourceIdentifier sets the SourceIdentifier field's value. +func (s *AddSourceIdentifierToSubscriptionInput) SetSourceIdentifier(v string) *AddSourceIdentifierToSubscriptionInput { + s.SourceIdentifier = &v + return s +} + +// SetSubscriptionName sets the SubscriptionName field's value. +func (s *AddSourceIdentifierToSubscriptionInput) SetSubscriptionName(v string) *AddSourceIdentifierToSubscriptionInput { + s.SubscriptionName = &v + return s +} + +type AddSourceIdentifierToSubscriptionOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful invocation of the DescribeEventSubscriptions + // action. + EventSubscription *EventSubscription `type:"structure"` +} + +// String returns the string representation +func (s AddSourceIdentifierToSubscriptionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AddSourceIdentifierToSubscriptionOutput) GoString() string { + return s.String() +} + +// SetEventSubscription sets the EventSubscription field's value. +func (s *AddSourceIdentifierToSubscriptionOutput) SetEventSubscription(v *EventSubscription) *AddSourceIdentifierToSubscriptionOutput { + s.EventSubscription = v + return s +} + +type AddTagsToResourceInput struct { + _ struct{} `type:"structure"` + + // The Amazon Neptune resource that the tags are added to. This value is an + // Amazon Resource Name (ARN). For information about creating an ARN, see Constructing + // an Amazon Resource Name (ARN) (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html#tagging.ARN.Constructing). + // + // ResourceName is a required field + ResourceName *string `type:"string" required:"true"` + + // The tags to be assigned to the Amazon Neptune resource. + // + // Tags is a required field + Tags []*Tag `locationNameList:"Tag" type:"list" required:"true"` +} + +// String returns the string representation +func (s AddTagsToResourceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AddTagsToResourceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AddTagsToResourceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AddTagsToResourceInput"} + if s.ResourceName == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceName")) + } + if s.Tags == nil { + invalidParams.Add(request.NewErrParamRequired("Tags")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetResourceName sets the ResourceName field's value. +func (s *AddTagsToResourceInput) SetResourceName(v string) *AddTagsToResourceInput { + s.ResourceName = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *AddTagsToResourceInput) SetTags(v []*Tag) *AddTagsToResourceInput { + s.Tags = v + return s +} + +type AddTagsToResourceOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s AddTagsToResourceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AddTagsToResourceOutput) GoString() string { + return s.String() +} + +type ApplyPendingMaintenanceActionInput struct { + _ struct{} `type:"structure"` + + // The pending maintenance action to apply to this resource. + // + // Valid values: system-update, db-upgrade + // + // ApplyAction is a required field + ApplyAction *string `type:"string" required:"true"` + + // A value that specifies the type of opt-in request, or undoes an opt-in request. + // An opt-in request of type immediate can't be undone. + // + // Valid values: + // + // * immediate - Apply the maintenance action immediately. + // + // * next-maintenance - Apply the maintenance action during the next maintenance + // window for the resource. + // + // * undo-opt-in - Cancel any existing next-maintenance opt-in requests. + // + // OptInType is a required field + OptInType *string `type:"string" required:"true"` + + // The Amazon Resource Name (ARN) of the resource that the pending maintenance + // action applies to. For information about creating an ARN, see Constructing + // an Amazon Resource Name (ARN) (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html#tagging.ARN.Constructing). + // + // ResourceIdentifier is a required field + ResourceIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s ApplyPendingMaintenanceActionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ApplyPendingMaintenanceActionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ApplyPendingMaintenanceActionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ApplyPendingMaintenanceActionInput"} + if s.ApplyAction == nil { + invalidParams.Add(request.NewErrParamRequired("ApplyAction")) + } + if s.OptInType == nil { + invalidParams.Add(request.NewErrParamRequired("OptInType")) + } + if s.ResourceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetApplyAction sets the ApplyAction field's value. +func (s *ApplyPendingMaintenanceActionInput) SetApplyAction(v string) *ApplyPendingMaintenanceActionInput { + s.ApplyAction = &v + return s +} + +// SetOptInType sets the OptInType field's value. +func (s *ApplyPendingMaintenanceActionInput) SetOptInType(v string) *ApplyPendingMaintenanceActionInput { + s.OptInType = &v + return s +} + +// SetResourceIdentifier sets the ResourceIdentifier field's value. +func (s *ApplyPendingMaintenanceActionInput) SetResourceIdentifier(v string) *ApplyPendingMaintenanceActionInput { + s.ResourceIdentifier = &v + return s +} + +type ApplyPendingMaintenanceActionOutput struct { + _ struct{} `type:"structure"` + + // Describes the pending maintenance actions for a resource. + ResourcePendingMaintenanceActions *ResourcePendingMaintenanceActions `type:"structure"` +} + +// String returns the string representation +func (s ApplyPendingMaintenanceActionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ApplyPendingMaintenanceActionOutput) GoString() string { + return s.String() +} + +// SetResourcePendingMaintenanceActions sets the ResourcePendingMaintenanceActions field's value. +func (s *ApplyPendingMaintenanceActionOutput) SetResourcePendingMaintenanceActions(v *ResourcePendingMaintenanceActions) *ApplyPendingMaintenanceActionOutput { + s.ResourcePendingMaintenanceActions = v + return s +} + +// Contains Availability Zone information. +// +// This data type is used as an element in the following data type: +// +// * OrderableDBInstanceOption +type AvailabilityZone struct { + _ struct{} `type:"structure"` + + // The name of the availability zone. + Name *string `type:"string"` +} + +// String returns the string representation +func (s AvailabilityZone) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AvailabilityZone) GoString() string { + return s.String() +} + +// SetName sets the Name field's value. +func (s *AvailabilityZone) SetName(v string) *AvailabilityZone { + s.Name = &v + return s +} + +// This data type is used as a response element in the action DescribeDBEngineVersions. +type CharacterSet struct { + _ struct{} `type:"structure"` + + // The description of the character set. + CharacterSetDescription *string `type:"string"` + + // The name of the character set. + CharacterSetName *string `type:"string"` +} + +// String returns the string representation +func (s CharacterSet) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CharacterSet) GoString() string { + return s.String() +} + +// SetCharacterSetDescription sets the CharacterSetDescription field's value. +func (s *CharacterSet) SetCharacterSetDescription(v string) *CharacterSet { + s.CharacterSetDescription = &v + return s +} + +// SetCharacterSetName sets the CharacterSetName field's value. +func (s *CharacterSet) SetCharacterSetName(v string) *CharacterSet { + s.CharacterSetName = &v + return s +} + +// The configuration setting for the log types to be enabled for export to CloudWatch +// Logs for a specific DB instance or DB cluster. +type CloudwatchLogsExportConfiguration struct { + _ struct{} `type:"structure"` + + // The list of log types to disable. + DisableLogTypes []*string `type:"list"` + + // The list of log types to enable. + EnableLogTypes []*string `type:"list"` +} + +// String returns the string representation +func (s CloudwatchLogsExportConfiguration) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CloudwatchLogsExportConfiguration) GoString() string { + return s.String() +} + +// SetDisableLogTypes sets the DisableLogTypes field's value. +func (s *CloudwatchLogsExportConfiguration) SetDisableLogTypes(v []*string) *CloudwatchLogsExportConfiguration { + s.DisableLogTypes = v + return s +} + +// SetEnableLogTypes sets the EnableLogTypes field's value. +func (s *CloudwatchLogsExportConfiguration) SetEnableLogTypes(v []*string) *CloudwatchLogsExportConfiguration { + s.EnableLogTypes = v + return s +} + +type CopyDBClusterParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The identifier or Amazon Resource Name (ARN) for the source DB cluster parameter + // group. For information about creating an ARN, see Constructing an Amazon + // Resource Name (ARN) (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html#tagging.ARN.Constructing). + // + // Constraints: + // + // * Must specify a valid DB cluster parameter group. + // + // * If the source DB cluster parameter group is in the same AWS Region as + // the copy, specify a valid DB parameter group identifier, for example my-db-cluster-param-group, + // or a valid ARN. + // + // * If the source DB parameter group is in a different AWS Region than the + // copy, specify a valid DB cluster parameter group ARN, for example arn:aws:rds:us-east-1:123456789012:cluster-pg:custom-cluster-group1. + // + // SourceDBClusterParameterGroupIdentifier is a required field + SourceDBClusterParameterGroupIdentifier *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // A description for the copied DB cluster parameter group. + // + // TargetDBClusterParameterGroupDescription is a required field + TargetDBClusterParameterGroupDescription *string `type:"string" required:"true"` + + // The identifier for the copied DB cluster parameter group. + // + // Constraints: + // + // * Cannot be null, empty, or blank + // + // * Must contain from 1 to 255 letters, numbers, or hyphens + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // Example: my-cluster-param-group1 + // + // TargetDBClusterParameterGroupIdentifier is a required field + TargetDBClusterParameterGroupIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s CopyDBClusterParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CopyDBClusterParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CopyDBClusterParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CopyDBClusterParameterGroupInput"} + if s.SourceDBClusterParameterGroupIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SourceDBClusterParameterGroupIdentifier")) + } + if s.TargetDBClusterParameterGroupDescription == nil { + invalidParams.Add(request.NewErrParamRequired("TargetDBClusterParameterGroupDescription")) + } + if s.TargetDBClusterParameterGroupIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("TargetDBClusterParameterGroupIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSourceDBClusterParameterGroupIdentifier sets the SourceDBClusterParameterGroupIdentifier field's value. +func (s *CopyDBClusterParameterGroupInput) SetSourceDBClusterParameterGroupIdentifier(v string) *CopyDBClusterParameterGroupInput { + s.SourceDBClusterParameterGroupIdentifier = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CopyDBClusterParameterGroupInput) SetTags(v []*Tag) *CopyDBClusterParameterGroupInput { + s.Tags = v + return s +} + +// SetTargetDBClusterParameterGroupDescription sets the TargetDBClusterParameterGroupDescription field's value. +func (s *CopyDBClusterParameterGroupInput) SetTargetDBClusterParameterGroupDescription(v string) *CopyDBClusterParameterGroupInput { + s.TargetDBClusterParameterGroupDescription = &v + return s +} + +// SetTargetDBClusterParameterGroupIdentifier sets the TargetDBClusterParameterGroupIdentifier field's value. +func (s *CopyDBClusterParameterGroupInput) SetTargetDBClusterParameterGroupIdentifier(v string) *CopyDBClusterParameterGroupInput { + s.TargetDBClusterParameterGroupIdentifier = &v + return s +} + +type CopyDBClusterParameterGroupOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster parameter group. + // + // This data type is used as a response element in the DescribeDBClusterParameterGroups + // action. + DBClusterParameterGroup *DBClusterParameterGroup `type:"structure"` +} + +// String returns the string representation +func (s CopyDBClusterParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CopyDBClusterParameterGroupOutput) GoString() string { + return s.String() +} + +// SetDBClusterParameterGroup sets the DBClusterParameterGroup field's value. +func (s *CopyDBClusterParameterGroupOutput) SetDBClusterParameterGroup(v *DBClusterParameterGroup) *CopyDBClusterParameterGroupOutput { + s.DBClusterParameterGroup = v + return s +} + +type CopyDBClusterSnapshotInput struct { + _ struct{} `type:"structure"` + + // True to copy all tags from the source DB cluster snapshot to the target DB + // cluster snapshot, and otherwise false. The default is false. + CopyTags *bool `type:"boolean"` + + // The AWS AWS KMS key ID for an encrypted DB cluster snapshot. The KMS key + // ID is the Amazon Resource Name (ARN), KMS key identifier, or the KMS key + // alias for the KMS encryption key. + // + // If you copy an unencrypted DB cluster snapshot and specify a value for the + // KmsKeyId parameter, Amazon Neptune encrypts the target DB cluster snapshot + // using the specified KMS encryption key. + // + // If you copy an encrypted DB cluster snapshot from your AWS account, you can + // specify a value for KmsKeyId to encrypt the copy with a new KMS encryption + // key. If you don't specify a value for KmsKeyId, then the copy of the DB cluster + // snapshot is encrypted with the same KMS key as the source DB cluster snapshot. + // + // If you copy an encrypted DB cluster snapshot that is shared from another + // AWS account, then you must specify a value for KmsKeyId. + // + // To copy an encrypted DB cluster snapshot to another AWS Region, you must + // set KmsKeyId to the KMS key ID you want to use to encrypt the copy of the + // DB cluster snapshot in the destination AWS Region. KMS encryption keys are + // specific to the AWS Region that they are created in, and you can't use encryption + // keys from one AWS Region in another AWS Region. + KmsKeyId *string `type:"string"` + + // The URL that contains a Signature Version 4 signed request for the CopyDBClusterSnapshot + // API action in the AWS Region that contains the source DB cluster snapshot + // to copy. The PreSignedUrl parameter must be used when copying an encrypted + // DB cluster snapshot from another AWS Region. + // + // The pre-signed URL must be a valid request for the CopyDBSClusterSnapshot + // API action that can be executed in the source AWS Region that contains the + // encrypted DB cluster snapshot to be copied. The pre-signed URL request must + // contain the following parameter values: + // + // * KmsKeyId - The AWS KMS key identifier for the key to use to encrypt + // the copy of the DB cluster snapshot in the destination AWS Region. This + // is the same identifier for both the CopyDBClusterSnapshot action that + // is called in the destination AWS Region, and the action contained in the + // pre-signed URL. + // + // * DestinationRegion - The name of the AWS Region that the DB cluster snapshot + // will be created in. + // + // * SourceDBClusterSnapshotIdentifier - The DB cluster snapshot identifier + // for the encrypted DB cluster snapshot to be copied. This identifier must + // be in the Amazon Resource Name (ARN) format for the source AWS Region. + // For example, if you are copying an encrypted DB cluster snapshot from + // the us-west-2 AWS Region, then your SourceDBClusterSnapshotIdentifier + // looks like the following example: arn:aws:rds:us-west-2:123456789012:cluster-snapshot:neptune-cluster1-snapshot-20161115. + // + // To learn how to generate a Signature Version 4 signed request, see Authenticating + // Requests: Using Query Parameters (AWS Signature Version 4) (http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html) + // and Signature Version 4 Signing Process (http://docs.aws.amazon.com/general/latest/gr/signature-version-4.html). + PreSignedUrl *string `type:"string"` + + // The identifier of the DB cluster snapshot to copy. This parameter is not + // case-sensitive. + // + // You can't copy an encrypted, shared DB cluster snapshot from one AWS Region + // to another. + // + // Constraints: + // + // * Must specify a valid system snapshot in the "available" state. + // + // * If the source snapshot is in the same AWS Region as the copy, specify + // a valid DB snapshot identifier. + // + // * If the source snapshot is in a different AWS Region than the copy, specify + // a valid DB cluster snapshot ARN. + // + // Example: my-cluster-snapshot1 + // + // SourceDBClusterSnapshotIdentifier is a required field + SourceDBClusterSnapshotIdentifier *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // The identifier of the new DB cluster snapshot to create from the source DB + // cluster snapshot. This parameter is not case-sensitive. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens. + // + // * First character must be a letter. + // + // * Cannot end with a hyphen or contain two consecutive hyphens. + // + // Example: my-cluster-snapshot2 + // + // TargetDBClusterSnapshotIdentifier is a required field + TargetDBClusterSnapshotIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s CopyDBClusterSnapshotInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CopyDBClusterSnapshotInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CopyDBClusterSnapshotInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CopyDBClusterSnapshotInput"} + if s.SourceDBClusterSnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SourceDBClusterSnapshotIdentifier")) + } + if s.TargetDBClusterSnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("TargetDBClusterSnapshotIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCopyTags sets the CopyTags field's value. +func (s *CopyDBClusterSnapshotInput) SetCopyTags(v bool) *CopyDBClusterSnapshotInput { + s.CopyTags = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *CopyDBClusterSnapshotInput) SetKmsKeyId(v string) *CopyDBClusterSnapshotInput { + s.KmsKeyId = &v + return s +} + +// SetPreSignedUrl sets the PreSignedUrl field's value. +func (s *CopyDBClusterSnapshotInput) SetPreSignedUrl(v string) *CopyDBClusterSnapshotInput { + s.PreSignedUrl = &v + return s +} + +// SetSourceDBClusterSnapshotIdentifier sets the SourceDBClusterSnapshotIdentifier field's value. +func (s *CopyDBClusterSnapshotInput) SetSourceDBClusterSnapshotIdentifier(v string) *CopyDBClusterSnapshotInput { + s.SourceDBClusterSnapshotIdentifier = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CopyDBClusterSnapshotInput) SetTags(v []*Tag) *CopyDBClusterSnapshotInput { + s.Tags = v + return s +} + +// SetTargetDBClusterSnapshotIdentifier sets the TargetDBClusterSnapshotIdentifier field's value. +func (s *CopyDBClusterSnapshotInput) SetTargetDBClusterSnapshotIdentifier(v string) *CopyDBClusterSnapshotInput { + s.TargetDBClusterSnapshotIdentifier = &v + return s +} + +type CopyDBClusterSnapshotOutput struct { + _ struct{} `type:"structure"` + + // Contains the details for an Amazon Neptune DB cluster snapshot + // + // This data type is used as a response element in the DescribeDBClusterSnapshots + // action. + DBClusterSnapshot *DBClusterSnapshot `type:"structure"` +} + +// String returns the string representation +func (s CopyDBClusterSnapshotOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CopyDBClusterSnapshotOutput) GoString() string { + return s.String() +} + +// SetDBClusterSnapshot sets the DBClusterSnapshot field's value. +func (s *CopyDBClusterSnapshotOutput) SetDBClusterSnapshot(v *DBClusterSnapshot) *CopyDBClusterSnapshotOutput { + s.DBClusterSnapshot = v + return s +} + +type CopyDBParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The identifier or ARN for the source DB parameter group. For information + // about creating an ARN, see Constructing an Amazon Resource Name (ARN) (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html#tagging.ARN.Constructing). + // + // Constraints: + // + // * Must specify a valid DB parameter group. + // + // * Must specify a valid DB parameter group identifier, for example my-db-param-group, + // or a valid ARN. + // + // SourceDBParameterGroupIdentifier is a required field + SourceDBParameterGroupIdentifier *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // A description for the copied DB parameter group. + // + // TargetDBParameterGroupDescription is a required field + TargetDBParameterGroupDescription *string `type:"string" required:"true"` + + // The identifier for the copied DB parameter group. + // + // Constraints: + // + // * Cannot be null, empty, or blank + // + // * Must contain from 1 to 255 letters, numbers, or hyphens + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // Example: my-db-parameter-group + // + // TargetDBParameterGroupIdentifier is a required field + TargetDBParameterGroupIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s CopyDBParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CopyDBParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CopyDBParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CopyDBParameterGroupInput"} + if s.SourceDBParameterGroupIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SourceDBParameterGroupIdentifier")) + } + if s.TargetDBParameterGroupDescription == nil { + invalidParams.Add(request.NewErrParamRequired("TargetDBParameterGroupDescription")) + } + if s.TargetDBParameterGroupIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("TargetDBParameterGroupIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSourceDBParameterGroupIdentifier sets the SourceDBParameterGroupIdentifier field's value. +func (s *CopyDBParameterGroupInput) SetSourceDBParameterGroupIdentifier(v string) *CopyDBParameterGroupInput { + s.SourceDBParameterGroupIdentifier = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CopyDBParameterGroupInput) SetTags(v []*Tag) *CopyDBParameterGroupInput { + s.Tags = v + return s +} + +// SetTargetDBParameterGroupDescription sets the TargetDBParameterGroupDescription field's value. +func (s *CopyDBParameterGroupInput) SetTargetDBParameterGroupDescription(v string) *CopyDBParameterGroupInput { + s.TargetDBParameterGroupDescription = &v + return s +} + +// SetTargetDBParameterGroupIdentifier sets the TargetDBParameterGroupIdentifier field's value. +func (s *CopyDBParameterGroupInput) SetTargetDBParameterGroupIdentifier(v string) *CopyDBParameterGroupInput { + s.TargetDBParameterGroupIdentifier = &v + return s +} + +type CopyDBParameterGroupOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB parameter group. + // + // This data type is used as a response element in the DescribeDBParameterGroups + // action. + DBParameterGroup *DBParameterGroup `type:"structure"` +} + +// String returns the string representation +func (s CopyDBParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CopyDBParameterGroupOutput) GoString() string { + return s.String() +} + +// SetDBParameterGroup sets the DBParameterGroup field's value. +func (s *CopyDBParameterGroupOutput) SetDBParameterGroup(v *DBParameterGroup) *CopyDBParameterGroupOutput { + s.DBParameterGroup = v + return s +} + +type CreateDBClusterInput struct { + _ struct{} `type:"structure"` + + // A list of EC2 Availability Zones that instances in the DB cluster can be + // created in. + AvailabilityZones []*string `locationNameList:"AvailabilityZone" type:"list"` + + // The number of days for which automated backups are retained. You must specify + // a minimum value of 1. + // + // Default: 1 + // + // Constraints: + // + // * Must be a value from 1 to 35 + BackupRetentionPeriod *int64 `type:"integer"` + + // A value that indicates that the DB cluster should be associated with the + // specified CharacterSet. + CharacterSetName *string `type:"string"` + + // The DB cluster identifier. This parameter is stored as a lowercase string. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens. + // + // * First character must be a letter. + // + // * Cannot end with a hyphen or contain two consecutive hyphens. + // + // Example: my-cluster1 + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The name of the DB cluster parameter group to associate with this DB cluster. + // If this argument is omitted, the default is used. + // + // Constraints: + // + // * If supplied, must match the name of an existing DBClusterParameterGroup. + DBClusterParameterGroupName *string `type:"string"` + + // A DB subnet group to associate with this DB cluster. + // + // Constraints: Must match the name of an existing DBSubnetGroup. Must not be + // default. + // + // Example: mySubnetgroup + DBSubnetGroupName *string `type:"string"` + + // The name for your database of up to 64 alpha-numeric characters. If you do + // not provide a name, Amazon Neptune will not create a database in the DB cluster + // you are creating. + DatabaseName *string `type:"string"` + + // True to enable mapping of AWS Identity and Access Management (IAM) accounts + // to database accounts, and otherwise false. + // + // Default: false + EnableIAMDatabaseAuthentication *bool `type:"boolean"` + + // The name of the database engine to be used for this DB cluster. + // + // Valid Values: neptune + // + // Engine is a required field + Engine *string `type:"string" required:"true"` + + // The version number of the database engine to use. + // + // Example: 1.0.1 + EngineVersion *string `type:"string"` + + // The AWS KMS key identifier for an encrypted DB cluster. + // + // The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption + // key. If you are creating a DB cluster with the same AWS account that owns + // the KMS encryption key used to encrypt the new DB cluster, then you can use + // the KMS key alias instead of the ARN for the KMS encryption key. + // + // If an encryption key is not specified in KmsKeyId: + // + // * If ReplicationSourceIdentifier identifies an encrypted source, then + // Amazon Neptune will use the encryption key used to encrypt the source. + // Otherwise, Amazon Neptune will use your default encryption key. + // + // * If the StorageEncrypted parameter is true and ReplicationSourceIdentifier + // is not specified, then Amazon Neptune will use your default encryption + // key. + // + // AWS KMS creates the default encryption key for your AWS account. Your AWS + // account has a different default encryption key for each AWS Region. + // + // If you create a Read Replica of an encrypted DB cluster in another AWS Region, + // you must set KmsKeyId to a KMS key ID that is valid in the destination AWS + // Region. This key is used to encrypt the Read Replica in that AWS Region. + KmsKeyId *string `type:"string"` + + // The password for the master database user. This password can contain any + // printable ASCII character except "/", """, or "@". + // + // Constraints: Must contain from 8 to 41 characters. + MasterUserPassword *string `type:"string"` + + // The name of the master user for the DB cluster. + // + // Constraints: + // + // * Must be 1 to 16 letters or numbers. + // + // * First character must be a letter. + // + // * Cannot be a reserved word for the chosen database engine. + MasterUsername *string `type:"string"` + + // A value that indicates that the DB cluster should be associated with the + // specified option group. + // + // Permanent options can't be removed from an option group. The option group + // can't be removed from a DB cluster once it is associated with a DB cluster. + OptionGroupName *string `type:"string"` + + // The port number on which the instances in the DB cluster accept connections. + // + // Default: 8182 + Port *int64 `type:"integer"` + + // A URL that contains a Signature Version 4 signed request for the CreateDBCluster + // action to be called in the source AWS Region where the DB cluster is replicated + // from. You only need to specify PreSignedUrl when you are performing cross-region + // replication from an encrypted DB cluster. + // + // The pre-signed URL must be a valid request for the CreateDBCluster API action + // that can be executed in the source AWS Region that contains the encrypted + // DB cluster to be copied. + // + // The pre-signed URL request must contain the following parameter values: + // + // * KmsKeyId - The AWS KMS key identifier for the key to use to encrypt + // the copy of the DB cluster in the destination AWS Region. This should + // refer to the same KMS key for both the CreateDBCluster action that is + // called in the destination AWS Region, and the action contained in the + // pre-signed URL. + // + // * DestinationRegion - The name of the AWS Region that Read Replica will + // be created in. + // + // * ReplicationSourceIdentifier - The DB cluster identifier for the encrypted + // DB cluster to be copied. This identifier must be in the Amazon Resource + // Name (ARN) format for the source AWS Region. For example, if you are copying + // an encrypted DB cluster from the us-west-2 AWS Region, then your ReplicationSourceIdentifier + // would look like Example: arn:aws:rds:us-west-2:123456789012:cluster:neptune-cluster1. + // + // To learn how to generate a Signature Version 4 signed request, see Authenticating + // Requests: Using Query Parameters (AWS Signature Version 4) (http://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-query-string-auth.html) + // and Signature Version 4 Signing Process (http://docs.aws.amazon.com/general/latest/gr/signature-version-4.html). + PreSignedUrl *string `type:"string"` + + // The daily time range during which automated backups are created if automated + // backups are enabled using the BackupRetentionPeriod parameter. + // + // The default is a 30-minute window selected at random from an 8-hour block + // of time for each AWS Region. To see the time blocks available, see Adjusting + // the Preferred Maintenance Window (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/AdjustingTheMaintenanceWindow.html) + // in the Amazon Neptune User Guide. + // + // Constraints: + // + // * Must be in the format hh24:mi-hh24:mi. + // + // * Must be in Universal Coordinated Time (UTC). + // + // * Must not conflict with the preferred maintenance window. + // + // * Must be at least 30 minutes. + PreferredBackupWindow *string `type:"string"` + + // The weekly time range during which system maintenance can occur, in Universal + // Coordinated Time (UTC). + // + // Format: ddd:hh24:mi-ddd:hh24:mi + // + // The default is a 30-minute window selected at random from an 8-hour block + // of time for each AWS Region, occurring on a random day of the week. To see + // the time blocks available, see Adjusting the Preferred Maintenance Window + // (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/AdjustingTheMaintenanceWindow.html) + // in the Amazon Neptune User Guide. + // + // Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun. + // + // Constraints: Minimum 30-minute window. + PreferredMaintenanceWindow *string `type:"string"` + + // The Amazon Resource Name (ARN) of the source DB instance or DB cluster if + // this DB cluster is created as a Read Replica. + ReplicationSourceIdentifier *string `type:"string"` + + // Specifies whether the DB cluster is encrypted. + StorageEncrypted *bool `type:"boolean"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // A list of EC2 VPC security groups to associate with this DB cluster. + VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` +} + +// String returns the string representation +func (s CreateDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateDBClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateDBClusterInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + if s.Engine == nil { + invalidParams.Add(request.NewErrParamRequired("Engine")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAvailabilityZones sets the AvailabilityZones field's value. +func (s *CreateDBClusterInput) SetAvailabilityZones(v []*string) *CreateDBClusterInput { + s.AvailabilityZones = v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *CreateDBClusterInput) SetBackupRetentionPeriod(v int64) *CreateDBClusterInput { + s.BackupRetentionPeriod = &v + return s +} + +// SetCharacterSetName sets the CharacterSetName field's value. +func (s *CreateDBClusterInput) SetCharacterSetName(v string) *CreateDBClusterInput { + s.CharacterSetName = &v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *CreateDBClusterInput) SetDBClusterIdentifier(v string) *CreateDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *CreateDBClusterInput) SetDBClusterParameterGroupName(v string) *CreateDBClusterInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *CreateDBClusterInput) SetDBSubnetGroupName(v string) *CreateDBClusterInput { + s.DBSubnetGroupName = &v + return s +} + +// SetDatabaseName sets the DatabaseName field's value. +func (s *CreateDBClusterInput) SetDatabaseName(v string) *CreateDBClusterInput { + s.DatabaseName = &v + return s +} + +// SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. +func (s *CreateDBClusterInput) SetEnableIAMDatabaseAuthentication(v bool) *CreateDBClusterInput { + s.EnableIAMDatabaseAuthentication = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *CreateDBClusterInput) SetEngine(v string) *CreateDBClusterInput { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *CreateDBClusterInput) SetEngineVersion(v string) *CreateDBClusterInput { + s.EngineVersion = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *CreateDBClusterInput) SetKmsKeyId(v string) *CreateDBClusterInput { + s.KmsKeyId = &v + return s +} + +// SetMasterUserPassword sets the MasterUserPassword field's value. +func (s *CreateDBClusterInput) SetMasterUserPassword(v string) *CreateDBClusterInput { + s.MasterUserPassword = &v + return s +} + +// SetMasterUsername sets the MasterUsername field's value. +func (s *CreateDBClusterInput) SetMasterUsername(v string) *CreateDBClusterInput { + s.MasterUsername = &v + return s +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *CreateDBClusterInput) SetOptionGroupName(v string) *CreateDBClusterInput { + s.OptionGroupName = &v + return s +} + +// SetPort sets the Port field's value. +func (s *CreateDBClusterInput) SetPort(v int64) *CreateDBClusterInput { + s.Port = &v + return s +} + +// SetPreSignedUrl sets the PreSignedUrl field's value. +func (s *CreateDBClusterInput) SetPreSignedUrl(v string) *CreateDBClusterInput { + s.PreSignedUrl = &v + return s +} + +// SetPreferredBackupWindow sets the PreferredBackupWindow field's value. +func (s *CreateDBClusterInput) SetPreferredBackupWindow(v string) *CreateDBClusterInput { + s.PreferredBackupWindow = &v + return s +} + +// SetPreferredMaintenanceWindow sets the PreferredMaintenanceWindow field's value. +func (s *CreateDBClusterInput) SetPreferredMaintenanceWindow(v string) *CreateDBClusterInput { + s.PreferredMaintenanceWindow = &v + return s +} + +// SetReplicationSourceIdentifier sets the ReplicationSourceIdentifier field's value. +func (s *CreateDBClusterInput) SetReplicationSourceIdentifier(v string) *CreateDBClusterInput { + s.ReplicationSourceIdentifier = &v + return s +} + +// SetStorageEncrypted sets the StorageEncrypted field's value. +func (s *CreateDBClusterInput) SetStorageEncrypted(v bool) *CreateDBClusterInput { + s.StorageEncrypted = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateDBClusterInput) SetTags(v []*Tag) *CreateDBClusterInput { + s.Tags = v + return s +} + +// SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. +func (s *CreateDBClusterInput) SetVpcSecurityGroupIds(v []*string) *CreateDBClusterInput { + s.VpcSecurityGroupIds = v + return s +} + +type CreateDBClusterOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s CreateDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBClusterOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *CreateDBClusterOutput) SetDBCluster(v *DBCluster) *CreateDBClusterOutput { + s.DBCluster = v + return s +} + +type CreateDBClusterParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster parameter group. + // + // Constraints: + // + // * Must match the name of an existing DBClusterParameterGroup. + // + // This value is stored as a lowercase string. + // + // DBClusterParameterGroupName is a required field + DBClusterParameterGroupName *string `type:"string" required:"true"` + + // The DB cluster parameter group family name. A DB cluster parameter group + // can be associated with one and only one DB cluster parameter group family, + // and can be applied only to a DB cluster running a database engine and engine + // version compatible with that DB cluster parameter group family. + // + // DBParameterGroupFamily is a required field + DBParameterGroupFamily *string `type:"string" required:"true"` + + // The description for the DB cluster parameter group. + // + // Description is a required field + Description *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` +} + +// String returns the string representation +func (s CreateDBClusterParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBClusterParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateDBClusterParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateDBClusterParameterGroupInput"} + if s.DBClusterParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterParameterGroupName")) + } + if s.DBParameterGroupFamily == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupFamily")) + } + if s.Description == nil { + invalidParams.Add(request.NewErrParamRequired("Description")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *CreateDBClusterParameterGroupInput) SetDBClusterParameterGroupName(v string) *CreateDBClusterParameterGroupInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *CreateDBClusterParameterGroupInput) SetDBParameterGroupFamily(v string) *CreateDBClusterParameterGroupInput { + s.DBParameterGroupFamily = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *CreateDBClusterParameterGroupInput) SetDescription(v string) *CreateDBClusterParameterGroupInput { + s.Description = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateDBClusterParameterGroupInput) SetTags(v []*Tag) *CreateDBClusterParameterGroupInput { + s.Tags = v + return s +} + +type CreateDBClusterParameterGroupOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster parameter group. + // + // This data type is used as a response element in the DescribeDBClusterParameterGroups + // action. + DBClusterParameterGroup *DBClusterParameterGroup `type:"structure"` +} + +// String returns the string representation +func (s CreateDBClusterParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBClusterParameterGroupOutput) GoString() string { + return s.String() +} + +// SetDBClusterParameterGroup sets the DBClusterParameterGroup field's value. +func (s *CreateDBClusterParameterGroupOutput) SetDBClusterParameterGroup(v *DBClusterParameterGroup) *CreateDBClusterParameterGroupOutput { + s.DBClusterParameterGroup = v + return s +} + +type CreateDBClusterSnapshotInput struct { + _ struct{} `type:"structure"` + + // The identifier of the DB cluster to create a snapshot for. This parameter + // is not case-sensitive. + // + // Constraints: + // + // * Must match the identifier of an existing DBCluster. + // + // Example: my-cluster1 + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The identifier of the DB cluster snapshot. This parameter is stored as a + // lowercase string. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens. + // + // * First character must be a letter. + // + // * Cannot end with a hyphen or contain two consecutive hyphens. + // + // Example: my-cluster1-snapshot1 + // + // DBClusterSnapshotIdentifier is a required field + DBClusterSnapshotIdentifier *string `type:"string" required:"true"` + + // The tags to be assigned to the DB cluster snapshot. + Tags []*Tag `locationNameList:"Tag" type:"list"` +} + +// String returns the string representation +func (s CreateDBClusterSnapshotInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBClusterSnapshotInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateDBClusterSnapshotInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateDBClusterSnapshotInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + if s.DBClusterSnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterSnapshotIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *CreateDBClusterSnapshotInput) SetDBClusterIdentifier(v string) *CreateDBClusterSnapshotInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *CreateDBClusterSnapshotInput) SetDBClusterSnapshotIdentifier(v string) *CreateDBClusterSnapshotInput { + s.DBClusterSnapshotIdentifier = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateDBClusterSnapshotInput) SetTags(v []*Tag) *CreateDBClusterSnapshotInput { + s.Tags = v + return s +} + +type CreateDBClusterSnapshotOutput struct { + _ struct{} `type:"structure"` + + // Contains the details for an Amazon Neptune DB cluster snapshot + // + // This data type is used as a response element in the DescribeDBClusterSnapshots + // action. + DBClusterSnapshot *DBClusterSnapshot `type:"structure"` +} + +// String returns the string representation +func (s CreateDBClusterSnapshotOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBClusterSnapshotOutput) GoString() string { + return s.String() +} + +// SetDBClusterSnapshot sets the DBClusterSnapshot field's value. +func (s *CreateDBClusterSnapshotOutput) SetDBClusterSnapshot(v *DBClusterSnapshot) *CreateDBClusterSnapshotOutput { + s.DBClusterSnapshot = v + return s +} + +type CreateDBInstanceInput struct { + _ struct{} `type:"structure"` + + // The amount of storage (in gibibytes) to allocate for the DB instance. + // + // Type: Integer + // + // Not applicable. Neptune cluster volumes automatically grow as the amount + // of data in your database increases, though you are only charged for the space + // that you use in a Neptune cluster volume. + AllocatedStorage *int64 `type:"integer"` + + // Indicates that minor engine upgrades are applied automatically to the DB + // instance during the maintenance window. + // + // Default: true + AutoMinorVersionUpgrade *bool `type:"boolean"` + + // The EC2 Availability Zone that the DB instance is created in. + // + // Default: A random, system-chosen Availability Zone in the endpoint's AWS + // Region. + // + // Example: us-east-1d + // + // Constraint: The AvailabilityZone parameter can't be specified if the MultiAZ + // parameter is set to true. The specified Availability Zone must be in the + // same AWS Region as the current endpoint. + AvailabilityZone *string `type:"string"` + + // The number of days for which automated backups are retained. + // + // Not applicable. The retention period for automated backups is managed by + // the DB cluster. For more information, see CreateDBCluster. + // + // Default: 1 + // + // Constraints: + // + // * Must be a value from 0 to 35 + // + // * Cannot be set to 0 if the DB instance is a source to Read Replicas + BackupRetentionPeriod *int64 `type:"integer"` + + // Indicates that the DB instance should be associated with the specified CharacterSet. + // + // Not applicable. The character set is managed by the DB cluster. For more + // information, see CreateDBCluster. + CharacterSetName *string `type:"string"` + + // True to copy all tags from the DB instance to snapshots of the DB instance, + // and otherwise false. The default is false. + CopyTagsToSnapshot *bool `type:"boolean"` + + // The identifier of the DB cluster that the instance will belong to. + // + // For information on creating a DB cluster, see CreateDBCluster. + // + // Type: String + DBClusterIdentifier *string `type:"string"` + + // The compute and memory capacity of the DB instance, for example, db.m4.large. + // Not all DB instance classes are available in all AWS Regions. + // + // DBInstanceClass is a required field + DBInstanceClass *string `type:"string" required:"true"` + + // The DB instance identifier. This parameter is stored as a lowercase string. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens. + // + // * First character must be a letter. + // + // * Cannot end with a hyphen or contain two consecutive hyphens. + // + // Example: mydbinstance + // + // DBInstanceIdentifier is a required field + DBInstanceIdentifier *string `type:"string" required:"true"` + + // The database name. + // + // Type: String + DBName *string `type:"string"` + + // The name of the DB parameter group to associate with this DB instance. If + // this argument is omitted, the default DBParameterGroup for the specified + // engine is used. + // + // Constraints: + // + // * Must be 1 to 255 letters, numbers, or hyphens. + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + DBParameterGroupName *string `type:"string"` + + // A list of DB security groups to associate with this DB instance. + // + // Default: The default DB security group for the database engine. + DBSecurityGroups []*string `locationNameList:"DBSecurityGroupName" type:"list"` + + // A DB subnet group to associate with this DB instance. + // + // If there is no DB subnet group, then it is a non-VPC DB instance. + DBSubnetGroupName *string `type:"string"` + + // Specify the Active Directory Domain to create the instance in. + Domain *string `type:"string"` + + // Specify the name of the IAM role to be used when making API calls to the + // Directory Service. + DomainIAMRoleName *string `type:"string"` + + // The list of log types that need to be enabled for exporting to CloudWatch + // Logs. + EnableCloudwatchLogsExports []*string `type:"list"` + + // True to enable AWS Identity and Access Management (IAM) authentication for + // Neptune. + // + // Default: false + EnableIAMDatabaseAuthentication *bool `type:"boolean"` + + // True to enable Performance Insights for the DB instance, and otherwise false. + EnablePerformanceInsights *bool `type:"boolean"` + + // The name of the database engine to be used for this instance. + // + // Valid Values: neptune + // + // Engine is a required field + Engine *string `type:"string" required:"true"` + + // The version number of the database engine to use. + EngineVersion *string `type:"string"` + + // The amount of Provisioned IOPS (input/output operations per second) to be + // initially allocated for the DB instance. + Iops *int64 `type:"integer"` + + // The AWS KMS key identifier for an encrypted DB instance. + // + // The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption + // key. If you are creating a DB instance with the same AWS account that owns + // the KMS encryption key used to encrypt the new DB instance, then you can + // use the KMS key alias instead of the ARN for the KM encryption key. + // + // Not applicable. The KMS key identifier is managed by the DB cluster. For + // more information, see CreateDBCluster. + // + // If the StorageEncrypted parameter is true, and you do not specify a value + // for the KmsKeyId parameter, then Amazon Neptune will use your default encryption + // key. AWS KMS creates the default encryption key for your AWS account. Your + // AWS account has a different default encryption key for each AWS Region. + KmsKeyId *string `type:"string"` + + // License model information for this DB instance. + // + // Valid values: license-included | bring-your-own-license | general-public-license + LicenseModel *string `type:"string"` + + // The password for the master user. The password can include any printable + // ASCII character except "/", """, or "@". + // + // Not used. + MasterUserPassword *string `type:"string"` + + // The name for the master user. Not used. + MasterUsername *string `type:"string"` + + // The interval, in seconds, between points when Enhanced Monitoring metrics + // are collected for the DB instance. To disable collecting Enhanced Monitoring + // metrics, specify 0. The default is 0. + // + // If MonitoringRoleArn is specified, then you must also set MonitoringInterval + // to a value other than 0. + // + // Valid Values: 0, 1, 5, 10, 15, 30, 60 + MonitoringInterval *int64 `type:"integer"` + + // The ARN for the IAM role that permits Neptune to send enhanced monitoring + // metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. + // + // If MonitoringInterval is set to a value other than 0, then you must supply + // a MonitoringRoleArn value. + MonitoringRoleArn *string `type:"string"` + + // Specifies if the DB instance is a Multi-AZ deployment. You can't set the + // AvailabilityZone parameter if the MultiAZ parameter is set to true. + MultiAZ *bool `type:"boolean"` + + // Indicates that the DB instance should be associated with the specified option + // group. + // + // Permanent options, such as the TDE option for Oracle Advanced Security TDE, + // can't be removed from an option group, and that option group can't be removed + // from a DB instance once it is associated with a DB instance + OptionGroupName *string `type:"string"` + + // The AWS KMS key identifier for encryption of Performance Insights data. The + // KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the + // KMS key alias for the KMS encryption key. + PerformanceInsightsKMSKeyId *string `type:"string"` + + // The port number on which the database accepts connections. + // + // Not applicable. The port is managed by the DB cluster. For more information, + // see CreateDBCluster. + // + // Default: 8182 + // + // Type: Integer + Port *int64 `type:"integer"` + + // The daily time range during which automated backups are created. + // + // Not applicable. The daily time range for creating automated backups is managed + // by the DB cluster. For more information, see CreateDBCluster. + PreferredBackupWindow *string `type:"string"` + + // The time range each week during which system maintenance can occur, in Universal + // Coordinated Time (UTC). + // + // Format: ddd:hh24:mi-ddd:hh24:mi + // + // The default is a 30-minute window selected at random from an 8-hour block + // of time for each AWS Region, occurring on a random day of the week. + // + // Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun. + // + // Constraints: Minimum 30-minute window. + PreferredMaintenanceWindow *string `type:"string"` + + // A value that specifies the order in which an Read Replica is promoted to + // the primary instance after a failure of the existing primary instance. + // + // Default: 1 + // + // Valid Values: 0 - 15 + PromotionTier *int64 `type:"integer"` + + // This parameter is not supported. + PubliclyAccessible *bool `deprecated:"true" type:"boolean"` + + // Specifies whether the DB instance is encrypted. + // + // Not applicable. The encryption for DB instances is managed by the DB cluster. + // For more information, see CreateDBCluster. + // + // Default: false + StorageEncrypted *bool `type:"boolean"` + + // Specifies the storage type to be associated with the DB instance. + // + // Not applicable. Storage is managed by the DB Cluster. + StorageType *string `type:"string"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // The ARN from the key store with which to associate the instance for TDE encryption. + TdeCredentialArn *string `type:"string"` + + // The password for the given ARN from the key store in order to access the + // device. + TdeCredentialPassword *string `type:"string"` + + // The time zone of the DB instance. + Timezone *string `type:"string"` + + // A list of EC2 VPC security groups to associate with this DB instance. + // + // Not applicable. The associated list of EC2 VPC security groups is managed + // by the DB cluster. For more information, see CreateDBCluster. + // + // Default: The default EC2 VPC security group for the DB subnet group's VPC. + VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` +} + +// String returns the string representation +func (s CreateDBInstanceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBInstanceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateDBInstanceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateDBInstanceInput"} + if s.DBInstanceClass == nil { + invalidParams.Add(request.NewErrParamRequired("DBInstanceClass")) + } + if s.DBInstanceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBInstanceIdentifier")) + } + if s.Engine == nil { + invalidParams.Add(request.NewErrParamRequired("Engine")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAllocatedStorage sets the AllocatedStorage field's value. +func (s *CreateDBInstanceInput) SetAllocatedStorage(v int64) *CreateDBInstanceInput { + s.AllocatedStorage = &v + return s +} + +// SetAutoMinorVersionUpgrade sets the AutoMinorVersionUpgrade field's value. +func (s *CreateDBInstanceInput) SetAutoMinorVersionUpgrade(v bool) *CreateDBInstanceInput { + s.AutoMinorVersionUpgrade = &v + return s +} + +// SetAvailabilityZone sets the AvailabilityZone field's value. +func (s *CreateDBInstanceInput) SetAvailabilityZone(v string) *CreateDBInstanceInput { + s.AvailabilityZone = &v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *CreateDBInstanceInput) SetBackupRetentionPeriod(v int64) *CreateDBInstanceInput { + s.BackupRetentionPeriod = &v + return s +} + +// SetCharacterSetName sets the CharacterSetName field's value. +func (s *CreateDBInstanceInput) SetCharacterSetName(v string) *CreateDBInstanceInput { + s.CharacterSetName = &v + return s +} + +// SetCopyTagsToSnapshot sets the CopyTagsToSnapshot field's value. +func (s *CreateDBInstanceInput) SetCopyTagsToSnapshot(v bool) *CreateDBInstanceInput { + s.CopyTagsToSnapshot = &v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *CreateDBInstanceInput) SetDBClusterIdentifier(v string) *CreateDBInstanceInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBInstanceClass sets the DBInstanceClass field's value. +func (s *CreateDBInstanceInput) SetDBInstanceClass(v string) *CreateDBInstanceInput { + s.DBInstanceClass = &v + return s +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *CreateDBInstanceInput) SetDBInstanceIdentifier(v string) *CreateDBInstanceInput { + s.DBInstanceIdentifier = &v + return s +} + +// SetDBName sets the DBName field's value. +func (s *CreateDBInstanceInput) SetDBName(v string) *CreateDBInstanceInput { + s.DBName = &v + return s +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *CreateDBInstanceInput) SetDBParameterGroupName(v string) *CreateDBInstanceInput { + s.DBParameterGroupName = &v + return s +} + +// SetDBSecurityGroups sets the DBSecurityGroups field's value. +func (s *CreateDBInstanceInput) SetDBSecurityGroups(v []*string) *CreateDBInstanceInput { + s.DBSecurityGroups = v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *CreateDBInstanceInput) SetDBSubnetGroupName(v string) *CreateDBInstanceInput { + s.DBSubnetGroupName = &v + return s +} + +// SetDomain sets the Domain field's value. +func (s *CreateDBInstanceInput) SetDomain(v string) *CreateDBInstanceInput { + s.Domain = &v + return s +} + +// SetDomainIAMRoleName sets the DomainIAMRoleName field's value. +func (s *CreateDBInstanceInput) SetDomainIAMRoleName(v string) *CreateDBInstanceInput { + s.DomainIAMRoleName = &v + return s +} + +// SetEnableCloudwatchLogsExports sets the EnableCloudwatchLogsExports field's value. +func (s *CreateDBInstanceInput) SetEnableCloudwatchLogsExports(v []*string) *CreateDBInstanceInput { + s.EnableCloudwatchLogsExports = v + return s +} + +// SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. +func (s *CreateDBInstanceInput) SetEnableIAMDatabaseAuthentication(v bool) *CreateDBInstanceInput { + s.EnableIAMDatabaseAuthentication = &v + return s +} + +// SetEnablePerformanceInsights sets the EnablePerformanceInsights field's value. +func (s *CreateDBInstanceInput) SetEnablePerformanceInsights(v bool) *CreateDBInstanceInput { + s.EnablePerformanceInsights = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *CreateDBInstanceInput) SetEngine(v string) *CreateDBInstanceInput { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *CreateDBInstanceInput) SetEngineVersion(v string) *CreateDBInstanceInput { + s.EngineVersion = &v + return s +} + +// SetIops sets the Iops field's value. +func (s *CreateDBInstanceInput) SetIops(v int64) *CreateDBInstanceInput { + s.Iops = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *CreateDBInstanceInput) SetKmsKeyId(v string) *CreateDBInstanceInput { + s.KmsKeyId = &v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *CreateDBInstanceInput) SetLicenseModel(v string) *CreateDBInstanceInput { + s.LicenseModel = &v + return s +} + +// SetMasterUserPassword sets the MasterUserPassword field's value. +func (s *CreateDBInstanceInput) SetMasterUserPassword(v string) *CreateDBInstanceInput { + s.MasterUserPassword = &v + return s +} + +// SetMasterUsername sets the MasterUsername field's value. +func (s *CreateDBInstanceInput) SetMasterUsername(v string) *CreateDBInstanceInput { + s.MasterUsername = &v + return s +} + +// SetMonitoringInterval sets the MonitoringInterval field's value. +func (s *CreateDBInstanceInput) SetMonitoringInterval(v int64) *CreateDBInstanceInput { + s.MonitoringInterval = &v + return s +} + +// SetMonitoringRoleArn sets the MonitoringRoleArn field's value. +func (s *CreateDBInstanceInput) SetMonitoringRoleArn(v string) *CreateDBInstanceInput { + s.MonitoringRoleArn = &v + return s +} + +// SetMultiAZ sets the MultiAZ field's value. +func (s *CreateDBInstanceInput) SetMultiAZ(v bool) *CreateDBInstanceInput { + s.MultiAZ = &v + return s +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *CreateDBInstanceInput) SetOptionGroupName(v string) *CreateDBInstanceInput { + s.OptionGroupName = &v + return s +} + +// SetPerformanceInsightsKMSKeyId sets the PerformanceInsightsKMSKeyId field's value. +func (s *CreateDBInstanceInput) SetPerformanceInsightsKMSKeyId(v string) *CreateDBInstanceInput { + s.PerformanceInsightsKMSKeyId = &v + return s +} + +// SetPort sets the Port field's value. +func (s *CreateDBInstanceInput) SetPort(v int64) *CreateDBInstanceInput { + s.Port = &v + return s +} + +// SetPreferredBackupWindow sets the PreferredBackupWindow field's value. +func (s *CreateDBInstanceInput) SetPreferredBackupWindow(v string) *CreateDBInstanceInput { + s.PreferredBackupWindow = &v + return s +} + +// SetPreferredMaintenanceWindow sets the PreferredMaintenanceWindow field's value. +func (s *CreateDBInstanceInput) SetPreferredMaintenanceWindow(v string) *CreateDBInstanceInput { + s.PreferredMaintenanceWindow = &v + return s +} + +// SetPromotionTier sets the PromotionTier field's value. +func (s *CreateDBInstanceInput) SetPromotionTier(v int64) *CreateDBInstanceInput { + s.PromotionTier = &v + return s +} + +// SetPubliclyAccessible sets the PubliclyAccessible field's value. +func (s *CreateDBInstanceInput) SetPubliclyAccessible(v bool) *CreateDBInstanceInput { + s.PubliclyAccessible = &v + return s +} + +// SetStorageEncrypted sets the StorageEncrypted field's value. +func (s *CreateDBInstanceInput) SetStorageEncrypted(v bool) *CreateDBInstanceInput { + s.StorageEncrypted = &v + return s +} + +// SetStorageType sets the StorageType field's value. +func (s *CreateDBInstanceInput) SetStorageType(v string) *CreateDBInstanceInput { + s.StorageType = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateDBInstanceInput) SetTags(v []*Tag) *CreateDBInstanceInput { + s.Tags = v + return s +} + +// SetTdeCredentialArn sets the TdeCredentialArn field's value. +func (s *CreateDBInstanceInput) SetTdeCredentialArn(v string) *CreateDBInstanceInput { + s.TdeCredentialArn = &v + return s +} + +// SetTdeCredentialPassword sets the TdeCredentialPassword field's value. +func (s *CreateDBInstanceInput) SetTdeCredentialPassword(v string) *CreateDBInstanceInput { + s.TdeCredentialPassword = &v + return s +} + +// SetTimezone sets the Timezone field's value. +func (s *CreateDBInstanceInput) SetTimezone(v string) *CreateDBInstanceInput { + s.Timezone = &v + return s +} + +// SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. +func (s *CreateDBInstanceInput) SetVpcSecurityGroupIds(v []*string) *CreateDBInstanceInput { + s.VpcSecurityGroupIds = v + return s +} + +type CreateDBInstanceOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB instance. + // + // This data type is used as a response element in the DescribeDBInstances action. + DBInstance *DBInstance `type:"structure"` +} + +// String returns the string representation +func (s CreateDBInstanceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBInstanceOutput) GoString() string { + return s.String() +} + +// SetDBInstance sets the DBInstance field's value. +func (s *CreateDBInstanceOutput) SetDBInstance(v *DBInstance) *CreateDBInstanceOutput { + s.DBInstance = v + return s +} + +type CreateDBParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The DB parameter group family name. A DB parameter group can be associated + // with one and only one DB parameter group family, and can be applied only + // to a DB instance running a database engine and engine version compatible + // with that DB parameter group family. + // + // DBParameterGroupFamily is a required field + DBParameterGroupFamily *string `type:"string" required:"true"` + + // The name of the DB parameter group. + // + // Constraints: + // + // * Must be 1 to 255 letters, numbers, or hyphens. + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // This value is stored as a lowercase string. + // + // DBParameterGroupName is a required field + DBParameterGroupName *string `type:"string" required:"true"` + + // The description for the DB parameter group. + // + // Description is a required field + Description *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` +} + +// String returns the string representation +func (s CreateDBParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateDBParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateDBParameterGroupInput"} + if s.DBParameterGroupFamily == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupFamily")) + } + if s.DBParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupName")) + } + if s.Description == nil { + invalidParams.Add(request.NewErrParamRequired("Description")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *CreateDBParameterGroupInput) SetDBParameterGroupFamily(v string) *CreateDBParameterGroupInput { + s.DBParameterGroupFamily = &v + return s +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *CreateDBParameterGroupInput) SetDBParameterGroupName(v string) *CreateDBParameterGroupInput { + s.DBParameterGroupName = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *CreateDBParameterGroupInput) SetDescription(v string) *CreateDBParameterGroupInput { + s.Description = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateDBParameterGroupInput) SetTags(v []*Tag) *CreateDBParameterGroupInput { + s.Tags = v + return s +} + +type CreateDBParameterGroupOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB parameter group. + // + // This data type is used as a response element in the DescribeDBParameterGroups + // action. + DBParameterGroup *DBParameterGroup `type:"structure"` +} + +// String returns the string representation +func (s CreateDBParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBParameterGroupOutput) GoString() string { + return s.String() +} + +// SetDBParameterGroup sets the DBParameterGroup field's value. +func (s *CreateDBParameterGroupOutput) SetDBParameterGroup(v *DBParameterGroup) *CreateDBParameterGroupOutput { + s.DBParameterGroup = v + return s +} + +type CreateDBSubnetGroupInput struct { + _ struct{} `type:"structure"` + + // The description for the DB subnet group. + // + // DBSubnetGroupDescription is a required field + DBSubnetGroupDescription *string `type:"string" required:"true"` + + // The name for the DB subnet group. This value is stored as a lowercase string. + // + // Constraints: Must contain no more than 255 letters, numbers, periods, underscores, + // spaces, or hyphens. Must not be default. + // + // Example: mySubnetgroup + // + // DBSubnetGroupName is a required field + DBSubnetGroupName *string `type:"string" required:"true"` + + // The EC2 Subnet IDs for the DB subnet group. + // + // SubnetIds is a required field + SubnetIds []*string `locationNameList:"SubnetIdentifier" type:"list" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` +} + +// String returns the string representation +func (s CreateDBSubnetGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBSubnetGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateDBSubnetGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateDBSubnetGroupInput"} + if s.DBSubnetGroupDescription == nil { + invalidParams.Add(request.NewErrParamRequired("DBSubnetGroupDescription")) + } + if s.DBSubnetGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBSubnetGroupName")) + } + if s.SubnetIds == nil { + invalidParams.Add(request.NewErrParamRequired("SubnetIds")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBSubnetGroupDescription sets the DBSubnetGroupDescription field's value. +func (s *CreateDBSubnetGroupInput) SetDBSubnetGroupDescription(v string) *CreateDBSubnetGroupInput { + s.DBSubnetGroupDescription = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *CreateDBSubnetGroupInput) SetDBSubnetGroupName(v string) *CreateDBSubnetGroupInput { + s.DBSubnetGroupName = &v + return s +} + +// SetSubnetIds sets the SubnetIds field's value. +func (s *CreateDBSubnetGroupInput) SetSubnetIds(v []*string) *CreateDBSubnetGroupInput { + s.SubnetIds = v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateDBSubnetGroupInput) SetTags(v []*Tag) *CreateDBSubnetGroupInput { + s.Tags = v + return s +} + +type CreateDBSubnetGroupOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB subnet group. + // + // This data type is used as a response element in the DescribeDBSubnetGroups + // action. + DBSubnetGroup *DBSubnetGroup `type:"structure"` +} + +// String returns the string representation +func (s CreateDBSubnetGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateDBSubnetGroupOutput) GoString() string { + return s.String() +} + +// SetDBSubnetGroup sets the DBSubnetGroup field's value. +func (s *CreateDBSubnetGroupOutput) SetDBSubnetGroup(v *DBSubnetGroup) *CreateDBSubnetGroupOutput { + s.DBSubnetGroup = v + return s +} + +type CreateEventSubscriptionInput struct { + _ struct{} `type:"structure"` + + // A Boolean value; set to true to activate the subscription, set to false to + // create the subscription but not active it. + Enabled *bool `type:"boolean"` + + // A list of event categories for a SourceType that you want to subscribe to. + // You can see a list of the categories for a given SourceType by using the + // DescribeEventCategories action. + EventCategories []*string `locationNameList:"EventCategory" type:"list"` + + // The Amazon Resource Name (ARN) of the SNS topic created for event notification. + // The ARN is created by Amazon SNS when you create a topic and subscribe to + // it. + // + // SnsTopicArn is a required field + SnsTopicArn *string `type:"string" required:"true"` + + // The list of identifiers of the event sources for which events are returned. + // If not specified, then all sources are included in the response. An identifier + // must begin with a letter and must contain only ASCII letters, digits, and + // hyphens; it can't end with a hyphen or contain two consecutive hyphens. + // + // Constraints: + // + // * If SourceIds are supplied, SourceType must also be provided. + // + // * If the source type is a DB instance, then a DBInstanceIdentifier must + // be supplied. + // + // * If the source type is a DB security group, a DBSecurityGroupName must + // be supplied. + // + // * If the source type is a DB parameter group, a DBParameterGroupName must + // be supplied. + // + // * If the source type is a DB snapshot, a DBSnapshotIdentifier must be + // supplied. + SourceIds []*string `locationNameList:"SourceId" type:"list"` + + // The type of source that is generating the events. For example, if you want + // to be notified of events generated by a DB instance, you would set this parameter + // to db-instance. if this value is not specified, all events are returned. + // + // Valid values: db-instance | db-cluster | db-parameter-group | db-security-group + // | db-snapshot | db-cluster-snapshot + SourceType *string `type:"string"` + + // The name of the subscription. + // + // Constraints: The name must be less than 255 characters. + // + // SubscriptionName is a required field + SubscriptionName *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` +} + +// String returns the string representation +func (s CreateEventSubscriptionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateEventSubscriptionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateEventSubscriptionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateEventSubscriptionInput"} + if s.SnsTopicArn == nil { + invalidParams.Add(request.NewErrParamRequired("SnsTopicArn")) + } + if s.SubscriptionName == nil { + invalidParams.Add(request.NewErrParamRequired("SubscriptionName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetEnabled sets the Enabled field's value. +func (s *CreateEventSubscriptionInput) SetEnabled(v bool) *CreateEventSubscriptionInput { + s.Enabled = &v + return s +} + +// SetEventCategories sets the EventCategories field's value. +func (s *CreateEventSubscriptionInput) SetEventCategories(v []*string) *CreateEventSubscriptionInput { + s.EventCategories = v + return s +} + +// SetSnsTopicArn sets the SnsTopicArn field's value. +func (s *CreateEventSubscriptionInput) SetSnsTopicArn(v string) *CreateEventSubscriptionInput { + s.SnsTopicArn = &v + return s +} + +// SetSourceIds sets the SourceIds field's value. +func (s *CreateEventSubscriptionInput) SetSourceIds(v []*string) *CreateEventSubscriptionInput { + s.SourceIds = v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *CreateEventSubscriptionInput) SetSourceType(v string) *CreateEventSubscriptionInput { + s.SourceType = &v + return s +} + +// SetSubscriptionName sets the SubscriptionName field's value. +func (s *CreateEventSubscriptionInput) SetSubscriptionName(v string) *CreateEventSubscriptionInput { + s.SubscriptionName = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateEventSubscriptionInput) SetTags(v []*Tag) *CreateEventSubscriptionInput { + s.Tags = v + return s +} + +type CreateEventSubscriptionOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful invocation of the DescribeEventSubscriptions + // action. + EventSubscription *EventSubscription `type:"structure"` +} + +// String returns the string representation +func (s CreateEventSubscriptionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateEventSubscriptionOutput) GoString() string { + return s.String() +} + +// SetEventSubscription sets the EventSubscription field's value. +func (s *CreateEventSubscriptionOutput) SetEventSubscription(v *EventSubscription) *CreateEventSubscriptionOutput { + s.EventSubscription = v + return s +} + +// Contains the details of an Amazon Neptune DB cluster. +// +// This data type is used as a response element in the DescribeDBClusters action. +type DBCluster struct { + _ struct{} `type:"structure"` + + // AllocatedStorage always returns 1, because Neptune DB cluster storage size + // is not fixed, but instead automatically adjusts as needed. + AllocatedStorage *int64 `type:"integer"` + + // Provides a list of the AWS Identity and Access Management (IAM) roles that + // are associated with the DB cluster. IAM roles that are associated with a + // DB cluster grant permission for the DB cluster to access other AWS services + // on your behalf. + AssociatedRoles []*DBClusterRole `locationNameList:"DBClusterRole" type:"list"` + + // Provides the list of EC2 Availability Zones that instances in the DB cluster + // can be created in. + AvailabilityZones []*string `locationNameList:"AvailabilityZone" type:"list"` + + // Specifies the number of days for which automatic DB snapshots are retained. + BackupRetentionPeriod *int64 `type:"integer"` + + // If present, specifies the name of the character set that this cluster is + // associated with. + CharacterSetName *string `type:"string"` + + // Identifies the clone group to which the DB cluster is associated. + CloneGroupId *string `type:"string"` + + // Specifies the time when the DB cluster was created, in Universal Coordinated + // Time (UTC). + ClusterCreateTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // The Amazon Resource Name (ARN) for the DB cluster. + DBClusterArn *string `type:"string"` + + // Contains a user-supplied DB cluster identifier. This identifier is the unique + // key that identifies a DB cluster. + DBClusterIdentifier *string `type:"string"` + + // Provides the list of instances that make up the DB cluster. + DBClusterMembers []*DBClusterMember `locationNameList:"DBClusterMember" type:"list"` + + // Provides the list of option group memberships for this DB cluster. + DBClusterOptionGroupMemberships []*DBClusterOptionGroupStatus `locationNameList:"DBClusterOptionGroup" type:"list"` + + // Specifies the name of the DB cluster parameter group for the DB cluster. + DBClusterParameterGroup *string `type:"string"` + + // Specifies information on the subnet group associated with the DB cluster, + // including the name, description, and subnets in the subnet group. + DBSubnetGroup *string `type:"string"` + + // Contains the name of the initial database of this DB cluster that was provided + // at create time, if one was specified when the DB cluster was created. This + // same name is returned for the life of the DB cluster. + DatabaseName *string `type:"string"` + + // The AWS Region-unique, immutable identifier for the DB cluster. This identifier + // is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB + // cluster is accessed. + DbClusterResourceId *string `type:"string"` + + // Specifies the earliest time to which a database can be restored with point-in-time + // restore. + EarliestRestorableTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Specifies the connection endpoint for the primary instance of the DB cluster. + Endpoint *string `type:"string"` + + // Provides the name of the database engine to be used for this DB cluster. + Engine *string `type:"string"` + + // Indicates the database engine version. + EngineVersion *string `type:"string"` + + // Specifies the ID that Amazon Route 53 assigns when you create a hosted zone. + HostedZoneId *string `type:"string"` + + // True if mapping of AWS Identity and Access Management (IAM) accounts to database + // accounts is enabled, and otherwise false. + IAMDatabaseAuthenticationEnabled *bool `type:"boolean"` + + // If StorageEncrypted is true, the AWS KMS key identifier for the encrypted + // DB cluster. + KmsKeyId *string `type:"string"` + + // Specifies the latest time to which a database can be restored with point-in-time + // restore. + LatestRestorableTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Contains the master username for the DB cluster. + MasterUsername *string `type:"string"` + + // Specifies whether the DB cluster has instances in multiple Availability Zones. + MultiAZ *bool `type:"boolean"` + + // Specifies the progress of the operation as a percentage. + PercentProgress *string `type:"string"` + + // Specifies the port that the database engine is listening on. + Port *int64 `type:"integer"` + + // Specifies the daily time range during which automated backups are created + // if automated backups are enabled, as determined by the BackupRetentionPeriod. + PreferredBackupWindow *string `type:"string"` + + // Specifies the weekly time range during which system maintenance can occur, + // in Universal Coordinated Time (UTC). + PreferredMaintenanceWindow *string `type:"string"` + + // Contains one or more identifiers of the Read Replicas associated with this + // DB cluster. + ReadReplicaIdentifiers []*string `locationNameList:"ReadReplicaIdentifier" type:"list"` + + // The reader endpoint for the DB cluster. The reader endpoint for a DB cluster + // load-balances connections across the Read Replicas that are available in + // a DB cluster. As clients request new connections to the reader endpoint, + // Neptune distributes the connection requests among the Read Replicas in the + // DB cluster. This functionality can help balance your read workload across + // multiple Read Replicas in your DB cluster. + // + // If a failover occurs, and the Read Replica that you are connected to is promoted + // to be the primary instance, your connection is dropped. To continue sending + // your read workload to other Read Replicas in the cluster, you can then reconnect + // to the reader endpoint. + ReaderEndpoint *string `type:"string"` + + // Contains the identifier of the source DB cluster if this DB cluster is a + // Read Replica. + ReplicationSourceIdentifier *string `type:"string"` + + // Specifies the current state of this DB cluster. + Status *string `type:"string"` + + // Specifies whether the DB cluster is encrypted. + StorageEncrypted *bool `type:"boolean"` + + // Provides a list of VPC security groups that the DB cluster belongs to. + VpcSecurityGroups []*VpcSecurityGroupMembership `locationNameList:"VpcSecurityGroupMembership" type:"list"` +} + +// String returns the string representation +func (s DBCluster) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBCluster) GoString() string { + return s.String() +} + +// SetAllocatedStorage sets the AllocatedStorage field's value. +func (s *DBCluster) SetAllocatedStorage(v int64) *DBCluster { + s.AllocatedStorage = &v + return s +} + +// SetAssociatedRoles sets the AssociatedRoles field's value. +func (s *DBCluster) SetAssociatedRoles(v []*DBClusterRole) *DBCluster { + s.AssociatedRoles = v + return s +} + +// SetAvailabilityZones sets the AvailabilityZones field's value. +func (s *DBCluster) SetAvailabilityZones(v []*string) *DBCluster { + s.AvailabilityZones = v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *DBCluster) SetBackupRetentionPeriod(v int64) *DBCluster { + s.BackupRetentionPeriod = &v + return s +} + +// SetCharacterSetName sets the CharacterSetName field's value. +func (s *DBCluster) SetCharacterSetName(v string) *DBCluster { + s.CharacterSetName = &v + return s +} + +// SetCloneGroupId sets the CloneGroupId field's value. +func (s *DBCluster) SetCloneGroupId(v string) *DBCluster { + s.CloneGroupId = &v + return s +} + +// SetClusterCreateTime sets the ClusterCreateTime field's value. +func (s *DBCluster) SetClusterCreateTime(v time.Time) *DBCluster { + s.ClusterCreateTime = &v + return s +} + +// SetDBClusterArn sets the DBClusterArn field's value. +func (s *DBCluster) SetDBClusterArn(v string) *DBCluster { + s.DBClusterArn = &v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *DBCluster) SetDBClusterIdentifier(v string) *DBCluster { + s.DBClusterIdentifier = &v + return s +} + +// SetDBClusterMembers sets the DBClusterMembers field's value. +func (s *DBCluster) SetDBClusterMembers(v []*DBClusterMember) *DBCluster { + s.DBClusterMembers = v + return s +} + +// SetDBClusterOptionGroupMemberships sets the DBClusterOptionGroupMemberships field's value. +func (s *DBCluster) SetDBClusterOptionGroupMemberships(v []*DBClusterOptionGroupStatus) *DBCluster { + s.DBClusterOptionGroupMemberships = v + return s +} + +// SetDBClusterParameterGroup sets the DBClusterParameterGroup field's value. +func (s *DBCluster) SetDBClusterParameterGroup(v string) *DBCluster { + s.DBClusterParameterGroup = &v + return s +} + +// SetDBSubnetGroup sets the DBSubnetGroup field's value. +func (s *DBCluster) SetDBSubnetGroup(v string) *DBCluster { + s.DBSubnetGroup = &v + return s +} + +// SetDatabaseName sets the DatabaseName field's value. +func (s *DBCluster) SetDatabaseName(v string) *DBCluster { + s.DatabaseName = &v + return s +} + +// SetDbClusterResourceId sets the DbClusterResourceId field's value. +func (s *DBCluster) SetDbClusterResourceId(v string) *DBCluster { + s.DbClusterResourceId = &v + return s +} + +// SetEarliestRestorableTime sets the EarliestRestorableTime field's value. +func (s *DBCluster) SetEarliestRestorableTime(v time.Time) *DBCluster { + s.EarliestRestorableTime = &v + return s +} + +// SetEndpoint sets the Endpoint field's value. +func (s *DBCluster) SetEndpoint(v string) *DBCluster { + s.Endpoint = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *DBCluster) SetEngine(v string) *DBCluster { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *DBCluster) SetEngineVersion(v string) *DBCluster { + s.EngineVersion = &v + return s +} + +// SetHostedZoneId sets the HostedZoneId field's value. +func (s *DBCluster) SetHostedZoneId(v string) *DBCluster { + s.HostedZoneId = &v + return s +} + +// SetIAMDatabaseAuthenticationEnabled sets the IAMDatabaseAuthenticationEnabled field's value. +func (s *DBCluster) SetIAMDatabaseAuthenticationEnabled(v bool) *DBCluster { + s.IAMDatabaseAuthenticationEnabled = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *DBCluster) SetKmsKeyId(v string) *DBCluster { + s.KmsKeyId = &v + return s +} + +// SetLatestRestorableTime sets the LatestRestorableTime field's value. +func (s *DBCluster) SetLatestRestorableTime(v time.Time) *DBCluster { + s.LatestRestorableTime = &v + return s +} + +// SetMasterUsername sets the MasterUsername field's value. +func (s *DBCluster) SetMasterUsername(v string) *DBCluster { + s.MasterUsername = &v + return s +} + +// SetMultiAZ sets the MultiAZ field's value. +func (s *DBCluster) SetMultiAZ(v bool) *DBCluster { + s.MultiAZ = &v + return s +} + +// SetPercentProgress sets the PercentProgress field's value. +func (s *DBCluster) SetPercentProgress(v string) *DBCluster { + s.PercentProgress = &v + return s +} + +// SetPort sets the Port field's value. +func (s *DBCluster) SetPort(v int64) *DBCluster { + s.Port = &v + return s +} + +// SetPreferredBackupWindow sets the PreferredBackupWindow field's value. +func (s *DBCluster) SetPreferredBackupWindow(v string) *DBCluster { + s.PreferredBackupWindow = &v + return s +} + +// SetPreferredMaintenanceWindow sets the PreferredMaintenanceWindow field's value. +func (s *DBCluster) SetPreferredMaintenanceWindow(v string) *DBCluster { + s.PreferredMaintenanceWindow = &v + return s +} + +// SetReadReplicaIdentifiers sets the ReadReplicaIdentifiers field's value. +func (s *DBCluster) SetReadReplicaIdentifiers(v []*string) *DBCluster { + s.ReadReplicaIdentifiers = v + return s +} + +// SetReaderEndpoint sets the ReaderEndpoint field's value. +func (s *DBCluster) SetReaderEndpoint(v string) *DBCluster { + s.ReaderEndpoint = &v + return s +} + +// SetReplicationSourceIdentifier sets the ReplicationSourceIdentifier field's value. +func (s *DBCluster) SetReplicationSourceIdentifier(v string) *DBCluster { + s.ReplicationSourceIdentifier = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DBCluster) SetStatus(v string) *DBCluster { + s.Status = &v + return s +} + +// SetStorageEncrypted sets the StorageEncrypted field's value. +func (s *DBCluster) SetStorageEncrypted(v bool) *DBCluster { + s.StorageEncrypted = &v + return s +} + +// SetVpcSecurityGroups sets the VpcSecurityGroups field's value. +func (s *DBCluster) SetVpcSecurityGroups(v []*VpcSecurityGroupMembership) *DBCluster { + s.VpcSecurityGroups = v + return s +} + +// Contains information about an instance that is part of a DB cluster. +type DBClusterMember struct { + _ struct{} `type:"structure"` + + // Specifies the status of the DB cluster parameter group for this member of + // the DB cluster. + DBClusterParameterGroupStatus *string `type:"string"` + + // Specifies the instance identifier for this member of the DB cluster. + DBInstanceIdentifier *string `type:"string"` + + // Value that is true if the cluster member is the primary instance for the + // DB cluster and false otherwise. + IsClusterWriter *bool `type:"boolean"` + + // A value that specifies the order in which a Read Replica is promoted to the + // primary instance after a failure of the existing primary instance. + PromotionTier *int64 `type:"integer"` +} + +// String returns the string representation +func (s DBClusterMember) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterMember) GoString() string { + return s.String() +} + +// SetDBClusterParameterGroupStatus sets the DBClusterParameterGroupStatus field's value. +func (s *DBClusterMember) SetDBClusterParameterGroupStatus(v string) *DBClusterMember { + s.DBClusterParameterGroupStatus = &v + return s +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *DBClusterMember) SetDBInstanceIdentifier(v string) *DBClusterMember { + s.DBInstanceIdentifier = &v + return s +} + +// SetIsClusterWriter sets the IsClusterWriter field's value. +func (s *DBClusterMember) SetIsClusterWriter(v bool) *DBClusterMember { + s.IsClusterWriter = &v + return s +} + +// SetPromotionTier sets the PromotionTier field's value. +func (s *DBClusterMember) SetPromotionTier(v int64) *DBClusterMember { + s.PromotionTier = &v + return s +} + +// Contains status information for a DB cluster option group. +type DBClusterOptionGroupStatus struct { + _ struct{} `type:"structure"` + + // Specifies the name of the DB cluster option group. + DBClusterOptionGroupName *string `type:"string"` + + // Specifies the status of the DB cluster option group. + Status *string `type:"string"` +} + +// String returns the string representation +func (s DBClusterOptionGroupStatus) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterOptionGroupStatus) GoString() string { + return s.String() +} + +// SetDBClusterOptionGroupName sets the DBClusterOptionGroupName field's value. +func (s *DBClusterOptionGroupStatus) SetDBClusterOptionGroupName(v string) *DBClusterOptionGroupStatus { + s.DBClusterOptionGroupName = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DBClusterOptionGroupStatus) SetStatus(v string) *DBClusterOptionGroupStatus { + s.Status = &v + return s +} + +// Contains the details of an Amazon Neptune DB cluster parameter group. +// +// This data type is used as a response element in the DescribeDBClusterParameterGroups +// action. +type DBClusterParameterGroup struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) for the DB cluster parameter group. + DBClusterParameterGroupArn *string `type:"string"` + + // Provides the name of the DB cluster parameter group. + DBClusterParameterGroupName *string `type:"string"` + + // Provides the name of the DB parameter group family that this DB cluster parameter + // group is compatible with. + DBParameterGroupFamily *string `type:"string"` + + // Provides the customer-specified description for this DB cluster parameter + // group. + Description *string `type:"string"` +} + +// String returns the string representation +func (s DBClusterParameterGroup) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterParameterGroup) GoString() string { + return s.String() +} + +// SetDBClusterParameterGroupArn sets the DBClusterParameterGroupArn field's value. +func (s *DBClusterParameterGroup) SetDBClusterParameterGroupArn(v string) *DBClusterParameterGroup { + s.DBClusterParameterGroupArn = &v + return s +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *DBClusterParameterGroup) SetDBClusterParameterGroupName(v string) *DBClusterParameterGroup { + s.DBClusterParameterGroupName = &v + return s +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *DBClusterParameterGroup) SetDBParameterGroupFamily(v string) *DBClusterParameterGroup { + s.DBParameterGroupFamily = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *DBClusterParameterGroup) SetDescription(v string) *DBClusterParameterGroup { + s.Description = &v + return s +} + +// Describes an AWS Identity and Access Management (IAM) role that is associated +// with a DB cluster. +type DBClusterRole struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the IAM role that is associated with the + // DB cluster. + RoleArn *string `type:"string"` + + // Describes the state of association between the IAM role and the DB cluster. + // The Status property returns one of the following values: + // + // * ACTIVE - the IAM role ARN is associated with the DB cluster and can + // be used to access other AWS services on your behalf. + // + // * PENDING - the IAM role ARN is being associated with the DB cluster. + // + // * INVALID - the IAM role ARN is associated with the DB cluster, but the + // DB cluster is unable to assume the IAM role in order to access other AWS + // services on your behalf. + Status *string `type:"string"` +} + +// String returns the string representation +func (s DBClusterRole) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterRole) GoString() string { + return s.String() +} + +// SetRoleArn sets the RoleArn field's value. +func (s *DBClusterRole) SetRoleArn(v string) *DBClusterRole { + s.RoleArn = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DBClusterRole) SetStatus(v string) *DBClusterRole { + s.Status = &v + return s +} + +// Contains the details for an Amazon Neptune DB cluster snapshot +// +// This data type is used as a response element in the DescribeDBClusterSnapshots +// action. +type DBClusterSnapshot struct { + _ struct{} `type:"structure"` + + // Specifies the allocated storage size in gibibytes (GiB). + AllocatedStorage *int64 `type:"integer"` + + // Provides the list of EC2 Availability Zones that instances in the DB cluster + // snapshot can be restored in. + AvailabilityZones []*string `locationNameList:"AvailabilityZone" type:"list"` + + // Specifies the time when the DB cluster was created, in Universal Coordinated + // Time (UTC). + ClusterCreateTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Specifies the DB cluster identifier of the DB cluster that this DB cluster + // snapshot was created from. + DBClusterIdentifier *string `type:"string"` + + // The Amazon Resource Name (ARN) for the DB cluster snapshot. + DBClusterSnapshotArn *string `type:"string"` + + // Specifies the identifier for the DB cluster snapshot. + DBClusterSnapshotIdentifier *string `type:"string"` + + // Specifies the name of the database engine. + Engine *string `type:"string"` + + // Provides the version of the database engine for this DB cluster snapshot. + EngineVersion *string `type:"string"` + + // True if mapping of AWS Identity and Access Management (IAM) accounts to database + // accounts is enabled, and otherwise false. + IAMDatabaseAuthenticationEnabled *bool `type:"boolean"` + + // If StorageEncrypted is true, the AWS KMS key identifier for the encrypted + // DB cluster snapshot. + KmsKeyId *string `type:"string"` + + // Provides the license model information for this DB cluster snapshot. + LicenseModel *string `type:"string"` + + // Provides the master username for the DB cluster snapshot. + MasterUsername *string `type:"string"` + + // Specifies the percentage of the estimated data that has been transferred. + PercentProgress *int64 `type:"integer"` + + // Specifies the port that the DB cluster was listening on at the time of the + // snapshot. + Port *int64 `type:"integer"` + + // Provides the time when the snapshot was taken, in Universal Coordinated Time + // (UTC). + SnapshotCreateTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Provides the type of the DB cluster snapshot. + SnapshotType *string `type:"string"` + + // If the DB cluster snapshot was copied from a source DB cluster snapshot, + // the Amazon Resource Name (ARN) for the source DB cluster snapshot, otherwise, + // a null value. + SourceDBClusterSnapshotArn *string `type:"string"` + + // Specifies the status of this DB cluster snapshot. + Status *string `type:"string"` + + // Specifies whether the DB cluster snapshot is encrypted. + StorageEncrypted *bool `type:"boolean"` + + // Provides the VPC ID associated with the DB cluster snapshot. + VpcId *string `type:"string"` +} + +// String returns the string representation +func (s DBClusterSnapshot) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterSnapshot) GoString() string { + return s.String() +} + +// SetAllocatedStorage sets the AllocatedStorage field's value. +func (s *DBClusterSnapshot) SetAllocatedStorage(v int64) *DBClusterSnapshot { + s.AllocatedStorage = &v + return s +} + +// SetAvailabilityZones sets the AvailabilityZones field's value. +func (s *DBClusterSnapshot) SetAvailabilityZones(v []*string) *DBClusterSnapshot { + s.AvailabilityZones = v + return s +} + +// SetClusterCreateTime sets the ClusterCreateTime field's value. +func (s *DBClusterSnapshot) SetClusterCreateTime(v time.Time) *DBClusterSnapshot { + s.ClusterCreateTime = &v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *DBClusterSnapshot) SetDBClusterIdentifier(v string) *DBClusterSnapshot { + s.DBClusterIdentifier = &v + return s +} + +// SetDBClusterSnapshotArn sets the DBClusterSnapshotArn field's value. +func (s *DBClusterSnapshot) SetDBClusterSnapshotArn(v string) *DBClusterSnapshot { + s.DBClusterSnapshotArn = &v + return s +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *DBClusterSnapshot) SetDBClusterSnapshotIdentifier(v string) *DBClusterSnapshot { + s.DBClusterSnapshotIdentifier = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *DBClusterSnapshot) SetEngine(v string) *DBClusterSnapshot { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *DBClusterSnapshot) SetEngineVersion(v string) *DBClusterSnapshot { + s.EngineVersion = &v + return s +} + +// SetIAMDatabaseAuthenticationEnabled sets the IAMDatabaseAuthenticationEnabled field's value. +func (s *DBClusterSnapshot) SetIAMDatabaseAuthenticationEnabled(v bool) *DBClusterSnapshot { + s.IAMDatabaseAuthenticationEnabled = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *DBClusterSnapshot) SetKmsKeyId(v string) *DBClusterSnapshot { + s.KmsKeyId = &v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *DBClusterSnapshot) SetLicenseModel(v string) *DBClusterSnapshot { + s.LicenseModel = &v + return s +} + +// SetMasterUsername sets the MasterUsername field's value. +func (s *DBClusterSnapshot) SetMasterUsername(v string) *DBClusterSnapshot { + s.MasterUsername = &v + return s +} + +// SetPercentProgress sets the PercentProgress field's value. +func (s *DBClusterSnapshot) SetPercentProgress(v int64) *DBClusterSnapshot { + s.PercentProgress = &v + return s +} + +// SetPort sets the Port field's value. +func (s *DBClusterSnapshot) SetPort(v int64) *DBClusterSnapshot { + s.Port = &v + return s +} + +// SetSnapshotCreateTime sets the SnapshotCreateTime field's value. +func (s *DBClusterSnapshot) SetSnapshotCreateTime(v time.Time) *DBClusterSnapshot { + s.SnapshotCreateTime = &v + return s +} + +// SetSnapshotType sets the SnapshotType field's value. +func (s *DBClusterSnapshot) SetSnapshotType(v string) *DBClusterSnapshot { + s.SnapshotType = &v + return s +} + +// SetSourceDBClusterSnapshotArn sets the SourceDBClusterSnapshotArn field's value. +func (s *DBClusterSnapshot) SetSourceDBClusterSnapshotArn(v string) *DBClusterSnapshot { + s.SourceDBClusterSnapshotArn = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DBClusterSnapshot) SetStatus(v string) *DBClusterSnapshot { + s.Status = &v + return s +} + +// SetStorageEncrypted sets the StorageEncrypted field's value. +func (s *DBClusterSnapshot) SetStorageEncrypted(v bool) *DBClusterSnapshot { + s.StorageEncrypted = &v + return s +} + +// SetVpcId sets the VpcId field's value. +func (s *DBClusterSnapshot) SetVpcId(v string) *DBClusterSnapshot { + s.VpcId = &v + return s +} + +// Contains the name and values of a manual DB cluster snapshot attribute. +// +// Manual DB cluster snapshot attributes are used to authorize other AWS accounts +// to restore a manual DB cluster snapshot. For more information, see the ModifyDBClusterSnapshotAttribute +// API action. +type DBClusterSnapshotAttribute struct { + _ struct{} `type:"structure"` + + // The name of the manual DB cluster snapshot attribute. + // + // The attribute named restore refers to the list of AWS accounts that have + // permission to copy or restore the manual DB cluster snapshot. For more information, + // see the ModifyDBClusterSnapshotAttribute API action. + AttributeName *string `type:"string"` + + // The value(s) for the manual DB cluster snapshot attribute. + // + // If the AttributeName field is set to restore, then this element returns a + // list of IDs of the AWS accounts that are authorized to copy or restore the + // manual DB cluster snapshot. If a value of all is in the list, then the manual + // DB cluster snapshot is public and available for any AWS account to copy or + // restore. + AttributeValues []*string `locationNameList:"AttributeValue" type:"list"` +} + +// String returns the string representation +func (s DBClusterSnapshotAttribute) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterSnapshotAttribute) GoString() string { + return s.String() +} + +// SetAttributeName sets the AttributeName field's value. +func (s *DBClusterSnapshotAttribute) SetAttributeName(v string) *DBClusterSnapshotAttribute { + s.AttributeName = &v + return s +} + +// SetAttributeValues sets the AttributeValues field's value. +func (s *DBClusterSnapshotAttribute) SetAttributeValues(v []*string) *DBClusterSnapshotAttribute { + s.AttributeValues = v + return s +} + +// Contains the results of a successful call to the DescribeDBClusterSnapshotAttributes +// API action. +// +// Manual DB cluster snapshot attributes are used to authorize other AWS accounts +// to copy or restore a manual DB cluster snapshot. For more information, see +// the ModifyDBClusterSnapshotAttribute API action. +type DBClusterSnapshotAttributesResult struct { + _ struct{} `type:"structure"` + + // The list of attributes and values for the manual DB cluster snapshot. + DBClusterSnapshotAttributes []*DBClusterSnapshotAttribute `locationNameList:"DBClusterSnapshotAttribute" type:"list"` + + // The identifier of the manual DB cluster snapshot that the attributes apply + // to. + DBClusterSnapshotIdentifier *string `type:"string"` +} + +// String returns the string representation +func (s DBClusterSnapshotAttributesResult) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBClusterSnapshotAttributesResult) GoString() string { + return s.String() +} + +// SetDBClusterSnapshotAttributes sets the DBClusterSnapshotAttributes field's value. +func (s *DBClusterSnapshotAttributesResult) SetDBClusterSnapshotAttributes(v []*DBClusterSnapshotAttribute) *DBClusterSnapshotAttributesResult { + s.DBClusterSnapshotAttributes = v + return s +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *DBClusterSnapshotAttributesResult) SetDBClusterSnapshotIdentifier(v string) *DBClusterSnapshotAttributesResult { + s.DBClusterSnapshotIdentifier = &v + return s +} + +// This data type is used as a response element in the action DescribeDBEngineVersions. +type DBEngineVersion struct { + _ struct{} `type:"structure"` + + // The description of the database engine. + DBEngineDescription *string `type:"string"` + + // The description of the database engine version. + DBEngineVersionDescription *string `type:"string"` + + // The name of the DB parameter group family for the database engine. + DBParameterGroupFamily *string `type:"string"` + + // The default character set for new instances of this engine version, if the + // CharacterSetName parameter of the CreateDBInstance API is not specified. + DefaultCharacterSet *CharacterSet `type:"structure"` + + // The name of the database engine. + Engine *string `type:"string"` + + // The version number of the database engine. + EngineVersion *string `type:"string"` + + // The types of logs that the database engine has available for export to CloudWatch + // Logs. + ExportableLogTypes []*string `type:"list"` + + // A list of the character sets supported by this engine for the CharacterSetName + // parameter of the CreateDBInstance action. + SupportedCharacterSets []*CharacterSet `locationNameList:"CharacterSet" type:"list"` + + // A list of the time zones supported by this engine for the Timezone parameter + // of the CreateDBInstance action. + SupportedTimezones []*Timezone `locationNameList:"Timezone" type:"list"` + + // A value that indicates whether the engine version supports exporting the + // log types specified by ExportableLogTypes to CloudWatch Logs. + SupportsLogExportsToCloudwatchLogs *bool `type:"boolean"` + + // Indicates whether the database engine version supports read replicas. + SupportsReadReplica *bool `type:"boolean"` + + // A list of engine versions that this database engine version can be upgraded + // to. + ValidUpgradeTarget []*UpgradeTarget `locationNameList:"UpgradeTarget" type:"list"` +} + +// String returns the string representation +func (s DBEngineVersion) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBEngineVersion) GoString() string { + return s.String() +} + +// SetDBEngineDescription sets the DBEngineDescription field's value. +func (s *DBEngineVersion) SetDBEngineDescription(v string) *DBEngineVersion { + s.DBEngineDescription = &v + return s +} + +// SetDBEngineVersionDescription sets the DBEngineVersionDescription field's value. +func (s *DBEngineVersion) SetDBEngineVersionDescription(v string) *DBEngineVersion { + s.DBEngineVersionDescription = &v + return s +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *DBEngineVersion) SetDBParameterGroupFamily(v string) *DBEngineVersion { + s.DBParameterGroupFamily = &v + return s +} + +// SetDefaultCharacterSet sets the DefaultCharacterSet field's value. +func (s *DBEngineVersion) SetDefaultCharacterSet(v *CharacterSet) *DBEngineVersion { + s.DefaultCharacterSet = v + return s +} + +// SetEngine sets the Engine field's value. +func (s *DBEngineVersion) SetEngine(v string) *DBEngineVersion { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *DBEngineVersion) SetEngineVersion(v string) *DBEngineVersion { + s.EngineVersion = &v + return s +} + +// SetExportableLogTypes sets the ExportableLogTypes field's value. +func (s *DBEngineVersion) SetExportableLogTypes(v []*string) *DBEngineVersion { + s.ExportableLogTypes = v + return s +} + +// SetSupportedCharacterSets sets the SupportedCharacterSets field's value. +func (s *DBEngineVersion) SetSupportedCharacterSets(v []*CharacterSet) *DBEngineVersion { + s.SupportedCharacterSets = v + return s +} + +// SetSupportedTimezones sets the SupportedTimezones field's value. +func (s *DBEngineVersion) SetSupportedTimezones(v []*Timezone) *DBEngineVersion { + s.SupportedTimezones = v + return s +} + +// SetSupportsLogExportsToCloudwatchLogs sets the SupportsLogExportsToCloudwatchLogs field's value. +func (s *DBEngineVersion) SetSupportsLogExportsToCloudwatchLogs(v bool) *DBEngineVersion { + s.SupportsLogExportsToCloudwatchLogs = &v + return s +} + +// SetSupportsReadReplica sets the SupportsReadReplica field's value. +func (s *DBEngineVersion) SetSupportsReadReplica(v bool) *DBEngineVersion { + s.SupportsReadReplica = &v + return s +} + +// SetValidUpgradeTarget sets the ValidUpgradeTarget field's value. +func (s *DBEngineVersion) SetValidUpgradeTarget(v []*UpgradeTarget) *DBEngineVersion { + s.ValidUpgradeTarget = v + return s +} + +// Contains the details of an Amazon Neptune DB instance. +// +// This data type is used as a response element in the DescribeDBInstances action. +type DBInstance struct { + _ struct{} `type:"structure"` + + // Specifies the allocated storage size specified in gibibytes. + AllocatedStorage *int64 `type:"integer"` + + // Indicates that minor version patches are applied automatically. + AutoMinorVersionUpgrade *bool `type:"boolean"` + + // Specifies the name of the Availability Zone the DB instance is located in. + AvailabilityZone *string `type:"string"` + + // Specifies the number of days for which automatic DB snapshots are retained. + BackupRetentionPeriod *int64 `type:"integer"` + + // The identifier of the CA certificate for this DB instance. + CACertificateIdentifier *string `type:"string"` + + // If present, specifies the name of the character set that this instance is + // associated with. + CharacterSetName *string `type:"string"` + + // Specifies whether tags are copied from the DB instance to snapshots of the + // DB instance. + CopyTagsToSnapshot *bool `type:"boolean"` + + // If the DB instance is a member of a DB cluster, contains the name of the + // DB cluster that the DB instance is a member of. + DBClusterIdentifier *string `type:"string"` + + // The Amazon Resource Name (ARN) for the DB instance. + DBInstanceArn *string `type:"string"` + + // Contains the name of the compute and memory capacity class of the DB instance. + DBInstanceClass *string `type:"string"` + + // Contains a user-supplied database identifier. This identifier is the unique + // key that identifies a DB instance. + DBInstanceIdentifier *string `type:"string"` + + // Specifies the current state of this database. + DBInstanceStatus *string `type:"string"` + + // The database name. + DBName *string `type:"string"` + + // Provides the list of DB parameter groups applied to this DB instance. + DBParameterGroups []*DBParameterGroupStatus `locationNameList:"DBParameterGroup" type:"list"` + + // Provides List of DB security group elements containing only DBSecurityGroup.Name + // and DBSecurityGroup.Status subelements. + DBSecurityGroups []*DBSecurityGroupMembership `locationNameList:"DBSecurityGroup" type:"list"` + + // Specifies information on the subnet group associated with the DB instance, + // including the name, description, and subnets in the subnet group. + DBSubnetGroup *DBSubnetGroup `type:"structure"` + + // Specifies the port that the DB instance listens on. If the DB instance is + // part of a DB cluster, this can be a different port than the DB cluster port. + DbInstancePort *int64 `type:"integer"` + + // The AWS Region-unique, immutable identifier for the DB instance. This identifier + // is found in AWS CloudTrail log entries whenever the AWS KMS key for the DB + // instance is accessed. + DbiResourceId *string `type:"string"` + + // Not supported + DomainMemberships []*DomainMembership `locationNameList:"DomainMembership" type:"list"` + + // A list of log types that this DB instance is configured to export to CloudWatch + // Logs. + EnabledCloudwatchLogsExports []*string `type:"list"` + + // Specifies the connection endpoint. + Endpoint *Endpoint `type:"structure"` + + // Provides the name of the database engine to be used for this DB instance. + Engine *string `type:"string"` + + // Indicates the database engine version. + EngineVersion *string `type:"string"` + + // The Amazon Resource Name (ARN) of the Amazon CloudWatch Logs log stream that + // receives the Enhanced Monitoring metrics data for the DB instance. + EnhancedMonitoringResourceArn *string `type:"string"` + + // True if AWS Identity and Access Management (IAM) authentication is enabled, + // and otherwise false. + IAMDatabaseAuthenticationEnabled *bool `type:"boolean"` + + // Provides the date and time the DB instance was created. + InstanceCreateTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Specifies the Provisioned IOPS (I/O operations per second) value. + Iops *int64 `type:"integer"` + + // If StorageEncrypted is true, the AWS KMS key identifier for the encrypted + // DB instance. + KmsKeyId *string `type:"string"` + + // Specifies the latest time to which a database can be restored with point-in-time + // restore. + LatestRestorableTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // License model information for this DB instance. + LicenseModel *string `type:"string"` + + // Contains the master username for the DB instance. + MasterUsername *string `type:"string"` + + // The interval, in seconds, between points when Enhanced Monitoring metrics + // are collected for the DB instance. + MonitoringInterval *int64 `type:"integer"` + + // The ARN for the IAM role that permits Neptune to send Enhanced Monitoring + // metrics to Amazon CloudWatch Logs. + MonitoringRoleArn *string `type:"string"` + + // Specifies if the DB instance is a Multi-AZ deployment. + MultiAZ *bool `type:"boolean"` + + // Provides the list of option group memberships for this DB instance. + OptionGroupMemberships []*OptionGroupMembership `locationNameList:"OptionGroupMembership" type:"list"` + + // Specifies that changes to the DB instance are pending. This element is only + // included when changes are pending. Specific changes are identified by subelements. + PendingModifiedValues *PendingModifiedValues `type:"structure"` + + // True if Performance Insights is enabled for the DB instance, and otherwise + // false. + PerformanceInsightsEnabled *bool `type:"boolean"` + + // The AWS KMS key identifier for encryption of Performance Insights data. The + // KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the + // KMS key alias for the KMS encryption key. + PerformanceInsightsKMSKeyId *string `type:"string"` + + // Specifies the daily time range during which automated backups are created + // if automated backups are enabled, as determined by the BackupRetentionPeriod. + PreferredBackupWindow *string `type:"string"` + + // Specifies the weekly time range during which system maintenance can occur, + // in Universal Coordinated Time (UTC). + PreferredMaintenanceWindow *string `type:"string"` + + // A value that specifies the order in which a Read Replica is promoted to the + // primary instance after a failure of the existing primary instance. + PromotionTier *int64 `type:"integer"` + + // This parameter is not supported. + PubliclyAccessible *bool `deprecated:"true" type:"boolean"` + + // Contains one or more identifiers of DB clusters that are Read Replicas of + // this DB instance. + ReadReplicaDBClusterIdentifiers []*string `locationNameList:"ReadReplicaDBClusterIdentifier" type:"list"` + + // Contains one or more identifiers of the Read Replicas associated with this + // DB instance. + ReadReplicaDBInstanceIdentifiers []*string `locationNameList:"ReadReplicaDBInstanceIdentifier" type:"list"` + + // Contains the identifier of the source DB instance if this DB instance is + // a Read Replica. + ReadReplicaSourceDBInstanceIdentifier *string `type:"string"` + + // If present, specifies the name of the secondary Availability Zone for a DB + // instance with multi-AZ support. + SecondaryAvailabilityZone *string `type:"string"` + + // The status of a Read Replica. If the instance is not a Read Replica, this + // is blank. + StatusInfos []*DBInstanceStatusInfo `locationNameList:"DBInstanceStatusInfo" type:"list"` + + // Specifies whether the DB instance is encrypted. + StorageEncrypted *bool `type:"boolean"` + + // Specifies the storage type associated with DB instance. + StorageType *string `type:"string"` + + // The ARN from the key store with which the instance is associated for TDE + // encryption. + TdeCredentialArn *string `type:"string"` + + // Not supported. + Timezone *string `type:"string"` + + // Provides a list of VPC security group elements that the DB instance belongs + // to. + VpcSecurityGroups []*VpcSecurityGroupMembership `locationNameList:"VpcSecurityGroupMembership" type:"list"` +} + +// String returns the string representation +func (s DBInstance) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBInstance) GoString() string { + return s.String() +} + +// SetAllocatedStorage sets the AllocatedStorage field's value. +func (s *DBInstance) SetAllocatedStorage(v int64) *DBInstance { + s.AllocatedStorage = &v + return s +} + +// SetAutoMinorVersionUpgrade sets the AutoMinorVersionUpgrade field's value. +func (s *DBInstance) SetAutoMinorVersionUpgrade(v bool) *DBInstance { + s.AutoMinorVersionUpgrade = &v + return s +} + +// SetAvailabilityZone sets the AvailabilityZone field's value. +func (s *DBInstance) SetAvailabilityZone(v string) *DBInstance { + s.AvailabilityZone = &v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *DBInstance) SetBackupRetentionPeriod(v int64) *DBInstance { + s.BackupRetentionPeriod = &v + return s +} + +// SetCACertificateIdentifier sets the CACertificateIdentifier field's value. +func (s *DBInstance) SetCACertificateIdentifier(v string) *DBInstance { + s.CACertificateIdentifier = &v + return s +} + +// SetCharacterSetName sets the CharacterSetName field's value. +func (s *DBInstance) SetCharacterSetName(v string) *DBInstance { + s.CharacterSetName = &v + return s +} + +// SetCopyTagsToSnapshot sets the CopyTagsToSnapshot field's value. +func (s *DBInstance) SetCopyTagsToSnapshot(v bool) *DBInstance { + s.CopyTagsToSnapshot = &v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *DBInstance) SetDBClusterIdentifier(v string) *DBInstance { + s.DBClusterIdentifier = &v + return s +} + +// SetDBInstanceArn sets the DBInstanceArn field's value. +func (s *DBInstance) SetDBInstanceArn(v string) *DBInstance { + s.DBInstanceArn = &v + return s +} + +// SetDBInstanceClass sets the DBInstanceClass field's value. +func (s *DBInstance) SetDBInstanceClass(v string) *DBInstance { + s.DBInstanceClass = &v + return s +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *DBInstance) SetDBInstanceIdentifier(v string) *DBInstance { + s.DBInstanceIdentifier = &v + return s +} + +// SetDBInstanceStatus sets the DBInstanceStatus field's value. +func (s *DBInstance) SetDBInstanceStatus(v string) *DBInstance { + s.DBInstanceStatus = &v + return s +} + +// SetDBName sets the DBName field's value. +func (s *DBInstance) SetDBName(v string) *DBInstance { + s.DBName = &v + return s +} + +// SetDBParameterGroups sets the DBParameterGroups field's value. +func (s *DBInstance) SetDBParameterGroups(v []*DBParameterGroupStatus) *DBInstance { + s.DBParameterGroups = v + return s +} + +// SetDBSecurityGroups sets the DBSecurityGroups field's value. +func (s *DBInstance) SetDBSecurityGroups(v []*DBSecurityGroupMembership) *DBInstance { + s.DBSecurityGroups = v + return s +} + +// SetDBSubnetGroup sets the DBSubnetGroup field's value. +func (s *DBInstance) SetDBSubnetGroup(v *DBSubnetGroup) *DBInstance { + s.DBSubnetGroup = v + return s +} + +// SetDbInstancePort sets the DbInstancePort field's value. +func (s *DBInstance) SetDbInstancePort(v int64) *DBInstance { + s.DbInstancePort = &v + return s +} + +// SetDbiResourceId sets the DbiResourceId field's value. +func (s *DBInstance) SetDbiResourceId(v string) *DBInstance { + s.DbiResourceId = &v + return s +} + +// SetDomainMemberships sets the DomainMemberships field's value. +func (s *DBInstance) SetDomainMemberships(v []*DomainMembership) *DBInstance { + s.DomainMemberships = v + return s +} + +// SetEnabledCloudwatchLogsExports sets the EnabledCloudwatchLogsExports field's value. +func (s *DBInstance) SetEnabledCloudwatchLogsExports(v []*string) *DBInstance { + s.EnabledCloudwatchLogsExports = v + return s +} + +// SetEndpoint sets the Endpoint field's value. +func (s *DBInstance) SetEndpoint(v *Endpoint) *DBInstance { + s.Endpoint = v + return s +} + +// SetEngine sets the Engine field's value. +func (s *DBInstance) SetEngine(v string) *DBInstance { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *DBInstance) SetEngineVersion(v string) *DBInstance { + s.EngineVersion = &v + return s +} + +// SetEnhancedMonitoringResourceArn sets the EnhancedMonitoringResourceArn field's value. +func (s *DBInstance) SetEnhancedMonitoringResourceArn(v string) *DBInstance { + s.EnhancedMonitoringResourceArn = &v + return s +} + +// SetIAMDatabaseAuthenticationEnabled sets the IAMDatabaseAuthenticationEnabled field's value. +func (s *DBInstance) SetIAMDatabaseAuthenticationEnabled(v bool) *DBInstance { + s.IAMDatabaseAuthenticationEnabled = &v + return s +} + +// SetInstanceCreateTime sets the InstanceCreateTime field's value. +func (s *DBInstance) SetInstanceCreateTime(v time.Time) *DBInstance { + s.InstanceCreateTime = &v + return s +} + +// SetIops sets the Iops field's value. +func (s *DBInstance) SetIops(v int64) *DBInstance { + s.Iops = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *DBInstance) SetKmsKeyId(v string) *DBInstance { + s.KmsKeyId = &v + return s +} + +// SetLatestRestorableTime sets the LatestRestorableTime field's value. +func (s *DBInstance) SetLatestRestorableTime(v time.Time) *DBInstance { + s.LatestRestorableTime = &v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *DBInstance) SetLicenseModel(v string) *DBInstance { + s.LicenseModel = &v + return s +} + +// SetMasterUsername sets the MasterUsername field's value. +func (s *DBInstance) SetMasterUsername(v string) *DBInstance { + s.MasterUsername = &v + return s +} + +// SetMonitoringInterval sets the MonitoringInterval field's value. +func (s *DBInstance) SetMonitoringInterval(v int64) *DBInstance { + s.MonitoringInterval = &v + return s +} + +// SetMonitoringRoleArn sets the MonitoringRoleArn field's value. +func (s *DBInstance) SetMonitoringRoleArn(v string) *DBInstance { + s.MonitoringRoleArn = &v + return s +} + +// SetMultiAZ sets the MultiAZ field's value. +func (s *DBInstance) SetMultiAZ(v bool) *DBInstance { + s.MultiAZ = &v + return s +} + +// SetOptionGroupMemberships sets the OptionGroupMemberships field's value. +func (s *DBInstance) SetOptionGroupMemberships(v []*OptionGroupMembership) *DBInstance { + s.OptionGroupMemberships = v + return s +} + +// SetPendingModifiedValues sets the PendingModifiedValues field's value. +func (s *DBInstance) SetPendingModifiedValues(v *PendingModifiedValues) *DBInstance { + s.PendingModifiedValues = v + return s +} + +// SetPerformanceInsightsEnabled sets the PerformanceInsightsEnabled field's value. +func (s *DBInstance) SetPerformanceInsightsEnabled(v bool) *DBInstance { + s.PerformanceInsightsEnabled = &v + return s +} + +// SetPerformanceInsightsKMSKeyId sets the PerformanceInsightsKMSKeyId field's value. +func (s *DBInstance) SetPerformanceInsightsKMSKeyId(v string) *DBInstance { + s.PerformanceInsightsKMSKeyId = &v + return s +} + +// SetPreferredBackupWindow sets the PreferredBackupWindow field's value. +func (s *DBInstance) SetPreferredBackupWindow(v string) *DBInstance { + s.PreferredBackupWindow = &v + return s +} + +// SetPreferredMaintenanceWindow sets the PreferredMaintenanceWindow field's value. +func (s *DBInstance) SetPreferredMaintenanceWindow(v string) *DBInstance { + s.PreferredMaintenanceWindow = &v + return s +} + +// SetPromotionTier sets the PromotionTier field's value. +func (s *DBInstance) SetPromotionTier(v int64) *DBInstance { + s.PromotionTier = &v + return s +} + +// SetPubliclyAccessible sets the PubliclyAccessible field's value. +func (s *DBInstance) SetPubliclyAccessible(v bool) *DBInstance { + s.PubliclyAccessible = &v + return s +} + +// SetReadReplicaDBClusterIdentifiers sets the ReadReplicaDBClusterIdentifiers field's value. +func (s *DBInstance) SetReadReplicaDBClusterIdentifiers(v []*string) *DBInstance { + s.ReadReplicaDBClusterIdentifiers = v + return s +} + +// SetReadReplicaDBInstanceIdentifiers sets the ReadReplicaDBInstanceIdentifiers field's value. +func (s *DBInstance) SetReadReplicaDBInstanceIdentifiers(v []*string) *DBInstance { + s.ReadReplicaDBInstanceIdentifiers = v + return s +} + +// SetReadReplicaSourceDBInstanceIdentifier sets the ReadReplicaSourceDBInstanceIdentifier field's value. +func (s *DBInstance) SetReadReplicaSourceDBInstanceIdentifier(v string) *DBInstance { + s.ReadReplicaSourceDBInstanceIdentifier = &v + return s +} + +// SetSecondaryAvailabilityZone sets the SecondaryAvailabilityZone field's value. +func (s *DBInstance) SetSecondaryAvailabilityZone(v string) *DBInstance { + s.SecondaryAvailabilityZone = &v + return s +} + +// SetStatusInfos sets the StatusInfos field's value. +func (s *DBInstance) SetStatusInfos(v []*DBInstanceStatusInfo) *DBInstance { + s.StatusInfos = v + return s +} + +// SetStorageEncrypted sets the StorageEncrypted field's value. +func (s *DBInstance) SetStorageEncrypted(v bool) *DBInstance { + s.StorageEncrypted = &v + return s +} + +// SetStorageType sets the StorageType field's value. +func (s *DBInstance) SetStorageType(v string) *DBInstance { + s.StorageType = &v + return s +} + +// SetTdeCredentialArn sets the TdeCredentialArn field's value. +func (s *DBInstance) SetTdeCredentialArn(v string) *DBInstance { + s.TdeCredentialArn = &v + return s +} + +// SetTimezone sets the Timezone field's value. +func (s *DBInstance) SetTimezone(v string) *DBInstance { + s.Timezone = &v + return s +} + +// SetVpcSecurityGroups sets the VpcSecurityGroups field's value. +func (s *DBInstance) SetVpcSecurityGroups(v []*VpcSecurityGroupMembership) *DBInstance { + s.VpcSecurityGroups = v + return s +} + +// Provides a list of status information for a DB instance. +type DBInstanceStatusInfo struct { + _ struct{} `type:"structure"` + + // Details of the error if there is an error for the instance. If the instance + // is not in an error state, this value is blank. + Message *string `type:"string"` + + // Boolean value that is true if the instance is operating normally, or false + // if the instance is in an error state. + Normal *bool `type:"boolean"` + + // Status of the DB instance. For a StatusType of read replica, the values can + // be replicating, error, stopped, or terminated. + Status *string `type:"string"` + + // This value is currently "read replication." + StatusType *string `type:"string"` +} + +// String returns the string representation +func (s DBInstanceStatusInfo) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBInstanceStatusInfo) GoString() string { + return s.String() +} + +// SetMessage sets the Message field's value. +func (s *DBInstanceStatusInfo) SetMessage(v string) *DBInstanceStatusInfo { + s.Message = &v + return s +} + +// SetNormal sets the Normal field's value. +func (s *DBInstanceStatusInfo) SetNormal(v bool) *DBInstanceStatusInfo { + s.Normal = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DBInstanceStatusInfo) SetStatus(v string) *DBInstanceStatusInfo { + s.Status = &v + return s +} + +// SetStatusType sets the StatusType field's value. +func (s *DBInstanceStatusInfo) SetStatusType(v string) *DBInstanceStatusInfo { + s.StatusType = &v + return s +} + +// Contains the details of an Amazon Neptune DB parameter group. +// +// This data type is used as a response element in the DescribeDBParameterGroups +// action. +type DBParameterGroup struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) for the DB parameter group. + DBParameterGroupArn *string `type:"string"` + + // Provides the name of the DB parameter group family that this DB parameter + // group is compatible with. + DBParameterGroupFamily *string `type:"string"` + + // Provides the name of the DB parameter group. + DBParameterGroupName *string `type:"string"` + + // Provides the customer-specified description for this DB parameter group. + Description *string `type:"string"` +} + +// String returns the string representation +func (s DBParameterGroup) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBParameterGroup) GoString() string { + return s.String() +} + +// SetDBParameterGroupArn sets the DBParameterGroupArn field's value. +func (s *DBParameterGroup) SetDBParameterGroupArn(v string) *DBParameterGroup { + s.DBParameterGroupArn = &v + return s +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *DBParameterGroup) SetDBParameterGroupFamily(v string) *DBParameterGroup { + s.DBParameterGroupFamily = &v + return s +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *DBParameterGroup) SetDBParameterGroupName(v string) *DBParameterGroup { + s.DBParameterGroupName = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *DBParameterGroup) SetDescription(v string) *DBParameterGroup { + s.Description = &v + return s +} + +// The status of the DB parameter group. +// +// This data type is used as a response element in the following actions: +// +// * CreateDBInstance +// +// * DeleteDBInstance +// +// * ModifyDBInstance +// +// * RebootDBInstance +type DBParameterGroupStatus struct { + _ struct{} `type:"structure"` + + // The name of the DP parameter group. + DBParameterGroupName *string `type:"string"` + + // The status of parameter updates. + ParameterApplyStatus *string `type:"string"` +} + +// String returns the string representation +func (s DBParameterGroupStatus) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBParameterGroupStatus) GoString() string { + return s.String() +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *DBParameterGroupStatus) SetDBParameterGroupName(v string) *DBParameterGroupStatus { + s.DBParameterGroupName = &v + return s +} + +// SetParameterApplyStatus sets the ParameterApplyStatus field's value. +func (s *DBParameterGroupStatus) SetParameterApplyStatus(v string) *DBParameterGroupStatus { + s.ParameterApplyStatus = &v + return s +} + +// This data type is used as a response element in the following actions: +// +// * ModifyDBInstance +// +// * RebootDBInstance +type DBSecurityGroupMembership struct { + _ struct{} `type:"structure"` + + // The name of the DB security group. + DBSecurityGroupName *string `type:"string"` + + // The status of the DB security group. + Status *string `type:"string"` +} + +// String returns the string representation +func (s DBSecurityGroupMembership) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBSecurityGroupMembership) GoString() string { + return s.String() +} + +// SetDBSecurityGroupName sets the DBSecurityGroupName field's value. +func (s *DBSecurityGroupMembership) SetDBSecurityGroupName(v string) *DBSecurityGroupMembership { + s.DBSecurityGroupName = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DBSecurityGroupMembership) SetStatus(v string) *DBSecurityGroupMembership { + s.Status = &v + return s +} + +// Contains the details of an Amazon Neptune DB subnet group. +// +// This data type is used as a response element in the DescribeDBSubnetGroups +// action. +type DBSubnetGroup struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) for the DB subnet group. + DBSubnetGroupArn *string `type:"string"` + + // Provides the description of the DB subnet group. + DBSubnetGroupDescription *string `type:"string"` + + // The name of the DB subnet group. + DBSubnetGroupName *string `type:"string"` + + // Provides the status of the DB subnet group. + SubnetGroupStatus *string `type:"string"` + + // Contains a list of Subnet elements. + Subnets []*Subnet `locationNameList:"Subnet" type:"list"` + + // Provides the VpcId of the DB subnet group. + VpcId *string `type:"string"` +} + +// String returns the string representation +func (s DBSubnetGroup) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DBSubnetGroup) GoString() string { + return s.String() +} + +// SetDBSubnetGroupArn sets the DBSubnetGroupArn field's value. +func (s *DBSubnetGroup) SetDBSubnetGroupArn(v string) *DBSubnetGroup { + s.DBSubnetGroupArn = &v + return s +} + +// SetDBSubnetGroupDescription sets the DBSubnetGroupDescription field's value. +func (s *DBSubnetGroup) SetDBSubnetGroupDescription(v string) *DBSubnetGroup { + s.DBSubnetGroupDescription = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *DBSubnetGroup) SetDBSubnetGroupName(v string) *DBSubnetGroup { + s.DBSubnetGroupName = &v + return s +} + +// SetSubnetGroupStatus sets the SubnetGroupStatus field's value. +func (s *DBSubnetGroup) SetSubnetGroupStatus(v string) *DBSubnetGroup { + s.SubnetGroupStatus = &v + return s +} + +// SetSubnets sets the Subnets field's value. +func (s *DBSubnetGroup) SetSubnets(v []*Subnet) *DBSubnetGroup { + s.Subnets = v + return s +} + +// SetVpcId sets the VpcId field's value. +func (s *DBSubnetGroup) SetVpcId(v string) *DBSubnetGroup { + s.VpcId = &v + return s +} + +type DeleteDBClusterInput struct { + _ struct{} `type:"structure"` + + // The DB cluster identifier for the DB cluster to be deleted. This parameter + // isn't case-sensitive. + // + // Constraints: + // + // * Must match an existing DBClusterIdentifier. + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The DB cluster snapshot identifier of the new DB cluster snapshot created + // when SkipFinalSnapshot is set to false. + // + // Specifying this parameter and also setting the SkipFinalShapshot parameter + // to true results in an error. + // + // Constraints: + // + // * Must be 1 to 255 letters, numbers, or hyphens. + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + FinalDBSnapshotIdentifier *string `type:"string"` + + // Determines whether a final DB cluster snapshot is created before the DB cluster + // is deleted. If true is specified, no DB cluster snapshot is created. If false + // is specified, a DB cluster snapshot is created before the DB cluster is deleted. + // + // You must specify a FinalDBSnapshotIdentifier parameter if SkipFinalSnapshot + // is false. + // + // Default: false + SkipFinalSnapshot *bool `type:"boolean"` +} + +// String returns the string representation +func (s DeleteDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteDBClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteDBClusterInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *DeleteDBClusterInput) SetDBClusterIdentifier(v string) *DeleteDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +// SetFinalDBSnapshotIdentifier sets the FinalDBSnapshotIdentifier field's value. +func (s *DeleteDBClusterInput) SetFinalDBSnapshotIdentifier(v string) *DeleteDBClusterInput { + s.FinalDBSnapshotIdentifier = &v + return s +} + +// SetSkipFinalSnapshot sets the SkipFinalSnapshot field's value. +func (s *DeleteDBClusterInput) SetSkipFinalSnapshot(v bool) *DeleteDBClusterInput { + s.SkipFinalSnapshot = &v + return s +} + +type DeleteDBClusterOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s DeleteDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBClusterOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *DeleteDBClusterOutput) SetDBCluster(v *DBCluster) *DeleteDBClusterOutput { + s.DBCluster = v + return s +} + +type DeleteDBClusterParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster parameter group. + // + // Constraints: + // + // * Must be the name of an existing DB cluster parameter group. + // + // * You can't delete a default DB cluster parameter group. + // + // * Cannot be associated with any DB clusters. + // + // DBClusterParameterGroupName is a required field + DBClusterParameterGroupName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteDBClusterParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBClusterParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteDBClusterParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteDBClusterParameterGroupInput"} + if s.DBClusterParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterParameterGroupName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *DeleteDBClusterParameterGroupInput) SetDBClusterParameterGroupName(v string) *DeleteDBClusterParameterGroupInput { + s.DBClusterParameterGroupName = &v + return s +} + +type DeleteDBClusterParameterGroupOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeleteDBClusterParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBClusterParameterGroupOutput) GoString() string { + return s.String() +} + +type DeleteDBClusterSnapshotInput struct { + _ struct{} `type:"structure"` + + // The identifier of the DB cluster snapshot to delete. + // + // Constraints: Must be the name of an existing DB cluster snapshot in the available + // state. + // + // DBClusterSnapshotIdentifier is a required field + DBClusterSnapshotIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteDBClusterSnapshotInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBClusterSnapshotInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteDBClusterSnapshotInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteDBClusterSnapshotInput"} + if s.DBClusterSnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterSnapshotIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *DeleteDBClusterSnapshotInput) SetDBClusterSnapshotIdentifier(v string) *DeleteDBClusterSnapshotInput { + s.DBClusterSnapshotIdentifier = &v + return s +} + +type DeleteDBClusterSnapshotOutput struct { + _ struct{} `type:"structure"` + + // Contains the details for an Amazon Neptune DB cluster snapshot + // + // This data type is used as a response element in the DescribeDBClusterSnapshots + // action. + DBClusterSnapshot *DBClusterSnapshot `type:"structure"` +} + +// String returns the string representation +func (s DeleteDBClusterSnapshotOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBClusterSnapshotOutput) GoString() string { + return s.String() +} + +// SetDBClusterSnapshot sets the DBClusterSnapshot field's value. +func (s *DeleteDBClusterSnapshotOutput) SetDBClusterSnapshot(v *DBClusterSnapshot) *DeleteDBClusterSnapshotOutput { + s.DBClusterSnapshot = v + return s +} + +type DeleteDBInstanceInput struct { + _ struct{} `type:"structure"` + + // The DB instance identifier for the DB instance to be deleted. This parameter + // isn't case-sensitive. + // + // Constraints: + // + // * Must match the name of an existing DB instance. + // + // DBInstanceIdentifier is a required field + DBInstanceIdentifier *string `type:"string" required:"true"` + + // The DBSnapshotIdentifier of the new DBSnapshot created when SkipFinalSnapshot + // is set to false. + // + // Specifying this parameter and also setting the SkipFinalShapshot parameter + // to true results in an error. + // + // Constraints: + // + // * Must be 1 to 255 letters or numbers. + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // * Cannot be specified when deleting a Read Replica. + FinalDBSnapshotIdentifier *string `type:"string"` + + // Determines whether a final DB snapshot is created before the DB instance + // is deleted. If true is specified, no DBSnapshot is created. If false is specified, + // a DB snapshot is created before the DB instance is deleted. + // + // Note that when a DB instance is in a failure state and has a status of 'failed', + // 'incompatible-restore', or 'incompatible-network', it can only be deleted + // when the SkipFinalSnapshot parameter is set to "true". + // + // Specify true when deleting a Read Replica. + // + // The FinalDBSnapshotIdentifier parameter must be specified if SkipFinalSnapshot + // is false. + // + // Default: false + SkipFinalSnapshot *bool `type:"boolean"` +} + +// String returns the string representation +func (s DeleteDBInstanceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBInstanceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteDBInstanceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteDBInstanceInput"} + if s.DBInstanceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBInstanceIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *DeleteDBInstanceInput) SetDBInstanceIdentifier(v string) *DeleteDBInstanceInput { + s.DBInstanceIdentifier = &v + return s +} + +// SetFinalDBSnapshotIdentifier sets the FinalDBSnapshotIdentifier field's value. +func (s *DeleteDBInstanceInput) SetFinalDBSnapshotIdentifier(v string) *DeleteDBInstanceInput { + s.FinalDBSnapshotIdentifier = &v + return s +} + +// SetSkipFinalSnapshot sets the SkipFinalSnapshot field's value. +func (s *DeleteDBInstanceInput) SetSkipFinalSnapshot(v bool) *DeleteDBInstanceInput { + s.SkipFinalSnapshot = &v + return s +} + +type DeleteDBInstanceOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB instance. + // + // This data type is used as a response element in the DescribeDBInstances action. + DBInstance *DBInstance `type:"structure"` +} + +// String returns the string representation +func (s DeleteDBInstanceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBInstanceOutput) GoString() string { + return s.String() +} + +// SetDBInstance sets the DBInstance field's value. +func (s *DeleteDBInstanceOutput) SetDBInstance(v *DBInstance) *DeleteDBInstanceOutput { + s.DBInstance = v + return s +} + +type DeleteDBParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB parameter group. + // + // Constraints: + // + // * Must be the name of an existing DB parameter group + // + // * You can't delete a default DB parameter group + // + // * Cannot be associated with any DB instances + // + // DBParameterGroupName is a required field + DBParameterGroupName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteDBParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteDBParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteDBParameterGroupInput"} + if s.DBParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *DeleteDBParameterGroupInput) SetDBParameterGroupName(v string) *DeleteDBParameterGroupInput { + s.DBParameterGroupName = &v + return s +} + +type DeleteDBParameterGroupOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeleteDBParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBParameterGroupOutput) GoString() string { + return s.String() +} + +type DeleteDBSubnetGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the database subnet group to delete. + // + // You can't delete the default subnet group. + // + // Constraints: + // + // Constraints: Must match the name of an existing DBSubnetGroup. Must not be + // default. + // + // Example: mySubnetgroup + // + // DBSubnetGroupName is a required field + DBSubnetGroupName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteDBSubnetGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBSubnetGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteDBSubnetGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteDBSubnetGroupInput"} + if s.DBSubnetGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBSubnetGroupName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *DeleteDBSubnetGroupInput) SetDBSubnetGroupName(v string) *DeleteDBSubnetGroupInput { + s.DBSubnetGroupName = &v + return s +} + +type DeleteDBSubnetGroupOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DeleteDBSubnetGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteDBSubnetGroupOutput) GoString() string { + return s.String() +} + +type DeleteEventSubscriptionInput struct { + _ struct{} `type:"structure"` + + // The name of the event notification subscription you want to delete. + // + // SubscriptionName is a required field + SubscriptionName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteEventSubscriptionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteEventSubscriptionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteEventSubscriptionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteEventSubscriptionInput"} + if s.SubscriptionName == nil { + invalidParams.Add(request.NewErrParamRequired("SubscriptionName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSubscriptionName sets the SubscriptionName field's value. +func (s *DeleteEventSubscriptionInput) SetSubscriptionName(v string) *DeleteEventSubscriptionInput { + s.SubscriptionName = &v + return s +} + +type DeleteEventSubscriptionOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful invocation of the DescribeEventSubscriptions + // action. + EventSubscription *EventSubscription `type:"structure"` +} + +// String returns the string representation +func (s DeleteEventSubscriptionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteEventSubscriptionOutput) GoString() string { + return s.String() +} + +// SetEventSubscription sets the EventSubscription field's value. +func (s *DeleteEventSubscriptionOutput) SetEventSubscription(v *EventSubscription) *DeleteEventSubscriptionOutput { + s.EventSubscription = v + return s +} + +type DescribeDBClusterParameterGroupsInput struct { + _ struct{} `type:"structure"` + + // The name of a specific DB cluster parameter group to return details for. + // + // Constraints: + // + // * If supplied, must match the name of an existing DBClusterParameterGroup. + DBClusterParameterGroupName *string `type:"string"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBClusterParameterGroups + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeDBClusterParameterGroupsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterParameterGroupsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBClusterParameterGroupsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBClusterParameterGroupsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *DescribeDBClusterParameterGroupsInput) SetDBClusterParameterGroupName(v string) *DescribeDBClusterParameterGroupsInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBClusterParameterGroupsInput) SetFilters(v []*Filter) *DescribeDBClusterParameterGroupsInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClusterParameterGroupsInput) SetMarker(v string) *DescribeDBClusterParameterGroupsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBClusterParameterGroupsInput) SetMaxRecords(v int64) *DescribeDBClusterParameterGroupsInput { + s.MaxRecords = &v + return s +} + +type DescribeDBClusterParameterGroupsOutput struct { + _ struct{} `type:"structure"` + + // A list of DB cluster parameter groups. + DBClusterParameterGroups []*DBClusterParameterGroup `locationNameList:"DBClusterParameterGroup" type:"list"` + + // An optional pagination token provided by a previous DescribeDBClusterParameterGroups + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBClusterParameterGroupsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterParameterGroupsOutput) GoString() string { + return s.String() +} + +// SetDBClusterParameterGroups sets the DBClusterParameterGroups field's value. +func (s *DescribeDBClusterParameterGroupsOutput) SetDBClusterParameterGroups(v []*DBClusterParameterGroup) *DescribeDBClusterParameterGroupsOutput { + s.DBClusterParameterGroups = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClusterParameterGroupsOutput) SetMarker(v string) *DescribeDBClusterParameterGroupsOutput { + s.Marker = &v + return s +} + +type DescribeDBClusterParametersInput struct { + _ struct{} `type:"structure"` + + // The name of a specific DB cluster parameter group to return parameter details + // for. + // + // Constraints: + // + // * If supplied, must match the name of an existing DBClusterParameterGroup. + // + // DBClusterParameterGroupName is a required field + DBClusterParameterGroupName *string `type:"string" required:"true"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBClusterParameters + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // A value that indicates to return only parameters for a specific source. Parameter + // sources can be engine, service, or customer. + Source *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBClusterParametersInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterParametersInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBClusterParametersInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBClusterParametersInput"} + if s.DBClusterParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterParameterGroupName")) + } + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *DescribeDBClusterParametersInput) SetDBClusterParameterGroupName(v string) *DescribeDBClusterParametersInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBClusterParametersInput) SetFilters(v []*Filter) *DescribeDBClusterParametersInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClusterParametersInput) SetMarker(v string) *DescribeDBClusterParametersInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBClusterParametersInput) SetMaxRecords(v int64) *DescribeDBClusterParametersInput { + s.MaxRecords = &v + return s +} + +// SetSource sets the Source field's value. +func (s *DescribeDBClusterParametersInput) SetSource(v string) *DescribeDBClusterParametersInput { + s.Source = &v + return s +} + +// Provides details about a DB cluster parameter group including the parameters +// in the DB cluster parameter group. +type DescribeDBClusterParametersOutput struct { + _ struct{} `type:"structure"` + + // An optional pagination token provided by a previous DescribeDBClusterParameters + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords . + Marker *string `type:"string"` + + // Provides a list of parameters for the DB cluster parameter group. + Parameters []*Parameter `locationNameList:"Parameter" type:"list"` +} + +// String returns the string representation +func (s DescribeDBClusterParametersOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterParametersOutput) GoString() string { + return s.String() +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClusterParametersOutput) SetMarker(v string) *DescribeDBClusterParametersOutput { + s.Marker = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *DescribeDBClusterParametersOutput) SetParameters(v []*Parameter) *DescribeDBClusterParametersOutput { + s.Parameters = v + return s +} + +type DescribeDBClusterSnapshotAttributesInput struct { + _ struct{} `type:"structure"` + + // The identifier for the DB cluster snapshot to describe the attributes for. + // + // DBClusterSnapshotIdentifier is a required field + DBClusterSnapshotIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeDBClusterSnapshotAttributesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterSnapshotAttributesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBClusterSnapshotAttributesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBClusterSnapshotAttributesInput"} + if s.DBClusterSnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterSnapshotIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *DescribeDBClusterSnapshotAttributesInput) SetDBClusterSnapshotIdentifier(v string) *DescribeDBClusterSnapshotAttributesInput { + s.DBClusterSnapshotIdentifier = &v + return s +} + +type DescribeDBClusterSnapshotAttributesOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful call to the DescribeDBClusterSnapshotAttributes + // API action. + // + // Manual DB cluster snapshot attributes are used to authorize other AWS accounts + // to copy or restore a manual DB cluster snapshot. For more information, see + // the ModifyDBClusterSnapshotAttribute API action. + DBClusterSnapshotAttributesResult *DBClusterSnapshotAttributesResult `type:"structure"` +} + +// String returns the string representation +func (s DescribeDBClusterSnapshotAttributesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterSnapshotAttributesOutput) GoString() string { + return s.String() +} + +// SetDBClusterSnapshotAttributesResult sets the DBClusterSnapshotAttributesResult field's value. +func (s *DescribeDBClusterSnapshotAttributesOutput) SetDBClusterSnapshotAttributesResult(v *DBClusterSnapshotAttributesResult) *DescribeDBClusterSnapshotAttributesOutput { + s.DBClusterSnapshotAttributesResult = v + return s +} + +type DescribeDBClusterSnapshotsInput struct { + _ struct{} `type:"structure"` + + // The ID of the DB cluster to retrieve the list of DB cluster snapshots for. + // This parameter can't be used in conjunction with the DBClusterSnapshotIdentifier + // parameter. This parameter is not case-sensitive. + // + // Constraints: + // + // * If supplied, must match the identifier of an existing DBCluster. + DBClusterIdentifier *string `type:"string"` + + // A specific DB cluster snapshot identifier to describe. This parameter can't + // be used in conjunction with the DBClusterIdentifier parameter. This value + // is stored as a lowercase string. + // + // Constraints: + // + // * If supplied, must match the identifier of an existing DBClusterSnapshot. + // + // * If this identifier is for an automated snapshot, the SnapshotType parameter + // must also be specified. + DBClusterSnapshotIdentifier *string `type:"string"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // True to include manual DB cluster snapshots that are public and can be copied + // or restored by any AWS account, and otherwise false. The default is false. + // The default is false. + // + // You can share a manual DB cluster snapshot as public by using the ModifyDBClusterSnapshotAttribute + // API action. + IncludePublic *bool `type:"boolean"` + + // True to include shared manual DB cluster snapshots from other AWS accounts + // that this AWS account has been given permission to copy or restore, and otherwise + // false. The default is false. + // + // You can give an AWS account permission to restore a manual DB cluster snapshot + // from another AWS account by the ModifyDBClusterSnapshotAttribute API action. + IncludeShared *bool `type:"boolean"` + + // An optional pagination token provided by a previous DescribeDBClusterSnapshots + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // The type of DB cluster snapshots to be returned. You can specify one of the + // following values: + // + // * automated - Return all DB cluster snapshots that have been automatically + // taken by Amazon Neptune for my AWS account. + // + // * manual - Return all DB cluster snapshots that have been taken by my + // AWS account. + // + // * shared - Return all manual DB cluster snapshots that have been shared + // to my AWS account. + // + // * public - Return all DB cluster snapshots that have been marked as public. + // + // If you don't specify a SnapshotType value, then both automated and manual + // DB cluster snapshots are returned. You can include shared DB cluster snapshots + // with these results by setting the IncludeShared parameter to true. You can + // include public DB cluster snapshots with these results by setting the IncludePublic + // parameter to true. + // + // The IncludeShared and IncludePublic parameters don't apply for SnapshotType + // values of manual or automated. The IncludePublic parameter doesn't apply + // when SnapshotType is set to shared. The IncludeShared parameter doesn't apply + // when SnapshotType is set to public. + SnapshotType *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBClusterSnapshotsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterSnapshotsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBClusterSnapshotsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBClusterSnapshotsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *DescribeDBClusterSnapshotsInput) SetDBClusterIdentifier(v string) *DescribeDBClusterSnapshotsInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *DescribeDBClusterSnapshotsInput) SetDBClusterSnapshotIdentifier(v string) *DescribeDBClusterSnapshotsInput { + s.DBClusterSnapshotIdentifier = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBClusterSnapshotsInput) SetFilters(v []*Filter) *DescribeDBClusterSnapshotsInput { + s.Filters = v + return s +} + +// SetIncludePublic sets the IncludePublic field's value. +func (s *DescribeDBClusterSnapshotsInput) SetIncludePublic(v bool) *DescribeDBClusterSnapshotsInput { + s.IncludePublic = &v + return s +} + +// SetIncludeShared sets the IncludeShared field's value. +func (s *DescribeDBClusterSnapshotsInput) SetIncludeShared(v bool) *DescribeDBClusterSnapshotsInput { + s.IncludeShared = &v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClusterSnapshotsInput) SetMarker(v string) *DescribeDBClusterSnapshotsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBClusterSnapshotsInput) SetMaxRecords(v int64) *DescribeDBClusterSnapshotsInput { + s.MaxRecords = &v + return s +} + +// SetSnapshotType sets the SnapshotType field's value. +func (s *DescribeDBClusterSnapshotsInput) SetSnapshotType(v string) *DescribeDBClusterSnapshotsInput { + s.SnapshotType = &v + return s +} + +// Provides a list of DB cluster snapshots for the user as the result of a call +// to the DescribeDBClusterSnapshots action. +type DescribeDBClusterSnapshotsOutput struct { + _ struct{} `type:"structure"` + + // Provides a list of DB cluster snapshots for the user. + DBClusterSnapshots []*DBClusterSnapshot `locationNameList:"DBClusterSnapshot" type:"list"` + + // An optional pagination token provided by a previous DescribeDBClusterSnapshots + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBClusterSnapshotsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClusterSnapshotsOutput) GoString() string { + return s.String() +} + +// SetDBClusterSnapshots sets the DBClusterSnapshots field's value. +func (s *DescribeDBClusterSnapshotsOutput) SetDBClusterSnapshots(v []*DBClusterSnapshot) *DescribeDBClusterSnapshotsOutput { + s.DBClusterSnapshots = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClusterSnapshotsOutput) SetMarker(v string) *DescribeDBClusterSnapshotsOutput { + s.Marker = &v + return s +} + +type DescribeDBClustersInput struct { + _ struct{} `type:"structure"` + + // The user-supplied DB cluster identifier. If this parameter is specified, + // information from only the specific DB cluster is returned. This parameter + // isn't case-sensitive. + // + // Constraints: + // + // * If supplied, must match an existing DBClusterIdentifier. + DBClusterIdentifier *string `type:"string"` + + // A filter that specifies one or more DB clusters to describe. + // + // Supported filters: + // + // * db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon + // Resource Names (ARNs). The results list will only include information + // about the DB clusters identified by these ARNs. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBClusters request. + // If this parameter is specified, the response includes only records beyond + // the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeDBClustersInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClustersInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBClustersInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBClustersInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *DescribeDBClustersInput) SetDBClusterIdentifier(v string) *DescribeDBClustersInput { + s.DBClusterIdentifier = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBClustersInput) SetFilters(v []*Filter) *DescribeDBClustersInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClustersInput) SetMarker(v string) *DescribeDBClustersInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBClustersInput) SetMaxRecords(v int64) *DescribeDBClustersInput { + s.MaxRecords = &v + return s +} + +// Contains the result of a successful invocation of the DescribeDBClusters +// action. +type DescribeDBClustersOutput struct { + _ struct{} `type:"structure"` + + // Contains a list of DB clusters for the user. + DBClusters []*DBCluster `locationNameList:"DBCluster" type:"list"` + + // A pagination token that can be used in a subsequent DescribeDBClusters request. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBClustersOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBClustersOutput) GoString() string { + return s.String() +} + +// SetDBClusters sets the DBClusters field's value. +func (s *DescribeDBClustersOutput) SetDBClusters(v []*DBCluster) *DescribeDBClustersOutput { + s.DBClusters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBClustersOutput) SetMarker(v string) *DescribeDBClustersOutput { + s.Marker = &v + return s +} + +type DescribeDBEngineVersionsInput struct { + _ struct{} `type:"structure"` + + // The name of a specific DB parameter group family to return details for. + // + // Constraints: + // + // * If supplied, must match an existing DBParameterGroupFamily. + DBParameterGroupFamily *string `type:"string"` + + // Indicates that only the default version of the specified engine or engine + // and major version combination is returned. + DefaultOnly *bool `type:"boolean"` + + // The database engine to return. + Engine *string `type:"string"` + + // The database engine version to return. + // + // Example: 5.1.49 + EngineVersion *string `type:"string"` + + // Not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // If this parameter is specified and the requested engine supports the CharacterSetName + // parameter for CreateDBInstance, the response includes a list of supported + // character sets for each engine version. + ListSupportedCharacterSets *bool `type:"boolean"` + + // If this parameter is specified and the requested engine supports the TimeZone + // parameter for CreateDBInstance, the response includes a list of supported + // time zones for each engine version. + ListSupportedTimezones *bool `type:"boolean"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the marker, up to + // the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more than the + // MaxRecords value is available, a pagination token called a marker is included + // in the response so that the following results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeDBEngineVersionsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBEngineVersionsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBEngineVersionsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBEngineVersionsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *DescribeDBEngineVersionsInput) SetDBParameterGroupFamily(v string) *DescribeDBEngineVersionsInput { + s.DBParameterGroupFamily = &v + return s +} + +// SetDefaultOnly sets the DefaultOnly field's value. +func (s *DescribeDBEngineVersionsInput) SetDefaultOnly(v bool) *DescribeDBEngineVersionsInput { + s.DefaultOnly = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *DescribeDBEngineVersionsInput) SetEngine(v string) *DescribeDBEngineVersionsInput { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *DescribeDBEngineVersionsInput) SetEngineVersion(v string) *DescribeDBEngineVersionsInput { + s.EngineVersion = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBEngineVersionsInput) SetFilters(v []*Filter) *DescribeDBEngineVersionsInput { + s.Filters = v + return s +} + +// SetListSupportedCharacterSets sets the ListSupportedCharacterSets field's value. +func (s *DescribeDBEngineVersionsInput) SetListSupportedCharacterSets(v bool) *DescribeDBEngineVersionsInput { + s.ListSupportedCharacterSets = &v + return s +} + +// SetListSupportedTimezones sets the ListSupportedTimezones field's value. +func (s *DescribeDBEngineVersionsInput) SetListSupportedTimezones(v bool) *DescribeDBEngineVersionsInput { + s.ListSupportedTimezones = &v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBEngineVersionsInput) SetMarker(v string) *DescribeDBEngineVersionsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBEngineVersionsInput) SetMaxRecords(v int64) *DescribeDBEngineVersionsInput { + s.MaxRecords = &v + return s +} + +// Contains the result of a successful invocation of the DescribeDBEngineVersions +// action. +type DescribeDBEngineVersionsOutput struct { + _ struct{} `type:"structure"` + + // A list of DBEngineVersion elements. + DBEngineVersions []*DBEngineVersion `locationNameList:"DBEngineVersion" type:"list"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the marker, up to + // the value specified by MaxRecords. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBEngineVersionsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBEngineVersionsOutput) GoString() string { + return s.String() +} + +// SetDBEngineVersions sets the DBEngineVersions field's value. +func (s *DescribeDBEngineVersionsOutput) SetDBEngineVersions(v []*DBEngineVersion) *DescribeDBEngineVersionsOutput { + s.DBEngineVersions = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBEngineVersionsOutput) SetMarker(v string) *DescribeDBEngineVersionsOutput { + s.Marker = &v + return s +} + +type DescribeDBInstancesInput struct { + _ struct{} `type:"structure"` + + // The user-supplied instance identifier. If this parameter is specified, information + // from only the specific DB instance is returned. This parameter isn't case-sensitive. + // + // Constraints: + // + // * If supplied, must match the identifier of an existing DBInstance. + DBInstanceIdentifier *string `type:"string"` + + // A filter that specifies one or more DB instances to describe. + // + // Supported filters: + // + // * db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon + // Resource Names (ARNs). The results list will only include information + // about the DB instances associated with the DB clusters identified by these + // ARNs. + // + // * db-instance-id - Accepts DB instance identifiers and DB instance Amazon + // Resource Names (ARNs). The results list will only include information + // about the DB instances identified by these ARNs. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBInstances request. + // If this parameter is specified, the response includes only records beyond + // the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeDBInstancesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBInstancesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBInstancesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBInstancesInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *DescribeDBInstancesInput) SetDBInstanceIdentifier(v string) *DescribeDBInstancesInput { + s.DBInstanceIdentifier = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBInstancesInput) SetFilters(v []*Filter) *DescribeDBInstancesInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBInstancesInput) SetMarker(v string) *DescribeDBInstancesInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBInstancesInput) SetMaxRecords(v int64) *DescribeDBInstancesInput { + s.MaxRecords = &v + return s +} + +// Contains the result of a successful invocation of the DescribeDBInstances +// action. +type DescribeDBInstancesOutput struct { + _ struct{} `type:"structure"` + + // A list of DBInstance instances. + DBInstances []*DBInstance `locationNameList:"DBInstance" type:"list"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the marker, up to + // the value specified by MaxRecords . + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBInstancesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBInstancesOutput) GoString() string { + return s.String() +} + +// SetDBInstances sets the DBInstances field's value. +func (s *DescribeDBInstancesOutput) SetDBInstances(v []*DBInstance) *DescribeDBInstancesOutput { + s.DBInstances = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBInstancesOutput) SetMarker(v string) *DescribeDBInstancesOutput { + s.Marker = &v + return s +} + +type DescribeDBParameterGroupsInput struct { + _ struct{} `type:"structure"` + + // The name of a specific DB parameter group to return details for. + // + // Constraints: + // + // * If supplied, must match the name of an existing DBClusterParameterGroup. + DBParameterGroupName *string `type:"string"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBParameterGroups + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeDBParameterGroupsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBParameterGroupsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBParameterGroupsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBParameterGroupsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *DescribeDBParameterGroupsInput) SetDBParameterGroupName(v string) *DescribeDBParameterGroupsInput { + s.DBParameterGroupName = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBParameterGroupsInput) SetFilters(v []*Filter) *DescribeDBParameterGroupsInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBParameterGroupsInput) SetMarker(v string) *DescribeDBParameterGroupsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBParameterGroupsInput) SetMaxRecords(v int64) *DescribeDBParameterGroupsInput { + s.MaxRecords = &v + return s +} + +// Contains the result of a successful invocation of the DescribeDBParameterGroups +// action. +type DescribeDBParameterGroupsOutput struct { + _ struct{} `type:"structure"` + + // A list of DBParameterGroup instances. + DBParameterGroups []*DBParameterGroup `locationNameList:"DBParameterGroup" type:"list"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the marker, up to + // the value specified by MaxRecords. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBParameterGroupsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBParameterGroupsOutput) GoString() string { + return s.String() +} + +// SetDBParameterGroups sets the DBParameterGroups field's value. +func (s *DescribeDBParameterGroupsOutput) SetDBParameterGroups(v []*DBParameterGroup) *DescribeDBParameterGroupsOutput { + s.DBParameterGroups = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBParameterGroupsOutput) SetMarker(v string) *DescribeDBParameterGroupsOutput { + s.Marker = &v + return s +} + +type DescribeDBParametersInput struct { + _ struct{} `type:"structure"` + + // The name of a specific DB parameter group to return details for. + // + // Constraints: + // + // * If supplied, must match the name of an existing DBParameterGroup. + // + // DBParameterGroupName is a required field + DBParameterGroupName *string `type:"string" required:"true"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBParameters + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // The parameter types to return. + // + // Default: All parameter types returned + // + // Valid Values: user | system | engine-default + Source *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBParametersInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBParametersInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBParametersInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBParametersInput"} + if s.DBParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupName")) + } + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *DescribeDBParametersInput) SetDBParameterGroupName(v string) *DescribeDBParametersInput { + s.DBParameterGroupName = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBParametersInput) SetFilters(v []*Filter) *DescribeDBParametersInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBParametersInput) SetMarker(v string) *DescribeDBParametersInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBParametersInput) SetMaxRecords(v int64) *DescribeDBParametersInput { + s.MaxRecords = &v + return s +} + +// SetSource sets the Source field's value. +func (s *DescribeDBParametersInput) SetSource(v string) *DescribeDBParametersInput { + s.Source = &v + return s +} + +// Contains the result of a successful invocation of the DescribeDBParameters +// action. +type DescribeDBParametersOutput struct { + _ struct{} `type:"structure"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the marker, up to + // the value specified by MaxRecords. + Marker *string `type:"string"` + + // A list of Parameter values. + Parameters []*Parameter `locationNameList:"Parameter" type:"list"` +} + +// String returns the string representation +func (s DescribeDBParametersOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBParametersOutput) GoString() string { + return s.String() +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBParametersOutput) SetMarker(v string) *DescribeDBParametersOutput { + s.Marker = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *DescribeDBParametersOutput) SetParameters(v []*Parameter) *DescribeDBParametersOutput { + s.Parameters = v + return s +} + +type DescribeDBSubnetGroupsInput struct { + _ struct{} `type:"structure"` + + // The name of the DB subnet group to return details for. + DBSubnetGroupName *string `type:"string"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeDBSubnetGroups + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeDBSubnetGroupsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBSubnetGroupsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDBSubnetGroupsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDBSubnetGroupsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *DescribeDBSubnetGroupsInput) SetDBSubnetGroupName(v string) *DescribeDBSubnetGroupsInput { + s.DBSubnetGroupName = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeDBSubnetGroupsInput) SetFilters(v []*Filter) *DescribeDBSubnetGroupsInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBSubnetGroupsInput) SetMarker(v string) *DescribeDBSubnetGroupsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeDBSubnetGroupsInput) SetMaxRecords(v int64) *DescribeDBSubnetGroupsInput { + s.MaxRecords = &v + return s +} + +// Contains the result of a successful invocation of the DescribeDBSubnetGroups +// action. +type DescribeDBSubnetGroupsOutput struct { + _ struct{} `type:"structure"` + + // A list of DBSubnetGroup instances. + DBSubnetGroups []*DBSubnetGroup `locationNameList:"DBSubnetGroup" type:"list"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the marker, up to + // the value specified by MaxRecords. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDBSubnetGroupsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDBSubnetGroupsOutput) GoString() string { + return s.String() +} + +// SetDBSubnetGroups sets the DBSubnetGroups field's value. +func (s *DescribeDBSubnetGroupsOutput) SetDBSubnetGroups(v []*DBSubnetGroup) *DescribeDBSubnetGroupsOutput { + s.DBSubnetGroups = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeDBSubnetGroupsOutput) SetMarker(v string) *DescribeDBSubnetGroupsOutput { + s.Marker = &v + return s +} + +type DescribeEngineDefaultClusterParametersInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster parameter group family to return engine parameter + // information for. + // + // DBParameterGroupFamily is a required field + DBParameterGroupFamily *string `type:"string" required:"true"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeEngineDefaultClusterParameters + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeEngineDefaultClusterParametersInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEngineDefaultClusterParametersInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeEngineDefaultClusterParametersInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeEngineDefaultClusterParametersInput"} + if s.DBParameterGroupFamily == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupFamily")) + } + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *DescribeEngineDefaultClusterParametersInput) SetDBParameterGroupFamily(v string) *DescribeEngineDefaultClusterParametersInput { + s.DBParameterGroupFamily = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeEngineDefaultClusterParametersInput) SetFilters(v []*Filter) *DescribeEngineDefaultClusterParametersInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeEngineDefaultClusterParametersInput) SetMarker(v string) *DescribeEngineDefaultClusterParametersInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeEngineDefaultClusterParametersInput) SetMaxRecords(v int64) *DescribeEngineDefaultClusterParametersInput { + s.MaxRecords = &v + return s +} + +type DescribeEngineDefaultClusterParametersOutput struct { + _ struct{} `type:"structure"` + + // Contains the result of a successful invocation of the DescribeEngineDefaultParameters + // action. + EngineDefaults *EngineDefaults `type:"structure"` +} + +// String returns the string representation +func (s DescribeEngineDefaultClusterParametersOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEngineDefaultClusterParametersOutput) GoString() string { + return s.String() +} + +// SetEngineDefaults sets the EngineDefaults field's value. +func (s *DescribeEngineDefaultClusterParametersOutput) SetEngineDefaults(v *EngineDefaults) *DescribeEngineDefaultClusterParametersOutput { + s.EngineDefaults = v + return s +} + +type DescribeEngineDefaultParametersInput struct { + _ struct{} `type:"structure"` + + // The name of the DB parameter group family. + // + // DBParameterGroupFamily is a required field + DBParameterGroupFamily *string `type:"string" required:"true"` + + // Not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeEngineDefaultParameters + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` +} + +// String returns the string representation +func (s DescribeEngineDefaultParametersInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEngineDefaultParametersInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeEngineDefaultParametersInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeEngineDefaultParametersInput"} + if s.DBParameterGroupFamily == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupFamily")) + } + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *DescribeEngineDefaultParametersInput) SetDBParameterGroupFamily(v string) *DescribeEngineDefaultParametersInput { + s.DBParameterGroupFamily = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeEngineDefaultParametersInput) SetFilters(v []*Filter) *DescribeEngineDefaultParametersInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeEngineDefaultParametersInput) SetMarker(v string) *DescribeEngineDefaultParametersInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeEngineDefaultParametersInput) SetMaxRecords(v int64) *DescribeEngineDefaultParametersInput { + s.MaxRecords = &v + return s +} + +type DescribeEngineDefaultParametersOutput struct { + _ struct{} `type:"structure"` + + // Contains the result of a successful invocation of the DescribeEngineDefaultParameters + // action. + EngineDefaults *EngineDefaults `type:"structure"` +} + +// String returns the string representation +func (s DescribeEngineDefaultParametersOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEngineDefaultParametersOutput) GoString() string { + return s.String() +} + +// SetEngineDefaults sets the EngineDefaults field's value. +func (s *DescribeEngineDefaultParametersOutput) SetEngineDefaults(v *EngineDefaults) *DescribeEngineDefaultParametersOutput { + s.EngineDefaults = v + return s +} + +type DescribeEventCategoriesInput struct { + _ struct{} `type:"structure"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // The type of source that is generating the events. + // + // Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot + SourceType *string `type:"string"` +} + +// String returns the string representation +func (s DescribeEventCategoriesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEventCategoriesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeEventCategoriesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeEventCategoriesInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilters sets the Filters field's value. +func (s *DescribeEventCategoriesInput) SetFilters(v []*Filter) *DescribeEventCategoriesInput { + s.Filters = v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *DescribeEventCategoriesInput) SetSourceType(v string) *DescribeEventCategoriesInput { + s.SourceType = &v + return s +} + +// Data returned from the DescribeEventCategories action. +type DescribeEventCategoriesOutput struct { + _ struct{} `type:"structure"` + + // A list of EventCategoriesMap data types. + EventCategoriesMapList []*EventCategoriesMap `locationNameList:"EventCategoriesMap" type:"list"` +} + +// String returns the string representation +func (s DescribeEventCategoriesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEventCategoriesOutput) GoString() string { + return s.String() +} + +// SetEventCategoriesMapList sets the EventCategoriesMapList field's value. +func (s *DescribeEventCategoriesOutput) SetEventCategoriesMapList(v []*EventCategoriesMap) *DescribeEventCategoriesOutput { + s.EventCategoriesMapList = v + return s +} + +type DescribeEventSubscriptionsInput struct { + _ struct{} `type:"structure"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords . + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // The name of the event notification subscription you want to describe. + SubscriptionName *string `type:"string"` +} + +// String returns the string representation +func (s DescribeEventSubscriptionsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEventSubscriptionsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeEventSubscriptionsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeEventSubscriptionsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilters sets the Filters field's value. +func (s *DescribeEventSubscriptionsInput) SetFilters(v []*Filter) *DescribeEventSubscriptionsInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeEventSubscriptionsInput) SetMarker(v string) *DescribeEventSubscriptionsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeEventSubscriptionsInput) SetMaxRecords(v int64) *DescribeEventSubscriptionsInput { + s.MaxRecords = &v + return s +} + +// SetSubscriptionName sets the SubscriptionName field's value. +func (s *DescribeEventSubscriptionsInput) SetSubscriptionName(v string) *DescribeEventSubscriptionsInput { + s.SubscriptionName = &v + return s +} + +// Data returned by the DescribeEventSubscriptions action. +type DescribeEventSubscriptionsOutput struct { + _ struct{} `type:"structure"` + + // A list of EventSubscriptions data types. + EventSubscriptionsList []*EventSubscription `locationNameList:"EventSubscription" type:"list"` + + // An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeEventSubscriptionsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEventSubscriptionsOutput) GoString() string { + return s.String() +} + +// SetEventSubscriptionsList sets the EventSubscriptionsList field's value. +func (s *DescribeEventSubscriptionsOutput) SetEventSubscriptionsList(v []*EventSubscription) *DescribeEventSubscriptionsOutput { + s.EventSubscriptionsList = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeEventSubscriptionsOutput) SetMarker(v string) *DescribeEventSubscriptionsOutput { + s.Marker = &v + return s +} + +type DescribeEventsInput struct { + _ struct{} `type:"structure"` + + // The number of minutes to retrieve events for. + // + // Default: 60 + Duration *int64 `type:"integer"` + + // The end of the time interval for which to retrieve events, specified in ISO + // 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia + // page. (http://en.wikipedia.org/wiki/ISO_8601) + // + // Example: 2009-07-08T18:00Z + EndTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // A list of event categories that trigger notifications for a event notification + // subscription. + EventCategories []*string `locationNameList:"EventCategory" type:"list"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribeEvents request. + // If this parameter is specified, the response includes only records beyond + // the marker, up to the value specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // The identifier of the event source for which events are returned. If not + // specified, then all sources are included in the response. + // + // Constraints: + // + // * If SourceIdentifier is supplied, SourceType must also be provided. + // + // * If the source type is DBInstance, then a DBInstanceIdentifier must be + // supplied. + // + // * If the source type is DBSecurityGroup, a DBSecurityGroupName must be + // supplied. + // + // * If the source type is DBParameterGroup, a DBParameterGroupName must + // be supplied. + // + // * If the source type is DBSnapshot, a DBSnapshotIdentifier must be supplied. + // + // * Cannot end with a hyphen or contain two consecutive hyphens. + SourceIdentifier *string `type:"string"` + + // The event source to retrieve events for. If no value is specified, all events + // are returned. + SourceType *string `type:"string" enum:"SourceType"` + + // The beginning of the time interval to retrieve events for, specified in ISO + // 8601 format. For more information about ISO 8601, go to the ISO8601 Wikipedia + // page. (http://en.wikipedia.org/wiki/ISO_8601) + // + // Example: 2009-07-08T18:00Z + StartTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` +} + +// String returns the string representation +func (s DescribeEventsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEventsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeEventsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeEventsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDuration sets the Duration field's value. +func (s *DescribeEventsInput) SetDuration(v int64) *DescribeEventsInput { + s.Duration = &v + return s +} + +// SetEndTime sets the EndTime field's value. +func (s *DescribeEventsInput) SetEndTime(v time.Time) *DescribeEventsInput { + s.EndTime = &v + return s +} + +// SetEventCategories sets the EventCategories field's value. +func (s *DescribeEventsInput) SetEventCategories(v []*string) *DescribeEventsInput { + s.EventCategories = v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeEventsInput) SetFilters(v []*Filter) *DescribeEventsInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeEventsInput) SetMarker(v string) *DescribeEventsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeEventsInput) SetMaxRecords(v int64) *DescribeEventsInput { + s.MaxRecords = &v + return s +} + +// SetSourceIdentifier sets the SourceIdentifier field's value. +func (s *DescribeEventsInput) SetSourceIdentifier(v string) *DescribeEventsInput { + s.SourceIdentifier = &v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *DescribeEventsInput) SetSourceType(v string) *DescribeEventsInput { + s.SourceType = &v + return s +} + +// SetStartTime sets the StartTime field's value. +func (s *DescribeEventsInput) SetStartTime(v time.Time) *DescribeEventsInput { + s.StartTime = &v + return s +} + +// Contains the result of a successful invocation of the DescribeEvents action. +type DescribeEventsOutput struct { + _ struct{} `type:"structure"` + + // A list of Event instances. + Events []*Event `locationNameList:"Event" type:"list"` + + // An optional pagination token provided by a previous Events request. If this + // parameter is specified, the response includes only records beyond the marker, + // up to the value specified by MaxRecords . + Marker *string `type:"string"` +} + +// String returns the string representation +func (s DescribeEventsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEventsOutput) GoString() string { + return s.String() +} + +// SetEvents sets the Events field's value. +func (s *DescribeEventsOutput) SetEvents(v []*Event) *DescribeEventsOutput { + s.Events = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeEventsOutput) SetMarker(v string) *DescribeEventsOutput { + s.Marker = &v + return s +} + +type DescribeOrderableDBInstanceOptionsInput struct { + _ struct{} `type:"structure"` + + // The DB instance class filter value. Specify this parameter to show only the + // available offerings matching the specified DB instance class. + DBInstanceClass *string `type:"string"` + + // The name of the engine to retrieve DB instance options for. + // + // Engine is a required field + Engine *string `type:"string" required:"true"` + + // The engine version filter value. Specify this parameter to show only the + // available offerings matching the specified engine version. + EngineVersion *string `type:"string"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // The license model filter value. Specify this parameter to show only the available + // offerings matching the specified license model. + LicenseModel *string `type:"string"` + + // An optional pagination token provided by a previous DescribeOrderableDBInstanceOptions + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords . + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // The VPC filter value. Specify this parameter to show only the available VPC + // or non-VPC offerings. + Vpc *bool `type:"boolean"` +} + +// String returns the string representation +func (s DescribeOrderableDBInstanceOptionsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeOrderableDBInstanceOptionsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeOrderableDBInstanceOptionsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeOrderableDBInstanceOptionsInput"} + if s.Engine == nil { + invalidParams.Add(request.NewErrParamRequired("Engine")) + } + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBInstanceClass sets the DBInstanceClass field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetDBInstanceClass(v string) *DescribeOrderableDBInstanceOptionsInput { + s.DBInstanceClass = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetEngine(v string) *DescribeOrderableDBInstanceOptionsInput { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetEngineVersion(v string) *DescribeOrderableDBInstanceOptionsInput { + s.EngineVersion = &v + return s +} + +// SetFilters sets the Filters field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetFilters(v []*Filter) *DescribeOrderableDBInstanceOptionsInput { + s.Filters = v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetLicenseModel(v string) *DescribeOrderableDBInstanceOptionsInput { + s.LicenseModel = &v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetMarker(v string) *DescribeOrderableDBInstanceOptionsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetMaxRecords(v int64) *DescribeOrderableDBInstanceOptionsInput { + s.MaxRecords = &v + return s +} + +// SetVpc sets the Vpc field's value. +func (s *DescribeOrderableDBInstanceOptionsInput) SetVpc(v bool) *DescribeOrderableDBInstanceOptionsInput { + s.Vpc = &v + return s +} + +// Contains the result of a successful invocation of the DescribeOrderableDBInstanceOptions +// action. +type DescribeOrderableDBInstanceOptionsOutput struct { + _ struct{} `type:"structure"` + + // An optional pagination token provided by a previous OrderableDBInstanceOptions + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to the value specified by MaxRecords . + Marker *string `type:"string"` + + // An OrderableDBInstanceOption structure containing information about orderable + // options for the DB instance. + OrderableDBInstanceOptions []*OrderableDBInstanceOption `locationNameList:"OrderableDBInstanceOption" type:"list"` +} + +// String returns the string representation +func (s DescribeOrderableDBInstanceOptionsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeOrderableDBInstanceOptionsOutput) GoString() string { + return s.String() +} + +// SetMarker sets the Marker field's value. +func (s *DescribeOrderableDBInstanceOptionsOutput) SetMarker(v string) *DescribeOrderableDBInstanceOptionsOutput { + s.Marker = &v + return s +} + +// SetOrderableDBInstanceOptions sets the OrderableDBInstanceOptions field's value. +func (s *DescribeOrderableDBInstanceOptionsOutput) SetOrderableDBInstanceOptions(v []*OrderableDBInstanceOption) *DescribeOrderableDBInstanceOptionsOutput { + s.OrderableDBInstanceOptions = v + return s +} + +type DescribePendingMaintenanceActionsInput struct { + _ struct{} `type:"structure"` + + // A filter that specifies one or more resources to return pending maintenance + // actions for. + // + // Supported filters: + // + // * db-cluster-id - Accepts DB cluster identifiers and DB cluster Amazon + // Resource Names (ARNs). The results list will only include pending maintenance + // actions for the DB clusters identified by these ARNs. + // + // * db-instance-id - Accepts DB instance identifiers and DB instance ARNs. + // The results list will only include pending maintenance actions for the + // DB instances identified by these ARNs. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // An optional pagination token provided by a previous DescribePendingMaintenanceActions + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to a number of records specified by MaxRecords. + Marker *string `type:"string"` + + // The maximum number of records to include in the response. If more records + // exist than the specified MaxRecords value, a pagination token called a marker + // is included in the response so that the remaining results can be retrieved. + // + // Default: 100 + // + // Constraints: Minimum 20, maximum 100. + MaxRecords *int64 `type:"integer"` + + // The ARN of a resource to return pending maintenance actions for. + ResourceIdentifier *string `type:"string"` +} + +// String returns the string representation +func (s DescribePendingMaintenanceActionsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribePendingMaintenanceActionsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribePendingMaintenanceActionsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribePendingMaintenanceActionsInput"} + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilters sets the Filters field's value. +func (s *DescribePendingMaintenanceActionsInput) SetFilters(v []*Filter) *DescribePendingMaintenanceActionsInput { + s.Filters = v + return s +} + +// SetMarker sets the Marker field's value. +func (s *DescribePendingMaintenanceActionsInput) SetMarker(v string) *DescribePendingMaintenanceActionsInput { + s.Marker = &v + return s +} + +// SetMaxRecords sets the MaxRecords field's value. +func (s *DescribePendingMaintenanceActionsInput) SetMaxRecords(v int64) *DescribePendingMaintenanceActionsInput { + s.MaxRecords = &v + return s +} + +// SetResourceIdentifier sets the ResourceIdentifier field's value. +func (s *DescribePendingMaintenanceActionsInput) SetResourceIdentifier(v string) *DescribePendingMaintenanceActionsInput { + s.ResourceIdentifier = &v + return s +} + +// Data returned from the DescribePendingMaintenanceActions action. +type DescribePendingMaintenanceActionsOutput struct { + _ struct{} `type:"structure"` + + // An optional pagination token provided by a previous DescribePendingMaintenanceActions + // request. If this parameter is specified, the response includes only records + // beyond the marker, up to a number of records specified by MaxRecords. + Marker *string `type:"string"` + + // A list of the pending maintenance actions for the resource. + PendingMaintenanceActions []*ResourcePendingMaintenanceActions `locationNameList:"ResourcePendingMaintenanceActions" type:"list"` +} + +// String returns the string representation +func (s DescribePendingMaintenanceActionsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribePendingMaintenanceActionsOutput) GoString() string { + return s.String() +} + +// SetMarker sets the Marker field's value. +func (s *DescribePendingMaintenanceActionsOutput) SetMarker(v string) *DescribePendingMaintenanceActionsOutput { + s.Marker = &v + return s +} + +// SetPendingMaintenanceActions sets the PendingMaintenanceActions field's value. +func (s *DescribePendingMaintenanceActionsOutput) SetPendingMaintenanceActions(v []*ResourcePendingMaintenanceActions) *DescribePendingMaintenanceActionsOutput { + s.PendingMaintenanceActions = v + return s +} + +type DescribeValidDBInstanceModificationsInput struct { + _ struct{} `type:"structure"` + + // The customer identifier or the ARN of your DB instance. + // + // DBInstanceIdentifier is a required field + DBInstanceIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeValidDBInstanceModificationsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeValidDBInstanceModificationsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeValidDBInstanceModificationsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeValidDBInstanceModificationsInput"} + if s.DBInstanceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBInstanceIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *DescribeValidDBInstanceModificationsInput) SetDBInstanceIdentifier(v string) *DescribeValidDBInstanceModificationsInput { + s.DBInstanceIdentifier = &v + return s +} + +type DescribeValidDBInstanceModificationsOutput struct { + _ struct{} `type:"structure"` + + // Information about valid modifications that you can make to your DB instance. + // Contains the result of a successful call to the DescribeValidDBInstanceModifications + // action. You can use this information when you call ModifyDBInstance. + ValidDBInstanceModificationsMessage *ValidDBInstanceModificationsMessage `type:"structure"` +} + +// String returns the string representation +func (s DescribeValidDBInstanceModificationsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeValidDBInstanceModificationsOutput) GoString() string { + return s.String() +} + +// SetValidDBInstanceModificationsMessage sets the ValidDBInstanceModificationsMessage field's value. +func (s *DescribeValidDBInstanceModificationsOutput) SetValidDBInstanceModificationsMessage(v *ValidDBInstanceModificationsMessage) *DescribeValidDBInstanceModificationsOutput { + s.ValidDBInstanceModificationsMessage = v + return s +} + +// An Active Directory Domain membership record associated with the DB instance. +type DomainMembership struct { + _ struct{} `type:"structure"` + + // The identifier of the Active Directory Domain. + Domain *string `type:"string"` + + // The fully qualified domain name of the Active Directory Domain. + FQDN *string `type:"string"` + + // The name of the IAM role to be used when making API calls to the Directory + // Service. + IAMRoleName *string `type:"string"` + + // The status of the DB instance's Active Directory Domain membership, such + // as joined, pending-join, failed etc). + Status *string `type:"string"` +} + +// String returns the string representation +func (s DomainMembership) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DomainMembership) GoString() string { + return s.String() +} + +// SetDomain sets the Domain field's value. +func (s *DomainMembership) SetDomain(v string) *DomainMembership { + s.Domain = &v + return s +} + +// SetFQDN sets the FQDN field's value. +func (s *DomainMembership) SetFQDN(v string) *DomainMembership { + s.FQDN = &v + return s +} + +// SetIAMRoleName sets the IAMRoleName field's value. +func (s *DomainMembership) SetIAMRoleName(v string) *DomainMembership { + s.IAMRoleName = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *DomainMembership) SetStatus(v string) *DomainMembership { + s.Status = &v + return s +} + +// A range of double values. +type DoubleRange struct { + _ struct{} `type:"structure"` + + // The minimum value in the range. + From *float64 `type:"double"` + + // The maximum value in the range. + To *float64 `type:"double"` +} + +// String returns the string representation +func (s DoubleRange) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DoubleRange) GoString() string { + return s.String() +} + +// SetFrom sets the From field's value. +func (s *DoubleRange) SetFrom(v float64) *DoubleRange { + s.From = &v + return s +} + +// SetTo sets the To field's value. +func (s *DoubleRange) SetTo(v float64) *DoubleRange { + s.To = &v + return s +} + +// This data type is used as a response element in the following actions: +// +// * CreateDBInstance +// +// * DescribeDBInstances +// +// * DeleteDBInstance +type Endpoint struct { + _ struct{} `type:"structure"` + + // Specifies the DNS address of the DB instance. + Address *string `type:"string"` + + // Specifies the ID that Amazon Route 53 assigns when you create a hosted zone. + HostedZoneId *string `type:"string"` + + // Specifies the port that the database engine is listening on. + Port *int64 `type:"integer"` +} + +// String returns the string representation +func (s Endpoint) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Endpoint) GoString() string { + return s.String() +} + +// SetAddress sets the Address field's value. +func (s *Endpoint) SetAddress(v string) *Endpoint { + s.Address = &v + return s +} + +// SetHostedZoneId sets the HostedZoneId field's value. +func (s *Endpoint) SetHostedZoneId(v string) *Endpoint { + s.HostedZoneId = &v + return s +} + +// SetPort sets the Port field's value. +func (s *Endpoint) SetPort(v int64) *Endpoint { + s.Port = &v + return s +} + +// Contains the result of a successful invocation of the DescribeEngineDefaultParameters +// action. +type EngineDefaults struct { + _ struct{} `type:"structure"` + + // Specifies the name of the DB parameter group family that the engine default + // parameters apply to. + DBParameterGroupFamily *string `type:"string"` + + // An optional pagination token provided by a previous EngineDefaults request. + // If this parameter is specified, the response includes only records beyond + // the marker, up to the value specified by MaxRecords . + Marker *string `type:"string"` + + // Contains a list of engine default parameters. + Parameters []*Parameter `locationNameList:"Parameter" type:"list"` +} + +// String returns the string representation +func (s EngineDefaults) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EngineDefaults) GoString() string { + return s.String() +} + +// SetDBParameterGroupFamily sets the DBParameterGroupFamily field's value. +func (s *EngineDefaults) SetDBParameterGroupFamily(v string) *EngineDefaults { + s.DBParameterGroupFamily = &v + return s +} + +// SetMarker sets the Marker field's value. +func (s *EngineDefaults) SetMarker(v string) *EngineDefaults { + s.Marker = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *EngineDefaults) SetParameters(v []*Parameter) *EngineDefaults { + s.Parameters = v + return s +} + +// This data type is used as a response element in the DescribeEvents action. +type Event struct { + _ struct{} `type:"structure"` + + // Specifies the date and time of the event. + Date *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Specifies the category for the event. + EventCategories []*string `locationNameList:"EventCategory" type:"list"` + + // Provides the text of this event. + Message *string `type:"string"` + + // The Amazon Resource Name (ARN) for the event. + SourceArn *string `type:"string"` + + // Provides the identifier for the source of the event. + SourceIdentifier *string `type:"string"` + + // Specifies the source type for this event. + SourceType *string `type:"string" enum:"SourceType"` +} + +// String returns the string representation +func (s Event) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Event) GoString() string { + return s.String() +} + +// SetDate sets the Date field's value. +func (s *Event) SetDate(v time.Time) *Event { + s.Date = &v + return s +} + +// SetEventCategories sets the EventCategories field's value. +func (s *Event) SetEventCategories(v []*string) *Event { + s.EventCategories = v + return s +} + +// SetMessage sets the Message field's value. +func (s *Event) SetMessage(v string) *Event { + s.Message = &v + return s +} + +// SetSourceArn sets the SourceArn field's value. +func (s *Event) SetSourceArn(v string) *Event { + s.SourceArn = &v + return s +} + +// SetSourceIdentifier sets the SourceIdentifier field's value. +func (s *Event) SetSourceIdentifier(v string) *Event { + s.SourceIdentifier = &v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *Event) SetSourceType(v string) *Event { + s.SourceType = &v + return s +} + +// Contains the results of a successful invocation of the DescribeEventCategories +// action. +type EventCategoriesMap struct { + _ struct{} `type:"structure"` + + // The event categories for the specified source type + EventCategories []*string `locationNameList:"EventCategory" type:"list"` + + // The source type that the returned categories belong to + SourceType *string `type:"string"` +} + +// String returns the string representation +func (s EventCategoriesMap) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EventCategoriesMap) GoString() string { + return s.String() +} + +// SetEventCategories sets the EventCategories field's value. +func (s *EventCategoriesMap) SetEventCategories(v []*string) *EventCategoriesMap { + s.EventCategories = v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *EventCategoriesMap) SetSourceType(v string) *EventCategoriesMap { + s.SourceType = &v + return s +} + +// Contains the results of a successful invocation of the DescribeEventSubscriptions +// action. +type EventSubscription struct { + _ struct{} `type:"structure"` + + // The event notification subscription Id. + CustSubscriptionId *string `type:"string"` + + // The AWS customer account associated with the event notification subscription. + CustomerAwsId *string `type:"string"` + + // A Boolean value indicating if the subscription is enabled. True indicates + // the subscription is enabled. + Enabled *bool `type:"boolean"` + + // A list of event categories for the event notification subscription. + EventCategoriesList []*string `locationNameList:"EventCategory" type:"list"` + + // The Amazon Resource Name (ARN) for the event subscription. + EventSubscriptionArn *string `type:"string"` + + // The topic ARN of the event notification subscription. + SnsTopicArn *string `type:"string"` + + // A list of source IDs for the event notification subscription. + SourceIdsList []*string `locationNameList:"SourceId" type:"list"` + + // The source type for the event notification subscription. + SourceType *string `type:"string"` + + // The status of the event notification subscription. + // + // Constraints: + // + // Can be one of the following: creating | modifying | deleting | active | no-permission + // | topic-not-exist + // + // The status "no-permission" indicates that Neptune no longer has permission + // to post to the SNS topic. The status "topic-not-exist" indicates that the + // topic was deleted after the subscription was created. + Status *string `type:"string"` + + // The time the event notification subscription was created. + SubscriptionCreationTime *string `type:"string"` +} + +// String returns the string representation +func (s EventSubscription) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EventSubscription) GoString() string { + return s.String() +} + +// SetCustSubscriptionId sets the CustSubscriptionId field's value. +func (s *EventSubscription) SetCustSubscriptionId(v string) *EventSubscription { + s.CustSubscriptionId = &v + return s +} + +// SetCustomerAwsId sets the CustomerAwsId field's value. +func (s *EventSubscription) SetCustomerAwsId(v string) *EventSubscription { + s.CustomerAwsId = &v + return s +} + +// SetEnabled sets the Enabled field's value. +func (s *EventSubscription) SetEnabled(v bool) *EventSubscription { + s.Enabled = &v + return s +} + +// SetEventCategoriesList sets the EventCategoriesList field's value. +func (s *EventSubscription) SetEventCategoriesList(v []*string) *EventSubscription { + s.EventCategoriesList = v + return s +} + +// SetEventSubscriptionArn sets the EventSubscriptionArn field's value. +func (s *EventSubscription) SetEventSubscriptionArn(v string) *EventSubscription { + s.EventSubscriptionArn = &v + return s +} + +// SetSnsTopicArn sets the SnsTopicArn field's value. +func (s *EventSubscription) SetSnsTopicArn(v string) *EventSubscription { + s.SnsTopicArn = &v + return s +} + +// SetSourceIdsList sets the SourceIdsList field's value. +func (s *EventSubscription) SetSourceIdsList(v []*string) *EventSubscription { + s.SourceIdsList = v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *EventSubscription) SetSourceType(v string) *EventSubscription { + s.SourceType = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *EventSubscription) SetStatus(v string) *EventSubscription { + s.Status = &v + return s +} + +// SetSubscriptionCreationTime sets the SubscriptionCreationTime field's value. +func (s *EventSubscription) SetSubscriptionCreationTime(v string) *EventSubscription { + s.SubscriptionCreationTime = &v + return s +} + +type FailoverDBClusterInput struct { + _ struct{} `type:"structure"` + + // A DB cluster identifier to force a failover for. This parameter is not case-sensitive. + // + // Constraints: + // + // * Must match the identifier of an existing DBCluster. + DBClusterIdentifier *string `type:"string"` + + // The name of the instance to promote to the primary instance. + // + // You must specify the instance identifier for an Read Replica in the DB cluster. + // For example, mydbcluster-replica1. + TargetDBInstanceIdentifier *string `type:"string"` +} + +// String returns the string representation +func (s FailoverDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FailoverDBClusterInput) GoString() string { + return s.String() +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *FailoverDBClusterInput) SetDBClusterIdentifier(v string) *FailoverDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +// SetTargetDBInstanceIdentifier sets the TargetDBInstanceIdentifier field's value. +func (s *FailoverDBClusterInput) SetTargetDBInstanceIdentifier(v string) *FailoverDBClusterInput { + s.TargetDBInstanceIdentifier = &v + return s +} + +type FailoverDBClusterOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s FailoverDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FailoverDBClusterOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *FailoverDBClusterOutput) SetDBCluster(v *DBCluster) *FailoverDBClusterOutput { + s.DBCluster = v + return s +} + +// This type is not currently supported. +type Filter struct { + _ struct{} `type:"structure"` + + // This parameter is not currently supported. + // + // Name is a required field + Name *string `type:"string" required:"true"` + + // This parameter is not currently supported. + // + // Values is a required field + Values []*string `locationNameList:"Value" type:"list" required:"true"` +} + +// String returns the string representation +func (s Filter) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Filter) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *Filter) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "Filter"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + if s.Values == nil { + invalidParams.Add(request.NewErrParamRequired("Values")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *Filter) SetName(v string) *Filter { + s.Name = &v + return s +} + +// SetValues sets the Values field's value. +func (s *Filter) SetValues(v []*string) *Filter { + s.Values = v + return s +} + +type ListTagsForResourceInput struct { + _ struct{} `type:"structure"` + + // This parameter is not currently supported. + Filters []*Filter `locationNameList:"Filter" type:"list"` + + // The Amazon Neptune resource with tags to be listed. This value is an Amazon + // Resource Name (ARN). For information about creating an ARN, see Constructing + // an Amazon Resource Name (ARN) (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html#tagging.ARN.Constructing). + // + // ResourceName is a required field + ResourceName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s ListTagsForResourceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListTagsForResourceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListTagsForResourceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListTagsForResourceInput"} + if s.ResourceName == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceName")) + } + if s.Filters != nil { + for i, v := range s.Filters { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Filters", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilters sets the Filters field's value. +func (s *ListTagsForResourceInput) SetFilters(v []*Filter) *ListTagsForResourceInput { + s.Filters = v + return s +} + +// SetResourceName sets the ResourceName field's value. +func (s *ListTagsForResourceInput) SetResourceName(v string) *ListTagsForResourceInput { + s.ResourceName = &v + return s +} + +type ListTagsForResourceOutput struct { + _ struct{} `type:"structure"` + + // List of tags returned by the ListTagsForResource operation. + TagList []*Tag `locationNameList:"Tag" type:"list"` +} + +// String returns the string representation +func (s ListTagsForResourceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListTagsForResourceOutput) GoString() string { + return s.String() +} + +// SetTagList sets the TagList field's value. +func (s *ListTagsForResourceOutput) SetTagList(v []*Tag) *ListTagsForResourceOutput { + s.TagList = v + return s +} + +type ModifyDBClusterInput struct { + _ struct{} `type:"structure"` + + // A value that specifies whether the modifications in this request and any + // pending modifications are asynchronously applied as soon as possible, regardless + // of the PreferredMaintenanceWindow setting for the DB cluster. If this parameter + // is set to false, changes to the DB cluster are applied during the next maintenance + // window. + // + // The ApplyImmediately parameter only affects the NewDBClusterIdentifier and + // MasterUserPassword values. If you set the ApplyImmediately parameter value + // to false, then changes to the NewDBClusterIdentifier and MasterUserPassword + // values are applied during the next maintenance window. All other changes + // are applied immediately, regardless of the value of the ApplyImmediately + // parameter. + // + // Default: false + ApplyImmediately *bool `type:"boolean"` + + // The number of days for which automated backups are retained. You must specify + // a minimum value of 1. + // + // Default: 1 + // + // Constraints: + // + // * Must be a value from 1 to 35 + BackupRetentionPeriod *int64 `type:"integer"` + + // The DB cluster identifier for the cluster being modified. This parameter + // is not case-sensitive. + // + // Constraints: + // + // * Must match the identifier of an existing DBCluster. + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The name of the DB cluster parameter group to use for the DB cluster. + DBClusterParameterGroupName *string `type:"string"` + + // True to enable mapping of AWS Identity and Access Management (IAM) accounts + // to database accounts, and otherwise false. + // + // Default: false + EnableIAMDatabaseAuthentication *bool `type:"boolean"` + + // The version number of the database engine to which you want to upgrade. Changing + // this parameter results in an outage. The change is applied during the next + // maintenance window unless the ApplyImmediately parameter is set to true. + // + // For a list of valid engine versions, see CreateDBInstance, or call DescribeDBEngineVersions. + EngineVersion *string `type:"string"` + + // The new password for the master database user. This password can contain + // any printable ASCII character except "/", """, or "@". + // + // Constraints: Must contain from 8 to 41 characters. + MasterUserPassword *string `type:"string"` + + // The new DB cluster identifier for the DB cluster when renaming a DB cluster. + // This value is stored as a lowercase string. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens + // + // * The first character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // Example: my-cluster2 + NewDBClusterIdentifier *string `type:"string"` + + // A value that indicates that the DB cluster should be associated with the + // specified option group. Changing this parameter doesn't result in an outage + // except in the following case, and the change is applied during the next maintenance + // window unless the ApplyImmediately parameter is set to true for this request. + // If the parameter change results in an option group that enables OEM, this + // change can cause a brief (sub-second) period during which new connections + // are rejected but existing connections are not interrupted. + // + // Permanent options can't be removed from an option group. The option group + // can't be removed from a DB cluster once it is associated with a DB cluster. + OptionGroupName *string `type:"string"` + + // The port number on which the DB cluster accepts connections. + // + // Constraints: Value must be 1150-65535 + // + // Default: The same port as the original DB cluster. + Port *int64 `type:"integer"` + + // The daily time range during which automated backups are created if automated + // backups are enabled, using the BackupRetentionPeriod parameter. + // + // The default is a 30-minute window selected at random from an 8-hour block + // of time for each AWS Region. + // + // Constraints: + // + // * Must be in the format hh24:mi-hh24:mi. + // + // * Must be in Universal Coordinated Time (UTC). + // + // * Must not conflict with the preferred maintenance window. + // + // * Must be at least 30 minutes. + PreferredBackupWindow *string `type:"string"` + + // The weekly time range during which system maintenance can occur, in Universal + // Coordinated Time (UTC). + // + // Format: ddd:hh24:mi-ddd:hh24:mi + // + // The default is a 30-minute window selected at random from an 8-hour block + // of time for each AWS Region, occurring on a random day of the week. + // + // Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun. + // + // Constraints: Minimum 30-minute window. + PreferredMaintenanceWindow *string `type:"string"` + + // A list of VPC security groups that the DB cluster will belong to. + VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` +} + +// String returns the string representation +func (s ModifyDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyDBClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyDBClusterInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetApplyImmediately sets the ApplyImmediately field's value. +func (s *ModifyDBClusterInput) SetApplyImmediately(v bool) *ModifyDBClusterInput { + s.ApplyImmediately = &v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *ModifyDBClusterInput) SetBackupRetentionPeriod(v int64) *ModifyDBClusterInput { + s.BackupRetentionPeriod = &v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *ModifyDBClusterInput) SetDBClusterIdentifier(v string) *ModifyDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *ModifyDBClusterInput) SetDBClusterParameterGroupName(v string) *ModifyDBClusterInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. +func (s *ModifyDBClusterInput) SetEnableIAMDatabaseAuthentication(v bool) *ModifyDBClusterInput { + s.EnableIAMDatabaseAuthentication = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *ModifyDBClusterInput) SetEngineVersion(v string) *ModifyDBClusterInput { + s.EngineVersion = &v + return s +} + +// SetMasterUserPassword sets the MasterUserPassword field's value. +func (s *ModifyDBClusterInput) SetMasterUserPassword(v string) *ModifyDBClusterInput { + s.MasterUserPassword = &v + return s +} + +// SetNewDBClusterIdentifier sets the NewDBClusterIdentifier field's value. +func (s *ModifyDBClusterInput) SetNewDBClusterIdentifier(v string) *ModifyDBClusterInput { + s.NewDBClusterIdentifier = &v + return s +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *ModifyDBClusterInput) SetOptionGroupName(v string) *ModifyDBClusterInput { + s.OptionGroupName = &v + return s +} + +// SetPort sets the Port field's value. +func (s *ModifyDBClusterInput) SetPort(v int64) *ModifyDBClusterInput { + s.Port = &v + return s +} + +// SetPreferredBackupWindow sets the PreferredBackupWindow field's value. +func (s *ModifyDBClusterInput) SetPreferredBackupWindow(v string) *ModifyDBClusterInput { + s.PreferredBackupWindow = &v + return s +} + +// SetPreferredMaintenanceWindow sets the PreferredMaintenanceWindow field's value. +func (s *ModifyDBClusterInput) SetPreferredMaintenanceWindow(v string) *ModifyDBClusterInput { + s.PreferredMaintenanceWindow = &v + return s +} + +// SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. +func (s *ModifyDBClusterInput) SetVpcSecurityGroupIds(v []*string) *ModifyDBClusterInput { + s.VpcSecurityGroupIds = v + return s +} + +type ModifyDBClusterOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s ModifyDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBClusterOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *ModifyDBClusterOutput) SetDBCluster(v *DBCluster) *ModifyDBClusterOutput { + s.DBCluster = v + return s +} + +type ModifyDBClusterParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster parameter group to modify. + // + // DBClusterParameterGroupName is a required field + DBClusterParameterGroupName *string `type:"string" required:"true"` + + // A list of parameters in the DB cluster parameter group to modify. + // + // Parameters is a required field + Parameters []*Parameter `locationNameList:"Parameter" type:"list" required:"true"` +} + +// String returns the string representation +func (s ModifyDBClusterParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBClusterParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyDBClusterParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyDBClusterParameterGroupInput"} + if s.DBClusterParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterParameterGroupName")) + } + if s.Parameters == nil { + invalidParams.Add(request.NewErrParamRequired("Parameters")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *ModifyDBClusterParameterGroupInput) SetDBClusterParameterGroupName(v string) *ModifyDBClusterParameterGroupInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *ModifyDBClusterParameterGroupInput) SetParameters(v []*Parameter) *ModifyDBClusterParameterGroupInput { + s.Parameters = v + return s +} + +type ModifyDBClusterSnapshotAttributeInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster snapshot attribute to modify. + // + // To manage authorization for other AWS accounts to copy or restore a manual + // DB cluster snapshot, set this value to restore. + // + // AttributeName is a required field + AttributeName *string `type:"string" required:"true"` + + // The identifier for the DB cluster snapshot to modify the attributes for. + // + // DBClusterSnapshotIdentifier is a required field + DBClusterSnapshotIdentifier *string `type:"string" required:"true"` + + // A list of DB cluster snapshot attributes to add to the attribute specified + // by AttributeName. + // + // To authorize other AWS accounts to copy or restore a manual DB cluster snapshot, + // set this list to include one or more AWS account IDs, or all to make the + // manual DB cluster snapshot restorable by any AWS account. Do not add the + // all value for any manual DB cluster snapshots that contain private information + // that you don't want available to all AWS accounts. + ValuesToAdd []*string `locationNameList:"AttributeValue" type:"list"` + + // A list of DB cluster snapshot attributes to remove from the attribute specified + // by AttributeName. + // + // To remove authorization for other AWS accounts to copy or restore a manual + // DB cluster snapshot, set this list to include one or more AWS account identifiers, + // or all to remove authorization for any AWS account to copy or restore the + // DB cluster snapshot. If you specify all, an AWS account whose account ID + // is explicitly added to the restore attribute can still copy or restore a + // manual DB cluster snapshot. + ValuesToRemove []*string `locationNameList:"AttributeValue" type:"list"` +} + +// String returns the string representation +func (s ModifyDBClusterSnapshotAttributeInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBClusterSnapshotAttributeInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyDBClusterSnapshotAttributeInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyDBClusterSnapshotAttributeInput"} + if s.AttributeName == nil { + invalidParams.Add(request.NewErrParamRequired("AttributeName")) + } + if s.DBClusterSnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterSnapshotIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAttributeName sets the AttributeName field's value. +func (s *ModifyDBClusterSnapshotAttributeInput) SetAttributeName(v string) *ModifyDBClusterSnapshotAttributeInput { + s.AttributeName = &v + return s +} + +// SetDBClusterSnapshotIdentifier sets the DBClusterSnapshotIdentifier field's value. +func (s *ModifyDBClusterSnapshotAttributeInput) SetDBClusterSnapshotIdentifier(v string) *ModifyDBClusterSnapshotAttributeInput { + s.DBClusterSnapshotIdentifier = &v + return s +} + +// SetValuesToAdd sets the ValuesToAdd field's value. +func (s *ModifyDBClusterSnapshotAttributeInput) SetValuesToAdd(v []*string) *ModifyDBClusterSnapshotAttributeInput { + s.ValuesToAdd = v + return s +} + +// SetValuesToRemove sets the ValuesToRemove field's value. +func (s *ModifyDBClusterSnapshotAttributeInput) SetValuesToRemove(v []*string) *ModifyDBClusterSnapshotAttributeInput { + s.ValuesToRemove = v + return s +} + +type ModifyDBClusterSnapshotAttributeOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful call to the DescribeDBClusterSnapshotAttributes + // API action. + // + // Manual DB cluster snapshot attributes are used to authorize other AWS accounts + // to copy or restore a manual DB cluster snapshot. For more information, see + // the ModifyDBClusterSnapshotAttribute API action. + DBClusterSnapshotAttributesResult *DBClusterSnapshotAttributesResult `type:"structure"` +} + +// String returns the string representation +func (s ModifyDBClusterSnapshotAttributeOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBClusterSnapshotAttributeOutput) GoString() string { + return s.String() +} + +// SetDBClusterSnapshotAttributesResult sets the DBClusterSnapshotAttributesResult field's value. +func (s *ModifyDBClusterSnapshotAttributeOutput) SetDBClusterSnapshotAttributesResult(v *DBClusterSnapshotAttributesResult) *ModifyDBClusterSnapshotAttributeOutput { + s.DBClusterSnapshotAttributesResult = v + return s +} + +type ModifyDBInstanceInput struct { + _ struct{} `type:"structure"` + + // The new amount of storage (in gibibytes) to allocate for the DB instance. + // + // Not applicable. Storage is managed by the DB Cluster. + AllocatedStorage *int64 `type:"integer"` + + // Indicates that major version upgrades are allowed. Changing this parameter + // doesn't result in an outage and the change is asynchronously applied as soon + // as possible. + // + // Constraints: This parameter must be set to true when specifying a value for + // the EngineVersion parameter that is a different major version than the DB + // instance's current version. + AllowMajorVersionUpgrade *bool `type:"boolean"` + + // Specifies whether the modifications in this request and any pending modifications + // are asynchronously applied as soon as possible, regardless of the PreferredMaintenanceWindow + // setting for the DB instance. + // + // If this parameter is set to false, changes to the DB instance are applied + // during the next maintenance window. Some parameter changes can cause an outage + // and are applied on the next call to RebootDBInstance, or the next failure + // reboot. + // + // Default: false + ApplyImmediately *bool `type:"boolean"` + + // Indicates that minor version upgrades are applied automatically to the DB + // instance during the maintenance window. Changing this parameter doesn't result + // in an outage except in the following case and the change is asynchronously + // applied as soon as possible. An outage will result if this parameter is set + // to true during the maintenance window, and a newer minor version is available, + // and Neptune has enabled auto patching for that engine version. + AutoMinorVersionUpgrade *bool `type:"boolean"` + + // The number of days to retain automated backups. Setting this parameter to + // a positive number enables backups. Setting this parameter to 0 disables automated + // backups. + // + // Not applicable. The retention period for automated backups is managed by + // the DB cluster. For more information, see ModifyDBCluster. + // + // Default: Uses existing setting + BackupRetentionPeriod *int64 `type:"integer"` + + // Indicates the certificate that needs to be associated with the instance. + CACertificateIdentifier *string `type:"string"` + + // The configuration setting for the log types to be enabled for export to CloudWatch + // Logs for a specific DB instance or DB cluster. + CloudwatchLogsExportConfiguration *CloudwatchLogsExportConfiguration `type:"structure"` + + // True to copy all tags from the DB instance to snapshots of the DB instance, + // and otherwise false. The default is false. + CopyTagsToSnapshot *bool `type:"boolean"` + + // The new compute and memory capacity of the DB instance, for example, db.m4.large. + // Not all DB instance classes are available in all AWS Regions. + // + // If you modify the DB instance class, an outage occurs during the change. + // The change is applied during the next maintenance window, unless ApplyImmediately + // is specified as true for this request. + // + // Default: Uses existing setting + DBInstanceClass *string `type:"string"` + + // The DB instance identifier. This value is stored as a lowercase string. + // + // Constraints: + // + // * Must match the identifier of an existing DBInstance. + // + // DBInstanceIdentifier is a required field + DBInstanceIdentifier *string `type:"string" required:"true"` + + // The name of the DB parameter group to apply to the DB instance. Changing + // this setting doesn't result in an outage. The parameter group name itself + // is changed immediately, but the actual parameter changes are not applied + // until you reboot the instance without failover. The db instance will NOT + // be rebooted automatically and the parameter changes will NOT be applied during + // the next maintenance window. + // + // Default: Uses existing setting + // + // Constraints: The DB parameter group must be in the same DB parameter group + // family as this DB instance. + DBParameterGroupName *string `type:"string"` + + // The port number on which the database accepts connections. + // + // The value of the DBPortNumber parameter must not match any of the port values + // specified for options in the option group for the DB instance. + // + // Your database will restart when you change the DBPortNumber value regardless + // of the value of the ApplyImmediately parameter. + // + // Default: 8182 + DBPortNumber *int64 `type:"integer"` + + // A list of DB security groups to authorize on this DB instance. Changing this + // setting doesn't result in an outage and the change is asynchronously applied + // as soon as possible. + // + // Constraints: + // + // * If supplied, must match existing DBSecurityGroups. + DBSecurityGroups []*string `locationNameList:"DBSecurityGroupName" type:"list"` + + // The new DB subnet group for the DB instance. You can use this parameter to + // move your DB instance to a different VPC. + // + // Changing the subnet group causes an outage during the change. The change + // is applied during the next maintenance window, unless you specify true for + // the ApplyImmediately parameter. + // + // Constraints: If supplied, must match the name of an existing DBSubnetGroup. + // + // Example: mySubnetGroup + DBSubnetGroupName *string `type:"string"` + + // Not supported. + Domain *string `type:"string"` + + // Not supported + DomainIAMRoleName *string `type:"string"` + + // True to enable mapping of AWS Identity and Access Management (IAM) accounts + // to database accounts, and otherwise false. + // + // You can enable IAM database authentication for the following database engines + // + // Not applicable. Mapping AWS IAM accounts to database accounts is managed + // by the DB cluster. For more information, see ModifyDBCluster. + // + // Default: false + EnableIAMDatabaseAuthentication *bool `type:"boolean"` + + // True to enable Performance Insights for the DB instance, and otherwise false. + EnablePerformanceInsights *bool `type:"boolean"` + + // The version number of the database engine to upgrade to. Changing this parameter + // results in an outage and the change is applied during the next maintenance + // window unless the ApplyImmediately parameter is set to true for this request. + // + // For major version upgrades, if a nondefault DB parameter group is currently + // in use, a new DB parameter group in the DB parameter group family for the + // new engine version must be specified. The new DB parameter group can be the + // default for that DB parameter group family. + EngineVersion *string `type:"string"` + + // The new Provisioned IOPS (I/O operations per second) value for the instance. + // + // Changing this setting doesn't result in an outage and the change is applied + // during the next maintenance window unless the ApplyImmediately parameter + // is set to true for this request. + // + // Default: Uses existing setting + Iops *int64 `type:"integer"` + + // The license model for the DB instance. + // + // Valid values: license-included | bring-your-own-license | general-public-license + LicenseModel *string `type:"string"` + + // The new password for the master user. The password can include any printable + // ASCII character except "/", """, or "@". + // + // Not applicable. + // + // Default: Uses existing setting + MasterUserPassword *string `type:"string"` + + // The interval, in seconds, between points when Enhanced Monitoring metrics + // are collected for the DB instance. To disable collecting Enhanced Monitoring + // metrics, specify 0. The default is 0. + // + // If MonitoringRoleArn is specified, then you must also set MonitoringInterval + // to a value other than 0. + // + // Valid Values: 0, 1, 5, 10, 15, 30, 60 + MonitoringInterval *int64 `type:"integer"` + + // The ARN for the IAM role that permits Neptune to send enhanced monitoring + // metrics to Amazon CloudWatch Logs. For example, arn:aws:iam:123456789012:role/emaccess. + // + // If MonitoringInterval is set to a value other than 0, then you must supply + // a MonitoringRoleArn value. + MonitoringRoleArn *string `type:"string"` + + // Specifies if the DB instance is a Multi-AZ deployment. Changing this parameter + // doesn't result in an outage and the change is applied during the next maintenance + // window unless the ApplyImmediately parameter is set to true for this request. + MultiAZ *bool `type:"boolean"` + + // The new DB instance identifier for the DB instance when renaming a DB instance. + // When you change the DB instance identifier, an instance reboot will occur + // immediately if you set Apply Immediately to true, or will occur during the + // next maintenance window if Apply Immediately to false. This value is stored + // as a lowercase string. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens. + // + // * The first character must be a letter. + // + // * Cannot end with a hyphen or contain two consecutive hyphens. + // + // Example: mydbinstance + NewDBInstanceIdentifier *string `type:"string"` + + // Indicates that the DB instance should be associated with the specified option + // group. Changing this parameter doesn't result in an outage except in the + // following case and the change is applied during the next maintenance window + // unless the ApplyImmediately parameter is set to true for this request. If + // the parameter change results in an option group that enables OEM, this change + // can cause a brief (sub-second) period during which new connections are rejected + // but existing connections are not interrupted. + // + // Permanent options, such as the TDE option for Oracle Advanced Security TDE, + // can't be removed from an option group, and that option group can't be removed + // from a DB instance once it is associated with a DB instance + OptionGroupName *string `type:"string"` + + // The AWS KMS key identifier for encryption of Performance Insights data. The + // KMS key ID is the Amazon Resource Name (ARN), KMS key identifier, or the + // KMS key alias for the KMS encryption key. + PerformanceInsightsKMSKeyId *string `type:"string"` + + // The daily time range during which automated backups are created if automated + // backups are enabled. + // + // Not applicable. The daily time range for creating automated backups is managed + // by the DB cluster. For more information, see ModifyDBCluster. + // + // Constraints: + // + // * Must be in the format hh24:mi-hh24:mi + // + // * Must be in Universal Time Coordinated (UTC) + // + // * Must not conflict with the preferred maintenance window + // + // * Must be at least 30 minutes + PreferredBackupWindow *string `type:"string"` + + // The weekly time range (in UTC) during which system maintenance can occur, + // which might result in an outage. Changing this parameter doesn't result in + // an outage, except in the following situation, and the change is asynchronously + // applied as soon as possible. If there are pending actions that cause a reboot, + // and the maintenance window is changed to include the current time, then changing + // this parameter will cause a reboot of the DB instance. If moving this window + // to the current time, there must be at least 30 minutes between the current + // time and end of the window to ensure pending changes are applied. + // + // Default: Uses existing setting + // + // Format: ddd:hh24:mi-ddd:hh24:mi + // + // Valid Days: Mon | Tue | Wed | Thu | Fri | Sat | Sun + // + // Constraints: Must be at least 30 minutes + PreferredMaintenanceWindow *string `type:"string"` + + // A value that specifies the order in which a Read Replica is promoted to the + // primary instance after a failure of the existing primary instance. + // + // Default: 1 + // + // Valid Values: 0 - 15 + PromotionTier *int64 `type:"integer"` + + // This parameter is not supported. + PubliclyAccessible *bool `deprecated:"true" type:"boolean"` + + // Specifies the storage type to be associated with the DB instance. + // + // If you specify Provisioned IOPS (io1), you must also include a value for + // the Iops parameter. + // + // If you choose to migrate your DB instance from using standard storage to + // using Provisioned IOPS, or from using Provisioned IOPS to using standard + // storage, the process can take time. The duration of the migration depends + // on several factors such as database load, storage size, storage type (standard + // or Provisioned IOPS), amount of IOPS provisioned (if any), and the number + // of prior scale storage operations. Typical migration times are under 24 hours, + // but the process can take up to several days in some cases. During the migration, + // the DB instance is available for use, but might experience performance degradation. + // While the migration takes place, nightly backups for the instance are suspended. + // No other Amazon Neptune operations can take place for the instance, including + // modifying the instance, rebooting the instance, deleting the instance, creating + // a Read Replica for the instance, and creating a DB snapshot of the instance. + // + // Valid values: standard | gp2 | io1 + // + // Default: io1 if the Iops parameter is specified, otherwise standard + StorageType *string `type:"string"` + + // The ARN from the key store with which to associate the instance for TDE encryption. + TdeCredentialArn *string `type:"string"` + + // The password for the given ARN from the key store in order to access the + // device. + TdeCredentialPassword *string `type:"string"` + + // A list of EC2 VPC security groups to authorize on this DB instance. This + // change is asynchronously applied as soon as possible. + // + // Not applicable. The associated list of EC2 VPC security groups is managed + // by the DB cluster. For more information, see ModifyDBCluster. + // + // Constraints: + // + // * If supplied, must match existing VpcSecurityGroupIds. + VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` +} + +// String returns the string representation +func (s ModifyDBInstanceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBInstanceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyDBInstanceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyDBInstanceInput"} + if s.DBInstanceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBInstanceIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAllocatedStorage sets the AllocatedStorage field's value. +func (s *ModifyDBInstanceInput) SetAllocatedStorage(v int64) *ModifyDBInstanceInput { + s.AllocatedStorage = &v + return s +} + +// SetAllowMajorVersionUpgrade sets the AllowMajorVersionUpgrade field's value. +func (s *ModifyDBInstanceInput) SetAllowMajorVersionUpgrade(v bool) *ModifyDBInstanceInput { + s.AllowMajorVersionUpgrade = &v + return s +} + +// SetApplyImmediately sets the ApplyImmediately field's value. +func (s *ModifyDBInstanceInput) SetApplyImmediately(v bool) *ModifyDBInstanceInput { + s.ApplyImmediately = &v + return s +} + +// SetAutoMinorVersionUpgrade sets the AutoMinorVersionUpgrade field's value. +func (s *ModifyDBInstanceInput) SetAutoMinorVersionUpgrade(v bool) *ModifyDBInstanceInput { + s.AutoMinorVersionUpgrade = &v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *ModifyDBInstanceInput) SetBackupRetentionPeriod(v int64) *ModifyDBInstanceInput { + s.BackupRetentionPeriod = &v + return s +} + +// SetCACertificateIdentifier sets the CACertificateIdentifier field's value. +func (s *ModifyDBInstanceInput) SetCACertificateIdentifier(v string) *ModifyDBInstanceInput { + s.CACertificateIdentifier = &v + return s +} + +// SetCloudwatchLogsExportConfiguration sets the CloudwatchLogsExportConfiguration field's value. +func (s *ModifyDBInstanceInput) SetCloudwatchLogsExportConfiguration(v *CloudwatchLogsExportConfiguration) *ModifyDBInstanceInput { + s.CloudwatchLogsExportConfiguration = v + return s +} + +// SetCopyTagsToSnapshot sets the CopyTagsToSnapshot field's value. +func (s *ModifyDBInstanceInput) SetCopyTagsToSnapshot(v bool) *ModifyDBInstanceInput { + s.CopyTagsToSnapshot = &v + return s +} + +// SetDBInstanceClass sets the DBInstanceClass field's value. +func (s *ModifyDBInstanceInput) SetDBInstanceClass(v string) *ModifyDBInstanceInput { + s.DBInstanceClass = &v + return s +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *ModifyDBInstanceInput) SetDBInstanceIdentifier(v string) *ModifyDBInstanceInput { + s.DBInstanceIdentifier = &v + return s +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *ModifyDBInstanceInput) SetDBParameterGroupName(v string) *ModifyDBInstanceInput { + s.DBParameterGroupName = &v + return s +} + +// SetDBPortNumber sets the DBPortNumber field's value. +func (s *ModifyDBInstanceInput) SetDBPortNumber(v int64) *ModifyDBInstanceInput { + s.DBPortNumber = &v + return s +} + +// SetDBSecurityGroups sets the DBSecurityGroups field's value. +func (s *ModifyDBInstanceInput) SetDBSecurityGroups(v []*string) *ModifyDBInstanceInput { + s.DBSecurityGroups = v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *ModifyDBInstanceInput) SetDBSubnetGroupName(v string) *ModifyDBInstanceInput { + s.DBSubnetGroupName = &v + return s +} + +// SetDomain sets the Domain field's value. +func (s *ModifyDBInstanceInput) SetDomain(v string) *ModifyDBInstanceInput { + s.Domain = &v + return s +} + +// SetDomainIAMRoleName sets the DomainIAMRoleName field's value. +func (s *ModifyDBInstanceInput) SetDomainIAMRoleName(v string) *ModifyDBInstanceInput { + s.DomainIAMRoleName = &v + return s +} + +// SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. +func (s *ModifyDBInstanceInput) SetEnableIAMDatabaseAuthentication(v bool) *ModifyDBInstanceInput { + s.EnableIAMDatabaseAuthentication = &v + return s +} + +// SetEnablePerformanceInsights sets the EnablePerformanceInsights field's value. +func (s *ModifyDBInstanceInput) SetEnablePerformanceInsights(v bool) *ModifyDBInstanceInput { + s.EnablePerformanceInsights = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *ModifyDBInstanceInput) SetEngineVersion(v string) *ModifyDBInstanceInput { + s.EngineVersion = &v + return s +} + +// SetIops sets the Iops field's value. +func (s *ModifyDBInstanceInput) SetIops(v int64) *ModifyDBInstanceInput { + s.Iops = &v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *ModifyDBInstanceInput) SetLicenseModel(v string) *ModifyDBInstanceInput { + s.LicenseModel = &v + return s +} + +// SetMasterUserPassword sets the MasterUserPassword field's value. +func (s *ModifyDBInstanceInput) SetMasterUserPassword(v string) *ModifyDBInstanceInput { + s.MasterUserPassword = &v + return s +} + +// SetMonitoringInterval sets the MonitoringInterval field's value. +func (s *ModifyDBInstanceInput) SetMonitoringInterval(v int64) *ModifyDBInstanceInput { + s.MonitoringInterval = &v + return s +} + +// SetMonitoringRoleArn sets the MonitoringRoleArn field's value. +func (s *ModifyDBInstanceInput) SetMonitoringRoleArn(v string) *ModifyDBInstanceInput { + s.MonitoringRoleArn = &v + return s +} + +// SetMultiAZ sets the MultiAZ field's value. +func (s *ModifyDBInstanceInput) SetMultiAZ(v bool) *ModifyDBInstanceInput { + s.MultiAZ = &v + return s +} + +// SetNewDBInstanceIdentifier sets the NewDBInstanceIdentifier field's value. +func (s *ModifyDBInstanceInput) SetNewDBInstanceIdentifier(v string) *ModifyDBInstanceInput { + s.NewDBInstanceIdentifier = &v + return s +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *ModifyDBInstanceInput) SetOptionGroupName(v string) *ModifyDBInstanceInput { + s.OptionGroupName = &v + return s +} + +// SetPerformanceInsightsKMSKeyId sets the PerformanceInsightsKMSKeyId field's value. +func (s *ModifyDBInstanceInput) SetPerformanceInsightsKMSKeyId(v string) *ModifyDBInstanceInput { + s.PerformanceInsightsKMSKeyId = &v + return s +} + +// SetPreferredBackupWindow sets the PreferredBackupWindow field's value. +func (s *ModifyDBInstanceInput) SetPreferredBackupWindow(v string) *ModifyDBInstanceInput { + s.PreferredBackupWindow = &v + return s +} + +// SetPreferredMaintenanceWindow sets the PreferredMaintenanceWindow field's value. +func (s *ModifyDBInstanceInput) SetPreferredMaintenanceWindow(v string) *ModifyDBInstanceInput { + s.PreferredMaintenanceWindow = &v + return s +} + +// SetPromotionTier sets the PromotionTier field's value. +func (s *ModifyDBInstanceInput) SetPromotionTier(v int64) *ModifyDBInstanceInput { + s.PromotionTier = &v + return s +} + +// SetPubliclyAccessible sets the PubliclyAccessible field's value. +func (s *ModifyDBInstanceInput) SetPubliclyAccessible(v bool) *ModifyDBInstanceInput { + s.PubliclyAccessible = &v + return s +} + +// SetStorageType sets the StorageType field's value. +func (s *ModifyDBInstanceInput) SetStorageType(v string) *ModifyDBInstanceInput { + s.StorageType = &v + return s +} + +// SetTdeCredentialArn sets the TdeCredentialArn field's value. +func (s *ModifyDBInstanceInput) SetTdeCredentialArn(v string) *ModifyDBInstanceInput { + s.TdeCredentialArn = &v + return s +} + +// SetTdeCredentialPassword sets the TdeCredentialPassword field's value. +func (s *ModifyDBInstanceInput) SetTdeCredentialPassword(v string) *ModifyDBInstanceInput { + s.TdeCredentialPassword = &v + return s +} + +// SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. +func (s *ModifyDBInstanceInput) SetVpcSecurityGroupIds(v []*string) *ModifyDBInstanceInput { + s.VpcSecurityGroupIds = v + return s +} + +type ModifyDBInstanceOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB instance. + // + // This data type is used as a response element in the DescribeDBInstances action. + DBInstance *DBInstance `type:"structure"` +} + +// String returns the string representation +func (s ModifyDBInstanceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBInstanceOutput) GoString() string { + return s.String() +} + +// SetDBInstance sets the DBInstance field's value. +func (s *ModifyDBInstanceOutput) SetDBInstance(v *DBInstance) *ModifyDBInstanceOutput { + s.DBInstance = v + return s +} + +type ModifyDBParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB parameter group. + // + // Constraints: + // + // * If supplied, must match the name of an existing DBParameterGroup. + // + // DBParameterGroupName is a required field + DBParameterGroupName *string `type:"string" required:"true"` + + // An array of parameter names, values, and the apply method for the parameter + // update. At least one parameter name, value, and apply method must be supplied; + // subsequent arguments are optional. A maximum of 20 parameters can be modified + // in a single request. + // + // Valid Values (for the application method): immediate | pending-reboot + // + // You can use the immediate value with dynamic parameters only. You can use + // the pending-reboot value for both dynamic and static parameters, and changes + // are applied when you reboot the DB instance without failover. + // + // Parameters is a required field + Parameters []*Parameter `locationNameList:"Parameter" type:"list" required:"true"` +} + +// String returns the string representation +func (s ModifyDBParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyDBParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyDBParameterGroupInput"} + if s.DBParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupName")) + } + if s.Parameters == nil { + invalidParams.Add(request.NewErrParamRequired("Parameters")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *ModifyDBParameterGroupInput) SetDBParameterGroupName(v string) *ModifyDBParameterGroupInput { + s.DBParameterGroupName = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *ModifyDBParameterGroupInput) SetParameters(v []*Parameter) *ModifyDBParameterGroupInput { + s.Parameters = v + return s +} + +type ModifyDBSubnetGroupInput struct { + _ struct{} `type:"structure"` + + // The description for the DB subnet group. + DBSubnetGroupDescription *string `type:"string"` + + // The name for the DB subnet group. This value is stored as a lowercase string. + // You can't modify the default subnet group. + // + // Constraints: Must match the name of an existing DBSubnetGroup. Must not be + // default. + // + // Example: mySubnetgroup + // + // DBSubnetGroupName is a required field + DBSubnetGroupName *string `type:"string" required:"true"` + + // The EC2 subnet IDs for the DB subnet group. + // + // SubnetIds is a required field + SubnetIds []*string `locationNameList:"SubnetIdentifier" type:"list" required:"true"` +} + +// String returns the string representation +func (s ModifyDBSubnetGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBSubnetGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyDBSubnetGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyDBSubnetGroupInput"} + if s.DBSubnetGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBSubnetGroupName")) + } + if s.SubnetIds == nil { + invalidParams.Add(request.NewErrParamRequired("SubnetIds")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBSubnetGroupDescription sets the DBSubnetGroupDescription field's value. +func (s *ModifyDBSubnetGroupInput) SetDBSubnetGroupDescription(v string) *ModifyDBSubnetGroupInput { + s.DBSubnetGroupDescription = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *ModifyDBSubnetGroupInput) SetDBSubnetGroupName(v string) *ModifyDBSubnetGroupInput { + s.DBSubnetGroupName = &v + return s +} + +// SetSubnetIds sets the SubnetIds field's value. +func (s *ModifyDBSubnetGroupInput) SetSubnetIds(v []*string) *ModifyDBSubnetGroupInput { + s.SubnetIds = v + return s +} + +type ModifyDBSubnetGroupOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB subnet group. + // + // This data type is used as a response element in the DescribeDBSubnetGroups + // action. + DBSubnetGroup *DBSubnetGroup `type:"structure"` +} + +// String returns the string representation +func (s ModifyDBSubnetGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyDBSubnetGroupOutput) GoString() string { + return s.String() +} + +// SetDBSubnetGroup sets the DBSubnetGroup field's value. +func (s *ModifyDBSubnetGroupOutput) SetDBSubnetGroup(v *DBSubnetGroup) *ModifyDBSubnetGroupOutput { + s.DBSubnetGroup = v + return s +} + +type ModifyEventSubscriptionInput struct { + _ struct{} `type:"structure"` + + // A Boolean value; set to true to activate the subscription. + Enabled *bool `type:"boolean"` + + // A list of event categories for a SourceType that you want to subscribe to. + // You can see a list of the categories for a given SourceType by using the + // DescribeEventCategories action. + EventCategories []*string `locationNameList:"EventCategory" type:"list"` + + // The Amazon Resource Name (ARN) of the SNS topic created for event notification. + // The ARN is created by Amazon SNS when you create a topic and subscribe to + // it. + SnsTopicArn *string `type:"string"` + + // The type of source that is generating the events. For example, if you want + // to be notified of events generated by a DB instance, you would set this parameter + // to db-instance. if this value is not specified, all events are returned. + // + // Valid values: db-instance | db-parameter-group | db-security-group | db-snapshot + SourceType *string `type:"string"` + + // The name of the event notification subscription. + // + // SubscriptionName is a required field + SubscriptionName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s ModifyEventSubscriptionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyEventSubscriptionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ModifyEventSubscriptionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ModifyEventSubscriptionInput"} + if s.SubscriptionName == nil { + invalidParams.Add(request.NewErrParamRequired("SubscriptionName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetEnabled sets the Enabled field's value. +func (s *ModifyEventSubscriptionInput) SetEnabled(v bool) *ModifyEventSubscriptionInput { + s.Enabled = &v + return s +} + +// SetEventCategories sets the EventCategories field's value. +func (s *ModifyEventSubscriptionInput) SetEventCategories(v []*string) *ModifyEventSubscriptionInput { + s.EventCategories = v + return s +} + +// SetSnsTopicArn sets the SnsTopicArn field's value. +func (s *ModifyEventSubscriptionInput) SetSnsTopicArn(v string) *ModifyEventSubscriptionInput { + s.SnsTopicArn = &v + return s +} + +// SetSourceType sets the SourceType field's value. +func (s *ModifyEventSubscriptionInput) SetSourceType(v string) *ModifyEventSubscriptionInput { + s.SourceType = &v + return s +} + +// SetSubscriptionName sets the SubscriptionName field's value. +func (s *ModifyEventSubscriptionInput) SetSubscriptionName(v string) *ModifyEventSubscriptionInput { + s.SubscriptionName = &v + return s +} + +type ModifyEventSubscriptionOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful invocation of the DescribeEventSubscriptions + // action. + EventSubscription *EventSubscription `type:"structure"` +} + +// String returns the string representation +func (s ModifyEventSubscriptionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ModifyEventSubscriptionOutput) GoString() string { + return s.String() +} + +// SetEventSubscription sets the EventSubscription field's value. +func (s *ModifyEventSubscriptionOutput) SetEventSubscription(v *EventSubscription) *ModifyEventSubscriptionOutput { + s.EventSubscription = v + return s +} + +// Provides information on the option groups the DB instance is a member of. +type OptionGroupMembership struct { + _ struct{} `type:"structure"` + + // The name of the option group that the instance belongs to. + OptionGroupName *string `type:"string"` + + // The status of the DB instance's option group membership. Valid values are: + // in-sync, pending-apply, pending-removal, pending-maintenance-apply, pending-maintenance-removal, + // applying, removing, and failed. + Status *string `type:"string"` +} + +// String returns the string representation +func (s OptionGroupMembership) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s OptionGroupMembership) GoString() string { + return s.String() +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *OptionGroupMembership) SetOptionGroupName(v string) *OptionGroupMembership { + s.OptionGroupName = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *OptionGroupMembership) SetStatus(v string) *OptionGroupMembership { + s.Status = &v + return s +} + +// Contains a list of available options for a DB instance. +// +// This data type is used as a response element in the DescribeOrderableDBInstanceOptions +// action. +type OrderableDBInstanceOption struct { + _ struct{} `type:"structure"` + + // A list of Availability Zones for a DB instance. + AvailabilityZones []*AvailabilityZone `locationNameList:"AvailabilityZone" type:"list"` + + // The DB instance class for a DB instance. + DBInstanceClass *string `type:"string"` + + // The engine type of a DB instance. + Engine *string `type:"string"` + + // The engine version of a DB instance. + EngineVersion *string `type:"string"` + + // The license model for a DB instance. + LicenseModel *string `type:"string"` + + // Maximum total provisioned IOPS for a DB instance. + MaxIopsPerDbInstance *int64 `type:"integer"` + + // Maximum provisioned IOPS per GiB for a DB instance. + MaxIopsPerGib *float64 `type:"double"` + + // Maximum storage size for a DB instance. + MaxStorageSize *int64 `type:"integer"` + + // Minimum total provisioned IOPS for a DB instance. + MinIopsPerDbInstance *int64 `type:"integer"` + + // Minimum provisioned IOPS per GiB for a DB instance. + MinIopsPerGib *float64 `type:"double"` + + // Minimum storage size for a DB instance. + MinStorageSize *int64 `type:"integer"` + + // Indicates whether a DB instance is Multi-AZ capable. + MultiAZCapable *bool `type:"boolean"` + + // Indicates whether a DB instance can have a Read Replica. + ReadReplicaCapable *bool `type:"boolean"` + + // Indicates the storage type for a DB instance. + StorageType *string `type:"string"` + + // Indicates whether a DB instance supports Enhanced Monitoring at intervals + // from 1 to 60 seconds. + SupportsEnhancedMonitoring *bool `type:"boolean"` + + // Indicates whether a DB instance supports IAM database authentication. + SupportsIAMDatabaseAuthentication *bool `type:"boolean"` + + // Indicates whether a DB instance supports provisioned IOPS. + SupportsIops *bool `type:"boolean"` + + // True if a DB instance supports Performance Insights, otherwise false. + SupportsPerformanceInsights *bool `type:"boolean"` + + // Indicates whether a DB instance supports encrypted storage. + SupportsStorageEncryption *bool `type:"boolean"` + + // Indicates whether a DB instance is in a VPC. + Vpc *bool `type:"boolean"` +} + +// String returns the string representation +func (s OrderableDBInstanceOption) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s OrderableDBInstanceOption) GoString() string { + return s.String() +} + +// SetAvailabilityZones sets the AvailabilityZones field's value. +func (s *OrderableDBInstanceOption) SetAvailabilityZones(v []*AvailabilityZone) *OrderableDBInstanceOption { + s.AvailabilityZones = v + return s +} + +// SetDBInstanceClass sets the DBInstanceClass field's value. +func (s *OrderableDBInstanceOption) SetDBInstanceClass(v string) *OrderableDBInstanceOption { + s.DBInstanceClass = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *OrderableDBInstanceOption) SetEngine(v string) *OrderableDBInstanceOption { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *OrderableDBInstanceOption) SetEngineVersion(v string) *OrderableDBInstanceOption { + s.EngineVersion = &v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *OrderableDBInstanceOption) SetLicenseModel(v string) *OrderableDBInstanceOption { + s.LicenseModel = &v + return s +} + +// SetMaxIopsPerDbInstance sets the MaxIopsPerDbInstance field's value. +func (s *OrderableDBInstanceOption) SetMaxIopsPerDbInstance(v int64) *OrderableDBInstanceOption { + s.MaxIopsPerDbInstance = &v + return s +} + +// SetMaxIopsPerGib sets the MaxIopsPerGib field's value. +func (s *OrderableDBInstanceOption) SetMaxIopsPerGib(v float64) *OrderableDBInstanceOption { + s.MaxIopsPerGib = &v + return s +} + +// SetMaxStorageSize sets the MaxStorageSize field's value. +func (s *OrderableDBInstanceOption) SetMaxStorageSize(v int64) *OrderableDBInstanceOption { + s.MaxStorageSize = &v + return s +} + +// SetMinIopsPerDbInstance sets the MinIopsPerDbInstance field's value. +func (s *OrderableDBInstanceOption) SetMinIopsPerDbInstance(v int64) *OrderableDBInstanceOption { + s.MinIopsPerDbInstance = &v + return s +} + +// SetMinIopsPerGib sets the MinIopsPerGib field's value. +func (s *OrderableDBInstanceOption) SetMinIopsPerGib(v float64) *OrderableDBInstanceOption { + s.MinIopsPerGib = &v + return s +} + +// SetMinStorageSize sets the MinStorageSize field's value. +func (s *OrderableDBInstanceOption) SetMinStorageSize(v int64) *OrderableDBInstanceOption { + s.MinStorageSize = &v + return s +} + +// SetMultiAZCapable sets the MultiAZCapable field's value. +func (s *OrderableDBInstanceOption) SetMultiAZCapable(v bool) *OrderableDBInstanceOption { + s.MultiAZCapable = &v + return s +} + +// SetReadReplicaCapable sets the ReadReplicaCapable field's value. +func (s *OrderableDBInstanceOption) SetReadReplicaCapable(v bool) *OrderableDBInstanceOption { + s.ReadReplicaCapable = &v + return s +} + +// SetStorageType sets the StorageType field's value. +func (s *OrderableDBInstanceOption) SetStorageType(v string) *OrderableDBInstanceOption { + s.StorageType = &v + return s +} + +// SetSupportsEnhancedMonitoring sets the SupportsEnhancedMonitoring field's value. +func (s *OrderableDBInstanceOption) SetSupportsEnhancedMonitoring(v bool) *OrderableDBInstanceOption { + s.SupportsEnhancedMonitoring = &v + return s +} + +// SetSupportsIAMDatabaseAuthentication sets the SupportsIAMDatabaseAuthentication field's value. +func (s *OrderableDBInstanceOption) SetSupportsIAMDatabaseAuthentication(v bool) *OrderableDBInstanceOption { + s.SupportsIAMDatabaseAuthentication = &v + return s +} + +// SetSupportsIops sets the SupportsIops field's value. +func (s *OrderableDBInstanceOption) SetSupportsIops(v bool) *OrderableDBInstanceOption { + s.SupportsIops = &v + return s +} + +// SetSupportsPerformanceInsights sets the SupportsPerformanceInsights field's value. +func (s *OrderableDBInstanceOption) SetSupportsPerformanceInsights(v bool) *OrderableDBInstanceOption { + s.SupportsPerformanceInsights = &v + return s +} + +// SetSupportsStorageEncryption sets the SupportsStorageEncryption field's value. +func (s *OrderableDBInstanceOption) SetSupportsStorageEncryption(v bool) *OrderableDBInstanceOption { + s.SupportsStorageEncryption = &v + return s +} + +// SetVpc sets the Vpc field's value. +func (s *OrderableDBInstanceOption) SetVpc(v bool) *OrderableDBInstanceOption { + s.Vpc = &v + return s +} + +// This data type is used as a request parameter in the ModifyDBParameterGroup +// and ResetDBParameterGroup actions. +// +// This data type is used as a response element in the DescribeEngineDefaultParameters +// and DescribeDBParameters actions. +type Parameter struct { + _ struct{} `type:"structure"` + + // Specifies the valid range of values for the parameter. + AllowedValues *string `type:"string"` + + // Indicates when to apply parameter updates. + ApplyMethod *string `type:"string" enum:"ApplyMethod"` + + // Specifies the engine specific parameters type. + ApplyType *string `type:"string"` + + // Specifies the valid data type for the parameter. + DataType *string `type:"string"` + + // Provides a description of the parameter. + Description *string `type:"string"` + + // Indicates whether (true) or not (false) the parameter can be modified. Some + // parameters have security or operational implications that prevent them from + // being changed. + IsModifiable *bool `type:"boolean"` + + // The earliest engine version to which the parameter can apply. + MinimumEngineVersion *string `type:"string"` + + // Specifies the name of the parameter. + ParameterName *string `type:"string"` + + // Specifies the value of the parameter. + ParameterValue *string `type:"string"` + + // Indicates the source of the parameter value. + Source *string `type:"string"` +} + +// String returns the string representation +func (s Parameter) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Parameter) GoString() string { + return s.String() +} + +// SetAllowedValues sets the AllowedValues field's value. +func (s *Parameter) SetAllowedValues(v string) *Parameter { + s.AllowedValues = &v + return s +} + +// SetApplyMethod sets the ApplyMethod field's value. +func (s *Parameter) SetApplyMethod(v string) *Parameter { + s.ApplyMethod = &v + return s +} + +// SetApplyType sets the ApplyType field's value. +func (s *Parameter) SetApplyType(v string) *Parameter { + s.ApplyType = &v + return s +} + +// SetDataType sets the DataType field's value. +func (s *Parameter) SetDataType(v string) *Parameter { + s.DataType = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *Parameter) SetDescription(v string) *Parameter { + s.Description = &v + return s +} + +// SetIsModifiable sets the IsModifiable field's value. +func (s *Parameter) SetIsModifiable(v bool) *Parameter { + s.IsModifiable = &v + return s +} + +// SetMinimumEngineVersion sets the MinimumEngineVersion field's value. +func (s *Parameter) SetMinimumEngineVersion(v string) *Parameter { + s.MinimumEngineVersion = &v + return s +} + +// SetParameterName sets the ParameterName field's value. +func (s *Parameter) SetParameterName(v string) *Parameter { + s.ParameterName = &v + return s +} + +// SetParameterValue sets the ParameterValue field's value. +func (s *Parameter) SetParameterValue(v string) *Parameter { + s.ParameterValue = &v + return s +} + +// SetSource sets the Source field's value. +func (s *Parameter) SetSource(v string) *Parameter { + s.Source = &v + return s +} + +// A list of the log types whose configuration is still pending. In other words, +// these log types are in the process of being activated or deactivated. +type PendingCloudwatchLogsExports struct { + _ struct{} `type:"structure"` + + // Log types that are in the process of being enabled. After they are enabled, + // these log types are exported to CloudWatch Logs. + LogTypesToDisable []*string `type:"list"` + + // Log types that are in the process of being deactivated. After they are deactivated, + // these log types aren't exported to CloudWatch Logs. + LogTypesToEnable []*string `type:"list"` +} + +// String returns the string representation +func (s PendingCloudwatchLogsExports) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PendingCloudwatchLogsExports) GoString() string { + return s.String() +} + +// SetLogTypesToDisable sets the LogTypesToDisable field's value. +func (s *PendingCloudwatchLogsExports) SetLogTypesToDisable(v []*string) *PendingCloudwatchLogsExports { + s.LogTypesToDisable = v + return s +} + +// SetLogTypesToEnable sets the LogTypesToEnable field's value. +func (s *PendingCloudwatchLogsExports) SetLogTypesToEnable(v []*string) *PendingCloudwatchLogsExports { + s.LogTypesToEnable = v + return s +} + +// Provides information about a pending maintenance action for a resource. +type PendingMaintenanceAction struct { + _ struct{} `type:"structure"` + + // The type of pending maintenance action that is available for the resource. + Action *string `type:"string"` + + // The date of the maintenance window when the action is applied. The maintenance + // action is applied to the resource during its first maintenance window after + // this date. If this date is specified, any next-maintenance opt-in requests + // are ignored. + AutoAppliedAfterDate *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // The effective date when the pending maintenance action is applied to the + // resource. This date takes into account opt-in requests received from the + // ApplyPendingMaintenanceAction API, the AutoAppliedAfterDate, and the ForcedApplyDate. + // This value is blank if an opt-in request has not been received and nothing + // has been specified as AutoAppliedAfterDate or ForcedApplyDate. + CurrentApplyDate *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // A description providing more detail about the maintenance action. + Description *string `type:"string"` + + // The date when the maintenance action is automatically applied. The maintenance + // action is applied to the resource on this date regardless of the maintenance + // window for the resource. If this date is specified, any immediate opt-in + // requests are ignored. + ForcedApplyDate *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // Indicates the type of opt-in request that has been received for the resource. + OptInStatus *string `type:"string"` +} + +// String returns the string representation +func (s PendingMaintenanceAction) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PendingMaintenanceAction) GoString() string { + return s.String() +} + +// SetAction sets the Action field's value. +func (s *PendingMaintenanceAction) SetAction(v string) *PendingMaintenanceAction { + s.Action = &v + return s +} + +// SetAutoAppliedAfterDate sets the AutoAppliedAfterDate field's value. +func (s *PendingMaintenanceAction) SetAutoAppliedAfterDate(v time.Time) *PendingMaintenanceAction { + s.AutoAppliedAfterDate = &v + return s +} + +// SetCurrentApplyDate sets the CurrentApplyDate field's value. +func (s *PendingMaintenanceAction) SetCurrentApplyDate(v time.Time) *PendingMaintenanceAction { + s.CurrentApplyDate = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *PendingMaintenanceAction) SetDescription(v string) *PendingMaintenanceAction { + s.Description = &v + return s +} + +// SetForcedApplyDate sets the ForcedApplyDate field's value. +func (s *PendingMaintenanceAction) SetForcedApplyDate(v time.Time) *PendingMaintenanceAction { + s.ForcedApplyDate = &v + return s +} + +// SetOptInStatus sets the OptInStatus field's value. +func (s *PendingMaintenanceAction) SetOptInStatus(v string) *PendingMaintenanceAction { + s.OptInStatus = &v + return s +} + +// This data type is used as a response element in the ModifyDBInstance action. +type PendingModifiedValues struct { + _ struct{} `type:"structure"` + + // Contains the new AllocatedStorage size for the DB instance that will be applied + // or is currently being applied. + AllocatedStorage *int64 `type:"integer"` + + // Specifies the pending number of days for which automated backups are retained. + BackupRetentionPeriod *int64 `type:"integer"` + + // Specifies the identifier of the CA certificate for the DB instance. + CACertificateIdentifier *string `type:"string"` + + // Contains the new DBInstanceClass for the DB instance that will be applied + // or is currently being applied. + DBInstanceClass *string `type:"string"` + + // Contains the new DBInstanceIdentifier for the DB instance that will be applied + // or is currently being applied. + DBInstanceIdentifier *string `type:"string"` + + // The new DB subnet group for the DB instance. + DBSubnetGroupName *string `type:"string"` + + // Indicates the database engine version. + EngineVersion *string `type:"string"` + + // Specifies the new Provisioned IOPS value for the DB instance that will be + // applied or is currently being applied. + Iops *int64 `type:"integer"` + + // The license model for the DB instance. + // + // Valid values: license-included | bring-your-own-license | general-public-license + LicenseModel *string `type:"string"` + + // Contains the pending or currently-in-progress change of the master credentials + // for the DB instance. + MasterUserPassword *string `type:"string"` + + // Indicates that the Single-AZ DB instance is to change to a Multi-AZ deployment. + MultiAZ *bool `type:"boolean"` + + // A list of the log types whose configuration is still pending. In other words, + // these log types are in the process of being activated or deactivated. + PendingCloudwatchLogsExports *PendingCloudwatchLogsExports `type:"structure"` + + // Specifies the pending port for the DB instance. + Port *int64 `type:"integer"` + + // Specifies the storage type to be associated with the DB instance. + StorageType *string `type:"string"` +} + +// String returns the string representation +func (s PendingModifiedValues) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PendingModifiedValues) GoString() string { + return s.String() +} + +// SetAllocatedStorage sets the AllocatedStorage field's value. +func (s *PendingModifiedValues) SetAllocatedStorage(v int64) *PendingModifiedValues { + s.AllocatedStorage = &v + return s +} + +// SetBackupRetentionPeriod sets the BackupRetentionPeriod field's value. +func (s *PendingModifiedValues) SetBackupRetentionPeriod(v int64) *PendingModifiedValues { + s.BackupRetentionPeriod = &v + return s +} + +// SetCACertificateIdentifier sets the CACertificateIdentifier field's value. +func (s *PendingModifiedValues) SetCACertificateIdentifier(v string) *PendingModifiedValues { + s.CACertificateIdentifier = &v + return s +} + +// SetDBInstanceClass sets the DBInstanceClass field's value. +func (s *PendingModifiedValues) SetDBInstanceClass(v string) *PendingModifiedValues { + s.DBInstanceClass = &v + return s +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *PendingModifiedValues) SetDBInstanceIdentifier(v string) *PendingModifiedValues { + s.DBInstanceIdentifier = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *PendingModifiedValues) SetDBSubnetGroupName(v string) *PendingModifiedValues { + s.DBSubnetGroupName = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *PendingModifiedValues) SetEngineVersion(v string) *PendingModifiedValues { + s.EngineVersion = &v + return s +} + +// SetIops sets the Iops field's value. +func (s *PendingModifiedValues) SetIops(v int64) *PendingModifiedValues { + s.Iops = &v + return s +} + +// SetLicenseModel sets the LicenseModel field's value. +func (s *PendingModifiedValues) SetLicenseModel(v string) *PendingModifiedValues { + s.LicenseModel = &v + return s +} + +// SetMasterUserPassword sets the MasterUserPassword field's value. +func (s *PendingModifiedValues) SetMasterUserPassword(v string) *PendingModifiedValues { + s.MasterUserPassword = &v + return s +} + +// SetMultiAZ sets the MultiAZ field's value. +func (s *PendingModifiedValues) SetMultiAZ(v bool) *PendingModifiedValues { + s.MultiAZ = &v + return s +} + +// SetPendingCloudwatchLogsExports sets the PendingCloudwatchLogsExports field's value. +func (s *PendingModifiedValues) SetPendingCloudwatchLogsExports(v *PendingCloudwatchLogsExports) *PendingModifiedValues { + s.PendingCloudwatchLogsExports = v + return s +} + +// SetPort sets the Port field's value. +func (s *PendingModifiedValues) SetPort(v int64) *PendingModifiedValues { + s.Port = &v + return s +} + +// SetStorageType sets the StorageType field's value. +func (s *PendingModifiedValues) SetStorageType(v string) *PendingModifiedValues { + s.StorageType = &v + return s +} + +type PromoteReadReplicaDBClusterInput struct { + _ struct{} `type:"structure"` + + // The identifier of the DB cluster Read Replica to promote. This parameter + // is not case-sensitive. + // + // Constraints: + // + // * Must match the identifier of an existing DBCluster Read Replica. + // + // Example: my-cluster-replica1 + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s PromoteReadReplicaDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PromoteReadReplicaDBClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *PromoteReadReplicaDBClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "PromoteReadReplicaDBClusterInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *PromoteReadReplicaDBClusterInput) SetDBClusterIdentifier(v string) *PromoteReadReplicaDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +type PromoteReadReplicaDBClusterOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s PromoteReadReplicaDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PromoteReadReplicaDBClusterOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *PromoteReadReplicaDBClusterOutput) SetDBCluster(v *DBCluster) *PromoteReadReplicaDBClusterOutput { + s.DBCluster = v + return s +} + +// A range of integer values. +type Range struct { + _ struct{} `type:"structure"` + + // The minimum value in the range. + From *int64 `type:"integer"` + + // The step value for the range. For example, if you have a range of 5,000 to + // 10,000, with a step value of 1,000, the valid values start at 5,000 and step + // up by 1,000. Even though 7,500 is within the range, it isn't a valid value + // for the range. The valid values are 5,000, 6,000, 7,000, 8,000... + Step *int64 `type:"integer"` + + // The maximum value in the range. + To *int64 `type:"integer"` +} + +// String returns the string representation +func (s Range) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Range) GoString() string { + return s.String() +} + +// SetFrom sets the From field's value. +func (s *Range) SetFrom(v int64) *Range { + s.From = &v + return s +} + +// SetStep sets the Step field's value. +func (s *Range) SetStep(v int64) *Range { + s.Step = &v + return s +} + +// SetTo sets the To field's value. +func (s *Range) SetTo(v int64) *Range { + s.To = &v + return s +} + +type RebootDBInstanceInput struct { + _ struct{} `type:"structure"` + + // The DB instance identifier. This parameter is stored as a lowercase string. + // + // Constraints: + // + // * Must match the identifier of an existing DBInstance. + // + // DBInstanceIdentifier is a required field + DBInstanceIdentifier *string `type:"string" required:"true"` + + // When true, the reboot is conducted through a MultiAZ failover. + // + // Constraint: You can't specify true if the instance is not configured for + // MultiAZ. + ForceFailover *bool `type:"boolean"` +} + +// String returns the string representation +func (s RebootDBInstanceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RebootDBInstanceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RebootDBInstanceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RebootDBInstanceInput"} + if s.DBInstanceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBInstanceIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBInstanceIdentifier sets the DBInstanceIdentifier field's value. +func (s *RebootDBInstanceInput) SetDBInstanceIdentifier(v string) *RebootDBInstanceInput { + s.DBInstanceIdentifier = &v + return s +} + +// SetForceFailover sets the ForceFailover field's value. +func (s *RebootDBInstanceInput) SetForceFailover(v bool) *RebootDBInstanceInput { + s.ForceFailover = &v + return s +} + +type RebootDBInstanceOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB instance. + // + // This data type is used as a response element in the DescribeDBInstances action. + DBInstance *DBInstance `type:"structure"` +} + +// String returns the string representation +func (s RebootDBInstanceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RebootDBInstanceOutput) GoString() string { + return s.String() +} + +// SetDBInstance sets the DBInstance field's value. +func (s *RebootDBInstanceOutput) SetDBInstance(v *DBInstance) *RebootDBInstanceOutput { + s.DBInstance = v + return s +} + +type RemoveRoleFromDBClusterInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster to disassociate the IAM role from. + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The Amazon Resource Name (ARN) of the IAM role to disassociate from the DB + // cluster, for example arn:aws:iam::123456789012:role/NeptuneAccessRole. + // + // RoleArn is a required field + RoleArn *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s RemoveRoleFromDBClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RemoveRoleFromDBClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RemoveRoleFromDBClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RemoveRoleFromDBClusterInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + if s.RoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("RoleArn")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *RemoveRoleFromDBClusterInput) SetDBClusterIdentifier(v string) *RemoveRoleFromDBClusterInput { + s.DBClusterIdentifier = &v + return s +} + +// SetRoleArn sets the RoleArn field's value. +func (s *RemoveRoleFromDBClusterInput) SetRoleArn(v string) *RemoveRoleFromDBClusterInput { + s.RoleArn = &v + return s +} + +type RemoveRoleFromDBClusterOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s RemoveRoleFromDBClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RemoveRoleFromDBClusterOutput) GoString() string { + return s.String() +} + +type RemoveSourceIdentifierFromSubscriptionInput struct { + _ struct{} `type:"structure"` + + // The source identifier to be removed from the subscription, such as the DB + // instance identifier for a DB instance or the name of a security group. + // + // SourceIdentifier is a required field + SourceIdentifier *string `type:"string" required:"true"` + + // The name of the event notification subscription you want to remove a source + // identifier from. + // + // SubscriptionName is a required field + SubscriptionName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s RemoveSourceIdentifierFromSubscriptionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RemoveSourceIdentifierFromSubscriptionInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RemoveSourceIdentifierFromSubscriptionInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RemoveSourceIdentifierFromSubscriptionInput"} + if s.SourceIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SourceIdentifier")) + } + if s.SubscriptionName == nil { + invalidParams.Add(request.NewErrParamRequired("SubscriptionName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSourceIdentifier sets the SourceIdentifier field's value. +func (s *RemoveSourceIdentifierFromSubscriptionInput) SetSourceIdentifier(v string) *RemoveSourceIdentifierFromSubscriptionInput { + s.SourceIdentifier = &v + return s +} + +// SetSubscriptionName sets the SubscriptionName field's value. +func (s *RemoveSourceIdentifierFromSubscriptionInput) SetSubscriptionName(v string) *RemoveSourceIdentifierFromSubscriptionInput { + s.SubscriptionName = &v + return s +} + +type RemoveSourceIdentifierFromSubscriptionOutput struct { + _ struct{} `type:"structure"` + + // Contains the results of a successful invocation of the DescribeEventSubscriptions + // action. + EventSubscription *EventSubscription `type:"structure"` +} + +// String returns the string representation +func (s RemoveSourceIdentifierFromSubscriptionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RemoveSourceIdentifierFromSubscriptionOutput) GoString() string { + return s.String() +} + +// SetEventSubscription sets the EventSubscription field's value. +func (s *RemoveSourceIdentifierFromSubscriptionOutput) SetEventSubscription(v *EventSubscription) *RemoveSourceIdentifierFromSubscriptionOutput { + s.EventSubscription = v + return s +} + +type RemoveTagsFromResourceInput struct { + _ struct{} `type:"structure"` + + // The Amazon Neptune resource that the tags are removed from. This value is + // an Amazon Resource Name (ARN). For information about creating an ARN, see + // Constructing an Amazon Resource Name (ARN) (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html#tagging.ARN.Constructing). + // + // ResourceName is a required field + ResourceName *string `type:"string" required:"true"` + + // The tag key (name) of the tag to be removed. + // + // TagKeys is a required field + TagKeys []*string `type:"list" required:"true"` +} + +// String returns the string representation +func (s RemoveTagsFromResourceInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RemoveTagsFromResourceInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RemoveTagsFromResourceInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RemoveTagsFromResourceInput"} + if s.ResourceName == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceName")) + } + if s.TagKeys == nil { + invalidParams.Add(request.NewErrParamRequired("TagKeys")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetResourceName sets the ResourceName field's value. +func (s *RemoveTagsFromResourceInput) SetResourceName(v string) *RemoveTagsFromResourceInput { + s.ResourceName = &v + return s +} + +// SetTagKeys sets the TagKeys field's value. +func (s *RemoveTagsFromResourceInput) SetTagKeys(v []*string) *RemoveTagsFromResourceInput { + s.TagKeys = v + return s +} + +type RemoveTagsFromResourceOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s RemoveTagsFromResourceOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RemoveTagsFromResourceOutput) GoString() string { + return s.String() +} + +type ResetDBClusterParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster parameter group to reset. + // + // DBClusterParameterGroupName is a required field + DBClusterParameterGroupName *string `type:"string" required:"true"` + + // A list of parameter names in the DB cluster parameter group to reset to the + // default values. You can't use this parameter if the ResetAllParameters parameter + // is set to true. + Parameters []*Parameter `locationNameList:"Parameter" type:"list"` + + // A value that is set to true to reset all parameters in the DB cluster parameter + // group to their default values, and false otherwise. You can't use this parameter + // if there is a list of parameter names specified for the Parameters parameter. + ResetAllParameters *bool `type:"boolean"` +} + +// String returns the string representation +func (s ResetDBClusterParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResetDBClusterParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ResetDBClusterParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ResetDBClusterParameterGroupInput"} + if s.DBClusterParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterParameterGroupName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *ResetDBClusterParameterGroupInput) SetDBClusterParameterGroupName(v string) *ResetDBClusterParameterGroupInput { + s.DBClusterParameterGroupName = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *ResetDBClusterParameterGroupInput) SetParameters(v []*Parameter) *ResetDBClusterParameterGroupInput { + s.Parameters = v + return s +} + +// SetResetAllParameters sets the ResetAllParameters field's value. +func (s *ResetDBClusterParameterGroupInput) SetResetAllParameters(v bool) *ResetDBClusterParameterGroupInput { + s.ResetAllParameters = &v + return s +} + +type ResetDBClusterParameterGroupOutput struct { + _ struct{} `type:"structure"` + + // The name of the DB cluster parameter group. + // + // Constraints: + // + // * Must be 1 to 255 letters or numbers. + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // This value is stored as a lowercase string. + DBClusterParameterGroupName *string `type:"string"` +} + +// String returns the string representation +func (s ResetDBClusterParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResetDBClusterParameterGroupOutput) GoString() string { + return s.String() +} + +// SetDBClusterParameterGroupName sets the DBClusterParameterGroupName field's value. +func (s *ResetDBClusterParameterGroupOutput) SetDBClusterParameterGroupName(v string) *ResetDBClusterParameterGroupOutput { + s.DBClusterParameterGroupName = &v + return s +} + +type ResetDBParameterGroupInput struct { + _ struct{} `type:"structure"` + + // The name of the DB parameter group. + // + // Constraints: + // + // * Must match the name of an existing DBParameterGroup. + // + // DBParameterGroupName is a required field + DBParameterGroupName *string `type:"string" required:"true"` + + // To reset the entire DB parameter group, specify the DBParameterGroup name + // and ResetAllParameters parameters. To reset specific parameters, provide + // a list of the following: ParameterName and ApplyMethod. A maximum of 20 parameters + // can be modified in a single request. + // + // Valid Values (for Apply method): pending-reboot + Parameters []*Parameter `locationNameList:"Parameter" type:"list"` + + // Specifies whether (true) or not (false) to reset all parameters in the DB + // parameter group to default values. + // + // Default: true + ResetAllParameters *bool `type:"boolean"` +} + +// String returns the string representation +func (s ResetDBParameterGroupInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResetDBParameterGroupInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ResetDBParameterGroupInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ResetDBParameterGroupInput"} + if s.DBParameterGroupName == nil { + invalidParams.Add(request.NewErrParamRequired("DBParameterGroupName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *ResetDBParameterGroupInput) SetDBParameterGroupName(v string) *ResetDBParameterGroupInput { + s.DBParameterGroupName = &v + return s +} + +// SetParameters sets the Parameters field's value. +func (s *ResetDBParameterGroupInput) SetParameters(v []*Parameter) *ResetDBParameterGroupInput { + s.Parameters = v + return s +} + +// SetResetAllParameters sets the ResetAllParameters field's value. +func (s *ResetDBParameterGroupInput) SetResetAllParameters(v bool) *ResetDBParameterGroupInput { + s.ResetAllParameters = &v + return s +} + +// Contains the result of a successful invocation of the ModifyDBParameterGroup +// or ResetDBParameterGroup action. +type ResetDBParameterGroupOutput struct { + _ struct{} `type:"structure"` + + // Provides the name of the DB parameter group. + DBParameterGroupName *string `type:"string"` +} + +// String returns the string representation +func (s ResetDBParameterGroupOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResetDBParameterGroupOutput) GoString() string { + return s.String() +} + +// SetDBParameterGroupName sets the DBParameterGroupName field's value. +func (s *ResetDBParameterGroupOutput) SetDBParameterGroupName(v string) *ResetDBParameterGroupOutput { + s.DBParameterGroupName = &v + return s +} + +// Describes the pending maintenance actions for a resource. +type ResourcePendingMaintenanceActions struct { + _ struct{} `type:"structure"` + + // A list that provides details about the pending maintenance actions for the + // resource. + PendingMaintenanceActionDetails []*PendingMaintenanceAction `locationNameList:"PendingMaintenanceAction" type:"list"` + + // The ARN of the resource that has pending maintenance actions. + ResourceIdentifier *string `type:"string"` +} + +// String returns the string representation +func (s ResourcePendingMaintenanceActions) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResourcePendingMaintenanceActions) GoString() string { + return s.String() +} + +// SetPendingMaintenanceActionDetails sets the PendingMaintenanceActionDetails field's value. +func (s *ResourcePendingMaintenanceActions) SetPendingMaintenanceActionDetails(v []*PendingMaintenanceAction) *ResourcePendingMaintenanceActions { + s.PendingMaintenanceActionDetails = v + return s +} + +// SetResourceIdentifier sets the ResourceIdentifier field's value. +func (s *ResourcePendingMaintenanceActions) SetResourceIdentifier(v string) *ResourcePendingMaintenanceActions { + s.ResourceIdentifier = &v + return s +} + +type RestoreDBClusterFromSnapshotInput struct { + _ struct{} `type:"structure"` + + // Provides the list of EC2 Availability Zones that instances in the restored + // DB cluster can be created in. + AvailabilityZones []*string `locationNameList:"AvailabilityZone" type:"list"` + + // The name of the DB cluster to create from the DB snapshot or DB cluster snapshot. + // This parameter isn't case-sensitive. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // Example: my-snapshot-id + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The name of the DB subnet group to use for the new DB cluster. + // + // Constraints: If supplied, must match the name of an existing DBSubnetGroup. + // + // Example: mySubnetgroup + DBSubnetGroupName *string `type:"string"` + + // The database name for the restored DB cluster. + DatabaseName *string `type:"string"` + + // True to enable mapping of AWS Identity and Access Management (IAM) accounts + // to database accounts, and otherwise false. + // + // Default: false + EnableIAMDatabaseAuthentication *bool `type:"boolean"` + + // The database engine to use for the new DB cluster. + // + // Default: The same as source + // + // Constraint: Must be compatible with the engine of the source + // + // Engine is a required field + Engine *string `type:"string" required:"true"` + + // The version of the database engine to use for the new DB cluster. + EngineVersion *string `type:"string"` + + // The AWS KMS key identifier to use when restoring an encrypted DB cluster + // from a DB snapshot or DB cluster snapshot. + // + // The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption + // key. If you are restoring a DB cluster with the same AWS account that owns + // the KMS encryption key used to encrypt the new DB cluster, then you can use + // the KMS key alias instead of the ARN for the KMS encryption key. + // + // If you do not specify a value for the KmsKeyId parameter, then the following + // will occur: + // + // * If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is encrypted, + // then the restored DB cluster is encrypted using the KMS key that was used + // to encrypt the DB snapshot or DB cluster snapshot. + // + // * If the DB snapshot or DB cluster snapshot in SnapshotIdentifier is not + // encrypted, then the restored DB cluster is not encrypted. + KmsKeyId *string `type:"string"` + + // The name of the option group to use for the restored DB cluster. + OptionGroupName *string `type:"string"` + + // The port number on which the new DB cluster accepts connections. + // + // Constraints: Value must be 1150-65535 + // + // Default: The same port as the original DB cluster. + Port *int64 `type:"integer"` + + // The identifier for the DB snapshot or DB cluster snapshot to restore from. + // + // You can use either the name or the Amazon Resource Name (ARN) to specify + // a DB cluster snapshot. However, you can use only the ARN to specify a DB + // snapshot. + // + // Constraints: + // + // * Must match the identifier of an existing Snapshot. + // + // SnapshotIdentifier is a required field + SnapshotIdentifier *string `type:"string" required:"true"` + + // The tags to be assigned to the restored DB cluster. + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // A list of VPC security groups that the new DB cluster will belong to. + VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` +} + +// String returns the string representation +func (s RestoreDBClusterFromSnapshotInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RestoreDBClusterFromSnapshotInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RestoreDBClusterFromSnapshotInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RestoreDBClusterFromSnapshotInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + if s.Engine == nil { + invalidParams.Add(request.NewErrParamRequired("Engine")) + } + if s.SnapshotIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SnapshotIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAvailabilityZones sets the AvailabilityZones field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetAvailabilityZones(v []*string) *RestoreDBClusterFromSnapshotInput { + s.AvailabilityZones = v + return s +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetDBClusterIdentifier(v string) *RestoreDBClusterFromSnapshotInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetDBSubnetGroupName(v string) *RestoreDBClusterFromSnapshotInput { + s.DBSubnetGroupName = &v + return s +} + +// SetDatabaseName sets the DatabaseName field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetDatabaseName(v string) *RestoreDBClusterFromSnapshotInput { + s.DatabaseName = &v + return s +} + +// SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetEnableIAMDatabaseAuthentication(v bool) *RestoreDBClusterFromSnapshotInput { + s.EnableIAMDatabaseAuthentication = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetEngine(v string) *RestoreDBClusterFromSnapshotInput { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetEngineVersion(v string) *RestoreDBClusterFromSnapshotInput { + s.EngineVersion = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetKmsKeyId(v string) *RestoreDBClusterFromSnapshotInput { + s.KmsKeyId = &v + return s +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetOptionGroupName(v string) *RestoreDBClusterFromSnapshotInput { + s.OptionGroupName = &v + return s +} + +// SetPort sets the Port field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetPort(v int64) *RestoreDBClusterFromSnapshotInput { + s.Port = &v + return s +} + +// SetSnapshotIdentifier sets the SnapshotIdentifier field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetSnapshotIdentifier(v string) *RestoreDBClusterFromSnapshotInput { + s.SnapshotIdentifier = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetTags(v []*Tag) *RestoreDBClusterFromSnapshotInput { + s.Tags = v + return s +} + +// SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetVpcSecurityGroupIds(v []*string) *RestoreDBClusterFromSnapshotInput { + s.VpcSecurityGroupIds = v + return s +} + +type RestoreDBClusterFromSnapshotOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s RestoreDBClusterFromSnapshotOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RestoreDBClusterFromSnapshotOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *RestoreDBClusterFromSnapshotOutput) SetDBCluster(v *DBCluster) *RestoreDBClusterFromSnapshotOutput { + s.DBCluster = v + return s +} + +type RestoreDBClusterToPointInTimeInput struct { + _ struct{} `type:"structure"` + + // The name of the new DB cluster to be created. + // + // Constraints: + // + // * Must contain from 1 to 63 letters, numbers, or hyphens + // + // * First character must be a letter + // + // * Cannot end with a hyphen or contain two consecutive hyphens + // + // DBClusterIdentifier is a required field + DBClusterIdentifier *string `type:"string" required:"true"` + + // The DB subnet group name to use for the new DB cluster. + // + // Constraints: If supplied, must match the name of an existing DBSubnetGroup. + // + // Example: mySubnetgroup + DBSubnetGroupName *string `type:"string"` + + // True to enable mapping of AWS Identity and Access Management (IAM) accounts + // to database accounts, and otherwise false. + // + // Default: false + EnableIAMDatabaseAuthentication *bool `type:"boolean"` + + // The AWS KMS key identifier to use when restoring an encrypted DB cluster + // from an encrypted DB cluster. + // + // The KMS key identifier is the Amazon Resource Name (ARN) for the KMS encryption + // key. If you are restoring a DB cluster with the same AWS account that owns + // the KMS encryption key used to encrypt the new DB cluster, then you can use + // the KMS key alias instead of the ARN for the KMS encryption key. + // + // You can restore to a new DB cluster and encrypt the new DB cluster with a + // KMS key that is different than the KMS key used to encrypt the source DB + // cluster. The new DB cluster is encrypted with the KMS key identified by the + // KmsKeyId parameter. + // + // If you do not specify a value for the KmsKeyId parameter, then the following + // will occur: + // + // * If the DB cluster is encrypted, then the restored DB cluster is encrypted + // using the KMS key that was used to encrypt the source DB cluster. + // + // * If the DB cluster is not encrypted, then the restored DB cluster is + // not encrypted. + // + // If DBClusterIdentifier refers to a DB cluster that is not encrypted, then + // the restore request is rejected. + KmsKeyId *string `type:"string"` + + // The name of the option group for the new DB cluster. + OptionGroupName *string `type:"string"` + + // The port number on which the new DB cluster accepts connections. + // + // Constraints: Value must be 1150-65535 + // + // Default: The same port as the original DB cluster. + Port *int64 `type:"integer"` + + // The date and time to restore the DB cluster to. + // + // Valid Values: Value must be a time in Universal Coordinated Time (UTC) format + // + // Constraints: + // + // * Must be before the latest restorable time for the DB instance + // + // * Must be specified if UseLatestRestorableTime parameter is not provided + // + // * Cannot be specified if UseLatestRestorableTime parameter is true + // + // * Cannot be specified if RestoreType parameter is copy-on-write + // + // Example: 2015-03-07T23:45:00Z + RestoreToTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + + // The type of restore to be performed. You can specify one of the following + // values: + // + // * full-copy - The new DB cluster is restored as a full copy of the source + // DB cluster. + // + // * copy-on-write - The new DB cluster is restored as a clone of the source + // DB cluster. + // + // Constraints: You can't specify copy-on-write if the engine version of the + // source DB cluster is earlier than 1.11. + // + // If you don't specify a RestoreType value, then the new DB cluster is restored + // as a full copy of the source DB cluster. + RestoreType *string `type:"string"` + + // The identifier of the source DB cluster from which to restore. + // + // Constraints: + // + // * Must match the identifier of an existing DBCluster. + // + // SourceDBClusterIdentifier is a required field + SourceDBClusterIdentifier *string `type:"string" required:"true"` + + // A list of tags. For more information, see Tagging Amazon Neptune Resources + // (http://docs.aws.amazon.com/neptune/latest/UserGuide/tagging.ARN.html). + Tags []*Tag `locationNameList:"Tag" type:"list"` + + // A value that is set to true to restore the DB cluster to the latest restorable + // backup time, and false otherwise. + // + // Default: false + // + // Constraints: Cannot be specified if RestoreToTime parameter is provided. + UseLatestRestorableTime *bool `type:"boolean"` + + // A list of VPC security groups that the new DB cluster belongs to. + VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` +} + +// String returns the string representation +func (s RestoreDBClusterToPointInTimeInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RestoreDBClusterToPointInTimeInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *RestoreDBClusterToPointInTimeInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "RestoreDBClusterToPointInTimeInput"} + if s.DBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("DBClusterIdentifier")) + } + if s.SourceDBClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("SourceDBClusterIdentifier")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDBClusterIdentifier sets the DBClusterIdentifier field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetDBClusterIdentifier(v string) *RestoreDBClusterToPointInTimeInput { + s.DBClusterIdentifier = &v + return s +} + +// SetDBSubnetGroupName sets the DBSubnetGroupName field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetDBSubnetGroupName(v string) *RestoreDBClusterToPointInTimeInput { + s.DBSubnetGroupName = &v + return s +} + +// SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetEnableIAMDatabaseAuthentication(v bool) *RestoreDBClusterToPointInTimeInput { + s.EnableIAMDatabaseAuthentication = &v + return s +} + +// SetKmsKeyId sets the KmsKeyId field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetKmsKeyId(v string) *RestoreDBClusterToPointInTimeInput { + s.KmsKeyId = &v + return s +} + +// SetOptionGroupName sets the OptionGroupName field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetOptionGroupName(v string) *RestoreDBClusterToPointInTimeInput { + s.OptionGroupName = &v + return s +} + +// SetPort sets the Port field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetPort(v int64) *RestoreDBClusterToPointInTimeInput { + s.Port = &v + return s +} + +// SetRestoreToTime sets the RestoreToTime field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetRestoreToTime(v time.Time) *RestoreDBClusterToPointInTimeInput { + s.RestoreToTime = &v + return s +} + +// SetRestoreType sets the RestoreType field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetRestoreType(v string) *RestoreDBClusterToPointInTimeInput { + s.RestoreType = &v + return s +} + +// SetSourceDBClusterIdentifier sets the SourceDBClusterIdentifier field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetSourceDBClusterIdentifier(v string) *RestoreDBClusterToPointInTimeInput { + s.SourceDBClusterIdentifier = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetTags(v []*Tag) *RestoreDBClusterToPointInTimeInput { + s.Tags = v + return s +} + +// SetUseLatestRestorableTime sets the UseLatestRestorableTime field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetUseLatestRestorableTime(v bool) *RestoreDBClusterToPointInTimeInput { + s.UseLatestRestorableTime = &v + return s +} + +// SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetVpcSecurityGroupIds(v []*string) *RestoreDBClusterToPointInTimeInput { + s.VpcSecurityGroupIds = v + return s +} + +type RestoreDBClusterToPointInTimeOutput struct { + _ struct{} `type:"structure"` + + // Contains the details of an Amazon Neptune DB cluster. + // + // This data type is used as a response element in the DescribeDBClusters action. + DBCluster *DBCluster `type:"structure"` +} + +// String returns the string representation +func (s RestoreDBClusterToPointInTimeOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RestoreDBClusterToPointInTimeOutput) GoString() string { + return s.String() +} + +// SetDBCluster sets the DBCluster field's value. +func (s *RestoreDBClusterToPointInTimeOutput) SetDBCluster(v *DBCluster) *RestoreDBClusterToPointInTimeOutput { + s.DBCluster = v + return s +} + +// This data type is used as a response element in the DescribeDBSubnetGroups +// action. +type Subnet struct { + _ struct{} `type:"structure"` + + // Contains Availability Zone information. + // + // This data type is used as an element in the following data type: + // + // * OrderableDBInstanceOption + SubnetAvailabilityZone *AvailabilityZone `type:"structure"` + + // Specifies the identifier of the subnet. + SubnetIdentifier *string `type:"string"` + + // Specifies the status of the subnet. + SubnetStatus *string `type:"string"` +} + +// String returns the string representation +func (s Subnet) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Subnet) GoString() string { + return s.String() +} + +// SetSubnetAvailabilityZone sets the SubnetAvailabilityZone field's value. +func (s *Subnet) SetSubnetAvailabilityZone(v *AvailabilityZone) *Subnet { + s.SubnetAvailabilityZone = v + return s +} + +// SetSubnetIdentifier sets the SubnetIdentifier field's value. +func (s *Subnet) SetSubnetIdentifier(v string) *Subnet { + s.SubnetIdentifier = &v + return s +} + +// SetSubnetStatus sets the SubnetStatus field's value. +func (s *Subnet) SetSubnetStatus(v string) *Subnet { + s.SubnetStatus = &v + return s +} + +// Metadata assigned to an Amazon Neptune resource consisting of a key-value +// pair. +type Tag struct { + _ struct{} `type:"structure"` + + // A key is the required name of the tag. The string value can be from 1 to + // 128 Unicode characters in length and can't be prefixed with "aws:" or "rds:". + // The string can only contain only the set of Unicode letters, digits, white-space, + // '_', '.', '/', '=', '+', '-' (Java regex: "^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-]*)$"). + Key *string `type:"string"` + + // A value is the optional value of the tag. The string value can be from 1 + // to 256 Unicode characters in length and can't be prefixed with "aws:" or + // "rds:". The string can only contain only the set of Unicode letters, digits, + // white-space, '_', '.', '/', '=', '+', '-' (Java regex: "^([\\p{L}\\p{Z}\\p{N}_.:/=+\\-]*)$"). + Value *string `type:"string"` +} + +// String returns the string representation +func (s Tag) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Tag) GoString() string { + return s.String() +} + +// SetKey sets the Key field's value. +func (s *Tag) SetKey(v string) *Tag { + s.Key = &v + return s +} + +// SetValue sets the Value field's value. +func (s *Tag) SetValue(v string) *Tag { + s.Value = &v + return s +} + +// A time zone associated with a DBInstance. This data type is an element in +// the response to the DescribeDBInstances, and the DescribeDBEngineVersions +// actions. +type Timezone struct { + _ struct{} `type:"structure"` + + // The name of the time zone. + TimezoneName *string `type:"string"` +} + +// String returns the string representation +func (s Timezone) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Timezone) GoString() string { + return s.String() +} + +// SetTimezoneName sets the TimezoneName field's value. +func (s *Timezone) SetTimezoneName(v string) *Timezone { + s.TimezoneName = &v + return s +} + +// The version of the database engine that a DB instance can be upgraded to. +type UpgradeTarget struct { + _ struct{} `type:"structure"` + + // A value that indicates whether the target version is applied to any source + // DB instances that have AutoMinorVersionUpgrade set to true. + AutoUpgrade *bool `type:"boolean"` + + // The version of the database engine that a DB instance can be upgraded to. + Description *string `type:"string"` + + // The name of the upgrade target database engine. + Engine *string `type:"string"` + + // The version number of the upgrade target database engine. + EngineVersion *string `type:"string"` + + // A value that indicates whether a database engine is upgraded to a major version. + IsMajorVersionUpgrade *bool `type:"boolean"` +} + +// String returns the string representation +func (s UpgradeTarget) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpgradeTarget) GoString() string { + return s.String() +} + +// SetAutoUpgrade sets the AutoUpgrade field's value. +func (s *UpgradeTarget) SetAutoUpgrade(v bool) *UpgradeTarget { + s.AutoUpgrade = &v + return s +} + +// SetDescription sets the Description field's value. +func (s *UpgradeTarget) SetDescription(v string) *UpgradeTarget { + s.Description = &v + return s +} + +// SetEngine sets the Engine field's value. +func (s *UpgradeTarget) SetEngine(v string) *UpgradeTarget { + s.Engine = &v + return s +} + +// SetEngineVersion sets the EngineVersion field's value. +func (s *UpgradeTarget) SetEngineVersion(v string) *UpgradeTarget { + s.EngineVersion = &v + return s +} + +// SetIsMajorVersionUpgrade sets the IsMajorVersionUpgrade field's value. +func (s *UpgradeTarget) SetIsMajorVersionUpgrade(v bool) *UpgradeTarget { + s.IsMajorVersionUpgrade = &v + return s +} + +// Information about valid modifications that you can make to your DB instance. +// Contains the result of a successful call to the DescribeValidDBInstanceModifications +// action. You can use this information when you call ModifyDBInstance. +type ValidDBInstanceModificationsMessage struct { + _ struct{} `type:"structure"` + + // Valid storage options for your DB instance. + Storage []*ValidStorageOptions `locationNameList:"ValidStorageOptions" type:"list"` +} + +// String returns the string representation +func (s ValidDBInstanceModificationsMessage) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ValidDBInstanceModificationsMessage) GoString() string { + return s.String() +} + +// SetStorage sets the Storage field's value. +func (s *ValidDBInstanceModificationsMessage) SetStorage(v []*ValidStorageOptions) *ValidDBInstanceModificationsMessage { + s.Storage = v + return s +} + +// Information about valid modifications that you can make to your DB instance. +// Contains the result of a successful call to the DescribeValidDBInstanceModifications +// action. +type ValidStorageOptions struct { + _ struct{} `type:"structure"` + + // The valid range of Provisioned IOPS to gibibytes of storage multiplier. For + // example, 3-10, which means that provisioned IOPS can be between 3 and 10 + // times storage. + IopsToStorageRatio []*DoubleRange `locationNameList:"DoubleRange" type:"list"` + + // The valid range of provisioned IOPS. For example, 1000-20000. + ProvisionedIops []*Range `locationNameList:"Range" type:"list"` + + // The valid range of storage in gibibytes. For example, 100 to 16384. + StorageSize []*Range `locationNameList:"Range" type:"list"` + + // The valid storage types for your DB instance. For example, gp2, io1. + StorageType *string `type:"string"` +} + +// String returns the string representation +func (s ValidStorageOptions) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ValidStorageOptions) GoString() string { + return s.String() +} + +// SetIopsToStorageRatio sets the IopsToStorageRatio field's value. +func (s *ValidStorageOptions) SetIopsToStorageRatio(v []*DoubleRange) *ValidStorageOptions { + s.IopsToStorageRatio = v + return s +} + +// SetProvisionedIops sets the ProvisionedIops field's value. +func (s *ValidStorageOptions) SetProvisionedIops(v []*Range) *ValidStorageOptions { + s.ProvisionedIops = v + return s +} + +// SetStorageSize sets the StorageSize field's value. +func (s *ValidStorageOptions) SetStorageSize(v []*Range) *ValidStorageOptions { + s.StorageSize = v + return s +} + +// SetStorageType sets the StorageType field's value. +func (s *ValidStorageOptions) SetStorageType(v string) *ValidStorageOptions { + s.StorageType = &v + return s +} + +// This data type is used as a response element for queries on VPC security +// group membership. +type VpcSecurityGroupMembership struct { + _ struct{} `type:"structure"` + + // The status of the VPC security group. + Status *string `type:"string"` + + // The name of the VPC security group. + VpcSecurityGroupId *string `type:"string"` +} + +// String returns the string representation +func (s VpcSecurityGroupMembership) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s VpcSecurityGroupMembership) GoString() string { + return s.String() +} + +// SetStatus sets the Status field's value. +func (s *VpcSecurityGroupMembership) SetStatus(v string) *VpcSecurityGroupMembership { + s.Status = &v + return s +} + +// SetVpcSecurityGroupId sets the VpcSecurityGroupId field's value. +func (s *VpcSecurityGroupMembership) SetVpcSecurityGroupId(v string) *VpcSecurityGroupMembership { + s.VpcSecurityGroupId = &v + return s +} + +const ( + // ApplyMethodImmediate is a ApplyMethod enum value + ApplyMethodImmediate = "immediate" + + // ApplyMethodPendingReboot is a ApplyMethod enum value + ApplyMethodPendingReboot = "pending-reboot" +) + +const ( + // SourceTypeDbInstance is a SourceType enum value + SourceTypeDbInstance = "db-instance" + + // SourceTypeDbParameterGroup is a SourceType enum value + SourceTypeDbParameterGroup = "db-parameter-group" + + // SourceTypeDbSecurityGroup is a SourceType enum value + SourceTypeDbSecurityGroup = "db-security-group" + + // SourceTypeDbSnapshot is a SourceType enum value + SourceTypeDbSnapshot = "db-snapshot" + + // SourceTypeDbCluster is a SourceType enum value + SourceTypeDbCluster = "db-cluster" + + // SourceTypeDbClusterSnapshot is a SourceType enum value + SourceTypeDbClusterSnapshot = "db-cluster-snapshot" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/neptune/doc.go b/vendor/github.com/aws/aws-sdk-go/service/neptune/doc.go new file mode 100644 index 000000000..503af8466 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/neptune/doc.go @@ -0,0 +1,48 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package neptune provides the client and types for making API +// requests to Amazon Neptune. +// +// Amazon Neptune is a fast, reliable, fully-managed graph database service +// that makes it easy to build and run applications that work with highly connected +// datasets. The core of Amazon Neptune is a purpose-built, high-performance +// graph database engine optimized for storing billions of relationships and +// querying the graph with milliseconds latency. Amazon Neptune supports popular +// graph models Property Graph and W3C's RDF, and their respective query languages +// Apache TinkerPop Gremlin and SPARQL, allowing you to easily build queries +// that efficiently navigate highly connected datasets. Neptune powers graph +// use cases such as recommendation engines, fraud detection, knowledge graphs, +// drug discovery, and network security. +// +// This interface reference for Amazon Neptune contains documentation for a +// programming or command line interface you can use to manage Amazon Neptune. +// Note that Amazon Neptune is asynchronous, which means that some interfaces +// might require techniques such as polling or callback functions to determine +// when a command has been applied. In this reference, the parameter descriptions +// indicate whether a command is applied immediately, on the next instance reboot, +// or during the maintenance window. The reference structure is as follows, +// and we list following some related topics from the user guide. +// +// Amazon Neptune API Reference +// +// See https://docs.aws.amazon.com/goto/WebAPI/neptune-2014-10-31 for more information on this service. +// +// See neptune package documentation for more information. +// https://docs.aws.amazon.com/sdk-for-go/api/service/neptune/ +// +// Using the Client +// +// To contact Amazon Neptune with the SDK use the New function to create +// a new service client. With that client you can make API requests to the service. +// These clients are safe to use concurrently. +// +// See the SDK's documentation for more information on how to use the SDK. +// https://docs.aws.amazon.com/sdk-for-go/api/ +// +// See aws.Config documentation for more information on configuring SDK clients. +// https://docs.aws.amazon.com/sdk-for-go/api/aws/#Config +// +// See the Amazon Neptune client Neptune for more +// information on creating client for this service. +// https://docs.aws.amazon.com/sdk-for-go/api/service/neptune/#New +package neptune diff --git a/vendor/github.com/aws/aws-sdk-go/service/neptune/errors.go b/vendor/github.com/aws/aws-sdk-go/service/neptune/errors.go new file mode 100644 index 000000000..8c0c3f967 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/neptune/errors.go @@ -0,0 +1,361 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package neptune + +const ( + + // ErrCodeAuthorizationNotFoundFault for service response error code + // "AuthorizationNotFound". + // + // Specified CIDRIP or EC2 security group is not authorized for the specified + // DB security group. + // + // Neptune may not also be authorized via IAM to perform necessary actions on + // your behalf. + ErrCodeAuthorizationNotFoundFault = "AuthorizationNotFound" + + // ErrCodeCertificateNotFoundFault for service response error code + // "CertificateNotFound". + // + // CertificateIdentifier does not refer to an existing certificate. + ErrCodeCertificateNotFoundFault = "CertificateNotFound" + + // ErrCodeDBClusterAlreadyExistsFault for service response error code + // "DBClusterAlreadyExistsFault". + // + // User already has a DB cluster with the given identifier. + ErrCodeDBClusterAlreadyExistsFault = "DBClusterAlreadyExistsFault" + + // ErrCodeDBClusterNotFoundFault for service response error code + // "DBClusterNotFoundFault". + // + // DBClusterIdentifier does not refer to an existing DB cluster. + ErrCodeDBClusterNotFoundFault = "DBClusterNotFoundFault" + + // ErrCodeDBClusterParameterGroupNotFoundFault for service response error code + // "DBClusterParameterGroupNotFound". + // + // DBClusterParameterGroupName does not refer to an existing DB Cluster parameter + // group. + ErrCodeDBClusterParameterGroupNotFoundFault = "DBClusterParameterGroupNotFound" + + // ErrCodeDBClusterQuotaExceededFault for service response error code + // "DBClusterQuotaExceededFault". + // + // User attempted to create a new DB cluster and the user has already reached + // the maximum allowed DB cluster quota. + ErrCodeDBClusterQuotaExceededFault = "DBClusterQuotaExceededFault" + + // ErrCodeDBClusterRoleAlreadyExistsFault for service response error code + // "DBClusterRoleAlreadyExists". + // + // The specified IAM role Amazon Resource Name (ARN) is already associated with + // the specified DB cluster. + ErrCodeDBClusterRoleAlreadyExistsFault = "DBClusterRoleAlreadyExists" + + // ErrCodeDBClusterRoleNotFoundFault for service response error code + // "DBClusterRoleNotFound". + // + // The specified IAM role Amazon Resource Name (ARN) is not associated with + // the specified DB cluster. + ErrCodeDBClusterRoleNotFoundFault = "DBClusterRoleNotFound" + + // ErrCodeDBClusterRoleQuotaExceededFault for service response error code + // "DBClusterRoleQuotaExceeded". + // + // You have exceeded the maximum number of IAM roles that can be associated + // with the specified DB cluster. + ErrCodeDBClusterRoleQuotaExceededFault = "DBClusterRoleQuotaExceeded" + + // ErrCodeDBClusterSnapshotAlreadyExistsFault for service response error code + // "DBClusterSnapshotAlreadyExistsFault". + // + // User already has a DB cluster snapshot with the given identifier. + ErrCodeDBClusterSnapshotAlreadyExistsFault = "DBClusterSnapshotAlreadyExistsFault" + + // ErrCodeDBClusterSnapshotNotFoundFault for service response error code + // "DBClusterSnapshotNotFoundFault". + // + // DBClusterSnapshotIdentifier does not refer to an existing DB cluster snapshot. + ErrCodeDBClusterSnapshotNotFoundFault = "DBClusterSnapshotNotFoundFault" + + // ErrCodeDBInstanceAlreadyExistsFault for service response error code + // "DBInstanceAlreadyExists". + // + // User already has a DB instance with the given identifier. + ErrCodeDBInstanceAlreadyExistsFault = "DBInstanceAlreadyExists" + + // ErrCodeDBInstanceNotFoundFault for service response error code + // "DBInstanceNotFound". + // + // DBInstanceIdentifier does not refer to an existing DB instance. + ErrCodeDBInstanceNotFoundFault = "DBInstanceNotFound" + + // ErrCodeDBParameterGroupAlreadyExistsFault for service response error code + // "DBParameterGroupAlreadyExists". + // + // A DB parameter group with the same name exists. + ErrCodeDBParameterGroupAlreadyExistsFault = "DBParameterGroupAlreadyExists" + + // ErrCodeDBParameterGroupNotFoundFault for service response error code + // "DBParameterGroupNotFound". + // + // DBParameterGroupName does not refer to an existing DB parameter group. + ErrCodeDBParameterGroupNotFoundFault = "DBParameterGroupNotFound" + + // ErrCodeDBParameterGroupQuotaExceededFault for service response error code + // "DBParameterGroupQuotaExceeded". + // + // Request would result in user exceeding the allowed number of DB parameter + // groups. + ErrCodeDBParameterGroupQuotaExceededFault = "DBParameterGroupQuotaExceeded" + + // ErrCodeDBSecurityGroupNotFoundFault for service response error code + // "DBSecurityGroupNotFound". + // + // DBSecurityGroupName does not refer to an existing DB security group. + ErrCodeDBSecurityGroupNotFoundFault = "DBSecurityGroupNotFound" + + // ErrCodeDBSnapshotAlreadyExistsFault for service response error code + // "DBSnapshotAlreadyExists". + // + // DBSnapshotIdentifier is already used by an existing snapshot. + ErrCodeDBSnapshotAlreadyExistsFault = "DBSnapshotAlreadyExists" + + // ErrCodeDBSnapshotNotFoundFault for service response error code + // "DBSnapshotNotFound". + // + // DBSnapshotIdentifier does not refer to an existing DB snapshot. + ErrCodeDBSnapshotNotFoundFault = "DBSnapshotNotFound" + + // ErrCodeDBSubnetGroupAlreadyExistsFault for service response error code + // "DBSubnetGroupAlreadyExists". + // + // DBSubnetGroupName is already used by an existing DB subnet group. + ErrCodeDBSubnetGroupAlreadyExistsFault = "DBSubnetGroupAlreadyExists" + + // ErrCodeDBSubnetGroupDoesNotCoverEnoughAZs for service response error code + // "DBSubnetGroupDoesNotCoverEnoughAZs". + // + // Subnets in the DB subnet group should cover at least two Availability Zones + // unless there is only one Availability Zone. + ErrCodeDBSubnetGroupDoesNotCoverEnoughAZs = "DBSubnetGroupDoesNotCoverEnoughAZs" + + // ErrCodeDBSubnetGroupNotFoundFault for service response error code + // "DBSubnetGroupNotFoundFault". + // + // DBSubnetGroupName does not refer to an existing DB subnet group. + ErrCodeDBSubnetGroupNotFoundFault = "DBSubnetGroupNotFoundFault" + + // ErrCodeDBSubnetGroupQuotaExceededFault for service response error code + // "DBSubnetGroupQuotaExceeded". + // + // Request would result in user exceeding the allowed number of DB subnet groups. + ErrCodeDBSubnetGroupQuotaExceededFault = "DBSubnetGroupQuotaExceeded" + + // ErrCodeDBSubnetQuotaExceededFault for service response error code + // "DBSubnetQuotaExceededFault". + // + // Request would result in user exceeding the allowed number of subnets in a + // DB subnet groups. + ErrCodeDBSubnetQuotaExceededFault = "DBSubnetQuotaExceededFault" + + // ErrCodeDBUpgradeDependencyFailureFault for service response error code + // "DBUpgradeDependencyFailure". + // + // The DB upgrade failed because a resource the DB depends on could not be modified. + ErrCodeDBUpgradeDependencyFailureFault = "DBUpgradeDependencyFailure" + + // ErrCodeDomainNotFoundFault for service response error code + // "DomainNotFoundFault". + // + // Domain does not refer to an existing Active Directory Domain. + ErrCodeDomainNotFoundFault = "DomainNotFoundFault" + + // ErrCodeEventSubscriptionQuotaExceededFault for service response error code + // "EventSubscriptionQuotaExceeded". + ErrCodeEventSubscriptionQuotaExceededFault = "EventSubscriptionQuotaExceeded" + + // ErrCodeInstanceQuotaExceededFault for service response error code + // "InstanceQuotaExceeded". + // + // Request would result in user exceeding the allowed number of DB instances. + ErrCodeInstanceQuotaExceededFault = "InstanceQuotaExceeded" + + // ErrCodeInsufficientDBClusterCapacityFault for service response error code + // "InsufficientDBClusterCapacityFault". + // + // The DB cluster does not have enough capacity for the current operation. + ErrCodeInsufficientDBClusterCapacityFault = "InsufficientDBClusterCapacityFault" + + // ErrCodeInsufficientDBInstanceCapacityFault for service response error code + // "InsufficientDBInstanceCapacity". + // + // Specified DB instance class is not available in the specified Availability + // Zone. + ErrCodeInsufficientDBInstanceCapacityFault = "InsufficientDBInstanceCapacity" + + // ErrCodeInsufficientStorageClusterCapacityFault for service response error code + // "InsufficientStorageClusterCapacity". + // + // There is insufficient storage available for the current action. You may be + // able to resolve this error by updating your subnet group to use different + // Availability Zones that have more storage available. + ErrCodeInsufficientStorageClusterCapacityFault = "InsufficientStorageClusterCapacity" + + // ErrCodeInvalidDBClusterSnapshotStateFault for service response error code + // "InvalidDBClusterSnapshotStateFault". + // + // The supplied value is not a valid DB cluster snapshot state. + ErrCodeInvalidDBClusterSnapshotStateFault = "InvalidDBClusterSnapshotStateFault" + + // ErrCodeInvalidDBClusterStateFault for service response error code + // "InvalidDBClusterStateFault". + // + // The DB cluster is not in a valid state. + ErrCodeInvalidDBClusterStateFault = "InvalidDBClusterStateFault" + + // ErrCodeInvalidDBInstanceStateFault for service response error code + // "InvalidDBInstanceState". + // + // The specified DB instance is not in the available state. + ErrCodeInvalidDBInstanceStateFault = "InvalidDBInstanceState" + + // ErrCodeInvalidDBParameterGroupStateFault for service response error code + // "InvalidDBParameterGroupState". + // + // The DB parameter group is in use or is in an invalid state. If you are attempting + // to delete the parameter group, you cannot delete it when the parameter group + // is in this state. + ErrCodeInvalidDBParameterGroupStateFault = "InvalidDBParameterGroupState" + + // ErrCodeInvalidDBSecurityGroupStateFault for service response error code + // "InvalidDBSecurityGroupState". + // + // The state of the DB security group does not allow deletion. + ErrCodeInvalidDBSecurityGroupStateFault = "InvalidDBSecurityGroupState" + + // ErrCodeInvalidDBSnapshotStateFault for service response error code + // "InvalidDBSnapshotState". + // + // The state of the DB snapshot does not allow deletion. + ErrCodeInvalidDBSnapshotStateFault = "InvalidDBSnapshotState" + + // ErrCodeInvalidDBSubnetGroupStateFault for service response error code + // "InvalidDBSubnetGroupStateFault". + // + // The DB subnet group cannot be deleted because it is in use. + ErrCodeInvalidDBSubnetGroupStateFault = "InvalidDBSubnetGroupStateFault" + + // ErrCodeInvalidDBSubnetStateFault for service response error code + // "InvalidDBSubnetStateFault". + // + // The DB subnet is not in the available state. + ErrCodeInvalidDBSubnetStateFault = "InvalidDBSubnetStateFault" + + // ErrCodeInvalidEventSubscriptionStateFault for service response error code + // "InvalidEventSubscriptionState". + ErrCodeInvalidEventSubscriptionStateFault = "InvalidEventSubscriptionState" + + // ErrCodeInvalidRestoreFault for service response error code + // "InvalidRestoreFault". + // + // Cannot restore from vpc backup to non-vpc DB instance. + ErrCodeInvalidRestoreFault = "InvalidRestoreFault" + + // ErrCodeInvalidSubnet for service response error code + // "InvalidSubnet". + // + // The requested subnet is invalid, or multiple subnets were requested that + // are not all in a common VPC. + ErrCodeInvalidSubnet = "InvalidSubnet" + + // ErrCodeInvalidVPCNetworkStateFault for service response error code + // "InvalidVPCNetworkStateFault". + // + // DB subnet group does not cover all Availability Zones after it is created + // because users' change. + ErrCodeInvalidVPCNetworkStateFault = "InvalidVPCNetworkStateFault" + + // ErrCodeKMSKeyNotAccessibleFault for service response error code + // "KMSKeyNotAccessibleFault". + // + // Error accessing KMS key. + ErrCodeKMSKeyNotAccessibleFault = "KMSKeyNotAccessibleFault" + + // ErrCodeOptionGroupNotFoundFault for service response error code + // "OptionGroupNotFoundFault". + ErrCodeOptionGroupNotFoundFault = "OptionGroupNotFoundFault" + + // ErrCodeProvisionedIopsNotAvailableInAZFault for service response error code + // "ProvisionedIopsNotAvailableInAZFault". + // + // Provisioned IOPS not available in the specified Availability Zone. + ErrCodeProvisionedIopsNotAvailableInAZFault = "ProvisionedIopsNotAvailableInAZFault" + + // ErrCodeResourceNotFoundFault for service response error code + // "ResourceNotFoundFault". + // + // The specified resource ID was not found. + ErrCodeResourceNotFoundFault = "ResourceNotFoundFault" + + // ErrCodeSNSInvalidTopicFault for service response error code + // "SNSInvalidTopic". + ErrCodeSNSInvalidTopicFault = "SNSInvalidTopic" + + // ErrCodeSNSNoAuthorizationFault for service response error code + // "SNSNoAuthorization". + ErrCodeSNSNoAuthorizationFault = "SNSNoAuthorization" + + // ErrCodeSNSTopicArnNotFoundFault for service response error code + // "SNSTopicArnNotFound". + ErrCodeSNSTopicArnNotFoundFault = "SNSTopicArnNotFound" + + // ErrCodeSharedSnapshotQuotaExceededFault for service response error code + // "SharedSnapshotQuotaExceeded". + // + // You have exceeded the maximum number of accounts that you can share a manual + // DB snapshot with. + ErrCodeSharedSnapshotQuotaExceededFault = "SharedSnapshotQuotaExceeded" + + // ErrCodeSnapshotQuotaExceededFault for service response error code + // "SnapshotQuotaExceeded". + // + // Request would result in user exceeding the allowed number of DB snapshots. + ErrCodeSnapshotQuotaExceededFault = "SnapshotQuotaExceeded" + + // ErrCodeSourceNotFoundFault for service response error code + // "SourceNotFound". + ErrCodeSourceNotFoundFault = "SourceNotFound" + + // ErrCodeStorageQuotaExceededFault for service response error code + // "StorageQuotaExceeded". + // + // Request would result in user exceeding the allowed amount of storage available + // across all DB instances. + ErrCodeStorageQuotaExceededFault = "StorageQuotaExceeded" + + // ErrCodeStorageTypeNotSupportedFault for service response error code + // "StorageTypeNotSupported". + // + // StorageType specified cannot be associated with the DB Instance. + ErrCodeStorageTypeNotSupportedFault = "StorageTypeNotSupported" + + // ErrCodeSubnetAlreadyInUse for service response error code + // "SubnetAlreadyInUse". + // + // The DB subnet is already in use in the Availability Zone. + ErrCodeSubnetAlreadyInUse = "SubnetAlreadyInUse" + + // ErrCodeSubscriptionAlreadyExistFault for service response error code + // "SubscriptionAlreadyExist". + ErrCodeSubscriptionAlreadyExistFault = "SubscriptionAlreadyExist" + + // ErrCodeSubscriptionCategoryNotFoundFault for service response error code + // "SubscriptionCategoryNotFound". + ErrCodeSubscriptionCategoryNotFoundFault = "SubscriptionCategoryNotFound" + + // ErrCodeSubscriptionNotFoundFault for service response error code + // "SubscriptionNotFound". + ErrCodeSubscriptionNotFoundFault = "SubscriptionNotFound" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/neptune/neptuneiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/neptune/neptuneiface/interface.go new file mode 100644 index 000000000..a83b20711 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/neptune/neptuneiface/interface.go @@ -0,0 +1,325 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package neptuneiface provides an interface to enable mocking the Amazon Neptune service client +// for testing your code. +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. +package neptuneiface + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/service/neptune" +) + +// NeptuneAPI provides an interface to enable mocking the +// neptune.Neptune service client's API operation, +// paginators, and waiters. This make unit testing your code that calls out +// to the SDK's service client's calls easier. +// +// The best way to use this interface is so the SDK's service client's calls +// can be stubbed out for unit testing your code with the SDK without needing +// to inject custom request handlers into the SDK's request pipeline. +// +// // myFunc uses an SDK service client to make a request to +// // Amazon Neptune. +// func myFunc(svc neptuneiface.NeptuneAPI) bool { +// // Make svc.AddRoleToDBCluster request +// } +// +// func main() { +// sess := session.New() +// svc := neptune.New(sess) +// +// myFunc(svc) +// } +// +// In your _test.go file: +// +// // Define a mock struct to be used in your unit tests of myFunc. +// type mockNeptuneClient struct { +// neptuneiface.NeptuneAPI +// } +// func (m *mockNeptuneClient) AddRoleToDBCluster(input *neptune.AddRoleToDBClusterInput) (*neptune.AddRoleToDBClusterOutput, error) { +// // mock response/functionality +// } +// +// func TestMyFunc(t *testing.T) { +// // Setup Test +// mockSvc := &mockNeptuneClient{} +// +// myfunc(mockSvc) +// +// // Verify myFunc's functionality +// } +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. Its suggested to use the pattern above for testing, or using +// tooling to generate mocks to satisfy the interfaces. +type NeptuneAPI interface { + AddRoleToDBCluster(*neptune.AddRoleToDBClusterInput) (*neptune.AddRoleToDBClusterOutput, error) + AddRoleToDBClusterWithContext(aws.Context, *neptune.AddRoleToDBClusterInput, ...request.Option) (*neptune.AddRoleToDBClusterOutput, error) + AddRoleToDBClusterRequest(*neptune.AddRoleToDBClusterInput) (*request.Request, *neptune.AddRoleToDBClusterOutput) + + AddSourceIdentifierToSubscription(*neptune.AddSourceIdentifierToSubscriptionInput) (*neptune.AddSourceIdentifierToSubscriptionOutput, error) + AddSourceIdentifierToSubscriptionWithContext(aws.Context, *neptune.AddSourceIdentifierToSubscriptionInput, ...request.Option) (*neptune.AddSourceIdentifierToSubscriptionOutput, error) + AddSourceIdentifierToSubscriptionRequest(*neptune.AddSourceIdentifierToSubscriptionInput) (*request.Request, *neptune.AddSourceIdentifierToSubscriptionOutput) + + AddTagsToResource(*neptune.AddTagsToResourceInput) (*neptune.AddTagsToResourceOutput, error) + AddTagsToResourceWithContext(aws.Context, *neptune.AddTagsToResourceInput, ...request.Option) (*neptune.AddTagsToResourceOutput, error) + AddTagsToResourceRequest(*neptune.AddTagsToResourceInput) (*request.Request, *neptune.AddTagsToResourceOutput) + + ApplyPendingMaintenanceAction(*neptune.ApplyPendingMaintenanceActionInput) (*neptune.ApplyPendingMaintenanceActionOutput, error) + ApplyPendingMaintenanceActionWithContext(aws.Context, *neptune.ApplyPendingMaintenanceActionInput, ...request.Option) (*neptune.ApplyPendingMaintenanceActionOutput, error) + ApplyPendingMaintenanceActionRequest(*neptune.ApplyPendingMaintenanceActionInput) (*request.Request, *neptune.ApplyPendingMaintenanceActionOutput) + + CopyDBClusterParameterGroup(*neptune.CopyDBClusterParameterGroupInput) (*neptune.CopyDBClusterParameterGroupOutput, error) + CopyDBClusterParameterGroupWithContext(aws.Context, *neptune.CopyDBClusterParameterGroupInput, ...request.Option) (*neptune.CopyDBClusterParameterGroupOutput, error) + CopyDBClusterParameterGroupRequest(*neptune.CopyDBClusterParameterGroupInput) (*request.Request, *neptune.CopyDBClusterParameterGroupOutput) + + CopyDBClusterSnapshot(*neptune.CopyDBClusterSnapshotInput) (*neptune.CopyDBClusterSnapshotOutput, error) + CopyDBClusterSnapshotWithContext(aws.Context, *neptune.CopyDBClusterSnapshotInput, ...request.Option) (*neptune.CopyDBClusterSnapshotOutput, error) + CopyDBClusterSnapshotRequest(*neptune.CopyDBClusterSnapshotInput) (*request.Request, *neptune.CopyDBClusterSnapshotOutput) + + CopyDBParameterGroup(*neptune.CopyDBParameterGroupInput) (*neptune.CopyDBParameterGroupOutput, error) + CopyDBParameterGroupWithContext(aws.Context, *neptune.CopyDBParameterGroupInput, ...request.Option) (*neptune.CopyDBParameterGroupOutput, error) + CopyDBParameterGroupRequest(*neptune.CopyDBParameterGroupInput) (*request.Request, *neptune.CopyDBParameterGroupOutput) + + CreateDBCluster(*neptune.CreateDBClusterInput) (*neptune.CreateDBClusterOutput, error) + CreateDBClusterWithContext(aws.Context, *neptune.CreateDBClusterInput, ...request.Option) (*neptune.CreateDBClusterOutput, error) + CreateDBClusterRequest(*neptune.CreateDBClusterInput) (*request.Request, *neptune.CreateDBClusterOutput) + + CreateDBClusterParameterGroup(*neptune.CreateDBClusterParameterGroupInput) (*neptune.CreateDBClusterParameterGroupOutput, error) + CreateDBClusterParameterGroupWithContext(aws.Context, *neptune.CreateDBClusterParameterGroupInput, ...request.Option) (*neptune.CreateDBClusterParameterGroupOutput, error) + CreateDBClusterParameterGroupRequest(*neptune.CreateDBClusterParameterGroupInput) (*request.Request, *neptune.CreateDBClusterParameterGroupOutput) + + CreateDBClusterSnapshot(*neptune.CreateDBClusterSnapshotInput) (*neptune.CreateDBClusterSnapshotOutput, error) + CreateDBClusterSnapshotWithContext(aws.Context, *neptune.CreateDBClusterSnapshotInput, ...request.Option) (*neptune.CreateDBClusterSnapshotOutput, error) + CreateDBClusterSnapshotRequest(*neptune.CreateDBClusterSnapshotInput) (*request.Request, *neptune.CreateDBClusterSnapshotOutput) + + CreateDBInstance(*neptune.CreateDBInstanceInput) (*neptune.CreateDBInstanceOutput, error) + CreateDBInstanceWithContext(aws.Context, *neptune.CreateDBInstanceInput, ...request.Option) (*neptune.CreateDBInstanceOutput, error) + CreateDBInstanceRequest(*neptune.CreateDBInstanceInput) (*request.Request, *neptune.CreateDBInstanceOutput) + + CreateDBParameterGroup(*neptune.CreateDBParameterGroupInput) (*neptune.CreateDBParameterGroupOutput, error) + CreateDBParameterGroupWithContext(aws.Context, *neptune.CreateDBParameterGroupInput, ...request.Option) (*neptune.CreateDBParameterGroupOutput, error) + CreateDBParameterGroupRequest(*neptune.CreateDBParameterGroupInput) (*request.Request, *neptune.CreateDBParameterGroupOutput) + + CreateDBSubnetGroup(*neptune.CreateDBSubnetGroupInput) (*neptune.CreateDBSubnetGroupOutput, error) + CreateDBSubnetGroupWithContext(aws.Context, *neptune.CreateDBSubnetGroupInput, ...request.Option) (*neptune.CreateDBSubnetGroupOutput, error) + CreateDBSubnetGroupRequest(*neptune.CreateDBSubnetGroupInput) (*request.Request, *neptune.CreateDBSubnetGroupOutput) + + CreateEventSubscription(*neptune.CreateEventSubscriptionInput) (*neptune.CreateEventSubscriptionOutput, error) + CreateEventSubscriptionWithContext(aws.Context, *neptune.CreateEventSubscriptionInput, ...request.Option) (*neptune.CreateEventSubscriptionOutput, error) + CreateEventSubscriptionRequest(*neptune.CreateEventSubscriptionInput) (*request.Request, *neptune.CreateEventSubscriptionOutput) + + DeleteDBCluster(*neptune.DeleteDBClusterInput) (*neptune.DeleteDBClusterOutput, error) + DeleteDBClusterWithContext(aws.Context, *neptune.DeleteDBClusterInput, ...request.Option) (*neptune.DeleteDBClusterOutput, error) + DeleteDBClusterRequest(*neptune.DeleteDBClusterInput) (*request.Request, *neptune.DeleteDBClusterOutput) + + DeleteDBClusterParameterGroup(*neptune.DeleteDBClusterParameterGroupInput) (*neptune.DeleteDBClusterParameterGroupOutput, error) + DeleteDBClusterParameterGroupWithContext(aws.Context, *neptune.DeleteDBClusterParameterGroupInput, ...request.Option) (*neptune.DeleteDBClusterParameterGroupOutput, error) + DeleteDBClusterParameterGroupRequest(*neptune.DeleteDBClusterParameterGroupInput) (*request.Request, *neptune.DeleteDBClusterParameterGroupOutput) + + DeleteDBClusterSnapshot(*neptune.DeleteDBClusterSnapshotInput) (*neptune.DeleteDBClusterSnapshotOutput, error) + DeleteDBClusterSnapshotWithContext(aws.Context, *neptune.DeleteDBClusterSnapshotInput, ...request.Option) (*neptune.DeleteDBClusterSnapshotOutput, error) + DeleteDBClusterSnapshotRequest(*neptune.DeleteDBClusterSnapshotInput) (*request.Request, *neptune.DeleteDBClusterSnapshotOutput) + + DeleteDBInstance(*neptune.DeleteDBInstanceInput) (*neptune.DeleteDBInstanceOutput, error) + DeleteDBInstanceWithContext(aws.Context, *neptune.DeleteDBInstanceInput, ...request.Option) (*neptune.DeleteDBInstanceOutput, error) + DeleteDBInstanceRequest(*neptune.DeleteDBInstanceInput) (*request.Request, *neptune.DeleteDBInstanceOutput) + + DeleteDBParameterGroup(*neptune.DeleteDBParameterGroupInput) (*neptune.DeleteDBParameterGroupOutput, error) + DeleteDBParameterGroupWithContext(aws.Context, *neptune.DeleteDBParameterGroupInput, ...request.Option) (*neptune.DeleteDBParameterGroupOutput, error) + DeleteDBParameterGroupRequest(*neptune.DeleteDBParameterGroupInput) (*request.Request, *neptune.DeleteDBParameterGroupOutput) + + DeleteDBSubnetGroup(*neptune.DeleteDBSubnetGroupInput) (*neptune.DeleteDBSubnetGroupOutput, error) + DeleteDBSubnetGroupWithContext(aws.Context, *neptune.DeleteDBSubnetGroupInput, ...request.Option) (*neptune.DeleteDBSubnetGroupOutput, error) + DeleteDBSubnetGroupRequest(*neptune.DeleteDBSubnetGroupInput) (*request.Request, *neptune.DeleteDBSubnetGroupOutput) + + DeleteEventSubscription(*neptune.DeleteEventSubscriptionInput) (*neptune.DeleteEventSubscriptionOutput, error) + DeleteEventSubscriptionWithContext(aws.Context, *neptune.DeleteEventSubscriptionInput, ...request.Option) (*neptune.DeleteEventSubscriptionOutput, error) + DeleteEventSubscriptionRequest(*neptune.DeleteEventSubscriptionInput) (*request.Request, *neptune.DeleteEventSubscriptionOutput) + + DescribeDBClusterParameterGroups(*neptune.DescribeDBClusterParameterGroupsInput) (*neptune.DescribeDBClusterParameterGroupsOutput, error) + DescribeDBClusterParameterGroupsWithContext(aws.Context, *neptune.DescribeDBClusterParameterGroupsInput, ...request.Option) (*neptune.DescribeDBClusterParameterGroupsOutput, error) + DescribeDBClusterParameterGroupsRequest(*neptune.DescribeDBClusterParameterGroupsInput) (*request.Request, *neptune.DescribeDBClusterParameterGroupsOutput) + + DescribeDBClusterParameters(*neptune.DescribeDBClusterParametersInput) (*neptune.DescribeDBClusterParametersOutput, error) + DescribeDBClusterParametersWithContext(aws.Context, *neptune.DescribeDBClusterParametersInput, ...request.Option) (*neptune.DescribeDBClusterParametersOutput, error) + DescribeDBClusterParametersRequest(*neptune.DescribeDBClusterParametersInput) (*request.Request, *neptune.DescribeDBClusterParametersOutput) + + DescribeDBClusterSnapshotAttributes(*neptune.DescribeDBClusterSnapshotAttributesInput) (*neptune.DescribeDBClusterSnapshotAttributesOutput, error) + DescribeDBClusterSnapshotAttributesWithContext(aws.Context, *neptune.DescribeDBClusterSnapshotAttributesInput, ...request.Option) (*neptune.DescribeDBClusterSnapshotAttributesOutput, error) + DescribeDBClusterSnapshotAttributesRequest(*neptune.DescribeDBClusterSnapshotAttributesInput) (*request.Request, *neptune.DescribeDBClusterSnapshotAttributesOutput) + + DescribeDBClusterSnapshots(*neptune.DescribeDBClusterSnapshotsInput) (*neptune.DescribeDBClusterSnapshotsOutput, error) + DescribeDBClusterSnapshotsWithContext(aws.Context, *neptune.DescribeDBClusterSnapshotsInput, ...request.Option) (*neptune.DescribeDBClusterSnapshotsOutput, error) + DescribeDBClusterSnapshotsRequest(*neptune.DescribeDBClusterSnapshotsInput) (*request.Request, *neptune.DescribeDBClusterSnapshotsOutput) + + DescribeDBClusters(*neptune.DescribeDBClustersInput) (*neptune.DescribeDBClustersOutput, error) + DescribeDBClustersWithContext(aws.Context, *neptune.DescribeDBClustersInput, ...request.Option) (*neptune.DescribeDBClustersOutput, error) + DescribeDBClustersRequest(*neptune.DescribeDBClustersInput) (*request.Request, *neptune.DescribeDBClustersOutput) + + DescribeDBEngineVersions(*neptune.DescribeDBEngineVersionsInput) (*neptune.DescribeDBEngineVersionsOutput, error) + DescribeDBEngineVersionsWithContext(aws.Context, *neptune.DescribeDBEngineVersionsInput, ...request.Option) (*neptune.DescribeDBEngineVersionsOutput, error) + DescribeDBEngineVersionsRequest(*neptune.DescribeDBEngineVersionsInput) (*request.Request, *neptune.DescribeDBEngineVersionsOutput) + + DescribeDBEngineVersionsPages(*neptune.DescribeDBEngineVersionsInput, func(*neptune.DescribeDBEngineVersionsOutput, bool) bool) error + DescribeDBEngineVersionsPagesWithContext(aws.Context, *neptune.DescribeDBEngineVersionsInput, func(*neptune.DescribeDBEngineVersionsOutput, bool) bool, ...request.Option) error + + DescribeDBInstances(*neptune.DescribeDBInstancesInput) (*neptune.DescribeDBInstancesOutput, error) + DescribeDBInstancesWithContext(aws.Context, *neptune.DescribeDBInstancesInput, ...request.Option) (*neptune.DescribeDBInstancesOutput, error) + DescribeDBInstancesRequest(*neptune.DescribeDBInstancesInput) (*request.Request, *neptune.DescribeDBInstancesOutput) + + DescribeDBInstancesPages(*neptune.DescribeDBInstancesInput, func(*neptune.DescribeDBInstancesOutput, bool) bool) error + DescribeDBInstancesPagesWithContext(aws.Context, *neptune.DescribeDBInstancesInput, func(*neptune.DescribeDBInstancesOutput, bool) bool, ...request.Option) error + + DescribeDBParameterGroups(*neptune.DescribeDBParameterGroupsInput) (*neptune.DescribeDBParameterGroupsOutput, error) + DescribeDBParameterGroupsWithContext(aws.Context, *neptune.DescribeDBParameterGroupsInput, ...request.Option) (*neptune.DescribeDBParameterGroupsOutput, error) + DescribeDBParameterGroupsRequest(*neptune.DescribeDBParameterGroupsInput) (*request.Request, *neptune.DescribeDBParameterGroupsOutput) + + DescribeDBParameterGroupsPages(*neptune.DescribeDBParameterGroupsInput, func(*neptune.DescribeDBParameterGroupsOutput, bool) bool) error + DescribeDBParameterGroupsPagesWithContext(aws.Context, *neptune.DescribeDBParameterGroupsInput, func(*neptune.DescribeDBParameterGroupsOutput, bool) bool, ...request.Option) error + + DescribeDBParameters(*neptune.DescribeDBParametersInput) (*neptune.DescribeDBParametersOutput, error) + DescribeDBParametersWithContext(aws.Context, *neptune.DescribeDBParametersInput, ...request.Option) (*neptune.DescribeDBParametersOutput, error) + DescribeDBParametersRequest(*neptune.DescribeDBParametersInput) (*request.Request, *neptune.DescribeDBParametersOutput) + + DescribeDBParametersPages(*neptune.DescribeDBParametersInput, func(*neptune.DescribeDBParametersOutput, bool) bool) error + DescribeDBParametersPagesWithContext(aws.Context, *neptune.DescribeDBParametersInput, func(*neptune.DescribeDBParametersOutput, bool) bool, ...request.Option) error + + DescribeDBSubnetGroups(*neptune.DescribeDBSubnetGroupsInput) (*neptune.DescribeDBSubnetGroupsOutput, error) + DescribeDBSubnetGroupsWithContext(aws.Context, *neptune.DescribeDBSubnetGroupsInput, ...request.Option) (*neptune.DescribeDBSubnetGroupsOutput, error) + DescribeDBSubnetGroupsRequest(*neptune.DescribeDBSubnetGroupsInput) (*request.Request, *neptune.DescribeDBSubnetGroupsOutput) + + DescribeDBSubnetGroupsPages(*neptune.DescribeDBSubnetGroupsInput, func(*neptune.DescribeDBSubnetGroupsOutput, bool) bool) error + DescribeDBSubnetGroupsPagesWithContext(aws.Context, *neptune.DescribeDBSubnetGroupsInput, func(*neptune.DescribeDBSubnetGroupsOutput, bool) bool, ...request.Option) error + + DescribeEngineDefaultClusterParameters(*neptune.DescribeEngineDefaultClusterParametersInput) (*neptune.DescribeEngineDefaultClusterParametersOutput, error) + DescribeEngineDefaultClusterParametersWithContext(aws.Context, *neptune.DescribeEngineDefaultClusterParametersInput, ...request.Option) (*neptune.DescribeEngineDefaultClusterParametersOutput, error) + DescribeEngineDefaultClusterParametersRequest(*neptune.DescribeEngineDefaultClusterParametersInput) (*request.Request, *neptune.DescribeEngineDefaultClusterParametersOutput) + + DescribeEngineDefaultParameters(*neptune.DescribeEngineDefaultParametersInput) (*neptune.DescribeEngineDefaultParametersOutput, error) + DescribeEngineDefaultParametersWithContext(aws.Context, *neptune.DescribeEngineDefaultParametersInput, ...request.Option) (*neptune.DescribeEngineDefaultParametersOutput, error) + DescribeEngineDefaultParametersRequest(*neptune.DescribeEngineDefaultParametersInput) (*request.Request, *neptune.DescribeEngineDefaultParametersOutput) + + DescribeEngineDefaultParametersPages(*neptune.DescribeEngineDefaultParametersInput, func(*neptune.DescribeEngineDefaultParametersOutput, bool) bool) error + DescribeEngineDefaultParametersPagesWithContext(aws.Context, *neptune.DescribeEngineDefaultParametersInput, func(*neptune.DescribeEngineDefaultParametersOutput, bool) bool, ...request.Option) error + + DescribeEventCategories(*neptune.DescribeEventCategoriesInput) (*neptune.DescribeEventCategoriesOutput, error) + DescribeEventCategoriesWithContext(aws.Context, *neptune.DescribeEventCategoriesInput, ...request.Option) (*neptune.DescribeEventCategoriesOutput, error) + DescribeEventCategoriesRequest(*neptune.DescribeEventCategoriesInput) (*request.Request, *neptune.DescribeEventCategoriesOutput) + + DescribeEventSubscriptions(*neptune.DescribeEventSubscriptionsInput) (*neptune.DescribeEventSubscriptionsOutput, error) + DescribeEventSubscriptionsWithContext(aws.Context, *neptune.DescribeEventSubscriptionsInput, ...request.Option) (*neptune.DescribeEventSubscriptionsOutput, error) + DescribeEventSubscriptionsRequest(*neptune.DescribeEventSubscriptionsInput) (*request.Request, *neptune.DescribeEventSubscriptionsOutput) + + DescribeEventSubscriptionsPages(*neptune.DescribeEventSubscriptionsInput, func(*neptune.DescribeEventSubscriptionsOutput, bool) bool) error + DescribeEventSubscriptionsPagesWithContext(aws.Context, *neptune.DescribeEventSubscriptionsInput, func(*neptune.DescribeEventSubscriptionsOutput, bool) bool, ...request.Option) error + + DescribeEvents(*neptune.DescribeEventsInput) (*neptune.DescribeEventsOutput, error) + DescribeEventsWithContext(aws.Context, *neptune.DescribeEventsInput, ...request.Option) (*neptune.DescribeEventsOutput, error) + DescribeEventsRequest(*neptune.DescribeEventsInput) (*request.Request, *neptune.DescribeEventsOutput) + + DescribeEventsPages(*neptune.DescribeEventsInput, func(*neptune.DescribeEventsOutput, bool) bool) error + DescribeEventsPagesWithContext(aws.Context, *neptune.DescribeEventsInput, func(*neptune.DescribeEventsOutput, bool) bool, ...request.Option) error + + DescribeOrderableDBInstanceOptions(*neptune.DescribeOrderableDBInstanceOptionsInput) (*neptune.DescribeOrderableDBInstanceOptionsOutput, error) + DescribeOrderableDBInstanceOptionsWithContext(aws.Context, *neptune.DescribeOrderableDBInstanceOptionsInput, ...request.Option) (*neptune.DescribeOrderableDBInstanceOptionsOutput, error) + DescribeOrderableDBInstanceOptionsRequest(*neptune.DescribeOrderableDBInstanceOptionsInput) (*request.Request, *neptune.DescribeOrderableDBInstanceOptionsOutput) + + DescribeOrderableDBInstanceOptionsPages(*neptune.DescribeOrderableDBInstanceOptionsInput, func(*neptune.DescribeOrderableDBInstanceOptionsOutput, bool) bool) error + DescribeOrderableDBInstanceOptionsPagesWithContext(aws.Context, *neptune.DescribeOrderableDBInstanceOptionsInput, func(*neptune.DescribeOrderableDBInstanceOptionsOutput, bool) bool, ...request.Option) error + + DescribePendingMaintenanceActions(*neptune.DescribePendingMaintenanceActionsInput) (*neptune.DescribePendingMaintenanceActionsOutput, error) + DescribePendingMaintenanceActionsWithContext(aws.Context, *neptune.DescribePendingMaintenanceActionsInput, ...request.Option) (*neptune.DescribePendingMaintenanceActionsOutput, error) + DescribePendingMaintenanceActionsRequest(*neptune.DescribePendingMaintenanceActionsInput) (*request.Request, *neptune.DescribePendingMaintenanceActionsOutput) + + DescribeValidDBInstanceModifications(*neptune.DescribeValidDBInstanceModificationsInput) (*neptune.DescribeValidDBInstanceModificationsOutput, error) + DescribeValidDBInstanceModificationsWithContext(aws.Context, *neptune.DescribeValidDBInstanceModificationsInput, ...request.Option) (*neptune.DescribeValidDBInstanceModificationsOutput, error) + DescribeValidDBInstanceModificationsRequest(*neptune.DescribeValidDBInstanceModificationsInput) (*request.Request, *neptune.DescribeValidDBInstanceModificationsOutput) + + FailoverDBCluster(*neptune.FailoverDBClusterInput) (*neptune.FailoverDBClusterOutput, error) + FailoverDBClusterWithContext(aws.Context, *neptune.FailoverDBClusterInput, ...request.Option) (*neptune.FailoverDBClusterOutput, error) + FailoverDBClusterRequest(*neptune.FailoverDBClusterInput) (*request.Request, *neptune.FailoverDBClusterOutput) + + ListTagsForResource(*neptune.ListTagsForResourceInput) (*neptune.ListTagsForResourceOutput, error) + ListTagsForResourceWithContext(aws.Context, *neptune.ListTagsForResourceInput, ...request.Option) (*neptune.ListTagsForResourceOutput, error) + ListTagsForResourceRequest(*neptune.ListTagsForResourceInput) (*request.Request, *neptune.ListTagsForResourceOutput) + + ModifyDBCluster(*neptune.ModifyDBClusterInput) (*neptune.ModifyDBClusterOutput, error) + ModifyDBClusterWithContext(aws.Context, *neptune.ModifyDBClusterInput, ...request.Option) (*neptune.ModifyDBClusterOutput, error) + ModifyDBClusterRequest(*neptune.ModifyDBClusterInput) (*request.Request, *neptune.ModifyDBClusterOutput) + + ModifyDBClusterParameterGroup(*neptune.ModifyDBClusterParameterGroupInput) (*neptune.ResetDBClusterParameterGroupOutput, error) + ModifyDBClusterParameterGroupWithContext(aws.Context, *neptune.ModifyDBClusterParameterGroupInput, ...request.Option) (*neptune.ResetDBClusterParameterGroupOutput, error) + ModifyDBClusterParameterGroupRequest(*neptune.ModifyDBClusterParameterGroupInput) (*request.Request, *neptune.ResetDBClusterParameterGroupOutput) + + ModifyDBClusterSnapshotAttribute(*neptune.ModifyDBClusterSnapshotAttributeInput) (*neptune.ModifyDBClusterSnapshotAttributeOutput, error) + ModifyDBClusterSnapshotAttributeWithContext(aws.Context, *neptune.ModifyDBClusterSnapshotAttributeInput, ...request.Option) (*neptune.ModifyDBClusterSnapshotAttributeOutput, error) + ModifyDBClusterSnapshotAttributeRequest(*neptune.ModifyDBClusterSnapshotAttributeInput) (*request.Request, *neptune.ModifyDBClusterSnapshotAttributeOutput) + + ModifyDBInstance(*neptune.ModifyDBInstanceInput) (*neptune.ModifyDBInstanceOutput, error) + ModifyDBInstanceWithContext(aws.Context, *neptune.ModifyDBInstanceInput, ...request.Option) (*neptune.ModifyDBInstanceOutput, error) + ModifyDBInstanceRequest(*neptune.ModifyDBInstanceInput) (*request.Request, *neptune.ModifyDBInstanceOutput) + + ModifyDBParameterGroup(*neptune.ModifyDBParameterGroupInput) (*neptune.ResetDBParameterGroupOutput, error) + ModifyDBParameterGroupWithContext(aws.Context, *neptune.ModifyDBParameterGroupInput, ...request.Option) (*neptune.ResetDBParameterGroupOutput, error) + ModifyDBParameterGroupRequest(*neptune.ModifyDBParameterGroupInput) (*request.Request, *neptune.ResetDBParameterGroupOutput) + + ModifyDBSubnetGroup(*neptune.ModifyDBSubnetGroupInput) (*neptune.ModifyDBSubnetGroupOutput, error) + ModifyDBSubnetGroupWithContext(aws.Context, *neptune.ModifyDBSubnetGroupInput, ...request.Option) (*neptune.ModifyDBSubnetGroupOutput, error) + ModifyDBSubnetGroupRequest(*neptune.ModifyDBSubnetGroupInput) (*request.Request, *neptune.ModifyDBSubnetGroupOutput) + + ModifyEventSubscription(*neptune.ModifyEventSubscriptionInput) (*neptune.ModifyEventSubscriptionOutput, error) + ModifyEventSubscriptionWithContext(aws.Context, *neptune.ModifyEventSubscriptionInput, ...request.Option) (*neptune.ModifyEventSubscriptionOutput, error) + ModifyEventSubscriptionRequest(*neptune.ModifyEventSubscriptionInput) (*request.Request, *neptune.ModifyEventSubscriptionOutput) + + PromoteReadReplicaDBCluster(*neptune.PromoteReadReplicaDBClusterInput) (*neptune.PromoteReadReplicaDBClusterOutput, error) + PromoteReadReplicaDBClusterWithContext(aws.Context, *neptune.PromoteReadReplicaDBClusterInput, ...request.Option) (*neptune.PromoteReadReplicaDBClusterOutput, error) + PromoteReadReplicaDBClusterRequest(*neptune.PromoteReadReplicaDBClusterInput) (*request.Request, *neptune.PromoteReadReplicaDBClusterOutput) + + RebootDBInstance(*neptune.RebootDBInstanceInput) (*neptune.RebootDBInstanceOutput, error) + RebootDBInstanceWithContext(aws.Context, *neptune.RebootDBInstanceInput, ...request.Option) (*neptune.RebootDBInstanceOutput, error) + RebootDBInstanceRequest(*neptune.RebootDBInstanceInput) (*request.Request, *neptune.RebootDBInstanceOutput) + + RemoveRoleFromDBCluster(*neptune.RemoveRoleFromDBClusterInput) (*neptune.RemoveRoleFromDBClusterOutput, error) + RemoveRoleFromDBClusterWithContext(aws.Context, *neptune.RemoveRoleFromDBClusterInput, ...request.Option) (*neptune.RemoveRoleFromDBClusterOutput, error) + RemoveRoleFromDBClusterRequest(*neptune.RemoveRoleFromDBClusterInput) (*request.Request, *neptune.RemoveRoleFromDBClusterOutput) + + RemoveSourceIdentifierFromSubscription(*neptune.RemoveSourceIdentifierFromSubscriptionInput) (*neptune.RemoveSourceIdentifierFromSubscriptionOutput, error) + RemoveSourceIdentifierFromSubscriptionWithContext(aws.Context, *neptune.RemoveSourceIdentifierFromSubscriptionInput, ...request.Option) (*neptune.RemoveSourceIdentifierFromSubscriptionOutput, error) + RemoveSourceIdentifierFromSubscriptionRequest(*neptune.RemoveSourceIdentifierFromSubscriptionInput) (*request.Request, *neptune.RemoveSourceIdentifierFromSubscriptionOutput) + + RemoveTagsFromResource(*neptune.RemoveTagsFromResourceInput) (*neptune.RemoveTagsFromResourceOutput, error) + RemoveTagsFromResourceWithContext(aws.Context, *neptune.RemoveTagsFromResourceInput, ...request.Option) (*neptune.RemoveTagsFromResourceOutput, error) + RemoveTagsFromResourceRequest(*neptune.RemoveTagsFromResourceInput) (*request.Request, *neptune.RemoveTagsFromResourceOutput) + + ResetDBClusterParameterGroup(*neptune.ResetDBClusterParameterGroupInput) (*neptune.ResetDBClusterParameterGroupOutput, error) + ResetDBClusterParameterGroupWithContext(aws.Context, *neptune.ResetDBClusterParameterGroupInput, ...request.Option) (*neptune.ResetDBClusterParameterGroupOutput, error) + ResetDBClusterParameterGroupRequest(*neptune.ResetDBClusterParameterGroupInput) (*request.Request, *neptune.ResetDBClusterParameterGroupOutput) + + ResetDBParameterGroup(*neptune.ResetDBParameterGroupInput) (*neptune.ResetDBParameterGroupOutput, error) + ResetDBParameterGroupWithContext(aws.Context, *neptune.ResetDBParameterGroupInput, ...request.Option) (*neptune.ResetDBParameterGroupOutput, error) + ResetDBParameterGroupRequest(*neptune.ResetDBParameterGroupInput) (*request.Request, *neptune.ResetDBParameterGroupOutput) + + RestoreDBClusterFromSnapshot(*neptune.RestoreDBClusterFromSnapshotInput) (*neptune.RestoreDBClusterFromSnapshotOutput, error) + RestoreDBClusterFromSnapshotWithContext(aws.Context, *neptune.RestoreDBClusterFromSnapshotInput, ...request.Option) (*neptune.RestoreDBClusterFromSnapshotOutput, error) + RestoreDBClusterFromSnapshotRequest(*neptune.RestoreDBClusterFromSnapshotInput) (*request.Request, *neptune.RestoreDBClusterFromSnapshotOutput) + + RestoreDBClusterToPointInTime(*neptune.RestoreDBClusterToPointInTimeInput) (*neptune.RestoreDBClusterToPointInTimeOutput, error) + RestoreDBClusterToPointInTimeWithContext(aws.Context, *neptune.RestoreDBClusterToPointInTimeInput, ...request.Option) (*neptune.RestoreDBClusterToPointInTimeOutput, error) + RestoreDBClusterToPointInTimeRequest(*neptune.RestoreDBClusterToPointInTimeInput) (*request.Request, *neptune.RestoreDBClusterToPointInTimeOutput) + + WaitUntilDBInstanceAvailable(*neptune.DescribeDBInstancesInput) error + WaitUntilDBInstanceAvailableWithContext(aws.Context, *neptune.DescribeDBInstancesInput, ...request.WaiterOption) error + + WaitUntilDBInstanceDeleted(*neptune.DescribeDBInstancesInput) error + WaitUntilDBInstanceDeletedWithContext(aws.Context, *neptune.DescribeDBInstancesInput, ...request.WaiterOption) error +} + +var _ NeptuneAPI = (*neptune.Neptune)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/neptune/service.go b/vendor/github.com/aws/aws-sdk-go/service/neptune/service.go new file mode 100644 index 000000000..3ddc5e5fb --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/neptune/service.go @@ -0,0 +1,98 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package neptune + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/signer/v4" + "github.com/aws/aws-sdk-go/private/protocol/query" +) + +// Neptune provides the API operation methods for making requests to +// Amazon Neptune. See this package's package overview docs +// for details on the service. +// +// Neptune methods are safe to use concurrently. It is not safe to +// modify mutate any of the struct's properties though. +type Neptune struct { + *client.Client +} + +// Used for custom client initialization logic +var initClient func(*client.Client) + +// Used for custom request initialization logic +var initRequest func(*request.Request) + +// Service information constants +const ( + ServiceName = "rds" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Neptune" // ServiceID is a unique identifer of a specific service. +) + +// New creates a new instance of the Neptune client with a session. +// If additional configuration is needed for the client instance use the optional +// aws.Config parameter to add your extra config. +// +// Example: +// // Create a Neptune client from just a session. +// svc := neptune.New(mySession) +// +// // Create a Neptune client with additional configuration +// svc := neptune.New(mySession, aws.NewConfig().WithRegion("us-west-2")) +func New(p client.ConfigProvider, cfgs ...*aws.Config) *Neptune { + c := p.ClientConfig(EndpointsID, cfgs...) + if c.SigningNameDerived || len(c.SigningName) == 0 { + c.SigningName = "rds" + } + return newClient(*c.Config, c.Handlers, c.Endpoint, c.SigningRegion, c.SigningName) +} + +// newClient creates, initializes and returns a new service client instance. +func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion, signingName string) *Neptune { + svc := &Neptune{ + Client: client.New( + cfg, + metadata.ClientInfo{ + ServiceName: ServiceName, + ServiceID: ServiceID, + SigningName: signingName, + SigningRegion: signingRegion, + Endpoint: endpoint, + APIVersion: "2014-10-31", + }, + handlers, + ), + } + + // Handlers + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(query.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(query.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(query.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(query.UnmarshalErrorHandler) + + // Run custom client initialization if present + if initClient != nil { + initClient(svc.Client) + } + + return svc +} + +// newRequest creates a new request for a Neptune operation and runs any +// custom request initialization. +func (c *Neptune) newRequest(op *request.Operation, params, data interface{}) *request.Request { + req := c.NewRequest(op, params, data) + + // Run custom request initialization if present + if initRequest != nil { + initRequest(req) + } + + return req +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/neptune/waiters.go b/vendor/github.com/aws/aws-sdk-go/service/neptune/waiters.go new file mode 100644 index 000000000..acc80183c --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/neptune/waiters.go @@ -0,0 +1,152 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package neptune + +import ( + "time" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/request" +) + +// WaitUntilDBInstanceAvailable uses the Amazon Neptune API operation +// DescribeDBInstances to wait for a condition to be met before returning. +// If the condition is not met within the max attempt window, an error will +// be returned. +func (c *Neptune) WaitUntilDBInstanceAvailable(input *DescribeDBInstancesInput) error { + return c.WaitUntilDBInstanceAvailableWithContext(aws.BackgroundContext(), input) +} + +// WaitUntilDBInstanceAvailableWithContext is an extended version of WaitUntilDBInstanceAvailable. +// With the support for passing in a context and options to configure the +// Waiter and the underlying request options. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) WaitUntilDBInstanceAvailableWithContext(ctx aws.Context, input *DescribeDBInstancesInput, opts ...request.WaiterOption) error { + w := request.Waiter{ + Name: "WaitUntilDBInstanceAvailable", + MaxAttempts: 60, + Delay: request.ConstantWaiterDelay(30 * time.Second), + Acceptors: []request.WaiterAcceptor{ + { + State: request.SuccessWaiterState, + Matcher: request.PathAllWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "available", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "deleted", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "deleting", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "failed", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "incompatible-restore", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "incompatible-parameters", + }, + }, + Logger: c.Config.Logger, + NewRequest: func(opts []request.Option) (*request.Request, error) { + var inCpy *DescribeDBInstancesInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBInstancesRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + w.ApplyOptions(opts...) + + return w.WaitWithContext(ctx) +} + +// WaitUntilDBInstanceDeleted uses the Amazon Neptune API operation +// DescribeDBInstances to wait for a condition to be met before returning. +// If the condition is not met within the max attempt window, an error will +// be returned. +func (c *Neptune) WaitUntilDBInstanceDeleted(input *DescribeDBInstancesInput) error { + return c.WaitUntilDBInstanceDeletedWithContext(aws.BackgroundContext(), input) +} + +// WaitUntilDBInstanceDeletedWithContext is an extended version of WaitUntilDBInstanceDeleted. +// With the support for passing in a context and options to configure the +// Waiter and the underlying request options. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Neptune) WaitUntilDBInstanceDeletedWithContext(ctx aws.Context, input *DescribeDBInstancesInput, opts ...request.WaiterOption) error { + w := request.Waiter{ + Name: "WaitUntilDBInstanceDeleted", + MaxAttempts: 60, + Delay: request.ConstantWaiterDelay(30 * time.Second), + Acceptors: []request.WaiterAcceptor{ + { + State: request.SuccessWaiterState, + Matcher: request.PathAllWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "deleted", + }, + { + State: request.SuccessWaiterState, + Matcher: request.ErrorWaiterMatch, + Expected: "DBInstanceNotFound", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "creating", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "modifying", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "rebooting", + }, + { + State: request.FailureWaiterState, + Matcher: request.PathAnyWaiterMatch, Argument: "DBInstances[].DBInstanceStatus", + Expected: "resetting-master-credentials", + }, + }, + Logger: c.Config.Logger, + NewRequest: func(opts []request.Option) (*request.Request, error) { + var inCpy *DescribeDBInstancesInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.DescribeDBInstancesRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + w.ApplyOptions(opts...) + + return w.WaitWithContext(ctx) +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/opsworks/service.go b/vendor/github.com/aws/aws-sdk-go/service/opsworks/service.go index 3e94bb43d..a1a8307ca 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/opsworks/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/opsworks/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "opsworks" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "opsworks" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "OpsWorks" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the OpsWorks client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/opsworkscm/service.go b/vendor/github.com/aws/aws-sdk-go/service/opsworkscm/service.go index 2c10985ba..590398c63 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/opsworkscm/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/opsworkscm/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "opsworks-cm" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "opsworks-cm" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "OpsWorksCM" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the OpsWorksCM client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/organizations/service.go b/vendor/github.com/aws/aws-sdk-go/service/organizations/service.go index 0ca4f04f1..565c1715f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/organizations/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/organizations/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "organizations" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "organizations" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Organizations" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Organizations client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/pi/api.go b/vendor/github.com/aws/aws-sdk-go/service/pi/api.go new file mode 100644 index 000000000..4fd430a8e --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/pi/api.go @@ -0,0 +1,1088 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package pi + +import ( + "fmt" + "time" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/awsutil" + "github.com/aws/aws-sdk-go/aws/request" +) + +const opDescribeDimensionKeys = "DescribeDimensionKeys" + +// DescribeDimensionKeysRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDimensionKeys operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDimensionKeys for more information on using the DescribeDimensionKeys +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDimensionKeysRequest method. +// req, resp := client.DescribeDimensionKeysRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/pi-2018-02-27/DescribeDimensionKeys +func (c *PI) DescribeDimensionKeysRequest(input *DescribeDimensionKeysInput) (req *request.Request, output *DescribeDimensionKeysOutput) { + op := &request.Operation{ + Name: opDescribeDimensionKeys, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDimensionKeysInput{} + } + + output = &DescribeDimensionKeysOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDimensionKeys API operation for AWS Performance Insights. +// +// For a specific time period, retrieve the top N dimension keys for a metric. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Performance Insights's +// API operation DescribeDimensionKeys for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidArgumentException "InvalidArgumentException" +// One of the arguments provided is invalid for this request. +// +// * ErrCodeInternalServiceError "InternalServiceError" +// The request failed due to an unknown error. +// +// * ErrCodeNotAuthorizedException "NotAuthorizedException" +// The user is not authorized to perform this request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/pi-2018-02-27/DescribeDimensionKeys +func (c *PI) DescribeDimensionKeys(input *DescribeDimensionKeysInput) (*DescribeDimensionKeysOutput, error) { + req, out := c.DescribeDimensionKeysRequest(input) + return out, req.Send() +} + +// DescribeDimensionKeysWithContext is the same as DescribeDimensionKeys with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDimensionKeys for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *PI) DescribeDimensionKeysWithContext(ctx aws.Context, input *DescribeDimensionKeysInput, opts ...request.Option) (*DescribeDimensionKeysOutput, error) { + req, out := c.DescribeDimensionKeysRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opGetResourceMetrics = "GetResourceMetrics" + +// GetResourceMetricsRequest generates a "aws/request.Request" representing the +// client's request for the GetResourceMetrics operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See GetResourceMetrics for more information on using the GetResourceMetrics +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the GetResourceMetricsRequest method. +// req, resp := client.GetResourceMetricsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/pi-2018-02-27/GetResourceMetrics +func (c *PI) GetResourceMetricsRequest(input *GetResourceMetricsInput) (req *request.Request, output *GetResourceMetricsOutput) { + op := &request.Operation{ + Name: opGetResourceMetrics, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &GetResourceMetricsInput{} + } + + output = &GetResourceMetricsOutput{} + req = c.newRequest(op, input, output) + return +} + +// GetResourceMetrics API operation for AWS Performance Insights. +// +// Retrieve Performance Insights metrics for a set of data sources, over a time +// period. You can provide specific dimension groups and dimensions, and provide +// aggregation and filtering criteria for each group. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Performance Insights's +// API operation GetResourceMetrics for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidArgumentException "InvalidArgumentException" +// One of the arguments provided is invalid for this request. +// +// * ErrCodeInternalServiceError "InternalServiceError" +// The request failed due to an unknown error. +// +// * ErrCodeNotAuthorizedException "NotAuthorizedException" +// The user is not authorized to perform this request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/pi-2018-02-27/GetResourceMetrics +func (c *PI) GetResourceMetrics(input *GetResourceMetricsInput) (*GetResourceMetricsOutput, error) { + req, out := c.GetResourceMetricsRequest(input) + return out, req.Send() +} + +// GetResourceMetricsWithContext is the same as GetResourceMetrics with the addition of +// the ability to pass a context and additional request options. +// +// See GetResourceMetrics for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *PI) GetResourceMetricsWithContext(ctx aws.Context, input *GetResourceMetricsInput, opts ...request.Option) (*GetResourceMetricsOutput, error) { + req, out := c.GetResourceMetricsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// A timestamp, and a single numerical value, which together represent a measurement +// at a particular point in time. +type DataPoint struct { + _ struct{} `type:"structure"` + + // The time, in epoch format, associated with a particular Value. + // + // Timestamp is a required field + Timestamp *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` + + // The actual value associated with a particular Timestamp. + // + // Value is a required field + Value *float64 `type:"double" required:"true"` +} + +// String returns the string representation +func (s DataPoint) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DataPoint) GoString() string { + return s.String() +} + +// SetTimestamp sets the Timestamp field's value. +func (s *DataPoint) SetTimestamp(v time.Time) *DataPoint { + s.Timestamp = &v + return s +} + +// SetValue sets the Value field's value. +func (s *DataPoint) SetValue(v float64) *DataPoint { + s.Value = &v + return s +} + +type DescribeDimensionKeysInput struct { + _ struct{} `type:"structure"` + + // The date and time specifying the end of the requested time series data. The + // value specified is exclusive - data points less than (but not equal to) EndTime + // will be returned. + // + // The value for EndTime must be later than the value for StartTime. + // + // EndTime is a required field + EndTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` + + // One or more filters to apply in the request. Restrictions: + // + // * Any number of filters by the same dimension, as specified in the GroupBy + // or Partition parameters. + // + // * A single filter for any other dimension in this dimension group. + Filter map[string]*string `type:"map"` + + // A specification for how to aggregate the data points from a query result. + // You must specify a valid dimension group. Performance Insights will return + // all of the dimensions within that group, unless you provide the names of + // specific dimensions within that group. You can also request that Performance + // Insights return a limited number of values for a dimension. + // + // GroupBy is a required field + GroupBy *DimensionGroup `type:"structure" required:"true"` + + // An immutable, AWS Region-unique identifier for a data source. Performance + // Insights gathers metrics from this data source. + // + // To use an Amazon RDS instance as a data source, you specify its DbiResourceId + // value - for example: db-FAIHNTYBKTGAUSUZQYPDS2GW4A + // + // Identifier is a required field + Identifier *string `type:"string" required:"true"` + + // The maximum number of items to return in the response. If more items exist + // than the specified MaxRecords value, a pagination token is included in the + // response so that the remaining results can be retrieved. + MaxResults *int64 `type:"integer"` + + // The name of a Performance Insights metric to be measured. + // + // Valid values for Metric are: + // + // * db.load.avg - a scaled representation of the number of active sessions + // for the database engine. + // + // * db.sampledload.avg - the raw number of active sessions for the database + // engine. + // + // Metric is a required field + Metric *string `type:"string" required:"true"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the token, up to + // the value specified by MaxRecords. + NextToken *string `type:"string"` + + // For each dimension specified in GroupBy, specify a secondary dimension to + // further subdivide the partition keys in the response. + PartitionBy *DimensionGroup `type:"structure"` + + // The granularity, in seconds, of the data points returned from Performance + // Insights. A period can be as short as one second, or as long as one day (86400 + // seconds). Valid values are: + // + // * 1 (one second) + // + // * 60 (one minute) + // + // * 300 (five minutes) + // + // * 3600 (one hour) + // + // * 86400 (twenty-four hours) + // + // If you don't specify PeriodInSeconds, then Performance Insights will choose + // a value for you, with a goal of returning roughly 100-200 data points in + // the response. + PeriodInSeconds *int64 `type:"integer"` + + // The AWS service for which Performance Insights will return metrics. The only + // valid value for ServiceType is: RDS + // + // ServiceType is a required field + ServiceType *string `type:"string" required:"true" enum:"ServiceType"` + + // The date and time specifying the beginning of the requested time series data. + // You can't specify a StartTime that's earlier than 7 days ago. The value specified + // is inclusive - data points equal to or greater than StartTime will be returned. + // + // The value for StartTime must be earlier than the value for EndTime. + // + // StartTime is a required field + StartTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` +} + +// String returns the string representation +func (s DescribeDimensionKeysInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDimensionKeysInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeDimensionKeysInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeDimensionKeysInput"} + if s.EndTime == nil { + invalidParams.Add(request.NewErrParamRequired("EndTime")) + } + if s.GroupBy == nil { + invalidParams.Add(request.NewErrParamRequired("GroupBy")) + } + if s.Identifier == nil { + invalidParams.Add(request.NewErrParamRequired("Identifier")) + } + if s.Metric == nil { + invalidParams.Add(request.NewErrParamRequired("Metric")) + } + if s.ServiceType == nil { + invalidParams.Add(request.NewErrParamRequired("ServiceType")) + } + if s.StartTime == nil { + invalidParams.Add(request.NewErrParamRequired("StartTime")) + } + if s.GroupBy != nil { + if err := s.GroupBy.Validate(); err != nil { + invalidParams.AddNested("GroupBy", err.(request.ErrInvalidParams)) + } + } + if s.PartitionBy != nil { + if err := s.PartitionBy.Validate(); err != nil { + invalidParams.AddNested("PartitionBy", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetEndTime sets the EndTime field's value. +func (s *DescribeDimensionKeysInput) SetEndTime(v time.Time) *DescribeDimensionKeysInput { + s.EndTime = &v + return s +} + +// SetFilter sets the Filter field's value. +func (s *DescribeDimensionKeysInput) SetFilter(v map[string]*string) *DescribeDimensionKeysInput { + s.Filter = v + return s +} + +// SetGroupBy sets the GroupBy field's value. +func (s *DescribeDimensionKeysInput) SetGroupBy(v *DimensionGroup) *DescribeDimensionKeysInput { + s.GroupBy = v + return s +} + +// SetIdentifier sets the Identifier field's value. +func (s *DescribeDimensionKeysInput) SetIdentifier(v string) *DescribeDimensionKeysInput { + s.Identifier = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *DescribeDimensionKeysInput) SetMaxResults(v int64) *DescribeDimensionKeysInput { + s.MaxResults = &v + return s +} + +// SetMetric sets the Metric field's value. +func (s *DescribeDimensionKeysInput) SetMetric(v string) *DescribeDimensionKeysInput { + s.Metric = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *DescribeDimensionKeysInput) SetNextToken(v string) *DescribeDimensionKeysInput { + s.NextToken = &v + return s +} + +// SetPartitionBy sets the PartitionBy field's value. +func (s *DescribeDimensionKeysInput) SetPartitionBy(v *DimensionGroup) *DescribeDimensionKeysInput { + s.PartitionBy = v + return s +} + +// SetPeriodInSeconds sets the PeriodInSeconds field's value. +func (s *DescribeDimensionKeysInput) SetPeriodInSeconds(v int64) *DescribeDimensionKeysInput { + s.PeriodInSeconds = &v + return s +} + +// SetServiceType sets the ServiceType field's value. +func (s *DescribeDimensionKeysInput) SetServiceType(v string) *DescribeDimensionKeysInput { + s.ServiceType = &v + return s +} + +// SetStartTime sets the StartTime field's value. +func (s *DescribeDimensionKeysInput) SetStartTime(v time.Time) *DescribeDimensionKeysInput { + s.StartTime = &v + return s +} + +type DescribeDimensionKeysOutput struct { + _ struct{} `type:"structure"` + + // The end time for the returned dimension keys, after alignment to a granular + // boundary (as specified by PeriodInSeconds). AlignedEndTime will be greater + // than or equal to the value of the user-specified Endtime. + AlignedEndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The start time for the returned dimension keys, after alignment to a granular + // boundary (as specified by PeriodInSeconds). AlignedStartTime will be less + // than or equal to the value of the user-specified StartTime. + AlignedStartTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The dimension keys that were requested. + Keys []*DimensionKeyDescription `type:"list"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the token, up to + // the value specified by MaxRecords. + NextToken *string `type:"string"` + + // If PartitionBy was present in the request, PartitionKeys contains the breakdown + // of dimension keys by the specified partitions. + PartitionKeys []*ResponsePartitionKey `type:"list"` +} + +// String returns the string representation +func (s DescribeDimensionKeysOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDimensionKeysOutput) GoString() string { + return s.String() +} + +// SetAlignedEndTime sets the AlignedEndTime field's value. +func (s *DescribeDimensionKeysOutput) SetAlignedEndTime(v time.Time) *DescribeDimensionKeysOutput { + s.AlignedEndTime = &v + return s +} + +// SetAlignedStartTime sets the AlignedStartTime field's value. +func (s *DescribeDimensionKeysOutput) SetAlignedStartTime(v time.Time) *DescribeDimensionKeysOutput { + s.AlignedStartTime = &v + return s +} + +// SetKeys sets the Keys field's value. +func (s *DescribeDimensionKeysOutput) SetKeys(v []*DimensionKeyDescription) *DescribeDimensionKeysOutput { + s.Keys = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *DescribeDimensionKeysOutput) SetNextToken(v string) *DescribeDimensionKeysOutput { + s.NextToken = &v + return s +} + +// SetPartitionKeys sets the PartitionKeys field's value. +func (s *DescribeDimensionKeysOutput) SetPartitionKeys(v []*ResponsePartitionKey) *DescribeDimensionKeysOutput { + s.PartitionKeys = v + return s +} + +// A logical grouping of Performance Insights metrics for a related subject +// area. For example, the db.sql dimension group consists of the following dimensions: +// db.sql.id, db.sql.db_id, db.sql.statement, and db.sql.tokenized_id. +type DimensionGroup struct { + _ struct{} `type:"structure"` + + // A list of specific dimensions from a dimension group. If this parameter is + // not present, then it signifies that all of the dimensions in the group were + // requested, or are present in the response. + // + // Valid values for elements in the Dimensions array are: + // + // * db.user.id + // + // * db.user.name + // + // * db.host.id + // + // * db.host.name + // + // * db.sql.id + // + // * db.sql.db_id + // + // * db.sql.statement + // + // * db.sql.tokenized_id + // + // * db.sql_tokenized.id + // + // * db.sql_tokenized.db_id + // + // * db.sql_tokenized.statement + // + // * db.wait_event.name + // + // * db.wait_event.type + // + // * db.wait_event_type.name + Dimensions []*string `min:"1" type:"list"` + + // The name of the dimension group. Valid values are: + // + // * db.user + // + // * db.host + // + // * db.sql + // + // * db.sql_tokenized + // + // * db.wait_event + // + // * db.wait_event_type + // + // Group is a required field + Group *string `type:"string" required:"true"` + + // The maximum number of items to fetch for this dimension group. + Limit *int64 `min:"1" type:"integer"` +} + +// String returns the string representation +func (s DimensionGroup) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DimensionGroup) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DimensionGroup) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DimensionGroup"} + if s.Dimensions != nil && len(s.Dimensions) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Dimensions", 1)) + } + if s.Group == nil { + invalidParams.Add(request.NewErrParamRequired("Group")) + } + if s.Limit != nil && *s.Limit < 1 { + invalidParams.Add(request.NewErrParamMinValue("Limit", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDimensions sets the Dimensions field's value. +func (s *DimensionGroup) SetDimensions(v []*string) *DimensionGroup { + s.Dimensions = v + return s +} + +// SetGroup sets the Group field's value. +func (s *DimensionGroup) SetGroup(v string) *DimensionGroup { + s.Group = &v + return s +} + +// SetLimit sets the Limit field's value. +func (s *DimensionGroup) SetLimit(v int64) *DimensionGroup { + s.Limit = &v + return s +} + +// An array of descriptions and aggregated values for each dimension within +// a dimension group. +type DimensionKeyDescription struct { + _ struct{} `type:"structure"` + + // A map of name-value pairs for the dimensions in the group. + Dimensions map[string]*string `type:"map"` + + // If PartitionBy was specified, PartitionKeys contains the dimensions that + // were. + Partitions []*float64 `type:"list"` + + // The aggregated metric value for the dimension(s), over the requested time + // range. + Total *float64 `type:"double"` +} + +// String returns the string representation +func (s DimensionKeyDescription) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DimensionKeyDescription) GoString() string { + return s.String() +} + +// SetDimensions sets the Dimensions field's value. +func (s *DimensionKeyDescription) SetDimensions(v map[string]*string) *DimensionKeyDescription { + s.Dimensions = v + return s +} + +// SetPartitions sets the Partitions field's value. +func (s *DimensionKeyDescription) SetPartitions(v []*float64) *DimensionKeyDescription { + s.Partitions = v + return s +} + +// SetTotal sets the Total field's value. +func (s *DimensionKeyDescription) SetTotal(v float64) *DimensionKeyDescription { + s.Total = &v + return s +} + +type GetResourceMetricsInput struct { + _ struct{} `type:"structure"` + + // The date and time specifiying the end of the requested time series data. + // The value specified is exclusive - data points less than (but not equal to) + // EndTime will be returned. + // + // The value for EndTime must be later than the value for StartTime. + // + // EndTime is a required field + EndTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` + + // An immutable, AWS Region-unique identifier for a data source. Performance + // Insights gathers metrics from this data source. + // + // To use an Amazon RDS instance as a data source, you specify its DbiResourceId + // value - for example: db-FAIHNTYBKTGAUSUZQYPDS2GW4A + // + // Identifier is a required field + Identifier *string `type:"string" required:"true"` + + // The maximum number of items to return in the response. If more items exist + // than the specified MaxRecords value, a pagination token is included in the + // response so that the remaining results can be retrieved. + MaxResults *int64 `type:"integer"` + + // An array of one or more queries to perform. Each query must specify a Performance + // Insights metric, and can optionally specify aggregation and filtering criteria. + // + // MetricQueries is a required field + MetricQueries []*MetricQuery `min:"1" type:"list" required:"true"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the token, up to + // the value specified by MaxRecords. + NextToken *string `type:"string"` + + // The granularity, in seconds, of the data points returned from Performance + // Insights. A period can be as short as one second, or as long as one day (86400 + // seconds). Valid values are: + // + // * 1 (one second) + // + // * 60 (one minute) + // + // * 300 (five minutes) + // + // * 3600 (one hour) + // + // * 86400 (twenty-four hours) + // + // If you don't specify PeriodInSeconds, then Performance Insights will choose + // a value for you, with a goal of returning roughly 100-200 data points in + // the response. + PeriodInSeconds *int64 `type:"integer"` + + // The AWS service for which Performance Insights will return metrics. The only + // valid value for ServiceType is: RDS + // + // ServiceType is a required field + ServiceType *string `type:"string" required:"true" enum:"ServiceType"` + + // The date and time specifying the beginning of the requested time series data. + // You can't specify a StartTime that's earlier than 7 days ago. The value specified + // is inclusive - data points equal to or greater than StartTime will be returned. + // + // The value for StartTime must be earlier than the value for EndTime. + // + // StartTime is a required field + StartTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` +} + +// String returns the string representation +func (s GetResourceMetricsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetResourceMetricsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *GetResourceMetricsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "GetResourceMetricsInput"} + if s.EndTime == nil { + invalidParams.Add(request.NewErrParamRequired("EndTime")) + } + if s.Identifier == nil { + invalidParams.Add(request.NewErrParamRequired("Identifier")) + } + if s.MetricQueries == nil { + invalidParams.Add(request.NewErrParamRequired("MetricQueries")) + } + if s.MetricQueries != nil && len(s.MetricQueries) < 1 { + invalidParams.Add(request.NewErrParamMinLen("MetricQueries", 1)) + } + if s.ServiceType == nil { + invalidParams.Add(request.NewErrParamRequired("ServiceType")) + } + if s.StartTime == nil { + invalidParams.Add(request.NewErrParamRequired("StartTime")) + } + if s.MetricQueries != nil { + for i, v := range s.MetricQueries { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "MetricQueries", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetEndTime sets the EndTime field's value. +func (s *GetResourceMetricsInput) SetEndTime(v time.Time) *GetResourceMetricsInput { + s.EndTime = &v + return s +} + +// SetIdentifier sets the Identifier field's value. +func (s *GetResourceMetricsInput) SetIdentifier(v string) *GetResourceMetricsInput { + s.Identifier = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *GetResourceMetricsInput) SetMaxResults(v int64) *GetResourceMetricsInput { + s.MaxResults = &v + return s +} + +// SetMetricQueries sets the MetricQueries field's value. +func (s *GetResourceMetricsInput) SetMetricQueries(v []*MetricQuery) *GetResourceMetricsInput { + s.MetricQueries = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *GetResourceMetricsInput) SetNextToken(v string) *GetResourceMetricsInput { + s.NextToken = &v + return s +} + +// SetPeriodInSeconds sets the PeriodInSeconds field's value. +func (s *GetResourceMetricsInput) SetPeriodInSeconds(v int64) *GetResourceMetricsInput { + s.PeriodInSeconds = &v + return s +} + +// SetServiceType sets the ServiceType field's value. +func (s *GetResourceMetricsInput) SetServiceType(v string) *GetResourceMetricsInput { + s.ServiceType = &v + return s +} + +// SetStartTime sets the StartTime field's value. +func (s *GetResourceMetricsInput) SetStartTime(v time.Time) *GetResourceMetricsInput { + s.StartTime = &v + return s +} + +type GetResourceMetricsOutput struct { + _ struct{} `type:"structure"` + + // The end time for the returned metrics, after alignment to a granular boundary + // (as specified by PeriodInSeconds). AlignedEndTime will be greater than or + // equal to the value of the user-specified Endtime. + AlignedEndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The start time for the returned metrics, after alignment to a granular boundary + // (as specified by PeriodInSeconds). AlignedStartTime will be less than or + // equal to the value of the user-specified StartTime. + AlignedStartTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // An immutable, AWS Region-unique identifier for a data source. Performance + // Insights gathers metrics from this data source. + // + // To use an Amazon RDS instance as a data source, you specify its DbiResourceId + // value - for example: db-FAIHNTYBKTGAUSUZQYPDS2GW4A + Identifier *string `type:"string"` + + // An array of metric results,, where each array element contains all of the + // data points for a particular dimension. + MetricList []*MetricKeyDataPoints `type:"list"` + + // An optional pagination token provided by a previous request. If this parameter + // is specified, the response includes only records beyond the token, up to + // the value specified by MaxRecords. + NextToken *string `type:"string"` +} + +// String returns the string representation +func (s GetResourceMetricsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetResourceMetricsOutput) GoString() string { + return s.String() +} + +// SetAlignedEndTime sets the AlignedEndTime field's value. +func (s *GetResourceMetricsOutput) SetAlignedEndTime(v time.Time) *GetResourceMetricsOutput { + s.AlignedEndTime = &v + return s +} + +// SetAlignedStartTime sets the AlignedStartTime field's value. +func (s *GetResourceMetricsOutput) SetAlignedStartTime(v time.Time) *GetResourceMetricsOutput { + s.AlignedStartTime = &v + return s +} + +// SetIdentifier sets the Identifier field's value. +func (s *GetResourceMetricsOutput) SetIdentifier(v string) *GetResourceMetricsOutput { + s.Identifier = &v + return s +} + +// SetMetricList sets the MetricList field's value. +func (s *GetResourceMetricsOutput) SetMetricList(v []*MetricKeyDataPoints) *GetResourceMetricsOutput { + s.MetricList = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *GetResourceMetricsOutput) SetNextToken(v string) *GetResourceMetricsOutput { + s.NextToken = &v + return s +} + +// A time-ordered series of data points, correpsonding to a dimension of a Performance +// Insights metric. +type MetricKeyDataPoints struct { + _ struct{} `type:"structure"` + + // An array of timestamp-value pairs, representing measurements over a period + // of time. + DataPoints []*DataPoint `type:"list"` + + // The dimension(s) to which the data points apply. + Key *ResponseResourceMetricKey `type:"structure"` +} + +// String returns the string representation +func (s MetricKeyDataPoints) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s MetricKeyDataPoints) GoString() string { + return s.String() +} + +// SetDataPoints sets the DataPoints field's value. +func (s *MetricKeyDataPoints) SetDataPoints(v []*DataPoint) *MetricKeyDataPoints { + s.DataPoints = v + return s +} + +// SetKey sets the Key field's value. +func (s *MetricKeyDataPoints) SetKey(v *ResponseResourceMetricKey) *MetricKeyDataPoints { + s.Key = v + return s +} + +// A single query to be processed. You must provide the metric to query. If +// no other parameters are specified, Performance Insights returns all of the +// data points for that metric. You can optionally request that the data points +// be aggregated by dimension group ( GroupBy), and return only those data points +// that match your criteria (Filter). +type MetricQuery struct { + _ struct{} `type:"structure"` + + // One or more filters to apply in the request. Restrictions: + // + // * Any number of filters by the same dimension, as specified in the GroupBy + // parameter. + // + // * A single filter for any other dimension in this dimension group. + Filter map[string]*string `type:"map"` + + // A specification for how to aggregate the data points from a query result. + // You must specify a valid dimension group. Performance Insights will return + // all of the dimensions within that group, unless you provide the names of + // specific dimensions within that group. You can also request that Performance + // Insights return a limited number of values for a dimension. + GroupBy *DimensionGroup `type:"structure"` + + // The name of a Performance Insights metric to be measured. + // + // Valid values for Metric are: + // + // * db.load.avg - a scaled representation of the number of active sessions + // for the database engine. + // + // * db.sampledload.avg - the raw number of active sessions for the database + // engine. + // + // Metric is a required field + Metric *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s MetricQuery) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s MetricQuery) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *MetricQuery) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "MetricQuery"} + if s.Metric == nil { + invalidParams.Add(request.NewErrParamRequired("Metric")) + } + if s.GroupBy != nil { + if err := s.GroupBy.Validate(); err != nil { + invalidParams.AddNested("GroupBy", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFilter sets the Filter field's value. +func (s *MetricQuery) SetFilter(v map[string]*string) *MetricQuery { + s.Filter = v + return s +} + +// SetGroupBy sets the GroupBy field's value. +func (s *MetricQuery) SetGroupBy(v *DimensionGroup) *MetricQuery { + s.GroupBy = v + return s +} + +// SetMetric sets the Metric field's value. +func (s *MetricQuery) SetMetric(v string) *MetricQuery { + s.Metric = &v + return s +} + +// If PartitionBy was specified in a DescribeDimensionKeys request, the dimensions +// are returned in an array. Each element in the array specifies one dimension. +type ResponsePartitionKey struct { + _ struct{} `type:"structure"` + + // A dimension map that contains the dimension(s) for this partition. + // + // Dimensions is a required field + Dimensions map[string]*string `type:"map" required:"true"` +} + +// String returns the string representation +func (s ResponsePartitionKey) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResponsePartitionKey) GoString() string { + return s.String() +} + +// SetDimensions sets the Dimensions field's value. +func (s *ResponsePartitionKey) SetDimensions(v map[string]*string) *ResponsePartitionKey { + s.Dimensions = v + return s +} + +// An object describing a Performance Insights metric and one or more dimensions +// for that metric. +type ResponseResourceMetricKey struct { + _ struct{} `type:"structure"` + + // The valid dimensions for the metric. + Dimensions map[string]*string `type:"map"` + + // The name of a Performance Insights metric to be measured. + // + // Valid values for Metric are: + // + // * db.load.avg - a scaled representation of the number of active sessions + // for the database engine. + // + // * db.sampledload.avg - the raw number of active sessions for the database + // engine. + // + // Metric is a required field + Metric *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s ResponseResourceMetricKey) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResponseResourceMetricKey) GoString() string { + return s.String() +} + +// SetDimensions sets the Dimensions field's value. +func (s *ResponseResourceMetricKey) SetDimensions(v map[string]*string) *ResponseResourceMetricKey { + s.Dimensions = v + return s +} + +// SetMetric sets the Metric field's value. +func (s *ResponseResourceMetricKey) SetMetric(v string) *ResponseResourceMetricKey { + s.Metric = &v + return s +} + +const ( + // ServiceTypeRds is a ServiceType enum value + ServiceTypeRds = "RDS" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/pi/doc.go b/vendor/github.com/aws/aws-sdk-go/service/pi/doc.go new file mode 100644 index 000000000..45015fc5c --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/pi/doc.go @@ -0,0 +1,44 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package pi provides the client and types for making API +// requests to AWS Performance Insights. +// +// AWS Performance Insights enables you to monitor and explore different dimensions +// of database load based on data captured from a running RDS instance. The +// guide provides detailed information about Performance Insights data types, +// parameters and errors. For more information about Performance Insights capabilities +// see Using Amazon RDS Performance Insights (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_PerfInsights.html) +// in the Amazon RDS User Guide. +// +// The AWS Performance Insights API provides visibility into the performance +// of your RDS instance, when Performance Insights is enabled for supported +// engine types. While Amazon CloudWatch provides the authoritative source for +// AWS service vended monitoring metrics, AWS Performance Insights offers a +// domain-specific view of database load measured as Average Active Sessions +// and provided to API consumers as a 2-dimensional time-series dataset. The +// time dimension of the data provides DB load data for each time point in the +// queried time range, and each time point decomposes overall load in relation +// to the requested dimensions, such as SQL, Wait-event, User or Host, measured +// at that time point. +// +// See https://docs.aws.amazon.com/goto/WebAPI/pi-2018-02-27 for more information on this service. +// +// See pi package documentation for more information. +// https://docs.aws.amazon.com/sdk-for-go/api/service/pi/ +// +// Using the Client +// +// To contact AWS Performance Insights with the SDK use the New function to create +// a new service client. With that client you can make API requests to the service. +// These clients are safe to use concurrently. +// +// See the SDK's documentation for more information on how to use the SDK. +// https://docs.aws.amazon.com/sdk-for-go/api/ +// +// See aws.Config documentation for more information on configuring SDK clients. +// https://docs.aws.amazon.com/sdk-for-go/api/aws/#Config +// +// See the AWS Performance Insights client PI for more +// information on creating client for this service. +// https://docs.aws.amazon.com/sdk-for-go/api/service/pi/#New +package pi diff --git a/vendor/github.com/aws/aws-sdk-go/service/pi/errors.go b/vendor/github.com/aws/aws-sdk-go/service/pi/errors.go new file mode 100644 index 000000000..869b1ed98 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/pi/errors.go @@ -0,0 +1,24 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package pi + +const ( + + // ErrCodeInternalServiceError for service response error code + // "InternalServiceError". + // + // The request failed due to an unknown error. + ErrCodeInternalServiceError = "InternalServiceError" + + // ErrCodeInvalidArgumentException for service response error code + // "InvalidArgumentException". + // + // One of the arguments provided is invalid for this request. + ErrCodeInvalidArgumentException = "InvalidArgumentException" + + // ErrCodeNotAuthorizedException for service response error code + // "NotAuthorizedException". + // + // The user is not authorized to perform this request. + ErrCodeNotAuthorizedException = "NotAuthorizedException" +) diff --git a/vendor/github.com/aws/aws-sdk-go/service/pi/piiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/pi/piiface/interface.go new file mode 100644 index 000000000..4694bef94 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/pi/piiface/interface.go @@ -0,0 +1,72 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +// Package piiface provides an interface to enable mocking the AWS Performance Insights service client +// for testing your code. +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. +package piiface + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/service/pi" +) + +// PIAPI provides an interface to enable mocking the +// pi.PI service client's API operation, +// paginators, and waiters. This make unit testing your code that calls out +// to the SDK's service client's calls easier. +// +// The best way to use this interface is so the SDK's service client's calls +// can be stubbed out for unit testing your code with the SDK without needing +// to inject custom request handlers into the SDK's request pipeline. +// +// // myFunc uses an SDK service client to make a request to +// // AWS Performance Insights. +// func myFunc(svc piiface.PIAPI) bool { +// // Make svc.DescribeDimensionKeys request +// } +// +// func main() { +// sess := session.New() +// svc := pi.New(sess) +// +// myFunc(svc) +// } +// +// In your _test.go file: +// +// // Define a mock struct to be used in your unit tests of myFunc. +// type mockPIClient struct { +// piiface.PIAPI +// } +// func (m *mockPIClient) DescribeDimensionKeys(input *pi.DescribeDimensionKeysInput) (*pi.DescribeDimensionKeysOutput, error) { +// // mock response/functionality +// } +// +// func TestMyFunc(t *testing.T) { +// // Setup Test +// mockSvc := &mockPIClient{} +// +// myfunc(mockSvc) +// +// // Verify myFunc's functionality +// } +// +// It is important to note that this interface will have breaking changes +// when the service model is updated and adds new API operations, paginators, +// and waiters. Its suggested to use the pattern above for testing, or using +// tooling to generate mocks to satisfy the interfaces. +type PIAPI interface { + DescribeDimensionKeys(*pi.DescribeDimensionKeysInput) (*pi.DescribeDimensionKeysOutput, error) + DescribeDimensionKeysWithContext(aws.Context, *pi.DescribeDimensionKeysInput, ...request.Option) (*pi.DescribeDimensionKeysOutput, error) + DescribeDimensionKeysRequest(*pi.DescribeDimensionKeysInput) (*request.Request, *pi.DescribeDimensionKeysOutput) + + GetResourceMetrics(*pi.GetResourceMetricsInput) (*pi.GetResourceMetricsOutput, error) + GetResourceMetricsWithContext(aws.Context, *pi.GetResourceMetricsInput, ...request.Option) (*pi.GetResourceMetricsOutput, error) + GetResourceMetricsRequest(*pi.GetResourceMetricsInput) (*request.Request, *pi.GetResourceMetricsOutput) +} + +var _ PIAPI = (*pi.PI)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/pi/service.go b/vendor/github.com/aws/aws-sdk-go/service/pi/service.go new file mode 100644 index 000000000..968a9f854 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/pi/service.go @@ -0,0 +1,100 @@ +// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. + +package pi + +import ( + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/client" + "github.com/aws/aws-sdk-go/aws/client/metadata" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/signer/v4" + "github.com/aws/aws-sdk-go/private/protocol/jsonrpc" +) + +// PI provides the API operation methods for making requests to +// AWS Performance Insights. See this package's package overview docs +// for details on the service. +// +// PI methods are safe to use concurrently. It is not safe to +// modify mutate any of the struct's properties though. +type PI struct { + *client.Client +} + +// Used for custom client initialization logic +var initClient func(*client.Client) + +// Used for custom request initialization logic +var initRequest func(*request.Request) + +// Service information constants +const ( + ServiceName = "pi" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "PI" // ServiceID is a unique identifer of a specific service. +) + +// New creates a new instance of the PI client with a session. +// If additional configuration is needed for the client instance use the optional +// aws.Config parameter to add your extra config. +// +// Example: +// // Create a PI client from just a session. +// svc := pi.New(mySession) +// +// // Create a PI client with additional configuration +// svc := pi.New(mySession, aws.NewConfig().WithRegion("us-west-2")) +func New(p client.ConfigProvider, cfgs ...*aws.Config) *PI { + c := p.ClientConfig(EndpointsID, cfgs...) + if c.SigningNameDerived || len(c.SigningName) == 0 { + c.SigningName = "pi" + } + return newClient(*c.Config, c.Handlers, c.Endpoint, c.SigningRegion, c.SigningName) +} + +// newClient creates, initializes and returns a new service client instance. +func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion, signingName string) *PI { + svc := &PI{ + Client: client.New( + cfg, + metadata.ClientInfo{ + ServiceName: ServiceName, + ServiceID: ServiceID, + SigningName: signingName, + SigningRegion: signingRegion, + Endpoint: endpoint, + APIVersion: "2018-02-27", + JSONVersion: "1.1", + TargetPrefix: "PerformanceInsightsv20180227", + }, + handlers, + ), + } + + // Handlers + svc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler) + svc.Handlers.Build.PushBackNamed(jsonrpc.BuildHandler) + svc.Handlers.Unmarshal.PushBackNamed(jsonrpc.UnmarshalHandler) + svc.Handlers.UnmarshalMeta.PushBackNamed(jsonrpc.UnmarshalMetaHandler) + svc.Handlers.UnmarshalError.PushBackNamed(jsonrpc.UnmarshalErrorHandler) + + // Run custom client initialization if present + if initClient != nil { + initClient(svc.Client) + } + + return svc +} + +// newRequest creates a new request for a PI operation and runs any +// custom request initialization. +func (c *PI) newRequest(op *request.Operation, params, data interface{}) *request.Request { + req := c.NewRequest(op, params, data) + + // Run custom request initialization if present + if initRequest != nil { + initRequest(req) + } + + return req +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/pinpoint/service.go b/vendor/github.com/aws/aws-sdk-go/service/pinpoint/service.go index 34533af39..9bee611da 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/pinpoint/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/pinpoint/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "pinpoint" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "pinpoint" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Pinpoint" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Pinpoint client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/polly/api.go b/vendor/github.com/aws/aws-sdk-go/service/polly/api.go index 020d42910..ed69a69f8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/polly/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/polly/api.go @@ -538,8 +538,8 @@ func (c *Polly) SynthesizeSpeechRequest(input *SynthesizeSpeechInput) (req *requ // Returned Error Codes: // * ErrCodeTextLengthExceededException "TextLengthExceededException" // The value of the "Text" parameter is longer than the accepted limits. The -// limit for input text is a maximum of 3000 characters total, of which no more -// than 1500 can be billed characters. SSML tags are not counted as billed characters. +// limit for input text is a maximum of 6000 characters total, of which no more +// than 3000 can be billed characters. SSML tags are not counted as billed characters. // // * ErrCodeInvalidSampleRateException "InvalidSampleRateException" // The specified sample rate is not valid. @@ -1477,6 +1477,9 @@ const ( // VoiceIdCeline is a VoiceId enum value VoiceIdCeline = "Celine" + // VoiceIdLea is a VoiceId enum value + VoiceIdLea = "Lea" + // VoiceIdMathieu is a VoiceId enum value VoiceIdMathieu = "Mathieu" diff --git a/vendor/github.com/aws/aws-sdk-go/service/polly/errors.go b/vendor/github.com/aws/aws-sdk-go/service/polly/errors.go index 9a53fb6b0..e2cccb782 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/polly/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/polly/errors.go @@ -83,8 +83,8 @@ const ( // "TextLengthExceededException". // // The value of the "Text" parameter is longer than the accepted limits. The - // limit for input text is a maximum of 3000 characters total, of which no more - // than 1500 can be billed characters. SSML tags are not counted as billed characters. + // limit for input text is a maximum of 6000 characters total, of which no more + // than 3000 can be billed characters. SSML tags are not counted as billed characters. ErrCodeTextLengthExceededException = "TextLengthExceededException" // ErrCodeUnsupportedPlsAlphabetException for service response error code diff --git a/vendor/github.com/aws/aws-sdk-go/service/polly/service.go b/vendor/github.com/aws/aws-sdk-go/service/polly/service.go index 0e7b93f48..fa9fa6ec9 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/polly/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/polly/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "polly" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "polly" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Polly" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Polly client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/pricing/service.go b/vendor/github.com/aws/aws-sdk-go/service/pricing/service.go index b4b8b815a..90ff33d0a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/pricing/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/pricing/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "api.pricing" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "api.pricing" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Pricing" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Pricing client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/api.go b/vendor/github.com/aws/aws-sdk-go/service/rds/api.go index 66002f1a8..51b8dcf6b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rds/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/api.go @@ -1838,6 +1838,9 @@ func (c *RDS) CreateDBSecurityGroupRequest(input *CreateDBSecurityGroupInput) (r // Creates a new DB security group. DB security groups control access to a DB // instance. // +// A DB security group controls access to EC2-Classic DB instances that are +// not in a VPC. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -8858,6 +8861,10 @@ func (c *RDS) RestoreDBClusterFromSnapshotRequest(input *RestoreDBClusterFromSna // * ErrCodeKMSKeyNotAccessibleFault "KMSKeyNotAccessibleFault" // An error occurred accessing an AWS KMS key. // +// * ErrCodeDBClusterParameterGroupNotFoundFault "DBClusterParameterGroupNotFound" +// DBClusterParameterGroupName doesn't refer to an existing DB cluster parameter +// group. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/rds-2014-10-31/RestoreDBClusterFromSnapshot func (c *RDS) RestoreDBClusterFromSnapshot(input *RestoreDBClusterFromSnapshotInput) (*RestoreDBClusterFromSnapshotOutput, error) { req, out := c.RestoreDBClusterFromSnapshotRequest(input) @@ -9002,6 +9009,10 @@ func (c *RDS) RestoreDBClusterToPointInTimeRequest(input *RestoreDBClusterToPoin // The request would result in the user exceeding the allowed amount of storage // available across all DB instances. // +// * ErrCodeDBClusterParameterGroupNotFoundFault "DBClusterParameterGroupNotFound" +// DBClusterParameterGroupName doesn't refer to an existing DB cluster parameter +// group. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/rds-2014-10-31/RestoreDBClusterToPointInTime func (c *RDS) RestoreDBClusterToPointInTime(input *RestoreDBClusterToPointInTimeInput) (*RestoreDBClusterToPointInTimeOutput, error) { req, out := c.RestoreDBClusterToPointInTimeRequest(input) @@ -9162,6 +9173,9 @@ func (c *RDS) RestoreDBInstanceFromDBSnapshotRequest(input *RestoreDBInstanceFro // * ErrCodeDomainNotFoundFault "DomainNotFoundFault" // Domain doesn't refer to an existing Active Directory domain. // +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName doesn't refer to an existing DB parameter group. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/rds-2014-10-31/RestoreDBInstanceFromDBSnapshot func (c *RDS) RestoreDBInstanceFromDBSnapshot(input *RestoreDBInstanceFromDBSnapshotInput) (*RestoreDBInstanceFromDBSnapshotOutput, error) { req, out := c.RestoreDBInstanceFromDBSnapshotRequest(input) @@ -9374,7 +9388,7 @@ func (c *RDS) RestoreDBInstanceToPointInTimeRequest(input *RestoreDBInstanceToPo // the BackupRetentionPeriod property. // // The target database is created with most of the original configuration, but -// in a system-selected availability zone, with the default security group, +// in a system-selected Availability Zone, with the default security group, // the default subnet group, and the default DB parameter group. By default, // the new DB instance is created as a single-AZ deployment except when the // instance is a SQL Server instance that has an option group that is associated @@ -9459,6 +9473,9 @@ func (c *RDS) RestoreDBInstanceToPointInTimeRequest(input *RestoreDBInstanceToPo // * ErrCodeDomainNotFoundFault "DomainNotFoundFault" // Domain doesn't refer to an existing Active Directory domain. // +// * ErrCodeDBParameterGroupNotFoundFault "DBParameterGroupNotFound" +// DBParameterGroupName doesn't refer to an existing DB parameter group. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/rds-2014-10-31/RestoreDBInstanceToPointInTime func (c *RDS) RestoreDBInstanceToPointInTime(input *RestoreDBInstanceToPointInTimeInput) (*RestoreDBInstanceToPointInTimeOutput, error) { req, out := c.RestoreDBInstanceToPointInTimeRequest(input) @@ -10279,7 +10296,7 @@ func (s *AuthorizeDBSecurityGroupIngressOutput) SetDBSecurityGroup(v *DBSecurity type AvailabilityZone struct { _ struct{} `type:"structure"` - // The name of the availability zone. + // The name of the Availability Zone. Name *string `type:"string"` } @@ -10299,6 +10316,53 @@ func (s *AvailabilityZone) SetName(v string) *AvailabilityZone { return s } +// Contains the available processor feature information for the DB instance +// class of a DB instance. +// +// For more information, see Configuring the Processor of the DB Instance Class +// (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Concepts.DBInstanceClass.html#USER_ConfigureProcessor) +// in the Amazon RDS User Guide. +type AvailableProcessorFeature struct { + _ struct{} `type:"structure"` + + // The allowed values for the processor feature of the DB instance class. + AllowedValues *string `type:"string"` + + // The default value for the processor feature of the DB instance class. + DefaultValue *string `type:"string"` + + // The name of the processor feature. Valid names are coreCount and threadsPerCore. + Name *string `type:"string"` +} + +// String returns the string representation +func (s AvailableProcessorFeature) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AvailableProcessorFeature) GoString() string { + return s.String() +} + +// SetAllowedValues sets the AllowedValues field's value. +func (s *AvailableProcessorFeature) SetAllowedValues(v string) *AvailableProcessorFeature { + s.AllowedValues = &v + return s +} + +// SetDefaultValue sets the DefaultValue field's value. +func (s *AvailableProcessorFeature) SetDefaultValue(v string) *AvailableProcessorFeature { + s.DefaultValue = &v + return s +} + +// SetName sets the Name field's value. +func (s *AvailableProcessorFeature) SetName(v string) *AvailableProcessorFeature { + s.Name = &v + return s +} + type BacktrackDBClusterInput struct { _ struct{} `type:"structure"` @@ -11506,6 +11570,10 @@ type CreateDBClusterInput struct { // DestinationRegion is used for presigning the request to a given region. DestinationRegion *string `type:"string"` + // The list of log types that need to be enabled for exporting to CloudWatch + // Logs. + EnableCloudwatchLogsExports []*string `type:"list"` + // True to enable mapping of AWS Identity and Access Management (IAM) accounts // to database accounts, and otherwise false. // @@ -11749,6 +11817,12 @@ func (s *CreateDBClusterInput) SetDestinationRegion(v string) *CreateDBClusterIn return s } +// SetEnableCloudwatchLogsExports sets the EnableCloudwatchLogsExports field's value. +func (s *CreateDBClusterInput) SetEnableCloudwatchLogsExports(v []*string) *CreateDBClusterInput { + s.EnableCloudwatchLogsExports = v + return s +} + // SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. func (s *CreateDBClusterInput) SetEnableIAMDatabaseAuthentication(v bool) *CreateDBClusterInput { s.EnableIAMDatabaseAuthentication = &v @@ -12376,6 +12450,9 @@ type CreateDBInstanceInput struct { EnableIAMDatabaseAuthentication *bool `type:"boolean"` // True to enable Performance Insights for the DB instance, and otherwise false. + // + // For more information, see Using Amazon Performance Insights (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_PerfInsights.html) + // in the Amazon Relational Database Service User Guide. EnablePerformanceInsights *bool `type:"boolean"` // The name of the database engine to be used for this instance. @@ -12417,9 +12494,11 @@ type CreateDBInstanceInput struct { // The version number of the database engine to use. // - // The following are the database engines and major and minor versions that - // are available with Amazon RDS. Not every database engine is available for - // every AWS Region. + // For a list of valid engine versions, call DescribeDBEngineVersions. + // + // The following are the database engines and links to information about the + // major and minor versions that are available with Amazon RDS. Not every database + // engine is available for every AWS Region. // // Amazon Aurora // @@ -12428,98 +12507,28 @@ type CreateDBInstanceInput struct { // // MariaDB // - // * 10.2.12 (supported in all AWS Regions) + // See MariaDB on Amazon RDS Versions (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_MariaDB.html#MariaDB.Concepts.VersionMgmt) + // in the Amazon RDS User Guide. // - // * 10.2.11 (supported in all AWS Regions) + // Microsoft SQL Server // - // 10.1.31 (supported in all AWS Regions) - // - // * 10.1.26 (supported in all AWS Regions) - // - // * 10.1.23 (supported in all AWS Regions) - // - // * 10.1.19 (supported in all AWS Regions) - // - // * 10.1.14 (supported in all AWS Regions except us-east-2) - // - // * 10.0.34 (supported in all AWS Regions) - // - // * 10.0.32 (supported in all AWS Regions) - // - // * 10.0.31 (supported in all AWS Regions) - // - // * 10.0.28 (supported in all AWS Regions) - // - // * 10.0.24 (supported in all AWS Regions) - // - // * 10.0.17 (supported in all AWS Regions except us-east-2, ca-central-1, - // eu-west-2) - // - // Microsoft SQL Server 2017 - // - // * 14.00.1000.169.v1 (supported for all editions, and all AWS Regions) - // - // Microsoft SQL Server 2016 - // - // * 13.00.4451.0.v1 (supported for all editions, and all AWS Regions) - // - // * 13.00.4422.0.v1 (supported for all editions, and all AWS Regions) - // - // * 13.00.2164.0.v1 (supported for all editions, and all AWS Regions) - // - // Microsoft SQL Server 2014 - // - // * 12.00.5546.0.v1 (supported for all editions, and all AWS Regions) - // - // * 12.00.5000.0.v1 (supported for all editions, and all AWS Regions) - // - // * 12.00.4422.0.v1 (supported for all editions except Enterprise Edition, - // and all AWS Regions except ca-central-1 and eu-west-2) - // - // Microsoft SQL Server 2012 - // - // * 11.00.6594.0.v1 (supported for all editions, and all AWS Regions) - // - // * 11.00.6020.0.v1 (supported for all editions, and all AWS Regions) - // - // * 11.00.5058.0.v1 (supported for all editions, and all AWS Regions except - // us-east-2, ca-central-1, and eu-west-2) - // - // * 11.00.2100.60.v1 (supported for all editions, and all AWS Regions except - // us-east-2, ca-central-1, and eu-west-2) - // - // Microsoft SQL Server 2008 R2 - // - // * 10.50.6529.0.v1 (supported for all editions, and all AWS Regions except - // us-east-2, ca-central-1, and eu-west-2) - // - // * 10.50.6000.34.v1 (supported for all editions, and all AWS Regions except - // us-east-2, ca-central-1, and eu-west-2) - // - // * 10.50.2789.0.v1 (supported for all editions, and all AWS Regions except - // us-east-2, ca-central-1, and eu-west-2) + // See Version and Feature Support on Amazon RDS (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_SQLServer.html#SQLServer.Concepts.General.FeatureSupport) + // in the Amazon RDS User Guide. // // MySQL // - // * 5.7.21 (supported in all AWS regions) + // See MySQL on Amazon RDS Versions (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_MySQL.html#MySQL.Concepts.VersionMgmt) + // in the Amazon RDS User Guide. // - // * 5.7.19 (supported in all AWS regions) + // Oracle // - // * 5.7.17 (supported in all AWS regions) + // See Oracle Database Engine Release Notes (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Appendix.Oracle.PatchComposition.html) + // in the Amazon RDS User Guide. // - // * 5.7.16 (supported in all AWS regions) + // PostgreSQL // - // 5.6.39(supported in all AWS Regions) - // - // 5.6.37(supported in all AWS Regions) - // - // 5.6.35(supported in all AWS Regions) - // - // 5.6.34(supported in all AWS Regions) - // - // 5.6.29(supported in all AWS Regions) - // - // 5.6.27 + // See Supported PostgreSQL Database Versions (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_PostgreSQL.html#PostgreSQL.Concepts.General.DBVersions) + // in the Amazon RDS User Guide. EngineVersion *string `type:"string"` // The amount of Provisioned IOPS (input/output operations per second) to be @@ -12687,6 +12696,10 @@ type CreateDBInstanceInput struct { // KMS key alias for the KMS encryption key. PerformanceInsightsKMSKeyId *string `type:"string"` + // The amount of time, in days, to retain Performance Insights data. Valid values + // are 7 or 731 (2 years). + PerformanceInsightsRetentionPeriod *int64 `type:"integer"` + // The port number on which the database accepts connections. // // MySQL @@ -12773,6 +12786,10 @@ type CreateDBInstanceInput struct { // Constraints: Minimum 30-minute window. PreferredMaintenanceWindow *string `type:"string"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // A value that specifies the order in which an Aurora Replica is promoted to // the primary instance after a failure of the existing primary instance. For // more information, see Fault Tolerance for an Aurora DB Cluster (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Aurora.Managing.html#Aurora.Managing.FaultTolerance). @@ -13053,6 +13070,12 @@ func (s *CreateDBInstanceInput) SetPerformanceInsightsKMSKeyId(v string) *Create return s } +// SetPerformanceInsightsRetentionPeriod sets the PerformanceInsightsRetentionPeriod field's value. +func (s *CreateDBInstanceInput) SetPerformanceInsightsRetentionPeriod(v int64) *CreateDBInstanceInput { + s.PerformanceInsightsRetentionPeriod = &v + return s +} + // SetPort sets the Port field's value. func (s *CreateDBInstanceInput) SetPort(v int64) *CreateDBInstanceInput { s.Port = &v @@ -13071,6 +13094,12 @@ func (s *CreateDBInstanceInput) SetPreferredMaintenanceWindow(v string) *CreateD return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *CreateDBInstanceInput) SetProcessorFeatures(v []*ProcessorFeature) *CreateDBInstanceInput { + s.ProcessorFeatures = v + return s +} + // SetPromotionTier sets the PromotionTier field's value. func (s *CreateDBInstanceInput) SetPromotionTier(v int64) *CreateDBInstanceInput { s.PromotionTier = &v @@ -13234,6 +13263,9 @@ type CreateDBInstanceReadReplicaInput struct { EnableIAMDatabaseAuthentication *bool `type:"boolean"` // True to enable Performance Insights for the read replica, and otherwise false. + // + // For more information, see Using Amazon Performance Insights (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_PerfInsights.html) + // in the Amazon Relational Database Service User Guide. EnablePerformanceInsights *bool `type:"boolean"` // The amount of Provisioned IOPS (input/output operations per second) to be @@ -13293,6 +13325,10 @@ type CreateDBInstanceReadReplicaInput struct { // KMS key alias for the KMS encryption key. PerformanceInsightsKMSKeyId *string `type:"string"` + // The amount of time, in days, to retain Performance Insights data. Valid values + // are 7 or 731 (2 years). + PerformanceInsightsRetentionPeriod *int64 `type:"integer"` + // The port number that the DB instance uses for connections. // // Default: Inherits from the source DB instance @@ -13343,6 +13379,10 @@ type CreateDBInstanceReadReplicaInput struct { // and Signature Version 4 Signing Process (http://docs.aws.amazon.com/general/latest/gr/signature-version-4.html). PreSignedUrl *string `type:"string"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // Specifies the accessibility options for the DB instance. A value of true // specifies an Internet-facing instance with a publicly resolvable DNS name, // which resolves to a public IP address. A value of false specifies an internal @@ -13405,6 +13445,10 @@ type CreateDBInstanceReadReplicaInput struct { // A list of tags. For more information, see Tagging Amazon RDS Resources (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_Tagging.html). Tags []*Tag `locationNameList:"Tag" type:"list"` + + // A value that specifies that the DB instance class of the DB instance uses + // its default processor features. + UseDefaultProcessorFeatures *bool `type:"boolean"` } // String returns the string representation @@ -13535,6 +13579,12 @@ func (s *CreateDBInstanceReadReplicaInput) SetPerformanceInsightsKMSKeyId(v stri return s } +// SetPerformanceInsightsRetentionPeriod sets the PerformanceInsightsRetentionPeriod field's value. +func (s *CreateDBInstanceReadReplicaInput) SetPerformanceInsightsRetentionPeriod(v int64) *CreateDBInstanceReadReplicaInput { + s.PerformanceInsightsRetentionPeriod = &v + return s +} + // SetPort sets the Port field's value. func (s *CreateDBInstanceReadReplicaInput) SetPort(v int64) *CreateDBInstanceReadReplicaInput { s.Port = &v @@ -13547,6 +13597,12 @@ func (s *CreateDBInstanceReadReplicaInput) SetPreSignedUrl(v string) *CreateDBIn return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *CreateDBInstanceReadReplicaInput) SetProcessorFeatures(v []*ProcessorFeature) *CreateDBInstanceReadReplicaInput { + s.ProcessorFeatures = v + return s +} + // SetPubliclyAccessible sets the PubliclyAccessible field's value. func (s *CreateDBInstanceReadReplicaInput) SetPubliclyAccessible(v bool) *CreateDBInstanceReadReplicaInput { s.PubliclyAccessible = &v @@ -13577,6 +13633,12 @@ func (s *CreateDBInstanceReadReplicaInput) SetTags(v []*Tag) *CreateDBInstanceRe return s } +// SetUseDefaultProcessorFeatures sets the UseDefaultProcessorFeatures field's value. +func (s *CreateDBInstanceReadReplicaInput) SetUseDefaultProcessorFeatures(v bool) *CreateDBInstanceReadReplicaInput { + s.UseDefaultProcessorFeatures = &v + return s +} + type CreateDBInstanceReadReplicaOutput struct { _ struct{} `type:"structure"` @@ -13610,6 +13672,13 @@ type CreateDBParameterGroupInput struct { // to a DB instance running a database engine and engine version compatible // with that DB parameter group family. // + // To list all of the available parameter group families, use the following + // command: + // + // aws rds describe-db-engine-versions --query "DBEngineVersions[].DBParameterGroupFamily" + // + // The output contains duplicates. + // // DBParameterGroupFamily is a required field DBParameterGroupFamily *string `type:"string" required:"true"` @@ -14380,6 +14449,10 @@ type DBCluster struct { // restore. EarliestRestorableTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` + // A list of log types that this DB cluster is configured to export to CloudWatch + // Logs. + EnabledCloudwatchLogsExports []*string `type:"list"` + // Specifies the connection endpoint for the primary instance of the DB cluster. Endpoint *string `type:"string"` @@ -14579,6 +14652,12 @@ func (s *DBCluster) SetEarliestRestorableTime(v time.Time) *DBCluster { return s } +// SetEnabledCloudwatchLogsExports sets the EnabledCloudwatchLogsExports field's value. +func (s *DBCluster) SetEnabledCloudwatchLogsExports(v []*string) *DBCluster { + s.EnabledCloudwatchLogsExports = v + return s +} + // SetEndpoint sets the Endpoint field's value. func (s *DBCluster) SetEndpoint(v string) *DBCluster { s.Endpoint = &v @@ -15502,6 +15581,10 @@ type DBInstance struct { // KMS key alias for the KMS encryption key. PerformanceInsightsKMSKeyId *string `type:"string"` + // The amount of time, in days, to retain Performance Insights data. Valid values + // are 7 or 731 (2 years). + PerformanceInsightsRetentionPeriod *int64 `type:"integer"` + // Specifies the daily time range during which automated backups are created // if automated backups are enabled, as determined by the BackupRetentionPeriod. PreferredBackupWindow *string `type:"string"` @@ -15510,6 +15593,10 @@ type DBInstance struct { // in Universal Coordinated Time (UTC). PreferredMaintenanceWindow *string `type:"string"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // A value that specifies the order in which an Aurora Replica is promoted to // the primary instance after a failure of the existing primary instance. For // more information, see Fault Tolerance for an Aurora DB Cluster (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Aurora.Managing.html#Aurora.Managing.FaultTolerance). @@ -15811,6 +15898,12 @@ func (s *DBInstance) SetPerformanceInsightsKMSKeyId(v string) *DBInstance { return s } +// SetPerformanceInsightsRetentionPeriod sets the PerformanceInsightsRetentionPeriod field's value. +func (s *DBInstance) SetPerformanceInsightsRetentionPeriod(v int64) *DBInstance { + s.PerformanceInsightsRetentionPeriod = &v + return s +} + // SetPreferredBackupWindow sets the PreferredBackupWindow field's value. func (s *DBInstance) SetPreferredBackupWindow(v string) *DBInstance { s.PreferredBackupWindow = &v @@ -15823,6 +15916,12 @@ func (s *DBInstance) SetPreferredMaintenanceWindow(v string) *DBInstance { return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *DBInstance) SetProcessorFeatures(v []*ProcessorFeature) *DBInstance { + s.ProcessorFeatures = v + return s +} + // SetPromotionTier sets the PromotionTier field's value. func (s *DBInstance) SetPromotionTier(v int64) *DBInstance { s.PromotionTier = &v @@ -16261,6 +16360,10 @@ type DBSnapshot struct { // of the snapshot. Port *int64 `type:"integer"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance when the DB snapshot was created. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // Provides the time when the snapshot was taken, in Universal Coordinated Time // (UTC). SnapshotCreateTime *time.Time `type:"timestamp" timestampFormat:"iso8601"` @@ -16405,6 +16508,12 @@ func (s *DBSnapshot) SetPort(v int64) *DBSnapshot { return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *DBSnapshot) SetProcessorFeatures(v []*ProcessorFeature) *DBSnapshot { + s.ProcessorFeatures = v + return s +} + // SetSnapshotCreateTime sets the SnapshotCreateTime field's value. func (s *DBSnapshot) SetSnapshotCreateTime(v time.Time) *DBSnapshot { s.SnapshotCreateTime = &v @@ -20846,7 +20955,9 @@ type DescribeReservedDBInstancesOfferingsInput struct { OfferingType *string `type:"string"` // Product description filter value. Specify this parameter to show only the - // available offerings matching the specified product description. + // available offerings that contain the specified product description. + // + // The results show offerings that partially match the filter value. ProductDescription *string `type:"string"` // The offering identifier filter value. Specify this parameter to show only @@ -22082,6 +22193,10 @@ type ModifyDBClusterInput struct { // * Must be a value from 1 to 35 BackupRetentionPeriod *int64 `type:"integer"` + // The configuration setting for the log types to be enabled for export to CloudWatch + // Logs for a specific DB cluster. + CloudwatchLogsExportConfiguration *CloudwatchLogsExportConfiguration `type:"structure"` + // The DB cluster identifier for the cluster being modified. This parameter // is not case-sensitive. // @@ -22105,7 +22220,7 @@ type ModifyDBClusterInput struct { // this parameter results in an outage. The change is applied during the next // maintenance window unless the ApplyImmediately parameter is set to true. // - // For a list of valid engine versions, see CreateDBInstance, or call DescribeDBEngineVersions. + // For a list of valid engine versions, see CreateDBCluster, or call DescribeDBEngineVersions. EngineVersion *string `type:"string"` // The new password for the master database user. This password can contain @@ -22227,6 +22342,12 @@ func (s *ModifyDBClusterInput) SetBackupRetentionPeriod(v int64) *ModifyDBCluste return s } +// SetCloudwatchLogsExportConfiguration sets the CloudwatchLogsExportConfiguration field's value. +func (s *ModifyDBClusterInput) SetCloudwatchLogsExportConfiguration(v *CloudwatchLogsExportConfiguration) *ModifyDBClusterInput { + s.CloudwatchLogsExportConfiguration = v + return s +} + // SetDBClusterIdentifier sets the DBClusterIdentifier field's value. func (s *ModifyDBClusterInput) SetDBClusterIdentifier(v string) *ModifyDBClusterInput { s.DBClusterIdentifier = &v @@ -22566,7 +22687,7 @@ type ModifyDBInstanceInput struct { CACertificateIdentifier *string `type:"string"` // The configuration setting for the log types to be enabled for export to CloudWatch - // Logs for a specific DB instance or DB cluster. + // Logs for a specific DB instance. CloudwatchLogsExportConfiguration *CloudwatchLogsExportConfiguration `type:"structure"` // True to copy all tags from the DB instance to snapshots of the DB instance, @@ -22708,6 +22829,9 @@ type ModifyDBInstanceInput struct { EnableIAMDatabaseAuthentication *bool `type:"boolean"` // True to enable Performance Insights for the DB instance, and otherwise false. + // + // For more information, see Using Amazon Performance Insights (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_PerfInsights.html) + // in the Amazon Relational Database Service User Guide. EnablePerformanceInsights *bool `type:"boolean"` // The version number of the database engine to upgrade to. Changing this parameter @@ -22719,7 +22843,8 @@ type ModifyDBInstanceInput struct { // new engine version must be specified. The new DB parameter group can be the // default for that DB parameter group family. // - // For a list of valid engine versions, see CreateDBInstance. + // For information about valid engine versions, see CreateDBInstance, or call + // DescribeDBEngineVersions. EngineVersion *string `type:"string"` // The new Provisioned IOPS (I/O operations per second) value for the RDS instance. @@ -22855,6 +22980,10 @@ type ModifyDBInstanceInput struct { // KMS key alias for the KMS encryption key. PerformanceInsightsKMSKeyId *string `type:"string"` + // The amount of time, in days, to retain Performance Insights data. Valid values + // are 7 or 731 (2 years). + PerformanceInsightsRetentionPeriod *int64 `type:"integer"` + // The daily time range during which automated backups are created if automated // backups are enabled, as determined by the BackupRetentionPeriod parameter. // Changing this parameter doesn't result in an outage and the change is asynchronously @@ -22894,6 +23023,10 @@ type ModifyDBInstanceInput struct { // Constraints: Must be at least 30 minutes PreferredMaintenanceWindow *string `type:"string"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // A value that specifies the order in which an Aurora Replica is promoted to // the primary instance after a failure of the existing primary instance. For // more information, see Fault Tolerance for an Aurora DB Cluster (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Aurora.Managing.html#Aurora.Managing.FaultTolerance). @@ -22949,6 +23082,10 @@ type ModifyDBInstanceInput struct { // device. TdeCredentialPassword *string `type:"string"` + // A value that specifies that the DB instance class of the DB instance uses + // its default processor features. + UseDefaultProcessorFeatures *bool `type:"boolean"` + // A list of EC2 VPC security groups to authorize on this DB instance. This // change is asynchronously applied as soon as possible. // @@ -23154,6 +23291,12 @@ func (s *ModifyDBInstanceInput) SetPerformanceInsightsKMSKeyId(v string) *Modify return s } +// SetPerformanceInsightsRetentionPeriod sets the PerformanceInsightsRetentionPeriod field's value. +func (s *ModifyDBInstanceInput) SetPerformanceInsightsRetentionPeriod(v int64) *ModifyDBInstanceInput { + s.PerformanceInsightsRetentionPeriod = &v + return s +} + // SetPreferredBackupWindow sets the PreferredBackupWindow field's value. func (s *ModifyDBInstanceInput) SetPreferredBackupWindow(v string) *ModifyDBInstanceInput { s.PreferredBackupWindow = &v @@ -23166,6 +23309,12 @@ func (s *ModifyDBInstanceInput) SetPreferredMaintenanceWindow(v string) *ModifyD return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *ModifyDBInstanceInput) SetProcessorFeatures(v []*ProcessorFeature) *ModifyDBInstanceInput { + s.ProcessorFeatures = v + return s +} + // SetPromotionTier sets the PromotionTier field's value. func (s *ModifyDBInstanceInput) SetPromotionTier(v int64) *ModifyDBInstanceInput { s.PromotionTier = &v @@ -23196,6 +23345,12 @@ func (s *ModifyDBInstanceInput) SetTdeCredentialPassword(v string) *ModifyDBInst return s } +// SetUseDefaultProcessorFeatures sets the UseDefaultProcessorFeatures field's value. +func (s *ModifyDBInstanceInput) SetUseDefaultProcessorFeatures(v bool) *ModifyDBInstanceInput { + s.UseDefaultProcessorFeatures = &v + return s +} + // SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. func (s *ModifyDBInstanceInput) SetVpcSecurityGroupIds(v []*string) *ModifyDBInstanceInput { s.VpcSecurityGroupIds = v @@ -24504,6 +24659,10 @@ type OrderableDBInstanceOption struct { // A list of Availability Zones for a DB instance. AvailabilityZones []*AvailabilityZone `locationNameList:"AvailabilityZone" type:"list"` + // A list of the available processor features for the DB instance class of a + // DB instance. + AvailableProcessorFeatures []*AvailableProcessorFeature `locationNameList:"AvailableProcessorFeature" type:"list"` + // The DB instance class for a DB instance. DBInstanceClass *string `type:"string"` @@ -24579,6 +24738,12 @@ func (s *OrderableDBInstanceOption) SetAvailabilityZones(v []*AvailabilityZone) return s } +// SetAvailableProcessorFeatures sets the AvailableProcessorFeatures field's value. +func (s *OrderableDBInstanceOption) SetAvailableProcessorFeatures(v []*AvailableProcessorFeature) *OrderableDBInstanceOption { + s.AvailableProcessorFeatures = v + return s +} + // SetDBInstanceClass sets the DBInstanceClass field's value. func (s *OrderableDBInstanceOption) SetDBInstanceClass(v string) *OrderableDBInstanceOption { s.DBInstanceClass = &v @@ -24970,6 +25135,10 @@ type PendingModifiedValues struct { // Specifies the pending port for the DB instance. Port *int64 `type:"integer"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // Specifies the storage type to be associated with the DB instance. StorageType *string `type:"string"` } @@ -25062,12 +25231,85 @@ func (s *PendingModifiedValues) SetPort(v int64) *PendingModifiedValues { return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *PendingModifiedValues) SetProcessorFeatures(v []*ProcessorFeature) *PendingModifiedValues { + s.ProcessorFeatures = v + return s +} + // SetStorageType sets the StorageType field's value. func (s *PendingModifiedValues) SetStorageType(v string) *PendingModifiedValues { s.StorageType = &v return s } +// Contains the processor features of a DB instance class. +// +// To specify the number of CPU cores, use the coreCount feature name for the +// Name parameter. To specify the number of threads per core, use the threadsPerCore +// feature name for the Name parameter. +// +// You can set the processor features of the DB instance class for a DB instance +// when you call one of the following actions: +// +// * CreateDBInstance +// +// * ModifyDBInstance +// +// * RestoreDBInstanceFromDBSnapshot +// +// * RestoreDBInstanceFromS3 +// +// * RestoreDBInstanceToPointInTime +// +// You can view the valid processor values for a particular instance class by +// calling the DescribeOrderableDBInstanceOptions action and specifying the +// instance class for the DBInstanceClass parameter. +// +// In addition, you can use the following actions for DB instance class processor +// information: +// +// * DescribeDBInstances +// +// * DescribeDBSnapshots +// +// * DescribeValidDBInstanceModifications +// +// For more information, see Configuring the Processor of the DB Instance Class +// (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/Concepts.DBInstanceClass.html#USER_ConfigureProcessor) +// in the Amazon RDS User Guide. +type ProcessorFeature struct { + _ struct{} `type:"structure"` + + // The name of the processor feature. Valid names are coreCount and threadsPerCore. + Name *string `type:"string"` + + // The value of a processor feature name. + Value *string `type:"string"` +} + +// String returns the string representation +func (s ProcessorFeature) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ProcessorFeature) GoString() string { + return s.String() +} + +// SetName sets the Name field's value. +func (s *ProcessorFeature) SetName(v string) *ProcessorFeature { + s.Name = &v + return s +} + +// SetValue sets the Value field's value. +func (s *ProcessorFeature) SetValue(v string) *ProcessorFeature { + s.Value = &v + return s +} + type PromoteReadReplicaDBClusterInput struct { _ struct{} `type:"structure"` @@ -26218,6 +26460,10 @@ type RestoreDBClusterFromS3Input struct { // The database name for the restored DB cluster. DatabaseName *string `type:"string"` + // The list of logs that the restored DB cluster is to export to CloudWatch + // Logs. + EnableCloudwatchLogsExports []*string `type:"list"` + // True to enable mapping of AWS Identity and Access Management (IAM) accounts // to database accounts, and otherwise false. // @@ -26462,6 +26708,12 @@ func (s *RestoreDBClusterFromS3Input) SetDatabaseName(v string) *RestoreDBCluste return s } +// SetEnableCloudwatchLogsExports sets the EnableCloudwatchLogsExports field's value. +func (s *RestoreDBClusterFromS3Input) SetEnableCloudwatchLogsExports(v []*string) *RestoreDBClusterFromS3Input { + s.EnableCloudwatchLogsExports = v + return s +} + // SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. func (s *RestoreDBClusterFromS3Input) SetEnableIAMDatabaseAuthentication(v bool) *RestoreDBClusterFromS3Input { s.EnableIAMDatabaseAuthentication = &v @@ -26639,6 +26891,10 @@ type RestoreDBClusterFromSnapshotInput struct { // The database name for the restored DB cluster. DatabaseName *string `type:"string"` + // The list of logs that the restored DB cluster is to export to CloudWatch + // Logs. + EnableCloudwatchLogsExports []*string `type:"list"` + // True to enable mapping of AWS Identity and Access Management (IAM) accounts // to database accounts, and otherwise false. // @@ -26765,6 +27021,12 @@ func (s *RestoreDBClusterFromSnapshotInput) SetDatabaseName(v string) *RestoreDB return s } +// SetEnableCloudwatchLogsExports sets the EnableCloudwatchLogsExports field's value. +func (s *RestoreDBClusterFromSnapshotInput) SetEnableCloudwatchLogsExports(v []*string) *RestoreDBClusterFromSnapshotInput { + s.EnableCloudwatchLogsExports = v + return s +} + // SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. func (s *RestoreDBClusterFromSnapshotInput) SetEnableIAMDatabaseAuthentication(v bool) *RestoreDBClusterFromSnapshotInput { s.EnableIAMDatabaseAuthentication = &v @@ -26878,6 +27140,10 @@ type RestoreDBClusterToPointInTimeInput struct { // Example: mySubnetgroup DBSubnetGroupName *string `type:"string"` + // The list of logs that the restored DB cluster is to export to CloudWatch + // Logs. + EnableCloudwatchLogsExports []*string `type:"list"` + // True to enable mapping of AWS Identity and Access Management (IAM) accounts // to database accounts, and otherwise false. // @@ -26915,9 +27181,9 @@ type RestoreDBClusterToPointInTimeInput struct { // The port number on which the new DB cluster accepts connections. // - // Constraints: Value must be 1150-65535 + // Constraints: A value from 1150-65535. // - // Default: The same port as the original DB cluster. + // Default: The default port for the engine. Port *int64 `type:"integer"` // The date and time to restore the DB cluster to. @@ -27021,6 +27287,12 @@ func (s *RestoreDBClusterToPointInTimeInput) SetDBSubnetGroupName(v string) *Res return s } +// SetEnableCloudwatchLogsExports sets the EnableCloudwatchLogsExports field's value. +func (s *RestoreDBClusterToPointInTimeInput) SetEnableCloudwatchLogsExports(v []*string) *RestoreDBClusterToPointInTimeInput { + s.EnableCloudwatchLogsExports = v + return s +} + // SetEnableIAMDatabaseAuthentication sets the EnableIAMDatabaseAuthentication field's value. func (s *RestoreDBClusterToPointInTimeInput) SetEnableIAMDatabaseAuthentication(v bool) *RestoreDBClusterToPointInTimeInput { s.EnableIAMDatabaseAuthentication = &v @@ -27272,6 +27544,10 @@ type RestoreDBInstanceFromDBSnapshotInput struct { // Constraints: Value must be 1150-65535 Port *int64 `type:"integer"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // Specifies the accessibility options for the DB instance. A value of true // specifies an Internet-facing instance with a publicly resolvable DNS name, // which resolves to a public IP address. A value of false specifies an internal @@ -27308,6 +27584,10 @@ type RestoreDBInstanceFromDBSnapshotInput struct { // The password for the given ARN from the key store in order to access the // device. TdeCredentialPassword *string `type:"string"` + + // A value that specifies that the DB instance class of the DB instance uses + // its default processor features. + UseDefaultProcessorFeatures *bool `type:"boolean"` } // String returns the string representation @@ -27444,6 +27724,12 @@ func (s *RestoreDBInstanceFromDBSnapshotInput) SetPort(v int64) *RestoreDBInstan return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *RestoreDBInstanceFromDBSnapshotInput) SetProcessorFeatures(v []*ProcessorFeature) *RestoreDBInstanceFromDBSnapshotInput { + s.ProcessorFeatures = v + return s +} + // SetPubliclyAccessible sets the PubliclyAccessible field's value. func (s *RestoreDBInstanceFromDBSnapshotInput) SetPubliclyAccessible(v bool) *RestoreDBInstanceFromDBSnapshotInput { s.PubliclyAccessible = &v @@ -27474,6 +27760,12 @@ func (s *RestoreDBInstanceFromDBSnapshotInput) SetTdeCredentialPassword(v string return s } +// SetUseDefaultProcessorFeatures sets the UseDefaultProcessorFeatures field's value. +func (s *RestoreDBInstanceFromDBSnapshotInput) SetUseDefaultProcessorFeatures(v bool) *RestoreDBInstanceFromDBSnapshotInput { + s.UseDefaultProcessorFeatures = &v + return s +} + type RestoreDBInstanceFromDBSnapshotOutput struct { _ struct{} `type:"structure"` @@ -27596,6 +27888,9 @@ type RestoreDBInstanceFromS3Input struct { EnableIAMDatabaseAuthentication *bool `type:"boolean"` // True to enable Performance Insights for the DB instance, and otherwise false. + // + // For more information, see Using Amazon Performance Insights (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_PerfInsights.html) + // in the Amazon Relational Database Service User Guide. EnablePerformanceInsights *bool `type:"boolean"` // The name of the database engine to be used for this instance. @@ -27606,7 +27901,8 @@ type RestoreDBInstanceFromS3Input struct { Engine *string `type:"string" required:"true"` // The version number of the database engine to use. Choose the latest minor - // version of your database engine as specified in CreateDBInstance. + // version of your database engine. For information about engine versions, see + // CreateDBInstance, or call DescribeDBEngineVersions. EngineVersion *string `type:"string"` // The amount of Provisioned IOPS (input/output operations per second) to allocate @@ -27682,6 +27978,10 @@ type RestoreDBInstanceFromS3Input struct { // the KMS key alias for the KMS encryption key. PerformanceInsightsKMSKeyId *string `type:"string"` + // The amount of time, in days, to retain Performance Insights data. Valid values + // are 7 or 731 (2 years). + PerformanceInsightsRetentionPeriod *int64 `type:"integer"` + // The port number on which the database accepts connections. // // Type: Integer @@ -27722,6 +28022,10 @@ type RestoreDBInstanceFromS3Input struct { // * Must be at least 30 minutes. PreferredMaintenanceWindow *string `type:"string"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // Specifies whether the DB instance is publicly accessible or not. For more // information, see CreateDBInstance. PubliclyAccessible *bool `type:"boolean"` @@ -27770,6 +28074,10 @@ type RestoreDBInstanceFromS3Input struct { // see Tagging Amazon RDS Resources (http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/USER_Tagging.html). Tags []*Tag `locationNameList:"Tag" type:"list"` + // A value that specifies that the DB instance class of the DB instance uses + // its default processor features. + UseDefaultProcessorFeatures *bool `type:"boolean"` + // A list of VPC security groups to associate with this DB instance. VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` } @@ -27971,6 +28279,12 @@ func (s *RestoreDBInstanceFromS3Input) SetPerformanceInsightsKMSKeyId(v string) return s } +// SetPerformanceInsightsRetentionPeriod sets the PerformanceInsightsRetentionPeriod field's value. +func (s *RestoreDBInstanceFromS3Input) SetPerformanceInsightsRetentionPeriod(v int64) *RestoreDBInstanceFromS3Input { + s.PerformanceInsightsRetentionPeriod = &v + return s +} + // SetPort sets the Port field's value. func (s *RestoreDBInstanceFromS3Input) SetPort(v int64) *RestoreDBInstanceFromS3Input { s.Port = &v @@ -27989,6 +28303,12 @@ func (s *RestoreDBInstanceFromS3Input) SetPreferredMaintenanceWindow(v string) * return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *RestoreDBInstanceFromS3Input) SetProcessorFeatures(v []*ProcessorFeature) *RestoreDBInstanceFromS3Input { + s.ProcessorFeatures = v + return s +} + // SetPubliclyAccessible sets the PubliclyAccessible field's value. func (s *RestoreDBInstanceFromS3Input) SetPubliclyAccessible(v bool) *RestoreDBInstanceFromS3Input { s.PubliclyAccessible = &v @@ -28043,6 +28363,12 @@ func (s *RestoreDBInstanceFromS3Input) SetTags(v []*Tag) *RestoreDBInstanceFromS return s } +// SetUseDefaultProcessorFeatures sets the UseDefaultProcessorFeatures field's value. +func (s *RestoreDBInstanceFromS3Input) SetUseDefaultProcessorFeatures(v bool) *RestoreDBInstanceFromS3Input { + s.UseDefaultProcessorFeatures = &v + return s +} + // SetVpcSecurityGroupIds sets the VpcSecurityGroupIds field's value. func (s *RestoreDBInstanceFromS3Input) SetVpcSecurityGroupIds(v []*string) *RestoreDBInstanceFromS3Input { s.VpcSecurityGroupIds = v @@ -28207,6 +28533,10 @@ type RestoreDBInstanceToPointInTimeInput struct { // Default: The same port as the original DB instance. Port *int64 `type:"integer"` + // The number of CPU cores and the number of threads per core for the DB instance + // class of the DB instance. + ProcessorFeatures []*ProcessorFeature `locationNameList:"ProcessorFeature" type:"list"` + // Specifies the accessibility options for the DB instance. A value of true // specifies an Internet-facing instance with a publicly resolvable DNS name, // which resolves to a public IP address. A value of false specifies an internal @@ -28279,6 +28609,10 @@ type RestoreDBInstanceToPointInTimeInput struct { // device. TdeCredentialPassword *string `type:"string"` + // A value that specifies that the DB instance class of the DB instance uses + // its default processor features. + UseDefaultProcessorFeatures *bool `type:"boolean"` + // Specifies whether (true) or not (false) the DB instance is restored from // the latest backup time. // @@ -28410,6 +28744,12 @@ func (s *RestoreDBInstanceToPointInTimeInput) SetPort(v int64) *RestoreDBInstanc return s } +// SetProcessorFeatures sets the ProcessorFeatures field's value. +func (s *RestoreDBInstanceToPointInTimeInput) SetProcessorFeatures(v []*ProcessorFeature) *RestoreDBInstanceToPointInTimeInput { + s.ProcessorFeatures = v + return s +} + // SetPubliclyAccessible sets the PubliclyAccessible field's value. func (s *RestoreDBInstanceToPointInTimeInput) SetPubliclyAccessible(v bool) *RestoreDBInstanceToPointInTimeInput { s.PubliclyAccessible = &v @@ -28458,6 +28798,12 @@ func (s *RestoreDBInstanceToPointInTimeInput) SetTdeCredentialPassword(v string) return s } +// SetUseDefaultProcessorFeatures sets the UseDefaultProcessorFeatures field's value. +func (s *RestoreDBInstanceToPointInTimeInput) SetUseDefaultProcessorFeatures(v bool) *RestoreDBInstanceToPointInTimeInput { + s.UseDefaultProcessorFeatures = &v + return s +} + // SetUseLatestRestorableTime sets the UseLatestRestorableTime field's value. func (s *RestoreDBInstanceToPointInTimeInput) SetUseLatestRestorableTime(v bool) *RestoreDBInstanceToPointInTimeInput { s.UseLatestRestorableTime = &v @@ -28959,6 +29305,9 @@ type ValidDBInstanceModificationsMessage struct { // Valid storage options for your DB instance. Storage []*ValidStorageOptions `locationNameList:"ValidStorageOptions" type:"list"` + + // Valid processor features for your DB instance. + ValidProcessorFeatures []*AvailableProcessorFeature `locationNameList:"AvailableProcessorFeature" type:"list"` } // String returns the string representation @@ -28977,6 +29326,12 @@ func (s *ValidDBInstanceModificationsMessage) SetStorage(v []*ValidStorageOption return s } +// SetValidProcessorFeatures sets the ValidProcessorFeatures field's value. +func (s *ValidDBInstanceModificationsMessage) SetValidProcessorFeatures(v []*AvailableProcessorFeature) *ValidDBInstanceModificationsMessage { + s.ValidProcessorFeatures = v + return s +} + // Information about valid modifications that you can make to your DB instance. // Contains the result of a successful call to the DescribeValidDBInstanceModifications // action. diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/examples_test.go b/vendor/github.com/aws/aws-sdk-go/service/rds/examples_test.go index 3efa0534d..5765eea11 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rds/examples_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/examples_test.go @@ -2803,6 +2803,8 @@ func ExampleRDS_RestoreDBClusterFromSnapshot_shared00() { fmt.Println(rds.ErrCodeOptionGroupNotFoundFault, aerr.Error()) case rds.ErrCodeKMSKeyNotAccessibleFault: fmt.Println(rds.ErrCodeKMSKeyNotAccessibleFault, aerr.Error()) + case rds.ErrCodeDBClusterParameterGroupNotFoundFault: + fmt.Println(rds.ErrCodeDBClusterParameterGroupNotFoundFault, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -2865,6 +2867,8 @@ func ExampleRDS_RestoreDBClusterToPointInTime_shared00() { fmt.Println(rds.ErrCodeOptionGroupNotFoundFault, aerr.Error()) case rds.ErrCodeStorageQuotaExceededFault: fmt.Println(rds.ErrCodeStorageQuotaExceededFault, aerr.Error()) + case rds.ErrCodeDBClusterParameterGroupNotFoundFault: + fmt.Println(rds.ErrCodeDBClusterParameterGroupNotFoundFault, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -2929,6 +2933,8 @@ func ExampleRDS_RestoreDBInstanceFromDBSnapshot_shared00() { fmt.Println(rds.ErrCodeDBSecurityGroupNotFoundFault, aerr.Error()) case rds.ErrCodeDomainNotFoundFault: fmt.Println(rds.ErrCodeDomainNotFoundFault, aerr.Error()) + case rds.ErrCodeDBParameterGroupNotFoundFault: + fmt.Println(rds.ErrCodeDBParameterGroupNotFoundFault, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -2997,6 +3003,8 @@ func ExampleRDS_RestoreDBInstanceToPointInTime_shared00() { fmt.Println(rds.ErrCodeDBSecurityGroupNotFoundFault, aerr.Error()) case rds.ErrCodeDomainNotFoundFault: fmt.Println(rds.ErrCodeDomainNotFoundFault, aerr.Error()) + case rds.ErrCodeDBParameterGroupNotFoundFault: + fmt.Println(rds.ErrCodeDBParameterGroupNotFoundFault, aerr.Error()) default: fmt.Println(aerr.Error()) } diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/builder.go b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/builder.go new file mode 100644 index 000000000..552acd130 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/builder.go @@ -0,0 +1,127 @@ +package rdsutils + +import ( + "fmt" + "net/url" + + "github.com/aws/aws-sdk-go/aws/awserr" + "github.com/aws/aws-sdk-go/aws/credentials" +) + +// ConnectionFormat is the type of connection that will be +// used to connect to the database +type ConnectionFormat string + +// ConnectionFormat enums +const ( + NoConnectionFormat ConnectionFormat = "" + TCPFormat ConnectionFormat = "tcp" +) + +// ErrNoConnectionFormat will be returned during build if no format had been +// specified +var ErrNoConnectionFormat = awserr.New("NoConnectionFormat", "No connection format was specified", nil) + +// ConnectionStringBuilder is a builder that will construct a connection +// string with the provided parameters. params field is required to have +// a tls specification and allowCleartextPasswords must be set to true. +type ConnectionStringBuilder struct { + dbName string + endpoint string + region string + user string + creds *credentials.Credentials + + connectFormat ConnectionFormat + params url.Values +} + +// NewConnectionStringBuilder will return an ConnectionStringBuilder +func NewConnectionStringBuilder(endpoint, region, dbUser, dbName string, creds *credentials.Credentials) ConnectionStringBuilder { + return ConnectionStringBuilder{ + dbName: dbName, + endpoint: endpoint, + region: region, + user: dbUser, + creds: creds, + } +} + +// WithEndpoint will return a builder with the given endpoint +func (b ConnectionStringBuilder) WithEndpoint(endpoint string) ConnectionStringBuilder { + b.endpoint = endpoint + return b +} + +// WithRegion will return a builder with the given region +func (b ConnectionStringBuilder) WithRegion(region string) ConnectionStringBuilder { + b.region = region + return b +} + +// WithUser will return a builder with the given user +func (b ConnectionStringBuilder) WithUser(user string) ConnectionStringBuilder { + b.user = user + return b +} + +// WithDBName will return a builder with the given database name +func (b ConnectionStringBuilder) WithDBName(dbName string) ConnectionStringBuilder { + b.dbName = dbName + return b +} + +// WithParams will return a builder with the given params. The parameters +// will be included in the connection query string +// +// Example: +// v := url.Values{} +// v.Add("tls", "rds") +// b := rdsutils.NewConnectionBuilder(endpoint, region, user, dbname, creds) +// connectStr, err := b.WithParams(v).WithTCPFormat().Build() +func (b ConnectionStringBuilder) WithParams(params url.Values) ConnectionStringBuilder { + b.params = params + return b +} + +// WithFormat will return a builder with the given connection format +func (b ConnectionStringBuilder) WithFormat(f ConnectionFormat) ConnectionStringBuilder { + b.connectFormat = f + return b +} + +// WithTCPFormat will set the format to TCP and return the modified builder +func (b ConnectionStringBuilder) WithTCPFormat() ConnectionStringBuilder { + return b.WithFormat(TCPFormat) +} + +// Build will return a new connection string that can be used to open a connection +// to the desired database. +// +// Example: +// b := rdsutils.NewConnectionStringBuilder(endpoint, region, user, dbname, creds) +// connectStr, err := b.WithTCPFormat().Build() +// if err != nil { +// panic(err) +// } +// const dbType = "mysql" +// db, err := sql.Open(dbType, connectStr) +func (b ConnectionStringBuilder) Build() (string, error) { + if b.connectFormat == NoConnectionFormat { + return "", ErrNoConnectionFormat + } + + authToken, err := BuildAuthToken(b.endpoint, b.region, b.user, b.creds) + if err != nil { + return "", err + } + + connectionStr := fmt.Sprintf("%s:%s@%s(%s)/%s", + b.user, authToken, string(b.connectFormat), b.endpoint, b.dbName, + ) + + if len(b.params) > 0 { + connectionStr = fmt.Sprintf("%s?%s", connectionStr, b.params.Encode()) + } + return connectionStr, nil +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/builder_test.go b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/builder_test.go new file mode 100644 index 000000000..9bc3b8463 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/builder_test.go @@ -0,0 +1,58 @@ +package rdsutils_test + +import ( + "net/url" + "regexp" + "testing" + + "github.com/aws/aws-sdk-go/aws/credentials" + "github.com/aws/aws-sdk-go/service/rds/rdsutils" +) + +func TestConnectionStringBuilder(t *testing.T) { + cases := []struct { + user string + endpoint string + region string + dbName string + values url.Values + format rdsutils.ConnectionFormat + creds *credentials.Credentials + + expectedErr error + expectedConnectRegex string + }{ + { + user: "foo", + endpoint: "foo.bar", + region: "region", + dbName: "name", + format: rdsutils.NoConnectionFormat, + creds: credentials.NewStaticCredentials("AKID", "SECRET", "SESSION"), + expectedErr: rdsutils.ErrNoConnectionFormat, + expectedConnectRegex: "", + }, + { + user: "foo", + endpoint: "foo.bar", + region: "region", + dbName: "name", + format: rdsutils.TCPFormat, + creds: credentials.NewStaticCredentials("AKID", "SECRET", "SESSION"), + expectedConnectRegex: `^foo:foo.bar\?Action=connect\&DBUser=foo.*\@tcp\(foo.bar\)/name`, + }, + } + + for _, c := range cases { + b := rdsutils.NewConnectionStringBuilder(c.endpoint, c.region, c.user, c.dbName, c.creds) + connectStr, err := b.WithFormat(c.format).Build() + + if e, a := c.expectedErr, err; e != a { + t.Errorf("expected %v error, but received %v", e, a) + } + + if re, a := regexp.MustCompile(c.expectedConnectRegex), connectStr; !re.MatchString(a) { + t.Errorf("expect %s to match %s", re, a) + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/connect.go b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/connect.go index cfa923188..be248a8d4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/connect.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/connect.go @@ -9,16 +9,13 @@ import ( "github.com/aws/aws-sdk-go/aws/signer/v4" ) -// BuildAuthToken will return a authentication token for the database's connect -// based on the RDS database endpoint, AWS region, IAM user or role, and AWS credentials. +// BuildAuthToken will return an authorization token used as the password for a DB +// connection. // -// Endpoint consists of the hostname and port, IE hostname:port, of the RDS database. -// Region is the AWS region the RDS database is in and where the authentication token -// will be generated for. DbUser is the IAM user or role the request will be authenticated -// for. The creds is the AWS credentials the authentication token is signed with. -// -// An error is returned if the authentication token is unable to be signed with -// the credentials, or the endpoint is not a valid URL. +// * endpoint - Endpoint consists of the port needed to connect to the DB. : +// * region - Region is the location of where the DB is +// * dbUser - User account within the database to sign in with +// * creds - Credentials to be signed with // // The following example shows how to use BuildAuthToken to create an authentication // token for connecting to a MySQL database in RDS. @@ -27,12 +24,12 @@ import ( // // // Create the MySQL DNS string for the DB connection // // user:password@protocol(endpoint)/dbname? -// dnsStr = fmt.Sprintf("%s:%s@tcp(%s)/%s?tls=true", +// connectStr = fmt.Sprintf("%s:%s@tcp(%s)/%s?allowCleartextPasswords=true&tls=rds", // dbUser, authToken, dbEndpoint, dbName, // ) // // // Use db to perform SQL operations on database -// db, err := sql.Open("mysql", dnsStr) +// db, err := sql.Open("mysql", connectStr) // // See http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/UsingWithRDS.IAMDBAuth.html // for more information on using IAM database authentication with RDS. diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/doc.go b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/doc.go new file mode 100644 index 000000000..d1ff2511a --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/doc.go @@ -0,0 +1,18 @@ +// Package rdsutils is used to generate authentication tokens used to +// connect to a givent Amazon Relational Database Service (RDS) database. +// +// Before using the authentication please visit the docs here to ensure +// the database has the proper policies to allow for IAM token authentication. +// https://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/UsingWithRDS.IAMDBAuth.html#UsingWithRDS.IAMDBAuth.Availability +// +// When building the connection string, there are two required parameters that are needed to be set on the query. +// * tls +// * allowCleartextPasswords must be set to true +// +// Example creating a basic auth token with the builder: +// v := url.Values{} +// v.Add("tls", "tls_profile_name") +// v.Add("allowCleartextPasswords", "true") +// b := rdsutils.NewConnectionStringBuilder(endpoint, region, user, dbname, creds) +// connectStr, err := b.WithTCPFormat().WithParams(v).Build() +package rdsutils diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/example_test.go b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/example_test.go new file mode 100644 index 000000000..d320a2e79 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/rdsutils/example_test.go @@ -0,0 +1,119 @@ +// +build example,exclude + +package rdsutils_test + +import ( + "crypto/tls" + "crypto/x509" + "database/sql" + "flag" + "fmt" + "io/ioutil" + "net/http" + "net/url" + "os" + + "github.com/go-sql-driver/mysql" + + "github.com/aws/aws-sdk-go/aws/credentials/stscreds" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/service/rds/rdsutils" +) + +// ExampleConnectionStringBuilder contains usage of assuming a role and using +// that to build the auth token. +// Usage: +// ./main -user "iamuser" -dbname "foo" -region "us-west-2" -rolearn "arn" -endpoint "dbendpoint" -port 3306 +func ExampleConnectionStringBuilder() { + userPtr := flag.String("user", "", "user of the credentials") + regionPtr := flag.String("region", "us-east-1", "region to be used when grabbing sts creds") + roleArnPtr := flag.String("rolearn", "", "role arn to be used when grabbing sts creds") + endpointPtr := flag.String("endpoint", "", "DB endpoint to be connected to") + portPtr := flag.Int("port", 3306, "DB port to be connected to") + tablePtr := flag.String("table", "test_table", "DB table to query against") + dbNamePtr := flag.String("dbname", "", "DB name to query against") + flag.Parse() + + // Check required flags. Will exit with status code 1 if + // required field isn't set. + if err := requiredFlags( + userPtr, + regionPtr, + roleArnPtr, + endpointPtr, + portPtr, + dbNamePtr, + ); err != nil { + fmt.Printf("Error: %v\n\n", err) + flag.PrintDefaults() + os.Exit(1) + } + + err := registerRDSMysqlCerts(http.DefaultClient) + if err != nil { + panic(err) + } + + sess := session.Must(session.NewSession()) + creds := stscreds.NewCredentials(sess, *roleArnPtr) + + v := url.Values{} + // required fields for DB connection + v.Add("tls", "rds") + v.Add("allowCleartextPasswords", "true") + endpoint := fmt.Sprintf("%s:%d", *endpointPtr, *portPtr) + + b := rdsutils.NewConnectionStringBuilder(endpoint, *regionPtr, *userPtr, *dbNamePtr, creds) + connectStr, err := b.WithTCPFormat().WithParams(v).Build() + + const dbType = "mysql" + db, err := sql.Open(dbType, connectStr) + // if an error is encountered here, then most likely security groups are incorrect + // in the database. + if err != nil { + panic(fmt.Errorf("failed to open connection to the database")) + } + + rows, err := db.Query(fmt.Sprintf("SELECT * FROM %s LIMIT 1", *tablePtr)) + if err != nil { + panic(fmt.Errorf("failed to select from table, %q, with %v", *tablePtr, err)) + } + + for rows.Next() { + columns, err := rows.Columns() + if err != nil { + panic(fmt.Errorf("failed to read columns from row: %v", err)) + } + + fmt.Printf("rows colums:\n%d\n", len(columns)) + } +} + +func requiredFlags(flags ...interface{}) error { + for _, f := range flags { + switch f.(type) { + case nil: + return fmt.Errorf("one or more required flags were not set") + } + } + return nil +} + +func registerRDSMysqlCerts(c *http.Client) error { + resp, err := c.Get("https://s3.amazonaws.com/rds-downloads/rds-combined-ca-bundle.pem") + if err != nil { + return err + } + + pem, err := ioutil.ReadAll(resp.Body) + if err != nil { + return err + } + + rootCertPool := x509.NewCertPool() + if ok := rootCertPool.AppendCertsFromPEM(pem); !ok { + return fmt.Errorf("failed to append cert to cert pool!") + } + + return mysql.RegisterTLSConfig("rds", &tls.Config{RootCAs: rootCertPool, InsecureSkipVerify: true}) +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/rds/service.go b/vendor/github.com/aws/aws-sdk-go/service/rds/service.go index 2e2ec2e97..f2d0efaf7 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rds/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rds/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "rds" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "rds" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "RDS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the RDS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/redshift/api.go b/vendor/github.com/aws/aws-sdk-go/service/redshift/api.go index 7fcc803f1..a2d952afb 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/redshift/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/redshift/api.go @@ -1376,7 +1376,7 @@ func (c *Redshift) CreateTagsRequest(input *CreateTagsInput) (req *request.Reque // // Adds one or more tags to a specified resource. // -// A resource can have up to 10 tags. If you try to create more than 10 tags +// A resource can have up to 50 tags. If you try to create more than 50 tags // for a resource, you will receive an error and the attempt will fail. // // If you specify a key that already exists for the resource, the value for @@ -14248,7 +14248,7 @@ type ModifyClusterInput struct { PubliclyAccessible *bool `type:"boolean"` // A list of virtual private cloud (VPC) security groups to be associated with - // the cluster. + // the cluster. This change is asynchronously applied as soon as possible. VpcSecurityGroupIds []*string `locationNameList:"VpcSecurityGroupId" type:"list"` } diff --git a/vendor/github.com/aws/aws-sdk-go/service/redshift/service.go b/vendor/github.com/aws/aws-sdk-go/service/redshift/service.go index b859170a4..a750d141c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/redshift/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/redshift/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "redshift" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "redshift" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Redshift" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Redshift client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/rekognition/api.go b/vendor/github.com/aws/aws-sdk-go/service/rekognition/api.go index 86c8fbce6..504d82be0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rekognition/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rekognition/api.go @@ -90,7 +90,8 @@ func (c *Rekognition) CompareFacesRequest(input *CompareFacesInput) (req *reques // This is a stateless API operation. That is, data returned by this operation // doesn't persist. // -// For an example, see faces-compare-images. +// For an example, see Comparing Faces in Images in the Amazon Rekognition Developer +// Guide. // // This operation requires permissions to perform the rekognition:CompareFaces // action. @@ -112,7 +113,7 @@ func (c *Rekognition) CompareFacesRequest(input *CompareFacesInput) (req *reques // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -302,8 +303,9 @@ func (c *Rekognition) CreateStreamProcessorRequest(input *CreateStreamProcessorI // Creates an Amazon Rekognition stream processor that you can use to detect // and recognize faces in a streaming video. // -// Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. -// Rekognition Video sends analysis results to Amazon Kinesis Data Streams. +// Amazon Rekognition Video is a consumer of live video from Amazon Kinesis +// Video Streams. Amazon Rekognition Video sends analysis results to Amazon +// Kinesis Data Streams. // // You provide as input a Kinesis video stream (Input) and a Kinesis data stream // (Output) stream. You also specify the face recognition criteria in Settings. @@ -339,10 +341,10 @@ func (c *Rekognition) CreateStreamProcessorRequest(input *CreateStreamProcessorI // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeResourceInUseException "ResourceInUseException" // @@ -824,8 +826,6 @@ func (c *Rekognition) DetectFacesRequest(input *DetectFacesInput) (req *request. // This is a stateless API operation. That is, the operation does not persist // any data. // -// For an example, see procedure-detecting-faces-in-images. -// // This operation requires permissions to perform the rekognition:DetectFaces // action. // @@ -846,7 +846,7 @@ func (c *Rekognition) DetectFacesRequest(input *DetectFacesInput) (req *request. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -931,11 +931,14 @@ func (c *Rekognition) DetectLabelsRequest(input *DetectLabelsInput) (req *reques // Detects instances of real-world entities within an image (JPEG or PNG) provided // as input. This includes objects like flower, tree, and table; events like // wedding, graduation, and birthday party; and concepts like landscape, evening, -// and nature. For an example, see images-s3. +// and nature. +// +// For an example, see Analyzing Images Stored in an Amazon S3 Bucket in the +// Amazon Rekognition Developer Guide. // // DetectLabels does not support the detection of activities. However, activity // detection is supported for label detection in videos. For more information, -// see . +// see StartLabelDetection in the Amazon Rekognition Developer Guide. // // You pass the input image as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon @@ -1000,7 +1003,7 @@ func (c *Rekognition) DetectLabelsRequest(input *DetectLabelsInput) (req *reques // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -1088,8 +1091,10 @@ func (c *Rekognition) DetectModerationLabelsRequest(input *DetectModerationLabel // images containing suggestive content. // // To filter images, use the labels returned by DetectModerationLabels to determine -// which types of content are appropriate. For information about moderation -// labels, see moderation. +// which types of content are appropriate. +// +// For information about moderation labels, see Detecting Unsafe Content in +// the Amazon Rekognition Developer Guide. // // You pass the input image either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon @@ -1113,7 +1118,7 @@ func (c *Rekognition) DetectModerationLabelsRequest(input *DetectModerationLabel // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -1225,7 +1230,8 @@ func (c *Rekognition) DetectTextRequest(input *DetectTextInput) (req *request.Re // To be detected, text must be within +/- 30 degrees orientation of the horizontal // axis. // -// For more information, see text-detection. +// For more information, see DetectText in the Amazon Rekognition Developer +// Guide. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1244,7 +1250,7 @@ func (c *Rekognition) DetectTextRequest(input *DetectTextInput) (req *request.Re // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -1329,7 +1335,10 @@ func (c *Rekognition) GetCelebrityInfoRequest(input *GetCelebrityInfoInput) (req // Gets the name and additional information about a celebrity based on his or // her Rekognition ID. The additional information is returned as an array of // URLs. If there is no additional information about the celebrity, this list -// is empty. For more information, see get-celebrity-info-procedure. +// is empty. +// +// For more information, see Recognizing Celebrities in an Image in the Amazon +// Rekognition Developer Guide. // // This operation requires permissions to perform the rekognition:GetCelebrityInfo // action. @@ -1432,18 +1441,20 @@ func (c *Rekognition) GetCelebrityRecognitionRequest(input *GetCelebrityRecognit // GetCelebrityRecognition API operation for Amazon Rekognition. // -// Gets the celebrity recognition results for a Rekognition Video analysis started -// by . +// Gets the celebrity recognition results for a Amazon Rekognition Video analysis +// started by . // // Celebrity recognition in a video is an asynchronous operation. Analysis is // started by a call to which returns a job identifier (JobId). When the celebrity -// recognition operation finishes, Rekognition Video publishes a completion +// recognition operation finishes, Amazon Rekognition Video publishes a completion // status to the Amazon Simple Notification Service topic registered in the // initial call to StartCelebrityRecognition. To get the results of the celebrity // recognition analysis, first check that the status value published to the // Amazon SNS topic is SUCCEEDED. If so, call GetCelebrityDetection and pass // the job identifier (JobId) from the initial call to StartCelebrityDetection. -// For more information, see video. +// +// For more information, see Working With Stored Videos in the Amazon Rekognition +// Developer Guide. // // GetCelebrityRecognition returns detected celebrities and the time(s) they // are detected in an array (Celebrities) of objects. Each CelebrityRecognition @@ -1453,7 +1464,7 @@ func (c *Rekognition) GetCelebrityRecognitionRequest(input *GetCelebrityRecognit // GetCelebrityRecognition only returns the default facial attributes (BoundingBox, // Confidence, Landmarks, Pose, and Quality). The other facial attributes listed // in the Face object of the following response syntax are not returned. For -// more information, see . +// more information, see FaceDetail in the Amazon Rekognition Developer Guide. // // By default, the Celebrities array is sorted by time (milliseconds from the // start of the video). You can also sort the array by celebrity by specifying @@ -1623,17 +1634,20 @@ func (c *Rekognition) GetContentModerationRequest(input *GetContentModerationInp // GetContentModeration API operation for Amazon Rekognition. // -// Gets the content moderation analysis results for a Rekognition Video analysis -// started by . +// Gets the content moderation analysis results for a Amazon Rekognition Video +// analysis started by . // // Content moderation analysis of a video is an asynchronous operation. You // start analysis by calling . which returns a job identifier (JobId). When -// analysis finishes, Rekognition Video publishes a completion status to the -// Amazon Simple Notification Service topic registered in the initial call to -// StartContentModeration. To get the results of the content moderation analysis, -// first check that the status value published to the Amazon SNS topic is SUCCEEDED. -// If so, call GetCelebrityDetection and pass the job identifier (JobId) from -// the initial call to StartCelebrityDetection. For more information, see video. +// analysis finishes, Amazon Rekognition Video publishes a completion status +// to the Amazon Simple Notification Service topic registered in the initial +// call to StartContentModeration. To get the results of the content moderation +// analysis, first check that the status value published to the Amazon SNS topic +// is SUCCEEDED. If so, call GetCelebrityDetection and pass the job identifier +// (JobId) from the initial call to StartCelebrityDetection. +// +// For more information, see Working with Stored Videos in the Amazon Rekognition +// Devlopers Guide. // // GetContentModeration returns detected content moderation labels, and the // time they are detected, in an array, ModerationLabels, of objects. @@ -1650,7 +1664,8 @@ func (c *Rekognition) GetContentModerationRequest(input *GetContentModerationInp // and populate the NextToken request parameter with the value of NextToken // returned from the previous call to GetContentModeration. // -// For more information, see moderation. +// For more information, see Detecting Unsafe Content in the Amazon Rekognition +// Developer Guide. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1803,15 +1818,16 @@ func (c *Rekognition) GetFaceDetectionRequest(input *GetFaceDetectionInput) (req // GetFaceDetection API operation for Amazon Rekognition. // -// Gets face detection results for a Rekognition Video analysis started by . +// Gets face detection results for a Amazon Rekognition Video analysis started +// by . // -// Face detection with Rekognition Video is an asynchronous operation. You start -// face detection by calling which returns a job identifier (JobId). When the -// face detection operation finishes, Rekognition Video publishes a completion -// status to the Amazon Simple Notification Service topic registered in the -// initial call to StartFaceDetection. To get the results of the face detection -// operation, first check that the status value published to the Amazon SNS -// topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from +// Face detection with Amazon Rekognition Video is an asynchronous operation. +// You start face detection by calling which returns a job identifier (JobId). +// When the face detection operation finishes, Amazon Rekognition Video publishes +// a completion status to the Amazon Simple Notification Service topic registered +// in the initial call to StartFaceDetection. To get the results of the face +// detection operation, first check that the status value published to the Amazon +// SNS topic is SUCCEEDED. If so, call and pass the job identifier (JobId) from // the initial call to StartFaceDetection. // // GetFaceDetection returns an array of detected faces (Faces) sorted by the @@ -1975,19 +1991,21 @@ func (c *Rekognition) GetFaceSearchRequest(input *GetFaceSearchInput) (req *requ // GetFaceSearch API operation for Amazon Rekognition. // -// Gets the face search results for Rekognition Video face search started by -// . The search returns faces in a collection that match the faces of persons +// Gets the face search results for Amazon Rekognition Video face search started +// by . The search returns faces in a collection that match the faces of persons // detected in a video. It also includes the time(s) that faces are matched // in the video. // // Face search in a video is an asynchronous operation. You start face search // by calling to which returns a job identifier (JobId). When the search operation -// finishes, Rekognition Video publishes a completion status to the Amazon Simple -// Notification Service topic registered in the initial call to StartFaceSearch. +// finishes, Amazon Rekognition Video publishes a completion status to the Amazon +// Simple Notification Service topic registered in the initial call to StartFaceSearch. // To get the search results, first check that the status value published to // the Amazon SNS topic is SUCCEEDED. If so, call GetFaceSearch and pass the -// job identifier (JobId) from the initial call to StartFaceSearch. For more -// information, see collections. +// job identifier (JobId) from the initial call to StartFaceSearch. +// +// For more information, see Searching Faces in a Collection in the Amazon Rekognition +// Developer Guide. // // The search results are retured in an array, Persons, of objects. EachPersonMatch // element contains details about the matching faces in the input collection, @@ -1997,7 +2015,7 @@ func (c *Rekognition) GetFaceSearchRequest(input *GetFaceSearchInput) (req *requ // GetFaceSearch only returns the default facial attributes (BoundingBox, Confidence, // Landmarks, Pose, and Quality). The other facial attributes listed in the // Face object of the following response syntax are not returned. For more information, -// see . +// see FaceDetail in the Amazon Rekognition Developer Guide. // // By default, the Persons array is sorted by the time, in milliseconds from // the start of the video, persons are matched. You can also sort by persons @@ -2154,7 +2172,7 @@ func (c *Rekognition) GetLabelDetectionRequest(input *GetLabelDetectionInput) (r // GetLabelDetection API operation for Amazon Rekognition. // -// Gets the label detection results of a Rekognition Video analysis started +// Gets the label detection results of a Amazon Rekognition Video analysis started // by . // // The label detection operation is started by a call to which returns a job @@ -2331,13 +2349,13 @@ func (c *Rekognition) GetPersonTrackingRequest(input *GetPersonTrackingInput) (r // GetPersonTracking API operation for Amazon Rekognition. // -// Gets the person tracking results of a Rekognition Video analysis started +// Gets the person tracking results of a Amazon Rekognition Video analysis started // by . // // The person detection operation is started by a call to StartPersonTracking // which returns a job identifier (JobId). When the person detection operation -// finishes, Rekognition Video publishes a completion status to the Amazon Simple -// Notification Service topic registered in the initial call to StartPersonTracking. +// finishes, Amazon Rekognition Video publishes a completion status to the Amazon +// Simple Notification Service topic registered in the initial call to StartPersonTracking. // // To get the results of the person tracking operation, first check that the // status value published to the Amazon SNS topic is SUCCEEDED. If so, call @@ -2348,8 +2366,10 @@ func (c *Rekognition) GetPersonTrackingRequest(input *GetPersonTrackingInput) (r // // GetPersonTracking only returns the default facial attributes (BoundingBox, // Confidence, Landmarks, Pose, and Quality). The other facial attributes listed -// in the Face object of the following response syntax are not returned. For -// more information, see . +// in the Face object of the following response syntax are not returned. +// +// For more information, see FaceDetail in the Amazon Rekognition Developer +// Guide. // // By default, the array is sorted by the time(s) a person is tracked in the // video. You can sort by tracked persons by specifying INDEX for the SortBy @@ -2519,7 +2539,10 @@ func (c *Rekognition) IndexFacesRequest(input *IndexFacesInput) (req *request.Re // the 15 largest faces in the input image. Later versions of the face detection // model index the 100 largest faces in the input image. To determine which // version of the model you are using, check the the value of FaceModelVersion -// in the response from IndexFaces. For more information, see face-detection-model. +// in the response from IndexFaces. +// +// For more information, see Model Versioning in the Amazon Rekognition Developer +// Guide. // // If you provide the optional ExternalImageID for the input image you provided, // Amazon Rekognition associates this ID with all faces that it detects. When @@ -2538,6 +2561,9 @@ func (c *Rekognition) IndexFacesRequest(input *IndexFacesInput) (req *request.Re // image, specify the same collection, and use the same external ID in the IndexFaces // operation, Amazon Rekognition doesn't save duplicate face metadata. // +// For more information, see Adding Faces to a Collection in the Amazon Rekognition +// Developer Guide. +// // The input image is passed either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must @@ -2563,7 +2589,7 @@ func (c *Rekognition) IndexFacesRequest(input *IndexFacesInput) (req *request.Re // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -2658,7 +2684,8 @@ func (c *Rekognition) ListCollectionsRequest(input *ListCollectionsInput) (req * // the response also provides a NextToken that you can use in the subsequent // request to fetch the next set of collection IDs. // -// For an example, see list-collection-procedure. +// For an example, see Listing Collections in the Amazon Rekognition Developer +// Guide. // // This operation requires permissions to perform the rekognition:ListCollections // action. @@ -2816,7 +2843,8 @@ func (c *Rekognition) ListFacesRequest(input *ListFacesInput) (req *request.Requ // // Returns metadata for faces in the specified collection. This metadata includes // information such as the bounding box coordinates, the confidence (that the -// bounding box contains a face), and face ID. For an example, see list-faces-in-collection-procedure. +// bounding box contains a face), and face ID. For an example, see Listing Faces +// in a Collection in the Amazon Rekognition Developer Guide. // // This operation requires permissions to perform the rekognition:ListFaces // action. @@ -3117,7 +3145,7 @@ func (c *Rekognition) RecognizeCelebritiesRequest(input *RecognizeCelebritiesInp // RecognizeCelebrities API operation for Amazon Rekognition. // // Returns an array of celebrities recognized in the input image. For more information, -// see celebrities. +// see Recognizing Celebrities in the Amazon Rekognition Developer Guide. // // RecognizeCelebrities returns the 100 largest faces in the image. It lists // recognized celebrities in the CelebrityFaces array and unrecognized faces @@ -3141,7 +3169,8 @@ func (c *Rekognition) RecognizeCelebritiesRequest(input *RecognizeCelebritiesInp // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // -// For an example, see celebrities-procedure-image. +// For an example, see Recognizing Celebrities in an Image in the Amazon Rekognition +// Developer Guide. // // This operation requires permissions to perform the rekognition:RecognizeCelebrities // operation. @@ -3166,7 +3195,7 @@ func (c *Rekognition) RecognizeCelebritiesRequest(input *RecognizeCelebritiesInp // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -3262,7 +3291,8 @@ func (c *Rekognition) SearchFacesRequest(input *SearchFacesInput) (req *request. // response also includes a confidence value for each face match, indicating // the confidence that the specific face matches the input face. // -// For an example, see search-face-with-id-procedure. +// For an example, see Searching for a Face Using Its Face ID in the Amazon +// Rekognition Developer Guide. // // This operation requires permissions to perform the rekognition:SearchFaces // action. @@ -3383,7 +3413,8 @@ func (c *Rekognition) SearchFacesByImageRequest(input *SearchFacesByImageInput) // confidence level that the bounding box contains a face) of the face that // Amazon Rekognition used for the input image. // -// For an example, see search-face-with-image-procedure. +// For an example, Searching for a Face Using an Image in the Amazon Rekognition +// Developer Guide. // // This operation requires permissions to perform the rekognition:SearchFacesByImage // action. @@ -3405,7 +3436,7 @@ func (c *Rekognition) SearchFacesByImageRequest(input *SearchFacesByImageInput) // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see -// limits. +// Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. @@ -3492,16 +3523,18 @@ func (c *Rekognition) StartCelebrityRecognitionRequest(input *StartCelebrityReco // // Starts asynchronous recognition of celebrities in a stored video. // -// Rekognition Video can detect celebrities in a video must be stored in an -// Amazon S3 bucket. Use Video to specify the bucket name and the filename of -// the video. StartCelebrityRecognition returns a job identifier (JobId) which -// you use to get the results of the analysis. When celebrity recognition analysis -// is finished, Rekognition Video publishes a completion status to the Amazon -// Simple Notification Service topic that you specify in NotificationChannel. +// Amazon Rekognition Video can detect celebrities in a video must be stored +// in an Amazon S3 bucket. Use Video to specify the bucket name and the filename +// of the video. StartCelebrityRecognition returns a job identifier (JobId) +// which you use to get the results of the analysis. When celebrity recognition +// analysis is finished, Amazon Rekognition Video publishes a completion status +// to the Amazon Simple Notification Service topic that you specify in NotificationChannel. // To get the results of the celebrity recognition analysis, first check that // the status value published to the Amazon SNS topic is SUCCEEDED. If so, call // and pass the job identifier (JobId) from the initial call to StartCelebrityRecognition. -// For more information, see celebrities. +// +// For more information, see Recognizing Celebrities in the Amazon Rekognition +// Developer Guide. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -3539,10 +3572,10 @@ func (c *Rekognition) StartCelebrityRecognitionRequest(input *StartCelebrityReco // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your @@ -3614,17 +3647,19 @@ func (c *Rekognition) StartContentModerationRequest(input *StartContentModeratio // Starts asynchronous detection of explicit or suggestive adult content in // a stored video. // -// Rekognition Video can moderate content in a video stored in an Amazon S3 -// bucket. Use Video to specify the bucket name and the filename of the video. +// Amazon Rekognition Video can moderate content in a video stored in an Amazon +// S3 bucket. Use Video to specify the bucket name and the filename of the video. // StartContentModeration returns a job identifier (JobId) which you use to // get the results of the analysis. When content moderation analysis is finished, -// Rekognition Video publishes a completion status to the Amazon Simple Notification -// Service topic that you specify in NotificationChannel. +// Amazon Rekognition Video publishes a completion status to the Amazon Simple +// Notification Service topic that you specify in NotificationChannel. // // To get the results of the content moderation analysis, first check that the // status value published to the Amazon SNS topic is SUCCEEDED. If so, call // and pass the job identifier (JobId) from the initial call to StartContentModeration. -// For more information, see moderation. +// +// For more information, see Detecting Unsafe Content in the Amazon Rekognition +// Developer Guide. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -3662,10 +3697,10 @@ func (c *Rekognition) StartContentModerationRequest(input *StartContentModeratio // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your @@ -3736,15 +3771,18 @@ func (c *Rekognition) StartFaceDetectionRequest(input *StartFaceDetectionInput) // // Starts asynchronous detection of faces in a stored video. // -// Rekognition Video can detect faces in a video stored in an Amazon S3 bucket. -// Use Video to specify the bucket name and the filename of the video. StartFaceDetection -// returns a job identifier (JobId) that you use to get the results of the operation. -// When face detection is finished, Rekognition Video publishes a completion -// status to the Amazon Simple Notification Service topic that you specify in -// NotificationChannel. To get the results of the label detection operation, -// first check that the status value published to the Amazon SNS topic is SUCCEEDED. -// If so, call and pass the job identifier (JobId) from the initial call to -// StartFaceDetection. For more information, see faces-video. +// Amazon Rekognition Video can detect faces in a video stored in an Amazon +// S3 bucket. Use Video to specify the bucket name and the filename of the video. +// StartFaceDetection returns a job identifier (JobId) that you use to get the +// results of the operation. When face detection is finished, Amazon Rekognition +// Video publishes a completion status to the Amazon Simple Notification Service +// topic that you specify in NotificationChannel. To get the results of the +// label detection operation, first check that the status value published to +// the Amazon SNS topic is SUCCEEDED. If so, call and pass the job identifier +// (JobId) from the initial call to StartFaceDetection. +// +// For more information, see Detecting Faces in a Stored Video in the Amazon +// Rekognition Developer Guide. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -3782,10 +3820,10 @@ func (c *Rekognition) StartFaceDetectionRequest(input *StartFaceDetectionInput) // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your @@ -3860,11 +3898,11 @@ func (c *Rekognition) StartFaceSearchRequest(input *StartFaceSearchInput) (req * // The video must be stored in an Amazon S3 bucket. Use Video to specify the // bucket name and the filename of the video. StartFaceSearch returns a job // identifier (JobId) which you use to get the search results once the search -// has completed. When searching is finished, Rekognition Video publishes a -// completion status to the Amazon Simple Notification Service topic that you -// specify in NotificationChannel. To get the search results, first check that -// the status value published to the Amazon SNS topic is SUCCEEDED. If so, call -// and pass the job identifier (JobId) from the initial call to StartFaceSearch. +// has completed. When searching is finished, Amazon Rekognition Video publishes +// a completion status to the Amazon Simple Notification Service topic that +// you specify in NotificationChannel. To get the search results, first check +// that the status value published to the Amazon SNS topic is SUCCEEDED. If +// so, call and pass the job identifier (JobId) from the initial call to StartFaceSearch. // For more information, see collections-search-person. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -3903,10 +3941,10 @@ func (c *Rekognition) StartFaceSearchRequest(input *StartFaceSearchInput) (req * // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. @@ -3980,16 +4018,16 @@ func (c *Rekognition) StartLabelDetectionRequest(input *StartLabelDetectionInput // // Starts asynchronous detection of labels in a stored video. // -// Rekognition Video can detect labels in a video. Labels are instances of real-world -// entities. This includes objects like flower, tree, and table; events like -// wedding, graduation, and birthday party; concepts like landscape, evening, -// and nature; and activities like a person getting out of a car or a person -// skiing. +// Amazon Rekognition Video can detect labels in a video. Labels are instances +// of real-world entities. This includes objects like flower, tree, and table; +// events like wedding, graduation, and birthday party; concepts like landscape, +// evening, and nature; and activities like a person getting out of a car or +// a person skiing. // // The video must be stored in an Amazon S3 bucket. Use Video to specify the // bucket name and the filename of the video. StartLabelDetection returns a // job identifier (JobId) which you use to get the results of the operation. -// When label detection is finished, Rekognition Video publishes a completion +// When label detection is finished, Amazon Rekognition Video publishes a completion // status to the Amazon Simple Notification Service topic that you specify in // NotificationChannel. // @@ -4033,10 +4071,10 @@ func (c *Rekognition) StartLabelDetectionRequest(input *StartLabelDetectionInput // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your @@ -4107,12 +4145,12 @@ func (c *Rekognition) StartPersonTrackingRequest(input *StartPersonTrackingInput // // Starts the asynchronous tracking of persons in a stored video. // -// Rekognition Video can track persons in a video stored in an Amazon S3 bucket. -// Use Video to specify the bucket name and the filename of the video. StartPersonTracking -// returns a job identifier (JobId) which you use to get the results of the -// operation. When label detection is finished, Amazon Rekognition publishes -// a completion status to the Amazon Simple Notification Service topic that -// you specify in NotificationChannel. +// Amazon Rekognition Video can track persons in a video stored in an Amazon +// S3 bucket. Use Video to specify the bucket name and the filename of the video. +// StartPersonTracking returns a job identifier (JobId) which you use to get +// the results of the operation. When label detection is finished, Amazon Rekognition +// publishes a completion status to the Amazon Simple Notification Service topic +// that you specify in NotificationChannel. // // To get the results of the person detection operation, first check that the // status value published to the Amazon SNS topic is SUCCEEDED. If so, call @@ -4154,10 +4192,10 @@ func (c *Rekognition) StartPersonTrackingRequest(input *StartPersonTrackingInput // // * ErrCodeLimitExceededException "LimitExceededException" // An Amazon Rekognition service limit was exceeded. For example, if you start -// too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, -// for example) will raise a LimitExceededException exception (HTTP status code: -// 400) until the number of concurrently running jobs is below the Amazon Rekognition -// service limit. +// too many Amazon Rekognition Video jobs concurrently, calls to start operations +// (StartLabelDetection, for example) will raise a LimitExceededException exception +// (HTTP status code: 400) until the number of concurrently running jobs is +// below the Amazon Rekognition service limit. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your @@ -4654,7 +4692,8 @@ func (s *CelebrityDetail) SetUrls(v []*string) *CelebrityDetail { } // Information about a detected celebrity and the time the celebrity was detected -// in a stored video. For more information, see . +// in a stored video. For more information, see GetCelebrityRecognition in the +// Amazon Rekognition Developer Guide. type CelebrityRecognition struct { _ struct{} `type:"structure"` @@ -5116,8 +5155,8 @@ type CreateStreamProcessorInput struct { // Name is a required field Name *string `min:"1" type:"string" required:"true"` - // Kinesis data stream stream to which Rekognition Video puts the analysis results. - // If you are using the AWS CLI, the parameter name is StreamProcessorOutput. + // Kinesis data stream stream to which Amazon Rekognition Video puts the analysis + // results. If you are using the AWS CLI, the parameter name is StreamProcessorOutput. // // Output is a required field Output *StreamProcessorOutput `type:"structure" required:"true"` @@ -5488,7 +5527,7 @@ type DescribeStreamProcessorOutput struct { // Name of the stream processor. Name *string `min:"1" type:"string"` - // Kinesis data stream to which Rekognition Video puts the analysis results. + // Kinesis data stream to which Amazon Rekognition Video puts the analysis results. Output *StreamProcessorOutput `type:"structure"` // ARN of the IAM role that allows access to the stream processor. @@ -6110,11 +6149,11 @@ func (s *Face) SetImageId(v string) *Face { // facial attributes. The default attributes are BoundingBox, Confidence, Landmarks, // Pose, and Quality. // -// is the only Rekognition Video stored video operation that can return a FaceDetail -// object with all attributes. To specify which attributes to return, use the -// FaceAttributes input parameter for . The following Rekognition Video operations -// return only the default attributes. The corresponding Start operations don't -// have a FaceAttributes input parameter. +// is the only Amazon Rekognition Video stored video operation that can return +// a FaceDetail object with all attributes. To specify which attributes to return, +// use the FaceAttributes input parameter for . The following Amazon Rekognition +// Video operations return only the default attributes. The corresponding Start +// operations don't have a FaceAttributes input parameter. // // * GetCelebrityRecognition // @@ -6122,9 +6161,9 @@ func (s *Face) SetImageId(v string) *Face { // // * GetFaceSearch // -// The Rekognition Image and operations can return all facial attributes. To -// specify which attributes to return, use the Attributes input parameter for -// DetectFaces. For IndexFaces, use the DetectAttributes input parameter. +// The Amazon Rekognition Image and operations can return all facial attributes. +// To specify which attributes to return, use the Attributes input parameter +// for DetectFaces. For IndexFaces, use the DetectAttributes input parameter. type FaceDetail struct { _ struct{} `type:"structure"` @@ -6591,8 +6630,8 @@ type GetCelebrityRecognitionInput struct { MaxResults *int64 `min:"1" type:"integer"` // If the previous response was incomplete (because there is more recognized - // celebrities to retrieve), Rekognition Video returns a pagination token in - // the response. You can use this pagination token to retrieve the next set + // celebrities to retrieve), Amazon Rekognition Video returns a pagination token + // in the response. You can use this pagination token to retrieve the next set // of celebrities. NextToken *string `type:"string"` @@ -6664,16 +6703,16 @@ type GetCelebrityRecognitionOutput struct { // The current status of the celebrity recognition job. JobStatus *string `type:"string" enum:"VideoJobStatus"` - // If the response is truncated, Rekognition Video returns this token that you - // can use in the subsequent request to retrieve the next set of celebrities. + // If the response is truncated, Amazon Rekognition Video returns this token + // that you can use in the subsequent request to retrieve the next set of celebrities. NextToken *string `type:"string"` // If the job fails, StatusMessage provides a descriptive error message. StatusMessage *string `type:"string"` - // Information about a video that Rekognition Video analyzed. Videometadata - // is returned in every page of paginated responses from a Rekognition Video - // operation. + // Information about a video that Amazon Rekognition Video analyzed. Videometadata + // is returned in every page of paginated responses from a Amazon Rekognition + // Video operation. VideoMetadata *VideoMetadata `type:"structure"` } @@ -6805,8 +6844,8 @@ type GetContentModerationOutput struct { // The detected moderation labels and the time(s) they were detected. ModerationLabels []*ContentModerationDetection `type:"list"` - // If the response is truncated, Rekognition Video returns this token that you - // can use in the subsequent request to retrieve the next set of moderation + // If the response is truncated, Amazon Rekognition Video returns this token + // that you can use in the subsequent request to retrieve the next set of moderation // labels. NextToken *string `type:"string"` @@ -6873,7 +6912,7 @@ type GetFaceDetectionInput struct { MaxResults *int64 `min:"1" type:"integer"` // If the previous response was incomplete (because there are more faces to - // retrieve), Rekognition Video returns a pagination token in the response. + // retrieve), Amazon Rekognition Video returns a pagination token in the response. // You can use this pagination token to retrieve the next set of faces. NextToken *string `type:"string"` } @@ -6943,7 +6982,7 @@ type GetFaceDetectionOutput struct { // If the job fails, StatusMessage provides a descriptive error message. StatusMessage *string `type:"string"` - // Information about a video that Rekognition Video analyzed. Videometadata + // Information about a video that Amazon Rekognition Video analyzed. Videometadata // is returned in every page of paginated responses from a Amazon Rekognition // video operation. VideoMetadata *VideoMetadata `type:"structure"` @@ -7004,8 +7043,9 @@ type GetFaceSearchInput struct { MaxResults *int64 `min:"1" type:"integer"` // If the previous response was incomplete (because there is more search results - // to retrieve), Rekognition Video returns a pagination token in the response. - // You can use this pagination token to retrieve the next set of search results. + // to retrieve), Amazon Rekognition Video returns a pagination token in the + // response. You can use this pagination token to retrieve the next set of search + // results. NextToken *string `type:"string"` // Sort to use for grouping faces in the response. Use TIMESTAMP to group faces @@ -7072,8 +7112,9 @@ type GetFaceSearchOutput struct { // The current status of the face search job. JobStatus *string `type:"string" enum:"VideoJobStatus"` - // If the response is truncated, Rekognition Video returns this token that you - // can use in the subsequent request to retrieve the next set of search results. + // If the response is truncated, Amazon Rekognition Video returns this token + // that you can use in the subsequent request to retrieve the next set of search + // results. NextToken *string `type:"string"` // An array of persons, , in the video whose face(s) match the face(s) in an @@ -7088,8 +7129,8 @@ type GetFaceSearchOutput struct { StatusMessage *string `type:"string"` // Information about a video that Amazon Rekognition analyzed. Videometadata - // is returned in every page of paginated responses from a Rekognition Video - // operation. + // is returned in every page of paginated responses from a Amazon Rekognition + // Video operation. VideoMetadata *VideoMetadata `type:"structure"` } @@ -7148,7 +7189,7 @@ type GetLabelDetectionInput struct { MaxResults *int64 `min:"1" type:"integer"` // If the previous response was incomplete (because there are more labels to - // retrieve), Rekognition Video returns a pagination token in the response. + // retrieve), Amazon Rekognition Video returns a pagination token in the response. // You can use this pagination token to retrieve the next set of labels. NextToken *string `type:"string"` @@ -7223,14 +7264,14 @@ type GetLabelDetectionOutput struct { // label was detected. Labels []*LabelDetection `type:"list"` - // If the response is truncated, Rekognition Video returns this token that you - // can use in the subsequent request to retrieve the next set of labels. + // If the response is truncated, Amazon Rekognition Video returns this token + // that you can use in the subsequent request to retrieve the next set of labels. NextToken *string `type:"string"` // If the job fails, StatusMessage provides a descriptive error message. StatusMessage *string `type:"string"` - // Information about a video that Rekognition Video analyzed. Videometadata + // Information about a video that Amazon Rekognition Video analyzed. Videometadata // is returned in every page of paginated responses from a Amazon Rekognition // video operation. VideoMetadata *VideoMetadata `type:"structure"` @@ -7291,7 +7332,7 @@ type GetPersonTrackingInput struct { MaxResults *int64 `min:"1" type:"integer"` // If the previous response was incomplete (because there are more persons to - // retrieve), Rekognition Video returns a pagination token in the response. + // retrieve), Amazon Rekognition Video returns a pagination token in the response. // You can use this pagination token to retrieve the next set of persons. NextToken *string `type:"string"` @@ -7361,8 +7402,8 @@ type GetPersonTrackingOutput struct { // The current status of the person tracking job. JobStatus *string `type:"string" enum:"VideoJobStatus"` - // If the response is truncated, Rekognition Video returns this token that you - // can use in the subsequent request to retrieve the next set of persons. + // If the response is truncated, Amazon Rekognition Video returns this token + // that you can use in the subsequent request to retrieve the next set of persons. NextToken *string `type:"string"` // An array of the persons detected in the video and the times they are tracked @@ -7373,9 +7414,9 @@ type GetPersonTrackingOutput struct { // If the job fails, StatusMessage provides a descriptive error message. StatusMessage *string `type:"string"` - // Information about a video that Rekognition Video analyzed. Videometadata - // is returned in every page of paginated responses from a Rekognition Video - // operation. + // Information about a video that Amazon Rekognition Video analyzed. Videometadata + // is returned in every page of paginated responses from a Amazon Rekognition + // Video operation. VideoMetadata *VideoMetadata `type:"structure"` } @@ -7425,8 +7466,10 @@ func (s *GetPersonTrackingOutput) SetVideoMetadata(v *VideoMetadata) *GetPersonT // For example, you would use the Bytes property to pass an image loaded from // a local file system. Image bytes passed by using the Bytes property must // be base64-encoded. Your code may not need to encode image bytes if you are -// using an AWS SDK to call Rekognition API operations. For more information, -// see images-bytes. +// using an AWS SDK to call Rekognition API operations. +// +// For more information, see Analyzing an Image Loaded from a Local File System +// in the Amazon Rekognition Developer Guide. // // You pass images stored in an S3 bucket to a Rekognition API operation by // using the S3Object property. Images stored in an S3 bucket do not need to @@ -7441,7 +7484,8 @@ func (s *GetPersonTrackingOutput) SetVideoMetadata(v *VideoMetadata) *GetPersonT // property. // // For Amazon Rekognition to process an S3 object, the user must have permission -// to access the S3 object. For more information, see manage-access-resource-policies. +// to access the S3 object. For more information, see Resource Based Policies +// in the Amazon Rekognition Developer Guide. type Image struct { _ struct{} `type:"structure"` @@ -7629,7 +7673,7 @@ type IndexFacesOutput struct { FaceModelVersion *string `type:"string"` // An array of faces detected and added to the collection. For more information, - // see collections-index-faces. + // see Searching Faces in a Collection in the Amazon Rekognition Developer Guide. FaceRecords []*FaceRecord `type:"list"` // The orientation of the input image (counterclockwise direction). If your @@ -7675,7 +7719,7 @@ func (s *IndexFacesOutput) SetOrientationCorrection(v string) *IndexFacesOutput // The Kinesis data stream Amazon Rekognition to which the analysis results // of a Amazon Rekognition stream processor are streamed. For more information, -// see . +// see CreateStreamProcessor in the Amazon Rekognition Developer Guide. type KinesisDataStream struct { _ struct{} `type:"structure"` @@ -7700,7 +7744,8 @@ func (s *KinesisDataStream) SetArn(v string) *KinesisDataStream { } // Kinesis video stream stream that provides the source streaming video for -// a Rekognition Video stream processor. For more information, see . +// a Amazon Rekognition Video stream processor. For more information, see CreateStreamProcessor +// in the Amazon Rekognition Developer Guide. type KinesisVideoStream struct { _ struct{} `type:"structure"` @@ -8022,13 +8067,14 @@ func (s *ListFacesOutput) SetNextToken(v string) *ListFacesOutput { type ListStreamProcessorsInput struct { _ struct{} `type:"structure"` - // Maximum number of stream processors you want Rekognition Video to return - // in the response. The default is 1000. + // Maximum number of stream processors you want Amazon Rekognition Video to + // return in the response. The default is 1000. MaxResults *int64 `min:"1" type:"integer"` // If the previous response was incomplete (because there are more stream processors - // to retrieve), Rekognition Video returns a pagination token in the response. - // You can use this pagination token to retrieve the next set of stream processors. + // to retrieve), Amazon Rekognition Video returns a pagination token in the + // response. You can use this pagination token to retrieve the next set of stream + // processors. NextToken *string `type:"string"` } @@ -8070,8 +8116,9 @@ func (s *ListStreamProcessorsInput) SetNextToken(v string) *ListStreamProcessors type ListStreamProcessorsOutput struct { _ struct{} `type:"structure"` - // If the response is truncated, Rekognition Video returns this token that you - // can use in the subsequent request to retrieve the next set of stream processors. + // If the response is truncated, Amazon Rekognition Video returns this token + // that you can use in the subsequent request to retrieve the next set of stream + // processors. NextToken *string `type:"string"` // List of stream processors that you have created. @@ -8102,7 +8149,8 @@ func (s *ListStreamProcessorsOutput) SetStreamProcessors(v []*StreamProcessor) * // Provides information about a single type of moderated content found in an // image or video. Each type of moderated content has a label within a hierarchical -// taxonomy. For more information, see moderation. +// taxonomy. For more information, see Detecting Unsafe Content in the Amazon +// Rekognition Developer Guide. type ModerationLabel struct { _ struct{} `type:"structure"` @@ -8322,7 +8370,10 @@ func (s *PersonDetail) SetIndex(v int64) *PersonDetail { // Details and tracking information for a single time a person is tracked in // a video. Amazon Rekognition operations that track persons return an array // of PersonDetection objects with elements for each time a person is tracked -// in a video. For more information, see . +// in a video. +// +// For more information, see API_GetPersonTracking in the Amazon Rekognition +// Developer Guide. type PersonDetection struct { _ struct{} `type:"structure"` @@ -8410,7 +8461,8 @@ func (s *PersonMatch) SetTimestamp(v int64) *PersonMatch { // the (350,50) pixel coordinate on the image. // // An array of Point objects, Polygon, is returned by . Polygon represents a -// fine-grained polygon around detected text. For more information, see . +// fine-grained polygon around detected text. For more information, see Geometry +// in the Amazon Rekognition Developer Guide. type Point struct { _ struct{} `type:"structure"` @@ -8589,7 +8641,8 @@ func (s *RecognizeCelebritiesOutput) SetUnrecognizedFaces(v []*ComparedFace) *Re // you use for Amazon Rekognition operations. // // For Amazon Rekognition to process an S3 object, the user must have permission -// to access the S3 object. For more information, see manage-access-resource-policies. +// to access the S3 object. For more information, see Resource Based Policies +// in the Amazon Rekognition Developer Guide. type S3Object struct { _ struct{} `type:"structure"` @@ -8958,8 +9011,8 @@ type StartCelebrityRecognitionInput struct { // published to the Amazon Simple Notification Service topic. JobTag *string `min:"1" type:"string"` - // The Amazon SNS topic ARN that you want Rekognition Video to publish the completion - // status of the celebrity recognition analysis to. + // The Amazon SNS topic ARN that you want Amazon Rekognition Video to publish + // the completion status of the celebrity recognition analysis to. NotificationChannel *NotificationChannel `type:"structure"` // The video in which you want to recognize celebrities. The video must be stored @@ -9077,8 +9130,8 @@ type StartContentModerationInput struct { // specified value. MinConfidence *float64 `type:"float"` - // The Amazon SNS topic ARN that you want Rekognition Video to publish the completion - // status of the content moderation analysis to. + // The Amazon SNS topic ARN that you want Amazon Rekognition Video to publish + // the completion status of the content moderation analysis to. NotificationChannel *NotificationChannel `type:"structure"` // The video in which you want to moderate content. The video must be stored @@ -9202,8 +9255,8 @@ type StartFaceDetectionInput struct { // published to the Amazon Simple Notification Service topic. JobTag *string `min:"1" type:"string"` - // The ARN of the Amazon SNS topic to which you want Rekognition Video to publish - // the completion status of the face detection operation. + // The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video + // to publish the completion status of the face detection operation. NotificationChannel *NotificationChannel `type:"structure"` // The video in which you want to detect faces. The video must be stored in @@ -9328,8 +9381,8 @@ type StartFaceSearchInput struct { // published to the Amazon Simple Notification Service topic. JobTag *string `min:"1" type:"string"` - // The ARN of the Amazon SNS topic to which you want Rekognition Video to publish - // the completion status of the search. + // The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video + // to publish the completion status of the search. NotificationChannel *NotificationChannel `type:"structure"` // The video you want to search. The video must be stored in an Amazon S3 bucket. @@ -9456,18 +9509,18 @@ type StartLabelDetectionInput struct { // published to the Amazon Simple Notification Service topic. JobTag *string `min:"1" type:"string"` - // Specifies the minimum confidence that Rekognition Video must have in order - // to return a detected label. Confidence represents how certain Amazon Rekognition - // is that a label is correctly identified.0 is the lowest confidence. 100 is - // the highest confidence. Rekognition Video doesn't return any labels with - // a confidence level lower than this specified value. + // Specifies the minimum confidence that Amazon Rekognition Video must have + // in order to return a detected label. Confidence represents how certain Amazon + // Rekognition is that a label is correctly identified.0 is the lowest confidence. + // 100 is the highest confidence. Amazon Rekognition Video doesn't return any + // labels with a confidence level lower than this specified value. // // If you don't specify MinConfidence, the operation returns labels with confidence // values greater than or equal to 50 percent. MinConfidence *float64 `type:"float"` - // The Amazon SNS topic ARN you want Rekognition Video to publish the completion - // status of the label detection operation to. + // The Amazon SNS topic ARN you want Amazon Rekognition Video to publish the + // completion status of the label detection operation to. NotificationChannel *NotificationChannel `type:"structure"` // The video in which you want to detect labels. The video must be stored in @@ -9583,8 +9636,8 @@ type StartPersonTrackingInput struct { // published to the Amazon Simple Notification Service topic. JobTag *string `min:"1" type:"string"` - // The Amazon SNS topic ARN you want Rekognition Video to publish the completion - // status of the people detection operation to. + // The Amazon SNS topic ARN you want Amazon Rekognition Video to publish the + // completion status of the people detection operation to. NotificationChannel *NotificationChannel `type:"structure"` // The video in which you want to detect people. The video must be stored in @@ -9851,9 +9904,10 @@ func (s *StreamProcessorInput) SetKinesisVideoStream(v *KinesisVideoStream) *Str return s } -// Information about the Amazon Kinesis Data Streams stream to which a Rekognition -// Video stream processor streams the results of a video analysis. For more -// information, see . +// Information about the Amazon Kinesis Data Streams stream to which a Amazon +// Rekognition Video stream processor streams the results of a video analysis. +// For more information, see CreateStreamProcessor in the Amazon Rekognition +// Developer Guide. type StreamProcessorOutput struct { _ struct{} `type:"structure"` @@ -9962,7 +10016,8 @@ func (s *Sunglasses) SetValue(v bool) *Sunglasses { // the word appears. The word Id is also an index for the word within a line // of words. // -// For more information, see text-detection. +// For more information, see Detecting Text in the Amazon Rekognition Developer +// Guide. type TextDetection struct { _ struct{} `type:"structure"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/rekognition/errors.go b/vendor/github.com/aws/aws-sdk-go/service/rekognition/errors.go index 75c4bbacc..8d4d4325c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rekognition/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rekognition/errors.go @@ -22,7 +22,7 @@ const ( // "ImageTooLargeException". // // The input image size exceeds the allowed limit. For more information, see - // limits. + // Limits in Amazon Rekognition in the Amazon Rekognition Developer Guide. ErrCodeImageTooLargeException = "ImageTooLargeException" // ErrCodeInternalServerError for service response error code @@ -60,10 +60,10 @@ const ( // "LimitExceededException". // // An Amazon Rekognition service limit was exceeded. For example, if you start - // too many Rekognition Video jobs concurrently, calls to start operations (StartLabelDetection, - // for example) will raise a LimitExceededException exception (HTTP status code: - // 400) until the number of concurrently running jobs is below the Amazon Rekognition - // service limit. + // too many Amazon Rekognition Video jobs concurrently, calls to start operations + // (StartLabelDetection, for example) will raise a LimitExceededException exception + // (HTTP status code: 400) until the number of concurrently running jobs is + // below the Amazon Rekognition service limit. ErrCodeLimitExceededException = "LimitExceededException" // ErrCodeProvisionedThroughputExceededException for service response error code diff --git a/vendor/github.com/aws/aws-sdk-go/service/rekognition/service.go b/vendor/github.com/aws/aws-sdk-go/service/rekognition/service.go index 97f89579f..435c25636 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/rekognition/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/rekognition/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "rekognition" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "rekognition" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Rekognition" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Rekognition client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/resourcegroups/service.go b/vendor/github.com/aws/aws-sdk-go/service/resourcegroups/service.go index bcff04e00..46a19ff23 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/resourcegroups/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/resourcegroups/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "resource-groups" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "resource-groups" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Resource Groups" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ResourceGroups client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/resourcegroupstaggingapi/service.go b/vendor/github.com/aws/aws-sdk-go/service/resourcegroupstaggingapi/service.go index 554a10d1e..705db6b0d 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/resourcegroupstaggingapi/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/resourcegroupstaggingapi/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "tagging" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "tagging" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Resource Groups Tagging API" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ResourceGroupsTaggingAPI client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/route53/service.go b/vendor/github.com/aws/aws-sdk-go/service/route53/service.go index 98ba1c8f8..dd22cb2cd 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/route53/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/route53/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "route53" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "route53" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Route 53" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Route53 client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/route53domains/service.go b/vendor/github.com/aws/aws-sdk-go/service/route53domains/service.go index 299c0770c..21629dea0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/route53domains/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/route53domains/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "route53domains" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "route53domains" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Route 53 Domains" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Route53Domains client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/s3/api.go b/vendor/github.com/aws/aws-sdk-go/service/s3/api.go index a27823fdf..f83dcb8ec 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/s3/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/s3/api.go @@ -3,14 +3,21 @@ package s3 import ( + "bytes" "fmt" "io" + "sync" + "sync/atomic" "time" "github.com/aws/aws-sdk-go/aws" "github.com/aws/aws-sdk-go/aws/awsutil" + "github.com/aws/aws-sdk-go/aws/client" "github.com/aws/aws-sdk-go/aws/request" "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/eventstream" + "github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi" + "github.com/aws/aws-sdk-go/private/protocol/rest" "github.com/aws/aws-sdk-go/private/protocol/restxml" ) @@ -6017,6 +6024,88 @@ func (c *S3) RestoreObjectWithContext(ctx aws.Context, input *RestoreObjectInput return out, req.Send() } +const opSelectObjectContent = "SelectObjectContent" + +// SelectObjectContentRequest generates a "aws/request.Request" representing the +// client's request for the SelectObjectContent operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See SelectObjectContent for more information on using the SelectObjectContent +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the SelectObjectContentRequest method. +// req, resp := client.SelectObjectContentRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/s3-2006-03-01/SelectObjectContent +func (c *S3) SelectObjectContentRequest(input *SelectObjectContentInput) (req *request.Request, output *SelectObjectContentOutput) { + op := &request.Operation{ + Name: opSelectObjectContent, + HTTPMethod: "POST", + HTTPPath: "/{Bucket}/{Key+}?select&select-type=2", + } + + if input == nil { + input = &SelectObjectContentInput{} + } + + output = &SelectObjectContentOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Send.Swap(client.LogHTTPResponseHandler.Name, client.LogHTTPResponseHeaderHandler) + req.Handlers.Unmarshal.Swap(restxml.UnmarshalHandler.Name, rest.UnmarshalHandler) + req.Handlers.Unmarshal.PushBack(output.runEventStreamLoop) + return +} + +// SelectObjectContent API operation for Amazon Simple Storage Service. +// +// This operation filters the contents of an Amazon S3 object based on a simple +// Structured Query Language (SQL) statement. In the request, along with the +// SQL expression, you must also specify a data serialization format (JSON or +// CSV) of the object. Amazon S3 uses this to parse object data into records, +// and returns only records that match the specified SQL expression. You must +// also specify the data serialization format for the response. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Simple Storage Service's +// API operation SelectObjectContent for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/s3-2006-03-01/SelectObjectContent +func (c *S3) SelectObjectContent(input *SelectObjectContentInput) (*SelectObjectContentOutput, error) { + req, out := c.SelectObjectContentRequest(input) + return out, req.Send() +} + +// SelectObjectContentWithContext is the same as SelectObjectContent with the addition of +// the ability to pass a context and additional request options. +// +// See SelectObjectContent for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *S3) SelectObjectContentWithContext(ctx aws.Context, input *SelectObjectContentInput, opts ...request.Option) (*SelectObjectContentOutput, error) { + req, out := c.SelectObjectContentRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opUploadPart = "UploadPart" // UploadPartRequest generates a "aws/request.Request" representing the @@ -6977,6 +7066,11 @@ func (s *CORSRule) SetMaxAgeSeconds(v int64) *CORSRule { type CSVInput struct { _ struct{} `type:"structure"` + // Specifies that CSV field values may contain quoted record delimiters and + // such records should be allowed. Default value is FALSE. Setting this value + // to TRUE may lower performance. + AllowQuotedRecordDelimiter *bool `type:"boolean"` + // Single character used to indicate a row should be ignored when present at // the start of a row. Comments *string `type:"string"` @@ -7008,6 +7102,12 @@ func (s CSVInput) GoString() string { return s.String() } +// SetAllowQuotedRecordDelimiter sets the AllowQuotedRecordDelimiter field's value. +func (s *CSVInput) SetAllowQuotedRecordDelimiter(v bool) *CSVInput { + s.AllowQuotedRecordDelimiter = &v + return s +} + // SetComments sets the Comments field's value. func (s *CSVInput) SetComments(v string) *CSVInput { s.Comments = &v @@ -7474,6 +7574,32 @@ func (s *Condition) SetKeyPrefixEquals(v string) *Condition { return s } +type ContinuationEvent struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s ContinuationEvent) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ContinuationEvent) GoString() string { + return s.String() +} + +// The ContinuationEvent is and event in the SelectObjectContentEventStream group of events. +func (s *ContinuationEvent) eventSelectObjectContentEventStream() {} + +// UnmarshalEvent unmarshals the EventStream Message into the ContinuationEvent value. +// This method is only used internally within the SDK's EventStream handling. +func (s *ContinuationEvent) UnmarshalEvent( + payloadUnmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + return nil +} + type CopyObjectInput struct { _ struct{} `type:"structure"` @@ -9919,6 +10045,32 @@ func (s *EncryptionConfiguration) SetReplicaKmsKeyID(v string) *EncryptionConfig return s } +type EndEvent struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s EndEvent) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EndEvent) GoString() string { + return s.String() +} + +// The EndEvent is and event in the SelectObjectContentEventStream group of events. +func (s *EndEvent) eventSelectObjectContentEventStream() {} + +// UnmarshalEvent unmarshals the EventStream Message into the EndEvent value. +// This method is only used internally within the SDK's EventStream handling. +func (s *EndEvent) UnmarshalEvent( + payloadUnmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + return nil +} + type Error struct { _ struct{} `type:"structure"` @@ -16380,6 +16532,87 @@ func (s *Part) SetSize(v int64) *Part { return s } +type Progress struct { + _ struct{} `type:"structure"` + + // Current number of uncompressed object bytes processed. + BytesProcessed *int64 `type:"long"` + + // Current number of bytes of records payload data returned. + BytesReturned *int64 `type:"long"` + + // Current number of object bytes scanned. + BytesScanned *int64 `type:"long"` +} + +// String returns the string representation +func (s Progress) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Progress) GoString() string { + return s.String() +} + +// SetBytesProcessed sets the BytesProcessed field's value. +func (s *Progress) SetBytesProcessed(v int64) *Progress { + s.BytesProcessed = &v + return s +} + +// SetBytesReturned sets the BytesReturned field's value. +func (s *Progress) SetBytesReturned(v int64) *Progress { + s.BytesReturned = &v + return s +} + +// SetBytesScanned sets the BytesScanned field's value. +func (s *Progress) SetBytesScanned(v int64) *Progress { + s.BytesScanned = &v + return s +} + +type ProgressEvent struct { + _ struct{} `type:"structure" payload:"Details"` + + // The Progress event details. + Details *Progress `locationName:"Details" type:"structure"` +} + +// String returns the string representation +func (s ProgressEvent) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ProgressEvent) GoString() string { + return s.String() +} + +// SetDetails sets the Details field's value. +func (s *ProgressEvent) SetDetails(v *Progress) *ProgressEvent { + s.Details = v + return s +} + +// The ProgressEvent is and event in the SelectObjectContentEventStream group of events. +func (s *ProgressEvent) eventSelectObjectContentEventStream() {} + +// UnmarshalEvent unmarshals the EventStream Message into the ProgressEvent value. +// This method is only used internally within the SDK's EventStream handling. +func (s *ProgressEvent) UnmarshalEvent( + payloadUnmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + if err := payloadUnmarshaler.UnmarshalPayload( + bytes.NewReader(msg.Payload), s, + ); err != nil { + return fmt.Errorf("failed to unmarshal payload, %v", err) + } + return nil +} + type PutBucketAccelerateConfigurationInput struct { _ struct{} `type:"structure" payload:"AccelerateConfiguration"` @@ -18622,6 +18855,45 @@ func (s *QueueConfigurationDeprecated) SetQueue(v string) *QueueConfigurationDep return s } +type RecordsEvent struct { + _ struct{} `type:"structure" payload:"Payload"` + + // The byte array of partial, one or more result records. + // + // Payload is automatically base64 encoded/decoded by the SDK. + Payload []byte `type:"blob"` +} + +// String returns the string representation +func (s RecordsEvent) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RecordsEvent) GoString() string { + return s.String() +} + +// SetPayload sets the Payload field's value. +func (s *RecordsEvent) SetPayload(v []byte) *RecordsEvent { + s.Payload = v + return s +} + +// The RecordsEvent is and event in the SelectObjectContentEventStream group of events. +func (s *RecordsEvent) eventSelectObjectContentEventStream() {} + +// UnmarshalEvent unmarshals the EventStream Message into the RecordsEvent value. +// This method is only used internally within the SDK's EventStream handling. +func (s *RecordsEvent) UnmarshalEvent( + payloadUnmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + s.Payload = make([]byte, len(msg.Payload)) + copy(s.Payload, msg.Payload) + return nil +} + type Redirect struct { _ struct{} `type:"structure"` @@ -18939,6 +19211,30 @@ func (s *RequestPaymentConfiguration) SetPayer(v string) *RequestPaymentConfigur return s } +type RequestProgress struct { + _ struct{} `type:"structure"` + + // Specifies whether periodic QueryProgress frames should be sent. Valid values: + // TRUE, FALSE. Default value: FALSE. + Enabled *bool `type:"boolean"` +} + +// String returns the string representation +func (s RequestProgress) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s RequestProgress) GoString() string { + return s.String() +} + +// SetEnabled sets the Enabled field's value. +func (s *RequestProgress) SetEnabled(v bool) *RequestProgress { + s.Enabled = &v + return s +} + type RestoreObjectInput struct { _ struct{} `type:"structure" payload:"RestoreRequest"` @@ -19392,6 +19688,436 @@ func (s SSES3) GoString() string { return s.String() } +// SelectObjectContentEventStream provides handling of EventStreams for +// the SelectObjectContent API. +// +// Use this type to receive SelectObjectContentEventStream events. The events +// can be read from the Events channel member. +// +// The events that can be received are: +// +// * ContinuationEvent +// * EndEvent +// * ProgressEvent +// * RecordsEvent +// * StatsEvent +type SelectObjectContentEventStream struct { + // Reader is the EventStream reader for the SelectObjectContentEventStream + // events. This value is automatically set by the SDK when the API call is made + // Use this member when unit testing your code with the SDK to mock out the + // EventStream Reader. + // + // Must not be nil. + Reader SelectObjectContentEventStreamReader + + // StreamCloser is the io.Closer for the EventStream connection. For HTTP + // EventStream this is the response Body. The stream will be closed when + // the Close method of the EventStream is called. + StreamCloser io.Closer +} + +// Close closes the EventStream. This will also cause the Events channel to be +// closed. You can use the closing of the Events channel to terminate your +// application's read from the API's EventStream. +// +// Will close the underlying EventStream reader. For EventStream over HTTP +// connection this will also close the HTTP connection. +// +// Close must be called when done using the EventStream API. Not calling Close +// may result in resource leaks. +func (es *SelectObjectContentEventStream) Close() (err error) { + es.Reader.Close() + return es.Err() +} + +// Err returns any error that occurred while reading EventStream Events from +// the service API's response. Returns nil if there were no errors. +func (es *SelectObjectContentEventStream) Err() error { + if err := es.Reader.Err(); err != nil { + return err + } + es.StreamCloser.Close() + + return nil +} + +// Events returns a channel to read EventStream Events from the +// SelectObjectContent API. +// +// These events are: +// +// * ContinuationEvent +// * EndEvent +// * ProgressEvent +// * RecordsEvent +// * StatsEvent +func (es *SelectObjectContentEventStream) Events() <-chan SelectObjectContentEventStreamEvent { + return es.Reader.Events() +} + +// SelectObjectContentEventStreamEvent groups together all EventStream +// events read from the SelectObjectContent API. +// +// These events are: +// +// * ContinuationEvent +// * EndEvent +// * ProgressEvent +// * RecordsEvent +// * StatsEvent +type SelectObjectContentEventStreamEvent interface { + eventSelectObjectContentEventStream() +} + +// SelectObjectContentEventStreamReader provides the interface for reading EventStream +// Events from the SelectObjectContent API. The +// default implementation for this interface will be SelectObjectContentEventStream. +// +// The reader's Close method must allow multiple concurrent calls. +// +// These events are: +// +// * ContinuationEvent +// * EndEvent +// * ProgressEvent +// * RecordsEvent +// * StatsEvent +type SelectObjectContentEventStreamReader interface { + // Returns a channel of events as they are read from the event stream. + Events() <-chan SelectObjectContentEventStreamEvent + + // Close will close the underlying event stream reader. For event stream over + // HTTP this will also close the HTTP connection. + Close() error + + // Returns any error that has occured while reading from the event stream. + Err() error +} + +type readSelectObjectContentEventStream struct { + eventReader *eventstreamapi.EventReader + stream chan SelectObjectContentEventStreamEvent + errVal atomic.Value + + done chan struct{} + closeOnce sync.Once +} + +func newReadSelectObjectContentEventStream( + reader io.ReadCloser, + unmarshalers request.HandlerList, + logger aws.Logger, + logLevel aws.LogLevelType, +) *readSelectObjectContentEventStream { + r := &readSelectObjectContentEventStream{ + stream: make(chan SelectObjectContentEventStreamEvent), + done: make(chan struct{}), + } + + r.eventReader = eventstreamapi.NewEventReader( + reader, + protocol.HandlerPayloadUnmarshal{ + Unmarshalers: unmarshalers, + }, + r.unmarshalerForEventType, + ) + r.eventReader.UseLogger(logger, logLevel) + + return r +} + +// Close will close the underlying event stream reader. For EventStream over +// HTTP this will also close the HTTP connection. +func (r *readSelectObjectContentEventStream) Close() error { + r.closeOnce.Do(r.safeClose) + + return r.Err() +} + +func (r *readSelectObjectContentEventStream) safeClose() { + close(r.done) + err := r.eventReader.Close() + if err != nil { + r.errVal.Store(err) + } +} + +func (r *readSelectObjectContentEventStream) Err() error { + if v := r.errVal.Load(); v != nil { + return v.(error) + } + + return nil +} + +func (r *readSelectObjectContentEventStream) Events() <-chan SelectObjectContentEventStreamEvent { + return r.stream +} + +func (r *readSelectObjectContentEventStream) readEventStream() { + defer close(r.stream) + + for { + event, err := r.eventReader.ReadEvent() + if err != nil { + if err == io.EOF { + return + } + select { + case <-r.done: + // If closed already ignore the error + return + default: + } + r.errVal.Store(err) + return + } + + select { + case r.stream <- event.(SelectObjectContentEventStreamEvent): + case <-r.done: + return + } + } +} + +func (r *readSelectObjectContentEventStream) unmarshalerForEventType( + eventType string, +) (eventstreamapi.Unmarshaler, error) { + switch eventType { + case "Cont": + return &ContinuationEvent{}, nil + + case "End": + return &EndEvent{}, nil + + case "Progress": + return &ProgressEvent{}, nil + + case "Records": + return &RecordsEvent{}, nil + + case "Stats": + return &StatsEvent{}, nil + default: + return nil, fmt.Errorf( + "unknown event type name, %s, for SelectObjectContentEventStream", eventType) + } +} + +// Request to filter the contents of an Amazon S3 object based on a simple Structured +// Query Language (SQL) statement. In the request, along with the SQL expression, +// you must also specify a data serialization format (JSON or CSV) of the object. +// Amazon S3 uses this to parse object data into records, and returns only records +// that match the specified SQL expression. You must also specify the data serialization +// format for the response. For more information, go to S3Select API Documentation +// (https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html) +type SelectObjectContentInput struct { + _ struct{} `locationName:"SelectObjectContentRequest" type:"structure" xmlURI:"http://s3.amazonaws.com/doc/2006-03-01/"` + + // The S3 Bucket. + // + // Bucket is a required field + Bucket *string `location:"uri" locationName:"Bucket" type:"string" required:"true"` + + // The expression that is used to query the object. + // + // Expression is a required field + Expression *string `type:"string" required:"true"` + + // The type of the provided expression (e.g., SQL). + // + // ExpressionType is a required field + ExpressionType *string `type:"string" required:"true" enum:"ExpressionType"` + + // Describes the format of the data in the object that is being queried. + // + // InputSerialization is a required field + InputSerialization *InputSerialization `type:"structure" required:"true"` + + // The Object Key. + // + // Key is a required field + Key *string `location:"uri" locationName:"Key" min:"1" type:"string" required:"true"` + + // Describes the format of the data that you want Amazon S3 to return in response. + // + // OutputSerialization is a required field + OutputSerialization *OutputSerialization `type:"structure" required:"true"` + + // Specifies if periodic request progress information should be enabled. + RequestProgress *RequestProgress `type:"structure"` + + // The SSE Algorithm used to encrypt the object. For more information, go to + // Server-Side Encryption (Using Customer-Provided Encryption Keys (https://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html) + SSECustomerAlgorithm *string `location:"header" locationName:"x-amz-server-side-encryption-customer-algorithm" type:"string"` + + // The SSE Customer Key. For more information, go to Server-Side Encryption + // (Using Customer-Provided Encryption Keys (https://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html) + SSECustomerKey *string `location:"header" locationName:"x-amz-server-side-encryption-customer-key" type:"string"` + + // The SSE Customer Key MD5. For more information, go to Server-Side Encryption + // (Using Customer-Provided Encryption Keys (https://docs.aws.amazon.com/AmazonS3/latest/dev/ServerSideEncryptionCustomerKeys.html) + SSECustomerKeyMD5 *string `location:"header" locationName:"x-amz-server-side-encryption-customer-key-MD5" type:"string"` +} + +// String returns the string representation +func (s SelectObjectContentInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SelectObjectContentInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *SelectObjectContentInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "SelectObjectContentInput"} + if s.Bucket == nil { + invalidParams.Add(request.NewErrParamRequired("Bucket")) + } + if s.Expression == nil { + invalidParams.Add(request.NewErrParamRequired("Expression")) + } + if s.ExpressionType == nil { + invalidParams.Add(request.NewErrParamRequired("ExpressionType")) + } + if s.InputSerialization == nil { + invalidParams.Add(request.NewErrParamRequired("InputSerialization")) + } + if s.Key == nil { + invalidParams.Add(request.NewErrParamRequired("Key")) + } + if s.Key != nil && len(*s.Key) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Key", 1)) + } + if s.OutputSerialization == nil { + invalidParams.Add(request.NewErrParamRequired("OutputSerialization")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetBucket sets the Bucket field's value. +func (s *SelectObjectContentInput) SetBucket(v string) *SelectObjectContentInput { + s.Bucket = &v + return s +} + +func (s *SelectObjectContentInput) getBucket() (v string) { + if s.Bucket == nil { + return v + } + return *s.Bucket +} + +// SetExpression sets the Expression field's value. +func (s *SelectObjectContentInput) SetExpression(v string) *SelectObjectContentInput { + s.Expression = &v + return s +} + +// SetExpressionType sets the ExpressionType field's value. +func (s *SelectObjectContentInput) SetExpressionType(v string) *SelectObjectContentInput { + s.ExpressionType = &v + return s +} + +// SetInputSerialization sets the InputSerialization field's value. +func (s *SelectObjectContentInput) SetInputSerialization(v *InputSerialization) *SelectObjectContentInput { + s.InputSerialization = v + return s +} + +// SetKey sets the Key field's value. +func (s *SelectObjectContentInput) SetKey(v string) *SelectObjectContentInput { + s.Key = &v + return s +} + +// SetOutputSerialization sets the OutputSerialization field's value. +func (s *SelectObjectContentInput) SetOutputSerialization(v *OutputSerialization) *SelectObjectContentInput { + s.OutputSerialization = v + return s +} + +// SetRequestProgress sets the RequestProgress field's value. +func (s *SelectObjectContentInput) SetRequestProgress(v *RequestProgress) *SelectObjectContentInput { + s.RequestProgress = v + return s +} + +// SetSSECustomerAlgorithm sets the SSECustomerAlgorithm field's value. +func (s *SelectObjectContentInput) SetSSECustomerAlgorithm(v string) *SelectObjectContentInput { + s.SSECustomerAlgorithm = &v + return s +} + +// SetSSECustomerKey sets the SSECustomerKey field's value. +func (s *SelectObjectContentInput) SetSSECustomerKey(v string) *SelectObjectContentInput { + s.SSECustomerKey = &v + return s +} + +func (s *SelectObjectContentInput) getSSECustomerKey() (v string) { + if s.SSECustomerKey == nil { + return v + } + return *s.SSECustomerKey +} + +// SetSSECustomerKeyMD5 sets the SSECustomerKeyMD5 field's value. +func (s *SelectObjectContentInput) SetSSECustomerKeyMD5(v string) *SelectObjectContentInput { + s.SSECustomerKeyMD5 = &v + return s +} + +type SelectObjectContentOutput struct { + _ struct{} `type:"structure" payload:"Payload"` + + // Use EventStream to use the API's stream. + EventStream *SelectObjectContentEventStream `type:"structure"` +} + +// String returns the string representation +func (s SelectObjectContentOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SelectObjectContentOutput) GoString() string { + return s.String() +} + +// SetEventStream sets the EventStream field's value. +func (s *SelectObjectContentOutput) SetEventStream(v *SelectObjectContentEventStream) *SelectObjectContentOutput { + s.EventStream = v + return s +} + +func (s *SelectObjectContentOutput) runEventStreamLoop(r *request.Request) { + if r.Error != nil { + return + } + reader := newReadSelectObjectContentEventStream( + r.HTTPResponse.Body, + r.Handlers.UnmarshalStream, + r.Config.Logger, + r.Config.LogLevel.Value(), + ) + go reader.readEventStream() + + eventStream := &SelectObjectContentEventStream{ + StreamCloser: r.HTTPResponse.Body, + Reader: reader, + } + s.EventStream = eventStream +} + // Describes the parameters for Select job types. type SelectParameters struct { _ struct{} `type:"structure"` @@ -19696,6 +20422,87 @@ func (s *SseKmsEncryptedObjects) SetStatus(v string) *SseKmsEncryptedObjects { return s } +type Stats struct { + _ struct{} `type:"structure"` + + // Total number of uncompressed object bytes processed. + BytesProcessed *int64 `type:"long"` + + // Total number of bytes of records payload data returned. + BytesReturned *int64 `type:"long"` + + // Total number of object bytes scanned. + BytesScanned *int64 `type:"long"` +} + +// String returns the string representation +func (s Stats) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Stats) GoString() string { + return s.String() +} + +// SetBytesProcessed sets the BytesProcessed field's value. +func (s *Stats) SetBytesProcessed(v int64) *Stats { + s.BytesProcessed = &v + return s +} + +// SetBytesReturned sets the BytesReturned field's value. +func (s *Stats) SetBytesReturned(v int64) *Stats { + s.BytesReturned = &v + return s +} + +// SetBytesScanned sets the BytesScanned field's value. +func (s *Stats) SetBytesScanned(v int64) *Stats { + s.BytesScanned = &v + return s +} + +type StatsEvent struct { + _ struct{} `type:"structure" payload:"Details"` + + // The Stats event details. + Details *Stats `locationName:"Details" type:"structure"` +} + +// String returns the string representation +func (s StatsEvent) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StatsEvent) GoString() string { + return s.String() +} + +// SetDetails sets the Details field's value. +func (s *StatsEvent) SetDetails(v *Stats) *StatsEvent { + s.Details = v + return s +} + +// The StatsEvent is and event in the SelectObjectContentEventStream group of events. +func (s *StatsEvent) eventSelectObjectContentEventStream() {} + +// UnmarshalEvent unmarshals the EventStream Message into the StatsEvent value. +// This method is only used internally within the SDK's EventStream handling. +func (s *StatsEvent) UnmarshalEvent( + payloadUnmarshaler protocol.PayloadUnmarshaler, + msg eventstream.Message, +) error { + if err := payloadUnmarshaler.UnmarshalPayload( + bytes.NewReader(msg.Payload), s, + ); err != nil { + return fmt.Errorf("failed to unmarshal payload, %v", err) + } + return nil +} + type StorageClassAnalysis struct { _ struct{} `type:"structure"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/s3/eventstream_test.go b/vendor/github.com/aws/aws-sdk-go/service/s3/eventstream_test.go new file mode 100644 index 000000000..35da7181b --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/s3/eventstream_test.go @@ -0,0 +1,297 @@ +package s3 + +import ( + "bytes" + "encoding/csv" + "fmt" + "io" + "io/ioutil" + "net/http" + "os" + "reflect" + "testing" + + "github.com/aws/aws-sdk-go/aws" + "github.com/aws/aws-sdk-go/aws/corehandlers" + "github.com/aws/aws-sdk-go/aws/request" + "github.com/aws/aws-sdk-go/aws/session" + "github.com/aws/aws-sdk-go/awstesting/unit" + "github.com/aws/aws-sdk-go/private/protocol" + "github.com/aws/aws-sdk-go/private/protocol/eventstream" + "github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamapi" + "github.com/aws/aws-sdk-go/private/protocol/eventstream/eventstreamtest" + "github.com/aws/aws-sdk-go/private/protocol/restxml" +) + +func mockSelectObjectContentEventStreamsOutputEvents() ([]SelectObjectContentEventStreamEvent, []eventstream.Message) { + + expectEvents := []SelectObjectContentEventStreamEvent{ + &ContinuationEvent{}, + &EndEvent{}, + &ProgressEvent{ + Details: &Progress{ + BytesProcessed: aws.Int64(123), + BytesReturned: aws.Int64(456), + BytesScanned: aws.Int64(789), + }, + }, + &RecordsEvent{ + Payload: []byte("abc123"), + }, + &StatsEvent{ + Details: &Stats{ + BytesProcessed: aws.Int64(123), + BytesReturned: aws.Int64(456), + BytesScanned: aws.Int64(789), + }, + }, + } + + var marshalers request.HandlerList + marshalers.PushBackNamed(restxml.BuildHandler) + payloadMarshaler := protocol.HandlerPayloadMarshal{ + Marshalers: marshalers, + } + + eventMsgs := []eventstream.Message{ + { + Headers: eventstream.Headers{ + eventstreamtest.EventMessageTypeHeader, + { + Name: eventstreamapi.EventTypeHeader, + Value: eventstream.StringValue("Cont"), + }, + }, + }, + { + Headers: eventstream.Headers{ + eventstreamtest.EventMessageTypeHeader, + { + Name: eventstreamapi.EventTypeHeader, + Value: eventstream.StringValue("End"), + }, + }, + }, + { + Headers: eventstream.Headers{ + eventstreamtest.EventMessageTypeHeader, + { + Name: eventstreamapi.EventTypeHeader, + Value: eventstream.StringValue("Progress"), + }, + }, + Payload: eventstreamtest.MarshalEventPayload(payloadMarshaler, expectEvents[2]), + }, + { + Headers: eventstream.Headers{ + eventstreamtest.EventMessageTypeHeader, + { + Name: eventstreamapi.EventTypeHeader, + Value: eventstream.StringValue("Records"), + }, + }, + Payload: eventstreamtest.MarshalEventPayload(payloadMarshaler, expectEvents[3]), + }, + { + Headers: eventstream.Headers{ + eventstreamtest.EventMessageTypeHeader, + { + Name: eventstreamapi.EventTypeHeader, + Value: eventstream.StringValue("Stats"), + }, + }, + Payload: eventstreamtest.MarshalEventPayload(payloadMarshaler, expectEvents[4]), + }, + } + + return expectEvents, eventMsgs +} + +func TestSelectObjectContentEventStream(t *testing.T) { + expectEvents, eventMsgs := mockSelectObjectContentEventStreamsOutputEvents() + sess, cleanupFn, err := eventstreamtest.SetupEventStreamSession(t, + eventstreamtest.ServeEventStream{ + T: t, + Events: eventMsgs, + }, + false, + ) + if err != nil { + t.Fatalf("expect no error, %v", err) + } + defer cleanupFn() + + svc := New(sess) + + params := &SelectObjectContentInput{} + resp, err := svc.SelectObjectContent(params) + if err != nil { + t.Fatalf("expect no error got, %v", err) + } + defer resp.EventStream.Close() + + var i int + for event := range resp.EventStream.Events() { + if event == nil { + t.Errorf("%d, expect event, got nil", i) + } + if e, a := expectEvents[i], event; !reflect.DeepEqual(e, a) { + t.Errorf("%d, expect %T %v, got %T %v", i, e, e, a, a) + } + i++ + } + + if err := resp.EventStream.Err(); err != nil { + t.Errorf("expect no error, %v", err) + } +} + +func TestSelectObjectContentEventStream_Close(t *testing.T) { + _, eventMsgs := mockSelectObjectContentEventStreamsOutputEvents() + sess, cleanupFn, err := eventstreamtest.SetupEventStreamSession(t, + eventstreamtest.ServeEventStream{ + T: t, + Events: eventMsgs, + }, + false, + ) + if err != nil { + t.Fatalf("expect no error, %v", err) + } + defer cleanupFn() + + svc := New(sess) + + params := &SelectObjectContentInput{} + resp, err := svc.SelectObjectContent(params) + if err != nil { + t.Fatalf("expect no error got, %v", err) + } + + resp.EventStream.Close() + + if err := resp.EventStream.Err(); err != nil { + t.Errorf("expect no error, %v", err) + } +} + +func ExampleS3_SelectObjectContent() { + sess := session.Must(session.NewSession()) + svc := New(sess) + + /* + Example myObjectKey CSV content: + + name,number + gopher,0 + ᵷodɥǝɹ,1 + */ + + // Make the Select Object Content API request using the object uploaded. + resp, err := svc.SelectObjectContent(&SelectObjectContentInput{ + Bucket: aws.String("myBucket"), + Key: aws.String("myObjectKey"), + Expression: aws.String("SELECT name FROM S3Object WHERE cast(number as int) < 1"), + ExpressionType: aws.String(ExpressionTypeSql), + InputSerialization: &InputSerialization{ + CSV: &CSVInput{ + FileHeaderInfo: aws.String(FileHeaderInfoUse), + }, + }, + OutputSerialization: &OutputSerialization{ + CSV: &CSVOutput{}, + }, + }) + if err != nil { + fmt.Fprintf(os.Stderr, "failed making API request, %v\n", err) + return + } + defer resp.EventStream.Close() + + results, resultWriter := io.Pipe() + go func() { + defer resultWriter.Close() + for event := range resp.EventStream.Events() { + switch e := event.(type) { + case *RecordsEvent: + resultWriter.Write(e.Payload) + case *StatsEvent: + fmt.Printf("Processed %d bytes\n", e.Details.BytesProcessed) + } + } + }() + + // Printout the results + resReader := csv.NewReader(results) + for { + record, err := resReader.Read() + if err == io.EOF { + break + } + fmt.Println(record) + } + + if err := resp.EventStream.Err(); err != nil { + fmt.Fprintf(os.Stderr, "reading from event stream failed, %v\n", err) + } +} + +func BenchmarkSelectObjectContentEventStream(b *testing.B) { + _, eventMsgs := mockSelectObjectContentEventStreamsOutputEvents() + var buf bytes.Buffer + encoder := eventstream.NewEncoder(&buf) + for _, msg := range eventMsgs { + if err := encoder.Encode(msg); err != nil { + b.Fatalf("failed to encode message, %v", err) + } + } + stream := &loopReader{source: bytes.NewReader(buf.Bytes())} + + sess := unit.Session + svc := New(sess, &aws.Config{ + Endpoint: aws.String("https://example.com"), + DisableParamValidation: aws.Bool(true), + }) + svc.Handlers.Send.Swap(corehandlers.SendHandler.Name, + request.NamedHandler{Name: "mockSend", + Fn: func(r *request.Request) { + r.HTTPResponse = &http.Response{ + Status: "200 OK", + StatusCode: 200, + Header: http.Header{}, + Body: ioutil.NopCloser(stream), + } + }, + }, + ) + + params := &SelectObjectContentInput{} + resp, err := svc.SelectObjectContent(params) + if err != nil { + b.Fatalf("failed to create request, %v", err) + } + defer resp.EventStream.Close() + b.ResetTimer() + + for i := 0; i < b.N; i++ { + if err = resp.EventStream.Err(); err != nil { + b.Fatalf("expect no error, got %v", err) + } + event := <-resp.EventStream.Events() + if event == nil { + b.Fatalf("expect event, got nil, %v, %d", resp.EventStream.Err(), i) + } + } +} + +type loopReader struct { + source *bytes.Reader +} + +func (c *loopReader) Read(p []byte) (int, error) { + if c.source.Len() == 0 { + c.source.Seek(0, 0) + } + + return c.source.Read(p) +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/s3/host_style_bucket_test.go b/vendor/github.com/aws/aws-sdk-go/service/s3/host_style_bucket_test.go index 094fa6527..e7e459855 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/s3/host_style_bucket_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/s3/host_style_bucket_test.go @@ -1,7 +1,10 @@ package s3_test import ( + "encoding/json" "net/url" + "os" + "path/filepath" "strings" "testing" @@ -125,3 +128,52 @@ func TestHostStyleBucketGetBucketLocation(t *testing.T) { t.Errorf("expect %s to be in %s", e, a) } } + +func TestVirtualHostStyleSuite(t *testing.T) { + f, err := os.Open(filepath.Join("testdata", "virtual_host.json")) + if err != nil { + t.Fatalf("expect no error, %v", err) + } + + cases := []struct { + Bucket string + Region string + UseDualStack bool + UseS3Accelerate bool + ConfiguredAddressingStyle string + + ExpectedURI string + }{} + + decoder := json.NewDecoder(f) + if err := decoder.Decode(&cases); err != nil { + t.Fatalf("expect no error, %v", err) + } + + const testPathStyle = "path" + for i, c := range cases { + svc := s3.New(unit.Session, &aws.Config{ + Region: &c.Region, + UseDualStack: &c.UseDualStack, + S3UseAccelerate: &c.UseS3Accelerate, + S3ForcePathStyle: aws.Bool(c.ConfiguredAddressingStyle == testPathStyle), + }) + + req, _ := svc.HeadBucketRequest(&s3.HeadBucketInput{ + Bucket: &c.Bucket, + }) + req.Build() + if req.Error != nil { + t.Fatalf("expect no error, got %v", req.Error) + } + + // Trim trailing '/' that are added by the SDK but not in the tests. + actualURI := strings.TrimRightFunc( + req.HTTPRequest.URL.String(), + func(r rune) bool { return r == '/' }, + ) + if e, a := c.ExpectedURI, actualURI; e != a { + t.Errorf("%d, expect\n%s\nurl to be\n%s", i, e, a) + } + } +} diff --git a/vendor/github.com/aws/aws-sdk-go/service/s3/s3iface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/s3/s3iface/interface.go index 28c30d976..cc427882b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/s3/s3iface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/s3/s3iface/interface.go @@ -375,6 +375,10 @@ type S3API interface { RestoreObjectWithContext(aws.Context, *s3.RestoreObjectInput, ...request.Option) (*s3.RestoreObjectOutput, error) RestoreObjectRequest(*s3.RestoreObjectInput) (*request.Request, *s3.RestoreObjectOutput) + SelectObjectContent(*s3.SelectObjectContentInput) (*s3.SelectObjectContentOutput, error) + SelectObjectContentWithContext(aws.Context, *s3.SelectObjectContentInput, ...request.Option) (*s3.SelectObjectContentOutput, error) + SelectObjectContentRequest(*s3.SelectObjectContentInput) (*request.Request, *s3.SelectObjectContentOutput) + UploadPart(*s3.UploadPartInput) (*s3.UploadPartOutput, error) UploadPartWithContext(aws.Context, *s3.UploadPartInput, ...request.Option) (*s3.UploadPartOutput, error) UploadPartRequest(*s3.UploadPartInput) (*request.Request, *s3.UploadPartOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/s3/service.go b/vendor/github.com/aws/aws-sdk-go/service/s3/service.go index 614e477d3..20de53f29 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/s3/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/s3/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "s3" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "s3" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "S3" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the S3 client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, @@ -71,6 +73,8 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio svc.Handlers.UnmarshalMeta.PushBackNamed(restxml.UnmarshalMetaHandler) svc.Handlers.UnmarshalError.PushBackNamed(restxml.UnmarshalErrorHandler) + svc.Handlers.UnmarshalStream.PushBackNamed(restxml.UnmarshalHandler) + // Run custom client initialization if present if initClient != nil { initClient(svc.Client) diff --git a/vendor/github.com/aws/aws-sdk-go/service/s3/testdata/virtual_host.json b/vendor/github.com/aws/aws-sdk-go/service/s3/testdata/virtual_host.json new file mode 100644 index 000000000..7399a26c5 --- /dev/null +++ b/vendor/github.com/aws/aws-sdk-go/service/s3/testdata/virtual_host.json @@ -0,0 +1,178 @@ +[ + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3.amazonaws.com", + "Region": "us-east-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3.us-west-1.amazonaws.com", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-with-number-1", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-with-number-1.s3.us-west-1.amazonaws.com", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3.cn-north-1.amazonaws.com.cn", + "Region": "cn-north-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "BucketName", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.amazonaws.com/BucketName", + "Region": "us-east-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "BucketName", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.amazonaws.com/BucketName", + "Region": "us-east-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket_name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.us-west-1.amazonaws.com/bucket_name", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket.name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.us-west-1.amazonaws.com/bucket.name", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "-bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.us-west-1.amazonaws.com/-bucket-name", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name-", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.us-west-1.amazonaws.com/bucket-name-", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "aa", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.us-west-1.amazonaws.com/aa", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.us-west-1.amazonaws.com/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3-accelerate.amazonaws.com", + "Region": "us-east-1", + "UseDualstack": false, + "UseS3Accelerate": true + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3-accelerate.amazonaws.com", + "Region": "us-west-1", + "UseDualstack": false, + "UseS3Accelerate": true + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3.dualstack.us-east-1.amazonaws.com", + "Region": "us-east-1", + "UseDualstack": true, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3.dualstack.us-west-2.amazonaws.com", + "Region": "us-west-2", + "UseDualstack": true, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket.name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://s3.dualstack.us-west-2.amazonaws.com/bucket.name", + "Region": "us-west-2", + "UseDualstack": true, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "default", + "ExpectedUri": "https://bucket-name.s3-accelerate.dualstack.amazonaws.com", + "Region": "us-east-1", + "UseDualstack": true, + "UseS3Accelerate": true + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "path", + "ExpectedUri": "https://s3.amazonaws.com/bucket-name", + "Region": "us-east-1", + "UseDualstack": false, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "path", + "ExpectedUri": "https://bucket-name.s3-accelerate.amazonaws.com", + "Region": "us-east-1", + "UseDualstack": false, + "UseS3Accelerate": true + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "path", + "ExpectedUri": "https://s3.dualstack.us-east-1.amazonaws.com/bucket-name", + "Region": "us-east-1", + "UseDualstack": true, + "UseS3Accelerate": false + }, + { + "Bucket": "bucket-name", + "ConfiguredAddressingStyle": "path", + "ExpectedUri": "https://bucket-name.s3-accelerate.dualstack.amazonaws.com", + "Region": "us-east-1", + "UseDualstack": true, + "UseS3Accelerate": true + } +] diff --git a/vendor/github.com/aws/aws-sdk-go/service/sagemaker/api.go b/vendor/github.com/aws/aws-sdk-go/service/sagemaker/api.go index a2686f952..f9fcf9314 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sagemaker/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sagemaker/api.go @@ -291,6 +291,89 @@ func (c *SageMaker) CreateEndpointConfigWithContext(ctx aws.Context, input *Crea return out, req.Send() } +const opCreateHyperParameterTuningJob = "CreateHyperParameterTuningJob" + +// CreateHyperParameterTuningJobRequest generates a "aws/request.Request" representing the +// client's request for the CreateHyperParameterTuningJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateHyperParameterTuningJob for more information on using the CreateHyperParameterTuningJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateHyperParameterTuningJobRequest method. +// req, resp := client.CreateHyperParameterTuningJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateHyperParameterTuningJob +func (c *SageMaker) CreateHyperParameterTuningJobRequest(input *CreateHyperParameterTuningJobInput) (req *request.Request, output *CreateHyperParameterTuningJobOutput) { + op := &request.Operation{ + Name: opCreateHyperParameterTuningJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateHyperParameterTuningJobInput{} + } + + output = &CreateHyperParameterTuningJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateHyperParameterTuningJob API operation for Amazon SageMaker Service. +// +// Starts a hyperparameter tuning job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon SageMaker Service's +// API operation CreateHyperParameterTuningJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceInUse "ResourceInUse" +// Resource being accessed is in use. +// +// * ErrCodeResourceLimitExceeded "ResourceLimitExceeded" +// You have exceeded an Amazon SageMaker resource limit. For example, you might +// have too many training jobs created. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/CreateHyperParameterTuningJob +func (c *SageMaker) CreateHyperParameterTuningJob(input *CreateHyperParameterTuningJobInput) (*CreateHyperParameterTuningJobOutput, error) { + req, out := c.CreateHyperParameterTuningJobRequest(input) + return out, req.Send() +} + +// CreateHyperParameterTuningJobWithContext is the same as CreateHyperParameterTuningJob with the addition of +// the ability to pass a context and additional request options. +// +// See CreateHyperParameterTuningJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) CreateHyperParameterTuningJobWithContext(ctx aws.Context, input *CreateHyperParameterTuningJobInput, opts ...request.Option) (*CreateHyperParameterTuningJobOutput, error) { + req, out := c.CreateHyperParameterTuningJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opCreateModel = "CreateModel" // CreateModelRequest generates a "aws/request.Request" representing the @@ -648,7 +731,7 @@ func (c *SageMaker) CreatePresignedNotebookInstanceUrlRequest(input *CreatePresi // CreatePresignedNotebookInstanceUrl API operation for Amazon SageMaker Service. // -// Returns a URL that you can use to connect to the Juypter server from a notebook +// Returns a URL that you can use to connect to the Jupyter server from a notebook // instance. In the Amazon SageMaker console, when you choose Open next to a // notebook instance, Amazon SageMaker opens a new tab showing the Jupyter server // home page from the notebook instance. The console uses this API to get the @@ -923,7 +1006,7 @@ func (c *SageMaker) DeleteEndpointConfigRequest(input *DeleteEndpointConfigInput // DeleteEndpointConfig API operation for Amazon SageMaker Service. // -// Deletes an endpoint configuration. The DeleteEndpoingConfig API deletes only +// Deletes an endpoint configuration. The DeleteEndpointConfig API deletes only // the specified configuration. It does not delete endpoints created using the // configuration. // @@ -1416,6 +1499,85 @@ func (c *SageMaker) DescribeEndpointConfigWithContext(ctx aws.Context, input *De return out, req.Send() } +const opDescribeHyperParameterTuningJob = "DescribeHyperParameterTuningJob" + +// DescribeHyperParameterTuningJobRequest generates a "aws/request.Request" representing the +// client's request for the DescribeHyperParameterTuningJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeHyperParameterTuningJob for more information on using the DescribeHyperParameterTuningJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeHyperParameterTuningJobRequest method. +// req, resp := client.DescribeHyperParameterTuningJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeHyperParameterTuningJob +func (c *SageMaker) DescribeHyperParameterTuningJobRequest(input *DescribeHyperParameterTuningJobInput) (req *request.Request, output *DescribeHyperParameterTuningJobOutput) { + op := &request.Operation{ + Name: opDescribeHyperParameterTuningJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeHyperParameterTuningJobInput{} + } + + output = &DescribeHyperParameterTuningJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeHyperParameterTuningJob API operation for Amazon SageMaker Service. +// +// Gets a description of a hyperparameter tuning job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon SageMaker Service's +// API operation DescribeHyperParameterTuningJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFound "ResourceNotFound" +// Resource being access is not found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/DescribeHyperParameterTuningJob +func (c *SageMaker) DescribeHyperParameterTuningJob(input *DescribeHyperParameterTuningJobInput) (*DescribeHyperParameterTuningJobOutput, error) { + req, out := c.DescribeHyperParameterTuningJobRequest(input) + return out, req.Send() +} + +// DescribeHyperParameterTuningJobWithContext is the same as DescribeHyperParameterTuningJob with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeHyperParameterTuningJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) DescribeHyperParameterTuningJobWithContext(ctx aws.Context, input *DescribeHyperParameterTuningJobInput, opts ...request.Option) (*DescribeHyperParameterTuningJobOutput, error) { + req, out := c.DescribeHyperParameterTuningJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDescribeModel = "DescribeModel" // DescribeModelRequest generates a "aws/request.Request" representing the @@ -1979,6 +2141,137 @@ func (c *SageMaker) ListEndpointsPagesWithContext(ctx aws.Context, input *ListEn return p.Err() } +const opListHyperParameterTuningJobs = "ListHyperParameterTuningJobs" + +// ListHyperParameterTuningJobsRequest generates a "aws/request.Request" representing the +// client's request for the ListHyperParameterTuningJobs operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListHyperParameterTuningJobs for more information on using the ListHyperParameterTuningJobs +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListHyperParameterTuningJobsRequest method. +// req, resp := client.ListHyperParameterTuningJobsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListHyperParameterTuningJobs +func (c *SageMaker) ListHyperParameterTuningJobsRequest(input *ListHyperParameterTuningJobsInput) (req *request.Request, output *ListHyperParameterTuningJobsOutput) { + op := &request.Operation{ + Name: opListHyperParameterTuningJobs, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListHyperParameterTuningJobsInput{} + } + + output = &ListHyperParameterTuningJobsOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListHyperParameterTuningJobs API operation for Amazon SageMaker Service. +// +// Gets a list of objects that describe the hyperparameter tuning jobs launched +// in your account. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon SageMaker Service's +// API operation ListHyperParameterTuningJobs for usage and error information. +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListHyperParameterTuningJobs +func (c *SageMaker) ListHyperParameterTuningJobs(input *ListHyperParameterTuningJobsInput) (*ListHyperParameterTuningJobsOutput, error) { + req, out := c.ListHyperParameterTuningJobsRequest(input) + return out, req.Send() +} + +// ListHyperParameterTuningJobsWithContext is the same as ListHyperParameterTuningJobs with the addition of +// the ability to pass a context and additional request options. +// +// See ListHyperParameterTuningJobs for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) ListHyperParameterTuningJobsWithContext(ctx aws.Context, input *ListHyperParameterTuningJobsInput, opts ...request.Option) (*ListHyperParameterTuningJobsOutput, error) { + req, out := c.ListHyperParameterTuningJobsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListHyperParameterTuningJobsPages iterates over the pages of a ListHyperParameterTuningJobs operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListHyperParameterTuningJobs method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListHyperParameterTuningJobs operation. +// pageNum := 0 +// err := client.ListHyperParameterTuningJobsPages(params, +// func(page *ListHyperParameterTuningJobsOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *SageMaker) ListHyperParameterTuningJobsPages(input *ListHyperParameterTuningJobsInput, fn func(*ListHyperParameterTuningJobsOutput, bool) bool) error { + return c.ListHyperParameterTuningJobsPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListHyperParameterTuningJobsPagesWithContext same as ListHyperParameterTuningJobsPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) ListHyperParameterTuningJobsPagesWithContext(ctx aws.Context, input *ListHyperParameterTuningJobsInput, fn func(*ListHyperParameterTuningJobsOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListHyperParameterTuningJobsInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListHyperParameterTuningJobsRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListHyperParameterTuningJobsOutput), !p.HasNextPage()) + } + return p.Err() +} + const opListModels = "ListModels" // ListModelsRequest generates a "aws/request.Request" representing the @@ -2631,6 +2924,142 @@ func (c *SageMaker) ListTrainingJobsPagesWithContext(ctx aws.Context, input *Lis return p.Err() } +const opListTrainingJobsForHyperParameterTuningJob = "ListTrainingJobsForHyperParameterTuningJob" + +// ListTrainingJobsForHyperParameterTuningJobRequest generates a "aws/request.Request" representing the +// client's request for the ListTrainingJobsForHyperParameterTuningJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See ListTrainingJobsForHyperParameterTuningJob for more information on using the ListTrainingJobsForHyperParameterTuningJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the ListTrainingJobsForHyperParameterTuningJobRequest method. +// req, resp := client.ListTrainingJobsForHyperParameterTuningJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListTrainingJobsForHyperParameterTuningJob +func (c *SageMaker) ListTrainingJobsForHyperParameterTuningJobRequest(input *ListTrainingJobsForHyperParameterTuningJobInput) (req *request.Request, output *ListTrainingJobsForHyperParameterTuningJobOutput) { + op := &request.Operation{ + Name: opListTrainingJobsForHyperParameterTuningJob, + HTTPMethod: "POST", + HTTPPath: "/", + Paginator: &request.Paginator{ + InputTokens: []string{"NextToken"}, + OutputTokens: []string{"NextToken"}, + LimitToken: "MaxResults", + TruncationToken: "", + }, + } + + if input == nil { + input = &ListTrainingJobsForHyperParameterTuningJobInput{} + } + + output = &ListTrainingJobsForHyperParameterTuningJobOutput{} + req = c.newRequest(op, input, output) + return +} + +// ListTrainingJobsForHyperParameterTuningJob API operation for Amazon SageMaker Service. +// +// Gets a list of objects that describe the training jobs that a hyperparameter +// tuning job launched. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon SageMaker Service's +// API operation ListTrainingJobsForHyperParameterTuningJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFound "ResourceNotFound" +// Resource being access is not found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/ListTrainingJobsForHyperParameterTuningJob +func (c *SageMaker) ListTrainingJobsForHyperParameterTuningJob(input *ListTrainingJobsForHyperParameterTuningJobInput) (*ListTrainingJobsForHyperParameterTuningJobOutput, error) { + req, out := c.ListTrainingJobsForHyperParameterTuningJobRequest(input) + return out, req.Send() +} + +// ListTrainingJobsForHyperParameterTuningJobWithContext is the same as ListTrainingJobsForHyperParameterTuningJob with the addition of +// the ability to pass a context and additional request options. +// +// See ListTrainingJobsForHyperParameterTuningJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) ListTrainingJobsForHyperParameterTuningJobWithContext(ctx aws.Context, input *ListTrainingJobsForHyperParameterTuningJobInput, opts ...request.Option) (*ListTrainingJobsForHyperParameterTuningJobOutput, error) { + req, out := c.ListTrainingJobsForHyperParameterTuningJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +// ListTrainingJobsForHyperParameterTuningJobPages iterates over the pages of a ListTrainingJobsForHyperParameterTuningJob operation, +// calling the "fn" function with the response data for each page. To stop +// iterating, return false from the fn function. +// +// See ListTrainingJobsForHyperParameterTuningJob method for more information on how to use this operation. +// +// Note: This operation can generate multiple requests to a service. +// +// // Example iterating over at most 3 pages of a ListTrainingJobsForHyperParameterTuningJob operation. +// pageNum := 0 +// err := client.ListTrainingJobsForHyperParameterTuningJobPages(params, +// func(page *ListTrainingJobsForHyperParameterTuningJobOutput, lastPage bool) bool { +// pageNum++ +// fmt.Println(page) +// return pageNum <= 3 +// }) +// +func (c *SageMaker) ListTrainingJobsForHyperParameterTuningJobPages(input *ListTrainingJobsForHyperParameterTuningJobInput, fn func(*ListTrainingJobsForHyperParameterTuningJobOutput, bool) bool) error { + return c.ListTrainingJobsForHyperParameterTuningJobPagesWithContext(aws.BackgroundContext(), input, fn) +} + +// ListTrainingJobsForHyperParameterTuningJobPagesWithContext same as ListTrainingJobsForHyperParameterTuningJobPages except +// it takes a Context and allows setting request options on the pages. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) ListTrainingJobsForHyperParameterTuningJobPagesWithContext(ctx aws.Context, input *ListTrainingJobsForHyperParameterTuningJobInput, fn func(*ListTrainingJobsForHyperParameterTuningJobOutput, bool) bool, opts ...request.Option) error { + p := request.Pagination{ + NewRequest: func() (*request.Request, error) { + var inCpy *ListTrainingJobsForHyperParameterTuningJobInput + if input != nil { + tmp := *input + inCpy = &tmp + } + req, _ := c.ListTrainingJobsForHyperParameterTuningJobRequest(inCpy) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return req, nil + }, + } + + cont := true + for p.Next() && cont { + cont = fn(p.Page().(*ListTrainingJobsForHyperParameterTuningJobOutput), !p.HasNextPage()) + } + return p.Err() +} + const opStartNotebookInstance = "StartNotebookInstance" // StartNotebookInstanceRequest generates a "aws/request.Request" representing the @@ -2717,6 +3146,93 @@ func (c *SageMaker) StartNotebookInstanceWithContext(ctx aws.Context, input *Sta return out, req.Send() } +const opStopHyperParameterTuningJob = "StopHyperParameterTuningJob" + +// StopHyperParameterTuningJobRequest generates a "aws/request.Request" representing the +// client's request for the StopHyperParameterTuningJob operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See StopHyperParameterTuningJob for more information on using the StopHyperParameterTuningJob +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the StopHyperParameterTuningJobRequest method. +// req, resp := client.StopHyperParameterTuningJobRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StopHyperParameterTuningJob +func (c *SageMaker) StopHyperParameterTuningJobRequest(input *StopHyperParameterTuningJobInput) (req *request.Request, output *StopHyperParameterTuningJobOutput) { + op := &request.Operation{ + Name: opStopHyperParameterTuningJob, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &StopHyperParameterTuningJobInput{} + } + + output = &StopHyperParameterTuningJobOutput{} + req = c.newRequest(op, input, output) + req.Handlers.Unmarshal.Remove(jsonrpc.UnmarshalHandler) + req.Handlers.Unmarshal.PushBackNamed(protocol.UnmarshalDiscardBodyHandler) + return +} + +// StopHyperParameterTuningJob API operation for Amazon SageMaker Service. +// +// Stops a running hyperparameter tuning job and all running training jobs that +// the tuning job launched. +// +// All model artifacts output from the training jobs are stored in Amazon Simple +// Storage Service (Amazon S3). All data that the training jobs write toAmazon +// CloudWatch Logs are still available in CloudWatch. After the tuning job moves +// to the Stopped state, it releases all reserved resources for the tuning job. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon SageMaker Service's +// API operation StopHyperParameterTuningJob for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFound "ResourceNotFound" +// Resource being access is not found. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/sagemaker-2017-07-24/StopHyperParameterTuningJob +func (c *SageMaker) StopHyperParameterTuningJob(input *StopHyperParameterTuningJobInput) (*StopHyperParameterTuningJobOutput, error) { + req, out := c.StopHyperParameterTuningJobRequest(input) + return out, req.Send() +} + +// StopHyperParameterTuningJobWithContext is the same as StopHyperParameterTuningJob with the addition of +// the ability to pass a context and additional request options. +// +// See StopHyperParameterTuningJob for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SageMaker) StopHyperParameterTuningJobWithContext(ctx aws.Context, input *StopHyperParameterTuningJobInput, opts ...request.Option) (*StopHyperParameterTuningJobOutput, error) { + req, out := c.StopHyperParameterTuningJobRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opStopNotebookInstance = "StopNotebookInstance" // StopNotebookInstanceRequest generates a "aws/request.Request" representing the @@ -3393,6 +3909,62 @@ func (s *AlgorithmSpecification) SetTrainingInputMode(v string) *AlgorithmSpecif return s } +// A list of categorical hyperparameters to tune. +type CategoricalParameterRange struct { + _ struct{} `type:"structure"` + + // The name of the categorical hyperparameter to tune. + // + // Name is a required field + Name *string `type:"string" required:"true"` + + // A list of the categories for the hyperparameter. + // + // Values is a required field + Values []*string `min:"1" type:"list" required:"true"` +} + +// String returns the string representation +func (s CategoricalParameterRange) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CategoricalParameterRange) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CategoricalParameterRange) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CategoricalParameterRange"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + if s.Values == nil { + invalidParams.Add(request.NewErrParamRequired("Values")) + } + if s.Values != nil && len(s.Values) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Values", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *CategoricalParameterRange) SetName(v string) *CategoricalParameterRange { + s.Name = &v + return s +} + +// SetValues sets the Values field's value. +func (s *CategoricalParameterRange) SetValues(v []*string) *CategoricalParameterRange { + s.Values = v + return s +} + // A channel is a named input source that training algorithms can consume. type Channel struct { _ struct{} `type:"structure"` @@ -3403,7 +3975,7 @@ type Channel struct { ChannelName *string `min:"1" type:"string" required:"true"` // If training data is compressed, the compression type. The default value is - // None. CompressionType is used only in PIPE input mode. In FILE mode, leave + // None. CompressionType is used only in Pipe input mode. In File mode, leave // this field unset or set it to None. CompressionType *string `type:"string" enum:"CompressionType"` @@ -3416,7 +3988,7 @@ type Channel struct { DataSource *DataSource `type:"structure" required:"true"` // Specify RecordIO as the value when input data is in raw format but the training - // algorithm requires the RecordIO format, in which caseAmazon SageMaker wraps + // algorithm requires the RecordIO format, in which case, Amazon SageMaker wraps // each individual S3 object in a RecordIO record. If the input data is already // in RecordIO format, you don't need to set this attribute. For more information, // see Create a Dataset Using RecordIO (https://mxnet.incubator.apache.org/how_to/recordio.html?highlight=im2rec) @@ -3561,6 +4133,75 @@ func (s *ContainerDefinition) SetModelDataUrl(v string) *ContainerDefinition { return s } +// A list of continuous hyperparameters to tune. +type ContinuousParameterRange struct { + _ struct{} `type:"structure"` + + // The maximum value for the hyperparameter. The tuning job uses floating-point + // values between MinValue value and this value for tuning. + // + // MaxValue is a required field + MaxValue *string `type:"string" required:"true"` + + // The minimum value for the hyperparameter. The tuning job uses floating-point + // values between this value and MaxValuefor tuning. + // + // MinValue is a required field + MinValue *string `type:"string" required:"true"` + + // The name of the continuous hyperparameter to tune. + // + // Name is a required field + Name *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s ContinuousParameterRange) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ContinuousParameterRange) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ContinuousParameterRange) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ContinuousParameterRange"} + if s.MaxValue == nil { + invalidParams.Add(request.NewErrParamRequired("MaxValue")) + } + if s.MinValue == nil { + invalidParams.Add(request.NewErrParamRequired("MinValue")) + } + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMaxValue sets the MaxValue field's value. +func (s *ContinuousParameterRange) SetMaxValue(v string) *ContinuousParameterRange { + s.MaxValue = &v + return s +} + +// SetMinValue sets the MinValue field's value. +func (s *ContinuousParameterRange) SetMinValue(v string) *ContinuousParameterRange { + s.MinValue = &v + return s +} + +// SetName sets the Name field's value. +func (s *ContinuousParameterRange) SetName(v string) *ContinuousParameterRange { + s.Name = &v + return s +} + type CreateEndpointConfigInput struct { _ struct{} `type:"structure"` @@ -3786,6 +4427,139 @@ func (s *CreateEndpointOutput) SetEndpointArn(v string) *CreateEndpointOutput { return s } +type CreateHyperParameterTuningJobInput struct { + _ struct{} `type:"structure"` + + // The object that describes the tuning job, including the search strategy, + // metric used to evaluate training jobs, ranges of parameters to search, and + // resource limits for the tuning job. + // + // HyperParameterTuningJobConfig is a required field + HyperParameterTuningJobConfig *HyperParameterTuningJobConfig `type:"structure" required:"true"` + + // The name of the tuning job. This name is the prefix for the names of all + // training jobs that this tuning job launches. The name must be unique within + // the same AWS account and AWS Region. Names are not case sensitive, and must + // be between 1-32 characters. + // + // HyperParameterTuningJobName is a required field + HyperParameterTuningJobName *string `min:"1" type:"string" required:"true"` + + // An array of key-value pairs. You can use tags to categorize your AWS resources + // in different ways, for example, by purpose, owner, or environment. For more + // information, see Using Cost Allocation Tags (http://docs.aws.amazon.com//awsaccountbilling/latest/aboutv2/cost-alloc-tags.html#allocation-what) + // in the AWS Billing and Cost Management User Guide. + Tags []*Tag `type:"list"` + + // The object that describes the training jobs that this tuning job launches, + // including static hyperparameters, input data configuration, output data configuration, + // resource configuration, and stopping condition. + // + // TrainingJobDefinition is a required field + TrainingJobDefinition *HyperParameterTrainingJobDefinition `type:"structure" required:"true"` +} + +// String returns the string representation +func (s CreateHyperParameterTuningJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateHyperParameterTuningJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateHyperParameterTuningJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateHyperParameterTuningJobInput"} + if s.HyperParameterTuningJobConfig == nil { + invalidParams.Add(request.NewErrParamRequired("HyperParameterTuningJobConfig")) + } + if s.HyperParameterTuningJobName == nil { + invalidParams.Add(request.NewErrParamRequired("HyperParameterTuningJobName")) + } + if s.HyperParameterTuningJobName != nil && len(*s.HyperParameterTuningJobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("HyperParameterTuningJobName", 1)) + } + if s.TrainingJobDefinition == nil { + invalidParams.Add(request.NewErrParamRequired("TrainingJobDefinition")) + } + if s.HyperParameterTuningJobConfig != nil { + if err := s.HyperParameterTuningJobConfig.Validate(); err != nil { + invalidParams.AddNested("HyperParameterTuningJobConfig", err.(request.ErrInvalidParams)) + } + } + if s.Tags != nil { + for i, v := range s.Tags { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "Tags", i), err.(request.ErrInvalidParams)) + } + } + } + if s.TrainingJobDefinition != nil { + if err := s.TrainingJobDefinition.Validate(); err != nil { + invalidParams.AddNested("TrainingJobDefinition", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetHyperParameterTuningJobConfig sets the HyperParameterTuningJobConfig field's value. +func (s *CreateHyperParameterTuningJobInput) SetHyperParameterTuningJobConfig(v *HyperParameterTuningJobConfig) *CreateHyperParameterTuningJobInput { + s.HyperParameterTuningJobConfig = v + return s +} + +// SetHyperParameterTuningJobName sets the HyperParameterTuningJobName field's value. +func (s *CreateHyperParameterTuningJobInput) SetHyperParameterTuningJobName(v string) *CreateHyperParameterTuningJobInput { + s.HyperParameterTuningJobName = &v + return s +} + +// SetTags sets the Tags field's value. +func (s *CreateHyperParameterTuningJobInput) SetTags(v []*Tag) *CreateHyperParameterTuningJobInput { + s.Tags = v + return s +} + +// SetTrainingJobDefinition sets the TrainingJobDefinition field's value. +func (s *CreateHyperParameterTuningJobInput) SetTrainingJobDefinition(v *HyperParameterTrainingJobDefinition) *CreateHyperParameterTuningJobInput { + s.TrainingJobDefinition = v + return s +} + +type CreateHyperParameterTuningJobOutput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the tuning job. + // + // HyperParameterTuningJobArn is a required field + HyperParameterTuningJobArn *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s CreateHyperParameterTuningJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateHyperParameterTuningJobOutput) GoString() string { + return s.String() +} + +// SetHyperParameterTuningJobArn sets the HyperParameterTuningJobArn field's value. +func (s *CreateHyperParameterTuningJobOutput) SetHyperParameterTuningJobArn(v string) *CreateHyperParameterTuningJobOutput { + s.HyperParameterTuningJobArn = &v + return s +} + type CreateModelInput struct { _ struct{} `type:"structure"` @@ -3794,6 +4568,9 @@ type CreateModelInput struct { // instances. Deploying on ML compute instances is part of model hosting. For // more information, see Amazon SageMaker Roles (http://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-roles.html). // + // To be able to pass this role to Amazon SageMaker, the caller of this API + // must have the iam:PassRole permission. + // // ExecutionRoleArn is a required field ExecutionRoleArn *string `min:"20" type:"string" required:"true"` @@ -3815,7 +4592,7 @@ type CreateModelInput struct { Tags []*Tag `type:"list"` // A object that specifies the VPC that you want your model to connect to. Control - // access to and from your training container by configuring the VPC. For more + // access to and from your model container by configuring the VPC. For more // information, see host-vpc. VpcConfig *VpcConfig `type:"structure"` } @@ -3966,6 +4743,9 @@ type CreateNotebookInstanceInput struct { // permissions to assume this role. For more information, see Amazon SageMaker // Roles (http://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-roles.html). // + // To be able to pass this role to Amazon SageMaker, the caller of this API + // must have the iam:PassRole permission. + // // RoleArn is a required field RoleArn *string `min:"20" type:"string" required:"true"` @@ -4285,10 +5065,10 @@ type CreateTrainingJobInput struct { // AlgorithmSpecification is a required field AlgorithmSpecification *AlgorithmSpecification `type:"structure" required:"true"` - // Algorithm-specific parameters. You set hyperparameters before you start the - // learning process. Hyperparameters influence the quality of the model. For - // a list of hyperparameters for each training algorithm provided by Amazon - // SageMaker, see Algorithms (http://docs.aws.amazon.com/sagemaker/latest/dg/algos.html). + // Algorithm-specific parameters that influence the quality of the model. You + // set hyperparameters before you start the learning process. For a list of + // hyperparameters for each training algorithm provided by Amazon SageMaker, + // see Algorithms (http://docs.aws.amazon.com/sagemaker/latest/dg/algos.html). // // You can specify a maximum of 100 hyperparameters. Each hyperparameter is // a key-value pair. Each key and value is limited to 256 characters, as specified @@ -4339,6 +5119,9 @@ type CreateTrainingJobInput struct { // these tasks to an IAM role. For more information, see Amazon SageMaker Roles // (http://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-roles.html). // + // To be able to pass this role to Amazon SageMaker, the caller of this API + // must have the iam:PassRole permission. + // // RoleArn is a required field RoleArn *string `min:"20" type:"string" required:"true"` @@ -5169,6 +5952,190 @@ func (s *DescribeEndpointOutput) SetProductionVariants(v []*ProductionVariantSum return s } +type DescribeHyperParameterTuningJobInput struct { + _ struct{} `type:"structure"` + + // The name of the tuning job to describe. + // + // HyperParameterTuningJobName is a required field + HyperParameterTuningJobName *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeHyperParameterTuningJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeHyperParameterTuningJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeHyperParameterTuningJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeHyperParameterTuningJobInput"} + if s.HyperParameterTuningJobName == nil { + invalidParams.Add(request.NewErrParamRequired("HyperParameterTuningJobName")) + } + if s.HyperParameterTuningJobName != nil && len(*s.HyperParameterTuningJobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("HyperParameterTuningJobName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetHyperParameterTuningJobName sets the HyperParameterTuningJobName field's value. +func (s *DescribeHyperParameterTuningJobInput) SetHyperParameterTuningJobName(v string) *DescribeHyperParameterTuningJobInput { + s.HyperParameterTuningJobName = &v + return s +} + +type DescribeHyperParameterTuningJobOutput struct { + _ struct{} `type:"structure"` + + // A object that describes the training job that completed with the best current + // . + BestTrainingJob *HyperParameterTrainingJobSummary `type:"structure"` + + // The date and time that the tuning job started. + // + // CreationTime is a required field + CreationTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` + + // If the tuning job failed, the reason it failed. + FailureReason *string `type:"string"` + + // The date and time that the tuning job ended. + HyperParameterTuningEndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The Amazon Resource Name (ARN) of the tuning job. + // + // HyperParameterTuningJobArn is a required field + HyperParameterTuningJobArn *string `type:"string" required:"true"` + + // The object that specifies the configuration of the tuning job. + // + // HyperParameterTuningJobConfig is a required field + HyperParameterTuningJobConfig *HyperParameterTuningJobConfig `type:"structure" required:"true"` + + // The name of the tuning job. + // + // HyperParameterTuningJobName is a required field + HyperParameterTuningJobName *string `min:"1" type:"string" required:"true"` + + // The status of the tuning job: InProgress, Completed, Failed, Stopping, or + // Stopped. + // + // HyperParameterTuningJobStatus is a required field + HyperParameterTuningJobStatus *string `type:"string" required:"true" enum:"HyperParameterTuningJobStatus"` + + // The date and time that the status of the tuning job was modified. + LastModifiedTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The object that specifies the number of training jobs, categorized by the + // status of their final objective metric, that this tuning job launched. + // + // ObjectiveStatusCounters is a required field + ObjectiveStatusCounters *ObjectiveStatusCounters `type:"structure" required:"true"` + + // The object that specifies the definition of the training jobs that this tuning + // job launches. + // + // TrainingJobDefinition is a required field + TrainingJobDefinition *HyperParameterTrainingJobDefinition `type:"structure" required:"true"` + + // The object that specifies the number of training jobs, categorized by status, + // that this tuning job launched. + // + // TrainingJobStatusCounters is a required field + TrainingJobStatusCounters *TrainingJobStatusCounters `type:"structure" required:"true"` +} + +// String returns the string representation +func (s DescribeHyperParameterTuningJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeHyperParameterTuningJobOutput) GoString() string { + return s.String() +} + +// SetBestTrainingJob sets the BestTrainingJob field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetBestTrainingJob(v *HyperParameterTrainingJobSummary) *DescribeHyperParameterTuningJobOutput { + s.BestTrainingJob = v + return s +} + +// SetCreationTime sets the CreationTime field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetCreationTime(v time.Time) *DescribeHyperParameterTuningJobOutput { + s.CreationTime = &v + return s +} + +// SetFailureReason sets the FailureReason field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetFailureReason(v string) *DescribeHyperParameterTuningJobOutput { + s.FailureReason = &v + return s +} + +// SetHyperParameterTuningEndTime sets the HyperParameterTuningEndTime field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetHyperParameterTuningEndTime(v time.Time) *DescribeHyperParameterTuningJobOutput { + s.HyperParameterTuningEndTime = &v + return s +} + +// SetHyperParameterTuningJobArn sets the HyperParameterTuningJobArn field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetHyperParameterTuningJobArn(v string) *DescribeHyperParameterTuningJobOutput { + s.HyperParameterTuningJobArn = &v + return s +} + +// SetHyperParameterTuningJobConfig sets the HyperParameterTuningJobConfig field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetHyperParameterTuningJobConfig(v *HyperParameterTuningJobConfig) *DescribeHyperParameterTuningJobOutput { + s.HyperParameterTuningJobConfig = v + return s +} + +// SetHyperParameterTuningJobName sets the HyperParameterTuningJobName field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetHyperParameterTuningJobName(v string) *DescribeHyperParameterTuningJobOutput { + s.HyperParameterTuningJobName = &v + return s +} + +// SetHyperParameterTuningJobStatus sets the HyperParameterTuningJobStatus field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetHyperParameterTuningJobStatus(v string) *DescribeHyperParameterTuningJobOutput { + s.HyperParameterTuningJobStatus = &v + return s +} + +// SetLastModifiedTime sets the LastModifiedTime field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetLastModifiedTime(v time.Time) *DescribeHyperParameterTuningJobOutput { + s.LastModifiedTime = &v + return s +} + +// SetObjectiveStatusCounters sets the ObjectiveStatusCounters field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetObjectiveStatusCounters(v *ObjectiveStatusCounters) *DescribeHyperParameterTuningJobOutput { + s.ObjectiveStatusCounters = v + return s +} + +// SetTrainingJobDefinition sets the TrainingJobDefinition field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetTrainingJobDefinition(v *HyperParameterTrainingJobDefinition) *DescribeHyperParameterTuningJobOutput { + s.TrainingJobDefinition = v + return s +} + +// SetTrainingJobStatusCounters sets the TrainingJobStatusCounters field's value. +func (s *DescribeHyperParameterTuningJobOutput) SetTrainingJobStatusCounters(v *TrainingJobStatusCounters) *DescribeHyperParameterTuningJobOutput { + s.TrainingJobStatusCounters = v + return s +} + type DescribeModelInput struct { _ struct{} `type:"structure"` @@ -5693,7 +6660,11 @@ type DescribeTrainingJobOutput struct { // StoppingCondition is a required field StoppingCondition *StoppingCondition `type:"structure" required:"true"` - // A timestamp that indicates when model training ended. + // Indicates the time when the training job ends on training instances. You + // are billed for the time interval between the value of TrainingStartTime and + // this time. For successful jobs and stopped jobs, this is the time after model + // artifacts are uploaded. For failed jobs, this is the time when Amazon SageMaker + // detects a job failure. TrainingEndTime *time.Time `type:"timestamp" timestampFormat:"unix"` // The Amazon Resource Name (ARN) of the training job. @@ -5728,9 +6699,17 @@ type DescribeTrainingJobOutput struct { // TrainingJobStatus is a required field TrainingJobStatus *string `type:"string" required:"true" enum:"TrainingJobStatus"` - // A timestamp that indicates when training started. + // Indicates the time when the training job starts on training instances. You + // are billed for the time interval between this time and the value of TrainingEndTime. + // The start time in CloudWatch Logs might be later than this time. The difference + // is due to the time it takes to download the training data and to the size + // of the training container. TrainingStartTime *time.Time `type:"timestamp" timestampFormat:"unix"` + // The Amazon Resource Name (ARN) of the associated hyperparameter tuning job + // if the training job was launched by a hyperparameter tuning job. + TuningJobArn *string `type:"string"` + // A object that specifies the VPC that this training job has access to. For // more information, see train-vpc. VpcConfig *VpcConfig `type:"structure"` @@ -5848,6 +6827,12 @@ func (s *DescribeTrainingJobOutput) SetTrainingStartTime(v time.Time) *DescribeT return s } +// SetTuningJobArn sets the TuningJobArn field's value. +func (s *DescribeTrainingJobOutput) SetTuningJobArn(v string) *DescribeTrainingJobOutput { + s.TuningJobArn = &v + return s +} + // SetVpcConfig sets the VpcConfig field's value. func (s *DescribeTrainingJobOutput) SetVpcConfig(v *VpcConfig) *DescribeTrainingJobOutput { s.VpcConfig = v @@ -6032,6 +7017,810 @@ func (s *EndpointSummary) SetLastModifiedTime(v time.Time) *EndpointSummary { return s } +// Shows the final value for the objective metric for a training job that was +// launched by a hyperparameter tuning job. You define the objective metric +// in the HyperParameterTuningJobObjective parameter of HyperParameterTuningJobConfig. +type FinalHyperParameterTuningJobObjectiveMetric struct { + _ struct{} `type:"structure"` + + // The name of the objective metric. + // + // MetricName is a required field + MetricName *string `min:"1" type:"string" required:"true"` + + // Whether to minimize or maximize the objective metric. Valid values are Minimize + // and Maximize. + Type *string `type:"string" enum:"HyperParameterTuningJobObjectiveType"` + + // The value of the objective metric. + // + // Value is a required field + Value *float64 `type:"float" required:"true"` +} + +// String returns the string representation +func (s FinalHyperParameterTuningJobObjectiveMetric) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FinalHyperParameterTuningJobObjectiveMetric) GoString() string { + return s.String() +} + +// SetMetricName sets the MetricName field's value. +func (s *FinalHyperParameterTuningJobObjectiveMetric) SetMetricName(v string) *FinalHyperParameterTuningJobObjectiveMetric { + s.MetricName = &v + return s +} + +// SetType sets the Type field's value. +func (s *FinalHyperParameterTuningJobObjectiveMetric) SetType(v string) *FinalHyperParameterTuningJobObjectiveMetric { + s.Type = &v + return s +} + +// SetValue sets the Value field's value. +func (s *FinalHyperParameterTuningJobObjectiveMetric) SetValue(v float64) *FinalHyperParameterTuningJobObjectiveMetric { + s.Value = &v + return s +} + +// Specifies which training algorithm to use for training jobs that a hyperparameter +// tuning job launches and the metrics to monitor. +type HyperParameterAlgorithmSpecification struct { + _ struct{} `type:"structure"` + + // An array of objects that specify the metrics that the algorithm emits. + MetricDefinitions []*MetricDefinition `type:"list"` + + // The registry path of the Docker image that contains the training algorithm. + // For information about Docker registry paths for built-in algorithms, see + // sagemaker-algo-docker-registry-paths. + // + // TrainingImage is a required field + TrainingImage *string `type:"string" required:"true"` + + // The input mode that the algorithm supports: File or Pipe. In File input mode, + // Amazon SageMaker downloads the training data from Amazon S3 to the storage + // volume that is attached to the training instance and mounts the directory + // to the Docker volume for the training container. In Pipe input mode, Amazon + // SageMaker streams data directly from Amazon S3 to the container. + // + // If you specify File mode, make sure that you provision the storage volume + // that is attached to the training instance with enough capacity to accommodate + // the training data downloaded from Amazon S3, the model artifacts, and intermediate + // information. + // + // For more information about input modes, see Algorithms (http://docs.aws.amazon.com/sagemaker/latest/dg/algos.html) + // + // TrainingInputMode is a required field + TrainingInputMode *string `type:"string" required:"true" enum:"TrainingInputMode"` +} + +// String returns the string representation +func (s HyperParameterAlgorithmSpecification) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HyperParameterAlgorithmSpecification) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *HyperParameterAlgorithmSpecification) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HyperParameterAlgorithmSpecification"} + if s.TrainingImage == nil { + invalidParams.Add(request.NewErrParamRequired("TrainingImage")) + } + if s.TrainingInputMode == nil { + invalidParams.Add(request.NewErrParamRequired("TrainingInputMode")) + } + if s.MetricDefinitions != nil { + for i, v := range s.MetricDefinitions { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "MetricDefinitions", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMetricDefinitions sets the MetricDefinitions field's value. +func (s *HyperParameterAlgorithmSpecification) SetMetricDefinitions(v []*MetricDefinition) *HyperParameterAlgorithmSpecification { + s.MetricDefinitions = v + return s +} + +// SetTrainingImage sets the TrainingImage field's value. +func (s *HyperParameterAlgorithmSpecification) SetTrainingImage(v string) *HyperParameterAlgorithmSpecification { + s.TrainingImage = &v + return s +} + +// SetTrainingInputMode sets the TrainingInputMode field's value. +func (s *HyperParameterAlgorithmSpecification) SetTrainingInputMode(v string) *HyperParameterAlgorithmSpecification { + s.TrainingInputMode = &v + return s +} + +// Defines the training jobs launched by a hyperparameter tuning job. +type HyperParameterTrainingJobDefinition struct { + _ struct{} `type:"structure"` + + // The object that specifies the algorithm to use for the training jobs that + // the tuning job launches. + // + // AlgorithmSpecification is a required field + AlgorithmSpecification *HyperParameterAlgorithmSpecification `type:"structure" required:"true"` + + // An array of objects that specify the input for the training jobs that the + // tuning job launches. + // + // InputDataConfig is a required field + InputDataConfig []*Channel `min:"1" type:"list" required:"true"` + + // Specifies the path to the Amazon S3 bucket where you store model artifacts + // from the training jobs that the tuning job launches. + // + // OutputDataConfig is a required field + OutputDataConfig *OutputDataConfig `type:"structure" required:"true"` + + // The resources, including the compute instances and storage volumes, to use + // for the training jobs that the tuning job launches. + // + // Storage volumes store model artifacts and incremental states. Training algorithms + // might also use storage volumes for scratch space. If you want Amazon SageMaker + // to use the storage volume to store the training data, choose File as the + // TrainingInputMode in the algorithm specification. For distributed training + // algorithms, specify an instance count greater than 1. + // + // ResourceConfig is a required field + ResourceConfig *ResourceConfig `type:"structure" required:"true"` + + // The Amazon Resource Name (ARN) of the IAM role associated with the training + // jobs that the tuning job launches. + // + // RoleArn is a required field + RoleArn *string `min:"20" type:"string" required:"true"` + + // Specifies the values of hyperparameters that do not change for the tuning + // job. + StaticHyperParameters map[string]*string `type:"map"` + + // Sets a maximum duration for the training jobs that the tuning job launches. + // Use this parameter to limit model training costs. + // + // To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal. This + // delays job termination for 120 seconds. Algorithms might use this 120-second + // window to save the model artifacts. + // + // When Amazon SageMaker terminates a job because the stopping condition has + // been met, training algorithms provided by Amazon SageMaker save the intermediate + // results of the job. + // + // StoppingCondition is a required field + StoppingCondition *StoppingCondition `type:"structure" required:"true"` + + // The object that specifies the VPC that you want the training jobs that this + // hyperparameter tuning job launches to connect to. Control access to and from + // your training container by configuring the VPC. For more information, see + // train-vpc. + VpcConfig *VpcConfig `type:"structure"` +} + +// String returns the string representation +func (s HyperParameterTrainingJobDefinition) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HyperParameterTrainingJobDefinition) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *HyperParameterTrainingJobDefinition) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HyperParameterTrainingJobDefinition"} + if s.AlgorithmSpecification == nil { + invalidParams.Add(request.NewErrParamRequired("AlgorithmSpecification")) + } + if s.InputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("InputDataConfig")) + } + if s.InputDataConfig != nil && len(s.InputDataConfig) < 1 { + invalidParams.Add(request.NewErrParamMinLen("InputDataConfig", 1)) + } + if s.OutputDataConfig == nil { + invalidParams.Add(request.NewErrParamRequired("OutputDataConfig")) + } + if s.ResourceConfig == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceConfig")) + } + if s.RoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("RoleArn")) + } + if s.RoleArn != nil && len(*s.RoleArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("RoleArn", 20)) + } + if s.StoppingCondition == nil { + invalidParams.Add(request.NewErrParamRequired("StoppingCondition")) + } + if s.AlgorithmSpecification != nil { + if err := s.AlgorithmSpecification.Validate(); err != nil { + invalidParams.AddNested("AlgorithmSpecification", err.(request.ErrInvalidParams)) + } + } + if s.InputDataConfig != nil { + for i, v := range s.InputDataConfig { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "InputDataConfig", i), err.(request.ErrInvalidParams)) + } + } + } + if s.OutputDataConfig != nil { + if err := s.OutputDataConfig.Validate(); err != nil { + invalidParams.AddNested("OutputDataConfig", err.(request.ErrInvalidParams)) + } + } + if s.ResourceConfig != nil { + if err := s.ResourceConfig.Validate(); err != nil { + invalidParams.AddNested("ResourceConfig", err.(request.ErrInvalidParams)) + } + } + if s.StoppingCondition != nil { + if err := s.StoppingCondition.Validate(); err != nil { + invalidParams.AddNested("StoppingCondition", err.(request.ErrInvalidParams)) + } + } + if s.VpcConfig != nil { + if err := s.VpcConfig.Validate(); err != nil { + invalidParams.AddNested("VpcConfig", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAlgorithmSpecification sets the AlgorithmSpecification field's value. +func (s *HyperParameterTrainingJobDefinition) SetAlgorithmSpecification(v *HyperParameterAlgorithmSpecification) *HyperParameterTrainingJobDefinition { + s.AlgorithmSpecification = v + return s +} + +// SetInputDataConfig sets the InputDataConfig field's value. +func (s *HyperParameterTrainingJobDefinition) SetInputDataConfig(v []*Channel) *HyperParameterTrainingJobDefinition { + s.InputDataConfig = v + return s +} + +// SetOutputDataConfig sets the OutputDataConfig field's value. +func (s *HyperParameterTrainingJobDefinition) SetOutputDataConfig(v *OutputDataConfig) *HyperParameterTrainingJobDefinition { + s.OutputDataConfig = v + return s +} + +// SetResourceConfig sets the ResourceConfig field's value. +func (s *HyperParameterTrainingJobDefinition) SetResourceConfig(v *ResourceConfig) *HyperParameterTrainingJobDefinition { + s.ResourceConfig = v + return s +} + +// SetRoleArn sets the RoleArn field's value. +func (s *HyperParameterTrainingJobDefinition) SetRoleArn(v string) *HyperParameterTrainingJobDefinition { + s.RoleArn = &v + return s +} + +// SetStaticHyperParameters sets the StaticHyperParameters field's value. +func (s *HyperParameterTrainingJobDefinition) SetStaticHyperParameters(v map[string]*string) *HyperParameterTrainingJobDefinition { + s.StaticHyperParameters = v + return s +} + +// SetStoppingCondition sets the StoppingCondition field's value. +func (s *HyperParameterTrainingJobDefinition) SetStoppingCondition(v *StoppingCondition) *HyperParameterTrainingJobDefinition { + s.StoppingCondition = v + return s +} + +// SetVpcConfig sets the VpcConfig field's value. +func (s *HyperParameterTrainingJobDefinition) SetVpcConfig(v *VpcConfig) *HyperParameterTrainingJobDefinition { + s.VpcConfig = v + return s +} + +// Specifies summary information about a training job. +type HyperParameterTrainingJobSummary struct { + _ struct{} `type:"structure"` + + // The date and time that the training job was created. + // + // CreationTime is a required field + CreationTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` + + // The reason that the + FailureReason *string `type:"string"` + + // The object that specifies the value of the objective metric of the tuning + // job that launched this training job. + FinalHyperParameterTuningJobObjectiveMetric *FinalHyperParameterTuningJobObjectiveMetric `type:"structure"` + + // The status of the objective metric for the training job: + // + // * Succeeded: The final objective metric for the training job was evaluated + // by the hyperparameter tuning job and used in the hyperparameter tuning + // process. + // + // * Pending: The training job is in progress and evaluation of its final + // objective metric is pending. + // + // * Failed: The final objective metric for the training job was not evaluated, + // and was not used in the hyperparameter tuning process. This typically + // occurs when the training job failed or did not emit an objective metric. + ObjectiveStatus *string `type:"string" enum:"ObjectiveStatus"` + + // The date and time that the training job ended. + TrainingEndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The Amazon Resource Name (ARN) of the training job. + // + // TrainingJobArn is a required field + TrainingJobArn *string `type:"string" required:"true"` + + // The name of the training job. + // + // TrainingJobName is a required field + TrainingJobName *string `min:"1" type:"string" required:"true"` + + // The status of the training job. + // + // TrainingJobStatus is a required field + TrainingJobStatus *string `type:"string" required:"true" enum:"TrainingJobStatus"` + + // The date and time that the training job started. + TrainingStartTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // A list of the hyperparameters for which you specified ranges to search. + // + // TunedHyperParameters is a required field + TunedHyperParameters map[string]*string `type:"map" required:"true"` +} + +// String returns the string representation +func (s HyperParameterTrainingJobSummary) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HyperParameterTrainingJobSummary) GoString() string { + return s.String() +} + +// SetCreationTime sets the CreationTime field's value. +func (s *HyperParameterTrainingJobSummary) SetCreationTime(v time.Time) *HyperParameterTrainingJobSummary { + s.CreationTime = &v + return s +} + +// SetFailureReason sets the FailureReason field's value. +func (s *HyperParameterTrainingJobSummary) SetFailureReason(v string) *HyperParameterTrainingJobSummary { + s.FailureReason = &v + return s +} + +// SetFinalHyperParameterTuningJobObjectiveMetric sets the FinalHyperParameterTuningJobObjectiveMetric field's value. +func (s *HyperParameterTrainingJobSummary) SetFinalHyperParameterTuningJobObjectiveMetric(v *FinalHyperParameterTuningJobObjectiveMetric) *HyperParameterTrainingJobSummary { + s.FinalHyperParameterTuningJobObjectiveMetric = v + return s +} + +// SetObjectiveStatus sets the ObjectiveStatus field's value. +func (s *HyperParameterTrainingJobSummary) SetObjectiveStatus(v string) *HyperParameterTrainingJobSummary { + s.ObjectiveStatus = &v + return s +} + +// SetTrainingEndTime sets the TrainingEndTime field's value. +func (s *HyperParameterTrainingJobSummary) SetTrainingEndTime(v time.Time) *HyperParameterTrainingJobSummary { + s.TrainingEndTime = &v + return s +} + +// SetTrainingJobArn sets the TrainingJobArn field's value. +func (s *HyperParameterTrainingJobSummary) SetTrainingJobArn(v string) *HyperParameterTrainingJobSummary { + s.TrainingJobArn = &v + return s +} + +// SetTrainingJobName sets the TrainingJobName field's value. +func (s *HyperParameterTrainingJobSummary) SetTrainingJobName(v string) *HyperParameterTrainingJobSummary { + s.TrainingJobName = &v + return s +} + +// SetTrainingJobStatus sets the TrainingJobStatus field's value. +func (s *HyperParameterTrainingJobSummary) SetTrainingJobStatus(v string) *HyperParameterTrainingJobSummary { + s.TrainingJobStatus = &v + return s +} + +// SetTrainingStartTime sets the TrainingStartTime field's value. +func (s *HyperParameterTrainingJobSummary) SetTrainingStartTime(v time.Time) *HyperParameterTrainingJobSummary { + s.TrainingStartTime = &v + return s +} + +// SetTunedHyperParameters sets the TunedHyperParameters field's value. +func (s *HyperParameterTrainingJobSummary) SetTunedHyperParameters(v map[string]*string) *HyperParameterTrainingJobSummary { + s.TunedHyperParameters = v + return s +} + +// Configures a hyperparameter tuning job. +type HyperParameterTuningJobConfig struct { + _ struct{} `type:"structure"` + + // The object that specifies the objective metric for this tuning job. + // + // HyperParameterTuningJobObjective is a required field + HyperParameterTuningJobObjective *HyperParameterTuningJobObjective `type:"structure" required:"true"` + + // The object that specifies the ranges of hyperparameters that this tuning + // job searches. + // + // ParameterRanges is a required field + ParameterRanges *ParameterRanges `type:"structure" required:"true"` + + // The object that specifies the maximum number of training jobs and parallel + // training jobs for this tuning job. + // + // ResourceLimits is a required field + ResourceLimits *ResourceLimits `type:"structure" required:"true"` + + // Specifies the search strategy for hyperparameters. Currently, the only valid + // value is Bayesian. + // + // Strategy is a required field + Strategy *string `type:"string" required:"true" enum:"HyperParameterTuningJobStrategyType"` +} + +// String returns the string representation +func (s HyperParameterTuningJobConfig) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HyperParameterTuningJobConfig) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *HyperParameterTuningJobConfig) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HyperParameterTuningJobConfig"} + if s.HyperParameterTuningJobObjective == nil { + invalidParams.Add(request.NewErrParamRequired("HyperParameterTuningJobObjective")) + } + if s.ParameterRanges == nil { + invalidParams.Add(request.NewErrParamRequired("ParameterRanges")) + } + if s.ResourceLimits == nil { + invalidParams.Add(request.NewErrParamRequired("ResourceLimits")) + } + if s.Strategy == nil { + invalidParams.Add(request.NewErrParamRequired("Strategy")) + } + if s.HyperParameterTuningJobObjective != nil { + if err := s.HyperParameterTuningJobObjective.Validate(); err != nil { + invalidParams.AddNested("HyperParameterTuningJobObjective", err.(request.ErrInvalidParams)) + } + } + if s.ParameterRanges != nil { + if err := s.ParameterRanges.Validate(); err != nil { + invalidParams.AddNested("ParameterRanges", err.(request.ErrInvalidParams)) + } + } + if s.ResourceLimits != nil { + if err := s.ResourceLimits.Validate(); err != nil { + invalidParams.AddNested("ResourceLimits", err.(request.ErrInvalidParams)) + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetHyperParameterTuningJobObjective sets the HyperParameterTuningJobObjective field's value. +func (s *HyperParameterTuningJobConfig) SetHyperParameterTuningJobObjective(v *HyperParameterTuningJobObjective) *HyperParameterTuningJobConfig { + s.HyperParameterTuningJobObjective = v + return s +} + +// SetParameterRanges sets the ParameterRanges field's value. +func (s *HyperParameterTuningJobConfig) SetParameterRanges(v *ParameterRanges) *HyperParameterTuningJobConfig { + s.ParameterRanges = v + return s +} + +// SetResourceLimits sets the ResourceLimits field's value. +func (s *HyperParameterTuningJobConfig) SetResourceLimits(v *ResourceLimits) *HyperParameterTuningJobConfig { + s.ResourceLimits = v + return s +} + +// SetStrategy sets the Strategy field's value. +func (s *HyperParameterTuningJobConfig) SetStrategy(v string) *HyperParameterTuningJobConfig { + s.Strategy = &v + return s +} + +// Defines the objective metric for a hyperparameter tuning job. Hyperparameter +// tuning uses the value of this metric to evaluate the training jobs it launches, +// and returns the training job that results in either the highest or lowest +// value for this metric, depending on the value you specify for the Type parameter. +type HyperParameterTuningJobObjective struct { + _ struct{} `type:"structure"` + + // The name of the metric to use for the objective metric. + // + // MetricName is a required field + MetricName *string `min:"1" type:"string" required:"true"` + + // Whether to minimize or maximize the objective metric. + // + // Type is a required field + Type *string `type:"string" required:"true" enum:"HyperParameterTuningJobObjectiveType"` +} + +// String returns the string representation +func (s HyperParameterTuningJobObjective) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HyperParameterTuningJobObjective) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *HyperParameterTuningJobObjective) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "HyperParameterTuningJobObjective"} + if s.MetricName == nil { + invalidParams.Add(request.NewErrParamRequired("MetricName")) + } + if s.MetricName != nil && len(*s.MetricName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("MetricName", 1)) + } + if s.Type == nil { + invalidParams.Add(request.NewErrParamRequired("Type")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMetricName sets the MetricName field's value. +func (s *HyperParameterTuningJobObjective) SetMetricName(v string) *HyperParameterTuningJobObjective { + s.MetricName = &v + return s +} + +// SetType sets the Type field's value. +func (s *HyperParameterTuningJobObjective) SetType(v string) *HyperParameterTuningJobObjective { + s.Type = &v + return s +} + +// Provides summary information about a hyperparameter tuning job. +type HyperParameterTuningJobSummary struct { + _ struct{} `type:"structure"` + + // The date and time that the tuning job was created. + // + // CreationTime is a required field + CreationTime *time.Time `type:"timestamp" timestampFormat:"unix" required:"true"` + + // The date and time that the tuning job ended. + HyperParameterTuningEndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The Amazon Resource Name (ARN) of the tuning job. + // + // HyperParameterTuningJobArn is a required field + HyperParameterTuningJobArn *string `type:"string" required:"true"` + + // The name of the tuning job. + // + // HyperParameterTuningJobName is a required field + HyperParameterTuningJobName *string `min:"1" type:"string" required:"true"` + + // The status of the tuning job. + // + // HyperParameterTuningJobStatus is a required field + HyperParameterTuningJobStatus *string `type:"string" required:"true" enum:"HyperParameterTuningJobStatus"` + + // The date and time that the tuning job was modified. + LastModifiedTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The object that specifies the numbers of training jobs, categorized by objective + // metric status, that this tuning job launched. + // + // ObjectiveStatusCounters is a required field + ObjectiveStatusCounters *ObjectiveStatusCounters `type:"structure" required:"true"` + + // The object that specifies the maximum number of training jobs and parallel + // training jobs allowed for this tuning job. + ResourceLimits *ResourceLimits `type:"structure"` + + // Specifies the search strategy hyperparameter tuning uses to choose which + // hyperparameters to use for each iteration. Currently, the only valid value + // is Bayesian. + // + // Strategy is a required field + Strategy *string `type:"string" required:"true" enum:"HyperParameterTuningJobStrategyType"` + + // The object that specifies the numbers of training jobs, categorized by status, + // that this tuning job launched. + // + // TrainingJobStatusCounters is a required field + TrainingJobStatusCounters *TrainingJobStatusCounters `type:"structure" required:"true"` +} + +// String returns the string representation +func (s HyperParameterTuningJobSummary) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s HyperParameterTuningJobSummary) GoString() string { + return s.String() +} + +// SetCreationTime sets the CreationTime field's value. +func (s *HyperParameterTuningJobSummary) SetCreationTime(v time.Time) *HyperParameterTuningJobSummary { + s.CreationTime = &v + return s +} + +// SetHyperParameterTuningEndTime sets the HyperParameterTuningEndTime field's value. +func (s *HyperParameterTuningJobSummary) SetHyperParameterTuningEndTime(v time.Time) *HyperParameterTuningJobSummary { + s.HyperParameterTuningEndTime = &v + return s +} + +// SetHyperParameterTuningJobArn sets the HyperParameterTuningJobArn field's value. +func (s *HyperParameterTuningJobSummary) SetHyperParameterTuningJobArn(v string) *HyperParameterTuningJobSummary { + s.HyperParameterTuningJobArn = &v + return s +} + +// SetHyperParameterTuningJobName sets the HyperParameterTuningJobName field's value. +func (s *HyperParameterTuningJobSummary) SetHyperParameterTuningJobName(v string) *HyperParameterTuningJobSummary { + s.HyperParameterTuningJobName = &v + return s +} + +// SetHyperParameterTuningJobStatus sets the HyperParameterTuningJobStatus field's value. +func (s *HyperParameterTuningJobSummary) SetHyperParameterTuningJobStatus(v string) *HyperParameterTuningJobSummary { + s.HyperParameterTuningJobStatus = &v + return s +} + +// SetLastModifiedTime sets the LastModifiedTime field's value. +func (s *HyperParameterTuningJobSummary) SetLastModifiedTime(v time.Time) *HyperParameterTuningJobSummary { + s.LastModifiedTime = &v + return s +} + +// SetObjectiveStatusCounters sets the ObjectiveStatusCounters field's value. +func (s *HyperParameterTuningJobSummary) SetObjectiveStatusCounters(v *ObjectiveStatusCounters) *HyperParameterTuningJobSummary { + s.ObjectiveStatusCounters = v + return s +} + +// SetResourceLimits sets the ResourceLimits field's value. +func (s *HyperParameterTuningJobSummary) SetResourceLimits(v *ResourceLimits) *HyperParameterTuningJobSummary { + s.ResourceLimits = v + return s +} + +// SetStrategy sets the Strategy field's value. +func (s *HyperParameterTuningJobSummary) SetStrategy(v string) *HyperParameterTuningJobSummary { + s.Strategy = &v + return s +} + +// SetTrainingJobStatusCounters sets the TrainingJobStatusCounters field's value. +func (s *HyperParameterTuningJobSummary) SetTrainingJobStatusCounters(v *TrainingJobStatusCounters) *HyperParameterTuningJobSummary { + s.TrainingJobStatusCounters = v + return s +} + +// For a hyperparameter of the integer type, specifies the range that a hyperparameter +// tuning job searches. +type IntegerParameterRange struct { + _ struct{} `type:"structure"` + + // The maximum value of the hyperparameter to search. + // + // MaxValue is a required field + MaxValue *string `type:"string" required:"true"` + + // The minimum value of the hyperparameter to search. + // + // MinValue is a required field + MinValue *string `type:"string" required:"true"` + + // The name of the hyperparameter to search. + // + // Name is a required field + Name *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s IntegerParameterRange) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s IntegerParameterRange) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *IntegerParameterRange) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "IntegerParameterRange"} + if s.MaxValue == nil { + invalidParams.Add(request.NewErrParamRequired("MaxValue")) + } + if s.MinValue == nil { + invalidParams.Add(request.NewErrParamRequired("MinValue")) + } + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMaxValue sets the MaxValue field's value. +func (s *IntegerParameterRange) SetMaxValue(v string) *IntegerParameterRange { + s.MaxValue = &v + return s +} + +// SetMinValue sets the MinValue field's value. +func (s *IntegerParameterRange) SetMinValue(v string) *IntegerParameterRange { + s.MinValue = &v + return s +} + +// SetName sets the Name field's value. +func (s *IntegerParameterRange) SetName(v string) *IntegerParameterRange { + s.Name = &v + return s +} + type ListEndpointConfigsInput struct { _ struct{} `type:"structure"` @@ -6321,6 +8110,167 @@ func (s *ListEndpointsOutput) SetNextToken(v string) *ListEndpointsOutput { return s } +type ListHyperParameterTuningJobsInput struct { + _ struct{} `type:"structure"` + + // A filter that returns only tuning jobs that were created after the specified + // time. + CreationTimeAfter *time.Time `type:"timestamp" timestampFormat:"unix"` + + // A filter that returns only tuning jobs that were created before the specified + // time. + CreationTimeBefore *time.Time `type:"timestamp" timestampFormat:"unix"` + + // A filter that returns only tuning jobs that were modified after the specified + // time. + LastModifiedTimeAfter *time.Time `type:"timestamp" timestampFormat:"unix"` + + // A filter that returns only tuning jobs that were modified before the specified + // time. + LastModifiedTimeBefore *time.Time `type:"timestamp" timestampFormat:"unix"` + + // The maximum number of tuning jobs to return. + MaxResults *int64 `min:"1" type:"integer"` + + // A string in the tuning job name. This filter returns only tuning jobs whose + // name contains the specified string. + NameContains *string `type:"string"` + + // If the result of the previous ListHyperParameterTuningJobs request was truncated, + // the response includes a NextToken. To retrieve the next set of tuning jobs, + // use the token in the next request. + NextToken *string `type:"string"` + + // The field to sort results by. The default is Name. + SortBy *string `type:"string" enum:"HyperParameterTuningJobSortByOptions"` + + // The sort order for results. The default is Ascending. + SortOrder *string `type:"string" enum:"SortOrder"` + + // A filter that returns only tuning jobs with the specified status. + StatusEquals *string `type:"string" enum:"HyperParameterTuningJobStatus"` +} + +// String returns the string representation +func (s ListHyperParameterTuningJobsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListHyperParameterTuningJobsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListHyperParameterTuningJobsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListHyperParameterTuningJobsInput"} + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCreationTimeAfter sets the CreationTimeAfter field's value. +func (s *ListHyperParameterTuningJobsInput) SetCreationTimeAfter(v time.Time) *ListHyperParameterTuningJobsInput { + s.CreationTimeAfter = &v + return s +} + +// SetCreationTimeBefore sets the CreationTimeBefore field's value. +func (s *ListHyperParameterTuningJobsInput) SetCreationTimeBefore(v time.Time) *ListHyperParameterTuningJobsInput { + s.CreationTimeBefore = &v + return s +} + +// SetLastModifiedTimeAfter sets the LastModifiedTimeAfter field's value. +func (s *ListHyperParameterTuningJobsInput) SetLastModifiedTimeAfter(v time.Time) *ListHyperParameterTuningJobsInput { + s.LastModifiedTimeAfter = &v + return s +} + +// SetLastModifiedTimeBefore sets the LastModifiedTimeBefore field's value. +func (s *ListHyperParameterTuningJobsInput) SetLastModifiedTimeBefore(v time.Time) *ListHyperParameterTuningJobsInput { + s.LastModifiedTimeBefore = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListHyperParameterTuningJobsInput) SetMaxResults(v int64) *ListHyperParameterTuningJobsInput { + s.MaxResults = &v + return s +} + +// SetNameContains sets the NameContains field's value. +func (s *ListHyperParameterTuningJobsInput) SetNameContains(v string) *ListHyperParameterTuningJobsInput { + s.NameContains = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListHyperParameterTuningJobsInput) SetNextToken(v string) *ListHyperParameterTuningJobsInput { + s.NextToken = &v + return s +} + +// SetSortBy sets the SortBy field's value. +func (s *ListHyperParameterTuningJobsInput) SetSortBy(v string) *ListHyperParameterTuningJobsInput { + s.SortBy = &v + return s +} + +// SetSortOrder sets the SortOrder field's value. +func (s *ListHyperParameterTuningJobsInput) SetSortOrder(v string) *ListHyperParameterTuningJobsInput { + s.SortOrder = &v + return s +} + +// SetStatusEquals sets the StatusEquals field's value. +func (s *ListHyperParameterTuningJobsInput) SetStatusEquals(v string) *ListHyperParameterTuningJobsInput { + s.StatusEquals = &v + return s +} + +type ListHyperParameterTuningJobsOutput struct { + _ struct{} `type:"structure"` + + // A list of objects that describe the tuning jobs that the ListHyperParameterTuningJobs + // request returned. + // + // HyperParameterTuningJobSummaries is a required field + HyperParameterTuningJobSummaries []*HyperParameterTuningJobSummary `type:"list" required:"true"` + + // If the result of this ListHyperParameterTuningJobs request was truncated, + // the response includes a NextToken. To retrieve the next set of tuning jobs, + // use the token in the next request. + NextToken *string `type:"string"` +} + +// String returns the string representation +func (s ListHyperParameterTuningJobsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListHyperParameterTuningJobsOutput) GoString() string { + return s.String() +} + +// SetHyperParameterTuningJobSummaries sets the HyperParameterTuningJobSummaries field's value. +func (s *ListHyperParameterTuningJobsOutput) SetHyperParameterTuningJobSummaries(v []*HyperParameterTuningJobSummary) *ListHyperParameterTuningJobsOutput { + s.HyperParameterTuningJobSummaries = v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListHyperParameterTuningJobsOutput) SetNextToken(v string) *ListHyperParameterTuningJobsOutput { + s.NextToken = &v + return s +} + type ListModelsInput struct { _ struct{} `type:"structure"` @@ -6866,6 +8816,134 @@ func (s *ListTagsOutput) SetTags(v []*Tag) *ListTagsOutput { return s } +type ListTrainingJobsForHyperParameterTuningJobInput struct { + _ struct{} `type:"structure"` + + // The name of the tuning job whose training jobs you want to list. + // + // HyperParameterTuningJobName is a required field + HyperParameterTuningJobName *string `min:"1" type:"string" required:"true"` + + // The maximum number of training jobs to return. + MaxResults *int64 `min:"1" type:"integer"` + + // If the result of the previous ListTrainingJobsForHyperParameterTuningJob + // request was truncated, the response includes a NextToken. To retrieve the + // next set of training jobs, use the token in the next request. + NextToken *string `type:"string"` + + // The field to sort results by. The default is Name. + SortBy *string `type:"string" enum:"TrainingJobSortByOptions"` + + // The sort order for results. The default is Ascending. + SortOrder *string `type:"string" enum:"SortOrder"` + + // A filter that returns only training jobs with the specified status. + StatusEquals *string `type:"string" enum:"TrainingJobStatus"` +} + +// String returns the string representation +func (s ListTrainingJobsForHyperParameterTuningJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListTrainingJobsForHyperParameterTuningJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ListTrainingJobsForHyperParameterTuningJobInput"} + if s.HyperParameterTuningJobName == nil { + invalidParams.Add(request.NewErrParamRequired("HyperParameterTuningJobName")) + } + if s.HyperParameterTuningJobName != nil && len(*s.HyperParameterTuningJobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("HyperParameterTuningJobName", 1)) + } + if s.MaxResults != nil && *s.MaxResults < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxResults", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetHyperParameterTuningJobName sets the HyperParameterTuningJobName field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) SetHyperParameterTuningJobName(v string) *ListTrainingJobsForHyperParameterTuningJobInput { + s.HyperParameterTuningJobName = &v + return s +} + +// SetMaxResults sets the MaxResults field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) SetMaxResults(v int64) *ListTrainingJobsForHyperParameterTuningJobInput { + s.MaxResults = &v + return s +} + +// SetNextToken sets the NextToken field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) SetNextToken(v string) *ListTrainingJobsForHyperParameterTuningJobInput { + s.NextToken = &v + return s +} + +// SetSortBy sets the SortBy field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) SetSortBy(v string) *ListTrainingJobsForHyperParameterTuningJobInput { + s.SortBy = &v + return s +} + +// SetSortOrder sets the SortOrder field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) SetSortOrder(v string) *ListTrainingJobsForHyperParameterTuningJobInput { + s.SortOrder = &v + return s +} + +// SetStatusEquals sets the StatusEquals field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobInput) SetStatusEquals(v string) *ListTrainingJobsForHyperParameterTuningJobInput { + s.StatusEquals = &v + return s +} + +type ListTrainingJobsForHyperParameterTuningJobOutput struct { + _ struct{} `type:"structure"` + + // If the result of this ListTrainingJobsForHyperParameterTuningJob request + // was truncated, the response includes a NextToken. To retrieve the next set + // of training jobs, use the token in the next request. + NextToken *string `type:"string"` + + // A list of objects that describe the training jobs that the ListTrainingJobsForHyperParameterTuningJob + // request returned. + // + // TrainingJobSummaries is a required field + TrainingJobSummaries []*HyperParameterTrainingJobSummary `type:"list" required:"true"` +} + +// String returns the string representation +func (s ListTrainingJobsForHyperParameterTuningJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ListTrainingJobsForHyperParameterTuningJobOutput) GoString() string { + return s.String() +} + +// SetNextToken sets the NextToken field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobOutput) SetNextToken(v string) *ListTrainingJobsForHyperParameterTuningJobOutput { + s.NextToken = &v + return s +} + +// SetTrainingJobSummaries sets the TrainingJobSummaries field's value. +func (s *ListTrainingJobsForHyperParameterTuningJobOutput) SetTrainingJobSummaries(v []*HyperParameterTrainingJobSummary) *ListTrainingJobsForHyperParameterTuningJobOutput { + s.TrainingJobSummaries = v + return s +} + type ListTrainingJobsInput struct { _ struct{} `type:"structure"` @@ -6887,8 +8965,8 @@ type ListTrainingJobsInput struct { // The maximum number of training jobs to return in the response. MaxResults *int64 `min:"1" type:"integer"` - // A string in the training job name. This filter returns only models whose - // name contains the specified string. + // A string in the training job name. This filter returns only training jobs + // whose name contains the specified string. NameContains *string `type:"string"` // If the result of the previous ListTrainingJobs request was truncated, the @@ -7024,6 +9102,70 @@ func (s *ListTrainingJobsOutput) SetTrainingJobSummaries(v []*TrainingJobSummary return s } +// Specifies a metric that the training algorithm writes to stderr or stdout. +// Amazon SageMakerHyperparamter tuning captures all defined metrics. You specify +// one metric that a hyperparameter tuning job uses as its objective metric +// to choose the best training job. +type MetricDefinition struct { + _ struct{} `type:"structure"` + + // The name of the metric. + // + // Name is a required field + Name *string `min:"1" type:"string" required:"true"` + + // A regular expression that searches the output of a training job and gets + // the value of the metric. For more information about using regular expressions + // to define metrics, see hpo-define-metrics. + // + // Regex is a required field + Regex *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s MetricDefinition) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s MetricDefinition) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *MetricDefinition) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "MetricDefinition"} + if s.Name == nil { + invalidParams.Add(request.NewErrParamRequired("Name")) + } + if s.Name != nil && len(*s.Name) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Name", 1)) + } + if s.Regex == nil { + invalidParams.Add(request.NewErrParamRequired("Regex")) + } + if s.Regex != nil && len(*s.Regex) < 1 { + invalidParams.Add(request.NewErrParamMinLen("Regex", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetName sets the Name field's value. +func (s *MetricDefinition) SetName(v string) *MetricDefinition { + s.Name = &v + return s +} + +// SetRegex sets the Regex field's value. +func (s *MetricDefinition) SetRegex(v string) *MetricDefinition { + s.Regex = &v + return s +} + // Provides information about the location that is configured for storing model // artifacts. type ModelArtifacts struct { @@ -7302,6 +9444,56 @@ func (s *NotebookInstanceSummary) SetUrl(v string) *NotebookInstanceSummary { return s } +// Specifies the number of training jobs that this hyperparameter tuning job +// launched, categorized by the status of their objective metric. The objective +// metric status shows whether the final objective metric for the training job +// has been evaluated by the tuning job and used in the hyperparameter tuning +// process. +type ObjectiveStatusCounters struct { + _ struct{} `type:"structure"` + + // The number of training jobs whose final objective metric was not evaluated + // and used in the hyperparameter tuning process. This typically occurs when + // the training job failed or did not emit an objective metric. + Failed *int64 `type:"integer"` + + // The number of training jobs that are in progress and pending evaluation of + // their final objective metric. + Pending *int64 `type:"integer"` + + // The number of training jobs whose final objective metric was evaluated by + // the hyperparameter tuning job and used in the hyperparameter tuning process. + Succeeded *int64 `type:"integer"` +} + +// String returns the string representation +func (s ObjectiveStatusCounters) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ObjectiveStatusCounters) GoString() string { + return s.String() +} + +// SetFailed sets the Failed field's value. +func (s *ObjectiveStatusCounters) SetFailed(v int64) *ObjectiveStatusCounters { + s.Failed = &v + return s +} + +// SetPending sets the Pending field's value. +func (s *ObjectiveStatusCounters) SetPending(v int64) *ObjectiveStatusCounters { + s.Pending = &v + return s +} + +// SetSucceeded sets the Succeeded field's value. +func (s *ObjectiveStatusCounters) SetSucceeded(v int64) *ObjectiveStatusCounters { + s.Succeeded = &v + return s +} + // Provides information about how to store model training results (model artifacts). type OutputDataConfig struct { _ struct{} `type:"structure"` @@ -7309,10 +9501,9 @@ type OutputDataConfig struct { // The AWS Key Management Service (AWS KMS) key that Amazon SageMaker uses to // encrypt the model artifacts at rest using Amazon S3 server-side encryption. // - // If the configuration of the output S3 bucket requires server-side encryption - // for objects, and you don't provide the KMS key ID, Amazon SageMaker uses - // the default service key. For more information, see KMS-Managed Encryption - // Keys (https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingKMSEncryption.html) + // If you don't provide the KMS key ID, Amazon SageMaker uses the default KMS + // key for Amazon S3 for your role's account. For more information, see KMS-Managed + // Encryption Keys (https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingKMSEncryption.html) // in Amazon Simple Storage Service developer guide. // // The KMS key policy must grant permission to the IAM role you specify in your @@ -7362,6 +9553,92 @@ func (s *OutputDataConfig) SetS3OutputPath(v string) *OutputDataConfig { return s } +// Specifies ranges of integer, continuous, and categorical hyperparameters +// that a hyperparameter tuning job searches. +type ParameterRanges struct { + _ struct{} `type:"structure"` + + // The array of CategoricalParameterRange objects that specify ranges of categorical + // hyperparameters that a hyperparameter tuning job searches. + CategoricalParameterRanges []*CategoricalParameterRange `type:"list"` + + // The array of ContinuousParameterRange objects that specify ranges of continuous + // hyperparameters that a hyperparameter tuning job searches. + ContinuousParameterRanges []*ContinuousParameterRange `type:"list"` + + // The array of IntegerParameterRange objects that specify ranges of integer + // hyperparameters that a hyperparameter tuning job searches. + IntegerParameterRanges []*IntegerParameterRange `type:"list"` +} + +// String returns the string representation +func (s ParameterRanges) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ParameterRanges) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ParameterRanges) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ParameterRanges"} + if s.CategoricalParameterRanges != nil { + for i, v := range s.CategoricalParameterRanges { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "CategoricalParameterRanges", i), err.(request.ErrInvalidParams)) + } + } + } + if s.ContinuousParameterRanges != nil { + for i, v := range s.ContinuousParameterRanges { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "ContinuousParameterRanges", i), err.(request.ErrInvalidParams)) + } + } + } + if s.IntegerParameterRanges != nil { + for i, v := range s.IntegerParameterRanges { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "IntegerParameterRanges", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCategoricalParameterRanges sets the CategoricalParameterRanges field's value. +func (s *ParameterRanges) SetCategoricalParameterRanges(v []*CategoricalParameterRange) *ParameterRanges { + s.CategoricalParameterRanges = v + return s +} + +// SetContinuousParameterRanges sets the ContinuousParameterRanges field's value. +func (s *ParameterRanges) SetContinuousParameterRanges(v []*ContinuousParameterRange) *ParameterRanges { + s.ContinuousParameterRanges = v + return s +} + +// SetIntegerParameterRanges sets the IntegerParameterRanges field's value. +func (s *ParameterRanges) SetIntegerParameterRanges(v []*IntegerParameterRange) *ParameterRanges { + s.IntegerParameterRanges = v + return s +} + // Identifies a model that you want to host and the resources to deploy for // hosting it. If you are deploying multiple models, tell Amazon SageMaker how // to distribute traffic among the models by specifying variant weights. @@ -7624,6 +9901,68 @@ func (s *ResourceConfig) SetVolumeSizeInGB(v int64) *ResourceConfig { return s } +// Specifies the maximum number of training jobs and parallel training jobs +// that a hyperparameter tuning job can launch. +type ResourceLimits struct { + _ struct{} `type:"structure"` + + // The maximum number of training jobs that a hyperparameter tuning job can + // launch. + // + // MaxNumberOfTrainingJobs is a required field + MaxNumberOfTrainingJobs *int64 `min:"1" type:"integer" required:"true"` + + // The maximum number of concurrent training jobs that a hyperparameter tuning + // job can launch. + // + // MaxParallelTrainingJobs is a required field + MaxParallelTrainingJobs *int64 `min:"1" type:"integer" required:"true"` +} + +// String returns the string representation +func (s ResourceLimits) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s ResourceLimits) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *ResourceLimits) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "ResourceLimits"} + if s.MaxNumberOfTrainingJobs == nil { + invalidParams.Add(request.NewErrParamRequired("MaxNumberOfTrainingJobs")) + } + if s.MaxNumberOfTrainingJobs != nil && *s.MaxNumberOfTrainingJobs < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxNumberOfTrainingJobs", 1)) + } + if s.MaxParallelTrainingJobs == nil { + invalidParams.Add(request.NewErrParamRequired("MaxParallelTrainingJobs")) + } + if s.MaxParallelTrainingJobs != nil && *s.MaxParallelTrainingJobs < 1 { + invalidParams.Add(request.NewErrParamMinValue("MaxParallelTrainingJobs", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetMaxNumberOfTrainingJobs sets the MaxNumberOfTrainingJobs field's value. +func (s *ResourceLimits) SetMaxNumberOfTrainingJobs(v int64) *ResourceLimits { + s.MaxNumberOfTrainingJobs = &v + return s +} + +// SetMaxParallelTrainingJobs sets the MaxParallelTrainingJobs field's value. +func (s *ResourceLimits) SetMaxParallelTrainingJobs(v int64) *ResourceLimits { + s.MaxParallelTrainingJobs = &v + return s +} + // Describes the S3 data source. type S3DataSource struct { _ struct{} `type:"structure"` @@ -7794,6 +10133,61 @@ func (s StartNotebookInstanceOutput) GoString() string { return s.String() } +type StopHyperParameterTuningJobInput struct { + _ struct{} `type:"structure"` + + // The name of the tuning job to stop. + // + // HyperParameterTuningJobName is a required field + HyperParameterTuningJobName *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s StopHyperParameterTuningJobInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopHyperParameterTuningJobInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *StopHyperParameterTuningJobInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "StopHyperParameterTuningJobInput"} + if s.HyperParameterTuningJobName == nil { + invalidParams.Add(request.NewErrParamRequired("HyperParameterTuningJobName")) + } + if s.HyperParameterTuningJobName != nil && len(*s.HyperParameterTuningJobName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("HyperParameterTuningJobName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetHyperParameterTuningJobName sets the HyperParameterTuningJobName field's value. +func (s *StopHyperParameterTuningJobInput) SetHyperParameterTuningJobName(v string) *StopHyperParameterTuningJobInput { + s.HyperParameterTuningJobName = &v + return s +} + +type StopHyperParameterTuningJobOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s StopHyperParameterTuningJobOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s StopHyperParameterTuningJobOutput) GoString() string { + return s.String() +} + type StopNotebookInstanceInput struct { _ struct{} `type:"structure"` @@ -8009,6 +10403,72 @@ func (s *Tag) SetValue(v string) *Tag { return s } +// The numbers of training jobs launched by a hyperparameter tuning job, categorized +// by status. +type TrainingJobStatusCounters struct { + _ struct{} `type:"structure"` + + // The number of completed training jobs launched by a hyperparameter tuning + // job. + Completed *int64 `type:"integer"` + + // The number of in-progress training jobs launched by a hyperparameter tuning + // job. + InProgress *int64 `type:"integer"` + + // The number of training jobs that failed and can't be retried. A failed training + // job can't be retried if it failed because a client error occurred. + NonRetryableError *int64 `type:"integer"` + + // The number of training jobs that failed, but can be retried. A failed training + // job can be retried only if it failed because an internal service error occurred. + RetryableError *int64 `type:"integer"` + + // The number of training jobs launched by a hyperparameter tuning job that + // were manually stopped. + Stopped *int64 `type:"integer"` +} + +// String returns the string representation +func (s TrainingJobStatusCounters) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s TrainingJobStatusCounters) GoString() string { + return s.String() +} + +// SetCompleted sets the Completed field's value. +func (s *TrainingJobStatusCounters) SetCompleted(v int64) *TrainingJobStatusCounters { + s.Completed = &v + return s +} + +// SetInProgress sets the InProgress field's value. +func (s *TrainingJobStatusCounters) SetInProgress(v int64) *TrainingJobStatusCounters { + s.InProgress = &v + return s +} + +// SetNonRetryableError sets the NonRetryableError field's value. +func (s *TrainingJobStatusCounters) SetNonRetryableError(v int64) *TrainingJobStatusCounters { + s.NonRetryableError = &v + return s +} + +// SetRetryableError sets the RetryableError field's value. +func (s *TrainingJobStatusCounters) SetRetryableError(v int64) *TrainingJobStatusCounters { + s.RetryableError = &v + return s +} + +// SetStopped sets the Stopped field's value. +func (s *TrainingJobStatusCounters) SetStopped(v int64) *TrainingJobStatusCounters { + s.Stopped = &v + return s +} + // Provides summary information about a training job. type TrainingJobSummary struct { _ struct{} `type:"structure"` @@ -8266,7 +10726,12 @@ type UpdateNotebookInstanceInput struct { // NotebookInstanceName is a required field NotebookInstanceName *string `type:"string" required:"true"` - // Amazon Resource Name (ARN) of the IAM role to associate with the instance. + // The Amazon Resource Name (ARN) of the IAM role that Amazon SageMaker can + // assume to access the notebook instance. For more information, see Amazon + // SageMaker Roles (http://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-roles.html). + // + // To be able to pass this role to Amazon SageMaker, the caller of this API + // must have the iam:PassRole permission. RoleArn *string `min:"20" type:"string"` } @@ -8540,6 +11005,49 @@ const ( EndpointStatusFailed = "Failed" ) +const ( + // HyperParameterTuningJobObjectiveTypeMaximize is a HyperParameterTuningJobObjectiveType enum value + HyperParameterTuningJobObjectiveTypeMaximize = "Maximize" + + // HyperParameterTuningJobObjectiveTypeMinimize is a HyperParameterTuningJobObjectiveType enum value + HyperParameterTuningJobObjectiveTypeMinimize = "Minimize" +) + +const ( + // HyperParameterTuningJobSortByOptionsName is a HyperParameterTuningJobSortByOptions enum value + HyperParameterTuningJobSortByOptionsName = "Name" + + // HyperParameterTuningJobSortByOptionsStatus is a HyperParameterTuningJobSortByOptions enum value + HyperParameterTuningJobSortByOptionsStatus = "Status" + + // HyperParameterTuningJobSortByOptionsCreationTime is a HyperParameterTuningJobSortByOptions enum value + HyperParameterTuningJobSortByOptionsCreationTime = "CreationTime" +) + +const ( + // HyperParameterTuningJobStatusCompleted is a HyperParameterTuningJobStatus enum value + HyperParameterTuningJobStatusCompleted = "Completed" + + // HyperParameterTuningJobStatusInProgress is a HyperParameterTuningJobStatus enum value + HyperParameterTuningJobStatusInProgress = "InProgress" + + // HyperParameterTuningJobStatusFailed is a HyperParameterTuningJobStatus enum value + HyperParameterTuningJobStatusFailed = "Failed" + + // HyperParameterTuningJobStatusStopped is a HyperParameterTuningJobStatus enum value + HyperParameterTuningJobStatusStopped = "Stopped" + + // HyperParameterTuningJobStatusStopping is a HyperParameterTuningJobStatus enum value + HyperParameterTuningJobStatusStopping = "Stopping" +) + +// The strategy hyperparameter tuning uses to find the best combination of hyperparameters +// for your model. Currently, the only supported value is Bayesian. +const ( + // HyperParameterTuningJobStrategyTypeBayesian is a HyperParameterTuningJobStrategyType enum value + HyperParameterTuningJobStrategyTypeBayesian = "Bayesian" +) + const ( // InstanceTypeMlT2Medium is a InstanceType enum value InstanceTypeMlT2Medium = "ml.t2.medium" @@ -8653,6 +11161,17 @@ const ( NotebookInstanceStatusDeleting = "Deleting" ) +const ( + // ObjectiveStatusSucceeded is a ObjectiveStatus enum value + ObjectiveStatusSucceeded = "Succeeded" + + // ObjectiveStatusPending is a ObjectiveStatus enum value + ObjectiveStatusPending = "Pending" + + // ObjectiveStatusFailed is a ObjectiveStatus enum value + ObjectiveStatusFailed = "Failed" +) + const ( // OrderKeyAscending is a OrderKey enum value OrderKeyAscending = "Ascending" @@ -8919,6 +11438,20 @@ const ( TrainingInstanceTypeMlC518xlarge = "ml.c5.18xlarge" ) +const ( + // TrainingJobSortByOptionsName is a TrainingJobSortByOptions enum value + TrainingJobSortByOptionsName = "Name" + + // TrainingJobSortByOptionsCreationTime is a TrainingJobSortByOptions enum value + TrainingJobSortByOptionsCreationTime = "CreationTime" + + // TrainingJobSortByOptionsStatus is a TrainingJobSortByOptions enum value + TrainingJobSortByOptionsStatus = "Status" + + // TrainingJobSortByOptionsFinalObjectiveMetricValue is a TrainingJobSortByOptions enum value + TrainingJobSortByOptionsFinalObjectiveMetricValue = "FinalObjectiveMetricValue" +) + const ( // TrainingJobStatusInProgress is a TrainingJobStatus enum value TrainingJobStatusInProgress = "InProgress" diff --git a/vendor/github.com/aws/aws-sdk-go/service/sagemaker/sagemakeriface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/sagemaker/sagemakeriface/interface.go index fa3344bc0..7b4af5a2c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sagemaker/sagemakeriface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sagemaker/sagemakeriface/interface.go @@ -72,6 +72,10 @@ type SageMakerAPI interface { CreateEndpointConfigWithContext(aws.Context, *sagemaker.CreateEndpointConfigInput, ...request.Option) (*sagemaker.CreateEndpointConfigOutput, error) CreateEndpointConfigRequest(*sagemaker.CreateEndpointConfigInput) (*request.Request, *sagemaker.CreateEndpointConfigOutput) + CreateHyperParameterTuningJob(*sagemaker.CreateHyperParameterTuningJobInput) (*sagemaker.CreateHyperParameterTuningJobOutput, error) + CreateHyperParameterTuningJobWithContext(aws.Context, *sagemaker.CreateHyperParameterTuningJobInput, ...request.Option) (*sagemaker.CreateHyperParameterTuningJobOutput, error) + CreateHyperParameterTuningJobRequest(*sagemaker.CreateHyperParameterTuningJobInput) (*request.Request, *sagemaker.CreateHyperParameterTuningJobOutput) + CreateModel(*sagemaker.CreateModelInput) (*sagemaker.CreateModelOutput, error) CreateModelWithContext(aws.Context, *sagemaker.CreateModelInput, ...request.Option) (*sagemaker.CreateModelOutput, error) CreateModelRequest(*sagemaker.CreateModelInput) (*request.Request, *sagemaker.CreateModelOutput) @@ -124,6 +128,10 @@ type SageMakerAPI interface { DescribeEndpointConfigWithContext(aws.Context, *sagemaker.DescribeEndpointConfigInput, ...request.Option) (*sagemaker.DescribeEndpointConfigOutput, error) DescribeEndpointConfigRequest(*sagemaker.DescribeEndpointConfigInput) (*request.Request, *sagemaker.DescribeEndpointConfigOutput) + DescribeHyperParameterTuningJob(*sagemaker.DescribeHyperParameterTuningJobInput) (*sagemaker.DescribeHyperParameterTuningJobOutput, error) + DescribeHyperParameterTuningJobWithContext(aws.Context, *sagemaker.DescribeHyperParameterTuningJobInput, ...request.Option) (*sagemaker.DescribeHyperParameterTuningJobOutput, error) + DescribeHyperParameterTuningJobRequest(*sagemaker.DescribeHyperParameterTuningJobInput) (*request.Request, *sagemaker.DescribeHyperParameterTuningJobOutput) + DescribeModel(*sagemaker.DescribeModelInput) (*sagemaker.DescribeModelOutput, error) DescribeModelWithContext(aws.Context, *sagemaker.DescribeModelInput, ...request.Option) (*sagemaker.DescribeModelOutput, error) DescribeModelRequest(*sagemaker.DescribeModelInput) (*request.Request, *sagemaker.DescribeModelOutput) @@ -154,6 +162,13 @@ type SageMakerAPI interface { ListEndpointsPages(*sagemaker.ListEndpointsInput, func(*sagemaker.ListEndpointsOutput, bool) bool) error ListEndpointsPagesWithContext(aws.Context, *sagemaker.ListEndpointsInput, func(*sagemaker.ListEndpointsOutput, bool) bool, ...request.Option) error + ListHyperParameterTuningJobs(*sagemaker.ListHyperParameterTuningJobsInput) (*sagemaker.ListHyperParameterTuningJobsOutput, error) + ListHyperParameterTuningJobsWithContext(aws.Context, *sagemaker.ListHyperParameterTuningJobsInput, ...request.Option) (*sagemaker.ListHyperParameterTuningJobsOutput, error) + ListHyperParameterTuningJobsRequest(*sagemaker.ListHyperParameterTuningJobsInput) (*request.Request, *sagemaker.ListHyperParameterTuningJobsOutput) + + ListHyperParameterTuningJobsPages(*sagemaker.ListHyperParameterTuningJobsInput, func(*sagemaker.ListHyperParameterTuningJobsOutput, bool) bool) error + ListHyperParameterTuningJobsPagesWithContext(aws.Context, *sagemaker.ListHyperParameterTuningJobsInput, func(*sagemaker.ListHyperParameterTuningJobsOutput, bool) bool, ...request.Option) error + ListModels(*sagemaker.ListModelsInput) (*sagemaker.ListModelsOutput, error) ListModelsWithContext(aws.Context, *sagemaker.ListModelsInput, ...request.Option) (*sagemaker.ListModelsOutput, error) ListModelsRequest(*sagemaker.ListModelsInput) (*request.Request, *sagemaker.ListModelsOutput) @@ -189,10 +204,21 @@ type SageMakerAPI interface { ListTrainingJobsPages(*sagemaker.ListTrainingJobsInput, func(*sagemaker.ListTrainingJobsOutput, bool) bool) error ListTrainingJobsPagesWithContext(aws.Context, *sagemaker.ListTrainingJobsInput, func(*sagemaker.ListTrainingJobsOutput, bool) bool, ...request.Option) error + ListTrainingJobsForHyperParameterTuningJob(*sagemaker.ListTrainingJobsForHyperParameterTuningJobInput) (*sagemaker.ListTrainingJobsForHyperParameterTuningJobOutput, error) + ListTrainingJobsForHyperParameterTuningJobWithContext(aws.Context, *sagemaker.ListTrainingJobsForHyperParameterTuningJobInput, ...request.Option) (*sagemaker.ListTrainingJobsForHyperParameterTuningJobOutput, error) + ListTrainingJobsForHyperParameterTuningJobRequest(*sagemaker.ListTrainingJobsForHyperParameterTuningJobInput) (*request.Request, *sagemaker.ListTrainingJobsForHyperParameterTuningJobOutput) + + ListTrainingJobsForHyperParameterTuningJobPages(*sagemaker.ListTrainingJobsForHyperParameterTuningJobInput, func(*sagemaker.ListTrainingJobsForHyperParameterTuningJobOutput, bool) bool) error + ListTrainingJobsForHyperParameterTuningJobPagesWithContext(aws.Context, *sagemaker.ListTrainingJobsForHyperParameterTuningJobInput, func(*sagemaker.ListTrainingJobsForHyperParameterTuningJobOutput, bool) bool, ...request.Option) error + StartNotebookInstance(*sagemaker.StartNotebookInstanceInput) (*sagemaker.StartNotebookInstanceOutput, error) StartNotebookInstanceWithContext(aws.Context, *sagemaker.StartNotebookInstanceInput, ...request.Option) (*sagemaker.StartNotebookInstanceOutput, error) StartNotebookInstanceRequest(*sagemaker.StartNotebookInstanceInput) (*request.Request, *sagemaker.StartNotebookInstanceOutput) + StopHyperParameterTuningJob(*sagemaker.StopHyperParameterTuningJobInput) (*sagemaker.StopHyperParameterTuningJobOutput, error) + StopHyperParameterTuningJobWithContext(aws.Context, *sagemaker.StopHyperParameterTuningJobInput, ...request.Option) (*sagemaker.StopHyperParameterTuningJobOutput, error) + StopHyperParameterTuningJobRequest(*sagemaker.StopHyperParameterTuningJobInput) (*request.Request, *sagemaker.StopHyperParameterTuningJobOutput) + StopNotebookInstance(*sagemaker.StopNotebookInstanceInput) (*sagemaker.StopNotebookInstanceOutput, error) StopNotebookInstanceWithContext(aws.Context, *sagemaker.StopNotebookInstanceInput, ...request.Option) (*sagemaker.StopNotebookInstanceOutput, error) StopNotebookInstanceRequest(*sagemaker.StopNotebookInstanceInput) (*request.Request, *sagemaker.StopNotebookInstanceOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/sagemaker/service.go b/vendor/github.com/aws/aws-sdk-go/service/sagemaker/service.go index c3e00b837..f21c94c79 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sagemaker/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sagemaker/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "sagemaker" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "sagemaker" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SageMaker" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SageMaker client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/sagemakerruntime/service.go b/vendor/github.com/aws/aws-sdk-go/service/sagemakerruntime/service.go index d6fc2166a..8fcc7a88c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sagemakerruntime/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sagemakerruntime/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "runtime.sagemaker" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "runtime.sagemaker" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SageMaker Runtime" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SageMakerRuntime client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/api.go b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/api.go index d4bba27c2..3c89d60c4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/api.go @@ -119,9 +119,16 @@ func (c *SecretsManager) CancelRotateSecretRequest(input *CancelRotateSecretInpu // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/CancelRotateSecret func (c *SecretsManager) CancelRotateSecret(input *CancelRotateSecretInput) (*CancelRotateSecretOutput, error) { @@ -208,23 +215,24 @@ func (c *SecretsManager) CreateSecretRequest(input *CreateSecretInput) (req *req // // If you call an operation that needs to encrypt or decrypt the SecretString // or SecretBinary for a secret in the same account as the calling user and -// that secret doesn't specify a KMS encryption key, Secrets Manager uses the -// account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. -// If this key doesn't already exist in your account then Secrets Manager creates -// it for you automatically. All users in the same AWS account automatically -// have access to use the default CMK. Note that if an Secrets Manager API call -// results in AWS having to create the account's AWS-managed CMK, it can result -// in a one-time significant delay in returning the result. +// that secret doesn't specify a AWS KMS encryption key, Secrets Manager uses +// the account's default AWS managed customer master key (CMK) with the alias +// aws/secretsmanager. If this key doesn't already exist in your account then +// Secrets Manager creates it for you automatically. All users in the same AWS +// account automatically have access to use the default CMK. Note that if an +// Secrets Manager API call results in AWS having to create the account's AWS-managed +// CMK, it can result in a one-time significant delay in returning the result. // // If the secret is in a different AWS account from the credentials calling // an API that requires encryption or decryption of the secret value then you -// must create and use a custom KMS CMK because you can't access the default +// must create and use a custom AWS KMS CMK because you can't access the default // CMK for the account using credentials from a different AWS account. Store // the ARN of the CMK in the secret when you create the secret or when you update // it by including it in the KMSKeyId. If you call an API that must encrypt // or decrypt SecretString or SecretBinary using credentials from a different -// account then the KMS key policy must grant cross-account access to that other -// account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations. +// account then the AWS KMS key policy must grant cross-account access to that +// other account's user or role for both the kms:GenerateDataKey and kms:Decrypt +// operations. // // Minimum permissions // @@ -232,13 +240,13 @@ func (c *SecretsManager) CreateSecretRequest(input *CreateSecretInput) (req *req // // * secretsmanager:CreateSecret // -// * kms:GenerateDataKey - needed only if you use a customer-created KMS -// key to encrypt the secret. You do not need this permission to use the -// account's default AWS managed CMK for Secrets Manager. +// * kms:GenerateDataKey - needed only if you use a customer-managed AWS +// KMS key to encrypt the secret. You do not need this permission to use +// the account's default AWS managed CMK for Secrets Manager. // -// * kms:Decrypt - needed only if you use a customer-created KMS key to encrypt -// the secret. You do not need this permission to use the account's default -// AWS managed CMK for Secrets Manager. +// * kms:Decrypt - needed only if you use a customer-managed AWS KMS key +// to encrypt the secret. You do not need this permission to use the account's +// default AWS managed CMK for Secrets Manager. // // Related operations // @@ -271,9 +279,16 @@ func (c *SecretsManager) CreateSecretRequest(input *CreateSecretInput) (req *req // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeLimitExceededException "LimitExceededException" // The request failed because it would exceed one of the Secrets Manager internal @@ -297,6 +312,9 @@ func (c *SecretsManager) CreateSecretRequest(input *CreateSecretInput) (req *req // * ErrCodeInternalServiceError "InternalServiceError" // An error occurred on the server side. // +// * ErrCodePreconditionNotMetException "PreconditionNotMetException" +// The request failed because you did not complete all the prerequisite steps. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/CreateSecret func (c *SecretsManager) CreateSecret(input *CreateSecretInput) (*CreateSecretOutput, error) { req, out := c.CreateSecretRequest(input) @@ -319,6 +337,116 @@ func (c *SecretsManager) CreateSecretWithContext(ctx aws.Context, input *CreateS return out, req.Send() } +const opDeleteResourcePolicy = "DeleteResourcePolicy" + +// DeleteResourcePolicyRequest generates a "aws/request.Request" representing the +// client's request for the DeleteResourcePolicy operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DeleteResourcePolicy for more information on using the DeleteResourcePolicy +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DeleteResourcePolicyRequest method. +// req, resp := client.DeleteResourcePolicyRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/DeleteResourcePolicy +func (c *SecretsManager) DeleteResourcePolicyRequest(input *DeleteResourcePolicyInput) (req *request.Request, output *DeleteResourcePolicyOutput) { + op := &request.Operation{ + Name: opDeleteResourcePolicy, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DeleteResourcePolicyInput{} + } + + output = &DeleteResourcePolicyOutput{} + req = c.newRequest(op, input, output) + return +} + +// DeleteResourcePolicy API operation for AWS Secrets Manager. +// +// Deletes the resource-based permission policy that's attached to the secret. +// +// Minimum permissions +// +// To run this command, you must have the following permissions: +// +// * secretsmanager:DeleteResourcePolicy +// +// Related operations +// +// * To attach a resource policy to a secret, use PutResourcePolicy. +// +// * To retrieve the current resource-based policy that's attached to a secret, +// use GetResourcePolicy. +// +// * To list all of the currently available secrets, use ListSecrets. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Secrets Manager's +// API operation DeleteResourcePolicy for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// We can't find the resource that you asked for. +// +// * ErrCodeInternalServiceError "InternalServiceError" +// An error occurred on the server side. +// +// * ErrCodeInvalidRequestException "InvalidRequestException" +// You provided a parameter value that is not valid for the current state of +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/DeleteResourcePolicy +func (c *SecretsManager) DeleteResourcePolicy(input *DeleteResourcePolicyInput) (*DeleteResourcePolicyOutput, error) { + req, out := c.DeleteResourcePolicyRequest(input) + return out, req.Send() +} + +// DeleteResourcePolicyWithContext is the same as DeleteResourcePolicy with the addition of +// the ability to pass a context and additional request options. +// +// See DeleteResourcePolicy for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SecretsManager) DeleteResourcePolicyWithContext(ctx aws.Context, input *DeleteResourcePolicyInput, opts ...request.Option) (*DeleteResourcePolicyOutput, error) { + req, out := c.DeleteResourcePolicyRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDeleteSecret = "DeleteSecret" // DeleteSecretRequest generates a "aws/request.Request" representing the @@ -416,9 +544,16 @@ func (c *SecretsManager) DeleteSecretRequest(input *DeleteSecretInput) (req *req // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeInternalServiceError "InternalServiceError" // An error occurred on the server side. @@ -613,9 +748,16 @@ func (c *SecretsManager) GetRandomPasswordRequest(input *GetRandomPasswordInput) // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeInternalServiceError "InternalServiceError" // An error occurred on the server side. @@ -642,6 +784,119 @@ func (c *SecretsManager) GetRandomPasswordWithContext(ctx aws.Context, input *Ge return out, req.Send() } +const opGetResourcePolicy = "GetResourcePolicy" + +// GetResourcePolicyRequest generates a "aws/request.Request" representing the +// client's request for the GetResourcePolicy operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See GetResourcePolicy for more information on using the GetResourcePolicy +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the GetResourcePolicyRequest method. +// req, resp := client.GetResourcePolicyRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/GetResourcePolicy +func (c *SecretsManager) GetResourcePolicyRequest(input *GetResourcePolicyInput) (req *request.Request, output *GetResourcePolicyOutput) { + op := &request.Operation{ + Name: opGetResourcePolicy, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &GetResourcePolicyInput{} + } + + output = &GetResourcePolicyOutput{} + req = c.newRequest(op, input, output) + return +} + +// GetResourcePolicy API operation for AWS Secrets Manager. +// +// Retrieves the JSON text of the resource-based policy document that's attached +// to the specified secret. The JSON request string input and response output +// are shown formatted with white space and line breaks for better readability. +// Submit your input as a single line JSON string. +// +// Minimum permissions +// +// To run this command, you must have the following permissions: +// +// * secretsmanager:GetResourcePolicy +// +// Related operations +// +// * To attach a resource policy to a secret, use PutResourcePolicy. +// +// * To delete the resource-based policy that's attached to a secret, use +// DeleteResourcePolicy. +// +// * To list all of the currently available secrets, use ListSecrets. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Secrets Manager's +// API operation GetResourcePolicy for usage and error information. +// +// Returned Error Codes: +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// We can't find the resource that you asked for. +// +// * ErrCodeInternalServiceError "InternalServiceError" +// An error occurred on the server side. +// +// * ErrCodeInvalidRequestException "InvalidRequestException" +// You provided a parameter value that is not valid for the current state of +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/GetResourcePolicy +func (c *SecretsManager) GetResourcePolicy(input *GetResourcePolicyInput) (*GetResourcePolicyOutput, error) { + req, out := c.GetResourcePolicyRequest(input) + return out, req.Send() +} + +// GetResourcePolicyWithContext is the same as GetResourcePolicy with the addition of +// the ability to pass a context and additional request options. +// +// See GetResourcePolicy for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SecretsManager) GetResourcePolicyWithContext(ctx aws.Context, input *GetResourcePolicyInput, opts ...request.Option) (*GetResourcePolicyOutput, error) { + req, out := c.GetResourcePolicyRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opGetSecretValue = "GetSecretValue" // GetSecretValueRequest generates a "aws/request.Request" representing the @@ -695,8 +950,8 @@ func (c *SecretsManager) GetSecretValueRequest(input *GetSecretValueInput) (req // // * secretsmanager:GetSecretValue // -// * kms:Decrypt - required only if you use a customer-created KMS key to -// encrypt the secret. You do not need this permission to use the account's +// * kms:Decrypt - required only if you use a customer-managed AWS KMS key +// to encrypt the secret. You do not need this permission to use the account's // default AWS managed CMK for Secrets Manager. // // Related operations @@ -722,9 +977,16 @@ func (c *SecretsManager) GetSecretValueRequest(input *GetSecretValueInput) (req // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeDecryptionFailure "DecryptionFailure" // Secrets Manager can't decrypt the protected secret text using the provided @@ -1075,6 +1337,132 @@ func (c *SecretsManager) ListSecretsPagesWithContext(ctx aws.Context, input *Lis return p.Err() } +const opPutResourcePolicy = "PutResourcePolicy" + +// PutResourcePolicyRequest generates a "aws/request.Request" representing the +// client's request for the PutResourcePolicy operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See PutResourcePolicy for more information on using the PutResourcePolicy +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the PutResourcePolicyRequest method. +// req, resp := client.PutResourcePolicyRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/PutResourcePolicy +func (c *SecretsManager) PutResourcePolicyRequest(input *PutResourcePolicyInput) (req *request.Request, output *PutResourcePolicyOutput) { + op := &request.Operation{ + Name: opPutResourcePolicy, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &PutResourcePolicyInput{} + } + + output = &PutResourcePolicyOutput{} + req = c.newRequest(op, input, output) + return +} + +// PutResourcePolicy API operation for AWS Secrets Manager. +// +// Attaches the contents of the specified resource-based permission policy to +// a secret. A resource-based policy is optional. Alternatively, you can use +// IAM identity-based policies that specify the secret's Amazon Resource Name +// (ARN) in the policy statement's Resources element. You can also use a combination +// of both identity-based and resource-based policies. The affected users and +// roles receive the permissions that are permitted by all of the relevant policies. +// For more information, see Using Resource-Based Policies for AWS Secrets Manager +// (http://docs.aws.amazon.com/secretsmanager/latest/userguide/auth-and-access_resource-based-policies.html). +// For the complete description of the AWS policy syntax and grammar, see IAM +// JSON Policy Reference (http://docs.aws.amazon.com/IAM/latest/UserGuide/reference_policies.html) +// in the IAM User Guide. +// +// Minimum permissions +// +// To run this command, you must have the following permissions: +// +// * secretsmanager:PutResourcePolicy +// +// Related operations +// +// * To retrieve the resource policy that's attached to a secret, use GetResourcePolicy. +// +// * To delete the resource-based policy that's attached to a secret, use +// DeleteResourcePolicy. +// +// * To list all of the currently available secrets, use ListSecrets. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Secrets Manager's +// API operation PutResourcePolicy for usage and error information. +// +// Returned Error Codes: +// * ErrCodeMalformedPolicyDocumentException "MalformedPolicyDocumentException" +// The policy document that you provided isn't valid. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// We can't find the resource that you asked for. +// +// * ErrCodeInvalidParameterException "InvalidParameterException" +// You provided an invalid value for a parameter. +// +// * ErrCodeInternalServiceError "InternalServiceError" +// An error occurred on the server side. +// +// * ErrCodeInvalidRequestException "InvalidRequestException" +// You provided a parameter value that is not valid for the current state of +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/PutResourcePolicy +func (c *SecretsManager) PutResourcePolicy(input *PutResourcePolicyInput) (*PutResourcePolicyOutput, error) { + req, out := c.PutResourcePolicyRequest(input) + return out, req.Send() +} + +// PutResourcePolicyWithContext is the same as PutResourcePolicy with the addition of +// the ability to pass a context and additional request options. +// +// See PutResourcePolicy for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *SecretsManager) PutResourcePolicyWithContext(ctx aws.Context, input *PutResourcePolicyInput, opts ...request.Option) (*PutResourcePolicyOutput, error) { + req, out := c.PutResourcePolicyRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opPutSecretValue = "PutSecretValue" // PutSecretValueRequest generates a "aws/request.Request" representing the @@ -1149,23 +1537,24 @@ func (c *SecretsManager) PutSecretValueRequest(input *PutSecretValueInput) (req // // If you call an operation that needs to encrypt or decrypt the SecretString // or SecretBinary for a secret in the same account as the calling user and -// that secret doesn't specify a KMS encryption key, Secrets Manager uses the -// account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. -// If this key doesn't already exist in your account then Secrets Manager creates -// it for you automatically. All users in the same AWS account automatically -// have access to use the default CMK. Note that if an Secrets Manager API call -// results in AWS having to create the account's AWS-managed CMK, it can result -// in a one-time significant delay in returning the result. +// that secret doesn't specify a AWS KMS encryption key, Secrets Manager uses +// the account's default AWS managed customer master key (CMK) with the alias +// aws/secretsmanager. If this key doesn't already exist in your account then +// Secrets Manager creates it for you automatically. All users in the same AWS +// account automatically have access to use the default CMK. Note that if an +// Secrets Manager API call results in AWS having to create the account's AWS-managed +// CMK, it can result in a one-time significant delay in returning the result. // // If the secret is in a different AWS account from the credentials calling // an API that requires encryption or decryption of the secret value then you -// must create and use a custom KMS CMK because you can't access the default +// must create and use a custom AWS KMS CMK because you can't access the default // CMK for the account using credentials from a different AWS account. Store // the ARN of the CMK in the secret when you create the secret or when you update // it by including it in the KMSKeyId. If you call an API that must encrypt // or decrypt SecretString or SecretBinary using credentials from a different -// account then the KMS key policy must grant cross-account access to that other -// account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations. +// account then the AWS KMS key policy must grant cross-account access to that +// other account's user or role for both the kms:GenerateDataKey and kms:Decrypt +// operations. // // Minimum permissions // @@ -1173,13 +1562,9 @@ func (c *SecretsManager) PutSecretValueRequest(input *PutSecretValueInput) (req // // * secretsmanager:PutSecretValue // -// * kms:GenerateDataKey - needed only if you use a customer-created KMS -// key to encrypt the secret. You do not need this permission to use the -// account's AWS managed CMK for Secrets Manager. -// -// * kms:Encrypt - needed only if you use a customer-created KMS key to encrypt -// the secret. You do not need this permission to use the account's AWS managed -// CMK for Secrets Manager. +// * kms:GenerateDataKey - needed only if you use a customer-managed AWS +// KMS key to encrypt the secret. You do not need this permission to use +// the account's default AWS managed CMK for Secrets Manager. // // Related operations // @@ -1205,9 +1590,16 @@ func (c *SecretsManager) PutSecretValueRequest(input *PutSecretValueInput) (req // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeLimitExceededException "LimitExceededException" // The request failed because it would exceed one of the Secrets Manager internal @@ -1323,9 +1715,16 @@ func (c *SecretsManager) RestoreSecretRequest(input *RestoreSecretInput) (req *r // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeInternalServiceError "InternalServiceError" // An error occurred on the server side. @@ -1465,9 +1864,16 @@ func (c *SecretsManager) RotateSecretRequest(input *RotateSecretInput) (req *req // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/RotateSecret func (c *SecretsManager) RotateSecret(input *RotateSecretInput) (*RotateSecretOutput, error) { @@ -1592,6 +1998,19 @@ func (c *SecretsManager) TagResourceRequest(input *TagResourceInput) (req *reque // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // We can't find the resource that you asked for. // +// * ErrCodeInvalidRequestException "InvalidRequestException" +// You provided a parameter value that is not valid for the current state of +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. +// // * ErrCodeInvalidParameterException "InvalidParameterException" // You provided an invalid value for a parameter. // @@ -1700,6 +2119,19 @@ func (c *SecretsManager) UntagResourceRequest(input *UntagResourceInput) (req *r // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // We can't find the resource that you asked for. // +// * ErrCodeInvalidRequestException "InvalidRequestException" +// You provided a parameter value that is not valid for the current state of +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. +// // * ErrCodeInvalidParameterException "InvalidParameterException" // You provided an invalid value for a parameter. // @@ -1793,23 +2225,24 @@ func (c *SecretsManager) UpdateSecretRequest(input *UpdateSecretInput) (req *req // // If you call an operation that needs to encrypt or decrypt the SecretString // or SecretBinary for a secret in the same account as the calling user and -// that secret doesn't specify a KMS encryption key, Secrets Manager uses the -// account's default AWS managed customer master key (CMK) with the alias aws/secretsmanager. -// If this key doesn't already exist in your account then Secrets Manager creates -// it for you automatically. All users in the same AWS account automatically -// have access to use the default CMK. Note that if an Secrets Manager API call -// results in AWS having to create the account's AWS-managed CMK, it can result -// in a one-time significant delay in returning the result. +// that secret doesn't specify a AWS KMS encryption key, Secrets Manager uses +// the account's default AWS managed customer master key (CMK) with the alias +// aws/secretsmanager. If this key doesn't already exist in your account then +// Secrets Manager creates it for you automatically. All users in the same AWS +// account automatically have access to use the default CMK. Note that if an +// Secrets Manager API call results in AWS having to create the account's AWS-managed +// CMK, it can result in a one-time significant delay in returning the result. // // If the secret is in a different AWS account from the credentials calling // an API that requires encryption or decryption of the secret value then you -// must create and use a custom KMS CMK because you can't access the default +// must create and use a custom AWS KMS CMK because you can't access the default // CMK for the account using credentials from a different AWS account. Store // the ARN of the CMK in the secret when you create the secret or when you update // it by including it in the KMSKeyId. If you call an API that must encrypt // or decrypt SecretString or SecretBinary using credentials from a different -// account then the KMS key policy must grant cross-account access to that other -// account's user or role for both the kms:GenerateDataKey and kms:Decrypt operations. +// account then the AWS KMS key policy must grant cross-account access to that +// other account's user or role for both the kms:GenerateDataKey and kms:Decrypt +// operations. // // Minimum permissions // @@ -1817,12 +2250,12 @@ func (c *SecretsManager) UpdateSecretRequest(input *UpdateSecretInput) (req *req // // * secretsmanager:UpdateSecret // -// * kms:GenerateDataKey - needed only if you use a custom KMS key to encrypt -// the secret. You do not need this permission to use the account's AWS managed -// CMK for Secrets Manager. +// * kms:GenerateDataKey - needed only if you use a custom AWS KMS key to +// encrypt the secret. You do not need this permission to use the account's +// AWS managed CMK for Secrets Manager. // -// * kms:Decrypt - needed only if you use a custom KMS key to encrypt the -// secret. You do not need this permission to use the account's AWS managed +// * kms:Decrypt - needed only if you use a custom AWS KMS key to encrypt +// the secret. You do not need this permission to use the account's AWS managed // CMK for Secrets Manager. // // Related operations @@ -1848,9 +2281,16 @@ func (c *SecretsManager) UpdateSecretRequest(input *UpdateSecretInput) (req *req // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeLimitExceededException "LimitExceededException" // The request failed because it would exceed one of the Secrets Manager internal @@ -1874,6 +2314,9 @@ func (c *SecretsManager) UpdateSecretRequest(input *UpdateSecretInput) (req *req // * ErrCodeInternalServiceError "InternalServiceError" // An error occurred on the server side. // +// * ErrCodePreconditionNotMetException "PreconditionNotMetException" +// The request failed because you did not complete all the prerequisite steps. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/secretsmanager-2017-10-17/UpdateSecret func (c *SecretsManager) UpdateSecret(input *UpdateSecretInput) (*UpdateSecretOutput, error) { req, out := c.UpdateSecretRequest(input) @@ -1990,9 +2433,16 @@ func (c *SecretsManager) UpdateSecretVersionStageRequest(input *UpdateSecretVers // // * ErrCodeInvalidRequestException "InvalidRequestException" // You provided a parameter value that is not valid for the current state of -// the resource. For example, if you try to enable rotation on a secret, you -// must already have a Lambda function ARN configured or included as a parameter -// in this call. +// the resource. +// +// Possible causes: +// +// * You tried to perform the operation on a secret that's currently marked +// deleted. +// +// * You tried to enable rotation on a secret that doesn't already have a +// Lambda function ARN configured and you didn't include such an ARN as a +// parameter in this call. // // * ErrCodeLimitExceededException "LimitExceededException" // The request failed because it would exceed one of the Secrets Manager internal @@ -2120,8 +2570,8 @@ type CreateSecretInput struct { // // If you use the AWS CLI or one of the AWS SDK to call this operation, then // you can leave this parameter empty. The CLI or SDK generates a random UUID - // for you and includes as the value for this parameter in the request. If you - // don't use the SDK and instead generate a raw HTTP request to the Secrets + // for you and includes it as the value for this parameter in the request. If + // you don't use the SDK and instead generate a raw HTTP request to the Secrets // Manager service endpoint, then you must generate a ClientRequestToken yourself // for the new version and include that value in the request. // @@ -2148,14 +2598,18 @@ type CreateSecretInput struct { // (Optional) Specifies a user-provided description of the secret. Description *string `type:"string"` - // (Optional) Specifies the ARN or alias of the AWS KMS customer master key - // (CMK) to be used to encrypt the SecretString or SecretBinary values in the - // versions stored in this secret. + // (Optional) Specifies the ARN, Key ID, or alias of the AWS KMS customer master + // key (CMK) to be used to encrypt the SecretString or SecretBinary values in + // the versions stored in this secret. + // + // You can specify any of the supported ways to identify a AWS KMS key ID. If + // you need to reference a CMK in a different account, you can use only the + // key ARN or the alias ARN. // // If you don't specify this value, then Secrets Manager defaults to using the - // AWS account's default CMK (the one named aws/secretsmanager). If a KMS CMK - // with that name doesn't yet exist, then Secrets Manager creates it for you - // automatically the first time it needs to encrypt a version's SecretString + // AWS account's default CMK (the one named aws/secretsmanager). If a AWS KMS + // CMK with that name doesn't yet exist, then Secrets Manager creates it for + // you automatically the first time it needs to encrypt a version's SecretString // or SecretBinary fields. // // You can use the account's default CMK to encrypt and decrypt only if you @@ -2166,6 +2620,9 @@ type CreateSecretInput struct { // Specifies the friendly name of the new secret. // + // The secret name must be ASCII letters, digits, or the following characters + // : /_+=.@- + // // Name is a required field Name *string `min:"1" type:"string" required:"true"` @@ -2383,6 +2840,82 @@ func (s *CreateSecretOutput) SetVersionId(v string) *CreateSecretOutput { return s } +type DeleteResourcePolicyInput struct { + _ struct{} `type:"structure"` + + // Specifies the secret that you want to delete the attached resource-based + // policy for. You can specify either the Amazon Resource Name (ARN) or the + // friendly name of the secret. + // + // SecretId is a required field + SecretId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s DeleteResourcePolicyInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteResourcePolicyInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DeleteResourcePolicyInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DeleteResourcePolicyInput"} + if s.SecretId == nil { + invalidParams.Add(request.NewErrParamRequired("SecretId")) + } + if s.SecretId != nil && len(*s.SecretId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("SecretId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSecretId sets the SecretId field's value. +func (s *DeleteResourcePolicyInput) SetSecretId(v string) *DeleteResourcePolicyInput { + s.SecretId = &v + return s +} + +type DeleteResourcePolicyOutput struct { + _ struct{} `type:"structure"` + + // The ARN of the secret that the resource-based policy was deleted for. + ARN *string `min:"20" type:"string"` + + // The friendly name of the secret that the resource-based policy was deleted + // for. + Name *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s DeleteResourcePolicyOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DeleteResourcePolicyOutput) GoString() string { + return s.String() +} + +// SetARN sets the ARN field's value. +func (s *DeleteResourcePolicyOutput) SetARN(v string) *DeleteResourcePolicyOutput { + s.ARN = &v + return s +} + +// SetName sets the Name field's value. +func (s *DeleteResourcePolicyOutput) SetName(v string) *DeleteResourcePolicyOutput { + s.Name = &v + return s +} + type DeleteSecretInput struct { _ struct{} `type:"structure"` @@ -2544,7 +3077,7 @@ type DescribeSecretOutput struct { // The ARN or alias of the AWS KMS customer master key (CMK) that's used to // encrypt the SecretString or SecretBinary fields in each version of the secret. // If you don't provide a key, then Secrets Manager defaults to encrypting the - // secret fields with the default KMS CMK (the one named awssecretsmanager) + // secret fields with the default AWS KMS CMK (the one named awssecretsmanager) // for this account. KmsKeyId *string `type:"string"` @@ -2813,6 +3346,97 @@ func (s *GetRandomPasswordOutput) SetRandomPassword(v string) *GetRandomPassword return s } +type GetResourcePolicyInput struct { + _ struct{} `type:"structure"` + + // Specifies the secret that you want to retrieve the attached resource-based + // policy for. You can specify either the Amazon Resource Name (ARN) or the + // friendly name of the secret. + // + // SecretId is a required field + SecretId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s GetResourcePolicyInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetResourcePolicyInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *GetResourcePolicyInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "GetResourcePolicyInput"} + if s.SecretId == nil { + invalidParams.Add(request.NewErrParamRequired("SecretId")) + } + if s.SecretId != nil && len(*s.SecretId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("SecretId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetSecretId sets the SecretId field's value. +func (s *GetResourcePolicyInput) SetSecretId(v string) *GetResourcePolicyInput { + s.SecretId = &v + return s +} + +type GetResourcePolicyOutput struct { + _ struct{} `type:"structure"` + + // The ARN of the secret that the resource-based policy was retrieved for. + ARN *string `min:"20" type:"string"` + + // The friendly name of the secret that the resource-based policy was retrieved + // for. + Name *string `min:"1" type:"string"` + + // A JSON-formatted string that describes the permissions that are associated + // with the attached secret. These permissions are combined with any permissions + // that are associated with the user or role that attempts to access this secret. + // The combined permissions specify who can access the secret and what actions + // they can perform. For more information, see Authentication and Access Control + // for AWS Secrets Manager (http://docs.aws.amazon.com/secretsmanager/latest/userguide/auth-and-access.html) + // in the AWS Secrets Manager User Guide. + ResourcePolicy *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s GetResourcePolicyOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s GetResourcePolicyOutput) GoString() string { + return s.String() +} + +// SetARN sets the ARN field's value. +func (s *GetResourcePolicyOutput) SetARN(v string) *GetResourcePolicyOutput { + s.ARN = &v + return s +} + +// SetName sets the Name field's value. +func (s *GetResourcePolicyOutput) SetName(v string) *GetResourcePolicyOutput { + s.Name = &v + return s +} + +// SetResourcePolicy sets the ResourcePolicy field's value. +func (s *GetResourcePolicyOutput) SetResourcePolicy(v string) *GetResourcePolicyOutput { + s.ResourcePolicy = &v + return s +} + type GetSecretValueInput struct { _ struct{} `type:"structure"` @@ -3244,6 +3868,103 @@ func (s *ListSecretsOutput) SetSecretList(v []*SecretListEntry) *ListSecretsOutp return s } +type PutResourcePolicyInput struct { + _ struct{} `type:"structure"` + + // A JSON-formatted string that's constructed according to the grammar and syntax + // for an AWS resource-based policy. The policy in the string identifies who + // can access or manage this secret and its versions. For information on how + // to format a JSON parameter for the various command line tool environments, + // see Using JSON for Parameters (http://docs.aws.amazon.com/cli/latest/userguide/cli-using-param.html#cli-using-param-json) + // in the AWS CLI User Guide. + // + // ResourcePolicy is a required field + ResourcePolicy *string `min:"1" type:"string" required:"true"` + + // Specifies the secret that you want to attach the resource-based policy to. + // You can specify either the ARN or the friendly name of the secret. + // + // SecretId is a required field + SecretId *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s PutResourcePolicyInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PutResourcePolicyInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *PutResourcePolicyInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "PutResourcePolicyInput"} + if s.ResourcePolicy == nil { + invalidParams.Add(request.NewErrParamRequired("ResourcePolicy")) + } + if s.ResourcePolicy != nil && len(*s.ResourcePolicy) < 1 { + invalidParams.Add(request.NewErrParamMinLen("ResourcePolicy", 1)) + } + if s.SecretId == nil { + invalidParams.Add(request.NewErrParamRequired("SecretId")) + } + if s.SecretId != nil && len(*s.SecretId) < 1 { + invalidParams.Add(request.NewErrParamMinLen("SecretId", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetResourcePolicy sets the ResourcePolicy field's value. +func (s *PutResourcePolicyInput) SetResourcePolicy(v string) *PutResourcePolicyInput { + s.ResourcePolicy = &v + return s +} + +// SetSecretId sets the SecretId field's value. +func (s *PutResourcePolicyInput) SetSecretId(v string) *PutResourcePolicyInput { + s.SecretId = &v + return s +} + +type PutResourcePolicyOutput struct { + _ struct{} `type:"structure"` + + // The ARN of the secret that the resource-based policy was retrieved for. + ARN *string `min:"20" type:"string"` + + // The friendly name of the secret that the resource-based policy was retrieved + // for. + Name *string `min:"1" type:"string"` +} + +// String returns the string representation +func (s PutResourcePolicyOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s PutResourcePolicyOutput) GoString() string { + return s.String() +} + +// SetARN sets the ARN field's value. +func (s *PutResourcePolicyOutput) SetARN(v string) *PutResourcePolicyOutput { + s.ARN = &v + return s +} + +// SetName sets the Name field's value. +func (s *PutResourcePolicyOutput) SetName(v string) *PutResourcePolicyOutput { + s.Name = &v + return s +} + type PutSecretValueInput struct { _ struct{} `type:"structure"` @@ -3720,7 +4441,7 @@ type SecretListEntry struct { // The Amazon Resource Name (ARN) of the secret. // // For more information about ARNs in Secrets Manager, see Policy Resources - // (http://docs.aws.amazon.com/http:/docs.aws.amazon.com/secretsmanager/latest/userguide/reference_iam-permissions.html#iam-resources) + // (http://docs.aws.amazon.com/secretsmanager/latest/userguide/reference_iam-permissions.html#iam-resources) // in the AWS Secrets Manager User Guide. ARN *string `min:"20" type:"string"` @@ -4170,14 +4891,14 @@ type UpdateSecretInput struct { // (Optional) Specifies a user-provided description of the secret. Description *string `type:"string"` - // (Optional) Specifies the ARN or alias of the KMS customer master key (CMK) - // to be used to encrypt the protected text in the versions of this secret. + // (Optional) Specifies the ARN or alias of the AWS KMS customer master key + // (CMK) to be used to encrypt the protected text in the versions of this secret. // // If you don't specify this value, then Secrets Manager defaults to using the - // default CMK in the account (the one named aws/secretsmanager). If a KMS CMK - // with that name doesn't exist, then Secrets Manager creates it for you automatically - // the first time it needs to encrypt a version's Plaintext or PlaintextString - // fields. + // default CMK in the account (the one named aws/secretsmanager). If a AWS KMS + // CMK with that name doesn't exist, then Secrets Manager creates it for you + // automatically the first time it needs to encrypt a version's Plaintext or + // PlaintextString fields. // // You can only use the account's default CMK to encrypt and decrypt if you // call this operation using credentials from the same account that owns the diff --git a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/errors.go b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/errors.go index 46b28d334..b3c0c48fd 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/errors.go @@ -42,9 +42,16 @@ const ( // "InvalidRequestException". // // You provided a parameter value that is not valid for the current state of - // the resource. For example, if you try to enable rotation on a secret, you - // must already have a Lambda function ARN configured or included as a parameter - // in this call. + // the resource. + // + // Possible causes: + // + // * You tried to perform the operation on a secret that's currently marked + // deleted. + // + // * You tried to enable rotation on a secret that doesn't already have a + // Lambda function ARN configured and you didn't include such an ARN as a + // parameter in this call. ErrCodeInvalidRequestException = "InvalidRequestException" // ErrCodeLimitExceededException for service response error code @@ -60,6 +67,12 @@ const ( // The policy document that you provided isn't valid. ErrCodeMalformedPolicyDocumentException = "MalformedPolicyDocumentException" + // ErrCodePreconditionNotMetException for service response error code + // "PreconditionNotMetException". + // + // The request failed because you did not complete all the prerequisite steps. + ErrCodePreconditionNotMetException = "PreconditionNotMetException" + // ErrCodeResourceExistsException for service response error code // "ResourceExistsException". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/examples_test.go b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/examples_test.go index fd2f25180..a532e32d8 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/examples_test.go +++ b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/examples_test.go @@ -95,6 +95,42 @@ func ExampleSecretsManager_CreateSecret_shared00() { fmt.Println(secretsmanager.ErrCodeMalformedPolicyDocumentException, aerr.Error()) case secretsmanager.ErrCodeInternalServiceError: fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodePreconditionNotMetException: + fmt.Println(secretsmanager.ErrCodePreconditionNotMetException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} + +// To delete the resource-based policy attached to a secret +// +// The following example shows how to delete the resource-based policy that is attached +// to a secret. +func ExampleSecretsManager_DeleteResourcePolicy_shared00() { + svc := secretsmanager.New(session.New()) + input := &secretsmanager.DeleteResourcePolicyInput{ + SecretId: aws.String("MyTestDatabaseSecret"), + } + + result, err := svc.DeleteResourcePolicy(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case secretsmanager.ErrCodeResourceNotFoundException: + fmt.Println(secretsmanager.ErrCodeResourceNotFoundException, aerr.Error()) + case secretsmanager.ErrCodeInternalServiceError: + fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodeInvalidRequestException: + fmt.Println(secretsmanager.ErrCodeInvalidRequestException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -216,6 +252,40 @@ func ExampleSecretsManager_GetRandomPassword_shared00() { fmt.Println(result) } +// To retrieve the resource-based policy attached to a secret +// +// The following example shows how to retrieve the resource-based policy that is attached +// to a secret. +func ExampleSecretsManager_GetResourcePolicy_shared00() { + svc := secretsmanager.New(session.New()) + input := &secretsmanager.GetResourcePolicyInput{ + SecretId: aws.String("MyTestDatabaseSecret"), + } + + result, err := svc.GetResourcePolicy(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case secretsmanager.ErrCodeResourceNotFoundException: + fmt.Println(secretsmanager.ErrCodeResourceNotFoundException, aerr.Error()) + case secretsmanager.ErrCodeInternalServiceError: + fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodeInvalidRequestException: + fmt.Println(secretsmanager.ErrCodeInvalidRequestException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} + // To retrieve the encrypted secret value of a secret // // The following example shows how to retrieve the secret string value from the version @@ -323,6 +393,44 @@ func ExampleSecretsManager_ListSecrets_shared00() { fmt.Println(result) } +// To add a resource-based policy to a secret +// +// The following example shows how to add a resource-based policy to a secret. +func ExampleSecretsManager_PutResourcePolicy_shared00() { + svc := secretsmanager.New(session.New()) + input := &secretsmanager.PutResourcePolicyInput{ + ResourcePolicy: aws.String("{\n\"Version\":\"2012-10-17\",\n\"Statement\":[{\n\"Effect\":\"Allow\",\n\"Principal\":{\n\"AWS\":\"arn:aws:iam::123456789012:root\"\n},\n\"Action\":\"secretsmanager:GetSecretValue\",\n\"Resource\":\"*\"\n}]\n}"), + SecretId: aws.String("MyTestDatabaseSecret"), + } + + result, err := svc.PutResourcePolicy(input) + if err != nil { + if aerr, ok := err.(awserr.Error); ok { + switch aerr.Code() { + case secretsmanager.ErrCodeMalformedPolicyDocumentException: + fmt.Println(secretsmanager.ErrCodeMalformedPolicyDocumentException, aerr.Error()) + case secretsmanager.ErrCodeResourceNotFoundException: + fmt.Println(secretsmanager.ErrCodeResourceNotFoundException, aerr.Error()) + case secretsmanager.ErrCodeInvalidParameterException: + fmt.Println(secretsmanager.ErrCodeInvalidParameterException, aerr.Error()) + case secretsmanager.ErrCodeInternalServiceError: + fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodeInvalidRequestException: + fmt.Println(secretsmanager.ErrCodeInvalidRequestException, aerr.Error()) + default: + fmt.Println(aerr.Error()) + } + } else { + // Print the error, cast err to awserr.Error to get the Code and + // Message from an error. + fmt.Println(err.Error()) + } + return + } + + fmt.Println(result) +} + // To store a secret value in a new version of a secret // // The following example shows how to create a new version of the secret. Alternatively, @@ -509,6 +617,8 @@ func ExampleSecretsManager_TagResource_shared00() { switch aerr.Code() { case secretsmanager.ErrCodeResourceNotFoundException: fmt.Println(secretsmanager.ErrCodeResourceNotFoundException, aerr.Error()) + case secretsmanager.ErrCodeInvalidRequestException: + fmt.Println(secretsmanager.ErrCodeInvalidRequestException, aerr.Error()) case secretsmanager.ErrCodeInvalidParameterException: fmt.Println(secretsmanager.ErrCodeInvalidParameterException, aerr.Error()) case secretsmanager.ErrCodeInternalServiceError: @@ -548,6 +658,8 @@ func ExampleSecretsManager_UntagResource_shared00() { switch aerr.Code() { case secretsmanager.ErrCodeResourceNotFoundException: fmt.Println(secretsmanager.ErrCodeResourceNotFoundException, aerr.Error()) + case secretsmanager.ErrCodeInvalidRequestException: + fmt.Println(secretsmanager.ErrCodeInvalidRequestException, aerr.Error()) case secretsmanager.ErrCodeInvalidParameterException: fmt.Println(secretsmanager.ErrCodeInvalidParameterException, aerr.Error()) case secretsmanager.ErrCodeInternalServiceError: @@ -597,6 +709,8 @@ func ExampleSecretsManager_UpdateSecret_shared00() { fmt.Println(secretsmanager.ErrCodeMalformedPolicyDocumentException, aerr.Error()) case secretsmanager.ErrCodeInternalServiceError: fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodePreconditionNotMetException: + fmt.Println(secretsmanager.ErrCodePreconditionNotMetException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -642,6 +756,8 @@ func ExampleSecretsManager_UpdateSecret_shared01() { fmt.Println(secretsmanager.ErrCodeMalformedPolicyDocumentException, aerr.Error()) case secretsmanager.ErrCodeInternalServiceError: fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodePreconditionNotMetException: + fmt.Println(secretsmanager.ErrCodePreconditionNotMetException, aerr.Error()) default: fmt.Println(aerr.Error()) } @@ -687,6 +803,8 @@ func ExampleSecretsManager_UpdateSecret_shared02() { fmt.Println(secretsmanager.ErrCodeMalformedPolicyDocumentException, aerr.Error()) case secretsmanager.ErrCodeInternalServiceError: fmt.Println(secretsmanager.ErrCodeInternalServiceError, aerr.Error()) + case secretsmanager.ErrCodePreconditionNotMetException: + fmt.Println(secretsmanager.ErrCodePreconditionNotMetException, aerr.Error()) default: fmt.Println(aerr.Error()) } diff --git a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/secretsmanageriface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/secretsmanageriface/interface.go index 36445d156..cf7b46f96 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/secretsmanageriface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/secretsmanageriface/interface.go @@ -68,6 +68,10 @@ type SecretsManagerAPI interface { CreateSecretWithContext(aws.Context, *secretsmanager.CreateSecretInput, ...request.Option) (*secretsmanager.CreateSecretOutput, error) CreateSecretRequest(*secretsmanager.CreateSecretInput) (*request.Request, *secretsmanager.CreateSecretOutput) + DeleteResourcePolicy(*secretsmanager.DeleteResourcePolicyInput) (*secretsmanager.DeleteResourcePolicyOutput, error) + DeleteResourcePolicyWithContext(aws.Context, *secretsmanager.DeleteResourcePolicyInput, ...request.Option) (*secretsmanager.DeleteResourcePolicyOutput, error) + DeleteResourcePolicyRequest(*secretsmanager.DeleteResourcePolicyInput) (*request.Request, *secretsmanager.DeleteResourcePolicyOutput) + DeleteSecret(*secretsmanager.DeleteSecretInput) (*secretsmanager.DeleteSecretOutput, error) DeleteSecretWithContext(aws.Context, *secretsmanager.DeleteSecretInput, ...request.Option) (*secretsmanager.DeleteSecretOutput, error) DeleteSecretRequest(*secretsmanager.DeleteSecretInput) (*request.Request, *secretsmanager.DeleteSecretOutput) @@ -80,6 +84,10 @@ type SecretsManagerAPI interface { GetRandomPasswordWithContext(aws.Context, *secretsmanager.GetRandomPasswordInput, ...request.Option) (*secretsmanager.GetRandomPasswordOutput, error) GetRandomPasswordRequest(*secretsmanager.GetRandomPasswordInput) (*request.Request, *secretsmanager.GetRandomPasswordOutput) + GetResourcePolicy(*secretsmanager.GetResourcePolicyInput) (*secretsmanager.GetResourcePolicyOutput, error) + GetResourcePolicyWithContext(aws.Context, *secretsmanager.GetResourcePolicyInput, ...request.Option) (*secretsmanager.GetResourcePolicyOutput, error) + GetResourcePolicyRequest(*secretsmanager.GetResourcePolicyInput) (*request.Request, *secretsmanager.GetResourcePolicyOutput) + GetSecretValue(*secretsmanager.GetSecretValueInput) (*secretsmanager.GetSecretValueOutput, error) GetSecretValueWithContext(aws.Context, *secretsmanager.GetSecretValueInput, ...request.Option) (*secretsmanager.GetSecretValueOutput, error) GetSecretValueRequest(*secretsmanager.GetSecretValueInput) (*request.Request, *secretsmanager.GetSecretValueOutput) @@ -98,6 +106,10 @@ type SecretsManagerAPI interface { ListSecretsPages(*secretsmanager.ListSecretsInput, func(*secretsmanager.ListSecretsOutput, bool) bool) error ListSecretsPagesWithContext(aws.Context, *secretsmanager.ListSecretsInput, func(*secretsmanager.ListSecretsOutput, bool) bool, ...request.Option) error + PutResourcePolicy(*secretsmanager.PutResourcePolicyInput) (*secretsmanager.PutResourcePolicyOutput, error) + PutResourcePolicyWithContext(aws.Context, *secretsmanager.PutResourcePolicyInput, ...request.Option) (*secretsmanager.PutResourcePolicyOutput, error) + PutResourcePolicyRequest(*secretsmanager.PutResourcePolicyInput) (*request.Request, *secretsmanager.PutResourcePolicyOutput) + PutSecretValue(*secretsmanager.PutSecretValueInput) (*secretsmanager.PutSecretValueOutput, error) PutSecretValueWithContext(aws.Context, *secretsmanager.PutSecretValueInput, ...request.Option) (*secretsmanager.PutSecretValueOutput, error) PutSecretValueRequest(*secretsmanager.PutSecretValueInput) (*request.Request, *secretsmanager.PutSecretValueOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/service.go b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/service.go index 10e885da2..c4758e96d 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/secretsmanager/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "secretsmanager" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "secretsmanager" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Secrets Manager" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SecretsManager client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/serverlessapplicationrepository/service.go b/vendor/github.com/aws/aws-sdk-go/service/serverlessapplicationrepository/service.go index fd5437fc9..78f8f47ce 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/serverlessapplicationrepository/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/serverlessapplicationrepository/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "serverlessrepo" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "serverlessrepo" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ServerlessApplicationRepository" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ServerlessApplicationRepository client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/servicecatalog/service.go b/vendor/github.com/aws/aws-sdk-go/service/servicecatalog/service.go index 05ced0ca3..f15a5b802 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/servicecatalog/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/servicecatalog/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "servicecatalog" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "servicecatalog" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Service Catalog" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ServiceCatalog client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/servicediscovery/service.go b/vendor/github.com/aws/aws-sdk-go/service/servicediscovery/service.go index 3a0c82b22..b5c177a62 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/servicediscovery/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/servicediscovery/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "servicediscovery" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "servicediscovery" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "ServiceDiscovery" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the ServiceDiscovery client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/ses/api.go b/vendor/github.com/aws/aws-sdk-go/service/ses/api.go index 88c520d0a..9231afe1c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ses/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ses/api.go @@ -246,7 +246,7 @@ func (c *SES) CreateConfigurationSetEventDestinationRequest(input *CreateConfigu // Creates a configuration set event destination. // // When you create or update an event destination, you must provide one, and -// only one, destination. The destination can be Amazon CloudWatch, Amazon Kinesis +// only one, destination. The destination can be CloudWatch, Amazon Kinesis // Firehose, or Amazon Simple Notification Service (Amazon SNS). // // An event destination is the AWS service to which Amazon SES publishes the @@ -357,9 +357,8 @@ func (c *SES) CreateConfigurationSetTrackingOptionsRequest(input *CreateConfigur // // By default, images and links used for tracking open and click events are // hosted on domains operated by Amazon SES. You can configure a subdomain of -// your own to handle these events. For information about using configuration -// sets, see Configuring Custom Domains to Handle Open and Click Tracking (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) -// in the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). +// your own to handle these events. For information about using custom domains, +// see the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -455,7 +454,7 @@ func (c *SES) CreateCustomVerificationEmailTemplateRequest(input *CreateCustomVe // Creates a new custom verification email template. // // For more information about custom verification email templates, see Using -// Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// Custom Verification Email Templates (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -855,8 +854,8 @@ func (c *SES) CreateTemplateRequest(input *CreateTemplateInput) (req *request.Re // Indicates that a resource could not be created because of a naming conflict. // // * ErrCodeInvalidTemplateException "InvalidTemplate" -// Indicates that a template could not be created because it contained invalid -// JSON. +// Indicates that the template that you specified could not be rendered. This +// issue may occur when a template refers to a partial that does not exist. // // * ErrCodeLimitExceededException "LimitExceeded" // Indicates that a resource could not be created because of service limits. @@ -1103,9 +1102,8 @@ func (c *SES) DeleteConfigurationSetTrackingOptionsRequest(input *DeleteConfigur // // By default, images and links used for tracking open and click events are // hosted on domains operated by Amazon SES. You can configure a subdomain of -// your own to handle these events. For information about using configuration -// sets, see Configuring Custom Domains to Handle Open and Click Tracking (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) -// in the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). +// your own to handle these events. For information about using custom domains, +// see the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html). // // Deleting this kind of association will result in emails sent using the specified // configuration set to capture open and click events using the standard, Amazon @@ -1196,7 +1194,7 @@ func (c *SES) DeleteCustomVerificationEmailTemplateRequest(input *DeleteCustomVe // Deletes an existing custom verification email template. // // For more information about custom verification email templates, see Using -// Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// Custom Verification Email Templates (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -2170,7 +2168,8 @@ func (c *SES) GetAccountSendingEnabledRequest(input *GetAccountSendingEnabledInp // GetAccountSendingEnabled API operation for Amazon Simple Email Service. // -// Returns the email sending status of the Amazon SES account. +// Returns the email sending status of the Amazon SES account for the current +// region. // // You can execute this operation no more than once per second. // @@ -2250,7 +2249,7 @@ func (c *SES) GetCustomVerificationEmailTemplateRequest(input *GetCustomVerifica // specify. // // For more information about custom verification email templates, see Using -// Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// Custom Verification Email Templates (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -2844,7 +2843,7 @@ func (c *SES) GetSendStatisticsRequest(input *GetSendStatisticsInput) (req *requ // GetSendStatistics API operation for Amazon Simple Email Service. // -// Provides sending statistics for the Amazon SES account. The result is a list +// Provides sending statistics for the current AWS Region. The result is a list // of data points, representing the last two weeks of sending activity. Each // data point in the list contains statistics for a 15-minute period of time. // @@ -3006,8 +3005,8 @@ func (c *SES) ListConfigurationSetsRequest(input *ListConfigurationSetsInput) (r // ListConfigurationSets API operation for Amazon Simple Email Service. // // Provides a list of the configuration sets associated with your Amazon SES -// account. For information about using configuration sets, see Monitoring Your -// Amazon SES Sending Activity (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/monitor-sending-activity.html) +// account in the current AWS Region. For information about using configuration +// sets, see Monitoring Your Amazon SES Sending Activity (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/monitor-sending-activity.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. This operation @@ -3095,10 +3094,11 @@ func (c *SES) ListCustomVerificationEmailTemplatesRequest(input *ListCustomVerif // ListCustomVerificationEmailTemplates API operation for Amazon Simple Email Service. // -// Lists the existing custom verification email templates for your account. +// Lists the existing custom verification email templates for your account in +// the current AWS Region. // // For more information about custom verification email templates, see Using -// Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// Custom Verification Email Templates (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -3232,7 +3232,8 @@ func (c *SES) ListIdentitiesRequest(input *ListIdentitiesInput) (req *request.Re // ListIdentities API operation for Amazon Simple Email Service. // // Returns a list containing all of the identities (email addresses and domains) -// for your AWS account, regardless of verification status. +// for your AWS account in the current AWS Region, regardless of verification +// status. // // You can execute this operation no more than once per second. // @@ -3443,7 +3444,8 @@ func (c *SES) ListReceiptFiltersRequest(input *ListReceiptFiltersInput) (req *re // ListReceiptFilters API operation for Amazon Simple Email Service. // -// Lists the IP address filters associated with your AWS account. +// Lists the IP address filters associated with your AWS account in the current +// AWS Region. // // For information about managing IP address filters, see the Amazon SES Developer // Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/receiving-email-managing-ip-filters.html). @@ -3522,10 +3524,10 @@ func (c *SES) ListReceiptRuleSetsRequest(input *ListReceiptRuleSetsInput) (req * // ListReceiptRuleSets API operation for Amazon Simple Email Service. // -// Lists the receipt rule sets that exist under your AWS account. If there are -// additional receipt rule sets to be retrieved, you will receive a NextToken -// that you can provide to the next call to ListReceiptRuleSets to retrieve -// the additional entries. +// Lists the receipt rule sets that exist under your AWS account in the current +// AWS Region. If there are additional receipt rule sets to be retrieved, you +// will receive a NextToken that you can provide to the next call to ListReceiptRuleSets +// to retrieve the additional entries. // // For information about managing receipt rule sets, see the Amazon SES Developer // Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/receiving-email-managing-receipt-rule-sets.html). @@ -3604,7 +3606,8 @@ func (c *SES) ListTemplatesRequest(input *ListTemplatesInput) (req *request.Requ // ListTemplates API operation for Amazon Simple Email Service. // -// Lists the email templates present in your Amazon SES account. +// Lists the email templates present in your Amazon SES account in the current +// AWS Region. // // You can execute this operation no more than once per second. // @@ -4156,12 +4159,13 @@ func (c *SES) SendCustomVerificationEmailRequest(input *SendCustomVerificationEm // SendCustomVerificationEmail API operation for Amazon Simple Email Service. // // Adds an email address to the list of identities for your Amazon SES account -// and attempts to verify it. As a result of executing this operation, a customized -// verification email is sent to the specified address. +// in the current AWS Region and attempts to verify it. As a result of executing +// this operation, a customized verification email is sent to the specified +// address. // // To use this operation, you must first create a custom verification email // template. For more information about creating and using custom verification -// email templates, see Using Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// email templates, see Using Custom Verification Email Templates (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -4595,6 +4599,17 @@ func (c *SES) SendTemplatedEmailRequest(input *SendTemplatedEmailInput) (req *re // message will be rejected, even if the message contains other recipients // that are valid. // +// If your call to the SendTemplatedEmail operation includes all of the required +// parameters, Amazon SES accepts it and returns a Message ID. However, if Amazon +// SES can't render the email because the template contains errors, it doesn't +// send the email. Additionally, because it already accepted the message, Amazon +// SES doesn't return a message stating that it was unable to send the email. +// +// For these reasons, we highly recommend that you set up Amazon SES to send +// you notifications when Rendering Failure events occur. For more information, +// see Sending Personalized Email Using the Amazon SES API (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/send-personalized-email-api.html) +// in the Amazon Simple Email Service Developer Guide. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -5389,10 +5404,11 @@ func (c *SES) UpdateAccountSendingEnabledRequest(input *UpdateAccountSendingEnab // UpdateAccountSendingEnabled API operation for Amazon Simple Email Service. // -// Enables or disables email sending across your entire Amazon SES account. -// You can use this operation in conjunction with Amazon CloudWatch alarms to -// temporarily pause email sending across your Amazon SES account when reputation -// metrics (such as your bounce on complaint rate) reach certain thresholds. +// Enables or disables email sending across your entire Amazon SES account in +// the current AWS Region. You can use this operation in conjunction with Amazon +// CloudWatch alarms to temporarily pause email sending across your Amazon SES +// account in a given AWS Region when reputation metrics (such as your bounce +// or complaint rates) reach certain thresholds. // // You can execute this operation no more than once per second. // @@ -5576,10 +5592,10 @@ func (c *SES) UpdateConfigurationSetReputationMetricsEnabledRequest(input *Updat // UpdateConfigurationSetReputationMetricsEnabled API operation for Amazon Simple Email Service. // // Enables or disables the publishing of reputation metrics for emails sent -// using a specific configuration set. Reputation metrics include bounce and -// complaint rates. These metrics are published to Amazon CloudWatch. By using -// Amazon CloudWatch, you can create alarms when bounce or complaint rates exceed -// a certain threshold. +// using a specific configuration set in a given AWS Region. Reputation metrics +// include bounce and complaint rates. These metrics are published to Amazon +// CloudWatch. By using CloudWatch, you can create alarms when bounce or complaint +// rates exceed certain thresholds. // // You can execute this operation no more than once per second. // @@ -5663,10 +5679,10 @@ func (c *SES) UpdateConfigurationSetSendingEnabledRequest(input *UpdateConfigura // UpdateConfigurationSetSendingEnabled API operation for Amazon Simple Email Service. // // Enables or disables email sending for messages sent using a specific configuration -// set. You can use this operation in conjunction with Amazon CloudWatch alarms -// to temporarily pause email sending for a configuration set when the reputation -// metrics for that configuration set (such as your bounce on complaint rate) -// reach certain thresholds. +// set in a given AWS Region. You can use this operation in conjunction with +// Amazon CloudWatch alarms to temporarily pause email sending for a configuration +// set when the reputation metrics for that configuration set (such as your +// bounce on complaint rate) exceed certain thresholds. // // You can execute this operation no more than once per second. // @@ -5752,9 +5768,8 @@ func (c *SES) UpdateConfigurationSetTrackingOptionsRequest(input *UpdateConfigur // // By default, images and links used for tracking open and click events are // hosted on domains operated by Amazon SES. You can configure a subdomain of -// your own to handle these events. For information about using configuration -// sets, see Configuring Custom Domains to Handle Open and Click Tracking (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) -// in the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). +// your own to handle these events. For information about using custom domains, +// see the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -5849,7 +5864,7 @@ func (c *SES) UpdateCustomVerificationEmailTemplateRequest(input *UpdateCustomVe // Updates an existing custom verification email template. // // For more information about custom verification email templates, see Using -// Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// Custom Verification Email Templates (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -6067,8 +6082,8 @@ func (c *SES) UpdateTemplateRequest(input *UpdateTemplateInput) (req *request.Re // SES account. // // * ErrCodeInvalidTemplateException "InvalidTemplate" -// Indicates that a template could not be created because it contained invalid -// JSON. +// Indicates that the template that you specified could not be rendered. This +// issue may occur when a template refers to a partial that does not exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/email-2010-12-01/UpdateTemplate func (c *SES) UpdateTemplate(input *UpdateTemplateInput) (*UpdateTemplateOutput, error) { @@ -6224,9 +6239,9 @@ func (c *SES) VerifyDomainIdentityRequest(input *VerifyDomainIdentityInput) (req // VerifyDomainIdentity API operation for Amazon Simple Email Service. // -// Adds a domain to the list of identities for your Amazon SES account and attempts -// to verify it. For more information about verifying domains, see Verifying -// Email Addresses and Domains (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/verify-addresses-and-domains.html) +// Adds a domain to the list of identities for your Amazon SES account in the +// current AWS Region and attempts to verify it. For more information about +// verifying domains, see Verifying Email Addresses and Domains (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/verify-addresses-and-domains.html) // in the Amazon SES Developer Guide. // // You can execute this operation no more than once per second. @@ -6380,8 +6395,8 @@ func (c *SES) VerifyEmailIdentityRequest(input *VerifyEmailIdentityInput) (req * // VerifyEmailIdentity API operation for Amazon Simple Email Service. // // Adds an email address to the list of identities for your Amazon SES account -// and attempts to verify it. As a result of executing this operation, a verification -// email is sent to the specified address. +// in the current AWS region and attempts to verify it. As a result of executing +// this operation, a verification email is sent to the specified address. // // You can execute this operation no more than once per second. // @@ -7348,8 +7363,8 @@ type CreateConfigurationSetTrackingOptionsInput struct { // emails. // // For more information, see Configuring Custom Domains to Handle Open and Click - // Tracking (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) - // in the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). + // Tracking (ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) + // in the Amazon SES Developer Guide. // // TrackingOptions is a required field TrackingOptions *TrackingOptions `type:"structure" required:"true"` @@ -9094,12 +9109,12 @@ func (s GetAccountSendingEnabledInput) GoString() string { } // Represents a request to return the email sending status for your Amazon SES -// account. +// account in the current AWS Region. type GetAccountSendingEnabledOutput struct { _ struct{} `type:"structure"` // Describes whether email sending is enabled or disabled for your Amazon SES - // account. + // account in the current AWS Region. Enabled *bool `type:"boolean"` } @@ -10237,7 +10252,7 @@ func (s *ListConfigurationSetsOutput) SetNextToken(v string) *ListConfigurationS // for your account. // // For more information about custom verification email templates, see Using -// Custom Verification Email Templates (https://docs.aws.amazon.com/ses/latest/DeveloperGuide/custom-verification-emails.html) +// Custom Verification Email Templates (ses/latest/DeveloperGuide/custom-verification-emails.html) // in the Amazon SES Developer Guide. type ListCustomVerificationEmailTemplatesInput struct { _ struct{} `type:"structure"` @@ -11088,7 +11103,7 @@ type ReceiptAction struct { StopAction *StopAction `type:"structure"` // Calls Amazon WorkMail and, optionally, publishes a notification to Amazon - // SNS. + // Amazon SNS. WorkmailAction *WorkmailAction `type:"structure"` } @@ -11787,10 +11802,10 @@ type S3Action struct { // using Amazon S3 server-side encryption. This means that you must use the // Amazon S3 encryption client to decrypt the email after retrieving it from // Amazon S3, as the service has no access to use your AWS KMS keys for decryption. - // This encryption client is currently available with the AWS Java SDK (http://aws.amazon.com/sdk-for-java/) - // and AWS Ruby SDK (http://aws.amazon.com/sdk-for-ruby/) only. For more information - // about client-side encryption using AWS KMS master keys, see the Amazon S3 - // Developer Guide (AmazonS3/latest/dev/UsingClientSideEncryption.html). + // This encryption client is currently available with the AWS SDK for Java (http://aws.amazon.com/sdk-for-java/) + // and AWS SDK for Ruby (http://aws.amazon.com/sdk-for-ruby/) only. For more + // information about client-side encryption using AWS KMS master keys, see the + // Amazon S3 Developer Guide (http://docs.aws.amazon.com/AmazonS3/latest/dev/UsingClientSideEncryption.html). KmsKeyArn *string `type:"string"` // The key prefix of the Amazon S3 bucket. The key prefix is similar to a directory @@ -13935,8 +13950,8 @@ func (s *TestRenderTemplateOutput) SetRenderedTemplate(v string) *TestRenderTemp // emails. // // For more information, see Configuring Custom Domains to Handle Open and Click -// Tracking (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) -// in the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). +// Tracking (ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) +// in the Amazon SES Developer Guide. type TrackingOptions struct { _ struct{} `type:"structure"` @@ -13967,7 +13982,7 @@ type UpdateAccountSendingEnabledInput struct { _ struct{} `type:"structure"` // Describes whether email sending is enabled or disabled for your Amazon SES - // account. + // account in the current AWS Region. Enabled *bool `type:"boolean"` } @@ -14231,8 +14246,8 @@ type UpdateConfigurationSetTrackingOptionsInput struct { // emails. // // For more information, see Configuring Custom Domains to Handle Open and Click - // Tracking (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) - // in the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). + // Tracking (ses/latest/DeveloperGuide/configure-custom-open-click-domains.html) + // in the Amazon SES Developer Guide. // // TrackingOptions is a required field TrackingOptions *TrackingOptions `type:"structure" required:"true"` diff --git a/vendor/github.com/aws/aws-sdk-go/service/ses/doc.go b/vendor/github.com/aws/aws-sdk-go/service/ses/doc.go index 0050868b2..6ba270a75 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ses/doc.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ses/doc.go @@ -3,9 +3,10 @@ // Package ses provides the client and types for making API // requests to Amazon Simple Email Service. // -// This is the API Reference for Amazon Simple Email Service (https://aws.amazon.com/ses/) -// (Amazon SES). This documentation is intended to be used in conjunction with -// the Amazon SES Developer Guide (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). +// This document contains reference information for the Amazon Simple Email +// Service (https://aws.amazon.com/ses/) (Amazon SES) API, version 2010-12-01. +// This document is best used in conjunction with the Amazon SES Developer Guide +// (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/Welcome.html). // // For a list of Amazon SES endpoints to use in service requests, see Regions // and Amazon SES (http://docs.aws.amazon.com/ses/latest/DeveloperGuide/regions.html) diff --git a/vendor/github.com/aws/aws-sdk-go/service/ses/errors.go b/vendor/github.com/aws/aws-sdk-go/service/ses/errors.go index 2d03c2440..dd94d6351 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ses/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ses/errors.go @@ -158,8 +158,8 @@ const ( // ErrCodeInvalidTemplateException for service response error code // "InvalidTemplate". // - // Indicates that a template could not be created because it contained invalid - // JSON. + // Indicates that the template that you specified could not be rendered. This + // issue may occur when a template refers to a partial that does not exist. ErrCodeInvalidTemplateException = "InvalidTemplate" // ErrCodeInvalidTrackingOptionsException for service response error code diff --git a/vendor/github.com/aws/aws-sdk-go/service/ses/service.go b/vendor/github.com/aws/aws-sdk-go/service/ses/service.go index b71b2322e..0e33b771f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ses/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ses/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "email" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "email" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SES" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SES client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/sfn/service.go b/vendor/github.com/aws/aws-sdk-go/service/sfn/service.go index f9d12a58f..2436268f0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sfn/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sfn/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "states" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "states" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SFN" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SFN client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/shield/api.go b/vendor/github.com/aws/aws-sdk-go/service/shield/api.go index 2d29ee909..f470e74b0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/shield/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/shield/api.go @@ -3,6 +3,7 @@ package shield import ( + "fmt" "time" "github.com/aws/aws-sdk-go/aws" @@ -10,6 +11,249 @@ import ( "github.com/aws/aws-sdk-go/aws/request" ) +const opAssociateDRTLogBucket = "AssociateDRTLogBucket" + +// AssociateDRTLogBucketRequest generates a "aws/request.Request" representing the +// client's request for the AssociateDRTLogBucket operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AssociateDRTLogBucket for more information on using the AssociateDRTLogBucket +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AssociateDRTLogBucketRequest method. +// req, resp := client.AssociateDRTLogBucketRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/AssociateDRTLogBucket +func (c *Shield) AssociateDRTLogBucketRequest(input *AssociateDRTLogBucketInput) (req *request.Request, output *AssociateDRTLogBucketOutput) { + op := &request.Operation{ + Name: opAssociateDRTLogBucket, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AssociateDRTLogBucketInput{} + } + + output = &AssociateDRTLogBucketOutput{} + req = c.newRequest(op, input, output) + return +} + +// AssociateDRTLogBucket API operation for AWS Shield. +// +// Authorizes the DDoS Response team (DRT) to access the specified Amazon S3 +// bucket containing your flow logs. You can associate up to 10 Amazon S3 buckets +// with your subscription. +// +// To use the services of the DRT and make an AssociateDRTLogBucket request, +// you must be subscribed to the Business Support plan (https://aws.amazon.com/premiumsupport/business-support/) +// or the Enterprise Support plan (https://aws.amazon.com/premiumsupport/enterprise-support/). +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation AssociateDRTLogBucket for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeInvalidOperationException "InvalidOperationException" +// Exception that indicates that the operation would not cause any change to +// occur. +// +// * ErrCodeNoAssociatedRoleException "NoAssociatedRoleException" +// The ARN of the role that you specifed does not exist. +// +// * ErrCodeLimitsExceededException "LimitsExceededException" +// Exception that indicates that the operation would exceed a limit. +// +// Type is the type of limit that would be exceeded. +// +// Limit is the threshold that would be exceeded. +// +// * ErrCodeInvalidParameterException "InvalidParameterException" +// Exception that indicates that the parameters passed to the API are invalid. +// +// * ErrCodeAccessDeniedForDependencyException "AccessDeniedForDependencyException" +// In order to grant the necessary access to the DDoS Response Team, the user +// submitting AssociateDRTRole must have the iam:PassRole permission. This error +// indicates the user did not have the appropriate permissions. For more information, +// see Granting a User Permissions to Pass a Role to an AWS Service (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_passrole.html). +// +// * ErrCodeOptimisticLockException "OptimisticLockException" +// Exception that indicates that the protection state has been modified by another +// client. You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/AssociateDRTLogBucket +func (c *Shield) AssociateDRTLogBucket(input *AssociateDRTLogBucketInput) (*AssociateDRTLogBucketOutput, error) { + req, out := c.AssociateDRTLogBucketRequest(input) + return out, req.Send() +} + +// AssociateDRTLogBucketWithContext is the same as AssociateDRTLogBucket with the addition of +// the ability to pass a context and additional request options. +// +// See AssociateDRTLogBucket for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) AssociateDRTLogBucketWithContext(ctx aws.Context, input *AssociateDRTLogBucketInput, opts ...request.Option) (*AssociateDRTLogBucketOutput, error) { + req, out := c.AssociateDRTLogBucketRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opAssociateDRTRole = "AssociateDRTRole" + +// AssociateDRTRoleRequest generates a "aws/request.Request" representing the +// client's request for the AssociateDRTRole operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See AssociateDRTRole for more information on using the AssociateDRTRole +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the AssociateDRTRoleRequest method. +// req, resp := client.AssociateDRTRoleRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/AssociateDRTRole +func (c *Shield) AssociateDRTRoleRequest(input *AssociateDRTRoleInput) (req *request.Request, output *AssociateDRTRoleOutput) { + op := &request.Operation{ + Name: opAssociateDRTRole, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &AssociateDRTRoleInput{} + } + + output = &AssociateDRTRoleOutput{} + req = c.newRequest(op, input, output) + return +} + +// AssociateDRTRole API operation for AWS Shield. +// +// Authorizes the DDoS Response team (DRT), using the specified role, to access +// your AWS account to assist with DDoS attack mitigation during potential attacks. +// This enables the DRT to inspect your AWS WAF configuration and create or +// update AWS WAF rules and web ACLs. +// +// You can associate only one RoleArn with your subscription. If you submit +// an AssociateDRTRole request for an account that already has an associated +// role, the new RoleArn will replace the existing RoleArn. +// +// Prior to making the AssociateDRTRole request, you must attach the AWSShieldDRTAccessPolicy +// (https://console.aws.amazon.com/iam/home?#/policies/arn:aws:iam::aws:policy/service-role/AWSShieldDRTAccessPolicy) +// managed policy to the role you will specify in the request. For more information +// see Attaching and Detaching IAM Policies ( https://docs.aws.amazon.com/IAM/latest/UserGuide/access_policies_manage-attach-detach.html). +// The role must also trust the service principal drt.shield.amazonaws.com. +// For more information, see IAM JSON Policy Elements: Principal (https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_policies_elements_principal.html). +// +// The DRT will have access only to your AWS WAF and Shield resources. By submitting +// this request, you authorize the DRT to inspect your AWS WAF and Shield configuration +// and create and update AWS WAF rules and web ACLs on your behalf. The DRT +// takes these actions only if explicitly authorized by you. +// +// You must have the iam:PassRole permission to make an AssociateDRTRole request. +// For more information, see Granting a User Permissions to Pass a Role to an +// AWS Service (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_passrole.html). +// +// To use the services of the DRT and make an AssociateDRTRole request, you +// must be subscribed to the Business Support plan (https://aws.amazon.com/premiumsupport/business-support/) +// or the Enterprise Support plan (https://aws.amazon.com/premiumsupport/enterprise-support/). +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation AssociateDRTRole for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeInvalidOperationException "InvalidOperationException" +// Exception that indicates that the operation would not cause any change to +// occur. +// +// * ErrCodeInvalidParameterException "InvalidParameterException" +// Exception that indicates that the parameters passed to the API are invalid. +// +// * ErrCodeAccessDeniedForDependencyException "AccessDeniedForDependencyException" +// In order to grant the necessary access to the DDoS Response Team, the user +// submitting AssociateDRTRole must have the iam:PassRole permission. This error +// indicates the user did not have the appropriate permissions. For more information, +// see Granting a User Permissions to Pass a Role to an AWS Service (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_passrole.html). +// +// * ErrCodeOptimisticLockException "OptimisticLockException" +// Exception that indicates that the protection state has been modified by another +// client. You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/AssociateDRTRole +func (c *Shield) AssociateDRTRole(input *AssociateDRTRoleInput) (*AssociateDRTRoleOutput, error) { + req, out := c.AssociateDRTRoleRequest(input) + return out, req.Send() +} + +// AssociateDRTRoleWithContext is the same as AssociateDRTRole with the addition of +// the ability to pass a context and additional request options. +// +// See AssociateDRTRole for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) AssociateDRTRoleWithContext(ctx aws.Context, input *AssociateDRTRoleInput, opts ...request.Option) (*AssociateDRTRoleOutput, error) { + req, out := c.AssociateDRTRoleRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opCreateProtection = "CreateProtection" // CreateProtectionRequest generates a "aws/request.Request" representing the @@ -58,6 +302,12 @@ func (c *Shield) CreateProtectionRequest(input *CreateProtectionInput) (req *req // be an Amazon CloudFront distribution, Elastic Load Balancing load balancer, // Elastic IP Address, or an Amazon Route 53 hosted zone. // +// You can add protection to only a single resource with each CreateProtection +// request. If you want to add protection to multiple resources at once, use +// the AWS WAF console (https://console.aws.amazon.com/waf/). For more information +// see Getting Started with AWS Shield Advanced (https://docs.aws.amazon.com/waf/latest/developerguide/getting-started-ddos.html) +// and Add AWS Shield Advanced Protection to more AWS Resources (https://docs.aws.amazon.com/waf/latest/developerguide/configure-new-protection.html). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -163,6 +413,15 @@ func (c *Shield) CreateSubscriptionRequest(input *CreateSubscriptionInput) (req // // Activates AWS Shield Advanced for an account. // +// As part of this request you can specify EmergencySettings that automaticaly +// grant the DDoS response team (DRT) needed permissions to assist you during +// a suspected DDoS attack. For more information see Authorize the DDoS Response +// Team to Create Rules and Web ACLs on Your Behalf (https://docs.aws.amazon.com/waf/latest/developerguide/authorize-DRT.html). +// +// When you initally create a subscription, your subscription is set to be automatically +// renewed at the end of the existing subscription period. You can change this +// by submitting an UpdateSubscription request. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -314,6 +573,9 @@ const opDeleteSubscription = "DeleteSubscription" // // See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DeleteSubscription func (c *Shield) DeleteSubscriptionRequest(input *DeleteSubscriptionInput) (req *request.Request, output *DeleteSubscriptionOutput) { + if c.Client.Config.Logger != nil { + c.Client.Config.Logger.Log("This operation, DeleteSubscription, has been deprecated") + } op := &request.Operation{ Name: opDeleteSubscription, HTTPMethod: "POST", @@ -348,8 +610,10 @@ func (c *Shield) DeleteSubscriptionRequest(input *DeleteSubscriptionInput) (req // You can retry the request. // // * ErrCodeLockedSubscriptionException "LockedSubscriptionException" -// Exception that indicates that the subscription you are trying to delete has -// not yet completed the 1-year commitment. You cannot delete this subscription. +// You are trying to update a subscription that has not yet completed the 1-year +// commitment. You can change the AutoRenew parameter during the last 30 days +// of your subscription. This exception indicates that you are attempting to +// change AutoRenew prior to that period. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // Exception indicating the specified resource does not exist. @@ -459,6 +723,175 @@ func (c *Shield) DescribeAttackWithContext(ctx aws.Context, input *DescribeAttac return out, req.Send() } +const opDescribeDRTAccess = "DescribeDRTAccess" + +// DescribeDRTAccessRequest generates a "aws/request.Request" representing the +// client's request for the DescribeDRTAccess operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeDRTAccess for more information on using the DescribeDRTAccess +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeDRTAccessRequest method. +// req, resp := client.DescribeDRTAccessRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DescribeDRTAccess +func (c *Shield) DescribeDRTAccessRequest(input *DescribeDRTAccessInput) (req *request.Request, output *DescribeDRTAccessOutput) { + op := &request.Operation{ + Name: opDescribeDRTAccess, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeDRTAccessInput{} + } + + output = &DescribeDRTAccessOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeDRTAccess API operation for AWS Shield. +// +// Returns the current role and list of Amazon S3 log buckets used by the DDoS +// Response team (DRT) to access your AWS account while assisting with attack +// mitigation. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation DescribeDRTAccess for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DescribeDRTAccess +func (c *Shield) DescribeDRTAccess(input *DescribeDRTAccessInput) (*DescribeDRTAccessOutput, error) { + req, out := c.DescribeDRTAccessRequest(input) + return out, req.Send() +} + +// DescribeDRTAccessWithContext is the same as DescribeDRTAccess with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeDRTAccess for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) DescribeDRTAccessWithContext(ctx aws.Context, input *DescribeDRTAccessInput, opts ...request.Option) (*DescribeDRTAccessOutput, error) { + req, out := c.DescribeDRTAccessRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeEmergencyContactSettings = "DescribeEmergencyContactSettings" + +// DescribeEmergencyContactSettingsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeEmergencyContactSettings operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeEmergencyContactSettings for more information on using the DescribeEmergencyContactSettings +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeEmergencyContactSettingsRequest method. +// req, resp := client.DescribeEmergencyContactSettingsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DescribeEmergencyContactSettings +func (c *Shield) DescribeEmergencyContactSettingsRequest(input *DescribeEmergencyContactSettingsInput) (req *request.Request, output *DescribeEmergencyContactSettingsOutput) { + op := &request.Operation{ + Name: opDescribeEmergencyContactSettings, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeEmergencyContactSettingsInput{} + } + + output = &DescribeEmergencyContactSettingsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeEmergencyContactSettings API operation for AWS Shield. +// +// Lists the email addresses that the DRT can use to contact you during a suspected +// attack. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation DescribeEmergencyContactSettings for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DescribeEmergencyContactSettings +func (c *Shield) DescribeEmergencyContactSettings(input *DescribeEmergencyContactSettingsInput) (*DescribeEmergencyContactSettingsOutput, error) { + req, out := c.DescribeEmergencyContactSettingsRequest(input) + return out, req.Send() +} + +// DescribeEmergencyContactSettingsWithContext is the same as DescribeEmergencyContactSettings with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeEmergencyContactSettings for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) DescribeEmergencyContactSettingsWithContext(ctx aws.Context, input *DescribeEmergencyContactSettingsInput, opts ...request.Option) (*DescribeEmergencyContactSettingsOutput, error) { + req, out := c.DescribeEmergencyContactSettingsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDescribeProtection = "DescribeProtection" // DescribeProtectionRequest generates a "aws/request.Request" representing the @@ -625,6 +1058,212 @@ func (c *Shield) DescribeSubscriptionWithContext(ctx aws.Context, input *Describ return out, req.Send() } +const opDisassociateDRTLogBucket = "DisassociateDRTLogBucket" + +// DisassociateDRTLogBucketRequest generates a "aws/request.Request" representing the +// client's request for the DisassociateDRTLogBucket operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DisassociateDRTLogBucket for more information on using the DisassociateDRTLogBucket +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DisassociateDRTLogBucketRequest method. +// req, resp := client.DisassociateDRTLogBucketRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DisassociateDRTLogBucket +func (c *Shield) DisassociateDRTLogBucketRequest(input *DisassociateDRTLogBucketInput) (req *request.Request, output *DisassociateDRTLogBucketOutput) { + op := &request.Operation{ + Name: opDisassociateDRTLogBucket, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DisassociateDRTLogBucketInput{} + } + + output = &DisassociateDRTLogBucketOutput{} + req = c.newRequest(op, input, output) + return +} + +// DisassociateDRTLogBucket API operation for AWS Shield. +// +// Removes the DDoS Response team's (DRT) access to the specified Amazon S3 +// bucket containing your flow logs. +// +// To make a DisassociateDRTLogBucket request, you must be subscribed to the +// Business Support plan (https://aws.amazon.com/premiumsupport/business-support/) +// or the Enterprise Support plan (https://aws.amazon.com/premiumsupport/enterprise-support/). +// However, if you are not subscribed to one of these support plans, but had +// been previously and had granted the DRT access to your account, you can submit +// a DisassociateDRTLogBucket request to remove this access. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation DisassociateDRTLogBucket for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeInvalidOperationException "InvalidOperationException" +// Exception that indicates that the operation would not cause any change to +// occur. +// +// * ErrCodeNoAssociatedRoleException "NoAssociatedRoleException" +// The ARN of the role that you specifed does not exist. +// +// * ErrCodeAccessDeniedForDependencyException "AccessDeniedForDependencyException" +// In order to grant the necessary access to the DDoS Response Team, the user +// submitting AssociateDRTRole must have the iam:PassRole permission. This error +// indicates the user did not have the appropriate permissions. For more information, +// see Granting a User Permissions to Pass a Role to an AWS Service (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_passrole.html). +// +// * ErrCodeOptimisticLockException "OptimisticLockException" +// Exception that indicates that the protection state has been modified by another +// client. You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DisassociateDRTLogBucket +func (c *Shield) DisassociateDRTLogBucket(input *DisassociateDRTLogBucketInput) (*DisassociateDRTLogBucketOutput, error) { + req, out := c.DisassociateDRTLogBucketRequest(input) + return out, req.Send() +} + +// DisassociateDRTLogBucketWithContext is the same as DisassociateDRTLogBucket with the addition of +// the ability to pass a context and additional request options. +// +// See DisassociateDRTLogBucket for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) DisassociateDRTLogBucketWithContext(ctx aws.Context, input *DisassociateDRTLogBucketInput, opts ...request.Option) (*DisassociateDRTLogBucketOutput, error) { + req, out := c.DisassociateDRTLogBucketRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDisassociateDRTRole = "DisassociateDRTRole" + +// DisassociateDRTRoleRequest generates a "aws/request.Request" representing the +// client's request for the DisassociateDRTRole operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DisassociateDRTRole for more information on using the DisassociateDRTRole +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DisassociateDRTRoleRequest method. +// req, resp := client.DisassociateDRTRoleRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DisassociateDRTRole +func (c *Shield) DisassociateDRTRoleRequest(input *DisassociateDRTRoleInput) (req *request.Request, output *DisassociateDRTRoleOutput) { + op := &request.Operation{ + Name: opDisassociateDRTRole, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DisassociateDRTRoleInput{} + } + + output = &DisassociateDRTRoleOutput{} + req = c.newRequest(op, input, output) + return +} + +// DisassociateDRTRole API operation for AWS Shield. +// +// Removes the DDoS Response team's (DRT) access to your AWS account. +// +// To make a DisassociateDRTRole request, you must be subscribed to the Business +// Support plan (https://aws.amazon.com/premiumsupport/business-support/) or +// the Enterprise Support plan (https://aws.amazon.com/premiumsupport/enterprise-support/). +// However, if you are not subscribed to one of these support plans, but had +// been previously and had granted the DRT access to your account, you can submit +// a DisassociateDRTRole request to remove this access. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation DisassociateDRTRole for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeInvalidOperationException "InvalidOperationException" +// Exception that indicates that the operation would not cause any change to +// occur. +// +// * ErrCodeOptimisticLockException "OptimisticLockException" +// Exception that indicates that the protection state has been modified by another +// client. You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/DisassociateDRTRole +func (c *Shield) DisassociateDRTRole(input *DisassociateDRTRoleInput) (*DisassociateDRTRoleOutput, error) { + req, out := c.DisassociateDRTRoleRequest(input) + return out, req.Send() +} + +// DisassociateDRTRoleWithContext is the same as DisassociateDRTRole with the addition of +// the ability to pass a context and additional request options. +// +// See DisassociateDRTRole for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) DisassociateDRTRoleWithContext(ctx aws.Context, input *DisassociateDRTRoleInput, opts ...request.Option) (*DisassociateDRTRoleOutput, error) { + req, out := c.DisassociateDRTRoleRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opGetSubscriptionState = "GetSubscriptionState" // GetSubscriptionStateRequest generates a "aws/request.Request" representing the @@ -854,6 +1493,10 @@ func (c *Shield) ListProtectionsRequest(input *ListProtectionsInput) (req *reque // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // Exception indicating the specified resource does not exist. // +// * ErrCodeInvalidPaginationTokenException "InvalidPaginationTokenException" +// Exception that indicates that the NextToken specified in the request is invalid. +// Submit the request using the NextToken value that was returned in the response. +// // See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/ListProtections func (c *Shield) ListProtections(input *ListProtectionsInput) (*ListProtectionsOutput, error) { req, out := c.ListProtectionsRequest(input) @@ -876,6 +1519,307 @@ func (c *Shield) ListProtectionsWithContext(ctx aws.Context, input *ListProtecti return out, req.Send() } +const opUpdateEmergencyContactSettings = "UpdateEmergencyContactSettings" + +// UpdateEmergencyContactSettingsRequest generates a "aws/request.Request" representing the +// client's request for the UpdateEmergencyContactSettings operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See UpdateEmergencyContactSettings for more information on using the UpdateEmergencyContactSettings +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the UpdateEmergencyContactSettingsRequest method. +// req, resp := client.UpdateEmergencyContactSettingsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/UpdateEmergencyContactSettings +func (c *Shield) UpdateEmergencyContactSettingsRequest(input *UpdateEmergencyContactSettingsInput) (req *request.Request, output *UpdateEmergencyContactSettingsOutput) { + op := &request.Operation{ + Name: opUpdateEmergencyContactSettings, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &UpdateEmergencyContactSettingsInput{} + } + + output = &UpdateEmergencyContactSettingsOutput{} + req = c.newRequest(op, input, output) + return +} + +// UpdateEmergencyContactSettings API operation for AWS Shield. +// +// Updates the details of the list of email addresses that the DRT can use to +// contact you during a suspected attack. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation UpdateEmergencyContactSettings for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeInvalidParameterException "InvalidParameterException" +// Exception that indicates that the parameters passed to the API are invalid. +// +// * ErrCodeOptimisticLockException "OptimisticLockException" +// Exception that indicates that the protection state has been modified by another +// client. You can retry the request. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/UpdateEmergencyContactSettings +func (c *Shield) UpdateEmergencyContactSettings(input *UpdateEmergencyContactSettingsInput) (*UpdateEmergencyContactSettingsOutput, error) { + req, out := c.UpdateEmergencyContactSettingsRequest(input) + return out, req.Send() +} + +// UpdateEmergencyContactSettingsWithContext is the same as UpdateEmergencyContactSettings with the addition of +// the ability to pass a context and additional request options. +// +// See UpdateEmergencyContactSettings for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) UpdateEmergencyContactSettingsWithContext(ctx aws.Context, input *UpdateEmergencyContactSettingsInput, opts ...request.Option) (*UpdateEmergencyContactSettingsOutput, error) { + req, out := c.UpdateEmergencyContactSettingsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opUpdateSubscription = "UpdateSubscription" + +// UpdateSubscriptionRequest generates a "aws/request.Request" representing the +// client's request for the UpdateSubscription operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See UpdateSubscription for more information on using the UpdateSubscription +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the UpdateSubscriptionRequest method. +// req, resp := client.UpdateSubscriptionRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/UpdateSubscription +func (c *Shield) UpdateSubscriptionRequest(input *UpdateSubscriptionInput) (req *request.Request, output *UpdateSubscriptionOutput) { + op := &request.Operation{ + Name: opUpdateSubscription, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &UpdateSubscriptionInput{} + } + + output = &UpdateSubscriptionOutput{} + req = c.newRequest(op, input, output) + return +} + +// UpdateSubscription API operation for AWS Shield. +// +// Updates the details of an existing subscription. Only enter values for parameters +// you want to change. Empty parameters are not updated. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Shield's +// API operation UpdateSubscription for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInternalErrorException "InternalErrorException" +// Exception that indicates that a problem occurred with the service infrastructure. +// You can retry the request. +// +// * ErrCodeLockedSubscriptionException "LockedSubscriptionException" +// You are trying to update a subscription that has not yet completed the 1-year +// commitment. You can change the AutoRenew parameter during the last 30 days +// of your subscription. This exception indicates that you are attempting to +// change AutoRenew prior to that period. +// +// * ErrCodeResourceNotFoundException "ResourceNotFoundException" +// Exception indicating the specified resource does not exist. +// +// * ErrCodeInvalidParameterException "InvalidParameterException" +// Exception that indicates that the parameters passed to the API are invalid. +// +// * ErrCodeOptimisticLockException "OptimisticLockException" +// Exception that indicates that the protection state has been modified by another +// client. You can retry the request. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/shield-2016-06-02/UpdateSubscription +func (c *Shield) UpdateSubscription(input *UpdateSubscriptionInput) (*UpdateSubscriptionOutput, error) { + req, out := c.UpdateSubscriptionRequest(input) + return out, req.Send() +} + +// UpdateSubscriptionWithContext is the same as UpdateSubscription with the addition of +// the ability to pass a context and additional request options. +// +// See UpdateSubscription for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *Shield) UpdateSubscriptionWithContext(ctx aws.Context, input *UpdateSubscriptionInput, opts ...request.Option) (*UpdateSubscriptionOutput, error) { + req, out := c.UpdateSubscriptionRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +type AssociateDRTLogBucketInput struct { + _ struct{} `type:"structure"` + + // The Amazon S3 bucket that contains your flow logs. + // + // LogBucket is a required field + LogBucket *string `min:"3" type:"string" required:"true"` +} + +// String returns the string representation +func (s AssociateDRTLogBucketInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateDRTLogBucketInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AssociateDRTLogBucketInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AssociateDRTLogBucketInput"} + if s.LogBucket == nil { + invalidParams.Add(request.NewErrParamRequired("LogBucket")) + } + if s.LogBucket != nil && len(*s.LogBucket) < 3 { + invalidParams.Add(request.NewErrParamMinLen("LogBucket", 3)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetLogBucket sets the LogBucket field's value. +func (s *AssociateDRTLogBucketInput) SetLogBucket(v string) *AssociateDRTLogBucketInput { + s.LogBucket = &v + return s +} + +type AssociateDRTLogBucketOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s AssociateDRTLogBucketOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateDRTLogBucketOutput) GoString() string { + return s.String() +} + +type AssociateDRTRoleInput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the role the DRT will use to access your + // AWS account. + // + // Prior to making the AssociateDRTRole request, you must attach the AWSShieldDRTAccessPolicy + // (https://console.aws.amazon.com/iam/home?#/policies/arn:aws:iam::aws:policy/service-role/AWSShieldDRTAccessPolicy) + // managed policy to this role. For more information see Attaching and Detaching + // IAM Policies ( https://docs.aws.amazon.com/IAM/latest/UserGuide/access_policies_manage-attach-detach.html). + // + // RoleArn is a required field + RoleArn *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s AssociateDRTRoleInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateDRTRoleInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *AssociateDRTRoleInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "AssociateDRTRoleInput"} + if s.RoleArn == nil { + invalidParams.Add(request.NewErrParamRequired("RoleArn")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetRoleArn sets the RoleArn field's value. +func (s *AssociateDRTRoleInput) SetRoleArn(v string) *AssociateDRTRoleInput { + s.RoleArn = &v + return s +} + +type AssociateDRTRoleOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s AssociateDRTRoleOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s AssociateDRTRoleOutput) GoString() string { + return s.String() +} + // The details of a DDoS attack. type AttackDetail struct { _ struct{} `type:"structure"` @@ -1360,7 +2304,7 @@ func (s DeleteProtectionOutput) GoString() string { } type DeleteSubscriptionInput struct { - _ struct{} `type:"structure"` + _ struct{} `deprecated:"true" type:"structure"` } // String returns the string representation @@ -1374,7 +2318,7 @@ func (s DeleteSubscriptionInput) GoString() string { } type DeleteSubscriptionOutput struct { - _ struct{} `type:"structure"` + _ struct{} `deprecated:"true" type:"structure"` } // String returns the string representation @@ -1451,6 +2395,91 @@ func (s *DescribeAttackOutput) SetAttack(v *AttackDetail) *DescribeAttackOutput return s } +type DescribeDRTAccessInput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DescribeDRTAccessInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDRTAccessInput) GoString() string { + return s.String() +} + +type DescribeDRTAccessOutput struct { + _ struct{} `type:"structure"` + + // The list of Amazon S3 buckets accessed by the DRT. + LogBucketList []*string `type:"list"` + + // The Amazon Resource Name (ARN) of the role the DRT used to access your AWS + // account. + RoleArn *string `type:"string"` +} + +// String returns the string representation +func (s DescribeDRTAccessOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeDRTAccessOutput) GoString() string { + return s.String() +} + +// SetLogBucketList sets the LogBucketList field's value. +func (s *DescribeDRTAccessOutput) SetLogBucketList(v []*string) *DescribeDRTAccessOutput { + s.LogBucketList = v + return s +} + +// SetRoleArn sets the RoleArn field's value. +func (s *DescribeDRTAccessOutput) SetRoleArn(v string) *DescribeDRTAccessOutput { + s.RoleArn = &v + return s +} + +type DescribeEmergencyContactSettingsInput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DescribeEmergencyContactSettingsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEmergencyContactSettingsInput) GoString() string { + return s.String() +} + +type DescribeEmergencyContactSettingsOutput struct { + _ struct{} `type:"structure"` + + // A list of email addresses that the DRT can use to contact you during a suspected + // attack. + EmergencyContactList []*EmergencyContact `type:"list"` +} + +// String returns the string representation +func (s DescribeEmergencyContactSettingsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeEmergencyContactSettingsOutput) GoString() string { + return s.String() +} + +// SetEmergencyContactList sets the EmergencyContactList field's value. +func (s *DescribeEmergencyContactSettingsOutput) SetEmergencyContactList(v []*EmergencyContact) *DescribeEmergencyContactSettingsOutput { + s.EmergencyContactList = v + return s +} + type DescribeProtectionInput struct { _ struct{} `type:"structure"` @@ -1552,6 +2581,129 @@ func (s *DescribeSubscriptionOutput) SetSubscription(v *Subscription) *DescribeS return s } +type DisassociateDRTLogBucketInput struct { + _ struct{} `type:"structure"` + + // The Amazon S3 bucket that contains your flow logs. + // + // LogBucket is a required field + LogBucket *string `min:"3" type:"string" required:"true"` +} + +// String returns the string representation +func (s DisassociateDRTLogBucketInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateDRTLogBucketInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DisassociateDRTLogBucketInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DisassociateDRTLogBucketInput"} + if s.LogBucket == nil { + invalidParams.Add(request.NewErrParamRequired("LogBucket")) + } + if s.LogBucket != nil && len(*s.LogBucket) < 3 { + invalidParams.Add(request.NewErrParamMinLen("LogBucket", 3)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetLogBucket sets the LogBucket field's value. +func (s *DisassociateDRTLogBucketInput) SetLogBucket(v string) *DisassociateDRTLogBucketInput { + s.LogBucket = &v + return s +} + +type DisassociateDRTLogBucketOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DisassociateDRTLogBucketOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateDRTLogBucketOutput) GoString() string { + return s.String() +} + +type DisassociateDRTRoleInput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DisassociateDRTRoleInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateDRTRoleInput) GoString() string { + return s.String() +} + +type DisassociateDRTRoleOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s DisassociateDRTRoleOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DisassociateDRTRoleOutput) GoString() string { + return s.String() +} + +// Contact information that the DRT can use to contact you during a suspected +// attack. +type EmergencyContact struct { + _ struct{} `type:"structure"` + + // An email address that the DRT can use to contact you during a suspected attack. + // + // EmailAddress is a required field + EmailAddress *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s EmergencyContact) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s EmergencyContact) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *EmergencyContact) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "EmergencyContact"} + if s.EmailAddress == nil { + invalidParams.Add(request.NewErrParamRequired("EmailAddress")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetEmailAddress sets the EmailAddress field's value. +func (s *EmergencyContact) SetEmailAddress(v string) *EmergencyContact { + s.EmailAddress = &v + return s +} + type GetSubscriptionStateInput struct { _ struct{} `type:"structure"` } @@ -1591,6 +2743,39 @@ func (s *GetSubscriptionStateOutput) SetSubscriptionState(v string) *GetSubscrip return s } +// Specifies how many protections of a given type you can create. +type Limit struct { + _ struct{} `type:"structure"` + + // The maximum number of protections that can be created for the specified Type. + Max *int64 `type:"long"` + + // The type of protection. + Type *string `type:"string"` +} + +// String returns the string representation +func (s Limit) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s Limit) GoString() string { + return s.String() +} + +// SetMax sets the Max field's value. +func (s *Limit) SetMax(v int64) *Limit { + s.Max = &v + return s +} + +// SetType sets the Type field's value. +func (s *Limit) SetType(v string) *Limit { + s.Type = &v + return s +} + type ListAttacksInput struct { _ struct{} `type:"structure"` @@ -1911,6 +3096,21 @@ func (s *SubResourceSummary) SetType(v string) *SubResourceSummary { type Subscription struct { _ struct{} `type:"structure"` + // If ENABLED, the subscription will be automatically renewed at the end of + // the existing subscription period. + // + // When you initally create a subscription, AutoRenew is set to ENABLED. You + // can change this by submitting an UpdateSubscription request. If the UpdateSubscription + // request does not included a value for AutoRenew, the existing value for AutoRenew + // remains unchanged. + AutoRenew *string `type:"string" enum:"AutoRenew"` + + // The date and time your subscription will end. + EndTime *time.Time `type:"timestamp" timestampFormat:"unix"` + + // Specifies how many protections of a given type you can create. + Limits []*Limit `type:"list"` + // The start time of the subscription, in Unix time in seconds. For more information // see timestamp (http://docs.aws.amazon.com/cli/latest/userguide/cli-using-param.html#parameter-types). StartTime *time.Time `type:"timestamp" timestampFormat:"unix"` @@ -1929,6 +3129,24 @@ func (s Subscription) GoString() string { return s.String() } +// SetAutoRenew sets the AutoRenew field's value. +func (s *Subscription) SetAutoRenew(v string) *Subscription { + s.AutoRenew = &v + return s +} + +// SetEndTime sets the EndTime field's value. +func (s *Subscription) SetEndTime(v time.Time) *Subscription { + s.EndTime = &v + return s +} + +// SetLimits sets the Limits field's value. +func (s *Subscription) SetLimits(v []*Limit) *Subscription { + s.Limits = v + return s +} + // SetStartTime sets the StartTime field's value. func (s *Subscription) SetStartTime(v time.Time) *Subscription { s.StartTime = &v @@ -2080,6 +3298,105 @@ func (s *TimeRange) SetToExclusive(v time.Time) *TimeRange { return s } +type UpdateEmergencyContactSettingsInput struct { + _ struct{} `type:"structure"` + + // A list of email addresses that the DRT can use to contact you during a suspected + // attack. + EmergencyContactList []*EmergencyContact `type:"list"` +} + +// String returns the string representation +func (s UpdateEmergencyContactSettingsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateEmergencyContactSettingsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *UpdateEmergencyContactSettingsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "UpdateEmergencyContactSettingsInput"} + if s.EmergencyContactList != nil { + for i, v := range s.EmergencyContactList { + if v == nil { + continue + } + if err := v.Validate(); err != nil { + invalidParams.AddNested(fmt.Sprintf("%s[%v]", "EmergencyContactList", i), err.(request.ErrInvalidParams)) + } + } + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetEmergencyContactList sets the EmergencyContactList field's value. +func (s *UpdateEmergencyContactSettingsInput) SetEmergencyContactList(v []*EmergencyContact) *UpdateEmergencyContactSettingsInput { + s.EmergencyContactList = v + return s +} + +type UpdateEmergencyContactSettingsOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s UpdateEmergencyContactSettingsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateEmergencyContactSettingsOutput) GoString() string { + return s.String() +} + +type UpdateSubscriptionInput struct { + _ struct{} `type:"structure"` + + // When you initally create a subscription, AutoRenew is set to ENABLED. If + // ENABLED, the subscription will be automatically renewed at the end of the + // existing subscription period. You can change this by submitting an UpdateSubscription + // request. If the UpdateSubscription request does not included a value for + // AutoRenew, the existing value for AutoRenew remains unchanged. + AutoRenew *string `type:"string" enum:"AutoRenew"` +} + +// String returns the string representation +func (s UpdateSubscriptionInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateSubscriptionInput) GoString() string { + return s.String() +} + +// SetAutoRenew sets the AutoRenew field's value. +func (s *UpdateSubscriptionInput) SetAutoRenew(v string) *UpdateSubscriptionInput { + s.AutoRenew = &v + return s +} + +type UpdateSubscriptionOutput struct { + _ struct{} `type:"structure"` +} + +// String returns the string representation +func (s UpdateSubscriptionOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateSubscriptionOutput) GoString() string { + return s.String() +} + const ( // AttackLayerNetwork is a AttackLayer enum value AttackLayerNetwork = "NETWORK" @@ -2108,6 +3425,14 @@ const ( AttackPropertyIdentifierSourceUserAgent = "SOURCE_USER_AGENT" ) +const ( + // AutoRenewEnabled is a AutoRenew enum value + AutoRenewEnabled = "ENABLED" + + // AutoRenewDisabled is a AutoRenew enum value + AutoRenewDisabled = "DISABLED" +) + const ( // SubResourceTypeIp is a SubResourceType enum value SubResourceTypeIp = "IP" diff --git a/vendor/github.com/aws/aws-sdk-go/service/shield/errors.go b/vendor/github.com/aws/aws-sdk-go/service/shield/errors.go index 3410e89cf..aff76eaa1 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/shield/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/shield/errors.go @@ -4,6 +4,15 @@ package shield const ( + // ErrCodeAccessDeniedForDependencyException for service response error code + // "AccessDeniedForDependencyException". + // + // In order to grant the necessary access to the DDoS Response Team, the user + // submitting AssociateDRTRole must have the iam:PassRole permission. This error + // indicates the user did not have the appropriate permissions. For more information, + // see Granting a User Permissions to Pass a Role to an AWS Service (https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_passrole.html). + ErrCodeAccessDeniedForDependencyException = "AccessDeniedForDependencyException" + // ErrCodeInternalErrorException for service response error code // "InternalErrorException". // @@ -18,6 +27,13 @@ const ( // occur. ErrCodeInvalidOperationException = "InvalidOperationException" + // ErrCodeInvalidPaginationTokenException for service response error code + // "InvalidPaginationTokenException". + // + // Exception that indicates that the NextToken specified in the request is invalid. + // Submit the request using the NextToken value that was returned in the response. + ErrCodeInvalidPaginationTokenException = "InvalidPaginationTokenException" + // ErrCodeInvalidParameterException for service response error code // "InvalidParameterException". // @@ -44,10 +60,18 @@ const ( // ErrCodeLockedSubscriptionException for service response error code // "LockedSubscriptionException". // - // Exception that indicates that the subscription you are trying to delete has - // not yet completed the 1-year commitment. You cannot delete this subscription. + // You are trying to update a subscription that has not yet completed the 1-year + // commitment. You can change the AutoRenew parameter during the last 30 days + // of your subscription. This exception indicates that you are attempting to + // change AutoRenew prior to that period. ErrCodeLockedSubscriptionException = "LockedSubscriptionException" + // ErrCodeNoAssociatedRoleException for service response error code + // "NoAssociatedRoleException". + // + // The ARN of the role that you specifed does not exist. + ErrCodeNoAssociatedRoleException = "NoAssociatedRoleException" + // ErrCodeOptimisticLockException for service response error code // "OptimisticLockException". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/shield/service.go b/vendor/github.com/aws/aws-sdk-go/service/shield/service.go index 31680396c..b7a62ef92 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/shield/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/shield/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "shield" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "shield" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Shield" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Shield client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/shield/shieldiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/shield/shieldiface/interface.go index 0850b967e..c325f026f 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/shield/shieldiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/shield/shieldiface/interface.go @@ -26,7 +26,7 @@ import ( // // myFunc uses an SDK service client to make a request to // // AWS Shield. // func myFunc(svc shieldiface.ShieldAPI) bool { -// // Make svc.CreateProtection request +// // Make svc.AssociateDRTLogBucket request // } // // func main() { @@ -42,7 +42,7 @@ import ( // type mockShieldClient struct { // shieldiface.ShieldAPI // } -// func (m *mockShieldClient) CreateProtection(input *shield.CreateProtectionInput) (*shield.CreateProtectionOutput, error) { +// func (m *mockShieldClient) AssociateDRTLogBucket(input *shield.AssociateDRTLogBucketInput) (*shield.AssociateDRTLogBucketOutput, error) { // // mock response/functionality // } // @@ -60,6 +60,14 @@ import ( // and waiters. Its suggested to use the pattern above for testing, or using // tooling to generate mocks to satisfy the interfaces. type ShieldAPI interface { + AssociateDRTLogBucket(*shield.AssociateDRTLogBucketInput) (*shield.AssociateDRTLogBucketOutput, error) + AssociateDRTLogBucketWithContext(aws.Context, *shield.AssociateDRTLogBucketInput, ...request.Option) (*shield.AssociateDRTLogBucketOutput, error) + AssociateDRTLogBucketRequest(*shield.AssociateDRTLogBucketInput) (*request.Request, *shield.AssociateDRTLogBucketOutput) + + AssociateDRTRole(*shield.AssociateDRTRoleInput) (*shield.AssociateDRTRoleOutput, error) + AssociateDRTRoleWithContext(aws.Context, *shield.AssociateDRTRoleInput, ...request.Option) (*shield.AssociateDRTRoleOutput, error) + AssociateDRTRoleRequest(*shield.AssociateDRTRoleInput) (*request.Request, *shield.AssociateDRTRoleOutput) + CreateProtection(*shield.CreateProtectionInput) (*shield.CreateProtectionOutput, error) CreateProtectionWithContext(aws.Context, *shield.CreateProtectionInput, ...request.Option) (*shield.CreateProtectionOutput, error) CreateProtectionRequest(*shield.CreateProtectionInput) (*request.Request, *shield.CreateProtectionOutput) @@ -80,6 +88,14 @@ type ShieldAPI interface { DescribeAttackWithContext(aws.Context, *shield.DescribeAttackInput, ...request.Option) (*shield.DescribeAttackOutput, error) DescribeAttackRequest(*shield.DescribeAttackInput) (*request.Request, *shield.DescribeAttackOutput) + DescribeDRTAccess(*shield.DescribeDRTAccessInput) (*shield.DescribeDRTAccessOutput, error) + DescribeDRTAccessWithContext(aws.Context, *shield.DescribeDRTAccessInput, ...request.Option) (*shield.DescribeDRTAccessOutput, error) + DescribeDRTAccessRequest(*shield.DescribeDRTAccessInput) (*request.Request, *shield.DescribeDRTAccessOutput) + + DescribeEmergencyContactSettings(*shield.DescribeEmergencyContactSettingsInput) (*shield.DescribeEmergencyContactSettingsOutput, error) + DescribeEmergencyContactSettingsWithContext(aws.Context, *shield.DescribeEmergencyContactSettingsInput, ...request.Option) (*shield.DescribeEmergencyContactSettingsOutput, error) + DescribeEmergencyContactSettingsRequest(*shield.DescribeEmergencyContactSettingsInput) (*request.Request, *shield.DescribeEmergencyContactSettingsOutput) + DescribeProtection(*shield.DescribeProtectionInput) (*shield.DescribeProtectionOutput, error) DescribeProtectionWithContext(aws.Context, *shield.DescribeProtectionInput, ...request.Option) (*shield.DescribeProtectionOutput, error) DescribeProtectionRequest(*shield.DescribeProtectionInput) (*request.Request, *shield.DescribeProtectionOutput) @@ -88,6 +104,14 @@ type ShieldAPI interface { DescribeSubscriptionWithContext(aws.Context, *shield.DescribeSubscriptionInput, ...request.Option) (*shield.DescribeSubscriptionOutput, error) DescribeSubscriptionRequest(*shield.DescribeSubscriptionInput) (*request.Request, *shield.DescribeSubscriptionOutput) + DisassociateDRTLogBucket(*shield.DisassociateDRTLogBucketInput) (*shield.DisassociateDRTLogBucketOutput, error) + DisassociateDRTLogBucketWithContext(aws.Context, *shield.DisassociateDRTLogBucketInput, ...request.Option) (*shield.DisassociateDRTLogBucketOutput, error) + DisassociateDRTLogBucketRequest(*shield.DisassociateDRTLogBucketInput) (*request.Request, *shield.DisassociateDRTLogBucketOutput) + + DisassociateDRTRole(*shield.DisassociateDRTRoleInput) (*shield.DisassociateDRTRoleOutput, error) + DisassociateDRTRoleWithContext(aws.Context, *shield.DisassociateDRTRoleInput, ...request.Option) (*shield.DisassociateDRTRoleOutput, error) + DisassociateDRTRoleRequest(*shield.DisassociateDRTRoleInput) (*request.Request, *shield.DisassociateDRTRoleOutput) + GetSubscriptionState(*shield.GetSubscriptionStateInput) (*shield.GetSubscriptionStateOutput, error) GetSubscriptionStateWithContext(aws.Context, *shield.GetSubscriptionStateInput, ...request.Option) (*shield.GetSubscriptionStateOutput, error) GetSubscriptionStateRequest(*shield.GetSubscriptionStateInput) (*request.Request, *shield.GetSubscriptionStateOutput) @@ -99,6 +123,14 @@ type ShieldAPI interface { ListProtections(*shield.ListProtectionsInput) (*shield.ListProtectionsOutput, error) ListProtectionsWithContext(aws.Context, *shield.ListProtectionsInput, ...request.Option) (*shield.ListProtectionsOutput, error) ListProtectionsRequest(*shield.ListProtectionsInput) (*request.Request, *shield.ListProtectionsOutput) + + UpdateEmergencyContactSettings(*shield.UpdateEmergencyContactSettingsInput) (*shield.UpdateEmergencyContactSettingsOutput, error) + UpdateEmergencyContactSettingsWithContext(aws.Context, *shield.UpdateEmergencyContactSettingsInput, ...request.Option) (*shield.UpdateEmergencyContactSettingsOutput, error) + UpdateEmergencyContactSettingsRequest(*shield.UpdateEmergencyContactSettingsInput) (*request.Request, *shield.UpdateEmergencyContactSettingsOutput) + + UpdateSubscription(*shield.UpdateSubscriptionInput) (*shield.UpdateSubscriptionOutput, error) + UpdateSubscriptionWithContext(aws.Context, *shield.UpdateSubscriptionInput, ...request.Option) (*shield.UpdateSubscriptionOutput, error) + UpdateSubscriptionRequest(*shield.UpdateSubscriptionInput) (*request.Request, *shield.UpdateSubscriptionOutput) } var _ ShieldAPI = (*shield.Shield)(nil) diff --git a/vendor/github.com/aws/aws-sdk-go/service/simpledb/service.go b/vendor/github.com/aws/aws-sdk-go/service/simpledb/service.go index fb3a0e3cd..d4de27413 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/simpledb/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/simpledb/service.go @@ -30,8 +30,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "sdb" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "sdb" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SimpleDB" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SimpleDB client with a session. @@ -56,6 +57,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/sms/service.go b/vendor/github.com/aws/aws-sdk-go/service/sms/service.go index f7e192ca9..87f786868 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sms/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sms/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "sms" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "sms" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SMS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SMS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/snowball/service.go b/vendor/github.com/aws/aws-sdk-go/service/snowball/service.go index d037541ba..4c4eb58a5 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/snowball/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/snowball/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "snowball" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "snowball" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Snowball" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Snowball client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/sns/api.go b/vendor/github.com/aws/aws-sdk-go/service/sns/api.go index 41ea5fcbc..b69a5ed8a 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sns/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sns/api.go @@ -1376,6 +1376,8 @@ func (c *SNS) ListEndpointsByPlatformApplicationRequest(input *ListEndpointsByPl // are no more records to return, NextToken will be null. For more information, // see Using Amazon SNS Mobile Push Notifications (http://docs.aws.amazon.com/sns/latest/dg/SNSMobilePush.html). // +// This action is throttled at 30 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -1624,6 +1626,8 @@ func (c *SNS) ListPlatformApplicationsRequest(input *ListPlatformApplicationsInp // no more records to return, NextToken will be null. For more information, // see Using Amazon SNS Mobile Push Notifications (http://docs.aws.amazon.com/sns/latest/dg/SNSMobilePush.html). // +// This action is throttled at 15 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -1768,6 +1772,8 @@ func (c *SNS) ListSubscriptionsRequest(input *ListSubscriptionsInput) (req *requ // is also returned. Use the NextToken parameter in a new ListSubscriptions // call to get further results. // +// This action is throttled at 30 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -1912,6 +1918,8 @@ func (c *SNS) ListSubscriptionsByTopicRequest(input *ListSubscriptionsByTopicInp // a NextToken is also returned. Use the NextToken parameter in a new ListSubscriptionsByTopic // call to get further results. // +// This action is throttled at 30 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -2058,6 +2066,8 @@ func (c *SNS) ListTopicsRequest(input *ListTopicsInput) (req *request.Request, o // of topics, up to 100. If there are more topics, a NextToken is also returned. // Use the NextToken parameter in a new ListTopics call to get further results. // +// This action is throttled at 30 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -2283,10 +2293,15 @@ func (c *SNS) PublishRequest(input *PublishInput) (req *request.Request, output // Publish API operation for Amazon Simple Notification Service. // -// Sends a message to all of a topic's subscribed endpoints. When a messageId -// is returned, the message has been saved and Amazon SNS will attempt to deliver -// it to the topic's subscribers shortly. The format of the outgoing message -// to each subscribed endpoint depends on the notification protocol. +// Sends a message to an Amazon SNS topic or sends a text message (SMS message) +// directly to a phone number. +// +// If you send a message to a topic, Amazon SNS delivers the message to each +// endpoint that is subscribed to the topic. The format of the message depends +// on the notification protocol for each subscribed endpoint. +// +// When a messageId is returned, the message has been saved and Amazon SNS will +// attempt to deliver it shortly. // // To use the Publish action for sending a message to a mobile endpoint, such // as an app on a Kindle device or mobile phone, you must specify the EndpointArn @@ -2764,7 +2779,8 @@ func (c *SNS) SetSubscriptionAttributesRequest(input *SetSubscriptionAttributesI // SetSubscriptionAttributes API operation for Amazon Simple Notification Service. // -// Allows a subscription owner to set an attribute of the topic to a new value. +// Allows a subscription owner to set an attribute of the subscription to a +// new value. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2777,6 +2793,11 @@ func (c *SNS) SetSubscriptionAttributesRequest(input *SetSubscriptionAttributesI // * ErrCodeInvalidParameterException "InvalidParameter" // Indicates that a request parameter does not comply with the associated constraints. // +// * ErrCodeFilterPolicyLimitExceededException "FilterPolicyLimitExceeded" +// Indicates that the number of filter polices in your AWS account exceeds the +// limit. To add more filter polices, submit an SNS Limit Increase case in the +// AWS Support Center. +// // * ErrCodeInternalErrorException "InternalError" // Indicates an internal service error. // @@ -2947,6 +2968,8 @@ func (c *SNS) SubscribeRequest(input *SubscribeInput) (req *request.Request, out // the ConfirmSubscription action with the token from the confirmation message. // Confirmation tokens are valid for three days. // +// This action is throttled at 100 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -2958,6 +2981,11 @@ func (c *SNS) SubscribeRequest(input *SubscribeInput) (req *request.Request, out // * ErrCodeSubscriptionLimitExceededException "SubscriptionLimitExceeded" // Indicates that the customer already owns the maximum allowed number of subscriptions. // +// * ErrCodeFilterPolicyLimitExceededException "FilterPolicyLimitExceeded" +// Indicates that the number of filter polices in your AWS account exceeds the +// limit. To add more filter polices, submit an SNS Limit Increase case in the +// AWS Support Center. +// // * ErrCodeInvalidParameterException "InvalidParameter" // Indicates that a request parameter does not comply with the associated constraints. // @@ -3045,6 +3073,8 @@ func (c *SNS) UnsubscribeRequest(input *UnsubscribeInput) (req *request.Request, // message is delivered to the endpoint, so that the endpoint owner can easily // resubscribe to the topic if the Unsubscribe request was unintended. // +// This action is throttled at 100 transactions per second (TPS). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -4047,21 +4077,31 @@ type GetSubscriptionAttributesOutput struct { // A map of the subscription's attributes. Attributes in this map include the // following: // - // * SubscriptionArn -- the subscription's ARN - // - // * TopicArn -- the topic ARN that the subscription is associated with - // - // * Owner -- the AWS account ID of the subscription's owner - // // * ConfirmationWasAuthenticated -- true if the subscription confirmation - // request was authenticated + // request was authenticated. // - // * DeliveryPolicy -- the JSON serialization of the subscription's delivery - // policy + // * DeliveryPolicy -- The JSON serialization of the subscription's delivery + // policy. // - // * EffectiveDeliveryPolicy -- the JSON serialization of the effective delivery + // * EffectiveDeliveryPolicy -- The JSON serialization of the effective delivery // policy that takes into account the topic delivery policy and account system - // defaults + // defaults. + // + // * FilterPolicy -- The filter policy JSON that is assigned to the subscription. + // + // * Owner -- The AWS account ID of the subscription's owner. + // + // * PendingConfirmation -- true if the subscription hasn't been confirmed. + // To confirm a pending subscription, call the ConfirmSubscription action + // with a confirmation token. + // + // * RawMessageDelivery -- true if raw message delivery is enabled for the + // subscription. Raw messages are free of JSON formatting and can be sent + // to HTTP/S and Amazon SQS endpoints. + // + // * SubscriptionArn -- The subscription's ARN. + // + // * TopicArn -- The topic ARN that the subscription is associated with. Attributes map[string]*string `type:"map"` } @@ -4586,8 +4626,9 @@ type MessageAttributeValue struct { // BinaryValue is automatically base64 encoded/decoded by the SDK. BinaryValue []byte `type:"blob"` - // Amazon SNS supports the following logical data types: String, Number, and - // Binary. For more information, see Message Attribute Data Types (http://docs.aws.amazon.com/sns/latest/dg/SNSMessageAttributes.html#SNSMessageAttributes.DataTypes). + // Amazon SNS supports the following logical data types: String, String.Array, + // Number, and Binary. For more information, see Message Attribute Data Types + // (http://docs.aws.amazon.com/sns/latest/dg/SNSMessageAttributes.html#SNSMessageAttributes.DataTypes). // // DataType is a required field DataType *string `type:"string" required:"true"` @@ -4729,17 +4770,26 @@ func (s *PlatformApplication) SetPlatformApplicationArn(v string) *PlatformAppli type PublishInput struct { _ struct{} `type:"structure"` - // The message you want to send to the topic. - // - // If you want to send the same message to all transport protocols, include - // the text of the message as a String value. + // The message you want to send. // + // If you are publishing to a topic and you want to send the same message to + // all transport protocols, include the text of the message as a String value. // If you want to send different messages for each transport protocol, set the // value of the MessageStructure parameter to json and use a JSON object for // the Message parameter. // - // Constraints: Messages must be UTF-8 encoded strings at most 256 KB in size - // (262144 bytes, not 262144 characters). + // Constraints: + // + // With the exception of SMS, messages must be UTF-8 encoded strings and at + // most 256 KB in size (262144 bytes, not 262144 characters). + // + // * For SMS, each message can contain up to 140 bytes, and the character + // limit depends on the encoding scheme. For example, an SMS message can + // contain 160 GSM characters, 140 ASCII characters, or 70 UCS-2 characters. + // If you publish a message that exceeds the size limit, Amazon SNS sends + // it as multiple messages, each fitting within the size limit. Messages + // are not cut off in the middle of a word but on whole-word boundaries. + // The total size limit for a single SMS publish action is 1600 bytes. // // JSON-specific constraints: // @@ -5186,8 +5236,10 @@ type SetSMSAttributesInput struct { // costs that exceed your limit. // // By default, the spend limit is set to the maximum allowed by Amazon SNS. - // If you want to exceed the maximum, contact AWS Support (https://aws.amazon.com/premiumsupport/) - // or your AWS sales representative for a service limit increase. + // If you want to raise the limit, submit an SNS Limit Increase case (https://console.aws.amazon.com/support/home#/case/create?issueType=service-limit-increase&limitType=service-code-sns). + // For New limit value, enter your desired monthly spend limit. In the Use Case + // Description field, explain that you are requesting an SMS monthly spend limit + // increase. // // DeliveryStatusIAMRole – The ARN of the IAM role that allows Amazon SNS to // write logs about SMS deliveries in CloudWatch Logs. For each SMS message @@ -5301,7 +5353,7 @@ type SetSubscriptionAttributesInput struct { // The name of the attribute you want to set. Only a subset of the subscriptions // attributes are mutable. // - // Valid values: DeliveryPolicy | RawMessageDelivery + // Valid values: DeliveryPolicy | FilterPolicy | RawMessageDelivery // // AttributeName is a required field AttributeName *string `type:"string" required:"true"` @@ -5456,6 +5508,10 @@ func (s SetTopicAttributesOutput) GoString() string { type SubscribeInput struct { _ struct{} `type:"structure"` + // Assigns attributes to the subscription as a map of key-value pairs. You can + // assign any attribute that is supported by the SetSubscriptionAttributes action. + Attributes map[string]*string `type:"map"` + // The endpoint that you want to receive notifications. Endpoints vary by protocol: // // * For the http protocol, the endpoint is an URL beginning with "http://" @@ -5499,6 +5555,21 @@ type SubscribeInput struct { // Protocol is a required field Protocol *string `type:"string" required:"true"` + // Sets whether the response from the Subscribe request includes the subscription + // ARN, even if the subscription is not yet confirmed. + // + // If you set this parameter to false, the response includes the ARN for confirmed + // subscriptions, but it includes an ARN value of "pending subscription" for + // subscriptions that are not yet confirmed. A subscription becomes confirmed + // when the subscriber calls the ConfirmSubscription action with a confirmation + // token. + // + // If you set this parameter to true, the response includes the ARN in all cases, + // even if the subscription is not yet confirmed. + // + // The default value is false. + ReturnSubscriptionArn *bool `type:"boolean"` + // The ARN of the topic you want to subscribe to. // // TopicArn is a required field @@ -5531,6 +5602,12 @@ func (s *SubscribeInput) Validate() error { return nil } +// SetAttributes sets the Attributes field's value. +func (s *SubscribeInput) SetAttributes(v map[string]*string) *SubscribeInput { + s.Attributes = v + return s +} + // SetEndpoint sets the Endpoint field's value. func (s *SubscribeInput) SetEndpoint(v string) *SubscribeInput { s.Endpoint = &v @@ -5543,6 +5620,12 @@ func (s *SubscribeInput) SetProtocol(v string) *SubscribeInput { return s } +// SetReturnSubscriptionArn sets the ReturnSubscriptionArn field's value. +func (s *SubscribeInput) SetReturnSubscriptionArn(v bool) *SubscribeInput { + s.ReturnSubscriptionArn = &v + return s +} + // SetTopicArn sets the TopicArn field's value. func (s *SubscribeInput) SetTopicArn(v string) *SubscribeInput { s.TopicArn = &v @@ -5553,8 +5636,10 @@ func (s *SubscribeInput) SetTopicArn(v string) *SubscribeInput { type SubscribeOutput struct { _ struct{} `type:"structure"` - // The ARN of the subscription, if the service was able to create a subscription - // immediately (without requiring endpoint owner confirmation). + // The ARN of the subscription if it is confirmed, or the string "pending confirmation" + // if the subscription requires confirmation. However, if the API request parameter + // ReturnSubscriptionArn is true, then the value is always the subscription + // ARN, even if the subscription requires confirmation. SubscriptionArn *string `type:"string"` } diff --git a/vendor/github.com/aws/aws-sdk-go/service/sns/errors.go b/vendor/github.com/aws/aws-sdk-go/service/sns/errors.go index 60e503f25..9d2993bcc 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sns/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sns/errors.go @@ -16,6 +16,14 @@ const ( // Exception error indicating endpoint disabled. ErrCodeEndpointDisabledException = "EndpointDisabled" + // ErrCodeFilterPolicyLimitExceededException for service response error code + // "FilterPolicyLimitExceeded". + // + // Indicates that the number of filter polices in your AWS account exceeds the + // limit. To add more filter polices, submit an SNS Limit Increase case in the + // AWS Support Center. + ErrCodeFilterPolicyLimitExceededException = "FilterPolicyLimitExceeded" + // ErrCodeInternalErrorException for service response error code // "InternalError". // diff --git a/vendor/github.com/aws/aws-sdk-go/service/sns/service.go b/vendor/github.com/aws/aws-sdk-go/service/sns/service.go index fb75ea60a..96d7c8ba0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sns/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sns/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "sns" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "sns" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SNS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SNS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/sqs/service.go b/vendor/github.com/aws/aws-sdk-go/service/sqs/service.go index 50b5f4b80..d463ecf0d 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sqs/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sqs/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "sqs" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "sqs" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SQS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SQS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/ssm/api.go b/vendor/github.com/aws/aws-sdk-go/service/ssm/api.go index f47913f27..b47687138 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ssm/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ssm/api.go @@ -190,12 +190,12 @@ func (c *SSM) CancelCommandRequest(input *CancelCommandInput) (req *request.Requ // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -390,12 +390,12 @@ func (c *SSM) CreateAssociationRequest(input *CreateAssociationInput) (req *requ // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -517,12 +517,12 @@ func (c *SSM) CreateAssociationBatchRequest(input *CreateAssociationBatchInput) // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -907,8 +907,8 @@ func (c *SSM) CreateResourceDataSyncRequest(input *CreateResourceDataSyncInput) // you enable encryption in Amazon S3 to ensure secure data storage. We also // recommend that you secure access to the Amazon S3 bucket by creating a restrictive // bucket policy. To view an example of a restrictive Amazon S3 bucket policy -// for Resource Data Sync, see Configuring Resource Data Sync for Inventory -// (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-inventory-configuring.html#sysman-inventory-datasync). +// for Resource Data Sync, see Create a Resource Data Sync for Inventory (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-inventory-datasync-create.html) +// in the AWS Systems Manager User Guide. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1119,12 +1119,12 @@ func (c *SSM) DeleteAssociationRequest(input *DeleteAssociationInput) (req *requ // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -1799,7 +1799,7 @@ func (c *SSM) DeregisterManagedInstanceRequest(input *DeregisterManagedInstanceI // // Removes the server or virtual machine from the list of registered servers. // You can reregister the instance again at any time. If you don't plan to use -// Run Command on the server, we suggest uninstalling the SSM Agent first. +// Run Command on the server, we suggest uninstalling SSM Agent first. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -1814,12 +1814,12 @@ func (c *SSM) DeregisterManagedInstanceRequest(input *DeregisterManagedInstanceI // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -2329,12 +2329,12 @@ func (c *SSM) DescribeAssociationRequest(input *DescribeAssociationInput) (req * // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -2856,12 +2856,12 @@ func (c *SSM) DescribeEffectiveInstanceAssociationsRequest(input *DescribeEffect // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -3050,12 +3050,12 @@ func (c *SSM) DescribeInstanceAssociationsStatusRequest(input *DescribeInstanceA // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -3142,6 +3142,10 @@ func (c *SSM) DescribeInstanceInformationRequest(input *DescribeInstanceInformat // information for all your instances. If you specify an instance ID that is // not valid or an instance that you do not own, you receive an error. // +// The IamRole field for this API action is the Amazon Identity and Access Management +// (IAM) role assigned to on-premises instances. This call does not return the +// IAM role for Amazon EC2 instances. +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -3158,12 +3162,12 @@ func (c *SSM) DescribeInstanceInformationRequest(input *DescribeInstanceInformat // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -3481,12 +3485,12 @@ func (c *SSM) DescribeInstancePatchesRequest(input *DescribeInstancePatchesInput // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -4655,12 +4659,12 @@ func (c *SSM) GetCommandInvocationRequest(input *GetCommandInvocationInput) (req // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -5603,7 +5607,8 @@ func (c *SSM) GetParameterRequest(input *GetParameterInput) (req *request.Reques // GetParameter API operation for Amazon Simple Systems Manager (SSM). // -// Get information about a parameter by using the parameter name. +// Get information about a parameter by using the parameter name. Don't confuse +// this API action with the GetParameters API action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -5836,7 +5841,8 @@ func (c *SSM) GetParametersRequest(input *GetParametersInput) (req *request.Requ // GetParameters API operation for Amazon Simple Systems Manager (SSM). // -// Get details of a parameter. +// Get details of a parameter. Don't confuse this API action with the GetParameter +// API action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -5925,7 +5931,8 @@ func (c *SSM) GetParametersByPathRequest(input *GetParametersByPathInput) (req * // GetParametersByPath API operation for Amazon Simple Systems Manager (SSM). // // Retrieve parameters in a specific hierarchy. For more information, see Working -// with Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-working.html). +// with Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-working.html) +// in the AWS Systems Manager User Guide. // // Request results are returned on a best-effort basis. If you specify MaxResults // in the request, the response includes information up to the limit specified. @@ -6503,12 +6510,12 @@ func (c *SSM) ListCommandInvocationsRequest(input *ListCommandInvocationsInput) // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -6661,12 +6668,12 @@ func (c *SSM) ListCommandsRequest(input *ListCommandsInput) (req *request.Reques // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -7222,12 +7229,12 @@ func (c *SSM) ListInventoryEntriesRequest(input *ListInventoryEntriesInput) (req // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -7832,12 +7839,12 @@ func (c *SSM) PutInventoryRequest(input *PutInventoryInput) (req *request.Reques // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -7973,7 +7980,8 @@ func (c *SSM) PutParameterRequest(input *PutParameterInput) (req *request.Reques // // * ErrCodeHierarchyLevelLimitExceededException "HierarchyLevelLimitExceededException" // A hierarchy can have a maximum of 15 levels. For more information, see Working -// with Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-working.html). +// with Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-working.html) +// in the AWS Systems Manager User Guide. // // * ErrCodeHierarchyTypeMismatchException "HierarchyTypeMismatchException" // Parameter Store does not support changing a parameter type in a hierarchy. @@ -8650,12 +8658,12 @@ func (c *SSM) SendCommandRequest(input *SendCommandInput) (req *request.Request, // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -9085,12 +9093,12 @@ func (c *SSM) UpdateAssociationStatusRequest(input *UpdateAssociationStatusInput // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -9673,12 +9681,12 @@ func (c *SSM) UpdateManagedInstanceRoleRequest(input *UpdateManagedInstanceRoleI // // You do not have permission to access the instance. // -// The SSM Agent is not running. On managed instances and Linux instances, verify +// SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // -// The SSM Agent or EC2Config service is not registered to the SSM endpoint. -// Try reinstalling the SSM Agent or EC2Config service. +// SSM Agent or EC2Config service is not registered to the SSM endpoint. Try +// reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. @@ -11108,10 +11116,62 @@ func (s CancelCommandOutput) GoString() string { return s.String() } +// Configuration options for sending command output to CloudWatch Logs. +type CloudWatchOutputConfig struct { + _ struct{} `type:"structure"` + + // The name of the CloudWatch log group where you want to send command output. + // If you don't specify a group name, Systems Manager automatically creates + // a log group for you. The log group uses the following naming format: aws/ssm/SystemsManagerDocumentName. + CloudWatchLogGroupName *string `min:"1" type:"string"` + + // Enables Systems Manager to send command output to CloudWatch Logs. + CloudWatchOutputEnabled *bool `type:"boolean"` +} + +// String returns the string representation +func (s CloudWatchOutputConfig) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CloudWatchOutputConfig) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CloudWatchOutputConfig) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CloudWatchOutputConfig"} + if s.CloudWatchLogGroupName != nil && len(*s.CloudWatchLogGroupName) < 1 { + invalidParams.Add(request.NewErrParamMinLen("CloudWatchLogGroupName", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetCloudWatchLogGroupName sets the CloudWatchLogGroupName field's value. +func (s *CloudWatchOutputConfig) SetCloudWatchLogGroupName(v string) *CloudWatchOutputConfig { + s.CloudWatchLogGroupName = &v + return s +} + +// SetCloudWatchOutputEnabled sets the CloudWatchOutputEnabled field's value. +func (s *CloudWatchOutputConfig) SetCloudWatchOutputEnabled(v bool) *CloudWatchOutputConfig { + s.CloudWatchOutputEnabled = &v + return s +} + // Describes a command request. type Command struct { _ struct{} `type:"structure"` + // CloudWatch Logs information where you want Systems Manager to send the command + // output. + CloudWatchOutputConfig *CloudWatchOutputConfig `type:"structure"` + // A unique identifier for this command. CommandId *string `min:"36" type:"string"` @@ -11124,6 +11184,9 @@ type Command struct { // Timed Out, Delivery Timed Out, Canceled, Terminated, or Undeliverable. CompletedCount *int64 `type:"integer"` + // The number of targets for which the status is Delivery Timed Out. + DeliveryTimedOutCount *int64 `type:"integer"` + // The name of the document requested for execution. DocumentName *string `type:"string"` @@ -11144,15 +11207,17 @@ type Command struct { // The maximum number of instances that are allowed to execute the command at // the same time. You can specify a number of instances, such as 10, or a percentage // of instances, such as 10%. The default value is 50. For more information - // about how to use MaxConcurrency, see Executing a Command Using Systems Manager - // Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/run-command.html). + // about how to use MaxConcurrency, see Executing Commands Using Systems Manager + // Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/run-command.html) + // in the AWS Systems Manager User Guide. MaxConcurrency *string `min:"1" type:"string"` // The maximum number of errors allowed before the system stops sending the // command to additional targets. You can specify a number of errors, such as // 10, or a percentage or errors, such as 10%. The default value is 0. For more - // information about how to use MaxErrors, see Executing a Command Using Systems - // Manager Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/run-command.html). + // information about how to use MaxErrors, see Executing Commands Using Systems + // Manager Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/run-command.html) + // in the AWS Systems Manager User Guide. MaxErrors *string `min:"1" type:"string"` // Configurations for sending notifications about command status changes. @@ -11187,8 +11252,10 @@ type Command struct { // A detailed status of the command execution. StatusDetails includes more information // than Status because it includes states resulting from error and concurrency // control parameters. StatusDetails can show different results than Status. - // For more information about these statuses, see Run Command Status (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-about-status.html). - // StatusDetails can be one of the following values: + // For more information about these statuses, see Understanding Command Statuses + // (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-commands.html) + // in the AWS Systems Manager User Guide. StatusDetails can be one of the following + // values: // // * Pending: The command has not been sent to any instances. // @@ -11238,6 +11305,12 @@ func (s Command) GoString() string { return s.String() } +// SetCloudWatchOutputConfig sets the CloudWatchOutputConfig field's value. +func (s *Command) SetCloudWatchOutputConfig(v *CloudWatchOutputConfig) *Command { + s.CloudWatchOutputConfig = v + return s +} + // SetCommandId sets the CommandId field's value. func (s *Command) SetCommandId(v string) *Command { s.CommandId = &v @@ -11256,6 +11329,12 @@ func (s *Command) SetCompletedCount(v int64) *Command { return s } +// SetDeliveryTimedOutCount sets the DeliveryTimedOutCount field's value. +func (s *Command) SetDeliveryTimedOutCount(v int64) *Command { + s.DeliveryTimedOutCount = &v + return s +} + // SetDocumentName sets the DocumentName field's value. func (s *Command) SetDocumentName(v string) *Command { s.DocumentName = &v @@ -11428,6 +11507,10 @@ func (s *CommandFilter) SetValue(v string) *CommandFilter { type CommandInvocation struct { _ struct{} `type:"structure"` + // CloudWatch Logs information where you want Systems Manager to send the command + // output. + CloudWatchOutputConfig *CloudWatchOutputConfig `type:"structure"` + // The command against which this invocation was requested. CommandId *string `min:"36" type:"string"` @@ -11481,8 +11564,9 @@ type CommandInvocation struct { // targeted by the command). StatusDetails includes more information than Status // because it includes states resulting from error and concurrency control parameters. // StatusDetails can show different results than Status. For more information - // about these statuses, see Run Command Status (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-about-status.html). - // StatusDetails can be one of the following values: + // about these statuses, see Understanding Command Statuses (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-commands.html) + // in the AWS Systems Manager User Guide. StatusDetails can be one of the following + // values: // // * Pending: The command has not been sent to the instance. // @@ -11535,6 +11619,12 @@ func (s CommandInvocation) GoString() string { return s.String() } +// SetCloudWatchOutputConfig sets the CloudWatchOutputConfig field's value. +func (s *CommandInvocation) SetCloudWatchOutputConfig(v *CloudWatchOutputConfig) *CommandInvocation { + s.CloudWatchOutputConfig = v + return s +} + // SetCommandId sets the CommandId field's value. func (s *CommandInvocation) SetCommandId(v string) *CommandInvocation { s.CommandId = &v @@ -11697,8 +11787,10 @@ type CommandPlugin struct { // A detailed status of the plugin execution. StatusDetails includes more information // than Status because it includes states resulting from error and concurrency // control parameters. StatusDetails can show different results than Status. - // For more information about these statuses, see Run Command Status (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-about-status.html). - // StatusDetails can be one of the following values: + // For more information about these statuses, see Understanding Command Statuses + // (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-commands.html) + // in the AWS Systems Manager User Guide. StatusDetails can be one of the following + // values: // // * Pending: The command has not been sent to the instance. // @@ -13481,7 +13573,8 @@ type DeleteInventoryOutput struct { DeletionId *string `type:"string"` // A summary of the delete operation. For more information about this summary, - // see Understanding the Delete Inventory Summary (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-inventory-delete.html#sysman-inventory-delete-summary). + // see Understanding the Delete Inventory Summary (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-inventory-delete.html#sysman-inventory-delete-summary) + // in the AWS Systems Manager User Guide. DeletionSummary *InventoryDeletionSummary `type:"structure"` // The name of the inventory data type specified in the request. @@ -17656,6 +17749,9 @@ func (s *GetCommandInvocationInput) SetPluginName(v string) *GetCommandInvocatio type GetCommandInvocationOutput struct { _ struct{} `type:"structure"` + // CloudWatch Logs information where Systems Manager sent the command output. + CloudWatchOutputConfig *CloudWatchOutputConfig `type:"structure"` + // The parent command ID of the invocation plugin. CommandId *string `min:"36" type:"string"` @@ -17726,9 +17822,10 @@ type GetCommandInvocationOutput struct { // A detailed status of the command execution for an invocation. StatusDetails // includes more information than Status because it includes states resulting // from error and concurrency control parameters. StatusDetails can show different - // results than Status. For more information about these statuses, see Run Command - // Status (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-about-status.html). - // StatusDetails can be one of the following values: + // results than Status. For more information about these statuses, see Understanding + // Command Statuses (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-commands.html) + // in the AWS Systems Manager User Guide. StatusDetails can be one of the following + // values: // // * Pending: The command has not been sent to the instance. // @@ -17783,6 +17880,12 @@ func (s GetCommandInvocationOutput) GoString() string { return s.String() } +// SetCloudWatchOutputConfig sets the CloudWatchOutputConfig field's value. +func (s *GetCommandInvocationOutput) SetCloudWatchOutputConfig(v *CloudWatchOutputConfig) *GetCommandInvocationOutput { + s.CloudWatchOutputConfig = v + return s +} + // SetCommandId sets the CommandId field's value. func (s *GetCommandInvocationOutput) SetCommandId(v string) *GetCommandInvocationOutput { s.CommandId = &v @@ -20255,7 +20358,7 @@ type InstanceInformation struct { // The activation ID created by Systems Manager when the server or VM was registered. ActivationId *string `type:"string"` - // The version of the SSM Agent running on your Linux instance. + // The version of SSM Agent running on your Linux instance. AgentVersion *string `type:"string"` // Information about the association. @@ -20270,14 +20373,15 @@ type InstanceInformation struct { // The IP address of the managed instance. IPAddress *string `min:"1" type:"string"` - // The Amazon Identity and Access Management (IAM) role assigned to EC2 instances - // or managed instances. + // The Amazon Identity and Access Management (IAM) role assigned to the on-premises + // Systems Manager managed instances. This call does not return the IAM role + // for Amazon EC2 instances. IamRole *string `type:"string"` // The instance ID. InstanceId *string `type:"string"` - // Indicates whether latest version of the SSM Agent is running on your instance. + // Indicates whether latest version of SSM Agent is running on your instance. // Some older versions of Windows Server use the EC2Config service to process // SSM requests. For this reason, this field does not indicate whether or not // the latest version is installed on Windows managed instances. @@ -20295,7 +20399,7 @@ type InstanceInformation struct { // The name of the managed instance. Name *string `type:"string"` - // Connection status of the SSM Agent. + // Connection status of SSM Agent. PingStatus *string `type:"string" enum:"PingStatus"` // The name of the operating system platform running on your instance. @@ -20857,7 +20961,8 @@ type InventoryDeletionStatusItem struct { DeletionStartTime *time.Time `type:"timestamp" timestampFormat:"unix"` // Information about the delete operation. For more information about this summary, - // see Understanding the Delete Inventory Summary (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-inventory-delete.html#sysman-inventory-delete-summary). + // see Understanding the Delete Inventory Summary (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-inventory-delete.html#sysman-inventory-delete-summary) + // in the AWS Systems Manager User Guide. DeletionSummary *InventoryDeletionSummary `type:"structure"` // The status of the operation. Possible values are InProgress and Complete. @@ -23999,8 +24104,8 @@ type NotificationConfig struct { // The different events for which you can receive notifications. These events // include the following: All (events), InProgress, Success, TimedOut, Cancelled, - // Failed. To learn more about these events, see Setting Up Events and Notifications - // (http://docs.aws.amazon.com/systems-manager/latest/userguide/monitor-commands.html) + // Failed. To learn more about these events, see Configuring Amazon SNS Notifications + // for Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/rc-sns-notifications.html) // in the AWS Systems Manager User Guide. NotificationEvents []*string `type:"list"` @@ -24280,6 +24385,8 @@ func (s *ParameterMetadata) SetVersion(v int64) *ParameterMetadata { } // One or more filters. Use a filter to return a more specific list of results. +// +// The Name field can't be used with the GetParametersByPath API action. type ParameterStringFilter struct { _ struct{} `type:"structure"` @@ -24725,6 +24832,10 @@ func (s *PatchComplianceData) SetTitle(v string) *PatchComplianceData { // // * WindowsServer2016 // +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// // Supported key:CLASSIFICATION // // Supported values: @@ -24776,6 +24887,10 @@ func (s *PatchComplianceData) SetTitle(v string) *PatchComplianceData { // // * Ubuntu16.04 // +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// // Supported key:PRIORITY // // Supported values: @@ -24829,6 +24944,54 @@ func (s *PatchComplianceData) SetTitle(v string) *PatchComplianceData { // // * AmazonLinux2017.09 // +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// +// Supported key:CLASSIFICATION +// +// Supported values: +// +// * Security +// +// * Bugfix +// +// * Enhancement +// +// * Recommended +// +// * Newpackage +// +// Supported key:SEVERITY +// +// Supported values: +// +// * Critical +// +// * Important +// +// * Medium +// +// * Low +// +// Amazon Linux 2 Operating Systems +// +// The supported keys for Amazon Linux 2 operating systems are PRODUCT, CLASSIFICATION, +// and SEVERITY. See the following lists for valid values for each of these +// keys. +// +// Supported key:PRODUCT +// +// Supported values: +// +// * AmazonLinux2 +// +// * AmazonLinux2.0 +// +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// // Supported key:CLASSIFICATION // // Supported values: @@ -24885,6 +25048,10 @@ func (s *PatchComplianceData) SetTitle(v string) *PatchComplianceData { // // * RedhatEnterpriseLinux7.4 // +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// // Supported key:CLASSIFICATION // // Supported values: @@ -24941,6 +25108,10 @@ func (s *PatchComplianceData) SetTitle(v string) *PatchComplianceData { // // * Suse12.9 // +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// // Supported key:CLASSIFICATION // // Supported values: @@ -24999,6 +25170,10 @@ func (s *PatchComplianceData) SetTitle(v string) *PatchComplianceData { // // * CentOS7.4 // +// * * +// +// Use a wildcard character (*) to target all supported operating system versions. +// // Supported key:CLASSIFICATION // // Supported values: @@ -25218,7 +25393,8 @@ type PatchRule struct { _ struct{} `type:"structure"` // The number of days after the release date of each patch matched by the rule - // the patch is marked as approved in the patch baseline. + // that the patch is marked as approved in the patch baseline. For example, + // a value of 7 means that patches are approved seven days after they are released. // // ApproveAfterDays is a required field ApproveAfterDays *int64 `type:"integer" required:"true"` @@ -25730,8 +25906,8 @@ type PutParameterInput struct { // The fully qualified name includes the complete hierarchy of the parameter // path and name. For example: /Dev/DBServer/MySQL/db-string13 // - // For information about parameter name requirements and restrictions, see About - // Creating Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-su-create.html#sysman-paramstore-su-create-about) + // For information about parameter name requirements and restrictions, see Creating + // Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-su-create.html) // in the AWS Systems Manager User Guide. // // The maximum length constraint listed below includes capacity for additional @@ -27061,6 +27237,9 @@ func (s SendAutomationSignalOutput) GoString() string { type SendCommandInput struct { _ struct{} `type:"structure"` + // Enables Systems Manager to send Run Command output to Amazon CloudWatch Logs. + CloudWatchOutputConfig *CloudWatchOutputConfig `type:"structure"` + // User-specified information about the command, such as a brief description // of what the command should do. Comment *string `type:"string"` @@ -27081,28 +27260,40 @@ type SendCommandInput struct { // DocumentName is a required field DocumentName *string `type:"string" required:"true"` - // The SSM document version to use in the request. You can specify Default, - // Latest, or a specific version number. + // The SSM document version to use in the request. You can specify $DEFAULT, + // $LATEST, or a specific version number. If you execute commands by using the + // AWS CLI, then you must escape the first two options by using a backslash. + // If you specify a version number, then you don't need to use the backslash. + // For example: + // + // --document-version "\$DEFAULT" + // + // --document-version "\$LATEST" + // + // --document-version "3" DocumentVersion *string `type:"string"` // The instance IDs where the command should execute. You can specify a maximum // of 50 IDs. If you prefer not to list individual instance IDs, you can instead // send commands to a fleet of instances using the Targets parameter, which // accepts EC2 tags. For more information about how to use Targets, see Sending - // Commands to a Fleet (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html). + // Commands to a Fleet (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html) + // in the AWS Systems Manager User Guide. InstanceIds []*string `type:"list"` // (Optional) The maximum number of instances that are allowed to execute the // command at the same time. You can specify a number such as 10 or a percentage // such as 10%. The default value is 50. For more information about how to use - // MaxConcurrency, see Using Concurrency Controls (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-velocity.html). + // MaxConcurrency, see Using Concurrency Controls (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html#send-commands-velocity) + // in the AWS Systems Manager User Guide. MaxConcurrency *string `min:"1" type:"string"` // The maximum number of errors allowed without the command failing. When the // command fails one more time beyond the value of MaxErrors, the systems stops // sending the command to additional targets. You can specify a number like // 10 or a percentage like 10%. The default value is 0. For more information - // about how to use MaxErrors, see Using Error Controls (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-maxerrors.html). + // about how to use MaxErrors, see Using Error Controls (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html#send-commands-maxerrors) + // in the AWS Systems Manager User Guide. MaxErrors *string `min:"1" type:"string"` // Configurations for sending notifications. @@ -27129,7 +27320,8 @@ type SendCommandInput struct { // (Optional) An array of search criteria that targets instances using a Key,Value // combination that you specify. Targets is required if you don't provide one // or more instance IDs in the call. For more information about how to use Targets, - // see Sending Commands to a Fleet (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html). + // see Sending Commands to a Fleet (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html) + // in the AWS Systems Manager User Guide. Targets []*Target `type:"list"` // If this time is reached and the command has not already started executing, @@ -27168,6 +27360,11 @@ func (s *SendCommandInput) Validate() error { if s.TimeoutSeconds != nil && *s.TimeoutSeconds < 30 { invalidParams.Add(request.NewErrParamMinValue("TimeoutSeconds", 30)) } + if s.CloudWatchOutputConfig != nil { + if err := s.CloudWatchOutputConfig.Validate(); err != nil { + invalidParams.AddNested("CloudWatchOutputConfig", err.(request.ErrInvalidParams)) + } + } if s.Targets != nil { for i, v := range s.Targets { if v == nil { @@ -27185,6 +27382,12 @@ func (s *SendCommandInput) Validate() error { return nil } +// SetCloudWatchOutputConfig sets the CloudWatchOutputConfig field's value. +func (s *SendCommandInput) SetCloudWatchOutputConfig(v *CloudWatchOutputConfig) *SendCommandInput { + s.CloudWatchOutputConfig = v + return s +} + // SetComment sets the Comment field's value. func (s *SendCommandInput) SetComment(v string) *SendCommandInput { s.Comment = &v @@ -27923,14 +28126,16 @@ type Target struct { // User-defined criteria for sending commands that target instances that meet // the criteria. Key can be tag: or InstanceIds. For more information // about how to send commands that target instances using Key,Value parameters, - // see Executing a Command Using Systems Manager Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html). + // see Targeting Multiple Instances (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html#send-commands-targeting) + // in the AWS Systems Manager User Guide. Key *string `min:"1" type:"string"` // User-defined criteria that maps to Key. For example, if you specified tag:ServerRole, // you could specify value:WebServer to execute a command on instances that // include Amazon EC2 tags of ServerRole,WebServer. For more information about // how to send commands that target instances using Key,Value parameters, see - // Executing a Command Using Systems Manager Run Command (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html). + // Sending Commands to a Fleet (http://docs.aws.amazon.com/systems-manager/latest/userguide/send-commands-multiple.html) + // in the AWS Systems Manager User Guide. Values []*string `type:"list"` } @@ -30027,6 +30232,9 @@ const ( // OperatingSystemAmazonLinux is a OperatingSystem enum value OperatingSystemAmazonLinux = "AMAZON_LINUX" + // OperatingSystemAmazonLinux2 is a OperatingSystem enum value + OperatingSystemAmazonLinux2 = "AMAZON_LINUX_2" + // OperatingSystemUbuntu is a OperatingSystem enum value OperatingSystemUbuntu = "UBUNTU" diff --git a/vendor/github.com/aws/aws-sdk-go/service/ssm/doc.go b/vendor/github.com/aws/aws-sdk-go/service/ssm/doc.go index 4f18dadcb..6964adba0 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ssm/doc.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ssm/doc.go @@ -15,7 +15,8 @@ // (http://docs.aws.amazon.com/systems-manager/latest/userguide/). // // To get started, verify prerequisites and configure managed instances. For -// more information, see Systems Manager Prerequisites (http://docs.aws.amazon.com/systems-manager/latest/userguide/systems-manager-setting-up.html). +// more information, see Systems Manager Prerequisites (http://docs.aws.amazon.com/systems-manager/latest/userguide/systems-manager-setting-up.html) +// in the AWS Systems Manager User Guide. // // For information about other API actions you can perform on Amazon EC2 instances, // see the Amazon EC2 API Reference (http://docs.aws.amazon.com/AWSEC2/latest/APIReference/). diff --git a/vendor/github.com/aws/aws-sdk-go/service/ssm/errors.go b/vendor/github.com/aws/aws-sdk-go/service/ssm/errors.go index 3321d9988..4cb00aed6 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ssm/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ssm/errors.go @@ -151,7 +151,8 @@ const ( // "HierarchyLevelLimitExceededException". // // A hierarchy can have a maximum of 15 levels. For more information, see Working - // with Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-working.html). + // with Systems Manager Parameters (http://docs.aws.amazon.com/systems-manager/latest/userguide/sysman-paramstore-working.html) + // in the AWS Systems Manager User Guide. ErrCodeHierarchyLevelLimitExceededException = "HierarchyLevelLimitExceededException" // ErrCodeHierarchyTypeMismatchException for service response error code @@ -305,12 +306,12 @@ const ( // // You do not have permission to access the instance. // - // The SSM Agent is not running. On managed instances and Linux instances, verify + // SSM Agent is not running. On managed instances and Linux instances, verify // that the SSM Agent is running. On EC2 Windows instances, verify that the // EC2Config service is running. // - // The SSM Agent or EC2Config service is not registered to the SSM endpoint. - // Try reinstalling the SSM Agent or EC2Config service. + // SSM Agent or EC2Config service is not registered to the SSM endpoint. Try + // reinstalling SSM Agent or EC2Config service. // // The instance is not in valid state. Valid states are: Running, Pending, Stopped, // Stopping. Invalid states are: Shutting-down and Terminated. diff --git a/vendor/github.com/aws/aws-sdk-go/service/ssm/service.go b/vendor/github.com/aws/aws-sdk-go/service/ssm/service.go index d414fb7d8..9a6b8f71c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/ssm/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/ssm/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "ssm" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "ssm" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SSM" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SSM client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/storagegateway/api.go b/vendor/github.com/aws/aws-sdk-go/service/storagegateway/api.go index 5825ae878..25a3628b6 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/storagegateway/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/storagegateway/api.go @@ -789,9 +789,9 @@ func (c *StorageGateway) CreateNFSFileShareRequest(input *CreateNFSFileShareInpu // CreateNFSFileShare API operation for AWS Storage Gateway. // -// Creates a file share on an existing file gateway. In Storage Gateway, a file -// share is a file system mount point backed by Amazon S3 cloud storage. Storage -// Gateway exposes file shares using a Network File System (NFS) interface. +// Creates a Network File System (NFS) file share on an existing file gateway. +// In Storage Gateway, a file share is a file system mount point backed by Amazon +// S3 cloud storage. Storage Gateway exposes file shares using a NFS interface. // This operation is only supported in the file gateway type. // // File gateway requires AWS Security Token Service (AWS STS) to be activated @@ -841,6 +841,102 @@ func (c *StorageGateway) CreateNFSFileShareWithContext(ctx aws.Context, input *C return out, req.Send() } +const opCreateSMBFileShare = "CreateSMBFileShare" + +// CreateSMBFileShareRequest generates a "aws/request.Request" representing the +// client's request for the CreateSMBFileShare operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See CreateSMBFileShare for more information on using the CreateSMBFileShare +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the CreateSMBFileShareRequest method. +// req, resp := client.CreateSMBFileShareRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/CreateSMBFileShare +func (c *StorageGateway) CreateSMBFileShareRequest(input *CreateSMBFileShareInput) (req *request.Request, output *CreateSMBFileShareOutput) { + op := &request.Operation{ + Name: opCreateSMBFileShare, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &CreateSMBFileShareInput{} + } + + output = &CreateSMBFileShareOutput{} + req = c.newRequest(op, input, output) + return +} + +// CreateSMBFileShare API operation for AWS Storage Gateway. +// +// Creates a Server Message Block (SMB) file share on an existing file gateway. +// In Storage Gateway, a file share is a file system mount point backed by Amazon +// S3 cloud storage. Storage Gateway expose file shares using a SMB interface. +// This operation is only supported in the file gateway type. +// +// File gateway requires AWS Security Token Service (AWS STS) to be activated +// to enable you create a file share. Make sure AWS STS is activated in the +// region you are creating your file gateway in. If AWS STS is not activated +// in the region, activate it. For information about how to activate AWS STS, +// see Activating and Deactivating AWS STS in an AWS Region in the AWS Identity +// and Access Management User Guide. +// +// File gateway does not support creating hard or symbolic links on a file share. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Storage Gateway's +// API operation CreateSMBFileShare for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidGatewayRequestException "InvalidGatewayRequestException" +// An exception occurred because an invalid gateway request was issued to the +// service. For more information, see the error and message fields. +// +// * ErrCodeInternalServerError "InternalServerError" +// An internal server error has occurred during the request. For more information, +// see the error and message fields. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/CreateSMBFileShare +func (c *StorageGateway) CreateSMBFileShare(input *CreateSMBFileShareInput) (*CreateSMBFileShareOutput, error) { + req, out := c.CreateSMBFileShareRequest(input) + return out, req.Send() +} + +// CreateSMBFileShareWithContext is the same as CreateSMBFileShare with the addition of +// the ability to pass a context and additional request options. +// +// See CreateSMBFileShare for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *StorageGateway) CreateSMBFileShareWithContext(ctx aws.Context, input *CreateSMBFileShareInput, opts ...request.Option) (*CreateSMBFileShareOutput, error) { + req, out := c.CreateSMBFileShareRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opCreateSnapshot = "CreateSnapshot" // CreateSnapshotRequest generates a "aws/request.Request" representing the @@ -2618,8 +2714,9 @@ func (c *StorageGateway) DescribeNFSFileSharesRequest(input *DescribeNFSFileShar // DescribeNFSFileShares API operation for AWS Storage Gateway. // -// Gets a description for one or more file shares from a file gateway. This -// operation is only supported in the file gateway type. +// Gets a description for one or more Network File System (NFS) file shares +// from a file gateway. This operation is only supported in the file gateway +// type. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2659,6 +2756,177 @@ func (c *StorageGateway) DescribeNFSFileSharesWithContext(ctx aws.Context, input return out, req.Send() } +const opDescribeSMBFileShares = "DescribeSMBFileShares" + +// DescribeSMBFileSharesRequest generates a "aws/request.Request" representing the +// client's request for the DescribeSMBFileShares operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeSMBFileShares for more information on using the DescribeSMBFileShares +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeSMBFileSharesRequest method. +// req, resp := client.DescribeSMBFileSharesRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/DescribeSMBFileShares +func (c *StorageGateway) DescribeSMBFileSharesRequest(input *DescribeSMBFileSharesInput) (req *request.Request, output *DescribeSMBFileSharesOutput) { + op := &request.Operation{ + Name: opDescribeSMBFileShares, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeSMBFileSharesInput{} + } + + output = &DescribeSMBFileSharesOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeSMBFileShares API operation for AWS Storage Gateway. +// +// Gets a description for one or more Server Message Block (SMB) file shares +// from a file gateway. This operation is only supported in the file gateway +// type. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Storage Gateway's +// API operation DescribeSMBFileShares for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidGatewayRequestException "InvalidGatewayRequestException" +// An exception occurred because an invalid gateway request was issued to the +// service. For more information, see the error and message fields. +// +// * ErrCodeInternalServerError "InternalServerError" +// An internal server error has occurred during the request. For more information, +// see the error and message fields. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/DescribeSMBFileShares +func (c *StorageGateway) DescribeSMBFileShares(input *DescribeSMBFileSharesInput) (*DescribeSMBFileSharesOutput, error) { + req, out := c.DescribeSMBFileSharesRequest(input) + return out, req.Send() +} + +// DescribeSMBFileSharesWithContext is the same as DescribeSMBFileShares with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeSMBFileShares for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *StorageGateway) DescribeSMBFileSharesWithContext(ctx aws.Context, input *DescribeSMBFileSharesInput, opts ...request.Option) (*DescribeSMBFileSharesOutput, error) { + req, out := c.DescribeSMBFileSharesRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + +const opDescribeSMBSettings = "DescribeSMBSettings" + +// DescribeSMBSettingsRequest generates a "aws/request.Request" representing the +// client's request for the DescribeSMBSettings operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See DescribeSMBSettings for more information on using the DescribeSMBSettings +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the DescribeSMBSettingsRequest method. +// req, resp := client.DescribeSMBSettingsRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/DescribeSMBSettings +func (c *StorageGateway) DescribeSMBSettingsRequest(input *DescribeSMBSettingsInput) (req *request.Request, output *DescribeSMBSettingsOutput) { + op := &request.Operation{ + Name: opDescribeSMBSettings, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &DescribeSMBSettingsInput{} + } + + output = &DescribeSMBSettingsOutput{} + req = c.newRequest(op, input, output) + return +} + +// DescribeSMBSettings API operation for AWS Storage Gateway. +// +// Gets a description of a Server Message Block (SMB) file share settings from +// a file gateway. This operation is only supported in the file gateway type. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Storage Gateway's +// API operation DescribeSMBSettings for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidGatewayRequestException "InvalidGatewayRequestException" +// An exception occurred because an invalid gateway request was issued to the +// service. For more information, see the error and message fields. +// +// * ErrCodeInternalServerError "InternalServerError" +// An internal server error has occurred during the request. For more information, +// see the error and message fields. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/DescribeSMBSettings +func (c *StorageGateway) DescribeSMBSettings(input *DescribeSMBSettingsInput) (*DescribeSMBSettingsOutput, error) { + req, out := c.DescribeSMBSettingsRequest(input) + return out, req.Send() +} + +// DescribeSMBSettingsWithContext is the same as DescribeSMBSettings with the addition of +// the ability to pass a context and additional request options. +// +// See DescribeSMBSettings for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *StorageGateway) DescribeSMBSettingsWithContext(ctx aws.Context, input *DescribeSMBSettingsInput, opts ...request.Option) (*DescribeSMBSettingsOutput, error) { + req, out := c.DescribeSMBSettingsRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opDescribeSnapshotSchedule = "DescribeSnapshotSchedule" // DescribeSnapshotScheduleRequest generates a "aws/request.Request" representing the @@ -3680,6 +3948,91 @@ func (c *StorageGateway) DisableGatewayWithContext(ctx aws.Context, input *Disab return out, req.Send() } +const opJoinDomain = "JoinDomain" + +// JoinDomainRequest generates a "aws/request.Request" representing the +// client's request for the JoinDomain operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See JoinDomain for more information on using the JoinDomain +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the JoinDomainRequest method. +// req, resp := client.JoinDomainRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/JoinDomain +func (c *StorageGateway) JoinDomainRequest(input *JoinDomainInput) (req *request.Request, output *JoinDomainOutput) { + op := &request.Operation{ + Name: opJoinDomain, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &JoinDomainInput{} + } + + output = &JoinDomainOutput{} + req = c.newRequest(op, input, output) + return +} + +// JoinDomain API operation for AWS Storage Gateway. +// +// Adds a file gateway to an Active Directory domain. This operation is only +// supported in the file gateway type that supports the SMB file protocol. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Storage Gateway's +// API operation JoinDomain for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidGatewayRequestException "InvalidGatewayRequestException" +// An exception occurred because an invalid gateway request was issued to the +// service. For more information, see the error and message fields. +// +// * ErrCodeInternalServerError "InternalServerError" +// An internal server error has occurred during the request. For more information, +// see the error and message fields. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/JoinDomain +func (c *StorageGateway) JoinDomain(input *JoinDomainInput) (*JoinDomainOutput, error) { + req, out := c.JoinDomainRequest(input) + return out, req.Send() +} + +// JoinDomainWithContext is the same as JoinDomain with the addition of +// the ability to pass a context and additional request options. +// +// See JoinDomain for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *StorageGateway) JoinDomainWithContext(ctx aws.Context, input *JoinDomainInput, opts ...request.Option) (*JoinDomainOutput, error) { + req, out := c.JoinDomainRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opListFileShares = "ListFileShares" // ListFileSharesRequest generates a "aws/request.Request" representing the @@ -5153,6 +5506,91 @@ func (c *StorageGateway) SetLocalConsolePasswordWithContext(ctx aws.Context, inp return out, req.Send() } +const opSetSMBGuestPassword = "SetSMBGuestPassword" + +// SetSMBGuestPasswordRequest generates a "aws/request.Request" representing the +// client's request for the SetSMBGuestPassword operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See SetSMBGuestPassword for more information on using the SetSMBGuestPassword +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the SetSMBGuestPasswordRequest method. +// req, resp := client.SetSMBGuestPasswordRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/SetSMBGuestPassword +func (c *StorageGateway) SetSMBGuestPasswordRequest(input *SetSMBGuestPasswordInput) (req *request.Request, output *SetSMBGuestPasswordOutput) { + op := &request.Operation{ + Name: opSetSMBGuestPassword, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &SetSMBGuestPasswordInput{} + } + + output = &SetSMBGuestPasswordOutput{} + req = c.newRequest(op, input, output) + return +} + +// SetSMBGuestPassword API operation for AWS Storage Gateway. +// +// Sets the password for the guest user “smbguest”. "smbguest" is the user when +// the Authentication method for the file share is “GuestAccess”. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Storage Gateway's +// API operation SetSMBGuestPassword for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidGatewayRequestException "InvalidGatewayRequestException" +// An exception occurred because an invalid gateway request was issued to the +// service. For more information, see the error and message fields. +// +// * ErrCodeInternalServerError "InternalServerError" +// An internal server error has occurred during the request. For more information, +// see the error and message fields. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/SetSMBGuestPassword +func (c *StorageGateway) SetSMBGuestPassword(input *SetSMBGuestPasswordInput) (*SetSMBGuestPasswordOutput, error) { + req, out := c.SetSMBGuestPasswordRequest(input) + return out, req.Send() +} + +// SetSMBGuestPasswordWithContext is the same as SetSMBGuestPassword with the addition of +// the ability to pass a context and additional request options. +// +// See SetSMBGuestPassword for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *StorageGateway) SetSMBGuestPasswordWithContext(ctx aws.Context, input *SetSMBGuestPasswordInput, opts ...request.Option) (*SetSMBGuestPasswordOutput, error) { + req, out := c.SetSMBGuestPasswordRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opShutdownGateway = "ShutdownGateway" // ShutdownGatewayRequest generates a "aws/request.Request" representing the @@ -5852,8 +6290,8 @@ func (c *StorageGateway) UpdateNFSFileShareRequest(input *UpdateNFSFileShareInpu // UpdateNFSFileShare API operation for AWS Storage Gateway. // -// Updates a file share. This operation is only supported in the file gateway -// type. +// Updates a Network File System (NFS) file share. This operation is only supported +// in the file gateway type. // // To leave a file share field unchanged, set the corresponding input field // to null. @@ -5911,6 +6349,103 @@ func (c *StorageGateway) UpdateNFSFileShareWithContext(ctx aws.Context, input *U return out, req.Send() } +const opUpdateSMBFileShare = "UpdateSMBFileShare" + +// UpdateSMBFileShareRequest generates a "aws/request.Request" representing the +// client's request for the UpdateSMBFileShare operation. The "output" return +// value will be populated with the request's response once the request completes +// successfuly. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See UpdateSMBFileShare for more information on using the UpdateSMBFileShare +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the UpdateSMBFileShareRequest method. +// req, resp := client.UpdateSMBFileShareRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/UpdateSMBFileShare +func (c *StorageGateway) UpdateSMBFileShareRequest(input *UpdateSMBFileShareInput) (req *request.Request, output *UpdateSMBFileShareOutput) { + op := &request.Operation{ + Name: opUpdateSMBFileShare, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &UpdateSMBFileShareInput{} + } + + output = &UpdateSMBFileShareOutput{} + req = c.newRequest(op, input, output) + return +} + +// UpdateSMBFileShare API operation for AWS Storage Gateway. +// +// Updates a Server Message Block (SMB) file share. This operation is only supported +// in the file gateway type. +// +// To leave a file share field unchanged, set the corresponding input field +// to null. This operation is only supported in the file gateway type. +// +// File gateway requires AWS Security Token Service (AWS STS) to be activated +// to enable you create a file share. Make sure AWS STS is activated in the +// region you are creating your file gateway in. If AWS STS is not activated +// in the region, activate it. For information about how to activate AWS STS, +// see Activating and Deactivating AWS STS in an AWS Region in the AWS Identity +// and Access Management User Guide. +// +// File gateway does not support creating hard or symbolic links on a file share. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for AWS Storage Gateway's +// API operation UpdateSMBFileShare for usage and error information. +// +// Returned Error Codes: +// * ErrCodeInvalidGatewayRequestException "InvalidGatewayRequestException" +// An exception occurred because an invalid gateway request was issued to the +// service. For more information, see the error and message fields. +// +// * ErrCodeInternalServerError "InternalServerError" +// An internal server error has occurred during the request. For more information, +// see the error and message fields. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/storagegateway-2013-06-30/UpdateSMBFileShare +func (c *StorageGateway) UpdateSMBFileShare(input *UpdateSMBFileShareInput) (*UpdateSMBFileShareOutput, error) { + req, out := c.UpdateSMBFileShareRequest(input) + return out, req.Send() +} + +// UpdateSMBFileShareWithContext is the same as UpdateSMBFileShare with the addition of +// the ability to pass a context and additional request options. +// +// See UpdateSMBFileShare for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *StorageGateway) UpdateSMBFileShareWithContext(ctx aws.Context, input *UpdateSMBFileShareInput, opts ...request.Option) (*UpdateSMBFileShareOutput, error) { + req, out := c.UpdateSMBFileShareRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opUpdateSnapshotSchedule = "UpdateSnapshotSchedule" // UpdateSnapshotScheduleRequest generates a "aws/request.Request" representing the @@ -6643,6 +7178,10 @@ type CachediSCSIVolume struct { // don’t have this time stamp. CreatedDate *time.Time `type:"timestamp" timestampFormat:"unix"` + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + // If the cached volume was created from a snapshot, this field contains the // snapshot ID used, e.g. snap-78e22663. Otherwise, this field is not included. SourceSnapshotId *string `type:"string"` @@ -6694,6 +7233,12 @@ func (s *CachediSCSIVolume) SetCreatedDate(v time.Time) *CachediSCSIVolume { return s } +// SetKMSKey sets the KMSKey field's value. +func (s *CachediSCSIVolume) SetKMSKey(v string) *CachediSCSIVolume { + s.KMSKey = &v + return s +} + // SetSourceSnapshotId sets the SourceSnapshotId field's value. func (s *CachediSCSIVolume) SetSourceSnapshotId(v string) *CachediSCSIVolume { s.SourceSnapshotId = &v @@ -6980,6 +7525,9 @@ func (s *ChapInfo) SetTargetARN(v string) *ChapInfo { type CreateCachediSCSIVolumeInput struct { _ struct{} `type:"structure"` + // A unique identifier that you use to retry a request. If you retry a request, + // use the same ClientToken you specified in the initial request. + // // ClientToken is a required field ClientToken *string `min:"5" type:"string" required:"true"` @@ -6989,9 +7537,28 @@ type CreateCachediSCSIVolumeInput struct { // GatewayARN is a required field GatewayARN *string `min:"50" type:"string" required:"true"` + // True to use Amazon S3 server side encryption with your own AWS KMS key, or + // false to use a key managed by Amazon S3. Optional. + KMSEncrypted *bool `type:"boolean"` + + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + + // The network interface of the gateway on which to expose the iSCSI target. + // Only IPv4 addresses are accepted. Use DescribeGatewayInformation to get a + // list of the network interfaces available on a gateway. + // + // Valid Values: A valid IP address. + // // NetworkInterfaceId is a required field NetworkInterfaceId *string `type:"string" required:"true"` + // The snapshot ID (e.g. "snap-1122aabb") of the snapshot to restore as the + // new cached volume. Specify this field if you want to create the iSCSI storage + // volume from a snapshot otherwise do not include this field. To list snapshots + // for your account use DescribeSnapshots (http://docs.aws.amazon.com/AWSEC2/latest/APIReference/ApiReference-query-DescribeSnapshots.html) + // in the Amazon Elastic Compute Cloud API Reference. SnapshotId *string `type:"string"` // The ARN for an existing volume. Specifying this ARN makes the new volume @@ -7000,9 +7567,16 @@ type CreateCachediSCSIVolumeInput struct { // than the size of the existing volume, in bytes. SourceVolumeARN *string `min:"50" type:"string"` + // The name of the iSCSI target used by initiators to connect to the target + // and as a suffix for the target ARN. For example, specifying TargetName as + // myvolume results in the target ARN of arn:aws:storagegateway:us-east-2:111122223333:gateway/sgw-12A3456B/target/iqn.1997-05.com.amazon:myvolume. + // The target name must be unique across all volumes of a gateway. + // // TargetName is a required field TargetName *string `min:"1" type:"string" required:"true"` + // The size of the volume in bytes. + // // VolumeSizeInBytes is a required field VolumeSizeInBytes *int64 `type:"long" required:"true"` } @@ -7032,6 +7606,9 @@ func (s *CreateCachediSCSIVolumeInput) Validate() error { if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) } + if s.KMSKey != nil && len(*s.KMSKey) < 20 { + invalidParams.Add(request.NewErrParamMinLen("KMSKey", 20)) + } if s.NetworkInterfaceId == nil { invalidParams.Add(request.NewErrParamRequired("NetworkInterfaceId")) } @@ -7066,6 +7643,18 @@ func (s *CreateCachediSCSIVolumeInput) SetGatewayARN(v string) *CreateCachediSCS return s } +// SetKMSEncrypted sets the KMSEncrypted field's value. +func (s *CreateCachediSCSIVolumeInput) SetKMSEncrypted(v bool) *CreateCachediSCSIVolumeInput { + s.KMSEncrypted = &v + return s +} + +// SetKMSKey sets the KMSKey field's value. +func (s *CreateCachediSCSIVolumeInput) SetKMSKey(v string) *CreateCachediSCSIVolumeInput { + s.KMSKey = &v + return s +} + // SetNetworkInterfaceId sets the NetworkInterfaceId field's value. func (s *CreateCachediSCSIVolumeInput) SetNetworkInterfaceId(v string) *CreateCachediSCSIVolumeInput { s.NetworkInterfaceId = &v @@ -7099,8 +7688,11 @@ func (s *CreateCachediSCSIVolumeInput) SetVolumeSizeInBytes(v int64) *CreateCach type CreateCachediSCSIVolumeOutput struct { _ struct{} `type:"structure"` + // he Amazon Resource Name (ARN) of the volume target that includes the iSCSI + // name that initiators can use to connect to the target. TargetARN *string `min:"50" type:"string"` + // The Amazon Resource Name (ARN) of the configured volume. VolumeARN *string `min:"50" type:"string"` } @@ -7141,8 +7733,8 @@ type CreateNFSFileShareInput struct { ClientToken *string `min:"5" type:"string" required:"true"` // The default storage class for objects put into an Amazon S3 bucket by file - // gateway. Possible values are S3_STANDARD or S3_STANDARD_IA. If this field - // is not populated, the default value S3_STANDARD is used. Optional. + // gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. + // If this field is not populated, the default value S3_STANDARD is used. Optional. DefaultStorageClass *string `min:"5" type:"string"` // The Amazon Resource Name (ARN) of the file gateway on which you want to create @@ -7160,8 +7752,8 @@ type CreateNFSFileShareInput struct { // false to use a key managed by Amazon S3. Optional. KMSEncrypted *bool `type:"boolean"` - // The KMS key used for Amazon S3 server side encryption. This value can only - // be set when KmsEncrypted is true. Optional. + // The Amazon Resource Name (ARN) KMS key used for Amazon S3 server side encryption. + // This value can only be set when KMSEncrypted is true. Optional. KMSKey *string `min:"20" type:"string"` // The ARN of the backed storage used for storing file data. @@ -7370,6 +7962,239 @@ func (s *CreateNFSFileShareOutput) SetFileShareARN(v string) *CreateNFSFileShare return s } +// CreateSMBFileShareInput +type CreateSMBFileShareInput struct { + _ struct{} `type:"structure"` + + // The authentication method that users use to access the file share. + // + // Valid values: "ActiveDirectory" or "GuestAccess". The default is "ActiveDirectory". + Authentication *string `min:"5" type:"string"` + + // A unique string value that you supply that is used by file gateway to ensure + // idempotent file share creation. + // + // ClientToken is a required field + ClientToken *string `min:"5" type:"string" required:"true"` + + // The default storage class for objects put into an Amazon S3 bucket by file + // gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. + // If this field is not populated, the default value S3_STANDARD is used. Optional. + DefaultStorageClass *string `min:"5" type:"string"` + + // The Amazon Resource Name (ARN) of the file gateway on which you want to create + // a file share. + // + // GatewayARN is a required field + GatewayARN *string `min:"50" type:"string" required:"true"` + + // Enables guessing of the MIME type for uploaded objects based on file extensions. + // Set this value to true to enable MIME type guessing, and otherwise to false. + // The default value is true. + GuessMIMETypeEnabled *bool `type:"boolean"` + + // A list of users in the Active Directory that are not allowed to access the + // file share. Can only be set if Authentication is set to "ActiveDirectory". + InvalidUserList []*string `type:"list"` + + // True to use Amazon S3 server side encryption with your own AWS KMS key, or + // false to use a key managed by Amazon S3. Optional. + KMSEncrypted *bool `type:"boolean"` + + // The Amazon Resource Name (ARN) KMS key used for Amazon S3 server side encryption. + // This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + + // The ARN of the backed storage used for storing file data. + // + // LocationARN is a required field + LocationARN *string `min:"16" type:"string" required:"true"` + + // Sets the access control list permission for objects in the Amazon S3 bucket + // that a file gateway puts objects into. The default value is "private". + ObjectACL *string `type:"string" enum:"ObjectACL"` + + // Sets the write status of a file share. This value is true if the write status + // is read-only, and otherwise false. + ReadOnly *bool `type:"boolean"` + + // Sets who pays the cost of the request and the data download from the Amazon + // S3 bucket. Set this value to true if you want the requester to pay instead + // of the bucket owner, and otherwise to false. + RequesterPays *bool `type:"boolean"` + + // The ARN of the AWS Identity and Access Management (IAM) role that a file + // gateway assumes when it accesses the underlying storage. + // + // Role is a required field + Role *string `min:"20" type:"string" required:"true"` + + // A list of users in the Active Directory that are allowed to access the file + // share. Can only be set if Authentication is set to "ActiveDirectory". + ValidUserList []*string `type:"list"` +} + +// String returns the string representation +func (s CreateSMBFileShareInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateSMBFileShareInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *CreateSMBFileShareInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "CreateSMBFileShareInput"} + if s.Authentication != nil && len(*s.Authentication) < 5 { + invalidParams.Add(request.NewErrParamMinLen("Authentication", 5)) + } + if s.ClientToken == nil { + invalidParams.Add(request.NewErrParamRequired("ClientToken")) + } + if s.ClientToken != nil && len(*s.ClientToken) < 5 { + invalidParams.Add(request.NewErrParamMinLen("ClientToken", 5)) + } + if s.DefaultStorageClass != nil && len(*s.DefaultStorageClass) < 5 { + invalidParams.Add(request.NewErrParamMinLen("DefaultStorageClass", 5)) + } + if s.GatewayARN == nil { + invalidParams.Add(request.NewErrParamRequired("GatewayARN")) + } + if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { + invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) + } + if s.KMSKey != nil && len(*s.KMSKey) < 20 { + invalidParams.Add(request.NewErrParamMinLen("KMSKey", 20)) + } + if s.LocationARN == nil { + invalidParams.Add(request.NewErrParamRequired("LocationARN")) + } + if s.LocationARN != nil && len(*s.LocationARN) < 16 { + invalidParams.Add(request.NewErrParamMinLen("LocationARN", 16)) + } + if s.Role == nil { + invalidParams.Add(request.NewErrParamRequired("Role")) + } + if s.Role != nil && len(*s.Role) < 20 { + invalidParams.Add(request.NewErrParamMinLen("Role", 20)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetAuthentication sets the Authentication field's value. +func (s *CreateSMBFileShareInput) SetAuthentication(v string) *CreateSMBFileShareInput { + s.Authentication = &v + return s +} + +// SetClientToken sets the ClientToken field's value. +func (s *CreateSMBFileShareInput) SetClientToken(v string) *CreateSMBFileShareInput { + s.ClientToken = &v + return s +} + +// SetDefaultStorageClass sets the DefaultStorageClass field's value. +func (s *CreateSMBFileShareInput) SetDefaultStorageClass(v string) *CreateSMBFileShareInput { + s.DefaultStorageClass = &v + return s +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *CreateSMBFileShareInput) SetGatewayARN(v string) *CreateSMBFileShareInput { + s.GatewayARN = &v + return s +} + +// SetGuessMIMETypeEnabled sets the GuessMIMETypeEnabled field's value. +func (s *CreateSMBFileShareInput) SetGuessMIMETypeEnabled(v bool) *CreateSMBFileShareInput { + s.GuessMIMETypeEnabled = &v + return s +} + +// SetInvalidUserList sets the InvalidUserList field's value. +func (s *CreateSMBFileShareInput) SetInvalidUserList(v []*string) *CreateSMBFileShareInput { + s.InvalidUserList = v + return s +} + +// SetKMSEncrypted sets the KMSEncrypted field's value. +func (s *CreateSMBFileShareInput) SetKMSEncrypted(v bool) *CreateSMBFileShareInput { + s.KMSEncrypted = &v + return s +} + +// SetKMSKey sets the KMSKey field's value. +func (s *CreateSMBFileShareInput) SetKMSKey(v string) *CreateSMBFileShareInput { + s.KMSKey = &v + return s +} + +// SetLocationARN sets the LocationARN field's value. +func (s *CreateSMBFileShareInput) SetLocationARN(v string) *CreateSMBFileShareInput { + s.LocationARN = &v + return s +} + +// SetObjectACL sets the ObjectACL field's value. +func (s *CreateSMBFileShareInput) SetObjectACL(v string) *CreateSMBFileShareInput { + s.ObjectACL = &v + return s +} + +// SetReadOnly sets the ReadOnly field's value. +func (s *CreateSMBFileShareInput) SetReadOnly(v bool) *CreateSMBFileShareInput { + s.ReadOnly = &v + return s +} + +// SetRequesterPays sets the RequesterPays field's value. +func (s *CreateSMBFileShareInput) SetRequesterPays(v bool) *CreateSMBFileShareInput { + s.RequesterPays = &v + return s +} + +// SetRole sets the Role field's value. +func (s *CreateSMBFileShareInput) SetRole(v string) *CreateSMBFileShareInput { + s.Role = &v + return s +} + +// SetValidUserList sets the ValidUserList field's value. +func (s *CreateSMBFileShareInput) SetValidUserList(v []*string) *CreateSMBFileShareInput { + s.ValidUserList = v + return s +} + +// CreateSMBFileShareOutput +type CreateSMBFileShareOutput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the newly created file share. + FileShareARN *string `min:"50" type:"string"` +} + +// String returns the string representation +func (s CreateSMBFileShareOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s CreateSMBFileShareOutput) GoString() string { + return s.String() +} + +// SetFileShareARN sets the FileShareARN field's value. +func (s *CreateSMBFileShareOutput) SetFileShareARN(v string) *CreateSMBFileShareOutput { + s.FileShareARN = &v + return s +} + type CreateSnapshotFromVolumeRecoveryPointInput struct { _ struct{} `type:"structure"` @@ -7757,6 +8582,14 @@ type CreateTapeWithBarcodeInput struct { // GatewayARN is a required field GatewayARN *string `min:"50" type:"string" required:"true"` + // True to use Amazon S3 server side encryption with your own AWS KMS key, or + // false to use a key managed by Amazon S3. Optional. + KMSEncrypted *bool `type:"boolean"` + + // The Amazon Resource Name (ARN) of the KMS Key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + // The barcode that you want to assign to the tape. // // Barcodes cannot be reused. This includes barcodes used for tapes that have @@ -7792,6 +8625,9 @@ func (s *CreateTapeWithBarcodeInput) Validate() error { if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) } + if s.KMSKey != nil && len(*s.KMSKey) < 20 { + invalidParams.Add(request.NewErrParamMinLen("KMSKey", 20)) + } if s.TapeBarcode == nil { invalidParams.Add(request.NewErrParamRequired("TapeBarcode")) } @@ -7814,6 +8650,18 @@ func (s *CreateTapeWithBarcodeInput) SetGatewayARN(v string) *CreateTapeWithBarc return s } +// SetKMSEncrypted sets the KMSEncrypted field's value. +func (s *CreateTapeWithBarcodeInput) SetKMSEncrypted(v bool) *CreateTapeWithBarcodeInput { + s.KMSEncrypted = &v + return s +} + +// SetKMSKey sets the KMSKey field's value. +func (s *CreateTapeWithBarcodeInput) SetKMSKey(v string) *CreateTapeWithBarcodeInput { + s.KMSKey = &v + return s +} + // SetTapeBarcode sets the TapeBarcode field's value. func (s *CreateTapeWithBarcodeInput) SetTapeBarcode(v string) *CreateTapeWithBarcodeInput { s.TapeBarcode = &v @@ -7870,6 +8718,14 @@ type CreateTapesInput struct { // GatewayARN is a required field GatewayARN *string `min:"50" type:"string" required:"true"` + // True to use Amazon S3 server side encryption with your own AWS KMS key, or + // false to use a key managed by Amazon S3. Optional. + KMSEncrypted *bool `type:"boolean"` + + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + // The number of virtual tapes that you want to create. // // NumTapesToCreate is a required field @@ -7917,6 +8773,9 @@ func (s *CreateTapesInput) Validate() error { if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) } + if s.KMSKey != nil && len(*s.KMSKey) < 20 { + invalidParams.Add(request.NewErrParamMinLen("KMSKey", 20)) + } if s.NumTapesToCreate == nil { invalidParams.Add(request.NewErrParamRequired("NumTapesToCreate")) } @@ -7951,6 +8810,18 @@ func (s *CreateTapesInput) SetGatewayARN(v string) *CreateTapesInput { return s } +// SetKMSEncrypted sets the KMSEncrypted field's value. +func (s *CreateTapesInput) SetKMSEncrypted(v bool) *CreateTapesInput { + s.KMSEncrypted = &v + return s +} + +// SetKMSKey sets the KMSKey field's value. +func (s *CreateTapesInput) SetKMSKey(v string) *CreateTapesInput { + s.KMSKey = &v + return s +} + // SetNumTapesToCreate sets the NumTapesToCreate field's value. func (s *CreateTapesInput) SetNumTapesToCreate(v int64) *CreateTapesInput { s.NumTapesToCreate = &v @@ -9286,6 +10157,158 @@ func (s *DescribeNFSFileSharesOutput) SetNFSFileShareInfoList(v []*NFSFileShareI return s } +// DescribeSMBFileSharesInput +type DescribeSMBFileSharesInput struct { + _ struct{} `type:"structure"` + + // An array containing the Amazon Resource Name (ARN) of each file share to + // be described. + // + // FileShareARNList is a required field + FileShareARNList []*string `min:"1" type:"list" required:"true"` +} + +// String returns the string representation +func (s DescribeSMBFileSharesInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeSMBFileSharesInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeSMBFileSharesInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeSMBFileSharesInput"} + if s.FileShareARNList == nil { + invalidParams.Add(request.NewErrParamRequired("FileShareARNList")) + } + if s.FileShareARNList != nil && len(s.FileShareARNList) < 1 { + invalidParams.Add(request.NewErrParamMinLen("FileShareARNList", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetFileShareARNList sets the FileShareARNList field's value. +func (s *DescribeSMBFileSharesInput) SetFileShareARNList(v []*string) *DescribeSMBFileSharesInput { + s.FileShareARNList = v + return s +} + +// DescribeSMBFileSharesOutput +type DescribeSMBFileSharesOutput struct { + _ struct{} `type:"structure"` + + // An array containing a description for each requested file share. + SMBFileShareInfoList []*SMBFileShareInfo `type:"list"` +} + +// String returns the string representation +func (s DescribeSMBFileSharesOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeSMBFileSharesOutput) GoString() string { + return s.String() +} + +// SetSMBFileShareInfoList sets the SMBFileShareInfoList field's value. +func (s *DescribeSMBFileSharesOutput) SetSMBFileShareInfoList(v []*SMBFileShareInfo) *DescribeSMBFileSharesOutput { + s.SMBFileShareInfoList = v + return s +} + +type DescribeSMBSettingsInput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation + // to return a list of gateways for your account and region. + // + // GatewayARN is a required field + GatewayARN *string `min:"50" type:"string" required:"true"` +} + +// String returns the string representation +func (s DescribeSMBSettingsInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeSMBSettingsInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *DescribeSMBSettingsInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "DescribeSMBSettingsInput"} + if s.GatewayARN == nil { + invalidParams.Add(request.NewErrParamRequired("GatewayARN")) + } + if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { + invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *DescribeSMBSettingsInput) SetGatewayARN(v string) *DescribeSMBSettingsInput { + s.GatewayARN = &v + return s +} + +type DescribeSMBSettingsOutput struct { + _ struct{} `type:"structure"` + + // The name of the domain that the gateway is joined to. + DomainName *string `type:"string"` + + // The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation + // to return a list of gateways for your account and region. + GatewayARN *string `min:"50" type:"string"` + + // This value is true if a password for the guest user “smbguest” is set, and + // otherwise false. + SMBGuestPasswordSet *bool `type:"boolean"` +} + +// String returns the string representation +func (s DescribeSMBSettingsOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s DescribeSMBSettingsOutput) GoString() string { + return s.String() +} + +// SetDomainName sets the DomainName field's value. +func (s *DescribeSMBSettingsOutput) SetDomainName(v string) *DescribeSMBSettingsOutput { + s.DomainName = &v + return s +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *DescribeSMBSettingsOutput) SetGatewayARN(v string) *DescribeSMBSettingsOutput { + s.GatewayARN = &v + return s +} + +// SetSMBGuestPasswordSet sets the SMBGuestPasswordSet field's value. +func (s *DescribeSMBSettingsOutput) SetSMBGuestPasswordSet(v bool) *DescribeSMBSettingsOutput { + s.SMBGuestPasswordSet = &v + return s +} + // A JSON object containing the DescribeSnapshotScheduleInput$VolumeARN of the // volume. type DescribeSnapshotScheduleInput struct { @@ -10345,6 +11368,9 @@ type FileShareInfo struct { // and DELETING. FileShareStatus *string `min:"3" type:"string"` + // The type of the file share. + FileShareType *string `type:"string" enum:"FileShareType"` + // The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation // to return a list of gateways for your account and region. GatewayARN *string `min:"50" type:"string"` @@ -10378,6 +11404,12 @@ func (s *FileShareInfo) SetFileShareStatus(v string) *FileShareInfo { return s } +// SetFileShareType sets the FileShareType field's value. +func (s *FileShareInfo) SetFileShareType(v string) *FileShareInfo { + s.FileShareType = &v + return s +} + // SetGatewayARN sets the GatewayARN field's value. func (s *FileShareInfo) SetGatewayARN(v string) *FileShareInfo { s.GatewayARN = &v @@ -10449,6 +11481,117 @@ func (s *GatewayInfo) SetGatewayType(v string) *GatewayInfo { return s } +// JoinDomainInput +type JoinDomainInput struct { + _ struct{} `type:"structure"` + + // The name of the domain that you want the gateway to join. + // + // DomainName is a required field + DomainName *string `type:"string" required:"true"` + + // The unique Amazon Resource Name of the file gateway you want to add to the + // Active Directory domain. + // + // GatewayARN is a required field + GatewayARN *string `min:"50" type:"string" required:"true"` + + // Sets the password of the user who has permission to add the gateway to the + // Active Directory domain. + // + // Password is a required field + Password *string `type:"string" required:"true"` + + // Sets the user name of user who has permission to add the gateway to the Active + // Directory domain. + // + // UserName is a required field + UserName *string `type:"string" required:"true"` +} + +// String returns the string representation +func (s JoinDomainInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s JoinDomainInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *JoinDomainInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "JoinDomainInput"} + if s.DomainName == nil { + invalidParams.Add(request.NewErrParamRequired("DomainName")) + } + if s.GatewayARN == nil { + invalidParams.Add(request.NewErrParamRequired("GatewayARN")) + } + if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { + invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) + } + if s.Password == nil { + invalidParams.Add(request.NewErrParamRequired("Password")) + } + if s.UserName == nil { + invalidParams.Add(request.NewErrParamRequired("UserName")) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDomainName sets the DomainName field's value. +func (s *JoinDomainInput) SetDomainName(v string) *JoinDomainInput { + s.DomainName = &v + return s +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *JoinDomainInput) SetGatewayARN(v string) *JoinDomainInput { + s.GatewayARN = &v + return s +} + +// SetPassword sets the Password field's value. +func (s *JoinDomainInput) SetPassword(v string) *JoinDomainInput { + s.Password = &v + return s +} + +// SetUserName sets the UserName field's value. +func (s *JoinDomainInput) SetUserName(v string) *JoinDomainInput { + s.UserName = &v + return s +} + +// JoinDomainOutput +type JoinDomainOutput struct { + _ struct{} `type:"structure"` + + // The unique Amazon Resource Name of the gateway that joined the domain. + GatewayARN *string `min:"50" type:"string"` +} + +// String returns the string representation +func (s JoinDomainOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s JoinDomainOutput) GoString() string { + return s.String() +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *JoinDomainOutput) SetGatewayARN(v string) *JoinDomainOutput { + s.GatewayARN = &v + return s +} + // ListFileShareInput type ListFileSharesInput struct { _ struct{} `type:"structure"` @@ -11192,11 +12335,12 @@ func (s *ListVolumesOutput) SetVolumeInfos(v []*VolumeInfo) *ListVolumesOutput { return s } -// Describes file share default values. Files and folders stored as Amazon S3 -// objects in S3 buckets don't, by default, have Unix file permissions assigned -// to them. Upon discovery in an S3 bucket by Storage Gateway, the S3 objects -// that represent files and folders are assigned these default Unix permissions. -// This operation is only supported in the file gateway type. +// Describes Network File System (NFS) file share default values. Files and +// folders stored as Amazon S3 objects in S3 buckets don't, by default, have +// Unix file permissions assigned to them. Upon discovery in an S3 bucket by +// Storage Gateway, the S3 objects that represent files and folders are assigned +// these default Unix permissions. This operation is only supported in the file +// gateway type. type NFSFileShareDefaults struct { _ struct{} `type:"structure"` @@ -11279,8 +12423,8 @@ type NFSFileShareInfo struct { ClientList []*string `min:"1" type:"list"` // The default storage class for objects put into an Amazon S3 bucket by file - // gateway. Possible values are S3_STANDARD or S3_STANDARD_IA. If this field - // is not populated, the default value S3_STANDARD is used. Optional. + // gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. + // If this field is not populated, the default value S3_STANDARD is used. Optional. DefaultStorageClass *string `min:"5" type:"string"` // The Amazon Resource Name (ARN) of the file share. @@ -11306,17 +12450,19 @@ type NFSFileShareInfo struct { // to use a key managed by Amazon S3. Optional. KMSEncrypted *bool `type:"boolean"` - // The ARN of the KMS key used for Amazon S3 server side encryption. + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. KMSKey *string `min:"20" type:"string"` // The ARN of the backend storage used for storing file data. LocationARN *string `min:"16" type:"string"` - // Describes file share default values. Files and folders stored as Amazon S3 - // objects in S3 buckets don't, by default, have Unix file permissions assigned - // to them. Upon discovery in an S3 bucket by Storage Gateway, the S3 objects - // that represent files and folders are assigned these default Unix permissions. - // This operation is only supported in the file gateway type. + // Describes Network File System (NFS) file share default values. Files and + // folders stored as Amazon S3 objects in S3 buckets don't, by default, have + // Unix file permissions assigned to them. Upon discovery in an S3 bucket by + // Storage Gateway, the S3 objects that represent files and folders are assigned + // these default Unix permissions. This operation is only supported in the file + // gateway type. NFSFileShareDefaults *NFSFileShareDefaults `type:"structure"` // Sets the access control list permission for objects in the S3 bucket that @@ -11968,6 +13114,192 @@ func (s *RetrieveTapeRecoveryPointOutput) SetTapeARN(v string) *RetrieveTapeReco return s } +// The Windows file permissions and ownership information assigned, by default, +// to native S3 objects when file gateway discovers them in S3 buckets. This +// operation is only supported in file gateways. +type SMBFileShareInfo struct { + _ struct{} `type:"structure"` + + // The authentication method of the file share. Valid values: "ActiveDirectory" + // or "GuestAccess". The default is "ActiveDirectory". + Authentication *string `min:"5" type:"string"` + + // The default storage class for objects put into an Amazon S3 bucket by file + // gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. + // If this field is not populated, the default value S3_STANDARD is used. Optional. + DefaultStorageClass *string `min:"5" type:"string"` + + // The Amazon Resource Name (ARN) of the file share. + FileShareARN *string `min:"50" type:"string"` + + // The ID of the file share. + FileShareId *string `min:"12" type:"string"` + + // The status of the file share. Possible values are CREATING, UPDATING, AVAILABLE + // and DELETING. + FileShareStatus *string `min:"3" type:"string"` + + // The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation + // to return a list of gateways for your account and region. + GatewayARN *string `min:"50" type:"string"` + + // Enables guessing of the MIME type for uploaded objects based on file extensions. + // Set this value to true to enable MIME type guessing, and otherwise to false. + // The default value is true. + GuessMIMETypeEnabled *bool `type:"boolean"` + + // A list of users in the Active Directory that are not allowed to access the + // file share. Can only be set if Authentication is set to "ActiveDirectory". + InvalidUserList []*string `type:"list"` + + // True to use Amazon S3 server side encryption with your own KMS key, or false + // to use a key managed by Amazon S3. Optional. + KMSEncrypted *bool `type:"boolean"` + + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + + // The ARN of the backend storage used for storing file data. + LocationARN *string `min:"16" type:"string"` + + // Sets the access control list permission for objects in the S3 bucket that + // a file gateway puts objects into. The default value is "private". + ObjectACL *string `type:"string" enum:"ObjectACL"` + + // The file share path used by the SMB client to identify the mount point. + Path *string `type:"string"` + + // Sets the write status of a file share. This value is true if the write status + // is read-only, and otherwise false. + ReadOnly *bool `type:"boolean"` + + // Sets who pays the cost of the request and the data download from the Amazon + // S3 bucket. Set this value to true if you want the requester to pay instead + // of the bucket owner, and otherwise to false. + RequesterPays *bool `type:"boolean"` + + // The ARN of the IAM role that file gateway assumes when it accesses the underlying + // storage. + Role *string `min:"20" type:"string"` + + // A list of users in the Active Directory that are allowed to access the file + // share. Can only be set if Authentication is set to "ActiveDirectory". + ValidUserList []*string `type:"list"` +} + +// String returns the string representation +func (s SMBFileShareInfo) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SMBFileShareInfo) GoString() string { + return s.String() +} + +// SetAuthentication sets the Authentication field's value. +func (s *SMBFileShareInfo) SetAuthentication(v string) *SMBFileShareInfo { + s.Authentication = &v + return s +} + +// SetDefaultStorageClass sets the DefaultStorageClass field's value. +func (s *SMBFileShareInfo) SetDefaultStorageClass(v string) *SMBFileShareInfo { + s.DefaultStorageClass = &v + return s +} + +// SetFileShareARN sets the FileShareARN field's value. +func (s *SMBFileShareInfo) SetFileShareARN(v string) *SMBFileShareInfo { + s.FileShareARN = &v + return s +} + +// SetFileShareId sets the FileShareId field's value. +func (s *SMBFileShareInfo) SetFileShareId(v string) *SMBFileShareInfo { + s.FileShareId = &v + return s +} + +// SetFileShareStatus sets the FileShareStatus field's value. +func (s *SMBFileShareInfo) SetFileShareStatus(v string) *SMBFileShareInfo { + s.FileShareStatus = &v + return s +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *SMBFileShareInfo) SetGatewayARN(v string) *SMBFileShareInfo { + s.GatewayARN = &v + return s +} + +// SetGuessMIMETypeEnabled sets the GuessMIMETypeEnabled field's value. +func (s *SMBFileShareInfo) SetGuessMIMETypeEnabled(v bool) *SMBFileShareInfo { + s.GuessMIMETypeEnabled = &v + return s +} + +// SetInvalidUserList sets the InvalidUserList field's value. +func (s *SMBFileShareInfo) SetInvalidUserList(v []*string) *SMBFileShareInfo { + s.InvalidUserList = v + return s +} + +// SetKMSEncrypted sets the KMSEncrypted field's value. +func (s *SMBFileShareInfo) SetKMSEncrypted(v bool) *SMBFileShareInfo { + s.KMSEncrypted = &v + return s +} + +// SetKMSKey sets the KMSKey field's value. +func (s *SMBFileShareInfo) SetKMSKey(v string) *SMBFileShareInfo { + s.KMSKey = &v + return s +} + +// SetLocationARN sets the LocationARN field's value. +func (s *SMBFileShareInfo) SetLocationARN(v string) *SMBFileShareInfo { + s.LocationARN = &v + return s +} + +// SetObjectACL sets the ObjectACL field's value. +func (s *SMBFileShareInfo) SetObjectACL(v string) *SMBFileShareInfo { + s.ObjectACL = &v + return s +} + +// SetPath sets the Path field's value. +func (s *SMBFileShareInfo) SetPath(v string) *SMBFileShareInfo { + s.Path = &v + return s +} + +// SetReadOnly sets the ReadOnly field's value. +func (s *SMBFileShareInfo) SetReadOnly(v bool) *SMBFileShareInfo { + s.ReadOnly = &v + return s +} + +// SetRequesterPays sets the RequesterPays field's value. +func (s *SMBFileShareInfo) SetRequesterPays(v bool) *SMBFileShareInfo { + s.RequesterPays = &v + return s +} + +// SetRole sets the Role field's value. +func (s *SMBFileShareInfo) SetRole(v string) *SMBFileShareInfo { + s.Role = &v + return s +} + +// SetValidUserList sets the ValidUserList field's value. +func (s *SMBFileShareInfo) SetValidUserList(v []*string) *SMBFileShareInfo { + s.ValidUserList = v + return s +} + // SetLocalConsolePasswordInput type SetLocalConsolePasswordInput struct { _ struct{} `type:"structure"` @@ -12052,6 +13384,90 @@ func (s *SetLocalConsolePasswordOutput) SetGatewayARN(v string) *SetLocalConsole return s } +// SetSMBGuestPasswordInput +type SetSMBGuestPasswordInput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the file gateway the SMB file share is + // associated with. + // + // GatewayARN is a required field + GatewayARN *string `min:"50" type:"string" required:"true"` + + // The password you want to set for your SMB Server. + // + // Password is a required field + Password *string `min:"6" type:"string" required:"true"` +} + +// String returns the string representation +func (s SetSMBGuestPasswordInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SetSMBGuestPasswordInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *SetSMBGuestPasswordInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "SetSMBGuestPasswordInput"} + if s.GatewayARN == nil { + invalidParams.Add(request.NewErrParamRequired("GatewayARN")) + } + if s.GatewayARN != nil && len(*s.GatewayARN) < 50 { + invalidParams.Add(request.NewErrParamMinLen("GatewayARN", 50)) + } + if s.Password == nil { + invalidParams.Add(request.NewErrParamRequired("Password")) + } + if s.Password != nil && len(*s.Password) < 6 { + invalidParams.Add(request.NewErrParamMinLen("Password", 6)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *SetSMBGuestPasswordInput) SetGatewayARN(v string) *SetSMBGuestPasswordInput { + s.GatewayARN = &v + return s +} + +// SetPassword sets the Password field's value. +func (s *SetSMBGuestPasswordInput) SetPassword(v string) *SetSMBGuestPasswordInput { + s.Password = &v + return s +} + +type SetSMBGuestPasswordOutput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the gateway. Use the ListGateways operation + // to return a list of gateways for your account and region. + GatewayARN *string `min:"50" type:"string"` +} + +// String returns the string representation +func (s SetSMBGuestPasswordOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s SetSMBGuestPasswordOutput) GoString() string { + return s.String() +} + +// SetGatewayARN sets the GatewayARN field's value. +func (s *SetSMBGuestPasswordOutput) SetGatewayARN(v string) *SetSMBGuestPasswordOutput { + s.GatewayARN = &v + return s +} + // A JSON object containing the of the gateway to shut down. type ShutdownGatewayInput struct { _ struct{} `type:"structure"` @@ -12378,6 +13794,10 @@ func (s *Tag) SetValue(v string) *Tag { type Tape struct { _ struct{} `type:"structure"` + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + // For archiving virtual tapes, indicates how much data remains to be uploaded // before archiving is complete. // @@ -12419,6 +13839,12 @@ func (s Tape) GoString() string { return s.String() } +// SetKMSKey sets the KMSKey field's value. +func (s *Tape) SetKMSKey(v string) *Tape { + s.KMSKey = &v + return s +} + // SetProgress sets the Progress field's value. func (s *Tape) SetProgress(v float64) *Tape { s.Progress = &v @@ -12473,10 +13899,14 @@ type TapeArchive struct { // The time that the archiving of the virtual tape was completed. // - // The string format of the completion time is in the ISO8601 extended YYYY-MM-DD'T'HH:MM:SS'Z' + // The default time stamp format is in the ISO8601 extended YYYY-MM-DD'T'HH:MM:SS'Z' // format. CompletionTime *time.Time `type:"timestamp" timestampFormat:"unix"` + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + // The Amazon Resource Name (ARN) of the tape gateway that the virtual tape // is being retrieved to. // @@ -12489,6 +13919,7 @@ type TapeArchive struct { // The barcode that identifies the archived virtual tape. TapeBarcode *string `min:"7" type:"string"` + // The date the virtual tape was created. TapeCreatedDate *time.Time `type:"timestamp" timestampFormat:"unix"` // The size, in bytes, of the archived virtual tape. @@ -12519,6 +13950,12 @@ func (s *TapeArchive) SetCompletionTime(v time.Time) *TapeArchive { return s } +// SetKMSKey sets the KMSKey field's value. +func (s *TapeArchive) SetKMSKey(v string) *TapeArchive { + s.KMSKey = &v + return s +} + // SetRetrievedTo sets the RetrievedTo field's value. func (s *TapeArchive) SetRetrievedTo(v string) *TapeArchive { s.RetrievedTo = &v @@ -12632,8 +14069,8 @@ type TapeRecoveryPointInfo struct { // The time when the point-in-time view of the virtual tape was replicated for // later recovery. // - // The string format of the tape recovery point time is in the ISO8601 extended - // YYYY-MM-DD'T'HH:MM:SS'Z' format. + // The default time stamp format of the tape recovery point time is in the ISO8601 + // extended YYYY-MM-DD'T'HH:MM:SS'Z' format. TapeRecoveryPointTime *time.Time `type:"timestamp" timestampFormat:"unix"` // The size, in bytes, of the virtual tapes to recover. @@ -13210,8 +14647,8 @@ type UpdateNFSFileShareInput struct { ClientList []*string `min:"1" type:"list"` // The default storage class for objects put into an Amazon S3 bucket by a file - // gateway. Possible values are S3_STANDARD or S3_STANDARD_IA. If this field - // is not populated, the default value S3_STANDARD is used. Optional. + // gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. + // If this field is not populated, the default value S3_STANDARD is used. Optional. DefaultStorageClass *string `min:"5" type:"string"` // The Amazon Resource Name (ARN) of the file share to be updated. @@ -13228,8 +14665,8 @@ type UpdateNFSFileShareInput struct { // false to use a key managed by Amazon S3. Optional. KMSEncrypted *bool `type:"boolean"` - // The KMS key used for Amazon S3 server side encryption. This value can only - // be set when KmsEncrypted is true. Optional. + // The Amazon Resource Name (ARN) of the KMS key used for Amazon S3 server side + // encryption. This value can only be set when KMSEncrypted is true. Optional. KMSKey *string `min:"20" type:"string"` // The default values for the file share. Optional. @@ -13391,6 +14828,171 @@ func (s *UpdateNFSFileShareOutput) SetFileShareARN(v string) *UpdateNFSFileShare return s } +// UpdateSMBFileShareInput +type UpdateSMBFileShareInput struct { + _ struct{} `type:"structure"` + + // The default storage class for objects put into an Amazon S3 bucket by file + // gateway. Possible values are S3_STANDARD, S3_STANDARD_IA or S3_ONEZONE_IA. + // If this field is not populated, the default value S3_STANDARD is used. Optional. + DefaultStorageClass *string `min:"5" type:"string"` + + // The Amazon Resource Name (ARN) of the SMB file share you want to update. + // + // FileShareARN is a required field + FileShareARN *string `min:"50" type:"string" required:"true"` + + // Enables guessing of the MIME type for uploaded objects based on file extensions. + // Set this value to true to enable MIME type guessing, and otherwise to false. + // The default value is true. + GuessMIMETypeEnabled *bool `type:"boolean"` + + // A list of users in the Active Directory that are not allowed to access the + // file share. Can only be set if Authentication is set to "ActiveDirectory". + InvalidUserList []*string `type:"list"` + + // True to use Amazon S3 server side encryption with your own AWS KMS key, or + // false to use a key managed by Amazon S3. Optional. + KMSEncrypted *bool `type:"boolean"` + + // The Amazon Resource Name (ARN) KMS key used for Amazon S3 server side encryption. + // This value can only be set when KMSEncrypted is true. Optional. + KMSKey *string `min:"20" type:"string"` + + // Sets the access control list permission for objects in the Amazon S3 bucket + // that a file gateway puts objects into. The default value is "private". + ObjectACL *string `type:"string" enum:"ObjectACL"` + + // Sets the write status of a file share. This value is true if the write status + // is read-only, and otherwise false. + ReadOnly *bool `type:"boolean"` + + // Sets who pays the cost of the request and the data download from the Amazon + // S3 bucket. Set this value to true if you want the requester to pay instead + // of the bucket owner, and otherwise to false. + RequesterPays *bool `type:"boolean"` + + // A list of users in the Active Directory that are allowed to access the file + // share. Can only be set if Authentication is set to "ActiveDirectory". + ValidUserList []*string `type:"list"` +} + +// String returns the string representation +func (s UpdateSMBFileShareInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateSMBFileShareInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *UpdateSMBFileShareInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "UpdateSMBFileShareInput"} + if s.DefaultStorageClass != nil && len(*s.DefaultStorageClass) < 5 { + invalidParams.Add(request.NewErrParamMinLen("DefaultStorageClass", 5)) + } + if s.FileShareARN == nil { + invalidParams.Add(request.NewErrParamRequired("FileShareARN")) + } + if s.FileShareARN != nil && len(*s.FileShareARN) < 50 { + invalidParams.Add(request.NewErrParamMinLen("FileShareARN", 50)) + } + if s.KMSKey != nil && len(*s.KMSKey) < 20 { + invalidParams.Add(request.NewErrParamMinLen("KMSKey", 20)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetDefaultStorageClass sets the DefaultStorageClass field's value. +func (s *UpdateSMBFileShareInput) SetDefaultStorageClass(v string) *UpdateSMBFileShareInput { + s.DefaultStorageClass = &v + return s +} + +// SetFileShareARN sets the FileShareARN field's value. +func (s *UpdateSMBFileShareInput) SetFileShareARN(v string) *UpdateSMBFileShareInput { + s.FileShareARN = &v + return s +} + +// SetGuessMIMETypeEnabled sets the GuessMIMETypeEnabled field's value. +func (s *UpdateSMBFileShareInput) SetGuessMIMETypeEnabled(v bool) *UpdateSMBFileShareInput { + s.GuessMIMETypeEnabled = &v + return s +} + +// SetInvalidUserList sets the InvalidUserList field's value. +func (s *UpdateSMBFileShareInput) SetInvalidUserList(v []*string) *UpdateSMBFileShareInput { + s.InvalidUserList = v + return s +} + +// SetKMSEncrypted sets the KMSEncrypted field's value. +func (s *UpdateSMBFileShareInput) SetKMSEncrypted(v bool) *UpdateSMBFileShareInput { + s.KMSEncrypted = &v + return s +} + +// SetKMSKey sets the KMSKey field's value. +func (s *UpdateSMBFileShareInput) SetKMSKey(v string) *UpdateSMBFileShareInput { + s.KMSKey = &v + return s +} + +// SetObjectACL sets the ObjectACL field's value. +func (s *UpdateSMBFileShareInput) SetObjectACL(v string) *UpdateSMBFileShareInput { + s.ObjectACL = &v + return s +} + +// SetReadOnly sets the ReadOnly field's value. +func (s *UpdateSMBFileShareInput) SetReadOnly(v bool) *UpdateSMBFileShareInput { + s.ReadOnly = &v + return s +} + +// SetRequesterPays sets the RequesterPays field's value. +func (s *UpdateSMBFileShareInput) SetRequesterPays(v bool) *UpdateSMBFileShareInput { + s.RequesterPays = &v + return s +} + +// SetValidUserList sets the ValidUserList field's value. +func (s *UpdateSMBFileShareInput) SetValidUserList(v []*string) *UpdateSMBFileShareInput { + s.ValidUserList = v + return s +} + +// UpdateSMBFileShareOutput +type UpdateSMBFileShareOutput struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the updated SMB file share. + FileShareARN *string `min:"50" type:"string"` +} + +// String returns the string representation +func (s UpdateSMBFileShareOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s UpdateSMBFileShareOutput) GoString() string { + return s.String() +} + +// SetFileShareARN sets the FileShareARN field's value. +func (s *UpdateSMBFileShareOutput) SetFileShareARN(v string) *UpdateSMBFileShareOutput { + s.FileShareARN = &v + return s +} + // A JSON object containing one or more of the following fields: // // * UpdateSnapshotScheduleInput$Description @@ -14030,6 +15632,15 @@ const ( ErrorCodeVolumeNotReady = "VolumeNotReady" ) +// The type of the file share. +const ( + // FileShareTypeNfs is a FileShareType enum value + FileShareTypeNfs = "NFS" + + // FileShareTypeSmb is a FileShareType enum value + FileShareTypeSmb = "SMB" +) + // Sets the access control list permission for objects in the S3 bucket that // a file gateway puts objects into. The default value is "private". const ( diff --git a/vendor/github.com/aws/aws-sdk-go/service/storagegateway/service.go b/vendor/github.com/aws/aws-sdk-go/service/storagegateway/service.go index 11c486c5f..9a0c08f69 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/storagegateway/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/storagegateway/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "storagegateway" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "storagegateway" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Storage Gateway" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the StorageGateway client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/storagegateway/storagegatewayiface/interface.go b/vendor/github.com/aws/aws-sdk-go/service/storagegateway/storagegatewayiface/interface.go index a2d2f4902..4b2128c2c 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/storagegateway/storagegatewayiface/interface.go +++ b/vendor/github.com/aws/aws-sdk-go/service/storagegateway/storagegatewayiface/interface.go @@ -96,6 +96,10 @@ type StorageGatewayAPI interface { CreateNFSFileShareWithContext(aws.Context, *storagegateway.CreateNFSFileShareInput, ...request.Option) (*storagegateway.CreateNFSFileShareOutput, error) CreateNFSFileShareRequest(*storagegateway.CreateNFSFileShareInput) (*request.Request, *storagegateway.CreateNFSFileShareOutput) + CreateSMBFileShare(*storagegateway.CreateSMBFileShareInput) (*storagegateway.CreateSMBFileShareOutput, error) + CreateSMBFileShareWithContext(aws.Context, *storagegateway.CreateSMBFileShareInput, ...request.Option) (*storagegateway.CreateSMBFileShareOutput, error) + CreateSMBFileShareRequest(*storagegateway.CreateSMBFileShareInput) (*request.Request, *storagegateway.CreateSMBFileShareOutput) + CreateSnapshot(*storagegateway.CreateSnapshotInput) (*storagegateway.CreateSnapshotOutput, error) CreateSnapshotWithContext(aws.Context, *storagegateway.CreateSnapshotInput, ...request.Option) (*storagegateway.CreateSnapshotOutput, error) CreateSnapshotRequest(*storagegateway.CreateSnapshotInput) (*request.Request, *storagegateway.CreateSnapshotOutput) @@ -176,6 +180,14 @@ type StorageGatewayAPI interface { DescribeNFSFileSharesWithContext(aws.Context, *storagegateway.DescribeNFSFileSharesInput, ...request.Option) (*storagegateway.DescribeNFSFileSharesOutput, error) DescribeNFSFileSharesRequest(*storagegateway.DescribeNFSFileSharesInput) (*request.Request, *storagegateway.DescribeNFSFileSharesOutput) + DescribeSMBFileShares(*storagegateway.DescribeSMBFileSharesInput) (*storagegateway.DescribeSMBFileSharesOutput, error) + DescribeSMBFileSharesWithContext(aws.Context, *storagegateway.DescribeSMBFileSharesInput, ...request.Option) (*storagegateway.DescribeSMBFileSharesOutput, error) + DescribeSMBFileSharesRequest(*storagegateway.DescribeSMBFileSharesInput) (*request.Request, *storagegateway.DescribeSMBFileSharesOutput) + + DescribeSMBSettings(*storagegateway.DescribeSMBSettingsInput) (*storagegateway.DescribeSMBSettingsOutput, error) + DescribeSMBSettingsWithContext(aws.Context, *storagegateway.DescribeSMBSettingsInput, ...request.Option) (*storagegateway.DescribeSMBSettingsOutput, error) + DescribeSMBSettingsRequest(*storagegateway.DescribeSMBSettingsInput) (*request.Request, *storagegateway.DescribeSMBSettingsOutput) + DescribeSnapshotSchedule(*storagegateway.DescribeSnapshotScheduleInput) (*storagegateway.DescribeSnapshotScheduleOutput, error) DescribeSnapshotScheduleWithContext(aws.Context, *storagegateway.DescribeSnapshotScheduleInput, ...request.Option) (*storagegateway.DescribeSnapshotScheduleOutput, error) DescribeSnapshotScheduleRequest(*storagegateway.DescribeSnapshotScheduleInput) (*request.Request, *storagegateway.DescribeSnapshotScheduleOutput) @@ -224,6 +236,10 @@ type StorageGatewayAPI interface { DisableGatewayWithContext(aws.Context, *storagegateway.DisableGatewayInput, ...request.Option) (*storagegateway.DisableGatewayOutput, error) DisableGatewayRequest(*storagegateway.DisableGatewayInput) (*request.Request, *storagegateway.DisableGatewayOutput) + JoinDomain(*storagegateway.JoinDomainInput) (*storagegateway.JoinDomainOutput, error) + JoinDomainWithContext(aws.Context, *storagegateway.JoinDomainInput, ...request.Option) (*storagegateway.JoinDomainOutput, error) + JoinDomainRequest(*storagegateway.JoinDomainInput) (*request.Request, *storagegateway.JoinDomainOutput) + ListFileShares(*storagegateway.ListFileSharesInput) (*storagegateway.ListFileSharesOutput, error) ListFileSharesWithContext(aws.Context, *storagegateway.ListFileSharesInput, ...request.Option) (*storagegateway.ListFileSharesOutput, error) ListFileSharesRequest(*storagegateway.ListFileSharesInput) (*request.Request, *storagegateway.ListFileSharesOutput) @@ -290,6 +306,10 @@ type StorageGatewayAPI interface { SetLocalConsolePasswordWithContext(aws.Context, *storagegateway.SetLocalConsolePasswordInput, ...request.Option) (*storagegateway.SetLocalConsolePasswordOutput, error) SetLocalConsolePasswordRequest(*storagegateway.SetLocalConsolePasswordInput) (*request.Request, *storagegateway.SetLocalConsolePasswordOutput) + SetSMBGuestPassword(*storagegateway.SetSMBGuestPasswordInput) (*storagegateway.SetSMBGuestPasswordOutput, error) + SetSMBGuestPasswordWithContext(aws.Context, *storagegateway.SetSMBGuestPasswordInput, ...request.Option) (*storagegateway.SetSMBGuestPasswordOutput, error) + SetSMBGuestPasswordRequest(*storagegateway.SetSMBGuestPasswordInput) (*request.Request, *storagegateway.SetSMBGuestPasswordOutput) + ShutdownGateway(*storagegateway.ShutdownGatewayInput) (*storagegateway.ShutdownGatewayOutput, error) ShutdownGatewayWithContext(aws.Context, *storagegateway.ShutdownGatewayInput, ...request.Option) (*storagegateway.ShutdownGatewayOutput, error) ShutdownGatewayRequest(*storagegateway.ShutdownGatewayInput) (*request.Request, *storagegateway.ShutdownGatewayOutput) @@ -322,6 +342,10 @@ type StorageGatewayAPI interface { UpdateNFSFileShareWithContext(aws.Context, *storagegateway.UpdateNFSFileShareInput, ...request.Option) (*storagegateway.UpdateNFSFileShareOutput, error) UpdateNFSFileShareRequest(*storagegateway.UpdateNFSFileShareInput) (*request.Request, *storagegateway.UpdateNFSFileShareOutput) + UpdateSMBFileShare(*storagegateway.UpdateSMBFileShareInput) (*storagegateway.UpdateSMBFileShareOutput, error) + UpdateSMBFileShareWithContext(aws.Context, *storagegateway.UpdateSMBFileShareInput, ...request.Option) (*storagegateway.UpdateSMBFileShareOutput, error) + UpdateSMBFileShareRequest(*storagegateway.UpdateSMBFileShareInput) (*request.Request, *storagegateway.UpdateSMBFileShareOutput) + UpdateSnapshotSchedule(*storagegateway.UpdateSnapshotScheduleInput) (*storagegateway.UpdateSnapshotScheduleOutput, error) UpdateSnapshotScheduleWithContext(aws.Context, *storagegateway.UpdateSnapshotScheduleInput, ...request.Option) (*storagegateway.UpdateSnapshotScheduleOutput, error) UpdateSnapshotScheduleRequest(*storagegateway.UpdateSnapshotScheduleInput) (*request.Request, *storagegateway.UpdateSnapshotScheduleOutput) diff --git a/vendor/github.com/aws/aws-sdk-go/service/sts/service.go b/vendor/github.com/aws/aws-sdk-go/service/sts/service.go index 1ee5839e0..185c914d1 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/sts/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/sts/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "sts" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "sts" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "STS" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the STS client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/support/service.go b/vendor/github.com/aws/aws-sdk-go/service/support/service.go index e22b2d375..dde3d2b7b 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/support/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/support/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "support" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "support" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Support" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Support client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/swf/service.go b/vendor/github.com/aws/aws-sdk-go/service/swf/service.go index ac85f412c..014d89a52 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/swf/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/swf/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "swf" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "swf" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "SWF" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the SWF client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/transcribeservice/service.go b/vendor/github.com/aws/aws-sdk-go/service/transcribeservice/service.go index f164d4295..bcecf55d5 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/transcribeservice/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/transcribeservice/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "transcribe" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "transcribe" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Transcribe Service" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the TranscribeService client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/translate/service.go b/vendor/github.com/aws/aws-sdk-go/service/translate/service.go index 45d734996..819b0f6d4 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/translate/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/translate/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "translate" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "translate" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "Translate" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the Translate client with a session. @@ -58,6 +59,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/waf/api.go b/vendor/github.com/aws/aws-sdk-go/service/waf/api.go index 3fd63c681..5f8d42c67 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/waf/api.go +++ b/vendor/github.com/aws/aws-sdk-go/service/waf/api.go @@ -86,18 +86,18 @@ func (c *WAF) CreateByteMatchSetRequest(input *CreateByteMatchSetInput) (req *re // API operation CreateByteMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -126,11 +126,11 @@ func (c *WAF) CreateByteMatchSetRequest(input *CreateByteMatchSetInput) (req *re // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -232,22 +232,22 @@ func (c *WAF) CreateGeoMatchSetRequest(input *CreateGeoMatchSetInput) (req *requ // API operation CreateGeoMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -276,7 +276,7 @@ func (c *WAF) CreateGeoMatchSetRequest(input *CreateGeoMatchSetInput) (req *requ // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -379,22 +379,22 @@ func (c *WAF) CreateIPSetRequest(input *CreateIPSetInput) (req *request.Request, // API operation CreateIPSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -423,7 +423,7 @@ func (c *WAF) CreateIPSetRequest(input *CreateIPSetInput) (req *request.Request, // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -565,18 +565,18 @@ func (c *WAF) CreateRateBasedRuleRequest(input *CreateRateBasedRuleInput) (req * // API operation CreateRateBasedRule for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -605,7 +605,7 @@ func (c *WAF) CreateRateBasedRuleRequest(input *CreateRateBasedRuleInput) (req * // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -709,18 +709,18 @@ func (c *WAF) CreateRegexMatchSetRequest(input *CreateRegexMatchSetInput) (req * // API operation CreateRegexMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -820,18 +820,18 @@ func (c *WAF) CreateRegexPatternSetRequest(input *CreateRegexPatternSetInput) (r // API operation CreateRegexPatternSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -948,18 +948,18 @@ func (c *WAF) CreateRuleRequest(input *CreateRuleInput) (req *request.Request, o // API operation CreateRule for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -988,7 +988,7 @@ func (c *WAF) CreateRuleRequest(input *CreateRuleInput) (req *request.Request, o // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -1083,18 +1083,18 @@ func (c *WAF) CreateRuleGroupRequest(input *CreateRuleGroupInput) (req *request. // API operation CreateRuleGroup for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -1198,22 +1198,22 @@ func (c *WAF) CreateSizeConstraintSetRequest(input *CreateSizeConstraintSetInput // API operation CreateSizeConstraintSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -1242,7 +1242,7 @@ func (c *WAF) CreateSizeConstraintSetRequest(input *CreateSizeConstraintSetInput // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -1342,18 +1342,18 @@ func (c *WAF) CreateSqlInjectionMatchSetRequest(input *CreateSqlInjectionMatchSe // API operation CreateSqlInjectionMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -1382,11 +1382,11 @@ func (c *WAF) CreateSqlInjectionMatchSetRequest(input *CreateSqlInjectionMatchSe // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -1498,22 +1498,22 @@ func (c *WAF) CreateWebACLRequest(input *CreateWebACLInput) (req *request.Reques // API operation CreateWebACL for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -1542,7 +1542,7 @@ func (c *WAF) CreateWebACLRequest(input *CreateWebACLInput) (req *request.Reques // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -1643,18 +1643,18 @@ func (c *WAF) CreateXssMatchSetRequest(input *CreateXssMatchSetInput) (req *requ // API operation CreateXssMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -1683,11 +1683,11 @@ func (c *WAF) CreateXssMatchSetRequest(input *CreateXssMatchSetInput) (req *requ // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -1783,18 +1783,18 @@ func (c *WAF) DeleteByteMatchSetRequest(input *DeleteByteMatchSetInput) (req *re // API operation DeleteByteMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -1802,11 +1802,11 @@ func (c *WAF) DeleteByteMatchSetRequest(input *DeleteByteMatchSetInput) (req *re // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -1909,22 +1909,22 @@ func (c *WAF) DeleteGeoMatchSetRequest(input *DeleteGeoMatchSetInput) (req *requ // API operation DeleteGeoMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -1932,7 +1932,7 @@ func (c *WAF) DeleteGeoMatchSetRequest(input *DeleteGeoMatchSetInput) (req *requ // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2035,22 +2035,22 @@ func (c *WAF) DeleteIPSetRequest(input *DeleteIPSetInput) (req *request.Request, // API operation DeleteIPSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2058,7 +2058,7 @@ func (c *WAF) DeleteIPSetRequest(input *DeleteIPSetInput) (req *request.Request, // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2150,15 +2150,15 @@ func (c *WAF) DeletePermissionPolicyRequest(input *DeletePermissionPolicyInput) // API operation DeletePermissionPolicy for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/DeletePermissionPolicy @@ -2252,22 +2252,22 @@ func (c *WAF) DeleteRateBasedRuleRequest(input *DeleteRateBasedRuleInput) (req * // API operation DeleteRateBasedRule for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2275,7 +2275,7 @@ func (c *WAF) DeleteRateBasedRuleRequest(input *DeleteRateBasedRuleInput) (req * // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2379,18 +2379,18 @@ func (c *WAF) DeleteRegexMatchSetRequest(input *DeleteRegexMatchSetInput) (req * // API operation DeleteRegexMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2398,11 +2398,11 @@ func (c *WAF) DeleteRegexMatchSetRequest(input *DeleteRegexMatchSetInput) (req * // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2494,18 +2494,18 @@ func (c *WAF) DeleteRegexPatternSetRequest(input *DeleteRegexPatternSetInput) (r // API operation DeleteRegexPatternSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2513,11 +2513,11 @@ func (c *WAF) DeleteRegexPatternSetRequest(input *DeleteRegexPatternSetInput) (r // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2620,22 +2620,22 @@ func (c *WAF) DeleteRuleRequest(input *DeleteRuleInput) (req *request.Request, o // API operation DeleteRule for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2643,7 +2643,7 @@ func (c *WAF) DeleteRuleRequest(input *DeleteRuleInput) (req *request.Request, o // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2745,18 +2745,18 @@ func (c *WAF) DeleteRuleGroupRequest(input *DeleteRuleGroupInput) (req *request. // API operation DeleteRuleGroup for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2764,7 +2764,7 @@ func (c *WAF) DeleteRuleGroupRequest(input *DeleteRuleGroupInput) (req *request. // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2868,22 +2868,22 @@ func (c *WAF) DeleteSizeConstraintSetRequest(input *DeleteSizeConstraintSetInput // API operation DeleteSizeConstraintSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -2891,7 +2891,7 @@ func (c *WAF) DeleteSizeConstraintSetRequest(input *DeleteSizeConstraintSetInput // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -2996,18 +2996,18 @@ func (c *WAF) DeleteSqlInjectionMatchSetRequest(input *DeleteSqlInjectionMatchSe // API operation DeleteSqlInjectionMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -3015,11 +3015,11 @@ func (c *WAF) DeleteSqlInjectionMatchSetRequest(input *DeleteSqlInjectionMatchSe // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -3119,22 +3119,22 @@ func (c *WAF) DeleteWebACLRequest(input *DeleteWebACLInput) (req *request.Reques // API operation DeleteWebACL for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -3142,7 +3142,7 @@ func (c *WAF) DeleteWebACLRequest(input *DeleteWebACLInput) (req *request.Reques // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -3246,18 +3246,18 @@ func (c *WAF) DeleteXssMatchSetRequest(input *DeleteXssMatchSetInput) (req *requ // API operation DeleteXssMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -3265,11 +3265,11 @@ func (c *WAF) DeleteXssMatchSetRequest(input *DeleteXssMatchSetInput) (req *requ // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonEmptyEntityException "NonEmptyEntityException" +// * ErrCodeNonEmptyEntityException "WAFNonEmptyEntityException" // The operation failed because you tried to delete an object that isn't empty. // For example: // @@ -3359,15 +3359,15 @@ func (c *WAF) GetByteMatchSetRequest(input *GetByteMatchSetInput) (req *request. // API operation GetByteMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetByteMatchSet @@ -3460,7 +3460,7 @@ func (c *WAF) GetChangeTokenRequest(input *GetChangeTokenInput) (req *request.Re // API operation GetChangeToken for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // @@ -3550,10 +3550,10 @@ func (c *WAF) GetChangeTokenStatusRequest(input *GetChangeTokenStatusInput) (req // API operation GetChangeTokenStatus for usage and error information. // // Returned Error Codes: -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // @@ -3633,15 +3633,15 @@ func (c *WAF) GetGeoMatchSetRequest(input *GetGeoMatchSetInput) (req *request.Re // API operation GetGeoMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetGeoMatchSet @@ -3720,15 +3720,15 @@ func (c *WAF) GetIPSetRequest(input *GetIPSetInput) (req *request.Request, outpu // API operation GetIPSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetIPSet @@ -3807,11 +3807,11 @@ func (c *WAF) GetPermissionPolicyRequest(input *GetPermissionPolicyInput) (req * // API operation GetPermissionPolicy for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetPermissionPolicy @@ -3891,15 +3891,15 @@ func (c *WAF) GetRateBasedRuleRequest(input *GetRateBasedRuleInput) (req *reques // API operation GetRateBasedRule for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetRateBasedRule @@ -3981,18 +3981,18 @@ func (c *WAF) GetRateBasedRuleManagedKeysRequest(input *GetRateBasedRuleManagedK // API operation GetRateBasedRuleManagedKeys for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -4097,15 +4097,15 @@ func (c *WAF) GetRegexMatchSetRequest(input *GetRegexMatchSetInput) (req *reques // API operation GetRegexMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetRegexMatchSet @@ -4184,15 +4184,15 @@ func (c *WAF) GetRegexPatternSetRequest(input *GetRegexPatternSetInput) (req *re // API operation GetRegexPatternSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetRegexPatternSet @@ -4272,15 +4272,15 @@ func (c *WAF) GetRuleRequest(input *GetRuleInput) (req *request.Request, output // API operation GetRule for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetRule @@ -4362,11 +4362,11 @@ func (c *WAF) GetRuleGroupRequest(input *GetRuleGroupInput) (req *request.Reques // API operation GetRuleGroup for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetRuleGroup @@ -4455,10 +4455,10 @@ func (c *WAF) GetSampledRequestsRequest(input *GetSampledRequestsInput) (req *re // API operation GetSampledRequests for usage and error information. // // Returned Error Codes: -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // @@ -4538,15 +4538,15 @@ func (c *WAF) GetSizeConstraintSetRequest(input *GetSizeConstraintSetInput) (req // API operation GetSizeConstraintSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetSizeConstraintSet @@ -4625,15 +4625,15 @@ func (c *WAF) GetSqlInjectionMatchSetRequest(input *GetSqlInjectionMatchSetInput // API operation GetSqlInjectionMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetSqlInjectionMatchSet @@ -4712,15 +4712,15 @@ func (c *WAF) GetWebACLRequest(input *GetWebACLInput) (req *request.Request, out // API operation GetWebACL for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetWebACL @@ -4799,15 +4799,15 @@ func (c *WAF) GetXssMatchSetRequest(input *GetXssMatchSetInput) (req *request.Re // API operation GetXssMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/GetXssMatchSet @@ -4886,14 +4886,14 @@ func (c *WAF) ListActivatedRulesInRuleGroupRequest(input *ListActivatedRulesInRu // API operation ListActivatedRulesInRuleGroup for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -4998,11 +4998,11 @@ func (c *WAF) ListByteMatchSetsRequest(input *ListByteMatchSetsInput) (req *requ // API operation ListByteMatchSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5082,11 +5082,11 @@ func (c *WAF) ListGeoMatchSetsRequest(input *ListGeoMatchSetsInput) (req *reques // API operation ListGeoMatchSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5166,11 +5166,11 @@ func (c *WAF) ListIPSetsRequest(input *ListIPSetsInput) (req *request.Request, o // API operation ListIPSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5250,11 +5250,11 @@ func (c *WAF) ListRateBasedRulesRequest(input *ListRateBasedRulesInput) (req *re // API operation ListRateBasedRules for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5334,11 +5334,11 @@ func (c *WAF) ListRegexMatchSetsRequest(input *ListRegexMatchSetsInput) (req *re // API operation ListRegexMatchSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5418,11 +5418,11 @@ func (c *WAF) ListRegexPatternSetsRequest(input *ListRegexPatternSetsInput) (req // API operation ListRegexPatternSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5502,7 +5502,7 @@ func (c *WAF) ListRuleGroupsRequest(input *ListRuleGroupsInput) (req *request.Re // API operation ListRuleGroups for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // @@ -5582,11 +5582,11 @@ func (c *WAF) ListRulesRequest(input *ListRulesInput) (req *request.Request, out // API operation ListRules for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5666,11 +5666,11 @@ func (c *WAF) ListSizeConstraintSetsRequest(input *ListSizeConstraintSetsInput) // API operation ListSizeConstraintSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5750,11 +5750,11 @@ func (c *WAF) ListSqlInjectionMatchSetsRequest(input *ListSqlInjectionMatchSetsI // API operation ListSqlInjectionMatchSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -5834,10 +5834,10 @@ func (c *WAF) ListSubscribedRuleGroupsRequest(input *ListSubscribedRuleGroupsInp // API operation ListSubscribedRuleGroups for usage and error information. // // Returned Error Codes: -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // @@ -5917,11 +5917,11 @@ func (c *WAF) ListWebACLsRequest(input *ListWebACLsInput) (req *request.Request, // API operation ListWebACLs for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -6001,11 +6001,11 @@ func (c *WAF) ListXssMatchSetsRequest(input *ListXssMatchSetsInput) (req *reques // API operation ListXssMatchSets for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -6110,18 +6110,18 @@ func (c *WAF) PutPermissionPolicyRequest(input *PutPermissionPolicyInput) (req * // API operation PutPermissionPolicy for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInvalidPermissionPolicyException "InvalidPermissionPolicyException" +// * ErrCodeInvalidPermissionPolicyException "WAFInvalidPermissionPolicyException" // The operation failed because the specified policy is not in the proper format. // // The policy is subject to the following restrictions: @@ -6256,15 +6256,15 @@ func (c *WAF) UpdateByteMatchSetRequest(input *UpdateByteMatchSetInput) (req *re // API operation UpdateByteMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -6285,7 +6285,7 @@ func (c *WAF) UpdateByteMatchSetRequest(input *UpdateByteMatchSetInput) (req *re // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -6314,7 +6314,7 @@ func (c *WAF) UpdateByteMatchSetRequest(input *UpdateByteMatchSetInput) (req *re // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -6330,14 +6330,14 @@ func (c *WAF) UpdateByteMatchSetRequest(input *UpdateByteMatchSetInput) (req *re // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -6446,19 +6446,19 @@ func (c *WAF) UpdateGeoMatchSetRequest(input *UpdateGeoMatchSetInput) (req *requ // API operation UpdateGeoMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -6479,7 +6479,7 @@ func (c *WAF) UpdateGeoMatchSetRequest(input *UpdateGeoMatchSetInput) (req *requ // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -6508,7 +6508,7 @@ func (c *WAF) UpdateGeoMatchSetRequest(input *UpdateGeoMatchSetInput) (req *requ // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -6524,10 +6524,10 @@ func (c *WAF) UpdateGeoMatchSetRequest(input *UpdateGeoMatchSetInput) (req *requ // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -6535,7 +6535,7 @@ func (c *WAF) UpdateGeoMatchSetRequest(input *UpdateGeoMatchSetInput) (req *requ // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -6665,19 +6665,19 @@ func (c *WAF) UpdateIPSetRequest(input *UpdateIPSetInput) (req *request.Request, // API operation UpdateIPSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -6698,7 +6698,7 @@ func (c *WAF) UpdateIPSetRequest(input *UpdateIPSetInput) (req *request.Request, // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -6727,7 +6727,7 @@ func (c *WAF) UpdateIPSetRequest(input *UpdateIPSetInput) (req *request.Request, // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -6743,10 +6743,10 @@ func (c *WAF) UpdateIPSetRequest(input *UpdateIPSetInput) (req *request.Request, // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -6754,7 +6754,7 @@ func (c *WAF) UpdateIPSetRequest(input *UpdateIPSetInput) (req *request.Request, // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -6873,19 +6873,19 @@ func (c *WAF) UpdateRateBasedRuleRequest(input *UpdateRateBasedRuleInput) (req * // API operation UpdateRateBasedRule for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -6906,7 +6906,7 @@ func (c *WAF) UpdateRateBasedRuleRequest(input *UpdateRateBasedRuleInput) (req * // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -6935,7 +6935,7 @@ func (c *WAF) UpdateRateBasedRuleRequest(input *UpdateRateBasedRuleInput) (req * // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -6951,10 +6951,10 @@ func (c *WAF) UpdateRateBasedRuleRequest(input *UpdateRateBasedRuleInput) (req * // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -6962,7 +6962,7 @@ func (c *WAF) UpdateRateBasedRuleRequest(input *UpdateRateBasedRuleInput) (req * // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -7077,27 +7077,27 @@ func (c *WAF) UpdateRegexMatchSetRequest(input *UpdateRegexMatchSetInput) (req * // API operation UpdateRegexMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeDisallowedNameException "DisallowedNameException" +// * ErrCodeDisallowedNameException "WAFDisallowedNameException" // The name specified is invalid. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) // in the AWS WAF Developer Guide. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -7113,7 +7113,7 @@ func (c *WAF) UpdateRegexMatchSetRequest(input *UpdateRegexMatchSetInput) (req * // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -7134,7 +7134,7 @@ func (c *WAF) UpdateRegexMatchSetRequest(input *UpdateRegexMatchSetInput) (req * // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // @@ -7244,24 +7244,24 @@ func (c *WAF) UpdateRegexPatternSetRequest(input *UpdateRegexPatternSetInput) (r // API operation UpdateRegexPatternSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) // in the AWS WAF Developer Guide. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -7277,7 +7277,7 @@ func (c *WAF) UpdateRegexPatternSetRequest(input *UpdateRegexPatternSetInput) (r // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -7298,11 +7298,11 @@ func (c *WAF) UpdateRegexPatternSetRequest(input *UpdateRegexPatternSetInput) (r // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidRegexPatternException "InvalidRegexPatternException" +// * ErrCodeInvalidRegexPatternException "WAFInvalidRegexPatternException" // The regular expression (regex) you specified in RegexPatternString is invalid. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/UpdateRegexPatternSet @@ -7413,19 +7413,19 @@ func (c *WAF) UpdateRuleRequest(input *UpdateRuleInput) (req *request.Request, o // API operation UpdateRule for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -7446,7 +7446,7 @@ func (c *WAF) UpdateRuleRequest(input *UpdateRuleInput) (req *request.Request, o // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -7475,7 +7475,7 @@ func (c *WAF) UpdateRuleRequest(input *UpdateRuleInput) (req *request.Request, o // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -7491,10 +7491,10 @@ func (c *WAF) UpdateRuleRequest(input *UpdateRuleInput) (req *request.Request, o // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -7502,7 +7502,7 @@ func (c *WAF) UpdateRuleRequest(input *UpdateRuleInput) (req *request.Request, o // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -7606,15 +7606,15 @@ func (c *WAF) UpdateRuleGroupRequest(input *UpdateRuleGroupInput) (req *request. // API operation UpdateRuleGroup for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -7630,10 +7630,10 @@ func (c *WAF) UpdateRuleGroupRequest(input *UpdateRuleGroupInput) (req *request. // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -7654,13 +7654,13 @@ func (c *WAF) UpdateRuleGroupRequest(input *UpdateRuleGroupInput) (req *request. // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) // in the AWS WAF Developer Guide. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -7803,19 +7803,19 @@ func (c *WAF) UpdateSizeConstraintSetRequest(input *UpdateSizeConstraintSetInput // API operation UpdateSizeConstraintSet for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -7836,7 +7836,7 @@ func (c *WAF) UpdateSizeConstraintSetRequest(input *UpdateSizeConstraintSetInput // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -7865,7 +7865,7 @@ func (c *WAF) UpdateSizeConstraintSetRequest(input *UpdateSizeConstraintSetInput // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -7881,10 +7881,10 @@ func (c *WAF) UpdateSizeConstraintSetRequest(input *UpdateSizeConstraintSetInput // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -7892,7 +7892,7 @@ func (c *WAF) UpdateSizeConstraintSetRequest(input *UpdateSizeConstraintSetInput // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -8005,15 +8005,15 @@ func (c *WAF) UpdateSqlInjectionMatchSetRequest(input *UpdateSqlInjectionMatchSe // API operation UpdateSqlInjectionMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -8034,7 +8034,7 @@ func (c *WAF) UpdateSqlInjectionMatchSetRequest(input *UpdateSqlInjectionMatchSe // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -8063,7 +8063,7 @@ func (c *WAF) UpdateSqlInjectionMatchSetRequest(input *UpdateSqlInjectionMatchSe // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -8079,14 +8079,14 @@ func (c *WAF) UpdateSqlInjectionMatchSetRequest(input *UpdateSqlInjectionMatchSe // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) @@ -8215,19 +8215,19 @@ func (c *WAF) UpdateWebACLRequest(input *UpdateWebACLInput) (req *request.Reques // API operation UpdateWebACL for usage and error information. // // Returned Error Codes: -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -8248,7 +8248,7 @@ func (c *WAF) UpdateWebACLRequest(input *UpdateWebACLInput) (req *request.Reques // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -8277,7 +8277,7 @@ func (c *WAF) UpdateWebACLRequest(input *UpdateWebACLInput) (req *request.Reques // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -8293,10 +8293,10 @@ func (c *WAF) UpdateWebACLRequest(input *UpdateWebACLInput) (req *request.Reques // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeReferencedItemException "ReferencedItemException" +// * ErrCodeReferencedItemException "WAFReferencedItemException" // The operation failed because you tried to delete an object that is still // in use. For example: // @@ -8304,13 +8304,13 @@ func (c *WAF) UpdateWebACLRequest(input *UpdateWebACLInput) (req *request.Reques // // * You tried to delete a Rule that is still referenced by a WebACL. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) // in the AWS WAF Developer Guide. // -// * ErrCodeSubscriptionNotFoundException "SubscriptionNotFoundException" +// * ErrCodeSubscriptionNotFoundException "WAFSubscriptionNotFoundException" // The specified subscription does not exist. // // See also, https://docs.aws.amazon.com/goto/WebAPI/waf-2015-08-24/UpdateWebACL @@ -8420,15 +8420,15 @@ func (c *WAF) UpdateXssMatchSetRequest(input *UpdateXssMatchSetInput) (req *requ // API operation UpdateXssMatchSet for usage and error information. // // Returned Error Codes: -// * ErrCodeInternalErrorException "InternalErrorException" +// * ErrCodeInternalErrorException "WAFInternalErrorException" // The operation failed because of a system problem, even though the request // was valid. Retry your request. // -// * ErrCodeInvalidAccountException "InvalidAccountException" +// * ErrCodeInvalidAccountException "WAFInvalidAccountException" // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. // -// * ErrCodeInvalidOperationException "InvalidOperationException" +// * ErrCodeInvalidOperationException "WAFInvalidOperationException" // The operation failed because there was nothing to do. For example: // // * You tried to remove a Rule from a WebACL, but the Rule isn't in the @@ -8449,7 +8449,7 @@ func (c *WAF) UpdateXssMatchSetRequest(input *UpdateXssMatchSetInput) (req *requ // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. // -// * ErrCodeInvalidParameterException "InvalidParameterException" +// * ErrCodeInvalidParameterException "WAFInvalidParameterException" // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: // @@ -8478,7 +8478,7 @@ func (c *WAF) UpdateXssMatchSetRequest(input *UpdateXssMatchSetInput) (req *requ // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. // -// * ErrCodeNonexistentContainerException "NonexistentContainerException" +// * ErrCodeNonexistentContainerException "WAFNonexistentContainerException" // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: // @@ -8494,14 +8494,14 @@ func (c *WAF) UpdateXssMatchSetRequest(input *UpdateXssMatchSetInput) (req *requ // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. // -// * ErrCodeNonexistentItemException "NonexistentItemException" +// * ErrCodeNonexistentItemException "WAFNonexistentItemException" // The operation failed because the referenced object doesn't exist. // -// * ErrCodeStaleDataException "StaleDataException" +// * ErrCodeStaleDataException "WAFStaleDataException" // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. // -// * ErrCodeLimitsExceededException "LimitsExceededException" +// * ErrCodeLimitsExceededException "WAFLimitsExceededException" // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) diff --git a/vendor/github.com/aws/aws-sdk-go/service/waf/errors.go b/vendor/github.com/aws/aws-sdk-go/service/waf/errors.go index 1e85bed52..97850b5df 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/waf/errors.go +++ b/vendor/github.com/aws/aws-sdk-go/service/waf/errors.go @@ -5,27 +5,27 @@ package waf const ( // ErrCodeDisallowedNameException for service response error code - // "DisallowedNameException". + // "WAFDisallowedNameException". // // The name specified is invalid. - ErrCodeDisallowedNameException = "DisallowedNameException" + ErrCodeDisallowedNameException = "WAFDisallowedNameException" // ErrCodeInternalErrorException for service response error code - // "InternalErrorException". + // "WAFInternalErrorException". // // The operation failed because of a system problem, even though the request // was valid. Retry your request. - ErrCodeInternalErrorException = "InternalErrorException" + ErrCodeInternalErrorException = "WAFInternalErrorException" // ErrCodeInvalidAccountException for service response error code - // "InvalidAccountException". + // "WAFInvalidAccountException". // // The operation failed because you tried to create, update, or delete an object // by using an invalid account identifier. - ErrCodeInvalidAccountException = "InvalidAccountException" + ErrCodeInvalidAccountException = "WAFInvalidAccountException" // ErrCodeInvalidOperationException for service response error code - // "InvalidOperationException". + // "WAFInvalidOperationException". // // The operation failed because there was nothing to do. For example: // @@ -46,10 +46,10 @@ const ( // // * You tried to add a ByteMatchTuple to a ByteMatchSet, but the ByteMatchTuple // already exists in the specified WebACL. - ErrCodeInvalidOperationException = "InvalidOperationException" + ErrCodeInvalidOperationException = "WAFInvalidOperationException" // ErrCodeInvalidParameterException for service response error code - // "InvalidParameterException". + // "WAFInvalidParameterException". // // The operation failed because AWS WAF didn't recognize a parameter in the // request. For example: @@ -78,10 +78,10 @@ const ( // // * Your request references an ARN that is malformed, or corresponds to // a resource with which a web ACL cannot be associated. - ErrCodeInvalidParameterException = "InvalidParameterException" + ErrCodeInvalidParameterException = "WAFInvalidParameterException" // ErrCodeInvalidPermissionPolicyException for service response error code - // "InvalidPermissionPolicyException". + // "WAFInvalidPermissionPolicyException". // // The operation failed because the specified policy is not in the proper format. // @@ -104,25 +104,25 @@ const ( // * The user making the request must be the owner of the RuleGroup. // // * Your policy must be composed using IAM Policy version 2012-10-17. - ErrCodeInvalidPermissionPolicyException = "InvalidPermissionPolicyException" + ErrCodeInvalidPermissionPolicyException = "WAFInvalidPermissionPolicyException" // ErrCodeInvalidRegexPatternException for service response error code - // "InvalidRegexPatternException". + // "WAFInvalidRegexPatternException". // // The regular expression (regex) you specified in RegexPatternString is invalid. - ErrCodeInvalidRegexPatternException = "InvalidRegexPatternException" + ErrCodeInvalidRegexPatternException = "WAFInvalidRegexPatternException" // ErrCodeLimitsExceededException for service response error code - // "LimitsExceededException". + // "WAFLimitsExceededException". // // The operation exceeds a resource limit, for example, the maximum number of // WebACL objects that you can create for an AWS account. For more information, // see Limits (http://docs.aws.amazon.com/waf/latest/developerguide/limits.html) // in the AWS WAF Developer Guide. - ErrCodeLimitsExceededException = "LimitsExceededException" + ErrCodeLimitsExceededException = "WAFLimitsExceededException" // ErrCodeNonEmptyEntityException for service response error code - // "NonEmptyEntityException". + // "WAFNonEmptyEntityException". // // The operation failed because you tried to delete an object that isn't empty. // For example: @@ -136,10 +136,10 @@ const ( // objects. // // * You tried to delete an IPSet that references one or more IP addresses. - ErrCodeNonEmptyEntityException = "NonEmptyEntityException" + ErrCodeNonEmptyEntityException = "WAFNonEmptyEntityException" // ErrCodeNonexistentContainerException for service response error code - // "NonexistentContainerException". + // "WAFNonexistentContainerException". // // The operation failed because you tried to add an object to or delete an object // from another object that doesn't exist. For example: @@ -155,16 +155,16 @@ const ( // // * You tried to add a ByteMatchTuple to or delete a ByteMatchTuple from // a ByteMatchSet that doesn't exist. - ErrCodeNonexistentContainerException = "NonexistentContainerException" + ErrCodeNonexistentContainerException = "WAFNonexistentContainerException" // ErrCodeNonexistentItemException for service response error code - // "NonexistentItemException". + // "WAFNonexistentItemException". // // The operation failed because the referenced object doesn't exist. - ErrCodeNonexistentItemException = "NonexistentItemException" + ErrCodeNonexistentItemException = "WAFNonexistentItemException" // ErrCodeReferencedItemException for service response error code - // "ReferencedItemException". + // "WAFReferencedItemException". // // The operation failed because you tried to delete an object that is still // in use. For example: @@ -172,18 +172,18 @@ const ( // * You tried to delete a ByteMatchSet that is still referenced by a Rule. // // * You tried to delete a Rule that is still referenced by a WebACL. - ErrCodeReferencedItemException = "ReferencedItemException" + ErrCodeReferencedItemException = "WAFReferencedItemException" // ErrCodeStaleDataException for service response error code - // "StaleDataException". + // "WAFStaleDataException". // // The operation failed because you tried to create, update, or delete an object // by using a change token that has already been used. - ErrCodeStaleDataException = "StaleDataException" + ErrCodeStaleDataException = "WAFStaleDataException" // ErrCodeSubscriptionNotFoundException for service response error code - // "SubscriptionNotFoundException". + // "WAFSubscriptionNotFoundException". // // The specified subscription does not exist. - ErrCodeSubscriptionNotFoundException = "SubscriptionNotFoundException" + ErrCodeSubscriptionNotFoundException = "WAFSubscriptionNotFoundException" ) diff --git a/vendor/github.com/aws/aws-sdk-go/service/waf/service.go b/vendor/github.com/aws/aws-sdk-go/service/waf/service.go index 91648a19f..09bf43d9e 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/waf/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/waf/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "waf" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "waf" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "WAF" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the WAF client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/wafregional/service.go b/vendor/github.com/aws/aws-sdk-go/service/wafregional/service.go index 8f0e70947..3a267ae63 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/wafregional/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/wafregional/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "waf-regional" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "waf-regional" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "WAF Regional" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the WAFRegional client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/workdocs/service.go b/vendor/github.com/aws/aws-sdk-go/service/workdocs/service.go index e1313753d..3daf66fac 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/workdocs/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/workdocs/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "workdocs" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "workdocs" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "WorkDocs" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the WorkDocs client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/workmail/service.go b/vendor/github.com/aws/aws-sdk-go/service/workmail/service.go index 5a5b01b43..7cc5520d3 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/workmail/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/workmail/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "workmail" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "workmail" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "WorkMail" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the WorkMail client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/workspaces/service.go b/vendor/github.com/aws/aws-sdk-go/service/workspaces/service.go index 452417031..38e1cc2ee 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/workspaces/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/workspaces/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "workspaces" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "workspaces" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "WorkSpaces" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the WorkSpaces client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/aws/aws-sdk-go/service/xray/service.go b/vendor/github.com/aws/aws-sdk-go/service/xray/service.go index 12a8a295f..fdc5ea329 100644 --- a/vendor/github.com/aws/aws-sdk-go/service/xray/service.go +++ b/vendor/github.com/aws/aws-sdk-go/service/xray/service.go @@ -29,8 +29,9 @@ var initRequest func(*request.Request) // Service information constants const ( - ServiceName = "xray" // Service endpoint prefix API calls made to. - EndpointsID = ServiceName // Service ID for Regions and Endpoints metadata. + ServiceName = "xray" // Name of service. + EndpointsID = ServiceName // ID to lookup a service endpoint with. + ServiceID = "XRay" // ServiceID is a unique identifer of a specific service. ) // New creates a new instance of the XRay client with a session. @@ -55,6 +56,7 @@ func newClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegio cfg, metadata.ClientInfo{ ServiceName: ServiceName, + ServiceID: ServiceID, SigningName: signingName, SigningRegion: signingRegion, Endpoint: endpoint, diff --git a/vendor/github.com/coreos/etcd/.gitignore b/vendor/github.com/coreos/etcd/.gitignore index 8abb0c8e2..b055a9853 100644 --- a/vendor/github.com/coreos/etcd/.gitignore +++ b/vendor/github.com/coreos/etcd/.gitignore @@ -1,6 +1,8 @@ /agent-* /coverage /covdir +/docs +/vendor /gopath /gopath.proto /go-bindata @@ -16,4 +18,4 @@ *.test hack/tls-setup/certs .idea -*.bak \ No newline at end of file +*.bak diff --git a/vendor/github.com/coreos/etcd/.travis.yml b/vendor/github.com/coreos/etcd/.travis.yml index 5412b9d36..c19901527 100644 --- a/vendor/github.com/coreos/etcd/.travis.yml +++ b/vendor/github.com/coreos/etcd/.travis.yml @@ -6,7 +6,7 @@ sudo: required services: docker go: -- 1.9.6 +- 1.9.7 notifications: on_success: never diff --git a/vendor/github.com/coreos/etcd/Makefile b/vendor/github.com/coreos/etcd/Makefile index 3eb4678c2..4036e5cd3 100644 --- a/vendor/github.com/coreos/etcd/Makefile +++ b/vendor/github.com/coreos/etcd/Makefile @@ -20,12 +20,16 @@ clean: rm -f ./codecov rm -rf ./agent-* rm -rf ./covdir + rm -f ./*.coverprofile rm -f ./*.log rm -f ./bin/Dockerfile-release rm -rf ./bin/*.etcd + rm -rf ./default.etcd + rm -rf ./tests/e2e/default.etcd rm -rf ./gopath rm -rf ./gopath.proto rm -rf ./release + rm -f ./snapshot/localhost:* rm -f ./integration/127.0.0.1:* ./integration/localhost:* rm -f ./clientv3/integration/127.0.0.1:* ./clientv3/integration/localhost:* rm -f ./clientv3/ordering/127.0.0.1:* ./clientv3/ordering/localhost:* @@ -46,7 +50,8 @@ docker-remove: -GO_VERSION ?= 1.10.1 +# GO_VERSION ?= 1.10.3 +GO_VERSION ?= 1.9.6 ETCD_VERSION ?= $(shell git rev-parse --short HEAD || echo "GitNotFound") TEST_SUFFIX = $(shell date +%s | base64 | head -c 15) @@ -61,16 +66,16 @@ endif # Example: # GO_VERSION=1.8.7 make build-docker-test -# GO_VERSION=1.9.5 make build-docker-test +# GO_VERSION=1.9.7 make build-docker-test # make build-docker-test # # gcloud docker -- login -u _json_key -p "$(cat /etc/gcp-key-etcd-development.json)" https://gcr.io # GO_VERSION=1.8.7 make push-docker-test -# GO_VERSION=1.9.5 make push-docker-test +# GO_VERSION=1.9.7 make push-docker-test # make push-docker-test # # gsutil -m acl ch -u allUsers:R -r gs://artifacts.etcd-development.appspot.com -# GO_VERSION=1.9.5 make pull-docker-test +# GO_VERSION=1.9.7 make pull-docker-test # make pull-docker-test build-docker-test: diff --git a/vendor/github.com/coreos/etcd/auth/jwt_test.go b/vendor/github.com/coreos/etcd/auth/jwt_test.go index 35e11dee2..42724a2f6 100644 --- a/vendor/github.com/coreos/etcd/auth/jwt_test.go +++ b/vendor/github.com/coreos/etcd/auth/jwt_test.go @@ -16,6 +16,7 @@ package auth import ( "context" + "fmt" "testing" ) @@ -92,3 +93,8 @@ func TestJWTBad(t *testing.T) { } opts["priv-key"] = jwtPrivKey } + +// testJWTOpts is useful for passing to NewTokenProvider which requires a string. +func testJWTOpts() string { + return fmt.Sprintf("%s,pub-key=%s,priv-key=%s,sign-method=RS256", tokenTypeJWT, jwtPubKey, jwtPrivKey) +} diff --git a/vendor/github.com/coreos/etcd/auth/nop.go b/vendor/github.com/coreos/etcd/auth/nop.go new file mode 100644 index 000000000..d4378747b --- /dev/null +++ b/vendor/github.com/coreos/etcd/auth/nop.go @@ -0,0 +1,35 @@ +// Copyright 2018 The etcd Authors +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package auth + +import ( + "context" +) + +type tokenNop struct{} + +func (t *tokenNop) enable() {} +func (t *tokenNop) disable() {} +func (t *tokenNop) invalidateUser(string) {} +func (t *tokenNop) genTokenPrefix() (string, error) { return "", nil } +func (t *tokenNop) info(ctx context.Context, token string, rev uint64) (*AuthInfo, bool) { + return nil, false +} +func (t *tokenNop) assign(ctx context.Context, username string, revision uint64) (string, error) { + return "", ErrAuthFailed +} +func newTokenProviderNop() (*tokenNop, error) { + return &tokenNop{}, nil +} diff --git a/vendor/github.com/coreos/etcd/auth/store.go b/vendor/github.com/coreos/etcd/auth/store.go index d379ffa8e..c2455a903 100644 --- a/vendor/github.com/coreos/etcd/auth/store.go +++ b/vendor/github.com/coreos/etcd/auth/store.go @@ -73,6 +73,9 @@ const ( rootUser = "root" rootRole = "root" + tokenTypeSimple = "simple" + tokenTypeJWT = "jwt" + revBytesLen = 8 ) @@ -1050,11 +1053,15 @@ func NewTokenProvider(tokenOpts string, indexWaiter func(uint64) <-chan struct{} } switch tokenType { - case "simple": + case tokenTypeSimple: plog.Warningf("simple token is not cryptographically signed") return newTokenProviderSimple(indexWaiter), nil - case "jwt": + + case tokenTypeJWT: return newTokenProviderJWT(typeSpecificOpts) + + case "": + return newTokenProviderNop() default: plog.Errorf("unknown token type: %s", tokenType) return nil, ErrInvalidAuthOpts @@ -1067,7 +1074,7 @@ func (as *authStore) WithRoot(ctx context.Context) context.Context { } var ctxForAssign context.Context - if ts := as.tokenProvider.(*tokenSimple); ts != nil { + if ts, ok := as.tokenProvider.(*tokenSimple); ok && ts != nil { ctx1 := context.WithValue(ctx, AuthenticateParamIndex{}, uint64(0)) prefix, err := ts.genTokenPrefix() if err != nil { diff --git a/vendor/github.com/coreos/etcd/auth/store_test.go b/vendor/github.com/coreos/etcd/auth/store_test.go index 3e9b0da96..02c8539d3 100644 --- a/vendor/github.com/coreos/etcd/auth/store_test.go +++ b/vendor/github.com/coreos/etcd/auth/store_test.go @@ -48,7 +48,7 @@ func TestNewAuthStoreRevision(t *testing.T) { b, tPath := backend.NewDefaultTmpBackend() defer os.Remove(tPath) - tp, err := NewTokenProvider("simple", dummyIndexWaiter) + tp, err := NewTokenProvider(tokenTypeSimple, dummyIndexWaiter) if err != nil { t.Fatal(err) } @@ -76,7 +76,7 @@ func TestNewAuthStoreRevision(t *testing.T) { func setupAuthStore(t *testing.T) (store *authStore, teardownfunc func(t *testing.T)) { b, tPath := backend.NewDefaultTmpBackend() - tp, err := NewTokenProvider("simple", dummyIndexWaiter) + tp, err := NewTokenProvider(tokenTypeSimple, dummyIndexWaiter) if err != nil { t.Fatal(err) } @@ -513,7 +513,7 @@ func TestAuthInfoFromCtxRace(t *testing.T) { b, tPath := backend.NewDefaultTmpBackend() defer os.Remove(tPath) - tp, err := NewTokenProvider("simple", dummyIndexWaiter) + tp, err := NewTokenProvider(tokenTypeSimple, dummyIndexWaiter) if err != nil { t.Fatal(err) } @@ -579,7 +579,7 @@ func TestRecoverFromSnapshot(t *testing.T) { as.Close() - tp, err := NewTokenProvider("simple", dummyIndexWaiter) + tp, err := NewTokenProvider(tokenTypeSimple, dummyIndexWaiter) if err != nil { t.Fatal(err) } @@ -661,7 +661,7 @@ func TestRolesOrder(t *testing.T) { b, tPath := backend.NewDefaultTmpBackend() defer os.Remove(tPath) - tp, err := NewTokenProvider("simple", dummyIndexWaiter) + tp, err := NewTokenProvider(tokenTypeSimple, dummyIndexWaiter) if err != nil { t.Fatal(err) } @@ -702,12 +702,21 @@ func TestRolesOrder(t *testing.T) { } } -// TestAuthInfoFromCtxWithRoot ensures "WithRoot" properly embeds token in the context. -func TestAuthInfoFromCtxWithRoot(t *testing.T) { +func TestAuthInfoFromCtxWithRootSimple(t *testing.T) { + testAuthInfoFromCtxWithRoot(t, tokenTypeSimple) +} + +func TestAuthInfoFromCtxWithRootJWT(t *testing.T) { + opts := testJWTOpts() + testAuthInfoFromCtxWithRoot(t, opts) +} + +// testAuthInfoFromCtxWithRoot ensures "WithRoot" properly embeds token in the context. +func testAuthInfoFromCtxWithRoot(t *testing.T, opts string) { b, tPath := backend.NewDefaultTmpBackend() defer os.Remove(tPath) - tp, err := NewTokenProvider("simple", dummyIndexWaiter) + tp, err := NewTokenProvider(opts, dummyIndexWaiter) if err != nil { t.Fatal(err) } diff --git a/vendor/github.com/coreos/etcd/clientv3/client.go b/vendor/github.com/coreos/etcd/clientv3/client.go index a2c313d35..713280776 100644 --- a/vendor/github.com/coreos/etcd/clientv3/client.go +++ b/vendor/github.com/coreos/etcd/clientv3/client.go @@ -529,6 +529,20 @@ func isHaltErr(ctx context.Context, err error) bool { return ev.Code() != codes.Unavailable && ev.Code() != codes.Internal } +// isUnavailableErr returns true if the given error is an unavailable error +func isUnavailableErr(ctx context.Context, err error) bool { + if ctx != nil && ctx.Err() != nil { + return false + } + if err == nil { + return false + } + ev, _ := status.FromError(err) + // Unavailable codes mean the system will be right back. + // (e.g., can't connect, lost leader) + return ev.Code() == codes.Unavailable +} + func toErr(ctx context.Context, err error) error { if err == nil { return nil diff --git a/vendor/github.com/coreos/etcd/clientv3/watch.go b/vendor/github.com/coreos/etcd/clientv3/watch.go index f606a1af9..d7633850e 100644 --- a/vendor/github.com/coreos/etcd/clientv3/watch.go +++ b/vendor/github.com/coreos/etcd/clientv3/watch.go @@ -769,10 +769,13 @@ func (w *watchGrpcStream) joinSubstreams() { } } +var maxBackoff = 100 * time.Millisecond + // openWatchClient retries opening a watch client until success or halt. // manually retry in case "ws==nil && err==nil" // TODO: remove FailFast=false func (w *watchGrpcStream) openWatchClient() (ws pb.Watch_WatchClient, err error) { + backoff := time.Millisecond for { select { case <-w.ctx.Done(): @@ -788,6 +791,17 @@ func (w *watchGrpcStream) openWatchClient() (ws pb.Watch_WatchClient, err error) if isHaltErr(w.ctx, err) { return nil, v3rpc.Error(err) } + if isUnavailableErr(w.ctx, err) { + // retry, but backoff + if backoff < maxBackoff { + // 25% backoff factor + backoff = backoff + backoff/4 + if backoff > maxBackoff { + backoff = maxBackoff + } + } + time.Sleep(backoff) + } } return ws, nil } diff --git a/vendor/github.com/coreos/etcd/e2e/cluster_test.go b/vendor/github.com/coreos/etcd/e2e/cluster_test.go index 69896dd48..cfdb8c80a 100644 --- a/vendor/github.com/coreos/etcd/e2e/cluster_test.go +++ b/vendor/github.com/coreos/etcd/e2e/cluster_test.go @@ -112,6 +112,8 @@ type etcdProcessClusterConfig struct { isClientAutoTLS bool isClientCRL bool + cipherSuites []string + forceNewCluster bool initialToken string quotaBackendBytes int64 @@ -296,6 +298,10 @@ func (cfg *etcdProcessClusterConfig) tlsArgs() (args []string) { args = append(args, "--client-crl-file", crlPath, "--client-cert-auth") } + if len(cfg.cipherSuites) > 0 { + args = append(args, "--cipher-suites", strings.Join(cfg.cipherSuites, ",")) + } + return args } diff --git a/vendor/github.com/coreos/etcd/e2e/ctl_v3_move_leader_test.go b/vendor/github.com/coreos/etcd/e2e/ctl_v3_move_leader_test.go index eb2afda5f..0a7f168b2 100644 --- a/vendor/github.com/coreos/etcd/e2e/ctl_v3_move_leader_test.go +++ b/vendor/github.com/coreos/etcd/e2e/ctl_v3_move_leader_test.go @@ -16,6 +16,7 @@ package e2e import ( "context" + "crypto/tls" "fmt" "os" "testing" @@ -23,19 +24,42 @@ import ( "github.com/coreos/etcd/clientv3" "github.com/coreos/etcd/pkg/testutil" + "github.com/coreos/etcd/pkg/transport" "github.com/coreos/etcd/pkg/types" ) -func TestCtlV3MoveLeader(t *testing.T) { +func TestCtlV3MoveLeaderSecure(t *testing.T) { + testCtlV3MoveLeader(t, configTLS) +} + +func TestCtlV3MoveLeaderInsecure(t *testing.T) { + testCtlV3MoveLeader(t, configNoTLS) +} + +func testCtlV3MoveLeader(t *testing.T, cfg etcdProcessClusterConfig) { defer testutil.AfterTest(t) - epc := setupEtcdctlTest(t, &configNoTLS, true) + epc := setupEtcdctlTest(t, &cfg, true) defer func() { if errC := epc.Close(); errC != nil { t.Fatalf("error closing etcd processes (%v)", errC) } }() + var tcfg *tls.Config + if cfg.clientTLS == clientTLS { + tinfo := transport.TLSInfo{ + CertFile: certPath, + KeyFile: privateKeyPath, + TrustedCAFile: caPath, + } + var err error + tcfg, err = tinfo.ClientConfig() + if err != nil { + t.Fatal(err) + } + } + var leadIdx int var leaderID uint64 var transferee uint64 @@ -43,6 +67,7 @@ func TestCtlV3MoveLeader(t *testing.T) { cli, err := clientv3.New(clientv3.Config{ Endpoints: []string{ep}, DialTimeout: 3 * time.Second, + TLS: tcfg, }) if err != nil { t.Fatal(err) diff --git a/vendor/github.com/coreos/etcd/e2e/v2_curl_test.go b/vendor/github.com/coreos/etcd/e2e/v2_curl_test.go index 9f350a3e3..fa6a78c02 100644 --- a/vendor/github.com/coreos/etcd/e2e/v2_curl_test.go +++ b/vendor/github.com/coreos/etcd/e2e/v2_curl_test.go @@ -127,6 +127,8 @@ type cURLReq struct { header string metricsURLScheme string + + ciphers string } // cURLPrefixArgs builds the beginning of a curl command for a given key @@ -165,6 +167,10 @@ func cURLPrefixArgs(clus *etcdProcessCluster, method string, req cURLReq) []stri cmdArgs = append(cmdArgs, "-H", req.header) } + if req.ciphers != "" { + cmdArgs = append(cmdArgs, "--ciphers", req.ciphers) + } + switch method { case "POST", "PUT": dt := req.value diff --git a/vendor/github.com/coreos/etcd/e2e/v3_curl_test.go b/vendor/github.com/coreos/etcd/e2e/v3_curl_test.go index 778860955..2c8aec355 100644 --- a/vendor/github.com/coreos/etcd/e2e/v3_curl_test.go +++ b/vendor/github.com/coreos/etcd/e2e/v3_curl_test.go @@ -17,6 +17,7 @@ package e2e import ( "encoding/base64" "encoding/json" + "fmt" "path" "strconv" "testing" @@ -24,6 +25,7 @@ import ( epb "github.com/coreos/etcd/etcdserver/api/v3election/v3electionpb" pb "github.com/coreos/etcd/etcdserver/etcdserverpb" "github.com/coreos/etcd/pkg/testutil" + "github.com/coreos/etcd/version" "github.com/grpc-ecosystem/grpc-gateway/runtime" ) @@ -390,3 +392,45 @@ type campaignResponse struct { Lease string `json:"lease,omitempty"` } `json:"leader,omitempty"` } + +func TestV3CurlCipherSuitesValid(t *testing.T) { testV3CurlCipherSuites(t, true) } +func TestV3CurlCipherSuitesMismatch(t *testing.T) { testV3CurlCipherSuites(t, false) } +func testV3CurlCipherSuites(t *testing.T, valid bool) { + cc := configClientTLS + cc.clusterSize = 1 + cc.cipherSuites = []string{ + "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256", + "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384", + "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256", + "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384", + "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305", + "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305", + } + testFunc := cipherSuiteTestValid + if !valid { + testFunc = cipherSuiteTestMismatch + } + testCtl(t, testFunc, withCfg(cc)) +} + +func cipherSuiteTestValid(cx ctlCtx) { + if err := cURLGet(cx.epc, cURLReq{ + endpoint: "/metrics", + expected: fmt.Sprintf(`etcd_server_version{server_version="%s"} 1`, version.Version), + metricsURLScheme: cx.cfg.metricsURLScheme, + ciphers: "ECDHE-RSA-AES128-GCM-SHA256", // TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 + }); err != nil { + cx.t.Fatalf("failed get with curl (%v)", err) + } +} + +func cipherSuiteTestMismatch(cx ctlCtx) { + if err := cURLGet(cx.epc, cURLReq{ + endpoint: "/metrics", + expected: "alert handshake failure", + metricsURLScheme: cx.cfg.metricsURLScheme, + ciphers: "ECDHE-RSA-DES-CBC3-SHA", // TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA + }); err != nil { + cx.t.Fatalf("failed get with curl (%v)", err) + } +} diff --git a/vendor/github.com/coreos/etcd/embed/config.go b/vendor/github.com/coreos/etcd/embed/config.go index 888bcb0b2..835e051a7 100644 --- a/vendor/github.com/coreos/etcd/embed/config.go +++ b/vendor/github.com/coreos/etcd/embed/config.go @@ -31,6 +31,7 @@ import ( "github.com/coreos/etcd/pkg/cors" "github.com/coreos/etcd/pkg/netutil" "github.com/coreos/etcd/pkg/srv" + "github.com/coreos/etcd/pkg/tlsutil" "github.com/coreos/etcd/pkg/transport" "github.com/coreos/etcd/pkg/types" @@ -183,6 +184,11 @@ type Config struct { PeerTLSInfo transport.TLSInfo PeerAutoTLS bool + // CipherSuites is a list of supported TLS cipher suites between + // client/server and peers. If empty, Go auto-populates the list. + // Note that cipher suites are prioritized in the given order. + CipherSuites []string `json:"cipher-suites"` + // debug Debug bool `json:"debug"` @@ -426,6 +432,24 @@ func (cfg *configYAML) configFromFile(path string) error { return cfg.Validate() } +func updateCipherSuites(tls *transport.TLSInfo, ss []string) error { + if len(tls.CipherSuites) > 0 && len(ss) > 0 { + return fmt.Errorf("TLSInfo.CipherSuites is already specified (given %v)", ss) + } + if len(ss) > 0 { + cs := make([]uint16, len(ss)) + for i, s := range ss { + var ok bool + cs[i], ok = tlsutil.GetCipherSuite(s) + if !ok { + return fmt.Errorf("unexpected TLS cipher suite %q", s) + } + } + tls.CipherSuites = cs + } + return nil +} + // Validate ensures that '*embed.Config' fields are properly configured. func (cfg *Config) Validate() error { if err := checkBindURLs(cfg.LPUrls); err != nil { @@ -562,31 +586,41 @@ func (cfg Config) defaultClientHost() bool { } func (cfg *Config) ClientSelfCert() (err error) { - if cfg.ClientAutoTLS && cfg.ClientTLSInfo.Empty() { - chosts := make([]string, len(cfg.LCUrls)) - for i, u := range cfg.LCUrls { - chosts[i] = u.Host - } - cfg.ClientTLSInfo, err = transport.SelfCert(filepath.Join(cfg.Dir, "fixtures", "client"), chosts) - return err - } else if cfg.ClientAutoTLS { - plog.Warningf("ignoring client auto TLS since certs given") + if !cfg.ClientAutoTLS { + return nil } - return nil + if !cfg.ClientTLSInfo.Empty() { + plog.Warningf("ignoring client auto TLS since certs given") + return nil + } + chosts := make([]string, len(cfg.LCUrls)) + for i, u := range cfg.LCUrls { + chosts[i] = u.Host + } + cfg.ClientTLSInfo, err = transport.SelfCert(filepath.Join(cfg.Dir, "fixtures", "client"), chosts) + if err != nil { + return err + } + return updateCipherSuites(&cfg.ClientTLSInfo, cfg.CipherSuites) } func (cfg *Config) PeerSelfCert() (err error) { - if cfg.PeerAutoTLS && cfg.PeerTLSInfo.Empty() { - phosts := make([]string, len(cfg.LPUrls)) - for i, u := range cfg.LPUrls { - phosts[i] = u.Host - } - cfg.PeerTLSInfo, err = transport.SelfCert(filepath.Join(cfg.Dir, "fixtures", "peer"), phosts) - return err - } else if cfg.PeerAutoTLS { - plog.Warningf("ignoring peer auto TLS since certs given") + if !cfg.PeerAutoTLS { + return nil } - return nil + if !cfg.PeerTLSInfo.Empty() { + plog.Warningf("ignoring peer auto TLS since certs given") + return nil + } + phosts := make([]string, len(cfg.LPUrls)) + for i, u := range cfg.LPUrls { + phosts[i] = u.Host + } + cfg.PeerTLSInfo, err = transport.SelfCert(filepath.Join(cfg.Dir, "fixtures", "peer"), phosts) + if err != nil { + return err + } + return updateCipherSuites(&cfg.PeerTLSInfo, cfg.CipherSuites) } // UpdateDefaultClusterFromName updates cluster advertise URLs with, if available, default host, diff --git a/vendor/github.com/coreos/etcd/embed/etcd.go b/vendor/github.com/coreos/etcd/embed/etcd.go index e2a10d4cb..bd848a713 100644 --- a/vendor/github.com/coreos/etcd/embed/etcd.go +++ b/vendor/github.com/coreos/etcd/embed/etcd.go @@ -42,7 +42,7 @@ import ( "github.com/coreos/etcd/rafthttp" "github.com/coreos/pkg/capnslog" - "github.com/grpc-ecosystem/go-grpc-prometheus" + grpc_prometheus "github.com/grpc-ecosystem/go-grpc-prometheus" "github.com/soheilhy/cmux" "google.golang.org/grpc" "google.golang.org/grpc/keepalive" @@ -302,6 +302,9 @@ func stopServers(ctx context.Context, ss *servers) { func (e *Etcd) Err() <-chan error { return e.errc } func startPeerListeners(cfg *Config) (peers []*peerListener, err error) { + if err = updateCipherSuites(&cfg.PeerTLSInfo, cfg.CipherSuites); err != nil { + return nil, err + } if err = cfg.PeerSelfCert(); err != nil { plog.Fatalf("could not get certs (%v)", err) } @@ -387,6 +390,9 @@ func (e *Etcd) servePeers() (err error) { } func startClientListeners(cfg *Config) (sctxs map[string]*serveCtx, err error) { + if err = updateCipherSuites(&cfg.ClientTLSInfo, cfg.CipherSuites); err != nil { + return nil, err + } if err = cfg.ClientSelfCert(); err != nil { plog.Fatalf("could not get certs (%v)", err) } diff --git a/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/global.go b/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/global.go index e52442ff8..6b2f05e3e 100644 --- a/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/global.go +++ b/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/global.go @@ -109,11 +109,12 @@ func (*discardValue) Type() string { return "" } func clientConfigFromCmd(cmd *cobra.Command) *clientConfig { fs := cmd.InheritedFlags() - - // silence "pkg/flags: unrecognized environment variable ETCDCTL_WATCH_KEY=foo" warnings - // silence "pkg/flags: unrecognized environment variable ETCDCTL_WATCH_RANGE_END=bar" warnings - fs.AddFlag(&pflag.Flag{Name: "watch-key", Value: &discardValue{}}) - fs.AddFlag(&pflag.Flag{Name: "watch-range-end", Value: &discardValue{}}) + if strings.HasPrefix(cmd.Use, "watch") { + // silence "pkg/flags: unrecognized environment variable ETCDCTL_WATCH_KEY=foo" warnings + // silence "pkg/flags: unrecognized environment variable ETCDCTL_WATCH_RANGE_END=bar" warnings + fs.AddFlag(&pflag.Flag{Name: "watch-key", Value: &discardValue{}}) + fs.AddFlag(&pflag.Flag{Name: "watch-range-end", Value: &discardValue{}}) + } flags.SetPflagsFromEnv("ETCDCTL", fs) debug, err := cmd.Flags().GetBool("debug") diff --git a/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/move_leader_command.go b/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/move_leader_command.go index 3bd44255e..5fdecf95a 100644 --- a/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/move_leader_command.go +++ b/vendor/github.com/coreos/etcd/etcdctl/ctlv3/command/move_leader_command.go @@ -17,7 +17,6 @@ package command import ( "fmt" "strconv" - "time" "github.com/coreos/etcd/clientv3" "github.com/spf13/cobra" @@ -53,16 +52,12 @@ func transferLeadershipCommandFunc(cmd *cobra.Command, args []string) { var leaderCli *clientv3.Client var leaderID uint64 for _, ep := range eps { - cli, err := clientv3.New(clientv3.Config{ - Endpoints: []string{ep}, - DialTimeout: 3 * time.Second, - }) - if err != nil { - ExitWithError(ExitError, err) - } - resp, err := cli.Status(ctx, ep) - if err != nil { - ExitWithError(ExitError, err) + cfg := clientConfigFromCmd(cmd) + cfg.endpoints = []string{ep} + cli := cfg.mustClient() + resp, serr := cli.Status(ctx, ep) + if serr != nil { + ExitWithError(ExitError, serr) } if resp.Header.GetMemberId() == resp.Leader { diff --git a/vendor/github.com/coreos/etcd/etcdmain/config.go b/vendor/github.com/coreos/etcd/etcdmain/config.go index 3028c6581..2a5faa746 100644 --- a/vendor/github.com/coreos/etcd/etcdmain/config.go +++ b/vendor/github.com/coreos/etcd/etcdmain/config.go @@ -190,6 +190,8 @@ func newConfig() *config { fs.StringVar(&cfg.ec.PeerTLSInfo.CRLFile, "peer-crl-file", "", "Path to the peer certificate revocation list file.") fs.StringVar(&cfg.ec.PeerTLSInfo.AllowedCN, "peer-cert-allowed-cn", "", "Allowed CN for inter peer authentication.") + fs.Var(flags.NewStringsValueV2(""), "cipher-suites", "Comma-separated list of supported TLS cipher suites between client/server and peers (empty will be auto-populated by Go).") + // logging fs.BoolVar(&cfg.ec.Debug, "debug", false, "Enable debug-level logging for etcd.") fs.StringVar(&cfg.ec.LogPkgLevels, "log-package-levels", "", "Specify a particular log level for each etcd package (eg: 'etcdmain=CRITICAL,etcdserver=DEBUG').") @@ -275,6 +277,8 @@ func (cfg *config) configFromCmdLine() error { cfg.ec.ListenMetricsUrls = []url.URL(u) } + cfg.ec.CipherSuites = flags.StringsFromFlagV2(cfg.cf.flagSet, "cipher-suites") + cfg.ec.ClusterState = cfg.cf.clusterState.String() cfg.cp.Fallback = cfg.cf.fallback.String() cfg.cp.Proxy = cfg.cf.proxy.String() diff --git a/vendor/github.com/coreos/etcd/etcdmain/help.go b/vendor/github.com/coreos/etcd/etcdmain/help.go index 82d270460..c64dab3bc 100644 --- a/vendor/github.com/coreos/etcd/etcdmain/help.go +++ b/vendor/github.com/coreos/etcd/etcdmain/help.go @@ -160,6 +160,8 @@ security flags: peer TLS using self-generated certificates if --peer-key-file and --peer-cert-file are not provided. --peer-crl-file '' path to the peer certificate revocation list file. + --cipher-suites '' + comma-separated list of supported TLS cipher suites between client/server and peers (empty will be auto-populated by Go). logging flags diff --git a/vendor/github.com/coreos/etcd/etcdserver/apply.go b/vendor/github.com/coreos/etcd/etcdserver/apply.go index 5be2ec0e7..93e78e390 100644 --- a/vendor/github.com/coreos/etcd/etcdserver/apply.go +++ b/vendor/github.com/coreos/etcd/etcdserver/apply.go @@ -107,9 +107,10 @@ func (s *EtcdServer) newApplierV3() applierV3 { } func (a *applierV3backend) Apply(r *pb.InternalRaftRequest) *applyResult { - defer warnOfExpensiveRequest(time.Now(), r) - ar := &applyResult{} + defer func(start time.Time) { + warnOfExpensiveRequest(start, &pb.InternalRaftStringer{Request: r}, ar.resp, ar.err) + }(time.Now()) // call into a.s.applyV3.F instead of a.F so upper appliers can check individual calls switch { diff --git a/vendor/github.com/coreos/etcd/etcdserver/apply_v2.go b/vendor/github.com/coreos/etcd/etcdserver/apply_v2.go index 00597261a..a49b6823d 100644 --- a/vendor/github.com/coreos/etcd/etcdserver/apply_v2.go +++ b/vendor/github.com/coreos/etcd/etcdserver/apply_v2.go @@ -107,7 +107,7 @@ func (a *applierV2store) Sync(r *RequestV2) Response { // applyV2Request interprets r as a call to store.X and returns a Response interpreted // from store.Event func (s *EtcdServer) applyV2Request(r *RequestV2) Response { - defer warnOfExpensiveRequest(time.Now(), r) + defer warnOfExpensiveRequest(time.Now(), r, nil, nil) switch r.Method { case "POST": diff --git a/vendor/github.com/coreos/etcd/etcdserver/etcdserverpb/raft_internal_stringer.go b/vendor/github.com/coreos/etcd/etcdserver/etcdserverpb/raft_internal_stringer.go new file mode 100644 index 000000000..ec6b6397b --- /dev/null +++ b/vendor/github.com/coreos/etcd/etcdserver/etcdserverpb/raft_internal_stringer.go @@ -0,0 +1,183 @@ +// Copyright 2018 The etcd Authors +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package etcdserverpb + +import ( + "fmt" + "strings" + + proto "github.com/golang/protobuf/proto" +) + +// InternalRaftStringer implements custom proto Stringer: +// redact password, replace value fields with value_size fields. +type InternalRaftStringer struct { + Request *InternalRaftRequest +} + +func (as *InternalRaftStringer) String() string { + switch { + case as.Request.LeaseGrant != nil: + return fmt.Sprintf("header:<%s> lease_grant:", + as.Request.Header.String(), + as.Request.LeaseGrant.TTL, + as.Request.LeaseGrant.ID, + ) + case as.Request.LeaseRevoke != nil: + return fmt.Sprintf("header:<%s> lease_revoke:", + as.Request.Header.String(), + as.Request.LeaseRevoke.ID, + ) + case as.Request.Authenticate != nil: + return fmt.Sprintf("header:<%s> authenticate:", + as.Request.Header.String(), + as.Request.Authenticate.Name, + as.Request.Authenticate.SimpleToken, + ) + case as.Request.AuthUserAdd != nil: + return fmt.Sprintf("header:<%s> auth_user_add:", + as.Request.Header.String(), + as.Request.AuthUserAdd.Name, + ) + case as.Request.AuthUserChangePassword != nil: + return fmt.Sprintf("header:<%s> auth_user_change_password:", + as.Request.Header.String(), + as.Request.AuthUserChangePassword.Name, + ) + case as.Request.Put != nil: + return fmt.Sprintf("header:<%s> put:<%s>", + as.Request.Header.String(), + newLoggablePutRequest(as.Request.Put).String(), + ) + case as.Request.Txn != nil: + return fmt.Sprintf("header:<%s> txn:<%s>", + as.Request.Header.String(), + NewLoggableTxnRequest(as.Request.Txn).String(), + ) + default: + // nothing to redact + } + return as.Request.String() +} + +// txnRequestStringer implements a custom proto String to replace value bytes fields with value size +// fields in any nested txn and put operations. +type txnRequestStringer struct { + Request *TxnRequest +} + +func NewLoggableTxnRequest(request *TxnRequest) *txnRequestStringer { + return &txnRequestStringer{request} +} + +func (as *txnRequestStringer) String() string { + var compare []string + for _, c := range as.Request.Compare { + switch cv := c.TargetUnion.(type) { + case *Compare_Value: + compare = append(compare, newLoggableValueCompare(c, cv).String()) + default: + // nothing to redact + compare = append(compare, c.String()) + } + } + var success []string + for _, s := range as.Request.Success { + success = append(success, newLoggableRequestOp(s).String()) + } + var failure []string + for _, f := range as.Request.Failure { + failure = append(failure, newLoggableRequestOp(f).String()) + } + return fmt.Sprintf("compare:<%s> success:<%s> failure:<%s>", + strings.Join(compare, " "), + strings.Join(success, " "), + strings.Join(failure, " "), + ) +} + +// requestOpStringer implements a custom proto String to replace value bytes fields with value +// size fields in any nested txn and put operations. +type requestOpStringer struct { + Op *RequestOp +} + +func newLoggableRequestOp(op *RequestOp) *requestOpStringer { + return &requestOpStringer{op} +} + +func (as *requestOpStringer) String() string { + switch op := as.Op.Request.(type) { + case *RequestOp_RequestPut: + return fmt.Sprintf("request_put:<%s>", newLoggablePutRequest(op.RequestPut).String()) + case *RequestOp_RequestTxn: + return fmt.Sprintf("request_txn:<%s>", NewLoggableTxnRequest(op.RequestTxn).String()) + default: + // nothing to redact + } + return as.Op.String() +} + +// loggableValueCompare implements a custom proto String for Compare.Value union member types to +// replace the value bytes field with a value size field. +// To preserve proto encoding of the key and range_end bytes, a faked out proto type is used here. +type loggableValueCompare struct { + Result Compare_CompareResult `protobuf:"varint,1,opt,name=result,proto3,enum=etcdserverpb.Compare_CompareResult"` + Target Compare_CompareTarget `protobuf:"varint,2,opt,name=target,proto3,enum=etcdserverpb.Compare_CompareTarget"` + Key []byte `protobuf:"bytes,3,opt,name=key,proto3"` + ValueSize int `protobuf:"bytes,7,opt,name=value_size,proto3"` + RangeEnd []byte `protobuf:"bytes,64,opt,name=range_end,proto3"` +} + +func newLoggableValueCompare(c *Compare, cv *Compare_Value) *loggableValueCompare { + return &loggableValueCompare{ + c.Result, + c.Target, + c.Key, + len(cv.Value), + c.RangeEnd, + } +} + +func (m *loggableValueCompare) Reset() { *m = loggableValueCompare{} } +func (m *loggableValueCompare) String() string { return proto.CompactTextString(m) } +func (*loggableValueCompare) ProtoMessage() {} + +// loggablePutRequest implements a custom proto String to replace value bytes field with a value +// size field. +// To preserve proto encoding of the key bytes, a faked out proto type is used here. +type loggablePutRequest struct { + Key []byte `protobuf:"bytes,1,opt,name=key,proto3"` + ValueSize int `protobuf:"varint,2,opt,name=value_size,proto3"` + Lease int64 `protobuf:"varint,3,opt,name=lease,proto3"` + PrevKv bool `protobuf:"varint,4,opt,name=prev_kv,proto3"` + IgnoreValue bool `protobuf:"varint,5,opt,name=ignore_value,proto3"` + IgnoreLease bool `protobuf:"varint,6,opt,name=ignore_lease,proto3"` +} + +func newLoggablePutRequest(request *PutRequest) *loggablePutRequest { + return &loggablePutRequest{ + request.Key, + len(request.Value), + request.Lease, + request.PrevKv, + request.IgnoreValue, + request.IgnoreLease, + } +} + +func (m *loggablePutRequest) Reset() { *m = loggablePutRequest{} } +func (m *loggablePutRequest) String() string { return proto.CompactTextString(m) } +func (*loggablePutRequest) ProtoMessage() {} diff --git a/vendor/github.com/coreos/etcd/etcdserver/util.go b/vendor/github.com/coreos/etcd/etcdserver/util.go index f1ba2dbec..708fee4cb 100644 --- a/vendor/github.com/coreos/etcd/etcdserver/util.go +++ b/vendor/github.com/coreos/etcd/etcdserver/util.go @@ -16,11 +16,15 @@ package etcdserver import ( "fmt" + "reflect" + "strings" "time" + pb "github.com/coreos/etcd/etcdserver/etcdserverpb" "github.com/coreos/etcd/etcdserver/membership" "github.com/coreos/etcd/pkg/types" "github.com/coreos/etcd/rafthttp" + "github.com/golang/protobuf/proto" ) // isConnectedToQuorumSince checks whether the local member is connected to the @@ -97,18 +101,54 @@ func (nc *notifier) notify(err error) { close(nc.c) } -func warnOfExpensiveRequest(now time.Time, stringer fmt.Stringer) { - warnOfExpensiveGenericRequest(now, stringer, "") +func warnOfExpensiveRequest(now time.Time, reqStringer fmt.Stringer, respMsg proto.Message, err error) { + var resp string + if !isNil(respMsg) { + resp = fmt.Sprintf("size:%d", proto.Size(respMsg)) + } + warnOfExpensiveGenericRequest(now, reqStringer, "", resp, err) } -func warnOfExpensiveReadOnlyRangeRequest(now time.Time, stringer fmt.Stringer) { - warnOfExpensiveGenericRequest(now, stringer, "read-only range ") +func warnOfExpensiveReadOnlyTxnRequest(now time.Time, r *pb.TxnRequest, txnResponse *pb.TxnResponse, err error) { + reqStringer := pb.NewLoggableTxnRequest(r) + var resp string + if !isNil(txnResponse) { + var resps []string + for _, r := range txnResponse.Responses { + switch op := r.Response.(type) { + case *pb.ResponseOp_ResponseRange: + resps = append(resps, fmt.Sprintf("range_response_count:%d", len(op.ResponseRange.Kvs))) + default: + // only range responses should be in a read only txn request + } + } + resp = fmt.Sprintf("responses:<%s> size:%d", strings.Join(resps, " "), proto.Size(txnResponse)) + } + warnOfExpensiveGenericRequest(now, reqStringer, "read-only range ", resp, err) } -func warnOfExpensiveGenericRequest(now time.Time, stringer fmt.Stringer, prefix string) { +func warnOfExpensiveReadOnlyRangeRequest(now time.Time, reqStringer fmt.Stringer, rangeResponse *pb.RangeResponse, err error) { + var resp string + if !isNil(rangeResponse) { + resp = fmt.Sprintf("range_response_count:%d size:%d", len(rangeResponse.Kvs), proto.Size(rangeResponse)) + } + warnOfExpensiveGenericRequest(now, reqStringer, "read-only range ", resp, err) +} + +func warnOfExpensiveGenericRequest(now time.Time, reqStringer fmt.Stringer, prefix string, resp string, err error) { // TODO: add metrics d := time.Since(now) if d > warnApplyDuration { - plog.Warningf("%srequest %q took too long (%v) to execute", prefix, stringer.String(), d) + var result string + if err != nil { + result = fmt.Sprintf("error:%v", err) + } else { + result = resp + } + plog.Warningf("%srequest %q with result %q took too long (%v) to execute", prefix, reqStringer.String(), result, d) } } + +func isNil(msg proto.Message) bool { + return msg == nil || reflect.ValueOf(msg).IsNil() +} diff --git a/vendor/github.com/coreos/etcd/etcdserver/v3_server.go b/vendor/github.com/coreos/etcd/etcdserver/v3_server.go index b3f773274..773dc95ba 100644 --- a/vendor/github.com/coreos/etcd/etcdserver/v3_server.go +++ b/vendor/github.com/coreos/etcd/etcdserver/v3_server.go @@ -84,23 +84,26 @@ type Authenticator interface { } func (s *EtcdServer) Range(ctx context.Context, r *pb.RangeRequest) (*pb.RangeResponse, error) { - defer warnOfExpensiveReadOnlyRangeRequest(time.Now(), r) + var resp *pb.RangeResponse + var err error + defer func(start time.Time) { + warnOfExpensiveReadOnlyRangeRequest(start, r, resp, err) + }(time.Now()) if !r.Serializable { - err := s.linearizableReadNotify(ctx) + err = s.linearizableReadNotify(ctx) if err != nil { return nil, err } } - var resp *pb.RangeResponse - var err error chk := func(ai *auth.AuthInfo) error { return s.authStore.IsRangePermitted(ai, r.Key, r.RangeEnd) } get := func() { resp, err = s.applyV3Base.Range(nil, r) } if serr := s.doSerialize(ctx, chk, get); serr != nil { - return nil, serr + err = serr + return nil, err } return resp, err } @@ -135,7 +138,9 @@ func (s *EtcdServer) Txn(ctx context.Context, r *pb.TxnRequest) (*pb.TxnResponse return checkTxnAuth(s.authStore, ai, r) } - defer warnOfExpensiveReadOnlyRangeRequest(time.Now(), r) + defer func(start time.Time) { + warnOfExpensiveReadOnlyTxnRequest(start, r, resp, err) + }(time.Now()) get := func() { resp, err = s.applyV3Base.Txn(r) } if serr := s.doSerialize(ctx, chk, get); serr != nil { diff --git a/vendor/github.com/coreos/etcd/integration/v3_tls_test.go b/vendor/github.com/coreos/etcd/integration/v3_tls_test.go new file mode 100644 index 000000000..5ebe7ad35 --- /dev/null +++ b/vendor/github.com/coreos/etcd/integration/v3_tls_test.go @@ -0,0 +1,71 @@ +// Copyright 2018 The etcd Authors +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package integration + +import ( + "context" + "crypto/tls" + "testing" + "time" + + "github.com/coreos/etcd/clientv3" + "github.com/coreos/etcd/pkg/testutil" +) + +func TestTLSClientCipherSuitesValid(t *testing.T) { testTLSCipherSuites(t, true) } +func TestTLSClientCipherSuitesMismatch(t *testing.T) { testTLSCipherSuites(t, false) } + +// testTLSCipherSuites ensures mismatching client-side cipher suite +// fail TLS handshake with the server. +func testTLSCipherSuites(t *testing.T, valid bool) { + defer testutil.AfterTest(t) + + cipherSuites := []uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305, + tls.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305, + } + srvTLS, cliTLS := testTLSInfo, testTLSInfo + if valid { + srvTLS.CipherSuites, cliTLS.CipherSuites = cipherSuites, cipherSuites + } else { + srvTLS.CipherSuites, cliTLS.CipherSuites = cipherSuites[:2], cipherSuites[2:] + } + + clus := NewClusterV3(t, &ClusterConfig{Size: 1, ClientTLS: &srvTLS}) + defer clus.Terminate(t) + + cc, err := cliTLS.ClientConfig() + if err != nil { + t.Fatal(err) + } + cli, cerr := clientv3.New(clientv3.Config{ + Endpoints: []string{clus.Members[0].GRPCAddr()}, + DialTimeout: time.Second, + TLS: cc, + }) + if cli != nil { + cli.Close() + } + if !valid && cerr != context.DeadlineExceeded { + t.Fatalf("expected %v with TLS handshake failure, got %v", context.DeadlineExceeded, cerr) + } + if valid && cerr != nil { + t.Fatalf("expected TLS handshake success, got %v", cerr) + } +} diff --git a/vendor/github.com/coreos/etcd/mvcc/watchable_store.go b/vendor/github.com/coreos/etcd/mvcc/watchable_store.go index 14bb14ce2..78df19326 100644 --- a/vendor/github.com/coreos/etcd/mvcc/watchable_store.go +++ b/vendor/github.com/coreos/etcd/mvcc/watchable_store.go @@ -192,6 +192,7 @@ func (s *watchableStore) Restore(b backend.Backend) error { } for wa := range s.synced.watchers { + wa.restore = true s.unsynced.add(wa) } s.synced = newWatcherGroup() @@ -482,6 +483,14 @@ type watcher struct { // compacted is set when the watcher is removed because of compaction compacted bool + // restore is true when the watcher is being restored from leader snapshot + // which means that this watcher has just been moved from "synced" to "unsynced" + // watcher group, possibly with a future revision when it was first added + // to the synced watcher + // "unsynced" watcher revision must always be <= current revision, + // except when the watcher were to be moved from "synced" watcher group + restore bool + // minRev is the minimum revision update the watcher will accept minRev int64 id WatchID diff --git a/vendor/github.com/coreos/etcd/mvcc/watchable_store_test.go b/vendor/github.com/coreos/etcd/mvcc/watchable_store_test.go index 84a0b6620..dc96d5336 100644 --- a/vendor/github.com/coreos/etcd/mvcc/watchable_store_test.go +++ b/vendor/github.com/coreos/etcd/mvcc/watchable_store_test.go @@ -338,6 +338,62 @@ func TestWatchRestore(t *testing.T) { t.Run("RunSyncWatchLoopBeforeRestore", test(time.Millisecond*120)) // longer than default waitDuration } +// TestWatchRestoreSyncedWatcher tests such a case that: +// 1. watcher is created with a future revision "math.MaxInt64 - 2" +// 2. watcher with a future revision is added to "synced" watcher group +// 3. restore/overwrite storage with snapshot of a higher lasat revision +// 4. restore operation moves "synced" to "unsynced" watcher group +// 5. choose the watcher from step 1, without panic +func TestWatchRestoreSyncedWatcher(t *testing.T) { + b1, b1Path := backend.NewDefaultTmpBackend() + s1 := newWatchableStore(b1, &lease.FakeLessor{}, nil) + defer cleanup(s1, b1, b1Path) + + b2, b2Path := backend.NewDefaultTmpBackend() + s2 := newWatchableStore(b2, &lease.FakeLessor{}, nil) + defer cleanup(s2, b2, b2Path) + + testKey, testValue := []byte("foo"), []byte("bar") + rev := s1.Put(testKey, testValue, lease.NoLease) + startRev := rev + 2 + + // create a watcher with a future revision + // add to "synced" watcher group (startRev > s.store.currentRev) + w1 := s1.NewWatchStream() + w1.Watch(testKey, nil, startRev) + + // make "s2" ends up with a higher last revision + s2.Put(testKey, testValue, lease.NoLease) + s2.Put(testKey, testValue, lease.NoLease) + + // overwrite storage with higher revisions + if err := s1.Restore(b2); err != nil { + t.Fatal(err) + } + + // wait for next "syncWatchersLoop" iteration + // and the unsynced watcher should be chosen + time.Sleep(2 * time.Second) + + // trigger events for "startRev" + s1.Put(testKey, testValue, lease.NoLease) + + select { + case resp := <-w1.Chan(): + if resp.Revision != startRev { + t.Fatalf("resp.Revision expect %d, got %d", startRev, resp.Revision) + } + if len(resp.Events) != 1 { + t.Fatalf("len(resp.Events) expect 1, got %d", len(resp.Events)) + } + if resp.Events[0].Kv.ModRevision != startRev { + t.Fatalf("resp.Events[0].Kv.ModRevision expect %d, got %d", startRev, resp.Events[0].Kv.ModRevision) + } + case <-time.After(time.Second): + t.Fatal("failed to receive event in 1 second") + } +} + // TestWatchBatchUnsynced tests batching on unsynced watchers func TestWatchBatchUnsynced(t *testing.T) { b, tmpPath := backend.NewDefaultTmpBackend() diff --git a/vendor/github.com/coreos/etcd/mvcc/watcher_group.go b/vendor/github.com/coreos/etcd/mvcc/watcher_group.go index 6ef1d0ce8..b65c7bc5e 100644 --- a/vendor/github.com/coreos/etcd/mvcc/watcher_group.go +++ b/vendor/github.com/coreos/etcd/mvcc/watcher_group.go @@ -15,6 +15,7 @@ package mvcc import ( + "fmt" "math" "github.com/coreos/etcd/mvcc/mvccpb" @@ -238,7 +239,15 @@ func (wg *watcherGroup) chooseAll(curRev, compactRev int64) int64 { minRev := int64(math.MaxInt64) for w := range wg.watchers { if w.minRev > curRev { - panic("watcher current revision should not exceed current revision") + // after network partition, possibly choosing future revision watcher from restore operation + // with watch key "proxy-namespace__lostleader" and revision "math.MaxInt64 - 2" + // do not panic when such watcher had been moved from "synced" watcher during restore operation + if !w.restore { + panic(fmt.Errorf("watcher minimum revision %d should not exceed current revision %d", w.minRev, curRev)) + } + + // mark 'restore' done, since it's chosen + w.restore = false } if w.minRev < compactRev { select { diff --git a/vendor/github.com/coreos/etcd/pkg/flags/strings.go b/vendor/github.com/coreos/etcd/pkg/flags/strings.go index 40ee43253..89bdf9506 100644 --- a/vendor/github.com/coreos/etcd/pkg/flags/strings.go +++ b/vendor/github.com/coreos/etcd/pkg/flags/strings.go @@ -14,7 +14,12 @@ package flags -import "errors" +import ( + "errors" + "flag" + "sort" + "strings" +) // NewStringsFlag creates a new string flag for which any one of the given // strings is a valid value, and any other value is an error. @@ -47,3 +52,34 @@ func (ss *StringsFlag) Set(s string) error { func (ss *StringsFlag) String() string { return ss.val } + +// StringsValueV2 wraps "sort.StringSlice". +type StringsValueV2 sort.StringSlice + +// Set parses a command line set of strings, separated by comma. +// Implements "flag.Value" interface. +func (ss *StringsValueV2) Set(s string) error { + *ss = strings.Split(s, ",") + return nil +} + +// String implements "flag.Value" interface. +func (ss *StringsValueV2) String() string { return strings.Join(*ss, ",") } + +// NewStringsValueV2 implements string slice as "flag.Value" interface. +// Given value is to be separated by comma. +func NewStringsValueV2(s string) (ss *StringsValueV2) { + if s == "" { + return &StringsValueV2{} + } + ss = new(StringsValueV2) + if err := ss.Set(s); err != nil { + plog.Panicf("new StringsValueV2 should never fail: %v", err) + } + return ss +} + +// StringsFromFlagV2 returns a string slice from the flag. +func StringsFromFlagV2(fs *flag.FlagSet, flagName string) []string { + return []string(*fs.Lookup(flagName).Value.(*StringsValueV2)) +} diff --git a/vendor/github.com/coreos/etcd/pkg/tlsutil/cipher_suites.go b/vendor/github.com/coreos/etcd/pkg/tlsutil/cipher_suites.go new file mode 100644 index 000000000..b5916bb54 --- /dev/null +++ b/vendor/github.com/coreos/etcd/pkg/tlsutil/cipher_suites.go @@ -0,0 +1,51 @@ +// Copyright 2018 The etcd Authors +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package tlsutil + +import "crypto/tls" + +// cipher suites implemented by Go +// https://github.com/golang/go/blob/dev.boringcrypto.go1.10/src/crypto/tls/cipher_suites.go +var cipherSuites = map[string]uint16{ + "TLS_RSA_WITH_RC4_128_SHA": tls.TLS_RSA_WITH_RC4_128_SHA, + "TLS_RSA_WITH_3DES_EDE_CBC_SHA": tls.TLS_RSA_WITH_3DES_EDE_CBC_SHA, + "TLS_RSA_WITH_AES_128_CBC_SHA": tls.TLS_RSA_WITH_AES_128_CBC_SHA, + "TLS_RSA_WITH_AES_256_CBC_SHA": tls.TLS_RSA_WITH_AES_256_CBC_SHA, + "TLS_RSA_WITH_AES_128_CBC_SHA256": tls.TLS_RSA_WITH_AES_128_CBC_SHA256, + "TLS_RSA_WITH_AES_128_GCM_SHA256": tls.TLS_RSA_WITH_AES_128_GCM_SHA256, + "TLS_RSA_WITH_AES_256_GCM_SHA384": tls.TLS_RSA_WITH_AES_256_GCM_SHA384, + "TLS_ECDHE_ECDSA_WITH_RC4_128_SHA": tls.TLS_ECDHE_ECDSA_WITH_RC4_128_SHA, + "TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA": tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA, + "TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA": tls.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA, + "TLS_ECDHE_RSA_WITH_RC4_128_SHA": tls.TLS_ECDHE_RSA_WITH_RC4_128_SHA, + "TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA": tls.TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA, + "TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA": tls.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA, + "TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA": tls.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA, + "TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256": tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256, + "TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256": tls.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256, + "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256": tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, + "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256": tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384": tls.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, + "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384": tls.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, + "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305": tls.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305, + "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305": tls.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305, +} + +// GetCipherSuite returns the corresponding cipher suite, +// and boolean value if it is supported. +func GetCipherSuite(s string) (uint16, bool) { + v, ok := cipherSuites[s] + return v, ok +} diff --git a/vendor/github.com/coreos/etcd/pkg/tlsutil/cipher_suites_test.go b/vendor/github.com/coreos/etcd/pkg/tlsutil/cipher_suites_test.go new file mode 100644 index 000000000..ff6d97ffe --- /dev/null +++ b/vendor/github.com/coreos/etcd/pkg/tlsutil/cipher_suites_test.go @@ -0,0 +1,42 @@ +// Copyright 2018 The etcd Authors +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package tlsutil + +import ( + "go/importer" + "reflect" + "strings" + "testing" +) + +func TestGetCipherSuites(t *testing.T) { + pkg, err := importer.For("source", nil).Import("crypto/tls") + if err != nil { + t.Fatal(err) + } + cm := make(map[string]uint16) + for _, s := range pkg.Scope().Names() { + if strings.HasPrefix(s, "TLS_RSA_") || strings.HasPrefix(s, "TLS_ECDHE_") { + v, ok := GetCipherSuite(s) + if !ok { + t.Fatalf("Go implements missing cipher suite %q (%v)", s, v) + } + cm[s] = v + } + } + if !reflect.DeepEqual(cm, cipherSuites) { + t.Fatalf("found unmatched cipher suites %v (Go) != %v", cm, cipherSuites) + } +} diff --git a/vendor/github.com/coreos/etcd/pkg/transport/listener.go b/vendor/github.com/coreos/etcd/pkg/transport/listener.go index 1df18dbb8..48655063f 100644 --- a/vendor/github.com/coreos/etcd/pkg/transport/listener.go +++ b/vendor/github.com/coreos/etcd/pkg/transport/listener.go @@ -72,6 +72,11 @@ type TLSInfo struct { // connection will be closed immediately afterwards. HandshakeFailure func(*tls.Conn, error) + // CipherSuites is a list of supported cipher suites. + // If empty, Go auto-populates it by default. + // Note that cipher suites are prioritized in the given order. + CipherSuites []uint16 + selfCert bool // parseFunc exists to simplify testing. Typically, parseFunc @@ -178,6 +183,10 @@ func (info TLSInfo) baseConfig() (*tls.Config, error) { ServerName: info.ServerName, } + if len(info.CipherSuites) > 0 { + cfg.CipherSuites = info.CipherSuites + } + if info.AllowedCN != "" { cfg.VerifyPeerCertificate = func(rawCerts [][]byte, verifiedChains [][]*x509.Certificate) error { for _, chains := range verifiedChains { diff --git a/vendor/github.com/coreos/etcd/pkg/transport/transport_test.go b/vendor/github.com/coreos/etcd/pkg/transport/transport_test.go new file mode 100644 index 000000000..f0860f8e7 --- /dev/null +++ b/vendor/github.com/coreos/etcd/pkg/transport/transport_test.go @@ -0,0 +1,73 @@ +// Copyright 2018 The etcd Authors +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package transport + +import ( + "crypto/tls" + "net/http" + "strings" + "testing" + "time" +) + +// TestNewTransportTLSInvalidCipherSuites expects a client with invalid +// cipher suites fail to handshake with the server. +func TestNewTransportTLSInvalidCipherSuites(t *testing.T) { + tlsInfo, del, err := createSelfCert() + if err != nil { + t.Fatalf("unable to create cert: %v", err) + } + defer del() + + cipherSuites := []uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305, + tls.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305, + } + + // make server and client have unmatched cipher suites + srvTLS, cliTLS := *tlsInfo, *tlsInfo + srvTLS.CipherSuites, cliTLS.CipherSuites = cipherSuites[:2], cipherSuites[2:] + + ln, err := NewListener("127.0.0.1:0", "https", &srvTLS) + if err != nil { + t.Fatalf("unexpected NewListener error: %v", err) + } + defer ln.Close() + + donec := make(chan struct{}) + go func() { + ln.Accept() + donec <- struct{}{} + }() + go func() { + tr, err := NewTransport(cliTLS, 3*time.Second) + if err != nil { + t.Fatalf("unexpected NewTransport error: %v", err) + } + cli := &http.Client{Transport: tr} + _, gerr := cli.Get("https://" + ln.Addr().String()) + if gerr == nil || !strings.Contains(gerr.Error(), "tls: handshake failure") { + t.Fatal("expected client TLS handshake error") + } + ln.Close() + donec <- struct{}{} + }() + <-donec + <-donec +} diff --git a/vendor/github.com/coreos/etcd/scripts/release b/vendor/github.com/coreos/etcd/scripts/release index d256566df..c71ca398e 100755 --- a/vendor/github.com/coreos/etcd/scripts/release +++ b/vendor/github.com/coreos/etcd/scripts/release @@ -48,6 +48,9 @@ main() { exit 1 fi + # Expected umask for etcd release artifacts + umask 022 + # Set up release directory. local reldir="/tmp/etcd-release-${VERSION}" if [ ! -d "${reldir}/etcd" ]; then @@ -58,8 +61,8 @@ main() { cd "${reldir}/etcd" # If a release version tag already exists, use it. - local remote_tag_exists=$(git ls-remote origin "refs/tags/${RELEASE_VERSION}" | grep -q "${RELEASE_VERSION}") - if ${remote_tag_exists}; then + local remote_tag_exists=$(git ls-remote origin "refs/tags/${RELEASE_VERSION}" | grep -c "${RELEASE_VERSION}") + if [ ${remote_tag_exists} -gt 0 ]; then echo "Release version tag exists on remote. Checking out refs/tags/${RELEASE_VERSION}" git checkout -q "tags/${RELEASE_VERSION}" fi @@ -73,7 +76,7 @@ main() { fi # If the release tag does not already exist remotely, create it. - if ! ${remote_tag_exists}; then + if [ ${remote_tag_exists} -eq 0 ]; then # Bump version/version.go to release version. local source_version=$(egrep "\s+Version\s*=" version/version.go | sed -e "s/.*\"\(.*\)\".*/\1/g") if [[ "${source_version}" != "${VERSION}" ]]; then @@ -109,7 +112,7 @@ main() { fi # Tag release. - if $(git tag --list | grep -q "${RELEASE_VERSION}"); then + if [ $(git tag --list | grep -c "${RELEASE_VERSION}") -gt 0 ]; then echo "Skipping tag step. git tag ${RELEASE_VERSION} already exists." else echo "Tagging release..." @@ -159,15 +162,20 @@ main() { else read -p "Publish etcd ${RELEASE_VERSION} docker images to quay.io [y/N]? " confirm [[ "${confirm,,}" == "y" ]] || exit 1 - until $(docker login quay.io); do - echo "Docker login failed, retrying." + for i in {1..5}; do + docker login quay.io && break + echo "login failed, retrying" done gcloud docker -- login -u _json_key -p "$(cat /etc/gcp-key-etcd-development.json)" https://gcr.io - - for TARGET_ARCH in "-arm64" "-ppc64le" ""; do - docker push quay.io/coreos/etcd:${RELEASE_VERSION}${TARGET_ARCH} - gcloud docker -- push gcr.io/etcd-development/etcd:${RELEASE_VERSION}${TARGET_ARCH} - done + + docker push quay.io/coreos/etcd:${RELEASE_VERSION} + gcloud docker -- push gcr.io/etcd-development/etcd:${RELEASE_VERSION} + if [ "${MINOR_VERSION}" != "3.1" ]; then + for TARGET_ARCH in "-arm64" "-ppc64le"; do + docker push quay.io/coreos/etcd:${RELEASE_VERSION}${TARGET_ARCH} + gcloud docker -- push gcr.io/etcd-development/etcd:${RELEASE_VERSION}${TARGET_ARCH} + done + fi gsutil -m acl ch -u allUsers:R -r gs://artifacts.etcd-development.appspot.com docker tag quay.io/coreos/etcd:${RELEASE_VERSION} quay.io/coreos/etcd:v${MINOR_VERSION} @@ -202,7 +210,6 @@ main() { # TODO: signing process echo "" echo "WARNING: The release has not been signed and published to github. This must be done manually." - echo "WARNING: version/version.go has not been updated to ${RELEASE_VERSION}+git. This must be done manually." echo "" echo "Success." exit 0 diff --git a/vendor/github.com/coreos/etcd/tests/semaphore.test.bash b/vendor/github.com/coreos/etcd/tests/semaphore.test.bash index 8dba70213..caa9b09d7 100755 --- a/vendor/github.com/coreos/etcd/tests/semaphore.test.bash +++ b/vendor/github.com/coreos/etcd/tests/semaphore.test.bash @@ -5,17 +5,34 @@ if ! [[ "$0" =~ "tests/semaphore.test.bash" ]]; then exit 255 fi -TEST_SUFFIX=$(date +%s | base64 | head -c 15) +<&1 | tee test-${TEST_SUFFIX}.log" - -! egrep "(--- FAIL:|panic: test timed out|appears to have leaked)" -B50 -A10 test-${TEST_SUFFIX}.log +echo "Running tests with" ${TEST_OPTS} +if [ "${TEST_OPTS}" == "PASSES='build grpcproxy'" ]; then + echo "Skip proxy tests for this branch!" + exit 0 +elif [ "${TEST_OPTS}" == "coverage" ]; then + echo "Skip coverage tests for this branch!" + exit 0 +else + sudo HOST_TMP_DIR=/tmp TEST_OPTS="${TEST_OPTS}" make docker-test +fi diff --git a/vendor/github.com/coreos/etcd/version/version.go b/vendor/github.com/coreos/etcd/version/version.go index ea04ae567..03ef91d43 100644 --- a/vendor/github.com/coreos/etcd/version/version.go +++ b/vendor/github.com/coreos/etcd/version/version.go @@ -26,7 +26,7 @@ import ( var ( // MinClusterVersion is the min cluster version this etcd binary is compatible with. MinClusterVersion = "3.0.0" - Version = "3.3.5" + Version = "3.3.8" APIVersion = "unknown" // Git SHA Value will be set during build diff --git a/vendor/github.com/coreos/go-semver/.travis.yml b/vendor/github.com/coreos/go-semver/.travis.yml deleted file mode 100644 index 05f548c9a..000000000 --- a/vendor/github.com/coreos/go-semver/.travis.yml +++ /dev/null @@ -1,8 +0,0 @@ -language: go -sudo: false -go: - - 1.4 - - 1.5 - - 1.6 - - tip -script: cd semver && go test diff --git a/vendor/github.com/coreos/go-semver/LICENSE b/vendor/github.com/coreos/go-semver/LICENSE deleted file mode 100644 index d64569567..000000000 --- a/vendor/github.com/coreos/go-semver/LICENSE +++ /dev/null @@ -1,202 +0,0 @@ - - Apache License - Version 2.0, January 2004 - http://www.apache.org/licenses/ - - TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION - - 1. Definitions. - - "License" shall mean the terms and conditions for use, reproduction, - and distribution as defined by Sections 1 through 9 of this document. - - "Licensor" shall mean the copyright owner or entity authorized by - the copyright owner that is granting the License. - - "Legal Entity" shall mean the union of the acting entity and all - other entities that control, are controlled by, or are under common - control with that entity. For the purposes of this definition, - "control" means (i) the power, direct or indirect, to cause the - direction or management of such entity, whether by contract or - otherwise, or (ii) ownership of fifty percent (50%) or more of the - outstanding shares, or (iii) beneficial ownership of such entity. - - "You" (or "Your") shall mean an individual or Legal Entity - exercising permissions granted by this License. - - "Source" form shall mean the preferred form for making modifications, - including but not limited to software source code, documentation - source, and configuration files. - - "Object" form shall mean any form resulting from mechanical - transformation or translation of a Source form, including but - not limited to compiled object code, generated documentation, - and conversions to other media types. - - "Work" shall mean the work of authorship, whether in Source or - Object form, made available under the License, as indicated by a - copyright notice that is included in or attached to the work - (an example is provided in the Appendix below). - - "Derivative Works" shall mean any work, whether in Source or Object - form, that is based on (or derived from) the Work and for which the - editorial revisions, annotations, elaborations, or other modifications - represent, as a whole, an original work of authorship. For the purposes - of this License, Derivative Works shall not include works that remain - separable from, or merely link (or bind by name) to the interfaces of, - the Work and Derivative Works thereof. - - "Contribution" shall mean any work of authorship, including - the original version of the Work and any modifications or additions - to that Work or Derivative Works thereof, that is intentionally - submitted to Licensor for inclusion in the Work by the copyright owner - or by an individual or Legal Entity authorized to submit on behalf of - the copyright owner. For the purposes of this definition, "submitted" - means any form of electronic, verbal, or written communication sent - to the Licensor or its representatives, including but not limited to - communication on electronic mailing lists, source code control systems, - and issue tracking systems that are managed by, or on behalf of, the - Licensor for the purpose of discussing and improving the Work, but - excluding communication that is conspicuously marked or otherwise - designated in writing by the copyright owner as "Not a Contribution." - - "Contributor" shall mean Licensor and any individual or Legal Entity - on behalf of whom a Contribution has been received by Licensor and - subsequently incorporated within the Work. - - 2. Grant of Copyright License. Subject to the terms and conditions of - this License, each Contributor hereby grants to You a perpetual, - worldwide, non-exclusive, no-charge, royalty-free, irrevocable - copyright license to reproduce, prepare Derivative Works of, - publicly display, publicly perform, sublicense, and distribute the - Work and such Derivative Works in Source or Object form. - - 3. Grant of Patent License. Subject to the terms and conditions of - this License, each Contributor hereby grants to You a perpetual, - worldwide, non-exclusive, no-charge, royalty-free, irrevocable - (except as stated in this section) patent license to make, have made, - use, offer to sell, sell, import, and otherwise transfer the Work, - where such license applies only to those patent claims licensable - by such Contributor that are necessarily infringed by their - Contribution(s) alone or by combination of their Contribution(s) - with the Work to which such Contribution(s) was submitted. If You - institute patent litigation against any entity (including a - cross-claim or counterclaim in a lawsuit) alleging that the Work - or a Contribution incorporated within the Work constitutes direct - or contributory patent infringement, then any patent licenses - granted to You under this License for that Work shall terminate - as of the date such litigation is filed. - - 4. Redistribution. You may reproduce and distribute copies of the - Work or Derivative Works thereof in any medium, with or without - modifications, and in Source or Object form, provided that You - meet the following conditions: - - (a) You must give any other recipients of the Work or - Derivative Works a copy of this License; and - - (b) You must cause any modified files to carry prominent notices - stating that You changed the files; and - - (c) You must retain, in the Source form of any Derivative Works - that You distribute, all copyright, patent, trademark, and - attribution notices from the Source form of the Work, - excluding those notices that do not pertain to any part of - the Derivative Works; and - - (d) If the Work includes a "NOTICE" text file as part of its - distribution, then any Derivative Works that You distribute must - include a readable copy of the attribution notices contained - within such NOTICE file, excluding those notices that do not - pertain to any part of the Derivative Works, in at least one - of the following places: within a NOTICE text file distributed - as part of the Derivative Works; within the Source form or - documentation, if provided along with the Derivative Works; or, - within a display generated by the Derivative Works, if and - wherever such third-party notices normally appear. The contents - of the NOTICE file are for informational purposes only and - do not modify the License. You may add Your own attribution - notices within Derivative Works that You distribute, alongside - or as an addendum to the NOTICE text from the Work, provided - that such additional attribution notices cannot be construed - as modifying the License. - - You may add Your own copyright statement to Your modifications and - may provide additional or different license terms and conditions - for use, reproduction, or distribution of Your modifications, or - for any such Derivative Works as a whole, provided Your use, - reproduction, and distribution of the Work otherwise complies with - the conditions stated in this License. - - 5. Submission of Contributions. Unless You explicitly state otherwise, - any Contribution intentionally submitted for inclusion in the Work - by You to the Licensor shall be under the terms and conditions of - this License, without any additional terms or conditions. - Notwithstanding the above, nothing herein shall supersede or modify - the terms of any separate license agreement you may have executed - with Licensor regarding such Contributions. - - 6. Trademarks. This License does not grant permission to use the trade - names, trademarks, service marks, or product names of the Licensor, - except as required for reasonable and customary use in describing the - origin of the Work and reproducing the content of the NOTICE file. - - 7. Disclaimer of Warranty. Unless required by applicable law or - agreed to in writing, Licensor provides the Work (and each - Contributor provides its Contributions) on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or - implied, including, without limitation, any warranties or conditions - of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A - PARTICULAR PURPOSE. You are solely responsible for determining the - appropriateness of using or redistributing the Work and assume any - risks associated with Your exercise of permissions under this License. - - 8. Limitation of Liability. In no event and under no legal theory, - whether in tort (including negligence), contract, or otherwise, - unless required by applicable law (such as deliberate and grossly - negligent acts) or agreed to in writing, shall any Contributor be - liable to You for damages, including any direct, indirect, special, - incidental, or consequential damages of any character arising as a - result of this License or out of the use or inability to use the - Work (including but not limited to damages for loss of goodwill, - work stoppage, computer failure or malfunction, or any and all - other commercial damages or losses), even if such Contributor - has been advised of the possibility of such damages. - - 9. Accepting Warranty or Additional Liability. While redistributing - the Work or Derivative Works thereof, You may choose to offer, - and charge a fee for, acceptance of support, warranty, indemnity, - or other liability obligations and/or rights consistent with this - License. However, in accepting such obligations, You may act only - on Your own behalf and on Your sole responsibility, not on behalf - of any other Contributor, and only if You agree to indemnify, - defend, and hold each Contributor harmless for any liability - incurred by, or claims asserted against, such Contributor by reason - of your accepting any such warranty or additional liability. - - END OF TERMS AND CONDITIONS - - APPENDIX: How to apply the Apache License to your work. - - To apply the Apache License to your work, attach the following - boilerplate notice, with the fields enclosed by brackets "[]" - replaced with your own identifying information. (Don't include - the brackets!) The text should be enclosed in the appropriate - comment syntax for the file format. We also recommend that a - file or class name and description of purpose be included on the - same "printed page" as the copyright notice for easier - identification within third-party archives. - - Copyright [yyyy] [name of copyright owner] - - Licensed under the Apache License, Version 2.0 (the "License"); - you may not use this file except in compliance with the License. - You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - - Unless required by applicable law or agreed to in writing, software - distributed under the License is distributed on an "AS IS" BASIS, - WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - See the License for the specific language governing permissions and - limitations under the License. diff --git a/vendor/github.com/coreos/go-semver/README.md b/vendor/github.com/coreos/go-semver/README.md deleted file mode 100644 index 5bc9263cf..000000000 --- a/vendor/github.com/coreos/go-semver/README.md +++ /dev/null @@ -1,28 +0,0 @@ -# go-semver - Semantic Versioning Library - -[![Build Status](https://travis-ci.org/coreos/go-semver.svg?branch=master)](https://travis-ci.org/coreos/go-semver) -[![GoDoc](https://godoc.org/github.com/coreos/go-semver/semver?status.svg)](https://godoc.org/github.com/coreos/go-semver/semver) - -go-semver is a [semantic versioning][semver] library for Go. It lets you parse -and compare two semantic version strings. - -[semver]: http://semver.org/ - -## Usage - -```go -vA := semver.New("1.2.3") -vB := semver.New("3.2.1") - -fmt.Printf("%s < %s == %t\n", vA, vB, vA.LessThan(*vB)) -``` - -## Example Application - -``` -$ go run example.go 1.2.3 3.2.1 -1.2.3 < 3.2.1 == true - -$ go run example.go 5.2.3 3.2.1 -5.2.3 < 3.2.1 == false -``` diff --git a/vendor/github.com/coreos/go-semver/example.go b/vendor/github.com/coreos/go-semver/example.go deleted file mode 100644 index fd2ee5af2..000000000 --- a/vendor/github.com/coreos/go-semver/example.go +++ /dev/null @@ -1,20 +0,0 @@ -package main - -import ( - "fmt" - "github.com/coreos/go-semver/semver" - "os" -) - -func main() { - vA, err := semver.NewVersion(os.Args[1]) - if err != nil { - fmt.Println(err.Error()) - } - vB, err := semver.NewVersion(os.Args[2]) - if err != nil { - fmt.Println(err.Error()) - } - - fmt.Printf("%s < %s == %t\n", vA, vB, vA.LessThan(*vB)) -} diff --git a/vendor/github.com/coreos/go-semver/semver/semver.go b/vendor/github.com/coreos/go-semver/semver/semver.go deleted file mode 100644 index 110fc23e1..000000000 --- a/vendor/github.com/coreos/go-semver/semver/semver.go +++ /dev/null @@ -1,268 +0,0 @@ -// Copyright 2013-2015 CoreOS, Inc. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// Semantic Versions http://semver.org -package semver - -import ( - "bytes" - "errors" - "fmt" - "strconv" - "strings" -) - -type Version struct { - Major int64 - Minor int64 - Patch int64 - PreRelease PreRelease - Metadata string -} - -type PreRelease string - -func splitOff(input *string, delim string) (val string) { - parts := strings.SplitN(*input, delim, 2) - - if len(parts) == 2 { - *input = parts[0] - val = parts[1] - } - - return val -} - -func New(version string) *Version { - return Must(NewVersion(version)) -} - -func NewVersion(version string) (*Version, error) { - v := Version{} - - if err := v.Set(version); err != nil { - return nil, err - } - - return &v, nil -} - -// Must is a helper for wrapping NewVersion and will panic if err is not nil. -func Must(v *Version, err error) *Version { - if err != nil { - panic(err) - } - return v -} - -// Set parses and updates v from the given version string. Implements flag.Value -func (v *Version) Set(version string) error { - metadata := splitOff(&version, "+") - preRelease := PreRelease(splitOff(&version, "-")) - dotParts := strings.SplitN(version, ".", 3) - - if len(dotParts) != 3 { - return fmt.Errorf("%s is not in dotted-tri format", version) - } - - parsed := make([]int64, 3, 3) - - for i, v := range dotParts[:3] { - val, err := strconv.ParseInt(v, 10, 64) - parsed[i] = val - if err != nil { - return err - } - } - - v.Metadata = metadata - v.PreRelease = preRelease - v.Major = parsed[0] - v.Minor = parsed[1] - v.Patch = parsed[2] - return nil -} - -func (v Version) String() string { - var buffer bytes.Buffer - - fmt.Fprintf(&buffer, "%d.%d.%d", v.Major, v.Minor, v.Patch) - - if v.PreRelease != "" { - fmt.Fprintf(&buffer, "-%s", v.PreRelease) - } - - if v.Metadata != "" { - fmt.Fprintf(&buffer, "+%s", v.Metadata) - } - - return buffer.String() -} - -func (v *Version) UnmarshalYAML(unmarshal func(interface{}) error) error { - var data string - if err := unmarshal(&data); err != nil { - return err - } - return v.Set(data) -} - -func (v Version) MarshalJSON() ([]byte, error) { - return []byte(`"` + v.String() + `"`), nil -} - -func (v *Version) UnmarshalJSON(data []byte) error { - l := len(data) - if l == 0 || string(data) == `""` { - return nil - } - if l < 2 || data[0] != '"' || data[l-1] != '"' { - return errors.New("invalid semver string") - } - return v.Set(string(data[1 : l-1])) -} - -// Compare tests if v is less than, equal to, or greater than versionB, -// returning -1, 0, or +1 respectively. -func (v Version) Compare(versionB Version) int { - if cmp := recursiveCompare(v.Slice(), versionB.Slice()); cmp != 0 { - return cmp - } - return preReleaseCompare(v, versionB) -} - -// Equal tests if v is equal to versionB. -func (v Version) Equal(versionB Version) bool { - return v.Compare(versionB) == 0 -} - -// LessThan tests if v is less than versionB. -func (v Version) LessThan(versionB Version) bool { - return v.Compare(versionB) < 0 -} - -// Slice converts the comparable parts of the semver into a slice of integers. -func (v Version) Slice() []int64 { - return []int64{v.Major, v.Minor, v.Patch} -} - -func (p PreRelease) Slice() []string { - preRelease := string(p) - return strings.Split(preRelease, ".") -} - -func preReleaseCompare(versionA Version, versionB Version) int { - a := versionA.PreRelease - b := versionB.PreRelease - - /* Handle the case where if two versions are otherwise equal it is the - * one without a PreRelease that is greater */ - if len(a) == 0 && (len(b) > 0) { - return 1 - } else if len(b) == 0 && (len(a) > 0) { - return -1 - } - - // If there is a prerelease, check and compare each part. - return recursivePreReleaseCompare(a.Slice(), b.Slice()) -} - -func recursiveCompare(versionA []int64, versionB []int64) int { - if len(versionA) == 0 { - return 0 - } - - a := versionA[0] - b := versionB[0] - - if a > b { - return 1 - } else if a < b { - return -1 - } - - return recursiveCompare(versionA[1:], versionB[1:]) -} - -func recursivePreReleaseCompare(versionA []string, versionB []string) int { - // A larger set of pre-release fields has a higher precedence than a smaller set, - // if all of the preceding identifiers are equal. - if len(versionA) == 0 { - if len(versionB) > 0 { - return -1 - } - return 0 - } else if len(versionB) == 0 { - // We're longer than versionB so return 1. - return 1 - } - - a := versionA[0] - b := versionB[0] - - aInt := false - bInt := false - - aI, err := strconv.Atoi(versionA[0]) - if err == nil { - aInt = true - } - - bI, err := strconv.Atoi(versionB[0]) - if err == nil { - bInt = true - } - - // Handle Integer Comparison - if aInt && bInt { - if aI > bI { - return 1 - } else if aI < bI { - return -1 - } - } - - // Handle String Comparison - if a > b { - return 1 - } else if a < b { - return -1 - } - - return recursivePreReleaseCompare(versionA[1:], versionB[1:]) -} - -// BumpMajor increments the Major field by 1 and resets all other fields to their default values -func (v *Version) BumpMajor() { - v.Major += 1 - v.Minor = 0 - v.Patch = 0 - v.PreRelease = PreRelease("") - v.Metadata = "" -} - -// BumpMinor increments the Minor field by 1 and resets all other fields to their default values -func (v *Version) BumpMinor() { - v.Minor += 1 - v.Patch = 0 - v.PreRelease = PreRelease("") - v.Metadata = "" -} - -// BumpPatch increments the Patch field by 1 and resets all other fields to their default values -func (v *Version) BumpPatch() { - v.Patch += 1 - v.PreRelease = PreRelease("") - v.Metadata = "" -} diff --git a/vendor/github.com/coreos/go-semver/semver/semver_test.go b/vendor/github.com/coreos/go-semver/semver/semver_test.go deleted file mode 100644 index 876c68e04..000000000 --- a/vendor/github.com/coreos/go-semver/semver/semver_test.go +++ /dev/null @@ -1,370 +0,0 @@ -// Copyright 2013-2015 CoreOS, Inc. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -package semver - -import ( - "bytes" - "encoding/json" - "errors" - "flag" - "fmt" - "math/rand" - "reflect" - "testing" - "time" - - "gopkg.in/yaml.v2" -) - -type fixture struct { - GreaterVersion string - LesserVersion string -} - -var fixtures = []fixture{ - fixture{"0.0.0", "0.0.0-foo"}, - fixture{"0.0.1", "0.0.0"}, - fixture{"1.0.0", "0.9.9"}, - fixture{"0.10.0", "0.9.0"}, - fixture{"0.99.0", "0.10.0"}, - fixture{"2.0.0", "1.2.3"}, - fixture{"0.0.0", "0.0.0-foo"}, - fixture{"0.0.1", "0.0.0"}, - fixture{"1.0.0", "0.9.9"}, - fixture{"0.10.0", "0.9.0"}, - fixture{"0.99.0", "0.10.0"}, - fixture{"2.0.0", "1.2.3"}, - fixture{"0.0.0", "0.0.0-foo"}, - fixture{"0.0.1", "0.0.0"}, - fixture{"1.0.0", "0.9.9"}, - fixture{"0.10.0", "0.9.0"}, - fixture{"0.99.0", "0.10.0"}, - fixture{"2.0.0", "1.2.3"}, - fixture{"1.2.3", "1.2.3-asdf"}, - fixture{"1.2.3", "1.2.3-4"}, - fixture{"1.2.3", "1.2.3-4-foo"}, - fixture{"1.2.3-5-foo", "1.2.3-5"}, - fixture{"1.2.3-5", "1.2.3-4"}, - fixture{"1.2.3-5-foo", "1.2.3-5-Foo"}, - fixture{"3.0.0", "2.7.2+asdf"}, - fixture{"3.0.0+foobar", "2.7.2"}, - fixture{"1.2.3-a.10", "1.2.3-a.5"}, - fixture{"1.2.3-a.b", "1.2.3-a.5"}, - fixture{"1.2.3-a.b", "1.2.3-a"}, - fixture{"1.2.3-a.b.c.10.d.5", "1.2.3-a.b.c.5.d.100"}, - fixture{"1.0.0", "1.0.0-rc.1"}, - fixture{"1.0.0-rc.2", "1.0.0-rc.1"}, - fixture{"1.0.0-rc.1", "1.0.0-beta.11"}, - fixture{"1.0.0-beta.11", "1.0.0-beta.2"}, - fixture{"1.0.0-beta.2", "1.0.0-beta"}, - fixture{"1.0.0-beta", "1.0.0-alpha.beta"}, - fixture{"1.0.0-alpha.beta", "1.0.0-alpha.1"}, - fixture{"1.0.0-alpha.1", "1.0.0-alpha"}, -} - -func TestCompare(t *testing.T) { - for _, v := range fixtures { - gt, err := NewVersion(v.GreaterVersion) - if err != nil { - t.Error(err) - } - - lt, err := NewVersion(v.LesserVersion) - if err != nil { - t.Error(err) - } - - if gt.LessThan(*lt) { - t.Errorf("%s should not be less than %s", gt, lt) - } - if gt.Equal(*lt) { - t.Errorf("%s should not be equal to %s", gt, lt) - } - if gt.Compare(*lt) <= 0 { - t.Errorf("%s should be greater than %s", gt, lt) - } - if !lt.LessThan(*gt) { - t.Errorf("%s should be less than %s", lt, gt) - } - if !lt.Equal(*lt) { - t.Errorf("%s should be equal to %s", lt, lt) - } - if lt.Compare(*gt) > 0 { - t.Errorf("%s should not be greater than %s", lt, gt) - } - } -} - -func testString(t *testing.T, orig string, version *Version) { - if orig != version.String() { - t.Errorf("%s != %s", orig, version) - } -} - -func TestString(t *testing.T) { - for _, v := range fixtures { - gt, err := NewVersion(v.GreaterVersion) - if err != nil { - t.Error(err) - } - testString(t, v.GreaterVersion, gt) - - lt, err := NewVersion(v.LesserVersion) - if err != nil { - t.Error(err) - } - testString(t, v.LesserVersion, lt) - } -} - -func shuffleStringSlice(src []string) []string { - dest := make([]string, len(src)) - rand.Seed(time.Now().Unix()) - perm := rand.Perm(len(src)) - for i, v := range perm { - dest[v] = src[i] - } - return dest -} - -func TestSort(t *testing.T) { - sortedVersions := []string{"1.0.0", "1.0.2", "1.2.0", "3.1.1"} - unsortedVersions := shuffleStringSlice(sortedVersions) - - semvers := []*Version{} - for _, v := range unsortedVersions { - sv, err := NewVersion(v) - if err != nil { - t.Fatal(err) - } - semvers = append(semvers, sv) - } - - Sort(semvers) - - for idx, sv := range semvers { - if sv.String() != sortedVersions[idx] { - t.Fatalf("incorrect sort at index %v", idx) - } - } -} - -func TestBumpMajor(t *testing.T) { - version, _ := NewVersion("1.0.0") - version.BumpMajor() - if version.Major != 2 { - t.Fatalf("bumping major on 1.0.0 resulted in %v", version) - } - - version, _ = NewVersion("1.5.2") - version.BumpMajor() - if version.Minor != 0 && version.Patch != 0 { - t.Fatalf("bumping major on 1.5.2 resulted in %v", version) - } - - version, _ = NewVersion("1.0.0+build.1-alpha.1") - version.BumpMajor() - if version.PreRelease != "" && version.PreRelease != "" { - t.Fatalf("bumping major on 1.0.0+build.1-alpha.1 resulted in %v", version) - } -} - -func TestBumpMinor(t *testing.T) { - version, _ := NewVersion("1.0.0") - version.BumpMinor() - - if version.Major != 1 { - t.Fatalf("bumping minor on 1.0.0 resulted in %v", version) - } - - if version.Minor != 1 { - t.Fatalf("bumping major on 1.0.0 resulted in %v", version) - } - - version, _ = NewVersion("1.0.0+build.1-alpha.1") - version.BumpMinor() - if version.PreRelease != "" && version.PreRelease != "" { - t.Fatalf("bumping major on 1.0.0+build.1-alpha.1 resulted in %v", version) - } -} - -func TestBumpPatch(t *testing.T) { - version, _ := NewVersion("1.0.0") - version.BumpPatch() - - if version.Major != 1 { - t.Fatalf("bumping minor on 1.0.0 resulted in %v", version) - } - - if version.Minor != 0 { - t.Fatalf("bumping major on 1.0.0 resulted in %v", version) - } - - if version.Patch != 1 { - t.Fatalf("bumping major on 1.0.0 resulted in %v", version) - } - - version, _ = NewVersion("1.0.0+build.1-alpha.1") - version.BumpPatch() - if version.PreRelease != "" && version.PreRelease != "" { - t.Fatalf("bumping major on 1.0.0+build.1-alpha.1 resulted in %v", version) - } -} - -func TestMust(t *testing.T) { - tests := []struct { - versionStr string - - version *Version - recov interface{} - }{ - { - versionStr: "1.0.0", - version: &Version{Major: 1}, - }, - { - versionStr: "version number", - recov: errors.New("version number is not in dotted-tri format"), - }, - } - - for _, tt := range tests { - func() { - defer func() { - recov := recover() - if !reflect.DeepEqual(tt.recov, recov) { - t.Fatalf("incorrect panic for %q: want %v, got %v", tt.versionStr, tt.recov, recov) - } - }() - - version := Must(NewVersion(tt.versionStr)) - if !reflect.DeepEqual(tt.version, version) { - t.Fatalf("incorrect version for %q: want %+v, got %+v", tt.versionStr, tt.version, version) - } - }() - } -} - -type fixtureJSON struct { - GreaterVersion *Version - LesserVersion *Version -} - -func TestJSON(t *testing.T) { - fj := make([]fixtureJSON, len(fixtures)) - for i, v := range fixtures { - var err error - fj[i].GreaterVersion, err = NewVersion(v.GreaterVersion) - if err != nil { - t.Fatal(err) - } - fj[i].LesserVersion, err = NewVersion(v.LesserVersion) - if err != nil { - t.Fatal(err) - } - } - - fromStrings, err := json.Marshal(fixtures) - if err != nil { - t.Fatal(err) - } - fromVersions, err := json.Marshal(fj) - if err != nil { - t.Fatal(err) - } - if !bytes.Equal(fromStrings, fromVersions) { - t.Errorf("Expected: %s", fromStrings) - t.Errorf("Unexpected: %s", fromVersions) - } - - fromJson := make([]fixtureJSON, 0, len(fj)) - err = json.Unmarshal(fromStrings, &fromJson) - if err != nil { - t.Fatal(err) - } - if !reflect.DeepEqual(fromJson, fj) { - t.Error("Expected: ", fj) - t.Error("Unexpected: ", fromJson) - } -} - -func TestYAML(t *testing.T) { - document, err := yaml.Marshal(fixtures) - if err != nil { - t.Fatal(err) - } - - expected := make([]fixtureJSON, len(fixtures)) - for i, v := range fixtures { - var err error - expected[i].GreaterVersion, err = NewVersion(v.GreaterVersion) - if err != nil { - t.Fatal(err) - } - expected[i].LesserVersion, err = NewVersion(v.LesserVersion) - if err != nil { - t.Fatal(err) - } - } - - fromYAML := make([]fixtureJSON, 0, len(fixtures)) - err = yaml.Unmarshal(document, &fromYAML) - if err != nil { - t.Fatal(err) - } - - if !reflect.DeepEqual(fromYAML, expected) { - t.Error("Expected: ", expected) - t.Error("Unexpected: ", fromYAML) - } -} - -func TestBadInput(t *testing.T) { - bad := []string{ - "1.2", - "1.2.3x", - "0x1.3.4", - "-1.2.3", - "1.2.3.4", - } - for _, b := range bad { - if _, err := NewVersion(b); err == nil { - t.Error("Improperly accepted value: ", b) - } - } -} - -func TestFlag(t *testing.T) { - v := Version{} - f := flag.NewFlagSet("version", flag.ContinueOnError) - f.Var(&v, "version", "set version") - - if err := f.Set("version", "1.2.3"); err != nil { - t.Fatal(err) - } - - if v.String() != "1.2.3" { - t.Errorf("Set wrong value %q", v) - } -} - -func ExampleVersion_LessThan() { - vA := New("1.2.3") - vB := New("3.2.1") - - fmt.Printf("%s < %s == %t\n", vA, vB, vA.LessThan(*vB)) - // Output: - // 1.2.3 < 3.2.1 == true -} diff --git a/vendor/github.com/go-ini/ini/.travis.yml b/vendor/github.com/go-ini/ini/.travis.yml index 75fe7b74b..3cb77dc9a 100644 --- a/vendor/github.com/go-ini/ini/.travis.yml +++ b/vendor/github.com/go-ini/ini/.travis.yml @@ -13,4 +13,5 @@ script: - go get github.com/smartystreets/goconvey - mkdir -p $HOME/gopath/src/gopkg.in - ln -s $HOME/gopath/src/github.com/go-ini/ini $HOME/gopath/src/gopkg.in/ini.v1 + - cd $HOME/gopath/src/gopkg.in/ini.v1 - go test -v -cover -race diff --git a/vendor/github.com/go-ini/ini/ini.go b/vendor/github.com/go-ini/ini/ini.go index d98353229..15ebc8f72 100644 --- a/vendor/github.com/go-ini/ini/ini.go +++ b/vendor/github.com/go-ini/ini/ini.go @@ -32,7 +32,7 @@ const ( // Maximum allowed depth when recursively substituing variable names. _DEPTH_VALUES = 99 - _VERSION = "1.36.0" + _VERSION = "1.37.0" ) // Version returns current package version literal. @@ -145,6 +145,11 @@ type LoadOptions struct { // Relevant quote: Values can also span multiple lines, as long as they are indented deeper // than the first line of the value. AllowPythonMultilineValues bool + // SpaceBeforeInlineComment indicates whether to allow comment symbols (\# and \;) inside value. + // Docs: https://docs.python.org/2/library/configparser.html + // Quote: Comments may appear on their own in an otherwise empty line, or may be entered in lines holding values or section names. + // In the latter case, they need to be preceded by a whitespace character to be recognized as a comment. + SpaceBeforeInlineComment bool // UnescapeValueDoubleQuotes indicates whether to unescape double quotes inside value to regular format // when value is surrounded by double quotes, e.g. key="a \"value\"" => key=a "value" UnescapeValueDoubleQuotes bool diff --git a/vendor/github.com/go-ini/ini/ini_test.go b/vendor/github.com/go-ini/ini/ini_test.go index 6b764018a..d68c7caa3 100644 --- a/vendor/github.com/go-ini/ini/ini_test.go +++ b/vendor/github.com/go-ini/ini/ini_test.go @@ -65,7 +65,7 @@ NAME = Unknwon So(err, ShouldBeNil) So(f, ShouldNotBeNil) - // Vaildate values make sure all sources are loaded correctly + // Validate values make sure all sources are loaded correctly sec := f.Section("") So(sec.Key("NAME").String(), ShouldEqual, "ini") So(sec.Key("VERSION").String(), ShouldEqual, "v1") @@ -88,7 +88,23 @@ NAME = Unknwon }) }) - Convey("Cant't parse small python-compatible INI files", t, func() { + Convey("Can't properly parse INI files containing `#` or `;` in value", t, func() { + f, err := ini.Load([]byte(` + [author] + NAME = U#n#k#n#w#o#n + GITHUB = U;n;k;n;w;o;n + `)) + So(err, ShouldBeNil) + So(f, ShouldNotBeNil) + + sec := f.Section("author") + nameValue := sec.Key("NAME").String() + githubValue := sec.Key("GITHUB").String() + So(nameValue, ShouldEqual, "U") + So(githubValue, ShouldEqual, "U") + }) + + Convey("Can't parse small python-compatible INI files", t, func() { f, err := ini.Load([]byte(` [long] long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- @@ -103,7 +119,7 @@ long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- So(err.Error(), ShouldEqual, "key-value delimiter not found: foo\n") }) - Convey("Cant't parse big python-compatible INI files", t, func() { + Convey("Can't parse big python-compatible INI files", t, func() { f, err := ini.Load([]byte(` [long] long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- @@ -324,12 +340,14 @@ key2=c\d\`)) IgnoreInlineComment: true, }, []byte(` key1=value ;comment -key2=value2 #comment2`)) +key2=value2 #comment2 +key3=val#ue #comment3`)) So(err, ShouldBeNil) So(f, ShouldNotBeNil) So(f.Section("").Key("key1").String(), ShouldEqual, `value ;comment`) So(f.Section("").Key("key2").String(), ShouldEqual, `value2 #comment2`) + So(f.Section("").Key("key3").String(), ShouldEqual, `val#ue #comment3`) Convey("Inverse case", func() { f, err := ini.LoadSources(ini.LoadOptions{AllowPythonMultilineValues: true}, []byte(` @@ -461,11 +479,11 @@ key = test value more text }, []byte(` [long] long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- - foo - bar - foobar - barfoo - -----END RSA PRIVATE KEY----- + foo + bar + foobar + barfoo + -----END RSA PRIVATE KEY----- `)) So(err, ShouldBeNil) So(f, ShouldNotBeNil) @@ -481,103 +499,103 @@ long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- }, []byte(` [long] long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- - 1foo - 2bar - 3foobar - 4barfoo - 5foo - 6bar - 7foobar - 8barfoo - 9foo - 10bar - 11foobar - 12barfoo - 13foo - 14bar - 15foobar - 16barfoo - 17foo - 18bar - 19foobar - 20barfoo - 21foo - 22bar - 23foobar - 24barfoo - 25foo - 26bar - 27foobar - 28barfoo - 29foo - 30bar - 31foobar - 32barfoo - 33foo - 34bar - 35foobar - 36barfoo - 37foo - 38bar - 39foobar - 40barfoo - 41foo - 42bar - 43foobar - 44barfoo - 45foo - 46bar - 47foobar - 48barfoo - 49foo - 50bar - 51foobar - 52barfoo - 53foo - 54bar - 55foobar - 56barfoo - 57foo - 58bar - 59foobar - 60barfoo - 61foo - 62bar - 63foobar - 64barfoo - 65foo - 66bar - 67foobar - 68barfoo - 69foo - 70bar - 71foobar - 72barfoo - 73foo - 74bar - 75foobar - 76barfoo - 77foo - 78bar - 79foobar - 80barfoo - 81foo - 82bar - 83foobar - 84barfoo - 85foo - 86bar - 87foobar - 88barfoo - 89foo - 90bar - 91foobar - 92barfoo - 93foo - 94bar - 95foobar - 96barfoo - -----END RSA PRIVATE KEY----- + 1foo + 2bar + 3foobar + 4barfoo + 5foo + 6bar + 7foobar + 8barfoo + 9foo + 10bar + 11foobar + 12barfoo + 13foo + 14bar + 15foobar + 16barfoo + 17foo + 18bar + 19foobar + 20barfoo + 21foo + 22bar + 23foobar + 24barfoo + 25foo + 26bar + 27foobar + 28barfoo + 29foo + 30bar + 31foobar + 32barfoo + 33foo + 34bar + 35foobar + 36barfoo + 37foo + 38bar + 39foobar + 40barfoo + 41foo + 42bar + 43foobar + 44barfoo + 45foo + 46bar + 47foobar + 48barfoo + 49foo + 50bar + 51foobar + 52barfoo + 53foo + 54bar + 55foobar + 56barfoo + 57foo + 58bar + 59foobar + 60barfoo + 61foo + 62bar + 63foobar + 64barfoo + 65foo + 66bar + 67foobar + 68barfoo + 69foo + 70bar + 71foobar + 72barfoo + 73foo + 74bar + 75foobar + 76barfoo + 77foo + 78bar + 79foobar + 80barfoo + 81foo + 82bar + 83foobar + 84barfoo + 85foo + 86bar + 87foobar + 88barfoo + 89foo + 90bar + 91foobar + 92barfoo + 93foo + 94bar + 95foobar + 96barfoo + -----END RSA PRIVATE KEY----- `)) So(err, ShouldBeNil) So(f, ShouldNotBeNil) @@ -734,6 +752,44 @@ my lesson state data – 1111111111111111111000000000000000001110000 So(err, ShouldNotBeNil) }) }) + + Convey("And false `SpaceBeforeInlineComment`", func() { + Convey("Can't parse INI files containing `#` or `;` in value", func() { + f, err := ini.LoadSources( + ini.LoadOptions{AllowPythonMultilineValues: false, SpaceBeforeInlineComment: false}, + []byte(` +[author] +NAME = U#n#k#n#w#o#n +GITHUB = U;n;k;n;w;o;n +`)) + So(err, ShouldBeNil) + So(f, ShouldNotBeNil) + sec := f.Section("author") + nameValue := sec.Key("NAME").String() + githubValue := sec.Key("GITHUB").String() + So(nameValue, ShouldEqual, "U") + So(githubValue, ShouldEqual, "U") + }) + }) + + Convey("And true `SpaceBeforeInlineComment`", func() { + Convey("Can parse INI files containing `#` or `;` in value", func() { + f, err := ini.LoadSources( + ini.LoadOptions{AllowPythonMultilineValues: false, SpaceBeforeInlineComment: true}, + []byte(` +[author] +NAME = U#n#k#n#w#o#n +GITHUB = U;n;k;n;w;o;n +`)) + So(err, ShouldBeNil) + So(f, ShouldNotBeNil) + sec := f.Section("author") + nameValue := sec.Key("NAME").String() + githubValue := sec.Key("GITHUB").String() + So(nameValue, ShouldEqual, "U#n#k#n#w#o#n") + So(githubValue, ShouldEqual, "U;n;k;n;w;o;n") + }) + }) }) Convey("with false `AllowPythonMultilineValues`", func() { @@ -806,12 +862,14 @@ key2=c\d\`)) IgnoreInlineComment: true, }, []byte(` key1=value ;comment -key2=value2 #comment2`)) +key2=value2 #comment2 +key3=val#ue #comment3`)) So(err, ShouldBeNil) So(f, ShouldNotBeNil) So(f.Section("").Key("key1").String(), ShouldEqual, `value ;comment`) So(f.Section("").Key("key2").String(), ShouldEqual, `value2 #comment2`) + So(f.Section("").Key("key3").String(), ShouldEqual, `val#ue #comment3`) Convey("Inverse case", func() { f, err := ini.LoadSources(ini.LoadOptions{AllowPythonMultilineValues: false}, []byte(` @@ -935,122 +993,122 @@ key = test value more text So(f.Section("").Key("key").String(), ShouldEqual, `test value more text`) }) - Convey("Cant't parse small python-compatible INI files", func() { + Convey("Can't parse small python-compatible INI files", func() { f, err := ini.LoadSources(ini.LoadOptions{AllowPythonMultilineValues: false}, []byte(` [long] long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- - foo - bar - foobar - barfoo - -----END RSA PRIVATE KEY----- + foo + bar + foobar + barfoo + -----END RSA PRIVATE KEY----- `)) So(err, ShouldNotBeNil) So(f, ShouldBeNil) So(err.Error(), ShouldEqual, "key-value delimiter not found: foo\n") }) - Convey("Cant't parse big python-compatible INI files", func() { + Convey("Can't parse big python-compatible INI files", func() { f, err := ini.LoadSources(ini.LoadOptions{AllowPythonMultilineValues: false}, []byte(` [long] long_rsa_private_key = -----BEGIN RSA PRIVATE KEY----- - 1foo - 2bar - 3foobar - 4barfoo - 5foo - 6bar - 7foobar - 8barfoo - 9foo - 10bar - 11foobar - 12barfoo - 13foo - 14bar - 15foobar - 16barfoo - 17foo - 18bar - 19foobar - 20barfoo - 21foo - 22bar - 23foobar - 24barfoo - 25foo - 26bar - 27foobar - 28barfoo - 29foo - 30bar - 31foobar - 32barfoo - 33foo - 34bar - 35foobar - 36barfoo - 37foo - 38bar - 39foobar - 40barfoo - 41foo - 42bar - 43foobar - 44barfoo - 45foo - 46bar - 47foobar - 48barfoo - 49foo - 50bar - 51foobar - 52barfoo - 53foo - 54bar - 55foobar - 56barfoo - 57foo - 58bar - 59foobar - 60barfoo - 61foo - 62bar - 63foobar - 64barfoo - 65foo - 66bar - 67foobar - 68barfoo - 69foo - 70bar - 71foobar - 72barfoo - 73foo - 74bar - 75foobar - 76barfoo - 77foo - 78bar - 79foobar - 80barfoo - 81foo - 82bar - 83foobar - 84barfoo - 85foo - 86bar - 87foobar - 88barfoo - 89foo - 90bar - 91foobar - 92barfoo - 93foo - 94bar - 95foobar - 96barfoo - -----END RSA PRIVATE KEY----- + 1foo + 2bar + 3foobar + 4barfoo + 5foo + 6bar + 7foobar + 8barfoo + 9foo + 10bar + 11foobar + 12barfoo + 13foo + 14bar + 15foobar + 16barfoo + 17foo + 18bar + 19foobar + 20barfoo + 21foo + 22bar + 23foobar + 24barfoo + 25foo + 26bar + 27foobar + 28barfoo + 29foo + 30bar + 31foobar + 32barfoo + 33foo + 34bar + 35foobar + 36barfoo + 37foo + 38bar + 39foobar + 40barfoo + 41foo + 42bar + 43foobar + 44barfoo + 45foo + 46bar + 47foobar + 48barfoo + 49foo + 50bar + 51foobar + 52barfoo + 53foo + 54bar + 55foobar + 56barfoo + 57foo + 58bar + 59foobar + 60barfoo + 61foo + 62bar + 63foobar + 64barfoo + 65foo + 66bar + 67foobar + 68barfoo + 69foo + 70bar + 71foobar + 72barfoo + 73foo + 74bar + 75foobar + 76barfoo + 77foo + 78bar + 79foobar + 80barfoo + 81foo + 82bar + 83foobar + 84barfoo + 85foo + 86bar + 87foobar + 88barfoo + 89foo + 90bar + 91foobar + 92barfoo + 93foo + 94bar + 95foobar + 96barfoo + -----END RSA PRIVATE KEY----- `)) So(err, ShouldNotBeNil) So(f, ShouldBeNil) @@ -1109,6 +1167,44 @@ my lesson state data – 1111111111111111111000000000000000001110000 So(err, ShouldNotBeNil) }) }) + + Convey("And false `SpaceBeforeInlineComment`", func() { + Convey("Can't parse INI files containing `#` or `;` in value", func() { + f, err := ini.LoadSources( + ini.LoadOptions{AllowPythonMultilineValues: true, SpaceBeforeInlineComment: false}, + []byte(` +[author] +NAME = U#n#k#n#w#o#n +GITHUB = U;n;k;n;w;o;n +`)) + So(err, ShouldBeNil) + So(f, ShouldNotBeNil) + sec := f.Section("author") + nameValue := sec.Key("NAME").String() + githubValue := sec.Key("GITHUB").String() + So(nameValue, ShouldEqual, "U") + So(githubValue, ShouldEqual, "U") + }) + }) + + Convey("And true `SpaceBeforeInlineComment`", func() { + Convey("Can parse INI files containing `#` or `;` in value", func() { + f, err := ini.LoadSources( + ini.LoadOptions{AllowPythonMultilineValues: true, SpaceBeforeInlineComment: true}, + []byte(` +[author] +NAME = U#n#k#n#w#o#n +GITHUB = U;n;k;n;w;o;n +`)) + So(err, ShouldBeNil) + So(f, ShouldNotBeNil) + sec := f.Section("author") + nameValue := sec.Key("NAME").String() + githubValue := sec.Key("GITHUB").String() + So(nameValue, ShouldEqual, "U#n#k#n#w#o#n") + So(githubValue, ShouldEqual, "U;n;k;n;w;o;n") + }) + }) }) }) } diff --git a/vendor/github.com/go-ini/ini/parser.go b/vendor/github.com/go-ini/ini/parser.go index 826e893c0..d5aa2db60 100644 --- a/vendor/github.com/go-ini/ini/parser.go +++ b/vendor/github.com/go-ini/ini/parser.go @@ -198,7 +198,7 @@ func hasSurroundedQuote(in string, quote byte) bool { func (p *parser) readValue(in []byte, parserBufferSize int, - ignoreContinuation, ignoreInlineComment, unescapeValueDoubleQuotes, unescapeValueCommentSymbols, allowPythonMultilines bool) (string, error) { + ignoreContinuation, ignoreInlineComment, unescapeValueDoubleQuotes, unescapeValueCommentSymbols, allowPythonMultilines, spaceBeforeInlineComment bool) (string, error) { line := strings.TrimLeftFunc(string(in), unicode.IsSpace) if len(line) == 0 { @@ -240,11 +240,22 @@ func (p *parser) readValue(in []byte, // Check if ignore inline comment if !ignoreInlineComment { - i := strings.IndexAny(line, "#;") + var i int + if spaceBeforeInlineComment { + i = strings.Index(line, " #") + if i == -1 { + i = strings.Index(line, " ;") + } + + } else { + i = strings.IndexAny(line, "#;") + } + if i > -1 { p.comment.WriteString(line[i:]) line = strings.TrimSpace(line[:i]) } + } // Trim single and double quotes @@ -429,7 +440,8 @@ func (f *File) parse(reader io.Reader) (err error) { f.options.IgnoreInlineComment, f.options.UnescapeValueDoubleQuotes, f.options.UnescapeValueCommentSymbols, - f.options.AllowPythonMultilineValues) + f.options.AllowPythonMultilineValues, + f.options.SpaceBeforeInlineComment) if err != nil { return err } @@ -458,7 +470,8 @@ func (f *File) parse(reader io.Reader) (err error) { f.options.IgnoreInlineComment, f.options.UnescapeValueDoubleQuotes, f.options.UnescapeValueCommentSymbols, - f.options.AllowPythonMultilineValues) + f.options.AllowPythonMultilineValues, + f.options.SpaceBeforeInlineComment) if err != nil { return err } diff --git a/vendor/github.com/golang/snappy/cmd/snappytool/main.go b/vendor/github.com/golang/snappy/cmd/snappytool/main.go new file mode 100644 index 000000000..b0f44c3fb --- /dev/null +++ b/vendor/github.com/golang/snappy/cmd/snappytool/main.go @@ -0,0 +1,46 @@ +package main + +import ( + "errors" + "flag" + "io/ioutil" + "os" + + "github.com/golang/snappy" +) + +var ( + decode = flag.Bool("d", false, "decode") + encode = flag.Bool("e", false, "encode") +) + +func run() error { + flag.Parse() + if *decode == *encode { + return errors.New("exactly one of -d or -e must be given") + } + + in, err := ioutil.ReadAll(os.Stdin) + if err != nil { + return err + } + + out := []byte(nil) + if *decode { + out, err = snappy.Decode(nil, in) + if err != nil { + return err + } + } else { + out = snappy.Encode(nil, in) + } + _, err = os.Stdout.Write(out) + return err +} + +func main() { + if err := run(); err != nil { + os.Stderr.WriteString(err.Error() + "\n") + os.Exit(1) + } +} diff --git a/vendor/github.com/golang/snappy/cmd/snappytool/main.cpp b/vendor/github.com/golang/snappy/misc/main.cpp similarity index 97% rename from vendor/github.com/golang/snappy/cmd/snappytool/main.cpp rename to vendor/github.com/golang/snappy/misc/main.cpp index fc31f5173..24a3d9a91 100644 --- a/vendor/github.com/golang/snappy/cmd/snappytool/main.cpp +++ b/vendor/github.com/golang/snappy/misc/main.cpp @@ -1,4 +1,6 @@ /* +This is a C version of the cmd/snappytool Go program. + To build the snappytool binary: g++ main.cpp /usr/lib/libsnappy.a -o snappytool or, if you have built the C++ snappy library from source: diff --git a/vendor/github.com/golang/snappy/snappy.go b/vendor/github.com/golang/snappy/snappy.go index 0cf5e379c..ece692ea4 100644 --- a/vendor/github.com/golang/snappy/snappy.go +++ b/vendor/github.com/golang/snappy/snappy.go @@ -2,10 +2,21 @@ // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. -// Package snappy implements the snappy block-based compression format. -// It aims for very high speeds and reasonable compression. +// Package snappy implements the Snappy compression format. It aims for very +// high speeds and reasonable compression. // -// The C++ snappy implementation is at https://github.com/google/snappy +// There are actually two Snappy formats: block and stream. They are related, +// but different: trying to decompress block-compressed data as a Snappy stream +// will fail, and vice versa. The block format is the Decode and Encode +// functions and the stream format is the Reader and Writer types. +// +// The block format, the more common case, is used when the complete size (the +// number of bytes) of the original data is known upfront, at the time +// compression starts. The stream format, also known as the framing format, is +// for when that isn't always true. +// +// The canonical, C++ implementation is at https://github.com/google/snappy and +// it only implements the block format. package snappy // import "github.com/golang/snappy" import ( diff --git a/vendor/github.com/googleapis/gnostic/.gitignore b/vendor/github.com/googleapis/gnostic/.gitignore index 63149fdda..2a9d09b79 100644 --- a/vendor/github.com/googleapis/gnostic/.gitignore +++ b/vendor/github.com/googleapis/gnostic/.gitignore @@ -1,3 +1,5 @@ +# IntelliJ IDEA +.idea # Eclipse .checkstyle .project @@ -5,6 +7,10 @@ # Swift .build Packages +# Node +node_modules +package-lock.json +bundle.json # vi *.swp # vscode @@ -12,3 +18,4 @@ Packages .DS_Store *~ Package.resolved +extensions/sample/generated diff --git a/vendor/github.com/googleapis/gnostic/.travis.yml b/vendor/github.com/googleapis/gnostic/.travis.yml index d31126d16..c3d2791fc 100644 --- a/vendor/github.com/googleapis/gnostic/.travis.yml +++ b/vendor/github.com/googleapis/gnostic/.travis.yml @@ -32,13 +32,16 @@ script: - pushd plugins/gnostic-go-generator/examples/v2.0/bookstore - make test - popd + - pushd plugins/gnostic-go-generator/examples/v2.0/sample + - make test + - popd - pushd plugins/gnostic-go-generator/examples/v3.0/bookstore - make test - popd - export PATH=.:$HOME/local/bin:$PATH - export LD_LIBRARY_PATH=$HOME/local/lib - pushd plugins/gnostic-swift-generator - - make + - make install - cd examples/bookstore - make - .build/debug/Server & diff --git a/vendor/github.com/googleapis/gnostic/COMPILE-PROTOS.sh b/vendor/github.com/googleapis/gnostic/COMPILE-PROTOS.sh index 017dab670..584878a10 100755 --- a/vendor/github.com/googleapis/gnostic/COMPILE-PROTOS.sh +++ b/vendor/github.com/googleapis/gnostic/COMPILE-PROTOS.sh @@ -22,6 +22,7 @@ protoc \ OpenAPIv2/OpenAPIv2.proto protoc \ +-I.:$GOPATH/src \ --go_out=:. \ plugins/plugin.proto diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go b/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go index 0e32451a3..5351f36f3 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.go @@ -7106,20 +7106,20 @@ func (m *Any) ToRawInfo() interface{} { func (m *ApiKeySecurity) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7130,14 +7130,14 @@ func (m *ApiKeySecurity) ToRawInfo() interface{} { func (m *BasicAuthenticationSecurity) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7148,24 +7148,24 @@ func (m *BasicAuthenticationSecurity) ToRawInfo() interface{} { func (m *BodyParameter) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.Schema != nil { - info = append(info, yaml.MapItem{"schema", m.Schema.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) } // &{Name:schema Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7176,17 +7176,17 @@ func (m *BodyParameter) ToRawInfo() interface{} { func (m *Contact) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.Email != "" { - info = append(info, yaml.MapItem{"email", m.Email}) + info = append(info, yaml.MapItem{Key: "email", Value: m.Email}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7198,7 +7198,7 @@ func (m *Default) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:false Description:} @@ -7210,7 +7210,7 @@ func (m *Definitions) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSchema StringEnumValues:[] MapType:Schema Repeated:true Pattern: Implicit:true Description:} @@ -7221,41 +7221,41 @@ func (m *Definitions) ToRawInfo() interface{} { func (m *Document) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Swagger != "" { - info = append(info, yaml.MapItem{"swagger", m.Swagger}) + info = append(info, yaml.MapItem{Key: "swagger", Value: m.Swagger}) } if m.Info != nil { - info = append(info, yaml.MapItem{"info", m.Info.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "info", Value: m.Info.ToRawInfo()}) } // &{Name:info Type:Info StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Host != "" { - info = append(info, yaml.MapItem{"host", m.Host}) + info = append(info, yaml.MapItem{Key: "host", Value: m.Host}) } if m.BasePath != "" { - info = append(info, yaml.MapItem{"basePath", m.BasePath}) + info = append(info, yaml.MapItem{Key: "basePath", Value: m.BasePath}) } if len(m.Schemes) != 0 { - info = append(info, yaml.MapItem{"schemes", m.Schemes}) + info = append(info, yaml.MapItem{Key: "schemes", Value: m.Schemes}) } if len(m.Consumes) != 0 { - info = append(info, yaml.MapItem{"consumes", m.Consumes}) + info = append(info, yaml.MapItem{Key: "consumes", Value: m.Consumes}) } if len(m.Produces) != 0 { - info = append(info, yaml.MapItem{"produces", m.Produces}) + info = append(info, yaml.MapItem{Key: "produces", Value: m.Produces}) } if m.Paths != nil { - info = append(info, yaml.MapItem{"paths", m.Paths.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "paths", Value: m.Paths.ToRawInfo()}) } // &{Name:paths Type:Paths StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Definitions != nil { - info = append(info, yaml.MapItem{"definitions", m.Definitions.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "definitions", Value: m.Definitions.ToRawInfo()}) } // &{Name:definitions Type:Definitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Parameters != nil { - info = append(info, yaml.MapItem{"parameters", m.Parameters.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "parameters", Value: m.Parameters.ToRawInfo()}) } // &{Name:parameters Type:ParameterDefinitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Responses != nil { - info = append(info, yaml.MapItem{"responses", m.Responses.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "responses", Value: m.Responses.ToRawInfo()}) } // &{Name:responses Type:ResponseDefinitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Security) != 0 { @@ -7263,11 +7263,11 @@ func (m *Document) ToRawInfo() interface{} { for _, item := range m.Security { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"security", items}) + info = append(info, yaml.MapItem{Key: "security", Value: items}) } // &{Name:security Type:SecurityRequirement StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.SecurityDefinitions != nil { - info = append(info, yaml.MapItem{"securityDefinitions", m.SecurityDefinitions.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "securityDefinitions", Value: m.SecurityDefinitions.ToRawInfo()}) } // &{Name:securityDefinitions Type:SecurityDefinitions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Tags) != 0 { @@ -7275,16 +7275,16 @@ func (m *Document) ToRawInfo() interface{} { for _, item := range m.Tags { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"tags", items}) + info = append(info, yaml.MapItem{Key: "tags", Value: items}) } // &{Name:tags Type:Tag StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7296,7 +7296,7 @@ func (m *Examples) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:true Description:} @@ -7307,14 +7307,14 @@ func (m *Examples) ToRawInfo() interface{} { func (m *ExternalDocs) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7325,38 +7325,38 @@ func (m *ExternalDocs) ToRawInfo() interface{} { func (m *FileSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Title != "" { - info = append(info, yaml.MapItem{"title", m.Title}) + info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Required) != 0 { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.ReadOnly != false { - info = append(info, yaml.MapItem{"readOnly", m.ReadOnly}) + info = append(info, yaml.MapItem{Key: "readOnly", Value: m.ReadOnly}) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{"example", m.Example.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) } // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7367,81 +7367,81 @@ func (m *FileSchema) ToRawInfo() interface{} { func (m *FormDataParameterSubSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.AllowEmptyValue != false { - info = append(info, yaml.MapItem{"allowEmptyValue", m.AllowEmptyValue}) + info = append(info, yaml.MapItem{Key: "allowEmptyValue", Value: m.AllowEmptyValue}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{"collectionFormat", m.CollectionFormat}) + info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7452,69 +7452,69 @@ func (m *FormDataParameterSubSchema) ToRawInfo() interface{} { func (m *Header) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{"collectionFormat", m.CollectionFormat}) + info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7525,78 +7525,78 @@ func (m *Header) ToRawInfo() interface{} { func (m *HeaderParameterSubSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{"collectionFormat", m.CollectionFormat}) + info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7608,7 +7608,7 @@ func (m *Headers) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedHeader StringEnumValues:[] MapType:Header Repeated:true Pattern: Implicit:true Description:} @@ -7619,28 +7619,28 @@ func (m *Headers) ToRawInfo() interface{} { func (m *Info) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Title != "" { - info = append(info, yaml.MapItem{"title", m.Title}) + info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) } if m.Version != "" { - info = append(info, yaml.MapItem{"version", m.Version}) + info = append(info, yaml.MapItem{Key: "version", Value: m.Version}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.TermsOfService != "" { - info = append(info, yaml.MapItem{"termsOfService", m.TermsOfService}) + info = append(info, yaml.MapItem{Key: "termsOfService", Value: m.TermsOfService}) } if m.Contact != nil { - info = append(info, yaml.MapItem{"contact", m.Contact.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "contact", Value: m.Contact.ToRawInfo()}) } // &{Name:contact Type:Contact StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.License != nil { - info = append(info, yaml.MapItem{"license", m.License.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "license", Value: m.License.ToRawInfo()}) } // &{Name:license Type:License StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7655,7 +7655,7 @@ func (m *ItemsItem) ToRawInfo() interface{} { for _, item := range m.Schema { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"schema", items}) + info = append(info, yaml.MapItem{Key: "schema", Value: items}) } // &{Name:schema Type:Schema StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} return info @@ -7665,10 +7665,10 @@ func (m *ItemsItem) ToRawInfo() interface{} { func (m *JsonReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } return info } @@ -7677,14 +7677,14 @@ func (m *JsonReference) ToRawInfo() interface{} { func (m *License) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7695,7 +7695,7 @@ func (m *License) ToRawInfo() interface{} { func (m *NamedAny) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7705,7 +7705,7 @@ func (m *NamedAny) ToRawInfo() interface{} { func (m *NamedHeader) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Header StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7715,7 +7715,7 @@ func (m *NamedHeader) ToRawInfo() interface{} { func (m *NamedParameter) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Parameter StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7725,7 +7725,7 @@ func (m *NamedParameter) ToRawInfo() interface{} { func (m *NamedPathItem) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:PathItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7735,7 +7735,7 @@ func (m *NamedPathItem) ToRawInfo() interface{} { func (m *NamedResponse) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Response StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7745,7 +7745,7 @@ func (m *NamedResponse) ToRawInfo() interface{} { func (m *NamedResponseValue) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:ResponseValue StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7755,7 +7755,7 @@ func (m *NamedResponseValue) ToRawInfo() interface{} { func (m *NamedSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7765,7 +7765,7 @@ func (m *NamedSchema) ToRawInfo() interface{} { func (m *NamedSecurityDefinitionsItem) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:SecurityDefinitionsItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7775,10 +7775,10 @@ func (m *NamedSecurityDefinitionsItem) ToRawInfo() interface{} { func (m *NamedString) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Value != "" { - info = append(info, yaml.MapItem{"value", m.Value}) + info = append(info, yaml.MapItem{Key: "value", Value: m.Value}) } return info } @@ -7787,7 +7787,7 @@ func (m *NamedString) ToRawInfo() interface{} { func (m *NamedStringArray) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:StringArray StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7824,27 +7824,27 @@ func (m *NonBodyParameter) ToRawInfo() interface{} { func (m *Oauth2AccessCodeSecurity) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Flow != "" { - info = append(info, yaml.MapItem{"flow", m.Flow}) + info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) } if m.Scopes != nil { - info = append(info, yaml.MapItem{"scopes", m.Scopes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) } // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.AuthorizationUrl != "" { - info = append(info, yaml.MapItem{"authorizationUrl", m.AuthorizationUrl}) + info = append(info, yaml.MapItem{Key: "authorizationUrl", Value: m.AuthorizationUrl}) } if m.TokenUrl != "" { - info = append(info, yaml.MapItem{"tokenUrl", m.TokenUrl}) + info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7855,24 +7855,24 @@ func (m *Oauth2AccessCodeSecurity) ToRawInfo() interface{} { func (m *Oauth2ApplicationSecurity) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Flow != "" { - info = append(info, yaml.MapItem{"flow", m.Flow}) + info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) } if m.Scopes != nil { - info = append(info, yaml.MapItem{"scopes", m.Scopes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) } // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.TokenUrl != "" { - info = append(info, yaml.MapItem{"tokenUrl", m.TokenUrl}) + info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7883,24 +7883,24 @@ func (m *Oauth2ApplicationSecurity) ToRawInfo() interface{} { func (m *Oauth2ImplicitSecurity) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Flow != "" { - info = append(info, yaml.MapItem{"flow", m.Flow}) + info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) } if m.Scopes != nil { - info = append(info, yaml.MapItem{"scopes", m.Scopes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) } // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.AuthorizationUrl != "" { - info = append(info, yaml.MapItem{"authorizationUrl", m.AuthorizationUrl}) + info = append(info, yaml.MapItem{Key: "authorizationUrl", Value: m.AuthorizationUrl}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7911,24 +7911,24 @@ func (m *Oauth2ImplicitSecurity) ToRawInfo() interface{} { func (m *Oauth2PasswordSecurity) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Flow != "" { - info = append(info, yaml.MapItem{"flow", m.Flow}) + info = append(info, yaml.MapItem{Key: "flow", Value: m.Flow}) } if m.Scopes != nil { - info = append(info, yaml.MapItem{"scopes", m.Scopes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) } // &{Name:scopes Type:Oauth2Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.TokenUrl != "" { - info = append(info, yaml.MapItem{"tokenUrl", m.TokenUrl}) + info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7946,56 +7946,56 @@ func (m *Oauth2Scopes) ToRawInfo() interface{} { func (m *Operation) ToRawInfo() interface{} { info := yaml.MapSlice{} if len(m.Tags) != 0 { - info = append(info, yaml.MapItem{"tags", m.Tags}) + info = append(info, yaml.MapItem{Key: "tags", Value: m.Tags}) } if m.Summary != "" { - info = append(info, yaml.MapItem{"summary", m.Summary}) + info = append(info, yaml.MapItem{Key: "summary", Value: m.Summary}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.OperationId != "" { - info = append(info, yaml.MapItem{"operationId", m.OperationId}) + info = append(info, yaml.MapItem{Key: "operationId", Value: m.OperationId}) } if len(m.Produces) != 0 { - info = append(info, yaml.MapItem{"produces", m.Produces}) + info = append(info, yaml.MapItem{Key: "produces", Value: m.Produces}) } if len(m.Consumes) != 0 { - info = append(info, yaml.MapItem{"consumes", m.Consumes}) + info = append(info, yaml.MapItem{Key: "consumes", Value: m.Consumes}) } if len(m.Parameters) != 0 { items := make([]interface{}, 0) for _, item := range m.Parameters { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"parameters", items}) + info = append(info, yaml.MapItem{Key: "parameters", Value: items}) } // &{Name:parameters Type:ParametersItem StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:The parameters needed to send a valid API call.} if m.Responses != nil { - info = append(info, yaml.MapItem{"responses", m.Responses.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "responses", Value: m.Responses.ToRawInfo()}) } // &{Name:responses Type:Responses StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Schemes) != 0 { - info = append(info, yaml.MapItem{"schemes", m.Schemes}) + info = append(info, yaml.MapItem{Key: "schemes", Value: m.Schemes}) } if m.Deprecated != false { - info = append(info, yaml.MapItem{"deprecated", m.Deprecated}) + info = append(info, yaml.MapItem{Key: "deprecated", Value: m.Deprecated}) } if len(m.Security) != 0 { items := make([]interface{}, 0) for _, item := range m.Security { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"security", items}) + info = append(info, yaml.MapItem{Key: "security", Value: items}) } // &{Name:security Type:SecurityRequirement StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8024,7 +8024,7 @@ func (m *ParameterDefinitions) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedParameter StringEnumValues:[] MapType:Parameter Repeated:true Pattern: Implicit:true Description:} @@ -8052,34 +8052,34 @@ func (m *ParametersItem) ToRawInfo() interface{} { func (m *PathItem) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.Get != nil { - info = append(info, yaml.MapItem{"get", m.Get.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "get", Value: m.Get.ToRawInfo()}) } // &{Name:get Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Put != nil { - info = append(info, yaml.MapItem{"put", m.Put.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "put", Value: m.Put.ToRawInfo()}) } // &{Name:put Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Post != nil { - info = append(info, yaml.MapItem{"post", m.Post.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "post", Value: m.Post.ToRawInfo()}) } // &{Name:post Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Delete != nil { - info = append(info, yaml.MapItem{"delete", m.Delete.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "delete", Value: m.Delete.ToRawInfo()}) } // &{Name:delete Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Options != nil { - info = append(info, yaml.MapItem{"options", m.Options.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "options", Value: m.Options.ToRawInfo()}) } // &{Name:options Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Head != nil { - info = append(info, yaml.MapItem{"head", m.Head.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "head", Value: m.Head.ToRawInfo()}) } // &{Name:head Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Patch != nil { - info = append(info, yaml.MapItem{"patch", m.Patch.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "patch", Value: m.Patch.ToRawInfo()}) } // &{Name:patch Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Parameters) != 0 { @@ -8087,12 +8087,12 @@ func (m *PathItem) ToRawInfo() interface{} { for _, item := range m.Parameters { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"parameters", items}) + info = append(info, yaml.MapItem{Key: "parameters", Value: items}) } // &{Name:parameters Type:ParametersItem StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:The parameters needed to send a valid API call.} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8103,78 +8103,78 @@ func (m *PathItem) ToRawInfo() interface{} { func (m *PathParameterSubSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{"collectionFormat", m.CollectionFormat}) + info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8186,13 +8186,13 @@ func (m *Paths) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} if m.Path != nil { for _, item := range m.Path { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:Path Type:NamedPathItem StringEnumValues:[] MapType:PathItem Repeated:true Pattern:^/ Implicit:true Description:} @@ -8203,66 +8203,66 @@ func (m *Paths) ToRawInfo() interface{} { func (m *PrimitivesItems) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{"collectionFormat", m.CollectionFormat}) + info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8274,7 +8274,7 @@ func (m *Properties) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSchema StringEnumValues:[] MapType:Schema Repeated:true Pattern: Implicit:true Description:} @@ -8285,81 +8285,81 @@ func (m *Properties) ToRawInfo() interface{} { func (m *QueryParameterSubSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.AllowEmptyValue != false { - info = append(info, yaml.MapItem{"allowEmptyValue", m.AllowEmptyValue}) + info = append(info, yaml.MapItem{Key: "allowEmptyValue", Value: m.AllowEmptyValue}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:PrimitivesItems StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.CollectionFormat != "" { - info = append(info, yaml.MapItem{"collectionFormat", m.CollectionFormat}) + info = append(info, yaml.MapItem{Key: "collectionFormat", Value: m.CollectionFormat}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8370,23 +8370,23 @@ func (m *QueryParameterSubSchema) ToRawInfo() interface{} { func (m *Response) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Schema != nil { - info = append(info, yaml.MapItem{"schema", m.Schema.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) } // &{Name:schema Type:SchemaItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Headers != nil { - info = append(info, yaml.MapItem{"headers", m.Headers.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "headers", Value: m.Headers.ToRawInfo()}) } // &{Name:headers Type:Headers StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Examples != nil { - info = append(info, yaml.MapItem{"examples", m.Examples.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "examples", Value: m.Examples.ToRawInfo()}) } // &{Name:examples Type:Examples StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8398,7 +8398,7 @@ func (m *ResponseDefinitions) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedResponse StringEnumValues:[] MapType:Response Repeated:true Pattern: Implicit:true Description:} @@ -8427,13 +8427,13 @@ func (m *Responses) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.ResponseCode != nil { for _, item := range m.ResponseCode { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:ResponseCode Type:NamedResponseValue StringEnumValues:[] MapType:ResponseValue Repeated:true Pattern:^([0-9]{3})$|^(default)$ Implicit:true Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8444,80 +8444,80 @@ func (m *Responses) ToRawInfo() interface{} { func (m *Schema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Title != "" { - info = append(info, yaml.MapItem{"title", m.Title}) + info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if m.MaxProperties != 0 { - info = append(info, yaml.MapItem{"maxProperties", m.MaxProperties}) + info = append(info, yaml.MapItem{Key: "maxProperties", Value: m.MaxProperties}) } if m.MinProperties != 0 { - info = append(info, yaml.MapItem{"minProperties", m.MinProperties}) + info = append(info, yaml.MapItem{Key: "minProperties", Value: m.MinProperties}) } if len(m.Required) != 0 { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.AdditionalProperties != nil { - info = append(info, yaml.MapItem{"additionalProperties", m.AdditionalProperties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "additionalProperties", Value: m.AdditionalProperties.ToRawInfo()}) } // &{Name:additionalProperties Type:AdditionalPropertiesItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Type != nil { if len(m.Type.Value) == 1 { - info = append(info, yaml.MapItem{"type", m.Type.Value[0]}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type.Value[0]}) } else { - info = append(info, yaml.MapItem{"type", m.Type.Value}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type.Value}) } } // &{Name:type Type:TypeItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} @@ -8526,7 +8526,7 @@ func (m *Schema) ToRawInfo() interface{} { for _, item := range m.Items.Schema { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"items", items[0]}) + info = append(info, yaml.MapItem{Key: "items", Value: items[0]}) } // &{Name:items Type:ItemsItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.AllOf) != 0 { @@ -8534,34 +8534,34 @@ func (m *Schema) ToRawInfo() interface{} { for _, item := range m.AllOf { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"allOf", items}) + info = append(info, yaml.MapItem{Key: "allOf", Value: items}) } // &{Name:allOf Type:Schema StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.Properties != nil { - info = append(info, yaml.MapItem{"properties", m.Properties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "properties", Value: m.Properties.ToRawInfo()}) } // &{Name:properties Type:Properties StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Discriminator != "" { - info = append(info, yaml.MapItem{"discriminator", m.Discriminator}) + info = append(info, yaml.MapItem{Key: "discriminator", Value: m.Discriminator}) } if m.ReadOnly != false { - info = append(info, yaml.MapItem{"readOnly", m.ReadOnly}) + info = append(info, yaml.MapItem{Key: "readOnly", Value: m.ReadOnly}) } if m.Xml != nil { - info = append(info, yaml.MapItem{"xml", m.Xml.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "xml", Value: m.Xml.ToRawInfo()}) } // &{Name:xml Type:Xml StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{"example", m.Example.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) } // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8590,7 +8590,7 @@ func (m *SecurityDefinitions) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSecurityDefinitionsItem StringEnumValues:[] MapType:SecurityDefinitionsItem Repeated:true Pattern: Implicit:true Description:} @@ -8639,7 +8639,7 @@ func (m *SecurityRequirement) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedStringArray StringEnumValues:[] MapType:StringArray Repeated:true Pattern: Implicit:true Description:} @@ -8655,18 +8655,18 @@ func (m *StringArray) ToRawInfo() interface{} { func (m *Tag) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8677,7 +8677,7 @@ func (m *Tag) ToRawInfo() interface{} { func (m *TypeItem) ToRawInfo() interface{} { info := yaml.MapSlice{} if len(m.Value) != 0 { - info = append(info, yaml.MapItem{"value", m.Value}) + info = append(info, yaml.MapItem{Key: "value", Value: m.Value}) } return info } @@ -8687,7 +8687,7 @@ func (m *VendorExtension) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:true Description:} @@ -8698,23 +8698,23 @@ func (m *VendorExtension) ToRawInfo() interface{} { func (m *Xml) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Namespace != "" { - info = append(info, yaml.MapItem{"namespace", m.Namespace}) + info = append(info, yaml.MapItem{Key: "namespace", Value: m.Namespace}) } if m.Prefix != "" { - info = append(info, yaml.MapItem{"prefix", m.Prefix}) + info = append(info, yaml.MapItem{Key: "prefix", Value: m.Prefix}) } if m.Attribute != false { - info = append(info, yaml.MapItem{"attribute", m.Attribute}) + info = append(info, yaml.MapItem{Key: "attribute", Value: m.Attribute}) } if m.Wrapped != false { - info = append(info, yaml.MapItem{"wrapped", m.Wrapped}) + info = append(info, yaml.MapItem{Key: "wrapped", Value: m.Wrapped}) } if m.VendorExtension != nil { for _, item := range m.VendorExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:VendorExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go b/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go index 37da7df25..a030fa676 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-go. +// Code generated by protoc-gen-go. DO NOT EDIT. // source: OpenAPIv2/OpenAPIv2.proto -// DO NOT EDIT! /* Package openapi_v2 is a generated protocol buffer package. @@ -4257,7 +4256,7 @@ func init() { proto.RegisterFile("OpenAPIv2/OpenAPIv2.proto", fileDescriptor0) } var fileDescriptor0 = []byte{ // 3129 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x09, 0x6e, 0x88, 0x02, 0xff, 0xec, 0x3b, 0x4b, 0x73, 0x1c, 0x57, + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xec, 0x3b, 0x4b, 0x73, 0x1c, 0x57, 0xd5, 0xf3, 0x7e, 0x1c, 0x69, 0x46, 0xa3, 0x96, 0x2c, 0xb7, 0x24, 0xc7, 0x71, 0xe4, 0x3c, 0x6c, 0xe7, 0xb3, 0x9c, 0x4f, 0x29, 0x48, 0x05, 0x2a, 0x05, 0xf2, 0xab, 0xc6, 0xc4, 0x44, 0x4a, 0xcb, 0x0e, 0x09, 0x04, 0xba, 0xae, 0x66, 0xee, 0x48, 0x9d, 0x74, 0xf7, 0x6d, 0x77, 0xf7, 0xc8, 0x1a, diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.go b/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.go index 2559f8237..5c8ed9c3e 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.go +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.go @@ -971,26 +971,6 @@ func NewExampleOrReference(in interface{}, context *compiler.Context) (*ExampleO return x, compiler.NewErrorGroupOrNil(errors) } -// NewExamples creates an object of type Examples if possible, returning an error if not. -func NewExamples(in interface{}, context *compiler.Context) (*Examples, error) { - errors := make([]error, 0) - x := &Examples{} - m, ok := compiler.UnpackMap(in) - if !ok { - message := fmt.Sprintf("has unexpected value: %+v (%T)", in, in) - errors = append(errors, compiler.NewError(context, message)) - } else { - allowedKeys := []string{} - var allowedPatterns []*regexp.Regexp - invalidKeys := compiler.InvalidKeysInMap(m, allowedKeys, allowedPatterns) - if len(invalidKeys) > 0 { - message := fmt.Sprintf("has invalid %s: %+v", compiler.PluralProperties(len(invalidKeys)), strings.Join(invalidKeys, ", ")) - errors = append(errors, compiler.NewError(context, message)) - } - } - return x, compiler.NewErrorGroupOrNil(errors) -} - // NewExamplesOrReferences creates an object of type ExamplesOrReferences if possible, returning an error if not. func NewExamplesOrReferences(in interface{}, context *compiler.Context) (*ExamplesOrReferences, error) { errors := make([]error, 0) @@ -5380,12 +5360,6 @@ func (m *ExampleOrReference) ResolveReferences(root string) (interface{}, error) return nil, compiler.NewErrorGroupOrNil(errors) } -// ResolveReferences resolves references found inside Examples objects. -func (m *Examples) ResolveReferences(root string) (interface{}, error) { - errors := make([]error, 0) - return nil, compiler.NewErrorGroupOrNil(errors) -} - // ResolveReferences resolves references found inside ExamplesOrReferences objects. func (m *ExamplesOrReferences) ResolveReferences(root string) (interface{}, error) { errors := make([]error, 0) @@ -6742,7 +6716,7 @@ func (m *AnysOrExpressions) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedAnyOrExpression StringEnumValues:[] MapType:AnyOrExpression Repeated:true Pattern: Implicit:true Description:} @@ -6754,13 +6728,13 @@ func (m *Callback) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Path != nil { for _, item := range m.Path { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:Path Type:NamedPathItem StringEnumValues:[] MapType:PathItem Repeated:true Pattern:^ Implicit:true Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -6789,7 +6763,7 @@ func (m *CallbacksOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedCallbackOrReference StringEnumValues:[] MapType:CallbackOrReference Repeated:true Pattern: Implicit:true Description:} @@ -6800,44 +6774,44 @@ func (m *CallbacksOrReferences) ToRawInfo() interface{} { func (m *Components) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Schemas != nil { - info = append(info, yaml.MapItem{"schemas", m.Schemas.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schemas", Value: m.Schemas.ToRawInfo()}) } // &{Name:schemas Type:SchemasOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Responses != nil { - info = append(info, yaml.MapItem{"responses", m.Responses.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "responses", Value: m.Responses.ToRawInfo()}) } // &{Name:responses Type:ResponsesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Parameters != nil { - info = append(info, yaml.MapItem{"parameters", m.Parameters.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "parameters", Value: m.Parameters.ToRawInfo()}) } // &{Name:parameters Type:ParametersOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Examples != nil { - info = append(info, yaml.MapItem{"examples", m.Examples.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "examples", Value: m.Examples.ToRawInfo()}) } // &{Name:examples Type:ExamplesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.RequestBodies != nil { - info = append(info, yaml.MapItem{"requestBodies", m.RequestBodies.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "requestBodies", Value: m.RequestBodies.ToRawInfo()}) } // &{Name:requestBodies Type:RequestBodiesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Headers != nil { - info = append(info, yaml.MapItem{"headers", m.Headers.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "headers", Value: m.Headers.ToRawInfo()}) } // &{Name:headers Type:HeadersOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SecuritySchemes != nil { - info = append(info, yaml.MapItem{"securitySchemes", m.SecuritySchemes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "securitySchemes", Value: m.SecuritySchemes.ToRawInfo()}) } // &{Name:securitySchemes Type:SecuritySchemesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Links != nil { - info = append(info, yaml.MapItem{"links", m.Links.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "links", Value: m.Links.ToRawInfo()}) } // &{Name:links Type:LinksOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Callbacks != nil { - info = append(info, yaml.MapItem{"callbacks", m.Callbacks.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "callbacks", Value: m.Callbacks.ToRawInfo()}) } // &{Name:callbacks Type:CallbacksOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -6848,17 +6822,17 @@ func (m *Components) ToRawInfo() interface{} { func (m *Contact) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.Email != "" { - info = append(info, yaml.MapItem{"email", m.Email}) + info = append(info, yaml.MapItem{Key: "email", Value: m.Email}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -6888,10 +6862,10 @@ func (m *DefaultType) ToRawInfo() interface{} { func (m *Discriminator) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.PropertyName != "" { - info = append(info, yaml.MapItem{"propertyName", m.PropertyName}) + info = append(info, yaml.MapItem{Key: "propertyName", Value: m.PropertyName}) } if m.Mapping != nil { - info = append(info, yaml.MapItem{"mapping", m.Mapping.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "mapping", Value: m.Mapping.ToRawInfo()}) } // &{Name:mapping Type:Strings StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} return info @@ -6901,10 +6875,10 @@ func (m *Discriminator) ToRawInfo() interface{} { func (m *Document) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Openapi != "" { - info = append(info, yaml.MapItem{"openapi", m.Openapi}) + info = append(info, yaml.MapItem{Key: "openapi", Value: m.Openapi}) } if m.Info != nil { - info = append(info, yaml.MapItem{"info", m.Info.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "info", Value: m.Info.ToRawInfo()}) } // &{Name:info Type:Info StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Servers) != 0 { @@ -6912,15 +6886,15 @@ func (m *Document) ToRawInfo() interface{} { for _, item := range m.Servers { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"servers", items}) + info = append(info, yaml.MapItem{Key: "servers", Value: items}) } // &{Name:servers Type:Server StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.Paths != nil { - info = append(info, yaml.MapItem{"paths", m.Paths.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "paths", Value: m.Paths.ToRawInfo()}) } // &{Name:paths Type:Paths StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Components != nil { - info = append(info, yaml.MapItem{"components", m.Components.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "components", Value: m.Components.ToRawInfo()}) } // &{Name:components Type:Components StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Security) != 0 { @@ -6928,7 +6902,7 @@ func (m *Document) ToRawInfo() interface{} { for _, item := range m.Security { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"security", items}) + info = append(info, yaml.MapItem{Key: "security", Value: items}) } // &{Name:security Type:SecurityRequirement StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if len(m.Tags) != 0 { @@ -6936,16 +6910,16 @@ func (m *Document) ToRawInfo() interface{} { for _, item := range m.Tags { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"tags", items}) + info = append(info, yaml.MapItem{Key: "tags", Value: items}) } // &{Name:tags Type:Tag StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -6956,24 +6930,24 @@ func (m *Document) ToRawInfo() interface{} { func (m *Encoding) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.ContentType != "" { - info = append(info, yaml.MapItem{"contentType", m.ContentType}) + info = append(info, yaml.MapItem{Key: "contentType", Value: m.ContentType}) } if m.Headers != nil { - info = append(info, yaml.MapItem{"headers", m.Headers.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "headers", Value: m.Headers.ToRawInfo()}) } // &{Name:headers Type:HeadersOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Style != "" { - info = append(info, yaml.MapItem{"style", m.Style}) + info = append(info, yaml.MapItem{Key: "style", Value: m.Style}) } if m.Explode != false { - info = append(info, yaml.MapItem{"explode", m.Explode}) + info = append(info, yaml.MapItem{Key: "explode", Value: m.Explode}) } if m.AllowReserved != false { - info = append(info, yaml.MapItem{"allowReserved", m.AllowReserved}) + info = append(info, yaml.MapItem{Key: "allowReserved", Value: m.AllowReserved}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -6985,7 +6959,7 @@ func (m *Encodings) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedEncoding StringEnumValues:[] MapType:Encoding Repeated:true Pattern: Implicit:true Description:} @@ -6996,18 +6970,18 @@ func (m *Encodings) ToRawInfo() interface{} { func (m *Example) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Summary != "" { - info = append(info, yaml.MapItem{"summary", m.Summary}) + info = append(info, yaml.MapItem{Key: "summary", Value: m.Summary}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } // &{Name:value Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ExternalValue != "" { - info = append(info, yaml.MapItem{"externalValue", m.ExternalValue}) + info = append(info, yaml.MapItem{Key: "externalValue", Value: m.ExternalValue}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7031,18 +7005,12 @@ func (m *ExampleOrReference) ToRawInfo() interface{} { return nil } -// ToRawInfo returns a description of Examples suitable for JSON or YAML export. -func (m *Examples) ToRawInfo() interface{} { - info := yaml.MapSlice{} - return info -} - // ToRawInfo returns a description of ExamplesOrReferences suitable for JSON or YAML export. func (m *ExamplesOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedExampleOrReference StringEnumValues:[] MapType:ExampleOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7054,7 +7022,7 @@ func (m *Expression) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:true Description:} @@ -7065,14 +7033,14 @@ func (m *Expression) ToRawInfo() interface{} { func (m *ExternalDocs) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7083,45 +7051,45 @@ func (m *ExternalDocs) ToRawInfo() interface{} { func (m *Header) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.Deprecated != false { - info = append(info, yaml.MapItem{"deprecated", m.Deprecated}) + info = append(info, yaml.MapItem{Key: "deprecated", Value: m.Deprecated}) } if m.AllowEmptyValue != false { - info = append(info, yaml.MapItem{"allowEmptyValue", m.AllowEmptyValue}) + info = append(info, yaml.MapItem{Key: "allowEmptyValue", Value: m.AllowEmptyValue}) } if m.Style != "" { - info = append(info, yaml.MapItem{"style", m.Style}) + info = append(info, yaml.MapItem{Key: "style", Value: m.Style}) } if m.Explode != false { - info = append(info, yaml.MapItem{"explode", m.Explode}) + info = append(info, yaml.MapItem{Key: "explode", Value: m.Explode}) } if m.AllowReserved != false { - info = append(info, yaml.MapItem{"allowReserved", m.AllowReserved}) + info = append(info, yaml.MapItem{Key: "allowReserved", Value: m.AllowReserved}) } if m.Schema != nil { - info = append(info, yaml.MapItem{"schema", m.Schema.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) } // &{Name:schema Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{"example", m.Example.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) } // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Examples != nil { - info = append(info, yaml.MapItem{"examples", m.Examples.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "examples", Value: m.Examples.ToRawInfo()}) } // &{Name:examples Type:ExamplesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Content != nil { - info = append(info, yaml.MapItem{"content", m.Content.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "content", Value: m.Content.ToRawInfo()}) } // &{Name:content Type:MediaTypes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7150,7 +7118,7 @@ func (m *HeadersOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedHeaderOrReference StringEnumValues:[] MapType:HeaderOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7161,28 +7129,28 @@ func (m *HeadersOrReferences) ToRawInfo() interface{} { func (m *Info) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Title != "" { - info = append(info, yaml.MapItem{"title", m.Title}) + info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.TermsOfService != "" { - info = append(info, yaml.MapItem{"termsOfService", m.TermsOfService}) + info = append(info, yaml.MapItem{Key: "termsOfService", Value: m.TermsOfService}) } if m.Contact != nil { - info = append(info, yaml.MapItem{"contact", m.Contact.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "contact", Value: m.Contact.ToRawInfo()}) } // &{Name:contact Type:Contact StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.License != nil { - info = append(info, yaml.MapItem{"license", m.License.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "license", Value: m.License.ToRawInfo()}) } // &{Name:license Type:License StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Version != "" { - info = append(info, yaml.MapItem{"version", m.Version}) + info = append(info, yaml.MapItem{Key: "version", Value: m.Version}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7197,7 +7165,7 @@ func (m *ItemsItem) ToRawInfo() interface{} { for _, item := range m.SchemaOrReference { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"schemaOrReference", items}) + info = append(info, yaml.MapItem{Key: "schemaOrReference", Value: items}) } // &{Name:schemaOrReference Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} return info @@ -7207,14 +7175,14 @@ func (m *ItemsItem) ToRawInfo() interface{} { func (m *License) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7225,29 +7193,29 @@ func (m *License) ToRawInfo() interface{} { func (m *Link) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.OperationRef != "" { - info = append(info, yaml.MapItem{"operationRef", m.OperationRef}) + info = append(info, yaml.MapItem{Key: "operationRef", Value: m.OperationRef}) } if m.OperationId != "" { - info = append(info, yaml.MapItem{"operationId", m.OperationId}) + info = append(info, yaml.MapItem{Key: "operationId", Value: m.OperationId}) } if m.Parameters != nil { - info = append(info, yaml.MapItem{"parameters", m.Parameters.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "parameters", Value: m.Parameters.ToRawInfo()}) } // &{Name:parameters Type:AnysOrExpressions StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.RequestBody != nil { - info = append(info, yaml.MapItem{"requestBody", m.RequestBody.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "requestBody", Value: m.RequestBody.ToRawInfo()}) } // &{Name:requestBody Type:AnyOrExpression StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Server != nil { - info = append(info, yaml.MapItem{"server", m.Server.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "server", Value: m.Server.ToRawInfo()}) } // &{Name:server Type:Server StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7276,7 +7244,7 @@ func (m *LinksOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedLinkOrReference StringEnumValues:[] MapType:LinkOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7287,24 +7255,24 @@ func (m *LinksOrReferences) ToRawInfo() interface{} { func (m *MediaType) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Schema != nil { - info = append(info, yaml.MapItem{"schema", m.Schema.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) } // &{Name:schema Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{"example", m.Example.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) } // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Examples != nil { - info = append(info, yaml.MapItem{"examples", m.Examples.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "examples", Value: m.Examples.ToRawInfo()}) } // &{Name:examples Type:ExamplesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Encoding != nil { - info = append(info, yaml.MapItem{"encoding", m.Encoding.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "encoding", Value: m.Encoding.ToRawInfo()}) } // &{Name:encoding Type:Encodings StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7316,7 +7284,7 @@ func (m *MediaTypes) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedMediaType StringEnumValues:[] MapType:MediaType Repeated:true Pattern: Implicit:true Description:} @@ -7327,7 +7295,7 @@ func (m *MediaTypes) ToRawInfo() interface{} { func (m *NamedAny) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7337,7 +7305,7 @@ func (m *NamedAny) ToRawInfo() interface{} { func (m *NamedAnyOrExpression) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:AnyOrExpression StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7347,7 +7315,7 @@ func (m *NamedAnyOrExpression) ToRawInfo() interface{} { func (m *NamedCallbackOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:CallbackOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7357,7 +7325,7 @@ func (m *NamedCallbackOrReference) ToRawInfo() interface{} { func (m *NamedEncoding) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Encoding StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7367,7 +7335,7 @@ func (m *NamedEncoding) ToRawInfo() interface{} { func (m *NamedExampleOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:ExampleOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7377,7 +7345,7 @@ func (m *NamedExampleOrReference) ToRawInfo() interface{} { func (m *NamedHeaderOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:HeaderOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7387,7 +7355,7 @@ func (m *NamedHeaderOrReference) ToRawInfo() interface{} { func (m *NamedLinkOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:LinkOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7397,7 +7365,7 @@ func (m *NamedLinkOrReference) ToRawInfo() interface{} { func (m *NamedMediaType) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:MediaType StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7407,7 +7375,7 @@ func (m *NamedMediaType) ToRawInfo() interface{} { func (m *NamedParameterOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:ParameterOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7417,7 +7385,7 @@ func (m *NamedParameterOrReference) ToRawInfo() interface{} { func (m *NamedPathItem) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:PathItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7427,7 +7395,7 @@ func (m *NamedPathItem) ToRawInfo() interface{} { func (m *NamedRequestBodyOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:RequestBodyOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7437,7 +7405,7 @@ func (m *NamedRequestBodyOrReference) ToRawInfo() interface{} { func (m *NamedResponseOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:ResponseOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7447,7 +7415,7 @@ func (m *NamedResponseOrReference) ToRawInfo() interface{} { func (m *NamedSchemaOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7457,7 +7425,7 @@ func (m *NamedSchemaOrReference) ToRawInfo() interface{} { func (m *NamedSecuritySchemeOrReference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:SecuritySchemeOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7467,7 +7435,7 @@ func (m *NamedSecuritySchemeOrReference) ToRawInfo() interface{} { func (m *NamedServerVariable) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:ServerVariable StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -7477,10 +7445,10 @@ func (m *NamedServerVariable) ToRawInfo() interface{} { func (m *NamedString) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Value != "" { - info = append(info, yaml.MapItem{"value", m.Value}) + info = append(info, yaml.MapItem{Key: "value", Value: m.Value}) } return info } @@ -7489,21 +7457,21 @@ func (m *NamedString) ToRawInfo() interface{} { func (m *OauthFlow) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AuthorizationUrl != "" { - info = append(info, yaml.MapItem{"authorizationUrl", m.AuthorizationUrl}) + info = append(info, yaml.MapItem{Key: "authorizationUrl", Value: m.AuthorizationUrl}) } if m.TokenUrl != "" { - info = append(info, yaml.MapItem{"tokenUrl", m.TokenUrl}) + info = append(info, yaml.MapItem{Key: "tokenUrl", Value: m.TokenUrl}) } if m.RefreshUrl != "" { - info = append(info, yaml.MapItem{"refreshUrl", m.RefreshUrl}) + info = append(info, yaml.MapItem{Key: "refreshUrl", Value: m.RefreshUrl}) } if m.Scopes != nil { - info = append(info, yaml.MapItem{"scopes", m.Scopes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) } // &{Name:scopes Type:Strings StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7514,24 +7482,24 @@ func (m *OauthFlow) ToRawInfo() interface{} { func (m *OauthFlows) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Implicit != nil { - info = append(info, yaml.MapItem{"implicit", m.Implicit.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "implicit", Value: m.Implicit.ToRawInfo()}) } // &{Name:implicit Type:OauthFlow StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Password != nil { - info = append(info, yaml.MapItem{"password", m.Password.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "password", Value: m.Password.ToRawInfo()}) } // &{Name:password Type:OauthFlow StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ClientCredentials != nil { - info = append(info, yaml.MapItem{"clientCredentials", m.ClientCredentials.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "clientCredentials", Value: m.ClientCredentials.ToRawInfo()}) } // &{Name:clientCredentials Type:OauthFlow StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.AuthorizationCode != nil { - info = append(info, yaml.MapItem{"authorizationCode", m.AuthorizationCode.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "authorizationCode", Value: m.AuthorizationCode.ToRawInfo()}) } // &{Name:authorizationCode Type:OauthFlow StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7543,7 +7511,7 @@ func (m *Object) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern: Implicit:true Description:} @@ -7554,50 +7522,50 @@ func (m *Object) ToRawInfo() interface{} { func (m *Operation) ToRawInfo() interface{} { info := yaml.MapSlice{} if len(m.Tags) != 0 { - info = append(info, yaml.MapItem{"tags", m.Tags}) + info = append(info, yaml.MapItem{Key: "tags", Value: m.Tags}) } if m.Summary != "" { - info = append(info, yaml.MapItem{"summary", m.Summary}) + info = append(info, yaml.MapItem{Key: "summary", Value: m.Summary}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.OperationId != "" { - info = append(info, yaml.MapItem{"operationId", m.OperationId}) + info = append(info, yaml.MapItem{Key: "operationId", Value: m.OperationId}) } if len(m.Parameters) != 0 { items := make([]interface{}, 0) for _, item := range m.Parameters { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"parameters", items}) + info = append(info, yaml.MapItem{Key: "parameters", Value: items}) } // &{Name:parameters Type:ParameterOrReference StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.RequestBody != nil { - info = append(info, yaml.MapItem{"requestBody", m.RequestBody.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "requestBody", Value: m.RequestBody.ToRawInfo()}) } // &{Name:requestBody Type:RequestBodyOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Responses != nil { - info = append(info, yaml.MapItem{"responses", m.Responses.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "responses", Value: m.Responses.ToRawInfo()}) } // &{Name:responses Type:Responses StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Callbacks != nil { - info = append(info, yaml.MapItem{"callbacks", m.Callbacks.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "callbacks", Value: m.Callbacks.ToRawInfo()}) } // &{Name:callbacks Type:CallbacksOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Deprecated != false { - info = append(info, yaml.MapItem{"deprecated", m.Deprecated}) + info = append(info, yaml.MapItem{Key: "deprecated", Value: m.Deprecated}) } if len(m.Security) != 0 { items := make([]interface{}, 0) for _, item := range m.Security { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"security", items}) + info = append(info, yaml.MapItem{Key: "security", Value: items}) } // &{Name:security Type:SecurityRequirement StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if len(m.Servers) != 0 { @@ -7605,12 +7573,12 @@ func (m *Operation) ToRawInfo() interface{} { for _, item := range m.Servers { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"servers", items}) + info = append(info, yaml.MapItem{Key: "servers", Value: items}) } // &{Name:servers Type:Server StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7621,51 +7589,51 @@ func (m *Operation) ToRawInfo() interface{} { func (m *Parameter) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.Deprecated != false { - info = append(info, yaml.MapItem{"deprecated", m.Deprecated}) + info = append(info, yaml.MapItem{Key: "deprecated", Value: m.Deprecated}) } if m.AllowEmptyValue != false { - info = append(info, yaml.MapItem{"allowEmptyValue", m.AllowEmptyValue}) + info = append(info, yaml.MapItem{Key: "allowEmptyValue", Value: m.AllowEmptyValue}) } if m.Style != "" { - info = append(info, yaml.MapItem{"style", m.Style}) + info = append(info, yaml.MapItem{Key: "style", Value: m.Style}) } if m.Explode != false { - info = append(info, yaml.MapItem{"explode", m.Explode}) + info = append(info, yaml.MapItem{Key: "explode", Value: m.Explode}) } if m.AllowReserved != false { - info = append(info, yaml.MapItem{"allowReserved", m.AllowReserved}) + info = append(info, yaml.MapItem{Key: "allowReserved", Value: m.AllowReserved}) } if m.Schema != nil { - info = append(info, yaml.MapItem{"schema", m.Schema.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schema", Value: m.Schema.ToRawInfo()}) } // &{Name:schema Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{"example", m.Example.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) } // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Examples != nil { - info = append(info, yaml.MapItem{"examples", m.Examples.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "examples", Value: m.Examples.ToRawInfo()}) } // &{Name:examples Type:ExamplesOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Content != nil { - info = append(info, yaml.MapItem{"content", m.Content.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "content", Value: m.Content.ToRawInfo()}) } // &{Name:content Type:MediaTypes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7694,7 +7662,7 @@ func (m *ParametersOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedParameterOrReference StringEnumValues:[] MapType:ParameterOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7705,44 +7673,44 @@ func (m *ParametersOrReferences) ToRawInfo() interface{} { func (m *PathItem) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.Summary != "" { - info = append(info, yaml.MapItem{"summary", m.Summary}) + info = append(info, yaml.MapItem{Key: "summary", Value: m.Summary}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Get != nil { - info = append(info, yaml.MapItem{"get", m.Get.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "get", Value: m.Get.ToRawInfo()}) } // &{Name:get Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Put != nil { - info = append(info, yaml.MapItem{"put", m.Put.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "put", Value: m.Put.ToRawInfo()}) } // &{Name:put Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Post != nil { - info = append(info, yaml.MapItem{"post", m.Post.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "post", Value: m.Post.ToRawInfo()}) } // &{Name:post Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Delete != nil { - info = append(info, yaml.MapItem{"delete", m.Delete.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "delete", Value: m.Delete.ToRawInfo()}) } // &{Name:delete Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Options != nil { - info = append(info, yaml.MapItem{"options", m.Options.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "options", Value: m.Options.ToRawInfo()}) } // &{Name:options Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Head != nil { - info = append(info, yaml.MapItem{"head", m.Head.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "head", Value: m.Head.ToRawInfo()}) } // &{Name:head Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Patch != nil { - info = append(info, yaml.MapItem{"patch", m.Patch.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "patch", Value: m.Patch.ToRawInfo()}) } // &{Name:patch Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Trace != nil { - info = append(info, yaml.MapItem{"trace", m.Trace.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "trace", Value: m.Trace.ToRawInfo()}) } // &{Name:trace Type:Operation StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Servers) != 0 { @@ -7750,7 +7718,7 @@ func (m *PathItem) ToRawInfo() interface{} { for _, item := range m.Servers { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"servers", items}) + info = append(info, yaml.MapItem{Key: "servers", Value: items}) } // &{Name:servers Type:Server StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if len(m.Parameters) != 0 { @@ -7758,12 +7726,12 @@ func (m *PathItem) ToRawInfo() interface{} { for _, item := range m.Parameters { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"parameters", items}) + info = append(info, yaml.MapItem{Key: "parameters", Value: items}) } // &{Name:parameters Type:ParameterOrReference StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7775,13 +7743,13 @@ func (m *Paths) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Path != nil { for _, item := range m.Path { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:Path Type:NamedPathItem StringEnumValues:[] MapType:PathItem Repeated:true Pattern:^/ Implicit:true Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7793,7 +7761,7 @@ func (m *Properties) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSchemaOrReference StringEnumValues:[] MapType:SchemaOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7804,7 +7772,7 @@ func (m *Properties) ToRawInfo() interface{} { func (m *Reference) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } return info } @@ -7814,7 +7782,7 @@ func (m *RequestBodiesOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedRequestBodyOrReference StringEnumValues:[] MapType:RequestBodyOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7825,18 +7793,18 @@ func (m *RequestBodiesOrReferences) ToRawInfo() interface{} { func (m *RequestBody) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Content != nil { - info = append(info, yaml.MapItem{"content", m.Content.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "content", Value: m.Content.ToRawInfo()}) } // &{Name:content Type:MediaTypes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7864,23 +7832,23 @@ func (m *RequestBodyOrReference) ToRawInfo() interface{} { func (m *Response) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Headers != nil { - info = append(info, yaml.MapItem{"headers", m.Headers.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "headers", Value: m.Headers.ToRawInfo()}) } // &{Name:headers Type:HeadersOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Content != nil { - info = append(info, yaml.MapItem{"content", m.Content.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "content", Value: m.Content.ToRawInfo()}) } // &{Name:content Type:MediaTypes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Links != nil { - info = append(info, yaml.MapItem{"links", m.Links.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "links", Value: m.Links.ToRawInfo()}) } // &{Name:links Type:LinksOrReferences StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7908,18 +7876,18 @@ func (m *ResponseOrReference) ToRawInfo() interface{} { func (m *Responses) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:ResponseOrReference StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ResponseOrReference != nil { for _, item := range m.ResponseOrReference { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:ResponseOrReference Type:NamedResponseOrReference StringEnumValues:[] MapType:ResponseOrReference Repeated:true Pattern:^([0-9X]{3})$ Implicit:true Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -7931,7 +7899,7 @@ func (m *ResponsesOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedResponseOrReference StringEnumValues:[] MapType:ResponseOrReference Repeated:true Pattern: Implicit:true Description:} @@ -7942,95 +7910,95 @@ func (m *ResponsesOrReferences) ToRawInfo() interface{} { func (m *Schema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Nullable != false { - info = append(info, yaml.MapItem{"nullable", m.Nullable}) + info = append(info, yaml.MapItem{Key: "nullable", Value: m.Nullable}) } if m.Discriminator != nil { - info = append(info, yaml.MapItem{"discriminator", m.Discriminator.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "discriminator", Value: m.Discriminator.ToRawInfo()}) } // &{Name:discriminator Type:Discriminator StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ReadOnly != false { - info = append(info, yaml.MapItem{"readOnly", m.ReadOnly}) + info = append(info, yaml.MapItem{Key: "readOnly", Value: m.ReadOnly}) } if m.WriteOnly != false { - info = append(info, yaml.MapItem{"writeOnly", m.WriteOnly}) + info = append(info, yaml.MapItem{Key: "writeOnly", Value: m.WriteOnly}) } if m.Xml != nil { - info = append(info, yaml.MapItem{"xml", m.Xml.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "xml", Value: m.Xml.ToRawInfo()}) } // &{Name:xml Type:Xml StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Example != nil { - info = append(info, yaml.MapItem{"example", m.Example.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "example", Value: m.Example.ToRawInfo()}) } // &{Name:example Type:Any StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Deprecated != false { - info = append(info, yaml.MapItem{"deprecated", m.Deprecated}) + info = append(info, yaml.MapItem{Key: "deprecated", Value: m.Deprecated}) } if m.Title != "" { - info = append(info, yaml.MapItem{"title", m.Title}) + info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) } if m.MultipleOf != 0.0 { - info = append(info, yaml.MapItem{"multipleOf", m.MultipleOf}) + info = append(info, yaml.MapItem{Key: "multipleOf", Value: m.MultipleOf}) } if m.Maximum != 0.0 { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if m.ExclusiveMaximum != false { - info = append(info, yaml.MapItem{"exclusiveMaximum", m.ExclusiveMaximum}) + info = append(info, yaml.MapItem{Key: "exclusiveMaximum", Value: m.ExclusiveMaximum}) } if m.Minimum != 0.0 { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.ExclusiveMinimum != false { - info = append(info, yaml.MapItem{"exclusiveMinimum", m.ExclusiveMinimum}) + info = append(info, yaml.MapItem{Key: "exclusiveMinimum", Value: m.ExclusiveMinimum}) } if m.MaxLength != 0 { - info = append(info, yaml.MapItem{"maxLength", m.MaxLength}) + info = append(info, yaml.MapItem{Key: "maxLength", Value: m.MaxLength}) } if m.MinLength != 0 { - info = append(info, yaml.MapItem{"minLength", m.MinLength}) + info = append(info, yaml.MapItem{Key: "minLength", Value: m.MinLength}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.MaxItems != 0 { - info = append(info, yaml.MapItem{"maxItems", m.MaxItems}) + info = append(info, yaml.MapItem{Key: "maxItems", Value: m.MaxItems}) } if m.MinItems != 0 { - info = append(info, yaml.MapItem{"minItems", m.MinItems}) + info = append(info, yaml.MapItem{Key: "minItems", Value: m.MinItems}) } if m.UniqueItems != false { - info = append(info, yaml.MapItem{"uniqueItems", m.UniqueItems}) + info = append(info, yaml.MapItem{Key: "uniqueItems", Value: m.UniqueItems}) } if m.MaxProperties != 0 { - info = append(info, yaml.MapItem{"maxProperties", m.MaxProperties}) + info = append(info, yaml.MapItem{Key: "maxProperties", Value: m.MaxProperties}) } if m.MinProperties != 0 { - info = append(info, yaml.MapItem{"minProperties", m.MinProperties}) + info = append(info, yaml.MapItem{Key: "minProperties", Value: m.MinProperties}) } if len(m.Required) != 0 { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if len(m.Enum) != 0 { items := make([]interface{}, 0) for _, item := range m.Enum { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"enum", items}) + info = append(info, yaml.MapItem{Key: "enum", Value: items}) } // &{Name:enum Type:Any StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if len(m.AllOf) != 0 { items := make([]interface{}, 0) for _, item := range m.AllOf { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"allOf", items}) + info = append(info, yaml.MapItem{Key: "allOf", Value: items}) } // &{Name:allOf Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if len(m.OneOf) != 0 { @@ -8038,7 +8006,7 @@ func (m *Schema) ToRawInfo() interface{} { for _, item := range m.OneOf { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"oneOf", items}) + info = append(info, yaml.MapItem{Key: "oneOf", Value: items}) } // &{Name:oneOf Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if len(m.AnyOf) != 0 { @@ -8046,11 +8014,11 @@ func (m *Schema) ToRawInfo() interface{} { for _, item := range m.AnyOf { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"anyOf", items}) + info = append(info, yaml.MapItem{Key: "anyOf", Value: items}) } // &{Name:anyOf Type:SchemaOrReference StringEnumValues:[] MapType: Repeated:true Pattern: Implicit:false Description:} if m.Not != nil { - info = append(info, yaml.MapItem{"not", m.Not.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "not", Value: m.Not.ToRawInfo()}) } // &{Name:not Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Items != nil { @@ -8058,30 +8026,30 @@ func (m *Schema) ToRawInfo() interface{} { for _, item := range m.Items.SchemaOrReference { items = append(items, item.ToRawInfo()) } - info = append(info, yaml.MapItem{"items", items[0]}) + info = append(info, yaml.MapItem{Key: "items", Value: items[0]}) } // &{Name:items Type:ItemsItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Properties != nil { - info = append(info, yaml.MapItem{"properties", m.Properties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "properties", Value: m.Properties.ToRawInfo()}) } // &{Name:properties Type:Properties StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.AdditionalProperties != nil { - info = append(info, yaml.MapItem{"additionalProperties", m.AdditionalProperties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "additionalProperties", Value: m.AdditionalProperties.ToRawInfo()}) } // &{Name:additionalProperties Type:AdditionalPropertiesItem StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Default != nil { - info = append(info, yaml.MapItem{"default", m.Default.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default.ToRawInfo()}) } // &{Name:default Type:DefaultType StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8110,7 +8078,7 @@ func (m *SchemasOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSchemaOrReference StringEnumValues:[] MapType:SchemaOrReference Repeated:true Pattern: Implicit:true Description:} @@ -8127,33 +8095,33 @@ func (m *SecurityRequirement) ToRawInfo() interface{} { func (m *SecurityScheme) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.In != "" { - info = append(info, yaml.MapItem{"in", m.In}) + info = append(info, yaml.MapItem{Key: "in", Value: m.In}) } if m.Scheme != "" { - info = append(info, yaml.MapItem{"scheme", m.Scheme}) + info = append(info, yaml.MapItem{Key: "scheme", Value: m.Scheme}) } if m.BearerFormat != "" { - info = append(info, yaml.MapItem{"bearerFormat", m.BearerFormat}) + info = append(info, yaml.MapItem{Key: "bearerFormat", Value: m.BearerFormat}) } if m.Flows != nil { - info = append(info, yaml.MapItem{"flows", m.Flows.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "flows", Value: m.Flows.ToRawInfo()}) } // &{Name:flows Type:OauthFlows StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.OpenIdConnectUrl != "" { - info = append(info, yaml.MapItem{"openIdConnectUrl", m.OpenIdConnectUrl}) + info = append(info, yaml.MapItem{Key: "openIdConnectUrl", Value: m.OpenIdConnectUrl}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8182,7 +8150,7 @@ func (m *SecuritySchemesOrReferences) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSecuritySchemeOrReference StringEnumValues:[] MapType:SecuritySchemeOrReference Repeated:true Pattern: Implicit:true Description:} @@ -8193,18 +8161,18 @@ func (m *SecuritySchemesOrReferences) ToRawInfo() interface{} { func (m *Server) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Url != "" { - info = append(info, yaml.MapItem{"url", m.Url}) + info = append(info, yaml.MapItem{Key: "url", Value: m.Url}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Variables != nil { - info = append(info, yaml.MapItem{"variables", m.Variables.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "variables", Value: m.Variables.ToRawInfo()}) } // &{Name:variables Type:ServerVariables StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8215,17 +8183,17 @@ func (m *Server) ToRawInfo() interface{} { func (m *ServerVariable) ToRawInfo() interface{} { info := yaml.MapSlice{} if len(m.Enum) != 0 { - info = append(info, yaml.MapItem{"enum", m.Enum}) + info = append(info, yaml.MapItem{Key: "enum", Value: m.Enum}) } if m.Default != "" { - info = append(info, yaml.MapItem{"default", m.Default}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8237,7 +8205,7 @@ func (m *ServerVariables) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedServerVariable StringEnumValues:[] MapType:ServerVariable Repeated:true Pattern: Implicit:true Description:} @@ -8279,18 +8247,18 @@ func (m *Strings) ToRawInfo() interface{} { func (m *Tag) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.ExternalDocs != nil { - info = append(info, yaml.MapItem{"externalDocs", m.ExternalDocs.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "externalDocs", Value: m.ExternalDocs.ToRawInfo()}) } // &{Name:externalDocs Type:ExternalDocs StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} @@ -8301,23 +8269,23 @@ func (m *Tag) ToRawInfo() interface{} { func (m *Xml) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Namespace != "" { - info = append(info, yaml.MapItem{"namespace", m.Namespace}) + info = append(info, yaml.MapItem{Key: "namespace", Value: m.Namespace}) } if m.Prefix != "" { - info = append(info, yaml.MapItem{"prefix", m.Prefix}) + info = append(info, yaml.MapItem{Key: "prefix", Value: m.Prefix}) } if m.Attribute != false { - info = append(info, yaml.MapItem{"attribute", m.Attribute}) + info = append(info, yaml.MapItem{Key: "attribute", Value: m.Attribute}) } if m.Wrapped != false { - info = append(info, yaml.MapItem{"wrapped", m.Wrapped}) + info = append(info, yaml.MapItem{Key: "wrapped", Value: m.Wrapped}) } if m.SpecificationExtension != nil { for _, item := range m.SpecificationExtension { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:SpecificationExtension Type:NamedAny StringEnumValues:[] MapType:Any Repeated:true Pattern:^x- Implicit:true Description:} diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.pb.go b/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.pb.go index 73aeb6f78..26e98c7c2 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.pb.go +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-go. +// Code generated by protoc-gen-go. DO NOT EDIT. // source: OpenAPIv3/OpenAPIv3.proto -// DO NOT EDIT! /* Package openapi_v3 is a generated protocol buffer package. @@ -25,7 +24,6 @@ It has these top-level messages: Encodings Example ExampleOrReference - Examples ExamplesOrReferences Expression ExternalDocs @@ -1170,14 +1168,6 @@ func _ExampleOrReference_OneofSizer(msg proto.Message) (n int) { return n } -type Examples struct { -} - -func (m *Examples) Reset() { *m = Examples{} } -func (m *Examples) String() string { return proto.CompactTextString(m) } -func (*Examples) ProtoMessage() {} -func (*Examples) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{16} } - type ExamplesOrReferences struct { AdditionalProperties []*NamedExampleOrReference `protobuf:"bytes,1,rep,name=additional_properties,json=additionalProperties" json:"additional_properties,omitempty"` } @@ -1185,7 +1175,7 @@ type ExamplesOrReferences struct { func (m *ExamplesOrReferences) Reset() { *m = ExamplesOrReferences{} } func (m *ExamplesOrReferences) String() string { return proto.CompactTextString(m) } func (*ExamplesOrReferences) ProtoMessage() {} -func (*ExamplesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{17} } +func (*ExamplesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{16} } func (m *ExamplesOrReferences) GetAdditionalProperties() []*NamedExampleOrReference { if m != nil { @@ -1201,7 +1191,7 @@ type Expression struct { func (m *Expression) Reset() { *m = Expression{} } func (m *Expression) String() string { return proto.CompactTextString(m) } func (*Expression) ProtoMessage() {} -func (*Expression) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{18} } +func (*Expression) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{17} } func (m *Expression) GetAdditionalProperties() []*NamedAny { if m != nil { @@ -1220,7 +1210,7 @@ type ExternalDocs struct { func (m *ExternalDocs) Reset() { *m = ExternalDocs{} } func (m *ExternalDocs) String() string { return proto.CompactTextString(m) } func (*ExternalDocs) ProtoMessage() {} -func (*ExternalDocs) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{19} } +func (*ExternalDocs) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{18} } func (m *ExternalDocs) GetDescription() string { if m != nil { @@ -1262,7 +1252,7 @@ type Header struct { func (m *Header) Reset() { *m = Header{} } func (m *Header) String() string { return proto.CompactTextString(m) } func (*Header) ProtoMessage() {} -func (*Header) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{20} } +func (*Header) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{19} } func (m *Header) GetDescription() string { if m != nil { @@ -1358,7 +1348,7 @@ type HeaderOrReference struct { func (m *HeaderOrReference) Reset() { *m = HeaderOrReference{} } func (m *HeaderOrReference) String() string { return proto.CompactTextString(m) } func (*HeaderOrReference) ProtoMessage() {} -func (*HeaderOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{21} } +func (*HeaderOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{20} } type isHeaderOrReference_Oneof interface { isHeaderOrReference_Oneof() @@ -1476,7 +1466,7 @@ type HeadersOrReferences struct { func (m *HeadersOrReferences) Reset() { *m = HeadersOrReferences{} } func (m *HeadersOrReferences) String() string { return proto.CompactTextString(m) } func (*HeadersOrReferences) ProtoMessage() {} -func (*HeadersOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{22} } +func (*HeadersOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{21} } func (m *HeadersOrReferences) GetAdditionalProperties() []*NamedHeaderOrReference { if m != nil { @@ -1499,7 +1489,7 @@ type Info struct { func (m *Info) Reset() { *m = Info{} } func (m *Info) String() string { return proto.CompactTextString(m) } func (*Info) ProtoMessage() {} -func (*Info) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{23} } +func (*Info) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{22} } func (m *Info) GetTitle() string { if m != nil { @@ -1557,7 +1547,7 @@ type ItemsItem struct { func (m *ItemsItem) Reset() { *m = ItemsItem{} } func (m *ItemsItem) String() string { return proto.CompactTextString(m) } func (*ItemsItem) ProtoMessage() {} -func (*ItemsItem) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{24} } +func (*ItemsItem) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{23} } func (m *ItemsItem) GetSchemaOrReference() []*SchemaOrReference { if m != nil { @@ -1576,7 +1566,7 @@ type License struct { func (m *License) Reset() { *m = License{} } func (m *License) String() string { return proto.CompactTextString(m) } func (*License) ProtoMessage() {} -func (*License) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{25} } +func (*License) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{24} } func (m *License) GetName() string { if m != nil { @@ -1613,7 +1603,7 @@ type Link struct { func (m *Link) Reset() { *m = Link{} } func (m *Link) String() string { return proto.CompactTextString(m) } func (*Link) ProtoMessage() {} -func (*Link) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{26} } +func (*Link) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{25} } func (m *Link) GetOperationRef() string { if m != nil { @@ -1674,7 +1664,7 @@ type LinkOrReference struct { func (m *LinkOrReference) Reset() { *m = LinkOrReference{} } func (m *LinkOrReference) String() string { return proto.CompactTextString(m) } func (*LinkOrReference) ProtoMessage() {} -func (*LinkOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{27} } +func (*LinkOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{26} } type isLinkOrReference_Oneof interface { isLinkOrReference_Oneof() @@ -1792,7 +1782,7 @@ type LinksOrReferences struct { func (m *LinksOrReferences) Reset() { *m = LinksOrReferences{} } func (m *LinksOrReferences) String() string { return proto.CompactTextString(m) } func (*LinksOrReferences) ProtoMessage() {} -func (*LinksOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{28} } +func (*LinksOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{27} } func (m *LinksOrReferences) GetAdditionalProperties() []*NamedLinkOrReference { if m != nil { @@ -1813,7 +1803,7 @@ type MediaType struct { func (m *MediaType) Reset() { *m = MediaType{} } func (m *MediaType) String() string { return proto.CompactTextString(m) } func (*MediaType) ProtoMessage() {} -func (*MediaType) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{29} } +func (*MediaType) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{28} } func (m *MediaType) GetSchema() *SchemaOrReference { if m != nil { @@ -1857,7 +1847,7 @@ type MediaTypes struct { func (m *MediaTypes) Reset() { *m = MediaTypes{} } func (m *MediaTypes) String() string { return proto.CompactTextString(m) } func (*MediaTypes) ProtoMessage() {} -func (*MediaTypes) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{30} } +func (*MediaTypes) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{29} } func (m *MediaTypes) GetAdditionalProperties() []*NamedMediaType { if m != nil { @@ -1877,7 +1867,7 @@ type NamedAny struct { func (m *NamedAny) Reset() { *m = NamedAny{} } func (m *NamedAny) String() string { return proto.CompactTextString(m) } func (*NamedAny) ProtoMessage() {} -func (*NamedAny) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{31} } +func (*NamedAny) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{30} } func (m *NamedAny) GetName() string { if m != nil { @@ -1904,7 +1894,7 @@ type NamedAnyOrExpression struct { func (m *NamedAnyOrExpression) Reset() { *m = NamedAnyOrExpression{} } func (m *NamedAnyOrExpression) String() string { return proto.CompactTextString(m) } func (*NamedAnyOrExpression) ProtoMessage() {} -func (*NamedAnyOrExpression) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{32} } +func (*NamedAnyOrExpression) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{31} } func (m *NamedAnyOrExpression) GetName() string { if m != nil { @@ -1931,7 +1921,7 @@ type NamedCallbackOrReference struct { func (m *NamedCallbackOrReference) Reset() { *m = NamedCallbackOrReference{} } func (m *NamedCallbackOrReference) String() string { return proto.CompactTextString(m) } func (*NamedCallbackOrReference) ProtoMessage() {} -func (*NamedCallbackOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{33} } +func (*NamedCallbackOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{32} } func (m *NamedCallbackOrReference) GetName() string { if m != nil { @@ -1958,7 +1948,7 @@ type NamedEncoding struct { func (m *NamedEncoding) Reset() { *m = NamedEncoding{} } func (m *NamedEncoding) String() string { return proto.CompactTextString(m) } func (*NamedEncoding) ProtoMessage() {} -func (*NamedEncoding) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{34} } +func (*NamedEncoding) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{33} } func (m *NamedEncoding) GetName() string { if m != nil { @@ -1985,7 +1975,7 @@ type NamedExampleOrReference struct { func (m *NamedExampleOrReference) Reset() { *m = NamedExampleOrReference{} } func (m *NamedExampleOrReference) String() string { return proto.CompactTextString(m) } func (*NamedExampleOrReference) ProtoMessage() {} -func (*NamedExampleOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{35} } +func (*NamedExampleOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{34} } func (m *NamedExampleOrReference) GetName() string { if m != nil { @@ -2012,7 +2002,7 @@ type NamedHeaderOrReference struct { func (m *NamedHeaderOrReference) Reset() { *m = NamedHeaderOrReference{} } func (m *NamedHeaderOrReference) String() string { return proto.CompactTextString(m) } func (*NamedHeaderOrReference) ProtoMessage() {} -func (*NamedHeaderOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{36} } +func (*NamedHeaderOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{35} } func (m *NamedHeaderOrReference) GetName() string { if m != nil { @@ -2039,7 +2029,7 @@ type NamedLinkOrReference struct { func (m *NamedLinkOrReference) Reset() { *m = NamedLinkOrReference{} } func (m *NamedLinkOrReference) String() string { return proto.CompactTextString(m) } func (*NamedLinkOrReference) ProtoMessage() {} -func (*NamedLinkOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{37} } +func (*NamedLinkOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{36} } func (m *NamedLinkOrReference) GetName() string { if m != nil { @@ -2066,7 +2056,7 @@ type NamedMediaType struct { func (m *NamedMediaType) Reset() { *m = NamedMediaType{} } func (m *NamedMediaType) String() string { return proto.CompactTextString(m) } func (*NamedMediaType) ProtoMessage() {} -func (*NamedMediaType) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{38} } +func (*NamedMediaType) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{37} } func (m *NamedMediaType) GetName() string { if m != nil { @@ -2093,7 +2083,7 @@ type NamedParameterOrReference struct { func (m *NamedParameterOrReference) Reset() { *m = NamedParameterOrReference{} } func (m *NamedParameterOrReference) String() string { return proto.CompactTextString(m) } func (*NamedParameterOrReference) ProtoMessage() {} -func (*NamedParameterOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{39} } +func (*NamedParameterOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{38} } func (m *NamedParameterOrReference) GetName() string { if m != nil { @@ -2120,7 +2110,7 @@ type NamedPathItem struct { func (m *NamedPathItem) Reset() { *m = NamedPathItem{} } func (m *NamedPathItem) String() string { return proto.CompactTextString(m) } func (*NamedPathItem) ProtoMessage() {} -func (*NamedPathItem) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{40} } +func (*NamedPathItem) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{39} } func (m *NamedPathItem) GetName() string { if m != nil { @@ -2147,7 +2137,7 @@ type NamedRequestBodyOrReference struct { func (m *NamedRequestBodyOrReference) Reset() { *m = NamedRequestBodyOrReference{} } func (m *NamedRequestBodyOrReference) String() string { return proto.CompactTextString(m) } func (*NamedRequestBodyOrReference) ProtoMessage() {} -func (*NamedRequestBodyOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{41} } +func (*NamedRequestBodyOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{40} } func (m *NamedRequestBodyOrReference) GetName() string { if m != nil { @@ -2174,7 +2164,7 @@ type NamedResponseOrReference struct { func (m *NamedResponseOrReference) Reset() { *m = NamedResponseOrReference{} } func (m *NamedResponseOrReference) String() string { return proto.CompactTextString(m) } func (*NamedResponseOrReference) ProtoMessage() {} -func (*NamedResponseOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{42} } +func (*NamedResponseOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{41} } func (m *NamedResponseOrReference) GetName() string { if m != nil { @@ -2201,7 +2191,7 @@ type NamedSchemaOrReference struct { func (m *NamedSchemaOrReference) Reset() { *m = NamedSchemaOrReference{} } func (m *NamedSchemaOrReference) String() string { return proto.CompactTextString(m) } func (*NamedSchemaOrReference) ProtoMessage() {} -func (*NamedSchemaOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{43} } +func (*NamedSchemaOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{42} } func (m *NamedSchemaOrReference) GetName() string { if m != nil { @@ -2228,7 +2218,7 @@ type NamedSecuritySchemeOrReference struct { func (m *NamedSecuritySchemeOrReference) Reset() { *m = NamedSecuritySchemeOrReference{} } func (m *NamedSecuritySchemeOrReference) String() string { return proto.CompactTextString(m) } func (*NamedSecuritySchemeOrReference) ProtoMessage() {} -func (*NamedSecuritySchemeOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{44} } +func (*NamedSecuritySchemeOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{43} } func (m *NamedSecuritySchemeOrReference) GetName() string { if m != nil { @@ -2255,7 +2245,7 @@ type NamedServerVariable struct { func (m *NamedServerVariable) Reset() { *m = NamedServerVariable{} } func (m *NamedServerVariable) String() string { return proto.CompactTextString(m) } func (*NamedServerVariable) ProtoMessage() {} -func (*NamedServerVariable) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{45} } +func (*NamedServerVariable) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{44} } func (m *NamedServerVariable) GetName() string { if m != nil { @@ -2282,7 +2272,7 @@ type NamedString struct { func (m *NamedString) Reset() { *m = NamedString{} } func (m *NamedString) String() string { return proto.CompactTextString(m) } func (*NamedString) ProtoMessage() {} -func (*NamedString) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{46} } +func (*NamedString) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{45} } func (m *NamedString) GetName() string { if m != nil { @@ -2310,7 +2300,7 @@ type OauthFlow struct { func (m *OauthFlow) Reset() { *m = OauthFlow{} } func (m *OauthFlow) String() string { return proto.CompactTextString(m) } func (*OauthFlow) ProtoMessage() {} -func (*OauthFlow) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{47} } +func (*OauthFlow) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{46} } func (m *OauthFlow) GetAuthorizationUrl() string { if m != nil { @@ -2359,7 +2349,7 @@ type OauthFlows struct { func (m *OauthFlows) Reset() { *m = OauthFlows{} } func (m *OauthFlows) String() string { return proto.CompactTextString(m) } func (*OauthFlows) ProtoMessage() {} -func (*OauthFlows) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{48} } +func (*OauthFlows) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{47} } func (m *OauthFlows) GetImplicit() *OauthFlow { if m != nil { @@ -2403,7 +2393,7 @@ type Object struct { func (m *Object) Reset() { *m = Object{} } func (m *Object) String() string { return proto.CompactTextString(m) } func (*Object) ProtoMessage() {} -func (*Object) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{49} } +func (*Object) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{48} } func (m *Object) GetAdditionalProperties() []*NamedAny { if m != nil { @@ -2432,7 +2422,7 @@ type Operation struct { func (m *Operation) Reset() { *m = Operation{} } func (m *Operation) String() string { return proto.CompactTextString(m) } func (*Operation) ProtoMessage() {} -func (*Operation) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{50} } +func (*Operation) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{49} } func (m *Operation) GetTags() []string { if m != nil { @@ -2546,7 +2536,7 @@ type Parameter struct { func (m *Parameter) Reset() { *m = Parameter{} } func (m *Parameter) String() string { return proto.CompactTextString(m) } func (*Parameter) ProtoMessage() {} -func (*Parameter) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{51} } +func (*Parameter) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{50} } func (m *Parameter) GetName() string { if m != nil { @@ -2656,7 +2646,7 @@ type ParameterOrReference struct { func (m *ParameterOrReference) Reset() { *m = ParameterOrReference{} } func (m *ParameterOrReference) String() string { return proto.CompactTextString(m) } func (*ParameterOrReference) ProtoMessage() {} -func (*ParameterOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{52} } +func (*ParameterOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{51} } type isParameterOrReference_Oneof interface { isParameterOrReference_Oneof() @@ -2774,7 +2764,7 @@ type ParametersOrReferences struct { func (m *ParametersOrReferences) Reset() { *m = ParametersOrReferences{} } func (m *ParametersOrReferences) String() string { return proto.CompactTextString(m) } func (*ParametersOrReferences) ProtoMessage() {} -func (*ParametersOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{53} } +func (*ParametersOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{52} } func (m *ParametersOrReferences) GetAdditionalProperties() []*NamedParameterOrReference { if m != nil { @@ -2804,7 +2794,7 @@ type PathItem struct { func (m *PathItem) Reset() { *m = PathItem{} } func (m *PathItem) String() string { return proto.CompactTextString(m) } func (*PathItem) ProtoMessage() {} -func (*PathItem) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{54} } +func (*PathItem) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{53} } func (m *PathItem) GetXRef() string { if m != nil { @@ -2913,7 +2903,7 @@ type Paths struct { func (m *Paths) Reset() { *m = Paths{} } func (m *Paths) String() string { return proto.CompactTextString(m) } func (*Paths) ProtoMessage() {} -func (*Paths) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{55} } +func (*Paths) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{54} } func (m *Paths) GetPath() []*NamedPathItem { if m != nil { @@ -2936,7 +2926,7 @@ type Properties struct { func (m *Properties) Reset() { *m = Properties{} } func (m *Properties) String() string { return proto.CompactTextString(m) } func (*Properties) ProtoMessage() {} -func (*Properties) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{56} } +func (*Properties) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{55} } func (m *Properties) GetAdditionalProperties() []*NamedSchemaOrReference { if m != nil { @@ -2953,7 +2943,7 @@ type Reference struct { func (m *Reference) Reset() { *m = Reference{} } func (m *Reference) String() string { return proto.CompactTextString(m) } func (*Reference) ProtoMessage() {} -func (*Reference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{57} } +func (*Reference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{56} } func (m *Reference) GetXRef() string { if m != nil { @@ -2969,7 +2959,7 @@ type RequestBodiesOrReferences struct { func (m *RequestBodiesOrReferences) Reset() { *m = RequestBodiesOrReferences{} } func (m *RequestBodiesOrReferences) String() string { return proto.CompactTextString(m) } func (*RequestBodiesOrReferences) ProtoMessage() {} -func (*RequestBodiesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{58} } +func (*RequestBodiesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{57} } func (m *RequestBodiesOrReferences) GetAdditionalProperties() []*NamedRequestBodyOrReference { if m != nil { @@ -2989,7 +2979,7 @@ type RequestBody struct { func (m *RequestBody) Reset() { *m = RequestBody{} } func (m *RequestBody) String() string { return proto.CompactTextString(m) } func (*RequestBody) ProtoMessage() {} -func (*RequestBody) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{59} } +func (*RequestBody) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{58} } func (m *RequestBody) GetDescription() string { if m != nil { @@ -3029,7 +3019,7 @@ type RequestBodyOrReference struct { func (m *RequestBodyOrReference) Reset() { *m = RequestBodyOrReference{} } func (m *RequestBodyOrReference) String() string { return proto.CompactTextString(m) } func (*RequestBodyOrReference) ProtoMessage() {} -func (*RequestBodyOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{60} } +func (*RequestBodyOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{59} } type isRequestBodyOrReference_Oneof interface { isRequestBodyOrReference_Oneof() @@ -3152,7 +3142,7 @@ type Response struct { func (m *Response) Reset() { *m = Response{} } func (m *Response) String() string { return proto.CompactTextString(m) } func (*Response) ProtoMessage() {} -func (*Response) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{61} } +func (*Response) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{60} } func (m *Response) GetDescription() string { if m != nil { @@ -3199,7 +3189,7 @@ type ResponseOrReference struct { func (m *ResponseOrReference) Reset() { *m = ResponseOrReference{} } func (m *ResponseOrReference) String() string { return proto.CompactTextString(m) } func (*ResponseOrReference) ProtoMessage() {} -func (*ResponseOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{62} } +func (*ResponseOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{61} } type isResponseOrReference_Oneof interface { isResponseOrReference_Oneof() @@ -3320,7 +3310,7 @@ type Responses struct { func (m *Responses) Reset() { *m = Responses{} } func (m *Responses) String() string { return proto.CompactTextString(m) } func (*Responses) ProtoMessage() {} -func (*Responses) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{63} } +func (*Responses) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{62} } func (m *Responses) GetDefault() *ResponseOrReference { if m != nil { @@ -3350,7 +3340,7 @@ type ResponsesOrReferences struct { func (m *ResponsesOrReferences) Reset() { *m = ResponsesOrReferences{} } func (m *ResponsesOrReferences) String() string { return proto.CompactTextString(m) } func (*ResponsesOrReferences) ProtoMessage() {} -func (*ResponsesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{64} } +func (*ResponsesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{63} } func (m *ResponsesOrReferences) GetAdditionalProperties() []*NamedResponseOrReference { if m != nil { @@ -3402,7 +3392,7 @@ type Schema struct { func (m *Schema) Reset() { *m = Schema{} } func (m *Schema) String() string { return proto.CompactTextString(m) } func (*Schema) ProtoMessage() {} -func (*Schema) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{65} } +func (*Schema) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{64} } func (m *Schema) GetNullable() bool { if m != nil { @@ -3666,7 +3656,7 @@ type SchemaOrReference struct { func (m *SchemaOrReference) Reset() { *m = SchemaOrReference{} } func (m *SchemaOrReference) String() string { return proto.CompactTextString(m) } func (*SchemaOrReference) ProtoMessage() {} -func (*SchemaOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{66} } +func (*SchemaOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{65} } type isSchemaOrReference_Oneof interface { isSchemaOrReference_Oneof() @@ -3784,7 +3774,7 @@ type SchemasOrReferences struct { func (m *SchemasOrReferences) Reset() { *m = SchemasOrReferences{} } func (m *SchemasOrReferences) String() string { return proto.CompactTextString(m) } func (*SchemasOrReferences) ProtoMessage() {} -func (*SchemasOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{67} } +func (*SchemasOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{66} } func (m *SchemasOrReferences) GetAdditionalProperties() []*NamedSchemaOrReference { if m != nil { @@ -3800,7 +3790,7 @@ type SecurityRequirement struct { func (m *SecurityRequirement) Reset() { *m = SecurityRequirement{} } func (m *SecurityRequirement) String() string { return proto.CompactTextString(m) } func (*SecurityRequirement) ProtoMessage() {} -func (*SecurityRequirement) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{68} } +func (*SecurityRequirement) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{67} } // Defines a security scheme that can be used by the operations. Supported schemes are HTTP authentication, an API key (either as a header or as a query parameter), OAuth2's common flows (implicit, password, application and access code) as defined in RFC6749, and OpenID Connect Discovery. type SecurityScheme struct { @@ -3818,7 +3808,7 @@ type SecurityScheme struct { func (m *SecurityScheme) Reset() { *m = SecurityScheme{} } func (m *SecurityScheme) String() string { return proto.CompactTextString(m) } func (*SecurityScheme) ProtoMessage() {} -func (*SecurityScheme) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{69} } +func (*SecurityScheme) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{68} } func (m *SecurityScheme) GetType() string { if m != nil { @@ -3893,7 +3883,7 @@ type SecuritySchemeOrReference struct { func (m *SecuritySchemeOrReference) Reset() { *m = SecuritySchemeOrReference{} } func (m *SecuritySchemeOrReference) String() string { return proto.CompactTextString(m) } func (*SecuritySchemeOrReference) ProtoMessage() {} -func (*SecuritySchemeOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{70} } +func (*SecuritySchemeOrReference) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{69} } type isSecuritySchemeOrReference_Oneof interface { isSecuritySchemeOrReference_Oneof() @@ -4011,7 +4001,7 @@ type SecuritySchemesOrReferences struct { func (m *SecuritySchemesOrReferences) Reset() { *m = SecuritySchemesOrReferences{} } func (m *SecuritySchemesOrReferences) String() string { return proto.CompactTextString(m) } func (*SecuritySchemesOrReferences) ProtoMessage() {} -func (*SecuritySchemesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{71} } +func (*SecuritySchemesOrReferences) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{70} } func (m *SecuritySchemesOrReferences) GetAdditionalProperties() []*NamedSecuritySchemeOrReference { if m != nil { @@ -4031,7 +4021,7 @@ type Server struct { func (m *Server) Reset() { *m = Server{} } func (m *Server) String() string { return proto.CompactTextString(m) } func (*Server) ProtoMessage() {} -func (*Server) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{72} } +func (*Server) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{71} } func (m *Server) GetUrl() string { if m != nil { @@ -4072,7 +4062,7 @@ type ServerVariable struct { func (m *ServerVariable) Reset() { *m = ServerVariable{} } func (m *ServerVariable) String() string { return proto.CompactTextString(m) } func (*ServerVariable) ProtoMessage() {} -func (*ServerVariable) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{73} } +func (*ServerVariable) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{72} } func (m *ServerVariable) GetEnum() []string { if m != nil { @@ -4109,7 +4099,7 @@ type ServerVariables struct { func (m *ServerVariables) Reset() { *m = ServerVariables{} } func (m *ServerVariables) String() string { return proto.CompactTextString(m) } func (*ServerVariables) ProtoMessage() {} -func (*ServerVariables) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{74} } +func (*ServerVariables) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{73} } func (m *ServerVariables) GetAdditionalProperties() []*NamedServerVariable { if m != nil { @@ -4130,7 +4120,7 @@ type SpecificationExtension struct { func (m *SpecificationExtension) Reset() { *m = SpecificationExtension{} } func (m *SpecificationExtension) String() string { return proto.CompactTextString(m) } func (*SpecificationExtension) ProtoMessage() {} -func (*SpecificationExtension) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{75} } +func (*SpecificationExtension) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{74} } type isSpecificationExtension_Oneof interface { isSpecificationExtension_Oneof() @@ -4268,7 +4258,7 @@ type StringArray struct { func (m *StringArray) Reset() { *m = StringArray{} } func (m *StringArray) String() string { return proto.CompactTextString(m) } func (*StringArray) ProtoMessage() {} -func (*StringArray) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{76} } +func (*StringArray) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{75} } func (m *StringArray) GetValue() []string { if m != nil { @@ -4284,7 +4274,7 @@ type Strings struct { func (m *Strings) Reset() { *m = Strings{} } func (m *Strings) String() string { return proto.CompactTextString(m) } func (*Strings) ProtoMessage() {} -func (*Strings) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{77} } +func (*Strings) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{76} } func (m *Strings) GetAdditionalProperties() []*NamedString { if m != nil { @@ -4304,7 +4294,7 @@ type Tag struct { func (m *Tag) Reset() { *m = Tag{} } func (m *Tag) String() string { return proto.CompactTextString(m) } func (*Tag) ProtoMessage() {} -func (*Tag) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{78} } +func (*Tag) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{77} } func (m *Tag) GetName() string { if m != nil { @@ -4347,7 +4337,7 @@ type Xml struct { func (m *Xml) Reset() { *m = Xml{} } func (m *Xml) String() string { return proto.CompactTextString(m) } func (*Xml) ProtoMessage() {} -func (*Xml) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{79} } +func (*Xml) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{78} } func (m *Xml) GetName() string { if m != nil { @@ -4408,7 +4398,6 @@ func init() { proto.RegisterType((*Encodings)(nil), "openapi.v3.Encodings") proto.RegisterType((*Example)(nil), "openapi.v3.Example") proto.RegisterType((*ExampleOrReference)(nil), "openapi.v3.ExampleOrReference") - proto.RegisterType((*Examples)(nil), "openapi.v3.Examples") proto.RegisterType((*ExamplesOrReferences)(nil), "openapi.v3.ExamplesOrReferences") proto.RegisterType((*Expression)(nil), "openapi.v3.Expression") proto.RegisterType((*ExternalDocs)(nil), "openapi.v3.ExternalDocs") @@ -4477,8 +4466,8 @@ func init() { func init() { proto.RegisterFile("OpenAPIv3/OpenAPIv3.proto", fileDescriptor0) } var fileDescriptor0 = []byte{ - // 3544 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x09, 0x6e, 0x88, 0x02, 0xff, 0xc4, 0x5b, 0x49, 0x6f, 0x1c, 0x47, + // 3539 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xc4, 0x5b, 0x49, 0x6f, 0x1c, 0x47, 0x96, 0x66, 0xd6, 0x5e, 0xaf, 0xb8, 0x06, 0x17, 0x95, 0x24, 0x4b, 0xa2, 0x48, 0xc9, 0x92, 0xb5, 0xda, 0x92, 0xe5, 0xb1, 0x61, 0x7b, 0x3c, 0x94, 0x44, 0x83, 0x02, 0x24, 0x97, 0x26, 0x25, 0xdb, 0x1a, 0x7b, 0x8c, 0x42, 0x30, 0x2b, 0x8a, 0x4c, 0x2b, 0x37, 0x65, 0x66, 0x49, 0xac, 0x39, 0x18, @@ -4552,152 +4541,152 @@ var fileDescriptor0 = []byte{ 0xea, 0x93, 0x11, 0xe4, 0x45, 0x98, 0x8f, 0x44, 0x8d, 0xe3, 0x2b, 0x8c, 0x57, 0x24, 0x80, 0x9f, 0x31, 0x58, 0xc1, 0x0e, 0x56, 0x27, 0xd8, 0xc1, 0x6f, 0x35, 0x40, 0x62, 0x91, 0xc9, 0x80, 0xe9, 0x26, 0x3d, 0x68, 0xd6, 0x2a, 0x22, 0x87, 0x65, 0x85, 0xe3, 0xa5, 0xe6, 0x4e, 0xa0, 0x8e, 0x1d, - 0x2d, 0x01, 0x34, 0xa4, 0x3b, 0xdf, 0xf0, 0x60, 0x45, 0xe5, 0xda, 0xd1, 0xb3, 0xe2, 0x03, 0xde, - 0xcc, 0x1e, 0x70, 0x66, 0x61, 0x39, 0x47, 0xfd, 0x39, 0x40, 0x22, 0x48, 0x7f, 0x50, 0x3c, 0x8e, - 0x7a, 0x83, 0xd5, 0x8c, 0xff, 0x5f, 0x83, 0xd9, 0xa4, 0x7d, 0x18, 0x15, 0x17, 0x2d, 0x2b, 0x2e, - 0x59, 0xa7, 0x57, 0x70, 0xe4, 0xe5, 0x09, 0x8e, 0xfc, 0xbf, 0x2b, 0x50, 0xe3, 0xba, 0x7f, 0x84, - 0xd9, 0x9c, 0x82, 0x86, 0xcf, 0xcd, 0x67, 0x8f, 0xfb, 0x38, 0x3d, 0xfa, 0x46, 0x67, 0x01, 0x7a, - 0xc4, 0xf3, 0x89, 0x81, 0x43, 0xd2, 0x63, 0xd2, 0xdd, 0xd0, 0x13, 0x2d, 0xe8, 0x0a, 0x2c, 0x71, - 0x9b, 0x40, 0x6c, 0x2f, 0x1c, 0x26, 0x84, 0xba, 0xa1, 0x2f, 0x30, 0xc2, 0x36, 0x6d, 0xe7, 0x62, - 0x1d, 0xd9, 0x9b, 0x6a, 0x8e, 0xbd, 0xa9, 0x1d, 0x66, 0x6f, 0xea, 0x2a, 0x7b, 0x73, 0x07, 0x6a, - 0x3c, 0xa2, 0x55, 0xc5, 0x49, 0x99, 0x2b, 0xa3, 0x2e, 0xc0, 0xe8, 0x8d, 0x58, 0xfc, 0x9b, 0x6a, - 0xa5, 0x8d, 0x04, 0x3f, 0x19, 0xa3, 0xc2, 0xd8, 0x31, 0xea, 0x9b, 0x50, 0x17, 0x06, 0xbb, 0xdd, - 0xca, 0xba, 0xbd, 0x47, 0xa4, 0x67, 0x62, 0x6a, 0xc9, 0x03, 0x5d, 0xc2, 0x8a, 0x84, 0x61, 0x76, - 0x02, 0x61, 0xf8, 0x4f, 0x0d, 0x96, 0xb8, 0x30, 0x24, 0xd5, 0xff, 0x1a, 0xd4, 0xb8, 0x23, 0x10, - 0xda, 0x8f, 0xb2, 0x7e, 0x83, 0x06, 0x34, 0x1c, 0x73, 0x6c, 0xdd, 0x77, 0x60, 0x59, 0xe1, 0x8b, - 0xd0, 0xe7, 0xc5, 0x6a, 0xb8, 0x91, 0x59, 0x67, 0x66, 0x1d, 0x39, 0x4a, 0xf9, 0x63, 0x09, 0x2a, - 0x34, 0xa8, 0xa1, 0x42, 0x17, 0x9a, 0xa1, 0x25, 0x7d, 0x27, 0xff, 0x38, 0x82, 0x45, 0xbf, 0x0c, - 0x8b, 0x21, 0xf1, 0xed, 0xa0, 0xeb, 0xf6, 0xbb, 0x54, 0xd0, 0x4c, 0x43, 0xfa, 0xc9, 0x79, 0xd6, - 0xde, 0xe9, 0x3f, 0xe1, 0xad, 0x34, 0x58, 0x33, 0x78, 0x80, 0x2b, 0x82, 0xa0, 0xe5, 0x74, 0x58, - 0xc3, 0x48, 0xba, 0xc4, 0x50, 0xb8, 0x65, 0x1a, 0xc4, 0x09, 0x88, 0x88, 0x82, 0x96, 0xd3, 0x71, - 0x3d, 0x23, 0xe9, 0x12, 0x43, 0xd5, 0x83, 0x06, 0x5a, 0xdc, 0x7b, 0x32, 0xaf, 0x24, 0x3e, 0x8b, - 0xa4, 0xa4, 0x3e, 0x81, 0x94, 0x7c, 0x01, 0x4d, 0x7a, 0x69, 0xe7, 0x39, 0x99, 0x47, 0x79, 0x39, - 0x99, 0xf2, 0xe1, 0x0a, 0x96, 0xcd, 0xc8, 0x6c, 0x7c, 0x03, 0x75, 0xb1, 0xb0, 0x23, 0xde, 0x01, - 0xa6, 0x6c, 0x0e, 0xff, 0x5c, 0x82, 0x0a, 0xbd, 0x31, 0xd1, 0xe8, 0x9b, 0x8a, 0x08, 0xe7, 0xe9, - 0x93, 0xbe, 0x8c, 0xbe, 0xa3, 0x46, 0x9d, 0xf4, 0x69, 0xd4, 0x15, 0x83, 0xcc, 0x9e, 0x94, 0x8e, - 0xa8, 0xed, 0x41, 0x0f, 0x7d, 0xa8, 0xb8, 0xf9, 0x9e, 0x19, 0xb1, 0x1f, 0xe9, 0xcc, 0x4c, 0xea, - 0xd2, 0xfb, 0xcf, 0x30, 0x9b, 0xb8, 0xb6, 0x0e, 0x85, 0xdc, 0x9c, 0x1e, 0x61, 0x90, 0xca, 0xcb, - 0xb4, 0xe2, 0xab, 0x6a, 0x26, 0x20, 0xa9, 0x66, 0xc5, 0xf7, 0x0a, 0xd4, 0x78, 0x88, 0x2e, 0x6e, - 0xb2, 0xaa, 0x20, 0x5e, 0x20, 0xa6, 0x2d, 0x48, 0x43, 0x58, 0xa0, 0x7b, 0x9d, 0xb4, 0x35, 0xaf, - 0x43, 0x85, 0xde, 0x4e, 0x85, 0xa5, 0x59, 0x1c, 0xbd, 0xc8, 0xee, 0xcc, 0xe8, 0x8c, 0x7e, 0x6c, - 0x2b, 0xf3, 0x35, 0x2c, 0x65, 0x2e, 0xc6, 0xe3, 0xa7, 0xc4, 0x46, 0x66, 0x9f, 0x63, 0x61, 0x7e, - 0x5d, 0x82, 0x66, 0x64, 0xbc, 0x13, 0x4e, 0x48, 0x9b, 0xd0, 0x09, 0x95, 0xc6, 0x70, 0x42, 0xe5, - 0xb1, 0x9d, 0xd0, 0x5b, 0xd0, 0x20, 0x22, 0x14, 0x16, 0xd2, 0x96, 0xda, 0xd8, 0x28, 0xc2, 0xd6, - 0x23, 0xd8, 0xb4, 0xa3, 0xd0, 0xaf, 0x00, 0x62, 0x5f, 0x87, 0x3a, 0xc5, 0x87, 0x72, 0x2a, 0xc3, - 0x3a, 0xea, 0x9b, 0x73, 0x1c, 0xdb, 0xd0, 0x90, 0x53, 0x50, 0xda, 0x98, 0x28, 0x42, 0x2f, 0x15, - 0x45, 0xe8, 0x1b, 0x5f, 0xc1, 0x8a, 0x2a, 0x2d, 0xaa, 0x64, 0xf9, 0x56, 0x9a, 0x65, 0xa1, 0xfa, - 0x0a, 0xf6, 0x04, 0xda, 0x79, 0xe9, 0x3e, 0xe5, 0x10, 0x77, 0xd2, 0x43, 0x9c, 0x53, 0x65, 0x72, - 0x92, 0x72, 0x25, 0x86, 0xe9, 0xc0, 0x5c, 0xea, 0xf6, 0xa3, 0xe4, 0x7d, 0x25, 0xcd, 0x7b, 0x45, - 0x25, 0x0f, 0x92, 0xa1, 0x01, 0x27, 0x72, 0xa2, 0x6d, 0x25, 0xeb, 0xb7, 0xd3, 0xac, 0xcf, 0x2a, - 0x04, 0x55, 0x31, 0x6b, 0x0c, 0x6b, 0x6a, 0x1f, 0xaf, 0x1c, 0xe3, 0x76, 0x7a, 0x8c, 0x33, 0xd9, - 0xf0, 0x45, 0x31, 0x84, 0x3c, 0xde, 0x51, 0x03, 0x35, 0xee, 0xf1, 0x8e, 0x9a, 0x08, 0xc1, 0xfe, - 0x5f, 0x61, 0x3e, 0x2d, 0xac, 0x4a, 0xc6, 0x57, 0xd3, 0x8c, 0x57, 0x95, 0xe1, 0xa0, 0x64, 0xb9, - 0x07, 0x27, 0x45, 0x52, 0x5d, 0x78, 0x8f, 0xc3, 0xa6, 0xfd, 0x4e, 0x9a, 0xfb, 0xba, 0x32, 0x1f, - 0x5b, 0x20, 0x33, 0x32, 0x7b, 0x3f, 0xb6, 0xcc, 0x44, 0x69, 0x7f, 0xc1, 0xf0, 0x39, 0x9c, 0x66, - 0x0c, 0xe3, 0xf4, 0xeb, 0xf0, 0xb0, 0xb9, 0xbf, 0x9b, 0x66, 0xbf, 0xa1, 0xce, 0xe2, 0x0e, 0x15, - 0xb3, 0x97, 0x8a, 0x25, 0x73, 0xd6, 0xc7, 0x51, 0x2c, 0x05, 0x8f, 0x51, 0x11, 0xcd, 0x58, 0xf4, - 0xb1, 0x45, 0x34, 0xeb, 0x13, 0xc4, 0x10, 0x2f, 0xe0, 0x2c, 0x1f, 0x22, 0x95, 0x0d, 0x3e, 0x6c, - 0xa8, 0xf7, 0xd3, 0x43, 0x5d, 0xcc, 0x4f, 0x3a, 0x2b, 0x86, 0xfc, 0x12, 0x96, 0xc5, 0x90, 0x34, - 0x1e, 0xf8, 0x0c, 0xfb, 0x26, 0xde, 0xb5, 0xd4, 0xe3, 0xbc, 0x99, 0x1e, 0xe7, 0x54, 0x36, 0xac, - 0x90, 0xdd, 0x25, 0xf3, 0x7f, 0x82, 0x16, 0x67, 0xce, 0xb2, 0x96, 0x4a, 0xa6, 0x2b, 0x49, 0xa6, - 0x4d, 0xd9, 0xf1, 0xaf, 0x1a, 0x34, 0x3b, 0x78, 0x10, 0xee, 0x7f, 0x6c, 0xb9, 0xaf, 0xd0, 0x55, - 0x58, 0xa2, 0xff, 0xbb, 0xbe, 0xf9, 0x1f, 0xdc, 0x1b, 0xd1, 0x88, 0x91, 0x33, 0x59, 0x4c, 0x11, - 0x3e, 0xf5, 0x2d, 0x74, 0x1a, 0x9a, 0xa1, 0xfb, 0x9c, 0x70, 0x10, 0x67, 0xda, 0x60, 0x0d, 0x94, - 0x78, 0x0e, 0x5a, 0x3e, 0xe9, 0xfb, 0x24, 0xd8, 0x67, 0x64, 0x1e, 0xd4, 0x83, 0x68, 0xa2, 0x80, - 0xab, 0xd4, 0x97, 0xbb, 0x5e, 0x54, 0x90, 0x50, 0x26, 0x5f, 0x05, 0x64, 0xda, 0x5e, 0xf2, 0xf7, - 0x25, 0x80, 0x68, 0xd1, 0xcc, 0x6d, 0x9b, 0xb6, 0x67, 0x99, 0x86, 0x19, 0x8a, 0xc0, 0x22, 0x65, - 0x2d, 0x22, 0xa4, 0x1e, 0xc1, 0x68, 0x17, 0x0f, 0x07, 0xc1, 0x2b, 0xd7, 0xef, 0xa9, 0x0c, 0x4c, - 0xa2, 0x8b, 0x84, 0xa1, 0xfb, 0x80, 0x0c, 0xcb, 0x24, 0x4e, 0xd8, 0x35, 0x7c, 0xd2, 0x23, 0x4e, - 0x68, 0x62, 0x4b, 0x06, 0x19, 0x39, 0x9d, 0x97, 0x78, 0x87, 0x7b, 0x31, 0x9e, 0x72, 0x49, 0x9f, - 0x90, 0x21, 0x73, 0x88, 0xf9, 0x5c, 0x52, 0x1d, 0xee, 0xd1, 0x4b, 0xff, 0x94, 0xf7, 0xf3, 0x09, - 0xd4, 0x3a, 0xbb, 0x5f, 0x13, 0x23, 0x9c, 0x66, 0xc6, 0xe7, 0xfb, 0x2a, 0x34, 0x3b, 0xf2, 0x36, - 0x40, 0x25, 0x9a, 0x25, 0x99, 0x29, 0x9f, 0xa6, 0xc8, 0x29, 0x27, 0x72, 0x89, 0xa5, 0xc2, 0x5c, - 0x62, 0x39, 0x1b, 0xba, 0x67, 0xf2, 0xd1, 0x95, 0xb1, 0xf2, 0xd1, 0xa3, 0xb7, 0x97, 0x6a, 0xf6, - 0xf6, 0xf2, 0x2f, 0xa9, 0xdb, 0x4b, 0x2d, 0x1b, 0x06, 0x2b, 0xfd, 0x44, 0xf2, 0x02, 0xb3, 0x3d, - 0x72, 0x81, 0xa9, 0x1f, 0xd9, 0x5e, 0xa7, 0xee, 0x31, 0xb7, 0x93, 0x15, 0xc8, 0x86, 0x2a, 0xde, - 0x17, 0xc4, 0x64, 0xd5, 0xf1, 0xd8, 0x15, 0xae, 0x74, 0x4e, 0x0b, 0x32, 0x39, 0xad, 0x64, 0xc9, - 0xa1, 0x35, 0x6e, 0xc9, 0x21, 0x51, 0x3e, 0x99, 0x3d, 0xbc, 0x7c, 0x52, 0x20, 0xed, 0x73, 0x13, - 0x48, 0xfb, 0x6f, 0x2a, 0xd0, 0x8c, 0xce, 0x4e, 0x69, 0x6a, 0xe7, 0xa1, 0x64, 0xca, 0x7c, 0x47, - 0xc9, 0x74, 0x8e, 0x20, 0x8e, 0xc9, 0xec, 0x60, 0xa5, 0x30, 0x3b, 0x58, 0x3d, 0x5a, 0x76, 0xb0, - 0x76, 0x48, 0x76, 0xb0, 0x9e, 0x93, 0x1d, 0x6c, 0x1c, 0x96, 0x1d, 0x6c, 0x16, 0x67, 0x07, 0x61, - 0xc2, 0x8b, 0x59, 0x6b, 0x8c, 0x8b, 0xd9, 0xec, 0x71, 0xb2, 0x83, 0x73, 0xc7, 0xce, 0x0e, 0xce, - 0x4f, 0x20, 0x33, 0xff, 0xa3, 0xc1, 0x8a, 0x32, 0xb8, 0xbc, 0x03, 0xcd, 0x48, 0xe3, 0x55, 0xce, - 0x27, 0xea, 0x44, 0x2f, 0xe3, 0x11, 0xf2, 0xd8, 0x77, 0xf8, 0x10, 0xd6, 0xd4, 0xcf, 0x06, 0xd0, - 0x17, 0xc5, 0x16, 0xfc, 0xa2, 0xe2, 0x21, 0x8a, 0xc2, 0x8c, 0xa9, 0x4d, 0xfa, 0xdf, 0x2a, 0xd0, - 0x88, 0x22, 0xdf, 0x25, 0xa8, 0x24, 0x92, 0x43, 0x65, 0x9d, 0xf4, 0x8f, 0x65, 0xd0, 0x2f, 0x41, - 0x79, 0x8f, 0x84, 0x4a, 0x4f, 0x18, 0xa5, 0x9d, 0x28, 0x82, 0x02, 0xbd, 0x41, 0x28, 0xd2, 0x82, - 0x79, 0x40, 0x6f, 0x10, 0xa2, 0x37, 0xa0, 0xe2, 0xb9, 0x41, 0x28, 0x72, 0x3b, 0x39, 0x48, 0x06, - 0x41, 0xd7, 0xa1, 0xd6, 0x23, 0x16, 0x09, 0x89, 0xb0, 0xd1, 0x39, 0x60, 0x01, 0x42, 0x37, 0xa1, - 0xee, 0xb2, 0x59, 0x2b, 0xed, 0x71, 0x8c, 0x97, 0x28, 0x3a, 0x95, 0x7d, 0x82, 0x7b, 0xc2, 0x10, - 0xe7, 0x4d, 0x85, 0x42, 0xe8, 0xbd, 0xc7, 0xc3, 0xa1, 0xb1, 0x2f, 0x34, 0x31, 0x07, 0xcb, 0x31, - 0x14, 0x1c, 0xfa, 0xd8, 0x90, 0xea, 0x97, 0x07, 0x66, 0x98, 0x31, 0x8d, 0x6e, 0xda, 0xfd, 0xcd, - 0x4d, 0xe0, 0xfe, 0xa6, 0xac, 0x82, 0xdf, 0x6a, 0x50, 0x65, 0xc5, 0xf2, 0x7f, 0xf0, 0xd3, 0x2a, - 0x02, 0x10, 0xab, 0xc4, 0xf8, 0xb9, 0xf9, 0xac, 0x35, 0x55, 0xeb, 0xda, 0x59, 0x68, 0xc6, 0x56, - 0x26, 0xab, 0x6b, 0x1b, 0x43, 0x38, 0x99, 0xfb, 0x64, 0x07, 0xfd, 0x7b, 0xf1, 0xac, 0x2e, 0x65, - 0x66, 0x95, 0x13, 0x87, 0xa8, 0xa7, 0xf6, 0x3b, 0x0d, 0x5a, 0x7a, 0x7e, 0xa2, 0x55, 0x51, 0x3c, - 0x4b, 0xd8, 0xef, 0xd2, 0xd1, 0xec, 0x77, 0xd2, 0xa1, 0x96, 0x47, 0x1c, 0xea, 0x94, 0x5f, 0xb9, - 0x7c, 0xaf, 0xc1, 0x5a, 0xce, 0xf5, 0xfb, 0x83, 0x91, 0x08, 0x8e, 0x1b, 0xf8, 0x13, 0x39, 0x11, - 0xdc, 0xce, 0x4c, 0x3a, 0x70, 0x3b, 0xae, 0x91, 0xff, 0x65, 0x09, 0x1a, 0x32, 0xb8, 0x3b, 0xc2, - 0x26, 0x1f, 0xe3, 0x91, 0x43, 0xe2, 0x7c, 0xca, 0x47, 0x3b, 0x9f, 0xe8, 0xd9, 0x55, 0x65, 0x8c, - 0x67, 0x57, 0x53, 0xbe, 0xb6, 0x7c, 0xa7, 0xc1, 0xb2, 0x2a, 0x95, 0x71, 0x8b, 0xca, 0x0e, 0x6f, - 0x56, 0x3d, 0x72, 0x94, 0x5d, 0x76, 0x66, 0xf4, 0x08, 0x77, 0xec, 0xb3, 0xfa, 0x8b, 0x46, 0xf5, - 0x55, 0x46, 0xdf, 0xef, 0x41, 0xbd, 0xc7, 0x1f, 0x4e, 0xa9, 0xde, 0x1b, 0xaa, 0x52, 0x27, 0x12, - 0x8f, 0x9e, 0xc1, 0xaa, 0x9c, 0x54, 0xba, 0xac, 0x54, 0xca, 0x79, 0x14, 0xa9, 0xe2, 0xb6, 0xec, - 0x2b, 0xb6, 0x65, 0xca, 0xe5, 0x22, 0x1f, 0x56, 0x95, 0x6f, 0x1f, 0xc7, 0x7f, 0xd6, 0xa9, 0x5a, - 0x41, 0x4e, 0xc1, 0xb2, 0x05, 0x35, 0x6e, 0x40, 0xa9, 0x81, 0x70, 0x06, 0x96, 0x85, 0x77, 0x45, - 0xd5, 0xb2, 0xa1, 0x47, 0xdf, 0xe8, 0x23, 0x98, 0xeb, 0x25, 0xdf, 0x93, 0x89, 0x03, 0x4d, 0x79, - 0x8a, 0xd4, 0x83, 0x33, 0x3d, 0x8d, 0x47, 0xa7, 0xa9, 0x34, 0xe0, 0x5e, 0xd7, 0x75, 0xac, 0x61, - 0x6c, 0x7e, 0x70, 0xaf, 0xe3, 0x58, 0x43, 0x74, 0x06, 0xe0, 0x95, 0x6f, 0x86, 0x84, 0x53, 0x79, - 0xb4, 0xdf, 0x64, 0x2d, 0x8c, 0x7c, 0x1e, 0xca, 0x07, 0xb6, 0x25, 0xe2, 0x93, 0x54, 0x40, 0xfc, - 0xcc, 0xb6, 0x74, 0x4a, 0xcb, 0x5e, 0x5e, 0x6b, 0x63, 0x5d, 0x5e, 0x13, 0x61, 0x77, 0xfd, 0x90, - 0xb0, 0x3b, 0x7d, 0xf7, 0x68, 0x64, 0xee, 0x1e, 0x51, 0xe1, 0xb7, 0x99, 0x2c, 0xfc, 0x9e, 0x83, - 0x96, 0x3d, 0xb0, 0x42, 0xd3, 0xb3, 0x48, 0xd7, 0xed, 0xb3, 0x48, 0x44, 0xd3, 0x41, 0x36, 0x75, - 0x58, 0xa0, 0x67, 0xe3, 0x03, 0xd3, 0x1e, 0xd8, 0x2c, 0xf2, 0xd0, 0x74, 0xf9, 0x89, 0xae, 0xc2, - 0x12, 0x39, 0x30, 0xac, 0x41, 0x60, 0xbe, 0x24, 0x5d, 0x89, 0x99, 0x65, 0xe3, 0x2e, 0x46, 0x84, - 0x47, 0x02, 0x4c, 0xd9, 0x98, 0x0e, 0x83, 0xcc, 0x09, 0x36, 0xfc, 0x73, 0x84, 0x8d, 0xc0, 0xcc, - 0x8f, 0xb2, 0x11, 0xe0, 0x33, 0x00, 0x36, 0x3e, 0xe8, 0x5a, 0xc4, 0xd9, 0x0b, 0xf7, 0xdb, 0x0b, - 0xeb, 0xda, 0xe5, 0xb2, 0xde, 0xb4, 0xf1, 0xc1, 0x43, 0xd6, 0xc0, 0xc8, 0xa6, 0x23, 0xc9, 0x8b, - 0x82, 0x6c, 0x3a, 0x82, 0xdc, 0x86, 0xba, 0x87, 0x43, 0xba, 0xbd, 0xed, 0x25, 0x1e, 0xb4, 0x8a, - 0x4f, 0x2a, 0x05, 0x94, 0xaf, 0x19, 0x12, 0x3b, 0x68, 0x23, 0xd6, 0xaf, 0x61, 0xe3, 0x03, 0x56, - 0x00, 0x66, 0x44, 0xd3, 0x11, 0xc4, 0x65, 0x41, 0x34, 0x1d, 0x4e, 0x3c, 0x0f, 0xb3, 0x03, 0xc7, - 0x7c, 0x31, 0x20, 0x82, 0xbe, 0xc2, 0x66, 0xde, 0xe2, 0x6d, 0x1c, 0x72, 0x11, 0xe6, 0x29, 0xf3, - 0x84, 0x7a, 0xac, 0x32, 0x26, 0x73, 0x36, 0x3e, 0x48, 0xc4, 0x17, 0x14, 0x66, 0x3a, 0x49, 0xd8, - 0x9a, 0x80, 0x99, 0x4e, 0x02, 0x96, 0x74, 0x97, 0x27, 0x58, 0x8a, 0x25, 0x76, 0x97, 0x9b, 0x50, - 0x21, 0xce, 0xc0, 0x6e, 0xb7, 0xb3, 0xef, 0xfb, 0xa8, 0xac, 0x30, 0x22, 0xcb, 0xcf, 0x0c, 0x3d, - 0xd2, 0x3e, 0xc9, 0xaf, 0xc1, 0xf4, 0x7f, 0xf4, 0x36, 0xd4, 0xb0, 0x65, 0x51, 0x09, 0x38, 0x75, - 0x94, 0x92, 0x76, 0x15, 0x5b, 0x56, 0xa7, 0x4f, 0x7b, 0xb9, 0x0e, 0x93, 0x9b, 0xd3, 0x47, 0xea, - 0xe5, 0x3a, 0x84, 0xf7, 0xc2, 0xce, 0x90, 0xf6, 0x7a, 0xed, 0x68, 0x63, 0x39, 0xc3, 0x4e, 0x1f, - 0x5d, 0x80, 0xb2, 0xe3, 0x86, 0xed, 0x33, 0x8a, 0xea, 0x2d, 0xeb, 0xa2, 0x53, 0x32, 0x8d, 0x92, - 0xf9, 0x31, 0x9c, 0xcd, 0xda, 0xf5, 0xa8, 0x9a, 0xaf, 0x73, 0x0c, 0x7a, 0x07, 0x20, 0xb1, 0xd9, - 0xe7, 0xb2, 0xde, 0x30, 0xde, 0x75, 0x3d, 0x81, 0xcc, 0xb7, 0x7a, 0xeb, 0x8c, 0x45, 0xca, 0xea, - 0xe5, 0xfd, 0xca, 0x43, 0x6d, 0xf5, 0xd0, 0x5b, 0xb1, 0x37, 0x39, 0x9f, 0x0d, 0x40, 0x12, 0x2f, - 0x74, 0x63, 0x2f, 0x32, 0x12, 0x2d, 0x6c, 0x64, 0xa3, 0x85, 0x35, 0xa8, 0xf5, 0x5d, 0xdf, 0xc6, - 0x61, 0x7b, 0x93, 0x11, 0xc5, 0x57, 0x91, 0x97, 0xb8, 0x30, 0xe1, 0xb3, 0x9a, 0x6c, 0x1d, 0xe0, - 0xda, 0x48, 0x21, 0x58, 0x71, 0x74, 0xec, 0x9d, 0x30, 0x37, 0xf5, 0x53, 0x78, 0x56, 0xa3, 0x78, - 0xe2, 0xff, 0xf3, 0x85, 0xee, 0xab, 0xb0, 0xac, 0x48, 0x7f, 0x6d, 0xfc, 0xa1, 0x04, 0xf3, 0xe9, - 0x2a, 0x43, 0xa4, 0x75, 0x5a, 0x42, 0xeb, 0x0e, 0x7f, 0x75, 0x23, 0x53, 0x56, 0xe5, 0x4c, 0xca, - 0xaa, 0x12, 0xa5, 0xac, 0xd6, 0xc4, 0x0e, 0xcb, 0x77, 0x64, 0xe2, 0x0b, 0x6d, 0xc2, 0xdc, 0x2e, - 0xc1, 0x3e, 0xf1, 0xbb, 0xe2, 0xf4, 0xf9, 0x7b, 0x99, 0x59, 0xde, 0xf8, 0x31, 0x97, 0x81, 0x6b, - 0x50, 0xed, 0x5b, 0xee, 0x2b, 0xf9, 0x38, 0x7f, 0x4d, 0x99, 0x97, 0x0e, 0x74, 0x0e, 0x42, 0xd7, - 0x61, 0x99, 0xd2, 0xbb, 0x66, 0xaf, 0x6b, 0xb8, 0x8e, 0x43, 0x8c, 0x90, 0x95, 0x0c, 0x1a, 0xbc, - 0xec, 0x40, 0x49, 0x0f, 0x7a, 0xf7, 0x38, 0xe1, 0xd3, 0xe2, 0x57, 0x2b, 0x93, 0x3c, 0x05, 0xfe, - 0x41, 0x83, 0x93, 0xf9, 0x55, 0xa0, 0x6d, 0x58, 0x18, 0xf9, 0xc5, 0x81, 0x90, 0xb8, 0x53, 0xf9, - 0xb5, 0x9f, 0x9d, 0x19, 0x7d, 0x3e, 0xfd, 0x2b, 0x83, 0x63, 0x4b, 0xe0, 0x37, 0x70, 0xba, 0xe0, - 0x47, 0x0d, 0xa8, 0x5b, 0x2c, 0x89, 0x57, 0xb2, 0x92, 0x98, 0x5b, 0xac, 0x52, 0x4b, 0xe4, 0x6f, - 0x35, 0xa8, 0xf1, 0x0b, 0xbe, 0x7c, 0x46, 0xa4, 0xc5, 0xcf, 0x88, 0x0e, 0x17, 0xb8, 0xf7, 0xa0, - 0xf9, 0x52, 0x14, 0xac, 0x64, 0xc5, 0xe3, 0x74, 0x7e, 0x4d, 0x2b, 0xd0, 0x63, 0xf4, 0xb4, 0xef, - 0x6a, 0x3f, 0x6a, 0x54, 0x87, 0x46, 0x0b, 0x70, 0xcc, 0xbd, 0x89, 0xca, 0x02, 0xf3, 0x66, 0xed, - 0xd8, 0x62, 0x8a, 0x44, 0x54, 0x8e, 0x61, 0x54, 0x24, 0xa2, 0xa6, 0x3c, 0xe3, 0x3d, 0x58, 0x18, - 0xd9, 0x1e, 0xf4, 0xb4, 0xf8, 0xb8, 0xcf, 0x29, 0x8e, 0x3b, 0x55, 0x33, 0x54, 0x9f, 0xf1, 0x0b, - 0x58, 0x7b, 0xa2, 0x9c, 0xc2, 0xcf, 0xf7, 0xbb, 0x8d, 0x4d, 0x68, 0xf1, 0x4a, 0xdf, 0x96, 0xef, - 0xe3, 0x61, 0x5c, 0xa1, 0xe4, 0x47, 0x21, 0x2a, 0x94, 0x9f, 0x43, 0x5d, 0x94, 0x03, 0xd1, 0xc3, - 0xe2, 0x85, 0x9f, 0xc8, 0x2e, 0x9c, 0x75, 0xcc, 0x59, 0xf0, 0x4f, 0x1a, 0x94, 0x9f, 0x62, 0x75, - 0xb1, 0xf4, 0x70, 0x99, 0xce, 0xc4, 0xe0, 0xe5, 0x69, 0xfd, 0xa0, 0x61, 0x12, 0x29, 0xf9, 0xa3, - 0x06, 0xe5, 0x67, 0xb6, 0xa5, 0x5c, 0xcb, 0x6b, 0xd0, 0xa4, 0x7f, 0x03, 0x0f, 0x1b, 0xb2, 0xf8, - 0x1b, 0x37, 0x50, 0x43, 0xef, 0xf9, 0xa4, 0x6f, 0x1e, 0x08, 0x59, 0x16, 0x5f, 0xb4, 0x17, 0x0e, - 0x43, 0xdf, 0xdc, 0x1d, 0x84, 0xf2, 0xad, 0x71, 0xdc, 0x40, 0x15, 0xe4, 0x95, 0x8f, 0x3d, 0x2f, - 0x2a, 0x48, 0xc8, 0xcf, 0x29, 0xff, 0x30, 0xe1, 0xee, 0xeb, 0x30, 0xef, 0xfa, 0x7b, 0xb2, 0x4b, - 0xf7, 0xe5, 0xed, 0xbb, 0xb3, 0xe2, 0xe7, 0xb1, 0x8f, 0x7d, 0x37, 0x74, 0x1f, 0x6b, 0x3f, 0x94, - 0xca, 0x9d, 0xad, 0x27, 0xbb, 0x35, 0xf6, 0x5b, 0xd2, 0xdb, 0x7f, 0x0f, 0x00, 0x00, 0xff, 0xff, - 0x29, 0xfd, 0x14, 0x6b, 0x47, 0x3b, 0x00, 0x00, + 0x2d, 0x79, 0xb0, 0xa2, 0x72, 0xe7, 0xe8, 0x59, 0xf1, 0xa1, 0x6e, 0x66, 0x0f, 0x35, 0xb3, 0x98, + 0x9c, 0xe3, 0xfd, 0x1c, 0x20, 0x11, 0x98, 0x3f, 0x28, 0x1e, 0x47, 0xbd, 0xa9, 0x6a, 0xc6, 0xff, + 0xaf, 0xc1, 0x6c, 0xd2, 0x26, 0x8c, 0x8a, 0x88, 0x96, 0x15, 0x91, 0xac, 0xa3, 0x2b, 0x38, 0xe6, + 0xf2, 0x04, 0xc7, 0xfc, 0xdf, 0x15, 0xa8, 0x71, 0x7d, 0x3f, 0xc2, 0x6c, 0x4e, 0x41, 0xc3, 0xe7, + 0x26, 0xb3, 0xc7, 0xfd, 0x9a, 0x1e, 0x7d, 0xa3, 0xb3, 0x00, 0x3d, 0xe2, 0xf9, 0xc4, 0xc0, 0x21, + 0xe9, 0x31, 0x89, 0x6e, 0xe8, 0x89, 0x16, 0x74, 0x05, 0x96, 0xb8, 0x1d, 0x20, 0xb6, 0x17, 0x0e, + 0x13, 0x82, 0xdc, 0xd0, 0x17, 0x18, 0x61, 0x9b, 0xb6, 0x73, 0x51, 0x8e, 0x6c, 0x4c, 0x35, 0xc7, + 0xc6, 0xd4, 0x0e, 0xb3, 0x31, 0x75, 0x95, 0x8d, 0xb9, 0x03, 0x35, 0x1e, 0xc5, 0xaa, 0x62, 0xa3, + 0xcc, 0x35, 0x51, 0x17, 0x60, 0xf4, 0x46, 0x2c, 0xf2, 0x4d, 0xb5, 0xa2, 0x46, 0xc2, 0x9e, 0x8c, + 0x4b, 0x61, 0xec, 0xb8, 0xf4, 0x4d, 0xa8, 0x0b, 0x23, 0xdd, 0x6e, 0x65, 0x5d, 0xdd, 0x23, 0xd2, + 0x33, 0x31, 0xb5, 0xde, 0x81, 0x2e, 0x61, 0x45, 0xc2, 0x30, 0x3b, 0x81, 0x30, 0xfc, 0xa7, 0x06, + 0x4b, 0x5c, 0x18, 0x92, 0x2a, 0x7f, 0x0d, 0x6a, 0xdc, 0xf8, 0x0b, 0x8d, 0x47, 0x59, 0x5f, 0x41, + 0x83, 0x18, 0x8e, 0x39, 0xb6, 0xbe, 0x3b, 0xb0, 0xac, 0xf0, 0x3f, 0xe8, 0xf3, 0x62, 0x35, 0xdc, + 0xc8, 0xac, 0x33, 0xb3, 0x8e, 0x1c, 0xa5, 0xfc, 0xb1, 0x04, 0x15, 0x1a, 0xc8, 0x50, 0xa1, 0x0b, + 0xcd, 0xd0, 0x92, 0xfe, 0x92, 0x7f, 0x1c, 0xc1, 0x8a, 0x5f, 0x86, 0xc5, 0x90, 0xf8, 0x76, 0xd0, + 0x75, 0xfb, 0x5d, 0x2a, 0x68, 0xa6, 0x21, 0x7d, 0xe3, 0x3c, 0x6b, 0xef, 0xf4, 0x9f, 0xf0, 0x56, + 0x1a, 0xa0, 0x19, 0x3c, 0xa8, 0x15, 0x81, 0xcf, 0x72, 0x3a, 0x94, 0x61, 0x24, 0x5d, 0x62, 0x28, + 0xdc, 0x32, 0x0d, 0xe2, 0x04, 0x44, 0x44, 0x3e, 0xcb, 0xe9, 0x58, 0x9e, 0x91, 0x74, 0x89, 0xa1, + 0xea, 0x41, 0x83, 0x2b, 0xee, 0x31, 0x99, 0x27, 0x12, 0x9f, 0x45, 0x52, 0x52, 0x9f, 0x40, 0x4a, + 0xbe, 0x80, 0x26, 0xbd, 0xa8, 0xf3, 0x3c, 0xcc, 0xa3, 0xbc, 0x3c, 0x4c, 0xf9, 0x70, 0x05, 0xcb, + 0x66, 0x61, 0x36, 0xbe, 0x81, 0xba, 0x58, 0xd8, 0x11, 0xe3, 0xfe, 0x29, 0x9b, 0xc3, 0x3f, 0x97, + 0xa0, 0x42, 0x6f, 0x49, 0x34, 0xe2, 0xa6, 0x22, 0xc2, 0x79, 0xfa, 0xa4, 0x2f, 0x23, 0xee, 0xa8, + 0x51, 0x27, 0x7d, 0x1a, 0x69, 0xc5, 0x20, 0xb3, 0x27, 0xa5, 0x23, 0x6a, 0x7b, 0xd0, 0x43, 0x1f, + 0x2a, 0x6e, 0xbb, 0x67, 0x46, 0xec, 0x47, 0x3a, 0x1b, 0x93, 0xba, 0xe8, 0xfe, 0x33, 0xcc, 0x26, + 0xae, 0xaa, 0x43, 0x21, 0x37, 0xa7, 0x47, 0x18, 0xa4, 0x72, 0x31, 0xad, 0xf8, 0x7a, 0x9a, 0x09, + 0x42, 0xaa, 0x59, 0xf1, 0xbd, 0x02, 0x35, 0x1e, 0x96, 0x8b, 0xdb, 0xab, 0x2a, 0x70, 0x17, 0x88, + 0x69, 0x0b, 0xd2, 0x10, 0x16, 0xe8, 0x5e, 0x27, 0x6d, 0xcd, 0xeb, 0x50, 0xa1, 0x37, 0x52, 0x61, + 0x69, 0x16, 0x47, 0x2f, 0xaf, 0x3b, 0x33, 0x3a, 0xa3, 0x1f, 0xdb, 0xca, 0x7c, 0x0d, 0x4b, 0x99, + 0xcb, 0xf0, 0xf8, 0x69, 0xb0, 0x91, 0xd9, 0xe7, 0x58, 0x98, 0x5f, 0x97, 0xa0, 0x19, 0x19, 0xef, + 0x84, 0x13, 0xd2, 0x26, 0x74, 0x42, 0xa5, 0x31, 0x9c, 0x50, 0x79, 0x6c, 0x27, 0xf4, 0x16, 0x34, + 0x88, 0x08, 0x7f, 0x85, 0xb4, 0xa5, 0x36, 0x36, 0x8a, 0xaa, 0xf5, 0x08, 0x36, 0xed, 0xc8, 0xf3, + 0x2b, 0x80, 0xd8, 0xd7, 0xa1, 0x4e, 0xf1, 0xa1, 0x9c, 0xca, 0xb0, 0x8e, 0xfa, 0xe6, 0x1c, 0xc7, + 0x36, 0x34, 0xe4, 0x14, 0x94, 0x36, 0x26, 0x8a, 0xca, 0x4b, 0x45, 0x51, 0xf9, 0xc6, 0x57, 0xb0, + 0xa2, 0x4a, 0x85, 0x2a, 0x59, 0xbe, 0x95, 0x66, 0x59, 0xa8, 0xbe, 0x82, 0x3d, 0x81, 0x76, 0x5e, + 0x8a, 0x4f, 0x39, 0xc4, 0x9d, 0xf4, 0x10, 0xe7, 0x54, 0xd9, 0x9b, 0xa4, 0x5c, 0x89, 0x61, 0x3a, + 0x30, 0x97, 0xba, 0xf1, 0x28, 0x79, 0x5f, 0x49, 0xf3, 0x5e, 0x51, 0xc9, 0x83, 0x64, 0x68, 0xc0, + 0x89, 0x9c, 0x68, 0x5b, 0xc9, 0xfa, 0xed, 0x34, 0xeb, 0xb3, 0x0a, 0x41, 0x55, 0xcc, 0x1a, 0xc3, + 0x9a, 0xda, 0xc7, 0x2b, 0xc7, 0xb8, 0x9d, 0x1e, 0xe3, 0x4c, 0x36, 0x7c, 0x51, 0x0c, 0x21, 0x8f, + 0x77, 0xd4, 0x40, 0x8d, 0x7b, 0xbc, 0xa3, 0x26, 0x42, 0xb0, 0xff, 0x57, 0x98, 0x4f, 0x0b, 0xab, + 0x92, 0xf1, 0xd5, 0x34, 0xe3, 0x55, 0x65, 0x38, 0x28, 0x59, 0xee, 0xc1, 0x49, 0x91, 0x48, 0x17, + 0xde, 0xe3, 0xb0, 0x69, 0xbf, 0x93, 0xe6, 0xbe, 0xae, 0xcc, 0xc1, 0x16, 0xc8, 0x8c, 0xcc, 0xd8, + 0x8f, 0x2d, 0x33, 0x51, 0xaa, 0x5f, 0x30, 0x7c, 0x0e, 0xa7, 0x19, 0xc3, 0x38, 0xe5, 0x3a, 0x3c, + 0x6c, 0xee, 0xef, 0xa6, 0xd9, 0x6f, 0xa8, 0x33, 0xb7, 0x43, 0xc5, 0xec, 0xa5, 0x62, 0xc9, 0x3c, + 0xf5, 0x71, 0x14, 0x4b, 0xc1, 0x63, 0x54, 0x44, 0x33, 0x16, 0x7d, 0x6c, 0x11, 0xcd, 0xfa, 0x04, + 0x31, 0xc4, 0x0b, 0x38, 0xcb, 0x87, 0x48, 0x65, 0x80, 0x0f, 0x1b, 0xea, 0xfd, 0xf4, 0x50, 0x17, + 0xf3, 0x13, 0xcd, 0x8a, 0x21, 0xbf, 0x84, 0x65, 0x31, 0x24, 0x8d, 0x07, 0x3e, 0xc3, 0xbe, 0x89, + 0x77, 0x2d, 0xf5, 0x38, 0x6f, 0xa6, 0xc7, 0x39, 0x95, 0x0d, 0x2b, 0x64, 0x77, 0xc9, 0xfc, 0x9f, + 0xa0, 0xc5, 0x99, 0xb3, 0x4c, 0xa5, 0x92, 0xe9, 0x4a, 0x92, 0x69, 0x53, 0x76, 0xfc, 0xab, 0x06, + 0xcd, 0x0e, 0x1e, 0x84, 0xfb, 0x1f, 0x5b, 0xee, 0x2b, 0x74, 0x15, 0x96, 0xe8, 0xff, 0xae, 0x6f, + 0xfe, 0x07, 0xf7, 0x46, 0x34, 0x62, 0xe4, 0x4c, 0x16, 0x53, 0x84, 0x4f, 0x7d, 0x0b, 0x9d, 0x86, + 0x66, 0xe8, 0x3e, 0x27, 0x1c, 0xc4, 0x99, 0x36, 0x58, 0x03, 0x25, 0x9e, 0x83, 0x96, 0x4f, 0xfa, + 0x3e, 0x09, 0xf6, 0x19, 0x99, 0x07, 0xf5, 0x20, 0x9a, 0x28, 0xe0, 0x2a, 0xf5, 0xe5, 0xae, 0x17, + 0x15, 0x21, 0x94, 0x09, 0x57, 0x01, 0x99, 0xb6, 0x97, 0xfc, 0x7d, 0x09, 0x20, 0x5a, 0x34, 0x73, + 0xdb, 0xa6, 0xed, 0x59, 0xa6, 0x61, 0x86, 0x22, 0xb0, 0x48, 0x59, 0x8b, 0x08, 0xa9, 0x47, 0x30, + 0xda, 0xc5, 0xc3, 0x41, 0xf0, 0xca, 0xf5, 0x7b, 0x2a, 0x03, 0x93, 0xe8, 0x22, 0x61, 0xe8, 0x3e, + 0x20, 0xc3, 0x32, 0x89, 0x13, 0x76, 0x0d, 0x9f, 0xf4, 0x88, 0x13, 0x9a, 0xd8, 0x92, 0x41, 0x46, + 0x4e, 0xe7, 0x25, 0xde, 0xe1, 0x5e, 0x8c, 0xa7, 0x5c, 0xd2, 0x27, 0x64, 0xc8, 0xbc, 0x61, 0x3e, + 0x97, 0x54, 0x87, 0x7b, 0xf4, 0xd2, 0x3f, 0xe5, 0xfd, 0x7c, 0x02, 0xb5, 0xce, 0xee, 0xd7, 0xc4, + 0x08, 0xa7, 0x99, 0xf1, 0xf9, 0xbe, 0x0a, 0xcd, 0x8e, 0xbc, 0x0d, 0x50, 0x89, 0x66, 0x89, 0x65, + 0xca, 0xa7, 0x29, 0xf2, 0xc8, 0x89, 0xfc, 0x61, 0xa9, 0x30, 0x7f, 0x58, 0xce, 0x86, 0xee, 0x99, + 0x1c, 0x74, 0x65, 0xac, 0x1c, 0xf4, 0xe8, 0xed, 0xa5, 0x9a, 0xbd, 0xbd, 0xfc, 0x4b, 0xea, 0xf6, + 0x52, 0xcb, 0x86, 0xc1, 0x4a, 0x3f, 0x91, 0xbc, 0xc0, 0x6c, 0x8f, 0x5c, 0x60, 0xea, 0x47, 0xb6, + 0xd7, 0xa9, 0x7b, 0xcc, 0xed, 0x64, 0xd5, 0xb1, 0xa1, 0x8a, 0xf7, 0x05, 0x31, 0x59, 0x69, 0x3c, + 0x76, 0x55, 0x2b, 0x9d, 0xd3, 0x82, 0x4c, 0x4e, 0x2b, 0x59, 0x66, 0x68, 0x8d, 0x5b, 0x66, 0x48, + 0x94, 0x4c, 0x66, 0x0f, 0x2f, 0x99, 0x14, 0x48, 0xfb, 0xdc, 0x04, 0xd2, 0xfe, 0x9b, 0x0a, 0x34, + 0xa3, 0xb3, 0x53, 0x9a, 0xda, 0x79, 0x28, 0x99, 0x32, 0xdf, 0x51, 0x32, 0x9d, 0x23, 0x88, 0x63, + 0x32, 0x3b, 0x58, 0x29, 0xcc, 0x0e, 0x56, 0x8f, 0x96, 0x1d, 0xac, 0x1d, 0x92, 0x1d, 0xac, 0xe7, + 0x64, 0x07, 0x1b, 0x87, 0x65, 0x07, 0x9b, 0xc5, 0xd9, 0x41, 0x98, 0xf0, 0x62, 0xd6, 0x1a, 0xe3, + 0x62, 0x36, 0x7b, 0x9c, 0xec, 0xe0, 0xdc, 0xb1, 0xb3, 0x83, 0xf3, 0x13, 0xc8, 0xcc, 0xff, 0x68, + 0xb0, 0xa2, 0x0c, 0x2e, 0xef, 0x40, 0x33, 0xd2, 0x78, 0x95, 0xf3, 0x89, 0x3a, 0xd1, 0xcb, 0x78, + 0x84, 0x3c, 0xf6, 0x1d, 0x3e, 0x84, 0x35, 0xf5, 0x53, 0x01, 0xf4, 0x45, 0xb1, 0x05, 0xbf, 0xa8, + 0x78, 0x7c, 0xa2, 0x30, 0x63, 0x6a, 0x93, 0xfe, 0xb7, 0x0a, 0x34, 0xa2, 0xc8, 0x77, 0x09, 0x2a, + 0x89, 0xe4, 0x50, 0x59, 0x27, 0xfd, 0x63, 0x19, 0xf4, 0x4b, 0x50, 0xde, 0x23, 0xa1, 0xd2, 0x13, + 0x46, 0x69, 0x27, 0x8a, 0xa0, 0x40, 0x6f, 0x10, 0x8a, 0xb4, 0x60, 0x1e, 0xd0, 0x1b, 0x84, 0xe8, + 0x0d, 0xa8, 0x78, 0x6e, 0x10, 0x8a, 0xdc, 0x4e, 0x0e, 0x92, 0x41, 0xd0, 0x75, 0xa8, 0xf5, 0x88, + 0x45, 0x42, 0x22, 0x6c, 0x74, 0x0e, 0x58, 0x80, 0xd0, 0x4d, 0xa8, 0xbb, 0x6c, 0xd6, 0x4a, 0x7b, + 0x1c, 0xe3, 0x25, 0x8a, 0x4e, 0x65, 0x9f, 0xe0, 0x9e, 0x30, 0xc4, 0x79, 0x53, 0xa1, 0x10, 0x7a, + 0xef, 0xf1, 0x70, 0x68, 0xec, 0x0b, 0x4d, 0xcc, 0xc1, 0x72, 0x0c, 0x05, 0x87, 0x3e, 0x36, 0xa4, + 0xfa, 0xe5, 0x81, 0x19, 0x66, 0x4c, 0xa3, 0x9b, 0x76, 0x7f, 0x73, 0x13, 0xb8, 0xbf, 0x29, 0xab, + 0xe0, 0xb7, 0x1a, 0x54, 0x59, 0x81, 0xfc, 0x1f, 0xfc, 0x9c, 0x8a, 0x00, 0xc4, 0x2a, 0x31, 0x7e, + 0x6e, 0x3e, 0x6b, 0x4d, 0xd5, 0xba, 0x76, 0x16, 0x9a, 0xb1, 0x95, 0xc9, 0xea, 0xda, 0xc6, 0x10, + 0x4e, 0xe6, 0x3e, 0xd3, 0x41, 0xff, 0x5e, 0x3c, 0xab, 0x4b, 0x99, 0x59, 0xe5, 0xc4, 0x21, 0xea, + 0xa9, 0xfd, 0x4e, 0x83, 0x96, 0x9e, 0x9f, 0x68, 0x55, 0x14, 0xcf, 0x12, 0xf6, 0xbb, 0x74, 0x34, + 0xfb, 0x9d, 0x74, 0xa8, 0xe5, 0x11, 0x87, 0x3a, 0xe5, 0x97, 0x2d, 0xdf, 0x6b, 0xb0, 0x96, 0x73, + 0xfd, 0xfe, 0x60, 0x24, 0x82, 0xe3, 0x06, 0xfe, 0x44, 0x4e, 0x04, 0xb7, 0x33, 0x93, 0x0e, 0xdc, + 0x8e, 0x6b, 0xe4, 0x7f, 0x59, 0x82, 0x86, 0x0c, 0xee, 0x8e, 0xb0, 0xc9, 0xc7, 0x78, 0xd8, 0x90, + 0x38, 0x9f, 0xf2, 0xd1, 0xce, 0x27, 0x7a, 0x6a, 0x55, 0x19, 0xe3, 0xa9, 0xd5, 0x94, 0xaf, 0x2d, + 0xdf, 0x69, 0xb0, 0xac, 0x4a, 0x65, 0xdc, 0xa2, 0xb2, 0xc3, 0x9b, 0x55, 0x0f, 0x1b, 0x65, 0x97, + 0x9d, 0x19, 0x3d, 0xc2, 0x1d, 0xfb, 0xac, 0xfe, 0xa2, 0x51, 0x7d, 0x95, 0xd1, 0xf7, 0x7b, 0x50, + 0xef, 0xf1, 0xc7, 0x52, 0xaa, 0x37, 0x86, 0xaa, 0xd4, 0x89, 0xc4, 0xa3, 0x67, 0xb0, 0x2a, 0x27, + 0x95, 0x2e, 0x2b, 0x95, 0x72, 0x1e, 0x42, 0xaa, 0xb8, 0x2d, 0xfb, 0x8a, 0x6d, 0x99, 0x72, 0xb9, + 0xc8, 0x87, 0x55, 0xe5, 0x7b, 0xc7, 0xf1, 0x9f, 0x72, 0xaa, 0x56, 0x90, 0x53, 0xb0, 0x6c, 0x41, + 0x8d, 0x1b, 0x50, 0x6a, 0x20, 0x9c, 0x81, 0x65, 0xe1, 0x5d, 0x51, 0xb5, 0x6c, 0xe8, 0xd1, 0x37, + 0xfa, 0x08, 0xe6, 0x7a, 0xc9, 0x37, 0x64, 0xe2, 0x40, 0x53, 0x9e, 0x22, 0xf5, 0xc8, 0x4c, 0x4f, + 0xe3, 0xd1, 0x69, 0x2a, 0x0d, 0xb8, 0xd7, 0x75, 0x1d, 0x6b, 0x18, 0x9b, 0x1f, 0xdc, 0xeb, 0x38, + 0xd6, 0x10, 0x9d, 0x01, 0x78, 0xe5, 0x9b, 0x21, 0xe1, 0x54, 0x1e, 0xed, 0x37, 0x59, 0x0b, 0x23, + 0x9f, 0x87, 0xf2, 0x81, 0x6d, 0x89, 0xf8, 0x24, 0x15, 0x10, 0x3f, 0xb3, 0x2d, 0x9d, 0xd2, 0xb2, + 0x97, 0xd7, 0xda, 0x58, 0x97, 0xd7, 0x44, 0xd8, 0x5d, 0x3f, 0x24, 0xec, 0x4e, 0xdf, 0x3d, 0x1a, + 0x99, 0xbb, 0x47, 0x54, 0xf8, 0x6d, 0x26, 0x0b, 0xbf, 0xe7, 0xa0, 0x65, 0x0f, 0xac, 0xd0, 0xf4, + 0x2c, 0xd2, 0x75, 0xfb, 0x2c, 0x12, 0xd1, 0x74, 0x90, 0x4d, 0x1d, 0x16, 0xe8, 0xd9, 0xf8, 0xc0, + 0xb4, 0x07, 0x36, 0x8b, 0x3c, 0x34, 0x5d, 0x7e, 0xa2, 0xab, 0xb0, 0x44, 0x0e, 0x0c, 0x6b, 0x10, + 0x98, 0x2f, 0x49, 0x57, 0x62, 0x66, 0xd9, 0xb8, 0x8b, 0x11, 0xe1, 0x91, 0x00, 0x53, 0x36, 0xa6, + 0xc3, 0x20, 0x73, 0x82, 0x0d, 0xff, 0x1c, 0x61, 0x23, 0x30, 0xf3, 0xa3, 0x6c, 0x04, 0xf8, 0x0c, + 0x80, 0x8d, 0x0f, 0xba, 0x16, 0x71, 0xf6, 0xc2, 0xfd, 0xf6, 0xc2, 0xba, 0x76, 0xb9, 0xac, 0x37, + 0x6d, 0x7c, 0xf0, 0x90, 0x35, 0x30, 0xb2, 0xe9, 0x48, 0xf2, 0xa2, 0x20, 0x9b, 0x8e, 0x20, 0xb7, + 0xa1, 0xee, 0xe1, 0x90, 0x6e, 0x6f, 0x7b, 0x89, 0x07, 0xad, 0xe2, 0x93, 0x4a, 0x01, 0xe5, 0x6b, + 0x86, 0xc4, 0x0e, 0xda, 0x88, 0xf5, 0x6b, 0xd8, 0xf8, 0x80, 0x15, 0x80, 0x19, 0xd1, 0x74, 0x04, + 0x71, 0x59, 0x10, 0x4d, 0x87, 0x13, 0xcf, 0xc3, 0xec, 0xc0, 0x31, 0x5f, 0x0c, 0x88, 0xa0, 0xaf, + 0xb0, 0x99, 0xb7, 0x78, 0x1b, 0x87, 0x5c, 0x84, 0x79, 0xca, 0x3c, 0xa1, 0x1e, 0xab, 0x8c, 0xc9, + 0x9c, 0x8d, 0x0f, 0x12, 0xf1, 0x05, 0x85, 0x99, 0x4e, 0x12, 0xb6, 0x26, 0x60, 0xa6, 0x93, 0x80, + 0x25, 0xdd, 0xe5, 0x09, 0x96, 0x62, 0x89, 0xdd, 0xe5, 0x26, 0x54, 0x88, 0x33, 0xb0, 0xdb, 0xed, + 0xec, 0x9b, 0x3e, 0x2a, 0x2b, 0x8c, 0xc8, 0xf2, 0x33, 0x43, 0x8f, 0xb4, 0x4f, 0xf2, 0x6b, 0x30, + 0xfd, 0x1f, 0xbd, 0x0d, 0x35, 0x6c, 0x59, 0x54, 0x02, 0x4e, 0x1d, 0xa5, 0xa4, 0x5d, 0xc5, 0x96, + 0xd5, 0xe9, 0xd3, 0x5e, 0xae, 0xc3, 0xe4, 0xe6, 0xf4, 0x91, 0x7a, 0xb9, 0x0e, 0xe1, 0xbd, 0xb0, + 0x33, 0xa4, 0xbd, 0x5e, 0x3b, 0xda, 0x58, 0xce, 0xb0, 0xd3, 0x47, 0x17, 0xa0, 0xec, 0xb8, 0x61, + 0xfb, 0x8c, 0xa2, 0x7a, 0xcb, 0xba, 0xe8, 0x94, 0x4c, 0xa3, 0x64, 0x7e, 0x0c, 0x67, 0xb3, 0x76, + 0x3d, 0xaa, 0xe6, 0xeb, 0x1c, 0x83, 0xde, 0x01, 0x48, 0x6c, 0xf6, 0xb9, 0xac, 0x37, 0x8c, 0x77, + 0x5d, 0x4f, 0x20, 0xf3, 0xad, 0xde, 0x3a, 0x63, 0x91, 0xb2, 0x7a, 0x79, 0xbf, 0xec, 0x50, 0x5b, + 0x3d, 0xf4, 0x56, 0xec, 0x4d, 0xce, 0x67, 0x03, 0x90, 0xc4, 0xab, 0xdc, 0xd8, 0x8b, 0x8c, 0x44, + 0x0b, 0x1b, 0xd9, 0x68, 0x61, 0x0d, 0x6a, 0x7d, 0xd7, 0xb7, 0x71, 0xd8, 0xde, 0x64, 0x44, 0xf1, + 0x55, 0xe4, 0x25, 0x2e, 0x4c, 0xf8, 0xac, 0x26, 0x5b, 0x07, 0xb8, 0x36, 0x52, 0x08, 0x56, 0x1c, + 0x1d, 0x7b, 0x1b, 0xcc, 0x4d, 0xfd, 0x14, 0x9e, 0xd5, 0x28, 0x9e, 0xf5, 0xff, 0x7c, 0xa1, 0xfb, + 0x2a, 0x2c, 0x2b, 0xd2, 0x5f, 0x1b, 0x7f, 0x28, 0xc1, 0x7c, 0xba, 0xca, 0x10, 0x69, 0x9d, 0x96, + 0xd0, 0xba, 0xc3, 0x5f, 0xdd, 0xc8, 0x94, 0x55, 0x39, 0x93, 0xb2, 0xaa, 0x44, 0x29, 0xab, 0x35, + 0xb1, 0xc3, 0xf2, 0x1d, 0x99, 0xf8, 0x42, 0x9b, 0x30, 0xb7, 0x4b, 0xb0, 0x4f, 0xfc, 0xae, 0x38, + 0x7d, 0xfe, 0x5e, 0x66, 0x96, 0x37, 0x7e, 0xcc, 0x65, 0xe0, 0x1a, 0x54, 0xfb, 0x96, 0xfb, 0x4a, + 0x3e, 0xc8, 0x5f, 0x53, 0xe6, 0xa5, 0x03, 0x9d, 0x83, 0xd0, 0x75, 0x58, 0xa6, 0xf4, 0xae, 0xd9, + 0xeb, 0x1a, 0xae, 0xe3, 0x10, 0x23, 0x64, 0x25, 0x83, 0x06, 0x2f, 0x3b, 0x50, 0xd2, 0x83, 0xde, + 0x3d, 0x4e, 0xf8, 0xb4, 0xf8, 0xd5, 0xca, 0x24, 0xcf, 0x7f, 0x7f, 0xd0, 0xe0, 0x64, 0x7e, 0x15, + 0x68, 0x1b, 0x16, 0x46, 0x7e, 0x65, 0x20, 0x24, 0xee, 0x54, 0x7e, 0xed, 0x67, 0x67, 0x46, 0x9f, + 0x4f, 0xff, 0xb2, 0xe0, 0xd8, 0x12, 0xf8, 0x0d, 0x9c, 0x2e, 0xf8, 0x21, 0x03, 0xea, 0x16, 0x4b, + 0xe2, 0x95, 0xac, 0x24, 0xe6, 0x16, 0xab, 0xd4, 0x12, 0xf9, 0x5b, 0x0d, 0x6a, 0xfc, 0x82, 0x2f, + 0x9f, 0x11, 0x69, 0xf1, 0x33, 0xa2, 0xc3, 0x05, 0xee, 0x3d, 0x68, 0xbe, 0x14, 0x05, 0x2b, 0x59, + 0xf1, 0x38, 0x9d, 0x5f, 0xd3, 0x0a, 0xf4, 0x18, 0x3d, 0xed, 0xbb, 0xda, 0x8f, 0x1a, 0xd5, 0xa1, + 0xd1, 0x02, 0x1c, 0x73, 0x6f, 0xa2, 0xb2, 0xc0, 0xbc, 0x59, 0x3b, 0xb6, 0x98, 0x22, 0x11, 0x95, + 0x63, 0x18, 0x15, 0x89, 0xa8, 0x29, 0xcf, 0x78, 0x0f, 0x16, 0x46, 0xb6, 0x07, 0x3d, 0x2d, 0x3e, + 0xee, 0x73, 0x8a, 0xe3, 0x4e, 0xd5, 0x0c, 0xd5, 0x67, 0xfc, 0x02, 0xd6, 0x9e, 0x28, 0xa7, 0xf0, + 0xf3, 0xfd, 0x56, 0x63, 0x13, 0x5a, 0xbc, 0xd2, 0xb7, 0xe5, 0xfb, 0x78, 0x18, 0x57, 0x28, 0xf9, + 0x51, 0x88, 0x0a, 0xe5, 0xe7, 0x50, 0x17, 0xe5, 0x40, 0xf4, 0xb0, 0x78, 0xe1, 0x27, 0xb2, 0x0b, + 0x67, 0x1d, 0x73, 0x16, 0xfc, 0x93, 0x06, 0xe5, 0xa7, 0x58, 0x5d, 0x2c, 0x3d, 0x5c, 0xa6, 0x33, + 0x31, 0x78, 0x79, 0x5a, 0x3f, 0x62, 0x98, 0x44, 0x4a, 0xfe, 0xa8, 0x41, 0xf9, 0x99, 0x6d, 0x29, + 0xd7, 0xf2, 0x1a, 0x34, 0xe9, 0xdf, 0xc0, 0xc3, 0x86, 0x2c, 0xfe, 0xc6, 0x0d, 0xd4, 0xd0, 0x7b, + 0x3e, 0xe9, 0x9b, 0x07, 0x42, 0x96, 0xc5, 0x17, 0xed, 0x85, 0xc3, 0xd0, 0x37, 0x77, 0x07, 0xa1, + 0x7c, 0x6b, 0x1c, 0x37, 0x50, 0x05, 0x79, 0xe5, 0x63, 0xcf, 0x8b, 0x0a, 0x12, 0xf2, 0x73, 0xca, + 0x3f, 0x46, 0xb8, 0xfb, 0x3a, 0xcc, 0xbb, 0xfe, 0x9e, 0xec, 0xd2, 0x7d, 0x79, 0xfb, 0xee, 0xac, + 0xf8, 0x49, 0xec, 0x63, 0xdf, 0x0d, 0xdd, 0xc7, 0xda, 0x0f, 0xa5, 0x72, 0x67, 0xeb, 0xc9, 0x6e, + 0x8d, 0xfd, 0x7e, 0xf4, 0xf6, 0xdf, 0x03, 0x00, 0x00, 0xff, 0xff, 0xcc, 0x96, 0x1a, 0x1f, 0x3b, + 0x3b, 0x00, 0x00, } diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.proto b/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.proto index f9a75c52c..dc2c1ec09 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.proto +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.proto @@ -158,9 +158,6 @@ message ExampleOrReference { } } -message Examples { -} - message ExamplesOrReferences { repeated NamedExampleOrReference additional_properties = 1; } diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/openapi-3.0.json b/vendor/github.com/googleapis/gnostic/OpenAPIv3/openapi-3.0.json index c15fc2e5b..aad0d1b65 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv3/openapi-3.0.json +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/openapi-3.0.json @@ -102,10 +102,12 @@ "type": "string" }, "url": { - "type": "string" + "type": "string", + "format": "uri" }, "email": { - "type": "string" + "type": "string", + "format": "email" } } }, @@ -692,11 +694,6 @@ } } }, - "examples": { - "type": "object", - "description": "", - "additionalProperties": false - }, "reference": { "type": "object", "description": "A simple object to allow referencing other components in the specification, internally and externally. The Reference Object is defined by JSON Reference and follows the same structure, behavior and rules. For this specification, reference resolution is accomplished as defined by the JSON Reference specification and not by the JSON Schema specification.", diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.md b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.0.md similarity index 100% rename from vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.md rename to vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.0.md diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.1.md b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.1.md new file mode 100644 index 000000000..0a8a83435 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/3.0.1.md @@ -0,0 +1,3380 @@ +# OpenAPI Specification + +#### Version 3.0.1 + +The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD", "SHOULD NOT", "RECOMMENDED", "NOT RECOMMENDED", "MAY", and "OPTIONAL" in this document are to be interpreted as described in [BCP 14](https://tools.ietf.org/html/bcp14) [RFC2119](https://tools.ietf.org/html/rfc2119) [RFC8174](https://tools.ietf.org/html/rfc8174) when, and only when, they appear in all capitals, as shown here. + +This document is licensed under [The Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0.html). + +## Introduction + +The OpenAPI Specification (OAS) defines a standard, language-agnostic interface to RESTful APIs which allows both humans and computers to discover and understand the capabilities of the service without access to source code, documentation, or through network traffic inspection. When properly defined, a consumer can understand and interact with the remote service with a minimal amount of implementation logic. + +An OpenAPI definition can then be used by documentation generation tools to display the API, code generation tools to generate servers and clients in various programming languages, testing tools, and many other use cases. + +## Table of Contents + + +- [Definitions](#definitions) + - [OpenAPI Document](#oasDocument) + - [Path Templating](#pathTemplating) + - [Media Types](#mediaTypes) + - [HTTP Status Codes](#httpCodes) +- [Specification](#specification) + - [Versions](#versions) + - [Format](#format) + - [Document Structure](#documentStructure) + - [Data Types](#dataTypes) + - [Rich Text Formatting](#richText) + - [Relative References In URLs](#relativeReferences) + - [Schema](#schema) + - [OpenAPI Object](#oasObject) + - [Info Object](#infoObject) + - [Contact Object](#contactObject) + - [License Object](#licenseObject) + - [Server Object](#serverObject) + - [Server Variable Object](#serverVariableObject) + - [Components Object](#componentsObject) + - [Paths Object](#pathsObject) + - [Path Item Object](#pathItemObject) + - [Operation Object](#operationObject) + - [External Documentation Object](#externalDocumentationObject) + - [Parameter Object](#parameterObject) + - [Request Body Object](#requestBodyObject) + - [Media Type Object](#mediaTypeObject) + - [Encoding Object](#encodingObject) + - [Responses Object](#responsesObject) + - [Response Object](#responseObject) + - [Callback Object](#callbackObject) + - [Example Object](#exampleObject) + - [Link Object](#linkObject) + - [Header Object](#headerObject) + - [Tag Object](#tagObject) + - [Reference Object](#referenceObject) + - [Schema Object](#schemaObject) + - [Discriminator Object](#discriminatorObject) + - [XML Object](#xmlObject) + - [Security Scheme Object](#securitySchemeObject) + - [OAuth Flows Object](#oauthFlowsObject) + - [OAuth Flow Object](#oauthFlowObject) + - [Security Requirement Object](#securityRequirementObject) + - [Specification Extensions](#specificationExtensions) + - [Security Filtering](#securityFiltering) +- [Appendix A: Revision History](#revisionHistory) + + + + +## Definitions + +##### OpenAPI Document +A document (or set of documents) that defines or describes an API. An OpenAPI definition uses and conforms to the OpenAPI Specification. + +##### Path Templating +Path templating refers to the usage of curly braces ({}) to mark a section of a URL path as replaceable using path parameters. + +##### Media Types +Media type definitions are spread across several resources. +The media type definitions SHOULD be in compliance with [RFC6838](https://tools.ietf.org/html/rfc6838). + +Some examples of possible media type definitions: +``` + text/plain; charset=utf-8 + application/json + application/vnd.github+json + application/vnd.github.v3+json + application/vnd.github.v3.raw+json + application/vnd.github.v3.text+json + application/vnd.github.v3.html+json + application/vnd.github.v3.full+json + application/vnd.github.v3.diff + application/vnd.github.v3.patch +``` +##### HTTP Status Codes +The HTTP Status Codes are used to indicate the status of the executed operation. +The available status codes are defined by [RFC7231](https://tools.ietf.org/html/rfc7231#section-6) and registered status codes are listed in the [IANA Status Code Registry](https://www.iana.org/assignments/http-status-codes/http-status-codes.xhtml). + +## Specification + +### Versions + +The OpenAPI Specification is versioned using [Semantic Versioning 2.0.0](https://semver.org/spec/v2.0.0.html) (semver) and follows the semver specification. + +The `major`.`minor` portion of the semver (for example `3.0`) SHALL designate the OAS feature set. Typically, *`.patch`* versions address errors in this document, not the feature set. Tooling which supports OAS 3.0 SHOULD be compatible with all OAS 3.0.\* versions. The patch version SHOULD NOT be considered by tooling, making no distinction between `3.0.0` and `3.0.1` for example. + +Subsequent minor version releases of the OpenAPI Specification (incrementing the `minor` version number) SHOULD NOT interfere with tooling developed to a lower minor version and same major version. Thus a hypothetical `3.1.0` specification SHOULD be usable with tooling designed for `3.0.0`. + +An OpenAPI document compatible with OAS 3.\*.\* contains a required [`openapi`](#oasVersion) field which designates the semantic version of the OAS that it uses. (OAS 2.0 documents contain a top-level version field named [`swagger`](https://github.com/OAI/OpenAPI-Specification/blob/master/versions/2.0.md#swaggerObject) and value `"2.0"`.) + +### Format + +An OpenAPI document that conforms to the OpenAPI Specification is itself a JSON object, which may be represented either in JSON or YAML format. + +For example, if a field has an array value, the JSON array representation will be used: + +```json +{ + "field": [ 1, 2, 3 ] +} +``` +All field names in the specification are **case sensitive**. + +The schema exposes two types of fields: Fixed fields, which have a declared name, and Patterned fields, which declare a regex pattern for the field name. + +Patterned fields MUST have unique names within the containing object. + +In order to preserve the ability to round-trip between YAML and JSON formats, YAML version [1.2](http://www.yaml.org/spec/1.2/spec.html) is RECOMMENDED along with some additional constraints: + +- Tags MUST be limited to those allowed by the [JSON Schema ruleset](http://www.yaml.org/spec/1.2/spec.html#id2803231). +- Keys used in YAML maps MUST be limited to a scalar string, as defined by the [YAML Failsafe schema ruleset](http://yaml.org/spec/1.2/spec.html#id2802346). + +**Note:** While APIs may be defined by OpenAPI documents in either YAML or JSON format, the API request and response bodies and other content are not required to be JSON or YAML. + +### Document Structure + +An OpenAPI document MAY be made up of a single document or be divided into multiple, connected parts at the discretion of the user. In the latter case, `$ref` fields MUST be used in the specification to reference those parts as follows from the [JSON Schema](http://json-schema.org) definitions. + +It is RECOMMENDED that the root OpenAPI document be named: `openapi.json` or `openapi.yaml`. + +### Data Types + +Primitive data types in the OAS are based on the types supported by the [JSON Schema Specification Wright Draft 00](https://tools.ietf.org/html/draft-wright-json-schema-00#section-4.2). +Note that `integer` as a type is also supported and is defined as a JSON number without a fraction or exponent part. +`null` is not supported as a type (see [`nullable`](#schemaNullable) for an alternative solution). +Models are defined using the [Schema Object](#schemaObject), which is an extended subset of JSON Schema Specification Wright Draft 00. + +Primitives have an optional modifier property: `format`. +OAS uses several known formats to define in fine detail the data type being used. +However, to support documentation needs, the `format` property is an open `string`-valued property, and can have any value. +Formats such as `"email"`, `"uuid"`, and so on, MAY be used even though undefined by this specification. +Types that are not accompanied by a `format` property follow the type definition in the JSON Schema. Tools that do not recognize a specific `format` MAY default back to the `type` alone, as if the `format` is not specified. + +The formats defined by the OAS are: + +Common Name | [`type`](#dataTypes) | [`format`](#dataTypeFormat) | Comments +----------- | ------ | -------- | -------- +integer | `integer` | `int32` | signed 32 bits +long | `integer` | `int64` | signed 64 bits +float | `number` | `float` | | +double | `number` | `double` | | +string | `string` | | | +byte | `string` | `byte` | base64 encoded characters +binary | `string` | `binary` | any sequence of octets +boolean | `boolean` | | | +date | `string` | `date` | As defined by `full-date` - [RFC3339](https://xml2rfc.ietf.org/public/rfc/html/rfc3339.html#anchor14) +dateTime | `string` | `date-time` | As defined by `date-time` - [RFC3339](https://xml2rfc.ietf.org/public/rfc/html/rfc3339.html#anchor14) +password | `string` | `password` | A hint to UIs to obscure input. + +### Rich Text Formatting +Throughout the specification `description` fields are noted as supporting CommonMark markdown formatting. +Where OpenAPI tooling renders rich text it MUST support, at a minimum, markdown syntax as described by [CommonMark 0.27](http://spec.commonmark.org/0.27/). Tooling MAY choose to ignore some CommonMark features to address security concerns. + +### Relative References in URLs + +Unless specified otherwise, all properties that are URLs MAY be relative references as defined by [RFC3986](https://tools.ietf.org/html/rfc3986#section-4.2). +Relative references are resolved using the URLs defined in the [`Server Object`](#serverObject) as a Base URI. + +Relative references used in `$ref` are processed as per [JSON Reference](https://tools.ietf.org/html/draft-pbryan-zyp-json-ref-03), using the URL of the current document as the base URI. See also the [Reference Object](#referenceObject). + +### Schema + +In the following description, if a field is not explicitly **REQUIRED** or described with a MUST or SHALL, it can be considered OPTIONAL. + +#### OpenAPI Object + +This is the root document object of the [OpenAPI document](#oasDocument). + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +openapi | `string` | **REQUIRED**. This string MUST be the [semantic version number](https://semver.org/spec/v2.0.0.html) of the [OpenAPI Specification version](#versions) that the OpenAPI document uses. The `openapi` field SHOULD be used by tooling specifications and clients to interpret the OpenAPI document. This is *not* related to the API [`info.version`](#infoVersion) string. +info | [Info Object](#infoObject) | **REQUIRED**. Provides metadata about the API. The metadata MAY be used by tooling as required. +servers | [[Server Object](#serverObject)] | An array of Server Objects, which provide connectivity information to a target server. If the `servers` property is not provided, or is an empty array, the default value would be a [Server Object](#serverObject) with a [url](#serverUrl) value of `/`. +paths | [Paths Object](#pathsObject) | **REQUIRED**. The available paths and operations for the API. +components | [Components Object](#componentsObject) | An element to hold various schemas for the specification. +security | [[Security Requirement Object](#securityRequirementObject)] | A declaration of which security mechanisms can be used across the API. The list of values includes alternative security requirement objects that can be used. Only one of the security requirement objects need to be satisfied to authorize a request. Individual operations can override this definition. +tags | [[Tag Object](#tagObject)] | A list of tags used by the specification with additional metadata. The order of the tags can be used to reflect on their order by the parsing tools. Not all tags that are used by the [Operation Object](#operationObject) must be declared. The tags that are not declared MAY be organized randomly or based on the tools' logic. Each tag name in the list MUST be unique. +externalDocs | [External Documentation Object](#externalDocumentationObject) | Additional external documentation. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +#### Info Object + +The object provides metadata about the API. +The metadata MAY be used by the clients if needed, and MAY be presented in editing or documentation generation tools for convenience. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +title | `string` | **REQUIRED**. The title of the application. +description | `string` | A short description of the application. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +termsOfService | `string` | A URL to the Terms of Service for the API. MUST be in the format of a URL. +contact | [Contact Object](#contactObject) | The contact information for the exposed API. +license | [License Object](#licenseObject) | The license information for the exposed API. +version | `string` | **REQUIRED**. The version of the OpenAPI document (which is distinct from the [OpenAPI Specification version](#oasVersion) or the API implementation version). + + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Info Object Example: + +```json +{ + "title": "Sample Pet Store App", + "description": "This is a sample server for a pet store.", + "termsOfService": "http://example.com/terms/", + "contact": { + "name": "API Support", + "url": "http://www.example.com/support", + "email": "support@example.com" + }, + "license": { + "name": "Apache 2.0", + "url": "https://www.apache.org/licenses/LICENSE-2.0.html" + }, + "version": "1.0.1" +} +``` + +```yaml +title: Sample Pet Store App +description: This is a sample server for a pet store. +termsOfService: http://example.com/terms/ +contact: + name: API Support + url: http://www.example.com/support + email: support@example.com +license: + name: Apache 2.0 + url: https://www.apache.org/licenses/LICENSE-2.0.html +version: 1.0.1 +``` + +#### Contact Object + +Contact information for the exposed API. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +name | `string` | The identifying name of the contact person/organization. +url | `string` | The URL pointing to the contact information. MUST be in the format of a URL. +email | `string` | The email address of the contact person/organization. MUST be in the format of an email address. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Contact Object Example: + +```json +{ + "name": "API Support", + "url": "http://www.example.com/support", + "email": "support@example.com" +} +``` + +```yaml +name: API Support +url: http://www.example.com/support +email: support@example.com +``` + +#### License Object + +License information for the exposed API. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +name | `string` | **REQUIRED**. The license name used for the API. +url | `string` | A URL to the license used for the API. MUST be in the format of a URL. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### License Object Example: + +```json +{ + "name": "Apache 2.0", + "url": "https://www.apache.org/licenses/LICENSE-2.0.html" +} +``` + +```yaml +name: Apache 2.0 +url: https://www.apache.org/licenses/LICENSE-2.0.html +``` + +#### Server Object + +An object representing a Server. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +url | `string` | **REQUIRED**. A URL to the target host. This URL supports Server Variables and MAY be relative, to indicate that the host location is relative to the location where the OpenAPI document is being served. Variable substitutions will be made when a variable is named in `{`brackets`}`. +description | `string` | An optional string describing the host designated by the URL. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +variables | Map[`string`, [Server Variable Object](#serverVariableObject)] | A map between a variable name and its value. The value is used for substitution in the server's URL template. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Server Object Example + +A single server would be described as: + +```json +{ + "url": "https://development.gigantic-server.com/v1", + "description": "Development server" +} +``` + +```yaml +url: https://development.gigantic-server.com/v1 +description: Development server +``` + +The following shows how multiple servers can be described, for example, at the OpenAPI Object's [`servers`](#oasServers): + +```json +{ + "servers": [ + { + "url": "https://development.gigantic-server.com/v1", + "description": "Development server" + }, + { + "url": "https://staging.gigantic-server.com/v1", + "description": "Staging server" + }, + { + "url": "https://api.gigantic-server.com/v1", + "description": "Production server" + } + ] +} +``` + +```yaml +servers: +- url: https://development.gigantic-server.com/v1 + description: Development server +- url: https://staging.gigantic-server.com/v1 + description: Staging server +- url: https://api.gigantic-server.com/v1 + description: Production server +``` + +The following shows how variables can be used for a server configuration: + +```json +{ + "servers": [ + { + "url": "https://{username}.gigantic-server.com:{port}/{basePath}", + "description": "The production API server", + "variables": { + "username": { + "default": "demo", + "description": "this value is assigned by the service provider, in this example `gigantic-server.com`" + }, + "port": { + "enum": [ + "8443", + "443" + ], + "default": "8443" + }, + "basePath": { + "default": "v2" + } + } + } + ] +} +``` + +```yaml +servers: +- url: https://{username}.gigantic-server.com:{port}/{basePath} + description: The production API server + variables: + username: + # note! no enum here means it is an open value + default: demo + description: this value is assigned by the service provider, in this example `gigantic-server.com` + port: + enum: + - '8443' + - '443' + default: '8443' + basePath: + # open meaning there is the opportunity to use special base paths as assigned by the provider, default is `v2` + default: v2 +``` + + +#### Server Variable Object + +An object representing a Server Variable for server URL template substitution. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +enum | [`string`] | An enumeration of string values to be used if the substitution options are from a limited set. +default | `string` | **REQUIRED**. The default value to use for substitution, and to send, if an alternate value is _not_ supplied. Unlike the [Schema Object's](#schemaObject) `default`, this value MUST be provided by the consumer. +description | `string` | An optional description for the server variable. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +#### Components Object + +Holds a set of reusable objects for different aspects of the OAS. +All objects defined within the components object will have no effect on the API unless they are explicitly referenced from properties outside the components object. + + +##### Fixed Fields + +Field Name | Type | Description +---|:---|--- + schemas | Map[`string`, [Schema Object](#schemaObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Schema Objects](#schemaObject). + responses | Map[`string`, [Response Object](#responseObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Response Objects](#responseObject). + parameters | Map[`string`, [Parameter Object](#parameterObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Parameter Objects](#parameterObject). + examples | Map[`string`, [Example Object](#exampleObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Example Objects](#exampleObject). + requestBodies | Map[`string`, [Request Body Object](#requestBodyObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Request Body Objects](#requestBodyObject). + headers | Map[`string`, [Header Object](#headerObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Header Objects](#headerObject). + securitySchemes| Map[`string`, [Security Scheme Object](#securitySchemeObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Security Scheme Objects](#securitySchemeObject). + links | Map[`string`, [Link Object](#linkObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Link Objects](#linkObject). + callbacks | Map[`string`, [Callback Object](#callbackObject) \| [Reference Object](#referenceObject)] | An object to hold reusable [Callback Objects](#callbackObject). + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +All the fixed fields declared above are objects that MUST use keys that match the regular expression: `^[a-zA-Z0-9\.\-_]+$`. + +Field Name Examples: + +``` +User +User_1 +User_Name +user-name +my.org.User +``` + +##### Components Object Example + +```json +"components": { + "schemas": { + "Category": { + "type": "object", + "properties": { + "id": { + "type": "integer", + "format": "int64" + }, + "name": { + "type": "string" + } + } + }, + "Tag": { + "type": "object", + "properties": { + "id": { + "type": "integer", + "format": "int64" + }, + "name": { + "type": "string" + } + } + } + }, + "parameters": { + "skipParam": { + "name": "skip", + "in": "query", + "description": "number of items to skip", + "required": true, + "schema": { + "type": "integer", + "format": "int32" + } + }, + "limitParam": { + "name": "limit", + "in": "query", + "description": "max records to return", + "required": true, + "schema" : { + "type": "integer", + "format": "int32" + } + } + }, + "responses": { + "NotFound": { + "description": "Entity not found." + }, + "IllegalInput": { + "description": "Illegal input for operation." + }, + "GeneralError": { + "description": "General Error", + "content": { + "application/json": { + "schema": { + "$ref": "#/components/schemas/GeneralError" + } + } + } + } + }, + "securitySchemes": { + "api_key": { + "type": "apiKey", + "name": "api_key", + "in": "header" + }, + "petstore_auth": { + "type": "oauth2", + "flows": { + "implicit": { + "authorizationUrl": "http://example.org/api/oauth/dialog", + "scopes": { + "write:pets": "modify pets in your account", + "read:pets": "read your pets" + } + } + } + } + } +} +``` + +```yaml +components: + schemas: + Category: + type: object + properties: + id: + type: integer + format: int64 + name: + type: string + Tag: + type: object + properties: + id: + type: integer + format: int64 + name: + type: string + parameters: + skipParam: + name: skip + in: query + description: number of items to skip + required: true + schema: + type: integer + format: int32 + limitParam: + name: limit + in: query + description: max records to return + required: true + schema: + type: integer + format: int32 + responses: + NotFound: + description: Entity not found. + IllegalInput: + description: Illegal input for operation. + GeneralError: + description: General Error + content: + application/json: + schema: + $ref: '#/components/schemas/GeneralError' + securitySchemes: + api_key: + type: apiKey + name: api_key + in: header + petstore_auth: + type: oauth2 + flows: + implicit: + authorizationUrl: http://example.org/api/oauth/dialog + scopes: + write:pets: modify pets in your account + read:pets: read your pets +``` + + +#### Paths Object + +Holds the relative paths to the individual endpoints and their operations. +The path is appended to the URL from the [`Server Object`](#serverObject) in order to construct the full URL. The Paths MAY be empty, due to [ACL constraints](#securityFiltering). + +##### Patterned Fields + +Field Pattern | Type | Description +---|:---:|--- +/{path} | [Path Item Object](#pathItemObject) | A relative path to an individual endpoint. The field name MUST begin with a slash. The path is **appended** (no relative URL resolution) to the expanded URL from the [`Server Object`](#serverObject)'s `url` field in order to construct the full URL. [Path templating](#pathTemplating) is allowed. When matching URLs, concrete (non-templated) paths would be matched before their templated counterparts. Templated paths with the same hierarchy but different templated names MUST NOT exist as they are identical. In case of ambiguous matching, it's up to the tooling to decide which one to use. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Path Templating Matching + +Assuming the following paths, the concrete definition, `/pets/mine`, will be matched first if used: + +``` + /pets/{petId} + /pets/mine +``` + +The following paths are considered identical and invalid: + +``` + /pets/{petId} + /pets/{name} +``` + +The following may lead to ambiguous resolution: + +``` + /{entity}/me + /books/{id} +``` + +##### Paths Object Example + +```json +{ + "/pets": { + "get": { + "description": "Returns all pets from the system that the user has access to", + "responses": { + "200": { + "description": "A list of pets.", + "content": { + "application/json": { + "schema": { + "type": "array", + "items": { + "$ref": "#/components/schemas/pet" + } + } + } + } + } + } + } + } +} +``` + +```yaml +/pets: + get: + description: Returns all pets from the system that the user has access to + responses: + '200': + description: A list of pets. + content: + application/json: + schema: + type: array + items: + $ref: '#/components/schemas/pet' +``` + +#### Path Item Object + +Describes the operations available on a single path. +A Path Item MAY be empty, due to [ACL constraints](#securityFiltering). +The path itself is still exposed to the documentation viewer but they will not know which operations and parameters are available. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +$ref | `string` | Allows for an external definition of this path item. The referenced structure MUST be in the format of a [Path Item Object](#pathItemObject). If there are conflicts between the referenced definition and this Path Item's definition, the behavior is *undefined*. +summary| `string` | An optional, string summary, intended to apply to all operations in this path. +description | `string` | An optional, string description, intended to apply to all operations in this path. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +get | [Operation Object](#operationObject) | A definition of a GET operation on this path. +put | [Operation Object](#operationObject) | A definition of a PUT operation on this path. +post | [Operation Object](#operationObject) | A definition of a POST operation on this path. +delete | [Operation Object](#operationObject) | A definition of a DELETE operation on this path. +options | [Operation Object](#operationObject) | A definition of a OPTIONS operation on this path. +head | [Operation Object](#operationObject) | A definition of a HEAD operation on this path. +patch | [Operation Object](#operationObject) | A definition of a PATCH operation on this path. +trace | [Operation Object](#operationObject) | A definition of a TRACE operation on this path. +servers | [[Server Object](#serverObject)] | An alternative `server` array to service all operations in this path. +parameters | [[Parameter Object](#parameterObject) \| [Reference Object](#referenceObject)] | A list of parameters that are applicable for all the operations described under this path. These parameters can be overridden at the operation level, but cannot be removed there. The list MUST NOT include duplicated parameters. A unique parameter is defined by a combination of a [name](#parameterName) and [location](#parameterIn). The list can use the [Reference Object](#referenceObject) to link to parameters that are defined at the [OpenAPI Object's components/parameters](#componentsParameters). + + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Path Item Object Example + +```json +{ + "get": { + "description": "Returns pets based on ID", + "summary": "Find pets by ID", + "operationId": "getPetsById", + "responses": { + "200": { + "description": "pet response", + "content": { + "*/*": { + "schema": { + "type": "array", + "items": { + "$ref": "#/components/schemas/Pet" + } + } + } + } + }, + "default": { + "description": "error payload", + "content": { + "text/html": { + "schema": { + "$ref": "#/components/schemas/ErrorModel" + } + } + } + } + } + }, + "parameters": [ + { + "name": "id", + "in": "path", + "description": "ID of pet to use", + "required": true, + "schema": { + "type": "array", + "items": { + "type": "string" + } + }, + "style": "simple" + } + ] +} +``` + +```yaml +get: + description: Returns pets based on ID + summary: Find pets by ID + operationId: getPetsById + responses: + '200': + description: pet response + content: + '*/*' : + schema: + type: array + items: + $ref: '#/components/schemas/Pet' + default: + description: error payload + content: + 'text/html': + schema: + $ref: '#/components/schemas/ErrorModel' +parameters: +- name: id + in: path + description: ID of pet to use + required: true + schema: + type: array + style: simple + items: + type: string +``` + +#### Operation Object + +Describes a single API operation on a path. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +tags | [`string`] | A list of tags for API documentation control. Tags can be used for logical grouping of operations by resources or any other qualifier. +summary | `string` | A short summary of what the operation does. +description | `string` | A verbose explanation of the operation behavior. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +externalDocs | [External Documentation Object](#externalDocumentationObject) | Additional external documentation for this operation. +operationId | `string` | Unique string used to identify the operation. The id MUST be unique among all operations described in the API. Tools and libraries MAY use the operationId to uniquely identify an operation, therefore, it is RECOMMENDED to follow common programming naming conventions. +parameters | [[Parameter Object](#parameterObject) \| [Reference Object](#referenceObject)] | A list of parameters that are applicable for this operation. If a parameter is already defined at the [Path Item](#pathItemParameters), the new definition will override it but can never remove it. The list MUST NOT include duplicated parameters. A unique parameter is defined by a combination of a [name](#parameterName) and [location](#parameterIn). The list can use the [Reference Object](#referenceObject) to link to parameters that are defined at the [OpenAPI Object's components/parameters](#componentsParameters). +requestBody | [Request Body Object](#requestBodyObject) \| [Reference Object](#referenceObject) | The request body applicable for this operation. The `requestBody` is only supported in HTTP methods where the HTTP 1.1 specification [RFC7231](https://tools.ietf.org/html/rfc7231#section-4.3.1) has explicitly defined semantics for request bodies. In other cases where the HTTP spec is vague, `requestBody` SHALL be ignored by consumers. +responses | [Responses Object](#responsesObject) | **REQUIRED**. The list of possible responses as they are returned from executing this operation. +callbacks | Map[`string`, [Callback Object](#callbackObject) \| [Reference Object](#referenceObject)] | A map of possible out-of band callbacks related to the parent operation. The key is a unique identifier for the Callback Object. Each value in the map is a [Callback Object](#callbackObject) that describes a request that may be initiated by the API provider and the expected responses. The key value used to identify the callback object is an expression, evaluated at runtime, that identifies a URL to use for the callback operation. +deprecated | `boolean` | Declares this operation to be deprecated. Consumers SHOULD refrain from usage of the declared operation. Default value is `false`. +security | [[Security Requirement Object](#securityRequirementObject)] | A declaration of which security mechanisms can be used for this operation. The list of values includes alternative security requirement objects that can be used. Only one of the security requirement objects need to be satisfied to authorize a request. This definition overrides any declared top-level [`security`](#oasSecurity). To remove a top-level security declaration, an empty array can be used. +servers | [[Server Object](#serverObject)] | An alternative `server` array to service this operation. If an alternative `server` object is specified at the Path Item Object or Root level, it will be overridden by this value. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Operation Object Example + +```json +{ + "tags": [ + "pet" + ], + "summary": "Updates a pet in the store with form data", + "operationId": "updatePetWithForm", + "parameters": [ + { + "name": "petId", + "in": "path", + "description": "ID of pet that needs to be updated", + "required": true, + "schema": { + "type": "string" + } + } + ], + "requestBody": { + "content": { + "application/x-www-form-urlencoded": { + "schema": { + "type": "object", + "properties": { + "name": { + "description": "Updated name of the pet", + "type": "string" + }, + "status": { + "description": "Updated status of the pet", + "type": "string" + } + }, + "required": ["status"] + } + } + } + }, + "responses": { + "200": { + "description": "Pet updated.", + "content": { + "application/json": {}, + "application/xml": {} + } + }, + "405": { + "description": "Invalid input", + "content": { + "application/json": {}, + "application/xml": {} + } + } + }, + "security": [ + { + "petstore_auth": [ + "write:pets", + "read:pets" + ] + } + ] +} +``` + +```yaml +tags: +- pet +summary: Updates a pet in the store with form data +operationId: updatePetWithForm +parameters: +- name: petId + in: path + description: ID of pet that needs to be updated + required: true + schema: + type: string +requestBody: + content: + 'application/x-www-form-urlencoded': + schema: + properties: + name: + description: Updated name of the pet + type: string + status: + description: Updated status of the pet + type: string + required: + - status +responses: + '200': + description: Pet updated. + content: + 'application/json': {} + 'application/xml': {} + '405': + description: Invalid input + content: + 'application/json': {} + 'application/xml': {} +security: +- petstore_auth: + - write:pets + - read:pets +``` + + +#### External Documentation Object + +Allows referencing an external resource for extended documentation. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +description | `string` | A short description of the target documentation. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +url | `string` | **REQUIRED**. The URL for the target documentation. Value MUST be in the format of a URL. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### External Documentation Object Example + +```json +{ + "description": "Find more info here", + "url": "https://example.com" +} +``` + +```yaml +description: Find more info here +url: https://example.com +``` + +#### Parameter Object + +Describes a single operation parameter. + +A unique parameter is defined by a combination of a [name](#parameterName) and [location](#parameterIn). + +##### Parameter Locations +There are four possible parameter locations specified by the `in` field: +* path - Used together with [Path Templating](#pathTemplating), where the parameter value is actually part of the operation's URL. This does not include the host or base path of the API. For example, in `/items/{itemId}`, the path parameter is `itemId`. +* query - Parameters that are appended to the URL. For example, in `/items?id=###`, the query parameter is `id`. +* header - Custom headers that are expected as part of the request. Note that [RFC7230](https://tools.ietf.org/html/rfc7230#page-22) states header names are case insensitive. +* cookie - Used to pass a specific cookie value to the API. + + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +name | `string` | **REQUIRED**. The name of the parameter. Parameter names are *case sensitive*.
  • If [`in`](#parameterIn) is `"path"`, the `name` field MUST correspond to the associated path segment from the [path](#pathsPath) field in the [Paths Object](#pathsObject). See [Path Templating](#pathTemplating) for further information.
  • If [`in`](#parameterIn) is `"header"` and the `name` field is `"Accept"`, `"Content-Type"` or `"Authorization"`, the parameter definition SHALL be ignored.
  • For all other cases, the `name` corresponds to the parameter name used by the [`in`](#parameterIn) property.
+in | `string` | **REQUIRED**. The location of the parameter. Possible values are "query", "header", "path" or "cookie". +description | `string` | A brief description of the parameter. This could contain examples of use. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +required | `boolean` | Determines whether this parameter is mandatory. If the [parameter location](#parameterIn) is "path", this property is **REQUIRED** and its value MUST be `true`. Otherwise, the property MAY be included and its default value is `false`. + deprecated | `boolean` | Specifies that a parameter is deprecated and SHOULD be transitioned out of usage. + allowEmptyValue | `boolean` | Sets the ability to pass empty-valued parameters. This is valid only for `query` parameters and allows sending a parameter with an empty value. Default value is `false`. If [`style`](#parameterStyle) is used, and if behavior is `n/a` (cannot be serialized), the value of `allowEmptyValue` SHALL be ignored. + +The rules for serialization of the parameter are specified in one of two ways. +For simpler scenarios, a [`schema`](#parameterSchema) and [`style`](#parameterStyle) can describe the structure and syntax of the parameter. + +Field Name | Type | Description +---|:---:|--- +style | `string` | Describes how the parameter value will be serialized depending on the type of the parameter value. Default values (based on value of `in`): for `query` - `form`; for `path` - `simple`; for `header` - `simple`; for `cookie` - `form`. +explode | `boolean` | When this is true, parameter values of type `array` or `object` generate separate parameters for each value of the array or key-value pair of the map. For other types of parameters this property has no effect. When [`style`](#parameterStyle) is `form`, the default value is `true`. For all other styles, the default value is `false`. +allowReserved | `boolean` | Determines whether the parameter value SHOULD allow reserved characters, as defined by [RFC3986](https://tools.ietf.org/html/rfc3986#section-2.2) `:/?#[]@!$&'()*+,;=` to be included without percent-encoding. This property only applies to parameters with an `in` value of `query`. The default value is `false`. +schema | [Schema Object](#schemaObject) \| [Reference Object](#referenceObject) | The schema defining the type used for the parameter. +example | Any | Example of the media type. The example SHOULD match the specified schema and encoding properties if present. The `example` field is mutually exclusive of the `examples` field. Furthermore, if referencing a `schema` which contains an example, the `example` value SHALL _override_ the example provided by the schema. To represent examples of media types that cannot naturally be represented in JSON or YAML, a string value can contain the example with escaping where necessary. +examples | Map[ `string`, [Example Object](#exampleObject) \| [Reference Object](#referenceObject)] | Examples of the media type. Each example SHOULD contain a value in the correct format as specified in the parameter encoding. The `examples` field is mutually exclusive of the `example` field. Furthermore, if referencing a `schema` which contains an example, the `examples` value SHALL _override_ the example provided by the schema. + +For more complex scenarios, the [`content`](#parameterContent) property can define the media type and schema of the parameter. +A parameter MUST contain either a `schema` property, or a `content` property, but not both. +When `example` or `examples` are provided in conjunction with the `schema` object, the example MUST follow the prescribed serialization strategy for the parameter. + + +Field Name | Type | Description +---|:---:|--- +content | Map[`string`, [Media Type Object](#mediaTypeObject)] | A map containing the representations for the parameter. The key is the media type and the value describes it. The map MUST only contain one entry. + +##### Style Values + +In order to support common ways of serializing simple parameters, a set of `style` values are defined. + +`style` | [`type`](#dataTypes) | `in` | Comments +----------- | ------ | -------- | -------- +matrix | `primitive`, `array`, `object` | `path` | Path-style parameters defined by [RFC6570](https://tools.ietf.org/html/rfc6570#section-3.2.7) +label | `primitive`, `array`, `object` | `path` | Label style parameters defined by [RFC6570](https://tools.ietf.org/html/rfc6570#section-3.2.5) +form | `primitive`, `array`, `object` | `query`, `cookie` | Form style parameters defined by [RFC6570](https://tools.ietf.org/html/rfc6570#section-3.2.8). This option replaces `collectionFormat` with a `csv` (when `explode` is false) or `multi` (when `explode` is true) value from OpenAPI 2.0. +simple | `array` | `path`, `header` | Simple style parameters defined by [RFC6570](https://tools.ietf.org/html/rfc6570#section-3.2.2). This option replaces `collectionFormat` with a `csv` value from OpenAPI 2.0. +spaceDelimited | `array` | `query` | Space separated array values. This option replaces `collectionFormat` equal to `ssv` from OpenAPI 2.0. +pipeDelimited | `array` | `query` | Pipe separated array values. This option replaces `collectionFormat` equal to `pipes` from OpenAPI 2.0. +deepObject | `object` | `query` | Provides a simple way of rendering nested objects using form parameters. + + +##### Style Examples + +Assume a parameter named `color` has one of the following values: + +``` + string -> "blue" + array -> ["blue","black","brown"] + object -> { "R": 100, "G": 200, "B": 150 } +``` +The following table shows examples of rendering differences for each value. + +[`style`](#dataTypeFormat) | `explode` | `empty` | `string` | `array` | `object` +----------- | ------ | -------- | -------- | --------|------- +matrix | false | ;color | ;color=blue | ;color=blue,black,brown | ;color=R,100,G,200,B,150 +matrix | true | ;color | ;color=blue | ;color=blue;color=black;color=brown | ;R=100;G=200;B=150 +label | false | . | .blue | .blue.black.brown | .R.100.G.200.B.150 +label | true | . | .blue | .blue.black.brown | .R=100.G=200.B=150 +form | false | color= | color=blue | color=blue,black,brown | color=R,100,G,200,B,150 +form | true | color= | color=blue | color=blue&color=black&color=brown | R=100&G=200&B=150 +simple | false | n/a | blue | blue,black,brown | R,100,G,200,B,150 +simple | true | n/a | blue | blue,black,brown | R=100,G=200,B=150 +spaceDelimited | false | n/a | n/a | blue%20black%20brown | R%20100%20G%20200%20B%20150 +pipeDelimited | false | n/a | n/a | blue\|black\|brown | R\|100\|G\|200|G\|150 +deepObject | true | n/a | n/a | n/a | color[R]=100&color[G]=200&color[B]=150 + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Parameter Object Examples + +A header parameter with an array of 64 bit integer numbers: + +```json +{ + "name": "token", + "in": "header", + "description": "token to be passed as a header", + "required": true, + "schema": { + "type": "array", + "items": { + "type": "integer", + "format": "int64" + } + }, + "style": "simple" +} +``` + +```yaml +name: token +in: header +description: token to be passed as a header +required: true +schema: + type: array + items: + type: integer + format: int64 +style: simple +``` + +A path parameter of a string value: +```json +{ + "name": "username", + "in": "path", + "description": "username to fetch", + "required": true, + "schema": { + "type": "string" + } +} +``` + +```yaml +name: username +in: path +description: username to fetch +required: true +schema: + type: string +``` + +An optional query parameter of a string value, allowing multiple values by repeating the query parameter: +```json +{ + "name": "id", + "in": "query", + "description": "ID of the object to fetch", + "required": false, + "schema": { + "type": "array", + "items": { + "type": "string" + } + }, + "style": "form", + "explode": true +} +``` + +```yaml +name: id +in: query +description: ID of the object to fetch +required: false +schema: + type: array + items: + type: string +style: form +explode: true +``` + +A free-form query parameter, allowing undefined parameters of a specific type: +```json +{ + "in": "query", + "name": "freeForm", + "schema": { + "type": "object", + "additionalProperties": { + "type": "integer" + }, + }, + "style": "form" +} +``` + +```yaml +in: query +name: freeForm +schema: + type: object + additionalProperties: + type: integer +style: form +``` + +A complex parameter using `content` to define serialization: + +```json +{ + "in": "query", + "name": "coordinates", + "content": { + "application/json": { + "schema": { + "type": "object", + "required": [ + "lat", + "long" + ], + "properties": { + "lat": { + "type": "number" + }, + "long": { + "type": "number" + } + } + } + } + } +} +``` + +```yaml +in: query +name: coordinates +content: + application/json: + schema: + type: object + required: + - lat + - long + properties: + lat: + type: number + long: + type: number +``` + +#### Request Body Object + +Describes a single request body. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +description | `string` | A brief description of the request body. This could contain examples of use. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +content | Map[`string`, [Media Type Object](#mediaTypeObject)] | **REQUIRED**. The content of the request body. The key is a media type or [media type range](https://tools.ietf.org/html/rfc7231#appendix-D) and the value describes it. For requests that match multiple keys, only the most specific key is applicable. e.g. text/plain overrides text/* +required | `boolean` | Determines if the request body is required in the request. Defaults to `false`. + + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Request Body Examples + +A request body with a referenced model definition. +```json +{ + "description": "user to add to the system", + "content": { + "application/json": { + "schema": { + "$ref": "#/components/schemas/User" + }, + "examples": { + "user" : { + "summary": "User Example", + "externalValue": "http://foo.bar/examples/user-example.json" + } + } + }, + "application/xml": { + "schema": { + "$ref": "#/components/schemas/User" + }, + "examples": { + "user" : { + "summary": "User example in XML", + "externalValue": "http://foo.bar/examples/user-example.xml" + } + } + }, + "text/plain": { + "examples": { + "user" : { + "summary": "User example in Plain text", + "externalValue": "http://foo.bar/examples/user-example.txt" + } + } + }, + "*/*": { + "examples": { + "user" : { + "summary": "User example in other format", + "externalValue": "http://foo.bar/examples/user-example.whatever" + } + } + } + } +} +``` + +```yaml +description: user to add to the system +content: + 'application/json': + schema: + $ref: '#/components/schemas/User' + examples: + user: + summary: User Example + externalValue: 'http://foo.bar/examples/user-example.json' + 'application/xml': + schema: + $ref: '#/components/schemas/User' + examples: + user: + summary: User Example in XML + externalValue: 'http://foo.bar/examples/user-example.xml' + 'text/plain': + examples: + user: + summary: User example in text plain format + externalValue: 'http://foo.bar/examples/user-example.txt' + '*/*': + examples: + user: + summary: User example in other format + externalValue: 'http://foo.bar/examples/user-example.whatever' +``` + +A body parameter that is an array of string values: +```json +{ + "description": "user to add to the system", + "content": { + "text/plain": { + "schema": { + "type": "array", + "items": { + "type": "string" + } + } + } + } +} +``` + +```yaml +description: user to add to the system +required: true +content: + text/plain: + schema: + type: array + items: + type: string +``` + + +#### Media Type Object +Each Media Type Object provides schema and examples for the media type identified by its key. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +schema | [Schema Object](#schemaObject) \| [Reference Object](#referenceObject) | The schema defining the type used for the request body. +example | Any | Example of the media type. The example object SHOULD be in the correct format as specified by the media type. The `example` field is mutually exclusive of the `examples` field. Furthermore, if referencing a `schema` which contains an example, the `example` value SHALL _override_ the example provided by the schema. +examples | Map[ `string`, [Example Object](#exampleObject) \| [Reference Object](#referenceObject)] | Examples of the media type. Each example object SHOULD match the media type and specified schema if present. The `examples` field is mutually exclusive of the `example` field. Furthermore, if referencing a `schema` which contains an example, the `examples` value SHALL _override_ the example provided by the schema. +encoding | Map[`string`, [Encoding Object](#encodingObject)] | A map between a property name and its encoding information. The key, being the property name, MUST exist in the schema as a property. The encoding object SHALL only apply to `requestBody` objects when the media type is `multipart` or `application/x-www-form-urlencoded`. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Media Type Examples + +```js +{ + "application/json": { + "schema": { + "$ref": "#/components/schemas/Pet" + }, + "examples": { + "cat" : { + "summary": "An example of a cat", + "value": + { + "name": "Fluffy", + "petType": "Cat", + "color": "White", + "gender": "male", + "breed": "Persian" + } + }, + "dog": { + "summary": "An example of a dog with a cat's name", + "value" : { + "name": "Puma", + "petType": "Dog", + "color": "Black", + "gender": "Female", + "breed": "Mixed" + }, + "frog": { + "$ref": "#/components/examples/frog-example" + } + } + } + } +} +``` + +```yaml +application/json: + schema: + $ref: "#/components/schemas/Pet" + examples: + cat: + summary: An example of a cat + value: + name: Fluffy + petType: Cat + color: White + gender: male + breed: Persian + dog: + summary: An example of a dog with a cat's name + value: + name: Puma + petType: Dog + color: Black + gender: Female + breed: Mixed + frog: + $ref: "#/components/examples/frog-example" +``` + +##### Considerations for File Uploads + +In contrast with the 2.0 specification, `file` input/output content in OpenAPI is described with the same semantics as any other schema type. Specifically: + +```yaml +# content transferred with base64 encoding +schema: + type: string + format: base64 +``` + +```yaml +# content transferred in binary (octet-stream): +schema: + type: string + format: binary +``` + +These examples apply to either input payloads of file uploads or response payloads. + +A `requestBody` for submitting a file in a `POST` operation may look like the following example: + +```yaml +requestBody: + content: + application/octet-stream: + # any media type is accepted, functionally equivalent to `*/*` + schema: + # a binary file of any type + type: string + format: binary +``` + +In addition, specific media types MAY be specified: + +```yaml +# multiple, specific media types may be specified: +requestBody: + content: + # a binary file of type png or jpeg + 'image/jpeg': + schema: + type: string + format: binary + 'image/png': + schema: + type: string + format: binary +``` + +To upload multiple files, a `multipart` media type MUST be used: + +```yaml +requestBody: + content: + multipart/form-data: + schema: + properties: + # The property name 'file' will be used for all files. + file: + type: array + items: + type: string + format: binary + +``` + +##### Support for x-www-form-urlencoded Request Bodies + +To submit content using form url encoding via [RFC1866](https://tools.ietf.org/html/rfc1866), the following +definition may be used: + +```yaml +requestBody: + content: + application/x-www-form-urlencoded: + schema: + type: object + properties: + id: + type: string + format: uuid + address: + # complex types are stringified to support RFC 1866 + type: object + properties: {} +``` + +In this example, the contents in the `requestBody` MUST be stringified per [RFC1866](https://tools.ietf.org/html/rfc1866/) when passed to the server. In addition, the `address` field complex object will be stringified. + +When passing complex objects in the `application/x-www-form-urlencoded` content type, the default serialization strategy of such properties is described in the [`Encoding Object`](#encodingObject)'s [`style`](#encodingStyle) property as `form`. + +##### Special Considerations for `multipart` Content + +It is common to use `multipart/form-data` as a `Content-Type` when transferring request bodies to operations. In contrast to 2.0, a `schema` is REQUIRED to define the input parameters to the operation when using `multipart` content. This supports complex structures as well as supporting mechanisms for multiple file uploads. + +When passing in `multipart` types, boundaries MAY be used to separate sections of the content being transferred — thus, the following default `Content-Type`s are defined for `multipart`: + +* If the property is a primitive, or an array of primitive values, the default Content-Type is `text/plain` +* If the property is complex, or an array of complex values, the default Content-Type is `application/json` +* If the property is a `type: string` with `format: binary` or `format: base64` (aka a file object), the default Content-Type is `application/octet-stream` + + +Examples: + +```yaml +requestBody: + content: + multipart/form-data: + schema: + type: object + properties: + id: + type: string + format: uuid + address: + # default Content-Type for objects is `application/json` + type: object + properties: {} + profileImage: + # default Content-Type for string/binary is `application/octet-stream` + type: string + format: binary + children: + # default Content-Type for arrays is based on the `inner` type (text/plain here) + type: array + items: + type: string + addresses: + # default Content-Type for arrays is based on the `inner` type (object shown, so `application/json` in this example) + type: array + items: + type: '#/components/schemas/Address' +``` + +An `encoding` attribute is introduced to give you control over the serialization of parts of `multipart` request bodies. This attribute is _only_ applicable to `multipart` and `application/x-www-form-urlencoded` request bodies. + +#### Encoding Object + +A single encoding definition applied to a single schema property. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +contentType | `string` | The Content-Type for encoding a specific property. Default value depends on the property type: for `string` with `format` being `binary` – `application/octet-stream`; for other primitive types – `text/plain`; for `object` - `application/json`; for `array` – the default is defined based on the inner type. The value can be a specific media type (e.g. `application/json`), a wildcard media type (e.g. `image/*`), or a comma-separated list of the two types. +headers | Map[`string`, [Header Object](#headerObject) \| [Reference Object](#referenceObject)] | A map allowing additional information to be provided as headers, for example `Content-Disposition`. `Content-Type` is described separately and SHALL be ignored in this section. This property SHALL be ignored if the request body media type is not a `multipart`. +style | `string` | Describes how a specific property value will be serialized depending on its type. See [Parameter Object](#parameterObject) for details on the [`style`](#parameterStyle) property. The behavior follows the same values as `query` parameters, including default values. This property SHALL be ignored if the request body media type is not `application/x-www-form-urlencoded`. +explode | `boolean` | When this is true, property values of type `array` or `object` generate separate parameters for each value of the array, or key-value-pair of the map. For other types of properties this property has no effect. When [`style`](#encodingStyle) is `form`, the default value is `true`. For all other styles, the default value is `false`. This property SHALL be ignored if the request body media type is not `application/x-www-form-urlencoded`. +allowReserved | `boolean` | Determines whether the parameter value SHOULD allow reserved characters, as defined by [RFC3986](https://tools.ietf.org/html/rfc3986#section-2.2) `:/?#[]@!$&'()*+,;=` to be included without percent-encoding. The default value is `false`. This property SHALL be ignored if the request body media type is not `application/x-www-form-urlencoded`. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Encoding Object Example + +```yaml +requestBody: + content: + multipart/mixed: + schema: + type: object + properties: + id: + # default is text/plain + type: string + format: uuid + address: + # default is application/json + type: object + properties: {} + historyMetadata: + # need to declare XML format! + description: metadata in XML format + type: object + properties: {} + profileImage: + # default is application/octet-stream, need to declare an image type only! + type: string + format: binary + encoding: + historyMetadata: + # require XML Content-Type in utf-8 encoding + contentType: application/xml; charset=utf-8 + profileImage: + # only accept png/jpeg + contentType: image/png, image/jpeg + headers: + X-Rate-Limit-Limit: + description: The number of allowed requests in the current period + schema: + type: integer +``` + +#### Responses Object + +A container for the expected responses of an operation. +The container maps a HTTP response code to the expected response. + +The documentation is not necessarily expected to cover all possible HTTP response codes because they may not be known in advance. +However, documentation is expected to cover a successful operation response and any known errors. + +The `default` MAY be used as a default response object for all HTTP codes +that are not covered individually by the specification. + +The `Responses Object` MUST contain at least one response code, and it +SHOULD be the response for a successful operation call. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +default | [Response Object](#responseObject) \| [Reference Object](#referenceObject) | The documentation of responses other than the ones declared for specific HTTP response codes. Use this field to cover undeclared responses. A [Reference Object](#referenceObject) can link to a response that the [OpenAPI Object's components/responses](#componentsResponses) section defines. + +##### Patterned Fields +Field Pattern | Type | Description +---|:---:|--- +[HTTP Status Code](#httpCodes) | [Response Object](#responseObject) \| [Reference Object](#referenceObject) | Any [HTTP status code](#httpCodes) can be used as the property name, but only one property per code, to describe the expected response for that HTTP status code. A [Reference Object](#referenceObject) can link to a response that is defined in the [OpenAPI Object's components/responses](#componentsResponses) section. This field MUST be enclosed in quotation marks (for example, "200") for compatibility between JSON and YAML. To define a range of response codes, this field MAY contain the uppercase wildcard character `X`. For example, `2XX` represents all response codes between `[200-299]`. The following range definitions are allowed: `1XX`, `2XX`, `3XX`, `4XX`, and `5XX`. If a response range is defined using an explicit code, the explicit code definition takes precedence over the range definition for that code. + + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Responses Object Example + +A 200 response for a successful operation and a default response for others (implying an error): + +```json +{ + "200": { + "description": "a pet to be returned", + "content": { + "application/json": { + "schema": { + "$ref": "#/components/schemas/Pet" + } + } + } + }, + "default": { + "description": "Unexpected error", + "content": { + "application/json": { + "schema": { + "$ref": "#/components/schemas/ErrorModel" + } + } + } + } +} +``` + +```yaml +'200': + description: a pet to be returned + content: + application/json: + schema: + $ref: '#/components/schemas/Pet' +default: + description: Unexpected error + content: + application/json: + schema: + $ref: '#/components/schemas/ErrorModel' +``` + +#### Response Object +Describes a single response from an API Operation, including design-time, static +`links` to operations based on the response. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +description | `string` | **REQUIRED**. A short description of the response. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +headers | Map[`string`, [Header Object](#headerObject) \| [Reference Object](#referenceObject)] | Maps a header name to its definition. [RFC7230](https://tools.ietf.org/html/rfc7230#page-22) states header names are case insensitive. If a response header is defined with the name `"Content-Type"`, it SHALL be ignored. +content | Map[`string`, [Media Type Object](#mediaTypeObject)] | A map containing descriptions of potential response payloads. The key is a media type or [media type range](https://tools.ietf.org/html/rfc7231#appendix-D) and the value describes it. For responses that match multiple keys, only the most specific key is applicable. e.g. text/plain overrides text/* +links | Map[`string`, [Link Object](#linkObject) \| [Reference Object](#referenceObject)] | A map of operations links that can be followed from the response. The key of the map is a short name for the link, following the naming constraints of the names for [Component Objects](#componentsObject). + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Response Object Examples + +Response of an array of a complex type: + +```json +{ + "description": "A complex object array response", + "content": { + "application/json": { + "schema": { + "type": "array", + "items": { + "$ref": "#/components/schemas/VeryComplexType" + } + } + } + } +} +``` + +```yaml +description: A complex object array response +content: + application/json: + schema: + type: array + items: + $ref: '#/components/schemas/VeryComplexType' +``` + +Response with a string type: + +```json +{ + "description": "A simple string response", + "content": { + "text/plain": { + "schema": { + "type": "string" + } + } + } + +} +``` + +```yaml +description: A simple string response +content: + text/plain: + schema: + type: string +``` + +Plain text response with headers: + +```json +{ + "description": "A simple string response", + "content": { + "text/plain": { + "schema": { + "type": "string" + } + } + }, + "headers": { + "X-Rate-Limit-Limit": { + "description": "The number of allowed requests in the current period", + "schema": { + "type": "integer" + } + }, + "X-Rate-Limit-Remaining": { + "description": "The number of remaining requests in the current period", + "schema": { + "type": "integer" + } + }, + "X-Rate-Limit-Reset": { + "description": "The number of seconds left in the current period", + "schema": { + "type": "integer" + } + } + } +} +``` + +```yaml +description: A simple string response +content: + text/plain: + schema: + type: string + example: 'whoa!' +headers: + X-Rate-Limit-Limit: + description: The number of allowed requests in the current period + schema: + type: integer + X-Rate-Limit-Remaining: + description: The number of remaining requests in the current period + schema: + type: integer + X-Rate-Limit-Reset: + description: The number of seconds left in the current period + schema: + type: integer +``` + +Response with no return value: + +```json +{ + "description": "object created" +} +``` + +```yaml +description: object created +``` + +#### Callback Object + +A map of possible out-of band callbacks related to the parent operation. +Each value in the map is a [Path Item Object](#pathItemObject) that describes a set of requests that may be initiated by the API provider and the expected responses. +The key value used to identify the callback object is an expression, evaluated at runtime, that identifies a URL to use for the callback operation. + +##### Patterned Fields +Field Pattern | Type | Description +---|:---:|--- +{expression} | [Path Item Object](#pathItemObject) | A Path Item Object used to define a callback request and expected responses. A [complete example](../examples/v3.0/callback-example.yaml) is available. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Key Expression + +The key that identifies the [Path Item Object](#pathItemObject) is a [runtime expression](#runtimeExpression) that can be evaluated in the context of a runtime HTTP request/response to identify the URL to be used for the callback request. +A simple example might be `$request.body#/url`. +However, using a [runtime expression](#runtimeExpression) the complete HTTP message can be accessed. +This includes accessing any part of a body that a JSON Pointer [RFC6901](https://tools.ietf.org/html/rfc6901) can reference. + +For example, given the following HTTP request: + +```http +POST /subscribe/myevent?queryUrl=http://clientdomain.com/stillrunning HTTP/1.1 +Host: example.org +Content-Type: application/json +Content-Length: 187 + +{ + "failedUrl" : "http://clientdomain.com/failed", + "successUrls" : [ + "http://clientdomain.com/fast", + "http://clientdomain.com/medium", + "http://clientdomain.com/slow" + ] +} + +201 Created +Location: http://example.org/subscription/1 +``` + +The following examples show how the various expressions evaluate, assuming the callback operation has a path parameter named `eventType` and a query parameter named `queryUrl`. + +Expression | Value +---|:--- +$url | http://example.org/subscribe/myevent?queryUrl=http://clientdomain.com/stillrunning +$method | POST +$request.path.eventType | myevent +$request.query.queryUrl | http://clientdomain.com/stillrunning +$request.header.content-Type | application/json +$request.body#/failedUrl | http://clientdomain.com/stillrunning +$request.body#/successUrls/2 | http://clientdomain.com/medium +$response.header.Location | http://example.org/subscription/1 + + +##### Callback Object Example + +The following example shows a callback to the URL specified by the `id` and `email` property in the request body. + +```yaml +myWebhook: + 'http://notificationServer.com?transactionId={$request.body#/id}&email={$request.body#/email}': + post: + requestBody: + description: Callback payload + content: + 'application/json': + schema: + $ref: '#/components/schemas/SomePayload' + responses: + '200': + description: webhook successfully processed and no retries will be performed +``` + + +#### Example Object + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +summary | `string` | Short description for the example. +description | `string` | Long description for the example. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +value | Any | Embedded literal example. The `value` field and `externalValue` field are mutually exclusive. To represent examples of media types that cannot naturally represented in JSON or YAML, use a string value to contain the example, escaping where necessary. +externalValue | `string` | A URL that points to the literal example. This provides the capability to reference examples that cannot easily be included in JSON or YAML documents. The `value` field and `externalValue` field are mutually exclusive. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +In all cases, the example value is expected to be compatible with the type schema +of its associated value. Tooling implementations MAY choose to +validate compatibility automatically, and reject the example value(s) if incompatible. + +##### Example Object Example + +```yaml +# in a model +schemas: + properties: + name: + type: string + examples: + name: + $ref: http://example.org/petapi-examples/openapi.json#/components/examples/name-example + +# in a request body: + requestBody: + content: + 'application/json': + schema: + $ref: '#/components/schemas/Address' + examples: + foo: + summary: A foo example + value: {"foo": "bar"} + bar: + summary: A bar example + value: {"bar": "baz"} + 'application/xml': + examples: + xmlExample: + summary: This is an example in XML + externalValue: 'http://example.org/examples/address-example.xml' + 'text/plain': + examples: + textExample: + summary: This is a text example + externalValue: 'http://foo.bar/examples/address-example.txt' + + +# in a parameter + parameters: + - name: 'zipCode' + in: 'query' + schema: + type: 'string' + format: 'zip-code' + examples: + zip-example: + $ref: '#/components/examples/zip-example' + +# in a response + responses: + '200': + description: your car appointment has been booked + content: + application/json: + schema: + $ref: '#/components/schemas/SuccessResponse' + examples: + confirmation-success: + $ref: '#/components/examples/confirmation-success' +``` + + +#### Link Object + +The `Link object` represents a possible design-time link for a response. +The presence of a link does not guarantee the caller's ability to successfully invoke it, rather it provides a known relationship and traversal mechanism between responses and other operations. + +Unlike _dynamic_ links (i.e. links provided **in** the response payload), the OAS linking mechanism does not require link information in the runtime response. + +For computing links, and providing instructions to execute them, a [runtime expression](#runtimeExpression) is used for accessing values in an operation and using them as parameters while invoking the linked operation. + +##### Fixed Fields + +Field Name | Type | Description +---|:---:|--- +operationRef | `string` | A relative or absolute reference to an OAS operation. This field is mutually exclusive of the `operationId` field, and MUST point to an [Operation Object](#operationObject). Relative `operationRef` values MAY be used to locate an existing [Operation Object](#operationObject) in the OpenAPI definition. +operationId | `string` | The name of an _existing_, resolvable OAS operation, as defined with a unique `operationId`. This field is mutually exclusive of the `operationRef` field. +parameters | Map[`string`, Any \| [{expression}](#runtimeExpression)] | A map representing parameters to pass to an operation as specified with `operationId` or identified via `operationRef`. The key is the parameter name to be used, whereas the value can be a constant or an expression to be evaluated and passed to the linked operation. The parameter name can be qualified using the [parameter location](#parameterIn) `[{in}.]{name}` for operations that use the same parameter name in different locations (e.g. path.id). +requestBody | Any \| [{expression}](#runtimeExpression) | A literal value or [{expression}](#runtimeExpression) to use as a request body when calling the target operation. +description | `string` | A description of the link. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +server | [Server Object](#serverObject) | A server object to be used by the target operation. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +A linked operation MUST be identified using either an `operationRef` or `operationId`. +In the case of an `operationId`, it MUST be unique and resolved in the scope of the OAS document. +Because of the potential for name clashes, the `operationRef` syntax is preferred +for specifications with external references. + +##### Examples + +Computing a link from a request operation where the `$request.path.id` is used to pass a request parameter to the linked operation. + +```yaml +paths: + /users/{id}: + parameters: + - name: id + in: path + required: true + description: the user identifier, as userId + schema: + type: string + get: + responses: + '200': + description: the user being returned + content: + application/json: + schema: + type: object + properties: + uuid: # the unique user id + type: string + format: uuid + links: + address: + # the target link operationId + operationId: getUserAddress + parameters: + # get the `id` field from the request path parameter named `id` + userId: $request.path.id + # the path item of the linked operation + /users/{userid}/address: + parameters: + - name: userid + in: path + required: true + description: the user identifier, as userId + schema: + type: string + # linked operation + get: + operationId: getUserAddress + responses: + '200': + description: the user's address +``` + +When a runtime expression fails to evaluate, no parameter value is passed to the target operation. + +Values from the response body can be used to drive a linked operation. + +```yaml +links: + address: + operationId: getUserAddressByUUID + parameters: + # get the `uuid` field from the `uuid` field in the response body + userUuid: $response.body#/uuid +``` + +Clients follow all links at their discretion. +Neither permissions, nor the capability to make a successful call to that link, is guaranteed +solely by the existence of a relationship. + + +##### OperationRef Examples + +As references to `operationId` MAY NOT be possible (the `operationId` is an optional +value), references MAY also be made through a relative `operationRef`: + +```yaml +links: + UserRepositories: + # returns array of '#/components/schemas/repository' + operationRef: '#/paths/~12.0~1repositories~1{username}/get' + parameters: + username: $response.body#/username +``` + +or an absolute `operationRef`: + +```yaml +links: + UserRepositories: + # returns array of '#/components/schemas/repository' + operationRef: 'https://na2.gigantic-server.com/#/paths/~12.0~1repositories~1{username}/get' + parameters: + username: $response.body#/username +``` + +Note that in the use of `operationRef`, the _escaped forward-slash_ is necessary when +using JSON references. + + +##### Runtime Expressions + +Runtime expressions allow defining values based on information that will only be available within the HTTP message in an actual API call. +This mechanism is used by [Link Objects](#linkObject) and [Callback Objects](#callbackObject). + +The runtime expression is defined by the following [ABNF](https://tools.ietf.org/html/rfc5234) syntax + +``` + expression = ( "$url" | "$method" | "$statusCode" | "$request." source | "$response." source ) + source = ( header-reference | query-reference | path-reference | body-reference ) + header-reference = "header." token + query-reference = "query." name + path-reference = "path." name + body-reference = "body" ["#" fragment] + fragment = a JSON Pointer [RFC 6901](https://tools.ietf.org/html/rfc6901) + name = *( char ) + char = as per RFC [7159](https://tools.ietf.org/html/rfc7159#section-7) + token = as per RFC [7230](https://tools.ietf.org/html/rfc7230#section-3.2.6) +``` + +The `name` identifier is case-sensitive, whereas `token` is not. + +The table below provides examples of runtime expressions and examples of their use in a value: + +##### Examples + +Source Location | example expression | notes +---|:---|:---| +HTTP Method | `$method` | The allowable values for the `$method` will be those for the HTTP operation. +Requested media type | `$request.header.accept` | +Request parameter | `$request.path.id` | Request parameters MUST be declared in the `parameters` section of the parent operation or they cannot be evaluated. This includes request headers. +Request body property | `$request.body#/user/uuid` | In operations which accept payloads, references may be made to portions of the `requestBody` or the entire body. +Request URL | `$url` | +Response value | `$response.body#/status` | In operations which return payloads, references may be made to portions of the response body or the entire body. +Response header | `$response.header.Server` | Single header values only are available + +Runtime expressions preserve the type of the referenced value. +Expressions can be embedded into string values by surrounding the expression with `{}` curly braces. + +#### Header Object + +The Header Object follows the structure of the [Parameter Object](#parameterObject) with the following changes: + +1. `name` MUST NOT be specified, it is given in the corresponding `headers` map. +1. `in` MUST NOT be specified, it is implicitly in `header`. +1. All traits that are affected by the location MUST be applicable to a location of `header` (for example, [`style`](#parameterStyle)). + +##### Header Object Example + +A simple header of type `integer`: + +```json +{ + "description": "The number of allowed requests in the current period", + "schema": { + "type": "integer" + } +} +``` + +```yaml +description: The number of allowed requests in the current period +schema: + type: integer +``` + +#### Tag Object + +Adds metadata to a single tag that is used by the [Operation Object](#operationObject). +It is not mandatory to have a Tag Object per tag defined in the Operation Object instances. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +name | `string` | **REQUIRED**. The name of the tag. +description | `string` | A short description for the tag. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +externalDocs | [External Documentation Object](#externalDocumentationObject) | Additional external documentation for this tag. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Tag Object Example + +```json +{ + "name": "pet", + "description": "Pets operations" +} +``` + +```yaml +name: pet +description: Pets operations +``` + + +#### Reference Object + +A simple object to allow referencing other components in the specification, internally and externally. + +The Reference Object is defined by [JSON Reference](https://tools.ietf.org/html/draft-pbryan-zyp-json-ref-03) and follows the same structure, behavior and rules. + +For this specification, reference resolution is accomplished as defined by the JSON Reference specification and not by the JSON Schema specification. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +$ref | `string` | **REQUIRED**. The reference string. + +This object cannot be extended with additional properties and any properties added SHALL be ignored. + +##### Reference Object Example + +```json +{ + "$ref": "#/components/schemas/Pet" +} +``` + +```yaml +$ref: '#/components/schemas/Pet' +``` + +##### Relative Schema Document Example +```json +{ + "$ref": "Pet.json" +} +``` + +```yaml +$ref: Pet.yaml +``` + +##### Relative Documents With Embedded Schema Example +```json +{ + "$ref": "definitions.json#/Pet" +} +``` + +```yaml +$ref: definitions.yaml#/Pet +``` + +#### Schema Object + +The Schema Object allows the definition of input and output data types. +These types can be objects, but also primitives and arrays. +This object is an extended subset of the [JSON Schema Specification Wright Draft 00](http://json-schema.org/). + +For more information about the properties, see [JSON Schema Core](https://tools.ietf.org/html/draft-wright-json-schema-00) and [JSON Schema Validation](https://tools.ietf.org/html/draft-wright-json-schema-validation-00). +Unless stated otherwise, the property definitions follow the JSON Schema. + +##### Properties + +The following properties are taken directly from the JSON Schema definition and follow the same specifications: + +- title +- multipleOf +- maximum +- exclusiveMaximum +- minimum +- exclusiveMinimum +- maxLength +- minLength +- pattern (This string SHOULD be a valid regular expression, according to the [ECMA 262 regular expression](https://www.ecma-international.org/ecma-262/5.1/#sec-7.8.5) dialect) +- maxItems +- minItems +- uniqueItems +- maxProperties +- minProperties +- required +- enum + +The following properties are taken from the JSON Schema definition but their definitions were adjusted to the OpenAPI Specification. +- type - Value MUST be a string. Multiple types via an array are not supported. +- allOf - Inline or referenced schema MUST be of a [Schema Object](#schemaObject) and not a standard JSON Schema. +- oneOf - Inline or referenced schema MUST be of a [Schema Object](#schemaObject) and not a standard JSON Schema. +- anyOf - Inline or referenced schema MUST be of a [Schema Object](#schemaObject) and not a standard JSON Schema. +- not - Inline or referenced schema MUST be of a [Schema Object](#schemaObject) and not a standard JSON Schema. +- items - Value MUST be an object and not an array. Inline or referenced schema MUST be of a [Schema Object](#schemaObject) and not a standard JSON Schema. `items` MUST be present if the `type` is `array`. +- properties - Property definitions MUST be a [Schema Object](#schemaObject) and not a standard JSON Schema (inline or referenced). +- additionalProperties - Value can be boolean or object. Inline or referenced schema MUST be of a [Schema Object](#schemaObject) and not a standard JSON Schema. +- description - [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +- format - See [Data Type Formats](#dataTypeFormat) for further details. While relying on JSON Schema's defined formats, the OAS offers a few additional predefined formats. +- default - The default value represents what would be assumed by the consumer of the input as the value of the schema if one is not provided. Unlike JSON Schema, the value MUST conform to the defined type for the Schema Object defined at the same level. For example, if `type` is `string`, then `default` can be `"foo"` but cannot be `1`. + +Alternatively, any time a Schema Object can be used, a [Reference Object](#referenceObject) can be used in its place. This allows referencing definitions instead of defining them inline. + +Additional properties defined by the JSON Schema specification that are not mentioned here are strictly unsupported. + +Other than the JSON Schema subset fields, the following fields MAY be used for further schema documentation: + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +nullable | `boolean` | Allows sending a `null` value for the defined schema. Default value is `false`. +discriminator | [Discriminator Object](#discriminatorObject) | Adds support for polymorphism. The discriminator is an object name that is used to differentiate between other schemas which may satisfy the payload description. See [Composition and Inheritance](#schemaComposition) for more details. +readOnly | `boolean` | Relevant only for Schema `"properties"` definitions. Declares the property as "read only". This means that it MAY be sent as part of a response but SHOULD NOT be sent as part of the request. If the property is marked as `readOnly` being `true` and is in the `required` list, the `required` will take effect on the response only. A property MUST NOT be marked as both `readOnly` and `writeOnly` being `true`. Default value is `false`. +writeOnly | `boolean` | Relevant only for Schema `"properties"` definitions. Declares the property as "write only". Therefore, it MAY be sent as part of a request but SHOULD NOT be sent as part of the response. If the property is marked as `writeOnly` being `true` and is in the `required` list, the `required` will take effect on the request only. A property MUST NOT be marked as both `readOnly` and `writeOnly` being `true`. Default value is `false`. +xml | [XML Object](#xmlObject) | This MAY be used only on properties schemas. It has no effect on root schemas. Adds additional metadata to describe the XML representation of this property. +externalDocs | [External Documentation Object](#externalDocumentationObject) | Additional external documentation for this schema. +example | Any | A free-form property to include an example of an instance for this schema. To represent examples that cannot be naturally represented in JSON or YAML, a string value can be used to contain the example with escaping where necessary. + deprecated | `boolean` | Specifies that a schema is deprecated and SHOULD be transitioned out of usage. Default value is `false`. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +###### Composition and Inheritance (Polymorphism) + +The OpenAPI Specification allows combining and extending model definitions using the `allOf` property of JSON Schema, in effect offering model composition. +`allOf` takes an array of object definitions that are validated *independently* but together compose a single object. + +While composition offers model extensibility, it does not imply a hierarchy between the models. +To support polymorphism, the OpenAPI Specification adds the `discriminator` field. +When used, the `discriminator` will be the name of the property that decides which schema definition validates the structure of the model. +As such, the `discriminator` field MUST be a required field. +There are two ways to define the value of a discriminator for an inheriting instance. +- Use the schema name. +- Override the schema name by overriding the property with a new value. If a new value exists, this takes precedence over the schema name. +As such, inline schema definitions, which do not have a given id, *cannot* be used in polymorphism. + +###### XML Modeling + +The [xml](#schemaXml) property allows extra definitions when translating the JSON definition to XML. +The [XML Object](#xmlObject) contains additional information about the available options. + +##### Schema Object Examples + +###### Primitive Sample + +```json +{ + "type": "string", + "format": "email" +} +``` + +```yaml +type: string +format: email +``` + +###### Simple Model + +```json +{ + "type": "object", + "required": [ + "name" + ], + "properties": { + "name": { + "type": "string" + }, + "address": { + "$ref": "#/components/schemas/Address" + }, + "age": { + "type": "integer", + "format": "int32", + "minimum": 0 + } + } +} +``` + +```yaml +type: object +required: +- name +properties: + name: + type: string + address: + $ref: '#/components/schemas/Address' + age: + type: integer + format: int32 + minimum: 0 +``` + +###### Model with Map/Dictionary Properties + +For a simple string to string mapping: + +```json +{ + "type": "object", + "additionalProperties": { + "type": "string" + } +} +``` + +```yaml +type: object +additionalProperties: + type: string +``` + +For a string to model mapping: + +```json +{ + "type": "object", + "additionalProperties": { + "$ref": "#/components/schemas/ComplexModel" + } +} +``` + +```yaml +type: object +additionalProperties: + $ref: '#/components/schemas/ComplexModel' +``` + +###### Model with Example + +```json +{ + "type": "object", + "properties": { + "id": { + "type": "integer", + "format": "int64" + }, + "name": { + "type": "string" + } + }, + "required": [ + "name" + ], + "example": { + "name": "Puma", + "id": 1 + } +} +``` + +```yaml +type: object +properties: + id: + type: integer + format: int64 + name: + type: string +required: +- name +example: + name: Puma + id: 1 +``` + +###### Models with Composition + +```json +{ + "components": { + "schemas": { + "ErrorModel": { + "type": "object", + "required": [ + "message", + "code" + ], + "properties": { + "message": { + "type": "string" + }, + "code": { + "type": "integer", + "minimum": 100, + "maximum": 600 + } + } + }, + "ExtendedErrorModel": { + "allOf": [ + { + "$ref": "#/components/schemas/ErrorModel" + }, + { + "type": "object", + "required": [ + "rootCause" + ], + "properties": { + "rootCause": { + "type": "string" + } + } + } + ] + } + } + } +} +``` + +```yaml +components: + schemas: + ErrorModel: + type: object + required: + - message + - code + properties: + message: + type: string + code: + type: integer + minimum: 100 + maximum: 600 + ExtendedErrorModel: + allOf: + - $ref: '#/components/schemas/ErrorModel' + - type: object + required: + - rootCause + properties: + rootCause: + type: string +``` + +###### Models with Polymorphism Support + +```json +{ + "components": { + "schemas": { + "Pet": { + "type": "object", + "discriminator": { + "propertyName": "petType" + }, + "properties": { + "name": { + "type": "string" + }, + "petType": { + "type": "string" + } + }, + "required": [ + "name", + "petType" + ] + }, + "Cat": { + "description": "A representation of a cat. Note that `Cat` will be used as the discriminator value.", + "allOf": [ + { + "$ref": "#/components/schemas/Pet" + }, + { + "type": "object", + "properties": { + "huntingSkill": { + "type": "string", + "description": "The measured skill for hunting", + "default": "lazy", + "enum": [ + "clueless", + "lazy", + "adventurous", + "aggressive" + ] + } + }, + "required": [ + "huntingSkill" + ] + } + ] + }, + "Dog": { + "description": "A representation of a dog. Note that `Dog` will be used as the discriminator value.", + "allOf": [ + { + "$ref": "#/components/schemas/Pet" + }, + { + "type": "object", + "properties": { + "packSize": { + "type": "integer", + "format": "int32", + "description": "the size of the pack the dog is from", + "default": 0, + "minimum": 0 + } + }, + "required": [ + "packSize" + ] + } + ] + } + } + } +} +``` + +```yaml +components: + schemas: + Pet: + type: object + discriminator: + propertyName: petType + properties: + name: + type: string + petType: + type: string + required: + - name + - petType + Cat: ## "Cat" will be used as the discriminator value + description: A representation of a cat + allOf: + - $ref: '#/components/schemas/Pet' + - type: object + properties: + huntingSkill: + type: string + description: The measured skill for hunting + enum: + - clueless + - lazy + - adventurous + - aggressive + required: + - huntingSkill + Dog: ## "Dog" will be used as the discriminator value + description: A representation of a dog + allOf: + - $ref: '#/components/schemas/Pet' + - type: object + properties: + packSize: + type: integer + format: int32 + description: the size of the pack the dog is from + default: 0 + minimum: 0 + required: + - packSize +``` + +#### Discriminator Object + +When request bodies or response payloads may be one of a number of different schemas, a `discriminator` object can be used to aid in serialization, deserialization, and validation. The discriminator is a specific object in a schema which is used to inform the consumer of the specification of an alternative schema based on the value associated with it. + +When using the discriminator, _inline_ schemas will not be considered. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +propertyName | `string` | **REQUIRED**. The name of the property in the payload that will hold the discriminator value. + mapping | Map[`string`, `string`] | An object to hold mappings between payload values and schema names or references. + +The discriminator attribute is legal only when using one of the composite keywords `oneOf`, `anyOf`, `allOf`. + +In OAS 3.0, a response payload MAY be described to be exactly one of any number of types: + +``` +MyResponseType: + oneOf: + - $ref: '#/components/schemas/Cat' + - $ref: '#/components/schemas/Dog' + - $ref: '#/components/schemas/Lizard' +``` + +which means the payload _MUST_, by validation, match exactly one of the schemas described by `Cat`, `Dog`, or `Lizard`. In this case, a discriminator MAY act as a "hint" to shortcut validation and selection of the matching schema which may be a costly operation, depending on the complexity of the schema. We can then describe exactly which field tells us which schema to use: + + +``` +MyResponseType: + oneOf: + - $ref: '#/components/schemas/Cat' + - $ref: '#/components/schemas/Dog' + - $ref: '#/components/schemas/Lizard' + discriminator: + propertyName: pet_type +``` + +The expectation now is that a property with name `pet_type` _MUST_ be present in the response payload, and the value will correspond to the name of a schema defined in the OAS document. Thus the response payload: + +``` +{ + "id": 12345, + "pet_type": "Cat" +} +``` + +Will indicate that the `Cat` schema be used in conjunction with this payload. + +In scenarios where the value of the discriminator field does not match the schema name or implicit mapping is not possible, an optional `mapping` definition MAY be used: + +``` +MyResponseType: + oneOf: + - $ref: '#/components/schemas/Cat' + - $ref: '#/components/schemas/Dog' + - $ref: '#/components/schemas/Lizard' + - $ref: 'https://gigantic-server.com/schemas/Monster/schema.json' + discriminator: + propertyName: pet_type + mapping: + dog: '#/components/schemas/Dog' + monster: 'https://gigantic-server.com/schemas/Monster/schema.json' +``` + +Here the discriminator _value_ of `dog` will map to the schema `#/components/schemas/Dog`, rather than the default (implicit) value of `Dog`. If the discriminator _value_ does not match an implicit or explicit mapping, no schema can be determined and validation SHOULD fail. Mapping keys MUST be string values, but tooling MAY convert response values to strings for comparison. + +When used in conjunction with the `anyOf` construct, the use of the discriminator can avoid ambiguity where multiple schemas may satisfy a single payload. + +In both the `oneOf` and `anyOf` use cases, all possible schemas MUST be listed explicitly. To avoid redundancy, the discriminator MAY be added to a parent schema definition, and all schemas comprising the parent schema in an `allOf` construct may be used as an alternate schema. + +For example: + +``` +components: + schemas: + Pet: + type: object + required: + - pet_type + properties: + pet_type: + type: string + discriminator: + propertyName: pet_type + mapping: + cachorro: Dog + Cat: + allOf: + - $ref: '#/components/schemas/Pet' + - type: object + # all other properties specific to a `Cat` + properties: + name: + type: string + Dog: + allOf: + - $ref: '#/components/schemas/Pet' + - type: object + # all other properties specific to a `Dog` + properties: + bark: + type: string + Lizard: + allOf: + - $ref: '#/components/schemas/Pet' + - type: object + # all other properties specific to a `Lizard` + properties: + lovesRocks: + type: boolean +``` + +a payload like this: + +``` +{ + "pet_type": "Cat", + "name": "misty" +} +``` + +will indicate that the `Cat` schema be used. Likewise this schema: + +``` +{ + "pet_type": "cachorro", + "bark": "soft" +} +``` + +will map to `Dog` because of the definition in the `mappings` element. + + +#### XML Object + +A metadata object that allows for more fine-tuned XML model definitions. + +When using arrays, XML element names are *not* inferred (for singular/plural forms) and the `name` property SHOULD be used to add that information. +See examples for expected behavior. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +name | `string` | Replaces the name of the element/attribute used for the described schema property. When defined within `items`, it will affect the name of the individual XML elements within the list. When defined alongside `type` being `array` (outside the `items`), it will affect the wrapping element and only if `wrapped` is `true`. If `wrapped` is `false`, it will be ignored. +namespace | `string` | The URI of the namespace definition. Value MUST be in the form of an absolute URI. +prefix | `string` | The prefix to be used for the [name](#xmlName). +attribute | `boolean` | Declares whether the property definition translates to an attribute instead of an element. Default value is `false`. +wrapped | `boolean` | MAY be used only for an array definition. Signifies whether the array is wrapped (for example, ``) or unwrapped (``). Default value is `false`. The definition takes effect only when defined alongside `type` being `array` (outside the `items`). + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### XML Object Examples + +The examples of the XML object definitions are included inside a property definition of a [Schema Object](#schemaObject) with a sample of the XML representation of it. + +###### No XML Element + +Basic string property: + +```json +{ + "animals": { + "type": "string" + } +} +``` + +```yaml +animals: + type: string +``` + +```xml +... +``` + +Basic string array property ([`wrapped`](#xmlWrapped) is `false` by default): + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string" + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string +``` + +```xml +... +... +... +``` + +###### XML Name Replacement + +```json +{ + "animals": { + "type": "string", + "xml": { + "name": "animal" + } + } +} +``` + +```yaml +animals: + type: string + xml: + name: animal +``` + +```xml +... +``` + + +###### XML Attribute, Prefix and Namespace + +In this example, a full model definition is shown. + +```json +{ + "Person": { + "type": "object", + "properties": { + "id": { + "type": "integer", + "format": "int32", + "xml": { + "attribute": true + } + }, + "name": { + "type": "string", + "xml": { + "namespace": "http://example.com/schema/sample", + "prefix": "sample" + } + } + } + } +} +``` + +```yaml +Person: + type: object + properties: + id: + type: integer + format: int32 + xml: + attribute: true + name: + type: string + xml: + namespace: http://example.com/schema/sample + prefix: sample +``` + +```xml + + example + +``` + +###### XML Arrays + +Changing the element names: + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string", + "xml": { + "name": "animal" + } + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string + xml: + name: animal +``` + +```xml +value +value +``` + +The external `name` property has no effect on the XML: + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string", + "xml": { + "name": "animal" + } + }, + "xml": { + "name": "aliens" + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string + xml: + name: animal + xml: + name: aliens +``` + +```xml +value +value +``` + +Even when the array is wrapped, if a name is not explicitly defined, the same name will be used both internally and externally: + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string" + }, + "xml": { + "wrapped": true + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string + xml: + wrapped: true +``` + +```xml + + value + value + +``` + +To overcome the naming problem in the example above, the following definition can be used: + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string", + "xml": { + "name": "animal" + } + }, + "xml": { + "wrapped": true + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string + xml: + name: animal + xml: + wrapped: true +``` + +```xml + + value + value + +``` + +Affecting both internal and external names: + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string", + "xml": { + "name": "animal" + } + }, + "xml": { + "name": "aliens", + "wrapped": true + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string + xml: + name: animal + xml: + name: aliens + wrapped: true +``` + +```xml + + value + value + +``` + +If we change the external element but not the internal ones: + +```json +{ + "animals": { + "type": "array", + "items": { + "type": "string" + }, + "xml": { + "name": "aliens", + "wrapped": true + } + } +} +``` + +```yaml +animals: + type: array + items: + type: string + xml: + name: aliens + wrapped: true +``` + +```xml + + value + value + +``` + +#### Security Scheme Object + +Defines a security scheme that can be used by the operations. +Supported schemes are HTTP authentication, an API key (either as a header or as a query parameter), OAuth2's common flows (implicit, password, application and access code) as defined in [RFC6749](https://tools.ietf.org/html/rfc6749), and [OpenID Connect Discovery](https://tools.ietf.org/html/draft-ietf-oauth-discovery-06). + +##### Fixed Fields +Field Name | Type | Applies To | Description +---|:---:|---|--- +type | `string` | Any | **REQUIRED**. The type of the security scheme. Valid values are `"apiKey"`, `"http"`, `"oauth2"`, `"openIdConnect"`. +description | `string` | Any | A short description for security scheme. [CommonMark syntax](http://spec.commonmark.org/) MAY be used for rich text representation. +name | `string` | `apiKey` | **REQUIRED**. The name of the header, query or cookie parameter to be used. +in | `string` | `apiKey` | **REQUIRED**. The location of the API key. Valid values are `"query"`, `"header"` or `"cookie"`. +scheme | `string` | `http` | **REQUIRED**. The name of the HTTP Authorization scheme to be used in the [Authorization header as defined in RFC7235](https://tools.ietf.org/html/rfc7235#section-5.1). +bearerFormat | `string` | `http` (`"bearer"`) | A hint to the client to identify how the bearer token is formatted. Bearer tokens are usually generated by an authorization server, so this information is primarily for documentation purposes. +flows | [OAuth Flows Object](#oauthFlowsObject) | `oauth2` | **REQUIRED**. An object containing configuration information for the flow types supported. +openIdConnectUrl | `string` | `openIdConnect` | **REQUIRED**. OpenId Connect URL to discover OAuth2 configuration values. This MUST be in the form of a URL. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### Security Scheme Object Example + +###### Basic Authentication Sample + +```json +{ + "type": "http", + "scheme": "basic" +} +``` + +```yaml +type: http +scheme: basic +``` + +###### API Key Sample + +```json +{ + "type": "apiKey", + "name": "api_key", + "in": "header" +} +``` + +```yaml +type: apiKey +name: api_key +in: header +``` + +###### JWT Bearer Sample + +```json +{ + "type": "http", + "scheme": "bearer", + "bearerFormat": "JWT", +} +``` + +```yaml +type: http +scheme: bearer +bearerFormat: JWT +``` + +###### Implicit OAuth2 Sample + +```json +{ + "type": "oauth2", + "flows": { + "implicit": { + "authorizationUrl": "https://example.com/api/oauth/dialog", + "scopes": { + "write:pets": "modify pets in your account", + "read:pets": "read your pets" + } + } + } +} +``` + +```yaml +type: oauth2 +flows: + implicit: + authorizationUrl: https://example.com/api/oauth/dialog + scopes: + write:pets: modify pets in your account + read:pets: read your pets +``` + +#### OAuth Flows Object + +Allows configuration of the supported OAuth Flows. + +##### Fixed Fields +Field Name | Type | Description +---|:---:|--- +implicit| [OAuth Flow Object](#oauthFlowObject) | Configuration for the OAuth Implicit flow +password| [OAuth Flow Object](#oauthFlowObject) | Configuration for the OAuth Resource Owner Password flow +clientCredentials| [OAuth Flow Object](#oauthFlowObject) | Configuration for the OAuth Client Credentials flow. Previously called `application` in OpenAPI 2.0. +authorizationCode| [OAuth Flow Object](#oauthFlowObject) | Configuration for the OAuth Authorization Code flow. Previously called `accessCode` in OpenAPI 2.0. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +#### OAuth Flow Object + +Configuration details for a supported OAuth Flow + +##### Fixed Fields +Field Name | Type | Applies To | Description +---|:---:|---|--- +authorizationUrl | `string` | `oauth2` (`"implicit"`, `"authorizationCode"`) | **REQUIRED**. The authorization URL to be used for this flow. This MUST be in the form of a URL. +tokenUrl | `string` | `oauth2` (`"password"`, `"clientCredentials"`, `"authorizationCode"`) | **REQUIRED**. The token URL to be used for this flow. This MUST be in the form of a URL. +refreshUrl | `string` | `oauth2` | The URL to be used for obtaining refresh tokens. This MUST be in the form of a URL. +scopes | Map[`string`, `string`] | `oauth2` | **REQUIRED**. The available scopes for the OAuth2 security scheme. A map between the scope name and a short description for it. + +This object MAY be extended with [Specification Extensions](#specificationExtensions). + +##### OAuth Flow Object Examples + +```JSON +{ + "type": "oauth2", + "flows": { + "implicit": { + "authorizationUrl": "https://example.com/api/oauth/dialog", + "scopes": { + "write:pets": "modify pets in your account", + "read:pets": "read your pets" + } + }, + "authorizationCode": { + "authorizationUrl": "https://example.com/api/oauth/dialog", + "tokenUrl": "https://example.com/api/oauth/token", + "scopes": { + "write:pets": "modify pets in your account", + "read:pets": "read your pets" + } + } + } +} +``` + +```YAML +type: oauth2 +flows: + implicit: + authorizationUrl: https://example.com/api/oauth/dialog + scopes: + write:pets: modify pets in your account + read:pets: read your pets + authorizationCode: + authorizationUrl: https://example.com/api/oauth/dialog + tokenUrl: https://example.com/api/oauth/token + scopes: + write:pets: modify pets in your account + read:pets: read your pets +``` + + +#### Security Requirement Object + +Lists the required security schemes to execute this operation. +The name used for each property MUST correspond to a security scheme declared in the [Security Schemes](#componentsSecuritySchemes) under the [Components Object](#componentsObject). + +Security Requirement Objects that contain multiple schemes require that all schemes MUST be satisfied for a request to be authorized. +This enables support for scenarios where multiple query parameters or HTTP headers are required to convey security information. + +When a list of Security Requirement Objects is defined on the [Open API object](#oasObject) or [Operation Object](#operationObject), only one of Security Requirement Objects in the list needs to be satisfied to authorize the request. + +##### Patterned Fields + +Field Pattern | Type | Description +---|:---:|--- +{name} | [`string`] | Each name MUST correspond to a security scheme which is declared in the [Security Schemes](#componentsSecuritySchemes) under the [Components Object](#componentsObject). If the security scheme is of type `"oauth2"` or `"openIdConnect"`, then the value is a list of scope names required for the execution. For other security scheme types, the array MUST be empty. + +##### Security Requirement Object Examples + +###### Non-OAuth2 Security Requirement + +```json +{ + "api_key": [] +} +``` + +```yaml +api_key: [] +``` + +###### OAuth2 Security Requirement + +```json +{ + "petstore_auth": [ + "write:pets", + "read:pets" + ] +} +``` + +```yaml +petstore_auth: +- write:pets +- read:pets +``` + +### Specification Extensions + +While the OpenAPI Specification tries to accommodate most use cases, additional data can be added to extend the specification at certain points. + +The extensions properties are implemented as patterned fields that are always prefixed by `"x-"`. + +Field Pattern | Type | Description +---|:---:|--- +^x- | Any | Allows extensions to the OpenAPI Schema. The field name MUST begin with `x-`, for example, `x-internal-id`. The value can be `null`, a primitive, an array or an object. Can have any valid JSON format value. + +The extensions may or may not be supported by the available tooling, but those may be extended as well to add requested support (if tools are internal or open-sourced). + +### Security Filtering + +Some objects in the OpenAPI Specification MAY be declared and remain empty, or be completely removed, even though they are inherently the core of the API documentation. + +The reasoning is to allow an additional layer of access control over the documentation. +While not part of the specification itself, certain libraries MAY choose to allow access to parts of the documentation based on some form of authentication/authorization. + +Two examples of this: + +1. The [Paths Object](#pathsObject) MAY be empty. It may be counterintuitive, but this may tell the viewer that they got to the right place, but can't access any documentation. They'd still have access to the [Info Object](#infoObject) which may contain additional information regarding authentication. +2. The [Path Item Object](#pathItemObject) MAY be empty. In this case, the viewer will be aware that the path exists, but will not be able to see any of its operations or parameters. This is different than hiding the path itself from the [Paths Object](#pathsObject), so the user will not be aware of its existence. This allows the documentation provider to finely control what the viewer can see. + +## Appendix A: Revision History + +Version | Date | Notes +--- | --- | --- +3.0.1 | 2017-12-06 | Patch release of the OpenAPI Specification 3.0.1 +3.0.0 | 2017-07-26 | Release of the OpenAPI Specification 3.0.0 +3.0.0-rc2 | 2017-06-16 | rc2 of the 3.0 specification +3.0.0-rc1 | 2017-04-27 | rc1 of the 3.0 specification +3.0.0-rc0 | 2017-02-28 | Implementer's Draft of the 3.0 specification +2.0 | 2015-12-31 | Donation of Swagger 2.0 to the Open API Initiative +2.0 | 2014-09-08 | Release of Swagger 2.0 +1.2 | 2014-03-14 | Initial release of the formal document. +1.1 | 2012-08-22 | Release of Swagger 1.1 +1.0 | 2011-08-10 | First release of the Swagger Specification diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/README.md b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/README.md index 3987bd7b9..9ac297a5c 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/README.md +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/README.md @@ -4,10 +4,7 @@ This directory contains a support tool that reads (scrapes) the Markdown text specification for OpenAPI 3.0 and builds a corresponding JSON schema. -It also contains "3.0.md", a local copy of the OpenAPI specification -with modifications that fix minor inconsistencies and make it easier -to read. We hope to have these changes merged into the official -document. +It also contains "3.0.1.md", a local copy of the OpenAPI specification. ## Disclaimer diff --git a/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/main.go b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/main.go index 91459310a..1f3d03784 100644 --- a/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/main.go +++ b/vendor/github.com/googleapis/gnostic/OpenAPIv3/schema-generator/main.go @@ -303,7 +303,7 @@ func (m *SchemaModel) objectWithID(id string) *SchemaObject { // NewSchemaModel returns a new SchemaModel. func NewSchemaModel(filename string) (schemaModel *SchemaModel, err error) { - b, err := ioutil.ReadFile("3.0.md") + b, err := ioutil.ReadFile("3.0.1.md") if err != nil { return nil, err } @@ -837,6 +837,19 @@ func main() { *(contentObject.PatternProperties) = append(*(contentObject.PatternProperties), namedSchema) } + // fix the contact object + contactObject := schema.DefinitionWithName("contact") + if contactObject != nil { + emailProperty := contactObject.PropertyWithName("email") + if emailProperty != nil { + emailProperty.Format = stringptr("email"); + } + urlProperty := contactObject.PropertyWithName("url") + if urlProperty != nil { + urlProperty.Format = stringptr("uri"); + } + } + // write the updated schema output := schema.JSONString() err = ioutil.WriteFile("schema.json", []byte(output), 0644) diff --git a/vendor/github.com/googleapis/gnostic/README.md b/vendor/github.com/googleapis/gnostic/README.md index d350f3f01..97de63590 100644 --- a/vendor/github.com/googleapis/gnostic/README.md +++ b/vendor/github.com/googleapis/gnostic/README.md @@ -61,8 +61,14 @@ buffers. Pre-generated versions of these files are in the OpenAPIv2 directory. 3. [Optional] Generate Protocol Buffer support code. A pre-generated version of this file is checked into the OpenAPIv2 directory. -This step requires a local installation of protoc, the Protocol Buffer Compiler. +This step requires a local installation of protoc, the Protocol Buffer Compiler, +and the Go protoc plugin. You can get protoc [here](https://github.com/google/protobuf). +You can install the plugin with this command: + + go get -u github.com/golang/protobuf/protoc-gen-go + +Then use the following to recompile the Gnostic Protocol Buffer models: ./COMPILE-PROTOS.sh @@ -74,7 +80,7 @@ You can get protoc [here](https://github.com/google/protobuf). 5. Run **gnostic**. This will create a file in the current directory named "petstore.pb" that contains a binary Protocol Buffer description of a sample API. - gnostic --pb-out=. examples/petstore.json + gnostic --pb-out=. examples/v2.0/json/petstore.json 6. You can also compile files that you specify with a URL. Here's another way to compile the previous example. This time we're creating "petstore.text", which contains a textual representation of the @@ -92,7 +98,7 @@ that reports some basic information about an API. The "-" causes the plugin to write its output to stdout. go install github.com/googleapis/gnostic/plugins/gnostic-go-sample - gnostic examples/petstore.json --go-sample-out=- + gnostic examples/v2.0/json/petstore.json --go-sample-out=- ## Copyright diff --git a/vendor/github.com/googleapis/gnostic/apps/disco/main.go b/vendor/github.com/googleapis/gnostic/apps/disco/main.go index 4f6135c2f..71e1c0584 100644 --- a/vendor/github.com/googleapis/gnostic/apps/disco/main.go +++ b/vendor/github.com/googleapis/gnostic/apps/disco/main.go @@ -151,7 +151,7 @@ Usage: func handleExportArgumentsForBytes(arguments map[string]interface{}, bytes []byte) (handled bool, err error) { // Unpack the discovery document. - info, err := compiler.ReadInfoFromBytes("source", bytes) + info, err := compiler.ReadInfoFromBytes("", bytes) if err != nil { return true, err } diff --git a/vendor/github.com/googleapis/gnostic/apps/disco/openapiv3.go b/vendor/github.com/googleapis/gnostic/apps/disco/openapiv3.go index 763014aef..6c274572a 100644 --- a/vendor/github.com/googleapis/gnostic/apps/disco/openapiv3.go +++ b/vendor/github.com/googleapis/gnostic/apps/disco/openapiv3.go @@ -18,8 +18,8 @@ import ( "net/url" "strings" - discovery "github.com/googleapis/gnostic/discovery" openapi3 "github.com/googleapis/gnostic/OpenAPIv3" + discovery "github.com/googleapis/gnostic/discovery" ) func pathForMethod(path string) string { @@ -59,9 +59,9 @@ func buildOpenAPI3SchemaOrReferenceForSchema(schema *discovery.Schema) *openapi3 } } if schema.Items != nil { - s2 := buildOpenAPI3SchemaOrReferenceForSchema(schema.Items) - s.Items = &openapi3.ItemsItem{} - s.Items.SchemaOrReference = append(s.Items.SchemaOrReference, s2) + s.Items = &openapi3.ItemsItem{ + SchemaOrReference: []*openapi3.SchemaOrReference{buildOpenAPI3SchemaOrReferenceForSchema(schema.Items)}, + } } if (schema.Properties != nil) && (len(schema.Properties.AdditionalProperties) > 0) { s.Properties = &openapi3.Properties{} diff --git a/vendor/github.com/googleapis/gnostic/apps/report-messages/README.md b/vendor/github.com/googleapis/gnostic/apps/report-messages/README.md new file mode 100644 index 000000000..e352b0fee --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/apps/report-messages/README.md @@ -0,0 +1,5 @@ +# report-messages tool + +This directory contains a command-line tool that provides a text +report listing the messages in a gnostic messages file. + diff --git a/vendor/github.com/googleapis/gnostic/apps/report-messages/main.go b/vendor/github.com/googleapis/gnostic/apps/report-messages/main.go new file mode 100644 index 000000000..b11e50b1a --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/apps/report-messages/main.go @@ -0,0 +1,69 @@ +// Copyright 2018 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Filter and display messages produced by gnostic invocations. +package main + +import ( + "flag" + "fmt" + "io/ioutil" + "os" + + "github.com/golang/protobuf/proto" + "github.com/googleapis/gnostic/printer" + + plugins "github.com/googleapis/gnostic/plugins" +) + +func readMessagesFromFileWithName(filename string) *plugins.Messages { + data, err := ioutil.ReadFile(filename) + if err != nil { + fmt.Printf("File error: %v\n", err) + os.Exit(1) + } + messages := &plugins.Messages{} + err = proto.Unmarshal(data, messages) + if err != nil { + panic(err) + } + return messages +} + +func printMessages(code *printer.Code, messages *plugins.Messages) { + for _, message := range messages.Messages { + line := fmt.Sprintf("%-7s %-14s %s %+v", + message.Level, + message.Code, + message.Text, + message.Keys) + code.Print(line) + } +} + +func main() { + flag.Parse() + args := flag.Args() + + if len(args) != 1 { + fmt.Printf("Usage: report-messages \n") + return + } + + messages := readMessagesFromFileWithName(args[0]) + + code := &printer.Code{} + printMessages(code, messages) + fmt.Printf("%s", code) +} diff --git a/vendor/github.com/googleapis/gnostic/compiler/reader.go b/vendor/github.com/googleapis/gnostic/compiler/reader.go index 2d4b3303d..c954a2d9b 100644 --- a/vendor/github.com/googleapis/gnostic/compiler/reader.go +++ b/vendor/github.com/googleapis/gnostic/compiler/reader.go @@ -110,7 +110,9 @@ func ReadInfoFromBytes(filename string, bytes []byte) (interface{}, error) { if err != nil { return nil, err } - infoCache[filename] = info + if len(filename) > 0 { + infoCache[filename] = info + } return info, nil } diff --git a/vendor/github.com/googleapis/gnostic/discovery/discovery.go b/vendor/github.com/googleapis/gnostic/discovery/discovery.go index 6e2788028..43b078175 100644 --- a/vendor/github.com/googleapis/gnostic/discovery/discovery.go +++ b/vendor/github.com/googleapis/gnostic/discovery/discovery.go @@ -2124,7 +2124,7 @@ func (m *StringArray) ResolveReferences(root string) (interface{}, error) { func (m *Annotations) ToRawInfo() interface{} { info := yaml.MapSlice{} if len(m.Required) != 0 { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } return info } @@ -2154,7 +2154,7 @@ func (m *Any) ToRawInfo() interface{} { func (m *Auth) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Oauth2 != nil { - info = append(info, yaml.MapItem{"oauth2", m.Oauth2.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "oauth2", Value: m.Oauth2.ToRawInfo()}) } // &{Name:oauth2 Type:Oauth2 StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} return info @@ -2164,100 +2164,100 @@ func (m *Auth) ToRawInfo() interface{} { func (m *Document) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Kind != "" { - info = append(info, yaml.MapItem{"kind", m.Kind}) + info = append(info, yaml.MapItem{Key: "kind", Value: m.Kind}) } if m.DiscoveryVersion != "" { - info = append(info, yaml.MapItem{"discoveryVersion", m.DiscoveryVersion}) + info = append(info, yaml.MapItem{Key: "discoveryVersion", Value: m.DiscoveryVersion}) } if m.Id != "" { - info = append(info, yaml.MapItem{"id", m.Id}) + info = append(info, yaml.MapItem{Key: "id", Value: m.Id}) } if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } if m.Version != "" { - info = append(info, yaml.MapItem{"version", m.Version}) + info = append(info, yaml.MapItem{Key: "version", Value: m.Version}) } if m.Revision != "" { - info = append(info, yaml.MapItem{"revision", m.Revision}) + info = append(info, yaml.MapItem{Key: "revision", Value: m.Revision}) } if m.Title != "" { - info = append(info, yaml.MapItem{"title", m.Title}) + info = append(info, yaml.MapItem{Key: "title", Value: m.Title}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Icons != nil { - info = append(info, yaml.MapItem{"icons", m.Icons.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "icons", Value: m.Icons.ToRawInfo()}) } // &{Name:icons Type:Icons StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.DocumentationLink != "" { - info = append(info, yaml.MapItem{"documentationLink", m.DocumentationLink}) + info = append(info, yaml.MapItem{Key: "documentationLink", Value: m.DocumentationLink}) } if len(m.Labels) != 0 { - info = append(info, yaml.MapItem{"labels", m.Labels}) + info = append(info, yaml.MapItem{Key: "labels", Value: m.Labels}) } if m.Protocol != "" { - info = append(info, yaml.MapItem{"protocol", m.Protocol}) + info = append(info, yaml.MapItem{Key: "protocol", Value: m.Protocol}) } if m.BaseUrl != "" { - info = append(info, yaml.MapItem{"baseUrl", m.BaseUrl}) + info = append(info, yaml.MapItem{Key: "baseUrl", Value: m.BaseUrl}) } if m.BasePath != "" { - info = append(info, yaml.MapItem{"basePath", m.BasePath}) + info = append(info, yaml.MapItem{Key: "basePath", Value: m.BasePath}) } if m.RootUrl != "" { - info = append(info, yaml.MapItem{"rootUrl", m.RootUrl}) + info = append(info, yaml.MapItem{Key: "rootUrl", Value: m.RootUrl}) } if m.ServicePath != "" { - info = append(info, yaml.MapItem{"servicePath", m.ServicePath}) + info = append(info, yaml.MapItem{Key: "servicePath", Value: m.ServicePath}) } if m.BatchPath != "" { - info = append(info, yaml.MapItem{"batchPath", m.BatchPath}) + info = append(info, yaml.MapItem{Key: "batchPath", Value: m.BatchPath}) } if m.Parameters != nil { - info = append(info, yaml.MapItem{"parameters", m.Parameters.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "parameters", Value: m.Parameters.ToRawInfo()}) } // &{Name:parameters Type:Parameters StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Auth != nil { - info = append(info, yaml.MapItem{"auth", m.Auth.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "auth", Value: m.Auth.ToRawInfo()}) } // &{Name:auth Type:Auth StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Features) != 0 { - info = append(info, yaml.MapItem{"features", m.Features}) + info = append(info, yaml.MapItem{Key: "features", Value: m.Features}) } if m.Schemas != nil { - info = append(info, yaml.MapItem{"schemas", m.Schemas.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "schemas", Value: m.Schemas.ToRawInfo()}) } // &{Name:schemas Type:Schemas StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Methods != nil { - info = append(info, yaml.MapItem{"methods", m.Methods.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "methods", Value: m.Methods.ToRawInfo()}) } // &{Name:methods Type:Methods StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Resources != nil { - info = append(info, yaml.MapItem{"resources", m.Resources.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "resources", Value: m.Resources.ToRawInfo()}) } // &{Name:resources Type:Resources StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Etag != "" { - info = append(info, yaml.MapItem{"etag", m.Etag}) + info = append(info, yaml.MapItem{Key: "etag", Value: m.Etag}) } if m.OwnerDomain != "" { - info = append(info, yaml.MapItem{"ownerDomain", m.OwnerDomain}) + info = append(info, yaml.MapItem{Key: "ownerDomain", Value: m.OwnerDomain}) } if m.OwnerName != "" { - info = append(info, yaml.MapItem{"ownerName", m.OwnerName}) + info = append(info, yaml.MapItem{Key: "ownerName", Value: m.OwnerName}) } if m.VersionModule != false { - info = append(info, yaml.MapItem{"version_module", m.VersionModule}) + info = append(info, yaml.MapItem{Key: "version_module", Value: m.VersionModule}) } if m.CanonicalName != "" { - info = append(info, yaml.MapItem{"canonicalName", m.CanonicalName}) + info = append(info, yaml.MapItem{Key: "canonicalName", Value: m.CanonicalName}) } if m.FullyEncodeReservedExpansion != false { - info = append(info, yaml.MapItem{"fullyEncodeReservedExpansion", m.FullyEncodeReservedExpansion}) + info = append(info, yaml.MapItem{Key: "fullyEncodeReservedExpansion", Value: m.FullyEncodeReservedExpansion}) } if m.PackagePath != "" { - info = append(info, yaml.MapItem{"packagePath", m.PackagePath}) + info = append(info, yaml.MapItem{Key: "packagePath", Value: m.PackagePath}) } return info } @@ -2266,10 +2266,10 @@ func (m *Document) ToRawInfo() interface{} { func (m *Icons) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.X16 != "" { - info = append(info, yaml.MapItem{"x16", m.X16}) + info = append(info, yaml.MapItem{Key: "x16", Value: m.X16}) } if m.X32 != "" { - info = append(info, yaml.MapItem{"x32", m.X32}) + info = append(info, yaml.MapItem{Key: "x32", Value: m.X32}) } return info } @@ -2278,17 +2278,17 @@ func (m *Icons) ToRawInfo() interface{} { func (m *MediaUpload) ToRawInfo() interface{} { info := yaml.MapSlice{} if len(m.Accept) != 0 { - info = append(info, yaml.MapItem{"accept", m.Accept}) + info = append(info, yaml.MapItem{Key: "accept", Value: m.Accept}) } if m.MaxSize != "" { - info = append(info, yaml.MapItem{"maxSize", m.MaxSize}) + info = append(info, yaml.MapItem{Key: "maxSize", Value: m.MaxSize}) } if m.Protocols != nil { - info = append(info, yaml.MapItem{"protocols", m.Protocols.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "protocols", Value: m.Protocols.ToRawInfo()}) } // &{Name:protocols Type:Protocols StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SupportsSubscription != false { - info = append(info, yaml.MapItem{"supportsSubscription", m.SupportsSubscription}) + info = append(info, yaml.MapItem{Key: "supportsSubscription", Value: m.SupportsSubscription}) } return info } @@ -2297,56 +2297,56 @@ func (m *MediaUpload) ToRawInfo() interface{} { func (m *Method) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Id != "" { - info = append(info, yaml.MapItem{"id", m.Id}) + info = append(info, yaml.MapItem{Key: "id", Value: m.Id}) } if m.Path != "" { - info = append(info, yaml.MapItem{"path", m.Path}) + info = append(info, yaml.MapItem{Key: "path", Value: m.Path}) } if m.HttpMethod != "" { - info = append(info, yaml.MapItem{"httpMethod", m.HttpMethod}) + info = append(info, yaml.MapItem{Key: "httpMethod", Value: m.HttpMethod}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Parameters != nil { - info = append(info, yaml.MapItem{"parameters", m.Parameters.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "parameters", Value: m.Parameters.ToRawInfo()}) } // &{Name:parameters Type:Parameters StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.ParameterOrder) != 0 { - info = append(info, yaml.MapItem{"parameterOrder", m.ParameterOrder}) + info = append(info, yaml.MapItem{Key: "parameterOrder", Value: m.ParameterOrder}) } if m.Request != nil { - info = append(info, yaml.MapItem{"request", m.Request.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "request", Value: m.Request.ToRawInfo()}) } // &{Name:request Type:Request StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Response != nil { - info = append(info, yaml.MapItem{"response", m.Response.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "response", Value: m.Response.ToRawInfo()}) } // &{Name:response Type:Response StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if len(m.Scopes) != 0 { - info = append(info, yaml.MapItem{"scopes", m.Scopes}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes}) } if m.SupportsMediaDownload != false { - info = append(info, yaml.MapItem{"supportsMediaDownload", m.SupportsMediaDownload}) + info = append(info, yaml.MapItem{Key: "supportsMediaDownload", Value: m.SupportsMediaDownload}) } if m.SupportsMediaUpload != false { - info = append(info, yaml.MapItem{"supportsMediaUpload", m.SupportsMediaUpload}) + info = append(info, yaml.MapItem{Key: "supportsMediaUpload", Value: m.SupportsMediaUpload}) } if m.UseMediaDownloadService != false { - info = append(info, yaml.MapItem{"useMediaDownloadService", m.UseMediaDownloadService}) + info = append(info, yaml.MapItem{Key: "useMediaDownloadService", Value: m.UseMediaDownloadService}) } if m.MediaUpload != nil { - info = append(info, yaml.MapItem{"mediaUpload", m.MediaUpload.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "mediaUpload", Value: m.MediaUpload.ToRawInfo()}) } // &{Name:mediaUpload Type:MediaUpload StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.SupportsSubscription != false { - info = append(info, yaml.MapItem{"supportsSubscription", m.SupportsSubscription}) + info = append(info, yaml.MapItem{Key: "supportsSubscription", Value: m.SupportsSubscription}) } if m.FlatPath != "" { - info = append(info, yaml.MapItem{"flatPath", m.FlatPath}) + info = append(info, yaml.MapItem{Key: "flatPath", Value: m.FlatPath}) } if m.EtagRequired != false { - info = append(info, yaml.MapItem{"etagRequired", m.EtagRequired}) + info = append(info, yaml.MapItem{Key: "etagRequired", Value: m.EtagRequired}) } return info } @@ -2356,7 +2356,7 @@ func (m *Methods) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedMethod StringEnumValues:[] MapType:Method Repeated:true Pattern: Implicit:true Description:} @@ -2367,7 +2367,7 @@ func (m *Methods) ToRawInfo() interface{} { func (m *NamedMethod) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Method StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -2377,7 +2377,7 @@ func (m *NamedMethod) ToRawInfo() interface{} { func (m *NamedParameter) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Parameter StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -2387,7 +2387,7 @@ func (m *NamedParameter) ToRawInfo() interface{} { func (m *NamedResource) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Resource StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -2397,7 +2397,7 @@ func (m *NamedResource) ToRawInfo() interface{} { func (m *NamedSchema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -2407,7 +2407,7 @@ func (m *NamedSchema) ToRawInfo() interface{} { func (m *NamedScope) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Name != "" { - info = append(info, yaml.MapItem{"name", m.Name}) + info = append(info, yaml.MapItem{Key: "name", Value: m.Name}) } // &{Name:value Type:Scope StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:Mapped value} return info @@ -2417,7 +2417,7 @@ func (m *NamedScope) ToRawInfo() interface{} { func (m *Oauth2) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Scopes != nil { - info = append(info, yaml.MapItem{"scopes", m.Scopes.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "scopes", Value: m.Scopes.ToRawInfo()}) } // &{Name:scopes Type:Scopes StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} return info @@ -2427,61 +2427,61 @@ func (m *Oauth2) ToRawInfo() interface{} { func (m *Parameter) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Id != "" { - info = append(info, yaml.MapItem{"id", m.Id}) + info = append(info, yaml.MapItem{Key: "id", Value: m.Id}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Default != "" { - info = append(info, yaml.MapItem{"default", m.Default}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default}) } if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.Minimum != "" { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.Maximum != "" { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if len(m.Enum) != 0 { - info = append(info, yaml.MapItem{"enum", m.Enum}) + info = append(info, yaml.MapItem{Key: "enum", Value: m.Enum}) } if len(m.EnumDescriptions) != 0 { - info = append(info, yaml.MapItem{"enumDescriptions", m.EnumDescriptions}) + info = append(info, yaml.MapItem{Key: "enumDescriptions", Value: m.EnumDescriptions}) } if m.Repeated != false { - info = append(info, yaml.MapItem{"repeated", m.Repeated}) + info = append(info, yaml.MapItem{Key: "repeated", Value: m.Repeated}) } if m.Location != "" { - info = append(info, yaml.MapItem{"location", m.Location}) + info = append(info, yaml.MapItem{Key: "location", Value: m.Location}) } if m.Properties != nil { - info = append(info, yaml.MapItem{"properties", m.Properties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "properties", Value: m.Properties.ToRawInfo()}) } // &{Name:properties Type:Schemas StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.AdditionalProperties != nil { - info = append(info, yaml.MapItem{"additionalProperties", m.AdditionalProperties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "additionalProperties", Value: m.AdditionalProperties.ToRawInfo()}) } // &{Name:additionalProperties Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Annotations != nil { - info = append(info, yaml.MapItem{"annotations", m.Annotations.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "annotations", Value: m.Annotations.ToRawInfo()}) } // &{Name:annotations Type:Annotations StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} return info @@ -2492,7 +2492,7 @@ func (m *Parameters) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedParameter StringEnumValues:[] MapType:Parameter Repeated:true Pattern: Implicit:true Description:} @@ -2503,11 +2503,11 @@ func (m *Parameters) ToRawInfo() interface{} { func (m *Protocols) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Simple != nil { - info = append(info, yaml.MapItem{"simple", m.Simple.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "simple", Value: m.Simple.ToRawInfo()}) } // &{Name:simple Type:Simple StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Resumable != nil { - info = append(info, yaml.MapItem{"resumable", m.Resumable.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "resumable", Value: m.Resumable.ToRawInfo()}) } // &{Name:resumable Type:Resumable StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} return info @@ -2517,10 +2517,10 @@ func (m *Protocols) ToRawInfo() interface{} { func (m *Request) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.ParameterName != "" { - info = append(info, yaml.MapItem{"parameterName", m.ParameterName}) + info = append(info, yaml.MapItem{Key: "parameterName", Value: m.ParameterName}) } return info } @@ -2529,11 +2529,11 @@ func (m *Request) ToRawInfo() interface{} { func (m *Resource) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Methods != nil { - info = append(info, yaml.MapItem{"methods", m.Methods.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "methods", Value: m.Methods.ToRawInfo()}) } // &{Name:methods Type:Methods StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Resources != nil { - info = append(info, yaml.MapItem{"resources", m.Resources.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "resources", Value: m.Resources.ToRawInfo()}) } // &{Name:resources Type:Resources StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} return info @@ -2544,7 +2544,7 @@ func (m *Resources) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedResource StringEnumValues:[] MapType:Resource Repeated:true Pattern: Implicit:true Description:} @@ -2555,7 +2555,7 @@ func (m *Resources) ToRawInfo() interface{} { func (m *Response) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } return info } @@ -2564,10 +2564,10 @@ func (m *Response) ToRawInfo() interface{} { func (m *Resumable) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Multipart != false { - info = append(info, yaml.MapItem{"multipart", m.Multipart}) + info = append(info, yaml.MapItem{Key: "multipart", Value: m.Multipart}) } if m.Path != "" { - info = append(info, yaml.MapItem{"path", m.Path}) + info = append(info, yaml.MapItem{Key: "path", Value: m.Path}) } return info } @@ -2576,65 +2576,65 @@ func (m *Resumable) ToRawInfo() interface{} { func (m *Schema) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Id != "" { - info = append(info, yaml.MapItem{"id", m.Id}) + info = append(info, yaml.MapItem{Key: "id", Value: m.Id}) } if m.Type != "" { - info = append(info, yaml.MapItem{"type", m.Type}) + info = append(info, yaml.MapItem{Key: "type", Value: m.Type}) } if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } if m.Default != "" { - info = append(info, yaml.MapItem{"default", m.Default}) + info = append(info, yaml.MapItem{Key: "default", Value: m.Default}) } if m.Required != false { - info = append(info, yaml.MapItem{"required", m.Required}) + info = append(info, yaml.MapItem{Key: "required", Value: m.Required}) } if m.Format != "" { - info = append(info, yaml.MapItem{"format", m.Format}) + info = append(info, yaml.MapItem{Key: "format", Value: m.Format}) } if m.Pattern != "" { - info = append(info, yaml.MapItem{"pattern", m.Pattern}) + info = append(info, yaml.MapItem{Key: "pattern", Value: m.Pattern}) } if m.Minimum != "" { - info = append(info, yaml.MapItem{"minimum", m.Minimum}) + info = append(info, yaml.MapItem{Key: "minimum", Value: m.Minimum}) } if m.Maximum != "" { - info = append(info, yaml.MapItem{"maximum", m.Maximum}) + info = append(info, yaml.MapItem{Key: "maximum", Value: m.Maximum}) } if len(m.Enum) != 0 { - info = append(info, yaml.MapItem{"enum", m.Enum}) + info = append(info, yaml.MapItem{Key: "enum", Value: m.Enum}) } if len(m.EnumDescriptions) != 0 { - info = append(info, yaml.MapItem{"enumDescriptions", m.EnumDescriptions}) + info = append(info, yaml.MapItem{Key: "enumDescriptions", Value: m.EnumDescriptions}) } if m.Repeated != false { - info = append(info, yaml.MapItem{"repeated", m.Repeated}) + info = append(info, yaml.MapItem{Key: "repeated", Value: m.Repeated}) } if m.Location != "" { - info = append(info, yaml.MapItem{"location", m.Location}) + info = append(info, yaml.MapItem{Key: "location", Value: m.Location}) } if m.Properties != nil { - info = append(info, yaml.MapItem{"properties", m.Properties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "properties", Value: m.Properties.ToRawInfo()}) } // &{Name:properties Type:Schemas StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.AdditionalProperties != nil { - info = append(info, yaml.MapItem{"additionalProperties", m.AdditionalProperties.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "additionalProperties", Value: m.AdditionalProperties.ToRawInfo()}) } // &{Name:additionalProperties Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.Items != nil { - info = append(info, yaml.MapItem{"items", m.Items.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "items", Value: m.Items.ToRawInfo()}) } // &{Name:items Type:Schema StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.XRef != "" { - info = append(info, yaml.MapItem{"$ref", m.XRef}) + info = append(info, yaml.MapItem{Key: "$ref", Value: m.XRef}) } if m.Annotations != nil { - info = append(info, yaml.MapItem{"annotations", m.Annotations.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: "annotations", Value: m.Annotations.ToRawInfo()}) } // &{Name:annotations Type:Annotations StringEnumValues:[] MapType: Repeated:false Pattern: Implicit:false Description:} if m.ReadOnly != false { - info = append(info, yaml.MapItem{"readOnly", m.ReadOnly}) + info = append(info, yaml.MapItem{Key: "readOnly", Value: m.ReadOnly}) } return info } @@ -2644,7 +2644,7 @@ func (m *Schemas) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedSchema StringEnumValues:[] MapType:Schema Repeated:true Pattern: Implicit:true Description:} @@ -2655,7 +2655,7 @@ func (m *Schemas) ToRawInfo() interface{} { func (m *Scope) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Description != "" { - info = append(info, yaml.MapItem{"description", m.Description}) + info = append(info, yaml.MapItem{Key: "description", Value: m.Description}) } return info } @@ -2665,7 +2665,7 @@ func (m *Scopes) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.AdditionalProperties != nil { for _, item := range m.AdditionalProperties { - info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()}) + info = append(info, yaml.MapItem{Key: item.Name, Value: item.Value.ToRawInfo()}) } } // &{Name:additionalProperties Type:NamedScope StringEnumValues:[] MapType:Scope Repeated:true Pattern: Implicit:true Description:} @@ -2676,10 +2676,10 @@ func (m *Scopes) ToRawInfo() interface{} { func (m *Simple) ToRawInfo() interface{} { info := yaml.MapSlice{} if m.Multipart != false { - info = append(info, yaml.MapItem{"multipart", m.Multipart}) + info = append(info, yaml.MapItem{Key: "multipart", Value: m.Multipart}) } if m.Path != "" { - info = append(info, yaml.MapItem{"path", m.Path}) + info = append(info, yaml.MapItem{Key: "path", Value: m.Path}) } return info } diff --git a/vendor/github.com/googleapis/gnostic/discovery/discovery.pb.go b/vendor/github.com/googleapis/gnostic/discovery/discovery.pb.go index bbfe676ca..b17e68627 100644 --- a/vendor/github.com/googleapis/gnostic/discovery/discovery.pb.go +++ b/vendor/github.com/googleapis/gnostic/discovery/discovery.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-go. +// Code generated by protoc-gen-go. DO NOT EDIT. // source: discovery/discovery.proto -// DO NOT EDIT! /* Package discovery_v1 is a generated protocol buffer package. @@ -1307,7 +1306,7 @@ func init() { proto.RegisterFile("discovery/discovery.proto", fileDescriptor0) } var fileDescriptor0 = []byte{ // 1683 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x09, 0x6e, 0x88, 0x02, 0xff, 0xa4, 0x58, 0x5f, 0x53, 0xdc, 0xc8, + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x58, 0x5f, 0x53, 0xdc, 0xc8, 0x11, 0xaf, 0x85, 0xfd, 0xdb, 0xbb, 0x0b, 0xcb, 0xf0, 0x4f, 0x60, 0x9c, 0x23, 0xba, 0xba, 0x04, 0x72, 0x77, 0x50, 0xc6, 0x39, 0x57, 0x2a, 0x4e, 0x1e, 0x70, 0xe0, 0xc1, 0x95, 0x60, 0xb0, 0x28, 0x27, 0x95, 0x87, 0x94, 0x6a, 0x90, 0x66, 0x59, 0x15, 0x92, 0x46, 0x19, 0x8d, 0x30, 0xeb, 0x8f, diff --git a/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go b/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go index 7c6b91496..749ff7841 100644 --- a/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go +++ b/vendor/github.com/googleapis/gnostic/extensions/extension.pb.go @@ -1,6 +1,5 @@ -// Code generated by protoc-gen-go. +// Code generated by protoc-gen-go. DO NOT EDIT. // source: extension.proto -// DO NOT EDIT! /* Package openapiextension_v1 is a generated protocol buffer package. @@ -78,7 +77,7 @@ func (m *Version) GetSuffix() string { // An encoded Request is written to the ExtensionHandler's stdin. type ExtensionHandlerRequest struct { // The OpenAPI descriptions that were explicitly listed on the command line. - // The specifications will appear in the order they are specified to openapic. + // The specifications will appear in the order they are specified to gnostic. Wrapper *Wrapper `protobuf:"bytes,1,opt,name=wrapper" json:"wrapper,omitempty"` // The version number of openapi compiler. CompilerVersion *Version `protobuf:"bytes,3,opt,name=compiler_version,json=compilerVersion" json:"compiler_version,omitempty"` @@ -192,28 +191,28 @@ func init() { func init() { proto.RegisterFile("extension.proto", fileDescriptor0) } var fileDescriptor0 = []byte{ - // 355 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x74, 0x91, 0x4d, 0x4b, 0xf3, 0x40, - 0x1c, 0xc4, 0x49, 0xdf, 0xf2, 0x64, 0x1f, 0xb4, 0xb2, 0x16, 0x8d, 0xe2, 0xa1, 0x04, 0x84, 0x22, - 0xb8, 0xa5, 0x0a, 0xde, 0x5b, 0x28, 0xea, 0xc5, 0x96, 0x3d, 0xd4, 0x9b, 0x65, 0x9b, 0xfe, 0xdb, - 0x46, 0x92, 0xdd, 0x75, 0xf3, 0x62, 0xfb, 0x55, 0x3c, 0xfa, 0x49, 0x25, 0xbb, 0xd9, 0x7a, 0x50, - 0x6f, 0x99, 0x1f, 0x93, 0xfc, 0x67, 0x26, 0xa8, 0x0d, 0xdb, 0x0c, 0x78, 0x1a, 0x09, 0x4e, 0xa4, - 0x12, 0x99, 0xc0, 0xc7, 0x42, 0x02, 0x67, 0x32, 0xfa, 0xe6, 0xc5, 0xe0, 0xfc, 0x6c, 0x2d, 0xc4, - 0x3a, 0x86, 0xbe, 0xb6, 0x2c, 0xf2, 0x55, 0x9f, 0xf1, 0x9d, 0xf1, 0x07, 0x21, 0x72, 0x67, 0xa0, - 0x4a, 0x23, 0xee, 0xa0, 0x66, 0xc2, 0x5e, 0x85, 0xf2, 0x9d, 0xae, 0xd3, 0x6b, 0x52, 0x23, 0x34, - 0x8d, 0xb8, 0x50, 0x7e, 0xad, 0xa2, 0xa5, 0x28, 0xa9, 0x64, 0x59, 0xb8, 0xf1, 0xeb, 0x86, 0x6a, - 0x81, 0x4f, 0x50, 0x2b, 0xcd, 0x57, 0xab, 0x68, 0xeb, 0x37, 0xba, 0x4e, 0xcf, 0xa3, 0x95, 0x0a, - 0x3e, 0x1c, 0x74, 0x3a, 0xb6, 0x81, 0x1e, 0x18, 0x5f, 0xc6, 0xa0, 0x28, 0xbc, 0xe5, 0x90, 0x66, - 0xf8, 0x0e, 0xb9, 0xef, 0x8a, 0x49, 0x09, 0xe6, 0xee, 0xff, 0x9b, 0x0b, 0xf2, 0x4b, 0x05, 0xf2, - 0x6c, 0x3c, 0xd4, 0x9a, 0xf1, 0x3d, 0x3a, 0x0a, 0x45, 0x22, 0xa3, 0x18, 0xd4, 0xbc, 0x30, 0x0d, - 0x74, 0x98, 0xbf, 0x3e, 0x50, 0xb5, 0xa4, 0x6d, 0xfb, 0x56, 0x05, 0x82, 0x02, 0xf9, 0x3f, 0xb3, - 0xa5, 0x52, 0xf0, 0x14, 0xb0, 0x8f, 0xdc, 0x8d, 0x46, 0x4b, 0x1d, 0xee, 0x1f, 0xb5, 0xb2, 0x1c, - 0x00, 0x94, 0xd2, 0xb3, 0xd4, 0x7b, 0x1e, 0x35, 0x02, 0x5f, 0xa1, 0x66, 0xc1, 0xe2, 0x1c, 0xaa, - 0x24, 0x1d, 0x62, 0x86, 0x27, 0x76, 0x78, 0x32, 0xe4, 0x3b, 0x6a, 0x2c, 0xc1, 0x0b, 0x72, 0xab, - 0x52, 0xe5, 0x19, 0x5b, 0xc1, 0xd1, 0xc3, 0x59, 0x89, 0x2f, 0xd1, 0xe1, 0xbe, 0xc5, 0x9c, 0xb3, - 0x04, 0xf4, 0x6f, 0xf0, 0xe8, 0xc1, 0x9e, 0x3e, 0xb1, 0x04, 0x30, 0x46, 0x8d, 0x1d, 0x4b, 0x62, - 0x7d, 0xd6, 0xa3, 0xfa, 0x79, 0x74, 0x8d, 0xda, 0x42, 0xad, 0xed, 0x16, 0x21, 0x29, 0x06, 0x23, - 0x3c, 0x91, 0xc0, 0x87, 0xd3, 0xc7, 0x7d, 0xdf, 0xd9, 0x60, 0xea, 0x7c, 0xd6, 0xea, 0x93, 0xe1, - 0x78, 0xd1, 0xd2, 0x19, 0x6f, 0xbf, 0x02, 0x00, 0x00, 0xff, 0xff, 0xfc, 0x56, 0x40, 0x4d, 0x52, - 0x02, 0x00, 0x00, + // 357 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x74, 0x91, 0x4d, 0x4b, 0xc3, 0x40, + 0x18, 0x84, 0x49, 0xbf, 0x62, 0x56, 0x6c, 0x65, 0x2d, 0x1a, 0xc5, 0x43, 0x09, 0x08, 0x45, 0x64, + 0x4b, 0x15, 0xbc, 0xb7, 0x50, 0xd4, 0x8b, 0x2d, 0x7b, 0xa8, 0x37, 0xcb, 0x36, 0x7d, 0x9b, 0x46, + 0x92, 0xdd, 0x75, 0xf3, 0x61, 0xfb, 0x57, 0x3c, 0xfa, 0x4b, 0x25, 0xbb, 0x49, 0x3d, 0xa8, 0xb7, + 0xcc, 0xc3, 0x24, 0xef, 0xcc, 0x04, 0x75, 0x60, 0x9b, 0x02, 0x4f, 0x42, 0xc1, 0x89, 0x54, 0x22, + 0x15, 0xf8, 0x44, 0x48, 0xe0, 0x4c, 0x86, 0x3f, 0x3c, 0x1f, 0x5e, 0x9c, 0x07, 0x42, 0x04, 0x11, + 0x0c, 0xb4, 0x65, 0x99, 0xad, 0x07, 0x8c, 0xef, 0x8c, 0xdf, 0xf3, 0x91, 0x3d, 0x07, 0x55, 0x18, + 0x71, 0x17, 0x35, 0x63, 0xf6, 0x26, 0x94, 0x6b, 0xf5, 0xac, 0x7e, 0x93, 0x1a, 0xa1, 0x69, 0xc8, + 0x85, 0x72, 0x6b, 0x25, 0x2d, 0x44, 0x41, 0x25, 0x4b, 0xfd, 0x8d, 0x5b, 0x37, 0x54, 0x0b, 0x7c, + 0x8a, 0x5a, 0x49, 0xb6, 0x5e, 0x87, 0x5b, 0xb7, 0xd1, 0xb3, 0xfa, 0x0e, 0x2d, 0x95, 0xf7, 0x69, + 0xa1, 0xb3, 0x49, 0x15, 0xe8, 0x91, 0xf1, 0x55, 0x04, 0x8a, 0xc2, 0x7b, 0x06, 0x49, 0x8a, 0xef, + 0x91, 0xfd, 0xa1, 0x98, 0x94, 0x60, 0xee, 0x1e, 0xde, 0x5e, 0x92, 0x3f, 0x2a, 0x90, 0x17, 0xe3, + 0xa1, 0x95, 0x19, 0x3f, 0xa0, 0x63, 0x5f, 0xc4, 0x32, 0x8c, 0x40, 0x2d, 0x72, 0xd3, 0x40, 0x87, + 0xf9, 0xef, 0x03, 0x65, 0x4b, 0xda, 0xa9, 0xde, 0x2a, 0x81, 0x97, 0x23, 0xf7, 0x77, 0xb6, 0x44, + 0x0a, 0x9e, 0x00, 0x76, 0x91, 0xbd, 0xd1, 0x68, 0xa5, 0xc3, 0x1d, 0xd0, 0x4a, 0x16, 0x03, 0x80, + 0x52, 0x7a, 0x96, 0x7a, 0xdf, 0xa1, 0x46, 0xe0, 0x6b, 0xd4, 0xcc, 0x59, 0x94, 0x41, 0x99, 0xa4, + 0x4b, 0xcc, 0xf0, 0xa4, 0x1a, 0x9e, 0x8c, 0xf8, 0x8e, 0x1a, 0x8b, 0xf7, 0x8a, 0xec, 0xb2, 0x54, + 0x71, 0xa6, 0xaa, 0x60, 0xe9, 0xe1, 0x2a, 0x89, 0xaf, 0x50, 0x7b, 0xdf, 0x62, 0xc1, 0x59, 0x0c, + 0xfa, 0x37, 0x38, 0xf4, 0x68, 0x4f, 0x9f, 0x59, 0x0c, 0x18, 0xa3, 0xc6, 0x8e, 0xc5, 0x91, 0x3e, + 0xeb, 0x50, 0xfd, 0x3c, 0xbe, 0x41, 0x6d, 0xa1, 0x02, 0x12, 0x70, 0x91, 0xa4, 0xa1, 0x4f, 0xf2, + 0xe1, 0x18, 0x4f, 0x25, 0xf0, 0xd1, 0xec, 0x69, 0x5f, 0x77, 0x3e, 0x9c, 0x59, 0x5f, 0xb5, 0xfa, + 0x74, 0x34, 0x59, 0xb6, 0x74, 0xc4, 0xbb, 0xef, 0x00, 0x00, 0x00, 0xff, 0xff, 0x84, 0x5c, 0x6b, + 0x80, 0x51, 0x02, 0x00, 0x00, } diff --git a/vendor/github.com/googleapis/gnostic/extensions/extension.proto b/vendor/github.com/googleapis/gnostic/extensions/extension.proto index 806760a13..04856f913 100644 --- a/vendor/github.com/googleapis/gnostic/extensions/extension.proto +++ b/vendor/github.com/googleapis/gnostic/extensions/extension.proto @@ -29,7 +29,7 @@ option java_multiple_files = true; option java_outer_classname = "OpenAPIExtensionV1"; // The Java package name must be proto package name with proper prefix. -option java_package = "org.openapic.v1"; +option java_package = "org.gnostic.v1"; // A reasonable prefix for the Objective-C symbols generated from the package. // It should at a minimum be 3 characters long, all uppercase, and convention @@ -53,7 +53,7 @@ message Version { message ExtensionHandlerRequest { // The OpenAPI descriptions that were explicitly listed on the command line. - // The specifications will appear in the order they are specified to openapic. + // The specifications will appear in the order they are specified to gnostic. Wrapper wrapper = 1; // The version number of openapi compiler. diff --git a/vendor/github.com/googleapis/gnostic/generate-gnostic/generate-compiler.go b/vendor/github.com/googleapis/gnostic/generate-gnostic/generate-compiler.go index 5cc48ec0f..db988213b 100644 --- a/vendor/github.com/googleapis/gnostic/generate-gnostic/generate-compiler.go +++ b/vendor/github.com/googleapis/gnostic/generate-gnostic/generate-compiler.go @@ -804,44 +804,44 @@ func (domain *Domain) generateToRawInfoMethodForType(code *printer.Code, typeNam propertyName := propertyModel.Name if !propertyModel.Repeated { code.Print("if m.%s != \"\" {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } else { code.Print("if len(m.%s) != 0 {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } case "bool": propertyName := propertyModel.Name if !propertyModel.Repeated { code.Print("if m.%s != false {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } else { code.Print("if len(m.%s) != 0 {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } case "int": propertyName := propertyModel.Name if !propertyModel.Repeated { code.Print("if m.%s != 0 {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } else { code.Print("if len(m.%s) != 0 {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } case "float": propertyName := propertyModel.Name if !propertyModel.Repeated { code.Print("if m.%s != 0.0 {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } else { code.Print("if len(m.%s) != 0 {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s})", propertyName, propertyModel.FieldName()) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s})", propertyName, propertyModel.FieldName()) code.Print("}") } default: @@ -852,9 +852,9 @@ func (domain *Domain) generateToRawInfoMethodForType(code *printer.Code, typeNam code.Print("if m.%s != nil {", propertyModel.FieldName()) if propertyModel.Type == "TypeItem" { code.Print("if len(m.Type.Value) == 1 {") - code.Print("info = append(info, yaml.MapItem{\"type\", m.Type.Value[0]})") + code.Print("info = append(info, yaml.MapItem{Key:\"type\", Value:m.Type.Value[0]})") code.Print("} else {") - code.Print("info = append(info, yaml.MapItem{\"type\", m.Type.Value})") + code.Print("info = append(info, yaml.MapItem{Key:\"type\", Value:m.Type.Value})") code.Print("}") } else if propertyModel.Type == "ItemsItem" { code.Print("items := make([]interface{}, 0)") @@ -865,9 +865,9 @@ func (domain *Domain) generateToRawInfoMethodForType(code *printer.Code, typeNam } code.Print(" items = append(items, item.ToRawInfo())") code.Print("}") - code.Print("info = append(info, yaml.MapItem{\"items\", items[0]})") + code.Print("info = append(info, yaml.MapItem{Key:\"items\", Value:items[0]})") } else { - code.Print("info = append(info, yaml.MapItem{\"%s\", m.%s.ToRawInfo()})", + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:m.%s.ToRawInfo()})", propertyName, propertyModel.FieldName()) } code.Print("}") @@ -877,7 +877,7 @@ func (domain *Domain) generateToRawInfoMethodForType(code *printer.Code, typeNam } else if propertyModel.MapType != "" { code.Print("if m.%s != nil {", propertyModel.FieldName()) code.Print("for _, item := range m.%s {", propertyModel.FieldName()) - code.Print("info = append(info, yaml.MapItem{item.Name, item.Value.ToRawInfo()})") + code.Print("info = append(info, yaml.MapItem{Key:item.Name, Value:item.Value.ToRawInfo()})") code.Print("}") code.Print("}") code.Print("// %+v", propertyModel) @@ -887,7 +887,7 @@ func (domain *Domain) generateToRawInfoMethodForType(code *printer.Code, typeNam code.Print("for _, item := range m.%s {", propertyModel.FieldName()) code.Print("items = append(items, item.ToRawInfo())") code.Print("}") - code.Print("info = append(info, yaml.MapItem{\"%s\", items})", propertyName) + code.Print("info = append(info, yaml.MapItem{Key:\"%s\", Value:items})", propertyName) code.Print("}") code.Print("// %+v", propertyModel) } diff --git a/vendor/github.com/googleapis/gnostic/gnostic.go b/vendor/github.com/googleapis/gnostic/gnostic.go index 80e050503..960041521 100644 --- a/vendor/github.com/googleapis/gnostic/gnostic.go +++ b/vendor/github.com/googleapis/gnostic/gnostic.go @@ -41,6 +41,7 @@ import ( "path/filepath" "regexp" "strings" + "time" "github.com/golang/protobuf/proto" "github.com/googleapis/gnostic/OpenAPIv2" @@ -92,7 +93,7 @@ type pluginCall struct { } // Invokes a plugin. -func (p *pluginCall) perform(document proto.Message, sourceFormat int, sourceName string) error { +func (p *pluginCall) perform(document proto.Message, sourceFormat int, sourceName string, timePlugins bool) ([]*plugins.Message, error) { if p.Name != "" { request := &plugins.Request{} @@ -112,7 +113,7 @@ func (p *pluginCall) perform(document proto.Message, sourceFormat int, sourceNam // invocationRegex := regexp.MustCompile(`^([\w-_\/\.]+=[\w-_\/\.]+(,[\w-_\/\.]+=[\w-_\/\.]+)*:)?[^,:=]+$`) if !invocationRegex.Match([]byte(p.Invocation)) { - return fmt.Errorf("Invalid invocation of %s: %s", executableName, invocation) + return nil, fmt.Errorf("Invalid invocation of %s: %s", executableName, invocation) } invocationParts := strings.Split(p.Invocation, ":") @@ -145,11 +146,21 @@ func (p *pluginCall) perform(document proto.Message, sourceFormat int, sourceNam request.SourceName = sourceName switch sourceFormat { case SourceFormatOpenAPI2: - request.Openapi2 = document.(*openapi_v2.Document) - request.Surface, _ = surface.NewModelFromOpenAPI2(request.Openapi2) + request.AddModel("openapi.v2.Document", document) + // include experimental API surface model + surfaceModel, err := surface.NewModelFromOpenAPI2(document.(*openapi_v2.Document)) + if err == nil { + request.AddModel("surface.v1.Model", surfaceModel) + } case SourceFormatOpenAPI3: - request.Openapi3 = document.(*openapi_v3.Document) - request.Surface, _ = surface.NewModelFromOpenAPI3(request.Openapi3) + request.AddModel("openapi.v3.Document", document) + // include experimental API surface model + surfaceModel, err := surface.NewModelFromOpenAPI3(document.(*openapi_v3.Document)) + if err == nil { + request.AddModel("surface.v1.Model", surfaceModel) + } + case SourceFormatDiscovery: + request.AddModel("discovery.v1.Document", document) default: } @@ -158,19 +169,27 @@ func (p *pluginCall) perform(document proto.Message, sourceFormat int, sourceNam cmd := exec.Command(executableName, "-plugin") cmd.Stdin = bytes.NewReader(requestBytes) cmd.Stderr = os.Stderr + pluginStartTime := time.Now() output, err := cmd.Output() + pluginElapsedTime := time.Since(pluginStartTime) + if timePlugins { + fmt.Printf("> %s (%s)\n", executableName, pluginElapsedTime) + } if err != nil { - return err + return nil, err } response := &plugins.Response{} err = proto.Unmarshal(output, response) if err != nil { - return err + // Gnostic expects plugins to only write the + // response message to stdout. Be sure that + // any logging messages are written to stderr only. + return nil, errors.New("Invalid plugin response (plugins must write log messages to stderr, not stdout).") } - plugins.HandleResponse(response, outputLocation) + return response.Messages, nil } - return nil + return nil, nil } func isFile(path string) bool { @@ -233,10 +252,12 @@ type Gnostic struct { yamlOutputPath string jsonOutputPath string errorOutputPath string + messageOutputPath string resolveReferences bool pluginCalls []*pluginCall extensionHandlers []compiler.ExtensionHandler sourceFormat int + timePlugins bool } // Initialize a structure to store global application state. @@ -244,20 +265,27 @@ func newGnostic() *Gnostic { g := &Gnostic{} // Option fields initialize to their default values. g.usage = ` -Usage: gnostic OPENAPI_SOURCE [OPTIONS] - OPENAPI_SOURCE is the filename or URL of an OpenAPI description to read. +Usage: gnostic SOURCE [OPTIONS] + SOURCE is the filename or URL of an API description. Options: --pb-out=PATH Write a binary proto to the specified location. --text-out=PATH Write a text proto to the specified location. --json-out=PATH Write a json API description to the specified location. --yaml-out=PATH Write a yaml API description to the specified location. --errors-out=PATH Write compilation errors to the specified location. - --PLUGIN-out=PATH Run the plugin named gnostic_PLUGIN and write results + --messages-out=PATH Write messages generated by plugins to the specified + location. Messages from all plugin invocations are + written to a single common file. + --PLUGIN-out=PATH Run the plugin named gnostic-PLUGIN and write results to the specified location. + --PLUGIN Run the plugin named gnostic-PLUGIN but don't write any + results. Used for plugins that return messages only. + PLUGIN must not match any other gnostic option. --x-EXTENSION Use the extension named gnostic-x-EXTENSION to process OpenAPI specification extensions. --resolve-refs Explicitly resolve $ref references. This could have problems with recursive definitions. + --time-plugins Report plugin runtimes. ` // Initialize internal structures. g.pluginCalls = make([]*pluginCall, 0) @@ -292,6 +320,8 @@ func (g *Gnostic) readOptions() { g.yamlOutputPath = invocation case "errors": g.errorOutputPath = invocation + case "messages": + g.messageOutputPath = invocation default: p := &pluginCall{Name: pluginName, Invocation: invocation} g.pluginCalls = append(g.pluginCalls, p) @@ -302,6 +332,13 @@ func (g *Gnostic) readOptions() { g.extensionHandlers = append(g.extensionHandlers, extensionHandler) } else if arg == "--resolve-refs" { g.resolveReferences = true + } else if arg == "--time-plugins" { + g.timePlugins = true + } else if arg[0] == '-' && arg[1] == '-' { + // try letting the option specify a plugin with no output files (or unwanted output files) + // this is useful for calling plugins like linters that only return messages + p := &pluginCall{Name: arg[2:len(arg)], Invocation: "!"} + g.pluginCalls = append(g.pluginCalls, p) } else if arg[0] == '-' { fmt.Fprintf(os.Stderr, "Unknown option: %s.\n%s\n", arg, g.usage) os.Exit(-1) @@ -467,6 +504,17 @@ func (g *Gnostic) writeJSONYAMLOutput(message proto.Message) { } } +// Write messages. +func (g *Gnostic) writeMessagesOutput(message proto.Message) { + protoBytes, err := proto.Marshal(message) + if err != nil { + writeFile(g.messageOutputPath, g.errorBytes(err), g.sourceName, "errors") + defer os.Exit(-1) + } else { + writeFile(g.messageOutputPath, protoBytes, g.sourceName, "messages.pb") + } +} + // Perform all actions specified in the command-line options. func (g *Gnostic) performActions(message proto.Message) (err error) { // Optionally resolve internal references. @@ -490,17 +538,29 @@ func (g *Gnostic) performActions(message proto.Message) (err error) { if g.textOutputPath != "" { g.writeTextOutput(message) } - // Optionaly write document in yaml and/or json formats. + // Optionally write document in yaml and/or json formats. if g.yamlOutputPath != "" || g.jsonOutputPath != "" { g.writeJSONYAMLOutput(message) } // Call all specified plugins. + messages := make([]*plugins.Message, 0) for _, p := range g.pluginCalls { - err := p.perform(message, g.sourceFormat, g.sourceName) + pluginMessages, err := p.perform(message, g.sourceFormat, g.sourceName, g.timePlugins) if err != nil { writeFile(g.errorOutputPath, g.errorBytes(err), g.sourceName, "errors") defer os.Exit(-1) // run all plugins, even when some have errors } + messages = append(messages, pluginMessages...) + } + if g.messageOutputPath != "" { + g.writeMessagesOutput(&plugins.Messages{Messages: messages}) + } else { + // Print any messages from the plugins + if len(messages) > 0 { + for _, message := range messages { + fmt.Printf("%+v\n", message) + } + } } return nil } diff --git a/vendor/github.com/googleapis/gnostic/jsonschema/writer.go b/vendor/github.com/googleapis/gnostic/jsonschema/writer.go index c82f3064b..3187c455e 100644 --- a/vendor/github.com/googleapis/gnostic/jsonschema/writer.go +++ b/vendor/github.com/googleapis/gnostic/jsonschema/writer.go @@ -50,7 +50,7 @@ func renderMap(info interface{}, indent string) (result string) { case []string: result += renderStringArray(value, innerIndent) default: - result += fmt.Sprintf("???MapItem(%+v, %T)", value, value) + result += fmt.Sprintf("???MapItem(Key:%+v, Value:%T)", value, value) } if i < len(pairs)-1 { result += "," @@ -106,7 +106,7 @@ func renderStringArray(array []string, indent string) (result string) { return result } -func render(info yaml.MapSlice) string { +func Render(info yaml.MapSlice) string { return renderMap(info, "") + "\n" } @@ -223,106 +223,106 @@ func schemaArrayValue(array *[]*Schema) []interface{} { func (schema *Schema) jsonValue() yaml.MapSlice { m := yaml.MapSlice{} if schema.Title != nil { - m = append(m, yaml.MapItem{"title", *schema.Title}) + m = append(m, yaml.MapItem{Key: "title", Value: *schema.Title}) } if schema.ID != nil { - m = append(m, yaml.MapItem{"id", *schema.ID}) + m = append(m, yaml.MapItem{Key: "id", Value: *schema.ID}) } if schema.Schema != nil { - m = append(m, yaml.MapItem{"$schema", *schema.Schema}) + m = append(m, yaml.MapItem{Key: "$schema", Value: *schema.Schema}) } if schema.Type != nil { - m = append(m, yaml.MapItem{"type", schema.Type.jsonValue()}) + m = append(m, yaml.MapItem{Key: "type", Value: schema.Type.jsonValue()}) } if schema.Items != nil { - m = append(m, yaml.MapItem{"items", schema.Items.jsonValue()}) + m = append(m, yaml.MapItem{Key: "items", Value: schema.Items.jsonValue()}) } if schema.Description != nil { - m = append(m, yaml.MapItem{"description", *schema.Description}) + m = append(m, yaml.MapItem{Key: "description", Value: *schema.Description}) } if schema.Required != nil { - m = append(m, yaml.MapItem{"required", *schema.Required}) + m = append(m, yaml.MapItem{Key: "required", Value: *schema.Required}) } if schema.AdditionalProperties != nil { - m = append(m, yaml.MapItem{"additionalProperties", schema.AdditionalProperties.jsonValue()}) + m = append(m, yaml.MapItem{Key: "additionalProperties", Value: schema.AdditionalProperties.jsonValue()}) } if schema.PatternProperties != nil { - m = append(m, yaml.MapItem{"patternProperties", namedSchemaArrayValue(schema.PatternProperties)}) + m = append(m, yaml.MapItem{Key: "patternProperties", Value: namedSchemaArrayValue(schema.PatternProperties)}) } if schema.Properties != nil { - m = append(m, yaml.MapItem{"properties", namedSchemaArrayValue(schema.Properties)}) + m = append(m, yaml.MapItem{Key: "properties", Value: namedSchemaArrayValue(schema.Properties)}) } if schema.Dependencies != nil { - m = append(m, yaml.MapItem{"dependencies", namedSchemaOrStringArrayValue(schema.Dependencies)}) + m = append(m, yaml.MapItem{Key: "dependencies", Value: namedSchemaOrStringArrayValue(schema.Dependencies)}) } if schema.Ref != nil { - m = append(m, yaml.MapItem{"$ref", *schema.Ref}) + m = append(m, yaml.MapItem{Key: "$ref", Value: *schema.Ref}) } if schema.MultipleOf != nil { - m = append(m, yaml.MapItem{"multipleOf", schema.MultipleOf.jsonValue()}) + m = append(m, yaml.MapItem{Key: "multipleOf", Value: schema.MultipleOf.jsonValue()}) } if schema.Maximum != nil { - m = append(m, yaml.MapItem{"maximum", schema.Maximum.jsonValue()}) + m = append(m, yaml.MapItem{Key: "maximum", Value: schema.Maximum.jsonValue()}) } if schema.ExclusiveMaximum != nil { - m = append(m, yaml.MapItem{"exclusiveMaximum", *schema.ExclusiveMaximum}) + m = append(m, yaml.MapItem{Key: "exclusiveMaximum", Value: schema.ExclusiveMaximum}) } if schema.Minimum != nil { - m = append(m, yaml.MapItem{"minimum", schema.Minimum.jsonValue()}) + m = append(m, yaml.MapItem{Key: "minimum", Value: schema.Minimum.jsonValue()}) } if schema.ExclusiveMinimum != nil { - m = append(m, yaml.MapItem{"exclusiveMinimum", *schema.ExclusiveMinimum}) + m = append(m, yaml.MapItem{Key: "exclusiveMinimum", Value: schema.ExclusiveMinimum}) } if schema.MaxLength != nil { - m = append(m, yaml.MapItem{"maxLength", *schema.MaxLength}) + m = append(m, yaml.MapItem{Key: "maxLength", Value: *schema.MaxLength}) } if schema.MinLength != nil { - m = append(m, yaml.MapItem{"minLength", *schema.MinLength}) + m = append(m, yaml.MapItem{Key: "minLength", Value: *schema.MinLength}) } if schema.Pattern != nil { - m = append(m, yaml.MapItem{"pattern", *schema.Pattern}) + m = append(m, yaml.MapItem{Key: "pattern", Value: *schema.Pattern}) } if schema.AdditionalItems != nil { - m = append(m, yaml.MapItem{"additionalItems", schema.AdditionalItems.jsonValue()}) + m = append(m, yaml.MapItem{Key: "additionalItems", Value: schema.AdditionalItems.jsonValue()}) } if schema.MaxItems != nil { - m = append(m, yaml.MapItem{"maxItems", *schema.MaxItems}) + m = append(m, yaml.MapItem{Key: "maxItems", Value: *schema.MaxItems}) } if schema.MinItems != nil { - m = append(m, yaml.MapItem{"minItems", *schema.MinItems}) + m = append(m, yaml.MapItem{Key: "minItems", Value: *schema.MinItems}) } if schema.UniqueItems != nil { - m = append(m, yaml.MapItem{"uniqueItems", *schema.UniqueItems}) + m = append(m, yaml.MapItem{Key: "uniqueItems", Value: *schema.UniqueItems}) } if schema.MaxProperties != nil { - m = append(m, yaml.MapItem{"maxProperties", *schema.MaxProperties}) + m = append(m, yaml.MapItem{Key: "maxProperties", Value: *schema.MaxProperties}) } if schema.MinProperties != nil { - m = append(m, yaml.MapItem{"minProperties", *schema.MinProperties}) + m = append(m, yaml.MapItem{Key: "minProperties", Value: *schema.MinProperties}) } if schema.Enumeration != nil { - m = append(m, yaml.MapItem{"enum", schemaEnumArrayValue(schema.Enumeration)}) + m = append(m, yaml.MapItem{Key: "enum", Value: schemaEnumArrayValue(schema.Enumeration)}) } if schema.AllOf != nil { - m = append(m, yaml.MapItem{"allOf", schemaArrayValue(schema.AllOf)}) + m = append(m, yaml.MapItem{Key: "allOf", Value: schemaArrayValue(schema.AllOf)}) } if schema.AnyOf != nil { - m = append(m, yaml.MapItem{"anyOf", schemaArrayValue(schema.AnyOf)}) + m = append(m, yaml.MapItem{Key: "anyOf", Value: schemaArrayValue(schema.AnyOf)}) } if schema.OneOf != nil { - m = append(m, yaml.MapItem{"oneOf", schemaArrayValue(schema.OneOf)}) + m = append(m, yaml.MapItem{Key: "oneOf", Value: schemaArrayValue(schema.OneOf)}) } if schema.Not != nil { - m = append(m, yaml.MapItem{"not", schema.Not.jsonValue()}) + m = append(m, yaml.MapItem{Key: "not", Value: schema.Not.jsonValue()}) } if schema.Definitions != nil { - m = append(m, yaml.MapItem{"definitions", namedSchemaArrayValue(schema.Definitions)}) + m = append(m, yaml.MapItem{Key: "definitions", Value: namedSchemaArrayValue(schema.Definitions)}) } if schema.Default != nil { - m = append(m, yaml.MapItem{"default", *schema.Default}) + m = append(m, yaml.MapItem{Key: "default", Value: *schema.Default}) } if schema.Format != nil { - m = append(m, yaml.MapItem{"format", *schema.Format}) + m = append(m, yaml.MapItem{Key: "format", Value: *schema.Format}) } return m } @@ -330,5 +330,5 @@ func (schema *Schema) jsonValue() yaml.MapSlice { // JSONString returns a json representation of a schema. func (schema *Schema) JSONString() string { info := schema.jsonValue() - return render(info) + return Render(info) } diff --git a/vendor/github.com/googleapis/gnostic/linters/README.md b/vendor/github.com/googleapis/gnostic/linters/README.md new file mode 100644 index 000000000..39813f3a9 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/README.md @@ -0,0 +1,20 @@ +# Linters + +This directory contains linters that can be used to check Gnostic models. + +Linters are plugins that generate no files but instead return messages +in their responses. Each message can include a level, an identifier, text, +and a key path in an API description associated with that message. +Messages are collected by gnostic and written to a common output file, +allowing multiple linter plugins to be invoked in a single gnostic +run. + +The following invocation runs the `gnostic-lint-paths` and +`gnostic-lint-descriptions` plugins and writes their messages +to a file named `lint.pb`. + +``` +% gnostic examples/v2.0/yaml/petstore.yaml --lint-paths --lint-descriptions --messages-out=lint.pb +``` + +Message files can be displayed using the `report-messages` tool in the `apps` directory. diff --git a/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/README.md b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/README.md new file mode 100644 index 000000000..6731400ed --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/README.md @@ -0,0 +1,9 @@ +# gnostic-lint-descriptions + +This directory contains a `gnostic` plugin that analyzes an OpenAPI +description for factors that might influence code generation and other +API automation. + +The plugin can be invoked like this: + + gnostic bookstore.json --lint-descriptions diff --git a/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/linter_v2.go b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/linter_v2.go new file mode 100644 index 000000000..721262cef --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/linter_v2.go @@ -0,0 +1,180 @@ +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package main + +import ( + openapi "github.com/googleapis/gnostic/OpenAPIv2" + plugins "github.com/googleapis/gnostic/plugins" +) + +// DocumentLinter contains information collected about an API description. +type DocumentLinterV2 struct { + document *openapi.Document `json:"-"` +} + +func (d *DocumentLinterV2) Run() []*plugins.Message { + return d.analyzeDocument(d.document) +} + +// NewDocumentLinter builds a new DocumentLinter object. +func NewDocumentLinterV2(document *openapi.Document) *DocumentLinterV2 { + return &DocumentLinterV2{document: document} +} + +// Analyze an OpenAPI description. +func (s *DocumentLinterV2) analyzeDocument(document *openapi.Document) []*plugins.Message { + messages := make([]*plugins.Message, 0, 0) + for _, pair := range document.Paths.Path { + path := pair.Value + if path.Get != nil { + messages = append(messages, s.analyzeOperation([]string{"paths", pair.Name, "get"}, path.Get)...) + } + if path.Post != nil { + messages = append(messages, s.analyzeOperation([]string{"paths", pair.Name, "post"}, path.Post)...) + } + if path.Put != nil { + messages = append(messages, s.analyzeOperation([]string{"paths", pair.Name, "put"}, path.Put)...) + } + if path.Delete != nil { + messages = append(messages, s.analyzeOperation([]string{"paths", pair.Name, "delete"}, path.Delete)...) + } + } + if document.Definitions != nil { + for _, pair := range document.Definitions.AdditionalProperties { + definition := pair.Value + messages = append(messages, s.analyzeDefinition([]string{"definitions", pair.Name}, definition)...) + } + } + return messages +} + +func (s *DocumentLinterV2) analyzeOperation(keys []string, operation *openapi.Operation) []*plugins.Message { + messages := make([]*plugins.Message, 0) + + if operation.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Operation has no description.", + Keys: keys}) + } + for _, parameter := range operation.Parameters { + p := parameter.GetParameter() + if p != nil { + b := p.GetBodyParameter() + if b != nil && b.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Parameter has no description.", + Keys: append(keys, []string{"responses", b.Name}...)}) + } + n := p.GetNonBodyParameter() + if n != nil { + hp := n.GetHeaderParameterSubSchema() + if hp != nil && hp.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Parameter has no description.", + Keys: append(keys, []string{"responses", hp.Name}...)}) + } + fp := n.GetFormDataParameterSubSchema() + if fp != nil && fp.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Parameter has no description.", + Keys: append(keys, []string{"responses", fp.Name}...)}) + } + qp := n.GetQueryParameterSubSchema() + if qp != nil && qp.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Parameter has no description.", + Keys: append(keys, []string{"responses", qp.Name}...)}) + } + pp := n.GetPathParameterSubSchema() + if pp != nil && pp.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Parameter has no description.", + Keys: append(keys, []string{"responses", pp.Name}...)}) + } + } + } + } + for _, pair := range operation.Responses.ResponseCode { + value := pair.Value + response := value.GetResponse() + if response != nil { + responseSchema := response.Schema + responseSchemaSchema := responseSchema.GetSchema() + if responseSchemaSchema != nil && responseSchemaSchema.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Response has no description.", + Keys: append(keys, []string{"responses", pair.Name}...)}) + } + responseFileSchema := responseSchema.GetFileSchema() + if responseFileSchema != nil && responseFileSchema.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Response has no description.", + Keys: append(keys, []string{"responses", pair.Name}...)}) + } + } + } + return messages +} + +// Analyze a definition in an OpenAPI description. +func (s *DocumentLinterV2) analyzeDefinition(keys []string, definition *openapi.Schema) []*plugins.Message { + messages := make([]*plugins.Message, 0) + if definition.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Definition has no description.", + Keys: keys}) + } + + if definition.Properties != nil { + for _, pair := range definition.Properties.AdditionalProperties { + propertySchema := pair.Value + if propertySchema.Description == "" { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_WARNING, + Code: "NODESCRIPTION", + Text: "Property has no description.", + Keys: append(keys, []string{"properties", pair.Name}...)}) + } + } + } + return messages +} diff --git a/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/linter_v3.go b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/linter_v3.go new file mode 100644 index 000000000..67891005c --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/linter_v3.go @@ -0,0 +1,33 @@ +// Copyright 2017 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package main + +import ( + openapi "github.com/googleapis/gnostic/OpenAPIv3" + plugins "github.com/googleapis/gnostic/plugins" +) + +// DocumentLinter contains information collected about an API description. +type DocumentLinterV3 struct { +} + +func (d *DocumentLinterV3) Run() []*plugins.Message { + return nil +} + +// NewDocumentLinter builds a new DocumentLinter object. +func NewDocumentLinterV3(document *openapi.Document) *DocumentLinterV3 { + return &DocumentLinterV3{} +} diff --git a/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/main.go b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/main.go new file mode 100644 index 000000000..707a53099 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-descriptions/main.go @@ -0,0 +1,61 @@ +// Copyright 2017 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// gnostic_lint is a tool for analyzing OpenAPI descriptions. +// +// It scans an API description and checks it against a set of +// coding style guidelines. +// +// Results are returned in a JSON structure. +package main + +import ( + "github.com/golang/protobuf/proto" + openapiv2 "github.com/googleapis/gnostic/OpenAPIv2" + openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" + plugins "github.com/googleapis/gnostic/plugins" +) + +type DocumentLinter interface { + Run() []*plugins.Message +} + +// This is the main function for the plugin. +func main() { + env, err := plugins.NewEnvironment() + env.RespondAndExitIfError(err) + + var linter DocumentLinter + + for _, model := range env.Request.Models { + switch model.TypeUrl { + case "openapi.v2.Document": + documentv2 := &openapiv2.Document{} + err = proto.Unmarshal(model.Value, documentv2) + if err == nil { + linter = NewDocumentLinterV2(documentv2) + env.Response.Messages = linter.Run() + } + case "openapi.v3.Document": + documentv3 := &openapiv3.Document{} + err = proto.Unmarshal(model.Value, documentv3) + if err == nil { + linter = NewDocumentLinterV3(documentv3) + env.Response.Messages = linter.Run() + } + } + } + + env.RespondAndExit() +} diff --git a/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-paths/README.md b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-paths/README.md new file mode 100644 index 000000000..6563fd3bb --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-paths/README.md @@ -0,0 +1,9 @@ +# gnostic-lint-paths + +This directory contains a `gnostic` plugin that analyzes an OpenAPI +description for factors that might influence code generation and other +API automation. + +The plugin can be invoked like this: + + gnostic bookstore.json --lint-paths diff --git a/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-paths/main.go b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-paths/main.go new file mode 100644 index 000000000..edf9a53f0 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/go/gnostic-lint-paths/main.go @@ -0,0 +1,77 @@ +// Copyright 2018 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// gnostic-lint-paths is a tool for analyzing paths in OpenAPI descriptions. +// +// It scans an API description and checks it against a set of coding style guidelines. +package main + +import ( + "github.com/golang/protobuf/proto" + openapiv2 "github.com/googleapis/gnostic/OpenAPIv2" + openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" + plugins "github.com/googleapis/gnostic/plugins" +) + +func checkPathsV2(document *openapiv2.Document, messages []*plugins.Message) []*plugins.Message { + for _, pair := range document.Paths.Path { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_INFO, + Code: "PATH", + Text: pair.Name, + Keys: []string{"paths", pair.Name}}) + } + return messages +} + +func checkPathsV3(document *openapiv3.Document, messages []*plugins.Message) []*plugins.Message { + for _, pair := range document.Paths.Path { + messages = append(messages, + &plugins.Message{ + Level: plugins.Message_INFO, + Code: "PATH", + Text: pair.Name, + Keys: []string{"paths", pair.Name}}) + } + return messages +} + +func main() { + env, err := plugins.NewEnvironment() + env.RespondAndExitIfError(err) + + messages := make([]*plugins.Message, 0, 0) + + for _, model := range env.Request.Models { + switch model.TypeUrl { + case "openapi.v2.Document": + documentv2 := &openapiv2.Document{} + err = proto.Unmarshal(model.Value, documentv2) + if err == nil { + messages = checkPathsV2(documentv2, messages) + } + case "openapi.v3.Document": + documentv3 := &openapiv3.Document{} + err = proto.Unmarshal(model.Value, documentv3) + if err == nil { + messages = checkPathsV3(documentv3, messages) + } + } + } + + env.RespondAndExitIfError(err) + env.Response.Messages = messages + env.RespondAndExit() +} diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/Makefile b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/Makefile new file mode 100644 index 000000000..9327ae262 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/Makefile @@ -0,0 +1,20 @@ + +GNOSTIC = $(GOPATH)/src/github.com/googleapis/gnostic + +plugin: + node_modules/.bin/pbjs -t json \ + $(GNOSTIC)/OpenAPIv2/OpenAPIv2.proto \ + $(GNOSTIC)/OpenAPIv3/OpenAPIv3.proto \ + $(GNOSTIC)/discovery/discovery.proto \ + $(GNOSTIC)/surface/surface.proto \ + $(GNOSTIC)/plugins/plugin.proto \ + > bundle.json + node_modules/.bin/nexe gnostic-lint-operations.js + +run: plugin + gnostic $(GNOSTIC)/examples/v2.0/yaml/petstore.yaml --lint-operations + +setup: + npm install protobufjs + npm install get-stdin + npm install nexe diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/README.md b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/README.md new file mode 100644 index 000000000..e932dd70b --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/README.md @@ -0,0 +1,15 @@ +This directory contains a gnostic linter written with node. + +It is built using [dcodeIO/Protobuf.js](https://github.com/dcodeIO/ProtoBuf.js). + +### SETUP + +- Install node. +- Run `make setup` to install node dependencies. + +### TRY IT + +- Run `make run` to test-run the plugin. + + + diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/gnostic-lint-operations.js b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/gnostic-lint-operations.js new file mode 100644 index 000000000..6899d1067 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-operations/gnostic-lint-operations.js @@ -0,0 +1,47 @@ +// import libraries +const protobuf = require("protobufjs") +const getStdin = require('get-stdin') + +// import messages +const root = protobuf.Root.fromJSON(require("./bundle.json")) +const Request = root.lookupType("gnostic.plugin.v1.Request") +const Response = root.lookupType("gnostic.plugin.v1.Response") +const Document = root.lookupType("openapi.v2.Document") + +getStdin.buffer().then(buffer => { + const request = Request.decode(buffer) + messages = [] + for (var j in request.models) { + const m = request.models[j] + if (m.type_url == "openapi.v2.Document") { + const openapi2 = Document.decode(m.value) + const paths = openapi2.paths.path + for (var i in paths) { + const path = paths[i] + //console.error('path %s\n\n', path.name) + const getOperation = path.value.get + if (getOperation && getOperation.operationId == "") { + messages.push({level:3, code:"NOOPERATIONID", text:"No operation id.", keys:["paths", path.name, "get"]}) + } + const postOperation = path.value.post + if (postOperation && postOperation.operationId == "") { + messages.push({level:3, code:"NOOPERATIONID", text:"No operation id.", keys:["paths", path.name, "post"]}) + } + //console.error('get %s\n\n', JSON.stringify(getOperation)) + } + } + } + + const payload = { + messages: messages + } + + // Verify the payload if necessary (i.e. when possibly incomplete or invalid) + const errMsg = Response.verify(payload) + if (errMsg) + throw Error(errMsg) + + const message = Response.create(payload) + process.stdout.write(Response.encode(message).finish()) + +}).catch(err => console.error(err)) diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/Makefile b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/Makefile new file mode 100644 index 000000000..61de67ee3 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/Makefile @@ -0,0 +1,19 @@ + +GNOSTIC = $(GOPATH)/src/github.com/googleapis/gnostic + +plugin: + node_modules/.bin/pbjs -t json \ + $(GNOSTIC)/OpenAPIv2/OpenAPIv2.proto \ + $(GNOSTIC)/OpenAPIv3/OpenAPIv3.proto \ + $(GNOSTIC)/discovery/discovery.proto \ + $(GNOSTIC)/surface/surface.proto \ + $(GNOSTIC)/plugins/plugin.proto \ + > bundle.json + node_modules/.bin/nexe gnostic-lint-responses.js + mv gnostic-lint-responses $(GOPATH)/bin + +run: plugin + gnostic $(GNOSTIC)/examples/v2.0/yaml/petstore.yaml --lint-responses + +setup: + npm install diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/README.md b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/README.md new file mode 100644 index 000000000..e932dd70b --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/README.md @@ -0,0 +1,15 @@ +This directory contains a gnostic linter written with node. + +It is built using [dcodeIO/Protobuf.js](https://github.com/dcodeIO/ProtoBuf.js). + +### SETUP + +- Install node. +- Run `make setup` to install node dependencies. + +### TRY IT + +- Run `make run` to test-run the plugin. + + + diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/gnostic-lint-responses.js b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/gnostic-lint-responses.js new file mode 100644 index 000000000..1c389a497 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/gnostic-lint-responses.js @@ -0,0 +1,83 @@ +// import libraries +const protobuf = require("protobufjs"); +const getStdin = require("get-stdin"); +const find = require("lodash/find"); +const forEach = require("lodash/forEach"); +const pick = require("lodash/pick"); + +// import messages +const root = protobuf.Root.fromJSON(require("./bundle.json")); +const Request = root.lookupType("gnostic.plugin.v1.Request"); +const Response = root.lookupType("gnostic.plugin.v1.Response"); +const Document = root.lookupType("openapi.v2.Document"); + +getStdin.buffer().then(buffer => { + const request = Request.decode(buffer); + var messages = []; + for (var j in request.models) { + const m = request.models[j]; + if (m.type_url == "openapi.v2.Document") { + const openapi2 = Document.decode(m.value); + const paths = openapi2.paths.path; + for (var i in paths) { + const path = paths[i]; + // console.error('path %s\n\n', path.name) + + // Arrays MUST NOT be returned as the top-level structure in a response body. + let pathOps = pick(path.value, ["get","head","post", "put", "patch", "delete", "options"]); + forEach(pathOps, (op, opKey) => { + if (op != null) { + forEach(op.responses.responseCode, responseObj => { + // console.error('responseObj is %j', responseObj) + name = responseObj.name; + response = responseObj.value.response; + if (response.schema && response.schema.schema) { + if (!response.schema.schema._ref) { + if ( + response.schema.schema.type != null && + response.schema.schema.type.value == "array" + ) { + messages.push({ + level: 3, + code: "NO_ARRAY_RESPONSES", + text: "Arrays MUST NOT be returned as the top-level structure in a response body.", + keys: ["paths", path.name, opKey, "responses", name, "schema"] + }); + } + } else { + let schemaName = response.schema.schema._ref.match(/#\/definitions\/(\w+)/); + if (schemaName) { + const definitions = openapi2.definitions.additionalProperties; + const schemaKvp = find(definitions, {name: schemaName[1] + }); + //console.error('schemaKvp.value.type = %s', schemaKvp.value.type.value) + if (schemaKvp && schemaKvp.value.type && schemaKvp.value.type.value.indexOf("array") >= 0) { + messages.push({ + level: 3, + code: "NO_ARRAY_RESPONSES", + text: "Arrays MUST NOT be returned as the top-level structure in a response body.", + keys: ["paths", path.name, opKey, "responses", name, "schema" ] + }); + } + } + } + } + }); + } + }); + } + } + } + + const payload = { + messages: messages + }; + + // Verify the payload if necessary (i.e. when possibly incomplete or invalid) + const errMsg = Response.verify(payload); + if (errMsg) throw Error(errMsg); + + const message = Response.create(payload); + process.stdout.write(Response.encode(message).finish()); +}) +.catch(err => console.error(err)); diff --git a/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/package.json b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/package.json new file mode 100644 index 000000000..e6b1d0158 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/node/gnostic-lint-responses/package.json @@ -0,0 +1,19 @@ +{ + "name": "gnostic-lint-responses", + "version": "1.0.0", + "description": "Gnostic linter plugin to check responses", + "main": "gnostic-lint-responses.js", + "repository": "git@github.com:googleapis/gnostic.git", + "author": "Mike Kistler", + "license": "Apache-2.0", + "dependencies": { + "get-stdin": "^5.0.1", + "lodash": "^4.17.5", + "nexe": "^2.0.0-rc.24", + "protobufjs": "^6.8.4" + }, + "devDependencies": {}, + "scripts": { + "test": "echo \"Error: no test specified\" && exit 1" + } +} diff --git a/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Makefile b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Makefile new file mode 100644 index 000000000..b7a141125 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Makefile @@ -0,0 +1,18 @@ + +TGT=gnostic-lint-responses-swift + +BINDIR=.build/debug + +all: + swift build + +install: all + cp $(BINDIR)/$(TGT) $(GOPATH)/bin/$(TGT) + +clean : + rm -rf Packages + rm -rf .build + rm -f Package.pins Package.resolved + +run: + gnostic ../../../examples/v2.0/yaml/petstore.yaml --lint-responses-swift diff --git a/vendor/github.com/coreos/go-semver/semver/sort.go b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Package.swift similarity index 53% rename from vendor/github.com/coreos/go-semver/semver/sort.go rename to vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Package.swift index e256b41a5..9877a4180 100644 --- a/vendor/github.com/coreos/go-semver/semver/sort.go +++ b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Package.swift @@ -1,10 +1,10 @@ -// Copyright 2013-2015 CoreOS, Inc. +// Copyright 2018 Google Inc. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // -// http://www.apache.org/licenses/LICENSE-2.0 +// http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, @@ -12,27 +12,15 @@ // See the License for the specific language governing permissions and // limitations under the License. -package semver +import PackageDescription -import ( - "sort" +let package = Package( + name: "gnostic-lint-responses-swift", + targets: [ + Target(name: "gnostic-lint-responses-swift", dependencies: [ "Gnostic" ]), + Target(name: "Gnostic") + ], + dependencies: [ + .Package(url: "https://github.com/apple/swift-protobuf.git", Version(1,0,2)) + ] ) - -type Versions []*Version - -func (s Versions) Len() int { - return len(s) -} - -func (s Versions) Swap(i, j int) { - s[i], s[j] = s[j], s[i] -} - -func (s Versions) Less(i, j int) bool { - return s[i].LessThan(*s[j]) -} - -// Sort sorts the given slice of Version -func Sort(versions []*Version) { - sort.Sort(Versions(versions)) -} diff --git a/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/README.md b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/README.md new file mode 100644 index 000000000..178f6a237 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/README.md @@ -0,0 +1,12 @@ +This directory contains a gnostic linter written with Swift. + +### SETUP + +- Install protoc (https://github.com/google/protobuf) +- Install the Swift Protocol Buffer plugin (https://github.com/apple/swift-protobuf) +- Run `sh compile-protos` to generate Protocol Buffer support code. +- Run `make install` to build and install the plugin. + +### TRY IT + +- Run `make run` to test-run the plugin. diff --git a/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Sources/gnostic-lint-responses-swift/io.swift b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Sources/gnostic-lint-responses-swift/io.swift new file mode 100644 index 000000000..38d5f34a1 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Sources/gnostic-lint-responses-swift/io.swift @@ -0,0 +1,77 @@ +// Copyright 2018 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import Foundation + +// The I/O code below is derived from Apple's swift-protobuf project. +// https://github.com/apple/swift-protobuf +// BEGIN swift-protobuf derivation + +#if os(Linux) + import Glibc +#else + import Darwin.C +#endif + +enum PluginError: Error { + /// Raised for any errors reading the input + case readFailure +} + +// Alias clib's write() so Stdout.write(bytes:) can call it. +private let _write = write + +class Stdin { + static func readall() throws -> Data { + let fd: Int32 = 0 + let buffSize = 32 + var buff = [UInt8]() + while true { + var fragment = [UInt8](repeating: 0, count: buffSize) + let count = read(fd, &fragment, buffSize) + if count < 0 { + throw PluginError.readFailure + } + if count < buffSize { + buff += fragment[0..) -> () in + _ = _write(1, p, bytes.count) + } + } + static func write(_ string: String) { + self.write(bytes:string.data(using:.utf8)!) + } +} + +class Stderr { + static func write(bytes: Data) { + bytes.withUnsafeBytes { (p: UnsafePointer) -> () in + _ = _write(2, p, bytes.count) + } + } + static func write(_ string: String) { + self.write(bytes:string.data(using:.utf8)!) + } +} + +// END swift-protobuf derivation diff --git a/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Sources/gnostic-lint-responses-swift/main.swift b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Sources/gnostic-lint-responses-swift/main.swift new file mode 100644 index 000000000..affbc7012 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/Sources/gnostic-lint-responses-swift/main.swift @@ -0,0 +1,116 @@ +// Copyright 2018 Google Inc. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import Foundation +import Gnostic + +extension Gnostic_Plugin_V1_Response { + mutating func message(level:Gnostic_Plugin_V1_Message.Level, + code:String, + text:String, + path:[String]=[]) { + var message = Gnostic_Plugin_V1_Message() + message.level = level + message.code = code + message.text = text + message.keys = path + messages.append(message) + } +} + +class ResponseLinter { + var document : Openapi_V2_Document = Openapi_V2_Document() + + func run(_ request : Gnostic_Plugin_V1_Request, + _ response : inout Gnostic_Plugin_V1_Response) throws { + for model in request.models { + if model.typeURL == "openapi.v2.Document" { + let document = try Openapi_V2_Document(serializedData: model.value) + self.document = document + for pair in document.paths.path { + let path = ["paths", pair.name] + let v = pair.value + if v.hasGet { + checkOperation(v.get, path:path + ["get"], response:&response) + } + if v.hasPost { + checkOperation(v.post, path:path + ["post"], response:&response) + } + if v.hasPut { + checkOperation(v.put, path:path + ["put"], response:&response) + } + if v.hasDelete { + checkOperation(v.delete, path:path + ["delete"], response:&response) + } + } + } + } + } + + func checkOperation(_ operation:Openapi_V2_Operation, + path:[String], + response:inout Gnostic_Plugin_V1_Response) { + for responseCode in operation.responses.responseCode { + let code = responseCode.name + if responseCode.value.response.hasSchema { + var schema = responseCode.value.response.schema.schema + if schema.ref != "" { + if let resolvedSchema = resolveReference(schema.ref) { + schema = resolvedSchema + } + } + checkSchemaType(schema, path: path + ["responses", code, "schema"], response: &response) + } + } + } + + func checkSchemaType(_ schema:Openapi_V2_Schema, + path:[String], + response:inout Gnostic_Plugin_V1_Response) { + if schema.hasType { + for type in schema.type.value { + if type == "array" { + response.message( + level: .error, + code: "NO_ARRAY_RESPONSES", + text: "Arrays MUST NOT be returned as the top-level structure in a response body.", + path: path) + } + } + } + } + + func resolveReference(_ reference:String) -> Openapi_V2_Schema? { + let prefix = "#/definitions/" + if reference.hasPrefix(prefix) { + let schemaName = reference.dropFirst(prefix.count) + for pair in document.definitions.additionalProperties { + if pair.name == schemaName { + return pair.value + } + } + } + return nil + } +} + +func main() throws { + let request = try Gnostic_Plugin_V1_Request(serializedData: Stdin.readall()) + var response = Gnostic_Plugin_V1_Response() + try ResponseLinter().run(request, &response) + let serializedResponse = try response.serializedData() + Stdout.write(bytes: serializedResponse) +} + +try main() diff --git a/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/compile-protos b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/compile-protos new file mode 100644 index 000000000..61aa9f897 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/linters/swift/gnostic-lint-responses-swift/compile-protos @@ -0,0 +1,54 @@ +#!/bin/sh +# +# Copyright 2018 Google Inc. All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +# Use this script to run protoc and swift-proto to generate +# support code for gnostic protos. + +GNOSTIC=$GOPATH/src/github.com/googleapis/gnostic + +PROTOS=( + plugins/plugin.proto + OpenAPIv2/OpenAPIv2.proto + OpenAPIv3/OpenAPIv3.proto + surface/surface.proto + discovery/discovery.proto +) + +mkdir -p Sources/Gnostic + +# remove old compiled pb files +rm -rf Sources/Gnostic/*.pb.swift + +# remove any prior compilations +rm -rf Sources/Gnostic/github.com + +# compile protos +for proto in "${PROTOS[@]}" +do + echo "COMPILING $proto" + protoc $GNOSTIC/$proto \ + --swift_opt=Visibility=Public \ + --swift_out=Sources/Gnostic \ + --proto_path=$GOPATH/src + +# relocate compiled protos +find Sources/Gnostic/github.com -name "*.pb.swift" -exec mv {} Sources/Gnostic \; + +# remove scaffolding of compilation +rm -rf Sources/Gnostic/github.com + +done + diff --git a/vendor/github.com/googleapis/gnostic/plugins/environment.go b/vendor/github.com/googleapis/gnostic/plugins/environment.go index 9a3bf9865..79a0867f2 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/environment.go +++ b/vendor/github.com/googleapis/gnostic/plugins/environment.go @@ -1,6 +1,7 @@ package gnostic_plugin_v1 import ( + "errors" "flag" "fmt" "io" @@ -10,9 +11,12 @@ import ( "path" "github.com/golang/protobuf/proto" + "github.com/golang/protobuf/ptypes/any" openapiv2 "github.com/googleapis/gnostic/OpenAPIv2" openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" + discovery "github.com/googleapis/gnostic/discovery" + surface "github.com/googleapis/gnostic/surface" ) // Environment contains the environment of a plugin call. @@ -77,7 +81,7 @@ When the -plugin option is specified, these flags are ignored.`) } // Log the invocation. - log.Printf("Running plugin %s", env.Invocation) + //log.Printf("Running plugin %s", env.Invocation) env.Request = request @@ -98,20 +102,36 @@ When the -plugin option is specified, these flags are ignored.`) documentv2 := &openapiv2.Document{} err = proto.Unmarshal(apiData, documentv2) if err == nil { - env.Request.Openapi2 = documentv2 - } else { - // ignore deserialization errors + env.Request.AddModel("openapi.v2.Document", documentv2) + // include experimental API surface model + surfaceModel, err := surface.NewModelFromOpenAPI2(documentv2) + if err != nil { + env.Request.AddModel("surface.v1.Model", surfaceModel) + } + return env, err } - - // Then try to unmarshal OpenAPI v3. + // If that failed, ignore deserialization errors and try to unmarshal OpenAPI v3. documentv3 := &openapiv3.Document{} err = proto.Unmarshal(apiData, documentv3) if err == nil { - env.Request.Openapi3 = documentv3 - } else { - // ignore deserialization errors + env.Request.AddModel("openapi.v3.Document", documentv3) + // include experimental API surface model + surfaceModel, err := surface.NewModelFromOpenAPI3(documentv3) + if err != nil { + env.Request.AddModel("surface.v1.Model", surfaceModel) + } + return env, err } - + // If that failed, ignore deserialization errors and try to unmarshal a Discovery document. + discoveryDocument := &discovery.Document{} + err = proto.Unmarshal(apiData, discoveryDocument) + if err == nil { + env.Request.AddModel("discovery.v1.Document", discoveryDocument) + return env, err + } + // If we get here, we don't know what we got + err = errors.New("Unrecognized format for input") + return env, err } return env, err } @@ -172,6 +192,12 @@ func HandleResponse(response *Response, outputLocation string) error { return nil } +func (request *Request) AddModel(modelType string, model proto.Message) error { + modelBytes, err := proto.Marshal(model) + request.Models = append(request.Models, &any.Any{TypeUrl: modelType, Value: modelBytes}) + return err +} + func isFile(path string) bool { fileInfo, err := os.Stat(path) if err != nil { diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-analyze/main.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-analyze/main.go index 09a86f7bd..ad7716151 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-analyze/main.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-analyze/main.go @@ -30,9 +30,12 @@ import ( "path" "strings" - "github.com/golang/protobuf/proto" - plugins "github.com/googleapis/gnostic/plugins" "github.com/googleapis/gnostic/plugins/gnostic-analyze/statistics" + + "github.com/golang/protobuf/proto" + openapiv2 "github.com/googleapis/gnostic/OpenAPIv2" + openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" + plugins "github.com/googleapis/gnostic/plugins" ) // Record an error, then serialize and return a response. @@ -56,14 +59,24 @@ func main() { env.RespondAndExitIfError(err) var stats *statistics.DocumentStatistics - if env.Request.Openapi2 != nil { - // Analyze the API document. - stats = statistics.NewDocumentStatistics(env.Request.SourceName, env.Request.Openapi2) - } - if env.Request.Openapi3 != nil { - // Analyze the API document. - stats = statistics.NewDocumentStatisticsV3(env.Request.SourceName, env.Request.Openapi3) + for _, model := range env.Request.Models { + switch model.TypeUrl { + case "openapi.v2.Document": + documentv2 := &openapiv2.Document{} + err = proto.Unmarshal(model.Value, documentv2) + if err == nil { + // Analyze the API document. + stats = statistics.NewDocumentStatistics(env.Request.SourceName, documentv2) + } + case "openapi.v3.Document": + documentv3 := &openapiv3.Document{} + err = proto.Unmarshal(model.Value, documentv3) + if err == nil { + // Analyze the API document. + stats = statistics.NewDocumentStatisticsV3(env.Request.SourceName, documentv3) + } + } } if stats != nil { diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/apis_guru/main.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/apis_guru/main.go index 2a2b35451..72ff44c59 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/apis_guru/main.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/apis_guru/main.go @@ -1,9 +1,13 @@ +// +build ignore +// This file is omitted when getting with `go get github.com/googleapis/gnostic/...` + package main import ( "fmt" - "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/apis_guru/apis_guru" "sort" + + "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/apis_guru/apis_guru" ) func main() { diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/bookstore/bookstore/bookstore.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/bookstore/bookstore/bookstore.go new file mode 100644 index 000000000..1f7dec7db --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/bookstore/bookstore/bookstore.go @@ -0,0 +1,19 @@ +/* + Copyright 2017 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +// Package bookstore exists to allow this repo to work with recursive go get. +// It will be filled in with auto generated code. +package bookstore diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/Makefile b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/Makefile new file mode 100644 index 000000000..4b3fe2c0d --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/Makefile @@ -0,0 +1,20 @@ +build: + go get golang.org/x/tools/cmd/goimports + go install github.com/googleapis/gnostic + go install github.com/googleapis/gnostic/plugins/gnostic-go-generator + rm -f $(GOPATH)/bin/gnostic-go-client $(GOPATH)/bin/gnostic-go-server + ln -s $(GOPATH)/bin/gnostic-go-generator $(GOPATH)/bin/gnostic-go-client + ln -s $(GOPATH)/bin/gnostic-go-generator $(GOPATH)/bin/gnostic-go-server + +all: build + gnostic sample.yaml --go-generator-out=sample + +clean: + rm -rf sample service/service + +test: all + killall service; true # ignore errors due to no matching processes + cd service; go get .; go build; ./service & + go test + killall service + diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/README.md b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/README.md new file mode 100644 index 000000000..f4ce24fcc --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/README.md @@ -0,0 +1,24 @@ +# API Sample + +This directory contains an OpenAPI description of a sample API +that exercises various OpenAPI features. + +Use this example to try the `gnostic-go-generator` plugin, which implements +`gnostic-go-client` and `gnostic-go-server` for generating API client and +server code, respectively. + +Run "make all" to build and install `gnostic` and the Go plugins. +It will generate both client and server code. The API client and +server code will be in the `sample` package. + +The `service` directory contains additional code that completes the server. +To build and run the service, `cd service` and do the following: + + go get . + go build + ./service & + +To test the service with the generated client, go back up to the top-level +directory and run `go test`. The test in `sample_test.go` uses client +code generated in `sample` to verify the service. + diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample.yaml b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample.yaml new file mode 100644 index 000000000..6235a3861 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample.yaml @@ -0,0 +1,67 @@ +swagger: '2.0' +schemes: + - https +host: sample.io +basePath: / +info: + title: sample.io + version: '1.0' +consumes: + - application/json +produces: + - application/json;charset=UTF-8 +securityDefinitions: + api_key: + in: query + name: key + type: apiKey +paths: + /sample/{id}: + get: + operationId: "GetSample" + parameters: + - description: identifier + in: path + name: id + required: true + type: string + responses: + '200': + description: sample response + schema: + $ref: '#/definitions/Sample' + '401': + description: User doesn't have a valid session. + schema: + $ref: '#/definitions/APIError' + '404': + description: Unable to find supplied extractor ID. + schema: + $ref: '#/definitions/APIError' + security: + - api_key: [] + summary: Get a sample response + tags: + - sample + - demo +definitions: + APIError: + properties: + code: + description: Internal error code + format: int + type: integer + message: + description: A message containing a brief description of the error + type: string + type: object + Sample: + properties: + id: + type: string + thing: + type: object + count: + format: int32 + type: integer + type: object diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample/sample.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample/sample.go new file mode 100644 index 000000000..18cd0ddc8 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample/sample.go @@ -0,0 +1,19 @@ +/* + Copyright 2017 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +// Package sample exists to allow this repo to work with recursive go get. +// It will be filled in with auto generated code. +package sample diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample_test.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample_test.go new file mode 100644 index 000000000..4d6a44ec8 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample_test.go @@ -0,0 +1,68 @@ +/* + Copyright 2018 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package test + +import ( + "fmt" + "net/http" + "strings" + "testing" + + "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample" +) + +const service = "http://localhost:8080" + +func TestSample(t *testing.T) { + // create a client + s := sample.NewClient(service, nil) + // verify a sample request + { + message := "hello world" + response, err := s.GetSample(message) + if err != nil { + t.Log("get sample failed") + t.Fail() + } + if response.OK.Id != message || response.OK.Count != int32(len(message)) { + t.Log(fmt.Sprintf("get sample received %+v", response.OK)) + t.Fail() + } + if (response == nil) || (response.OK == nil) { + t.Log(fmt.Sprintf("get sample failed %+v", response.OK)) + t.Fail() + } + } + // verify the handling of an invalid request + { + req, err := http.NewRequest("GET", service+"/unsupported", strings.NewReader("")) + if err != nil { + t.Log("bad request failed") + return + } + resp, err := http.DefaultClient.Do(req) + if err != nil { + return + } + // we expect a 404 (Not Found) code + if resp.StatusCode != 404 { + t.Log("bad request failed") + t.Fail() + } + return + } +} diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/app.yaml b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/app.yaml new file mode 100644 index 000000000..7f0272d47 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/app.yaml @@ -0,0 +1,9 @@ +application: sample +version: 1 +runtime: go +api_version: go1 +handlers: +- url: /.* + script: _go_app +- url: / + static_dir: static diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/init.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/init.go new file mode 100644 index 000000000..13d3dc1a4 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/init.go @@ -0,0 +1,27 @@ +/* + Copyright 2018 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package main + +import ( + "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample" +) + +// init() is called when the package is loaded +// this allows this app to be trivially deployed to Google App Engine, which does not call main() +func init() { + sample.Initialize(NewService()) +} diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/main.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/main.go new file mode 100644 index 000000000..e65195400 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/main.go @@ -0,0 +1,34 @@ +// +build !appengine + +// This file is omitted when the app is built for Google App Engine + +/* + Copyright 2018 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package main + +import ( + "log" + + "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample" +) + +func main() { + err := sample.ServeHTTP(":8080") + if err != nil { + log.Printf("%v", err) + } +} diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/service.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/service.go new file mode 100644 index 000000000..6d66cea7c --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/service/service.go @@ -0,0 +1,38 @@ +/* + Copyright 2018 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +package main + +import ( + "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/sample/sample" +) + +// +// The Service type implements a sample service. +// +type Service struct{} + +func NewService() *Service { + return &Service{} +} + +func (service *Service) GetSample(parameters *sample.GetSampleParameters, responses *sample.GetSampleResponses) (err error) { + (*responses).OK = &sample.Sample{ + Id: parameters.Id, + Thing: map[string]interface{}{"thing": 123}, + Count: int32(len(parameters.Id))} + return err +} diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/main.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/main.go index 1bd3448ee..81c357921 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/main.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/main.go @@ -2,6 +2,7 @@ package main import ( "fmt" + "github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/xkcd" ) diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/xkcd/xkcd.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/xkcd/xkcd.go new file mode 100644 index 000000000..7309ab188 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v2.0/xkcd/xkcd/xkcd.go @@ -0,0 +1,19 @@ +/* + Copyright 2017 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +// Package xkcd exists to allow this repo to work with recursive go get. +// It will be filled in with auto generated code. +package xkcd diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v3.0/bookstore/bookstore/bookstore.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v3.0/bookstore/bookstore/bookstore.go new file mode 100644 index 000000000..1f7dec7db --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v3.0/bookstore/bookstore/bookstore.go @@ -0,0 +1,19 @@ +/* + Copyright 2017 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +// Package bookstore exists to allow this repo to work with recursive go get. +// It will be filled in with auto generated code. +package bookstore diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v3.0/urlshortener/urlshortener/urlshortener.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v3.0/urlshortener/urlshortener/urlshortener.go new file mode 100644 index 000000000..5084f9c12 --- /dev/null +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/examples/v3.0/urlshortener/urlshortener/urlshortener.go @@ -0,0 +1,19 @@ +/* + Copyright 2017 Google Inc. All Rights Reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. +*/ + +// Package urlshortener exists to allow this repo to work with recursive go get. +// It will be filled in with auto generated code. +package urlshortener diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/language.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/language.go index 704370874..aa5e58feb 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/language.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/language.go @@ -16,8 +16,8 @@ package main import ( surface "github.com/googleapis/gnostic/surface" - "unicode" "strings" + "unicode" ) type GoLanguageModel struct{} @@ -49,7 +49,7 @@ func (language *GoLanguageModel) Prepare(model *surface.Model) { f.NativeType = "int64" } case "object": - f.NativeType = "{}interface" + f.NativeType = "interface{}" case "string": f.NativeType = "string" default: @@ -88,6 +88,8 @@ func goFieldName(name string) string { // avoid integers if name == "200" { return "OK" + } else if unicode.IsDigit(rune(name[0])) { + return "Code" + name } return name } diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/main.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/main.go index cbdab39d1..47d0cf692 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/main.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/main.go @@ -21,7 +21,9 @@ import ( "errors" "strings" + "github.com/golang/protobuf/proto" plugins "github.com/googleapis/gnostic/plugins" + surface "github.com/googleapis/gnostic/surface" ) // This is the main function for the code generation plugin. @@ -42,30 +44,33 @@ func main() { files = []string{"client.go", "server.go", "provider.go", "types.go", "constants.go"} } - // Get the code surface model. - model := env.Request.Surface + for _, model := range env.Request.Models { + switch model.TypeUrl { + case "surface.v1.Model": + surfaceModel := &surface.Model{} + err = proto.Unmarshal(model.Value, surfaceModel) + if err == nil { + // Customize the code surface model for Go + NewGoLanguageModel().Prepare(surfaceModel) - if model == nil { - err = errors.New("No generated code surface model is available.") - env.RespondAndExitIfError(err) + modelJSON, _ := json.MarshalIndent(surfaceModel, "", " ") + modelFile := &plugins.File{Name: "model.json", Data: modelJSON} + env.Response.Files = append(env.Response.Files, modelFile) + + // Create the renderer. + renderer, err := NewServiceRenderer(surfaceModel) + renderer.Package = packageName + env.RespondAndExitIfError(err) + + // Run the renderer to generate files and add them to the response object. + err = renderer.Render(env.Response, files) + env.RespondAndExitIfError(err) + + // Return with success. + env.RespondAndExit() + } + } } - - // Customize the code surface model for Go - NewGoLanguageModel().Prepare(model) - - modelJSON, _ := json.MarshalIndent(model, "", " ") - modelFile := &plugins.File{Name: "model.json", Data: modelJSON} - env.Response.Files = append(env.Response.Files, modelFile) - - // Create the renderer. - renderer, err := NewServiceRenderer(model) - renderer.Package = packageName + err = errors.New("No generated code surface model is available.") env.RespondAndExitIfError(err) - - // Run the renderer to generate files and add them to the response object. - err = renderer.Render(env.Response, files) - env.RespondAndExitIfError(err) - - // Return with success. - env.RespondAndExit() } diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_client.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_client.go index 6d93b4e7b..363e6648c 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_client.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_client.go @@ -109,7 +109,9 @@ func (renderer *Renderer) RenderClient() ([]byte, error) { if method.Method == "POST" { f.WriteLine(`body := new(bytes.Buffer)`) - f.WriteLine(`json.NewEncoder(body).Encode(` + parametersType.FieldWithPosition(surface.Position_BODY).Name + `)`) + if parametersType != nil { + f.WriteLine(`json.NewEncoder(body).Encode(` + parametersType.FieldWithPosition(surface.Position_BODY).Name + `)`) + } f.WriteLine(`req, err := http.NewRequest("` + method.Method + `", path, body)`) f.WriteLine(`reqHeaders := make(http.Header)`) f.WriteLine(`reqHeaders.Set("Content-Type", "application/json")`) diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_types.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_types.go index 2d7ab1e5f..e77707e1c 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_types.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-go-generator/render_types.go @@ -43,7 +43,7 @@ func (renderer *Renderer) RenderTypes() ([]byte, error) { } else if modelType.Kind == surface.TypeKind_OBJECT { f.WriteLine(`type ` + modelType.TypeName + ` map[string]` + modelType.ContentType) } else { - f.WriteLine(`type ` + modelType.TypeName + ` struct {}`) + f.WriteLine(`type ` + modelType.TypeName + ` interface {}`) } } return f.Bytes(), nil diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-summary/main.go b/vendor/github.com/googleapis/gnostic/plugins/gnostic-summary/main.go index 8172da3e2..90864886b 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-summary/main.go +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-summary/main.go @@ -17,14 +17,17 @@ package main import ( - openapi2 "github.com/googleapis/gnostic/OpenAPIv2" - openapi3 "github.com/googleapis/gnostic/OpenAPIv3" + "log" + + "github.com/golang/protobuf/proto" + openapiv2 "github.com/googleapis/gnostic/OpenAPIv2" + openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" plugins "github.com/googleapis/gnostic/plugins" "github.com/googleapis/gnostic/printer" ) // generate a simple report of an OpenAPI document's contents -func printDocumentV2(code *printer.Code, document *openapi2.Document) { +func printDocumentV2(code *printer.Code, document *openapiv2.Document) { code.Print("Swagger: %+v", document.Swagger) code.Print("Host: %+v", document.Host) code.Print("BasePath: %+v", document.BasePath) @@ -57,7 +60,7 @@ func printDocumentV2(code *printer.Code, document *openapi2.Document) { } // generate a simple report of an OpenAPI document's contents -func printDocumentV3(code *printer.Code, document *openapi3.Document) { +func printDocumentV3(code *printer.Code, document *openapiv3.Document) { code.Print("OpenAPI: %+v", document.Openapi) code.Print("Servers: %+v", document.Servers) if document.Info != nil { @@ -92,20 +95,28 @@ func printDocumentV3(code *printer.Code, document *openapi3.Document) { func main() { env, err := plugins.NewEnvironment() env.RespondAndExitIfError(err) - code := &printer.Code{} - switch { - case env.Request.Openapi2 != nil: - printDocumentV2(code, env.Request.Openapi2) - case env.Request.Openapi3 != nil: - printDocumentV3(code, env.Request.Openapi3) - default: + for _, model := range env.Request.Models { + log.Printf("model %s", model.TypeUrl) + switch model.TypeUrl { + case "openapi.v2.Document": + documentv2 := &openapiv2.Document{} + err = proto.Unmarshal(model.Value, documentv2) + if err == nil { + printDocumentV2(code, documentv2) + } + case "openapi.v3.Document": + documentv3 := &openapiv3.Document{} + err = proto.Unmarshal(model.Value, documentv3) + if err == nil { + printDocumentV3(code, documentv3) + } + } } file := &plugins.File{ Name: "summary.txt", Data: []byte(code.String()), } env.Response.Files = append(env.Response.Files, file) - env.RespondAndExit() } diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Makefile b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Makefile index fcd9ea1bc..4e492d00b 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Makefile +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Makefile @@ -1,11 +1,13 @@ +BINDIR=.build/debug + all: swift build - cp .build/debug/gnostic-swift-generator gnostic-swift-generator - rm -f gnostic-swift-client gnostic-swift-server - ln -s gnostic-swift-generator gnostic-swift-client - ln -s gnostic-swift-generator gnostic-swift-server + +install: all + cp $(BINDIR)/gnostic-swift-generator $(GOPATH)/bin/gnostic-swift-generator + cp $(BINDIR)/gnostic-swift-generator $(GOPATH)/bin/gnostic-swift-client + cp $(BINDIR)/gnostic-swift-generator $(GOPATH)/bin/gnostic-swift-server clean: rm -rf .build Packages - rm -rf gnostic-swift-client gnostic-swift-server gnostic-swift-generator diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/README.md b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/README.md index afa873980..fc4ef6fa5 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/README.md +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/README.md @@ -1,10 +1,10 @@ # OpenAPI Swift Generator Plugin -This directory contains an `openapic` plugin that can be used to generate a Swift client library and scaffolding for a Swift server for an API with an OpenAPI description. +This directory contains a `gnostic` plugin that can be used to generate a Swift client library and scaffolding for a Swift server for an API with an OpenAPI description. The plugin can be invoked like this: - openapic bookstore.json --swift_generator_out=Bookstore + gnostic bookstore.json --swift-generator-out=Bookstore Where `Bookstore` is the name of a directory where the generated code will be written. diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/Gnostic/plugin.pb.swift b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/Gnostic/plugin.pb.swift index 9e19a082f..d0602027a 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/Gnostic/plugin.pb.swift +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/Gnostic/plugin.pb.swift @@ -171,45 +171,11 @@ public struct Gnostic_Plugin_V1_Request: SwiftProtobuf.Message { /// Clears the value of `compilerVersion`. Subsequent reads from it will return its default value. public mutating func clearCompilerVersion() {_storage._compilerVersion = nil} - /// OpenAPI v2 API representation - public var openapi2: Openapi_V2_Document { - get {return _storage._openapi2 ?? Openapi_V2_Document()} - set {_uniqueStorage()._openapi2 = newValue} + /// API models + public var models: [SwiftProtobuf.Google_Protobuf_Any] { + get {return _storage._models} + set {_uniqueStorage()._models = newValue} } - /// Returns true if `openapi2` has been explicitly set. - public var hasOpenapi2: Bool {return _storage._openapi2 != nil} - /// Clears the value of `openapi2`. Subsequent reads from it will return its default value. - public mutating func clearOpenapi2() {_storage._openapi2 = nil} - - /// OpenAPI v3 API representation - public var openapi3: Openapi_V3_Document { - get {return _storage._openapi3 ?? Openapi_V3_Document()} - set {_uniqueStorage()._openapi3 = newValue} - } - /// Returns true if `openapi3` has been explicitly set. - public var hasOpenapi3: Bool {return _storage._openapi3 != nil} - /// Clears the value of `openapi3`. Subsequent reads from it will return its default value. - public mutating func clearOpenapi3() {_storage._openapi3 = nil} - - /// Discovery API representation - public var discovery: Discovery_V1_Document { - get {return _storage._discovery ?? Discovery_V1_Document()} - set {_uniqueStorage()._discovery = newValue} - } - /// Returns true if `discovery` has been explicitly set. - public var hasDiscovery: Bool {return _storage._discovery != nil} - /// Clears the value of `discovery`. Subsequent reads from it will return its default value. - public mutating func clearDiscovery() {_storage._discovery = nil} - - /// generated code surface representation - public var surface: Surface_V1_Model { - get {return _storage._surface ?? Surface_V1_Model()} - set {_uniqueStorage()._surface = newValue} - } - /// Returns true if `surface` has been explicitly set. - public var hasSurface: Bool {return _storage._surface != nil} - /// Clears the value of `surface`. Subsequent reads from it will return its default value. - public mutating func clearSurface() {_storage._surface = nil} public var unknownFields = SwiftProtobuf.UnknownStorage() @@ -228,10 +194,7 @@ public struct Gnostic_Plugin_V1_Request: SwiftProtobuf.Message { case 2: try decoder.decodeSingularStringField(value: &_storage._outputPath) case 3: try decoder.decodeRepeatedMessageField(value: &_storage._parameters) case 4: try decoder.decodeSingularMessageField(value: &_storage._compilerVersion) - case 5: try decoder.decodeSingularMessageField(value: &_storage._openapi2) - case 6: try decoder.decodeSingularMessageField(value: &_storage._openapi3) - case 7: try decoder.decodeSingularMessageField(value: &_storage._discovery) - case 8: try decoder.decodeSingularMessageField(value: &_storage._surface) + case 5: try decoder.decodeRepeatedMessageField(value: &_storage._models) default: break } } @@ -256,17 +219,8 @@ public struct Gnostic_Plugin_V1_Request: SwiftProtobuf.Message { if let v = _storage._compilerVersion { try visitor.visitSingularMessageField(value: v, fieldNumber: 4) } - if let v = _storage._openapi2 { - try visitor.visitSingularMessageField(value: v, fieldNumber: 5) - } - if let v = _storage._openapi3 { - try visitor.visitSingularMessageField(value: v, fieldNumber: 6) - } - if let v = _storage._discovery { - try visitor.visitSingularMessageField(value: v, fieldNumber: 7) - } - if let v = _storage._surface { - try visitor.visitSingularMessageField(value: v, fieldNumber: 8) + if !_storage._models.isEmpty { + try visitor.visitRepeatedMessageField(value: _storage._models, fieldNumber: 5) } } try unknownFields.traverse(visitor: &visitor) @@ -284,13 +238,13 @@ public struct Gnostic_Plugin_V1_Response: SwiftProtobuf.Message { /// even if it reports an error in this way. /// /// This should be used to indicate errors which prevent the plugin from - /// operating as intended. Errors which indicate a problem in openapic + /// operating as intended. Errors which indicate a problem in gnostic /// itself -- such as the input Document being unparseable -- should be /// reported by writing a message to stderr and exiting with a non-zero /// status code. public var errors: [String] = [] - /// file output, each file will be written by openapic to an appropriate location. + /// file output, each file will be written by gnostic to an appropriate location. public var files: [Gnostic_Plugin_V1_File] = [] public var unknownFields = SwiftProtobuf.UnknownStorage() @@ -411,10 +365,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S 2: .standard(proto: "output_path"), 3: .same(proto: "parameters"), 4: .standard(proto: "compiler_version"), - 5: .same(proto: "openapi2"), - 6: .same(proto: "openapi3"), - 7: .same(proto: "discovery"), - 8: .same(proto: "surface"), + 5: .same(proto: "models"), ] fileprivate class _StorageClass { @@ -422,10 +373,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S var _outputPath: String = String() var _parameters: [Gnostic_Plugin_V1_Parameter] = [] var _compilerVersion: Gnostic_Plugin_V1_Version? = nil - var _openapi2: Openapi_V2_Document? = nil - var _openapi3: Openapi_V3_Document? = nil - var _discovery: Discovery_V1_Document? = nil - var _surface: Surface_V1_Model? = nil + var _models: [SwiftProtobuf.Google_Protobuf_Any] = [] static let defaultInstance = _StorageClass() @@ -436,10 +384,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S _outputPath = source._outputPath _parameters = source._parameters _compilerVersion = source._compilerVersion - _openapi2 = source._openapi2 - _openapi3 = source._openapi3 - _discovery = source._discovery - _surface = source._surface + _models = source._models } } @@ -457,10 +402,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S if _storage._outputPath != other_storage._outputPath {return false} if _storage._parameters != other_storage._parameters {return false} if _storage._compilerVersion != other_storage._compilerVersion {return false} - if _storage._openapi2 != other_storage._openapi2 {return false} - if _storage._openapi3 != other_storage._openapi3 {return false} - if _storage._discovery != other_storage._discovery {return false} - if _storage._surface != other_storage._surface {return false} + if _storage._models != other_storage._models {return false} return true } if !storagesAreEqual {return false} diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/gnostic-swift-generator/main.swift b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/gnostic-swift-generator/main.swift index a6e5bfb55..5cf9d9dcf 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/gnostic-swift-generator/main.swift +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/Sources/gnostic-swift-generator/main.swift @@ -26,15 +26,24 @@ func main() throws { let request = try Gnostic_Plugin_V1_Request(serializedData:rawRequest) var response = Gnostic_Plugin_V1_Response() + + var openapiv2 : Openapi_V2_Document? + var surface : Surface_V1_Model? + + for model in request.models { + if model.typeURL == "openapi.v2.Document" { + openapiv2 = try Openapi_V2_Document(serializedData: model.value) + } else if model.typeURL == "surface.v1.Model" { + surface = try Surface_V1_Model(serializedData: model.value) + } + } + - if request.hasOpenapi2 && request.hasSurface { - let document = request.openapi2 - let surface = request.surface - - Log("\(request.surface)") - + if let openapiv2 = openapiv2, + let surface = surface { + // build the service renderer - let renderer = ServiceRenderer(surface:surface, document:document) + let renderer = ServiceRenderer(surface:surface, document:openapiv2) // generate the desired files var filenames : [String] diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/Makefile b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/Makefile index 6d145622d..13ba05359 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/Makefile +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/Makefile @@ -1,7 +1,5 @@ all: - rm -f gnostic-swift-generator - ln -s ../../gnostic-swift-generator gnostic bookstore.json --swift-generator-out=Sources/Bookstore swift build diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/README.md b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/README.md index ee9fb3bad..7a719a6cf 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/README.md +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-generator/examples/bookstore/README.md @@ -2,11 +2,11 @@ This directory contains an OpenAPI description of a simple bookstore API. -Use this example to try the `openapi_swift_generator` plugin, which +Use this example to try the `gnostic-swift-generator` plugin, which generates Swift code that implements an API client and server for an OpenAPI description. -Run `make all` to build and install `openapic` and the Swift plugin. +Run `make all` to build and install `gnostic` and the Swift plugin. It will generate both client and server code. The API client and server code will be in the `Sources/Bookstore` package. diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Makefile b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Makefile index 8b67d8d33..7d36c923e 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Makefile +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Makefile @@ -1,7 +1,14 @@ +TGT=gnostic-swift-sample + +BINDIR=.build/debug + all: swift build +install: all + cp $(BINDIR)/$(TGT) $(GOPATH)/bin/$(TGT) + clean : rm -rf Packages rm -rf .build diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/Gnostic/plugin.pb.swift b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/Gnostic/plugin.pb.swift index 9e19a082f..d0602027a 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/Gnostic/plugin.pb.swift +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/Gnostic/plugin.pb.swift @@ -171,45 +171,11 @@ public struct Gnostic_Plugin_V1_Request: SwiftProtobuf.Message { /// Clears the value of `compilerVersion`. Subsequent reads from it will return its default value. public mutating func clearCompilerVersion() {_storage._compilerVersion = nil} - /// OpenAPI v2 API representation - public var openapi2: Openapi_V2_Document { - get {return _storage._openapi2 ?? Openapi_V2_Document()} - set {_uniqueStorage()._openapi2 = newValue} + /// API models + public var models: [SwiftProtobuf.Google_Protobuf_Any] { + get {return _storage._models} + set {_uniqueStorage()._models = newValue} } - /// Returns true if `openapi2` has been explicitly set. - public var hasOpenapi2: Bool {return _storage._openapi2 != nil} - /// Clears the value of `openapi2`. Subsequent reads from it will return its default value. - public mutating func clearOpenapi2() {_storage._openapi2 = nil} - - /// OpenAPI v3 API representation - public var openapi3: Openapi_V3_Document { - get {return _storage._openapi3 ?? Openapi_V3_Document()} - set {_uniqueStorage()._openapi3 = newValue} - } - /// Returns true if `openapi3` has been explicitly set. - public var hasOpenapi3: Bool {return _storage._openapi3 != nil} - /// Clears the value of `openapi3`. Subsequent reads from it will return its default value. - public mutating func clearOpenapi3() {_storage._openapi3 = nil} - - /// Discovery API representation - public var discovery: Discovery_V1_Document { - get {return _storage._discovery ?? Discovery_V1_Document()} - set {_uniqueStorage()._discovery = newValue} - } - /// Returns true if `discovery` has been explicitly set. - public var hasDiscovery: Bool {return _storage._discovery != nil} - /// Clears the value of `discovery`. Subsequent reads from it will return its default value. - public mutating func clearDiscovery() {_storage._discovery = nil} - - /// generated code surface representation - public var surface: Surface_V1_Model { - get {return _storage._surface ?? Surface_V1_Model()} - set {_uniqueStorage()._surface = newValue} - } - /// Returns true if `surface` has been explicitly set. - public var hasSurface: Bool {return _storage._surface != nil} - /// Clears the value of `surface`. Subsequent reads from it will return its default value. - public mutating func clearSurface() {_storage._surface = nil} public var unknownFields = SwiftProtobuf.UnknownStorage() @@ -228,10 +194,7 @@ public struct Gnostic_Plugin_V1_Request: SwiftProtobuf.Message { case 2: try decoder.decodeSingularStringField(value: &_storage._outputPath) case 3: try decoder.decodeRepeatedMessageField(value: &_storage._parameters) case 4: try decoder.decodeSingularMessageField(value: &_storage._compilerVersion) - case 5: try decoder.decodeSingularMessageField(value: &_storage._openapi2) - case 6: try decoder.decodeSingularMessageField(value: &_storage._openapi3) - case 7: try decoder.decodeSingularMessageField(value: &_storage._discovery) - case 8: try decoder.decodeSingularMessageField(value: &_storage._surface) + case 5: try decoder.decodeRepeatedMessageField(value: &_storage._models) default: break } } @@ -256,17 +219,8 @@ public struct Gnostic_Plugin_V1_Request: SwiftProtobuf.Message { if let v = _storage._compilerVersion { try visitor.visitSingularMessageField(value: v, fieldNumber: 4) } - if let v = _storage._openapi2 { - try visitor.visitSingularMessageField(value: v, fieldNumber: 5) - } - if let v = _storage._openapi3 { - try visitor.visitSingularMessageField(value: v, fieldNumber: 6) - } - if let v = _storage._discovery { - try visitor.visitSingularMessageField(value: v, fieldNumber: 7) - } - if let v = _storage._surface { - try visitor.visitSingularMessageField(value: v, fieldNumber: 8) + if !_storage._models.isEmpty { + try visitor.visitRepeatedMessageField(value: _storage._models, fieldNumber: 5) } } try unknownFields.traverse(visitor: &visitor) @@ -284,13 +238,13 @@ public struct Gnostic_Plugin_V1_Response: SwiftProtobuf.Message { /// even if it reports an error in this way. /// /// This should be used to indicate errors which prevent the plugin from - /// operating as intended. Errors which indicate a problem in openapic + /// operating as intended. Errors which indicate a problem in gnostic /// itself -- such as the input Document being unparseable -- should be /// reported by writing a message to stderr and exiting with a non-zero /// status code. public var errors: [String] = [] - /// file output, each file will be written by openapic to an appropriate location. + /// file output, each file will be written by gnostic to an appropriate location. public var files: [Gnostic_Plugin_V1_File] = [] public var unknownFields = SwiftProtobuf.UnknownStorage() @@ -411,10 +365,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S 2: .standard(proto: "output_path"), 3: .same(proto: "parameters"), 4: .standard(proto: "compiler_version"), - 5: .same(proto: "openapi2"), - 6: .same(proto: "openapi3"), - 7: .same(proto: "discovery"), - 8: .same(proto: "surface"), + 5: .same(proto: "models"), ] fileprivate class _StorageClass { @@ -422,10 +373,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S var _outputPath: String = String() var _parameters: [Gnostic_Plugin_V1_Parameter] = [] var _compilerVersion: Gnostic_Plugin_V1_Version? = nil - var _openapi2: Openapi_V2_Document? = nil - var _openapi3: Openapi_V3_Document? = nil - var _discovery: Discovery_V1_Document? = nil - var _surface: Surface_V1_Model? = nil + var _models: [SwiftProtobuf.Google_Protobuf_Any] = [] static let defaultInstance = _StorageClass() @@ -436,10 +384,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S _outputPath = source._outputPath _parameters = source._parameters _compilerVersion = source._compilerVersion - _openapi2 = source._openapi2 - _openapi3 = source._openapi3 - _discovery = source._discovery - _surface = source._surface + _models = source._models } } @@ -457,10 +402,7 @@ extension Gnostic_Plugin_V1_Request: SwiftProtobuf._MessageImplementationBase, S if _storage._outputPath != other_storage._outputPath {return false} if _storage._parameters != other_storage._parameters {return false} if _storage._compilerVersion != other_storage._compilerVersion {return false} - if _storage._openapi2 != other_storage._openapi2 {return false} - if _storage._openapi3 != other_storage._openapi3 {return false} - if _storage._discovery != other_storage._discovery {return false} - if _storage._surface != other_storage._surface {return false} + if _storage._models != other_storage._models {return false} return true } if !storagesAreEqual {return false} diff --git a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/gnostic-swift-sample/main.swift b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/gnostic-swift-sample/main.swift index e7bec7a4a..037de5059 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/gnostic-swift-sample/main.swift +++ b/vendor/github.com/googleapis/gnostic/plugins/gnostic-swift-sample/Sources/gnostic-swift-sample/main.swift @@ -55,16 +55,19 @@ func main() throws { var response = Gnostic_Plugin_V1_Response() let rawRequest = try Stdin.readall() let request = try Gnostic_Plugin_V1_Request(serializedData: rawRequest) - if request.hasOpenapi2 { - let document = request.openapi2 - let report = printDocument(document:document, name:request.sourceName) - if let reportData = report.data(using:.utf8) { - var file = Gnostic_Plugin_V1_File() - file.name = "report.txt" - file.data = reportData - response.files.append(file) + for model in request.models { + if model.typeURL == "openapi.v2.Document" { + let document = try Openapi_V2_Document(serializedData: model.value) + let report = printDocument(document:document, name:request.sourceName) + if let reportData = report.data(using:.utf8) { + var file = Gnostic_Plugin_V1_File() + file.name = "report.txt" + file.data = reportData + response.files.append(file) + } } - } + } + let serializedResponse = try response.serializedData() Stdout.write(bytes: serializedResponse) } diff --git a/vendor/github.com/googleapis/gnostic/plugins/plugin.pb.go b/vendor/github.com/googleapis/gnostic/plugins/plugin.pb.go index 2cc3d0c5e..89094baa9 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/plugin.pb.go +++ b/vendor/github.com/googleapis/gnostic/plugins/plugin.pb.go @@ -1,17 +1,18 @@ -// Code generated by protoc-gen-go. -// source: plugin.proto -// DO NOT EDIT! +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: plugins/plugin.proto /* Package gnostic_plugin_v1 is a generated protocol buffer package. It is generated from these files: - plugin.proto + plugins/plugin.proto It has these top-level messages: Version Parameter Request + Message + Messages Response File */ @@ -20,10 +21,7 @@ package gnostic_plugin_v1 import proto "github.com/golang/protobuf/proto" import fmt "fmt" import math "math" -import openapi_v2 "github.com/googleapis/gnostic/OpenAPIv2" -import openapi_v3 "github.com/googleapis/gnostic/OpenAPIv3" -import discovery_v1 "github.com/googleapis/gnostic/discovery" -import surface_v1 "github.com/googleapis/gnostic/surface" +import google_protobuf "github.com/golang/protobuf/ptypes/any" // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal @@ -36,6 +34,36 @@ var _ = math.Inf // proto package needs to be updated. const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package +type Message_Level int32 + +const ( + Message_UNKNOWN Message_Level = 0 + Message_INFO Message_Level = 1 + Message_WARNING Message_Level = 2 + Message_ERROR Message_Level = 3 + Message_FATAL Message_Level = 4 +) + +var Message_Level_name = map[int32]string{ + 0: "UNKNOWN", + 1: "INFO", + 2: "WARNING", + 3: "ERROR", + 4: "FATAL", +} +var Message_Level_value = map[string]int32{ + "UNKNOWN": 0, + "INFO": 1, + "WARNING": 2, + "ERROR": 3, + "FATAL": 4, +} + +func (x Message_Level) String() string { + return proto.EnumName(Message_Level_name, int32(x)) +} +func (Message_Level) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{3, 0} } + // The version number of gnostic. type Version struct { Major int32 `protobuf:"varint,1,opt,name=major" json:"major,omitempty"` @@ -116,14 +144,8 @@ type Request struct { Parameters []*Parameter `protobuf:"bytes,3,rep,name=parameters" json:"parameters,omitempty"` // The version number of gnostic. CompilerVersion *Version `protobuf:"bytes,4,opt,name=compiler_version,json=compilerVersion" json:"compiler_version,omitempty"` - // OpenAPI v2 API representation - Openapi2 *openapi_v2.Document `protobuf:"bytes,5,opt,name=openapi2" json:"openapi2,omitempty"` - // OpenAPI v3 API representation - Openapi3 *openapi_v3.Document `protobuf:"bytes,6,opt,name=openapi3" json:"openapi3,omitempty"` - // Discovery API representation - Discovery *discovery_v1.Document `protobuf:"bytes,7,opt,name=discovery" json:"discovery,omitempty"` - // generated code surface representation - Surface *surface_v1.Model `protobuf:"bytes,8,opt,name=surface" json:"surface,omitempty"` + // API models + Models []*google_protobuf.Any `protobuf:"bytes,5,rep,name=models" json:"models,omitempty"` } func (m *Request) Reset() { *m = Request{} } @@ -159,30 +181,70 @@ func (m *Request) GetCompilerVersion() *Version { return nil } -func (m *Request) GetOpenapi2() *openapi_v2.Document { +func (m *Request) GetModels() []*google_protobuf.Any { if m != nil { - return m.Openapi2 + return m.Models } return nil } -func (m *Request) GetOpenapi3() *openapi_v3.Document { +// Plugins can return messages to be collated and reported by gnostic. +type Message struct { + // message severity + Level Message_Level `protobuf:"varint,1,opt,name=level,enum=gnostic.plugin.v1.Message_Level" json:"level,omitempty"` + // a unique message identifier + Code string `protobuf:"bytes,2,opt,name=code" json:"code,omitempty"` + // message text + Text string `protobuf:"bytes,3,opt,name=text" json:"text,omitempty"` + // an associated key path in an API description + Keys []string `protobuf:"bytes,4,rep,name=keys" json:"keys,omitempty"` +} + +func (m *Message) Reset() { *m = Message{} } +func (m *Message) String() string { return proto.CompactTextString(m) } +func (*Message) ProtoMessage() {} +func (*Message) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{3} } + +func (m *Message) GetLevel() Message_Level { if m != nil { - return m.Openapi3 + return m.Level + } + return Message_UNKNOWN +} + +func (m *Message) GetCode() string { + if m != nil { + return m.Code + } + return "" +} + +func (m *Message) GetText() string { + if m != nil { + return m.Text + } + return "" +} + +func (m *Message) GetKeys() []string { + if m != nil { + return m.Keys } return nil } -func (m *Request) GetDiscovery() *discovery_v1.Document { - if m != nil { - return m.Discovery - } - return nil +type Messages struct { + Messages []*Message `protobuf:"bytes,1,rep,name=messages" json:"messages,omitempty"` } -func (m *Request) GetSurface() *surface_v1.Model { +func (m *Messages) Reset() { *m = Messages{} } +func (m *Messages) String() string { return proto.CompactTextString(m) } +func (*Messages) ProtoMessage() {} +func (*Messages) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{4} } + +func (m *Messages) GetMessages() []*Message { if m != nil { - return m.Surface + return m.Messages } return nil } @@ -194,19 +256,21 @@ type Response struct { // even if it reports an error in this way. // // This should be used to indicate errors which prevent the plugin from - // operating as intended. Errors which indicate a problem in openapic + // operating as intended. Errors which indicate a problem in gnostic // itself -- such as the input Document being unparseable -- should be // reported by writing a message to stderr and exiting with a non-zero // status code. Errors []string `protobuf:"bytes,1,rep,name=errors" json:"errors,omitempty"` - // file output, each file will be written by openapic to an appropriate location. + // file output, each file will be written by gnostic to an appropriate location. Files []*File `protobuf:"bytes,2,rep,name=files" json:"files,omitempty"` + // informational messages to be collected and reported by gnostic. + Messages []*Message `protobuf:"bytes,3,rep,name=messages" json:"messages,omitempty"` } func (m *Response) Reset() { *m = Response{} } func (m *Response) String() string { return proto.CompactTextString(m) } func (*Response) ProtoMessage() {} -func (*Response) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{3} } +func (*Response) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{5} } func (m *Response) GetErrors() []string { if m != nil { @@ -222,6 +286,13 @@ func (m *Response) GetFiles() []*File { return nil } +func (m *Response) GetMessages() []*Message { + if m != nil { + return m.Messages + } + return nil +} + // File describes a file generated by a plugin. type File struct { // name of the file @@ -233,7 +304,7 @@ type File struct { func (m *File) Reset() { *m = File{} } func (m *File) String() string { return proto.CompactTextString(m) } func (*File) ProtoMessage() {} -func (*File) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{4} } +func (*File) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{6} } func (m *File) GetName() string { if m != nil { @@ -253,44 +324,49 @@ func init() { proto.RegisterType((*Version)(nil), "gnostic.plugin.v1.Version") proto.RegisterType((*Parameter)(nil), "gnostic.plugin.v1.Parameter") proto.RegisterType((*Request)(nil), "gnostic.plugin.v1.Request") + proto.RegisterType((*Message)(nil), "gnostic.plugin.v1.Message") + proto.RegisterType((*Messages)(nil), "gnostic.plugin.v1.Messages") proto.RegisterType((*Response)(nil), "gnostic.plugin.v1.Response") proto.RegisterType((*File)(nil), "gnostic.plugin.v1.File") + proto.RegisterEnum("gnostic.plugin.v1.Message_Level", Message_Level_name, Message_Level_value) } -func init() { proto.RegisterFile("plugin.proto", fileDescriptor0) } +func init() { proto.RegisterFile("plugins/plugin.proto", fileDescriptor0) } var fileDescriptor0 = []byte{ - // 499 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x92, 0x4f, 0x8f, 0xd3, 0x3c, - 0x10, 0xc6, 0xd5, 0xa6, 0x7f, 0xa7, 0xfb, 0xbe, 0xb0, 0xd6, 0x6a, 0xb1, 0x56, 0x48, 0x5b, 0xf5, - 0x42, 0x25, 0x84, 0x4b, 0x1b, 0x10, 0x17, 0x2e, 0xac, 0x80, 0x15, 0x07, 0x76, 0x83, 0x0f, 0x5c, - 0x2b, 0x6f, 0xea, 0xa6, 0x46, 0x49, 0x6c, 0x6c, 0x27, 0x82, 0x4f, 0xc2, 0x9d, 0x4f, 0x8a, 0x62, - 0x27, 0x4d, 0xc5, 0x56, 0x68, 0x4f, 0x9d, 0xe7, 0xd1, 0xfc, 0x9e, 0xb8, 0x33, 0x03, 0x27, 0x2a, - 0x2d, 0x12, 0x91, 0x13, 0xa5, 0xa5, 0x95, 0xe8, 0x34, 0xc9, 0xa5, 0xb1, 0x22, 0x26, 0xb5, 0x5b, - 0x2e, 0x2f, 0xde, 0x24, 0xc2, 0xee, 0x8a, 0x3b, 0x12, 0xcb, 0x6c, 0x91, 0x48, 0x99, 0xa4, 0x9c, - 0x29, 0x61, 0x16, 0x75, 0xe3, 0xe2, 0x56, 0xf1, 0xfc, 0x5d, 0xf4, 0xa9, 0x5c, 0xb5, 0x95, 0xcf, - 0x7a, 0x28, 0x18, 0xb6, 0xd5, 0xc3, 0xc0, 0x8d, 0x30, 0xb1, 0x2c, 0xb9, 0xfe, 0xd9, 0x56, 0x35, - 0x18, 0xfe, 0x1b, 0x34, 0x85, 0xde, 0xb2, 0x98, 0x37, 0xbf, 0x1e, 0x9a, 0xc5, 0x30, 0xfc, 0xca, - 0xb5, 0x11, 0x32, 0x47, 0x67, 0xd0, 0xcf, 0xd8, 0x37, 0xa9, 0x71, 0x67, 0xda, 0x99, 0xf7, 0xa9, - 0x17, 0xce, 0x15, 0xb9, 0xd4, 0xb8, 0x5b, 0xbb, 0x95, 0xa8, 0x5c, 0xc5, 0x6c, 0xbc, 0xc3, 0x81, - 0x77, 0x9d, 0x40, 0xe7, 0x30, 0x30, 0xc5, 0x76, 0x2b, 0x7e, 0xe0, 0xde, 0xb4, 0x33, 0x1f, 0xd3, - 0x5a, 0xcd, 0x5e, 0xc3, 0x38, 0x62, 0x9a, 0x65, 0xdc, 0x72, 0x8d, 0x10, 0xf4, 0x72, 0x96, 0x71, - 0xf7, 0x95, 0x31, 0x75, 0x75, 0x15, 0x57, 0xb2, 0xb4, 0xe0, 0xee, 0x23, 0x63, 0xea, 0xc5, 0xec, - 0x57, 0x00, 0x43, 0xca, 0xbf, 0x17, 0xdc, 0x58, 0x74, 0x09, 0x13, 0x23, 0x0b, 0x1d, 0xf3, 0xf5, - 0x01, 0x0c, 0xde, 0xba, 0xa9, 0x22, 0x2e, 0x61, 0x22, 0x0b, 0xab, 0x0a, 0xbb, 0x56, 0xcc, 0xee, - 0xea, 0x20, 0xf0, 0x56, 0xc4, 0xec, 0x0e, 0xbd, 0x05, 0x50, 0xcd, 0x23, 0x0c, 0x0e, 0xa6, 0xc1, - 0x7c, 0xb2, 0x7a, 0x4a, 0xee, 0x6d, 0x9c, 0xec, 0x5f, 0x4a, 0x0f, 0xfa, 0xd1, 0x07, 0x78, 0x1c, - 0xcb, 0x4c, 0x89, 0x94, 0xeb, 0x75, 0xe9, 0x07, 0xe6, 0xfe, 0xe4, 0x64, 0x75, 0x71, 0x24, 0xa3, - 0x1e, 0x29, 0x7d, 0xd4, 0x30, 0xcd, 0x8c, 0x5f, 0xc2, 0x48, 0x2a, 0x9e, 0x33, 0x25, 0x56, 0xb8, - 0xef, 0xf0, 0x33, 0x52, 0x1b, 0xa4, 0x5c, 0x91, 0xf7, 0x32, 0x2e, 0x32, 0x9e, 0x5b, 0xba, 0xef, - 0x3a, 0x20, 0x42, 0x3c, 0xf8, 0x9b, 0x08, 0xef, 0x13, 0x21, 0x7a, 0x05, 0xe3, 0xfd, 0x69, 0xe0, - 0xa1, 0x43, 0xce, 0x49, 0x7b, 0x2c, 0xe5, 0xb2, 0x85, 0xda, 0x46, 0xf4, 0x1c, 0x86, 0xf5, 0x65, - 0xe0, 0x91, 0x63, 0x4e, 0x49, 0x73, 0x29, 0xe5, 0x92, 0x7c, 0x96, 0x1b, 0x9e, 0xd2, 0xa6, 0x63, - 0xf6, 0x05, 0x46, 0x94, 0x1b, 0x25, 0x73, 0xc3, 0xab, 0xa5, 0x73, 0xad, 0xa5, 0x36, 0xb8, 0x33, - 0x0d, 0xaa, 0xa5, 0x7b, 0x85, 0x5e, 0x40, 0x7f, 0x2b, 0x52, 0x6e, 0x70, 0xd7, 0x8d, 0xfa, 0xc9, - 0x91, 0x31, 0x7d, 0x14, 0x29, 0xa7, 0xbe, 0x6b, 0x46, 0xa0, 0x57, 0xc9, 0xa3, 0xe7, 0x81, 0xa0, - 0xb7, 0x61, 0x96, 0xb9, 0xa5, 0x9e, 0x50, 0x57, 0x5f, 0x3d, 0x83, 0xff, 0xa5, 0x4e, 0xf6, 0xa1, - 0xe5, 0xf2, 0xea, 0xbf, 0x6b, 0x5f, 0x47, 0x2e, 0x3f, 0xea, 0xfc, 0xee, 0x06, 0xd7, 0x37, 0xb7, - 0x77, 0x03, 0x77, 0xe8, 0xe1, 0x9f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x40, 0x96, 0xff, 0x27, 0xeb, - 0x03, 0x00, 0x00, + // 539 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x52, 0xc1, 0x6e, 0xd3, 0x40, + 0x10, 0xc5, 0xb1, 0xdd, 0xc4, 0x13, 0x28, 0x66, 0x55, 0x81, 0xa9, 0x90, 0x1a, 0xf9, 0x42, 0x0e, + 0xe0, 0xa8, 0x41, 0xf4, 0xc4, 0x25, 0x91, 0x9a, 0xa8, 0xa2, 0x38, 0xd1, 0x0a, 0xe8, 0x31, 0xda, + 0x3a, 0x1b, 0xc7, 0x60, 0x7b, 0xcd, 0xee, 0x3a, 0x6a, 0x3e, 0x81, 0xdf, 0xe0, 0x4b, 0xf8, 0x32, + 0x84, 0x76, 0xd7, 0x89, 0x8a, 0x08, 0x07, 0x4e, 0x7e, 0xf3, 0xf4, 0xfc, 0x66, 0xe6, 0xed, 0xc0, + 0x49, 0x95, 0xd7, 0x69, 0x56, 0x8a, 0x81, 0xf9, 0x46, 0x15, 0x67, 0x92, 0xa1, 0x27, 0x69, 0xc9, + 0x84, 0xcc, 0x92, 0xa8, 0x61, 0x37, 0xe7, 0xa7, 0xcf, 0x53, 0xc6, 0xd2, 0x9c, 0x0e, 0xb4, 0xe0, + 0xb6, 0x5e, 0x0d, 0x48, 0xb9, 0x35, 0xea, 0x30, 0x81, 0xf6, 0x67, 0xca, 0x45, 0xc6, 0x4a, 0x74, + 0x02, 0x6e, 0x41, 0xbe, 0x30, 0x1e, 0x58, 0x3d, 0xab, 0xef, 0x62, 0x53, 0x68, 0x36, 0x2b, 0x19, + 0x0f, 0x5a, 0x0d, 0xab, 0x0a, 0xc5, 0x56, 0x44, 0x26, 0xeb, 0xc0, 0x36, 0xac, 0x2e, 0xd0, 0x53, + 0x38, 0x12, 0xf5, 0x6a, 0x95, 0xdd, 0x05, 0x4e, 0xcf, 0xea, 0x7b, 0xb8, 0xa9, 0xc2, 0xb7, 0xe0, + 0xcd, 0x09, 0x27, 0x05, 0x95, 0x94, 0x23, 0x04, 0x4e, 0x49, 0x0a, 0xaa, 0xbb, 0x78, 0x58, 0x63, + 0x65, 0xb7, 0x21, 0x79, 0x4d, 0x75, 0x13, 0x0f, 0x9b, 0x22, 0xfc, 0x65, 0x41, 0x1b, 0xd3, 0x6f, + 0x35, 0x15, 0x12, 0x9d, 0x41, 0x57, 0xb0, 0x9a, 0x27, 0x74, 0x71, 0xef, 0x67, 0x30, 0x54, 0xac, + 0x2c, 0xce, 0xa0, 0xcb, 0x6a, 0x59, 0xd5, 0x72, 0x51, 0x11, 0xb9, 0x6e, 0x8c, 0xc0, 0x50, 0x73, + 0x22, 0xd7, 0xe8, 0x1d, 0x40, 0xb5, 0x1b, 0x42, 0x04, 0x76, 0xcf, 0xee, 0x77, 0x87, 0x2f, 0xa2, + 0xbf, 0xc2, 0x8a, 0xf6, 0x93, 0xe2, 0x7b, 0x7a, 0x74, 0x09, 0x7e, 0xc2, 0x8a, 0x2a, 0xcb, 0x29, + 0x5f, 0x6c, 0x4c, 0x60, 0x7a, 0xc9, 0xee, 0xf0, 0xf4, 0x80, 0x47, 0x13, 0x29, 0x7e, 0xbc, 0xfb, + 0x67, 0x97, 0xf1, 0x2b, 0x38, 0x2a, 0xd8, 0x92, 0xe6, 0x22, 0x70, 0xf5, 0x00, 0x27, 0x91, 0x79, + 0x9a, 0x68, 0xf7, 0x34, 0xd1, 0xa8, 0xdc, 0xe2, 0x46, 0x13, 0xfe, 0xb4, 0xa0, 0xfd, 0x81, 0x0a, + 0x41, 0x52, 0x8a, 0x2e, 0xc0, 0xcd, 0xe9, 0x86, 0xe6, 0x7a, 0xf5, 0xe3, 0x61, 0xef, 0x40, 0xd7, + 0x46, 0x1a, 0x5d, 0x2b, 0x1d, 0x36, 0x72, 0x15, 0x77, 0xc2, 0x96, 0xbb, 0x64, 0x35, 0x56, 0x9c, + 0xa4, 0x77, 0x52, 0x3f, 0x9e, 0x87, 0x35, 0x56, 0xdc, 0x57, 0xba, 0x15, 0x81, 0xd3, 0xb3, 0x15, + 0xa7, 0x70, 0x38, 0x02, 0x57, 0x7b, 0xa1, 0x2e, 0xb4, 0x3f, 0xc5, 0xef, 0xe3, 0xd9, 0x4d, 0xec, + 0x3f, 0x40, 0x1d, 0x70, 0xae, 0xe2, 0xc9, 0xcc, 0xb7, 0x14, 0x7d, 0x33, 0xc2, 0xf1, 0x55, 0x3c, + 0xf5, 0x5b, 0xc8, 0x03, 0xf7, 0x12, 0xe3, 0x19, 0xf6, 0x6d, 0x05, 0x27, 0xa3, 0x8f, 0xa3, 0x6b, + 0xdf, 0x09, 0xc7, 0xd0, 0x69, 0xc6, 0x12, 0xe8, 0x02, 0x3a, 0x45, 0x83, 0x03, 0x4b, 0xaf, 0x7f, + 0xfa, 0xef, 0x2d, 0xf0, 0x5e, 0x1b, 0x7e, 0xb7, 0xa0, 0x83, 0xa9, 0xa8, 0x58, 0x29, 0xa8, 0xba, + 0x31, 0xca, 0x39, 0xe3, 0xc6, 0xc2, 0xc3, 0x4d, 0x85, 0x5e, 0x83, 0xbb, 0xca, 0x72, 0x2a, 0x82, + 0x96, 0x76, 0x7e, 0x76, 0xc0, 0x79, 0x92, 0xe5, 0x14, 0x1b, 0xd5, 0x1f, 0xb3, 0xd8, 0xff, 0x31, + 0x4b, 0x04, 0x8e, 0xb2, 0x39, 0x78, 0xc5, 0x08, 0x9c, 0x25, 0x91, 0x44, 0x47, 0xfd, 0x10, 0x6b, + 0x3c, 0x7e, 0x09, 0xc7, 0x8c, 0xa7, 0x7b, 0xeb, 0xcd, 0xf9, 0xf8, 0xd1, 0xd4, 0xe0, 0xb9, 0xee, + 0x32, 0xb7, 0x7e, 0xb4, 0xec, 0x69, 0x3c, 0xbb, 0x3d, 0xd2, 0x17, 0xf0, 0xe6, 0x77, 0x00, 0x00, + 0x00, 0xff, 0xff, 0x97, 0xa0, 0x65, 0xe7, 0xd5, 0x03, 0x00, 0x00, } diff --git a/vendor/github.com/googleapis/gnostic/plugins/plugin.proto b/vendor/github.com/googleapis/gnostic/plugins/plugin.proto index b546f36d7..078e1e091 100644 --- a/vendor/github.com/googleapis/gnostic/plugins/plugin.proto +++ b/vendor/github.com/googleapis/gnostic/plugins/plugin.proto @@ -22,10 +22,7 @@ syntax = "proto3"; -import "github.com/googleapis/gnostic/OpenAPIv2/OpenAPIv2.proto"; -import "github.com/googleapis/gnostic/OpenAPIv3/OpenAPIv3.proto"; -import "github.com/googleapis/gnostic/discovery/discovery.proto"; -import "github.com/googleapis/gnostic/surface/surface.proto"; +import "google/protobuf/any.proto"; package gnostic.plugin.v1; @@ -85,17 +82,36 @@ message Request { // The version number of gnostic. Version compiler_version = 4; - // OpenAPI v2 API representation - openapi.v2.Document openapi2 = 5; + // API models + repeated google.protobuf.Any models = 5; +} - // OpenAPI v3 API representation - openapi.v3.Document openapi3 = 6; +// Plugins can return messages to be collated and reported by gnostic. +message Message { - // Discovery API representation - discovery.v1.Document discovery = 7; - - // generated code surface representation - surface.v1.Model surface = 8; + enum Level { + UNKNOWN = 0; + INFO = 1; + WARNING = 2; + ERROR = 3; + FATAL = 4; + } + + // message severity + Level level = 1; + + // a unique message identifier + string code = 2; + + // message text + string text = 3; + + // an associated key path in an API description + repeated string keys = 4; +} + +message Messages { + repeated Message messages = 1; } // The plugin writes an encoded Response to stdout. @@ -106,14 +122,17 @@ message Response { // even if it reports an error in this way. // // This should be used to indicate errors which prevent the plugin from - // operating as intended. Errors which indicate a problem in openapic + // operating as intended. Errors which indicate a problem in gnostic // itself -- such as the input Document being unparseable -- should be // reported by writing a message to stderr and exiting with a non-zero // status code. repeated string errors = 1; - // file output, each file will be written by openapic to an appropriate location. + // file output, each file will be written by gnostic to an appropriate location. repeated File files = 2; + + // informational messages to be collected and reported by gnostic. + repeated Message messages = 3; } // File describes a file generated by a plugin. diff --git a/vendor/github.com/googleapis/gnostic/surface/model_openapiv2.go b/vendor/github.com/googleapis/gnostic/surface/model_openapiv2.go index 14c4eccf0..903d27ea7 100644 --- a/vendor/github.com/googleapis/gnostic/surface/model_openapiv2.go +++ b/vendor/github.com/googleapis/gnostic/surface/model_openapiv2.go @@ -99,7 +99,9 @@ func (b *OpenAPI2Builder) buildTypeFromDefinition(name string, schema *openapiv2 // If the schema has no fixed properties and additional properties of a specified type, // generate a map pointing to objects of that type. t.Kind = TypeKind_OBJECT - t.ContentType = typeForRef(schema.AdditionalProperties.GetSchema().XRef) + if schema.AdditionalProperties.GetSchema() != nil { + t.ContentType = typeForRef(schema.AdditionalProperties.GetSchema().XRef) + } } } return t, err diff --git a/vendor/github.com/googleapis/gnostic/surface/model_openapiv3.go b/vendor/github.com/googleapis/gnostic/surface/model_openapiv3.go index e0892cf1b..eb82b27e4 100644 --- a/vendor/github.com/googleapis/gnostic/surface/model_openapiv3.go +++ b/vendor/github.com/googleapis/gnostic/surface/model_openapiv3.go @@ -19,8 +19,9 @@ import ( "fmt" "log" - openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" "strings" + + openapiv3 "github.com/googleapis/gnostic/OpenAPIv3" ) // NewModelFromOpenAPIv3 builds a model of an API service for use in code generation. @@ -270,13 +271,14 @@ func (b *OpenAPI3Builder) typeForSchema(schema *openapiv3.Schema) (kind FieldKin case "array": if schema.Items != nil { // we have an array.., but of what? - items := schema.Items.SchemaOrReference - if len(items) == 1 { - if items[0].GetReference().GetXRef() != "" { - return FieldKind_ARRAY, typeForRef(items[0].GetReference().GetXRef()), format - } else if items[0].GetSchema().Type == "string" { + items := schema.Items + if items != nil { + a := items.GetSchemaOrReference() + if a[0].GetReference().GetXRef() != "" { + return FieldKind_ARRAY, typeForRef(a[0].GetReference().GetXRef()), format + } else if a[0].GetSchema().Type == "string" { return FieldKind_ARRAY, "string", format - } else if items[0].GetSchema().Type == "object" { + } else if a[0].GetSchema().Type == "object" { return FieldKind_ARRAY, "interface{}", format } } diff --git a/vendor/github.com/imdario/mergo/issue66_test.go b/vendor/github.com/imdario/mergo/issue66_test.go index 23fa5e271..9e4bccedc 100644 --- a/vendor/github.com/imdario/mergo/issue66_test.go +++ b/vendor/github.com/imdario/mergo/issue66_test.go @@ -20,6 +20,25 @@ func TestPrivateSlice(t *testing.T) { if err := Merge(&p1, p2); err != nil { t.Fatalf("Error during the merge: %v", err) } + if len(p1.PublicStrings) != 3 { + t.Error("5 elements should be in 'PublicStrings' field") + } + if len(p1.privateStrings) != 2 { + t.Error("2 elements should be in 'privateStrings' field") + } +} + +func TestPrivateSliceWithAppendSlice(t *testing.T) { + p1 := PrivateSliceTest66{ + PublicStrings: []string{"one", "two", "three"}, + privateStrings: []string{"four", "five"}, + } + p2 := PrivateSliceTest66{ + PublicStrings: []string{"six", "seven"}, + } + if err := Merge(&p1, p2, WithAppendSlice); err != nil { + t.Fatalf("Error during the merge: %v", err) + } if len(p1.PublicStrings) != 5 { t.Error("5 elements should be in 'PublicStrings' field") } diff --git a/vendor/github.com/imdario/mergo/merge.go b/vendor/github.com/imdario/mergo/merge.go index f0e17924a..706b22069 100644 --- a/vendor/github.com/imdario/mergo/merge.go +++ b/vendor/github.com/imdario/mergo/merge.go @@ -103,7 +103,15 @@ func deepMerge(dst, src reflect.Value, visited map[uintptr]*visit, depth int, co case reflect.Ptr: fallthrough case reflect.Map: - if err = deepMerge(dstElement, srcElement, visited, depth+1, config); err != nil { + srcMapElm := srcElement + dstMapElm := dstElement + if srcMapElm.CanInterface() { + srcMapElm = reflect.ValueOf(srcMapElm.Interface()) + if dstMapElm.IsValid() { + dstMapElm = reflect.ValueOf(dstMapElm.Interface()) + } + } + if err = deepMerge(dstMapElm, srcMapElm, visited, depth+1, config); err != nil { return } case reflect.Slice: @@ -116,7 +124,11 @@ func deepMerge(dst, src reflect.Value, visited map[uintptr]*visit, depth int, co dstSlice = reflect.ValueOf(dstElement.Interface()) } - dstSlice = reflect.AppendSlice(dstSlice, srcSlice) + if !isEmptyValue(src) && (overwrite || isEmptyValue(dst)) && !config.AppendSlice { + dstSlice = srcSlice + } else if config.AppendSlice { + dstSlice = reflect.AppendSlice(dstSlice, srcSlice) + } dst.SetMapIndex(key, dstSlice) } } @@ -124,7 +136,7 @@ func deepMerge(dst, src reflect.Value, visited map[uintptr]*visit, depth int, co continue } - if srcElement.IsValid() && (overwrite || (!dstElement.IsValid() || isEmptyValue(dst))) { + if srcElement.IsValid() && (overwrite || (!dstElement.IsValid() || isEmptyValue(dstElement))) { if dst.IsNil() { dst.Set(reflect.MakeMap(dst.Type())) } @@ -137,7 +149,7 @@ func deepMerge(dst, src reflect.Value, visited map[uintptr]*visit, depth int, co } if !isEmptyValue(src) && (overwrite || isEmptyValue(dst)) && !config.AppendSlice { dst.Set(src) - } else { + } else if config.AppendSlice { dst.Set(reflect.AppendSlice(dst, src)) } case reflect.Ptr: diff --git a/vendor/github.com/imdario/mergo/mergo.go b/vendor/github.com/imdario/mergo/mergo.go index 785618cd0..a82fea2fd 100644 --- a/vendor/github.com/imdario/mergo/mergo.go +++ b/vendor/github.com/imdario/mergo/mergo.go @@ -45,7 +45,12 @@ func isEmptyValue(v reflect.Value) bool { return v.Uint() == 0 case reflect.Float32, reflect.Float64: return v.Float() == 0 - case reflect.Interface, reflect.Ptr, reflect.Func: + case reflect.Interface, reflect.Ptr: + if v.IsNil() { + return true + } + return isEmptyValue(v.Elem()) + case reflect.Func: return v.IsNil() case reflect.Invalid: return true diff --git a/vendor/github.com/imdario/mergo/mergo_test.go b/vendor/github.com/imdario/mergo/mergo_test.go index 854513886..d77753843 100644 --- a/vendor/github.com/imdario/mergo/mergo_test.go +++ b/vendor/github.com/imdario/mergo/mergo_test.go @@ -6,11 +6,12 @@ package mergo import ( - "gopkg.in/yaml.v2" "io/ioutil" "reflect" "testing" "time" + + "gopkg.in/yaml.v2" ) type simpleTest struct { @@ -225,13 +226,13 @@ func TestPointerStructNil(t *testing.T) { } } -func testSlice(t *testing.T, a []int, b []int) { +func testSlice(t *testing.T, a []int, b []int, e []int, opts ...func(*Config)) { + t.Helper() bc := b - e := append(a, b...) sa := sliceTest{a} sb := sliceTest{b} - if err := Merge(&sa, sb); err != nil { + if err := Merge(&sa, sb, opts...); err != nil { t.FailNow() } if !reflect.DeepEqual(sb.S, bc) { @@ -243,14 +244,14 @@ func testSlice(t *testing.T, a []int, b []int) { ma := map[string][]int{"S": a} mb := map[string][]int{"S": b} - if err := Merge(&ma, mb); err != nil { + if err := Merge(&ma, mb, opts...); err != nil { t.FailNow() } if !reflect.DeepEqual(mb["S"], bc) { - t.Fatalf("Source slice was modified %d != %d", mb["S"], bc) + t.Fatalf("map value: Source slice was modified %d != %d", mb["S"], bc) } if !reflect.DeepEqual(ma["S"], e) { - t.Fatalf("b not merged in a proper way %d != %d", ma["S"], e) + t.Fatalf("map value: b not merged in a proper way %d != %d", ma["S"], e) } if a == nil { @@ -261,10 +262,10 @@ func testSlice(t *testing.T, a []int, b []int) { t.FailNow() } if !reflect.DeepEqual(mb["S"], bc) { - t.Fatalf("Source slice was modified %d != %d", mb["S"], bc) + t.Fatalf("missing dst key: Source slice was modified %d != %d", mb["S"], bc) } if !reflect.DeepEqual(ma["S"], e) { - t.Fatalf("b not merged in a proper way %d != %d", ma["S"], e) + t.Fatalf("missing dst key: b not merged in a proper way %d != %d", ma["S"], e) } } @@ -276,20 +277,25 @@ func testSlice(t *testing.T, a []int, b []int) { t.FailNow() } if !reflect.DeepEqual(mb["S"], bc) { - t.Fatalf("Source slice was modified %d != %d", mb["S"], bc) + t.Fatalf("missing src key: Source slice was modified %d != %d", mb["S"], bc) } if !reflect.DeepEqual(ma["S"], e) { - t.Fatalf("b not merged in a proper way %d != %d", ma["S"], e) + t.Fatalf("missing src key: b not merged in a proper way %d != %d", ma["S"], e) } } } func TestSlice(t *testing.T) { - testSlice(t, nil, []int{1, 2, 3}) - testSlice(t, []int{}, []int{1, 2, 3}) - testSlice(t, []int{1}, []int{2, 3}) - testSlice(t, []int{1}, []int{}) - testSlice(t, []int{1}, nil) + testSlice(t, nil, []int{1, 2, 3}, []int{1, 2, 3}) + testSlice(t, []int{}, []int{1, 2, 3}, []int{1, 2, 3}) + testSlice(t, []int{1}, []int{2, 3}, []int{1}) + testSlice(t, []int{1}, []int{}, []int{1}) + testSlice(t, []int{1}, nil, []int{1}) + testSlice(t, nil, []int{1, 2, 3}, []int{1, 2, 3}, WithAppendSlice) + testSlice(t, []int{}, []int{1, 2, 3}, []int{1, 2, 3}, WithAppendSlice) + testSlice(t, []int{1}, []int{2, 3}, []int{1, 2, 3}, WithAppendSlice) + testSlice(t, []int{1}, []int{}, []int{1}, WithAppendSlice) + testSlice(t, []int{1}, nil, []int{1}, WithAppendSlice) } func TestEmptyMaps(t *testing.T) { diff --git a/vendor/github.com/imdario/mergo/pr80_test.go b/vendor/github.com/imdario/mergo/pr80_test.go new file mode 100644 index 000000000..0b3220f3b --- /dev/null +++ b/vendor/github.com/imdario/mergo/pr80_test.go @@ -0,0 +1,18 @@ +package mergo + +import ( + "testing" +) + +type mapInterface map[string]interface{} + +func TestMergeMapsEmptyString(t *testing.T) { + a := mapInterface{"s": ""} + b := mapInterface{"s": "foo"} + if err := Merge(&a, b); err != nil { + t.Fatal(err) + } + if a["s"] != "foo" { + t.Fatalf("b not merged in properly: a.s.Value(%s) != expected(%s)", a["s"], "foo") + } +} diff --git a/vendor/github.com/imdario/mergo/pr81_test.go b/vendor/github.com/imdario/mergo/pr81_test.go new file mode 100644 index 000000000..e90e923fe --- /dev/null +++ b/vendor/github.com/imdario/mergo/pr81_test.go @@ -0,0 +1,42 @@ +package mergo + +import ( + "testing" +) + +func TestMapInterfaceWithMultipleLayer(t *testing.T) { + m1 := map[string]interface{}{ + "k1": map[string]interface{}{ + "k1.1": "v1", + }, + } + + m2 := map[string]interface{}{ + "k1": map[string]interface{}{ + "k1.1": "v2", + "k1.2": "v3", + }, + } + + if err := Map(&m1, m2, WithOverride); err != nil { + t.Fatalf("Error merging: %v", err) + } + + // Check overwrite of sub map works + expected := "v2" + actual := m1["k1"].(map[string]interface{})["k1.1"].(string) + if actual != expected { + t.Fatalf("Expected %v but got %v", + expected, + actual) + } + + // Check new key is merged + expected = "v3" + actual = m1["k1"].(map[string]interface{})["k1.2"].(string) + if actual != expected { + t.Fatalf("Expected %v but got %v", + expected, + actual) + } +} diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/.travis.yml b/vendor/github.com/matttproud/golang_protobuf_extensions/.travis.yml index f1309c9f8..5db258039 100644 --- a/vendor/github.com/matttproud/golang_protobuf_extensions/.travis.yml +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/.travis.yml @@ -1,2 +1,8 @@ language: go +go: + - 1.5 + - 1.6 + - tip + +script: make -f Makefile.TRAVIS diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/Makefile.TRAVIS b/vendor/github.com/matttproud/golang_protobuf_extensions/Makefile.TRAVIS new file mode 100644 index 000000000..24f9649e2 --- /dev/null +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/Makefile.TRAVIS @@ -0,0 +1,15 @@ +all: build cover test vet + +build: + go build -v ./... + +cover: test + $(MAKE) -C pbutil cover + +test: build + go test -v ./... + +vet: build + go vet -v ./... + +.PHONY: build cover test vet diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/.gitignore b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/.gitignore new file mode 100644 index 000000000..e16fb946b --- /dev/null +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/.gitignore @@ -0,0 +1 @@ +cover.dat diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/Makefile b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/Makefile new file mode 100644 index 000000000..81be21437 --- /dev/null +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/Makefile @@ -0,0 +1,7 @@ +all: + +cover: + go test -cover -v -coverprofile=cover.dat ./... + go tool cover -func cover.dat + +.PHONY: cover diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/all_test.go b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/all_test.go index 5c463722d..a793c8856 100644 --- a/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/all_test.go +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/all_test.go @@ -18,14 +18,15 @@ import ( "bytes" "testing" - . "github.com/golang/protobuf/proto" - . "github.com/golang/protobuf/proto/testdata" + "github.com/golang/protobuf/proto" + + . "github.com/matttproud/golang_protobuf_extensions/testdata" ) func TestWriteDelimited(t *testing.T) { t.Parallel() for _, test := range []struct { - msg Message + msg proto.Message buf []byte n int err error @@ -42,7 +43,7 @@ func TestWriteDelimited(t *testing.T) { }, { msg: &Strings{ - StringField: String(`This is my gigantic, unhappy string. It exceeds + StringField: proto.String(`This is my gigantic, unhappy string. It exceeds the encoding size of a single byte varint. We are using it to fuzz test the correctness of the header decoding mechanisms, which may prove problematic. I expect it may. Let's hope you enjoy testing as much as we do.`), @@ -82,7 +83,7 @@ func TestReadDelimited(t *testing.T) { t.Parallel() for _, test := range []struct { buf []byte - msg Message + msg proto.Message n int err error }{ @@ -116,7 +117,7 @@ func TestReadDelimited(t *testing.T) { 106, 111, 121, 32, 116, 101, 115, 116, 105, 110, 103, 32, 97, 115, 32, 109, 117, 99, 104, 32, 97, 115, 32, 119, 101, 32, 100, 111, 46}, msg: &Strings{ - StringField: String(`This is my gigantic, unhappy string. It exceeds + StringField: proto.String(`This is my gigantic, unhappy string. It exceeds the encoding size of a single byte varint. We are using it to fuzz test the correctness of the header decoding mechanisms, which may prove problematic. I expect it may. Let's hope you enjoy testing as much as we do.`), @@ -124,12 +125,12 @@ I expect it may. Let's hope you enjoy testing as much as we do.`), n: 271, }, } { - msg := Clone(test.msg) + msg := proto.Clone(test.msg) msg.Reset() if n, err := ReadDelimited(bytes.NewBuffer(test.buf), msg); n != test.n || err != test.err { t.Fatalf("ReadDelimited(%v, msg) = %v, %v; want %v, %v", test.buf, n, err, test.n, test.err) } - if !Equal(msg, test.msg) { + if !proto.Equal(msg, test.msg) { t.Fatalf("ReadDelimited(%v, msg); msg = %v; want %v", test.buf, msg, test.msg) } } @@ -137,12 +138,12 @@ I expect it may. Let's hope you enjoy testing as much as we do.`), func TestEndToEndValid(t *testing.T) { t.Parallel() - for _, test := range [][]Message{ + for _, test := range [][]proto.Message{ {&Empty{}}, {&GoEnum{Foo: FOO_FOO1.Enum()}, &Empty{}, &GoEnum{Foo: FOO_FOO1.Enum()}}, {&GoEnum{Foo: FOO_FOO1.Enum()}}, {&Strings{ - StringField: String(`This is my gigantic, unhappy string. It exceeds + StringField: proto.String(`This is my gigantic, unhappy string. It exceeds the encoding size of a single byte varint. We are using it to fuzz test the correctness of the header decoding mechanisms, which may prove problematic. I expect it may. Let's hope you enjoy testing as much as we do.`), @@ -161,12 +162,12 @@ I expect it may. Let's hope you enjoy testing as much as we do.`), } var read int for i, msg := range test { - out := Clone(msg) + out := proto.Clone(msg) out.Reset() n, _ := ReadDelimited(&buf, out) // Decide to do EOF checking? read += n - if !Equal(out, msg) { + if !proto.Equal(out, msg) { t.Fatalf("out = %v; want %v[%d] = %#v", out, test, i, msg) } } diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/fixtures_test.go b/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/fixtures_test.go deleted file mode 100644 index d6d9b2559..000000000 --- a/vendor/github.com/matttproud/golang_protobuf_extensions/pbutil/fixtures_test.go +++ /dev/null @@ -1,103 +0,0 @@ -// Copyright 2010 The Go Authors. All rights reserved. -// http://github.com/golang/protobuf/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package pbutil - -import ( - . "github.com/golang/protobuf/proto" - . "github.com/golang/protobuf/proto/testdata" -) - -// FROM https://github.com/golang/protobuf/blob/master/proto/all_test.go. - -func initGoTestField() *GoTestField { - f := new(GoTestField) - f.Label = String("label") - f.Type = String("type") - return f -} - -// These are all structurally equivalent but the tag numbers differ. -// (It's remarkable that required, optional, and repeated all have -// 8 letters.) -func initGoTest_RequiredGroup() *GoTest_RequiredGroup { - return &GoTest_RequiredGroup{ - RequiredField: String("required"), - } -} - -func initGoTest_OptionalGroup() *GoTest_OptionalGroup { - return &GoTest_OptionalGroup{ - RequiredField: String("optional"), - } -} - -func initGoTest_RepeatedGroup() *GoTest_RepeatedGroup { - return &GoTest_RepeatedGroup{ - RequiredField: String("repeated"), - } -} - -func initGoTest(setdefaults bool) *GoTest { - pb := new(GoTest) - if setdefaults { - pb.F_BoolDefaulted = Bool(Default_GoTest_F_BoolDefaulted) - pb.F_Int32Defaulted = Int32(Default_GoTest_F_Int32Defaulted) - pb.F_Int64Defaulted = Int64(Default_GoTest_F_Int64Defaulted) - pb.F_Fixed32Defaulted = Uint32(Default_GoTest_F_Fixed32Defaulted) - pb.F_Fixed64Defaulted = Uint64(Default_GoTest_F_Fixed64Defaulted) - pb.F_Uint32Defaulted = Uint32(Default_GoTest_F_Uint32Defaulted) - pb.F_Uint64Defaulted = Uint64(Default_GoTest_F_Uint64Defaulted) - pb.F_FloatDefaulted = Float32(Default_GoTest_F_FloatDefaulted) - pb.F_DoubleDefaulted = Float64(Default_GoTest_F_DoubleDefaulted) - pb.F_StringDefaulted = String(Default_GoTest_F_StringDefaulted) - pb.F_BytesDefaulted = Default_GoTest_F_BytesDefaulted - pb.F_Sint32Defaulted = Int32(Default_GoTest_F_Sint32Defaulted) - pb.F_Sint64Defaulted = Int64(Default_GoTest_F_Sint64Defaulted) - } - - pb.Kind = GoTest_TIME.Enum() - pb.RequiredField = initGoTestField() - pb.F_BoolRequired = Bool(true) - pb.F_Int32Required = Int32(3) - pb.F_Int64Required = Int64(6) - pb.F_Fixed32Required = Uint32(32) - pb.F_Fixed64Required = Uint64(64) - pb.F_Uint32Required = Uint32(3232) - pb.F_Uint64Required = Uint64(6464) - pb.F_FloatRequired = Float32(3232) - pb.F_DoubleRequired = Float64(6464) - pb.F_StringRequired = String("string") - pb.F_BytesRequired = []byte("bytes") - pb.F_Sint32Required = Int32(-32) - pb.F_Sint64Required = Int64(-64) - pb.Requiredgroup = initGoTest_RequiredGroup() - - return pb -} diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/README.THIRD_PARTY b/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/README.THIRD_PARTY new file mode 100644 index 000000000..0c1f84246 --- /dev/null +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/README.THIRD_PARTY @@ -0,0 +1,4 @@ +test.pb.go and test.proto are third-party data. + +SOURCE: https://github.com/golang/protobuf +REVISION: bf531ff1a004f24ee53329dfd5ce0b41bfdc17df diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/test.pb.go b/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/test.pb.go new file mode 100644 index 000000000..772adcb62 --- /dev/null +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/test.pb.go @@ -0,0 +1,4029 @@ +// Code generated by protoc-gen-go. +// source: test.proto +// DO NOT EDIT! + +/* +Package testdata is a generated protocol buffer package. + +It is generated from these files: + test.proto + +It has these top-level messages: + GoEnum + GoTestField + GoTest + GoSkipTest + NonPackedTest + PackedTest + MaxTag + OldMessage + NewMessage + InnerMessage + OtherMessage + RequiredInnerMessage + MyMessage + Ext + ComplexExtension + DefaultsMessage + MyMessageSet + Empty + MessageList + Strings + Defaults + SubDefaults + RepeatedEnum + MoreRepeated + GroupOld + GroupNew + FloatingPoint + MessageWithMap + Oneof + Communique +*/ +package testdata + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +const _ = proto.ProtoPackageIsVersion1 + +type FOO int32 + +const ( + FOO_FOO1 FOO = 1 +) + +var FOO_name = map[int32]string{ + 1: "FOO1", +} +var FOO_value = map[string]int32{ + "FOO1": 1, +} + +func (x FOO) Enum() *FOO { + p := new(FOO) + *p = x + return p +} +func (x FOO) String() string { + return proto.EnumName(FOO_name, int32(x)) +} +func (x *FOO) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(FOO_value, data, "FOO") + if err != nil { + return err + } + *x = FOO(value) + return nil +} +func (FOO) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{0} } + +// An enum, for completeness. +type GoTest_KIND int32 + +const ( + GoTest_VOID GoTest_KIND = 0 + // Basic types + GoTest_BOOL GoTest_KIND = 1 + GoTest_BYTES GoTest_KIND = 2 + GoTest_FINGERPRINT GoTest_KIND = 3 + GoTest_FLOAT GoTest_KIND = 4 + GoTest_INT GoTest_KIND = 5 + GoTest_STRING GoTest_KIND = 6 + GoTest_TIME GoTest_KIND = 7 + // Groupings + GoTest_TUPLE GoTest_KIND = 8 + GoTest_ARRAY GoTest_KIND = 9 + GoTest_MAP GoTest_KIND = 10 + // Table types + GoTest_TABLE GoTest_KIND = 11 + // Functions + GoTest_FUNCTION GoTest_KIND = 12 +) + +var GoTest_KIND_name = map[int32]string{ + 0: "VOID", + 1: "BOOL", + 2: "BYTES", + 3: "FINGERPRINT", + 4: "FLOAT", + 5: "INT", + 6: "STRING", + 7: "TIME", + 8: "TUPLE", + 9: "ARRAY", + 10: "MAP", + 11: "TABLE", + 12: "FUNCTION", +} +var GoTest_KIND_value = map[string]int32{ + "VOID": 0, + "BOOL": 1, + "BYTES": 2, + "FINGERPRINT": 3, + "FLOAT": 4, + "INT": 5, + "STRING": 6, + "TIME": 7, + "TUPLE": 8, + "ARRAY": 9, + "MAP": 10, + "TABLE": 11, + "FUNCTION": 12, +} + +func (x GoTest_KIND) Enum() *GoTest_KIND { + p := new(GoTest_KIND) + *p = x + return p +} +func (x GoTest_KIND) String() string { + return proto.EnumName(GoTest_KIND_name, int32(x)) +} +func (x *GoTest_KIND) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(GoTest_KIND_value, data, "GoTest_KIND") + if err != nil { + return err + } + *x = GoTest_KIND(value) + return nil +} +func (GoTest_KIND) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{2, 0} } + +type MyMessage_Color int32 + +const ( + MyMessage_RED MyMessage_Color = 0 + MyMessage_GREEN MyMessage_Color = 1 + MyMessage_BLUE MyMessage_Color = 2 +) + +var MyMessage_Color_name = map[int32]string{ + 0: "RED", + 1: "GREEN", + 2: "BLUE", +} +var MyMessage_Color_value = map[string]int32{ + "RED": 0, + "GREEN": 1, + "BLUE": 2, +} + +func (x MyMessage_Color) Enum() *MyMessage_Color { + p := new(MyMessage_Color) + *p = x + return p +} +func (x MyMessage_Color) String() string { + return proto.EnumName(MyMessage_Color_name, int32(x)) +} +func (x *MyMessage_Color) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(MyMessage_Color_value, data, "MyMessage_Color") + if err != nil { + return err + } + *x = MyMessage_Color(value) + return nil +} +func (MyMessage_Color) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{12, 0} } + +type DefaultsMessage_DefaultsEnum int32 + +const ( + DefaultsMessage_ZERO DefaultsMessage_DefaultsEnum = 0 + DefaultsMessage_ONE DefaultsMessage_DefaultsEnum = 1 + DefaultsMessage_TWO DefaultsMessage_DefaultsEnum = 2 +) + +var DefaultsMessage_DefaultsEnum_name = map[int32]string{ + 0: "ZERO", + 1: "ONE", + 2: "TWO", +} +var DefaultsMessage_DefaultsEnum_value = map[string]int32{ + "ZERO": 0, + "ONE": 1, + "TWO": 2, +} + +func (x DefaultsMessage_DefaultsEnum) Enum() *DefaultsMessage_DefaultsEnum { + p := new(DefaultsMessage_DefaultsEnum) + *p = x + return p +} +func (x DefaultsMessage_DefaultsEnum) String() string { + return proto.EnumName(DefaultsMessage_DefaultsEnum_name, int32(x)) +} +func (x *DefaultsMessage_DefaultsEnum) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(DefaultsMessage_DefaultsEnum_value, data, "DefaultsMessage_DefaultsEnum") + if err != nil { + return err + } + *x = DefaultsMessage_DefaultsEnum(value) + return nil +} +func (DefaultsMessage_DefaultsEnum) EnumDescriptor() ([]byte, []int) { + return fileDescriptor0, []int{15, 0} +} + +type Defaults_Color int32 + +const ( + Defaults_RED Defaults_Color = 0 + Defaults_GREEN Defaults_Color = 1 + Defaults_BLUE Defaults_Color = 2 +) + +var Defaults_Color_name = map[int32]string{ + 0: "RED", + 1: "GREEN", + 2: "BLUE", +} +var Defaults_Color_value = map[string]int32{ + "RED": 0, + "GREEN": 1, + "BLUE": 2, +} + +func (x Defaults_Color) Enum() *Defaults_Color { + p := new(Defaults_Color) + *p = x + return p +} +func (x Defaults_Color) String() string { + return proto.EnumName(Defaults_Color_name, int32(x)) +} +func (x *Defaults_Color) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(Defaults_Color_value, data, "Defaults_Color") + if err != nil { + return err + } + *x = Defaults_Color(value) + return nil +} +func (Defaults_Color) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{20, 0} } + +type RepeatedEnum_Color int32 + +const ( + RepeatedEnum_RED RepeatedEnum_Color = 1 +) + +var RepeatedEnum_Color_name = map[int32]string{ + 1: "RED", +} +var RepeatedEnum_Color_value = map[string]int32{ + "RED": 1, +} + +func (x RepeatedEnum_Color) Enum() *RepeatedEnum_Color { + p := new(RepeatedEnum_Color) + *p = x + return p +} +func (x RepeatedEnum_Color) String() string { + return proto.EnumName(RepeatedEnum_Color_name, int32(x)) +} +func (x *RepeatedEnum_Color) UnmarshalJSON(data []byte) error { + value, err := proto.UnmarshalJSONEnum(RepeatedEnum_Color_value, data, "RepeatedEnum_Color") + if err != nil { + return err + } + *x = RepeatedEnum_Color(value) + return nil +} +func (RepeatedEnum_Color) EnumDescriptor() ([]byte, []int) { return fileDescriptor0, []int{22, 0} } + +type GoEnum struct { + Foo *FOO `protobuf:"varint,1,req,name=foo,enum=testdata.FOO" json:"foo,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoEnum) Reset() { *m = GoEnum{} } +func (m *GoEnum) String() string { return proto.CompactTextString(m) } +func (*GoEnum) ProtoMessage() {} +func (*GoEnum) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{0} } + +func (m *GoEnum) GetFoo() FOO { + if m != nil && m.Foo != nil { + return *m.Foo + } + return FOO_FOO1 +} + +type GoTestField struct { + Label *string `protobuf:"bytes,1,req,name=Label,json=label" json:"Label,omitempty"` + Type *string `protobuf:"bytes,2,req,name=Type,json=type" json:"Type,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoTestField) Reset() { *m = GoTestField{} } +func (m *GoTestField) String() string { return proto.CompactTextString(m) } +func (*GoTestField) ProtoMessage() {} +func (*GoTestField) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{1} } + +func (m *GoTestField) GetLabel() string { + if m != nil && m.Label != nil { + return *m.Label + } + return "" +} + +func (m *GoTestField) GetType() string { + if m != nil && m.Type != nil { + return *m.Type + } + return "" +} + +type GoTest struct { + // Some typical parameters + Kind *GoTest_KIND `protobuf:"varint,1,req,name=Kind,json=kind,enum=testdata.GoTest_KIND" json:"Kind,omitempty"` + Table *string `protobuf:"bytes,2,opt,name=Table,json=table" json:"Table,omitempty"` + Param *int32 `protobuf:"varint,3,opt,name=Param,json=param" json:"Param,omitempty"` + // Required, repeated and optional foreign fields. + RequiredField *GoTestField `protobuf:"bytes,4,req,name=RequiredField,json=requiredField" json:"RequiredField,omitempty"` + RepeatedField []*GoTestField `protobuf:"bytes,5,rep,name=RepeatedField,json=repeatedField" json:"RepeatedField,omitempty"` + OptionalField *GoTestField `protobuf:"bytes,6,opt,name=OptionalField,json=optionalField" json:"OptionalField,omitempty"` + // Required fields of all basic types + F_BoolRequired *bool `protobuf:"varint,10,req,name=F_Bool_required,json=fBoolRequired" json:"F_Bool_required,omitempty"` + F_Int32Required *int32 `protobuf:"varint,11,req,name=F_Int32_required,json=fInt32Required" json:"F_Int32_required,omitempty"` + F_Int64Required *int64 `protobuf:"varint,12,req,name=F_Int64_required,json=fInt64Required" json:"F_Int64_required,omitempty"` + F_Fixed32Required *uint32 `protobuf:"fixed32,13,req,name=F_Fixed32_required,json=fFixed32Required" json:"F_Fixed32_required,omitempty"` + F_Fixed64Required *uint64 `protobuf:"fixed64,14,req,name=F_Fixed64_required,json=fFixed64Required" json:"F_Fixed64_required,omitempty"` + F_Uint32Required *uint32 `protobuf:"varint,15,req,name=F_Uint32_required,json=fUint32Required" json:"F_Uint32_required,omitempty"` + F_Uint64Required *uint64 `protobuf:"varint,16,req,name=F_Uint64_required,json=fUint64Required" json:"F_Uint64_required,omitempty"` + F_FloatRequired *float32 `protobuf:"fixed32,17,req,name=F_Float_required,json=fFloatRequired" json:"F_Float_required,omitempty"` + F_DoubleRequired *float64 `protobuf:"fixed64,18,req,name=F_Double_required,json=fDoubleRequired" json:"F_Double_required,omitempty"` + F_StringRequired *string `protobuf:"bytes,19,req,name=F_String_required,json=fStringRequired" json:"F_String_required,omitempty"` + F_BytesRequired []byte `protobuf:"bytes,101,req,name=F_Bytes_required,json=fBytesRequired" json:"F_Bytes_required,omitempty"` + F_Sint32Required *int32 `protobuf:"zigzag32,102,req,name=F_Sint32_required,json=fSint32Required" json:"F_Sint32_required,omitempty"` + F_Sint64Required *int64 `protobuf:"zigzag64,103,req,name=F_Sint64_required,json=fSint64Required" json:"F_Sint64_required,omitempty"` + // Repeated fields of all basic types + F_BoolRepeated []bool `protobuf:"varint,20,rep,name=F_Bool_repeated,json=fBoolRepeated" json:"F_Bool_repeated,omitempty"` + F_Int32Repeated []int32 `protobuf:"varint,21,rep,name=F_Int32_repeated,json=fInt32Repeated" json:"F_Int32_repeated,omitempty"` + F_Int64Repeated []int64 `protobuf:"varint,22,rep,name=F_Int64_repeated,json=fInt64Repeated" json:"F_Int64_repeated,omitempty"` + F_Fixed32Repeated []uint32 `protobuf:"fixed32,23,rep,name=F_Fixed32_repeated,json=fFixed32Repeated" json:"F_Fixed32_repeated,omitempty"` + F_Fixed64Repeated []uint64 `protobuf:"fixed64,24,rep,name=F_Fixed64_repeated,json=fFixed64Repeated" json:"F_Fixed64_repeated,omitempty"` + F_Uint32Repeated []uint32 `protobuf:"varint,25,rep,name=F_Uint32_repeated,json=fUint32Repeated" json:"F_Uint32_repeated,omitempty"` + F_Uint64Repeated []uint64 `protobuf:"varint,26,rep,name=F_Uint64_repeated,json=fUint64Repeated" json:"F_Uint64_repeated,omitempty"` + F_FloatRepeated []float32 `protobuf:"fixed32,27,rep,name=F_Float_repeated,json=fFloatRepeated" json:"F_Float_repeated,omitempty"` + F_DoubleRepeated []float64 `protobuf:"fixed64,28,rep,name=F_Double_repeated,json=fDoubleRepeated" json:"F_Double_repeated,omitempty"` + F_StringRepeated []string `protobuf:"bytes,29,rep,name=F_String_repeated,json=fStringRepeated" json:"F_String_repeated,omitempty"` + F_BytesRepeated [][]byte `protobuf:"bytes,201,rep,name=F_Bytes_repeated,json=fBytesRepeated" json:"F_Bytes_repeated,omitempty"` + F_Sint32Repeated []int32 `protobuf:"zigzag32,202,rep,name=F_Sint32_repeated,json=fSint32Repeated" json:"F_Sint32_repeated,omitempty"` + F_Sint64Repeated []int64 `protobuf:"zigzag64,203,rep,name=F_Sint64_repeated,json=fSint64Repeated" json:"F_Sint64_repeated,omitempty"` + // Optional fields of all basic types + F_BoolOptional *bool `protobuf:"varint,30,opt,name=F_Bool_optional,json=fBoolOptional" json:"F_Bool_optional,omitempty"` + F_Int32Optional *int32 `protobuf:"varint,31,opt,name=F_Int32_optional,json=fInt32Optional" json:"F_Int32_optional,omitempty"` + F_Int64Optional *int64 `protobuf:"varint,32,opt,name=F_Int64_optional,json=fInt64Optional" json:"F_Int64_optional,omitempty"` + F_Fixed32Optional *uint32 `protobuf:"fixed32,33,opt,name=F_Fixed32_optional,json=fFixed32Optional" json:"F_Fixed32_optional,omitempty"` + F_Fixed64Optional *uint64 `protobuf:"fixed64,34,opt,name=F_Fixed64_optional,json=fFixed64Optional" json:"F_Fixed64_optional,omitempty"` + F_Uint32Optional *uint32 `protobuf:"varint,35,opt,name=F_Uint32_optional,json=fUint32Optional" json:"F_Uint32_optional,omitempty"` + F_Uint64Optional *uint64 `protobuf:"varint,36,opt,name=F_Uint64_optional,json=fUint64Optional" json:"F_Uint64_optional,omitempty"` + F_FloatOptional *float32 `protobuf:"fixed32,37,opt,name=F_Float_optional,json=fFloatOptional" json:"F_Float_optional,omitempty"` + F_DoubleOptional *float64 `protobuf:"fixed64,38,opt,name=F_Double_optional,json=fDoubleOptional" json:"F_Double_optional,omitempty"` + F_StringOptional *string `protobuf:"bytes,39,opt,name=F_String_optional,json=fStringOptional" json:"F_String_optional,omitempty"` + F_BytesOptional []byte `protobuf:"bytes,301,opt,name=F_Bytes_optional,json=fBytesOptional" json:"F_Bytes_optional,omitempty"` + F_Sint32Optional *int32 `protobuf:"zigzag32,302,opt,name=F_Sint32_optional,json=fSint32Optional" json:"F_Sint32_optional,omitempty"` + F_Sint64Optional *int64 `protobuf:"zigzag64,303,opt,name=F_Sint64_optional,json=fSint64Optional" json:"F_Sint64_optional,omitempty"` + // Default-valued fields of all basic types + F_BoolDefaulted *bool `protobuf:"varint,40,opt,name=F_Bool_defaulted,json=fBoolDefaulted,def=1" json:"F_Bool_defaulted,omitempty"` + F_Int32Defaulted *int32 `protobuf:"varint,41,opt,name=F_Int32_defaulted,json=fInt32Defaulted,def=32" json:"F_Int32_defaulted,omitempty"` + F_Int64Defaulted *int64 `protobuf:"varint,42,opt,name=F_Int64_defaulted,json=fInt64Defaulted,def=64" json:"F_Int64_defaulted,omitempty"` + F_Fixed32Defaulted *uint32 `protobuf:"fixed32,43,opt,name=F_Fixed32_defaulted,json=fFixed32Defaulted,def=320" json:"F_Fixed32_defaulted,omitempty"` + F_Fixed64Defaulted *uint64 `protobuf:"fixed64,44,opt,name=F_Fixed64_defaulted,json=fFixed64Defaulted,def=640" json:"F_Fixed64_defaulted,omitempty"` + F_Uint32Defaulted *uint32 `protobuf:"varint,45,opt,name=F_Uint32_defaulted,json=fUint32Defaulted,def=3200" json:"F_Uint32_defaulted,omitempty"` + F_Uint64Defaulted *uint64 `protobuf:"varint,46,opt,name=F_Uint64_defaulted,json=fUint64Defaulted,def=6400" json:"F_Uint64_defaulted,omitempty"` + F_FloatDefaulted *float32 `protobuf:"fixed32,47,opt,name=F_Float_defaulted,json=fFloatDefaulted,def=314159" json:"F_Float_defaulted,omitempty"` + F_DoubleDefaulted *float64 `protobuf:"fixed64,48,opt,name=F_Double_defaulted,json=fDoubleDefaulted,def=271828" json:"F_Double_defaulted,omitempty"` + F_StringDefaulted *string `protobuf:"bytes,49,opt,name=F_String_defaulted,json=fStringDefaulted,def=hello, \"world!\"\n" json:"F_String_defaulted,omitempty"` + F_BytesDefaulted []byte `protobuf:"bytes,401,opt,name=F_Bytes_defaulted,json=fBytesDefaulted,def=Bignose" json:"F_Bytes_defaulted,omitempty"` + F_Sint32Defaulted *int32 `protobuf:"zigzag32,402,opt,name=F_Sint32_defaulted,json=fSint32Defaulted,def=-32" json:"F_Sint32_defaulted,omitempty"` + F_Sint64Defaulted *int64 `protobuf:"zigzag64,403,opt,name=F_Sint64_defaulted,json=fSint64Defaulted,def=-64" json:"F_Sint64_defaulted,omitempty"` + // Packed repeated fields (no string or bytes). + F_BoolRepeatedPacked []bool `protobuf:"varint,50,rep,packed,name=F_Bool_repeated_packed,json=fBoolRepeatedPacked" json:"F_Bool_repeated_packed,omitempty"` + F_Int32RepeatedPacked []int32 `protobuf:"varint,51,rep,packed,name=F_Int32_repeated_packed,json=fInt32RepeatedPacked" json:"F_Int32_repeated_packed,omitempty"` + F_Int64RepeatedPacked []int64 `protobuf:"varint,52,rep,packed,name=F_Int64_repeated_packed,json=fInt64RepeatedPacked" json:"F_Int64_repeated_packed,omitempty"` + F_Fixed32RepeatedPacked []uint32 `protobuf:"fixed32,53,rep,packed,name=F_Fixed32_repeated_packed,json=fFixed32RepeatedPacked" json:"F_Fixed32_repeated_packed,omitempty"` + F_Fixed64RepeatedPacked []uint64 `protobuf:"fixed64,54,rep,packed,name=F_Fixed64_repeated_packed,json=fFixed64RepeatedPacked" json:"F_Fixed64_repeated_packed,omitempty"` + F_Uint32RepeatedPacked []uint32 `protobuf:"varint,55,rep,packed,name=F_Uint32_repeated_packed,json=fUint32RepeatedPacked" json:"F_Uint32_repeated_packed,omitempty"` + F_Uint64RepeatedPacked []uint64 `protobuf:"varint,56,rep,packed,name=F_Uint64_repeated_packed,json=fUint64RepeatedPacked" json:"F_Uint64_repeated_packed,omitempty"` + F_FloatRepeatedPacked []float32 `protobuf:"fixed32,57,rep,packed,name=F_Float_repeated_packed,json=fFloatRepeatedPacked" json:"F_Float_repeated_packed,omitempty"` + F_DoubleRepeatedPacked []float64 `protobuf:"fixed64,58,rep,packed,name=F_Double_repeated_packed,json=fDoubleRepeatedPacked" json:"F_Double_repeated_packed,omitempty"` + F_Sint32RepeatedPacked []int32 `protobuf:"zigzag32,502,rep,packed,name=F_Sint32_repeated_packed,json=fSint32RepeatedPacked" json:"F_Sint32_repeated_packed,omitempty"` + F_Sint64RepeatedPacked []int64 `protobuf:"zigzag64,503,rep,packed,name=F_Sint64_repeated_packed,json=fSint64RepeatedPacked" json:"F_Sint64_repeated_packed,omitempty"` + Requiredgroup *GoTest_RequiredGroup `protobuf:"group,70,req,name=RequiredGroup,json=requiredgroup" json:"requiredgroup,omitempty"` + Repeatedgroup []*GoTest_RepeatedGroup `protobuf:"group,80,rep,name=RepeatedGroup,json=repeatedgroup" json:"repeatedgroup,omitempty"` + Optionalgroup *GoTest_OptionalGroup `protobuf:"group,90,opt,name=OptionalGroup,json=optionalgroup" json:"optionalgroup,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoTest) Reset() { *m = GoTest{} } +func (m *GoTest) String() string { return proto.CompactTextString(m) } +func (*GoTest) ProtoMessage() {} +func (*GoTest) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{2} } + +const Default_GoTest_F_BoolDefaulted bool = true +const Default_GoTest_F_Int32Defaulted int32 = 32 +const Default_GoTest_F_Int64Defaulted int64 = 64 +const Default_GoTest_F_Fixed32Defaulted uint32 = 320 +const Default_GoTest_F_Fixed64Defaulted uint64 = 640 +const Default_GoTest_F_Uint32Defaulted uint32 = 3200 +const Default_GoTest_F_Uint64Defaulted uint64 = 6400 +const Default_GoTest_F_FloatDefaulted float32 = 314159 +const Default_GoTest_F_DoubleDefaulted float64 = 271828 +const Default_GoTest_F_StringDefaulted string = "hello, \"world!\"\n" + +var Default_GoTest_F_BytesDefaulted []byte = []byte("Bignose") + +const Default_GoTest_F_Sint32Defaulted int32 = -32 +const Default_GoTest_F_Sint64Defaulted int64 = -64 + +func (m *GoTest) GetKind() GoTest_KIND { + if m != nil && m.Kind != nil { + return *m.Kind + } + return GoTest_VOID +} + +func (m *GoTest) GetTable() string { + if m != nil && m.Table != nil { + return *m.Table + } + return "" +} + +func (m *GoTest) GetParam() int32 { + if m != nil && m.Param != nil { + return *m.Param + } + return 0 +} + +func (m *GoTest) GetRequiredField() *GoTestField { + if m != nil { + return m.RequiredField + } + return nil +} + +func (m *GoTest) GetRepeatedField() []*GoTestField { + if m != nil { + return m.RepeatedField + } + return nil +} + +func (m *GoTest) GetOptionalField() *GoTestField { + if m != nil { + return m.OptionalField + } + return nil +} + +func (m *GoTest) GetF_BoolRequired() bool { + if m != nil && m.F_BoolRequired != nil { + return *m.F_BoolRequired + } + return false +} + +func (m *GoTest) GetF_Int32Required() int32 { + if m != nil && m.F_Int32Required != nil { + return *m.F_Int32Required + } + return 0 +} + +func (m *GoTest) GetF_Int64Required() int64 { + if m != nil && m.F_Int64Required != nil { + return *m.F_Int64Required + } + return 0 +} + +func (m *GoTest) GetF_Fixed32Required() uint32 { + if m != nil && m.F_Fixed32Required != nil { + return *m.F_Fixed32Required + } + return 0 +} + +func (m *GoTest) GetF_Fixed64Required() uint64 { + if m != nil && m.F_Fixed64Required != nil { + return *m.F_Fixed64Required + } + return 0 +} + +func (m *GoTest) GetF_Uint32Required() uint32 { + if m != nil && m.F_Uint32Required != nil { + return *m.F_Uint32Required + } + return 0 +} + +func (m *GoTest) GetF_Uint64Required() uint64 { + if m != nil && m.F_Uint64Required != nil { + return *m.F_Uint64Required + } + return 0 +} + +func (m *GoTest) GetF_FloatRequired() float32 { + if m != nil && m.F_FloatRequired != nil { + return *m.F_FloatRequired + } + return 0 +} + +func (m *GoTest) GetF_DoubleRequired() float64 { + if m != nil && m.F_DoubleRequired != nil { + return *m.F_DoubleRequired + } + return 0 +} + +func (m *GoTest) GetF_StringRequired() string { + if m != nil && m.F_StringRequired != nil { + return *m.F_StringRequired + } + return "" +} + +func (m *GoTest) GetF_BytesRequired() []byte { + if m != nil { + return m.F_BytesRequired + } + return nil +} + +func (m *GoTest) GetF_Sint32Required() int32 { + if m != nil && m.F_Sint32Required != nil { + return *m.F_Sint32Required + } + return 0 +} + +func (m *GoTest) GetF_Sint64Required() int64 { + if m != nil && m.F_Sint64Required != nil { + return *m.F_Sint64Required + } + return 0 +} + +func (m *GoTest) GetF_BoolRepeated() []bool { + if m != nil { + return m.F_BoolRepeated + } + return nil +} + +func (m *GoTest) GetF_Int32Repeated() []int32 { + if m != nil { + return m.F_Int32Repeated + } + return nil +} + +func (m *GoTest) GetF_Int64Repeated() []int64 { + if m != nil { + return m.F_Int64Repeated + } + return nil +} + +func (m *GoTest) GetF_Fixed32Repeated() []uint32 { + if m != nil { + return m.F_Fixed32Repeated + } + return nil +} + +func (m *GoTest) GetF_Fixed64Repeated() []uint64 { + if m != nil { + return m.F_Fixed64Repeated + } + return nil +} + +func (m *GoTest) GetF_Uint32Repeated() []uint32 { + if m != nil { + return m.F_Uint32Repeated + } + return nil +} + +func (m *GoTest) GetF_Uint64Repeated() []uint64 { + if m != nil { + return m.F_Uint64Repeated + } + return nil +} + +func (m *GoTest) GetF_FloatRepeated() []float32 { + if m != nil { + return m.F_FloatRepeated + } + return nil +} + +func (m *GoTest) GetF_DoubleRepeated() []float64 { + if m != nil { + return m.F_DoubleRepeated + } + return nil +} + +func (m *GoTest) GetF_StringRepeated() []string { + if m != nil { + return m.F_StringRepeated + } + return nil +} + +func (m *GoTest) GetF_BytesRepeated() [][]byte { + if m != nil { + return m.F_BytesRepeated + } + return nil +} + +func (m *GoTest) GetF_Sint32Repeated() []int32 { + if m != nil { + return m.F_Sint32Repeated + } + return nil +} + +func (m *GoTest) GetF_Sint64Repeated() []int64 { + if m != nil { + return m.F_Sint64Repeated + } + return nil +} + +func (m *GoTest) GetF_BoolOptional() bool { + if m != nil && m.F_BoolOptional != nil { + return *m.F_BoolOptional + } + return false +} + +func (m *GoTest) GetF_Int32Optional() int32 { + if m != nil && m.F_Int32Optional != nil { + return *m.F_Int32Optional + } + return 0 +} + +func (m *GoTest) GetF_Int64Optional() int64 { + if m != nil && m.F_Int64Optional != nil { + return *m.F_Int64Optional + } + return 0 +} + +func (m *GoTest) GetF_Fixed32Optional() uint32 { + if m != nil && m.F_Fixed32Optional != nil { + return *m.F_Fixed32Optional + } + return 0 +} + +func (m *GoTest) GetF_Fixed64Optional() uint64 { + if m != nil && m.F_Fixed64Optional != nil { + return *m.F_Fixed64Optional + } + return 0 +} + +func (m *GoTest) GetF_Uint32Optional() uint32 { + if m != nil && m.F_Uint32Optional != nil { + return *m.F_Uint32Optional + } + return 0 +} + +func (m *GoTest) GetF_Uint64Optional() uint64 { + if m != nil && m.F_Uint64Optional != nil { + return *m.F_Uint64Optional + } + return 0 +} + +func (m *GoTest) GetF_FloatOptional() float32 { + if m != nil && m.F_FloatOptional != nil { + return *m.F_FloatOptional + } + return 0 +} + +func (m *GoTest) GetF_DoubleOptional() float64 { + if m != nil && m.F_DoubleOptional != nil { + return *m.F_DoubleOptional + } + return 0 +} + +func (m *GoTest) GetF_StringOptional() string { + if m != nil && m.F_StringOptional != nil { + return *m.F_StringOptional + } + return "" +} + +func (m *GoTest) GetF_BytesOptional() []byte { + if m != nil { + return m.F_BytesOptional + } + return nil +} + +func (m *GoTest) GetF_Sint32Optional() int32 { + if m != nil && m.F_Sint32Optional != nil { + return *m.F_Sint32Optional + } + return 0 +} + +func (m *GoTest) GetF_Sint64Optional() int64 { + if m != nil && m.F_Sint64Optional != nil { + return *m.F_Sint64Optional + } + return 0 +} + +func (m *GoTest) GetF_BoolDefaulted() bool { + if m != nil && m.F_BoolDefaulted != nil { + return *m.F_BoolDefaulted + } + return Default_GoTest_F_BoolDefaulted +} + +func (m *GoTest) GetF_Int32Defaulted() int32 { + if m != nil && m.F_Int32Defaulted != nil { + return *m.F_Int32Defaulted + } + return Default_GoTest_F_Int32Defaulted +} + +func (m *GoTest) GetF_Int64Defaulted() int64 { + if m != nil && m.F_Int64Defaulted != nil { + return *m.F_Int64Defaulted + } + return Default_GoTest_F_Int64Defaulted +} + +func (m *GoTest) GetF_Fixed32Defaulted() uint32 { + if m != nil && m.F_Fixed32Defaulted != nil { + return *m.F_Fixed32Defaulted + } + return Default_GoTest_F_Fixed32Defaulted +} + +func (m *GoTest) GetF_Fixed64Defaulted() uint64 { + if m != nil && m.F_Fixed64Defaulted != nil { + return *m.F_Fixed64Defaulted + } + return Default_GoTest_F_Fixed64Defaulted +} + +func (m *GoTest) GetF_Uint32Defaulted() uint32 { + if m != nil && m.F_Uint32Defaulted != nil { + return *m.F_Uint32Defaulted + } + return Default_GoTest_F_Uint32Defaulted +} + +func (m *GoTest) GetF_Uint64Defaulted() uint64 { + if m != nil && m.F_Uint64Defaulted != nil { + return *m.F_Uint64Defaulted + } + return Default_GoTest_F_Uint64Defaulted +} + +func (m *GoTest) GetF_FloatDefaulted() float32 { + if m != nil && m.F_FloatDefaulted != nil { + return *m.F_FloatDefaulted + } + return Default_GoTest_F_FloatDefaulted +} + +func (m *GoTest) GetF_DoubleDefaulted() float64 { + if m != nil && m.F_DoubleDefaulted != nil { + return *m.F_DoubleDefaulted + } + return Default_GoTest_F_DoubleDefaulted +} + +func (m *GoTest) GetF_StringDefaulted() string { + if m != nil && m.F_StringDefaulted != nil { + return *m.F_StringDefaulted + } + return Default_GoTest_F_StringDefaulted +} + +func (m *GoTest) GetF_BytesDefaulted() []byte { + if m != nil && m.F_BytesDefaulted != nil { + return m.F_BytesDefaulted + } + return append([]byte(nil), Default_GoTest_F_BytesDefaulted...) +} + +func (m *GoTest) GetF_Sint32Defaulted() int32 { + if m != nil && m.F_Sint32Defaulted != nil { + return *m.F_Sint32Defaulted + } + return Default_GoTest_F_Sint32Defaulted +} + +func (m *GoTest) GetF_Sint64Defaulted() int64 { + if m != nil && m.F_Sint64Defaulted != nil { + return *m.F_Sint64Defaulted + } + return Default_GoTest_F_Sint64Defaulted +} + +func (m *GoTest) GetF_BoolRepeatedPacked() []bool { + if m != nil { + return m.F_BoolRepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Int32RepeatedPacked() []int32 { + if m != nil { + return m.F_Int32RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Int64RepeatedPacked() []int64 { + if m != nil { + return m.F_Int64RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Fixed32RepeatedPacked() []uint32 { + if m != nil { + return m.F_Fixed32RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Fixed64RepeatedPacked() []uint64 { + if m != nil { + return m.F_Fixed64RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Uint32RepeatedPacked() []uint32 { + if m != nil { + return m.F_Uint32RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Uint64RepeatedPacked() []uint64 { + if m != nil { + return m.F_Uint64RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_FloatRepeatedPacked() []float32 { + if m != nil { + return m.F_FloatRepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_DoubleRepeatedPacked() []float64 { + if m != nil { + return m.F_DoubleRepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Sint32RepeatedPacked() []int32 { + if m != nil { + return m.F_Sint32RepeatedPacked + } + return nil +} + +func (m *GoTest) GetF_Sint64RepeatedPacked() []int64 { + if m != nil { + return m.F_Sint64RepeatedPacked + } + return nil +} + +func (m *GoTest) GetRequiredgroup() *GoTest_RequiredGroup { + if m != nil { + return m.Requiredgroup + } + return nil +} + +func (m *GoTest) GetRepeatedgroup() []*GoTest_RepeatedGroup { + if m != nil { + return m.Repeatedgroup + } + return nil +} + +func (m *GoTest) GetOptionalgroup() *GoTest_OptionalGroup { + if m != nil { + return m.Optionalgroup + } + return nil +} + +// Required, repeated, and optional groups. +type GoTest_RequiredGroup struct { + RequiredField *string `protobuf:"bytes,71,req,name=RequiredField,json=requiredField" json:"RequiredField,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoTest_RequiredGroup) Reset() { *m = GoTest_RequiredGroup{} } +func (m *GoTest_RequiredGroup) String() string { return proto.CompactTextString(m) } +func (*GoTest_RequiredGroup) ProtoMessage() {} +func (*GoTest_RequiredGroup) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{2, 0} } + +func (m *GoTest_RequiredGroup) GetRequiredField() string { + if m != nil && m.RequiredField != nil { + return *m.RequiredField + } + return "" +} + +type GoTest_RepeatedGroup struct { + RequiredField *string `protobuf:"bytes,81,req,name=RequiredField,json=requiredField" json:"RequiredField,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoTest_RepeatedGroup) Reset() { *m = GoTest_RepeatedGroup{} } +func (m *GoTest_RepeatedGroup) String() string { return proto.CompactTextString(m) } +func (*GoTest_RepeatedGroup) ProtoMessage() {} +func (*GoTest_RepeatedGroup) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{2, 1} } + +func (m *GoTest_RepeatedGroup) GetRequiredField() string { + if m != nil && m.RequiredField != nil { + return *m.RequiredField + } + return "" +} + +type GoTest_OptionalGroup struct { + RequiredField *string `protobuf:"bytes,91,req,name=RequiredField,json=requiredField" json:"RequiredField,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoTest_OptionalGroup) Reset() { *m = GoTest_OptionalGroup{} } +func (m *GoTest_OptionalGroup) String() string { return proto.CompactTextString(m) } +func (*GoTest_OptionalGroup) ProtoMessage() {} +func (*GoTest_OptionalGroup) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{2, 2} } + +func (m *GoTest_OptionalGroup) GetRequiredField() string { + if m != nil && m.RequiredField != nil { + return *m.RequiredField + } + return "" +} + +// For testing skipping of unrecognized fields. +// Numbers are all big, larger than tag numbers in GoTestField, +// the message used in the corresponding test. +type GoSkipTest struct { + SkipInt32 *int32 `protobuf:"varint,11,req,name=skip_int32,json=skipInt32" json:"skip_int32,omitempty"` + SkipFixed32 *uint32 `protobuf:"fixed32,12,req,name=skip_fixed32,json=skipFixed32" json:"skip_fixed32,omitempty"` + SkipFixed64 *uint64 `protobuf:"fixed64,13,req,name=skip_fixed64,json=skipFixed64" json:"skip_fixed64,omitempty"` + SkipString *string `protobuf:"bytes,14,req,name=skip_string,json=skipString" json:"skip_string,omitempty"` + Skipgroup *GoSkipTest_SkipGroup `protobuf:"group,15,req,name=SkipGroup,json=skipgroup" json:"skipgroup,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoSkipTest) Reset() { *m = GoSkipTest{} } +func (m *GoSkipTest) String() string { return proto.CompactTextString(m) } +func (*GoSkipTest) ProtoMessage() {} +func (*GoSkipTest) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{3} } + +func (m *GoSkipTest) GetSkipInt32() int32 { + if m != nil && m.SkipInt32 != nil { + return *m.SkipInt32 + } + return 0 +} + +func (m *GoSkipTest) GetSkipFixed32() uint32 { + if m != nil && m.SkipFixed32 != nil { + return *m.SkipFixed32 + } + return 0 +} + +func (m *GoSkipTest) GetSkipFixed64() uint64 { + if m != nil && m.SkipFixed64 != nil { + return *m.SkipFixed64 + } + return 0 +} + +func (m *GoSkipTest) GetSkipString() string { + if m != nil && m.SkipString != nil { + return *m.SkipString + } + return "" +} + +func (m *GoSkipTest) GetSkipgroup() *GoSkipTest_SkipGroup { + if m != nil { + return m.Skipgroup + } + return nil +} + +type GoSkipTest_SkipGroup struct { + GroupInt32 *int32 `protobuf:"varint,16,req,name=group_int32,json=groupInt32" json:"group_int32,omitempty"` + GroupString *string `protobuf:"bytes,17,req,name=group_string,json=groupString" json:"group_string,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GoSkipTest_SkipGroup) Reset() { *m = GoSkipTest_SkipGroup{} } +func (m *GoSkipTest_SkipGroup) String() string { return proto.CompactTextString(m) } +func (*GoSkipTest_SkipGroup) ProtoMessage() {} +func (*GoSkipTest_SkipGroup) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{3, 0} } + +func (m *GoSkipTest_SkipGroup) GetGroupInt32() int32 { + if m != nil && m.GroupInt32 != nil { + return *m.GroupInt32 + } + return 0 +} + +func (m *GoSkipTest_SkipGroup) GetGroupString() string { + if m != nil && m.GroupString != nil { + return *m.GroupString + } + return "" +} + +// For testing packed/non-packed decoder switching. +// A serialized instance of one should be deserializable as the other. +type NonPackedTest struct { + A []int32 `protobuf:"varint,1,rep,name=a" json:"a,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *NonPackedTest) Reset() { *m = NonPackedTest{} } +func (m *NonPackedTest) String() string { return proto.CompactTextString(m) } +func (*NonPackedTest) ProtoMessage() {} +func (*NonPackedTest) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{4} } + +func (m *NonPackedTest) GetA() []int32 { + if m != nil { + return m.A + } + return nil +} + +type PackedTest struct { + B []int32 `protobuf:"varint,1,rep,packed,name=b" json:"b,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *PackedTest) Reset() { *m = PackedTest{} } +func (m *PackedTest) String() string { return proto.CompactTextString(m) } +func (*PackedTest) ProtoMessage() {} +func (*PackedTest) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{5} } + +func (m *PackedTest) GetB() []int32 { + if m != nil { + return m.B + } + return nil +} + +type MaxTag struct { + // Maximum possible tag number. + LastField *string `protobuf:"bytes,536870911,opt,name=last_field,json=lastField" json:"last_field,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MaxTag) Reset() { *m = MaxTag{} } +func (m *MaxTag) String() string { return proto.CompactTextString(m) } +func (*MaxTag) ProtoMessage() {} +func (*MaxTag) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{6} } + +func (m *MaxTag) GetLastField() string { + if m != nil && m.LastField != nil { + return *m.LastField + } + return "" +} + +type OldMessage struct { + Nested *OldMessage_Nested `protobuf:"bytes,1,opt,name=nested" json:"nested,omitempty"` + Num *int32 `protobuf:"varint,2,opt,name=num" json:"num,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *OldMessage) Reset() { *m = OldMessage{} } +func (m *OldMessage) String() string { return proto.CompactTextString(m) } +func (*OldMessage) ProtoMessage() {} +func (*OldMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{7} } + +func (m *OldMessage) GetNested() *OldMessage_Nested { + if m != nil { + return m.Nested + } + return nil +} + +func (m *OldMessage) GetNum() int32 { + if m != nil && m.Num != nil { + return *m.Num + } + return 0 +} + +type OldMessage_Nested struct { + Name *string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *OldMessage_Nested) Reset() { *m = OldMessage_Nested{} } +func (m *OldMessage_Nested) String() string { return proto.CompactTextString(m) } +func (*OldMessage_Nested) ProtoMessage() {} +func (*OldMessage_Nested) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{7, 0} } + +func (m *OldMessage_Nested) GetName() string { + if m != nil && m.Name != nil { + return *m.Name + } + return "" +} + +// NewMessage is wire compatible with OldMessage; +// imagine it as a future version. +type NewMessage struct { + Nested *NewMessage_Nested `protobuf:"bytes,1,opt,name=nested" json:"nested,omitempty"` + // This is an int32 in OldMessage. + Num *int64 `protobuf:"varint,2,opt,name=num" json:"num,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *NewMessage) Reset() { *m = NewMessage{} } +func (m *NewMessage) String() string { return proto.CompactTextString(m) } +func (*NewMessage) ProtoMessage() {} +func (*NewMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{8} } + +func (m *NewMessage) GetNested() *NewMessage_Nested { + if m != nil { + return m.Nested + } + return nil +} + +func (m *NewMessage) GetNum() int64 { + if m != nil && m.Num != nil { + return *m.Num + } + return 0 +} + +type NewMessage_Nested struct { + Name *string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + FoodGroup *string `protobuf:"bytes,2,opt,name=food_group,json=foodGroup" json:"food_group,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *NewMessage_Nested) Reset() { *m = NewMessage_Nested{} } +func (m *NewMessage_Nested) String() string { return proto.CompactTextString(m) } +func (*NewMessage_Nested) ProtoMessage() {} +func (*NewMessage_Nested) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{8, 0} } + +func (m *NewMessage_Nested) GetName() string { + if m != nil && m.Name != nil { + return *m.Name + } + return "" +} + +func (m *NewMessage_Nested) GetFoodGroup() string { + if m != nil && m.FoodGroup != nil { + return *m.FoodGroup + } + return "" +} + +type InnerMessage struct { + Host *string `protobuf:"bytes,1,req,name=host" json:"host,omitempty"` + Port *int32 `protobuf:"varint,2,opt,name=port,def=4000" json:"port,omitempty"` + Connected *bool `protobuf:"varint,3,opt,name=connected" json:"connected,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *InnerMessage) Reset() { *m = InnerMessage{} } +func (m *InnerMessage) String() string { return proto.CompactTextString(m) } +func (*InnerMessage) ProtoMessage() {} +func (*InnerMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{9} } + +const Default_InnerMessage_Port int32 = 4000 + +func (m *InnerMessage) GetHost() string { + if m != nil && m.Host != nil { + return *m.Host + } + return "" +} + +func (m *InnerMessage) GetPort() int32 { + if m != nil && m.Port != nil { + return *m.Port + } + return Default_InnerMessage_Port +} + +func (m *InnerMessage) GetConnected() bool { + if m != nil && m.Connected != nil { + return *m.Connected + } + return false +} + +type OtherMessage struct { + Key *int64 `protobuf:"varint,1,opt,name=key" json:"key,omitempty"` + Value []byte `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Weight *float32 `protobuf:"fixed32,3,opt,name=weight" json:"weight,omitempty"` + Inner *InnerMessage `protobuf:"bytes,4,opt,name=inner" json:"inner,omitempty"` + XXX_extensions map[int32]proto.Extension `json:"-"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *OtherMessage) Reset() { *m = OtherMessage{} } +func (m *OtherMessage) String() string { return proto.CompactTextString(m) } +func (*OtherMessage) ProtoMessage() {} +func (*OtherMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{10} } + +var extRange_OtherMessage = []proto.ExtensionRange{ + {100, 536870911}, +} + +func (*OtherMessage) ExtensionRangeArray() []proto.ExtensionRange { + return extRange_OtherMessage +} +func (m *OtherMessage) ExtensionMap() map[int32]proto.Extension { + if m.XXX_extensions == nil { + m.XXX_extensions = make(map[int32]proto.Extension) + } + return m.XXX_extensions +} + +func (m *OtherMessage) GetKey() int64 { + if m != nil && m.Key != nil { + return *m.Key + } + return 0 +} + +func (m *OtherMessage) GetValue() []byte { + if m != nil { + return m.Value + } + return nil +} + +func (m *OtherMessage) GetWeight() float32 { + if m != nil && m.Weight != nil { + return *m.Weight + } + return 0 +} + +func (m *OtherMessage) GetInner() *InnerMessage { + if m != nil { + return m.Inner + } + return nil +} + +type RequiredInnerMessage struct { + LeoFinallyWonAnOscar *InnerMessage `protobuf:"bytes,1,req,name=leo_finally_won_an_oscar,json=leoFinallyWonAnOscar" json:"leo_finally_won_an_oscar,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *RequiredInnerMessage) Reset() { *m = RequiredInnerMessage{} } +func (m *RequiredInnerMessage) String() string { return proto.CompactTextString(m) } +func (*RequiredInnerMessage) ProtoMessage() {} +func (*RequiredInnerMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{11} } + +func (m *RequiredInnerMessage) GetLeoFinallyWonAnOscar() *InnerMessage { + if m != nil { + return m.LeoFinallyWonAnOscar + } + return nil +} + +type MyMessage struct { + Count *int32 `protobuf:"varint,1,req,name=count" json:"count,omitempty"` + Name *string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Quote *string `protobuf:"bytes,3,opt,name=quote" json:"quote,omitempty"` + Pet []string `protobuf:"bytes,4,rep,name=pet" json:"pet,omitempty"` + Inner *InnerMessage `protobuf:"bytes,5,opt,name=inner" json:"inner,omitempty"` + Others []*OtherMessage `protobuf:"bytes,6,rep,name=others" json:"others,omitempty"` + WeMustGoDeeper *RequiredInnerMessage `protobuf:"bytes,13,opt,name=we_must_go_deeper,json=weMustGoDeeper" json:"we_must_go_deeper,omitempty"` + RepInner []*InnerMessage `protobuf:"bytes,12,rep,name=rep_inner,json=repInner" json:"rep_inner,omitempty"` + Bikeshed *MyMessage_Color `protobuf:"varint,7,opt,name=bikeshed,enum=testdata.MyMessage_Color" json:"bikeshed,omitempty"` + Somegroup *MyMessage_SomeGroup `protobuf:"group,8,opt,name=SomeGroup,json=somegroup" json:"somegroup,omitempty"` + // This field becomes [][]byte in the generated code. + RepBytes [][]byte `protobuf:"bytes,10,rep,name=rep_bytes,json=repBytes" json:"rep_bytes,omitempty"` + Bigfloat *float64 `protobuf:"fixed64,11,opt,name=bigfloat" json:"bigfloat,omitempty"` + XXX_extensions map[int32]proto.Extension `json:"-"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MyMessage) Reset() { *m = MyMessage{} } +func (m *MyMessage) String() string { return proto.CompactTextString(m) } +func (*MyMessage) ProtoMessage() {} +func (*MyMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{12} } + +var extRange_MyMessage = []proto.ExtensionRange{ + {100, 536870911}, +} + +func (*MyMessage) ExtensionRangeArray() []proto.ExtensionRange { + return extRange_MyMessage +} +func (m *MyMessage) ExtensionMap() map[int32]proto.Extension { + if m.XXX_extensions == nil { + m.XXX_extensions = make(map[int32]proto.Extension) + } + return m.XXX_extensions +} + +func (m *MyMessage) GetCount() int32 { + if m != nil && m.Count != nil { + return *m.Count + } + return 0 +} + +func (m *MyMessage) GetName() string { + if m != nil && m.Name != nil { + return *m.Name + } + return "" +} + +func (m *MyMessage) GetQuote() string { + if m != nil && m.Quote != nil { + return *m.Quote + } + return "" +} + +func (m *MyMessage) GetPet() []string { + if m != nil { + return m.Pet + } + return nil +} + +func (m *MyMessage) GetInner() *InnerMessage { + if m != nil { + return m.Inner + } + return nil +} + +func (m *MyMessage) GetOthers() []*OtherMessage { + if m != nil { + return m.Others + } + return nil +} + +func (m *MyMessage) GetWeMustGoDeeper() *RequiredInnerMessage { + if m != nil { + return m.WeMustGoDeeper + } + return nil +} + +func (m *MyMessage) GetRepInner() []*InnerMessage { + if m != nil { + return m.RepInner + } + return nil +} + +func (m *MyMessage) GetBikeshed() MyMessage_Color { + if m != nil && m.Bikeshed != nil { + return *m.Bikeshed + } + return MyMessage_RED +} + +func (m *MyMessage) GetSomegroup() *MyMessage_SomeGroup { + if m != nil { + return m.Somegroup + } + return nil +} + +func (m *MyMessage) GetRepBytes() [][]byte { + if m != nil { + return m.RepBytes + } + return nil +} + +func (m *MyMessage) GetBigfloat() float64 { + if m != nil && m.Bigfloat != nil { + return *m.Bigfloat + } + return 0 +} + +type MyMessage_SomeGroup struct { + GroupField *int32 `protobuf:"varint,9,opt,name=group_field,json=groupField" json:"group_field,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MyMessage_SomeGroup) Reset() { *m = MyMessage_SomeGroup{} } +func (m *MyMessage_SomeGroup) String() string { return proto.CompactTextString(m) } +func (*MyMessage_SomeGroup) ProtoMessage() {} +func (*MyMessage_SomeGroup) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{12, 0} } + +func (m *MyMessage_SomeGroup) GetGroupField() int32 { + if m != nil && m.GroupField != nil { + return *m.GroupField + } + return 0 +} + +type Ext struct { + Data *string `protobuf:"bytes,1,opt,name=data" json:"data,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *Ext) Reset() { *m = Ext{} } +func (m *Ext) String() string { return proto.CompactTextString(m) } +func (*Ext) ProtoMessage() {} +func (*Ext) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{13} } + +func (m *Ext) GetData() string { + if m != nil && m.Data != nil { + return *m.Data + } + return "" +} + +var E_Ext_More = &proto.ExtensionDesc{ + ExtendedType: (*MyMessage)(nil), + ExtensionType: (*Ext)(nil), + Field: 103, + Name: "testdata.Ext.more", + Tag: "bytes,103,opt,name=more", +} + +var E_Ext_Text = &proto.ExtensionDesc{ + ExtendedType: (*MyMessage)(nil), + ExtensionType: (*string)(nil), + Field: 104, + Name: "testdata.Ext.text", + Tag: "bytes,104,opt,name=text", +} + +var E_Ext_Number = &proto.ExtensionDesc{ + ExtendedType: (*MyMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 105, + Name: "testdata.Ext.number", + Tag: "varint,105,opt,name=number", +} + +type ComplexExtension struct { + First *int32 `protobuf:"varint,1,opt,name=first" json:"first,omitempty"` + Second *int32 `protobuf:"varint,2,opt,name=second" json:"second,omitempty"` + Third []int32 `protobuf:"varint,3,rep,name=third" json:"third,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *ComplexExtension) Reset() { *m = ComplexExtension{} } +func (m *ComplexExtension) String() string { return proto.CompactTextString(m) } +func (*ComplexExtension) ProtoMessage() {} +func (*ComplexExtension) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{14} } + +func (m *ComplexExtension) GetFirst() int32 { + if m != nil && m.First != nil { + return *m.First + } + return 0 +} + +func (m *ComplexExtension) GetSecond() int32 { + if m != nil && m.Second != nil { + return *m.Second + } + return 0 +} + +func (m *ComplexExtension) GetThird() []int32 { + if m != nil { + return m.Third + } + return nil +} + +type DefaultsMessage struct { + XXX_extensions map[int32]proto.Extension `json:"-"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *DefaultsMessage) Reset() { *m = DefaultsMessage{} } +func (m *DefaultsMessage) String() string { return proto.CompactTextString(m) } +func (*DefaultsMessage) ProtoMessage() {} +func (*DefaultsMessage) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{15} } + +var extRange_DefaultsMessage = []proto.ExtensionRange{ + {100, 536870911}, +} + +func (*DefaultsMessage) ExtensionRangeArray() []proto.ExtensionRange { + return extRange_DefaultsMessage +} +func (m *DefaultsMessage) ExtensionMap() map[int32]proto.Extension { + if m.XXX_extensions == nil { + m.XXX_extensions = make(map[int32]proto.Extension) + } + return m.XXX_extensions +} + +type MyMessageSet struct { + XXX_extensions map[int32]proto.Extension `json:"-"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MyMessageSet) Reset() { *m = MyMessageSet{} } +func (m *MyMessageSet) String() string { return proto.CompactTextString(m) } +func (*MyMessageSet) ProtoMessage() {} +func (*MyMessageSet) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{16} } + +func (m *MyMessageSet) Marshal() ([]byte, error) { + return proto.MarshalMessageSet(m.ExtensionMap()) +} +func (m *MyMessageSet) Unmarshal(buf []byte) error { + return proto.UnmarshalMessageSet(buf, m.ExtensionMap()) +} +func (m *MyMessageSet) MarshalJSON() ([]byte, error) { + return proto.MarshalMessageSetJSON(m.XXX_extensions) +} +func (m *MyMessageSet) UnmarshalJSON(buf []byte) error { + return proto.UnmarshalMessageSetJSON(buf, m.XXX_extensions) +} + +// ensure MyMessageSet satisfies proto.Marshaler and proto.Unmarshaler +var _ proto.Marshaler = (*MyMessageSet)(nil) +var _ proto.Unmarshaler = (*MyMessageSet)(nil) + +var extRange_MyMessageSet = []proto.ExtensionRange{ + {100, 2147483646}, +} + +func (*MyMessageSet) ExtensionRangeArray() []proto.ExtensionRange { + return extRange_MyMessageSet +} +func (m *MyMessageSet) ExtensionMap() map[int32]proto.Extension { + if m.XXX_extensions == nil { + m.XXX_extensions = make(map[int32]proto.Extension) + } + return m.XXX_extensions +} + +type Empty struct { + XXX_unrecognized []byte `json:"-"` +} + +func (m *Empty) Reset() { *m = Empty{} } +func (m *Empty) String() string { return proto.CompactTextString(m) } +func (*Empty) ProtoMessage() {} +func (*Empty) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{17} } + +type MessageList struct { + Message []*MessageList_Message `protobuf:"group,1,rep,name=Message,json=message" json:"message,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MessageList) Reset() { *m = MessageList{} } +func (m *MessageList) String() string { return proto.CompactTextString(m) } +func (*MessageList) ProtoMessage() {} +func (*MessageList) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{18} } + +func (m *MessageList) GetMessage() []*MessageList_Message { + if m != nil { + return m.Message + } + return nil +} + +type MessageList_Message struct { + Name *string `protobuf:"bytes,2,req,name=name" json:"name,omitempty"` + Count *int32 `protobuf:"varint,3,req,name=count" json:"count,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MessageList_Message) Reset() { *m = MessageList_Message{} } +func (m *MessageList_Message) String() string { return proto.CompactTextString(m) } +func (*MessageList_Message) ProtoMessage() {} +func (*MessageList_Message) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{18, 0} } + +func (m *MessageList_Message) GetName() string { + if m != nil && m.Name != nil { + return *m.Name + } + return "" +} + +func (m *MessageList_Message) GetCount() int32 { + if m != nil && m.Count != nil { + return *m.Count + } + return 0 +} + +type Strings struct { + StringField *string `protobuf:"bytes,1,opt,name=string_field,json=stringField" json:"string_field,omitempty"` + BytesField []byte `protobuf:"bytes,2,opt,name=bytes_field,json=bytesField" json:"bytes_field,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *Strings) Reset() { *m = Strings{} } +func (m *Strings) String() string { return proto.CompactTextString(m) } +func (*Strings) ProtoMessage() {} +func (*Strings) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{19} } + +func (m *Strings) GetStringField() string { + if m != nil && m.StringField != nil { + return *m.StringField + } + return "" +} + +func (m *Strings) GetBytesField() []byte { + if m != nil { + return m.BytesField + } + return nil +} + +type Defaults struct { + // Default-valued fields of all basic types. + // Same as GoTest, but copied here to make testing easier. + F_Bool *bool `protobuf:"varint,1,opt,name=F_Bool,json=fBool,def=1" json:"F_Bool,omitempty"` + F_Int32 *int32 `protobuf:"varint,2,opt,name=F_Int32,json=fInt32,def=32" json:"F_Int32,omitempty"` + F_Int64 *int64 `protobuf:"varint,3,opt,name=F_Int64,json=fInt64,def=64" json:"F_Int64,omitempty"` + F_Fixed32 *uint32 `protobuf:"fixed32,4,opt,name=F_Fixed32,json=fFixed32,def=320" json:"F_Fixed32,omitempty"` + F_Fixed64 *uint64 `protobuf:"fixed64,5,opt,name=F_Fixed64,json=fFixed64,def=640" json:"F_Fixed64,omitempty"` + F_Uint32 *uint32 `protobuf:"varint,6,opt,name=F_Uint32,json=fUint32,def=3200" json:"F_Uint32,omitempty"` + F_Uint64 *uint64 `protobuf:"varint,7,opt,name=F_Uint64,json=fUint64,def=6400" json:"F_Uint64,omitempty"` + F_Float *float32 `protobuf:"fixed32,8,opt,name=F_Float,json=fFloat,def=314159" json:"F_Float,omitempty"` + F_Double *float64 `protobuf:"fixed64,9,opt,name=F_Double,json=fDouble,def=271828" json:"F_Double,omitempty"` + F_String *string `protobuf:"bytes,10,opt,name=F_String,json=fString,def=hello, \"world!\"\n" json:"F_String,omitempty"` + F_Bytes []byte `protobuf:"bytes,11,opt,name=F_Bytes,json=fBytes,def=Bignose" json:"F_Bytes,omitempty"` + F_Sint32 *int32 `protobuf:"zigzag32,12,opt,name=F_Sint32,json=fSint32,def=-32" json:"F_Sint32,omitempty"` + F_Sint64 *int64 `protobuf:"zigzag64,13,opt,name=F_Sint64,json=fSint64,def=-64" json:"F_Sint64,omitempty"` + F_Enum *Defaults_Color `protobuf:"varint,14,opt,name=F_Enum,json=fEnum,enum=testdata.Defaults_Color,def=1" json:"F_Enum,omitempty"` + // More fields with crazy defaults. + F_Pinf *float32 `protobuf:"fixed32,15,opt,name=F_Pinf,json=fPinf,def=inf" json:"F_Pinf,omitempty"` + F_Ninf *float32 `protobuf:"fixed32,16,opt,name=F_Ninf,json=fNinf,def=-inf" json:"F_Ninf,omitempty"` + F_Nan *float32 `protobuf:"fixed32,17,opt,name=F_Nan,json=fNan,def=nan" json:"F_Nan,omitempty"` + // Sub-message. + Sub *SubDefaults `protobuf:"bytes,18,opt,name=sub" json:"sub,omitempty"` + // Redundant but explicit defaults. + StrZero *string `protobuf:"bytes,19,opt,name=str_zero,json=strZero,def=" json:"str_zero,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *Defaults) Reset() { *m = Defaults{} } +func (m *Defaults) String() string { return proto.CompactTextString(m) } +func (*Defaults) ProtoMessage() {} +func (*Defaults) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{20} } + +const Default_Defaults_F_Bool bool = true +const Default_Defaults_F_Int32 int32 = 32 +const Default_Defaults_F_Int64 int64 = 64 +const Default_Defaults_F_Fixed32 uint32 = 320 +const Default_Defaults_F_Fixed64 uint64 = 640 +const Default_Defaults_F_Uint32 uint32 = 3200 +const Default_Defaults_F_Uint64 uint64 = 6400 +const Default_Defaults_F_Float float32 = 314159 +const Default_Defaults_F_Double float64 = 271828 +const Default_Defaults_F_String string = "hello, \"world!\"\n" + +var Default_Defaults_F_Bytes []byte = []byte("Bignose") + +const Default_Defaults_F_Sint32 int32 = -32 +const Default_Defaults_F_Sint64 int64 = -64 +const Default_Defaults_F_Enum Defaults_Color = Defaults_GREEN + +var Default_Defaults_F_Pinf float32 = float32(math.Inf(1)) +var Default_Defaults_F_Ninf float32 = float32(math.Inf(-1)) +var Default_Defaults_F_Nan float32 = float32(math.NaN()) + +func (m *Defaults) GetF_Bool() bool { + if m != nil && m.F_Bool != nil { + return *m.F_Bool + } + return Default_Defaults_F_Bool +} + +func (m *Defaults) GetF_Int32() int32 { + if m != nil && m.F_Int32 != nil { + return *m.F_Int32 + } + return Default_Defaults_F_Int32 +} + +func (m *Defaults) GetF_Int64() int64 { + if m != nil && m.F_Int64 != nil { + return *m.F_Int64 + } + return Default_Defaults_F_Int64 +} + +func (m *Defaults) GetF_Fixed32() uint32 { + if m != nil && m.F_Fixed32 != nil { + return *m.F_Fixed32 + } + return Default_Defaults_F_Fixed32 +} + +func (m *Defaults) GetF_Fixed64() uint64 { + if m != nil && m.F_Fixed64 != nil { + return *m.F_Fixed64 + } + return Default_Defaults_F_Fixed64 +} + +func (m *Defaults) GetF_Uint32() uint32 { + if m != nil && m.F_Uint32 != nil { + return *m.F_Uint32 + } + return Default_Defaults_F_Uint32 +} + +func (m *Defaults) GetF_Uint64() uint64 { + if m != nil && m.F_Uint64 != nil { + return *m.F_Uint64 + } + return Default_Defaults_F_Uint64 +} + +func (m *Defaults) GetF_Float() float32 { + if m != nil && m.F_Float != nil { + return *m.F_Float + } + return Default_Defaults_F_Float +} + +func (m *Defaults) GetF_Double() float64 { + if m != nil && m.F_Double != nil { + return *m.F_Double + } + return Default_Defaults_F_Double +} + +func (m *Defaults) GetF_String() string { + if m != nil && m.F_String != nil { + return *m.F_String + } + return Default_Defaults_F_String +} + +func (m *Defaults) GetF_Bytes() []byte { + if m != nil && m.F_Bytes != nil { + return m.F_Bytes + } + return append([]byte(nil), Default_Defaults_F_Bytes...) +} + +func (m *Defaults) GetF_Sint32() int32 { + if m != nil && m.F_Sint32 != nil { + return *m.F_Sint32 + } + return Default_Defaults_F_Sint32 +} + +func (m *Defaults) GetF_Sint64() int64 { + if m != nil && m.F_Sint64 != nil { + return *m.F_Sint64 + } + return Default_Defaults_F_Sint64 +} + +func (m *Defaults) GetF_Enum() Defaults_Color { + if m != nil && m.F_Enum != nil { + return *m.F_Enum + } + return Default_Defaults_F_Enum +} + +func (m *Defaults) GetF_Pinf() float32 { + if m != nil && m.F_Pinf != nil { + return *m.F_Pinf + } + return Default_Defaults_F_Pinf +} + +func (m *Defaults) GetF_Ninf() float32 { + if m != nil && m.F_Ninf != nil { + return *m.F_Ninf + } + return Default_Defaults_F_Ninf +} + +func (m *Defaults) GetF_Nan() float32 { + if m != nil && m.F_Nan != nil { + return *m.F_Nan + } + return Default_Defaults_F_Nan +} + +func (m *Defaults) GetSub() *SubDefaults { + if m != nil { + return m.Sub + } + return nil +} + +func (m *Defaults) GetStrZero() string { + if m != nil && m.StrZero != nil { + return *m.StrZero + } + return "" +} + +type SubDefaults struct { + N *int64 `protobuf:"varint,1,opt,name=n,def=7" json:"n,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *SubDefaults) Reset() { *m = SubDefaults{} } +func (m *SubDefaults) String() string { return proto.CompactTextString(m) } +func (*SubDefaults) ProtoMessage() {} +func (*SubDefaults) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{21} } + +const Default_SubDefaults_N int64 = 7 + +func (m *SubDefaults) GetN() int64 { + if m != nil && m.N != nil { + return *m.N + } + return Default_SubDefaults_N +} + +type RepeatedEnum struct { + Color []RepeatedEnum_Color `protobuf:"varint,1,rep,name=color,enum=testdata.RepeatedEnum_Color" json:"color,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *RepeatedEnum) Reset() { *m = RepeatedEnum{} } +func (m *RepeatedEnum) String() string { return proto.CompactTextString(m) } +func (*RepeatedEnum) ProtoMessage() {} +func (*RepeatedEnum) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{22} } + +func (m *RepeatedEnum) GetColor() []RepeatedEnum_Color { + if m != nil { + return m.Color + } + return nil +} + +type MoreRepeated struct { + Bools []bool `protobuf:"varint,1,rep,name=bools" json:"bools,omitempty"` + BoolsPacked []bool `protobuf:"varint,2,rep,packed,name=bools_packed,json=boolsPacked" json:"bools_packed,omitempty"` + Ints []int32 `protobuf:"varint,3,rep,name=ints" json:"ints,omitempty"` + IntsPacked []int32 `protobuf:"varint,4,rep,packed,name=ints_packed,json=intsPacked" json:"ints_packed,omitempty"` + Int64SPacked []int64 `protobuf:"varint,7,rep,packed,name=int64s_packed,json=int64sPacked" json:"int64s_packed,omitempty"` + Strings []string `protobuf:"bytes,5,rep,name=strings" json:"strings,omitempty"` + Fixeds []uint32 `protobuf:"fixed32,6,rep,name=fixeds" json:"fixeds,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MoreRepeated) Reset() { *m = MoreRepeated{} } +func (m *MoreRepeated) String() string { return proto.CompactTextString(m) } +func (*MoreRepeated) ProtoMessage() {} +func (*MoreRepeated) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{23} } + +func (m *MoreRepeated) GetBools() []bool { + if m != nil { + return m.Bools + } + return nil +} + +func (m *MoreRepeated) GetBoolsPacked() []bool { + if m != nil { + return m.BoolsPacked + } + return nil +} + +func (m *MoreRepeated) GetInts() []int32 { + if m != nil { + return m.Ints + } + return nil +} + +func (m *MoreRepeated) GetIntsPacked() []int32 { + if m != nil { + return m.IntsPacked + } + return nil +} + +func (m *MoreRepeated) GetInt64SPacked() []int64 { + if m != nil { + return m.Int64SPacked + } + return nil +} + +func (m *MoreRepeated) GetStrings() []string { + if m != nil { + return m.Strings + } + return nil +} + +func (m *MoreRepeated) GetFixeds() []uint32 { + if m != nil { + return m.Fixeds + } + return nil +} + +type GroupOld struct { + G *GroupOld_G `protobuf:"group,101,opt,name=G,json=g" json:"g,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GroupOld) Reset() { *m = GroupOld{} } +func (m *GroupOld) String() string { return proto.CompactTextString(m) } +func (*GroupOld) ProtoMessage() {} +func (*GroupOld) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{24} } + +func (m *GroupOld) GetG() *GroupOld_G { + if m != nil { + return m.G + } + return nil +} + +type GroupOld_G struct { + X *int32 `protobuf:"varint,2,opt,name=x" json:"x,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GroupOld_G) Reset() { *m = GroupOld_G{} } +func (m *GroupOld_G) String() string { return proto.CompactTextString(m) } +func (*GroupOld_G) ProtoMessage() {} +func (*GroupOld_G) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{24, 0} } + +func (m *GroupOld_G) GetX() int32 { + if m != nil && m.X != nil { + return *m.X + } + return 0 +} + +type GroupNew struct { + G *GroupNew_G `protobuf:"group,101,opt,name=G,json=g" json:"g,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GroupNew) Reset() { *m = GroupNew{} } +func (m *GroupNew) String() string { return proto.CompactTextString(m) } +func (*GroupNew) ProtoMessage() {} +func (*GroupNew) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{25} } + +func (m *GroupNew) GetG() *GroupNew_G { + if m != nil { + return m.G + } + return nil +} + +type GroupNew_G struct { + X *int32 `protobuf:"varint,2,opt,name=x" json:"x,omitempty"` + Y *int32 `protobuf:"varint,3,opt,name=y" json:"y,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *GroupNew_G) Reset() { *m = GroupNew_G{} } +func (m *GroupNew_G) String() string { return proto.CompactTextString(m) } +func (*GroupNew_G) ProtoMessage() {} +func (*GroupNew_G) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{25, 0} } + +func (m *GroupNew_G) GetX() int32 { + if m != nil && m.X != nil { + return *m.X + } + return 0 +} + +func (m *GroupNew_G) GetY() int32 { + if m != nil && m.Y != nil { + return *m.Y + } + return 0 +} + +type FloatingPoint struct { + F *float64 `protobuf:"fixed64,1,req,name=f" json:"f,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *FloatingPoint) Reset() { *m = FloatingPoint{} } +func (m *FloatingPoint) String() string { return proto.CompactTextString(m) } +func (*FloatingPoint) ProtoMessage() {} +func (*FloatingPoint) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{26} } + +func (m *FloatingPoint) GetF() float64 { + if m != nil && m.F != nil { + return *m.F + } + return 0 +} + +type MessageWithMap struct { + NameMapping map[int32]string `protobuf:"bytes,1,rep,name=name_mapping,json=nameMapping" json:"name_mapping,omitempty" protobuf_key:"varint,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + MsgMapping map[int64]*FloatingPoint `protobuf:"bytes,2,rep,name=msg_mapping,json=msgMapping" json:"msg_mapping,omitempty" protobuf_key:"zigzag64,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ByteMapping map[bool][]byte `protobuf:"bytes,3,rep,name=byte_mapping,json=byteMapping" json:"byte_mapping,omitempty" protobuf_key:"varint,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + StrToStr map[string]string `protobuf:"bytes,4,rep,name=str_to_str,json=strToStr" json:"str_to_str,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *MessageWithMap) Reset() { *m = MessageWithMap{} } +func (m *MessageWithMap) String() string { return proto.CompactTextString(m) } +func (*MessageWithMap) ProtoMessage() {} +func (*MessageWithMap) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{27} } + +func (m *MessageWithMap) GetNameMapping() map[int32]string { + if m != nil { + return m.NameMapping + } + return nil +} + +func (m *MessageWithMap) GetMsgMapping() map[int64]*FloatingPoint { + if m != nil { + return m.MsgMapping + } + return nil +} + +func (m *MessageWithMap) GetByteMapping() map[bool][]byte { + if m != nil { + return m.ByteMapping + } + return nil +} + +func (m *MessageWithMap) GetStrToStr() map[string]string { + if m != nil { + return m.StrToStr + } + return nil +} + +type Oneof struct { + // Types that are valid to be assigned to Union: + // *Oneof_F_Bool + // *Oneof_F_Int32 + // *Oneof_F_Int64 + // *Oneof_F_Fixed32 + // *Oneof_F_Fixed64 + // *Oneof_F_Uint32 + // *Oneof_F_Uint64 + // *Oneof_F_Float + // *Oneof_F_Double + // *Oneof_F_String + // *Oneof_F_Bytes + // *Oneof_F_Sint32 + // *Oneof_F_Sint64 + // *Oneof_F_Enum + // *Oneof_F_Message + // *Oneof_FGroup + // *Oneof_F_Largest_Tag + Union isOneof_Union `protobuf_oneof:"union"` + // Types that are valid to be assigned to Tormato: + // *Oneof_Value + Tormato isOneof_Tormato `protobuf_oneof:"tormato"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *Oneof) Reset() { *m = Oneof{} } +func (m *Oneof) String() string { return proto.CompactTextString(m) } +func (*Oneof) ProtoMessage() {} +func (*Oneof) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{28} } + +type isOneof_Union interface { + isOneof_Union() +} +type isOneof_Tormato interface { + isOneof_Tormato() +} + +type Oneof_F_Bool struct { + F_Bool bool `protobuf:"varint,1,opt,name=F_Bool,json=fBool,oneof"` +} +type Oneof_F_Int32 struct { + F_Int32 int32 `protobuf:"varint,2,opt,name=F_Int32,json=fInt32,oneof"` +} +type Oneof_F_Int64 struct { + F_Int64 int64 `protobuf:"varint,3,opt,name=F_Int64,json=fInt64,oneof"` +} +type Oneof_F_Fixed32 struct { + F_Fixed32 uint32 `protobuf:"fixed32,4,opt,name=F_Fixed32,json=fFixed32,oneof"` +} +type Oneof_F_Fixed64 struct { + F_Fixed64 uint64 `protobuf:"fixed64,5,opt,name=F_Fixed64,json=fFixed64,oneof"` +} +type Oneof_F_Uint32 struct { + F_Uint32 uint32 `protobuf:"varint,6,opt,name=F_Uint32,json=fUint32,oneof"` +} +type Oneof_F_Uint64 struct { + F_Uint64 uint64 `protobuf:"varint,7,opt,name=F_Uint64,json=fUint64,oneof"` +} +type Oneof_F_Float struct { + F_Float float32 `protobuf:"fixed32,8,opt,name=F_Float,json=fFloat,oneof"` +} +type Oneof_F_Double struct { + F_Double float64 `protobuf:"fixed64,9,opt,name=F_Double,json=fDouble,oneof"` +} +type Oneof_F_String struct { + F_String string `protobuf:"bytes,10,opt,name=F_String,json=fString,oneof"` +} +type Oneof_F_Bytes struct { + F_Bytes []byte `protobuf:"bytes,11,opt,name=F_Bytes,json=fBytes,oneof"` +} +type Oneof_F_Sint32 struct { + F_Sint32 int32 `protobuf:"zigzag32,12,opt,name=F_Sint32,json=fSint32,oneof"` +} +type Oneof_F_Sint64 struct { + F_Sint64 int64 `protobuf:"zigzag64,13,opt,name=F_Sint64,json=fSint64,oneof"` +} +type Oneof_F_Enum struct { + F_Enum MyMessage_Color `protobuf:"varint,14,opt,name=F_Enum,json=fEnum,enum=testdata.MyMessage_Color,oneof"` +} +type Oneof_F_Message struct { + F_Message *GoTestField `protobuf:"bytes,15,opt,name=F_Message,json=fMessage,oneof"` +} +type Oneof_FGroup struct { + FGroup *Oneof_F_Group `protobuf:"group,16,opt,name=F_Group,json=fGroup,oneof"` +} +type Oneof_F_Largest_Tag struct { + F_Largest_Tag int32 `protobuf:"varint,536870911,opt,name=F_Largest_Tag,json=fLargestTag,oneof"` +} +type Oneof_Value struct { + Value int32 `protobuf:"varint,100,opt,name=value,oneof"` +} + +func (*Oneof_F_Bool) isOneof_Union() {} +func (*Oneof_F_Int32) isOneof_Union() {} +func (*Oneof_F_Int64) isOneof_Union() {} +func (*Oneof_F_Fixed32) isOneof_Union() {} +func (*Oneof_F_Fixed64) isOneof_Union() {} +func (*Oneof_F_Uint32) isOneof_Union() {} +func (*Oneof_F_Uint64) isOneof_Union() {} +func (*Oneof_F_Float) isOneof_Union() {} +func (*Oneof_F_Double) isOneof_Union() {} +func (*Oneof_F_String) isOneof_Union() {} +func (*Oneof_F_Bytes) isOneof_Union() {} +func (*Oneof_F_Sint32) isOneof_Union() {} +func (*Oneof_F_Sint64) isOneof_Union() {} +func (*Oneof_F_Enum) isOneof_Union() {} +func (*Oneof_F_Message) isOneof_Union() {} +func (*Oneof_FGroup) isOneof_Union() {} +func (*Oneof_F_Largest_Tag) isOneof_Union() {} +func (*Oneof_Value) isOneof_Tormato() {} + +func (m *Oneof) GetUnion() isOneof_Union { + if m != nil { + return m.Union + } + return nil +} +func (m *Oneof) GetTormato() isOneof_Tormato { + if m != nil { + return m.Tormato + } + return nil +} + +func (m *Oneof) GetF_Bool() bool { + if x, ok := m.GetUnion().(*Oneof_F_Bool); ok { + return x.F_Bool + } + return false +} + +func (m *Oneof) GetF_Int32() int32 { + if x, ok := m.GetUnion().(*Oneof_F_Int32); ok { + return x.F_Int32 + } + return 0 +} + +func (m *Oneof) GetF_Int64() int64 { + if x, ok := m.GetUnion().(*Oneof_F_Int64); ok { + return x.F_Int64 + } + return 0 +} + +func (m *Oneof) GetF_Fixed32() uint32 { + if x, ok := m.GetUnion().(*Oneof_F_Fixed32); ok { + return x.F_Fixed32 + } + return 0 +} + +func (m *Oneof) GetF_Fixed64() uint64 { + if x, ok := m.GetUnion().(*Oneof_F_Fixed64); ok { + return x.F_Fixed64 + } + return 0 +} + +func (m *Oneof) GetF_Uint32() uint32 { + if x, ok := m.GetUnion().(*Oneof_F_Uint32); ok { + return x.F_Uint32 + } + return 0 +} + +func (m *Oneof) GetF_Uint64() uint64 { + if x, ok := m.GetUnion().(*Oneof_F_Uint64); ok { + return x.F_Uint64 + } + return 0 +} + +func (m *Oneof) GetF_Float() float32 { + if x, ok := m.GetUnion().(*Oneof_F_Float); ok { + return x.F_Float + } + return 0 +} + +func (m *Oneof) GetF_Double() float64 { + if x, ok := m.GetUnion().(*Oneof_F_Double); ok { + return x.F_Double + } + return 0 +} + +func (m *Oneof) GetF_String() string { + if x, ok := m.GetUnion().(*Oneof_F_String); ok { + return x.F_String + } + return "" +} + +func (m *Oneof) GetF_Bytes() []byte { + if x, ok := m.GetUnion().(*Oneof_F_Bytes); ok { + return x.F_Bytes + } + return nil +} + +func (m *Oneof) GetF_Sint32() int32 { + if x, ok := m.GetUnion().(*Oneof_F_Sint32); ok { + return x.F_Sint32 + } + return 0 +} + +func (m *Oneof) GetF_Sint64() int64 { + if x, ok := m.GetUnion().(*Oneof_F_Sint64); ok { + return x.F_Sint64 + } + return 0 +} + +func (m *Oneof) GetF_Enum() MyMessage_Color { + if x, ok := m.GetUnion().(*Oneof_F_Enum); ok { + return x.F_Enum + } + return MyMessage_RED +} + +func (m *Oneof) GetF_Message() *GoTestField { + if x, ok := m.GetUnion().(*Oneof_F_Message); ok { + return x.F_Message + } + return nil +} + +func (m *Oneof) GetFGroup() *Oneof_F_Group { + if x, ok := m.GetUnion().(*Oneof_FGroup); ok { + return x.FGroup + } + return nil +} + +func (m *Oneof) GetF_Largest_Tag() int32 { + if x, ok := m.GetUnion().(*Oneof_F_Largest_Tag); ok { + return x.F_Largest_Tag + } + return 0 +} + +func (m *Oneof) GetValue() int32 { + if x, ok := m.GetTormato().(*Oneof_Value); ok { + return x.Value + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Oneof) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Oneof_OneofMarshaler, _Oneof_OneofUnmarshaler, _Oneof_OneofSizer, []interface{}{ + (*Oneof_F_Bool)(nil), + (*Oneof_F_Int32)(nil), + (*Oneof_F_Int64)(nil), + (*Oneof_F_Fixed32)(nil), + (*Oneof_F_Fixed64)(nil), + (*Oneof_F_Uint32)(nil), + (*Oneof_F_Uint64)(nil), + (*Oneof_F_Float)(nil), + (*Oneof_F_Double)(nil), + (*Oneof_F_String)(nil), + (*Oneof_F_Bytes)(nil), + (*Oneof_F_Sint32)(nil), + (*Oneof_F_Sint64)(nil), + (*Oneof_F_Enum)(nil), + (*Oneof_F_Message)(nil), + (*Oneof_FGroup)(nil), + (*Oneof_F_Largest_Tag)(nil), + (*Oneof_Value)(nil), + } +} + +func _Oneof_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Oneof) + // union + switch x := m.Union.(type) { + case *Oneof_F_Bool: + t := uint64(0) + if x.F_Bool { + t = 1 + } + b.EncodeVarint(1<<3 | proto.WireVarint) + b.EncodeVarint(t) + case *Oneof_F_Int32: + b.EncodeVarint(2<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.F_Int32)) + case *Oneof_F_Int64: + b.EncodeVarint(3<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.F_Int64)) + case *Oneof_F_Fixed32: + b.EncodeVarint(4<<3 | proto.WireFixed32) + b.EncodeFixed32(uint64(x.F_Fixed32)) + case *Oneof_F_Fixed64: + b.EncodeVarint(5<<3 | proto.WireFixed64) + b.EncodeFixed64(uint64(x.F_Fixed64)) + case *Oneof_F_Uint32: + b.EncodeVarint(6<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.F_Uint32)) + case *Oneof_F_Uint64: + b.EncodeVarint(7<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.F_Uint64)) + case *Oneof_F_Float: + b.EncodeVarint(8<<3 | proto.WireFixed32) + b.EncodeFixed32(uint64(math.Float32bits(x.F_Float))) + case *Oneof_F_Double: + b.EncodeVarint(9<<3 | proto.WireFixed64) + b.EncodeFixed64(math.Float64bits(x.F_Double)) + case *Oneof_F_String: + b.EncodeVarint(10<<3 | proto.WireBytes) + b.EncodeStringBytes(x.F_String) + case *Oneof_F_Bytes: + b.EncodeVarint(11<<3 | proto.WireBytes) + b.EncodeRawBytes(x.F_Bytes) + case *Oneof_F_Sint32: + b.EncodeVarint(12<<3 | proto.WireVarint) + b.EncodeZigzag32(uint64(x.F_Sint32)) + case *Oneof_F_Sint64: + b.EncodeVarint(13<<3 | proto.WireVarint) + b.EncodeZigzag64(uint64(x.F_Sint64)) + case *Oneof_F_Enum: + b.EncodeVarint(14<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.F_Enum)) + case *Oneof_F_Message: + b.EncodeVarint(15<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.F_Message); err != nil { + return err + } + case *Oneof_FGroup: + b.EncodeVarint(16<<3 | proto.WireStartGroup) + if err := b.Marshal(x.FGroup); err != nil { + return err + } + b.EncodeVarint(16<<3 | proto.WireEndGroup) + case *Oneof_F_Largest_Tag: + b.EncodeVarint(536870911<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.F_Largest_Tag)) + case nil: + default: + return fmt.Errorf("Oneof.Union has unexpected type %T", x) + } + // tormato + switch x := m.Tormato.(type) { + case *Oneof_Value: + b.EncodeVarint(100<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Value)) + case nil: + default: + return fmt.Errorf("Oneof.Tormato has unexpected type %T", x) + } + return nil +} + +func _Oneof_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Oneof) + switch tag { + case 1: // union.F_Bool + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Bool{x != 0} + return true, err + case 2: // union.F_Int32 + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Int32{int32(x)} + return true, err + case 3: // union.F_Int64 + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Int64{int64(x)} + return true, err + case 4: // union.F_Fixed32 + if wire != proto.WireFixed32 { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeFixed32() + m.Union = &Oneof_F_Fixed32{uint32(x)} + return true, err + case 5: // union.F_Fixed64 + if wire != proto.WireFixed64 { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeFixed64() + m.Union = &Oneof_F_Fixed64{x} + return true, err + case 6: // union.F_Uint32 + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Uint32{uint32(x)} + return true, err + case 7: // union.F_Uint64 + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Uint64{x} + return true, err + case 8: // union.F_Float + if wire != proto.WireFixed32 { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeFixed32() + m.Union = &Oneof_F_Float{math.Float32frombits(uint32(x))} + return true, err + case 9: // union.F_Double + if wire != proto.WireFixed64 { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeFixed64() + m.Union = &Oneof_F_Double{math.Float64frombits(x)} + return true, err + case 10: // union.F_String + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.Union = &Oneof_F_String{x} + return true, err + case 11: // union.F_Bytes + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeRawBytes(true) + m.Union = &Oneof_F_Bytes{x} + return true, err + case 12: // union.F_Sint32 + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeZigzag32() + m.Union = &Oneof_F_Sint32{int32(x)} + return true, err + case 13: // union.F_Sint64 + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeZigzag64() + m.Union = &Oneof_F_Sint64{int64(x)} + return true, err + case 14: // union.F_Enum + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Enum{MyMessage_Color(x)} + return true, err + case 15: // union.F_Message + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(GoTestField) + err := b.DecodeMessage(msg) + m.Union = &Oneof_F_Message{msg} + return true, err + case 16: // union.f_group + if wire != proto.WireStartGroup { + return true, proto.ErrInternalBadWireType + } + msg := new(Oneof_F_Group) + err := b.DecodeGroup(msg) + m.Union = &Oneof_FGroup{msg} + return true, err + case 536870911: // union.F_Largest_Tag + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Oneof_F_Largest_Tag{int32(x)} + return true, err + case 100: // tormato.value + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Tormato = &Oneof_Value{int32(x)} + return true, err + default: + return false, nil + } +} + +func _Oneof_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Oneof) + // union + switch x := m.Union.(type) { + case *Oneof_F_Bool: + n += proto.SizeVarint(1<<3 | proto.WireVarint) + n += 1 + case *Oneof_F_Int32: + n += proto.SizeVarint(2<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.F_Int32)) + case *Oneof_F_Int64: + n += proto.SizeVarint(3<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.F_Int64)) + case *Oneof_F_Fixed32: + n += proto.SizeVarint(4<<3 | proto.WireFixed32) + n += 4 + case *Oneof_F_Fixed64: + n += proto.SizeVarint(5<<3 | proto.WireFixed64) + n += 8 + case *Oneof_F_Uint32: + n += proto.SizeVarint(6<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.F_Uint32)) + case *Oneof_F_Uint64: + n += proto.SizeVarint(7<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.F_Uint64)) + case *Oneof_F_Float: + n += proto.SizeVarint(8<<3 | proto.WireFixed32) + n += 4 + case *Oneof_F_Double: + n += proto.SizeVarint(9<<3 | proto.WireFixed64) + n += 8 + case *Oneof_F_String: + n += proto.SizeVarint(10<<3 | proto.WireBytes) + n += proto.SizeVarint(uint64(len(x.F_String))) + n += len(x.F_String) + case *Oneof_F_Bytes: + n += proto.SizeVarint(11<<3 | proto.WireBytes) + n += proto.SizeVarint(uint64(len(x.F_Bytes))) + n += len(x.F_Bytes) + case *Oneof_F_Sint32: + n += proto.SizeVarint(12<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64((uint32(x.F_Sint32) << 1) ^ uint32((int32(x.F_Sint32) >> 31)))) + case *Oneof_F_Sint64: + n += proto.SizeVarint(13<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(uint64(x.F_Sint64<<1) ^ uint64((int64(x.F_Sint64) >> 63)))) + case *Oneof_F_Enum: + n += proto.SizeVarint(14<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.F_Enum)) + case *Oneof_F_Message: + s := proto.Size(x.F_Message) + n += proto.SizeVarint(15<<3 | proto.WireBytes) + n += proto.SizeVarint(uint64(s)) + n += s + case *Oneof_FGroup: + n += proto.SizeVarint(16<<3 | proto.WireStartGroup) + n += proto.Size(x.FGroup) + n += proto.SizeVarint(16<<3 | proto.WireEndGroup) + case *Oneof_F_Largest_Tag: + n += proto.SizeVarint(536870911<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.F_Largest_Tag)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + // tormato + switch x := m.Tormato.(type) { + case *Oneof_Value: + n += proto.SizeVarint(100<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.Value)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +type Oneof_F_Group struct { + X *int32 `protobuf:"varint,17,opt,name=x" json:"x,omitempty"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *Oneof_F_Group) Reset() { *m = Oneof_F_Group{} } +func (m *Oneof_F_Group) String() string { return proto.CompactTextString(m) } +func (*Oneof_F_Group) ProtoMessage() {} +func (*Oneof_F_Group) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{28, 0} } + +func (m *Oneof_F_Group) GetX() int32 { + if m != nil && m.X != nil { + return *m.X + } + return 0 +} + +type Communique struct { + MakeMeCry *bool `protobuf:"varint,1,opt,name=make_me_cry,json=makeMeCry" json:"make_me_cry,omitempty"` + // This is a oneof, called "union". + // + // Types that are valid to be assigned to Union: + // *Communique_Number + // *Communique_Name + // *Communique_Data + // *Communique_TempC + // *Communique_Col + // *Communique_Msg + Union isCommunique_Union `protobuf_oneof:"union"` + XXX_unrecognized []byte `json:"-"` +} + +func (m *Communique) Reset() { *m = Communique{} } +func (m *Communique) String() string { return proto.CompactTextString(m) } +func (*Communique) ProtoMessage() {} +func (*Communique) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{29} } + +type isCommunique_Union interface { + isCommunique_Union() +} + +type Communique_Number struct { + Number int32 `protobuf:"varint,5,opt,name=number,oneof"` +} +type Communique_Name struct { + Name string `protobuf:"bytes,6,opt,name=name,oneof"` +} +type Communique_Data struct { + Data []byte `protobuf:"bytes,7,opt,name=data,oneof"` +} +type Communique_TempC struct { + TempC float64 `protobuf:"fixed64,8,opt,name=temp_c,json=tempC,oneof"` +} +type Communique_Col struct { + Col MyMessage_Color `protobuf:"varint,9,opt,name=col,enum=testdata.MyMessage_Color,oneof"` +} +type Communique_Msg struct { + Msg *Strings `protobuf:"bytes,10,opt,name=msg,oneof"` +} + +func (*Communique_Number) isCommunique_Union() {} +func (*Communique_Name) isCommunique_Union() {} +func (*Communique_Data) isCommunique_Union() {} +func (*Communique_TempC) isCommunique_Union() {} +func (*Communique_Col) isCommunique_Union() {} +func (*Communique_Msg) isCommunique_Union() {} + +func (m *Communique) GetUnion() isCommunique_Union { + if m != nil { + return m.Union + } + return nil +} + +func (m *Communique) GetMakeMeCry() bool { + if m != nil && m.MakeMeCry != nil { + return *m.MakeMeCry + } + return false +} + +func (m *Communique) GetNumber() int32 { + if x, ok := m.GetUnion().(*Communique_Number); ok { + return x.Number + } + return 0 +} + +func (m *Communique) GetName() string { + if x, ok := m.GetUnion().(*Communique_Name); ok { + return x.Name + } + return "" +} + +func (m *Communique) GetData() []byte { + if x, ok := m.GetUnion().(*Communique_Data); ok { + return x.Data + } + return nil +} + +func (m *Communique) GetTempC() float64 { + if x, ok := m.GetUnion().(*Communique_TempC); ok { + return x.TempC + } + return 0 +} + +func (m *Communique) GetCol() MyMessage_Color { + if x, ok := m.GetUnion().(*Communique_Col); ok { + return x.Col + } + return MyMessage_RED +} + +func (m *Communique) GetMsg() *Strings { + if x, ok := m.GetUnion().(*Communique_Msg); ok { + return x.Msg + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Communique) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Communique_OneofMarshaler, _Communique_OneofUnmarshaler, _Communique_OneofSizer, []interface{}{ + (*Communique_Number)(nil), + (*Communique_Name)(nil), + (*Communique_Data)(nil), + (*Communique_TempC)(nil), + (*Communique_Col)(nil), + (*Communique_Msg)(nil), + } +} + +func _Communique_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Communique) + // union + switch x := m.Union.(type) { + case *Communique_Number: + b.EncodeVarint(5<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Number)) + case *Communique_Name: + b.EncodeVarint(6<<3 | proto.WireBytes) + b.EncodeStringBytes(x.Name) + case *Communique_Data: + b.EncodeVarint(7<<3 | proto.WireBytes) + b.EncodeRawBytes(x.Data) + case *Communique_TempC: + b.EncodeVarint(8<<3 | proto.WireFixed64) + b.EncodeFixed64(math.Float64bits(x.TempC)) + case *Communique_Col: + b.EncodeVarint(9<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Col)) + case *Communique_Msg: + b.EncodeVarint(10<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.Msg); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("Communique.Union has unexpected type %T", x) + } + return nil +} + +func _Communique_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Communique) + switch tag { + case 5: // union.number + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Communique_Number{int32(x)} + return true, err + case 6: // union.name + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.Union = &Communique_Name{x} + return true, err + case 7: // union.data + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeRawBytes(true) + m.Union = &Communique_Data{x} + return true, err + case 8: // union.temp_c + if wire != proto.WireFixed64 { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeFixed64() + m.Union = &Communique_TempC{math.Float64frombits(x)} + return true, err + case 9: // union.col + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.Union = &Communique_Col{MyMessage_Color(x)} + return true, err + case 10: // union.msg + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(Strings) + err := b.DecodeMessage(msg) + m.Union = &Communique_Msg{msg} + return true, err + default: + return false, nil + } +} + +func _Communique_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Communique) + // union + switch x := m.Union.(type) { + case *Communique_Number: + n += proto.SizeVarint(5<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.Number)) + case *Communique_Name: + n += proto.SizeVarint(6<<3 | proto.WireBytes) + n += proto.SizeVarint(uint64(len(x.Name))) + n += len(x.Name) + case *Communique_Data: + n += proto.SizeVarint(7<<3 | proto.WireBytes) + n += proto.SizeVarint(uint64(len(x.Data))) + n += len(x.Data) + case *Communique_TempC: + n += proto.SizeVarint(8<<3 | proto.WireFixed64) + n += 8 + case *Communique_Col: + n += proto.SizeVarint(9<<3 | proto.WireVarint) + n += proto.SizeVarint(uint64(x.Col)) + case *Communique_Msg: + s := proto.Size(x.Msg) + n += proto.SizeVarint(10<<3 | proto.WireBytes) + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +var E_Greeting = &proto.ExtensionDesc{ + ExtendedType: (*MyMessage)(nil), + ExtensionType: ([]string)(nil), + Field: 106, + Name: "testdata.greeting", + Tag: "bytes,106,rep,name=greeting", +} + +var E_Complex = &proto.ExtensionDesc{ + ExtendedType: (*OtherMessage)(nil), + ExtensionType: (*ComplexExtension)(nil), + Field: 200, + Name: "testdata.complex", + Tag: "bytes,200,opt,name=complex", +} + +var E_RComplex = &proto.ExtensionDesc{ + ExtendedType: (*OtherMessage)(nil), + ExtensionType: ([]*ComplexExtension)(nil), + Field: 201, + Name: "testdata.r_complex", + Tag: "bytes,201,rep,name=r_complex,json=rComplex", +} + +var E_NoDefaultDouble = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*float64)(nil), + Field: 101, + Name: "testdata.no_default_double", + Tag: "fixed64,101,opt,name=no_default_double,json=noDefaultDouble", +} + +var E_NoDefaultFloat = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*float32)(nil), + Field: 102, + Name: "testdata.no_default_float", + Tag: "fixed32,102,opt,name=no_default_float,json=noDefaultFloat", +} + +var E_NoDefaultInt32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 103, + Name: "testdata.no_default_int32", + Tag: "varint,103,opt,name=no_default_int32,json=noDefaultInt32", +} + +var E_NoDefaultInt64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int64)(nil), + Field: 104, + Name: "testdata.no_default_int64", + Tag: "varint,104,opt,name=no_default_int64,json=noDefaultInt64", +} + +var E_NoDefaultUint32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint32)(nil), + Field: 105, + Name: "testdata.no_default_uint32", + Tag: "varint,105,opt,name=no_default_uint32,json=noDefaultUint32", +} + +var E_NoDefaultUint64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint64)(nil), + Field: 106, + Name: "testdata.no_default_uint64", + Tag: "varint,106,opt,name=no_default_uint64,json=noDefaultUint64", +} + +var E_NoDefaultSint32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 107, + Name: "testdata.no_default_sint32", + Tag: "zigzag32,107,opt,name=no_default_sint32,json=noDefaultSint32", +} + +var E_NoDefaultSint64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int64)(nil), + Field: 108, + Name: "testdata.no_default_sint64", + Tag: "zigzag64,108,opt,name=no_default_sint64,json=noDefaultSint64", +} + +var E_NoDefaultFixed32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint32)(nil), + Field: 109, + Name: "testdata.no_default_fixed32", + Tag: "fixed32,109,opt,name=no_default_fixed32,json=noDefaultFixed32", +} + +var E_NoDefaultFixed64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint64)(nil), + Field: 110, + Name: "testdata.no_default_fixed64", + Tag: "fixed64,110,opt,name=no_default_fixed64,json=noDefaultFixed64", +} + +var E_NoDefaultSfixed32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 111, + Name: "testdata.no_default_sfixed32", + Tag: "fixed32,111,opt,name=no_default_sfixed32,json=noDefaultSfixed32", +} + +var E_NoDefaultSfixed64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int64)(nil), + Field: 112, + Name: "testdata.no_default_sfixed64", + Tag: "fixed64,112,opt,name=no_default_sfixed64,json=noDefaultSfixed64", +} + +var E_NoDefaultBool = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*bool)(nil), + Field: 113, + Name: "testdata.no_default_bool", + Tag: "varint,113,opt,name=no_default_bool,json=noDefaultBool", +} + +var E_NoDefaultString = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*string)(nil), + Field: 114, + Name: "testdata.no_default_string", + Tag: "bytes,114,opt,name=no_default_string,json=noDefaultString", +} + +var E_NoDefaultBytes = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: ([]byte)(nil), + Field: 115, + Name: "testdata.no_default_bytes", + Tag: "bytes,115,opt,name=no_default_bytes,json=noDefaultBytes", +} + +var E_NoDefaultEnum = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*DefaultsMessage_DefaultsEnum)(nil), + Field: 116, + Name: "testdata.no_default_enum", + Tag: "varint,116,opt,name=no_default_enum,json=noDefaultEnum,enum=testdata.DefaultsMessage_DefaultsEnum", +} + +var E_DefaultDouble = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*float64)(nil), + Field: 201, + Name: "testdata.default_double", + Tag: "fixed64,201,opt,name=default_double,json=defaultDouble,def=3.1415", +} + +var E_DefaultFloat = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*float32)(nil), + Field: 202, + Name: "testdata.default_float", + Tag: "fixed32,202,opt,name=default_float,json=defaultFloat,def=3.14", +} + +var E_DefaultInt32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 203, + Name: "testdata.default_int32", + Tag: "varint,203,opt,name=default_int32,json=defaultInt32,def=42", +} + +var E_DefaultInt64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int64)(nil), + Field: 204, + Name: "testdata.default_int64", + Tag: "varint,204,opt,name=default_int64,json=defaultInt64,def=43", +} + +var E_DefaultUint32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint32)(nil), + Field: 205, + Name: "testdata.default_uint32", + Tag: "varint,205,opt,name=default_uint32,json=defaultUint32,def=44", +} + +var E_DefaultUint64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint64)(nil), + Field: 206, + Name: "testdata.default_uint64", + Tag: "varint,206,opt,name=default_uint64,json=defaultUint64,def=45", +} + +var E_DefaultSint32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 207, + Name: "testdata.default_sint32", + Tag: "zigzag32,207,opt,name=default_sint32,json=defaultSint32,def=46", +} + +var E_DefaultSint64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int64)(nil), + Field: 208, + Name: "testdata.default_sint64", + Tag: "zigzag64,208,opt,name=default_sint64,json=defaultSint64,def=47", +} + +var E_DefaultFixed32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint32)(nil), + Field: 209, + Name: "testdata.default_fixed32", + Tag: "fixed32,209,opt,name=default_fixed32,json=defaultFixed32,def=48", +} + +var E_DefaultFixed64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*uint64)(nil), + Field: 210, + Name: "testdata.default_fixed64", + Tag: "fixed64,210,opt,name=default_fixed64,json=defaultFixed64,def=49", +} + +var E_DefaultSfixed32 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int32)(nil), + Field: 211, + Name: "testdata.default_sfixed32", + Tag: "fixed32,211,opt,name=default_sfixed32,json=defaultSfixed32,def=50", +} + +var E_DefaultSfixed64 = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*int64)(nil), + Field: 212, + Name: "testdata.default_sfixed64", + Tag: "fixed64,212,opt,name=default_sfixed64,json=defaultSfixed64,def=51", +} + +var E_DefaultBool = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*bool)(nil), + Field: 213, + Name: "testdata.default_bool", + Tag: "varint,213,opt,name=default_bool,json=defaultBool,def=1", +} + +var E_DefaultString = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*string)(nil), + Field: 214, + Name: "testdata.default_string", + Tag: "bytes,214,opt,name=default_string,json=defaultString,def=Hello, string", +} + +var E_DefaultBytes = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: ([]byte)(nil), + Field: 215, + Name: "testdata.default_bytes", + Tag: "bytes,215,opt,name=default_bytes,json=defaultBytes,def=Hello, bytes", +} + +var E_DefaultEnum = &proto.ExtensionDesc{ + ExtendedType: (*DefaultsMessage)(nil), + ExtensionType: (*DefaultsMessage_DefaultsEnum)(nil), + Field: 216, + Name: "testdata.default_enum", + Tag: "varint,216,opt,name=default_enum,json=defaultEnum,enum=testdata.DefaultsMessage_DefaultsEnum,def=1", +} + +var E_X201 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 201, + Name: "testdata.x201", + Tag: "bytes,201,opt,name=x201", +} + +var E_X202 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 202, + Name: "testdata.x202", + Tag: "bytes,202,opt,name=x202", +} + +var E_X203 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 203, + Name: "testdata.x203", + Tag: "bytes,203,opt,name=x203", +} + +var E_X204 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 204, + Name: "testdata.x204", + Tag: "bytes,204,opt,name=x204", +} + +var E_X205 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 205, + Name: "testdata.x205", + Tag: "bytes,205,opt,name=x205", +} + +var E_X206 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 206, + Name: "testdata.x206", + Tag: "bytes,206,opt,name=x206", +} + +var E_X207 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 207, + Name: "testdata.x207", + Tag: "bytes,207,opt,name=x207", +} + +var E_X208 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 208, + Name: "testdata.x208", + Tag: "bytes,208,opt,name=x208", +} + +var E_X209 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 209, + Name: "testdata.x209", + Tag: "bytes,209,opt,name=x209", +} + +var E_X210 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 210, + Name: "testdata.x210", + Tag: "bytes,210,opt,name=x210", +} + +var E_X211 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 211, + Name: "testdata.x211", + Tag: "bytes,211,opt,name=x211", +} + +var E_X212 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 212, + Name: "testdata.x212", + Tag: "bytes,212,opt,name=x212", +} + +var E_X213 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 213, + Name: "testdata.x213", + Tag: "bytes,213,opt,name=x213", +} + +var E_X214 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 214, + Name: "testdata.x214", + Tag: "bytes,214,opt,name=x214", +} + +var E_X215 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 215, + Name: "testdata.x215", + Tag: "bytes,215,opt,name=x215", +} + +var E_X216 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 216, + Name: "testdata.x216", + Tag: "bytes,216,opt,name=x216", +} + +var E_X217 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 217, + Name: "testdata.x217", + Tag: "bytes,217,opt,name=x217", +} + +var E_X218 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 218, + Name: "testdata.x218", + Tag: "bytes,218,opt,name=x218", +} + +var E_X219 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 219, + Name: "testdata.x219", + Tag: "bytes,219,opt,name=x219", +} + +var E_X220 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 220, + Name: "testdata.x220", + Tag: "bytes,220,opt,name=x220", +} + +var E_X221 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 221, + Name: "testdata.x221", + Tag: "bytes,221,opt,name=x221", +} + +var E_X222 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 222, + Name: "testdata.x222", + Tag: "bytes,222,opt,name=x222", +} + +var E_X223 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 223, + Name: "testdata.x223", + Tag: "bytes,223,opt,name=x223", +} + +var E_X224 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 224, + Name: "testdata.x224", + Tag: "bytes,224,opt,name=x224", +} + +var E_X225 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 225, + Name: "testdata.x225", + Tag: "bytes,225,opt,name=x225", +} + +var E_X226 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 226, + Name: "testdata.x226", + Tag: "bytes,226,opt,name=x226", +} + +var E_X227 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 227, + Name: "testdata.x227", + Tag: "bytes,227,opt,name=x227", +} + +var E_X228 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 228, + Name: "testdata.x228", + Tag: "bytes,228,opt,name=x228", +} + +var E_X229 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 229, + Name: "testdata.x229", + Tag: "bytes,229,opt,name=x229", +} + +var E_X230 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 230, + Name: "testdata.x230", + Tag: "bytes,230,opt,name=x230", +} + +var E_X231 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 231, + Name: "testdata.x231", + Tag: "bytes,231,opt,name=x231", +} + +var E_X232 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 232, + Name: "testdata.x232", + Tag: "bytes,232,opt,name=x232", +} + +var E_X233 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 233, + Name: "testdata.x233", + Tag: "bytes,233,opt,name=x233", +} + +var E_X234 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 234, + Name: "testdata.x234", + Tag: "bytes,234,opt,name=x234", +} + +var E_X235 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 235, + Name: "testdata.x235", + Tag: "bytes,235,opt,name=x235", +} + +var E_X236 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 236, + Name: "testdata.x236", + Tag: "bytes,236,opt,name=x236", +} + +var E_X237 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 237, + Name: "testdata.x237", + Tag: "bytes,237,opt,name=x237", +} + +var E_X238 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 238, + Name: "testdata.x238", + Tag: "bytes,238,opt,name=x238", +} + +var E_X239 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 239, + Name: "testdata.x239", + Tag: "bytes,239,opt,name=x239", +} + +var E_X240 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 240, + Name: "testdata.x240", + Tag: "bytes,240,opt,name=x240", +} + +var E_X241 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 241, + Name: "testdata.x241", + Tag: "bytes,241,opt,name=x241", +} + +var E_X242 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 242, + Name: "testdata.x242", + Tag: "bytes,242,opt,name=x242", +} + +var E_X243 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 243, + Name: "testdata.x243", + Tag: "bytes,243,opt,name=x243", +} + +var E_X244 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 244, + Name: "testdata.x244", + Tag: "bytes,244,opt,name=x244", +} + +var E_X245 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 245, + Name: "testdata.x245", + Tag: "bytes,245,opt,name=x245", +} + +var E_X246 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 246, + Name: "testdata.x246", + Tag: "bytes,246,opt,name=x246", +} + +var E_X247 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 247, + Name: "testdata.x247", + Tag: "bytes,247,opt,name=x247", +} + +var E_X248 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 248, + Name: "testdata.x248", + Tag: "bytes,248,opt,name=x248", +} + +var E_X249 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 249, + Name: "testdata.x249", + Tag: "bytes,249,opt,name=x249", +} + +var E_X250 = &proto.ExtensionDesc{ + ExtendedType: (*MyMessageSet)(nil), + ExtensionType: (*Empty)(nil), + Field: 250, + Name: "testdata.x250", + Tag: "bytes,250,opt,name=x250", +} + +func init() { + proto.RegisterType((*GoEnum)(nil), "testdata.GoEnum") + proto.RegisterType((*GoTestField)(nil), "testdata.GoTestField") + proto.RegisterType((*GoTest)(nil), "testdata.GoTest") + proto.RegisterType((*GoTest_RequiredGroup)(nil), "testdata.GoTest.RequiredGroup") + proto.RegisterType((*GoTest_RepeatedGroup)(nil), "testdata.GoTest.RepeatedGroup") + proto.RegisterType((*GoTest_OptionalGroup)(nil), "testdata.GoTest.OptionalGroup") + proto.RegisterType((*GoSkipTest)(nil), "testdata.GoSkipTest") + proto.RegisterType((*GoSkipTest_SkipGroup)(nil), "testdata.GoSkipTest.SkipGroup") + proto.RegisterType((*NonPackedTest)(nil), "testdata.NonPackedTest") + proto.RegisterType((*PackedTest)(nil), "testdata.PackedTest") + proto.RegisterType((*MaxTag)(nil), "testdata.MaxTag") + proto.RegisterType((*OldMessage)(nil), "testdata.OldMessage") + proto.RegisterType((*OldMessage_Nested)(nil), "testdata.OldMessage.Nested") + proto.RegisterType((*NewMessage)(nil), "testdata.NewMessage") + proto.RegisterType((*NewMessage_Nested)(nil), "testdata.NewMessage.Nested") + proto.RegisterType((*InnerMessage)(nil), "testdata.InnerMessage") + proto.RegisterType((*OtherMessage)(nil), "testdata.OtherMessage") + proto.RegisterType((*RequiredInnerMessage)(nil), "testdata.RequiredInnerMessage") + proto.RegisterType((*MyMessage)(nil), "testdata.MyMessage") + proto.RegisterType((*MyMessage_SomeGroup)(nil), "testdata.MyMessage.SomeGroup") + proto.RegisterType((*Ext)(nil), "testdata.Ext") + proto.RegisterType((*ComplexExtension)(nil), "testdata.ComplexExtension") + proto.RegisterType((*DefaultsMessage)(nil), "testdata.DefaultsMessage") + proto.RegisterType((*MyMessageSet)(nil), "testdata.MyMessageSet") + proto.RegisterType((*Empty)(nil), "testdata.Empty") + proto.RegisterType((*MessageList)(nil), "testdata.MessageList") + proto.RegisterType((*MessageList_Message)(nil), "testdata.MessageList.Message") + proto.RegisterType((*Strings)(nil), "testdata.Strings") + proto.RegisterType((*Defaults)(nil), "testdata.Defaults") + proto.RegisterType((*SubDefaults)(nil), "testdata.SubDefaults") + proto.RegisterType((*RepeatedEnum)(nil), "testdata.RepeatedEnum") + proto.RegisterType((*MoreRepeated)(nil), "testdata.MoreRepeated") + proto.RegisterType((*GroupOld)(nil), "testdata.GroupOld") + proto.RegisterType((*GroupOld_G)(nil), "testdata.GroupOld.G") + proto.RegisterType((*GroupNew)(nil), "testdata.GroupNew") + proto.RegisterType((*GroupNew_G)(nil), "testdata.GroupNew.G") + proto.RegisterType((*FloatingPoint)(nil), "testdata.FloatingPoint") + proto.RegisterType((*MessageWithMap)(nil), "testdata.MessageWithMap") + proto.RegisterType((*Oneof)(nil), "testdata.Oneof") + proto.RegisterType((*Oneof_F_Group)(nil), "testdata.Oneof.F_Group") + proto.RegisterType((*Communique)(nil), "testdata.Communique") + proto.RegisterEnum("testdata.FOO", FOO_name, FOO_value) + proto.RegisterEnum("testdata.GoTest_KIND", GoTest_KIND_name, GoTest_KIND_value) + proto.RegisterEnum("testdata.MyMessage_Color", MyMessage_Color_name, MyMessage_Color_value) + proto.RegisterEnum("testdata.DefaultsMessage_DefaultsEnum", DefaultsMessage_DefaultsEnum_name, DefaultsMessage_DefaultsEnum_value) + proto.RegisterEnum("testdata.Defaults_Color", Defaults_Color_name, Defaults_Color_value) + proto.RegisterEnum("testdata.RepeatedEnum_Color", RepeatedEnum_Color_name, RepeatedEnum_Color_value) + proto.RegisterExtension(E_Ext_More) + proto.RegisterExtension(E_Ext_Text) + proto.RegisterExtension(E_Ext_Number) + proto.RegisterExtension(E_Greeting) + proto.RegisterExtension(E_Complex) + proto.RegisterExtension(E_RComplex) + proto.RegisterExtension(E_NoDefaultDouble) + proto.RegisterExtension(E_NoDefaultFloat) + proto.RegisterExtension(E_NoDefaultInt32) + proto.RegisterExtension(E_NoDefaultInt64) + proto.RegisterExtension(E_NoDefaultUint32) + proto.RegisterExtension(E_NoDefaultUint64) + proto.RegisterExtension(E_NoDefaultSint32) + proto.RegisterExtension(E_NoDefaultSint64) + proto.RegisterExtension(E_NoDefaultFixed32) + proto.RegisterExtension(E_NoDefaultFixed64) + proto.RegisterExtension(E_NoDefaultSfixed32) + proto.RegisterExtension(E_NoDefaultSfixed64) + proto.RegisterExtension(E_NoDefaultBool) + proto.RegisterExtension(E_NoDefaultString) + proto.RegisterExtension(E_NoDefaultBytes) + proto.RegisterExtension(E_NoDefaultEnum) + proto.RegisterExtension(E_DefaultDouble) + proto.RegisterExtension(E_DefaultFloat) + proto.RegisterExtension(E_DefaultInt32) + proto.RegisterExtension(E_DefaultInt64) + proto.RegisterExtension(E_DefaultUint32) + proto.RegisterExtension(E_DefaultUint64) + proto.RegisterExtension(E_DefaultSint32) + proto.RegisterExtension(E_DefaultSint64) + proto.RegisterExtension(E_DefaultFixed32) + proto.RegisterExtension(E_DefaultFixed64) + proto.RegisterExtension(E_DefaultSfixed32) + proto.RegisterExtension(E_DefaultSfixed64) + proto.RegisterExtension(E_DefaultBool) + proto.RegisterExtension(E_DefaultString) + proto.RegisterExtension(E_DefaultBytes) + proto.RegisterExtension(E_DefaultEnum) + proto.RegisterExtension(E_X201) + proto.RegisterExtension(E_X202) + proto.RegisterExtension(E_X203) + proto.RegisterExtension(E_X204) + proto.RegisterExtension(E_X205) + proto.RegisterExtension(E_X206) + proto.RegisterExtension(E_X207) + proto.RegisterExtension(E_X208) + proto.RegisterExtension(E_X209) + proto.RegisterExtension(E_X210) + proto.RegisterExtension(E_X211) + proto.RegisterExtension(E_X212) + proto.RegisterExtension(E_X213) + proto.RegisterExtension(E_X214) + proto.RegisterExtension(E_X215) + proto.RegisterExtension(E_X216) + proto.RegisterExtension(E_X217) + proto.RegisterExtension(E_X218) + proto.RegisterExtension(E_X219) + proto.RegisterExtension(E_X220) + proto.RegisterExtension(E_X221) + proto.RegisterExtension(E_X222) + proto.RegisterExtension(E_X223) + proto.RegisterExtension(E_X224) + proto.RegisterExtension(E_X225) + proto.RegisterExtension(E_X226) + proto.RegisterExtension(E_X227) + proto.RegisterExtension(E_X228) + proto.RegisterExtension(E_X229) + proto.RegisterExtension(E_X230) + proto.RegisterExtension(E_X231) + proto.RegisterExtension(E_X232) + proto.RegisterExtension(E_X233) + proto.RegisterExtension(E_X234) + proto.RegisterExtension(E_X235) + proto.RegisterExtension(E_X236) + proto.RegisterExtension(E_X237) + proto.RegisterExtension(E_X238) + proto.RegisterExtension(E_X239) + proto.RegisterExtension(E_X240) + proto.RegisterExtension(E_X241) + proto.RegisterExtension(E_X242) + proto.RegisterExtension(E_X243) + proto.RegisterExtension(E_X244) + proto.RegisterExtension(E_X245) + proto.RegisterExtension(E_X246) + proto.RegisterExtension(E_X247) + proto.RegisterExtension(E_X248) + proto.RegisterExtension(E_X249) + proto.RegisterExtension(E_X250) +} + +var fileDescriptor0 = []byte{ + // 4407 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x09, 0x6e, 0x88, 0x02, 0xff, 0x94, 0x5a, 0x59, 0x77, 0xdb, 0x48, + 0x76, 0x36, 0xc0, 0xfd, 0x92, 0x12, 0xa1, 0xb2, 0xda, 0x4d, 0x4b, 0x5e, 0x60, 0xce, 0x74, 0x37, + 0xbd, 0x69, 0x24, 0x10, 0xa2, 0x6d, 0xba, 0xd3, 0xe7, 0x78, 0xa1, 0x64, 0x9d, 0xb1, 0x44, 0x05, + 0x52, 0x77, 0x9f, 0xe9, 0x3c, 0xf0, 0x50, 0x22, 0x48, 0xb3, 0x4d, 0x02, 0x34, 0x09, 0xc5, 0x52, + 0xf2, 0xd2, 0x2f, 0xc9, 0x6b, 0xb6, 0x97, 0xbc, 0xe6, 0x29, 0x4f, 0x49, 0xce, 0xc9, 0x9f, 0x48, + 0xba, 0x7b, 0xd6, 0x9e, 0x35, 0xeb, 0x64, 0x5f, 0x26, 0xfb, 0x36, 0x93, 0xe4, 0xa5, 0xe7, 0xd4, + 0xad, 0x02, 0x50, 0x00, 0x09, 0x48, 0x7e, 0x12, 0x51, 0xf5, 0x7d, 0xb7, 0x6e, 0x15, 0xbe, 0xba, + 0xb7, 0x6e, 0x41, 0x00, 0x8e, 0x39, 0x71, 0x56, 0x46, 0x63, 0xdb, 0xb1, 0x49, 0x96, 0xfe, 0xee, + 0xb4, 0x9d, 0x76, 0xf9, 0x3a, 0xa4, 0x37, 0xed, 0x86, 0x75, 0x34, 0x24, 0x57, 0x21, 0xd1, 0xb5, + 0xed, 0x92, 0xa4, 0xca, 0x95, 0x79, 0x6d, 0x6e, 0xc5, 0x45, 0xac, 0x6c, 0x34, 0x9b, 0x06, 0xed, + 0x29, 0xdf, 0x81, 0xfc, 0xa6, 0xbd, 0x6f, 0x4e, 0x9c, 0x8d, 0xbe, 0x39, 0xe8, 0x90, 0x45, 0x48, + 0x3d, 0x6d, 0x1f, 0x98, 0x03, 0x64, 0xe4, 0x8c, 0xd4, 0x80, 0x3e, 0x10, 0x02, 0xc9, 0xfd, 0x93, + 0x91, 0x59, 0x92, 0xb1, 0x31, 0xe9, 0x9c, 0x8c, 0xcc, 0xf2, 0xaf, 0x5c, 0xa1, 0x83, 0x50, 0x26, + 0xb9, 0x0e, 0xc9, 0x2f, 0xf7, 0xad, 0x0e, 0x1f, 0xe5, 0x35, 0x7f, 0x14, 0xd6, 0xbf, 0xf2, 0xe5, + 0xad, 0x9d, 0xc7, 0x46, 0xf2, 0x79, 0xdf, 0x42, 0xfb, 0xfb, 0xed, 0x83, 0x01, 0x35, 0x25, 0x51, + 0xfb, 0x0e, 0x7d, 0xa0, 0xad, 0xbb, 0xed, 0x71, 0x7b, 0x58, 0x4a, 0xa8, 0x52, 0x25, 0x65, 0xa4, + 0x46, 0xf4, 0x81, 0xdc, 0x87, 0x39, 0xc3, 0x7c, 0x71, 0xd4, 0x1f, 0x9b, 0x1d, 0x74, 0xae, 0x94, + 0x54, 0xe5, 0x4a, 0x7e, 0xda, 0x3e, 0x76, 0x1a, 0x73, 0x63, 0x11, 0xcb, 0xc8, 0x23, 0xb3, 0xed, + 0xb8, 0xe4, 0x94, 0x9a, 0x88, 0x25, 0x0b, 0x58, 0x4a, 0x6e, 0x8e, 0x9c, 0xbe, 0x6d, 0xb5, 0x07, + 0x8c, 0x9c, 0x56, 0xa5, 0x18, 0xb2, 0x2d, 0x62, 0xc9, 0x9b, 0x50, 0xdc, 0x68, 0x3d, 0xb4, 0xed, + 0x41, 0xcb, 0xf5, 0xa8, 0x04, 0xaa, 0x5c, 0xc9, 0x1a, 0x73, 0x5d, 0xda, 0xea, 0x4e, 0x89, 0x54, + 0x40, 0xd9, 0x68, 0x6d, 0x59, 0x4e, 0x55, 0xf3, 0x81, 0x79, 0x55, 0xae, 0xa4, 0x8c, 0xf9, 0x2e, + 0x36, 0x4f, 0x21, 0x6b, 0xba, 0x8f, 0x2c, 0xa8, 0x72, 0x25, 0xc1, 0x90, 0x35, 0xdd, 0x43, 0xde, + 0x02, 0xb2, 0xd1, 0xda, 0xe8, 0x1f, 0x9b, 0x1d, 0xd1, 0xea, 0x9c, 0x2a, 0x57, 0x32, 0x86, 0xd2, + 0xe5, 0x1d, 0x33, 0xd0, 0xa2, 0xe5, 0x79, 0x55, 0xae, 0xa4, 0x5d, 0xb4, 0x60, 0xfb, 0x06, 0x2c, + 0x6c, 0xb4, 0xde, 0xed, 0x07, 0x1d, 0x2e, 0xaa, 0x72, 0x65, 0xce, 0x28, 0x76, 0x59, 0xfb, 0x34, + 0x56, 0x34, 0xac, 0xa8, 0x72, 0x25, 0xc9, 0xb1, 0x82, 0x5d, 0x9c, 0xdd, 0xc6, 0xc0, 0x6e, 0x3b, + 0x3e, 0x74, 0x41, 0x95, 0x2b, 0xb2, 0x31, 0xdf, 0xc5, 0xe6, 0xa0, 0xd5, 0xc7, 0xf6, 0xd1, 0xc1, + 0xc0, 0xf4, 0xa1, 0x44, 0x95, 0x2b, 0x92, 0x51, 0xec, 0xb2, 0xf6, 0x20, 0x76, 0xcf, 0x19, 0xf7, + 0xad, 0x9e, 0x8f, 0x3d, 0x8f, 0xfa, 0x2d, 0x76, 0x59, 0x7b, 0xd0, 0x83, 0x87, 0x27, 0x8e, 0x39, + 0xf1, 0xa1, 0xa6, 0x2a, 0x57, 0x0a, 0xc6, 0x7c, 0x17, 0x9b, 0x43, 0x56, 0x43, 0x6b, 0xd0, 0x55, + 0xe5, 0xca, 0x02, 0xb5, 0x3a, 0x63, 0x0d, 0xf6, 0x42, 0x6b, 0xd0, 0x53, 0xe5, 0x0a, 0xe1, 0x58, + 0x61, 0x0d, 0x44, 0xcd, 0x30, 0x21, 0x96, 0x16, 0xd5, 0x84, 0xa0, 0x19, 0xd6, 0x18, 0xd4, 0x0c, + 0x07, 0xbe, 0xa6, 0x26, 0x44, 0xcd, 0x84, 0x90, 0x38, 0x38, 0x47, 0x5e, 0x50, 0x13, 0xa2, 0x66, + 0x38, 0x32, 0xa4, 0x19, 0x8e, 0x7d, 0x5d, 0x4d, 0x04, 0x35, 0x33, 0x85, 0x16, 0x2d, 0x97, 0xd4, + 0x44, 0x50, 0x33, 0x1c, 0x1d, 0xd4, 0x0c, 0x07, 0x5f, 0x54, 0x13, 0x01, 0xcd, 0x84, 0xb1, 0xa2, + 0xe1, 0x25, 0x35, 0x11, 0xd0, 0x8c, 0x38, 0x3b, 0x57, 0x33, 0x1c, 0xba, 0xac, 0x26, 0x44, 0xcd, + 0x88, 0x56, 0x3d, 0xcd, 0x70, 0xe8, 0x25, 0x35, 0x11, 0xd0, 0x8c, 0x88, 0xf5, 0x34, 0xc3, 0xb1, + 0x97, 0xd5, 0x44, 0x40, 0x33, 0x1c, 0x7b, 0x5d, 0xd4, 0x0c, 0x87, 0x7e, 0x2c, 0xa9, 0x09, 0x51, + 0x34, 0x1c, 0x7a, 0x33, 0x20, 0x1a, 0x8e, 0xfd, 0x84, 0x62, 0x45, 0xd5, 0x84, 0xc1, 0xe2, 0x2a, + 0x7c, 0x4a, 0xc1, 0xa2, 0x6c, 0x38, 0xd8, 0x97, 0x8d, 0x1b, 0x82, 0x4a, 0x57, 0x54, 0xc9, 0x93, + 0x8d, 0x1b, 0xc3, 0x44, 0xd9, 0x78, 0xc0, 0xab, 0x18, 0x6a, 0xb9, 0x6c, 0xa6, 0x90, 0x35, 0xdd, + 0x47, 0xaa, 0xaa, 0xe4, 0xcb, 0xc6, 0x43, 0x06, 0x64, 0xe3, 0x61, 0xaf, 0xa9, 0x92, 0x28, 0x9b, + 0x19, 0x68, 0xd1, 0x72, 0x59, 0x95, 0x44, 0xd9, 0x78, 0x68, 0x51, 0x36, 0x1e, 0xf8, 0x0b, 0xaa, + 0x24, 0xc8, 0x66, 0x1a, 0x2b, 0x1a, 0xfe, 0xa2, 0x2a, 0x09, 0xb2, 0x09, 0xce, 0x8e, 0xc9, 0xc6, + 0x83, 0xbe, 0xa1, 0x4a, 0xbe, 0x6c, 0x82, 0x56, 0xb9, 0x6c, 0x3c, 0xe8, 0x9b, 0xaa, 0x24, 0xc8, + 0x26, 0x88, 0xe5, 0xb2, 0xf1, 0xb0, 0x6f, 0x61, 0x7e, 0x73, 0x65, 0xe3, 0x61, 0x05, 0xd9, 0x78, + 0xd0, 0xdf, 0xa1, 0xb9, 0xd0, 0x93, 0x8d, 0x07, 0x15, 0x65, 0xe3, 0x61, 0x7f, 0x97, 0x62, 0x7d, + 0xd9, 0x4c, 0x83, 0xc5, 0x55, 0xf8, 0x3d, 0x0a, 0xf6, 0x65, 0xe3, 0x81, 0x57, 0xd0, 0x09, 0x2a, + 0x9b, 0x8e, 0xd9, 0x6d, 0x1f, 0x0d, 0xa8, 0xc4, 0x2a, 0x54, 0x37, 0xf5, 0xa4, 0x33, 0x3e, 0x32, + 0xa9, 0x27, 0xb6, 0x3d, 0x78, 0xec, 0xf6, 0x91, 0x15, 0x6a, 0x9c, 0xc9, 0xc7, 0x27, 0x5c, 0xa7, + 0xfa, 0xa9, 0xcb, 0x55, 0xcd, 0x28, 0x32, 0x0d, 0x4d, 0xe3, 0x6b, 0xba, 0x80, 0xbf, 0x41, 0x55, + 0x54, 0x97, 0x6b, 0x3a, 0xc3, 0xd7, 0x74, 0x1f, 0x5f, 0x85, 0xf3, 0xbe, 0x94, 0x7c, 0xc6, 0x4d, + 0xaa, 0xa5, 0x7a, 0xa2, 0xaa, 0xad, 0x1a, 0x0b, 0xae, 0xa0, 0x66, 0x91, 0x02, 0xc3, 0xdc, 0xa2, + 0x92, 0xaa, 0x27, 0x6a, 0xba, 0x47, 0x12, 0x47, 0xd2, 0xa8, 0x0c, 0xb9, 0xb0, 0x7c, 0xce, 0x6d, + 0xaa, 0xac, 0x7a, 0xb2, 0xaa, 0xad, 0xae, 0x1a, 0x0a, 0xd7, 0xd7, 0x0c, 0x4e, 0x60, 0x9c, 0x15, + 0xaa, 0xb0, 0x7a, 0xb2, 0xa6, 0x7b, 0x9c, 0xe0, 0x38, 0x0b, 0xae, 0xd0, 0x7c, 0xca, 0x97, 0xa8, + 0xd2, 0xea, 0xe9, 0xea, 0x9a, 0xbe, 0xb6, 0x7e, 0xcf, 0x28, 0x32, 0xc5, 0xf9, 0x1c, 0x9d, 0x8e, + 0xc3, 0x25, 0xe7, 0x93, 0x56, 0xa9, 0xe6, 0xea, 0x69, 0xed, 0xce, 0xda, 0x5d, 0xed, 0xae, 0xa1, + 0x70, 0xed, 0xf9, 0xac, 0x77, 0x28, 0x8b, 0x8b, 0xcf, 0x67, 0xad, 0x51, 0xf5, 0xd5, 0x95, 0x67, + 0xe6, 0x60, 0x60, 0xdf, 0x52, 0xcb, 0x2f, 0xed, 0xf1, 0xa0, 0x73, 0xad, 0x0c, 0x86, 0xc2, 0xf5, + 0x28, 0x8e, 0xba, 0xe0, 0x0a, 0xd2, 0xa7, 0xff, 0x1a, 0x3d, 0x87, 0x15, 0xea, 0x99, 0x87, 0xfd, + 0x9e, 0x65, 0x4f, 0x4c, 0xa3, 0xc8, 0xa4, 0x19, 0x5a, 0x93, 0xbd, 0xf0, 0x3a, 0xfe, 0x3a, 0xa5, + 0x2d, 0xd4, 0x13, 0xb7, 0xab, 0x1a, 0x1d, 0x69, 0xd6, 0x3a, 0xee, 0x85, 0xd7, 0xf1, 0x37, 0x28, + 0x87, 0xd4, 0x13, 0xb7, 0x6b, 0x3a, 0xe7, 0x88, 0xeb, 0x78, 0x07, 0x2e, 0x84, 0xf2, 0x62, 0x6b, + 0xd4, 0x3e, 0x7c, 0x6e, 0x76, 0x4a, 0x1a, 0x4d, 0x8f, 0x0f, 0x65, 0x45, 0x32, 0xce, 0x07, 0x52, + 0xe4, 0x2e, 0x76, 0x93, 0x7b, 0xf0, 0x7a, 0x38, 0x51, 0xba, 0xcc, 0x2a, 0xcd, 0x97, 0xc8, 0x5c, + 0x0c, 0xe6, 0xcc, 0x10, 0x55, 0x08, 0xc0, 0x2e, 0x55, 0xa7, 0x09, 0xd4, 0xa7, 0xfa, 0x91, 0x98, + 0x53, 0x7f, 0x06, 0x2e, 0x4e, 0xa7, 0x52, 0x97, 0xbc, 0x4e, 0x33, 0x2a, 0x92, 0x2f, 0x84, 0xb3, + 0xea, 0x14, 0x7d, 0xc6, 0xd8, 0x35, 0x9a, 0x62, 0x45, 0xfa, 0xd4, 0xe8, 0xf7, 0xa1, 0x34, 0x95, + 0x6c, 0x5d, 0xf6, 0x1d, 0x9a, 0x73, 0x91, 0xfd, 0x5a, 0x28, 0xef, 0x86, 0xc9, 0x33, 0x86, 0xbe, + 0x4b, 0x93, 0xb0, 0x40, 0x9e, 0x1a, 0x19, 0x97, 0x2c, 0x98, 0x8e, 0x5d, 0xee, 0x3d, 0x9a, 0x95, + 0xf9, 0x92, 0x05, 0x32, 0xb3, 0x38, 0x6e, 0x28, 0x3f, 0xbb, 0xdc, 0x3a, 0x4d, 0xd3, 0x7c, 0xdc, + 0x60, 0xaa, 0xe6, 0xe4, 0xb7, 0x29, 0x79, 0x6f, 0xf6, 0x8c, 0x7f, 0x9c, 0xa0, 0x09, 0x96, 0xb3, + 0xf7, 0x66, 0x4d, 0xd9, 0x63, 0xcf, 0x98, 0xf2, 0x4f, 0x28, 0x9b, 0x08, 0xec, 0xa9, 0x39, 0x3f, + 0x06, 0xaf, 0xe2, 0xe8, 0x8d, 0xed, 0xa3, 0x51, 0x69, 0x43, 0x95, 0x2b, 0xa0, 0x5d, 0x99, 0xaa, + 0x7e, 0xdc, 0x43, 0xde, 0x26, 0x45, 0x19, 0x41, 0x12, 0xb3, 0xc2, 0xec, 0x32, 0x2b, 0xbb, 0x6a, + 0x22, 0xc2, 0x0a, 0x43, 0x79, 0x56, 0x04, 0x12, 0xb5, 0xe2, 0x06, 0x7d, 0x66, 0xe5, 0x03, 0x55, + 0x9a, 0x69, 0xc5, 0x4d, 0x01, 0xdc, 0x4a, 0x80, 0xb4, 0xb4, 0xee, 0xd7, 0x5b, 0xd8, 0x4f, 0xbe, + 0x18, 0x2e, 0xc0, 0x36, 0xf1, 0xfc, 0x1c, 0xac, 0xb4, 0x18, 0x4d, 0x70, 0x6e, 0x9a, 0xf6, 0xb3, + 0x11, 0xb4, 0x80, 0x37, 0xd3, 0xb4, 0x9f, 0x9b, 0x41, 0x2b, 0xff, 0xa6, 0x04, 0x49, 0x5a, 0x4f, + 0x92, 0x2c, 0x24, 0xdf, 0x6b, 0x6e, 0x3d, 0x56, 0xce, 0xd1, 0x5f, 0x0f, 0x9b, 0xcd, 0xa7, 0x8a, + 0x44, 0x72, 0x90, 0x7a, 0xf8, 0x95, 0xfd, 0xc6, 0x9e, 0x22, 0x93, 0x22, 0xe4, 0x37, 0xb6, 0x76, + 0x36, 0x1b, 0xc6, 0xae, 0xb1, 0xb5, 0xb3, 0xaf, 0x24, 0x68, 0xdf, 0xc6, 0xd3, 0xe6, 0x83, 0x7d, + 0x25, 0x49, 0x32, 0x90, 0xa0, 0x6d, 0x29, 0x02, 0x90, 0xde, 0xdb, 0x37, 0xb6, 0x76, 0x36, 0x95, + 0x34, 0xb5, 0xb2, 0xbf, 0xb5, 0xdd, 0x50, 0x32, 0x14, 0xb9, 0xff, 0xee, 0xee, 0xd3, 0x86, 0x92, + 0xa5, 0x3f, 0x1f, 0x18, 0xc6, 0x83, 0xaf, 0x28, 0x39, 0x4a, 0xda, 0x7e, 0xb0, 0xab, 0x00, 0x76, + 0x3f, 0x78, 0xf8, 0xb4, 0xa1, 0xe4, 0x49, 0x01, 0xb2, 0x1b, 0xef, 0xee, 0x3c, 0xda, 0xdf, 0x6a, + 0xee, 0x28, 0x85, 0xf2, 0x6f, 0xc9, 0x00, 0x9b, 0xf6, 0xde, 0xf3, 0xfe, 0x08, 0xab, 0xe2, 0xcb, + 0x00, 0x93, 0xe7, 0xfd, 0x51, 0x0b, 0xa5, 0xc7, 0x2b, 0xbb, 0x1c, 0x6d, 0xc1, 0xa0, 0x43, 0xae, + 0x41, 0x01, 0xbb, 0xbb, 0x2c, 0x14, 0x60, 0x41, 0x97, 0x31, 0xf2, 0xb4, 0x8d, 0x47, 0x87, 0x20, + 0xa4, 0xa6, 0x63, 0x1d, 0x97, 0x16, 0x20, 0x35, 0x9d, 0x5c, 0x05, 0x7c, 0x6c, 0x4d, 0x30, 0xac, + 0x63, 0xed, 0x96, 0x33, 0x70, 0x5c, 0x16, 0xe8, 0xc9, 0xdb, 0x80, 0x63, 0x32, 0x59, 0x14, 0xa7, + 0x25, 0xea, 0xba, 0xbb, 0x42, 0x7f, 0x30, 0x59, 0xf8, 0x84, 0xa5, 0x26, 0xe4, 0xbc, 0x76, 0x3a, + 0x16, 0xb6, 0xf2, 0x19, 0x29, 0x38, 0x23, 0xc0, 0x26, 0x6f, 0x4a, 0x0c, 0xc0, 0xbd, 0x59, 0x40, + 0x6f, 0x18, 0x89, 0xb9, 0x53, 0xbe, 0x0c, 0x73, 0x3b, 0xb6, 0xc5, 0xb6, 0x10, 0xae, 0x52, 0x01, + 0xa4, 0x76, 0x49, 0xc2, 0x12, 0x46, 0x6a, 0x97, 0xaf, 0x00, 0x08, 0x7d, 0x0a, 0x48, 0x07, 0xac, + 0x0f, 0x37, 0xa2, 0x74, 0x50, 0xbe, 0x09, 0xe9, 0xed, 0xf6, 0xf1, 0x7e, 0xbb, 0x47, 0xae, 0x01, + 0x0c, 0xda, 0x13, 0xa7, 0xd5, 0x45, 0xa9, 0x7c, 0xfe, 0xf9, 0xe7, 0x9f, 0x4b, 0x78, 0xe2, 0xca, + 0xd1, 0x56, 0x26, 0x95, 0x17, 0x00, 0xcd, 0x41, 0x67, 0xdb, 0x9c, 0x4c, 0xda, 0x3d, 0x93, 0x54, + 0x21, 0x6d, 0x99, 0x13, 0x9a, 0x72, 0x24, 0x2c, 0xe6, 0x97, 0xfd, 0x55, 0xf0, 0x51, 0x2b, 0x3b, + 0x08, 0x31, 0x38, 0x94, 0x28, 0x90, 0xb0, 0x8e, 0x86, 0x78, 0x59, 0x91, 0x32, 0xe8, 0xcf, 0xa5, + 0x4b, 0x90, 0x66, 0x18, 0x42, 0x20, 0x69, 0xb5, 0x87, 0x66, 0x89, 0x8d, 0x8b, 0xbf, 0xcb, 0xbf, + 0x2a, 0x01, 0xec, 0x98, 0x2f, 0xcf, 0x30, 0xa6, 0x8f, 0x8a, 0x19, 0x33, 0xc1, 0xc6, 0xbc, 0x1f, + 0x37, 0x26, 0xd5, 0x59, 0xd7, 0xb6, 0x3b, 0x2d, 0xf6, 0x8a, 0xd9, 0xbd, 0x4a, 0x8e, 0xb6, 0xe0, + 0x5b, 0x2b, 0x7f, 0x00, 0x85, 0x2d, 0xcb, 0x32, 0xc7, 0xae, 0x4f, 0x04, 0x92, 0xcf, 0xec, 0x89, + 0xc3, 0x2f, 0x78, 0xf0, 0x37, 0x29, 0x41, 0x72, 0x64, 0x8f, 0x1d, 0x36, 0xcf, 0x7a, 0x52, 0x5f, + 0x5d, 0x5d, 0x35, 0xb0, 0x85, 0x5c, 0x82, 0xdc, 0xa1, 0x6d, 0x59, 0xe6, 0x21, 0x9d, 0x44, 0x02, + 0x6b, 0x0b, 0xbf, 0xa1, 0xfc, 0xcb, 0x12, 0x14, 0x9a, 0xce, 0x33, 0xdf, 0xb8, 0x02, 0x89, 0xe7, + 0xe6, 0x09, 0xba, 0x97, 0x30, 0xe8, 0x4f, 0xb2, 0x08, 0xa9, 0x9f, 0x6f, 0x0f, 0x8e, 0xd8, 0x85, + 0x4f, 0xc1, 0x60, 0x0f, 0xe4, 0x02, 0xa4, 0x5f, 0x9a, 0xfd, 0xde, 0x33, 0x07, 0x6d, 0xca, 0x06, + 0x7f, 0x22, 0xb7, 0x20, 0xd5, 0xa7, 0xce, 0x96, 0x92, 0xb8, 0x5e, 0x17, 0xfc, 0xf5, 0x12, 0xe7, + 0x60, 0x30, 0xd0, 0x8d, 0x6c, 0xb6, 0xa3, 0x7c, 0xf4, 0xd1, 0x47, 0x1f, 0xc9, 0xe5, 0x2e, 0x2c, + 0xba, 0xb1, 0x23, 0x30, 0xd9, 0x1d, 0x28, 0x0d, 0x4c, 0xbb, 0xd5, 0xed, 0x5b, 0xed, 0xc1, 0xe0, + 0xa4, 0xf5, 0xd2, 0xb6, 0x5a, 0x6d, 0xab, 0x65, 0x4f, 0x0e, 0xdb, 0x63, 0x5c, 0x80, 0xe8, 0x21, + 0x16, 0x07, 0xa6, 0xbd, 0xc1, 0x68, 0xef, 0xdb, 0xd6, 0x03, 0xab, 0x49, 0x39, 0xe5, 0x3f, 0x48, + 0x42, 0x6e, 0xfb, 0xc4, 0xb5, 0xbe, 0x08, 0xa9, 0x43, 0xfb, 0xc8, 0x62, 0x6b, 0x99, 0x32, 0xd8, + 0x83, 0xf7, 0x8e, 0x64, 0xe1, 0x1d, 0x2d, 0x42, 0xea, 0xc5, 0x91, 0xed, 0x98, 0x38, 0xdd, 0x9c, + 0xc1, 0x1e, 0xe8, 0x6a, 0x8d, 0x4c, 0xa7, 0x94, 0xc4, 0x0a, 0x93, 0xfe, 0xf4, 0xe7, 0x9f, 0x3a, + 0xc3, 0xfc, 0xc9, 0x0a, 0xa4, 0x6d, 0xba, 0xfa, 0x93, 0x52, 0x1a, 0x2f, 0xb7, 0x04, 0xb8, 0xf8, + 0x56, 0x0c, 0x8e, 0x22, 0x5b, 0xb0, 0xf0, 0xd2, 0x6c, 0x0d, 0x8f, 0x26, 0x4e, 0xab, 0x67, 0xb7, + 0x3a, 0xa6, 0x39, 0x32, 0xc7, 0xa5, 0x39, 0x1c, 0x49, 0x88, 0x09, 0xb3, 0x16, 0xd2, 0x98, 0x7f, + 0x69, 0x6e, 0x1f, 0x4d, 0x9c, 0x4d, 0xfb, 0x31, 0xb2, 0x48, 0x15, 0x72, 0x63, 0x93, 0x46, 0x02, + 0xea, 0x6c, 0x21, 0x3c, 0x7a, 0x80, 0x9a, 0x1d, 0x9b, 0x23, 0x6c, 0x20, 0xeb, 0x90, 0x3d, 0xe8, + 0x3f, 0x37, 0x27, 0xcf, 0xcc, 0x4e, 0x29, 0xa3, 0x4a, 0x95, 0x79, 0xed, 0xa2, 0xcf, 0xf1, 0x96, + 0x75, 0xe5, 0x91, 0x3d, 0xb0, 0xc7, 0x86, 0x07, 0x25, 0xf7, 0x21, 0x37, 0xb1, 0x87, 0x26, 0xd3, + 0x77, 0x16, 0x33, 0xdb, 0xe5, 0x59, 0xbc, 0x3d, 0x7b, 0x68, 0xba, 0x11, 0xcc, 0xc5, 0x93, 0x65, + 0xe6, 0xe8, 0x01, 0x3d, 0xbf, 0x96, 0x00, 0xeb, 0x73, 0xea, 0x10, 0x9e, 0x67, 0xc9, 0x12, 0x75, + 0xa8, 0xd7, 0xa5, 0xc7, 0x92, 0x52, 0x1e, 0x8b, 0x3b, 0xef, 0x79, 0xe9, 0x16, 0xe4, 0x3c, 0x83, + 0x7e, 0xe8, 0x63, 0xe1, 0x26, 0x87, 0xf1, 0x80, 0x85, 0x3e, 0x16, 0x6b, 0xde, 0x80, 0x14, 0xba, + 0x4d, 0xd3, 0x84, 0xd1, 0xa0, 0x59, 0x29, 0x07, 0xa9, 0x4d, 0xa3, 0xd1, 0xd8, 0x51, 0x24, 0x4c, + 0x50, 0x4f, 0xdf, 0x6d, 0x28, 0xb2, 0xa0, 0xd8, 0xdf, 0x96, 0x20, 0xd1, 0x38, 0x46, 0xb5, 0xd0, + 0x69, 0xb8, 0x3b, 0x9a, 0xfe, 0xd6, 0x6a, 0x90, 0x1c, 0xda, 0x63, 0x93, 0x9c, 0x9f, 0x31, 0xcb, + 0x52, 0x0f, 0xdf, 0x97, 0x70, 0x95, 0xdb, 0x38, 0x76, 0x0c, 0xc4, 0x6b, 0x6f, 0x41, 0xd2, 0x31, + 0x8f, 0x9d, 0xd9, 0xbc, 0x67, 0x6c, 0x00, 0x0a, 0xd0, 0x6e, 0x42, 0xda, 0x3a, 0x1a, 0x1e, 0x98, + 0xe3, 0xd9, 0xd0, 0x3e, 0x4e, 0x8f, 0x43, 0xca, 0xef, 0x81, 0xf2, 0xc8, 0x1e, 0x8e, 0x06, 0xe6, + 0x71, 0xe3, 0xd8, 0x31, 0xad, 0x49, 0xdf, 0xb6, 0xa8, 0x9e, 0xbb, 0xfd, 0x31, 0x46, 0x11, 0xbc, + 0xb0, 0xc5, 0x07, 0xba, 0xab, 0x27, 0xe6, 0xa1, 0x6d, 0x75, 0x78, 0xc0, 0xe4, 0x4f, 0x14, 0xed, + 0x3c, 0xeb, 0x8f, 0x69, 0x00, 0xa1, 0x71, 0x9e, 0x3d, 0x94, 0x37, 0xa1, 0xc8, 0x0f, 0xfa, 0x13, + 0x3e, 0x70, 0xf9, 0x06, 0x14, 0xdc, 0x26, 0xbc, 0xbd, 0xce, 0x42, 0xf2, 0x83, 0x86, 0xd1, 0x54, + 0xce, 0xd1, 0x65, 0x6d, 0xee, 0x34, 0x14, 0x89, 0xfe, 0xd8, 0x7f, 0xbf, 0x19, 0x58, 0xca, 0x4b, + 0x50, 0xf0, 0x7c, 0xdf, 0x33, 0x1d, 0xec, 0xa1, 0x09, 0x21, 0x53, 0x97, 0xb3, 0x52, 0x39, 0x03, + 0xa9, 0xc6, 0x70, 0xe4, 0x9c, 0x94, 0x7f, 0x11, 0xf2, 0x1c, 0xf4, 0xb4, 0x3f, 0x71, 0xc8, 0x1d, + 0xc8, 0x0c, 0xf9, 0x7c, 0x25, 0x3c, 0x73, 0x89, 0x9a, 0xf2, 0x71, 0xee, 0x6f, 0xc3, 0x45, 0x2f, + 0x55, 0x21, 0x23, 0xc4, 0x52, 0xbe, 0xd5, 0x65, 0x71, 0xab, 0xb3, 0xa0, 0x90, 0x10, 0x82, 0x42, + 0x79, 0x1b, 0x32, 0x2c, 0x03, 0x4e, 0x30, 0xab, 0xb3, 0x7a, 0x8d, 0x89, 0x89, 0xbd, 0xf9, 0x3c, + 0x6b, 0x63, 0x57, 0xc8, 0x57, 0x21, 0x8f, 0x82, 0xe5, 0x08, 0x16, 0x3a, 0x01, 0x9b, 0x98, 0xdc, + 0x7e, 0x3f, 0x05, 0x59, 0x77, 0xa5, 0xc8, 0x32, 0xa4, 0x59, 0x91, 0x84, 0xa6, 0xdc, 0x22, 0x3e, + 0x85, 0x65, 0x11, 0x59, 0x86, 0x0c, 0x2f, 0x84, 0x78, 0x74, 0xa7, 0x15, 0x7b, 0x9a, 0x15, 0x3e, + 0x5e, 0x67, 0x4d, 0xc7, 0xc0, 0xc4, 0xca, 0xf3, 0x34, 0x2b, 0x6d, 0x88, 0x0a, 0x39, 0xaf, 0x98, + 0xc1, 0x78, 0xcc, 0x6b, 0xf1, 0xac, 0x5b, 0xbd, 0x08, 0x88, 0x9a, 0x8e, 0x11, 0x8b, 0x17, 0xde, + 0xd9, 0xae, 0x7f, 0x3c, 0xc9, 0xba, 0x25, 0x09, 0xde, 0xa1, 0xbb, 0x55, 0x76, 0x86, 0x17, 0x21, + 0x3e, 0xa0, 0xa6, 0x63, 0x48, 0x70, 0x4b, 0xea, 0x0c, 0x2f, 0x34, 0xc8, 0x55, 0xea, 0x22, 0x16, + 0x0e, 0xb8, 0xf5, 0xfd, 0xfa, 0x39, 0xcd, 0xca, 0x09, 0x72, 0x8d, 0x5a, 0x60, 0xd5, 0x01, 0xee, + 0x4b, 0xbf, 0x58, 0xce, 0xf0, 0xa2, 0x81, 0xdc, 0xa4, 0x10, 0xb6, 0xfc, 0x25, 0x88, 0xa8, 0x8c, + 0x33, 0xbc, 0x32, 0x26, 0x2a, 0x1d, 0x10, 0xc3, 0x03, 0x86, 0x04, 0xa1, 0x0a, 0x4e, 0xb3, 0x2a, + 0x98, 0x5c, 0x41, 0x73, 0x6c, 0x52, 0x05, 0xbf, 0xe2, 0xcd, 0xf0, 0x2a, 0xc3, 0xef, 0xc7, 0x23, + 0x9b, 0x57, 0xdd, 0x66, 0x78, 0x1d, 0x41, 0x6a, 0xf4, 0x7d, 0x51, 0x7d, 0x97, 0xe6, 0x31, 0x08, + 0x96, 0x7c, 0xe1, 0xb9, 0xef, 0x94, 0xc5, 0xc0, 0x3a, 0x8b, 0x20, 0x46, 0xaa, 0x8b, 0xbb, 0x61, + 0x89, 0xf2, 0x76, 0xfb, 0x56, 0xb7, 0x54, 0xc4, 0x95, 0x48, 0xf4, 0xad, 0xae, 0x91, 0xea, 0xd2, + 0x16, 0xa6, 0x81, 0x1d, 0xda, 0xa7, 0x60, 0x5f, 0xf2, 0x36, 0xeb, 0xa4, 0x4d, 0xa4, 0x04, 0xa9, + 0x8d, 0xd6, 0x4e, 0xdb, 0x2a, 0x2d, 0x30, 0x9e, 0xd5, 0xb6, 0x8c, 0x64, 0x77, 0xa7, 0x6d, 0x91, + 0xb7, 0x20, 0x31, 0x39, 0x3a, 0x28, 0x91, 0xf0, 0xe7, 0x8d, 0xbd, 0xa3, 0x03, 0xd7, 0x15, 0x83, + 0x22, 0xc8, 0x32, 0x64, 0x27, 0xce, 0xb8, 0xf5, 0x0b, 0xe6, 0xd8, 0x2e, 0x9d, 0xc7, 0x25, 0x3c, + 0x67, 0x64, 0x26, 0xce, 0xf8, 0x03, 0x73, 0x6c, 0x9f, 0x31, 0xf8, 0x95, 0xaf, 0x40, 0x5e, 0xb0, + 0x4b, 0x8a, 0x20, 0x59, 0xec, 0xa4, 0x50, 0x97, 0xee, 0x18, 0x92, 0x55, 0xde, 0x87, 0x82, 0x5b, + 0x48, 0xe0, 0x7c, 0x35, 0xba, 0x93, 0x06, 0xf6, 0x18, 0xf7, 0xe7, 0xbc, 0x76, 0x49, 0x4c, 0x51, + 0x3e, 0x8c, 0xa7, 0x0b, 0x06, 0x2d, 0x2b, 0x21, 0x57, 0xa4, 0xf2, 0x0f, 0x25, 0x28, 0x6c, 0xdb, + 0x63, 0xff, 0x96, 0x77, 0x11, 0x52, 0x07, 0xb6, 0x3d, 0x98, 0xa0, 0xd9, 0xac, 0xc1, 0x1e, 0xc8, + 0x1b, 0x50, 0xc0, 0x1f, 0x6e, 0x01, 0x28, 0x7b, 0xf7, 0x0b, 0x79, 0x6c, 0xe7, 0x55, 0x1f, 0x81, + 0x64, 0xdf, 0x72, 0x26, 0x3c, 0x92, 0xe1, 0x6f, 0xf2, 0x05, 0xc8, 0xd3, 0xbf, 0x2e, 0x33, 0xe9, + 0x1d, 0x58, 0x81, 0x36, 0x73, 0xe2, 0x5b, 0x30, 0x87, 0x6f, 0xdf, 0x83, 0x65, 0xbc, 0xbb, 0x84, + 0x02, 0xeb, 0xe0, 0xc0, 0x12, 0x64, 0x58, 0x28, 0x98, 0xe0, 0x27, 0xab, 0x9c, 0xe1, 0x3e, 0xd2, + 0xf0, 0x8a, 0x95, 0x00, 0x4b, 0xf7, 0x19, 0x83, 0x3f, 0x95, 0x1f, 0x40, 0x16, 0xb3, 0x54, 0x73, + 0xd0, 0x21, 0x65, 0x90, 0x7a, 0x25, 0x13, 0x73, 0xe4, 0xa2, 0x70, 0xcc, 0xe7, 0xdd, 0x2b, 0x9b, + 0x86, 0xd4, 0x5b, 0x5a, 0x00, 0x69, 0x93, 0x9e, 0xbb, 0x8f, 0x79, 0x98, 0x96, 0x8e, 0xcb, 0x4d, + 0x6e, 0x62, 0xc7, 0x7c, 0x19, 0x67, 0x62, 0xc7, 0x7c, 0xc9, 0x4c, 0x5c, 0x9d, 0x32, 0x41, 0x9f, + 0x4e, 0xf8, 0xf7, 0x3b, 0xe9, 0x84, 0x9e, 0xf3, 0x71, 0x7b, 0xf6, 0xad, 0xde, 0xae, 0xdd, 0xb7, + 0xf0, 0x9c, 0xdf, 0xc5, 0x73, 0x92, 0x64, 0x48, 0xdd, 0xf2, 0x67, 0x49, 0x98, 0xe7, 0x41, 0xf4, + 0xfd, 0xbe, 0xf3, 0x6c, 0xbb, 0x3d, 0x22, 0x4f, 0xa1, 0x40, 0xe3, 0x67, 0x6b, 0xd8, 0x1e, 0x8d, + 0xe8, 0x46, 0x95, 0xf0, 0x50, 0x71, 0x7d, 0x2a, 0x28, 0x73, 0xfc, 0xca, 0x4e, 0x7b, 0x68, 0x6e, + 0x33, 0x6c, 0xc3, 0x72, 0xc6, 0x27, 0x46, 0xde, 0xf2, 0x5b, 0xc8, 0x16, 0xe4, 0x87, 0x93, 0x9e, + 0x67, 0x4c, 0x46, 0x63, 0x95, 0x48, 0x63, 0xdb, 0x93, 0x5e, 0xc0, 0x16, 0x0c, 0xbd, 0x06, 0xea, + 0x18, 0x8d, 0xbc, 0x9e, 0xad, 0xc4, 0x29, 0x8e, 0xd1, 0x20, 0x11, 0x74, 0xec, 0xc0, 0x6f, 0x21, + 0x8f, 0x01, 0xe8, 0x46, 0x72, 0x6c, 0x5a, 0x24, 0xa1, 0x56, 0xf2, 0xda, 0x9b, 0x91, 0xb6, 0xf6, + 0x9c, 0xf1, 0xbe, 0xbd, 0xe7, 0x8c, 0x99, 0x21, 0xba, 0x05, 0xf1, 0x71, 0xe9, 0x1d, 0x50, 0xc2, + 0xf3, 0x17, 0xcf, 0xde, 0xa9, 0x19, 0x67, 0xef, 0x1c, 0x3f, 0x7b, 0xd7, 0xe5, 0xbb, 0xd2, 0xd2, + 0x7b, 0x50, 0x0c, 0x4d, 0x59, 0xa4, 0x13, 0x46, 0xbf, 0x2d, 0xd2, 0xf3, 0xda, 0xeb, 0xc2, 0xd7, + 0x63, 0xf1, 0xd5, 0x8a, 0x76, 0xdf, 0x01, 0x25, 0x3c, 0x7d, 0xd1, 0x70, 0x36, 0xa6, 0x26, 0x40, + 0xfe, 0x7d, 0x98, 0x0b, 0x4c, 0x59, 0x24, 0xe7, 0x4e, 0x99, 0x54, 0xf9, 0x97, 0x52, 0x90, 0x6a, + 0x5a, 0xa6, 0xdd, 0x25, 0xaf, 0x07, 0x33, 0xe2, 0x93, 0x73, 0x6e, 0x36, 0xbc, 0x18, 0xca, 0x86, + 0x4f, 0xce, 0x79, 0xb9, 0xf0, 0x62, 0x28, 0x17, 0xba, 0x5d, 0x35, 0x9d, 0x5c, 0x9e, 0xca, 0x84, + 0x4f, 0xce, 0x09, 0x69, 0xf0, 0xf2, 0x54, 0x1a, 0xf4, 0xbb, 0x6b, 0x3a, 0x0d, 0x9d, 0xc1, 0x1c, + 0xf8, 0xe4, 0x9c, 0x9f, 0xff, 0x96, 0xc3, 0xf9, 0xcf, 0xeb, 0xac, 0xe9, 0xcc, 0x25, 0x21, 0xf7, + 0xa1, 0x4b, 0x2c, 0xeb, 0x2d, 0x87, 0xb3, 0x1e, 0xf2, 0x78, 0xbe, 0x5b, 0x0e, 0xe7, 0x3b, 0xec, + 0xe4, 0xf9, 0xed, 0x62, 0x28, 0xbf, 0xa1, 0x51, 0x96, 0xd8, 0x96, 0xc3, 0x89, 0x8d, 0xf1, 0x04, + 0x4f, 0xc5, 0xac, 0xe6, 0x75, 0xd6, 0x74, 0xa2, 0x85, 0x52, 0x5a, 0xf4, 0xb9, 0x1e, 0xdf, 0x05, + 0x86, 0x77, 0x9d, 0x2e, 0x9b, 0x7b, 0xe4, 0x2c, 0xc6, 0x7c, 0x60, 0xc7, 0xd5, 0x74, 0x8f, 0x5c, + 0x1a, 0x64, 0xba, 0xbc, 0xd4, 0x55, 0x30, 0x46, 0x09, 0xb2, 0xc4, 0x97, 0xbf, 0xb2, 0xd1, 0xc2, + 0x58, 0x85, 0xf3, 0x62, 0xa7, 0xf7, 0x0a, 0xcc, 0x6d, 0xb4, 0x9e, 0xb6, 0xc7, 0x3d, 0x73, 0xe2, + 0xb4, 0xf6, 0xdb, 0x3d, 0xef, 0xba, 0x80, 0xbe, 0xff, 0x7c, 0x97, 0xf7, 0xec, 0xb7, 0x7b, 0xe4, + 0x82, 0x2b, 0xae, 0x0e, 0xf6, 0x4a, 0x5c, 0x5e, 0x4b, 0xaf, 0xd3, 0x45, 0x63, 0xc6, 0x30, 0xea, + 0x2d, 0xf0, 0xa8, 0xf7, 0x30, 0x03, 0xa9, 0x23, 0xab, 0x6f, 0x5b, 0x0f, 0x73, 0x90, 0x71, 0xec, + 0xf1, 0xb0, 0xed, 0xd8, 0xe5, 0x1f, 0x49, 0x00, 0x8f, 0xec, 0xe1, 0xf0, 0xc8, 0xea, 0xbf, 0x38, + 0x32, 0xc9, 0x15, 0xc8, 0x0f, 0xdb, 0xcf, 0xcd, 0xd6, 0xd0, 0x6c, 0x1d, 0x8e, 0xdd, 0x7d, 0x90, + 0xa3, 0x4d, 0xdb, 0xe6, 0xa3, 0xf1, 0x09, 0x29, 0xb9, 0x87, 0x71, 0xd4, 0x0e, 0x4a, 0x92, 0x1f, + 0xce, 0x17, 0xf9, 0xf1, 0x32, 0xcd, 0xdf, 0xa1, 0x7b, 0xc0, 0x64, 0x15, 0x43, 0x86, 0xbf, 0x3d, + 0x7c, 0xa2, 0x92, 0x77, 0xcc, 0xe1, 0xa8, 0x75, 0x88, 0x52, 0xa1, 0x72, 0x48, 0xd1, 0xe7, 0x47, + 0xe4, 0x36, 0x24, 0x0e, 0xed, 0x01, 0x8a, 0xe4, 0x94, 0xf7, 0x42, 0x71, 0xe4, 0x0d, 0x48, 0x0c, + 0x27, 0x4c, 0x36, 0x79, 0x6d, 0x41, 0x38, 0x11, 0xb0, 0x24, 0x44, 0x61, 0xc3, 0x49, 0xcf, 0x9b, + 0xf7, 0x8d, 0x22, 0x24, 0x36, 0x9a, 0x4d, 0x9a, 0xe5, 0x37, 0x9a, 0xcd, 0x35, 0x45, 0xaa, 0x7f, + 0x09, 0xb2, 0xbd, 0xb1, 0x69, 0xd2, 0xf0, 0x30, 0xbb, 0xba, 0xf8, 0x10, 0xb3, 0x9a, 0x07, 0xaa, + 0x6f, 0x43, 0xe6, 0x90, 0xd5, 0x17, 0x24, 0xa2, 0x80, 0x2d, 0xfd, 0x21, 0xbb, 0x3e, 0x59, 0xf2, + 0xbb, 0xc3, 0x15, 0x89, 0xe1, 0xda, 0xa8, 0xef, 0x42, 0x6e, 0xdc, 0x3a, 0xcd, 0xe0, 0xc7, 0x2c, + 0xbb, 0xc4, 0x19, 0xcc, 0x8e, 0x79, 0x53, 0xbd, 0x01, 0x0b, 0x96, 0xed, 0x7e, 0xb2, 0x68, 0x75, + 0xd8, 0x1e, 0xbb, 0x38, 0x7d, 0x68, 0x73, 0x8d, 0x9b, 0xec, 0x33, 0xa1, 0x65, 0xf3, 0x0e, 0xb6, + 0x2b, 0xeb, 0x8f, 0x40, 0x11, 0xcc, 0x60, 0x91, 0x19, 0x67, 0xa5, 0xcb, 0xbe, 0x4b, 0x7a, 0x56, + 0x70, 0xdf, 0x87, 0x8c, 0xb0, 0x9d, 0x19, 0x63, 0xa4, 0xc7, 0x3e, 0xf2, 0x7a, 0x46, 0x30, 0xd4, + 0x4d, 0x1b, 0xa1, 0xb1, 0x26, 0xda, 0xc8, 0x33, 0xf6, 0xfd, 0x57, 0x34, 0x52, 0xd3, 0x43, 0xab, + 0x72, 0x74, 0xaa, 0x2b, 0x7d, 0xf6, 0xf9, 0xd6, 0xb3, 0xc2, 0x02, 0xe0, 0x0c, 0x33, 0xf1, 0xce, + 0x7c, 0xc8, 0xbe, 0xec, 0x06, 0xcc, 0x4c, 0x79, 0x33, 0x39, 0xd5, 0x9b, 0xe7, 0xec, 0x33, 0xaa, + 0x67, 0x66, 0x6f, 0x96, 0x37, 0x93, 0x53, 0xbd, 0x19, 0xb0, 0x0f, 0xac, 0x01, 0x33, 0x35, 0xbd, + 0xbe, 0x09, 0x44, 0x7c, 0xd5, 0x3c, 0x4f, 0xc4, 0xd8, 0x19, 0xb2, 0xcf, 0xe6, 0xfe, 0xcb, 0x66, + 0x94, 0x59, 0x86, 0xe2, 0x1d, 0xb2, 0xd8, 0x17, 0xf5, 0xa0, 0xa1, 0x9a, 0x5e, 0xdf, 0x82, 0xf3, + 0xe2, 0xc4, 0xce, 0xe0, 0x92, 0xad, 0x4a, 0x95, 0xa2, 0xb1, 0xe0, 0x4f, 0x8d, 0x73, 0x66, 0x9a, + 0x8a, 0x77, 0x6a, 0xa4, 0x4a, 0x15, 0x65, 0xca, 0x54, 0x4d, 0xaf, 0x3f, 0x80, 0xa2, 0x60, 0xea, + 0x00, 0x33, 0x74, 0xb4, 0x99, 0x17, 0xec, 0x5f, 0x1b, 0x3c, 0x33, 0x34, 0xa3, 0x87, 0xdf, 0x18, + 0xcf, 0x71, 0xd1, 0x46, 0xc6, 0xec, 0xbb, 0xbc, 0xef, 0x0b, 0x32, 0x42, 0x5b, 0x02, 0x2b, 0xed, + 0x38, 0x2b, 0x13, 0xf6, 0xc5, 0xde, 0x77, 0x85, 0x12, 0xea, 0xfd, 0xc0, 0x74, 0x4c, 0x9a, 0xe4, + 0x62, 0x6c, 0x38, 0x18, 0x91, 0xdf, 0x8c, 0x04, 0xac, 0x88, 0x57, 0x21, 0xc2, 0xb4, 0xe9, 0x63, + 0x7d, 0x0b, 0xe6, 0xcf, 0x1e, 0x90, 0x3e, 0x96, 0x58, 0x5d, 0x5c, 0x5d, 0xa1, 0xa5, 0xb3, 0x31, + 0xd7, 0x09, 0xc4, 0xa5, 0x06, 0xcc, 0x9d, 0x39, 0x28, 0x7d, 0x22, 0xb1, 0xea, 0x92, 0x5a, 0x32, + 0x0a, 0x9d, 0x60, 0x64, 0x9a, 0x3b, 0x73, 0x58, 0xfa, 0x54, 0x62, 0x57, 0x11, 0xba, 0xe6, 0x19, + 0x71, 0x23, 0xd3, 0xdc, 0x99, 0xc3, 0xd2, 0x57, 0x59, 0xed, 0x28, 0xeb, 0x55, 0xd1, 0x08, 0xc6, + 0x82, 0xf9, 0xb3, 0x87, 0xa5, 0xaf, 0x49, 0x78, 0x2d, 0x21, 0xeb, 0xba, 0xb7, 0x2e, 0x5e, 0x64, + 0x9a, 0x3f, 0x7b, 0x58, 0xfa, 0xba, 0x84, 0x97, 0x17, 0xb2, 0xbe, 0x1e, 0x30, 0x13, 0xf4, 0xe6, + 0xf4, 0xb0, 0xf4, 0x0d, 0x09, 0xef, 0x13, 0x64, 0xbd, 0xe6, 0x99, 0xd9, 0x9b, 0xf2, 0xe6, 0xf4, + 0xb0, 0xf4, 0x4d, 0x3c, 0xc5, 0xd7, 0x65, 0xfd, 0x4e, 0xc0, 0x0c, 0x46, 0xa6, 0xe2, 0x2b, 0x84, + 0xa5, 0x6f, 0x49, 0x78, 0xed, 0x23, 0xeb, 0x77, 0x0d, 0x77, 0x74, 0x3f, 0x32, 0x15, 0x5f, 0x21, + 0x2c, 0x7d, 0x26, 0xe1, 0xed, 0x90, 0xac, 0xdf, 0x0b, 0x1a, 0xc2, 0xc8, 0xa4, 0xbc, 0x4a, 0x58, + 0xfa, 0x36, 0xb5, 0x54, 0xac, 0xcb, 0xeb, 0xab, 0x86, 0xeb, 0x80, 0x10, 0x99, 0x94, 0x57, 0x09, + 0x4b, 0xdf, 0xa1, 0xa6, 0x94, 0xba, 0xbc, 0xbe, 0x16, 0x32, 0x55, 0xd3, 0xeb, 0x8f, 0xa0, 0x70, + 0xd6, 0xb0, 0xf4, 0x5d, 0xf1, 0xd6, 0x2d, 0xdf, 0x11, 0x62, 0xd3, 0xae, 0xf0, 0xce, 0x4e, 0x0d, + 0x4c, 0xdf, 0xc3, 0x1a, 0xa7, 0x3e, 0xf7, 0x84, 0xdd, 0x4c, 0x31, 0x82, 0xff, 0xfa, 0x58, 0x98, + 0xda, 0xf6, 0xf7, 0xc7, 0xa9, 0x31, 0xea, 0xfb, 0x12, 0x5e, 0x5f, 0x15, 0xb8, 0x41, 0xc4, 0x7b, + 0x3b, 0x85, 0x05, 0xac, 0x0f, 0xfd, 0x59, 0x9e, 0x16, 0xad, 0x7e, 0x20, 0xbd, 0x4a, 0xb8, 0xaa, + 0x27, 0x9a, 0x3b, 0x0d, 0x6f, 0x31, 0xb0, 0xe5, 0x6d, 0x48, 0x1e, 0x6b, 0xab, 0x6b, 0xe2, 0x91, + 0x4c, 0xbc, 0xb5, 0x65, 0x41, 0x2a, 0xaf, 0x15, 0x85, 0x8b, 0xed, 0xe1, 0xc8, 0x39, 0x31, 0x90, + 0xc5, 0xd9, 0x5a, 0x24, 0xfb, 0x93, 0x18, 0xb6, 0xc6, 0xd9, 0xd5, 0x48, 0xf6, 0xa7, 0x31, 0xec, + 0x2a, 0x67, 0xeb, 0x91, 0xec, 0xaf, 0xc6, 0xb0, 0x75, 0xce, 0x5e, 0x8f, 0x64, 0x7f, 0x2d, 0x86, + 0xbd, 0xce, 0xd9, 0xb5, 0x48, 0xf6, 0xd7, 0x63, 0xd8, 0x35, 0xce, 0xbe, 0x13, 0xc9, 0xfe, 0x46, + 0x0c, 0xfb, 0x0e, 0x67, 0xdf, 0x8d, 0x64, 0x7f, 0x33, 0x86, 0x7d, 0x97, 0xb3, 0xef, 0x45, 0xb2, + 0xbf, 0x15, 0xc3, 0xbe, 0xc7, 0xd8, 0x6b, 0xab, 0x91, 0xec, 0xcf, 0xa2, 0xd9, 0x6b, 0xab, 0x9c, + 0x1d, 0xad, 0xb5, 0x6f, 0xc7, 0xb0, 0xb9, 0xd6, 0xd6, 0xa2, 0xb5, 0xf6, 0x9d, 0x18, 0x36, 0xd7, + 0xda, 0x5a, 0xb4, 0xd6, 0xbe, 0x1b, 0xc3, 0xe6, 0x5a, 0x5b, 0x8b, 0xd6, 0xda, 0xf7, 0x62, 0xd8, + 0x5c, 0x6b, 0x6b, 0xd1, 0x5a, 0xfb, 0x7e, 0x0c, 0x9b, 0x6b, 0x6d, 0x2d, 0x5a, 0x6b, 0x3f, 0x88, + 0x61, 0x73, 0xad, 0xad, 0x45, 0x6b, 0xed, 0x8f, 0x62, 0xd8, 0x5c, 0x6b, 0x6b, 0xd1, 0x5a, 0xfb, + 0xe3, 0x18, 0x36, 0xd7, 0xda, 0x5a, 0xb4, 0xd6, 0xfe, 0x24, 0x86, 0xcd, 0xb5, 0xa6, 0x45, 0x6b, + 0xed, 0x4f, 0xa3, 0xd9, 0x1a, 0xd7, 0x9a, 0x16, 0xad, 0xb5, 0x3f, 0x8b, 0x61, 0x73, 0xad, 0x69, + 0xd1, 0x5a, 0xfb, 0xf3, 0x18, 0x36, 0xd7, 0x9a, 0x16, 0xad, 0xb5, 0x1f, 0xc6, 0xb0, 0xb9, 0xd6, + 0xb4, 0x68, 0xad, 0xfd, 0x45, 0x0c, 0x9b, 0x6b, 0x4d, 0x8b, 0xd6, 0xda, 0x5f, 0xc6, 0xb0, 0xb9, + 0xd6, 0xb4, 0x68, 0xad, 0xfd, 0x55, 0x0c, 0x9b, 0x6b, 0x4d, 0x8b, 0xd6, 0xda, 0x5f, 0xc7, 0xb0, + 0xb9, 0xd6, 0xb4, 0x68, 0xad, 0xfd, 0x4d, 0x0c, 0x9b, 0x6b, 0x4d, 0x8b, 0xd6, 0xda, 0xdf, 0xc6, + 0xb0, 0xb9, 0xd6, 0xaa, 0xd1, 0x5a, 0xfb, 0xbb, 0x68, 0x76, 0x95, 0x6b, 0xad, 0x1a, 0xad, 0xb5, + 0xbf, 0x8f, 0x61, 0x73, 0xad, 0x55, 0xa3, 0xb5, 0xf6, 0x0f, 0x31, 0x6c, 0xae, 0xb5, 0x6a, 0xb4, + 0xd6, 0xfe, 0x31, 0x86, 0xcd, 0xb5, 0x56, 0x8d, 0xd6, 0xda, 0x8f, 0x62, 0xd8, 0x5c, 0x6b, 0xd5, + 0x68, 0xad, 0xfd, 0x53, 0x0c, 0x9b, 0x6b, 0xad, 0x1a, 0xad, 0xb5, 0x7f, 0x8e, 0x61, 0x73, 0xad, + 0x55, 0xa3, 0xb5, 0xf6, 0x2f, 0x31, 0x6c, 0xae, 0xb5, 0x6a, 0xb4, 0xd6, 0xfe, 0x35, 0x86, 0xcd, + 0xb5, 0x56, 0x8d, 0xd6, 0xda, 0xbf, 0xc5, 0xb0, 0xb9, 0xd6, 0xf4, 0x68, 0xad, 0xfd, 0x7b, 0x34, + 0x5b, 0xe7, 0x5a, 0xd3, 0xa3, 0xb5, 0xf6, 0x1f, 0x31, 0x6c, 0xae, 0x35, 0x3d, 0x5a, 0x6b, 0xff, + 0x19, 0xc3, 0xe6, 0x5a, 0xd3, 0xa3, 0xb5, 0xf6, 0x5f, 0x31, 0x6c, 0xae, 0x35, 0x3d, 0x5a, 0x6b, + 0xff, 0x1d, 0xc3, 0xe6, 0x5a, 0xd3, 0xa3, 0xb5, 0xf6, 0x3f, 0x31, 0x6c, 0xae, 0x35, 0x3d, 0x5a, + 0x6b, 0x3f, 0x8e, 0x61, 0x73, 0xad, 0xe9, 0xd1, 0x5a, 0xfb, 0x49, 0x0c, 0x9b, 0x6b, 0x4d, 0x8f, + 0xd6, 0xda, 0xff, 0xc6, 0xb0, 0xb9, 0xd6, 0xf4, 0x68, 0xad, 0xfd, 0x5f, 0x0c, 0x9b, 0x6b, 0x6d, + 0x3d, 0x5a, 0x6b, 0xff, 0x1f, 0xcd, 0x5e, 0x5f, 0xfd, 0x69, 0x00, 0x00, 0x00, 0xff, 0xff, 0x81, + 0x23, 0xc6, 0xe6, 0xc6, 0x38, 0x00, 0x00, +} diff --git a/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/test.proto b/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/test.proto new file mode 100644 index 000000000..f60711369 --- /dev/null +++ b/vendor/github.com/matttproud/golang_protobuf_extensions/testdata/test.proto @@ -0,0 +1,540 @@ +// Go support for Protocol Buffers - Google's data interchange format +// +// Copyright 2010 The Go Authors. All rights reserved. +// https://github.com/golang/protobuf +// +// Redistribution and use in source and binary forms, with or without +// modification, are permitted provided that the following conditions are +// met: +// +// * Redistributions of source code must retain the above copyright +// notice, this list of conditions and the following disclaimer. +// * Redistributions in binary form must reproduce the above +// copyright notice, this list of conditions and the following disclaimer +// in the documentation and/or other materials provided with the +// distribution. +// * Neither the name of Google Inc. nor the names of its +// contributors may be used to endorse or promote products derived from +// this software without specific prior written permission. +// +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +// A feature-rich test file for the protocol compiler and libraries. + +syntax = "proto2"; + +package testdata; + +enum FOO { FOO1 = 1; }; + +message GoEnum { + required FOO foo = 1; +} + +message GoTestField { + required string Label = 1; + required string Type = 2; +} + +message GoTest { + // An enum, for completeness. + enum KIND { + VOID = 0; + + // Basic types + BOOL = 1; + BYTES = 2; + FINGERPRINT = 3; + FLOAT = 4; + INT = 5; + STRING = 6; + TIME = 7; + + // Groupings + TUPLE = 8; + ARRAY = 9; + MAP = 10; + + // Table types + TABLE = 11; + + // Functions + FUNCTION = 12; // last tag + }; + + // Some typical parameters + required KIND Kind = 1; + optional string Table = 2; + optional int32 Param = 3; + + // Required, repeated and optional foreign fields. + required GoTestField RequiredField = 4; + repeated GoTestField RepeatedField = 5; + optional GoTestField OptionalField = 6; + + // Required fields of all basic types + required bool F_Bool_required = 10; + required int32 F_Int32_required = 11; + required int64 F_Int64_required = 12; + required fixed32 F_Fixed32_required = 13; + required fixed64 F_Fixed64_required = 14; + required uint32 F_Uint32_required = 15; + required uint64 F_Uint64_required = 16; + required float F_Float_required = 17; + required double F_Double_required = 18; + required string F_String_required = 19; + required bytes F_Bytes_required = 101; + required sint32 F_Sint32_required = 102; + required sint64 F_Sint64_required = 103; + + // Repeated fields of all basic types + repeated bool F_Bool_repeated = 20; + repeated int32 F_Int32_repeated = 21; + repeated int64 F_Int64_repeated = 22; + repeated fixed32 F_Fixed32_repeated = 23; + repeated fixed64 F_Fixed64_repeated = 24; + repeated uint32 F_Uint32_repeated = 25; + repeated uint64 F_Uint64_repeated = 26; + repeated float F_Float_repeated = 27; + repeated double F_Double_repeated = 28; + repeated string F_String_repeated = 29; + repeated bytes F_Bytes_repeated = 201; + repeated sint32 F_Sint32_repeated = 202; + repeated sint64 F_Sint64_repeated = 203; + + // Optional fields of all basic types + optional bool F_Bool_optional = 30; + optional int32 F_Int32_optional = 31; + optional int64 F_Int64_optional = 32; + optional fixed32 F_Fixed32_optional = 33; + optional fixed64 F_Fixed64_optional = 34; + optional uint32 F_Uint32_optional = 35; + optional uint64 F_Uint64_optional = 36; + optional float F_Float_optional = 37; + optional double F_Double_optional = 38; + optional string F_String_optional = 39; + optional bytes F_Bytes_optional = 301; + optional sint32 F_Sint32_optional = 302; + optional sint64 F_Sint64_optional = 303; + + // Default-valued fields of all basic types + optional bool F_Bool_defaulted = 40 [default=true]; + optional int32 F_Int32_defaulted = 41 [default=32]; + optional int64 F_Int64_defaulted = 42 [default=64]; + optional fixed32 F_Fixed32_defaulted = 43 [default=320]; + optional fixed64 F_Fixed64_defaulted = 44 [default=640]; + optional uint32 F_Uint32_defaulted = 45 [default=3200]; + optional uint64 F_Uint64_defaulted = 46 [default=6400]; + optional float F_Float_defaulted = 47 [default=314159.]; + optional double F_Double_defaulted = 48 [default=271828.]; + optional string F_String_defaulted = 49 [default="hello, \"world!\"\n"]; + optional bytes F_Bytes_defaulted = 401 [default="Bignose"]; + optional sint32 F_Sint32_defaulted = 402 [default = -32]; + optional sint64 F_Sint64_defaulted = 403 [default = -64]; + + // Packed repeated fields (no string or bytes). + repeated bool F_Bool_repeated_packed = 50 [packed=true]; + repeated int32 F_Int32_repeated_packed = 51 [packed=true]; + repeated int64 F_Int64_repeated_packed = 52 [packed=true]; + repeated fixed32 F_Fixed32_repeated_packed = 53 [packed=true]; + repeated fixed64 F_Fixed64_repeated_packed = 54 [packed=true]; + repeated uint32 F_Uint32_repeated_packed = 55 [packed=true]; + repeated uint64 F_Uint64_repeated_packed = 56 [packed=true]; + repeated float F_Float_repeated_packed = 57 [packed=true]; + repeated double F_Double_repeated_packed = 58 [packed=true]; + repeated sint32 F_Sint32_repeated_packed = 502 [packed=true]; + repeated sint64 F_Sint64_repeated_packed = 503 [packed=true]; + + // Required, repeated, and optional groups. + required group RequiredGroup = 70 { + required string RequiredField = 71; + }; + + repeated group RepeatedGroup = 80 { + required string RequiredField = 81; + }; + + optional group OptionalGroup = 90 { + required string RequiredField = 91; + }; +} + +// For testing skipping of unrecognized fields. +// Numbers are all big, larger than tag numbers in GoTestField, +// the message used in the corresponding test. +message GoSkipTest { + required int32 skip_int32 = 11; + required fixed32 skip_fixed32 = 12; + required fixed64 skip_fixed64 = 13; + required string skip_string = 14; + required group SkipGroup = 15 { + required int32 group_int32 = 16; + required string group_string = 17; + } +} + +// For testing packed/non-packed decoder switching. +// A serialized instance of one should be deserializable as the other. +message NonPackedTest { + repeated int32 a = 1; +} + +message PackedTest { + repeated int32 b = 1 [packed=true]; +} + +message MaxTag { + // Maximum possible tag number. + optional string last_field = 536870911; +} + +message OldMessage { + message Nested { + optional string name = 1; + } + optional Nested nested = 1; + + optional int32 num = 2; +} + +// NewMessage is wire compatible with OldMessage; +// imagine it as a future version. +message NewMessage { + message Nested { + optional string name = 1; + optional string food_group = 2; + } + optional Nested nested = 1; + + // This is an int32 in OldMessage. + optional int64 num = 2; +} + +// Smaller tests for ASCII formatting. + +message InnerMessage { + required string host = 1; + optional int32 port = 2 [default=4000]; + optional bool connected = 3; +} + +message OtherMessage { + optional int64 key = 1; + optional bytes value = 2; + optional float weight = 3; + optional InnerMessage inner = 4; + + extensions 100 to max; +} + +message RequiredInnerMessage { + required InnerMessage leo_finally_won_an_oscar = 1; +} + +message MyMessage { + required int32 count = 1; + optional string name = 2; + optional string quote = 3; + repeated string pet = 4; + optional InnerMessage inner = 5; + repeated OtherMessage others = 6; + optional RequiredInnerMessage we_must_go_deeper = 13; + repeated InnerMessage rep_inner = 12; + + enum Color { + RED = 0; + GREEN = 1; + BLUE = 2; + }; + optional Color bikeshed = 7; + + optional group SomeGroup = 8 { + optional int32 group_field = 9; + } + + // This field becomes [][]byte in the generated code. + repeated bytes rep_bytes = 10; + + optional double bigfloat = 11; + + extensions 100 to max; +} + +message Ext { + extend MyMessage { + optional Ext more = 103; + optional string text = 104; + optional int32 number = 105; + } + + optional string data = 1; +} + +extend MyMessage { + repeated string greeting = 106; +} + +message ComplexExtension { + optional int32 first = 1; + optional int32 second = 2; + repeated int32 third = 3; +} + +extend OtherMessage { + optional ComplexExtension complex = 200; + repeated ComplexExtension r_complex = 201; +} + +message DefaultsMessage { + enum DefaultsEnum { + ZERO = 0; + ONE = 1; + TWO = 2; + }; + extensions 100 to max; +} + +extend DefaultsMessage { + optional double no_default_double = 101; + optional float no_default_float = 102; + optional int32 no_default_int32 = 103; + optional int64 no_default_int64 = 104; + optional uint32 no_default_uint32 = 105; + optional uint64 no_default_uint64 = 106; + optional sint32 no_default_sint32 = 107; + optional sint64 no_default_sint64 = 108; + optional fixed32 no_default_fixed32 = 109; + optional fixed64 no_default_fixed64 = 110; + optional sfixed32 no_default_sfixed32 = 111; + optional sfixed64 no_default_sfixed64 = 112; + optional bool no_default_bool = 113; + optional string no_default_string = 114; + optional bytes no_default_bytes = 115; + optional DefaultsMessage.DefaultsEnum no_default_enum = 116; + + optional double default_double = 201 [default = 3.1415]; + optional float default_float = 202 [default = 3.14]; + optional int32 default_int32 = 203 [default = 42]; + optional int64 default_int64 = 204 [default = 43]; + optional uint32 default_uint32 = 205 [default = 44]; + optional uint64 default_uint64 = 206 [default = 45]; + optional sint32 default_sint32 = 207 [default = 46]; + optional sint64 default_sint64 = 208 [default = 47]; + optional fixed32 default_fixed32 = 209 [default = 48]; + optional fixed64 default_fixed64 = 210 [default = 49]; + optional sfixed32 default_sfixed32 = 211 [default = 50]; + optional sfixed64 default_sfixed64 = 212 [default = 51]; + optional bool default_bool = 213 [default = true]; + optional string default_string = 214 [default = "Hello, string"]; + optional bytes default_bytes = 215 [default = "Hello, bytes"]; + optional DefaultsMessage.DefaultsEnum default_enum = 216 [default = ONE]; +} + +message MyMessageSet { + option message_set_wire_format = true; + extensions 100 to max; +} + +message Empty { +} + +extend MyMessageSet { + optional Empty x201 = 201; + optional Empty x202 = 202; + optional Empty x203 = 203; + optional Empty x204 = 204; + optional Empty x205 = 205; + optional Empty x206 = 206; + optional Empty x207 = 207; + optional Empty x208 = 208; + optional Empty x209 = 209; + optional Empty x210 = 210; + optional Empty x211 = 211; + optional Empty x212 = 212; + optional Empty x213 = 213; + optional Empty x214 = 214; + optional Empty x215 = 215; + optional Empty x216 = 216; + optional Empty x217 = 217; + optional Empty x218 = 218; + optional Empty x219 = 219; + optional Empty x220 = 220; + optional Empty x221 = 221; + optional Empty x222 = 222; + optional Empty x223 = 223; + optional Empty x224 = 224; + optional Empty x225 = 225; + optional Empty x226 = 226; + optional Empty x227 = 227; + optional Empty x228 = 228; + optional Empty x229 = 229; + optional Empty x230 = 230; + optional Empty x231 = 231; + optional Empty x232 = 232; + optional Empty x233 = 233; + optional Empty x234 = 234; + optional Empty x235 = 235; + optional Empty x236 = 236; + optional Empty x237 = 237; + optional Empty x238 = 238; + optional Empty x239 = 239; + optional Empty x240 = 240; + optional Empty x241 = 241; + optional Empty x242 = 242; + optional Empty x243 = 243; + optional Empty x244 = 244; + optional Empty x245 = 245; + optional Empty x246 = 246; + optional Empty x247 = 247; + optional Empty x248 = 248; + optional Empty x249 = 249; + optional Empty x250 = 250; +} + +message MessageList { + repeated group Message = 1 { + required string name = 2; + required int32 count = 3; + } +} + +message Strings { + optional string string_field = 1; + optional bytes bytes_field = 2; +} + +message Defaults { + enum Color { + RED = 0; + GREEN = 1; + BLUE = 2; + } + + // Default-valued fields of all basic types. + // Same as GoTest, but copied here to make testing easier. + optional bool F_Bool = 1 [default=true]; + optional int32 F_Int32 = 2 [default=32]; + optional int64 F_Int64 = 3 [default=64]; + optional fixed32 F_Fixed32 = 4 [default=320]; + optional fixed64 F_Fixed64 = 5 [default=640]; + optional uint32 F_Uint32 = 6 [default=3200]; + optional uint64 F_Uint64 = 7 [default=6400]; + optional float F_Float = 8 [default=314159.]; + optional double F_Double = 9 [default=271828.]; + optional string F_String = 10 [default="hello, \"world!\"\n"]; + optional bytes F_Bytes = 11 [default="Bignose"]; + optional sint32 F_Sint32 = 12 [default=-32]; + optional sint64 F_Sint64 = 13 [default=-64]; + optional Color F_Enum = 14 [default=GREEN]; + + // More fields with crazy defaults. + optional float F_Pinf = 15 [default=inf]; + optional float F_Ninf = 16 [default=-inf]; + optional float F_Nan = 17 [default=nan]; + + // Sub-message. + optional SubDefaults sub = 18; + + // Redundant but explicit defaults. + optional string str_zero = 19 [default=""]; +} + +message SubDefaults { + optional int64 n = 1 [default=7]; +} + +message RepeatedEnum { + enum Color { + RED = 1; + } + repeated Color color = 1; +} + +message MoreRepeated { + repeated bool bools = 1; + repeated bool bools_packed = 2 [packed=true]; + repeated int32 ints = 3; + repeated int32 ints_packed = 4 [packed=true]; + repeated int64 int64s_packed = 7 [packed=true]; + repeated string strings = 5; + repeated fixed32 fixeds = 6; +} + +// GroupOld and GroupNew have the same wire format. +// GroupNew has a new field inside a group. + +message GroupOld { + optional group G = 101 { + optional int32 x = 2; + } +} + +message GroupNew { + optional group G = 101 { + optional int32 x = 2; + optional int32 y = 3; + } +} + +message FloatingPoint { + required double f = 1; +} + +message MessageWithMap { + map name_mapping = 1; + map msg_mapping = 2; + map byte_mapping = 3; + map str_to_str = 4; +} + +message Oneof { + oneof union { + bool F_Bool = 1; + int32 F_Int32 = 2; + int64 F_Int64 = 3; + fixed32 F_Fixed32 = 4; + fixed64 F_Fixed64 = 5; + uint32 F_Uint32 = 6; + uint64 F_Uint64 = 7; + float F_Float = 8; + double F_Double = 9; + string F_String = 10; + bytes F_Bytes = 11; + sint32 F_Sint32 = 12; + sint64 F_Sint64 = 13; + MyMessage.Color F_Enum = 14; + GoTestField F_Message = 15; + group F_Group = 16 { + optional int32 x = 17; + } + int32 F_Largest_Tag = 536870911; + } + + oneof tormato { + int32 value = 100; + } +} + +message Communique { + optional bool make_me_cry = 1; + + // This is a oneof, called "union". + oneof union { + int32 number = 5; + string name = 6; + bytes data = 7; + double temp_c = 8; + MyMessage.Color col = 9; + Strings msg = 10; + } +} diff --git a/vendor/github.com/pierrec/lz4/.gitignore b/vendor/github.com/pierrec/lz4/.gitignore index c2bb6e4af..e48bab32a 100644 --- a/vendor/github.com/pierrec/lz4/.gitignore +++ b/vendor/github.com/pierrec/lz4/.gitignore @@ -29,3 +29,5 @@ Temporary Items .apdisk # End of https://www.gitignore.io/api/macos + +lz4c/lz4c diff --git a/vendor/github.com/pierrec/lz4/.travis.yml b/vendor/github.com/pierrec/lz4/.travis.yml index 78be21cc8..b2c806d57 100644 --- a/vendor/github.com/pierrec/lz4/.travis.yml +++ b/vendor/github.com/pierrec/lz4/.travis.yml @@ -1,8 +1,18 @@ language: go go: - - 1.x + - 1.8.x + - 1.9.x + - 1.10.x + - master + +matrix: + fast_finish: true + allow_failures: + - go: master + +sudo: false script: - go test -v -cpu=2 - - go test -v -cpu=2 -race \ No newline at end of file + - go test -v -cpu=2 -race diff --git a/vendor/github.com/pierrec/lz4/README.md b/vendor/github.com/pierrec/lz4/README.md index dd3c9d47e..f2f2764c2 100644 --- a/vendor/github.com/pierrec/lz4/README.md +++ b/vendor/github.com/pierrec/lz4/README.md @@ -1,8 +1,7 @@ [![godoc](https://godoc.org/github.com/pierrec/lz4?status.png)](https://godoc.org/github.com/pierrec/lz4) -[![Build Status](https://travis-ci.org/pierrec/lz4.svg?branch=master)](https://travis-ci.org/pierrec/lz4) # lz4 -LZ4 compression and decompression in pure Go +LZ4 compression and decompression in pure Go. ## Usage @@ -11,21 +10,13 @@ import "github.com/pierrec/lz4" ``` ## Description - Package lz4 implements reading and writing lz4 compressed data (a frame), -as specified in http://fastcompression.blogspot.fr/2013/04/lz4-streaming-format-final.html, -using an io.Reader (decompression) and io.Writer (compression). -It is designed to minimize memory usage while maximizing throughput by being able to -[de]compress data concurrently. - -The Reader and the Writer support concurrent processing provided the supplied buffers are -large enough (in multiples of BlockMaxSize) and there is no block dependency. -Reader.WriteTo and Writer.ReadFrom do leverage the concurrency transparently. -The runtime.GOMAXPROCS() value is used to apply concurrency or not. - +as specified in http://fastcompression.blogspot.fr/2013/04/lz4-streaming-format-final.html. Although the block level compression and decompression functions are exposed and are fully compatible with the lz4 block format definition, they are low level and should not be used directly. + For a complete description of an lz4 compressed block, see: http://fastcompression.blogspot.fr/2011/05/lz4-explained.html See https://github.com/Cyan4973/lz4 for the reference C implementation. + diff --git a/vendor/github.com/pierrec/lz4/bench_test.go b/vendor/github.com/pierrec/lz4/bench_test.go new file mode 100644 index 000000000..72f1eaaef --- /dev/null +++ b/vendor/github.com/pierrec/lz4/bench_test.go @@ -0,0 +1,119 @@ +package lz4_test + +import ( + "bytes" + "io" + "io/ioutil" + "testing" + + "github.com/pierrec/lz4" +) + +func BenchmarkCompress(b *testing.B) { + var hashTable [1 << 16]int + buf := make([]byte, len(pg1661)) + + b.ReportAllocs() + b.ResetTimer() + + for i := 0; i < b.N; i++ { + lz4.CompressBlock(pg1661, buf, hashTable[:]) + } +} + +func BenchmarkCompressHC(b *testing.B) { + buf := make([]byte, len(pg1661)) + + b.ReportAllocs() + b.ResetTimer() + + for i := 0; i < b.N; i++ { + lz4.CompressBlockHC(pg1661, buf, 16) + } +} + +func BenchmarkUncompress(b *testing.B) { + buf := make([]byte, len(pg1661)) + + b.ReportAllocs() + b.ResetTimer() + + for i := 0; i < b.N; i++ { + lz4.UncompressBlock(pg1661LZ4, buf) + } +} + +func mustLoadFile(f string) []byte { + b, err := ioutil.ReadFile(f) + if err != nil { + panic(err) + } + return b +} + +var ( + pg1661 = mustLoadFile("testdata/pg1661.txt") + digits = mustLoadFile("testdata/e.txt") + twain = mustLoadFile("testdata/Mark.Twain-Tom.Sawyer.txt") + random = mustLoadFile("testdata/random.data") + pg1661LZ4 = mustLoadFile("testdata/pg1661.txt.lz4") + digitsLZ4 = mustLoadFile("testdata/e.txt.lz4") + twainLZ4 = mustLoadFile("testdata/Mark.Twain-Tom.Sawyer.txt.lz4") + randomLZ4 = mustLoadFile("testdata/random.data.lz4") +) + +func benchmarkUncompress(b *testing.B, compressed []byte) { + r := bytes.NewReader(compressed) + zr := lz4.NewReader(r) + + // Determine the uncompressed size of testfile. + uncompressedSize, err := io.Copy(ioutil.Discard, zr) + if err != nil { + b.Fatal(err) + } + + b.SetBytes(uncompressedSize) + b.ReportAllocs() + b.ResetTimer() + + for i := 0; i < b.N; i++ { + r.Reset(compressed) + zr.Reset(r) + io.Copy(ioutil.Discard, zr) + } +} + +func BenchmarkUncompressPg1661(b *testing.B) { benchmarkUncompress(b, pg1661LZ4) } +func BenchmarkUncompressDigits(b *testing.B) { benchmarkUncompress(b, digitsLZ4) } +func BenchmarkUncompressTwain(b *testing.B) { benchmarkUncompress(b, twainLZ4) } +func BenchmarkUncompressRand(b *testing.B) { benchmarkUncompress(b, randomLZ4) } + +func benchmarkCompress(b *testing.B, uncompressed []byte) { + w := bytes.NewBuffer(nil) + zw := lz4.NewWriter(w) + r := bytes.NewReader(uncompressed) + + // Determine the compressed size of testfile. + compressedSize, err := io.Copy(zw, r) + if err != nil { + b.Fatal(err) + } + if err := zw.Close(); err != nil { + b.Fatal(err) + } + + b.SetBytes(compressedSize) + b.ReportAllocs() + b.ResetTimer() + + for i := 0; i < b.N; i++ { + r.Reset(uncompressed) + zw.Reset(w) + io.Copy(zw, r) + } +} + +func BenchmarkCompressPg1661(b *testing.B) { benchmarkCompress(b, pg1661) } +func BenchmarkCompressDigits(b *testing.B) { benchmarkCompress(b, digits) } +func BenchmarkCompressTwain(b *testing.B) { benchmarkCompress(b, twain) } +func BenchmarkCompressRand(b *testing.B) { benchmarkCompress(b, random) } diff --git a/vendor/github.com/pierrec/lz4/block.go b/vendor/github.com/pierrec/lz4/block.go index 44e3eaaac..ef24f17e5 100644 --- a/vendor/github.com/pierrec/lz4/block.go +++ b/vendor/github.com/pierrec/lz4/block.go @@ -5,188 +5,164 @@ import ( "errors" ) -// block represents a frame data block. -// Used when compressing or decompressing frame blocks concurrently. -type block struct { - compressed bool - zdata []byte // compressed data - data []byte // decompressed data - offset int // offset within the data as with block dependency the 64Kb window is prepended to it - checksum uint32 // compressed data checksum - err error // error while [de]compressing -} - var ( - // ErrInvalidSource is returned by UncompressBlock when a compressed block is corrupted. - ErrInvalidSource = errors.New("lz4: invalid source") - // ErrShortBuffer is returned by UncompressBlock, CompressBlock or CompressBlockHC when - // the supplied buffer for [de]compression is too small. - ErrShortBuffer = errors.New("lz4: short buffer") + // ErrInvalidSourceShortBuffer is returned by UncompressBlock or CompressBLock when a compressed + // block is corrupted or the destination buffer is not large enough for the uncompressed data. + ErrInvalidSourceShortBuffer = errors.New("lz4: invalid source or destination buffer too short") + // ErrInvalid is returned when reading an invalid LZ4 archive. + ErrInvalid = errors.New("lz4: bad magic number") ) +// blockHash hashes 4 bytes into a value < winSize. +func blockHash(x uint32) uint32 { + const hasher uint32 = 2654435761 // Knuth multiplicative hash. + return x * hasher >> hashShift +} + // CompressBlockBound returns the maximum size of a given buffer of size n, when not compressible. func CompressBlockBound(n int) int { return n + n/255 + 16 } -// UncompressBlock decompresses the source buffer into the destination one, -// starting at the di index and returning the decompressed size. +// UncompressBlock uncompresses the source buffer into the destination one, +// and returns the uncompressed size. // // The destination buffer must be sized appropriately. // // An error is returned if the source data is invalid or the destination buffer is too small. -func UncompressBlock(src, dst []byte, di int) (int, error) { - si, sn, di0 := 0, len(src), di +func UncompressBlock(src, dst []byte) (si int, err error) { + defer func() { + // It is now faster to let the runtime panic and recover on out of bound slice access + // than checking indices as we go along. + if recover() != nil { + err = ErrInvalidSourceShortBuffer + } + }() + sn := len(src) if sn == 0 { return 0, nil } + var di int for { - // literals and match lengths (token) - lLen := int(src[si] >> 4) - mLen := int(src[si] & 0xF) - if si++; si == sn { - return di, ErrInvalidSource - } + // Literals and match lengths (token). + b := int(src[si]) + si++ - // literals - if lLen > 0 { + // Literals. + if lLen := b >> 4; lLen > 0 { if lLen == 0xF { for src[si] == 0xFF { lLen += 0xFF - if si++; si == sn { - return di - di0, ErrInvalidSource - } + si++ } lLen += int(src[si]) - if si++; si == sn { - return di - di0, ErrInvalidSource - } + si++ } - if len(dst)-di < lLen || si+lLen > sn { - return di - di0, ErrShortBuffer - } - di += copy(dst[di:], src[si:si+lLen]) + i := si + si += lLen + di += copy(dst[di:], src[i:si]) - if si += lLen; si >= sn { - return di - di0, nil + if si >= sn { + return di, nil } } - if si += 2; si >= sn { - return di, ErrInvalidSource - } - offset := int(src[si-2]) | int(src[si-1])<<8 - if di-offset < 0 || offset == 0 { - return di - di0, ErrInvalidSource - } + si++ + _ = src[si] // Bound check elimination. + offset := int(src[si-1]) | int(src[si])<<8 + si++ - // match + // Match. + mLen := b & 0xF if mLen == 0xF { for src[si] == 0xFF { mLen += 0xFF - if si++; si == sn { - return di - di0, ErrInvalidSource - } + si++ } mLen += int(src[si]) - if si++; si == sn { - return di - di0, ErrInvalidSource - } - } - // minimum match length is 4 - mLen += 4 - if len(dst)-di <= mLen { - return di - di0, ErrShortBuffer + si++ } + mLen += minMatch - // copy the match (NB. match is at least 4 bytes long) - if mLen >= offset { + // Copy the match. + i := di - offset + if offset > 0 && mLen >= offset { + // Efficiently copy the match dst[di-offset:di] into the dst slice. bytesToCopy := offset * (mLen / offset) - // Efficiently copy the match dst[di-offset:di] into the slice - // dst[di:di+bytesToCopy] - expanded := dst[di-offset : di+bytesToCopy] - n := offset - for n <= bytesToCopy+offset { + expanded := dst[i:] + for n := offset; n <= bytesToCopy+offset; n *= 2 { copy(expanded[n:], expanded[:n]) - n *= 2 } di += bytesToCopy mLen -= bytesToCopy } - - di += copy(dst[di:], dst[di-offset:di-offset+mLen]) + di += copy(dst[di:], dst[i:i+mLen]) } } -// CompressBlock compresses the source buffer starting at soffet into the destination one. +// CompressBlock compresses the source buffer into the destination one. // This is the fast version of LZ4 compression and also the default one. +// The size of hashTable must be at least 64Kb. // // The size of the compressed data is returned. If it is 0 and no error, then the data is incompressible. // // An error is returned if the destination buffer is too small. -func CompressBlock(src, dst []byte, soffset int) (int, error) { +func CompressBlock(src, dst []byte, hashTable []int) (di int, err error) { + defer func() { + if recover() != nil { + err = ErrInvalidSourceShortBuffer + } + }() + sn, dn := len(src)-mfLimit, len(dst) - if sn <= 0 || dn == 0 || soffset >= sn { + if sn <= 0 || dn == 0 { return 0, nil } - var si, di int + var si int - // fast scan strategy: - // we only need a hash table to store the last sequences (4 bytes) - var hashTable [1 << hashLog]int - var hashShift = uint((minMatch * 8) - hashLog) + // Fast scan strategy: the hash table only stores the last 4 bytes sequences. + // const accInit = 1 << skipStrength - // Initialise the hash table with the first 64Kb of the input buffer - // (used when compressing dependent blocks) - for si < soffset { - h := binary.LittleEndian.Uint32(src[si:]) * hasher >> hashShift - si++ + anchor := si // Position of the current literals. + // acc := accInit // Variable step: improves performance on non-compressible data. + + for si < sn { + // Hash the next 4 bytes (sequence)... + match := binary.LittleEndian.Uint32(src[si:]) + h := blockHash(match) + + ref := hashTable[h] hashTable[h] = si - } - - anchor := si - fma := 1 << skipStrength - for si < sn-minMatch { - // hash the next 4 bytes (sequence)... - h := binary.LittleEndian.Uint32(src[si:]) * hasher >> hashShift - // -1 to separate existing entries from new ones - ref := hashTable[h] - 1 - // ...and store the position of the hash in the hash table (+1 to compensate the -1 upon saving) - hashTable[h] = si + 1 - // no need to check the last 3 bytes in the first literal 4 bytes as - // this guarantees that the next match, if any, is compressed with - // a lower size, since to have some compression we must have: - // ll+ml-overlap > 1 + (ll-15)/255 + (ml-4-15)/255 + 2 (uncompressed size>compressed size) - // => ll+ml>3+2*overlap => ll+ml>= 4+2*overlap - // and by definition we do have: - // ll >= 1, ml >= 4 - // => ll+ml >= 5 - // => so overlap must be 0 - - // the sequence is new, out of bound (64kb) or not valid: try next sequence - if ref < 0 || fma&(1<>winSizeLog > 0 || - src[ref] != src[si] || - src[ref+1] != src[si+1] || - src[ref+2] != src[si+2] || - src[ref+3] != src[si+3] { - // variable step: improves performance on non-compressible data - si += fma >> skipStrength - fma++ + if ref >= sn { // Invalid reference (dirty hashtable). + si++ continue } - // match found - fma = 1 << skipStrength - lLen := si - anchor offset := si - ref + if offset <= 0 || offset >= winSize || // Out of window. + match != binary.LittleEndian.Uint32(src[ref:]) { // Hash collision on different matches. + // si += acc >> skipStrength + // acc++ + si++ + continue + } - // encode match length part 1 + // Match found. + // acc = accInit + lLen := si - anchor // Literal length. + + // Encode match length part 1. si += minMatch - mLen := si // match length has minMatch already - for si <= sn && src[si] == src[si-offset] { + mLen := si // Match length has minMatch already. + // Find the longest match, first looking by batches of 8 bytes. + for si < sn && binary.LittleEndian.Uint64(src[si:]) == binary.LittleEndian.Uint64(src[si-offset:]) { + si += 8 + } + // Then byte by byte. + for si < sn && src[si] == src[si-offset] { si++ } + mLen = si - mLen if mLen < 0xF { dst[di] = byte(mLen) @@ -194,169 +170,160 @@ func CompressBlock(src, dst []byte, soffset int) (int, error) { dst[di] = 0xF } - // encode literals length + // Encode literals length. if lLen < 0xF { dst[di] |= byte(lLen << 4) } else { dst[di] |= 0xF0 - if di++; di == dn { - return di, ErrShortBuffer - } + di++ l := lLen - 0xF for ; l >= 0xFF; l -= 0xFF { dst[di] = 0xFF - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } dst[di] = byte(l) } - if di++; di == dn { - return di, ErrShortBuffer - } + di++ - // literals - if di+lLen >= dn { - return di, ErrShortBuffer - } - di += copy(dst[di:], src[anchor:anchor+lLen]) + // Literals. + copy(dst[di:], src[anchor:anchor+lLen]) + di += lLen + 2 anchor = si - // encode offset - if di += 2; di >= dn { - return di, ErrShortBuffer - } + // Encode offset. + _ = dst[di] // Bound check elimination. dst[di-2], dst[di-1] = byte(offset), byte(offset>>8) - // encode match length part 2 + // Encode match length part 2. if mLen >= 0xF { for mLen -= 0xF; mLen >= 0xFF; mLen -= 0xFF { dst[di] = 0xFF - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } dst[di] = byte(mLen) - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } } if anchor == 0 { - // incompressible + // Incompressible. return 0, nil } - // last literals + // Last literals. lLen := len(src) - anchor if lLen < 0xF { dst[di] = byte(lLen << 4) } else { dst[di] = 0xF0 - if di++; di == dn { - return di, ErrShortBuffer - } - lLen -= 0xF - for ; lLen >= 0xFF; lLen -= 0xFF { + di++ + for lLen -= 0xF; lLen >= 0xFF; lLen -= 0xFF { dst[di] = 0xFF - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } dst[di] = byte(lLen) } - if di++; di == dn { - return di, ErrShortBuffer - } + di++ - // write literals - src = src[anchor:] - switch n := di + len(src); { - case n > dn: - return di, ErrShortBuffer - case n >= sn: - // incompressible + // Write the last literals. + if di >= anchor { + // Incompressible. return 0, nil } - di += copy(dst[di:], src) + di += copy(dst[di:], src[anchor:]) return di, nil } -// CompressBlockHC compresses the source buffer starting at soffet into the destination one. +// CompressBlockHC compresses the source buffer src into the destination dst +// with max search depth (use 0 or negative value for no max). +// // CompressBlockHC compression ratio is better than CompressBlock but it is also slower. // // The size of the compressed data is returned. If it is 0 and no error, then the data is not compressible. // // An error is returned if the destination buffer is too small. -func CompressBlockHC(src, dst []byte, soffset int) (int, error) { +func CompressBlockHC(src, dst []byte, depth int) (di int, err error) { + defer func() { + if recover() != nil { + err = ErrInvalidSourceShortBuffer + } + }() + sn, dn := len(src)-mfLimit, len(dst) - if sn <= 0 || dn == 0 || soffset >= sn { + if sn <= 0 || dn == 0 { return 0, nil } - var si, di int + var si int - // Hash Chain strategy: - // we need a hash table and a chain table - // the chain table cannot contain more entries than the window size (64Kb entries) - var hashTable [1 << hashLog]int - var chainTable [winSize]int - var hashShift = uint((minMatch * 8) - hashLog) + // hashTable: stores the last position found for a given hash + // chaingTable: stores previous positions for a given hash + var hashTable, chainTable [winSize]int - // Initialise the hash table with the first 64Kb of the input buffer - // (used when compressing dependent blocks) - for si < soffset { - h := binary.LittleEndian.Uint32(src[si:]) * hasher >> hashShift - chainTable[si&winMask] = hashTable[h] - si++ - hashTable[h] = si + if depth <= 0 { + depth = winSize } anchor := si - for si < sn-minMatch { - // hash the next 4 bytes (sequence)... - h := binary.LittleEndian.Uint32(src[si:]) * hasher >> hashShift + for si < sn { + // Hash the next 4 bytes (sequence). + match := binary.LittleEndian.Uint32(src[si:]) + h := blockHash(match) - // follow the chain until out of window and give the longest match + // Follow the chain until out of window and give the longest match. mLen := 0 offset := 0 - for next := hashTable[h] - 1; next > 0 && next > si-winSize; next = chainTable[next&winMask] - 1 { - // the first (mLen==0) or next byte (mLen>=minMatch) at current match length must match to improve on the match length - if src[next+mLen] == src[si+mLen] { - for ml := 0; ; ml++ { - if src[next+ml] != src[si+ml] || si+ml > sn { - // found a longer match, keep its position and length - if mLen < ml && ml >= minMatch { - mLen = ml - offset = si - next - } - break - } - } + for next, try := hashTable[h], depth; try > 0 && next > 0 && si-next < winSize; next = chainTable[next&winMask] { + // The first (mLen==0) or next byte (mLen>=minMatch) at current match length + // must match to improve on the match length. + if src[next+mLen] != src[si+mLen] { + continue } + ml := 0 + // Compare the current position with a previous with the same hash. + for ml < sn-si && binary.LittleEndian.Uint64(src[next+ml:]) == binary.LittleEndian.Uint64(src[si+ml:]) { + ml += 8 + } + for ml < sn-si && src[next+ml] == src[si+ml] { + ml++ + } + if ml+1 < minMatch || ml <= mLen { + // Match too small (> hashShift + // Match found. + // Update hash/chain tables with overlapping bytes: + // si already hashed, add everything from si+1 up to the match length. + winStart := si + 1 + if ws := si + mLen - winSize; ws > winStart { + winStart = ws + } + for si, ml := winStart, si+mLen; si < ml; { + match >>= 8 + match |= uint32(src[si+3]) << 24 + h := blockHash(match) chainTable[si&winMask] = hashTable[h] - si++ hashTable[h] = si + si++ } lLen := si - anchor si += mLen - mLen -= minMatch // match length does not include minMatch + mLen -= minMatch // Match length does not include minMatch. if mLen < 0xF { dst[di] = byte(mLen) @@ -364,91 +331,67 @@ func CompressBlockHC(src, dst []byte, soffset int) (int, error) { dst[di] = 0xF } - // encode literals length + // Encode literals length. if lLen < 0xF { dst[di] |= byte(lLen << 4) } else { dst[di] |= 0xF0 - if di++; di == dn { - return di, ErrShortBuffer - } + di++ l := lLen - 0xF for ; l >= 0xFF; l -= 0xFF { dst[di] = 0xFF - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } dst[di] = byte(l) } - if di++; di == dn { - return di, ErrShortBuffer - } + di++ - // literals - if di+lLen >= dn { - return di, ErrShortBuffer - } - di += copy(dst[di:], src[anchor:anchor+lLen]) + // Literals. + copy(dst[di:], src[anchor:anchor+lLen]) + di += lLen anchor = si - // encode offset - if di += 2; di >= dn { - return di, ErrShortBuffer - } + // Encode offset. + di += 2 dst[di-2], dst[di-1] = byte(offset), byte(offset>>8) - // encode match length part 2 + // Encode match length part 2. if mLen >= 0xF { for mLen -= 0xF; mLen >= 0xFF; mLen -= 0xFF { dst[di] = 0xFF - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } dst[di] = byte(mLen) - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } } if anchor == 0 { - // incompressible + // Incompressible. return 0, nil } - // last literals + // Last literals. lLen := len(src) - anchor if lLen < 0xF { dst[di] = byte(lLen << 4) } else { dst[di] = 0xF0 - if di++; di == dn { - return di, ErrShortBuffer - } + di++ lLen -= 0xF for ; lLen >= 0xFF; lLen -= 0xFF { dst[di] = 0xFF - if di++; di == dn { - return di, ErrShortBuffer - } + di++ } dst[di] = byte(lLen) } - if di++; di == dn { - return di, ErrShortBuffer - } + di++ - // write literals - src = src[anchor:] - switch n := di + len(src); { - case n > dn: - return di, ErrShortBuffer - case n >= sn: - // incompressible + // Write the last literals. + if di >= anchor { + // Incompressible. return 0, nil } - di += copy(dst[di:], src) + di += copy(dst[di:], src[anchor:]) return di, nil } diff --git a/vendor/github.com/pierrec/lz4/block_test.go b/vendor/github.com/pierrec/lz4/block_test.go new file mode 100644 index 000000000..312160bb5 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/block_test.go @@ -0,0 +1,98 @@ +//+build go1.9 + +package lz4_test + +import ( + "fmt" + "io/ioutil" + "reflect" + "testing" + + "github.com/pierrec/lz4" +) + +type testcase struct { + file string + compressible bool + src []byte +} + +var rawFiles = []testcase{ + // {"testdata/207326ba-36f8-11e7-954a-aca46ba8ca73.png", true, nil}, + {"testdata/e.txt", true, nil}, + {"testdata/gettysburg.txt", true, nil}, + {"testdata/Mark.Twain-Tom.Sawyer.txt", true, nil}, + {"testdata/pg1661.txt", true, nil}, + {"testdata/pi.txt", true, nil}, + {"testdata/random.data", false, nil}, + {"testdata/repeat.txt", true, nil}, +} + +func TestCompressUncompressBlock(t *testing.T) { + type compressor func(s, d []byte) (int, error) + + run := func(tc testcase, compress compressor) int { + t.Helper() + src := tc.src + + // Compress the data. + zbuf := make([]byte, lz4.CompressBlockBound(len(src))) + n, err := compress(src, zbuf) + if err != nil { + t.Error(err) + return 0 + } + zbuf = zbuf[:n] + + // Make sure that it was actually compressed unless not compressible. + if !tc.compressible { + return 0 + } + + if n == 0 || n >= len(src) { + t.Errorf("data not compressed: %d/%d", n, len(src)) + return 0 + } + + // Uncompress the data. + buf := make([]byte, len(src)) + n, err = lz4.UncompressBlock(zbuf, buf) + if err != nil { + t.Fatal(err) + } + buf = buf[:n] + if !reflect.DeepEqual(src, buf) { + t.Error("uncompressed compressed data not matching initial input") + return 0 + } + + return len(zbuf) + } + + for _, tc := range rawFiles { + src, err := ioutil.ReadFile(tc.file) + if err != nil { + t.Fatal(err) + } + tc.src = src + + var n, nhc int + t.Run("", func(t *testing.T) { + tc := tc + t.Run(tc.file, func(t *testing.T) { + t.Parallel() + n = run(tc, func(src, dst []byte) (int, error) { + var ht [1 << 16]int + return lz4.CompressBlock(src, dst, ht[:]) + }) + }) + t.Run(fmt.Sprintf("%s HC", tc.file), func(t *testing.T) { + t.Parallel() + nhc = run(tc, func(src, dst []byte) (int, error) { + return lz4.CompressBlockHC(src, dst, -1) + }) + }) + }) + fmt.Printf("%-40s: %8d / %8d / %8d\n", tc.file, n, nhc, len(src)) + } +} diff --git a/vendor/github.com/pierrec/lz4/debug.go b/vendor/github.com/pierrec/lz4/debug.go new file mode 100644 index 000000000..5a5b1b950 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/debug.go @@ -0,0 +1,21 @@ +// +build lz4debug + +package lz4 + +import ( + "fmt" + "os" + "path/filepath" + "runtime" +) + +func debug(args ...interface{}) { + _, file, line, _ := runtime.Caller(1) + file = filepath.Base(file) + + f := fmt.Sprintf("LZ4: %s:%d %s", file, line, args[0]) + if f[len(f)-1] != '\n' { + f += "\n" + } + fmt.Fprintf(os.Stderr, f, args[1:]...) +} diff --git a/vendor/github.com/pierrec/lz4/debug_stub.go b/vendor/github.com/pierrec/lz4/debug_stub.go new file mode 100644 index 000000000..f57b90cf7 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/debug_stub.go @@ -0,0 +1,5 @@ +// +build !lz4debug + +package lz4 + +func debug(args ...interface{}) {} diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/0519d86e62cc577b98e9a4836b071ba1692c7674-30 b/vendor/github.com/pierrec/lz4/fuzz/corpus/0519d86e62cc577b98e9a4836b071ba1692c7674-30 new file mode 100644 index 000000000..b1287982e Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/0519d86e62cc577b98e9a4836b071ba1692c7674-30 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/0608f9eba5e6fd4d70241a81a6950ca51d78eb64-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/0608f9eba5e6fd4d70241a81a6950ca51d78eb64-33 new file mode 100644 index 000000000..67cb6d4a1 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/0608f9eba5e6fd4d70241a81a6950ca51d78eb64-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/17871030a73ac4d12ada652948135cb4639d679c-34 b/vendor/github.com/pierrec/lz4/fuzz/corpus/17871030a73ac4d12ada652948135cb4639d679c-34 new file mode 100644 index 000000000..68d5a7cb5 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/17871030a73ac4d12ada652948135cb4639d679c-34 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/1971e6ed6c6f6069fc2a9ed3038101e89bbcc381-26 b/vendor/github.com/pierrec/lz4/fuzz/corpus/1971e6ed6c6f6069fc2a9ed3038101e89bbcc381-26 new file mode 100644 index 000000000..073f10356 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/1971e6ed6c6f6069fc2a9ed3038101e89bbcc381-26 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/1a58f02dc83ac8315a85babdea6d757cbff2bb03-30 b/vendor/github.com/pierrec/lz4/fuzz/corpus/1a58f02dc83ac8315a85babdea6d757cbff2bb03-30 new file mode 100644 index 000000000..d589b761c Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/1a58f02dc83ac8315a85babdea6d757cbff2bb03-30 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/1a5a08b67764facaad851b9f1cbc5cfb31b7fb56-29 b/vendor/github.com/pierrec/lz4/fuzz/corpus/1a5a08b67764facaad851b9f1cbc5cfb31b7fb56-29 new file mode 100644 index 000000000..bd27105c3 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/1a5a08b67764facaad851b9f1cbc5cfb31b7fb56-29 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/1c944d5065b1a2b30e412604a14aa52565a5765b-35 b/vendor/github.com/pierrec/lz4/fuzz/corpus/1c944d5065b1a2b30e412604a14aa52565a5765b-35 new file mode 100644 index 000000000..da130d44b Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/1c944d5065b1a2b30e412604a14aa52565a5765b-35 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/2065ba3177c7dc5047742faa7158b3faeaac1f3c-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/2065ba3177c7dc5047742faa7158b3faeaac1f3c-32 new file mode 100644 index 000000000..8f24c4c02 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/2065ba3177c7dc5047742faa7158b3faeaac1f3c-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/21c8be1bb9eeea5b141500dee4987ab7fbd40d4a-23 b/vendor/github.com/pierrec/lz4/fuzz/corpus/21c8be1bb9eeea5b141500dee4987ab7fbd40d4a-23 new file mode 100644 index 000000000..222b09581 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/21c8be1bb9eeea5b141500dee4987ab7fbd40d4a-23 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/31c6c22708d346ed9e936fa7e77c8d9ab6da8d1e-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/31c6c22708d346ed9e936fa7e77c8d9ab6da8d1e-33 new file mode 100644 index 000000000..c1465e33d Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/31c6c22708d346ed9e936fa7e77c8d9ab6da8d1e-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/344d38ec2ec90cb617e809439938b4cbf3b11f02-10 b/vendor/github.com/pierrec/lz4/fuzz/corpus/344d38ec2ec90cb617e809439938b4cbf3b11f02-10 new file mode 100644 index 000000000..32922f358 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/344d38ec2ec90cb617e809439938b4cbf3b11f02-10 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/352631eab692c4a2c378b231fb3407ebcc0c3039-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/352631eab692c4a2c378b231fb3407ebcc0c3039-33 new file mode 100644 index 000000000..36b04114f Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/352631eab692c4a2c378b231fb3407ebcc0c3039-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/396146e06d3a4b2468d080f89ab5862348073424-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/396146e06d3a4b2468d080f89ab5862348073424-28 new file mode 100644 index 000000000..8df46b92e Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/396146e06d3a4b2468d080f89ab5862348073424-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/3b6fd6da48bb34284390a75e22940e7234dbbd28-34 b/vendor/github.com/pierrec/lz4/fuzz/corpus/3b6fd6da48bb34284390a75e22940e7234dbbd28-34 new file mode 100644 index 000000000..de4e0a82c Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/3b6fd6da48bb34284390a75e22940e7234dbbd28-34 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/4114fd99aaa4dc95365dc4bbcb3c9a8a03434a5a-29 b/vendor/github.com/pierrec/lz4/fuzz/corpus/4114fd99aaa4dc95365dc4bbcb3c9a8a03434a5a-29 new file mode 100644 index 000000000..4c8ea2b5f Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/4114fd99aaa4dc95365dc4bbcb3c9a8a03434a5a-29 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/4131f155339a3476898088b065b8588a2b28278e-26 b/vendor/github.com/pierrec/lz4/fuzz/corpus/4131f155339a3476898088b065b8588a2b28278e-26 new file mode 100644 index 000000000..a97a916d9 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/4131f155339a3476898088b065b8588a2b28278e-26 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/42544ff3318fe86dd466e9a05068e752a1057fcc-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/42544ff3318fe86dd466e9a05068e752a1057fcc-32 new file mode 100644 index 000000000..a71836ea2 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/42544ff3318fe86dd466e9a05068e752a1057fcc-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/4a14a3883f5c8819405319e8fb96234f5746a0ef-22 b/vendor/github.com/pierrec/lz4/fuzz/corpus/4a14a3883f5c8819405319e8fb96234f5746a0ef-22 new file mode 100644 index 000000000..398d0cfe1 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/4a14a3883f5c8819405319e8fb96234f5746a0ef-22 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/51075c34f23d161fb97edcf6f1b73ee6005009a0-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/51075c34f23d161fb97edcf6f1b73ee6005009a0-28 new file mode 100644 index 000000000..4b93cd1b8 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/51075c34f23d161fb97edcf6f1b73ee6005009a0-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/517d39f406222f0a0121b7a1961953204674c251-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/517d39f406222f0a0121b7a1961953204674c251-33 new file mode 100644 index 000000000..4b42d4b50 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/517d39f406222f0a0121b7a1961953204674c251-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/5e19e298d051aac48b7683dc24577b46268b630c-35 b/vendor/github.com/pierrec/lz4/fuzz/corpus/5e19e298d051aac48b7683dc24577b46268b630c-35 new file mode 100644 index 000000000..a39d26172 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/5e19e298d051aac48b7683dc24577b46268b630c-35 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/5f946423d1138924933334c6e5d3eb13e1020e9c-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/5f946423d1138924933334c6e5d3eb13e1020e9c-33 new file mode 100644 index 000000000..fe8e779fa Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/5f946423d1138924933334c6e5d3eb13e1020e9c-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/633df0cd78621cd45067a58d23c6ed67bb1b60cb-31 b/vendor/github.com/pierrec/lz4/fuzz/corpus/633df0cd78621cd45067a58d23c6ed67bb1b60cb-31 new file mode 100644 index 000000000..57f89bcb4 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/633df0cd78621cd45067a58d23c6ed67bb1b60cb-31 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/66c34847568ac9cb3ccbb8be26f494988a3e0628-7 b/vendor/github.com/pierrec/lz4/fuzz/corpus/66c34847568ac9cb3ccbb8be26f494988a3e0628-7 new file mode 100644 index 000000000..11972bf19 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/66c34847568ac9cb3ccbb8be26f494988a3e0628-7 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/67534dbd68040fb9a8867e6af384d33ea323758b-29 b/vendor/github.com/pierrec/lz4/fuzz/corpus/67534dbd68040fb9a8867e6af384d33ea323758b-29 new file mode 100644 index 000000000..1755a5123 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/67534dbd68040fb9a8867e6af384d33ea323758b-29 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/68612136c2017f9caf87122155f82a25f57c2d2a-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/68612136c2017f9caf87122155f82a25f57c2d2a-32 new file mode 100644 index 000000000..9c0fac860 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/68612136c2017f9caf87122155f82a25f57c2d2a-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/6981397d97c481e39d563d43916377fb3c74c60e-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/6981397d97c481e39d563d43916377fb3c74c60e-28 new file mode 100644 index 000000000..cfb90f6fe Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/6981397d97c481e39d563d43916377fb3c74c60e-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/69c2accb74456005e2a9bbef15ccad3d076f2124-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/69c2accb74456005e2a9bbef15ccad3d076f2124-28 new file mode 100644 index 000000000..1e3d09709 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/69c2accb74456005e2a9bbef15ccad3d076f2124-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/69fcd886042d5c3ebe89afd561782ac25619e35b-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/69fcd886042d5c3ebe89afd561782ac25619e35b-27 new file mode 100644 index 000000000..10a5a7102 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/69fcd886042d5c3ebe89afd561782ac25619e35b-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/6b72fdd9989971ecc3b50c34ee420f56a03e1026-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/6b72fdd9989971ecc3b50c34ee420f56a03e1026-27 new file mode 100644 index 000000000..3eec51590 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/6b72fdd9989971ecc3b50c34ee420f56a03e1026-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/72c738d7492d3055c6fe7391198422984b9e4702-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/72c738d7492d3055c6fe7391198422984b9e4702-32 new file mode 100644 index 000000000..00aa56e44 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/72c738d7492d3055c6fe7391198422984b9e4702-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/764571571e4d46f4397ed534d0160718ce578da4-26 b/vendor/github.com/pierrec/lz4/fuzz/corpus/764571571e4d46f4397ed534d0160718ce578da4-26 new file mode 100644 index 000000000..08edac5ff Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/764571571e4d46f4397ed534d0160718ce578da4-26 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/78e59daada9b9be755d1b508dd392fa9fc6fa9c2-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/78e59daada9b9be755d1b508dd392fa9fc6fa9c2-27 new file mode 100644 index 000000000..21ad88cce Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/78e59daada9b9be755d1b508dd392fa9fc6fa9c2-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/78ef686662a059f053f80c1c63c2921deff073fb-31 b/vendor/github.com/pierrec/lz4/fuzz/corpus/78ef686662a059f053f80c1c63c2921deff073fb-31 new file mode 100644 index 000000000..a8400b752 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/78ef686662a059f053f80c1c63c2921deff073fb-31 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/7a0fc8dacceae32a59589711dce63800085c22c7-23 b/vendor/github.com/pierrec/lz4/fuzz/corpus/7a0fc8dacceae32a59589711dce63800085c22c7-23 new file mode 100644 index 000000000..adcd6885a Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/7a0fc8dacceae32a59589711dce63800085c22c7-23 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/7b919213d591e6ce4355c635dc1ecc0d8e78befe-30 b/vendor/github.com/pierrec/lz4/fuzz/corpus/7b919213d591e6ce4355c635dc1ecc0d8e78befe-30 new file mode 100644 index 000000000..2c7bc2760 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/7b919213d591e6ce4355c635dc1ecc0d8e78befe-30 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/7f8c3b163798c8d5e1b65e03f411b56b6c9384bb-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/7f8c3b163798c8d5e1b65e03f411b56b6c9384bb-28 new file mode 100644 index 000000000..d4eb29f09 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/7f8c3b163798c8d5e1b65e03f411b56b6c9384bb-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/82a499521f34b6a9aff3b71d5f8bfd358933a4b2-36 b/vendor/github.com/pierrec/lz4/fuzz/corpus/82a499521f34b6a9aff3b71d5f8bfd358933a4b2-36 new file mode 100644 index 000000000..945f5fc94 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/82a499521f34b6a9aff3b71d5f8bfd358933a4b2-36 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/82c627991d65c5c4e88c9ccac39be082cca40765-24 b/vendor/github.com/pierrec/lz4/fuzz/corpus/82c627991d65c5c4e88c9ccac39be082cca40765-24 new file mode 100644 index 000000000..bfa3172a7 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/82c627991d65c5c4e88c9ccac39be082cca40765-24 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/8435aa58e67c4de798375b44c11bffa5b680f615-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/8435aa58e67c4de798375b44c11bffa5b680f615-32 new file mode 100644 index 000000000..d18b93e2b Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/8435aa58e67c4de798375b44c11bffa5b680f615-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/87caf7737ebb025ec2d908224818ceb2bc76b658-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/87caf7737ebb025ec2d908224818ceb2bc76b658-28 new file mode 100644 index 000000000..10431de1f Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/87caf7737ebb025ec2d908224818ceb2bc76b658-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/8d70b7de160bbef22ab46f798d687a69dbda772c-5 b/vendor/github.com/pierrec/lz4/fuzz/corpus/8d70b7de160bbef22ab46f798d687a69dbda772c-5 new file mode 100644 index 000000000..cfba1be2e --- /dev/null +++ b/vendor/github.com/pierrec/lz4/fuzz/corpus/8d70b7de160bbef22ab46f798d687a69dbda772c-5 @@ -0,0 +1 @@ +"MbT \ No newline at end of file diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/8f4788d30edd22ebcfef0e52bbf9e8c3d1e8d7e9-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/8f4788d30edd22ebcfef0e52bbf9e8c3d1e8d7e9-27 new file mode 100644 index 000000000..4cc2e6ed2 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/8f4788d30edd22ebcfef0e52bbf9e8c3d1e8d7e9-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/92d41e4fca52311e848fac274144b6293d9260f7-34 b/vendor/github.com/pierrec/lz4/fuzz/corpus/92d41e4fca52311e848fac274144b6293d9260f7-34 new file mode 100644 index 000000000..d2c123a2e Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/92d41e4fca52311e848fac274144b6293d9260f7-34 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/955c823909722e2693dd7cea3eadc17833dddf86-24 b/vendor/github.com/pierrec/lz4/fuzz/corpus/955c823909722e2693dd7cea3eadc17833dddf86-24 new file mode 100644 index 000000000..3dae9dc25 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/955c823909722e2693dd7cea3eadc17833dddf86-24 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/95ca8da5556065f33b46c2c8186c2f1cebb1b5da-29 b/vendor/github.com/pierrec/lz4/fuzz/corpus/95ca8da5556065f33b46c2c8186c2f1cebb1b5da-29 new file mode 100644 index 000000000..c95c989da Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/95ca8da5556065f33b46c2c8186c2f1cebb1b5da-29 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/995d50f1cb750cbf038246d6cb0cf8db11d7e60e-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/995d50f1cb750cbf038246d6cb0cf8db11d7e60e-33 new file mode 100644 index 000000000..9d5311e86 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/995d50f1cb750cbf038246d6cb0cf8db11d7e60e-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/9a5ab6c72a445b3b27129004d2a1a417cd4d8440-26 b/vendor/github.com/pierrec/lz4/fuzz/corpus/9a5ab6c72a445b3b27129004d2a1a417cd4d8440-26 new file mode 100644 index 000000000..f0c11e54d Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/9a5ab6c72a445b3b27129004d2a1a417cd4d8440-26 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/9e160ae007fc11092a3fd877ebe706c4d841db49-19 b/vendor/github.com/pierrec/lz4/fuzz/corpus/9e160ae007fc11092a3fd877ebe706c4d841db49-19 new file mode 100644 index 000000000..5c90b1562 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/9e160ae007fc11092a3fd877ebe706c4d841db49-19 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/a97d9bf241e8ec73f99205b32c24fcd64194f0b9-8 b/vendor/github.com/pierrec/lz4/fuzz/corpus/a97d9bf241e8ec73f99205b32c24fcd64194f0b9-8 new file mode 100644 index 000000000..cd8d11eea Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/a97d9bf241e8ec73f99205b32c24fcd64194f0b9-8 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/b53101ec4348e9c329c13e22790ffde246743030-35 b/vendor/github.com/pierrec/lz4/fuzz/corpus/b53101ec4348e9c329c13e22790ffde246743030-35 new file mode 100644 index 000000000..f0910f13e Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/b53101ec4348e9c329c13e22790ffde246743030-35 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/b58429fd1107617191026029cf327b2ebed963bb-18 b/vendor/github.com/pierrec/lz4/fuzz/corpus/b58429fd1107617191026029cf327b2ebed963bb-18 new file mode 100644 index 000000000..c26170392 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/b58429fd1107617191026029cf327b2ebed963bb-18 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/b92c70d3f12e67c69ba5db9ad491b7a4e075ece8-7 b/vendor/github.com/pierrec/lz4/fuzz/corpus/b92c70d3f12e67c69ba5db9ad491b7a4e075ece8-7 new file mode 100644 index 000000000..b113b1c5f Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/b92c70d3f12e67c69ba5db9ad491b7a4e075ece8-7 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/bc3ac4aae07cba8d7f657a8739d1774e44bde613-31 b/vendor/github.com/pierrec/lz4/fuzz/corpus/bc3ac4aae07cba8d7f657a8739d1774e44bde613-31 new file mode 100644 index 000000000..b52bd3110 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/bc3ac4aae07cba8d7f657a8739d1774e44bde613-31 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/bdc123d9da19a7ae0ff87ca0741002fbd8bb2cca-34 b/vendor/github.com/pierrec/lz4/fuzz/corpus/bdc123d9da19a7ae0ff87ca0741002fbd8bb2cca-34 new file mode 100644 index 000000000..b6d6b05ac Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/bdc123d9da19a7ae0ff87ca0741002fbd8bb2cca-34 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/c1972d0c898848e6188b69bcdbb7d14fcc780ee5-26 b/vendor/github.com/pierrec/lz4/fuzz/corpus/c1972d0c898848e6188b69bcdbb7d14fcc780ee5-26 new file mode 100644 index 000000000..79651e1cf Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/c1972d0c898848e6188b69bcdbb7d14fcc780ee5-26 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/c42ae63ab9584753959f4692cef9fd8513b54691-30 b/vendor/github.com/pierrec/lz4/fuzz/corpus/c42ae63ab9584753959f4692cef9fd8513b54691-30 new file mode 100644 index 000000000..4a9048820 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/c42ae63ab9584753959f4692cef9fd8513b54691-30 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/c8b01a7ea9c1b84e4ee5eb68121c64f183e7ea10-9 b/vendor/github.com/pierrec/lz4/fuzz/corpus/c8b01a7ea9c1b84e4ee5eb68121c64f183e7ea10-9 new file mode 100644 index 000000000..56aee0515 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/c8b01a7ea9c1b84e4ee5eb68121c64f183e7ea10-9 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/cb1314cc880a1a389cedf5c16cc4b8ad505b4506-23 b/vendor/github.com/pierrec/lz4/fuzz/corpus/cb1314cc880a1a389cedf5c16cc4b8ad505b4506-23 new file mode 100644 index 000000000..29567cd5e Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/cb1314cc880a1a389cedf5c16cc4b8ad505b4506-23 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/ceb22e7f581d85ed876e3d61da7df65da8954bf2-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/ceb22e7f581d85ed876e3d61da7df65da8954bf2-32 new file mode 100644 index 000000000..3f4ae7812 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/ceb22e7f581d85ed876e3d61da7df65da8954bf2-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/d8873ec9a0344ea23f70d1ffd78c2fd0435b9885-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/d8873ec9a0344ea23f70d1ffd78c2fd0435b9885-27 new file mode 100644 index 000000000..c8252dc0d Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/d8873ec9a0344ea23f70d1ffd78c2fd0435b9885-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/da3418e70658be491531ef6524f6ef7984ff9e96-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/da3418e70658be491531ef6524f6ef7984ff9e96-27 new file mode 100644 index 000000000..676412a82 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/da3418e70658be491531ef6524f6ef7984ff9e96-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/daffc68f738bd5945de9c7babd4e01cc4438fae8-31 b/vendor/github.com/pierrec/lz4/fuzz/corpus/daffc68f738bd5945de9c7babd4e01cc4438fae8-31 new file mode 100644 index 000000000..7f067483b Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/daffc68f738bd5945de9c7babd4e01cc4438fae8-31 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/df5bd5044e9b74c648b5f5fcb4dbdf953175f9f9-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/df5bd5044e9b74c648b5f5fcb4dbdf953175f9f9-27 new file mode 100644 index 000000000..eb27e25ee Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/df5bd5044e9b74c648b5f5fcb4dbdf953175f9f9-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/e22a5ac115e8bfd3468c9e6ad73ea11b8743798a-30 b/vendor/github.com/pierrec/lz4/fuzz/corpus/e22a5ac115e8bfd3468c9e6ad73ea11b8743798a-30 new file mode 100644 index 000000000..fad764ac0 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/e22a5ac115e8bfd3468c9e6ad73ea11b8743798a-30 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/e544de8de59a005934dd4b7fd465c5bb0046482e-26 b/vendor/github.com/pierrec/lz4/fuzz/corpus/e544de8de59a005934dd4b7fd465c5bb0046482e-26 new file mode 100644 index 000000000..2f3eeb066 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/e544de8de59a005934dd4b7fd465c5bb0046482e-26 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/e7f55f4c85203100c3cd819cdc87abb0e9e86374-32 b/vendor/github.com/pierrec/lz4/fuzz/corpus/e7f55f4c85203100c3cd819cdc87abb0e9e86374-32 new file mode 100644 index 000000000..92934ff4f Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/e7f55f4c85203100c3cd819cdc87abb0e9e86374-32 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/ea83e3b78398628e8a85e2e618fa956c0ffbd733-35 b/vendor/github.com/pierrec/lz4/fuzz/corpus/ea83e3b78398628e8a85e2e618fa956c0ffbd733-35 new file mode 100644 index 000000000..0918d6f8a Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/ea83e3b78398628e8a85e2e618fa956c0ffbd733-35 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/eb967d9cb0407c2328bbdbf98b5602274452d900-23 b/vendor/github.com/pierrec/lz4/fuzz/corpus/eb967d9cb0407c2328bbdbf98b5602274452d900-23 new file mode 100644 index 000000000..2eb1c2c61 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/eb967d9cb0407c2328bbdbf98b5602274452d900-23 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/ec93fb54ce508e132c89b6637913f84c3c78bafd-29 b/vendor/github.com/pierrec/lz4/fuzz/corpus/ec93fb54ce508e132c89b6637913f84c3c78bafd-29 new file mode 100644 index 000000000..74632e7c7 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/ec93fb54ce508e132c89b6637913f84c3c78bafd-29 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/efd3db86b12d209db7f0b24281a2cccebff526cd-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/efd3db86b12d209db7f0b24281a2cccebff526cd-33 new file mode 100644 index 000000000..11350777a Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/efd3db86b12d209db7f0b24281a2cccebff526cd-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/f31dcf6e3044e050f2396b601ebe420e89749c07-27 b/vendor/github.com/pierrec/lz4/fuzz/corpus/f31dcf6e3044e050f2396b601ebe420e89749c07-27 new file mode 100644 index 000000000..2fe704fcf Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/f31dcf6e3044e050f2396b601ebe420e89749c07-27 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/f3f49f3016c41052be090544cf110c322bc7ef63-24 b/vendor/github.com/pierrec/lz4/fuzz/corpus/f3f49f3016c41052be090544cf110c322bc7ef63-24 new file mode 100644 index 000000000..283f07800 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/f3f49f3016c41052be090544cf110c322bc7ef63-24 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/f4003ca01b90a4ee1be5701a5dd7d5f04e00c8f8-28 b/vendor/github.com/pierrec/lz4/fuzz/corpus/f4003ca01b90a4ee1be5701a5dd7d5f04e00c8f8-28 new file mode 100644 index 000000000..7dbc6dfb6 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/f4003ca01b90a4ee1be5701a5dd7d5f04e00c8f8-28 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/f5ecb47dfd92bb0564588beefd03ffcb0bbdae54-29 b/vendor/github.com/pierrec/lz4/fuzz/corpus/f5ecb47dfd92bb0564588beefd03ffcb0bbdae54-29 new file mode 100644 index 000000000..ec0539ee8 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/f5ecb47dfd92bb0564588beefd03ffcb0bbdae54-29 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/corpus/f9bcd3660c355799a865fedd15cb27a18591f244-33 b/vendor/github.com/pierrec/lz4/fuzz/corpus/f9bcd3660c355799a865fedd15cb27a18591f244-33 new file mode 100644 index 000000000..5af2a422b Binary files /dev/null and b/vendor/github.com/pierrec/lz4/fuzz/corpus/f9bcd3660c355799a865fedd15cb27a18591f244-33 differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968 b/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968 deleted file mode 100644 index 2b0938aa3..000000000 Binary files a/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968 and /dev/null differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968.output b/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968.output deleted file mode 100644 index 62f54ca14..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968.output +++ /dev/null @@ -1,51 +0,0 @@ -program hanged (timeout 10 seconds) - -SIGABRT: abort -PC=0x5e9e2 m=0 - -goroutine 1 [running]: -github.com/pierrec/lz4.UncompressBlock(0x820282830, 0x6, 0x6, 0x82032e000, 0x10000, 0x10000, 0x0, 0x6, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/block.go:104 +0xec2 fp=0x8202b59d8 sp=0x8202b5900 -github.com/pierrec/lz4.(*Reader).decompressBlock(0x8203085a0, 0x820290240, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/reader.go:271 +0x189 fp=0x8202b5a48 sp=0x8202b59d8 -github.com/pierrec/lz4.(*Reader).Read(0x8203085a0, 0x82030b400, 0x200, 0x200, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/reader.go:188 +0x1156 fp=0x8202b5c38 sp=0x8202b5a48 -bytes.(*Buffer).ReadFrom(0x8202b5d68, 0x882042d260, 0x8203085a0, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/bytes/buffer.go:173 +0x3db fp=0x8202b5ce8 sp=0x8202b5c38 -io/ioutil.readAll(0x882042d260, 0x8203085a0, 0x200, 0x0, 0x0, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/io/ioutil/ioutil.go:33 +0x1ed fp=0x8202b5de0 sp=0x8202b5ce8 -io/ioutil.ReadAll(0x882042d260, 0x8203085a0, 0x0, 0x0, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/io/ioutil/ioutil.go:42 +0x80 fp=0x8202b5e28 sp=0x8202b5de0 -github.com/pierrec/lz4/fuzz.Fuzz(0x8820479000, 0x1b, 0x200000, 0x3) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/fuzz/lz4.go:11 +0x15f fp=0x8202b5ea0 sp=0x8202b5e28 -github.com/dvyukov/go-fuzz/go-fuzz-dep.Main(0x1a7f18) - /Users/pierrecurto/sandbox/src/github.com/dvyukov/go-fuzz/go-fuzz-dep/main.go:47 +0x14c fp=0x8202b5f40 sp=0x8202b5ea0 -main.main() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/go-fuzz-main/main.go:10 +0x23 fp=0x8202b5f50 sp=0x8202b5f40 -runtime.main() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/runtime/proc.go:111 +0x2b0 fp=0x8202b5fa0 sp=0x8202b5f50 -runtime.goexit() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/runtime/asm_amd64.s:1696 +0x1 fp=0x8202b5fa8 sp=0x8202b5fa0 - -rax 0x0 -rbx 0x0 -rcx 0x0 -rdx 0x82032e000 -rdi 0x82032e000 -rsi 0x82032e000 -rbp 0x0 -rsp 0x8202b5900 -r8 0x10000 -r9 0x82032e000 -r10 0x10000 -r11 0x82032e000 -r12 0x10000 -r13 0x10000 -r14 0x1 -r15 0x8 -rip 0x5e9e2 -rflags 0x206 -cs 0x2b -fs 0x0 -gs 0x0 -exit status 2 \ No newline at end of file diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968.quoted b/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968.quoted deleted file mode 100644 index 4b42bc150..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/crashers/0b8f7fcd1f53d5bd839e5728ba92db050f5e0968.quoted +++ /dev/null @@ -1,2 +0,0 @@ - "\x04\"M\x18M@\x00\x00B*M\f\x00'\x01\x06\x00\x00\x00\x00" + - "\x00\x00\x16\xe3\x00\x10\x1e" diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b b/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b deleted file mode 100644 index 501c796d7..000000000 Binary files a/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b and /dev/null differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b.output b/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b.output deleted file mode 100644 index ab8d5754a..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b.output +++ /dev/null @@ -1,54 +0,0 @@ -program hanged (timeout 10 seconds) - -SIGABRT: abort -PC=0x5669b m=0 - -goroutine 0 [idle]: -runtime.mach_semaphore_wait(0x703, 0x7fff5fbff9a8, 0x8202fe401, 0x0, 0x1, 0x238cc0, 0x49b09, 0xffffffffffffffff, 0x600, 0x7fff5fbff90c, ...) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/sys_darwin_amd64.s:407 +0xb -runtime.semasleep1(0xffffffffffffffff, 0x600) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/os1_darwin.go:385 +0xe5 -runtime.semasleep.func1() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/os1_darwin.go:401 +0x29 -runtime.systemstack(0x7fff5fbff910) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/asm_amd64.s:278 +0xab -runtime.semasleep(0xffffffffffffffff, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/os1_darwin.go:402 +0x36 -runtime.notesleep(0x239110) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/lock_sema.go:169 +0x100 -runtime.stopm() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/proc1.go:1128 +0x112 -runtime.findrunnable(0x8202a4000, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/proc1.go:1530 +0x69e -runtime.schedule() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/proc1.go:1639 +0x267 -runtime.park_m(0x82028af00) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/proc1.go:1698 +0x18b -runtime.mcall(0x7fff5fbffa90) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build890014890/src/runtime/asm_amd64.s:204 +0x5b - -goroutine 1 [running]: - goroutine running on other thread; stack unavailable - -rax 0xe -rbx 0x703 -rcx 0x7fff5fbff898 -rdx 0x7fff5fbff910 -rdi 0x703 -rsi 0x238cc0 -rbp 0x239000 -rsp 0x7fff5fbff898 -r8 0x239000 -r9 0x8820290330 -r10 0x25ee08c1e -r11 0x286 -r12 0x0 -r13 0x6d9e8a1cfd40 -r14 0x13fde99489843000 -r15 0x238960 -rip 0x5669b -rflags 0x286 -cs 0x7 -fs 0x0 -gs 0x0 -exit status 2 \ No newline at end of file diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b.quoted b/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b.quoted deleted file mode 100644 index f79103ecc..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/crashers/169b44c5a64fec4d8e969d25d3e4764c9c3b604b.quoted +++ /dev/null @@ -1,4 +0,0 @@ - "\x04\"M\x18na\x84Ƚ\xbf\xef]\x00\x01\x00\x02\x00\x00\x00\x18" + - "N\x02funcn\x02\x00\x00\x00\x18n\x02\x00\x00\x00\x18\x00\x02" + - "\x00\x00\x00\x18n\x02\x00\x00\x00\x80|\x18n\x00\x18n\x02\x00\x00\x00" + - "\x18n\x02\x00\x18n" diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458 b/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458 deleted file mode 100644 index 722be7db3..000000000 Binary files a/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458 and /dev/null differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458.output b/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458.output deleted file mode 100644 index af036901d..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458.output +++ /dev/null @@ -1,23 +0,0 @@ -panic: runtime error: slice bounds out of range [recovered] - panic: runtime error: slice bounds out of range - -goroutine 1 [running]: -io/ioutil.readAll.func1(0x8202b1e10) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/io/ioutil/ioutil.go:30 +0x228 -github.com/pierrec/lz4.(*Reader).readBlock(0x820312000, 0x820316000, 0x10000, 0x10000, 0x8202900c0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/reader.go:241 +0xc62 -github.com/pierrec/lz4.(*Reader).Read(0x820312000, 0x820314000, 0x200, 0x200, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/reader.go:178 +0x7a6 -bytes.(*Buffer).ReadFrom(0x8202b1d68, 0x88204290f0, 0x820312000, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/bytes/buffer.go:173 +0x3db -io/ioutil.readAll(0x88204290f0, 0x820312000, 0x200, 0x0, 0x0, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/io/ioutil/ioutil.go:33 +0x1ed -io/ioutil.ReadAll(0x88204290f0, 0x820312000, 0x0, 0x0, 0x0, 0x0, 0x0) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/io/ioutil/ioutil.go:42 +0x80 -github.com/pierrec/lz4/fuzz.Fuzz(0x8820479000, 0x13, 0x200000, 0x8202900b8) - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/github.com/pierrec/lz4/fuzz/lz4.go:11 +0x15f -github.com/dvyukov/go-fuzz/go-fuzz-dep.Main(0x1a7f18) - /Users/pierrecurto/sandbox/src/github.com/dvyukov/go-fuzz/go-fuzz-dep/main.go:47 +0x14c -main.main() - /var/folders/bw/wf4p9qr50pg23qb4py4028140000gp/T/go-fuzz-build320605510/src/go-fuzz-main/main.go:10 +0x23 -exit status 2 \ No newline at end of file diff --git a/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458.quoted b/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458.quoted deleted file mode 100644 index 4f6ce344f..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/crashers/ea0a00651ba4143c05fe7b5c85f69fe16a29a458.quoted +++ /dev/null @@ -1 +0,0 @@ - "\x04\"M\x18M@\x00\x00B*M\f\x00'\x01000\xe4" diff --git a/vendor/github.com/pierrec/lz4/fuzz/lz4-fuzz.zip b/vendor/github.com/pierrec/lz4/fuzz/lz4-fuzz.zip index 4abfa3d3e..a7783c029 100644 Binary files a/vendor/github.com/pierrec/lz4/fuzz/lz4-fuzz.zip and b/vendor/github.com/pierrec/lz4/fuzz/lz4-fuzz.zip differ diff --git a/vendor/github.com/pierrec/lz4/fuzz/suppressions/a596442269a13f32d85889a173f2d36187a768c6 b/vendor/github.com/pierrec/lz4/fuzz/suppressions/a596442269a13f32d85889a173f2d36187a768c6 deleted file mode 100644 index 1f1915235..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/suppressions/a596442269a13f32d85889a173f2d36187a768c6 +++ /dev/null @@ -1 +0,0 @@ -SIGABRT: abort diff --git a/vendor/github.com/pierrec/lz4/fuzz/suppressions/d159e91cdd6fcbee9e37460f96c597b70c590886 b/vendor/github.com/pierrec/lz4/fuzz/suppressions/d159e91cdd6fcbee9e37460f96c597b70c590886 deleted file mode 100644 index 381c311ec..000000000 --- a/vendor/github.com/pierrec/lz4/fuzz/suppressions/d159e91cdd6fcbee9e37460f96c597b70c590886 +++ /dev/null @@ -1,10 +0,0 @@ -panic: runtime error: slice bounds out of range [recovered] -io/ioutil.readAll.func1 -github.com/pierrec/lz4.(*Reader).readBlock -github.com/pierrec/lz4.(*Reader).Read -bytes.(*Buffer).ReadFrom -io/ioutil.readAll -io/ioutil.ReadAll -github.com/pierrec/lz4/fuzz.Fuzz -github.com/dvyukov/go-fuzz/go-fuzz-dep.Main -main.main diff --git a/vendor/github.com/pierrec/lz4/internal/xxh32/xxh32zero.go b/vendor/github.com/pierrec/lz4/internal/xxh32/xxh32zero.go new file mode 100644 index 000000000..850a6fdf6 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/internal/xxh32/xxh32zero.go @@ -0,0 +1,222 @@ +// Package xxh32 implements the very fast XXH hashing algorithm (32 bits version). +// (https://github.com/Cyan4973/XXH/) +package xxh32 + +import ( + "encoding/binary" +) + +const ( + prime32_1 uint32 = 2654435761 + prime32_2 uint32 = 2246822519 + prime32_3 uint32 = 3266489917 + prime32_4 uint32 = 668265263 + prime32_5 uint32 = 374761393 + + prime32_1plus2 uint32 = 606290984 + prime32_minus1 uint32 = 1640531535 +) + +// XXHZero represents an xxhash32 object with seed 0. +type XXHZero struct { + v1 uint32 + v2 uint32 + v3 uint32 + v4 uint32 + totalLen uint64 + buf [16]byte + bufused int +} + +// Sum appends the current hash to b and returns the resulting slice. +// It does not change the underlying hash state. +func (xxh XXHZero) Sum(b []byte) []byte { + h32 := xxh.Sum32() + return append(b, byte(h32), byte(h32>>8), byte(h32>>16), byte(h32>>24)) +} + +// Reset resets the Hash to its initial state. +func (xxh *XXHZero) Reset() { + xxh.v1 = prime32_1plus2 + xxh.v2 = prime32_2 + xxh.v3 = 0 + xxh.v4 = prime32_minus1 + xxh.totalLen = 0 + xxh.bufused = 0 +} + +// Size returns the number of bytes returned by Sum(). +func (xxh *XXHZero) Size() int { + return 4 +} + +// BlockSize gives the minimum number of bytes accepted by Write(). +func (xxh *XXHZero) BlockSize() int { + return 1 +} + +// Write adds input bytes to the Hash. +// It never returns an error. +func (xxh *XXHZero) Write(input []byte) (int, error) { + if xxh.totalLen == 0 { + xxh.Reset() + } + n := len(input) + m := xxh.bufused + + xxh.totalLen += uint64(n) + + r := len(xxh.buf) - m + if n < r { + copy(xxh.buf[m:], input) + xxh.bufused += len(input) + return n, nil + } + + p := 0 + // Causes compiler to work directly from registers instead of stack: + v1, v2, v3, v4 := xxh.v1, xxh.v2, xxh.v3, xxh.v4 + if m > 0 { + // some data left from previous update + copy(xxh.buf[xxh.bufused:], input[:r]) + xxh.bufused += len(input) - r + + // fast rotl(13) + buf := xxh.buf[:16] // BCE hint. + v1 = rol13(v1+binary.LittleEndian.Uint32(buf[:])*prime32_2) * prime32_1 + v2 = rol13(v2+binary.LittleEndian.Uint32(buf[4:])*prime32_2) * prime32_1 + v3 = rol13(v3+binary.LittleEndian.Uint32(buf[8:])*prime32_2) * prime32_1 + v4 = rol13(v4+binary.LittleEndian.Uint32(buf[12:])*prime32_2) * prime32_1 + p = r + xxh.bufused = 0 + } + + for n := n - 16; p <= n; p += 16 { + sub := input[p:][:16] //BCE hint for compiler + v1 = rol13(v1+binary.LittleEndian.Uint32(sub[:])*prime32_2) * prime32_1 + v2 = rol13(v2+binary.LittleEndian.Uint32(sub[4:])*prime32_2) * prime32_1 + v3 = rol13(v3+binary.LittleEndian.Uint32(sub[8:])*prime32_2) * prime32_1 + v4 = rol13(v4+binary.LittleEndian.Uint32(sub[12:])*prime32_2) * prime32_1 + } + xxh.v1, xxh.v2, xxh.v3, xxh.v4 = v1, v2, v3, v4 + + copy(xxh.buf[xxh.bufused:], input[p:]) + xxh.bufused += len(input) - p + + return n, nil +} + +// Sum32 returns the 32 bits Hash value. +func (xxh *XXHZero) Sum32() uint32 { + h32 := uint32(xxh.totalLen) + if h32 >= 16 { + h32 += rol1(xxh.v1) + rol7(xxh.v2) + rol12(xxh.v3) + rol18(xxh.v4) + } else { + h32 += prime32_5 + } + + p := 0 + n := xxh.bufused + buf := xxh.buf + for n := n - 4; p <= n; p += 4 { + h32 += binary.LittleEndian.Uint32(buf[p:p+4]) * prime32_3 + h32 = rol17(h32) * prime32_4 + } + for ; p < n; p++ { + h32 += uint32(buf[p]) * prime32_5 + h32 = rol11(h32) * prime32_1 + } + + h32 ^= h32 >> 15 + h32 *= prime32_2 + h32 ^= h32 >> 13 + h32 *= prime32_3 + h32 ^= h32 >> 16 + + return h32 +} + +// ChecksumZero returns the 32bits Hash value. +func ChecksumZero(input []byte) uint32 { + n := len(input) + h32 := uint32(n) + + if n < 16 { + h32 += prime32_5 + } else { + v1 := prime32_1plus2 + v2 := prime32_2 + v3 := uint32(0) + v4 := prime32_minus1 + p := 0 + for n := n - 16; p <= n; p += 16 { + sub := input[p:][:16] //BCE hint for compiler + v1 = rol13(v1+binary.LittleEndian.Uint32(sub[:])*prime32_2) * prime32_1 + v2 = rol13(v2+binary.LittleEndian.Uint32(sub[4:])*prime32_2) * prime32_1 + v3 = rol13(v3+binary.LittleEndian.Uint32(sub[8:])*prime32_2) * prime32_1 + v4 = rol13(v4+binary.LittleEndian.Uint32(sub[12:])*prime32_2) * prime32_1 + } + input = input[p:] + n -= p + h32 += rol1(v1) + rol7(v2) + rol12(v3) + rol18(v4) + } + + p := 0 + for n := n - 4; p <= n; p += 4 { + h32 += binary.LittleEndian.Uint32(input[p:p+4]) * prime32_3 + h32 = rol17(h32) * prime32_4 + } + for p < n { + h32 += uint32(input[p]) * prime32_5 + h32 = rol11(h32) * prime32_1 + p++ + } + + h32 ^= h32 >> 15 + h32 *= prime32_2 + h32 ^= h32 >> 13 + h32 *= prime32_3 + h32 ^= h32 >> 16 + + return h32 +} + +// Uint32Zero hashes x with seed 0. +func Uint32Zero(x uint32) uint32 { + h := prime32_5 + 4 + x*prime32_3 + h = rol17(h) * prime32_4 + h ^= h >> 15 + h *= prime32_2 + h ^= h >> 13 + h *= prime32_3 + h ^= h >> 16 + return h +} + +func rol1(u uint32) uint32 { + return u<<1 | u>>31 +} + +func rol7(u uint32) uint32 { + return u<<7 | u>>25 +} + +func rol11(u uint32) uint32 { + return u<<11 | u>>21 +} + +func rol12(u uint32) uint32 { + return u<<12 | u>>20 +} + +func rol13(u uint32) uint32 { + return u<<13 | u>>19 +} + +func rol17(u uint32) uint32 { + return u<<17 | u>>15 +} + +func rol18(u uint32) uint32 { + return u<<18 | u>>14 +} diff --git a/vendor/github.com/pierrec/xxHash/xxHash32/xxHash32_test.go b/vendor/github.com/pierrec/lz4/internal/xxh32/xxh32zero_test.go similarity index 57% rename from vendor/github.com/pierrec/xxHash/xxHash32/xxHash32_test.go rename to vendor/github.com/pierrec/lz4/internal/xxh32/xxh32zero_test.go index 63e054869..afec7ab1e 100644 --- a/vendor/github.com/pierrec/xxHash/xxHash32/xxHash32_test.go +++ b/vendor/github.com/pierrec/lz4/internal/xxh32/xxh32zero_test.go @@ -1,4 +1,4 @@ -package xxHash32_test +package xxh32_test import ( "encoding/binary" @@ -6,7 +6,7 @@ import ( "hash/fnv" "testing" - "github.com/pierrec/xxHash/xxHash32" + "github.com/pierrec/lz4/internal/xxh32" ) type test struct { @@ -15,20 +15,20 @@ type test struct { } var testdata = []test{ - {0x02cc5d05, "", ""}, - {0x550d7456, "a", ""}, - {0x4999fc53, "ab", ""}, - {0x32d153ff, "abc", ""}, - {0xa3643705, "abcd", ""}, - {0x9738f19b, "abcde", ""}, - {0x8b7cd587, "abcdef", ""}, - {0x9dd093b3, "abcdefg", ""}, - {0x0bb3c6bb, "abcdefgh", ""}, - {0xd03c13fd, "abcdefghi", ""}, - {0x8b988cfe, "abcdefghij", ""}, + // {0x02cc5d05, "", ""}, + // {0x550d7456, "a", ""}, + // {0x4999fc53, "ab", ""}, + // {0x32d153ff, "abc", ""}, + // {0xa3643705, "abcd", ""}, + // {0x9738f19b, "abcde", ""}, + // {0x8b7cd587, "abcdef", ""}, + // {0x9dd093b3, "abcdefg", ""}, + // {0x0bb3c6bb, "abcdefgh", ""}, + // {0xd03c13fd, "abcdefghi", ""}, + // {0x8b988cfe, "abcdefghij", ""}, {0x9d2d8b62, "abcdefghijklmnop", ""}, - {0x42ae804d, "abcdefghijklmnopqrstuvwxyz0123456789", ""}, - {0x62b4ed00, "Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.", ""}, + // {0x42ae804d, "abcdefghijklmnopqrstuvwxyz0123456789", ""}, + // {0x62b4ed00, "Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.", ""}, } func init() { @@ -42,39 +42,39 @@ func init() { } } -func TestBlockSize(t *testing.T) { - xxh := xxHash32.New(0) +func TestZeroBlockSize(t *testing.T) { + var xxh xxh32.XXHZero if s := xxh.BlockSize(); s <= 0 { t.Errorf("invalid BlockSize: %d", s) } } -func TestSize(t *testing.T) { - xxh := xxHash32.New(0) +func TestZeroSize(t *testing.T) { + var xxh xxh32.XXHZero if s := xxh.Size(); s != 4 { t.Errorf("invalid Size: got %d expected 4", s) } } -func TestData(t *testing.T) { +func TestZeroData(t *testing.T) { for i, td := range testdata { - xxh := xxHash32.New(0) + var xxh xxh32.XXHZero data := []byte(td.data) xxh.Write(data) if h := xxh.Sum32(); h != td.sum { t.Errorf("test %d: xxh32(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) t.FailNow() } - if h := xxHash32.Checksum(data, 0); h != td.sum { + if h := xxh32.ChecksumZero(data); h != td.sum { t.Errorf("test %d: xxh32(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) t.FailNow() } } } -func TestSplitData(t *testing.T) { +func TestZeroSplitData(t *testing.T) { for i, td := range testdata { - xxh := xxHash32.New(0) + var xxh xxh32.XXHZero data := []byte(td.data) l := len(data) / 2 xxh.Write(data[0:l]) @@ -87,9 +87,9 @@ func TestSplitData(t *testing.T) { } } -func TestSum(t *testing.T) { +func TestZeroSum(t *testing.T) { for i, td := range testdata { - xxh := xxHash32.New(0) + var xxh xxh32.XXHZero data := []byte(td.data) xxh.Write(data) b := xxh.Sum(data) @@ -100,8 +100,19 @@ func TestSum(t *testing.T) { } } -func TestReset(t *testing.T) { - xxh := xxHash32.New(0) +func TestZeroChecksum(t *testing.T) { + for i, td := range testdata { + data := []byte(td.data) + h := xxh32.ChecksumZero(data) + if h != td.sum { + t.Errorf("test %d: xxh32(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) + t.FailNow() + } + } +} + +func TestZeroReset(t *testing.T) { + var xxh xxh32.XXHZero for i, td := range testdata { xxh.Write([]byte(td.data)) h := xxh.Sum32() @@ -119,7 +130,7 @@ func TestReset(t *testing.T) { var testdata1 = []byte(testdata[len(testdata)-1].data) func Benchmark_XXH32(b *testing.B) { - h := xxHash32.New(0) + var h xxh32.XXHZero for n := 0; n < b.N; n++ { h.Write(testdata1) h.Sum32() @@ -129,7 +140,7 @@ func Benchmark_XXH32(b *testing.B) { func Benchmark_XXH32_Checksum(b *testing.B) { for n := 0; n < b.N; n++ { - xxHash32.Checksum(testdata1, 0) + xxh32.ChecksumZero(testdata1) } } diff --git a/vendor/github.com/pierrec/lz4/lz4.go b/vendor/github.com/pierrec/lz4/lz4.go index ddb82f66f..35802756c 100644 --- a/vendor/github.com/pierrec/lz4/lz4.go +++ b/vendor/github.com/pierrec/lz4/lz4.go @@ -1,13 +1,5 @@ // Package lz4 implements reading and writing lz4 compressed data (a frame), -// as specified in http://fastcompression.blogspot.fr/2013/04/lz4-streaming-format-final.html, -// using an io.Reader (decompression) and io.Writer (compression). -// It is designed to minimize memory usage while maximizing throughput by being able to -// [de]compress data concurrently. -// -// The Reader and the Writer support concurrent processing provided the supplied buffers are -// large enough (in multiples of BlockMaxSize) and there is no block dependency. -// Reader.WriteTo and Writer.ReadFrom do leverage the concurrency transparently. -// The runtime.GOMAXPROCS() value is used to apply concurrency or not. +// as specified in http://fastcompression.blogspot.fr/2013/04/lz4-streaming-format-final.html. // // Although the block level compression and decompression functions are exposed and are fully compatible // with the lz4 block format definition, they are low level and should not be used directly. @@ -15,29 +7,25 @@ // http://fastcompression.blogspot.fr/2011/05/lz4-explained.html // // See https://github.com/Cyan4973/lz4 for the reference C implementation. +// package lz4 -import ( - "hash" - "sync" - - "github.com/pierrec/xxHash/xxHash32" -) - const ( // Extension is the LZ4 frame file name extension Extension = ".lz4" // Version is the LZ4 frame format version Version = 1 - frameMagic = uint32(0x184D2204) - frameSkipMagic = uint32(0x184D2A50) + frameMagic uint32 = 0x184D2204 + frameSkipMagic uint32 = 0x184D2A50 // The following constants are used to setup the compression algorithm. - minMatch = 4 // the minimum size of the match sequence size (4 bytes) - winSizeLog = 16 // LZ4 64Kb window size limit - winSize = 1 << winSizeLog - winMask = winSize - 1 // 64Kb window of previous data for dependent blocks + minMatch = 4 // the minimum size of the match sequence size (4 bytes) + winSizeLog = 16 // LZ4 64Kb window size limit + winSize = 1 << winSizeLog + winMask = winSize - 1 // 64Kb window of previous data for dependent blocks + compressedBlockFlag = 1 << 31 + compressedBlockMask = compressedBlockFlag - 1 // hashLog determines the size of the hash table used to quickly find a previous match position. // Its value influences the compression speed and memory usage, the lower the faster, @@ -49,13 +37,13 @@ const ( mfLimit = 8 + minMatch // The last match cannot start within the last 12 bytes. skipStrength = 6 // variable step for fast scan - - hasher = uint32(2654435761) // prime number used to hash minMatch ) // map the block max size id with its value in bytes: 64Kb, 256Kb, 1Mb and 4Mb. -var bsMapID = map[byte]int{4: 64 << 10, 5: 256 << 10, 6: 1 << 20, 7: 4 << 20} -var bsMapValue = map[int]byte{} +var ( + bsMapID = map[byte]int{4: 64 << 10, 5: 256 << 10, 6: 1 << 20, 7: 4 << 20} + bsMapValue = make(map[int]byte, len(bsMapID)) +) // Reversed. func init() { @@ -65,41 +53,16 @@ func init() { } // Header describes the various flags that can be set on a Writer or obtained from a Reader. -// The default values match those of the LZ4 frame format definition (http://fastcompression.blogspot.com/2013/04/lz4-streaming-format-final.html). +// The default values match those of the LZ4 frame format definition +// (http://fastcompression.blogspot.com/2013/04/lz4-streaming-format-final.html). // // NB. in a Reader, in case of concatenated frames, the Header values may change between Read() calls. -// It is the caller responsibility to check them if necessary (typically when using the Reader concurrency). +// It is the caller responsibility to check them if necessary. type Header struct { - BlockDependency bool // compressed blocks are dependent (one block depends on the last 64Kb of the previous one) - BlockChecksum bool // compressed blocks are checksumed - NoChecksum bool // frame checksum - BlockMaxSize int // the size of the decompressed data block (one of [64KB, 256KB, 1MB, 4MB]). Default=4MB. - Size uint64 // the frame total size. It is _not_ computed by the Writer. - HighCompression bool // use high compression (only for the Writer) - done bool // whether the descriptor was processed (Read or Write and checked) - // Removed as not supported - // Dict bool // a dictionary id is to be used - // DictID uint32 // the dictionary id read from the frame, if any. -} - -// xxhPool wraps the standard pool for xxHash items. -// Putting items back in the pool automatically resets them. -type xxhPool struct { - sync.Pool -} - -func (p *xxhPool) Get() hash.Hash32 { - return p.Pool.Get().(hash.Hash32) -} - -func (p *xxhPool) Put(h hash.Hash32) { - h.Reset() - p.Pool.Put(h) -} - -// hashPool is used by readers and writers and contains xxHash items. -var hashPool = xxhPool{ - Pool: sync.Pool{ - New: func() interface{} { return xxHash32.New(0) }, - }, + BlockChecksum bool // Compressed blocks checksum flag. + NoChecksum bool // Frame checksum flag. + BlockMaxSize int // Size of the uncompressed data block (one of [64KB, 256KB, 1MB, 4MB]). Default=4MB. + Size uint64 // Frame total size. It is _not_ computed by the Writer. + CompressionLevel int // Compression level (higher is better, use 0 for fastest compression). + done bool // Header processed flag (Read or Write and checked). } diff --git a/vendor/github.com/pierrec/lz4/lz4_go1.10.go b/vendor/github.com/pierrec/lz4/lz4_go1.10.go new file mode 100644 index 000000000..9a0fb0070 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/lz4_go1.10.go @@ -0,0 +1,29 @@ +//+build go1.10 + +package lz4 + +import ( + "fmt" + "strings" +) + +func (h Header) String() string { + var s strings.Builder + + s.WriteString(fmt.Sprintf("%T{", h)) + if h.BlockChecksum { + s.WriteString("BlockChecksum: true ") + } + if h.NoChecksum { + s.WriteString("NoChecksum: true ") + } + if bs := h.BlockMaxSize; bs != 0 && bs != 4<<20 { + s.WriteString(fmt.Sprintf("BlockMaxSize: %d ", bs)) + } + if l := h.CompressionLevel; l != 0 { + s.WriteString(fmt.Sprintf("CompressionLevel: %d ", l)) + } + s.WriteByte('}') + + return s.String() +} diff --git a/vendor/github.com/pierrec/lz4/lz4_notgo1.10.go b/vendor/github.com/pierrec/lz4/lz4_notgo1.10.go new file mode 100644 index 000000000..12c761a2e --- /dev/null +++ b/vendor/github.com/pierrec/lz4/lz4_notgo1.10.go @@ -0,0 +1,29 @@ +//+build !go1.10 + +package lz4 + +import ( + "bytes" + "fmt" +) + +func (h Header) String() string { + var s bytes.Buffer + + s.WriteString(fmt.Sprintf("%T{", h)) + if h.BlockChecksum { + s.WriteString("BlockChecksum: true ") + } + if h.NoChecksum { + s.WriteString("NoChecksum: true ") + } + if bs := h.BlockMaxSize; bs != 0 && bs != 4<<20 { + s.WriteString(fmt.Sprintf("BlockMaxSize: %d ", bs)) + } + if l := h.CompressionLevel; l != 0 { + s.WriteString(fmt.Sprintf("CompressionLevel: %d ", l)) + } + s.WriteByte('}') + + return s.String() +} diff --git a/vendor/github.com/pierrec/lz4/lz4_test.go b/vendor/github.com/pierrec/lz4/lz4_test.go deleted file mode 100644 index a196c8d5b..000000000 --- a/vendor/github.com/pierrec/lz4/lz4_test.go +++ /dev/null @@ -1,727 +0,0 @@ -package lz4_test - -import ( - "bytes" - "crypto/rand" - "encoding/binary" - "fmt" - "io" - "io/ioutil" - "math/big" - "os" - "reflect" - "testing" - - "github.com/pierrec/lz4" -) - -// testBuffer wraps bytes.Buffer to remove the WriteTo() and ReadFrom() methods. -type testBuffer struct { - buf *bytes.Buffer -} - -func (b *testBuffer) Read(buf []byte) (int, error) { - return b.buf.Read(buf) -} - -func (b *testBuffer) Write(buf []byte) (int, error) { - return b.buf.Write(buf) -} - -func (b *testBuffer) Len() int { - return b.buf.Len() -} - -func (b *testBuffer) Bytes() []byte { - return b.buf.Bytes() -} - -// testData represents a test data item. It is really used to provide a human readable label to a slice of bytes. -type testData struct { - label string - data []byte -} - -// testHeader represents a test data item. It is really used to provide a human readable label to an LZ4 header. -type testHeader struct { - label string - header lz4.Header -} - -// compareHeaders... compares 2 lz4 headers. -func compareHeaders(h, hh lz4.Header, t *testing.T) { - ok := true - if h.BlockDependency != hh.BlockDependency { - t.Errorf("BlockDependency: expected %v, got %v", h.BlockDependency, hh.BlockDependency) - ok = false - } - if h.BlockChecksum != hh.BlockChecksum { - t.Errorf("BlockChecksum: expected %v, got %v", h.BlockChecksum, hh.BlockChecksum) - ok = false - } - if h.NoChecksum != hh.NoChecksum { - t.Errorf("NoChecksum: expected %v, got %v", h.NoChecksum, hh.NoChecksum) - ok = false - } - if h.BlockMaxSize != hh.BlockMaxSize { - t.Errorf("BlockMaxSize: expected %d, got %d", h.BlockMaxSize, hh.BlockMaxSize) - ok = false - } - if h.Size != hh.Size { - t.Errorf("Size: expected %d, got %d", h.Size, hh.Size) - ok = false - } - // if h.Dict != hh.Dict { - // t.Errorf("Dict: expected %d, got %d", h.Dict, hh.Dict) - // ok = false - // } - // if h.DictID != hh.DictID { - // t.Errorf("DictID: expected %d, got %d", h.DictID, hh.DictID) - // ok = false - // } - if !ok { - t.FailNow() - } -} - -var ( - lorem = []byte("Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.") - // Initial data items used for testing. More are added with random and other kind of data. - testDataItems = []testData{ - {"empty", nil}, - { - "small pattern", - []byte("aaaaaaaaaaaaaaaaaaa"), - }, - { - "small pattern long", - []byte("aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa"), - }, - { - "medium pattern", - []byte("abcdefghijklmnopqabcdefghijklmnopq"), - }, - { - "lorem", - lorem, - }, - } - testHeaderItems = []testHeader{} -) - -// Build the list of all possible headers with the default values + the ones defined in the map. -func buildHeaders(options map[string][]interface{}) []testHeader { - testHeaderItems := make([]testHeader, 1) - for fieldName, fieldData := range options { - for _, o := range fieldData { - for _, d := range testHeaderItems { - s := reflect.ValueOf(&d.header).Elem() - t := s.Type() - for i := 0; i < s.NumField(); i++ { - if t.Field(i).Name == fieldName { - switch f := s.Field(i); f.Kind() { - case reflect.Bool: - f.SetBool(o.(bool)) - case reflect.Int: - f.SetInt(int64(o.(int))) - case reflect.Int64: - switch o.(type) { - case int: - f.SetInt(int64(o.(int))) - default: - f.SetInt(o.(int64)) - } - case reflect.Uint32: - switch o.(type) { - case int: - f.SetUint(uint64(o.(int))) - default: - f.SetUint(uint64(o.(uint32))) - } - case reflect.Uint64: - switch o.(type) { - case int: - f.SetUint(uint64(o.(int))) - default: - f.SetUint(o.(uint64)) - } - default: - panic(fmt.Sprintf("unsupported type: %v", f.Kind())) - } - d.label = fmt.Sprintf("%+v", d.header) - testHeaderItems = append(testHeaderItems, d) - break - } - } - } - } - } - - for i, n := 0, len(testHeaderItems); i < n; { - testHeaderItem := testHeaderItems[i] - // remove the 0 BlockMaxSize value as it is invalid and we have provisioned all possible values already. - if testHeaderItem.header.BlockMaxSize == 0 { - n-- - testHeaderItems[i], testHeaderItems = testHeaderItems[n], testHeaderItems[:n] - } else { - testHeaderItem.label = fmt.Sprintf("%+v", testHeaderItem) - i++ - } - } - - return testHeaderItems -} - -// Generate all possible LZ4 headers. -func init() { - // Only set the relevant headers having an impact on the comrpession. - seed := map[string][]interface{}{ - "BlockDependency": {true}, - "BlockChecksum": {true}, - "NoChecksum": {true}, - // "Dict": {true}, - // Enabling this substantially increase the testing time. - // As this test is not really required it is disabled. - // "HighCompression": {true}, - } - for _, bms := range lz4.BlockMaxSizeItems { - seed["BlockMaxSize"] = append(seed["BlockMaxSize"], bms) - } - testHeaderItems = buildHeaders(seed) -} - -// Initialize the test data with various sizes of uncompressible and compressible data. -func init() { - maxSize := 10 << 20 // > max block max size of 4Mb - - // repeated data with very high compression ratio - repeat := make([]byte, maxSize) - for i := copy(repeat, lorem); i < len(repeat); { - i += copy(repeat[i:], repeat[:i]) - } - - // repeated data with small compression ratio - repeatlow := make([]byte, maxSize) - for i := 0; i < len(repeatlow); { - i += copy(repeatlow[i:], lorem) - // randomly skip some bytes to make sure the pattern does not repeat too much - n, _ := rand.Int(rand.Reader, big.NewInt(int64(10))) - i += int(n.Int64()) - } - - // random data: low to no compression - random := make([]byte, maxSize) - if _, err := rand.Read(random); err != nil { - panic(fmt.Sprintf("cannot initialize random data for size %d", maxSize)) - } - - // generate some test data with various sizes and kind of data: all valid block max sizes + others - for _, size := range lz4.BlockMaxSizeItems { - testDataItems = append( - testDataItems, - testData{fmt.Sprintf("random %d", size), random[:size]}, - testData{fmt.Sprintf("random < %d", size), random[:size/3]}, - testData{fmt.Sprintf("repeated %d", size), repeat[:size]}, - testData{fmt.Sprintf("repeated < %d", size), repeat[:size/3]}, - ) - } -} - -// Test low levels core functions: -// a. compress and compare with supplied data if any -// b. decompress the previous data and compare it with the original one -func TestBlock(t *testing.T) { - for _, compress := range []func([]byte, []byte, int) (int, error){ - lz4.CompressBlock, - lz4.CompressBlockHC, - } { - for _, item := range testDataItems { - data := item.data - z := make([]byte, lz4.CompressBlockBound(len(data))) - n, err := compress(data, z, 0) - if n == 0 { // not compressible - continue - } - if err != nil { - t.Errorf("CompressBlock: %s", err) - t.FailNow() - } - z = z[:n] - d := make([]byte, len(data)) - n, err = lz4.UncompressBlock(z, d, 0) - if err != nil { - t.Errorf("UncompressBlock: %s", err) - t.FailNow() - } - d = d[:n] - if !bytes.Equal(d, data) { - t.Errorf("invalid decompressed data: %s: %s", item.label, string(d)) - t.FailNow() - } - } - } -} - -func TestBlockCompression(t *testing.T) { - input := make([]byte, 64*1024) - - for i := 0; i < 64*1024; i += 1 { - input[i] = byte(i & 0x1) - } - output := make([]byte, 64*1024) - - c, err := lz4.CompressBlock(input, output, 0) - - if err != nil { - t.Fatal(err) - } - - if c == 0 { - t.Fatal("cannot compress compressible data") - } -} - -func BenchmarkUncompressBlock(b *testing.B) { - d := make([]byte, len(lorem)) - z := make([]byte, len(lorem)) - n, err := lz4.CompressBlock(lorem, z, 0) - if err != nil { - b.Errorf("CompressBlock: %s", err) - b.FailNow() - } - z = z[:n] - for i := 0; i < b.N; i++ { - lz4.UncompressBlock(z, d, 0) - } -} - -func BenchmarkUncompressConstantBlock(b *testing.B) { - d := make([]byte, 4096) - z := make([]byte, 4096) - source := make([]byte, 4096) - n, err := lz4.CompressBlock(source, z, 0) - if err != nil { - b.Errorf("CompressBlock: %s", err) - b.FailNow() - } - z = z[:n] - for i := 0; i < b.N; i++ { - lz4.UncompressBlock(z, d, 0) - } -} - -func BenchmarkCompressBlock(b *testing.B) { - d := append([]byte{}, lorem...) - z := make([]byte, len(lorem)) - n, err := lz4.CompressBlock(d, z, 0) - if err != nil { - b.Errorf("CompressBlock: %s", err) - b.FailNow() - } - z = z[:n] - for i := 0; i < b.N; i++ { - d = append([]byte{}, lorem...) - lz4.CompressBlock(d, z, 0) - } -} - -func BenchmarkCompressConstantBlock(b *testing.B) { - d := make([]byte, 4096) - z := make([]byte, 4096) - n, err := lz4.CompressBlock(d, z, 0) - if err != nil { - b.Errorf("CompressBlock: %s", err) - b.FailNow() - } - z = z[:n] - for i := 0; i < b.N; i++ { - lz4.CompressBlock(d, z, 0) - } -} - -func BenchmarkCompressBlockHC(b *testing.B) { - d := append([]byte{}, lorem...) - z := make([]byte, len(lorem)) - n, err := lz4.CompressBlockHC(d, z, 0) - if err != nil { - b.Errorf("CompressBlock: %s", err) - b.FailNow() - } - z = z[:n] - for i := 0; i < b.N; i++ { - d = append([]byte{}, lorem...) - lz4.CompressBlockHC(d, z, 0) - } -} -func BenchmarkCompressEndToEnd(b *testing.B) { - w := lz4.NewWriter(ioutil.Discard) - b.ResetTimer() - for i := 0; i < b.N; i++ { - if _, err := w.Write(lorem); err != nil { - b.Fatal(err) - } - } -} - -// TestNoWrite compresses without any call to Write() (empty frame). -// It does so checking all possible headers. -func TestNoWrite(t *testing.T) { - // that is 2*2*2*2*2*2^4 = 512 headers! - seed := map[string][]interface{}{ - "BlockDependency": {true}, - "BlockChecksum": {true}, - "NoChecksum": {true}, - "Size": {999}, - // "Dict": {true}, - // Enabling this substantially increase the testing time. - // As this test is not really required it is disabled. - // "HighCompression": {true}, - } - for _, bms := range lz4.BlockMaxSizeItems { - seed["BlockMaxSize"] = append(seed["BlockMaxSize"], bms) - } - testHeaderItems := buildHeaders(seed) - - for _, h := range testHeaderItems { - rw := bytes.NewBuffer(nil) - - w := lz4.NewWriter(rw) - w.Header = h.header - if err := w.Close(); err != nil { - t.Errorf("Close(): unexpected error: %v", err) - t.FailNow() - } - - r := lz4.NewReader(rw) - n, err := r.Read(nil) - if err != nil { - t.Errorf("Read(): unexpected error: %v", err) - t.FailNow() - } - if n != 0 { - t.Errorf("expected 0 bytes read, got %d", n) - t.FailNow() - } - - buf := make([]byte, 16) - n, err = r.Read(buf) - if err != nil && err != io.EOF { - t.Errorf("Read(): unexpected error: %v", err) - t.FailNow() - } - if n != 0 { - t.Errorf("expected 0 bytes read, got %d", n) - t.FailNow() - } - } -} - -// TestReset tests that the Reset() method resets the header on the Reader and Writer. -func TestReset(t *testing.T) { - h := lz4.Header{ - BlockDependency: true, - BlockChecksum: true, - NoChecksum: true, - BlockMaxSize: 123, - Size: 999, - // Dict: true, - // DictID: 555, - } - dh := lz4.Header{} - - w := lz4.NewWriter(nil) - w.Header = h - w.Reset(nil) - compareHeaders(w.Header, dh, t) - - r := lz4.NewReader(nil) - r.Header = h - r.Reset(nil) - compareHeaders(r.Header, dh, t) -} - -// TestFrame compresses and decompresses LZ4 streams with various input data and options. -func TestFrame(t *testing.T) { - for _, tdata := range testDataItems { - data := tdata.data - t.Run(tdata.label, func(t *testing.T) { - t.Parallel() - // test various options - for _, headerItem := range testHeaderItems { - tag := tdata.label + ": " + headerItem.label - rw := bytes.NewBuffer(nil) - - // Set all options to non default values and compress - w := lz4.NewWriter(rw) - w.Header = headerItem.header - - n, err := w.Write(data) - if err != nil { - t.Errorf("%s: Write(): unexpected error: %v", tag, err) - t.FailNow() - } - if n != len(data) { - t.Errorf("%s: Write(): expected %d bytes written, got %d", tag, len(data), n) - t.FailNow() - } - if err = w.Close(); err != nil { - t.Errorf("%s: Close(): unexpected error: %v", tag, err) - t.FailNow() - } - - // Decompress - r := lz4.NewReader(rw) - n, err = r.Read(nil) - if err != nil { - t.Errorf("%s: Read(): unexpected error: %v", tag, err) - t.FailNow() - } - if n != 0 { - t.Errorf("%s: Read(): expected 0 bytes read, got %d", tag, n) - } - - buf := make([]byte, len(data)) - n, err = r.Read(buf) - if err != nil && err != io.EOF { - t.Errorf("%s: Read(): unexpected error: %v", tag, err) - t.FailNow() - } - if n != len(data) { - t.Errorf("%s: Read(): expected %d bytes read, got %d", tag, len(data), n) - } - buf = buf[:n] - if !bytes.Equal(buf, data) { - t.Errorf("%s: decompress(compress(data)) != data (%d/%d)", tag, len(buf), len(data)) - t.FailNow() - } - - compareHeaders(w.Header, r.Header, t) - } - }) - } -} - -// TestReadFromWriteTo tests the Reader.WriteTo() and Writer.ReadFrom() methods. -func TestReadFromWriteTo(t *testing.T) { - for _, tdata := range testDataItems { - data := tdata.data - - t.Run(tdata.label, func(t *testing.T) { - t.Parallel() - // test various options - for _, headerItem := range testHeaderItems { - tag := "ReadFromWriteTo: " + tdata.label + ": " + headerItem.label - dbuf := bytes.NewBuffer(data) - - zbuf := bytes.NewBuffer(nil) - w := lz4.NewWriter(zbuf) - w.Header = headerItem.header - if _, err := w.ReadFrom(dbuf); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - if err := w.Close(); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - buf := bytes.NewBuffer(nil) - r := lz4.NewReader(zbuf) - if _, err := r.WriteTo(buf); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - if !bytes.Equal(buf.Bytes(), data) { - t.Errorf("%s: decompress(compress(data)) != data (%d/%d)", tag, buf.Len(), len(data)) - t.FailNow() - } - } - }) - } -} - -// TestCopy will use io.Copy and avoid using Reader.WriteTo() and Writer.ReadFrom(). -func TestCopy(t *testing.T) { - for _, tdata := range testDataItems { - data := tdata.data - t.Run(tdata.label, func(t *testing.T) { - t.Parallel() - - w := lz4.NewWriter(nil) - r := lz4.NewReader(nil) - // test various options - for _, headerItem := range testHeaderItems { - tag := "io.Copy: " + tdata.label + ": " + headerItem.label - dbuf := &testBuffer{bytes.NewBuffer(data)} - - zbuf := bytes.NewBuffer(nil) - w.Reset(zbuf) - w.Header = headerItem.header - if _, err := io.Copy(w, dbuf); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - if err := w.Close(); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - buf := &testBuffer{bytes.NewBuffer(nil)} - r.Reset(zbuf) - if _, err := io.Copy(buf, r); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - if !bytes.Equal(buf.Bytes(), data) { - t.Errorf("%s: decompress(compress(data)) != data (%d/%d)", tag, buf.Len(), len(data)) - t.FailNow() - } - } - }) - } -} - -func TestSkippable(t *testing.T) { - w := lz4.NewWriter(nil) - r := lz4.NewReader(nil) - - skippable := make([]byte, 1<<20) - binary.LittleEndian.PutUint32(skippable, lz4.FrameSkipMagic) - binary.LittleEndian.PutUint32(skippable[4:], uint32(len(skippable)-8)) - - buf := make([]byte, len(lorem)) - - tag := "skippable first" - zbuf := bytes.NewBuffer(skippable) - w.Reset(zbuf) - w.Write(lorem) - w.Close() - - r.Reset(zbuf) - if _, err := r.Read(buf); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - tag = "skippable last" - zbuf = bytes.NewBuffer(nil) - w.Reset(zbuf) - w.Write(lorem) - w.Close() - zbuf.Write(skippable) - - r.Reset(zbuf) - if _, err := r.Read(buf); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - - tag = "skippable middle" - zbuf = bytes.NewBuffer(nil) - w.Reset(zbuf) - w.Write(lorem) - zbuf.Write(skippable) - w.Write(lorem) - w.Close() - - r.Reset(zbuf) - if _, err := r.Read(buf); err != nil { - t.Errorf("%s: unexpected error: %s", tag, err) - t.FailNow() - } - -} - -func TestWrittenCountAfterBufferedWrite(t *testing.T) { - w := lz4.NewWriter(bytes.NewBuffer(nil)) - w.Header.BlockDependency = true - - if n, _ := w.Write([]byte{1}); n != 1 { - t.Errorf("expected to write 1 byte, wrote %d", n) - t.FailNow() - } - - forcesWrite := make([]byte, 1<<16) - - if n, _ := w.Write(forcesWrite); n != len(forcesWrite) { - t.Errorf("expected to write %d bytes, wrote %d", len(forcesWrite), n) - t.FailNow() - } -} - -func TestWrittenBlocksExactlyWindowSize(t *testing.T) { - input := make([]byte, 128*1024) - - copy(input[64*1024-1:], []byte{1, 2, 3, 4, 1, 2, 3, 4}) - - output := writeReadChunked(t, input, 64*1024) - - if !bytes.Equal(input, output) { - t.Errorf("output is not equal to source input") - t.FailNow() - } -} - -func TestWrittenBlocksLessThanWindowSize(t *testing.T) { - input := make([]byte, 80*1024) - - copy(input[64*1024-1:], []byte{1, 2, 3, 4, 1, 2, 3, 4}) - copy(input[72*1024-1:], []byte{5, 6, 7, 8, 5, 6, 7, 8}) - - output := writeReadChunked(t, input, 8*1024) - if !bytes.Equal(input, output) { - t.Errorf("output is not equal to source input") - t.FailNow() - } -} - -func writeReadChunked(t *testing.T, in []byte, chunkSize int) []byte { - compressed := bytes.NewBuffer(nil) - w := lz4.NewWriter(compressed) - w.Header.BlockDependency = true - - buf := bytes.NewBuffer(in) - for buf.Len() > 0 { - _, err := w.Write(buf.Next(chunkSize)) - if err != nil { - t.Errorf("unexpected error: %v", err) - t.FailNow() - } - } - - r := lz4.NewReader(compressed) - out := make([]byte, len(in)) - _, err := io.ReadFull(r, out) - if err != nil { - t.Errorf("unexpected error: %v", err) - t.FailNow() - } - return out -} - -func TestMultiBlockWrite(t *testing.T) { - f, err := os.Open("testdata/207326ba-36f8-11e7-954a-aca46ba8ca73.png") - if err != nil { - t.Fatal(err) - } - defer f.Close() - - zbuf := bytes.NewBuffer(nil) - zw := lz4.NewWriter(zbuf) - if _, err := io.Copy(zw, f); err != nil { - t.Fatal(err) - } - if err := zw.Flush(); err != nil { - t.Fatal(err) - } - - buf := bytes.NewBuffer(nil) - zr := lz4.NewReader(zbuf) - if _, err := io.Copy(buf, zr); err != nil { - t.Fatal(err) - } -} diff --git a/vendor/github.com/pierrec/lz4/lz4c/main.go b/vendor/github.com/pierrec/lz4/lz4c/main.go index 048ab5004..001007589 100644 --- a/vendor/github.com/pierrec/lz4/lz4c/main.go +++ b/vendor/github.com/pierrec/lz4/lz4c/main.go @@ -14,19 +14,21 @@ import ( "strings" "github.com/pierrec/lz4" + "github.com/pkg/profile" ) func main() { // Process command line arguments var ( - blockMaxSizeDefault = 4 << 20 - flagStdout = flag.Bool("c", false, "output to stdout") - flagDecompress = flag.Bool("d", false, "decompress flag") - flagBlockMaxSize = flag.Int("B", blockMaxSizeDefault, "block max size [64Kb,256Kb,1Mb,4Mb]") - flagBlockDependency = flag.Bool("BD", false, "enable block dependency") - flagBlockChecksum = flag.Bool("BX", false, "enable block checksum") - flagStreamChecksum = flag.Bool("Sx", false, "disable stream checksum") - flagHighCompression = flag.Bool("9", false, "enabled high compression") + blockMaxSizeDefault = 4 << 20 + flagStdout = flag.Bool("c", false, "output to stdout") + flagDecompress = flag.Bool("d", false, "decompress flag") + flagBlockMaxSize = flag.Int("B", blockMaxSizeDefault, "block max size [64Kb,256Kb,1Mb,4Mb]") + flagBlockChecksum = flag.Bool("BX", false, "enable block checksum") + flagStreamChecksum = flag.Bool("Sx", false, "disable stream checksum") + flagCompressionLevel = flag.Int("l", 0, "compression level (0=fastest)") + profileName = flag.String("p", "", "path to the profile file") + mode = flag.String("profile.mode", "", "enable profiling mode, one of [cpu, mem, mutex, block]") ) flag.Usage = func() { fmt.Fprintf(os.Stderr, "Usage:\n\t%s [arg] [input]...\n\tNo input means [de]compress stdin to stdout\n\n", os.Args[0]) @@ -34,17 +36,31 @@ func main() { } flag.Parse() + ppath := profile.ProfilePath(*profileName) + switch *mode { + case "cpu": + defer profile.Start(profile.CPUProfile, ppath).Stop() + case "mem": + defer profile.Start(profile.MemProfile, ppath).Stop() + case "mutex": + defer profile.Start(profile.MutexProfile, ppath).Stop() + case "block": + defer profile.Start(profile.BlockProfile, ppath).Stop() + default: + // do nothing + } + flag.Parse() + // Use all CPUs runtime.GOMAXPROCS(runtime.NumCPU()) zr := lz4.NewReader(nil) zw := lz4.NewWriter(nil) zh := lz4.Header{ - BlockDependency: *flagBlockDependency, - BlockChecksum: *flagBlockChecksum, - BlockMaxSize: *flagBlockMaxSize, - NoChecksum: *flagStreamChecksum, - HighCompression: *flagHighCompression, + BlockChecksum: *flagBlockChecksum, + BlockMaxSize: *flagBlockMaxSize, + NoChecksum: *flagStreamChecksum, + CompressionLevel: *flagCompressionLevel, } worker := func(in io.Reader, out io.Writer) { diff --git a/vendor/github.com/pierrec/lz4/reader.go b/vendor/github.com/pierrec/lz4/reader.go index 9f7fd6042..c76567601 100644 --- a/vendor/github.com/pierrec/lz4/reader.go +++ b/vendor/github.com/pierrec/lz4/reader.go @@ -2,44 +2,33 @@ package lz4 import ( "encoding/binary" - "errors" "fmt" - "hash" "io" "io/ioutil" - "runtime" - "sync" - "sync/atomic" + + "github.com/pierrec/lz4/internal/xxh32" ) -// ErrInvalid is returned when the data being read is not an LZ4 archive -// (LZ4 magic number detection failed). -var ErrInvalid = errors.New("invalid lz4 data") - -// errEndOfBlock is returned by readBlock when it has reached the last block of the frame. -// It is not an error. -var errEndOfBlock = errors.New("end of block") - // Reader implements the LZ4 frame decoder. // The Header is set after the first call to Read(). // The Header may change between Read() calls in case of concatenated frames. type Reader struct { - Pos int64 // position within the source Header - src io.Reader - checksum hash.Hash32 // frame hash - wg sync.WaitGroup // decompressing go routine wait group - data []byte // buffered decompressed data - window []byte // 64Kb decompressed data window + + buf [8]byte // Scrap buffer. + pos int64 // Current position in src. + src io.Reader // Source. + zdata []byte // Compressed data. + data []byte // Uncompressed data. + idx int // Index of unread bytes into data. + checksum xxh32.XXHZero // Frame hash. } // NewReader returns a new LZ4 frame decoder. // No access to the underlying io.Reader is performed. func NewReader(src io.Reader) *Reader { - return &Reader{ - src: src, - checksum: hashPool.Get(), - } + r := &Reader{src: src} + return r } // readHeader checks the frame magic number and parses the frame descriptoz. @@ -48,88 +37,91 @@ func NewReader(src io.Reader) *Reader { func (z *Reader) readHeader(first bool) error { defer z.checksum.Reset() + buf := z.buf[:] for { - var magic uint32 - if err := binary.Read(z.src, binary.LittleEndian, &magic); err != nil { + magic, err := z.readUint32() + if err != nil { + z.pos += 4 if !first && err == io.ErrUnexpectedEOF { return io.EOF } return err } - z.Pos += 4 - if magic>>8 == frameSkipMagic>>8 { - var skipSize uint32 - if err := binary.Read(z.src, binary.LittleEndian, &skipSize); err != nil { - return err - } - z.Pos += 4 - m, err := io.CopyN(ioutil.Discard, z.src, int64(skipSize)) - z.Pos += m - if err != nil { - return err - } - continue + if magic == frameMagic { + break } - if magic != frameMagic { + if magic>>8 != frameSkipMagic>>8 { return ErrInvalid } - break + skipSize, err := z.readUint32() + if err != nil { + return err + } + z.pos += 4 + m, err := io.CopyN(ioutil.Discard, z.src, int64(skipSize)) + if err != nil { + return err + } + z.pos += m } - // header - var buf [8]byte + // Header. if _, err := io.ReadFull(z.src, buf[:2]); err != nil { return err } - z.Pos += 2 + z.pos += 8 b := buf[0] - if b>>6 != Version { - return fmt.Errorf("lz4.Read: invalid version: got %d expected %d", b>>6, Version) + if v := b >> 6; v != Version { + return fmt.Errorf("lz4: invalid version: got %d; expected %d", v, Version) + } + if b>>5&1 == 0 { + return fmt.Errorf("lz4: block dependency not supported") } - z.BlockDependency = b>>5&1 == 0 z.BlockChecksum = b>>4&1 > 0 frameSize := b>>3&1 > 0 z.NoChecksum = b>>2&1 == 0 - // z.Dict = b&1 > 0 bmsID := buf[1] >> 4 & 0x7 bSize, ok := bsMapID[bmsID] if !ok { - return fmt.Errorf("lz4.Read: invalid block max size: %d", bmsID) + return fmt.Errorf("lz4: invalid block max size ID: %d", bmsID) } z.BlockMaxSize = bSize + // Allocate the compressed/uncompressed buffers. + // The compressed buffer cannot exceed the uncompressed one. + if n := 2 * bSize; cap(z.zdata) < n { + z.zdata = make([]byte, n, n) + } + debug("header block max size id=%d size=%d", bmsID, bSize) + z.zdata = z.zdata[:bSize] + z.data = z.zdata[:cap(z.zdata)][bSize:] + z.idx = len(z.data) + z.checksum.Write(buf[0:2]) if frameSize { - if err := binary.Read(z.src, binary.LittleEndian, &z.Size); err != nil { + buf := buf[:8] + if _, err := io.ReadFull(z.src, buf); err != nil { return err } - z.Pos += 8 - binary.LittleEndian.PutUint64(buf[:], z.Size) - z.checksum.Write(buf[0:8]) + z.Size = binary.LittleEndian.Uint64(buf) + z.pos += 8 + z.checksum.Write(buf) } - // if z.Dict { - // if err := binary.Read(z.src, binary.LittleEndian, &z.DictID); err != nil { - // return err - // } - // z.Pos += 4 - // binary.LittleEndian.PutUint32(buf[:], z.DictID) - // z.checksum.Write(buf[0:4]) - // } - - // header checksum + // Header checksum. if _, err := io.ReadFull(z.src, buf[:1]); err != nil { return err } - z.Pos++ + z.pos++ if h := byte(z.checksum.Sum32() >> 8 & 0xFF); h != buf[0] { - return fmt.Errorf("lz4.Read: invalid header checksum: got %v expected %v", buf[0], h) + return fmt.Errorf("lz4: invalid header checksum: got %x; expected %x", buf[0], h) } z.Header.done = true + debug("header read: %v", z.Header) return nil } @@ -139,177 +131,123 @@ func (z *Reader) readHeader(first bool) error { // Since there can be multiple streams concatenated, Header values may // change between calls to Read(). If that is the case, no data is actually read from // the underlying io.Reader, to allow for potential input buffer resizing. -// -// Data is buffered if the input buffer is too small, and exhausted upon successive calls. -// -// If the buffer is large enough (typically in multiples of BlockMaxSize) and there is -// no block dependency, then the data will be decompressed concurrently based on the GOMAXPROCS value. -func (z *Reader) Read(buf []byte) (n int, err error) { +func (z *Reader) Read(buf []byte) (int, error) { + debug("Read buf len=%d", len(buf)) if !z.Header.done { - if err = z.readHeader(true); err != nil { - return + if err := z.readHeader(true); err != nil { + return 0, err } + debug("header read OK compressed buffer %d / %d uncompressed buffer %d : %d index=%d", + len(z.zdata), cap(z.zdata), len(z.data), cap(z.data), z.idx) } if len(buf) == 0 { - return + return 0, nil } - // exhaust remaining data from previous Read() - if len(z.data) > 0 { - n = copy(buf, z.data) - z.data = z.data[n:] - if len(z.data) == 0 { - z.data = nil - } - return - } - - // Break up the input buffer into BlockMaxSize blocks with at least one block. - // Then decompress into each of them concurrently if possible (no dependency). - // In case of dependency, the first block will be missing the window (except on the - // very first call), the rest will have it already since it comes from the previous block. - wbuf := buf - zn := (len(wbuf) + z.BlockMaxSize - 1) / z.BlockMaxSize - zblocks := make([]block, zn) - for zi, abort := 0, uint32(0); zi < zn && atomic.LoadUint32(&abort) == 0; zi++ { - zb := &zblocks[zi] - // last block may be too small - if len(wbuf) < z.BlockMaxSize+len(z.window) { - wbuf = make([]byte, z.BlockMaxSize+len(z.window)) - } - copy(wbuf, z.window) - if zb.err = z.readBlock(wbuf, zb); zb.err != nil { - break - } - wbuf = wbuf[z.BlockMaxSize:] - if !z.BlockDependency { - z.wg.Add(1) - go z.decompressBlock(zb, &abort) - continue - } - // cannot decompress concurrently when dealing with block dependency - z.decompressBlock(zb, nil) - // the last block may not contain enough data - if len(z.window) == 0 { - z.window = make([]byte, winSize) - } - if len(zb.data) >= winSize { - copy(z.window, zb.data[len(zb.data)-winSize:]) - } else { - copy(z.window, z.window[len(zb.data):]) - copy(z.window[len(zb.data)+1:], zb.data) - } - } - z.wg.Wait() - - // since a block size may be less then BlockMaxSize, trim the decompressed buffers - for _, zb := range zblocks { - if zb.err != nil { - if zb.err == errEndOfBlock { - return n, z.close() - } - return n, zb.err - } - bLen := len(zb.data) - if !z.NoChecksum { - z.checksum.Write(zb.data) - } - m := copy(buf[n:], zb.data) - // buffer the remaining data (this is necessarily the last block) - if m < bLen { - z.data = zb.data[m:] - } - n += m - } - - return -} - -// readBlock reads an entire frame block from the frame. -// The input buffer is the one that will receive the decompressed data. -// If the end of the frame is detected, it returns the errEndOfBlock error. -func (z *Reader) readBlock(buf []byte, b *block) error { - var bLen uint32 - if err := binary.Read(z.src, binary.LittleEndian, &bLen); err != nil { - return err - } - atomic.AddInt64(&z.Pos, 4) - - switch { - case bLen == 0: - return errEndOfBlock - case bLen&(1<<31) == 0: - b.compressed = true - b.data = buf - b.zdata = make([]byte, bLen) - default: - bLen = bLen & (1<<31 - 1) - if int(bLen) > len(buf) { - return fmt.Errorf("lz4.Read: invalid block size: %d", bLen) - } - b.data = buf[:bLen] - b.zdata = buf[:bLen] - } - if _, err := io.ReadFull(z.src, b.zdata); err != nil { - return err - } - - if z.BlockChecksum { - if err := binary.Read(z.src, binary.LittleEndian, &b.checksum); err != nil { - return err - } - xxh := hashPool.Get() - defer hashPool.Put(xxh) - xxh.Write(b.zdata) - if h := xxh.Sum32(); h != b.checksum { - return fmt.Errorf("lz4.Read: invalid block checksum: got %x expected %x", h, b.checksum) - } - } - - return nil -} - -// decompressBlock decompresses a frame block. -// In case of an error, the block err is set with it and abort is set to 1. -func (z *Reader) decompressBlock(b *block, abort *uint32) { - if abort != nil { - defer z.wg.Done() - } - if b.compressed { - n := len(z.window) - m, err := UncompressBlock(b.zdata, b.data, n) + if z.idx == len(z.data) { + // No data ready for reading, process the next block. + debug("reading block from writer") + // Block length: 0 = end of frame, highest bit set: uncompressed. + bLen, err := z.readUint32() if err != nil { - if abort != nil { - atomic.StoreUint32(abort, 1) + return 0, err + } + z.pos += 4 + + if bLen == 0 { + // End of frame reached. + if !z.NoChecksum { + // Validate the frame checksum. + checksum, err := z.readUint32() + if err != nil { + return 0, err + } + debug("frame checksum got=%x / want=%x", z.checksum.Sum32(), checksum) + z.pos += 4 + if h := z.checksum.Sum32(); checksum != h { + return 0, fmt.Errorf("lz4: invalid frame checksum: got %x; expected %x", h, checksum) + } } - b.err = err - return - } - b.data = b.data[n : n+m] - } - atomic.AddInt64(&z.Pos, int64(len(b.data))) -} -// close validates the frame checksum (if any) and checks the next frame (if any). -func (z *Reader) close() error { - if !z.NoChecksum { - var checksum uint32 - if err := binary.Read(z.src, binary.LittleEndian, &checksum); err != nil { - return err + // Get ready for the next concatenated frame and keep the position. + pos := z.pos + z.Reset(z.src) + z.pos = pos + + // Since multiple frames can be concatenated, check for more. + return 0, z.readHeader(false) } - if checksum != z.checksum.Sum32() { - return fmt.Errorf("lz4.Read: invalid frame checksum: got %x expected %x", z.checksum.Sum32(), checksum) + + debug("raw block size %d", bLen) + if bLen&compressedBlockFlag > 0 { + // Uncompressed block. + bLen &= compressedBlockMask + debug("uncompressed block size %d", bLen) + if int(bLen) > cap(z.data) { + return 0, fmt.Errorf("lz4: invalid block size: %d", bLen) + } + z.data = z.data[:bLen] + if _, err := io.ReadFull(z.src, z.data); err != nil { + return 0, err + } + z.pos += int64(bLen) + + if z.BlockChecksum { + checksum, err := z.readUint32() + if err != nil { + return 0, err + } + z.pos += 4 + + if h := xxh32.ChecksumZero(z.data); h != checksum { + return 0, fmt.Errorf("lz4: invalid block checksum: got %x; expected %x", h, checksum) + } + } + + } else { + // Compressed block. + debug("compressed block size %d", bLen) + if int(bLen) > cap(z.data) { + return 0, fmt.Errorf("lz4: invalid block size: %d", bLen) + } + zdata := z.zdata[:bLen] + if _, err := io.ReadFull(z.src, zdata); err != nil { + return 0, err + } + z.pos += int64(bLen) + + if z.BlockChecksum { + checksum, err := z.readUint32() + if err != nil { + return 0, err + } + z.pos += 4 + + if h := xxh32.ChecksumZero(zdata); h != checksum { + return 0, fmt.Errorf("lz4: invalid block checksum: got %x; expected %x", h, checksum) + } + } + + n, err := UncompressBlock(zdata, z.data) + if err != nil { + return 0, err + } + z.data = z.data[:n] } + + if !z.NoChecksum { + z.checksum.Write(z.data) + debug("current frame checksum %x", z.checksum.Sum32()) + } + z.idx = 0 } - // get ready for the next concatenated frame, but do not change the position - pos := z.Pos - z.Reset(z.src) - z.Pos = pos + n := copy(buf, z.data[z.idx:]) + z.idx += n + debug("copied %d bytes to input", n) - // since multiple frames can be concatenated, check for another one - return z.readHeader(false) + return n, nil } // Reset discards the Reader's state and makes it equivalent to the @@ -317,48 +255,19 @@ func (z *Reader) close() error { // This permits reusing a Reader rather than allocating a new one. func (z *Reader) Reset(r io.Reader) { z.Header = Header{} - z.Pos = 0 + z.pos = 0 z.src = r + z.zdata = z.zdata[:0] + z.data = z.data[:0] + z.idx = 0 z.checksum.Reset() - z.data = nil - z.window = nil } -// WriteTo decompresses the data from the underlying io.Reader and writes it to the io.Writer. -// Returns the number of bytes written. -func (z *Reader) WriteTo(w io.Writer) (n int64, err error) { - cpus := runtime.GOMAXPROCS(0) - var buf []byte - - // The initial buffer being nil, the first Read will be only read the compressed frame options. - // The buffer can then be sized appropriately to support maximum concurrency decompression. - // If multiple frames are concatenated, Read() will return with no data decompressed but with - // potentially changed options. The buffer will be resized accordingly, always trying to - // maximize concurrency. - for { - nsize := 0 - // the block max size can change if multiple streams are concatenated. - // Check it after every Read(). - if z.BlockDependency { - // in case of dependency, we cannot decompress concurrently, - // so allocate the minimum buffer + window size - nsize = len(z.window) + z.BlockMaxSize - } else { - // if no dependency, allocate a buffer large enough for concurrent decompression - nsize = cpus * z.BlockMaxSize - } - if nsize != len(buf) { - buf = make([]byte, nsize) - } - - m, er := z.Read(buf) - if er != nil && er != io.EOF { - return n, er - } - m, err = w.Write(buf[:m]) - n += int64(m) - if err != nil || er == io.EOF { - return - } - } +// readUint32 reads an uint32 into the supplied buffer. +// The idea is to make use of the already allocated buffers avoiding additional allocations. +func (z *Reader) readUint32() (uint32, error) { + buf := z.buf[:4] + _, err := io.ReadFull(z.src, buf) + x := binary.LittleEndian.Uint32(buf) + return x, err } diff --git a/vendor/github.com/pierrec/lz4/reader_test.go b/vendor/github.com/pierrec/lz4/reader_test.go new file mode 100644 index 000000000..9165b51f7 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/reader_test.go @@ -0,0 +1,59 @@ +package lz4_test + +import ( + "bytes" + "io" + "io/ioutil" + "os" + "reflect" + "strings" + "testing" + + "github.com/pierrec/lz4" +) + +func TestReader(t *testing.T) { + goldenFiles := []string{ + "testdata/e.txt.lz4", + "testdata/gettysburg.txt.lz4", + "testdata/Mark.Twain-Tom.Sawyer.txt.lz4", + "testdata/pg1661.txt.lz4", + "testdata/pi.txt.lz4", + "testdata/random.data.lz4", + "testdata/repeat.txt.lz4", + } + + for _, fname := range goldenFiles { + t.Run(fname, func(t *testing.T) { + fname := fname + t.Parallel() + + f, err := os.Open(fname) + if err != nil { + t.Fatal(err) + } + defer f.Close() + + rawfile := strings.TrimSuffix(fname, ".lz4") + raw, err := ioutil.ReadFile(rawfile) + if err != nil { + t.Fatal(err) + } + + var out bytes.Buffer + zr := lz4.NewReader(f) + n, err := io.Copy(&out, zr) + if err != nil { + t.Fatal(err) + } + + if got, want := int(n), len(raw); got != want { + t.Errorf("invalid sizes: got %d; want %d", got, want) + } + + if got, want := out.Bytes(), raw; !reflect.DeepEqual(got, want) { + t.Fatal("uncompressed data does not match original") + } + }) + } +} diff --git a/vendor/github.com/pierrec/lz4/testdata/207326ba-36f8-11e7-954a-aca46ba8ca73.png b/vendor/github.com/pierrec/lz4/testdata/207326ba-36f8-11e7-954a-aca46ba8ca73.png deleted file mode 100644 index 9489dfa61..000000000 Binary files a/vendor/github.com/pierrec/lz4/testdata/207326ba-36f8-11e7-954a-aca46ba8ca73.png and /dev/null differ diff --git a/vendor/github.com/pierrec/lz4/testdata/Mark.Twain-Tom.Sawyer.txt b/vendor/github.com/pierrec/lz4/testdata/Mark.Twain-Tom.Sawyer.txt new file mode 100644 index 000000000..c9106fd52 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/Mark.Twain-Tom.Sawyer.txt @@ -0,0 +1,8465 @@ +Produced by David Widger. The previous edition was updated by Jose +Menendez. + + + + + + THE ADVENTURES OF TOM SAWYER + BY + MARK TWAIN + (Samuel Langhorne Clemens) + + + + + P R E F A C E + +MOST of the adventures recorded in this book really occurred; one or +two were experiences of my own, the rest those of boys who were +schoolmates of mine. Huck Finn is drawn from life; Tom Sawyer also, but +not from an individual--he is a combination of the characteristics of +three boys whom I knew, and therefore belongs to the composite order of +architecture. + +The odd superstitions touched upon were all prevalent among children +and slaves in the West at the period of this story--that is to say, +thirty or forty years ago. + +Although my book is intended mainly for the entertainment of boys and +girls, I hope it will not be shunned by men and women on that account, +for part of my plan has been to try to pleasantly remind adults of what +they once were themselves, and of how they felt and thought and talked, +and what queer enterprises they sometimes engaged in. + + THE AUTHOR. + +HARTFORD, 1876. + + + + T O M S A W Y E R + + + +CHAPTER I + +"TOM!" + +No answer. + +"TOM!" + +No answer. + +"What's gone with that boy, I wonder? You TOM!" + +No answer. + +The old lady pulled her spectacles down and looked over them about the +room; then she put them up and looked out under them. She seldom or +never looked THROUGH them for so small a thing as a boy; they were her +state pair, the pride of her heart, and were built for "style," not +service--she could have seen through a pair of stove-lids just as well. +She looked perplexed for a moment, and then said, not fiercely, but +still loud enough for the furniture to hear: + +"Well, I lay if I get hold of you I'll--" + +She did not finish, for by this time she was bending down and punching +under the bed with the broom, and so she needed breath to punctuate the +punches with. She resurrected nothing but the cat. + +"I never did see the beat of that boy!" + +She went to the open door and stood in it and looked out among the +tomato vines and "jimpson" weeds that constituted the garden. No Tom. +So she lifted up her voice at an angle calculated for distance and +shouted: + +"Y-o-u-u TOM!" + +There was a slight noise behind her and she turned just in time to +seize a small boy by the slack of his roundabout and arrest his flight. + +"There! I might 'a' thought of that closet. What you been doing in +there?" + +"Nothing." + +"Nothing! Look at your hands. And look at your mouth. What IS that +truck?" + +"I don't know, aunt." + +"Well, I know. It's jam--that's what it is. Forty times I've said if +you didn't let that jam alone I'd skin you. Hand me that switch." + +The switch hovered in the air--the peril was desperate-- + +"My! Look behind you, aunt!" + +The old lady whirled round, and snatched her skirts out of danger. The +lad fled on the instant, scrambled up the high board-fence, and +disappeared over it. + +His aunt Polly stood surprised a moment, and then broke into a gentle +laugh. + +"Hang the boy, can't I never learn anything? Ain't he played me tricks +enough like that for me to be looking out for him by this time? But old +fools is the biggest fools there is. Can't learn an old dog new tricks, +as the saying is. But my goodness, he never plays them alike, two days, +and how is a body to know what's coming? He 'pears to know just how +long he can torment me before I get my dander up, and he knows if he +can make out to put me off for a minute or make me laugh, it's all down +again and I can't hit him a lick. I ain't doing my duty by that boy, +and that's the Lord's truth, goodness knows. Spare the rod and spile +the child, as the Good Book says. I'm a laying up sin and suffering for +us both, I know. He's full of the Old Scratch, but laws-a-me! he's my +own dead sister's boy, poor thing, and I ain't got the heart to lash +him, somehow. Every time I let him off, my conscience does hurt me so, +and every time I hit him my old heart most breaks. Well-a-well, man +that is born of woman is of few days and full of trouble, as the +Scripture says, and I reckon it's so. He'll play hookey this evening, * +and [* Southwestern for "afternoon"] I'll just be obleeged to make him +work, to-morrow, to punish him. It's mighty hard to make him work +Saturdays, when all the boys is having holiday, but he hates work more +than he hates anything else, and I've GOT to do some of my duty by him, +or I'll be the ruination of the child." + +Tom did play hookey, and he had a very good time. He got back home +barely in season to help Jim, the small colored boy, saw next-day's +wood and split the kindlings before supper--at least he was there in +time to tell his adventures to Jim while Jim did three-fourths of the +work. Tom's younger brother (or rather half-brother) Sid was already +through with his part of the work (picking up chips), for he was a +quiet boy, and had no adventurous, troublesome ways. + +While Tom was eating his supper, and stealing sugar as opportunity +offered, Aunt Polly asked him questions that were full of guile, and +very deep--for she wanted to trap him into damaging revealments. Like +many other simple-hearted souls, it was her pet vanity to believe she +was endowed with a talent for dark and mysterious diplomacy, and she +loved to contemplate her most transparent devices as marvels of low +cunning. Said she: + +"Tom, it was middling warm in school, warn't it?" + +"Yes'm." + +"Powerful warm, warn't it?" + +"Yes'm." + +"Didn't you want to go in a-swimming, Tom?" + +A bit of a scare shot through Tom--a touch of uncomfortable suspicion. +He searched Aunt Polly's face, but it told him nothing. So he said: + +"No'm--well, not very much." + +The old lady reached out her hand and felt Tom's shirt, and said: + +"But you ain't too warm now, though." And it flattered her to reflect +that she had discovered that the shirt was dry without anybody knowing +that that was what she had in her mind. But in spite of her, Tom knew +where the wind lay, now. So he forestalled what might be the next move: + +"Some of us pumped on our heads--mine's damp yet. See?" + +Aunt Polly was vexed to think she had overlooked that bit of +circumstantial evidence, and missed a trick. Then she had a new +inspiration: + +"Tom, you didn't have to undo your shirt collar where I sewed it, to +pump on your head, did you? Unbutton your jacket!" + +The trouble vanished out of Tom's face. He opened his jacket. His +shirt collar was securely sewed. + +"Bother! Well, go 'long with you. I'd made sure you'd played hookey +and been a-swimming. But I forgive ye, Tom. I reckon you're a kind of a +singed cat, as the saying is--better'n you look. THIS time." + +She was half sorry her sagacity had miscarried, and half glad that Tom +had stumbled into obedient conduct for once. + +But Sidney said: + +"Well, now, if I didn't think you sewed his collar with white thread, +but it's black." + +"Why, I did sew it with white! Tom!" + +But Tom did not wait for the rest. As he went out at the door he said: + +"Siddy, I'll lick you for that." + +In a safe place Tom examined two large needles which were thrust into +the lapels of his jacket, and had thread bound about them--one needle +carried white thread and the other black. He said: + +"She'd never noticed if it hadn't been for Sid. Confound it! sometimes +she sews it with white, and sometimes she sews it with black. I wish to +geeminy she'd stick to one or t'other--I can't keep the run of 'em. But +I bet you I'll lam Sid for that. I'll learn him!" + +He was not the Model Boy of the village. He knew the model boy very +well though--and loathed him. + +Within two minutes, or even less, he had forgotten all his troubles. +Not because his troubles were one whit less heavy and bitter to him +than a man's are to a man, but because a new and powerful interest bore +them down and drove them out of his mind for the time--just as men's +misfortunes are forgotten in the excitement of new enterprises. This +new interest was a valued novelty in whistling, which he had just +acquired from a negro, and he was suffering to practise it undisturbed. +It consisted in a peculiar bird-like turn, a sort of liquid warble, +produced by touching the tongue to the roof of the mouth at short +intervals in the midst of the music--the reader probably remembers how +to do it, if he has ever been a boy. Diligence and attention soon gave +him the knack of it, and he strode down the street with his mouth full +of harmony and his soul full of gratitude. He felt much as an +astronomer feels who has discovered a new planet--no doubt, as far as +strong, deep, unalloyed pleasure is concerned, the advantage was with +the boy, not the astronomer. + +The summer evenings were long. It was not dark, yet. Presently Tom +checked his whistle. A stranger was before him--a boy a shade larger +than himself. A new-comer of any age or either sex was an impressive +curiosity in the poor little shabby village of St. Petersburg. This boy +was well dressed, too--well dressed on a week-day. This was simply +astounding. His cap was a dainty thing, his close-buttoned blue cloth +roundabout was new and natty, and so were his pantaloons. He had shoes +on--and it was only Friday. He even wore a necktie, a bright bit of +ribbon. He had a citified air about him that ate into Tom's vitals. The +more Tom stared at the splendid marvel, the higher he turned up his +nose at his finery and the shabbier and shabbier his own outfit seemed +to him to grow. Neither boy spoke. If one moved, the other moved--but +only sidewise, in a circle; they kept face to face and eye to eye all +the time. Finally Tom said: + +"I can lick you!" + +"I'd like to see you try it." + +"Well, I can do it." + +"No you can't, either." + +"Yes I can." + +"No you can't." + +"I can." + +"You can't." + +"Can!" + +"Can't!" + +An uncomfortable pause. Then Tom said: + +"What's your name?" + +"'Tisn't any of your business, maybe." + +"Well I 'low I'll MAKE it my business." + +"Well why don't you?" + +"If you say much, I will." + +"Much--much--MUCH. There now." + +"Oh, you think you're mighty smart, DON'T you? I could lick you with +one hand tied behind me, if I wanted to." + +"Well why don't you DO it? You SAY you can do it." + +"Well I WILL, if you fool with me." + +"Oh yes--I've seen whole families in the same fix." + +"Smarty! You think you're SOME, now, DON'T you? Oh, what a hat!" + +"You can lump that hat if you don't like it. I dare you to knock it +off--and anybody that'll take a dare will suck eggs." + +"You're a liar!" + +"You're another." + +"You're a fighting liar and dasn't take it up." + +"Aw--take a walk!" + +"Say--if you give me much more of your sass I'll take and bounce a +rock off'n your head." + +"Oh, of COURSE you will." + +"Well I WILL." + +"Well why don't you DO it then? What do you keep SAYING you will for? +Why don't you DO it? It's because you're afraid." + +"I AIN'T afraid." + +"You are." + +"I ain't." + +"You are." + +Another pause, and more eying and sidling around each other. Presently +they were shoulder to shoulder. Tom said: + +"Get away from here!" + +"Go away yourself!" + +"I won't." + +"I won't either." + +So they stood, each with a foot placed at an angle as a brace, and +both shoving with might and main, and glowering at each other with +hate. But neither could get an advantage. After struggling till both +were hot and flushed, each relaxed his strain with watchful caution, +and Tom said: + +"You're a coward and a pup. I'll tell my big brother on you, and he +can thrash you with his little finger, and I'll make him do it, too." + +"What do I care for your big brother? I've got a brother that's bigger +than he is--and what's more, he can throw him over that fence, too." +[Both brothers were imaginary.] + +"That's a lie." + +"YOUR saying so don't make it so." + +Tom drew a line in the dust with his big toe, and said: + +"I dare you to step over that, and I'll lick you till you can't stand +up. Anybody that'll take a dare will steal sheep." + +The new boy stepped over promptly, and said: + +"Now you said you'd do it, now let's see you do it." + +"Don't you crowd me now; you better look out." + +"Well, you SAID you'd do it--why don't you do it?" + +"By jingo! for two cents I WILL do it." + +The new boy took two broad coppers out of his pocket and held them out +with derision. Tom struck them to the ground. In an instant both boys +were rolling and tumbling in the dirt, gripped together like cats; and +for the space of a minute they tugged and tore at each other's hair and +clothes, punched and scratched each other's nose, and covered +themselves with dust and glory. Presently the confusion took form, and +through the fog of battle Tom appeared, seated astride the new boy, and +pounding him with his fists. "Holler 'nuff!" said he. + +The boy only struggled to free himself. He was crying--mainly from rage. + +"Holler 'nuff!"--and the pounding went on. + +At last the stranger got out a smothered "'Nuff!" and Tom let him up +and said: + +"Now that'll learn you. Better look out who you're fooling with next +time." + +The new boy went off brushing the dust from his clothes, sobbing, +snuffling, and occasionally looking back and shaking his head and +threatening what he would do to Tom the "next time he caught him out." +To which Tom responded with jeers, and started off in high feather, and +as soon as his back was turned the new boy snatched up a stone, threw +it and hit him between the shoulders and then turned tail and ran like +an antelope. Tom chased the traitor home, and thus found out where he +lived. He then held a position at the gate for some time, daring the +enemy to come outside, but the enemy only made faces at him through the +window and declined. At last the enemy's mother appeared, and called +Tom a bad, vicious, vulgar child, and ordered him away. So he went +away; but he said he "'lowed" to "lay" for that boy. + +He got home pretty late that night, and when he climbed cautiously in +at the window, he uncovered an ambuscade, in the person of his aunt; +and when she saw the state his clothes were in her resolution to turn +his Saturday holiday into captivity at hard labor became adamantine in +its firmness. + + + +CHAPTER II + +SATURDAY morning was come, and all the summer world was bright and +fresh, and brimming with life. There was a song in every heart; and if +the heart was young the music issued at the lips. There was cheer in +every face and a spring in every step. The locust-trees were in bloom +and the fragrance of the blossoms filled the air. Cardiff Hill, beyond +the village and above it, was green with vegetation and it lay just far +enough away to seem a Delectable Land, dreamy, reposeful, and inviting. + +Tom appeared on the sidewalk with a bucket of whitewash and a +long-handled brush. He surveyed the fence, and all gladness left him and +a deep melancholy settled down upon his spirit. Thirty yards of board +fence nine feet high. Life to him seemed hollow, and existence but a +burden. Sighing, he dipped his brush and passed it along the topmost +plank; repeated the operation; did it again; compared the insignificant +whitewashed streak with the far-reaching continent of unwhitewashed +fence, and sat down on a tree-box discouraged. Jim came skipping out at +the gate with a tin pail, and singing Buffalo Gals. Bringing water from +the town pump had always been hateful work in Tom's eyes, before, but +now it did not strike him so. He remembered that there was company at +the pump. White, mulatto, and negro boys and girls were always there +waiting their turns, resting, trading playthings, quarrelling, +fighting, skylarking. And he remembered that although the pump was only +a hundred and fifty yards off, Jim never got back with a bucket of +water under an hour--and even then somebody generally had to go after +him. Tom said: + +"Say, Jim, I'll fetch the water if you'll whitewash some." + +Jim shook his head and said: + +"Can't, Mars Tom. Ole missis, she tole me I got to go an' git dis +water an' not stop foolin' roun' wid anybody. She say she spec' Mars +Tom gwine to ax me to whitewash, an' so she tole me go 'long an' 'tend +to my own business--she 'lowed SHE'D 'tend to de whitewashin'." + +"Oh, never you mind what she said, Jim. That's the way she always +talks. Gimme the bucket--I won't be gone only a a minute. SHE won't +ever know." + +"Oh, I dasn't, Mars Tom. Ole missis she'd take an' tar de head off'n +me. 'Deed she would." + +"SHE! She never licks anybody--whacks 'em over the head with her +thimble--and who cares for that, I'd like to know. She talks awful, but +talk don't hurt--anyways it don't if she don't cry. Jim, I'll give you +a marvel. I'll give you a white alley!" + +Jim began to waver. + +"White alley, Jim! And it's a bully taw." + +"My! Dat's a mighty gay marvel, I tell you! But Mars Tom I's powerful +'fraid ole missis--" + +"And besides, if you will I'll show you my sore toe." + +Jim was only human--this attraction was too much for him. He put down +his pail, took the white alley, and bent over the toe with absorbing +interest while the bandage was being unwound. In another moment he was +flying down the street with his pail and a tingling rear, Tom was +whitewashing with vigor, and Aunt Polly was retiring from the field +with a slipper in her hand and triumph in her eye. + +But Tom's energy did not last. He began to think of the fun he had +planned for this day, and his sorrows multiplied. Soon the free boys +would come tripping along on all sorts of delicious expeditions, and +they would make a world of fun of him for having to work--the very +thought of it burnt him like fire. He got out his worldly wealth and +examined it--bits of toys, marbles, and trash; enough to buy an +exchange of WORK, maybe, but not half enough to buy so much as half an +hour of pure freedom. So he returned his straitened means to his +pocket, and gave up the idea of trying to buy the boys. At this dark +and hopeless moment an inspiration burst upon him! Nothing less than a +great, magnificent inspiration. + +He took up his brush and went tranquilly to work. Ben Rogers hove in +sight presently--the very boy, of all boys, whose ridicule he had been +dreading. Ben's gait was the hop-skip-and-jump--proof enough that his +heart was light and his anticipations high. He was eating an apple, and +giving a long, melodious whoop, at intervals, followed by a deep-toned +ding-dong-dong, ding-dong-dong, for he was personating a steamboat. As +he drew near, he slackened speed, took the middle of the street, leaned +far over to starboard and rounded to ponderously and with laborious +pomp and circumstance--for he was personating the Big Missouri, and +considered himself to be drawing nine feet of water. He was boat and +captain and engine-bells combined, so he had to imagine himself +standing on his own hurricane-deck giving the orders and executing them: + +"Stop her, sir! Ting-a-ling-ling!" The headway ran almost out, and he +drew up slowly toward the sidewalk. + +"Ship up to back! Ting-a-ling-ling!" His arms straightened and +stiffened down his sides. + +"Set her back on the stabboard! Ting-a-ling-ling! Chow! ch-chow-wow! +Chow!" His right hand, meantime, describing stately circles--for it was +representing a forty-foot wheel. + +"Let her go back on the labboard! Ting-a-lingling! Chow-ch-chow-chow!" +The left hand began to describe circles. + +"Stop the stabboard! Ting-a-ling-ling! Stop the labboard! Come ahead +on the stabboard! Stop her! Let your outside turn over slow! +Ting-a-ling-ling! Chow-ow-ow! Get out that head-line! LIVELY now! +Come--out with your spring-line--what're you about there! Take a turn +round that stump with the bight of it! Stand by that stage, now--let her +go! Done with the engines, sir! Ting-a-ling-ling! SH'T! S'H'T! SH'T!" +(trying the gauge-cocks). + +Tom went on whitewashing--paid no attention to the steamboat. Ben +stared a moment and then said: "Hi-YI! YOU'RE up a stump, ain't you!" + +No answer. Tom surveyed his last touch with the eye of an artist, then +he gave his brush another gentle sweep and surveyed the result, as +before. Ben ranged up alongside of him. Tom's mouth watered for the +apple, but he stuck to his work. Ben said: + +"Hello, old chap, you got to work, hey?" + +Tom wheeled suddenly and said: + +"Why, it's you, Ben! I warn't noticing." + +"Say--I'm going in a-swimming, I am. Don't you wish you could? But of +course you'd druther WORK--wouldn't you? Course you would!" + +Tom contemplated the boy a bit, and said: + +"What do you call work?" + +"Why, ain't THAT work?" + +Tom resumed his whitewashing, and answered carelessly: + +"Well, maybe it is, and maybe it ain't. All I know, is, it suits Tom +Sawyer." + +"Oh come, now, you don't mean to let on that you LIKE it?" + +The brush continued to move. + +"Like it? Well, I don't see why I oughtn't to like it. Does a boy get +a chance to whitewash a fence every day?" + +That put the thing in a new light. Ben stopped nibbling his apple. Tom +swept his brush daintily back and forth--stepped back to note the +effect--added a touch here and there--criticised the effect again--Ben +watching every move and getting more and more interested, more and more +absorbed. Presently he said: + +"Say, Tom, let ME whitewash a little." + +Tom considered, was about to consent; but he altered his mind: + +"No--no--I reckon it wouldn't hardly do, Ben. You see, Aunt Polly's +awful particular about this fence--right here on the street, you know +--but if it was the back fence I wouldn't mind and SHE wouldn't. Yes, +she's awful particular about this fence; it's got to be done very +careful; I reckon there ain't one boy in a thousand, maybe two +thousand, that can do it the way it's got to be done." + +"No--is that so? Oh come, now--lemme just try. Only just a little--I'd +let YOU, if you was me, Tom." + +"Ben, I'd like to, honest injun; but Aunt Polly--well, Jim wanted to +do it, but she wouldn't let him; Sid wanted to do it, and she wouldn't +let Sid. Now don't you see how I'm fixed? If you was to tackle this +fence and anything was to happen to it--" + +"Oh, shucks, I'll be just as careful. Now lemme try. Say--I'll give +you the core of my apple." + +"Well, here--No, Ben, now don't. I'm afeard--" + +"I'll give you ALL of it!" + +Tom gave up the brush with reluctance in his face, but alacrity in his +heart. And while the late steamer Big Missouri worked and sweated in +the sun, the retired artist sat on a barrel in the shade close by, +dangled his legs, munched his apple, and planned the slaughter of more +innocents. There was no lack of material; boys happened along every +little while; they came to jeer, but remained to whitewash. By the time +Ben was fagged out, Tom had traded the next chance to Billy Fisher for +a kite, in good repair; and when he played out, Johnny Miller bought in +for a dead rat and a string to swing it with--and so on, and so on, +hour after hour. And when the middle of the afternoon came, from being +a poor poverty-stricken boy in the morning, Tom was literally rolling +in wealth. He had besides the things before mentioned, twelve marbles, +part of a jews-harp, a piece of blue bottle-glass to look through, a +spool cannon, a key that wouldn't unlock anything, a fragment of chalk, +a glass stopper of a decanter, a tin soldier, a couple of tadpoles, six +fire-crackers, a kitten with only one eye, a brass doorknob, a +dog-collar--but no dog--the handle of a knife, four pieces of +orange-peel, and a dilapidated old window sash. + +He had had a nice, good, idle time all the while--plenty of company +--and the fence had three coats of whitewash on it! If he hadn't run out +of whitewash he would have bankrupted every boy in the village. + +Tom said to himself that it was not such a hollow world, after all. He +had discovered a great law of human action, without knowing it--namely, +that in order to make a man or a boy covet a thing, it is only +necessary to make the thing difficult to attain. If he had been a great +and wise philosopher, like the writer of this book, he would now have +comprehended that Work consists of whatever a body is OBLIGED to do, +and that Play consists of whatever a body is not obliged to do. And +this would help him to understand why constructing artificial flowers +or performing on a tread-mill is work, while rolling ten-pins or +climbing Mont Blanc is only amusement. There are wealthy gentlemen in +England who drive four-horse passenger-coaches twenty or thirty miles +on a daily line, in the summer, because the privilege costs them +considerable money; but if they were offered wages for the service, +that would turn it into work and then they would resign. + +The boy mused awhile over the substantial change which had taken place +in his worldly circumstances, and then wended toward headquarters to +report. + + + +CHAPTER III + +TOM presented himself before Aunt Polly, who was sitting by an open +window in a pleasant rearward apartment, which was bedroom, +breakfast-room, dining-room, and library, combined. The balmy summer +air, the restful quiet, the odor of the flowers, and the drowsing murmur +of the bees had had their effect, and she was nodding over her knitting +--for she had no company but the cat, and it was asleep in her lap. Her +spectacles were propped up on her gray head for safety. She had thought +that of course Tom had deserted long ago, and she wondered at seeing him +place himself in her power again in this intrepid way. He said: "Mayn't +I go and play now, aunt?" + +"What, a'ready? How much have you done?" + +"It's all done, aunt." + +"Tom, don't lie to me--I can't bear it." + +"I ain't, aunt; it IS all done." + +Aunt Polly placed small trust in such evidence. She went out to see +for herself; and she would have been content to find twenty per cent. +of Tom's statement true. When she found the entire fence whitewashed, +and not only whitewashed but elaborately coated and recoated, and even +a streak added to the ground, her astonishment was almost unspeakable. +She said: + +"Well, I never! There's no getting round it, you can work when you're +a mind to, Tom." And then she diluted the compliment by adding, "But +it's powerful seldom you're a mind to, I'm bound to say. Well, go 'long +and play; but mind you get back some time in a week, or I'll tan you." + +She was so overcome by the splendor of his achievement that she took +him into the closet and selected a choice apple and delivered it to +him, along with an improving lecture upon the added value and flavor a +treat took to itself when it came without sin through virtuous effort. +And while she closed with a happy Scriptural flourish, he "hooked" a +doughnut. + +Then he skipped out, and saw Sid just starting up the outside stairway +that led to the back rooms on the second floor. Clods were handy and +the air was full of them in a twinkling. They raged around Sid like a +hail-storm; and before Aunt Polly could collect her surprised faculties +and sally to the rescue, six or seven clods had taken personal effect, +and Tom was over the fence and gone. There was a gate, but as a general +thing he was too crowded for time to make use of it. His soul was at +peace, now that he had settled with Sid for calling attention to his +black thread and getting him into trouble. + +Tom skirted the block, and came round into a muddy alley that led by +the back of his aunt's cow-stable. He presently got safely beyond the +reach of capture and punishment, and hastened toward the public square +of the village, where two "military" companies of boys had met for +conflict, according to previous appointment. Tom was General of one of +these armies, Joe Harper (a bosom friend) General of the other. These +two great commanders did not condescend to fight in person--that being +better suited to the still smaller fry--but sat together on an eminence +and conducted the field operations by orders delivered through +aides-de-camp. Tom's army won a great victory, after a long and +hard-fought battle. Then the dead were counted, prisoners exchanged, +the terms of the next disagreement agreed upon, and the day for the +necessary battle appointed; after which the armies fell into line and +marched away, and Tom turned homeward alone. + +As he was passing by the house where Jeff Thatcher lived, he saw a new +girl in the garden--a lovely little blue-eyed creature with yellow hair +plaited into two long-tails, white summer frock and embroidered +pantalettes. The fresh-crowned hero fell without firing a shot. A +certain Amy Lawrence vanished out of his heart and left not even a +memory of herself behind. He had thought he loved her to distraction; +he had regarded his passion as adoration; and behold it was only a poor +little evanescent partiality. He had been months winning her; she had +confessed hardly a week ago; he had been the happiest and the proudest +boy in the world only seven short days, and here in one instant of time +she had gone out of his heart like a casual stranger whose visit is +done. + +He worshipped this new angel with furtive eye, till he saw that she +had discovered him; then he pretended he did not know she was present, +and began to "show off" in all sorts of absurd boyish ways, in order to +win her admiration. He kept up this grotesque foolishness for some +time; but by-and-by, while he was in the midst of some dangerous +gymnastic performances, he glanced aside and saw that the little girl +was wending her way toward the house. Tom came up to the fence and +leaned on it, grieving, and hoping she would tarry yet awhile longer. +She halted a moment on the steps and then moved toward the door. Tom +heaved a great sigh as she put her foot on the threshold. But his face +lit up, right away, for she tossed a pansy over the fence a moment +before she disappeared. + +The boy ran around and stopped within a foot or two of the flower, and +then shaded his eyes with his hand and began to look down street as if +he had discovered something of interest going on in that direction. +Presently he picked up a straw and began trying to balance it on his +nose, with his head tilted far back; and as he moved from side to side, +in his efforts, he edged nearer and nearer toward the pansy; finally +his bare foot rested upon it, his pliant toes closed upon it, and he +hopped away with the treasure and disappeared round the corner. But +only for a minute--only while he could button the flower inside his +jacket, next his heart--or next his stomach, possibly, for he was not +much posted in anatomy, and not hypercritical, anyway. + +He returned, now, and hung about the fence till nightfall, "showing +off," as before; but the girl never exhibited herself again, though Tom +comforted himself a little with the hope that she had been near some +window, meantime, and been aware of his attentions. Finally he strode +home reluctantly, with his poor head full of visions. + +All through supper his spirits were so high that his aunt wondered +"what had got into the child." He took a good scolding about clodding +Sid, and did not seem to mind it in the least. He tried to steal sugar +under his aunt's very nose, and got his knuckles rapped for it. He said: + +"Aunt, you don't whack Sid when he takes it." + +"Well, Sid don't torment a body the way you do. You'd be always into +that sugar if I warn't watching you." + +Presently she stepped into the kitchen, and Sid, happy in his +immunity, reached for the sugar-bowl--a sort of glorying over Tom which +was wellnigh unbearable. But Sid's fingers slipped and the bowl dropped +and broke. Tom was in ecstasies. In such ecstasies that he even +controlled his tongue and was silent. He said to himself that he would +not speak a word, even when his aunt came in, but would sit perfectly +still till she asked who did the mischief; and then he would tell, and +there would be nothing so good in the world as to see that pet model +"catch it." He was so brimful of exultation that he could hardly hold +himself when the old lady came back and stood above the wreck +discharging lightnings of wrath from over her spectacles. He said to +himself, "Now it's coming!" And the next instant he was sprawling on +the floor! The potent palm was uplifted to strike again when Tom cried +out: + +"Hold on, now, what 'er you belting ME for?--Sid broke it!" + +Aunt Polly paused, perplexed, and Tom looked for healing pity. But +when she got her tongue again, she only said: + +"Umf! Well, you didn't get a lick amiss, I reckon. You been into some +other audacious mischief when I wasn't around, like enough." + +Then her conscience reproached her, and she yearned to say something +kind and loving; but she judged that this would be construed into a +confession that she had been in the wrong, and discipline forbade that. +So she kept silence, and went about her affairs with a troubled heart. +Tom sulked in a corner and exalted his woes. He knew that in her heart +his aunt was on her knees to him, and he was morosely gratified by the +consciousness of it. He would hang out no signals, he would take notice +of none. He knew that a yearning glance fell upon him, now and then, +through a film of tears, but he refused recognition of it. He pictured +himself lying sick unto death and his aunt bending over him beseeching +one little forgiving word, but he would turn his face to the wall, and +die with that word unsaid. Ah, how would she feel then? And he pictured +himself brought home from the river, dead, with his curls all wet, and +his sore heart at rest. How she would throw herself upon him, and how +her tears would fall like rain, and her lips pray God to give her back +her boy and she would never, never abuse him any more! But he would lie +there cold and white and make no sign--a poor little sufferer, whose +griefs were at an end. He so worked upon his feelings with the pathos +of these dreams, that he had to keep swallowing, he was so like to +choke; and his eyes swam in a blur of water, which overflowed when he +winked, and ran down and trickled from the end of his nose. And such a +luxury to him was this petting of his sorrows, that he could not bear +to have any worldly cheeriness or any grating delight intrude upon it; +it was too sacred for such contact; and so, presently, when his cousin +Mary danced in, all alive with the joy of seeing home again after an +age-long visit of one week to the country, he got up and moved in +clouds and darkness out at one door as she brought song and sunshine in +at the other. + +He wandered far from the accustomed haunts of boys, and sought +desolate places that were in harmony with his spirit. A log raft in the +river invited him, and he seated himself on its outer edge and +contemplated the dreary vastness of the stream, wishing, the while, +that he could only be drowned, all at once and unconsciously, without +undergoing the uncomfortable routine devised by nature. Then he thought +of his flower. He got it out, rumpled and wilted, and it mightily +increased his dismal felicity. He wondered if she would pity him if she +knew? Would she cry, and wish that she had a right to put her arms +around his neck and comfort him? Or would she turn coldly away like all +the hollow world? This picture brought such an agony of pleasurable +suffering that he worked it over and over again in his mind and set it +up in new and varied lights, till he wore it threadbare. At last he +rose up sighing and departed in the darkness. + +About half-past nine or ten o'clock he came along the deserted street +to where the Adored Unknown lived; he paused a moment; no sound fell +upon his listening ear; a candle was casting a dull glow upon the +curtain of a second-story window. Was the sacred presence there? He +climbed the fence, threaded his stealthy way through the plants, till +he stood under that window; he looked up at it long, and with emotion; +then he laid him down on the ground under it, disposing himself upon +his back, with his hands clasped upon his breast and holding his poor +wilted flower. And thus he would die--out in the cold world, with no +shelter over his homeless head, no friendly hand to wipe the +death-damps from his brow, no loving face to bend pityingly over him +when the great agony came. And thus SHE would see him when she looked +out upon the glad morning, and oh! would she drop one little tear upon +his poor, lifeless form, would she heave one little sigh to see a bright +young life so rudely blighted, so untimely cut down? + +The window went up, a maid-servant's discordant voice profaned the +holy calm, and a deluge of water drenched the prone martyr's remains! + +The strangling hero sprang up with a relieving snort. There was a whiz +as of a missile in the air, mingled with the murmur of a curse, a sound +as of shivering glass followed, and a small, vague form went over the +fence and shot away in the gloom. + +Not long after, as Tom, all undressed for bed, was surveying his +drenched garments by the light of a tallow dip, Sid woke up; but if he +had any dim idea of making any "references to allusions," he thought +better of it and held his peace, for there was danger in Tom's eye. + +Tom turned in without the added vexation of prayers, and Sid made +mental note of the omission. + + + +CHAPTER IV + +THE sun rose upon a tranquil world, and beamed down upon the peaceful +village like a benediction. Breakfast over, Aunt Polly had family +worship: it began with a prayer built from the ground up of solid +courses of Scriptural quotations, welded together with a thin mortar of +originality; and from the summit of this she delivered a grim chapter +of the Mosaic Law, as from Sinai. + +Then Tom girded up his loins, so to speak, and went to work to "get +his verses." Sid had learned his lesson days before. Tom bent all his +energies to the memorizing of five verses, and he chose part of the +Sermon on the Mount, because he could find no verses that were shorter. +At the end of half an hour Tom had a vague general idea of his lesson, +but no more, for his mind was traversing the whole field of human +thought, and his hands were busy with distracting recreations. Mary +took his book to hear him recite, and he tried to find his way through +the fog: + +"Blessed are the--a--a--" + +"Poor"-- + +"Yes--poor; blessed are the poor--a--a--" + +"In spirit--" + +"In spirit; blessed are the poor in spirit, for they--they--" + +"THEIRS--" + +"For THEIRS. Blessed are the poor in spirit, for theirs is the kingdom +of heaven. Blessed are they that mourn, for they--they--" + +"Sh--" + +"For they--a--" + +"S, H, A--" + +"For they S, H--Oh, I don't know what it is!" + +"SHALL!" + +"Oh, SHALL! for they shall--for they shall--a--a--shall mourn--a--a-- +blessed are they that shall--they that--a--they that shall mourn, for +they shall--a--shall WHAT? Why don't you tell me, Mary?--what do you +want to be so mean for?" + +"Oh, Tom, you poor thick-headed thing, I'm not teasing you. I wouldn't +do that. You must go and learn it again. Don't you be discouraged, Tom, +you'll manage it--and if you do, I'll give you something ever so nice. +There, now, that's a good boy." + +"All right! What is it, Mary, tell me what it is." + +"Never you mind, Tom. You know if I say it's nice, it is nice." + +"You bet you that's so, Mary. All right, I'll tackle it again." + +And he did "tackle it again"--and under the double pressure of +curiosity and prospective gain he did it with such spirit that he +accomplished a shining success. Mary gave him a brand-new "Barlow" +knife worth twelve and a half cents; and the convulsion of delight that +swept his system shook him to his foundations. True, the knife would +not cut anything, but it was a "sure-enough" Barlow, and there was +inconceivable grandeur in that--though where the Western boys ever got +the idea that such a weapon could possibly be counterfeited to its +injury is an imposing mystery and will always remain so, perhaps. Tom +contrived to scarify the cupboard with it, and was arranging to begin +on the bureau, when he was called off to dress for Sunday-school. + +Mary gave him a tin basin of water and a piece of soap, and he went +outside the door and set the basin on a little bench there; then he +dipped the soap in the water and laid it down; turned up his sleeves; +poured out the water on the ground, gently, and then entered the +kitchen and began to wipe his face diligently on the towel behind the +door. But Mary removed the towel and said: + +"Now ain't you ashamed, Tom. You mustn't be so bad. Water won't hurt +you." + +Tom was a trifle disconcerted. The basin was refilled, and this time +he stood over it a little while, gathering resolution; took in a big +breath and began. When he entered the kitchen presently, with both eyes +shut and groping for the towel with his hands, an honorable testimony +of suds and water was dripping from his face. But when he emerged from +the towel, he was not yet satisfactory, for the clean territory stopped +short at his chin and his jaws, like a mask; below and beyond this line +there was a dark expanse of unirrigated soil that spread downward in +front and backward around his neck. Mary took him in hand, and when she +was done with him he was a man and a brother, without distinction of +color, and his saturated hair was neatly brushed, and its short curls +wrought into a dainty and symmetrical general effect. [He privately +smoothed out the curls, with labor and difficulty, and plastered his +hair close down to his head; for he held curls to be effeminate, and +his own filled his life with bitterness.] Then Mary got out a suit of +his clothing that had been used only on Sundays during two years--they +were simply called his "other clothes"--and so by that we know the +size of his wardrobe. The girl "put him to rights" after he had dressed +himself; she buttoned his neat roundabout up to his chin, turned his +vast shirt collar down over his shoulders, brushed him off and crowned +him with his speckled straw hat. He now looked exceedingly improved and +uncomfortable. He was fully as uncomfortable as he looked; for there +was a restraint about whole clothes and cleanliness that galled him. He +hoped that Mary would forget his shoes, but the hope was blighted; she +coated them thoroughly with tallow, as was the custom, and brought them +out. He lost his temper and said he was always being made to do +everything he didn't want to do. But Mary said, persuasively: + +"Please, Tom--that's a good boy." + +So he got into the shoes snarling. Mary was soon ready, and the three +children set out for Sunday-school--a place that Tom hated with his +whole heart; but Sid and Mary were fond of it. + +Sabbath-school hours were from nine to half-past ten; and then church +service. Two of the children always remained for the sermon +voluntarily, and the other always remained too--for stronger reasons. +The church's high-backed, uncushioned pews would seat about three +hundred persons; the edifice was but a small, plain affair, with a sort +of pine board tree-box on top of it for a steeple. At the door Tom +dropped back a step and accosted a Sunday-dressed comrade: + +"Say, Billy, got a yaller ticket?" + +"Yes." + +"What'll you take for her?" + +"What'll you give?" + +"Piece of lickrish and a fish-hook." + +"Less see 'em." + +Tom exhibited. They were satisfactory, and the property changed hands. +Then Tom traded a couple of white alleys for three red tickets, and +some small trifle or other for a couple of blue ones. He waylaid other +boys as they came, and went on buying tickets of various colors ten or +fifteen minutes longer. He entered the church, now, with a swarm of +clean and noisy boys and girls, proceeded to his seat and started a +quarrel with the first boy that came handy. The teacher, a grave, +elderly man, interfered; then turned his back a moment and Tom pulled a +boy's hair in the next bench, and was absorbed in his book when the boy +turned around; stuck a pin in another boy, presently, in order to hear +him say "Ouch!" and got a new reprimand from his teacher. Tom's whole +class were of a pattern--restless, noisy, and troublesome. When they +came to recite their lessons, not one of them knew his verses +perfectly, but had to be prompted all along. However, they worried +through, and each got his reward--in small blue tickets, each with a +passage of Scripture on it; each blue ticket was pay for two verses of +the recitation. Ten blue tickets equalled a red one, and could be +exchanged for it; ten red tickets equalled a yellow one; for ten yellow +tickets the superintendent gave a very plainly bound Bible (worth forty +cents in those easy times) to the pupil. How many of my readers would +have the industry and application to memorize two thousand verses, even +for a Dore Bible? And yet Mary had acquired two Bibles in this way--it +was the patient work of two years--and a boy of German parentage had +won four or five. He once recited three thousand verses without +stopping; but the strain upon his mental faculties was too great, and +he was little better than an idiot from that day forth--a grievous +misfortune for the school, for on great occasions, before company, the +superintendent (as Tom expressed it) had always made this boy come out +and "spread himself." Only the older pupils managed to keep their +tickets and stick to their tedious work long enough to get a Bible, and +so the delivery of one of these prizes was a rare and noteworthy +circumstance; the successful pupil was so great and conspicuous for +that day that on the spot every scholar's heart was fired with a fresh +ambition that often lasted a couple of weeks. It is possible that Tom's +mental stomach had never really hungered for one of those prizes, but +unquestionably his entire being had for many a day longed for the glory +and the eclat that came with it. + +In due course the superintendent stood up in front of the pulpit, with +a closed hymn-book in his hand and his forefinger inserted between its +leaves, and commanded attention. When a Sunday-school superintendent +makes his customary little speech, a hymn-book in the hand is as +necessary as is the inevitable sheet of music in the hand of a singer +who stands forward on the platform and sings a solo at a concert +--though why, is a mystery: for neither the hymn-book nor the sheet of +music is ever referred to by the sufferer. This superintendent was a +slim creature of thirty-five, with a sandy goatee and short sandy hair; +he wore a stiff standing-collar whose upper edge almost reached his +ears and whose sharp points curved forward abreast the corners of his +mouth--a fence that compelled a straight lookout ahead, and a turning +of the whole body when a side view was required; his chin was propped +on a spreading cravat which was as broad and as long as a bank-note, +and had fringed ends; his boot toes were turned sharply up, in the +fashion of the day, like sleigh-runners--an effect patiently and +laboriously produced by the young men by sitting with their toes +pressed against a wall for hours together. Mr. Walters was very earnest +of mien, and very sincere and honest at heart; and he held sacred +things and places in such reverence, and so separated them from worldly +matters, that unconsciously to himself his Sunday-school voice had +acquired a peculiar intonation which was wholly absent on week-days. He +began after this fashion: + +"Now, children, I want you all to sit up just as straight and pretty +as you can and give me all your attention for a minute or two. There +--that is it. That is the way good little boys and girls should do. I see +one little girl who is looking out of the window--I am afraid she +thinks I am out there somewhere--perhaps up in one of the trees making +a speech to the little birds. [Applausive titter.] I want to tell you +how good it makes me feel to see so many bright, clean little faces +assembled in a place like this, learning to do right and be good." And +so forth and so on. It is not necessary to set down the rest of the +oration. It was of a pattern which does not vary, and so it is familiar +to us all. + +The latter third of the speech was marred by the resumption of fights +and other recreations among certain of the bad boys, and by fidgetings +and whisperings that extended far and wide, washing even to the bases +of isolated and incorruptible rocks like Sid and Mary. But now every +sound ceased suddenly, with the subsidence of Mr. Walters' voice, and +the conclusion of the speech was received with a burst of silent +gratitude. + +A good part of the whispering had been occasioned by an event which +was more or less rare--the entrance of visitors: lawyer Thatcher, +accompanied by a very feeble and aged man; a fine, portly, middle-aged +gentleman with iron-gray hair; and a dignified lady who was doubtless +the latter's wife. The lady was leading a child. Tom had been restless +and full of chafings and repinings; conscience-smitten, too--he could +not meet Amy Lawrence's eye, he could not brook her loving gaze. But +when he saw this small new-comer his soul was all ablaze with bliss in +a moment. The next moment he was "showing off" with all his might +--cuffing boys, pulling hair, making faces--in a word, using every art +that seemed likely to fascinate a girl and win her applause. His +exaltation had but one alloy--the memory of his humiliation in this +angel's garden--and that record in sand was fast washing out, under +the waves of happiness that were sweeping over it now. + +The visitors were given the highest seat of honor, and as soon as Mr. +Walters' speech was finished, he introduced them to the school. The +middle-aged man turned out to be a prodigious personage--no less a one +than the county judge--altogether the most august creation these +children had ever looked upon--and they wondered what kind of material +he was made of--and they half wanted to hear him roar, and were half +afraid he might, too. He was from Constantinople, twelve miles away--so +he had travelled, and seen the world--these very eyes had looked upon +the county court-house--which was said to have a tin roof. The awe +which these reflections inspired was attested by the impressive silence +and the ranks of staring eyes. This was the great Judge Thatcher, +brother of their own lawyer. Jeff Thatcher immediately went forward, to +be familiar with the great man and be envied by the school. It would +have been music to his soul to hear the whisperings: + +"Look at him, Jim! He's a going up there. Say--look! he's a going to +shake hands with him--he IS shaking hands with him! By jings, don't you +wish you was Jeff?" + +Mr. Walters fell to "showing off," with all sorts of official +bustlings and activities, giving orders, delivering judgments, +discharging directions here, there, everywhere that he could find a +target. The librarian "showed off"--running hither and thither with his +arms full of books and making a deal of the splutter and fuss that +insect authority delights in. The young lady teachers "showed off" +--bending sweetly over pupils that were lately being boxed, lifting +pretty warning fingers at bad little boys and patting good ones +lovingly. The young gentlemen teachers "showed off" with small +scoldings and other little displays of authority and fine attention to +discipline--and most of the teachers, of both sexes, found business up +at the library, by the pulpit; and it was business that frequently had +to be done over again two or three times (with much seeming vexation). +The little girls "showed off" in various ways, and the little boys +"showed off" with such diligence that the air was thick with paper wads +and the murmur of scufflings. And above it all the great man sat and +beamed a majestic judicial smile upon all the house, and warmed himself +in the sun of his own grandeur--for he was "showing off," too. + +There was only one thing wanting to make Mr. Walters' ecstasy +complete, and that was a chance to deliver a Bible-prize and exhibit a +prodigy. Several pupils had a few yellow tickets, but none had enough +--he had been around among the star pupils inquiring. He would have given +worlds, now, to have that German lad back again with a sound mind. + +And now at this moment, when hope was dead, Tom Sawyer came forward +with nine yellow tickets, nine red tickets, and ten blue ones, and +demanded a Bible. This was a thunderbolt out of a clear sky. Walters +was not expecting an application from this source for the next ten +years. But there was no getting around it--here were the certified +checks, and they were good for their face. Tom was therefore elevated +to a place with the Judge and the other elect, and the great news was +announced from headquarters. It was the most stunning surprise of the +decade, and so profound was the sensation that it lifted the new hero +up to the judicial one's altitude, and the school had two marvels to +gaze upon in place of one. The boys were all eaten up with envy--but +those that suffered the bitterest pangs were those who perceived too +late that they themselves had contributed to this hated splendor by +trading tickets to Tom for the wealth he had amassed in selling +whitewashing privileges. These despised themselves, as being the dupes +of a wily fraud, a guileful snake in the grass. + +The prize was delivered to Tom with as much effusion as the +superintendent could pump up under the circumstances; but it lacked +somewhat of the true gush, for the poor fellow's instinct taught him +that there was a mystery here that could not well bear the light, +perhaps; it was simply preposterous that this boy had warehoused two +thousand sheaves of Scriptural wisdom on his premises--a dozen would +strain his capacity, without a doubt. + +Amy Lawrence was proud and glad, and she tried to make Tom see it in +her face--but he wouldn't look. She wondered; then she was just a grain +troubled; next a dim suspicion came and went--came again; she watched; +a furtive glance told her worlds--and then her heart broke, and she was +jealous, and angry, and the tears came and she hated everybody. Tom +most of all (she thought). + +Tom was introduced to the Judge; but his tongue was tied, his breath +would hardly come, his heart quaked--partly because of the awful +greatness of the man, but mainly because he was her parent. He would +have liked to fall down and worship him, if it were in the dark. The +Judge put his hand on Tom's head and called him a fine little man, and +asked him what his name was. The boy stammered, gasped, and got it out: + +"Tom." + +"Oh, no, not Tom--it is--" + +"Thomas." + +"Ah, that's it. I thought there was more to it, maybe. That's very +well. But you've another one I daresay, and you'll tell it to me, won't +you?" + +"Tell the gentleman your other name, Thomas," said Walters, "and say +sir. You mustn't forget your manners." + +"Thomas Sawyer--sir." + +"That's it! That's a good boy. Fine boy. Fine, manly little fellow. +Two thousand verses is a great many--very, very great many. And you +never can be sorry for the trouble you took to learn them; for +knowledge is worth more than anything there is in the world; it's what +makes great men and good men; you'll be a great man and a good man +yourself, some day, Thomas, and then you'll look back and say, It's all +owing to the precious Sunday-school privileges of my boyhood--it's all +owing to my dear teachers that taught me to learn--it's all owing to +the good superintendent, who encouraged me, and watched over me, and +gave me a beautiful Bible--a splendid elegant Bible--to keep and have +it all for my own, always--it's all owing to right bringing up! That is +what you will say, Thomas--and you wouldn't take any money for those +two thousand verses--no indeed you wouldn't. And now you wouldn't mind +telling me and this lady some of the things you've learned--no, I know +you wouldn't--for we are proud of little boys that learn. Now, no +doubt you know the names of all the twelve disciples. Won't you tell us +the names of the first two that were appointed?" + +Tom was tugging at a button-hole and looking sheepish. He blushed, +now, and his eyes fell. Mr. Walters' heart sank within him. He said to +himself, it is not possible that the boy can answer the simplest +question--why DID the Judge ask him? Yet he felt obliged to speak up +and say: + +"Answer the gentleman, Thomas--don't be afraid." + +Tom still hung fire. + +"Now I know you'll tell me," said the lady. "The names of the first +two disciples were--" + +"DAVID AND GOLIAH!" + +Let us draw the curtain of charity over the rest of the scene. + + + +CHAPTER V + +ABOUT half-past ten the cracked bell of the small church began to +ring, and presently the people began to gather for the morning sermon. +The Sunday-school children distributed themselves about the house and +occupied pews with their parents, so as to be under supervision. Aunt +Polly came, and Tom and Sid and Mary sat with her--Tom being placed +next the aisle, in order that he might be as far away from the open +window and the seductive outside summer scenes as possible. The crowd +filed up the aisles: the aged and needy postmaster, who had seen better +days; the mayor and his wife--for they had a mayor there, among other +unnecessaries; the justice of the peace; the widow Douglass, fair, +smart, and forty, a generous, good-hearted soul and well-to-do, her +hill mansion the only palace in the town, and the most hospitable and +much the most lavish in the matter of festivities that St. Petersburg +could boast; the bent and venerable Major and Mrs. Ward; lawyer +Riverson, the new notable from a distance; next the belle of the +village, followed by a troop of lawn-clad and ribbon-decked young +heart-breakers; then all the young clerks in town in a body--for they +had stood in the vestibule sucking their cane-heads, a circling wall of +oiled and simpering admirers, till the last girl had run their gantlet; +and last of all came the Model Boy, Willie Mufferson, taking as heedful +care of his mother as if she were cut glass. He always brought his +mother to church, and was the pride of all the matrons. The boys all +hated him, he was so good. And besides, he had been "thrown up to them" +so much. His white handkerchief was hanging out of his pocket behind, as +usual on Sundays--accidentally. Tom had no handkerchief, and he looked +upon boys who had as snobs. + +The congregation being fully assembled, now, the bell rang once more, +to warn laggards and stragglers, and then a solemn hush fell upon the +church which was only broken by the tittering and whispering of the +choir in the gallery. The choir always tittered and whispered all +through service. There was once a church choir that was not ill-bred, +but I have forgotten where it was, now. It was a great many years ago, +and I can scarcely remember anything about it, but I think it was in +some foreign country. + +The minister gave out the hymn, and read it through with a relish, in +a peculiar style which was much admired in that part of the country. +His voice began on a medium key and climbed steadily up till it reached +a certain point, where it bore with strong emphasis upon the topmost +word and then plunged down as if from a spring-board: + + Shall I be car-ri-ed toe the skies, on flow'ry BEDS of ease, + + Whilst others fight to win the prize, and sail thro' BLOODY seas? + +He was regarded as a wonderful reader. At church "sociables" he was +always called upon to read poetry; and when he was through, the ladies +would lift up their hands and let them fall helplessly in their laps, +and "wall" their eyes, and shake their heads, as much as to say, "Words +cannot express it; it is too beautiful, TOO beautiful for this mortal +earth." + +After the hymn had been sung, the Rev. Mr. Sprague turned himself into +a bulletin-board, and read off "notices" of meetings and societies and +things till it seemed that the list would stretch out to the crack of +doom--a queer custom which is still kept up in America, even in cities, +away here in this age of abundant newspapers. Often, the less there is +to justify a traditional custom, the harder it is to get rid of it. + +And now the minister prayed. A good, generous prayer it was, and went +into details: it pleaded for the church, and the little children of the +church; for the other churches of the village; for the village itself; +for the county; for the State; for the State officers; for the United +States; for the churches of the United States; for Congress; for the +President; for the officers of the Government; for poor sailors, tossed +by stormy seas; for the oppressed millions groaning under the heel of +European monarchies and Oriental despotisms; for such as have the light +and the good tidings, and yet have not eyes to see nor ears to hear +withal; for the heathen in the far islands of the sea; and closed with +a supplication that the words he was about to speak might find grace +and favor, and be as seed sown in fertile ground, yielding in time a +grateful harvest of good. Amen. + +There was a rustling of dresses, and the standing congregation sat +down. The boy whose history this book relates did not enjoy the prayer, +he only endured it--if he even did that much. He was restive all +through it; he kept tally of the details of the prayer, unconsciously +--for he was not listening, but he knew the ground of old, and the +clergyman's regular route over it--and when a little trifle of new +matter was interlarded, his ear detected it and his whole nature +resented it; he considered additions unfair, and scoundrelly. In the +midst of the prayer a fly had lit on the back of the pew in front of +him and tortured his spirit by calmly rubbing its hands together, +embracing its head with its arms, and polishing it so vigorously that +it seemed to almost part company with the body, and the slender thread +of a neck was exposed to view; scraping its wings with its hind legs +and smoothing them to its body as if they had been coat-tails; going +through its whole toilet as tranquilly as if it knew it was perfectly +safe. As indeed it was; for as sorely as Tom's hands itched to grab for +it they did not dare--he believed his soul would be instantly destroyed +if he did such a thing while the prayer was going on. But with the +closing sentence his hand began to curve and steal forward; and the +instant the "Amen" was out the fly was a prisoner of war. His aunt +detected the act and made him let it go. + +The minister gave out his text and droned along monotonously through +an argument that was so prosy that many a head by and by began to nod +--and yet it was an argument that dealt in limitless fire and brimstone +and thinned the predestined elect down to a company so small as to be +hardly worth the saving. Tom counted the pages of the sermon; after +church he always knew how many pages there had been, but he seldom knew +anything else about the discourse. However, this time he was really +interested for a little while. The minister made a grand and moving +picture of the assembling together of the world's hosts at the +millennium when the lion and the lamb should lie down together and a +little child should lead them. But the pathos, the lesson, the moral of +the great spectacle were lost upon the boy; he only thought of the +conspicuousness of the principal character before the on-looking +nations; his face lit with the thought, and he said to himself that he +wished he could be that child, if it was a tame lion. + +Now he lapsed into suffering again, as the dry argument was resumed. +Presently he bethought him of a treasure he had and got it out. It was +a large black beetle with formidable jaws--a "pinchbug," he called it. +It was in a percussion-cap box. The first thing the beetle did was to +take him by the finger. A natural fillip followed, the beetle went +floundering into the aisle and lit on its back, and the hurt finger +went into the boy's mouth. The beetle lay there working its helpless +legs, unable to turn over. Tom eyed it, and longed for it; but it was +safe out of his reach. Other people uninterested in the sermon found +relief in the beetle, and they eyed it too. Presently a vagrant poodle +dog came idling along, sad at heart, lazy with the summer softness and +the quiet, weary of captivity, sighing for change. He spied the beetle; +the drooping tail lifted and wagged. He surveyed the prize; walked +around it; smelt at it from a safe distance; walked around it again; +grew bolder, and took a closer smell; then lifted his lip and made a +gingerly snatch at it, just missing it; made another, and another; +began to enjoy the diversion; subsided to his stomach with the beetle +between his paws, and continued his experiments; grew weary at last, +and then indifferent and absent-minded. His head nodded, and little by +little his chin descended and touched the enemy, who seized it. There +was a sharp yelp, a flirt of the poodle's head, and the beetle fell a +couple of yards away, and lit on its back once more. The neighboring +spectators shook with a gentle inward joy, several faces went behind +fans and handkerchiefs, and Tom was entirely happy. The dog looked +foolish, and probably felt so; but there was resentment in his heart, +too, and a craving for revenge. So he went to the beetle and began a +wary attack on it again; jumping at it from every point of a circle, +lighting with his fore-paws within an inch of the creature, making even +closer snatches at it with his teeth, and jerking his head till his +ears flapped again. But he grew tired once more, after a while; tried +to amuse himself with a fly but found no relief; followed an ant +around, with his nose close to the floor, and quickly wearied of that; +yawned, sighed, forgot the beetle entirely, and sat down on it. Then +there was a wild yelp of agony and the poodle went sailing up the +aisle; the yelps continued, and so did the dog; he crossed the house in +front of the altar; he flew down the other aisle; he crossed before the +doors; he clamored up the home-stretch; his anguish grew with his +progress, till presently he was but a woolly comet moving in its orbit +with the gleam and the speed of light. At last the frantic sufferer +sheered from its course, and sprang into its master's lap; he flung it +out of the window, and the voice of distress quickly thinned away and +died in the distance. + +By this time the whole church was red-faced and suffocating with +suppressed laughter, and the sermon had come to a dead standstill. The +discourse was resumed presently, but it went lame and halting, all +possibility of impressiveness being at an end; for even the gravest +sentiments were constantly being received with a smothered burst of +unholy mirth, under cover of some remote pew-back, as if the poor +parson had said a rarely facetious thing. It was a genuine relief to +the whole congregation when the ordeal was over and the benediction +pronounced. + +Tom Sawyer went home quite cheerful, thinking to himself that there +was some satisfaction about divine service when there was a bit of +variety in it. He had but one marring thought; he was willing that the +dog should play with his pinchbug, but he did not think it was upright +in him to carry it off. + + + +CHAPTER VI + +MONDAY morning found Tom Sawyer miserable. Monday morning always found +him so--because it began another week's slow suffering in school. He +generally began that day with wishing he had had no intervening +holiday, it made the going into captivity and fetters again so much +more odious. + +Tom lay thinking. Presently it occurred to him that he wished he was +sick; then he could stay home from school. Here was a vague +possibility. He canvassed his system. No ailment was found, and he +investigated again. This time he thought he could detect colicky +symptoms, and he began to encourage them with considerable hope. But +they soon grew feeble, and presently died wholly away. He reflected +further. Suddenly he discovered something. One of his upper front teeth +was loose. This was lucky; he was about to begin to groan, as a +"starter," as he called it, when it occurred to him that if he came +into court with that argument, his aunt would pull it out, and that +would hurt. So he thought he would hold the tooth in reserve for the +present, and seek further. Nothing offered for some little time, and +then he remembered hearing the doctor tell about a certain thing that +laid up a patient for two or three weeks and threatened to make him +lose a finger. So the boy eagerly drew his sore toe from under the +sheet and held it up for inspection. But now he did not know the +necessary symptoms. However, it seemed well worth while to chance it, +so he fell to groaning with considerable spirit. + +But Sid slept on unconscious. + +Tom groaned louder, and fancied that he began to feel pain in the toe. + +No result from Sid. + +Tom was panting with his exertions by this time. He took a rest and +then swelled himself up and fetched a succession of admirable groans. + +Sid snored on. + +Tom was aggravated. He said, "Sid, Sid!" and shook him. This course +worked well, and Tom began to groan again. Sid yawned, stretched, then +brought himself up on his elbow with a snort, and began to stare at +Tom. Tom went on groaning. Sid said: + +"Tom! Say, Tom!" [No response.] "Here, Tom! TOM! What is the matter, +Tom?" And he shook him and looked in his face anxiously. + +Tom moaned out: + +"Oh, don't, Sid. Don't joggle me." + +"Why, what's the matter, Tom? I must call auntie." + +"No--never mind. It'll be over by and by, maybe. Don't call anybody." + +"But I must! DON'T groan so, Tom, it's awful. How long you been this +way?" + +"Hours. Ouch! Oh, don't stir so, Sid, you'll kill me." + +"Tom, why didn't you wake me sooner? Oh, Tom, DON'T! It makes my +flesh crawl to hear you. Tom, what is the matter?" + +"I forgive you everything, Sid. [Groan.] Everything you've ever done +to me. When I'm gone--" + +"Oh, Tom, you ain't dying, are you? Don't, Tom--oh, don't. Maybe--" + +"I forgive everybody, Sid. [Groan.] Tell 'em so, Sid. And Sid, you +give my window-sash and my cat with one eye to that new girl that's +come to town, and tell her--" + +But Sid had snatched his clothes and gone. Tom was suffering in +reality, now, so handsomely was his imagination working, and so his +groans had gathered quite a genuine tone. + +Sid flew down-stairs and said: + +"Oh, Aunt Polly, come! Tom's dying!" + +"Dying!" + +"Yes'm. Don't wait--come quick!" + +"Rubbage! I don't believe it!" + +But she fled up-stairs, nevertheless, with Sid and Mary at her heels. +And her face grew white, too, and her lip trembled. When she reached +the bedside she gasped out: + +"You, Tom! Tom, what's the matter with you?" + +"Oh, auntie, I'm--" + +"What's the matter with you--what is the matter with you, child?" + +"Oh, auntie, my sore toe's mortified!" + +The old lady sank down into a chair and laughed a little, then cried a +little, then did both together. This restored her and she said: + +"Tom, what a turn you did give me. Now you shut up that nonsense and +climb out of this." + +The groans ceased and the pain vanished from the toe. The boy felt a +little foolish, and he said: + +"Aunt Polly, it SEEMED mortified, and it hurt so I never minded my +tooth at all." + +"Your tooth, indeed! What's the matter with your tooth?" + +"One of them's loose, and it aches perfectly awful." + +"There, there, now, don't begin that groaning again. Open your mouth. +Well--your tooth IS loose, but you're not going to die about that. +Mary, get me a silk thread, and a chunk of fire out of the kitchen." + +Tom said: + +"Oh, please, auntie, don't pull it out. It don't hurt any more. I wish +I may never stir if it does. Please don't, auntie. I don't want to stay +home from school." + +"Oh, you don't, don't you? So all this row was because you thought +you'd get to stay home from school and go a-fishing? Tom, Tom, I love +you so, and you seem to try every way you can to break my old heart +with your outrageousness." By this time the dental instruments were +ready. The old lady made one end of the silk thread fast to Tom's tooth +with a loop and tied the other to the bedpost. Then she seized the +chunk of fire and suddenly thrust it almost into the boy's face. The +tooth hung dangling by the bedpost, now. + +But all trials bring their compensations. As Tom wended to school +after breakfast, he was the envy of every boy he met because the gap in +his upper row of teeth enabled him to expectorate in a new and +admirable way. He gathered quite a following of lads interested in the +exhibition; and one that had cut his finger and had been a centre of +fascination and homage up to this time, now found himself suddenly +without an adherent, and shorn of his glory. His heart was heavy, and +he said with a disdain which he did not feel that it wasn't anything to +spit like Tom Sawyer; but another boy said, "Sour grapes!" and he +wandered away a dismantled hero. + +Shortly Tom came upon the juvenile pariah of the village, Huckleberry +Finn, son of the town drunkard. Huckleberry was cordially hated and +dreaded by all the mothers of the town, because he was idle and lawless +and vulgar and bad--and because all their children admired him so, and +delighted in his forbidden society, and wished they dared to be like +him. Tom was like the rest of the respectable boys, in that he envied +Huckleberry his gaudy outcast condition, and was under strict orders +not to play with him. So he played with him every time he got a chance. +Huckleberry was always dressed in the cast-off clothes of full-grown +men, and they were in perennial bloom and fluttering with rags. His hat +was a vast ruin with a wide crescent lopped out of its brim; his coat, +when he wore one, hung nearly to his heels and had the rearward buttons +far down the back; but one suspender supported his trousers; the seat +of the trousers bagged low and contained nothing, the fringed legs +dragged in the dirt when not rolled up. + +Huckleberry came and went, at his own free will. He slept on doorsteps +in fine weather and in empty hogsheads in wet; he did not have to go to +school or to church, or call any being master or obey anybody; he could +go fishing or swimming when and where he chose, and stay as long as it +suited him; nobody forbade him to fight; he could sit up as late as he +pleased; he was always the first boy that went barefoot in the spring +and the last to resume leather in the fall; he never had to wash, nor +put on clean clothes; he could swear wonderfully. In a word, everything +that goes to make life precious that boy had. So thought every +harassed, hampered, respectable boy in St. Petersburg. + +Tom hailed the romantic outcast: + +"Hello, Huckleberry!" + +"Hello yourself, and see how you like it." + +"What's that you got?" + +"Dead cat." + +"Lemme see him, Huck. My, he's pretty stiff. Where'd you get him?" + +"Bought him off'n a boy." + +"What did you give?" + +"I give a blue ticket and a bladder that I got at the slaughter-house." + +"Where'd you get the blue ticket?" + +"Bought it off'n Ben Rogers two weeks ago for a hoop-stick." + +"Say--what is dead cats good for, Huck?" + +"Good for? Cure warts with." + +"No! Is that so? I know something that's better." + +"I bet you don't. What is it?" + +"Why, spunk-water." + +"Spunk-water! I wouldn't give a dern for spunk-water." + +"You wouldn't, wouldn't you? D'you ever try it?" + +"No, I hain't. But Bob Tanner did." + +"Who told you so!" + +"Why, he told Jeff Thatcher, and Jeff told Johnny Baker, and Johnny +told Jim Hollis, and Jim told Ben Rogers, and Ben told a nigger, and +the nigger told me. There now!" + +"Well, what of it? They'll all lie. Leastways all but the nigger. I +don't know HIM. But I never see a nigger that WOULDN'T lie. Shucks! Now +you tell me how Bob Tanner done it, Huck." + +"Why, he took and dipped his hand in a rotten stump where the +rain-water was." + +"In the daytime?" + +"Certainly." + +"With his face to the stump?" + +"Yes. Least I reckon so." + +"Did he say anything?" + +"I don't reckon he did. I don't know." + +"Aha! Talk about trying to cure warts with spunk-water such a blame +fool way as that! Why, that ain't a-going to do any good. You got to go +all by yourself, to the middle of the woods, where you know there's a +spunk-water stump, and just as it's midnight you back up against the +stump and jam your hand in and say: + + 'Barley-corn, barley-corn, injun-meal shorts, + Spunk-water, spunk-water, swaller these warts,' + +and then walk away quick, eleven steps, with your eyes shut, and then +turn around three times and walk home without speaking to anybody. +Because if you speak the charm's busted." + +"Well, that sounds like a good way; but that ain't the way Bob Tanner +done." + +"No, sir, you can bet he didn't, becuz he's the wartiest boy in this +town; and he wouldn't have a wart on him if he'd knowed how to work +spunk-water. I've took off thousands of warts off of my hands that way, +Huck. I play with frogs so much that I've always got considerable many +warts. Sometimes I take 'em off with a bean." + +"Yes, bean's good. I've done that." + +"Have you? What's your way?" + +"You take and split the bean, and cut the wart so as to get some +blood, and then you put the blood on one piece of the bean and take and +dig a hole and bury it 'bout midnight at the crossroads in the dark of +the moon, and then you burn up the rest of the bean. You see that piece +that's got the blood on it will keep drawing and drawing, trying to +fetch the other piece to it, and so that helps the blood to draw the +wart, and pretty soon off she comes." + +"Yes, that's it, Huck--that's it; though when you're burying it if you +say 'Down bean; off wart; come no more to bother me!' it's better. +That's the way Joe Harper does, and he's been nearly to Coonville and +most everywheres. But say--how do you cure 'em with dead cats?" + +"Why, you take your cat and go and get in the graveyard 'long about +midnight when somebody that was wicked has been buried; and when it's +midnight a devil will come, or maybe two or three, but you can't see +'em, you can only hear something like the wind, or maybe hear 'em talk; +and when they're taking that feller away, you heave your cat after 'em +and say, 'Devil follow corpse, cat follow devil, warts follow cat, I'm +done with ye!' That'll fetch ANY wart." + +"Sounds right. D'you ever try it, Huck?" + +"No, but old Mother Hopkins told me." + +"Well, I reckon it's so, then. Becuz they say she's a witch." + +"Say! Why, Tom, I KNOW she is. She witched pap. Pap says so his own +self. He come along one day, and he see she was a-witching him, so he +took up a rock, and if she hadn't dodged, he'd a got her. Well, that +very night he rolled off'n a shed wher' he was a layin drunk, and broke +his arm." + +"Why, that's awful. How did he know she was a-witching him?" + +"Lord, pap can tell, easy. Pap says when they keep looking at you +right stiddy, they're a-witching you. Specially if they mumble. Becuz +when they mumble they're saying the Lord's Prayer backards." + +"Say, Hucky, when you going to try the cat?" + +"To-night. I reckon they'll come after old Hoss Williams to-night." + +"But they buried him Saturday. Didn't they get him Saturday night?" + +"Why, how you talk! How could their charms work till midnight?--and +THEN it's Sunday. Devils don't slosh around much of a Sunday, I don't +reckon." + +"I never thought of that. That's so. Lemme go with you?" + +"Of course--if you ain't afeard." + +"Afeard! 'Tain't likely. Will you meow?" + +"Yes--and you meow back, if you get a chance. Last time, you kep' me +a-meowing around till old Hays went to throwing rocks at me and says +'Dern that cat!' and so I hove a brick through his window--but don't +you tell." + +"I won't. I couldn't meow that night, becuz auntie was watching me, +but I'll meow this time. Say--what's that?" + +"Nothing but a tick." + +"Where'd you get him?" + +"Out in the woods." + +"What'll you take for him?" + +"I don't know. I don't want to sell him." + +"All right. It's a mighty small tick, anyway." + +"Oh, anybody can run a tick down that don't belong to them. I'm +satisfied with it. It's a good enough tick for me." + +"Sho, there's ticks a plenty. I could have a thousand of 'em if I +wanted to." + +"Well, why don't you? Becuz you know mighty well you can't. This is a +pretty early tick, I reckon. It's the first one I've seen this year." + +"Say, Huck--I'll give you my tooth for him." + +"Less see it." + +Tom got out a bit of paper and carefully unrolled it. Huckleberry +viewed it wistfully. The temptation was very strong. At last he said: + +"Is it genuwyne?" + +Tom lifted his lip and showed the vacancy. + +"Well, all right," said Huckleberry, "it's a trade." + +Tom enclosed the tick in the percussion-cap box that had lately been +the pinchbug's prison, and the boys separated, each feeling wealthier +than before. + +When Tom reached the little isolated frame schoolhouse, he strode in +briskly, with the manner of one who had come with all honest speed. +He hung his hat on a peg and flung himself into his seat with +business-like alacrity. The master, throned on high in his great +splint-bottom arm-chair, was dozing, lulled by the drowsy hum of study. +The interruption roused him. + +"Thomas Sawyer!" + +Tom knew that when his name was pronounced in full, it meant trouble. + +"Sir!" + +"Come up here. Now, sir, why are you late again, as usual?" + +Tom was about to take refuge in a lie, when he saw two long tails of +yellow hair hanging down a back that he recognized by the electric +sympathy of love; and by that form was THE ONLY VACANT PLACE on the +girls' side of the schoolhouse. He instantly said: + +"I STOPPED TO TALK WITH HUCKLEBERRY FINN!" + +The master's pulse stood still, and he stared helplessly. The buzz of +study ceased. The pupils wondered if this foolhardy boy had lost his +mind. The master said: + +"You--you did what?" + +"Stopped to talk with Huckleberry Finn." + +There was no mistaking the words. + +"Thomas Sawyer, this is the most astounding confession I have ever +listened to. No mere ferule will answer for this offence. Take off your +jacket." + +The master's arm performed until it was tired and the stock of +switches notably diminished. Then the order followed: + +"Now, sir, go and sit with the girls! And let this be a warning to you." + +The titter that rippled around the room appeared to abash the boy, but +in reality that result was caused rather more by his worshipful awe of +his unknown idol and the dread pleasure that lay in his high good +fortune. He sat down upon the end of the pine bench and the girl +hitched herself away from him with a toss of her head. Nudges and winks +and whispers traversed the room, but Tom sat still, with his arms upon +the long, low desk before him, and seemed to study his book. + +By and by attention ceased from him, and the accustomed school murmur +rose upon the dull air once more. Presently the boy began to steal +furtive glances at the girl. She observed it, "made a mouth" at him and +gave him the back of her head for the space of a minute. When she +cautiously faced around again, a peach lay before her. She thrust it +away. Tom gently put it back. She thrust it away again, but with less +animosity. Tom patiently returned it to its place. Then she let it +remain. Tom scrawled on his slate, "Please take it--I got more." The +girl glanced at the words, but made no sign. Now the boy began to draw +something on the slate, hiding his work with his left hand. For a time +the girl refused to notice; but her human curiosity presently began to +manifest itself by hardly perceptible signs. The boy worked on, +apparently unconscious. The girl made a sort of noncommittal attempt to +see, but the boy did not betray that he was aware of it. At last she +gave in and hesitatingly whispered: + +"Let me see it." + +Tom partly uncovered a dismal caricature of a house with two gable +ends to it and a corkscrew of smoke issuing from the chimney. Then the +girl's interest began to fasten itself upon the work and she forgot +everything else. When it was finished, she gazed a moment, then +whispered: + +"It's nice--make a man." + +The artist erected a man in the front yard, that resembled a derrick. +He could have stepped over the house; but the girl was not +hypercritical; she was satisfied with the monster, and whispered: + +"It's a beautiful man--now make me coming along." + +Tom drew an hour-glass with a full moon and straw limbs to it and +armed the spreading fingers with a portentous fan. The girl said: + +"It's ever so nice--I wish I could draw." + +"It's easy," whispered Tom, "I'll learn you." + +"Oh, will you? When?" + +"At noon. Do you go home to dinner?" + +"I'll stay if you will." + +"Good--that's a whack. What's your name?" + +"Becky Thatcher. What's yours? Oh, I know. It's Thomas Sawyer." + +"That's the name they lick me by. I'm Tom when I'm good. You call me +Tom, will you?" + +"Yes." + +Now Tom began to scrawl something on the slate, hiding the words from +the girl. But she was not backward this time. She begged to see. Tom +said: + +"Oh, it ain't anything." + +"Yes it is." + +"No it ain't. You don't want to see." + +"Yes I do, indeed I do. Please let me." + +"You'll tell." + +"No I won't--deed and deed and double deed won't." + +"You won't tell anybody at all? Ever, as long as you live?" + +"No, I won't ever tell ANYbody. Now let me." + +"Oh, YOU don't want to see!" + +"Now that you treat me so, I WILL see." And she put her small hand +upon his and a little scuffle ensued, Tom pretending to resist in +earnest but letting his hand slip by degrees till these words were +revealed: "I LOVE YOU." + +"Oh, you bad thing!" And she hit his hand a smart rap, but reddened +and looked pleased, nevertheless. + +Just at this juncture the boy felt a slow, fateful grip closing on his +ear, and a steady lifting impulse. In that wise he was borne across the +house and deposited in his own seat, under a peppering fire of giggles +from the whole school. Then the master stood over him during a few +awful moments, and finally moved away to his throne without saying a +word. But although Tom's ear tingled, his heart was jubilant. + +As the school quieted down Tom made an honest effort to study, but the +turmoil within him was too great. In turn he took his place in the +reading class and made a botch of it; then in the geography class and +turned lakes into mountains, mountains into rivers, and rivers into +continents, till chaos was come again; then in the spelling class, and +got "turned down," by a succession of mere baby words, till he brought +up at the foot and yielded up the pewter medal which he had worn with +ostentation for months. + + + +CHAPTER VII + +THE harder Tom tried to fasten his mind on his book, the more his +ideas wandered. So at last, with a sigh and a yawn, he gave it up. It +seemed to him that the noon recess would never come. The air was +utterly dead. There was not a breath stirring. It was the sleepiest of +sleepy days. The drowsing murmur of the five and twenty studying +scholars soothed the soul like the spell that is in the murmur of bees. +Away off in the flaming sunshine, Cardiff Hill lifted its soft green +sides through a shimmering veil of heat, tinted with the purple of +distance; a few birds floated on lazy wing high in the air; no other +living thing was visible but some cows, and they were asleep. Tom's +heart ached to be free, or else to have something of interest to do to +pass the dreary time. His hand wandered into his pocket and his face +lit up with a glow of gratitude that was prayer, though he did not know +it. Then furtively the percussion-cap box came out. He released the +tick and put him on the long flat desk. The creature probably glowed +with a gratitude that amounted to prayer, too, at this moment, but it +was premature: for when he started thankfully to travel off, Tom turned +him aside with a pin and made him take a new direction. + +Tom's bosom friend sat next him, suffering just as Tom had been, and +now he was deeply and gratefully interested in this entertainment in an +instant. This bosom friend was Joe Harper. The two boys were sworn +friends all the week, and embattled enemies on Saturdays. Joe took a +pin out of his lapel and began to assist in exercising the prisoner. +The sport grew in interest momently. Soon Tom said that they were +interfering with each other, and neither getting the fullest benefit of +the tick. So he put Joe's slate on the desk and drew a line down the +middle of it from top to bottom. + +"Now," said he, "as long as he is on your side you can stir him up and +I'll let him alone; but if you let him get away and get on my side, +you're to leave him alone as long as I can keep him from crossing over." + +"All right, go ahead; start him up." + +The tick escaped from Tom, presently, and crossed the equator. Joe +harassed him awhile, and then he got away and crossed back again. This +change of base occurred often. While one boy was worrying the tick with +absorbing interest, the other would look on with interest as strong, +the two heads bowed together over the slate, and the two souls dead to +all things else. At last luck seemed to settle and abide with Joe. The +tick tried this, that, and the other course, and got as excited and as +anxious as the boys themselves, but time and again just as he would +have victory in his very grasp, so to speak, and Tom's fingers would be +twitching to begin, Joe's pin would deftly head him off, and keep +possession. At last Tom could stand it no longer. The temptation was +too strong. So he reached out and lent a hand with his pin. Joe was +angry in a moment. Said he: + +"Tom, you let him alone." + +"I only just want to stir him up a little, Joe." + +"No, sir, it ain't fair; you just let him alone." + +"Blame it, I ain't going to stir him much." + +"Let him alone, I tell you." + +"I won't!" + +"You shall--he's on my side of the line." + +"Look here, Joe Harper, whose is that tick?" + +"I don't care whose tick he is--he's on my side of the line, and you +sha'n't touch him." + +"Well, I'll just bet I will, though. He's my tick and I'll do what I +blame please with him, or die!" + +A tremendous whack came down on Tom's shoulders, and its duplicate on +Joe's; and for the space of two minutes the dust continued to fly from +the two jackets and the whole school to enjoy it. The boys had been too +absorbed to notice the hush that had stolen upon the school awhile +before when the master came tiptoeing down the room and stood over +them. He had contemplated a good part of the performance before he +contributed his bit of variety to it. + +When school broke up at noon, Tom flew to Becky Thatcher, and +whispered in her ear: + +"Put on your bonnet and let on you're going home; and when you get to +the corner, give the rest of 'em the slip, and turn down through the +lane and come back. I'll go the other way and come it over 'em the same +way." + +So the one went off with one group of scholars, and the other with +another. In a little while the two met at the bottom of the lane, and +when they reached the school they had it all to themselves. Then they +sat together, with a slate before them, and Tom gave Becky the pencil +and held her hand in his, guiding it, and so created another surprising +house. When the interest in art began to wane, the two fell to talking. +Tom was swimming in bliss. He said: + +"Do you love rats?" + +"No! I hate them!" + +"Well, I do, too--LIVE ones. But I mean dead ones, to swing round your +head with a string." + +"No, I don't care for rats much, anyway. What I like is chewing-gum." + +"Oh, I should say so! I wish I had some now." + +"Do you? I've got some. I'll let you chew it awhile, but you must give +it back to me." + +That was agreeable, so they chewed it turn about, and dangled their +legs against the bench in excess of contentment. + +"Was you ever at a circus?" said Tom. + +"Yes, and my pa's going to take me again some time, if I'm good." + +"I been to the circus three or four times--lots of times. Church ain't +shucks to a circus. There's things going on at a circus all the time. +I'm going to be a clown in a circus when I grow up." + +"Oh, are you! That will be nice. They're so lovely, all spotted up." + +"Yes, that's so. And they get slathers of money--most a dollar a day, +Ben Rogers says. Say, Becky, was you ever engaged?" + +"What's that?" + +"Why, engaged to be married." + +"No." + +"Would you like to?" + +"I reckon so. I don't know. What is it like?" + +"Like? Why it ain't like anything. You only just tell a boy you won't +ever have anybody but him, ever ever ever, and then you kiss and that's +all. Anybody can do it." + +"Kiss? What do you kiss for?" + +"Why, that, you know, is to--well, they always do that." + +"Everybody?" + +"Why, yes, everybody that's in love with each other. Do you remember +what I wrote on the slate?" + +"Ye--yes." + +"What was it?" + +"I sha'n't tell you." + +"Shall I tell YOU?" + +"Ye--yes--but some other time." + +"No, now." + +"No, not now--to-morrow." + +"Oh, no, NOW. Please, Becky--I'll whisper it, I'll whisper it ever so +easy." + +Becky hesitating, Tom took silence for consent, and passed his arm +about her waist and whispered the tale ever so softly, with his mouth +close to her ear. And then he added: + +"Now you whisper it to me--just the same." + +She resisted, for a while, and then said: + +"You turn your face away so you can't see, and then I will. But you +mustn't ever tell anybody--WILL you, Tom? Now you won't, WILL you?" + +"No, indeed, indeed I won't. Now, Becky." + +He turned his face away. She bent timidly around till her breath +stirred his curls and whispered, "I--love--you!" + +Then she sprang away and ran around and around the desks and benches, +with Tom after her, and took refuge in a corner at last, with her +little white apron to her face. Tom clasped her about her neck and +pleaded: + +"Now, Becky, it's all done--all over but the kiss. Don't you be afraid +of that--it ain't anything at all. Please, Becky." And he tugged at her +apron and the hands. + +By and by she gave up, and let her hands drop; her face, all glowing +with the struggle, came up and submitted. Tom kissed the red lips and +said: + +"Now it's all done, Becky. And always after this, you know, you ain't +ever to love anybody but me, and you ain't ever to marry anybody but +me, ever never and forever. Will you?" + +"No, I'll never love anybody but you, Tom, and I'll never marry +anybody but you--and you ain't to ever marry anybody but me, either." + +"Certainly. Of course. That's PART of it. And always coming to school +or when we're going home, you're to walk with me, when there ain't +anybody looking--and you choose me and I choose you at parties, because +that's the way you do when you're engaged." + +"It's so nice. I never heard of it before." + +"Oh, it's ever so gay! Why, me and Amy Lawrence--" + +The big eyes told Tom his blunder and he stopped, confused. + +"Oh, Tom! Then I ain't the first you've ever been engaged to!" + +The child began to cry. Tom said: + +"Oh, don't cry, Becky, I don't care for her any more." + +"Yes, you do, Tom--you know you do." + +Tom tried to put his arm about her neck, but she pushed him away and +turned her face to the wall, and went on crying. Tom tried again, with +soothing words in his mouth, and was repulsed again. Then his pride was +up, and he strode away and went outside. He stood about, restless and +uneasy, for a while, glancing at the door, every now and then, hoping +she would repent and come to find him. But she did not. Then he began +to feel badly and fear that he was in the wrong. It was a hard struggle +with him to make new advances, now, but he nerved himself to it and +entered. She was still standing back there in the corner, sobbing, with +her face to the wall. Tom's heart smote him. He went to her and stood a +moment, not knowing exactly how to proceed. Then he said hesitatingly: + +"Becky, I--I don't care for anybody but you." + +No reply--but sobs. + +"Becky"--pleadingly. "Becky, won't you say something?" + +More sobs. + +Tom got out his chiefest jewel, a brass knob from the top of an +andiron, and passed it around her so that she could see it, and said: + +"Please, Becky, won't you take it?" + +She struck it to the floor. Then Tom marched out of the house and over +the hills and far away, to return to school no more that day. Presently +Becky began to suspect. She ran to the door; he was not in sight; she +flew around to the play-yard; he was not there. Then she called: + +"Tom! Come back, Tom!" + +She listened intently, but there was no answer. She had no companions +but silence and loneliness. So she sat down to cry again and upbraid +herself; and by this time the scholars began to gather again, and she +had to hide her griefs and still her broken heart and take up the cross +of a long, dreary, aching afternoon, with none among the strangers +about her to exchange sorrows with. + + + +CHAPTER VIII + +TOM dodged hither and thither through lanes until he was well out of +the track of returning scholars, and then fell into a moody jog. He +crossed a small "branch" two or three times, because of a prevailing +juvenile superstition that to cross water baffled pursuit. Half an hour +later he was disappearing behind the Douglas mansion on the summit of +Cardiff Hill, and the schoolhouse was hardly distinguishable away off +in the valley behind him. He entered a dense wood, picked his pathless +way to the centre of it, and sat down on a mossy spot under a spreading +oak. There was not even a zephyr stirring; the dead noonday heat had +even stilled the songs of the birds; nature lay in a trance that was +broken by no sound but the occasional far-off hammering of a +woodpecker, and this seemed to render the pervading silence and sense +of loneliness the more profound. The boy's soul was steeped in +melancholy; his feelings were in happy accord with his surroundings. He +sat long with his elbows on his knees and his chin in his hands, +meditating. It seemed to him that life was but a trouble, at best, and +he more than half envied Jimmy Hodges, so lately released; it must be +very peaceful, he thought, to lie and slumber and dream forever and +ever, with the wind whispering through the trees and caressing the +grass and the flowers over the grave, and nothing to bother and grieve +about, ever any more. If he only had a clean Sunday-school record he +could be willing to go, and be done with it all. Now as to this girl. +What had he done? Nothing. He had meant the best in the world, and been +treated like a dog--like a very dog. She would be sorry some day--maybe +when it was too late. Ah, if he could only die TEMPORARILY! + +But the elastic heart of youth cannot be compressed into one +constrained shape long at a time. Tom presently began to drift +insensibly back into the concerns of this life again. What if he turned +his back, now, and disappeared mysteriously? What if he went away--ever +so far away, into unknown countries beyond the seas--and never came +back any more! How would she feel then! The idea of being a clown +recurred to him now, only to fill him with disgust. For frivolity and +jokes and spotted tights were an offense, when they intruded themselves +upon a spirit that was exalted into the vague august realm of the +romantic. No, he would be a soldier, and return after long years, all +war-worn and illustrious. No--better still, he would join the Indians, +and hunt buffaloes and go on the warpath in the mountain ranges and the +trackless great plains of the Far West, and away in the future come +back a great chief, bristling with feathers, hideous with paint, and +prance into Sunday-school, some drowsy summer morning, with a +bloodcurdling war-whoop, and sear the eyeballs of all his companions +with unappeasable envy. But no, there was something gaudier even than +this. He would be a pirate! That was it! NOW his future lay plain +before him, and glowing with unimaginable splendor. How his name would +fill the world, and make people shudder! How gloriously he would go +plowing the dancing seas, in his long, low, black-hulled racer, the +Spirit of the Storm, with his grisly flag flying at the fore! And at +the zenith of his fame, how he would suddenly appear at the old village +and stalk into church, brown and weather-beaten, in his black velvet +doublet and trunks, his great jack-boots, his crimson sash, his belt +bristling with horse-pistols, his crime-rusted cutlass at his side, his +slouch hat with waving plumes, his black flag unfurled, with the skull +and crossbones on it, and hear with swelling ecstasy the whisperings, +"It's Tom Sawyer the Pirate!--the Black Avenger of the Spanish Main!" + +Yes, it was settled; his career was determined. He would run away from +home and enter upon it. He would start the very next morning. Therefore +he must now begin to get ready. He would collect his resources +together. He went to a rotten log near at hand and began to dig under +one end of it with his Barlow knife. He soon struck wood that sounded +hollow. He put his hand there and uttered this incantation impressively: + +"What hasn't come here, come! What's here, stay here!" + +Then he scraped away the dirt, and exposed a pine shingle. He took it +up and disclosed a shapely little treasure-house whose bottom and sides +were of shingles. In it lay a marble. Tom's astonishment was boundless! +He scratched his head with a perplexed air, and said: + +"Well, that beats anything!" + +Then he tossed the marble away pettishly, and stood cogitating. The +truth was, that a superstition of his had failed, here, which he and +all his comrades had always looked upon as infallible. If you buried a +marble with certain necessary incantations, and left it alone a +fortnight, and then opened the place with the incantation he had just +used, you would find that all the marbles you had ever lost had +gathered themselves together there, meantime, no matter how widely they +had been separated. But now, this thing had actually and unquestionably +failed. Tom's whole structure of faith was shaken to its foundations. +He had many a time heard of this thing succeeding but never of its +failing before. It did not occur to him that he had tried it several +times before, himself, but could never find the hiding-places +afterward. He puzzled over the matter some time, and finally decided +that some witch had interfered and broken the charm. He thought he +would satisfy himself on that point; so he searched around till he +found a small sandy spot with a little funnel-shaped depression in it. +He laid himself down and put his mouth close to this depression and +called-- + +"Doodle-bug, doodle-bug, tell me what I want to know! Doodle-bug, +doodle-bug, tell me what I want to know!" + +The sand began to work, and presently a small black bug appeared for a +second and then darted under again in a fright. + +"He dasn't tell! So it WAS a witch that done it. I just knowed it." + +He well knew the futility of trying to contend against witches, so he +gave up discouraged. But it occurred to him that he might as well have +the marble he had just thrown away, and therefore he went and made a +patient search for it. But he could not find it. Now he went back to +his treasure-house and carefully placed himself just as he had been +standing when he tossed the marble away; then he took another marble +from his pocket and tossed it in the same way, saying: + +"Brother, go find your brother!" + +He watched where it stopped, and went there and looked. But it must +have fallen short or gone too far; so he tried twice more. The last +repetition was successful. The two marbles lay within a foot of each +other. + +Just here the blast of a toy tin trumpet came faintly down the green +aisles of the forest. Tom flung off his jacket and trousers, turned a +suspender into a belt, raked away some brush behind the rotten log, +disclosing a rude bow and arrow, a lath sword and a tin trumpet, and in +a moment had seized these things and bounded away, barelegged, with +fluttering shirt. He presently halted under a great elm, blew an +answering blast, and then began to tiptoe and look warily out, this way +and that. He said cautiously--to an imaginary company: + +"Hold, my merry men! Keep hid till I blow." + +Now appeared Joe Harper, as airily clad and elaborately armed as Tom. +Tom called: + +"Hold! Who comes here into Sherwood Forest without my pass?" + +"Guy of Guisborne wants no man's pass. Who art thou that--that--" + +"Dares to hold such language," said Tom, prompting--for they talked +"by the book," from memory. + +"Who art thou that dares to hold such language?" + +"I, indeed! I am Robin Hood, as thy caitiff carcase soon shall know." + +"Then art thou indeed that famous outlaw? Right gladly will I dispute +with thee the passes of the merry wood. Have at thee!" + +They took their lath swords, dumped their other traps on the ground, +struck a fencing attitude, foot to foot, and began a grave, careful +combat, "two up and two down." Presently Tom said: + +"Now, if you've got the hang, go it lively!" + +So they "went it lively," panting and perspiring with the work. By and +by Tom shouted: + +"Fall! fall! Why don't you fall?" + +"I sha'n't! Why don't you fall yourself? You're getting the worst of +it." + +"Why, that ain't anything. I can't fall; that ain't the way it is in +the book. The book says, 'Then with one back-handed stroke he slew poor +Guy of Guisborne.' You're to turn around and let me hit you in the +back." + +There was no getting around the authorities, so Joe turned, received +the whack and fell. + +"Now," said Joe, getting up, "you got to let me kill YOU. That's fair." + +"Why, I can't do that, it ain't in the book." + +"Well, it's blamed mean--that's all." + +"Well, say, Joe, you can be Friar Tuck or Much the miller's son, and +lam me with a quarter-staff; or I'll be the Sheriff of Nottingham and +you be Robin Hood a little while and kill me." + +This was satisfactory, and so these adventures were carried out. Then +Tom became Robin Hood again, and was allowed by the treacherous nun to +bleed his strength away through his neglected wound. And at last Joe, +representing a whole tribe of weeping outlaws, dragged him sadly forth, +gave his bow into his feeble hands, and Tom said, "Where this arrow +falls, there bury poor Robin Hood under the greenwood tree." Then he +shot the arrow and fell back and would have died, but he lit on a +nettle and sprang up too gaily for a corpse. + +The boys dressed themselves, hid their accoutrements, and went off +grieving that there were no outlaws any more, and wondering what modern +civilization could claim to have done to compensate for their loss. +They said they would rather be outlaws a year in Sherwood Forest than +President of the United States forever. + + + +CHAPTER IX + +AT half-past nine, that night, Tom and Sid were sent to bed, as usual. +They said their prayers, and Sid was soon asleep. Tom lay awake and +waited, in restless impatience. When it seemed to him that it must be +nearly daylight, he heard the clock strike ten! This was despair. He +would have tossed and fidgeted, as his nerves demanded, but he was +afraid he might wake Sid. So he lay still, and stared up into the dark. +Everything was dismally still. By and by, out of the stillness, little, +scarcely perceptible noises began to emphasize themselves. The ticking +of the clock began to bring itself into notice. Old beams began to +crack mysteriously. The stairs creaked faintly. Evidently spirits were +abroad. A measured, muffled snore issued from Aunt Polly's chamber. And +now the tiresome chirping of a cricket that no human ingenuity could +locate, began. Next the ghastly ticking of a deathwatch in the wall at +the bed's head made Tom shudder--it meant that somebody's days were +numbered. Then the howl of a far-off dog rose on the night air, and was +answered by a fainter howl from a remoter distance. Tom was in an +agony. At last he was satisfied that time had ceased and eternity +begun; he began to doze, in spite of himself; the clock chimed eleven, +but he did not hear it. And then there came, mingling with his +half-formed dreams, a most melancholy caterwauling. The raising of a +neighboring window disturbed him. A cry of "Scat! you devil!" and the +crash of an empty bottle against the back of his aunt's woodshed +brought him wide awake, and a single minute later he was dressed and +out of the window and creeping along the roof of the "ell" on all +fours. He "meow'd" with caution once or twice, as he went; then jumped +to the roof of the woodshed and thence to the ground. Huckleberry Finn +was there, with his dead cat. The boys moved off and disappeared in the +gloom. At the end of half an hour they were wading through the tall +grass of the graveyard. + +It was a graveyard of the old-fashioned Western kind. It was on a +hill, about a mile and a half from the village. It had a crazy board +fence around it, which leaned inward in places, and outward the rest of +the time, but stood upright nowhere. Grass and weeds grew rank over the +whole cemetery. All the old graves were sunken in, there was not a +tombstone on the place; round-topped, worm-eaten boards staggered over +the graves, leaning for support and finding none. "Sacred to the memory +of" So-and-So had been painted on them once, but it could no longer +have been read, on the most of them, now, even if there had been light. + +A faint wind moaned through the trees, and Tom feared it might be the +spirits of the dead, complaining at being disturbed. The boys talked +little, and only under their breath, for the time and the place and the +pervading solemnity and silence oppressed their spirits. They found the +sharp new heap they were seeking, and ensconced themselves within the +protection of three great elms that grew in a bunch within a few feet +of the grave. + +Then they waited in silence for what seemed a long time. The hooting +of a distant owl was all the sound that troubled the dead stillness. +Tom's reflections grew oppressive. He must force some talk. So he said +in a whisper: + +"Hucky, do you believe the dead people like it for us to be here?" + +Huckleberry whispered: + +"I wisht I knowed. It's awful solemn like, AIN'T it?" + +"I bet it is." + +There was a considerable pause, while the boys canvassed this matter +inwardly. Then Tom whispered: + +"Say, Hucky--do you reckon Hoss Williams hears us talking?" + +"O' course he does. Least his sperrit does." + +Tom, after a pause: + +"I wish I'd said Mister Williams. But I never meant any harm. +Everybody calls him Hoss." + +"A body can't be too partic'lar how they talk 'bout these-yer dead +people, Tom." + +This was a damper, and conversation died again. + +Presently Tom seized his comrade's arm and said: + +"Sh!" + +"What is it, Tom?" And the two clung together with beating hearts. + +"Sh! There 'tis again! Didn't you hear it?" + +"I--" + +"There! Now you hear it." + +"Lord, Tom, they're coming! They're coming, sure. What'll we do?" + +"I dono. Think they'll see us?" + +"Oh, Tom, they can see in the dark, same as cats. I wisht I hadn't +come." + +"Oh, don't be afeard. I don't believe they'll bother us. We ain't +doing any harm. If we keep perfectly still, maybe they won't notice us +at all." + +"I'll try to, Tom, but, Lord, I'm all of a shiver." + +"Listen!" + +The boys bent their heads together and scarcely breathed. A muffled +sound of voices floated up from the far end of the graveyard. + +"Look! See there!" whispered Tom. "What is it?" + +"It's devil-fire. Oh, Tom, this is awful." + +Some vague figures approached through the gloom, swinging an +old-fashioned tin lantern that freckled the ground with innumerable +little spangles of light. Presently Huckleberry whispered with a +shudder: + +"It's the devils sure enough. Three of 'em! Lordy, Tom, we're goners! +Can you pray?" + +"I'll try, but don't you be afeard. They ain't going to hurt us. 'Now +I lay me down to sleep, I--'" + +"Sh!" + +"What is it, Huck?" + +"They're HUMANS! One of 'em is, anyway. One of 'em's old Muff Potter's +voice." + +"No--'tain't so, is it?" + +"I bet I know it. Don't you stir nor budge. He ain't sharp enough to +notice us. Drunk, the same as usual, likely--blamed old rip!" + +"All right, I'll keep still. Now they're stuck. Can't find it. Here +they come again. Now they're hot. Cold again. Hot again. Red hot! +They're p'inted right, this time. Say, Huck, I know another o' them +voices; it's Injun Joe." + +"That's so--that murderin' half-breed! I'd druther they was devils a +dern sight. What kin they be up to?" + +The whisper died wholly out, now, for the three men had reached the +grave and stood within a few feet of the boys' hiding-place. + +"Here it is," said the third voice; and the owner of it held the +lantern up and revealed the face of young Doctor Robinson. + +Potter and Injun Joe were carrying a handbarrow with a rope and a +couple of shovels on it. They cast down their load and began to open +the grave. The doctor put the lantern at the head of the grave and came +and sat down with his back against one of the elm trees. He was so +close the boys could have touched him. + +"Hurry, men!" he said, in a low voice; "the moon might come out at any +moment." + +They growled a response and went on digging. For some time there was +no noise but the grating sound of the spades discharging their freight +of mould and gravel. It was very monotonous. Finally a spade struck +upon the coffin with a dull woody accent, and within another minute or +two the men had hoisted it out on the ground. They pried off the lid +with their shovels, got out the body and dumped it rudely on the +ground. The moon drifted from behind the clouds and exposed the pallid +face. The barrow was got ready and the corpse placed on it, covered +with a blanket, and bound to its place with the rope. Potter took out a +large spring-knife and cut off the dangling end of the rope and then +said: + +"Now the cussed thing's ready, Sawbones, and you'll just out with +another five, or here she stays." + +"That's the talk!" said Injun Joe. + +"Look here, what does this mean?" said the doctor. "You required your +pay in advance, and I've paid you." + +"Yes, and you done more than that," said Injun Joe, approaching the +doctor, who was now standing. "Five years ago you drove me away from +your father's kitchen one night, when I come to ask for something to +eat, and you said I warn't there for any good; and when I swore I'd get +even with you if it took a hundred years, your father had me jailed for +a vagrant. Did you think I'd forget? The Injun blood ain't in me for +nothing. And now I've GOT you, and you got to SETTLE, you know!" + +He was threatening the doctor, with his fist in his face, by this +time. The doctor struck out suddenly and stretched the ruffian on the +ground. Potter dropped his knife, and exclaimed: + +"Here, now, don't you hit my pard!" and the next moment he had +grappled with the doctor and the two were struggling with might and +main, trampling the grass and tearing the ground with their heels. +Injun Joe sprang to his feet, his eyes flaming with passion, snatched +up Potter's knife, and went creeping, catlike and stooping, round and +round about the combatants, seeking an opportunity. All at once the +doctor flung himself free, seized the heavy headboard of Williams' +grave and felled Potter to the earth with it--and in the same instant +the half-breed saw his chance and drove the knife to the hilt in the +young man's breast. He reeled and fell partly upon Potter, flooding him +with his blood, and in the same moment the clouds blotted out the +dreadful spectacle and the two frightened boys went speeding away in +the dark. + +Presently, when the moon emerged again, Injun Joe was standing over +the two forms, contemplating them. The doctor murmured inarticulately, +gave a long gasp or two and was still. The half-breed muttered: + +"THAT score is settled--damn you." + +Then he robbed the body. After which he put the fatal knife in +Potter's open right hand, and sat down on the dismantled coffin. Three +--four--five minutes passed, and then Potter began to stir and moan. His +hand closed upon the knife; he raised it, glanced at it, and let it +fall, with a shudder. Then he sat up, pushing the body from him, and +gazed at it, and then around him, confusedly. His eyes met Joe's. + +"Lord, how is this, Joe?" he said. + +"It's a dirty business," said Joe, without moving. + +"What did you do it for?" + +"I! I never done it!" + +"Look here! That kind of talk won't wash." + +Potter trembled and grew white. + +"I thought I'd got sober. I'd no business to drink to-night. But it's +in my head yet--worse'n when we started here. I'm all in a muddle; +can't recollect anything of it, hardly. Tell me, Joe--HONEST, now, old +feller--did I do it? Joe, I never meant to--'pon my soul and honor, I +never meant to, Joe. Tell me how it was, Joe. Oh, it's awful--and him +so young and promising." + +"Why, you two was scuffling, and he fetched you one with the headboard +and you fell flat; and then up you come, all reeling and staggering +like, and snatched the knife and jammed it into him, just as he fetched +you another awful clip--and here you've laid, as dead as a wedge til +now." + +"Oh, I didn't know what I was a-doing. I wish I may die this minute if +I did. It was all on account of the whiskey and the excitement, I +reckon. I never used a weepon in my life before, Joe. I've fought, but +never with weepons. They'll all say that. Joe, don't tell! Say you +won't tell, Joe--that's a good feller. I always liked you, Joe, and +stood up for you, too. Don't you remember? You WON'T tell, WILL you, +Joe?" And the poor creature dropped on his knees before the stolid +murderer, and clasped his appealing hands. + +"No, you've always been fair and square with me, Muff Potter, and I +won't go back on you. There, now, that's as fair as a man can say." + +"Oh, Joe, you're an angel. I'll bless you for this the longest day I +live." And Potter began to cry. + +"Come, now, that's enough of that. This ain't any time for blubbering. +You be off yonder way and I'll go this. Move, now, and don't leave any +tracks behind you." + +Potter started on a trot that quickly increased to a run. The +half-breed stood looking after him. He muttered: + +"If he's as much stunned with the lick and fuddled with the rum as he +had the look of being, he won't think of the knife till he's gone so +far he'll be afraid to come back after it to such a place by himself +--chicken-heart!" + +Two or three minutes later the murdered man, the blanketed corpse, the +lidless coffin, and the open grave were under no inspection but the +moon's. The stillness was complete again, too. + + + +CHAPTER X + +THE two boys flew on and on, toward the village, speechless with +horror. They glanced backward over their shoulders from time to time, +apprehensively, as if they feared they might be followed. Every stump +that started up in their path seemed a man and an enemy, and made them +catch their breath; and as they sped by some outlying cottages that lay +near the village, the barking of the aroused watch-dogs seemed to give +wings to their feet. + +"If we can only get to the old tannery before we break down!" +whispered Tom, in short catches between breaths. "I can't stand it much +longer." + +Huckleberry's hard pantings were his only reply, and the boys fixed +their eyes on the goal of their hopes and bent to their work to win it. +They gained steadily on it, and at last, breast to breast, they burst +through the open door and fell grateful and exhausted in the sheltering +shadows beyond. By and by their pulses slowed down, and Tom whispered: + +"Huckleberry, what do you reckon'll come of this?" + +"If Doctor Robinson dies, I reckon hanging'll come of it." + +"Do you though?" + +"Why, I KNOW it, Tom." + +Tom thought a while, then he said: + +"Who'll tell? We?" + +"What are you talking about? S'pose something happened and Injun Joe +DIDN'T hang? Why, he'd kill us some time or other, just as dead sure as +we're a laying here." + +"That's just what I was thinking to myself, Huck." + +"If anybody tells, let Muff Potter do it, if he's fool enough. He's +generally drunk enough." + +Tom said nothing--went on thinking. Presently he whispered: + +"Huck, Muff Potter don't know it. How can he tell?" + +"What's the reason he don't know it?" + +"Because he'd just got that whack when Injun Joe done it. D'you reckon +he could see anything? D'you reckon he knowed anything?" + +"By hokey, that's so, Tom!" + +"And besides, look-a-here--maybe that whack done for HIM!" + +"No, 'taint likely, Tom. He had liquor in him; I could see that; and +besides, he always has. Well, when pap's full, you might take and belt +him over the head with a church and you couldn't phase him. He says so, +his own self. So it's the same with Muff Potter, of course. But if a +man was dead sober, I reckon maybe that whack might fetch him; I dono." + +After another reflective silence, Tom said: + +"Hucky, you sure you can keep mum?" + +"Tom, we GOT to keep mum. You know that. That Injun devil wouldn't +make any more of drownding us than a couple of cats, if we was to +squeak 'bout this and they didn't hang him. Now, look-a-here, Tom, less +take and swear to one another--that's what we got to do--swear to keep +mum." + +"I'm agreed. It's the best thing. Would you just hold hands and swear +that we--" + +"Oh no, that wouldn't do for this. That's good enough for little +rubbishy common things--specially with gals, cuz THEY go back on you +anyway, and blab if they get in a huff--but there orter be writing +'bout a big thing like this. And blood." + +Tom's whole being applauded this idea. It was deep, and dark, and +awful; the hour, the circumstances, the surroundings, were in keeping +with it. He picked up a clean pine shingle that lay in the moonlight, +took a little fragment of "red keel" out of his pocket, got the moon on +his work, and painfully scrawled these lines, emphasizing each slow +down-stroke by clamping his tongue between his teeth, and letting up +the pressure on the up-strokes. [See next page.] + + "Huck Finn and + Tom Sawyer swears + they will keep mum + about This and They + wish They may Drop + down dead in Their + Tracks if They ever + Tell and Rot." + +Huckleberry was filled with admiration of Tom's facility in writing, +and the sublimity of his language. He at once took a pin from his lapel +and was going to prick his flesh, but Tom said: + +"Hold on! Don't do that. A pin's brass. It might have verdigrease on +it." + +"What's verdigrease?" + +"It's p'ison. That's what it is. You just swaller some of it once +--you'll see." + +So Tom unwound the thread from one of his needles, and each boy +pricked the ball of his thumb and squeezed out a drop of blood. In +time, after many squeezes, Tom managed to sign his initials, using the +ball of his little finger for a pen. Then he showed Huckleberry how to +make an H and an F, and the oath was complete. They buried the shingle +close to the wall, with some dismal ceremonies and incantations, and +the fetters that bound their tongues were considered to be locked and +the key thrown away. + +A figure crept stealthily through a break in the other end of the +ruined building, now, but they did not notice it. + +"Tom," whispered Huckleberry, "does this keep us from EVER telling +--ALWAYS?" + +"Of course it does. It don't make any difference WHAT happens, we got +to keep mum. We'd drop down dead--don't YOU know that?" + +"Yes, I reckon that's so." + +They continued to whisper for some little time. Presently a dog set up +a long, lugubrious howl just outside--within ten feet of them. The boys +clasped each other suddenly, in an agony of fright. + +"Which of us does he mean?" gasped Huckleberry. + +"I dono--peep through the crack. Quick!" + +"No, YOU, Tom!" + +"I can't--I can't DO it, Huck!" + +"Please, Tom. There 'tis again!" + +"Oh, lordy, I'm thankful!" whispered Tom. "I know his voice. It's Bull +Harbison." * + +[* If Mr. Harbison owned a slave named Bull, Tom would have spoken of +him as "Harbison's Bull," but a son or a dog of that name was "Bull +Harbison."] + +"Oh, that's good--I tell you, Tom, I was most scared to death; I'd a +bet anything it was a STRAY dog." + +The dog howled again. The boys' hearts sank once more. + +"Oh, my! that ain't no Bull Harbison!" whispered Huckleberry. "DO, Tom!" + +Tom, quaking with fear, yielded, and put his eye to the crack. His +whisper was hardly audible when he said: + +"Oh, Huck, IT S A STRAY DOG!" + +"Quick, Tom, quick! Who does he mean?" + +"Huck, he must mean us both--we're right together." + +"Oh, Tom, I reckon we're goners. I reckon there ain't no mistake 'bout +where I'LL go to. I been so wicked." + +"Dad fetch it! This comes of playing hookey and doing everything a +feller's told NOT to do. I might a been good, like Sid, if I'd a tried +--but no, I wouldn't, of course. But if ever I get off this time, I lay +I'll just WALLER in Sunday-schools!" And Tom began to snuffle a little. + +"YOU bad!" and Huckleberry began to snuffle too. "Consound it, Tom +Sawyer, you're just old pie, 'longside o' what I am. Oh, LORDY, lordy, +lordy, I wisht I only had half your chance." + +Tom choked off and whispered: + +"Look, Hucky, look! He's got his BACK to us!" + +Hucky looked, with joy in his heart. + +"Well, he has, by jingoes! Did he before?" + +"Yes, he did. But I, like a fool, never thought. Oh, this is bully, +you know. NOW who can he mean?" + +The howling stopped. Tom pricked up his ears. + +"Sh! What's that?" he whispered. + +"Sounds like--like hogs grunting. No--it's somebody snoring, Tom." + +"That IS it! Where 'bouts is it, Huck?" + +"I bleeve it's down at 'tother end. Sounds so, anyway. Pap used to +sleep there, sometimes, 'long with the hogs, but laws bless you, he +just lifts things when HE snores. Besides, I reckon he ain't ever +coming back to this town any more." + +The spirit of adventure rose in the boys' souls once more. + +"Hucky, do you das't to go if I lead?" + +"I don't like to, much. Tom, s'pose it's Injun Joe!" + +Tom quailed. But presently the temptation rose up strong again and the +boys agreed to try, with the understanding that they would take to +their heels if the snoring stopped. So they went tiptoeing stealthily +down, the one behind the other. When they had got to within five steps +of the snorer, Tom stepped on a stick, and it broke with a sharp snap. +The man moaned, writhed a little, and his face came into the moonlight. +It was Muff Potter. The boys' hearts had stood still, and their hopes +too, when the man moved, but their fears passed away now. They tiptoed +out, through the broken weather-boarding, and stopped at a little +distance to exchange a parting word. That long, lugubrious howl rose on +the night air again! They turned and saw the strange dog standing +within a few feet of where Potter was lying, and FACING Potter, with +his nose pointing heavenward. + +"Oh, geeminy, it's HIM!" exclaimed both boys, in a breath. + +"Say, Tom--they say a stray dog come howling around Johnny Miller's +house, 'bout midnight, as much as two weeks ago; and a whippoorwill +come in and lit on the banisters and sung, the very same evening; and +there ain't anybody dead there yet." + +"Well, I know that. And suppose there ain't. Didn't Gracie Miller fall +in the kitchen fire and burn herself terrible the very next Saturday?" + +"Yes, but she ain't DEAD. And what's more, she's getting better, too." + +"All right, you wait and see. She's a goner, just as dead sure as Muff +Potter's a goner. That's what the niggers say, and they know all about +these kind of things, Huck." + +Then they separated, cogitating. When Tom crept in at his bedroom +window the night was almost spent. He undressed with excessive caution, +and fell asleep congratulating himself that nobody knew of his +escapade. He was not aware that the gently-snoring Sid was awake, and +had been so for an hour. + +When Tom awoke, Sid was dressed and gone. There was a late look in the +light, a late sense in the atmosphere. He was startled. Why had he not +been called--persecuted till he was up, as usual? The thought filled +him with bodings. Within five minutes he was dressed and down-stairs, +feeling sore and drowsy. The family were still at table, but they had +finished breakfast. There was no voice of rebuke; but there were +averted eyes; there was a silence and an air of solemnity that struck a +chill to the culprit's heart. He sat down and tried to seem gay, but it +was up-hill work; it roused no smile, no response, and he lapsed into +silence and let his heart sink down to the depths. + +After breakfast his aunt took him aside, and Tom almost brightened in +the hope that he was going to be flogged; but it was not so. His aunt +wept over him and asked him how he could go and break her old heart so; +and finally told him to go on, and ruin himself and bring her gray +hairs with sorrow to the grave, for it was no use for her to try any +more. This was worse than a thousand whippings, and Tom's heart was +sorer now than his body. He cried, he pleaded for forgiveness, promised +to reform over and over again, and then received his dismissal, feeling +that he had won but an imperfect forgiveness and established but a +feeble confidence. + +He left the presence too miserable to even feel revengeful toward Sid; +and so the latter's prompt retreat through the back gate was +unnecessary. He moped to school gloomy and sad, and took his flogging, +along with Joe Harper, for playing hookey the day before, with the air +of one whose heart was busy with heavier woes and wholly dead to +trifles. Then he betook himself to his seat, rested his elbows on his +desk and his jaws in his hands, and stared at the wall with the stony +stare of suffering that has reached the limit and can no further go. +His elbow was pressing against some hard substance. After a long time +he slowly and sadly changed his position, and took up this object with +a sigh. It was in a paper. He unrolled it. A long, lingering, colossal +sigh followed, and his heart broke. It was his brass andiron knob! + +This final feather broke the camel's back. + + + +CHAPTER XI + +CLOSE upon the hour of noon the whole village was suddenly electrified +with the ghastly news. No need of the as yet undreamed-of telegraph; +the tale flew from man to man, from group to group, from house to +house, with little less than telegraphic speed. Of course the +schoolmaster gave holiday for that afternoon; the town would have +thought strangely of him if he had not. + +A gory knife had been found close to the murdered man, and it had been +recognized by somebody as belonging to Muff Potter--so the story ran. +And it was said that a belated citizen had come upon Potter washing +himself in the "branch" about one or two o'clock in the morning, and +that Potter had at once sneaked off--suspicious circumstances, +especially the washing which was not a habit with Potter. It was also +said that the town had been ransacked for this "murderer" (the public +are not slow in the matter of sifting evidence and arriving at a +verdict), but that he could not be found. Horsemen had departed down +all the roads in every direction, and the Sheriff "was confident" that +he would be captured before night. + +All the town was drifting toward the graveyard. Tom's heartbreak +vanished and he joined the procession, not because he would not a +thousand times rather go anywhere else, but because an awful, +unaccountable fascination drew him on. Arrived at the dreadful place, +he wormed his small body through the crowd and saw the dismal +spectacle. It seemed to him an age since he was there before. Somebody +pinched his arm. He turned, and his eyes met Huckleberry's. Then both +looked elsewhere at once, and wondered if anybody had noticed anything +in their mutual glance. But everybody was talking, and intent upon the +grisly spectacle before them. + +"Poor fellow!" "Poor young fellow!" "This ought to be a lesson to +grave robbers!" "Muff Potter'll hang for this if they catch him!" This +was the drift of remark; and the minister said, "It was a judgment; His +hand is here." + +Now Tom shivered from head to heel; for his eye fell upon the stolid +face of Injun Joe. At this moment the crowd began to sway and struggle, +and voices shouted, "It's him! it's him! he's coming himself!" + +"Who? Who?" from twenty voices. + +"Muff Potter!" + +"Hallo, he's stopped!--Look out, he's turning! Don't let him get away!" + +People in the branches of the trees over Tom's head said he wasn't +trying to get away--he only looked doubtful and perplexed. + +"Infernal impudence!" said a bystander; "wanted to come and take a +quiet look at his work, I reckon--didn't expect any company." + +The crowd fell apart, now, and the Sheriff came through, +ostentatiously leading Potter by the arm. The poor fellow's face was +haggard, and his eyes showed the fear that was upon him. When he stood +before the murdered man, he shook as with a palsy, and he put his face +in his hands and burst into tears. + +"I didn't do it, friends," he sobbed; "'pon my word and honor I never +done it." + +"Who's accused you?" shouted a voice. + +This shot seemed to carry home. Potter lifted his face and looked +around him with a pathetic hopelessness in his eyes. He saw Injun Joe, +and exclaimed: + +"Oh, Injun Joe, you promised me you'd never--" + +"Is that your knife?" and it was thrust before him by the Sheriff. + +Potter would have fallen if they had not caught him and eased him to +the ground. Then he said: + +"Something told me 't if I didn't come back and get--" He shuddered; +then waved his nerveless hand with a vanquished gesture and said, "Tell +'em, Joe, tell 'em--it ain't any use any more." + +Then Huckleberry and Tom stood dumb and staring, and heard the +stony-hearted liar reel off his serene statement, they expecting every +moment that the clear sky would deliver God's lightnings upon his head, +and wondering to see how long the stroke was delayed. And when he had +finished and still stood alive and whole, their wavering impulse to +break their oath and save the poor betrayed prisoner's life faded and +vanished away, for plainly this miscreant had sold himself to Satan and +it would be fatal to meddle with the property of such a power as that. + +"Why didn't you leave? What did you want to come here for?" somebody +said. + +"I couldn't help it--I couldn't help it," Potter moaned. "I wanted to +run away, but I couldn't seem to come anywhere but here." And he fell +to sobbing again. + +Injun Joe repeated his statement, just as calmly, a few minutes +afterward on the inquest, under oath; and the boys, seeing that the +lightnings were still withheld, were confirmed in their belief that Joe +had sold himself to the devil. He was now become, to them, the most +balefully interesting object they had ever looked upon, and they could +not take their fascinated eyes from his face. + +They inwardly resolved to watch him nights, when opportunity should +offer, in the hope of getting a glimpse of his dread master. + +Injun Joe helped to raise the body of the murdered man and put it in a +wagon for removal; and it was whispered through the shuddering crowd +that the wound bled a little! The boys thought that this happy +circumstance would turn suspicion in the right direction; but they were +disappointed, for more than one villager remarked: + +"It was within three feet of Muff Potter when it done it." + +Tom's fearful secret and gnawing conscience disturbed his sleep for as +much as a week after this; and at breakfast one morning Sid said: + +"Tom, you pitch around and talk in your sleep so much that you keep me +awake half the time." + +Tom blanched and dropped his eyes. + +"It's a bad sign," said Aunt Polly, gravely. "What you got on your +mind, Tom?" + +"Nothing. Nothing 't I know of." But the boy's hand shook so that he +spilled his coffee. + +"And you do talk such stuff," Sid said. "Last night you said, 'It's +blood, it's blood, that's what it is!' You said that over and over. And +you said, 'Don't torment me so--I'll tell!' Tell WHAT? What is it +you'll tell?" + +Everything was swimming before Tom. There is no telling what might +have happened, now, but luckily the concern passed out of Aunt Polly's +face and she came to Tom's relief without knowing it. She said: + +"Sho! It's that dreadful murder. I dream about it most every night +myself. Sometimes I dream it's me that done it." + +Mary said she had been affected much the same way. Sid seemed +satisfied. Tom got out of the presence as quick as he plausibly could, +and after that he complained of toothache for a week, and tied up his +jaws every night. He never knew that Sid lay nightly watching, and +frequently slipped the bandage free and then leaned on his elbow +listening a good while at a time, and afterward slipped the bandage +back to its place again. Tom's distress of mind wore off gradually and +the toothache grew irksome and was discarded. If Sid really managed to +make anything out of Tom's disjointed mutterings, he kept it to himself. + +It seemed to Tom that his schoolmates never would get done holding +inquests on dead cats, and thus keeping his trouble present to his +mind. Sid noticed that Tom never was coroner at one of these inquiries, +though it had been his habit to take the lead in all new enterprises; +he noticed, too, that Tom never acted as a witness--and that was +strange; and Sid did not overlook the fact that Tom even showed a +marked aversion to these inquests, and always avoided them when he +could. Sid marvelled, but said nothing. However, even inquests went out +of vogue at last, and ceased to torture Tom's conscience. + +Every day or two, during this time of sorrow, Tom watched his +opportunity and went to the little grated jail-window and smuggled such +small comforts through to the "murderer" as he could get hold of. The +jail was a trifling little brick den that stood in a marsh at the edge +of the village, and no guards were afforded for it; indeed, it was +seldom occupied. These offerings greatly helped to ease Tom's +conscience. + +The villagers had a strong desire to tar-and-feather Injun Joe and +ride him on a rail, for body-snatching, but so formidable was his +character that nobody could be found who was willing to take the lead +in the matter, so it was dropped. He had been careful to begin both of +his inquest-statements with the fight, without confessing the +grave-robbery that preceded it; therefore it was deemed wisest not +to try the case in the courts at present. + + + +CHAPTER XII + +ONE of the reasons why Tom's mind had drifted away from its secret +troubles was, that it had found a new and weighty matter to interest +itself about. Becky Thatcher had stopped coming to school. Tom had +struggled with his pride a few days, and tried to "whistle her down the +wind," but failed. He began to find himself hanging around her father's +house, nights, and feeling very miserable. She was ill. What if she +should die! There was distraction in the thought. He no longer took an +interest in war, nor even in piracy. The charm of life was gone; there +was nothing but dreariness left. He put his hoop away, and his bat; +there was no joy in them any more. His aunt was concerned. She began to +try all manner of remedies on him. She was one of those people who are +infatuated with patent medicines and all new-fangled methods of +producing health or mending it. She was an inveterate experimenter in +these things. When something fresh in this line came out she was in a +fever, right away, to try it; not on herself, for she was never ailing, +but on anybody else that came handy. She was a subscriber for all the +"Health" periodicals and phrenological frauds; and the solemn ignorance +they were inflated with was breath to her nostrils. All the "rot" they +contained about ventilation, and how to go to bed, and how to get up, +and what to eat, and what to drink, and how much exercise to take, and +what frame of mind to keep one's self in, and what sort of clothing to +wear, was all gospel to her, and she never observed that her +health-journals of the current month customarily upset everything they +had recommended the month before. She was as simple-hearted and honest +as the day was long, and so she was an easy victim. She gathered +together her quack periodicals and her quack medicines, and thus armed +with death, went about on her pale horse, metaphorically speaking, with +"hell following after." But she never suspected that she was not an +angel of healing and the balm of Gilead in disguise, to the suffering +neighbors. + +The water treatment was new, now, and Tom's low condition was a +windfall to her. She had him out at daylight every morning, stood him +up in the woodshed and drowned him with a deluge of cold water; then +she scrubbed him down with a towel like a file, and so brought him to; +then she rolled him up in a wet sheet and put him away under blankets +till she sweated his soul clean and "the yellow stains of it came +through his pores"--as Tom said. + +Yet notwithstanding all this, the boy grew more and more melancholy +and pale and dejected. She added hot baths, sitz baths, shower baths, +and plunges. The boy remained as dismal as a hearse. She began to +assist the water with a slim oatmeal diet and blister-plasters. She +calculated his capacity as she would a jug's, and filled him up every +day with quack cure-alls. + +Tom had become indifferent to persecution by this time. This phase +filled the old lady's heart with consternation. This indifference must +be broken up at any cost. Now she heard of Pain-killer for the first +time. She ordered a lot at once. She tasted it and was filled with +gratitude. It was simply fire in a liquid form. She dropped the water +treatment and everything else, and pinned her faith to Pain-killer. She +gave Tom a teaspoonful and watched with the deepest anxiety for the +result. Her troubles were instantly at rest, her soul at peace again; +for the "indifference" was broken up. The boy could not have shown a +wilder, heartier interest, if she had built a fire under him. + +Tom felt that it was time to wake up; this sort of life might be +romantic enough, in his blighted condition, but it was getting to have +too little sentiment and too much distracting variety about it. So he +thought over various plans for relief, and finally hit pon that of +professing to be fond of Pain-killer. He asked for it so often that he +became a nuisance, and his aunt ended by telling him to help himself +and quit bothering her. If it had been Sid, she would have had no +misgivings to alloy her delight; but since it was Tom, she watched the +bottle clandestinely. She found that the medicine did really diminish, +but it did not occur to her that the boy was mending the health of a +crack in the sitting-room floor with it. + +One day Tom was in the act of dosing the crack when his aunt's yellow +cat came along, purring, eying the teaspoon avariciously, and begging +for a taste. Tom said: + +"Don't ask for it unless you want it, Peter." + +But Peter signified that he did want it. + +"You better make sure." + +Peter was sure. + +"Now you've asked for it, and I'll give it to you, because there ain't +anything mean about me; but if you find you don't like it, you mustn't +blame anybody but your own self." + +Peter was agreeable. So Tom pried his mouth open and poured down the +Pain-killer. Peter sprang a couple of yards in the air, and then +delivered a war-whoop and set off round and round the room, banging +against furniture, upsetting flower-pots, and making general havoc. +Next he rose on his hind feet and pranced around, in a frenzy of +enjoyment, with his head over his shoulder and his voice proclaiming +his unappeasable happiness. Then he went tearing around the house again +spreading chaos and destruction in his path. Aunt Polly entered in time +to see him throw a few double summersets, deliver a final mighty +hurrah, and sail through the open window, carrying the rest of the +flower-pots with him. The old lady stood petrified with astonishment, +peering over her glasses; Tom lay on the floor expiring with laughter. + +"Tom, what on earth ails that cat?" + +"I don't know, aunt," gasped the boy. + +"Why, I never see anything like it. What did make him act so?" + +"Deed I don't know, Aunt Polly; cats always act so when they're having +a good time." + +"They do, do they?" There was something in the tone that made Tom +apprehensive. + +"Yes'm. That is, I believe they do." + +"You DO?" + +"Yes'm." + +The old lady was bending down, Tom watching, with interest emphasized +by anxiety. Too late he divined her "drift." The handle of the telltale +teaspoon was visible under the bed-valance. Aunt Polly took it, held it +up. Tom winced, and dropped his eyes. Aunt Polly raised him by the +usual handle--his ear--and cracked his head soundly with her thimble. + +"Now, sir, what did you want to treat that poor dumb beast so, for?" + +"I done it out of pity for him--because he hadn't any aunt." + +"Hadn't any aunt!--you numskull. What has that got to do with it?" + +"Heaps. Because if he'd had one she'd a burnt him out herself! She'd a +roasted his bowels out of him 'thout any more feeling than if he was a +human!" + +Aunt Polly felt a sudden pang of remorse. This was putting the thing +in a new light; what was cruelty to a cat MIGHT be cruelty to a boy, +too. She began to soften; she felt sorry. Her eyes watered a little, +and she put her hand on Tom's head and said gently: + +"I was meaning for the best, Tom. And, Tom, it DID do you good." + +Tom looked up in her face with just a perceptible twinkle peeping +through his gravity. + +"I know you was meaning for the best, aunty, and so was I with Peter. +It done HIM good, too. I never see him get around so since--" + +"Oh, go 'long with you, Tom, before you aggravate me again. And you +try and see if you can't be a good boy, for once, and you needn't take +any more medicine." + +Tom reached school ahead of time. It was noticed that this strange +thing had been occurring every day latterly. And now, as usual of late, +he hung about the gate of the schoolyard instead of playing with his +comrades. He was sick, he said, and he looked it. He tried to seem to +be looking everywhere but whither he really was looking--down the road. +Presently Jeff Thatcher hove in sight, and Tom's face lighted; he gazed +a moment, and then turned sorrowfully away. When Jeff arrived, Tom +accosted him; and "led up" warily to opportunities for remark about +Becky, but the giddy lad never could see the bait. Tom watched and +watched, hoping whenever a frisking frock came in sight, and hating the +owner of it as soon as he saw she was not the right one. At last frocks +ceased to appear, and he dropped hopelessly into the dumps; he entered +the empty schoolhouse and sat down to suffer. Then one more frock +passed in at the gate, and Tom's heart gave a great bound. The next +instant he was out, and "going on" like an Indian; yelling, laughing, +chasing boys, jumping over the fence at risk of life and limb, throwing +handsprings, standing on his head--doing all the heroic things he could +conceive of, and keeping a furtive eye out, all the while, to see if +Becky Thatcher was noticing. But she seemed to be unconscious of it +all; she never looked. Could it be possible that she was not aware that +he was there? He carried his exploits to her immediate vicinity; came +war-whooping around, snatched a boy's cap, hurled it to the roof of the +schoolhouse, broke through a group of boys, tumbling them in every +direction, and fell sprawling, himself, under Becky's nose, almost +upsetting her--and she turned, with her nose in the air, and he heard +her say: "Mf! some people think they're mighty smart--always showing +off!" + +Tom's cheeks burned. He gathered himself up and sneaked off, crushed +and crestfallen. + + + +CHAPTER XIII + +TOM'S mind was made up now. He was gloomy and desperate. He was a +forsaken, friendless boy, he said; nobody loved him; when they found +out what they had driven him to, perhaps they would be sorry; he had +tried to do right and get along, but they would not let him; since +nothing would do them but to be rid of him, let it be so; and let them +blame HIM for the consequences--why shouldn't they? What right had the +friendless to complain? Yes, they had forced him to it at last: he +would lead a life of crime. There was no choice. + +By this time he was far down Meadow Lane, and the bell for school to +"take up" tinkled faintly upon his ear. He sobbed, now, to think he +should never, never hear that old familiar sound any more--it was very +hard, but it was forced on him; since he was driven out into the cold +world, he must submit--but he forgave them. Then the sobs came thick +and fast. + +Just at this point he met his soul's sworn comrade, Joe Harper +--hard-eyed, and with evidently a great and dismal purpose in his heart. +Plainly here were "two souls with but a single thought." Tom, wiping +his eyes with his sleeve, began to blubber out something about a +resolution to escape from hard usage and lack of sympathy at home by +roaming abroad into the great world never to return; and ended by +hoping that Joe would not forget him. + +But it transpired that this was a request which Joe had just been +going to make of Tom, and had come to hunt him up for that purpose. His +mother had whipped him for drinking some cream which he had never +tasted and knew nothing about; it was plain that she was tired of him +and wished him to go; if she felt that way, there was nothing for him +to do but succumb; he hoped she would be happy, and never regret having +driven her poor boy out into the unfeeling world to suffer and die. + +As the two boys walked sorrowing along, they made a new compact to +stand by each other and be brothers and never separate till death +relieved them of their troubles. Then they began to lay their plans. +Joe was for being a hermit, and living on crusts in a remote cave, and +dying, some time, of cold and want and grief; but after listening to +Tom, he conceded that there were some conspicuous advantages about a +life of crime, and so he consented to be a pirate. + +Three miles below St. Petersburg, at a point where the Mississippi +River was a trifle over a mile wide, there was a long, narrow, wooded +island, with a shallow bar at the head of it, and this offered well as +a rendezvous. It was not inhabited; it lay far over toward the further +shore, abreast a dense and almost wholly unpeopled forest. So Jackson's +Island was chosen. Who were to be the subjects of their piracies was a +matter that did not occur to them. Then they hunted up Huckleberry +Finn, and he joined them promptly, for all careers were one to him; he +was indifferent. They presently separated to meet at a lonely spot on +the river-bank two miles above the village at the favorite hour--which +was midnight. There was a small log raft there which they meant to +capture. Each would bring hooks and lines, and such provision as he +could steal in the most dark and mysterious way--as became outlaws. And +before the afternoon was done, they had all managed to enjoy the sweet +glory of spreading the fact that pretty soon the town would "hear +something." All who got this vague hint were cautioned to "be mum and +wait." + +About midnight Tom arrived with a boiled ham and a few trifles, +and stopped in a dense undergrowth on a small bluff overlooking the +meeting-place. It was starlight, and very still. The mighty river lay +like an ocean at rest. Tom listened a moment, but no sound disturbed the +quiet. Then he gave a low, distinct whistle. It was answered from under +the bluff. Tom whistled twice more; these signals were answered in the +same way. Then a guarded voice said: + +"Who goes there?" + +"Tom Sawyer, the Black Avenger of the Spanish Main. Name your names." + +"Huck Finn the Red-Handed, and Joe Harper the Terror of the Seas." Tom +had furnished these titles, from his favorite literature. + +"'Tis well. Give the countersign." + +Two hoarse whispers delivered the same awful word simultaneously to +the brooding night: + +"BLOOD!" + +Then Tom tumbled his ham over the bluff and let himself down after it, +tearing both skin and clothes to some extent in the effort. There was +an easy, comfortable path along the shore under the bluff, but it +lacked the advantages of difficulty and danger so valued by a pirate. + +The Terror of the Seas had brought a side of bacon, and had about worn +himself out with getting it there. Finn the Red-Handed had stolen a +skillet and a quantity of half-cured leaf tobacco, and had also brought +a few corn-cobs to make pipes with. But none of the pirates smoked or +"chewed" but himself. The Black Avenger of the Spanish Main said it +would never do to start without some fire. That was a wise thought; +matches were hardly known there in that day. They saw a fire +smouldering upon a great raft a hundred yards above, and they went +stealthily thither and helped themselves to a chunk. They made an +imposing adventure of it, saying, "Hist!" every now and then, and +suddenly halting with finger on lip; moving with hands on imaginary +dagger-hilts; and giving orders in dismal whispers that if "the foe" +stirred, to "let him have it to the hilt," because "dead men tell no +tales." They knew well enough that the raftsmen were all down at the +village laying in stores or having a spree, but still that was no +excuse for their conducting this thing in an unpiratical way. + +They shoved off, presently, Tom in command, Huck at the after oar and +Joe at the forward. Tom stood amidships, gloomy-browed, and with folded +arms, and gave his orders in a low, stern whisper: + +"Luff, and bring her to the wind!" + +"Aye-aye, sir!" + +"Steady, steady-y-y-y!" + +"Steady it is, sir!" + +"Let her go off a point!" + +"Point it is, sir!" + +As the boys steadily and monotonously drove the raft toward mid-stream +it was no doubt understood that these orders were given only for +"style," and were not intended to mean anything in particular. + +"What sail's she carrying?" + +"Courses, tops'ls, and flying-jib, sir." + +"Send the r'yals up! Lay out aloft, there, half a dozen of ye +--foretopmaststuns'l! Lively, now!" + +"Aye-aye, sir!" + +"Shake out that maintogalans'l! Sheets and braces! NOW my hearties!" + +"Aye-aye, sir!" + +"Hellum-a-lee--hard a port! Stand by to meet her when she comes! Port, +port! NOW, men! With a will! Stead-y-y-y!" + +"Steady it is, sir!" + +The raft drew beyond the middle of the river; the boys pointed her +head right, and then lay on their oars. The river was not high, so +there was not more than a two or three mile current. Hardly a word was +said during the next three-quarters of an hour. Now the raft was +passing before the distant town. Two or three glimmering lights showed +where it lay, peacefully sleeping, beyond the vague vast sweep of +star-gemmed water, unconscious of the tremendous event that was happening. +The Black Avenger stood still with folded arms, "looking his last" upon +the scene of his former joys and his later sufferings, and wishing +"she" could see him now, abroad on the wild sea, facing peril and death +with dauntless heart, going to his doom with a grim smile on his lips. +It was but a small strain on his imagination to remove Jackson's Island +beyond eyeshot of the village, and so he "looked his last" with a +broken and satisfied heart. The other pirates were looking their last, +too; and they all looked so long that they came near letting the +current drift them out of the range of the island. But they discovered +the danger in time, and made shift to avert it. About two o'clock in +the morning the raft grounded on the bar two hundred yards above the +head of the island, and they waded back and forth until they had landed +their freight. Part of the little raft's belongings consisted of an old +sail, and this they spread over a nook in the bushes for a tent to +shelter their provisions; but they themselves would sleep in the open +air in good weather, as became outlaws. + +They built a fire against the side of a great log twenty or thirty +steps within the sombre depths of the forest, and then cooked some +bacon in the frying-pan for supper, and used up half of the corn "pone" +stock they had brought. It seemed glorious sport to be feasting in that +wild, free way in the virgin forest of an unexplored and uninhabited +island, far from the haunts of men, and they said they never would +return to civilization. The climbing fire lit up their faces and threw +its ruddy glare upon the pillared tree-trunks of their forest temple, +and upon the varnished foliage and festooning vines. + +When the last crisp slice of bacon was gone, and the last allowance of +corn pone devoured, the boys stretched themselves out on the grass, +filled with contentment. They could have found a cooler place, but they +would not deny themselves such a romantic feature as the roasting +camp-fire. + +"AIN'T it gay?" said Joe. + +"It's NUTS!" said Tom. "What would the boys say if they could see us?" + +"Say? Well, they'd just die to be here--hey, Hucky!" + +"I reckon so," said Huckleberry; "anyways, I'm suited. I don't want +nothing better'n this. I don't ever get enough to eat, gen'ally--and +here they can't come and pick at a feller and bullyrag him so." + +"It's just the life for me," said Tom. "You don't have to get up, +mornings, and you don't have to go to school, and wash, and all that +blame foolishness. You see a pirate don't have to do ANYTHING, Joe, +when he's ashore, but a hermit HE has to be praying considerable, and +then he don't have any fun, anyway, all by himself that way." + +"Oh yes, that's so," said Joe, "but I hadn't thought much about it, +you know. I'd a good deal rather be a pirate, now that I've tried it." + +"You see," said Tom, "people don't go much on hermits, nowadays, like +they used to in old times, but a pirate's always respected. And a +hermit's got to sleep on the hardest place he can find, and put +sackcloth and ashes on his head, and stand out in the rain, and--" + +"What does he put sackcloth and ashes on his head for?" inquired Huck. + +"I dono. But they've GOT to do it. Hermits always do. You'd have to do +that if you was a hermit." + +"Dern'd if I would," said Huck. + +"Well, what would you do?" + +"I dono. But I wouldn't do that." + +"Why, Huck, you'd HAVE to. How'd you get around it?" + +"Why, I just wouldn't stand it. I'd run away." + +"Run away! Well, you WOULD be a nice old slouch of a hermit. You'd be +a disgrace." + +The Red-Handed made no response, being better employed. He had +finished gouging out a cob, and now he fitted a weed stem to it, loaded +it with tobacco, and was pressing a coal to the charge and blowing a +cloud of fragrant smoke--he was in the full bloom of luxurious +contentment. The other pirates envied him this majestic vice, and +secretly resolved to acquire it shortly. Presently Huck said: + +"What does pirates have to do?" + +Tom said: + +"Oh, they have just a bully time--take ships and burn them, and get +the money and bury it in awful places in their island where there's +ghosts and things to watch it, and kill everybody in the ships--make +'em walk a plank." + +"And they carry the women to the island," said Joe; "they don't kill +the women." + +"No," assented Tom, "they don't kill the women--they're too noble. And +the women's always beautiful, too. + +"And don't they wear the bulliest clothes! Oh no! All gold and silver +and di'monds," said Joe, with enthusiasm. + +"Who?" said Huck. + +"Why, the pirates." + +Huck scanned his own clothing forlornly. + +"I reckon I ain't dressed fitten for a pirate," said he, with a +regretful pathos in his voice; "but I ain't got none but these." + +But the other boys told him the fine clothes would come fast enough, +after they should have begun their adventures. They made him understand +that his poor rags would do to begin with, though it was customary for +wealthy pirates to start with a proper wardrobe. + +Gradually their talk died out and drowsiness began to steal upon the +eyelids of the little waifs. The pipe dropped from the fingers of the +Red-Handed, and he slept the sleep of the conscience-free and the +weary. The Terror of the Seas and the Black Avenger of the Spanish Main +had more difficulty in getting to sleep. They said their prayers +inwardly, and lying down, since there was nobody there with authority +to make them kneel and recite aloud; in truth, they had a mind not to +say them at all, but they were afraid to proceed to such lengths as +that, lest they might call down a sudden and special thunderbolt from +heaven. Then at once they reached and hovered upon the imminent verge +of sleep--but an intruder came, now, that would not "down." It was +conscience. They began to feel a vague fear that they had been doing +wrong to run away; and next they thought of the stolen meat, and then +the real torture came. They tried to argue it away by reminding +conscience that they had purloined sweetmeats and apples scores of +times; but conscience was not to be appeased by such thin +plausibilities; it seemed to them, in the end, that there was no +getting around the stubborn fact that taking sweetmeats was only +"hooking," while taking bacon and hams and such valuables was plain +simple stealing--and there was a command against that in the Bible. So +they inwardly resolved that so long as they remained in the business, +their piracies should not again be sullied with the crime of stealing. +Then conscience granted a truce, and these curiously inconsistent +pirates fell peacefully to sleep. + + + +CHAPTER XIV + +WHEN Tom awoke in the morning, he wondered where he was. He sat up and +rubbed his eyes and looked around. Then he comprehended. It was the +cool gray dawn, and there was a delicious sense of repose and peace in +the deep pervading calm and silence of the woods. Not a leaf stirred; +not a sound obtruded upon great Nature's meditation. Beaded dewdrops +stood upon the leaves and grasses. A white layer of ashes covered the +fire, and a thin blue breath of smoke rose straight into the air. Joe +and Huck still slept. + +Now, far away in the woods a bird called; another answered; presently +the hammering of a woodpecker was heard. Gradually the cool dim gray of +the morning whitened, and as gradually sounds multiplied and life +manifested itself. The marvel of Nature shaking off sleep and going to +work unfolded itself to the musing boy. A little green worm came +crawling over a dewy leaf, lifting two-thirds of his body into the air +from time to time and "sniffing around," then proceeding again--for he +was measuring, Tom said; and when the worm approached him, of its own +accord, he sat as still as a stone, with his hopes rising and falling, +by turns, as the creature still came toward him or seemed inclined to +go elsewhere; and when at last it considered a painful moment with its +curved body in the air and then came decisively down upon Tom's leg and +began a journey over him, his whole heart was glad--for that meant that +he was going to have a new suit of clothes--without the shadow of a +doubt a gaudy piratical uniform. Now a procession of ants appeared, +from nowhere in particular, and went about their labors; one struggled +manfully by with a dead spider five times as big as itself in its arms, +and lugged it straight up a tree-trunk. A brown spotted lady-bug +climbed the dizzy height of a grass blade, and Tom bent down close to +it and said, "Lady-bug, lady-bug, fly away home, your house is on fire, +your children's alone," and she took wing and went off to see about it +--which did not surprise the boy, for he knew of old that this insect was +credulous about conflagrations, and he had practised upon its +simplicity more than once. A tumblebug came next, heaving sturdily at +its ball, and Tom touched the creature, to see it shut its legs against +its body and pretend to be dead. The birds were fairly rioting by this +time. A catbird, the Northern mocker, lit in a tree over Tom's head, +and trilled out her imitations of her neighbors in a rapture of +enjoyment; then a shrill jay swept down, a flash of blue flame, and +stopped on a twig almost within the boy's reach, cocked his head to one +side and eyed the strangers with a consuming curiosity; a gray squirrel +and a big fellow of the "fox" kind came skurrying along, sitting up at +intervals to inspect and chatter at the boys, for the wild things had +probably never seen a human being before and scarcely knew whether to +be afraid or not. All Nature was wide awake and stirring, now; long +lances of sunlight pierced down through the dense foliage far and near, +and a few butterflies came fluttering upon the scene. + +Tom stirred up the other pirates and they all clattered away with a +shout, and in a minute or two were stripped and chasing after and +tumbling over each other in the shallow limpid water of the white +sandbar. They felt no longing for the little village sleeping in the +distance beyond the majestic waste of water. A vagrant current or a +slight rise in the river had carried off their raft, but this only +gratified them, since its going was something like burning the bridge +between them and civilization. + +They came back to camp wonderfully refreshed, glad-hearted, and +ravenous; and they soon had the camp-fire blazing up again. Huck found +a spring of clear cold water close by, and the boys made cups of broad +oak or hickory leaves, and felt that water, sweetened with such a +wildwood charm as that, would be a good enough substitute for coffee. +While Joe was slicing bacon for breakfast, Tom and Huck asked him to +hold on a minute; they stepped to a promising nook in the river-bank +and threw in their lines; almost immediately they had reward. Joe had +not had time to get impatient before they were back again with some +handsome bass, a couple of sun-perch and a small catfish--provisions +enough for quite a family. They fried the fish with the bacon, and were +astonished; for no fish had ever seemed so delicious before. They did +not know that the quicker a fresh-water fish is on the fire after he is +caught the better he is; and they reflected little upon what a sauce +open-air sleeping, open-air exercise, bathing, and a large ingredient +of hunger make, too. + +They lay around in the shade, after breakfast, while Huck had a smoke, +and then went off through the woods on an exploring expedition. They +tramped gayly along, over decaying logs, through tangled underbrush, +among solemn monarchs of the forest, hung from their crowns to the +ground with a drooping regalia of grape-vines. Now and then they came +upon snug nooks carpeted with grass and jeweled with flowers. + +They found plenty of things to be delighted with, but nothing to be +astonished at. They discovered that the island was about three miles +long and a quarter of a mile wide, and that the shore it lay closest to +was only separated from it by a narrow channel hardly two hundred yards +wide. They took a swim about every hour, so it was close upon the +middle of the afternoon when they got back to camp. They were too +hungry to stop to fish, but they fared sumptuously upon cold ham, and +then threw themselves down in the shade to talk. But the talk soon +began to drag, and then died. The stillness, the solemnity that brooded +in the woods, and the sense of loneliness, began to tell upon the +spirits of the boys. They fell to thinking. A sort of undefined longing +crept upon them. This took dim shape, presently--it was budding +homesickness. Even Finn the Red-Handed was dreaming of his doorsteps +and empty hogsheads. But they were all ashamed of their weakness, and +none was brave enough to speak his thought. + +For some time, now, the boys had been dully conscious of a peculiar +sound in the distance, just as one sometimes is of the ticking of a +clock which he takes no distinct note of. But now this mysterious sound +became more pronounced, and forced a recognition. The boys started, +glanced at each other, and then each assumed a listening attitude. +There was a long silence, profound and unbroken; then a deep, sullen +boom came floating down out of the distance. + +"What is it!" exclaimed Joe, under his breath. + +"I wonder," said Tom in a whisper. + +"'Tain't thunder," said Huckleberry, in an awed tone, "becuz thunder--" + +"Hark!" said Tom. "Listen--don't talk." + +They waited a time that seemed an age, and then the same muffled boom +troubled the solemn hush. + +"Let's go and see." + +They sprang to their feet and hurried to the shore toward the town. +They parted the bushes on the bank and peered out over the water. The +little steam ferryboat was about a mile below the village, drifting +with the current. Her broad deck seemed crowded with people. There were +a great many skiffs rowing about or floating with the stream in the +neighborhood of the ferryboat, but the boys could not determine what +the men in them were doing. Presently a great jet of white smoke burst +from the ferryboat's side, and as it expanded and rose in a lazy cloud, +that same dull throb of sound was borne to the listeners again. + +"I know now!" exclaimed Tom; "somebody's drownded!" + +"That's it!" said Huck; "they done that last summer, when Bill Turner +got drownded; they shoot a cannon over the water, and that makes him +come up to the top. Yes, and they take loaves of bread and put +quicksilver in 'em and set 'em afloat, and wherever there's anybody +that's drownded, they'll float right there and stop." + +"Yes, I've heard about that," said Joe. "I wonder what makes the bread +do that." + +"Oh, it ain't the bread, so much," said Tom; "I reckon it's mostly +what they SAY over it before they start it out." + +"But they don't say anything over it," said Huck. "I've seen 'em and +they don't." + +"Well, that's funny," said Tom. "But maybe they say it to themselves. +Of COURSE they do. Anybody might know that." + +The other boys agreed that there was reason in what Tom said, because +an ignorant lump of bread, uninstructed by an incantation, could not be +expected to act very intelligently when set upon an errand of such +gravity. + +"By jings, I wish I was over there, now," said Joe. + +"I do too" said Huck "I'd give heaps to know who it is." + +The boys still listened and watched. Presently a revealing thought +flashed through Tom's mind, and he exclaimed: + +"Boys, I know who's drownded--it's us!" + +They felt like heroes in an instant. Here was a gorgeous triumph; they +were missed; they were mourned; hearts were breaking on their account; +tears were being shed; accusing memories of unkindness to these poor +lost lads were rising up, and unavailing regrets and remorse were being +indulged; and best of all, the departed were the talk of the whole +town, and the envy of all the boys, as far as this dazzling notoriety +was concerned. This was fine. It was worth while to be a pirate, after +all. + +As twilight drew on, the ferryboat went back to her accustomed +business and the skiffs disappeared. The pirates returned to camp. They +were jubilant with vanity over their new grandeur and the illustrious +trouble they were making. They caught fish, cooked supper and ate it, +and then fell to guessing at what the village was thinking and saying +about them; and the pictures they drew of the public distress on their +account were gratifying to look upon--from their point of view. But +when the shadows of night closed them in, they gradually ceased to +talk, and sat gazing into the fire, with their minds evidently +wandering elsewhere. The excitement was gone, now, and Tom and Joe +could not keep back thoughts of certain persons at home who were not +enjoying this fine frolic as much as they were. Misgivings came; they +grew troubled and unhappy; a sigh or two escaped, unawares. By and by +Joe timidly ventured upon a roundabout "feeler" as to how the others +might look upon a return to civilization--not right now, but-- + +Tom withered him with derision! Huck, being uncommitted as yet, joined +in with Tom, and the waverer quickly "explained," and was glad to get +out of the scrape with as little taint of chicken-hearted homesickness +clinging to his garments as he could. Mutiny was effectually laid to +rest for the moment. + +As the night deepened, Huck began to nod, and presently to snore. Joe +followed next. Tom lay upon his elbow motionless, for some time, +watching the two intently. At last he got up cautiously, on his knees, +and went searching among the grass and the flickering reflections flung +by the camp-fire. He picked up and inspected several large +semi-cylinders of the thin white bark of a sycamore, and finally chose +two which seemed to suit him. Then he knelt by the fire and painfully +wrote something upon each of these with his "red keel"; one he rolled up +and put in his jacket pocket, and the other he put in Joe's hat and +removed it to a little distance from the owner. And he also put into the +hat certain schoolboy treasures of almost inestimable value--among them +a lump of chalk, an India-rubber ball, three fishhooks, and one of that +kind of marbles known as a "sure 'nough crystal." Then he tiptoed his +way cautiously among the trees till he felt that he was out of hearing, +and straightway broke into a keen run in the direction of the sandbar. + + + +CHAPTER XV + +A FEW minutes later Tom was in the shoal water of the bar, wading +toward the Illinois shore. Before the depth reached his middle he was +half-way over; the current would permit no more wading, now, so he +struck out confidently to swim the remaining hundred yards. He swam +quartering upstream, but still was swept downward rather faster than he +had expected. However, he reached the shore finally, and drifted along +till he found a low place and drew himself out. He put his hand on his +jacket pocket, found his piece of bark safe, and then struck through +the woods, following the shore, with streaming garments. Shortly before +ten o'clock he came out into an open place opposite the village, and +saw the ferryboat lying in the shadow of the trees and the high bank. +Everything was quiet under the blinking stars. He crept down the bank, +watching with all his eyes, slipped into the water, swam three or four +strokes and climbed into the skiff that did "yawl" duty at the boat's +stern. He laid himself down under the thwarts and waited, panting. + +Presently the cracked bell tapped and a voice gave the order to "cast +off." A minute or two later the skiff's head was standing high up, +against the boat's swell, and the voyage was begun. Tom felt happy in +his success, for he knew it was the boat's last trip for the night. At +the end of a long twelve or fifteen minutes the wheels stopped, and Tom +slipped overboard and swam ashore in the dusk, landing fifty yards +downstream, out of danger of possible stragglers. + +He flew along unfrequented alleys, and shortly found himself at his +aunt's back fence. He climbed over, approached the "ell," and looked in +at the sitting-room window, for a light was burning there. There sat +Aunt Polly, Sid, Mary, and Joe Harper's mother, grouped together, +talking. They were by the bed, and the bed was between them and the +door. Tom went to the door and began to softly lift the latch; then he +pressed gently and the door yielded a crack; he continued pushing +cautiously, and quaking every time it creaked, till he judged he might +squeeze through on his knees; so he put his head through and began, +warily. + +"What makes the candle blow so?" said Aunt Polly. Tom hurried up. +"Why, that door's open, I believe. Why, of course it is. No end of +strange things now. Go 'long and shut it, Sid." + +Tom disappeared under the bed just in time. He lay and "breathed" +himself for a time, and then crept to where he could almost touch his +aunt's foot. + +"But as I was saying," said Aunt Polly, "he warn't BAD, so to say +--only mischEEvous. Only just giddy, and harum-scarum, you know. He +warn't any more responsible than a colt. HE never meant any harm, and +he was the best-hearted boy that ever was"--and she began to cry. + +"It was just so with my Joe--always full of his devilment, and up to +every kind of mischief, but he was just as unselfish and kind as he +could be--and laws bless me, to think I went and whipped him for taking +that cream, never once recollecting that I throwed it out myself +because it was sour, and I never to see him again in this world, never, +never, never, poor abused boy!" And Mrs. Harper sobbed as if her heart +would break. + +"I hope Tom's better off where he is," said Sid, "but if he'd been +better in some ways--" + +"SID!" Tom felt the glare of the old lady's eye, though he could not +see it. "Not a word against my Tom, now that he's gone! God'll take +care of HIM--never you trouble YOURself, sir! Oh, Mrs. Harper, I don't +know how to give him up! I don't know how to give him up! He was such a +comfort to me, although he tormented my old heart out of me, 'most." + +"The Lord giveth and the Lord hath taken away--Blessed be the name of +the Lord! But it's so hard--Oh, it's so hard! Only last Saturday my +Joe busted a firecracker right under my nose and I knocked him +sprawling. Little did I know then, how soon--Oh, if it was to do over +again I'd hug him and bless him for it." + +"Yes, yes, yes, I know just how you feel, Mrs. Harper, I know just +exactly how you feel. No longer ago than yesterday noon, my Tom took +and filled the cat full of Pain-killer, and I did think the cretur +would tear the house down. And God forgive me, I cracked Tom's head +with my thimble, poor boy, poor dead boy. But he's out of all his +troubles now. And the last words I ever heard him say was to reproach--" + +But this memory was too much for the old lady, and she broke entirely +down. Tom was snuffling, now, himself--and more in pity of himself than +anybody else. He could hear Mary crying, and putting in a kindly word +for him from time to time. He began to have a nobler opinion of himself +than ever before. Still, he was sufficiently touched by his aunt's +grief to long to rush out from under the bed and overwhelm her with +joy--and the theatrical gorgeousness of the thing appealed strongly to +his nature, too, but he resisted and lay still. + +He went on listening, and gathered by odds and ends that it was +conjectured at first that the boys had got drowned while taking a swim; +then the small raft had been missed; next, certain boys said the +missing lads had promised that the village should "hear something" +soon; the wise-heads had "put this and that together" and decided that +the lads had gone off on that raft and would turn up at the next town +below, presently; but toward noon the raft had been found, lodged +against the Missouri shore some five or six miles below the village +--and then hope perished; they must be drowned, else hunger would have +driven them home by nightfall if not sooner. It was believed that the +search for the bodies had been a fruitless effort merely because the +drowning must have occurred in mid-channel, since the boys, being good +swimmers, would otherwise have escaped to shore. This was Wednesday +night. If the bodies continued missing until Sunday, all hope would be +given over, and the funerals would be preached on that morning. Tom +shuddered. + +Mrs. Harper gave a sobbing good-night and turned to go. Then with a +mutual impulse the two bereaved women flung themselves into each +other's arms and had a good, consoling cry, and then parted. Aunt Polly +was tender far beyond her wont, in her good-night to Sid and Mary. Sid +snuffled a bit and Mary went off crying with all her heart. + +Aunt Polly knelt down and prayed for Tom so touchingly, so +appealingly, and with such measureless love in her words and her old +trembling voice, that he was weltering in tears again, long before she +was through. + +He had to keep still long after she went to bed, for she kept making +broken-hearted ejaculations from time to time, tossing unrestfully, and +turning over. But at last she was still, only moaning a little in her +sleep. Now the boy stole out, rose gradually by the bedside, shaded the +candle-light with his hand, and stood regarding her. His heart was full +of pity for her. He took out his sycamore scroll and placed it by the +candle. But something occurred to him, and he lingered considering. His +face lighted with a happy solution of his thought; he put the bark +hastily in his pocket. Then he bent over and kissed the faded lips, and +straightway made his stealthy exit, latching the door behind him. + +He threaded his way back to the ferry landing, found nobody at large +there, and walked boldly on board the boat, for he knew she was +tenantless except that there was a watchman, who always turned in and +slept like a graven image. He untied the skiff at the stern, slipped +into it, and was soon rowing cautiously upstream. When he had pulled a +mile above the village, he started quartering across and bent himself +stoutly to his work. He hit the landing on the other side neatly, for +this was a familiar bit of work to him. He was moved to capture the +skiff, arguing that it might be considered a ship and therefore +legitimate prey for a pirate, but he knew a thorough search would be +made for it and that might end in revelations. So he stepped ashore and +entered the woods. + +He sat down and took a long rest, torturing himself meanwhile to keep +awake, and then started warily down the home-stretch. The night was far +spent. It was broad daylight before he found himself fairly abreast the +island bar. He rested again until the sun was well up and gilding the +great river with its splendor, and then he plunged into the stream. A +little later he paused, dripping, upon the threshold of the camp, and +heard Joe say: + +"No, Tom's true-blue, Huck, and he'll come back. He won't desert. He +knows that would be a disgrace to a pirate, and Tom's too proud for +that sort of thing. He's up to something or other. Now I wonder what?" + +"Well, the things is ours, anyway, ain't they?" + +"Pretty near, but not yet, Huck. The writing says they are if he ain't +back here to breakfast." + +"Which he is!" exclaimed Tom, with fine dramatic effect, stepping +grandly into camp. + +A sumptuous breakfast of bacon and fish was shortly provided, and as +the boys set to work upon it, Tom recounted (and adorned) his +adventures. They were a vain and boastful company of heroes when the +tale was done. Then Tom hid himself away in a shady nook to sleep till +noon, and the other pirates got ready to fish and explore. + + + +CHAPTER XVI + +AFTER dinner all the gang turned out to hunt for turtle eggs on the +bar. They went about poking sticks into the sand, and when they found a +soft place they went down on their knees and dug with their hands. +Sometimes they would take fifty or sixty eggs out of one hole. They +were perfectly round white things a trifle smaller than an English +walnut. They had a famous fried-egg feast that night, and another on +Friday morning. + +After breakfast they went whooping and prancing out on the bar, and +chased each other round and round, shedding clothes as they went, until +they were naked, and then continued the frolic far away up the shoal +water of the bar, against the stiff current, which latter tripped their +legs from under them from time to time and greatly increased the fun. +And now and then they stooped in a group and splashed water in each +other's faces with their palms, gradually approaching each other, with +averted faces to avoid the strangling sprays, and finally gripping and +struggling till the best man ducked his neighbor, and then they all +went under in a tangle of white legs and arms and came up blowing, +sputtering, laughing, and gasping for breath at one and the same time. + +When they were well exhausted, they would run out and sprawl on the +dry, hot sand, and lie there and cover themselves up with it, and by +and by break for the water again and go through the original +performance once more. Finally it occurred to them that their naked +skin represented flesh-colored "tights" very fairly; so they drew a +ring in the sand and had a circus--with three clowns in it, for none +would yield this proudest post to his neighbor. + +Next they got their marbles and played "knucks" and "ring-taw" and +"keeps" till that amusement grew stale. Then Joe and Huck had another +swim, but Tom would not venture, because he found that in kicking off +his trousers he had kicked his string of rattlesnake rattles off his +ankle, and he wondered how he had escaped cramp so long without the +protection of this mysterious charm. He did not venture again until he +had found it, and by that time the other boys were tired and ready to +rest. They gradually wandered apart, dropped into the "dumps," and fell +to gazing longingly across the wide river to where the village lay +drowsing in the sun. Tom found himself writing "BECKY" in the sand with +his big toe; he scratched it out, and was angry with himself for his +weakness. But he wrote it again, nevertheless; he could not help it. He +erased it once more and then took himself out of temptation by driving +the other boys together and joining them. + +But Joe's spirits had gone down almost beyond resurrection. He was so +homesick that he could hardly endure the misery of it. The tears lay +very near the surface. Huck was melancholy, too. Tom was downhearted, +but tried hard not to show it. He had a secret which he was not ready +to tell, yet, but if this mutinous depression was not broken up soon, +he would have to bring it out. He said, with a great show of +cheerfulness: + +"I bet there's been pirates on this island before, boys. We'll explore +it again. They've hid treasures here somewhere. How'd you feel to light +on a rotten chest full of gold and silver--hey?" + +But it roused only faint enthusiasm, which faded out, with no reply. +Tom tried one or two other seductions; but they failed, too. It was +discouraging work. Joe sat poking up the sand with a stick and looking +very gloomy. Finally he said: + +"Oh, boys, let's give it up. I want to go home. It's so lonesome." + +"Oh no, Joe, you'll feel better by and by," said Tom. "Just think of +the fishing that's here." + +"I don't care for fishing. I want to go home." + +"But, Joe, there ain't such another swimming-place anywhere." + +"Swimming's no good. I don't seem to care for it, somehow, when there +ain't anybody to say I sha'n't go in. I mean to go home." + +"Oh, shucks! Baby! You want to see your mother, I reckon." + +"Yes, I DO want to see my mother--and you would, too, if you had one. +I ain't any more baby than you are." And Joe snuffled a little. + +"Well, we'll let the cry-baby go home to his mother, won't we, Huck? +Poor thing--does it want to see its mother? And so it shall. You like +it here, don't you, Huck? We'll stay, won't we?" + +Huck said, "Y-e-s"--without any heart in it. + +"I'll never speak to you again as long as I live," said Joe, rising. +"There now!" And he moved moodily away and began to dress himself. + +"Who cares!" said Tom. "Nobody wants you to. Go 'long home and get +laughed at. Oh, you're a nice pirate. Huck and me ain't cry-babies. +We'll stay, won't we, Huck? Let him go if he wants to. I reckon we can +get along without him, per'aps." + +But Tom was uneasy, nevertheless, and was alarmed to see Joe go +sullenly on with his dressing. And then it was discomforting to see +Huck eying Joe's preparations so wistfully, and keeping up such an +ominous silence. Presently, without a parting word, Joe began to wade +off toward the Illinois shore. Tom's heart began to sink. He glanced at +Huck. Huck could not bear the look, and dropped his eyes. Then he said: + +"I want to go, too, Tom. It was getting so lonesome anyway, and now +it'll be worse. Let's us go, too, Tom." + +"I won't! You can all go, if you want to. I mean to stay." + +"Tom, I better go." + +"Well, go 'long--who's hendering you." + +Huck began to pick up his scattered clothes. He said: + +"Tom, I wisht you'd come, too. Now you think it over. We'll wait for +you when we get to shore." + +"Well, you'll wait a blame long time, that's all." + +Huck started sorrowfully away, and Tom stood looking after him, with a +strong desire tugging at his heart to yield his pride and go along too. +He hoped the boys would stop, but they still waded slowly on. It +suddenly dawned on Tom that it was become very lonely and still. He +made one final struggle with his pride, and then darted after his +comrades, yelling: + +"Wait! Wait! I want to tell you something!" + +They presently stopped and turned around. When he got to where they +were, he began unfolding his secret, and they listened moodily till at +last they saw the "point" he was driving at, and then they set up a +war-whoop of applause and said it was "splendid!" and said if he had +told them at first, they wouldn't have started away. He made a plausible +excuse; but his real reason had been the fear that not even the secret +would keep them with him any very great length of time, and so he had +meant to hold it in reserve as a last seduction. + +The lads came gayly back and went at their sports again with a will, +chattering all the time about Tom's stupendous plan and admiring the +genius of it. After a dainty egg and fish dinner, Tom said he wanted to +learn to smoke, now. Joe caught at the idea and said he would like to +try, too. So Huck made pipes and filled them. These novices had never +smoked anything before but cigars made of grape-vine, and they "bit" +the tongue, and were not considered manly anyway. + +Now they stretched themselves out on their elbows and began to puff, +charily, and with slender confidence. The smoke had an unpleasant +taste, and they gagged a little, but Tom said: + +"Why, it's just as easy! If I'd a knowed this was all, I'd a learnt +long ago." + +"So would I," said Joe. "It's just nothing." + +"Why, many a time I've looked at people smoking, and thought well I +wish I could do that; but I never thought I could," said Tom. + +"That's just the way with me, hain't it, Huck? You've heard me talk +just that way--haven't you, Huck? I'll leave it to Huck if I haven't." + +"Yes--heaps of times," said Huck. + +"Well, I have too," said Tom; "oh, hundreds of times. Once down by the +slaughter-house. Don't you remember, Huck? Bob Tanner was there, and +Johnny Miller, and Jeff Thatcher, when I said it. Don't you remember, +Huck, 'bout me saying that?" + +"Yes, that's so," said Huck. "That was the day after I lost a white +alley. No, 'twas the day before." + +"There--I told you so," said Tom. "Huck recollects it." + +"I bleeve I could smoke this pipe all day," said Joe. "I don't feel +sick." + +"Neither do I," said Tom. "I could smoke it all day. But I bet you +Jeff Thatcher couldn't." + +"Jeff Thatcher! Why, he'd keel over just with two draws. Just let him +try it once. HE'D see!" + +"I bet he would. And Johnny Miller--I wish could see Johnny Miller +tackle it once." + +"Oh, don't I!" said Joe. "Why, I bet you Johnny Miller couldn't any +more do this than nothing. Just one little snifter would fetch HIM." + +"'Deed it would, Joe. Say--I wish the boys could see us now." + +"So do I." + +"Say--boys, don't say anything about it, and some time when they're +around, I'll come up to you and say, 'Joe, got a pipe? I want a smoke.' +And you'll say, kind of careless like, as if it warn't anything, you'll +say, 'Yes, I got my OLD pipe, and another one, but my tobacker ain't +very good.' And I'll say, 'Oh, that's all right, if it's STRONG +enough.' And then you'll out with the pipes, and we'll light up just as +ca'm, and then just see 'em look!" + +"By jings, that'll be gay, Tom! I wish it was NOW!" + +"So do I! And when we tell 'em we learned when we was off pirating, +won't they wish they'd been along?" + +"Oh, I reckon not! I'll just BET they will!" + +So the talk ran on. But presently it began to flag a trifle, and grow +disjointed. The silences widened; the expectoration marvellously +increased. Every pore inside the boys' cheeks became a spouting +fountain; they could scarcely bail out the cellars under their tongues +fast enough to prevent an inundation; little overflowings down their +throats occurred in spite of all they could do, and sudden retchings +followed every time. Both boys were looking very pale and miserable, +now. Joe's pipe dropped from his nerveless fingers. Tom's followed. +Both fountains were going furiously and both pumps bailing with might +and main. Joe said feebly: + +"I've lost my knife. I reckon I better go and find it." + +Tom said, with quivering lips and halting utterance: + +"I'll help you. You go over that way and I'll hunt around by the +spring. No, you needn't come, Huck--we can find it." + +So Huck sat down again, and waited an hour. Then he found it lonesome, +and went to find his comrades. They were wide apart in the woods, both +very pale, both fast asleep. But something informed him that if they +had had any trouble they had got rid of it. + +They were not talkative at supper that night. They had a humble look, +and when Huck prepared his pipe after the meal and was going to prepare +theirs, they said no, they were not feeling very well--something they +ate at dinner had disagreed with them. + +About midnight Joe awoke, and called the boys. There was a brooding +oppressiveness in the air that seemed to bode something. The boys +huddled themselves together and sought the friendly companionship of +the fire, though the dull dead heat of the breathless atmosphere was +stifling. They sat still, intent and waiting. The solemn hush +continued. Beyond the light of the fire everything was swallowed up in +the blackness of darkness. Presently there came a quivering glow that +vaguely revealed the foliage for a moment and then vanished. By and by +another came, a little stronger. Then another. Then a faint moan came +sighing through the branches of the forest and the boys felt a fleeting +breath upon their cheeks, and shuddered with the fancy that the Spirit +of the Night had gone by. There was a pause. Now a weird flash turned +night into day and showed every little grass-blade, separate and +distinct, that grew about their feet. And it showed three white, +startled faces, too. A deep peal of thunder went rolling and tumbling +down the heavens and lost itself in sullen rumblings in the distance. A +sweep of chilly air passed by, rustling all the leaves and snowing the +flaky ashes broadcast about the fire. Another fierce glare lit up the +forest and an instant crash followed that seemed to rend the tree-tops +right over the boys' heads. They clung together in terror, in the thick +gloom that followed. A few big rain-drops fell pattering upon the +leaves. + +"Quick! boys, go for the tent!" exclaimed Tom. + +They sprang away, stumbling over roots and among vines in the dark, no +two plunging in the same direction. A furious blast roared through the +trees, making everything sing as it went. One blinding flash after +another came, and peal on peal of deafening thunder. And now a +drenching rain poured down and the rising hurricane drove it in sheets +along the ground. The boys cried out to each other, but the roaring +wind and the booming thunder-blasts drowned their voices utterly. +However, one by one they straggled in at last and took shelter under +the tent, cold, scared, and streaming with water; but to have company +in misery seemed something to be grateful for. They could not talk, the +old sail flapped so furiously, even if the other noises would have +allowed them. The tempest rose higher and higher, and presently the +sail tore loose from its fastenings and went winging away on the blast. +The boys seized each others' hands and fled, with many tumblings and +bruises, to the shelter of a great oak that stood upon the river-bank. +Now the battle was at its highest. Under the ceaseless conflagration of +lightning that flamed in the skies, everything below stood out in +clean-cut and shadowless distinctness: the bending trees, the billowy +river, white with foam, the driving spray of spume-flakes, the dim +outlines of the high bluffs on the other side, glimpsed through the +drifting cloud-rack and the slanting veil of rain. Every little while +some giant tree yielded the fight and fell crashing through the younger +growth; and the unflagging thunder-peals came now in ear-splitting +explosive bursts, keen and sharp, and unspeakably appalling. The storm +culminated in one matchless effort that seemed likely to tear the island +to pieces, burn it up, drown it to the tree-tops, blow it away, and +deafen every creature in it, all at one and the same moment. It was a +wild night for homeless young heads to be out in. + +But at last the battle was done, and the forces retired with weaker +and weaker threatenings and grumblings, and peace resumed her sway. The +boys went back to camp, a good deal awed; but they found there was +still something to be thankful for, because the great sycamore, the +shelter of their beds, was a ruin, now, blasted by the lightnings, and +they were not under it when the catastrophe happened. + +Everything in camp was drenched, the camp-fire as well; for they were +but heedless lads, like their generation, and had made no provision +against rain. Here was matter for dismay, for they were soaked through +and chilled. They were eloquent in their distress; but they presently +discovered that the fire had eaten so far up under the great log it had +been built against (where it curved upward and separated itself from +the ground), that a handbreadth or so of it had escaped wetting; so +they patiently wrought until, with shreds and bark gathered from the +under sides of sheltered logs, they coaxed the fire to burn again. Then +they piled on great dead boughs till they had a roaring furnace, and +were glad-hearted once more. They dried their boiled ham and had a +feast, and after that they sat by the fire and expanded and glorified +their midnight adventure until morning, for there was not a dry spot to +sleep on, anywhere around. + +As the sun began to steal in upon the boys, drowsiness came over them, +and they went out on the sandbar and lay down to sleep. They got +scorched out by and by, and drearily set about getting breakfast. After +the meal they felt rusty, and stiff-jointed, and a little homesick once +more. Tom saw the signs, and fell to cheering up the pirates as well as +he could. But they cared nothing for marbles, or circus, or swimming, +or anything. He reminded them of the imposing secret, and raised a ray +of cheer. While it lasted, he got them interested in a new device. This +was to knock off being pirates, for a while, and be Indians for a +change. They were attracted by this idea; so it was not long before +they were stripped, and striped from head to heel with black mud, like +so many zebras--all of them chiefs, of course--and then they went +tearing through the woods to attack an English settlement. + +By and by they separated into three hostile tribes, and darted upon +each other from ambush with dreadful war-whoops, and killed and scalped +each other by thousands. It was a gory day. Consequently it was an +extremely satisfactory one. + +They assembled in camp toward supper-time, hungry and happy; but now a +difficulty arose--hostile Indians could not break the bread of +hospitality together without first making peace, and this was a simple +impossibility without smoking a pipe of peace. There was no other +process that ever they had heard of. Two of the savages almost wished +they had remained pirates. However, there was no other way; so with +such show of cheerfulness as they could muster they called for the pipe +and took their whiff as it passed, in due form. + +And behold, they were glad they had gone into savagery, for they had +gained something; they found that they could now smoke a little without +having to go and hunt for a lost knife; they did not get sick enough to +be seriously uncomfortable. They were not likely to fool away this high +promise for lack of effort. No, they practised cautiously, after +supper, with right fair success, and so they spent a jubilant evening. +They were prouder and happier in their new acquirement than they would +have been in the scalping and skinning of the Six Nations. We will +leave them to smoke and chatter and brag, since we have no further use +for them at present. + + + +CHAPTER XVII + +BUT there was no hilarity in the little town that same tranquil +Saturday afternoon. The Harpers, and Aunt Polly's family, were being +put into mourning, with great grief and many tears. An unusual quiet +possessed the village, although it was ordinarily quiet enough, in all +conscience. The villagers conducted their concerns with an absent air, +and talked little; but they sighed often. The Saturday holiday seemed a +burden to the children. They had no heart in their sports, and +gradually gave them up. + +In the afternoon Becky Thatcher found herself moping about the +deserted schoolhouse yard, and feeling very melancholy. But she found +nothing there to comfort her. She soliloquized: + +"Oh, if I only had a brass andiron-knob again! But I haven't got +anything now to remember him by." And she choked back a little sob. + +Presently she stopped, and said to herself: + +"It was right here. Oh, if it was to do over again, I wouldn't say +that--I wouldn't say it for the whole world. But he's gone now; I'll +never, never, never see him any more." + +This thought broke her down, and she wandered away, with tears rolling +down her cheeks. Then quite a group of boys and girls--playmates of +Tom's and Joe's--came by, and stood looking over the paling fence and +talking in reverent tones of how Tom did so-and-so the last time they +saw him, and how Joe said this and that small trifle (pregnant with +awful prophecy, as they could easily see now!)--and each speaker +pointed out the exact spot where the lost lads stood at the time, and +then added something like "and I was a-standing just so--just as I am +now, and as if you was him--I was as close as that--and he smiled, just +this way--and then something seemed to go all over me, like--awful, you +know--and I never thought what it meant, of course, but I can see now!" + +Then there was a dispute about who saw the dead boys last in life, and +many claimed that dismal distinction, and offered evidences, more or +less tampered with by the witness; and when it was ultimately decided +who DID see the departed last, and exchanged the last words with them, +the lucky parties took upon themselves a sort of sacred importance, and +were gaped at and envied by all the rest. One poor chap, who had no +other grandeur to offer, said with tolerably manifest pride in the +remembrance: + +"Well, Tom Sawyer he licked me once." + +But that bid for glory was a failure. Most of the boys could say that, +and so that cheapened the distinction too much. The group loitered +away, still recalling memories of the lost heroes, in awed voices. + +When the Sunday-school hour was finished, the next morning, the bell +began to toll, instead of ringing in the usual way. It was a very still +Sabbath, and the mournful sound seemed in keeping with the musing hush +that lay upon nature. The villagers began to gather, loitering a moment +in the vestibule to converse in whispers about the sad event. But there +was no whispering in the house; only the funereal rustling of dresses +as the women gathered to their seats disturbed the silence there. None +could remember when the little church had been so full before. There +was finally a waiting pause, an expectant dumbness, and then Aunt Polly +entered, followed by Sid and Mary, and they by the Harper family, all +in deep black, and the whole congregation, the old minister as well, +rose reverently and stood until the mourners were seated in the front +pew. There was another communing silence, broken at intervals by +muffled sobs, and then the minister spread his hands abroad and prayed. +A moving hymn was sung, and the text followed: "I am the Resurrection +and the Life." + +As the service proceeded, the clergyman drew such pictures of the +graces, the winning ways, and the rare promise of the lost lads that +every soul there, thinking he recognized these pictures, felt a pang in +remembering that he had persistently blinded himself to them always +before, and had as persistently seen only faults and flaws in the poor +boys. The minister related many a touching incident in the lives of the +departed, too, which illustrated their sweet, generous natures, and the +people could easily see, now, how noble and beautiful those episodes +were, and remembered with grief that at the time they occurred they had +seemed rank rascalities, well deserving of the cowhide. The +congregation became more and more moved, as the pathetic tale went on, +till at last the whole company broke down and joined the weeping +mourners in a chorus of anguished sobs, the preacher himself giving way +to his feelings, and crying in the pulpit. + +There was a rustle in the gallery, which nobody noticed; a moment +later the church door creaked; the minister raised his streaming eyes +above his handkerchief, and stood transfixed! First one and then +another pair of eyes followed the minister's, and then almost with one +impulse the congregation rose and stared while the three dead boys came +marching up the aisle, Tom in the lead, Joe next, and Huck, a ruin of +drooping rags, sneaking sheepishly in the rear! They had been hid in +the unused gallery listening to their own funeral sermon! + +Aunt Polly, Mary, and the Harpers threw themselves upon their restored +ones, smothered them with kisses and poured out thanksgivings, while +poor Huck stood abashed and uncomfortable, not knowing exactly what to +do or where to hide from so many unwelcoming eyes. He wavered, and +started to slink away, but Tom seized him and said: + +"Aunt Polly, it ain't fair. Somebody's got to be glad to see Huck." + +"And so they shall. I'm glad to see him, poor motherless thing!" And +the loving attentions Aunt Polly lavished upon him were the one thing +capable of making him more uncomfortable than he was before. + +Suddenly the minister shouted at the top of his voice: "Praise God +from whom all blessings flow--SING!--and put your hearts in it!" + +And they did. Old Hundred swelled up with a triumphant burst, and +while it shook the rafters Tom Sawyer the Pirate looked around upon the +envying juveniles about him and confessed in his heart that this was +the proudest moment of his life. + +As the "sold" congregation trooped out they said they would almost be +willing to be made ridiculous again to hear Old Hundred sung like that +once more. + +Tom got more cuffs and kisses that day--according to Aunt Polly's +varying moods--than he had earned before in a year; and he hardly knew +which expressed the most gratefulness to God and affection for himself. + + + +CHAPTER XVIII + +THAT was Tom's great secret--the scheme to return home with his +brother pirates and attend their own funerals. They had paddled over to +the Missouri shore on a log, at dusk on Saturday, landing five or six +miles below the village; they had slept in the woods at the edge of the +town till nearly daylight, and had then crept through back lanes and +alleys and finished their sleep in the gallery of the church among a +chaos of invalided benches. + +At breakfast, Monday morning, Aunt Polly and Mary were very loving to +Tom, and very attentive to his wants. There was an unusual amount of +talk. In the course of it Aunt Polly said: + +"Well, I don't say it wasn't a fine joke, Tom, to keep everybody +suffering 'most a week so you boys had a good time, but it is a pity +you could be so hard-hearted as to let me suffer so. If you could come +over on a log to go to your funeral, you could have come over and give +me a hint some way that you warn't dead, but only run off." + +"Yes, you could have done that, Tom," said Mary; "and I believe you +would if you had thought of it." + +"Would you, Tom?" said Aunt Polly, her face lighting wistfully. "Say, +now, would you, if you'd thought of it?" + +"I--well, I don't know. 'Twould 'a' spoiled everything." + +"Tom, I hoped you loved me that much," said Aunt Polly, with a grieved +tone that discomforted the boy. "It would have been something if you'd +cared enough to THINK of it, even if you didn't DO it." + +"Now, auntie, that ain't any harm," pleaded Mary; "it's only Tom's +giddy way--he is always in such a rush that he never thinks of +anything." + +"More's the pity. Sid would have thought. And Sid would have come and +DONE it, too. Tom, you'll look back, some day, when it's too late, and +wish you'd cared a little more for me when it would have cost you so +little." + +"Now, auntie, you know I do care for you," said Tom. + +"I'd know it better if you acted more like it." + +"I wish now I'd thought," said Tom, with a repentant tone; "but I +dreamt about you, anyway. That's something, ain't it?" + +"It ain't much--a cat does that much--but it's better than nothing. +What did you dream?" + +"Why, Wednesday night I dreamt that you was sitting over there by the +bed, and Sid was sitting by the woodbox, and Mary next to him." + +"Well, so we did. So we always do. I'm glad your dreams could take +even that much trouble about us." + +"And I dreamt that Joe Harper's mother was here." + +"Why, she was here! Did you dream any more?" + +"Oh, lots. But it's so dim, now." + +"Well, try to recollect--can't you?" + +"Somehow it seems to me that the wind--the wind blowed the--the--" + +"Try harder, Tom! The wind did blow something. Come!" + +Tom pressed his fingers on his forehead an anxious minute, and then +said: + +"I've got it now! I've got it now! It blowed the candle!" + +"Mercy on us! Go on, Tom--go on!" + +"And it seems to me that you said, 'Why, I believe that that door--'" + +"Go ON, Tom!" + +"Just let me study a moment--just a moment. Oh, yes--you said you +believed the door was open." + +"As I'm sitting here, I did! Didn't I, Mary! Go on!" + +"And then--and then--well I won't be certain, but it seems like as if +you made Sid go and--and--" + +"Well? Well? What did I make him do, Tom? What did I make him do?" + +"You made him--you--Oh, you made him shut it." + +"Well, for the land's sake! I never heard the beat of that in all my +days! Don't tell ME there ain't anything in dreams, any more. Sereny +Harper shall know of this before I'm an hour older. I'd like to see her +get around THIS with her rubbage 'bout superstition. Go on, Tom!" + +"Oh, it's all getting just as bright as day, now. Next you said I +warn't BAD, only mischeevous and harum-scarum, and not any more +responsible than--than--I think it was a colt, or something." + +"And so it was! Well, goodness gracious! Go on, Tom!" + +"And then you began to cry." + +"So I did. So I did. Not the first time, neither. And then--" + +"Then Mrs. Harper she began to cry, and said Joe was just the same, +and she wished she hadn't whipped him for taking cream when she'd +throwed it out her own self--" + +"Tom! The sperrit was upon you! You was a prophesying--that's what you +was doing! Land alive, go on, Tom!" + +"Then Sid he said--he said--" + +"I don't think I said anything," said Sid. + +"Yes you did, Sid," said Mary. + +"Shut your heads and let Tom go on! What did he say, Tom?" + +"He said--I THINK he said he hoped I was better off where I was gone +to, but if I'd been better sometimes--" + +"THERE, d'you hear that! It was his very words!" + +"And you shut him up sharp." + +"I lay I did! There must 'a' been an angel there. There WAS an angel +there, somewheres!" + +"And Mrs. Harper told about Joe scaring her with a firecracker, and +you told about Peter and the Painkiller--" + +"Just as true as I live!" + +"And then there was a whole lot of talk 'bout dragging the river for +us, and 'bout having the funeral Sunday, and then you and old Miss +Harper hugged and cried, and she went." + +"It happened just so! It happened just so, as sure as I'm a-sitting in +these very tracks. Tom, you couldn't told it more like if you'd 'a' +seen it! And then what? Go on, Tom!" + +"Then I thought you prayed for me--and I could see you and hear every +word you said. And you went to bed, and I was so sorry that I took and +wrote on a piece of sycamore bark, 'We ain't dead--we are only off +being pirates,' and put it on the table by the candle; and then you +looked so good, laying there asleep, that I thought I went and leaned +over and kissed you on the lips." + +"Did you, Tom, DID you! I just forgive you everything for that!" And +she seized the boy in a crushing embrace that made him feel like the +guiltiest of villains. + +"It was very kind, even though it was only a--dream," Sid soliloquized +just audibly. + +"Shut up, Sid! A body does just the same in a dream as he'd do if he +was awake. Here's a big Milum apple I've been saving for you, Tom, if +you was ever found again--now go 'long to school. I'm thankful to the +good God and Father of us all I've got you back, that's long-suffering +and merciful to them that believe on Him and keep His word, though +goodness knows I'm unworthy of it, but if only the worthy ones got His +blessings and had His hand to help them over the rough places, there's +few enough would smile here or ever enter into His rest when the long +night comes. Go 'long Sid, Mary, Tom--take yourselves off--you've +hendered me long enough." + +The children left for school, and the old lady to call on Mrs. Harper +and vanquish her realism with Tom's marvellous dream. Sid had better +judgment than to utter the thought that was in his mind as he left the +house. It was this: "Pretty thin--as long a dream as that, without any +mistakes in it!" + +What a hero Tom was become, now! He did not go skipping and prancing, +but moved with a dignified swagger as became a pirate who felt that the +public eye was on him. And indeed it was; he tried not to seem to see +the looks or hear the remarks as he passed along, but they were food +and drink to him. Smaller boys than himself flocked at his heels, as +proud to be seen with him, and tolerated by him, as if he had been the +drummer at the head of a procession or the elephant leading a menagerie +into town. Boys of his own size pretended not to know he had been away +at all; but they were consuming with envy, nevertheless. They would +have given anything to have that swarthy suntanned skin of his, and his +glittering notoriety; and Tom would not have parted with either for a +circus. + +At school the children made so much of him and of Joe, and delivered +such eloquent admiration from their eyes, that the two heroes were not +long in becoming insufferably "stuck-up." They began to tell their +adventures to hungry listeners--but they only began; it was not a thing +likely to have an end, with imaginations like theirs to furnish +material. And finally, when they got out their pipes and went serenely +puffing around, the very summit of glory was reached. + +Tom decided that he could be independent of Becky Thatcher now. Glory +was sufficient. He would live for glory. Now that he was distinguished, +maybe she would be wanting to "make up." Well, let her--she should see +that he could be as indifferent as some other people. Presently she +arrived. Tom pretended not to see her. He moved away and joined a group +of boys and girls and began to talk. Soon he observed that she was +tripping gayly back and forth with flushed face and dancing eyes, +pretending to be busy chasing schoolmates, and screaming with laughter +when she made a capture; but he noticed that she always made her +captures in his vicinity, and that she seemed to cast a conscious eye +in his direction at such times, too. It gratified all the vicious +vanity that was in him; and so, instead of winning him, it only "set +him up" the more and made him the more diligent to avoid betraying that +he knew she was about. Presently she gave over skylarking, and moved +irresolutely about, sighing once or twice and glancing furtively and +wistfully toward Tom. Then she observed that now Tom was talking more +particularly to Amy Lawrence than to any one else. She felt a sharp +pang and grew disturbed and uneasy at once. She tried to go away, but +her feet were treacherous, and carried her to the group instead. She +said to a girl almost at Tom's elbow--with sham vivacity: + +"Why, Mary Austin! you bad girl, why didn't you come to Sunday-school?" + +"I did come--didn't you see me?" + +"Why, no! Did you? Where did you sit?" + +"I was in Miss Peters' class, where I always go. I saw YOU." + +"Did you? Why, it's funny I didn't see you. I wanted to tell you about +the picnic." + +"Oh, that's jolly. Who's going to give it?" + +"My ma's going to let me have one." + +"Oh, goody; I hope she'll let ME come." + +"Well, she will. The picnic's for me. She'll let anybody come that I +want, and I want you." + +"That's ever so nice. When is it going to be?" + +"By and by. Maybe about vacation." + +"Oh, won't it be fun! You going to have all the girls and boys?" + +"Yes, every one that's friends to me--or wants to be"; and she glanced +ever so furtively at Tom, but he talked right along to Amy Lawrence +about the terrible storm on the island, and how the lightning tore the +great sycamore tree "all to flinders" while he was "standing within +three feet of it." + +"Oh, may I come?" said Grace Miller. + +"Yes." + +"And me?" said Sally Rogers. + +"Yes." + +"And me, too?" said Susy Harper. "And Joe?" + +"Yes." + +And so on, with clapping of joyful hands till all the group had begged +for invitations but Tom and Amy. Then Tom turned coolly away, still +talking, and took Amy with him. Becky's lips trembled and the tears +came to her eyes; she hid these signs with a forced gayety and went on +chattering, but the life had gone out of the picnic, now, and out of +everything else; she got away as soon as she could and hid herself and +had what her sex call "a good cry." Then she sat moody, with wounded +pride, till the bell rang. She roused up, now, with a vindictive cast +in her eye, and gave her plaited tails a shake and said she knew what +SHE'D do. + +At recess Tom continued his flirtation with Amy with jubilant +self-satisfaction. And he kept drifting about to find Becky and lacerate +her with the performance. At last he spied her, but there was a sudden +falling of his mercury. She was sitting cosily on a little bench behind +the schoolhouse looking at a picture-book with Alfred Temple--and so +absorbed were they, and their heads so close together over the book, +that they did not seem to be conscious of anything in the world besides. +Jealousy ran red-hot through Tom's veins. He began to hate himself for +throwing away the chance Becky had offered for a reconciliation. He +called himself a fool, and all the hard names he could think of. He +wanted to cry with vexation. Amy chatted happily along, as they walked, +for her heart was singing, but Tom's tongue had lost its function. He +did not hear what Amy was saying, and whenever she paused expectantly he +could only stammer an awkward assent, which was as often misplaced as +otherwise. He kept drifting to the rear of the schoolhouse, again and +again, to sear his eyeballs with the hateful spectacle there. He could +not help it. And it maddened him to see, as he thought he saw, that +Becky Thatcher never once suspected that he was even in the land of the +living. But she did see, nevertheless; and she knew she was winning her +fight, too, and was glad to see him suffer as she had suffered. + +Amy's happy prattle became intolerable. Tom hinted at things he had to +attend to; things that must be done; and time was fleeting. But in +vain--the girl chirped on. Tom thought, "Oh, hang her, ain't I ever +going to get rid of her?" At last he must be attending to those +things--and she said artlessly that she would be "around" when school +let out. And he hastened away, hating her for it. + +"Any other boy!" Tom thought, grating his teeth. "Any boy in the whole +town but that Saint Louis smarty that thinks he dresses so fine and is +aristocracy! Oh, all right, I licked you the first day you ever saw +this town, mister, and I'll lick you again! You just wait till I catch +you out! I'll just take and--" + +And he went through the motions of thrashing an imaginary boy +--pummelling the air, and kicking and gouging. "Oh, you do, do you? You +holler 'nough, do you? Now, then, let that learn you!" And so the +imaginary flogging was finished to his satisfaction. + +Tom fled home at noon. His conscience could not endure any more of +Amy's grateful happiness, and his jealousy could bear no more of the +other distress. Becky resumed her picture inspections with Alfred, but +as the minutes dragged along and no Tom came to suffer, her triumph +began to cloud and she lost interest; gravity and absent-mindedness +followed, and then melancholy; two or three times she pricked up her +ear at a footstep, but it was a false hope; no Tom came. At last she +grew entirely miserable and wished she hadn't carried it so far. When +poor Alfred, seeing that he was losing her, he did not know how, kept +exclaiming: "Oh, here's a jolly one! look at this!" she lost patience +at last, and said, "Oh, don't bother me! I don't care for them!" and +burst into tears, and got up and walked away. + +Alfred dropped alongside and was going to try to comfort her, but she +said: + +"Go away and leave me alone, can't you! I hate you!" + +So the boy halted, wondering what he could have done--for she had said +she would look at pictures all through the nooning--and she walked on, +crying. Then Alfred went musing into the deserted schoolhouse. He was +humiliated and angry. He easily guessed his way to the truth--the girl +had simply made a convenience of him to vent her spite upon Tom Sawyer. +He was far from hating Tom the less when this thought occurred to him. +He wished there was some way to get that boy into trouble without much +risk to himself. Tom's spelling-book fell under his eye. Here was his +opportunity. He gratefully opened to the lesson for the afternoon and +poured ink upon the page. + +Becky, glancing in at a window behind him at the moment, saw the act, +and moved on, without discovering herself. She started homeward, now, +intending to find Tom and tell him; Tom would be thankful and their +troubles would be healed. Before she was half way home, however, she +had changed her mind. The thought of Tom's treatment of her when she +was talking about her picnic came scorching back and filled her with +shame. She resolved to let him get whipped on the damaged +spelling-book's account, and to hate him forever, into the bargain. + + + +CHAPTER XIX + +TOM arrived at home in a dreary mood, and the first thing his aunt +said to him showed him that he had brought his sorrows to an +unpromising market: + +"Tom, I've a notion to skin you alive!" + +"Auntie, what have I done?" + +"Well, you've done enough. Here I go over to Sereny Harper, like an +old softy, expecting I'm going to make her believe all that rubbage +about that dream, when lo and behold you she'd found out from Joe that +you was over here and heard all the talk we had that night. Tom, I +don't know what is to become of a boy that will act like that. It makes +me feel so bad to think you could let me go to Sereny Harper and make +such a fool of myself and never say a word." + +This was a new aspect of the thing. His smartness of the morning had +seemed to Tom a good joke before, and very ingenious. It merely looked +mean and shabby now. He hung his head and could not think of anything +to say for a moment. Then he said: + +"Auntie, I wish I hadn't done it--but I didn't think." + +"Oh, child, you never think. You never think of anything but your own +selfishness. You could think to come all the way over here from +Jackson's Island in the night to laugh at our troubles, and you could +think to fool me with a lie about a dream; but you couldn't ever think +to pity us and save us from sorrow." + +"Auntie, I know now it was mean, but I didn't mean to be mean. I +didn't, honest. And besides, I didn't come over here to laugh at you +that night." + +"What did you come for, then?" + +"It was to tell you not to be uneasy about us, because we hadn't got +drownded." + +"Tom, Tom, I would be the thankfullest soul in this world if I could +believe you ever had as good a thought as that, but you know you never +did--and I know it, Tom." + +"Indeed and 'deed I did, auntie--I wish I may never stir if I didn't." + +"Oh, Tom, don't lie--don't do it. It only makes things a hundred times +worse." + +"It ain't a lie, auntie; it's the truth. I wanted to keep you from +grieving--that was all that made me come." + +"I'd give the whole world to believe that--it would cover up a power +of sins, Tom. I'd 'most be glad you'd run off and acted so bad. But it +ain't reasonable; because, why didn't you tell me, child?" + +"Why, you see, when you got to talking about the funeral, I just got +all full of the idea of our coming and hiding in the church, and I +couldn't somehow bear to spoil it. So I just put the bark back in my +pocket and kept mum." + +"What bark?" + +"The bark I had wrote on to tell you we'd gone pirating. I wish, now, +you'd waked up when I kissed you--I do, honest." + +The hard lines in his aunt's face relaxed and a sudden tenderness +dawned in her eyes. + +"DID you kiss me, Tom?" + +"Why, yes, I did." + +"Are you sure you did, Tom?" + +"Why, yes, I did, auntie--certain sure." + +"What did you kiss me for, Tom?" + +"Because I loved you so, and you laid there moaning and I was so sorry." + +The words sounded like truth. The old lady could not hide a tremor in +her voice when she said: + +"Kiss me again, Tom!--and be off with you to school, now, and don't +bother me any more." + +The moment he was gone, she ran to a closet and got out the ruin of a +jacket which Tom had gone pirating in. Then she stopped, with it in her +hand, and said to herself: + +"No, I don't dare. Poor boy, I reckon he's lied about it--but it's a +blessed, blessed lie, there's such a comfort come from it. I hope the +Lord--I KNOW the Lord will forgive him, because it was such +goodheartedness in him to tell it. But I don't want to find out it's a +lie. I won't look." + +She put the jacket away, and stood by musing a minute. Twice she put +out her hand to take the garment again, and twice she refrained. Once +more she ventured, and this time she fortified herself with the +thought: "It's a good lie--it's a good lie--I won't let it grieve me." +So she sought the jacket pocket. A moment later she was reading Tom's +piece of bark through flowing tears and saying: "I could forgive the +boy, now, if he'd committed a million sins!" + + + +CHAPTER XX + +THERE was something about Aunt Polly's manner, when she kissed Tom, +that swept away his low spirits and made him lighthearted and happy +again. He started to school and had the luck of coming upon Becky +Thatcher at the head of Meadow Lane. His mood always determined his +manner. Without a moment's hesitation he ran to her and said: + +"I acted mighty mean to-day, Becky, and I'm so sorry. I won't ever, +ever do that way again, as long as ever I live--please make up, won't +you?" + +The girl stopped and looked him scornfully in the face: + +"I'll thank you to keep yourself TO yourself, Mr. Thomas Sawyer. I'll +never speak to you again." + +She tossed her head and passed on. Tom was so stunned that he had not +even presence of mind enough to say "Who cares, Miss Smarty?" until the +right time to say it had gone by. So he said nothing. But he was in a +fine rage, nevertheless. He moped into the schoolyard wishing she were +a boy, and imagining how he would trounce her if she were. He presently +encountered her and delivered a stinging remark as he passed. She +hurled one in return, and the angry breach was complete. It seemed to +Becky, in her hot resentment, that she could hardly wait for school to +"take in," she was so impatient to see Tom flogged for the injured +spelling-book. If she had had any lingering notion of exposing Alfred +Temple, Tom's offensive fling had driven it entirely away. + +Poor girl, she did not know how fast she was nearing trouble herself. +The master, Mr. Dobbins, had reached middle age with an unsatisfied +ambition. The darling of his desires was, to be a doctor, but poverty +had decreed that he should be nothing higher than a village +schoolmaster. Every day he took a mysterious book out of his desk and +absorbed himself in it at times when no classes were reciting. He kept +that book under lock and key. There was not an urchin in school but was +perishing to have a glimpse of it, but the chance never came. Every boy +and girl had a theory about the nature of that book; but no two +theories were alike, and there was no way of getting at the facts in +the case. Now, as Becky was passing by the desk, which stood near the +door, she noticed that the key was in the lock! It was a precious +moment. She glanced around; found herself alone, and the next instant +she had the book in her hands. The title-page--Professor Somebody's +ANATOMY--carried no information to her mind; so she began to turn the +leaves. She came at once upon a handsomely engraved and colored +frontispiece--a human figure, stark naked. At that moment a shadow fell +on the page and Tom Sawyer stepped in at the door and caught a glimpse +of the picture. Becky snatched at the book to close it, and had the +hard luck to tear the pictured page half down the middle. She thrust +the volume into the desk, turned the key, and burst out crying with +shame and vexation. + +"Tom Sawyer, you are just as mean as you can be, to sneak up on a +person and look at what they're looking at." + +"How could I know you was looking at anything?" + +"You ought to be ashamed of yourself, Tom Sawyer; you know you're +going to tell on me, and oh, what shall I do, what shall I do! I'll be +whipped, and I never was whipped in school." + +Then she stamped her little foot and said: + +"BE so mean if you want to! I know something that's going to happen. +You just wait and you'll see! Hateful, hateful, hateful!"--and she +flung out of the house with a new explosion of crying. + +Tom stood still, rather flustered by this onslaught. Presently he said +to himself: + +"What a curious kind of a fool a girl is! Never been licked in school! +Shucks! What's a licking! That's just like a girl--they're so +thin-skinned and chicken-hearted. Well, of course I ain't going to tell +old Dobbins on this little fool, because there's other ways of getting +even on her, that ain't so mean; but what of it? Old Dobbins will ask +who it was tore his book. Nobody'll answer. Then he'll do just the way +he always does--ask first one and then t'other, and when he comes to the +right girl he'll know it, without any telling. Girls' faces always tell +on them. They ain't got any backbone. She'll get licked. Well, it's a +kind of a tight place for Becky Thatcher, because there ain't any way +out of it." Tom conned the thing a moment longer, and then added: "All +right, though; she'd like to see me in just such a fix--let her sweat it +out!" + +Tom joined the mob of skylarking scholars outside. In a few moments +the master arrived and school "took in." Tom did not feel a strong +interest in his studies. Every time he stole a glance at the girls' +side of the room Becky's face troubled him. Considering all things, he +did not want to pity her, and yet it was all he could do to help it. He +could get up no exultation that was really worthy the name. Presently +the spelling-book discovery was made, and Tom's mind was entirely full +of his own matters for a while after that. Becky roused up from her +lethargy of distress and showed good interest in the proceedings. She +did not expect that Tom could get out of his trouble by denying that he +spilt the ink on the book himself; and she was right. The denial only +seemed to make the thing worse for Tom. Becky supposed she would be +glad of that, and she tried to believe she was glad of it, but she +found she was not certain. When the worst came to the worst, she had an +impulse to get up and tell on Alfred Temple, but she made an effort and +forced herself to keep still--because, said she to herself, "he'll tell +about me tearing the picture sure. I wouldn't say a word, not to save +his life!" + +Tom took his whipping and went back to his seat not at all +broken-hearted, for he thought it was possible that he had unknowingly +upset the ink on the spelling-book himself, in some skylarking bout--he +had denied it for form's sake and because it was custom, and had stuck +to the denial from principle. + +A whole hour drifted by, the master sat nodding in his throne, the air +was drowsy with the hum of study. By and by, Mr. Dobbins straightened +himself up, yawned, then unlocked his desk, and reached for his book, +but seemed undecided whether to take it out or leave it. Most of the +pupils glanced up languidly, but there were two among them that watched +his movements with intent eyes. Mr. Dobbins fingered his book absently +for a while, then took it out and settled himself in his chair to read! +Tom shot a glance at Becky. He had seen a hunted and helpless rabbit +look as she did, with a gun levelled at its head. Instantly he forgot +his quarrel with her. Quick--something must be done! done in a flash, +too! But the very imminence of the emergency paralyzed his invention. +Good!--he had an inspiration! He would run and snatch the book, spring +through the door and fly. But his resolution shook for one little +instant, and the chance was lost--the master opened the volume. If Tom +only had the wasted opportunity back again! Too late. There was no help +for Becky now, he said. The next moment the master faced the school. +Every eye sank under his gaze. There was that in it which smote even +the innocent with fear. There was silence while one might count ten +--the master was gathering his wrath. Then he spoke: "Who tore this book?" + +There was not a sound. One could have heard a pin drop. The stillness +continued; the master searched face after face for signs of guilt. + +"Benjamin Rogers, did you tear this book?" + +A denial. Another pause. + +"Joseph Harper, did you?" + +Another denial. Tom's uneasiness grew more and more intense under the +slow torture of these proceedings. The master scanned the ranks of +boys--considered a while, then turned to the girls: + +"Amy Lawrence?" + +A shake of the head. + +"Gracie Miller?" + +The same sign. + +"Susan Harper, did you do this?" + +Another negative. The next girl was Becky Thatcher. Tom was trembling +from head to foot with excitement and a sense of the hopelessness of +the situation. + +"Rebecca Thatcher" [Tom glanced at her face--it was white with terror] +--"did you tear--no, look me in the face" [her hands rose in appeal] +--"did you tear this book?" + +A thought shot like lightning through Tom's brain. He sprang to his +feet and shouted--"I done it!" + +The school stared in perplexity at this incredible folly. Tom stood a +moment, to gather his dismembered faculties; and when he stepped +forward to go to his punishment the surprise, the gratitude, the +adoration that shone upon him out of poor Becky's eyes seemed pay +enough for a hundred floggings. Inspired by the splendor of his own +act, he took without an outcry the most merciless flaying that even Mr. +Dobbins had ever administered; and also received with indifference the +added cruelty of a command to remain two hours after school should be +dismissed--for he knew who would wait for him outside till his +captivity was done, and not count the tedious time as loss, either. + +Tom went to bed that night planning vengeance against Alfred Temple; +for with shame and repentance Becky had told him all, not forgetting +her own treachery; but even the longing for vengeance had to give way, +soon, to pleasanter musings, and he fell asleep at last with Becky's +latest words lingering dreamily in his ear-- + +"Tom, how COULD you be so noble!" + + + +CHAPTER XXI + +VACATION was approaching. The schoolmaster, always severe, grew +severer and more exacting than ever, for he wanted the school to make a +good showing on "Examination" day. His rod and his ferule were seldom +idle now--at least among the smaller pupils. Only the biggest boys, and +young ladies of eighteen and twenty, escaped lashing. Mr. Dobbins' +lashings were very vigorous ones, too; for although he carried, under +his wig, a perfectly bald and shiny head, he had only reached middle +age, and there was no sign of feebleness in his muscle. As the great +day approached, all the tyranny that was in him came to the surface; he +seemed to take a vindictive pleasure in punishing the least +shortcomings. The consequence was, that the smaller boys spent their +days in terror and suffering and their nights in plotting revenge. They +threw away no opportunity to do the master a mischief. But he kept +ahead all the time. The retribution that followed every vengeful +success was so sweeping and majestic that the boys always retired from +the field badly worsted. At last they conspired together and hit upon a +plan that promised a dazzling victory. They swore in the sign-painter's +boy, told him the scheme, and asked his help. He had his own reasons +for being delighted, for the master boarded in his father's family and +had given the boy ample cause to hate him. The master's wife would go +on a visit to the country in a few days, and there would be nothing to +interfere with the plan; the master always prepared himself for great +occasions by getting pretty well fuddled, and the sign-painter's boy +said that when the dominie had reached the proper condition on +Examination Evening he would "manage the thing" while he napped in his +chair; then he would have him awakened at the right time and hurried +away to school. + +In the fulness of time the interesting occasion arrived. At eight in +the evening the schoolhouse was brilliantly lighted, and adorned with +wreaths and festoons of foliage and flowers. The master sat throned in +his great chair upon a raised platform, with his blackboard behind him. +He was looking tolerably mellow. Three rows of benches on each side and +six rows in front of him were occupied by the dignitaries of the town +and by the parents of the pupils. To his left, back of the rows of +citizens, was a spacious temporary platform upon which were seated the +scholars who were to take part in the exercises of the evening; rows of +small boys, washed and dressed to an intolerable state of discomfort; +rows of gawky big boys; snowbanks of girls and young ladies clad in +lawn and muslin and conspicuously conscious of their bare arms, their +grandmothers' ancient trinkets, their bits of pink and blue ribbon and +the flowers in their hair. All the rest of the house was filled with +non-participating scholars. + +The exercises began. A very little boy stood up and sheepishly +recited, "You'd scarce expect one of my age to speak in public on the +stage," etc.--accompanying himself with the painfully exact and +spasmodic gestures which a machine might have used--supposing the +machine to be a trifle out of order. But he got through safely, though +cruelly scared, and got a fine round of applause when he made his +manufactured bow and retired. + +A little shamefaced girl lisped, "Mary had a little lamb," etc., +performed a compassion-inspiring curtsy, got her meed of applause, and +sat down flushed and happy. + +Tom Sawyer stepped forward with conceited confidence and soared into +the unquenchable and indestructible "Give me liberty or give me death" +speech, with fine fury and frantic gesticulation, and broke down in the +middle of it. A ghastly stage-fright seized him, his legs quaked under +him and he was like to choke. True, he had the manifest sympathy of the +house but he had the house's silence, too, which was even worse than +its sympathy. The master frowned, and this completed the disaster. Tom +struggled awhile and then retired, utterly defeated. There was a weak +attempt at applause, but it died early. + +"The Boy Stood on the Burning Deck" followed; also "The Assyrian Came +Down," and other declamatory gems. Then there were reading exercises, +and a spelling fight. The meagre Latin class recited with honor. The +prime feature of the evening was in order, now--original "compositions" +by the young ladies. Each in her turn stepped forward to the edge of +the platform, cleared her throat, held up her manuscript (tied with +dainty ribbon), and proceeded to read, with labored attention to +"expression" and punctuation. The themes were the same that had been +illuminated upon similar occasions by their mothers before them, their +grandmothers, and doubtless all their ancestors in the female line +clear back to the Crusades. "Friendship" was one; "Memories of Other +Days"; "Religion in History"; "Dream Land"; "The Advantages of +Culture"; "Forms of Political Government Compared and Contrasted"; +"Melancholy"; "Filial Love"; "Heart Longings," etc., etc. + +A prevalent feature in these compositions was a nursed and petted +melancholy; another was a wasteful and opulent gush of "fine language"; +another was a tendency to lug in by the ears particularly prized words +and phrases until they were worn entirely out; and a peculiarity that +conspicuously marked and marred them was the inveterate and intolerable +sermon that wagged its crippled tail at the end of each and every one +of them. No matter what the subject might be, a brain-racking effort +was made to squirm it into some aspect or other that the moral and +religious mind could contemplate with edification. The glaring +insincerity of these sermons was not sufficient to compass the +banishment of the fashion from the schools, and it is not sufficient +to-day; it never will be sufficient while the world stands, perhaps. +There is no school in all our land where the young ladies do not feel +obliged to close their compositions with a sermon; and you will find +that the sermon of the most frivolous and the least religious girl in +the school is always the longest and the most relentlessly pious. But +enough of this. Homely truth is unpalatable. + +Let us return to the "Examination." The first composition that was +read was one entitled "Is this, then, Life?" Perhaps the reader can +endure an extract from it: + + "In the common walks of life, with what delightful + emotions does the youthful mind look forward to some + anticipated scene of festivity! Imagination is busy + sketching rose-tinted pictures of joy. In fancy, the + voluptuous votary of fashion sees herself amid the + festive throng, 'the observed of all observers.' Her + graceful form, arrayed in snowy robes, is whirling + through the mazes of the joyous dance; her eye is + brightest, her step is lightest in the gay assembly. + + "In such delicious fancies time quickly glides by, + and the welcome hour arrives for her entrance into + the Elysian world, of which she has had such bright + dreams. How fairy-like does everything appear to + her enchanted vision! Each new scene is more charming + than the last. But after a while she finds that + beneath this goodly exterior, all is vanity, the + flattery which once charmed her soul, now grates + harshly upon her ear; the ball-room has lost its + charms; and with wasted health and imbittered heart, + she turns away with the conviction that earthly + pleasures cannot satisfy the longings of the soul!" + +And so forth and so on. There was a buzz of gratification from time to +time during the reading, accompanied by whispered ejaculations of "How +sweet!" "How eloquent!" "So true!" etc., and after the thing had closed +with a peculiarly afflicting sermon the applause was enthusiastic. + +Then arose a slim, melancholy girl, whose face had the "interesting" +paleness that comes of pills and indigestion, and read a "poem." Two +stanzas of it will do: + + "A MISSOURI MAIDEN'S FAREWELL TO ALABAMA + + "Alabama, good-bye! I love thee well! + But yet for a while do I leave thee now! + Sad, yes, sad thoughts of thee my heart doth swell, + And burning recollections throng my brow! + For I have wandered through thy flowery woods; + Have roamed and read near Tallapoosa's stream; + Have listened to Tallassee's warring floods, + And wooed on Coosa's side Aurora's beam. + + "Yet shame I not to bear an o'er-full heart, + Nor blush to turn behind my tearful eyes; + 'Tis from no stranger land I now must part, + 'Tis to no strangers left I yield these sighs. + Welcome and home were mine within this State, + Whose vales I leave--whose spires fade fast from me + And cold must be mine eyes, and heart, and tete, + When, dear Alabama! they turn cold on thee!" + +There were very few there who knew what "tete" meant, but the poem was +very satisfactory, nevertheless. + +Next appeared a dark-complexioned, black-eyed, black-haired young +lady, who paused an impressive moment, assumed a tragic expression, and +began to read in a measured, solemn tone: + + "A VISION + + "Dark and tempestuous was night. Around the + throne on high not a single star quivered; but + the deep intonations of the heavy thunder + constantly vibrated upon the ear; whilst the + terrific lightning revelled in angry mood + through the cloudy chambers of heaven, seeming + to scorn the power exerted over its terror by + the illustrious Franklin! Even the boisterous + winds unanimously came forth from their mystic + homes, and blustered about as if to enhance by + their aid the wildness of the scene. + + "At such a time, so dark, so dreary, for human + sympathy my very spirit sighed; but instead thereof, + + "'My dearest friend, my counsellor, my comforter + and guide--My joy in grief, my second bliss + in joy,' came to my side. She moved like one of + those bright beings pictured in the sunny walks + of fancy's Eden by the romantic and young, a + queen of beauty unadorned save by her own + transcendent loveliness. So soft was her step, it + failed to make even a sound, and but for the + magical thrill imparted by her genial touch, as + other unobtrusive beauties, she would have glided + away un-perceived--unsought. A strange sadness + rested upon her features, like icy tears upon + the robe of December, as she pointed to the + contending elements without, and bade me contemplate + the two beings presented." + +This nightmare occupied some ten pages of manuscript and wound up with +a sermon so destructive of all hope to non-Presbyterians that it took +the first prize. This composition was considered to be the very finest +effort of the evening. The mayor of the village, in delivering the +prize to the author of it, made a warm speech in which he said that it +was by far the most "eloquent" thing he had ever listened to, and that +Daniel Webster himself might well be proud of it. + +It may be remarked, in passing, that the number of compositions in +which the word "beauteous" was over-fondled, and human experience +referred to as "life's page," was up to the usual average. + +Now the master, mellow almost to the verge of geniality, put his chair +aside, turned his back to the audience, and began to draw a map of +America on the blackboard, to exercise the geography class upon. But he +made a sad business of it with his unsteady hand, and a smothered +titter rippled over the house. He knew what the matter was, and set +himself to right it. He sponged out lines and remade them; but he only +distorted them more than ever, and the tittering was more pronounced. +He threw his entire attention upon his work, now, as if determined not +to be put down by the mirth. He felt that all eyes were fastened upon +him; he imagined he was succeeding, and yet the tittering continued; it +even manifestly increased. And well it might. There was a garret above, +pierced with a scuttle over his head; and down through this scuttle +came a cat, suspended around the haunches by a string; she had a rag +tied about her head and jaws to keep her from mewing; as she slowly +descended she curved upward and clawed at the string, she swung +downward and clawed at the intangible air. The tittering rose higher +and higher--the cat was within six inches of the absorbed teacher's +head--down, down, a little lower, and she grabbed his wig with her +desperate claws, clung to it, and was snatched up into the garret in an +instant with her trophy still in her possession! And how the light did +blaze abroad from the master's bald pate--for the sign-painter's boy +had GILDED it! + +That broke up the meeting. The boys were avenged. Vacation had come. + + NOTE:--The pretended "compositions" quoted in + this chapter are taken without alteration from a + volume entitled "Prose and Poetry, by a Western + Lady"--but they are exactly and precisely after + the schoolgirl pattern, and hence are much + happier than any mere imitations could be. + + + +CHAPTER XXII + +TOM joined the new order of Cadets of Temperance, being attracted by +the showy character of their "regalia." He promised to abstain from +smoking, chewing, and profanity as long as he remained a member. Now he +found out a new thing--namely, that to promise not to do a thing is the +surest way in the world to make a body want to go and do that very +thing. Tom soon found himself tormented with a desire to drink and +swear; the desire grew to be so intense that nothing but the hope of a +chance to display himself in his red sash kept him from withdrawing +from the order. Fourth of July was coming; but he soon gave that up +--gave it up before he had worn his shackles over forty-eight hours--and +fixed his hopes upon old Judge Frazer, justice of the peace, who was +apparently on his deathbed and would have a big public funeral, since +he was so high an official. During three days Tom was deeply concerned +about the Judge's condition and hungry for news of it. Sometimes his +hopes ran high--so high that he would venture to get out his regalia +and practise before the looking-glass. But the Judge had a most +discouraging way of fluctuating. At last he was pronounced upon the +mend--and then convalescent. Tom was disgusted; and felt a sense of +injury, too. He handed in his resignation at once--and that night the +Judge suffered a relapse and died. Tom resolved that he would never +trust a man like that again. + +The funeral was a fine thing. The Cadets paraded in a style calculated +to kill the late member with envy. Tom was a free boy again, however +--there was something in that. He could drink and swear, now--but found +to his surprise that he did not want to. The simple fact that he could, +took the desire away, and the charm of it. + +Tom presently wondered to find that his coveted vacation was beginning +to hang a little heavily on his hands. + +He attempted a diary--but nothing happened during three days, and so +he abandoned it. + +The first of all the negro minstrel shows came to town, and made a +sensation. Tom and Joe Harper got up a band of performers and were +happy for two days. + +Even the Glorious Fourth was in some sense a failure, for it rained +hard, there was no procession in consequence, and the greatest man in +the world (as Tom supposed), Mr. Benton, an actual United States +Senator, proved an overwhelming disappointment--for he was not +twenty-five feet high, nor even anywhere in the neighborhood of it. + +A circus came. The boys played circus for three days afterward in +tents made of rag carpeting--admission, three pins for boys, two for +girls--and then circusing was abandoned. + +A phrenologist and a mesmerizer came--and went again and left the +village duller and drearier than ever. + +There were some boys-and-girls' parties, but they were so few and so +delightful that they only made the aching voids between ache the harder. + +Becky Thatcher was gone to her Constantinople home to stay with her +parents during vacation--so there was no bright side to life anywhere. + +The dreadful secret of the murder was a chronic misery. It was a very +cancer for permanency and pain. + +Then came the measles. + +During two long weeks Tom lay a prisoner, dead to the world and its +happenings. He was very ill, he was interested in nothing. When he got +upon his feet at last and moved feebly down-town, a melancholy change +had come over everything and every creature. There had been a +"revival," and everybody had "got religion," not only the adults, but +even the boys and girls. Tom went about, hoping against hope for the +sight of one blessed sinful face, but disappointment crossed him +everywhere. He found Joe Harper studying a Testament, and turned sadly +away from the depressing spectacle. He sought Ben Rogers, and found him +visiting the poor with a basket of tracts. He hunted up Jim Hollis, who +called his attention to the precious blessing of his late measles as a +warning. Every boy he encountered added another ton to his depression; +and when, in desperation, he flew for refuge at last to the bosom of +Huckleberry Finn and was received with a Scriptural quotation, his +heart broke and he crept home and to bed realizing that he alone of all +the town was lost, forever and forever. + +And that night there came on a terrific storm, with driving rain, +awful claps of thunder and blinding sheets of lightning. He covered his +head with the bedclothes and waited in a horror of suspense for his +doom; for he had not the shadow of a doubt that all this hubbub was +about him. He believed he had taxed the forbearance of the powers above +to the extremity of endurance and that this was the result. It might +have seemed to him a waste of pomp and ammunition to kill a bug with a +battery of artillery, but there seemed nothing incongruous about the +getting up such an expensive thunderstorm as this to knock the turf +from under an insect like himself. + +By and by the tempest spent itself and died without accomplishing its +object. The boy's first impulse was to be grateful, and reform. His +second was to wait--for there might not be any more storms. + +The next day the doctors were back; Tom had relapsed. The three weeks +he spent on his back this time seemed an entire age. When he got abroad +at last he was hardly grateful that he had been spared, remembering how +lonely was his estate, how companionless and forlorn he was. He drifted +listlessly down the street and found Jim Hollis acting as judge in a +juvenile court that was trying a cat for murder, in the presence of her +victim, a bird. He found Joe Harper and Huck Finn up an alley eating a +stolen melon. Poor lads! they--like Tom--had suffered a relapse. + + + +CHAPTER XXIII + +AT last the sleepy atmosphere was stirred--and vigorously: the murder +trial came on in the court. It became the absorbing topic of village +talk immediately. Tom could not get away from it. Every reference to +the murder sent a shudder to his heart, for his troubled conscience and +fears almost persuaded him that these remarks were put forth in his +hearing as "feelers"; he did not see how he could be suspected of +knowing anything about the murder, but still he could not be +comfortable in the midst of this gossip. It kept him in a cold shiver +all the time. He took Huck to a lonely place to have a talk with him. +It would be some relief to unseal his tongue for a little while; to +divide his burden of distress with another sufferer. Moreover, he +wanted to assure himself that Huck had remained discreet. + +"Huck, have you ever told anybody about--that?" + +"'Bout what?" + +"You know what." + +"Oh--'course I haven't." + +"Never a word?" + +"Never a solitary word, so help me. What makes you ask?" + +"Well, I was afeard." + +"Why, Tom Sawyer, we wouldn't be alive two days if that got found out. +YOU know that." + +Tom felt more comfortable. After a pause: + +"Huck, they couldn't anybody get you to tell, could they?" + +"Get me to tell? Why, if I wanted that half-breed devil to drownd me +they could get me to tell. They ain't no different way." + +"Well, that's all right, then. I reckon we're safe as long as we keep +mum. But let's swear again, anyway. It's more surer." + +"I'm agreed." + +So they swore again with dread solemnities. + +"What is the talk around, Huck? I've heard a power of it." + +"Talk? Well, it's just Muff Potter, Muff Potter, Muff Potter all the +time. It keeps me in a sweat, constant, so's I want to hide som'ers." + +"That's just the same way they go on round me. I reckon he's a goner. +Don't you feel sorry for him, sometimes?" + +"Most always--most always. He ain't no account; but then he hain't +ever done anything to hurt anybody. Just fishes a little, to get money +to get drunk on--and loafs around considerable; but lord, we all do +that--leastways most of us--preachers and such like. But he's kind of +good--he give me half a fish, once, when there warn't enough for two; +and lots of times he's kind of stood by me when I was out of luck." + +"Well, he's mended kites for me, Huck, and knitted hooks on to my +line. I wish we could get him out of there." + +"My! we couldn't get him out, Tom. And besides, 'twouldn't do any +good; they'd ketch him again." + +"Yes--so they would. But I hate to hear 'em abuse him so like the +dickens when he never done--that." + +"I do too, Tom. Lord, I hear 'em say he's the bloodiest looking +villain in this country, and they wonder he wasn't ever hung before." + +"Yes, they talk like that, all the time. I've heard 'em say that if he +was to get free they'd lynch him." + +"And they'd do it, too." + +The boys had a long talk, but it brought them little comfort. As the +twilight drew on, they found themselves hanging about the neighborhood +of the little isolated jail, perhaps with an undefined hope that +something would happen that might clear away their difficulties. But +nothing happened; there seemed to be no angels or fairies interested in +this luckless captive. + +The boys did as they had often done before--went to the cell grating +and gave Potter some tobacco and matches. He was on the ground floor +and there were no guards. + +His gratitude for their gifts had always smote their consciences +before--it cut deeper than ever, this time. They felt cowardly and +treacherous to the last degree when Potter said: + +"You've been mighty good to me, boys--better'n anybody else in this +town. And I don't forget it, I don't. Often I says to myself, says I, +'I used to mend all the boys' kites and things, and show 'em where the +good fishin' places was, and befriend 'em what I could, and now they've +all forgot old Muff when he's in trouble; but Tom don't, and Huck +don't--THEY don't forget him, says I, 'and I don't forget them.' Well, +boys, I done an awful thing--drunk and crazy at the time--that's the +only way I account for it--and now I got to swing for it, and it's +right. Right, and BEST, too, I reckon--hope so, anyway. Well, we won't +talk about that. I don't want to make YOU feel bad; you've befriended +me. But what I want to say, is, don't YOU ever get drunk--then you won't +ever get here. Stand a litter furder west--so--that's it; it's a prime +comfort to see faces that's friendly when a body's in such a muck of +trouble, and there don't none come here but yourn. Good friendly +faces--good friendly faces. Git up on one another's backs and let me +touch 'em. That's it. Shake hands--yourn'll come through the bars, but +mine's too big. Little hands, and weak--but they've helped Muff Potter +a power, and they'd help him more if they could." + +Tom went home miserable, and his dreams that night were full of +horrors. The next day and the day after, he hung about the court-room, +drawn by an almost irresistible impulse to go in, but forcing himself +to stay out. Huck was having the same experience. They studiously +avoided each other. Each wandered away, from time to time, but the same +dismal fascination always brought them back presently. Tom kept his +ears open when idlers sauntered out of the court-room, but invariably +heard distressing news--the toils were closing more and more +relentlessly around poor Potter. At the end of the second day the +village talk was to the effect that Injun Joe's evidence stood firm and +unshaken, and that there was not the slightest question as to what the +jury's verdict would be. + +Tom was out late, that night, and came to bed through the window. He +was in a tremendous state of excitement. It was hours before he got to +sleep. All the village flocked to the court-house the next morning, for +this was to be the great day. Both sexes were about equally represented +in the packed audience. After a long wait the jury filed in and took +their places; shortly afterward, Potter, pale and haggard, timid and +hopeless, was brought in, with chains upon him, and seated where all +the curious eyes could stare at him; no less conspicuous was Injun Joe, +stolid as ever. There was another pause, and then the judge arrived and +the sheriff proclaimed the opening of the court. The usual whisperings +among the lawyers and gathering together of papers followed. These +details and accompanying delays worked up an atmosphere of preparation +that was as impressive as it was fascinating. + +Now a witness was called who testified that he found Muff Potter +washing in the brook, at an early hour of the morning that the murder +was discovered, and that he immediately sneaked away. After some +further questioning, counsel for the prosecution said: + +"Take the witness." + +The prisoner raised his eyes for a moment, but dropped them again when +his own counsel said: + +"I have no questions to ask him." + +The next witness proved the finding of the knife near the corpse. +Counsel for the prosecution said: + +"Take the witness." + +"I have no questions to ask him," Potter's lawyer replied. + +A third witness swore he had often seen the knife in Potter's +possession. + +"Take the witness." + +Counsel for Potter declined to question him. The faces of the audience +began to betray annoyance. Did this attorney mean to throw away his +client's life without an effort? + +Several witnesses deposed concerning Potter's guilty behavior when +brought to the scene of the murder. They were allowed to leave the +stand without being cross-questioned. + +Every detail of the damaging circumstances that occurred in the +graveyard upon that morning which all present remembered so well was +brought out by credible witnesses, but none of them were cross-examined +by Potter's lawyer. The perplexity and dissatisfaction of the house +expressed itself in murmurs and provoked a reproof from the bench. +Counsel for the prosecution now said: + +"By the oaths of citizens whose simple word is above suspicion, we +have fastened this awful crime, beyond all possibility of question, +upon the unhappy prisoner at the bar. We rest our case here." + +A groan escaped from poor Potter, and he put his face in his hands and +rocked his body softly to and fro, while a painful silence reigned in +the court-room. Many men were moved, and many women's compassion +testified itself in tears. Counsel for the defence rose and said: + +"Your honor, in our remarks at the opening of this trial, we +foreshadowed our purpose to prove that our client did this fearful deed +while under the influence of a blind and irresponsible delirium +produced by drink. We have changed our mind. We shall not offer that +plea." [Then to the clerk:] "Call Thomas Sawyer!" + +A puzzled amazement awoke in every face in the house, not even +excepting Potter's. Every eye fastened itself with wondering interest +upon Tom as he rose and took his place upon the stand. The boy looked +wild enough, for he was badly scared. The oath was administered. + +"Thomas Sawyer, where were you on the seventeenth of June, about the +hour of midnight?" + +Tom glanced at Injun Joe's iron face and his tongue failed him. The +audience listened breathless, but the words refused to come. After a +few moments, however, the boy got a little of his strength back, and +managed to put enough of it into his voice to make part of the house +hear: + +"In the graveyard!" + +"A little bit louder, please. Don't be afraid. You were--" + +"In the graveyard." + +A contemptuous smile flitted across Injun Joe's face. + +"Were you anywhere near Horse Williams' grave?" + +"Yes, sir." + +"Speak up--just a trifle louder. How near were you?" + +"Near as I am to you." + +"Were you hidden, or not?" + +"I was hid." + +"Where?" + +"Behind the elms that's on the edge of the grave." + +Injun Joe gave a barely perceptible start. + +"Any one with you?" + +"Yes, sir. I went there with--" + +"Wait--wait a moment. Never mind mentioning your companion's name. We +will produce him at the proper time. Did you carry anything there with +you." + +Tom hesitated and looked confused. + +"Speak out, my boy--don't be diffident. The truth is always +respectable. What did you take there?" + +"Only a--a--dead cat." + +There was a ripple of mirth, which the court checked. + +"We will produce the skeleton of that cat. Now, my boy, tell us +everything that occurred--tell it in your own way--don't skip anything, +and don't be afraid." + +Tom began--hesitatingly at first, but as he warmed to his subject his +words flowed more and more easily; in a little while every sound ceased +but his own voice; every eye fixed itself upon him; with parted lips +and bated breath the audience hung upon his words, taking no note of +time, rapt in the ghastly fascinations of the tale. The strain upon +pent emotion reached its climax when the boy said: + +"--and as the doctor fetched the board around and Muff Potter fell, +Injun Joe jumped with the knife and--" + +Crash! Quick as lightning the half-breed sprang for a window, tore his +way through all opposers, and was gone! + + + +CHAPTER XXIV + +TOM was a glittering hero once more--the pet of the old, the envy of +the young. His name even went into immortal print, for the village +paper magnified him. There were some that believed he would be +President, yet, if he escaped hanging. + +As usual, the fickle, unreasoning world took Muff Potter to its bosom +and fondled him as lavishly as it had abused him before. But that sort +of conduct is to the world's credit; therefore it is not well to find +fault with it. + +Tom's days were days of splendor and exultation to him, but his nights +were seasons of horror. Injun Joe infested all his dreams, and always +with doom in his eye. Hardly any temptation could persuade the boy to +stir abroad after nightfall. Poor Huck was in the same state of +wretchedness and terror, for Tom had told the whole story to the lawyer +the night before the great day of the trial, and Huck was sore afraid +that his share in the business might leak out, yet, notwithstanding +Injun Joe's flight had saved him the suffering of testifying in court. +The poor fellow had got the attorney to promise secrecy, but what of +that? Since Tom's harassed conscience had managed to drive him to the +lawyer's house by night and wring a dread tale from lips that had been +sealed with the dismalest and most formidable of oaths, Huck's +confidence in the human race was well-nigh obliterated. + +Daily Muff Potter's gratitude made Tom glad he had spoken; but nightly +he wished he had sealed up his tongue. + +Half the time Tom was afraid Injun Joe would never be captured; the +other half he was afraid he would be. He felt sure he never could draw +a safe breath again until that man was dead and he had seen the corpse. + +Rewards had been offered, the country had been scoured, but no Injun +Joe was found. One of those omniscient and awe-inspiring marvels, a +detective, came up from St. Louis, moused around, shook his head, +looked wise, and made that sort of astounding success which members of +that craft usually achieve. That is to say, he "found a clew." But you +can't hang a "clew" for murder, and so after that detective had got +through and gone home, Tom felt just as insecure as he was before. + +The slow days drifted on, and each left behind it a slightly lightened +weight of apprehension. + + + +CHAPTER XXV + +THERE comes a time in every rightly-constructed boy's life when he has +a raging desire to go somewhere and dig for hidden treasure. This +desire suddenly came upon Tom one day. He sallied out to find Joe +Harper, but failed of success. Next he sought Ben Rogers; he had gone +fishing. Presently he stumbled upon Huck Finn the Red-Handed. Huck +would answer. Tom took him to a private place and opened the matter to +him confidentially. Huck was willing. Huck was always willing to take a +hand in any enterprise that offered entertainment and required no +capital, for he had a troublesome superabundance of that sort of time +which is not money. "Where'll we dig?" said Huck. + +"Oh, most anywhere." + +"Why, is it hid all around?" + +"No, indeed it ain't. It's hid in mighty particular places, Huck +--sometimes on islands, sometimes in rotten chests under the end of a +limb of an old dead tree, just where the shadow falls at midnight; but +mostly under the floor in ha'nted houses." + +"Who hides it?" + +"Why, robbers, of course--who'd you reckon? Sunday-school +sup'rintendents?" + +"I don't know. If 'twas mine I wouldn't hide it; I'd spend it and have +a good time." + +"So would I. But robbers don't do that way. They always hide it and +leave it there." + +"Don't they come after it any more?" + +"No, they think they will, but they generally forget the marks, or +else they die. Anyway, it lays there a long time and gets rusty; and by +and by somebody finds an old yellow paper that tells how to find the +marks--a paper that's got to be ciphered over about a week because it's +mostly signs and hy'roglyphics." + +"Hyro--which?" + +"Hy'roglyphics--pictures and things, you know, that don't seem to mean +anything." + +"Have you got one of them papers, Tom?" + +"No." + +"Well then, how you going to find the marks?" + +"I don't want any marks. They always bury it under a ha'nted house or +on an island, or under a dead tree that's got one limb sticking out. +Well, we've tried Jackson's Island a little, and we can try it again +some time; and there's the old ha'nted house up the Still-House branch, +and there's lots of dead-limb trees--dead loads of 'em." + +"Is it under all of them?" + +"How you talk! No!" + +"Then how you going to know which one to go for?" + +"Go for all of 'em!" + +"Why, Tom, it'll take all summer." + +"Well, what of that? Suppose you find a brass pot with a hundred +dollars in it, all rusty and gray, or rotten chest full of di'monds. +How's that?" + +Huck's eyes glowed. + +"That's bully. Plenty bully enough for me. Just you gimme the hundred +dollars and I don't want no di'monds." + +"All right. But I bet you I ain't going to throw off on di'monds. Some +of 'em's worth twenty dollars apiece--there ain't any, hardly, but's +worth six bits or a dollar." + +"No! Is that so?" + +"Cert'nly--anybody'll tell you so. Hain't you ever seen one, Huck?" + +"Not as I remember." + +"Oh, kings have slathers of them." + +"Well, I don' know no kings, Tom." + +"I reckon you don't. But if you was to go to Europe you'd see a raft +of 'em hopping around." + +"Do they hop?" + +"Hop?--your granny! No!" + +"Well, what did you say they did, for?" + +"Shucks, I only meant you'd SEE 'em--not hopping, of course--what do +they want to hop for?--but I mean you'd just see 'em--scattered around, +you know, in a kind of a general way. Like that old humpbacked Richard." + +"Richard? What's his other name?" + +"He didn't have any other name. Kings don't have any but a given name." + +"No?" + +"But they don't." + +"Well, if they like it, Tom, all right; but I don't want to be a king +and have only just a given name, like a nigger. But say--where you +going to dig first?" + +"Well, I don't know. S'pose we tackle that old dead-limb tree on the +hill t'other side of Still-House branch?" + +"I'm agreed." + +So they got a crippled pick and a shovel, and set out on their +three-mile tramp. They arrived hot and panting, and threw themselves +down in the shade of a neighboring elm to rest and have a smoke. + +"I like this," said Tom. + +"So do I." + +"Say, Huck, if we find a treasure here, what you going to do with your +share?" + +"Well, I'll have pie and a glass of soda every day, and I'll go to +every circus that comes along. I bet I'll have a gay time." + +"Well, ain't you going to save any of it?" + +"Save it? What for?" + +"Why, so as to have something to live on, by and by." + +"Oh, that ain't any use. Pap would come back to thish-yer town some +day and get his claws on it if I didn't hurry up, and I tell you he'd +clean it out pretty quick. What you going to do with yourn, Tom?" + +"I'm going to buy a new drum, and a sure-'nough sword, and a red +necktie and a bull pup, and get married." + +"Married!" + +"That's it." + +"Tom, you--why, you ain't in your right mind." + +"Wait--you'll see." + +"Well, that's the foolishest thing you could do. Look at pap and my +mother. Fight! Why, they used to fight all the time. I remember, mighty +well." + +"That ain't anything. The girl I'm going to marry won't fight." + +"Tom, I reckon they're all alike. They'll all comb a body. Now you +better think 'bout this awhile. I tell you you better. What's the name +of the gal?" + +"It ain't a gal at all--it's a girl." + +"It's all the same, I reckon; some says gal, some says girl--both's +right, like enough. Anyway, what's her name, Tom?" + +"I'll tell you some time--not now." + +"All right--that'll do. Only if you get married I'll be more lonesomer +than ever." + +"No you won't. You'll come and live with me. Now stir out of this and +we'll go to digging." + +They worked and sweated for half an hour. No result. They toiled +another half-hour. Still no result. Huck said: + +"Do they always bury it as deep as this?" + +"Sometimes--not always. Not generally. I reckon we haven't got the +right place." + +So they chose a new spot and began again. The labor dragged a little, +but still they made progress. They pegged away in silence for some +time. Finally Huck leaned on his shovel, swabbed the beaded drops from +his brow with his sleeve, and said: + +"Where you going to dig next, after we get this one?" + +"I reckon maybe we'll tackle the old tree that's over yonder on +Cardiff Hill back of the widow's." + +"I reckon that'll be a good one. But won't the widow take it away from +us, Tom? It's on her land." + +"SHE take it away! Maybe she'd like to try it once. Whoever finds one +of these hid treasures, it belongs to him. It don't make any difference +whose land it's on." + +That was satisfactory. The work went on. By and by Huck said: + +"Blame it, we must be in the wrong place again. What do you think?" + +"It is mighty curious, Huck. I don't understand it. Sometimes witches +interfere. I reckon maybe that's what's the trouble now." + +"Shucks! Witches ain't got no power in the daytime." + +"Well, that's so. I didn't think of that. Oh, I know what the matter +is! What a blamed lot of fools we are! You got to find out where the +shadow of the limb falls at midnight, and that's where you dig!" + +"Then consound it, we've fooled away all this work for nothing. Now +hang it all, we got to come back in the night. It's an awful long way. +Can you get out?" + +"I bet I will. We've got to do it to-night, too, because if somebody +sees these holes they'll know in a minute what's here and they'll go +for it." + +"Well, I'll come around and maow to-night." + +"All right. Let's hide the tools in the bushes." + +The boys were there that night, about the appointed time. They sat in +the shadow waiting. It was a lonely place, and an hour made solemn by +old traditions. Spirits whispered in the rustling leaves, ghosts lurked +in the murky nooks, the deep baying of a hound floated up out of the +distance, an owl answered with his sepulchral note. The boys were +subdued by these solemnities, and talked little. By and by they judged +that twelve had come; they marked where the shadow fell, and began to +dig. Their hopes commenced to rise. Their interest grew stronger, and +their industry kept pace with it. The hole deepened and still deepened, +but every time their hearts jumped to hear the pick strike upon +something, they only suffered a new disappointment. It was only a stone +or a chunk. At last Tom said: + +"It ain't any use, Huck, we're wrong again." + +"Well, but we CAN'T be wrong. We spotted the shadder to a dot." + +"I know it, but then there's another thing." + +"What's that?". + +"Why, we only guessed at the time. Like enough it was too late or too +early." + +Huck dropped his shovel. + +"That's it," said he. "That's the very trouble. We got to give this +one up. We can't ever tell the right time, and besides this kind of +thing's too awful, here this time of night with witches and ghosts +a-fluttering around so. I feel as if something's behind me all the time; +and I'm afeard to turn around, becuz maybe there's others in front +a-waiting for a chance. I been creeping all over, ever since I got here." + +"Well, I've been pretty much so, too, Huck. They most always put in a +dead man when they bury a treasure under a tree, to look out for it." + +"Lordy!" + +"Yes, they do. I've always heard that." + +"Tom, I don't like to fool around much where there's dead people. A +body's bound to get into trouble with 'em, sure." + +"I don't like to stir 'em up, either. S'pose this one here was to +stick his skull out and say something!" + +"Don't Tom! It's awful." + +"Well, it just is. Huck, I don't feel comfortable a bit." + +"Say, Tom, let's give this place up, and try somewheres else." + +"All right, I reckon we better." + +"What'll it be?" + +Tom considered awhile; and then said: + +"The ha'nted house. That's it!" + +"Blame it, I don't like ha'nted houses, Tom. Why, they're a dern sight +worse'n dead people. Dead people might talk, maybe, but they don't come +sliding around in a shroud, when you ain't noticing, and peep over your +shoulder all of a sudden and grit their teeth, the way a ghost does. I +couldn't stand such a thing as that, Tom--nobody could." + +"Yes, but, Huck, ghosts don't travel around only at night. They won't +hender us from digging there in the daytime." + +"Well, that's so. But you know mighty well people don't go about that +ha'nted house in the day nor the night." + +"Well, that's mostly because they don't like to go where a man's been +murdered, anyway--but nothing's ever been seen around that house except +in the night--just some blue lights slipping by the windows--no regular +ghosts." + +"Well, where you see one of them blue lights flickering around, Tom, +you can bet there's a ghost mighty close behind it. It stands to +reason. Becuz you know that they don't anybody but ghosts use 'em." + +"Yes, that's so. But anyway they don't come around in the daytime, so +what's the use of our being afeard?" + +"Well, all right. We'll tackle the ha'nted house if you say so--but I +reckon it's taking chances." + +They had started down the hill by this time. There in the middle of +the moonlit valley below them stood the "ha'nted" house, utterly +isolated, its fences gone long ago, rank weeds smothering the very +doorsteps, the chimney crumbled to ruin, the window-sashes vacant, a +corner of the roof caved in. The boys gazed awhile, half expecting to +see a blue light flit past a window; then talking in a low tone, as +befitted the time and the circumstances, they struck far off to the +right, to give the haunted house a wide berth, and took their way +homeward through the woods that adorned the rearward side of Cardiff +Hill. + + + +CHAPTER XXVI + +ABOUT noon the next day the boys arrived at the dead tree; they had +come for their tools. Tom was impatient to go to the haunted house; +Huck was measurably so, also--but suddenly said: + +"Lookyhere, Tom, do you know what day it is?" + +Tom mentally ran over the days of the week, and then quickly lifted +his eyes with a startled look in them-- + +"My! I never once thought of it, Huck!" + +"Well, I didn't neither, but all at once it popped onto me that it was +Friday." + +"Blame it, a body can't be too careful, Huck. We might 'a' got into an +awful scrape, tackling such a thing on a Friday." + +"MIGHT! Better say we WOULD! There's some lucky days, maybe, but +Friday ain't." + +"Any fool knows that. I don't reckon YOU was the first that found it +out, Huck." + +"Well, I never said I was, did I? And Friday ain't all, neither. I had +a rotten bad dream last night--dreampt about rats." + +"No! Sure sign of trouble. Did they fight?" + +"No." + +"Well, that's good, Huck. When they don't fight it's only a sign that +there's trouble around, you know. All we got to do is to look mighty +sharp and keep out of it. We'll drop this thing for to-day, and play. +Do you know Robin Hood, Huck?" + +"No. Who's Robin Hood?" + +"Why, he was one of the greatest men that was ever in England--and the +best. He was a robber." + +"Cracky, I wisht I was. Who did he rob?" + +"Only sheriffs and bishops and rich people and kings, and such like. +But he never bothered the poor. He loved 'em. He always divided up with +'em perfectly square." + +"Well, he must 'a' been a brick." + +"I bet you he was, Huck. Oh, he was the noblest man that ever was. +They ain't any such men now, I can tell you. He could lick any man in +England, with one hand tied behind him; and he could take his yew bow +and plug a ten-cent piece every time, a mile and a half." + +"What's a YEW bow?" + +"I don't know. It's some kind of a bow, of course. And if he hit that +dime only on the edge he would set down and cry--and curse. But we'll +play Robin Hood--it's nobby fun. I'll learn you." + +"I'm agreed." + +So they played Robin Hood all the afternoon, now and then casting a +yearning eye down upon the haunted house and passing a remark about the +morrow's prospects and possibilities there. As the sun began to sink +into the west they took their way homeward athwart the long shadows of +the trees and soon were buried from sight in the forests of Cardiff +Hill. + +On Saturday, shortly after noon, the boys were at the dead tree again. +They had a smoke and a chat in the shade, and then dug a little in +their last hole, not with great hope, but merely because Tom said there +were so many cases where people had given up a treasure after getting +down within six inches of it, and then somebody else had come along and +turned it up with a single thrust of a shovel. The thing failed this +time, however, so the boys shouldered their tools and went away feeling +that they had not trifled with fortune, but had fulfilled all the +requirements that belong to the business of treasure-hunting. + +When they reached the haunted house there was something so weird and +grisly about the dead silence that reigned there under the baking sun, +and something so depressing about the loneliness and desolation of the +place, that they were afraid, for a moment, to venture in. Then they +crept to the door and took a trembling peep. They saw a weed-grown, +floorless room, unplastered, an ancient fireplace, vacant windows, a +ruinous staircase; and here, there, and everywhere hung ragged and +abandoned cobwebs. They presently entered, softly, with quickened +pulses, talking in whispers, ears alert to catch the slightest sound, +and muscles tense and ready for instant retreat. + +In a little while familiarity modified their fears and they gave the +place a critical and interested examination, rather admiring their own +boldness, and wondering at it, too. Next they wanted to look up-stairs. +This was something like cutting off retreat, but they got to daring +each other, and of course there could be but one result--they threw +their tools into a corner and made the ascent. Up there were the same +signs of decay. In one corner they found a closet that promised +mystery, but the promise was a fraud--there was nothing in it. Their +courage was up now and well in hand. They were about to go down and +begin work when-- + +"Sh!" said Tom. + +"What is it?" whispered Huck, blanching with fright. + +"Sh!... There!... Hear it?" + +"Yes!... Oh, my! Let's run!" + +"Keep still! Don't you budge! They're coming right toward the door." + +The boys stretched themselves upon the floor with their eyes to +knot-holes in the planking, and lay waiting, in a misery of fear. + +"They've stopped.... No--coming.... Here they are. Don't whisper +another word, Huck. My goodness, I wish I was out of this!" + +Two men entered. Each boy said to himself: "There's the old deaf and +dumb Spaniard that's been about town once or twice lately--never saw +t'other man before." + +"T'other" was a ragged, unkempt creature, with nothing very pleasant +in his face. The Spaniard was wrapped in a serape; he had bushy white +whiskers; long white hair flowed from under his sombrero, and he wore +green goggles. When they came in, "t'other" was talking in a low voice; +they sat down on the ground, facing the door, with their backs to the +wall, and the speaker continued his remarks. His manner became less +guarded and his words more distinct as he proceeded: + +"No," said he, "I've thought it all over, and I don't like it. It's +dangerous." + +"Dangerous!" grunted the "deaf and dumb" Spaniard--to the vast +surprise of the boys. "Milksop!" + +This voice made the boys gasp and quake. It was Injun Joe's! There was +silence for some time. Then Joe said: + +"What's any more dangerous than that job up yonder--but nothing's come +of it." + +"That's different. Away up the river so, and not another house about. +'Twon't ever be known that we tried, anyway, long as we didn't succeed." + +"Well, what's more dangerous than coming here in the daytime!--anybody +would suspicion us that saw us." + +"I know that. But there warn't any other place as handy after that +fool of a job. I want to quit this shanty. I wanted to yesterday, only +it warn't any use trying to stir out of here, with those infernal boys +playing over there on the hill right in full view." + +"Those infernal boys" quaked again under the inspiration of this +remark, and thought how lucky it was that they had remembered it was +Friday and concluded to wait a day. They wished in their hearts they +had waited a year. + +The two men got out some food and made a luncheon. After a long and +thoughtful silence, Injun Joe said: + +"Look here, lad--you go back up the river where you belong. Wait there +till you hear from me. I'll take the chances on dropping into this town +just once more, for a look. We'll do that 'dangerous' job after I've +spied around a little and think things look well for it. Then for +Texas! We'll leg it together!" + +This was satisfactory. Both men presently fell to yawning, and Injun +Joe said: + +"I'm dead for sleep! It's your turn to watch." + +He curled down in the weeds and soon began to snore. His comrade +stirred him once or twice and he became quiet. Presently the watcher +began to nod; his head drooped lower and lower, both men began to snore +now. + +The boys drew a long, grateful breath. Tom whispered: + +"Now's our chance--come!" + +Huck said: + +"I can't--I'd die if they was to wake." + +Tom urged--Huck held back. At last Tom rose slowly and softly, and +started alone. But the first step he made wrung such a hideous creak +from the crazy floor that he sank down almost dead with fright. He +never made a second attempt. The boys lay there counting the dragging +moments till it seemed to them that time must be done and eternity +growing gray; and then they were grateful to note that at last the sun +was setting. + +Now one snore ceased. Injun Joe sat up, stared around--smiled grimly +upon his comrade, whose head was drooping upon his knees--stirred him +up with his foot and said: + +"Here! YOU'RE a watchman, ain't you! All right, though--nothing's +happened." + +"My! have I been asleep?" + +"Oh, partly, partly. Nearly time for us to be moving, pard. What'll we +do with what little swag we've got left?" + +"I don't know--leave it here as we've always done, I reckon. No use to +take it away till we start south. Six hundred and fifty in silver's +something to carry." + +"Well--all right--it won't matter to come here once more." + +"No--but I'd say come in the night as we used to do--it's better." + +"Yes: but look here; it may be a good while before I get the right +chance at that job; accidents might happen; 'tain't in such a very good +place; we'll just regularly bury it--and bury it deep." + +"Good idea," said the comrade, who walked across the room, knelt down, +raised one of the rearward hearth-stones and took out a bag that +jingled pleasantly. He subtracted from it twenty or thirty dollars for +himself and as much for Injun Joe, and passed the bag to the latter, +who was on his knees in the corner, now, digging with his bowie-knife. + +The boys forgot all their fears, all their miseries in an instant. +With gloating eyes they watched every movement. Luck!--the splendor of +it was beyond all imagination! Six hundred dollars was money enough to +make half a dozen boys rich! Here was treasure-hunting under the +happiest auspices--there would not be any bothersome uncertainty as to +where to dig. They nudged each other every moment--eloquent nudges and +easily understood, for they simply meant--"Oh, but ain't you glad NOW +we're here!" + +Joe's knife struck upon something. + +"Hello!" said he. + +"What is it?" said his comrade. + +"Half-rotten plank--no, it's a box, I believe. Here--bear a hand and +we'll see what it's here for. Never mind, I've broke a hole." + +He reached his hand in and drew it out-- + +"Man, it's money!" + +The two men examined the handful of coins. They were gold. The boys +above were as excited as themselves, and as delighted. + +Joe's comrade said: + +"We'll make quick work of this. There's an old rusty pick over amongst +the weeds in the corner the other side of the fireplace--I saw it a +minute ago." + +He ran and brought the boys' pick and shovel. Injun Joe took the pick, +looked it over critically, shook his head, muttered something to +himself, and then began to use it. The box was soon unearthed. It was +not very large; it was iron bound and had been very strong before the +slow years had injured it. The men contemplated the treasure awhile in +blissful silence. + +"Pard, there's thousands of dollars here," said Injun Joe. + +"'Twas always said that Murrel's gang used to be around here one +summer," the stranger observed. + +"I know it," said Injun Joe; "and this looks like it, I should say." + +"Now you won't need to do that job." + +The half-breed frowned. Said he: + +"You don't know me. Least you don't know all about that thing. 'Tain't +robbery altogether--it's REVENGE!" and a wicked light flamed in his +eyes. "I'll need your help in it. When it's finished--then Texas. Go +home to your Nance and your kids, and stand by till you hear from me." + +"Well--if you say so; what'll we do with this--bury it again?" + +"Yes. [Ravishing delight overhead.] NO! by the great Sachem, no! +[Profound distress overhead.] I'd nearly forgot. That pick had fresh +earth on it! [The boys were sick with terror in a moment.] What +business has a pick and a shovel here? What business with fresh earth +on them? Who brought them here--and where are they gone? Have you heard +anybody?--seen anybody? What! bury it again and leave them to come and +see the ground disturbed? Not exactly--not exactly. We'll take it to my +den." + +"Why, of course! Might have thought of that before. You mean Number +One?" + +"No--Number Two--under the cross. The other place is bad--too common." + +"All right. It's nearly dark enough to start." + +Injun Joe got up and went about from window to window cautiously +peeping out. Presently he said: + +"Who could have brought those tools here? Do you reckon they can be +up-stairs?" + +The boys' breath forsook them. Injun Joe put his hand on his knife, +halted a moment, undecided, and then turned toward the stairway. The +boys thought of the closet, but their strength was gone. The steps came +creaking up the stairs--the intolerable distress of the situation woke +the stricken resolution of the lads--they were about to spring for the +closet, when there was a crash of rotten timbers and Injun Joe landed +on the ground amid the debris of the ruined stairway. He gathered +himself up cursing, and his comrade said: + +"Now what's the use of all that? If it's anybody, and they're up +there, let them STAY there--who cares? If they want to jump down, now, +and get into trouble, who objects? It will be dark in fifteen minutes +--and then let them follow us if they want to. I'm willing. In my +opinion, whoever hove those things in here caught a sight of us and +took us for ghosts or devils or something. I'll bet they're running +yet." + +Joe grumbled awhile; then he agreed with his friend that what daylight +was left ought to be economized in getting things ready for leaving. +Shortly afterward they slipped out of the house in the deepening +twilight, and moved toward the river with their precious box. + +Tom and Huck rose up, weak but vastly relieved, and stared after them +through the chinks between the logs of the house. Follow? Not they. +They were content to reach ground again without broken necks, and take +the townward track over the hill. They did not talk much. They were too +much absorbed in hating themselves--hating the ill luck that made them +take the spade and the pick there. But for that, Injun Joe never would +have suspected. He would have hidden the silver with the gold to wait +there till his "revenge" was satisfied, and then he would have had the +misfortune to find that money turn up missing. Bitter, bitter luck that +the tools were ever brought there! + +They resolved to keep a lookout for that Spaniard when he should come +to town spying out for chances to do his revengeful job, and follow him +to "Number Two," wherever that might be. Then a ghastly thought +occurred to Tom. + +"Revenge? What if he means US, Huck!" + +"Oh, don't!" said Huck, nearly fainting. + +They talked it all over, and as they entered town they agreed to +believe that he might possibly mean somebody else--at least that he +might at least mean nobody but Tom, since only Tom had testified. + +Very, very small comfort it was to Tom to be alone in danger! Company +would be a palpable improvement, he thought. + + + +CHAPTER XXVII + +THE adventure of the day mightily tormented Tom's dreams that night. +Four times he had his hands on that rich treasure and four times it +wasted to nothingness in his fingers as sleep forsook him and +wakefulness brought back the hard reality of his misfortune. As he lay +in the early morning recalling the incidents of his great adventure, he +noticed that they seemed curiously subdued and far away--somewhat as if +they had happened in another world, or in a time long gone by. Then it +occurred to him that the great adventure itself must be a dream! There +was one very strong argument in favor of this idea--namely, that the +quantity of coin he had seen was too vast to be real. He had never seen +as much as fifty dollars in one mass before, and he was like all boys +of his age and station in life, in that he imagined that all references +to "hundreds" and "thousands" were mere fanciful forms of speech, and +that no such sums really existed in the world. He never had supposed +for a moment that so large a sum as a hundred dollars was to be found +in actual money in any one's possession. If his notions of hidden +treasure had been analyzed, they would have been found to consist of a +handful of real dimes and a bushel of vague, splendid, ungraspable +dollars. + +But the incidents of his adventure grew sensibly sharper and clearer +under the attrition of thinking them over, and so he presently found +himself leaning to the impression that the thing might not have been a +dream, after all. This uncertainty must be swept away. He would snatch +a hurried breakfast and go and find Huck. Huck was sitting on the +gunwale of a flatboat, listlessly dangling his feet in the water and +looking very melancholy. Tom concluded to let Huck lead up to the +subject. If he did not do it, then the adventure would be proved to +have been only a dream. + +"Hello, Huck!" + +"Hello, yourself." + +Silence, for a minute. + +"Tom, if we'd 'a' left the blame tools at the dead tree, we'd 'a' got +the money. Oh, ain't it awful!" + +"'Tain't a dream, then, 'tain't a dream! Somehow I most wish it was. +Dog'd if I don't, Huck." + +"What ain't a dream?" + +"Oh, that thing yesterday. I been half thinking it was." + +"Dream! If them stairs hadn't broke down you'd 'a' seen how much dream +it was! I've had dreams enough all night--with that patch-eyed Spanish +devil going for me all through 'em--rot him!" + +"No, not rot him. FIND him! Track the money!" + +"Tom, we'll never find him. A feller don't have only one chance for +such a pile--and that one's lost. I'd feel mighty shaky if I was to see +him, anyway." + +"Well, so'd I; but I'd like to see him, anyway--and track him out--to +his Number Two." + +"Number Two--yes, that's it. I been thinking 'bout that. But I can't +make nothing out of it. What do you reckon it is?" + +"I dono. It's too deep. Say, Huck--maybe it's the number of a house!" + +"Goody!... No, Tom, that ain't it. If it is, it ain't in this +one-horse town. They ain't no numbers here." + +"Well, that's so. Lemme think a minute. Here--it's the number of a +room--in a tavern, you know!" + +"Oh, that's the trick! They ain't only two taverns. We can find out +quick." + +"You stay here, Huck, till I come." + +Tom was off at once. He did not care to have Huck's company in public +places. He was gone half an hour. He found that in the best tavern, No. +2 had long been occupied by a young lawyer, and was still so occupied. +In the less ostentatious house, No. 2 was a mystery. The +tavern-keeper's young son said it was kept locked all the time, and he +never saw anybody go into it or come out of it except at night; he did +not know any particular reason for this state of things; had had some +little curiosity, but it was rather feeble; had made the most of the +mystery by entertaining himself with the idea that that room was +"ha'nted"; had noticed that there was a light in there the night before. + +"That's what I've found out, Huck. I reckon that's the very No. 2 +we're after." + +"I reckon it is, Tom. Now what you going to do?" + +"Lemme think." + +Tom thought a long time. Then he said: + +"I'll tell you. The back door of that No. 2 is the door that comes out +into that little close alley between the tavern and the old rattle trap +of a brick store. Now you get hold of all the door-keys you can find, +and I'll nip all of auntie's, and the first dark night we'll go there +and try 'em. And mind you, keep a lookout for Injun Joe, because he +said he was going to drop into town and spy around once more for a +chance to get his revenge. If you see him, you just follow him; and if +he don't go to that No. 2, that ain't the place." + +"Lordy, I don't want to foller him by myself!" + +"Why, it'll be night, sure. He mightn't ever see you--and if he did, +maybe he'd never think anything." + +"Well, if it's pretty dark I reckon I'll track him. I dono--I dono. +I'll try." + +"You bet I'll follow him, if it's dark, Huck. Why, he might 'a' found +out he couldn't get his revenge, and be going right after that money." + +"It's so, Tom, it's so. I'll foller him; I will, by jingoes!" + +"Now you're TALKING! Don't you ever weaken, Huck, and I won't." + + + +CHAPTER XXVIII + +THAT night Tom and Huck were ready for their adventure. They hung +about the neighborhood of the tavern until after nine, one watching the +alley at a distance and the other the tavern door. Nobody entered the +alley or left it; nobody resembling the Spaniard entered or left the +tavern door. The night promised to be a fair one; so Tom went home with +the understanding that if a considerable degree of darkness came on, +Huck was to come and "maow," whereupon he would slip out and try the +keys. But the night remained clear, and Huck closed his watch and +retired to bed in an empty sugar hogshead about twelve. + +Tuesday the boys had the same ill luck. Also Wednesday. But Thursday +night promised better. Tom slipped out in good season with his aunt's +old tin lantern, and a large towel to blindfold it with. He hid the +lantern in Huck's sugar hogshead and the watch began. An hour before +midnight the tavern closed up and its lights (the only ones +thereabouts) were put out. No Spaniard had been seen. Nobody had +entered or left the alley. Everything was auspicious. The blackness of +darkness reigned, the perfect stillness was interrupted only by +occasional mutterings of distant thunder. + +Tom got his lantern, lit it in the hogshead, wrapped it closely in the +towel, and the two adventurers crept in the gloom toward the tavern. +Huck stood sentry and Tom felt his way into the alley. Then there was a +season of waiting anxiety that weighed upon Huck's spirits like a +mountain. He began to wish he could see a flash from the lantern--it +would frighten him, but it would at least tell him that Tom was alive +yet. It seemed hours since Tom had disappeared. Surely he must have +fainted; maybe he was dead; maybe his heart had burst under terror and +excitement. In his uneasiness Huck found himself drawing closer and +closer to the alley; fearing all sorts of dreadful things, and +momentarily expecting some catastrophe to happen that would take away +his breath. There was not much to take away, for he seemed only able to +inhale it by thimblefuls, and his heart would soon wear itself out, the +way it was beating. Suddenly there was a flash of light and Tom came +tearing by him: "Run!" said he; "run, for your life!" + +He needn't have repeated it; once was enough; Huck was making thirty +or forty miles an hour before the repetition was uttered. The boys +never stopped till they reached the shed of a deserted slaughter-house +at the lower end of the village. Just as they got within its shelter +the storm burst and the rain poured down. As soon as Tom got his breath +he said: + +"Huck, it was awful! I tried two of the keys, just as soft as I could; +but they seemed to make such a power of racket that I couldn't hardly +get my breath I was so scared. They wouldn't turn in the lock, either. +Well, without noticing what I was doing, I took hold of the knob, and +open comes the door! It warn't locked! I hopped in, and shook off the +towel, and, GREAT CAESAR'S GHOST!" + +"What!--what'd you see, Tom?" + +"Huck, I most stepped onto Injun Joe's hand!" + +"No!" + +"Yes! He was lying there, sound asleep on the floor, with his old +patch on his eye and his arms spread out." + +"Lordy, what did you do? Did he wake up?" + +"No, never budged. Drunk, I reckon. I just grabbed that towel and +started!" + +"I'd never 'a' thought of the towel, I bet!" + +"Well, I would. My aunt would make me mighty sick if I lost it." + +"Say, Tom, did you see that box?" + +"Huck, I didn't wait to look around. I didn't see the box, I didn't +see the cross. I didn't see anything but a bottle and a tin cup on the +floor by Injun Joe; yes, I saw two barrels and lots more bottles in the +room. Don't you see, now, what's the matter with that ha'nted room?" + +"How?" + +"Why, it's ha'nted with whiskey! Maybe ALL the Temperance Taverns have +got a ha'nted room, hey, Huck?" + +"Well, I reckon maybe that's so. Who'd 'a' thought such a thing? But +say, Tom, now's a mighty good time to get that box, if Injun Joe's +drunk." + +"It is, that! You try it!" + +Huck shuddered. + +"Well, no--I reckon not." + +"And I reckon not, Huck. Only one bottle alongside of Injun Joe ain't +enough. If there'd been three, he'd be drunk enough and I'd do it." + +There was a long pause for reflection, and then Tom said: + +"Lookyhere, Huck, less not try that thing any more till we know Injun +Joe's not in there. It's too scary. Now, if we watch every night, we'll +be dead sure to see him go out, some time or other, and then we'll +snatch that box quicker'n lightning." + +"Well, I'm agreed. I'll watch the whole night long, and I'll do it +every night, too, if you'll do the other part of the job." + +"All right, I will. All you got to do is to trot up Hooper Street a +block and maow--and if I'm asleep, you throw some gravel at the window +and that'll fetch me." + +"Agreed, and good as wheat!" + +"Now, Huck, the storm's over, and I'll go home. It'll begin to be +daylight in a couple of hours. You go back and watch that long, will +you?" + +"I said I would, Tom, and I will. I'll ha'nt that tavern every night +for a year! I'll sleep all day and I'll stand watch all night." + +"That's all right. Now, where you going to sleep?" + +"In Ben Rogers' hayloft. He lets me, and so does his pap's nigger man, +Uncle Jake. I tote water for Uncle Jake whenever he wants me to, and +any time I ask him he gives me a little something to eat if he can +spare it. That's a mighty good nigger, Tom. He likes me, becuz I don't +ever act as if I was above him. Sometime I've set right down and eat +WITH him. But you needn't tell that. A body's got to do things when +he's awful hungry he wouldn't want to do as a steady thing." + +"Well, if I don't want you in the daytime, I'll let you sleep. I won't +come bothering around. Any time you see something's up, in the night, +just skip right around and maow." + + + +CHAPTER XXIX + +THE first thing Tom heard on Friday morning was a glad piece of news +--Judge Thatcher's family had come back to town the night before. Both +Injun Joe and the treasure sunk into secondary importance for a moment, +and Becky took the chief place in the boy's interest. He saw her and +they had an exhausting good time playing "hi-spy" and "gully-keeper" +with a crowd of their school-mates. The day was completed and crowned +in a peculiarly satisfactory way: Becky teased her mother to appoint +the next day for the long-promised and long-delayed picnic, and she +consented. The child's delight was boundless; and Tom's not more +moderate. The invitations were sent out before sunset, and straightway +the young folks of the village were thrown into a fever of preparation +and pleasurable anticipation. Tom's excitement enabled him to keep +awake until a pretty late hour, and he had good hopes of hearing Huck's +"maow," and of having his treasure to astonish Becky and the picnickers +with, next day; but he was disappointed. No signal came that night. + +Morning came, eventually, and by ten or eleven o'clock a giddy and +rollicking company were gathered at Judge Thatcher's, and everything +was ready for a start. It was not the custom for elderly people to mar +the picnics with their presence. The children were considered safe +enough under the wings of a few young ladies of eighteen and a few +young gentlemen of twenty-three or thereabouts. The old steam ferryboat +was chartered for the occasion; presently the gay throng filed up the +main street laden with provision-baskets. Sid was sick and had to miss +the fun; Mary remained at home to entertain him. The last thing Mrs. +Thatcher said to Becky, was: + +"You'll not get back till late. Perhaps you'd better stay all night +with some of the girls that live near the ferry-landing, child." + +"Then I'll stay with Susy Harper, mamma." + +"Very well. And mind and behave yourself and don't be any trouble." + +Presently, as they tripped along, Tom said to Becky: + +"Say--I'll tell you what we'll do. 'Stead of going to Joe Harper's +we'll climb right up the hill and stop at the Widow Douglas'. She'll +have ice-cream! She has it most every day--dead loads of it. And she'll +be awful glad to have us." + +"Oh, that will be fun!" + +Then Becky reflected a moment and said: + +"But what will mamma say?" + +"How'll she ever know?" + +The girl turned the idea over in her mind, and said reluctantly: + +"I reckon it's wrong--but--" + +"But shucks! Your mother won't know, and so what's the harm? All she +wants is that you'll be safe; and I bet you she'd 'a' said go there if +she'd 'a' thought of it. I know she would!" + +The Widow Douglas' splendid hospitality was a tempting bait. It and +Tom's persuasions presently carried the day. So it was decided to say +nothing anybody about the night's programme. Presently it occurred to +Tom that maybe Huck might come this very night and give the signal. The +thought took a deal of the spirit out of his anticipations. Still he +could not bear to give up the fun at Widow Douglas'. And why should he +give it up, he reasoned--the signal did not come the night before, so +why should it be any more likely to come to-night? The sure fun of the +evening outweighed the uncertain treasure; and, boy-like, he determined +to yield to the stronger inclination and not allow himself to think of +the box of money another time that day. + +Three miles below town the ferryboat stopped at the mouth of a woody +hollow and tied up. The crowd swarmed ashore and soon the forest +distances and craggy heights echoed far and near with shoutings and +laughter. All the different ways of getting hot and tired were gone +through with, and by-and-by the rovers straggled back to camp fortified +with responsible appetites, and then the destruction of the good things +began. After the feast there was a refreshing season of rest and chat +in the shade of spreading oaks. By-and-by somebody shouted: + +"Who's ready for the cave?" + +Everybody was. Bundles of candles were procured, and straightway there +was a general scamper up the hill. The mouth of the cave was up the +hillside--an opening shaped like a letter A. Its massive oaken door +stood unbarred. Within was a small chamber, chilly as an ice-house, and +walled by Nature with solid limestone that was dewy with a cold sweat. +It was romantic and mysterious to stand here in the deep gloom and look +out upon the green valley shining in the sun. But the impressiveness of +the situation quickly wore off, and the romping began again. The moment +a candle was lighted there was a general rush upon the owner of it; a +struggle and a gallant defence followed, but the candle was soon +knocked down or blown out, and then there was a glad clamor of laughter +and a new chase. But all things have an end. By-and-by the procession +went filing down the steep descent of the main avenue, the flickering +rank of lights dimly revealing the lofty walls of rock almost to their +point of junction sixty feet overhead. This main avenue was not more +than eight or ten feet wide. Every few steps other lofty and still +narrower crevices branched from it on either hand--for McDougal's cave +was but a vast labyrinth of crooked aisles that ran into each other and +out again and led nowhere. It was said that one might wander days and +nights together through its intricate tangle of rifts and chasms, and +never find the end of the cave; and that he might go down, and down, +and still down, into the earth, and it was just the same--labyrinth +under labyrinth, and no end to any of them. No man "knew" the cave. +That was an impossible thing. Most of the young men knew a portion of +it, and it was not customary to venture much beyond this known portion. +Tom Sawyer knew as much of the cave as any one. + +The procession moved along the main avenue some three-quarters of a +mile, and then groups and couples began to slip aside into branch +avenues, fly along the dismal corridors, and take each other by +surprise at points where the corridors joined again. Parties were able +to elude each other for the space of half an hour without going beyond +the "known" ground. + +By-and-by, one group after another came straggling back to the mouth +of the cave, panting, hilarious, smeared from head to foot with tallow +drippings, daubed with clay, and entirely delighted with the success of +the day. Then they were astonished to find that they had been taking no +note of time and that night was about at hand. The clanging bell had +been calling for half an hour. However, this sort of close to the day's +adventures was romantic and therefore satisfactory. When the ferryboat +with her wild freight pushed into the stream, nobody cared sixpence for +the wasted time but the captain of the craft. + +Huck was already upon his watch when the ferryboat's lights went +glinting past the wharf. He heard no noise on board, for the young +people were as subdued and still as people usually are who are nearly +tired to death. He wondered what boat it was, and why she did not stop +at the wharf--and then he dropped her out of his mind and put his +attention upon his business. The night was growing cloudy and dark. Ten +o'clock came, and the noise of vehicles ceased, scattered lights began +to wink out, all straggling foot-passengers disappeared, the village +betook itself to its slumbers and left the small watcher alone with the +silence and the ghosts. Eleven o'clock came, and the tavern lights were +put out; darkness everywhere, now. Huck waited what seemed a weary long +time, but nothing happened. His faith was weakening. Was there any use? +Was there really any use? Why not give it up and turn in? + +A noise fell upon his ear. He was all attention in an instant. The +alley door closed softly. He sprang to the corner of the brick store. +The next moment two men brushed by him, and one seemed to have +something under his arm. It must be that box! So they were going to +remove the treasure. Why call Tom now? It would be absurd--the men +would get away with the box and never be found again. No, he would +stick to their wake and follow them; he would trust to the darkness for +security from discovery. So communing with himself, Huck stepped out +and glided along behind the men, cat-like, with bare feet, allowing +them to keep just far enough ahead not to be invisible. + +They moved up the river street three blocks, then turned to the left +up a cross-street. They went straight ahead, then, until they came to +the path that led up Cardiff Hill; this they took. They passed by the +old Welshman's house, half-way up the hill, without hesitating, and +still climbed upward. Good, thought Huck, they will bury it in the old +quarry. But they never stopped at the quarry. They passed on, up the +summit. They plunged into the narrow path between the tall sumach +bushes, and were at once hidden in the gloom. Huck closed up and +shortened his distance, now, for they would never be able to see him. +He trotted along awhile; then slackened his pace, fearing he was +gaining too fast; moved on a piece, then stopped altogether; listened; +no sound; none, save that he seemed to hear the beating of his own +heart. The hooting of an owl came over the hill--ominous sound! But no +footsteps. Heavens, was everything lost! He was about to spring with +winged feet, when a man cleared his throat not four feet from him! +Huck's heart shot into his throat, but he swallowed it again; and then +he stood there shaking as if a dozen agues had taken charge of him at +once, and so weak that he thought he must surely fall to the ground. He +knew where he was. He knew he was within five steps of the stile +leading into Widow Douglas' grounds. Very well, he thought, let them +bury it there; it won't be hard to find. + +Now there was a voice--a very low voice--Injun Joe's: + +"Damn her, maybe she's got company--there's lights, late as it is." + +"I can't see any." + +This was that stranger's voice--the stranger of the haunted house. A +deadly chill went to Huck's heart--this, then, was the "revenge" job! +His thought was, to fly. Then he remembered that the Widow Douglas had +been kind to him more than once, and maybe these men were going to +murder her. He wished he dared venture to warn her; but he knew he +didn't dare--they might come and catch him. He thought all this and +more in the moment that elapsed between the stranger's remark and Injun +Joe's next--which was-- + +"Because the bush is in your way. Now--this way--now you see, don't +you?" + +"Yes. Well, there IS company there, I reckon. Better give it up." + +"Give it up, and I just leaving this country forever! Give it up and +maybe never have another chance. I tell you again, as I've told you +before, I don't care for her swag--you may have it. But her husband was +rough on me--many times he was rough on me--and mainly he was the +justice of the peace that jugged me for a vagrant. And that ain't all. +It ain't a millionth part of it! He had me HORSEWHIPPED!--horsewhipped +in front of the jail, like a nigger!--with all the town looking on! +HORSEWHIPPED!--do you understand? He took advantage of me and died. But +I'll take it out of HER." + +"Oh, don't kill her! Don't do that!" + +"Kill? Who said anything about killing? I would kill HIM if he was +here; but not her. When you want to get revenge on a woman you don't +kill her--bosh! you go for her looks. You slit her nostrils--you notch +her ears like a sow!" + +"By God, that's--" + +"Keep your opinion to yourself! It will be safest for you. I'll tie +her to the bed. If she bleeds to death, is that my fault? I'll not cry, +if she does. My friend, you'll help me in this thing--for MY sake +--that's why you're here--I mightn't be able alone. If you flinch, I'll +kill you. Do you understand that? And if I have to kill you, I'll kill +her--and then I reckon nobody'll ever know much about who done this +business." + +"Well, if it's got to be done, let's get at it. The quicker the +better--I'm all in a shiver." + +"Do it NOW? And company there? Look here--I'll get suspicious of you, +first thing you know. No--we'll wait till the lights are out--there's +no hurry." + +Huck felt that a silence was going to ensue--a thing still more awful +than any amount of murderous talk; so he held his breath and stepped +gingerly back; planted his foot carefully and firmly, after balancing, +one-legged, in a precarious way and almost toppling over, first on one +side and then on the other. He took another step back, with the same +elaboration and the same risks; then another and another, and--a twig +snapped under his foot! His breath stopped and he listened. There was +no sound--the stillness was perfect. His gratitude was measureless. Now +he turned in his tracks, between the walls of sumach bushes--turned +himself as carefully as if he were a ship--and then stepped quickly but +cautiously along. When he emerged at the quarry he felt secure, and so +he picked up his nimble heels and flew. Down, down he sped, till he +reached the Welshman's. He banged at the door, and presently the heads +of the old man and his two stalwart sons were thrust from windows. + +"What's the row there? Who's banging? What do you want?" + +"Let me in--quick! I'll tell everything." + +"Why, who are you?" + +"Huckleberry Finn--quick, let me in!" + +"Huckleberry Finn, indeed! It ain't a name to open many doors, I +judge! But let him in, lads, and let's see what's the trouble." + +"Please don't ever tell I told you," were Huck's first words when he +got in. "Please don't--I'd be killed, sure--but the widow's been good +friends to me sometimes, and I want to tell--I WILL tell if you'll +promise you won't ever say it was me." + +"By George, he HAS got something to tell, or he wouldn't act so!" +exclaimed the old man; "out with it and nobody here'll ever tell, lad." + +Three minutes later the old man and his sons, well armed, were up the +hill, and just entering the sumach path on tiptoe, their weapons in +their hands. Huck accompanied them no further. He hid behind a great +bowlder and fell to listening. There was a lagging, anxious silence, +and then all of a sudden there was an explosion of firearms and a cry. + +Huck waited for no particulars. He sprang away and sped down the hill +as fast as his legs could carry him. + + + +CHAPTER XXX + +AS the earliest suspicion of dawn appeared on Sunday morning, Huck +came groping up the hill and rapped gently at the old Welshman's door. +The inmates were asleep, but it was a sleep that was set on a +hair-trigger, on account of the exciting episode of the night. A call +came from a window: + +"Who's there!" + +Huck's scared voice answered in a low tone: + +"Please let me in! It's only Huck Finn!" + +"It's a name that can open this door night or day, lad!--and welcome!" + +These were strange words to the vagabond boy's ears, and the +pleasantest he had ever heard. He could not recollect that the closing +word had ever been applied in his case before. The door was quickly +unlocked, and he entered. Huck was given a seat and the old man and his +brace of tall sons speedily dressed themselves. + +"Now, my boy, I hope you're good and hungry, because breakfast will be +ready as soon as the sun's up, and we'll have a piping hot one, too +--make yourself easy about that! I and the boys hoped you'd turn up and +stop here last night." + +"I was awful scared," said Huck, "and I run. I took out when the +pistols went off, and I didn't stop for three mile. I've come now becuz +I wanted to know about it, you know; and I come before daylight becuz I +didn't want to run across them devils, even if they was dead." + +"Well, poor chap, you do look as if you'd had a hard night of it--but +there's a bed here for you when you've had your breakfast. No, they +ain't dead, lad--we are sorry enough for that. You see we knew right +where to put our hands on them, by your description; so we crept along +on tiptoe till we got within fifteen feet of them--dark as a cellar +that sumach path was--and just then I found I was going to sneeze. It +was the meanest kind of luck! I tried to keep it back, but no use +--'twas bound to come, and it did come! I was in the lead with my pistol +raised, and when the sneeze started those scoundrels a-rustling to get +out of the path, I sung out, 'Fire boys!' and blazed away at the place +where the rustling was. So did the boys. But they were off in a jiffy, +those villains, and we after them, down through the woods. I judge we +never touched them. They fired a shot apiece as they started, but their +bullets whizzed by and didn't do us any harm. As soon as we lost the +sound of their feet we quit chasing, and went down and stirred up the +constables. They got a posse together, and went off to guard the river +bank, and as soon as it is light the sheriff and a gang are going to +beat up the woods. My boys will be with them presently. I wish we had +some sort of description of those rascals--'twould help a good deal. +But you couldn't see what they were like, in the dark, lad, I suppose?" + +"Oh yes; I saw them down-town and follered them." + +"Splendid! Describe them--describe them, my boy!" + +"One's the old deaf and dumb Spaniard that's ben around here once or +twice, and t'other's a mean-looking, ragged--" + +"That's enough, lad, we know the men! Happened on them in the woods +back of the widow's one day, and they slunk away. Off with you, boys, +and tell the sheriff--get your breakfast to-morrow morning!" + +The Welshman's sons departed at once. As they were leaving the room +Huck sprang up and exclaimed: + +"Oh, please don't tell ANYbody it was me that blowed on them! Oh, +please!" + +"All right if you say it, Huck, but you ought to have the credit of +what you did." + +"Oh no, no! Please don't tell!" + +When the young men were gone, the old Welshman said: + +"They won't tell--and I won't. But why don't you want it known?" + +Huck would not explain, further than to say that he already knew too +much about one of those men and would not have the man know that he +knew anything against him for the whole world--he would be killed for +knowing it, sure. + +The old man promised secrecy once more, and said: + +"How did you come to follow these fellows, lad? Were they looking +suspicious?" + +Huck was silent while he framed a duly cautious reply. Then he said: + +"Well, you see, I'm a kind of a hard lot,--least everybody says so, +and I don't see nothing agin it--and sometimes I can't sleep much, on +account of thinking about it and sort of trying to strike out a new way +of doing. That was the way of it last night. I couldn't sleep, and so I +come along up-street 'bout midnight, a-turning it all over, and when I +got to that old shackly brick store by the Temperance Tavern, I backed +up agin the wall to have another think. Well, just then along comes +these two chaps slipping along close by me, with something under their +arm, and I reckoned they'd stole it. One was a-smoking, and t'other one +wanted a light; so they stopped right before me and the cigars lit up +their faces and I see that the big one was the deaf and dumb Spaniard, +by his white whiskers and the patch on his eye, and t'other one was a +rusty, ragged-looking devil." + +"Could you see the rags by the light of the cigars?" + +This staggered Huck for a moment. Then he said: + +"Well, I don't know--but somehow it seems as if I did." + +"Then they went on, and you--" + +"Follered 'em--yes. That was it. I wanted to see what was up--they +sneaked along so. I dogged 'em to the widder's stile, and stood in the +dark and heard the ragged one beg for the widder, and the Spaniard +swear he'd spile her looks just as I told you and your two--" + +"What! The DEAF AND DUMB man said all that!" + +Huck had made another terrible mistake! He was trying his best to keep +the old man from getting the faintest hint of who the Spaniard might +be, and yet his tongue seemed determined to get him into trouble in +spite of all he could do. He made several efforts to creep out of his +scrape, but the old man's eye was upon him and he made blunder after +blunder. Presently the Welshman said: + +"My boy, don't be afraid of me. I wouldn't hurt a hair of your head +for all the world. No--I'd protect you--I'd protect you. This Spaniard +is not deaf and dumb; you've let that slip without intending it; you +can't cover that up now. You know something about that Spaniard that +you want to keep dark. Now trust me--tell me what it is, and trust me +--I won't betray you." + +Huck looked into the old man's honest eyes a moment, then bent over +and whispered in his ear: + +"'Tain't a Spaniard--it's Injun Joe!" + +The Welshman almost jumped out of his chair. In a moment he said: + +"It's all plain enough, now. When you talked about notching ears and +slitting noses I judged that that was your own embellishment, because +white men don't take that sort of revenge. But an Injun! That's a +different matter altogether." + +During breakfast the talk went on, and in the course of it the old man +said that the last thing which he and his sons had done, before going +to bed, was to get a lantern and examine the stile and its vicinity for +marks of blood. They found none, but captured a bulky bundle of-- + +"Of WHAT?" + +If the words had been lightning they could not have leaped with a more +stunning suddenness from Huck's blanched lips. His eyes were staring +wide, now, and his breath suspended--waiting for the answer. The +Welshman started--stared in return--three seconds--five seconds--ten +--then replied: + +"Of burglar's tools. Why, what's the MATTER with you?" + +Huck sank back, panting gently, but deeply, unutterably grateful. The +Welshman eyed him gravely, curiously--and presently said: + +"Yes, burglar's tools. That appears to relieve you a good deal. But +what did give you that turn? What were YOU expecting we'd found?" + +Huck was in a close place--the inquiring eye was upon him--he would +have given anything for material for a plausible answer--nothing +suggested itself--the inquiring eye was boring deeper and deeper--a +senseless reply offered--there was no time to weigh it, so at a venture +he uttered it--feebly: + +"Sunday-school books, maybe." + +Poor Huck was too distressed to smile, but the old man laughed loud +and joyously, shook up the details of his anatomy from head to foot, +and ended by saying that such a laugh was money in a-man's pocket, +because it cut down the doctor's bill like everything. Then he added: + +"Poor old chap, you're white and jaded--you ain't well a bit--no +wonder you're a little flighty and off your balance. But you'll come +out of it. Rest and sleep will fetch you out all right, I hope." + +Huck was irritated to think he had been such a goose and betrayed such +a suspicious excitement, for he had dropped the idea that the parcel +brought from the tavern was the treasure, as soon as he had heard the +talk at the widow's stile. He had only thought it was not the treasure, +however--he had not known that it wasn't--and so the suggestion of a +captured bundle was too much for his self-possession. But on the whole +he felt glad the little episode had happened, for now he knew beyond +all question that that bundle was not THE bundle, and so his mind was +at rest and exceedingly comfortable. In fact, everything seemed to be +drifting just in the right direction, now; the treasure must be still +in No. 2, the men would be captured and jailed that day, and he and Tom +could seize the gold that night without any trouble or any fear of +interruption. + +Just as breakfast was completed there was a knock at the door. Huck +jumped for a hiding-place, for he had no mind to be connected even +remotely with the late event. The Welshman admitted several ladies and +gentlemen, among them the Widow Douglas, and noticed that groups of +citizens were climbing up the hill--to stare at the stile. So the news +had spread. The Welshman had to tell the story of the night to the +visitors. The widow's gratitude for her preservation was outspoken. + +"Don't say a word about it, madam. There's another that you're more +beholden to than you are to me and my boys, maybe, but he don't allow +me to tell his name. We wouldn't have been there but for him." + +Of course this excited a curiosity so vast that it almost belittled +the main matter--but the Welshman allowed it to eat into the vitals of +his visitors, and through them be transmitted to the whole town, for he +refused to part with his secret. When all else had been learned, the +widow said: + +"I went to sleep reading in bed and slept straight through all that +noise. Why didn't you come and wake me?" + +"We judged it warn't worth while. Those fellows warn't likely to come +again--they hadn't any tools left to work with, and what was the use of +waking you up and scaring you to death? My three negro men stood guard +at your house all the rest of the night. They've just come back." + +More visitors came, and the story had to be told and retold for a +couple of hours more. + +There was no Sabbath-school during day-school vacation, but everybody +was early at church. The stirring event was well canvassed. News came +that not a sign of the two villains had been yet discovered. When the +sermon was finished, Judge Thatcher's wife dropped alongside of Mrs. +Harper as she moved down the aisle with the crowd and said: + +"Is my Becky going to sleep all day? I just expected she would be +tired to death." + +"Your Becky?" + +"Yes," with a startled look--"didn't she stay with you last night?" + +"Why, no." + +Mrs. Thatcher turned pale, and sank into a pew, just as Aunt Polly, +talking briskly with a friend, passed by. Aunt Polly said: + +"Good-morning, Mrs. Thatcher. Good-morning, Mrs. Harper. I've got a +boy that's turned up missing. I reckon my Tom stayed at your house last +night--one of you. And now he's afraid to come to church. I've got to +settle with him." + +Mrs. Thatcher shook her head feebly and turned paler than ever. + +"He didn't stay with us," said Mrs. Harper, beginning to look uneasy. +A marked anxiety came into Aunt Polly's face. + +"Joe Harper, have you seen my Tom this morning?" + +"No'm." + +"When did you see him last?" + +Joe tried to remember, but was not sure he could say. The people had +stopped moving out of church. Whispers passed along, and a boding +uneasiness took possession of every countenance. Children were +anxiously questioned, and young teachers. They all said they had not +noticed whether Tom and Becky were on board the ferryboat on the +homeward trip; it was dark; no one thought of inquiring if any one was +missing. One young man finally blurted out his fear that they were +still in the cave! Mrs. Thatcher swooned away. Aunt Polly fell to +crying and wringing her hands. + +The alarm swept from lip to lip, from group to group, from street to +street, and within five minutes the bells were wildly clanging and the +whole town was up! The Cardiff Hill episode sank into instant +insignificance, the burglars were forgotten, horses were saddled, +skiffs were manned, the ferryboat ordered out, and before the horror +was half an hour old, two hundred men were pouring down highroad and +river toward the cave. + +All the long afternoon the village seemed empty and dead. Many women +visited Aunt Polly and Mrs. Thatcher and tried to comfort them. They +cried with them, too, and that was still better than words. All the +tedious night the town waited for news; but when the morning dawned at +last, all the word that came was, "Send more candles--and send food." +Mrs. Thatcher was almost crazed; and Aunt Polly, also. Judge Thatcher +sent messages of hope and encouragement from the cave, but they +conveyed no real cheer. + +The old Welshman came home toward daylight, spattered with +candle-grease, smeared with clay, and almost worn out. He found Huck +still in the bed that had been provided for him, and delirious with +fever. The physicians were all at the cave, so the Widow Douglas came +and took charge of the patient. She said she would do her best by him, +because, whether he was good, bad, or indifferent, he was the Lord's, +and nothing that was the Lord's was a thing to be neglected. The +Welshman said Huck had good spots in him, and the widow said: + +"You can depend on it. That's the Lord's mark. He don't leave it off. +He never does. Puts it somewhere on every creature that comes from his +hands." + +Early in the forenoon parties of jaded men began to straggle into the +village, but the strongest of the citizens continued searching. All the +news that could be gained was that remotenesses of the cavern were +being ransacked that had never been visited before; that every corner +and crevice was going to be thoroughly searched; that wherever one +wandered through the maze of passages, lights were to be seen flitting +hither and thither in the distance, and shoutings and pistol-shots sent +their hollow reverberations to the ear down the sombre aisles. In one +place, far from the section usually traversed by tourists, the names +"BECKY & TOM" had been found traced upon the rocky wall with +candle-smoke, and near at hand a grease-soiled bit of ribbon. Mrs. +Thatcher recognized the ribbon and cried over it. She said it was the +last relic she should ever have of her child; and that no other memorial +of her could ever be so precious, because this one parted latest from +the living body before the awful death came. Some said that now and +then, in the cave, a far-away speck of light would glimmer, and then a +glorious shout would burst forth and a score of men go trooping down the +echoing aisle--and then a sickening disappointment always followed; the +children were not there; it was only a searcher's light. + +Three dreadful days and nights dragged their tedious hours along, and +the village sank into a hopeless stupor. No one had heart for anything. +The accidental discovery, just made, that the proprietor of the +Temperance Tavern kept liquor on his premises, scarcely fluttered the +public pulse, tremendous as the fact was. In a lucid interval, Huck +feebly led up to the subject of taverns, and finally asked--dimly +dreading the worst--if anything had been discovered at the Temperance +Tavern since he had been ill. + +"Yes," said the widow. + +Huck started up in bed, wild-eyed: + +"What? What was it?" + +"Liquor!--and the place has been shut up. Lie down, child--what a turn +you did give me!" + +"Only tell me just one thing--only just one--please! Was it Tom Sawyer +that found it?" + +The widow burst into tears. "Hush, hush, child, hush! I've told you +before, you must NOT talk. You are very, very sick!" + +Then nothing but liquor had been found; there would have been a great +powwow if it had been the gold. So the treasure was gone forever--gone +forever! But what could she be crying about? Curious that she should +cry. + +These thoughts worked their dim way through Huck's mind, and under the +weariness they gave him he fell asleep. The widow said to herself: + +"There--he's asleep, poor wreck. Tom Sawyer find it! Pity but somebody +could find Tom Sawyer! Ah, there ain't many left, now, that's got hope +enough, or strength enough, either, to go on searching." + + + +CHAPTER XXXI + +NOW to return to Tom and Becky's share in the picnic. They tripped +along the murky aisles with the rest of the company, visiting the +familiar wonders of the cave--wonders dubbed with rather +over-descriptive names, such as "The Drawing-Room," "The Cathedral," +"Aladdin's Palace," and so on. Presently the hide-and-seek frolicking +began, and Tom and Becky engaged in it with zeal until the exertion +began to grow a trifle wearisome; then they wandered down a sinuous +avenue holding their candles aloft and reading the tangled web-work of +names, dates, post-office addresses, and mottoes with which the rocky +walls had been frescoed (in candle-smoke). Still drifting along and +talking, they scarcely noticed that they were now in a part of the cave +whose walls were not frescoed. They smoked their own names under an +overhanging shelf and moved on. Presently they came to a place where a +little stream of water, trickling over a ledge and carrying a limestone +sediment with it, had, in the slow-dragging ages, formed a laced and +ruffled Niagara in gleaming and imperishable stone. Tom squeezed his +small body behind it in order to illuminate it for Becky's +gratification. He found that it curtained a sort of steep natural +stairway which was enclosed between narrow walls, and at once the +ambition to be a discoverer seized him. Becky responded to his call, +and they made a smoke-mark for future guidance, and started upon their +quest. They wound this way and that, far down into the secret depths of +the cave, made another mark, and branched off in search of novelties to +tell the upper world about. In one place they found a spacious cavern, +from whose ceiling depended a multitude of shining stalactites of the +length and circumference of a man's leg; they walked all about it, +wondering and admiring, and presently left it by one of the numerous +passages that opened into it. This shortly brought them to a bewitching +spring, whose basin was incrusted with a frostwork of glittering +crystals; it was in the midst of a cavern whose walls were supported by +many fantastic pillars which had been formed by the joining of great +stalactites and stalagmites together, the result of the ceaseless +water-drip of centuries. Under the roof vast knots of bats had packed +themselves together, thousands in a bunch; the lights disturbed the +creatures and they came flocking down by hundreds, squeaking and +darting furiously at the candles. Tom knew their ways and the danger of +this sort of conduct. He seized Becky's hand and hurried her into the +first corridor that offered; and none too soon, for a bat struck +Becky's light out with its wing while she was passing out of the +cavern. The bats chased the children a good distance; but the fugitives +plunged into every new passage that offered, and at last got rid of the +perilous things. Tom found a subterranean lake, shortly, which +stretched its dim length away until its shape was lost in the shadows. +He wanted to explore its borders, but concluded that it would be best +to sit down and rest awhile, first. Now, for the first time, the deep +stillness of the place laid a clammy hand upon the spirits of the +children. Becky said: + +"Why, I didn't notice, but it seems ever so long since I heard any of +the others." + +"Come to think, Becky, we are away down below them--and I don't know +how far away north, or south, or east, or whichever it is. We couldn't +hear them here." + +Becky grew apprehensive. + +"I wonder how long we've been down here, Tom? We better start back." + +"Yes, I reckon we better. P'raps we better." + +"Can you find the way, Tom? It's all a mixed-up crookedness to me." + +"I reckon I could find it--but then the bats. If they put our candles +out it will be an awful fix. Let's try some other way, so as not to go +through there." + +"Well. But I hope we won't get lost. It would be so awful!" and the +girl shuddered at the thought of the dreadful possibilities. + +They started through a corridor, and traversed it in silence a long +way, glancing at each new opening, to see if there was anything +familiar about the look of it; but they were all strange. Every time +Tom made an examination, Becky would watch his face for an encouraging +sign, and he would say cheerily: + +"Oh, it's all right. This ain't the one, but we'll come to it right +away!" + +But he felt less and less hopeful with each failure, and presently +began to turn off into diverging avenues at sheer random, in desperate +hope of finding the one that was wanted. He still said it was "all +right," but there was such a leaden dread at his heart that the words +had lost their ring and sounded just as if he had said, "All is lost!" +Becky clung to his side in an anguish of fear, and tried hard to keep +back the tears, but they would come. At last she said: + +"Oh, Tom, never mind the bats, let's go back that way! We seem to get +worse and worse off all the time." + +"Listen!" said he. + +Profound silence; silence so deep that even their breathings were +conspicuous in the hush. Tom shouted. The call went echoing down the +empty aisles and died out in the distance in a faint sound that +resembled a ripple of mocking laughter. + +"Oh, don't do it again, Tom, it is too horrid," said Becky. + +"It is horrid, but I better, Becky; they might hear us, you know," and +he shouted again. + +The "might" was even a chillier horror than the ghostly laughter, it +so confessed a perishing hope. The children stood still and listened; +but there was no result. Tom turned upon the back track at once, and +hurried his steps. It was but a little while before a certain +indecision in his manner revealed another fearful fact to Becky--he +could not find his way back! + +"Oh, Tom, you didn't make any marks!" + +"Becky, I was such a fool! Such a fool! I never thought we might want +to come back! No--I can't find the way. It's all mixed up." + +"Tom, Tom, we're lost! we're lost! We never can get out of this awful +place! Oh, why DID we ever leave the others!" + +She sank to the ground and burst into such a frenzy of crying that Tom +was appalled with the idea that she might die, or lose her reason. He +sat down by her and put his arms around her; she buried her face in his +bosom, she clung to him, she poured out her terrors, her unavailing +regrets, and the far echoes turned them all to jeering laughter. Tom +begged her to pluck up hope again, and she said she could not. He fell +to blaming and abusing himself for getting her into this miserable +situation; this had a better effect. She said she would try to hope +again, she would get up and follow wherever he might lead if only he +would not talk like that any more. For he was no more to blame than +she, she said. + +So they moved on again--aimlessly--simply at random--all they could do +was to move, keep moving. For a little while, hope made a show of +reviving--not with any reason to back it, but only because it is its +nature to revive when the spring has not been taken out of it by age +and familiarity with failure. + +By-and-by Tom took Becky's candle and blew it out. This economy meant +so much! Words were not needed. Becky understood, and her hope died +again. She knew that Tom had a whole candle and three or four pieces in +his pockets--yet he must economize. + +By-and-by, fatigue began to assert its claims; the children tried to +pay attention, for it was dreadful to think of sitting down when time +was grown to be so precious, moving, in some direction, in any +direction, was at least progress and might bear fruit; but to sit down +was to invite death and shorten its pursuit. + +At last Becky's frail limbs refused to carry her farther. She sat +down. Tom rested with her, and they talked of home, and the friends +there, and the comfortable beds and, above all, the light! Becky cried, +and Tom tried to think of some way of comforting her, but all his +encouragements were grown threadbare with use, and sounded like +sarcasms. Fatigue bore so heavily upon Becky that she drowsed off to +sleep. Tom was grateful. He sat looking into her drawn face and saw it +grow smooth and natural under the influence of pleasant dreams; and +by-and-by a smile dawned and rested there. The peaceful face reflected +somewhat of peace and healing into his own spirit, and his thoughts +wandered away to bygone times and dreamy memories. While he was deep in +his musings, Becky woke up with a breezy little laugh--but it was +stricken dead upon her lips, and a groan followed it. + +"Oh, how COULD I sleep! I wish I never, never had waked! No! No, I +don't, Tom! Don't look so! I won't say it again." + +"I'm glad you've slept, Becky; you'll feel rested, now, and we'll find +the way out." + +"We can try, Tom; but I've seen such a beautiful country in my dream. +I reckon we are going there." + +"Maybe not, maybe not. Cheer up, Becky, and let's go on trying." + +They rose up and wandered along, hand in hand and hopeless. They tried +to estimate how long they had been in the cave, but all they knew was +that it seemed days and weeks, and yet it was plain that this could not +be, for their candles were not gone yet. A long time after this--they +could not tell how long--Tom said they must go softly and listen for +dripping water--they must find a spring. They found one presently, and +Tom said it was time to rest again. Both were cruelly tired, yet Becky +said she thought she could go a little farther. She was surprised to +hear Tom dissent. She could not understand it. They sat down, and Tom +fastened his candle to the wall in front of them with some clay. +Thought was soon busy; nothing was said for some time. Then Becky broke +the silence: + +"Tom, I am so hungry!" + +Tom took something out of his pocket. + +"Do you remember this?" said he. + +Becky almost smiled. + +"It's our wedding-cake, Tom." + +"Yes--I wish it was as big as a barrel, for it's all we've got." + +"I saved it from the picnic for us to dream on, Tom, the way grown-up +people do with wedding-cake--but it'll be our--" + +She dropped the sentence where it was. Tom divided the cake and Becky +ate with good appetite, while Tom nibbled at his moiety. There was +abundance of cold water to finish the feast with. By-and-by Becky +suggested that they move on again. Tom was silent a moment. Then he +said: + +"Becky, can you bear it if I tell you something?" + +Becky's face paled, but she thought she could. + +"Well, then, Becky, we must stay here, where there's water to drink. +That little piece is our last candle!" + +Becky gave loose to tears and wailings. Tom did what he could to +comfort her, but with little effect. At length Becky said: + +"Tom!" + +"Well, Becky?" + +"They'll miss us and hunt for us!" + +"Yes, they will! Certainly they will!" + +"Maybe they're hunting for us now, Tom." + +"Why, I reckon maybe they are. I hope they are." + +"When would they miss us, Tom?" + +"When they get back to the boat, I reckon." + +"Tom, it might be dark then--would they notice we hadn't come?" + +"I don't know. But anyway, your mother would miss you as soon as they +got home." + +A frightened look in Becky's face brought Tom to his senses and he saw +that he had made a blunder. Becky was not to have gone home that night! +The children became silent and thoughtful. In a moment a new burst of +grief from Becky showed Tom that the thing in his mind had struck hers +also--that the Sabbath morning might be half spent before Mrs. Thatcher +discovered that Becky was not at Mrs. Harper's. + +The children fastened their eyes upon their bit of candle and watched +it melt slowly and pitilessly away; saw the half inch of wick stand +alone at last; saw the feeble flame rise and fall, climb the thin +column of smoke, linger at its top a moment, and then--the horror of +utter darkness reigned! + +How long afterward it was that Becky came to a slow consciousness that +she was crying in Tom's arms, neither could tell. All that they knew +was, that after what seemed a mighty stretch of time, both awoke out of +a dead stupor of sleep and resumed their miseries once more. Tom said +it might be Sunday, now--maybe Monday. He tried to get Becky to talk, +but her sorrows were too oppressive, all her hopes were gone. Tom said +that they must have been missed long ago, and no doubt the search was +going on. He would shout and maybe some one would come. He tried it; +but in the darkness the distant echoes sounded so hideously that he +tried it no more. + +The hours wasted away, and hunger came to torment the captives again. +A portion of Tom's half of the cake was left; they divided and ate it. +But they seemed hungrier than before. The poor morsel of food only +whetted desire. + +By-and-by Tom said: + +"SH! Did you hear that?" + +Both held their breath and listened. There was a sound like the +faintest, far-off shout. Instantly Tom answered it, and leading Becky +by the hand, started groping down the corridor in its direction. +Presently he listened again; again the sound was heard, and apparently +a little nearer. + +"It's them!" said Tom; "they're coming! Come along, Becky--we're all +right now!" + +The joy of the prisoners was almost overwhelming. Their speed was +slow, however, because pitfalls were somewhat common, and had to be +guarded against. They shortly came to one and had to stop. It might be +three feet deep, it might be a hundred--there was no passing it at any +rate. Tom got down on his breast and reached as far down as he could. +No bottom. They must stay there and wait until the searchers came. They +listened; evidently the distant shoutings were growing more distant! a +moment or two more and they had gone altogether. The heart-sinking +misery of it! Tom whooped until he was hoarse, but it was of no use. He +talked hopefully to Becky; but an age of anxious waiting passed and no +sounds came again. + +The children groped their way back to the spring. The weary time +dragged on; they slept again, and awoke famished and woe-stricken. Tom +believed it must be Tuesday by this time. + +Now an idea struck him. There were some side passages near at hand. It +would be better to explore some of these than bear the weight of the +heavy time in idleness. He took a kite-line from his pocket, tied it to +a projection, and he and Becky started, Tom in the lead, unwinding the +line as he groped along. At the end of twenty steps the corridor ended +in a "jumping-off place." Tom got down on his knees and felt below, and +then as far around the corner as he could reach with his hands +conveniently; he made an effort to stretch yet a little farther to the +right, and at that moment, not twenty yards away, a human hand, holding +a candle, appeared from behind a rock! Tom lifted up a glorious shout, +and instantly that hand was followed by the body it belonged to--Injun +Joe's! Tom was paralyzed; he could not move. He was vastly gratified +the next moment, to see the "Spaniard" take to his heels and get +himself out of sight. Tom wondered that Joe had not recognized his +voice and come over and killed him for testifying in court. But the +echoes must have disguised the voice. Without doubt, that was it, he +reasoned. Tom's fright weakened every muscle in his body. He said to +himself that if he had strength enough to get back to the spring he +would stay there, and nothing should tempt him to run the risk of +meeting Injun Joe again. He was careful to keep from Becky what it was +he had seen. He told her he had only shouted "for luck." + +But hunger and wretchedness rise superior to fears in the long run. +Another tedious wait at the spring and another long sleep brought +changes. The children awoke tortured with a raging hunger. Tom believed +that it must be Wednesday or Thursday or even Friday or Saturday, now, +and that the search had been given over. He proposed to explore another +passage. He felt willing to risk Injun Joe and all other terrors. But +Becky was very weak. She had sunk into a dreary apathy and would not be +roused. She said she would wait, now, where she was, and die--it would +not be long. She told Tom to go with the kite-line and explore if he +chose; but she implored him to come back every little while and speak +to her; and she made him promise that when the awful time came, he +would stay by her and hold her hand until all was over. + +Tom kissed her, with a choking sensation in his throat, and made a +show of being confident of finding the searchers or an escape from the +cave; then he took the kite-line in his hand and went groping down one +of the passages on his hands and knees, distressed with hunger and sick +with bodings of coming doom. + + + +CHAPTER XXXII + +TUESDAY afternoon came, and waned to the twilight. The village of St. +Petersburg still mourned. The lost children had not been found. Public +prayers had been offered up for them, and many and many a private +prayer that had the petitioner's whole heart in it; but still no good +news came from the cave. The majority of the searchers had given up the +quest and gone back to their daily avocations, saying that it was plain +the children could never be found. Mrs. Thatcher was very ill, and a +great part of the time delirious. People said it was heartbreaking to +hear her call her child, and raise her head and listen a whole minute +at a time, then lay it wearily down again with a moan. Aunt Polly had +drooped into a settled melancholy, and her gray hair had grown almost +white. The village went to its rest on Tuesday night, sad and forlorn. + +Away in the middle of the night a wild peal burst from the village +bells, and in a moment the streets were swarming with frantic half-clad +people, who shouted, "Turn out! turn out! they're found! they're +found!" Tin pans and horns were added to the din, the population massed +itself and moved toward the river, met the children coming in an open +carriage drawn by shouting citizens, thronged around it, joined its +homeward march, and swept magnificently up the main street roaring +huzzah after huzzah! + +The village was illuminated; nobody went to bed again; it was the +greatest night the little town had ever seen. During the first half-hour +a procession of villagers filed through Judge Thatcher's house, seized +the saved ones and kissed them, squeezed Mrs. Thatcher's hand, tried to +speak but couldn't--and drifted out raining tears all over the place. + +Aunt Polly's happiness was complete, and Mrs. Thatcher's nearly so. It +would be complete, however, as soon as the messenger dispatched with +the great news to the cave should get the word to her husband. Tom lay +upon a sofa with an eager auditory about him and told the history of +the wonderful adventure, putting in many striking additions to adorn it +withal; and closed with a description of how he left Becky and went on +an exploring expedition; how he followed two avenues as far as his +kite-line would reach; how he followed a third to the fullest stretch of +the kite-line, and was about to turn back when he glimpsed a far-off +speck that looked like daylight; dropped the line and groped toward it, +pushed his head and shoulders through a small hole, and saw the broad +Mississippi rolling by! And if it had only happened to be night he would +not have seen that speck of daylight and would not have explored that +passage any more! He told how he went back for Becky and broke the good +news and she told him not to fret her with such stuff, for she was +tired, and knew she was going to die, and wanted to. He described how he +labored with her and convinced her; and how she almost died for joy when +she had groped to where she actually saw the blue speck of daylight; how +he pushed his way out at the hole and then helped her out; how they sat +there and cried for gladness; how some men came along in a skiff and Tom +hailed them and told them their situation and their famished condition; +how the men didn't believe the wild tale at first, "because," said they, +"you are five miles down the river below the valley the cave is in" +--then took them aboard, rowed to a house, gave them supper, made them +rest till two or three hours after dark and then brought them home. + +Before day-dawn, Judge Thatcher and the handful of searchers with him +were tracked out, in the cave, by the twine clews they had strung +behind them, and informed of the great news. + +Three days and nights of toil and hunger in the cave were not to be +shaken off at once, as Tom and Becky soon discovered. They were +bedridden all of Wednesday and Thursday, and seemed to grow more and +more tired and worn, all the time. Tom got about, a little, on +Thursday, was down-town Friday, and nearly as whole as ever Saturday; +but Becky did not leave her room until Sunday, and then she looked as +if she had passed through a wasting illness. + +Tom learned of Huck's sickness and went to see him on Friday, but +could not be admitted to the bedroom; neither could he on Saturday or +Sunday. He was admitted daily after that, but was warned to keep still +about his adventure and introduce no exciting topic. The Widow Douglas +stayed by to see that he obeyed. At home Tom learned of the Cardiff +Hill event; also that the "ragged man's" body had eventually been found +in the river near the ferry-landing; he had been drowned while trying +to escape, perhaps. + +About a fortnight after Tom's rescue from the cave, he started off to +visit Huck, who had grown plenty strong enough, now, to hear exciting +talk, and Tom had some that would interest him, he thought. Judge +Thatcher's house was on Tom's way, and he stopped to see Becky. The +Judge and some friends set Tom to talking, and some one asked him +ironically if he wouldn't like to go to the cave again. Tom said he +thought he wouldn't mind it. The Judge said: + +"Well, there are others just like you, Tom, I've not the least doubt. +But we have taken care of that. Nobody will get lost in that cave any +more." + +"Why?" + +"Because I had its big door sheathed with boiler iron two weeks ago, +and triple-locked--and I've got the keys." + +Tom turned as white as a sheet. + +"What's the matter, boy! Here, run, somebody! Fetch a glass of water!" + +The water was brought and thrown into Tom's face. + +"Ah, now you're all right. What was the matter with you, Tom?" + +"Oh, Judge, Injun Joe's in the cave!" + + + +CHAPTER XXXIII + +WITHIN a few minutes the news had spread, and a dozen skiff-loads of +men were on their way to McDougal's cave, and the ferryboat, well +filled with passengers, soon followed. Tom Sawyer was in the skiff that +bore Judge Thatcher. + +When the cave door was unlocked, a sorrowful sight presented itself in +the dim twilight of the place. Injun Joe lay stretched upon the ground, +dead, with his face close to the crack of the door, as if his longing +eyes had been fixed, to the latest moment, upon the light and the cheer +of the free world outside. Tom was touched, for he knew by his own +experience how this wretch had suffered. His pity was moved, but +nevertheless he felt an abounding sense of relief and security, now, +which revealed to him in a degree which he had not fully appreciated +before how vast a weight of dread had been lying upon him since the day +he lifted his voice against this bloody-minded outcast. + +Injun Joe's bowie-knife lay close by, its blade broken in two. The +great foundation-beam of the door had been chipped and hacked through, +with tedious labor; useless labor, too, it was, for the native rock +formed a sill outside it, and upon that stubborn material the knife had +wrought no effect; the only damage done was to the knife itself. But if +there had been no stony obstruction there the labor would have been +useless still, for if the beam had been wholly cut away Injun Joe could +not have squeezed his body under the door, and he knew it. So he had +only hacked that place in order to be doing something--in order to pass +the weary time--in order to employ his tortured faculties. Ordinarily +one could find half a dozen bits of candle stuck around in the crevices +of this vestibule, left there by tourists; but there were none now. The +prisoner had searched them out and eaten them. He had also contrived to +catch a few bats, and these, also, he had eaten, leaving only their +claws. The poor unfortunate had starved to death. In one place, near at +hand, a stalagmite had been slowly growing up from the ground for ages, +builded by the water-drip from a stalactite overhead. The captive had +broken off the stalagmite, and upon the stump had placed a stone, +wherein he had scooped a shallow hollow to catch the precious drop +that fell once in every three minutes with the dreary regularity of a +clock-tick--a dessertspoonful once in four and twenty hours. That drop +was falling when the Pyramids were new; when Troy fell; when the +foundations of Rome were laid; when Christ was crucified; when the +Conqueror created the British empire; when Columbus sailed; when the +massacre at Lexington was "news." It is falling now; it will still be +falling when all these things shall have sunk down the afternoon of +history, and the twilight of tradition, and been swallowed up in the +thick night of oblivion. Has everything a purpose and a mission? Did +this drop fall patiently during five thousand years to be ready for +this flitting human insect's need? and has it another important object +to accomplish ten thousand years to come? No matter. It is many and +many a year since the hapless half-breed scooped out the stone to catch +the priceless drops, but to this day the tourist stares longest at that +pathetic stone and that slow-dropping water when he comes to see the +wonders of McDougal's cave. Injun Joe's cup stands first in the list of +the cavern's marvels; even "Aladdin's Palace" cannot rival it. + +Injun Joe was buried near the mouth of the cave; and people flocked +there in boats and wagons from the towns and from all the farms and +hamlets for seven miles around; they brought their children, and all +sorts of provisions, and confessed that they had had almost as +satisfactory a time at the funeral as they could have had at the +hanging. + +This funeral stopped the further growth of one thing--the petition to +the governor for Injun Joe's pardon. The petition had been largely +signed; many tearful and eloquent meetings had been held, and a +committee of sappy women been appointed to go in deep mourning and wail +around the governor, and implore him to be a merciful ass and trample +his duty under foot. Injun Joe was believed to have killed five +citizens of the village, but what of that? If he had been Satan himself +there would have been plenty of weaklings ready to scribble their names +to a pardon-petition, and drip a tear on it from their permanently +impaired and leaky water-works. + +The morning after the funeral Tom took Huck to a private place to have +an important talk. Huck had learned all about Tom's adventure from the +Welshman and the Widow Douglas, by this time, but Tom said he reckoned +there was one thing they had not told him; that thing was what he +wanted to talk about now. Huck's face saddened. He said: + +"I know what it is. You got into No. 2 and never found anything but +whiskey. Nobody told me it was you; but I just knowed it must 'a' ben +you, soon as I heard 'bout that whiskey business; and I knowed you +hadn't got the money becuz you'd 'a' got at me some way or other and +told me even if you was mum to everybody else. Tom, something's always +told me we'd never get holt of that swag." + +"Why, Huck, I never told on that tavern-keeper. YOU know his tavern +was all right the Saturday I went to the picnic. Don't you remember you +was to watch there that night?" + +"Oh yes! Why, it seems 'bout a year ago. It was that very night that I +follered Injun Joe to the widder's." + +"YOU followed him?" + +"Yes--but you keep mum. I reckon Injun Joe's left friends behind him, +and I don't want 'em souring on me and doing me mean tricks. If it +hadn't ben for me he'd be down in Texas now, all right." + +Then Huck told his entire adventure in confidence to Tom, who had only +heard of the Welshman's part of it before. + +"Well," said Huck, presently, coming back to the main question, +"whoever nipped the whiskey in No. 2, nipped the money, too, I reckon +--anyways it's a goner for us, Tom." + +"Huck, that money wasn't ever in No. 2!" + +"What!" Huck searched his comrade's face keenly. "Tom, have you got on +the track of that money again?" + +"Huck, it's in the cave!" + +Huck's eyes blazed. + +"Say it again, Tom." + +"The money's in the cave!" + +"Tom--honest injun, now--is it fun, or earnest?" + +"Earnest, Huck--just as earnest as ever I was in my life. Will you go +in there with me and help get it out?" + +"I bet I will! I will if it's where we can blaze our way to it and not +get lost." + +"Huck, we can do that without the least little bit of trouble in the +world." + +"Good as wheat! What makes you think the money's--" + +"Huck, you just wait till we get in there. If we don't find it I'll +agree to give you my drum and every thing I've got in the world. I +will, by jings." + +"All right--it's a whiz. When do you say?" + +"Right now, if you say it. Are you strong enough?" + +"Is it far in the cave? I ben on my pins a little, three or four days, +now, but I can't walk more'n a mile, Tom--least I don't think I could." + +"It's about five mile into there the way anybody but me would go, +Huck, but there's a mighty short cut that they don't anybody but me +know about. Huck, I'll take you right to it in a skiff. I'll float the +skiff down there, and I'll pull it back again all by myself. You +needn't ever turn your hand over." + +"Less start right off, Tom." + +"All right. We want some bread and meat, and our pipes, and a little +bag or two, and two or three kite-strings, and some of these +new-fangled things they call lucifer matches. I tell you, many's +the time I wished I had some when I was in there before." + +A trifle after noon the boys borrowed a small skiff from a citizen who +was absent, and got under way at once. When they were several miles +below "Cave Hollow," Tom said: + +"Now you see this bluff here looks all alike all the way down from the +cave hollow--no houses, no wood-yards, bushes all alike. But do you see +that white place up yonder where there's been a landslide? Well, that's +one of my marks. We'll get ashore, now." + +They landed. + +"Now, Huck, where we're a-standing you could touch that hole I got out +of with a fishing-pole. See if you can find it." + +Huck searched all the place about, and found nothing. Tom proudly +marched into a thick clump of sumach bushes and said: + +"Here you are! Look at it, Huck; it's the snuggest hole in this +country. You just keep mum about it. All along I've been wanting to be +a robber, but I knew I'd got to have a thing like this, and where to +run across it was the bother. We've got it now, and we'll keep it +quiet, only we'll let Joe Harper and Ben Rogers in--because of course +there's got to be a Gang, or else there wouldn't be any style about it. +Tom Sawyer's Gang--it sounds splendid, don't it, Huck?" + +"Well, it just does, Tom. And who'll we rob?" + +"Oh, most anybody. Waylay people--that's mostly the way." + +"And kill them?" + +"No, not always. Hive them in the cave till they raise a ransom." + +"What's a ransom?" + +"Money. You make them raise all they can, off'n their friends; and +after you've kept them a year, if it ain't raised then you kill them. +That's the general way. Only you don't kill the women. You shut up the +women, but you don't kill them. They're always beautiful and rich, and +awfully scared. You take their watches and things, but you always take +your hat off and talk polite. They ain't anybody as polite as robbers +--you'll see that in any book. Well, the women get to loving you, and +after they've been in the cave a week or two weeks they stop crying and +after that you couldn't get them to leave. If you drove them out they'd +turn right around and come back. It's so in all the books." + +"Why, it's real bully, Tom. I believe it's better'n to be a pirate." + +"Yes, it's better in some ways, because it's close to home and +circuses and all that." + +By this time everything was ready and the boys entered the hole, Tom +in the lead. They toiled their way to the farther end of the tunnel, +then made their spliced kite-strings fast and moved on. A few steps +brought them to the spring, and Tom felt a shudder quiver all through +him. He showed Huck the fragment of candle-wick perched on a lump of +clay against the wall, and described how he and Becky had watched the +flame struggle and expire. + +The boys began to quiet down to whispers, now, for the stillness and +gloom of the place oppressed their spirits. They went on, and presently +entered and followed Tom's other corridor until they reached the +"jumping-off place." The candles revealed the fact that it was not +really a precipice, but only a steep clay hill twenty or thirty feet +high. Tom whispered: + +"Now I'll show you something, Huck." + +He held his candle aloft and said: + +"Look as far around the corner as you can. Do you see that? There--on +the big rock over yonder--done with candle-smoke." + +"Tom, it's a CROSS!" + +"NOW where's your Number Two? 'UNDER THE CROSS,' hey? Right yonder's +where I saw Injun Joe poke up his candle, Huck!" + +Huck stared at the mystic sign awhile, and then said with a shaky voice: + +"Tom, less git out of here!" + +"What! and leave the treasure?" + +"Yes--leave it. Injun Joe's ghost is round about there, certain." + +"No it ain't, Huck, no it ain't. It would ha'nt the place where he +died--away out at the mouth of the cave--five mile from here." + +"No, Tom, it wouldn't. It would hang round the money. I know the ways +of ghosts, and so do you." + +Tom began to fear that Huck was right. Misgivings gathered in his +mind. But presently an idea occurred to him-- + +"Lookyhere, Huck, what fools we're making of ourselves! Injun Joe's +ghost ain't a going to come around where there's a cross!" + +The point was well taken. It had its effect. + +"Tom, I didn't think of that. But that's so. It's luck for us, that +cross is. I reckon we'll climb down there and have a hunt for that box." + +Tom went first, cutting rude steps in the clay hill as he descended. +Huck followed. Four avenues opened out of the small cavern which the +great rock stood in. The boys examined three of them with no result. +They found a small recess in the one nearest the base of the rock, with +a pallet of blankets spread down in it; also an old suspender, some +bacon rind, and the well-gnawed bones of two or three fowls. But there +was no money-box. The lads searched and researched this place, but in +vain. Tom said: + +"He said UNDER the cross. Well, this comes nearest to being under the +cross. It can't be under the rock itself, because that sets solid on +the ground." + +They searched everywhere once more, and then sat down discouraged. +Huck could suggest nothing. By-and-by Tom said: + +"Lookyhere, Huck, there's footprints and some candle-grease on the +clay about one side of this rock, but not on the other sides. Now, +what's that for? I bet you the money IS under the rock. I'm going to +dig in the clay." + +"That ain't no bad notion, Tom!" said Huck with animation. + +Tom's "real Barlow" was out at once, and he had not dug four inches +before he struck wood. + +"Hey, Huck!--you hear that?" + +Huck began to dig and scratch now. Some boards were soon uncovered and +removed. They had concealed a natural chasm which led under the rock. +Tom got into this and held his candle as far under the rock as he +could, but said he could not see to the end of the rift. He proposed to +explore. He stooped and passed under; the narrow way descended +gradually. He followed its winding course, first to the right, then to +the left, Huck at his heels. Tom turned a short curve, by-and-by, and +exclaimed: + +"My goodness, Huck, lookyhere!" + +It was the treasure-box, sure enough, occupying a snug little cavern, +along with an empty powder-keg, a couple of guns in leather cases, two +or three pairs of old moccasins, a leather belt, and some other rubbish +well soaked with the water-drip. + +"Got it at last!" said Huck, ploughing among the tarnished coins with +his hand. "My, but we're rich, Tom!" + +"Huck, I always reckoned we'd get it. It's just too good to believe, +but we HAVE got it, sure! Say--let's not fool around here. Let's snake +it out. Lemme see if I can lift the box." + +It weighed about fifty pounds. Tom could lift it, after an awkward +fashion, but could not carry it conveniently. + +"I thought so," he said; "THEY carried it like it was heavy, that day +at the ha'nted house. I noticed that. I reckon I was right to think of +fetching the little bags along." + +The money was soon in the bags and the boys took it up to the cross +rock. + +"Now less fetch the guns and things," said Huck. + +"No, Huck--leave them there. They're just the tricks to have when we +go to robbing. We'll keep them there all the time, and we'll hold our +orgies there, too. It's an awful snug place for orgies." + +"What orgies?" + +"I dono. But robbers always have orgies, and of course we've got to +have them, too. Come along, Huck, we've been in here a long time. It's +getting late, I reckon. I'm hungry, too. We'll eat and smoke when we +get to the skiff." + +They presently emerged into the clump of sumach bushes, looked warily +out, found the coast clear, and were soon lunching and smoking in the +skiff. As the sun dipped toward the horizon they pushed out and got +under way. Tom skimmed up the shore through the long twilight, chatting +cheerily with Huck, and landed shortly after dark. + +"Now, Huck," said Tom, "we'll hide the money in the loft of the +widow's woodshed, and I'll come up in the morning and we'll count it +and divide, and then we'll hunt up a place out in the woods for it +where it will be safe. Just you lay quiet here and watch the stuff till +I run and hook Benny Taylor's little wagon; I won't be gone a minute." + +He disappeared, and presently returned with the wagon, put the two +small sacks into it, threw some old rags on top of them, and started +off, dragging his cargo behind him. When the boys reached the +Welshman's house, they stopped to rest. Just as they were about to move +on, the Welshman stepped out and said: + +"Hallo, who's that?" + +"Huck and Tom Sawyer." + +"Good! Come along with me, boys, you are keeping everybody waiting. +Here--hurry up, trot ahead--I'll haul the wagon for you. Why, it's not +as light as it might be. Got bricks in it?--or old metal?" + +"Old metal," said Tom. + +"I judged so; the boys in this town will take more trouble and fool +away more time hunting up six bits' worth of old iron to sell to the +foundry than they would to make twice the money at regular work. But +that's human nature--hurry along, hurry along!" + +The boys wanted to know what the hurry was about. + +"Never mind; you'll see, when we get to the Widow Douglas'." + +Huck said with some apprehension--for he was long used to being +falsely accused: + +"Mr. Jones, we haven't been doing nothing." + +The Welshman laughed. + +"Well, I don't know, Huck, my boy. I don't know about that. Ain't you +and the widow good friends?" + +"Yes. Well, she's ben good friends to me, anyway." + +"All right, then. What do you want to be afraid for?" + +This question was not entirely answered in Huck's slow mind before he +found himself pushed, along with Tom, into Mrs. Douglas' drawing-room. +Mr. Jones left the wagon near the door and followed. + +The place was grandly lighted, and everybody that was of any +consequence in the village was there. The Thatchers were there, the +Harpers, the Rogerses, Aunt Polly, Sid, Mary, the minister, the editor, +and a great many more, and all dressed in their best. The widow +received the boys as heartily as any one could well receive two such +looking beings. They were covered with clay and candle-grease. Aunt +Polly blushed crimson with humiliation, and frowned and shook her head +at Tom. Nobody suffered half as much as the two boys did, however. Mr. +Jones said: + +"Tom wasn't at home, yet, so I gave him up; but I stumbled on him and +Huck right at my door, and so I just brought them along in a hurry." + +"And you did just right," said the widow. "Come with me, boys." + +She took them to a bedchamber and said: + +"Now wash and dress yourselves. Here are two new suits of clothes +--shirts, socks, everything complete. They're Huck's--no, no thanks, +Huck--Mr. Jones bought one and I the other. But they'll fit both of you. +Get into them. We'll wait--come down when you are slicked up enough." + +Then she left. + + + +CHAPTER XXXIV + +HUCK said: "Tom, we can slope, if we can find a rope. The window ain't +high from the ground." + +"Shucks! what do you want to slope for?" + +"Well, I ain't used to that kind of a crowd. I can't stand it. I ain't +going down there, Tom." + +"Oh, bother! It ain't anything. I don't mind it a bit. I'll take care +of you." + +Sid appeared. + +"Tom," said he, "auntie has been waiting for you all the afternoon. +Mary got your Sunday clothes ready, and everybody's been fretting about +you. Say--ain't this grease and clay, on your clothes?" + +"Now, Mr. Siddy, you jist 'tend to your own business. What's all this +blow-out about, anyway?" + +"It's one of the widow's parties that she's always having. This time +it's for the Welshman and his sons, on account of that scrape they +helped her out of the other night. And say--I can tell you something, +if you want to know." + +"Well, what?" + +"Why, old Mr. Jones is going to try to spring something on the people +here to-night, but I overheard him tell auntie to-day about it, as a +secret, but I reckon it's not much of a secret now. Everybody knows +--the widow, too, for all she tries to let on she don't. Mr. Jones was +bound Huck should be here--couldn't get along with his grand secret +without Huck, you know!" + +"Secret about what, Sid?" + +"About Huck tracking the robbers to the widow's. I reckon Mr. Jones +was going to make a grand time over his surprise, but I bet you it will +drop pretty flat." + +Sid chuckled in a very contented and satisfied way. + +"Sid, was it you that told?" + +"Oh, never mind who it was. SOMEBODY told--that's enough." + +"Sid, there's only one person in this town mean enough to do that, and +that's you. If you had been in Huck's place you'd 'a' sneaked down the +hill and never told anybody on the robbers. You can't do any but mean +things, and you can't bear to see anybody praised for doing good ones. +There--no thanks, as the widow says"--and Tom cuffed Sid's ears and +helped him to the door with several kicks. "Now go and tell auntie if +you dare--and to-morrow you'll catch it!" + +Some minutes later the widow's guests were at the supper-table, and a +dozen children were propped up at little side-tables in the same room, +after the fashion of that country and that day. At the proper time Mr. +Jones made his little speech, in which he thanked the widow for the +honor she was doing himself and his sons, but said that there was +another person whose modesty-- + +And so forth and so on. He sprung his secret about Huck's share in the +adventure in the finest dramatic manner he was master of, but the +surprise it occasioned was largely counterfeit and not as clamorous and +effusive as it might have been under happier circumstances. However, +the widow made a pretty fair show of astonishment, and heaped so many +compliments and so much gratitude upon Huck that he almost forgot the +nearly intolerable discomfort of his new clothes in the entirely +intolerable discomfort of being set up as a target for everybody's gaze +and everybody's laudations. + +The widow said she meant to give Huck a home under her roof and have +him educated; and that when she could spare the money she would start +him in business in a modest way. Tom's chance was come. He said: + +"Huck don't need it. Huck's rich." + +Nothing but a heavy strain upon the good manners of the company kept +back the due and proper complimentary laugh at this pleasant joke. But +the silence was a little awkward. Tom broke it: + +"Huck's got money. Maybe you don't believe it, but he's got lots of +it. Oh, you needn't smile--I reckon I can show you. You just wait a +minute." + +Tom ran out of doors. The company looked at each other with a +perplexed interest--and inquiringly at Huck, who was tongue-tied. + +"Sid, what ails Tom?" said Aunt Polly. "He--well, there ain't ever any +making of that boy out. I never--" + +Tom entered, struggling with the weight of his sacks, and Aunt Polly +did not finish her sentence. Tom poured the mass of yellow coin upon +the table and said: + +"There--what did I tell you? Half of it's Huck's and half of it's mine!" + +The spectacle took the general breath away. All gazed, nobody spoke +for a moment. Then there was a unanimous call for an explanation. Tom +said he could furnish it, and he did. The tale was long, but brimful of +interest. There was scarcely an interruption from any one to break the +charm of its flow. When he had finished, Mr. Jones said: + +"I thought I had fixed up a little surprise for this occasion, but it +don't amount to anything now. This one makes it sing mighty small, I'm +willing to allow." + +The money was counted. The sum amounted to a little over twelve +thousand dollars. It was more than any one present had ever seen at one +time before, though several persons were there who were worth +considerably more than that in property. + + + +CHAPTER XXXV + +THE reader may rest satisfied that Tom's and Huck's windfall made a +mighty stir in the poor little village of St. Petersburg. So vast a +sum, all in actual cash, seemed next to incredible. It was talked +about, gloated over, glorified, until the reason of many of the +citizens tottered under the strain of the unhealthy excitement. Every +"haunted" house in St. Petersburg and the neighboring villages was +dissected, plank by plank, and its foundations dug up and ransacked for +hidden treasure--and not by boys, but men--pretty grave, unromantic +men, too, some of them. Wherever Tom and Huck appeared they were +courted, admired, stared at. The boys were not able to remember that +their remarks had possessed weight before; but now their sayings were +treasured and repeated; everything they did seemed somehow to be +regarded as remarkable; they had evidently lost the power of doing and +saying commonplace things; moreover, their past history was raked up +and discovered to bear marks of conspicuous originality. The village +paper published biographical sketches of the boys. + +The Widow Douglas put Huck's money out at six per cent., and Judge +Thatcher did the same with Tom's at Aunt Polly's request. Each lad had +an income, now, that was simply prodigious--a dollar for every week-day +in the year and half of the Sundays. It was just what the minister got +--no, it was what he was promised--he generally couldn't collect it. A +dollar and a quarter a week would board, lodge, and school a boy in +those old simple days--and clothe him and wash him, too, for that +matter. + +Judge Thatcher had conceived a great opinion of Tom. He said that no +commonplace boy would ever have got his daughter out of the cave. When +Becky told her father, in strict confidence, how Tom had taken her +whipping at school, the Judge was visibly moved; and when she pleaded +grace for the mighty lie which Tom had told in order to shift that +whipping from her shoulders to his own, the Judge said with a fine +outburst that it was a noble, a generous, a magnanimous lie--a lie that +was worthy to hold up its head and march down through history breast to +breast with George Washington's lauded Truth about the hatchet! Becky +thought her father had never looked so tall and so superb as when he +walked the floor and stamped his foot and said that. She went straight +off and told Tom about it. + +Judge Thatcher hoped to see Tom a great lawyer or a great soldier some +day. He said he meant to look to it that Tom should be admitted to the +National Military Academy and afterward trained in the best law school +in the country, in order that he might be ready for either career or +both. + +Huck Finn's wealth and the fact that he was now under the Widow +Douglas' protection introduced him into society--no, dragged him into +it, hurled him into it--and his sufferings were almost more than he +could bear. The widow's servants kept him clean and neat, combed and +brushed, and they bedded him nightly in unsympathetic sheets that had +not one little spot or stain which he could press to his heart and know +for a friend. He had to eat with a knife and fork; he had to use +napkin, cup, and plate; he had to learn his book, he had to go to +church; he had to talk so properly that speech was become insipid in +his mouth; whithersoever he turned, the bars and shackles of +civilization shut him in and bound him hand and foot. + +He bravely bore his miseries three weeks, and then one day turned up +missing. For forty-eight hours the widow hunted for him everywhere in +great distress. The public were profoundly concerned; they searched +high and low, they dragged the river for his body. Early the third +morning Tom Sawyer wisely went poking among some old empty hogsheads +down behind the abandoned slaughter-house, and in one of them he found +the refugee. Huck had slept there; he had just breakfasted upon some +stolen odds and ends of food, and was lying off, now, in comfort, with +his pipe. He was unkempt, uncombed, and clad in the same old ruin of +rags that had made him picturesque in the days when he was free and +happy. Tom routed him out, told him the trouble he had been causing, +and urged him to go home. Huck's face lost its tranquil content, and +took a melancholy cast. He said: + +"Don't talk about it, Tom. I've tried it, and it don't work; it don't +work, Tom. It ain't for me; I ain't used to it. The widder's good to +me, and friendly; but I can't stand them ways. She makes me get up just +at the same time every morning; she makes me wash, they comb me all to +thunder; she won't let me sleep in the woodshed; I got to wear them +blamed clothes that just smothers me, Tom; they don't seem to any air +git through 'em, somehow; and they're so rotten nice that I can't set +down, nor lay down, nor roll around anywher's; I hain't slid on a +cellar-door for--well, it 'pears to be years; I got to go to church and +sweat and sweat--I hate them ornery sermons! I can't ketch a fly in +there, I can't chaw. I got to wear shoes all Sunday. The widder eats by +a bell; she goes to bed by a bell; she gits up by a bell--everything's +so awful reg'lar a body can't stand it." + +"Well, everybody does that way, Huck." + +"Tom, it don't make no difference. I ain't everybody, and I can't +STAND it. It's awful to be tied up so. And grub comes too easy--I don't +take no interest in vittles, that way. I got to ask to go a-fishing; I +got to ask to go in a-swimming--dern'd if I hain't got to ask to do +everything. Well, I'd got to talk so nice it wasn't no comfort--I'd got +to go up in the attic and rip out awhile, every day, to git a taste in +my mouth, or I'd a died, Tom. The widder wouldn't let me smoke; she +wouldn't let me yell, she wouldn't let me gape, nor stretch, nor +scratch, before folks--" [Then with a spasm of special irritation and +injury]--"And dad fetch it, she prayed all the time! I never see such a +woman! I HAD to shove, Tom--I just had to. And besides, that school's +going to open, and I'd a had to go to it--well, I wouldn't stand THAT, +Tom. Looky here, Tom, being rich ain't what it's cracked up to be. It's +just worry and worry, and sweat and sweat, and a-wishing you was dead +all the time. Now these clothes suits me, and this bar'l suits me, and +I ain't ever going to shake 'em any more. Tom, I wouldn't ever got into +all this trouble if it hadn't 'a' ben for that money; now you just take +my sheer of it along with your'n, and gimme a ten-center sometimes--not +many times, becuz I don't give a dern for a thing 'thout it's tollable +hard to git--and you go and beg off for me with the widder." + +"Oh, Huck, you know I can't do that. 'Tain't fair; and besides if +you'll try this thing just a while longer you'll come to like it." + +"Like it! Yes--the way I'd like a hot stove if I was to set on it long +enough. No, Tom, I won't be rich, and I won't live in them cussed +smothery houses. I like the woods, and the river, and hogsheads, and +I'll stick to 'em, too. Blame it all! just as we'd got guns, and a +cave, and all just fixed to rob, here this dern foolishness has got to +come up and spile it all!" + +Tom saw his opportunity-- + +"Lookyhere, Huck, being rich ain't going to keep me back from turning +robber." + +"No! Oh, good-licks; are you in real dead-wood earnest, Tom?" + +"Just as dead earnest as I'm sitting here. But Huck, we can't let you +into the gang if you ain't respectable, you know." + +Huck's joy was quenched. + +"Can't let me in, Tom? Didn't you let me go for a pirate?" + +"Yes, but that's different. A robber is more high-toned than what a +pirate is--as a general thing. In most countries they're awful high up +in the nobility--dukes and such." + +"Now, Tom, hain't you always ben friendly to me? You wouldn't shet me +out, would you, Tom? You wouldn't do that, now, WOULD you, Tom?" + +"Huck, I wouldn't want to, and I DON'T want to--but what would people +say? Why, they'd say, 'Mph! Tom Sawyer's Gang! pretty low characters in +it!' They'd mean you, Huck. You wouldn't like that, and I wouldn't." + +Huck was silent for some time, engaged in a mental struggle. Finally +he said: + +"Well, I'll go back to the widder for a month and tackle it and see if +I can come to stand it, if you'll let me b'long to the gang, Tom." + +"All right, Huck, it's a whiz! Come along, old chap, and I'll ask the +widow to let up on you a little, Huck." + +"Will you, Tom--now will you? That's good. If she'll let up on some of +the roughest things, I'll smoke private and cuss private, and crowd +through or bust. When you going to start the gang and turn robbers?" + +"Oh, right off. We'll get the boys together and have the initiation +to-night, maybe." + +"Have the which?" + +"Have the initiation." + +"What's that?" + +"It's to swear to stand by one another, and never tell the gang's +secrets, even if you're chopped all to flinders, and kill anybody and +all his family that hurts one of the gang." + +"That's gay--that's mighty gay, Tom, I tell you." + +"Well, I bet it is. And all that swearing's got to be done at +midnight, in the lonesomest, awfulest place you can find--a ha'nted +house is the best, but they're all ripped up now." + +"Well, midnight's good, anyway, Tom." + +"Yes, so it is. And you've got to swear on a coffin, and sign it with +blood." + +"Now, that's something LIKE! Why, it's a million times bullier than +pirating. I'll stick to the widder till I rot, Tom; and if I git to be +a reg'lar ripper of a robber, and everybody talking 'bout it, I reckon +she'll be proud she snaked me in out of the wet." + + + +CONCLUSION + +SO endeth this chronicle. It being strictly a history of a BOY, it +must stop here; the story could not go much further without becoming +the history of a MAN. When one writes a novel about grown people, he +knows exactly where to stop--that is, with a marriage; but when he +writes of juveniles, he must stop where he best can. + +Most of the characters that perform in this book still live, and are +prosperous and happy. Some day it may seem worth while to take up the +story of the younger ones again and see what sort of men and women they +turned out to be; therefore it will be wisest not to reveal any of that +part of their lives at present. diff --git a/vendor/github.com/pierrec/lz4/testdata/Mark.Twain-Tom.Sawyer.txt.lz4 b/vendor/github.com/pierrec/lz4/testdata/Mark.Twain-Tom.Sawyer.txt.lz4 new file mode 100644 index 000000000..c7ff441fa Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/Mark.Twain-Tom.Sawyer.txt.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/testdata/README.txt b/vendor/github.com/pierrec/lz4/testdata/README.txt new file mode 100644 index 000000000..1ddabf183 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/README.txt @@ -0,0 +1 @@ +The golden data files (*.lz4) have been compressed with the original C implementation of lz4. \ No newline at end of file diff --git a/vendor/github.com/pierrec/lz4/testdata/e.txt b/vendor/github.com/pierrec/lz4/testdata/e.txt new file mode 100644 index 000000000..5ca186f14 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/e.txt @@ -0,0 +1 @@ +2.7182818284590452353602874713526624977572470936999595749669676277240766303535475945713821785251664274274663919320030599218174135966290435729003342952605956307381323286279434907632338298807531952510190115738341879307021540891499348841675092447614606680822648001684774118537423454424371075390777449920695517027618386062613313845830007520449338265602976067371132007093287091274437470472306969772093101416928368190255151086574637721112523897844250569536967707854499699679468644549059879316368892300987931277361782154249992295763514822082698951936680331825288693984964651058209392398294887933203625094431173012381970684161403970198376793206832823764648042953118023287825098194558153017567173613320698112509961818815930416903515988885193458072738667385894228792284998920868058257492796104841984443634632449684875602336248270419786232090021609902353043699418491463140934317381436405462531520961836908887070167683964243781405927145635490613031072085103837505101157477041718986106873969655212671546889570350354021234078498193343210681701210056278802351930332247450158539047304199577770935036604169973297250886876966403555707162268447162560798826517871341951246652010305921236677194325278675398558944896970964097545918569563802363701621120477427228364896134225164450781824423529486363721417402388934412479635743702637552944483379980161254922785092577825620926226483262779333865664816277251640191059004916449982893150566047258027786318641551956532442586982946959308019152987211725563475463964479101459040905862984967912874068705048958586717479854667757573205681288459205413340539220001137863009455606881667400169842055804033637953764520304024322566135278369511778838638744396625322498506549958862342818997077332761717839280349465014345588970719425863987727547109629537415211151368350627526023264847287039207643100595841166120545297030236472549296669381151373227536450988890313602057248176585118063036442812314965507047510254465011727211555194866850800368532281831521960037356252794495158284188294787610852639813955990067376482922443752871846245780361929819713991475644882626039033814418232625150974827987779964373089970388867782271383605772978824125611907176639465070633045279546618550966661856647097113444740160704626215680717481877844371436988218559670959102596862002353718588748569652200050311734392073211390803293634479727355955277349071783793421637012050054513263835440001863239914907054797780566978533580489669062951194324730995876552368128590413832411607226029983305353708761389396391779574540161372236187893652605381558415871869255386061647798340254351284396129460352913325942794904337299085731580290958631382683291477116396337092400316894586360606458459251269946557248391865642097526850823075442545993769170419777800853627309417101634349076964237222943523661255725088147792231519747780605696725380171807763603462459278778465850656050780844211529697521890874019660906651803516501792504619501366585436632712549639908549144200014574760819302212066024330096412704894390397177195180699086998606636583232278709376502260149291011517177635944602023249300280401867723910288097866605651183260043688508817157238669842242201024950551881694803221002515426494639812873677658927688163598312477886520141174110913601164995076629077943646005851941998560162647907615321038727557126992518275687989302761761146162549356495903798045838182323368612016243736569846703785853305275833337939907521660692380533698879565137285593883499894707416181550125397064648171946708348197214488898790676503795903669672494992545279033729636162658976039498576741397359441023744329709355477982629614591442936451428617158587339746791897571211956187385783644758448423555581050025611492391518893099463428413936080383091662818811503715284967059741625628236092168075150177725387402564253470879089137291722828611515915683725241630772254406337875931059826760944203261924285317018781772960235413060672136046000389661093647095141417185777014180606443636815464440053316087783143174440811949422975599314011888683314832802706553833004693290115744147563139997221703804617092894579096271662260740718749975359212756084414737823303270330168237193648002173285734935947564334129943024850235732214597843282641421684878721673367010615094243456984401873312810107945127223737886126058165668053714396127888732527373890392890506865324138062796025930387727697783792868409325365880733988457218746021005311483351323850047827169376218004904795597959290591655470505777514308175112698985188408718564026035305583737832422924185625644255022672155980274012617971928047139600689163828665277009752767069777036439260224372841840883251848770472638440379530166905465937461619323840363893131364327137688841026811219891275223056256756254701725086349765367288605966752740868627407912856576996313789753034660616669804218267724560530660773899624218340859882071864682623215080288286359746839654358856685503773131296587975810501214916207656769950659715344763470320853215603674828608378656803073062657633469774295634643716709397193060876963495328846833613038829431040800296873869117066666146800015121143442256023874474325250769387077775193299942137277211258843608715834835626961661980572526612206797540621062080649882918454395301529982092503005498257043390553570168653120526495614857249257386206917403695213533732531666345466588597286659451136441370331393672118569553952108458407244323835586063106806964924851232632699514603596037297253198368423363904632136710116192821711150282801604488058802382031981493096369596735832742024988245684941273860566491352526706046234450549227581151709314921879592718001940968866986837037302200475314338181092708030017205935530520700706072233999463990571311587099635777359027196285061146514837526209565346713290025994397663114545902685898979115837093419370441155121920117164880566945938131183843765620627846310490346293950029458341164824114969758326011800731699437393506966295712410273239138741754923071862454543222039552735295240245903805744502892246886285336542213815722131163288112052146489805180092024719391710555390113943316681515828843687606961102505171007392762385553386272553538830960671644662370922646809671254061869502143176211668140097595281493907222601112681153108387317617323235263605838173151034595736538223534992935822836851007810884634349983518404451704270189381994243410090575376257767571118090088164183319201962623416288166521374717325477727783488774366518828752156685719506371936565390389449366421764003121527870222366463635755503565576948886549500270853923617105502131147413744106134445544192101336172996285694899193369184729478580729156088510396781959429833186480756083679551496636448965592948187851784038773326247051945050419847742014183947731202815886845707290544057510601285258056594703046836344592652552137008068752009593453607316226118728173928074623094685367823106097921599360019946237993434210687813497346959246469752506246958616909178573976595199392993995567542714654910456860702099012606818704984178079173924071945996323060254707901774527513186809982284730860766536866855516467702911336827563107223346726113705490795365834538637196235856312618387156774118738527722922594743373785695538456246801013905727871016512966636764451872465653730402443684140814488732957847348490003019477888020460324660842875351848364959195082888323206522128104190448047247949291342284951970022601310430062410717971502793433263407995960531446053230488528972917659876016667811937932372453857209607582277178483361613582612896226118129455927462767137794487586753657544861407611931125958512655759734573015333642630767985443385761715333462325270572005303988289499034259566232975782488735029259166825894456894655992658454762694528780516501720674785417887982276806536650641910973434528878338621726156269582654478205672987756426325321594294418039943217000090542650763095588465895171709147607437136893319469090981904501290307099566226620303182649365733698419555776963787624918852865686607600566025605445711337286840205574416030837052312242587223438854123179481388550075689381124935386318635287083799845692619981794523364087429591180747453419551420351726184200845509170845682368200897739455842679214273477560879644279202708312150156406341341617166448069815483764491573900121217041547872591998943825364950514771379399147205219529079396137621107238494290616357604596231253506068537651423115349665683715116604220796394466621163255157729070978473156278277598788136491951257483328793771571459091064841642678309949723674420175862269402159407924480541255360431317992696739157542419296607312393763542139230617876753958711436104089409966089471418340698362993675362621545247298464213752891079884381306095552622720837518629837066787224430195793793786072107254277289071732854874374355781966511716618330881129120245204048682200072344035025448202834254187884653602591506445271657700044521097735585897622655484941621714989532383421600114062950718490427789258552743035221396835679018076406042138307308774460170842688272261177180842664333651780002171903449234264266292261456004337383868335555343453004264818473989215627086095650629340405264943244261445665921291225648893569655009154306426134252668472594914314239398845432486327461842846655985332312210466259890141712103446084271616619001257195870793217569698544013397622096749454185407118446433946990162698351607848924514058940946395267807354579700307051163682519487701189764002827648414160587206184185297189154019688253289309149665345753571427318482016384644832499037886069008072709327673127581966563941148961716832980455139729506687604740915420428429993541025829113502241690769431668574242522509026939034814856451303069925199590436384028429267412573422447765584177886171737265462085498294498946787350929581652632072258992368768457017823038096567883112289305809140572610865884845873101658151167533327674887014829167419701512559782572707406431808601428149024146780472327597684269633935773542930186739439716388611764209004068663398856841681003872389214483176070116684503887212364367043314091155733280182977988736590916659612402021778558854876176161989370794380056663364884365089144805571039765214696027662583599051987042300179465536788567430285974600143785483237068701190078499404930918919181649327259774030074879681484882342932023012128032327460392219687528340516906974194257614673978110715464186273369091584973185011183960482533518748438923177292613543024932562896371361977285456622924461644497284597867711574125670307871885109336344480149675240618536569532074170533486782754827815415561966911055101472799040386897220465550833170782394808785990501947563108984124144672821865459971596639015641941751820935932616316888380132758752601460507676098392625726411120135288591317848299475682472564885533357279772205543568126302535748216585414000805314820697137262149755576051890481622376790414926742600071045922695314835188137463887104273544767623577933993970632396604969145303273887874557905934937772320142954803345000695256980935282887783710670585567749481373858630385762823040694005665340584887527005308832459182183494318049834199639981458773435863115940570443683515285383609442955964360676090221741896883548131643997437764158365242234642619597390455450680695232850751868719449064767791886720306418630751053512149851051207313846648717547518382979990189317751550639981016466414592102406838294603208535554058147159273220677567669213664081505900806952540610628536408293276621931939933861623836069111767785448236129326858199965239275488427435414402884536455595124735546139403154952097397051896240157976832639450633230452192645049651735466775699295718989690470902730288544945416699791992948038254980285946029052763145580316514066229171223429375806143993484914362107993576737317948964252488813720435579287511385856973381976083524423240466778020948399639946684833774706725483618848273000648319163826022110555221246733323184463005504481849916996622087746140216157021029603318588727333298779352570182393861244026868339555870607758169954398469568540671174444932479519572159419645863736126915526457574786985964242176592896862383506370433939811671397544736228625506803682664135541448048997721373174119199970017293907303350869020922519124447393278376156321810842898207706974138707053266117683698647741787180202729412982310888796831880854367327806879771659111654224453806625861711729498038248879986504061563975629936962809358189761491017145343556659542757064194408833816841111166200759787244137082333917886114708228657531078536674695018462140736493917366254937783014074302668422150335117736471853872324040421037907750266020114814935482228916663640782450166815341213505278578539332606110249802273093636740213515386431693015267460536064351732154701091440650878823636764236831187390937464232609021646365627553976834019482932795750624399645272578624400375983422050808935129023122475970644105678361870877172333555465482598906861201410107222465904008553798235253885171623518256518482203125214950700378300411216212126052726059944320443056274522916128891766814160639131235975350390320077529587392412476451850809163911459296071156344204347133544720981178461451077872399140606290228276664309264900592249810291068759434533858330391178747575977065953570979640012224092199031158229259667913153991561438070129260780197022589662923368154312499412259460023399472228171056603931877226800493833148980338548909468685130789292064242819174795866199944411196208730498064385006852620258432842085582338566936649849720817046135376163584015342840674118587581546514598270228676671855309311923340191286170613364873183197560812569460089402953094429119590295968563923037689976327462283900735457144596414108229285922239332836210192822937243590283003884445701383771632056518351970100115722010956997890484964453434612129224964732356126321951155701565824427661599326463155806672053127596948538057364208384918887095176052287817339462747644656858900936266123311152910816041524100214195937349786431661556732702792109593543055579732660554677963552005378304619540636971842916168582734122217145885870814274090248185446421774876925093328785670674677381226752831653559245204578070541352576903253522738963847495646255940378924925007624386893776475310102323746733771474581625530698032499033676455430305274561512961214585944432150749051491453950981001388737926379964873728396416897555132275962011838248650746985492038097691932606437608743209385602815642849756549307909733854185583515789409814007691892389063090542534883896831762904120212949167195811935791203162514344096503132835216728021372415947344095498316138322505486708172221475138425166790445416617303200820330902895488808516797258495813407132180533988828139346049850532340472595097214331492586604248511405819579711564191458842833000525684776874305916390494306871343118796189637475503362820939949343690321031976898112055595369465424704173323895394046035325396758354395350516720261647961347790912327995264929045151148307923369382166010702872651938143844844532639517394110131152502750465749343063766541866128915264446926222884366299462732467958736383501937142786471398054038215513463223702071533134887083174146591492406359493020921122052610312390682941345696785958518393491382340884274312419099152870804332809132993078936867127413922890033069995875921815297612482409116951587789964090352577345938248232053055567238095022266790439614231852991989181065554412477204508510210071522352342792531266930108270633942321762570076323139159349709946933241013908779161651226804414809765618979735043151396066913258379033748620836695475083280318786707751177525663963479259219733577949555498655214193398170268639987388347010255262052312317215254062571636771270010760912281528326508984359568975961038372157726831170734552250194121701541318793651818502020877326906133592182000762327269503283827391243828198170871168108951187896746707073377869592565542713340052326706040004348843432902760360498027862160749469654989210474443927871934536701798673920803845633723311983855862638008516345597194441994344624761123844617615736242015935078520825600604101556889899501732554337298073561699861101908472096600708320280569917042590103876928658336557728758684250492690370934262028022399861803400211320742198642917383679176232826444645756330336556777374808644109969141827774253417010988435853189339175934511574023847292909015468559163792696196841000676598399744972047287881831200233383298030567865480871476464512824264478216644266616732096012564794514827125671326697067367144617795643752391742928503987022583734069852309190464967260243411270345611114149835783901793499713790913696706497637127248466613279908254305449295528594932793818341607827091326680865655921102733746700132583428715240835661522165574998431236278287106649401564670141943713823863454729606978693335973109537126499416282656463708490580151538205338326511289504938566468752921135932220265681856418260827538790002407915892646028490894922299966167437731347776134150965262448332709343898412056926145108857812249139616912534202918139898683901335795857624435194008943955180554746554000051766240202825944828833811886381749594284892013520090951007864941868256009273977667585642598378587497776669563350170748579027248701370264203283965756348010818356182372177082236423186591595883669487322411726504487268392328453010991677518376831599821263237123854357312681202445175401852132663740538802901249728180895021553100673598184430429105288459323064725590442355960551978839325930339572934663055160430923785677229293537208416693134575284011873746854691620648991164726909428982971065606801805807843600461866223562874591385185904416250663222249561448724413813849763797102676020845531824111963927941069619465426480006761727618115630063644321116224837379105623611358836334550102286170517890440570419577859833348463317921904494652923021469259756566389965893747728751393377105569802455757436190501772466214587592374418657530064998056688376964229825501195065837843125232135309371235243969149662310110328243570065781487677299160941153954063362752423712935549926713485031578238899567545287915578420483105749330060197958207739558522807307048950936235550769837881926357141779338750216344391014187576711938914416277109602859415809719913429313295145924373636456473035037374538503489286113141638094752301745088784885645741275003353303416138096560043105860548355773946625033230034341587814634602169235079216111013148948281895391028916816328709309713184139815427678818067628650978085718262117003140003377301581536334149093237034703637513354537634521050370995452942055232078817449370937677056009306353645510913481627378204985657055608784211964039972344556458607689515569686899384896439195225232309703301037277227710870564912966121061494072782442033414057441446459968236966118878411656290355117839944070961772567164919790168195234523807446299877664824873753313018142763910519234685081979001796519907050490865237442841652776611425351538665162781316090964802801234493372427866930894827913465443931965254154829494577875758599482099181824522449312077768250830768282335001597040419199560509705364696473142448453825888112602753909548852639708652339052941829691802357120545328231809270356491743371932080628731303589640570873779967845174740515317401384878082881006046388936711640477755985481263907504747295012609419990373721246201677030517790352952793168766305099837441859803498821239340919805055103821539827677291373138006715339240126954586376422065097810852907639079727841301764553247527073788764069366420012194745702358295481365781809867944020220280822637957006755393575808086318932075864444206644691649334467698180811716568665213389686173592450920801465312529777966137198695916451869432324246404401672381978020728394418264502183131483366019384891972317817154372192103946638473715630226701801343515930442853848941825678870721238520597263859224934763623122188113706307506918260109689069251417142514218153491532129077723748506635489170892850760234351768218355008829647410655814882049239533702270536705630750317499788187009989251020178015601042277836283644323729779929935160925884515772055232896978333126427671291093993103773425910592303277652667641874842441076564447767097790392324958416348527735171981064673837142742974468992320406932506062834468937543016787815320616009057693404906146176607094380110915443261929000745209895959201159412324102274845482605404361871836330268992858623582145643879695210235266673372434423091577183277565800211928270391042391966426911155333594569685782817020325495552528875464466074620294766116004435551604735044292127916358748473501590215522120388281168021413865865168464569964810015633741255098479730138656275460161279246359783661480163871602794405482710196290774543628092612567507181773641749763254436773503632580004042919906963117397787875081560227368824967077635559869284901628768699628053790181848148810833946900016380791075960745504688912686792812391148880036720729730801354431325347713094186717178607522981373539126772812593958220524289991371690685650421575056729991274177149279608831502358697816190894908487717722503860872618384947939757440664912760518878124233683125467278331513186758915668300679210215947336858591201395360301678110413444411030903388761520488296909104689167671555373346622545575975202624771242796225983278405833585897671474205724047439720232895903726148688388003174146490203843590358527993123871042845981608996101945691646983837718267264685264869172948414153004604004299585035164101899027529366867431834955447458124140190754681607770977920579383895378192128847409929537040546962226547278807248685508046571043123854873351653070570784584243335550958221912862797205455466267099131902370311779690892786623112661337671178512943059323281605826535623848164192144732543731002062738466812351691016359252588256806438946389880872735284406462208149513862275239938938734905082625472417781702582044129853760499827899020083498387362992498125742354568439023012261733665820546785671147973065077035475620567428300187473019197310881157516777005071432012726354601912460800451608108641835539669946936947322271670748972850464195392966434725254724357659192969949061670189061433616907056148280980363243454128229968275980226694045642181328624517549652147221620839824594576613342710564957193564431561774500828376935700995419541839029151033187933907614207467028867968594985439789457300768939890070073924697461812855764662265412913204052279071212820653775058280040897163467163709024906774736309136904002615646432159560910851092445162454420141442641660181385990017417408244245378610158433361777292580611159192008414091888191208858207627011483671760749046980914443057262211104583300789331698191603917150622792986282709446275915009683226345073725451366858172483498470080840163868209726371345205439802277866337293290829914010645589761697455978409211409167684020269370229231743334499986901841510888993165125090001163719114994852024821586396216294981753094623047604832399379391002142532996476235163569009445086058091202459904612118623318278614464727795523218635916551883057930657703331498510068357135624341881884405780028844018129031378653794869614630467726914552953690154167025838032477842272417994513653582260971652588356712133519546838335349801503269359798167463231847628306340588324731228951257944267639877946713121042763380872695738609314631539148548792514028885025189788076023838995615684850391995855029256054176767663145354058496296796781349420116003325874431438746248313850214980401681940795687219268462617287403480967931949965604299190281810597603263251746405016454606266765529010639868703668263299050577706266397868453584384057673298268163448646707439990917504018892319267557518354054956017732907127219134577524905771512773358423314008356080926962298894163047287780054743798498545562870729968407382937218623831766524716090967192007237658894226186550487552614557855898773008703234726418384831040394818743616224455286163287628541175946460497027724490799275146445792982549802258601001772437840167723166802004162547244179415547810554178036773553354467030326469619447560812831933095679685582771932031205941616693902049665352189672822671972640029493307384717544753761937017882976382487233361813499414541694736549254840633793674361541081593464960431603544354737728802361047743115330785159902977771499610274627769759612488879448609863349422852847651310277926279743981957617505591300993377368240510902583759345170015340522266144077237050890044496613295859536020556034009492820943862994618834790932894161098856594954213114335608810239423706087108026465913203560121875933791639666437282836752328391688865373751335794859860107569374889645657187292540448508624449947816273842517229343960137212406286783636675845331904743954740664015260871940915743955282773904303868772728262065663129387459875317749973799293043294371763801856280061141619563942414312254397099163565102848315765427037906837175764870230052388197498746636856292655058222887713221781440489538099681072143012394693530931524054081215705402274414521876541901428386744260011889041724570537470755550581632831687247110220353727166112304857340460879272501694701067831178927095527253222125224361673343366384756590949728221809418684074238351567868893421148203905824224324264643630201441787982022116248471657468291146315407563770222740135841109076078464780070182766336227978104546331131294044833570134869585165267459515187680033395522410548181767867772152798270250117195816577603549732923724732067853690257536233971216884390878879262188202305529937132397194333083536231248870386416194361506529551267334207198502259771408638122015980894363561808597010080081622557455039101321981979045520049618583777721048046635533806616517023595097133203631578945644487800945620369784973459902004606886572701865867757842758530645706617127194967371083950603267501532435909029491516973738110897934782297684100117657987098185725131372267749706609250481876835516003714638685918913011736805218743265426063700710595364425062760458252336880552521181566417553430681181548267844169315284408461087588214317641649835663127518728182948655658524206852221830755306118393326934164459415342651778653397980580828158806300749952897558204686612590853678738603318442905510689778698417735603118111677563872589911516803236547002987989628986181014596471307916144369564690909518788574398821730583884980809523077569358851616027719521488998358632323127308909861560777386006984035267826785387215920936255817889813416247486456433211043194821421299793188104636399541496539441501383868748384870224681829391860319598667962363489309283087840712400431022706137591368056518861313458307990705003607588327248867879324093380071864152853317943535073401891193638546730000660453783784472469288830546979000131248952100446949032058838294923613919284305249167833012980192255157050378521810552961623637523647962685751660066539364142273063001648652613891842243501797455993616794063303522111829071597538821839777552812981538570168702202620274678647916644030729018445497956399844836807851997088201407769199261674991148329821854382718946282165387064858588646221611410343570342878862979083418871606214430014533275029715104673156021000043869510583773779766003460887624861640938645252177935289947578496255243925598620521409052346250847830487046492688313289470553891357290706967599556298586669559721686506052072801342104355762779184021797626656484580261591407173477009039475168017709900129391137881248534255949312866653465033728846390649968460644741907524313323903404908195233044389559060547854954620263256676813262435925020249516275607080900436460421497025691488555265022810327762115842282433269528629137662675481993546118143913367579700141255870143319434764035725376914388899683088262844616425575034001428982557620386364384137906519612917777354183694676232982904981261717676191554292570438432239918482261744350470199171258214687683172646078959690569981353264435973965173473319484798758064137926885413552523275720457329477215706850016950046959758389373527538622664943456437071610511521617176237598050900553232154896062817794302268640579555845730600598376482703339859420098582351400179507104569019191359062304102336798080907240196312675268916362136351032648077232914950859151265812143823371072949148088472355286394195993455684156344577951727033374238129903260198160571971183950662758220321837136059718025940870615534713104482272716848395524105913605919812444978458110854511231668173534838253724825347636777581712867205865148285317273569069839935110763432091319780314031658897379628301178409806410175016511072932907832177487566289310650383806093372841399226733384778203302020700517188941706465146238366720632742644336612174011766914919235570905644803016342294301837655263108450172510307540942604409687066288066265900569082451407632599158164499361455172452057020443093722305550217222299706209749268609762787409626448772056043078634808885709143464793241536214303199965695610753570417207285334250171325558818113295504095217830139465216436594262960768570585698507157151317262928960072587601564840556088613165411835958628710665496282599535127193244635791046554389165150954187306071015034430609582302257455974944275067630926322529966338219395202927917973247094559691016402983683080426309910481567503623509654924302589575273521412445149542462972258510120707802110188106722347972579330653187713438466713807546383471635428854957610942841898601794658721444495198801550804042506452191484989920400007310672369944655246020908767882300064337725657385010969899058191290957079866699453765080407917852438222041070599278889267745752084287526377986730360561230710723922581504781379172731261234878334034473833573601973235946604273704635201327182592410906040097638585857716958419563109577748529579836844756803121874818202833941887076311731615289811756429711334181497218078040465077657204457082859417475114926179367379999220181789399433337731146911970737861041963986422166045588965683206701337505745038872111332436739840284188639147633491695114032583475841514170325690161784931455706904169858050217798497637014758914810543205854914100662201721719726878930012101267481270235940855162601689425111458499658315589660460091525797881670384625905383256920520425791378948827579603278877535466861441826827797651258953563761485994485049706638406266121957141911063246061774180577212381659872472432252969098533628440799030007594546281549235506086481557928961969617060715201589825299772803520002610888814176506636216905928021516429198484077446143617891415191517976537848282687018750030264867608433204658525470555882410254654806040437372771834769014720664234434374255514129178503032471263418076525187802925534774001104853996960549926508093910691337614841834884596365621526610332239417467064368340504749943339802285610313083038484571294767389856293937641914407036507544622061186499127249643799875806537850203753189972618014404667793050140301580709266213229273649718653952866567538572115133606114457222800851183757899219543063413692302293139751143702404830227357629039911794499248480915071002444078482866598579406525539141041497342780203520135419925977628178182825372022920108186449448349255421793982723279357095828748597126780783134286180750497175747373730296280477376908932558914598141724852658299510882230055223242218586191394795184220131553319634363922684259164168669438122537135960710031743651959027712571604588486044820674410935215327906816032054215967959066411120187618531256710150212239401285668608469435937408158536481912528004920724042172170913983123118054043277015835629513656274610248827706488865037765175678806872498861657094846665770674577000207144332525555736557083150320019082992096545498737419756608619533492312940263904930982014700371161829485939931199955070455381196711289367735249958182011774799788636393286405807810818657337668157893827656450642917396685579555053188715314552353070355994740186225988149854660737787698781542360397080977412361518245964026869979609564523828584235953564615185448165799966460648261396618720304839119560250381111550938420209894591555760083897989949964566262540514195610780090298667014635238532066032574466820259430618801773091109212741138269148784355679352572808875543164693077235363768226036080174040660997151176880434927489197133087822951123746632635635328517394189466510943745768270782209928468034684157443127739811044186762032954475468077511126663685479944460934809992951875666499902261686019672053749149951226823637895865245462813439289338365156536992413109638102559114643923805213907862893561660998836479175633176725856523591069520326895990054884753424160586689820067483163174286329119633399132709086065074595260357157323069712106423424081597068328707624437165532750228797802598690981111226558888151520837482450034463046505984569690276166958278982913613535306291331427881888249342136442417833519319786543940201465328083410341785272489879050919932369270996567133507711905899945951923990615156165480300145359212550696405345263823452155999210578191371030188979206408883974767667144727314254467923500524618849237455307575734902707342496298879996942094595961008702501329453325358045689285707241207965919809225550560061971283541270202072583994171175520920820151096509526685113897577150810849443508285458749912943857563115668324566827992991861539009255871716840495663991959154034218364537212023678608655364745175654879318925644085274489190918193411667583563439758886046349413111875241038425467937999203546910411935443113219136068129657568583611774564654674861061988591414805799318725367531243470335482637527081353105570818049642498584646147973467599315946514787025065271083508782350656532331797738656666181652390017664988485456054961300215776115255813396184027067814900350252876823607822107397102339146870159735868589015297010347780503292154014359595298683404657471756232196640515401477953167461726208727304820634652469109953327375561090578378455945469160223687689641425960164689647106348074109928546482353083540132332924864037318003195202317476206537726163717445360549726690601711176761047774971666890152163838974311714180622222345718567941507299526201086205084783127474791909996889937275229053674785020500038630036526218800670926674104806027341997756660029427941090400064654281074454007616429525362460261476180471744322889953285828397762184600967669267581270302806519535452053173536808954589902180783145775891280203970053633193821100095443241244197949192916205234421346395653840781209416214835001155883618421164283992454027590719621537570187067083731012246141362048926555668109467076386536083015847614512581588569610030337081197058344452874666198891534664244887911940711423940115986970795745946337170243268484864632018986352827092313047089215684758207753034387689978702323438584381125011714013265769320554911860153519551654627941175593967947958810333935413289702528893533748106257875620364294270257512121137330213811951395756419122685155962476203282038726342066227347868223036522019655729325905068134849292299647248229359787842720945578267329975853818536442370617353517653060396801087899490506654491544577952166038552398013798104340564182403396162494910454712104839439200945914647542424785991096900046541371091630096785951563947332190934511838669964622788855817353221326876634958059123761251203010983867841195725887799206041260049865895027247133146763722204388398558347770112599424691208308595666787531942465131444389971195968105937957532155524204659410081418351120174196853432672343271868099625045432475688702055341969199545300952644398446384346598830418262932239295612610045884644244285011551557765935780379565026806130721758672048541797157896401554276881090475899564605488362989140226580026134158039480357971019004151547655018391755772677897148793477372747525743898158705040701968215101218826088040084551332795162841280679678965570163917067779841529149397403158167896865448841319046368332179115059107813898261026271979696826411179918656038993895418928488851750122504754778999508544083983800725431468842988412616042682248823097788556495765424017114510393927980290997604904428832198976751320535115230545666467143795931915272680278210241540629795828828466355623580986725638200565215519951793551069127710538552661926903526081367717666435071213453983711357500975854405939558661737828297120544693182260401670308530911657973113259516101749193468250063285777004686987177255226525708428745733039859744230639751837209975339055095883623642814493247460522424051972825153787541962759327436278819283740253185668545040893929401040561666867664402868211607294830305236465560955351079987185041352121321534713770667681396211443891632403235741573773787908838267618458756361026435182951815392455211729022985278518025598478407179607904114472041476091765804302984501746867981277584971731733287305281134969591668387877072315968334322509070204019030503595891994666652037530271923764252552910347950343816357721698115464329245608951158732012675424975710520894362639501382962152214033621065422821876739580121286442788547491928976959315766891987305176388698461503354594898541849550251690616888419122873385522699976822609645007504500096116866129171093180282355042553653997166054753907348915189650027442328981181709248273610863801576007240601649547082331349361582435128299050405405333992577071321011503713898695076713447940748097845416328110406350804863393555238405735580863718763530261867971725608155328716436111474875107033512913923595452951407437943144900950809932872153235195999616750297532475931909938012968640379783553559071355708369947311923538531051736669154087312467233440702525006918026747725078958903448856673081487299464807786497709361969389290891718228134002845552513917355978456150353144603409441211512001738697261466786933733154341007587514908295822756919350542184106448264951943804240543255345965248373785310657979037977505031436474651422484768831323479762673689855474944277949916560108528257618964374464656819789319422077536824661110427671936481836360534108748971066866318805026555929568123959680449295166615409802610781691689418764353363449482900125929366840591370059526914934421861891742142561071896846626335874414976973921566392767687720145153302241853125308442727245771161505550519076276250016522166274796257424425420546785767478190959486500575711016264847833741198041625940813327229905891486422127968042984725356237202887830051788539737909455265135144073130049869453403245984236934627060242579432563660640597549471239092372458126154582526667304702319359866523378856244229188278436440434628094888288712101968642736370461639297485616780079779959696843367730352483047478240669928277140069031660709951473154191919911453182543906294573298686613524886500574780251977607442660798300291573030523199052185718628543687577860915726925232573171665625274275808460620177046433101212443409281314659760221360416223031167750085960128475289259463348312408766740128170543067985261868949895004918275008304998926472034986965363326210919830621495095877228260815566702155693484634079776879525038204442326697479264829899016938511552124688935873289878336267819361764023681714606495185508780596635354698788205094762016350757090024201498400967867845405354130050482404996646978558002628931826518708714613909521454987992300431779500489569529280112698632533646737179519363094399609176354568799002814515169743717518330632232942199132137614506411391269837128970829395360832883050256072727563548374205497856659895469089938558918441085605111510354367477810778500572718180809661542709143010161515013086522842238721618109043183163796046431523184434669799904865336375319295967726080853457652274714047941973192220960296582500937408249714373040087376988068797038047223488825819819025644086847749767508999164153502160223967816357097637814023962825054332801828798160046910336602415904504637333597488119998663995617171089911809851197616486499233594328274275983382931099806461605360243604040848379619072542165869409486682092396143083817303621520642297839982533698027039931804024928814430649614747600087654305571672697259114631990688823893005380061568007730984416061355843701277573463708822073792921409548717956947854414951731561828176343929570234710460088230637509877521391223419548471196982303169544468045517922669260631327498272520906329003279972932906827204647650366969765227673645419031639887433042226322021325368176044169612053532174352764937901877252263626883107879345194133825996368795020985033021472307603375442346871647223795507794130304865403488955400210765171630884759704098331306109510294140865574071074640401937347718815339902047036749084359309086354777210564861918603858715882024476138160390378532660185842568914109194464566162667753712365992832481865739251429498555141512136758288423285957759412684479036912662015308418041737698963759002546999454131659341985624780714434977201991702665380714107259910648709897259362243300706760476097690456341576573395549588448948093604077155688747288451838106069038026528318275560395905381507241627615047252487759578650784894547389096573312763852962664517004459626327934637721151028545472312880039058405918498833810711366073657536918428084655898982349219315205257478363855266205400703561310260405145079325925798227406012199249391735122145336707913500607486561657301854049217477162051678486507913573336334257685988361252720250944019430674728667983441293018131344299088234006652915385763779110955708000600143579956351811596764725075668367726052352939773016348235753572874236648294604770429166438403558846422370760111774821079625901180265548868995181239470625954254584491340203400196442965370643088660925268811549596291166168612036195319253262662271108142149856132646467211954801142455133946382385908540917878668826947602781853283155445565265933912487885639504644196022475186011405239187543742526581685003052301877096152411653980646785444273124462179491306502631062903402737260479940181929954454297256377507172705659271779285537195547433852182309492703218343678206382655341157162788603990157495208065443409462446634653253581574814022471260618973060860559065082163068709634119751925774318683671722139063093061019303182326666420628155129647685313861018672921889347039342072245556791239578260248978371473556820782675452142687314252252601795889759116238720807580527221031327444754083319215135934526961397220564699247718289310588394769170851420631557192703636345039529604362885088555160008371973526383838996789184600327073682083234847108471706160879195227388252347506380811606090840124222431476103563328940609282430125462013806032608121942876847907192546246309055749298781661271916548229644317263587524548607563020667656942355342774617635549231817456159185668061686428714964129290560130053913469569829490891003991259088290348791943368696942620662946948514931472688923571615032405542263391673583102728579723061998175868700492227418629077079508809336215346303842967525604369606110193842723883107587771653594778681499030978765900869583480043137176832954871752604714113064847270887246697164585218774442100900090916189819413456305028950484575822161887397443918833085509908566008543102796375247476265353031558684515120283396640547496946343986288291957510384781539068343717740714095628337554413567955424664601335663617305811711646062717854078898495334329100315985673932305693426085376230981047171826940937686754301837015557540822371538037838383342702379535934403549452173960327095407712107332936507766465603712364707109272580867897181182493799540477008369348889220963814281561595610931815183701135104790176383595168144627670903450457460997444500166918675661035889313483800512736411157304599205955471122443903196476642761038164285918037488354360663299436899730090925177601162043761411616688128178292382311221745850238080733727204908880095181889576314103157447684338100457385008523652069340710078955916549813037292944462306371284357984809871964143085146878525033128989319500645722582281175483887671061073178169281242483613796475692482076321356427357261609825142445262515952514875273805633150964052552659776922077806644338105562443538136258941809788015677378951310313157361136026047890761945591820289365770116416881703644242694283057457471567494391573593353763114830246668754727566653059819746822346578699972291792416156043557665183382167059157867799311835820189855730344883681934418305987021880502259192818047775223884407167894780414701414651073580452021499197980812095692195622632313741870979731320870864552236740416185590793816745658234353037283309503729022429802768451559528656923189798000383061378732434546500582722712325031420712488100290697226311129067629080951145758060270806092801504406139446350643069742785469477459876821004441453438033759717384777232052065301037861326418823586036569054773343070911759152582503029410738914441818378779490613137536794654893375260322906277631983337976816641721083140551864133302224787118511817036598365960493964571491686005656771360533192423185262166760222073368844844409234470948568027905894191829969467724456269443308241243846160408284006424867072583661011433404214473683453638496544701067827313169538435919120440283949541956874453676459875488726170687163109591315801609722382049772577307454562979127906177531663252857205858766376754282917933549923678212008601904369428956102301731743150352204665675088491593025926618816581008701658499456495586855628208747248318351516339189292646558880593601275151838235485893426165223086697314511412035659916934103076974774451947043836739600076578628245472064617380804602903639144493859012422380173377038154675297645596518492676039300171943042511794045679862114630138402371099347243455794730048929825402680821621522346560274258486595687074510352794291633405915025075992398611224340312056999780516223878772230396359709132856830486160362127579561601328561866388146004722200580017580282279272167842720649966956840905752590774886105493806116954293569077377792821084159737469613143291808510446953973485067590503662391722108732333169909603363771705474725026941732982890400239372879549386540463828596742216318201530139629734398479588628632934746650690284066719018081265539973675916799759010867483920062877888531102781695087545740384607594616919584610655963327283485609570305572502494416337066573150237126843581984154103154401008430380631442183776750349813408169325201240813452285974626715177152223063741359255747513535160669108359443999692315898156732033027129284241219651936303734407981204656795322986357374589031654007016472204989445629050395873788912680565516464274460174738175296313458739390484560414203426465560422112239134631023161290836446988901247285192778589195228773637440432659264672239982186452797664826673070168802722052338600372842903155828454593854349099449420750911108532138744823216151007808922516285123275724355101999038195993350032641446053470357293073912578481757987468353429629749652545426864234949270336399427519354240001973125098882419600095766257217621860474573769577649582201796258392376391717855799468922496750179251915218219624653575570564228220399546682648329822996167217080156801080799777126517156274295763666959661983507435667132218383358509536665806605597148376773866922551603463644386269977295750658468929599809168949981898588529537874489519527097766262684177088590284321676352132630838812766335363319004134332844347630067982023716933653652880580156390360562722752187272454764258840995216482554453662083811789117725225682611478014242896970967121967502094421226279437073328703410646312100557376727450271638975234111426287828736758358819056742163061523416789476056879277154789714326222041069587947186435439940738639948986836168919377836648327137363654676901173760246643082285362494712605173293777247276797635865806019396287718060679122426813922872134061694882029506831654589707623668302556167559477498715183426989208952182644710514911419441192277010977616645850068963849426165593473112961064282379048216056210094265076173838082479030510998790719611852832556787472942907151041468948104916751035295897242381802288151276582257190705537652455285511598636421244284176256230139538669970308943645907600684938040875210854159851278070333207779865635907968462191534944587677170063778573171211036517486371634098385626541555573292664616402279791195975248525300376741774056125700303625811704838385391207273191845064713669122576415213769896260940351804147432053600369234179035440735703058314741623452840188940808983125191307741823338981880316339159565954543405777784331681162551898060409183018907512170192983622897099598983405484962284289398469847938668614293324543983592637036699355184231661615244505980576745765335552338715678211466689996845227042954589710922163652573965950289645637766038988037941517917867910675199009966139206238732318786758420544279396366759104126821843375015743069045967947046685602358283919759975285865384338189120042853787549302768972168199113340697282255535300044743958830079799736518459131437946494086272149669719100359399974735262764126125995350902609540048669398955899487421379590802893196914845826873123710180229775301190684280440780938156598081694611679374425663244656799606363751546304833112722231812338371779800439731087402647536582575657351059978314264831879619843765495877803685261751835391844920488198629786329743136948511780579298636452193232481339393090754566368038513630619718033957979522539508697432546502659123585049283028832934489284591373621624852528877442891851104093746333590660233239711922814450735588373324057814862662207486215513375036775585494138678352928273109003823116855374520901095101174796663003330352534143230024288248051396631446632656081582045216883922312025671065388459503224002320453633895521539919011035217362720909565500846486605368975498478995875596103167696587161281951919668893326641203784750417081752273735270989343717167642329956935697166213782736138899530515711822960896394055380431939398453970864418654291655853168697537052760701061488025700785387150835779480952313152747735711713643356413242974208137266896149109564214803567792270566625834289773407718710649866150447478726164249976671481383053947984958938064202886667951943482750168192023591633247099185942520392818083953020434979919361853380201407072481627304313418985942503858404365993281651941497377286729589582881907490040331593436076189609669494800067194371424058105327517721952474344983414191979918179909864631583246021516575531754156198940698289315745851842783390581029411600498699307751428513021286202539508732388779357409781288187000829944831476678183644656510024467827445695591845768068704978044824105799710771577579093525803824227377612436908709875189149049904225568041463131309240101049368241449253427992201346380538342369643767428862595140146178201810734100565466708236854312816339049676558789901487477972479202502227218169405159042170892104287552188658308608452708423928652597536146290037780167001654671681605343292907573031466562485809639550080023347676187068086526878722783177420214068980703410506200235273632267291964034093571225623659496432076928058165514428643204955256838543079254299909353199329432966018220787933122323225928276556048763399988478426451731890365879756498207607478270258861409976050788036706732268192473513646356758611212953074644777149423343867876705824452296605797007134458987594126654609414211447540007211790607458330686866231309155780005966522736183536340439991445294960728379007338249976020630448806064574892740547730693971337007962746135534442514745423654662752252624869916077111131569725392943756732215758704952417232428206555322808868670153681482911738542735797154157943689491063759749151524510096986573825654899585216747260540468342338610760823605782941948009334370046866568258579827323875158302566720152604684361412652956519894291184887986819088277339147282063794512260294515707367105637720023427811802621502691790400488001808901847311751199425460594416773315777951735444490965752131026306836047140331442314298077895617051256930051804287472368435536402764392777908638966566390166776625678575354239947427919442544664643315554138265543388487778859972063679660692327601733858843763144148113561693030468420017434061395220072403658812798249143261731617813894970955038369479594617979829257740992171922783223006387384996138434398468502234780438733784470928703890536420557474836284616809363650973790900204118525835525201575239280826462555785658190226958376345342663420946214426672453987171047721482128157607275305173330963455909323664528978019175132987747952929099598069790148515839540444283988381797511245355548426126784217797728268989735007954505834273726937288386902125284843370917479603207479554080911491866208687184899550445210616155437083299502854903659617362726552868081324793106686855857401668022408227992433394360936223390321499357262507480617409173636062365464458476384647869520547719533384203403990244761056010612777546471464177412625548519830144627405538601855708359981544891286863480720710061787059669365218674805943569985859699554089329219507269337550235821561424994538234781138316591662683103065194730233419384164076823699357668723462219641322516076261161976034708844046473083172682611277723613381938490606534404043904909864126903479263503943531836741051762565704797064478004684323069430241749029731181951132935746854550484711078742905499870600373983113761544808189067620753424526993443755719446665453524088287267537759197074526286322840219629557247932987132852479994638938924943286917770190128914220188747760484939855471168524810559991574441551507431214406120333762869533792439547155394213121021954430556748370425907553004950664994802614794524739012802842646689229455664958621308118913500279654910344806150170407268010067948926855360944990373928383520627992820181576427054962997401900837493444950600754365525758905546552402103412862124809003162941975876195941956592556732874237856112669741771367104424821916671499611728903944393665340294226514575682907490402153401026923964977275904729573320027982816062130523130658731513076913832317193626664465502290735017347656293033318520949298475227462534564256702254695786484819977513326393221579478212493307051107367474918016345667888810782101151826314878755138027101379868751299375133303843885631415175908928986956197561123025310875057188962535763225834275763348421016668109884514141469311719314272028007223449941999003964948245457520704922091620614222912795322688239046498239081592961111003756999529251250673688233852648213896986384052437049402152187547825163347082430303521036927849762517317825860862215614519165573478940019558704784741658847364803865995119651409542615026615147651220820245816010801218275982577477652393859159165067449846149161165153821266726927461290533753163055654440793427876550267301214578324885948736899073512166118397877342715872870912311383472485146035661382188014840560716074652441118841800734067898587159273982452147328317214621907330492060817440914125388918087968538960627860118193099489240811702350413554126823863744341209267781729790694714759018264824761112414556423937732224538665992861551475342773370683344173073150805440138894084087253197595538897613986400165639906934600670780501058567196636796167140097031535132386972899001749862948883362389858632127176571330142071330179992326381982094042993377790345261665892577931395405145369730429462079488033141099249907113241694504241391265397274078984953073730364134893688060340009640631540701820289244667315059736321311926231179142794944897281477264038321021720718017561601025111179022163703476297572233435788863537030535008357679180120653016668316780269873860755423748298548246360981608957670421903145684942967286646362305101773132268579232832164818921732941553151386988781837232271364011755881332524294135348699384658137175857614330952147617551708342432434174779579226338663454959438736807839569911987059388085500837507984051126658973018149321061950769007587519836861526164087252594820126991923916722273718430385263107266000047367872474915828601694439920041571102706081507270147619679971490141639274282889578424398001497985658130305740620028554097382687819891158955487586486645709231721825870342960508203415938806006561845735081804032347750084214100574577342802985404049555529215986404933246481040773076611691605586804857302606467764258503301836174306413323887707999698641372275526317649662882467901094531117120243890323410259937511584651917675138077575448307953064925086002835629697045016137935696266759775923436166369375035368699454550392874449940328328128905560530091416446608691247256021455381248285307613556149618444364923014290938289373215312818797541139219415606631622784836152140668972661027123715779503062132916001988806369127647416567067485490795342762338253943990022498972883660263920518704790601584084302914787302246651371144395418253441269003331181914268070735159284180415100555199146564934872796969351992963117195821262627236458009708099166752820365818699111948365866102758375863322993225541477479210421324166848264953111826527351008031659958888814809945737293785681411438021523876706455063233067233939551964260397443829874822322662036352861302543796600943104500158604854027036789711934695579989189112302233381602302236277726084846296189550730850698061500281436425336666311433321645213882557346329366870956708432252564333895997812402164189946978348320376011613913855499933990786652305860332060641949298931012423081105800169745975038516887112037747631577311831360002742502722451570906304496369230938382329175076469684003556425503797106891999812319602533733677437970687713814747552190142928586781724044248049323750330957002929126630316970587409214456472022710796484778657310660832173093768033821742156446602190335203981531618935787083561603302255162155107179460621892674335641960083663483835896703409115513087820138723494714321400450513941428998350576038799343355677628023346565854351219361896876831439866735726040869511136649881229957801618882834124004126142251475184552502502640896823664946401177803776799157180146386554733265278569418005501363433953502870836220605121839418516239153709790768084909674194289061134979961034672077354959593868862427986411437928435620575955500144308051267664432183688321434583708549082240014585748228606859593502657405750939203135881722442164955416889785558265198046245527898343289578416968890756237467281044803018524217706136533236073856228166664597654076844715963930782091017090763377917711485205493367936868430832404126789220929930411890501756484917499452393770674524578019171841679541825554377930299249277892416277257788147974770446005423669346157135208417428211847353652367573702352791459837645712257646122605628127852169580892808988394594406165340521932514843306105322700231133680378433377389724881307874325614952744243584753011150345103737688223837573804282007358586938044331529253129961025096113761670187568525921208929131354473196308440066835155160913925692912175784379179004808848023029304392630921342768601226558630456913133560978156776098711809238440656353136182676923761613389237802972720736243967239854144480757286813436768000573823963610796223140429490728058551444771338682314499547929338131259971996894072233847404542592316639781608209399269744676323921370773991899853301483814622364299493902073285072098040905300059160091641710175605409814301906444379905831277826625762288108104414704097708248077905168225857235732665234414956169007985520848841886027352780861218049418060017941147110410688703738674378147161236141950474056521041002268987858525470689031657094677131822113205505046579701869337769278257145248837213394613987859786320048011792814546859096532616616068403160077901584946840224344163938313618742275417712170336151163782359059685168880561304838542087505126933144171705880517278127917564053282929427357971823360842784676292324980318169828654166132873909074116734612367109059236155113860447246378721244612580406931724769152219217409096880209008801535633471775664392125733993165330324425899852598966724744126503608416484160724482125980550754851232313331300621490042708542735985913041306918279258584509440150719217604794274047740253314305451367710311947544521321732225875550489799267468541529538871443696399406391099267018219539890685186755868574434469213792094590683677929528246795437302263472495359466300235998990248299853826140395410812427393530207575128774273992824866921285637240069184859771126480352376025469714309316636539718514623865421671429236191647402172547787238964043145364190541101514371773797752463632741619269990461595895793940622986041489302535678633503526382069821487003578061101552210224486633247184367035502326672749787730470216165019711937442505629639916559369593557640005236360445141148916155147776301876302136068825296274460238077523189646894043033182148655637014692476427395401909403584437251915352134557610698046469739424511797999048754951422010043090235713636892619493763602673645872492900162675597083797995647487354531686531900176427222751039446099641439322672532108666047912598938351926694497553568096931962642014042788365702610390456105151611792018698900673027082384103280213487456720062839744828713298223957579105420819286308176631987048287388639069922461848323992902685392499812367091421613488781501234093387999776097433615750910992585468475923085725368613605356762146929424264323906626708602846163376051573599050869800314239735368928435294958099434465414316189806451480849292695749412903363373410480943579407321266012450796613789442208485840536446021616517885568969302685188950832476793300404851688934411125834396590422211152736276278672366665845757559585409486248261694480201791748223085835007862255216359325125768382924978090431102048708975715033330963651576804501966025215527080352103848176167004443740572131294252820989545456276344353575741673638980108310579931697917916718271145837435222026387771805250290791645414791173616253155840768495583288190293564201219633684854080865928095131505012602919562576032932512847250469881908146475324342363863860247943921015193235101390117789997483527186469346024554247028375300033725403910085997650987642832802908445662021678362267272292737780213652404028817217012490974899454430826861772239385250883760749742195942655217301733355851389407457348144161511380845358039740277795072051893487170722955427683655826706766313911972211811528466502223383490906676554168336907959409404576472940901354356409277969379842065738891481990225399022315913388145851487225126560927576795873759207013915029216513720851137197522734365458411622066281660256333632074449918511469174455062297146086578736313585389023662557285424516018080487167823688885575325066254262367702604215835160174851981885460860036597606743233346410471991027562358645341748631726556391320606407754779439671383653877377610828300019937359760370467245737880967939894493795829602910746901609451288456550071458091887879542641820145369659962842686882363495879277007025298960996798975941955735253914237782443302746708282008722602053415292735847582937522487377937899136764642153727843553986244015856488692101644781661602962113570056638347990334049623875941092886778920270077504951511405782565295015024484968204744379710872943108541684540513016310902267112951959140520827546866418137305837933236150599142045255880213558474751516267815309465541240524091663857551298894834797423322854504140527354235070335984964593699534959698554244978249586929179182415068053002553370412778703476446244329205906832901886692400222391918714603175399666877477960121790688623311002908668305431787009355066944389131913333586368037447530664502418437136030852288582121720231274167009740351431532131803978033680228154223490183737494117973254478594157962104378787072154814091725163615415163381388912588517924237727229603497305533840942889918919161186249580560073570527227874940321250645426206304469470804277945973817146810395192821550688079136701210109944220737024613687196031491162370967939354636396448139025711768057799751751298979667073292674886430097398814873780767363792886767781170520534367705731566895899181530825761606591843760505051704242093231358724816618683821026679970982966436224723644898648976857100173643547336955619347638598187756855912376232580849341570570863450733443976604780386678461711520325115528237161469200634713570383377229877321365028868868859434051205798386937002783312365427450532283462669786446920780944052138528653384627970748017872477988461146015077617116261800781557915472305214759943058006652042710117125674185860274188801377931279938153727692612114066810156521441903567333926116697140453812010040811760123270513163743154487571768761575554916236601762880220601068655524141619314312671535587154866747899398685510873576261006923021359580838145290642217792987748784161516349497309700794368305080955621264592795333690631936594413261117944256602433064619312002953123619348034504503004315096798588111896950537335671086336886944665564112662287921812114121425167348136472449021275252555647623248505638391391630760976364990288930588053406631352470996993362568102360392264043588787550723319888417590521211390376609272658409023873553418516426444865247805763826160023858280693148922231457758783791564902227590699346481624734399733206013058796068136378152964615963260698744961105368384203105364183675373594176373955988088591188920114871545460924735613515979992999722298041707112256996310945945097765566409972722824015293663094891067963296735505830412258608050740410916678539569261234499102819759563955711753011823480304181029089719655278245770283085321733741593938595853203645590564229716679900322284081259569032886928291260139267587858284765599075828016611120063145411315144108875767081854894287737618991537664505164279985451077400771946398046265077776614053524831090497899859510873112620613018757108643735744708366215377470972660188656210681516328000908086198554303597948479869789466434027029290899143432223920333487108261968698934611177160561910681226015874410833093070377506876977485840324132474643763087889666151972556180371472590029550718424245405129246729039791532535999005557334600111693557020225722442772950263840538309433999383388018839553821540371447394465152512354603526742382254148328248990134023054550811390236768038649723899924257800315803725555410178461863478690646045865826036072306952576113184134225274786464852363324759102670562466350802553058142201552282050989197818420425028259521880098846231828512448393059455162005455907776121981297954040150653985341579053629101777939776957892084510979265382905626736402636703151957650493344879513766262192237185642999150828898080904189181015450813145034385734032579549707819385285699926238835221520814478940626889936085239827537174490903769904145555260249190126341431327373827075950390882531223536876389814182564965563294518709637484074360669912550026080424160562533591856230955376566866124027875883101021495284600804805028045254063691285010599912421270508133194975917146762267305044225075915290251742774636494555052325186322411388406191257012917881384181566918237215400893603475101448554254698937834239606460813666829750019379115061709452680984785152862123171377897417492087541064556959508967969794980679770961683057941674310519254486327358885118436597143583348756027405400165571178309126113117314169066606067613797690123141099672013123730329707678988740099317309687380126740538923612230370779727025191340850390101739924877352408881040807749924412635346413181858792480760553268122881584307471326768283097203149049868884456187976015468233715478415429742230166504759393312132256510189175368566338139736836336126010908419590215582111816677413843969205870515074254852744810154541079359513596653630049188769523677579147319184225806802539818418929888943038224766186405856591859943091324575886587044653095332668532261321209825839180538360814144791320319699276037194760191286674308615217243049852806380129834255379486287824758850820609389214668693729881191560115633701248675404205911464930888219050248857645752083363921499441937170268576222251074166230901665867067714568862793343153513505688216165112807318529333124070912343832502302341169501745502360505475824093175657701604884577017762183184615567978427541088499501610912720817913532406784267161792013428902861583277304794830971705537485109380418091491750245433432217445924133037928381694330975012918544596923388733288616144238100112755828623259628572648121538348900698511503485369544461542161283241700533583180520082915722904696365553178152398468725451306350506984981006205514844020769539324155096762680887603572463913955278222246439122592651921288446961107463586148252820017348957533954255019475442643148903233373926763409115527189768429887783617346613535388507656327107814312435018965109238453660236940276060642119384227665755210663671879603217527184404651560427289869560206997012906367847161654793068868305846508082886614111979138822898112498261434559408961813509226857611474609406147937240008842153535862052780125014270055274468359151840373309373580494342483940467505708347927948338133276237937844629209323999417593374917899786484958148818865149169302451512835579818112344900827168644548306546633975256079615935830821400021951611342337058359111545217293721664061708131602078213341260356852013161345136871600980378712556766143923146458085652084039744217352744813741215277475202259244561520365608268890193913957991844109971588312780020898275935898106482117936157951837937026741451400902833064466209280549839169261068975151083963132117128513257434964510681479694782619701483204392206140109523453209269311762298139422044308117317394338867965739135764377642819353621467837436136161591167926578700137748127848510041447845416464568496606699139509524527949914769441031612575776863713634644477006787131066832417871556281779122339077841275184193161188155887229676749605752053192594847679397486414128879475647133049543555044790277128690095643357913405127375570391806822344718167939329121448449553897728696601037841520390662890781218240141299368590465146519209198605347788576842696538459445700169758422531241268031418456268722581132040056433413524302102739213788415250475704533878002467378571470021087314693254557923134757243640544448132093266582986850659125571745568328831440322798049274104403921761438405750750288608423536966715191668510428001748971774811216784160854454400190449242294333666338347684438072624307319019363571067447363413698467328522605570126450123348367412135721830146848071241856625742852208909104583727386227300781566668914250733456373259567253354316171586533339843321723688126003809020585719930855573100508771533737446465211874481748868710652311198691114058503492239156755462142467550498676710264926176510110766876596258810039163948397811986615585196216487695936398904500383258041054420595482859955239065758108017936807080830518996468540836412752905182813744878769639548306385089756146421874889271294890398025623046812175145502330254086076115859321603465240763923593699949180470780496764486889980902123735780457040380820770357387588525976042434608851075199334470112741787878845674656640471901619633546770714090590826954225196409446319547658653032104723804625249971910690110456227579220926904132753699634145768795242244563973018311291451151322757841320376225862458224784696669785947914981610522628786944136373683125108310682898766123782697506343047263278453719024447970975017396831214493357290791648779915089163278018852504558488782722376705263811803792477835540018117452957747339714012352011459901984753358434861297092928529424139865507522507808919352104173963493428604871342370429572757862549365917805401652536330410692033704691093097588782938291296447890613200063096560747882082122140978472301680600835812336957051454650181292694364578357815608503303392466039553797630836137289498678842851139853615593352782103740733076818433040893624460576706096188294529171362940967592507631348636606011346115980434147450705511490716640635688739020690279453438236930531133440901381392849163507484449076828386687476663619303412376248380175840467851210698290605196112357188811150723607303158506622574566366740720668999061320627793994112805759798332878792144188725498543014546662945079670707688135022230580562225942983096887732856788971494623888272184647618153045844390967248232348259587963698908456664795754200195991919240707615823002328977439748112690476546256873684352229063217889227643289360535947903046811114130586348244566489159211382258867880972564351646404364328416076247766114349880319792230537889671148058968061594279189647401954989466232962162567264739015818692956765601444248501821713300527995551312539849919933907083138030214072556753022600033565715934283182650908979350869698950542635843046765145668997627989606295925119763672907762567862769469947280606094290314917493590511523235698715397127866718077578671910380368991445381484562682604003456798248689847811138328054940490519768008320299631757043011485087384048591850157264392187414592464617404735275250506783992273121600117160338604710710015235631159734711153198198710616109850375758965576728904060387168114313084172893710817412764581206119054145955378853200366615264923610030157044627231777788649806700723598889528747481372190175074700005571108178930354895017924552067329003818814068686247959272205591627902292600592107710510448103392878991286820705448979977319695574374529708195463942431669050083984398993036790655541596099324867822475424361758944371791403787168166189093900243862038610001362193667280872414291108080291896093127526202667881902085595708111853836166128848729527875143202956393295910508349687029060692838441522579419764824996318479414814660898281725690484184326061946254276693688953540732363428302189694947766126078346328490315128061501009539164530614554234923393806214007779256337619373052025699319099789404390847443596972052065999017828537676265683558625452697455260991024576619614037537859594506363227095122489241931813728141668427013096050734578659047904243852086508154491350136491698639048125666610843702294730266721499164849610746803261583352580352858275799038584091667618877199539888680431991650866887781701439663176815592262016991396613153738021294160006906947533431677802632207226265881842757216055461439677336258462997385077307751473833315101468395296411397329672457933540390136107395245686243008096720460995545708974893048753897955544443791303790422346037768729236001386569593952300768091377768847789746299699489949016141866131552200856673695770822720338936659590666350594330040363762591189195691561626122704788696510356062748423100605472091437069471661080277379848576543481249822444235828329813543645124092220896643987201997945619030397327254617823136363375927622656301565813545578319730419339269008282952718252138855126583037630477490625995514925943105307478901043009876580816508144862607975129633326675259272351611791836777128931053144471668835182920514343609292493191180249366051791485330421043899773019267686085347768149502299280938065840007311767895491286098112311307002535600347898600653805084532572431553654422067661352337408211307834360326940015926958459588297845649462271300855594293344520727007718206398887404742186697709349647758173683580193168322111365547392288184271373843690526638607662451284299368435082612881367358536293873792369928837047900484722240370919885912556341130849457067599032002751632513926694249485692320904596897775676762684224768120033279577059394613185252356456291805905295974791266162882381429824622654141067246487216174351317397697122228010100668178786776119825961537643641828573481088089988571570279722274734750248439022607880448075724807701621064670166965100202654371260046641935546165838945950143502160890185703558173661823437491622669077311800121188299737319891006060966841193266075165452741829459541189277264192546108246351931647783837078295218389645376236304858042774417907169146356546201215125418664885396161542055152375000426794253417764590821513675258479774465114750438460596325820468809667795709044645884673847481638045635188183210386594798204376334738389017759714236223057776395541011294523488098341476645559342209402059733452337956309441446698222457026367119493286653989491344225517746402732596722993581333110831711807234044326813737231209669052411856734897392234152750707954137453460386506786693396236535556479102508529284294227710593056660625152290924148057080971159783458351173168204129645967070633303569271821496292272073250126955216172649821895790908865085382490848904421755530946832055636316431893917626269931034289485184392539670922412565933079102365485294162132200251193795272480340133135247014182195618419055761030190199521647459734401211601239235679307823190770288415814605647291481745105388060109787505925537152356112290181284710137917215124667428500061818271276125025241876177485994084521492727902567005925854431027704636911098800554312457229683836980470864041706010966962231877065395275783874454229129966623016408054769705821417128636329650130416501278156397799631957412627634011130135082721772287129164002237230234809031485343677016544959380750634285293053131127965945266651960426350406454862543383772209428482543536823186182982713182489884498260285705690699045790998144649193654563259496570044689011049923939218088155626191834404362264965506449848521612498442375928443642612004256628602157801140467879662339228190804577624109076487087406157070486658398144845855803277997327929143195789110373530019873110486895656281917362036703039179710646309906285483702836118486672219457621775034511770110458001291255925462680537427727378863726783016568351092332280649908459179620305691566806180826586923920561895421631986004793961133953226395999749526798801074576466538377400437463695133685671362553184054638475191646737948743270916620098057717103475575333102702706317395612448413745782734376330101853438497450236265733191742446567787499665000938706441886733491099877926005340862442833450486907338279348425305698737469497333364267191968992849534561045719338665222471536681145666596959735075972188416698767321649331898967182978657974612216573922404856900225324160367805329990925438960169901664189038843548375648056012628830409421321300206164540821986138099462721214327234457806819925823202851398237118926541234460723597174777907172041523181575194793527456442984630888846385381068621715274531612303165705848974316209831401326306699896632888532682145204083110738032052784669279984003137878996525635126885368435559620598057278951754498694219326972133205286374577983487319388899574634252048213337552584571056619586932031563299451502519194559691231437579991138301656117185508816658756751184338145761060365142858427872190232598107834593970738225147111878311540875777560020664124562293239116606733386480367086953749244898068000217666674827426925968686433731916548717750106343608307376281613984107392410037196754833838054369880310983922140260514297591221159148505938770679068701351029862207502287721123345624421024715163941251258954337788492834236361124473822814504596821452253550035968325337489186278678359443979041598043992124889848660795045011701169092519383155609441705397900600291315024253848282782826223304151370929502192196508374714697845805550615914539506437316401173317807741497557116733034632008408954066541694665746735785483133770133628948904397670025863002540635264006601631712883920305576358989492412827022489373848906764385339931878608019223108328847459816417701264089078551777830131616162049792779670521847212730327970738223860581986744668610994383049960437407323195784473254857416239738852016202384784256163512597161783106850156299135559874758848151014815490937380933394074455700842090155903853444962128368313687375166780513082594599771257467939781491953642874321122421579851584491669362551569370916855252644720786527971466476760328471332985501945689772758983450586004316822658631176606237201721007922216410188299330808409384014213759697185976897042759041500946595252763487628135867117352364964121058854934496645898651826545634382851159137631569519895230262881794959971545221250667461174394884433312659432286710965281109501693028351496524082850120190831078678067061851145740970787563117610746428835593915985421673115153096948758378955979586132649569817205284291038172721213138681565524428109871168862743968021885581515367531218374119972919471325465199144188500672036481975944167950887487934416759598361960010994838744709079104099785974656112459851972157558134628546189728615020774374529539536929655449012953097288963767713353842429715394179547179095580120134210175150931491664699052366350233024087218654727629639065723341455005903913890253699317155917179823065162679744711857951506573868504088229934804445549850597823297898617029498418376255258757455303112991914341109413088238114443068843062655305601658801408561023324210300218460588586954418502977463085858496130037238190325162225570729975710727306066072916922978033647048840958711228045188511908718588299514331534128549297173849768523136276076868494780364948299904475715771141080958058141208956059471668626290036145602625334863284986816039463372436667112964460292915746181117789169695839947080954788863503281129626899231110099889317815313946681882028368363373822281414974006917942192888817139116283910295684918233358930813360131488748366464224381776081007739183393749346933644748150564933649323157235306109385796839902153381449126925350768211098738352197507736653475499431740580563099143218212547336281359488317681489194306530426029773885492974570569448783077945878865062970895499843760181694031056909587141386804846359853684034105948341788438963179956468815791937174656705047441528027712541569401365862097760735632832966564135817028088013546326104892768731829917950379944446328158595181380144716817284996793061814177131912099236282922612543236071226270324572637946863533391758737446552006008819975294017572421299723542069630427857950608911113416534893431149175314953530067419744979017235181671568754163484949491289001739377451431928382431183263265079530371177806185851153508809998200482761808307209649636476943066172549186143700971387567940218696710148540307471561091358933165600167252126542502898612259306484105898847129649230941215144563947889999327145875969555737090855150648002321476443037232466147111552578583071024936898814562568786834745518893385181791667579054210421036349316257870476543126790661216644142285017446278477132740595579600648343288827864837043456066966456899746910373987712891593313271266247505582258634928427718355831641593667712218537642376222104779338956378722902509543014182257180331300148113377736941508488867501893156994849838936052666818012783912005801431596441910546663236810148207799356523056490420711364192200177189107935243234322761787712568251126481332974354926568682748715986654943041648468220593921673359485057849622807932422649812705271398407720995707236227009245067665680069149966555737866411877079767754867028786431817941521796178310655030287157272282250812017060713380339641841211253856248920130010782462165136989511064611133562443838185366273563783436921279354709230119655914915800561707258518503167289370411936374780625824298250726464801821523430268081486978164824349353456855843696378384153838051184406043696871666416514036129729992912630842812149152469877429332305214999981829046119471676727503742221367186614654042534463141660649871499001000660041544868437352208483059495953182872280520828676300361091734508632133033647289584176588755345227938480297724485711815574893561311524926772006362198369980664159549388683836411891430443767715498026544959061738265591178545999378510861446014967645550103653971251138583505085112442517772923814396233043724036032603181442991365750246012787514117944901305803452199992701148071712847770301254994886841867572975189214295652512486943983729047410363121899124217339550688778643130750024823361832738729697376598820053895902935486054979802320400472236873557411858132734337978931582039412878989728973298812553514507641535360519462112217000676321611195841029252568536561813138784086477147099724553013170761712163186600291464501378587854802096244703771373587720086738054108140042311418525803293267396324596914044834665722042880679280616029884043400536534009706581694636096660911110968789751801325224478246957913251892122653056085866541115373584912790254654369020869419871125588453729063224423222287139122012248769976837147645598526739225904997885514250047585260297929306159913444898341973583316070107516452301310796620382579278533125161760789984630103493496981494261055367836366022561213767081421091373531780682420175737470287189310207606953355721704357535177461573524838432101571399813798596607129664438314791296359275429627129436142685922138993054980645399144588692472767598544271527788443836760149912897358259961869729756588978741082189422337344547375227693199222635973520722998387368484349176841191020246627479579564349615012657433845758638834735832242535328142047826934473129971189346354502994681747128179298167439644524956655532311649920677163664580318205849626132234652606175413532444702007661807418914040158148560001030119994109595492321434406067634769713089513389171050503856336503545166431774489640061738861761193622676890576955693918707703942304940038440622614449572516631017080642923345170422426679607075404028551182398361531383751432493056398381877995594942545196756559181968690885283434886050828529642437578712929439366177362830136595872723080969468398938676366226456791132977469812675226595621009318322081754694778878755356188335083870248295346078597023609865656376722755704495258739871812593441903785275571333409842450127258596692434317689018966145404453679047136294238156127656824247864736176671770647002431119711090007474065945650315375044177982192306323700872039212085499569681061379189029961178936752146022386905665481382858280449537530160921422195940638787074787991194920898374091788534417523064715030278397979864517336625329511775105559014160459873338186887977858817291976604516353353556047648420520888811722831990044504284486852338334530105533929637308039738230604714104525470094899407601215247602819963846343554852932377161410869591950786873276075400085220065031871239272857835807010762542769655355964789450166013816295177908531139811092831583216931563867459747449584385282701658246192092219529134323496779345585613140207765996142546463288677356891785576835169608392864188830094883324700447958316931533832382377876344426323456301679513671047510469669001217777128065522453689371871451567394733440447280450959433090683667110655953338602938000999949010642769859623260401863733572846679531229683156358145420890540651226419162015504500430562136991850941034609601030543816694795964585804425194905110733387679946734471718615647723811737035654917628707589456035519195603962301157866323750234725054461073979402475184415558178087962822231972692984516683306919505079993357259165675557294585962182052650473353712351623662770479333289322136141858785972771685682725303734836891911847197133753088446777943274857148827821608844765700041403499921376794209627560883081509438030705666022764678117533361028187800710219794428777313146387857817205661409023041499923248268982477222109852189758140879763486146763606368674611966620347304608917277240045953051376938375381543486981101990651706961774052218247422657652138152740612699012706880875386408669901461740890540981877671880076124151967064152117653084325544261017536348281196837493395825742541244634247233586360777980960199745187758845459645895956779558869098404768259253477849930457883128541747079059795909431627722327844578918694214929451540174214623240300841907975296782445969183509474202123617940309048634960534054931299919496087957952586977170236680033862505764938088740994009589948109397983231108838769236490221499111120870639202892490698435333152727991330986335454324971441378059132240814960156485679843966464780280409057580889190254236606774500413415794312112501275232250148067232979652230488493751166084976116412777395311302041566848265531411348993243747890268935173904043294851610659785832253168204202834993641595980197343889883020994152152288611175126686173051956249367180053845637855129171848417841594797435580617856680758491080185805695567990185198397660693358224779136504562705766735170961550493338390452612404395517449136885115987454340932040102218982707539212403241042424451570052968378815749468441508011138612561164102477190903050040240662278945607061512108266146098662040425010583978098192019726759010749924884966139441184159734610382401178556739080566483321039073867083298691078093495828888707110651559651222542929154212923108071159723275797510859911398076844732639426419452063138217862260999160086752446265457028969067192282283045169111363652774517975842147102219099906257373383472726498678244401048998507631630668050267115944636293525120269424810854530602810627264236538250773340575475701704367039596467715959261029438313074897245505729085688496091346323165819468660587092144653716755655531962091865952628448253731353698162517351930115341581171353292035873164168839107994000677266031617527582917398395852606454113318985505747847121053505795649095931672167565624818782002769963734155880000867852567422461511406015760115910256449002264980039498403358091309140197877843650167960167465370287466062584346329708303725980494653589318912163976013193079476972058034710553111117215859219066231028099212084069283091906017370764654655683413207556315315006453462321007133584907633048328153458698497332599801187479664273140279381289961720524540674695271948079930396730194274036466594154400092799908634806622334906695224044652158992864203435098858422692019340575496840904812955522654754650713532842543496616084954788090727649930252702815067862810825243222979985391759845188868387004477101866772159439708514664612871148749531862180941719676843144666435175837688436786081446319641912566574047718699160915550910878919431253671945651261878486910876729910565595155159739659034383628124629118117760949411880105946336671039049777312004243578115790429823045072038322781246413671297959415082918378213212876890545963586369344879749784841123274921331663162812456388238288715648447883142417650147980187858215768793063001153788998014623690135803753306246148576074932567807682651045738059018831237617271889933790487113395588485234240255002352200613574914318259142479829367775490496399350755839668967578364316618369307625603528602940662803255416535431518013714821941772672244005268401996533334184004345525296592918502940131600651124395297874364222806977720437363717873457948420238745151249157913139411148608416429347958793681868609689684640858334131017858142710955416293375915178392341303110543328703526599993904966822112768158316511246866451167351378214345336650598328347443536290312393672084593164394941881138607974670134709640378534907149089842317891739783650654751982883367395714360000003439863363212091718954899055748693397700245632475954504411422582410783866837655467400137324322809113692670682805397549111166171102397437749479335174036135005397581475520834285772800986189401984375446435081498218360112577632447389452051636938585136484259964518361856989088721789764694721246807900330925083496645841656554261294195108847197209106605105540933731954888406444080280579549008076040034154662137669606444293774985897353625591959618552448187940317374508256072895120945456562159540405425814886929842786582357673195799285293120866275922366115137445767916063621675267440451221051052090834707443986137829082352772895849625656881972792768694795806100573787084121444815034797422312103295359297822377134077549545477791813823542607184617108389097825964406170543546968567030745411634244134486308676327949177682923093183221341455482591367202823284396549001805653203960795517074496039006696990334199278212696767771835209083959545341866777944872740383733381985235884202840150981579594685874537989503257362809837592216229258598599123843993575573285028613155970362934249814178056461615863415338635077223269996508860870999964899373049307170967888740149746147542880387421250689212155876692242387434701120990859082164073576380817386959755176083877600277517253037133445654852635661720197563001580049790223419586738061442401502436288957503206533690825756785507020555105572381878574650371086308158185862815883054564662297694803970618265491385181326737485227188267917919091354407852685476254126683398240534022469989966652573155637645862251862823092085424412805997628505488913098331761884983352975136073772030571342739638126588567405013841074788943393996603591853934198416322617654857376671943132840050626295140357877264680649549355746326408186979718630218760025813995719923601345374229758918285167511358171472625828596940798518571870075823122317068134867930884899275181661399609753105295773584618525865211893339375771859916335112163441037910451845019023066893064178977808158101360449495409665363660370075881004450265734935127707426742578608784898185628869980851665713320835842613381142623855420315774246613108873106318111989880289722849790551075148403702290580483052731884959994156606537314021296702220821915862905952604040620011815269664910068587592655660567562963361434230232810747488395040380984981860056164646099819257616235478710913832967563761506732550860683433720438748186791668975746563456020002562889601191100980453350423842063824039434163502977688802779835087481178298349417211674919425601608685332435385951152061809031241698182079314615062073826097180458265687043623935757495737332781578904386011378078508110273049446611821957450170106059384336519458628360682108585130499820420578458577175933849015564447305834515291412561679970569657426139901681932056241927977282026714297258700193234337873153939403115411184101414292741703537542003698760608765500109345299007034032401334806388514095769557147190364152027721127070187421548123931953220997506553022646844227700020589045922742423904937051507367764629844971682121994198274794049092601715727439368569721862936007387077810797440975556627807371228030350048829843919546433753355787895064018998685060281902452191177018634505171087023903398550540704454189088472042376499749035038518949505897971286631644699407490959473411581934618336692169573605081585080837952036335619947691937965065016808710250735070825260046821242820434367245824478859256555487861614478717581068572356895150707602217433511627331709472765932413249132702425519391509083601346239612335001086614623850633127072987745618984384288764099836164964775714638573247333226653894523588365972955159905187411779288608760239306160016168434070611663449248395156319152882728822831375458678269830696691220130954815935450754923554167766876455212545681242936427474153815692219503331560151614492247512488957534835926226263545406704767033866410025277276800886383266629488582740369655329362236090572479794734434077704284318507901973469071141230364111729224929307731939309795452877412451183953480382210373644697046967493042810911797232448615413264031578430955396671061468083815548947146733652483679138566431084747848676243012018489329109615281108087617422779131629345494425395422727309645057976122885347393189600810965202090151104579377602529543130188938184010247010134929317443562883578609861545691161669857388024973756940558138630581099823372565164920155443216861690537054630176154809626620800633059320775897175589925862195462096455464624399535391743228225433267174308492508396461328929584567927365409119947616225155964704061297047759818551878441419948614013153859322060745185909608884280218943358691959604936409651570327527570641500776261323783648149005245481413195989296398441371781402764122087644989688629798910870164270169014007825748311598976330612951195680427485317886333041169767175063822135213839779138443325644288490872919067009802496281560626258636942322658490628628035057282983101266919109637258378149363774960594515216932644945188292639525772348420077356021656909077097264985642831778694777804964343991762549216500608626285329471055602670413384500507827390640287529864161287496473708235188892189612641279553536442286955430551308700009878557534223100547153412810957024870812654319123261956462149376527526356402127388765103883255007364899937167183280028398832319373301564123277185395654932422977953016534830128490677845037490891749347389015649588574802194996722621185874361039774946338633057887487405540005440439344888192044102134790034598411927024921557026873700970995205391930979319495883265922171508324621942300185974396706491149559411733728199869021311629886680267446443489233020607003821262841723679627307191405008084085703978151998148822390059948911946474438682533745889962375133378280532928272016815977970066488394482446332210928320504045983008943565954267256879714918703447338237767914829203283196838105907715727191903042365315650957464549643425328069510396558733549803850995143463506175361480050195045201350200180281506933241918267855737764414097080945745624854867704904368368717590918057269794010465019484853146726642978667687697789291431128505043098192949736165944259471754765135205245072597538577958372797702972231435199958499522344049394502115428867244188717409524554771867484911475031801773304689909317974472957035192387686405544278134169807249382219749124257510162187439772902147704638010731470653154201300583810458905006764557332998149945854655105526374914354195867992595981412218735238407957416123372264063860431988936249867649693592569592128495906254446474331759999685163660305216426770428154681777589339252115538590526823311608302751194384823861552852465010329467297198112105314125898165100120742688143577590825227466863206188376830450921784582526239594189673003640808624233657620979111641766331328852352062487922978959456450333733139422384778582717195412347860434376165241568717943562570215636666680088531006728947033079540804583324192188488870712275670333173939262509073556164513677064199539111948881240659821685787131385056850623094155206877987539740658484250135205615103489821873770245063583314243624807432542464195984647411575625441010389671576677263196442524931941806472423789334668561083789808830313571333157729435664956078125304917594015895146954965223118559669048559467607968190167266634650186182955669893965019614544401768162810604465068448139561667220729261210164692339016793399632833013163850830967942792934551268435760356901970523138364640961311774904600772840862214747547653221505518116489887879087780918009050706040061220010051271575991225725282523378026809030528461581739558198122397010092017202251606352922464781615533532275453264543087093320924631855976580561717446840450048285353396546862678852330044967795580761661801833668792312510460809773895565488962815089519622093675058841609752282328250433712970186608193748968699961301486924694482420723632912367052542145464162968910442981633373266871675946715392611950649224725627254543274193495995569590243279097174392258098103601486364409101491734183079646345064833303404765711827040276868271418084574998493392039317445402616663674646668754385093967129918067471909885312710726724428584870694307099756567949198418996425748884764622030325637751112534060087936904565779272035205921345924272965206683338510673615276261016026647772485083344719891986802656197236420847504962661607797092906844757798251795569758235084371746103310387911789239441630112634077535773520558040066982523191225570519133631407211349723226549151062961739050617857127509403623146700931176133132018631158730886798239298009805089491510788371194099750375473674305745187265414016446924576792185753680363289139664155342066705623272936001177781498886100830877849571709880858667023104043242526785955562077310543072298032125941107957349146684680220501816192150766649106862033378713826058987655210423668198670177861672671972374156917880001690656659046965316154923604061891820982414006103779407166342002735828911994182647812782659666207030384795881442790246669264032799404016800137293477301530941805070587421153284642203006550763966756168318897005152026656649929417382840327305940740147117478464839241225676523593418554066440983706083636457657081801664285044258224551650808864421212113914352453935225522162483791737330329812349528984098613273709957407786789349311975204237925022851375880436791854547836416773151821457226504640800104202100410766027807729152555503218182387221708112766208665317651926458452495269685376314437998340336947124447247796973890514941120010934140073794061859447165516612674930799374705772930521750426383798367668159183589049652163726492960837147204067428996276720315410211504333742057182854090136325721437592054640471894328548696883599785122262130812989581571391597464534806099601555877223193450760315411663112963843719400333736013305526352571490454327925190794007111504785378036370897340146753465517470747096935814912797188187854376797751675927822300312945518595042883902735494672667647506072643698761394806879080593531793001711000214417701504495496412454361656210150919997862972495905809191825255486358703529320142005857057855419217730505342687533799076038746689684283402648733290888881745453047194740939258407362058242849349024756883352446212456101562729065130618520732925434179252299417447855189995098959999877410951464170076989305620163502192692653166599093238118295411937545448509428621839424186218067457128099385258842631930670182098008050900019819621758458932516877698594110522845465835679362969619219080897536813210484518784516230623911878024604050824909336069998094776253792973597037759066145994638578378211017122446355845171941670344732162722443265914858595797823752976323442911242311368603724514438765801271594060878788638511089680883165505046309006148832545452819908256238805872042843941834687865142541377686054291079721004271658 diff --git a/vendor/github.com/pierrec/lz4/testdata/e.txt.lz4 b/vendor/github.com/pierrec/lz4/testdata/e.txt.lz4 new file mode 100644 index 000000000..e039fd6d0 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/e.txt.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/testdata/gettysburg.txt b/vendor/github.com/pierrec/lz4/testdata/gettysburg.txt new file mode 100644 index 000000000..2c9bcde36 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/gettysburg.txt @@ -0,0 +1,29 @@ + Four score and seven years ago our fathers brought forth on +this continent, a new nation, conceived in Liberty, and dedicated +to the proposition that all men are created equal. + Now we are engaged in a great Civil War, testing whether that +nation, or any nation so conceived and so dedicated, can long +endure. + We are met on a great battle-field of that war. + We have come to dedicate a portion of that field, as a final +resting place for those who here gave their lives that that +nation might live. It is altogether fitting and proper that +we should do this. + But, in a larger sense, we can not dedicate - we can not +consecrate - we can not hallow - this ground. + The brave men, living and dead, who struggled here, have +consecrated it, far above our poor power to add or detract. +The world will little note, nor long remember what we say here, +but it can never forget what they did here. + It is for us the living, rather, to be dedicated here to the +unfinished work which they who fought here have thus far so +nobly advanced. It is rather for us to be here dedicated to +the great task remaining before us - that from these honored +dead we take increased devotion to that cause for which they +gave the last full measure of devotion - + that we here highly resolve that these dead shall not have +died in vain - that this nation, under God, shall have a new +birth of freedom - and that government of the people, by the +people, for the people, shall not perish from this earth. + +Abraham Lincoln, November 19, 1863, Gettysburg, Pennsylvania diff --git a/vendor/github.com/pierrec/lz4/testdata/gettysburg.txt.lz4 b/vendor/github.com/pierrec/lz4/testdata/gettysburg.txt.lz4 new file mode 100644 index 000000000..00010be96 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/gettysburg.txt.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/testdata/pg1661.txt b/vendor/github.com/pierrec/lz4/testdata/pg1661.txt new file mode 100644 index 000000000..c4c313050 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/pg1661.txt @@ -0,0 +1,13052 @@ +Project Gutenberg's The Adventures of Sherlock Holmes, by Arthur Conan Doyle + +This eBook is for the use of anyone anywhere at no cost and with +almost no restrictions whatsoever. You may copy it, give it away or +re-use it under the terms of the Project Gutenberg License included +with this eBook or online at www.gutenberg.net + + +Title: The Adventures of Sherlock Holmes + +Author: Arthur Conan Doyle + +Posting Date: April 18, 2011 [EBook #1661] +First Posted: November 29, 2002 + +Language: English + + +*** START OF THIS PROJECT GUTENBERG EBOOK THE ADVENTURES OF SHERLOCK HOLMES *** + + + + +Produced by an anonymous Project Gutenberg volunteer and Jose Menendez + + + + + + + + + +THE ADVENTURES OF SHERLOCK HOLMES + +by + +SIR ARTHUR CONAN DOYLE + + + + I. A Scandal in Bohemia + II. The Red-headed League + III. A Case of Identity + IV. The Boscombe Valley Mystery + V. The Five Orange Pips + VI. The Man with the Twisted Lip + VII. The Adventure of the Blue Carbuncle +VIII. The Adventure of the Speckled Band + IX. The Adventure of the Engineer's Thumb + X. The Adventure of the Noble Bachelor + XI. The Adventure of the Beryl Coronet + XII. The Adventure of the Copper Beeches + + + + +ADVENTURE I. A SCANDAL IN BOHEMIA + +I. + +To Sherlock Holmes she is always THE woman. I have seldom heard +him mention her under any other name. In his eyes she eclipses +and predominates the whole of her sex. It was not that he felt +any emotion akin to love for Irene Adler. All emotions, and that +one particularly, were abhorrent to his cold, precise but +admirably balanced mind. He was, I take it, the most perfect +reasoning and observing machine that the world has seen, but as a +lover he would have placed himself in a false position. He never +spoke of the softer passions, save with a gibe and a sneer. They +were admirable things for the observer--excellent for drawing the +veil from men's motives and actions. But for the trained reasoner +to admit such intrusions into his own delicate and finely +adjusted temperament was to introduce a distracting factor which +might throw a doubt upon all his mental results. Grit in a +sensitive instrument, or a crack in one of his own high-power +lenses, would not be more disturbing than a strong emotion in a +nature such as his. And yet there was but one woman to him, and +that woman was the late Irene Adler, of dubious and questionable +memory. + +I had seen little of Holmes lately. My marriage had drifted us +away from each other. My own complete happiness, and the +home-centred interests which rise up around the man who first +finds himself master of his own establishment, were sufficient to +absorb all my attention, while Holmes, who loathed every form of +society with his whole Bohemian soul, remained in our lodgings in +Baker Street, buried among his old books, and alternating from +week to week between cocaine and ambition, the drowsiness of the +drug, and the fierce energy of his own keen nature. He was still, +as ever, deeply attracted by the study of crime, and occupied his +immense faculties and extraordinary powers of observation in +following out those clues, and clearing up those mysteries which +had been abandoned as hopeless by the official police. From time +to time I heard some vague account of his doings: of his summons +to Odessa in the case of the Trepoff murder, of his clearing up +of the singular tragedy of the Atkinson brothers at Trincomalee, +and finally of the mission which he had accomplished so +delicately and successfully for the reigning family of Holland. +Beyond these signs of his activity, however, which I merely +shared with all the readers of the daily press, I knew little of +my former friend and companion. + +One night--it was on the twentieth of March, 1888--I was +returning from a journey to a patient (for I had now returned to +civil practice), when my way led me through Baker Street. As I +passed the well-remembered door, which must always be associated +in my mind with my wooing, and with the dark incidents of the +Study in Scarlet, I was seized with a keen desire to see Holmes +again, and to know how he was employing his extraordinary powers. +His rooms were brilliantly lit, and, even as I looked up, I saw +his tall, spare figure pass twice in a dark silhouette against +the blind. He was pacing the room swiftly, eagerly, with his head +sunk upon his chest and his hands clasped behind him. To me, who +knew his every mood and habit, his attitude and manner told their +own story. He was at work again. He had risen out of his +drug-created dreams and was hot upon the scent of some new +problem. I rang the bell and was shown up to the chamber which +had formerly been in part my own. + +His manner was not effusive. It seldom was; but he was glad, I +think, to see me. With hardly a word spoken, but with a kindly +eye, he waved me to an armchair, threw across his case of cigars, +and indicated a spirit case and a gasogene in the corner. Then he +stood before the fire and looked me over in his singular +introspective fashion. + +"Wedlock suits you," he remarked. "I think, Watson, that you have +put on seven and a half pounds since I saw you." + +"Seven!" I answered. + +"Indeed, I should have thought a little more. Just a trifle more, +I fancy, Watson. And in practice again, I observe. You did not +tell me that you intended to go into harness." + +"Then, how do you know?" + +"I see it, I deduce it. How do I know that you have been getting +yourself very wet lately, and that you have a most clumsy and +careless servant girl?" + +"My dear Holmes," said I, "this is too much. You would certainly +have been burned, had you lived a few centuries ago. It is true +that I had a country walk on Thursday and came home in a dreadful +mess, but as I have changed my clothes I can't imagine how you +deduce it. As to Mary Jane, she is incorrigible, and my wife has +given her notice, but there, again, I fail to see how you work it +out." + +He chuckled to himself and rubbed his long, nervous hands +together. + +"It is simplicity itself," said he; "my eyes tell me that on the +inside of your left shoe, just where the firelight strikes it, +the leather is scored by six almost parallel cuts. Obviously they +have been caused by someone who has very carelessly scraped round +the edges of the sole in order to remove crusted mud from it. +Hence, you see, my double deduction that you had been out in vile +weather, and that you had a particularly malignant boot-slitting +specimen of the London slavey. As to your practice, if a +gentleman walks into my rooms smelling of iodoform, with a black +mark of nitrate of silver upon his right forefinger, and a bulge +on the right side of his top-hat to show where he has secreted +his stethoscope, I must be dull, indeed, if I do not pronounce +him to be an active member of the medical profession." + +I could not help laughing at the ease with which he explained his +process of deduction. "When I hear you give your reasons," I +remarked, "the thing always appears to me to be so ridiculously +simple that I could easily do it myself, though at each +successive instance of your reasoning I am baffled until you +explain your process. And yet I believe that my eyes are as good +as yours." + +"Quite so," he answered, lighting a cigarette, and throwing +himself down into an armchair. "You see, but you do not observe. +The distinction is clear. For example, you have frequently seen +the steps which lead up from the hall to this room." + +"Frequently." + +"How often?" + +"Well, some hundreds of times." + +"Then how many are there?" + +"How many? I don't know." + +"Quite so! You have not observed. And yet you have seen. That is +just my point. Now, I know that there are seventeen steps, +because I have both seen and observed. By-the-way, since you are +interested in these little problems, and since you are good +enough to chronicle one or two of my trifling experiences, you +may be interested in this." He threw over a sheet of thick, +pink-tinted note-paper which had been lying open upon the table. +"It came by the last post," said he. "Read it aloud." + +The note was undated, and without either signature or address. + +"There will call upon you to-night, at a quarter to eight +o'clock," it said, "a gentleman who desires to consult you upon a +matter of the very deepest moment. Your recent services to one of +the royal houses of Europe have shown that you are one who may +safely be trusted with matters which are of an importance which +can hardly be exaggerated. This account of you we have from all +quarters received. Be in your chamber then at that hour, and do +not take it amiss if your visitor wear a mask." + +"This is indeed a mystery," I remarked. "What do you imagine that +it means?" + +"I have no data yet. It is a capital mistake to theorize before +one has data. Insensibly one begins to twist facts to suit +theories, instead of theories to suit facts. But the note itself. +What do you deduce from it?" + +I carefully examined the writing, and the paper upon which it was +written. + +"The man who wrote it was presumably well to do," I remarked, +endeavouring to imitate my companion's processes. "Such paper +could not be bought under half a crown a packet. It is peculiarly +strong and stiff." + +"Peculiar--that is the very word," said Holmes. "It is not an +English paper at all. Hold it up to the light." + +I did so, and saw a large "E" with a small "g," a "P," and a +large "G" with a small "t" woven into the texture of the paper. + +"What do you make of that?" asked Holmes. + +"The name of the maker, no doubt; or his monogram, rather." + +"Not at all. The 'G' with the small 't' stands for +'Gesellschaft,' which is the German for 'Company.' It is a +customary contraction like our 'Co.' 'P,' of course, stands for +'Papier.' Now for the 'Eg.' Let us glance at our Continental +Gazetteer." He took down a heavy brown volume from his shelves. +"Eglow, Eglonitz--here we are, Egria. It is in a German-speaking +country--in Bohemia, not far from Carlsbad. 'Remarkable as being +the scene of the death of Wallenstein, and for its numerous +glass-factories and paper-mills.' Ha, ha, my boy, what do you +make of that?" His eyes sparkled, and he sent up a great blue +triumphant cloud from his cigarette. + +"The paper was made in Bohemia," I said. + +"Precisely. And the man who wrote the note is a German. Do you +note the peculiar construction of the sentence--'This account of +you we have from all quarters received.' A Frenchman or Russian +could not have written that. It is the German who is so +uncourteous to his verbs. It only remains, therefore, to discover +what is wanted by this German who writes upon Bohemian paper and +prefers wearing a mask to showing his face. And here he comes, if +I am not mistaken, to resolve all our doubts." + +As he spoke there was the sharp sound of horses' hoofs and +grating wheels against the curb, followed by a sharp pull at the +bell. Holmes whistled. + +"A pair, by the sound," said he. "Yes," he continued, glancing +out of the window. "A nice little brougham and a pair of +beauties. A hundred and fifty guineas apiece. There's money in +this case, Watson, if there is nothing else." + +"I think that I had better go, Holmes." + +"Not a bit, Doctor. Stay where you are. I am lost without my +Boswell. And this promises to be interesting. It would be a pity +to miss it." + +"But your client--" + +"Never mind him. I may want your help, and so may he. Here he +comes. Sit down in that armchair, Doctor, and give us your best +attention." + +A slow and heavy step, which had been heard upon the stairs and +in the passage, paused immediately outside the door. Then there +was a loud and authoritative tap. + +"Come in!" said Holmes. + +A man entered who could hardly have been less than six feet six +inches in height, with the chest and limbs of a Hercules. His +dress was rich with a richness which would, in England, be looked +upon as akin to bad taste. Heavy bands of astrakhan were slashed +across the sleeves and fronts of his double-breasted coat, while +the deep blue cloak which was thrown over his shoulders was lined +with flame-coloured silk and secured at the neck with a brooch +which consisted of a single flaming beryl. Boots which extended +halfway up his calves, and which were trimmed at the tops with +rich brown fur, completed the impression of barbaric opulence +which was suggested by his whole appearance. He carried a +broad-brimmed hat in his hand, while he wore across the upper +part of his face, extending down past the cheekbones, a black +vizard mask, which he had apparently adjusted that very moment, +for his hand was still raised to it as he entered. From the lower +part of the face he appeared to be a man of strong character, +with a thick, hanging lip, and a long, straight chin suggestive +of resolution pushed to the length of obstinacy. + +"You had my note?" he asked with a deep harsh voice and a +strongly marked German accent. "I told you that I would call." He +looked from one to the other of us, as if uncertain which to +address. + +"Pray take a seat," said Holmes. "This is my friend and +colleague, Dr. Watson, who is occasionally good enough to help me +in my cases. Whom have I the honour to address?" + +"You may address me as the Count Von Kramm, a Bohemian nobleman. +I understand that this gentleman, your friend, is a man of honour +and discretion, whom I may trust with a matter of the most +extreme importance. If not, I should much prefer to communicate +with you alone." + +I rose to go, but Holmes caught me by the wrist and pushed me +back into my chair. "It is both, or none," said he. "You may say +before this gentleman anything which you may say to me." + +The Count shrugged his broad shoulders. "Then I must begin," said +he, "by binding you both to absolute secrecy for two years; at +the end of that time the matter will be of no importance. At +present it is not too much to say that it is of such weight it +may have an influence upon European history." + +"I promise," said Holmes. + +"And I." + +"You will excuse this mask," continued our strange visitor. "The +august person who employs me wishes his agent to be unknown to +you, and I may confess at once that the title by which I have +just called myself is not exactly my own." + +"I was aware of it," said Holmes dryly. + +"The circumstances are of great delicacy, and every precaution +has to be taken to quench what might grow to be an immense +scandal and seriously compromise one of the reigning families of +Europe. To speak plainly, the matter implicates the great House +of Ormstein, hereditary kings of Bohemia." + +"I was also aware of that," murmured Holmes, settling himself +down in his armchair and closing his eyes. + +Our visitor glanced with some apparent surprise at the languid, +lounging figure of the man who had been no doubt depicted to him +as the most incisive reasoner and most energetic agent in Europe. +Holmes slowly reopened his eyes and looked impatiently at his +gigantic client. + +"If your Majesty would condescend to state your case," he +remarked, "I should be better able to advise you." + +The man sprang from his chair and paced up and down the room in +uncontrollable agitation. Then, with a gesture of desperation, he +tore the mask from his face and hurled it upon the ground. "You +are right," he cried; "I am the King. Why should I attempt to +conceal it?" + +"Why, indeed?" murmured Holmes. "Your Majesty had not spoken +before I was aware that I was addressing Wilhelm Gottsreich +Sigismond von Ormstein, Grand Duke of Cassel-Felstein, and +hereditary King of Bohemia." + +"But you can understand," said our strange visitor, sitting down +once more and passing his hand over his high white forehead, "you +can understand that I am not accustomed to doing such business in +my own person. Yet the matter was so delicate that I could not +confide it to an agent without putting myself in his power. I +have come incognito from Prague for the purpose of consulting +you." + +"Then, pray consult," said Holmes, shutting his eyes once more. + +"The facts are briefly these: Some five years ago, during a +lengthy visit to Warsaw, I made the acquaintance of the well-known +adventuress, Irene Adler. The name is no doubt familiar to you." + +"Kindly look her up in my index, Doctor," murmured Holmes without +opening his eyes. For many years he had adopted a system of +docketing all paragraphs concerning men and things, so that it +was difficult to name a subject or a person on which he could not +at once furnish information. In this case I found her biography +sandwiched in between that of a Hebrew rabbi and that of a +staff-commander who had written a monograph upon the deep-sea +fishes. + +"Let me see!" said Holmes. "Hum! Born in New Jersey in the year +1858. Contralto--hum! La Scala, hum! Prima donna Imperial Opera +of Warsaw--yes! Retired from operatic stage--ha! Living in +London--quite so! Your Majesty, as I understand, became entangled +with this young person, wrote her some compromising letters, and +is now desirous of getting those letters back." + +"Precisely so. But how--" + +"Was there a secret marriage?" + +"None." + +"No legal papers or certificates?" + +"None." + +"Then I fail to follow your Majesty. If this young person should +produce her letters for blackmailing or other purposes, how is +she to prove their authenticity?" + +"There is the writing." + +"Pooh, pooh! Forgery." + +"My private note-paper." + +"Stolen." + +"My own seal." + +"Imitated." + +"My photograph." + +"Bought." + +"We were both in the photograph." + +"Oh, dear! That is very bad! Your Majesty has indeed committed an +indiscretion." + +"I was mad--insane." + +"You have compromised yourself seriously." + +"I was only Crown Prince then. I was young. I am but thirty now." + +"It must be recovered." + +"We have tried and failed." + +"Your Majesty must pay. It must be bought." + +"She will not sell." + +"Stolen, then." + +"Five attempts have been made. Twice burglars in my pay ransacked +her house. Once we diverted her luggage when she travelled. Twice +she has been waylaid. There has been no result." + +"No sign of it?" + +"Absolutely none." + +Holmes laughed. "It is quite a pretty little problem," said he. + +"But a very serious one to me," returned the King reproachfully. + +"Very, indeed. And what does she propose to do with the +photograph?" + +"To ruin me." + +"But how?" + +"I am about to be married." + +"So I have heard." + +"To Clotilde Lothman von Saxe-Meningen, second daughter of the +King of Scandinavia. You may know the strict principles of her +family. She is herself the very soul of delicacy. A shadow of a +doubt as to my conduct would bring the matter to an end." + +"And Irene Adler?" + +"Threatens to send them the photograph. And she will do it. I +know that she will do it. You do not know her, but she has a soul +of steel. She has the face of the most beautiful of women, and +the mind of the most resolute of men. Rather than I should marry +another woman, there are no lengths to which she would not +go--none." + +"You are sure that she has not sent it yet?" + +"I am sure." + +"And why?" + +"Because she has said that she would send it on the day when the +betrothal was publicly proclaimed. That will be next Monday." + +"Oh, then we have three days yet," said Holmes with a yawn. "That +is very fortunate, as I have one or two matters of importance to +look into just at present. Your Majesty will, of course, stay in +London for the present?" + +"Certainly. You will find me at the Langham under the name of the +Count Von Kramm." + +"Then I shall drop you a line to let you know how we progress." + +"Pray do so. I shall be all anxiety." + +"Then, as to money?" + +"You have carte blanche." + +"Absolutely?" + +"I tell you that I would give one of the provinces of my kingdom +to have that photograph." + +"And for present expenses?" + +The King took a heavy chamois leather bag from under his cloak +and laid it on the table. + +"There are three hundred pounds in gold and seven hundred in +notes," he said. + +Holmes scribbled a receipt upon a sheet of his note-book and +handed it to him. + +"And Mademoiselle's address?" he asked. + +"Is Briony Lodge, Serpentine Avenue, St. John's Wood." + +Holmes took a note of it. "One other question," said he. "Was the +photograph a cabinet?" + +"It was." + +"Then, good-night, your Majesty, and I trust that we shall soon +have some good news for you. And good-night, Watson," he added, +as the wheels of the royal brougham rolled down the street. "If +you will be good enough to call to-morrow afternoon at three +o'clock I should like to chat this little matter over with you." + + +II. + +At three o'clock precisely I was at Baker Street, but Holmes had +not yet returned. The landlady informed me that he had left the +house shortly after eight o'clock in the morning. I sat down +beside the fire, however, with the intention of awaiting him, +however long he might be. I was already deeply interested in his +inquiry, for, though it was surrounded by none of the grim and +strange features which were associated with the two crimes which +I have already recorded, still, the nature of the case and the +exalted station of his client gave it a character of its own. +Indeed, apart from the nature of the investigation which my +friend had on hand, there was something in his masterly grasp of +a situation, and his keen, incisive reasoning, which made it a +pleasure to me to study his system of work, and to follow the +quick, subtle methods by which he disentangled the most +inextricable mysteries. So accustomed was I to his invariable +success that the very possibility of his failing had ceased to +enter into my head. + +It was close upon four before the door opened, and a +drunken-looking groom, ill-kempt and side-whiskered, with an +inflamed face and disreputable clothes, walked into the room. +Accustomed as I was to my friend's amazing powers in the use of +disguises, I had to look three times before I was certain that it +was indeed he. With a nod he vanished into the bedroom, whence he +emerged in five minutes tweed-suited and respectable, as of old. +Putting his hands into his pockets, he stretched out his legs in +front of the fire and laughed heartily for some minutes. + +"Well, really!" he cried, and then he choked and laughed again +until he was obliged to lie back, limp and helpless, in the +chair. + +"What is it?" + +"It's quite too funny. I am sure you could never guess how I +employed my morning, or what I ended by doing." + +"I can't imagine. I suppose that you have been watching the +habits, and perhaps the house, of Miss Irene Adler." + +"Quite so; but the sequel was rather unusual. I will tell you, +however. I left the house a little after eight o'clock this +morning in the character of a groom out of work. There is a +wonderful sympathy and freemasonry among horsey men. Be one of +them, and you will know all that there is to know. I soon found +Briony Lodge. It is a bijou villa, with a garden at the back, but +built out in front right up to the road, two stories. Chubb lock +to the door. Large sitting-room on the right side, well +furnished, with long windows almost to the floor, and those +preposterous English window fasteners which a child could open. +Behind there was nothing remarkable, save that the passage window +could be reached from the top of the coach-house. I walked round +it and examined it closely from every point of view, but without +noting anything else of interest. + +"I then lounged down the street and found, as I expected, that +there was a mews in a lane which runs down by one wall of the +garden. I lent the ostlers a hand in rubbing down their horses, +and received in exchange twopence, a glass of half and half, two +fills of shag tobacco, and as much information as I could desire +about Miss Adler, to say nothing of half a dozen other people in +the neighbourhood in whom I was not in the least interested, but +whose biographies I was compelled to listen to." + +"And what of Irene Adler?" I asked. + +"Oh, she has turned all the men's heads down in that part. She is +the daintiest thing under a bonnet on this planet. So say the +Serpentine-mews, to a man. She lives quietly, sings at concerts, +drives out at five every day, and returns at seven sharp for +dinner. Seldom goes out at other times, except when she sings. +Has only one male visitor, but a good deal of him. He is dark, +handsome, and dashing, never calls less than once a day, and +often twice. He is a Mr. Godfrey Norton, of the Inner Temple. See +the advantages of a cabman as a confidant. They had driven him +home a dozen times from Serpentine-mews, and knew all about him. +When I had listened to all they had to tell, I began to walk up +and down near Briony Lodge once more, and to think over my plan +of campaign. + +"This Godfrey Norton was evidently an important factor in the +matter. He was a lawyer. That sounded ominous. What was the +relation between them, and what the object of his repeated +visits? Was she his client, his friend, or his mistress? If the +former, she had probably transferred the photograph to his +keeping. If the latter, it was less likely. On the issue of this +question depended whether I should continue my work at Briony +Lodge, or turn my attention to the gentleman's chambers in the +Temple. It was a delicate point, and it widened the field of my +inquiry. I fear that I bore you with these details, but I have to +let you see my little difficulties, if you are to understand the +situation." + +"I am following you closely," I answered. + +"I was still balancing the matter in my mind when a hansom cab +drove up to Briony Lodge, and a gentleman sprang out. He was a +remarkably handsome man, dark, aquiline, and moustached--evidently +the man of whom I had heard. He appeared to be in a +great hurry, shouted to the cabman to wait, and brushed past the +maid who opened the door with the air of a man who was thoroughly +at home. + +"He was in the house about half an hour, and I could catch +glimpses of him in the windows of the sitting-room, pacing up and +down, talking excitedly, and waving his arms. Of her I could see +nothing. Presently he emerged, looking even more flurried than +before. As he stepped up to the cab, he pulled a gold watch from +his pocket and looked at it earnestly, 'Drive like the devil,' he +shouted, 'first to Gross & Hankey's in Regent Street, and then to +the Church of St. Monica in the Edgeware Road. Half a guinea if +you do it in twenty minutes!' + +"Away they went, and I was just wondering whether I should not do +well to follow them when up the lane came a neat little landau, +the coachman with his coat only half-buttoned, and his tie under +his ear, while all the tags of his harness were sticking out of +the buckles. It hadn't pulled up before she shot out of the hall +door and into it. I only caught a glimpse of her at the moment, +but she was a lovely woman, with a face that a man might die for. + +"'The Church of St. Monica, John,' she cried, 'and half a +sovereign if you reach it in twenty minutes.' + +"This was quite too good to lose, Watson. I was just balancing +whether I should run for it, or whether I should perch behind her +landau when a cab came through the street. The driver looked +twice at such a shabby fare, but I jumped in before he could +object. 'The Church of St. Monica,' said I, 'and half a sovereign +if you reach it in twenty minutes.' It was twenty-five minutes to +twelve, and of course it was clear enough what was in the wind. + +"My cabby drove fast. I don't think I ever drove faster, but the +others were there before us. The cab and the landau with their +steaming horses were in front of the door when I arrived. I paid +the man and hurried into the church. There was not a soul there +save the two whom I had followed and a surpliced clergyman, who +seemed to be expostulating with them. They were all three +standing in a knot in front of the altar. I lounged up the side +aisle like any other idler who has dropped into a church. +Suddenly, to my surprise, the three at the altar faced round to +me, and Godfrey Norton came running as hard as he could towards +me. + +"'Thank God,' he cried. 'You'll do. Come! Come!' + +"'What then?' I asked. + +"'Come, man, come, only three minutes, or it won't be legal.' + +"I was half-dragged up to the altar, and before I knew where I was +I found myself mumbling responses which were whispered in my ear, +and vouching for things of which I knew nothing, and generally +assisting in the secure tying up of Irene Adler, spinster, to +Godfrey Norton, bachelor. It was all done in an instant, and +there was the gentleman thanking me on the one side and the lady +on the other, while the clergyman beamed on me in front. It was +the most preposterous position in which I ever found myself in my +life, and it was the thought of it that started me laughing just +now. It seems that there had been some informality about their +license, that the clergyman absolutely refused to marry them +without a witness of some sort, and that my lucky appearance +saved the bridegroom from having to sally out into the streets in +search of a best man. The bride gave me a sovereign, and I mean +to wear it on my watch-chain in memory of the occasion." + +"This is a very unexpected turn of affairs," said I; "and what +then?" + +"Well, I found my plans very seriously menaced. It looked as if +the pair might take an immediate departure, and so necessitate +very prompt and energetic measures on my part. At the church +door, however, they separated, he driving back to the Temple, and +she to her own house. 'I shall drive out in the park at five as +usual,' she said as she left him. I heard no more. They drove +away in different directions, and I went off to make my own +arrangements." + +"Which are?" + +"Some cold beef and a glass of beer," he answered, ringing the +bell. "I have been too busy to think of food, and I am likely to +be busier still this evening. By the way, Doctor, I shall want +your co-operation." + +"I shall be delighted." + +"You don't mind breaking the law?" + +"Not in the least." + +"Nor running a chance of arrest?" + +"Not in a good cause." + +"Oh, the cause is excellent!" + +"Then I am your man." + +"I was sure that I might rely on you." + +"But what is it you wish?" + +"When Mrs. Turner has brought in the tray I will make it clear to +you. Now," he said as he turned hungrily on the simple fare that +our landlady had provided, "I must discuss it while I eat, for I +have not much time. It is nearly five now. In two hours we must +be on the scene of action. Miss Irene, or Madame, rather, returns +from her drive at seven. We must be at Briony Lodge to meet her." + +"And what then?" + +"You must leave that to me. I have already arranged what is to +occur. There is only one point on which I must insist. You must +not interfere, come what may. You understand?" + +"I am to be neutral?" + +"To do nothing whatever. There will probably be some small +unpleasantness. Do not join in it. It will end in my being +conveyed into the house. Four or five minutes afterwards the +sitting-room window will open. You are to station yourself close +to that open window." + +"Yes." + +"You are to watch me, for I will be visible to you." + +"Yes." + +"And when I raise my hand--so--you will throw into the room what +I give you to throw, and will, at the same time, raise the cry of +fire. You quite follow me?" + +"Entirely." + +"It is nothing very formidable," he said, taking a long cigar-shaped +roll from his pocket. "It is an ordinary plumber's smoke-rocket, +fitted with a cap at either end to make it self-lighting. +Your task is confined to that. When you raise your cry of fire, +it will be taken up by quite a number of people. You may then +walk to the end of the street, and I will rejoin you in ten +minutes. I hope that I have made myself clear?" + +"I am to remain neutral, to get near the window, to watch you, +and at the signal to throw in this object, then to raise the cry +of fire, and to wait you at the corner of the street." + +"Precisely." + +"Then you may entirely rely on me." + +"That is excellent. I think, perhaps, it is almost time that I +prepare for the new role I have to play." + +He disappeared into his bedroom and returned in a few minutes in +the character of an amiable and simple-minded Nonconformist +clergyman. His broad black hat, his baggy trousers, his white +tie, his sympathetic smile, and general look of peering and +benevolent curiosity were such as Mr. John Hare alone could have +equalled. It was not merely that Holmes changed his costume. His +expression, his manner, his very soul seemed to vary with every +fresh part that he assumed. The stage lost a fine actor, even as +science lost an acute reasoner, when he became a specialist in +crime. + +It was a quarter past six when we left Baker Street, and it still +wanted ten minutes to the hour when we found ourselves in +Serpentine Avenue. It was already dusk, and the lamps were just +being lighted as we paced up and down in front of Briony Lodge, +waiting for the coming of its occupant. The house was just such +as I had pictured it from Sherlock Holmes' succinct description, +but the locality appeared to be less private than I expected. On +the contrary, for a small street in a quiet neighbourhood, it was +remarkably animated. There was a group of shabbily dressed men +smoking and laughing in a corner, a scissors-grinder with his +wheel, two guardsmen who were flirting with a nurse-girl, and +several well-dressed young men who were lounging up and down with +cigars in their mouths. + +"You see," remarked Holmes, as we paced to and fro in front of +the house, "this marriage rather simplifies matters. The +photograph becomes a double-edged weapon now. The chances are +that she would be as averse to its being seen by Mr. Godfrey +Norton, as our client is to its coming to the eyes of his +princess. Now the question is, Where are we to find the +photograph?" + +"Where, indeed?" + +"It is most unlikely that she carries it about with her. It is +cabinet size. Too large for easy concealment about a woman's +dress. She knows that the King is capable of having her waylaid +and searched. Two attempts of the sort have already been made. We +may take it, then, that she does not carry it about with her." + +"Where, then?" + +"Her banker or her lawyer. There is that double possibility. But +I am inclined to think neither. Women are naturally secretive, +and they like to do their own secreting. Why should she hand it +over to anyone else? She could trust her own guardianship, but +she could not tell what indirect or political influence might be +brought to bear upon a business man. Besides, remember that she +had resolved to use it within a few days. It must be where she +can lay her hands upon it. It must be in her own house." + +"But it has twice been burgled." + +"Pshaw! They did not know how to look." + +"But how will you look?" + +"I will not look." + +"What then?" + +"I will get her to show me." + +"But she will refuse." + +"She will not be able to. But I hear the rumble of wheels. It is +her carriage. Now carry out my orders to the letter." + +As he spoke the gleam of the side-lights of a carriage came round +the curve of the avenue. It was a smart little landau which +rattled up to the door of Briony Lodge. As it pulled up, one of +the loafing men at the corner dashed forward to open the door in +the hope of earning a copper, but was elbowed away by another +loafer, who had rushed up with the same intention. A fierce +quarrel broke out, which was increased by the two guardsmen, who +took sides with one of the loungers, and by the scissors-grinder, +who was equally hot upon the other side. A blow was struck, and +in an instant the lady, who had stepped from her carriage, was +the centre of a little knot of flushed and struggling men, who +struck savagely at each other with their fists and sticks. Holmes +dashed into the crowd to protect the lady; but just as he reached +her he gave a cry and dropped to the ground, with the blood +running freely down his face. At his fall the guardsmen took to +their heels in one direction and the loungers in the other, while +a number of better-dressed people, who had watched the scuffle +without taking part in it, crowded in to help the lady and to +attend to the injured man. Irene Adler, as I will still call her, +had hurried up the steps; but she stood at the top with her +superb figure outlined against the lights of the hall, looking +back into the street. + +"Is the poor gentleman much hurt?" she asked. + +"He is dead," cried several voices. + +"No, no, there's life in him!" shouted another. "But he'll be +gone before you can get him to hospital." + +"He's a brave fellow," said a woman. "They would have had the +lady's purse and watch if it hadn't been for him. They were a +gang, and a rough one, too. Ah, he's breathing now." + +"He can't lie in the street. May we bring him in, marm?" + +"Surely. Bring him into the sitting-room. There is a comfortable +sofa. This way, please!" + +Slowly and solemnly he was borne into Briony Lodge and laid out +in the principal room, while I still observed the proceedings +from my post by the window. The lamps had been lit, but the +blinds had not been drawn, so that I could see Holmes as he lay +upon the couch. I do not know whether he was seized with +compunction at that moment for the part he was playing, but I +know that I never felt more heartily ashamed of myself in my life +than when I saw the beautiful creature against whom I was +conspiring, or the grace and kindliness with which she waited +upon the injured man. And yet it would be the blackest treachery +to Holmes to draw back now from the part which he had intrusted +to me. I hardened my heart, and took the smoke-rocket from under +my ulster. After all, I thought, we are not injuring her. We are +but preventing her from injuring another. + +Holmes had sat up upon the couch, and I saw him motion like a man +who is in need of air. A maid rushed across and threw open the +window. At the same instant I saw him raise his hand and at the +signal I tossed my rocket into the room with a cry of "Fire!" The +word was no sooner out of my mouth than the whole crowd of +spectators, well dressed and ill--gentlemen, ostlers, and +servant-maids--joined in a general shriek of "Fire!" Thick clouds +of smoke curled through the room and out at the open window. I +caught a glimpse of rushing figures, and a moment later the voice +of Holmes from within assuring them that it was a false alarm. +Slipping through the shouting crowd I made my way to the corner +of the street, and in ten minutes was rejoiced to find my +friend's arm in mine, and to get away from the scene of uproar. +He walked swiftly and in silence for some few minutes until we +had turned down one of the quiet streets which lead towards the +Edgeware Road. + +"You did it very nicely, Doctor," he remarked. "Nothing could +have been better. It is all right." + +"You have the photograph?" + +"I know where it is." + +"And how did you find out?" + +"She showed me, as I told you she would." + +"I am still in the dark." + +"I do not wish to make a mystery," said he, laughing. "The matter +was perfectly simple. You, of course, saw that everyone in the +street was an accomplice. They were all engaged for the evening." + +"I guessed as much." + +"Then, when the row broke out, I had a little moist red paint in +the palm of my hand. I rushed forward, fell down, clapped my hand +to my face, and became a piteous spectacle. It is an old trick." + +"That also I could fathom." + +"Then they carried me in. She was bound to have me in. What else +could she do? And into her sitting-room, which was the very room +which I suspected. It lay between that and her bedroom, and I was +determined to see which. They laid me on a couch, I motioned for +air, they were compelled to open the window, and you had your +chance." + +"How did that help you?" + +"It was all-important. When a woman thinks that her house is on +fire, her instinct is at once to rush to the thing which she +values most. It is a perfectly overpowering impulse, and I have +more than once taken advantage of it. In the case of the +Darlington substitution scandal it was of use to me, and also in +the Arnsworth Castle business. A married woman grabs at her baby; +an unmarried one reaches for her jewel-box. Now it was clear to +me that our lady of to-day had nothing in the house more precious +to her than what we are in quest of. She would rush to secure it. +The alarm of fire was admirably done. The smoke and shouting were +enough to shake nerves of steel. She responded beautifully. The +photograph is in a recess behind a sliding panel just above the +right bell-pull. She was there in an instant, and I caught a +glimpse of it as she half-drew it out. When I cried out that it +was a false alarm, she replaced it, glanced at the rocket, rushed +from the room, and I have not seen her since. I rose, and, making +my excuses, escaped from the house. I hesitated whether to +attempt to secure the photograph at once; but the coachman had +come in, and as he was watching me narrowly it seemed safer to +wait. A little over-precipitance may ruin all." + +"And now?" I asked. + +"Our quest is practically finished. I shall call with the King +to-morrow, and with you, if you care to come with us. We will be +shown into the sitting-room to wait for the lady, but it is +probable that when she comes she may find neither us nor the +photograph. It might be a satisfaction to his Majesty to regain +it with his own hands." + +"And when will you call?" + +"At eight in the morning. She will not be up, so that we shall +have a clear field. Besides, we must be prompt, for this marriage +may mean a complete change in her life and habits. I must wire to +the King without delay." + +We had reached Baker Street and had stopped at the door. He was +searching his pockets for the key when someone passing said: + +"Good-night, Mister Sherlock Holmes." + +There were several people on the pavement at the time, but the +greeting appeared to come from a slim youth in an ulster who had +hurried by. + +"I've heard that voice before," said Holmes, staring down the +dimly lit street. "Now, I wonder who the deuce that could have +been." + + +III. + +I slept at Baker Street that night, and we were engaged upon our +toast and coffee in the morning when the King of Bohemia rushed +into the room. + +"You have really got it!" he cried, grasping Sherlock Holmes by +either shoulder and looking eagerly into his face. + +"Not yet." + +"But you have hopes?" + +"I have hopes." + +"Then, come. I am all impatience to be gone." + +"We must have a cab." + +"No, my brougham is waiting." + +"Then that will simplify matters." We descended and started off +once more for Briony Lodge. + +"Irene Adler is married," remarked Holmes. + +"Married! When?" + +"Yesterday." + +"But to whom?" + +"To an English lawyer named Norton." + +"But she could not love him." + +"I am in hopes that she does." + +"And why in hopes?" + +"Because it would spare your Majesty all fear of future +annoyance. If the lady loves her husband, she does not love your +Majesty. If she does not love your Majesty, there is no reason +why she should interfere with your Majesty's plan." + +"It is true. And yet--Well! I wish she had been of my own +station! What a queen she would have made!" He relapsed into a +moody silence, which was not broken until we drew up in +Serpentine Avenue. + +The door of Briony Lodge was open, and an elderly woman stood +upon the steps. She watched us with a sardonic eye as we stepped +from the brougham. + +"Mr. Sherlock Holmes, I believe?" said she. + +"I am Mr. Holmes," answered my companion, looking at her with a +questioning and rather startled gaze. + +"Indeed! My mistress told me that you were likely to call. She +left this morning with her husband by the 5:15 train from Charing +Cross for the Continent." + +"What!" Sherlock Holmes staggered back, white with chagrin and +surprise. "Do you mean that she has left England?" + +"Never to return." + +"And the papers?" asked the King hoarsely. "All is lost." + +"We shall see." He pushed past the servant and rushed into the +drawing-room, followed by the King and myself. The furniture was +scattered about in every direction, with dismantled shelves and +open drawers, as if the lady had hurriedly ransacked them before +her flight. Holmes rushed at the bell-pull, tore back a small +sliding shutter, and, plunging in his hand, pulled out a +photograph and a letter. The photograph was of Irene Adler +herself in evening dress, the letter was superscribed to +"Sherlock Holmes, Esq. To be left till called for." My friend +tore it open and we all three read it together. It was dated at +midnight of the preceding night and ran in this way: + +"MY DEAR MR. SHERLOCK HOLMES,--You really did it very well. You +took me in completely. Until after the alarm of fire, I had not a +suspicion. But then, when I found how I had betrayed myself, I +began to think. I had been warned against you months ago. I had +been told that if the King employed an agent it would certainly +be you. And your address had been given me. Yet, with all this, +you made me reveal what you wanted to know. Even after I became +suspicious, I found it hard to think evil of such a dear, kind +old clergyman. But, you know, I have been trained as an actress +myself. Male costume is nothing new to me. I often take advantage +of the freedom which it gives. I sent John, the coachman, to +watch you, ran up stairs, got into my walking-clothes, as I call +them, and came down just as you departed. + +"Well, I followed you to your door, and so made sure that I was +really an object of interest to the celebrated Mr. Sherlock +Holmes. Then I, rather imprudently, wished you good-night, and +started for the Temple to see my husband. + +"We both thought the best resource was flight, when pursued by +so formidable an antagonist; so you will find the nest empty when +you call to-morrow. As to the photograph, your client may rest in +peace. I love and am loved by a better man than he. The King may +do what he will without hindrance from one whom he has cruelly +wronged. I keep it only to safeguard myself, and to preserve a +weapon which will always secure me from any steps which he might +take in the future. I leave a photograph which he might care to +possess; and I remain, dear Mr. Sherlock Holmes, + + "Very truly yours, + "IRENE NORTON, née ADLER." + +"What a woman--oh, what a woman!" cried the King of Bohemia, when +we had all three read this epistle. "Did I not tell you how quick +and resolute she was? Would she not have made an admirable queen? +Is it not a pity that she was not on my level?" + +"From what I have seen of the lady she seems indeed to be on a +very different level to your Majesty," said Holmes coldly. "I am +sorry that I have not been able to bring your Majesty's business +to a more successful conclusion." + +"On the contrary, my dear sir," cried the King; "nothing could be +more successful. I know that her word is inviolate. The +photograph is now as safe as if it were in the fire." + +"I am glad to hear your Majesty say so." + +"I am immensely indebted to you. Pray tell me in what way I can +reward you. This ring--" He slipped an emerald snake ring from +his finger and held it out upon the palm of his hand. + +"Your Majesty has something which I should value even more +highly," said Holmes. + +"You have but to name it." + +"This photograph!" + +The King stared at him in amazement. + +"Irene's photograph!" he cried. "Certainly, if you wish it." + +"I thank your Majesty. Then there is no more to be done in the +matter. I have the honour to wish you a very good-morning." He +bowed, and, turning away without observing the hand which the +King had stretched out to him, he set off in my company for his +chambers. + +And that was how a great scandal threatened to affect the kingdom +of Bohemia, and how the best plans of Mr. Sherlock Holmes were +beaten by a woman's wit. He used to make merry over the +cleverness of women, but I have not heard him do it of late. And +when he speaks of Irene Adler, or when he refers to her +photograph, it is always under the honourable title of the woman. + + + +ADVENTURE II. THE RED-HEADED LEAGUE + +I had called upon my friend, Mr. Sherlock Holmes, one day in the +autumn of last year and found him in deep conversation with a +very stout, florid-faced, elderly gentleman with fiery red hair. +With an apology for my intrusion, I was about to withdraw when +Holmes pulled me abruptly into the room and closed the door +behind me. + +"You could not possibly have come at a better time, my dear +Watson," he said cordially. + +"I was afraid that you were engaged." + +"So I am. Very much so." + +"Then I can wait in the next room." + +"Not at all. This gentleman, Mr. Wilson, has been my partner and +helper in many of my most successful cases, and I have no +doubt that he will be of the utmost use to me in yours also." + +The stout gentleman half rose from his chair and gave a bob of +greeting, with a quick little questioning glance from his small +fat-encircled eyes. + +"Try the settee," said Holmes, relapsing into his armchair and +putting his fingertips together, as was his custom when in +judicial moods. "I know, my dear Watson, that you share my love +of all that is bizarre and outside the conventions and humdrum +routine of everyday life. You have shown your relish for it by +the enthusiasm which has prompted you to chronicle, and, if you +will excuse my saying so, somewhat to embellish so many of my own +little adventures." + +"Your cases have indeed been of the greatest interest to me," I +observed. + +"You will remember that I remarked the other day, just before we +went into the very simple problem presented by Miss Mary +Sutherland, that for strange effects and extraordinary +combinations we must go to life itself, which is always far more +daring than any effort of the imagination." + +"A proposition which I took the liberty of doubting." + +"You did, Doctor, but none the less you must come round to my +view, for otherwise I shall keep on piling fact upon fact on you +until your reason breaks down under them and acknowledges me to +be right. Now, Mr. Jabez Wilson here has been good enough to call +upon me this morning, and to begin a narrative which promises to +be one of the most singular which I have listened to for some +time. You have heard me remark that the strangest and most unique +things are very often connected not with the larger but with the +smaller crimes, and occasionally, indeed, where there is room for +doubt whether any positive crime has been committed. As far as I +have heard it is impossible for me to say whether the present +case is an instance of crime or not, but the course of events is +certainly among the most singular that I have ever listened to. +Perhaps, Mr. Wilson, you would have the great kindness to +recommence your narrative. I ask you not merely because my friend +Dr. Watson has not heard the opening part but also because the +peculiar nature of the story makes me anxious to have every +possible detail from your lips. As a rule, when I have heard some +slight indication of the course of events, I am able to guide +myself by the thousands of other similar cases which occur to my +memory. In the present instance I am forced to admit that the +facts are, to the best of my belief, unique." + +The portly client puffed out his chest with an appearance of some +little pride and pulled a dirty and wrinkled newspaper from the +inside pocket of his greatcoat. As he glanced down the +advertisement column, with his head thrust forward and the paper +flattened out upon his knee, I took a good look at the man and +endeavoured, after the fashion of my companion, to read the +indications which might be presented by his dress or appearance. + +I did not gain very much, however, by my inspection. Our visitor +bore every mark of being an average commonplace British +tradesman, obese, pompous, and slow. He wore rather baggy grey +shepherd's check trousers, a not over-clean black frock-coat, +unbuttoned in the front, and a drab waistcoat with a heavy brassy +Albert chain, and a square pierced bit of metal dangling down as +an ornament. A frayed top-hat and a faded brown overcoat with a +wrinkled velvet collar lay upon a chair beside him. Altogether, +look as I would, there was nothing remarkable about the man save +his blazing red head, and the expression of extreme chagrin and +discontent upon his features. + +Sherlock Holmes' quick eye took in my occupation, and he shook +his head with a smile as he noticed my questioning glances. +"Beyond the obvious facts that he has at some time done manual +labour, that he takes snuff, that he is a Freemason, that he has +been in China, and that he has done a considerable amount of +writing lately, I can deduce nothing else." + +Mr. Jabez Wilson started up in his chair, with his forefinger +upon the paper, but his eyes upon my companion. + +"How, in the name of good-fortune, did you know all that, Mr. +Holmes?" he asked. "How did you know, for example, that I did +manual labour. It's as true as gospel, for I began as a ship's +carpenter." + +"Your hands, my dear sir. Your right hand is quite a size larger +than your left. You have worked with it, and the muscles are more +developed." + +"Well, the snuff, then, and the Freemasonry?" + +"I won't insult your intelligence by telling you how I read that, +especially as, rather against the strict rules of your order, you +use an arc-and-compass breastpin." + +"Ah, of course, I forgot that. But the writing?" + +"What else can be indicated by that right cuff so very shiny for +five inches, and the left one with the smooth patch near the +elbow where you rest it upon the desk?" + +"Well, but China?" + +"The fish that you have tattooed immediately above your right +wrist could only have been done in China. I have made a small +study of tattoo marks and have even contributed to the literature +of the subject. That trick of staining the fishes' scales of a +delicate pink is quite peculiar to China. When, in addition, I +see a Chinese coin hanging from your watch-chain, the matter +becomes even more simple." + +Mr. Jabez Wilson laughed heavily. "Well, I never!" said he. "I +thought at first that you had done something clever, but I see +that there was nothing in it, after all." + +"I begin to think, Watson," said Holmes, "that I make a mistake +in explaining. 'Omne ignotum pro magnifico,' you know, and my +poor little reputation, such as it is, will suffer shipwreck if I +am so candid. Can you not find the advertisement, Mr. Wilson?" + +"Yes, I have got it now," he answered with his thick red finger +planted halfway down the column. "Here it is. This is what began +it all. You just read it for yourself, sir." + +I took the paper from him and read as follows: + +"TO THE RED-HEADED LEAGUE: On account of the bequest of the late +Ezekiah Hopkins, of Lebanon, Pennsylvania, U. S. A., there is now +another vacancy open which entitles a member of the League to a +salary of 4 pounds a week for purely nominal services. All +red-headed men who are sound in body and mind and above the age +of twenty-one years, are eligible. Apply in person on Monday, at +eleven o'clock, to Duncan Ross, at the offices of the League, 7 +Pope's Court, Fleet Street." + +"What on earth does this mean?" I ejaculated after I had twice +read over the extraordinary announcement. + +Holmes chuckled and wriggled in his chair, as was his habit when +in high spirits. "It is a little off the beaten track, isn't it?" +said he. "And now, Mr. Wilson, off you go at scratch and tell us +all about yourself, your household, and the effect which this +advertisement had upon your fortunes. You will first make a note, +Doctor, of the paper and the date." + +"It is The Morning Chronicle of April 27, 1890. Just two months +ago." + +"Very good. Now, Mr. Wilson?" + +"Well, it is just as I have been telling you, Mr. Sherlock +Holmes," said Jabez Wilson, mopping his forehead; "I have a small +pawnbroker's business at Coburg Square, near the City. It's not a +very large affair, and of late years it has not done more than +just give me a living. I used to be able to keep two assistants, +but now I only keep one; and I would have a job to pay him but +that he is willing to come for half wages so as to learn the +business." + +"What is the name of this obliging youth?" asked Sherlock Holmes. + +"His name is Vincent Spaulding, and he's not such a youth, +either. It's hard to say his age. I should not wish a smarter +assistant, Mr. Holmes; and I know very well that he could better +himself and earn twice what I am able to give him. But, after +all, if he is satisfied, why should I put ideas in his head?" + +"Why, indeed? You seem most fortunate in having an employé who +comes under the full market price. It is not a common experience +among employers in this age. I don't know that your assistant is +not as remarkable as your advertisement." + +"Oh, he has his faults, too," said Mr. Wilson. "Never was such a +fellow for photography. Snapping away with a camera when he ought +to be improving his mind, and then diving down into the cellar +like a rabbit into its hole to develop his pictures. That is his +main fault, but on the whole he's a good worker. There's no vice +in him." + +"He is still with you, I presume?" + +"Yes, sir. He and a girl of fourteen, who does a bit of simple +cooking and keeps the place clean--that's all I have in the +house, for I am a widower and never had any family. We live very +quietly, sir, the three of us; and we keep a roof over our heads +and pay our debts, if we do nothing more. + +"The first thing that put us out was that advertisement. +Spaulding, he came down into the office just this day eight +weeks, with this very paper in his hand, and he says: + +"'I wish to the Lord, Mr. Wilson, that I was a red-headed man.' + +"'Why that?' I asks. + +"'Why,' says he, 'here's another vacancy on the League of the +Red-headed Men. It's worth quite a little fortune to any man who +gets it, and I understand that there are more vacancies than +there are men, so that the trustees are at their wits' end what +to do with the money. If my hair would only change colour, here's +a nice little crib all ready for me to step into.' + +"'Why, what is it, then?' I asked. You see, Mr. Holmes, I am a +very stay-at-home man, and as my business came to me instead of +my having to go to it, I was often weeks on end without putting +my foot over the door-mat. In that way I didn't know much of what +was going on outside, and I was always glad of a bit of news. + +"'Have you never heard of the League of the Red-headed Men?' he +asked with his eyes open. + +"'Never.' + +"'Why, I wonder at that, for you are eligible yourself for one +of the vacancies.' + +"'And what are they worth?' I asked. + +"'Oh, merely a couple of hundred a year, but the work is slight, +and it need not interfere very much with one's other +occupations.' + +"Well, you can easily think that that made me prick up my ears, +for the business has not been over-good for some years, and an +extra couple of hundred would have been very handy. + +"'Tell me all about it,' said I. + +"'Well,' said he, showing me the advertisement, 'you can see for +yourself that the League has a vacancy, and there is the address +where you should apply for particulars. As far as I can make out, +the League was founded by an American millionaire, Ezekiah +Hopkins, who was very peculiar in his ways. He was himself +red-headed, and he had a great sympathy for all red-headed men; +so when he died it was found that he had left his enormous +fortune in the hands of trustees, with instructions to apply the +interest to the providing of easy berths to men whose hair is of +that colour. From all I hear it is splendid pay and very little to +do.' + +"'But,' said I, 'there would be millions of red-headed men who +would apply.' + +"'Not so many as you might think,' he answered. 'You see it is +really confined to Londoners, and to grown men. This American had +started from London when he was young, and he wanted to do the +old town a good turn. Then, again, I have heard it is no use your +applying if your hair is light red, or dark red, or anything but +real bright, blazing, fiery red. Now, if you cared to apply, Mr. +Wilson, you would just walk in; but perhaps it would hardly be +worth your while to put yourself out of the way for the sake of a +few hundred pounds.' + +"Now, it is a fact, gentlemen, as you may see for yourselves, +that my hair is of a very full and rich tint, so that it seemed +to me that if there was to be any competition in the matter I +stood as good a chance as any man that I had ever met. Vincent +Spaulding seemed to know so much about it that I thought he might +prove useful, so I just ordered him to put up the shutters for +the day and to come right away with me. He was very willing to +have a holiday, so we shut the business up and started off for +the address that was given us in the advertisement. + +"I never hope to see such a sight as that again, Mr. Holmes. From +north, south, east, and west every man who had a shade of red in +his hair had tramped into the city to answer the advertisement. +Fleet Street was choked with red-headed folk, and Pope's Court +looked like a coster's orange barrow. I should not have thought +there were so many in the whole country as were brought together +by that single advertisement. Every shade of colour they +were--straw, lemon, orange, brick, Irish-setter, liver, clay; +but, as Spaulding said, there were not many who had the real +vivid flame-coloured tint. When I saw how many were waiting, I +would have given it up in despair; but Spaulding would not hear +of it. How he did it I could not imagine, but he pushed and +pulled and butted until he got me through the crowd, and right up +to the steps which led to the office. There was a double stream +upon the stair, some going up in hope, and some coming back +dejected; but we wedged in as well as we could and soon found +ourselves in the office." + +"Your experience has been a most entertaining one," remarked +Holmes as his client paused and refreshed his memory with a huge +pinch of snuff. "Pray continue your very interesting statement." + +"There was nothing in the office but a couple of wooden chairs +and a deal table, behind which sat a small man with a head that +was even redder than mine. He said a few words to each candidate +as he came up, and then he always managed to find some fault in +them which would disqualify them. Getting a vacancy did not seem +to be such a very easy matter, after all. However, when our turn +came the little man was much more favourable to me than to any of +the others, and he closed the door as we entered, so that he +might have a private word with us. + +"'This is Mr. Jabez Wilson,' said my assistant, 'and he is +willing to fill a vacancy in the League.' + +"'And he is admirably suited for it,' the other answered. 'He has +every requirement. I cannot recall when I have seen anything so +fine.' He took a step backward, cocked his head on one side, and +gazed at my hair until I felt quite bashful. Then suddenly he +plunged forward, wrung my hand, and congratulated me warmly on my +success. + +"'It would be injustice to hesitate,' said he. 'You will, +however, I am sure, excuse me for taking an obvious precaution.' +With that he seized my hair in both his hands, and tugged until I +yelled with the pain. 'There is water in your eyes,' said he as +he released me. 'I perceive that all is as it should be. But we +have to be careful, for we have twice been deceived by wigs and +once by paint. I could tell you tales of cobbler's wax which +would disgust you with human nature.' He stepped over to the +window and shouted through it at the top of his voice that the +vacancy was filled. A groan of disappointment came up from below, +and the folk all trooped away in different directions until there +was not a red-head to be seen except my own and that of the +manager. + +"'My name,' said he, 'is Mr. Duncan Ross, and I am myself one of +the pensioners upon the fund left by our noble benefactor. Are +you a married man, Mr. Wilson? Have you a family?' + +"I answered that I had not. + +"His face fell immediately. + +"'Dear me!' he said gravely, 'that is very serious indeed! I am +sorry to hear you say that. The fund was, of course, for the +propagation and spread of the red-heads as well as for their +maintenance. It is exceedingly unfortunate that you should be a +bachelor.' + +"My face lengthened at this, Mr. Holmes, for I thought that I was +not to have the vacancy after all; but after thinking it over for +a few minutes he said that it would be all right. + +"'In the case of another,' said he, 'the objection might be +fatal, but we must stretch a point in favour of a man with such a +head of hair as yours. When shall you be able to enter upon your +new duties?' + +"'Well, it is a little awkward, for I have a business already,' +said I. + +"'Oh, never mind about that, Mr. Wilson!' said Vincent Spaulding. +'I should be able to look after that for you.' + +"'What would be the hours?' I asked. + +"'Ten to two.' + +"Now a pawnbroker's business is mostly done of an evening, Mr. +Holmes, especially Thursday and Friday evening, which is just +before pay-day; so it would suit me very well to earn a little in +the mornings. Besides, I knew that my assistant was a good man, +and that he would see to anything that turned up. + +"'That would suit me very well,' said I. 'And the pay?' + +"'Is 4 pounds a week.' + +"'And the work?' + +"'Is purely nominal.' + +"'What do you call purely nominal?' + +"'Well, you have to be in the office, or at least in the +building, the whole time. If you leave, you forfeit your whole +position forever. The will is very clear upon that point. You +don't comply with the conditions if you budge from the office +during that time.' + +"'It's only four hours a day, and I should not think of leaving,' +said I. + +"'No excuse will avail,' said Mr. Duncan Ross; 'neither sickness +nor business nor anything else. There you must stay, or you lose +your billet.' + +"'And the work?' + +"'Is to copy out the "Encyclopaedia Britannica." There is the first +volume of it in that press. You must find your own ink, pens, and +blotting-paper, but we provide this table and chair. Will you be +ready to-morrow?' + +"'Certainly,' I answered. + +"'Then, good-bye, Mr. Jabez Wilson, and let me congratulate you +once more on the important position which you have been fortunate +enough to gain.' He bowed me out of the room and I went home with +my assistant, hardly knowing what to say or do, I was so pleased +at my own good fortune. + +"Well, I thought over the matter all day, and by evening I was in +low spirits again; for I had quite persuaded myself that the +whole affair must be some great hoax or fraud, though what its +object might be I could not imagine. It seemed altogether past +belief that anyone could make such a will, or that they would pay +such a sum for doing anything so simple as copying out the +'Encyclopaedia Britannica.' Vincent Spaulding did what he could to +cheer me up, but by bedtime I had reasoned myself out of the +whole thing. However, in the morning I determined to have a look +at it anyhow, so I bought a penny bottle of ink, and with a +quill-pen, and seven sheets of foolscap paper, I started off for +Pope's Court. + +"Well, to my surprise and delight, everything was as right as +possible. The table was set out ready for me, and Mr. Duncan Ross +was there to see that I got fairly to work. He started me off +upon the letter A, and then he left me; but he would drop in from +time to time to see that all was right with me. At two o'clock he +bade me good-day, complimented me upon the amount that I had +written, and locked the door of the office after me. + +"This went on day after day, Mr. Holmes, and on Saturday the +manager came in and planked down four golden sovereigns for my +week's work. It was the same next week, and the same the week +after. Every morning I was there at ten, and every afternoon I +left at two. By degrees Mr. Duncan Ross took to coming in only +once of a morning, and then, after a time, he did not come in at +all. Still, of course, I never dared to leave the room for an +instant, for I was not sure when he might come, and the billet +was such a good one, and suited me so well, that I would not risk +the loss of it. + +"Eight weeks passed away like this, and I had written about +Abbots and Archery and Armour and Architecture and Attica, and +hoped with diligence that I might get on to the B's before very +long. It cost me something in foolscap, and I had pretty nearly +filled a shelf with my writings. And then suddenly the whole +business came to an end." + +"To an end?" + +"Yes, sir. And no later than this morning. I went to my work as +usual at ten o'clock, but the door was shut and locked, with a +little square of cardboard hammered on to the middle of the +panel with a tack. Here it is, and you can read for yourself." + +He held up a piece of white cardboard about the size of a sheet +of note-paper. It read in this fashion: + + THE RED-HEADED LEAGUE + + IS + + DISSOLVED. + + October 9, 1890. + +Sherlock Holmes and I surveyed this curt announcement and the +rueful face behind it, until the comical side of the affair so +completely overtopped every other consideration that we both +burst out into a roar of laughter. + +"I cannot see that there is anything very funny," cried our +client, flushing up to the roots of his flaming head. "If you can +do nothing better than laugh at me, I can go elsewhere." + +"No, no," cried Holmes, shoving him back into the chair from +which he had half risen. "I really wouldn't miss your case for +the world. It is most refreshingly unusual. But there is, if you +will excuse my saying so, something just a little funny about it. +Pray what steps did you take when you found the card upon the +door?" + +"I was staggered, sir. I did not know what to do. Then I called +at the offices round, but none of them seemed to know anything +about it. Finally, I went to the landlord, who is an accountant +living on the ground-floor, and I asked him if he could tell me +what had become of the Red-headed League. He said that he had +never heard of any such body. Then I asked him who Mr. Duncan +Ross was. He answered that the name was new to him. + +"'Well,' said I, 'the gentleman at No. 4.' + +"'What, the red-headed man?' + +"'Yes.' + +"'Oh,' said he, 'his name was William Morris. He was a solicitor +and was using my room as a temporary convenience until his new +premises were ready. He moved out yesterday.' + +"'Where could I find him?' + +"'Oh, at his new offices. He did tell me the address. Yes, 17 +King Edward Street, near St. Paul's.' + +"I started off, Mr. Holmes, but when I got to that address it was +a manufactory of artificial knee-caps, and no one in it had ever +heard of either Mr. William Morris or Mr. Duncan Ross." + +"And what did you do then?" asked Holmes. + +"I went home to Saxe-Coburg Square, and I took the advice of my +assistant. But he could not help me in any way. He could only say +that if I waited I should hear by post. But that was not quite +good enough, Mr. Holmes. I did not wish to lose such a place +without a struggle, so, as I had heard that you were good enough +to give advice to poor folk who were in need of it, I came right +away to you." + +"And you did very wisely," said Holmes. "Your case is an +exceedingly remarkable one, and I shall be happy to look into it. +From what you have told me I think that it is possible that +graver issues hang from it than might at first sight appear." + +"Grave enough!" said Mr. Jabez Wilson. "Why, I have lost four +pound a week." + +"As far as you are personally concerned," remarked Holmes, "I do +not see that you have any grievance against this extraordinary +league. On the contrary, you are, as I understand, richer by some +30 pounds, to say nothing of the minute knowledge which you have +gained on every subject which comes under the letter A. You have +lost nothing by them." + +"No, sir. But I want to find out about them, and who they are, +and what their object was in playing this prank--if it was a +prank--upon me. It was a pretty expensive joke for them, for it +cost them two and thirty pounds." + +"We shall endeavour to clear up these points for you. And, first, +one or two questions, Mr. Wilson. This assistant of yours who +first called your attention to the advertisement--how long had he +been with you?" + +"About a month then." + +"How did he come?" + +"In answer to an advertisement." + +"Was he the only applicant?" + +"No, I had a dozen." + +"Why did you pick him?" + +"Because he was handy and would come cheap." + +"At half-wages, in fact." + +"Yes." + +"What is he like, this Vincent Spaulding?" + +"Small, stout-built, very quick in his ways, no hair on his face, +though he's not short of thirty. Has a white splash of acid upon +his forehead." + +Holmes sat up in his chair in considerable excitement. "I thought +as much," said he. "Have you ever observed that his ears are +pierced for earrings?" + +"Yes, sir. He told me that a gipsy had done it for him when he +was a lad." + +"Hum!" said Holmes, sinking back in deep thought. "He is still +with you?" + +"Oh, yes, sir; I have only just left him." + +"And has your business been attended to in your absence?" + +"Nothing to complain of, sir. There's never very much to do of a +morning." + +"That will do, Mr. Wilson. I shall be happy to give you an +opinion upon the subject in the course of a day or two. To-day is +Saturday, and I hope that by Monday we may come to a conclusion." + +"Well, Watson," said Holmes when our visitor had left us, "what +do you make of it all?" + +"I make nothing of it," I answered frankly. "It is a most +mysterious business." + +"As a rule," said Holmes, "the more bizarre a thing is the less +mysterious it proves to be. It is your commonplace, featureless +crimes which are really puzzling, just as a commonplace face is +the most difficult to identify. But I must be prompt over this +matter." + +"What are you going to do, then?" I asked. + +"To smoke," he answered. "It is quite a three pipe problem, and I +beg that you won't speak to me for fifty minutes." He curled +himself up in his chair, with his thin knees drawn up to his +hawk-like nose, and there he sat with his eyes closed and his +black clay pipe thrusting out like the bill of some strange bird. +I had come to the conclusion that he had dropped asleep, and +indeed was nodding myself, when he suddenly sprang out of his +chair with the gesture of a man who has made up his mind and put +his pipe down upon the mantelpiece. + +"Sarasate plays at the St. James's Hall this afternoon," he +remarked. "What do you think, Watson? Could your patients spare +you for a few hours?" + +"I have nothing to do to-day. My practice is never very +absorbing." + +"Then put on your hat and come. I am going through the City +first, and we can have some lunch on the way. I observe that +there is a good deal of German music on the programme, which is +rather more to my taste than Italian or French. It is +introspective, and I want to introspect. Come along!" + +We travelled by the Underground as far as Aldersgate; and a short +walk took us to Saxe-Coburg Square, the scene of the singular +story which we had listened to in the morning. It was a poky, +little, shabby-genteel place, where four lines of dingy +two-storied brick houses looked out into a small railed-in +enclosure, where a lawn of weedy grass and a few clumps of faded +laurel-bushes made a hard fight against a smoke-laden and +uncongenial atmosphere. Three gilt balls and a brown board with +"JABEZ WILSON" in white letters, upon a corner house, announced +the place where our red-headed client carried on his business. +Sherlock Holmes stopped in front of it with his head on one side +and looked it all over, with his eyes shining brightly between +puckered lids. Then he walked slowly up the street, and then down +again to the corner, still looking keenly at the houses. Finally +he returned to the pawnbroker's, and, having thumped vigorously +upon the pavement with his stick two or three times, he went up +to the door and knocked. It was instantly opened by a +bright-looking, clean-shaven young fellow, who asked him to step +in. + +"Thank you," said Holmes, "I only wished to ask you how you would +go from here to the Strand." + +"Third right, fourth left," answered the assistant promptly, +closing the door. + +"Smart fellow, that," observed Holmes as we walked away. "He is, +in my judgment, the fourth smartest man in London, and for daring +I am not sure that he has not a claim to be third. I have known +something of him before." + +"Evidently," said I, "Mr. Wilson's assistant counts for a good +deal in this mystery of the Red-headed League. I am sure that you +inquired your way merely in order that you might see him." + +"Not him." + +"What then?" + +"The knees of his trousers." + +"And what did you see?" + +"What I expected to see." + +"Why did you beat the pavement?" + +"My dear doctor, this is a time for observation, not for talk. We +are spies in an enemy's country. We know something of Saxe-Coburg +Square. Let us now explore the parts which lie behind it." + +The road in which we found ourselves as we turned round the +corner from the retired Saxe-Coburg Square presented as great a +contrast to it as the front of a picture does to the back. It was +one of the main arteries which conveyed the traffic of the City +to the north and west. The roadway was blocked with the immense +stream of commerce flowing in a double tide inward and outward, +while the footpaths were black with the hurrying swarm of +pedestrians. It was difficult to realise as we looked at the line +of fine shops and stately business premises that they really +abutted on the other side upon the faded and stagnant square +which we had just quitted. + +"Let me see," said Holmes, standing at the corner and glancing +along the line, "I should like just to remember the order of the +houses here. It is a hobby of mine to have an exact knowledge of +London. There is Mortimer's, the tobacconist, the little +newspaper shop, the Coburg branch of the City and Suburban Bank, +the Vegetarian Restaurant, and McFarlane's carriage-building +depot. That carries us right on to the other block. And now, +Doctor, we've done our work, so it's time we had some play. A +sandwich and a cup of coffee, and then off to violin-land, where +all is sweetness and delicacy and harmony, and there are no +red-headed clients to vex us with their conundrums." + +My friend was an enthusiastic musician, being himself not only a +very capable performer but a composer of no ordinary merit. All +the afternoon he sat in the stalls wrapped in the most perfect +happiness, gently waving his long, thin fingers in time to the +music, while his gently smiling face and his languid, dreamy eyes +were as unlike those of Holmes the sleuth-hound, Holmes the +relentless, keen-witted, ready-handed criminal agent, as it was +possible to conceive. In his singular character the dual nature +alternately asserted itself, and his extreme exactness and +astuteness represented, as I have often thought, the reaction +against the poetic and contemplative mood which occasionally +predominated in him. The swing of his nature took him from +extreme languor to devouring energy; and, as I knew well, he was +never so truly formidable as when, for days on end, he had been +lounging in his armchair amid his improvisations and his +black-letter editions. Then it was that the lust of the chase +would suddenly come upon him, and that his brilliant reasoning +power would rise to the level of intuition, until those who were +unacquainted with his methods would look askance at him as on a +man whose knowledge was not that of other mortals. When I saw him +that afternoon so enwrapped in the music at St. James's Hall I +felt that an evil time might be coming upon those whom he had set +himself to hunt down. + +"You want to go home, no doubt, Doctor," he remarked as we +emerged. + +"Yes, it would be as well." + +"And I have some business to do which will take some hours. This +business at Coburg Square is serious." + +"Why serious?" + +"A considerable crime is in contemplation. I have every reason to +believe that we shall be in time to stop it. But to-day being +Saturday rather complicates matters. I shall want your help +to-night." + +"At what time?" + +"Ten will be early enough." + +"I shall be at Baker Street at ten." + +"Very well. And, I say, Doctor, there may be some little danger, +so kindly put your army revolver in your pocket." He waved his +hand, turned on his heel, and disappeared in an instant among the +crowd. + +I trust that I am not more dense than my neighbours, but I was +always oppressed with a sense of my own stupidity in my dealings +with Sherlock Holmes. Here I had heard what he had heard, I had +seen what he had seen, and yet from his words it was evident that +he saw clearly not only what had happened but what was about to +happen, while to me the whole business was still confused and +grotesque. As I drove home to my house in Kensington I thought +over it all, from the extraordinary story of the red-headed +copier of the "Encyclopaedia" down to the visit to Saxe-Coburg +Square, and the ominous words with which he had parted from me. +What was this nocturnal expedition, and why should I go armed? +Where were we going, and what were we to do? I had the hint from +Holmes that this smooth-faced pawnbroker's assistant was a +formidable man--a man who might play a deep game. I tried to +puzzle it out, but gave it up in despair and set the matter aside +until night should bring an explanation. + +It was a quarter-past nine when I started from home and made my +way across the Park, and so through Oxford Street to Baker +Street. Two hansoms were standing at the door, and as I entered +the passage I heard the sound of voices from above. On entering +his room I found Holmes in animated conversation with two men, +one of whom I recognised as Peter Jones, the official police +agent, while the other was a long, thin, sad-faced man, with a +very shiny hat and oppressively respectable frock-coat. + +"Ha! Our party is complete," said Holmes, buttoning up his +pea-jacket and taking his heavy hunting crop from the rack. +"Watson, I think you know Mr. Jones, of Scotland Yard? Let me +introduce you to Mr. Merryweather, who is to be our companion in +to-night's adventure." + +"We're hunting in couples again, Doctor, you see," said Jones in +his consequential way. "Our friend here is a wonderful man for +starting a chase. All he wants is an old dog to help him to do +the running down." + +"I hope a wild goose may not prove to be the end of our chase," +observed Mr. Merryweather gloomily. + +"You may place considerable confidence in Mr. Holmes, sir," said +the police agent loftily. "He has his own little methods, which +are, if he won't mind my saying so, just a little too theoretical +and fantastic, but he has the makings of a detective in him. It +is not too much to say that once or twice, as in that business of +the Sholto murder and the Agra treasure, he has been more nearly +correct than the official force." + +"Oh, if you say so, Mr. Jones, it is all right," said the +stranger with deference. "Still, I confess that I miss my rubber. +It is the first Saturday night for seven-and-twenty years that I +have not had my rubber." + +"I think you will find," said Sherlock Holmes, "that you will +play for a higher stake to-night than you have ever done yet, and +that the play will be more exciting. For you, Mr. Merryweather, +the stake will be some 30,000 pounds; and for you, Jones, it will +be the man upon whom you wish to lay your hands." + +"John Clay, the murderer, thief, smasher, and forger. He's a +young man, Mr. Merryweather, but he is at the head of his +profession, and I would rather have my bracelets on him than on +any criminal in London. He's a remarkable man, is young John +Clay. His grandfather was a royal duke, and he himself has been +to Eton and Oxford. His brain is as cunning as his fingers, and +though we meet signs of him at every turn, we never know where to +find the man himself. He'll crack a crib in Scotland one week, +and be raising money to build an orphanage in Cornwall the next. +I've been on his track for years and have never set eyes on him +yet." + +"I hope that I may have the pleasure of introducing you to-night. +I've had one or two little turns also with Mr. John Clay, and I +agree with you that he is at the head of his profession. It is +past ten, however, and quite time that we started. If you two +will take the first hansom, Watson and I will follow in the +second." + +Sherlock Holmes was not very communicative during the long drive +and lay back in the cab humming the tunes which he had heard in +the afternoon. We rattled through an endless labyrinth of gas-lit +streets until we emerged into Farrington Street. + +"We are close there now," my friend remarked. "This fellow +Merryweather is a bank director, and personally interested in the +matter. I thought it as well to have Jones with us also. He is +not a bad fellow, though an absolute imbecile in his profession. +He has one positive virtue. He is as brave as a bulldog and as +tenacious as a lobster if he gets his claws upon anyone. Here we +are, and they are waiting for us." + +We had reached the same crowded thoroughfare in which we had +found ourselves in the morning. Our cabs were dismissed, and, +following the guidance of Mr. Merryweather, we passed down a +narrow passage and through a side door, which he opened for us. +Within there was a small corridor, which ended in a very massive +iron gate. This also was opened, and led down a flight of winding +stone steps, which terminated at another formidable gate. Mr. +Merryweather stopped to light a lantern, and then conducted us +down a dark, earth-smelling passage, and so, after opening a +third door, into a huge vault or cellar, which was piled all +round with crates and massive boxes. + +"You are not very vulnerable from above," Holmes remarked as he +held up the lantern and gazed about him. + +"Nor from below," said Mr. Merryweather, striking his stick upon +the flags which lined the floor. "Why, dear me, it sounds quite +hollow!" he remarked, looking up in surprise. + +"I must really ask you to be a little more quiet!" said Holmes +severely. "You have already imperilled the whole success of our +expedition. Might I beg that you would have the goodness to sit +down upon one of those boxes, and not to interfere?" + +The solemn Mr. Merryweather perched himself upon a crate, with a +very injured expression upon his face, while Holmes fell upon his +knees upon the floor and, with the lantern and a magnifying lens, +began to examine minutely the cracks between the stones. A few +seconds sufficed to satisfy him, for he sprang to his feet again +and put his glass in his pocket. + +"We have at least an hour before us," he remarked, "for they can +hardly take any steps until the good pawnbroker is safely in bed. +Then they will not lose a minute, for the sooner they do their +work the longer time they will have for their escape. We are at +present, Doctor--as no doubt you have divined--in the cellar of +the City branch of one of the principal London banks. Mr. +Merryweather is the chairman of directors, and he will explain to +you that there are reasons why the more daring criminals of +London should take a considerable interest in this cellar at +present." + +"It is our French gold," whispered the director. "We have had +several warnings that an attempt might be made upon it." + +"Your French gold?" + +"Yes. We had occasion some months ago to strengthen our resources +and borrowed for that purpose 30,000 napoleons from the Bank of +France. It has become known that we have never had occasion to +unpack the money, and that it is still lying in our cellar. The +crate upon which I sit contains 2,000 napoleons packed between +layers of lead foil. Our reserve of bullion is much larger at +present than is usually kept in a single branch office, and the +directors have had misgivings upon the subject." + +"Which were very well justified," observed Holmes. "And now it is +time that we arranged our little plans. I expect that within an +hour matters will come to a head. In the meantime Mr. +Merryweather, we must put the screen over that dark lantern." + +"And sit in the dark?" + +"I am afraid so. I had brought a pack of cards in my pocket, and +I thought that, as we were a partie carrée, you might have your +rubber after all. But I see that the enemy's preparations have +gone so far that we cannot risk the presence of a light. And, +first of all, we must choose our positions. These are daring men, +and though we shall take them at a disadvantage, they may do us +some harm unless we are careful. I shall stand behind this crate, +and do you conceal yourselves behind those. Then, when I flash a +light upon them, close in swiftly. If they fire, Watson, have no +compunction about shooting them down." + +I placed my revolver, cocked, upon the top of the wooden case +behind which I crouched. Holmes shot the slide across the front +of his lantern and left us in pitch darkness--such an absolute +darkness as I have never before experienced. The smell of hot +metal remained to assure us that the light was still there, ready +to flash out at a moment's notice. To me, with my nerves worked +up to a pitch of expectancy, there was something depressing and +subduing in the sudden gloom, and in the cold dank air of the +vault. + +"They have but one retreat," whispered Holmes. "That is back +through the house into Saxe-Coburg Square. I hope that you have +done what I asked you, Jones?" + +"I have an inspector and two officers waiting at the front door." + +"Then we have stopped all the holes. And now we must be silent +and wait." + +What a time it seemed! From comparing notes afterwards it was but +an hour and a quarter, yet it appeared to me that the night must +have almost gone and the dawn be breaking above us. My limbs +were weary and stiff, for I feared to change my position; yet my +nerves were worked up to the highest pitch of tension, and my +hearing was so acute that I could not only hear the gentle +breathing of my companions, but I could distinguish the deeper, +heavier in-breath of the bulky Jones from the thin, sighing note +of the bank director. From my position I could look over the case +in the direction of the floor. Suddenly my eyes caught the glint +of a light. + +At first it was but a lurid spark upon the stone pavement. Then +it lengthened out until it became a yellow line, and then, +without any warning or sound, a gash seemed to open and a hand +appeared, a white, almost womanly hand, which felt about in the +centre of the little area of light. For a minute or more the +hand, with its writhing fingers, protruded out of the floor. Then +it was withdrawn as suddenly as it appeared, and all was dark +again save the single lurid spark which marked a chink between +the stones. + +Its disappearance, however, was but momentary. With a rending, +tearing sound, one of the broad, white stones turned over upon +its side and left a square, gaping hole, through which streamed +the light of a lantern. Over the edge there peeped a clean-cut, +boyish face, which looked keenly about it, and then, with a hand +on either side of the aperture, drew itself shoulder-high and +waist-high, until one knee rested upon the edge. In another +instant he stood at the side of the hole and was hauling after +him a companion, lithe and small like himself, with a pale face +and a shock of very red hair. + +"It's all clear," he whispered. "Have you the chisel and the +bags? Great Scott! Jump, Archie, jump, and I'll swing for it!" + +Sherlock Holmes had sprung out and seized the intruder by the +collar. The other dived down the hole, and I heard the sound of +rending cloth as Jones clutched at his skirts. The light flashed +upon the barrel of a revolver, but Holmes' hunting crop came +down on the man's wrist, and the pistol clinked upon the stone +floor. + +"It's no use, John Clay," said Holmes blandly. "You have no +chance at all." + +"So I see," the other answered with the utmost coolness. "I fancy +that my pal is all right, though I see you have got his +coat-tails." + +"There are three men waiting for him at the door," said Holmes. + +"Oh, indeed! You seem to have done the thing very completely. I +must compliment you." + +"And I you," Holmes answered. "Your red-headed idea was very new +and effective." + +"You'll see your pal again presently," said Jones. "He's quicker +at climbing down holes than I am. Just hold out while I fix the +derbies." + +"I beg that you will not touch me with your filthy hands," +remarked our prisoner as the handcuffs clattered upon his wrists. +"You may not be aware that I have royal blood in my veins. Have +the goodness, also, when you address me always to say 'sir' and +'please.'" + +"All right," said Jones with a stare and a snigger. "Well, would +you please, sir, march upstairs, where we can get a cab to carry +your Highness to the police-station?" + +"That is better," said John Clay serenely. He made a sweeping bow +to the three of us and walked quietly off in the custody of the +detective. + +"Really, Mr. Holmes," said Mr. Merryweather as we followed them +from the cellar, "I do not know how the bank can thank you or +repay you. There is no doubt that you have detected and defeated +in the most complete manner one of the most determined attempts +at bank robbery that have ever come within my experience." + +"I have had one or two little scores of my own to settle with Mr. +John Clay," said Holmes. "I have been at some small expense over +this matter, which I shall expect the bank to refund, but beyond +that I am amply repaid by having had an experience which is in +many ways unique, and by hearing the very remarkable narrative of +the Red-headed League." + + +"You see, Watson," he explained in the early hours of the morning +as we sat over a glass of whisky and soda in Baker Street, "it +was perfectly obvious from the first that the only possible +object of this rather fantastic business of the advertisement of +the League, and the copying of the 'Encyclopaedia,' must be to get +this not over-bright pawnbroker out of the way for a number of +hours every day. It was a curious way of managing it, but, +really, it would be difficult to suggest a better. The method was +no doubt suggested to Clay's ingenious mind by the colour of his +accomplice's hair. The 4 pounds a week was a lure which must draw +him, and what was it to them, who were playing for thousands? +They put in the advertisement, one rogue has the temporary +office, the other rogue incites the man to apply for it, and +together they manage to secure his absence every morning in the +week. From the time that I heard of the assistant having come for +half wages, it was obvious to me that he had some strong motive +for securing the situation." + +"But how could you guess what the motive was?" + +"Had there been women in the house, I should have suspected a +mere vulgar intrigue. That, however, was out of the question. The +man's business was a small one, and there was nothing in his +house which could account for such elaborate preparations, and +such an expenditure as they were at. It must, then, be something +out of the house. What could it be? I thought of the assistant's +fondness for photography, and his trick of vanishing into the +cellar. The cellar! There was the end of this tangled clue. Then +I made inquiries as to this mysterious assistant and found that I +had to deal with one of the coolest and most daring criminals in +London. He was doing something in the cellar--something which +took many hours a day for months on end. What could it be, once +more? I could think of nothing save that he was running a tunnel +to some other building. + +"So far I had got when we went to visit the scene of action. I +surprised you by beating upon the pavement with my stick. I was +ascertaining whether the cellar stretched out in front or behind. +It was not in front. Then I rang the bell, and, as I hoped, the +assistant answered it. We have had some skirmishes, but we had +never set eyes upon each other before. I hardly looked at his +face. His knees were what I wished to see. You must yourself have +remarked how worn, wrinkled, and stained they were. They spoke of +those hours of burrowing. The only remaining point was what they +were burrowing for. I walked round the corner, saw the City and +Suburban Bank abutted on our friend's premises, and felt that I +had solved my problem. When you drove home after the concert I +called upon Scotland Yard and upon the chairman of the bank +directors, with the result that you have seen." + +"And how could you tell that they would make their attempt +to-night?" I asked. + +"Well, when they closed their League offices that was a sign that +they cared no longer about Mr. Jabez Wilson's presence--in other +words, that they had completed their tunnel. But it was essential +that they should use it soon, as it might be discovered, or the +bullion might be removed. Saturday would suit them better than +any other day, as it would give them two days for their escape. +For all these reasons I expected them to come to-night." + +"You reasoned it out beautifully," I exclaimed in unfeigned +admiration. "It is so long a chain, and yet every link rings +true." + +"It saved me from ennui," he answered, yawning. "Alas! I already +feel it closing in upon me. My life is spent in one long effort +to escape from the commonplaces of existence. These little +problems help me to do so." + +"And you are a benefactor of the race," said I. + +He shrugged his shoulders. "Well, perhaps, after all, it is of +some little use," he remarked. "'L'homme c'est rien--l'oeuvre +c'est tout,' as Gustave Flaubert wrote to George Sand." + + + +ADVENTURE III. A CASE OF IDENTITY + +"My dear fellow," said Sherlock Holmes as we sat on either side +of the fire in his lodgings at Baker Street, "life is infinitely +stranger than anything which the mind of man could invent. We +would not dare to conceive the things which are really mere +commonplaces of existence. If we could fly out of that window +hand in hand, hover over this great city, gently remove the +roofs, and peep in at the queer things which are going on, the +strange coincidences, the plannings, the cross-purposes, the +wonderful chains of events, working through generations, and +leading to the most outré results, it would make all fiction with +its conventionalities and foreseen conclusions most stale and +unprofitable." + +"And yet I am not convinced of it," I answered. "The cases which +come to light in the papers are, as a rule, bald enough, and +vulgar enough. We have in our police reports realism pushed to +its extreme limits, and yet the result is, it must be confessed, +neither fascinating nor artistic." + +"A certain selection and discretion must be used in producing a +realistic effect," remarked Holmes. "This is wanting in the +police report, where more stress is laid, perhaps, upon the +platitudes of the magistrate than upon the details, which to an +observer contain the vital essence of the whole matter. Depend +upon it, there is nothing so unnatural as the commonplace." + +I smiled and shook my head. "I can quite understand your thinking +so," I said. "Of course, in your position of unofficial adviser +and helper to everybody who is absolutely puzzled, throughout +three continents, you are brought in contact with all that is +strange and bizarre. But here"--I picked up the morning paper +from the ground--"let us put it to a practical test. Here is the +first heading upon which I come. 'A husband's cruelty to his +wife.' There is half a column of print, but I know without +reading it that it is all perfectly familiar to me. There is, of +course, the other woman, the drink, the push, the blow, the +bruise, the sympathetic sister or landlady. The crudest of +writers could invent nothing more crude." + +"Indeed, your example is an unfortunate one for your argument," +said Holmes, taking the paper and glancing his eye down it. "This +is the Dundas separation case, and, as it happens, I was engaged +in clearing up some small points in connection with it. The +husband was a teetotaler, there was no other woman, and the +conduct complained of was that he had drifted into the habit of +winding up every meal by taking out his false teeth and hurling +them at his wife, which, you will allow, is not an action likely +to occur to the imagination of the average story-teller. Take a +pinch of snuff, Doctor, and acknowledge that I have scored over +you in your example." + +He held out his snuffbox of old gold, with a great amethyst in +the centre of the lid. Its splendour was in such contrast to his +homely ways and simple life that I could not help commenting upon +it. + +"Ah," said he, "I forgot that I had not seen you for some weeks. +It is a little souvenir from the King of Bohemia in return for my +assistance in the case of the Irene Adler papers." + +"And the ring?" I asked, glancing at a remarkable brilliant which +sparkled upon his finger. + +"It was from the reigning family of Holland, though the matter in +which I served them was of such delicacy that I cannot confide it +even to you, who have been good enough to chronicle one or two of +my little problems." + +"And have you any on hand just now?" I asked with interest. + +"Some ten or twelve, but none which present any feature of +interest. They are important, you understand, without being +interesting. Indeed, I have found that it is usually in +unimportant matters that there is a field for the observation, +and for the quick analysis of cause and effect which gives the +charm to an investigation. The larger crimes are apt to be the +simpler, for the bigger the crime the more obvious, as a rule, is +the motive. In these cases, save for one rather intricate matter +which has been referred to me from Marseilles, there is nothing +which presents any features of interest. It is possible, however, +that I may have something better before very many minutes are +over, for this is one of my clients, or I am much mistaken." + +He had risen from his chair and was standing between the parted +blinds gazing down into the dull neutral-tinted London street. +Looking over his shoulder, I saw that on the pavement opposite +there stood a large woman with a heavy fur boa round her neck, +and a large curling red feather in a broad-brimmed hat which was +tilted in a coquettish Duchess of Devonshire fashion over her +ear. From under this great panoply she peeped up in a nervous, +hesitating fashion at our windows, while her body oscillated +backward and forward, and her fingers fidgeted with her glove +buttons. Suddenly, with a plunge, as of the swimmer who leaves +the bank, she hurried across the road, and we heard the sharp +clang of the bell. + +"I have seen those symptoms before," said Holmes, throwing his +cigarette into the fire. "Oscillation upon the pavement always +means an affaire de coeur. She would like advice, but is not sure +that the matter is not too delicate for communication. And yet +even here we may discriminate. When a woman has been seriously +wronged by a man she no longer oscillates, and the usual symptom +is a broken bell wire. Here we may take it that there is a love +matter, but that the maiden is not so much angry as perplexed, or +grieved. But here she comes in person to resolve our doubts." + +As he spoke there was a tap at the door, and the boy in buttons +entered to announce Miss Mary Sutherland, while the lady herself +loomed behind his small black figure like a full-sailed +merchant-man behind a tiny pilot boat. Sherlock Holmes welcomed +her with the easy courtesy for which he was remarkable, and, +having closed the door and bowed her into an armchair, he looked +her over in the minute and yet abstracted fashion which was +peculiar to him. + +"Do you not find," he said, "that with your short sight it is a +little trying to do so much typewriting?" + +"I did at first," she answered, "but now I know where the letters +are without looking." Then, suddenly realising the full purport +of his words, she gave a violent start and looked up, with fear +and astonishment upon her broad, good-humoured face. "You've +heard about me, Mr. Holmes," she cried, "else how could you know +all that?" + +"Never mind," said Holmes, laughing; "it is my business to know +things. Perhaps I have trained myself to see what others +overlook. If not, why should you come to consult me?" + +"I came to you, sir, because I heard of you from Mrs. Etherege, +whose husband you found so easy when the police and everyone had +given him up for dead. Oh, Mr. Holmes, I wish you would do as +much for me. I'm not rich, but still I have a hundred a year in +my own right, besides the little that I make by the machine, and +I would give it all to know what has become of Mr. Hosmer Angel." + +"Why did you come away to consult me in such a hurry?" asked +Sherlock Holmes, with his finger-tips together and his eyes to +the ceiling. + +Again a startled look came over the somewhat vacuous face of Miss +Mary Sutherland. "Yes, I did bang out of the house," she said, +"for it made me angry to see the easy way in which Mr. +Windibank--that is, my father--took it all. He would not go to +the police, and he would not go to you, and so at last, as he +would do nothing and kept on saying that there was no harm done, +it made me mad, and I just on with my things and came right away +to you." + +"Your father," said Holmes, "your stepfather, surely, since the +name is different." + +"Yes, my stepfather. I call him father, though it sounds funny, +too, for he is only five years and two months older than myself." + +"And your mother is alive?" + +"Oh, yes, mother is alive and well. I wasn't best pleased, Mr. +Holmes, when she married again so soon after father's death, and +a man who was nearly fifteen years younger than herself. Father +was a plumber in the Tottenham Court Road, and he left a tidy +business behind him, which mother carried on with Mr. Hardy, the +foreman; but when Mr. Windibank came he made her sell the +business, for he was very superior, being a traveller in wines. +They got 4700 pounds for the goodwill and interest, which wasn't +near as much as father could have got if he had been alive." + +I had expected to see Sherlock Holmes impatient under this +rambling and inconsequential narrative, but, on the contrary, he +had listened with the greatest concentration of attention. + +"Your own little income," he asked, "does it come out of the +business?" + +"Oh, no, sir. It is quite separate and was left me by my uncle +Ned in Auckland. It is in New Zealand stock, paying 4 1/2 per +cent. Two thousand five hundred pounds was the amount, but I can +only touch the interest." + +"You interest me extremely," said Holmes. "And since you draw so +large a sum as a hundred a year, with what you earn into the +bargain, you no doubt travel a little and indulge yourself in +every way. I believe that a single lady can get on very nicely +upon an income of about 60 pounds." + +"I could do with much less than that, Mr. Holmes, but you +understand that as long as I live at home I don't wish to be a +burden to them, and so they have the use of the money just while +I am staying with them. Of course, that is only just for the +time. Mr. Windibank draws my interest every quarter and pays it +over to mother, and I find that I can do pretty well with what I +earn at typewriting. It brings me twopence a sheet, and I can +often do from fifteen to twenty sheets in a day." + +"You have made your position very clear to me," said Holmes. +"This is my friend, Dr. Watson, before whom you can speak as +freely as before myself. Kindly tell us now all about your +connection with Mr. Hosmer Angel." + +A flush stole over Miss Sutherland's face, and she picked +nervously at the fringe of her jacket. "I met him first at the +gasfitters' ball," she said. "They used to send father tickets +when he was alive, and then afterwards they remembered us, and +sent them to mother. Mr. Windibank did not wish us to go. He +never did wish us to go anywhere. He would get quite mad if I +wanted so much as to join a Sunday-school treat. But this time I +was set on going, and I would go; for what right had he to +prevent? He said the folk were not fit for us to know, when all +father's friends were to be there. And he said that I had nothing +fit to wear, when I had my purple plush that I had never so much +as taken out of the drawer. At last, when nothing else would do, +he went off to France upon the business of the firm, but we went, +mother and I, with Mr. Hardy, who used to be our foreman, and it +was there I met Mr. Hosmer Angel." + +"I suppose," said Holmes, "that when Mr. Windibank came back from +France he was very annoyed at your having gone to the ball." + +"Oh, well, he was very good about it. He laughed, I remember, and +shrugged his shoulders, and said there was no use denying +anything to a woman, for she would have her way." + +"I see. Then at the gasfitters' ball you met, as I understand, a +gentleman called Mr. Hosmer Angel." + +"Yes, sir. I met him that night, and he called next day to ask if +we had got home all safe, and after that we met him--that is to +say, Mr. Holmes, I met him twice for walks, but after that father +came back again, and Mr. Hosmer Angel could not come to the house +any more." + +"No?" + +"Well, you know father didn't like anything of the sort. He +wouldn't have any visitors if he could help it, and he used to +say that a woman should be happy in her own family circle. But +then, as I used to say to mother, a woman wants her own circle to +begin with, and I had not got mine yet." + +"But how about Mr. Hosmer Angel? Did he make no attempt to see +you?" + +"Well, father was going off to France again in a week, and Hosmer +wrote and said that it would be safer and better not to see each +other until he had gone. We could write in the meantime, and he +used to write every day. I took the letters in in the morning, so +there was no need for father to know." + +"Were you engaged to the gentleman at this time?" + +"Oh, yes, Mr. Holmes. We were engaged after the first walk that +we took. Hosmer--Mr. Angel--was a cashier in an office in +Leadenhall Street--and--" + +"What office?" + +"That's the worst of it, Mr. Holmes, I don't know." + +"Where did he live, then?" + +"He slept on the premises." + +"And you don't know his address?" + +"No--except that it was Leadenhall Street." + +"Where did you address your letters, then?" + +"To the Leadenhall Street Post Office, to be left till called +for. He said that if they were sent to the office he would be +chaffed by all the other clerks about having letters from a lady, +so I offered to typewrite them, like he did his, but he wouldn't +have that, for he said that when I wrote them they seemed to come +from me, but when they were typewritten he always felt that the +machine had come between us. That will just show you how fond he +was of me, Mr. Holmes, and the little things that he would think +of." + +"It was most suggestive," said Holmes. "It has long been an axiom +of mine that the little things are infinitely the most important. +Can you remember any other little things about Mr. Hosmer Angel?" + +"He was a very shy man, Mr. Holmes. He would rather walk with me +in the evening than in the daylight, for he said that he hated to +be conspicuous. Very retiring and gentlemanly he was. Even his +voice was gentle. He'd had the quinsy and swollen glands when he +was young, he told me, and it had left him with a weak throat, +and a hesitating, whispering fashion of speech. He was always +well dressed, very neat and plain, but his eyes were weak, just +as mine are, and he wore tinted glasses against the glare." + +"Well, and what happened when Mr. Windibank, your stepfather, +returned to France?" + +"Mr. Hosmer Angel came to the house again and proposed that we +should marry before father came back. He was in dreadful earnest +and made me swear, with my hands on the Testament, that whatever +happened I would always be true to him. Mother said he was quite +right to make me swear, and that it was a sign of his passion. +Mother was all in his favour from the first and was even fonder +of him than I was. Then, when they talked of marrying within the +week, I began to ask about father; but they both said never to +mind about father, but just to tell him afterwards, and mother +said she would make it all right with him. I didn't quite like +that, Mr. Holmes. It seemed funny that I should ask his leave, as +he was only a few years older than me; but I didn't want to do +anything on the sly, so I wrote to father at Bordeaux, where the +company has its French offices, but the letter came back to me on +the very morning of the wedding." + +"It missed him, then?" + +"Yes, sir; for he had started to England just before it arrived." + +"Ha! that was unfortunate. Your wedding was arranged, then, for +the Friday. Was it to be in church?" + +"Yes, sir, but very quietly. It was to be at St. Saviour's, near +King's Cross, and we were to have breakfast afterwards at the St. +Pancras Hotel. Hosmer came for us in a hansom, but as there were +two of us he put us both into it and stepped himself into a +four-wheeler, which happened to be the only other cab in the +street. We got to the church first, and when the four-wheeler +drove up we waited for him to step out, but he never did, and +when the cabman got down from the box and looked there was no one +there! The cabman said that he could not imagine what had become +of him, for he had seen him get in with his own eyes. That was +last Friday, Mr. Holmes, and I have never seen or heard anything +since then to throw any light upon what became of him." + +"It seems to me that you have been very shamefully treated," said +Holmes. + +"Oh, no, sir! He was too good and kind to leave me so. Why, all +the morning he was saying to me that, whatever happened, I was to +be true; and that even if something quite unforeseen occurred to +separate us, I was always to remember that I was pledged to him, +and that he would claim his pledge sooner or later. It seemed +strange talk for a wedding-morning, but what has happened since +gives a meaning to it." + +"Most certainly it does. Your own opinion is, then, that some +unforeseen catastrophe has occurred to him?" + +"Yes, sir. I believe that he foresaw some danger, or else he +would not have talked so. And then I think that what he foresaw +happened." + +"But you have no notion as to what it could have been?" + +"None." + +"One more question. How did your mother take the matter?" + +"She was angry, and said that I was never to speak of the matter +again." + +"And your father? Did you tell him?" + +"Yes; and he seemed to think, with me, that something had +happened, and that I should hear of Hosmer again. As he said, +what interest could anyone have in bringing me to the doors of +the church, and then leaving me? Now, if he had borrowed my +money, or if he had married me and got my money settled on him, +there might be some reason, but Hosmer was very independent about +money and never would look at a shilling of mine. And yet, what +could have happened? And why could he not write? Oh, it drives me +half-mad to think of it, and I can't sleep a wink at night." She +pulled a little handkerchief out of her muff and began to sob +heavily into it. + +"I shall glance into the case for you," said Holmes, rising, "and +I have no doubt that we shall reach some definite result. Let the +weight of the matter rest upon me now, and do not let your mind +dwell upon it further. Above all, try to let Mr. Hosmer Angel +vanish from your memory, as he has done from your life." + +"Then you don't think I'll see him again?" + +"I fear not." + +"Then what has happened to him?" + +"You will leave that question in my hands. I should like an +accurate description of him and any letters of his which you can +spare." + +"I advertised for him in last Saturday's Chronicle," said she. +"Here is the slip and here are four letters from him." + +"Thank you. And your address?" + +"No. 31 Lyon Place, Camberwell." + +"Mr. Angel's address you never had, I understand. Where is your +father's place of business?" + +"He travels for Westhouse & Marbank, the great claret importers +of Fenchurch Street." + +"Thank you. You have made your statement very clearly. You will +leave the papers here, and remember the advice which I have given +you. Let the whole incident be a sealed book, and do not allow it +to affect your life." + +"You are very kind, Mr. Holmes, but I cannot do that. I shall be +true to Hosmer. He shall find me ready when he comes back." + +For all the preposterous hat and the vacuous face, there was +something noble in the simple faith of our visitor which +compelled our respect. She laid her little bundle of papers upon +the table and went her way, with a promise to come again whenever +she might be summoned. + +Sherlock Holmes sat silent for a few minutes with his fingertips +still pressed together, his legs stretched out in front of him, +and his gaze directed upward to the ceiling. Then he took down +from the rack the old and oily clay pipe, which was to him as a +counsellor, and, having lit it, he leaned back in his chair, with +the thick blue cloud-wreaths spinning up from him, and a look of +infinite languor in his face. + +"Quite an interesting study, that maiden," he observed. "I found +her more interesting than her little problem, which, by the way, +is rather a trite one. You will find parallel cases, if you +consult my index, in Andover in '77, and there was something of +the sort at The Hague last year. Old as is the idea, however, +there were one or two details which were new to me. But the +maiden herself was most instructive." + +"You appeared to read a good deal upon her which was quite +invisible to me," I remarked. + +"Not invisible but unnoticed, Watson. You did not know where to +look, and so you missed all that was important. I can never bring +you to realise the importance of sleeves, the suggestiveness of +thumb-nails, or the great issues that may hang from a boot-lace. +Now, what did you gather from that woman's appearance? Describe +it." + +"Well, she had a slate-coloured, broad-brimmed straw hat, with a +feather of a brickish red. Her jacket was black, with black beads +sewn upon it, and a fringe of little black jet ornaments. Her +dress was brown, rather darker than coffee colour, with a little +purple plush at the neck and sleeves. Her gloves were greyish and +were worn through at the right forefinger. Her boots I didn't +observe. She had small round, hanging gold earrings, and a +general air of being fairly well-to-do in a vulgar, comfortable, +easy-going way." + +Sherlock Holmes clapped his hands softly together and chuckled. + +"'Pon my word, Watson, you are coming along wonderfully. You have +really done very well indeed. It is true that you have missed +everything of importance, but you have hit upon the method, and +you have a quick eye for colour. Never trust to general +impressions, my boy, but concentrate yourself upon details. My +first glance is always at a woman's sleeve. In a man it is +perhaps better first to take the knee of the trouser. As you +observe, this woman had plush upon her sleeves, which is a most +useful material for showing traces. The double line a little +above the wrist, where the typewritist presses against the table, +was beautifully defined. The sewing-machine, of the hand type, +leaves a similar mark, but only on the left arm, and on the side +of it farthest from the thumb, instead of being right across the +broadest part, as this was. I then glanced at her face, and, +observing the dint of a pince-nez at either side of her nose, I +ventured a remark upon short sight and typewriting, which seemed +to surprise her." + +"It surprised me." + +"But, surely, it was obvious. I was then much surprised and +interested on glancing down to observe that, though the boots +which she was wearing were not unlike each other, they were +really odd ones; the one having a slightly decorated toe-cap, and +the other a plain one. One was buttoned only in the two lower +buttons out of five, and the other at the first, third, and +fifth. Now, when you see that a young lady, otherwise neatly +dressed, has come away from home with odd boots, half-buttoned, +it is no great deduction to say that she came away in a hurry." + +"And what else?" I asked, keenly interested, as I always was, by +my friend's incisive reasoning. + +"I noted, in passing, that she had written a note before leaving +home but after being fully dressed. You observed that her right +glove was torn at the forefinger, but you did not apparently see +that both glove and finger were stained with violet ink. She had +written in a hurry and dipped her pen too deep. It must have been +this morning, or the mark would not remain clear upon the finger. +All this is amusing, though rather elementary, but I must go back +to business, Watson. Would you mind reading me the advertised +description of Mr. Hosmer Angel?" + +I held the little printed slip to the light. + +"Missing," it said, "on the morning of the fourteenth, a gentleman +named Hosmer Angel. About five ft. seven in. in height; +strongly built, sallow complexion, black hair, a little bald in +the centre, bushy, black side-whiskers and moustache; tinted +glasses, slight infirmity of speech. Was dressed, when last seen, +in black frock-coat faced with silk, black waistcoat, gold Albert +chain, and grey Harris tweed trousers, with brown gaiters over +elastic-sided boots. Known to have been employed in an office in +Leadenhall Street. Anybody bringing--" + +"That will do," said Holmes. "As to the letters," he continued, +glancing over them, "they are very commonplace. Absolutely no +clue in them to Mr. Angel, save that he quotes Balzac once. There +is one remarkable point, however, which will no doubt strike +you." + +"They are typewritten," I remarked. + +"Not only that, but the signature is typewritten. Look at the +neat little 'Hosmer Angel' at the bottom. There is a date, you +see, but no superscription except Leadenhall Street, which is +rather vague. The point about the signature is very suggestive--in +fact, we may call it conclusive." + +"Of what?" + +"My dear fellow, is it possible you do not see how strongly it +bears upon the case?" + +"I cannot say that I do unless it were that he wished to be able +to deny his signature if an action for breach of promise were +instituted." + +"No, that was not the point. However, I shall write two letters, +which should settle the matter. One is to a firm in the City, the +other is to the young lady's stepfather, Mr. Windibank, asking +him whether he could meet us here at six o'clock tomorrow +evening. It is just as well that we should do business with the +male relatives. And now, Doctor, we can do nothing until the +answers to those letters come, so we may put our little problem +upon the shelf for the interim." + +I had had so many reasons to believe in my friend's subtle powers +of reasoning and extraordinary energy in action that I felt that +he must have some solid grounds for the assured and easy +demeanour with which he treated the singular mystery which he had +been called upon to fathom. Once only had I known him to fail, in +the case of the King of Bohemia and of the Irene Adler +photograph; but when I looked back to the weird business of the +Sign of Four, and the extraordinary circumstances connected with +the Study in Scarlet, I felt that it would be a strange tangle +indeed which he could not unravel. + +I left him then, still puffing at his black clay pipe, with the +conviction that when I came again on the next evening I would +find that he held in his hands all the clues which would lead up +to the identity of the disappearing bridegroom of Miss Mary +Sutherland. + +A professional case of great gravity was engaging my own +attention at the time, and the whole of next day I was busy at +the bedside of the sufferer. It was not until close upon six +o'clock that I found myself free and was able to spring into a +hansom and drive to Baker Street, half afraid that I might be too +late to assist at the dénouement of the little mystery. I found +Sherlock Holmes alone, however, half asleep, with his long, thin +form curled up in the recesses of his armchair. A formidable +array of bottles and test-tubes, with the pungent cleanly smell +of hydrochloric acid, told me that he had spent his day in the +chemical work which was so dear to him. + +"Well, have you solved it?" I asked as I entered. + +"Yes. It was the bisulphate of baryta." + +"No, no, the mystery!" I cried. + +"Oh, that! I thought of the salt that I have been working upon. +There was never any mystery in the matter, though, as I said +yesterday, some of the details are of interest. The only drawback +is that there is no law, I fear, that can touch the scoundrel." + +"Who was he, then, and what was his object in deserting Miss +Sutherland?" + +The question was hardly out of my mouth, and Holmes had not yet +opened his lips to reply, when we heard a heavy footfall in the +passage and a tap at the door. + +"This is the girl's stepfather, Mr. James Windibank," said +Holmes. "He has written to me to say that he would be here at +six. Come in!" + +The man who entered was a sturdy, middle-sized fellow, some +thirty years of age, clean-shaven, and sallow-skinned, with a +bland, insinuating manner, and a pair of wonderfully sharp and +penetrating grey eyes. He shot a questioning glance at each of +us, placed his shiny top-hat upon the sideboard, and with a +slight bow sidled down into the nearest chair. + +"Good-evening, Mr. James Windibank," said Holmes. "I think that +this typewritten letter is from you, in which you made an +appointment with me for six o'clock?" + +"Yes, sir. I am afraid that I am a little late, but I am not +quite my own master, you know. I am sorry that Miss Sutherland +has troubled you about this little matter, for I think it is far +better not to wash linen of the sort in public. It was quite +against my wishes that she came, but she is a very excitable, +impulsive girl, as you may have noticed, and she is not easily +controlled when she has made up her mind on a point. Of course, I +did not mind you so much, as you are not connected with the +official police, but it is not pleasant to have a family +misfortune like this noised abroad. Besides, it is a useless +expense, for how could you possibly find this Hosmer Angel?" + +"On the contrary," said Holmes quietly; "I have every reason to +believe that I will succeed in discovering Mr. Hosmer Angel." + +Mr. Windibank gave a violent start and dropped his gloves. "I am +delighted to hear it," he said. + +"It is a curious thing," remarked Holmes, "that a typewriter has +really quite as much individuality as a man's handwriting. Unless +they are quite new, no two of them write exactly alike. Some +letters get more worn than others, and some wear only on one +side. Now, you remark in this note of yours, Mr. Windibank, that +in every case there is some little slurring over of the 'e,' and +a slight defect in the tail of the 'r.' There are fourteen other +characteristics, but those are the more obvious." + +"We do all our correspondence with this machine at the office, +and no doubt it is a little worn," our visitor answered, glancing +keenly at Holmes with his bright little eyes. + +"And now I will show you what is really a very interesting study, +Mr. Windibank," Holmes continued. "I think of writing another +little monograph some of these days on the typewriter and its +relation to crime. It is a subject to which I have devoted some +little attention. I have here four letters which purport to come +from the missing man. They are all typewritten. In each case, not +only are the 'e's' slurred and the 'r's' tailless, but you will +observe, if you care to use my magnifying lens, that the fourteen +other characteristics to which I have alluded are there as well." + +Mr. Windibank sprang out of his chair and picked up his hat. "I +cannot waste time over this sort of fantastic talk, Mr. Holmes," +he said. "If you can catch the man, catch him, and let me know +when you have done it." + +"Certainly," said Holmes, stepping over and turning the key in +the door. "I let you know, then, that I have caught him!" + +"What! where?" shouted Mr. Windibank, turning white to his lips +and glancing about him like a rat in a trap. + +"Oh, it won't do--really it won't," said Holmes suavely. "There +is no possible getting out of it, Mr. Windibank. It is quite too +transparent, and it was a very bad compliment when you said that +it was impossible for me to solve so simple a question. That's +right! Sit down and let us talk it over." + +Our visitor collapsed into a chair, with a ghastly face and a +glitter of moisture on his brow. "It--it's not actionable," he +stammered. + +"I am very much afraid that it is not. But between ourselves, +Windibank, it was as cruel and selfish and heartless a trick in a +petty way as ever came before me. Now, let me just run over the +course of events, and you will contradict me if I go wrong." + +The man sat huddled up in his chair, with his head sunk upon his +breast, like one who is utterly crushed. Holmes stuck his feet up +on the corner of the mantelpiece and, leaning back with his hands +in his pockets, began talking, rather to himself, as it seemed, +than to us. + +"The man married a woman very much older than himself for her +money," said he, "and he enjoyed the use of the money of the +daughter as long as she lived with them. It was a considerable +sum, for people in their position, and the loss of it would have +made a serious difference. It was worth an effort to preserve it. +The daughter was of a good, amiable disposition, but affectionate +and warm-hearted in her ways, so that it was evident that with +her fair personal advantages, and her little income, she would +not be allowed to remain single long. Now her marriage would +mean, of course, the loss of a hundred a year, so what does her +stepfather do to prevent it? He takes the obvious course of +keeping her at home and forbidding her to seek the company of +people of her own age. But soon he found that that would not +answer forever. She became restive, insisted upon her rights, and +finally announced her positive intention of going to a certain +ball. What does her clever stepfather do then? He conceives an +idea more creditable to his head than to his heart. With the +connivance and assistance of his wife he disguised himself, +covered those keen eyes with tinted glasses, masked the face with +a moustache and a pair of bushy whiskers, sunk that clear voice +into an insinuating whisper, and doubly secure on account of the +girl's short sight, he appears as Mr. Hosmer Angel, and keeps off +other lovers by making love himself." + +"It was only a joke at first," groaned our visitor. "We never +thought that she would have been so carried away." + +"Very likely not. However that may be, the young lady was very +decidedly carried away, and, having quite made up her mind that +her stepfather was in France, the suspicion of treachery never +for an instant entered her mind. She was flattered by the +gentleman's attentions, and the effect was increased by the +loudly expressed admiration of her mother. Then Mr. Angel began +to call, for it was obvious that the matter should be pushed as +far as it would go if a real effect were to be produced. There +were meetings, and an engagement, which would finally secure the +girl's affections from turning towards anyone else. But the +deception could not be kept up forever. These pretended journeys +to France were rather cumbrous. The thing to do was clearly to +bring the business to an end in such a dramatic manner that it +would leave a permanent impression upon the young lady's mind and +prevent her from looking upon any other suitor for some time to +come. Hence those vows of fidelity exacted upon a Testament, and +hence also the allusions to a possibility of something happening +on the very morning of the wedding. James Windibank wished Miss +Sutherland to be so bound to Hosmer Angel, and so uncertain as to +his fate, that for ten years to come, at any rate, she would not +listen to another man. As far as the church door he brought her, +and then, as he could go no farther, he conveniently vanished +away by the old trick of stepping in at one door of a +four-wheeler and out at the other. I think that was the chain of +events, Mr. Windibank!" + +Our visitor had recovered something of his assurance while Holmes +had been talking, and he rose from his chair now with a cold +sneer upon his pale face. + +"It may be so, or it may not, Mr. Holmes," said he, "but if you +are so very sharp you ought to be sharp enough to know that it is +you who are breaking the law now, and not me. I have done nothing +actionable from the first, but as long as you keep that door +locked you lay yourself open to an action for assault and illegal +constraint." + +"The law cannot, as you say, touch you," said Holmes, unlocking +and throwing open the door, "yet there never was a man who +deserved punishment more. If the young lady has a brother or a +friend, he ought to lay a whip across your shoulders. By Jove!" +he continued, flushing up at the sight of the bitter sneer upon +the man's face, "it is not part of my duties to my client, but +here's a hunting crop handy, and I think I shall just treat +myself to--" He took two swift steps to the whip, but before he +could grasp it there was a wild clatter of steps upon the stairs, +the heavy hall door banged, and from the window we could see Mr. +James Windibank running at the top of his speed down the road. + +"There's a cold-blooded scoundrel!" said Holmes, laughing, as he +threw himself down into his chair once more. "That fellow will +rise from crime to crime until he does something very bad, and +ends on a gallows. The case has, in some respects, been not +entirely devoid of interest." + +"I cannot now entirely see all the steps of your reasoning," I +remarked. + +"Well, of course it was obvious from the first that this Mr. +Hosmer Angel must have some strong object for his curious +conduct, and it was equally clear that the only man who really +profited by the incident, as far as we could see, was the +stepfather. Then the fact that the two men were never together, +but that the one always appeared when the other was away, was +suggestive. So were the tinted spectacles and the curious voice, +which both hinted at a disguise, as did the bushy whiskers. My +suspicions were all confirmed by his peculiar action in +typewriting his signature, which, of course, inferred that his +handwriting was so familiar to her that she would recognise even +the smallest sample of it. You see all these isolated facts, +together with many minor ones, all pointed in the same +direction." + +"And how did you verify them?" + +"Having once spotted my man, it was easy to get corroboration. I +knew the firm for which this man worked. Having taken the printed +description. I eliminated everything from it which could be the +result of a disguise--the whiskers, the glasses, the voice, and I +sent it to the firm, with a request that they would inform me +whether it answered to the description of any of their +travellers. I had already noticed the peculiarities of the +typewriter, and I wrote to the man himself at his business +address asking him if he would come here. As I expected, his +reply was typewritten and revealed the same trivial but +characteristic defects. The same post brought me a letter from +Westhouse & Marbank, of Fenchurch Street, to say that the +description tallied in every respect with that of their employé, +James Windibank. Voilà tout!" + +"And Miss Sutherland?" + +"If I tell her she will not believe me. You may remember the old +Persian saying, 'There is danger for him who taketh the tiger +cub, and danger also for whoso snatches a delusion from a woman.' +There is as much sense in Hafiz as in Horace, and as much +knowledge of the world." + + + +ADVENTURE IV. THE BOSCOMBE VALLEY MYSTERY + +We were seated at breakfast one morning, my wife and I, when the +maid brought in a telegram. It was from Sherlock Holmes and ran +in this way: + +"Have you a couple of days to spare? Have just been wired for from +the west of England in connection with Boscombe Valley tragedy. +Shall be glad if you will come with me. Air and scenery perfect. +Leave Paddington by the 11:15." + +"What do you say, dear?" said my wife, looking across at me. +"Will you go?" + +"I really don't know what to say. I have a fairly long list at +present." + +"Oh, Anstruther would do your work for you. You have been looking +a little pale lately. I think that the change would do you good, +and you are always so interested in Mr. Sherlock Holmes' cases." + +"I should be ungrateful if I were not, seeing what I gained +through one of them," I answered. "But if I am to go, I must pack +at once, for I have only half an hour." + +My experience of camp life in Afghanistan had at least had the +effect of making me a prompt and ready traveller. My wants were +few and simple, so that in less than the time stated I was in a +cab with my valise, rattling away to Paddington Station. Sherlock +Holmes was pacing up and down the platform, his tall, gaunt +figure made even gaunter and taller by his long grey +travelling-cloak and close-fitting cloth cap. + +"It is really very good of you to come, Watson," said he. "It +makes a considerable difference to me, having someone with me on +whom I can thoroughly rely. Local aid is always either worthless +or else biassed. If you will keep the two corner seats I shall +get the tickets." + +We had the carriage to ourselves save for an immense litter of +papers which Holmes had brought with him. Among these he rummaged +and read, with intervals of note-taking and of meditation, until +we were past Reading. Then he suddenly rolled them all into a +gigantic ball and tossed them up onto the rack. + +"Have you heard anything of the case?" he asked. + +"Not a word. I have not seen a paper for some days." + +"The London press has not had very full accounts. I have just +been looking through all the recent papers in order to master the +particulars. It seems, from what I gather, to be one of those +simple cases which are so extremely difficult." + +"That sounds a little paradoxical." + +"But it is profoundly true. Singularity is almost invariably a +clue. The more featureless and commonplace a crime is, the more +difficult it is to bring it home. In this case, however, they +have established a very serious case against the son of the +murdered man." + +"It is a murder, then?" + +"Well, it is conjectured to be so. I shall take nothing for +granted until I have the opportunity of looking personally into +it. I will explain the state of things to you, as far as I have +been able to understand it, in a very few words. + +"Boscombe Valley is a country district not very far from Ross, in +Herefordshire. The largest landed proprietor in that part is a +Mr. John Turner, who made his money in Australia and returned +some years ago to the old country. One of the farms which he +held, that of Hatherley, was let to Mr. Charles McCarthy, who was +also an ex-Australian. The men had known each other in the +colonies, so that it was not unnatural that when they came to +settle down they should do so as near each other as possible. +Turner was apparently the richer man, so McCarthy became his +tenant but still remained, it seems, upon terms of perfect +equality, as they were frequently together. McCarthy had one son, +a lad of eighteen, and Turner had an only daughter of the same +age, but neither of them had wives living. They appear to have +avoided the society of the neighbouring English families and to +have led retired lives, though both the McCarthys were fond of +sport and were frequently seen at the race-meetings of the +neighbourhood. McCarthy kept two servants--a man and a girl. +Turner had a considerable household, some half-dozen at the +least. That is as much as I have been able to gather about the +families. Now for the facts. + +"On June 3rd, that is, on Monday last, McCarthy left his house at +Hatherley about three in the afternoon and walked down to the +Boscombe Pool, which is a small lake formed by the spreading out +of the stream which runs down the Boscombe Valley. He had been +out with his serving-man in the morning at Ross, and he had told +the man that he must hurry, as he had an appointment of +importance to keep at three. From that appointment he never came +back alive. + +"From Hatherley Farm-house to the Boscombe Pool is a quarter of a +mile, and two people saw him as he passed over this ground. One +was an old woman, whose name is not mentioned, and the other was +William Crowder, a game-keeper in the employ of Mr. Turner. Both +these witnesses depose that Mr. McCarthy was walking alone. The +game-keeper adds that within a few minutes of his seeing Mr. +McCarthy pass he had seen his son, Mr. James McCarthy, going the +same way with a gun under his arm. To the best of his belief, the +father was actually in sight at the time, and the son was +following him. He thought no more of the matter until he heard in +the evening of the tragedy that had occurred. + +"The two McCarthys were seen after the time when William Crowder, +the game-keeper, lost sight of them. The Boscombe Pool is thickly +wooded round, with just a fringe of grass and of reeds round the +edge. A girl of fourteen, Patience Moran, who is the daughter of +the lodge-keeper of the Boscombe Valley estate, was in one of the +woods picking flowers. She states that while she was there she +saw, at the border of the wood and close by the lake, Mr. +McCarthy and his son, and that they appeared to be having a +violent quarrel. She heard Mr. McCarthy the elder using very +strong language to his son, and she saw the latter raise up his +hand as if to strike his father. She was so frightened by their +violence that she ran away and told her mother when she reached +home that she had left the two McCarthys quarrelling near +Boscombe Pool, and that she was afraid that they were going to +fight. She had hardly said the words when young Mr. McCarthy came +running up to the lodge to say that he had found his father dead +in the wood, and to ask for the help of the lodge-keeper. He was +much excited, without either his gun or his hat, and his right +hand and sleeve were observed to be stained with fresh blood. On +following him they found the dead body stretched out upon the +grass beside the pool. The head had been beaten in by repeated +blows of some heavy and blunt weapon. The injuries were such as +might very well have been inflicted by the butt-end of his son's +gun, which was found lying on the grass within a few paces of the +body. Under these circumstances the young man was instantly +arrested, and a verdict of 'wilful murder' having been returned +at the inquest on Tuesday, he was on Wednesday brought before the +magistrates at Ross, who have referred the case to the next +Assizes. Those are the main facts of the case as they came out +before the coroner and the police-court." + +"I could hardly imagine a more damning case," I remarked. "If +ever circumstantial evidence pointed to a criminal it does so +here." + +"Circumstantial evidence is a very tricky thing," answered Holmes +thoughtfully. "It may seem to point very straight to one thing, +but if you shift your own point of view a little, you may find it +pointing in an equally uncompromising manner to something +entirely different. It must be confessed, however, that the case +looks exceedingly grave against the young man, and it is very +possible that he is indeed the culprit. There are several people +in the neighbourhood, however, and among them Miss Turner, the +daughter of the neighbouring landowner, who believe in his +innocence, and who have retained Lestrade, whom you may recollect +in connection with the Study in Scarlet, to work out the case in +his interest. Lestrade, being rather puzzled, has referred the +case to me, and hence it is that two middle-aged gentlemen are +flying westward at fifty miles an hour instead of quietly +digesting their breakfasts at home." + +"I am afraid," said I, "that the facts are so obvious that you +will find little credit to be gained out of this case." + +"There is nothing more deceptive than an obvious fact," he +answered, laughing. "Besides, we may chance to hit upon some +other obvious facts which may have been by no means obvious to +Mr. Lestrade. You know me too well to think that I am boasting +when I say that I shall either confirm or destroy his theory by +means which he is quite incapable of employing, or even of +understanding. To take the first example to hand, I very clearly +perceive that in your bedroom the window is upon the right-hand +side, and yet I question whether Mr. Lestrade would have noted +even so self-evident a thing as that." + +"How on earth--" + +"My dear fellow, I know you well. I know the military neatness +which characterises you. You shave every morning, and in this +season you shave by the sunlight; but since your shaving is less +and less complete as we get farther back on the left side, until +it becomes positively slovenly as we get round the angle of the +jaw, it is surely very clear that that side is less illuminated +than the other. I could not imagine a man of your habits looking +at himself in an equal light and being satisfied with such a +result. I only quote this as a trivial example of observation and +inference. Therein lies my métier, and it is just possible that +it may be of some service in the investigation which lies before +us. There are one or two minor points which were brought out in +the inquest, and which are worth considering." + +"What are they?" + +"It appears that his arrest did not take place at once, but after +the return to Hatherley Farm. On the inspector of constabulary +informing him that he was a prisoner, he remarked that he was not +surprised to hear it, and that it was no more than his deserts. +This observation of his had the natural effect of removing any +traces of doubt which might have remained in the minds of the +coroner's jury." + +"It was a confession," I ejaculated. + +"No, for it was followed by a protestation of innocence." + +"Coming on the top of such a damning series of events, it was at +least a most suspicious remark." + +"On the contrary," said Holmes, "it is the brightest rift which I +can at present see in the clouds. However innocent he might be, +he could not be such an absolute imbecile as not to see that the +circumstances were very black against him. Had he appeared +surprised at his own arrest, or feigned indignation at it, I +should have looked upon it as highly suspicious, because such +surprise or anger would not be natural under the circumstances, +and yet might appear to be the best policy to a scheming man. His +frank acceptance of the situation marks him as either an innocent +man, or else as a man of considerable self-restraint and +firmness. As to his remark about his deserts, it was also not +unnatural if you consider that he stood beside the dead body of +his father, and that there is no doubt that he had that very day +so far forgotten his filial duty as to bandy words with him, and +even, according to the little girl whose evidence is so +important, to raise his hand as if to strike him. The +self-reproach and contrition which are displayed in his remark +appear to me to be the signs of a healthy mind rather than of a +guilty one." + +I shook my head. "Many men have been hanged on far slighter +evidence," I remarked. + +"So they have. And many men have been wrongfully hanged." + +"What is the young man's own account of the matter?" + +"It is, I am afraid, not very encouraging to his supporters, +though there are one or two points in it which are suggestive. +You will find it here, and may read it for yourself." + +He picked out from his bundle a copy of the local Herefordshire +paper, and having turned down the sheet he pointed out the +paragraph in which the unfortunate young man had given his own +statement of what had occurred. I settled myself down in the +corner of the carriage and read it very carefully. It ran in this +way: + +"Mr. James McCarthy, the only son of the deceased, was then called +and gave evidence as follows: 'I had been away from home for +three days at Bristol, and had only just returned upon the +morning of last Monday, the 3rd. My father was absent from home at +the time of my arrival, and I was informed by the maid that he +had driven over to Ross with John Cobb, the groom. Shortly after +my return I heard the wheels of his trap in the yard, and, +looking out of my window, I saw him get out and walk rapidly out +of the yard, though I was not aware in which direction he was +going. I then took my gun and strolled out in the direction of +the Boscombe Pool, with the intention of visiting the rabbit +warren which is upon the other side. On my way I saw William +Crowder, the game-keeper, as he had stated in his evidence; but +he is mistaken in thinking that I was following my father. I had +no idea that he was in front of me. When about a hundred yards +from the pool I heard a cry of "Cooee!" which was a usual signal +between my father and myself. I then hurried forward, and found +him standing by the pool. He appeared to be much surprised at +seeing me and asked me rather roughly what I was doing there. A +conversation ensued which led to high words and almost to blows, +for my father was a man of a very violent temper. Seeing that his +passion was becoming ungovernable, I left him and returned +towards Hatherley Farm. I had not gone more than 150 yards, +however, when I heard a hideous outcry behind me, which caused me +to run back again. I found my father expiring upon the ground, +with his head terribly injured. I dropped my gun and held him in +my arms, but he almost instantly expired. I knelt beside him for +some minutes, and then made my way to Mr. Turner's lodge-keeper, +his house being the nearest, to ask for assistance. I saw no one +near my father when I returned, and I have no idea how he came by +his injuries. He was not a popular man, being somewhat cold and +forbidding in his manners, but he had, as far as I know, no +active enemies. I know nothing further of the matter.' + +"The Coroner: Did your father make any statement to you before +he died? + +"Witness: He mumbled a few words, but I could only catch some +allusion to a rat. + +"The Coroner: What did you understand by that? + +"Witness: It conveyed no meaning to me. I thought that he was +delirious. + +"The Coroner: What was the point upon which you and your father +had this final quarrel? + +"Witness: I should prefer not to answer. + +"The Coroner: I am afraid that I must press it. + +"Witness: It is really impossible for me to tell you. I can +assure you that it has nothing to do with the sad tragedy which +followed. + +"The Coroner: That is for the court to decide. I need not point +out to you that your refusal to answer will prejudice your case +considerably in any future proceedings which may arise. + +"Witness: I must still refuse. + +"The Coroner: I understand that the cry of 'Cooee' was a common +signal between you and your father? + +"Witness: It was. + +"The Coroner: How was it, then, that he uttered it before he saw +you, and before he even knew that you had returned from Bristol? + +"Witness (with considerable confusion): I do not know. + +"A Juryman: Did you see nothing which aroused your suspicions +when you returned on hearing the cry and found your father +fatally injured? + +"Witness: Nothing definite. + +"The Coroner: What do you mean? + +"Witness: I was so disturbed and excited as I rushed out into +the open, that I could think of nothing except of my father. Yet +I have a vague impression that as I ran forward something lay +upon the ground to the left of me. It seemed to me to be +something grey in colour, a coat of some sort, or a plaid perhaps. +When I rose from my father I looked round for it, but it was +gone. + +"'Do you mean that it disappeared before you went for help?' + +"'Yes, it was gone.' + +"'You cannot say what it was?' + +"'No, I had a feeling something was there.' + +"'How far from the body?' + +"'A dozen yards or so.' + +"'And how far from the edge of the wood?' + +"'About the same.' + +"'Then if it was removed it was while you were within a dozen +yards of it?' + +"'Yes, but with my back towards it.' + +"This concluded the examination of the witness." + +"I see," said I as I glanced down the column, "that the coroner +in his concluding remarks was rather severe upon young McCarthy. +He calls attention, and with reason, to the discrepancy about his +father having signalled to him before seeing him, also to his +refusal to give details of his conversation with his father, and +his singular account of his father's dying words. They are all, +as he remarks, very much against the son." + +Holmes laughed softly to himself and stretched himself out upon +the cushioned seat. "Both you and the coroner have been at some +pains," said he, "to single out the very strongest points in the +young man's favour. Don't you see that you alternately give him +credit for having too much imagination and too little? Too +little, if he could not invent a cause of quarrel which would +give him the sympathy of the jury; too much, if he evolved from +his own inner consciousness anything so outré as a dying +reference to a rat, and the incident of the vanishing cloth. No, +sir, I shall approach this case from the point of view that what +this young man says is true, and we shall see whither that +hypothesis will lead us. And now here is my pocket Petrarch, and +not another word shall I say of this case until we are on the +scene of action. We lunch at Swindon, and I see that we shall be +there in twenty minutes." + +It was nearly four o'clock when we at last, after passing through +the beautiful Stroud Valley, and over the broad gleaming Severn, +found ourselves at the pretty little country-town of Ross. A +lean, ferret-like man, furtive and sly-looking, was waiting for +us upon the platform. In spite of the light brown dustcoat and +leather-leggings which he wore in deference to his rustic +surroundings, I had no difficulty in recognising Lestrade, of +Scotland Yard. With him we drove to the Hereford Arms where a +room had already been engaged for us. + +"I have ordered a carriage," said Lestrade as we sat over a cup +of tea. "I knew your energetic nature, and that you would not be +happy until you had been on the scene of the crime." + +"It was very nice and complimentary of you," Holmes answered. "It +is entirely a question of barometric pressure." + +Lestrade looked startled. "I do not quite follow," he said. + +"How is the glass? Twenty-nine, I see. No wind, and not a cloud +in the sky. I have a caseful of cigarettes here which need +smoking, and the sofa is very much superior to the usual country +hotel abomination. I do not think that it is probable that I +shall use the carriage to-night." + +Lestrade laughed indulgently. "You have, no doubt, already formed +your conclusions from the newspapers," he said. "The case is as +plain as a pikestaff, and the more one goes into it the plainer +it becomes. Still, of course, one can't refuse a lady, and such a +very positive one, too. She has heard of you, and would have your +opinion, though I repeatedly told her that there was nothing +which you could do which I had not already done. Why, bless my +soul! here is her carriage at the door." + +He had hardly spoken before there rushed into the room one of the +most lovely young women that I have ever seen in my life. Her +violet eyes shining, her lips parted, a pink flush upon her +cheeks, all thought of her natural reserve lost in her +overpowering excitement and concern. + +"Oh, Mr. Sherlock Holmes!" she cried, glancing from one to the +other of us, and finally, with a woman's quick intuition, +fastening upon my companion, "I am so glad that you have come. I +have driven down to tell you so. I know that James didn't do it. +I know it, and I want you to start upon your work knowing it, +too. Never let yourself doubt upon that point. We have known each +other since we were little children, and I know his faults as no +one else does; but he is too tender-hearted to hurt a fly. Such a +charge is absurd to anyone who really knows him." + +"I hope we may clear him, Miss Turner," said Sherlock Holmes. +"You may rely upon my doing all that I can." + +"But you have read the evidence. You have formed some conclusion? +Do you not see some loophole, some flaw? Do you not yourself +think that he is innocent?" + +"I think that it is very probable." + +"There, now!" she cried, throwing back her head and looking +defiantly at Lestrade. "You hear! He gives me hopes." + +Lestrade shrugged his shoulders. "I am afraid that my colleague +has been a little quick in forming his conclusions," he said. + +"But he is right. Oh! I know that he is right. James never did +it. And about his quarrel with his father, I am sure that the +reason why he would not speak about it to the coroner was because +I was concerned in it." + +"In what way?" asked Holmes. + +"It is no time for me to hide anything. James and his father had +many disagreements about me. Mr. McCarthy was very anxious that +there should be a marriage between us. James and I have always +loved each other as brother and sister; but of course he is young +and has seen very little of life yet, and--and--well, he +naturally did not wish to do anything like that yet. So there +were quarrels, and this, I am sure, was one of them." + +"And your father?" asked Holmes. "Was he in favour of such a +union?" + +"No, he was averse to it also. No one but Mr. McCarthy was in +favour of it." A quick blush passed over her fresh young face as +Holmes shot one of his keen, questioning glances at her. + +"Thank you for this information," said he. "May I see your father +if I call to-morrow?" + +"I am afraid the doctor won't allow it." + +"The doctor?" + +"Yes, have you not heard? Poor father has never been strong for +years back, but this has broken him down completely. He has taken +to his bed, and Dr. Willows says that he is a wreck and that his +nervous system is shattered. Mr. McCarthy was the only man alive +who had known dad in the old days in Victoria." + +"Ha! In Victoria! That is important." + +"Yes, at the mines." + +"Quite so; at the gold-mines, where, as I understand, Mr. Turner +made his money." + +"Yes, certainly." + +"Thank you, Miss Turner. You have been of material assistance to +me." + +"You will tell me if you have any news to-morrow. No doubt you +will go to the prison to see James. Oh, if you do, Mr. Holmes, do +tell him that I know him to be innocent." + +"I will, Miss Turner." + +"I must go home now, for dad is very ill, and he misses me so if +I leave him. Good-bye, and God help you in your undertaking." She +hurried from the room as impulsively as she had entered, and we +heard the wheels of her carriage rattle off down the street. + +"I am ashamed of you, Holmes," said Lestrade with dignity after a +few minutes' silence. "Why should you raise up hopes which you +are bound to disappoint? I am not over-tender of heart, but I +call it cruel." + +"I think that I see my way to clearing James McCarthy," said +Holmes. "Have you an order to see him in prison?" + +"Yes, but only for you and me." + +"Then I shall reconsider my resolution about going out. We have +still time to take a train to Hereford and see him to-night?" + +"Ample." + +"Then let us do so. Watson, I fear that you will find it very +slow, but I shall only be away a couple of hours." + +I walked down to the station with them, and then wandered through +the streets of the little town, finally returning to the hotel, +where I lay upon the sofa and tried to interest myself in a +yellow-backed novel. The puny plot of the story was so thin, +however, when compared to the deep mystery through which we were +groping, and I found my attention wander so continually from the +action to the fact, that I at last flung it across the room and +gave myself up entirely to a consideration of the events of the +day. Supposing that this unhappy young man's story were +absolutely true, then what hellish thing, what absolutely +unforeseen and extraordinary calamity could have occurred between +the time when he parted from his father, and the moment when, +drawn back by his screams, he rushed into the glade? It was +something terrible and deadly. What could it be? Might not the +nature of the injuries reveal something to my medical instincts? +I rang the bell and called for the weekly county paper, which +contained a verbatim account of the inquest. In the surgeon's +deposition it was stated that the posterior third of the left +parietal bone and the left half of the occipital bone had been +shattered by a heavy blow from a blunt weapon. I marked the spot +upon my own head. Clearly such a blow must have been struck from +behind. That was to some extent in favour of the accused, as when +seen quarrelling he was face to face with his father. Still, it +did not go for very much, for the older man might have turned his +back before the blow fell. Still, it might be worth while to call +Holmes' attention to it. Then there was the peculiar dying +reference to a rat. What could that mean? It could not be +delirium. A man dying from a sudden blow does not commonly become +delirious. No, it was more likely to be an attempt to explain how +he met his fate. But what could it indicate? I cudgelled my +brains to find some possible explanation. And then the incident +of the grey cloth seen by young McCarthy. If that were true the +murderer must have dropped some part of his dress, presumably his +overcoat, in his flight, and must have had the hardihood to +return and to carry it away at the instant when the son was +kneeling with his back turned not a dozen paces off. What a +tissue of mysteries and improbabilities the whole thing was! I +did not wonder at Lestrade's opinion, and yet I had so much faith +in Sherlock Holmes' insight that I could not lose hope as long +as every fresh fact seemed to strengthen his conviction of young +McCarthy's innocence. + +It was late before Sherlock Holmes returned. He came back alone, +for Lestrade was staying in lodgings in the town. + +"The glass still keeps very high," he remarked as he sat down. +"It is of importance that it should not rain before we are able +to go over the ground. On the other hand, a man should be at his +very best and keenest for such nice work as that, and I did not +wish to do it when fagged by a long journey. I have seen young +McCarthy." + +"And what did you learn from him?" + +"Nothing." + +"Could he throw no light?" + +"None at all. I was inclined to think at one time that he knew +who had done it and was screening him or her, but I am convinced +now that he is as puzzled as everyone else. He is not a very +quick-witted youth, though comely to look at and, I should think, +sound at heart." + +"I cannot admire his taste," I remarked, "if it is indeed a fact +that he was averse to a marriage with so charming a young lady as +this Miss Turner." + +"Ah, thereby hangs a rather painful tale. This fellow is madly, +insanely, in love with her, but some two years ago, when he was +only a lad, and before he really knew her, for she had been away +five years at a boarding-school, what does the idiot do but get +into the clutches of a barmaid in Bristol and marry her at a +registry office? No one knows a word of the matter, but you can +imagine how maddening it must be to him to be upbraided for not +doing what he would give his very eyes to do, but what he knows +to be absolutely impossible. It was sheer frenzy of this sort +which made him throw his hands up into the air when his father, +at their last interview, was goading him on to propose to Miss +Turner. On the other hand, he had no means of supporting himself, +and his father, who was by all accounts a very hard man, would +have thrown him over utterly had he known the truth. It was with +his barmaid wife that he had spent the last three days in +Bristol, and his father did not know where he was. Mark that +point. It is of importance. Good has come out of evil, however, +for the barmaid, finding from the papers that he is in serious +trouble and likely to be hanged, has thrown him over utterly and +has written to him to say that she has a husband already in the +Bermuda Dockyard, so that there is really no tie between them. I +think that that bit of news has consoled young McCarthy for all +that he has suffered." + +"But if he is innocent, who has done it?" + +"Ah! who? I would call your attention very particularly to two +points. One is that the murdered man had an appointment with +someone at the pool, and that the someone could not have been his +son, for his son was away, and he did not know when he would +return. The second is that the murdered man was heard to cry +'Cooee!' before he knew that his son had returned. Those are the +crucial points upon which the case depends. And now let us talk +about George Meredith, if you please, and we shall leave all +minor matters until to-morrow." + +There was no rain, as Holmes had foretold, and the morning broke +bright and cloudless. At nine o'clock Lestrade called for us with +the carriage, and we set off for Hatherley Farm and the Boscombe +Pool. + +"There is serious news this morning," Lestrade observed. "It is +said that Mr. Turner, of the Hall, is so ill that his life is +despaired of." + +"An elderly man, I presume?" said Holmes. + +"About sixty; but his constitution has been shattered by his life +abroad, and he has been in failing health for some time. This +business has had a very bad effect upon him. He was an old friend +of McCarthy's, and, I may add, a great benefactor to him, for I +have learned that he gave him Hatherley Farm rent free." + +"Indeed! That is interesting," said Holmes. + +"Oh, yes! In a hundred other ways he has helped him. Everybody +about here speaks of his kindness to him." + +"Really! Does it not strike you as a little singular that this +McCarthy, who appears to have had little of his own, and to have +been under such obligations to Turner, should still talk of +marrying his son to Turner's daughter, who is, presumably, +heiress to the estate, and that in such a very cocksure manner, +as if it were merely a case of a proposal and all else would +follow? It is the more strange, since we know that Turner himself +was averse to the idea. The daughter told us as much. Do you not +deduce something from that?" + +"We have got to the deductions and the inferences," said +Lestrade, winking at me. "I find it hard enough to tackle facts, +Holmes, without flying away after theories and fancies." + +"You are right," said Holmes demurely; "you do find it very hard +to tackle the facts." + +"Anyhow, I have grasped one fact which you seem to find it +difficult to get hold of," replied Lestrade with some warmth. + +"And that is--" + +"That McCarthy senior met his death from McCarthy junior and that +all theories to the contrary are the merest moonshine." + +"Well, moonshine is a brighter thing than fog," said Holmes, +laughing. "But I am very much mistaken if this is not Hatherley +Farm upon the left." + +"Yes, that is it." It was a widespread, comfortable-looking +building, two-storied, slate-roofed, with great yellow blotches +of lichen upon the grey walls. The drawn blinds and the smokeless +chimneys, however, gave it a stricken look, as though the weight +of this horror still lay heavy upon it. We called at the door, +when the maid, at Holmes' request, showed us the boots which her +master wore at the time of his death, and also a pair of the +son's, though not the pair which he had then had. Having measured +these very carefully from seven or eight different points, Holmes +desired to be led to the court-yard, from which we all followed +the winding track which led to Boscombe Pool. + +Sherlock Holmes was transformed when he was hot upon such a scent +as this. Men who had only known the quiet thinker and logician of +Baker Street would have failed to recognise him. His face flushed +and darkened. His brows were drawn into two hard black lines, +while his eyes shone out from beneath them with a steely glitter. +His face was bent downward, his shoulders bowed, his lips +compressed, and the veins stood out like whipcord in his long, +sinewy neck. His nostrils seemed to dilate with a purely animal +lust for the chase, and his mind was so absolutely concentrated +upon the matter before him that a question or remark fell +unheeded upon his ears, or, at the most, only provoked a quick, +impatient snarl in reply. Swiftly and silently he made his way +along the track which ran through the meadows, and so by way of +the woods to the Boscombe Pool. It was damp, marshy ground, as is +all that district, and there were marks of many feet, both upon +the path and amid the short grass which bounded it on either +side. Sometimes Holmes would hurry on, sometimes stop dead, and +once he made quite a little detour into the meadow. Lestrade and +I walked behind him, the detective indifferent and contemptuous, +while I watched my friend with the interest which sprang from the +conviction that every one of his actions was directed towards a +definite end. + +The Boscombe Pool, which is a little reed-girt sheet of water +some fifty yards across, is situated at the boundary between the +Hatherley Farm and the private park of the wealthy Mr. Turner. +Above the woods which lined it upon the farther side we could see +the red, jutting pinnacles which marked the site of the rich +landowner's dwelling. On the Hatherley side of the pool the woods +grew very thick, and there was a narrow belt of sodden grass +twenty paces across between the edge of the trees and the reeds +which lined the lake. Lestrade showed us the exact spot at which +the body had been found, and, indeed, so moist was the ground, +that I could plainly see the traces which had been left by the +fall of the stricken man. To Holmes, as I could see by his eager +face and peering eyes, very many other things were to be read +upon the trampled grass. He ran round, like a dog who is picking +up a scent, and then turned upon my companion. + +"What did you go into the pool for?" he asked. + +"I fished about with a rake. I thought there might be some weapon +or other trace. But how on earth--" + +"Oh, tut, tut! I have no time! That left foot of yours with its +inward twist is all over the place. A mole could trace it, and +there it vanishes among the reeds. Oh, how simple it would all +have been had I been here before they came like a herd of buffalo +and wallowed all over it. Here is where the party with the +lodge-keeper came, and they have covered all tracks for six or +eight feet round the body. But here are three separate tracks of +the same feet." He drew out a lens and lay down upon his +waterproof to have a better view, talking all the time rather to +himself than to us. "These are young McCarthy's feet. Twice he +was walking, and once he ran swiftly, so that the soles are +deeply marked and the heels hardly visible. That bears out his +story. He ran when he saw his father on the ground. Then here are +the father's feet as he paced up and down. What is this, then? It +is the butt-end of the gun as the son stood listening. And this? +Ha, ha! What have we here? Tiptoes! tiptoes! Square, too, quite +unusual boots! They come, they go, they come again--of course +that was for the cloak. Now where did they come from?" He ran up +and down, sometimes losing, sometimes finding the track until we +were well within the edge of the wood and under the shadow of a +great beech, the largest tree in the neighbourhood. Holmes traced +his way to the farther side of this and lay down once more upon +his face with a little cry of satisfaction. For a long time he +remained there, turning over the leaves and dried sticks, +gathering up what seemed to me to be dust into an envelope and +examining with his lens not only the ground but even the bark of +the tree as far as he could reach. A jagged stone was lying among +the moss, and this also he carefully examined and retained. Then +he followed a pathway through the wood until he came to the +highroad, where all traces were lost. + +"It has been a case of considerable interest," he remarked, +returning to his natural manner. "I fancy that this grey house on +the right must be the lodge. I think that I will go in and have a +word with Moran, and perhaps write a little note. Having done +that, we may drive back to our luncheon. You may walk to the cab, +and I shall be with you presently." + +It was about ten minutes before we regained our cab and drove +back into Ross, Holmes still carrying with him the stone which he +had picked up in the wood. + +"This may interest you, Lestrade," he remarked, holding it out. +"The murder was done with it." + +"I see no marks." + +"There are none." + +"How do you know, then?" + +"The grass was growing under it. It had only lain there a few +days. There was no sign of a place whence it had been taken. It +corresponds with the injuries. There is no sign of any other +weapon." + +"And the murderer?" + +"Is a tall man, left-handed, limps with the right leg, wears +thick-soled shooting-boots and a grey cloak, smokes Indian +cigars, uses a cigar-holder, and carries a blunt pen-knife in his +pocket. There are several other indications, but these may be +enough to aid us in our search." + +Lestrade laughed. "I am afraid that I am still a sceptic," he +said. "Theories are all very well, but we have to deal with a +hard-headed British jury." + +"Nous verrons," answered Holmes calmly. "You work your own +method, and I shall work mine. I shall be busy this afternoon, +and shall probably return to London by the evening train." + +"And leave your case unfinished?" + +"No, finished." + +"But the mystery?" + +"It is solved." + +"Who was the criminal, then?" + +"The gentleman I describe." + +"But who is he?" + +"Surely it would not be difficult to find out. This is not such a +populous neighbourhood." + +Lestrade shrugged his shoulders. "I am a practical man," he said, +"and I really cannot undertake to go about the country looking +for a left-handed gentleman with a game leg. I should become the +laughing-stock of Scotland Yard." + +"All right," said Holmes quietly. "I have given you the chance. +Here are your lodgings. Good-bye. I shall drop you a line before +I leave." + +Having left Lestrade at his rooms, we drove to our hotel, where +we found lunch upon the table. Holmes was silent and buried in +thought with a pained expression upon his face, as one who finds +himself in a perplexing position. + +"Look here, Watson," he said when the cloth was cleared "just sit +down in this chair and let me preach to you for a little. I don't +know quite what to do, and I should value your advice. Light a +cigar and let me expound." + + "Pray do so." + +"Well, now, in considering this case there are two points about +young McCarthy's narrative which struck us both instantly, +although they impressed me in his favour and you against him. One +was the fact that his father should, according to his account, +cry 'Cooee!' before seeing him. The other was his singular dying +reference to a rat. He mumbled several words, you understand, but +that was all that caught the son's ear. Now from this double +point our research must commence, and we will begin it by +presuming that what the lad says is absolutely true." + +"What of this 'Cooee!' then?" + +"Well, obviously it could not have been meant for the son. The +son, as far as he knew, was in Bristol. It was mere chance that +he was within earshot. The 'Cooee!' was meant to attract the +attention of whoever it was that he had the appointment with. But +'Cooee' is a distinctly Australian cry, and one which is used +between Australians. There is a strong presumption that the +person whom McCarthy expected to meet him at Boscombe Pool was +someone who had been in Australia." + +"What of the rat, then?" + +Sherlock Holmes took a folded paper from his pocket and flattened +it out on the table. "This is a map of the Colony of Victoria," +he said. "I wired to Bristol for it last night." He put his hand +over part of the map. "What do you read?" + +"ARAT," I read. + +"And now?" He raised his hand. + +"BALLARAT." + +"Quite so. That was the word the man uttered, and of which his +son only caught the last two syllables. He was trying to utter +the name of his murderer. So and so, of Ballarat." + +"It is wonderful!" I exclaimed. + +"It is obvious. And now, you see, I had narrowed the field down +considerably. The possession of a grey garment was a third point +which, granting the son's statement to be correct, was a +certainty. We have come now out of mere vagueness to the definite +conception of an Australian from Ballarat with a grey cloak." + +"Certainly." + +"And one who was at home in the district, for the pool can only +be approached by the farm or by the estate, where strangers could +hardly wander." + +"Quite so." + +"Then comes our expedition of to-day. By an examination of the +ground I gained the trifling details which I gave to that +imbecile Lestrade, as to the personality of the criminal." + +"But how did you gain them?" + +"You know my method. It is founded upon the observation of +trifles." + +"His height I know that you might roughly judge from the length +of his stride. His boots, too, might be told from their traces." + +"Yes, they were peculiar boots." + +"But his lameness?" + +"The impression of his right foot was always less distinct than +his left. He put less weight upon it. Why? Because he limped--he +was lame." + +"But his left-handedness." + +"You were yourself struck by the nature of the injury as recorded +by the surgeon at the inquest. The blow was struck from +immediately behind, and yet was upon the left side. Now, how can +that be unless it were by a left-handed man? He had stood behind +that tree during the interview between the father and son. He had +even smoked there. I found the ash of a cigar, which my special +knowledge of tobacco ashes enables me to pronounce as an Indian +cigar. I have, as you know, devoted some attention to this, and +written a little monograph on the ashes of 140 different +varieties of pipe, cigar, and cigarette tobacco. Having found the +ash, I then looked round and discovered the stump among the moss +where he had tossed it. It was an Indian cigar, of the variety +which are rolled in Rotterdam." + +"And the cigar-holder?" + +"I could see that the end had not been in his mouth. Therefore he +used a holder. The tip had been cut off, not bitten off, but the +cut was not a clean one, so I deduced a blunt pen-knife." + +"Holmes," I said, "you have drawn a net round this man from which +he cannot escape, and you have saved an innocent human life as +truly as if you had cut the cord which was hanging him. I see the +direction in which all this points. The culprit is--" + +"Mr. John Turner," cried the hotel waiter, opening the door of +our sitting-room, and ushering in a visitor. + +The man who entered was a strange and impressive figure. His +slow, limping step and bowed shoulders gave the appearance of +decrepitude, and yet his hard, deep-lined, craggy features, and +his enormous limbs showed that he was possessed of unusual +strength of body and of character. His tangled beard, grizzled +hair, and outstanding, drooping eyebrows combined to give an air +of dignity and power to his appearance, but his face was of an +ashen white, while his lips and the corners of his nostrils were +tinged with a shade of blue. It was clear to me at a glance that +he was in the grip of some deadly and chronic disease. + +"Pray sit down on the sofa," said Holmes gently. "You had my +note?" + +"Yes, the lodge-keeper brought it up. You said that you wished to +see me here to avoid scandal." + +"I thought people would talk if I went to the Hall." + +"And why did you wish to see me?" He looked across at my +companion with despair in his weary eyes, as though his question +was already answered. + +"Yes," said Holmes, answering the look rather than the words. "It +is so. I know all about McCarthy." + +The old man sank his face in his hands. "God help me!" he cried. +"But I would not have let the young man come to harm. I give you +my word that I would have spoken out if it went against him at +the Assizes." + +"I am glad to hear you say so," said Holmes gravely. + +"I would have spoken now had it not been for my dear girl. It +would break her heart--it will break her heart when she hears +that I am arrested." + +"It may not come to that," said Holmes. + +"What?" + +"I am no official agent. I understand that it was your daughter +who required my presence here, and I am acting in her interests. +Young McCarthy must be got off, however." + +"I am a dying man," said old Turner. "I have had diabetes for +years. My doctor says it is a question whether I shall live a +month. Yet I would rather die under my own roof than in a gaol." + +Holmes rose and sat down at the table with his pen in his hand +and a bundle of paper before him. "Just tell us the truth," he +said. "I shall jot down the facts. You will sign it, and Watson +here can witness it. Then I could produce your confession at the +last extremity to save young McCarthy. I promise you that I shall +not use it unless it is absolutely needed." + +"It's as well," said the old man; "it's a question whether I +shall live to the Assizes, so it matters little to me, but I +should wish to spare Alice the shock. And now I will make the +thing clear to you; it has been a long time in the acting, but +will not take me long to tell. + +"You didn't know this dead man, McCarthy. He was a devil +incarnate. I tell you that. God keep you out of the clutches of +such a man as he. His grip has been upon me these twenty years, +and he has blasted my life. I'll tell you first how I came to be +in his power. + +"It was in the early '60's at the diggings. I was a young chap +then, hot-blooded and reckless, ready to turn my hand at +anything; I got among bad companions, took to drink, had no luck +with my claim, took to the bush, and in a word became what you +would call over here a highway robber. There were six of us, and +we had a wild, free life of it, sticking up a station from time +to time, or stopping the wagons on the road to the diggings. +Black Jack of Ballarat was the name I went under, and our party +is still remembered in the colony as the Ballarat Gang. + +"One day a gold convoy came down from Ballarat to Melbourne, and +we lay in wait for it and attacked it. There were six troopers +and six of us, so it was a close thing, but we emptied four of +their saddles at the first volley. Three of our boys were killed, +however, before we got the swag. I put my pistol to the head of +the wagon-driver, who was this very man McCarthy. I wish to the +Lord that I had shot him then, but I spared him, though I saw his +wicked little eyes fixed on my face, as though to remember every +feature. We got away with the gold, became wealthy men, and made +our way over to England without being suspected. There I parted +from my old pals and determined to settle down to a quiet and +respectable life. I bought this estate, which chanced to be in +the market, and I set myself to do a little good with my money, +to make up for the way in which I had earned it. I married, too, +and though my wife died young she left me my dear little Alice. +Even when she was just a baby her wee hand seemed to lead me down +the right path as nothing else had ever done. In a word, I turned +over a new leaf and did my best to make up for the past. All was +going well when McCarthy laid his grip upon me. + +"I had gone up to town about an investment, and I met him in +Regent Street with hardly a coat to his back or a boot to his +foot. + +"'Here we are, Jack,' says he, touching me on the arm; 'we'll be +as good as a family to you. There's two of us, me and my son, and +you can have the keeping of us. If you don't--it's a fine, +law-abiding country is England, and there's always a policeman +within hail.' + +"Well, down they came to the west country, there was no shaking +them off, and there they have lived rent free on my best land +ever since. There was no rest for me, no peace, no forgetfulness; +turn where I would, there was his cunning, grinning face at my +elbow. It grew worse as Alice grew up, for he soon saw I was more +afraid of her knowing my past than of the police. Whatever he +wanted he must have, and whatever it was I gave him without +question, land, money, houses, until at last he asked a thing +which I could not give. He asked for Alice. + +"His son, you see, had grown up, and so had my girl, and as I was +known to be in weak health, it seemed a fine stroke to him that +his lad should step into the whole property. But there I was +firm. I would not have his cursed stock mixed with mine; not that +I had any dislike to the lad, but his blood was in him, and that +was enough. I stood firm. McCarthy threatened. I braved him to do +his worst. We were to meet at the pool midway between our houses +to talk it over. + +"When I went down there I found him talking with his son, so I +smoked a cigar and waited behind a tree until he should be alone. +But as I listened to his talk all that was black and bitter in +me seemed to come uppermost. He was urging his son to marry my +daughter with as little regard for what she might think as if she +were a slut from off the streets. It drove me mad to think that I +and all that I held most dear should be in the power of such a +man as this. Could I not snap the bond? I was already a dying and +a desperate man. Though clear of mind and fairly strong of limb, +I knew that my own fate was sealed. But my memory and my girl! +Both could be saved if I could but silence that foul tongue. I +did it, Mr. Holmes. I would do it again. Deeply as I have sinned, +I have led a life of martyrdom to atone for it. But that my girl +should be entangled in the same meshes which held me was more +than I could suffer. I struck him down with no more compunction +than if he had been some foul and venomous beast. His cry brought +back his son; but I had gained the cover of the wood, though I +was forced to go back to fetch the cloak which I had dropped in +my flight. That is the true story, gentlemen, of all that +occurred." + +"Well, it is not for me to judge you," said Holmes as the old man +signed the statement which had been drawn out. "I pray that we +may never be exposed to such a temptation." + +"I pray not, sir. And what do you intend to do?" + +"In view of your health, nothing. You are yourself aware that you +will soon have to answer for your deed at a higher court than the +Assizes. I will keep your confession, and if McCarthy is +condemned I shall be forced to use it. If not, it shall never be +seen by mortal eye; and your secret, whether you be alive or +dead, shall be safe with us." + +"Farewell, then," said the old man solemnly. "Your own deathbeds, +when they come, will be the easier for the thought of the peace +which you have given to mine." Tottering and shaking in all his +giant frame, he stumbled slowly from the room. + +"God help us!" said Holmes after a long silence. "Why does fate +play such tricks with poor, helpless worms? I never hear of such +a case as this that I do not think of Baxter's words, and say, +'There, but for the grace of God, goes Sherlock Holmes.'" + +James McCarthy was acquitted at the Assizes on the strength of a +number of objections which had been drawn out by Holmes and +submitted to the defending counsel. Old Turner lived for seven +months after our interview, but he is now dead; and there is +every prospect that the son and daughter may come to live happily +together in ignorance of the black cloud which rests upon their +past. + + + +ADVENTURE V. THE FIVE ORANGE PIPS + +When I glance over my notes and records of the Sherlock Holmes +cases between the years '82 and '90, I am faced by so many which +present strange and interesting features that it is no easy +matter to know which to choose and which to leave. Some, however, +have already gained publicity through the papers, and others have +not offered a field for those peculiar qualities which my friend +possessed in so high a degree, and which it is the object of +these papers to illustrate. Some, too, have baffled his +analytical skill, and would be, as narratives, beginnings without +an ending, while others have been but partially cleared up, and +have their explanations founded rather upon conjecture and +surmise than on that absolute logical proof which was so dear to +him. There is, however, one of these last which was so remarkable +in its details and so startling in its results that I am tempted +to give some account of it in spite of the fact that there are +points in connection with it which never have been, and probably +never will be, entirely cleared up. + +The year '87 furnished us with a long series of cases of greater +or less interest, of which I retain the records. Among my +headings under this one twelve months I find an account of the +adventure of the Paradol Chamber, of the Amateur Mendicant +Society, who held a luxurious club in the lower vault of a +furniture warehouse, of the facts connected with the loss of the +British barque "Sophy Anderson", of the singular adventures of the +Grice Patersons in the island of Uffa, and finally of the +Camberwell poisoning case. In the latter, as may be remembered, +Sherlock Holmes was able, by winding up the dead man's watch, to +prove that it had been wound up two hours before, and that +therefore the deceased had gone to bed within that time--a +deduction which was of the greatest importance in clearing up the +case. All these I may sketch out at some future date, but none of +them present such singular features as the strange train of +circumstances which I have now taken up my pen to describe. + +It was in the latter days of September, and the equinoctial gales +had set in with exceptional violence. All day the wind had +screamed and the rain had beaten against the windows, so that +even here in the heart of great, hand-made London we were forced +to raise our minds for the instant from the routine of life and +to recognise the presence of those great elemental forces which +shriek at mankind through the bars of his civilisation, like +untamed beasts in a cage. As evening drew in, the storm grew +higher and louder, and the wind cried and sobbed like a child in +the chimney. Sherlock Holmes sat moodily at one side of the +fireplace cross-indexing his records of crime, while I at the +other was deep in one of Clark Russell's fine sea-stories until +the howl of the gale from without seemed to blend with the text, +and the splash of the rain to lengthen out into the long swash of +the sea waves. My wife was on a visit to her mother's, and for a +few days I was a dweller once more in my old quarters at Baker +Street. + +"Why," said I, glancing up at my companion, "that was surely the +bell. Who could come to-night? Some friend of yours, perhaps?" + +"Except yourself I have none," he answered. "I do not encourage +visitors." + +"A client, then?" + +"If so, it is a serious case. Nothing less would bring a man out +on such a day and at such an hour. But I take it that it is more +likely to be some crony of the landlady's." + +Sherlock Holmes was wrong in his conjecture, however, for there +came a step in the passage and a tapping at the door. He +stretched out his long arm to turn the lamp away from himself and +towards the vacant chair upon which a newcomer must sit. + +"Come in!" said he. + +The man who entered was young, some two-and-twenty at the +outside, well-groomed and trimly clad, with something of +refinement and delicacy in his bearing. The streaming umbrella +which he held in his hand, and his long shining waterproof told +of the fierce weather through which he had come. He looked about +him anxiously in the glare of the lamp, and I could see that his +face was pale and his eyes heavy, like those of a man who is +weighed down with some great anxiety. + +"I owe you an apology," he said, raising his golden pince-nez to +his eyes. "I trust that I am not intruding. I fear that I have +brought some traces of the storm and rain into your snug +chamber." + +"Give me your coat and umbrella," said Holmes. "They may rest +here on the hook and will be dry presently. You have come up from +the south-west, I see." + +"Yes, from Horsham." + +"That clay and chalk mixture which I see upon your toe caps is +quite distinctive." + +"I have come for advice." + +"That is easily got." + +"And help." + +"That is not always so easy." + +"I have heard of you, Mr. Holmes. I heard from Major Prendergast +how you saved him in the Tankerville Club scandal." + +"Ah, of course. He was wrongfully accused of cheating at cards." + +"He said that you could solve anything." + +"He said too much." + +"That you are never beaten." + +"I have been beaten four times--three times by men, and once by a +woman." + +"But what is that compared with the number of your successes?" + +"It is true that I have been generally successful." + +"Then you may be so with me." + +"I beg that you will draw your chair up to the fire and favour me +with some details as to your case." + +"It is no ordinary one." + +"None of those which come to me are. I am the last court of +appeal." + +"And yet I question, sir, whether, in all your experience, you +have ever listened to a more mysterious and inexplicable chain of +events than those which have happened in my own family." + +"You fill me with interest," said Holmes. "Pray give us the +essential facts from the commencement, and I can afterwards +question you as to those details which seem to me to be most +important." + +The young man pulled his chair up and pushed his wet feet out +towards the blaze. + +"My name," said he, "is John Openshaw, but my own affairs have, +as far as I can understand, little to do with this awful +business. It is a hereditary matter; so in order to give you an +idea of the facts, I must go back to the commencement of the +affair. + +"You must know that my grandfather had two sons--my uncle Elias +and my father Joseph. My father had a small factory at Coventry, +which he enlarged at the time of the invention of bicycling. He +was a patentee of the Openshaw unbreakable tire, and his business +met with such success that he was able to sell it and to retire +upon a handsome competence. + +"My uncle Elias emigrated to America when he was a young man and +became a planter in Florida, where he was reported to have done +very well. At the time of the war he fought in Jackson's army, +and afterwards under Hood, where he rose to be a colonel. When +Lee laid down his arms my uncle returned to his plantation, where +he remained for three or four years. About 1869 or 1870 he came +back to Europe and took a small estate in Sussex, near Horsham. +He had made a very considerable fortune in the States, and his +reason for leaving them was his aversion to the negroes, and his +dislike of the Republican policy in extending the franchise to +them. He was a singular man, fierce and quick-tempered, very +foul-mouthed when he was angry, and of a most retiring +disposition. During all the years that he lived at Horsham, I +doubt if ever he set foot in the town. He had a garden and two or +three fields round his house, and there he would take his +exercise, though very often for weeks on end he would never leave +his room. He drank a great deal of brandy and smoked very +heavily, but he would see no society and did not want any +friends, not even his own brother. + +"He didn't mind me; in fact, he took a fancy to me, for at the +time when he saw me first I was a youngster of twelve or so. This +would be in the year 1878, after he had been eight or nine years +in England. He begged my father to let me live with him and he +was very kind to me in his way. When he was sober he used to be +fond of playing backgammon and draughts with me, and he would +make me his representative both with the servants and with the +tradespeople, so that by the time that I was sixteen I was quite +master of the house. I kept all the keys and could go where I +liked and do what I liked, so long as I did not disturb him in +his privacy. There was one singular exception, however, for he +had a single room, a lumber-room up among the attics, which was +invariably locked, and which he would never permit either me or +anyone else to enter. With a boy's curiosity I have peeped +through the keyhole, but I was never able to see more than such a +collection of old trunks and bundles as would be expected in such +a room. + +"One day--it was in March, 1883--a letter with a foreign stamp +lay upon the table in front of the colonel's plate. It was not a +common thing for him to receive letters, for his bills were all +paid in ready money, and he had no friends of any sort. 'From +India!' said he as he took it up, 'Pondicherry postmark! What can +this be?' Opening it hurriedly, out there jumped five little +dried orange pips, which pattered down upon his plate. I began to +laugh at this, but the laugh was struck from my lips at the sight +of his face. His lip had fallen, his eyes were protruding, his +skin the colour of putty, and he glared at the envelope which he +still held in his trembling hand, 'K. K. K.!' he shrieked, and +then, 'My God, my God, my sins have overtaken me!' + +"'What is it, uncle?' I cried. + +"'Death,' said he, and rising from the table he retired to his +room, leaving me palpitating with horror. I took up the envelope +and saw scrawled in red ink upon the inner flap, just above the +gum, the letter K three times repeated. There was nothing else +save the five dried pips. What could be the reason of his +overpowering terror? I left the breakfast-table, and as I +ascended the stair I met him coming down with an old rusty key, +which must have belonged to the attic, in one hand, and a small +brass box, like a cashbox, in the other. + +"'They may do what they like, but I'll checkmate them still,' +said he with an oath. 'Tell Mary that I shall want a fire in my +room to-day, and send down to Fordham, the Horsham lawyer.' + +"I did as he ordered, and when the lawyer arrived I was asked to +step up to the room. The fire was burning brightly, and in the +grate there was a mass of black, fluffy ashes, as of burned +paper, while the brass box stood open and empty beside it. As I +glanced at the box I noticed, with a start, that upon the lid was +printed the treble K which I had read in the morning upon the +envelope. + +"'I wish you, John,' said my uncle, 'to witness my will. I leave +my estate, with all its advantages and all its disadvantages, to +my brother, your father, whence it will, no doubt, descend to +you. If you can enjoy it in peace, well and good! If you find you +cannot, take my advice, my boy, and leave it to your deadliest +enemy. I am sorry to give you such a two-edged thing, but I can't +say what turn things are going to take. Kindly sign the paper +where Mr. Fordham shows you.' + +"I signed the paper as directed, and the lawyer took it away with +him. The singular incident made, as you may think, the deepest +impression upon me, and I pondered over it and turned it every +way in my mind without being able to make anything of it. Yet I +could not shake off the vague feeling of dread which it left +behind, though the sensation grew less keen as the weeks passed +and nothing happened to disturb the usual routine of our lives. I +could see a change in my uncle, however. He drank more than ever, +and he was less inclined for any sort of society. Most of his +time he would spend in his room, with the door locked upon the +inside, but sometimes he would emerge in a sort of drunken frenzy +and would burst out of the house and tear about the garden with a +revolver in his hand, screaming out that he was afraid of no man, +and that he was not to be cooped up, like a sheep in a pen, by +man or devil. When these hot fits were over, however, he would +rush tumultuously in at the door and lock and bar it behind him, +like a man who can brazen it out no longer against the terror +which lies at the roots of his soul. At such times I have seen +his face, even on a cold day, glisten with moisture, as though it +were new raised from a basin. + +"Well, to come to an end of the matter, Mr. Holmes, and not to +abuse your patience, there came a night when he made one of those +drunken sallies from which he never came back. We found him, when +we went to search for him, face downward in a little +green-scummed pool, which lay at the foot of the garden. There +was no sign of any violence, and the water was but two feet deep, +so that the jury, having regard to his known eccentricity, +brought in a verdict of 'suicide.' But I, who knew how he winced +from the very thought of death, had much ado to persuade myself +that he had gone out of his way to meet it. The matter passed, +however, and my father entered into possession of the estate, and +of some 14,000 pounds, which lay to his credit at the bank." + +"One moment," Holmes interposed, "your statement is, I foresee, +one of the most remarkable to which I have ever listened. Let me +have the date of the reception by your uncle of the letter, and +the date of his supposed suicide." + +"The letter arrived on March 10, 1883. His death was seven weeks +later, upon the night of May 2nd." + +"Thank you. Pray proceed." + +"When my father took over the Horsham property, he, at my +request, made a careful examination of the attic, which had been +always locked up. We found the brass box there, although its +contents had been destroyed. On the inside of the cover was a +paper label, with the initials of K. K. K. repeated upon it, and +'Letters, memoranda, receipts, and a register' written beneath. +These, we presume, indicated the nature of the papers which had +been destroyed by Colonel Openshaw. For the rest, there was +nothing of much importance in the attic save a great many +scattered papers and note-books bearing upon my uncle's life in +America. Some of them were of the war time and showed that he had +done his duty well and had borne the repute of a brave soldier. +Others were of a date during the reconstruction of the Southern +states, and were mostly concerned with politics, for he had +evidently taken a strong part in opposing the carpet-bag +politicians who had been sent down from the North. + +"Well, it was the beginning of '84 when my father came to live at +Horsham, and all went as well as possible with us until the +January of '85. On the fourth day after the new year I heard my +father give a sharp cry of surprise as we sat together at the +breakfast-table. There he was, sitting with a newly opened +envelope in one hand and five dried orange pips in the +outstretched palm of the other one. He had always laughed at what +he called my cock-and-bull story about the colonel, but he looked +very scared and puzzled now that the same thing had come upon +himself. + +"'Why, what on earth does this mean, John?' he stammered. + +"My heart had turned to lead. 'It is K. K. K.,' said I. + +"He looked inside the envelope. 'So it is,' he cried. 'Here are +the very letters. But what is this written above them?' + +"'Put the papers on the sundial,' I read, peeping over his +shoulder. + +"'What papers? What sundial?' he asked. + +"'The sundial in the garden. There is no other,' said I; 'but the +papers must be those that are destroyed.' + +"'Pooh!' said he, gripping hard at his courage. 'We are in a +civilised land here, and we can't have tomfoolery of this kind. +Where does the thing come from?' + +"'From Dundee,' I answered, glancing at the postmark. + +"'Some preposterous practical joke,' said he. 'What have I to do +with sundials and papers? I shall take no notice of such +nonsense.' + +"'I should certainly speak to the police,' I said. + +"'And be laughed at for my pains. Nothing of the sort.' + +"'Then let me do so?' + +"'No, I forbid you. I won't have a fuss made about such +nonsense.' + +"It was in vain to argue with him, for he was a very obstinate +man. I went about, however, with a heart which was full of +forebodings. + +"On the third day after the coming of the letter my father went +from home to visit an old friend of his, Major Freebody, who is +in command of one of the forts upon Portsdown Hill. I was glad +that he should go, for it seemed to me that he was farther from +danger when he was away from home. In that, however, I was in +error. Upon the second day of his absence I received a telegram +from the major, imploring me to come at once. My father had +fallen over one of the deep chalk-pits which abound in the +neighbourhood, and was lying senseless, with a shattered skull. I +hurried to him, but he passed away without having ever recovered +his consciousness. He had, as it appears, been returning from +Fareham in the twilight, and as the country was unknown to him, +and the chalk-pit unfenced, the jury had no hesitation in +bringing in a verdict of 'death from accidental causes.' +Carefully as I examined every fact connected with his death, I +was unable to find anything which could suggest the idea of +murder. There were no signs of violence, no footmarks, no +robbery, no record of strangers having been seen upon the roads. +And yet I need not tell you that my mind was far from at ease, +and that I was well-nigh certain that some foul plot had been +woven round him. + +"In this sinister way I came into my inheritance. You will ask me +why I did not dispose of it? I answer, because I was well +convinced that our troubles were in some way dependent upon an +incident in my uncle's life, and that the danger would be as +pressing in one house as in another. + +"It was in January, '85, that my poor father met his end, and two +years and eight months have elapsed since then. During that time +I have lived happily at Horsham, and I had begun to hope that +this curse had passed away from the family, and that it had ended +with the last generation. I had begun to take comfort too soon, +however; yesterday morning the blow fell in the very shape in +which it had come upon my father." + +The young man took from his waistcoat a crumpled envelope, and +turning to the table he shook out upon it five little dried +orange pips. + +"This is the envelope," he continued. "The postmark is +London--eastern division. Within are the very words which were +upon my father's last message: 'K. K. K.'; and then 'Put the +papers on the sundial.'" + +"What have you done?" asked Holmes. + +"Nothing." + +"Nothing?" + +"To tell the truth"--he sank his face into his thin, white +hands--"I have felt helpless. I have felt like one of those poor +rabbits when the snake is writhing towards it. I seem to be in +the grasp of some resistless, inexorable evil, which no foresight +and no precautions can guard against." + +"Tut! tut!" cried Sherlock Holmes. "You must act, man, or you are +lost. Nothing but energy can save you. This is no time for +despair." + +"I have seen the police." + +"Ah!" + +"But they listened to my story with a smile. I am convinced that +the inspector has formed the opinion that the letters are all +practical jokes, and that the deaths of my relations were really +accidents, as the jury stated, and were not to be connected with +the warnings." + +Holmes shook his clenched hands in the air. "Incredible +imbecility!" he cried. + +"They have, however, allowed me a policeman, who may remain in +the house with me." + +"Has he come with you to-night?" + +"No. His orders were to stay in the house." + +Again Holmes raved in the air. + +"Why did you come to me," he cried, "and, above all, why did you +not come at once?" + +"I did not know. It was only to-day that I spoke to Major +Prendergast about my troubles and was advised by him to come to +you." + +"It is really two days since you had the letter. We should have +acted before this. You have no further evidence, I suppose, than +that which you have placed before us--no suggestive detail which +might help us?" + +"There is one thing," said John Openshaw. He rummaged in his coat +pocket, and, drawing out a piece of discoloured, blue-tinted +paper, he laid it out upon the table. "I have some remembrance," +said he, "that on the day when my uncle burned the papers I +observed that the small, unburned margins which lay amid the +ashes were of this particular colour. I found this single sheet +upon the floor of his room, and I am inclined to think that it +may be one of the papers which has, perhaps, fluttered out from +among the others, and in that way has escaped destruction. Beyond +the mention of pips, I do not see that it helps us much. I think +myself that it is a page from some private diary. The writing is +undoubtedly my uncle's." + +Holmes moved the lamp, and we both bent over the sheet of paper, +which showed by its ragged edge that it had indeed been torn from +a book. It was headed, "March, 1869," and beneath were the +following enigmatical notices: + +"4th. Hudson came. Same old platform. + +"7th. Set the pips on McCauley, Paramore, and + John Swain, of St. Augustine. + +"9th. McCauley cleared. + +"10th. John Swain cleared. + +"12th. Visited Paramore. All well." + +"Thank you!" said Holmes, folding up the paper and returning it +to our visitor. "And now you must on no account lose another +instant. We cannot spare time even to discuss what you have told +me. You must get home instantly and act." + +"What shall I do?" + +"There is but one thing to do. It must be done at once. You must +put this piece of paper which you have shown us into the brass +box which you have described. You must also put in a note to say +that all the other papers were burned by your uncle, and that +this is the only one which remains. You must assert that in such +words as will carry conviction with them. Having done this, you +must at once put the box out upon the sundial, as directed. Do +you understand?" + +"Entirely." + +"Do not think of revenge, or anything of the sort, at present. I +think that we may gain that by means of the law; but we have our +web to weave, while theirs is already woven. The first +consideration is to remove the pressing danger which threatens +you. The second is to clear up the mystery and to punish the +guilty parties." + +"I thank you," said the young man, rising and pulling on his +overcoat. "You have given me fresh life and hope. I shall +certainly do as you advise." + +"Do not lose an instant. And, above all, take care of yourself in +the meanwhile, for I do not think that there can be a doubt that +you are threatened by a very real and imminent danger. How do you +go back?" + +"By train from Waterloo." + +"It is not yet nine. The streets will be crowded, so I trust that +you may be in safety. And yet you cannot guard yourself too +closely." + +"I am armed." + +"That is well. To-morrow I shall set to work upon your case." + +"I shall see you at Horsham, then?" + +"No, your secret lies in London. It is there that I shall seek +it." + +"Then I shall call upon you in a day, or in two days, with news +as to the box and the papers. I shall take your advice in every +particular." He shook hands with us and took his leave. Outside +the wind still screamed and the rain splashed and pattered +against the windows. This strange, wild story seemed to have come +to us from amid the mad elements--blown in upon us like a sheet +of sea-weed in a gale--and now to have been reabsorbed by them +once more. + +Sherlock Holmes sat for some time in silence, with his head sunk +forward and his eyes bent upon the red glow of the fire. Then he +lit his pipe, and leaning back in his chair he watched the blue +smoke-rings as they chased each other up to the ceiling. + +"I think, Watson," he remarked at last, "that of all our cases we +have had none more fantastic than this." + +"Save, perhaps, the Sign of Four." + +"Well, yes. Save, perhaps, that. And yet this John Openshaw seems +to me to be walking amid even greater perils than did the +Sholtos." + +"But have you," I asked, "formed any definite conception as to +what these perils are?" + +"There can be no question as to their nature," he answered. + +"Then what are they? Who is this K. K. K., and why does he pursue +this unhappy family?" + +Sherlock Holmes closed his eyes and placed his elbows upon the +arms of his chair, with his finger-tips together. "The ideal +reasoner," he remarked, "would, when he had once been shown a +single fact in all its bearings, deduce from it not only all the +chain of events which led up to it but also all the results which +would follow from it. As Cuvier could correctly describe a whole +animal by the contemplation of a single bone, so the observer who +has thoroughly understood one link in a series of incidents +should be able to accurately state all the other ones, both +before and after. We have not yet grasped the results which the +reason alone can attain to. Problems may be solved in the study +which have baffled all those who have sought a solution by the +aid of their senses. To carry the art, however, to its highest +pitch, it is necessary that the reasoner should be able to +utilise all the facts which have come to his knowledge; and this +in itself implies, as you will readily see, a possession of all +knowledge, which, even in these days of free education and +encyclopaedias, is a somewhat rare accomplishment. It is not so +impossible, however, that a man should possess all knowledge +which is likely to be useful to him in his work, and this I have +endeavoured in my case to do. If I remember rightly, you on one +occasion, in the early days of our friendship, defined my limits +in a very precise fashion." + +"Yes," I answered, laughing. "It was a singular document. +Philosophy, astronomy, and politics were marked at zero, I +remember. Botany variable, geology profound as regards the +mud-stains from any region within fifty miles of town, chemistry +eccentric, anatomy unsystematic, sensational literature and crime +records unique, violin-player, boxer, swordsman, lawyer, and +self-poisoner by cocaine and tobacco. Those, I think, were the +main points of my analysis." + +Holmes grinned at the last item. "Well," he said, "I say now, as +I said then, that a man should keep his little brain-attic +stocked with all the furniture that he is likely to use, and the +rest he can put away in the lumber-room of his library, where he +can get it if he wants it. Now, for such a case as the one which +has been submitted to us to-night, we need certainly to muster +all our resources. Kindly hand me down the letter K of the +'American Encyclopaedia' which stands upon the shelf beside you. +Thank you. Now let us consider the situation and see what may be +deduced from it. In the first place, we may start with a strong +presumption that Colonel Openshaw had some very strong reason for +leaving America. Men at his time of life do not change all their +habits and exchange willingly the charming climate of Florida for +the lonely life of an English provincial town. His extreme love +of solitude in England suggests the idea that he was in fear of +someone or something, so we may assume as a working hypothesis +that it was fear of someone or something which drove him from +America. As to what it was he feared, we can only deduce that by +considering the formidable letters which were received by himself +and his successors. Did you remark the postmarks of those +letters?" + +"The first was from Pondicherry, the second from Dundee, and the +third from London." + +"From East London. What do you deduce from that?" + +"They are all seaports. That the writer was on board of a ship." + +"Excellent. We have already a clue. There can be no doubt that +the probability--the strong probability--is that the writer was +on board of a ship. And now let us consider another point. In the +case of Pondicherry, seven weeks elapsed between the threat and +its fulfilment, in Dundee it was only some three or four days. +Does that suggest anything?" + +"A greater distance to travel." + +"But the letter had also a greater distance to come." + +"Then I do not see the point." + +"There is at least a presumption that the vessel in which the man +or men are is a sailing-ship. It looks as if they always send +their singular warning or token before them when starting upon +their mission. You see how quickly the deed followed the sign +when it came from Dundee. If they had come from Pondicherry in a +steamer they would have arrived almost as soon as their letter. +But, as a matter of fact, seven weeks elapsed. I think that those +seven weeks represented the difference between the mail-boat which +brought the letter and the sailing vessel which brought the +writer." + +"It is possible." + +"More than that. It is probable. And now you see the deadly +urgency of this new case, and why I urged young Openshaw to +caution. The blow has always fallen at the end of the time which +it would take the senders to travel the distance. But this one +comes from London, and therefore we cannot count upon delay." + +"Good God!" I cried. "What can it mean, this relentless +persecution?" + +"The papers which Openshaw carried are obviously of vital +importance to the person or persons in the sailing-ship. I think +that it is quite clear that there must be more than one of them. +A single man could not have carried out two deaths in such a way +as to deceive a coroner's jury. There must have been several in +it, and they must have been men of resource and determination. +Their papers they mean to have, be the holder of them who it may. +In this way you see K. K. K. ceases to be the initials of an +individual and becomes the badge of a society." + +"But of what society?" + +"Have you never--" said Sherlock Holmes, bending forward and +sinking his voice--"have you never heard of the Ku Klux Klan?" + +"I never have." + +Holmes turned over the leaves of the book upon his knee. "Here it +is," said he presently: + +"'Ku Klux Klan. A name derived from the fanciful resemblance to +the sound produced by cocking a rifle. This terrible secret +society was formed by some ex-Confederate soldiers in the +Southern states after the Civil War, and it rapidly formed local +branches in different parts of the country, notably in Tennessee, +Louisiana, the Carolinas, Georgia, and Florida. Its power was +used for political purposes, principally for the terrorising of +the negro voters and the murdering and driving from the country +of those who were opposed to its views. Its outrages were usually +preceded by a warning sent to the marked man in some fantastic +but generally recognised shape--a sprig of oak-leaves in some +parts, melon seeds or orange pips in others. On receiving this +the victim might either openly abjure his former ways, or might +fly from the country. If he braved the matter out, death would +unfailingly come upon him, and usually in some strange and +unforeseen manner. So perfect was the organisation of the +society, and so systematic its methods, that there is hardly a +case upon record where any man succeeded in braving it with +impunity, or in which any of its outrages were traced home to the +perpetrators. For some years the organisation flourished in spite +of the efforts of the United States government and of the better +classes of the community in the South. Eventually, in the year +1869, the movement rather suddenly collapsed, although there have +been sporadic outbreaks of the same sort since that date.' + +"You will observe," said Holmes, laying down the volume, "that +the sudden breaking up of the society was coincident with the +disappearance of Openshaw from America with their papers. It may +well have been cause and effect. It is no wonder that he and his +family have some of the more implacable spirits upon their track. +You can understand that this register and diary may implicate +some of the first men in the South, and that there may be many +who will not sleep easy at night until it is recovered." + +"Then the page we have seen--" + +"Is such as we might expect. It ran, if I remember right, 'sent +the pips to A, B, and C'--that is, sent the society's warning to +them. Then there are successive entries that A and B cleared, or +left the country, and finally that C was visited, with, I fear, a +sinister result for C. Well, I think, Doctor, that we may let +some light into this dark place, and I believe that the only +chance young Openshaw has in the meantime is to do what I have +told him. There is nothing more to be said or to be done +to-night, so hand me over my violin and let us try to forget for +half an hour the miserable weather and the still more miserable +ways of our fellow-men." + + +It had cleared in the morning, and the sun was shining with a +subdued brightness through the dim veil which hangs over the +great city. Sherlock Holmes was already at breakfast when I came +down. + +"You will excuse me for not waiting for you," said he; "I have, I +foresee, a very busy day before me in looking into this case of +young Openshaw's." + +"What steps will you take?" I asked. + +"It will very much depend upon the results of my first inquiries. +I may have to go down to Horsham, after all." + +"You will not go there first?" + +"No, I shall commence with the City. Just ring the bell and the +maid will bring up your coffee." + +As I waited, I lifted the unopened newspaper from the table and +glanced my eye over it. It rested upon a heading which sent a +chill to my heart. + +"Holmes," I cried, "you are too late." + +"Ah!" said he, laying down his cup, "I feared as much. How was it +done?" He spoke calmly, but I could see that he was deeply moved. + +"My eye caught the name of Openshaw, and the heading 'Tragedy +Near Waterloo Bridge.' Here is the account: + +"Between nine and ten last night Police-Constable Cook, of the H +Division, on duty near Waterloo Bridge, heard a cry for help and +a splash in the water. The night, however, was extremely dark and +stormy, so that, in spite of the help of several passers-by, it +was quite impossible to effect a rescue. The alarm, however, was +given, and, by the aid of the water-police, the body was +eventually recovered. It proved to be that of a young gentleman +whose name, as it appears from an envelope which was found in his +pocket, was John Openshaw, and whose residence is near Horsham. +It is conjectured that he may have been hurrying down to catch +the last train from Waterloo Station, and that in his haste and +the extreme darkness he missed his path and walked over the edge +of one of the small landing-places for river steamboats. The body +exhibited no traces of violence, and there can be no doubt that +the deceased had been the victim of an unfortunate accident, +which should have the effect of calling the attention of the +authorities to the condition of the riverside landing-stages." + +We sat in silence for some minutes, Holmes more depressed and +shaken than I had ever seen him. + +"That hurts my pride, Watson," he said at last. "It is a petty +feeling, no doubt, but it hurts my pride. It becomes a personal +matter with me now, and, if God sends me health, I shall set my +hand upon this gang. That he should come to me for help, and that +I should send him away to his death--!" He sprang from his chair +and paced about the room in uncontrollable agitation, with a +flush upon his sallow cheeks and a nervous clasping and +unclasping of his long thin hands. + +"They must be cunning devils," he exclaimed at last. "How could +they have decoyed him down there? The Embankment is not on the +direct line to the station. The bridge, no doubt, was too +crowded, even on such a night, for their purpose. Well, Watson, +we shall see who will win in the long run. I am going out now!" + +"To the police?" + +"No; I shall be my own police. When I have spun the web they may +take the flies, but not before." + +All day I was engaged in my professional work, and it was late in +the evening before I returned to Baker Street. Sherlock Holmes +had not come back yet. It was nearly ten o'clock before he +entered, looking pale and worn. He walked up to the sideboard, +and tearing a piece from the loaf he devoured it voraciously, +washing it down with a long draught of water. + +"You are hungry," I remarked. + +"Starving. It had escaped my memory. I have had nothing since +breakfast." + +"Nothing?" + +"Not a bite. I had no time to think of it." + +"And how have you succeeded?" + +"Well." + +"You have a clue?" + +"I have them in the hollow of my hand. Young Openshaw shall not +long remain unavenged. Why, Watson, let us put their own devilish +trade-mark upon them. It is well thought of!" + +"What do you mean?" + +He took an orange from the cupboard, and tearing it to pieces he +squeezed out the pips upon the table. Of these he took five and +thrust them into an envelope. On the inside of the flap he wrote +"S. H. for J. O." Then he sealed it and addressed it to "Captain +James Calhoun, Barque 'Lone Star,' Savannah, Georgia." + +"That will await him when he enters port," said he, chuckling. +"It may give him a sleepless night. He will find it as sure a +precursor of his fate as Openshaw did before him." + +"And who is this Captain Calhoun?" + +"The leader of the gang. I shall have the others, but he first." + +"How did you trace it, then?" + +He took a large sheet of paper from his pocket, all covered with +dates and names. + +"I have spent the whole day," said he, "over Lloyd's registers +and files of the old papers, following the future career of every +vessel which touched at Pondicherry in January and February in +'83. There were thirty-six ships of fair tonnage which were +reported there during those months. Of these, one, the 'Lone Star,' +instantly attracted my attention, since, although it was reported +as having cleared from London, the name is that which is given to +one of the states of the Union." + +"Texas, I think." + +"I was not and am not sure which; but I knew that the ship must +have an American origin." + +"What then?" + +"I searched the Dundee records, and when I found that the barque +'Lone Star' was there in January, '85, my suspicion became a +certainty. I then inquired as to the vessels which lay at present +in the port of London." + +"Yes?" + +"The 'Lone Star' had arrived here last week. I went down to the +Albert Dock and found that she had been taken down the river by +the early tide this morning, homeward bound to Savannah. I wired +to Gravesend and learned that she had passed some time ago, and +as the wind is easterly I have no doubt that she is now past the +Goodwins and not very far from the Isle of Wight." + +"What will you do, then?" + +"Oh, I have my hand upon him. He and the two mates, are as I +learn, the only native-born Americans in the ship. The others are +Finns and Germans. I know, also, that they were all three away +from the ship last night. I had it from the stevedore who has +been loading their cargo. By the time that their sailing-ship +reaches Savannah the mail-boat will have carried this letter, and +the cable will have informed the police of Savannah that these +three gentlemen are badly wanted here upon a charge of murder." + +There is ever a flaw, however, in the best laid of human plans, +and the murderers of John Openshaw were never to receive the +orange pips which would show them that another, as cunning and as +resolute as themselves, was upon their track. Very long and very +severe were the equinoctial gales that year. We waited long for +news of the "Lone Star" of Savannah, but none ever reached us. We +did at last hear that somewhere far out in the Atlantic a +shattered stern-post of a boat was seen swinging in the trough +of a wave, with the letters "L. S." carved upon it, and that is +all which we shall ever know of the fate of the "Lone Star." + + + +ADVENTURE VI. THE MAN WITH THE TWISTED LIP + +Isa Whitney, brother of the late Elias Whitney, D.D., Principal +of the Theological College of St. George's, was much addicted to +opium. The habit grew upon him, as I understand, from some +foolish freak when he was at college; for having read De +Quincey's description of his dreams and sensations, he had +drenched his tobacco with laudanum in an attempt to produce the +same effects. He found, as so many more have done, that the +practice is easier to attain than to get rid of, and for many +years he continued to be a slave to the drug, an object of +mingled horror and pity to his friends and relatives. I can see +him now, with yellow, pasty face, drooping lids, and pin-point +pupils, all huddled in a chair, the wreck and ruin of a noble +man. + +One night--it was in June, '89--there came a ring to my bell, +about the hour when a man gives his first yawn and glances at the +clock. I sat up in my chair, and my wife laid her needle-work +down in her lap and made a little face of disappointment. + +"A patient!" said she. "You'll have to go out." + +I groaned, for I was newly come back from a weary day. + +We heard the door open, a few hurried words, and then quick steps +upon the linoleum. Our own door flew open, and a lady, clad in +some dark-coloured stuff, with a black veil, entered the room. + +"You will excuse my calling so late," she began, and then, +suddenly losing her self-control, she ran forward, threw her arms +about my wife's neck, and sobbed upon her shoulder. "Oh, I'm in +such trouble!" she cried; "I do so want a little help." + +"Why," said my wife, pulling up her veil, "it is Kate Whitney. +How you startled me, Kate! I had not an idea who you were when +you came in." + +"I didn't know what to do, so I came straight to you." That was +always the way. Folk who were in grief came to my wife like birds +to a light-house. + +"It was very sweet of you to come. Now, you must have some wine +and water, and sit here comfortably and tell us all about it. Or +should you rather that I sent James off to bed?" + +"Oh, no, no! I want the doctor's advice and help, too. It's about +Isa. He has not been home for two days. I am so frightened about +him!" + +It was not the first time that she had spoken to us of her +husband's trouble, to me as a doctor, to my wife as an old friend +and school companion. We soothed and comforted her by such words +as we could find. Did she know where her husband was? Was it +possible that we could bring him back to her? + +It seems that it was. She had the surest information that of late +he had, when the fit was on him, made use of an opium den in the +farthest east of the City. Hitherto his orgies had always been +confined to one day, and he had come back, twitching and +shattered, in the evening. But now the spell had been upon him +eight-and-forty hours, and he lay there, doubtless among the +dregs of the docks, breathing in the poison or sleeping off the +effects. There he was to be found, she was sure of it, at the Bar +of Gold, in Upper Swandam Lane. But what was she to do? How could +she, a young and timid woman, make her way into such a place and +pluck her husband out from among the ruffians who surrounded him? + +There was the case, and of course there was but one way out of +it. Might I not escort her to this place? And then, as a second +thought, why should she come at all? I was Isa Whitney's medical +adviser, and as such I had influence over him. I could manage it +better if I were alone. I promised her on my word that I would +send him home in a cab within two hours if he were indeed at the +address which she had given me. And so in ten minutes I had left +my armchair and cheery sitting-room behind me, and was speeding +eastward in a hansom on a strange errand, as it seemed to me at +the time, though the future only could show how strange it was to +be. + +But there was no great difficulty in the first stage of my +adventure. Upper Swandam Lane is a vile alley lurking behind the +high wharves which line the north side of the river to the east +of London Bridge. Between a slop-shop and a gin-shop, approached +by a steep flight of steps leading down to a black gap like the +mouth of a cave, I found the den of which I was in search. +Ordering my cab to wait, I passed down the steps, worn hollow in +the centre by the ceaseless tread of drunken feet; and by the +light of a flickering oil-lamp above the door I found the latch +and made my way into a long, low room, thick and heavy with the +brown opium smoke, and terraced with wooden berths, like the +forecastle of an emigrant ship. + +Through the gloom one could dimly catch a glimpse of bodies lying +in strange fantastic poses, bowed shoulders, bent knees, heads +thrown back, and chins pointing upward, with here and there a +dark, lack-lustre eye turned upon the newcomer. Out of the black +shadows there glimmered little red circles of light, now bright, +now faint, as the burning poison waxed or waned in the bowls of +the metal pipes. The most lay silent, but some muttered to +themselves, and others talked together in a strange, low, +monotonous voice, their conversation coming in gushes, and then +suddenly tailing off into silence, each mumbling out his own +thoughts and paying little heed to the words of his neighbour. At +the farther end was a small brazier of burning charcoal, beside +which on a three-legged wooden stool there sat a tall, thin old +man, with his jaw resting upon his two fists, and his elbows upon +his knees, staring into the fire. + +As I entered, a sallow Malay attendant had hurried up with a pipe +for me and a supply of the drug, beckoning me to an empty berth. + +"Thank you. I have not come to stay," said I. "There is a friend +of mine here, Mr. Isa Whitney, and I wish to speak with him." + +There was a movement and an exclamation from my right, and +peering through the gloom, I saw Whitney, pale, haggard, and +unkempt, staring out at me. + +"My God! It's Watson," said he. He was in a pitiable state of +reaction, with every nerve in a twitter. "I say, Watson, what +o'clock is it?" + +"Nearly eleven." + +"Of what day?" + +"Of Friday, June 19th." + +"Good heavens! I thought it was Wednesday. It is Wednesday. What +d'you want to frighten a chap for?" He sank his face onto his +arms and began to sob in a high treble key. + +"I tell you that it is Friday, man. Your wife has been waiting +this two days for you. You should be ashamed of yourself!" + +"So I am. But you've got mixed, Watson, for I have only been here +a few hours, three pipes, four pipes--I forget how many. But I'll +go home with you. I wouldn't frighten Kate--poor little Kate. +Give me your hand! Have you a cab?" + +"Yes, I have one waiting." + +"Then I shall go in it. But I must owe something. Find what I +owe, Watson. I am all off colour. I can do nothing for myself." + +I walked down the narrow passage between the double row of +sleepers, holding my breath to keep out the vile, stupefying +fumes of the drug, and looking about for the manager. As I passed +the tall man who sat by the brazier I felt a sudden pluck at my +skirt, and a low voice whispered, "Walk past me, and then look +back at me." The words fell quite distinctly upon my ear. I +glanced down. They could only have come from the old man at my +side, and yet he sat now as absorbed as ever, very thin, very +wrinkled, bent with age, an opium pipe dangling down from between +his knees, as though it had dropped in sheer lassitude from his +fingers. I took two steps forward and looked back. It took all my +self-control to prevent me from breaking out into a cry of +astonishment. He had turned his back so that none could see him +but I. His form had filled out, his wrinkles were gone, the dull +eyes had regained their fire, and there, sitting by the fire and +grinning at my surprise, was none other than Sherlock Holmes. He +made a slight motion to me to approach him, and instantly, as he +turned his face half round to the company once more, subsided +into a doddering, loose-lipped senility. + +"Holmes!" I whispered, "what on earth are you doing in this den?" + +"As low as you can," he answered; "I have excellent ears. If you +would have the great kindness to get rid of that sottish friend +of yours I should be exceedingly glad to have a little talk with +you." + +"I have a cab outside." + +"Then pray send him home in it. You may safely trust him, for he +appears to be too limp to get into any mischief. I should +recommend you also to send a note by the cabman to your wife to +say that you have thrown in your lot with me. If you will wait +outside, I shall be with you in five minutes." + +It was difficult to refuse any of Sherlock Holmes' requests, for +they were always so exceedingly definite, and put forward with +such a quiet air of mastery. I felt, however, that when Whitney +was once confined in the cab my mission was practically +accomplished; and for the rest, I could not wish anything better +than to be associated with my friend in one of those singular +adventures which were the normal condition of his existence. In a +few minutes I had written my note, paid Whitney's bill, led him +out to the cab, and seen him driven through the darkness. In a +very short time a decrepit figure had emerged from the opium den, +and I was walking down the street with Sherlock Holmes. For two +streets he shuffled along with a bent back and an uncertain foot. +Then, glancing quickly round, he straightened himself out and +burst into a hearty fit of laughter. + +"I suppose, Watson," said he, "that you imagine that I have added +opium-smoking to cocaine injections, and all the other little +weaknesses on which you have favoured me with your medical +views." + +"I was certainly surprised to find you there." + +"But not more so than I to find you." + +"I came to find a friend." + +"And I to find an enemy." + +"An enemy?" + +"Yes; one of my natural enemies, or, shall I say, my natural +prey. Briefly, Watson, I am in the midst of a very remarkable +inquiry, and I have hoped to find a clue in the incoherent +ramblings of these sots, as I have done before now. Had I been +recognised in that den my life would not have been worth an +hour's purchase; for I have used it before now for my own +purposes, and the rascally Lascar who runs it has sworn to have +vengeance upon me. There is a trap-door at the back of that +building, near the corner of Paul's Wharf, which could tell some +strange tales of what has passed through it upon the moonless +nights." + +"What! You do not mean bodies?" + +"Ay, bodies, Watson. We should be rich men if we had 1000 pounds +for every poor devil who has been done to death in that den. It +is the vilest murder-trap on the whole riverside, and I fear that +Neville St. Clair has entered it never to leave it more. But our +trap should be here." He put his two forefingers between his +teeth and whistled shrilly--a signal which was answered by a +similar whistle from the distance, followed shortly by the rattle +of wheels and the clink of horses' hoofs. + +"Now, Watson," said Holmes, as a tall dog-cart dashed up through +the gloom, throwing out two golden tunnels of yellow light from +its side lanterns. "You'll come with me, won't you?" + +"If I can be of use." + +"Oh, a trusty comrade is always of use; and a chronicler still +more so. My room at The Cedars is a double-bedded one." + +"The Cedars?" + +"Yes; that is Mr. St. Clair's house. I am staying there while I +conduct the inquiry." + +"Where is it, then?" + +"Near Lee, in Kent. We have a seven-mile drive before us." + +"But I am all in the dark." + +"Of course you are. You'll know all about it presently. Jump up +here. All right, John; we shall not need you. Here's half a +crown. Look out for me to-morrow, about eleven. Give her her +head. So long, then!" + +He flicked the horse with his whip, and we dashed away through +the endless succession of sombre and deserted streets, which +widened gradually, until we were flying across a broad +balustraded bridge, with the murky river flowing sluggishly +beneath us. Beyond lay another dull wilderness of bricks and +mortar, its silence broken only by the heavy, regular footfall of +the policeman, or the songs and shouts of some belated party of +revellers. A dull wrack was drifting slowly across the sky, and a +star or two twinkled dimly here and there through the rifts of +the clouds. Holmes drove in silence, with his head sunk upon his +breast, and the air of a man who is lost in thought, while I sat +beside him, curious to learn what this new quest might be which +seemed to tax his powers so sorely, and yet afraid to break in +upon the current of his thoughts. We had driven several miles, +and were beginning to get to the fringe of the belt of suburban +villas, when he shook himself, shrugged his shoulders, and lit up +his pipe with the air of a man who has satisfied himself that he +is acting for the best. + +"You have a grand gift of silence, Watson," said he. "It makes +you quite invaluable as a companion. 'Pon my word, it is a great +thing for me to have someone to talk to, for my own thoughts are +not over-pleasant. I was wondering what I should say to this dear +little woman to-night when she meets me at the door." + +"You forget that I know nothing about it." + +"I shall just have time to tell you the facts of the case before +we get to Lee. It seems absurdly simple, and yet, somehow I can +get nothing to go upon. There's plenty of thread, no doubt, but I +can't get the end of it into my hand. Now, I'll state the case +clearly and concisely to you, Watson, and maybe you can see a +spark where all is dark to me." + +"Proceed, then." + +"Some years ago--to be definite, in May, 1884--there came to Lee +a gentleman, Neville St. Clair by name, who appeared to have +plenty of money. He took a large villa, laid out the grounds very +nicely, and lived generally in good style. By degrees he made +friends in the neighbourhood, and in 1887 he married the daughter +of a local brewer, by whom he now has two children. He had no +occupation, but was interested in several companies and went into +town as a rule in the morning, returning by the 5:14 from Cannon +Street every night. Mr. St. Clair is now thirty-seven years of +age, is a man of temperate habits, a good husband, a very +affectionate father, and a man who is popular with all who know +him. I may add that his whole debts at the present moment, as far +as we have been able to ascertain, amount to 88 pounds 10s., while +he has 220 pounds standing to his credit in the Capital and +Counties Bank. There is no reason, therefore, to think that money +troubles have been weighing upon his mind. + +"Last Monday Mr. Neville St. Clair went into town rather earlier +than usual, remarking before he started that he had two important +commissions to perform, and that he would bring his little boy +home a box of bricks. Now, by the merest chance, his wife +received a telegram upon this same Monday, very shortly after his +departure, to the effect that a small parcel of considerable +value which she had been expecting was waiting for her at the +offices of the Aberdeen Shipping Company. Now, if you are well up +in your London, you will know that the office of the company is +in Fresno Street, which branches out of Upper Swandam Lane, where +you found me to-night. Mrs. St. Clair had her lunch, started for +the City, did some shopping, proceeded to the company's office, +got her packet, and found herself at exactly 4:35 walking through +Swandam Lane on her way back to the station. Have you followed me +so far?" + +"It is very clear." + +"If you remember, Monday was an exceedingly hot day, and Mrs. St. +Clair walked slowly, glancing about in the hope of seeing a cab, +as she did not like the neighbourhood in which she found herself. +While she was walking in this way down Swandam Lane, she suddenly +heard an ejaculation or cry, and was struck cold to see her +husband looking down at her and, as it seemed to her, beckoning +to her from a second-floor window. The window was open, and she +distinctly saw his face, which she describes as being terribly +agitated. He waved his hands frantically to her, and then +vanished from the window so suddenly that it seemed to her that +he had been plucked back by some irresistible force from behind. +One singular point which struck her quick feminine eye was that +although he wore some dark coat, such as he had started to town +in, he had on neither collar nor necktie. + +"Convinced that something was amiss with him, she rushed down the +steps--for the house was none other than the opium den in which +you found me to-night--and running through the front room she +attempted to ascend the stairs which led to the first floor. At +the foot of the stairs, however, she met this Lascar scoundrel of +whom I have spoken, who thrust her back and, aided by a Dane, who +acts as assistant there, pushed her out into the street. Filled +with the most maddening doubts and fears, she rushed down the +lane and, by rare good-fortune, met in Fresno Street a number of +constables with an inspector, all on their way to their beat. The +inspector and two men accompanied her back, and in spite of the +continued resistance of the proprietor, they made their way to +the room in which Mr. St. Clair had last been seen. There was no +sign of him there. In fact, in the whole of that floor there was +no one to be found save a crippled wretch of hideous aspect, who, +it seems, made his home there. Both he and the Lascar stoutly +swore that no one else had been in the front room during the +afternoon. So determined was their denial that the inspector was +staggered, and had almost come to believe that Mrs. St. Clair had +been deluded when, with a cry, she sprang at a small deal box +which lay upon the table and tore the lid from it. Out there fell +a cascade of children's bricks. It was the toy which he had +promised to bring home. + +"This discovery, and the evident confusion which the cripple +showed, made the inspector realise that the matter was serious. +The rooms were carefully examined, and results all pointed to an +abominable crime. The front room was plainly furnished as a +sitting-room and led into a small bedroom, which looked out upon +the back of one of the wharves. Between the wharf and the bedroom +window is a narrow strip, which is dry at low tide but is covered +at high tide with at least four and a half feet of water. The +bedroom window was a broad one and opened from below. On +examination traces of blood were to be seen upon the windowsill, +and several scattered drops were visible upon the wooden floor of +the bedroom. Thrust away behind a curtain in the front room were +all the clothes of Mr. Neville St. Clair, with the exception of +his coat. His boots, his socks, his hat, and his watch--all were +there. There were no signs of violence upon any of these +garments, and there were no other traces of Mr. Neville St. +Clair. Out of the window he must apparently have gone for no +other exit could be discovered, and the ominous bloodstains upon +the sill gave little promise that he could save himself by +swimming, for the tide was at its very highest at the moment of +the tragedy. + +"And now as to the villains who seemed to be immediately +implicated in the matter. The Lascar was known to be a man of the +vilest antecedents, but as, by Mrs. St. Clair's story, he was +known to have been at the foot of the stair within a very few +seconds of her husband's appearance at the window, he could +hardly have been more than an accessory to the crime. His defence +was one of absolute ignorance, and he protested that he had no +knowledge as to the doings of Hugh Boone, his lodger, and that he +could not account in any way for the presence of the missing +gentleman's clothes. + +"So much for the Lascar manager. Now for the sinister cripple who +lives upon the second floor of the opium den, and who was +certainly the last human being whose eyes rested upon Neville St. +Clair. His name is Hugh Boone, and his hideous face is one which +is familiar to every man who goes much to the City. He is a +professional beggar, though in order to avoid the police +regulations he pretends to a small trade in wax vestas. Some +little distance down Threadneedle Street, upon the left-hand +side, there is, as you may have remarked, a small angle in the +wall. Here it is that this creature takes his daily seat, +cross-legged with his tiny stock of matches on his lap, and as he +is a piteous spectacle a small rain of charity descends into the +greasy leather cap which lies upon the pavement beside him. I +have watched the fellow more than once before ever I thought of +making his professional acquaintance, and I have been surprised +at the harvest which he has reaped in a short time. His +appearance, you see, is so remarkable that no one can pass him +without observing him. A shock of orange hair, a pale face +disfigured by a horrible scar, which, by its contraction, has +turned up the outer edge of his upper lip, a bulldog chin, and a +pair of very penetrating dark eyes, which present a singular +contrast to the colour of his hair, all mark him out from amid +the common crowd of mendicants and so, too, does his wit, for he +is ever ready with a reply to any piece of chaff which may be +thrown at him by the passers-by. This is the man whom we now +learn to have been the lodger at the opium den, and to have been +the last man to see the gentleman of whom we are in quest." + +"But a cripple!" said I. "What could he have done single-handed +against a man in the prime of life?" + +"He is a cripple in the sense that he walks with a limp; but in +other respects he appears to be a powerful and well-nurtured man. +Surely your medical experience would tell you, Watson, that +weakness in one limb is often compensated for by exceptional +strength in the others." + +"Pray continue your narrative." + +"Mrs. St. Clair had fainted at the sight of the blood upon the +window, and she was escorted home in a cab by the police, as her +presence could be of no help to them in their investigations. +Inspector Barton, who had charge of the case, made a very careful +examination of the premises, but without finding anything which +threw any light upon the matter. One mistake had been made in not +arresting Boone instantly, as he was allowed some few minutes +during which he might have communicated with his friend the +Lascar, but this fault was soon remedied, and he was seized and +searched, without anything being found which could incriminate +him. There were, it is true, some blood-stains upon his right +shirt-sleeve, but he pointed to his ring-finger, which had been +cut near the nail, and explained that the bleeding came from +there, adding that he had been to the window not long before, and +that the stains which had been observed there came doubtless from +the same source. He denied strenuously having ever seen Mr. +Neville St. Clair and swore that the presence of the clothes in +his room was as much a mystery to him as to the police. As to +Mrs. St. Clair's assertion that she had actually seen her husband +at the window, he declared that she must have been either mad or +dreaming. He was removed, loudly protesting, to the +police-station, while the inspector remained upon the premises in +the hope that the ebbing tide might afford some fresh clue. + +"And it did, though they hardly found upon the mud-bank what they +had feared to find. It was Neville St. Clair's coat, and not +Neville St. Clair, which lay uncovered as the tide receded. And +what do you think they found in the pockets?" + +"I cannot imagine." + +"No, I don't think you would guess. Every pocket stuffed with +pennies and half-pennies--421 pennies and 270 half-pennies. It +was no wonder that it had not been swept away by the tide. But a +human body is a different matter. There is a fierce eddy between +the wharf and the house. It seemed likely enough that the +weighted coat had remained when the stripped body had been sucked +away into the river." + +"But I understand that all the other clothes were found in the +room. Would the body be dressed in a coat alone?" + +"No, sir, but the facts might be met speciously enough. Suppose +that this man Boone had thrust Neville St. Clair through the +window, there is no human eye which could have seen the deed. +What would he do then? It would of course instantly strike him +that he must get rid of the tell-tale garments. He would seize +the coat, then, and be in the act of throwing it out, when it +would occur to him that it would swim and not sink. He has little +time, for he has heard the scuffle downstairs when the wife tried +to force her way up, and perhaps he has already heard from his +Lascar confederate that the police are hurrying up the street. +There is not an instant to be lost. He rushes to some secret +hoard, where he has accumulated the fruits of his beggary, and he +stuffs all the coins upon which he can lay his hands into the +pockets to make sure of the coat's sinking. He throws it out, and +would have done the same with the other garments had not he heard +the rush of steps below, and only just had time to close the +window when the police appeared." + +"It certainly sounds feasible." + +"Well, we will take it as a working hypothesis for want of a +better. Boone, as I have told you, was arrested and taken to the +station, but it could not be shown that there had ever before +been anything against him. He had for years been known as a +professional beggar, but his life appeared to have been a very +quiet and innocent one. There the matter stands at present, and +the questions which have to be solved--what Neville St. Clair was +doing in the opium den, what happened to him when there, where is +he now, and what Hugh Boone had to do with his disappearance--are +all as far from a solution as ever. I confess that I cannot +recall any case within my experience which looked at the first +glance so simple and yet which presented such difficulties." + +While Sherlock Holmes had been detailing this singular series of +events, we had been whirling through the outskirts of the great +town until the last straggling houses had been left behind, and +we rattled along with a country hedge upon either side of us. +Just as he finished, however, we drove through two scattered +villages, where a few lights still glimmered in the windows. + +"We are on the outskirts of Lee," said my companion. "We have +touched on three English counties in our short drive, starting in +Middlesex, passing over an angle of Surrey, and ending in Kent. +See that light among the trees? That is The Cedars, and beside +that lamp sits a woman whose anxious ears have already, I have +little doubt, caught the clink of our horse's feet." + +"But why are you not conducting the case from Baker Street?" I +asked. + +"Because there are many inquiries which must be made out here. +Mrs. St. Clair has most kindly put two rooms at my disposal, and +you may rest assured that she will have nothing but a welcome for +my friend and colleague. I hate to meet her, Watson, when I have +no news of her husband. Here we are. Whoa, there, whoa!" + +We had pulled up in front of a large villa which stood within its +own grounds. A stable-boy had run out to the horse's head, and +springing down, I followed Holmes up the small, winding +gravel-drive which led to the house. As we approached, the door +flew open, and a little blonde woman stood in the opening, clad +in some sort of light mousseline de soie, with a touch of fluffy +pink chiffon at her neck and wrists. She stood with her figure +outlined against the flood of light, one hand upon the door, one +half-raised in her eagerness, her body slightly bent, her head +and face protruded, with eager eyes and parted lips, a standing +question. + +"Well?" she cried, "well?" And then, seeing that there were two +of us, she gave a cry of hope which sank into a groan as she saw +that my companion shook his head and shrugged his shoulders. + +"No good news?" + +"None." + +"No bad?" + +"No." + +"Thank God for that. But come in. You must be weary, for you have +had a long day." + +"This is my friend, Dr. Watson. He has been of most vital use to +me in several of my cases, and a lucky chance has made it +possible for me to bring him out and associate him with this +investigation." + +"I am delighted to see you," said she, pressing my hand warmly. +"You will, I am sure, forgive anything that may be wanting in our +arrangements, when you consider the blow which has come so +suddenly upon us." + +"My dear madam," said I, "I am an old campaigner, and if I were +not I can very well see that no apology is needed. If I can be of +any assistance, either to you or to my friend here, I shall be +indeed happy." + +"Now, Mr. Sherlock Holmes," said the lady as we entered a +well-lit dining-room, upon the table of which a cold supper had +been laid out, "I should very much like to ask you one or two +plain questions, to which I beg that you will give a plain +answer." + +"Certainly, madam." + +"Do not trouble about my feelings. I am not hysterical, nor given +to fainting. I simply wish to hear your real, real opinion." + +"Upon what point?" + +"In your heart of hearts, do you think that Neville is alive?" + +Sherlock Holmes seemed to be embarrassed by the question. +"Frankly, now!" she repeated, standing upon the rug and looking +keenly down at him as he leaned back in a basket-chair. + +"Frankly, then, madam, I do not." + +"You think that he is dead?" + +"I do." + +"Murdered?" + +"I don't say that. Perhaps." + +"And on what day did he meet his death?" + +"On Monday." + +"Then perhaps, Mr. Holmes, you will be good enough to explain how +it is that I have received a letter from him to-day." + +Sherlock Holmes sprang out of his chair as if he had been +galvanised. + +"What!" he roared. + +"Yes, to-day." She stood smiling, holding up a little slip of +paper in the air. + +"May I see it?" + +"Certainly." + +He snatched it from her in his eagerness, and smoothing it out +upon the table he drew over the lamp and examined it intently. I +had left my chair and was gazing at it over his shoulder. The +envelope was a very coarse one and was stamped with the Gravesend +postmark and with the date of that very day, or rather of the day +before, for it was considerably after midnight. + +"Coarse writing," murmured Holmes. "Surely this is not your +husband's writing, madam." + +"No, but the enclosure is." + +"I perceive also that whoever addressed the envelope had to go +and inquire as to the address." + +"How can you tell that?" + +"The name, you see, is in perfectly black ink, which has dried +itself. The rest is of the greyish colour, which shows that +blotting-paper has been used. If it had been written straight +off, and then blotted, none would be of a deep black shade. This +man has written the name, and there has then been a pause before +he wrote the address, which can only mean that he was not +familiar with it. It is, of course, a trifle, but there is +nothing so important as trifles. Let us now see the letter. Ha! +there has been an enclosure here!" + +"Yes, there was a ring. His signet-ring." + +"And you are sure that this is your husband's hand?" + +"One of his hands." + +"One?" + +"His hand when he wrote hurriedly. It is very unlike his usual +writing, and yet I know it well." + +"'Dearest do not be frightened. All will come well. There is a +huge error which it may take some little time to rectify. +Wait in patience.--NEVILLE.' Written in pencil upon the fly-leaf +of a book, octavo size, no water-mark. Hum! Posted to-day in +Gravesend by a man with a dirty thumb. Ha! And the flap has been +gummed, if I am not very much in error, by a person who had been +chewing tobacco. And you have no doubt that it is your husband's +hand, madam?" + +"None. Neville wrote those words." + +"And they were posted to-day at Gravesend. Well, Mrs. St. Clair, +the clouds lighten, though I should not venture to say that the +danger is over." + +"But he must be alive, Mr. Holmes." + +"Unless this is a clever forgery to put us on the wrong scent. +The ring, after all, proves nothing. It may have been taken from +him." + +"No, no; it is, it is his very own writing!" + +"Very well. It may, however, have been written on Monday and only +posted to-day." + +"That is possible." + +"If so, much may have happened between." + +"Oh, you must not discourage me, Mr. Holmes. I know that all is +well with him. There is so keen a sympathy between us that I +should know if evil came upon him. On the very day that I saw him +last he cut himself in the bedroom, and yet I in the dining-room +rushed upstairs instantly with the utmost certainty that +something had happened. Do you think that I would respond to such +a trifle and yet be ignorant of his death?" + +"I have seen too much not to know that the impression of a woman +may be more valuable than the conclusion of an analytical +reasoner. And in this letter you certainly have a very strong +piece of evidence to corroborate your view. But if your husband +is alive and able to write letters, why should he remain away +from you?" + +"I cannot imagine. It is unthinkable." + +"And on Monday he made no remarks before leaving you?" + +"No." + +"And you were surprised to see him in Swandam Lane?" + +"Very much so." + +"Was the window open?" + +"Yes." + +"Then he might have called to you?" + +"He might." + +"He only, as I understand, gave an inarticulate cry?" + +"Yes." + +"A call for help, you thought?" + +"Yes. He waved his hands." + +"But it might have been a cry of surprise. Astonishment at the +unexpected sight of you might cause him to throw up his hands?" + +"It is possible." + +"And you thought he was pulled back?" + +"He disappeared so suddenly." + +"He might have leaped back. You did not see anyone else in the +room?" + +"No, but this horrible man confessed to having been there, and +the Lascar was at the foot of the stairs." + +"Quite so. Your husband, as far as you could see, had his +ordinary clothes on?" + +"But without his collar or tie. I distinctly saw his bare +throat." + +"Had he ever spoken of Swandam Lane?" + +"Never." + +"Had he ever showed any signs of having taken opium?" + +"Never." + +"Thank you, Mrs. St. Clair. Those are the principal points about +which I wished to be absolutely clear. We shall now have a little +supper and then retire, for we may have a very busy day +to-morrow." + +A large and comfortable double-bedded room had been placed at our +disposal, and I was quickly between the sheets, for I was weary +after my night of adventure. Sherlock Holmes was a man, however, +who, when he had an unsolved problem upon his mind, would go for +days, and even for a week, without rest, turning it over, +rearranging his facts, looking at it from every point of view +until he had either fathomed it or convinced himself that his +data were insufficient. It was soon evident to me that he was now +preparing for an all-night sitting. He took off his coat and +waistcoat, put on a large blue dressing-gown, and then wandered +about the room collecting pillows from his bed and cushions from +the sofa and armchairs. With these he constructed a sort of +Eastern divan, upon which he perched himself cross-legged, with +an ounce of shag tobacco and a box of matches laid out in front +of him. In the dim light of the lamp I saw him sitting there, an +old briar pipe between his lips, his eyes fixed vacantly upon the +corner of the ceiling, the blue smoke curling up from him, +silent, motionless, with the light shining upon his strong-set +aquiline features. So he sat as I dropped off to sleep, and so he +sat when a sudden ejaculation caused me to wake up, and I found +the summer sun shining into the apartment. The pipe was still +between his lips, the smoke still curled upward, and the room was +full of a dense tobacco haze, but nothing remained of the heap of +shag which I had seen upon the previous night. + +"Awake, Watson?" he asked. + +"Yes." + +"Game for a morning drive?" + +"Certainly." + +"Then dress. No one is stirring yet, but I know where the +stable-boy sleeps, and we shall soon have the trap out." He +chuckled to himself as he spoke, his eyes twinkled, and he seemed +a different man to the sombre thinker of the previous night. + +As I dressed I glanced at my watch. It was no wonder that no one +was stirring. It was twenty-five minutes past four. I had hardly +finished when Holmes returned with the news that the boy was +putting in the horse. + +"I want to test a little theory of mine," said he, pulling on his +boots. "I think, Watson, that you are now standing in the +presence of one of the most absolute fools in Europe. I deserve +to be kicked from here to Charing Cross. But I think I have the +key of the affair now." + +"And where is it?" I asked, smiling. + +"In the bathroom," he answered. "Oh, yes, I am not joking," he +continued, seeing my look of incredulity. "I have just been +there, and I have taken it out, and I have got it in this +Gladstone bag. Come on, my boy, and we shall see whether it will +not fit the lock." + +We made our way downstairs as quietly as possible, and out into +the bright morning sunshine. In the road stood our horse and +trap, with the half-clad stable-boy waiting at the head. We both +sprang in, and away we dashed down the London Road. A few country +carts were stirring, bearing in vegetables to the metropolis, but +the lines of villas on either side were as silent and lifeless as +some city in a dream. + +"It has been in some points a singular case," said Holmes, +flicking the horse on into a gallop. "I confess that I have been +as blind as a mole, but it is better to learn wisdom late than +never to learn it at all." + +In town the earliest risers were just beginning to look sleepily +from their windows as we drove through the streets of the Surrey +side. Passing down the Waterloo Bridge Road we crossed over the +river, and dashing up Wellington Street wheeled sharply to the +right and found ourselves in Bow Street. Sherlock Holmes was well +known to the force, and the two constables at the door saluted +him. One of them held the horse's head while the other led us in. + +"Who is on duty?" asked Holmes. + +"Inspector Bradstreet, sir." + +"Ah, Bradstreet, how are you?" A tall, stout official had come +down the stone-flagged passage, in a peaked cap and frogged +jacket. "I wish to have a quiet word with you, Bradstreet." +"Certainly, Mr. Holmes. Step into my room here." It was a small, +office-like room, with a huge ledger upon the table, and a +telephone projecting from the wall. The inspector sat down at his +desk. + +"What can I do for you, Mr. Holmes?" + +"I called about that beggarman, Boone--the one who was charged +with being concerned in the disappearance of Mr. Neville St. +Clair, of Lee." + +"Yes. He was brought up and remanded for further inquiries." + +"So I heard. You have him here?" + +"In the cells." + +"Is he quiet?" + +"Oh, he gives no trouble. But he is a dirty scoundrel." + +"Dirty?" + +"Yes, it is all we can do to make him wash his hands, and his +face is as black as a tinker's. Well, when once his case has been +settled, he will have a regular prison bath; and I think, if you +saw him, you would agree with me that he needed it." + +"I should like to see him very much." + +"Would you? That is easily done. Come this way. You can leave +your bag." + +"No, I think that I'll take it." + +"Very good. Come this way, if you please." He led us down a +passage, opened a barred door, passed down a winding stair, and +brought us to a whitewashed corridor with a line of doors on each +side. + +"The third on the right is his," said the inspector. "Here it +is!" He quietly shot back a panel in the upper part of the door +and glanced through. + +"He is asleep," said he. "You can see him very well." + +We both put our eyes to the grating. The prisoner lay with his +face towards us, in a very deep sleep, breathing slowly and +heavily. He was a middle-sized man, coarsely clad as became his +calling, with a coloured shirt protruding through the rent in his +tattered coat. He was, as the inspector had said, extremely +dirty, but the grime which covered his face could not conceal its +repulsive ugliness. A broad wheal from an old scar ran right +across it from eye to chin, and by its contraction had turned up +one side of the upper lip, so that three teeth were exposed in a +perpetual snarl. A shock of very bright red hair grew low over +his eyes and forehead. + +"He's a beauty, isn't he?" said the inspector. + +"He certainly needs a wash," remarked Holmes. "I had an idea that +he might, and I took the liberty of bringing the tools with me." +He opened the Gladstone bag as he spoke, and took out, to my +astonishment, a very large bath-sponge. + +"He! he! You are a funny one," chuckled the inspector. + +"Now, if you will have the great goodness to open that door very +quietly, we will soon make him cut a much more respectable +figure." + +"Well, I don't know why not," said the inspector. "He doesn't +look a credit to the Bow Street cells, does he?" He slipped his +key into the lock, and we all very quietly entered the cell. The +sleeper half turned, and then settled down once more into a deep +slumber. Holmes stooped to the water-jug, moistened his sponge, +and then rubbed it twice vigorously across and down the +prisoner's face. + +"Let me introduce you," he shouted, "to Mr. Neville St. Clair, of +Lee, in the county of Kent." + +Never in my life have I seen such a sight. The man's face peeled +off under the sponge like the bark from a tree. Gone was the +coarse brown tint! Gone, too, was the horrid scar which had +seamed it across, and the twisted lip which had given the +repulsive sneer to the face! A twitch brought away the tangled +red hair, and there, sitting up in his bed, was a pale, +sad-faced, refined-looking man, black-haired and smooth-skinned, +rubbing his eyes and staring about him with sleepy bewilderment. +Then suddenly realising the exposure, he broke into a scream and +threw himself down with his face to the pillow. + +"Great heavens!" cried the inspector, "it is, indeed, the missing +man. I know him from the photograph." + +The prisoner turned with the reckless air of a man who abandons +himself to his destiny. "Be it so," said he. "And pray what am I +charged with?" + +"With making away with Mr. Neville St.-- Oh, come, you can't be +charged with that unless they make a case of attempted suicide of +it," said the inspector with a grin. "Well, I have been +twenty-seven years in the force, but this really takes the cake." + +"If I am Mr. Neville St. Clair, then it is obvious that no crime +has been committed, and that, therefore, I am illegally +detained." + +"No crime, but a very great error has been committed," said +Holmes. "You would have done better to have trusted your wife." + +"It was not the wife; it was the children," groaned the prisoner. +"God help me, I would not have them ashamed of their father. My +God! What an exposure! What can I do?" + +Sherlock Holmes sat down beside him on the couch and patted him +kindly on the shoulder. + +"If you leave it to a court of law to clear the matter up," said +he, "of course you can hardly avoid publicity. On the other hand, +if you convince the police authorities that there is no possible +case against you, I do not know that there is any reason that the +details should find their way into the papers. Inspector +Bradstreet would, I am sure, make notes upon anything which you +might tell us and submit it to the proper authorities. The case +would then never go into court at all." + +"God bless you!" cried the prisoner passionately. "I would have +endured imprisonment, ay, even execution, rather than have left +my miserable secret as a family blot to my children. + +"You are the first who have ever heard my story. My father was a +schoolmaster in Chesterfield, where I received an excellent +education. I travelled in my youth, took to the stage, and +finally became a reporter on an evening paper in London. One day +my editor wished to have a series of articles upon begging in the +metropolis, and I volunteered to supply them. There was the point +from which all my adventures started. It was only by trying +begging as an amateur that I could get the facts upon which to +base my articles. When an actor I had, of course, learned all the +secrets of making up, and had been famous in the green-room for +my skill. I took advantage now of my attainments. I painted my +face, and to make myself as pitiable as possible I made a good +scar and fixed one side of my lip in a twist by the aid of a +small slip of flesh-coloured plaster. Then with a red head of +hair, and an appropriate dress, I took my station in the business +part of the city, ostensibly as a match-seller but really as a +beggar. For seven hours I plied my trade, and when I returned +home in the evening I found to my surprise that I had received no +less than 26s. 4d. + +"I wrote my articles and thought little more of the matter until, +some time later, I backed a bill for a friend and had a writ +served upon me for 25 pounds. I was at my wit's end where to get +the money, but a sudden idea came to me. I begged a fortnight's +grace from the creditor, asked for a holiday from my employers, +and spent the time in begging in the City under my disguise. In +ten days I had the money and had paid the debt. + +"Well, you can imagine how hard it was to settle down to arduous +work at 2 pounds a week when I knew that I could earn as much in +a day by smearing my face with a little paint, laying my cap on +the ground, and sitting still. It was a long fight between my +pride and the money, but the dollars won at last, and I threw up +reporting and sat day after day in the corner which I had first +chosen, inspiring pity by my ghastly face and filling my pockets +with coppers. Only one man knew my secret. He was the keeper of a +low den in which I used to lodge in Swandam Lane, where I could +every morning emerge as a squalid beggar and in the evenings +transform myself into a well-dressed man about town. This fellow, +a Lascar, was well paid by me for his rooms, so that I knew that +my secret was safe in his possession. + +"Well, very soon I found that I was saving considerable sums of +money. I do not mean that any beggar in the streets of London +could earn 700 pounds a year--which is less than my average +takings--but I had exceptional advantages in my power of making +up, and also in a facility of repartee, which improved by +practice and made me quite a recognised character in the City. +All day a stream of pennies, varied by silver, poured in upon me, +and it was a very bad day in which I failed to take 2 pounds. + +"As I grew richer I grew more ambitious, took a house in the +country, and eventually married, without anyone having a +suspicion as to my real occupation. My dear wife knew that I had +business in the City. She little knew what. + +"Last Monday I had finished for the day and was dressing in my +room above the opium den when I looked out of my window and saw, +to my horror and astonishment, that my wife was standing in the +street, with her eyes fixed full upon me. I gave a cry of +surprise, threw up my arms to cover my face, and, rushing to my +confidant, the Lascar, entreated him to prevent anyone from +coming up to me. I heard her voice downstairs, but I knew that +she could not ascend. Swiftly I threw off my clothes, pulled on +those of a beggar, and put on my pigments and wig. Even a wife's +eyes could not pierce so complete a disguise. But then it +occurred to me that there might be a search in the room, and that +the clothes might betray me. I threw open the window, reopening +by my violence a small cut which I had inflicted upon myself in +the bedroom that morning. Then I seized my coat, which was +weighted by the coppers which I had just transferred to it from +the leather bag in which I carried my takings. I hurled it out of +the window, and it disappeared into the Thames. The other clothes +would have followed, but at that moment there was a rush of +constables up the stair, and a few minutes after I found, rather, +I confess, to my relief, that instead of being identified as Mr. +Neville St. Clair, I was arrested as his murderer. + +"I do not know that there is anything else for me to explain. I +was determined to preserve my disguise as long as possible, and +hence my preference for a dirty face. Knowing that my wife would +be terribly anxious, I slipped off my ring and confided it to the +Lascar at a moment when no constable was watching me, together +with a hurried scrawl, telling her that she had no cause to +fear." + +"That note only reached her yesterday," said Holmes. + +"Good God! What a week she must have spent!" + +"The police have watched this Lascar," said Inspector Bradstreet, +"and I can quite understand that he might find it difficult to +post a letter unobserved. Probably he handed it to some sailor +customer of his, who forgot all about it for some days." + +"That was it," said Holmes, nodding approvingly; "I have no doubt +of it. But have you never been prosecuted for begging?" + +"Many times; but what was a fine to me?" + +"It must stop here, however," said Bradstreet. "If the police are +to hush this thing up, there must be no more of Hugh Boone." + +"I have sworn it by the most solemn oaths which a man can take." + +"In that case I think that it is probable that no further steps +may be taken. But if you are found again, then all must come out. +I am sure, Mr. Holmes, that we are very much indebted to you for +having cleared the matter up. I wish I knew how you reach your +results." + +"I reached this one," said my friend, "by sitting upon five +pillows and consuming an ounce of shag. I think, Watson, that if +we drive to Baker Street we shall just be in time for breakfast." + + + +VII. THE ADVENTURE OF THE BLUE CARBUNCLE + +I had called upon my friend Sherlock Holmes upon the second +morning after Christmas, with the intention of wishing him the +compliments of the season. He was lounging upon the sofa in a +purple dressing-gown, a pipe-rack within his reach upon the +right, and a pile of crumpled morning papers, evidently newly +studied, near at hand. Beside the couch was a wooden chair, and +on the angle of the back hung a very seedy and disreputable +hard-felt hat, much the worse for wear, and cracked in several +places. A lens and a forceps lying upon the seat of the chair +suggested that the hat had been suspended in this manner for the +purpose of examination. + +"You are engaged," said I; "perhaps I interrupt you." + +"Not at all. I am glad to have a friend with whom I can discuss +my results. The matter is a perfectly trivial one"--he jerked his +thumb in the direction of the old hat--"but there are points in +connection with it which are not entirely devoid of interest and +even of instruction." + +I seated myself in his armchair and warmed my hands before his +crackling fire, for a sharp frost had set in, and the windows +were thick with the ice crystals. "I suppose," I remarked, "that, +homely as it looks, this thing has some deadly story linked on to +it--that it is the clue which will guide you in the solution of +some mystery and the punishment of some crime." + +"No, no. No crime," said Sherlock Holmes, laughing. "Only one of +those whimsical little incidents which will happen when you have +four million human beings all jostling each other within the +space of a few square miles. Amid the action and reaction of so +dense a swarm of humanity, every possible combination of events +may be expected to take place, and many a little problem will be +presented which may be striking and bizarre without being +criminal. We have already had experience of such." + +"So much so," I remarked, "that of the last six cases which I +have added to my notes, three have been entirely free of any +legal crime." + +"Precisely. You allude to my attempt to recover the Irene Adler +papers, to the singular case of Miss Mary Sutherland, and to the +adventure of the man with the twisted lip. Well, I have no doubt +that this small matter will fall into the same innocent category. +You know Peterson, the commissionaire?" + +"Yes." + +"It is to him that this trophy belongs." + +"It is his hat." + +"No, no, he found it. Its owner is unknown. I beg that you will +look upon it not as a battered billycock but as an intellectual +problem. And, first, as to how it came here. It arrived upon +Christmas morning, in company with a good fat goose, which is, I +have no doubt, roasting at this moment in front of Peterson's +fire. The facts are these: about four o'clock on Christmas +morning, Peterson, who, as you know, is a very honest fellow, was +returning from some small jollification and was making his way +homeward down Tottenham Court Road. In front of him he saw, in +the gaslight, a tallish man, walking with a slight stagger, and +carrying a white goose slung over his shoulder. As he reached the +corner of Goodge Street, a row broke out between this stranger +and a little knot of roughs. One of the latter knocked off the +man's hat, on which he raised his stick to defend himself and, +swinging it over his head, smashed the shop window behind him. +Peterson had rushed forward to protect the stranger from his +assailants; but the man, shocked at having broken the window, and +seeing an official-looking person in uniform rushing towards him, +dropped his goose, took to his heels, and vanished amid the +labyrinth of small streets which lie at the back of Tottenham +Court Road. The roughs had also fled at the appearance of +Peterson, so that he was left in possession of the field of +battle, and also of the spoils of victory in the shape of this +battered hat and a most unimpeachable Christmas goose." + +"Which surely he restored to their owner?" + +"My dear fellow, there lies the problem. It is true that 'For +Mrs. Henry Baker' was printed upon a small card which was tied to +the bird's left leg, and it is also true that the initials 'H. +B.' are legible upon the lining of this hat, but as there are +some thousands of Bakers, and some hundreds of Henry Bakers in +this city of ours, it is not easy to restore lost property to any +one of them." + +"What, then, did Peterson do?" + +"He brought round both hat and goose to me on Christmas morning, +knowing that even the smallest problems are of interest to me. +The goose we retained until this morning, when there were signs +that, in spite of the slight frost, it would be well that it +should be eaten without unnecessary delay. Its finder has carried +it off, therefore, to fulfil the ultimate destiny of a goose, +while I continue to retain the hat of the unknown gentleman who +lost his Christmas dinner." + +"Did he not advertise?" + +"No." + +"Then, what clue could you have as to his identity?" + +"Only as much as we can deduce." + +"From his hat?" + +"Precisely." + +"But you are joking. What can you gather from this old battered +felt?" + +"Here is my lens. You know my methods. What can you gather +yourself as to the individuality of the man who has worn this +article?" + +I took the tattered object in my hands and turned it over rather +ruefully. It was a very ordinary black hat of the usual round +shape, hard and much the worse for wear. The lining had been of +red silk, but was a good deal discoloured. There was no maker's +name; but, as Holmes had remarked, the initials "H. B." were +scrawled upon one side. It was pierced in the brim for a +hat-securer, but the elastic was missing. For the rest, it was +cracked, exceedingly dusty, and spotted in several places, +although there seemed to have been some attempt to hide the +discoloured patches by smearing them with ink. + +"I can see nothing," said I, handing it back to my friend. + +"On the contrary, Watson, you can see everything. You fail, +however, to reason from what you see. You are too timid in +drawing your inferences." + +"Then, pray tell me what it is that you can infer from this hat?" + +He picked it up and gazed at it in the peculiar introspective +fashion which was characteristic of him. "It is perhaps less +suggestive than it might have been," he remarked, "and yet there +are a few inferences which are very distinct, and a few others +which represent at least a strong balance of probability. That +the man was highly intellectual is of course obvious upon the +face of it, and also that he was fairly well-to-do within the +last three years, although he has now fallen upon evil days. He +had foresight, but has less now than formerly, pointing to a +moral retrogression, which, when taken with the decline of his +fortunes, seems to indicate some evil influence, probably drink, +at work upon him. This may account also for the obvious fact that +his wife has ceased to love him." + +"My dear Holmes!" + +"He has, however, retained some degree of self-respect," he +continued, disregarding my remonstrance. "He is a man who leads a +sedentary life, goes out little, is out of training entirely, is +middle-aged, has grizzled hair which he has had cut within the +last few days, and which he anoints with lime-cream. These are +the more patent facts which are to be deduced from his hat. Also, +by the way, that it is extremely improbable that he has gas laid +on in his house." + +"You are certainly joking, Holmes." + +"Not in the least. Is it possible that even now, when I give you +these results, you are unable to see how they are attained?" + +"I have no doubt that I am very stupid, but I must confess that I +am unable to follow you. For example, how did you deduce that +this man was intellectual?" + +For answer Holmes clapped the hat upon his head. It came right +over the forehead and settled upon the bridge of his nose. "It is +a question of cubic capacity," said he; "a man with so large a +brain must have something in it." + +"The decline of his fortunes, then?" + +"This hat is three years old. These flat brims curled at the edge +came in then. It is a hat of the very best quality. Look at the +band of ribbed silk and the excellent lining. If this man could +afford to buy so expensive a hat three years ago, and has had no +hat since, then he has assuredly gone down in the world." + +"Well, that is clear enough, certainly. But how about the +foresight and the moral retrogression?" + +Sherlock Holmes laughed. "Here is the foresight," said he putting +his finger upon the little disc and loop of the hat-securer. +"They are never sold upon hats. If this man ordered one, it is a +sign of a certain amount of foresight, since he went out of his +way to take this precaution against the wind. But since we see +that he has broken the elastic and has not troubled to replace +it, it is obvious that he has less foresight now than formerly, +which is a distinct proof of a weakening nature. On the other +hand, he has endeavoured to conceal some of these stains upon the +felt by daubing them with ink, which is a sign that he has not +entirely lost his self-respect." + +"Your reasoning is certainly plausible." + +"The further points, that he is middle-aged, that his hair is +grizzled, that it has been recently cut, and that he uses +lime-cream, are all to be gathered from a close examination of the +lower part of the lining. The lens discloses a large number of +hair-ends, clean cut by the scissors of the barber. They all +appear to be adhesive, and there is a distinct odour of +lime-cream. This dust, you will observe, is not the gritty, grey +dust of the street but the fluffy brown dust of the house, +showing that it has been hung up indoors most of the time, while +the marks of moisture upon the inside are proof positive that the +wearer perspired very freely, and could therefore, hardly be in +the best of training." + +"But his wife--you said that she had ceased to love him." + +"This hat has not been brushed for weeks. When I see you, my dear +Watson, with a week's accumulation of dust upon your hat, and +when your wife allows you to go out in such a state, I shall fear +that you also have been unfortunate enough to lose your wife's +affection." + +"But he might be a bachelor." + +"Nay, he was bringing home the goose as a peace-offering to his +wife. Remember the card upon the bird's leg." + +"You have an answer to everything. But how on earth do you deduce +that the gas is not laid on in his house?" + +"One tallow stain, or even two, might come by chance; but when I +see no less than five, I think that there can be little doubt +that the individual must be brought into frequent contact with +burning tallow--walks upstairs at night probably with his hat in +one hand and a guttering candle in the other. Anyhow, he never +got tallow-stains from a gas-jet. Are you satisfied?" + +"Well, it is very ingenious," said I, laughing; "but since, as +you said just now, there has been no crime committed, and no harm +done save the loss of a goose, all this seems to be rather a +waste of energy." + +Sherlock Holmes had opened his mouth to reply, when the door flew +open, and Peterson, the commissionaire, rushed into the apartment +with flushed cheeks and the face of a man who is dazed with +astonishment. + +"The goose, Mr. Holmes! The goose, sir!" he gasped. + +"Eh? What of it, then? Has it returned to life and flapped off +through the kitchen window?" Holmes twisted himself round upon +the sofa to get a fairer view of the man's excited face. + +"See here, sir! See what my wife found in its crop!" He held out +his hand and displayed upon the centre of the palm a brilliantly +scintillating blue stone, rather smaller than a bean in size, but +of such purity and radiance that it twinkled like an electric +point in the dark hollow of his hand. + +Sherlock Holmes sat up with a whistle. "By Jove, Peterson!" said +he, "this is treasure trove indeed. I suppose you know what you +have got?" + +"A diamond, sir? A precious stone. It cuts into glass as though +it were putty." + +"It's more than a precious stone. It is the precious stone." + +"Not the Countess of Morcar's blue carbuncle!" I ejaculated. + +"Precisely so. I ought to know its size and shape, seeing that I +have read the advertisement about it in The Times every day +lately. It is absolutely unique, and its value can only be +conjectured, but the reward offered of 1000 pounds is certainly +not within a twentieth part of the market price." + +"A thousand pounds! Great Lord of mercy!" The commissionaire +plumped down into a chair and stared from one to the other of us. + +"That is the reward, and I have reason to know that there are +sentimental considerations in the background which would induce +the Countess to part with half her fortune if she could but +recover the gem." + +"It was lost, if I remember aright, at the Hotel Cosmopolitan," I +remarked. + +"Precisely so, on December 22nd, just five days ago. John Horner, +a plumber, was accused of having abstracted it from the lady's +jewel-case. The evidence against him was so strong that the case +has been referred to the Assizes. I have some account of the +matter here, I believe." He rummaged amid his newspapers, +glancing over the dates, until at last he smoothed one out, +doubled it over, and read the following paragraph: + +"Hotel Cosmopolitan Jewel Robbery. John Horner, 26, plumber, was +brought up upon the charge of having upon the 22nd inst., +abstracted from the jewel-case of the Countess of Morcar the +valuable gem known as the blue carbuncle. James Ryder, +upper-attendant at the hotel, gave his evidence to the effect +that he had shown Horner up to the dressing-room of the Countess +of Morcar upon the day of the robbery in order that he might +solder the second bar of the grate, which was loose. He had +remained with Horner some little time, but had finally been +called away. On returning, he found that Horner had disappeared, +that the bureau had been forced open, and that the small morocco +casket in which, as it afterwards transpired, the Countess was +accustomed to keep her jewel, was lying empty upon the +dressing-table. Ryder instantly gave the alarm, and Horner was +arrested the same evening; but the stone could not be found +either upon his person or in his rooms. Catherine Cusack, maid to +the Countess, deposed to having heard Ryder's cry of dismay on +discovering the robbery, and to having rushed into the room, +where she found matters as described by the last witness. +Inspector Bradstreet, B division, gave evidence as to the arrest +of Horner, who struggled frantically, and protested his innocence +in the strongest terms. Evidence of a previous conviction for +robbery having been given against the prisoner, the magistrate +refused to deal summarily with the offence, but referred it to +the Assizes. Horner, who had shown signs of intense emotion +during the proceedings, fainted away at the conclusion and was +carried out of court." + +"Hum! So much for the police-court," said Holmes thoughtfully, +tossing aside the paper. "The question for us now to solve is the +sequence of events leading from a rifled jewel-case at one end to +the crop of a goose in Tottenham Court Road at the other. You +see, Watson, our little deductions have suddenly assumed a much +more important and less innocent aspect. Here is the stone; the +stone came from the goose, and the goose came from Mr. Henry +Baker, the gentleman with the bad hat and all the other +characteristics with which I have bored you. So now we must set +ourselves very seriously to finding this gentleman and +ascertaining what part he has played in this little mystery. To +do this, we must try the simplest means first, and these lie +undoubtedly in an advertisement in all the evening papers. If +this fail, I shall have recourse to other methods." + +"What will you say?" + +"Give me a pencil and that slip of paper. Now, then: 'Found at +the corner of Goodge Street, a goose and a black felt hat. Mr. +Henry Baker can have the same by applying at 6:30 this evening at +221B, Baker Street.' That is clear and concise." + +"Very. But will he see it?" + +"Well, he is sure to keep an eye on the papers, since, to a poor +man, the loss was a heavy one. He was clearly so scared by his +mischance in breaking the window and by the approach of Peterson +that he thought of nothing but flight, but since then he must +have bitterly regretted the impulse which caused him to drop his +bird. Then, again, the introduction of his name will cause him to +see it, for everyone who knows him will direct his attention to +it. Here you are, Peterson, run down to the advertising agency +and have this put in the evening papers." + +"In which, sir?" + +"Oh, in the Globe, Star, Pall Mall, St. James's, Evening News, +Standard, Echo, and any others that occur to you." + +"Very well, sir. And this stone?" + +"Ah, yes, I shall keep the stone. Thank you. And, I say, +Peterson, just buy a goose on your way back and leave it here +with me, for we must have one to give to this gentleman in place +of the one which your family is now devouring." + +When the commissionaire had gone, Holmes took up the stone and +held it against the light. "It's a bonny thing," said he. "Just +see how it glints and sparkles. Of course it is a nucleus and +focus of crime. Every good stone is. They are the devil's pet +baits. In the larger and older jewels every facet may stand for a +bloody deed. This stone is not yet twenty years old. It was found +in the banks of the Amoy River in southern China and is remarkable +in having every characteristic of the carbuncle, save that it is +blue in shade instead of ruby red. In spite of its youth, it has +already a sinister history. There have been two murders, a +vitriol-throwing, a suicide, and several robberies brought about +for the sake of this forty-grain weight of crystallised charcoal. +Who would think that so pretty a toy would be a purveyor to the +gallows and the prison? I'll lock it up in my strong box now and +drop a line to the Countess to say that we have it." + +"Do you think that this man Horner is innocent?" + +"I cannot tell." + +"Well, then, do you imagine that this other one, Henry Baker, had +anything to do with the matter?" + +"It is, I think, much more likely that Henry Baker is an +absolutely innocent man, who had no idea that the bird which he +was carrying was of considerably more value than if it were made +of solid gold. That, however, I shall determine by a very simple +test if we have an answer to our advertisement." + +"And you can do nothing until then?" + +"Nothing." + +"In that case I shall continue my professional round. But I shall +come back in the evening at the hour you have mentioned, for I +should like to see the solution of so tangled a business." + +"Very glad to see you. I dine at seven. There is a woodcock, I +believe. By the way, in view of recent occurrences, perhaps I +ought to ask Mrs. Hudson to examine its crop." + +I had been delayed at a case, and it was a little after half-past +six when I found myself in Baker Street once more. As I +approached the house I saw a tall man in a Scotch bonnet with a +coat which was buttoned up to his chin waiting outside in the +bright semicircle which was thrown from the fanlight. Just as I +arrived the door was opened, and we were shown up together to +Holmes' room. + +"Mr. Henry Baker, I believe," said he, rising from his armchair +and greeting his visitor with the easy air of geniality which he +could so readily assume. "Pray take this chair by the fire, Mr. +Baker. It is a cold night, and I observe that your circulation is +more adapted for summer than for winter. Ah, Watson, you have +just come at the right time. Is that your hat, Mr. Baker?" + +"Yes, sir, that is undoubtedly my hat." + +He was a large man with rounded shoulders, a massive head, and a +broad, intelligent face, sloping down to a pointed beard of +grizzled brown. A touch of red in nose and cheeks, with a slight +tremor of his extended hand, recalled Holmes' surmise as to his +habits. His rusty black frock-coat was buttoned right up in +front, with the collar turned up, and his lank wrists protruded +from his sleeves without a sign of cuff or shirt. He spoke in a +slow staccato fashion, choosing his words with care, and gave the +impression generally of a man of learning and letters who had had +ill-usage at the hands of fortune. + +"We have retained these things for some days," said Holmes, +"because we expected to see an advertisement from you giving your +address. I am at a loss to know now why you did not advertise." + +Our visitor gave a rather shamefaced laugh. "Shillings have not +been so plentiful with me as they once were," he remarked. "I had +no doubt that the gang of roughs who assaulted me had carried off +both my hat and the bird. I did not care to spend more money in a +hopeless attempt at recovering them." + +"Very naturally. By the way, about the bird, we were compelled to +eat it." + +"To eat it!" Our visitor half rose from his chair in his +excitement. + +"Yes, it would have been of no use to anyone had we not done so. +But I presume that this other goose upon the sideboard, which is +about the same weight and perfectly fresh, will answer your +purpose equally well?" + +"Oh, certainly, certainly," answered Mr. Baker with a sigh of +relief. + +"Of course, we still have the feathers, legs, crop, and so on of +your own bird, so if you wish--" + +The man burst into a hearty laugh. "They might be useful to me as +relics of my adventure," said he, "but beyond that I can hardly +see what use the disjecta membra of my late acquaintance are +going to be to me. No, sir, I think that, with your permission, I +will confine my attentions to the excellent bird which I perceive +upon the sideboard." + +Sherlock Holmes glanced sharply across at me with a slight shrug +of his shoulders. + +"There is your hat, then, and there your bird," said he. "By the +way, would it bore you to tell me where you got the other one +from? I am somewhat of a fowl fancier, and I have seldom seen a +better grown goose." + +"Certainly, sir," said Baker, who had risen and tucked his newly +gained property under his arm. "There are a few of us who +frequent the Alpha Inn, near the Museum--we are to be found in +the Museum itself during the day, you understand. This year our +good host, Windigate by name, instituted a goose club, by which, +on consideration of some few pence every week, we were each to +receive a bird at Christmas. My pence were duly paid, and the +rest is familiar to you. I am much indebted to you, sir, for a +Scotch bonnet is fitted neither to my years nor my gravity." With +a comical pomposity of manner he bowed solemnly to both of us and +strode off upon his way. + +"So much for Mr. Henry Baker," said Holmes when he had closed the +door behind him. "It is quite certain that he knows nothing +whatever about the matter. Are you hungry, Watson?" + +"Not particularly." + +"Then I suggest that we turn our dinner into a supper and follow +up this clue while it is still hot." + +"By all means." + +It was a bitter night, so we drew on our ulsters and wrapped +cravats about our throats. Outside, the stars were shining coldly +in a cloudless sky, and the breath of the passers-by blew out +into smoke like so many pistol shots. Our footfalls rang out +crisply and loudly as we swung through the doctors' quarter, +Wimpole Street, Harley Street, and so through Wigmore Street into +Oxford Street. In a quarter of an hour we were in Bloomsbury at +the Alpha Inn, which is a small public-house at the corner of one +of the streets which runs down into Holborn. Holmes pushed open +the door of the private bar and ordered two glasses of beer from +the ruddy-faced, white-aproned landlord. + +"Your beer should be excellent if it is as good as your geese," +said he. + +"My geese!" The man seemed surprised. + +"Yes. I was speaking only half an hour ago to Mr. Henry Baker, +who was a member of your goose club." + +"Ah! yes, I see. But you see, sir, them's not our geese." + +"Indeed! Whose, then?" + +"Well, I got the two dozen from a salesman in Covent Garden." + +"Indeed? I know some of them. Which was it?" + +"Breckinridge is his name." + +"Ah! I don't know him. Well, here's your good health landlord, +and prosperity to your house. Good-night." + +"Now for Mr. Breckinridge," he continued, buttoning up his coat +as we came out into the frosty air. "Remember, Watson that though +we have so homely a thing as a goose at one end of this chain, we +have at the other a man who will certainly get seven years' penal +servitude unless we can establish his innocence. It is possible +that our inquiry may but confirm his guilt; but, in any case, we +have a line of investigation which has been missed by the police, +and which a singular chance has placed in our hands. Let us +follow it out to the bitter end. Faces to the south, then, and +quick march!" + +We passed across Holborn, down Endell Street, and so through a +zigzag of slums to Covent Garden Market. One of the largest +stalls bore the name of Breckinridge upon it, and the proprietor +a horsey-looking man, with a sharp face and trim side-whiskers was +helping a boy to put up the shutters. + +"Good-evening. It's a cold night," said Holmes. + +The salesman nodded and shot a questioning glance at my +companion. + +"Sold out of geese, I see," continued Holmes, pointing at the +bare slabs of marble. + +"Let you have five hundred to-morrow morning." + +"That's no good." + +"Well, there are some on the stall with the gas-flare." + +"Ah, but I was recommended to you." + +"Who by?" + +"The landlord of the Alpha." + +"Oh, yes; I sent him a couple of dozen." + +"Fine birds they were, too. Now where did you get them from?" + +To my surprise the question provoked a burst of anger from the +salesman. + +"Now, then, mister," said he, with his head cocked and his arms +akimbo, "what are you driving at? Let's have it straight, now." + +"It is straight enough. I should like to know who sold you the +geese which you supplied to the Alpha." + +"Well then, I shan't tell you. So now!" + +"Oh, it is a matter of no importance; but I don't know why you +should be so warm over such a trifle." + +"Warm! You'd be as warm, maybe, if you were as pestered as I am. +When I pay good money for a good article there should be an end +of the business; but it's 'Where are the geese?' and 'Who did you +sell the geese to?' and 'What will you take for the geese?' One +would think they were the only geese in the world, to hear the +fuss that is made over them." + +"Well, I have no connection with any other people who have been +making inquiries," said Holmes carelessly. "If you won't tell us +the bet is off, that is all. But I'm always ready to back my +opinion on a matter of fowls, and I have a fiver on it that the +bird I ate is country bred." + +"Well, then, you've lost your fiver, for it's town bred," snapped +the salesman. + +"It's nothing of the kind." + +"I say it is." + +"I don't believe it." + +"D'you think you know more about fowls than I, who have handled +them ever since I was a nipper? I tell you, all those birds that +went to the Alpha were town bred." + +"You'll never persuade me to believe that." + +"Will you bet, then?" + +"It's merely taking your money, for I know that I am right. But +I'll have a sovereign on with you, just to teach you not to be +obstinate." + +The salesman chuckled grimly. "Bring me the books, Bill," said +he. + +The small boy brought round a small thin volume and a great +greasy-backed one, laying them out together beneath the hanging +lamp. + +"Now then, Mr. Cocksure," said the salesman, "I thought that I +was out of geese, but before I finish you'll find that there is +still one left in my shop. You see this little book?" + +"Well?" + +"That's the list of the folk from whom I buy. D'you see? Well, +then, here on this page are the country folk, and the numbers +after their names are where their accounts are in the big ledger. +Now, then! You see this other page in red ink? Well, that is a +list of my town suppliers. Now, look at that third name. Just +read it out to me." + +"Mrs. Oakshott, 117, Brixton Road--249," read Holmes. + +"Quite so. Now turn that up in the ledger." + +Holmes turned to the page indicated. "Here you are, 'Mrs. +Oakshott, 117, Brixton Road, egg and poultry supplier.'" + +"Now, then, what's the last entry?" + +"'December 22nd. Twenty-four geese at 7s. 6d.'" + +"Quite so. There you are. And underneath?" + +"'Sold to Mr. Windigate of the Alpha, at 12s.'" + +"What have you to say now?" + +Sherlock Holmes looked deeply chagrined. He drew a sovereign from +his pocket and threw it down upon the slab, turning away with the +air of a man whose disgust is too deep for words. A few yards off +he stopped under a lamp-post and laughed in the hearty, noiseless +fashion which was peculiar to him. + +"When you see a man with whiskers of that cut and the 'Pink 'un' +protruding out of his pocket, you can always draw him by a bet," +said he. "I daresay that if I had put 100 pounds down in front of +him, that man would not have given me such complete information +as was drawn from him by the idea that he was doing me on a +wager. Well, Watson, we are, I fancy, nearing the end of our +quest, and the only point which remains to be determined is +whether we should go on to this Mrs. Oakshott to-night, or +whether we should reserve it for to-morrow. It is clear from what +that surly fellow said that there are others besides ourselves +who are anxious about the matter, and I should--" + +His remarks were suddenly cut short by a loud hubbub which broke +out from the stall which we had just left. Turning round we saw a +little rat-faced fellow standing in the centre of the circle of +yellow light which was thrown by the swinging lamp, while +Breckinridge, the salesman, framed in the door of his stall, was +shaking his fists fiercely at the cringing figure. + +"I've had enough of you and your geese," he shouted. "I wish you +were all at the devil together. If you come pestering me any more +with your silly talk I'll set the dog at you. You bring Mrs. +Oakshott here and I'll answer her, but what have you to do with +it? Did I buy the geese off you?" + +"No; but one of them was mine all the same," whined the little +man. + +"Well, then, ask Mrs. Oakshott for it." + +"She told me to ask you." + +"Well, you can ask the King of Proosia, for all I care. I've had +enough of it. Get out of this!" He rushed fiercely forward, and +the inquirer flitted away into the darkness. + +"Ha! this may save us a visit to Brixton Road," whispered Holmes. +"Come with me, and we will see what is to be made of this +fellow." Striding through the scattered knots of people who +lounged round the flaring stalls, my companion speedily overtook +the little man and touched him upon the shoulder. He sprang +round, and I could see in the gas-light that every vestige of +colour had been driven from his face. + +"Who are you, then? What do you want?" he asked in a quavering +voice. + +"You will excuse me," said Holmes blandly, "but I could not help +overhearing the questions which you put to the salesman just now. +I think that I could be of assistance to you." + +"You? Who are you? How could you know anything of the matter?" + +"My name is Sherlock Holmes. It is my business to know what other +people don't know." + +"But you can know nothing of this?" + +"Excuse me, I know everything of it. You are endeavouring to +trace some geese which were sold by Mrs. Oakshott, of Brixton +Road, to a salesman named Breckinridge, by him in turn to Mr. +Windigate, of the Alpha, and by him to his club, of which Mr. +Henry Baker is a member." + +"Oh, sir, you are the very man whom I have longed to meet," cried +the little fellow with outstretched hands and quivering fingers. +"I can hardly explain to you how interested I am in this matter." + +Sherlock Holmes hailed a four-wheeler which was passing. "In that +case we had better discuss it in a cosy room rather than in this +wind-swept market-place," said he. "But pray tell me, before we +go farther, who it is that I have the pleasure of assisting." + +The man hesitated for an instant. "My name is John Robinson," he +answered with a sidelong glance. + +"No, no; the real name," said Holmes sweetly. "It is always +awkward doing business with an alias." + +A flush sprang to the white cheeks of the stranger. "Well then," +said he, "my real name is James Ryder." + +"Precisely so. Head attendant at the Hotel Cosmopolitan. Pray +step into the cab, and I shall soon be able to tell you +everything which you would wish to know." + +The little man stood glancing from one to the other of us with +half-frightened, half-hopeful eyes, as one who is not sure +whether he is on the verge of a windfall or of a catastrophe. +Then he stepped into the cab, and in half an hour we were back in +the sitting-room at Baker Street. Nothing had been said during +our drive, but the high, thin breathing of our new companion, and +the claspings and unclaspings of his hands, spoke of the nervous +tension within him. + +"Here we are!" said Holmes cheerily as we filed into the room. +"The fire looks very seasonable in this weather. You look cold, +Mr. Ryder. Pray take the basket-chair. I will just put on my +slippers before we settle this little matter of yours. Now, then! +You want to know what became of those geese?" + +"Yes, sir." + +"Or rather, I fancy, of that goose. It was one bird, I imagine in +which you were interested--white, with a black bar across the +tail." + +Ryder quivered with emotion. "Oh, sir," he cried, "can you tell +me where it went to?" + +"It came here." + +"Here?" + +"Yes, and a most remarkable bird it proved. I don't wonder that +you should take an interest in it. It laid an egg after it was +dead--the bonniest, brightest little blue egg that ever was seen. +I have it here in my museum." + +Our visitor staggered to his feet and clutched the mantelpiece +with his right hand. Holmes unlocked his strong-box and held up +the blue carbuncle, which shone out like a star, with a cold, +brilliant, many-pointed radiance. Ryder stood glaring with a +drawn face, uncertain whether to claim or to disown it. + +"The game's up, Ryder," said Holmes quietly. "Hold up, man, or +you'll be into the fire! Give him an arm back into his chair, +Watson. He's not got blood enough to go in for felony with +impunity. Give him a dash of brandy. So! Now he looks a little +more human. What a shrimp it is, to be sure!" + +For a moment he had staggered and nearly fallen, but the brandy +brought a tinge of colour into his cheeks, and he sat staring +with frightened eyes at his accuser. + +"I have almost every link in my hands, and all the proofs which I +could possibly need, so there is little which you need tell me. +Still, that little may as well be cleared up to make the case +complete. You had heard, Ryder, of this blue stone of the +Countess of Morcar's?" + +"It was Catherine Cusack who told me of it," said he in a +crackling voice. + +"I see--her ladyship's waiting-maid. Well, the temptation of +sudden wealth so easily acquired was too much for you, as it has +been for better men before you; but you were not very scrupulous +in the means you used. It seems to me, Ryder, that there is the +making of a very pretty villain in you. You knew that this man +Horner, the plumber, had been concerned in some such matter +before, and that suspicion would rest the more readily upon him. +What did you do, then? You made some small job in my lady's +room--you and your confederate Cusack--and you managed that he +should be the man sent for. Then, when he had left, you rifled +the jewel-case, raised the alarm, and had this unfortunate man +arrested. You then--" + +Ryder threw himself down suddenly upon the rug and clutched at my +companion's knees. "For God's sake, have mercy!" he shrieked. +"Think of my father! Of my mother! It would break their hearts. I +never went wrong before! I never will again. I swear it. I'll +swear it on a Bible. Oh, don't bring it into court! For Christ's +sake, don't!" + +"Get back into your chair!" said Holmes sternly. "It is very well +to cringe and crawl now, but you thought little enough of this +poor Horner in the dock for a crime of which he knew nothing." + +"I will fly, Mr. Holmes. I will leave the country, sir. Then the +charge against him will break down." + +"Hum! We will talk about that. And now let us hear a true account +of the next act. How came the stone into the goose, and how came +the goose into the open market? Tell us the truth, for there lies +your only hope of safety." + +Ryder passed his tongue over his parched lips. "I will tell you +it just as it happened, sir," said he. "When Horner had been +arrested, it seemed to me that it would be best for me to get +away with the stone at once, for I did not know at what moment +the police might not take it into their heads to search me and my +room. There was no place about the hotel where it would be safe. +I went out, as if on some commission, and I made for my sister's +house. She had married a man named Oakshott, and lived in Brixton +Road, where she fattened fowls for the market. All the way there +every man I met seemed to me to be a policeman or a detective; +and, for all that it was a cold night, the sweat was pouring down +my face before I came to the Brixton Road. My sister asked me +what was the matter, and why I was so pale; but I told her that I +had been upset by the jewel robbery at the hotel. Then I went +into the back yard and smoked a pipe and wondered what it would +be best to do. + +"I had a friend once called Maudsley, who went to the bad, and +has just been serving his time in Pentonville. One day he had met +me, and fell into talk about the ways of thieves, and how they +could get rid of what they stole. I knew that he would be true to +me, for I knew one or two things about him; so I made up my mind +to go right on to Kilburn, where he lived, and take him into my +confidence. He would show me how to turn the stone into money. +But how to get to him in safety? I thought of the agonies I had +gone through in coming from the hotel. I might at any moment be +seized and searched, and there would be the stone in my waistcoat +pocket. I was leaning against the wall at the time and looking at +the geese which were waddling about round my feet, and suddenly +an idea came into my head which showed me how I could beat the +best detective that ever lived. + +"My sister had told me some weeks before that I might have the +pick of her geese for a Christmas present, and I knew that she +was always as good as her word. I would take my goose now, and in +it I would carry my stone to Kilburn. There was a little shed in +the yard, and behind this I drove one of the birds--a fine big +one, white, with a barred tail. I caught it, and prying its bill +open, I thrust the stone down its throat as far as my finger +could reach. The bird gave a gulp, and I felt the stone pass +along its gullet and down into its crop. But the creature flapped +and struggled, and out came my sister to know what was the +matter. As I turned to speak to her the brute broke loose and +fluttered off among the others. + +"'Whatever were you doing with that bird, Jem?' says she. + +"'Well,' said I, 'you said you'd give me one for Christmas, and I +was feeling which was the fattest.' + +"'Oh,' says she, 'we've set yours aside for you--Jem's bird, we +call it. It's the big white one over yonder. There's twenty-six +of them, which makes one for you, and one for us, and two dozen +for the market.' + +"'Thank you, Maggie,' says I; 'but if it is all the same to you, +I'd rather have that one I was handling just now.' + +"'The other is a good three pound heavier,' said she, 'and we +fattened it expressly for you.' + +"'Never mind. I'll have the other, and I'll take it now,' said I. + +"'Oh, just as you like,' said she, a little huffed. 'Which is it +you want, then?' + +"'That white one with the barred tail, right in the middle of the +flock.' + +"'Oh, very well. Kill it and take it with you.' + +"Well, I did what she said, Mr. Holmes, and I carried the bird +all the way to Kilburn. I told my pal what I had done, for he was +a man that it was easy to tell a thing like that to. He laughed +until he choked, and we got a knife and opened the goose. My +heart turned to water, for there was no sign of the stone, and I +knew that some terrible mistake had occurred. I left the bird, +rushed back to my sister's, and hurried into the back yard. There +was not a bird to be seen there. + +"'Where are they all, Maggie?' I cried. + +"'Gone to the dealer's, Jem.' + +"'Which dealer's?' + +"'Breckinridge, of Covent Garden.' + +"'But was there another with a barred tail?' I asked, 'the same +as the one I chose?' + +"'Yes, Jem; there were two barred-tailed ones, and I could never +tell them apart.' + +"Well, then, of course I saw it all, and I ran off as hard as my +feet would carry me to this man Breckinridge; but he had sold the +lot at once, and not one word would he tell me as to where they +had gone. You heard him yourselves to-night. Well, he has always +answered me like that. My sister thinks that I am going mad. +Sometimes I think that I am myself. And now--and now I am myself +a branded thief, without ever having touched the wealth for which +I sold my character. God help me! God help me!" He burst into +convulsive sobbing, with his face buried in his hands. + +There was a long silence, broken only by his heavy breathing and +by the measured tapping of Sherlock Holmes' finger-tips upon the +edge of the table. Then my friend rose and threw open the door. + +"Get out!" said he. + +"What, sir! Oh, Heaven bless you!" + +"No more words. Get out!" + +And no more words were needed. There was a rush, a clatter upon +the stairs, the bang of a door, and the crisp rattle of running +footfalls from the street. + +"After all, Watson," said Holmes, reaching up his hand for his +clay pipe, "I am not retained by the police to supply their +deficiencies. If Horner were in danger it would be another thing; +but this fellow will not appear against him, and the case must +collapse. I suppose that I am commuting a felony, but it is just +possible that I am saving a soul. This fellow will not go wrong +again; he is too terribly frightened. Send him to gaol now, and +you make him a gaol-bird for life. Besides, it is the season of +forgiveness. Chance has put in our way a most singular and +whimsical problem, and its solution is its own reward. If you +will have the goodness to touch the bell, Doctor, we will begin +another investigation, in which, also a bird will be the chief +feature." + + + +VIII. THE ADVENTURE OF THE SPECKLED BAND + +On glancing over my notes of the seventy odd cases in which I +have during the last eight years studied the methods of my friend +Sherlock Holmes, I find many tragic, some comic, a large number +merely strange, but none commonplace; for, working as he did +rather for the love of his art than for the acquirement of +wealth, he refused to associate himself with any investigation +which did not tend towards the unusual, and even the fantastic. +Of all these varied cases, however, I cannot recall any which +presented more singular features than that which was associated +with the well-known Surrey family of the Roylotts of Stoke Moran. +The events in question occurred in the early days of my +association with Holmes, when we were sharing rooms as bachelors +in Baker Street. It is possible that I might have placed them +upon record before, but a promise of secrecy was made at the +time, from which I have only been freed during the last month by +the untimely death of the lady to whom the pledge was given. It +is perhaps as well that the facts should now come to light, for I +have reasons to know that there are widespread rumours as to the +death of Dr. Grimesby Roylott which tend to make the matter even +more terrible than the truth. + +It was early in April in the year '83 that I woke one morning to +find Sherlock Holmes standing, fully dressed, by the side of my +bed. He was a late riser, as a rule, and as the clock on the +mantelpiece showed me that it was only a quarter-past seven, I +blinked up at him in some surprise, and perhaps just a little +resentment, for I was myself regular in my habits. + +"Very sorry to knock you up, Watson," said he, "but it's the +common lot this morning. Mrs. Hudson has been knocked up, she +retorted upon me, and I on you." + +"What is it, then--a fire?" + +"No; a client. It seems that a young lady has arrived in a +considerable state of excitement, who insists upon seeing me. She +is waiting now in the sitting-room. Now, when young ladies wander +about the metropolis at this hour of the morning, and knock +sleepy people up out of their beds, I presume that it is +something very pressing which they have to communicate. Should it +prove to be an interesting case, you would, I am sure, wish to +follow it from the outset. I thought, at any rate, that I should +call you and give you the chance." + +"My dear fellow, I would not miss it for anything." + +I had no keener pleasure than in following Holmes in his +professional investigations, and in admiring the rapid +deductions, as swift as intuitions, and yet always founded on a +logical basis with which he unravelled the problems which were +submitted to him. I rapidly threw on my clothes and was ready in +a few minutes to accompany my friend down to the sitting-room. A +lady dressed in black and heavily veiled, who had been sitting in +the window, rose as we entered. + +"Good-morning, madam," said Holmes cheerily. "My name is Sherlock +Holmes. This is my intimate friend and associate, Dr. Watson, +before whom you can speak as freely as before myself. Ha! I am +glad to see that Mrs. Hudson has had the good sense to light the +fire. Pray draw up to it, and I shall order you a cup of hot +coffee, for I observe that you are shivering." + +"It is not cold which makes me shiver," said the woman in a low +voice, changing her seat as requested. + +"What, then?" + +"It is fear, Mr. Holmes. It is terror." She raised her veil as +she spoke, and we could see that she was indeed in a pitiable +state of agitation, her face all drawn and grey, with restless +frightened eyes, like those of some hunted animal. Her features +and figure were those of a woman of thirty, but her hair was shot +with premature grey, and her expression was weary and haggard. +Sherlock Holmes ran her over with one of his quick, +all-comprehensive glances. + +"You must not fear," said he soothingly, bending forward and +patting her forearm. "We shall soon set matters right, I have no +doubt. You have come in by train this morning, I see." + +"You know me, then?" + +"No, but I observe the second half of a return ticket in the palm +of your left glove. You must have started early, and yet you had +a good drive in a dog-cart, along heavy roads, before you reached +the station." + +The lady gave a violent start and stared in bewilderment at my +companion. + +"There is no mystery, my dear madam," said he, smiling. "The left +arm of your jacket is spattered with mud in no less than seven +places. The marks are perfectly fresh. There is no vehicle save a +dog-cart which throws up mud in that way, and then only when you +sit on the left-hand side of the driver." + +"Whatever your reasons may be, you are perfectly correct," said +she. "I started from home before six, reached Leatherhead at +twenty past, and came in by the first train to Waterloo. Sir, I +can stand this strain no longer; I shall go mad if it continues. +I have no one to turn to--none, save only one, who cares for me, +and he, poor fellow, can be of little aid. I have heard of you, +Mr. Holmes; I have heard of you from Mrs. Farintosh, whom you +helped in the hour of her sore need. It was from her that I had +your address. Oh, sir, do you not think that you could help me, +too, and at least throw a little light through the dense darkness +which surrounds me? At present it is out of my power to reward +you for your services, but in a month or six weeks I shall be +married, with the control of my own income, and then at least you +shall not find me ungrateful." + +Holmes turned to his desk and, unlocking it, drew out a small +case-book, which he consulted. + +"Farintosh," said he. "Ah yes, I recall the case; it was +concerned with an opal tiara. I think it was before your time, +Watson. I can only say, madam, that I shall be happy to devote +the same care to your case as I did to that of your friend. As to +reward, my profession is its own reward; but you are at liberty +to defray whatever expenses I may be put to, at the time which +suits you best. And now I beg that you will lay before us +everything that may help us in forming an opinion upon the +matter." + +"Alas!" replied our visitor, "the very horror of my situation +lies in the fact that my fears are so vague, and my suspicions +depend so entirely upon small points, which might seem trivial to +another, that even he to whom of all others I have a right to +look for help and advice looks upon all that I tell him about it +as the fancies of a nervous woman. He does not say so, but I can +read it from his soothing answers and averted eyes. But I have +heard, Mr. Holmes, that you can see deeply into the manifold +wickedness of the human heart. You may advise me how to walk amid +the dangers which encompass me." + +"I am all attention, madam." + +"My name is Helen Stoner, and I am living with my stepfather, who +is the last survivor of one of the oldest Saxon families in +England, the Roylotts of Stoke Moran, on the western border of +Surrey." + +Holmes nodded his head. "The name is familiar to me," said he. + +"The family was at one time among the richest in England, and the +estates extended over the borders into Berkshire in the north, +and Hampshire in the west. In the last century, however, four +successive heirs were of a dissolute and wasteful disposition, +and the family ruin was eventually completed by a gambler in the +days of the Regency. Nothing was left save a few acres of ground, +and the two-hundred-year-old house, which is itself crushed under +a heavy mortgage. The last squire dragged out his existence +there, living the horrible life of an aristocratic pauper; but +his only son, my stepfather, seeing that he must adapt himself to +the new conditions, obtained an advance from a relative, which +enabled him to take a medical degree and went out to Calcutta, +where, by his professional skill and his force of character, he +established a large practice. In a fit of anger, however, caused +by some robberies which had been perpetrated in the house, he +beat his native butler to death and narrowly escaped a capital +sentence. As it was, he suffered a long term of imprisonment and +afterwards returned to England a morose and disappointed man. + +"When Dr. Roylott was in India he married my mother, Mrs. Stoner, +the young widow of Major-General Stoner, of the Bengal Artillery. +My sister Julia and I were twins, and we were only two years old +at the time of my mother's re-marriage. She had a considerable +sum of money--not less than 1000 pounds a year--and this she +bequeathed to Dr. Roylott entirely while we resided with him, +with a provision that a certain annual sum should be allowed to +each of us in the event of our marriage. Shortly after our return +to England my mother died--she was killed eight years ago in a +railway accident near Crewe. Dr. Roylott then abandoned his +attempts to establish himself in practice in London and took us +to live with him in the old ancestral house at Stoke Moran. The +money which my mother had left was enough for all our wants, and +there seemed to be no obstacle to our happiness. + +"But a terrible change came over our stepfather about this time. +Instead of making friends and exchanging visits with our +neighbours, who had at first been overjoyed to see a Roylott of +Stoke Moran back in the old family seat, he shut himself up in +his house and seldom came out save to indulge in ferocious +quarrels with whoever might cross his path. Violence of temper +approaching to mania has been hereditary in the men of the +family, and in my stepfather's case it had, I believe, been +intensified by his long residence in the tropics. A series of +disgraceful brawls took place, two of which ended in the +police-court, until at last he became the terror of the village, +and the folks would fly at his approach, for he is a man of +immense strength, and absolutely uncontrollable in his anger. + +"Last week he hurled the local blacksmith over a parapet into a +stream, and it was only by paying over all the money which I +could gather together that I was able to avert another public +exposure. He had no friends at all save the wandering gipsies, +and he would give these vagabonds leave to encamp upon the few +acres of bramble-covered land which represent the family estate, +and would accept in return the hospitality of their tents, +wandering away with them sometimes for weeks on end. He has a +passion also for Indian animals, which are sent over to him by a +correspondent, and he has at this moment a cheetah and a baboon, +which wander freely over his grounds and are feared by the +villagers almost as much as their master. + +"You can imagine from what I say that my poor sister Julia and I +had no great pleasure in our lives. No servant would stay with +us, and for a long time we did all the work of the house. She was +but thirty at the time of her death, and yet her hair had already +begun to whiten, even as mine has." + +"Your sister is dead, then?" + +"She died just two years ago, and it is of her death that I wish +to speak to you. You can understand that, living the life which I +have described, we were little likely to see anyone of our own +age and position. We had, however, an aunt, my mother's maiden +sister, Miss Honoria Westphail, who lives near Harrow, and we +were occasionally allowed to pay short visits at this lady's +house. Julia went there at Christmas two years ago, and met there +a half-pay major of marines, to whom she became engaged. My +stepfather learned of the engagement when my sister returned and +offered no objection to the marriage; but within a fortnight of +the day which had been fixed for the wedding, the terrible event +occurred which has deprived me of my only companion." + +Sherlock Holmes had been leaning back in his chair with his eyes +closed and his head sunk in a cushion, but he half opened his +lids now and glanced across at his visitor. + +"Pray be precise as to details," said he. + +"It is easy for me to be so, for every event of that dreadful +time is seared into my memory. The manor-house is, as I have +already said, very old, and only one wing is now inhabited. The +bedrooms in this wing are on the ground floor, the sitting-rooms +being in the central block of the buildings. Of these bedrooms +the first is Dr. Roylott's, the second my sister's, and the third +my own. There is no communication between them, but they all open +out into the same corridor. Do I make myself plain?" + +"Perfectly so." + +"The windows of the three rooms open out upon the lawn. That +fatal night Dr. Roylott had gone to his room early, though we +knew that he had not retired to rest, for my sister was troubled +by the smell of the strong Indian cigars which it was his custom +to smoke. She left her room, therefore, and came into mine, where +she sat for some time, chatting about her approaching wedding. At +eleven o'clock she rose to leave me, but she paused at the door +and looked back. + +"'Tell me, Helen,' said she, 'have you ever heard anyone whistle +in the dead of the night?' + +"'Never,' said I. + +"'I suppose that you could not possibly whistle, yourself, in +your sleep?' + +"'Certainly not. But why?' + +"'Because during the last few nights I have always, about three +in the morning, heard a low, clear whistle. I am a light sleeper, +and it has awakened me. I cannot tell where it came from--perhaps +from the next room, perhaps from the lawn. I thought that I would +just ask you whether you had heard it.' + +"'No, I have not. It must be those wretched gipsies in the +plantation.' + +"'Very likely. And yet if it were on the lawn, I wonder that you +did not hear it also.' + +"'Ah, but I sleep more heavily than you.' + +"'Well, it is of no great consequence, at any rate.' She smiled +back at me, closed my door, and a few moments later I heard her +key turn in the lock." + +"Indeed," said Holmes. "Was it your custom always to lock +yourselves in at night?" + +"Always." + +"And why?" + +"I think that I mentioned to you that the doctor kept a cheetah +and a baboon. We had no feeling of security unless our doors were +locked." + +"Quite so. Pray proceed with your statement." + +"I could not sleep that night. A vague feeling of impending +misfortune impressed me. My sister and I, you will recollect, +were twins, and you know how subtle are the links which bind two +souls which are so closely allied. It was a wild night. The wind +was howling outside, and the rain was beating and splashing +against the windows. Suddenly, amid all the hubbub of the gale, +there burst forth the wild scream of a terrified woman. I knew +that it was my sister's voice. I sprang from my bed, wrapped a +shawl round me, and rushed into the corridor. As I opened my door +I seemed to hear a low whistle, such as my sister described, and +a few moments later a clanging sound, as if a mass of metal had +fallen. As I ran down the passage, my sister's door was unlocked, +and revolved slowly upon its hinges. I stared at it +horror-stricken, not knowing what was about to issue from it. By +the light of the corridor-lamp I saw my sister appear at the +opening, her face blanched with terror, her hands groping for +help, her whole figure swaying to and fro like that of a +drunkard. I ran to her and threw my arms round her, but at that +moment her knees seemed to give way and she fell to the ground. +She writhed as one who is in terrible pain, and her limbs were +dreadfully convulsed. At first I thought that she had not +recognised me, but as I bent over her she suddenly shrieked out +in a voice which I shall never forget, 'Oh, my God! Helen! It was +the band! The speckled band!' There was something else which she +would fain have said, and she stabbed with her finger into the +air in the direction of the doctor's room, but a fresh convulsion +seized her and choked her words. I rushed out, calling loudly for +my stepfather, and I met him hastening from his room in his +dressing-gown. When he reached my sister's side she was +unconscious, and though he poured brandy down her throat and sent +for medical aid from the village, all efforts were in vain, for +she slowly sank and died without having recovered her +consciousness. Such was the dreadful end of my beloved sister." + +"One moment," said Holmes, "are you sure about this whistle and +metallic sound? Could you swear to it?" + +"That was what the county coroner asked me at the inquiry. It is +my strong impression that I heard it, and yet, among the crash of +the gale and the creaking of an old house, I may possibly have +been deceived." + +"Was your sister dressed?" + +"No, she was in her night-dress. In her right hand was found the +charred stump of a match, and in her left a match-box." + +"Showing that she had struck a light and looked about her when +the alarm took place. That is important. And what conclusions did +the coroner come to?" + +"He investigated the case with great care, for Dr. Roylott's +conduct had long been notorious in the county, but he was unable +to find any satisfactory cause of death. My evidence showed that +the door had been fastened upon the inner side, and the windows +were blocked by old-fashioned shutters with broad iron bars, +which were secured every night. The walls were carefully sounded, +and were shown to be quite solid all round, and the flooring was +also thoroughly examined, with the same result. The chimney is +wide, but is barred up by four large staples. It is certain, +therefore, that my sister was quite alone when she met her end. +Besides, there were no marks of any violence upon her." + +"How about poison?" + +"The doctors examined her for it, but without success." + +"What do you think that this unfortunate lady died of, then?" + +"It is my belief that she died of pure fear and nervous shock, +though what it was that frightened her I cannot imagine." + +"Were there gipsies in the plantation at the time?" + +"Yes, there are nearly always some there." + +"Ah, and what did you gather from this allusion to a band--a +speckled band?" + +"Sometimes I have thought that it was merely the wild talk of +delirium, sometimes that it may have referred to some band of +people, perhaps to these very gipsies in the plantation. I do not +know whether the spotted handkerchiefs which so many of them wear +over their heads might have suggested the strange adjective which +she used." + +Holmes shook his head like a man who is far from being satisfied. + +"These are very deep waters," said he; "pray go on with your +narrative." + +"Two years have passed since then, and my life has been until +lately lonelier than ever. A month ago, however, a dear friend, +whom I have known for many years, has done me the honour to ask +my hand in marriage. His name is Armitage--Percy Armitage--the +second son of Mr. Armitage, of Crane Water, near Reading. My +stepfather has offered no opposition to the match, and we are to +be married in the course of the spring. Two days ago some repairs +were started in the west wing of the building, and my bedroom +wall has been pierced, so that I have had to move into the +chamber in which my sister died, and to sleep in the very bed in +which she slept. Imagine, then, my thrill of terror when last +night, as I lay awake, thinking over her terrible fate, I +suddenly heard in the silence of the night the low whistle which +had been the herald of her own death. I sprang up and lit the +lamp, but nothing was to be seen in the room. I was too shaken to +go to bed again, however, so I dressed, and as soon as it was +daylight I slipped down, got a dog-cart at the Crown Inn, which +is opposite, and drove to Leatherhead, from whence I have come on +this morning with the one object of seeing you and asking your +advice." + +"You have done wisely," said my friend. "But have you told me +all?" + +"Yes, all." + +"Miss Roylott, you have not. You are screening your stepfather." + +"Why, what do you mean?" + +For answer Holmes pushed back the frill of black lace which +fringed the hand that lay upon our visitor's knee. Five little +livid spots, the marks of four fingers and a thumb, were printed +upon the white wrist. + +"You have been cruelly used," said Holmes. + +The lady coloured deeply and covered over her injured wrist. "He +is a hard man," she said, "and perhaps he hardly knows his own +strength." + +There was a long silence, during which Holmes leaned his chin +upon his hands and stared into the crackling fire. + +"This is a very deep business," he said at last. "There are a +thousand details which I should desire to know before I decide +upon our course of action. Yet we have not a moment to lose. If +we were to come to Stoke Moran to-day, would it be possible for +us to see over these rooms without the knowledge of your +stepfather?" + +"As it happens, he spoke of coming into town to-day upon some +most important business. It is probable that he will be away all +day, and that there would be nothing to disturb you. We have a +housekeeper now, but she is old and foolish, and I could easily +get her out of the way." + +"Excellent. You are not averse to this trip, Watson?" + +"By no means." + +"Then we shall both come. What are you going to do yourself?" + +"I have one or two things which I would wish to do now that I am +in town. But I shall return by the twelve o'clock train, so as to +be there in time for your coming." + +"And you may expect us early in the afternoon. I have myself some +small business matters to attend to. Will you not wait and +breakfast?" + +"No, I must go. My heart is lightened already since I have +confided my trouble to you. I shall look forward to seeing you +again this afternoon." She dropped her thick black veil over her +face and glided from the room. + +"And what do you think of it all, Watson?" asked Sherlock Holmes, +leaning back in his chair. + +"It seems to me to be a most dark and sinister business." + +"Dark enough and sinister enough." + +"Yet if the lady is correct in saying that the flooring and walls +are sound, and that the door, window, and chimney are impassable, +then her sister must have been undoubtedly alone when she met her +mysterious end." + +"What becomes, then, of these nocturnal whistles, and what of the +very peculiar words of the dying woman?" + +"I cannot think." + +"When you combine the ideas of whistles at night, the presence of +a band of gipsies who are on intimate terms with this old doctor, +the fact that we have every reason to believe that the doctor has +an interest in preventing his stepdaughter's marriage, the dying +allusion to a band, and, finally, the fact that Miss Helen Stoner +heard a metallic clang, which might have been caused by one of +those metal bars that secured the shutters falling back into its +place, I think that there is good ground to think that the +mystery may be cleared along those lines." + +"But what, then, did the gipsies do?" + +"I cannot imagine." + +"I see many objections to any such theory." + +"And so do I. It is precisely for that reason that we are going +to Stoke Moran this day. I want to see whether the objections are +fatal, or if they may be explained away. But what in the name of +the devil!" + +The ejaculation had been drawn from my companion by the fact that +our door had been suddenly dashed open, and that a huge man had +framed himself in the aperture. His costume was a peculiar +mixture of the professional and of the agricultural, having a +black top-hat, a long frock-coat, and a pair of high gaiters, +with a hunting-crop swinging in his hand. So tall was he that his +hat actually brushed the cross bar of the doorway, and his +breadth seemed to span it across from side to side. A large face, +seared with a thousand wrinkles, burned yellow with the sun, and +marked with every evil passion, was turned from one to the other +of us, while his deep-set, bile-shot eyes, and his high, thin, +fleshless nose, gave him somewhat the resemblance to a fierce old +bird of prey. + +"Which of you is Holmes?" asked this apparition. + +"My name, sir; but you have the advantage of me," said my +companion quietly. + +"I am Dr. Grimesby Roylott, of Stoke Moran." + +"Indeed, Doctor," said Holmes blandly. "Pray take a seat." + +"I will do nothing of the kind. My stepdaughter has been here. I +have traced her. What has she been saying to you?" + +"It is a little cold for the time of the year," said Holmes. + +"What has she been saying to you?" screamed the old man +furiously. + +"But I have heard that the crocuses promise well," continued my +companion imperturbably. + +"Ha! You put me off, do you?" said our new visitor, taking a step +forward and shaking his hunting-crop. "I know you, you scoundrel! +I have heard of you before. You are Holmes, the meddler." + +My friend smiled. + +"Holmes, the busybody!" + +His smile broadened. + +"Holmes, the Scotland Yard Jack-in-office!" + +Holmes chuckled heartily. "Your conversation is most +entertaining," said he. "When you go out close the door, for +there is a decided draught." + +"I will go when I have said my say. Don't you dare to meddle with +my affairs. I know that Miss Stoner has been here. I traced her! +I am a dangerous man to fall foul of! See here." He stepped +swiftly forward, seized the poker, and bent it into a curve with +his huge brown hands. + +"See that you keep yourself out of my grip," he snarled, and +hurling the twisted poker into the fireplace he strode out of the +room. + +"He seems a very amiable person," said Holmes, laughing. "I am +not quite so bulky, but if he had remained I might have shown him +that my grip was not much more feeble than his own." As he spoke +he picked up the steel poker and, with a sudden effort, +straightened it out again. + +"Fancy his having the insolence to confound me with the official +detective force! This incident gives zest to our investigation, +however, and I only trust that our little friend will not suffer +from her imprudence in allowing this brute to trace her. And now, +Watson, we shall order breakfast, and afterwards I shall walk +down to Doctors' Commons, where I hope to get some data which may +help us in this matter." + + +It was nearly one o'clock when Sherlock Holmes returned from his +excursion. He held in his hand a sheet of blue paper, scrawled +over with notes and figures. + +"I have seen the will of the deceased wife," said he. "To +determine its exact meaning I have been obliged to work out the +present prices of the investments with which it is concerned. The +total income, which at the time of the wife's death was little +short of 1100 pounds, is now, through the fall in agricultural +prices, not more than 750 pounds. Each daughter can claim an +income of 250 pounds, in case of marriage. It is evident, +therefore, that if both girls had married, this beauty would have +had a mere pittance, while even one of them would cripple him to +a very serious extent. My morning's work has not been wasted, +since it has proved that he has the very strongest motives for +standing in the way of anything of the sort. And now, Watson, +this is too serious for dawdling, especially as the old man is +aware that we are interesting ourselves in his affairs; so if you +are ready, we shall call a cab and drive to Waterloo. I should be +very much obliged if you would slip your revolver into your +pocket. An Eley's No. 2 is an excellent argument with gentlemen +who can twist steel pokers into knots. That and a tooth-brush +are, I think, all that we need." + +At Waterloo we were fortunate in catching a train for +Leatherhead, where we hired a trap at the station inn and drove +for four or five miles through the lovely Surrey lanes. It was a +perfect day, with a bright sun and a few fleecy clouds in the +heavens. The trees and wayside hedges were just throwing out +their first green shoots, and the air was full of the pleasant +smell of the moist earth. To me at least there was a strange +contrast between the sweet promise of the spring and this +sinister quest upon which we were engaged. My companion sat in +the front of the trap, his arms folded, his hat pulled down over +his eyes, and his chin sunk upon his breast, buried in the +deepest thought. Suddenly, however, he started, tapped me on the +shoulder, and pointed over the meadows. + +"Look there!" said he. + +A heavily timbered park stretched up in a gentle slope, +thickening into a grove at the highest point. From amid the +branches there jutted out the grey gables and high roof-tree of a +very old mansion. + +"Stoke Moran?" said he. + +"Yes, sir, that be the house of Dr. Grimesby Roylott," remarked +the driver. + +"There is some building going on there," said Holmes; "that is +where we are going." + +"There's the village," said the driver, pointing to a cluster of +roofs some distance to the left; "but if you want to get to the +house, you'll find it shorter to get over this stile, and so by +the foot-path over the fields. There it is, where the lady is +walking." + +"And the lady, I fancy, is Miss Stoner," observed Holmes, shading +his eyes. "Yes, I think we had better do as you suggest." + +We got off, paid our fare, and the trap rattled back on its way +to Leatherhead. + +"I thought it as well," said Holmes as we climbed the stile, +"that this fellow should think we had come here as architects, or +on some definite business. It may stop his gossip. +Good-afternoon, Miss Stoner. You see that we have been as good as +our word." + +Our client of the morning had hurried forward to meet us with a +face which spoke her joy. "I have been waiting so eagerly for +you," she cried, shaking hands with us warmly. "All has turned +out splendidly. Dr. Roylott has gone to town, and it is unlikely +that he will be back before evening." + +"We have had the pleasure of making the doctor's acquaintance," +said Holmes, and in a few words he sketched out what had +occurred. Miss Stoner turned white to the lips as she listened. + +"Good heavens!" she cried, "he has followed me, then." + +"So it appears." + +"He is so cunning that I never know when I am safe from him. What +will he say when he returns?" + +"He must guard himself, for he may find that there is someone +more cunning than himself upon his track. You must lock yourself +up from him to-night. If he is violent, we shall take you away to +your aunt's at Harrow. Now, we must make the best use of our +time, so kindly take us at once to the rooms which we are to +examine." + +The building was of grey, lichen-blotched stone, with a high +central portion and two curving wings, like the claws of a crab, +thrown out on each side. In one of these wings the windows were +broken and blocked with wooden boards, while the roof was partly +caved in, a picture of ruin. The central portion was in little +better repair, but the right-hand block was comparatively modern, +and the blinds in the windows, with the blue smoke curling up +from the chimneys, showed that this was where the family resided. +Some scaffolding had been erected against the end wall, and the +stone-work had been broken into, but there were no signs of any +workmen at the moment of our visit. Holmes walked slowly up and +down the ill-trimmed lawn and examined with deep attention the +outsides of the windows. + +"This, I take it, belongs to the room in which you used to sleep, +the centre one to your sister's, and the one next to the main +building to Dr. Roylott's chamber?" + +"Exactly so. But I am now sleeping in the middle one." + +"Pending the alterations, as I understand. By the way, there does +not seem to be any very pressing need for repairs at that end +wall." + +"There were none. I believe that it was an excuse to move me from +my room." + +"Ah! that is suggestive. Now, on the other side of this narrow +wing runs the corridor from which these three rooms open. There +are windows in it, of course?" + +"Yes, but very small ones. Too narrow for anyone to pass +through." + +"As you both locked your doors at night, your rooms were +unapproachable from that side. Now, would you have the kindness +to go into your room and bar your shutters?" + +Miss Stoner did so, and Holmes, after a careful examination +through the open window, endeavoured in every way to force the +shutter open, but without success. There was no slit through +which a knife could be passed to raise the bar. Then with his +lens he tested the hinges, but they were of solid iron, built +firmly into the massive masonry. "Hum!" said he, scratching his +chin in some perplexity, "my theory certainly presents some +difficulties. No one could pass these shutters if they were +bolted. Well, we shall see if the inside throws any light upon +the matter." + +A small side door led into the whitewashed corridor from which +the three bedrooms opened. Holmes refused to examine the third +chamber, so we passed at once to the second, that in which Miss +Stoner was now sleeping, and in which her sister had met with her +fate. It was a homely little room, with a low ceiling and a +gaping fireplace, after the fashion of old country-houses. A +brown chest of drawers stood in one corner, a narrow +white-counterpaned bed in another, and a dressing-table on the +left-hand side of the window. These articles, with two small +wicker-work chairs, made up all the furniture in the room save +for a square of Wilton carpet in the centre. The boards round and +the panelling of the walls were of brown, worm-eaten oak, so old +and discoloured that it may have dated from the original building +of the house. Holmes drew one of the chairs into a corner and sat +silent, while his eyes travelled round and round and up and down, +taking in every detail of the apartment. + +"Where does that bell communicate with?" he asked at last +pointing to a thick bell-rope which hung down beside the bed, the +tassel actually lying upon the pillow. + +"It goes to the housekeeper's room." + +"It looks newer than the other things?" + +"Yes, it was only put there a couple of years ago." + +"Your sister asked for it, I suppose?" + +"No, I never heard of her using it. We used always to get what we +wanted for ourselves." + +"Indeed, it seemed unnecessary to put so nice a bell-pull there. +You will excuse me for a few minutes while I satisfy myself as to +this floor." He threw himself down upon his face with his lens in +his hand and crawled swiftly backward and forward, examining +minutely the cracks between the boards. Then he did the same with +the wood-work with which the chamber was panelled. Finally he +walked over to the bed and spent some time in staring at it and +in running his eye up and down the wall. Finally he took the +bell-rope in his hand and gave it a brisk tug. + +"Why, it's a dummy," said he. + +"Won't it ring?" + +"No, it is not even attached to a wire. This is very interesting. +You can see now that it is fastened to a hook just above where +the little opening for the ventilator is." + +"How very absurd! I never noticed that before." + +"Very strange!" muttered Holmes, pulling at the rope. "There are +one or two very singular points about this room. For example, +what a fool a builder must be to open a ventilator into another +room, when, with the same trouble, he might have communicated +with the outside air!" + +"That is also quite modern," said the lady. + +"Done about the same time as the bell-rope?" remarked Holmes. + +"Yes, there were several little changes carried out about that +time." + +"They seem to have been of a most interesting character--dummy +bell-ropes, and ventilators which do not ventilate. With your +permission, Miss Stoner, we shall now carry our researches into +the inner apartment." + +Dr. Grimesby Roylott's chamber was larger than that of his +step-daughter, but was as plainly furnished. A camp-bed, a small +wooden shelf full of books, mostly of a technical character, an +armchair beside the bed, a plain wooden chair against the wall, a +round table, and a large iron safe were the principal things +which met the eye. Holmes walked slowly round and examined each +and all of them with the keenest interest. + +"What's in here?" he asked, tapping the safe. + +"My stepfather's business papers." + +"Oh! you have seen inside, then?" + +"Only once, some years ago. I remember that it was full of +papers." + +"There isn't a cat in it, for example?" + +"No. What a strange idea!" + +"Well, look at this!" He took up a small saucer of milk which +stood on the top of it. + +"No; we don't keep a cat. But there is a cheetah and a baboon." + +"Ah, yes, of course! Well, a cheetah is just a big cat, and yet a +saucer of milk does not go very far in satisfying its wants, I +daresay. There is one point which I should wish to determine." He +squatted down in front of the wooden chair and examined the seat +of it with the greatest attention. + +"Thank you. That is quite settled," said he, rising and putting +his lens in his pocket. "Hullo! Here is something interesting!" + +The object which had caught his eye was a small dog lash hung on +one corner of the bed. The lash, however, was curled upon itself +and tied so as to make a loop of whipcord. + +"What do you make of that, Watson?" + +"It's a common enough lash. But I don't know why it should be +tied." + +"That is not quite so common, is it? Ah, me! it's a wicked world, +and when a clever man turns his brains to crime it is the worst +of all. I think that I have seen enough now, Miss Stoner, and +with your permission we shall walk out upon the lawn." + +I had never seen my friend's face so grim or his brow so dark as +it was when we turned from the scene of this investigation. We +had walked several times up and down the lawn, neither Miss +Stoner nor myself liking to break in upon his thoughts before he +roused himself from his reverie. + +"It is very essential, Miss Stoner," said he, "that you should +absolutely follow my advice in every respect." + +"I shall most certainly do so." + +"The matter is too serious for any hesitation. Your life may +depend upon your compliance." + +"I assure you that I am in your hands." + +"In the first place, both my friend and I must spend the night in +your room." + +Both Miss Stoner and I gazed at him in astonishment. + +"Yes, it must be so. Let me explain. I believe that that is the +village inn over there?" + +"Yes, that is the Crown." + +"Very good. Your windows would be visible from there?" + +"Certainly." + +"You must confine yourself to your room, on pretence of a +headache, when your stepfather comes back. Then when you hear him +retire for the night, you must open the shutters of your window, +undo the hasp, put your lamp there as a signal to us, and then +withdraw quietly with everything which you are likely to want +into the room which you used to occupy. I have no doubt that, in +spite of the repairs, you could manage there for one night." + +"Oh, yes, easily." + +"The rest you will leave in our hands." + +"But what will you do?" + +"We shall spend the night in your room, and we shall investigate +the cause of this noise which has disturbed you." + +"I believe, Mr. Holmes, that you have already made up your mind," +said Miss Stoner, laying her hand upon my companion's sleeve. + +"Perhaps I have." + +"Then, for pity's sake, tell me what was the cause of my sister's +death." + +"I should prefer to have clearer proofs before I speak." + +"You can at least tell me whether my own thought is correct, and +if she died from some sudden fright." + +"No, I do not think so. I think that there was probably some more +tangible cause. And now, Miss Stoner, we must leave you for if +Dr. Roylott returned and saw us our journey would be in vain. +Good-bye, and be brave, for if you will do what I have told you, +you may rest assured that we shall soon drive away the dangers +that threaten you." + +Sherlock Holmes and I had no difficulty in engaging a bedroom and +sitting-room at the Crown Inn. They were on the upper floor, and +from our window we could command a view of the avenue gate, and +of the inhabited wing of Stoke Moran Manor House. At dusk we saw +Dr. Grimesby Roylott drive past, his huge form looming up beside +the little figure of the lad who drove him. The boy had some +slight difficulty in undoing the heavy iron gates, and we heard +the hoarse roar of the doctor's voice and saw the fury with which +he shook his clinched fists at him. The trap drove on, and a few +minutes later we saw a sudden light spring up among the trees as +the lamp was lit in one of the sitting-rooms. + +"Do you know, Watson," said Holmes as we sat together in the +gathering darkness, "I have really some scruples as to taking you +to-night. There is a distinct element of danger." + +"Can I be of assistance?" + +"Your presence might be invaluable." + +"Then I shall certainly come." + +"It is very kind of you." + +"You speak of danger. You have evidently seen more in these rooms +than was visible to me." + +"No, but I fancy that I may have deduced a little more. I imagine +that you saw all that I did." + +"I saw nothing remarkable save the bell-rope, and what purpose +that could answer I confess is more than I can imagine." + +"You saw the ventilator, too?" + +"Yes, but I do not think that it is such a very unusual thing to +have a small opening between two rooms. It was so small that a +rat could hardly pass through." + +"I knew that we should find a ventilator before ever we came to +Stoke Moran." + +"My dear Holmes!" + +"Oh, yes, I did. You remember in her statement she said that her +sister could smell Dr. Roylott's cigar. Now, of course that +suggested at once that there must be a communication between the +two rooms. It could only be a small one, or it would have been +remarked upon at the coroner's inquiry. I deduced a ventilator." + +"But what harm can there be in that?" + +"Well, there is at least a curious coincidence of dates. A +ventilator is made, a cord is hung, and a lady who sleeps in the +bed dies. Does not that strike you?" + +"I cannot as yet see any connection." + +"Did you observe anything very peculiar about that bed?" + +"No." + +"It was clamped to the floor. Did you ever see a bed fastened +like that before?" + +"I cannot say that I have." + +"The lady could not move her bed. It must always be in the same +relative position to the ventilator and to the rope--or so we may +call it, since it was clearly never meant for a bell-pull." + +"Holmes," I cried, "I seem to see dimly what you are hinting at. +We are only just in time to prevent some subtle and horrible +crime." + +"Subtle enough and horrible enough. When a doctor does go wrong +he is the first of criminals. He has nerve and he has knowledge. +Palmer and Pritchard were among the heads of their profession. +This man strikes even deeper, but I think, Watson, that we shall +be able to strike deeper still. But we shall have horrors enough +before the night is over; for goodness' sake let us have a quiet +pipe and turn our minds for a few hours to something more +cheerful." + + +About nine o'clock the light among the trees was extinguished, +and all was dark in the direction of the Manor House. Two hours +passed slowly away, and then, suddenly, just at the stroke of +eleven, a single bright light shone out right in front of us. + +"That is our signal," said Holmes, springing to his feet; "it +comes from the middle window." + +As we passed out he exchanged a few words with the landlord, +explaining that we were going on a late visit to an acquaintance, +and that it was possible that we might spend the night there. A +moment later we were out on the dark road, a chill wind blowing +in our faces, and one yellow light twinkling in front of us +through the gloom to guide us on our sombre errand. + +There was little difficulty in entering the grounds, for +unrepaired breaches gaped in the old park wall. Making our way +among the trees, we reached the lawn, crossed it, and were about +to enter through the window when out from a clump of laurel +bushes there darted what seemed to be a hideous and distorted +child, who threw itself upon the grass with writhing limbs and +then ran swiftly across the lawn into the darkness. + +"My God!" I whispered; "did you see it?" + +Holmes was for the moment as startled as I. His hand closed like +a vice upon my wrist in his agitation. Then he broke into a low +laugh and put his lips to my ear. + +"It is a nice household," he murmured. "That is the baboon." + +I had forgotten the strange pets which the doctor affected. There +was a cheetah, too; perhaps we might find it upon our shoulders +at any moment. I confess that I felt easier in my mind when, +after following Holmes' example and slipping off my shoes, I +found myself inside the bedroom. My companion noiselessly closed +the shutters, moved the lamp onto the table, and cast his eyes +round the room. All was as we had seen it in the daytime. Then +creeping up to me and making a trumpet of his hand, he whispered +into my ear again so gently that it was all that I could do to +distinguish the words: + +"The least sound would be fatal to our plans." + +I nodded to show that I had heard. + +"We must sit without light. He would see it through the +ventilator." + +I nodded again. + +"Do not go asleep; your very life may depend upon it. Have your +pistol ready in case we should need it. I will sit on the side of +the bed, and you in that chair." + +I took out my revolver and laid it on the corner of the table. + +Holmes had brought up a long thin cane, and this he placed upon +the bed beside him. By it he laid the box of matches and the +stump of a candle. Then he turned down the lamp, and we were left +in darkness. + +How shall I ever forget that dreadful vigil? I could not hear a +sound, not even the drawing of a breath, and yet I knew that my +companion sat open-eyed, within a few feet of me, in the same +state of nervous tension in which I was myself. The shutters cut +off the least ray of light, and we waited in absolute darkness. + +From outside came the occasional cry of a night-bird, and once at +our very window a long drawn catlike whine, which told us that +the cheetah was indeed at liberty. Far away we could hear the +deep tones of the parish clock, which boomed out every quarter of +an hour. How long they seemed, those quarters! Twelve struck, and +one and two and three, and still we sat waiting silently for +whatever might befall. + +Suddenly there was the momentary gleam of a light up in the +direction of the ventilator, which vanished immediately, but was +succeeded by a strong smell of burning oil and heated metal. +Someone in the next room had lit a dark-lantern. I heard a gentle +sound of movement, and then all was silent once more, though the +smell grew stronger. For half an hour I sat with straining ears. +Then suddenly another sound became audible--a very gentle, +soothing sound, like that of a small jet of steam escaping +continually from a kettle. The instant that we heard it, Holmes +sprang from the bed, struck a match, and lashed furiously with +his cane at the bell-pull. + +"You see it, Watson?" he yelled. "You see it?" + +But I saw nothing. At the moment when Holmes struck the light I +heard a low, clear whistle, but the sudden glare flashing into my +weary eyes made it impossible for me to tell what it was at which +my friend lashed so savagely. I could, however, see that his face +was deadly pale and filled with horror and loathing. He had +ceased to strike and was gazing up at the ventilator when +suddenly there broke from the silence of the night the most +horrible cry to which I have ever listened. It swelled up louder +and louder, a hoarse yell of pain and fear and anger all mingled +in the one dreadful shriek. They say that away down in the +village, and even in the distant parsonage, that cry raised the +sleepers from their beds. It struck cold to our hearts, and I +stood gazing at Holmes, and he at me, until the last echoes of it +had died away into the silence from which it rose. + +"What can it mean?" I gasped. + +"It means that it is all over," Holmes answered. "And perhaps, +after all, it is for the best. Take your pistol, and we will +enter Dr. Roylott's room." + +With a grave face he lit the lamp and led the way down the +corridor. Twice he struck at the chamber door without any reply +from within. Then he turned the handle and entered, I at his +heels, with the cocked pistol in my hand. + +It was a singular sight which met our eyes. On the table stood a +dark-lantern with the shutter half open, throwing a brilliant +beam of light upon the iron safe, the door of which was ajar. +Beside this table, on the wooden chair, sat Dr. Grimesby Roylott +clad in a long grey dressing-gown, his bare ankles protruding +beneath, and his feet thrust into red heelless Turkish slippers. +Across his lap lay the short stock with the long lash which we +had noticed during the day. His chin was cocked upward and his +eyes were fixed in a dreadful, rigid stare at the corner of the +ceiling. Round his brow he had a peculiar yellow band, with +brownish speckles, which seemed to be bound tightly round his +head. As we entered he made neither sound nor motion. + +"The band! the speckled band!" whispered Holmes. + +I took a step forward. In an instant his strange headgear began +to move, and there reared itself from among his hair the squat +diamond-shaped head and puffed neck of a loathsome serpent. + +"It is a swamp adder!" cried Holmes; "the deadliest snake in +India. He has died within ten seconds of being bitten. Violence +does, in truth, recoil upon the violent, and the schemer falls +into the pit which he digs for another. Let us thrust this +creature back into its den, and we can then remove Miss Stoner to +some place of shelter and let the county police know what has +happened." + +As he spoke he drew the dog-whip swiftly from the dead man's lap, +and throwing the noose round the reptile's neck he drew it from +its horrid perch and, carrying it at arm's length, threw it into +the iron safe, which he closed upon it. + +Such are the true facts of the death of Dr. Grimesby Roylott, of +Stoke Moran. It is not necessary that I should prolong a +narrative which has already run to too great a length by telling +how we broke the sad news to the terrified girl, how we conveyed +her by the morning train to the care of her good aunt at Harrow, +of how the slow process of official inquiry came to the +conclusion that the doctor met his fate while indiscreetly +playing with a dangerous pet. The little which I had yet to learn +of the case was told me by Sherlock Holmes as we travelled back +next day. + +"I had," said he, "come to an entirely erroneous conclusion which +shows, my dear Watson, how dangerous it always is to reason from +insufficient data. The presence of the gipsies, and the use of +the word 'band,' which was used by the poor girl, no doubt, to +explain the appearance which she had caught a hurried glimpse of +by the light of her match, were sufficient to put me upon an +entirely wrong scent. I can only claim the merit that I instantly +reconsidered my position when, however, it became clear to me +that whatever danger threatened an occupant of the room could not +come either from the window or the door. My attention was +speedily drawn, as I have already remarked to you, to this +ventilator, and to the bell-rope which hung down to the bed. The +discovery that this was a dummy, and that the bed was clamped to +the floor, instantly gave rise to the suspicion that the rope was +there as a bridge for something passing through the hole and +coming to the bed. The idea of a snake instantly occurred to me, +and when I coupled it with my knowledge that the doctor was +furnished with a supply of creatures from India, I felt that I +was probably on the right track. The idea of using a form of +poison which could not possibly be discovered by any chemical +test was just such a one as would occur to a clever and ruthless +man who had had an Eastern training. The rapidity with which such +a poison would take effect would also, from his point of view, be +an advantage. It would be a sharp-eyed coroner, indeed, who could +distinguish the two little dark punctures which would show where +the poison fangs had done their work. Then I thought of the +whistle. Of course he must recall the snake before the morning +light revealed it to the victim. He had trained it, probably by +the use of the milk which we saw, to return to him when summoned. +He would put it through this ventilator at the hour that he +thought best, with the certainty that it would crawl down the +rope and land on the bed. It might or might not bite the +occupant, perhaps she might escape every night for a week, but +sooner or later she must fall a victim. + +"I had come to these conclusions before ever I had entered his +room. An inspection of his chair showed me that he had been in +the habit of standing on it, which of course would be necessary +in order that he should reach the ventilator. The sight of the +safe, the saucer of milk, and the loop of whipcord were enough to +finally dispel any doubts which may have remained. The metallic +clang heard by Miss Stoner was obviously caused by her stepfather +hastily closing the door of his safe upon its terrible occupant. +Having once made up my mind, you know the steps which I took in +order to put the matter to the proof. I heard the creature hiss +as I have no doubt that you did also, and I instantly lit the +light and attacked it." + +"With the result of driving it through the ventilator." + +"And also with the result of causing it to turn upon its master +at the other side. Some of the blows of my cane came home and +roused its snakish temper, so that it flew upon the first person +it saw. In this way I am no doubt indirectly responsible for Dr. +Grimesby Roylott's death, and I cannot say that it is likely to +weigh very heavily upon my conscience." + + + +IX. THE ADVENTURE OF THE ENGINEER'S THUMB + +Of all the problems which have been submitted to my friend, Mr. +Sherlock Holmes, for solution during the years of our intimacy, +there were only two which I was the means of introducing to his +notice--that of Mr. Hatherley's thumb, and that of Colonel +Warburton's madness. Of these the latter may have afforded a +finer field for an acute and original observer, but the other was +so strange in its inception and so dramatic in its details that +it may be the more worthy of being placed upon record, even if it +gave my friend fewer openings for those deductive methods of +reasoning by which he achieved such remarkable results. The story +has, I believe, been told more than once in the newspapers, but, +like all such narratives, its effect is much less striking when +set forth en bloc in a single half-column of print than when the +facts slowly evolve before your own eyes, and the mystery clears +gradually away as each new discovery furnishes a step which leads +on to the complete truth. At the time the circumstances made a +deep impression upon me, and the lapse of two years has hardly +served to weaken the effect. + +It was in the summer of '89, not long after my marriage, that the +events occurred which I am now about to summarise. I had returned +to civil practice and had finally abandoned Holmes in his Baker +Street rooms, although I continually visited him and occasionally +even persuaded him to forgo his Bohemian habits so far as to come +and visit us. My practice had steadily increased, and as I +happened to live at no very great distance from Paddington +Station, I got a few patients from among the officials. One of +these, whom I had cured of a painful and lingering disease, was +never weary of advertising my virtues and of endeavouring to send +me on every sufferer over whom he might have any influence. + +One morning, at a little before seven o'clock, I was awakened by +the maid tapping at the door to announce that two men had come +from Paddington and were waiting in the consulting-room. I +dressed hurriedly, for I knew by experience that railway cases +were seldom trivial, and hastened downstairs. As I descended, my +old ally, the guard, came out of the room and closed the door +tightly behind him. + +"I've got him here," he whispered, jerking his thumb over his +shoulder; "he's all right." + +"What is it, then?" I asked, for his manner suggested that it was +some strange creature which he had caged up in my room. + +"It's a new patient," he whispered. "I thought I'd bring him +round myself; then he couldn't slip away. There he is, all safe +and sound. I must go now, Doctor; I have my dooties, just the +same as you." And off he went, this trusty tout, without even +giving me time to thank him. + +I entered my consulting-room and found a gentleman seated by the +table. He was quietly dressed in a suit of heather tweed with a +soft cloth cap which he had laid down upon my books. Round one of +his hands he had a handkerchief wrapped, which was mottled all +over with bloodstains. He was young, not more than +five-and-twenty, I should say, with a strong, masculine face; but +he was exceedingly pale and gave me the impression of a man who +was suffering from some strong agitation, which it took all his +strength of mind to control. + +"I am sorry to knock you up so early, Doctor," said he, "but I +have had a very serious accident during the night. I came in by +train this morning, and on inquiring at Paddington as to where I +might find a doctor, a worthy fellow very kindly escorted me +here. I gave the maid a card, but I see that she has left it upon +the side-table." + +I took it up and glanced at it. "Mr. Victor Hatherley, hydraulic +engineer, 16A, Victoria Street (3rd floor)." That was the name, +style, and abode of my morning visitor. "I regret that I have +kept you waiting," said I, sitting down in my library-chair. "You +are fresh from a night journey, I understand, which is in itself +a monotonous occupation." + +"Oh, my night could not be called monotonous," said he, and +laughed. He laughed very heartily, with a high, ringing note, +leaning back in his chair and shaking his sides. All my medical +instincts rose up against that laugh. + +"Stop it!" I cried; "pull yourself together!" and I poured out +some water from a caraffe. + +It was useless, however. He was off in one of those hysterical +outbursts which come upon a strong nature when some great crisis +is over and gone. Presently he came to himself once more, very +weary and pale-looking. + +"I have been making a fool of myself," he gasped. + +"Not at all. Drink this." I dashed some brandy into the water, +and the colour began to come back to his bloodless cheeks. + +"That's better!" said he. "And now, Doctor, perhaps you would +kindly attend to my thumb, or rather to the place where my thumb +used to be." + +He unwound the handkerchief and held out his hand. It gave even +my hardened nerves a shudder to look at it. There were four +protruding fingers and a horrid red, spongy surface where the +thumb should have been. It had been hacked or torn right out from +the roots. + +"Good heavens!" I cried, "this is a terrible injury. It must have +bled considerably." + +"Yes, it did. I fainted when it was done, and I think that I must +have been senseless for a long time. When I came to I found that +it was still bleeding, so I tied one end of my handkerchief very +tightly round the wrist and braced it up with a twig." + +"Excellent! You should have been a surgeon." + +"It is a question of hydraulics, you see, and came within my own +province." + +"This has been done," said I, examining the wound, "by a very +heavy and sharp instrument." + +"A thing like a cleaver," said he. + +"An accident, I presume?" + +"By no means." + +"What! a murderous attack?" + +"Very murderous indeed." + +"You horrify me." + +I sponged the wound, cleaned it, dressed it, and finally covered +it over with cotton wadding and carbolised bandages. He lay back +without wincing, though he bit his lip from time to time. + +"How is that?" I asked when I had finished. + +"Capital! Between your brandy and your bandage, I feel a new man. +I was very weak, but I have had a good deal to go through." + +"Perhaps you had better not speak of the matter. It is evidently +trying to your nerves." + +"Oh, no, not now. I shall have to tell my tale to the police; +but, between ourselves, if it were not for the convincing +evidence of this wound of mine, I should be surprised if they +believed my statement, for it is a very extraordinary one, and I +have not much in the way of proof with which to back it up; and, +even if they believe me, the clues which I can give them are so +vague that it is a question whether justice will be done." + +"Ha!" cried I, "if it is anything in the nature of a problem +which you desire to see solved, I should strongly recommend you +to come to my friend, Mr. Sherlock Holmes, before you go to the +official police." + +"Oh, I have heard of that fellow," answered my visitor, "and I +should be very glad if he would take the matter up, though of +course I must use the official police as well. Would you give me +an introduction to him?" + +"I'll do better. I'll take you round to him myself." + +"I should be immensely obliged to you." + +"We'll call a cab and go together. We shall just be in time to +have a little breakfast with him. Do you feel equal to it?" + +"Yes; I shall not feel easy until I have told my story." + +"Then my servant will call a cab, and I shall be with you in an +instant." I rushed upstairs, explained the matter shortly to my +wife, and in five minutes was inside a hansom, driving with my +new acquaintance to Baker Street. + +Sherlock Holmes was, as I expected, lounging about his +sitting-room in his dressing-gown, reading the agony column of The +Times and smoking his before-breakfast pipe, which was composed +of all the plugs and dottles left from his smokes of the day +before, all carefully dried and collected on the corner of the +mantelpiece. He received us in his quietly genial fashion, +ordered fresh rashers and eggs, and joined us in a hearty meal. +When it was concluded he settled our new acquaintance upon the +sofa, placed a pillow beneath his head, and laid a glass of +brandy and water within his reach. + +"It is easy to see that your experience has been no common one, +Mr. Hatherley," said he. "Pray, lie down there and make yourself +absolutely at home. Tell us what you can, but stop when you are +tired and keep up your strength with a little stimulant." + +"Thank you," said my patient, "but I have felt another man since +the doctor bandaged me, and I think that your breakfast has +completed the cure. I shall take up as little of your valuable +time as possible, so I shall start at once upon my peculiar +experiences." + +Holmes sat in his big armchair with the weary, heavy-lidded +expression which veiled his keen and eager nature, while I sat +opposite to him, and we listened in silence to the strange story +which our visitor detailed to us. + +"You must know," said he, "that I am an orphan and a bachelor, +residing alone in lodgings in London. By profession I am a +hydraulic engineer, and I have had considerable experience of my +work during the seven years that I was apprenticed to Venner & +Matheson, the well-known firm, of Greenwich. Two years ago, +having served my time, and having also come into a fair sum of +money through my poor father's death, I determined to start in +business for myself and took professional chambers in Victoria +Street. + +"I suppose that everyone finds his first independent start in +business a dreary experience. To me it has been exceptionally so. +During two years I have had three consultations and one small +job, and that is absolutely all that my profession has brought +me. My gross takings amount to 27 pounds 10s. Every day, from +nine in the morning until four in the afternoon, I waited in my +little den, until at last my heart began to sink, and I came to +believe that I should never have any practice at all. + +"Yesterday, however, just as I was thinking of leaving the +office, my clerk entered to say there was a gentleman waiting who +wished to see me upon business. He brought up a card, too, with +the name of 'Colonel Lysander Stark' engraved upon it. Close at +his heels came the colonel himself, a man rather over the middle +size, but of an exceeding thinness. I do not think that I have +ever seen so thin a man. His whole face sharpened away into nose +and chin, and the skin of his cheeks was drawn quite tense over +his outstanding bones. Yet this emaciation seemed to be his +natural habit, and due to no disease, for his eye was bright, his +step brisk, and his bearing assured. He was plainly but neatly +dressed, and his age, I should judge, would be nearer forty than +thirty. + +"'Mr. Hatherley?' said he, with something of a German accent. +'You have been recommended to me, Mr. Hatherley, as being a man +who is not only proficient in his profession but is also discreet +and capable of preserving a secret.' + +"I bowed, feeling as flattered as any young man would at such an +address. 'May I ask who it was who gave me so good a character?' + +"'Well, perhaps it is better that I should not tell you that just +at this moment. I have it from the same source that you are both +an orphan and a bachelor and are residing alone in London.' + +"'That is quite correct,' I answered; 'but you will excuse me if +I say that I cannot see how all this bears upon my professional +qualifications. I understand that it was on a professional matter +that you wished to speak to me?' + +"'Undoubtedly so. But you will find that all I say is really to +the point. I have a professional commission for you, but absolute +secrecy is quite essential--absolute secrecy, you understand, and +of course we may expect that more from a man who is alone than +from one who lives in the bosom of his family.' + +"'If I promise to keep a secret,' said I, 'you may absolutely +depend upon my doing so.' + +"He looked very hard at me as I spoke, and it seemed to me that I +had never seen so suspicious and questioning an eye. + +"'Do you promise, then?' said he at last. + +"'Yes, I promise.' + +"'Absolute and complete silence before, during, and after? No +reference to the matter at all, either in word or writing?' + +"'I have already given you my word.' + +"'Very good.' He suddenly sprang up, and darting like lightning +across the room he flung open the door. The passage outside was +empty. + +"'That's all right,' said he, coming back. 'I know that clerks are +sometimes curious as to their master's affairs. Now we can talk +in safety.' He drew up his chair very close to mine and began to +stare at me again with the same questioning and thoughtful look. + +"A feeling of repulsion, and of something akin to fear had begun +to rise within me at the strange antics of this fleshless man. +Even my dread of losing a client could not restrain me from +showing my impatience. + +"'I beg that you will state your business, sir,' said I; 'my time +is of value.' Heaven forgive me for that last sentence, but the +words came to my lips. + +"'How would fifty guineas for a night's work suit you?' he asked. + +"'Most admirably.' + +"'I say a night's work, but an hour's would be nearer the mark. I +simply want your opinion about a hydraulic stamping machine which +has got out of gear. If you show us what is wrong we shall soon +set it right ourselves. What do you think of such a commission as +that?' + +"'The work appears to be light and the pay munificent.' + +"'Precisely so. We shall want you to come to-night by the last +train.' + +"'Where to?' + +"'To Eyford, in Berkshire. It is a little place near the borders +of Oxfordshire, and within seven miles of Reading. There is a +train from Paddington which would bring you there at about +11:15.' + +"'Very good.' + +"'I shall come down in a carriage to meet you.' + +"'There is a drive, then?' + +"'Yes, our little place is quite out in the country. It is a good +seven miles from Eyford Station.' + +"'Then we can hardly get there before midnight. I suppose there +would be no chance of a train back. I should be compelled to stop +the night.' + +"'Yes, we could easily give you a shake-down.' + +"'That is very awkward. Could I not come at some more convenient +hour?' + +"'We have judged it best that you should come late. It is to +recompense you for any inconvenience that we are paying to you, a +young and unknown man, a fee which would buy an opinion from the +very heads of your profession. Still, of course, if you would +like to draw out of the business, there is plenty of time to do +so.' + +"I thought of the fifty guineas, and of how very useful they +would be to me. 'Not at all,' said I, 'I shall be very happy to +accommodate myself to your wishes. I should like, however, to +understand a little more clearly what it is that you wish me to +do.' + +"'Quite so. It is very natural that the pledge of secrecy which +we have exacted from you should have aroused your curiosity. I +have no wish to commit you to anything without your having it all +laid before you. I suppose that we are absolutely safe from +eavesdroppers?' + +"'Entirely.' + +"'Then the matter stands thus. You are probably aware that +fuller's-earth is a valuable product, and that it is only found +in one or two places in England?' + +"'I have heard so.' + +"'Some little time ago I bought a small place--a very small +place--within ten miles of Reading. I was fortunate enough to +discover that there was a deposit of fuller's-earth in one of my +fields. On examining it, however, I found that this deposit was a +comparatively small one, and that it formed a link between two +very much larger ones upon the right and left--both of them, +however, in the grounds of my neighbours. These good people were +absolutely ignorant that their land contained that which was +quite as valuable as a gold-mine. Naturally, it was to my +interest to buy their land before they discovered its true value, +but unfortunately I had no capital by which I could do this. I +took a few of my friends into the secret, however, and they +suggested that we should quietly and secretly work our own little +deposit and that in this way we should earn the money which would +enable us to buy the neighbouring fields. This we have now been +doing for some time, and in order to help us in our operations we +erected a hydraulic press. This press, as I have already +explained, has got out of order, and we wish your advice upon the +subject. We guard our secret very jealously, however, and if it +once became known that we had hydraulic engineers coming to our +little house, it would soon rouse inquiry, and then, if the facts +came out, it would be good-bye to any chance of getting these +fields and carrying out our plans. That is why I have made you +promise me that you will not tell a human being that you are +going to Eyford to-night. I hope that I make it all plain?' + +"'I quite follow you,' said I. 'The only point which I could not +quite understand was what use you could make of a hydraulic press +in excavating fuller's-earth, which, as I understand, is dug out +like gravel from a pit.' + +"'Ah!' said he carelessly, 'we have our own process. We compress +the earth into bricks, so as to remove them without revealing +what they are. But that is a mere detail. I have taken you fully +into my confidence now, Mr. Hatherley, and I have shown you how I +trust you.' He rose as he spoke. 'I shall expect you, then, at +Eyford at 11:15.' + +"'I shall certainly be there.' + +"'And not a word to a soul.' He looked at me with a last long, +questioning gaze, and then, pressing my hand in a cold, dank +grasp, he hurried from the room. + +"Well, when I came to think it all over in cool blood I was very +much astonished, as you may both think, at this sudden commission +which had been intrusted to me. On the one hand, of course, I was +glad, for the fee was at least tenfold what I should have asked +had I set a price upon my own services, and it was possible that +this order might lead to other ones. On the other hand, the face +and manner of my patron had made an unpleasant impression upon +me, and I could not think that his explanation of the +fuller's-earth was sufficient to explain the necessity for my +coming at midnight, and his extreme anxiety lest I should tell +anyone of my errand. However, I threw all fears to the winds, ate +a hearty supper, drove to Paddington, and started off, having +obeyed to the letter the injunction as to holding my tongue. + +"At Reading I had to change not only my carriage but my station. +However, I was in time for the last train to Eyford, and I +reached the little dim-lit station after eleven o'clock. I was the +only passenger who got out there, and there was no one upon the +platform save a single sleepy porter with a lantern. As I passed +out through the wicket gate, however, I found my acquaintance of +the morning waiting in the shadow upon the other side. Without a +word he grasped my arm and hurried me into a carriage, the door +of which was standing open. He drew up the windows on either +side, tapped on the wood-work, and away we went as fast as the +horse could go." + +"One horse?" interjected Holmes. + +"Yes, only one." + +"Did you observe the colour?" + +"Yes, I saw it by the side-lights when I was stepping into the +carriage. It was a chestnut." + +"Tired-looking or fresh?" + +"Oh, fresh and glossy." + +"Thank you. I am sorry to have interrupted you. Pray continue +your most interesting statement." + +"Away we went then, and we drove for at least an hour. Colonel +Lysander Stark had said that it was only seven miles, but I +should think, from the rate that we seemed to go, and from the +time that we took, that it must have been nearer twelve. He sat +at my side in silence all the time, and I was aware, more than +once when I glanced in his direction, that he was looking at me +with great intensity. The country roads seem to be not very good +in that part of the world, for we lurched and jolted terribly. I +tried to look out of the windows to see something of where we +were, but they were made of frosted glass, and I could make out +nothing save the occasional bright blur of a passing light. Now +and then I hazarded some remark to break the monotony of the +journey, but the colonel answered only in monosyllables, and the +conversation soon flagged. At last, however, the bumping of the +road was exchanged for the crisp smoothness of a gravel-drive, +and the carriage came to a stand. Colonel Lysander Stark sprang +out, and, as I followed after him, pulled me swiftly into a porch +which gaped in front of us. We stepped, as it were, right out of +the carriage and into the hall, so that I failed to catch the +most fleeting glance of the front of the house. The instant that +I had crossed the threshold the door slammed heavily behind us, +and I heard faintly the rattle of the wheels as the carriage +drove away. + +"It was pitch dark inside the house, and the colonel fumbled +about looking for matches and muttering under his breath. +Suddenly a door opened at the other end of the passage, and a +long, golden bar of light shot out in our direction. It grew +broader, and a woman appeared with a lamp in her hand, which she +held above her head, pushing her face forward and peering at us. +I could see that she was pretty, and from the gloss with which +the light shone upon her dark dress I knew that it was a rich +material. She spoke a few words in a foreign tongue in a tone as +though asking a question, and when my companion answered in a +gruff monosyllable she gave such a start that the lamp nearly +fell from her hand. Colonel Stark went up to her, whispered +something in her ear, and then, pushing her back into the room +from whence she had come, he walked towards me again with the +lamp in his hand. + +"'Perhaps you will have the kindness to wait in this room for a +few minutes,' said he, throwing open another door. It was a +quiet, little, plainly furnished room, with a round table in the +centre, on which several German books were scattered. Colonel +Stark laid down the lamp on the top of a harmonium beside the +door. 'I shall not keep you waiting an instant,' said he, and +vanished into the darkness. + +"I glanced at the books upon the table, and in spite of my +ignorance of German I could see that two of them were treatises +on science, the others being volumes of poetry. Then I walked +across to the window, hoping that I might catch some glimpse of +the country-side, but an oak shutter, heavily barred, was folded +across it. It was a wonderfully silent house. There was an old +clock ticking loudly somewhere in the passage, but otherwise +everything was deadly still. A vague feeling of uneasiness began +to steal over me. Who were these German people, and what were +they doing living in this strange, out-of-the-way place? And +where was the place? I was ten miles or so from Eyford, that was +all I knew, but whether north, south, east, or west I had no +idea. For that matter, Reading, and possibly other large towns, +were within that radius, so the place might not be so secluded, +after all. Yet it was quite certain, from the absolute stillness, +that we were in the country. I paced up and down the room, +humming a tune under my breath to keep up my spirits and feeling +that I was thoroughly earning my fifty-guinea fee. + +"Suddenly, without any preliminary sound in the midst of the +utter stillness, the door of my room swung slowly open. The woman +was standing in the aperture, the darkness of the hall behind +her, the yellow light from my lamp beating upon her eager and +beautiful face. I could see at a glance that she was sick with +fear, and the sight sent a chill to my own heart. She held up one +shaking finger to warn me to be silent, and she shot a few +whispered words of broken English at me, her eyes glancing back, +like those of a frightened horse, into the gloom behind her. + +"'I would go,' said she, trying hard, as it seemed to me, to +speak calmly; 'I would go. I should not stay here. There is no +good for you to do.' + +"'But, madam,' said I, 'I have not yet done what I came for. I +cannot possibly leave until I have seen the machine.' + +"'It is not worth your while to wait,' she went on. 'You can pass +through the door; no one hinders.' And then, seeing that I smiled +and shook my head, she suddenly threw aside her constraint and +made a step forward, with her hands wrung together. 'For the love +of Heaven!' she whispered, 'get away from here before it is too +late!' + +"But I am somewhat headstrong by nature, and the more ready to +engage in an affair when there is some obstacle in the way. I +thought of my fifty-guinea fee, of my wearisome journey, and of +the unpleasant night which seemed to be before me. Was it all to +go for nothing? Why should I slink away without having carried +out my commission, and without the payment which was my due? This +woman might, for all I knew, be a monomaniac. With a stout +bearing, therefore, though her manner had shaken me more than I +cared to confess, I still shook my head and declared my intention +of remaining where I was. She was about to renew her entreaties +when a door slammed overhead, and the sound of several footsteps +was heard upon the stairs. She listened for an instant, threw up +her hands with a despairing gesture, and vanished as suddenly and +as noiselessly as she had come. + +"The newcomers were Colonel Lysander Stark and a short thick man +with a chinchilla beard growing out of the creases of his double +chin, who was introduced to me as Mr. Ferguson. + +"'This is my secretary and manager,' said the colonel. 'By the +way, I was under the impression that I left this door shut just +now. I fear that you have felt the draught.' + +"'On the contrary,' said I, 'I opened the door myself because I +felt the room to be a little close.' + +"He shot one of his suspicious looks at me. 'Perhaps we had +better proceed to business, then,' said he. 'Mr. Ferguson and I +will take you up to see the machine.' + +"'I had better put my hat on, I suppose.' + +"'Oh, no, it is in the house.' + +"'What, you dig fuller's-earth in the house?' + +"'No, no. This is only where we compress it. But never mind that. +All we wish you to do is to examine the machine and to let us +know what is wrong with it.' + +"We went upstairs together, the colonel first with the lamp, the +fat manager and I behind him. It was a labyrinth of an old house, +with corridors, passages, narrow winding staircases, and little +low doors, the thresholds of which were hollowed out by the +generations who had crossed them. There were no carpets and no +signs of any furniture above the ground floor, while the plaster +was peeling off the walls, and the damp was breaking through in +green, unhealthy blotches. I tried to put on as unconcerned an +air as possible, but I had not forgotten the warnings of the +lady, even though I disregarded them, and I kept a keen eye upon +my two companions. Ferguson appeared to be a morose and silent +man, but I could see from the little that he said that he was at +least a fellow-countryman. + +"Colonel Lysander Stark stopped at last before a low door, which +he unlocked. Within was a small, square room, in which the three +of us could hardly get at one time. Ferguson remained outside, +and the colonel ushered me in. + +"'We are now,' said he, 'actually within the hydraulic press, and +it would be a particularly unpleasant thing for us if anyone were +to turn it on. The ceiling of this small chamber is really the +end of the descending piston, and it comes down with the force of +many tons upon this metal floor. There are small lateral columns +of water outside which receive the force, and which transmit and +multiply it in the manner which is familiar to you. The machine +goes readily enough, but there is some stiffness in the working +of it, and it has lost a little of its force. Perhaps you will +have the goodness to look it over and to show us how we can set +it right.' + +"I took the lamp from him, and I examined the machine very +thoroughly. It was indeed a gigantic one, and capable of +exercising enormous pressure. When I passed outside, however, and +pressed down the levers which controlled it, I knew at once by +the whishing sound that there was a slight leakage, which allowed +a regurgitation of water through one of the side cylinders. An +examination showed that one of the india-rubber bands which was +round the head of a driving-rod had shrunk so as not quite to +fill the socket along which it worked. This was clearly the cause +of the loss of power, and I pointed it out to my companions, who +followed my remarks very carefully and asked several practical +questions as to how they should proceed to set it right. When I +had made it clear to them, I returned to the main chamber of the +machine and took a good look at it to satisfy my own curiosity. +It was obvious at a glance that the story of the fuller's-earth +was the merest fabrication, for it would be absurd to suppose +that so powerful an engine could be designed for so inadequate a +purpose. The walls were of wood, but the floor consisted of a +large iron trough, and when I came to examine it I could see a +crust of metallic deposit all over it. I had stooped and was +scraping at this to see exactly what it was when I heard a +muttered exclamation in German and saw the cadaverous face of the +colonel looking down at me. + +"'What are you doing there?' he asked. + +"I felt angry at having been tricked by so elaborate a story as +that which he had told me. 'I was admiring your fuller's-earth,' +said I; 'I think that I should be better able to advise you as to +your machine if I knew what the exact purpose was for which it +was used.' + +"The instant that I uttered the words I regretted the rashness of +my speech. His face set hard, and a baleful light sprang up in +his grey eyes. + +"'Very well,' said he, 'you shall know all about the machine.' He +took a step backward, slammed the little door, and turned the key +in the lock. I rushed towards it and pulled at the handle, but it +was quite secure, and did not give in the least to my kicks and +shoves. 'Hullo!' I yelled. 'Hullo! Colonel! Let me out!' + +"And then suddenly in the silence I heard a sound which sent my +heart into my mouth. It was the clank of the levers and the swish +of the leaking cylinder. He had set the engine at work. The lamp +still stood upon the floor where I had placed it when examining +the trough. By its light I saw that the black ceiling was coming +down upon me, slowly, jerkily, but, as none knew better than +myself, with a force which must within a minute grind me to a +shapeless pulp. I threw myself, screaming, against the door, and +dragged with my nails at the lock. I implored the colonel to let +me out, but the remorseless clanking of the levers drowned my +cries. The ceiling was only a foot or two above my head, and with +my hand upraised I could feel its hard, rough surface. Then it +flashed through my mind that the pain of my death would depend +very much upon the position in which I met it. If I lay on my +face the weight would come upon my spine, and I shuddered to +think of that dreadful snap. Easier the other way, perhaps; and +yet, had I the nerve to lie and look up at that deadly black +shadow wavering down upon me? Already I was unable to stand +erect, when my eye caught something which brought a gush of hope +back to my heart. + +"I have said that though the floor and ceiling were of iron, the +walls were of wood. As I gave a last hurried glance around, I saw +a thin line of yellow light between two of the boards, which +broadened and broadened as a small panel was pushed backward. For +an instant I could hardly believe that here was indeed a door +which led away from death. The next instant I threw myself +through, and lay half-fainting upon the other side. The panel had +closed again behind me, but the crash of the lamp, and a few +moments afterwards the clang of the two slabs of metal, told me +how narrow had been my escape. + +"I was recalled to myself by a frantic plucking at my wrist, and +I found myself lying upon the stone floor of a narrow corridor, +while a woman bent over me and tugged at me with her left hand, +while she held a candle in her right. It was the same good friend +whose warning I had so foolishly rejected. + +"'Come! come!' she cried breathlessly. 'They will be here in a +moment. They will see that you are not there. Oh, do not waste +the so-precious time, but come!' + +"This time, at least, I did not scorn her advice. I staggered to +my feet and ran with her along the corridor and down a winding +stair. The latter led to another broad passage, and just as we +reached it we heard the sound of running feet and the shouting of +two voices, one answering the other from the floor on which we +were and from the one beneath. My guide stopped and looked about +her like one who is at her wit's end. Then she threw open a door +which led into a bedroom, through the window of which the moon +was shining brightly. + +"'It is your only chance,' said she. 'It is high, but it may be +that you can jump it.' + +"As she spoke a light sprang into view at the further end of the +passage, and I saw the lean figure of Colonel Lysander Stark +rushing forward with a lantern in one hand and a weapon like a +butcher's cleaver in the other. I rushed across the bedroom, +flung open the window, and looked out. How quiet and sweet and +wholesome the garden looked in the moonlight, and it could not be +more than thirty feet down. I clambered out upon the sill, but I +hesitated to jump until I should have heard what passed between +my saviour and the ruffian who pursued me. If she were ill-used, +then at any risks I was determined to go back to her assistance. +The thought had hardly flashed through my mind before he was at +the door, pushing his way past her; but she threw her arms round +him and tried to hold him back. + +"'Fritz! Fritz!' she cried in English, 'remember your promise +after the last time. You said it should not be again. He will be +silent! Oh, he will be silent!' + +"'You are mad, Elise!' he shouted, struggling to break away from +her. 'You will be the ruin of us. He has seen too much. Let me +pass, I say!' He dashed her to one side, and, rushing to the +window, cut at me with his heavy weapon. I had let myself go, and +was hanging by the hands to the sill, when his blow fell. I was +conscious of a dull pain, my grip loosened, and I fell into the +garden below. + +"I was shaken but not hurt by the fall; so I picked myself up and +rushed off among the bushes as hard as I could run, for I +understood that I was far from being out of danger yet. Suddenly, +however, as I ran, a deadly dizziness and sickness came over me. +I glanced down at my hand, which was throbbing painfully, and +then, for the first time, saw that my thumb had been cut off and +that the blood was pouring from my wound. I endeavoured to tie my +handkerchief round it, but there came a sudden buzzing in my +ears, and next moment I fell in a dead faint among the +rose-bushes. + +"How long I remained unconscious I cannot tell. It must have been +a very long time, for the moon had sunk, and a bright morning was +breaking when I came to myself. My clothes were all sodden with +dew, and my coat-sleeve was drenched with blood from my wounded +thumb. The smarting of it recalled in an instant all the +particulars of my night's adventure, and I sprang to my feet with +the feeling that I might hardly yet be safe from my pursuers. But +to my astonishment, when I came to look round me, neither house +nor garden were to be seen. I had been lying in an angle of the +hedge close by the highroad, and just a little lower down was a +long building, which proved, upon my approaching it, to be the +very station at which I had arrived upon the previous night. Were +it not for the ugly wound upon my hand, all that had passed +during those dreadful hours might have been an evil dream. + +"Half dazed, I went into the station and asked about the morning +train. There would be one to Reading in less than an hour. The +same porter was on duty, I found, as had been there when I +arrived. I inquired of him whether he had ever heard of Colonel +Lysander Stark. The name was strange to him. Had he observed a +carriage the night before waiting for me? No, he had not. Was +there a police-station anywhere near? There was one about three +miles off. + +"It was too far for me to go, weak and ill as I was. I determined +to wait until I got back to town before telling my story to the +police. It was a little past six when I arrived, so I went first +to have my wound dressed, and then the doctor was kind enough to +bring me along here. I put the case into your hands and shall do +exactly what you advise." + +We both sat in silence for some little time after listening to +this extraordinary narrative. Then Sherlock Holmes pulled down +from the shelf one of the ponderous commonplace books in which he +placed his cuttings. + +"Here is an advertisement which will interest you," said he. "It +appeared in all the papers about a year ago. Listen to this: +'Lost, on the 9th inst., Mr. Jeremiah Hayling, aged +twenty-six, a hydraulic engineer. Left his lodgings at ten +o'clock at night, and has not been heard of since. Was +dressed in,' etc., etc. Ha! That represents the last time that +the colonel needed to have his machine overhauled, I fancy." + +"Good heavens!" cried my patient. "Then that explains what the +girl said." + +"Undoubtedly. It is quite clear that the colonel was a cool and +desperate man, who was absolutely determined that nothing should +stand in the way of his little game, like those out-and-out +pirates who will leave no survivor from a captured ship. Well, +every moment now is precious, so if you feel equal to it we shall +go down to Scotland Yard at once as a preliminary to starting for +Eyford." + +Some three hours or so afterwards we were all in the train +together, bound from Reading to the little Berkshire village. +There were Sherlock Holmes, the hydraulic engineer, Inspector +Bradstreet, of Scotland Yard, a plain-clothes man, and myself. +Bradstreet had spread an ordnance map of the county out upon the +seat and was busy with his compasses drawing a circle with Eyford +for its centre. + +"There you are," said he. "That circle is drawn at a radius of +ten miles from the village. The place we want must be somewhere +near that line. You said ten miles, I think, sir." + +"It was an hour's good drive." + +"And you think that they brought you back all that way when you +were unconscious?" + +"They must have done so. I have a confused memory, too, of having +been lifted and conveyed somewhere." + +"What I cannot understand," said I, "is why they should have +spared you when they found you lying fainting in the garden. +Perhaps the villain was softened by the woman's entreaties." + +"I hardly think that likely. I never saw a more inexorable face +in my life." + +"Oh, we shall soon clear up all that," said Bradstreet. "Well, I +have drawn my circle, and I only wish I knew at what point upon +it the folk that we are in search of are to be found." + +"I think I could lay my finger on it," said Holmes quietly. + +"Really, now!" cried the inspector, "you have formed your +opinion! Come, now, we shall see who agrees with you. I say it is +south, for the country is more deserted there." + +"And I say east," said my patient. + +"I am for west," remarked the plain-clothes man. "There are +several quiet little villages up there." + +"And I am for north," said I, "because there are no hills there, +and our friend says that he did not notice the carriage go up +any." + +"Come," cried the inspector, laughing; "it's a very pretty +diversity of opinion. We have boxed the compass among us. Who do +you give your casting vote to?" + +"You are all wrong." + +"But we can't all be." + +"Oh, yes, you can. This is my point." He placed his finger in the +centre of the circle. "This is where we shall find them." + +"But the twelve-mile drive?" gasped Hatherley. + +"Six out and six back. Nothing simpler. You say yourself that the +horse was fresh and glossy when you got in. How could it be that +if it had gone twelve miles over heavy roads?" + +"Indeed, it is a likely ruse enough," observed Bradstreet +thoughtfully. "Of course there can be no doubt as to the nature +of this gang." + +"None at all," said Holmes. "They are coiners on a large scale, +and have used the machine to form the amalgam which has taken the +place of silver." + +"We have known for some time that a clever gang was at work," +said the inspector. "They have been turning out half-crowns by +the thousand. We even traced them as far as Reading, but could +get no farther, for they had covered their traces in a way that +showed that they were very old hands. But now, thanks to this +lucky chance, I think that we have got them right enough." + +But the inspector was mistaken, for those criminals were not +destined to fall into the hands of justice. As we rolled into +Eyford Station we saw a gigantic column of smoke which streamed +up from behind a small clump of trees in the neighbourhood and +hung like an immense ostrich feather over the landscape. + +"A house on fire?" asked Bradstreet as the train steamed off +again on its way. + +"Yes, sir!" said the station-master. + +"When did it break out?" + +"I hear that it was during the night, sir, but it has got worse, +and the whole place is in a blaze." + +"Whose house is it?" + +"Dr. Becher's." + +"Tell me," broke in the engineer, "is Dr. Becher a German, very +thin, with a long, sharp nose?" + +The station-master laughed heartily. "No, sir, Dr. Becher is an +Englishman, and there isn't a man in the parish who has a +better-lined waistcoat. But he has a gentleman staying with him, +a patient, as I understand, who is a foreigner, and he looks as +if a little good Berkshire beef would do him no harm." + +The station-master had not finished his speech before we were all +hastening in the direction of the fire. The road topped a low +hill, and there was a great widespread whitewashed building in +front of us, spouting fire at every chink and window, while in +the garden in front three fire-engines were vainly striving to +keep the flames under. + +"That's it!" cried Hatherley, in intense excitement. "There is +the gravel-drive, and there are the rose-bushes where I lay. That +second window is the one that I jumped from." + +"Well, at least," said Holmes, "you have had your revenge upon +them. There can be no question that it was your oil-lamp which, +when it was crushed in the press, set fire to the wooden walls, +though no doubt they were too excited in the chase after you to +observe it at the time. Now keep your eyes open in this crowd for +your friends of last night, though I very much fear that they are +a good hundred miles off by now." + +And Holmes' fears came to be realised, for from that day to this +no word has ever been heard either of the beautiful woman, the +sinister German, or the morose Englishman. Early that morning a +peasant had met a cart containing several people and some very +bulky boxes driving rapidly in the direction of Reading, but +there all traces of the fugitives disappeared, and even Holmes' +ingenuity failed ever to discover the least clue as to their +whereabouts. + +The firemen had been much perturbed at the strange arrangements +which they had found within, and still more so by discovering a +newly severed human thumb upon a window-sill of the second floor. +About sunset, however, their efforts were at last successful, and +they subdued the flames, but not before the roof had fallen in, +and the whole place been reduced to such absolute ruin that, save +some twisted cylinders and iron piping, not a trace remained of +the machinery which had cost our unfortunate acquaintance so +dearly. Large masses of nickel and of tin were discovered stored +in an out-house, but no coins were to be found, which may have +explained the presence of those bulky boxes which have been +already referred to. + +How our hydraulic engineer had been conveyed from the garden to +the spot where he recovered his senses might have remained +forever a mystery were it not for the soft mould, which told us a +very plain tale. He had evidently been carried down by two +persons, one of whom had remarkably small feet and the other +unusually large ones. On the whole, it was most probable that the +silent Englishman, being less bold or less murderous than his +companion, had assisted the woman to bear the unconscious man out +of the way of danger. + +"Well," said our engineer ruefully as we took our seats to return +once more to London, "it has been a pretty business for me! I +have lost my thumb and I have lost a fifty-guinea fee, and what +have I gained?" + +"Experience," said Holmes, laughing. "Indirectly it may be of +value, you know; you have only to put it into words to gain the +reputation of being excellent company for the remainder of your +existence." + + + +X. THE ADVENTURE OF THE NOBLE BACHELOR + +The Lord St. Simon marriage, and its curious termination, have +long ceased to be a subject of interest in those exalted circles +in which the unfortunate bridegroom moves. Fresh scandals have +eclipsed it, and their more piquant details have drawn the +gossips away from this four-year-old drama. As I have reason to +believe, however, that the full facts have never been revealed to +the general public, and as my friend Sherlock Holmes had a +considerable share in clearing the matter up, I feel that no +memoir of him would be complete without some little sketch of +this remarkable episode. + +It was a few weeks before my own marriage, during the days when I +was still sharing rooms with Holmes in Baker Street, that he came +home from an afternoon stroll to find a letter on the table +waiting for him. I had remained indoors all day, for the weather +had taken a sudden turn to rain, with high autumnal winds, and +the Jezail bullet which I had brought back in one of my limbs as +a relic of my Afghan campaign throbbed with dull persistence. +With my body in one easy-chair and my legs upon another, I had +surrounded myself with a cloud of newspapers until at last, +saturated with the news of the day, I tossed them all aside and +lay listless, watching the huge crest and monogram upon the +envelope upon the table and wondering lazily who my friend's +noble correspondent could be. + +"Here is a very fashionable epistle," I remarked as he entered. +"Your morning letters, if I remember right, were from a +fish-monger and a tide-waiter." + +"Yes, my correspondence has certainly the charm of variety," he +answered, smiling, "and the humbler are usually the more +interesting. This looks like one of those unwelcome social +summonses which call upon a man either to be bored or to lie." + +He broke the seal and glanced over the contents. + +"Oh, come, it may prove to be something of interest, after all." + +"Not social, then?" + +"No, distinctly professional." + +"And from a noble client?" + +"One of the highest in England." + +"My dear fellow, I congratulate you." + +"I assure you, Watson, without affectation, that the status of my +client is a matter of less moment to me than the interest of his +case. It is just possible, however, that that also may not be +wanting in this new investigation. You have been reading the +papers diligently of late, have you not?" + +"It looks like it," said I ruefully, pointing to a huge bundle in +the corner. "I have had nothing else to do." + +"It is fortunate, for you will perhaps be able to post me up. I +read nothing except the criminal news and the agony column. The +latter is always instructive. But if you have followed recent +events so closely you must have read about Lord St. Simon and his +wedding?" + +"Oh, yes, with the deepest interest." + +"That is well. The letter which I hold in my hand is from Lord +St. Simon. I will read it to you, and in return you must turn +over these papers and let me have whatever bears upon the matter. +This is what he says: + +"'MY DEAR MR. SHERLOCK HOLMES:--Lord Backwater tells me that I +may place implicit reliance upon your judgment and discretion. I +have determined, therefore, to call upon you and to consult you +in reference to the very painful event which has occurred in +connection with my wedding. Mr. Lestrade, of Scotland Yard, is +acting already in the matter, but he assures me that he sees no +objection to your co-operation, and that he even thinks that +it might be of some assistance. I will call at four o'clock in +the afternoon, and, should you have any other engagement at that +time, I hope that you will postpone it, as this matter is of +paramount importance. Yours faithfully, ST. SIMON.' + +"It is dated from Grosvenor Mansions, written with a quill pen, +and the noble lord has had the misfortune to get a smear of ink +upon the outer side of his right little finger," remarked Holmes +as he folded up the epistle. + +"He says four o'clock. It is three now. He will be here in an +hour." + +"Then I have just time, with your assistance, to get clear upon +the subject. Turn over those papers and arrange the extracts in +their order of time, while I take a glance as to who our client +is." He picked a red-covered volume from a line of books of +reference beside the mantelpiece. "Here he is," said he, sitting +down and flattening it out upon his knee. "'Lord Robert Walsingham +de Vere St. Simon, second son of the Duke of Balmoral.' Hum! 'Arms: +Azure, three caltrops in chief over a fess sable. Born in 1846.' +He's forty-one years of age, which is mature for marriage. Was +Under-Secretary for the colonies in a late administration. The +Duke, his father, was at one time Secretary for Foreign Affairs. +They inherit Plantagenet blood by direct descent, and Tudor on +the distaff side. Ha! Well, there is nothing very instructive in +all this. I think that I must turn to you Watson, for something +more solid." + +"I have very little difficulty in finding what I want," said I, +"for the facts are quite recent, and the matter struck me as +remarkable. I feared to refer them to you, however, as I knew +that you had an inquiry on hand and that you disliked the +intrusion of other matters." + +"Oh, you mean the little problem of the Grosvenor Square +furniture van. That is quite cleared up now--though, indeed, it +was obvious from the first. Pray give me the results of your +newspaper selections." + +"Here is the first notice which I can find. It is in the personal +column of the Morning Post, and dates, as you see, some weeks +back: 'A marriage has been arranged,' it says, 'and will, if +rumour is correct, very shortly take place, between Lord Robert +St. Simon, second son of the Duke of Balmoral, and Miss Hatty +Doran, the only daughter of Aloysius Doran. Esq., of San +Francisco, Cal., U.S.A.' That is all." + +"Terse and to the point," remarked Holmes, stretching his long, +thin legs towards the fire. + +"There was a paragraph amplifying this in one of the society +papers of the same week. Ah, here it is: 'There will soon be a +call for protection in the marriage market, for the present +free-trade principle appears to tell heavily against our home +product. One by one the management of the noble houses of Great +Britain is passing into the hands of our fair cousins from across +the Atlantic. An important addition has been made during the last +week to the list of the prizes which have been borne away by +these charming invaders. Lord St. Simon, who has shown himself +for over twenty years proof against the little god's arrows, has +now definitely announced his approaching marriage with Miss Hatty +Doran, the fascinating daughter of a California millionaire. Miss +Doran, whose graceful figure and striking face attracted much +attention at the Westbury House festivities, is an only child, +and it is currently reported that her dowry will run to +considerably over the six figures, with expectancies for the +future. As it is an open secret that the Duke of Balmoral has +been compelled to sell his pictures within the last few years, +and as Lord St. Simon has no property of his own save the small +estate of Birchmoor, it is obvious that the Californian heiress +is not the only gainer by an alliance which will enable her to +make the easy and common transition from a Republican lady to a +British peeress.'" + +"Anything else?" asked Holmes, yawning. + +"Oh, yes; plenty. Then there is another note in the Morning Post +to say that the marriage would be an absolutely quiet one, that it +would be at St. George's, Hanover Square, that only half a dozen +intimate friends would be invited, and that the party would +return to the furnished house at Lancaster Gate which has been +taken by Mr. Aloysius Doran. Two days later--that is, on +Wednesday last--there is a curt announcement that the wedding had +taken place, and that the honeymoon would be passed at Lord +Backwater's place, near Petersfield. Those are all the notices +which appeared before the disappearance of the bride." + +"Before the what?" asked Holmes with a start. + +"The vanishing of the lady." + +"When did she vanish, then?" + +"At the wedding breakfast." + +"Indeed. This is more interesting than it promised to be; quite +dramatic, in fact." + +"Yes; it struck me as being a little out of the common." + +"They often vanish before the ceremony, and occasionally during +the honeymoon; but I cannot call to mind anything quite so prompt +as this. Pray let me have the details." + +"I warn you that they are very incomplete." + +"Perhaps we may make them less so." + +"Such as they are, they are set forth in a single article of a +morning paper of yesterday, which I will read to you. It is +headed, 'Singular Occurrence at a Fashionable Wedding': + +"'The family of Lord Robert St. Simon has been thrown into the +greatest consternation by the strange and painful episodes which +have taken place in connection with his wedding. The ceremony, as +shortly announced in the papers of yesterday, occurred on the +previous morning; but it is only now that it has been possible to +confirm the strange rumours which have been so persistently +floating about. In spite of the attempts of the friends to hush +the matter up, so much public attention has now been drawn to it +that no good purpose can be served by affecting to disregard what +is a common subject for conversation. + +"'The ceremony, which was performed at St. George's, Hanover +Square, was a very quiet one, no one being present save the +father of the bride, Mr. Aloysius Doran, the Duchess of Balmoral, +Lord Backwater, Lord Eustace and Lady Clara St. Simon (the +younger brother and sister of the bridegroom), and Lady Alicia +Whittington. The whole party proceeded afterwards to the house of +Mr. Aloysius Doran, at Lancaster Gate, where breakfast had been +prepared. It appears that some little trouble was caused by a +woman, whose name has not been ascertained, who endeavoured to +force her way into the house after the bridal party, alleging +that she had some claim upon Lord St. Simon. It was only after a +painful and prolonged scene that she was ejected by the butler +and the footman. The bride, who had fortunately entered the house +before this unpleasant interruption, had sat down to breakfast +with the rest, when she complained of a sudden indisposition and +retired to her room. Her prolonged absence having caused some +comment, her father followed her, but learned from her maid that +she had only come up to her chamber for an instant, caught up an +ulster and bonnet, and hurried down to the passage. One of the +footmen declared that he had seen a lady leave the house thus +apparelled, but had refused to credit that it was his mistress, +believing her to be with the company. On ascertaining that his +daughter had disappeared, Mr. Aloysius Doran, in conjunction with +the bridegroom, instantly put themselves in communication with +the police, and very energetic inquiries are being made, which +will probably result in a speedy clearing up of this very +singular business. Up to a late hour last night, however, nothing +had transpired as to the whereabouts of the missing lady. There +are rumours of foul play in the matter, and it is said that the +police have caused the arrest of the woman who had caused the +original disturbance, in the belief that, from jealousy or some +other motive, she may have been concerned in the strange +disappearance of the bride.'" + +"And is that all?" + +"Only one little item in another of the morning papers, but it is +a suggestive one." + +"And it is--" + +"That Miss Flora Millar, the lady who had caused the disturbance, +has actually been arrested. It appears that she was formerly a +danseuse at the Allegro, and that she has known the bridegroom +for some years. There are no further particulars, and the whole +case is in your hands now--so far as it has been set forth in the +public press." + +"And an exceedingly interesting case it appears to be. I would +not have missed it for worlds. But there is a ring at the bell, +Watson, and as the clock makes it a few minutes after four, I +have no doubt that this will prove to be our noble client. Do not +dream of going, Watson, for I very much prefer having a witness, +if only as a check to my own memory." + +"Lord Robert St. Simon," announced our page-boy, throwing open +the door. A gentleman entered, with a pleasant, cultured face, +high-nosed and pale, with something perhaps of petulance about +the mouth, and with the steady, well-opened eye of a man whose +pleasant lot it had ever been to command and to be obeyed. His +manner was brisk, and yet his general appearance gave an undue +impression of age, for he had a slight forward stoop and a little +bend of the knees as he walked. His hair, too, as he swept off +his very curly-brimmed hat, was grizzled round the edges and thin +upon the top. As to his dress, it was careful to the verge of +foppishness, with high collar, black frock-coat, white waistcoat, +yellow gloves, patent-leather shoes, and light-coloured gaiters. +He advanced slowly into the room, turning his head from left to +right, and swinging in his right hand the cord which held his +golden eyeglasses. + +"Good-day, Lord St. Simon," said Holmes, rising and bowing. "Pray +take the basket-chair. This is my friend and colleague, Dr. +Watson. Draw up a little to the fire, and we will talk this +matter over." + +"A most painful matter to me, as you can most readily imagine, +Mr. Holmes. I have been cut to the quick. I understand that you +have already managed several delicate cases of this sort, sir, +though I presume that they were hardly from the same class of +society." + +"No, I am descending." + +"I beg pardon." + +"My last client of the sort was a king." + +"Oh, really! I had no idea. And which king?" + +"The King of Scandinavia." + +"What! Had he lost his wife?" + +"You can understand," said Holmes suavely, "that I extend to the +affairs of my other clients the same secrecy which I promise to +you in yours." + +"Of course! Very right! very right! I'm sure I beg pardon. As to +my own case, I am ready to give you any information which may +assist you in forming an opinion." + +"Thank you. I have already learned all that is in the public +prints, nothing more. I presume that I may take it as correct--this +article, for example, as to the disappearance of the bride." + +Lord St. Simon glanced over it. "Yes, it is correct, as far as it +goes." + +"But it needs a great deal of supplementing before anyone could +offer an opinion. I think that I may arrive at my facts most +directly by questioning you." + +"Pray do so." + +"When did you first meet Miss Hatty Doran?" + +"In San Francisco, a year ago." + +"You were travelling in the States?" + +"Yes." + +"Did you become engaged then?" + +"No." + +"But you were on a friendly footing?" + +"I was amused by her society, and she could see that I was +amused." + +"Her father is very rich?" + +"He is said to be the richest man on the Pacific slope." + +"And how did he make his money?" + +"In mining. He had nothing a few years ago. Then he struck gold, +invested it, and came up by leaps and bounds." + +"Now, what is your own impression as to the young lady's--your +wife's character?" + +The nobleman swung his glasses a little faster and stared down +into the fire. "You see, Mr. Holmes," said he, "my wife was +twenty before her father became a rich man. During that time she +ran free in a mining camp and wandered through woods or +mountains, so that her education has come from Nature rather than +from the schoolmaster. She is what we call in England a tomboy, +with a strong nature, wild and free, unfettered by any sort of +traditions. She is impetuous--volcanic, I was about to say. She +is swift in making up her mind and fearless in carrying out her +resolutions. On the other hand, I would not have given her the +name which I have the honour to bear"--he gave a little stately +cough--"had not I thought her to be at bottom a noble woman. I +believe that she is capable of heroic self-sacrifice and that +anything dishonourable would be repugnant to her." + +"Have you her photograph?" + +"I brought this with me." He opened a locket and showed us the +full face of a very lovely woman. It was not a photograph but an +ivory miniature, and the artist had brought out the full effect +of the lustrous black hair, the large dark eyes, and the +exquisite mouth. Holmes gazed long and earnestly at it. Then he +closed the locket and handed it back to Lord St. Simon. + +"The young lady came to London, then, and you renewed your +acquaintance?" + +"Yes, her father brought her over for this last London season. I +met her several times, became engaged to her, and have now +married her." + +"She brought, I understand, a considerable dowry?" + +"A fair dowry. Not more than is usual in my family." + +"And this, of course, remains to you, since the marriage is a +fait accompli?" + +"I really have made no inquiries on the subject." + +"Very naturally not. Did you see Miss Doran on the day before the +wedding?" + +"Yes." + +"Was she in good spirits?" + +"Never better. She kept talking of what we should do in our +future lives." + +"Indeed! That is very interesting. And on the morning of the +wedding?" + +"She was as bright as possible--at least until after the +ceremony." + +"And did you observe any change in her then?" + +"Well, to tell the truth, I saw then the first signs that I had +ever seen that her temper was just a little sharp. The incident +however, was too trivial to relate and can have no possible +bearing upon the case." + +"Pray let us have it, for all that." + +"Oh, it is childish. She dropped her bouquet as we went towards +the vestry. She was passing the front pew at the time, and it +fell over into the pew. There was a moment's delay, but the +gentleman in the pew handed it up to her again, and it did not +appear to be the worse for the fall. Yet when I spoke to her of +the matter, she answered me abruptly; and in the carriage, on our +way home, she seemed absurdly agitated over this trifling cause." + +"Indeed! You say that there was a gentleman in the pew. Some of +the general public were present, then?" + +"Oh, yes. It is impossible to exclude them when the church is +open." + +"This gentleman was not one of your wife's friends?" + +"No, no; I call him a gentleman by courtesy, but he was quite a +common-looking person. I hardly noticed his appearance. But +really I think that we are wandering rather far from the point." + +"Lady St. Simon, then, returned from the wedding in a less +cheerful frame of mind than she had gone to it. What did she do +on re-entering her father's house?" + +"I saw her in conversation with her maid." + +"And who is her maid?" + +"Alice is her name. She is an American and came from California +with her." + +"A confidential servant?" + +"A little too much so. It seemed to me that her mistress allowed +her to take great liberties. Still, of course, in America they +look upon these things in a different way." + +"How long did she speak to this Alice?" + +"Oh, a few minutes. I had something else to think of." + +"You did not overhear what they said?" + +"Lady St. Simon said something about 'jumping a claim.' She was +accustomed to use slang of the kind. I have no idea what she +meant." + +"American slang is very expressive sometimes. And what did your +wife do when she finished speaking to her maid?" + +"She walked into the breakfast-room." + +"On your arm?" + +"No, alone. She was very independent in little matters like that. +Then, after we had sat down for ten minutes or so, she rose +hurriedly, muttered some words of apology, and left the room. She +never came back." + +"But this maid, Alice, as I understand, deposes that she went to +her room, covered her bride's dress with a long ulster, put on a +bonnet, and went out." + +"Quite so. And she was afterwards seen walking into Hyde Park in +company with Flora Millar, a woman who is now in custody, and who +had already made a disturbance at Mr. Doran's house that +morning." + +"Ah, yes. I should like a few particulars as to this young lady, +and your relations to her." + +Lord St. Simon shrugged his shoulders and raised his eyebrows. +"We have been on a friendly footing for some years--I may say on +a very friendly footing. She used to be at the Allegro. I have +not treated her ungenerously, and she had no just cause of +complaint against me, but you know what women are, Mr. Holmes. +Flora was a dear little thing, but exceedingly hot-headed and +devotedly attached to me. She wrote me dreadful letters when she +heard that I was about to be married, and, to tell the truth, the +reason why I had the marriage celebrated so quietly was that I +feared lest there might be a scandal in the church. She came to +Mr. Doran's door just after we returned, and she endeavoured to +push her way in, uttering very abusive expressions towards my +wife, and even threatening her, but I had foreseen the +possibility of something of the sort, and I had two police +fellows there in private clothes, who soon pushed her out again. +She was quiet when she saw that there was no good in making a +row." + +"Did your wife hear all this?" + +"No, thank goodness, she did not." + +"And she was seen walking with this very woman afterwards?" + +"Yes. That is what Mr. Lestrade, of Scotland Yard, looks upon as +so serious. It is thought that Flora decoyed my wife out and laid +some terrible trap for her." + +"Well, it is a possible supposition." + +"You think so, too?" + +"I did not say a probable one. But you do not yourself look upon +this as likely?" + +"I do not think Flora would hurt a fly." + +"Still, jealousy is a strange transformer of characters. Pray +what is your own theory as to what took place?" + +"Well, really, I came to seek a theory, not to propound one. I +have given you all the facts. Since you ask me, however, I may +say that it has occurred to me as possible that the excitement of +this affair, the consciousness that she had made so immense a +social stride, had the effect of causing some little nervous +disturbance in my wife." + +"In short, that she had become suddenly deranged?" + +"Well, really, when I consider that she has turned her back--I +will not say upon me, but upon so much that many have aspired to +without success--I can hardly explain it in any other fashion." + +"Well, certainly that is also a conceivable hypothesis," said +Holmes, smiling. "And now, Lord St. Simon, I think that I have +nearly all my data. May I ask whether you were seated at the +breakfast-table so that you could see out of the window?" + +"We could see the other side of the road and the Park." + +"Quite so. Then I do not think that I need to detain you longer. +I shall communicate with you." + +"Should you be fortunate enough to solve this problem," said our +client, rising. + +"I have solved it." + +"Eh? What was that?" + +"I say that I have solved it." + +"Where, then, is my wife?" + +"That is a detail which I shall speedily supply." + +Lord St. Simon shook his head. "I am afraid that it will take +wiser heads than yours or mine," he remarked, and bowing in a +stately, old-fashioned manner he departed. + +"It is very good of Lord St. Simon to honour my head by putting +it on a level with his own," said Sherlock Holmes, laughing. "I +think that I shall have a whisky and soda and a cigar after all +this cross-questioning. I had formed my conclusions as to the +case before our client came into the room." + +"My dear Holmes!" + +"I have notes of several similar cases, though none, as I +remarked before, which were quite as prompt. My whole examination +served to turn my conjecture into a certainty. Circumstantial +evidence is occasionally very convincing, as when you find a +trout in the milk, to quote Thoreau's example." + +"But I have heard all that you have heard." + +"Without, however, the knowledge of pre-existing cases which +serves me so well. There was a parallel instance in Aberdeen some +years back, and something on very much the same lines at Munich +the year after the Franco-Prussian War. It is one of these +cases--but, hullo, here is Lestrade! Good-afternoon, Lestrade! +You will find an extra tumbler upon the sideboard, and there are +cigars in the box." + +The official detective was attired in a pea-jacket and cravat, +which gave him a decidedly nautical appearance, and he carried a +black canvas bag in his hand. With a short greeting he seated +himself and lit the cigar which had been offered to him. + +"What's up, then?" asked Holmes with a twinkle in his eye. "You +look dissatisfied." + +"And I feel dissatisfied. It is this infernal St. Simon marriage +case. I can make neither head nor tail of the business." + +"Really! You surprise me." + +"Who ever heard of such a mixed affair? Every clue seems to slip +through my fingers. I have been at work upon it all day." + +"And very wet it seems to have made you," said Holmes laying his +hand upon the arm of the pea-jacket. + +"Yes, I have been dragging the Serpentine." + +"In heaven's name, what for?" + +"In search of the body of Lady St. Simon." + +Sherlock Holmes leaned back in his chair and laughed heartily. + +"Have you dragged the basin of Trafalgar Square fountain?" he +asked. + +"Why? What do you mean?" + +"Because you have just as good a chance of finding this lady in +the one as in the other." + +Lestrade shot an angry glance at my companion. "I suppose you +know all about it," he snarled. + +"Well, I have only just heard the facts, but my mind is made up." + +"Oh, indeed! Then you think that the Serpentine plays no part in +the matter?" + +"I think it very unlikely." + +"Then perhaps you will kindly explain how it is that we found +this in it?" He opened his bag as he spoke, and tumbled onto the +floor a wedding-dress of watered silk, a pair of white satin +shoes and a bride's wreath and veil, all discoloured and soaked +in water. "There," said he, putting a new wedding-ring upon the +top of the pile. "There is a little nut for you to crack, Master +Holmes." + +"Oh, indeed!" said my friend, blowing blue rings into the air. +"You dragged them from the Serpentine?" + +"No. They were found floating near the margin by a park-keeper. +They have been identified as her clothes, and it seemed to me +that if the clothes were there the body would not be far off." + +"By the same brilliant reasoning, every man's body is to be found +in the neighbourhood of his wardrobe. And pray what did you hope +to arrive at through this?" + +"At some evidence implicating Flora Millar in the disappearance." + +"I am afraid that you will find it difficult." + +"Are you, indeed, now?" cried Lestrade with some bitterness. "I +am afraid, Holmes, that you are not very practical with your +deductions and your inferences. You have made two blunders in as +many minutes. This dress does implicate Miss Flora Millar." + +"And how?" + +"In the dress is a pocket. In the pocket is a card-case. In the +card-case is a note. And here is the very note." He slapped it +down upon the table in front of him. "Listen to this: 'You will +see me when all is ready. Come at once. F.H.M.' Now my theory all +along has been that Lady St. Simon was decoyed away by Flora +Millar, and that she, with confederates, no doubt, was +responsible for her disappearance. Here, signed with her +initials, is the very note which was no doubt quietly slipped +into her hand at the door and which lured her within their +reach." + +"Very good, Lestrade," said Holmes, laughing. "You really are +very fine indeed. Let me see it." He took up the paper in a +listless way, but his attention instantly became riveted, and he +gave a little cry of satisfaction. "This is indeed important," +said he. + +"Ha! you find it so?" + +"Extremely so. I congratulate you warmly." + +Lestrade rose in his triumph and bent his head to look. "Why," he +shrieked, "you're looking at the wrong side!" + +"On the contrary, this is the right side." + +"The right side? You're mad! Here is the note written in pencil +over here." + +"And over here is what appears to be the fragment of a hotel +bill, which interests me deeply." + +"There's nothing in it. I looked at it before," said Lestrade. +"'Oct. 4th, rooms 8s., breakfast 2s. 6d., cocktail 1s., lunch 2s. +6d., glass sherry, 8d.' I see nothing in that." + +"Very likely not. It is most important, all the same. As to the +note, it is important also, or at least the initials are, so I +congratulate you again." + +"I've wasted time enough," said Lestrade, rising. "I believe in +hard work and not in sitting by the fire spinning fine theories. +Good-day, Mr. Holmes, and we shall see which gets to the bottom +of the matter first." He gathered up the garments, thrust them +into the bag, and made for the door. + +"Just one hint to you, Lestrade," drawled Holmes before his rival +vanished; "I will tell you the true solution of the matter. Lady +St. Simon is a myth. There is not, and there never has been, any +such person." + +Lestrade looked sadly at my companion. Then he turned to me, +tapped his forehead three times, shook his head solemnly, and +hurried away. + +He had hardly shut the door behind him when Holmes rose to put on +his overcoat. "There is something in what the fellow says about +outdoor work," he remarked, "so I think, Watson, that I must +leave you to your papers for a little." + +It was after five o'clock when Sherlock Holmes left me, but I had +no time to be lonely, for within an hour there arrived a +confectioner's man with a very large flat box. This he unpacked +with the help of a youth whom he had brought with him, and +presently, to my very great astonishment, a quite epicurean +little cold supper began to be laid out upon our humble +lodging-house mahogany. There were a couple of brace of cold +woodcock, a pheasant, a pâté de foie gras pie with a group of +ancient and cobwebby bottles. Having laid out all these luxuries, +my two visitors vanished away, like the genii of the Arabian +Nights, with no explanation save that the things had been paid +for and were ordered to this address. + +Just before nine o'clock Sherlock Holmes stepped briskly into the +room. His features were gravely set, but there was a light in his +eye which made me think that he had not been disappointed in his +conclusions. + +"They have laid the supper, then," he said, rubbing his hands. + +"You seem to expect company. They have laid for five." + +"Yes, I fancy we may have some company dropping in," said he. "I +am surprised that Lord St. Simon has not already arrived. Ha! I +fancy that I hear his step now upon the stairs." + +It was indeed our visitor of the afternoon who came bustling in, +dangling his glasses more vigorously than ever, and with a very +perturbed expression upon his aristocratic features. + +"My messenger reached you, then?" asked Holmes. + +"Yes, and I confess that the contents startled me beyond measure. +Have you good authority for what you say?" + +"The best possible." + +Lord St. Simon sank into a chair and passed his hand over his +forehead. + +"What will the Duke say," he murmured, "when he hears that one of +the family has been subjected to such humiliation?" + +"It is the purest accident. I cannot allow that there is any +humiliation." + +"Ah, you look on these things from another standpoint." + +"I fail to see that anyone is to blame. I can hardly see how the +lady could have acted otherwise, though her abrupt method of +doing it was undoubtedly to be regretted. Having no mother, she +had no one to advise her at such a crisis." + +"It was a slight, sir, a public slight," said Lord St. Simon, +tapping his fingers upon the table. + +"You must make allowance for this poor girl, placed in so +unprecedented a position." + +"I will make no allowance. I am very angry indeed, and I have +been shamefully used." + +"I think that I heard a ring," said Holmes. "Yes, there are steps +on the landing. If I cannot persuade you to take a lenient view +of the matter, Lord St. Simon, I have brought an advocate here +who may be more successful." He opened the door and ushered in a +lady and gentleman. "Lord St. Simon," said he "allow me to +introduce you to Mr. and Mrs. Francis Hay Moulton. The lady, I +think, you have already met." + +At the sight of these newcomers our client had sprung from his +seat and stood very erect, with his eyes cast down and his hand +thrust into the breast of his frock-coat, a picture of offended +dignity. The lady had taken a quick step forward and had held out +her hand to him, but he still refused to raise his eyes. It was +as well for his resolution, perhaps, for her pleading face was +one which it was hard to resist. + +"You're angry, Robert," said she. "Well, I guess you have every +cause to be." + +"Pray make no apology to me," said Lord St. Simon bitterly. + +"Oh, yes, I know that I have treated you real bad and that I +should have spoken to you before I went; but I was kind of +rattled, and from the time when I saw Frank here again I just +didn't know what I was doing or saying. I only wonder I didn't +fall down and do a faint right there before the altar." + +"Perhaps, Mrs. Moulton, you would like my friend and me to leave +the room while you explain this matter?" + +"If I may give an opinion," remarked the strange gentleman, +"we've had just a little too much secrecy over this business +already. For my part, I should like all Europe and America to +hear the rights of it." He was a small, wiry, sunburnt man, +clean-shaven, with a sharp face and alert manner. + +"Then I'll tell our story right away," said the lady. "Frank here +and I met in '84, in McQuire's camp, near the Rockies, where pa +was working a claim. We were engaged to each other, Frank and I; +but then one day father struck a rich pocket and made a pile, +while poor Frank here had a claim that petered out and came to +nothing. The richer pa grew the poorer was Frank; so at last pa +wouldn't hear of our engagement lasting any longer, and he took +me away to 'Frisco. Frank wouldn't throw up his hand, though; so +he followed me there, and he saw me without pa knowing anything +about it. It would only have made him mad to know, so we just +fixed it all up for ourselves. Frank said that he would go and +make his pile, too, and never come back to claim me until he had +as much as pa. So then I promised to wait for him to the end of +time and pledged myself not to marry anyone else while he lived. +'Why shouldn't we be married right away, then,' said he, 'and +then I will feel sure of you; and I won't claim to be your +husband until I come back?' Well, we talked it over, and he had +fixed it all up so nicely, with a clergyman all ready in waiting, +that we just did it right there; and then Frank went off to seek +his fortune, and I went back to pa. + +"The next I heard of Frank was that he was in Montana, and then +he went prospecting in Arizona, and then I heard of him from New +Mexico. After that came a long newspaper story about how a +miners' camp had been attacked by Apache Indians, and there was +my Frank's name among the killed. I fainted dead away, and I was +very sick for months after. Pa thought I had a decline and took +me to half the doctors in 'Frisco. Not a word of news came for a +year and more, so that I never doubted that Frank was really +dead. Then Lord St. Simon came to 'Frisco, and we came to London, +and a marriage was arranged, and pa was very pleased, but I felt +all the time that no man on this earth would ever take the place +in my heart that had been given to my poor Frank. + +"Still, if I had married Lord St. Simon, of course I'd have done +my duty by him. We can't command our love, but we can our +actions. I went to the altar with him with the intention to make +him just as good a wife as it was in me to be. But you may +imagine what I felt when, just as I came to the altar rails, I +glanced back and saw Frank standing and looking at me out of the +first pew. I thought it was his ghost at first; but when I looked +again there he was still, with a kind of question in his eyes, as +if to ask me whether I were glad or sorry to see him. I wonder I +didn't drop. I know that everything was turning round, and the +words of the clergyman were just like the buzz of a bee in my +ear. I didn't know what to do. Should I stop the service and make +a scene in the church? I glanced at him again, and he seemed to +know what I was thinking, for he raised his finger to his lips to +tell me to be still. Then I saw him scribble on a piece of paper, +and I knew that he was writing me a note. As I passed his pew on +the way out I dropped my bouquet over to him, and he slipped the +note into my hand when he returned me the flowers. It was only a +line asking me to join him when he made the sign to me to do so. +Of course I never doubted for a moment that my first duty was now +to him, and I determined to do just whatever he might direct. + +"When I got back I told my maid, who had known him in California, +and had always been his friend. I ordered her to say nothing, but +to get a few things packed and my ulster ready. I know I ought to +have spoken to Lord St. Simon, but it was dreadful hard before +his mother and all those great people. I just made up my mind to +run away and explain afterwards. I hadn't been at the table ten +minutes before I saw Frank out of the window at the other side of +the road. He beckoned to me and then began walking into the Park. +I slipped out, put on my things, and followed him. Some woman +came talking something or other about Lord St. Simon to +me--seemed to me from the little I heard as if he had a little +secret of his own before marriage also--but I managed to get away +from her and soon overtook Frank. We got into a cab together, and +away we drove to some lodgings he had taken in Gordon Square, and +that was my true wedding after all those years of waiting. Frank +had been a prisoner among the Apaches, had escaped, came on to +'Frisco, found that I had given him up for dead and had gone to +England, followed me there, and had come upon me at last on the +very morning of my second wedding." + +"I saw it in a paper," explained the American. "It gave the name +and the church but not where the lady lived." + +"Then we had a talk as to what we should do, and Frank was all +for openness, but I was so ashamed of it all that I felt as if I +should like to vanish away and never see any of them again--just +sending a line to pa, perhaps, to show him that I was alive. It +was awful to me to think of all those lords and ladies sitting +round that breakfast-table and waiting for me to come back. So +Frank took my wedding-clothes and things and made a bundle of +them, so that I should not be traced, and dropped them away +somewhere where no one could find them. It is likely that we +should have gone on to Paris to-morrow, only that this good +gentleman, Mr. Holmes, came round to us this evening, though how +he found us is more than I can think, and he showed us very +clearly and kindly that I was wrong and that Frank was right, and +that we should be putting ourselves in the wrong if we were so +secret. Then he offered to give us a chance of talking to Lord +St. Simon alone, and so we came right away round to his rooms at +once. Now, Robert, you have heard it all, and I am very sorry if +I have given you pain, and I hope that you do not think very +meanly of me." + +Lord St. Simon had by no means relaxed his rigid attitude, but +had listened with a frowning brow and a compressed lip to this +long narrative. + +"Excuse me," he said, "but it is not my custom to discuss my most +intimate personal affairs in this public manner." + +"Then you won't forgive me? You won't shake hands before I go?" + +"Oh, certainly, if it would give you any pleasure." He put out +his hand and coldly grasped that which she extended to him. + +"I had hoped," suggested Holmes, "that you would have joined us +in a friendly supper." + +"I think that there you ask a little too much," responded his +Lordship. "I may be forced to acquiesce in these recent +developments, but I can hardly be expected to make merry over +them. I think that with your permission I will now wish you all a +very good-night." He included us all in a sweeping bow and +stalked out of the room. + +"Then I trust that you at least will honour me with your +company," said Sherlock Holmes. "It is always a joy to meet an +American, Mr. Moulton, for I am one of those who believe that the +folly of a monarch and the blundering of a minister in far-gone +years will not prevent our children from being some day citizens +of the same world-wide country under a flag which shall be a +quartering of the Union Jack with the Stars and Stripes." + +"The case has been an interesting one," remarked Holmes when our +visitors had left us, "because it serves to show very clearly how +simple the explanation may be of an affair which at first sight +seems to be almost inexplicable. Nothing could be more natural +than the sequence of events as narrated by this lady, and nothing +stranger than the result when viewed, for instance, by Mr. +Lestrade of Scotland Yard." + +"You were not yourself at fault at all, then?" + +"From the first, two facts were very obvious to me, the one that +the lady had been quite willing to undergo the wedding ceremony, +the other that she had repented of it within a few minutes of +returning home. Obviously something had occurred during the +morning, then, to cause her to change her mind. What could that +something be? She could not have spoken to anyone when she was +out, for she had been in the company of the bridegroom. Had she +seen someone, then? If she had, it must be someone from America +because she had spent so short a time in this country that she +could hardly have allowed anyone to acquire so deep an influence +over her that the mere sight of him would induce her to change +her plans so completely. You see we have already arrived, by a +process of exclusion, at the idea that she might have seen an +American. Then who could this American be, and why should he +possess so much influence over her? It might be a lover; it might +be a husband. Her young womanhood had, I knew, been spent in +rough scenes and under strange conditions. So far I had got +before I ever heard Lord St. Simon's narrative. When he told us +of a man in a pew, of the change in the bride's manner, of so +transparent a device for obtaining a note as the dropping of a +bouquet, of her resort to her confidential maid, and of her very +significant allusion to claim-jumping--which in miners' parlance +means taking possession of that which another person has a prior +claim to--the whole situation became absolutely clear. She had +gone off with a man, and the man was either a lover or was a +previous husband--the chances being in favour of the latter." + +"And how in the world did you find them?" + +"It might have been difficult, but friend Lestrade held +information in his hands the value of which he did not himself +know. The initials were, of course, of the highest importance, +but more valuable still was it to know that within a week he had +settled his bill at one of the most select London hotels." + +"How did you deduce the select?" + +"By the select prices. Eight shillings for a bed and eightpence +for a glass of sherry pointed to one of the most expensive +hotels. There are not many in London which charge at that rate. +In the second one which I visited in Northumberland Avenue, I +learned by an inspection of the book that Francis H. Moulton, an +American gentleman, had left only the day before, and on looking +over the entries against him, I came upon the very items which I +had seen in the duplicate bill. His letters were to be forwarded +to 226 Gordon Square; so thither I travelled, and being fortunate +enough to find the loving couple at home, I ventured to give them +some paternal advice and to point out to them that it would be +better in every way that they should make their position a little +clearer both to the general public and to Lord St. Simon in +particular. I invited them to meet him here, and, as you see, I +made him keep the appointment." + +"But with no very good result," I remarked. "His conduct was +certainly not very gracious." + +"Ah, Watson," said Holmes, smiling, "perhaps you would not be +very gracious either, if, after all the trouble of wooing and +wedding, you found yourself deprived in an instant of wife and of +fortune. I think that we may judge Lord St. Simon very mercifully +and thank our stars that we are never likely to find ourselves in +the same position. Draw your chair up and hand me my violin, for +the only problem we have still to solve is how to while away +these bleak autumnal evenings." + + + +XI. THE ADVENTURE OF THE BERYL CORONET + +"Holmes," said I as I stood one morning in our bow-window looking +down the street, "here is a madman coming along. It seems rather +sad that his relatives should allow him to come out alone." + +My friend rose lazily from his armchair and stood with his hands +in the pockets of his dressing-gown, looking over my shoulder. It +was a bright, crisp February morning, and the snow of the day +before still lay deep upon the ground, shimmering brightly in the +wintry sun. Down the centre of Baker Street it had been ploughed +into a brown crumbly band by the traffic, but at either side and +on the heaped-up edges of the foot-paths it still lay as white as +when it fell. The grey pavement had been cleaned and scraped, but +was still dangerously slippery, so that there were fewer +passengers than usual. Indeed, from the direction of the +Metropolitan Station no one was coming save the single gentleman +whose eccentric conduct had drawn my attention. + +He was a man of about fifty, tall, portly, and imposing, with a +massive, strongly marked face and a commanding figure. He was +dressed in a sombre yet rich style, in black frock-coat, shining +hat, neat brown gaiters, and well-cut pearl-grey trousers. Yet +his actions were in absurd contrast to the dignity of his dress +and features, for he was running hard, with occasional little +springs, such as a weary man gives who is little accustomed to +set any tax upon his legs. As he ran he jerked his hands up and +down, waggled his head, and writhed his face into the most +extraordinary contortions. + +"What on earth can be the matter with him?" I asked. "He is +looking up at the numbers of the houses." + +"I believe that he is coming here," said Holmes, rubbing his +hands. + +"Here?" + +"Yes; I rather think he is coming to consult me professionally. I +think that I recognise the symptoms. Ha! did I not tell you?" As +he spoke, the man, puffing and blowing, rushed at our door and +pulled at our bell until the whole house resounded with the +clanging. + +A few moments later he was in our room, still puffing, still +gesticulating, but with so fixed a look of grief and despair in +his eyes that our smiles were turned in an instant to horror and +pity. For a while he could not get his words out, but swayed his +body and plucked at his hair like one who has been driven to the +extreme limits of his reason. Then, suddenly springing to his +feet, he beat his head against the wall with such force that we +both rushed upon him and tore him away to the centre of the room. +Sherlock Holmes pushed him down into the easy-chair and, sitting +beside him, patted his hand and chatted with him in the easy, +soothing tones which he knew so well how to employ. + +"You have come to me to tell your story, have you not?" said he. +"You are fatigued with your haste. Pray wait until you have +recovered yourself, and then I shall be most happy to look into +any little problem which you may submit to me." + +The man sat for a minute or more with a heaving chest, fighting +against his emotion. Then he passed his handkerchief over his +brow, set his lips tight, and turned his face towards us. + +"No doubt you think me mad?" said he. + +"I see that you have had some great trouble," responded Holmes. + +"God knows I have!--a trouble which is enough to unseat my +reason, so sudden and so terrible is it. Public disgrace I might +have faced, although I am a man whose character has never yet +borne a stain. Private affliction also is the lot of every man; +but the two coming together, and in so frightful a form, have +been enough to shake my very soul. Besides, it is not I alone. +The very noblest in the land may suffer unless some way be found +out of this horrible affair." + +"Pray compose yourself, sir," said Holmes, "and let me have a +clear account of who you are and what it is that has befallen +you." + +"My name," answered our visitor, "is probably familiar to your +ears. I am Alexander Holder, of the banking firm of Holder & +Stevenson, of Threadneedle Street." + +The name was indeed well known to us as belonging to the senior +partner in the second largest private banking concern in the City +of London. What could have happened, then, to bring one of the +foremost citizens of London to this most pitiable pass? We +waited, all curiosity, until with another effort he braced +himself to tell his story. + +"I feel that time is of value," said he; "that is why I hastened +here when the police inspector suggested that I should secure +your co-operation. I came to Baker Street by the Underground and +hurried from there on foot, for the cabs go slowly through this +snow. That is why I was so out of breath, for I am a man who +takes very little exercise. I feel better now, and I will put the +facts before you as shortly and yet as clearly as I can. + +"It is, of course, well known to you that in a successful banking +business as much depends upon our being able to find remunerative +investments for our funds as upon our increasing our connection +and the number of our depositors. One of our most lucrative means +of laying out money is in the shape of loans, where the security +is unimpeachable. We have done a good deal in this direction +during the last few years, and there are many noble families to +whom we have advanced large sums upon the security of their +pictures, libraries, or plate. + +"Yesterday morning I was seated in my office at the bank when a +card was brought in to me by one of the clerks. I started when I +saw the name, for it was that of none other than--well, perhaps +even to you I had better say no more than that it was a name +which is a household word all over the earth--one of the highest, +noblest, most exalted names in England. I was overwhelmed by the +honour and attempted, when he entered, to say so, but he plunged +at once into business with the air of a man who wishes to hurry +quickly through a disagreeable task. + +"'Mr. Holder,' said he, 'I have been informed that you are in the +habit of advancing money.' + +"'The firm does so when the security is good.' I answered. + +"'It is absolutely essential to me,' said he, 'that I should have +50,000 pounds at once. I could, of course, borrow so trifling a +sum ten times over from my friends, but I much prefer to make it +a matter of business and to carry out that business myself. In my +position you can readily understand that it is unwise to place +one's self under obligations.' + +"'For how long, may I ask, do you want this sum?' I asked. + +"'Next Monday I have a large sum due to me, and I shall then most +certainly repay what you advance, with whatever interest you +think it right to charge. But it is very essential to me that the +money should be paid at once.' + +"'I should be happy to advance it without further parley from my +own private purse,' said I, 'were it not that the strain would be +rather more than it could bear. If, on the other hand, I am to do +it in the name of the firm, then in justice to my partner I must +insist that, even in your case, every businesslike precaution +should be taken.' + +"'I should much prefer to have it so,' said he, raising up a +square, black morocco case which he had laid beside his chair. +'You have doubtless heard of the Beryl Coronet?' + +"'One of the most precious public possessions of the empire,' +said I. + +"'Precisely.' He opened the case, and there, imbedded in soft, +flesh-coloured velvet, lay the magnificent piece of jewellery +which he had named. 'There are thirty-nine enormous beryls,' said +he, 'and the price of the gold chasing is incalculable. The +lowest estimate would put the worth of the coronet at double the +sum which I have asked. I am prepared to leave it with you as my +security.' + +"I took the precious case into my hands and looked in some +perplexity from it to my illustrious client. + +"'You doubt its value?' he asked. + +"'Not at all. I only doubt--' + +"'The propriety of my leaving it. You may set your mind at rest +about that. I should not dream of doing so were it not absolutely +certain that I should be able in four days to reclaim it. It is a +pure matter of form. Is the security sufficient?' + +"'Ample.' + +"'You understand, Mr. Holder, that I am giving you a strong proof +of the confidence which I have in you, founded upon all that I +have heard of you. I rely upon you not only to be discreet and to +refrain from all gossip upon the matter but, above all, to +preserve this coronet with every possible precaution because I +need not say that a great public scandal would be caused if any +harm were to befall it. Any injury to it would be almost as +serious as its complete loss, for there are no beryls in the +world to match these, and it would be impossible to replace them. +I leave it with you, however, with every confidence, and I shall +call for it in person on Monday morning.' + +"Seeing that my client was anxious to leave, I said no more but, +calling for my cashier, I ordered him to pay over fifty 1000 +pound notes. When I was alone once more, however, with the +precious case lying upon the table in front of me, I could not +but think with some misgivings of the immense responsibility +which it entailed upon me. There could be no doubt that, as it +was a national possession, a horrible scandal would ensue if any +misfortune should occur to it. I already regretted having ever +consented to take charge of it. However, it was too late to alter +the matter now, so I locked it up in my private safe and turned +once more to my work. + +"When evening came I felt that it would be an imprudence to leave +so precious a thing in the office behind me. Bankers' safes had +been forced before now, and why should not mine be? If so, how +terrible would be the position in which I should find myself! I +determined, therefore, that for the next few days I would always +carry the case backward and forward with me, so that it might +never be really out of my reach. With this intention, I called a +cab and drove out to my house at Streatham, carrying the jewel +with me. I did not breathe freely until I had taken it upstairs +and locked it in the bureau of my dressing-room. + +"And now a word as to my household, Mr. Holmes, for I wish you to +thoroughly understand the situation. My groom and my page sleep +out of the house, and may be set aside altogether. I have three +maid-servants who have been with me a number of years and whose +absolute reliability is quite above suspicion. Another, Lucy +Parr, the second waiting-maid, has only been in my service a few +months. She came with an excellent character, however, and has +always given me satisfaction. She is a very pretty girl and has +attracted admirers who have occasionally hung about the place. +That is the only drawback which we have found to her, but we +believe her to be a thoroughly good girl in every way. + +"So much for the servants. My family itself is so small that it +will not take me long to describe it. I am a widower and have an +only son, Arthur. He has been a disappointment to me, Mr. +Holmes--a grievous disappointment. I have no doubt that I am +myself to blame. People tell me that I have spoiled him. Very +likely I have. When my dear wife died I felt that he was all I +had to love. I could not bear to see the smile fade even for a +moment from his face. I have never denied him a wish. Perhaps it +would have been better for both of us had I been sterner, but I +meant it for the best. + +"It was naturally my intention that he should succeed me in my +business, but he was not of a business turn. He was wild, +wayward, and, to speak the truth, I could not trust him in the +handling of large sums of money. When he was young he became a +member of an aristocratic club, and there, having charming +manners, he was soon the intimate of a number of men with long +purses and expensive habits. He learned to play heavily at cards +and to squander money on the turf, until he had again and again +to come to me and implore me to give him an advance upon his +allowance, that he might settle his debts of honour. He tried +more than once to break away from the dangerous company which he +was keeping, but each time the influence of his friend, Sir +George Burnwell, was enough to draw him back again. + +"And, indeed, I could not wonder that such a man as Sir George +Burnwell should gain an influence over him, for he has frequently +brought him to my house, and I have found myself that I could +hardly resist the fascination of his manner. He is older than +Arthur, a man of the world to his finger-tips, one who had been +everywhere, seen everything, a brilliant talker, and a man of +great personal beauty. Yet when I think of him in cold blood, far +away from the glamour of his presence, I am convinced from his +cynical speech and the look which I have caught in his eyes that +he is one who should be deeply distrusted. So I think, and so, +too, thinks my little Mary, who has a woman's quick insight into +character. + +"And now there is only she to be described. She is my niece; but +when my brother died five years ago and left her alone in the +world I adopted her, and have looked upon her ever since as my +daughter. She is a sunbeam in my house--sweet, loving, beautiful, +a wonderful manager and housekeeper, yet as tender and quiet and +gentle as a woman could be. She is my right hand. I do not know +what I could do without her. In only one matter has she ever gone +against my wishes. Twice my boy has asked her to marry him, for +he loves her devotedly, but each time she has refused him. I +think that if anyone could have drawn him into the right path it +would have been she, and that his marriage might have changed his +whole life; but now, alas! it is too late--forever too late! + +"Now, Mr. Holmes, you know the people who live under my roof, and +I shall continue with my miserable story. + +"When we were taking coffee in the drawing-room that night after +dinner, I told Arthur and Mary my experience, and of the precious +treasure which we had under our roof, suppressing only the name +of my client. Lucy Parr, who had brought in the coffee, had, I am +sure, left the room; but I cannot swear that the door was closed. +Mary and Arthur were much interested and wished to see the famous +coronet, but I thought it better not to disturb it. + +"'Where have you put it?' asked Arthur. + +"'In my own bureau.' + +"'Well, I hope to goodness the house won't be burgled during the +night.' said he. + +"'It is locked up,' I answered. + +"'Oh, any old key will fit that bureau. When I was a youngster I +have opened it myself with the key of the box-room cupboard.' + +"He often had a wild way of talking, so that I thought little of +what he said. He followed me to my room, however, that night with +a very grave face. + +"'Look here, dad,' said he with his eyes cast down, 'can you let +me have 200 pounds?' + +"'No, I cannot!' I answered sharply. 'I have been far too +generous with you in money matters.' + +"'You have been very kind,' said he, 'but I must have this money, +or else I can never show my face inside the club again.' + +"'And a very good thing, too!' I cried. + +"'Yes, but you would not have me leave it a dishonoured man,' +said he. 'I could not bear the disgrace. I must raise the money +in some way, and if you will not let me have it, then I must try +other means.' + +"I was very angry, for this was the third demand during the +month. 'You shall not have a farthing from me,' I cried, on which +he bowed and left the room without another word. + +"When he was gone I unlocked my bureau, made sure that my +treasure was safe, and locked it again. Then I started to go +round the house to see that all was secure--a duty which I +usually leave to Mary but which I thought it well to perform +myself that night. As I came down the stairs I saw Mary herself +at the side window of the hall, which she closed and fastened as +I approached. + +"'Tell me, dad,' said she, looking, I thought, a little +disturbed, 'did you give Lucy, the maid, leave to go out +to-night?' + +"'Certainly not.' + +"'She came in just now by the back door. I have no doubt that she +has only been to the side gate to see someone, but I think that +it is hardly safe and should be stopped.' + +"'You must speak to her in the morning, or I will if you prefer +it. Are you sure that everything is fastened?' + +"'Quite sure, dad.' + +"'Then, good-night.' I kissed her and went up to my bedroom +again, where I was soon asleep. + +"I am endeavouring to tell you everything, Mr. Holmes, which may +have any bearing upon the case, but I beg that you will question +me upon any point which I do not make clear." + +"On the contrary, your statement is singularly lucid." + +"I come to a part of my story now in which I should wish to be +particularly so. I am not a very heavy sleeper, and the anxiety +in my mind tended, no doubt, to make me even less so than usual. +About two in the morning, then, I was awakened by some sound in +the house. It had ceased ere I was wide awake, but it had left an +impression behind it as though a window had gently closed +somewhere. I lay listening with all my ears. Suddenly, to my +horror, there was a distinct sound of footsteps moving softly in +the next room. I slipped out of bed, all palpitating with fear, +and peeped round the corner of my dressing-room door. + +"'Arthur!' I screamed, 'you villain! you thief! How dare you +touch that coronet?' + +"The gas was half up, as I had left it, and my unhappy boy, +dressed only in his shirt and trousers, was standing beside the +light, holding the coronet in his hands. He appeared to be +wrenching at it, or bending it with all his strength. At my cry +he dropped it from his grasp and turned as pale as death. I +snatched it up and examined it. One of the gold corners, with +three of the beryls in it, was missing. + +"'You blackguard!' I shouted, beside myself with rage. 'You have +destroyed it! You have dishonoured me forever! Where are the +jewels which you have stolen?' + +"'Stolen!' he cried. + +"'Yes, thief!' I roared, shaking him by the shoulder. + +"'There are none missing. There cannot be any missing,' said he. + +"'There are three missing. And you know where they are. Must I +call you a liar as well as a thief? Did I not see you trying to +tear off another piece?' + +"'You have called me names enough,' said he, 'I will not stand it +any longer. I shall not say another word about this business, +since you have chosen to insult me. I will leave your house in +the morning and make my own way in the world.' + +"'You shall leave it in the hands of the police!' I cried +half-mad with grief and rage. 'I shall have this matter probed to +the bottom.' + +"'You shall learn nothing from me,' said he with a passion such +as I should not have thought was in his nature. 'If you choose to +call the police, let the police find what they can.' + +"By this time the whole house was astir, for I had raised my +voice in my anger. Mary was the first to rush into my room, and, +at the sight of the coronet and of Arthur's face, she read the +whole story and, with a scream, fell down senseless on the +ground. I sent the house-maid for the police and put the +investigation into their hands at once. When the inspector and a +constable entered the house, Arthur, who had stood sullenly with +his arms folded, asked me whether it was my intention to charge +him with theft. I answered that it had ceased to be a private +matter, but had become a public one, since the ruined coronet was +national property. I was determined that the law should have its +way in everything. + +"'At least,' said he, 'you will not have me arrested at once. It +would be to your advantage as well as mine if I might leave the +house for five minutes.' + +"'That you may get away, or perhaps that you may conceal what you +have stolen,' said I. And then, realising the dreadful position +in which I was placed, I implored him to remember that not only +my honour but that of one who was far greater than I was at +stake; and that he threatened to raise a scandal which would +convulse the nation. He might avert it all if he would but tell +me what he had done with the three missing stones. + +"'You may as well face the matter,' said I; 'you have been caught +in the act, and no confession could make your guilt more heinous. +If you but make such reparation as is in your power, by telling +us where the beryls are, all shall be forgiven and forgotten.' + +"'Keep your forgiveness for those who ask for it,' he answered, +turning away from me with a sneer. I saw that he was too hardened +for any words of mine to influence him. There was but one way for +it. I called in the inspector and gave him into custody. A search +was made at once not only of his person but of his room and of +every portion of the house where he could possibly have concealed +the gems; but no trace of them could be found, nor would the +wretched boy open his mouth for all our persuasions and our +threats. This morning he was removed to a cell, and I, after +going through all the police formalities, have hurried round to +you to implore you to use your skill in unravelling the matter. +The police have openly confessed that they can at present make +nothing of it. You may go to any expense which you think +necessary. I have already offered a reward of 1000 pounds. My +God, what shall I do! I have lost my honour, my gems, and my son +in one night. Oh, what shall I do!" + +He put a hand on either side of his head and rocked himself to +and fro, droning to himself like a child whose grief has got +beyond words. + +Sherlock Holmes sat silent for some few minutes, with his brows +knitted and his eyes fixed upon the fire. + +"Do you receive much company?" he asked. + +"None save my partner with his family and an occasional friend of +Arthur's. Sir George Burnwell has been several times lately. No +one else, I think." + +"Do you go out much in society?" + +"Arthur does. Mary and I stay at home. We neither of us care for +it." + +"That is unusual in a young girl." + +"She is of a quiet nature. Besides, she is not so very young. She +is four-and-twenty." + +"This matter, from what you say, seems to have been a shock to +her also." + +"Terrible! She is even more affected than I." + +"You have neither of you any doubt as to your son's guilt?" + +"How can we have when I saw him with my own eyes with the coronet +in his hands." + +"I hardly consider that a conclusive proof. Was the remainder of +the coronet at all injured?" + +"Yes, it was twisted." + +"Do you not think, then, that he might have been trying to +straighten it?" + +"God bless you! You are doing what you can for him and for me. +But it is too heavy a task. What was he doing there at all? If +his purpose were innocent, why did he not say so?" + +"Precisely. And if it were guilty, why did he not invent a lie? +His silence appears to me to cut both ways. There are several +singular points about the case. What did the police think of the +noise which awoke you from your sleep?" + +"They considered that it might be caused by Arthur's closing his +bedroom door." + +"A likely story! As if a man bent on felony would slam his door +so as to wake a household. What did they say, then, of the +disappearance of these gems?" + +"They are still sounding the planking and probing the furniture +in the hope of finding them." + +"Have they thought of looking outside the house?" + +"Yes, they have shown extraordinary energy. The whole garden has +already been minutely examined." + +"Now, my dear sir," said Holmes, "is it not obvious to you now +that this matter really strikes very much deeper than either you +or the police were at first inclined to think? It appeared to you +to be a simple case; to me it seems exceedingly complex. Consider +what is involved by your theory. You suppose that your son came +down from his bed, went, at great risk, to your dressing-room, +opened your bureau, took out your coronet, broke off by main +force a small portion of it, went off to some other place, +concealed three gems out of the thirty-nine, with such skill that +nobody can find them, and then returned with the other thirty-six +into the room in which he exposed himself to the greatest danger +of being discovered. I ask you now, is such a theory tenable?" + +"But what other is there?" cried the banker with a gesture of +despair. "If his motives were innocent, why does he not explain +them?" + +"It is our task to find that out," replied Holmes; "so now, if +you please, Mr. Holder, we will set off for Streatham together, +and devote an hour to glancing a little more closely into +details." + +My friend insisted upon my accompanying them in their expedition, +which I was eager enough to do, for my curiosity and sympathy +were deeply stirred by the story to which we had listened. I +confess that the guilt of the banker's son appeared to me to be +as obvious as it did to his unhappy father, but still I had such +faith in Holmes' judgment that I felt that there must be some +grounds for hope as long as he was dissatisfied with the accepted +explanation. He hardly spoke a word the whole way out to the +southern suburb, but sat with his chin upon his breast and his +hat drawn over his eyes, sunk in the deepest thought. Our client +appeared to have taken fresh heart at the little glimpse of hope +which had been presented to him, and he even broke into a +desultory chat with me over his business affairs. A short railway +journey and a shorter walk brought us to Fairbank, the modest +residence of the great financier. + +Fairbank was a good-sized square house of white stone, standing +back a little from the road. A double carriage-sweep, with a +snow-clad lawn, stretched down in front to two large iron gates +which closed the entrance. On the right side was a small wooden +thicket, which led into a narrow path between two neat hedges +stretching from the road to the kitchen door, and forming the +tradesmen's entrance. On the left ran a lane which led to the +stables, and was not itself within the grounds at all, being a +public, though little used, thoroughfare. Holmes left us standing +at the door and walked slowly all round the house, across the +front, down the tradesmen's path, and so round by the garden +behind into the stable lane. So long was he that Mr. Holder and I +went into the dining-room and waited by the fire until he should +return. We were sitting there in silence when the door opened and +a young lady came in. She was rather above the middle height, +slim, with dark hair and eyes, which seemed the darker against +the absolute pallor of her skin. I do not think that I have ever +seen such deadly paleness in a woman's face. Her lips, too, were +bloodless, but her eyes were flushed with crying. As she swept +silently into the room she impressed me with a greater sense of +grief than the banker had done in the morning, and it was the +more striking in her as she was evidently a woman of strong +character, with immense capacity for self-restraint. Disregarding +my presence, she went straight to her uncle and passed her hand +over his head with a sweet womanly caress. + +"You have given orders that Arthur should be liberated, have you +not, dad?" she asked. + +"No, no, my girl, the matter must be probed to the bottom." + +"But I am so sure that he is innocent. You know what woman's +instincts are. I know that he has done no harm and that you will +be sorry for having acted so harshly." + +"Why is he silent, then, if he is innocent?" + +"Who knows? Perhaps because he was so angry that you should +suspect him." + +"How could I help suspecting him, when I actually saw him with +the coronet in his hand?" + +"Oh, but he had only picked it up to look at it. Oh, do, do take +my word for it that he is innocent. Let the matter drop and say +no more. It is so dreadful to think of our dear Arthur in +prison!" + +"I shall never let it drop until the gems are found--never, Mary! +Your affection for Arthur blinds you as to the awful consequences +to me. Far from hushing the thing up, I have brought a gentleman +down from London to inquire more deeply into it." + +"This gentleman?" she asked, facing round to me. + +"No, his friend. He wished us to leave him alone. He is round in +the stable lane now." + +"The stable lane?" She raised her dark eyebrows. "What can he +hope to find there? Ah! this, I suppose, is he. I trust, sir, +that you will succeed in proving, what I feel sure is the truth, +that my cousin Arthur is innocent of this crime." + +"I fully share your opinion, and I trust, with you, that we may +prove it," returned Holmes, going back to the mat to knock the +snow from his shoes. "I believe I have the honour of addressing +Miss Mary Holder. Might I ask you a question or two?" + +"Pray do, sir, if it may help to clear this horrible affair up." + +"You heard nothing yourself last night?" + +"Nothing, until my uncle here began to speak loudly. I heard +that, and I came down." + +"You shut up the windows and doors the night before. Did you +fasten all the windows?" + +"Yes." + +"Were they all fastened this morning?" + +"Yes." + +"You have a maid who has a sweetheart? I think that you remarked +to your uncle last night that she had been out to see him?" + +"Yes, and she was the girl who waited in the drawing-room, and +who may have heard uncle's remarks about the coronet." + +"I see. You infer that she may have gone out to tell her +sweetheart, and that the two may have planned the robbery." + +"But what is the good of all these vague theories," cried the +banker impatiently, "when I have told you that I saw Arthur with +the coronet in his hands?" + +"Wait a little, Mr. Holder. We must come back to that. About this +girl, Miss Holder. You saw her return by the kitchen door, I +presume?" + +"Yes; when I went to see if the door was fastened for the night I +met her slipping in. I saw the man, too, in the gloom." + +"Do you know him?" + +"Oh, yes! he is the green-grocer who brings our vegetables round. +His name is Francis Prosper." + +"He stood," said Holmes, "to the left of the door--that is to +say, farther up the path than is necessary to reach the door?" + +"Yes, he did." + +"And he is a man with a wooden leg?" + +Something like fear sprang up in the young lady's expressive +black eyes. "Why, you are like a magician," said she. "How do you +know that?" She smiled, but there was no answering smile in +Holmes' thin, eager face. + +"I should be very glad now to go upstairs," said he. "I shall +probably wish to go over the outside of the house again. Perhaps +I had better take a look at the lower windows before I go up." + +He walked swiftly round from one to the other, pausing only at +the large one which looked from the hall onto the stable lane. +This he opened and made a very careful examination of the sill +with his powerful magnifying lens. "Now we shall go upstairs," +said he at last. + +The banker's dressing-room was a plainly furnished little +chamber, with a grey carpet, a large bureau, and a long mirror. +Holmes went to the bureau first and looked hard at the lock. + +"Which key was used to open it?" he asked. + +"That which my son himself indicated--that of the cupboard of the +lumber-room." + +"Have you it here?" + +"That is it on the dressing-table." + +Sherlock Holmes took it up and opened the bureau. + +"It is a noiseless lock," said he. "It is no wonder that it did +not wake you. This case, I presume, contains the coronet. We must +have a look at it." He opened the case, and taking out the diadem +he laid it upon the table. It was a magnificent specimen of the +jeweller's art, and the thirty-six stones were the finest that I +have ever seen. At one side of the coronet was a cracked edge, +where a corner holding three gems had been torn away. + +"Now, Mr. Holder," said Holmes, "here is the corner which +corresponds to that which has been so unfortunately lost. Might I +beg that you will break it off." + +The banker recoiled in horror. "I should not dream of trying," +said he. + +"Then I will." Holmes suddenly bent his strength upon it, but +without result. "I feel it give a little," said he; "but, though +I am exceptionally strong in the fingers, it would take me all my +time to break it. An ordinary man could not do it. Now, what do +you think would happen if I did break it, Mr. Holder? There would +be a noise like a pistol shot. Do you tell me that all this +happened within a few yards of your bed and that you heard +nothing of it?" + +"I do not know what to think. It is all dark to me." + +"But perhaps it may grow lighter as we go. What do you think, +Miss Holder?" + +"I confess that I still share my uncle's perplexity." + +"Your son had no shoes or slippers on when you saw him?" + +"He had nothing on save only his trousers and shirt." + +"Thank you. We have certainly been favoured with extraordinary +luck during this inquiry, and it will be entirely our own fault +if we do not succeed in clearing the matter up. With your +permission, Mr. Holder, I shall now continue my investigations +outside." + +He went alone, at his own request, for he explained that any +unnecessary footmarks might make his task more difficult. For an +hour or more he was at work, returning at last with his feet +heavy with snow and his features as inscrutable as ever. + +"I think that I have seen now all that there is to see, Mr. +Holder," said he; "I can serve you best by returning to my +rooms." + +"But the gems, Mr. Holmes. Where are they?" + +"I cannot tell." + +The banker wrung his hands. "I shall never see them again!" he +cried. "And my son? You give me hopes?" + +"My opinion is in no way altered." + +"Then, for God's sake, what was this dark business which was +acted in my house last night?" + +"If you can call upon me at my Baker Street rooms to-morrow +morning between nine and ten I shall be happy to do what I can to +make it clearer. I understand that you give me carte blanche to +act for you, provided only that I get back the gems, and that you +place no limit on the sum I may draw." + +"I would give my fortune to have them back." + +"Very good. I shall look into the matter between this and then. +Good-bye; it is just possible that I may have to come over here +again before evening." + +It was obvious to me that my companion's mind was now made up +about the case, although what his conclusions were was more than +I could even dimly imagine. Several times during our homeward +journey I endeavoured to sound him upon the point, but he always +glided away to some other topic, until at last I gave it over in +despair. It was not yet three when we found ourselves in our +rooms once more. He hurried to his chamber and was down again in +a few minutes dressed as a common loafer. With his collar turned +up, his shiny, seedy coat, his red cravat, and his worn boots, he +was a perfect sample of the class. + +"I think that this should do," said he, glancing into the glass +above the fireplace. "I only wish that you could come with me, +Watson, but I fear that it won't do. I may be on the trail in +this matter, or I may be following a will-o'-the-wisp, but I +shall soon know which it is. I hope that I may be back in a few +hours." He cut a slice of beef from the joint upon the sideboard, +sandwiched it between two rounds of bread, and thrusting this +rude meal into his pocket he started off upon his expedition. + +I had just finished my tea when he returned, evidently in +excellent spirits, swinging an old elastic-sided boot in his +hand. He chucked it down into a corner and helped himself to a +cup of tea. + +"I only looked in as I passed," said he. "I am going right on." + +"Where to?" + +"Oh, to the other side of the West End. It may be some time +before I get back. Don't wait up for me in case I should be +late." + +"How are you getting on?" + +"Oh, so so. Nothing to complain of. I have been out to Streatham +since I saw you last, but I did not call at the house. It is a +very sweet little problem, and I would not have missed it for a +good deal. However, I must not sit gossiping here, but must get +these disreputable clothes off and return to my highly +respectable self." + +I could see by his manner that he had stronger reasons for +satisfaction than his words alone would imply. His eyes twinkled, +and there was even a touch of colour upon his sallow cheeks. He +hastened upstairs, and a few minutes later I heard the slam of +the hall door, which told me that he was off once more upon his +congenial hunt. + +I waited until midnight, but there was no sign of his return, so +I retired to my room. It was no uncommon thing for him to be away +for days and nights on end when he was hot upon a scent, so that +his lateness caused me no surprise. I do not know at what hour he +came in, but when I came down to breakfast in the morning there +he was with a cup of coffee in one hand and the paper in the +other, as fresh and trim as possible. + +"You will excuse my beginning without you, Watson," said he, "but +you remember that our client has rather an early appointment this +morning." + +"Why, it is after nine now," I answered. "I should not be +surprised if that were he. I thought I heard a ring." + +It was, indeed, our friend the financier. I was shocked by the +change which had come over him, for his face which was naturally +of a broad and massive mould, was now pinched and fallen in, +while his hair seemed to me at least a shade whiter. He entered +with a weariness and lethargy which was even more painful than +his violence of the morning before, and he dropped heavily into +the armchair which I pushed forward for him. + +"I do not know what I have done to be so severely tried," said +he. "Only two days ago I was a happy and prosperous man, without +a care in the world. Now I am left to a lonely and dishonoured +age. One sorrow comes close upon the heels of another. My niece, +Mary, has deserted me." + +"Deserted you?" + +"Yes. Her bed this morning had not been slept in, her room was +empty, and a note for me lay upon the hall table. I had said to +her last night, in sorrow and not in anger, that if she had +married my boy all might have been well with him. Perhaps it was +thoughtless of me to say so. It is to that remark that she refers +in this note: + +"'MY DEAREST UNCLE:--I feel that I have brought trouble upon you, +and that if I had acted differently this terrible misfortune +might never have occurred. I cannot, with this thought in my +mind, ever again be happy under your roof, and I feel that I must +leave you forever. Do not worry about my future, for that is +provided for; and, above all, do not search for me, for it will +be fruitless labour and an ill-service to me. In life or in +death, I am ever your loving,--MARY.' + +"What could she mean by that note, Mr. Holmes? Do you think it +points to suicide?" + +"No, no, nothing of the kind. It is perhaps the best possible +solution. I trust, Mr. Holder, that you are nearing the end of +your troubles." + +"Ha! You say so! You have heard something, Mr. Holmes; you have +learned something! Where are the gems?" + +"You would not think 1000 pounds apiece an excessive sum for +them?" + +"I would pay ten." + +"That would be unnecessary. Three thousand will cover the matter. +And there is a little reward, I fancy. Have you your check-book? +Here is a pen. Better make it out for 4000 pounds." + +With a dazed face the banker made out the required check. Holmes +walked over to his desk, took out a little triangular piece of +gold with three gems in it, and threw it down upon the table. + +With a shriek of joy our client clutched it up. + +"You have it!" he gasped. "I am saved! I am saved!" + +The reaction of joy was as passionate as his grief had been, and +he hugged his recovered gems to his bosom. + +"There is one other thing you owe, Mr. Holder," said Sherlock +Holmes rather sternly. + +"Owe!" He caught up a pen. "Name the sum, and I will pay it." + +"No, the debt is not to me. You owe a very humble apology to that +noble lad, your son, who has carried himself in this matter as I +should be proud to see my own son do, should I ever chance to +have one." + +"Then it was not Arthur who took them?" + +"I told you yesterday, and I repeat to-day, that it was not." + +"You are sure of it! Then let us hurry to him at once to let him +know that the truth is known." + +"He knows it already. When I had cleared it all up I had an +interview with him, and finding that he would not tell me the +story, I told it to him, on which he had to confess that I was +right and to add the very few details which were not yet quite +clear to me. Your news of this morning, however, may open his +lips." + +"For heaven's sake, tell me, then, what is this extraordinary +mystery!" + +"I will do so, and I will show you the steps by which I reached +it. And let me say to you, first, that which it is hardest for me +to say and for you to hear: there has been an understanding +between Sir George Burnwell and your niece Mary. They have now +fled together." + +"My Mary? Impossible!" + +"It is unfortunately more than possible; it is certain. Neither +you nor your son knew the true character of this man when you +admitted him into your family circle. He is one of the most +dangerous men in England--a ruined gambler, an absolutely +desperate villain, a man without heart or conscience. Your niece +knew nothing of such men. When he breathed his vows to her, as he +had done to a hundred before her, she flattered herself that she +alone had touched his heart. The devil knows best what he said, +but at least she became his tool and was in the habit of seeing +him nearly every evening." + +"I cannot, and I will not, believe it!" cried the banker with an +ashen face. + +"I will tell you, then, what occurred in your house last night. +Your niece, when you had, as she thought, gone to your room, +slipped down and talked to her lover through the window which +leads into the stable lane. His footmarks had pressed right +through the snow, so long had he stood there. She told him of the +coronet. His wicked lust for gold kindled at the news, and he +bent her to his will. I have no doubt that she loved you, but +there are women in whom the love of a lover extinguishes all +other loves, and I think that she must have been one. She had +hardly listened to his instructions when she saw you coming +downstairs, on which she closed the window rapidly and told you +about one of the servants' escapade with her wooden-legged lover, +which was all perfectly true. + +"Your boy, Arthur, went to bed after his interview with you but +he slept badly on account of his uneasiness about his club debts. +In the middle of the night he heard a soft tread pass his door, +so he rose and, looking out, was surprised to see his cousin +walking very stealthily along the passage until she disappeared +into your dressing-room. Petrified with astonishment, the lad +slipped on some clothes and waited there in the dark to see what +would come of this strange affair. Presently she emerged from the +room again, and in the light of the passage-lamp your son saw +that she carried the precious coronet in her hands. She passed +down the stairs, and he, thrilling with horror, ran along and +slipped behind the curtain near your door, whence he could see +what passed in the hall beneath. He saw her stealthily open the +window, hand out the coronet to someone in the gloom, and then +closing it once more hurry back to her room, passing quite close +to where he stood hid behind the curtain. + +"As long as she was on the scene he could not take any action +without a horrible exposure of the woman whom he loved. But the +instant that she was gone he realised how crushing a misfortune +this would be for you, and how all-important it was to set it +right. He rushed down, just as he was, in his bare feet, opened +the window, sprang out into the snow, and ran down the lane, +where he could see a dark figure in the moonlight. Sir George +Burnwell tried to get away, but Arthur caught him, and there was +a struggle between them, your lad tugging at one side of the +coronet, and his opponent at the other. In the scuffle, your son +struck Sir George and cut him over the eye. Then something +suddenly snapped, and your son, finding that he had the coronet +in his hands, rushed back, closed the window, ascended to your +room, and had just observed that the coronet had been twisted in +the struggle and was endeavouring to straighten it when you +appeared upon the scene." + +"Is it possible?" gasped the banker. + +"You then roused his anger by calling him names at a moment when +he felt that he had deserved your warmest thanks. He could not +explain the true state of affairs without betraying one who +certainly deserved little enough consideration at his hands. He +took the more chivalrous view, however, and preserved her +secret." + +"And that was why she shrieked and fainted when she saw the +coronet," cried Mr. Holder. "Oh, my God! what a blind fool I have +been! And his asking to be allowed to go out for five minutes! +The dear fellow wanted to see if the missing piece were at the +scene of the struggle. How cruelly I have misjudged him!" + +"When I arrived at the house," continued Holmes, "I at once went +very carefully round it to observe if there were any traces in +the snow which might help me. I knew that none had fallen since +the evening before, and also that there had been a strong frost +to preserve impressions. I passed along the tradesmen's path, but +found it all trampled down and indistinguishable. Just beyond it, +however, at the far side of the kitchen door, a woman had stood +and talked with a man, whose round impressions on one side showed +that he had a wooden leg. I could even tell that they had been +disturbed, for the woman had run back swiftly to the door, as was +shown by the deep toe and light heel marks, while Wooden-leg had +waited a little, and then had gone away. I thought at the time +that this might be the maid and her sweetheart, of whom you had +already spoken to me, and inquiry showed it was so. I passed +round the garden without seeing anything more than random tracks, +which I took to be the police; but when I got into the stable +lane a very long and complex story was written in the snow in +front of me. + +"There was a double line of tracks of a booted man, and a second +double line which I saw with delight belonged to a man with naked +feet. I was at once convinced from what you had told me that the +latter was your son. The first had walked both ways, but the +other had run swiftly, and as his tread was marked in places over +the depression of the boot, it was obvious that he had passed +after the other. I followed them up and found they led to the +hall window, where Boots had worn all the snow away while +waiting. Then I walked to the other end, which was a hundred +yards or more down the lane. I saw where Boots had faced round, +where the snow was cut up as though there had been a struggle, +and, finally, where a few drops of blood had fallen, to show me +that I was not mistaken. Boots had then run down the lane, and +another little smudge of blood showed that it was he who had been +hurt. When he came to the highroad at the other end, I found that +the pavement had been cleared, so there was an end to that clue. + +"On entering the house, however, I examined, as you remember, the +sill and framework of the hall window with my lens, and I could +at once see that someone had passed out. I could distinguish the +outline of an instep where the wet foot had been placed in coming +in. I was then beginning to be able to form an opinion as to what +had occurred. A man had waited outside the window; someone had +brought the gems; the deed had been overseen by your son; he had +pursued the thief; had struggled with him; they had each tugged +at the coronet, their united strength causing injuries which +neither alone could have effected. He had returned with the +prize, but had left a fragment in the grasp of his opponent. So +far I was clear. The question now was, who was the man and who +was it brought him the coronet? + +"It is an old maxim of mine that when you have excluded the +impossible, whatever remains, however improbable, must be the +truth. Now, I knew that it was not you who had brought it down, +so there only remained your niece and the maids. But if it were +the maids, why should your son allow himself to be accused in +their place? There could be no possible reason. As he loved his +cousin, however, there was an excellent explanation why he should +retain her secret--the more so as the secret was a disgraceful +one. When I remembered that you had seen her at that window, and +how she had fainted on seeing the coronet again, my conjecture +became a certainty. + +"And who could it be who was her confederate? A lover evidently, +for who else could outweigh the love and gratitude which she must +feel to you? I knew that you went out little, and that your +circle of friends was a very limited one. But among them was Sir +George Burnwell. I had heard of him before as being a man of evil +reputation among women. It must have been he who wore those boots +and retained the missing gems. Even though he knew that Arthur +had discovered him, he might still flatter himself that he was +safe, for the lad could not say a word without compromising his +own family. + +"Well, your own good sense will suggest what measures I took +next. I went in the shape of a loafer to Sir George's house, +managed to pick up an acquaintance with his valet, learned that +his master had cut his head the night before, and, finally, at +the expense of six shillings, made all sure by buying a pair of +his cast-off shoes. With these I journeyed down to Streatham and +saw that they exactly fitted the tracks." + +"I saw an ill-dressed vagabond in the lane yesterday evening," +said Mr. Holder. + +"Precisely. It was I. I found that I had my man, so I came home +and changed my clothes. It was a delicate part which I had to +play then, for I saw that a prosecution must be avoided to avert +scandal, and I knew that so astute a villain would see that our +hands were tied in the matter. I went and saw him. At first, of +course, he denied everything. But when I gave him every +particular that had occurred, he tried to bluster and took down a +life-preserver from the wall. I knew my man, however, and I +clapped a pistol to his head before he could strike. Then he +became a little more reasonable. I told him that we would give +him a price for the stones he held--1000 pounds apiece. That +brought out the first signs of grief that he had shown. 'Why, +dash it all!' said he, 'I've let them go at six hundred for the +three!' I soon managed to get the address of the receiver who had +them, on promising him that there would be no prosecution. Off I +set to him, and after much chaffering I got our stones at 1000 +pounds apiece. Then I looked in upon your son, told him that all +was right, and eventually got to my bed about two o'clock, after +what I may call a really hard day's work." + +"A day which has saved England from a great public scandal," said +the banker, rising. "Sir, I cannot find words to thank you, but +you shall not find me ungrateful for what you have done. Your +skill has indeed exceeded all that I have heard of it. And now I +must fly to my dear boy to apologise to him for the wrong which I +have done him. As to what you tell me of poor Mary, it goes to my +very heart. Not even your skill can inform me where she is now." + +"I think that we may safely say," returned Holmes, "that she is +wherever Sir George Burnwell is. It is equally certain, too, that +whatever her sins are, they will soon receive a more than +sufficient punishment." + + + +XII. THE ADVENTURE OF THE COPPER BEECHES + +"To the man who loves art for its own sake," remarked Sherlock +Holmes, tossing aside the advertisement sheet of the Daily +Telegraph, "it is frequently in its least important and lowliest +manifestations that the keenest pleasure is to be derived. It is +pleasant to me to observe, Watson, that you have so far grasped +this truth that in these little records of our cases which you +have been good enough to draw up, and, I am bound to say, +occasionally to embellish, you have given prominence not so much +to the many causes célèbres and sensational trials in which I +have figured but rather to those incidents which may have been +trivial in themselves, but which have given room for those +faculties of deduction and of logical synthesis which I have made +my special province." + +"And yet," said I, smiling, "I cannot quite hold myself absolved +from the charge of sensationalism which has been urged against my +records." + +"You have erred, perhaps," he observed, taking up a glowing +cinder with the tongs and lighting with it the long cherry-wood +pipe which was wont to replace his clay when he was in a +disputatious rather than a meditative mood--"you have erred +perhaps in attempting to put colour and life into each of your +statements instead of confining yourself to the task of placing +upon record that severe reasoning from cause to effect which is +really the only notable feature about the thing." + +"It seems to me that I have done you full justice in the matter," +I remarked with some coldness, for I was repelled by the egotism +which I had more than once observed to be a strong factor in my +friend's singular character. + +"No, it is not selfishness or conceit," said he, answering, as +was his wont, my thoughts rather than my words. "If I claim full +justice for my art, it is because it is an impersonal thing--a +thing beyond myself. Crime is common. Logic is rare. Therefore it +is upon the logic rather than upon the crime that you should +dwell. You have degraded what should have been a course of +lectures into a series of tales." + +It was a cold morning of the early spring, and we sat after +breakfast on either side of a cheery fire in the old room at +Baker Street. A thick fog rolled down between the lines of +dun-coloured houses, and the opposing windows loomed like dark, +shapeless blurs through the heavy yellow wreaths. Our gas was lit +and shone on the white cloth and glimmer of china and metal, for +the table had not been cleared yet. Sherlock Holmes had been +silent all the morning, dipping continuously into the +advertisement columns of a succession of papers until at last, +having apparently given up his search, he had emerged in no very +sweet temper to lecture me upon my literary shortcomings. + +"At the same time," he remarked after a pause, during which he +had sat puffing at his long pipe and gazing down into the fire, +"you can hardly be open to a charge of sensationalism, for out of +these cases which you have been so kind as to interest yourself +in, a fair proportion do not treat of crime, in its legal sense, +at all. The small matter in which I endeavoured to help the King +of Bohemia, the singular experience of Miss Mary Sutherland, the +problem connected with the man with the twisted lip, and the +incident of the noble bachelor, were all matters which are +outside the pale of the law. But in avoiding the sensational, I +fear that you may have bordered on the trivial." + +"The end may have been so," I answered, "but the methods I hold +to have been novel and of interest." + +"Pshaw, my dear fellow, what do the public, the great unobservant +public, who could hardly tell a weaver by his tooth or a +compositor by his left thumb, care about the finer shades of +analysis and deduction! But, indeed, if you are trivial, I cannot +blame you, for the days of the great cases are past. Man, or at +least criminal man, has lost all enterprise and originality. As +to my own little practice, it seems to be degenerating into an +agency for recovering lost lead pencils and giving advice to +young ladies from boarding-schools. I think that I have touched +bottom at last, however. This note I had this morning marks my +zero-point, I fancy. Read it!" He tossed a crumpled letter across +to me. + +It was dated from Montague Place upon the preceding evening, and +ran thus: + +"DEAR MR. HOLMES:--I am very anxious to consult you as to whether +I should or should not accept a situation which has been offered +to me as governess. I shall call at half-past ten to-morrow if I +do not inconvenience you. Yours faithfully, + "VIOLET HUNTER." + +"Do you know the young lady?" I asked. + +"Not I." + +"It is half-past ten now." + +"Yes, and I have no doubt that is her ring." + +"It may turn out to be of more interest than you think. You +remember that the affair of the blue carbuncle, which appeared to +be a mere whim at first, developed into a serious investigation. +It may be so in this case, also." + +"Well, let us hope so. But our doubts will very soon be solved, +for here, unless I am much mistaken, is the person in question." + +As he spoke the door opened and a young lady entered the room. +She was plainly but neatly dressed, with a bright, quick face, +freckled like a plover's egg, and with the brisk manner of a +woman who has had her own way to make in the world. + +"You will excuse my troubling you, I am sure," said she, as my +companion rose to greet her, "but I have had a very strange +experience, and as I have no parents or relations of any sort +from whom I could ask advice, I thought that perhaps you would be +kind enough to tell me what I should do." + +"Pray take a seat, Miss Hunter. I shall be happy to do anything +that I can to serve you." + +I could see that Holmes was favourably impressed by the manner +and speech of his new client. He looked her over in his searching +fashion, and then composed himself, with his lids drooping and +his finger-tips together, to listen to her story. + +"I have been a governess for five years," said she, "in the +family of Colonel Spence Munro, but two months ago the colonel +received an appointment at Halifax, in Nova Scotia, and took his +children over to America with him, so that I found myself without +a situation. I advertised, and I answered advertisements, but +without success. At last the little money which I had saved began +to run short, and I was at my wit's end as to what I should do. + +"There is a well-known agency for governesses in the West End +called Westaway's, and there I used to call about once a week in +order to see whether anything had turned up which might suit me. +Westaway was the name of the founder of the business, but it is +really managed by Miss Stoper. She sits in her own little office, +and the ladies who are seeking employment wait in an anteroom, +and are then shown in one by one, when she consults her ledgers +and sees whether she has anything which would suit them. + +"Well, when I called last week I was shown into the little office +as usual, but I found that Miss Stoper was not alone. A +prodigiously stout man with a very smiling face and a great heavy +chin which rolled down in fold upon fold over his throat sat at +her elbow with a pair of glasses on his nose, looking very +earnestly at the ladies who entered. As I came in he gave quite a +jump in his chair and turned quickly to Miss Stoper. + +"'That will do,' said he; 'I could not ask for anything better. +Capital! capital!' He seemed quite enthusiastic and rubbed his +hands together in the most genial fashion. He was such a +comfortable-looking man that it was quite a pleasure to look at +him. + +"'You are looking for a situation, miss?' he asked. + +"'Yes, sir.' + +"'As governess?' + +"'Yes, sir.' + +"'And what salary do you ask?' + +"'I had 4 pounds a month in my last place with Colonel Spence +Munro.' + +"'Oh, tut, tut! sweating--rank sweating!' he cried, throwing his +fat hands out into the air like a man who is in a boiling +passion. 'How could anyone offer so pitiful a sum to a lady with +such attractions and accomplishments?' + +"'My accomplishments, sir, may be less than you imagine,' said I. +'A little French, a little German, music, and drawing--' + +"'Tut, tut!' he cried. 'This is all quite beside the question. +The point is, have you or have you not the bearing and deportment +of a lady? There it is in a nutshell. If you have not, you are +not fitted for the rearing of a child who may some day play a +considerable part in the history of the country. But if you have +why, then, how could any gentleman ask you to condescend to +accept anything under the three figures? Your salary with me, +madam, would commence at 100 pounds a year.' + +"You may imagine, Mr. Holmes, that to me, destitute as I was, +such an offer seemed almost too good to be true. The gentleman, +however, seeing perhaps the look of incredulity upon my face, +opened a pocket-book and took out a note. + +"'It is also my custom,' said he, smiling in the most pleasant +fashion until his eyes were just two little shining slits amid +the white creases of his face, 'to advance to my young ladies +half their salary beforehand, so that they may meet any little +expenses of their journey and their wardrobe.' + +"It seemed to me that I had never met so fascinating and so +thoughtful a man. As I was already in debt to my tradesmen, the +advance was a great convenience, and yet there was something +unnatural about the whole transaction which made me wish to know +a little more before I quite committed myself. + +"'May I ask where you live, sir?' said I. + +"'Hampshire. Charming rural place. The Copper Beeches, five miles +on the far side of Winchester. It is the most lovely country, my +dear young lady, and the dearest old country-house.' + +"'And my duties, sir? I should be glad to know what they would +be.' + +"'One child--one dear little romper just six years old. Oh, if +you could see him killing cockroaches with a slipper! Smack! +smack! smack! Three gone before you could wink!' He leaned back +in his chair and laughed his eyes into his head again. + +"I was a little startled at the nature of the child's amusement, +but the father's laughter made me think that perhaps he was +joking. + +"'My sole duties, then,' I asked, 'are to take charge of a single +child?' + +"'No, no, not the sole, not the sole, my dear young lady,' he +cried. 'Your duty would be, as I am sure your good sense would +suggest, to obey any little commands my wife might give, provided +always that they were such commands as a lady might with +propriety obey. You see no difficulty, heh?' + +"'I should be happy to make myself useful.' + +"'Quite so. In dress now, for example. We are faddy people, you +know--faddy but kind-hearted. If you were asked to wear any dress +which we might give you, you would not object to our little whim. +Heh?' + +"'No,' said I, considerably astonished at his words. + +"'Or to sit here, or sit there, that would not be offensive to +you?' + +"'Oh, no.' + +"'Or to cut your hair quite short before you come to us?' + +"I could hardly believe my ears. As you may observe, Mr. Holmes, +my hair is somewhat luxuriant, and of a rather peculiar tint of +chestnut. It has been considered artistic. I could not dream of +sacrificing it in this offhand fashion. + +"'I am afraid that that is quite impossible,' said I. He had been +watching me eagerly out of his small eyes, and I could see a +shadow pass over his face as I spoke. + +"'I am afraid that it is quite essential,' said he. 'It is a +little fancy of my wife's, and ladies' fancies, you know, madam, +ladies' fancies must be consulted. And so you won't cut your +hair?' + +"'No, sir, I really could not,' I answered firmly. + +"'Ah, very well; then that quite settles the matter. It is a +pity, because in other respects you would really have done very +nicely. In that case, Miss Stoper, I had best inspect a few more +of your young ladies.' + +"The manageress had sat all this while busy with her papers +without a word to either of us, but she glanced at me now with so +much annoyance upon her face that I could not help suspecting +that she had lost a handsome commission through my refusal. + +"'Do you desire your name to be kept upon the books?' she asked. + +"'If you please, Miss Stoper.' + +"'Well, really, it seems rather useless, since you refuse the +most excellent offers in this fashion,' said she sharply. 'You +can hardly expect us to exert ourselves to find another such +opening for you. Good-day to you, Miss Hunter.' She struck a gong +upon the table, and I was shown out by the page. + +"Well, Mr. Holmes, when I got back to my lodgings and found +little enough in the cupboard, and two or three bills upon the +table, I began to ask myself whether I had not done a very +foolish thing. After all, if these people had strange fads and +expected obedience on the most extraordinary matters, they were +at least ready to pay for their eccentricity. Very few +governesses in England are getting 100 pounds a year. Besides, +what use was my hair to me? Many people are improved by wearing +it short and perhaps I should be among the number. Next day I was +inclined to think that I had made a mistake, and by the day after +I was sure of it. I had almost overcome my pride so far as to go +back to the agency and inquire whether the place was still open +when I received this letter from the gentleman himself. I have it +here and I will read it to you: + + "'The Copper Beeches, near Winchester. +"'DEAR MISS HUNTER:--Miss Stoper has very kindly given me your +address, and I write from here to ask you whether you have +reconsidered your decision. My wife is very anxious that you +should come, for she has been much attracted by my description of +you. We are willing to give 30 pounds a quarter, or 120 pounds a +year, so as to recompense you for any little inconvenience which +our fads may cause you. They are not very exacting, after all. My +wife is fond of a particular shade of electric blue and would +like you to wear such a dress indoors in the morning. You need +not, however, go to the expense of purchasing one, as we have one +belonging to my dear daughter Alice (now in Philadelphia), which +would, I should think, fit you very well. Then, as to sitting +here or there, or amusing yourself in any manner indicated, that +need cause you no inconvenience. As regards your hair, it is no +doubt a pity, especially as I could not help remarking its beauty +during our short interview, but I am afraid that I must remain +firm upon this point, and I only hope that the increased salary +may recompense you for the loss. Your duties, as far as the child +is concerned, are very light. Now do try to come, and I shall +meet you with the dog-cart at Winchester. Let me know your train. +Yours faithfully, JEPHRO RUCASTLE.' + +"That is the letter which I have just received, Mr. Holmes, and +my mind is made up that I will accept it. I thought, however, +that before taking the final step I should like to submit the +whole matter to your consideration." + +"Well, Miss Hunter, if your mind is made up, that settles the +question," said Holmes, smiling. + +"But you would not advise me to refuse?" + +"I confess that it is not the situation which I should like to +see a sister of mine apply for." + +"What is the meaning of it all, Mr. Holmes?" + +"Ah, I have no data. I cannot tell. Perhaps you have yourself +formed some opinion?" + +"Well, there seems to me to be only one possible solution. Mr. +Rucastle seemed to be a very kind, good-natured man. Is it not +possible that his wife is a lunatic, that he desires to keep the +matter quiet for fear she should be taken to an asylum, and that +he humours her fancies in every way in order to prevent an +outbreak?" + +"That is a possible solution--in fact, as matters stand, it is +the most probable one. But in any case it does not seem to be a +nice household for a young lady." + +"But the money, Mr. Holmes, the money!" + +"Well, yes, of course the pay is good--too good. That is what +makes me uneasy. Why should they give you 120 pounds a year, when +they could have their pick for 40 pounds? There must be some +strong reason behind." + +"I thought that if I told you the circumstances you would +understand afterwards if I wanted your help. I should feel so +much stronger if I felt that you were at the back of me." + +"Oh, you may carry that feeling away with you. I assure you that +your little problem promises to be the most interesting which has +come my way for some months. There is something distinctly novel +about some of the features. If you should find yourself in doubt +or in danger--" + +"Danger! What danger do you foresee?" + +Holmes shook his head gravely. "It would cease to be a danger if +we could define it," said he. "But at any time, day or night, a +telegram would bring me down to your help." + +"That is enough." She rose briskly from her chair with the +anxiety all swept from her face. "I shall go down to Hampshire +quite easy in my mind now. I shall write to Mr. Rucastle at once, +sacrifice my poor hair to-night, and start for Winchester +to-morrow." With a few grateful words to Holmes she bade us both +good-night and bustled off upon her way. + +"At least," said I as we heard her quick, firm steps descending +the stairs, "she seems to be a young lady who is very well able +to take care of herself." + +"And she would need to be," said Holmes gravely. "I am much +mistaken if we do not hear from her before many days are past." + +It was not very long before my friend's prediction was fulfilled. +A fortnight went by, during which I frequently found my thoughts +turning in her direction and wondering what strange side-alley of +human experience this lonely woman had strayed into. The unusual +salary, the curious conditions, the light duties, all pointed to +something abnormal, though whether a fad or a plot, or whether +the man were a philanthropist or a villain, it was quite beyond +my powers to determine. As to Holmes, I observed that he sat +frequently for half an hour on end, with knitted brows and an +abstracted air, but he swept the matter away with a wave of his +hand when I mentioned it. "Data! data! data!" he cried +impatiently. "I can't make bricks without clay." And yet he would +always wind up by muttering that no sister of his should ever +have accepted such a situation. + +The telegram which we eventually received came late one night +just as I was thinking of turning in and Holmes was settling down +to one of those all-night chemical researches which he frequently +indulged in, when I would leave him stooping over a retort and a +test-tube at night and find him in the same position when I came +down to breakfast in the morning. He opened the yellow envelope, +and then, glancing at the message, threw it across to me. + +"Just look up the trains in Bradshaw," said he, and turned back +to his chemical studies. + +The summons was a brief and urgent one. + +"Please be at the Black Swan Hotel at Winchester at midday +to-morrow," it said. "Do come! I am at my wit's end. HUNTER." + +"Will you come with me?" asked Holmes, glancing up. + +"I should wish to." + +"Just look it up, then." + +"There is a train at half-past nine," said I, glancing over my +Bradshaw. "It is due at Winchester at 11:30." + +"That will do very nicely. Then perhaps I had better postpone my +analysis of the acetones, as we may need to be at our best in the +morning." + +By eleven o'clock the next day we were well upon our way to the +old English capital. Holmes had been buried in the morning papers +all the way down, but after we had passed the Hampshire border he +threw them down and began to admire the scenery. It was an ideal +spring day, a light blue sky, flecked with little fleecy white +clouds drifting across from west to east. The sun was shining +very brightly, and yet there was an exhilarating nip in the air, +which set an edge to a man's energy. All over the countryside, +away to the rolling hills around Aldershot, the little red and +grey roofs of the farm-steadings peeped out from amid the light +green of the new foliage. + +"Are they not fresh and beautiful?" I cried with all the +enthusiasm of a man fresh from the fogs of Baker Street. + +But Holmes shook his head gravely. + +"Do you know, Watson," said he, "that it is one of the curses of +a mind with a turn like mine that I must look at everything with +reference to my own special subject. You look at these scattered +houses, and you are impressed by their beauty. I look at them, +and the only thought which comes to me is a feeling of their +isolation and of the impunity with which crime may be committed +there." + +"Good heavens!" I cried. "Who would associate crime with these +dear old homesteads?" + +"They always fill me with a certain horror. It is my belief, +Watson, founded upon my experience, that the lowest and vilest +alleys in London do not present a more dreadful record of sin +than does the smiling and beautiful countryside." + +"You horrify me!" + +"But the reason is very obvious. The pressure of public opinion +can do in the town what the law cannot accomplish. There is no +lane so vile that the scream of a tortured child, or the thud of +a drunkard's blow, does not beget sympathy and indignation among +the neighbours, and then the whole machinery of justice is ever +so close that a word of complaint can set it going, and there is +but a step between the crime and the dock. But look at these +lonely houses, each in its own fields, filled for the most part +with poor ignorant folk who know little of the law. Think of the +deeds of hellish cruelty, the hidden wickedness which may go on, +year in, year out, in such places, and none the wiser. Had this +lady who appeals to us for help gone to live in Winchester, I +should never have had a fear for her. It is the five miles of +country which makes the danger. Still, it is clear that she is +not personally threatened." + +"No. If she can come to Winchester to meet us she can get away." + +"Quite so. She has her freedom." + +"What CAN be the matter, then? Can you suggest no explanation?" + +"I have devised seven separate explanations, each of which would +cover the facts as far as we know them. But which of these is +correct can only be determined by the fresh information which we +shall no doubt find waiting for us. Well, there is the tower of +the cathedral, and we shall soon learn all that Miss Hunter has +to tell." + +The Black Swan is an inn of repute in the High Street, at no +distance from the station, and there we found the young lady +waiting for us. She had engaged a sitting-room, and our lunch +awaited us upon the table. + +"I am so delighted that you have come," she said earnestly. "It +is so very kind of you both; but indeed I do not know what I +should do. Your advice will be altogether invaluable to me." + +"Pray tell us what has happened to you." + +"I will do so, and I must be quick, for I have promised Mr. +Rucastle to be back before three. I got his leave to come into +town this morning, though he little knew for what purpose." + +"Let us have everything in its due order." Holmes thrust his long +thin legs out towards the fire and composed himself to listen. + +"In the first place, I may say that I have met, on the whole, +with no actual ill-treatment from Mr. and Mrs. Rucastle. It is +only fair to them to say that. But I cannot understand them, and +I am not easy in my mind about them." + +"What can you not understand?" + +"Their reasons for their conduct. But you shall have it all just +as it occurred. When I came down, Mr. Rucastle met me here and +drove me in his dog-cart to the Copper Beeches. It is, as he +said, beautifully situated, but it is not beautiful in itself, +for it is a large square block of a house, whitewashed, but all +stained and streaked with damp and bad weather. There are grounds +round it, woods on three sides, and on the fourth a field which +slopes down to the Southampton highroad, which curves past about +a hundred yards from the front door. This ground in front belongs +to the house, but the woods all round are part of Lord +Southerton's preserves. A clump of copper beeches immediately in +front of the hall door has given its name to the place. + +"I was driven over by my employer, who was as amiable as ever, +and was introduced by him that evening to his wife and the child. +There was no truth, Mr. Holmes, in the conjecture which seemed to +us to be probable in your rooms at Baker Street. Mrs. Rucastle is +not mad. I found her to be a silent, pale-faced woman, much +younger than her husband, not more than thirty, I should think, +while he can hardly be less than forty-five. From their +conversation I have gathered that they have been married about +seven years, that he was a widower, and that his only child by +the first wife was the daughter who has gone to Philadelphia. Mr. +Rucastle told me in private that the reason why she had left them +was that she had an unreasoning aversion to her stepmother. As +the daughter could not have been less than twenty, I can quite +imagine that her position must have been uncomfortable with her +father's young wife. + +"Mrs. Rucastle seemed to me to be colourless in mind as well as +in feature. She impressed me neither favourably nor the reverse. +She was a nonentity. It was easy to see that she was passionately +devoted both to her husband and to her little son. Her light grey +eyes wandered continually from one to the other, noting every +little want and forestalling it if possible. He was kind to her +also in his bluff, boisterous fashion, and on the whole they +seemed to be a happy couple. And yet she had some secret sorrow, +this woman. She would often be lost in deep thought, with the +saddest look upon her face. More than once I have surprised her +in tears. I have thought sometimes that it was the disposition of +her child which weighed upon her mind, for I have never met so +utterly spoiled and so ill-natured a little creature. He is small +for his age, with a head which is quite disproportionately large. +His whole life appears to be spent in an alternation between +savage fits of passion and gloomy intervals of sulking. Giving +pain to any creature weaker than himself seems to be his one idea +of amusement, and he shows quite remarkable talent in planning +the capture of mice, little birds, and insects. But I would +rather not talk about the creature, Mr. Holmes, and, indeed, he +has little to do with my story." + +"I am glad of all details," remarked my friend, "whether they +seem to you to be relevant or not." + +"I shall try not to miss anything of importance. The one +unpleasant thing about the house, which struck me at once, was +the appearance and conduct of the servants. There are only two, a +man and his wife. Toller, for that is his name, is a rough, +uncouth man, with grizzled hair and whiskers, and a perpetual +smell of drink. Twice since I have been with them he has been +quite drunk, and yet Mr. Rucastle seemed to take no notice of it. +His wife is a very tall and strong woman with a sour face, as +silent as Mrs. Rucastle and much less amiable. They are a most +unpleasant couple, but fortunately I spend most of my time in the +nursery and my own room, which are next to each other in one +corner of the building. + +"For two days after my arrival at the Copper Beeches my life was +very quiet; on the third, Mrs. Rucastle came down just after +breakfast and whispered something to her husband. + +"'Oh, yes,' said he, turning to me, 'we are very much obliged to +you, Miss Hunter, for falling in with our whims so far as to cut +your hair. I assure you that it has not detracted in the tiniest +iota from your appearance. We shall now see how the electric-blue +dress will become you. You will find it laid out upon the bed in +your room, and if you would be so good as to put it on we should +both be extremely obliged.' + +"The dress which I found waiting for me was of a peculiar shade +of blue. It was of excellent material, a sort of beige, but it +bore unmistakable signs of having been worn before. It could not +have been a better fit if I had been measured for it. Both Mr. +and Mrs. Rucastle expressed a delight at the look of it, which +seemed quite exaggerated in its vehemence. They were waiting for +me in the drawing-room, which is a very large room, stretching +along the entire front of the house, with three long windows +reaching down to the floor. A chair had been placed close to the +central window, with its back turned towards it. In this I was +asked to sit, and then Mr. Rucastle, walking up and down on the +other side of the room, began to tell me a series of the funniest +stories that I have ever listened to. You cannot imagine how +comical he was, and I laughed until I was quite weary. Mrs. +Rucastle, however, who has evidently no sense of humour, never so +much as smiled, but sat with her hands in her lap, and a sad, +anxious look upon her face. After an hour or so, Mr. Rucastle +suddenly remarked that it was time to commence the duties of the +day, and that I might change my dress and go to little Edward in +the nursery. + +"Two days later this same performance was gone through under +exactly similar circumstances. Again I changed my dress, again I +sat in the window, and again I laughed very heartily at the funny +stories of which my employer had an immense répertoire, and which +he told inimitably. Then he handed me a yellow-backed novel, and +moving my chair a little sideways, that my own shadow might not +fall upon the page, he begged me to read aloud to him. I read for +about ten minutes, beginning in the heart of a chapter, and then +suddenly, in the middle of a sentence, he ordered me to cease and +to change my dress. + +"You can easily imagine, Mr. Holmes, how curious I became as to +what the meaning of this extraordinary performance could possibly +be. They were always very careful, I observed, to turn my face +away from the window, so that I became consumed with the desire +to see what was going on behind my back. At first it seemed to be +impossible, but I soon devised a means. My hand-mirror had been +broken, so a happy thought seized me, and I concealed a piece of +the glass in my handkerchief. On the next occasion, in the midst +of my laughter, I put my handkerchief up to my eyes, and was able +with a little management to see all that there was behind me. I +confess that I was disappointed. There was nothing. At least that +was my first impression. At the second glance, however, I +perceived that there was a man standing in the Southampton Road, +a small bearded man in a grey suit, who seemed to be looking in +my direction. The road is an important highway, and there are +usually people there. This man, however, was leaning against the +railings which bordered our field and was looking earnestly up. I +lowered my handkerchief and glanced at Mrs. Rucastle to find her +eyes fixed upon me with a most searching gaze. She said nothing, +but I am convinced that she had divined that I had a mirror in my +hand and had seen what was behind me. She rose at once. + +"'Jephro,' said she, 'there is an impertinent fellow upon the +road there who stares up at Miss Hunter.' + +"'No friend of yours, Miss Hunter?' he asked. + +"'No, I know no one in these parts.' + +"'Dear me! How very impertinent! Kindly turn round and motion to +him to go away.' + +"'Surely it would be better to take no notice.' + +"'No, no, we should have him loitering here always. Kindly turn +round and wave him away like that.' + +"I did as I was told, and at the same instant Mrs. Rucastle drew +down the blind. That was a week ago, and from that time I have +not sat again in the window, nor have I worn the blue dress, nor +seen the man in the road." + +"Pray continue," said Holmes. "Your narrative promises to be a +most interesting one." + +"You will find it rather disconnected, I fear, and there may +prove to be little relation between the different incidents of +which I speak. On the very first day that I was at the Copper +Beeches, Mr. Rucastle took me to a small outhouse which stands +near the kitchen door. As we approached it I heard the sharp +rattling of a chain, and the sound as of a large animal moving +about. + +"'Look in here!' said Mr. Rucastle, showing me a slit between two +planks. 'Is he not a beauty?' + +"I looked through and was conscious of two glowing eyes, and of a +vague figure huddled up in the darkness. + +"'Don't be frightened,' said my employer, laughing at the start +which I had given. 'It's only Carlo, my mastiff. I call him mine, +but really old Toller, my groom, is the only man who can do +anything with him. We feed him once a day, and not too much then, +so that he is always as keen as mustard. Toller lets him loose +every night, and God help the trespasser whom he lays his fangs +upon. For goodness' sake don't you ever on any pretext set your +foot over the threshold at night, for it's as much as your life +is worth.' + +"The warning was no idle one, for two nights later I happened to +look out of my bedroom window about two o'clock in the morning. +It was a beautiful moonlight night, and the lawn in front of the +house was silvered over and almost as bright as day. I was +standing, rapt in the peaceful beauty of the scene, when I was +aware that something was moving under the shadow of the copper +beeches. As it emerged into the moonshine I saw what it was. It +was a giant dog, as large as a calf, tawny tinted, with hanging +jowl, black muzzle, and huge projecting bones. It walked slowly +across the lawn and vanished into the shadow upon the other side. +That dreadful sentinel sent a chill to my heart which I do not +think that any burglar could have done. + +"And now I have a very strange experience to tell you. I had, as +you know, cut off my hair in London, and I had placed it in a +great coil at the bottom of my trunk. One evening, after the +child was in bed, I began to amuse myself by examining the +furniture of my room and by rearranging my own little things. +There was an old chest of drawers in the room, the two upper ones +empty and open, the lower one locked. I had filled the first two +with my linen, and as I had still much to pack away I was +naturally annoyed at not having the use of the third drawer. It +struck me that it might have been fastened by a mere oversight, +so I took out my bunch of keys and tried to open it. The very +first key fitted to perfection, and I drew the drawer open. There +was only one thing in it, but I am sure that you would never +guess what it was. It was my coil of hair. + +"I took it up and examined it. It was of the same peculiar tint, +and the same thickness. But then the impossibility of the thing +obtruded itself upon me. How could my hair have been locked in +the drawer? With trembling hands I undid my trunk, turned out the +contents, and drew from the bottom my own hair. I laid the two +tresses together, and I assure you that they were identical. Was +it not extraordinary? Puzzle as I would, I could make nothing at +all of what it meant. I returned the strange hair to the drawer, +and I said nothing of the matter to the Rucastles as I felt that +I had put myself in the wrong by opening a drawer which they had +locked. + +"I am naturally observant, as you may have remarked, Mr. Holmes, +and I soon had a pretty good plan of the whole house in my head. +There was one wing, however, which appeared not to be inhabited +at all. A door which faced that which led into the quarters of +the Tollers opened into this suite, but it was invariably locked. +One day, however, as I ascended the stair, I met Mr. Rucastle +coming out through this door, his keys in his hand, and a look on +his face which made him a very different person to the round, +jovial man to whom I was accustomed. His cheeks were red, his +brow was all crinkled with anger, and the veins stood out at his +temples with passion. He locked the door and hurried past me +without a word or a look. + +"This aroused my curiosity, so when I went out for a walk in the +grounds with my charge, I strolled round to the side from which I +could see the windows of this part of the house. There were four +of them in a row, three of which were simply dirty, while the +fourth was shuttered up. They were evidently all deserted. As I +strolled up and down, glancing at them occasionally, Mr. Rucastle +came out to me, looking as merry and jovial as ever. + +"'Ah!' said he, 'you must not think me rude if I passed you +without a word, my dear young lady. I was preoccupied with +business matters.' + +"I assured him that I was not offended. 'By the way,' said I, +'you seem to have quite a suite of spare rooms up there, and one +of them has the shutters up.' + +"He looked surprised and, as it seemed to me, a little startled +at my remark. + +"'Photography is one of my hobbies,' said he. 'I have made my +dark room up there. But, dear me! what an observant young lady we +have come upon. Who would have believed it? Who would have ever +believed it?' He spoke in a jesting tone, but there was no jest +in his eyes as he looked at me. I read suspicion there and +annoyance, but no jest. + +"Well, Mr. Holmes, from the moment that I understood that there +was something about that suite of rooms which I was not to know, +I was all on fire to go over them. It was not mere curiosity, +though I have my share of that. It was more a feeling of duty--a +feeling that some good might come from my penetrating to this +place. They talk of woman's instinct; perhaps it was woman's +instinct which gave me that feeling. At any rate, it was there, +and I was keenly on the lookout for any chance to pass the +forbidden door. + +"It was only yesterday that the chance came. I may tell you that, +besides Mr. Rucastle, both Toller and his wife find something to +do in these deserted rooms, and I once saw him carrying a large +black linen bag with him through the door. Recently he has been +drinking hard, and yesterday evening he was very drunk; and when +I came upstairs there was the key in the door. I have no doubt at +all that he had left it there. Mr. and Mrs. Rucastle were both +downstairs, and the child was with them, so that I had an +admirable opportunity. I turned the key gently in the lock, +opened the door, and slipped through. + +"There was a little passage in front of me, unpapered and +uncarpeted, which turned at a right angle at the farther end. +Round this corner were three doors in a line, the first and third +of which were open. They each led into an empty room, dusty and +cheerless, with two windows in the one and one in the other, so +thick with dirt that the evening light glimmered dimly through +them. The centre door was closed, and across the outside of it +had been fastened one of the broad bars of an iron bed, padlocked +at one end to a ring in the wall, and fastened at the other with +stout cord. The door itself was locked as well, and the key was +not there. This barricaded door corresponded clearly with the +shuttered window outside, and yet I could see by the glimmer from +beneath it that the room was not in darkness. Evidently there was +a skylight which let in light from above. As I stood in the +passage gazing at the sinister door and wondering what secret it +might veil, I suddenly heard the sound of steps within the room +and saw a shadow pass backward and forward against the little +slit of dim light which shone out from under the door. A mad, +unreasoning terror rose up in me at the sight, Mr. Holmes. My +overstrung nerves failed me suddenly, and I turned and ran--ran +as though some dreadful hand were behind me clutching at the +skirt of my dress. I rushed down the passage, through the door, +and straight into the arms of Mr. Rucastle, who was waiting +outside. + +"'So,' said he, smiling, 'it was you, then. I thought that it +must be when I saw the door open.' + +"'Oh, I am so frightened!' I panted. + +"'My dear young lady! my dear young lady!'--you cannot think how +caressing and soothing his manner was--'and what has frightened +you, my dear young lady?' + +"But his voice was just a little too coaxing. He overdid it. I +was keenly on my guard against him. + +"'I was foolish enough to go into the empty wing,' I answered. +'But it is so lonely and eerie in this dim light that I was +frightened and ran out again. Oh, it is so dreadfully still in +there!' + +"'Only that?' said he, looking at me keenly. + +"'Why, what did you think?' I asked. + +"'Why do you think that I lock this door?' + +"'I am sure that I do not know.' + +"'It is to keep people out who have no business there. Do you +see?' He was still smiling in the most amiable manner. + +"'I am sure if I had known--' + +"'Well, then, you know now. And if you ever put your foot over +that threshold again'--here in an instant the smile hardened into +a grin of rage, and he glared down at me with the face of a +demon--'I'll throw you to the mastiff.' + +"I was so terrified that I do not know what I did. I suppose that +I must have rushed past him into my room. I remember nothing +until I found myself lying on my bed trembling all over. Then I +thought of you, Mr. Holmes. I could not live there longer without +some advice. I was frightened of the house, of the man, of the +woman, of the servants, even of the child. They were all horrible +to me. If I could only bring you down all would be well. Of +course I might have fled from the house, but my curiosity was +almost as strong as my fears. My mind was soon made up. I would +send you a wire. I put on my hat and cloak, went down to the +office, which is about half a mile from the house, and then +returned, feeling very much easier. A horrible doubt came into my +mind as I approached the door lest the dog might be loose, but I +remembered that Toller had drunk himself into a state of +insensibility that evening, and I knew that he was the only one +in the household who had any influence with the savage creature, +or who would venture to set him free. I slipped in in safety and +lay awake half the night in my joy at the thought of seeing you. +I had no difficulty in getting leave to come into Winchester this +morning, but I must be back before three o'clock, for Mr. and +Mrs. Rucastle are going on a visit, and will be away all the +evening, so that I must look after the child. Now I have told you +all my adventures, Mr. Holmes, and I should be very glad if you +could tell me what it all means, and, above all, what I should +do." + +Holmes and I had listened spellbound to this extraordinary story. +My friend rose now and paced up and down the room, his hands in +his pockets, and an expression of the most profound gravity upon +his face. + +"Is Toller still drunk?" he asked. + +"Yes. I heard his wife tell Mrs. Rucastle that she could do +nothing with him." + +"That is well. And the Rucastles go out to-night?" + +"Yes." + +"Is there a cellar with a good strong lock?" + +"Yes, the wine-cellar." + +"You seem to me to have acted all through this matter like a very +brave and sensible girl, Miss Hunter. Do you think that you could +perform one more feat? I should not ask it of you if I did not +think you a quite exceptional woman." + +"I will try. What is it?" + +"We shall be at the Copper Beeches by seven o'clock, my friend +and I. The Rucastles will be gone by that time, and Toller will, +we hope, be incapable. There only remains Mrs. Toller, who might +give the alarm. If you could send her into the cellar on some +errand, and then turn the key upon her, you would facilitate +matters immensely." + +"I will do it." + +"Excellent! We shall then look thoroughly into the affair. Of +course there is only one feasible explanation. You have been +brought there to personate someone, and the real person is +imprisoned in this chamber. That is obvious. As to who this +prisoner is, I have no doubt that it is the daughter, Miss Alice +Rucastle, if I remember right, who was said to have gone to +America. You were chosen, doubtless, as resembling her in height, +figure, and the colour of your hair. Hers had been cut off, very +possibly in some illness through which she has passed, and so, of +course, yours had to be sacrificed also. By a curious chance you +came upon her tresses. The man in the road was undoubtedly some +friend of hers--possibly her fiancé--and no doubt, as you wore +the girl's dress and were so like her, he was convinced from your +laughter, whenever he saw you, and afterwards from your gesture, +that Miss Rucastle was perfectly happy, and that she no longer +desired his attentions. The dog is let loose at night to prevent +him from endeavouring to communicate with her. So much is fairly +clear. The most serious point in the case is the disposition of +the child." + +"What on earth has that to do with it?" I ejaculated. + +"My dear Watson, you as a medical man are continually gaining +light as to the tendencies of a child by the study of the +parents. Don't you see that the converse is equally valid. I have +frequently gained my first real insight into the character of +parents by studying their children. This child's disposition is +abnormally cruel, merely for cruelty's sake, and whether he +derives this from his smiling father, as I should suspect, or +from his mother, it bodes evil for the poor girl who is in their +power." + +"I am sure that you are right, Mr. Holmes," cried our client. "A +thousand things come back to me which make me certain that you +have hit it. Oh, let us lose not an instant in bringing help to +this poor creature." + +"We must be circumspect, for we are dealing with a very cunning +man. We can do nothing until seven o'clock. At that hour we shall +be with you, and it will not be long before we solve the +mystery." + +We were as good as our word, for it was just seven when we +reached the Copper Beeches, having put up our trap at a wayside +public-house. The group of trees, with their dark leaves shining +like burnished metal in the light of the setting sun, were +sufficient to mark the house even had Miss Hunter not been +standing smiling on the door-step. + +"Have you managed it?" asked Holmes. + +A loud thudding noise came from somewhere downstairs. "That is +Mrs. Toller in the cellar," said she. "Her husband lies snoring +on the kitchen rug. Here are his keys, which are the duplicates +of Mr. Rucastle's." + +"You have done well indeed!" cried Holmes with enthusiasm. "Now +lead the way, and we shall soon see the end of this black +business." + +We passed up the stair, unlocked the door, followed on down a +passage, and found ourselves in front of the barricade which Miss +Hunter had described. Holmes cut the cord and removed the +transverse bar. Then he tried the various keys in the lock, but +without success. No sound came from within, and at the silence +Holmes' face clouded over. + +"I trust that we are not too late," said he. "I think, Miss +Hunter, that we had better go in without you. Now, Watson, put +your shoulder to it, and we shall see whether we cannot make our +way in." + +It was an old rickety door and gave at once before our united +strength. Together we rushed into the room. It was empty. There +was no furniture save a little pallet bed, a small table, and a +basketful of linen. The skylight above was open, and the prisoner +gone. + +"There has been some villainy here," said Holmes; "this beauty +has guessed Miss Hunter's intentions and has carried his victim +off." + +"But how?" + +"Through the skylight. We shall soon see how he managed it." He +swung himself up onto the roof. "Ah, yes," he cried, "here's the +end of a long light ladder against the eaves. That is how he did +it." + +"But it is impossible," said Miss Hunter; "the ladder was not +there when the Rucastles went away." + +"He has come back and done it. I tell you that he is a clever and +dangerous man. I should not be very much surprised if this were +he whose step I hear now upon the stair. I think, Watson, that it +would be as well for you to have your pistol ready." + +The words were hardly out of his mouth before a man appeared at +the door of the room, a very fat and burly man, with a heavy +stick in his hand. Miss Hunter screamed and shrunk against the +wall at the sight of him, but Sherlock Holmes sprang forward and +confronted him. + +"You villain!" said he, "where's your daughter?" + +The fat man cast his eyes round, and then up at the open +skylight. + +"It is for me to ask you that," he shrieked, "you thieves! Spies +and thieves! I have caught you, have I? You are in my power. I'll +serve you!" He turned and clattered down the stairs as hard as he +could go. + +"He's gone for the dog!" cried Miss Hunter. + +"I have my revolver," said I. + +"Better close the front door," cried Holmes, and we all rushed +down the stairs together. We had hardly reached the hall when we +heard the baying of a hound, and then a scream of agony, with a +horrible worrying sound which it was dreadful to listen to. An +elderly man with a red face and shaking limbs came staggering out +at a side door. + +"My God!" he cried. "Someone has loosed the dog. It's not been +fed for two days. Quick, quick, or it'll be too late!" + +Holmes and I rushed out and round the angle of the house, with +Toller hurrying behind us. There was the huge famished brute, its +black muzzle buried in Rucastle's throat, while he writhed and +screamed upon the ground. Running up, I blew its brains out, and +it fell over with its keen white teeth still meeting in the great +creases of his neck. With much labour we separated them and +carried him, living but horribly mangled, into the house. We laid +him upon the drawing-room sofa, and having dispatched the sobered +Toller to bear the news to his wife, I did what I could to +relieve his pain. We were all assembled round him when the door +opened, and a tall, gaunt woman entered the room. + +"Mrs. Toller!" cried Miss Hunter. + +"Yes, miss. Mr. Rucastle let me out when he came back before he +went up to you. Ah, miss, it is a pity you didn't let me know +what you were planning, for I would have told you that your pains +were wasted." + +"Ha!" said Holmes, looking keenly at her. "It is clear that Mrs. +Toller knows more about this matter than anyone else." + +"Yes, sir, I do, and I am ready enough to tell what I know." + +"Then, pray, sit down, and let us hear it for there are several +points on which I must confess that I am still in the dark." + +"I will soon make it clear to you," said she; "and I'd have done +so before now if I could ha' got out from the cellar. If there's +police-court business over this, you'll remember that I was the +one that stood your friend, and that I was Miss Alice's friend +too. + +"She was never happy at home, Miss Alice wasn't, from the time +that her father married again. She was slighted like and had no +say in anything, but it never really became bad for her until +after she met Mr. Fowler at a friend's house. As well as I could +learn, Miss Alice had rights of her own by will, but she was so +quiet and patient, she was, that she never said a word about them +but just left everything in Mr. Rucastle's hands. He knew he was +safe with her; but when there was a chance of a husband coming +forward, who would ask for all that the law would give him, then +her father thought it time to put a stop on it. He wanted her to +sign a paper, so that whether she married or not, he could use +her money. When she wouldn't do it, he kept on worrying her until +she got brain-fever, and for six weeks was at death's door. Then +she got better at last, all worn to a shadow, and with her +beautiful hair cut off; but that didn't make no change in her +young man, and he stuck to her as true as man could be." + +"Ah," said Holmes, "I think that what you have been good enough +to tell us makes the matter fairly clear, and that I can deduce +all that remains. Mr. Rucastle then, I presume, took to this +system of imprisonment?" + +"Yes, sir." + +"And brought Miss Hunter down from London in order to get rid of +the disagreeable persistence of Mr. Fowler." + +"That was it, sir." + +"But Mr. Fowler being a persevering man, as a good seaman should +be, blockaded the house, and having met you succeeded by certain +arguments, metallic or otherwise, in convincing you that your +interests were the same as his." + +"Mr. Fowler was a very kind-spoken, free-handed gentleman," said +Mrs. Toller serenely. + +"And in this way he managed that your good man should have no +want of drink, and that a ladder should be ready at the moment +when your master had gone out." + +"You have it, sir, just as it happened." + +"I am sure we owe you an apology, Mrs. Toller," said Holmes, "for +you have certainly cleared up everything which puzzled us. And +here comes the country surgeon and Mrs. Rucastle, so I think, +Watson, that we had best escort Miss Hunter back to Winchester, +as it seems to me that our locus standi now is rather a +questionable one." + +And thus was solved the mystery of the sinister house with the +copper beeches in front of the door. Mr. Rucastle survived, but +was always a broken man, kept alive solely through the care of +his devoted wife. They still live with their old servants, who +probably know so much of Rucastle's past life that he finds it +difficult to part from them. Mr. Fowler and Miss Rucastle were +married, by special license, in Southampton the day after their +flight, and he is now the holder of a government appointment in +the island of Mauritius. As to Miss Violet Hunter, my friend +Holmes, rather to my disappointment, manifested no further +interest in her when once she had ceased to be the centre of one +of his problems, and she is now the head of a private school at +Walsall, where I believe that she has met with considerable success. + + + + + + + + + +End of the Project Gutenberg EBook of The Adventures of Sherlock Holmes, by +Arthur Conan Doyle + +*** END OF THIS PROJECT GUTENBERG EBOOK THE ADVENTURES OF SHERLOCK HOLMES *** + +***** This file should be named 1661-8.txt or 1661-8.zip ***** +This and all associated files of various formats will be found in: + http://www.gutenberg.org/1/6/6/1661/ + +Produced by an anonymous Project Gutenberg volunteer and Jose Menendez + +Updated editions will replace the previous one--the old editions +will be renamed. + +Creating the works from public domain print editions means that no +one owns a United States copyright in these works, so the Foundation +(and you!) can copy and distribute it in the United States without +permission and without paying copyright royalties. Special rules, +set forth in the General Terms of Use part of this license, apply to +copying and distributing Project Gutenberg-tm electronic works to +protect the PROJECT GUTENBERG-tm concept and trademark. Project +Gutenberg is a registered trademark, and may not be used if you +charge for the eBooks, unless you receive specific permission. If you +do not charge anything for copies of this eBook, complying with the +rules is very easy. You may use this eBook for nearly any purpose +such as creation of derivative works, reports, performances and +research. They may be modified and printed and given away--you may do +practically ANYTHING with public domain eBooks. Redistribution is +subject to the trademark license, especially commercial +redistribution. + + + +*** START: FULL LICENSE *** + +THE FULL PROJECT GUTENBERG LICENSE +PLEASE READ THIS BEFORE YOU DISTRIBUTE OR USE THIS WORK + +To protect the Project Gutenberg-tm mission of promoting the free +distribution of electronic works, by using or distributing this work +(or any other work associated in any way with the phrase "Project +Gutenberg"), you agree to comply with all the terms of the Full Project +Gutenberg-tm License (available with this file or online at +http://gutenberg.net/license). + + +Section 1. General Terms of Use and Redistributing Project Gutenberg-tm +electronic works + +1.A. By reading or using any part of this Project Gutenberg-tm +electronic work, you indicate that you have read, understand, agree to +and accept all the terms of this license and intellectual property +(trademark/copyright) agreement. If you do not agree to abide by all +the terms of this agreement, you must cease using and return or destroy +all copies of Project Gutenberg-tm electronic works in your possession. +If you paid a fee for obtaining a copy of or access to a Project +Gutenberg-tm electronic work and you do not agree to be bound by the +terms of this agreement, you may obtain a refund from the person or +entity to whom you paid the fee as set forth in paragraph 1.E.8. + +1.B. "Project Gutenberg" is a registered trademark. It may only be +used on or associated in any way with an electronic work by people who +agree to be bound by the terms of this agreement. There are a few +things that you can do with most Project Gutenberg-tm electronic works +even without complying with the full terms of this agreement. See +paragraph 1.C below. There are a lot of things you can do with Project +Gutenberg-tm electronic works if you follow the terms of this agreement +and help preserve free future access to Project Gutenberg-tm electronic +works. See paragraph 1.E below. + +1.C. The Project Gutenberg Literary Archive Foundation ("the Foundation" +or PGLAF), owns a compilation copyright in the collection of Project +Gutenberg-tm electronic works. Nearly all the individual works in the +collection are in the public domain in the United States. If an +individual work is in the public domain in the United States and you are +located in the United States, we do not claim a right to prevent you from +copying, distributing, performing, displaying or creating derivative +works based on the work as long as all references to Project Gutenberg +are removed. Of course, we hope that you will support the Project +Gutenberg-tm mission of promoting free access to electronic works by +freely sharing Project Gutenberg-tm works in compliance with the terms of +this agreement for keeping the Project Gutenberg-tm name associated with +the work. You can easily comply with the terms of this agreement by +keeping this work in the same format with its attached full Project +Gutenberg-tm License when you share it without charge with others. + +1.D. The copyright laws of the place where you are located also govern +what you can do with this work. Copyright laws in most countries are in +a constant state of change. If you are outside the United States, check +the laws of your country in addition to the terms of this agreement +before downloading, copying, displaying, performing, distributing or +creating derivative works based on this work or any other Project +Gutenberg-tm work. The Foundation makes no representations concerning +the copyright status of any work in any country outside the United +States. + +1.E. Unless you have removed all references to Project Gutenberg: + +1.E.1. The following sentence, with active links to, or other immediate +access to, the full Project Gutenberg-tm License must appear prominently +whenever any copy of a Project Gutenberg-tm work (any work on which the +phrase "Project Gutenberg" appears, or with which the phrase "Project +Gutenberg" is associated) is accessed, displayed, performed, viewed, +copied or distributed: + +This eBook is for the use of anyone anywhere at no cost and with +almost no restrictions whatsoever. You may copy it, give it away or +re-use it under the terms of the Project Gutenberg License included +with this eBook or online at www.gutenberg.net + +1.E.2. If an individual Project Gutenberg-tm electronic work is derived +from the public domain (does not contain a notice indicating that it is +posted with permission of the copyright holder), the work can be copied +and distributed to anyone in the United States without paying any fees +or charges. If you are redistributing or providing access to a work +with the phrase "Project Gutenberg" associated with or appearing on the +work, you must comply either with the requirements of paragraphs 1.E.1 +through 1.E.7 or obtain permission for the use of the work and the +Project Gutenberg-tm trademark as set forth in paragraphs 1.E.8 or +1.E.9. + +1.E.3. If an individual Project Gutenberg-tm electronic work is posted +with the permission of the copyright holder, your use and distribution +must comply with both paragraphs 1.E.1 through 1.E.7 and any additional +terms imposed by the copyright holder. Additional terms will be linked +to the Project Gutenberg-tm License for all works posted with the +permission of the copyright holder found at the beginning of this work. + +1.E.4. Do not unlink or detach or remove the full Project Gutenberg-tm +License terms from this work, or any files containing a part of this +work or any other work associated with Project Gutenberg-tm. + +1.E.5. Do not copy, display, perform, distribute or redistribute this +electronic work, or any part of this electronic work, without +prominently displaying the sentence set forth in paragraph 1.E.1 with +active links or immediate access to the full terms of the Project +Gutenberg-tm License. + +1.E.6. You may convert to and distribute this work in any binary, +compressed, marked up, nonproprietary or proprietary form, including any +word processing or hypertext form. However, if you provide access to or +distribute copies of a Project Gutenberg-tm work in a format other than +"Plain Vanilla ASCII" or other format used in the official version +posted on the official Project Gutenberg-tm web site (www.gutenberg.net), +you must, at no additional cost, fee or expense to the user, provide a +copy, a means of exporting a copy, or a means of obtaining a copy upon +request, of the work in its original "Plain Vanilla ASCII" or other +form. Any alternate format must include the full Project Gutenberg-tm +License as specified in paragraph 1.E.1. + +1.E.7. Do not charge a fee for access to, viewing, displaying, +performing, copying or distributing any Project Gutenberg-tm works +unless you comply with paragraph 1.E.8 or 1.E.9. + +1.E.8. You may charge a reasonable fee for copies of or providing +access to or distributing Project Gutenberg-tm electronic works provided +that + +- You pay a royalty fee of 20% of the gross profits you derive from + the use of Project Gutenberg-tm works calculated using the method + you already use to calculate your applicable taxes. The fee is + owed to the owner of the Project Gutenberg-tm trademark, but he + has agreed to donate royalties under this paragraph to the + Project Gutenberg Literary Archive Foundation. Royalty payments + must be paid within 60 days following each date on which you + prepare (or are legally required to prepare) your periodic tax + returns. Royalty payments should be clearly marked as such and + sent to the Project Gutenberg Literary Archive Foundation at the + address specified in Section 4, "Information about donations to + the Project Gutenberg Literary Archive Foundation." + +- You provide a full refund of any money paid by a user who notifies + you in writing (or by e-mail) within 30 days of receipt that s/he + does not agree to the terms of the full Project Gutenberg-tm + License. You must require such a user to return or + destroy all copies of the works possessed in a physical medium + and discontinue all use of and all access to other copies of + Project Gutenberg-tm works. + +- You provide, in accordance with paragraph 1.F.3, a full refund of any + money paid for a work or a replacement copy, if a defect in the + electronic work is discovered and reported to you within 90 days + of receipt of the work. + +- You comply with all other terms of this agreement for free + distribution of Project Gutenberg-tm works. + +1.E.9. If you wish to charge a fee or distribute a Project Gutenberg-tm +electronic work or group of works on different terms than are set +forth in this agreement, you must obtain permission in writing from +both the Project Gutenberg Literary Archive Foundation and Michael +Hart, the owner of the Project Gutenberg-tm trademark. Contact the +Foundation as set forth in Section 3 below. + +1.F. + +1.F.1. Project Gutenberg volunteers and employees expend considerable +effort to identify, do copyright research on, transcribe and proofread +public domain works in creating the Project Gutenberg-tm +collection. Despite these efforts, Project Gutenberg-tm electronic +works, and the medium on which they may be stored, may contain +"Defects," such as, but not limited to, incomplete, inaccurate or +corrupt data, transcription errors, a copyright or other intellectual +property infringement, a defective or damaged disk or other medium, a +computer virus, or computer codes that damage or cannot be read by +your equipment. + +1.F.2. LIMITED WARRANTY, DISCLAIMER OF DAMAGES - Except for the "Right +of Replacement or Refund" described in paragraph 1.F.3, the Project +Gutenberg Literary Archive Foundation, the owner of the Project +Gutenberg-tm trademark, and any other party distributing a Project +Gutenberg-tm electronic work under this agreement, disclaim all +liability to you for damages, costs and expenses, including legal +fees. YOU AGREE THAT YOU HAVE NO REMEDIES FOR NEGLIGENCE, STRICT +LIABILITY, BREACH OF WARRANTY OR BREACH OF CONTRACT EXCEPT THOSE +PROVIDED IN PARAGRAPH 1.F.3. YOU AGREE THAT THE FOUNDATION, THE +TRADEMARK OWNER, AND ANY DISTRIBUTOR UNDER THIS AGREEMENT WILL NOT BE +LIABLE TO YOU FOR ACTUAL, DIRECT, INDIRECT, CONSEQUENTIAL, PUNITIVE OR +INCIDENTAL DAMAGES EVEN IF YOU GIVE NOTICE OF THE POSSIBILITY OF SUCH +DAMAGE. + +1.F.3. LIMITED RIGHT OF REPLACEMENT OR REFUND - If you discover a +defect in this electronic work within 90 days of receiving it, you can +receive a refund of the money (if any) you paid for it by sending a +written explanation to the person you received the work from. If you +received the work on a physical medium, you must return the medium with +your written explanation. The person or entity that provided you with +the defective work may elect to provide a replacement copy in lieu of a +refund. If you received the work electronically, the person or entity +providing it to you may choose to give you a second opportunity to +receive the work electronically in lieu of a refund. If the second copy +is also defective, you may demand a refund in writing without further +opportunities to fix the problem. + +1.F.4. Except for the limited right of replacement or refund set forth +in paragraph 1.F.3, this work is provided to you 'AS-IS' WITH NO OTHER +WARRANTIES OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO +WARRANTIES OF MERCHANTIBILITY OR FITNESS FOR ANY PURPOSE. + +1.F.5. Some states do not allow disclaimers of certain implied +warranties or the exclusion or limitation of certain types of damages. +If any disclaimer or limitation set forth in this agreement violates the +law of the state applicable to this agreement, the agreement shall be +interpreted to make the maximum disclaimer or limitation permitted by +the applicable state law. The invalidity or unenforceability of any +provision of this agreement shall not void the remaining provisions. + +1.F.6. INDEMNITY - You agree to indemnify and hold the Foundation, the +trademark owner, any agent or employee of the Foundation, anyone +providing copies of Project Gutenberg-tm electronic works in accordance +with this agreement, and any volunteers associated with the production, +promotion and distribution of Project Gutenberg-tm electronic works, +harmless from all liability, costs and expenses, including legal fees, +that arise directly or indirectly from any of the following which you do +or cause to occur: (a) distribution of this or any Project Gutenberg-tm +work, (b) alteration, modification, or additions or deletions to any +Project Gutenberg-tm work, and (c) any Defect you cause. + + +Section 2. Information about the Mission of Project Gutenberg-tm + +Project Gutenberg-tm is synonymous with the free distribution of +electronic works in formats readable by the widest variety of computers +including obsolete, old, middle-aged and new computers. It exists +because of the efforts of hundreds of volunteers and donations from +people in all walks of life. + +Volunteers and financial support to provide volunteers with the +assistance they need are critical to reaching Project Gutenberg-tm's +goals and ensuring that the Project Gutenberg-tm collection will +remain freely available for generations to come. In 2001, the Project +Gutenberg Literary Archive Foundation was created to provide a secure +and permanent future for Project Gutenberg-tm and future generations. +To learn more about the Project Gutenberg Literary Archive Foundation +and how your efforts and donations can help, see Sections 3 and 4 +and the Foundation web page at http://www.pglaf.org. + + +Section 3. Information about the Project Gutenberg Literary Archive +Foundation + +The Project Gutenberg Literary Archive Foundation is a non profit +501(c)(3) educational corporation organized under the laws of the +state of Mississippi and granted tax exempt status by the Internal +Revenue Service. The Foundation's EIN or federal tax identification +number is 64-6221541. Its 501(c)(3) letter is posted at +http://pglaf.org/fundraising. Contributions to the Project Gutenberg +Literary Archive Foundation are tax deductible to the full extent +permitted by U.S. federal laws and your state's laws. + +The Foundation's principal office is located at 4557 Melan Dr. S. +Fairbanks, AK, 99712., but its volunteers and employees are scattered +throughout numerous locations. Its business office is located at +809 North 1500 West, Salt Lake City, UT 84116, (801) 596-1887, email +business@pglaf.org. Email contact links and up to date contact +information can be found at the Foundation's web site and official +page at http://pglaf.org + +For additional contact information: + Dr. Gregory B. Newby + Chief Executive and Director + gbnewby@pglaf.org + + +Section 4. Information about Donations to the Project Gutenberg +Literary Archive Foundation + +Project Gutenberg-tm depends upon and cannot survive without wide +spread public support and donations to carry out its mission of +increasing the number of public domain and licensed works that can be +freely distributed in machine readable form accessible by the widest +array of equipment including outdated equipment. Many small donations +($1 to $5,000) are particularly important to maintaining tax exempt +status with the IRS. + +The Foundation is committed to complying with the laws regulating +charities and charitable donations in all 50 states of the United +States. Compliance requirements are not uniform and it takes a +considerable effort, much paperwork and many fees to meet and keep up +with these requirements. We do not solicit donations in locations +where we have not received written confirmation of compliance. To +SEND DONATIONS or determine the status of compliance for any +particular state visit http://pglaf.org + +While we cannot and do not solicit contributions from states where we +have not met the solicitation requirements, we know of no prohibition +against accepting unsolicited donations from donors in such states who +approach us with offers to donate. + +International donations are gratefully accepted, but we cannot make +any statements concerning tax treatment of donations received from +outside the United States. U.S. laws alone swamp our small staff. + +Please check the Project Gutenberg Web pages for current donation +methods and addresses. Donations are accepted in a number of other +ways including including checks, online payments and credit card +donations. To donate, please visit: http://pglaf.org/donate + + +Section 5. General Information About Project Gutenberg-tm electronic +works. + +Professor Michael S. Hart is the originator of the Project Gutenberg-tm +concept of a library of electronic works that could be freely shared +with anyone. For thirty years, he produced and distributed Project +Gutenberg-tm eBooks with only a loose network of volunteer support. + + +Project Gutenberg-tm eBooks are often created from several printed +editions, all of which are confirmed as Public Domain in the U.S. +unless a copyright notice is included. Thus, we do not necessarily +keep eBooks in compliance with any particular paper edition. + + +Most people start at our Web site which has the main PG search facility: + + http://www.gutenberg.net + +This Web site includes information about Project Gutenberg-tm, +including how to make donations to the Project Gutenberg Literary +Archive Foundation, how to help produce our new eBooks, and how to +subscribe to our email newsletter to hear about new eBooks. diff --git a/vendor/github.com/pierrec/lz4/testdata/pg1661.txt.lz4 b/vendor/github.com/pierrec/lz4/testdata/pg1661.txt.lz4 new file mode 100644 index 000000000..bceb66cd9 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/pg1661.txt.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/testdata/pi.txt b/vendor/github.com/pierrec/lz4/testdata/pi.txt new file mode 100644 index 000000000..ca99bbc2a --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/pi.txt @@ -0,0 +1 @@ +3.1415926535897932384626433832795028841971693993751058209749445923078164062862089986280348253421170679821480865132823066470938446095505822317253594081284811174502841027019385211055596446229489549303819644288109756659334461284756482337867831652712019091456485669234603486104543266482133936072602491412737245870066063155881748815209209628292540917153643678925903600113305305488204665213841469519415116094330572703657595919530921861173819326117931051185480744623799627495673518857527248912279381830119491298336733624406566430860213949463952247371907021798609437027705392171762931767523846748184676694051320005681271452635608277857713427577896091736371787214684409012249534301465495853710507922796892589235420199561121290219608640344181598136297747713099605187072113499999983729780499510597317328160963185950244594553469083026425223082533446850352619311881710100031378387528865875332083814206171776691473035982534904287554687311595628638823537875937519577818577805321712268066130019278766111959092164201989380952572010654858632788659361533818279682303019520353018529689957736225994138912497217752834791315155748572424541506959508295331168617278558890750983817546374649393192550604009277016711390098488240128583616035637076601047101819429555961989467678374494482553797747268471040475346462080466842590694912933136770289891521047521620569660240580381501935112533824300355876402474964732639141992726042699227967823547816360093417216412199245863150302861829745557067498385054945885869269956909272107975093029553211653449872027559602364806654991198818347977535663698074265425278625518184175746728909777727938000816470600161452491921732172147723501414419735685481613611573525521334757418494684385233239073941433345477624168625189835694855620992192221842725502542568876717904946016534668049886272327917860857843838279679766814541009538837863609506800642251252051173929848960841284886269456042419652850222106611863067442786220391949450471237137869609563643719172874677646575739624138908658326459958133904780275900994657640789512694683983525957098258226205224894077267194782684826014769909026401363944374553050682034962524517493996514314298091906592509372216964615157098583874105978859597729754989301617539284681382686838689427741559918559252459539594310499725246808459872736446958486538367362226260991246080512438843904512441365497627807977156914359977001296160894416948685558484063534220722258284886481584560285060168427394522674676788952521385225499546667278239864565961163548862305774564980355936345681743241125150760694794510965960940252288797108931456691368672287489405601015033086179286809208747609178249385890097149096759852613655497818931297848216829989487226588048575640142704775551323796414515237462343645428584447952658678210511413547357395231134271661021359695362314429524849371871101457654035902799344037420073105785390621983874478084784896833214457138687519435064302184531910484810053706146806749192781911979399520614196634287544406437451237181921799983910159195618146751426912397489409071864942319615679452080951465502252316038819301420937621378559566389377870830390697920773467221825625996615014215030680384477345492026054146659252014974428507325186660021324340881907104863317346496514539057962685610055081066587969981635747363840525714591028970641401109712062804390397595156771577004203378699360072305587631763594218731251471205329281918261861258673215791984148488291644706095752706957220917567116722910981690915280173506712748583222871835209353965725121083579151369882091444210067510334671103141267111369908658516398315019701651511685171437657618351556508849099898599823873455283316355076479185358932261854896321329330898570642046752590709154814165498594616371802709819943099244889575712828905923233260972997120844335732654893823911932597463667305836041428138830320382490375898524374417029132765618093773444030707469211201913020330380197621101100449293215160842444859637669838952286847831235526582131449576857262433441893039686426243410773226978028073189154411010446823252716201052652272111660396665573092547110557853763466820653109896526918620564769312570586356620185581007293606598764861179104533488503461136576867532494416680396265797877185560845529654126654085306143444318586769751456614068007002378776591344017127494704205622305389945613140711270004078547332699390814546646458807972708266830634328587856983052358089330657574067954571637752542021149557615814002501262285941302164715509792592309907965473761255176567513575178296664547791745011299614890304639947132962107340437518957359614589019389713111790429782856475032031986915140287080859904801094121472213179476477726224142548545403321571853061422881375850430633217518297986622371721591607716692547487389866549494501146540628433663937900397692656721463853067360965712091807638327166416274888800786925602902284721040317211860820419000422966171196377921337575114959501566049631862947265473642523081770367515906735023507283540567040386743513622224771589150495309844489333096340878076932599397805419341447377441842631298608099888687413260472156951623965864573021631598193195167353812974167729478672422924654366800980676928238280689964004824354037014163149658979409243237896907069779422362508221688957383798623001593776471651228935786015881617557829735233446042815126272037343146531977774160319906655418763979293344195215413418994854447345673831624993419131814809277771038638773431772075456545322077709212019051660962804909263601975988281613323166636528619326686336062735676303544776280350450777235547105859548702790814356240145171806246436267945612753181340783303362542327839449753824372058353114771199260638133467768796959703098339130771098704085913374641442822772634659470474587847787201927715280731767907707157213444730605700733492436931138350493163128404251219256517980694113528013147013047816437885185290928545201165839341965621349143415956258658655705526904965209858033850722426482939728584783163057777560688876446248246857926039535277348030480290058760758251047470916439613626760449256274204208320856611906254543372131535958450687724602901618766795240616342522577195429162991930645537799140373404328752628889639958794757291746426357455254079091451357111369410911939325191076020825202618798531887705842972591677813149699009019211697173727847684726860849003377024242916513005005168323364350389517029893922334517220138128069650117844087451960121228599371623130171144484640903890644954440061986907548516026327505298349187407866808818338510228334508504860825039302133219715518430635455007668282949304137765527939751754613953984683393638304746119966538581538420568533862186725233402830871123282789212507712629463229563989898935821167456270102183564622013496715188190973038119800497340723961036854066431939509790190699639552453005450580685501956730229219139339185680344903982059551002263535361920419947455385938102343955449597783779023742161727111723643435439478221818528624085140066604433258885698670543154706965747458550332323342107301545940516553790686627333799585115625784322988273723198987571415957811196358330059408730681216028764962867446047746491599505497374256269010490377819868359381465741268049256487985561453723478673303904688383436346553794986419270563872931748723320837601123029911367938627089438799362016295154133714248928307220126901475466847653576164773794675200490757155527819653621323926406160136358155907422020203187277605277219005561484255518792530343513984425322341576233610642506390497500865627109535919465897514131034822769306247435363256916078154781811528436679570611086153315044521274739245449454236828860613408414863776700961207151249140430272538607648236341433462351897576645216413767969031495019108575984423919862916421939949072362346468441173940326591840443780513338945257423995082965912285085558215725031071257012668302402929525220118726767562204154205161841634847565169998116141010029960783869092916030288400269104140792886215078424516709087000699282120660418371806535567252532567532861291042487761825829765157959847035622262934860034158722980534989650226291748788202734209222245339856264766914905562842503912757710284027998066365825488926488025456610172967026640765590429099456815065265305371829412703369313785178609040708667114965583434347693385781711386455873678123014587687126603489139095620099393610310291616152881384379099042317473363948045759314931405297634757481193567091101377517210080315590248530906692037671922033229094334676851422144773793937517034436619910403375111735471918550464490263655128162288244625759163330391072253837421821408835086573917715096828874782656995995744906617583441375223970968340800535598491754173818839994469748676265516582765848358845314277568790029095170283529716344562129640435231176006651012412006597558512761785838292041974844236080071930457618932349229279650198751872127267507981255470958904556357921221033346697499235630254947802490114195212382815309114079073860251522742995818072471625916685451333123948049470791191532673430282441860414263639548000448002670496248201792896476697583183271314251702969234889627668440323260927524960357996469256504936818360900323809293459588970695365349406034021665443755890045632882250545255640564482465151875471196218443965825337543885690941130315095261793780029741207665147939425902989695946995565761218656196733786236256125216320862869222103274889218654364802296780705765615144632046927906821207388377814233562823608963208068222468012248261177185896381409183903673672220888321513755600372798394004152970028783076670944474560134556417254370906979396122571429894671543578468788614445812314593571984922528471605049221242470141214780573455105008019086996033027634787081081754501193071412233908663938339529425786905076431006383519834389341596131854347546495569781038293097164651438407007073604112373599843452251610507027056235266012764848308407611830130527932054274628654036036745328651057065874882256981579367897669742205750596834408697350201410206723585020072452256326513410559240190274216248439140359989535394590944070469120914093870012645600162374288021092764579310657922955249887275846101264836999892256959688159205600101655256375678566722796619885782794848855834397518744545512965634434803966420557982936804352202770984294232533022576341807039476994159791594530069752148293366555661567873640053666564165473217043903521329543529169414599041608753201868379370234888689479151071637852902345292440773659495630510074210871426134974595615138498713757047101787957310422969066670214498637464595280824369445789772330048764765241339075920434019634039114732023380715095222010682563427471646024335440051521266932493419673977041595683753555166730273900749729736354964533288869844061196496162773449518273695588220757355176651589855190986665393549481068873206859907540792342402300925900701731960362254756478940647548346647760411463233905651343306844953979070903023460461470961696886885014083470405460742958699138296682468185710318879065287036650832431974404771855678934823089431068287027228097362480939962706074726455399253994428081137369433887294063079261595995462624629707062594845569034711972996409089418059534393251236235508134949004364278527138315912568989295196427287573946914272534366941532361004537304881985517065941217352462589548730167600298865925786628561249665523533829428785425340483083307016537228563559152534784459818313411290019992059813522051173365856407826484942764411376393866924803118364453698589175442647399882284621844900877769776312795722672655562596282542765318300134070922334365779160128093179401718598599933849235495640057099558561134980252499066984233017350358044081168552653117099570899427328709258487894436460050410892266917835258707859512983441729535195378855345737426085902908176515578039059464087350612322611200937310804854852635722825768203416050484662775045003126200800799804925485346941469775164932709504934639382432227188515974054702148289711177792376122578873477188196825462981268685817050740272550263329044976277894423621674119186269439650671515779586756482399391760426017633870454990176143641204692182370764887834196896861181558158736062938603810171215855272668300823834046564758804051380801633638874216371406435495561868964112282140753302655100424104896783528588290243670904887118190909494533144218287661810310073547705498159680772009474696134360928614849417850171807793068108546900094458995279424398139213505586422196483491512639012803832001097738680662877923971801461343244572640097374257007359210031541508936793008169980536520276007277496745840028362405346037263416554259027601834840306811381855105979705664007509426087885735796037324514146786703688098806097164258497595138069309449401515422221943291302173912538355915031003330325111749156969174502714943315155885403922164097229101129035521815762823283182342548326111912800928252561902052630163911477247331485739107775874425387611746578671169414776421441111263583553871361011023267987756410246824032264834641766369806637857681349204530224081972785647198396308781543221166912246415911776732253264335686146186545222681268872684459684424161078540167681420808850280054143613146230821025941737562389942075713627516745731891894562835257044133543758575342698699472547031656613991999682628247270641336222178923903176085428943733935618891651250424404008952719837873864805847268954624388234375178852014395600571048119498842390606136957342315590796703461491434478863604103182350736502778590897578272731305048893989009923913503373250855982655867089242612429473670193907727130706869170926462548423240748550366080136046689511840093668609546325002145852930950000907151058236267293264537382104938724996699339424685516483261134146110680267446637334375340764294026682973865220935701626384648528514903629320199199688285171839536691345222444708045923966028171565515656661113598231122506289058549145097157553900243931535190902107119457300243880176615035270862602537881797519478061013715004489917210022201335013106016391541589578037117792775225978742891917915522417189585361680594741234193398420218745649256443462392531953135103311476394911995072858430658361935369329699289837914941939406085724863968836903265564364216644257607914710869984315733749648835292769328220762947282381537409961545598798259891093717126218283025848112389011968221429457667580718653806506487026133892822994972574530332838963818439447707794022843598834100358385423897354243956475556840952248445541392394100016207693636846776413017819659379971557468541946334893748439129742391433659360410035234377706588867781139498616478747140793263858738624732889645643598774667638479466504074111825658378878454858148962961273998413442726086061872455452360643153710112746809778704464094758280348769758948328241239292960582948619196670918958089833201210318430340128495116203534280144127617285830243559830032042024512072872535581195840149180969253395075778400067465526031446167050827682772223534191102634163157147406123850425845988419907611287258059113935689601431668283176323567325417073420817332230462987992804908514094790368878687894930546955703072619009502076433493359106024545086453628935456862958531315337183868265617862273637169757741830239860065914816164049449650117321313895747062088474802365371031150898427992754426853277974311395143574172219759799359685252285745263796289612691572357986620573408375766873884266405990993505000813375432454635967504844235284874701443545419576258473564216198134073468541117668831186544893776979566517279662326714810338643913751865946730024434500544995399742372328712494834706044063471606325830649829795510109541836235030309453097335834462839476304775645015008507578949548931393944899216125525597701436858943585877526379625597081677643800125436502371412783467926101995585224717220177723700417808419423948725406801556035998390548985723546745642390585850216719031395262944554391316631345308939062046784387785054239390524731362012947691874975191011472315289326772533918146607300089027768963114810902209724520759167297007850580717186381054967973100167870850694207092232908070383263453452038027860990556900134137182368370991949516489600755049341267876436746384902063964019766685592335654639138363185745698147196210841080961884605456039038455343729141446513474940784884423772175154334260306698831768331001133108690421939031080143784334151370924353013677631084913516156422698475074303297167469640666531527035325467112667522460551199581831963763707617991919203579582007595605302346267757943936307463056901080114942714100939136913810725813781357894005599500183542511841721360557275221035268037357265279224173736057511278872181908449006178013889710770822931002797665935838758909395688148560263224393726562472776037890814458837855019702843779362407825052704875816470324581290878395232453237896029841669225489649715606981192186584926770403956481278102179913217416305810554598801300484562997651121241536374515005635070127815926714241342103301566165356024733807843028655257222753049998837015348793008062601809623815161366903341111386538510919367393835229345888322550887064507539473952043968079067086806445096986548801682874343786126453815834280753061845485903798217994599681154419742536344399602902510015888272164745006820704193761584547123183460072629339550548239557137256840232268213012476794522644820910235647752723082081063518899152692889108455571126603965034397896278250016110153235160519655904211844949907789992007329476905868577878720982901352956613978884860509786085957017731298155314951681467176959760994210036183559138777817698458758104466283998806006162298486169353373865787735983361613384133853684211978938900185295691967804554482858483701170967212535338758621582310133103877668272115726949518179589754693992642197915523385766231676275475703546994148929041301863861194391962838870543677743224276809132365449485366768000001065262485473055861598999140170769838548318875014293890899506854530765116803337322265175662207526951791442252808165171667766727930354851542040238174608923283917032754257508676551178593950027933895920576682789677644531840404185540104351348389531201326378369283580827193783126549617459970567450718332065034556644034490453627560011250184335607361222765949278393706478426456763388188075656121689605041611390390639601620221536849410926053876887148379895599991120991646464411918568277004574243434021672276445589330127781586869525069499364610175685060167145354315814801054588605645501332037586454858403240298717093480910556211671546848477803944756979804263180991756422809873998766973237695737015808068229045992123661689025962730430679316531149401764737693873514093361833216142802149763399189835484875625298752423873077559555955465196394401821840998412489826236737714672260616336432964063357281070788758164043814850188411431885988276944901193212968271588841338694346828590066640806314077757725705630729400492940302420498416565479736705485580445865720227637840466823379852827105784319753541795011347273625774080213476826045022851579795797647467022840999561601569108903845824502679265942055503958792298185264800706837650418365620945554346135134152570065974881916341359556719649654032187271602648593049039787489589066127250794828276938953521753621850796297785146188432719223223810158744450528665238022532843891375273845892384422535472653098171578447834215822327020690287232330053862163479885094695472004795231120150432932266282727632177908840087861480221475376578105819702226309717495072127248479478169572961423658595782090830733233560348465318730293026659645013718375428897557971449924654038681799213893469244741985097334626793321072686870768062639919361965044099542167627840914669856925715074315740793805323925239477557441591845821562518192155233709607483329234921034514626437449805596103307994145347784574699992128599999399612281615219314888769388022281083001986016549416542616968586788372609587745676182507275992950893180521872924610867639958916145855058397274209809097817293239301067663868240401113040247007350857828724627134946368531815469690466968693925472519413992914652423857762550047485295476814795467007050347999588867695016124972282040303995463278830695976249361510102436555352230690612949388599015734661023712235478911292547696176005047974928060721268039226911027772261025441492215765045081206771735712027180242968106203776578837166909109418074487814049075517820385653909910477594141321543284406250301802757169650820964273484146957263978842560084531214065935809041271135920041975985136254796160632288736181367373244506079244117639975974619383584574915988097667447093006546342423460634237474666080431701260052055928493695941434081468529815053947178900451835755154125223590590687264878635752541911288877371766374860276606349603536794702692322971868327717393236192007774522126247518698334951510198642698878471719396649769070825217423365662725928440620430214113719922785269984698847702323823840055655517889087661360130477098438611687052310553149162517283732728676007248172987637569816335415074608838663640693470437206688651275688266149730788657015685016918647488541679154596507234287730699853713904300266530783987763850323818215535597323530686043010675760838908627049841888595138091030423595782495143988590113185835840667472370297149785084145853085781339156270760356390763947311455495832266945702494139831634332378975955680856836297253867913275055542524491943589128405045226953812179131914513500993846311774017971512283785460116035955402864405902496466930707769055481028850208085800878115773817191741776017330738554758006056014337743299012728677253043182519757916792969965041460706645712588834697979642931622965520168797300035646304579308840327480771811555330909887025505207680463034608658165394876951960044084820659673794731680864156456505300498816164905788311543454850526600698230931577765003780704661264706021457505793270962047825615247145918965223608396645624105195510522357239739512881816405978591427914816542632892004281609136937773722299983327082082969955737727375667615527113922588055201898876201141680054687365580633471603734291703907986396522961312801782679717289822936070288069087768660593252746378405397691848082041021944719713869256084162451123980620113184541244782050110798760717155683154078865439041210873032402010685341947230476666721749869868547076781205124736792479193150856444775379853799732234456122785843296846647513336573692387201464723679427870042503255589926884349592876124007558756946413705625140011797133166207153715436006876477318675587148783989081074295309410605969443158477539700943988394914432353668539209946879645066533985738887866147629443414010498889931600512076781035886116602029611936396821349607501116498327856353161451684576956871090029997698412632665023477167286573785790857466460772283415403114415294188047825438761770790430001566986776795760909966936075594965152736349811896413043311662774712338817406037317439705406703109676765748695358789670031925866259410510533584384656023391796749267844763708474978333655579007384191473198862713525954625181604342253729962863267496824058060296421146386436864224724887283434170441573482481833301640566959668866769563491416328426414974533349999480002669987588815935073578151958899005395120853510357261373640343675347141048360175464883004078464167452167371904831096767113443494819262681110739948250607394950735031690197318521195526356325843390998224986240670310768318446607291248747540316179699411397387765899868554170318847788675929026070043212666179192235209382278788809886335991160819235355570464634911320859189796132791319756490976000139962344455350143464268604644958624769094347048293294140411146540923988344435159133201077394411184074107684981066347241048239358274019449356651610884631256785297769734684303061462418035852933159734583038455410337010916767763742762102137013548544509263071901147318485749233181672072137279355679528443925481560913728128406333039373562420016045664557414588166052166608738748047243391212955877763906969037078828527753894052460758496231574369171131761347838827194168606625721036851321566478001476752310393578606896111259960281839309548709059073861351914591819510297327875571049729011487171897180046961697770017913919613791417162707018958469214343696762927459109940060084983568425201915593703701011049747339493877885989417433031785348707603221982970579751191440510994235883034546353492349826883624043327267415540301619505680654180939409982020609994140216890900708213307230896621197755306659188141191577836272927461561857103721724710095214236964830864102592887457999322374955191221951903424452307535133806856807354464995127203174487195403976107308060269906258076020292731455252078079914184290638844373499681458273372072663917670201183004648190002413083508846584152148991276106513741539435657211390328574918769094413702090517031487773461652879848235338297260136110984514841823808120540996125274580881099486972216128524897425555516076371675054896173016809613803811914361143992106380050832140987604599309324851025168294467260666138151745712559754953580239983146982203613380828499356705575524712902745397762140493182014658008021566536067765508783804304134310591804606800834591136640834887408005741272586704792258319127415739080914383138456424150940849133918096840251163991936853225557338966953749026620923261318855891580832455571948453875628786128859004106006073746501402627824027346962528217174941582331749239683530136178653673760642166778137739951006589528877427662636841830680190804609849809469763667335662282915132352788806157768278159588669180238940333076441912403412022316368577860357276941541778826435238131905028087018575047046312933353757285386605888904583111450773942935201994321971171642235005644042979892081594307167019857469273848653833436145794634175922573898588001698014757420542995801242958105456510831046297282937584161162532562516572498078492099897990620035936509934721582965174135798491047111660791587436986541222348341887722929446335178653856731962559852026072947674072616767145573649812105677716893484917660771705277187601199908144113058645577910525684304811440261938402322470939249802933550731845890355397133088446174107959162511714864874468611247605428673436709046678468670274091881014249711149657817724279347070216688295610877794405048437528443375108828264771978540006509704033021862556147332117771174413350281608840351781452541964320309576018694649088681545285621346988355444560249556668436602922195124830910605377201980218310103270417838665447181260397190688462370857518080035327047185659499476124248110999288679158969049563947624608424065930948621507690314987020673533848349550836366017848771060809804269247132410009464014373603265645184566792456669551001502298330798496079949882497061723674493612262229617908143114146609412341593593095854079139087208322733549572080757165171876599449856937956238755516175754380917805280294642004472153962807463602113294255916002570735628126387331060058910652457080244749375431841494014821199962764531068006631183823761639663180931444671298615527598201451410275600689297502463040173514891945763607893528555053173314164570504996443890936308438744847839616840518452732884032345202470568516465716477139323775517294795126132398229602394548579754586517458787713318138752959809412174227300352296508089177705068259248822322154938048371454781647213976820963320508305647920482085920475499857320388876391601995240918938945576768749730856955958010659526503036266159750662225084067428898265907510637563569968211510949669744580547288693631020367823250182323708459790111548472087618212477813266330412076216587312970811230758159821248639807212407868878114501655825136178903070860870198975889807456643955157415363193191981070575336633738038272152798849350397480015890519420879711308051233933221903466249917169150948541401871060354603794643379005890957721180804465743962806186717861017156740967662080295766577051291209907944304632892947306159510430902221439371849560634056189342513057268291465783293340524635028929175470872564842600349629611654138230077313327298305001602567240141851520418907011542885799208121984493156999059182011819733500126187728036812481995877070207532406361259313438595542547781961142935163561223496661522614735399674051584998603552953329245752388810136202347624669055816438967863097627365504724348643071218494373485300606387644566272186661701238127715621379746149861328744117714552444708997144522885662942440230184791205478498574521634696448973892062401943518310088283480249249085403077863875165911302873958787098100772718271874529013972836614842142871705531796543076504534324600536361472618180969976933486264077435199928686323835088756683595097265574815431940195576850437248001020413749831872259677387154958399718444907279141965845930083942637020875635398216962055324803212267498911402678528599673405242031091797899905718821949391320753431707980023736590985375520238911643467185582906853711897952626234492483392496342449714656846591248918556629589329909035239233333647435203707701010843880032907598342170185542283861617210417603011645918780539367447472059985023582891833692922337323999480437108419659473162654825748099482509991833006976569367159689364493348864744213500840700660883597235039532340179582557036016936990988671132109798897070517280755855191269930673099250704070245568507786790694766126298082251633136399521170984528092630375922426742575599892892783704744452189363203489415521044597261883800300677617931381399162058062701651024458869247649246891924612125310275731390840470007143561362316992371694848132554200914530410371354532966206392105479824392125172540132314902740585892063217589494345489068463993137570910346332714153162232805522972979538018801628590735729554162788676498274186164218789885741071649069191851162815285486794173638906653885764229158342500673612453849160674137340173572779956341043326883569507814931378007362354180070619180267328551191942676091221035987469241172837493126163395001239599240508454375698507957046222664619000103500490183034153545842833764378111988556318777792537201166718539541835984438305203762819440761594106820716970302285152250573126093046898423433152732131361216582808075212631547730604423774753505952287174402666389148817173086436111389069420279088143119448799417154042103412190847094080254023932942945493878640230512927119097513536000921971105412096683111516328705423028470073120658032626417116165957613272351566662536672718998534199895236884830999302757419916463841427077988708874229277053891227172486322028898425125287217826030500994510824783572905691988555467886079462805371227042466543192145281760741482403827835829719301017888345674167811398954750448339314689630763396657226727043393216745421824557062524797219978668542798977992339579057581890622525473582205236424850783407110144980478726691990186438822932305382318559732869780922253529591017341407334884761005564018242392192695062083183814546983923664613639891012102177095976704908305081854704194664371312299692358895384930136356576186106062228705599423371631021278457446463989738188566746260879482018647487672727222062676465338099801966883680994159075776852639865146253336312450536402610569605513183813174261184420189088853196356986962795036738424313011331753305329802016688817481342988681585577810343231753064784983210629718425184385534427620128234570716988530518326179641178579608888150329602290705614476220915094739035946646916235396809201394578175891088931992112260073928149169481615273842736264298098234063200244024495894456129167049508235812487391799648641133480324757775219708932772262349486015046652681439877051615317026696929704928316285504212898146706195331970269507214378230476875280287354126166391708245925170010714180854800636923259462019002278087409859771921805158532147392653251559035410209284665925299914353791825314545290598415817637058927906909896911164381187809435371521332261443625314490127454772695739393481546916311624928873574718824071503995009446731954316193855485207665738825139639163576723151005556037263394867208207808653734942440115799667507360711159351331959197120948964717553024531364770942094635696982226673775209945168450643623824211853534887989395673187806606107885440005508276570305587448541805778891719207881423351138662929667179643468760077047999537883387870348718021842437342112273940255717690819603092018240188427057046092622564178375265263358324240661253311529423457965569502506810018310900411245379015332966156970522379210325706937051090830789479999004999395322153622748476603613677697978567386584670936679588583788795625946464891376652199588286933801836011932368578558558195556042156250883650203322024513762158204618106705195330653060606501054887167245377942831338871631395596905832083416898476065607118347136218123246227258841990286142087284956879639325464285343075301105285713829643709990356948885285190402956047346131138263878897551788560424998748316382804046848618938189590542039889872650697620201995548412650005394428203930127481638158530396439925470201672759328574366661644110962566337305409219519675148328734808957477775278344221091073111351828046036347198185655572957144747682552857863349342858423118749440003229690697758315903858039353521358860079600342097547392296733310649395601812237812854584317605561733861126734780745850676063048229409653041118306671081893031108871728167519579675347188537229309616143204006381322465841111157758358581135018569047815368938137718472814751998350504781297718599084707621974605887423256995828892535041937958260616211842368768511418316068315867994601652057740529423053601780313357263267054790338401257305912339601880137825421927094767337191987287385248057421248921183470876629667207272325650565129333126059505777727542471241648312832982072361750574673870128209575544305968395555686861188397135522084452852640081252027665557677495969626612604565245684086139238265768583384698499778726706555191854468698469478495734622606294219624557085371272776523098955450193037732166649182578154677292005212667143463209637891852323215018976126034373684067194193037746880999296877582441047878123266253181845960453853543839114496775312864260925211537673258866722604042523491087026958099647595805794663973419064010036361904042033113579336542426303561457009011244800890020801478056603710154122328891465722393145076071670643556827437743965789067972687438473076346451677562103098604092717090951280863090297385044527182892749689212106670081648583395537735919136950153162018908887484210798706899114804669270650940762046502772528650728905328548561433160812693005693785417861096969202538865034577183176686885923681488475276498468821949739729707737187188400414323127636504814531122850990020742409255859252926103021067368154347015252348786351643976235860419194129697690405264832347009911154242601273438022089331096686367898694977994001260164227609260823493041180643829138347354679725399262338791582998486459271734059225620749105308531537182911681637219395188700957788181586850464507699343940987433514431626330317247747486897918209239480833143970840673084079589358108966564775859905563769525232653614424780230826811831037735887089240613031336477371011628214614661679404090518615260360092521947218890918107335871964142144478654899528582343947050079830388538860831035719306002771194558021911942899922722353458707566246926177663178855144350218287026685610665003531050216318206017609217984684936863161293727951873078972637353717150256378733579771808184878458866504335824377004147710414934927438457587107159731559439426412570270965125108115548247939403597681188117282472158250109496096625393395380922195591918188552678062149923172763163218339896938075616855911752998450132067129392404144593862398809381240452191484831646210147389182510109096773869066404158973610476436500068077105656718486281496371118832192445663945814491486165500495676982690308911185687986929470513524816091743243015383684707292898982846022237301452655679898627767968091469798378268764311598832109043715611299766521539635464420869197567370005738764978437686287681792497469438427465256316323005551304174227341646455127812784577772457520386543754282825671412885834544435132562054464241011037955464190581168623059644769587054072141985212106734332410756767575818456990693046047522770167005684543969234041711089888993416350585157887353430815520811772071880379104046983069578685473937656433631979786803671873079693924236321448450354776315670255390065423117920153464977929066241508328858395290542637687668968805033317227800185885069736232403894700471897619347344308437443759925034178807972235859134245813144049847701732361694719765715353197754997162785663119046912609182591249890367654176979903623755286526375733763526969344354400473067198868901968147428767790866979688522501636949856730217523132529265375896415171479559538784278499866456302878831962099830494519874396369070682762657485810439112232618794059941554063270131989895703761105323606298674803779153767511583043208498720920280929752649812569163425000522908872646925284666104665392171482080130502298052637836426959733707053922789153510568883938113249757071331029504430346715989448786847116438328050692507766274500122003526203709466023414648998390252588830148678162196775194583167718762757200505439794412459900771152051546199305098386982542846407255540927403132571632640792934183342147090412542533523248021932277075355546795871638358750181593387174236061551171013123525633485820365146141870049205704372018261733194715700867578539336078622739558185797587258744102542077105475361294047460100094095444959662881486915903899071865980563617137692227290764197755177720104276496949611056220592502420217704269622154958726453989227697660310524980855759471631075870133208861463266412591148633881220284440694169488261529577625325019870359870674380469821942056381255833436421949232275937221289056420943082352544084110864545369404969271494003319782861318186188811118408257865928757426384450059944229568586460481033015388911499486935436030221810943466764000022362550573631294626296096198760564259963946138692330837196265954739234624134597795748524647837980795693198650815977675350553918991151335252298736112779182748542008689539658359421963331502869561192012298889887006079992795411188269023078913107603617634779489432032102773359416908650071932804017163840644987871753756781185321328408216571107549528294974936214608215583205687232185574065161096274874375098092230211609982633033915469494644491004515280925089745074896760324090768983652940657920198315265410658136823791984090645712468948470209357761193139980246813405200394781949866202624008902150166163813538381515037735022966074627952910384068685569070157516624192987244482719429331004854824454580718897633003232525821581280327467962002814762431828622171054352898348208273451680186131719593324711074662228508710666117703465352839577625997744672185715816126411143271794347885990892808486694914139097716736900277758502686646540565950394867841110790116104008572744562938425494167594605487117235946429105850909950214958793112196135908315882620682332156153086833730838173279328196983875087083483880463884784418840031847126974543709373298362402875197920802321878744882872843727378017827008058782410749357514889978911739746129320351081432703251409030487462262942344327571260086642508333187688650756429271605525289544921537651751492196367181049435317858383453865255656640657251363575064353236508936790431702597878177190314867963840828810209461490079715137717099061954969640070867667102330048672631475510537231757114322317411411680622864206388906210192355223546711662137499693269321737043105987225039456574924616978260970253359475020913836673772894438696400028110344026084712899000746807764844088711341352503367877316797709372778682166117865344231732264637847697875144332095340001650692130546476890985050203015044880834261845208730530973189492916425322933612431514306578264070283898409841602950309241897120971601649265613413433422298827909921786042679812457285345801338260995877178113102167340256562744007296834066198480676615805021691833723680399027931606420436812079900316264449146190219458229690992122788553948783538305646864881655562294315673128274390826450611628942803501661336697824051770155219626522725455850738640585299830379180350432876703809252167907571204061237596327685674845079151147313440001832570344920909712435809447900462494313455028900680648704293534037436032625820535790118395649089354345101342969617545249573960621490288728932792520696535386396443225388327522499605986974759882329916263545973324445163755334377492928990581175786355555626937426910947117002165411718219750519831787137106051063795558588905568852887989084750915764639074693619881507814685262133252473837651192990156109189777922008705793396463827490680698769168197492365624226087154176100430608904377976678519661891404144925270480881971498801542057787006521594009289777601330756847966992955433656139847738060394368895887646054983871478968482805384701730871117761159663505039979343869339119789887109156541709133082607647406305711411098839388095481437828474528838368079418884342666222070438722887413947801017721392281911992365405516395893474263953824829609036900288359327745855060801317988407162446563997948275783650195514221551339281978226984278638391679715091262410548725700924070045488485692950448110738087996547481568913935380943474556972128919827177020766613602489581468119133614121258783895577357194986317210844398901423948496659251731388171602663261931065366535041473070804414939169363262373767777095850313255990095762731957308648042467701212327020533742667053142448208168130306397378736642483672539837487690980602182785786216512738563513290148903509883270617258932575363993979055729175160097615459044771692265806315111028038436017374742152476085152099016158582312571590733421736576267142390478279587281505095633092802668458937649649770232973641319060982740633531089792464242134583740901169391964250459128813403498810635400887596820054408364386516617880557608956896727531538081942077332597917278437625661184319891025007491829086475149794003160703845549465385946027452447466812314687943441610993338908992638411847425257044572517459325738989565185716575961481266020310797628254165590506042479114016957900338356574869252800743025623419498286467914476322774005529460903940177536335655471931000175430047504719144899841040015867946179241610016454716551337074073950260442769538553834397550548871099785205401175169747581344926079433689543783221172450687344231989878844128542064742809735625807066983106979935260693392135685881391214807354728463227784908087002467776303605551232386656295178853719673034634701222939581606792509153217489030840886516061119011498443412350124646928028805996134283511884715449771278473361766285062169778717743824362565711779450064477718370221999106695021656757644044997940765037999954845002710665987813603802314126836905783190460792765297277694043613023051787080546511542469395265127101052927070306673024447125973939950514628404767431363739978259184541176413327906460636584152927019030276017339474866960348694976541752429306040727005059039503148522921392575594845078867977925253931765156416197168443524369794447355964260633391055126826061595726217036698506473281266724521989060549880280782881429796336696744124805982192146339565745722102298677599746738126069367069134081559412016115960190237753525556300606247983261249881288192937343476862689219239777833910733106588256813777172328315329082525092733047850724977139448333892552081175608452966590553940965568541706001179857293813998258319293679100391844099286575605993598910002969864460974714718470101531283762631146774209145574041815908800064943237855839308530828305476076799524357391631221886057549673832243195650655460852881201902363644712703748634421727257879503428486312944916318475347531435041392096108796057730987201352484075057637199253650470908582513936863463863368042891767107602111159828875539940120076013947033661793715396306139863655492213741597905119083588290097656647300733879314678913181465109316761575821351424860442292445304113160652700974330088499034675405518640677342603583409608605533747362760935658853109760994238347382222087292464497684560579562516765574088410321731345627735856052358236389532038534024842273371639123973215995440828421666636023296545694703577184873442034227706653837387506169212768015766181095420097708363604361110592409117889540338021426523948929686439808926114635414571535194342850721353453018315875628275733898268898523557799295727645229391567477566676051087887648453493636068278050564622813598885879259940946446041705204470046315137975431737187756039815962647501410906658866162180038266989961965580587208639721176995219466789857011798332440601811575658074284182910615193917630059194314434605154047710570054339000182453117733718955857603607182860506356479979004139761808955363669603162193113250223851791672055180659263518036251214575926238369348222665895576994660491938112486609099798128571823494006615552196112207203092277646200999315244273589488710576623894693889446495093960330454340842102462401048723328750081749179875543879387381439894238011762700837196053094383940063756116458560943129517597713935396074322792489221267045808183313764165818269562105872892447740035947009268662659651422050630078592002488291860839743732353849083964326147000532423540647042089499210250404726781059083644007466380020870126664209457181702946752278540074508552377720890581683918446592829417018288233014971554235235911774818628592967605048203864343108779562892925405638946621948268711042828163893975711757786915430165058602965217459581988878680408110328432739867198621306205559855266036405046282152306154594474489908839081999738747452969810776201487134000122535522246695409315213115337915798026979555710508507473874750758068765376445782524432638046143042889235934852961058269382103498000405248407084403561167817170512813378805705643450616119330424440798260377951198548694559152051960093041271007277849301555038895360338261929343797081874320949914159593396368110627557295278004254863060054523839151068998913578820019411786535682149118528207852130125518518493711503422159542244511900207393539627400208110465530207932867254740543652717595893500716336076321614725815407642053020045340183572338292661915308354095120226329165054426123619197051613839357326693760156914429944943744856809775696303129588719161129294681884936338647392747601226964158848900965717086160598147204467428664208765334799858222090619802173211614230419477754990738738567941189824660913091691772274207233367635032678340586301930193242996397204445179288122854478211953530898910125342975524727635730226281382091807439748671453590778633530160821559911314144205091447293535022230817193663509346865858656314855575862447818620108711889760652969899269328178705576435143382060141077329261063431525337182243385263520217735440715281898137698755157574546939727150488469793619500477720970561793913828989845327426227288647108883270173723258818244658436249580592560338105215606206155713299156084892064340303395262263451454283678698288074251422567451806184149564686111635404971897682154227722479474033571527436819409892050113653400123846714296551867344153741615042563256713430247655125219218035780169240326699541746087592409207004669340396510178134857835694440760470232540755557764728450751826890418293966113310160131119077398632462778219023650660374041606724962490137433217246454097412995570529142438208076098364823465973886691349919784013108015581343979194852830436739012482082444814128095443773898320059864909159505322857914576884962578665885999179867520554558099004556461178755249370124553217170194282884617402736649978475508294228020232901221630102309772151569446427909802190826689868834263071609207914085197695235553488657743425277531197247430873043619511396119080030255878387644206085044730631299277888942729189727169890575925244679660189707482960949190648764693702750773866432391919042254290235318923377293166736086996228032557185308919284403805071030064776847863243191000223929785255372375566213644740096760539439838235764606992465260089090624105904215453927904411529580345334500256244101006359530039598864466169595626351878060688513723462707997327233134693971456285542615467650632465676620279245208581347717608521691340946520307673391841147504140168924121319826881568664561485380287539331160232292555618941042995335640095786495340935115266454024418775949316930560448686420862757201172319526405023099774567647838488973464317215980626787671838005247696884084989185086149003432403476742686245952395890358582135006450998178244636087317754378859677672919526111213859194725451400301180503437875277664402762618941017576872680428176623860680477885242887430259145247073950546525135339459598789619778911041890292943818567205070964606263541732944649576612651953495701860015412623962286413897796733329070567376962156498184506842263690367849555970026079867996261019039331263768556968767029295371162528005543100786408728939225714512481135778627664902425161990277471090335933309304948380597856628844787441469841499067123764789582263294904679812089984857163571087831191848630254501620929805829208334813638405421720056121989353669371336733392464416125223196943471206417375491216357008573694397305979709719726666642267431117762176403068681310351899112271339724036887000996862922546465006385288620393800504778276912835603372548255793912985251506829969107754257647488325341412132800626717094009098223529657957997803018282428490221470748111124018607613415150387569830918652780658896682362523937845272634530420418802508442363190383318384550522367992357752929106925043261446950109861088899914658551881873582528164302520939285258077969737620845637482114433988162710031703151334402309526351929588680690821355853680161000213740851154484912685841268695899174149133820578492800698255195740201818105641297250836070356851055331787840829000041552511865779453963317538532092149720526607831260281961164858098684587525129997404092797683176639914655386108937587952214971731728131517932904431121815871023518740757222100123768721944747209349312324107065080618562372526732540733324875754482967573450019321902199119960797989373383673242576103938985349278777473980508080015544764061053522202325409443567718794565430406735896491017610775948364540823486130254718476485189575836674399791508512858020607820554462991723202028222914886959399729974297471155371858924238493855858595407438104882624648788053304271463011941589896328792678327322456103852197011130466587100500083285177311776489735230926661234588873102883515626446023671996644554727608310118788389151149340939344750073025855814756190881398752357812331342279866503522725367171230756861045004548970360079569827626392344107146584895780241408158405229536937499710665594894459246286619963556350652623405339439142111271810691052290024657423604130093691889255865784668461215679554256605416005071276641766056874274200329577160643448606201239821698271723197826816628249938714995449137302051843669076723577400053932662622760323659751718925901801104290384274185507894887438832703063283279963007200698012244365116394086922220745320244624121155804354542064215121585056896157356414313068883443185280853975927734433655384188340303517822946253702015782157373265523185763554098954033236382319219892171177449469403678296185920803403867575834111518824177439145077366384071880489358256868542011645031357633355509440319236720348651010561049872726472131986543435450409131859513145181276437310438972507004981987052176272494065214619959232142314439776546708351714749367986186552791715824080651063799500184295938799158350171580759883784962257398512129810326379376218322456594236685376799113140108043139732335449090824910499143325843298821033984698141715756010829706583065211347076803680695322971990599904451209087275776225351040902392888779424630483280319132710495478599180196967835321464441189260631526618167443193550817081875477050802654025294109218264858213857526688155584113198560022135158887210365696087515063187533002942118682221893775546027227291290504292259787710667873840000616772154638441292371193521828499824350920891801685572798156421858191197490985730570332667646460728757430565372602768982373259745084479649545648030771598153955827779139373601717422996027353102768719449444917939785144631597314435351850491413941557329382048542123508173912549749819308714396615132942045919380106231421774199184060180347949887691051557905554806953878540066453375981862846419905220452803306263695626490910827627115903856995051246529996062855443838330327638599800792922846659503551211245284087516229060262011857775313747949362055496401073001348853150735487353905602908933526400713274732621960311773433943673385759124508149335736911664541281788171454023054750667136518258284898099512139193995633241336556777098003081910272040997148687418134667006094051021462690280449159646545330107754695413088714165312544813061192407821188690056027781824235022696189344352547633573536485619363254417756613981703930632872166905722259745209192917262199844409646158269456380239502837121686446561785235565164127712826918688615572716201474934052276946595712198314943381622114006936307430444173284786101777743837977037231795255434107223445512555589998646183876764903972461167959018100035098928641204195163551108763204267612979826529425882951141275841262732790798807559751851576841264742209479721843309352972665210015662514552994745127631550917636730259462132930190402837954246323258550301096706922720227074863419005438302650681214142135057154175057508639907673946335146209082888934938376439399256900604067311422093312195936202982972351163259386772241477911629572780752395056251581603133359382311500518626890530658368129988108663263271980611271548858798093487912913707498230575929091862939195014721197586067270092547718025750337730799397134539532646195269996596385654917590458333585799102012713204583903200853878881633637685182083727885131175227769609787962142372162545214591281831798216044111311671406914827170981015457781939202311563871950805024679725792497605772625913328559726371211201905720771409148645074094926718035815157571514050397610963846755569298970383547314100223802583468767350129775413279532060971154506484212185936490997917766874774481882870632315515865032898164228288232746866106592732197907162384642153489852476216789050260998045266483929542357287343977680495774091449538391575565485459058976495198513801007958010783759945775299196700547602252552034453988712538780171960718164078124847847257912407824544361682345239570689514272269750431873633263011103053423335821609333191218806608268341428910415173247216053355849993224548730778822905252324234861531520976938461042582849714963475341837562003014915703279685301868631572488401526639835689563634657435321783493199825542117308467745297085839507616458229630324424328237737450517028560698067889521768198156710781633405266759539424926280756968326107495323390536223090807081455919837355377748742029039018142937311529334644468151212945097596534306284215319445727118614900017650558177095302468875263250119705209476159416768727784472000192789137251841622857783792284439084301181121496366424659033634194540657183544771912446621259392656620306888520055599121235363718226922531781458792593750441448933981608657900876165024635197045828895481793756681046474614105142498870252139936870509372305447734112641354892806841059107716677821238332810262185587751312721179344448201440425745083063944738363793906283008973306241380614589414227694747931665717623182472168350678076487573420491557628217583972975134478990696589532548940335615613167403276472469212505759116251529654568544633498114317670257295661844775487469378464233737238981920662048511894378868224807279352022501796545343757274163910791972952950812942922205347717304184477915673991738418311710362524395716152714669005814700002633010452643547865903290733205468338872078735444762647925297690170912007874183736735087713376977683496344252419949951388315074877537433849458259765560996555954318040920178497184685497370696212088524377013853757681416632722412634423982152941645378000492507262765150789085071265997036708726692764308377229685985169122305037462744310852934305273078865283977335246017463527703205938179125396915621063637625882937571373840754406468964783100704580613446731271591194608435935825987782835266531151065041623295329047772174083559349723758552138048305090009646676088301540612824308740645594431853413755220166305812111033453120745086824339432159043594430312431227471385842030390106070940315235556172767994160020393975099897629335325855575624808996691829864222677502360193257974726742578211119734709402357457222271212526852384295874273501563660093188045493338989741571490544182559738080871565281430102670460284316819230392535297795765862414392701549740879273131051636119137577008929564823323648298263024607975875767745377160102490804624301856524161756655600160859121534556267602192689982855377872583145144082654583484409478463178777374794653580169960779405568701192328608041130904629350871827125934668712766694873899824598527786499569165464029458935064964335809824765965165142090986755203808309203230487342703468288751604071546653834619611223013759451579252696743642531927390036038608236450762698827497618723575476762889950752114804852527950845033958570838130476937881321123674281319487950228066320170022460331989671970649163741175854851878484012054844672588851401562725019821719066960812627785485964818369621410721714214986361918774754509650308957099470934337856981674465828267911940611956037845397855839240761276344105766751024307559814552786167815949657062559755074306521085301597908073343736079432866757890533483669555486803913433720156498834220893399971641479746938696905480089193067138057171505857307148815649920714086758259602876056459782423770242469805328056632787041926768467116266879463486950464507420219373945259262668613552940624781361206202636498199999498405143868285258956342264328707663299304891723400725471764188685351372332667877921738347541480022803392997357936152412755829569276837231234798989446274330454566790062032420516396282588443085438307201495672106460533238537203143242112607424485845094580494081820927639140008540422023556260218564348994145439950410980591817948882628052066441086319001688568155169229486203010738897181007709290590480749092427141018933542818429995988169660993836961644381528877214085268088757488293258735809905670755817017949161906114001908553744882726200936685604475596557476485674008177381703307380305476973609786543859382187220583902344443508867499866506040645874346005331827436296177862518081893144363251205107094690813586440519229512932450078833398788429339342435126343365204385812912834345297308652909783300671261798130316794385535726296998740359570458452230856390098913179475948752126397078375944861139451960286751210561638976008880092746115860800207803341591451797073036835196977766076373785333012024120112046988609209339085365773222392412449051532780950955866459477634482269986074813297302630975028812103517723124465095349653693090018637764094094349837313251321862080214809922685502948454661814715557444709669530177690434272031892770604717784527939160472281534379803539679861424370956683221491465438014593829277393396032754048009552231816667380357183932757077142046723838624617803976292377131209580789363841447929802588065522129262093623930637313496640186619510811583471173312025805866727639992763579078063818813069156366274125431259589936119647626101405563503399523140323113819656236327198961837254845333702062563464223952766943568376761368711962921818754576081617053031590728828700712313666308722754918661395773730546065997437810987649802414011242142773668082751390959313404155826266789510846776118665957660165998178089414985754976284387856100263796543178313634025135814161151902096499133548733131115022700681930135929595971640197196053625033558479980963488718039111612813595968565478868325856437896173159762002419621552896297904819822199462269487137462444729093456470028537694958859591606789282491054412515996300781368367490209374915732896270028656829344431342347351239298259166739503425995868970697267332582735903121288746660451461487850346142827765991608090398652575717263081833494441820193533385071292345774375579344062178711330063106003324053991693682603746176638565758877580201229366353270267100681261825172914608202541892885935244491070138206211553827793565296914576502048643282865557934707209634807372692141186895467322767751335690190153723669036865389161291688887876407525493494249733427181178892759931596719354758988097924525262363659036320070854440784544797348291802082044926670634420437555325050527522833778887040804033531923407685630109347772125639088640413101073817853338316038135280828119040832564401842053746792992622037698718018061122624490909242641985820861751177113789051609140381575003366424156095216328197122335023167422600567941281406217219641842705784328959802882335059828208196666249035857789940333152274817776952843681630088531769694783690580671064828083598046698841098135158654906933319522394363287923990534810987830274500172065433699066117784554364687723631844464768069142828004551074686645392805399409108754939166095731619715033166968309929466349142798780842257220697148875580637480308862995118473187124777291910070227588893486939456289515802965372150409603107761289831263589964893410247036036645058687287589051406841238124247386385427908282733827973326885504935874303160274749063129572349742611221517417153133618622410913869500688835898962349276317316478340077460886655598733382113829928776911495492184192087771606068472874673681886167507221017261103830671787856694812948785048943063086169948798703160515884108282351274153538513365895332948629494495061868514779105804696039069372662670386512905201137810858616188886947957607413585534585151768051973334433495230120395770739623771316030242887200537320998253008977618973129817881944671731160647231476248457551928732782825127182446807824215216469567819294098238926284943760248852279003620219386696482215628093605373178040863727268426696421929946819214908701707533361094791381804063287387593848269535583077395761447997270003472880182785281389503217986345216111066608839314053226944905455527867894417579202440021450780192099804461382547805858048442416404775031536054906591430078158372430123137511562284015838644270890718284816757527123846782459534334449622010096071051370608461801187543120725491334994247617115633321408934609156561550600317384218701570226103101916603887064661438897736318780940711527528174689576401581047016965247557740891644568677717158500583269943401677202156767724068128366565264122982439465133197359199709403275938502669557470231813203243716420586141033606524536939160050644953060161267822648942437397166717661231048975031885732165554988342121802846912529086101485527815277625623750456375769497734336846015607727035509629049392487088406281067943622418704747008368842671022558302403599841645951122485272633632645114017395248086194635840783753556885622317115520947223065437092606797351000565549381224575483728545711797393615756167641692895805257297522338558611388322171107362265816218842443178857488798109026653793426664216990914056536432249301334867988154886628665052346997235574738424830590423677143278792316422403877764330192600192284778313837632536121025336935812624086866699738275977365682227907215832478888642369346396164363308730139814211430306008730666164803678984091335926293402304324974926887831643602681011309570716141912830686577323532639653677390317661361315965553584999398600565155921936759977717933019744688148371103206503693192894521402650915465184309936553493337183425298433679915939417466223900389527673813330617747629574943868716978453767219493506590875711917720875477107189937960894774512654757501871194870738736785890200617373321075693302216320628432065671192096950585761173961632326217708945426214609858410237813215817727602222738133495410481003073275107799948991977963883530734443457532975914263768405442264784216063122769646967156473999043715903323906560726644116438605404838847161912109008701019130726071044114143241976796828547885524779476481802959736049439700479596040292746299203572099761950140348315380947714601056333446998820822120587281510729182971211917876424880354672316916541852256729234429187128163232596965413548589577133208339911288775917226115273379010341362085614577992398778325083550730199818459025958355989260553299673770491722454935329683300002230181517226575787524058832249085821280089747909326100762578770428656006996176212176845478996440705066241710213327486796237430229155358200780141165348065647488230615003392068983794766255036549822805329662862117930628430170492402301985719978948836897183043805182174419147660429752437251683435411217038631379411422095295885798060152938752753799030938871683572095760715221900279379292786303637268765822681241993384808166021603722154710143007377537792699069587121289288019052031601285861825494413353820784883465311632650407642428390870121015194231961652268422003711230464300673442064747718021353070124098860353399152667923871101706221865883573781210935179775604425634694999787251125440854522274810914874307259869602040275941178942581281882159952359658979181144077653354321757595255536158128001163846720319346507296807990793963714961774312119402021297573125165253768017359101557338153772001952444543620071848475663415407442328621060997613243487548847434539665981338717466093020535070271952983943271425371155766600025784423031073429551533945060486222764966687624079324353192992639253731076892135352572321080889819339168668278948281170472624501948409700975760920983724090074717973340788141825195842598096241747610138252643955135259311885045636264188300338539652435997416931322894719878308427600401368074703904097238473945834896186539790594118599310356168436869219485382055780395773881360679549900085123259442529724486666766834641402189915944565309423440650667851948417766779470472041958822043295380326310537494883122180391279678446100139726753892195119117836587662528083690053249004597410947068772912328214304635337283519953648274325833119144459017809607782883583730111857543659958982724531925310588115026307542571493943024453931870179923608166611305426253995833897942971602070338767815033010280120095997252222280801423571094760351925544434929986767817891045559063015953809761875920358937341978962358931125983902598310267193304189215109689156225069659119828323455503059081730735195503721665870288053992138576037035377105178021280129566841984140362872725623214428754302210909472721073474134975514190737043318276626177275996888826027225247133683353452816692779591328861381766349857728936900965749562287103024362590772412219094300871755692625758065709912016659622436080242870024547362036394841255954881727272473653467783647201918303998717627037515724649922289467932322693619177641614618795613956699567783068290316589699430767333508234990790624100202506134057344300695745474682175690441651540636584680463692621274211075399042188716127617787014258864825775223889184599523376292377915585744549477361295525952226578636462118377598473700347971408206994145580719080213590732269233100831759510659019121294795408603640757358750205890208704579670007055262505811420663907459215273309406823649441590891009220296680523325266198911311842016291631076894084723564366808182168657219688268358402785500782804043453710183651096951782335743030504852653738073531074185917705610397395062640355442275156101107261779370634723804990666922161971194259120445084641746383589938239946517395509000859479990136026674261494290066467115067175422177038774507673563742154782905911012619157555870238957001405117822646989944917908301795475876760168094100135837613578591356924455647764464178667115391951357696104864922490083446715486383054477914330097680486878348184672733758436892724310447406807685278625585165092088263813233623148733336714764520450876627614950389949504809560460989604329123358348859990294526400284994280878624039811814884767301216754161106629995553668193123287425702063738352020086863691311733469731741219153633246745325630871347302792174956227014687325867891734558379964351358800959350877556356248810493852999007675135513527792412429277488565888566513247302514710210575352516511814850902750476845518252096331899068527614435138213662152368890578786699432288816028377482035506016029894009119713850179871683633744139275973644017007014763706655703504338121113576415018451821413619823495159601064752712575935185304332875537783057509567425442684712219618709178560783936144511383335649103256405733898667178123972237519316430617013859539474367843392670986712452211189690840236327411496601243483098929941738030588417166613073040067588380432111555379440605497721705942821514886165672771240903387727745629097110134885184374118695655449745736845218066982911045058004299887953899027804383596282409421860556287788428802127553884803728640019441614257499904272009595204654170598104989967504511936471172772220436102614079750809686975176600237187748348016120310234680567112644766123747627852190241202569943534716226660893675219833111813511146503854895025120655772636145473604426859498074396932331297127377157347099713952291182653485155587137336629120242714302503763269501350911612952993785864681307226486008270881333538193703682598867893321238327053297625857382790097826460545598555131836688844628265133798491667839409761353766251798258249663458771950124384040359140849209733754642474488176184070023569580177410177696925077814893386672557898564589851056891960924398841569280696983352240225634570497312245269354193837004843183357196516626721575524193401933099018319309196582920969656247667683659647019595754739345514337413708761517323677204227385674279170698204549953095918872434939524094441678998846319845504852393662972079777452814399418256789457795712552426826089940863317371538896262889629402112108884427376568624527612130371017300785135715404533041507959447776143597437803742436646973247138410492124314138903579092416036406314038149831481905251720937103964026808994832572297954564042701757722904173234796073618787889913318305843069394825961318713816423467218730845133877219086975104942843769325024981656673816260615941768252509993741672883951744066932549653403101452225316189009235376486378482881344209870048096227171226407489571939002918573307460104360729190945767994614929290427981687729426487729952858434647775386906950148984133924540394144680263625402118614317031251117577642829914644533408920976961699098372652361768745605894704968170136974909523072082682887890730190018253425805343421705928713931737993142410852647390948284596418093614138475831136130576108462366837237695913492615824516221552134879244145041756848064120636520170386330129532777699023118648020067556905682295016354931992305914246396217025329747573114094220180199368035026495636955866425906762685687372110339156793839895765565193177883000241613539562437777840801748819373095020699900890899328088397430367736595524891300156633294077907139615464534088791510300651321934486673248275907946807879819425019582622320395131252014109960531260696555404248670549986786923021746989009547850725672978794769888831093487464426400718183160331655511534276155622405474473378049246214952133258527698847336269182649174338987824789278468918828054669982303689939783413747587025805716349413568433929396068192061773331791738208562436433635359863494496890781064019674074436583667071586924521182997893804077137501290858646578905771426833582768978554717687184427726120509266486102051535642840632368481807287940717127966820060727559555904040233178749447346454760628189541512139162918444297651066947969354016866010055196077687335396511614930937570968554559381513789569039251014953265628147011998326992200066392875374713135236421589265126204072887716578358405219646054105435443642166562244565042999010256586927279142752931172082793937751326106052881235373451068372939893580871243869385934389175713376300720319760816604464683937725806909237297523486702916910426369262090199605204121024077648190316014085863558427609537086558164273995349346546314504040199528537252004957805254656251154109252437991326262713609099402902262062836752132305065183934057450112099341464918433323646569371725914489324159006242020612885732926133596808726500045628284557574596592120530341310111827501306961509835515632004310784601906565493806542525229161991819959602752327702249855738824899882707465936355768582560518068964285376850772012220347920993936179268206590142165615925306737944568949070853263568196831861772268249911472615732035807646298116244013316737892788689229032593349861797021994981925739617673075834417098559222170171825712777534491508205278430904619460835217402005838672849709411023266953921445461066215006410674740207009189911951376466904481267253691537162290791385403937560077835153374167747942100384002308951850994548779039346122220865060160500351776264831611153325587705073541279249909859373473787081194253055121436979749914951860535920403830235716352727630874693219622190064260886183676103346002255477477813641012691906569686495012688376296907233961276287223041141813610060264044030035996988919945827397624114613744804059697062576764723766065541618574690527229238228275186799156983390747671146103022776606020061246876477728819096791613354019881402757992174167678799231603963569492851513633647219540611171767387372555728522940054361785176502307544693869307873499110352182532929726044553210797887711449898870911511237250604238753734841257086064069052058452122754533848008205302450456517669518576913200042816758054924811780519832646032445792829730129105318385636821206215531288668564956512613892261367064093953334570526986959692350353094224543865278677673027540402702246384483553239914751363441044050092330361271496081355490531539021002299595756583705381261965683144286057956696622154721695620870013727768536960840704833325132793112232507148630206951245395003735723346807094656483089209801534878705633491092366057554050864111521441481434630437273271045027768661953107858323334857840297160925215326092558932655600672124359464255065996771770388445396181632879614460817789272171836908880126778207430106422524634807454300476492885553409062185153654355474125476152769772667769772777058315801412185688011705028365275543214803488004442979998062157904564161957212784508928489806426497427090579129069217807298769477975112447305991406050629946894280931034216416629935614828130998870745292716048433630818404126469637925843094185442216359084576146078558562473814931427078266215185541603870206876980461747400808324343665382354555109449498431093494759944672673665352517662706772194183191977196378015702169933675083760057163454643671776723387588643405644871566964321041282595645349841388412890420682047007615596916843038999348366793542549210328113363184722592305554383058206941675629992013373175489122037230349072681068534454035993561823576312837767640631013125335212141994611869350833176587852047112364331226765129964171325217513553261867681942338790365468908001827135283584888444111761234101179918709236507184857856221021104009776994453121795022479578069506532965940383987369907240797679040826794007618729547835963492793904576973661643405359792219285870574957481696694062334272619733518136626063735982575552496509807260123668283605928341855848026958413772558970883789942910549800331113884603401939166122186696058491571485733568286149500019097591125218800396419762163559375743718011480559442298730418196808085647265713547612831629200449880315402105530597076666362749328308916880932359290081787411985738317192616728834918402429721290434965526942726402559641463525914348400675867690350382320572934132981593533044446496829441367323442158380761694831219333119819061096142952201536170298575105594326461468505452684975764807808009221335811378197749271768545075538328768874474591593731162470601091244609829424841287520224462594477638749491997840446829257360968534549843266536862844489365704111817793806441616531223600214918768769467398407517176307516849856359201486892943105940202457969622924566644881967576294349535326382171613395757790766370764569570259738800438415805894336137106551859987600754924187211714889295221737721146081154344982665479872580056674724051122007383459271575727715218589946948117940644466399432370044291140747218180224825837736017346685300744985564715420036123593397312914458591522887408719508708632218837288262822884631843717261903305777147651564143822306791847386039147683108141358275755853643597721650028277803713422869688787349795096031108899196143386664068450697420787700280509367203387232629637856038653216432348815557557018469089074647879122436375556668678067610544955017260791142930831285761254481944449473244819093795369008206384631678225064809531810406570254327604385703505922818919878065865412184299217273720955103242251079718077833042609086794273428955735559252723805511440438001239041687716445180226491681641927401106451622431101700056691121733189423400547959684669804298017362570406733282129962153684881404102194463424646220745575643960452985313071409084608499653767803793201899140865814662175319337665970114330608625009829566917638846056762972931464911493704624469351984039534449135141193667933301936617663652555149174982307987072280860859626112660504289296966535652516688885572112276802772743708917389639772257564890533401038855931125679991516589025016486961427207005916056166159702451989051832969278935550303934681219761582183980483960562523091462638447386296039848924386187298507775928792722068554807210497817653286210187476766897248841139560349480376727036316921007350834073865261684507482496448597428134936480372426116704266870831925040997615319076855770327421785010006441984124207396400139603601583810565928413684574119102736420274163723488214524101347716529603128408658419787951116511529827814620379139855006399960326591248525308493690313130100799977191362230866011099929142871249388541612038020411340188887219693477904497527454288072803509305828754420755134816660927879353566521255620139988249628478726214432362853676502591450468377635282587652139156480972141929675549384375582600253168536356731379262475878049445944183429172756988376226261846365452743497662411138451305481449836311789784489732076719508784158618879692955819733250699951402601511675529750575437810242238957925786562128432731202200716730574069286869363930186765958251326499145950260917069347519408975357464016830811798846452473618956056479426358070562563281189269663026479535951097127659136233180866921535788607812759910537171402204506186075374866306350591483916467656723205714516886170790984695932236724946737583099607042589220481550799132752088583781117685214269334786921895240622657921043620348852926267984013953216458791151579050460579710838983371864038024417511347226472547010794793996953554669619726763255229914654933499663234185951450360980344092212206712567698723427940708857070474293173329188523896721971353924492426178641188637790962814486917869468177591717150669111480020759432012061969637795103227089029566085562225452602610460736131368869009281721068198618553780982018471154163630326265699283424155023600978046417108525537612728905335045506135684143775854429677977014660294387687225115363801191758154028120818255606485410787933598921064427244898618961629413418001295130683638609294100083136673372153008352696235737175330738653338204842190308186449184093723944033405244909554558016406460761581010301767488475017661908692946098769201691202181688291040870709560951470416921147027413390052253340834812870353031023919699978597413908593605433599697075604460134242453682496098772581311024732798562072126572499003468293886872304895562253204463602639854225258416464324271611419817802482595563544907219226583863662663750835944314877635156145710745528016159677048442714194435183275698407552677926411261765250615965235457187956673170913319358761628255920783080185206890151504713340386100310055914817852110384754542933389188444120517943969970194112695119526564919594189975418393234647424290702718875223534393673633663200307232747037407123982562024662651974090199762452056198557625760008708173083288344381831070054514493545885422678578551915372292379555494333410174420169600090696415612732297770221217951868376359082255128816470021992348864043959153018464004714321186360622527011541122283802778538911098490201342741014121559769965438877197485376431158229838533123071751132961904559007938064276695819014842627991221792947987348901868471676503827328552059082984529806259250352128451925927986593506132961946796252373972565584157853744567558998032405492186962888490332560851455344391660226257775512916200772796852629387937530454181080729285891989715381797343496187232927614747850192611450413274873242970583408471112333746274617274626582415324271059322506255302314738759251724787322881491455915605036334575424233779160374952502493022351481961381162563911415610326844958072508273431765944054098269765269344579863479709743124498271933113863873159636361218623497261409556079920628316999420072054811525353393946076850019909886553861433495781650089961649079678142901148387645682174914075623767618453775144031475411206760160726460556859257799322070337333398916369504346690694828436629980037414527627716547623825546170883189810868806847853705536480469350958818025360529740793538676511195079373282083146268960071075175520614433784114549950136432446328193346389050936545714506900864483440180428363390513578157273973334537284263372174065775771079830517555721036795976901889958494130195999573017901240193908681356585539661941371794487632079868800371607303220547423572266896801882123424391885984168972277652194032493227314793669234004848976059037958094696041754279613782553781223947646147832926976545162290281701100437846038756544151739433960048915318817576650500951697402415644771293656614253949368884230517400129920556854289853897942669956777027089146513736892206104415481662156804219838476730871787590279209175900695273456682026513373111518000181434120962601658629821076663523361774007837783423709152644063054071807843358061072961105550020415131696373046849213356837265400307509829089364612047891114753037049893952833457824082817386441322710002968311940203323456420826473276233830294639378998375836554559919340866235090967961134004867027123176526663710778725111860354037554487418693519733656621772359229396776463251562023487570113795712096237723431370212031004965152111976013176419408203437348512852602913334915125083119802850177855710725373149139215709105130965059885999931560863655477403551898166733535880048214665099741433761182777723351910741217572841592580872591315074606025634903777263373914461377038021318347447301113032670296917335047701632106616227830027269283365584011791419447808748253360714403296252285775009808599609040936312635621328162071453406104224112083010008587264252112262480142647519426184325853386753874054743491072710049754281159466017136122590440158991600229827801796035194080046513534752698777609527839984368086908989197839693532179980139135442552717910225397010810632143048511378291498511381969143043497500189980681644412123273328307192824362406733196554692677851193152775113446468905504248113361434984604849051258345683266441528489713972376040328212660253516693914082049947320486021627759791771234751097502403078935759937715095021751693555827072533911892334070223832077585802137174778378778391015234132098489423459613692340497998279304144463162707214796117456975719681239291913740982925805561955207434243295982898980529233366415419256367380689494201471241340525072204061794355252555225008748790086568314542835167750542294803274783044056438581591952666758282929705226127628711040134801787224801789684052407924360582742467443076721645270313451354167649668901274786801010295133862698649748212118629040337691568576240699296372493097201628707200189835423690364149270236961938547372480329855045112089192879829874467864129159417531675602533435310626745254507114181483239880607297140234725520713490798398982355268723950909365667878992383712578976248755990443228895388377317348941122757071410959790047919301046740750411435381782464630795989555638991884773781341347070246747362112048986226991888517456251732519341352038115863350123913054441910073628447567514161050410973505852762044489190978901984315485280533985777844313933883994310444465669244550885946314081751220331390681596592510546858013133838152176418210433429788826119630443111388796258746090226130900849975430395771243230616906262919403921439740270894777663702488155499322458825979020631257436910946393252806241642476868495455324938017639371615636847859823715902385421265840615367228607131702674740131145261063765383390315921943469817605358380310612887852051546933639241088467632009567089718367490578163085158138161966882222047570437590614338040725853862083565176998426774523195824182683698270160237414938363496629351576854061397342746470899685618170160551104880971554859118617189668025973541705423985135560018720335079060946421271143993196046527424050882225359773481519135438571253258540493946010865793798058620143366078825219717809025817370870916460452727977153509910340736425020386386718220522879694458387652947951048660717390229327455426785669776865939923416834122274663015062155320502655341460995249356050854921756549134830958906536175693817637473644183378974229700703545206663170929607591989627732423090252397443861014263098687733913882518684316501027964911497737582888913450341148865948670215492101084328080783428089417298008983297536940644969903125399863919581601468995220880662285408414864274786281975546629278814621607171381880180840572084715868906836919393381864278454537956719272397972364651667592011057995663962598535512763558768140213409829016296873429850792471846056874828331381259161962476156902875901072733103299140623864608333378638257926302391590003557609032477281338887339178096966601469615031754226751125993315529674213336300222964906480934582008181061802100227664580400278213336758573019011371754672763059044353131319036092489097246427928455549913490005180295707082919052556781889913899625138662319380053611346224294610248954072404857123256628888931722116432947816190554868054943441034090680716088028227959686950133643814268252170472870863010137301155236861416908375675747637239763185757038109443390564564468524183028148107998376918512127201935044041804604721626939445788377090105974693219720558114078775989772072009689382249303236830515862657281114637996983137517937623215111252349734305240622105244234353732905655163406669506165892878218707756794176080712973781335187117931650033155523822487730653444179453415395202424449703410120874072188109388268167512042299404948179449472732894770111574139441228455521828424922240658752689172272780607116754046973008037039618787796694882555614674384392570115829546661358678671897661297311267200072971553613027503556167817765442287442114729881614802705243806817653573275578602505847084013208837932816008769081300492491473682517035382219619039014999523495387105997351143478292339499187936608692301375596368532373806703591144243268561512109404259582639301678017128669239283231057658851714020211196957064799814031505633045141564414623163763809904402816256917576489142569714163598439317433270237812336938043012892626375382667795034169334323607500248175741808750388475094939454896209740485442635637164995949920980884294790363666297526003243856352945844728944547166209297495496616877414120882130477022816116456044007236351581149729739218966737382647204722642221242016560150284971306332795814302516013694825567014780935790889657134926158161346901806965089556310121218491805847922720691871696316330044858020102860657858591269974637661741463934159569539554203314628026518951167938074573315759846086173702687867602943677780500244673391332431669880354073232388281847501051641331189537036488422690270478052742490603492082954755054003457160184072574536938145531175354210726557835615499874447480427323457880061873149341566046352979779455075359304795687209316724536547208381685855606043801977030764246083489876101345709394877002946175792061952549255757109038525171488525265671045349813419803390641529876343695420256080277614421914318921393908834543131769685101840103844472348948869520981943531906506555354617335814045544837884752526253949665869992058417652780125341033896469818642430034146791380619028059607854888010789705516946215228773090104467462497979992627120951684779568482583341402266477210843362437593741610536734041954738964197895425335036301861400951534766961476255651873823292468547356935802896011536791787303553159378363082248615177770541577576561759358512016692943111138863582159667618830326104164651714846979385422621687161400122378213779774131268977266712992025922017408770076956283473932201088159356286281928563571893384958850603853158179760679479840878360975960149733420572704603521790605647603285569276273495182203236144112584182426247712012035776388895974318232827871314608053533574494297621796789034568169889553518504478325616380709476951699086247100019748809205009521943632378719764870339223811540363475488626845956159755193765410115014067001226927474393888589943859730245414801061235908036274585288493563251585384383242493252666087588908318700709100237377106576985056433928854337658342596750653715005333514489908293887737352051459333049626531415141386124437935885070944688045486975358170212908490787347806814366323322819415827345671356443171537967818058195852464840084032909981943781718177302317003989733050495387356116261023999433259780126893432605584710278764901070923443884634011735556865903585244919370181041626208504299258697435817098133894045934471937493877624232409852832762266604942385129709453245586252103600829286649724174919141988966129558076770979594795306013119159011773943104209049079424448868513086844493705909026006120649425744710353547657859242708130410618546219881830090634588187038755856274911587375421064667951346487586771543838018521348281915812462599335160198935595167968932852205824799421034512715877163345222995418839680448835529753361286837225935390079201666941339091168758803988828869216002373257361588207163516271332810518187602104852180675526648673908900907195138058626735124312215691637902277328705410842037841525683288718046987952513073266340278519059417338920358540395677035611329354482585628287610610698229721420961993509331312171187891078766872044548876089410174798647137882462153955933333275562009439580434537919782280590395959927436913793778664940964048777841748336432684026282932406260081908081804390914556351936856063045089142289645219987798849347477729132797266027658401667890136490508741142126861969862044126965282981087045479861559545338021201155646979976785738920186243599326777689454060508218838227909833627167124490026761178498264377033002081844590009717235204331994708242098771514449751017055643029542821819670009202515615844174205933658148134902693111517093872260026458630561325605792560927332265579346280805683443921373688405650434307396574061017779370141424615493070741360805442100295600095663588977899267630517718781943706761498217564186590116160865408635391513039201316805769034172596453692350806417446562351523929050409479953184074862151210561833854566176652606393713658802521666223576132201941701372664966073252010771947931265282763302413805164907174565964853748354669194523580315301969160480994606814904037819829732360930087135760798621425422096419004367905479049930078372421581954535418371129368658430553842717628035279128821129308351575656599944741788438381565148434229858704245592434693295232821803508333726283791830216591836181554217157448465778420134329982594566884558266171979012180849480332448787258183774805522268151011371745368417870280274452442905474518234674919564188551244421337783521423865979925988203287085109338386829906571994614906290257427686038850511032638544540419184958866538545040571323629681069146814847869659166861842756798460041868762298055562963045953227923051616721591968675849523635298935788507746081537321454642984792310511676357749494622952569497660359473962430995343310404994209677883827002714478494069037073249106444151696053256560586778757417472110827435774315194060757983563629143326397812218946287447798119807225646714664054850131009656786314880090303749338875364183165134982546694673316118123364854397649325026179549357204305402182974871251107404011611405899911093062492312813116340549262571356721818628932786138833718028535056503591952741400869510926167541476792668032109237467087213606278332922386413619594121339278036118276324106004740971111048140003623342714514483334641675466354699731494756643423659493496845884551524150756376605086632827424794136062876041290644913828519456402643153225858624043141838669590633245063000392213192647625962691510904457695301444054618037857503036686212462278639752746667870121003392984873375014475600322100622358029343774955032037012738468163061026570300872275462966796880890587127676361066225722352229739206443093524327228100859973095132528630601105497915644791845004618046762408928925680912930592960642357021061524646205023248966593987324933967376952023991760898474571843531936646529125848064480196520162838795189499336759241485626136995945307287254532463291529110128763770605570609531377527751867923292134955245133089867969165129073841302167573238637575820080363575728002754490327953079900799442541108725693188014667935595834676432868876966610097395749967836593397846346959948950610490383647409504695226063858046758073069912290474089879166872117147527644711604401952718169508289733537148530928937046384420893299771125856840846608339934045689026787516008775461267988015465856522061210953490796707365539702576199431376639960606061106406959330828171876426043573425361756943784848495250108266488395159700490598380812105221111091943323951136051446459834210799058082093716464523127704023160072138543723461267260997870385657091998507595634613248460188409850194287687902268734556500519121546544063829253851276317663922050938345204300773017029940362615434001322763910912988327863920412300445551684054889809080779174636092439334912641164240093880746356607262336695842764583698268734815881961058571835767462009650526065929263548291499045768307210893245857073701660717398194485028842603963660746031184786225831056580870870305567595861341700745402965687634774176431051751036732869245558582082372038601781739405175130437994868822320044378043103170921034261674998000073016094814586374488778522273076330495383944345382770608760763542098445008306247630253572781032783461766970544287155315340016497076657195985041748199087201490875686037783591994719343352772947285537925787684832301101859365800717291186967617655053775030293033830706448912811412025506150896411007623824574488655182581058140345320124754723269087547507078577659732542844459353044992070014538748948226556442223696365544194225441338212225477497535494624827680533336983284156138692363443358553868471111430498248398991803165458638289353799130535222833430137953372954016257623228081138499491876144141322933767106563492528814528239506209022357876684650116660097382753660405446941653422239052108314585847035529352219928272760574821266065291385530345549744551470344939486863429459658431024190785923680224560763936784166270518555178702904073557304620639692453307795782245949710420188043000183881429008173039450507342787013124466860092778581811040911511729374873627887874907465285565434748886831064110051023020875107768918781525622735251550379532444857787277617001964853703555167655209119339343762866284619844026295252183678522367475108809781507098978413086245881522660963551401874495836926917799047120726494905737264286005211403581231076006699518536124862746756375896225299116496066876508261734178484789337295056739007878617925351440621045366250640463728815698232317500596261080921955211150859302955654967538862612972339914628358476048627627027309739202001432248707582337354915246085608210328882974183906478869923273691360048837436615223517058437705545210815513361262142911815615301758882573594892507108879262128641392443309383797333867806131795237315266773820858024701433527009243803266951742119507670884326346442749127558907746863582162166042741315170212458586056233631493164646913946562497471741958354218607748711057338458433689939645913740603382159352243594751626239188685307822821763983237306180204246560477527943104796189724299533029792497481684052893791044947004590864991872727345413508101983881864673609392571930511968645601855782450218231065889437986522432050677379966196955472440585922417953006820451795370043472451762893566770508490213107736625751697335527462302943031203596260953423574397249659211010657817826108745318874803187430823573699195156340957162700992444929749105489851519658664740148225106335367949737142510229341882585117371994499115097583746130105505064197721531929354875371191630262030328588658528480193509225875775597425276584011721342323648084027143356367542046375182552524944329657043861387865901965738802868401894087672816714137033661732650120578653915780703088714261519075001492576112927675193096728453971160213606303090542243966320674323582797889332324405779199278484633339777737655901870574806828678347965624146102899508487399692970750432753029972872297327934442988646412725348160603779707298299173029296308695801996312413304939350493325412355071054461182591141116454534710329881047844067780138077131465400099386306481266614330858206811395838319169545558259426895769841428893743467084107946318932539106963955780706021245974898293564613560788983472419979478564362042094613412387613198865352358312996862268948608408456655606876954501274486631405054735351746873009806322780468912246821460806727627708402402266155485024008952891657117617439020337584877842911289623247059191874691042005848326140677333751027195653994697162517248312230633919328707983800748485726516123434933273356664473358556430235280883924348278760886164943289399166399210488307847777048045728491456303353265070029588906265915498509407972767567129795010098229476228961891591441520032283878773485130979081019129267227103778898053964156362364169154985768408398468861684375407065121039062506128107663799047908879674778069738473170475253442156390387201238806323688037017949308954900776331523063548374256816653361606641980030188287123767481898330246836371488309259283375902278942588060087286038859168849730693948020511221766359138251524278670094406942355120201568377778851824670025651708509249623747726813694284350062938814429987905301056217375459182679973217735029368928065210025396268807498092643458011655715886700443503976505323478287327368840863540002740676783821963522226539290939807367391364082898722017776747168118195856133721583119054682936083236976113450281757830202934845982925000895682630271263295866292147653142233351793093387951357095346377183684092444422096319331295620305575517340067973740614162107923633423805646850092037167152642556371853889571416419772387422610596667396997173168169415435095283193556417705668622215217991151355639707143312893657553844648326201206424338016955862698561022460646069330793847858814367407000599769703649019273328826135329363112403650698652160638987250267238087403396744397830258296894256896741864336134979475245526291426522842419243083388103580053787023999542172113686550275341362211693140694669513186928102574795985605145005021715913317751609957865551981886193211282110709442287240442481153406055895958355815232012184605820563592699303478851132068626627588771446035996656108430725696500563064489187599466596772847171539573612108180841547273142661748933134174632662354222072600146012701206934639520564445543291662986660783089068118790090815295063626782075614388815781351134695366303878412092346942868730839320432333872775496805210302821544324723388845215343727250128589747691460808314404125868181540049187772287869801853454537006526655649170915429522756709222217474112062720656622989806032891672068743654948246108697367225547404812889242471854323605753411672850757552057131156697954584887398742228135887985840783135060548290551482785294891121905383195624228719484759407859398047901094194070671764439032730712135887385049993638838205501683402777496070276844880281912220636888636811043569529300652195528261526991271637277388418993287130563464688227398288763198645709836308917786487086676185485680047672552675414742851028145807403152992197814557756843681110185317498167016426647884090262682824448258027532094549915104518517716546311804904567985713257528117913656278158111288816562285876030875974963849435275676612168959261485030785362045274507752950631012480341804584059432926079854435620093708091821523920371790678121992280496069738238743312626730306795943960954957189577217915597300588693646845576676092450906088202212235719254536715191834872587423919410890444115959932760044506556206461164655665487594247369252336955993030355095817626176231849561906494839673002037763874369343999829430209147073618947932692762445186560239559053705128978163455423320114975994896278424327483788032701418676952621180975006405149755889650293004867605208010491537885413909424531691719987628941277221129464568294860281493181560249677887949813777216229359437811004448060797672429276249510784153446429150842764520002042769470698041775832209097020291657347251582904630910359037842977572651720877244740952267166306005469716387943171196873484688738186656751279298575016363411314627530499019135646823804329970695770150789337728658035712790913767420805655493624646 diff --git a/vendor/github.com/pierrec/lz4/testdata/pi.txt.lz4 b/vendor/github.com/pierrec/lz4/testdata/pi.txt.lz4 new file mode 100644 index 000000000..36accc5e1 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/pi.txt.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/testdata/random.data b/vendor/github.com/pierrec/lz4/testdata/random.data new file mode 100644 index 000000000..713184f44 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/random.data differ diff --git a/vendor/github.com/pierrec/lz4/testdata/random.data.lz4 b/vendor/github.com/pierrec/lz4/testdata/random.data.lz4 new file mode 100644 index 000000000..40b92699e Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/random.data.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/testdata/repeat.txt b/vendor/github.com/pierrec/lz4/testdata/repeat.txt new file mode 100644 index 000000000..679764d40 --- /dev/null +++ b/vendor/github.com/pierrec/lz4/testdata/repeat.txt @@ -0,0 +1 @@ +abcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyzabcdefghijklmnopqrstuvzxyz \ No newline at end of file diff --git a/vendor/github.com/pierrec/lz4/testdata/repeat.txt.lz4 b/vendor/github.com/pierrec/lz4/testdata/repeat.txt.lz4 new file mode 100644 index 000000000..0780e3696 Binary files /dev/null and b/vendor/github.com/pierrec/lz4/testdata/repeat.txt.lz4 differ diff --git a/vendor/github.com/pierrec/lz4/writer.go b/vendor/github.com/pierrec/lz4/writer.go index b1b712fe2..62a561703 100644 --- a/vendor/github.com/pierrec/lz4/writer.go +++ b/vendor/github.com/pierrec/lz4/writer.go @@ -3,21 +3,22 @@ package lz4 import ( "encoding/binary" "fmt" - "hash" "io" - "runtime" + + "github.com/pierrec/lz4/internal/xxh32" ) // Writer implements the LZ4 frame encoder. type Writer struct { Header - dst io.Writer - checksum hash.Hash32 // frame checksum - data []byte // data to be compressed, only used when dealing with block dependency as we need 64Kb to work with - window []byte // last 64KB of decompressed data (block dependency) + blockMaxSize buffer - zbCompressBuf []byte // buffer for compressing lz4 blocks - writeSizeBuf []byte // four-byte slice for writing checksums and sizes in writeblock + buf [19]byte // magic number(4) + header(flags(2)+[Size(8)+DictID(4)]+checksum(1)) does not exceed 19 bytes + dst io.Writer // Destination. + checksum xxh32.XXHZero // Frame checksum. + zdata []byte // Compressed data. + data []byte // Data to be compressed. + idx int // Index into data. + hashtable [winSize]int // Hash table used in CompressBlock(). } // NewWriter returns a new LZ4 frame encoder. @@ -25,38 +26,37 @@ type Writer struct { // The supplied Header is checked at the first Write. // It is ok to change it before the first Write but then not until a Reset() is performed. func NewWriter(dst io.Writer) *Writer { - return &Writer{ - dst: dst, - checksum: hashPool.Get(), - Header: Header{ - BlockMaxSize: 4 << 20, - }, - writeSizeBuf: make([]byte, 4), - } + return &Writer{dst: dst} } // writeHeader builds and writes the header (magic+header) to the underlying io.Writer. func (z *Writer) writeHeader() error { - // Default to 4Mb if BlockMaxSize is not set + // Default to 4Mb if BlockMaxSize is not set. if z.Header.BlockMaxSize == 0 { - z.Header.BlockMaxSize = 4 << 20 + z.Header.BlockMaxSize = bsMapID[7] } - // the only option that need to be validated - bSize, ok := bsMapValue[z.Header.BlockMaxSize] + // The only option that needs to be validated. + bSize := z.Header.BlockMaxSize + bSizeID, ok := bsMapValue[bSize] if !ok { - return fmt.Errorf("lz4: invalid block max size: %d", z.Header.BlockMaxSize) + return fmt.Errorf("lz4: invalid block max size: %d", bSize) } + // Allocate the compressed/uncompressed buffers. + // The compressed buffer cannot exceed the uncompressed one. + if n := 2 * bSize; cap(z.zdata) < n { + z.zdata = make([]byte, n, n) + } + z.zdata = z.zdata[:bSize] + z.data = z.zdata[:cap(z.zdata)][bSize:] + z.idx = 0 - // magic number(4) + header(flags(2)+[Size(8)+DictID(4)]+checksum(1)) does not exceed 19 bytes - // Size and DictID are optional - var buf [19]byte + // Size is optional. + buf := z.buf[:] - // set the fixed size data: magic number, block max size and flags + // Set the fixed size data: magic number, block max size and flags. binary.LittleEndian.PutUint32(buf[0:], frameMagic) flg := byte(Version << 6) - if !z.Header.BlockDependency { - flg |= 1 << 5 - } + flg |= 1 << 5 // No block dependency. if z.Header.BlockChecksum { flg |= 1 << 4 } @@ -66,254 +66,126 @@ func (z *Writer) writeHeader() error { if !z.Header.NoChecksum { flg |= 1 << 2 } - // if z.Header.Dict { - // flg |= 1 - // } buf[4] = flg - buf[5] = bSize << 4 + buf[5] = bSizeID << 4 - // current buffer size: magic(4) + flags(1) + block max size (1) + // Current buffer size: magic(4) + flags(1) + block max size (1). n := 6 - // optional items + // Optional items. if z.Header.Size > 0 { binary.LittleEndian.PutUint64(buf[n:], z.Header.Size) n += 8 } - // if z.Header.Dict { - // binary.LittleEndian.PutUint32(buf[n:], z.Header.DictID) - // n += 4 - // } - // header checksum includes the flags, block max size and optional Size and DictID - z.checksum.Write(buf[4:n]) - buf[n] = byte(z.checksum.Sum32() >> 8 & 0xFF) + // The header checksum includes the flags, block max size and optional Size. + buf[n] = byte(xxh32.ChecksumZero(buf[4:n]) >> 8 & 0xFF) z.checksum.Reset() - // header ready, write it out + // Header ready, write it out. if _, err := z.dst.Write(buf[0 : n+1]); err != nil { return err } z.Header.done = true - - // initialize buffers dependent on header info - z.zbCompressBuf = make([]byte, winSize+z.BlockMaxSize) + debug("wrote header %v", z.Header) return nil } // Write compresses data from the supplied buffer into the underlying io.Writer. // Write does not return until the data has been written. -// -// If the input buffer is large enough (typically in multiples of BlockMaxSize) -// the data will be compressed concurrently. -// -// Write never buffers any data unless in BlockDependency mode where it may -// do so until it has 64Kb of data, after which it never buffers any. -func (z *Writer) Write(buf []byte) (n int, err error) { +func (z *Writer) Write(buf []byte) (int, error) { if !z.Header.done { - if err = z.writeHeader(); err != nil { - return + if err := z.writeHeader(); err != nil { + return 0, err } } - if len(buf) == 0 { - return - } - if !z.NoChecksum { z.checksum.Write(buf) } + debug("input buffer len=%d index=%d", len(buf), z.idx) - // with block dependency, require at least 64Kb of data to work with - // not having 64Kb only matters initially to setup the first window - bl := 0 - if z.BlockDependency && len(z.window) == 0 { - bl = len(z.data) - z.data = append(z.data, buf...) - if len(z.data) < winSize { - return len(buf), nil - } - buf = z.data - z.data = nil - } - - // Break up the input buffer into BlockMaxSize blocks, provisioning the left over block. - // Then compress into each of them concurrently if possible (no dependency). - var ( - zb block - wbuf = buf - zn = len(wbuf) / z.BlockMaxSize - zi = 0 - leftover = len(buf) % z.BlockMaxSize - ) - -loop: - for zi < zn { - if z.BlockDependency { - if zi == 0 { - // first block does not have the window - zb.data = append(z.window, wbuf[:z.BlockMaxSize]...) - zb.offset = len(z.window) - wbuf = wbuf[z.BlockMaxSize-winSize:] - } else { - // set the uncompressed data including the window from previous block - zb.data = wbuf[:z.BlockMaxSize+winSize] - zb.offset = winSize - wbuf = wbuf[z.BlockMaxSize:] - } - } else { - zb.data = wbuf[:z.BlockMaxSize] - wbuf = wbuf[z.BlockMaxSize:] - } - - goto write - } - - // left over - if leftover > 0 { - zb = block{data: wbuf} - if z.BlockDependency { - if zn == 0 { - zb.data = append(z.window, zb.data...) - zb.offset = len(z.window) - } else { - zb.offset = winSize - } - } - - leftover = 0 - goto write - } - - if z.BlockDependency { - if len(z.window) == 0 { - z.window = make([]byte, winSize) - } - // last buffer may be shorter than the window - if len(buf) >= winSize { - copy(z.window, buf[len(buf)-winSize:]) - } else { - copy(z.window, z.window[len(buf):]) - copy(z.window[len(buf)+1:], buf) - } - } - - return - -write: - zb = z.compressBlock(zb) - _, err = z.writeBlock(zb) - - written := len(zb.data) - if bl > 0 { - if written >= bl { - written -= bl - bl = 0 - } else { - bl -= written - written = 0 - } - } - - n += written - // remove the window in zb.data - if z.BlockDependency { - if zi == 0 { - n -= len(z.window) - } else { - n -= winSize - } - } - if err != nil { - return - } - zi++ - goto loop -} - -// compressBlock compresses a block. -func (z *Writer) compressBlock(zb block) block { - // compressed block size cannot exceed the input's - var ( - n int - err error - zbuf = z.zbCompressBuf - ) - if z.HighCompression { - n, err = CompressBlockHC(zb.data, zbuf, zb.offset) - } else { - n, err = CompressBlock(zb.data, zbuf, zb.offset) - } - - // compressible and compressed size smaller than decompressed: ok! - if err == nil && n > 0 && len(zb.zdata) < len(zb.data) { - zb.compressed = true - zb.zdata = zbuf[:n] - } else { - zb.compressed = false - zb.zdata = zb.data[zb.offset:] - } - - if z.BlockChecksum { - xxh := hashPool.Get() - xxh.Write(zb.zdata) - zb.checksum = xxh.Sum32() - hashPool.Put(xxh) - } - - return zb -} - -// writeBlock writes a frame block to the underlying io.Writer (size, data). -func (z *Writer) writeBlock(zb block) (int, error) { - bLen := uint32(len(zb.zdata)) - if !zb.compressed { - bLen |= 1 << 31 - } - - n := 0 - - binary.LittleEndian.PutUint32(z.writeSizeBuf, bLen) - n, err := z.dst.Write(z.writeSizeBuf) - if err != nil { - return n, err - } - - m, err := z.dst.Write(zb.zdata) - n += m - if err != nil { - return n, err - } - - if z.BlockChecksum { - binary.LittleEndian.PutUint32(z.writeSizeBuf, zb.checksum) - m, err := z.dst.Write(z.writeSizeBuf) + var n int + for len(buf) > 0 { + // Accumulate the data to be compressed. + m := copy(z.data[z.idx:], buf) n += m + z.idx += m + buf = buf[m:] + debug("%d bytes copied to buf, current index %d", n, z.idx) - if err != nil { + if z.idx < len(z.data) { + // Buffer not filled. + debug("need more data for compression") + return n, nil + } + + // Buffer full. + if err := z.compressBlock(z.data); err != nil { return n, err } + z.idx = 0 } return n, nil } +// compressBlock compresses a block. +func (z *Writer) compressBlock(data []byte) error { + // The compressed block size cannot exceed the input's. + var zn int + var err error + + if level := z.Header.CompressionLevel; level != 0 { + zn, err = CompressBlockHC(data, z.zdata, level) + } else { + zn, err = CompressBlock(data, z.zdata, z.hashtable[:]) + } + + var zdata []byte + var bLen uint32 + debug("block compression %d => %d", len(data), zn) + if err == nil && zn > 0 && zn < len(data) { + // Compressible and compressed size smaller than uncompressed: ok! + bLen = uint32(zn) + zdata = z.zdata[:zn] + } else { + // Uncompressed block. + bLen = uint32(len(data)) | compressedBlockFlag + zdata = data + } + debug("block compression to be written len=%d data len=%d", bLen, len(zdata)) + + // Write the block. + if err := z.writeUint32(bLen); err != nil { + return err + } + if _, err := z.dst.Write(zdata); err != nil { + return err + } + + if z.BlockChecksum { + checksum := xxh32.ChecksumZero(zdata) + debug("block checksum %x", checksum) + if err := z.writeUint32(checksum); err != nil { + return err + } + } + debug("current frame checksum %x", z.checksum.Sum32()) + + return nil +} + // Flush flushes any pending compressed data to the underlying writer. // Flush does not return until the data has been written. // If the underlying writer returns an error, Flush returns that error. -// -// Flush is only required when in BlockDependency mode and the total of -// data written is less than 64Kb. func (z *Writer) Flush() error { - if len(z.data) == 0 { + debug("flush with index %d", z.idx) + if z.idx == 0 { return nil } - zb := z.compressBlock(block{data: z.data}) - if _, err := z.writeBlock(zb); err != nil { - return err - } - return nil + return z.compressBlock(z.data[:z.idx]) } // Close closes the Writer, flushing any unwritten data to the underlying io.Writer, but does not close the underlying io.Writer. @@ -324,19 +196,18 @@ func (z *Writer) Close() error { } } - // buffered data for the block dependency window - if z.BlockDependency && len(z.data) > 0 { - zb := block{data: z.data} - if _, err := z.writeBlock(z.compressBlock(zb)); err != nil { - return err - } + if err := z.Flush(); err != nil { + return err } - if err := binary.Write(z.dst, binary.LittleEndian, uint32(0)); err != nil { + debug("writing last empty block") + if err := z.writeUint32(0); err != nil { return err } if !z.NoChecksum { - if err := binary.Write(z.dst, binary.LittleEndian, z.checksum.Sum32()); err != nil { + checksum := z.checksum.Sum32() + debug("stream checksum %x", checksum) + if err := z.writeUint32(checksum); err != nil { return err } } @@ -350,28 +221,15 @@ func (z *Writer) Reset(w io.Writer) { z.Header = Header{} z.dst = w z.checksum.Reset() - z.data = nil - z.window = nil + z.zdata = z.zdata[:0] + z.data = z.data[:0] + z.idx = 0 } -// ReadFrom compresses the data read from the io.Reader and writes it to the underlying io.Writer. -// Returns the number of bytes read. -// It does not close the Writer. -func (z *Writer) ReadFrom(r io.Reader) (n int64, err error) { - cpus := runtime.GOMAXPROCS(0) - buf := make([]byte, cpus*z.BlockMaxSize) - for { - m, er := io.ReadFull(r, buf) - n += int64(m) - if er == nil || er == io.ErrUnexpectedEOF || er == io.EOF { - if _, err = z.Write(buf[:m]); err != nil { - return - } - if er == nil { - continue - } - return - } - return n, er - } +// writeUint32 writes a uint32 to the underlying writer. +func (z *Writer) writeUint32(x uint32) error { + buf := z.buf[:4] + binary.LittleEndian.PutUint32(buf, x) + _, err := z.dst.Write(buf) + return err } diff --git a/vendor/github.com/pierrec/lz4/writer_test.go b/vendor/github.com/pierrec/lz4/writer_test.go new file mode 100644 index 000000000..57fc4583d --- /dev/null +++ b/vendor/github.com/pierrec/lz4/writer_test.go @@ -0,0 +1,78 @@ +package lz4_test + +import ( + "bytes" + "fmt" + "io" + "io/ioutil" + "reflect" + "testing" + + "github.com/pierrec/lz4" +) + +func TestWriter(t *testing.T) { + goldenFiles := []string{ + "testdata/e.txt", + "testdata/gettysburg.txt", + "testdata/Mark.Twain-Tom.Sawyer.txt", + "testdata/pg1661.txt", + "testdata/pi.txt", + "testdata/random.data", + "testdata/repeat.txt", + } + + for _, fname := range goldenFiles { + for _, header := range []lz4.Header{ + {}, // Default header. + {BlockChecksum: true}, + {NoChecksum: true}, + {BlockMaxSize: 64 << 10}, // 64Kb + {CompressionLevel: 10}, + {Size: 123}, + } { + label := fmt.Sprintf("%s/%s", fname, header) + t.Run(label, func(t *testing.T) { + fname := fname + header := header + t.Parallel() + + raw, err := ioutil.ReadFile(fname) + if err != nil { + t.Fatal(err) + } + r := bytes.NewReader(raw) + + // Compress. + var zout bytes.Buffer + zw := lz4.NewWriter(&zout) + zw.Header = header + _, err = io.Copy(zw, r) + if err != nil { + t.Fatal(err) + } + err = zw.Close() + if err != nil { + t.Fatal(err) + } + + // Uncompress. + var out bytes.Buffer + zr := lz4.NewReader(&zout) + n, err := io.Copy(&out, zr) + if err != nil { + t.Fatal(err) + } + + // The uncompressed data must be the same as the initial input. + if got, want := int(n), len(raw); got != want { + t.Errorf("invalid sizes: got %d; want %d", got, want) + } + + if got, want := out.Bytes(), raw; !reflect.DeepEqual(got, want) { + t.Fatal("uncompressed data does not match original") + } + }) + } + } +} diff --git a/vendor/github.com/pierrec/xxHash/.travis.yml b/vendor/github.com/pierrec/xxHash/.travis.yml deleted file mode 100644 index 53e280eb7..000000000 --- a/vendor/github.com/pierrec/xxHash/.travis.yml +++ /dev/null @@ -1,11 +0,0 @@ -language: go - -go: - - 1.4 - - 1.5 - -script: - - go test -cpu=2 github.com/pierrec/xxHash/xxHash32 - - go test -cpu=2 github.com/pierrec/xxHash/xxHash64 - - go test -cpu=2 -race github.com/pierrec/xxHash/xxHash32 - - go test -cpu=2 -race github.com/pierrec/xxHash/xxHash64 diff --git a/vendor/github.com/pierrec/xxHash/LICENSE b/vendor/github.com/pierrec/xxHash/LICENSE deleted file mode 100644 index c1418f3f6..000000000 --- a/vendor/github.com/pierrec/xxHash/LICENSE +++ /dev/null @@ -1,28 +0,0 @@ -Copyright (c) 2014, Pierre Curto -All rights reserved. - -Redistribution and use in source and binary forms, with or without -modification, are permitted provided that the following conditions are met: - -* Redistributions of source code must retain the above copyright notice, this - list of conditions and the following disclaimer. - -* Redistributions in binary form must reproduce the above copyright notice, - this list of conditions and the following disclaimer in the documentation - and/or other materials provided with the distribution. - -* Neither the name of xxHash nor the names of its - contributors may be used to endorse or promote products derived from - this software without specific prior written permission. - -THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" -AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE -IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE -DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE -FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL -DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR -SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER -CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, -OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - diff --git a/vendor/github.com/pierrec/xxHash/README.md b/vendor/github.com/pierrec/xxHash/README.md deleted file mode 100644 index aafd28430..000000000 --- a/vendor/github.com/pierrec/xxHash/README.md +++ /dev/null @@ -1,36 +0,0 @@ -[![godoc](https://godoc.org/github.com/pierrec/xxHash?status.png)](https://godoc.org/github.com/pierrec/xxHash) -[![Build Status](https://travis-ci.org/pierrec/xxHash.svg?branch=master)](https://travis-ci.org/pierrec/xxHash) - -# Pure Go implementation of xxHash (32 and 64 bits versions) - -## Synopsis - -xxHash is a very fast hashing algorithm (see the details [here](https://github.com/Cyan4973/xxHash/)). -This package implements xxHash in pure [Go](http://www.golang.com). - - -## Usage - -This package follows the hash interfaces (hash.Hash32 and hash.Hash64). - -```go - import ( - "fmt" - - "github.com/pierrec/xxHash/xxHash32" - ) - - x := xxHash32.New(0xCAFE) // hash.Hash32 - x.Write([]byte("abc")) - x.Write([]byte("def")) - fmt.Printf("%x\n", x.Sum32()) - - x.Reset() - x.Write([]byte("abc")) - fmt.Printf("%x\n", x.Sum32()) -``` - -## Command line utility - -A simple command line utility is provided to hash files content under the xxhsum directory. - diff --git a/vendor/github.com/pierrec/xxHash/xxHash32/example_test.go b/vendor/github.com/pierrec/xxHash/xxHash32/example_test.go deleted file mode 100644 index bd912b04c..000000000 --- a/vendor/github.com/pierrec/xxHash/xxHash32/example_test.go +++ /dev/null @@ -1,21 +0,0 @@ -package xxHash32_test - -import ( - "bytes" - "fmt" - "github.com/pierrec/xxHash/xxHash32" -) - -func ExampleNew() { - buf := bytes.NewBufferString("this is a test") - x := xxHash32.New(0xCAFE) - x.Write(buf.Bytes()) - fmt.Printf("%x\n", x.Sum32()) - // Output: bb4f02bc -} - -func ExampleChecksum() { - buf := bytes.NewBufferString("this is a test") - fmt.Printf("%x\n", xxHash32.Checksum(buf.Bytes(), 0xCAFE)) - // Output: bb4f02bc -} diff --git a/vendor/github.com/pierrec/xxHash/xxHash32/xxHash32.go b/vendor/github.com/pierrec/xxHash/xxHash32/xxHash32.go deleted file mode 100644 index 411504e4b..000000000 --- a/vendor/github.com/pierrec/xxHash/xxHash32/xxHash32.go +++ /dev/null @@ -1,205 +0,0 @@ -// Package xxHash32 implements the very fast xxHash hashing algorithm (32 bits version). -// (https://github.com/Cyan4973/xxHash/) -package xxHash32 - -import "hash" - -const ( - prime32_1 = 2654435761 - prime32_2 = 2246822519 - prime32_3 = 3266489917 - prime32_4 = 668265263 - prime32_5 = 374761393 -) - -type xxHash struct { - seed uint32 - v1 uint32 - v2 uint32 - v3 uint32 - v4 uint32 - totalLen uint64 - buf [16]byte - bufused int -} - -// New returns a new Hash32 instance. -func New(seed uint32) hash.Hash32 { - xxh := &xxHash{seed: seed} - xxh.Reset() - return xxh -} - -// Sum appends the current hash to b and returns the resulting slice. -// It does not change the underlying hash state. -func (xxh xxHash) Sum(b []byte) []byte { - h32 := xxh.Sum32() - return append(b, byte(h32), byte(h32>>8), byte(h32>>16), byte(h32>>24)) -} - -// Reset resets the Hash to its initial state. -func (xxh *xxHash) Reset() { - xxh.v1 = xxh.seed + prime32_1 + prime32_2 - xxh.v2 = xxh.seed + prime32_2 - xxh.v3 = xxh.seed - xxh.v4 = xxh.seed - prime32_1 - xxh.totalLen = 0 - xxh.bufused = 0 -} - -// Size returns the number of bytes returned by Sum(). -func (xxh *xxHash) Size() int { - return 4 -} - -// BlockSize gives the minimum number of bytes accepted by Write(). -func (xxh *xxHash) BlockSize() int { - return 1 -} - -// Write adds input bytes to the Hash. -// It never returns an error. -func (xxh *xxHash) Write(input []byte) (int, error) { - n := len(input) - m := xxh.bufused - - xxh.totalLen += uint64(n) - - r := len(xxh.buf) - m - if n < r { - copy(xxh.buf[m:], input) - xxh.bufused += len(input) - return n, nil - } - - p := 0 - if m > 0 { - // some data left from previous update - copy(xxh.buf[xxh.bufused:], input[:r]) - xxh.bufused += len(input) - r - - // fast rotl(13) - p32 := xxh.v1 + (uint32(xxh.buf[p+3])<<24|uint32(xxh.buf[p+2])<<16|uint32(xxh.buf[p+1])<<8|uint32(xxh.buf[p]))*prime32_2 - xxh.v1 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - p32 = xxh.v2 + (uint32(xxh.buf[p+3])<<24|uint32(xxh.buf[p+2])<<16|uint32(xxh.buf[p+1])<<8|uint32(xxh.buf[p]))*prime32_2 - xxh.v2 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - p32 = xxh.v3 + (uint32(xxh.buf[p+3])<<24|uint32(xxh.buf[p+2])<<16|uint32(xxh.buf[p+1])<<8|uint32(xxh.buf[p]))*prime32_2 - xxh.v3 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - p32 = xxh.v4 + (uint32(xxh.buf[p+3])<<24|uint32(xxh.buf[p+2])<<16|uint32(xxh.buf[p+1])<<8|uint32(xxh.buf[p]))*prime32_2 - xxh.v4 = (p32<<13 | p32>>19) * prime32_1 - - p = r - xxh.bufused = 0 - } - - for n := n - 16; p <= n; { - p32 := xxh.v1 + (uint32(input[p+3])<<24|uint32(input[p+2])<<16|uint32(input[p+1])<<8|uint32(input[p]))*prime32_2 - xxh.v1 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - p32 = xxh.v2 + (uint32(input[p+3])<<24|uint32(input[p+2])<<16|uint32(input[p+1])<<8|uint32(input[p]))*prime32_2 - xxh.v2 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - p32 = xxh.v3 + (uint32(input[p+3])<<24|uint32(input[p+2])<<16|uint32(input[p+1])<<8|uint32(input[p]))*prime32_2 - xxh.v3 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - p32 = xxh.v4 + (uint32(input[p+3])<<24|uint32(input[p+2])<<16|uint32(input[p+1])<<8|uint32(input[p]))*prime32_2 - xxh.v4 = (p32<<13 | p32>>19) * prime32_1 - p += 4 - } - - copy(xxh.buf[xxh.bufused:], input[p:]) - xxh.bufused += len(input) - p - - return n, nil -} - -// Sum32 returns the 32 bits Hash value. -func (xxh *xxHash) Sum32() uint32 { - h32 := uint32(xxh.totalLen) - if xxh.totalLen >= 16 { - h32 += ((xxh.v1 << 1) | (xxh.v1 >> 31)) + - ((xxh.v2 << 7) | (xxh.v2 >> 25)) + - ((xxh.v3 << 12) | (xxh.v3 >> 20)) + - ((xxh.v4 << 18) | (xxh.v4 >> 14)) - } else { - h32 += xxh.seed + prime32_5 - } - - p := 0 - n := xxh.bufused - for n := n - 4; p <= n; p += 4 { - h32 += (uint32(xxh.buf[p+3])<<24 | uint32(xxh.buf[p+2])<<16 | uint32(xxh.buf[p+1])<<8 | uint32(xxh.buf[p])) * prime32_3 - h32 = ((h32 << 17) | (h32 >> 15)) * prime32_4 - } - for ; p < n; p++ { - h32 += uint32(xxh.buf[p]) * prime32_5 - h32 = ((h32 << 11) | (h32 >> 21)) * prime32_1 - } - - h32 ^= h32 >> 15 - h32 *= prime32_2 - h32 ^= h32 >> 13 - h32 *= prime32_3 - h32 ^= h32 >> 16 - - return h32 -} - -// Checksum returns the 32bits Hash value. -func Checksum(input []byte, seed uint32) uint32 { - n := len(input) - h32 := uint32(n) - - if n < 16 { - h32 += seed + prime32_5 - } else { - v1 := seed + prime32_1 + prime32_2 - v2 := seed + prime32_2 - v3 := seed - v4 := seed - prime32_1 - p := 0 - for p <= n-16 { - v1 += (uint32(input[p+3])<<24 | uint32(input[p+2])<<16 | uint32(input[p+1])<<8 | uint32(input[p])) * prime32_2 - v1 = (v1<<13 | v1>>19) * prime32_1 - p += 4 - v2 += (uint32(input[p+3])<<24 | uint32(input[p+2])<<16 | uint32(input[p+1])<<8 | uint32(input[p])) * prime32_2 - v2 = (v2<<13 | v2>>19) * prime32_1 - p += 4 - v3 += (uint32(input[p+3])<<24 | uint32(input[p+2])<<16 | uint32(input[p+1])<<8 | uint32(input[p])) * prime32_2 - v3 = (v3<<13 | v3>>19) * prime32_1 - p += 4 - v4 += (uint32(input[p+3])<<24 | uint32(input[p+2])<<16 | uint32(input[p+1])<<8 | uint32(input[p])) * prime32_2 - v4 = (v4<<13 | v4>>19) * prime32_1 - p += 4 - } - input = input[p:] - n -= p - h32 += ((v1 << 1) | (v1 >> 31)) + - ((v2 << 7) | (v2 >> 25)) + - ((v3 << 12) | (v3 >> 20)) + - ((v4 << 18) | (v4 >> 14)) - } - - p := 0 - for p <= n-4 { - h32 += (uint32(input[p+3])<<24 | uint32(input[p+2])<<16 | uint32(input[p+1])<<8 | uint32(input[p])) * prime32_3 - h32 = ((h32 << 17) | (h32 >> 15)) * prime32_4 - p += 4 - } - for p < n { - h32 += uint32(input[p]) * prime32_5 - h32 = ((h32 << 11) | (h32 >> 21)) * prime32_1 - p++ - } - - h32 ^= h32 >> 15 - h32 *= prime32_2 - h32 ^= h32 >> 13 - h32 *= prime32_3 - h32 ^= h32 >> 16 - - return h32 -} diff --git a/vendor/github.com/pierrec/xxHash/xxHash64/example_test.go b/vendor/github.com/pierrec/xxHash/xxHash64/example_test.go deleted file mode 100644 index c5fa56598..000000000 --- a/vendor/github.com/pierrec/xxHash/xxHash64/example_test.go +++ /dev/null @@ -1,21 +0,0 @@ -package xxHash64_test - -import ( - "bytes" - "fmt" - "github.com/pierrec/xxHash/xxHash64" -) - -func ExampleNew() { - buf := bytes.NewBufferString("this is a test") - x := xxHash64.New(0xCAFE) - x.Write(buf.Bytes()) - fmt.Printf("%x\n", x.Sum64()) - // Output: 4228c3215949e862 -} - -func ExampleChecksum() { - buf := bytes.NewBufferString("this is a test") - fmt.Printf("%x\n", xxHash64.Checksum(buf.Bytes(), 0xCAFE)) - // Output: 4228c3215949e862 -} diff --git a/vendor/github.com/pierrec/xxHash/xxHash64/xxHash64.go b/vendor/github.com/pierrec/xxHash/xxHash64/xxHash64.go deleted file mode 100644 index 2788e9506..000000000 --- a/vendor/github.com/pierrec/xxHash/xxHash64/xxHash64.go +++ /dev/null @@ -1,249 +0,0 @@ -// Package xxHash64 implements the very fast xxHash hashing algorithm (64 bits version). -// (https://github.com/Cyan4973/xxHash/) -package xxHash64 - -import "hash" - -const ( - prime64_1 = 11400714785074694791 - prime64_2 = 14029467366897019727 - prime64_3 = 1609587929392839161 - prime64_4 = 9650029242287828579 - prime64_5 = 2870177450012600261 -) - -type xxHash struct { - seed uint64 - v1 uint64 - v2 uint64 - v3 uint64 - v4 uint64 - totalLen uint64 - buf [32]byte - bufused int -} - -// New returns a new Hash64 instance. -func New(seed uint64) hash.Hash64 { - xxh := &xxHash{seed: seed} - xxh.Reset() - return xxh -} - -// Sum appends the current hash to b and returns the resulting slice. -// It does not change the underlying hash state. -func (xxh xxHash) Sum(b []byte) []byte { - h64 := xxh.Sum64() - return append(b, byte(h64), byte(h64>>8), byte(h64>>16), byte(h64>>24), byte(h64>>32), byte(h64>>40), byte(h64>>48), byte(h64>>56)) -} - -// Reset resets the Hash to its initial state. -func (xxh *xxHash) Reset() { - xxh.v1 = xxh.seed + prime64_1 + prime64_2 - xxh.v2 = xxh.seed + prime64_2 - xxh.v3 = xxh.seed - xxh.v4 = xxh.seed - prime64_1 - xxh.totalLen = 0 - xxh.bufused = 0 -} - -// Size returns the number of bytes returned by Sum(). -func (xxh *xxHash) Size() int { - return 8 -} - -// BlockSize gives the minimum number of bytes accepted by Write(). -func (xxh *xxHash) BlockSize() int { - return 1 -} - -// Write adds input bytes to the Hash. -// It never returns an error. -func (xxh *xxHash) Write(input []byte) (int, error) { - n := len(input) - m := xxh.bufused - - xxh.totalLen += uint64(n) - - r := len(xxh.buf) - m - if n < r { - copy(xxh.buf[m:], input) - xxh.bufused += len(input) - return n, nil - } - - p := 0 - if m > 0 { - // some data left from previous update - copy(xxh.buf[xxh.bufused:], input[:r]) - xxh.bufused += len(input) - r - - // fast rotl(31) - p64 := xxh.v1 + (uint64(xxh.buf[p+7])<<56|uint64(xxh.buf[p+6])<<48|uint64(xxh.buf[p+5])<<40|uint64(xxh.buf[p+4])<<32|uint64(xxh.buf[p+3])<<24|uint64(xxh.buf[p+2])<<16|uint64(xxh.buf[p+1])<<8|uint64(xxh.buf[p]))*prime64_2 - xxh.v1 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = xxh.v2 + (uint64(xxh.buf[p+7])<<56|uint64(xxh.buf[p+6])<<48|uint64(xxh.buf[p+5])<<40|uint64(xxh.buf[p+4])<<32|uint64(xxh.buf[p+3])<<24|uint64(xxh.buf[p+2])<<16|uint64(xxh.buf[p+1])<<8|uint64(xxh.buf[p]))*prime64_2 - xxh.v2 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = xxh.v3 + (uint64(xxh.buf[p+7])<<56|uint64(xxh.buf[p+6])<<48|uint64(xxh.buf[p+5])<<40|uint64(xxh.buf[p+4])<<32|uint64(xxh.buf[p+3])<<24|uint64(xxh.buf[p+2])<<16|uint64(xxh.buf[p+1])<<8|uint64(xxh.buf[p]))*prime64_2 - xxh.v3 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = xxh.v4 + (uint64(xxh.buf[p+7])<<56|uint64(xxh.buf[p+6])<<48|uint64(xxh.buf[p+5])<<40|uint64(xxh.buf[p+4])<<32|uint64(xxh.buf[p+3])<<24|uint64(xxh.buf[p+2])<<16|uint64(xxh.buf[p+1])<<8|uint64(xxh.buf[p]))*prime64_2 - xxh.v4 = (p64<<31 | p64>>33) * prime64_1 - - p = r - xxh.bufused = 0 - } - - for n := n - 32; p <= n; { - p64 := xxh.v1 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - xxh.v1 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = xxh.v2 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - xxh.v2 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = xxh.v3 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - xxh.v3 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = xxh.v4 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - xxh.v4 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - } - - copy(xxh.buf[xxh.bufused:], input[p:]) - xxh.bufused += len(input) - p - - return n, nil -} - -// Sum64 returns the 64bits Hash value. -func (xxh *xxHash) Sum64() uint64 { - var h64 uint64 - if xxh.totalLen >= 32 { - h64 = ((xxh.v1 << 1) | (xxh.v1 >> 63)) + - ((xxh.v2 << 7) | (xxh.v2 >> 57)) + - ((xxh.v3 << 12) | (xxh.v3 >> 52)) + - ((xxh.v4 << 18) | (xxh.v4 >> 46)) - - xxh.v1 *= prime64_2 - h64 ^= ((xxh.v1 << 31) | (xxh.v1 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 - - xxh.v2 *= prime64_2 - h64 ^= ((xxh.v2 << 31) | (xxh.v2 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 - - xxh.v3 *= prime64_2 - h64 ^= ((xxh.v3 << 31) | (xxh.v3 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 - - xxh.v4 *= prime64_2 - h64 ^= ((xxh.v4 << 31) | (xxh.v4 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 + xxh.totalLen - } else { - h64 = xxh.seed + prime64_5 + xxh.totalLen - } - - p := 0 - n := xxh.bufused - for n := n - 8; p <= n; p += 8 { - p64 := (uint64(xxh.buf[p+7])<<56 | uint64(xxh.buf[p+6])<<48 | uint64(xxh.buf[p+5])<<40 | uint64(xxh.buf[p+4])<<32 | uint64(xxh.buf[p+3])<<24 | uint64(xxh.buf[p+2])<<16 | uint64(xxh.buf[p+1])<<8 | uint64(xxh.buf[p])) * prime64_2 - h64 ^= ((p64 << 31) | (p64 >> 33)) * prime64_1 - h64 = ((h64<<27)|(h64>>37))*prime64_1 + prime64_4 - } - if p+4 <= n { - h64 ^= (uint64(xxh.buf[p+3])<<24 | uint64(xxh.buf[p+2])<<16 | uint64(xxh.buf[p+1])<<8 | uint64(xxh.buf[p])) * prime64_1 - h64 = ((h64<<23)|(h64>>41))*prime64_2 + prime64_3 - p += 4 - } - for ; p < n; p++ { - h64 ^= uint64(xxh.buf[p]) * prime64_5 - h64 = ((h64 << 11) | (h64 >> 53)) * prime64_1 - } - - h64 ^= h64 >> 33 - h64 *= prime64_2 - h64 ^= h64 >> 29 - h64 *= prime64_3 - h64 ^= h64 >> 32 - - return h64 -} - -// Checksum returns the 64bits Hash value. -func Checksum(input []byte, seed uint64) uint64 { - n := len(input) - var h64 uint64 - - if n >= 32 { - v1 := seed + prime64_1 + prime64_2 - v2 := seed + prime64_2 - v3 := seed - v4 := seed - prime64_1 - p := 0 - for n := n - 32; p <= n; { - p64 := v1 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - v1 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = v2 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - v2 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = v3 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - v3 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - p64 = v4 + (uint64(input[p+7])<<56|uint64(input[p+6])<<48|uint64(input[p+5])<<40|uint64(input[p+4])<<32|uint64(input[p+3])<<24|uint64(input[p+2])<<16|uint64(input[p+1])<<8|uint64(input[p]))*prime64_2 - v4 = (p64<<31 | p64>>33) * prime64_1 - p += 8 - } - - h64 = ((v1 << 1) | (v1 >> 63)) + - ((v2 << 7) | (v2 >> 57)) + - ((v3 << 12) | (v3 >> 52)) + - ((v4 << 18) | (v4 >> 46)) - - v1 *= prime64_2 - h64 ^= ((v1 << 31) | (v1 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 - - v2 *= prime64_2 - h64 ^= ((v2 << 31) | (v2 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 - - v3 *= prime64_2 - h64 ^= ((v3 << 31) | (v3 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 - - v4 *= prime64_2 - h64 ^= ((v4 << 31) | (v4 >> 33)) * prime64_1 - h64 = h64*prime64_1 + prime64_4 + uint64(n) - - input = input[p:] - n -= p - } else { - h64 = seed + prime64_5 + uint64(n) - } - - p := 0 - for n := n - 8; p <= n; p += 8 { - p64 := (uint64(input[p+7])<<56 | uint64(input[p+6])<<48 | uint64(input[p+5])<<40 | uint64(input[p+4])<<32 | uint64(input[p+3])<<24 | uint64(input[p+2])<<16 | uint64(input[p+1])<<8 | uint64(input[p])) * prime64_2 - h64 ^= ((p64 << 31) | (p64 >> 33)) * prime64_1 - h64 = ((h64<<27)|(h64>>37))*prime64_1 + prime64_4 - } - if p+4 <= n { - h64 ^= (uint64(input[p+3])<<24 | uint64(input[p+2])<<16 | uint64(input[p+1])<<8 | uint64(input[p])) * prime64_1 - h64 = ((h64<<23)|(h64>>41))*prime64_2 + prime64_3 - p += 4 - } - for ; p < n; p++ { - h64 ^= uint64(input[p]) * prime64_5 - h64 = ((h64 << 11) | (h64 >> 53)) * prime64_1 - } - - h64 ^= h64 >> 33 - h64 *= prime64_2 - h64 ^= h64 >> 29 - h64 *= prime64_3 - h64 ^= h64 >> 32 - - return h64 -} diff --git a/vendor/github.com/pierrec/xxHash/xxHash64/xxHash64_test.go b/vendor/github.com/pierrec/xxHash/xxHash64/xxHash64_test.go deleted file mode 100644 index e00a4ef4c..000000000 --- a/vendor/github.com/pierrec/xxHash/xxHash64/xxHash64_test.go +++ /dev/null @@ -1,150 +0,0 @@ -package xxHash64_test - -import ( - "encoding/binary" - "hash/crc64" - "hash/fnv" - "testing" - - "github.com/pierrec/xxHash/xxHash64" -) - -type test struct { - sum uint64 - data, printable string -} - -var testdata = []test{ - {0xef46db3751d8e999, "", ""}, - {0xd24ec4f1a98c6e5b, "a", ""}, - {0x65f708ca92d04a61, "ab", ""}, - {0x44bc2cf5ad770999, "abc", ""}, - {0xde0327b0d25d92cc, "abcd", ""}, - {0x07e3670c0c8dc7eb, "abcde", ""}, - {0xfa8afd82c423144d, "abcdef", ""}, - {0x1860940e2902822d, "abcdefg", ""}, - {0x3ad351775b4634b7, "abcdefgh", ""}, - {0x27f1a34fdbb95e13, "abcdefghi", ""}, - {0xd6287a1de5498bb2, "abcdefghij", ""}, - {0xbf2cd639b4143b80, "abcdefghijklmnopqrstuvwxyz012345", ""}, - {0x64f23ecf1609b766, "abcdefghijklmnopqrstuvwxyz0123456789", ""}, - {0xc5a8b11443765630, "Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.", ""}, -} - -func init() { - for i := range testdata { - d := &testdata[i] - if len(d.data) > 20 { - d.printable = d.data[:20] - } else { - d.printable = d.data - } - } -} - -func TestBlockSize(t *testing.T) { - xxh := xxHash64.New(0) - if s := xxh.BlockSize(); s <= 0 { - t.Errorf("invalid BlockSize: %d", s) - } -} - -func TestSize(t *testing.T) { - xxh := xxHash64.New(0) - if s := xxh.Size(); s != 8 { - t.Errorf("invalid Size: got %d expected 8", s) - } -} - -func TestData(t *testing.T) { - for i, td := range testdata { - xxh := xxHash64.New(0) - data := []byte(td.data) - xxh.Write(data) - if h := xxh.Sum64(); h != td.sum { - t.Errorf("test %d: xxh64(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) - t.FailNow() - } - if h := xxHash64.Checksum(data, 0); h != td.sum { - t.Errorf("test %d: xxh64(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) - t.FailNow() - } - } -} - -func TestSplitData(t *testing.T) { - for i, td := range testdata { - xxh := xxHash64.New(0) - data := []byte(td.data) - l := len(data) / 2 - xxh.Write(data[0:l]) - xxh.Write(data[l:]) - h := xxh.Sum64() - if h != td.sum { - t.Errorf("test %d: xxh64(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) - t.FailNow() - } - } -} - -func TestSum(t *testing.T) { - for i, td := range testdata { - xxh := xxHash64.New(0) - data := []byte(td.data) - xxh.Write(data) - b := xxh.Sum(data) - if h := binary.LittleEndian.Uint64(b[len(data):]); h != td.sum { - t.Errorf("test %d: xxh64(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) - t.FailNow() - } - } -} - -func TestReset(t *testing.T) { - xxh := xxHash64.New(0) - for i, td := range testdata { - xxh.Write([]byte(td.data)) - h := xxh.Sum64() - if h != td.sum { - t.Errorf("test %d: xxh64(%s)=0x%x expected 0x%x", i, td.printable, h, td.sum) - t.FailNow() - } - xxh.Reset() - } -} - -/////////////////////////////////////////////////////////////////////////////// -// Benchmarks -// -var testdata1 = []byte(testdata[len(testdata)-1].data) - -func Benchmark_XXH64(b *testing.B) { - h := xxHash64.New(0) - for n := 0; n < b.N; n++ { - h.Write(testdata1) - h.Sum64() - h.Reset() - } -} - -func Benchmark_XXH64_Checksum(b *testing.B) { - for n := 0; n < b.N; n++ { - xxHash64.Checksum(testdata1, 0) - } -} - -func Benchmark_CRC64(b *testing.B) { - t := crc64.MakeTable(0) - for i := 0; i < b.N; i++ { - crc64.Checksum(testdata1, t) - } -} - -func Benchmark_Fnv64(b *testing.B) { - h := fnv.New64() - for i := 0; i < b.N; i++ { - h.Write(testdata1) - h.Sum64() - h.Reset() - } -} diff --git a/vendor/github.com/pierrec/xxHash/xxhsum/main.go b/vendor/github.com/pierrec/xxHash/xxhsum/main.go deleted file mode 100644 index b73158bd9..000000000 --- a/vendor/github.com/pierrec/xxHash/xxhsum/main.go +++ /dev/null @@ -1,44 +0,0 @@ -// Command line interface to the xxHash32 and xxHash64 packages. -// Usage: -// xxHash [-mode 0] [-seed 123] filename1 [filename2...] -// where -// mode: hash mode (0=32bits, 1=64bits) (default=1) -// seed: seed to be used (default=0) -package main - -import ( - "flag" - "fmt" - "hash" - "io" - "os" - - "github.com/pierrec/xxHash/xxHash32" - "github.com/pierrec/xxHash/xxHash64" -) - -func main() { - seed := flag.Uint64("seed", 0, "seed value") - mode := flag.Int("mode", 1, "hash mode: 0=32bits, 1=64bits") - flag.Parse() - - var xxh hash.Hash - if *mode == 0 { - xxh = xxHash32.New(uint32(*seed)) - } else { - xxh = xxHash64.New(*seed) - } - - // Process each file in sequence - for _, filename := range flag.Args() { - inputFile, err := os.Open(filename) - if err != nil { - continue - } - if _, err := io.Copy(xxh, inputFile); err == nil { - fmt.Printf("%x %s\n", xxh.Sum(nil), filename) - } - inputFile.Close() - xxh.Reset() - } -} diff --git a/vendor/github.com/prometheus/common/config/http_config.go b/vendor/github.com/prometheus/common/config/http_config.go index 59a12e953..da5d59014 100644 --- a/vendor/github.com/prometheus/common/config/http_config.go +++ b/vendor/github.com/prometheus/common/config/http_config.go @@ -85,9 +85,6 @@ func (c *HTTPClientConfig) Validate() error { if c.BasicAuth != nil && (len(c.BearerToken) > 0 || len(c.BearerTokenFile) > 0) { return fmt.Errorf("at most one of basic_auth, bearer_token & bearer_token_file must be configured") } - if c.BasicAuth != nil && c.BasicAuth.Username == "" { - return fmt.Errorf("basic_auth requires a username") - } if c.BasicAuth != nil && (string(c.BasicAuth.Password) != "" && c.BasicAuth.PasswordFile != "") { return fmt.Errorf("at most one of basic_auth password & password_file must be configured") } diff --git a/vendor/github.com/prometheus/common/config/http_config_test.go b/vendor/github.com/prometheus/common/config/http_config_test.go index bd3771854..4639ae475 100644 --- a/vendor/github.com/prometheus/common/config/http_config_test.go +++ b/vendor/github.com/prometheus/common/config/http_config_test.go @@ -476,6 +476,32 @@ func TestBasicAuthNoPassword(t *testing.T) { } } +func TestBasicAuthNoUsername(t *testing.T) { + cfg, _, err := LoadHTTPConfigFile("testdata/http.conf.basic-auth.no-username.yaml") + if err != nil { + t.Errorf("Error loading HTTP client config: %v", err) + } + client, err := NewClientFromConfig(*cfg, "test") + if err != nil { + t.Errorf("Error creating HTTP Client: %v", err) + } + + rt, ok := client.Transport.(*basicAuthRoundTripper) + if !ok { + t.Fatalf("Error casting to basic auth transport, %v", client.Transport) + } + + if rt.username != "" { + t.Errorf("Got unexpected username: %s", rt.username) + } + if string(rt.password) != "secret" { + t.Errorf("Unexpected HTTP client password: %s", string(rt.password)) + } + if string(rt.passwordFile) != "" { + t.Errorf("Expected empty HTTP client passwordFile: %s", rt.passwordFile) + } +} + func TestBasicAuthPasswordFile(t *testing.T) { cfg, _, err := LoadHTTPConfigFile("testdata/http.conf.basic-auth.good.yaml") if err != nil { diff --git a/vendor/github.com/prometheus/common/config/testdata/http.conf.basic-auth.no-username.yaml b/vendor/github.com/prometheus/common/config/testdata/http.conf.basic-auth.no-username.yaml new file mode 100644 index 000000000..86e290372 --- /dev/null +++ b/vendor/github.com/prometheus/common/config/testdata/http.conf.basic-auth.no-username.yaml @@ -0,0 +1,2 @@ +basic_auth: + password: secret diff --git a/vendor/golang.org/x/crypto/acme/acme.go b/vendor/golang.org/x/crypto/acme/acme.go index 1f4fb69ed..8257ffb02 100644 --- a/vendor/golang.org/x/crypto/acme/acme.go +++ b/vendor/golang.org/x/crypto/acme/acme.go @@ -14,7 +14,6 @@ package acme import ( - "bytes" "context" "crypto" "crypto/ecdsa" @@ -23,6 +22,8 @@ import ( "crypto/sha256" "crypto/tls" "crypto/x509" + "crypto/x509/pkix" + "encoding/asn1" "encoding/base64" "encoding/hex" "encoding/json" @@ -33,7 +34,6 @@ import ( "io/ioutil" "math/big" "net/http" - "strconv" "strings" "sync" "time" @@ -42,6 +42,9 @@ import ( // LetsEncryptURL is the Directory endpoint of Let's Encrypt CA. const LetsEncryptURL = "https://acme-v01.api.letsencrypt.org/directory" +// idPeACMEIdentifierV1 is the OID for the ACME extension for the TLS-ALPN challenge. +var idPeACMEIdentifierV1 = asn1.ObjectIdentifier{1, 3, 6, 1, 5, 5, 7, 1, 30, 1} + const ( maxChainLen = 5 // max depth and breadth of a certificate chain maxCertSize = 1 << 20 // max size of a certificate, in bytes @@ -76,6 +79,22 @@ type Client struct { // will have no effect. DirectoryURL string + // RetryBackoff computes the duration after which the nth retry of a failed request + // should occur. The value of n for the first call on failure is 1. + // The values of r and resp are the request and response of the last failed attempt. + // If the returned value is negative or zero, no more retries are done and an error + // is returned to the caller of the original method. + // + // Requests which result in a 4xx client error are not retried, + // except for 400 Bad Request due to "bad nonce" errors and 429 Too Many Requests. + // + // If RetryBackoff is nil, a truncated exponential backoff algorithm + // with the ceiling of 10 seconds is used, where each subsequent retry n + // is done after either ("Retry-After" + jitter) or (2^n seconds + jitter), + // preferring the former if "Retry-After" header is found in the resp. + // The jitter is a random value up to 1 second. + RetryBackoff func(n int, r *http.Request, resp *http.Response) time.Duration + dirMu sync.Mutex // guards writes to dir dir *Directory // cached result of Client's Discover method @@ -99,15 +118,12 @@ func (c *Client) Discover(ctx context.Context) (Directory, error) { if dirURL == "" { dirURL = LetsEncryptURL } - res, err := c.get(ctx, dirURL) + res, err := c.get(ctx, dirURL, wantStatus(http.StatusOK)) if err != nil { return Directory{}, err } defer res.Body.Close() c.addNonce(res.Header) - if res.StatusCode != http.StatusOK { - return Directory{}, responseError(res) - } var v struct { Reg string `json:"new-reg"` @@ -166,14 +182,11 @@ func (c *Client) CreateCert(ctx context.Context, csr []byte, exp time.Duration, req.NotAfter = now.Add(exp).Format(time.RFC3339) } - res, err := c.retryPostJWS(ctx, c.Key, c.dir.CertURL, req) + res, err := c.post(ctx, c.Key, c.dir.CertURL, req, wantStatus(http.StatusCreated)) if err != nil { return nil, "", err } defer res.Body.Close() - if res.StatusCode != http.StatusCreated { - return nil, "", responseError(res) - } curl := res.Header.Get("Location") // cert permanent URL if res.ContentLength == 0 { @@ -196,26 +209,11 @@ func (c *Client) CreateCert(ctx context.Context, csr []byte, exp time.Duration, // Callers are encouraged to parse the returned value to ensure the certificate is valid // and has expected features. func (c *Client) FetchCert(ctx context.Context, url string, bundle bool) ([][]byte, error) { - for { - res, err := c.get(ctx, url) - if err != nil { - return nil, err - } - defer res.Body.Close() - if res.StatusCode == http.StatusOK { - return c.responseCert(ctx, res, bundle) - } - if res.StatusCode > 299 { - return nil, responseError(res) - } - d := retryAfter(res.Header.Get("Retry-After"), 3*time.Second) - select { - case <-time.After(d): - // retry - case <-ctx.Done(): - return nil, ctx.Err() - } + res, err := c.get(ctx, url, wantStatus(http.StatusOK)) + if err != nil { + return nil, err } + return c.responseCert(ctx, res, bundle) } // RevokeCert revokes a previously issued certificate cert, provided in DER format. @@ -241,14 +239,11 @@ func (c *Client) RevokeCert(ctx context.Context, key crypto.Signer, cert []byte, if key == nil { key = c.Key } - res, err := c.retryPostJWS(ctx, key, c.dir.RevokeURL, body) + res, err := c.post(ctx, key, c.dir.RevokeURL, body, wantStatus(http.StatusOK)) if err != nil { return err } defer res.Body.Close() - if res.StatusCode != http.StatusOK { - return responseError(res) - } return nil } @@ -329,14 +324,11 @@ func (c *Client) Authorize(ctx context.Context, domain string) (*Authorization, Resource: "new-authz", Identifier: authzID{Type: "dns", Value: domain}, } - res, err := c.retryPostJWS(ctx, c.Key, c.dir.AuthzURL, req) + res, err := c.post(ctx, c.Key, c.dir.AuthzURL, req, wantStatus(http.StatusCreated)) if err != nil { return nil, err } defer res.Body.Close() - if res.StatusCode != http.StatusCreated { - return nil, responseError(res) - } var v wireAuthz if err := json.NewDecoder(res.Body).Decode(&v); err != nil { @@ -353,14 +345,11 @@ func (c *Client) Authorize(ctx context.Context, domain string) (*Authorization, // If a caller needs to poll an authorization until its status is final, // see the WaitAuthorization method. func (c *Client) GetAuthorization(ctx context.Context, url string) (*Authorization, error) { - res, err := c.get(ctx, url) + res, err := c.get(ctx, url, wantStatus(http.StatusOK, http.StatusAccepted)) if err != nil { return nil, err } defer res.Body.Close() - if res.StatusCode != http.StatusOK && res.StatusCode != http.StatusAccepted { - return nil, responseError(res) - } var v wireAuthz if err := json.NewDecoder(res.Body).Decode(&v); err != nil { return nil, fmt.Errorf("acme: invalid response: %v", err) @@ -387,14 +376,11 @@ func (c *Client) RevokeAuthorization(ctx context.Context, url string) error { Status: "deactivated", Delete: true, } - res, err := c.retryPostJWS(ctx, c.Key, url, req) + res, err := c.post(ctx, c.Key, url, req, wantStatus(http.StatusOK)) if err != nil { return err } defer res.Body.Close() - if res.StatusCode != http.StatusOK { - return responseError(res) - } return nil } @@ -406,44 +392,42 @@ func (c *Client) RevokeAuthorization(ctx context.Context, url string) error { // In all other cases WaitAuthorization returns an error. // If the Status is StatusInvalid, the returned error is of type *AuthorizationError. func (c *Client) WaitAuthorization(ctx context.Context, url string) (*Authorization, error) { - sleep := sleeper(ctx) for { - res, err := c.get(ctx, url) + res, err := c.get(ctx, url, wantStatus(http.StatusOK, http.StatusAccepted)) if err != nil { return nil, err } - if res.StatusCode >= 400 && res.StatusCode <= 499 { - // Non-retriable error. For instance, Let's Encrypt may return 404 Not Found - // when requesting an expired authorization. - defer res.Body.Close() - return nil, responseError(res) - } - retry := res.Header.Get("Retry-After") - if res.StatusCode != http.StatusOK && res.StatusCode != http.StatusAccepted { - res.Body.Close() - if err := sleep(retry, 1); err != nil { - return nil, err - } - continue - } var raw wireAuthz err = json.NewDecoder(res.Body).Decode(&raw) res.Body.Close() - if err != nil { - if err := sleep(retry, 0); err != nil { - return nil, err - } - continue - } - if raw.Status == StatusValid { + switch { + case err != nil: + // Skip and retry. + case raw.Status == StatusValid: return raw.authorization(url), nil - } - if raw.Status == StatusInvalid { + case raw.Status == StatusInvalid: return nil, raw.error(url) } - if err := sleep(retry, 0); err != nil { - return nil, err + + // Exponential backoff is implemented in c.get above. + // This is just to prevent continuously hitting the CA + // while waiting for a final authorization status. + d := retryAfter(res.Header.Get("Retry-After")) + if d == 0 { + // Given that the fastest challenges TLS-SNI and HTTP-01 + // require a CA to make at least 1 network round trip + // and most likely persist a challenge state, + // this default delay seems reasonable. + d = time.Second + } + t := time.NewTimer(d) + select { + case <-ctx.Done(): + t.Stop() + return nil, ctx.Err() + case <-t.C: + // Retry. } } } @@ -452,14 +436,11 @@ func (c *Client) WaitAuthorization(ctx context.Context, url string) (*Authorizat // // A client typically polls a challenge status using this method. func (c *Client) GetChallenge(ctx context.Context, url string) (*Challenge, error) { - res, err := c.get(ctx, url) + res, err := c.get(ctx, url, wantStatus(http.StatusOK, http.StatusAccepted)) if err != nil { return nil, err } defer res.Body.Close() - if res.StatusCode != http.StatusOK && res.StatusCode != http.StatusAccepted { - return nil, responseError(res) - } v := wireChallenge{URI: url} if err := json.NewDecoder(res.Body).Decode(&v); err != nil { return nil, fmt.Errorf("acme: invalid response: %v", err) @@ -486,16 +467,14 @@ func (c *Client) Accept(ctx context.Context, chal *Challenge) (*Challenge, error Type: chal.Type, Auth: auth, } - res, err := c.retryPostJWS(ctx, c.Key, chal.URI, req) + res, err := c.post(ctx, c.Key, chal.URI, req, wantStatus( + http.StatusOK, // according to the spec + http.StatusAccepted, // Let's Encrypt: see https://goo.gl/WsJ7VT (acme-divergences.md) + )) if err != nil { return nil, err } defer res.Body.Close() - // Note: the protocol specifies 200 as the expected response code, but - // letsencrypt seems to be returning 202. - if res.StatusCode != http.StatusOK && res.StatusCode != http.StatusAccepted { - return nil, responseError(res) - } var v wireChallenge if err := json.NewDecoder(res.Body).Decode(&v); err != nil { @@ -552,7 +531,7 @@ func (c *Client) HTTP01ChallengePath(token string) string { // If no WithKey option is provided, a new ECDSA key is generated using P-256 curve. // // The returned certificate is valid for the next 24 hours and must be presented only when -// the server name of the client hello matches exactly the returned name value. +// the server name of the TLS ClientHello matches exactly the returned name value. func (c *Client) TLSSNI01ChallengeCert(token string, opt ...CertOption) (cert tls.Certificate, name string, err error) { ka, err := keyAuth(c.Key.Public(), token) if err != nil { @@ -579,7 +558,7 @@ func (c *Client) TLSSNI01ChallengeCert(token string, opt ...CertOption) (cert tl // If no WithKey option is provided, a new ECDSA key is generated using P-256 curve. // // The returned certificate is valid for the next 24 hours and must be presented only when -// the server name in the client hello matches exactly the returned name value. +// the server name in the TLS ClientHello matches exactly the returned name value. func (c *Client) TLSSNI02ChallengeCert(token string, opt ...CertOption) (cert tls.Certificate, name string, err error) { b := sha256.Sum256([]byte(token)) h := hex.EncodeToString(b[:]) @@ -600,6 +579,52 @@ func (c *Client) TLSSNI02ChallengeCert(token string, opt ...CertOption) (cert tl return cert, sanA, nil } +// TLSALPN01ChallengeCert creates a certificate for TLS-ALPN-01 challenge response. +// Servers can present the certificate to validate the challenge and prove control +// over a domain name. For more details on TLS-ALPN-01 see +// https://tools.ietf.org/html/draft-shoemaker-acme-tls-alpn-00#section-3 +// +// The token argument is a Challenge.Token value. +// If a WithKey option is provided, its private part signs the returned cert, +// and the public part is used to specify the signee. +// If no WithKey option is provided, a new ECDSA key is generated using P-256 curve. +// +// The returned certificate is valid for the next 24 hours and must be presented only when +// the server name in the TLS ClientHello matches the domain, and the special acme-tls/1 ALPN protocol +// has been specified. +func (c *Client) TLSALPN01ChallengeCert(token, domain string, opt ...CertOption) (cert tls.Certificate, err error) { + ka, err := keyAuth(c.Key.Public(), token) + if err != nil { + return tls.Certificate{}, err + } + shasum := sha256.Sum256([]byte(ka)) + extValue, err := asn1.Marshal(shasum[:]) + if err != nil { + return tls.Certificate{}, err + } + acmeExtension := pkix.Extension{ + Id: idPeACMEIdentifierV1, + Critical: true, + Value: extValue, + } + + tmpl := defaultTLSChallengeCertTemplate() + + var newOpt []CertOption + for _, o := range opt { + switch o := o.(type) { + case *certOptTemplate: + t := *(*x509.Certificate)(o) // shallow copy is ok + tmpl = &t + default: + newOpt = append(newOpt, o) + } + } + tmpl.ExtraExtensions = append(tmpl.ExtraExtensions, acmeExtension) + newOpt = append(newOpt, WithTemplate(tmpl)) + return tlsChallengeCert([]string{domain}, newOpt) +} + // doReg sends all types of registration requests. // The type of request is identified by typ argument, which is a "resource" // in the ACME spec terms. @@ -619,14 +644,14 @@ func (c *Client) doReg(ctx context.Context, url string, typ string, acct *Accoun req.Contact = acct.Contact req.Agreement = acct.AgreedTerms } - res, err := c.retryPostJWS(ctx, c.Key, url, req) + res, err := c.post(ctx, c.Key, url, req, wantStatus( + http.StatusOK, // updates and deletes + http.StatusCreated, // new account creation + )) if err != nil { return nil, err } defer res.Body.Close() - if res.StatusCode < 200 || res.StatusCode > 299 { - return nil, responseError(res) - } var v struct { Contact []string @@ -656,59 +681,6 @@ func (c *Client) doReg(ctx context.Context, url string, typ string, acct *Accoun }, nil } -// retryPostJWS will retry calls to postJWS if there is a badNonce error, -// clearing the stored nonces after each error. -// If the response was 4XX-5XX, then responseError is called on the body, -// the body is closed, and the error returned. -func (c *Client) retryPostJWS(ctx context.Context, key crypto.Signer, url string, body interface{}) (*http.Response, error) { - sleep := sleeper(ctx) - for { - res, err := c.postJWS(ctx, key, url, body) - if err != nil { - return nil, err - } - // handle errors 4XX-5XX with responseError - if res.StatusCode >= 400 && res.StatusCode <= 599 { - err := responseError(res) - res.Body.Close() - // according to spec badNonce is urn:ietf:params:acme:error:badNonce - // however, acme servers in the wild return their version of the error - // https://tools.ietf.org/html/draft-ietf-acme-acme-02#section-5.4 - if ae, ok := err.(*Error); ok && strings.HasSuffix(strings.ToLower(ae.ProblemType), ":badnonce") { - // clear any nonces that we might've stored that might now be - // considered bad - c.clearNonces() - retry := res.Header.Get("Retry-After") - if err := sleep(retry, 1); err != nil { - return nil, err - } - continue - } - return nil, err - } - return res, nil - } -} - -// postJWS signs the body with the given key and POSTs it to the provided url. -// The body argument must be JSON-serializable. -func (c *Client) postJWS(ctx context.Context, key crypto.Signer, url string, body interface{}) (*http.Response, error) { - nonce, err := c.popNonce(ctx, url) - if err != nil { - return nil, err - } - b, err := jwsEncodeJSON(body, key, nonce) - if err != nil { - return nil, err - } - res, err := c.post(ctx, url, "application/jose+json", bytes.NewReader(b)) - if err != nil { - return nil, err - } - c.addNonce(res.Header) - return res, nil -} - // popNonce returns a nonce value previously stored with c.addNonce // or fetches a fresh one from the given URL. func (c *Client) popNonce(ctx context.Context, url string) (string, error) { @@ -749,58 +721,12 @@ func (c *Client) addNonce(h http.Header) { c.nonces[v] = struct{}{} } -func (c *Client) httpClient() *http.Client { - if c.HTTPClient != nil { - return c.HTTPClient - } - return http.DefaultClient -} - -func (c *Client) get(ctx context.Context, urlStr string) (*http.Response, error) { - req, err := http.NewRequest("GET", urlStr, nil) - if err != nil { - return nil, err - } - return c.do(ctx, req) -} - -func (c *Client) head(ctx context.Context, urlStr string) (*http.Response, error) { - req, err := http.NewRequest("HEAD", urlStr, nil) - if err != nil { - return nil, err - } - return c.do(ctx, req) -} - -func (c *Client) post(ctx context.Context, urlStr, contentType string, body io.Reader) (*http.Response, error) { - req, err := http.NewRequest("POST", urlStr, body) - if err != nil { - return nil, err - } - req.Header.Set("Content-Type", contentType) - return c.do(ctx, req) -} - -func (c *Client) do(ctx context.Context, req *http.Request) (*http.Response, error) { - res, err := c.httpClient().Do(req.WithContext(ctx)) - if err != nil { - select { - case <-ctx.Done(): - // Prefer the unadorned context error. - // (The acme package had tests assuming this, previously from ctxhttp's - // behavior, predating net/http supporting contexts natively) - // TODO(bradfitz): reconsider this in the future. But for now this - // requires no test updates. - return nil, ctx.Err() - default: - return nil, err - } - } - return res, nil -} - func (c *Client) fetchNonce(ctx context.Context, url string) (string, error) { - resp, err := c.head(ctx, url) + r, err := http.NewRequest("HEAD", url, nil) + if err != nil { + return "", err + } + resp, err := c.doNoRetry(ctx, r) if err != nil { return "", err } @@ -852,24 +778,6 @@ func (c *Client) responseCert(ctx context.Context, res *http.Response, bundle bo return cert, nil } -// responseError creates an error of Error type from resp. -func responseError(resp *http.Response) error { - // don't care if ReadAll returns an error: - // json.Unmarshal will fail in that case anyway - b, _ := ioutil.ReadAll(resp.Body) - e := &wireError{Status: resp.StatusCode} - if err := json.Unmarshal(b, e); err != nil { - // this is not a regular error response: - // populate detail with anything we received, - // e.Status will already contain HTTP response code value - e.Detail = string(b) - if e.Detail == "" { - e.Detail = resp.Status - } - } - return e.error(resp.Header) -} - // chainCert fetches CA certificate chain recursively by following "up" links. // Each recursive call increments the depth by 1, resulting in an error // if the recursion level reaches maxChainLen. @@ -880,14 +788,11 @@ func (c *Client) chainCert(ctx context.Context, url string, depth int) ([][]byte return nil, errors.New("acme: certificate chain is too deep") } - res, err := c.get(ctx, url) + res, err := c.get(ctx, url, wantStatus(http.StatusOK)) if err != nil { return nil, err } defer res.Body.Close() - if res.StatusCode != http.StatusOK { - return nil, responseError(res) - } b, err := ioutil.ReadAll(io.LimitReader(res.Body, maxCertSize+1)) if err != nil { return nil, err @@ -932,65 +837,6 @@ func linkHeader(h http.Header, rel string) []string { return links } -// sleeper returns a function that accepts the Retry-After HTTP header value -// and an increment that's used with backoff to increasingly sleep on -// consecutive calls until the context is done. If the Retry-After header -// cannot be parsed, then backoff is used with a maximum sleep time of 10 -// seconds. -func sleeper(ctx context.Context) func(ra string, inc int) error { - var count int - return func(ra string, inc int) error { - count += inc - d := backoff(count, 10*time.Second) - d = retryAfter(ra, d) - wakeup := time.NewTimer(d) - defer wakeup.Stop() - select { - case <-ctx.Done(): - return ctx.Err() - case <-wakeup.C: - return nil - } - } -} - -// retryAfter parses a Retry-After HTTP header value, -// trying to convert v into an int (seconds) or use http.ParseTime otherwise. -// It returns d if v cannot be parsed. -func retryAfter(v string, d time.Duration) time.Duration { - if i, err := strconv.Atoi(v); err == nil { - return time.Duration(i) * time.Second - } - t, err := http.ParseTime(v) - if err != nil { - return d - } - return t.Sub(timeNow()) -} - -// backoff computes a duration after which an n+1 retry iteration should occur -// using truncated exponential backoff algorithm. -// -// The n argument is always bounded between 0 and 30. -// The max argument defines upper bound for the returned value. -func backoff(n int, max time.Duration) time.Duration { - if n < 0 { - n = 0 - } - if n > 30 { - n = 30 - } - var d time.Duration - if x, err := rand.Int(rand.Reader, big.NewInt(1000)); err == nil { - d = time.Duration(x.Int64()) * time.Millisecond - } - d += time.Duration(1< max { - return max - } - return d -} - // keyAuth generates a key authorization string for a given token. func keyAuth(pub crypto.PublicKey, token string) (string, error) { th, err := JWKThumbprint(pub) @@ -1000,15 +846,25 @@ func keyAuth(pub crypto.PublicKey, token string) (string, error) { return fmt.Sprintf("%s.%s", token, th), nil } +// defaultTLSChallengeCertTemplate is a template used to create challenge certs for TLS challenges. +func defaultTLSChallengeCertTemplate() *x509.Certificate { + return &x509.Certificate{ + SerialNumber: big.NewInt(1), + NotBefore: time.Now(), + NotAfter: time.Now().Add(24 * time.Hour), + BasicConstraintsValid: true, + KeyUsage: x509.KeyUsageKeyEncipherment | x509.KeyUsageDigitalSignature, + ExtKeyUsage: []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth}, + } +} + // tlsChallengeCert creates a temporary certificate for TLS-SNI challenges // with the given SANs and auto-generated public/private key pair. // The Subject Common Name is set to the first SAN to aid debugging. // To create a cert with a custom key pair, specify WithKey option. func tlsChallengeCert(san []string, opt []CertOption) (tls.Certificate, error) { - var ( - key crypto.Signer - tmpl *x509.Certificate - ) + var key crypto.Signer + tmpl := defaultTLSChallengeCertTemplate() for _, o := range opt { switch o := o.(type) { case *certOptKey: @@ -1017,7 +873,7 @@ func tlsChallengeCert(san []string, opt []CertOption) (tls.Certificate, error) { } key = o.key case *certOptTemplate: - var t = *(*x509.Certificate)(o) // shallow copy is ok + t := *(*x509.Certificate)(o) // shallow copy is ok tmpl = &t default: // package's fault, if we let this happen: @@ -1030,16 +886,6 @@ func tlsChallengeCert(san []string, opt []CertOption) (tls.Certificate, error) { return tls.Certificate{}, err } } - if tmpl == nil { - tmpl = &x509.Certificate{ - SerialNumber: big.NewInt(1), - NotBefore: time.Now(), - NotAfter: time.Now().Add(24 * time.Hour), - BasicConstraintsValid: true, - KeyUsage: x509.KeyUsageKeyEncipherment | x509.KeyUsageDigitalSignature, - ExtKeyUsage: []x509.ExtKeyUsage{x509.ExtKeyUsageServerAuth}, - } - } tmpl.DNSNames = san if len(san) > 0 { tmpl.Subject.CommonName = san[0] diff --git a/vendor/golang.org/x/crypto/acme/acme_test.go b/vendor/golang.org/x/crypto/acme/acme_test.go index 63cb79b98..ef1fe4782 100644 --- a/vendor/golang.org/x/crypto/acme/acme_test.go +++ b/vendor/golang.org/x/crypto/acme/acme_test.go @@ -13,9 +13,9 @@ import ( "crypto/x509" "crypto/x509/pkix" "encoding/base64" + "encoding/hex" "encoding/json" "fmt" - "io/ioutil" "math/big" "net/http" "net/http/httptest" @@ -485,88 +485,37 @@ func TestGetAuthorization(t *testing.T) { } func TestWaitAuthorization(t *testing.T) { - var count int - ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { - count++ - w.Header().Set("Retry-After", "0") - if count > 1 { - fmt.Fprintf(w, `{"status":"valid"}`) - return + t.Run("wait loop", func(t *testing.T) { + var count int + authz, err := runWaitAuthorization(context.Background(), t, func(w http.ResponseWriter, r *http.Request) { + count++ + w.Header().Set("Retry-After", "0") + if count > 1 { + fmt.Fprintf(w, `{"status":"valid"}`) + return + } + fmt.Fprintf(w, `{"status":"pending"}`) + }) + if err != nil { + t.Fatalf("non-nil error: %v", err) } - fmt.Fprintf(w, `{"status":"pending"}`) - })) - defer ts.Close() - - type res struct { - authz *Authorization - err error - } - done := make(chan res) - defer close(done) - go func() { - var client Client - a, err := client.WaitAuthorization(context.Background(), ts.URL) - done <- res{a, err} - }() - - select { - case <-time.After(5 * time.Second): - t.Fatal("WaitAuthz took too long to return") - case res := <-done: - if res.err != nil { - t.Fatalf("res.err = %v", res.err) - } - if res.authz == nil { - t.Fatal("res.authz is nil") - } - } -} - -func TestWaitAuthorizationInvalid(t *testing.T) { - ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { - fmt.Fprintf(w, `{"status":"invalid"}`) - })) - defer ts.Close() - - res := make(chan error) - defer close(res) - go func() { - var client Client - _, err := client.WaitAuthorization(context.Background(), ts.URL) - res <- err - }() - - select { - case <-time.After(3 * time.Second): - t.Fatal("WaitAuthz took too long to return") - case err := <-res: - if err == nil { - t.Error("err is nil") + if authz == nil { + t.Fatal("authz is nil") } + }) + t.Run("invalid status", func(t *testing.T) { + _, err := runWaitAuthorization(context.Background(), t, func(w http.ResponseWriter, r *http.Request) { + fmt.Fprintf(w, `{"status":"invalid"}`) + }) if _, ok := err.(*AuthorizationError); !ok { - t.Errorf("err is %T; want *AuthorizationError", err) + t.Errorf("err is %v (%T); want non-nil *AuthorizationError", err, err) } - } -} - -func TestWaitAuthorizationClientError(t *testing.T) { - const code = http.StatusBadRequest - ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { - w.WriteHeader(code) - })) - defer ts.Close() - - ch := make(chan error, 1) - go func() { - var client Client - _, err := client.WaitAuthorization(context.Background(), ts.URL) - ch <- err - }() - - select { - case <-time.After(3 * time.Second): - t.Fatal("WaitAuthz took too long to return") - case err := <-ch: + }) + t.Run("non-retriable error", func(t *testing.T) { + const code = http.StatusBadRequest + _, err := runWaitAuthorization(context.Background(), t, func(w http.ResponseWriter, r *http.Request) { + w.WriteHeader(code) + }) res, ok := err.(*Error) if !ok { t.Fatalf("err is %v (%T); want a non-nil *Error", err, err) @@ -574,34 +523,60 @@ func TestWaitAuthorizationClientError(t *testing.T) { if res.StatusCode != code { t.Errorf("res.StatusCode = %d; want %d", res.StatusCode, code) } + }) + for _, code := range []int{http.StatusTooManyRequests, http.StatusInternalServerError} { + t.Run(fmt.Sprintf("retriable %d error", code), func(t *testing.T) { + var count int + authz, err := runWaitAuthorization(context.Background(), t, func(w http.ResponseWriter, r *http.Request) { + count++ + w.Header().Set("Retry-After", "0") + if count > 1 { + fmt.Fprintf(w, `{"status":"valid"}`) + return + } + w.WriteHeader(code) + }) + if err != nil { + t.Fatalf("non-nil error: %v", err) + } + if authz == nil { + t.Fatal("authz is nil") + } + }) } -} - -func TestWaitAuthorizationCancel(t *testing.T) { - ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { - w.Header().Set("Retry-After", "60") - fmt.Fprintf(w, `{"status":"pending"}`) - })) - defer ts.Close() - - res := make(chan error) - defer close(res) - go func() { - var client Client + t.Run("context cancel", func(t *testing.T) { ctx, cancel := context.WithTimeout(context.Background(), 200*time.Millisecond) defer cancel() - _, err := client.WaitAuthorization(ctx, ts.URL) - res <- err - }() - - select { - case <-time.After(time.Second): - t.Fatal("WaitAuthz took too long to return") - case err := <-res: + _, err := runWaitAuthorization(ctx, t, func(w http.ResponseWriter, r *http.Request) { + w.Header().Set("Retry-After", "60") + fmt.Fprintf(w, `{"status":"pending"}`) + }) if err == nil { t.Error("err is nil") } + }) +} +func runWaitAuthorization(ctx context.Context, t *testing.T, h http.HandlerFunc) (*Authorization, error) { + t.Helper() + ts := httptest.NewServer(h) + defer ts.Close() + type res struct { + authz *Authorization + err error } + ch := make(chan res, 1) + go func() { + var client Client + a, err := client.WaitAuthorization(ctx, ts.URL) + ch <- res{a, err} + }() + select { + case <-time.After(3 * time.Second): + t.Fatal("WaitAuthorization took too long to return") + case v := <-ch: + return v.authz, v.err + } + panic("runWaitAuthorization: out of select") } func TestRevokeAuthorization(t *testing.T) { @@ -628,7 +603,7 @@ func TestRevokeAuthorization(t *testing.T) { t.Errorf("req.Delete is false") } case "/2": - w.WriteHeader(http.StatusInternalServerError) + w.WriteHeader(http.StatusBadRequest) } })) defer ts.Close() @@ -849,7 +824,7 @@ func TestFetchCertRetry(t *testing.T) { ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { if count < 1 { w.Header().Set("Retry-After", "0") - w.WriteHeader(http.StatusAccepted) + w.WriteHeader(http.StatusTooManyRequests) count++ return } @@ -1096,44 +1071,6 @@ func TestNonce_postJWS(t *testing.T) { } } -func TestRetryPostJWS(t *testing.T) { - var count int - ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { - count++ - w.Header().Set("Replay-Nonce", fmt.Sprintf("nonce%d", count)) - if r.Method == "HEAD" { - // We expect the client to do 2 head requests to fetch - // nonces, one to start and another after getting badNonce - return - } - - head, err := decodeJWSHead(r) - if err != nil { - t.Errorf("decodeJWSHead: %v", err) - } else if head.Nonce == "" { - t.Error("head.Nonce is empty") - } else if head.Nonce == "nonce1" { - // return a badNonce error to force the call to retry - w.WriteHeader(http.StatusBadRequest) - w.Write([]byte(`{"type":"urn:ietf:params:acme:error:badNonce"}`)) - return - } - // Make client.Authorize happy; we're not testing its result. - w.WriteHeader(http.StatusCreated) - w.Write([]byte(`{"status":"valid"}`)) - })) - defer ts.Close() - - client := Client{Key: testKey, dir: &Directory{AuthzURL: ts.URL}} - // This call will fail with badNonce, causing a retry - if _, err := client.Authorize(context.Background(), "example.com"); err != nil { - t.Errorf("client.Authorize 1: %v", err) - } - if count != 4 { - t.Errorf("total requests count: %d; want 4", count) - } -} - func TestLinkHeader(t *testing.T) { h := http.Header{"Link": { `;rel="next"`, @@ -1157,37 +1094,6 @@ func TestLinkHeader(t *testing.T) { } } -func TestErrorResponse(t *testing.T) { - s := `{ - "status": 400, - "type": "urn:acme:error:xxx", - "detail": "text" - }` - res := &http.Response{ - StatusCode: 400, - Status: "400 Bad Request", - Body: ioutil.NopCloser(strings.NewReader(s)), - Header: http.Header{"X-Foo": {"bar"}}, - } - err := responseError(res) - v, ok := err.(*Error) - if !ok { - t.Fatalf("err = %+v (%T); want *Error type", err, err) - } - if v.StatusCode != 400 { - t.Errorf("v.StatusCode = %v; want 400", v.StatusCode) - } - if v.ProblemType != "urn:acme:error:xxx" { - t.Errorf("v.ProblemType = %q; want urn:acme:error:xxx", v.ProblemType) - } - if v.Detail != "text" { - t.Errorf("v.Detail = %q; want text", v.Detail) - } - if !reflect.DeepEqual(v.Header, res.Header) { - t.Errorf("v.Header = %+v; want %+v", v.Header, res.Header) - } -} - func TestTLSSNI01ChallengeCert(t *testing.T) { const ( token = "evaGxfADs6pSRb2LAv9IZf17Dt3juxGJ-PCt92wr-oA" @@ -1255,6 +1161,58 @@ func TestTLSSNI02ChallengeCert(t *testing.T) { } } +func TestTLSALPN01ChallengeCert(t *testing.T) { + const ( + token = "evaGxfADs6pSRb2LAv9IZf17Dt3juxGJ-PCt92wr-oA" + keyAuth = "evaGxfADs6pSRb2LAv9IZf17Dt3juxGJ-PCt92wr-oA." + testKeyECThumbprint + // echo -n | shasum -a 256 + h = "0420dbbd5eefe7b4d06eb9d1d9f5acb4c7cda27d320e4b30332f0b6cb441734ad7b0" + domain = "example.com" + ) + + extValue, err := hex.DecodeString(h) + if err != nil { + t.Fatal(err) + } + + client := &Client{Key: testKeyEC} + tlscert, err := client.TLSALPN01ChallengeCert(token, domain) + if err != nil { + t.Fatal(err) + } + + if n := len(tlscert.Certificate); n != 1 { + t.Fatalf("len(tlscert.Certificate) = %d; want 1", n) + } + cert, err := x509.ParseCertificate(tlscert.Certificate[0]) + if err != nil { + t.Fatal(err) + } + names := []string{domain} + if !reflect.DeepEqual(cert.DNSNames, names) { + t.Fatalf("cert.DNSNames = %v;\nwant %v", cert.DNSNames, names) + } + if cn := cert.Subject.CommonName; cn != domain { + t.Errorf("CommonName = %q; want %q", cn, domain) + } + acmeExts := []pkix.Extension{} + for _, ext := range cert.Extensions { + if idPeACMEIdentifierV1.Equal(ext.Id) { + acmeExts = append(acmeExts, ext) + } + } + if len(acmeExts) != 1 { + t.Errorf("acmeExts = %v; want exactly one", acmeExts) + } + if !acmeExts[0].Critical { + t.Errorf("acmeExt.Critical = %v; want true", acmeExts[0].Critical) + } + if bytes.Compare(acmeExts[0].Value, extValue) != 0 { + t.Errorf("acmeExt.Value = %v; want %v", acmeExts[0].Value, extValue) + } + +} + func TestTLSChallengeCertOpt(t *testing.T) { key, err := rsa.GenerateKey(rand.Reader, 512) if err != nil { @@ -1353,28 +1311,3 @@ func TestDNS01ChallengeRecord(t *testing.T) { t.Errorf("val = %q; want %q", val, value) } } - -func TestBackoff(t *testing.T) { - tt := []struct{ min, max time.Duration }{ - {time.Second, 2 * time.Second}, - {2 * time.Second, 3 * time.Second}, - {4 * time.Second, 5 * time.Second}, - {8 * time.Second, 9 * time.Second}, - } - for i, test := range tt { - d := backoff(i, time.Minute) - if d < test.min || test.max < d { - t.Errorf("%d: d = %v; want between %v and %v", i, d, test.min, test.max) - } - } - - min, max := time.Second, 2*time.Second - if d := backoff(-1, time.Minute); d < min || max < d { - t.Errorf("d = %v; want between %v and %v", d, min, max) - } - - bound := 10 * time.Second - if d := backoff(100, bound); d != bound { - t.Errorf("d = %v; want %v", d, bound) - } -} diff --git a/vendor/golang.org/x/crypto/acme/autocert/autocert.go b/vendor/golang.org/x/crypto/acme/autocert/autocert.go index 263b29133..c8fa4e666 100644 --- a/vendor/golang.org/x/crypto/acme/autocert/autocert.go +++ b/vendor/golang.org/x/crypto/acme/autocert/autocert.go @@ -98,11 +98,11 @@ type Manager struct { // To always accept the terms, the callers can use AcceptTOS. Prompt func(tosURL string) bool - // Cache optionally stores and retrieves previously-obtained certificates. - // If nil, certs will only be cached for the lifetime of the Manager. + // Cache optionally stores and retrieves previously-obtained certificates + // and other state. If nil, certs will only be cached for the lifetime of + // the Manager. Multiple Managers can share the same Cache. // - // Manager passes the Cache certificates data encoded in PEM, with private/public - // parts combined in a single Cache.Put call, private key first. + // Using a persistent Cache, such as DirCache, is strongly recommended. Cache Cache // HostPolicy controls which domains the Manager will attempt @@ -127,8 +127,10 @@ type Manager struct { // Client is used to perform low-level operations, such as account registration // and requesting new certificates. + // // If Client is nil, a zero-value acme.Client is used with acme.LetsEncryptURL - // directory endpoint and a newly-generated ECDSA P-256 key. + // as directory endpoint. If the Client.Key is nil, a new ECDSA P-256 key is + // generated and, if Cache is not nil, stored in cache. // // Mutating the field after the first call of GetCertificate method will have no effect. Client *acme.Client @@ -140,22 +142,30 @@ type Manager struct { // If the Client's account key is already registered, Email is not used. Email string - // ForceRSA makes the Manager generate certificates with 2048-bit RSA keys. + // ForceRSA used to make the Manager generate RSA certificates. It is now ignored. // - // If false, a default is used. Currently the default - // is EC-based keys using the P-256 curve. + // Deprecated: the Manager will request the correct type of certificate based + // on what each client supports. ForceRSA bool + // ExtraExtensions are used when generating a new CSR (Certificate Request), + // thus allowing customization of the resulting certificate. + // For instance, TLS Feature Extension (RFC 7633) can be used + // to prevent an OCSP downgrade attack. + // + // The field value is passed to crypto/x509.CreateCertificateRequest + // in the template's ExtraExtensions field as is. + ExtraExtensions []pkix.Extension + clientMu sync.Mutex client *acme.Client // initialized by acmeClient method stateMu sync.Mutex - state map[string]*certState // keyed by domain name + state map[certKey]*certState // renewal tracks the set of domains currently running renewal timers. - // It is keyed by domain name. renewalMu sync.Mutex - renewal map[string]*domainRenewal + renewal map[certKey]*domainRenewal // tokensMu guards the rest of the fields: tryHTTP01, certTokens and httpTokens. tokensMu sync.RWMutex @@ -174,6 +184,23 @@ type Manager struct { certTokens map[string]*tls.Certificate } +// certKey is the key by which certificates are tracked in state, renewal and cache. +type certKey struct { + domain string // without trailing dot + isRSA bool // RSA cert for legacy clients (as opposed to default ECDSA) + isToken bool // tls-sni challenge token cert; key type is undefined regardless of isRSA +} + +func (c certKey) String() string { + if c.isToken { + return c.domain + "+token" + } + if c.isRSA { + return c.domain + "+rsa" + } + return c.domain +} + // GetCertificate implements the tls.Config.GetCertificate hook. // It provides a TLS certificate for hello.ServerName host, including answering // *.acme.invalid (TLS-SNI) challenges. All other fields of hello are ignored. @@ -194,7 +221,7 @@ func (m *Manager) GetCertificate(hello *tls.ClientHelloInfo) (*tls.Certificate, if !strings.Contains(strings.Trim(name, "."), ".") { return nil, errors.New("acme/autocert: server name component count invalid") } - if strings.ContainsAny(name, `/\`) { + if strings.ContainsAny(name, `+/\`) { return nil, errors.New("acme/autocert: server name contains invalid character") } @@ -210,7 +237,7 @@ func (m *Manager) GetCertificate(hello *tls.ClientHelloInfo) (*tls.Certificate, if cert := m.certTokens[name]; cert != nil { return cert, nil } - if cert, err := m.cacheGet(ctx, name); err == nil { + if cert, err := m.cacheGet(ctx, certKey{domain: name, isToken: true}); err == nil { return cert, nil } // TODO: cache error results? @@ -218,8 +245,11 @@ func (m *Manager) GetCertificate(hello *tls.ClientHelloInfo) (*tls.Certificate, } // regular domain - name = strings.TrimSuffix(name, ".") // golang.org/issue/18114 - cert, err := m.cert(ctx, name) + ck := certKey{ + domain: strings.TrimSuffix(name, "."), // golang.org/issue/18114 + isRSA: !supportsECDSA(hello), + } + cert, err := m.cert(ctx, ck) if err == nil { return cert, nil } @@ -231,14 +261,60 @@ func (m *Manager) GetCertificate(hello *tls.ClientHelloInfo) (*tls.Certificate, if err := m.hostPolicy()(ctx, name); err != nil { return nil, err } - cert, err = m.createCert(ctx, name) + cert, err = m.createCert(ctx, ck) if err != nil { return nil, err } - m.cachePut(ctx, name, cert) + m.cachePut(ctx, ck, cert) return cert, nil } +func supportsECDSA(hello *tls.ClientHelloInfo) bool { + // The "signature_algorithms" extension, if present, limits the key exchange + // algorithms allowed by the cipher suites. See RFC 5246, section 7.4.1.4.1. + if hello.SignatureSchemes != nil { + ecdsaOK := false + schemeLoop: + for _, scheme := range hello.SignatureSchemes { + const tlsECDSAWithSHA1 tls.SignatureScheme = 0x0203 // constant added in Go 1.10 + switch scheme { + case tlsECDSAWithSHA1, tls.ECDSAWithP256AndSHA256, + tls.ECDSAWithP384AndSHA384, tls.ECDSAWithP521AndSHA512: + ecdsaOK = true + break schemeLoop + } + } + if !ecdsaOK { + return false + } + } + if hello.SupportedCurves != nil { + ecdsaOK := false + for _, curve := range hello.SupportedCurves { + if curve == tls.CurveP256 { + ecdsaOK = true + break + } + } + if !ecdsaOK { + return false + } + } + for _, suite := range hello.CipherSuites { + switch suite { + case tls.TLS_ECDHE_ECDSA_WITH_RC4_128_SHA, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA, + tls.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256, + tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + tls.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384, + tls.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305: + return true + } + } + return false +} + // HTTPHandler configures the Manager to provision ACME "http-01" challenge responses. // It returns an http.Handler that responds to the challenges and must be // running on port 80. If it receives a request that is not an ACME challenge, @@ -304,16 +380,16 @@ func stripPort(hostport string) string { // cert returns an existing certificate either from m.state or cache. // If a certificate is found in cache but not in m.state, the latter will be filled // with the cached value. -func (m *Manager) cert(ctx context.Context, name string) (*tls.Certificate, error) { +func (m *Manager) cert(ctx context.Context, ck certKey) (*tls.Certificate, error) { m.stateMu.Lock() - if s, ok := m.state[name]; ok { + if s, ok := m.state[ck]; ok { m.stateMu.Unlock() s.RLock() defer s.RUnlock() return s.tlscert() } defer m.stateMu.Unlock() - cert, err := m.cacheGet(ctx, name) + cert, err := m.cacheGet(ctx, ck) if err != nil { return nil, err } @@ -322,25 +398,25 @@ func (m *Manager) cert(ctx context.Context, name string) (*tls.Certificate, erro return nil, errors.New("acme/autocert: private key cannot sign") } if m.state == nil { - m.state = make(map[string]*certState) + m.state = make(map[certKey]*certState) } s := &certState{ key: signer, cert: cert.Certificate, leaf: cert.Leaf, } - m.state[name] = s - go m.renew(name, s.key, s.leaf.NotAfter) + m.state[ck] = s + go m.renew(ck, s.key, s.leaf.NotAfter) return cert, nil } // cacheGet always returns a valid certificate, or an error otherwise. -// If a cached certficate exists but is not valid, ErrCacheMiss is returned. -func (m *Manager) cacheGet(ctx context.Context, domain string) (*tls.Certificate, error) { +// If a cached certificate exists but is not valid, ErrCacheMiss is returned. +func (m *Manager) cacheGet(ctx context.Context, ck certKey) (*tls.Certificate, error) { if m.Cache == nil { return nil, ErrCacheMiss } - data, err := m.Cache.Get(ctx, domain) + data, err := m.Cache.Get(ctx, ck.String()) if err != nil { return nil, err } @@ -371,7 +447,7 @@ func (m *Manager) cacheGet(ctx context.Context, domain string) (*tls.Certificate } // verify and create TLS cert - leaf, err := validCert(domain, pubDER, privKey) + leaf, err := validCert(ck, pubDER, privKey) if err != nil { return nil, ErrCacheMiss } @@ -383,7 +459,7 @@ func (m *Manager) cacheGet(ctx context.Context, domain string) (*tls.Certificate return tlscert, nil } -func (m *Manager) cachePut(ctx context.Context, domain string, tlscert *tls.Certificate) error { +func (m *Manager) cachePut(ctx context.Context, ck certKey, tlscert *tls.Certificate) error { if m.Cache == nil { return nil } @@ -415,7 +491,7 @@ func (m *Manager) cachePut(ctx context.Context, domain string, tlscert *tls.Cert } } - return m.Cache.Put(ctx, domain, buf.Bytes()) + return m.Cache.Put(ctx, ck.String(), buf.Bytes()) } func encodeECDSAKey(w io.Writer, key *ecdsa.PrivateKey) error { @@ -432,9 +508,9 @@ func encodeECDSAKey(w io.Writer, key *ecdsa.PrivateKey) error { // // If the domain is already being verified, it waits for the existing verification to complete. // Either way, createCert blocks for the duration of the whole process. -func (m *Manager) createCert(ctx context.Context, domain string) (*tls.Certificate, error) { +func (m *Manager) createCert(ctx context.Context, ck certKey) (*tls.Certificate, error) { // TODO: maybe rewrite this whole piece using sync.Once - state, err := m.certState(domain) + state, err := m.certState(ck) if err != nil { return nil, err } @@ -452,44 +528,44 @@ func (m *Manager) createCert(ctx context.Context, domain string) (*tls.Certifica defer state.Unlock() state.locked = false - der, leaf, err := m.authorizedCert(ctx, state.key, domain) + der, leaf, err := m.authorizedCert(ctx, state.key, ck) if err != nil { // Remove the failed state after some time, // making the manager call createCert again on the following TLS hello. time.AfterFunc(createCertRetryAfter, func() { - defer testDidRemoveState(domain) + defer testDidRemoveState(ck) m.stateMu.Lock() defer m.stateMu.Unlock() // Verify the state hasn't changed and it's still invalid // before deleting. - s, ok := m.state[domain] + s, ok := m.state[ck] if !ok { return } - if _, err := validCert(domain, s.cert, s.key); err == nil { + if _, err := validCert(ck, s.cert, s.key); err == nil { return } - delete(m.state, domain) + delete(m.state, ck) }) return nil, err } state.cert = der state.leaf = leaf - go m.renew(domain, state.key, state.leaf.NotAfter) + go m.renew(ck, state.key, state.leaf.NotAfter) return state.tlscert() } // certState returns a new or existing certState. // If a new certState is returned, state.exist is false and the state is locked. // The returned error is non-nil only in the case where a new state could not be created. -func (m *Manager) certState(domain string) (*certState, error) { +func (m *Manager) certState(ck certKey) (*certState, error) { m.stateMu.Lock() defer m.stateMu.Unlock() if m.state == nil { - m.state = make(map[string]*certState) + m.state = make(map[certKey]*certState) } // existing state - if state, ok := m.state[domain]; ok { + if state, ok := m.state[ck]; ok { return state, nil } @@ -498,7 +574,7 @@ func (m *Manager) certState(domain string) (*certState, error) { err error key crypto.Signer ) - if m.ForceRSA { + if ck.isRSA { key, err = rsa.GenerateKey(rand.Reader, 2048) } else { key, err = ecdsa.GenerateKey(elliptic.P256(), rand.Reader) @@ -512,22 +588,22 @@ func (m *Manager) certState(domain string) (*certState, error) { locked: true, } state.Lock() // will be unlocked by m.certState caller - m.state[domain] = state + m.state[ck] = state return state, nil } // authorizedCert starts the domain ownership verification process and requests a new cert upon success. // The key argument is the certificate private key. -func (m *Manager) authorizedCert(ctx context.Context, key crypto.Signer, domain string) (der [][]byte, leaf *x509.Certificate, err error) { +func (m *Manager) authorizedCert(ctx context.Context, key crypto.Signer, ck certKey) (der [][]byte, leaf *x509.Certificate, err error) { client, err := m.acmeClient(ctx) if err != nil { return nil, nil, err } - if err := m.verify(ctx, client, domain); err != nil { + if err := m.verify(ctx, client, ck.domain); err != nil { return nil, nil, err } - csr, err := certRequest(key, domain) + csr, err := certRequest(key, ck.domain, m.ExtraExtensions) if err != nil { return nil, nil, err } @@ -535,13 +611,25 @@ func (m *Manager) authorizedCert(ctx context.Context, key crypto.Signer, domain if err != nil { return nil, nil, err } - leaf, err = validCert(domain, der, key) + leaf, err = validCert(ck, der, key) if err != nil { return nil, nil, err } return der, leaf, nil } +// revokePendingAuthz revokes all authorizations idenfied by the elements of uri slice. +// It ignores revocation errors. +func (m *Manager) revokePendingAuthz(ctx context.Context, uri []string) { + client, err := m.acmeClient(ctx) + if err != nil { + return + } + for _, u := range uri { + client.RevokeAuthorization(ctx, u) + } +} + // verify runs the identifier (domain) authorization flow // using each applicable ACME challenge type. func (m *Manager) verify(ctx context.Context, client *acme.Client, domain string) error { @@ -554,6 +642,24 @@ func (m *Manager) verify(ctx context.Context, client *acme.Client, domain string } m.tokensMu.RUnlock() + // Keep track of pending authzs and revoke the ones that did not validate. + pendingAuthzs := make(map[string]bool) + defer func() { + var uri []string + for k, pending := range pendingAuthzs { + if pending { + uri = append(uri, k) + } + } + if len(uri) > 0 { + // Use "detached" background context. + // The revocations need not happen in the current verification flow. + go m.revokePendingAuthz(context.Background(), uri) + } + }() + + // errs accumulates challenge failure errors, printed if all fail + errs := make(map[*acme.Challenge]error) var nextTyp int // challengeType index of the next challenge type to try for { // Start domain authorization and get the challenge. @@ -570,6 +676,8 @@ func (m *Manager) verify(ctx context.Context, client *acme.Client, domain string return fmt.Errorf("acme/autocert: invalid authorization %q", authz.URI) } + pendingAuthzs[authz.URI] = true + // Pick the next preferred challenge. var chal *acme.Challenge for chal == nil && nextTyp < len(challengeTypes) { @@ -577,21 +685,30 @@ func (m *Manager) verify(ctx context.Context, client *acme.Client, domain string nextTyp++ } if chal == nil { - return fmt.Errorf("acme/autocert: unable to authorize %q; tried %q", domain, challengeTypes) + errorMsg := fmt.Sprintf("acme/autocert: unable to authorize %q", domain) + for chal, err := range errs { + errorMsg += fmt.Sprintf("; challenge %q failed with error: %v", chal.Type, err) + } + return errors.New(errorMsg) } cleanup, err := m.fulfill(ctx, client, chal) if err != nil { + errs[chal] = err continue } defer cleanup() if _, err := client.Accept(ctx, chal); err != nil { + errs[chal] = err continue } // A challenge is fulfilled and accepted: wait for the CA to validate. - if _, err := client.WaitAuthorization(ctx, authz.URI); err == nil { - return nil + if _, err := client.WaitAuthorization(ctx, authz.URI); err != nil { + errs[chal] = err + continue } + delete(pendingAuthzs, authz.URI) + return nil } } @@ -634,8 +751,8 @@ func pickChallenge(typ string, chal []*acme.Challenge) *acme.Challenge { return nil } -// putCertToken stores the cert under the named key in both m.certTokens map -// and m.Cache. +// putCertToken stores the token certificate with the specified name +// in both m.certTokens map and m.Cache. func (m *Manager) putCertToken(ctx context.Context, name string, cert *tls.Certificate) { m.tokensMu.Lock() defer m.tokensMu.Unlock() @@ -643,17 +760,18 @@ func (m *Manager) putCertToken(ctx context.Context, name string, cert *tls.Certi m.certTokens = make(map[string]*tls.Certificate) } m.certTokens[name] = cert - m.cachePut(ctx, name, cert) + m.cachePut(ctx, certKey{domain: name, isToken: true}, cert) } -// deleteCertToken removes the token certificate for the specified domain name +// deleteCertToken removes the token certificate with the specified name // from both m.certTokens map and m.Cache. func (m *Manager) deleteCertToken(name string) { m.tokensMu.Lock() defer m.tokensMu.Unlock() delete(m.certTokens, name) if m.Cache != nil { - m.Cache.Delete(context.Background(), name) + ck := certKey{domain: name, isToken: true} + m.Cache.Delete(context.Background(), ck.String()) } } @@ -704,7 +822,7 @@ func (m *Manager) deleteHTTPToken(tokenPath string) { // httpTokenCacheKey returns a key at which an http-01 token value may be stored // in the Manager's optional Cache. func httpTokenCacheKey(tokenPath string) string { - return "http-01-" + path.Base(tokenPath) + return path.Base(tokenPath) + "+http-01" } // renew starts a cert renewal timer loop, one per domain. @@ -715,18 +833,18 @@ func httpTokenCacheKey(tokenPath string) string { // // The key argument is a certificate private key. // The exp argument is the cert expiration time (NotAfter). -func (m *Manager) renew(domain string, key crypto.Signer, exp time.Time) { +func (m *Manager) renew(ck certKey, key crypto.Signer, exp time.Time) { m.renewalMu.Lock() defer m.renewalMu.Unlock() - if m.renewal[domain] != nil { + if m.renewal[ck] != nil { // another goroutine is already on it return } if m.renewal == nil { - m.renewal = make(map[string]*domainRenewal) + m.renewal = make(map[certKey]*domainRenewal) } - dr := &domainRenewal{m: m, domain: domain, key: key} - m.renewal[domain] = dr + dr := &domainRenewal{m: m, ck: ck, key: key} + m.renewal[ck] = dr dr.start(exp) } @@ -742,7 +860,10 @@ func (m *Manager) stopRenew() { } func (m *Manager) accountKey(ctx context.Context) (crypto.Signer, error) { - const keyName = "acme_account.key" + const keyName = "acme_account+key" + + // Previous versions of autocert stored the value under a different key. + const legacyKeyName = "acme_account.key" genKey := func() (*ecdsa.PrivateKey, error) { return ecdsa.GenerateKey(elliptic.P256(), rand.Reader) @@ -753,6 +874,9 @@ func (m *Manager) accountKey(ctx context.Context) (crypto.Signer, error) { } data, err := m.Cache.Get(ctx, keyName) + if err == ErrCacheMiss { + data, err = m.Cache.Get(ctx, legacyKeyName) + } if err == ErrCacheMiss { key, err := genKey() if err != nil { @@ -849,12 +973,12 @@ func (s *certState) tlscert() (*tls.Certificate, error) { }, nil } -// certRequest creates a certificate request for the given common name cn -// and optional SANs. -func certRequest(key crypto.Signer, cn string, san ...string) ([]byte, error) { +// certRequest generates a CSR for the given common name cn and optional SANs. +func certRequest(key crypto.Signer, cn string, ext []pkix.Extension, san ...string) ([]byte, error) { req := &x509.CertificateRequest{ - Subject: pkix.Name{CommonName: cn}, - DNSNames: san, + Subject: pkix.Name{CommonName: cn}, + DNSNames: san, + ExtraExtensions: ext, } return x509.CreateCertificateRequest(rand.Reader, req, key) } @@ -885,12 +1009,12 @@ func parsePrivateKey(der []byte) (crypto.Signer, error) { return nil, errors.New("acme/autocert: failed to parse private key") } -// validCert parses a cert chain provided as der argument and verifies the leaf, der[0], -// corresponds to the private key, as well as the domain match and expiration dates. -// It doesn't do any revocation checking. +// validCert parses a cert chain provided as der argument and verifies the leaf and der[0] +// correspond to the private key, the domain and key type match, and expiration dates +// are valid. It doesn't do any revocation checking. // // The returned value is the verified leaf cert. -func validCert(domain string, der [][]byte, key crypto.Signer) (leaf *x509.Certificate, err error) { +func validCert(ck certKey, der [][]byte, key crypto.Signer) (leaf *x509.Certificate, err error) { // parse public part(s) var n int for _, b := range der { @@ -902,7 +1026,7 @@ func validCert(domain string, der [][]byte, key crypto.Signer) (leaf *x509.Certi n += copy(pub[n:], b) } x509Cert, err := x509.ParseCertificates(pub) - if len(x509Cert) == 0 { + if err != nil || len(x509Cert) == 0 { return nil, errors.New("acme/autocert: no public key found") } // verify the leaf is not expired and matches the domain name @@ -914,10 +1038,10 @@ func validCert(domain string, der [][]byte, key crypto.Signer) (leaf *x509.Certi if now.After(leaf.NotAfter) { return nil, errors.New("acme/autocert: expired certificate") } - if err := leaf.VerifyHostname(domain); err != nil { + if err := leaf.VerifyHostname(ck.domain); err != nil { return nil, err } - // ensure the leaf corresponds to the private key + // ensure the leaf corresponds to the private key and matches the certKey type switch pub := leaf.PublicKey.(type) { case *rsa.PublicKey: prv, ok := key.(*rsa.PrivateKey) @@ -927,6 +1051,9 @@ func validCert(domain string, der [][]byte, key crypto.Signer) (leaf *x509.Certi if pub.N.Cmp(prv.N) != 0 { return nil, errors.New("acme/autocert: private key does not match public key") } + if !ck.isRSA && !ck.isToken { + return nil, errors.New("acme/autocert: key type does not match expected value") + } case *ecdsa.PublicKey: prv, ok := key.(*ecdsa.PrivateKey) if !ok { @@ -935,6 +1062,9 @@ func validCert(domain string, der [][]byte, key crypto.Signer) (leaf *x509.Certi if pub.X.Cmp(prv.X) != 0 || pub.Y.Cmp(prv.Y) != 0 { return nil, errors.New("acme/autocert: private key does not match public key") } + if ck.isRSA && !ck.isToken { + return nil, errors.New("acme/autocert: key type does not match expected value") + } default: return nil, errors.New("acme/autocert: unknown public key algorithm") } @@ -958,5 +1088,5 @@ var ( timeNow = time.Now // Called when a state is removed. - testDidRemoveState = func(domain string) {} + testDidRemoveState = func(certKey) {} ) diff --git a/vendor/golang.org/x/crypto/acme/autocert/autocert_test.go b/vendor/golang.org/x/crypto/acme/autocert/autocert_test.go index 2da1912e9..48ccd35bc 100644 --- a/vendor/golang.org/x/crypto/acme/autocert/autocert_test.go +++ b/vendor/golang.org/x/crypto/acme/autocert/autocert_test.go @@ -5,6 +5,7 @@ package autocert import ( + "bytes" "context" "crypto" "crypto/ecdsa" @@ -14,6 +15,7 @@ import ( "crypto/tls" "crypto/x509" "crypto/x509/pkix" + "encoding/asn1" "encoding/base64" "encoding/json" "fmt" @@ -31,6 +33,12 @@ import ( "golang.org/x/crypto/acme" ) +var ( + exampleDomain = "example.org" + exampleCertKey = certKey{domain: exampleDomain} + exampleCertKeyRSA = certKey{domain: exampleDomain, isRSA: true} +) + var discoTmpl = template.Must(template.New("disco").Parse(`{ "new-reg": "{{.}}/new-reg", "new-authz": "{{.}}/new-authz", @@ -64,6 +72,7 @@ var authzTmpl = template.Must(template.New("authz").Parse(`{ }`)) type memCache struct { + t *testing.T mu sync.Mutex keyData map[string][]byte } @@ -79,7 +88,26 @@ func (m *memCache) Get(ctx context.Context, key string) ([]byte, error) { return v, nil } +// filenameSafe returns whether all characters in s are printable ASCII +// and safe to use in a filename on most filesystems. +func filenameSafe(s string) bool { + for _, c := range s { + if c < 0x20 || c > 0x7E { + return false + } + switch c { + case '\\', '/', ':', '*', '?', '"', '<', '>', '|': + return false + } + } + return true +} + func (m *memCache) Put(ctx context.Context, key string, data []byte) error { + if !filenameSafe(key) { + m.t.Errorf("invalid characters in cache key %q", key) + } + m.mu.Lock() defer m.mu.Unlock() @@ -95,12 +123,29 @@ func (m *memCache) Delete(ctx context.Context, key string) error { return nil } -func newMemCache() *memCache { +func newMemCache(t *testing.T) *memCache { return &memCache{ + t: t, keyData: make(map[string][]byte), } } +func (m *memCache) numCerts() int { + m.mu.Lock() + defer m.mu.Unlock() + + res := 0 + for key := range m.keyData { + if strings.HasSuffix(key, "+token") || + strings.HasSuffix(key, "+key") || + strings.HasSuffix(key, "+http-01") { + continue + } + res++ + } + return res +} + func dummyCert(pub interface{}, san ...string) ([]byte, error) { return dateDummyCert(pub, time.Now(), time.Now().Add(90*24*time.Hour), san...) } @@ -137,53 +182,58 @@ func decodePayload(v interface{}, r io.Reader) error { return json.Unmarshal(payload, v) } +func clientHelloInfo(sni string, ecdsaSupport bool) *tls.ClientHelloInfo { + hello := &tls.ClientHelloInfo{ + ServerName: sni, + CipherSuites: []uint16{tls.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305}, + } + if ecdsaSupport { + hello.CipherSuites = append(hello.CipherSuites, tls.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305) + } + return hello +} + func TestGetCertificate(t *testing.T) { man := &Manager{Prompt: AcceptTOS} defer man.stopRenew() - hello := &tls.ClientHelloInfo{ServerName: "example.org"} + hello := clientHelloInfo("example.org", true) testGetCertificate(t, man, "example.org", hello) } func TestGetCertificate_trailingDot(t *testing.T) { man := &Manager{Prompt: AcceptTOS} defer man.stopRenew() - hello := &tls.ClientHelloInfo{ServerName: "example.org."} + hello := clientHelloInfo("example.org.", true) testGetCertificate(t, man, "example.org", hello) } func TestGetCertificate_ForceRSA(t *testing.T) { man := &Manager{ Prompt: AcceptTOS, - Cache: newMemCache(), + Cache: newMemCache(t), ForceRSA: true, } defer man.stopRenew() - hello := &tls.ClientHelloInfo{ServerName: "example.org"} - testGetCertificate(t, man, "example.org", hello) + hello := clientHelloInfo(exampleDomain, true) + testGetCertificate(t, man, exampleDomain, hello) - cert, err := man.cacheGet(context.Background(), "example.org") + // ForceRSA was deprecated and is now ignored. + cert, err := man.cacheGet(context.Background(), exampleCertKey) if err != nil { t.Fatalf("man.cacheGet: %v", err) } - if _, ok := cert.PrivateKey.(*rsa.PrivateKey); !ok { - t.Errorf("cert.PrivateKey is %T; want *rsa.PrivateKey", cert.PrivateKey) + if _, ok := cert.PrivateKey.(*ecdsa.PrivateKey); !ok { + t.Errorf("cert.PrivateKey is %T; want *ecdsa.PrivateKey", cert.PrivateKey) } } func TestGetCertificate_nilPrompt(t *testing.T) { man := &Manager{} defer man.stopRenew() - url, finish := startACMEServerStub(t, man, "example.org") + url, finish := startACMEServerStub(t, getCertificateFromManager(man, true), "example.org") defer finish() - key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) - if err != nil { - t.Fatal(err) - } - man.Client = &acme.Client{ - Key: key, - DirectoryURL: url, - } - hello := &tls.ClientHelloInfo{ServerName: "example.org"} + man.Client = &acme.Client{DirectoryURL: url} + hello := clientHelloInfo("example.org", true) if _, err := man.GetCertificate(hello); err == nil { t.Error("got certificate for example.org; wanted error") } @@ -197,7 +247,7 @@ func TestGetCertificate_expiredCache(t *testing.T) { } tmpl := &x509.Certificate{ SerialNumber: big.NewInt(1), - Subject: pkix.Name{CommonName: "example.org"}, + Subject: pkix.Name{CommonName: exampleDomain}, NotAfter: time.Now(), } pub, err := x509.CreateCertificate(rand.Reader, tmpl, tmpl, &pk.PublicKey, pk) @@ -209,16 +259,16 @@ func TestGetCertificate_expiredCache(t *testing.T) { PrivateKey: pk, } - man := &Manager{Prompt: AcceptTOS, Cache: newMemCache()} + man := &Manager{Prompt: AcceptTOS, Cache: newMemCache(t)} defer man.stopRenew() - if err := man.cachePut(context.Background(), "example.org", tlscert); err != nil { + if err := man.cachePut(context.Background(), exampleCertKey, tlscert); err != nil { t.Fatalf("man.cachePut: %v", err) } // The expired cached cert should trigger a new cert issuance // and return without an error. - hello := &tls.ClientHelloInfo{ServerName: "example.org"} - testGetCertificate(t, man, "example.org", hello) + hello := clientHelloInfo(exampleDomain, true) + testGetCertificate(t, man, exampleDomain, hello) } func TestGetCertificate_failedAttempt(t *testing.T) { @@ -227,7 +277,6 @@ func TestGetCertificate_failedAttempt(t *testing.T) { })) defer ts.Close() - const example = "example.org" d := createCertRetryAfter f := testDidRemoveState defer func() { @@ -236,51 +285,167 @@ func TestGetCertificate_failedAttempt(t *testing.T) { }() createCertRetryAfter = 0 done := make(chan struct{}) - testDidRemoveState = func(domain string) { - if domain != example { - t.Errorf("testDidRemoveState: domain = %q; want %q", domain, example) + testDidRemoveState = func(ck certKey) { + if ck != exampleCertKey { + t.Errorf("testDidRemoveState: domain = %v; want %v", ck, exampleCertKey) } close(done) } - key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) - if err != nil { - t.Fatal(err) - } man := &Manager{ Prompt: AcceptTOS, Client: &acme.Client{ - Key: key, DirectoryURL: ts.URL, }, } defer man.stopRenew() - hello := &tls.ClientHelloInfo{ServerName: example} + hello := clientHelloInfo(exampleDomain, true) if _, err := man.GetCertificate(hello); err == nil { t.Error("GetCertificate: err is nil") } select { case <-time.After(5 * time.Second): - t.Errorf("took too long to remove the %q state", example) + t.Errorf("took too long to remove the %q state", exampleCertKey) case <-done: man.stateMu.Lock() defer man.stateMu.Unlock() - if v, exist := man.state[example]; exist { - t.Errorf("state exists for %q: %+v", example, v) + if v, exist := man.state[exampleCertKey]; exist { + t.Errorf("state exists for %v: %+v", exampleCertKey, v) } } } +// testGetCertificate_tokenCache tests the fallback of token certificate fetches +// to cache when Manager.certTokens misses. ecdsaSupport refers to the CA when +// verifying the certificate token. +func testGetCertificate_tokenCache(t *testing.T, ecdsaSupport bool) { + man1 := &Manager{ + Cache: newMemCache(t), + Prompt: AcceptTOS, + } + defer man1.stopRenew() + man2 := &Manager{ + Cache: man1.Cache, + Prompt: AcceptTOS, + } + defer man2.stopRenew() + + // Send the verification request to a different Manager from the one that + // initiated the authorization, when they share caches. + url, finish := startACMEServerStub(t, getCertificateFromManager(man2, ecdsaSupport), "example.org") + defer finish() + man1.Client = &acme.Client{DirectoryURL: url} + hello := clientHelloInfo("example.org", true) + if _, err := man1.GetCertificate(hello); err != nil { + t.Error(err) + } + if _, err := man2.GetCertificate(hello); err != nil { + t.Error(err) + } +} + +func TestGetCertificate_tokenCache(t *testing.T) { + t.Run("ecdsaSupport=true", func(t *testing.T) { + testGetCertificate_tokenCache(t, true) + }) + t.Run("ecdsaSupport=false", func(t *testing.T) { + testGetCertificate_tokenCache(t, false) + }) +} + +func TestGetCertificate_ecdsaVsRSA(t *testing.T) { + cache := newMemCache(t) + man := &Manager{Prompt: AcceptTOS, Cache: cache} + defer man.stopRenew() + url, finish := startACMEServerStub(t, getCertificateFromManager(man, true), "example.org") + defer finish() + man.Client = &acme.Client{DirectoryURL: url} + + cert, err := man.GetCertificate(clientHelloInfo("example.org", true)) + if err != nil { + t.Error(err) + } + if _, ok := cert.Leaf.PublicKey.(*ecdsa.PublicKey); !ok { + t.Error("an ECDSA client was served a non-ECDSA certificate") + } + + cert, err = man.GetCertificate(clientHelloInfo("example.org", false)) + if err != nil { + t.Error(err) + } + if _, ok := cert.Leaf.PublicKey.(*rsa.PublicKey); !ok { + t.Error("a RSA client was served a non-RSA certificate") + } + + if _, err := man.GetCertificate(clientHelloInfo("example.org", true)); err != nil { + t.Error(err) + } + if _, err := man.GetCertificate(clientHelloInfo("example.org", false)); err != nil { + t.Error(err) + } + if numCerts := cache.numCerts(); numCerts != 2 { + t.Errorf("found %d certificates in cache; want %d", numCerts, 2) + } +} + +func TestGetCertificate_wrongCacheKeyType(t *testing.T) { + cache := newMemCache(t) + man := &Manager{Prompt: AcceptTOS, Cache: cache} + defer man.stopRenew() + url, finish := startACMEServerStub(t, getCertificateFromManager(man, true), exampleDomain) + defer finish() + man.Client = &acme.Client{DirectoryURL: url} + + // Make an RSA cert and cache it without suffix. + pk, err := rsa.GenerateKey(rand.Reader, 512) + if err != nil { + t.Fatal(err) + } + tmpl := &x509.Certificate{ + SerialNumber: big.NewInt(1), + Subject: pkix.Name{CommonName: exampleDomain}, + NotAfter: time.Now().Add(90 * 24 * time.Hour), + } + pub, err := x509.CreateCertificate(rand.Reader, tmpl, tmpl, &pk.PublicKey, pk) + if err != nil { + t.Fatal(err) + } + rsaCert := &tls.Certificate{ + Certificate: [][]byte{pub}, + PrivateKey: pk, + } + if err := man.cachePut(context.Background(), exampleCertKey, rsaCert); err != nil { + t.Fatalf("man.cachePut: %v", err) + } + + // The RSA cached cert should be silently ignored and replaced. + cert, err := man.GetCertificate(clientHelloInfo(exampleDomain, true)) + if err != nil { + t.Error(err) + } + if _, ok := cert.Leaf.PublicKey.(*ecdsa.PublicKey); !ok { + t.Error("an ECDSA client was served a non-ECDSA certificate") + } + if numCerts := cache.numCerts(); numCerts != 1 { + t.Errorf("found %d certificates in cache; want %d", numCerts, 1) + } +} + +func getCertificateFromManager(man *Manager, ecdsaSupport bool) func(string) error { + return func(sni string) error { + _, err := man.GetCertificate(clientHelloInfo(sni, ecdsaSupport)) + return err + } +} + // startACMEServerStub runs an ACME server // The domain argument is the expected domain name of a certificate request. -func startACMEServerStub(t *testing.T, man *Manager, domain string) (url string, finish func()) { +func startACMEServerStub(t *testing.T, getCertificate func(string) error, domain string) (url string, finish func()) { // echo token-02 | shasum -a 256 // then divide result in 2 parts separated by dot tokenCertName := "4e8eb87631187e9ff2153b56b13a4dec.13a35d002e485d60ff37354b32f665d9.token.acme.invalid" verifyTokenCert := func() { - hello := &tls.ClientHelloInfo{ServerName: tokenCertName} - _, err := man.GetCertificate(hello) - if err != nil { + if err := getCertificate(tokenCertName); err != nil { t.Errorf("verifyTokenCert: GetCertificate(%q): %v", tokenCertName, err) return } @@ -362,8 +527,7 @@ func startACMEServerStub(t *testing.T, man *Manager, domain string) (url string, tick := time.NewTicker(100 * time.Millisecond) defer tick.Stop() for { - hello := &tls.ClientHelloInfo{ServerName: tokenCertName} - if _, err := man.GetCertificate(hello); err != nil { + if err := getCertificate(tokenCertName); err != nil { return } select { @@ -387,21 +551,13 @@ func startACMEServerStub(t *testing.T, man *Manager, domain string) (url string, // tests man.GetCertificate flow using the provided hello argument. // The domain argument is the expected domain name of a certificate request. func testGetCertificate(t *testing.T, man *Manager, domain string, hello *tls.ClientHelloInfo) { - url, finish := startACMEServerStub(t, man, domain) + url, finish := startACMEServerStub(t, getCertificateFromManager(man, true), domain) defer finish() - - // use EC key to run faster on 386 - key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) - if err != nil { - t.Fatal(err) - } - man.Client = &acme.Client{ - Key: key, - DirectoryURL: url, - } + man.Client = &acme.Client{DirectoryURL: url} // simulate tls.Config.GetCertificate var tlscert *tls.Certificate + var err error done := make(chan struct{}) go func() { tlscert, err = man.GetCertificate(hello) @@ -445,7 +601,7 @@ func TestVerifyHTTP01(t *testing.T) { if w.Code != http.StatusOK { t.Errorf("http token: w.Code = %d; want %d", w.Code, http.StatusOK) } - if v := string(w.Body.Bytes()); !strings.HasPrefix(v, "token-http-01.") { + if v := w.Body.String(); !strings.HasPrefix(v, "token-http-01.") { t.Errorf("http token value = %q; want 'token-http-01.' prefix", v) } } @@ -505,18 +661,18 @@ func TestVerifyHTTP01(t *testing.T) { })) defer ca.Close() - key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) - if err != nil { - t.Fatal(err) - } m := &Manager{ Client: &acme.Client{ - Key: key, DirectoryURL: ca.URL, }, } http01 = m.HTTPHandler(nil) - if err := m.verify(context.Background(), m.Client, "example.org"); err != nil { + ctx := context.Background() + client, err := m.acmeClient(ctx) + if err != nil { + t.Fatalf("m.acmeClient: %v", err) + } + if err := m.verify(ctx, client, "example.org"); err != nil { t.Errorf("m.verify: %v", err) } // Only tls-sni-01, tls-sni-02 and http-01 must be accepted @@ -529,6 +685,111 @@ func TestVerifyHTTP01(t *testing.T) { } } +func TestRevokeFailedAuthz(t *testing.T) { + // Prefill authorization URIs expected to be revoked. + // The challenges are selected in a specific order, + // each tried within a newly created authorization. + // This means each authorization URI corresponds to a different challenge type. + revokedAuthz := map[string]bool{ + "/authz/0": false, // tls-sni-02 + "/authz/1": false, // tls-sni-01 + "/authz/2": false, // no viable challenge, but authz is created + } + + var authzCount int // num. of created authorizations + var revokeCount int // num. of revoked authorizations + done := make(chan struct{}) // closed when revokeCount is 3 + + // ACME CA server stub, only the needed bits. + // TODO: Merge this with startACMEServerStub, making it a configurable CA for testing. + var ca *httptest.Server + ca = httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.Header().Set("Replay-Nonce", "nonce") + if r.Method == "HEAD" { + // a nonce request + return + } + + switch r.URL.Path { + // Discovery. + case "/": + if err := discoTmpl.Execute(w, ca.URL); err != nil { + t.Errorf("discoTmpl: %v", err) + } + // Client key registration. + case "/new-reg": + w.Write([]byte("{}")) + // New domain authorization. + case "/new-authz": + w.Header().Set("Location", fmt.Sprintf("%s/authz/%d", ca.URL, authzCount)) + w.WriteHeader(http.StatusCreated) + if err := authzTmpl.Execute(w, ca.URL); err != nil { + t.Errorf("authzTmpl: %v", err) + } + authzCount++ + // tls-sni-02 challenge "accept" request. + case "/challenge/2": + // Refuse. + http.Error(w, "won't accept tls-sni-02 challenge", http.StatusBadRequest) + // tls-sni-01 challenge "accept" request. + case "/challenge/1": + // Accept but the authorization will be "expired". + w.Write([]byte("{}")) + // Authorization requests. + case "/authz/0", "/authz/1", "/authz/2": + // Revocation requests. + if r.Method == "POST" { + var req struct{ Status string } + if err := decodePayload(&req, r.Body); err != nil { + t.Errorf("%s: decodePayload: %v", r.URL, err) + } + switch req.Status { + case "deactivated": + revokedAuthz[r.URL.Path] = true + revokeCount++ + if revokeCount >= 3 { + // Last authorization is revoked. + defer close(done) + } + default: + t.Errorf("%s: req.Status = %q; want 'deactivated'", r.URL, req.Status) + } + w.Write([]byte(`{"status": "invalid"}`)) + return + } + // Authorization status requests. + // Simulate abandoned authorization, deleted by the CA. + w.WriteHeader(http.StatusNotFound) + default: + http.NotFound(w, r) + t.Errorf("unrecognized r.URL.Path: %s", r.URL.Path) + } + })) + defer ca.Close() + + m := &Manager{ + Client: &acme.Client{DirectoryURL: ca.URL}, + } + // Should fail and revoke 3 authorizations. + // The first 2 are tsl-sni-02 and tls-sni-01 challenges. + // The third time an authorization is created but no viable challenge is found. + // See revokedAuthz above for more explanation. + if _, err := m.createCert(context.Background(), exampleCertKey); err == nil { + t.Errorf("m.createCert returned nil error") + } + select { + case <-time.After(3 * time.Second): + t.Error("revocations took too long") + case <-done: + // revokeCount is at least 3. + } + for uri, ok := range revokedAuthz { + if !ok { + t.Errorf("%q authorization was not revoked", uri) + } + } +} + func TestHTTPHandlerDefaultFallback(t *testing.T) { tt := []struct { method, url string @@ -571,7 +832,7 @@ func TestHTTPHandlerDefaultFallback(t *testing.T) { } func TestAccountKeyCache(t *testing.T) { - m := Manager{Cache: newMemCache()} + m := Manager{Cache: newMemCache(t)} ctx := context.Background() k1, err := m.accountKey(ctx) if err != nil { @@ -587,36 +848,57 @@ func TestAccountKeyCache(t *testing.T) { } func TestCache(t *testing.T) { - privKey, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) + ecdsaKey, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) if err != nil { t.Fatal(err) } - tmpl := &x509.Certificate{ - SerialNumber: big.NewInt(1), - Subject: pkix.Name{CommonName: "example.org"}, - NotAfter: time.Now().Add(time.Hour), - } - pub, err := x509.CreateCertificate(rand.Reader, tmpl, tmpl, &privKey.PublicKey, privKey) + cert, err := dummyCert(ecdsaKey.Public(), exampleDomain) if err != nil { t.Fatal(err) } - tlscert := &tls.Certificate{ - Certificate: [][]byte{pub}, - PrivateKey: privKey, + ecdsaCert := &tls.Certificate{ + Certificate: [][]byte{cert}, + PrivateKey: ecdsaKey, } - man := &Manager{Cache: newMemCache()} + rsaKey, err := rsa.GenerateKey(rand.Reader, 512) + if err != nil { + t.Fatal(err) + } + cert, err = dummyCert(rsaKey.Public(), exampleDomain) + if err != nil { + t.Fatal(err) + } + rsaCert := &tls.Certificate{ + Certificate: [][]byte{cert}, + PrivateKey: rsaKey, + } + + man := &Manager{Cache: newMemCache(t)} defer man.stopRenew() ctx := context.Background() - if err := man.cachePut(ctx, "example.org", tlscert); err != nil { + + if err := man.cachePut(ctx, exampleCertKey, ecdsaCert); err != nil { t.Fatalf("man.cachePut: %v", err) } - res, err := man.cacheGet(ctx, "example.org") + if err := man.cachePut(ctx, exampleCertKeyRSA, rsaCert); err != nil { + t.Fatalf("man.cachePut: %v", err) + } + + res, err := man.cacheGet(ctx, exampleCertKey) if err != nil { t.Fatalf("man.cacheGet: %v", err) } - if res == nil { - t.Fatal("res is nil") + if res == nil || !bytes.Equal(res.Certificate[0], ecdsaCert.Certificate[0]) { + t.Errorf("man.cacheGet = %+v; want %+v", res, ecdsaCert) + } + + res, err = man.cacheGet(ctx, exampleCertKeyRSA) + if err != nil { + t.Fatalf("man.cacheGet: %v", err) + } + if res == nil || !bytes.Equal(res.Certificate[0], rsaCert.Certificate[0]) { + t.Errorf("man.cacheGet = %+v; want %+v", res, rsaCert) } } @@ -680,26 +962,28 @@ func TestValidCert(t *testing.T) { } tt := []struct { - domain string - key crypto.Signer - cert [][]byte - ok bool + ck certKey + key crypto.Signer + cert [][]byte + ok bool }{ - {"example.org", key1, [][]byte{cert1}, true}, - {"example.org", key3, [][]byte{cert3}, true}, - {"example.org", key1, [][]byte{cert1, cert2, cert3}, true}, - {"example.org", key1, [][]byte{cert1, {1}}, false}, - {"example.org", key1, [][]byte{{1}}, false}, - {"example.org", key1, [][]byte{cert2}, false}, - {"example.org", key2, [][]byte{cert1}, false}, - {"example.org", key1, [][]byte{cert3}, false}, - {"example.org", key3, [][]byte{cert1}, false}, - {"example.net", key1, [][]byte{cert1}, false}, - {"example.org", key1, [][]byte{early}, false}, - {"example.org", key1, [][]byte{expired}, false}, + {certKey{domain: "example.org"}, key1, [][]byte{cert1}, true}, + {certKey{domain: "example.org", isRSA: true}, key3, [][]byte{cert3}, true}, + {certKey{domain: "example.org"}, key1, [][]byte{cert1, cert2, cert3}, true}, + {certKey{domain: "example.org"}, key1, [][]byte{cert1, {1}}, false}, + {certKey{domain: "example.org"}, key1, [][]byte{{1}}, false}, + {certKey{domain: "example.org"}, key1, [][]byte{cert2}, false}, + {certKey{domain: "example.org"}, key2, [][]byte{cert1}, false}, + {certKey{domain: "example.org"}, key1, [][]byte{cert3}, false}, + {certKey{domain: "example.org"}, key3, [][]byte{cert1}, false}, + {certKey{domain: "example.net"}, key1, [][]byte{cert1}, false}, + {certKey{domain: "example.org"}, key1, [][]byte{early}, false}, + {certKey{domain: "example.org"}, key1, [][]byte{expired}, false}, + {certKey{domain: "example.org", isRSA: true}, key1, [][]byte{cert1}, false}, + {certKey{domain: "example.org"}, key3, [][]byte{cert3}, false}, } for i, test := range tt { - leaf, err := validCert(test.domain, test.cert, test.key) + leaf, err := validCert(test.ck, test.cert, test.key) if err != nil && test.ok { t.Errorf("%d: err = %v", i, err) } @@ -748,10 +1032,99 @@ func TestManagerGetCertificateBogusSNI(t *testing.T) { {"fo.o", "cache.Get of fo.o"}, } for _, tt := range tests { - _, err := m.GetCertificate(&tls.ClientHelloInfo{ServerName: tt.name}) + _, err := m.GetCertificate(clientHelloInfo(tt.name, true)) got := fmt.Sprint(err) if got != tt.wantErr { t.Errorf("GetCertificate(SNI = %q) = %q; want %q", tt.name, got, tt.wantErr) } } } + +func TestCertRequest(t *testing.T) { + key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) + if err != nil { + t.Fatal(err) + } + // An extension from RFC7633. Any will do. + ext := pkix.Extension{ + Id: asn1.ObjectIdentifier{1, 3, 6, 1, 5, 5, 7, 1}, + Value: []byte("dummy"), + } + b, err := certRequest(key, "example.org", []pkix.Extension{ext}, "san.example.org") + if err != nil { + t.Fatalf("certRequest: %v", err) + } + r, err := x509.ParseCertificateRequest(b) + if err != nil { + t.Fatalf("ParseCertificateRequest: %v", err) + } + var found bool + for _, v := range r.Extensions { + if v.Id.Equal(ext.Id) { + found = true + break + } + } + if !found { + t.Errorf("want %v in Extensions: %v", ext, r.Extensions) + } +} + +func TestSupportsECDSA(t *testing.T) { + tests := []struct { + CipherSuites []uint16 + SignatureSchemes []tls.SignatureScheme + SupportedCurves []tls.CurveID + ecdsaOk bool + }{ + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, + }, nil, nil, false}, + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + }, nil, nil, true}, + + // SignatureSchemes limits, not extends, CipherSuites + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, + }, []tls.SignatureScheme{ + tls.PKCS1WithSHA256, tls.ECDSAWithP256AndSHA256, + }, nil, false}, + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + }, []tls.SignatureScheme{ + tls.PKCS1WithSHA256, + }, nil, false}, + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + }, []tls.SignatureScheme{ + tls.PKCS1WithSHA256, tls.ECDSAWithP256AndSHA256, + }, nil, true}, + + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + }, []tls.SignatureScheme{ + tls.PKCS1WithSHA256, tls.ECDSAWithP256AndSHA256, + }, []tls.CurveID{ + tls.CurveP521, + }, false}, + {[]uint16{ + tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256, tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256, + }, []tls.SignatureScheme{ + tls.PKCS1WithSHA256, tls.ECDSAWithP256AndSHA256, + }, []tls.CurveID{ + tls.CurveP256, + tls.CurveP521, + }, true}, + } + for i, tt := range tests { + result := supportsECDSA(&tls.ClientHelloInfo{ + CipherSuites: tt.CipherSuites, + SignatureSchemes: tt.SignatureSchemes, + SupportedCurves: tt.SupportedCurves, + }) + if result != tt.ecdsaOk { + t.Errorf("%d: supportsECDSA = %v; want %v", i, result, tt.ecdsaOk) + } + } +} diff --git a/vendor/golang.org/x/crypto/acme/autocert/cache.go b/vendor/golang.org/x/crypto/acme/autocert/cache.go index 61a5fd239..aa9aa845c 100644 --- a/vendor/golang.org/x/crypto/acme/autocert/cache.go +++ b/vendor/golang.org/x/crypto/acme/autocert/cache.go @@ -16,10 +16,10 @@ import ( var ErrCacheMiss = errors.New("acme/autocert: certificate cache miss") // Cache is used by Manager to store and retrieve previously obtained certificates -// as opaque data. +// and other account data as opaque blobs. // -// The key argument of the methods refers to a domain name but need not be an FQDN. -// Cache implementations should not rely on the key naming pattern. +// Cache implementations should not rely on the key naming pattern. Keys can +// include any printable ASCII characters, except the following: \/:*?"<>| type Cache interface { // Get returns a certificate data for the specified key. // If there's no such key, Get returns ErrCacheMiss. diff --git a/vendor/golang.org/x/crypto/acme/autocert/renewal.go b/vendor/golang.org/x/crypto/acme/autocert/renewal.go index 3fa4d61a2..ef3e44e19 100644 --- a/vendor/golang.org/x/crypto/acme/autocert/renewal.go +++ b/vendor/golang.org/x/crypto/acme/autocert/renewal.go @@ -17,9 +17,9 @@ const renewJitter = time.Hour // domainRenewal tracks the state used by the periodic timers // renewing a single domain's cert. type domainRenewal struct { - m *Manager - domain string - key crypto.Signer + m *Manager + ck certKey + key crypto.Signer timerMu sync.Mutex timer *time.Timer @@ -77,7 +77,7 @@ func (dr *domainRenewal) updateState(state *certState) { dr.m.stateMu.Lock() defer dr.m.stateMu.Unlock() dr.key = state.key - dr.m.state[dr.domain] = state + dr.m.state[dr.ck] = state } // do is similar to Manager.createCert but it doesn't lock a Manager.state item. @@ -91,7 +91,7 @@ func (dr *domainRenewal) updateState(state *certState) { func (dr *domainRenewal) do(ctx context.Context) (time.Duration, error) { // a race is likely unavoidable in a distributed environment // but we try nonetheless - if tlscert, err := dr.m.cacheGet(ctx, dr.domain); err == nil { + if tlscert, err := dr.m.cacheGet(ctx, dr.ck); err == nil { next := dr.next(tlscert.Leaf.NotAfter) if next > dr.m.renewBefore()+renewJitter { signer, ok := tlscert.PrivateKey.(crypto.Signer) @@ -107,7 +107,7 @@ func (dr *domainRenewal) do(ctx context.Context) (time.Duration, error) { } } - der, leaf, err := dr.m.authorizedCert(ctx, dr.key, dr.domain) + der, leaf, err := dr.m.authorizedCert(ctx, dr.key, dr.ck) if err != nil { return 0, err } @@ -120,7 +120,7 @@ func (dr *domainRenewal) do(ctx context.Context) (time.Duration, error) { if err != nil { return 0, err } - if err := dr.m.cachePut(ctx, dr.domain, tlscert); err != nil { + if err := dr.m.cachePut(ctx, dr.ck, tlscert); err != nil { return 0, err } dr.updateState(state) diff --git a/vendor/golang.org/x/crypto/acme/autocert/renewal_test.go b/vendor/golang.org/x/crypto/acme/autocert/renewal_test.go index 6e88672bd..634305a13 100644 --- a/vendor/golang.org/x/crypto/acme/autocert/renewal_test.go +++ b/vendor/golang.org/x/crypto/acme/autocert/renewal_test.go @@ -48,8 +48,6 @@ func TestRenewalNext(t *testing.T) { } func TestRenewFromCache(t *testing.T) { - const domain = "example.org" - // ACME CA server stub var ca *httptest.Server ca = httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { @@ -84,7 +82,7 @@ func TestRenewFromCache(t *testing.T) { if err != nil { t.Fatalf("new-cert: CSR: %v", err) } - der, err := dummyCert(csr.PublicKey, domain) + der, err := dummyCert(csr.PublicKey, exampleDomain) if err != nil { t.Fatalf("new-cert: dummyCert: %v", err) } @@ -105,30 +103,28 @@ func TestRenewFromCache(t *testing.T) { })) defer ca.Close() - // use EC key to run faster on 386 - key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) - if err != nil { - t.Fatal(err) - } man := &Manager{ Prompt: AcceptTOS, - Cache: newMemCache(), + Cache: newMemCache(t), RenewBefore: 24 * time.Hour, Client: &acme.Client{ - Key: key, DirectoryURL: ca.URL, }, } defer man.stopRenew() // cache an almost expired cert + key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) + if err != nil { + t.Fatal(err) + } now := time.Now() - cert, err := dateDummyCert(key.Public(), now.Add(-2*time.Hour), now.Add(time.Minute), domain) + cert, err := dateDummyCert(key.Public(), now.Add(-2*time.Hour), now.Add(time.Minute), exampleDomain) if err != nil { t.Fatal(err) } tlscert := &tls.Certificate{PrivateKey: key, Certificate: [][]byte{cert}} - if err := man.cachePut(context.Background(), domain, tlscert); err != nil { + if err := man.cachePut(context.Background(), exampleCertKey, tlscert); err != nil { t.Fatal(err) } @@ -152,7 +148,7 @@ func TestRenewFromCache(t *testing.T) { // ensure the new cert is cached after := time.Now().Add(future) - tlscert, err := man.cacheGet(context.Background(), domain) + tlscert, err := man.cacheGet(context.Background(), exampleCertKey) if err != nil { t.Fatalf("man.cacheGet: %v", err) } @@ -163,9 +159,9 @@ func TestRenewFromCache(t *testing.T) { // verify the old cert is also replaced in memory man.stateMu.Lock() defer man.stateMu.Unlock() - s := man.state[domain] + s := man.state[exampleCertKey] if s == nil { - t.Fatalf("m.state[%q] is nil", domain) + t.Fatalf("m.state[%q] is nil", exampleCertKey) } tlscert, err = s.tlscert() if err != nil { @@ -177,7 +173,7 @@ func TestRenewFromCache(t *testing.T) { } // trigger renew - hello := &tls.ClientHelloInfo{ServerName: domain} + hello := clientHelloInfo(exampleDomain, true) if _, err := man.GetCertificate(hello); err != nil { t.Fatal(err) } @@ -191,19 +187,11 @@ func TestRenewFromCache(t *testing.T) { } func TestRenewFromCacheAlreadyRenewed(t *testing.T) { - const domain = "example.org" - - // use EC key to run faster on 386 - key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) - if err != nil { - t.Fatal(err) - } man := &Manager{ Prompt: AcceptTOS, - Cache: newMemCache(), + Cache: newMemCache(t), RenewBefore: 24 * time.Hour, Client: &acme.Client{ - Key: key, DirectoryURL: "invalid", }, } @@ -215,38 +203,42 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { t.Fatal(err) } now := time.Now() - newCert, err := dateDummyCert(newKey.Public(), now.Add(-2*time.Hour), now.Add(time.Hour*24*90), domain) + newCert, err := dateDummyCert(newKey.Public(), now.Add(-2*time.Hour), now.Add(time.Hour*24*90), exampleDomain) if err != nil { t.Fatal(err) } - newLeaf, err := validCert(domain, [][]byte{newCert}, newKey) + newLeaf, err := validCert(exampleCertKey, [][]byte{newCert}, newKey) if err != nil { t.Fatal(err) } newTLSCert := &tls.Certificate{PrivateKey: newKey, Certificate: [][]byte{newCert}, Leaf: newLeaf} - if err := man.cachePut(context.Background(), domain, newTLSCert); err != nil { + if err := man.cachePut(context.Background(), exampleCertKey, newTLSCert); err != nil { t.Fatal(err) } // set internal state to an almost expired cert - oldCert, err := dateDummyCert(key.Public(), now.Add(-2*time.Hour), now.Add(time.Minute), domain) + key, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) if err != nil { t.Fatal(err) } - oldLeaf, err := validCert(domain, [][]byte{oldCert}, key) + oldCert, err := dateDummyCert(key.Public(), now.Add(-2*time.Hour), now.Add(time.Minute), exampleDomain) + if err != nil { + t.Fatal(err) + } + oldLeaf, err := validCert(exampleCertKey, [][]byte{oldCert}, key) if err != nil { t.Fatal(err) } man.stateMu.Lock() if man.state == nil { - man.state = make(map[string]*certState) + man.state = make(map[certKey]*certState) } s := &certState{ key: key, cert: [][]byte{oldCert}, leaf: oldLeaf, } - man.state[domain] = s + man.state[exampleCertKey] = s man.stateMu.Unlock() // veriy the renewal accepted the newer cached cert @@ -267,7 +259,7 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { } // ensure the cached cert was not modified - tlscert, err := man.cacheGet(context.Background(), domain) + tlscert, err := man.cacheGet(context.Background(), exampleCertKey) if err != nil { t.Fatalf("man.cacheGet: %v", err) } @@ -278,9 +270,9 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { // verify the old cert is also replaced in memory man.stateMu.Lock() defer man.stateMu.Unlock() - s := man.state[domain] + s := man.state[exampleCertKey] if s == nil { - t.Fatalf("m.state[%q] is nil", domain) + t.Fatalf("m.state[%q] is nil", exampleCertKey) } stateKey := s.key.Public().(*ecdsa.PublicKey) if stateKey.X.Cmp(newKey.X) != 0 || stateKey.Y.Cmp(newKey.Y) != 0 { @@ -295,9 +287,9 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { } // verify the private key is replaced in the renewal state - r := man.renewal[domain] + r := man.renewal[exampleCertKey] if r == nil { - t.Fatalf("m.renewal[%q] is nil", domain) + t.Fatalf("m.renewal[%q] is nil", exampleCertKey) } renewalKey := r.key.Public().(*ecdsa.PublicKey) if renewalKey.X.Cmp(newKey.X) != 0 || renewalKey.Y.Cmp(newKey.Y) != 0 { @@ -307,7 +299,7 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { } // assert the expiring cert is returned from state - hello := &tls.ClientHelloInfo{ServerName: domain} + hello := clientHelloInfo(exampleDomain, true) tlscert, err := man.GetCertificate(hello) if err != nil { t.Fatal(err) @@ -317,7 +309,7 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { } // trigger renew - go man.renew(domain, s.key, s.leaf.NotAfter) + go man.renew(exampleCertKey, s.key, s.leaf.NotAfter) // wait for renew loop select { @@ -325,7 +317,7 @@ func TestRenewFromCacheAlreadyRenewed(t *testing.T) { t.Fatal("renew took too long to occur") case <-done: // assert the new cert is returned from state after renew - hello := &tls.ClientHelloInfo{ServerName: domain} + hello := clientHelloInfo(exampleDomain, true) tlscert, err := man.GetCertificate(hello) if err != nil { t.Fatal(err) diff --git a/vendor/golang.org/x/crypto/acme/http.go b/vendor/golang.org/x/crypto/acme/http.go new file mode 100644 index 000000000..a43ce6a5f --- /dev/null +++ b/vendor/golang.org/x/crypto/acme/http.go @@ -0,0 +1,281 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package acme + +import ( + "bytes" + "context" + "crypto" + "crypto/rand" + "encoding/json" + "fmt" + "io/ioutil" + "math/big" + "net/http" + "strconv" + "strings" + "time" +) + +// retryTimer encapsulates common logic for retrying unsuccessful requests. +// It is not safe for concurrent use. +type retryTimer struct { + // backoffFn provides backoff delay sequence for retries. + // See Client.RetryBackoff doc comment. + backoffFn func(n int, r *http.Request, res *http.Response) time.Duration + // n is the current retry attempt. + n int +} + +func (t *retryTimer) inc() { + t.n++ +} + +// backoff pauses the current goroutine as described in Client.RetryBackoff. +func (t *retryTimer) backoff(ctx context.Context, r *http.Request, res *http.Response) error { + d := t.backoffFn(t.n, r, res) + if d <= 0 { + return fmt.Errorf("acme: no more retries for %s; tried %d time(s)", r.URL, t.n) + } + wakeup := time.NewTimer(d) + defer wakeup.Stop() + select { + case <-ctx.Done(): + return ctx.Err() + case <-wakeup.C: + return nil + } +} + +func (c *Client) retryTimer() *retryTimer { + f := c.RetryBackoff + if f == nil { + f = defaultBackoff + } + return &retryTimer{backoffFn: f} +} + +// defaultBackoff provides default Client.RetryBackoff implementation +// using a truncated exponential backoff algorithm, +// as described in Client.RetryBackoff. +// +// The n argument is always bounded between 1 and 30. +// The returned value is always greater than 0. +func defaultBackoff(n int, r *http.Request, res *http.Response) time.Duration { + const max = 10 * time.Second + var jitter time.Duration + if x, err := rand.Int(rand.Reader, big.NewInt(1000)); err == nil { + // Set the minimum to 1ms to avoid a case where + // an invalid Retry-After value is parsed into 0 below, + // resulting in the 0 returned value which would unintentionally + // stop the retries. + jitter = (1 + time.Duration(x.Int64())) * time.Millisecond + } + if v, ok := res.Header["Retry-After"]; ok { + return retryAfter(v[0]) + jitter + } + + if n < 1 { + n = 1 + } + if n > 30 { + n = 30 + } + d := time.Duration(1< max { + return max + } + return d +} + +// retryAfter parses a Retry-After HTTP header value, +// trying to convert v into an int (seconds) or use http.ParseTime otherwise. +// It returns zero value if v cannot be parsed. +func retryAfter(v string) time.Duration { + if i, err := strconv.Atoi(v); err == nil { + return time.Duration(i) * time.Second + } + t, err := http.ParseTime(v) + if err != nil { + return 0 + } + return t.Sub(timeNow()) +} + +// resOkay is a function that reports whether the provided response is okay. +// It is expected to keep the response body unread. +type resOkay func(*http.Response) bool + +// wantStatus returns a function which reports whether the code +// matches the status code of a response. +func wantStatus(codes ...int) resOkay { + return func(res *http.Response) bool { + for _, code := range codes { + if code == res.StatusCode { + return true + } + } + return false + } +} + +// get issues an unsigned GET request to the specified URL. +// It returns a non-error value only when ok reports true. +// +// get retries unsuccessful attempts according to c.RetryBackoff +// until the context is done or a non-retriable error is received. +func (c *Client) get(ctx context.Context, url string, ok resOkay) (*http.Response, error) { + retry := c.retryTimer() + for { + req, err := http.NewRequest("GET", url, nil) + if err != nil { + return nil, err + } + res, err := c.doNoRetry(ctx, req) + switch { + case err != nil: + return nil, err + case ok(res): + return res, nil + case isRetriable(res.StatusCode): + retry.inc() + resErr := responseError(res) + res.Body.Close() + // Ignore the error value from retry.backoff + // and return the one from last retry, as received from the CA. + if retry.backoff(ctx, req, res) != nil { + return nil, resErr + } + default: + defer res.Body.Close() + return nil, responseError(res) + } + } +} + +// post issues a signed POST request in JWS format using the provided key +// to the specified URL. +// It returns a non-error value only when ok reports true. +// +// post retries unsuccessful attempts according to c.RetryBackoff +// until the context is done or a non-retriable error is received. +// It uses postNoRetry to make individual requests. +func (c *Client) post(ctx context.Context, key crypto.Signer, url string, body interface{}, ok resOkay) (*http.Response, error) { + retry := c.retryTimer() + for { + res, req, err := c.postNoRetry(ctx, key, url, body) + if err != nil { + return nil, err + } + if ok(res) { + return res, nil + } + resErr := responseError(res) + res.Body.Close() + switch { + // Check for bad nonce before isRetriable because it may have been returned + // with an unretriable response code such as 400 Bad Request. + case isBadNonce(resErr): + // Consider any previously stored nonce values to be invalid. + c.clearNonces() + case !isRetriable(res.StatusCode): + return nil, resErr + } + retry.inc() + // Ignore the error value from retry.backoff + // and return the one from last retry, as received from the CA. + if err := retry.backoff(ctx, req, res); err != nil { + return nil, resErr + } + } +} + +// postNoRetry signs the body with the given key and POSTs it to the provided url. +// The body argument must be JSON-serializable. +// It is used by c.post to retry unsuccessful attempts. +func (c *Client) postNoRetry(ctx context.Context, key crypto.Signer, url string, body interface{}) (*http.Response, *http.Request, error) { + nonce, err := c.popNonce(ctx, url) + if err != nil { + return nil, nil, err + } + b, err := jwsEncodeJSON(body, key, nonce) + if err != nil { + return nil, nil, err + } + req, err := http.NewRequest("POST", url, bytes.NewReader(b)) + if err != nil { + return nil, nil, err + } + req.Header.Set("Content-Type", "application/jose+json") + res, err := c.doNoRetry(ctx, req) + if err != nil { + return nil, nil, err + } + c.addNonce(res.Header) + return res, req, nil +} + +// doNoRetry issues a request req, replacing its context (if any) with ctx. +func (c *Client) doNoRetry(ctx context.Context, req *http.Request) (*http.Response, error) { + res, err := c.httpClient().Do(req.WithContext(ctx)) + if err != nil { + select { + case <-ctx.Done(): + // Prefer the unadorned context error. + // (The acme package had tests assuming this, previously from ctxhttp's + // behavior, predating net/http supporting contexts natively) + // TODO(bradfitz): reconsider this in the future. But for now this + // requires no test updates. + return nil, ctx.Err() + default: + return nil, err + } + } + return res, nil +} + +func (c *Client) httpClient() *http.Client { + if c.HTTPClient != nil { + return c.HTTPClient + } + return http.DefaultClient +} + +// isBadNonce reports whether err is an ACME "badnonce" error. +func isBadNonce(err error) bool { + // According to the spec badNonce is urn:ietf:params:acme:error:badNonce. + // However, ACME servers in the wild return their versions of the error. + // See https://tools.ietf.org/html/draft-ietf-acme-acme-02#section-5.4 + // and https://github.com/letsencrypt/boulder/blob/0e07eacb/docs/acme-divergences.md#section-66. + ae, ok := err.(*Error) + return ok && strings.HasSuffix(strings.ToLower(ae.ProblemType), ":badnonce") +} + +// isRetriable reports whether a request can be retried +// based on the response status code. +// +// Note that a "bad nonce" error is returned with a non-retriable 400 Bad Request code. +// Callers should parse the response and check with isBadNonce. +func isRetriable(code int) bool { + return code <= 399 || code >= 500 || code == http.StatusTooManyRequests +} + +// responseError creates an error of Error type from resp. +func responseError(resp *http.Response) error { + // don't care if ReadAll returns an error: + // json.Unmarshal will fail in that case anyway + b, _ := ioutil.ReadAll(resp.Body) + e := &wireError{Status: resp.StatusCode} + if err := json.Unmarshal(b, e); err != nil { + // this is not a regular error response: + // populate detail with anything we received, + // e.Status will already contain HTTP response code value + e.Detail = string(b) + if e.Detail == "" { + e.Detail = resp.Status + } + } + return e.error(resp.Header) +} diff --git a/vendor/golang.org/x/crypto/acme/http_test.go b/vendor/golang.org/x/crypto/acme/http_test.go new file mode 100644 index 000000000..15e401ba3 --- /dev/null +++ b/vendor/golang.org/x/crypto/acme/http_test.go @@ -0,0 +1,209 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package acme + +import ( + "context" + "fmt" + "io/ioutil" + "net/http" + "net/http/httptest" + "reflect" + "strings" + "testing" + "time" +) + +func TestDefaultBackoff(t *testing.T) { + tt := []struct { + nretry int + retryAfter string // Retry-After header + out time.Duration // expected min; max = min + jitter + }{ + {-1, "", time.Second}, // verify the lower bound is 1 + {0, "", time.Second}, // verify the lower bound is 1 + {100, "", 10 * time.Second}, // verify the ceiling + {1, "3600", time.Hour}, // verify the header value is used + {1, "", 1 * time.Second}, + {2, "", 2 * time.Second}, + {3, "", 4 * time.Second}, + {4, "", 8 * time.Second}, + } + for i, test := range tt { + r := httptest.NewRequest("GET", "/", nil) + resp := &http.Response{Header: http.Header{}} + if test.retryAfter != "" { + resp.Header.Set("Retry-After", test.retryAfter) + } + d := defaultBackoff(test.nretry, r, resp) + max := test.out + time.Second // + max jitter + if d < test.out || max < d { + t.Errorf("%d: defaultBackoff(%v) = %v; want between %v and %v", i, test.nretry, d, test.out, max) + } + } +} + +func TestErrorResponse(t *testing.T) { + s := `{ + "status": 400, + "type": "urn:acme:error:xxx", + "detail": "text" + }` + res := &http.Response{ + StatusCode: 400, + Status: "400 Bad Request", + Body: ioutil.NopCloser(strings.NewReader(s)), + Header: http.Header{"X-Foo": {"bar"}}, + } + err := responseError(res) + v, ok := err.(*Error) + if !ok { + t.Fatalf("err = %+v (%T); want *Error type", err, err) + } + if v.StatusCode != 400 { + t.Errorf("v.StatusCode = %v; want 400", v.StatusCode) + } + if v.ProblemType != "urn:acme:error:xxx" { + t.Errorf("v.ProblemType = %q; want urn:acme:error:xxx", v.ProblemType) + } + if v.Detail != "text" { + t.Errorf("v.Detail = %q; want text", v.Detail) + } + if !reflect.DeepEqual(v.Header, res.Header) { + t.Errorf("v.Header = %+v; want %+v", v.Header, res.Header) + } +} + +func TestPostWithRetries(t *testing.T) { + var count int + ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + count++ + w.Header().Set("Replay-Nonce", fmt.Sprintf("nonce%d", count)) + if r.Method == "HEAD" { + // We expect the client to do 2 head requests to fetch + // nonces, one to start and another after getting badNonce + return + } + + head, err := decodeJWSHead(r) + switch { + case err != nil: + t.Errorf("decodeJWSHead: %v", err) + case head.Nonce == "": + t.Error("head.Nonce is empty") + case head.Nonce == "nonce1": + // Return a badNonce error to force the call to retry. + w.Header().Set("Retry-After", "0") + w.WriteHeader(http.StatusBadRequest) + w.Write([]byte(`{"type":"urn:ietf:params:acme:error:badNonce"}`)) + return + } + // Make client.Authorize happy; we're not testing its result. + w.WriteHeader(http.StatusCreated) + w.Write([]byte(`{"status":"valid"}`)) + })) + defer ts.Close() + + client := &Client{Key: testKey, dir: &Directory{AuthzURL: ts.URL}} + // This call will fail with badNonce, causing a retry + if _, err := client.Authorize(context.Background(), "example.com"); err != nil { + t.Errorf("client.Authorize 1: %v", err) + } + if count != 4 { + t.Errorf("total requests count: %d; want 4", count) + } +} + +func TestRetryErrorType(t *testing.T) { + ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.Header().Set("Replay-Nonce", "nonce") + w.WriteHeader(http.StatusTooManyRequests) + w.Write([]byte(`{"type":"rateLimited"}`)) + })) + defer ts.Close() + + client := &Client{ + Key: testKey, + RetryBackoff: func(n int, r *http.Request, res *http.Response) time.Duration { + // Do no retries. + return 0 + }, + dir: &Directory{AuthzURL: ts.URL}, + } + + t.Run("post", func(t *testing.T) { + testRetryErrorType(t, func() error { + _, err := client.Authorize(context.Background(), "example.com") + return err + }) + }) + t.Run("get", func(t *testing.T) { + testRetryErrorType(t, func() error { + _, err := client.GetAuthorization(context.Background(), ts.URL) + return err + }) + }) +} + +func testRetryErrorType(t *testing.T, callClient func() error) { + t.Helper() + err := callClient() + if err == nil { + t.Fatal("client.Authorize returned nil error") + } + acmeErr, ok := err.(*Error) + if !ok { + t.Fatalf("err is %v (%T); want *Error", err, err) + } + if acmeErr.StatusCode != http.StatusTooManyRequests { + t.Errorf("acmeErr.StatusCode = %d; want %d", acmeErr.StatusCode, http.StatusTooManyRequests) + } + if acmeErr.ProblemType != "rateLimited" { + t.Errorf("acmeErr.ProblemType = %q; want 'rateLimited'", acmeErr.ProblemType) + } +} + +func TestRetryBackoffArgs(t *testing.T) { + const resCode = http.StatusInternalServerError + ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) { + w.Header().Set("Replay-Nonce", "test-nonce") + w.WriteHeader(resCode) + })) + defer ts.Close() + + // Canceled in backoff. + ctx, cancel := context.WithCancel(context.Background()) + + var nretry int + backoff := func(n int, r *http.Request, res *http.Response) time.Duration { + nretry++ + if n != nretry { + t.Errorf("n = %d; want %d", n, nretry) + } + if nretry == 3 { + cancel() + } + + if r == nil { + t.Error("r is nil") + } + if res.StatusCode != resCode { + t.Errorf("res.StatusCode = %d; want %d", res.StatusCode, resCode) + } + return time.Millisecond + } + + client := &Client{ + Key: testKey, + RetryBackoff: backoff, + dir: &Directory{AuthzURL: ts.URL}, + } + if _, err := client.Authorize(ctx, "example.com"); err == nil { + t.Error("err is nil") + } + if nretry != 3 { + t.Errorf("nretry = %d; want 3", nretry) + } +} diff --git a/vendor/golang.org/x/crypto/acme/types.go b/vendor/golang.org/x/crypto/acme/types.go index 3e199749e..54792c065 100644 --- a/vendor/golang.org/x/crypto/acme/types.go +++ b/vendor/golang.org/x/crypto/acme/types.go @@ -104,7 +104,7 @@ func RateLimit(err error) (time.Duration, bool) { if e.Header == nil { return 0, true } - return retryAfter(e.Header.Get("Retry-After"), 0), true + return retryAfter(e.Header.Get("Retry-After")), true } // Account is a user account. It is associated with a private key. @@ -296,8 +296,8 @@ func (e *wireError) error(h http.Header) *Error { } } -// CertOption is an optional argument type for the TLSSNIxChallengeCert methods for -// customizing a temporary certificate for TLS-SNI challenges. +// CertOption is an optional argument type for the TLS ChallengeCert methods for +// customizing a temporary certificate for TLS-based challenges. type CertOption interface { privateCertOpt() } @@ -317,7 +317,7 @@ func (*certOptKey) privateCertOpt() {} // WithTemplate creates an option for specifying a certificate template. // See x509.CreateCertificate for template usage details. // -// In TLSSNIxChallengeCert methods, the template is also used as parent, +// In TLS ChallengeCert methods, the template is also used as parent, // resulting in a self-signed certificate. // The DNSNames field of t is always overwritten for tls-sni challenge certs. func WithTemplate(t *x509.Certificate) CertOption { diff --git a/vendor/golang.org/x/crypto/bn256/bn256.go b/vendor/golang.org/x/crypto/bn256/bn256.go index 479e7316a..ff27febd6 100644 --- a/vendor/golang.org/x/crypto/bn256/bn256.go +++ b/vendor/golang.org/x/crypto/bn256/bn256.go @@ -109,7 +109,6 @@ func (e *G1) Marshal() []byte { xBytes := new(big.Int).Mod(e.p.x, p).Bytes() yBytes := new(big.Int).Mod(e.p.y, p).Bytes() - ret := make([]byte, numBytes*2) copy(ret[1*numBytes-len(xBytes):], xBytes) copy(ret[2*numBytes-len(yBytes):], yBytes) @@ -224,7 +223,6 @@ func (n *G2) Marshal() []byte { yxBytes := new(big.Int).Mod(n.p.y.x, p).Bytes() yyBytes := new(big.Int).Mod(n.p.y.y, p).Bytes() - ret := make([]byte, numBytes*4) copy(ret[1*numBytes-len(xxBytes):], xxBytes) copy(ret[2*numBytes-len(xyBytes):], xyBytes) diff --git a/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_amd64.go b/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_amd64.go index 07d18a32f..ec13d1388 100644 --- a/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_amd64.go +++ b/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_amd64.go @@ -9,6 +9,7 @@ package chacha20poly1305 import ( "encoding/binary" + "golang.org/x/crypto/internal/subtle" "golang.org/x/sys/cpu" ) @@ -55,6 +56,9 @@ func (c *chacha20poly1305) seal(dst, nonce, plaintext, additionalData []byte) [] setupState(&state, &c.key, nonce) ret, out := sliceForAppend(dst, len(plaintext)+16) + if subtle.InexactOverlap(out, plaintext) { + panic("chacha20poly1305: invalid buffer overlap") + } chacha20Poly1305Seal(out[:], state[:], plaintext, additionalData) return ret } @@ -69,6 +73,9 @@ func (c *chacha20poly1305) open(dst, nonce, ciphertext, additionalData []byte) ( ciphertext = ciphertext[:len(ciphertext)-16] ret, out := sliceForAppend(dst, len(ciphertext)) + if subtle.InexactOverlap(out, ciphertext) { + panic("chacha20poly1305: invalid buffer overlap") + } if !chacha20Poly1305Open(out, state[:], ciphertext, additionalData) { for i := range out { out[i] = 0 diff --git a/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_generic.go b/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_generic.go index 8d28ce27f..c27971216 100644 --- a/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_generic.go +++ b/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_generic.go @@ -8,6 +8,7 @@ import ( "encoding/binary" "golang.org/x/crypto/internal/chacha20" + "golang.org/x/crypto/internal/subtle" "golang.org/x/crypto/poly1305" ) @@ -17,6 +18,9 @@ func roundTo16(n int) int { func (c *chacha20poly1305) sealGeneric(dst, nonce, plaintext, additionalData []byte) []byte { ret, out := sliceForAppend(dst, len(plaintext)+poly1305.TagSize) + if subtle.InexactOverlap(out, plaintext) { + panic("chacha20poly1305: invalid buffer overlap") + } var polyKey [32]byte s := chacha20.New(c.key, [3]uint32{ @@ -62,6 +66,9 @@ func (c *chacha20poly1305) openGeneric(dst, nonce, ciphertext, additionalData [] binary.LittleEndian.PutUint64(polyInput[len(polyInput)-8:], uint64(len(ciphertext))) ret, out := sliceForAppend(dst, len(ciphertext)) + if subtle.InexactOverlap(out, ciphertext) { + panic("chacha20poly1305: invalid buffer overlap") + } if !poly1305.Verify(&tag, polyInput, &polyKey) { for i := range out { out[i] = 0 diff --git a/vendor/golang.org/x/crypto/ed25519/ed25519.go b/vendor/golang.org/x/crypto/ed25519/ed25519.go index a57771a1e..d6f683ba3 100644 --- a/vendor/golang.org/x/crypto/ed25519/ed25519.go +++ b/vendor/golang.org/x/crypto/ed25519/ed25519.go @@ -6,7 +6,10 @@ // https://ed25519.cr.yp.to/. // // These functions are also compatible with the “Ed25519” function defined in -// RFC 8032. +// RFC 8032. However, unlike RFC 8032's formulation, this package's private key +// representation includes a public key suffix to make multiple signing +// operations with the same key more efficient. This package refers to the RFC +// 8032 private key as the “seed”. package ed25519 // This code is a port of the public domain, “ref10” implementation of ed25519 @@ -31,6 +34,8 @@ const ( PrivateKeySize = 64 // SignatureSize is the size, in bytes, of signatures generated and verified by this package. SignatureSize = 64 + // SeedSize is the size, in bytes, of private key seeds. These are the private key representations used by RFC 8032. + SeedSize = 32 ) // PublicKey is the type of Ed25519 public keys. @@ -46,6 +51,15 @@ func (priv PrivateKey) Public() crypto.PublicKey { return PublicKey(publicKey) } +// Seed returns the private key seed corresponding to priv. It is provided for +// interoperability with RFC 8032. RFC 8032's private keys correspond to seeds +// in this package. +func (priv PrivateKey) Seed() []byte { + seed := make([]byte, SeedSize) + copy(seed, priv[:32]) + return seed +} + // Sign signs the given message with priv. // Ed25519 performs two passes over messages to be signed and therefore cannot // handle pre-hashed messages. Thus opts.HashFunc() must return zero to @@ -61,19 +75,33 @@ func (priv PrivateKey) Sign(rand io.Reader, message []byte, opts crypto.SignerOp // GenerateKey generates a public/private key pair using entropy from rand. // If rand is nil, crypto/rand.Reader will be used. -func GenerateKey(rand io.Reader) (publicKey PublicKey, privateKey PrivateKey, err error) { +func GenerateKey(rand io.Reader) (PublicKey, PrivateKey, error) { if rand == nil { rand = cryptorand.Reader } - privateKey = make([]byte, PrivateKeySize) - publicKey = make([]byte, PublicKeySize) - _, err = io.ReadFull(rand, privateKey[:32]) - if err != nil { + seed := make([]byte, SeedSize) + if _, err := io.ReadFull(rand, seed); err != nil { return nil, nil, err } - digest := sha512.Sum512(privateKey[:32]) + privateKey := NewKeyFromSeed(seed) + publicKey := make([]byte, PublicKeySize) + copy(publicKey, privateKey[32:]) + + return publicKey, privateKey, nil +} + +// NewKeyFromSeed calculates a private key from a seed. It will panic if +// len(seed) is not SeedSize. This function is provided for interoperability +// with RFC 8032. RFC 8032's private keys correspond to seeds in this +// package. +func NewKeyFromSeed(seed []byte) PrivateKey { + if l := len(seed); l != SeedSize { + panic("ed25519: bad seed length: " + strconv.Itoa(l)) + } + + digest := sha512.Sum512(seed) digest[0] &= 248 digest[31] &= 127 digest[31] |= 64 @@ -85,10 +113,11 @@ func GenerateKey(rand io.Reader) (publicKey PublicKey, privateKey PrivateKey, er var publicKeyBytes [32]byte A.ToBytes(&publicKeyBytes) + privateKey := make([]byte, PrivateKeySize) + copy(privateKey, seed) copy(privateKey[32:], publicKeyBytes[:]) - copy(publicKey, publicKeyBytes[:]) - return publicKey, privateKey, nil + return privateKey } // Sign signs the message with privateKey and returns a signature. It will diff --git a/vendor/golang.org/x/crypto/ed25519/ed25519_test.go b/vendor/golang.org/x/crypto/ed25519/ed25519_test.go index 5f946e996..80946036d 100644 --- a/vendor/golang.org/x/crypto/ed25519/ed25519_test.go +++ b/vendor/golang.org/x/crypto/ed25519/ed25519_test.go @@ -139,6 +139,19 @@ func TestGolden(t *testing.T) { if !Verify(pubKey, msg, sig2) { t.Errorf("signature failed to verify on line %d", lineNo) } + + priv2 := NewKeyFromSeed(priv[:32]) + if !bytes.Equal(priv[:], priv2) { + t.Errorf("recreating key pair gave different private key on line %d: %x vs %x", lineNo, priv[:], priv2) + } + + if pubKey2 := priv2.Public().(PublicKey); !bytes.Equal(pubKey, pubKey2) { + t.Errorf("recreating key pair gave different public key on line %d: %x vs %x", lineNo, pubKey, pubKey2) + } + + if seed := priv2.Seed(); !bytes.Equal(priv[:32], seed) { + t.Errorf("recreating key pair gave different seed on line %d: %x vs %x", lineNo, priv[:32], seed) + } } if err := scanner.Err(); err != nil { diff --git a/vendor/golang.org/x/crypto/internal/chacha20/chacha_generic.go b/vendor/golang.org/x/crypto/internal/chacha20/chacha_generic.go index 7ed1cd9b1..523751f7f 100644 --- a/vendor/golang.org/x/crypto/internal/chacha20/chacha_generic.go +++ b/vendor/golang.org/x/crypto/internal/chacha20/chacha_generic.go @@ -9,6 +9,8 @@ package chacha20 import ( "crypto/cipher" "encoding/binary" + + "golang.org/x/crypto/internal/subtle" ) // assert that *Cipher implements cipher.Stream @@ -41,6 +43,13 @@ func New(key [8]uint32, nonce [3]uint32) *Cipher { // the src buffers was passed in a single run. That is, Cipher // maintains state and does not reset at each XORKeyStream call. func (s *Cipher) XORKeyStream(dst, src []byte) { + if len(dst) < len(src) { + panic("chacha20: output smaller than input") + } + if subtle.InexactOverlap(dst[:len(src)], src) { + panic("chacha20: invalid buffer overlap") + } + // xor src with buffered keystream first if s.len != 0 { buf := s.buf[len(s.buf)-s.len:] diff --git a/vendor/golang.org/x/crypto/internal/subtle/aliasing.go b/vendor/golang.org/x/crypto/internal/subtle/aliasing.go new file mode 100644 index 000000000..f38797bfa --- /dev/null +++ b/vendor/golang.org/x/crypto/internal/subtle/aliasing.go @@ -0,0 +1,32 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build !appengine + +// Package subtle implements functions that are often useful in cryptographic +// code but require careful thought to use correctly. +package subtle // import "golang.org/x/crypto/internal/subtle" + +import "unsafe" + +// AnyOverlap reports whether x and y share memory at any (not necessarily +// corresponding) index. The memory beyond the slice length is ignored. +func AnyOverlap(x, y []byte) bool { + return len(x) > 0 && len(y) > 0 && + uintptr(unsafe.Pointer(&x[0])) <= uintptr(unsafe.Pointer(&y[len(y)-1])) && + uintptr(unsafe.Pointer(&y[0])) <= uintptr(unsafe.Pointer(&x[len(x)-1])) +} + +// InexactOverlap reports whether x and y share memory at any non-corresponding +// index. The memory beyond the slice length is ignored. Note that x and y can +// have different lengths and still not have any inexact overlap. +// +// InexactOverlap can be used to implement the requirements of the crypto/cipher +// AEAD, Block, BlockMode and Stream interfaces. +func InexactOverlap(x, y []byte) bool { + if len(x) == 0 || len(y) == 0 || &x[0] == &y[0] { + return false + } + return AnyOverlap(x, y) +} diff --git a/vendor/golang.org/x/crypto/internal/subtle/aliasing_appengine.go b/vendor/golang.org/x/crypto/internal/subtle/aliasing_appengine.go new file mode 100644 index 000000000..0cc4a8a64 --- /dev/null +++ b/vendor/golang.org/x/crypto/internal/subtle/aliasing_appengine.go @@ -0,0 +1,35 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build appengine + +// Package subtle implements functions that are often useful in cryptographic +// code but require careful thought to use correctly. +package subtle // import "golang.org/x/crypto/internal/subtle" + +// This is the Google App Engine standard variant based on reflect +// because the unsafe package and cgo are disallowed. + +import "reflect" + +// AnyOverlap reports whether x and y share memory at any (not necessarily +// corresponding) index. The memory beyond the slice length is ignored. +func AnyOverlap(x, y []byte) bool { + return len(x) > 0 && len(y) > 0 && + reflect.ValueOf(&x[0]).Pointer() <= reflect.ValueOf(&y[len(y)-1]).Pointer() && + reflect.ValueOf(&y[0]).Pointer() <= reflect.ValueOf(&x[len(x)-1]).Pointer() +} + +// InexactOverlap reports whether x and y share memory at any non-corresponding +// index. The memory beyond the slice length is ignored. Note that x and y can +// have different lengths and still not have any inexact overlap. +// +// InexactOverlap can be used to implement the requirements of the crypto/cipher +// AEAD, Block, BlockMode and Stream interfaces. +func InexactOverlap(x, y []byte) bool { + if len(x) == 0 || len(y) == 0 || &x[0] == &y[0] { + return false + } + return AnyOverlap(x, y) +} diff --git a/vendor/golang.org/x/crypto/internal/subtle/aliasing_test.go b/vendor/golang.org/x/crypto/internal/subtle/aliasing_test.go new file mode 100644 index 000000000..a5b62ff74 --- /dev/null +++ b/vendor/golang.org/x/crypto/internal/subtle/aliasing_test.go @@ -0,0 +1,50 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package subtle_test + +import ( + "testing" + + "golang.org/x/crypto/internal/subtle" +) + +var a, b [100]byte + +var aliasingTests = []struct { + x, y []byte + anyOverlap, inexactOverlap bool +}{ + {a[:], b[:], false, false}, + {a[:], b[:0], false, false}, + {a[:], b[:50], false, false}, + {a[40:50], a[50:60], false, false}, + {a[40:50], a[60:70], false, false}, + {a[:51], a[50:], true, true}, + {a[:], a[:], true, false}, + {a[:50], a[:60], true, false}, + {a[:], nil, false, false}, + {nil, nil, false, false}, + {a[:], a[:0], false, false}, + {a[:10], a[:10:20], true, false}, + {a[:10], a[5:10:20], true, true}, +} + +func testAliasing(t *testing.T, i int, x, y []byte, anyOverlap, inexactOverlap bool) { + any := subtle.AnyOverlap(x, y) + if any != anyOverlap { + t.Errorf("%d: wrong AnyOverlap result, expected %v, got %v", i, anyOverlap, any) + } + inexact := subtle.InexactOverlap(x, y) + if inexact != inexactOverlap { + t.Errorf("%d: wrong InexactOverlap result, expected %v, got %v", i, inexactOverlap, any) + } +} + +func TestAliasing(t *testing.T) { + for i, tt := range aliasingTests { + testAliasing(t, i, tt.x, tt.y, tt.anyOverlap, tt.inexactOverlap) + testAliasing(t, i, tt.y, tt.x, tt.anyOverlap, tt.inexactOverlap) + } +} diff --git a/vendor/golang.org/x/crypto/nacl/secretbox/secretbox.go b/vendor/golang.org/x/crypto/nacl/secretbox/secretbox.go index 53ee83cfb..a98d1bd45 100644 --- a/vendor/golang.org/x/crypto/nacl/secretbox/secretbox.go +++ b/vendor/golang.org/x/crypto/nacl/secretbox/secretbox.go @@ -35,6 +35,7 @@ This package is interoperable with NaCl: https://nacl.cr.yp.to/secretbox.html. package secretbox // import "golang.org/x/crypto/nacl/secretbox" import ( + "golang.org/x/crypto/internal/subtle" "golang.org/x/crypto/poly1305" "golang.org/x/crypto/salsa20/salsa" ) @@ -87,6 +88,9 @@ func Seal(out, message []byte, nonce *[24]byte, key *[32]byte) []byte { copy(poly1305Key[:], firstBlock[:]) ret, out := sliceForAppend(out, len(message)+poly1305.TagSize) + if subtle.AnyOverlap(out, message) { + panic("nacl: invalid buffer overlap") + } // We XOR up to 32 bytes of message with the keystream generated from // the first block. @@ -118,7 +122,7 @@ func Seal(out, message []byte, nonce *[24]byte, key *[32]byte) []byte { // Open authenticates and decrypts a box produced by Seal and appends the // message to out, which must not overlap box. The output will be Overhead // bytes smaller than box. -func Open(out []byte, box []byte, nonce *[24]byte, key *[32]byte) ([]byte, bool) { +func Open(out, box []byte, nonce *[24]byte, key *[32]byte) ([]byte, bool) { if len(box) < Overhead { return nil, false } @@ -143,6 +147,9 @@ func Open(out []byte, box []byte, nonce *[24]byte, key *[32]byte) ([]byte, bool) } ret, out := sliceForAppend(out, len(box)-Overhead) + if subtle.AnyOverlap(out, box) { + panic("nacl: invalid buffer overlap") + } // We XOR up to 32 bytes of box with the keystream generated from // the first block. diff --git a/vendor/golang.org/x/crypto/nacl/sign/sign.go b/vendor/golang.org/x/crypto/nacl/sign/sign.go index a9ac0a771..d07627019 100644 --- a/vendor/golang.org/x/crypto/nacl/sign/sign.go +++ b/vendor/golang.org/x/crypto/nacl/sign/sign.go @@ -24,6 +24,7 @@ import ( "io" "golang.org/x/crypto/ed25519" + "golang.org/x/crypto/internal/subtle" ) // Overhead is the number of bytes of overhead when signing a message. @@ -47,6 +48,9 @@ func GenerateKey(rand io.Reader) (publicKey *[32]byte, privateKey *[64]byte, err func Sign(out, message []byte, privateKey *[64]byte) []byte { sig := ed25519.Sign(ed25519.PrivateKey((*privateKey)[:]), message) ret, out := sliceForAppend(out, Overhead+len(message)) + if subtle.AnyOverlap(out, message) { + panic("nacl: invalid buffer overlap") + } copy(out, sig) copy(out[Overhead:], message) return ret @@ -63,6 +67,9 @@ func Open(out, signedMessage []byte, publicKey *[32]byte) ([]byte, bool) { return nil, false } ret, out := sliceForAppend(out, len(signedMessage)-Overhead) + if subtle.AnyOverlap(out, signedMessage) { + panic("nacl: invalid buffer overlap") + } copy(out, signedMessage[Overhead:]) return ret, true } diff --git a/vendor/golang.org/x/crypto/openpgp/keys.go b/vendor/golang.org/x/crypto/openpgp/keys.go index fd582a89c..a79a8c13a 100644 --- a/vendor/golang.org/x/crypto/openpgp/keys.go +++ b/vendor/golang.org/x/crypto/openpgp/keys.go @@ -346,22 +346,25 @@ EachPacket: switch pkt := p.(type) { case *packet.UserId: + // Make a new Identity object, that we might wind up throwing away. + // We'll only add it if we get a valid self-signature over this + // userID. current = new(Identity) current.Name = pkt.Id current.UserId = pkt - e.Identities[pkt.Id] = current for { p, err = packets.Next() if err == io.EOF { - return nil, io.ErrUnexpectedEOF + break EachPacket } else if err != nil { return nil, err } sig, ok := p.(*packet.Signature) if !ok { - return nil, errors.StructuralError("user ID packet not followed by self-signature") + packets.Unread(p) + continue EachPacket } if (sig.SigType == packet.SigTypePositiveCert || sig.SigType == packet.SigTypeGenericCert) && sig.IssuerKeyId != nil && *sig.IssuerKeyId == e.PrimaryKey.KeyId { @@ -369,9 +372,10 @@ EachPacket: return nil, errors.StructuralError("user ID self-signature invalid: " + err.Error()) } current.SelfSignature = sig - break + e.Identities[pkt.Id] = current + } else { + current.Signatures = append(current.Signatures, sig) } - current.Signatures = append(current.Signatures, sig) } case *packet.Signature: if pkt.SigType == packet.SigTypeKeyRevocation { @@ -500,6 +504,10 @@ func NewEntity(name, comment, email string, config *packet.Config) (*Entity, err IssuerKeyId: &e.PrimaryKey.KeyId, }, } + err = e.Identities[uid.Id].SelfSignature.SignUserId(uid.Id, e.PrimaryKey, e.PrivateKey, config) + if err != nil { + return nil, err + } // If the user passes in a DefaultHash via packet.Config, // set the PreferredHash for the SelfSignature. @@ -529,13 +537,16 @@ func NewEntity(name, comment, email string, config *packet.Config) (*Entity, err } e.Subkeys[0].PublicKey.IsSubkey = true e.Subkeys[0].PrivateKey.IsSubkey = true - + err = e.Subkeys[0].Sig.SignKey(e.Subkeys[0].PublicKey, e.PrivateKey, config) + if err != nil { + return nil, err + } return e, nil } -// SerializePrivate serializes an Entity, including private key material, to -// the given Writer. For now, it must only be used on an Entity returned from -// NewEntity. +// SerializePrivate serializes an Entity, including private key material, but +// excluding signatures from other entities, to the given Writer. +// Identities and subkeys are re-signed in case they changed since NewEntry. // If config is nil, sensible defaults will be used. func (e *Entity) SerializePrivate(w io.Writer, config *packet.Config) (err error) { err = e.PrivateKey.Serialize(w) @@ -573,8 +584,8 @@ func (e *Entity) SerializePrivate(w io.Writer, config *packet.Config) (err error return nil } -// Serialize writes the public part of the given Entity to w. (No private -// key material will be output). +// Serialize writes the public part of the given Entity to w, including +// signatures from other entities. No private key material will be output. func (e *Entity) Serialize(w io.Writer) error { err := e.PrimaryKey.Serialize(w) if err != nil { diff --git a/vendor/golang.org/x/crypto/openpgp/keys_test.go b/vendor/golang.org/x/crypto/openpgp/keys_test.go index 3a1550638..d877589ae 100644 --- a/vendor/golang.org/x/crypto/openpgp/keys_test.go +++ b/vendor/golang.org/x/crypto/openpgp/keys_test.go @@ -29,16 +29,16 @@ func TestKeyExpiry(t *testing.T) { // // So this should select the newest, non-expired encryption key. key, _ := entity.encryptionKey(time1) - if id := key.PublicKey.KeyIdShortString(); id != "96A672F5" { - t.Errorf("Expected key 1ABB25A0 at time %s, but got key %s", time1.Format(timeFormat), id) + if id, expected := key.PublicKey.KeyIdShortString(), "96A672F5"; id != expected { + t.Errorf("Expected key %s at time %s, but got key %s", expected, time1.Format(timeFormat), id) } // Once the first encryption subkey has expired, the second should be // selected. time2, _ := time.Parse(timeFormat, "2013-07-09") key, _ = entity.encryptionKey(time2) - if id := key.PublicKey.KeyIdShortString(); id != "96A672F5" { - t.Errorf("Expected key 96A672F5 at time %s, but got key %s", time2.Format(timeFormat), id) + if id, expected := key.PublicKey.KeyIdShortString(), "96A672F5"; id != expected { + t.Errorf("Expected key %s at time %s, but got key %s", expected, time2.Format(timeFormat), id) } // Once all the keys have expired, nothing should be returned. @@ -105,6 +105,33 @@ func TestGoodCrossSignature(t *testing.T) { } } +func TestRevokedUserID(t *testing.T) { + // This key contains 2 UIDs, one of which is revoked: + // [ultimate] (1) Golang Gopher + // [ revoked] (2) Golang Gopher + keys, err := ReadArmoredKeyRing(bytes.NewBufferString(revokedUserIDKey)) + if err != nil { + t.Fatal(err) + } + + if len(keys) != 1 { + t.Fatal("Failed to read key with a revoked user id") + } + + var identities []*Identity + for _, identity := range keys[0].Identities { + identities = append(identities, identity) + } + + if numIdentities, numExpected := len(identities), 1; numIdentities != numExpected { + t.Errorf("obtained %d identities, expected %d", numIdentities, numExpected) + } + + if identityName, expectedName := identities[0].Name, "Golang Gopher "; identityName != expectedName { + t.Errorf("obtained identity %s expected %s", identityName, expectedName) + } +} + // TestExternallyRevokableKey attempts to load and parse a key with a third party revocation permission. func TestExternallyRevocableKey(t *testing.T) { kring, err := ReadKeyRing(readerFromHex(subkeyUsageHex)) @@ -370,6 +397,20 @@ func TestNewEntityWithoutPreferredSymmetric(t *testing.T) { } } +func TestNewEntityPublicSerialization(t *testing.T) { + entity, err := NewEntity("Golang Gopher", "Test Key", "no-reply@golang.com", nil) + if err != nil { + t.Fatal(err) + } + serializedEntity := bytes.NewBuffer(nil) + entity.Serialize(serializedEntity) + + _, err = ReadEntity(packet.NewReader(bytes.NewBuffer(serializedEntity.Bytes()))) + if err != nil { + t.Fatal(err) + } +} + const expiringKeyHex = "988d0451d1ec5d010400ba3385721f2dc3f4ab096b2ee867ab77213f0a27a8538441c35d2fa225b08798a1439a66a5150e6bdc3f40f5d28d588c712394c632b6299f77db8c0d48d37903fb72ebd794d61be6aa774688839e5fdecfe06b2684cc115d240c98c66cb1ef22ae84e3aa0c2b0c28665c1e7d4d044e7f270706193f5223c8d44e0d70b7b8da830011010001b40f4578706972792074657374206b657988be041301020028050251d1ec5d021b03050900278d00060b090807030206150802090a0b0416020301021e01021780000a091072589ad75e237d8c033503fd10506d72837834eb7f994117740723adc39227104b0d326a1161871c0b415d25b4aedef946ca77ea4c05af9c22b32cf98be86ab890111fced1ee3f75e87b7cc3c00dc63bbc85dfab91c0dc2ad9de2c4d13a34659333a85c6acc1a669c5e1d6cecb0cf1e56c10e72d855ae177ddc9e766f9b2dda57ccbb75f57156438bbdb4e42b88d0451d1ec5d0104009c64906559866c5cb61578f5846a94fcee142a489c9b41e67b12bb54cfe86eb9bc8566460f9a720cb00d6526fbccfd4f552071a8e3f7744b1882d01036d811ee5a3fb91a1c568055758f43ba5d2c6a9676b012f3a1a89e47bbf624f1ad571b208f3cc6224eb378f1645dd3d47584463f9eadeacfd1ce6f813064fbfdcc4b5a53001101000188a504180102000f021b0c050251d1f06b050900093e89000a091072589ad75e237d8c20e00400ab8310a41461425b37889c4da28129b5fae6084fafbc0a47dd1adc74a264c6e9c9cc125f40462ee1433072a58384daef88c961c390ed06426a81b464a53194c4e291ddd7e2e2ba3efced01537d713bd111f48437bde2363446200995e8e0d4e528dda377fd1e8f8ede9c8e2198b393bd86852ce7457a7e3daf74d510461a5b77b88d0451d1ece8010400b3a519f83ab0010307e83bca895170acce8964a044190a2b368892f7a244758d9fc193482648acb1fb9780d28cc22d171931f38bb40279389fc9bf2110876d4f3db4fcfb13f22f7083877fe56592b3b65251312c36f83ffcb6d313c6a17f197dd471f0712aad15a8537b435a92471ba2e5b0c72a6c72536c3b567c558d7b6051001101000188a504180102000f021b0c050251d1f07b050900279091000a091072589ad75e237d8ce69e03fe286026afacf7c97ee20673864d4459a2240b5655219950643c7dba0ac384b1d4359c67805b21d98211f7b09c2a0ccf6410c8c04d4ff4a51293725d8d6570d9d8bb0e10c07d22357caeb49626df99c180be02d77d1fe8ed25e7a54481237646083a9f89a11566cd20b9e995b1487c5f9e02aeb434f3a1897cd416dd0a87861838da3e9e" const subkeyUsageHex = "988d04533a52bc010400d26af43085558f65b9e7dbc90cb9238015259aed5e954637adcfa2181548b2d0b60c65f1f42ec5081cbf1bc0a8aa4900acfb77070837c58f26012fbce297d70afe96e759ad63531f0037538e70dbf8e384569b9720d99d8eb39d8d0a2947233ed242436cb6ac7dfe74123354b3d0119b5c235d3dd9c9d6c004f8ffaf67ad8583001101000188b7041f010200210502533b8552170c8001ce094aa433f7040bb2ddf0be3893cb843d0fe70c020700000a0910a42704b92866382aa98404009d63d916a27543da4221c60087c33f1c44bec9998c5438018ed370cca4962876c748e94b73eb39c58eb698063f3fd6346d58dd2a11c0247934c4a9d71f24754f7468f96fb24c3e791dd2392b62f626148ad724189498cbf993db2df7c0cdc2d677c35da0f16cb16c9ce7c33b4de65a4a91b1d21a130ae9cc26067718910ef8e2b417556d627261203c756d627261407379642e65642e61753e88b80413010200220502533a52bc021b03060b090807030206150802090a0b0416020301021e01021780000a0910a42704b92866382a47840400c0c2bd04f5fca586de408b395b3c280a278259c93eaaa8b79a53b97003f8ed502a8a00446dd9947fb462677e4fcac0dac2f0701847d15130aadb6cd9e0705ea0cf5f92f129136c7be21a718d46c8e641eb7f044f2adae573e11ae423a0a9ca51324f03a8a2f34b91fa40c3cc764bee4dccadedb54c768ba0469b683ea53f1c29b88d04533a52bc01040099c92a5d6f8b744224da27bc2369127c35269b58bec179de6bbc038f749344222f85a31933224f26b70243c4e4b2d242f0c4777eaef7b5502f9dad6d8bf3aaeb471210674b74de2d7078af497d55f5cdad97c7bedfbc1b41e8065a97c9c3d344b21fc81d27723af8e374bc595da26ea242dccb6ae497be26eea57e563ed517e90011010001889f0418010200090502533a52bc021b0c000a0910a42704b92866382afa1403ff70284c2de8a043ff51d8d29772602fa98009b7861c540535f874f2c230af8caf5638151a636b21f8255003997ccd29747fdd06777bb24f9593bd7d98a3e887689bf902f999915fcc94625ae487e5d13e6616f89090ebc4fdc7eb5cad8943e4056995bb61c6af37f8043016876a958ec7ebf39c43d20d53b7f546cfa83e8d2604b88d04533b8283010400c0b529316dbdf58b4c54461e7e669dc11c09eb7f73819f178ccd4177b9182b91d138605fcf1e463262fabefa73f94a52b5e15d1904635541c7ea540f07050ce0fb51b73e6f88644cec86e91107c957a114f69554548a85295d2b70bd0b203992f76eb5d493d86d9eabcaa7ef3fc7db7e458438db3fcdb0ca1cc97c638439a9170011010001889f0418010200090502533b8283021b0c000a0910a42704b92866382adc6d0400cfff6258485a21675adb7a811c3e19ebca18851533f75a7ba317950b9997fda8d1a4c8c76505c08c04b6c2cc31dc704d33da36a21273f2b388a1a706f7c3378b66d887197a525936ed9a69acb57fe7f718133da85ec742001c5d1864e9c6c8ea1b94f1c3759cebfd93b18606066c063a63be86085b7e37bdbc65f9a915bf084bb901a204533b85cd110400aed3d2c52af2b38b5b67904b0ef73d6dd7aef86adb770e2b153cd22489654dcc91730892087bb9856ae2d9f7ed1eb48f214243fe86bfe87b349ebd7c30e630e49c07b21fdabf78b7a95c8b7f969e97e3d33f2e074c63552ba64a2ded7badc05ce0ea2be6d53485f6900c7860c7aa76560376ce963d7271b9b54638a4028b573f00a0d8854bfcdb04986141568046202192263b9b67350400aaa1049dbc7943141ef590a70dcb028d730371d92ea4863de715f7f0f16d168bd3dc266c2450457d46dcbbf0b071547e5fbee7700a820c3750b236335d8d5848adb3c0da010e998908dfd93d961480084f3aea20b247034f8988eccb5546efaa35a92d0451df3aaf1aee5aa36a4c4d462c760ecd9cebcabfbe1412b1f21450f203fd126687cd486496e971a87fd9e1a8a765fe654baa219a6871ab97768596ab05c26c1aeea8f1a2c72395a58dbc12ef9640d2b95784e974a4d2d5a9b17c25fedacfe551bda52602de8f6d2e48443f5dd1a2a2a8e6a5e70ecdb88cd6e766ad9745c7ee91d78cc55c3d06536b49c3fee6c3d0b6ff0fb2bf13a314f57c953b8f4d93bf88e70418010200090502533b85cd021b0200520910a42704b92866382a47200419110200060502533b85cd000a091042ce2c64bc0ba99214b2009e26b26852c8b13b10c35768e40e78fbbb48bd084100a0c79d9ea0844fa5853dd3c85ff3ecae6f2c9dd6c557aa04008bbbc964cd65b9b8299d4ebf31f41cc7264b8cf33a00e82c5af022331fac79efc9563a822497ba012953cefe2629f1242fcdcb911dbb2315985bab060bfd58261ace3c654bdbbe2e8ed27a46e836490145c86dc7bae15c011f7e1ffc33730109b9338cd9f483e7cef3d2f396aab5bd80efb6646d7e778270ee99d934d187dd98" const revokedKeyHex = "988d045331ce82010400c4fdf7b40a5477f206e6ee278eaef888ca73bf9128a9eef9f2f1ddb8b7b71a4c07cfa241f028a04edb405e4d916c61d6beabc333813dc7b484d2b3c52ee233c6a79b1eea4e9cc51596ba9cd5ac5aeb9df62d86ea051055b79d03f8a4fa9f38386f5bd17529138f3325d46801514ea9047977e0829ed728e68636802796801be10011010001889f04200102000905025331d0e3021d03000a0910a401d9f09a34f7c042aa040086631196405b7e6af71026b88e98012eab44aa9849f6ef3fa930c7c9f23deaedba9db1538830f8652fb7648ec3fcade8dbcbf9eaf428e83c6cbcc272201bfe2fbb90d41963397a7c0637a1a9d9448ce695d9790db2dc95433ad7be19eb3de72dacf1d6db82c3644c13eae2a3d072b99bb341debba012c5ce4006a7d34a1f4b94b444526567205265766f6b657220283c52656727732022424d204261726973746122204b657920262530305c303e5c29203c72656740626d626172697374612e636f2e61753e88b704130102002205025331ce82021b03060b090807030206150802090a0b0416020301021e01021780000a0910a401d9f09a34f7c0019c03f75edfbeb6a73e7225ad3cc52724e2872e04260d7daf0d693c170d8c4b243b8767bc7785763533febc62ec2600c30603c433c095453ede59ff2fcabeb84ce32e0ed9d5cf15ffcbc816202b64370d4d77c1e9077d74e94a16fb4fa2e5bec23a56d7a73cf275f91691ae1801a976fcde09e981a2f6327ac27ea1fecf3185df0d56889c04100102000605025331cfb5000a0910fe9645554e8266b64b4303fc084075396674fb6f778d302ac07cef6bc0b5d07b66b2004c44aef711cbac79617ef06d836b4957522d8772dd94bf41a2f4ac8b1ee6d70c57503f837445a74765a076d07b829b8111fc2a918423ddb817ead7ca2a613ef0bfb9c6b3562aec6c3cf3c75ef3031d81d95f6563e4cdcc9960bcb386c5d757b104fcca5fe11fc709df884604101102000605025331cfe7000a09107b15a67f0b3ddc0317f6009e360beea58f29c1d963a22b962b80788c3fa6c84e009d148cfde6b351469b8eae91187eff07ad9d08fcaab88d045331ce820104009f25e20a42b904f3fa555530fe5c46737cf7bd076c35a2a0d22b11f7e0b61a69320b768f4a80fe13980ce380d1cfc4a0cd8fbe2d2e2ef85416668b77208baa65bf973fe8e500e78cc310d7c8705cdb34328bf80e24f0385fce5845c33bc7943cf6b11b02348a23da0bf6428e57c05135f2dc6bd7c1ce325d666d5a5fd2fd5e410011010001889f04180102000905025331ce82021b0c000a0910a401d9f09a34f7c0418003fe34feafcbeaef348a800a0d908a7a6809cc7304017d820f70f0474d5e23cb17e38b67dc6dca282c6ca00961f4ec9edf2738d0f087b1d81e4871ef08e1798010863afb4eac4c44a376cb343be929c5be66a78cfd4456ae9ec6a99d97f4e1c3ff3583351db2147a65c0acef5c003fb544ab3a2e2dc4d43646f58b811a6c3a369d1f" @@ -467,3 +508,42 @@ SqLHvbKh2dL/RXymC3+rjPvQf5cup9bPxNMa6WagdYBNAfzWGtkVISeaQW+cTEp/ MtgVijRGXR/lGLGETPg2X3Afwn9N9bLMBkBprKgbBqU7lpaoPupxT61bL70= =vtbN -----END PGP PUBLIC KEY BLOCK-----` + +const revokedUserIDKey = `-----BEGIN PGP PUBLIC KEY BLOCK----- + +mQENBFsgO5EBCADhREPmcjsPkXe1z7ctvyWL0S7oa9JaoGZ9oPDHFDlQxd0qlX2e +DZJZDg0qYvVixmaULIulApq1puEsaJCn3lHUbHlb4PYKwLEywYXM28JN91KtLsz/ +uaEX2KC5WqeP40utmzkNLq+oRX/xnRMgwbO7yUNVG2UlEa6eI+xOXO3YtLdmJMBW +ClQ066ZnOIzEo1JxnIwha1CDBMWLLfOLrg6l8InUqaXbtEBbnaIYO6fXVXELUjkx +nmk7t/QOk0tXCy8muH9UDqJkwDUESY2l79XwBAcx9riX8vY7vwC34pm22fAUVLCJ +x1SJx0J8bkeNp38jKM2Zd9SUQqSbfBopQ4pPABEBAAG0I0dvbGFuZyBHb3BoZXIg +PG5vLXJlcGx5QGdvbGFuZy5jb20+iQFUBBMBCgA+FiEE5Ik5JLcNx6l6rZfw1oFy +9I6cUoMFAlsgO5ECGwMFCQPCZwAFCwkIBwMFFQoJCAsFFgIDAQACHgECF4AACgkQ +1oFy9I6cUoMIkwf8DNPeD23i4jRwd/pylbvxwZintZl1fSwTJW1xcOa1emXaEtX2 +depuqhP04fjlRQGfsYAQh7X9jOJxAHjTmhqFBi5sD7QvKU00cPFYbJ/JTx0B41bl +aXnSbGhRPh63QtEZL7ACAs+shwvvojJqysx7kyVRu0EW2wqjXdHwR/SJO6nhNBa2 +DXzSiOU/SUA42mmG+5kjF8Aabq9wPwT9wjraHShEweNerNMmOqJExBOy3yFeyDpa +XwEZFzBfOKoxFNkIaVf5GSdIUGhFECkGvBMB935khftmgR8APxdU4BE7XrXexFJU +8RCuPXonm4WQOwTWR0vQg64pb2WKAzZ8HhwTGbQiR29sYW5nIEdvcGhlciA8cmV2 +b2tlZEBnb2xhbmcuY29tPokBNgQwAQoAIBYhBOSJOSS3Dcepeq2X8NaBcvSOnFKD +BQJbIDv3Ah0AAAoJENaBcvSOnFKDfWMIAKhI/Tvu3h8fSUxp/gSAcduT6bC1JttG +0lYQ5ilKB/58lBUA5CO3ZrKDKlzW3M8VEcvohVaqeTMKeoQd5rCZq8KxHn/KvN6N +s85REfXfniCKfAbnGgVXX3kDmZ1g63pkxrFu0fDZjVDXC6vy+I0sGyI/Inro0Pzb +tvn0QCsxjapKK15BtmSrpgHgzVqVg0cUp8vqZeKFxarYbYB2idtGRci4b9tObOK0 +BSTVFy26+I/mrFGaPrySYiy2Kz5NMEcRhjmTxJ8jSwEr2O2sUR0yjbgUAXbTxDVE +/jg5fQZ1ACvBRQnB7LvMHcInbzjyeTM3FazkkSYQD6b97+dkWwb1iWG5AQ0EWyA7 +kQEIALkg04REDZo1JgdYV4x8HJKFS4xAYWbIva1ZPqvDNmZRUbQZR2+gpJGEwn7z +VofGvnOYiGW56AS5j31SFf5kro1+1bZQ5iOONBng08OOo58/l1hRseIIVGB5TGSa +PCdChKKHreJI6hS3mShxH6hdfFtiZuB45rwoaArMMsYcjaezLwKeLc396cpUwwcZ +snLUNd1Xu5EWEF2OdFkZ2a1qYdxBvAYdQf4+1Nr+NRIx1u1NS9c8jp3PuMOkrQEi +bNtc1v6v0Jy52mKLG4y7mC/erIkvkQBYJdxPaP7LZVaPYc3/xskcyijrJ/5ufoD8 +K71/ShtsZUXSQn9jlRaYR0EbojMAEQEAAYkBPAQYAQoAJhYhBOSJOSS3Dcepeq2X +8NaBcvSOnFKDBQJbIDuRAhsMBQkDwmcAAAoJENaBcvSOnFKDkFMIAIt64bVZ8x7+ +TitH1bR4pgcNkaKmgKoZz6FXu80+SnbuEt2NnDyf1cLOSimSTILpwLIuv9Uft5Pb +OraQbYt3xi9yrqdKqGLv80bxqK0NuryNkvh9yyx5WoG1iKqMj9/FjGghuPrRaT4l +QinNAghGVkEy1+aXGFrG2DsOC1FFI51CC2WVTzZ5RwR2GpiNRfESsU1rZAUqf/2V +yJl9bD5R4SUNy8oQmhOxi+gbhD4Ao34e4W0ilibslI/uawvCiOwlu5NGd8zv5n+U +heiQvzkApQup5c+BhH5zFDFdKJ2CBByxw9+7QjMFI/wgLixKuE0Ob2kAokXf7RlB +7qTZOahrETw= +=IKnw +-----END PGP PUBLIC KEY BLOCK-----` diff --git a/vendor/golang.org/x/crypto/salsa20/salsa20.go b/vendor/golang.org/x/crypto/salsa20/salsa20.go index 3ca67485b..6f9bb106c 100644 --- a/vendor/golang.org/x/crypto/salsa20/salsa20.go +++ b/vendor/golang.org/x/crypto/salsa20/salsa20.go @@ -24,6 +24,7 @@ package salsa20 // import "golang.org/x/crypto/salsa20" // TODO(agl): implement XORKeyStream12 and XORKeyStream8 - the reduced round variants of Salsa20. import ( + "golang.org/x/crypto/internal/subtle" "golang.org/x/crypto/salsa20/salsa" ) @@ -34,6 +35,9 @@ func XORKeyStream(out, in []byte, nonce []byte, key *[32]byte) { if len(out) < len(in) { panic("salsa20: output smaller than input") } + if subtle.InexactOverlap(out[:len(in)], in) { + panic("salsa20: invalid buffer overlap") + } var subNonce [16]byte diff --git a/vendor/golang.org/x/crypto/ssh/terminal/terminal_test.go b/vendor/golang.org/x/crypto/ssh/terminal/terminal_test.go index a27cdd667..d9b77c1c5 100644 --- a/vendor/golang.org/x/crypto/ssh/terminal/terminal_test.go +++ b/vendor/golang.org/x/crypto/ssh/terminal/terminal_test.go @@ -10,6 +10,7 @@ import ( "bytes" "io" "os" + "runtime" "testing" ) @@ -326,6 +327,11 @@ func TestMakeRawState(t *testing.T) { if err != nil { t.Fatalf("failed to get terminal state from GetState: %s", err) } + + if runtime.GOOS == "darwin" && (runtime.GOARCH == "arm" || runtime.GOARCH == "arm64") { + t.Skip("MakeRaw not allowed on iOS; skipping test") + } + defer Restore(fd, st) raw, err := MakeRaw(fd) if err != nil { diff --git a/vendor/golang.org/x/crypto/xts/xts.go b/vendor/golang.org/x/crypto/xts/xts.go index 92cbce99b..9654e1fc0 100644 --- a/vendor/golang.org/x/crypto/xts/xts.go +++ b/vendor/golang.org/x/crypto/xts/xts.go @@ -25,6 +25,8 @@ import ( "crypto/cipher" "encoding/binary" "errors" + + "golang.org/x/crypto/internal/subtle" ) // Cipher contains an expanded key structure. It doesn't contain mutable state @@ -64,6 +66,9 @@ func (c *Cipher) Encrypt(ciphertext, plaintext []byte, sectorNum uint64) { if len(plaintext)%blockSize != 0 { panic("xts: plaintext is not a multiple of the block size") } + if subtle.InexactOverlap(ciphertext[:len(plaintext)], plaintext) { + panic("xts: invalid buffer overlap") + } var tweak [blockSize]byte binary.LittleEndian.PutUint64(tweak[:8], sectorNum) @@ -95,6 +100,9 @@ func (c *Cipher) Decrypt(plaintext, ciphertext []byte, sectorNum uint64) { if len(ciphertext)%blockSize != 0 { panic("xts: ciphertext is not a multiple of the block size") } + if subtle.InexactOverlap(plaintext[:len(ciphertext)], ciphertext) { + panic("xts: invalid buffer overlap") + } var tweak [blockSize]byte binary.LittleEndian.PutUint64(tweak[:8], sectorNum) diff --git a/vendor/golang.org/x/net/http/httpguts/guts.go b/vendor/golang.org/x/net/http/httpguts/guts.go index 8255fd49b..e6cd0ced3 100644 --- a/vendor/golang.org/x/net/http/httpguts/guts.go +++ b/vendor/golang.org/x/net/http/httpguts/guts.go @@ -14,21 +14,6 @@ import ( "strings" ) -// SniffedContentType reports whether ct is a Content-Type that is known -// to cause client-side content sniffing. -// -// This provides just a partial implementation of mime.ParseMediaType -// with the assumption that the Content-Type is not attacker controlled. -func SniffedContentType(ct string) bool { - if i := strings.Index(ct, ";"); i != -1 { - ct = ct[:i] - } - ct = strings.ToLower(strings.TrimSpace(ct)) - return ct == "text/plain" || ct == "application/octet-stream" || - ct == "application/unknown" || ct == "unknown/unknown" || ct == "*/*" || - !strings.Contains(ct, "/") -} - // ValidTrailerHeader reports whether name is a valid header field name to appear // in trailers. // See RFC 7230, Section 4.1.2 diff --git a/vendor/golang.org/x/net/http2/flow.go b/vendor/golang.org/x/net/http2/flow.go index 957de2542..cea601fcd 100644 --- a/vendor/golang.org/x/net/http2/flow.go +++ b/vendor/golang.org/x/net/http2/flow.go @@ -41,10 +41,10 @@ func (f *flow) take(n int32) { // add adds n bytes (positive or negative) to the flow control window. // It returns false if the sum would exceed 2^31-1. func (f *flow) add(n int32) bool { - remain := (1<<31 - 1) - f.n - if n > remain { - return false + sum := f.n + n + if (sum > n) == (f.n > 0) { + f.n = sum + return true } - f.n += n - return true + return false } diff --git a/vendor/golang.org/x/net/http2/flow_test.go b/vendor/golang.org/x/net/http2/flow_test.go index 859adf5d1..7ae82c781 100644 --- a/vendor/golang.org/x/net/http2/flow_test.go +++ b/vendor/golang.org/x/net/http2/flow_test.go @@ -49,5 +49,39 @@ func TestFlowAdd(t *testing.T) { if f.add(1) { t.Fatal("adding 1 to max shouldn't be allowed") } +} + +func TestFlowAddOverflow(t *testing.T) { + var f flow + if !f.add(0) { + t.Fatal("failed to add 0") + } + if !f.add(-1) { + t.Fatal("failed to add -1") + } + if !f.add(0) { + t.Fatal("failed to add 0") + } + if !f.add(1) { + t.Fatal("failed to add 1") + } + if !f.add(1) { + t.Fatal("failed to add 1") + } + if !f.add(0) { + t.Fatal("failed to add 0") + } + if !f.add(-3) { + t.Fatal("failed to add -3") + } + if got, want := f.available(), int32(-2); got != want { + t.Fatalf("size = %d; want %d", got, want) + } + if !f.add(1<<31 - 1) { + t.Fatal("failed to add 2^31-1") + } + if got, want := f.available(), int32(1+-3+(1<<31-1)); got != want { + t.Fatalf("size = %d; want %d", got, want) + } } diff --git a/vendor/golang.org/x/net/http2/http2_test.go b/vendor/golang.org/x/net/http2/http2_test.go index 524877647..667db4887 100644 --- a/vendor/golang.org/x/net/http2/http2_test.go +++ b/vendor/golang.org/x/net/http2/http2_test.go @@ -14,6 +14,7 @@ import ( "strconv" "strings" "testing" + "time" "golang.org/x/net/http2/hpack" ) @@ -197,3 +198,30 @@ func TestSorterPoolAllocs(t *testing.T) { t.Logf("Keys allocs = %v; want <1", allocs) } } + +// waitCondition reports whether fn eventually returned true, +// checking immediately and then every checkEvery amount, +// until waitFor has elapsed, at which point it returns false. +func waitCondition(waitFor, checkEvery time.Duration, fn func() bool) bool { + deadline := time.Now().Add(waitFor) + for time.Now().Before(deadline) { + if fn() { + return true + } + time.Sleep(checkEvery) + } + return false +} + +// waitErrCondition is like waitCondition but with errors instead of bools. +func waitErrCondition(waitFor, checkEvery time.Duration, fn func() error) error { + deadline := time.Now().Add(waitFor) + var err error + for time.Now().Before(deadline) { + if err = fn(); err == nil { + return nil + } + time.Sleep(checkEvery) + } + return err +} diff --git a/vendor/golang.org/x/net/http2/server.go b/vendor/golang.org/x/net/http2/server.go index abf94e8d3..793899169 100644 --- a/vendor/golang.org/x/net/http2/server.go +++ b/vendor/golang.org/x/net/http2/server.go @@ -1608,7 +1608,10 @@ func (sc *serverConn) processData(f *DataFrame) error { // Sender sending more than they'd declared? if st.declBodyBytes != -1 && st.bodyBytes+int64(len(data)) > st.declBodyBytes { st.body.CloseWithError(fmt.Errorf("sender tried to send more than declared Content-Length of %d bytes", st.declBodyBytes)) - return streamError(id, ErrCodeStreamClosed) + // RFC 7540, sec 8.1.2.6: A request or response is also malformed if the + // value of a content-length header field does not equal the sum of the + // DATA frame payload lengths that form the body. + return streamError(id, ErrCodeProtocol) } if f.Length > 0 { // Check whether the client has flow control quota. @@ -2309,7 +2312,6 @@ func (rws *responseWriterState) writeChunk(p []byte) (n int, err error) { isHeadResp := rws.req.Method == "HEAD" if !rws.sentHeader { rws.sentHeader = true - var ctype, clen string if clen = rws.snapHeader.Get("Content-Length"); clen != "" { rws.snapHeader.Del("Content-Length") @@ -2323,7 +2325,6 @@ func (rws *responseWriterState) writeChunk(p []byte) (n int, err error) { if clen == "" && rws.handlerDone && bodyAllowedForStatus(rws.status) && (len(p) > 0 || !isHeadResp) { clen = strconv.Itoa(len(p)) } - _, hasContentType := rws.snapHeader["Content-Type"] if !hasContentType && bodyAllowedForStatus(rws.status) && len(p) > 0 { if cto := rws.snapHeader.Get("X-Content-Type-Options"); strings.EqualFold("nosniff", cto) { @@ -2336,20 +2337,6 @@ func (rws *responseWriterState) writeChunk(p []byte) (n int, err error) { ctype = http.DetectContentType(p) } } - - var noSniff bool - if bodyAllowedForStatus(rws.status) && (rws.sentContentLen > 0 || len(p) > 0) { - // If the content type triggers client-side sniffing on old browsers, - // attach a X-Content-Type-Options header if not present (or explicitly nil). - if _, ok := rws.snapHeader["X-Content-Type-Options"]; !ok { - if hasContentType { - noSniff = httpguts.SniffedContentType(rws.snapHeader.Get("Content-Type")) - } else if ctype != "" { - noSniff = httpguts.SniffedContentType(ctype) - } - } - } - var date string if _, ok := rws.snapHeader["Date"]; !ok { // TODO(bradfitz): be faster here, like net/http? measure. @@ -2360,6 +2347,19 @@ func (rws *responseWriterState) writeChunk(p []byte) (n int, err error) { foreachHeaderElement(v, rws.declareTrailer) } + // "Connection" headers aren't allowed in HTTP/2 (RFC 7540, 8.1.2.2), + // but respect "Connection" == "close" to mean sending a GOAWAY and tearing + // down the TCP connection when idle, like we do for HTTP/1. + // TODO: remove more Connection-specific header fields here, in addition + // to "Connection". + if _, ok := rws.snapHeader["Connection"]; ok { + v := rws.snapHeader.Get("Connection") + delete(rws.snapHeader, "Connection") + if v == "close" { + rws.conn.startGracefulShutdown() + } + } + endStream := (rws.handlerDone && !rws.hasTrailers() && len(p) == 0) || isHeadResp err = rws.conn.writeHeaders(rws.stream, &writeResHeaders{ streamID: rws.stream.id, @@ -2368,7 +2368,6 @@ func (rws *responseWriterState) writeChunk(p []byte) (n int, err error) { endStream: endStream, contentType: ctype, contentLength: clen, - noSniff: noSniff, date: date, }) if err != nil { diff --git a/vendor/golang.org/x/net/http2/server_test.go b/vendor/golang.org/x/net/http2/server_test.go index c48d8d387..02eb0dcce 100644 --- a/vendor/golang.org/x/net/http2/server_test.go +++ b/vendor/golang.org/x/net/http2/server_test.go @@ -1810,7 +1810,6 @@ func TestServer_Response_TransferEncoding_chunked(t *testing.T) { {":status", "200"}, {"content-type", "text/plain; charset=utf-8"}, {"content-length", strconv.Itoa(len(msg))}, - {"x-content-type-options", "nosniff"}, } if !reflect.DeepEqual(goth, wanth) { t.Errorf("Got headers %v; want %v", goth, wanth) @@ -1999,7 +1998,6 @@ func TestServer_Response_LargeWrite(t *testing.T) { wanth := [][2]string{ {":status", "200"}, {"content-type", "text/plain; charset=utf-8"}, // sniffed - {"x-content-type-options", "nosniff"}, // and no content-length } if !reflect.DeepEqual(goth, wanth) { @@ -2214,7 +2212,6 @@ func TestServer_Response_Automatic100Continue(t *testing.T) { {":status", "200"}, {"content-type", "text/plain; charset=utf-8"}, {"content-length", strconv.Itoa(len(reply))}, - {"x-content-type-options", "nosniff"}, } if !reflect.DeepEqual(goth, wanth) { t.Errorf("Got headers %v; want %v", goth, wanth) @@ -2938,7 +2935,6 @@ func testServerWritesTrailers(t *testing.T, withFlush bool) { {"trailer", "Transfer-Encoding, Content-Length, Trailer"}, {"content-type", "text/plain; charset=utf-8"}, {"content-length", "5"}, - {"x-content-type-options", "nosniff"}, } if !reflect.DeepEqual(goth, wanth) { t.Errorf("Header mismatch.\n got: %v\nwant: %v", goth, wanth) @@ -3330,7 +3326,6 @@ func TestServerNoDuplicateContentType(t *testing.T) { {":status", "200"}, {"content-type", ""}, {"content-length", "41"}, - {"x-content-type-options", "nosniff"}, } if !reflect.DeepEqual(headers, want) { t.Errorf("Headers mismatch.\n got: %q\nwant: %q\n", headers, want) @@ -3764,3 +3759,85 @@ func TestIssue20704Race(t *testing.T) { resp.Body.Close() } } + +func TestServer_Rejects_TooSmall(t *testing.T) { + testServerResponse(t, func(w http.ResponseWriter, r *http.Request) error { + ioutil.ReadAll(r.Body) + return nil + }, func(st *serverTester) { + st.writeHeaders(HeadersFrameParam{ + StreamID: 1, // clients send odd numbers + BlockFragment: st.encodeHeader( + ":method", "POST", + "content-length", "4", + ), + EndStream: false, // to say DATA frames are coming + EndHeaders: true, + }) + st.writeData(1, true, []byte("12345")) + + st.wantRSTStream(1, ErrCodeProtocol) + }) +} + +// Tests that a handler setting "Connection: close" results in a GOAWAY being sent, +// and the connection still completing. +func TestServerHandlerConnectionClose(t *testing.T) { + unblockHandler := make(chan bool, 1) + defer close(unblockHandler) // backup; in case of errors + testServerResponse(t, func(w http.ResponseWriter, r *http.Request) error { + w.Header().Set("Connection", "close") + w.Header().Set("Foo", "bar") + w.(http.Flusher).Flush() + <-unblockHandler + return nil + }, func(st *serverTester) { + st.writeHeaders(HeadersFrameParam{ + StreamID: 1, + BlockFragment: st.encodeHeader(), + EndStream: true, + EndHeaders: true, + }) + var sawGoAway bool + var sawRes bool + for { + f, err := st.readFrame() + if err == io.EOF { + break + } + if err != nil { + t.Fatal(err) + } + switch f := f.(type) { + case *GoAwayFrame: + sawGoAway = true + unblockHandler <- true + if f.LastStreamID != 1 || f.ErrCode != ErrCodeNo { + t.Errorf("unexpected GOAWAY frame: %v", summarizeFrame(f)) + } + case *HeadersFrame: + goth := st.decodeHeader(f.HeaderBlockFragment()) + if !sawGoAway { + t.Fatalf("unexpected Headers frame before GOAWAY: %s, %v", summarizeFrame(f), goth) + } + wanth := [][2]string{ + {":status", "200"}, + {"foo", "bar"}, + } + if !reflect.DeepEqual(goth, wanth) { + t.Errorf("got headers %v; want %v", goth, wanth) + } + sawRes = true + case *DataFrame: + if f.StreamID != 1 || !f.StreamEnded() || len(f.Data()) != 0 { + t.Errorf("unexpected DATA frame: %v", summarizeFrame(f)) + } + default: + t.Logf("unexpected frame: %v", summarizeFrame(f)) + } + } + if !sawRes { + t.Errorf("didn't see response") + } + }) +} diff --git a/vendor/golang.org/x/net/http2/transport_test.go b/vendor/golang.org/x/net/http2/transport_test.go index ed58ce870..17c09cab7 100644 --- a/vendor/golang.org/x/net/http2/transport_test.go +++ b/vendor/golang.org/x/net/http2/transport_test.go @@ -145,10 +145,9 @@ func TestTransport(t *testing.T) { t.Errorf("Status = %q; want %q", g, w) } wantHeader := http.Header{ - "Content-Length": []string{"3"}, - "X-Content-Type-Options": []string{"nosniff"}, - "Content-Type": []string{"text/plain; charset=utf-8"}, - "Date": []string{"XXX"}, // see cleanDate + "Content-Length": []string{"3"}, + "Content-Type": []string{"text/plain; charset=utf-8"}, + "Date": []string{"XXX"}, // see cleanDate } cleanDate(res) if !reflect.DeepEqual(res.Header, wantHeader) { @@ -2395,11 +2394,12 @@ func TestTransportHandlerBodyClose(t *testing.T) { } tr.CloseIdleConnections() - gd := runtime.NumGoroutine() - g0 - if gd > numReq/2 { + if !waitCondition(5*time.Second, 100*time.Millisecond, func() bool { + gd := runtime.NumGoroutine() - g0 + return gd < numReq/2 + }) { t.Errorf("appeared to leak goroutines") } - } // https://golang.org/issue/15930 diff --git a/vendor/golang.org/x/net/http2/write.go b/vendor/golang.org/x/net/http2/write.go index a5120412e..8a9711f6e 100644 --- a/vendor/golang.org/x/net/http2/write.go +++ b/vendor/golang.org/x/net/http2/write.go @@ -186,7 +186,6 @@ type writeResHeaders struct { date string contentType string contentLength string - noSniff bool } func encKV(enc *hpack.Encoder, k, v string) { @@ -223,9 +222,6 @@ func (w *writeResHeaders) writeFrame(ctx writeContext) error { if w.contentLength != "" { encKV(enc, "content-length", w.contentLength) } - if w.noSniff { - encKV(enc, "x-content-type-options", "nosniff") - } if w.date != "" { encKV(enc, "date", w.date) } diff --git a/vendor/golang.org/x/net/icmp/listen_stub.go b/vendor/golang.org/x/net/icmp/listen_stub.go index 668728d17..6b77ab487 100644 --- a/vendor/golang.org/x/net/icmp/listen_stub.go +++ b/vendor/golang.org/x/net/icmp/listen_stub.go @@ -2,7 +2,7 @@ // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. -// +build nacl plan9 +// +build js nacl plan9 package icmp diff --git a/vendor/golang.org/x/net/icmp/message.go b/vendor/golang.org/x/net/icmp/message.go index 46fe95ab0..c9329a2f6 100644 --- a/vendor/golang.org/x/net/icmp/message.go +++ b/vendor/golang.org/x/net/icmp/message.go @@ -25,7 +25,7 @@ import ( "golang.org/x/net/ipv6" ) -// BUG(mikio): This package is not implemented on NaCl and Plan 9. +// BUG(mikio): This package is not implemented on JS, NaCl and Plan 9. var ( errMessageTooShort = errors.New("message too short") diff --git a/vendor/golang.org/x/net/icmp/mpls.go b/vendor/golang.org/x/net/icmp/mpls.go index c31491748..f9f4841bc 100644 --- a/vendor/golang.org/x/net/icmp/mpls.go +++ b/vendor/golang.org/x/net/icmp/mpls.go @@ -6,7 +6,7 @@ package icmp import "encoding/binary" -// A MPLSLabel represents a MPLS label stack entry. +// MPLSLabel represents an MPLS label stack entry. type MPLSLabel struct { Label int // label value TC int // traffic class; formerly experimental use @@ -19,7 +19,7 @@ const ( typeIncomingMPLSLabelStack = 1 ) -// A MPLSLabelStack represents a MPLS label stack. +// MPLSLabelStack represents an MPLS label stack. type MPLSLabelStack struct { Class int // extension object class number Type int // extension object sub-type diff --git a/vendor/golang.org/x/net/internal/iana/const.go b/vendor/golang.org/x/net/internal/iana/const.go index 826633e1b..cea712fac 100644 --- a/vendor/golang.org/x/net/internal/iana/const.go +++ b/vendor/golang.org/x/net/internal/iana/const.go @@ -4,38 +4,34 @@ // Package iana provides protocol number resources managed by the Internet Assigned Numbers Authority (IANA). package iana // import "golang.org/x/net/internal/iana" -// Differentiated Services Field Codepoints (DSCP), Updated: 2017-05-12 +// Differentiated Services Field Codepoints (DSCP), Updated: 2018-05-04 const ( - DiffServCS0 = 0x0 // CS0 - DiffServCS1 = 0x20 // CS1 - DiffServCS2 = 0x40 // CS2 - DiffServCS3 = 0x60 // CS3 - DiffServCS4 = 0x80 // CS4 - DiffServCS5 = 0xa0 // CS5 - DiffServCS6 = 0xc0 // CS6 - DiffServCS7 = 0xe0 // CS7 - DiffServAF11 = 0x28 // AF11 - DiffServAF12 = 0x30 // AF12 - DiffServAF13 = 0x38 // AF13 - DiffServAF21 = 0x48 // AF21 - DiffServAF22 = 0x50 // AF22 - DiffServAF23 = 0x58 // AF23 - DiffServAF31 = 0x68 // AF31 - DiffServAF32 = 0x70 // AF32 - DiffServAF33 = 0x78 // AF33 - DiffServAF41 = 0x88 // AF41 - DiffServAF42 = 0x90 // AF42 - DiffServAF43 = 0x98 // AF43 - DiffServEF = 0xb8 // EF - DiffServVOICEADMIT = 0xb0 // VOICE-ADMIT -) - -// IPv4 TOS Byte and IPv6 Traffic Class Octet, Updated: 2001-09-06 -const ( - NotECNTransport = 0x0 // Not-ECT (Not ECN-Capable Transport) - ECNTransport1 = 0x1 // ECT(1) (ECN-Capable Transport(1)) - ECNTransport0 = 0x2 // ECT(0) (ECN-Capable Transport(0)) - CongestionExperienced = 0x3 // CE (Congestion Experienced) + DiffServCS0 = 0x00 // CS0 + DiffServCS1 = 0x20 // CS1 + DiffServCS2 = 0x40 // CS2 + DiffServCS3 = 0x60 // CS3 + DiffServCS4 = 0x80 // CS4 + DiffServCS5 = 0xa0 // CS5 + DiffServCS6 = 0xc0 // CS6 + DiffServCS7 = 0xe0 // CS7 + DiffServAF11 = 0x28 // AF11 + DiffServAF12 = 0x30 // AF12 + DiffServAF13 = 0x38 // AF13 + DiffServAF21 = 0x48 // AF21 + DiffServAF22 = 0x50 // AF22 + DiffServAF23 = 0x58 // AF23 + DiffServAF31 = 0x68 // AF31 + DiffServAF32 = 0x70 // AF32 + DiffServAF33 = 0x78 // AF33 + DiffServAF41 = 0x88 // AF41 + DiffServAF42 = 0x90 // AF42 + DiffServAF43 = 0x98 // AF43 + DiffServEF = 0xb8 // EF + DiffServVOICEADMIT = 0xb0 // VOICE-ADMIT + NotECNTransport = 0x00 // Not-ECT (Not ECN-Capable Transport) + ECNTransport1 = 0x01 // ECT(1) (ECN-Capable Transport(1)) + ECNTransport0 = 0x02 // ECT(0) (ECN-Capable Transport(0)) + CongestionExperienced = 0x03 // CE (Congestion Experienced) ) // Protocol Numbers, Updated: 2017-10-13 @@ -179,7 +175,7 @@ const ( ProtocolReserved = 255 // Reserved ) -// Address Family Numbers, Updated: 2016-10-25 +// Address Family Numbers, Updated: 2018-04-02 const ( AddrFamilyIPv4 = 1 // IP (IP version 4) AddrFamilyIPv6 = 2 // IP6 (IP version 6) diff --git a/vendor/golang.org/x/net/internal/iana/gen.go b/vendor/golang.org/x/net/internal/iana/gen.go index 2227e09e8..2a7661c27 100644 --- a/vendor/golang.org/x/net/internal/iana/gen.go +++ b/vendor/golang.org/x/net/internal/iana/gen.go @@ -31,16 +31,12 @@ var registries = []struct { "https://www.iana.org/assignments/dscp-registry/dscp-registry.xml", parseDSCPRegistry, }, - { - "https://www.iana.org/assignments/ipv4-tos-byte/ipv4-tos-byte.xml", - parseTOSTCByte, - }, { "https://www.iana.org/assignments/protocol-numbers/protocol-numbers.xml", parseProtocolNumbers, }, { - "http://www.iana.org/assignments/address-family-numbers/address-family-numbers.xml", + "https://www.iana.org/assignments/address-family-numbers/address-family-numbers.xml", parseAddrFamilyNumbers, }, } @@ -85,31 +81,39 @@ func parseDSCPRegistry(w io.Writer, r io.Reader) error { if err := dec.Decode(&dr); err != nil { return err } - drs := dr.escape() fmt.Fprintf(w, "// %s, Updated: %s\n", dr.Title, dr.Updated) fmt.Fprintf(w, "const (\n") - for _, dr := range drs { - fmt.Fprintf(w, "DiffServ%s = %#x", dr.Name, dr.Value) + for _, dr := range dr.escapeDSCP() { + fmt.Fprintf(w, "DiffServ%s = %#02x", dr.Name, dr.Value) fmt.Fprintf(w, "// %s\n", dr.OrigName) } + for _, er := range dr.escapeECN() { + fmt.Fprintf(w, "%s = %#02x", er.Descr, er.Value) + fmt.Fprintf(w, "// %s\n", er.OrigDescr) + } fmt.Fprintf(w, ")\n") return nil } type dscpRegistry struct { - XMLName xml.Name `xml:"registry"` - Title string `xml:"title"` - Updated string `xml:"updated"` - Note string `xml:"note"` - RegTitle string `xml:"registry>title"` - PoolRecords []struct { - Name string `xml:"name"` - Space string `xml:"space"` - } `xml:"registry>record"` - Records []struct { - Name string `xml:"name"` - Space string `xml:"space"` - } `xml:"registry>registry>record"` + XMLName xml.Name `xml:"registry"` + Title string `xml:"title"` + Updated string `xml:"updated"` + Note string `xml:"note"` + Registries []struct { + Title string `xml:"title"` + Registries []struct { + Title string `xml:"title"` + Records []struct { + Name string `xml:"name"` + Space string `xml:"space"` + } `xml:"record"` + } `xml:"registry"` + Records []struct { + Value string `xml:"value"` + Descr string `xml:"description"` + } `xml:"record"` + } `xml:"registry"` } type canonDSCPRecord struct { @@ -118,92 +122,84 @@ type canonDSCPRecord struct { Value int } -func (drr *dscpRegistry) escape() []canonDSCPRecord { - drs := make([]canonDSCPRecord, len(drr.Records)) - sr := strings.NewReplacer( - "+", "", - "-", "", - "/", "", - ".", "", - " ", "", - ) - for i, dr := range drr.Records { - s := strings.TrimSpace(dr.Name) - drs[i].OrigName = s - drs[i].Name = sr.Replace(s) - n, err := strconv.ParseUint(dr.Space, 2, 8) - if err != nil { +func (drr *dscpRegistry) escapeDSCP() []canonDSCPRecord { + var drs []canonDSCPRecord + for _, preg := range drr.Registries { + if !strings.Contains(preg.Title, "Differentiated Services Field Codepoints") { continue } - drs[i].Value = int(n) << 2 + for _, reg := range preg.Registries { + if !strings.Contains(reg.Title, "Pool 1 Codepoints") { + continue + } + drs = make([]canonDSCPRecord, len(reg.Records)) + sr := strings.NewReplacer( + "+", "", + "-", "", + "/", "", + ".", "", + " ", "", + ) + for i, dr := range reg.Records { + s := strings.TrimSpace(dr.Name) + drs[i].OrigName = s + drs[i].Name = sr.Replace(s) + n, err := strconv.ParseUint(dr.Space, 2, 8) + if err != nil { + continue + } + drs[i].Value = int(n) << 2 + } + } } return drs } -func parseTOSTCByte(w io.Writer, r io.Reader) error { - dec := xml.NewDecoder(r) - var ttb tosTCByte - if err := dec.Decode(&ttb); err != nil { - return err - } - trs := ttb.escape() - fmt.Fprintf(w, "// %s, Updated: %s\n", ttb.Title, ttb.Updated) - fmt.Fprintf(w, "const (\n") - for _, tr := range trs { - fmt.Fprintf(w, "%s = %#x", tr.Keyword, tr.Value) - fmt.Fprintf(w, "// %s\n", tr.OrigKeyword) - } - fmt.Fprintf(w, ")\n") - return nil +type canonECNRecord struct { + OrigDescr string + Descr string + Value int } -type tosTCByte struct { - XMLName xml.Name `xml:"registry"` - Title string `xml:"title"` - Updated string `xml:"updated"` - Note string `xml:"note"` - RegTitle string `xml:"registry>title"` - Records []struct { - Binary string `xml:"binary"` - Keyword string `xml:"keyword"` - } `xml:"registry>record"` -} - -type canonTOSTCByteRecord struct { - OrigKeyword string - Keyword string - Value int -} - -func (ttb *tosTCByte) escape() []canonTOSTCByteRecord { - trs := make([]canonTOSTCByteRecord, len(ttb.Records)) - sr := strings.NewReplacer( - "Capable", "", - "(", "", - ")", "", - "+", "", - "-", "", - "/", "", - ".", "", - " ", "", - ) - for i, tr := range ttb.Records { - s := strings.TrimSpace(tr.Keyword) - trs[i].OrigKeyword = s - ss := strings.Split(s, " ") - if len(ss) > 1 { - trs[i].Keyword = strings.Join(ss[1:], " ") - } else { - trs[i].Keyword = ss[0] - } - trs[i].Keyword = sr.Replace(trs[i].Keyword) - n, err := strconv.ParseUint(tr.Binary, 2, 8) - if err != nil { +func (drr *dscpRegistry) escapeECN() []canonECNRecord { + var ers []canonECNRecord + for _, reg := range drr.Registries { + if !strings.Contains(reg.Title, "ECN Field") { continue } - trs[i].Value = int(n) + ers = make([]canonECNRecord, len(reg.Records)) + sr := strings.NewReplacer( + "Capable", "", + "Not-ECT", "", + "ECT(1)", "", + "ECT(0)", "", + "CE", "", + "(", "", + ")", "", + "+", "", + "-", "", + "/", "", + ".", "", + " ", "", + ) + for i, er := range reg.Records { + s := strings.TrimSpace(er.Descr) + ers[i].OrigDescr = s + ss := strings.Split(s, " ") + if len(ss) > 1 { + ers[i].Descr = strings.Join(ss[1:], " ") + } else { + ers[i].Descr = ss[0] + } + ers[i].Descr = sr.Replace(er.Descr) + n, err := strconv.ParseUint(er.Value, 2, 8) + if err != nil { + continue + } + ers[i].Value = int(n) + } } - return trs + return ers } func parseProtocolNumbers(w io.Writer, r io.Reader) error { diff --git a/vendor/golang.org/x/net/internal/socks/dial_test.go b/vendor/golang.org/x/net/internal/socks/dial_test.go index 93101a61b..3a7a31bde 100644 --- a/vendor/golang.org/x/net/internal/socks/dial_test.go +++ b/vendor/golang.org/x/net/internal/socks/dial_test.go @@ -17,19 +17,11 @@ import ( "golang.org/x/net/internal/sockstest" ) -const ( - targetNetwork = "tcp6" - targetHostname = "fqdn.doesnotexist" - targetHostIP = "2001:db8::1" - targetPort = "5963" -) - func TestDial(t *testing.T) { t.Run("Connect", func(t *testing.T) { ss, err := sockstest.NewServer(sockstest.NoAuthRequired, sockstest.NoProxyRequired) if err != nil { - t.Error(err) - return + t.Fatal(err) } defer ss.Close() d := socks.NewDialer(ss.Addr().Network(), ss.Addr().String()) @@ -41,21 +33,45 @@ func TestDial(t *testing.T) { Username: "username", Password: "password", }).Authenticate - c, err := d.Dial(targetNetwork, net.JoinHostPort(targetHostIP, targetPort)) - if err == nil { - c.(*socks.Conn).BoundAddr() - c.Close() - } + c, err := d.DialContext(context.Background(), ss.TargetAddr().Network(), ss.TargetAddr().String()) if err != nil { - t.Error(err) - return + t.Fatal(err) + } + c.(*socks.Conn).BoundAddr() + c.Close() + }) + t.Run("ConnectWithConn", func(t *testing.T) { + ss, err := sockstest.NewServer(sockstest.NoAuthRequired, sockstest.NoProxyRequired) + if err != nil { + t.Fatal(err) + } + defer ss.Close() + c, err := net.Dial(ss.Addr().Network(), ss.Addr().String()) + if err != nil { + t.Fatal(err) + } + defer c.Close() + d := socks.NewDialer(ss.Addr().Network(), ss.Addr().String()) + d.AuthMethods = []socks.AuthMethod{ + socks.AuthMethodNotRequired, + socks.AuthMethodUsernamePassword, + } + d.Authenticate = (&socks.UsernamePassword{ + Username: "username", + Password: "password", + }).Authenticate + a, err := d.DialWithConn(context.Background(), c, ss.TargetAddr().Network(), ss.TargetAddr().String()) + if err != nil { + t.Fatal(err) + } + if _, ok := a.(*socks.Addr); !ok { + t.Fatalf("got %+v; want socks.Addr", a) } }) t.Run("Cancel", func(t *testing.T) { ss, err := sockstest.NewServer(sockstest.NoAuthRequired, blackholeCmdFunc) if err != nil { - t.Error(err) - return + t.Fatal(err) } defer ss.Close() d := socks.NewDialer(ss.Addr().Network(), ss.Addr().String()) @@ -63,7 +79,7 @@ func TestDial(t *testing.T) { defer cancel() dialErr := make(chan error) go func() { - c, err := d.DialContext(ctx, ss.TargetAddr().Network(), net.JoinHostPort(targetHostname, targetPort)) + c, err := d.DialContext(ctx, ss.TargetAddr().Network(), ss.TargetAddr().String()) if err == nil { c.Close() } @@ -73,41 +89,37 @@ func TestDial(t *testing.T) { cancel() err = <-dialErr if perr, nerr := parseDialError(err); perr != context.Canceled && nerr == nil { - t.Errorf("got %v; want context.Canceled or equivalent", err) - return + t.Fatalf("got %v; want context.Canceled or equivalent", err) } }) t.Run("Deadline", func(t *testing.T) { ss, err := sockstest.NewServer(sockstest.NoAuthRequired, blackholeCmdFunc) if err != nil { - t.Error(err) - return + t.Fatal(err) } defer ss.Close() d := socks.NewDialer(ss.Addr().Network(), ss.Addr().String()) ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(100*time.Millisecond)) defer cancel() - c, err := d.DialContext(ctx, ss.TargetAddr().Network(), net.JoinHostPort(targetHostname, targetPort)) + c, err := d.DialContext(ctx, ss.TargetAddr().Network(), ss.TargetAddr().String()) if err == nil { c.Close() } if perr, nerr := parseDialError(err); perr != context.DeadlineExceeded && nerr == nil { - t.Errorf("got %v; want context.DeadlineExceeded or equivalent", err) - return + t.Fatalf("got %v; want context.DeadlineExceeded or equivalent", err) } }) t.Run("WithRogueServer", func(t *testing.T) { ss, err := sockstest.NewServer(sockstest.NoAuthRequired, rogueCmdFunc) if err != nil { - t.Error(err) - return + t.Fatal(err) } defer ss.Close() d := socks.NewDialer(ss.Addr().Network(), ss.Addr().String()) for i := 0; i < 2*len(rogueCmdList); i++ { ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(100*time.Millisecond)) defer cancel() - c, err := d.DialContext(ctx, targetNetwork, net.JoinHostPort(targetHostIP, targetPort)) + c, err := d.DialContext(ctx, ss.TargetAddr().Network(), ss.TargetAddr().String()) if err == nil { t.Log(c.(*socks.Conn).BoundAddr()) c.Close() diff --git a/vendor/golang.org/x/net/internal/socks/socks.go b/vendor/golang.org/x/net/internal/socks/socks.go index 915859536..d93e699b4 100644 --- a/vendor/golang.org/x/net/internal/socks/socks.go +++ b/vendor/golang.org/x/net/internal/socks/socks.go @@ -75,7 +75,7 @@ const ( AuthMethodNotRequired AuthMethod = 0x00 // no authentication required AuthMethodUsernamePassword AuthMethod = 0x02 // use username/password - AuthMethodNoAcceptableMethods AuthMethod = 0xff // no acceptable authetication methods + AuthMethodNoAcceptableMethods AuthMethod = 0xff // no acceptable authentication methods StatusSucceeded Reply = 0x00 ) @@ -149,20 +149,13 @@ type Dialer struct { // See func Dial of the net package of standard library for a // description of the network and address parameters. func (d *Dialer) DialContext(ctx context.Context, network, address string) (net.Conn, error) { - switch network { - case "tcp", "tcp6", "tcp4": - default: + if err := d.validateTarget(network, address); err != nil { proxy, dst, _ := d.pathAddrs(address) - return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: errors.New("network not implemented")} - } - switch d.cmd { - case CmdConnect, cmdBind: - default: - proxy, dst, _ := d.pathAddrs(address) - return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: errors.New("command not implemented")} + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: err} } if ctx == nil { - ctx = context.Background() + proxy, dst, _ := d.pathAddrs(address) + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: errors.New("nil context")} } var err error var c net.Conn @@ -185,11 +178,69 @@ func (d *Dialer) DialContext(ctx context.Context, network, address string) (net. return &Conn{Conn: c, boundAddr: a}, nil } +// DialWithConn initiates a connection from SOCKS server to the target +// network and address using the connection c that is already +// connected to the SOCKS server. +// +// It returns the connection's local address assigned by the SOCKS +// server. +func (d *Dialer) DialWithConn(ctx context.Context, c net.Conn, network, address string) (net.Addr, error) { + if err := d.validateTarget(network, address); err != nil { + proxy, dst, _ := d.pathAddrs(address) + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: err} + } + if ctx == nil { + proxy, dst, _ := d.pathAddrs(address) + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: errors.New("nil context")} + } + a, err := d.connect(ctx, c, address) + if err != nil { + proxy, dst, _ := d.pathAddrs(address) + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: err} + } + return a, nil +} + // Dial connects to the provided address on the provided network. // -// Deprecated: Use DialContext instead. +// Unlike DialContext, it returns a raw transport connection instead +// of a forward proxy connection. +// +// Deprecated: Use DialContext or DialWithConn instead. func (d *Dialer) Dial(network, address string) (net.Conn, error) { - return d.DialContext(context.Background(), network, address) + if err := d.validateTarget(network, address); err != nil { + proxy, dst, _ := d.pathAddrs(address) + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: err} + } + var err error + var c net.Conn + if d.ProxyDial != nil { + c, err = d.ProxyDial(context.Background(), d.proxyNetwork, d.proxyAddress) + } else { + c, err = net.Dial(d.proxyNetwork, d.proxyAddress) + } + if err != nil { + proxy, dst, _ := d.pathAddrs(address) + return nil, &net.OpError{Op: d.cmd.String(), Net: network, Source: proxy, Addr: dst, Err: err} + } + if _, err := d.DialWithConn(context.Background(), c, network, address); err != nil { + return nil, err + } + return c, nil +} + +func (d *Dialer) validateTarget(network, address string) error { + switch network { + case "tcp", "tcp6", "tcp4": + default: + return errors.New("network not implemented") + } + switch d.cmd { + case CmdConnect, cmdBind: + default: + return errors.New("command not implemented") + } + return nil } func (d *Dialer) pathAddrs(address string) (proxy, dst net.Addr, err error) { diff --git a/vendor/golang.org/x/net/ipv4/doc.go b/vendor/golang.org/x/net/ipv4/doc.go index b43935a5a..d12455548 100644 --- a/vendor/golang.org/x/net/ipv4/doc.go +++ b/vendor/golang.org/x/net/ipv4/doc.go @@ -241,4 +241,4 @@ // IncludeSourceSpecificGroup may return an error. package ipv4 // import "golang.org/x/net/ipv4" -// BUG(mikio): This package is not implemented on NaCl and Plan 9. +// BUG(mikio): This package is not implemented on JS, NaCl and Plan 9. diff --git a/vendor/golang.org/x/net/ipv4/multicast_test.go b/vendor/golang.org/x/net/ipv4/multicast_test.go index bcf49736b..2136a3c6e 100644 --- a/vendor/golang.org/x/net/ipv4/multicast_test.go +++ b/vendor/golang.org/x/net/ipv4/multicast_test.go @@ -29,7 +29,7 @@ var packetConnReadWriteMulticastUDPTests = []struct { func TestPacketConnReadWriteMulticastUDP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "solaris", "windows": + case "js", "nacl", "plan9", "solaris", "windows": t.Skipf("not supported on %s", runtime.GOOS) } ifi := nettest.RoutedInterface("ip4", net.FlagUp|net.FlagMulticast|net.FlagLoopback) @@ -117,7 +117,7 @@ var packetConnReadWriteMulticastICMPTests = []struct { func TestPacketConnReadWriteMulticastICMP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "solaris", "windows": + case "js", "nacl", "plan9", "solaris", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if m, ok := nettest.SupportsRawIPSocket(); !ok { diff --git a/vendor/golang.org/x/net/ipv4/multicastlistener_test.go b/vendor/golang.org/x/net/ipv4/multicastlistener_test.go index e43fbbe08..ba2480e48 100644 --- a/vendor/golang.org/x/net/ipv4/multicastlistener_test.go +++ b/vendor/golang.org/x/net/ipv4/multicastlistener_test.go @@ -21,7 +21,7 @@ var udpMultipleGroupListenerTests = []net.Addr{ func TestUDPSinglePacketConnWithMultipleGroupListeners(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if testing.Short() { @@ -61,7 +61,7 @@ func TestUDPSinglePacketConnWithMultipleGroupListeners(t *testing.T) { func TestUDPMultiplePacketConnWithMultipleGroupListeners(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if testing.Short() { @@ -116,7 +116,7 @@ func TestUDPMultiplePacketConnWithMultipleGroupListeners(t *testing.T) { func TestUDPPerInterfaceSinglePacketConnWithSingleGroupListener(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if testing.Short() { @@ -172,7 +172,7 @@ func TestUDPPerInterfaceSinglePacketConnWithSingleGroupListener(t *testing.T) { func TestIPSingleRawConnWithSingleGroupListener(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if testing.Short() { @@ -217,7 +217,7 @@ func TestIPSingleRawConnWithSingleGroupListener(t *testing.T) { func TestIPPerInterfaceSingleRawConnWithSingleGroupListener(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if testing.Short() { diff --git a/vendor/golang.org/x/net/ipv4/multicastsockopt_test.go b/vendor/golang.org/x/net/ipv4/multicastsockopt_test.go index f7efac24c..c0a26800a 100644 --- a/vendor/golang.org/x/net/ipv4/multicastsockopt_test.go +++ b/vendor/golang.org/x/net/ipv4/multicastsockopt_test.go @@ -26,7 +26,7 @@ var packetConnMulticastSocketOptionTests = []struct { func TestPacketConnMulticastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9": + case "js", "nacl", "plan9": t.Skipf("not supported on %s", runtime.GOOS) } ifi := nettest.RoutedInterface("ip4", net.FlagUp|net.FlagMulticast|net.FlagLoopback) @@ -66,7 +66,7 @@ var rawConnMulticastSocketOptionTests = []struct { func TestRawConnMulticastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9": + case "js", "nacl", "plan9": t.Skipf("not supported on %s", runtime.GOOS) } if m, ok := nettest.SupportsRawIPSocket(); !ok { diff --git a/vendor/golang.org/x/net/ipv4/readwrite_go1_8_test.go b/vendor/golang.org/x/net/ipv4/readwrite_go1_8_test.go index 1cd926e7f..83bf927ed 100644 --- a/vendor/golang.org/x/net/ipv4/readwrite_go1_8_test.go +++ b/vendor/golang.org/x/net/ipv4/readwrite_go1_8_test.go @@ -22,7 +22,7 @@ import ( func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": b.Skipf("not supported on %s", runtime.GOOS) } @@ -126,7 +126,7 @@ func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { func TestPacketConnConcurrentReadWriteUnicast(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } diff --git a/vendor/golang.org/x/net/ipv4/readwrite_go1_9_test.go b/vendor/golang.org/x/net/ipv4/readwrite_go1_9_test.go index 365de022a..ef76c8a72 100644 --- a/vendor/golang.org/x/net/ipv4/readwrite_go1_9_test.go +++ b/vendor/golang.org/x/net/ipv4/readwrite_go1_9_test.go @@ -22,7 +22,7 @@ import ( func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": b.Skipf("not supported on %s", runtime.GOOS) } @@ -172,7 +172,7 @@ func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { func TestPacketConnConcurrentReadWriteUnicast(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } diff --git a/vendor/golang.org/x/net/ipv4/readwrite_test.go b/vendor/golang.org/x/net/ipv4/readwrite_test.go index 3896a8ae4..1b7f74884 100644 --- a/vendor/golang.org/x/net/ipv4/readwrite_test.go +++ b/vendor/golang.org/x/net/ipv4/readwrite_test.go @@ -61,7 +61,7 @@ func BenchmarkReadWriteUnicast(b *testing.B) { func TestPacketConnConcurrentReadWriteUnicastUDP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } diff --git a/vendor/golang.org/x/net/ipv4/unicast_test.go b/vendor/golang.org/x/net/ipv4/unicast_test.go index 02c089f00..274a558c7 100644 --- a/vendor/golang.org/x/net/ipv4/unicast_test.go +++ b/vendor/golang.org/x/net/ipv4/unicast_test.go @@ -20,7 +20,7 @@ import ( func TestPacketConnReadWriteUnicastUDP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } ifi := nettest.RoutedInterface("ip4", net.FlagUp|net.FlagLoopback) @@ -71,7 +71,7 @@ func TestPacketConnReadWriteUnicastUDP(t *testing.T) { func TestPacketConnReadWriteUnicastICMP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if m, ok := nettest.SupportsRawIPSocket(); !ok { @@ -157,7 +157,7 @@ func TestPacketConnReadWriteUnicastICMP(t *testing.T) { func TestRawConnReadWriteUnicastICMP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if m, ok := nettest.SupportsRawIPSocket(); !ok { diff --git a/vendor/golang.org/x/net/ipv4/unicastsockopt_test.go b/vendor/golang.org/x/net/ipv4/unicastsockopt_test.go index db5213b91..2bfda7ff6 100644 --- a/vendor/golang.org/x/net/ipv4/unicastsockopt_test.go +++ b/vendor/golang.org/x/net/ipv4/unicastsockopt_test.go @@ -16,7 +16,7 @@ import ( func TestConnUnicastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } ifi := nettest.RoutedInterface("ip4", net.FlagUp|net.FlagLoopback) @@ -62,7 +62,7 @@ var packetConnUnicastSocketOptionTests = []struct { func TestPacketConnUnicastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } ifi := nettest.RoutedInterface("ip4", net.FlagUp|net.FlagLoopback) @@ -88,7 +88,7 @@ func TestPacketConnUnicastSocketOptions(t *testing.T) { func TestRawConnUnicastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if m, ok := nettest.SupportsRawIPSocket(); !ok { diff --git a/vendor/golang.org/x/net/ipv6/doc.go b/vendor/golang.org/x/net/ipv6/doc.go index 664a97dea..cb4000623 100644 --- a/vendor/golang.org/x/net/ipv6/doc.go +++ b/vendor/golang.org/x/net/ipv6/doc.go @@ -240,4 +240,4 @@ // IncludeSourceSpecificGroup may return an error. package ipv6 // import "golang.org/x/net/ipv6" -// BUG(mikio): This package is not implemented on NaCl and Plan 9. +// BUG(mikio): This package is not implemented on JS, NaCl and Plan 9. diff --git a/vendor/golang.org/x/net/ipv6/icmp_test.go b/vendor/golang.org/x/net/ipv6/icmp_test.go index d8e9675dc..bdde2fdbe 100644 --- a/vendor/golang.org/x/net/ipv6/icmp_test.go +++ b/vendor/golang.org/x/net/ipv6/icmp_test.go @@ -34,7 +34,7 @@ func TestICMPString(t *testing.T) { func TestICMPFilter(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } @@ -61,7 +61,7 @@ func TestICMPFilter(t *testing.T) { func TestSetICMPFilter(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/multicast_test.go b/vendor/golang.org/x/net/ipv6/multicast_test.go index 69a21cd38..50442c368 100644 --- a/vendor/golang.org/x/net/ipv6/multicast_test.go +++ b/vendor/golang.org/x/net/ipv6/multicast_test.go @@ -29,7 +29,7 @@ var packetConnReadWriteMulticastUDPTests = []struct { func TestPacketConnReadWriteMulticastUDP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -129,7 +129,7 @@ var packetConnReadWriteMulticastICMPTests = []struct { func TestPacketConnReadWriteMulticastICMP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/multicastlistener_test.go b/vendor/golang.org/x/net/ipv6/multicastlistener_test.go index b27713e2f..1ee0549d7 100644 --- a/vendor/golang.org/x/net/ipv6/multicastlistener_test.go +++ b/vendor/golang.org/x/net/ipv6/multicastlistener_test.go @@ -21,7 +21,7 @@ var udpMultipleGroupListenerTests = []net.Addr{ func TestUDPSinglePacketConnWithMultipleGroupListeners(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -61,7 +61,7 @@ func TestUDPSinglePacketConnWithMultipleGroupListeners(t *testing.T) { func TestUDPMultiplePacketConnWithMultipleGroupListeners(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -116,7 +116,7 @@ func TestUDPMultiplePacketConnWithMultipleGroupListeners(t *testing.T) { func TestUDPPerInterfaceSinglePacketConnWithSingleGroupListener(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -172,7 +172,7 @@ func TestUDPPerInterfaceSinglePacketConnWithSingleGroupListener(t *testing.T) { func TestIPSinglePacketConnWithSingleGroupListener(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -216,7 +216,7 @@ func TestIPPerInterfaceSinglePacketConnWithSingleGroupListener(t *testing.T) { switch runtime.GOOS { case "darwin", "dragonfly", "openbsd": // platforms that return fe80::1%lo0: bind: can't assign requested address t.Skipf("not supported on %s", runtime.GOOS) - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/multicastsockopt_test.go b/vendor/golang.org/x/net/ipv6/multicastsockopt_test.go index 9e6b902d7..9906e8767 100644 --- a/vendor/golang.org/x/net/ipv6/multicastsockopt_test.go +++ b/vendor/golang.org/x/net/ipv6/multicastsockopt_test.go @@ -26,7 +26,7 @@ var packetConnMulticastSocketOptionTests = []struct { func TestPacketConnMulticastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/readwrite_go1_8_test.go b/vendor/golang.org/x/net/ipv6/readwrite_go1_8_test.go index c11d92ae9..b3fbda13a 100644 --- a/vendor/golang.org/x/net/ipv6/readwrite_go1_8_test.go +++ b/vendor/golang.org/x/net/ipv6/readwrite_go1_8_test.go @@ -22,7 +22,7 @@ import ( func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": b.Skipf("not supported on %s", runtime.GOOS) } @@ -123,7 +123,7 @@ func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { func TestPacketConnConcurrentReadWriteUnicast(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } diff --git a/vendor/golang.org/x/net/ipv6/readwrite_go1_9_test.go b/vendor/golang.org/x/net/ipv6/readwrite_go1_9_test.go index e2fd73370..7110619b8 100644 --- a/vendor/golang.org/x/net/ipv6/readwrite_go1_9_test.go +++ b/vendor/golang.org/x/net/ipv6/readwrite_go1_9_test.go @@ -22,7 +22,7 @@ import ( func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": b.Skipf("not supported on %s", runtime.GOOS) } @@ -169,7 +169,7 @@ func BenchmarkPacketConnReadWriteUnicast(b *testing.B) { func TestPacketConnConcurrentReadWriteUnicast(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } diff --git a/vendor/golang.org/x/net/ipv6/readwrite_test.go b/vendor/golang.org/x/net/ipv6/readwrite_test.go index 206b915ce..668208f22 100644 --- a/vendor/golang.org/x/net/ipv6/readwrite_test.go +++ b/vendor/golang.org/x/net/ipv6/readwrite_test.go @@ -65,7 +65,7 @@ func BenchmarkReadWriteUnicast(b *testing.B) { func TestPacketConnConcurrentReadWriteUnicastUDP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/sockopt_test.go b/vendor/golang.org/x/net/ipv6/sockopt_test.go index 774338dbf..add045e3e 100644 --- a/vendor/golang.org/x/net/ipv6/sockopt_test.go +++ b/vendor/golang.org/x/net/ipv6/sockopt_test.go @@ -19,7 +19,7 @@ var supportsIPv6 bool = nettest.SupportsIPv6() func TestConnInitiatorPathMTU(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -57,7 +57,7 @@ func TestConnInitiatorPathMTU(t *testing.T) { func TestConnResponderPathMTU(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -95,7 +95,7 @@ func TestConnResponderPathMTU(t *testing.T) { func TestPacketConnChecksum(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/unicast_test.go b/vendor/golang.org/x/net/ipv6/unicast_test.go index a0b7d9550..159bb39b9 100644 --- a/vendor/golang.org/x/net/ipv6/unicast_test.go +++ b/vendor/golang.org/x/net/ipv6/unicast_test.go @@ -20,7 +20,7 @@ import ( func TestPacketConnReadWriteUnicastUDP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -78,7 +78,7 @@ func TestPacketConnReadWriteUnicastUDP(t *testing.T) { func TestPacketConnReadWriteUnicastICMP(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/ipv6/unicastsockopt_test.go b/vendor/golang.org/x/net/ipv6/unicastsockopt_test.go index e175dccf5..69adefe8f 100644 --- a/vendor/golang.org/x/net/ipv6/unicastsockopt_test.go +++ b/vendor/golang.org/x/net/ipv6/unicastsockopt_test.go @@ -16,7 +16,7 @@ import ( func TestConnUnicastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { @@ -61,7 +61,7 @@ var packetConnUnicastSocketOptionTests = []struct { func TestPacketConnUnicastSocketOptions(t *testing.T) { switch runtime.GOOS { - case "nacl", "plan9", "windows": + case "js", "nacl", "plan9", "windows": t.Skipf("not supported on %s", runtime.GOOS) } if !supportsIPv6 { diff --git a/vendor/golang.org/x/net/publicsuffix/table.go b/vendor/golang.org/x/net/publicsuffix/table.go index a870b36cd..ca7f32c1f 100644 --- a/vendor/golang.org/x/net/publicsuffix/table.go +++ b/vendor/golang.org/x/net/publicsuffix/table.go @@ -2,7 +2,7 @@ package publicsuffix -const version = "publicsuffix.org's public_suffix_list.dat, git revision 0f3b07d9aab6d6c9fe74990af98316468d40f488 (2018-01-25T09:22:16Z)" +const version = "publicsuffix.org's public_suffix_list.dat, git revision 2225db8d9f4a2a27ec697c883360632fa0c16261 (2018-05-23T23:26:06Z)" const ( nodesBitsChildren = 10 @@ -23,464 +23,476 @@ const ( ) // numTLD is the number of top level domains. -const numTLD = 1551 +const numTLD = 1555 // Text is the combined text of all labels. -const text = "0emmafann-arboretumbriamallamaceiobihirosakikamijimatsuzaki234li" + - "ma-cityeatselinogradult3l3p0rtargets-itargivestbytomaritimekeepi" + - "ng120009guacuiababia-goracleaningroks-theatreeastcoastaldefencea" + - "tonsbergjemnes3-ap-northeast-1337bilbaogashimadachicagoboats3-we" + - "bsite-us-east-1billustrationikonanporovnopocznoppdalindesnes3-we" + - "bsite-us-west-1biobirdartcenterprisesakimobetsuitainairforcechir" + - "ealminamiechizeninohekinannestadiybirkenesoddtangenovaranzanpach" + - "igasakievennodesaarlandnpanasonicateringebuilderschmidtre-gaulda" + - "livornobirthplacebitballooningladefinimakanegasakinkobayashikaoi" + - "rminamifuranobjarkoybjerkreimbananarepublicasadelamonedatingjesd" + - "alimitediscountysvardolls3-eu-west-3utilitiesquare7bjugninomiyak" + - "onojorpelandrangedalombardiamonds3-website-us-west-2blancomedica" + - "ltanissettaipeiheijinuyamashinatsukigatakasagotpantheonsitebloom" + - "bergbauernuorochesterbloxcms5ybluedancebmoattachmentsakyotanabel" + - "lunord-aurdalvdalcesalangenirasakinvestmentsalondonetskarmoybmsa" + - "ltdalombardynamisches-dnsaludray-dnsupdaternopilawawebspacebmwed" + - "dinglassassinationalheritagebnpparibaselburgleezebnrwedeployboml" + - "oansalvadordalibabalsanagochihayaakasakawaharaholtalenvironmenta" + - "lconservationishiazainzais-a-candidatebondrayddnsfreebox-osascol" + - "i-picenordre-landraydnsalzburgliwicebonnishigobookinglobalashovh" + - "achinohedmarkarpaczeladzparaglidingloboavistaprintelligencebooml" + - "adbrokesamegawabootsamnangerboschaefflerdalwaysdatabaseballangen" + - "oamishirasatochigiessensiositelekommunikationishiharabostikaruiz" + - "awabostonakijinsekikogentinglogowegroweibolognagasukebotanicalga" + - "rdenishiizunazukis-a-catererbotanicgardenishikatakayamatsushigeb" + - "otanybouncemerckmsdnipropetrovskjervoyagebounty-fullensakerrypro" + - "pertiesampagespeedmobilizeroboutiquebecatholicaxiascolipicenodum" + - "inamiiselectjomemorialomzaporizhzheguris-a-celticsfanishikatsura" + - "git-repostfoldnavybozentsujiiebplacedekagaminord-odalondrinaples" + - "amsclubindalorenskogloppenzaolbia-tempio-olbiatempioolbialystokk" + - "embuchikumagayagawakuyabukihokumakogenglandrivelandrobaknoluokta" + - "chikawakkanaibetsubamericanfamilydscloudcontrolappspotagerbrandy" + - "winevalleybrasiliabrindisibenikebristoloseyouripirangapartmentsa" + - "msungmbhartiffanybritishcolumbialowiezachpomorskienishikawazukam" + - "itsuebroadcastlefrakkestadrudunsandvikcoromantovalle-d-aostathel" + - "lebroadwaybroke-itjxjavald-aostaplesanfranciscofreakunemurorange" + - "iseiyoichippubetsubetsugarugbyengerdalaskanittedallasalleasingle" + - "surancertmgretagajobojis-a-chefarmsteadupontariodejaneirodoybrok" + - "erbronnoysundurbanamexnetlifyis-a-conservativefsnillfjordurhambu" + - "rgminakamichiharabrothermesaverdeatnurembergmodellingmxn--0trq7p" + - "7nnishimerabrowsersafetymarketsangobrumunddalotenkawabrunelastic" + - "beanstalkarumaifarsundyndns-at-workinggrouparisor-fronishinomiya" + - "shironobrusselsanjotkmaxxn--11b4c3dyndns-blogdnsannanishinoomote" + - "gobruxellesannohelplfinancialottebryanskleppgafanquannefrankfurt" + - "ksatxn--12c1fe0bradescorporationishinoshimatsuurabrynewjerseybus" + - "kerudinewportlligatmparliamentoyosatoyonakagyokutoyokawabuzenish" + - "iokoppegardyndns-freeboxoslodingenishitosashimizunaminamibosognd" + - "alottokorozawabuzzweirbwfashionishiwakis-a-cpadualstackspace-to-" + - "rentalstomakomaibarabzhitomirumalatvuopmicrolightingrimstadyndns" + - "-homednsanokasaokaminokawanishiaizubangecommunitysnesardegnaroyc" + - "omobaracomparemarkerryhotelsardiniacompute-1computerhistoryofsci" + - "ence-fictioncomsecuritytacticsarlutskashiwazakiyosemitecondoshic" + - "hinohealth-carereformitakeharaconferenceconstructionconsuladohar" + - "uovatrani-andria-barletta-trani-andriaconsultanthropologyconsult" + - "ingvolluxembourgruecontactraniandriabarlettatraniandriacontagema" + - "tsubaracontemporaryarteducationalchikugojomedio-campidano-medioc" + - "ampidanomediocontractorskenconventureshinodearthdfcbankasukabedz" + - "in-the-bandaioiraseeklogest-mon-blogueurovisionionjukudoyamainte" + - "nancebetsuikidsmynasushiobarackmazerbaijan-mayenebakkeshibechamb" + - "agriculturennebudapest-a-la-masionthewifiat-band-campaniacooking" + - "channelsdvrdnsdojoetsuwanouchikujogaszczytnordlandyndns-weberlin" + - "colncoolkuszkolahppiacenzagancooperativano-frankivskodjeffersonc" + - "openhagencyclopedichernivtsiciliacorsicagliaribeiraokinawashiros" + - "atochiokinoshimaizuruhrcorvettemasekasumigaurawa-mazowszextraspa" + - "cekitagatajirissagamiharacosenzakopanerairguardiannakadomarinebr" + - "askaunjargalsaceocosidnsfor-better-thanawatchesarpsborguitarsaru" + - "futsunomiyawakasaikaitakoelncostumedizinhistorischesasayamacouch" + - "potatofriesasebofagecounciluxurycouponsaskatchewancoursesassaris" + - "-a-doctoraycq-acranbrookuwanalyticsaudacreditcardyndns-wikiracre" + - "ditunioncremonashgabadaddjaguarqhachiojiyahoooshikamaishimodatec" + - "rewhoswhokksundyndns-workisboringujoinvillewismillercricketrzync" + - "rimeast-kazakhstanangercrotonexus-3crownprovidercrsvparsauherady" + - "ndns1cruisesavannahgacryptonomichigangwoncuisinellair-traffic-co" + - "ntrolleyculturalcentertainmentranoycuneocupcakecuritibaghdadynns" + - "aves-the-whalessandria-trani-barletta-andriatranibarlettaandriac" + - "xn--12cfi8ixb8luzerncyberlevagangaviikanonjis-a-financialadvisor" + - "-aurdalvivanovodkamisatokashikiwakunigamiharufcfancymrussiacyona" + - "barulsandoycyoutheworkpccwiiheyakagefgushikamifuranorth-kazakhst" + - "anfhvalerfidonnakanotoddenfieldynvpnchernovtsykkylvenetogakushim" + - "otoganewyorkshirecipesaro-urbino-pesarourbinopesaromasvuotnakaiw" + - "amizawassamukawataricohdatsunanjoburgriwataraidyndns-iparmattele" + - "fonicapitalonewspaperfigueresinstagingxn--1ctwolominamatakkokami" + - "noyamaxunusualpersonfilateliafilegearfilminamimakis-a-geekaszuby" + - "finalfinancefineartscholarshipschoolfinlandyroyrvikingulenfinnoy" + - "firebaseappartis-a-greenfirenzefirestonefirmdaleirvikatowicefish" + - "ingolffanschulefitjarfitnessettlementransurlfjalerflesbergflickr" + - "agerotikakamigaharaflightschwarzgwangjuniperflirflogintohmalvika" + - "tsushikabeeldengeluidfloraflorencefloridavvesiidazaifudaigokasel" + - "jordfloripaderbornfloristanohatakahamamurogawaflorogerschweizflo" + - "wersciencecentersciencehistoryflynnhosting-clusterflynnhubarclay" + - "s3-sa-east-1fndfor-ourfor-someeresistancefor-theaterforexrothach" + - "irogatakamoriokalmykiaforgotdnscientistockholmestrandforli-cesen" + - "a-forlicesenaforlikescandynamic-dnscjohnsonforsaleitungsenforsan" + - "dasuoloftrapaniizafortalfortmissoulancashireggio-calabriafortwor" + - "thadanorthwesternmutualforuminamiminowafosnescotlandfotaruis-a-g" + - "urufoxfordebianfozorafredrikstadtvscrapper-sitefreeddnsgeekgalax" + - "yfreemasonryfreesitevadsochildrensgardenfreetlscrappingfreiburgf" + - "reightravelchannelfreseniuscountryestateofdelawarezzoologyfribou" + - "rgfriuli-v-giuliafriuli-ve-giuliafriuli-vegiuliafriuli-venezia-g" + - "iuliafriuli-veneziagiuliafriuli-vgiuliafriuliv-giuliafriulive-gi" + - "uliafriulivegiuliafriulivenezia-giuliafriuliveneziagiuliafriuliv" + - "giuliafrlfroganscrysechirurgiens-dentistes-en-francefrognfroland" + - "from-akrehamnfrom-alfrom-arfrom-azfrom-canonoichinomiyakefrom-co" + - "dynaliasdaburfrom-ctravelersinsurancefrom-dchiryukyuragifuchungb" + - "ukharafrom-dedyn-ip24from-flanderservegame-serversicherungfrom-g" + - "ausdalfrom-higashiagatsumagoianiafrom-iafrom-idfrom-ilfrom-inche" + - "onfrom-kservehalflifestylefrom-kyowariasahikawafrom-lancasterfro" + - "m-mangonohejis-a-hard-workerfrom-mdfrom-meethnologyfrom-mifunefr" + - "om-mnfrom-modalenfrom-mservehttpartnerservehumourfrom-mtnfrom-nc" + - "hitachinakagawatchandclockashibatakashimarumorimachidafrom-ndfro" + - "m-nefrom-nh-servebbserveirchitosetogitsuliguriafrom-njaworznotog" + - "awafrom-nminamiogunicomcastresindeviceserveminecraftrdfrom-nv-in" + - "foodnetworkshoppingfrom-nyfrom-ohtawaramotoineppuboliviajessheim" + - "periafrom-oketohnoshooguyfrom-orfrom-padovaksdalfrom-pratohobby-" + - "sitexashorokanaiefrom-rivnefrom-schoenbrunnfrom-sdfrom-tnfrom-tx" + - "n--1lqs03nfrom-utazuerichardlillehammerfeste-ipartservemp3from-v" + - "al-daostavalleyfrom-vtrentino-a-adigefrom-wafrom-wielunnerfrom-w" + - "valled-aostatoilfrom-wyfrosinonefrostalowa-wolawafroyahikobeardu" + - "baiduckdnservep2partyfstavernfujiiderafujikawaguchikonefujiminok" + - "amoenairtelecitychyattorneyagawakeisenbahnfujinomiyadafujiokayam" + - "angyshlakasamatsudontexistmein-vigorgefujisatoshonairtrafficplex" + - "us-1fujisawafujishiroishidakabiratoridefensells-for-lesservepics" + - "ervequakefujitsurugashimaringatlantakaharufujixeroxn--1lqs71dfuj" + - "iyoshidafukayabeatservesarcasmatartanddesignfukuchiyamadafukudom" + - "inichocolatelevisionissedalouvreisenisshingugefukuis-a-hunterfuk" + - "umitsubishigakirovogradoyfukuokazakiryuohadselfipasadenaritakura" + - "shikis-a-knightpointtokamachintaifun-dnsaliasiafukuroishikarikat" + - "urindalfukusakisarazurewebsiteshikagamiishibukawafukuyamagatakah" + - "ashimamakishiwadafunabashiriuchinadafunagatakahatakaishimogosenf" + - "unahashikamiamakusatsumasendaisennangoodyearfundaciofuoiskujukur" + - "iyamaniwakuratextileksvikatsuyamarylandfuosskoczowildlifedorainf" + - "racloudcontrolledogawarabikomaezakirunore-og-uvdalfurnitureggio-" + - "emilia-romagnakatombetsumitakagiizefurubirafurudonostiaarpassage" + - "nservicesettsurgeonshalloffameloyalistjordalshalsenfurukawais-a-" + - "landscaperfusodegaurafussaikisofukushimannorfolkebiblelveruminam" + - "isanrikubetsupportrentino-aadigefutabayamaguchinomigawafutboldly" + - "goingnowhere-for-morenakatsugawafuttsurugiminamitanefuturecmseva" + - "stopolefuturehostingfuturemailingfvgfylkesbiblackfridayfyresdalh" + - "angoutsystemscloudfunctionsevenassisicilyhannanmokuizumodenakaya" + - "mapassenger-associationhannosegawahanyuzenhapmirhareidsbergenhar" + - "stadharvestcelebrationhasamarburghasaminami-alpssells-itrentino-" + - "altoadigehashbanghasudahasura-appatriahasvikazohatogayaitakanabe" + - "autysfjordhatoyamazakitakamiizumisanofidelityhatsukaichikaiseis-" + - "a-linux-useranishiaritabashijonawatehattfjelldalhayashimamotobun" + - "gotakadapliernewmexicoalhazuminobusellsyourhomegoodsewilliamhill" + - "hbodoes-itvedestrandhelsinkitakatakanezawahembygdsforbundhemnesh" + - "aris-a-llamarriottrentino-s-tirollagrigentomologyeonggiehtavuoat" + - "nagaivuotnagaokakyotambabydgoszczecinemaceratabusebastopologyeon" + - "gnamegawakayamadridhemsedalhepforgeherokussldheroyhgtvalledaosta" + - "vangerhigashichichibunkyonanaoshimageandsoundandvisionhigashihir" + - "oshimanehigashiizumozakitakyushuaiahigashikagawahigashikagurasoe" + - "dahigashikawakitaaikitamihamadahigashikurumeguromskoghigashimats" + - "ushimarcheapaviancargodaddyn-vpnplus-2higashimatsuyamakitaakitad" + - "aitoigawahigashimurayamamotorcyclesharpfizerhigashinarusembokuki" + - "tamotosumy-routerhigashinehigashiomihachimanaustdalhigashiosakas" + - "ayamanakakogawahigashishirakawamatakaokaluganskydivinghigashisum" + - "iyoshikawaminamiaikitanakagusukumodernhigashitsunoshiroomurahiga" + - "shiurausukitashiobarahigashiyamatokoriyamanashifteditchyouripgfo" + - "ggiahigashiyodogawahigashiyoshinogaris-a-musicianhiraizumisatoka" + - "izukamakurazakitaurayasudahirakatashinagawahiranais-a-nascarfanh" + - "irarahiratsukagawahirayaizuwakamatsubushikusakadogawahistorichou" + - "seshawaiijimaritimoduminamiyamashirokawanabelembetsukubankazunow" + - "tvallee-aosteroyhitachiomiyagildeskaliszhitachiotagoperauniteroi" + - "zumizakisosakitagawahitraeumtgeradellogliastradinghjartdalhjelme" + - "landholeckobierzyceholidayhomeipharmacienshellaspeziahomelinkddi" + - "elddanuorrikuzentakataiwanairlinedre-eikerhomelinuxn--1qqw23ahom" + - "eofficehomesecuritymacaparecidahomesecuritypchofunatoriginsurecr" + - "eationiyodogawahomesenseminehomeunixn--2m4a15ehondahoneywellbein" + - "gzonehongotembaixadahonjyoitakarazukameokameyamatotakadahorninda" + - "lhorseoullensvanguardhortendofinternet-dnshimojis-a-nurservebeer" + - "hospitalhoteleshimokawahotmailhoyangerhoylandetroitskypehumaniti" + - "eshimokitayamahurdalhurumajis-a-painteractivegarsheis-a-patsfanh" + - "yllestadhyogoris-a-personaltrainerhyugawarahyundaiwafunejewelryj" + - "ewishartgalleryjfkharkovanylvenicejgorajlcube-serverrankoshigaya" + - "kumoldelmenhorstalbanshinichinanjlljmphilatelyjnjcphiladelphiaar" + - "eadmyblogsitejoyentrentino-sued-tiroljoyokaichibalatinoipifonymi" + - "nanojpmorganjpnjprshinjournalismailillesandefjordjurkoshunantank" + - "hmelnitskiyamarylhurstjohnkosugekotohiradomainshinjukumanokotour" + - "akouhokutamakis-a-techietis-a-photographerokuappharmacyshimonita" + - "yanagithubusercontentrentino-stirolkounosupplieshinkamigotoyohas" + - "himotottoris-a-therapistoiakouyamashikekouzushimashikis-an-accou" + - "ntantshimonosekikawakozagawakozakis-an-actorkozowinbarrel-of-kno" + - "wledgeologyonagoyaustrheimatunduhrennesoyolasitebizenakasatsunai" + - "rportland-4-salernoboribetsucks3-eu-central-1kpnkppspdnshinshino" + - "tsurgerykrasnodarkredstonekristiansandcatshinshirokristiansundkr" + - "odsheradkrokstadelvaldaostarnbergkrymincommbankhmelnytskyivaokum" + - "atorinokumejimasoykumenantokonamegatakatoris-an-actresshimosuwal" + - "kis-a-playerkunisakis-an-anarchistoricalsocietykunitachiarailway" + - "kunitomigusukumamotoyamashikokuchuokunneppugliakunstsammlungkuns" + - "tunddesignkuokgrouphoenixn--30rr7ykurehabmerkurgankurobelaudible" + - "borkangerkurogiminamiashigarakuroisoftwarendalenugkuromatsunais-" + - "an-artisteinkjerusalembroiderykurotakikawasakis-an-engineeringku" + - "shirogawakustanais-an-entertainerkusupplykutchanelkutnokuzumakis" + - "-bykvafjordkvalsundkvamfamberkeleykvanangenkvinesdalkvinnheradkv" + - "iteseidskogkvitsoykwpspiegelkzmitoyoakemiuramiyazumiyotamanomjon" + - "dalenmlbfanmonstermontrealestatefarmequipmentrentinoa-adigemonza" + - "-brianzaporizhzhiamonza-e-della-brianzapposhintomikasaharamonzab" + - "rianzaptokyotangotsukitahatakamatsukawamonzaebrianzaramonzaedell" + - "abrianzamoonscalemoparachutingmordoviamoriyamatsumotofukemoriyos" + - "himinamiawajikis-into-animeiwamarshallstatebankfhappoumormonmout" + - "hagakhanamigawamoroyamatsunomortgagemoscowindmillmoseushistorymo" + - "sjoenmoskeneshinyoshitomiokamogawamosshiojirishirifujiedamosvikn" + - "x-serveronamsskoganeis-a-rockstarachowicemoteginowaniihamatamaka" + - "wajimansionshioyanaizumoviemovimientolgamovistargardmtpchoyodoba" + - "shichikashukujitawaramtranbymuenstermuginozawaonsenmuikamisunaga" + - "wamukodairamulhouserveblogspotrentinoaadigemunakatanemuncienciam" + - "uosattemuphonefosshirahamatonbetsurnadalmurmanskolobrzegersundmu" + - "rotorcraftrentinoalto-adigemusashimurayamatsusakahoginankokubunj" + - "is-into-carshimotsukemusashinoharamuseetrentinoaltoadigemuseumve" + - "renigingmusicarbonia-iglesias-carboniaiglesiascarboniamutsuzawam" + - "y-vigorlicemy-wanggouvicenzamyactivedirectorymyasustor-elvdalmyc" + - "dn77-securecifedexhibitionmyddnskingmydissentrentinos-tirolmydro" + - "boehringerikemydshirakofuefukihaborokunohealthcareershiranukanag" + - "awamyeffectrentinostirolmyfirewallonieruchomoscienceandindustryn" + - "myfritzmyftpaccesshiraois-into-cartoonshimotsumamyhome-serversai" + - "lleshiraokananiimihoboleslawiechristiansburgrondarmykolaivaporcl" + - "oudmymailermymediapchristmasakinderoymyokohamamatsudamypephotogr" + - "aphysiomypetshiratakahagitlabormyphotoshibalestrandabergamoareke" + - "ymachinewhampshirebungoonombresciamypsxn--32vp30hagebostadmysecu" + - "ritycamerakermyshopblockshishikuis-into-gamessinazawamytis-a-boo" + - "kkeeperugiamytuleapiagetmyipictetrentinosud-tirolmyvnchromedicin" + - "akamagayachtsantabarbaramywireitrentinosudtirolpinkomaganepionee" + - "rpippulawypiszpittsburghofauskedsmokorsetagayasells-for-unzenpiw" + - "atepixolinopizzapkomakiyosunndalplanetariuminnesotaketakatsukis-" + - "certifieducatorahimeshimamateramobilyplantationplantshitaramapla" + - "tformshangrilanshizukuishimofusaitamatsukuris-lostre-toteneis-a-" + - "republicancerresearchaeologicaliforniaplaystationplazaplchungnam" + - "dalseidfjordyndns-mailucaniaplumbingoplurinacionalpmnpodzonepohl" + - "poivronpokerpokrovskomatsushimasfjordenpoliticarrierpolitiendapo" + - "lkowicepoltavalle-aostarostwodzislawindowshizuokanazawapomorzesz" + - "owinnershoujis-not-certifiedunetbankhakassiapordenonepornporsang" + - "erporsanguidell-ogliastraderporsgrunnanyokoshibahikariwanumatake" + - "tomisatoshimapoznanpraxis-a-bruinsfanprdpreservationpresidioprgm" + - "rprimeldalprincipeprivatizehealthinsuranceprochowiceproductionsh" + - "owaprofesionalprogressivegaskvolloabathsbchurchaseljeepsongdalen" + - "viknaharimalopolskanlandyndns-office-on-the-webcampinashikiminoh" + - "kurapromombetsurfbsbxn--12co0c3b4evalleaostaticsavonarusawaprope" + - "rtyprotectionprotonetrentinosued-tirolprudentialpruszkowioshowti" + - "memergencyahabahcavuotnagarahkkeravjuegoshikikonaikawachinaganoh" + - "aramcoachampionshiphoptobishimagentositecnologiaprzeworskogptplu" + - "sgardenpupictureshisognepvhaibarakitahiroshimaoris-a-lawyerpvtre" + - "ntinosuedtirolpwciprianiigataishinomakindlegnicafederationpzqldq" + - "ponqslgbtrentoyonezawaquicksyteshriramlidlugolekafjordquipelemen" + - "tsienarutomobellevuelosangelesjabbottrevisohughesigdalqvcirclego" + - "doesntexisteingeekashiharasrtroandinosaurepaircraftrogstadsrvare" + - "servecounterstrikestoragestordalstoregontrailroadstorfjordstorjd" + - "evcloudfrontdoorstpetersburgstreamsterdamnserverbaniastudiostudy" + - "ndns-at-homedepotenzamamidsundstuff-4-salestufftoread-booksnesir" + - "dalstuttgartromsakakinokiasusakis-savedsusonosuzakaniepcesuzukan" + - "makiwiensuzukis-slickharkivalleeaosteigensvalbardunloppacificirc" + - "ustomersveiosvelvikomvuxn--2scrj9choshibuyachiyodavvenjargaulard" + - "alowiczest-le-patronsvizzerasvn-reposjcbnlswedenswidnicartoonart" + - "decologiaswiebodzindianapolis-a-bloggerswiftcoverswinoujsciencea" + - "ndhistoryswisshikis-uberleetrentino-sud-tirolsynology-dslingtush" + - "uissier-justicetuvalle-daostatic-accessnoasaitotaltuxfamilytwmai" + - "lvenneslaskerrylogisticsokaneyamazoevestfoldvestnesokndalvestre-" + - "slidrepbodynathomebuiltrusteevestre-totennishiawakuravestvagoyve" + - "velstadvibo-valentiavibovalentiavideovillasnesoddenmarkhangelskj" + - "akdnepropetrovskiervaapsteiermarkongsvingervinnicasacamdvrcampin" + - "agrandebugattipschlesischesolarssonvinnytsiavipsinaappiemontevir" + - "giniavirtualvirtueeldomeindianmarketingvirtuelvisakegawaviterbok" + - "nowsitallvivoldavixn--3bst00misakis-foundationvlaanderenvladikav" + - "kazimierz-dolnyvladimirvlogoipilotshisuifuelblagdenesnaaseraling" + - "enkainanaejrietisalatinabenonichryslervolkswagentsolognevologdan" + - "skoninjambylvolvolkenkundenvolyngdalvossevangenvotevotingvotoyon" + - "owiwatsukiyonoticiaskimitsubatamibudejjuedischesapeakebayernrtrv" + - "arggatromsojamisonwloclawekonsulatrobeepilepsydneywmflabsolundbe" + - "ckommuneworldworse-thandawowitdkonskowolayangrouphilipsynology-d" + - "iskstationwpdevcloudwritesthisblogsytewroclawithgoogleapisa-hock" + - "eynutsiracusakatakinouewtcmisasaguris-gonewtfbx-ostrowwlkpmgunma" + - "nxn--1ck2e1barclaycards3-fips-us-gov-west-1wuozuwwwithyoutubenev" + - "entoeidsvollwzmiuwajimaxn--42c2d9axn--45br5cylxn--45brj9citadeli" + - "veryxn--45q11citichernigovernmentoyotaris-a-cubicle-slavellinota" + - "irestaurantoyotomiyazakis-a-democratoyotsukaidoxn--4gbriminingxn" + - "--4it168dxn--4it797kooris-a-soxfanxn--4pvxs4allxn--54b7fta0ccivi" + - "laviationxn--55qw42gxn--55qx5dxn--5js045dxn--5rtp49civilisationx" + - "n--5rtq34kopervikhersonxn--5su34j936bgsgxn--5tzm5gxn--6btw5axn--" + - "6frz82gxn--6orx2rxn--6qq986b3xlxn--7t0a264civilizationxn--80adxh" + - "ksolutionsilkomforbargainstitutelemarkarateu-1xn--80ao21axn--80a" + - "qecdr1axn--80asehdbarsyonlinewhollandiscoveryonaguniversityoriik" + - "aratsuginamikatagamilitaryoshiokaracoldwarmiastageu-2xn--80aswgx" + - "n--80audnedalnxn--8ltr62koryokamikawanehonbetsurutaharaxn--8pvr4" + - "uxn--8y0a063axn--90a3academiamicaaarborteaches-yogasawaracingxn-" + - "-90aeroportalaheadjudaicable-modemocraciaxn--90aishobarakawagoex" + - "n--90azhytomyrxn--9dbhblg6dietcimdbashkiriauthordalandeportenrig" + - "htathomeftpalmaseratibigawastronomy-gatewayokosukanzakiyosatokig" + - "awagrocerybnikahokutobamagazineat-url-o-g-i-natuurwetenschappena" + - "umburgjerdrumeteorappalermomahachijolstereportarumizusawaetnagah" + - "amaroygardendoftheinternetflixilovecollegefantasyleaguernseybolt" + - "arnobrzegyptianaturhistorisches3-ap-northeast-2ixboxenapponazure" + - "-mobile12hpaleobirabogadocscbgdyniabruzzoologicalvinklein-addram" + - "menuernberggfarmerseine164xn--9dbq2axn--9et52uxn--9krt00axn--and" + - "y-iraxn--aroport-byandexn--3ds443gxn--asky-iraxn--aurskog-hland-" + - "jnbasilicataniautomotiveconomiasakuchinotsuchiurakawalmartataran" + - "toyakokonoehimejibmdgcahcesuolocalhostrodawaraumalborkdalaziocea" + - "nographics3-eu-west-1xn--avery-yuasakuhokkaidoomdnsiskinkyotobet" + - "sumidatlanticivilwarmanagementoyouraxn--b-5gaxn--b4w605ferdxn--b" + - "ck1b9a5dre4claimsantacruzsantafedjejuifminamiizukamishihoronobea" + - "uxartsandcraftsantamariakexn--bdddj-mrabdxn--bearalvhki-y4axn--b" + - "erlevg-jxaxn--bhcavuotna-s4axn--bhccavuotna-k7axn--bidr-5nachika" + - "tsuuraxn--bievt-0qa2xn--bjarky-fyaotsurreyxn--bjddar-ptamayufuet" + - "tertdasnetzxn--blt-elabourxn--bmlo-graingerxn--bod-2natalxn--brn" + - "ny-wuacademy-firewall-gatewayxn--brnnysund-m8accident-investigat" + - "ion-aptibleaseating-organicbcn-north-1xn--brum-voagatrysiljanxn-" + - "-btsfjord-9zaxn--c1avgxn--c2br7gxn--c3s14misawaxn--cck2b3basketb" + - "allyngenhktatsunoddautoscanadaejeonbukarasjohkamikoaniikappueblo" + - "ckbustermezgoraugustowadaegubambleclerc66xn--cg4bkis-very-badajo" + - "zxn--ciqpnxn--clchc0ea0b2g2a9gcdn77-sslattumisconfusedxn--comuni" + - "caes-v6a2oxn--correios-e-telecomunicaes-ghc29axn--czr694batodayu" + - "kindustriaveroykeniwaizumiotsukumiyamazonawsadodgemologicallilly" + - "ombolzanord-frontiereviewskrakowebhostingjerstadotsuruokakegawau" + - "kraanghkepnogifts3-ap-southeast-2xn--czrs0tulanxesslupskommunalf" + - "orbundxn--czru2dxn--czrw28batsfjordishakotanhlfanhs3-us-gov-west" + - "-1xn--d1acj3bauhausposts-and-telecommunicationsncfdisrechtranaka" + - "muratajimidoriopretogoldpoint2thisamitsukeu-3xn--d1alfaromeoxn--" + - "d1atuneslzxn--d5qv7z876clanbibaidarmeniaxn--davvenjrga-y4axn--dj" + - "rs72d6uyxn--djty4kosaigawaxn--dnna-grajewolterskluwerxn--drbak-w" + - "uaxn--dyry-iraxn--e1a4cldmailuccapetownnews-stagingrongaxn--eckv" + - "dtc9dxn--efvn9somaxn--efvy88hair-surveillancexn--ehqz56nxn--elqq" + - "16hakatanortonxn--estv75gxn--eveni-0qa01gaxn--f6qx53axn--fct429k" + - "osakaerodromegallupinbarreauctionflfanfshostrowiecaseihichisobet" + - "suldalimoliserniaustraliaisondriobranconagawalesundemoneyokozebi" + - "nordreisa-geekaragandamusementashkentatamotors3-ap-southeast-1pa" + - "sswordd-dnshome-webservercellikes-piedmonticellocus-4xn--fhbeiar" + - "nxn--finny-yuaxn--fiq228c5hsomnarviikamitondabayashiogamagorizia" + - "xn--fiq64bbcasertairavennagatorockartuzyukuhashimoichinosekigaha" + - "ravocatanzarowebredirectmetacentrumetlifeinsurancempresashibetsu" + - "kuiitatebayashiibajddarchitecturealtydalipayomitanoceanographiqu" + - "emrevistanbulminamidaitomandalimanowarudaurskog-holandroverhalla" + - "-speziajudygarlanddnss3-ap-south-1kappchizippodhaleangaviikadena" + - "amesjevuemielno-ip6xn--fiqs8sooxn--fiqz9sopotritonxn--fjord-lrax" + - "n--fjq720axn--fl-ziaxn--flor-jraxn--flw351exn--fpcrj9c3dxn--frde" + - "-grandrapidsor-odalxn--frna-woaraisaijosoyrorosor-varangerxn--fr" + - "ya-hraxn--fzc2c9e2clickashiwaraxn--fzys8d69uvgmailxn--g2xx48clin" + - "ichernihivguccieszynissandnessjoenissayokkaichiropracticheltenha" + - "m-radio-opencraftrainingripescaravantaaxn--gckr3f0fbxosaxoxn--ge" + - "crj9cliniquenoharaxn--ggaviika-8ya47hakodatexn--gildeskl-g0axn--" + - "givuotna-8yasakaiminatoyookannamilanotteroyxn--gjvik-wuaxn--gk3a" + - "t1exn--gls-elacaixaxn--gmq050is-very-evillagexn--gmqw5axn--h-2fa" + - "ilxn--h1aeghakonexn--h2breg3evenesorfoldxn--h2brj9c8clintonoshoe" + - "santoandreamhostersanukis-a-designerimarnardalucernexn--h3cuzk1d" + - "igitalxn--hbmer-xqaxn--hcesuolo-7ya35bbtattoolsztynsettlers3-us-" + - "west-1xn--hery-iraxn--hgebostad-g3axn--hmmrfeasta-s4accident-pre" + - "vention-webhopenairbusantiquest-a-la-maisondre-landroidvagsoyeri" + - "cssonyoursidealerimo-i-ranadexeterxn--hnefoss-q1axn--hobl-iraxn-" + - "-holtlen-hxaxn--hpmir-xqaxn--hxt814exn--hyanger-q1axn--hylandet-" + - "54axn--i1b6b1a6a2exn--imr513nxn--indery-fyasugivingxn--io0a7is-v" + - "ery-goodhandsonxn--j1aefedorapeopleikangerxn--j1amhakubahccavuot" + - "nagareyamakeupowiathletajimabaridagawalbrzycharternidxn--j6w193g" + - "xn--jlq61u9w7bbvacationswatch-and-clockerhcloudns3-us-west-2xn--" + - "jlster-byasuokanraxn--jrpeland-54axn--jvr189mishimasudaxn--k7yn9" + - "5exn--karmy-yuaxn--kbrq7oxn--kcrx77d1x4axn--kfjord-iuaxn--klbu-w" + - "oaxn--klt787dxn--kltp7dxn--kltx9axn--klty5xn--3e0b707exn--koluok" + - "ta-7ya57hakuis-a-liberalxn--kprw13dxn--kpry57dxn--kpu716fedorapr" + - "ojectransportexn--kput3is-very-nicexn--krager-gyatomitamamuraxn-" + - "-kranghke-b0axn--krdsherad-m8axn--krehamn-dxaxn--krjohka-hwab49j" + - "dfastlylbarcelonagasakikuchikuseikarugamvikarasjokarasuyamarugam" + - "e-hostrolekamiminers3-external-1xn--ksnes-uuaxn--kvfjord-nxaxn--" + - "kvitsy-fyatsukanumazuryxn--kvnangen-k0axn--l-1fairwindsorocabals" + - "fjordxn--l1accentureklamborghinikis-very-sweetpepperxn--laheadju" + - "-7yatsushiroxn--langevg-jxaxn--lcvr32dxn--ldingen-q1axn--leagavi" + - "ika-52bentleyurihonjournalistgoryusuharavoues3-eu-west-2xn--lesu" + - "nd-huaxn--lgbbat1ad8jelenia-goraxn--lgrd-poacctunkongsbergxn--lh" + - "ppi-xqaxn--linds-pramericanarturystykanoyaltakasakiyokawaraxn--l" + - "ns-qlapyatigorskoseis-a-studentalxn--loabt-0qaxn--lrdal-sraxn--l" + - "renskog-54axn--lt-liaclothingdustkagoshimalselvendrellukowhaling" + - "rossetouchijiwadegreexn--lten-granexn--lury-iraxn--m3ch0j3axn--m" + - "ely-iraxn--merker-kuaxn--mgb2ddesorreisahayakawakamiichikawamisa" + - "toursimple-urlxn--mgb9awbfeiraquarellebesbyglandynulvikasuyanaga" + - "waxn--mgba3a3ejtuscanyxn--mgba3a4f16axn--mgba3a4franamizuholding" + - "smilevangerxn--mgba7c0bbn0axn--mgbaakc7dvfermochizukirkenesbscho" + - "koladenxn--mgbaam7a8hakusandiegooglecodespotrentino-alto-adigexn" + - "--mgbab2bdxn--mgbai9a5eva00beppublishproxyzjampagefrontappalmspr" + - "ingsakerxn--mgbai9azgqp6jeonnamerikawauexn--mgbayh7gpalacexn--mg" + - "bb9fbpobanazawaxn--mgbbh1a71exn--mgbc0a9azcgxn--mgbca7dzdoxn--mg" + - "berp4a5d4a87gxn--mgberp4a5d4arxn--mgbgu82axn--mgbi4ecexposedxn--" + - "mgbpl2fhskoleirfjordxn--mgbqly7c0a67fbcngroundhandlingroznyxn--m" + - "gbqly7cvafranziskanerdpolicexn--mgbt3dhdxn--mgbtf8flatangerxn--m" + - "gbtx2beskidyn-o-saurlandes3-website-ap-northeast-1xn--mgbx4cd0ab" + - "bvieeexn--mix082ferraraxn--mix891ferrarittoguraxn--mjndalen-64ax" + - "n--mk0axindigenaklodzkochikushinonsenergyxn--mk1bu44cnsaobernard" + - "ownloadyndns-picsaogoncartierxn--mkru45is-with-thebandovre-eiker" + - "xn--mlatvuopmi-s4axn--mli-tlaquilanciaxn--mlselv-iuaxn--moreke-j" + - "uaxn--mori-qsakuragawaxn--mosjen-eyawaraxn--mot-tlarvikosherbroo" + - "kegawaxn--mre-og-romsdal-qqbestbuyshouses3-website-ap-southeast-" + - "1xn--msy-ula0haldenxn--mtta-vrjjat-k7afamilycompanycntoystre-sli" + - "drettozawaxn--muost-0qaxn--mxtq1missilezajsklabudhabikinokawabar" + - "thaebaruminamiuonumassa-carrara-massacarraramassabusinessebykleg" + - "allocalhistoryggeelvinckaufenxn--ngbc5azdxn--ngbe9e0axn--ngbrxn-" + - "-3hcrj9cistrondheimmobilienxn--nit225koshimizumakizunokunimimata" + - "kasugais-a-teacherkassymantechnologyxn--nmesjevuemie-tcbaltimore" + - "-og-romsdalpha-myqnapcloudaccesscambridgestoneuesortlandxn--nnx3" + - "88axn--nodessakuraisleofmanchesterxn--nqv7fs00emaxn--nry-yla5gxn" + - "--ntso0iqx3axn--ntsq17gxn--nttery-byaeserveexchangexn--nvuotna-h" + - "waxn--nyqy26axn--o1achattanooganordkappimientakazakis-leetnedalx" + - "n--o3cw4halsaintlouis-a-anarchistoireggiocalabriaxn--o3cyx2axn--" + - "od0algxn--od0aq3betainaboxfusejnynysagaeroclubmedecincinnationwi" + - "dealstahaugesunderseaportsinfolldalabamagasakishimabaraogakibich" + - "uomutashinaindustriesteambulanceu-4xn--ogbpf8flekkefjordxn--oppe" + - "grd-ixaxn--ostery-fyawatahamaxn--osyro-wuaxn--p1acferreroticampo" + - "bassociatestinguovdageaidnuslivinghistoryxn--p1aissmarterthanyou" + - "xn--pbt977coguchikuzenxn--pgbs0dhlxn--porsgu-sta26fetsundynv6xn-" + - "-pssu33lxn--pssy2uxn--q9jyb4collectionxn--qcka1pmckinseyxn--qqqt" + - "11misugitokuyamatsumaebashikshacknetrentino-suedtirolxn--qxamune" + - "ustarhubsoruminternationalfirearmshintokushimaxn--rady-iraxn--rd" + - "al-poaxn--rde-ulavagiskexn--rdy-0nabariwchonanbuildingroks-thisa" + - "yamanobeokakudamatsuexn--rennesy-v1axn--rhkkervju-01aflakstadaok" + - "agakicks-assedicolognextdirectozsdeloittemp-dnsaotomelhusdecorat" + - "iveartsapodlasiellaktyubinskiptveterinairealtorlandyndns-remotew" + - "dyndns-serverdaluroyxn--rholt-mragowoodsideltaitogliattiresouthc" + - "arolinarvikomonoxn--rhqv96gxn--rht27zxn--rht3dxn--rht61exn--risa" + - "-5nativeamericanantiquesouthwestfalenxn--risr-iraxn--rland-uuaxn" + - "--rlingen-mxaxn--rmskog-byaxn--rny31hammarfeastafricapebretonami" + - "crosoftbankautokeinowruzhgorodeoxn--rovu88bhzcasinorddalindaskoy" + - "abearalvahkijobserverisignieznogataijinfinitintuitaxihuanikkoebe" + - "nhavnikolaevents3-website-ap-southeast-2xn--rros-granvindafjordx" + - "n--rskog-uuaxn--rst-0naturalhistorymuseumcenterxn--rsta-francais" + - "eharaxn--rvc1e0am3exn--ryken-vuaxn--ryrvik-byaxn--s-1faithruhere" + - "dumbrellajollamericanexpressexyxn--s9brj9colonialwilliamsburgrpa" + - "rocherkasyno-dsapporoxn--sandnessjen-ogbizxn--sandy-yuaxn--seral" + - "-lraxn--ses554gxn--sgne-gratangenxn--skierv-utazassnasabaerobati" + - "cketsowaxn--skjervy-v1axn--skjk-soaxn--sknit-yqaxn--sknland-fxax" + - "n--slat-5naturalsciencesnaturellespjelkavikomorotsukamiokamikita" + - "yamatsuris-a-socialistcgrouphdxn--slt-elabcgxn--smla-hraxn--smna" + - "-gratis-a-bulls-fanxn--snase-nraxn--sndre-land-0cbremangerxn--sn" + - "es-poaxn--snsa-roaxn--sr-aurdal-l8axn--sr-fron-q1axn--sr-odal-q1" + - "axn--sr-varanger-ggbieigersundivtasvuodnakaniikawatanaguraxauste" + - "vollavangenaval-d-aosta-valleyokotebinagisoccertificationavigati" + - "onavoibestadds3-ca-central-1xn--srfold-byaxn--srreisa-q1axn--sru" + - "m-grazxn--stfold-9xaxn--stjrdal-s1axn--stjrdalshalsen-sqbielawal" + - "terxn--stre-toten-zcbspreadbettingxn--t60b56axn--tckweatherchann" + - "elxn--tiq49xqyjetztrentino-sudtirolxn--tjme-hraxn--tn0agrinet-fr" + - "eakspydebergxn--tnsberg-q1axn--tor131oxn--trany-yuaxn--trgstad-r" + - "1axn--trna-woaxn--troms-zuaxn--tysvr-vraxn--uc0atvaroyxn--uc0ay4" + - "axn--uist22hamurakamigoris-a-libertarianxn--uisz3gxn--unjrga-rta" + - "obaomoriguchiharagusartsrlxn--unup4yxn--uuwu58axn--vads-jraxn--v" + - "ard-jraxn--vegrshei-c0axn--vermgensberater-ctbiellaakesvuemielec" + - "ceverbankareliancevje-og-hornnes3-website-eu-west-1xn--vermgensb" + - "eratung-pwbieszczadygeyachimataikikugawarszawashingtondclkariyam" + - "elbournexn--vestvgy-ixa6oxn--vg-yiabkhaziaxn--vgan-qoaxn--vgsy-q" + - "oa0jevnakershuscultureggioemiliaromagnamsosnowiechoseiroumuenche" + - "nxn--vgu402coloradoplateaudioxn--vhquvbarrell-of-knowledgeometre" + - "-experts-comptables3-us-east-2xn--vler-qoaxn--vre-eiker-k8axn--v" + - "rggt-xqadxn--vry-yla5gxn--vuq861bievatmallorcadaques3-website-sa" + - "-east-1xn--w4r85el8fhu5dnraxn--w4rs40lxn--wcvs22dxn--wgbh1columb" + - "usheyxn--wgbl6axn--xhq521bifukagawashtenawdev-myqnapcloudapplebt" + - "imnetzlgjovikarlsoyusuisserveftpanamatta-varjjatjeldsundivttasvu" + - "otnakanojohanamakinoharaxn--xkc2al3hye2axn--xkc2dl3a5ee0hangglid" + - "ingxn--y9a3aquariumitourismolangevagrarchaeologyeongbukmpspbaref" + - "ootballfinanzgorzeleccoffeedbackplaneapplinziiyamanouchikuhokury" + - "ugasakitchenayorovigovtateshinanomachimkentateyamaustinnavuotnar" + - "ashinobninsk12xn--yer-znaturbruksgymnxn--yfro4i67oxn--ygarden-p1" + - "axn--ygbi2ammxn--3oq18vl8pn36axn--ystre-slidre-ujbihorologyuucon" + - "nectjmaxxxfinityuzawaxn--zbx025dxn--zf0ao64axn--zf0avxn--3pxu8ko" + - "nyvelolxn--zfr164bikedagestangeorgeorgiaxperiaxz" +const text = "9guacuiababia-goracleaningroks-theatreebinagisobetsumidatlantica" + + "sertairanzanquannefrankfurtashkentatamotors3-ap-northeast-20001w" + + "wwebredirectmemsettsupport3l3p0rtargets-itargivestbytomaritimeke" + + "eping12038biomutashinaindustriabirdartcenterprisesakimobetsuitai" + + "nairforceoppdalimoliserniabirkenesoddtangenovaraholtalenikkoeben" + + "havnikolaevents3-website-eu-west-1birthplacebitballooningjovikar" + + "iyaltakasakiyosatokigawabjarkoyukuhashimoichinosekigaharabjerkre" + + "imbarclaycards3-eu-west-2bjugnieznord-aurdalpha-myqnapcloud66bla" + + "ckfridayurihonjournalisteinkjerusalembroideryusuharablancomedica" + + "ltanissettaipeiheijindustriesteamfamberkeleyusuisserveirchattano" + + "oganordkappanamatta-varjjatjmaxxxboxenapponazure-mobilebloomberg" + + "bauernuorochesterbloxcms3-website-sa-east-1bluedancebmoattachmen" + + "ts3-website-us-east-1bms3-website-us-west-1bmweddingladefensells" + + "-for-less3-website-us-west-2bnpparibaselburglassassinationalheri" + + "tagebnrwedeployuu2-localhostrowwlkpmgleezebomloabathsbcheltenham" + + "-radio-openairbusantiquest-a-la-maisondre-landroidivttasvuotnaka" + + "nojohanamakinoharabondiyuzawabonnikonanporovnobookingliwicebooml" + + "adbrokes5yboschaefflerdalvdalaskanittedallasallebesbyglandroverh" + + "alla-speziabostikarlsoybostonakijinsekikogentinglobalashovhachin" + + "ohedmarkarmoybotanicalgardeninohekinannestadnpanasonichernigover" + + "nmentjomemorialindaskvollindesnesakyotanabellunombresciabotanicg" + + "ardeninomiyakonojorpelandrangedalinkyard-cloudeitybotanybounceme" + + "rckmsdnipropetrovskjervoyagebounty-fullensakerrypropertiesalange" + + "nirasakinfinitintuitjxfinityboutiquebechernihivgubarclays3-eu-we" + + "st-3utilitiesquare7bozen-suedtirolkuszczytnord-frontierbplacedek" + + "agaminord-odalwaysdatabaseballangenoamishirasatochigiessensiosit" + + "elekommunikationishiazainuyamashinatsukigatakasagotembaixadabran" + + "dywinevalleybrasiliabrindisibenikebristoloseyouripirangap-northe" + + "ast-3britishcolumbialowiezachpomorskienishigovtkmaxxjavald-aosta" + + "plesalondonetskarpaczeladzlgloboavistaprintelligencebroadcastlef" + + "rakkestadray-dnstracebroadwaybroke-itksatxn--0trq7p7nnishiharabr" + + "okerbronnoysundrayddnsfreebox-osascoli-picenordlandraydnsupdater" + + "brothermesaverdealstahaugesunderseaportsinfolldalivornobrowsersa" + + "fetymarketsaltdalomzaporizhzhegurinvestmentsaludrivefsnillfjordr" + + "obaknoluoktagajobojinzais-a-candidatebrumunddalondrinaplesalvado" + + "rdalibabalsan-suedtirollagdenesnaaseralingenkainanaejrietisalati" + + "nabenonichernivtsiciliabrunelasticbeanstalkaruizawabrusselsalzbu" + + "rglogowegroweibolognagatorockartuzybruxellesamegawabryanskleppga" + + "fanpachigasakievennodesaarlandrudunsamnangerbrynewjerseybuskerud" + + "inewportlligatmparaglidingloppenzaolbia-tempio-olbiatempioolbial" + + "ystokkembuchikumagayagawakuyabukihokumakogenglandupontariodejane" + + "irodoybuzentsujiiebuzzparisor-fronishiizunazukis-a-catererbweirb" + + "zhitomirumalatvuopmicrolightinglugmbhartiffanycolumbusheycommuni" + + "tysvardoharuovatoystre-slidrettozawacomobaracomparemarkerryhotel" + + "sanokashiharacompute-1computerhistoryofscience-fictioncomsecurit" + + "ytacticsantabarbaracondoshichinohealth-carereformitakeharaconfer" + + "enceconstructionconsuladollsantacruzsantafedjejuifminamidaitoman" + + "dalucerneconsultanthropologyconsultingrossetouchihayaakasakawaha" + + "racontactozsdeloittemp-dnsantamariakecontagematsubaracontemporar" + + "yarteducationalchikugojomedio-campidano-mediocampidanomediocontr" + + "actorskenconventureshinodearthdfcbankashiwaracookingchannelsdvrd" + + "nsdojoetsuwanouchikujogaszkolahppiacenzagancoolukowfashioncooper" + + "ativano-frankivskoleikangercopenhagencyclopedichitachinakagawatc" + + "handclockarumaifarmsteadurhamburgmodellingmxn--11b4c3dyndns-at-w" + + "orkinggrouparliamentoyosatoyonakagyokutoyokawacorsicagliaribeira" + + "okinawashirosatochiokinoshimaizuruhrcorvettemasekashiwazakiyosem" + + "itecosenzakopanerairguardiannakadomarinebraskaunjargalsacertmgre" + + "tachikawakeisenbahncosidnsfor-better-thanawatchesantoandreamhost" + + "ersanukis-a-cubicle-slavellinotairestaurantrani-andria-barletta-" + + "trani-andriacostumedizinhistorischesaobernardownloadyndns-remote" + + "wdyndns-serverdaluroycouchpotatofriesaogoncartoonartdecologiacou" + + "ncilutskasukabedzin-the-bandaioiraseeklogesurancechirealmpmncoup" + + "onsaotomeloyalistjordalshalsencoursesapodlasiellaktyubinskiptvet" + + "erinairealtorlandyndns-webhopencraftraniandriabarlettatraniandri" + + "acq-acranbrookuwanalyticsapporocreditcardyndns-wikiracreditunion" + + "cremonashgabadaddjaguarqhachiojiyahoooshikamaishimodatecrewhalin" + + "groundhandlingroznycricketrzyncrimeast-kazakhstanangercrotonexus" + + "-3crowniparsardegnaroycrsvpartis-a-democratranoycruisesardiniacr" + + "yptonomichigangwoncuisinellair-traffic-controlleyculturalcentern" + + "opilawawhoswhokksundyndns-workisboringrpartsarluxembourgruecuneo" + + "cupcakecuritibaghdadyndns1cxn--12c1fe0bradescorporationcyberleva" + + "gangaviikanonjis-a-designercymrussiacyonabaruminamiechizencyouth" + + "eworkpccwiiheyakageferrarittoguraferreroticanonoichinomiyakefets" + + "undynservebbsarufutsunomiyawakasaikaitakoelnfguitarsassaris-a-do" + + "ctorayfhvalerfidonnakanotoddenfieldynuconnectransportefigueresin" + + "stagingujoinvillevangerfilateliafilegearfilminamiiselectrapaniiz" + + "afinalvivanovodkamisatokashikiwakunigamiharulminamiizukamishihor" + + "onobeauxartsandcraftsaudafinancefineartsauheradynv6finlandynvpnp" + + "lus-4finnoyfirebaseappartyfirenzefirestonefirmdaleirvikasumigaur" + + "awa-mazowszextraspacekitagatajirissagamiharafishingolffansavanna" + + "hgafitjarfitnessettlementravelchannelfjalerflesbergulenflickrage" + + "rotikakamigaharaflightsaves-the-whalessandria-trani-barletta-and" + + "riatranibarlettaandriaflirflogintohmalvikasuyamelbournefloraflor" + + "encefloridavvenjargaulardalfloripaderbornfloristanohatakahamamur" + + "ogawaflorogersavonarusawaflowersaxofltravelersinsuranceflynnhost" + + "ing-clusterflynnhubargainstitutelemarkarasjohkamikoaniikappueblo" + + "ckbustermezgorzeleccoffeedbackplaneapplegodoesntexisteingeekaras" + + "jokarasuyamarugame-hostrolekamiminers3-external-1fndyroyfor-ourf" + + "or-someeresistancefor-theaterforexrothachirogatakamoriokalmykiaf" + + "orgotdnsbschokokekschokoladenforsaleitungsenforsandasuololfortal" + + "fortmissoulancashireggio-calabriafortworthadanorthwesternmutualf" + + "orumzwildlifedorainfracloudcontrolappasadenaritakurashikis-a-fin" + + "ancialadvisor-aurdalfosnescholarshipschoolfotarivnefoxfordeatnur" + + "embergunmapartmentschulefozorafredrikstadtvschwarzgwangjuniperfr" + + "eeddnsgeekgalaxyfreedesktoperauniteroizumizakirovogradoyfreemaso" + + "nryfreesitevadsoccertificationfreetlschweizfreiburguovdageaidnul" + + "vikaszubyfreightrdfreseniuscountryestateofdelawarezzoologyfribou" + + "rgushikamifuranorth-kazakhstanfriuli-v-giuliafriuli-ve-giuliafri" + + "uli-vegiuliafriuli-venezia-giuliafriuli-veneziagiuliafriuli-vgiu" + + "liafriuliv-giuliafriulive-giuliafriulivegiuliafriulivenezia-giul" + + "iafriuliveneziagiuliafriulivgiuliafrlfrogansciencecenterscienceh" + + "istoryfrognfrolandfrom-akrehamnfrom-alfrom-arfrom-azfrom-capebre" + + "tonamicrosoftbankatowicefrom-codynaliasdaburfrom-ctrentin-sued-t" + + "irolfrom-dchitosetogitsuldalorenskogrimstadyndns-blogdnsampagesp" + + "eedmobilizerofrom-debianfrom-flanderscientistockholmestrandfrom-" + + "gausdalfrom-hichisochildrensgardenfrom-iafrom-idfrom-ilfrom-inch" + + "eonfrom-kscjohnsonfrom-kyowariasahikawafrom-lancasterfrom-mangon" + + "ohejis-a-geekatsushikabeeldengeluidfrom-mdfrom-meethnologyfrom-m" + + "ifunefrom-mnfrom-modalenfrom-mscotlandfrom-mtnfrom-nchocolatelev" + + "isionishikatsuragit-repostre-totendofinternet-dnsamsclubindalote" + + "nkawafrom-ndfrom-nefrom-nh-serveblogsitexashorokanaiefrom-njawor" + + "znotogawafrom-nminamimakis-a-greenfrom-nv-infoodnetworkshoppingw" + + "iddlewismillerfrom-nyfrom-ohkurafrom-oketohnoshooguyfrom-orfrom-" + + "padovaksdalfrom-pratohobby-sitextileksvikatsuyamarylandfrom-ris-" + + "a-gurunzenfrom-schoenbrunnfrom-sdfrom-tnfrom-txn--12co0c3b4evall" + + "eaostaticscrapper-sitefrom-utazuerichardlikescandynamic-dnscrapp" + + "ingxn--1ck2e1barreauctionavigationavoiiyamanouchikuhokuryugasaki" + + "tchenavuotnarashinoceanographics3-fips-us-gov-west-1from-val-dao" + + "stavalleyfrom-vtrentin-suedtirolfrom-wafrom-wielunnerfrom-wvalle" + + "d-aostatoilfrom-wyfrosinonefrostalowa-wolawafroyahikobeardubaidu" + + "ckdnscrysechofunatoriginsurecreationishikawazukamitsuefstavernfu" + + "jiiderafujikawaguchikonefujiminokamoenairtelecitychyattorneyagaw" + + "akkanaibetsubamericanfamilydscloudapplinzis-a-hard-workerfujinom" + + "iyadavvesiidattowebcampinashikiminohosteroyrvikingfujiokayamangy" + + "shlakasamatsudontexistmein-iservebeerfujisatoshonairtrafficplexu" + + "s-1fujisawafujishiroishidakabiratoridedyn-ip24fujitsurugashimani" + + "wakuratefujixeroxn--1ctwolominamatakkokaminoyamaxunusualpersonfu" + + "jiyoshidazaifudaigokaseljordfukayabeatserveminecraftrentino-a-ad" + + "igefukuchiyamadafukudominichonanbuildingripescaravantaafukuis-a-" + + "hunterfukumitsubishigakiryuohtawaramotoineppuboliviajessheimperi" + + "afukuokazakisarazurecontainerdpolicefukuroishikarikaturindalfuku" + + "sakishiwadafukuyamagatakahashimamakisofukushimannore-og-uvdalfun" + + "abashiriuchinadafunagatakahatakaishimogosenfunahashikamiamakusat" + + "sumasendaisennangoodyearfundaciofuoiskujukuriyamansionservemp3fu" + + "osskoczowilliamhillfurnitureggio-emilia-romagnakatsugawafurubira" + + "furudonostiaarpassagenservep2passenger-associationfurukawais-a-k" + + "nightpointtokamachintaifun-dnsaliasiafusodegaurafussaikisosakita" + + "gawafutabayamaguchinomigawafutboldlygoingnowhere-for-morenakayam" + + "anxn--1lqs03nfuttsurugiminamiminowafuturecmservepicservequakefut" + + "urehostingfuturemailingfvgfylkesbiblackbaudcdn77-securecifederat" + + "ionfyresdalhannanmokuizumodenaklodzkobierzycehannosegawahanyuzen" + + "hapmirhareidsbergenharstadharvestcelebrationhasamarcheapaviancar" + + "rierhasaminami-alpssells-itrentino-aadigehashbanghasudahasura-ap" + + "pfizerhasvikautokeinow-dnservesarcasmatartanddesignhatogayaitaka" + + "nabeautysneservicesevastopolehatoyamazakitakamiizumisanofidelity" + + "hatsukaichikaiseis-a-liberalhattfjelldalhayashimamotobungotakada" + + "pliernewmexicoalhazuminobusellsyourhomegoodsevenassisicilyhbodoe" + + "s-itvedestrandhelsinkitakatakanezawahembygdsforbundhemnesewinbar" + + "rel-of-knowledgeologyokozeu-1hemsedalhepforgeherokussldheroyhgtv" + + "alledaostavangerhigashiagatsumagoianiahigashichichibunkyonanaosh" + + "imageandsoundandvisionhigashihiroshimanehigashiizumozakitakyushu" + + "aiahigashikagawahigashikagurasoedahigashikawakitaaikitamihamadah" + + "igashikurumeguromskoghigashimatsushimaritimodernhigashimatsuyama" + + "kitaakitadaitoigawahigashimurayamamotorcyclesharis-a-libertarian" + + "higashinarusembokukitamotosumy-gatewayhigashinehigashiomihachima" + + "naustdalhigashiosakasayamanakakogawahigashishirakawamatakaokalug" + + "anskypehigashisumiyoshikawaminamiaikitanakagusukumoduminamioguni" + + "comcastresindevicesharpgfoggiahigashitsunoshiroomurahigashiuraus" + + "ukitashiobarahigashiyamatokoriyamanashifteditchyouripharmaciensh" + + "awaiijimarnardalhigashiyodogawahigashiyoshinogaris-a-linux-usera" + + "nishiaritabashijonawatehiraizumisatokaizukamakurazakitaurayasuda" + + "hirakatashinagawahiranais-a-llamarriottrentino-alto-adigehirarah" + + "iratsukagawahirayaizuwakamatsubushikusakadogawahistorichouseshel" + + "laspeziahitachiomiyagildeskaliszhitachiotagooglecodespotaruis-a-" + + "musicianhitraeumtgeradellogliastradinghjartdalhjelmelandholeckoc" + + "hikushinonsenergyholidayhomeipharmacyshimojis-a-nascarfanhomelin" + + "kitoolsztynsettlershimokawahomelinuxn--1lqs71dhomeofficehomesecu" + + "ritymacaparecidahomesecuritypchoseiroumuenchenishimerahomesensem" + + "inehomeunixn--1qqw23ahondahoneywellbeingzonehongopocznorfolkebib" + + "lelhonjyoitakarazukameokameyamatotakadahornindalhorseoullensvang" + + "uardhorteneis-a-nurservegame-serverhospitalhoteleshimokitayamaho" + + "tmailhoyangerhoylandetroitskazohumanitieshimonitayanagithubuserc" + + "ontentrentino-altoadigehurdalhurumajis-a-painteractivegaskimitsu" + + "batamibudejjuedischesapeakebayernrtrentino-s-tirolhyllestadhyogo" + + "ris-a-patsfanhyugawarahyundaiwafunejfkharkivanylvenicejlchoyodob" + + "ashichikashukujitawaravennakamagayachtsamsungriwataraidyndns-fre" + + "eboxosloftranakasatsunairportland-4-salernoboribetsucksandnessjo" + + "enishinomiyashironojlljmphilipsynology-diskstationjnjcphilatelyj" + + "oyentrentinoa-adigejoyokaichibalatinogiftshinichinanjpmorganjpnc" + + "hristiansburgroks-thisayamanobeokakudamatsuejprshinjournalismail" + + "illehammerfeste-iphoenixn--2m4a15ejurkoshimizumakizunokunimimata" + + "kasugais-a-studentalkoshunantankhersonkosugekotohiradomainshinsh" + + "irokotourakouhokutamakis-a-teacherkassymantechnologykounosupplie" + + "shintokushimakouyamashikekouzushimashikis-a-techietis-a-personal" + + "trainerkozagawakozakis-a-therapistoiakozowindmillkpnkppspdnshint" + + "omikasaharakrasnodarkredstonekristiansandcatshinyoshitomiokamoga" + + "wakristiansundkrodsheradkrokstadelvaldaostarnbergkryminamisanrik" + + "ubetsurfastpanelblagrarchaeologyeongbuklugsmileasinglest-mon-blo" + + "gueurovisionionjukudoyamaceratabusebastopologyeonggiehtavuoatnag" + + "aivuotnagaokakyotambabydgoszczecinemadridvagsoygardendoftheinter" + + "netflixilovecollegefantasyleaguernseykumatorinokumejimasoykumena" + + "ntokonamegatakatoris-an-accountantshimonosekikawakunisakis-an-ac" + + "torkunitachiarailwaykunitomigusukumamotoyamashikokuchuokunneppug" + + "liakunstsammlungkunstunddesignkuokgroupictetrentinoaadigekurehab" + + "merkurgankurobelaudiblebtimnetzkurogiminamiashigarakuroisoftware" + + "ndalenugkuromatsunais-an-actresshimosuwalkis-a-photographerokuap" + + "phdkurotakikawasakis-an-anarchistoricalsocietykushirogawakustana" + + "is-an-artisteigenkusupplykutchanelkutnokuzumakis-an-engineeringk" + + "vafjordkvalsundkvamlidlugolekafjordkvanangenkvinesdalkvinnheradk" + + "viteseidskogkvitsoykwpspectruminamitanekzmissilezajsklabudhabiki" + + "nokawabarthadselfipatriamisugitokuyamatsumaebashikshacknetrentin" + + "oalto-adigemitourismolangevagrigentomologyeongnamegawakayamagazi" + + "neat-urlmitoyoakemiuramiyazurewebsiteshikagamiishibukawamiyotama" + + "nomjondalenmlbfanmonstermontrealestatefarmequipmentrentinoaltoad" + + "igemonza-brianzaporizhzhiamonza-e-della-brianzapposhirahamatonbe" + + "tsurnadalmonzabrianzaptokyotangotpantheonsitemonzaebrianzaramonz" + + "aedellabrianzamoonscalemoparachutingmordoviamoriyamatsumotofukem" + + "oriyoshiminamiawajikis-foundationmormonmouthaebaruericssonyoursi" + + "degreemoroyamatsunomortgagemoscowindowshirakofuefukihaborokunohe" + + "althcareershiranukanagawamoseushistorymosjoenmoskeneshiraois-gon" + + "emosshiraokananiimihoboleslawiechromedicinakamurataishinomakindl" + + "egnicafedexhibitionishinoomotegomosvikmpspbarrell-of-knowledgeom" + + "etre-experts-comptables3-sa-east-1moteginowaniihamatamakawajimao" + + "ris-into-animeiwamarshallstatebankddielddanuorrikuzentakatajimid" + + "oriopretogoldpoint2thisamitsukemoviemovimientolgamovistargardmoz" + + "illa-iotrentinos-tirolmtranbymuenstermuginozawaonsenmuikamisunag" + + "awamukodairamulhouservehalflifestylemunakatanemuncienciamuosatte" + + "mupictureshiratakahagitlabormurmansknx-serverrankoshigayanagawam" + + "urotorcraftrentinostirolmusashimurayamatsusakahoginankokubunjis-" + + "into-carshimotsukemusashinoharamuseetrentinosued-tirolmuseumvere" + + "nigingmusicargodaddyn-vpndnshishikuis-into-cartoonshimotsumamuts" + + "uzawamy-vigorgemy-wanggouvicenzamyactivedirectorymyasustor-elvda" + + "lmycdn77-sslattuminamiuonumassa-carrara-massacarraramassabusines" + + "sebyklegalloanshiojirishirifujiedamydattolocalhistorymyddnskingm" + + "ydissentrentinosuedtirolmydroboehringerikemydshisognemyeffectren" + + "tinsued-tirolmyfirewallonieruchomoscienceandindustrynmyfritzmyft" + + "paccesshisuifuelveruminamiyamashirokawanabelembetsukubankhmelnit" + + "skiyamarylhurstgorymyhome-servermyjinomykolaivarggatrentinsuedti" + + "rolmymailermymediapchryslermyokohamamatsudamypepiemontemypetshit" + + "aramamyphotoshibalestrandabergamoarekeymachinewhampshirebungoono" + + "ipifonyminanomypiagetmyiphostfoldnavymypsxn--30rr7ymysecuritycam" + + "erakermyshopblockshizukuishimofusaitamatsukuris-into-gamessinaza" + + "wamytis-a-bookkeeperugiamytuleapilotshizuokanazawamyvnchungnamda" + + "lseidfjordyndns-homednsandoymywireitrentoyonezawapiszpittsburgho" + + "fficialpiwatepixolinopizzapkomaganeplanetariumincommbankhmelnyts" + + "kyivaporcloudnshinjukumanoplantationplantshoujis-lostrodawarapla" + + "tformshangrilanshowaplaystationplazaplchurcharternidyndns-iparma" + + "ttelefonicarbonia-iglesias-carboniaiglesiascarboniaplumbingoplur" + + "inacionalpodzonepohlpoivronpokerpokrovskomakiyosunndalpoliticart" + + "ierpolitiendapolkowicepoltavalle-aostarostwodzislawinnershowtime" + + "mergencyahabahcavuotnagareyamakeupowiathletajimabaridagawalbrzyc" + + "haritysfjordpomorzeszowioshriramsterdamnserverbaniapordenoneporn" + + "porsangerporsangugeporsgrunnanyokoshibahikariwanumataketomisatos" + + "himayfirstjohnpoznanpraxis-a-bruinsfanprdpreservationpresidioprg" + + "mrprimelhusdecorativeartsienarutomobellevuelosangelesjabbottrevi" + + "sohughesigdalprincipeprivatizehealthinsuranceprochowiceproductio" + + "nsilkomatsushimasfjordenprofesionalprogressivenneslaskerrylogist" + + "icsimple-urlpromombetsurgeonshalloffameldalpropertyprotectionpro" + + "tonetritonprudentialpruszkowitdkomforbarsycentertainmentattoocea" + + "nographiqueu-2przeworskogptplusgardenpupimientakazakis-leetnedal" + + "pvhagakhanamigawapvtroandinosaurepaircraftingvollombardynamische" + + "s-dnsirdalpwcircleverappspotagerpzqldqponqslgbtrogstadquicksytes" + + "tingquipelementslingqvcircustomerstoregontrailroadstorfjordstorj" + + "devcloudcontrolledstpetersburgstreamuneuesokaneyamazoestudiostud" + + "yndns-at-homedepotenzamamidsundstuff-4-salestufftoread-booksneso" + + "kndalstuttgartrusteesusakis-not-certifieducatorahimeshimamateram" + + "obilysusonosuzakaniepcesuzukanmakiwiensuzukis-savedunetbankfhapp" + + "ouslivinghistorysvalbardunloppacificistrondheimmobilienishinoshi" + + "matsuurasveiosvelvikomvuxn--2scrj9christmasakinderoysvizzerasvn-" + + "reposolarssonswedenswidnicasacamdvrcampinagrandebugattipschlesis" + + "chesologneswiebodzindianapolis-a-bloggerswiftcoverswinoujscience" + + "andhistoryswisshikis-slickhakassiasynology-dsolundbeckommunetush" + + "uissier-justicetuvalle-daostatic-accessootuxfamilytwmailvestre-s" + + "lidrepbodynathomebuiltrvbashkiriautoscanadaejeonbuk12vestre-tote" + + "nnishiawakuravestvagoyvevelstadvibo-valentiavibovalentiavideovil" + + "lasnesoddenmarkhangelskjakdnepropetrovskiervaapsteiermarkongsvin" + + "gervinnicasadelamonedatingvinnytsiavipsinaappinkolobrzegersundvi" + + "rginiavirtual-userveexchangevirtualuserveftpioneervirtueeldomein" + + "-vigorlicevirtuelvisakegawaviterboknowsitallvivoldavixn--32vp30h" + + "agebostadvlaanderenvladikavkazimierz-dolnyvladimirvlogoipippulaw" + + "yvolkswagentsor-odalvologdanskoninjambylvolvolkenkundenvolyngdal" + + "vossevangenvotevotingvotoyonowiwatsukiyonoticiaskoyabearalvahkij" + + "observeronagarahkkeravjuegoshikikonaikawachinaganoharamcoachampi" + + "onshiphoptobishimaintenancebetsuikidsmynasushiobarackmazerbaijan" + + "-mayenebakkeshibechambagriculturennebudapest-a-la-masionthewifia" + + "t-band-campaniawloclawekonskowolayangrouphotographysiowmflabsor-" + + "varangerworldworse-thandawowithgoogleapisa-hockeynutsiracusakata" + + "kinouewpdevcloudyclusterwritesthisblogsytewroclawithyoutuberspac" + + "ewtcminnesotaketakatsukis-an-entertainerwtfastvps-serverisignwuo" + + "zuwzmiuwajimaxn--3pxu8konsulatrobeepilepsydneyxn--42c2d9axn--45b" + + "r5cylxn--45brj9citichernovtsykkylvenetogakushimotoganewyorkshire" + + "cipesaro-urbino-pesarourbinopesaromasvuotnakaiwamizawassamukawat" + + "aricohdatsunanjoburgminakamichiharaxn--45q11civilaviationishioko" + + "ppegardyndns-mailottexn--4gbriminingxn--4it168dxn--4it797konyvel" + + "ombardiamondshinshinotsurgeryxn--4pvxs4allxn--54b7fta0ccivilisat" + + "ionishitosashimizunaminamibosogndalottokorozawaxn--55qw42gxn--55" + + "qx5dxn--5js045dxn--5rtp49civilizationishiwakis-a-chefarsundyndns" + + "-office-on-the-weberlincolnissandiegoxn--5rtq34kooris-a-socialis" + + "tcgrouphiladelphiaareadmyblogspotrentino-stirolxn--5su34j936bgsg" + + "xn--5tzm5gxn--6btw5axn--6frz82gxn--6orx2rxn--6qq986b3xlxn--7t0a2" + + "64civilwarmanagementoyotaparocherkasyno-dsandvikcoromantovalle-d" + + "-aostathellexn--80adxhksorfoldxn--80ao21axn--80aqecdr1axn--80ase" + + "hdbasilicataniaveroykeniwaizumiotsukumiyamazonawsadodgemological" + + "lavangenaval-d-aosta-valleyokotemrevistanbulsan-suedtirolaziobni" + + "nskaragandaustraliaisondriobranconagawalesundemoneyboltateshinan" + + "omachimkentateyamaurskog-holandingjerdrumetacentrumeteorappalerm" + + "omahachijolstereviewskrakowebspacempresashibetsukuibmdds3-ap-sou" + + "theast-1kappchizippodhaleangaviikadenaamesjevuemielno-ip6xn--80a" + + "swgxn--80audnedalnxn--8ltr62kopervikharkovaoxn--8pvr4uxn--8y0a06" + + "3axn--90a3academiamicaaarborteaches-yogasawaracingxn--90aeroport" + + "alabamagasakishimabaraogakibichuoxn--90aishobarakawagoexn--90azh" + + "ytomyravendbasketballyngenvironmentalconservationflfanfshostrowi" + + "ecasinorddalillesandefjordgcahcesuolocus-2xn--9dbhblg6dietcimdba" + + "todayolasiteu-3xn--9dbq2axn--9et52uxn--9krt00axn--andy-iraxn--ar" + + "oport-byandexn--3bst00minternationalfirearmshioyanaizutwentexn--" + + "asky-iraxn--aurskog-hland-jnbatsfjordiscountyombolzano-altoadige" + + "u-4xn--avery-yuasakuhokkaidoomdnsiskinkyotobetsulikes-piedmontic" + + "ellodingenxn--b-5gaxn--b4w605ferdxn--balsan-sudtirol-rqis-uberle" + + "etrentino-sued-tirolxn--bck1b9a5dre4claimsanfranciscofreakunemur" + + "orangeiseiyoichippubetsubetsugarugbyengerdalaheadjudygarlandyndn" + + "s-picsangoxn--bdddj-mrabdxn--bearalvhki-y4axn--berlevg-jxaxn--bh" + + "cavuotna-s4axn--bhccavuotna-k7axn--bidr-5nachikatsuuraxn--bievt-" + + "0qa2xn--bjarky-fyaotsurreyxn--bjddar-ptamayufuettertdasnetzxn--b" + + "lt-elabourxn--bmlo-graingerxn--bod-2natalxn--bozen-sudtirol-76ha" + + "ibarakitahiroshimarburgxn--brnny-wuacademy-firewall-gatewayxn--b" + + "rnnysund-m8accident-investigation-aptibleaseating-organicbcieszy" + + "nxn--brum-voagatrysiljanxn--btsfjord-9zaxn--bulsan-sudtirol-rqis" + + "-very-badajozxn--c1avgxn--c2br7gxn--c3s14misakis-byxn--cck2b3bau" + + "hausposts-and-telecommunicationsncfdiscoveryomitanoddavocatanzar" + + "ownproviderhcloudfunctions3-eu-central-1xn--cesena-forli-c2gxn--" + + "cesenaforli-0jgoraxn--cg4bkis-very-evillagexn--ciqpnxn--clchc0ea" + + "0b2g2a9gcdxn--comunicaes-v6a2oxn--correios-e-telecomunicaes-ghc2" + + "9axn--czr694bbcn-north-1xn--czrs0tulanxessolutionslupskommunalfo" + + "rbundxn--czru2dxn--czrw28bbtcp4xn--d1acj3bbvacationswatch-and-cl" + + "ockerxn--d1alfaromeoxn--d1atunesomaxn--d5qv7z876clanbibaidarmeni" + + "axn--davvenjrga-y4axn--djrs72d6uyxn--djty4koryokamikawanehonbets" + + "urutaharaxn--dnna-grajewolterskluwerxn--drbak-wuaxn--dyry-iraxn-" + + "-e1a4cldmailouvreisenissayokkaichiropractichirurgiens-dentistes-" + + "en-francexn--eckvdtc9dxn--efvn9sorocabalsfjordxn--efvy88hair-sur" + + "veillancexn--ehqz56nxn--elqq16hakatanortonxn--estv75gxn--eveni-0" + + "qa01gaxn--f6qx53axn--fct429kosaigawaxn--fhbeiarnxn--finny-yuaxn-" + + "-fiq228c5hsorreisahayakawakamiichikawamisatourslzxn--fiq64beneve" + + "ntoeidsvollillyonagoyavoues3-eu-west-1xn--fiqs8sortlandxn--fiqz9" + + "soruminiserversicherungxn--fjord-lraxn--fjq720axn--fl-ziaxn--flo" + + "r-jraxn--flw351exn--forli-cesena-41gxn--forlicesena-ujgxn--fpcrj" + + "9c3dxn--frde-grandrapidsoundcastronomy-routerxn--frna-woaraisaij" + + "osoyrorosouthcarolinarvikomonowtvareservehttphonefosshinkamigoto" + + "yohashimotottoris-a-republicancerresearchaeologicaliforniaxn--fr" + + "ya-hraxn--fzc2c9e2clickashibatakashimarumorimachidaxn--fzys8d69u" + + "vgmailxn--g2xx48clinichiryukyuragifuchungbukharaumalopolskanland" + + "urbanamexnetlifyis-a-celticsfanishikatakayamatsushigexn--gckr3f0" + + "fauskedsmokorsetagayasells-for-ufcfanxn--gecrj9cliniquenoharaxn-" + + "-ggaviika-8ya47hakodatexn--gildeskl-g0axn--givuotna-8yasakaimina" + + "toyookannamilanotteroyxn--gjvik-wuaxn--gk3at1exn--gls-elacaixaxn" + + "--gmq050is-very-goodhandsonxn--gmqw5axn--h-2failxn--h1aeghakonex" + + "n--h2breg3evenesouthwestfalenxn--h2brj9c8clintonoshoesanjotoyoto" + + "miyazakis-a-conservativegarsheis-a-cpadualstackspace-to-rentalst" + + "omakomaibaraxn--h3cuzk1digitalxn--hbmer-xqaxn--hcesuolo-7ya35ben" + + "tleyonaguniversityoriikarateverbankaratsuginamikatagamilitaryosh" + + "iokaracoldwarmiastagevje-og-hornnes3-us-east-2xn--hery-iraxn--hg" + + "ebostad-g3axn--hmmrfeasta-s4accident-prevention-webhostingxn--hn" + + "efoss-q1axn--hobl-iraxn--holtlen-hxaxn--hpmir-xqaxn--hxt814exn--" + + "hyanger-q1axn--hylandet-54axn--i1b6b1a6a2exn--imr513nxn--indery-" + + "fyasugivingxn--io0a7is-very-nicexn--j1aefbsbxn--12cfi8ixb8luxury" + + "xn--j1amhakubahccavuotnagasakikuchikuseikarugamvikaufenxn--j6w19" + + "3gxn--jlq61u9w7beppublishproxyzjampagefrontappalmspringsakerxn--" + + "jlster-byasuokanraxn--jrpeland-54axn--jvr189misasaguris-certifie" + + "dogawarabikomaezakirunordreisa-geekazunowruzhgorodeoxn--k7yn95ex" + + "n--karmy-yuaxn--kbrq7oxn--kcrx77d1x4axn--kfjord-iuaxn--klbu-woax" + + "n--klt787dxn--kltp7dxn--kltx9axn--klty5xn--3ds443gxn--koluokta-7" + + "ya57hakuis-a-landscaperxn--kprw13dxn--kpry57dxn--kpu716fbx-osasa" + + "yamaxn--kput3is-very-sweetpepperxn--krager-gyatomitamamuraxn--kr" + + "anghke-b0axn--krdsherad-m8axn--krehamn-dxaxn--krjohka-hwab49jdfa" + + "stlylbarefootballfinanzgorauthordalandeportenrightathomeftpalmas" + + "eratiitatebayashiibajddarchitecturealtydalces3-ca-central-1xn--k" + + "snes-uuaxn--kvfjord-nxaxn--kvitsy-fyatsukanumazuryxn--kvnangen-k" + + "0axn--l-1fairwindsowaxn--l1accentureklamborghinikis-with-theband" + + "ovre-eikerxn--laheadju-7yatsushiroxn--langevg-jxaxn--lcvr32dxn--" + + "ldingen-q1axn--leagaviika-52beskidyn-o-saurlandes3-us-gov-west-1" + + "xn--lesund-huaxn--lgbbat1ad8jelenia-goraxn--lgrd-poacctunkongsbe" + + "rgxn--lhppi-xqaxn--linds-pramericanarturystykanoyakumoldelmenhor" + + "stalbansomnarviikamitondabayashiogamagoriziaxn--lns-qlapyxn--loa" + + "bt-0qaxn--lrdal-sraxn--lrenskog-54axn--lt-liaclothingdustkagoshi" + + "malselvendrellowiczest-le-patronissedalucaniaxn--lten-granexn--l" + + "ury-iraxn--m3ch0j3axn--mely-iraxn--merker-kuaxn--mgb2ddespeedpar" + + "tnersnoasaitotalxn--mgb9awbfbxosasebofagexn--mgba3a3ejtuscanyxn-" + + "-mgba3a4f16axn--mgba3a4franamizuholdingspiegelxn--mgba7c0bbn0axn" + + "--mgbaakc7dvfedorapeopleirfjordynnsarpsborguidefinimakanegasakin" + + "kobayashikaoirminamifuranoxn--mgbaam7a8hakusanagochijiwadell-ogl" + + "iastraderxn--mgbab2bdxn--mgbai9a5eva00bestbuyshouses3-us-west-1x" + + "n--mgbai9azgqp6jeonnamerikawauexn--mgbayh7gpalacexn--mgbb9fbpoba" + + "nazawaxn--mgbbh1a71exn--mgbc0a9azcgxn--mgbca7dzdoxn--mgberp4a5d4" + + "a87gxn--mgberp4a5d4arxn--mgbgu82axn--mgbi4ecexposedxn--mgbpl2fhs" + + "kydivingxn--mgbqly7c0a67fbcngrondarxn--mgbqly7cvafranziskanerima" + + "ringatlantakaharuxn--mgbt3dhdxn--mgbtf8flatangerxn--mgbtx2betain" + + "aboxfusejnynysagaeroclubmedecincinnationwidealerimo-i-ranadexete" + + "rxn--mgbx4cd0abbvieeexn--mix082fedoraprojectransurluzernxn--mix8" + + "91feiraquarelleborkangerxn--mjndalen-64axn--mk0axindianmarketing" + + "xn--mk1bu44cnpyatigorskodjeffersonisshingucciprianiigataitogliat" + + "tiresannaniyodogawaxn--mkru45isleofmanchesterxn--mlatvuopmi-s4ax" + + "n--mli-tlaquilanciaxn--mlselv-iuaxn--moreke-juaxn--mori-qsakurag" + + "awaxn--mosjen-eyawaraxn--mot-tlarvikosakaerodromegallupinbarsyon" + + "linewhollandevelopmentaxihuanayorovigotsukitahatakamatsukawautom" + + "otiveconomiasakuchinotsuchiurakawalmartatsunobiraustinnatuurwete" + + "nschappenaumburgjerstadotsuruokakegawaukraanghkepnogataijibigawa" + + "etnagahamaroyereportatarantoyakokonoebinordre-landd-dnshome-webs" + + "ervercelliguriagrocerybnikahokutobamagentositecnologiajudaicable" + + "-modemocraciaugustowadaeguambulancebizenakatombetsumitakagiizehi" + + "mejibestaddnslivelanddnss3-ap-south-16-bambleclerc66xn--mre-og-r" + + "omsdal-qqbhzcateringebuilderschmidtre-gauldalimanowarudaxaustrhe" + + "imatunduhrennesoyokosukanzakiyokawaraustevoll-o-g-i-naturhistori" + + "sches3-ap-southeast-2ix4432-bananarepublicaseihicampobassociates" + + "t-iservecounterstrike12hpaleobihirosakikamijimatsuzaki234lima-ci" + + "tyeatselinogradultarnobrzegyptian4tarumizusawabogadocscbgdyniabr" + + "uzzoologicalvinklein-addrammenuernberggfarmerseine164-barcelonag" + + "asukeastcoastaldefenceatonsbergjemnes3-ap-northeast-1337xn--msy-" + + "ula0haldenxn--mtta-vrjjat-k7afamilycompanycnsannohelplfinancialu" + + "ccapitalonewspaperxn--muost-0qaxn--mxtq1misawaxn--ngbc5azdxn--ng" + + "be9e0axn--ngbrxn--3e0b707exn--nit225koseis-a-soxfanxn--nmesjevue" + + "mie-tcbaltimore-og-romsdalipayxn--nnx388axn--nodessakuraissmarte" + + "rthanyousrlxn--nqv7fs00emaxn--nry-yla5gxn--ntso0iqx3axn--ntsq17g" + + "xn--nttery-byaeservehumourxn--nvuotna-hwaxn--nyqy26axn--o1achase" + + "ljeepsongdalenviknaharimalborkdalxn--o3cw4halsaintlouis-a-anarch" + + "istoireggiocalabriaxn--o3cyx2axn--od0algxn--od0aq3bieigersundish" + + "akotanhktjeldsundisrechtrainingjesdalimitedivtasvuodnakaniikawat" + + "anaguraxn--ogbpf8flekkefjordxn--oppegrd-ixaxn--ostery-fyawataham" + + "axn--osyro-wuaxn--otu796dxn--p1acfermochizukirkenesaskatchewanxn" + + "--p1aiwchoshibuyachiyodattorelayxn--pbt977cntoyotsukaidoxn--pgbs" + + "0dhlxn--porsgu-sta26ferraraxn--pssu33lxn--pssy2uxn--q9jyb4coguch" + + "ikuzenxn--qcka1pmckinseyxn--qqqt11misconfusedxn--qxamusementdllc" + + "ube-serversaillespjelkavikomorotsukamiokamikitayamatsuris-a-rock" + + "starachowicexn--rady-iraxn--rdal-poaxn--rde-ulavagiskexn--rdy-0n" + + "abarixn--rennesy-v1axn--rhkkervju-01aflakstadaokagakicks-assedic" + + "ollectionxn--rholt-mragowoodsideltaiwanairlinedre-eikerxn--rhqv9" + + "6gxn--rht27zxn--rht3dxn--rht61exn--risa-5nativeamericanantiquesp" + + "readbettingxn--risr-iraxn--rland-uuaxn--rlingen-mxaxn--rmskog-by" + + "axn--rny31hammarfeastafricapetownnews-stagingxn--rovu88bielawalt" + + "erxn--rros-granvindafjordxn--rskog-uuaxn--rst-0naturalhistorymus" + + "eumcenterxn--rsta-francaiseharaxn--rvc1e0am3exn--ryken-vuaxn--ry" + + "rvik-byaxn--s-1faithruheredumbrellajollamericanexpressexyxn--s9b" + + "rj9colognextdirectoyouraxn--sandnessjen-ogbizxn--sandy-yuaxn--se" + + "ral-lraxn--ses554gxn--sgne-gratangenxn--skierv-utazassnasabaerob" + + "aticketspydebergxn--skjervy-v1axn--skjk-soaxn--sknit-yqaxn--sknl" + + "and-fxaxn--slat-5naturalsciencesnaturellesrtromsojamisonxn--slt-" + + "elabcgxn--smla-hraxn--smna-gratis-a-bulls-fanxn--snase-nraxn--sn" + + "dre-land-0cbremangerxn--snes-poaxn--snsa-roaxn--sr-aurdal-l8axn-" + + "-sr-fron-q1axn--sr-odal-q1axn--sr-varanger-ggbiellaakesvuemielec" + + "cexn--srfold-byaxn--srreisa-q1axn--srum-grazxn--stfold-9xaxn--st" + + "jrdal-s1axn--stjrdalshalsen-sqbieszczadygeyachimataikikugawarsza" + + "washingtondclkareliancexn--stre-toten-zcbsrvaroyxn--sudtirol-y0e" + + "mmafann-arboretumbriamallamaceioxn--t60b56axn--tckweatherchannel" + + "xn--tiq49xqyjetztrentino-suedtirolxn--tjme-hraxn--tn0agrinet-fre" + + "akstoragexn--tnsberg-q1axn--tor131oxn--trany-yuaxn--trentin-sud-" + + "tirol-tsjcbnlxn--trentin-sudtirol-b9ixn--trentino-sud-tirol-dcko" + + "sherbrookegawaxn--trentino-sudtirol-usjevnakershuscultureggioemi" + + "liaromagnamsskoganeis-a-playerxn--trentinosud-tirol-tsjewelryxn-" + + "-trentinosudtirol-b9ixn--trentinsud-tirol-98ixn--trentinsudtirol" + + "-rqixn--trgstad-r1axn--trna-woaxn--troms-zuaxn--tysvr-vraxn--uc0" + + "atvestfoldxn--uc0ay4axn--uist22hamurakamigoris-a-lawyerxn--uisz3" + + "gxn--unjrga-rtaobaomoriguchiharagusartstordalxn--unup4yxn--uuwu5" + + "8axn--vads-jraxn--vallee-aoste-i2gxn--vallee-d-aoste-43hangglidi" + + "ngxn--valleeaoste-6jgxn--valleedaoste-i2gxn--vard-jraxn--vegrshe" + + "i-c0axn--vermgensberater-ctbievatmallorcadaques3-us-west-2xn--ve" + + "rmgensberatung-pwbifukagawashtenawdev-myqnapcloudaccesscambridge" + + "stoneustarhubs3-website-ap-northeast-1xn--vestvgy-ixa6oxn--vg-yi" + + "abkhaziaxn--vgan-qoaxn--vgsy-qoa0jewishartgalleryxn--vgu402colon" + + "ialwilliamsburgrongaxn--vhquvestnesopotromsakakinokiaxn--vler-qo" + + "axn--vre-eiker-k8axn--vrggt-xqadxn--vry-yla5gxn--vuq861bihorolog" + + "yukindigenamsosnowiecatholicaxiascolipicenodumetlifeinsurancexn-" + + "-w4r85el8fhu5dnraxn--w4rs40lxn--wcvs22dxn--wgbh1coloradoplateaud" + + "ioxn--wgbl6axn--xhq521bikedagestangeorgeorgiaxn--xkc2al3hye2axn-" + + "-xkc2dl3a5ee0hangoutsystemscloudfrontdoorxn--y9a3aquariumishimas" + + "udaxn--yer-znaturbruksgymnxn--yfro4i67oxn--ygarden-p1axn--ygbi2a" + + "mmxn--3hcrj9citadeliveryggeelvinckasaokaminokawanishiaizubangexn" + + "--ystre-slidre-ujbilbaogashimadachicagoboats3-website-ap-southea" + + "st-1xn--zbx025dxn--zf0ao64axn--zf0avxn--3oq18vl8pn36axn--zfr164b" + + "illustrationhlfanhs3-website-ap-southeast-2xnbayxperiaxz" // nodes is the list of nodes. Each node is represented as a uint32, which // encodes the node's children, wildcard bit and node type (as an index into @@ -500,8517 +512,8694 @@ const text = "0emmafann-arboretumbriamallamaceiobihirosakikamijimatsuzaki234li" // [15 bits] text index // [ 6 bits] text length var nodes = [...]uint32{ - 0x31a803, - 0x284d84, - 0x382f06, - 0x2f37c3, - 0x2f37c6, - 0x37af86, - 0x3a7a03, - 0x31b604, - 0x322487, - 0x382b48, - 0x1a00742, - 0x32e147, - 0x3672c9, - 0x2b4eca, - 0x2b4ecb, - 0x232183, - 0x2ab9c6, - 0x238485, - 0x1e01482, - 0x203b44, - 0x260543, - 0x201485, - 0x2215842, - 0x332603, - 0x271b0c4, - 0x31fe05, - 0x2a00102, - 0x38194e, - 0x256483, - 0x39cbc6, - 0x2e03d02, - 0x2c8047, - 0x23e146, - 0x3205c42, - 0x257dc3, - 0x257dc4, - 0x357406, - 0x205d08, - 0x277146, - 0x302004, - 0x3600602, - 0x33acc9, - 0x211307, - 0x347986, - 0x3c1109, - 0x2c78c8, - 0x331004, - 0x241286, - 0x230106, - 0x3a00582, - 0x3a234f, - 0x21f4ce, - 0x226484, - 0x2c1545, - 0x31a705, - 0x2f6809, - 0x244689, - 0x357c07, - 0x22bbc6, - 0x206dc3, - 0x3e03942, - 0x21d6c3, - 0x220d4a, - 0x21fbc3, - 0x3bde45, - 0x2f2542, - 0x370749, - 0x4200282, - 0x216c84, - 0x2ef006, - 0x2bb6c5, - 0x2d7c04, - 0x4a14344, - 0x205583, - 0x2374c4, - 0x4e02b82, - 0x267184, - 0x527eac4, - 0x39004a, - 0x5600cc2, - 0x35c447, - 0x2774c8, - 0x6207ec2, - 0x340687, - 0x2bde44, - 0x2bde47, - 0x3b9605, - 0x339407, - 0x31ca86, - 0x325384, - 0x3314c5, - 0x298307, - 0x720fc02, - 0x335a43, - 0x21ab82, - 0x3aae43, - 0x7612442, - 0x27f485, - 0x7a023c2, - 0x293584, - 0x276005, - 0x2263c7, - 0x20974e, - 0x2391c4, - 0x238cc4, - 0x20b583, - 0x364209, - 0x30e2cb, - 0x259e48, - 0x3c0ec8, - 0x316488, - 0x215cc8, - 0x330e4a, - 0x339307, - 0x309d86, - 0x7e6e442, - 0x345243, - 0x355943, - 0x35d344, - 0x3a7a43, - 0x32f6c3, - 0x172a782, - 0x8203102, - 0x27b385, - 0x28df86, - 0x2a9f04, - 0x369187, - 0x23ce86, - 0x3806c4, - 0x3806c7, - 0x205a83, - 0x86c31c2, - 0x8b14902, - 0x8e21182, - 0x221186, - 0x9200882, - 0x286c45, - 0x32bcc3, - 0x3c6444, - 0x2e3804, - 0x2e3805, - 0x2053c3, - 0x96b6c03, - 0x9a09342, - 0x289b05, - 0x289b0b, - 0x20bd06, - 0x331f4b, - 0x22aa44, - 0x20cec9, - 0x20d784, - 0x9e0d9c2, - 0x20ef03, - 0x20fec3, - 0x1610702, - 0x2fb9c3, - 0x21070a, - 0xa200302, - 0x203dc5, - 0x2d400a, - 0x243384, - 0x210f03, - 0x212984, - 0x213b83, - 0x213b84, - 0x213b87, - 0x2153c5, - 0x215705, - 0x216d46, - 0x2170c6, - 0x217d43, - 0x21a708, - 0x212d43, - 0xa6004c2, - 0x22c3c8, - 0x3878cb, - 0x223088, - 0x225f06, - 0x227447, - 0x22a1c8, - 0xb604002, - 0xbaf21c2, - 0x23b388, - 0x3031c7, - 0x207a45, - 0x207a48, - 0x383c48, - 0x2fa9c3, - 0x22f384, - 0x35d382, - 0xbe2f582, - 0xc201bc2, - 0xca30502, - 0x230503, - 0xce03cc2, - 0x31b5c3, - 0x2f1b84, - 0x20bf83, - 0x335e04, - 0x322b8b, - 0x237c03, - 0x2db106, - 0x237c04, - 0x2e21ce, - 0x2669c5, - 0x33d7c8, - 0x251107, - 0x25110a, - 0x2342c3, - 0x34f747, - 0x30e485, - 0x2342c4, - 0x2d4b86, - 0x2d4b87, - 0x2d0204, - 0x37d587, - 0x209a84, - 0x340c44, - 0x340c46, - 0x25d944, - 0x39db46, - 0x207803, - 0x207808, - 0x21a988, - 0x238c83, - 0x2fb983, - 0x3a8c04, - 0x3ae4c3, - 0xd24d5c2, - 0xd6d2fc2, - 0x2083c3, - 0x205646, - 0x241383, - 0x354bc4, - 0xda4b182, - 0x24cb83, - 0x339c03, - 0x218882, - 0xde03c02, - 0x2c0b06, - 0x23c007, - 0x2eab45, - 0x38a504, - 0x2981c5, - 0x27e687, - 0x2d84c9, - 0x2dcd46, - 0x307788, - 0x2eaa46, - 0xe2010c2, - 0x2f1408, - 0x2f3e06, - 0x223a85, - 0x30fe07, - 0x310344, - 0x310345, - 0x2010c4, - 0x2010c8, - 0xe619382, - 0xea02642, - 0x3292c6, - 0x202648, - 0x34d485, - 0x34df06, - 0x350108, - 0x36d548, - 0xee1f8c5, - 0xf21d0c4, - 0x38ca87, - 0xf60d642, - 0xfaefa02, - 0x10e02c42, - 0x2ef105, - 0x373905, - 0x3c1546, - 0x3208c7, - 0x3973c7, - 0x1160be03, - 0x26f507, - 0x2b99c8, - 0x231a09, - 0x381b07, - 0x2321c7, - 0x232b08, - 0x233306, - 0x233dc6, - 0x234a0c, - 0x235e4a, - 0x2364c7, - 0x23834b, - 0x23be47, - 0x23be4e, - 0x1a23d104, - 0x23d744, - 0x23e847, - 0x2616c7, - 0x243806, - 0x243807, - 0x243c87, - 0x1a630a42, - 0x2449c6, - 0x2449ca, - 0x244f4b, - 0x246d07, - 0x2478c5, - 0x247c03, - 0x248146, - 0x248147, - 0x322643, - 0x1aa022c2, - 0x248a4a, - 0x1af68802, - 0x1b24d602, - 0x1b64afc2, - 0x1ba3e242, - 0x24cc85, - 0x24d2c4, - 0x1c204ac2, - 0x267205, - 0x245543, - 0x20d885, - 0x215bc4, - 0x20f984, - 0x209d86, - 0x2505c6, - 0x289d03, - 0x3b6d84, - 0x3ac2c3, - 0x1ca02e02, - 0x3582c4, - 0x3582c6, - 0x38d005, - 0x36e3c6, - 0x30ff08, - 0x227b84, - 0x397848, - 0x399a45, - 0x311708, - 0x36c6c6, - 0x265847, - 0x27b984, - 0x27b986, - 0x26f803, - 0x3917c3, - 0x20b648, - 0x31c684, - 0x354fc7, - 0x2d2906, - 0x2d2909, - 0x20a1c8, - 0x317908, - 0x338884, - 0x2067c3, - 0x23dd42, - 0x1da4c3c2, - 0x1de14202, - 0x207583, - 0x1e20a502, - 0x3225c4, - 0x2440c6, - 0x335b45, - 0x283403, - 0x234ec4, - 0x2b1a07, - 0x336bc3, - 0x37cfc8, - 0x21ea85, - 0x25f7c3, - 0x275f85, - 0x2760c4, - 0x2f9c06, - 0x222704, - 0x225986, - 0x226306, - 0x357d84, - 0x23c203, - 0x1e614582, - 0x238ac5, + 0x32b543, + 0x2872c4, + 0x2c8146, + 0x2f4d83, + 0x2f4d86, + 0x382346, + 0x3b2083, + 0x2d4484, + 0x393b47, + 0x2c7d88, + 0x1a000c2, + 0x1f3a8c7, + 0x36e6c9, + 0x2bf60a, + 0x2bf60b, + 0x22b8c3, + 0x2acc06, + 0x232005, + 0x220cd42, + 0x3d0744, + 0x25f303, + 0x393345, + 0x2605202, + 0x358d83, + 0x2b2be04, + 0x38afc5, + 0x2e20fc2, + 0x39670e, + 0x2513c3, + 0x3aa546, + 0x3200a82, + 0x2fa0c7, + 0x234846, + 0x3601102, + 0x27f3c3, + 0x27f3c4, + 0x20f2c6, + 0x204048, + 0x279986, + 0x309604, + 0x3a04342, + 0x3442c9, + 0x223087, + 0x399606, + 0x36b389, + 0x2d5588, + 0x32d484, + 0x238d06, + 0x35af06, + 0x3e01902, + 0x3ad9cf, + 0x27a34e, + 0x354144, + 0x2097c5, + 0x32b445, + 0x2f1789, + 0x240409, + 0x20fac7, + 0x201286, 0x2011c3, - 0x1ea05ec2, - 0x2319c3, - 0x21c8c5, - 0x237583, - 0x237589, - 0x1ee01f02, - 0x1f608ac2, - 0x289645, - 0x219286, - 0x37c8c6, - 0x2bfcc8, - 0x2bfccb, - 0x20568b, - 0x21c145, - 0x2ead45, - 0x2c3909, - 0x1603142, - 0x357f48, - 0x23e504, - 0x1fe01b02, - 0x20aac3, - 0x20661886, - 0x224fc8, - 0x20a003c2, - 0x307348, - 0x20e0a6c2, - 0x23994a, - 0x212c8d03, - 0x39f286, - 0x3b5048, - 0x389ac8, - 0x3ba046, - 0x377d47, - 0x3a2547, - 0x23fe0a, - 0x243404, - 0x352f84, - 0x366b89, - 0x21ba1d45, - 0x21f6c6, - 0x200143, - 0x255184, - 0x21e25784, - 0x323307, - 0x22f607, - 0x364044, - 0x2d3345, - 0x3c1608, - 0x37b847, - 0x38fc87, - 0x22208882, - 0x23b9c4, - 0x28e948, - 0x24e244, - 0x252944, - 0x253005, - 0x253147, - 0x22b509, - 0x254004, - 0x2547c9, - 0x254a08, - 0x254f04, - 0x254f07, - 0x226553c3, - 0x255547, - 0x1626d02, - 0x16ad402, - 0x255e86, - 0x2564c7, - 0x256b04, - 0x258487, - 0x258f47, - 0x259783, - 0x329982, - 0x205dc2, - 0x270003, - 0x270004, - 0x27000b, - 0x3c0fc8, - 0x25f184, - 0x25ad05, - 0x25cac7, - 0x25e5c5, - 0x30590a, - 0x25f0c3, - 0x22a12c42, - 0x212c44, - 0x261489, - 0x265183, - 0x265247, - 0x2f61c9, - 0x336308, - 0x25d1c3, - 0x27a247, - 0x27aa89, - 0x26be83, - 0x281b04, - 0x283c89, - 0x287dc6, - 0x2266c3, - 0x2039c2, - 0x241243, - 0x2ad207, - 0x383fc5, - 0x340346, - 0x268984, - 0x2dba05, - 0x220d03, - 0x217f86, - 0x20d0c2, - 0x3a3984, - 0x22e2ab02, - 0x22ab03, - 0x23201802, - 0x252843, - 0x217544, - 0x217547, - 0x3c6746, - 0x255e42, - 0x23629942, - 0x384384, - 0x23a30b82, - 0x23e01a42, - 0x337304, - 0x337305, - 0x201a45, - 0x35ab46, - 0x24208742, - 0x208745, - 0x2100c5, - 0x210ac3, - 0x213d06, - 0x214885, - 0x221102, - 0x34db45, - 0x221104, - 0x227ac3, - 0x227d03, - 0x2460ad82, - 0x298507, - 0x33a504, - 0x33a509, - 0x255084, - 0x281903, - 0x35b109, - 0x281908, - 0x24b0cc04, - 0x30cc06, - 0x2a2c83, - 0x20cb03, - 0x30e843, - 0x24eefe82, - 0x375502, - 0x25201402, - 0x32d8c8, - 0x327088, - 0x3a8046, - 0x2544c5, - 0x34f5c5, - 0x31e0c7, - 0x229985, - 0x25cd82, - 0x25694cc2, - 0x1602202, - 0x240a88, - 0x34e285, - 0x27ca84, - 0x2e7205, - 0x241d87, - 0x25efc4, - 0x248942, - 0x25a2dac2, - 0x33e704, - 0x226ec7, - 0x289fc7, - 0x3393c4, - 0x291003, - 0x238bc4, - 0x238bc8, - 0x234106, - 0x2d4a0a, - 0x22b3c4, - 0x291508, - 0x288204, - 0x227546, - 0x294c84, - 0x2ef406, - 0x33a7c9, - 0x26d007, - 0x34e1c3, - 0x25eebfc2, - 0x331203, - 0x207c82, - 0x2625c982, - 0x30cf06, - 0x371e48, - 0x2a44c7, - 0x2f7209, - 0x290ac9, - 0x2a61c5, - 0x2a73c9, - 0x2a7b85, - 0x2a7cc9, - 0x2a9045, - 0x2aa008, - 0x266598c4, - 0x26a598c7, - 0x232583, - 0x2aa207, - 0x232586, - 0x2aa5c7, - 0x2a0f45, - 0x2ca8c3, - 0x26e35c02, - 0x2ea984, - 0x27230bc2, - 0x276552c2, - 0x2f3ac6, - 0x277445, - 0x2acac7, - 0x326403, - 0x32f644, - 0x2130c3, - 0x23b0c3, - 0x27a07d02, - 0x28206202, - 0x37b084, - 0x329943, - 0x24b905, - 0x28603882, - 0x28e00c42, - 0x2e0586, - 0x31c7c4, - 0x385444, - 0x38544a, - 0x29601342, - 0x38e2ca, - 0x39e948, - 0x29a6ff84, - 0x201fc3, - 0x208c43, - 0x3165c9, - 0x267709, - 0x2a6e06, - 0x29e14bc3, - 0x214bc5, - 0x39434d, - 0x39eb06, - 0x20e84b, - 0x2a200802, - 0x220b88, - 0x2ca1a802, - 0x2ce00942, - 0x2c9a85, - 0x2d205842, - 0x21b147, - 0x2b0747, - 0x214a43, - 0x348148, - 0x2d601102, - 0x29f384, - 0x291203, - 0x325545, - 0x395983, - 0x245646, - 0x223504, - 0x2fb943, - 0x2aec03, - 0x2da03202, - 0x2eacc4, - 0x3af385, - 0x2ace07, - 0x277e03, - 0x2ad9c3, - 0x2ae803, - 0x16ae8c2, - 0x2ae8c3, - 0x2aeb83, - 0x2de0b0c2, - 0x39e304, - 0x2507c6, - 0x22a443, - 0x2af343, - 0x2e2b0102, - 0x2b0108, - 0x2b03c4, - 0x2ee8c6, - 0x256947, - 0x3845c6, - 0x2a4f04, - 0x3be01ec2, - 0x23244b, - 0x2ff28e, - 0x219e0f, - 0x2c7b83, - 0x3c65fe82, - 0x1647302, - 0x3ca00a82, - 0x25b4c3, - 0x205983, - 0x2d8746, - 0x2f1946, - 0x3c2147, - 0x2f9084, - 0x3ce193c2, - 0x3d21edc2, - 0x2425c5, - 0x2e44c7, - 0x37fd86, - 0x3d64d542, - 0x30de04, - 0x2b7b43, - 0x3da09602, - 0x3df63443, - 0x2b8444, - 0x2bd289, - 0x16c2482, - 0x3e20dd82, - 0x327e05, - 0x3e6c2702, - 0x3ea00682, - 0x352307, - 0x214fc9, - 0x36754b, - 0x3a2305, - 0x26ad09, - 0x37e806, - 0x20bd47, - 0x3ee074c4, - 0x348c89, - 0x337b07, - 0x224c87, - 0x230803, - 0x2afc46, - 0x30a7c7, - 0x20fbc3, - 0x2f0d46, - 0x3f6038c2, - 0x3fa0e402, - 0x3bec83, - 0x32f245, - 0x332807, - 0x222386, - 0x383f45, - 0x2f3f04, - 0x278f45, - 0x2f2144, - 0x3fe00f82, - 0x341587, - 0x2f2984, - 0x26a444, - 0x34694d, - 0x26a449, - 0x230b08, - 0x25c404, - 0x335ec5, - 0x20a047, - 0x341144, - 0x23cf47, - 0x204cc5, - 0x402a4e44, - 0x30bcc5, - 0x263e44, - 0x390706, - 0x3206c5, - 0x406291c2, - 0x210fc4, - 0x210fc5, - 0x35d8c6, - 0x343b85, - 0x25d144, - 0x3c6103, - 0x20eb46, - 0x22b705, - 0x22f045, - 0x3207c4, - 0x22b443, - 0x22b44c, - 0x40aacf02, - 0x40e0a5c2, - 0x41201542, - 0x20f003, - 0x20f004, - 0x41604482, - 0x30ae88, - 0x340405, - 0x236184, - 0x243686, - 0x41a0e302, - 0x41e1de42, - 0x422000c2, - 0x2b2cc5, - 0x294346, - 0x229304, - 0x357946, - 0x35c206, - 0x222a83, - 0x4272850a, - 0x26b085, - 0x28b003, - 0x228606, - 0x304789, - 0x228607, - 0x292288, - 0x2c7789, - 0x31d348, - 0x250e46, - 0x209703, - 0x42a6f582, - 0x392c08, - 0x42e54ac2, - 0x43201e42, - 0x20be83, - 0x2d8345, - 0x26ba04, - 0x3b6fc9, - 0x2ee004, - 0x21b388, - 0x20dc03, - 0x323004, - 0x2a5fc3, - 0x2192c8, - 0x346887, - 0x43a25242, - 0x290ec2, - 0x31a685, - 0x39cf89, - 0x21f743, - 0x27bfc4, - 0x394304, - 0x20a0c3, - 0x27d04a, - 0x43f7c0c2, - 0x44210f82, - 0x2c3143, - 0x37ea83, - 0x1600082, - 0x200083, - 0x44603282, - 0x44a05a02, - 0x44e1a484, - 0x322046, - 0x2e07c6, - 0x245e44, - 0x277043, - 0x345c03, - 0x2ec1c3, - 0x2452c6, - 0x341d05, - 0x2c32c7, - 0x2c6445, - 0x2c7d86, - 0x2c8708, - 0x2c8906, - 0x24efc4, - 0x29960b, - 0x2cb583, - 0x2cb585, - 0x2cba08, - 0x21a202, - 0x352602, - 0x4524cd02, - 0x4560d682, - 0x219403, - 0x45a6cd82, - 0x26cd83, - 0x2cbd04, - 0x2cc543, - 0x462168c2, - 0x466d0e06, - 0x25e446, - 0x46ad0f42, - 0x46e0ff02, - 0x47227d42, - 0x4763a3c2, - 0x47a1b5c2, - 0x47e047c2, - 0x20dec3, - 0x358645, - 0x2b6306, - 0x48226444, - 0x38ce0a, - 0x3a0546, - 0x21c384, - 0x277943, - 0x48e02f02, - 0x2032c2, - 0x26fb43, - 0x4920ec83, - 0x2e6747, - 0x3205c7, - 0x4aa70107, - 0x393f87, - 0x22cd03, - 0x3176ca, - 0x251304, - 0x397504, - 0x39750a, - 0x247705, - 0x4ae1f682, - 0x258443, - 0x4b202002, - 0x228803, - 0x3311c3, - 0x4ba02742, - 0x26f484, - 0x220704, - 0x2046c5, - 0x3080c5, - 0x34e4c6, - 0x34e846, - 0x4be53982, - 0x4c201382, - 0x2f8545, - 0x25e152, - 0x33f206, - 0x25e8c3, - 0x39d486, - 0x2a1f45, - 0x1604842, - 0x54610c82, - 0x35e3c3, - 0x210c83, - 0x27e483, - 0x54a0c502, - 0x381c43, - 0x54e06e02, - 0x200843, - 0x39e348, - 0x285603, - 0x2a6046, - 0x23ecc7, - 0x30b986, - 0x30b98b, - 0x21c2c7, - 0x2ea784, - 0x55601c82, - 0x340285, - 0x55a09cc3, - 0x292c83, - 0x239b45, - 0x3175c3, - 0x55f175c6, - 0x3580ca, - 0x245ac3, - 0x23e004, - 0x202586, - 0x223e86, - 0x56241d03, - 0x32f507, - 0x2a6d07, - 0x29ae85, - 0x311986, - 0x22b743, - 0x58e13f43, - 0x59206f02, - 0x21a244, - 0x207609, - 0x240887, - 0x229a85, - 0x247d04, - 0x26c7c8, - 0x273b85, - 0x59676405, - 0x284e49, - 0x347a43, - 0x24d584, - 0x59a0b182, - 0x219603, - 0x59e94742, - 0x299986, - 0x162bac2, - 0x5a2a47c2, - 0x2b2bc8, - 0x3a76c3, - 0x30bc07, - 0x2ce245, - 0x2b2785, - 0x2d8e4b, - 0x2d9846, - 0x2d9046, - 0x2dc486, - 0x279f04, - 0x2dc6c6, - 0x5a6f0248, - 0x237cc3, - 0x201f83, - 0x201f84, - 0x2ddbc4, - 0x2dde87, - 0x2df2c5, - 0x5aadf402, - 0x5ae08302, - 0x208305, - 0x2bb184, - 0x2e298b, - 0x2e3708, - 0x298804, - 0x230982, - 0x5b64e882, - 0x24e883, - 0x2e3f04, - 0x2e41c5, - 0x2e4d07, - 0x2e6d44, - 0x21c184, - 0x5ba057c2, - 0x36b449, - 0x2e84c5, - 0x3a25c5, - 0x2e9045, - 0x5be19543, - 0x2e9d04, - 0x2e9d0b, - 0x2ea0c4, - 0x2ea38b, - 0x2ec105, - 0x219f4a, - 0x2ecec8, - 0x2ed0ca, - 0x2ed983, - 0x2ed98a, - 0x5c21fc42, - 0x5c648602, - 0x209943, - 0x5caf1382, - 0x2f1383, - 0x5cf6c182, - 0x5d32c442, - 0x2f1fc4, - 0x21a846, - 0x357685, - 0x2f3d83, - 0x31adc6, - 0x34f085, - 0x250ac4, - 0x5d600382, - 0x2aefc4, - 0x2c358a, - 0x398a07, - 0x3477c6, - 0x24f3c7, - 0x244a03, - 0x2b8488, - 0x3a1f8b, - 0x2bd905, - 0x27c785, - 0x27c786, - 0x2dd704, - 0x3b5288, - 0x21d343, - 0x230004, - 0x230007, - 0x2f4a06, - 0x31fa06, - 0x2e200a, - 0x254844, - 0x31104a, - 0x5db364c6, - 0x3364c7, - 0x25ad87, - 0x273544, - 0x273549, - 0x250485, - 0x31cf4b, - 0x2e1283, - 0x225b43, - 0x5de20b43, - 0x2344c4, - 0x5e200982, - 0x3a3006, - 0x5e6ca645, - 0x39d6c5, - 0x258c46, - 0x29cd44, - 0x5ea07bc2, - 0x247c44, - 0x5ee16f02, - 0x224645, - 0x23f4c4, - 0x228d83, - 0x5f610cc2, - 0x210cc3, - 0x267b86, - 0x5fa00a02, - 0x2073c8, - 0x228484, - 0x228486, - 0x37f306, - 0x25cb84, - 0x20eac5, - 0x21cfc8, - 0x220f87, - 0x2227c7, - 0x2227cf, - 0x28e846, - 0x309bc3, - 0x398184, - 0x233744, - 0x2101c3, - 0x227684, - 0x34fd84, - 0x5fe030c2, - 0x289a43, - 0x372e83, - 0x60207c02, - 0x25c503, - 0x322683, - 0x21578a, - 0x207c07, - 0x2534cc, - 0x253786, - 0x255246, - 0x256647, - 0x60632f47, - 0x25c889, - 0x22c504, - 0x25eb04, - 0x60a09f82, - 0x60e01282, - 0x2e23c6, - 0x32f304, - 0x2d3146, - 0x2333c8, - 0x239444, - 0x21b186, - 0x37c885, - 0x285048, - 0x205883, - 0x28a685, - 0x290cc3, - 0x3a26c3, - 0x3a26c4, - 0x212c03, - 0x6125fd82, - 0x61603a42, - 0x2e1149, - 0x299885, - 0x2a1084, - 0x2a4a45, - 0x212384, - 0x393607, - 0x353f05, - 0x2702c4, - 0x2702c8, - 0x2e61c6, - 0x2e9f84, - 0x2ede88, - 0x2f27c7, - 0x61a04042, - 0x316244, - 0x210284, - 0x224e87, - 0x61e04044, - 0x2c9002, - 0x6220ed42, - 0x221b83, - 0x2d37c4, - 0x29bb43, - 0x2aacc5, - 0x6262e642, - 0x2fddc5, - 0x23a382, - 0x390c85, - 0x372005, - 0x62a04e02, - 0x339b84, - 0x62e06a42, - 0x3aba46, - 0x3a7346, - 0x39d0c8, - 0x2be888, - 0x2f3a44, - 0x303685, - 0x316049, - 0x2eadc4, - 0x358084, - 0x2b6983, - 0x6320fd85, - 0x2c2547, - 0x21fc84, - 0x3ae54d, - 0x2f4682, - 0x3b35c3, - 0x2f4683, - 0x63601b42, - 0x396e45, - 0x223747, - 0x2b9604, - 0x394047, - 0x2c7989, - 0x2c36c9, - 0x275247, - 0x202bc3, - 0x3a7508, - 0x25b949, - 0x2f5487, - 0x2f5805, - 0x2f6706, - 0x2f6d46, - 0x2f6ec5, - 0x26a545, - 0x63a00d42, - 0x2b7685, - 0x2b3a08, - 0x2c08c6, - 0x63e872c7, - 0x2ec344, - 0x2b8047, - 0x2f9206, - 0x6420a402, - 0x35d5c6, - 0x2fc9ca, - 0x2fd245, - 0x646da942, - 0x64a4eb02, - 0x30ab06, - 0x386548, - 0x64e8a187, - 0x6523c902, - 0x215c43, - 0x20c246, - 0x229144, - 0x3b2f86, - 0x201746, - 0x34290a, - 0x325e45, - 0x3559c6, - 0x39ec43, - 0x39ec44, - 0x202c02, - 0x31c743, - 0x6560f042, - 0x30b743, - 0x38e544, - 0x2b2484, - 0x38668a, - 0x214c43, - 0x277208, - 0x250f0a, - 0x23f747, - 0x300986, - 0x260484, - 0x21c242, - 0x2a3702, - 0x65a02982, - 0x238b83, - 0x25ab47, - 0x202987, - 0x284d04, - 0x3a4f87, - 0x2e4e06, - 0x221287, - 0x303304, - 0x399d85, - 0x292705, - 0x65e1b2c2, - 0x3c50c6, - 0x223443, - 0x22a4c2, - 0x22a4c6, - 0x66222342, - 0x6660e982, - 0x3bb945, - 0x66a27882, - 0x66e01c42, - 0x334e85, - 0x2c51c5, - 0x355a85, - 0x289043, - 0x244185, - 0x2d9907, - 0x2feb45, - 0x370005, - 0x33d8c4, - 0x310806, - 0x381d44, - 0x67202a82, - 0x67ee7585, - 0x2a3ac7, - 0x34f408, - 0x261146, - 0x26114d, - 0x2674c9, - 0x2674d2, - 0x300585, - 0x309f43, - 0x6820c202, - 0x30ee44, - 0x39eb83, - 0x33b0c5, - 0x2fdf05, - 0x68630882, - 0x25f803, - 0x68a51b02, - 0x692d6142, - 0x69602242, - 0x2a1d45, - 0x394183, - 0x3c4f08, - 0x69a0ad42, - 0x69e0c842, - 0x26f446, - 0x317c4a, - 0x20e043, - 0x25d0c3, - 0x337d03, - 0x6aa04182, - 0x78e0c542, - 0x79600d82, - 0x206d02, - 0x35d3c9, - 0x2c18c4, - 0x2a9348, - 0x79af3dc2, - 0x79e01ac2, - 0x2ea5c5, - 0x238788, - 0x39e488, - 0x268b8c, - 0x23f683, - 0x7a263802, - 0x7a611d82, - 0x270d46, - 0x301805, - 0x2787c3, - 0x253c06, - 0x301946, - 0x29bc83, - 0x303b03, - 0x303f46, - 0x304b84, - 0x239a46, - 0x214a05, - 0x214a0a, - 0x24c1c4, - 0x305244, - 0x305b8a, - 0x7aa04982, - 0x24c345, - 0x30798a, - 0x308305, - 0x308bc4, - 0x308cc6, - 0x308e44, - 0x2198c6, - 0x7ae308c2, - 0x2f3446, - 0x341ac5, - 0x325cc7, - 0x3adf46, - 0x256844, - 0x2d2387, - 0x328446, - 0x241a05, - 0x241a07, - 0x3aed07, - 0x3aed0e, - 0x2ebb06, - 0x23ce05, - 0x203f87, - 0x20ff43, - 0x20ff47, - 0x217945, - 0x22f484, - 0x22f5c2, - 0x246087, - 0x2f9104, - 0x244dc4, - 0x290d4b, - 0x220003, - 0x2e58c7, - 0x220004, - 0x2e6047, - 0x2903c3, - 0x33ca0d, - 0x3998c8, - 0x2297c4, - 0x2701c5, - 0x30b205, - 0x30b643, - 0x7b228382, - 0x30d5c3, - 0x30da83, - 0x321c04, - 0x27ab85, - 0x3c5247, - 0x39ecc6, - 0x37c1c3, - 0x22a60b, - 0x30e04b, - 0x2a5c8b, - 0x2fa5cb, - 0x2bd60a, - 0x30548b, - 0x3245cb, - 0x360a8c, - 0x384f4b, - 0x3c4351, - 0x3c5d4a, - 0x30f5cb, - 0x30f88c, - 0x30fb8b, - 0x31010a, - 0x311bca, - 0x312bce, - 0x31324b, - 0x31350a, - 0x3145d1, - 0x314a0a, - 0x314f0b, - 0x31544e, - 0x315d8c, - 0x316b8b, - 0x316e4e, - 0x3171cc, - 0x318d4a, - 0x31a04c, - 0x7b71a34a, - 0x31af48, - 0x31ba49, - 0x32368a, - 0x32390a, - 0x323b8b, - 0x328b4e, - 0x328ed1, - 0x330349, - 0x33058a, - 0x330bcb, - 0x332a4a, - 0x333296, - 0x334b8b, - 0x33784a, - 0x33818a, - 0x33908b, - 0x33ab49, - 0x33d389, - 0x33de0d, - 0x33e48b, - 0x33f38b, - 0x33fd4b, - 0x343d49, - 0x34438e, - 0x34500a, - 0x34a4ca, - 0x34a7ca, - 0x34afcb, - 0x34b80b, - 0x34bacd, - 0x34d18d, - 0x34d7d0, - 0x34dc8b, - 0x34f9cc, - 0x34fe8b, - 0x351e0b, - 0x35344e, - 0x353a0b, - 0x353a0d, - 0x35964b, - 0x35a0cf, - 0x35a48b, - 0x35acca, - 0x35b3c9, - 0x35ba89, - 0x35cd4b, - 0x35d00e, - 0x35e88b, - 0x35f64f, - 0x36160b, - 0x3618cb, - 0x361b8b, - 0x36238a, - 0x367149, - 0x36a18f, - 0x36f54c, - 0x37038c, - 0x37108e, - 0x37158f, - 0x37194e, - 0x3722d0, - 0x3726cf, - 0x3731ce, - 0x373f8c, - 0x374292, - 0x375211, - 0x375a0e, - 0x375e8e, - 0x3763ce, - 0x37674f, - 0x376b0e, - 0x376e93, - 0x377351, - 0x37778c, - 0x377a8e, - 0x377f0c, - 0x378513, - 0x378ed0, - 0x37970c, - 0x379a0c, - 0x379ecb, - 0x37ac8e, - 0x37b18b, - 0x37b5cb, - 0x37ca4c, - 0x3825ca, - 0x38474c, - 0x384a4c, - 0x384d49, - 0x387e0b, - 0x3880c8, - 0x388889, - 0x38888f, - 0x38a08b, - 0x7bb8afca, - 0x38e8cc, - 0x38fa89, - 0x390a48, - 0x39100b, - 0x39158b, - 0x39220a, - 0x39248b, - 0x39298c, - 0x393d48, - 0x39a40b, - 0x39d80b, - 0x3a114e, - 0x3a27cb, - 0x3a410b, - 0x3ae88b, - 0x3aeb49, - 0x3af08d, - 0x3b368a, - 0x3b45d7, - 0x3b5cd8, - 0x3b9749, - 0x3bb58b, - 0x3bc1d4, - 0x3bc6cb, - 0x3bcc4a, - 0x3bd14a, - 0x3bd3cb, - 0x3bf610, - 0x3bfa11, - 0x3c00ca, - 0x3c394d, - 0x3c404d, - 0x3c61cb, - 0x3c6a06, - 0x3c51c3, - 0x7bf74a03, - 0x2dd1c6, - 0x245a05, - 0x252087, - 0x324486, - 0x1601182, - 0x2cbe89, - 0x31abc4, - 0x2d8988, - 0x220a83, - 0x30ed87, - 0x201c02, - 0x2acb03, - 0x7c200dc2, - 0x2c4946, - 0x2c5c84, - 0x21a604, - 0x349a43, - 0x349a45, - 0x7cac2742, - 0x7cea8044, - 0x273487, - 0x7d22f442, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x20ae43, - 0x200742, - 0xcd588, - 0x202c42, - 0x30e843, - 0x21f743, - 0x20ec83, - 0xae43, - 0x241d03, - 0x207c03, - 0x32eb56, - 0x356d13, - 0x3a4e09, - 0x38c988, - 0x340109, - 0x307b06, - 0x33e750, - 0x248c93, - 0x2f4ac8, - 0x2a5687, - 0x2b6f87, - 0x278c8a, - 0x38e5c9, - 0x342549, - 0x28b30b, - 0x31ca86, - 0x20850a, - 0x225f06, - 0x31a7c3, - 0x298445, - 0x207808, - 0x3abb0d, - 0x2ef1cc, - 0x35cac7, - 0x312f0d, - 0x21d0c4, - 0x23478a, - 0x23598a, - 0x235e4a, - 0x21fa07, - 0x243507, - 0x245fc4, - 0x27b986, - 0x3264c4, - 0x2e01c8, - 0x2ee049, - 0x2bfcc6, - 0x2bfcc8, - 0x24944d, - 0x2c3909, - 0x389ac8, - 0x3a2547, - 0x2f1c0a, - 0x2564c6, - 0x260fc7, - 0x306a04, - 0x214707, - 0x3105ca, - 0x378a0e, - 0x229985, - 0x3bfe0b, - 0x300389, - 0x267709, - 0x2b0587, - 0x3694ca, - 0x224dc7, - 0x2ff3c9, - 0x31e5c8, - 0x239e8b, - 0x2d8345, - 0x2309ca, - 0x227b09, - 0x3abe0a, - 0x2c64cb, - 0x21460b, - 0x28b095, - 0x306745, - 0x3a25c5, - 0x2e9d0a, - 0x2a6f0a, - 0x300107, - 0x2388c3, - 0x2e2348, - 0x2cf00a, - 0x228486, - 0x25b789, - 0x285048, - 0x2e9f84, - 0x29bb49, - 0x2be888, - 0x36c607, - 0x2e7586, - 0x2a3ac7, - 0x2ac6c7, - 0x2450c5, - 0x2297cc, - 0x2701c5, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x202c42, - 0x20be03, - 0x20ec83, - 0x20ae43, - 0x241d03, - 0x20be03, - 0x20ec83, - 0xae43, - 0x285603, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0xae43, - 0x241d03, - 0xcd588, - 0x202c42, - 0x209d42, - 0x236082, - 0x201102, - 0x2013c2, - 0x2db482, - 0x460be03, - 0x237583, - 0x203d43, - 0x30e843, - 0x214bc3, - 0x21f743, - 0x2d1206, - 0x20ec83, - 0x241d03, - 0x238843, - 0xcd588, - 0x323584, - 0x322dc7, - 0x34a403, - 0x2402c4, - 0x21b903, - 0x283cc3, - 0x30e843, - 0x15da87, - 0x1221c4, - 0x121b83, - 0xf45, - 0x200742, - 0xb6c03, - 0x5a02c42, - 0x1488d09, - 0x891cd, - 0x8950d, - 0x236082, - 0x6ff84, - 0xf89, - 0x200342, - 0x5f8d588, - 0xe9484, - 0xcd588, - 0x1426502, - 0x1508546, - 0x233603, - 0x2b8283, - 0x660be03, - 0x234784, - 0x6a37583, - 0x6f0e843, - 0x207d02, - 0x26ff84, - 0x20ec83, - 0x2fbbc3, - 0x2056c2, - 0x241d03, - 0x21c4c2, - 0x2f1f03, - 0x200a02, - 0x29d2c3, - 0x26f883, - 0x20fc42, - 0xcd588, - 0x233603, - 0x2fbbc3, - 0x2056c2, - 0x2f1f03, - 0x200a02, - 0x29d2c3, - 0x26f883, - 0x20fc42, - 0x2f1f03, - 0x200a02, - 0x29d2c3, - 0x26f883, - 0x20fc42, - 0x20be03, - 0x2b6c03, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x214bc3, - 0x21f743, - 0x226444, - 0x20ec83, - 0x241d03, - 0x204bc2, - 0x219543, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x2b6c03, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x20ec83, - 0x241d03, - 0x2f5805, - 0x230882, - 0x200742, - 0xcd588, - 0x1455908, - 0x1367ca, - 0x30e843, - 0x200001, - 0x202081, - 0x200ec1, - 0x200f01, - 0x200f41, - 0x20d701, - 0x312181, - 0x203801, - 0x24b241, - 0x2021c1, - 0x200101, - 0x200301, - 0x117485, - 0xcd588, - 0x200781, - 0x2014c1, - 0x200041, - 0x200141, - 0x201401, - 0x200901, - 0x200541, - 0x200c01, - 0x200a81, - 0x200641, - 0x200081, - 0x2001c1, - 0x200341, - 0x201681, - 0x20ab41, - 0x2002c1, - 0x200a01, - 0x200401, - 0x200441, - 0x201ac1, - 0x203f81, - 0x20d601, - 0x201181, - 0x200dc1, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x202c42, - 0x20be03, - 0x237583, - 0x200342, - 0x241d03, - 0x15da87, - 0x1f847, - 0x29546, - 0x4160a, - 0x88348, - 0x5a588, - 0x5aa47, - 0x86, - 0xd61c5, - 0x14a345, - 0x7dac6, - 0x157206, - 0x28b304, - 0x340547, - 0xcd588, - 0x2d2484, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x31a548, - 0x31e084, - 0x2374c4, - 0x22aa44, - 0x270c47, - 0x2cde07, - 0x20be03, - 0x23d74b, - 0x31b7ca, - 0x31cd47, - 0x2fc048, - 0x3255c8, - 0x237583, - 0x346c47, - 0x203d43, - 0x37c208, - 0x335049, - 0x26ff84, - 0x214bc3, - 0x2dce48, - 0x21f743, - 0x2cb6ca, - 0x2d1206, - 0x3a0547, - 0x20ec83, - 0x2da606, - 0x309308, - 0x241d03, - 0x28d806, - 0x2e394d, - 0x2e49c8, - 0x2ea0cb, - 0x331e86, - 0x348087, - 0x20f605, - 0x2ef98a, - 0x22bfc5, - 0x36210a, - 0x230882, - 0x203f83, - 0x244dc4, - 0x2021c6, - 0x3a7a03, - 0x2af043, - 0x24be43, - 0x23b003, - 0x349183, - 0x200582, - 0x2d7285, - 0x2a6589, - 0x245743, - 0x205583, - 0x202fc3, - 0x200301, - 0x2a1a85, - 0x39da83, - 0x2053c3, - 0x22aa44, - 0x326443, - 0x214948, - 0x2ec443, - 0x302e8d, - 0x2ebbc8, - 0x21ab46, - 0x31c783, - 0x378983, - 0x381cc3, - 0xaa0be03, - 0x236dc8, - 0x23d744, - 0x246d03, - 0x2022c6, - 0x249bc8, - 0x202e03, - 0x2ef9c3, - 0x2319c3, - 0x237583, - 0x21d8c3, - 0x21e903, - 0x21a303, - 0x31c703, - 0x2b25c3, - 0x225783, - 0x370645, - 0x256c04, - 0x258107, - 0x329982, - 0x25a303, - 0x25d486, - 0x25ed03, - 0x25f3c3, - 0x276543, - 0x202043, - 0x323283, - 0x269687, - 0xaf0e843, - 0x2363c3, - 0x2096c3, - 0x204d03, - 0x26ff83, - 0x2f3783, - 0x374ac5, - 0x363fc3, - 0x246889, - 0x20b0c3, - 0x2fe203, - 0xb2527c3, - 0x286d03, - 0x21cd08, - 0x2a64c6, - 0x200706, - 0x29aa46, - 0x27a5c7, - 0x200c83, - 0x20be83, - 0x21f743, - 0x288446, - 0x21a202, - 0x29ea43, - 0x32dd05, - 0x20ec83, - 0x2a2e47, - 0x160ae43, - 0x24e483, - 0x21fa83, - 0x225e03, - 0x241d03, - 0x212e46, - 0x31d286, - 0x36aa43, - 0x22ba83, - 0x219543, - 0x253743, - 0x303b83, - 0x2f0603, - 0x2f20c3, - 0x34f085, - 0x24f3c3, - 0x2d3246, - 0x23eb08, - 0x225b43, - 0x341789, - 0x33a308, - 0x2110c8, - 0x21a185, - 0x32a38a, - 0x35400a, - 0x37cd8b, - 0x37d408, - 0x2fb903, - 0x2f2103, - 0x33b1c3, - 0x366d88, - 0x2f4e83, - 0x39ec44, - 0x261983, - 0x202983, - 0x22d483, - 0x26fcc3, - 0x238843, - 0x230882, - 0x22d0c3, - 0x23f683, - 0x305403, - 0x3065c4, - 0x244dc4, - 0x3be143, - 0xcd588, - 0x200742, - 0x200602, - 0x200582, - 0x203402, - 0x2023c2, - 0x200782, - 0x238c02, - 0x201b02, - 0x202542, - 0x2000c2, - 0x225242, - 0x20d682, - 0x26cd82, - 0x206f02, - 0x2db482, - 0x20b182, - 0x201f82, - 0x2057c2, - 0x2f5f42, - 0x208102, - 0x200982, - 0x219e82, - 0x207bc2, - 0x207c02, - 0x201282, - 0x20fd82, - 0x201c42, - 0x742, - 0x602, - 0x582, - 0x3402, - 0x23c2, - 0x782, - 0x38c02, - 0x1b02, - 0x2542, - 0xc2, - 0x25242, - 0xd682, - 0x6cd82, - 0x6f02, - 0xdb482, - 0xb182, - 0x1f82, - 0x57c2, - 0xf5f42, - 0x8102, - 0x982, - 0x19e82, - 0x7bc2, - 0x7c02, - 0x1282, - 0xfd82, - 0x1c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x3f82, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x202c42, - 0x241d03, - 0xc60be03, - 0x30e843, - 0x21f743, - 0xaff03, - 0x223b82, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0xaff03, - 0x241d03, - 0xdc2, - 0x142f49, - 0x202382, - 0x15bda05, - 0x2eaa02, - 0xcd588, - 0x2c42, - 0x23bfc2, - 0x200482, - 0x244482, - 0x21f682, - 0x253982, - 0x14a345, - 0x203082, - 0x2056c2, - 0x20c502, - 0x203042, - 0x20b182, - 0x392a82, - 0x20ed42, - 0x24eb42, - 0x15da87, - 0x120a8d, - 0xd6249, - 0x6898b, - 0xd97c8, - 0x60b89, - 0xfeec6, - 0x30e843, - 0xcd588, - 0x1221c4, - 0x121b83, - 0xf45, - 0xcd588, - 0x5b646, - 0xf89, - 0xab07, - 0x200742, - 0x28b304, - 0x202c42, - 0x20be03, - 0x209d42, - 0x237583, - 0x202542, - 0x2d2484, - 0x214bc3, - 0x254ac2, - 0x20ec83, - 0x200342, - 0x241d03, - 0x3a25c6, - 0x32414f, - 0x70ec03, - 0xcd588, - 0x202c42, - 0x203d43, - 0x30e843, - 0x21f743, - 0xae43, - 0x14ef74b, - 0x141650a, - 0x14eca47, - 0x78d4b, - 0xd7e45, - 0x15da87, - 0x202c42, - 0x20be03, - 0x30e843, - 0x20ec83, - 0x200742, - 0x211a42, - 0x209342, - 0xfe0be03, - 0x2442c2, - 0x237583, - 0x226d02, - 0x22ab02, - 0x30e843, - 0x25cd82, - 0x251942, - 0x2a8002, - 0x211742, - 0x28d302, - 0x2029c2, - 0x200902, - 0x2ebfc2, - 0x278142, - 0x25c982, - 0x2ad9c2, - 0x2fcdc2, - 0x223482, - 0x23d082, - 0x21f743, - 0x205a02, - 0x20ec83, - 0x211e82, - 0x2c9fc2, - 0x241d03, - 0x2457c2, - 0x207c02, - 0x209f82, - 0x203a42, - 0x204e02, - 0x2da942, - 0x21b2c2, - 0x251b02, - 0x2234c2, - 0x31350a, - 0x35acca, - 0x38bf0a, - 0x3c6b82, - 0x20f2c2, - 0x374a82, - 0x103358c9, - 0x1072f70a, - 0x14328c7, - 0x10a03fc2, - 0x1410983, - 0x3342, - 0x12f70a, - 0x253404, - 0x1120be03, - 0x237583, - 0x254a04, - 0x30e843, - 0x26ff84, - 0x214bc3, - 0x21f743, - 0x20ec83, - 0x1aec5, - 0x20ae43, - 0x241d03, - 0x24f3c3, - 0x203f83, - 0xcd588, - 0x1400004, - 0x149845, - 0x142f49, - 0xa8ca, - 0x119fc2, - 0x19cbc6, - 0x187251, - 0x11b358c9, - 0x1498c8, - 0x1c1948, - 0x1fbc7, - 0x282, - 0x11748b, - 0x18c40a, - 0x844a, - 0x2aa47, - 0xcd588, - 0x10c788, - 0xd547, - 0x18419a4b, - 0x1c787, - 0x4c2, - 0x5e87, - 0x23a8a, - 0x1f8cf, - 0x8308f, - 0xefa02, - 0x2c42, - 0x173908, - 0xf698a, - 0x12b48, - 0x5fcc8, - 0xd3708, - 0x2e02, - 0x1bda8f, - 0x9dc8b, - 0x7e948, - 0x3c2c7, - 0x127c0a, - 0xf400b, - 0x78449, - 0x127b07, - 0x12a48, - 0x1541cc, - 0x3a347, - 0x17a28a, - 0x67008, - 0xf6f8e, - 0x2954e, - 0x2a88b, - 0x2e28b, - 0x30e8b, - 0x50b89, - 0xe32cb, - 0xeb5cd, - 0x17d18b, - 0x198c8d, - 0x19900d, - 0x3cc4a, - 0x44c0b, - 0x4638b, - 0x49ec5, - 0x18828e50, - 0x15770f, - 0x3b4cf, - 0xfb1cd, - 0x39610, - 0xa6c2, - 0x18e071c8, - 0x1f6c8, - 0x192ec405, - 0x5400b, - 0x12e350, - 0x59c88, - 0x12c4a, - 0x2e449, - 0x66007, - 0x66347, - 0x66507, - 0x66887, - 0x67347, - 0x67947, - 0x68187, - 0x68547, - 0x68e87, - 0x69187, - 0x69847, - 0x69a07, - 0x69bc7, - 0x69d87, - 0x6a087, - 0x6a687, - 0x6af47, - 0x6b707, - 0x6bcc7, - 0x6bf87, - 0x6c147, - 0x6c447, - 0x6cc47, - 0x6ce47, - 0x6dd87, - 0x6df47, - 0x6e107, - 0x6ebc7, - 0x6f0c7, - 0x6fd87, - 0x70687, - 0x71147, - 0x71647, - 0x71807, - 0x71c07, - 0x72447, - 0x726c7, - 0x72ac7, - 0x72c87, - 0x72e47, - 0x73287, - 0x73e87, - 0x743c7, - 0x74947, - 0x74b07, - 0x74e87, - 0x75407, - 0xd0c2, - 0x5fdca, - 0xdc547, - 0x84785, - 0xb3111, - 0x10ac6, - 0x10cc0a, - 0x17378a, - 0x5b646, - 0xcb0b, - 0x1402, - 0x34111, - 0xb29c9, - 0x948c9, - 0xebfc2, - 0x71e8a, - 0xa5a89, - 0xa61cf, - 0xa67ce, - 0xa7708, - 0x552c2, - 0x549, - 0x18b4ce, - 0xfc6cc, - 0xdbe0f, - 0x1a814e, - 0x1840c, - 0x25589, - 0x26751, - 0x2f988, - 0x1109d2, - 0x1115cd, - 0x1545cd, - 0x43f8b, - 0x4bad5, - 0x52c49, - 0x5438a, - 0x5ee89, - 0x6b310, - 0x7cc8b, - 0x85ecf, - 0xf0c0b, - 0x16130c, - 0x1b2610, - 0x9208a, - 0x9e90d, - 0x9fc4e, - 0xa9bca, - 0xab6cc, - 0xac394, - 0xb2651, - 0xbb04b, - 0xe1ecf, - 0xca50d, - 0x1a720e, - 0x16c4cc, - 0x18618c, - 0xb234b, - 0xb428e, - 0xb4d50, - 0xb584b, - 0xbaa8d, - 0xbb4cf, - 0xbef4c, - 0xbfb4e, - 0xc0411, - 0xdff4c, - 0x10d8c7, - 0xc738d, - 0xd000c, - 0xd65d0, - 0xdb80d, - 0x18acc7, - 0xe6310, - 0xf9348, - 0xfd44b, - 0x17d9cf, - 0x142188, - 0x10ce0d, - 0x190c10, - 0xf5f89, - 0x196af346, - 0xb0303, - 0xb5b05, - 0x9602, - 0x143709, - 0x5c40a, - 0x106606, - 0x2098a, - 0x1991f309, - 0x264c3, - 0xd2711, - 0xd2b49, - 0xd3ec7, - 0x1873cb, - 0xdae90, - 0xdb34c, - 0xdc2c8, - 0xdcc45, - 0x11e748, - 0x1afe8a, - 0x26587, - 0x140947, - 0x1382, - 0x12f04a, - 0x3b809, - 0x71505, - 0xa2cca, - 0x8a0cf, - 0x4794b, - 0x174b8c, - 0x1a252, - 0x9df05, - 0xdf0c8, - 0x13a60a, - 0x19ee8f05, - 0x17478c, - 0x12c443, - 0x192a82, - 0xf258a, - 0x14f2d8c, - 0x3a6c8, - 0x198e48, - 0x15da07, - 0x16f02, - 0xa02, - 0x49fd0, - 0x653c7, - 0x1282, - 0x333cf, - 0x7dac6, - 0x79a8e, - 0xdeb8b, - 0x6e308, - 0xa9dc9, - 0xf5012, - 0x18998d, - 0x1be608, - 0x68849, - 0x6a20d, - 0x6c5c9, - 0x6c98b, - 0x6e4c8, - 0x73c88, - 0x76248, - 0x79dc9, - 0x79fca, - 0x7b48c, - 0x17010a, - 0x103bc7, - 0x2fdcd, - 0xf7a8b, - 0x11a9cc, - 0x1979c8, - 0x4d3c9, - 0x13d8d0, - 0xc842, - 0x521cd, - 0x4182, - 0xc542, - 0x103b0a, - 0x10cb0a, - 0x10ec8b, - 0x4654c, - 0x10c28a, - 0x10c50e, - 0x121ccd, - 0xb6a08, - 0xdc2, - 0x11e0340e, - 0x1272184e, - 0x12f4960a, - 0x13742c0e, - 0x13f374ce, - 0x147ac40c, - 0x14328c7, - 0x14328c9, - 0x1410983, - 0x14eb784c, - 0x15727309, - 0x15f69bc9, - 0x1660a6c9, - 0x3342, - 0x3351, - 0x121791, - 0x14954d, - 0x142b51, - 0x137411, - 0x1ac34f, - 0xb778f, - 0x12724c, - 0x169b0c, - 0xa60c, - 0x1654cd, - 0x10e595, - 0x5a00c, - 0x1ba48c, - 0x138c90, - 0x155e8c, - 0x15dc0c, - 0x17a659, - 0x180a19, - 0x19f3d9, - 0x1b57d4, - 0x1bbcd4, - 0x3ed4, - 0x4ed4, - 0xb814, - 0x16e5a0c9, - 0x17404189, - 0x17fba549, - 0x1222fb89, - 0x3342, - 0x12a2fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x1322fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x13a2fb89, - 0x3342, - 0x1422fb89, - 0x3342, - 0x14a2fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x1522fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x15a2fb89, - 0x3342, - 0x1622fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x16a2fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x1722fb89, - 0x3342, - 0x17a2fb89, - 0x3342, - 0x1822fb89, - 0x3342, - 0x3eca, - 0x3342, - 0x187245, - 0x18c404, - 0x340e, - 0x12184e, - 0x14960a, - 0x142c0e, - 0x1374ce, - 0x1ac40c, - 0xb784c, - 0x127309, - 0x169bc9, - 0xa6c9, - 0x5a0c9, - 0x4189, - 0x1ba549, - 0x10e78d, - 0x5189, - 0xbac9, - 0x116a84, - 0x118c44, - 0x13aa44, - 0x18e7c4, - 0x79004, - 0x98884, - 0x477c4, - 0x143c44, - 0x1fbc4, - 0x157cd03, - 0xa6c2, - 0x121cc3, - 0x2e02, - 0x200742, - 0x202c42, - 0x209d42, - 0x208882, - 0x202542, - 0x200342, - 0x200a02, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff83, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x20ec83, - 0x241d03, - 0x1a9c3, - 0x30e843, - 0x6ff84, - 0x200742, - 0x2b6c03, - 0x1be0be03, - 0x2394c7, - 0x30e843, - 0x20f003, - 0x226444, - 0x20ec83, - 0x241d03, - 0x22d50a, - 0x3a25c5, - 0x219543, - 0x20e982, - 0xcd588, - 0xcd588, - 0x2c42, - 0x129242, - 0x1c74660b, - 0x5fc5, - 0x1f8c5, - 0xf9fc6, - 0x1221c4, - 0x121b83, - 0xf45, - 0x117485, - 0xcd588, - 0x1c787, - 0xbe03, - 0x1ce41447, - 0x143146, - 0x1d149445, - 0x143207, - 0xf84a, - 0xf708, - 0x13407, - 0x68348, - 0x98647, - 0xf28f, - 0x47f87, - 0x4e786, - 0x12e350, - 0x12cf0f, - 0x1c009, - 0x106684, - 0x1d5432ce, - 0xa978c, - 0xf420a, - 0x785c7, - 0xd9f8a, - 0x11e909, - 0xada0c, - 0x1bdf0a, - 0x5cc0a, - 0xf89, - 0x106606, - 0x7868a, - 0x11d84a, - 0x9a209, - 0xd1fc8, - 0xd22c6, - 0xd6c0d, - 0xb7cc5, - 0xab07, - 0xfb709, - 0x1a3207, - 0x10bd94, - 0xfdb4b, - 0x7e78a, - 0xa358d, - 0xf283, - 0xf283, - 0x29546, - 0xf283, - 0xb6c03, - 0xcd588, - 0x2c42, - 0x54a04, - 0x5da83, - 0xf5805, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x205583, - 0x20be03, - 0x237583, - 0x203d43, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x294a83, - 0x203f83, - 0x205583, - 0x28b304, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x235cc3, - 0x20be03, - 0x237583, - 0x208883, - 0x203d43, - 0x30e843, - 0x26ff84, - 0x3c32c3, - 0x20be83, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x219543, - 0x20c283, - 0x1f20be03, - 0x237583, - 0x24e683, - 0x30e843, - 0x211343, - 0x20be83, - 0x241d03, - 0x2057c3, - 0x317f04, - 0xcd588, - 0x1fa0be03, - 0x237583, - 0x2a77c3, - 0x30e843, - 0x21f743, - 0x226444, - 0x20ec83, - 0x241d03, - 0x232f43, - 0xcd588, - 0x2020be03, - 0x237583, - 0x203d43, - 0x20ae43, - 0x241d03, - 0xcd588, - 0x14328c7, - 0x2b6c03, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x226444, - 0x20ec83, - 0x241d03, - 0x142f49, - 0x117485, - 0x15da87, - 0x10bfcb, - 0xd2f44, - 0xb7cc5, - 0x1455908, - 0xa7e0d, - 0x21676405, - 0x8f204, - 0x10ec3, - 0xf5e85, - 0x31cc45, - 0xcd588, - 0xf282, - 0x3a283, - 0xefec6, - 0x31b0c8, - 0x397247, - 0x28b304, - 0x346046, - 0x3699c6, - 0xcd588, - 0x312ec3, - 0x23aec9, - 0x265555, - 0x6555f, - 0x20be03, - 0x3ba052, - 0x10db06, - 0x14fc85, - 0x12c4a, - 0x2e449, - 0x3b9e0f, - 0x2d2484, - 0x225285, - 0x2fdfd0, - 0x38cb87, - 0x20ae43, - 0x310f08, - 0x157146, - 0x2a47ca, - 0x22d244, - 0x2e8943, - 0x3a25c6, - 0x20e982, - 0x3987cb, - 0xae43, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x2f0ec3, - 0x202c42, - 0xee203, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x20f003, - 0x227b03, - 0x241d03, - 0x202c42, - 0x20be03, - 0x237583, - 0x20ec83, - 0xae43, - 0x241d03, - 0x200742, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x1f8c5, - 0x28b304, - 0x20be03, - 0x237583, - 0x21a484, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0xaff03, - 0x241d03, - 0x20be03, - 0x237583, - 0x203d43, - 0x204d03, - 0x21f743, - 0x20ec83, - 0xae43, - 0x241d03, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x357d43, - 0x3cf83, - 0xf003, - 0x20ec83, - 0x241d03, - 0x31350a, - 0x333049, - 0x3524cb, - 0x352b4a, - 0x35acca, - 0x3686cb, - 0x37bfca, - 0x3825ca, - 0x38bf0a, - 0x38c18b, - 0x3afbc9, - 0x3b1a0a, - 0x3b1d8b, - 0x3bc98b, - 0x3c5b0a, - 0x20be03, - 0x237583, - 0x203d43, - 0x21f743, - 0x20ec83, - 0xae43, - 0x241d03, - 0x18754b, - 0x60308, - 0x14f209, - 0xcd588, - 0x20be03, - 0x266004, - 0x206302, - 0x226444, - 0x201485, - 0x205583, - 0x28b304, - 0x20be03, - 0x23d744, - 0x237583, - 0x254a04, - 0x2d2484, - 0x26ff84, - 0x20be83, - 0x20ec83, - 0x241d03, - 0x252385, - 0x235cc3, - 0x219543, - 0x2b5d83, - 0x2702c4, - 0x2020c4, - 0x3c0885, - 0xcd588, - 0x320f04, - 0x39db46, - 0x2010c4, - 0x202c42, - 0x38fd87, - 0x256087, - 0x252944, - 0x25e5c5, - 0x2dba05, - 0x232585, - 0x26ff84, - 0x27a688, - 0x23c806, - 0x3c5f88, - 0x278185, - 0x2d8345, - 0x251304, - 0x241d03, - 0x2e9484, - 0x367486, - 0x3a26c3, - 0x2702c4, - 0x362205, - 0x26e984, - 0x23fd84, - 0x20e982, - 0x397746, - 0x3a4b06, - 0x301805, - 0x200742, - 0x2b6c03, - 0x27e02c42, - 0x207344, - 0x202542, - 0x21f743, - 0x23a3c2, - 0x20ec83, - 0x200342, - 0x207c03, - 0x203f83, - 0xcd588, - 0xcd588, - 0x30e843, - 0x200742, - 0x28a02c42, - 0x30e843, - 0x2574c3, - 0x3c32c3, - 0x2168c4, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x200742, - 0x29202c42, - 0x20be03, - 0x20ec83, - 0xae43, - 0x241d03, - 0x982, - 0x20c202, - 0x230882, - 0x20f003, - 0x2e2d83, - 0x200742, - 0x117485, - 0xcd588, - 0x15da87, - 0x202c42, - 0x237583, - 0x254a04, - 0x206c03, - 0x30e843, - 0x204d03, - 0x21f743, - 0x20ec83, - 0x207783, - 0x241d03, - 0x2388c3, - 0xb5cd3, - 0x1b9994, - 0x15da87, - 0x102dc6, - 0x5c60b, - 0x29546, - 0x5a3c7, - 0x2809, - 0x195d4a, - 0x8820d, - 0x12078c, - 0x104fca, - 0x14a345, - 0xf888, - 0x7dac6, - 0x6ff06, - 0x157206, - 0x20a6c2, - 0x1c170c, - 0x18c5c7, - 0x282d1, - 0x20be03, - 0x682c5, - 0x8808, - 0x22644, - 0x2a507646, - 0xb3106, - 0xd95c6, - 0x8d5ca, - 0x19dac3, - 0x2aa48c44, - 0x27c5, - 0x15cc83, - 0x2ae38a07, - 0x1aec5, - 0xcbcc, - 0xed348, - 0x6f6cb, - 0x2b25168c, - 0x140d6c3, - 0xb8888, - 0x9db09, - 0x3ff48, - 0x14208c6, - 0x2b76d609, - 0xd7e4a, - 0x10d08, - 0xf9fc8, - 0x1fbc4, - 0x118b45, - 0x6f807, - 0x2ba6f803, - 0x2bf39c86, - 0x2c2e9d04, - 0x2c790207, - 0xf9fc4, - 0xf9fc4, - 0xf9fc4, - 0xf9fc4, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x200742, - 0x202c42, - 0x30e843, - 0x207d02, - 0x20ec83, - 0x241d03, - 0x207c03, - 0x37158f, - 0x37194e, - 0xcd588, - 0x20be03, - 0x49a07, - 0x237583, - 0x30e843, - 0x214bc3, - 0x20ec83, - 0x241d03, - 0x220443, - 0x322887, - 0x203d02, - 0x292889, - 0x200602, - 0x24a2cb, - 0x2cf44a, - 0x28d009, - 0x200182, - 0x3418c6, - 0x235295, - 0x24a415, - 0x236793, - 0x24a993, - 0x203942, - 0x222dc5, - 0x3ab48c, - 0x27410b, - 0x2a2205, - 0x203402, - 0x2f2542, - 0x37e706, - 0x200282, - 0x261bc6, - 0x212ecd, - 0x21ac4c, - 0x228ec4, - 0x200cc2, - 0x2149c2, - 0x310d88, - 0x2023c2, - 0x211446, - 0x35c704, - 0x235455, - 0x236913, - 0x2108c3, - 0x32508a, - 0x20df47, - 0x30eec9, - 0x2d9d07, - 0x314902, - 0x200882, - 0x3b4b46, - 0x2099c2, - 0xcd588, - 0x210702, - 0x200302, - 0x217a07, - 0x336087, - 0x21c485, - 0x2004c2, - 0x2da6c7, - 0x220488, - 0x204002, - 0x2f21c2, - 0x230502, - 0x203cc2, - 0x23e988, - 0x20bf83, - 0x25dc48, - 0x20bf8d, - 0x237c03, - 0x23bc48, - 0x237c0f, - 0x237fce, - 0x38feca, - 0x2d1311, - 0x2d1790, - 0x38360d, - 0x38394c, - 0x3452c7, - 0x325207, - 0x346109, - 0x228fc2, - 0x200782, - 0x25becc, - 0x25c1cb, - 0x203c02, - 0x2b2506, - 0x2010c2, - 0x202642, - 0x2efa02, - 0x202c42, - 0x231fc4, - 0x240647, - 0x230a42, - 0x245207, - 0x2475c7, - 0x21bc02, - 0x21b282, - 0x2498c5, - 0x204ac2, - 0x2e72ce, - 0x2a384d, - 0x237583, - 0x28400e, - 0x3b868d, - 0x348003, - 0x202ec2, - 0x2817c4, - 0x238c42, - 0x202e82, - 0x372a45, - 0x37b407, - 0x24d902, - 0x208882, - 0x254607, - 0x257688, - 0x329982, - 0x29df86, - 0x25bd4c, - 0x25c08b, - 0x212c42, - 0x26208f, - 0x262450, - 0x26284f, - 0x262c15, - 0x263154, - 0x26364e, - 0x2639ce, - 0x263d4f, - 0x26410e, - 0x264494, - 0x264993, - 0x264e4d, - 0x2755c9, - 0x289843, - 0x201802, - 0x215f45, - 0x206c06, - 0x202542, - 0x344e47, - 0x30e843, - 0x201402, - 0x36dfc8, - 0x2d1551, - 0x2d1990, - 0x200c42, - 0x270f87, - 0x205842, - 0x341287, - 0x209602, - 0x348f89, - 0x37e6c7, - 0x2a4b48, - 0x307486, - 0x2e2c83, - 0x326e05, - 0x20e402, - 0x202682, - 0x3b4f45, - 0x3c1485, - 0x200f82, - 0x214d03, - 0x26ea07, - 0x208007, - 0x2085c2, - 0x22e684, - 0x20b4c3, - 0x20b4c9, - 0x20f108, - 0x201542, - 0x204482, - 0x2e3547, - 0x33d705, - 0x293988, - 0x222a87, - 0x201cc3, - 0x298106, - 0x38348d, - 0x38380c, - 0x2e0646, - 0x200482, - 0x26f582, - 0x201e42, - 0x237a8f, - 0x237e8e, - 0x2dba87, - 0x200b82, - 0x3517c5, - 0x3517c6, - 0x203282, - 0x205a02, - 0x28ad86, - 0x292ac3, - 0x3411c6, - 0x2c3ec5, - 0x2c3ecd, - 0x2c4495, - 0x2c4e8c, - 0x2c59cd, - 0x2c5d92, - 0x20d682, - 0x26cd82, - 0x2047c2, - 0x21ce86, - 0x2fc586, - 0x201382, - 0x206c86, - 0x20c502, - 0x20d245, - 0x2013c2, - 0x2a3949, - 0x21d70c, - 0x21da4b, - 0x200342, - 0x258508, - 0x20cb42, - 0x206f02, - 0x271946, - 0x22fb05, - 0x31f507, - 0x250d85, - 0x2982c5, - 0x249a82, - 0x204c02, - 0x20b182, - 0x2dc107, - 0x24f4cd, - 0x24f84c, - 0x34f687, - 0x22bac2, - 0x201f82, - 0x23d488, - 0x343888, - 0x303d48, - 0x30cdc4, - 0x2b4507, - 0x2e3c83, - 0x24e882, - 0x204882, - 0x2e6b09, - 0x2f7387, - 0x2057c2, - 0x271d45, - 0x248602, - 0x209942, - 0x2bca43, - 0x2bca46, - 0x2f0602, - 0x2f1e82, - 0x201442, - 0x3b33c6, - 0x3454c7, - 0x205e42, - 0x200382, - 0x25da8f, - 0x283e4d, - 0x38b8ce, - 0x3b850c, - 0x2017c2, - 0x200502, - 0x3072c5, - 0x311d86, - 0x209002, - 0x208102, - 0x200982, - 0x222a04, - 0x2dcdc4, - 0x3c23c6, - 0x200a02, - 0x2b7307, - 0x231d03, - 0x231d08, - 0x2326c8, - 0x243e07, - 0x2ecbc6, - 0x204042, - 0x23e683, - 0x23e687, - 0x28a8c6, - 0x2f3045, - 0x30d148, - 0x206a42, - 0x341687, - 0x20fd82, - 0x2f4682, - 0x20c142, - 0x2f1149, - 0x20a402, - 0x201742, - 0x24adc3, - 0x325ec7, - 0x2040c2, - 0x21d88c, - 0x21db8b, - 0x2e06c6, - 0x35cbc5, - 0x227882, - 0x201c42, - 0x2ba046, - 0x22e983, - 0x331547, - 0x20cb82, - 0x202a82, - 0x235115, - 0x24a5d5, - 0x236653, - 0x24ab13, - 0x25d207, - 0x274548, - 0x274550, - 0x28744f, - 0x373ad3, - 0x28cdd2, - 0x292450, - 0x2b350f, - 0x2fd6d2, - 0x3af491, - 0x2af493, - 0x3938d2, - 0x2c3b0f, - 0x2cd74e, - 0x2cf252, - 0x2d09d1, - 0x2d3b0f, - 0x2d528e, - 0x2dc811, - 0x2dd7d0, - 0x2ed512, - 0x2f0f51, - 0x2f2206, - 0x2f3907, - 0x38e407, - 0x200d02, - 0x27efc5, - 0x3713c7, - 0x230882, - 0x20f6c2, - 0x22d0c5, - 0x200443, - 0x200446, - 0x24f68d, - 0x24f9cc, - 0x206d02, - 0x3ab30b, - 0x273fca, - 0x22358a, - 0x2b9489, - 0x2e530b, - 0x222bcd, - 0x2fe44c, - 0x2ec88a, - 0x27500c, - 0x294d4b, - 0x2a204c, - 0x2f968b, - 0x2b9e83, - 0x2f4f06, - 0x3b9942, - 0x2f3dc2, - 0x20e343, - 0x201ac2, - 0x207203, - 0x24ec86, - 0x262dc7, - 0x2ad706, - 0x2f6b48, - 0x343588, - 0x2ca146, - 0x211d82, - 0x3011cd, - 0x30150c, - 0x2d2547, - 0x304e07, - 0x23c242, - 0x219742, - 0x23e602, - 0x257a42, - 0x202c42, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x226444, - 0x20ec83, - 0x241d03, - 0x207c03, - 0x200742, - 0x201482, - 0x2e68ecc5, - 0x2ea8e4c5, - 0x2efb3086, - 0xcd588, - 0x2f2afb05, - 0x202c42, - 0x209d42, - 0x2f726285, - 0x2fa7cb85, - 0x2fe7d647, - 0x302867c9, - 0x30667d84, - 0x202542, - 0x201402, - 0x30b0dec5, - 0x30e95f49, - 0x31327988, - 0x316ac205, - 0x31af0707, - 0x31e21cc8, - 0x322def85, - 0x3266d246, - 0x32b6d849, - 0x32ed4ec8, - 0x332bf988, - 0x3369658a, - 0x33a75e04, - 0x33f7c545, - 0x342bc308, - 0x34727e05, - 0x217f42, - 0x34a061c3, - 0x34ea2606, - 0x35311408, - 0x356eee46, - 0x35b643c8, - 0x35eb6306, - 0x363c2f44, - 0x2032c2, - 0x366f1587, - 0x36aa8644, - 0x36e77e87, - 0x3723ecc7, - 0x200342, - 0x3769ae85, - 0x37a403c4, - 0x37ee1787, - 0x383a3387, - 0x38681606, - 0x38a7d205, - 0x38e96047, - 0x392e5a48, - 0x396162c7, - 0x39b94949, - 0x39ec51c5, - 0x3a2b4107, - 0x3a68e306, - 0x3aa941c8, - 0x227d8d, - 0x251989, - 0x272fcb, - 0x27ac8b, - 0x2a78cb, - 0x2da98b, - 0x311f8b, - 0x31224b, - 0x312889, - 0x31378b, - 0x313a4b, - 0x313fcb, - 0x314c8a, - 0x3151ca, - 0x3157cc, - 0x31938b, - 0x319dca, - 0x33080a, - 0x33b28e, - 0x33be8e, - 0x33c20a, - 0x33e14a, - 0x33eb4b, - 0x33ee0b, - 0x33fa8b, - 0x35edcb, - 0x35f3ca, - 0x36008b, - 0x36034a, - 0x3605ca, - 0x36084a, - 0x37d74b, - 0x3856cb, - 0x388f8e, - 0x38930b, - 0x391f4b, - 0x392e0b, - 0x39a6ca, - 0x39a949, - 0x39ab8a, - 0x39c6ca, - 0x3b06cb, - 0x3b204b, - 0x3b2a0a, - 0x3b390b, - 0x3b904b, - 0x3c554b, - 0x3ae7fd48, - 0x3b287989, - 0x3b69d989, - 0x3bad8988, - 0x34c805, - 0x200583, - 0x22a3c4, - 0x217c05, - 0x267ac6, - 0x26cfc5, - 0x286284, - 0x344d48, - 0x30b505, - 0x290604, - 0x2064c7, - 0x29cf0a, - 0x266b4a, - 0x2dbb87, - 0x20c4c7, - 0x2fd2c7, - 0x282187, - 0x2f8c45, - 0x3b6e46, - 0x386007, - 0x244e44, - 0x2df546, - 0x2df446, - 0x204745, - 0x3389c4, - 0x2975c6, - 0x29bfc7, - 0x22df06, - 0x27c8c7, - 0x250803, - 0x3912c6, - 0x234f05, - 0x27d747, - 0x26a84a, - 0x26e7c4, - 0x21bd88, - 0x2b8a49, - 0x2e0d07, - 0x319c46, - 0x258788, - 0x2ef589, - 0x30f084, - 0x33a484, - 0x29ef05, - 0x2ba648, - 0x2c2807, - 0x2b3e49, - 0x22dc08, - 0x2f2306, - 0x310806, - 0x297f88, - 0x362bc6, - 0x28e4c5, - 0x2816c6, - 0x278988, - 0x237986, - 0x25af0b, - 0x2c7c06, - 0x299b0d, - 0x369405, - 0x2a8506, - 0x21f085, - 0x331b49, - 0x3a6cc7, - 0x318308, - 0x2a1e46, - 0x298d89, - 0x33ffc6, - 0x26a7c5, - 0x24c486, - 0x288b86, - 0x2c6e49, - 0x31e2c6, - 0x29cc07, - 0x245e85, - 0x203983, - 0x25b085, - 0x299dc7, - 0x3ab746, - 0x369309, - 0x3b3086, - 0x26b146, - 0x213fc9, - 0x2810c9, - 0x29fac7, - 0x200908, - 0x2b2f49, - 0x27ec48, - 0x330a46, - 0x2d1d85, - 0x240c8a, - 0x26b1c6, - 0x239346, - 0x2cac05, - 0x2d4888, - 0x22b287, - 0x233f0a, - 0x254f86, - 0x251dc5, - 0x3324c6, - 0x224507, - 0x319b07, - 0x2835c5, - 0x26a985, - 0x395a06, - 0x3b8c06, - 0x2fa846, - 0x2bc7c4, - 0x280449, - 0x288806, - 0x2c814a, - 0x227248, - 0x36fd08, - 0x266b4a, - 0x212505, - 0x29bf05, - 0x2dd048, - 0x2c9688, - 0x233907, - 0x2ba946, - 0x32bf88, - 0x309507, - 0x27f348, - 0x2b5706, - 0x281e48, - 0x295586, - 0x278307, - 0x33a206, - 0x2975c6, - 0x22ecca, - 0x232046, - 0x2d1d89, - 0x2ee586, - 0x35c00a, - 0x3c2f49, - 0x27dd86, - 0x2b8304, - 0x21600d, - 0x287c07, - 0x239c06, - 0x2bf845, - 0x340045, - 0x37f306, - 0x2e15c9, - 0x2d4407, - 0x279406, - 0x306886, - 0x286309, - 0x2a3204, - 0x242544, - 0x3c2a88, - 0x24f046, - 0x271348, - 0x2e8008, - 0x29f447, - 0x3b6589, - 0x2faa47, - 0x2af9ca, - 0x2e79cf, - 0x31194a, - 0x3070c5, - 0x278bc5, - 0x218b05, - 0x35c647, - 0x2240c3, - 0x200b08, - 0x21e646, - 0x21e749, - 0x2d8646, - 0x2c8547, - 0x298b49, - 0x318208, - 0x2cacc7, - 0x30eb43, - 0x34c885, - 0x224045, - 0x2bc60b, - 0x327ec4, - 0x2d6884, - 0x276bc6, - 0x30f407, - 0x38f4ca, - 0x206247, - 0x20c347, - 0x27cb85, - 0x3c6485, - 0x282609, - 0x2975c6, - 0x2060cd, - 0x31e505, - 0x2b18c3, - 0x20b003, - 0x3a4d45, - 0x351305, - 0x258788, - 0x27a347, - 0x2422c6, - 0x29d606, - 0x22de05, - 0x237847, - 0x3c1d47, - 0x23c6c7, - 0x37c5ca, - 0x391388, - 0x2bc7c4, - 0x257bc7, - 0x27bb07, - 0x33f086, - 0x2692c7, - 0x2a1808, - 0x395f08, - 0x329b06, - 0x20c708, - 0x2cfbc4, - 0x386006, - 0x370d86, - 0x36bd46, - 0x277806, - 0x29b244, - 0x282246, - 0x2be246, - 0x297986, - 0x2060c6, - 0x20aec6, - 0x2a1646, - 0x2421c8, - 0x385a88, - 0x2cdac8, - 0x26d1c8, - 0x2dcfc6, - 0x212305, - 0x39e746, - 0x2ac285, - 0x396f87, - 0x22dcc5, - 0x213c03, - 0x38e045, - 0x33dd04, - 0x20b005, - 0x247643, - 0x33c4c7, - 0x30d708, - 0x27c986, - 0x2c930d, - 0x278b86, - 0x296f45, - 0x222083, - 0x2bbcc9, - 0x2a3386, - 0x291706, - 0x271e04, - 0x3118c7, - 0x23a1c6, - 0x2d46c5, - 0x21af83, - 0x3be4c4, - 0x27bcc6, - 0x3b6f44, - 0x370e88, - 0x3459c9, - 0x2317c9, - 0x29ed0a, - 0x2a05cd, - 0x2118c7, - 0x2391c6, - 0x20f984, - 0x2867c9, - 0x284ac8, - 0x287806, - 0x241906, - 0x2692c7, - 0x2d9346, - 0x22a046, - 0x347086, - 0x23ed4a, - 0x221cc8, - 0x22f885, - 0x2a2fc9, - 0x27f84a, - 0x2ff648, - 0x29b6c8, - 0x291688, - 0x29d24c, - 0x3124c5, - 0x29d888, - 0x385d86, - 0x24c9c6, - 0x35eb07, - 0x206145, - 0x281845, - 0x231689, - 0x2139c7, - 0x21e705, - 0x22aec7, - 0x20b003, - 0x2c2d45, - 0x2151c8, - 0x280d47, - 0x29b589, - 0x2e9f85, - 0x33e384, - 0x2a0288, - 0x2f16c7, - 0x2cae88, - 0x3aac88, - 0x2e1dc5, - 0x21e546, - 0x29d706, - 0x3a7009, - 0x2cb3c7, - 0x2ac8c6, - 0x206e87, - 0x239fc3, - 0x267d84, - 0x2cfcc5, - 0x2f3f84, - 0x246804, - 0x27ffc7, - 0x340d87, - 0x26dc84, - 0x29b3d0, - 0x31d507, - 0x3c6485, - 0x2561cc, - 0x224a04, - 0x2c4c88, - 0x278209, - 0x375886, - 0x240088, - 0x21ca84, - 0x276ec8, - 0x234506, - 0x22eb48, - 0x29a086, - 0x28854b, - 0x38ddc5, - 0x2cfb48, - 0x2173c4, - 0x345e0a, - 0x29b589, - 0x33a106, - 0x218bc8, - 0x25ed85, - 0x31dec4, - 0x2c4b86, - 0x23c588, - 0x27fd48, - 0x32c806, - 0x3c2344, - 0x240c06, - 0x2faac7, - 0x277d87, - 0x2692cf, - 0x205847, - 0x27de47, - 0x351685, - 0x35e345, - 0x29f789, - 0x382e46, - 0x27d885, - 0x2813c7, - 0x3934c8, - 0x2c7645, - 0x33a206, - 0x227088, - 0x2eee4a, - 0x3bf088, - 0x28ab07, - 0x2e7e06, - 0x2a2f86, - 0x202583, - 0x20de03, - 0x27fa09, - 0x2b2dc9, - 0x2c4a86, - 0x2e9f85, - 0x36bac8, - 0x218bc8, - 0x362d48, - 0x34710b, - 0x2c9547, - 0x309149, - 0x269548, - 0x350284, - 0x318648, - 0x28c889, - 0x2acbc5, - 0x35c547, - 0x267e05, - 0x27fc48, - 0x28eb4b, - 0x295d90, - 0x2a8145, - 0x21730c, - 0x242485, - 0x27cc03, - 0x2b1d06, - 0x2bd884, - 0x2404c6, - 0x29bfc7, - 0x227104, - 0x248688, - 0x2009cd, - 0x2dfc05, - 0x211904, - 0x28f244, - 0x28f249, - 0x2ae548, - 0x31bc47, - 0x234588, - 0x280508, - 0x279705, - 0x21f2c7, - 0x279687, - 0x23ac87, - 0x26a989, - 0x346dc9, - 0x272146, - 0x383b46, - 0x269506, - 0x33b6c5, - 0x3aa4c4, - 0x3bd646, - 0x3c4c46, - 0x279748, - 0x2241cb, - 0x26e687, - 0x20f984, - 0x23a106, - 0x2a1b47, - 0x335405, - 0x3583c5, - 0x223884, - 0x346d46, - 0x3bd6c8, - 0x2867c9, - 0x2091c6, - 0x2848c8, - 0x2d4786, - 0x350908, - 0x2ce58c, - 0x2795c6, - 0x296c0d, - 0x29708b, - 0x29ccc5, - 0x3c1e87, - 0x31e3c6, - 0x3199c8, - 0x2721c9, - 0x329dc8, - 0x3c6485, - 0x208947, - 0x27ed48, - 0x24ff89, - 0x2a5586, - 0x24da8a, - 0x319748, - 0x329c0b, - 0x2ccd8c, - 0x276fc8, - 0x27b286, - 0x21dfc8, - 0x2eeac7, - 0x205989, - 0x2f084d, - 0x2974c6, - 0x31dd48, - 0x385949, - 0x2bc8c8, - 0x281f48, - 0x2bec8c, - 0x2bff87, - 0x2c0a47, - 0x26a7c5, - 0x2b4807, - 0x393388, - 0x2c4c06, - 0x20904c, - 0x2ec1c8, - 0x2c8c48, - 0x250286, - 0x223dc7, - 0x272344, - 0x26d1c8, - 0x2b6d0c, - 0x28430c, - 0x307145, - 0x2047c7, - 0x3c22c6, - 0x223d46, - 0x331d08, - 0x367784, - 0x22df0b, - 0x2b744b, - 0x2e7e06, - 0x200847, - 0x322385, - 0x271285, - 0x22e046, - 0x25ed45, - 0x327e85, - 0x2c6c87, - 0x270a09, - 0x3b8dc4, - 0x25f405, - 0x2de045, - 0x2add08, - 0x2da405, - 0x287109, - 0x2c9ac7, - 0x2c9acb, - 0x24fbc6, - 0x241f09, - 0x338908, - 0x291f85, - 0x23ad88, - 0x346e08, - 0x2570c7, - 0x208e47, - 0x280049, - 0x22ea87, - 0x2aa389, - 0x2b7dcc, - 0x394848, - 0x2d4d09, - 0x2d6447, - 0x2805c9, - 0x340ec7, - 0x2cce88, - 0x3b6745, - 0x385f86, - 0x2bf888, - 0x30d3c8, - 0x27f709, - 0x327ec7, - 0x256d85, - 0x2301c9, - 0x201c46, - 0x28e304, - 0x326006, - 0x311288, - 0x328747, - 0x2243c8, - 0x20c7c9, - 0x325b87, - 0x29d0c6, - 0x3c1f44, - 0x38e0c9, - 0x21f148, - 0x250147, - 0x2adf86, - 0x224106, - 0x2392c4, - 0x26d846, - 0x20af83, - 0x38d949, - 0x38dd86, - 0x20ca45, - 0x29d606, - 0x2c7205, - 0x27f1c8, - 0x2ee987, - 0x2eb146, - 0x3262c6, - 0x36fd08, - 0x29f907, - 0x297505, - 0x29b1c8, - 0x3b2448, - 0x319748, - 0x242345, - 0x386006, - 0x231589, - 0x3a6e84, - 0x2c708b, - 0x229d4b, - 0x22f789, - 0x20b003, - 0x25cf45, - 0x22abc6, - 0x242cc8, - 0x34e904, - 0x27c986, - 0x37c709, - 0x2f0405, - 0x2c6bc6, - 0x2f16c6, - 0x20c984, - 0x2a86ca, - 0x20c988, - 0x30d3c6, - 0x2934c5, - 0x331287, - 0x351547, - 0x21e544, - 0x229f87, - 0x22dc84, - 0x22dc86, - 0x200b43, - 0x26a985, - 0x37dc85, - 0x364648, - 0x257d85, - 0x279309, - 0x26d007, - 0x26d00b, - 0x2a240c, - 0x2a2a0a, - 0x2f0707, - 0x205c83, - 0x2ebcc8, - 0x242505, - 0x2c76c5, - 0x34c944, - 0x2ccd86, - 0x278206, - 0x26d887, - 0x23f8cb, - 0x29b244, - 0x2d7404, - 0x2c2784, - 0x2c6986, - 0x227104, - 0x2ba748, - 0x34c745, - 0x258a45, - 0x362c87, - 0x3c1f89, - 0x351305, - 0x37f30a, - 0x245d89, - 0x2d698a, - 0x23ee89, - 0x3a5104, - 0x306945, - 0x2d9448, - 0x2e184b, - 0x29ef05, - 0x2f3206, - 0x247684, - 0x279846, - 0x325a09, - 0x2a1c47, - 0x3b3248, - 0x2a0946, - 0x2faa47, - 0x27fd48, - 0x37f886, - 0x334f84, - 0x371c87, - 0x361205, - 0x373507, - 0x21c984, - 0x31e346, - 0x2e5bc8, - 0x297248, - 0x2e44c7, - 0x24e388, - 0x295645, - 0x20ae44, - 0x266a48, - 0x24e484, - 0x208e45, - 0x2f8e44, - 0x309607, - 0x2888c7, - 0x280708, - 0x2cb006, - 0x257d05, - 0x279108, - 0x3bf288, - 0x29ec49, - 0x22a046, - 0x233f88, - 0x345c8a, - 0x335488, - 0x2def85, - 0x225446, - 0x245c48, - 0x208a0a, - 0x229207, - 0x285dc5, - 0x28e508, - 0x2cc2c4, - 0x2d4906, - 0x2c0dc8, - 0x20aec6, - 0x31fc48, - 0x25b247, - 0x2063c6, - 0x2b8304, - 0x2a6b87, - 0x2b0d44, - 0x3259c7, - 0x2a524d, - 0x22f805, - 0x2e13cb, - 0x284586, - 0x258608, - 0x248644, - 0x2ee246, - 0x27bcc6, - 0x21e307, - 0x2968cd, - 0x24b947, - 0x2b1808, - 0x286985, - 0x364e48, - 0x2c2786, - 0x2956c8, - 0x354486, - 0x336b47, - 0x2c5689, - 0x353e07, - 0x287ac8, - 0x2733c5, - 0x21c508, - 0x223c85, - 0x2f7505, - 0x23f105, - 0x24c4c3, - 0x277884, - 0x28e705, - 0x36d849, - 0x36b906, - 0x2a1908, - 0x208c05, - 0x2b46c7, - 0x29f14a, - 0x2c6b09, - 0x288a8a, - 0x2cdb48, - 0x22ad0c, - 0x28144d, - 0x34a703, - 0x31fb48, - 0x3be485, - 0x2eec06, - 0x318086, - 0x2deac5, - 0x206f89, - 0x3ab885, - 0x279108, - 0x25e046, - 0x3532c6, - 0x2a0149, - 0x3a0f87, - 0x28ee06, - 0x29f0c8, - 0x36bc48, - 0x2d8b87, - 0x2be3ce, - 0x2c29c5, - 0x24fe85, - 0x20adc8, - 0x3269c7, - 0x208f82, - 0x2be804, - 0x2403ca, - 0x250208, - 0x346f46, - 0x298c88, - 0x29d706, - 0x31da88, - 0x2ac8c8, - 0x2f74c4, - 0x2b4a85, - 0x6010c4, - 0x6010c4, - 0x6010c4, - 0x200a43, - 0x223f86, - 0x2795c6, - 0x29c98c, - 0x201343, - 0x21c986, - 0x200b04, - 0x2a3308, - 0x37c545, - 0x2404c6, - 0x2bc408, - 0x2cef86, - 0x2eb0c6, - 0x339f08, - 0x2cfd47, - 0x22e849, - 0x2a714a, - 0x211644, - 0x22dcc5, - 0x2b3e05, - 0x2c5406, - 0x211906, - 0x29c706, - 0x2f8686, - 0x22e984, - 0x22e98b, - 0x22d744, - 0x242085, - 0x2ab5c5, - 0x29f506, - 0x369808, - 0x281307, - 0x38dd04, - 0x2076c3, - 0x2cbdc5, - 0x22dac7, - 0x28120b, - 0x364547, - 0x2bc308, - 0x2b4bc7, - 0x26be06, - 0x251c48, - 0x26f24b, - 0x217b46, - 0x216b09, - 0x26f3c5, - 0x30eb43, - 0x2c6bc6, - 0x25b148, - 0x20c843, - 0x22dbc3, - 0x27fd46, - 0x29d706, - 0x36808a, - 0x27b2c5, - 0x27bb0b, - 0x29d54b, - 0x247b03, - 0x220043, - 0x2af944, - 0x2a88c7, - 0x25b1c4, - 0x240084, - 0x385c04, - 0x335788, - 0x293408, - 0x20dd89, - 0x2c5248, - 0x23f387, - 0x2060c6, - 0x2a154f, - 0x2c2b06, - 0x2cd084, - 0x29324a, - 0x22d9c7, - 0x2b0e46, - 0x28e349, - 0x20dd05, - 0x364785, - 0x20de46, - 0x21c643, - 0x2cc309, - 0x221e46, - 0x20c589, - 0x38f4c6, - 0x26a985, - 0x307545, - 0x205843, - 0x2a8a08, - 0x31be07, - 0x21e644, - 0x2a3188, - 0x24c744, - 0x39a286, - 0x2b1d06, - 0x2445c6, - 0x2cfa09, - 0x2c7645, - 0x2975c6, - 0x21afc9, - 0x393086, - 0x2a1646, - 0x395846, - 0x203a45, - 0x2f8e46, - 0x336b44, - 0x3b6745, - 0x2bf884, - 0x2b2206, - 0x31e4c4, - 0x200d03, - 0x284b85, - 0x238888, - 0x2509c7, - 0x34e989, - 0x285cc8, - 0x297d51, - 0x2f174a, - 0x2e7d47, - 0x396246, - 0x200b04, - 0x2bf988, - 0x26d9c8, - 0x297f0a, - 0x286ecd, - 0x24c486, - 0x33a006, - 0x2a6c46, - 0x283447, - 0x2b18c5, - 0x341987, - 0x2009c5, - 0x2c9c04, - 0x2a7586, - 0x26d6c7, - 0x2cc00d, - 0x245b87, - 0x344c48, - 0x279409, - 0x225346, - 0x2a5505, - 0x2fa084, - 0x311386, - 0x21e446, - 0x250386, - 0x299508, - 0x21d683, - 0x208d83, - 0x341f45, - 0x257e46, - 0x2ac885, - 0x2a0b48, - 0x29c18a, - 0x39e284, - 0x2a3308, - 0x291688, - 0x29f347, - 0x208cc9, - 0x2bc008, - 0x286847, - 0x385e86, - 0x20aeca, - 0x311408, - 0x3a6b09, - 0x2ae608, - 0x228089, - 0x396107, - 0x2fea85, - 0x347306, - 0x2c4a88, - 0x27a888, - 0x28de08, - 0x38ab08, - 0x242085, - 0x203bc4, - 0x236ec8, - 0x209784, - 0x23ec84, - 0x26a985, - 0x290647, - 0x3c1d49, - 0x21e107, - 0x214045, - 0x276dc6, - 0x35bc86, - 0x211a84, - 0x2a0486, - 0x257b44, - 0x2a11c6, - 0x3c1b06, - 0x2181c6, - 0x3c6485, - 0x2a0a07, - 0x205c83, - 0x216e89, - 0x36fb08, - 0x2866c4, - 0x2866cd, - 0x297348, - 0x2ddc88, - 0x3a6a86, - 0x2c5789, - 0x2c6b09, - 0x325705, - 0x29c28a, - 0x252a0a, - 0x25e6cc, - 0x25e846, - 0x277c06, - 0x2c2c86, - 0x372b09, - 0x2eee46, - 0x29f946, - 0x3ab946, - 0x26d1c8, - 0x24e386, - 0x2cca0b, - 0x2907c5, - 0x258a45, - 0x277e85, - 0x3c2806, - 0x20ae83, - 0x244546, - 0x245b07, - 0x2bf845, - 0x3108c5, - 0x340045, - 0x2f83c6, - 0x3257c4, - 0x327886, - 0x2bad89, - 0x3c268c, - 0x2c9948, - 0x23c504, - 0x2f8b46, - 0x284686, - 0x25b148, - 0x218bc8, - 0x3c2589, - 0x331287, - 0x24ed89, - 0x37ba46, - 0x230604, - 0x20d804, - 0x280344, - 0x27fd48, - 0x3c1b8a, - 0x351286, - 0x35e207, - 0x36e207, - 0x242005, - 0x2b3dc4, - 0x28c846, - 0x2b1906, - 0x23a283, - 0x36f947, - 0x3aab88, - 0x32584a, - 0x22ca48, - 0x3643c8, - 0x31e505, - 0x29cdc5, - 0x26e785, - 0x2423c6, - 0x243286, - 0x340cc5, - 0x38db89, - 0x2b3bcc, - 0x26e847, - 0x297f88, - 0x2dee05, - 0x6010c4, - 0x24d184, - 0x280e84, - 0x21b846, - 0x29e4ce, - 0x364807, - 0x283645, - 0x3a6e0c, - 0x2f8f47, - 0x26d647, - 0x2f4449, - 0x21be49, - 0x285dc5, - 0x36fb08, - 0x231589, - 0x319605, - 0x2bf788, - 0x221fc6, - 0x266cc6, - 0x3c2f44, - 0x28b688, - 0x225503, - 0x3875c4, - 0x2cbe45, - 0x388307, - 0x228785, - 0x345b49, - 0x2ab04d, - 0x2b0486, - 0x207704, - 0x2ba8c8, - 0x27084a, - 0x228b87, - 0x31ce85, - 0x23b3c3, - 0x29d70e, - 0x2a8b0c, - 0x2ff747, - 0x29e687, - 0x217b83, - 0x2eee85, - 0x280e85, - 0x299048, - 0x2963c9, - 0x23c406, - 0x25b1c4, - 0x2e7c86, - 0x23390b, - 0x38320c, - 0x33a8c7, - 0x2cccc5, - 0x3b2348, - 0x2d8945, - 0x293247, - 0x2f1587, - 0x245945, - 0x20ae83, - 0x335ac4, - 0x22a385, - 0x3b8cc5, - 0x3b8cc6, - 0x2b5308, - 0x26d6c7, - 0x318386, - 0x205c06, - 0x23f046, - 0x27e509, - 0x21f3c7, - 0x250646, - 0x383386, - 0x275d06, - 0x2a8605, - 0x3c53c6, - 0x3746c5, - 0x2da488, - 0x28ff4b, - 0x28c586, - 0x36e244, - 0x2e0409, - 0x26d004, - 0x221f48, - 0x326107, - 0x281e44, - 0x2bb308, - 0x2c0844, - 0x2a8644, - 0x286605, - 0x2dfc46, - 0x3356c7, - 0x27f283, - 0x29d185, - 0x32ce84, - 0x24fec6, - 0x325788, - 0x2b6c05, - 0x28fc09, - 0x2303c5, - 0x21c988, - 0x2312c7, - 0x38de88, - 0x2ba487, - 0x27df09, - 0x2820c6, - 0x305706, - 0x2b3084, - 0x2d7345, - 0x300a4c, - 0x277e87, - 0x278a87, - 0x36e0c8, - 0x2b0486, - 0x271484, - 0x30a244, - 0x27fec9, - 0x2c2d86, - 0x282687, - 0x277784, - 0x24d786, - 0x317bc5, - 0x2cab47, - 0x2cc986, - 0x24d949, - 0x383047, - 0x2692c7, - 0x29ffc6, - 0x24d6c5, - 0x27d1c8, - 0x221cc8, - 0x348546, - 0x2b6c45, - 0x349e86, - 0x206543, - 0x298ec9, - 0x29c48e, - 0x2ba1c8, - 0x24c848, - 0x34834b, - 0x28fe46, - 0x211584, - 0x281044, - 0x29c58a, - 0x217207, - 0x250705, - 0x216b09, - 0x2be305, - 0x23ecc7, - 0x24e304, - 0x2a9a47, - 0x2e7f08, - 0x2e0dc6, - 0x24c589, - 0x2bc10a, - 0x217186, - 0x296e86, - 0x2ab545, - 0x3898c5, - 0x347ac7, - 0x24cf08, - 0x317b08, - 0x2f74c6, - 0x3075c5, - 0x21168e, - 0x2bc7c4, - 0x298fc5, - 0x276749, - 0x382c48, - 0x28aa46, - 0x29accc, - 0x29bd90, - 0x29e10f, - 0x29f688, - 0x2f0707, - 0x3c6485, - 0x28e705, - 0x335549, - 0x28e709, - 0x240d06, - 0x29ef87, - 0x2d7245, - 0x34d589, - 0x33f106, - 0x2eec8d, - 0x280209, - 0x240084, - 0x2b9f48, - 0x236f89, - 0x351446, - 0x2ebec5, - 0x305706, - 0x3b3109, - 0x277608, - 0x212305, - 0x28b684, - 0x29ae8b, - 0x351305, - 0x242d46, - 0x281786, - 0x285206, - 0x28f64b, - 0x28fd09, - 0x205b45, - 0x396e87, - 0x2f16c6, - 0x240206, - 0x280c08, - 0x2dfd49, - 0x344a0c, - 0x22d8c8, - 0x308ec6, - 0x32c803, - 0x32a506, - 0x27ddc5, - 0x27be48, - 0x306fc6, - 0x2cad88, - 0x2062c5, - 0x27a585, - 0x365288, - 0x31dc07, - 0x317fc7, - 0x26d887, - 0x240088, - 0x2c5508, - 0x2b1206, - 0x2b2047, - 0x267c47, - 0x28f34a, - 0x256c83, - 0x3c2806, - 0x23c645, - 0x2403c4, - 0x279409, - 0x27de84, - 0x250a44, - 0x29a104, - 0x29e68b, - 0x31bd47, - 0x2118c5, - 0x295348, - 0x276dc6, - 0x276dc8, - 0x27b206, - 0x28b5c5, - 0x28b885, - 0x28d446, - 0x28dbc8, - 0x28e288, - 0x2795c6, - 0x29518f, - 0x298990, - 0x369405, - 0x205c83, - 0x2306c5, - 0x309088, - 0x28e609, - 0x319748, - 0x24c408, - 0x238d88, - 0x31be07, - 0x276a89, - 0x2caf88, - 0x28dac4, - 0x299f88, - 0x2addc9, - 0x2b38c7, - 0x299f04, - 0x21e1c8, - 0x2a07ca, - 0x2aff86, - 0x24c486, - 0x229f09, - 0x29bfc7, - 0x2c83c8, - 0x345608, - 0x294048, - 0x25d345, - 0x38a705, - 0x258a45, - 0x280e45, - 0x37ffc7, - 0x20ae85, - 0x2bf845, - 0x206d86, - 0x319687, - 0x2e1787, - 0x2a0ac6, - 0x2ce085, - 0x242d46, - 0x24c685, - 0x2d70c8, - 0x2ff5c4, - 0x393106, - 0x334e84, - 0x31dec8, - 0x22f10a, - 0x27a34c, - 0x23fac5, - 0x283506, - 0x344bc6, - 0x341e06, - 0x308f44, - 0x317e85, - 0x27b047, - 0x29c049, - 0x2c6f47, - 0x6010c4, - 0x6010c4, - 0x31bbc5, - 0x2cb984, - 0x29a68a, - 0x276c46, - 0x251e84, - 0x204745, - 0x36c3c5, - 0x2b1804, - 0x2813c7, - 0x230347, - 0x2c6988, - 0x31fec8, - 0x212309, - 0x26eec8, - 0x29a84b, - 0x2b7fc4, - 0x37b985, - 0x27d905, - 0x26d809, - 0x2dfd49, - 0x2e0308, - 0x22d748, - 0x29f504, - 0x2846c5, - 0x200583, - 0x2c53c5, - 0x297646, - 0x29620c, - 0x21f046, - 0x2ebdc6, - 0x28acc5, - 0x2f8448, - 0x35ec46, - 0x3963c6, - 0x24c486, - 0x22c7cc, - 0x250544, - 0x23f18a, - 0x28ac08, - 0x296047, - 0x32cd86, - 0x23c4c7, - 0x2e7885, - 0x2adf86, - 0x35aa46, - 0x366207, - 0x250a84, - 0x309705, - 0x276744, - 0x2c9c87, - 0x276988, - 0x277a8a, - 0x27ebc7, - 0x2a8207, - 0x2f0687, - 0x2d8a89, - 0x29620a, - 0x22e943, - 0x250985, - 0x218203, - 0x385c49, - 0x336dc8, - 0x351687, - 0x319849, - 0x221dc6, - 0x3b6808, - 0x33c445, - 0x3bf38a, - 0x200c89, - 0x3299c9, - 0x35eb07, - 0x26dac9, - 0x2180c8, - 0x3663c6, - 0x2836c8, - 0x203a47, - 0x22ea87, - 0x245d87, - 0x2e5a48, - 0x2f89c6, - 0x2a0585, - 0x27b047, - 0x296988, - 0x334e04, - 0x2c8004, - 0x28ed07, - 0x2acc47, - 0x23140a, - 0x366346, - 0x364c4a, - 0x2be747, - 0x2bc587, - 0x3097c4, - 0x2aa444, - 0x2caa46, - 0x23a444, - 0x23a44c, - 0x39ee05, - 0x218a09, - 0x337284, - 0x2b18c5, - 0x2707c8, - 0x239dc5, - 0x37f306, - 0x2311c4, - 0x2d02ca, - 0x2cb2c6, - 0x29180a, - 0x2162c7, - 0x224505, - 0x21c645, - 0x24204a, - 0x28dd45, - 0x29ed06, - 0x209784, - 0x2afac6, - 0x347b85, - 0x307086, - 0x2e44cc, - 0x218d4a, - 0x252b04, - 0x2060c6, - 0x29bfc7, - 0x2cc904, - 0x26d1c8, - 0x2f3106, - 0x211509, - 0x2db609, - 0x394949, - 0x2c7246, - 0x203b46, - 0x283807, - 0x38dac8, - 0x203949, - 0x31bd47, - 0x2954c6, - 0x2faac7, - 0x2a6b05, - 0x2bc7c4, - 0x2833c7, - 0x267e05, - 0x286545, - 0x31f747, - 0x245808, - 0x3b22c6, - 0x2977cd, - 0x29924f, - 0x29d54d, - 0x214084, - 0x238986, - 0x2d0688, - 0x3ab905, - 0x28f508, - 0x256f8a, - 0x240084, - 0x233b46, - 0x2cd107, - 0x2ca387, - 0x2cfe09, - 0x283685, - 0x2b1804, - 0x2b49ca, - 0x2bbbc9, - 0x26dbc7, - 0x297a86, - 0x351446, - 0x284606, - 0x371d46, - 0x2cf6cf, - 0x2d0549, - 0x24e386, - 0x354846, - 0x31ac09, - 0x2b2147, - 0x20be43, - 0x22c946, - 0x20de03, - 0x2de988, - 0x2fa907, - 0x29f889, - 0x2b1b88, - 0x318108, - 0x328006, - 0x21ef89, - 0x399b05, - 0x2b2204, - 0x2e8187, - 0x372b85, - 0x214084, - 0x211988, - 0x2174c4, - 0x2b1e87, - 0x30d686, - 0x395ac5, - 0x2ae608, - 0x35130b, - 0x2b4107, - 0x2422c6, - 0x2c2b84, - 0x2b6286, - 0x26a985, - 0x267e05, - 0x27cf49, - 0x280fc9, - 0x22eac4, - 0x22eb05, - 0x206105, - 0x3bf206, - 0x36fc08, - 0x2bdc86, - 0x3aa9cb, - 0x37570a, - 0x2ba585, - 0x28b906, - 0x39df85, - 0x345405, - 0x29b847, - 0x3c2a88, - 0x24ed84, - 0x39ce86, - 0x28e306, - 0x218287, - 0x30eb04, - 0x27bcc6, - 0x35c745, - 0x35c749, - 0x203d44, - 0x2b3f49, - 0x2795c6, - 0x2c0048, - 0x206105, - 0x36e305, - 0x307086, - 0x344909, - 0x21be49, - 0x2ebe46, - 0x382d48, - 0x2ab188, - 0x39df44, - 0x2b5504, - 0x2b5508, - 0x239d08, - 0x24ee89, - 0x2975c6, - 0x24c486, - 0x32be4d, - 0x27c986, - 0x2ce449, - 0x39e845, - 0x20de46, - 0x2941c8, - 0x3277c5, - 0x267c84, - 0x26a985, - 0x280908, - 0x29a449, - 0x276804, - 0x31e346, - 0x251f0a, - 0x2ff648, - 0x231589, - 0x25890a, - 0x3197c6, - 0x299408, - 0x293005, - 0x28ae88, - 0x2e7905, - 0x221c89, - 0x376189, - 0x23c442, - 0x26f3c5, - 0x2ebf86, - 0x279507, - 0x38c8c5, - 0x30d2c6, - 0x304c08, - 0x2b0486, - 0x2d9309, - 0x278b86, - 0x280a88, - 0x2a9605, - 0x382106, - 0x336c48, - 0x27fd48, - 0x396008, - 0x2f2388, - 0x3c53c4, - 0x21e583, - 0x2d9544, - 0x27edc6, - 0x2a6b44, - 0x24c787, - 0x3962c9, - 0x3c0485, - 0x345606, - 0x22c946, - 0x2b514b, - 0x2b0d86, - 0x293b06, - 0x393208, - 0x310806, - 0x224303, - 0x3c4e83, - 0x2bc7c4, - 0x233e85, - 0x2d45c7, - 0x276988, - 0x27698f, - 0x27af4b, - 0x36fa08, - 0x31e3c6, - 0x36fd0e, - 0x242483, - 0x2d4544, - 0x2b0d05, - 0x2b1686, - 0x28c94b, - 0x290706, - 0x227109, - 0x395ac5, - 0x2eccc8, - 0x20e688, - 0x21bd0c, - 0x29e6c6, - 0x2c5406, - 0x2e9f85, - 0x287888, - 0x27a345, - 0x350288, - 0x29b04a, - 0x29d989, - 0x6010c4, - 0x200742, - 0x3c202c42, - 0x202542, - 0x26ff84, - 0x201e42, - 0x21a484, - 0x2032c2, - 0x200342, - 0x207c02, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x2b6c03, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x20ec83, - 0x241d03, - 0x210143, - 0x28b304, - 0x20be03, - 0x23d744, - 0x237583, - 0x2d2484, - 0x30e843, - 0x38cb87, - 0x21f743, - 0x20ae43, - 0x310f08, - 0x241d03, - 0x2a47cb, - 0x2e8943, - 0x3a25c6, - 0x20e982, - 0x3987cb, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x241d03, - 0x2098c3, - 0x204543, - 0x200742, - 0xcd588, - 0x3574c5, - 0x267e88, - 0x2e2e08, - 0x202c42, - 0x3325c5, - 0x331707, - 0x201b02, - 0x248887, - 0x202542, - 0x256807, - 0x3c0b89, - 0x292bc8, - 0x293ec9, - 0x247342, - 0x269ec7, - 0x329844, - 0x3317c7, - 0x375607, - 0x25fe82, - 0x21f743, - 0x20d682, - 0x2032c2, - 0x200342, - 0x20b182, - 0x200382, - 0x207c02, - 0x2a9105, - 0x24d0c5, - 0x2c42, - 0x37583, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0xaff03, - 0x241d03, - 0x10c43, - 0x781, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x214bc3, - 0x20ec83, - 0xaff03, - 0x241d03, - 0x21a003, - 0x3f0eca46, - 0x6f803, - 0x7f685, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x9482, - 0xcd588, - 0xae43, - 0xaff03, - 0x4cec4, - 0xd8d45, - 0x200742, - 0x3a4c04, - 0x20be03, - 0x237583, - 0x30e843, - 0x3a2f03, - 0x232585, - 0x214bc3, - 0x20f003, - 0x20ec83, - 0x228803, - 0x241d03, - 0x207c03, - 0x2605c3, - 0x203f83, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x202c42, - 0x241d03, - 0xcd588, - 0x30e843, - 0xaff03, - 0xcd588, - 0xaff03, - 0x2b8283, - 0x20be03, - 0x234784, - 0x237583, - 0x30e843, - 0x207d02, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x207d02, - 0x20be83, - 0x20ec83, - 0x241d03, - 0x2e2d83, - 0x207c03, - 0x200742, - 0x202c42, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x3a25c5, - 0x9a2c6, - 0x28b304, - 0x20e982, - 0xcd588, - 0x200742, - 0x20288, - 0x132983, - 0x202c42, - 0x43490186, - 0x12b44, - 0x10bfcb, - 0x41546, - 0x1f847, - 0x237583, - 0x52748, - 0x30e843, - 0xef4c5, - 0xe84, - 0x222003, - 0x56c47, - 0xd4344, - 0x20ec83, - 0xafd44, - 0xaff03, - 0x241d03, - 0x2e9484, - 0xfdb48, - 0x157206, - 0x10d08, - 0x135fc5, - 0x126749, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ae43, - 0x241d03, - 0x2e8943, - 0x20e982, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff83, - 0x226444, - 0x20ec83, - 0xae43, - 0x241d03, - 0x20be03, - 0x237583, - 0x2d2484, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x3a25c6, - 0x237583, - 0x30e843, - 0x181c43, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x1f847, - 0xcd588, - 0x30e843, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x45e0be03, - 0x237583, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x200742, - 0x202c42, - 0x20be03, - 0x30e843, - 0x20ec83, - 0x200342, - 0x241d03, - 0x32e147, - 0x23b00b, - 0x205d03, - 0x2409c8, - 0x38d847, - 0x224906, - 0x2c1605, - 0x38e5c9, - 0x21f4c8, - 0x36b4c9, - 0x39b210, - 0x36b4cb, - 0x2f6809, - 0x207503, - 0x22bcc9, - 0x235b46, - 0x235b4c, - 0x357588, - 0x3c02c8, - 0x200289, - 0x2e244e, - 0x3c094b, - 0x34754c, - 0x205583, - 0x27b80c, - 0x205589, - 0x2fbf47, - 0x2374cc, - 0x2ad14a, - 0x253404, - 0x32a08d, - 0x27b6c8, - 0x21014d, - 0x28a7c6, - 0x28b30b, - 0x31d689, - 0x27a747, - 0x3c2c46, - 0x341409, - 0x32538a, - 0x313048, - 0x2e8544, - 0x332e87, - 0x249d47, - 0x277984, - 0x2d9a44, - 0x386dc9, - 0x364209, - 0x215cc8, - 0x2108c5, - 0x2ea8c5, - 0x20d106, - 0x329f49, - 0x25720d, - 0x2f3308, - 0x20d007, - 0x2c1688, - 0x23ce86, - 0x37ce44, - 0x286c45, - 0x203846, - 0x2043c4, - 0x205487, - 0x2081ca, - 0x213904, - 0x2170c6, - 0x217d49, - 0x217d4f, - 0x21870d, - 0x218fc6, - 0x21fe90, - 0x220286, - 0x220807, - 0x221447, - 0x22144f, - 0x222149, - 0x228d46, - 0x22a1c7, - 0x22a1c8, - 0x22b089, - 0x395b88, - 0x2de4c7, - 0x22cf43, - 0x3bcfc6, - 0x3bbb08, - 0x2e270a, - 0x387809, - 0x212743, - 0x331606, - 0x39ccca, - 0x2e4b47, - 0x2fbd8a, - 0x209a8e, - 0x222286, - 0x26f5c7, - 0x21bac6, - 0x205646, - 0x38a50b, - 0x34edca, - 0x309b0d, - 0x203c07, - 0x260648, - 0x260649, - 0x26064f, - 0x34eb0c, - 0x27c0c9, - 0x2d778e, - 0x38cc8a, - 0x293886, - 0x2f9f06, - 0x313ccc, - 0x315a8c, - 0x328308, - 0x353d07, - 0x26d545, - 0x290504, - 0x202c4e, - 0x266f84, - 0x3188c7, - 0x39270a, - 0x3a2a54, - 0x3b92cf, - 0x221608, - 0x3bce88, - 0x33984d, - 0x33984e, - 0x231a09, - 0x232b08, - 0x232b0f, - 0x2371cc, - 0x2371cf, - 0x2386c7, - 0x23dfca, - 0x22c54b, - 0x23f5c8, - 0x242707, - 0x2616cd, - 0x20a286, - 0x32a246, - 0x2443c9, - 0x2a6f88, - 0x249208, - 0x24920e, - 0x23b107, - 0x24b505, - 0x24cc85, - 0x204c44, - 0x224bc6, - 0x215bc8, - 0x322ec3, - 0x397e4e, - 0x261a88, - 0x2ae14b, - 0x301c07, - 0x2f7305, - 0x27b986, - 0x2aab07, - 0x2f4848, - 0x317909, - 0x20a505, - 0x284bc8, - 0x226e06, - 0x39caca, - 0x202b49, - 0x237589, - 0x23758b, - 0x3211c8, - 0x277849, - 0x210986, - 0x36db0a, - 0x2bf50a, - 0x23e1cc, - 0x21e907, - 0x2929ca, - 0x211d0b, - 0x211d19, - 0x30a988, - 0x3a2645, - 0x261886, - 0x26ba89, - 0x358706, - 0x2d340a, - 0x21f6c6, - 0x225784, - 0x2c380d, - 0x323307, - 0x225789, - 0x24e685, - 0x251548, - 0x252509, - 0x252944, - 0x253307, - 0x253308, - 0x253907, - 0x268688, - 0x257887, - 0x205dc5, - 0x25d60c, - 0x25de49, - 0x30590a, - 0x3a0e09, - 0x22bdc9, - 0x37924c, - 0x26004b, - 0x260dc8, - 0x261e88, - 0x265244, - 0x281b08, - 0x283c89, - 0x2ad207, - 0x217f86, - 0x31d907, - 0x3843c9, - 0x335c0b, - 0x2b6107, - 0x3c6847, - 0x216407, - 0x2100c4, - 0x2100c5, - 0x278845, - 0x34c04b, - 0x3ad084, - 0x320d08, - 0x28550a, - 0x226ec7, - 0x35b207, - 0x28c112, - 0x2a10c6, - 0x234106, - 0x2b658e, - 0x2a4a86, - 0x291508, - 0x291a8f, - 0x210508, - 0x38b748, - 0x2bb78a, - 0x2bb791, - 0x2a0d4e, - 0x242a0a, - 0x242a0c, - 0x232d07, - 0x232d10, - 0x3c4cc8, - 0x2a0f45, - 0x2aae0a, - 0x20440c, - 0x29580d, - 0x2fc446, - 0x2fc447, - 0x2fc44c, - 0x30460c, - 0x214bcc, - 0x2ab88b, - 0x3706c4, - 0x211dc4, - 0x388489, - 0x30a2c7, - 0x23dd89, - 0x2bf349, - 0x2ace07, - 0x2acfc6, - 0x2acfc9, - 0x2ad3c3, - 0x2b058a, - 0x31b487, - 0x3491cb, - 0x30998a, - 0x3298c4, - 0x316946, - 0x27ee49, - 0x23a2c4, - 0x39eeca, - 0x2425c5, - 0x2bcbc5, - 0x2bcbcd, - 0x2bcf0e, - 0x2d9685, - 0x32d506, - 0x3a21c7, - 0x25d88a, - 0x37a506, - 0x2e1304, - 0x303707, - 0x246a4b, - 0x23cf47, - 0x3c1a04, - 0x390706, - 0x39070d, - 0x38408c, - 0x20eb46, - 0x2f350a, - 0x27c686, - 0x285788, - 0x354b47, - 0x23618a, - 0x24b386, - 0x203b03, - 0x294346, - 0x3bb988, - 0x38860a, - 0x2cb107, - 0x2cb108, - 0x30df84, - 0x28c687, - 0x201cc8, - 0x2a12c8, - 0x2827c8, - 0x2b130a, - 0x2d8345, - 0x20be87, - 0x242853, - 0x25a706, - 0x21b388, - 0x227609, - 0x248748, - 0x32808b, - 0x318488, - 0x246b84, - 0x365386, - 0x311e06, - 0x2dfa89, - 0x382807, - 0x25d708, - 0x29c806, - 0x31f644, - 0x341d05, - 0x2c7e48, - 0x34398a, - 0x2c3488, - 0x2c8906, - 0x29960a, - 0x3b8e48, - 0x2cc708, - 0x2cd2c8, - 0x2cdd46, - 0x2d0886, - 0x3a08cc, - 0x2d0e10, - 0x29ea45, - 0x210308, - 0x394490, - 0x210310, - 0x39b08e, - 0x3a054e, - 0x3a0554, - 0x3a624f, - 0x3a6606, - 0x321391, - 0x31eb13, - 0x31ef88, - 0x3ab285, - 0x240f08, - 0x387b05, - 0x2da18c, - 0x22cd09, - 0x290349, - 0x22d187, - 0x251309, - 0x24f147, - 0x2f8cc6, - 0x286a47, - 0x2034c5, - 0x210c83, - 0x323089, - 0x2278c9, - 0x381c43, - 0x38c7c4, - 0x326f0d, - 0x347c8f, - 0x31f685, - 0x3175c6, - 0x225a47, - 0x357307, - 0x2f5886, - 0x2f588b, - 0x2a2bc5, - 0x25f106, - 0x2f9d87, - 0x258249, - 0x375d06, - 0x322285, - 0x374e4b, - 0x3be7c6, - 0x229a85, - 0x27dc08, - 0x2b2bc8, - 0x2aec8c, - 0x2aec90, - 0x2ae949, - 0x2bd487, - 0x2d8e4b, - 0x306746, - 0x2de38a, - 0x2df80b, - 0x2e094a, - 0x2e0bc6, - 0x2e2c45, - 0x31b1c6, - 0x2b7048, - 0x22d24a, - 0x3394dc, - 0x2e8a0c, - 0x2e8d08, - 0x3a25c5, - 0x361f87, - 0x209946, - 0x270bc5, - 0x21a846, - 0x2f5a48, - 0x2bbe47, - 0x2e2348, - 0x25a7ca, - 0x225b4c, - 0x20d309, - 0x345787, - 0x222a04, - 0x24cd46, - 0x38b2ca, - 0x2bf445, - 0x2110cc, - 0x213588, - 0x373608, - 0x2238cc, - 0x2dd30c, - 0x329409, - 0x329647, - 0x24398c, - 0x22c044, - 0x2482ca, - 0x3033cc, - 0x27280b, - 0x372f0b, - 0x253786, - 0x258d87, - 0x232f47, - 0x232f4f, - 0x2fce11, - 0x2d5b52, - 0x2590cd, - 0x2590ce, - 0x25940e, - 0x3a6408, - 0x3a6412, - 0x25eb08, - 0x38d487, - 0x2556ca, - 0x355cc8, - 0x2a4a45, - 0x37fe0a, - 0x220607, - 0x316244, - 0x221b83, - 0x378205, - 0x2bba07, - 0x307c47, - 0x295a0e, - 0x399e8d, - 0x39b5c9, - 0x20fd85, - 0x3b00c3, - 0x20ab06, - 0x25f705, - 0x2ae388, - 0x2b9609, - 0x2618c5, - 0x2618cf, - 0x2e2a87, - 0x38e505, - 0x3025ca, - 0x2b1506, - 0x25b949, - 0x2f640c, - 0x2f80c9, - 0x3be506, - 0x28530c, - 0x32c906, - 0x2fb508, - 0x2fba86, - 0x30ab06, - 0x2b0f04, - 0x30d603, - 0x38668a, - 0x216711, - 0x27c28a, - 0x272085, - 0x282347, - 0x25ab47, - 0x201dc4, - 0x201dcb, - 0x293d48, - 0x2ba046, - 0x36e145, - 0x33d8c4, - 0x362109, - 0x202a84, - 0x249047, - 0x300585, - 0x300587, - 0x2b67c5, - 0x3482c3, - 0x38d348, - 0x317c4a, - 0x27f283, - 0x35750a, - 0x3b3486, - 0x26164f, - 0x3b8349, - 0x397dd0, - 0x2effc8, - 0x2c8d49, - 0x296707, - 0x39068f, - 0x319c04, - 0x2d2504, - 0x220106, - 0x34f846, - 0x2d6e8a, - 0x253c06, - 0x352987, - 0x303f48, - 0x304147, - 0x3049c7, - 0x305b8a, - 0x3083cb, - 0x341ac5, - 0x2d5788, - 0x256703, - 0x3b6b0c, - 0x35d60f, - 0x26d34d, - 0x25e287, - 0x39b709, - 0x36de47, - 0x282c48, - 0x3a2c4c, - 0x2c8a48, - 0x2701c8, - 0x31c38e, - 0x333d94, - 0x3342a4, - 0x35308a, - 0x36becb, - 0x24f204, - 0x24f209, - 0x233bc8, - 0x24d305, - 0x3229ca, - 0x261cc7, - 0x31b0c4, - 0x2b6c03, - 0x20be03, - 0x23d744, - 0x237583, - 0x30e843, - 0x26ff84, - 0x214bc3, - 0x21f743, - 0x2d0e06, - 0x226444, - 0x20ec83, - 0x241d03, - 0x219543, - 0x200742, - 0x2b6c03, - 0x202c42, - 0x20be03, - 0x23d744, - 0x237583, - 0x30e843, - 0x214bc3, - 0x2d0e06, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x203d43, - 0x20ec83, - 0xaff03, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x226444, - 0x20ec83, - 0x241d03, - 0x200742, - 0x24be43, - 0x202c42, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x203cc2, - 0x24d5c2, - 0x202c42, - 0x20be03, - 0x20b2c2, - 0x201342, - 0x26ff84, - 0x21a484, - 0x227d42, - 0x226444, - 0x200342, - 0x241d03, - 0x219543, - 0x253786, - 0x230882, - 0x204182, - 0x228382, - 0x48610503, - 0x48a32d03, - 0x5b586, - 0x5b586, - 0x28b304, - 0x20ae43, - 0x15a4a, - 0x3ba4c, - 0x121ecc, - 0x7f48d, - 0x117485, - 0x2aa47, - 0x14ec6, - 0x19148, - 0x1c787, - 0x23288, - 0x1807ca, - 0x102c07, - 0x496d2fc5, - 0x133789, - 0x3c00b, - 0x18754b, - 0x1bdd08, - 0xf608a, - 0x8a34e, - 0x144854b, - 0x12b44, - 0x5f246, - 0x8808, - 0x7e948, - 0x3c2c7, - 0x910c7, - 0x78449, - 0x3a347, - 0x67008, - 0x100249, - 0x170bc4, - 0x191e05, - 0x12f34e, - 0xa964d, - 0x1f6c8, - 0x49b64046, - 0x4a564048, - 0x739c8, - 0x12e350, - 0x5978c, - 0x666c7, - 0x66e47, - 0x6abc7, - 0x704c7, - 0xd0c2, - 0x1807, - 0x14ef8c, - 0x11d107, - 0xa4686, - 0xa5a89, - 0xa7708, - 0x552c2, - 0x1342, - 0x3900b, - 0xafdc7, - 0x25589, - 0x52c49, - 0x142188, - 0xb0102, - 0x1970c9, - 0xc9f8a, - 0xc6209, - 0xd3909, - 0xd50c8, - 0xd6007, - 0xd82c9, - 0xda885, - 0xdae90, - 0x138ac6, - 0x14a345, - 0xeb78d, - 0x2bac6, - 0xe3d47, - 0xe9498, - 0x3a6c8, - 0x14640a, - 0x16f02, - 0x5f88d, - 0x1282, - 0x7dac6, - 0x8cc08, - 0x6e308, - 0xcd449, - 0x1be608, - 0x6f98e, - 0xab07, - 0xfe68d, - 0xf26c5, - 0x1588, - 0x1a1e08, - 0xfeec6, - 0xc842, - 0x157206, - 0xdc2, - 0x2c1, - 0x60a07, - 0x8b003, - 0x49ee9d04, - 0x4a294a43, - 0x101, - 0x13d06, - 0x101, - 0x301, - 0x13d06, - 0x8b003, - 0x140e3c5, - 0x253404, - 0x20be03, - 0x254a04, - 0x26ff84, - 0x20ec83, - 0x2274c5, - 0x21a003, - 0x24f3c3, - 0x2f5805, - 0x203f83, - 0x4b60be03, - 0x237583, - 0x30e843, - 0x200541, - 0x21f743, - 0x21a484, - 0x226444, - 0x20ec83, - 0x241d03, - 0x207c03, - 0xcd588, - 0x200742, - 0x2b6c03, - 0x202c42, - 0x20be03, - 0x237583, - 0x203d43, - 0x201342, - 0x26ff84, - 0x214bc3, - 0x21f743, - 0x20ec83, - 0x20ae43, - 0x241d03, - 0x203f83, - 0xcd588, - 0x35d382, - 0x1851c7, - 0x2c42, - 0x141c85, - 0x598cf, - 0x1455908, - 0x10430e, - 0x4c607402, - 0x31a848, - 0x307206, - 0x2c1146, - 0x306b87, - 0x4ca11a42, - 0x4cfb81c8, - 0x21ed8a, - 0x266148, - 0x200602, - 0x31b2c9, - 0x341b07, - 0x217f06, - 0x38d089, - 0x20bfc4, - 0x20e2c6, - 0x2f2904, - 0x270984, - 0x25cf89, - 0x3030c6, - 0x24d185, - 0x2faf45, - 0x2322c7, - 0x2be9c7, - 0x354984, - 0x306dc6, - 0x2ff205, - 0x309485, - 0x39dec5, - 0x2ea687, - 0x301a45, - 0x319149, - 0x336f85, - 0x2f4984, - 0x37a447, - 0x348a0e, - 0x3aaec9, - 0x2b6449, - 0x335246, - 0x2454c8, - 0x2ee34b, - 0x35bd8c, - 0x33b746, - 0x347407, - 0x2afbc5, - 0x2d9a4a, - 0x215dc9, - 0x366f09, - 0x332706, - 0x2f9b45, - 0x383105, - 0x338689, - 0x39e04b, - 0x275e86, - 0x343ec6, - 0x203104, - 0x28bdc6, - 0x24b588, - 0x3bb806, - 0x21d186, - 0x206888, - 0x209347, - 0x209509, - 0x20acc5, - 0xcd588, - 0x291044, - 0x304f44, - 0x210f05, - 0x3a8c49, - 0x226087, - 0x22608b, - 0x2288ca, - 0x22cc45, - 0x4d207a42, - 0x309847, - 0x4d62cf48, - 0x370a47, - 0x383e85, - 0x32654a, - 0x2c42, - 0x2fac0b, - 0x2579ca, - 0x2277c6, - 0x210d83, - 0x2a4f8d, - 0x3aa74c, - 0x3bedcd, - 0x24e2c5, - 0x37dd45, - 0x322f07, - 0x20b2c9, - 0x21ec86, - 0x253a85, - 0x2ced88, - 0x28bcc3, - 0x2e3108, - 0x28bcc8, - 0x2c2107, - 0x30f108, - 0x3aa549, - 0x286d47, - 0x23ab87, - 0x224788, - 0x38ae04, - 0x38ae07, - 0x28a6c8, - 0x353706, - 0x3b544f, - 0x2293c7, - 0x2de646, - 0x329785, - 0x228503, - 0x391c87, - 0x378183, - 0x253e86, - 0x2553c6, - 0x255b06, - 0x28fa05, - 0x268683, - 0x396d48, - 0x379c89, - 0x38eb4b, - 0x255c88, - 0x257545, - 0x258b85, - 0x4db29982, - 0x286b09, - 0x38d707, - 0x25f185, - 0x25ce87, - 0x25e9c6, - 0x371c05, - 0x25f54b, - 0x260dc4, - 0x265d05, - 0x265e47, - 0x275806, - 0x275c45, - 0x281d07, - 0x2829c7, - 0x2e1704, - 0x28990a, - 0x289dc8, - 0x293089, - 0x241245, - 0x364946, - 0x24b74a, - 0x2fae46, - 0x268fc7, - 0x292d4d, - 0x2a2709, - 0x3954c5, - 0x2031c7, - 0x31f148, - 0x336a08, - 0x209ec7, - 0x3be1c6, - 0x21d4c7, - 0x255083, - 0x303044, - 0x36e785, - 0x39fc47, - 0x3a4609, - 0x230cc8, - 0x33dc85, - 0x2363c4, - 0x253d45, - 0x39038d, - 0x211742, - 0x2bd986, - 0x27da06, - 0x2dac0a, - 0x381346, - 0x38b205, - 0x31ffc5, - 0x31ffc7, - 0x39c90c, - 0x27384a, - 0x28ba86, - 0x2c4d85, - 0x28bc06, - 0x28bf47, - 0x28d986, - 0x28f90c, - 0x38d1c9, - 0x4de14187, - 0x291e45, - 0x291e46, - 0x2944c8, - 0x247e85, - 0x2a3505, - 0x2a3c88, - 0x2a3e8a, - 0x4e278142, - 0x4e607c82, - 0x2d7485, - 0x2a6b43, - 0x2461c8, - 0x211b83, - 0x2a4104, - 0x25ba8b, - 0x211b88, - 0x2ce288, - 0x4eb1cb49, - 0x2a8e09, - 0x2a9546, - 0x2aa788, - 0x2aa989, - 0x2ab386, - 0x2ab505, - 0x24e0c6, - 0x2abfc9, - 0x3ac147, - 0x381fc6, - 0x2d8787, - 0x21eb07, - 0x34e204, - 0x4ee3a9c9, - 0x270e08, - 0x3b80c8, - 0x31f887, - 0x2c2f46, - 0x20b0c9, - 0x2f2bc7, - 0x33194a, - 0x364a88, - 0x3be307, - 0x20ed86, - 0x39d28a, - 0x372cc8, - 0x382ac5, - 0x22b9c5, - 0x30b047, - 0x36ac49, - 0x30280b, - 0x314248, - 0x337009, - 0x255f87, - 0x2b868c, - 0x2b8c8c, - 0x2b8f8a, - 0x2b920c, - 0x2c10c8, - 0x2c12c8, - 0x2c14c4, - 0x2c1889, - 0x2c1ac9, - 0x2c1d0a, - 0x2c1f89, - 0x2c22c7, - 0x3b4c4c, - 0x23d386, - 0x3c0708, - 0x2faf06, - 0x37fc46, - 0x3953c7, - 0x3ab0c8, - 0x349c4b, - 0x370907, - 0x35b849, - 0x3782c9, - 0x254b87, - 0x2f2b44, - 0x282487, - 0x2eaf46, - 0x215946, - 0x2f36c5, - 0x3720c8, - 0x290244, - 0x290246, - 0x27370b, - 0x2b0889, - 0x39ddc6, - 0x204cc9, - 0x2ea806, - 0x22e688, - 0x20b4c3, - 0x2f9cc5, - 0x21d2c9, - 0x228b05, - 0x30ae84, - 0x274d06, - 0x3993c5, - 0x259b06, - 0x308747, - 0x331086, - 0x23078b, - 0x36da07, - 0x256e46, - 0x348606, - 0x232386, - 0x354949, - 0x2e474a, - 0x2ba345, - 0x3be8cd, - 0x2a3f86, - 0x2e9106, - 0x397cc6, - 0x285705, - 0x2db187, - 0x2f75c7, - 0x207cce, - 0x21f743, - 0x2c2f09, - 0x358489, - 0x2d9e47, - 0x26c287, - 0x2a1445, - 0x2ae085, - 0x4f386f0f, - 0x2c8f87, - 0x2c9148, - 0x2c9884, - 0x2c9e46, - 0x4f64cd02, - 0x2cdfc6, - 0x2d0e06, - 0x349f8e, - 0x2e2f4a, - 0x3b8946, - 0x2ca24a, - 0x2065c9, - 0x231e85, - 0x344788, - 0x39a146, - 0x29aac8, - 0x3c2dc8, - 0x2a57cb, - 0x306c85, - 0x301ac8, - 0x2069cc, - 0x383d47, - 0x255606, - 0x27c4c8, - 0x224a88, - 0x4fa53982, - 0x20e08b, - 0x3361c9, - 0x21cb09, - 0x39dc47, - 0x38a7c8, - 0x4fe3ca88, - 0x21318b, - 0x342009, - 0x28394d, - 0x24e488, - 0x3518c8, - 0x502056c2, - 0x331404, - 0x50623b82, - 0x2f7e06, - 0x50a0a542, - 0x24fc8a, - 0x204b86, - 0x22e0c8, - 0x2be048, - 0x326cc6, - 0x398b46, - 0x2efd46, - 0x2ae305, - 0x240684, - 0x50e2e604, - 0x34c986, - 0x2a2247, - 0x5121c1c7, - 0x2e1bcb, - 0x348dc9, - 0x37dd8a, - 0x357ec4, - 0x320108, - 0x381d8d, - 0x2e6e49, - 0x2e7088, - 0x2e7709, - 0x2e9484, - 0x22c404, - 0x27d505, - 0x2ee68b, - 0x211b06, - 0x34c7c5, - 0x222449, - 0x306e88, - 0x29fb44, - 0x2d9bc9, - 0x326b05, - 0x2bea08, - 0x23b247, - 0x2b6848, - 0x27f046, - 0x207907, - 0x2d4109, - 0x374fc9, - 0x229b05, - 0x240305, - 0x51607482, - 0x2f4744, - 0x225dc5, - 0x292786, - 0x2f8305, - 0x297b87, - 0x34ca85, - 0x275844, - 0x335306, - 0x253b07, - 0x234fc6, - 0x384305, - 0x20e4c8, - 0x307405, - 0x20ef87, - 0x2154c9, - 0x2b09ca, - 0x34e587, - 0x34e58c, - 0x24d146, - 0x241b89, - 0x244885, - 0x247dc8, - 0x201283, - 0x210945, - 0x2eac05, - 0x257f47, - 0x51a12c02, - 0x398347, - 0x2f3c06, - 0x32fdc6, - 0x2f7f46, - 0x2249c6, - 0x2eb408, - 0x241045, - 0x2de707, - 0x2de70d, - 0x221b83, - 0x221b85, - 0x302387, - 0x398688, - 0x301f45, - 0x219788, - 0x23dc86, - 0x333947, - 0x3c0645, - 0x306d06, - 0x3a4c85, - 0x226bca, - 0x2fe986, - 0x22eec7, - 0x2f04c5, - 0x2ffc87, - 0x303684, - 0x30ae06, - 0x3446c5, - 0x357a0b, - 0x2eadc9, - 0x24bf4a, - 0x229b88, - 0x34c2c8, - 0x34cb8c, - 0x353847, - 0x36f808, - 0x387c08, - 0x394205, - 0x3a684a, - 0x3b00c9, - 0x51e01b42, - 0x3c6646, - 0x222e44, - 0x222e49, - 0x294f49, - 0x276587, - 0x2f9887, - 0x2bf1c9, - 0x285908, - 0x28590f, - 0x21dec6, - 0x2d2c8b, - 0x2f5645, - 0x2f5647, - 0x2f5c49, - 0x25bbc6, - 0x2d9b47, - 0x2d5ec5, - 0x234dc4, - 0x341006, - 0x226244, - 0x2df647, - 0x2ce808, - 0x522f9a48, - 0x2fa1c5, - 0x2fa307, - 0x24eb09, - 0x20de44, - 0x2473c8, - 0x52716788, - 0x201dc4, - 0x235fc8, - 0x3c2d04, - 0x3bebc9, - 0x21b2c5, - 0x52a0e982, - 0x21df05, - 0x2cb8c5, - 0x203008, - 0x238507, - 0x52e02a82, - 0x339e45, - 0x2cc586, - 0x249746, - 0x2f4708, - 0x2f4c88, - 0x2f82c6, - 0x30a146, - 0x385289, - 0x32fd06, - 0x29124b, - 0x3478c5, - 0x355c06, - 0x28e088, - 0x231bc6, - 0x20a386, - 0x219c4a, - 0x2a91ca, - 0x370c85, - 0x241107, - 0x30d0c6, - 0x53206d02, - 0x3024c7, - 0x260505, - 0x24b6c4, - 0x24b6c5, - 0x357dc6, - 0x271a47, - 0x220105, - 0x295004, - 0x2ad5c8, - 0x20a445, - 0x309fc7, - 0x3b1c45, - 0x226b05, - 0x268904, - 0x2abac9, - 0x2ff048, - 0x399286, - 0x2adc46, - 0x201ac6, - 0x536ff908, - 0x2ffb07, - 0x2ffe4d, - 0x30074c, - 0x300d49, - 0x300f89, - 0x53b65c82, - 0x3b7e83, - 0x2228c3, - 0x2eb005, - 0x39fd4a, - 0x32fbc6, - 0x3052c5, - 0x308904, - 0x30890b, - 0x323e4c, - 0x32488c, - 0x324b95, - 0x32754d, - 0x32a68f, - 0x32aa52, - 0x32aecf, - 0x32b292, - 0x32b713, - 0x32bbcd, - 0x32c18d, - 0x32c50e, - 0x32ca8e, - 0x32d2cc, - 0x32d68c, - 0x32dacb, - 0x32de4e, - 0x32e752, - 0x32f98c, - 0x32ff50, - 0x33ba12, - 0x33c68c, - 0x33cd4d, - 0x33d08c, - 0x33f651, - 0x34404d, - 0x34ac8d, - 0x34b28a, - 0x34b50c, - 0x34be0c, - 0x34c4cc, - 0x34ce8c, - 0x350493, - 0x350b10, - 0x350f10, - 0x351acd, - 0x3520cc, - 0x352dc9, - 0x35518d, - 0x3554d3, - 0x356491, - 0x3568d3, - 0x35888f, - 0x358c4c, - 0x358f4f, - 0x35930d, - 0x35990f, - 0x359cd0, - 0x35a74e, - 0x35df0e, - 0x35e490, - 0x35f08d, - 0x35fa0e, - 0x35fd8c, - 0x360d93, - 0x3628ce, - 0x362f50, - 0x363351, - 0x36378f, - 0x363b53, - 0x36580d, - 0x365b4f, - 0x365f0e, - 0x3665d0, - 0x3669c9, - 0x367d10, - 0x36830f, - 0x36898f, - 0x368d52, - 0x369e0e, - 0x36a80d, - 0x36ae8d, - 0x36b1cd, - 0x36c84d, - 0x36cb8d, - 0x36ced0, - 0x36d2cb, - 0x36e54c, - 0x36e8cc, - 0x36eecc, - 0x36f1ce, - 0x37bbd0, - 0x37e012, - 0x37e48b, - 0x37e98e, - 0x37ed0e, - 0x37f58e, - 0x37fa0b, - 0x53f80196, - 0x38104d, - 0x3814d4, - 0x38228d, - 0x386915, - 0x388c4d, - 0x3895cf, - 0x389ccf, - 0x38ee0f, - 0x38f1ce, - 0x38f74d, - 0x391891, - 0x394b8c, - 0x394e8c, - 0x39518b, - 0x39560c, - 0x39654f, - 0x396912, - 0x39950d, - 0x39ae0c, - 0x39b94c, - 0x39bc4d, - 0x39bf8f, - 0x39c34e, - 0x39fa0c, - 0x39ffcd, - 0x3a030b, - 0x3a0bcc, - 0x3a14cd, - 0x3a180e, - 0x3a1b89, - 0x3a3553, - 0x3a3a8d, - 0x3a3dcd, - 0x3a43cc, - 0x3a484e, - 0x3a520f, - 0x3a55cc, - 0x3a58cd, - 0x3a5c0f, - 0x3a5fcc, - 0x3a778c, - 0x3a7b0c, - 0x3a7e0c, - 0x3a84cd, - 0x3a8812, - 0x3a8e8c, - 0x3a918c, - 0x3a9491, - 0x3a98cf, - 0x3a9c8f, - 0x3aa053, - 0x3ac70e, - 0x3aca8f, - 0x3ace4c, - 0x543ad18e, - 0x3ad50f, - 0x3ad8d6, - 0x3ae0d2, - 0x3af8cc, - 0x3b030f, - 0x3b098d, - 0x3b0ccf, - 0x3b108c, - 0x3b138d, - 0x3b16cd, - 0x3b2c8e, - 0x3b3bcc, - 0x3b3ecc, - 0x3b41d0, - 0x3b7211, - 0x3b764b, - 0x3b7a8c, - 0x3b7d8e, - 0x3baa91, - 0x3baece, - 0x3bb24d, - 0x3c338b, - 0x3c3c8f, - 0x3c4794, - 0x25cd82, - 0x25cd82, - 0x2032c3, - 0x25cd82, - 0x2032c3, - 0x25cd82, - 0x2009c2, - 0x24e105, - 0x3ba78c, - 0x25cd82, - 0x25cd82, - 0x2009c2, - 0x25cd82, - 0x294b45, - 0x2b09c5, - 0x25cd82, - 0x25cd82, - 0x200302, - 0x294b45, - 0x328909, - 0x35618c, - 0x25cd82, - 0x25cd82, - 0x25cd82, - 0x25cd82, - 0x24e105, - 0x25cd82, - 0x25cd82, - 0x25cd82, - 0x25cd82, - 0x200302, - 0x328909, - 0x25cd82, - 0x25cd82, - 0x25cd82, - 0x2b09c5, - 0x25cd82, - 0x2b09c5, - 0x35618c, - 0x3ba78c, - 0x2b6c03, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x20ec83, - 0x241d03, - 0x1233c8, - 0x6fac4, - 0xae43, - 0x193708, - 0x200742, - 0x55202c42, - 0x246d03, - 0x252d84, - 0x206c03, - 0x3c24c4, - 0x234106, - 0x2137c3, - 0x2f9084, - 0x2f0b05, - 0x21f743, - 0x20ec83, - 0xaff03, - 0x241d03, - 0x22d50a, - 0x253786, - 0x37f08c, - 0xcd588, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x20be83, - 0x2d0e06, - 0x20ec83, - 0x241d03, - 0x219543, - 0xa4d48, - 0x117485, - 0x187689, - 0x15842, - 0x56793ec5, - 0x2aa47, - 0xaf148, - 0xd9ce, - 0x87e92, - 0x11b5cb, - 0x102d06, - 0x56ad2fc5, - 0x56ed2fcc, - 0x25147, - 0x15da87, - 0x1208ca, - 0x42ed0, - 0x149445, - 0x10bfcb, - 0x7e948, - 0x3c2c7, - 0xf400b, - 0x78449, - 0x127b07, - 0x3a347, - 0x760c7, - 0x3c206, - 0x67008, - 0x57429546, - 0xa964d, - 0x120290, - 0x5780a6c2, - 0x1f6c8, - 0x82dd0, - 0x15b5cc, - 0x57f61e0d, - 0x5fbc8, - 0x6b8c7, - 0x164f49, - 0x5b646, - 0x946c8, - 0xebfc2, - 0x71e8a, - 0x31047, - 0x11d107, - 0xa5a89, - 0xa7708, - 0xef4c5, - 0xe85ce, - 0x1260e, - 0x1b98f, - 0x25589, - 0x52c49, - 0x7e10b, - 0x8ef4f, - 0xabccc, - 0x1125cb, - 0x105848, - 0xe1ac7, - 0xf87c8, - 0x132c8b, - 0x15274c, - 0x15af0c, - 0x1625cc, - 0x16784d, - 0x142188, - 0xfcdc2, - 0x1970c9, - 0x133acb, - 0xc3146, - 0x12e28b, - 0xd560a, - 0xd61c5, - 0xdae90, - 0xdd606, - 0x140806, - 0x14a345, - 0x18a988, - 0xe3d47, - 0xe4007, - 0x1fcc7, - 0xf7c4a, - 0xaefca, - 0x7dac6, - 0x9088d, - 0x6e308, - 0x1be608, - 0x68849, - 0xb7cc5, - 0xf778c, - 0x167a4b, - 0x16ab84, - 0xfec89, - 0xfeec6, - 0x4cb06, - 0x4182, - 0x157206, - 0x14634b, - 0x10ac87, - 0xdc2, - 0xc5105, - 0x16704, - 0x781, - 0x7bc3, - 0x572a6d06, - 0x94a43, - 0x2542, - 0x31044, - 0x602, - 0x8b304, - 0xcc2, - 0x7ec2, - 0x3102, - 0x114902, - 0x3cc2, - 0xd2fc2, - 0x3c02, - 0xefa02, - 0x3e242, - 0x4ac2, - 0x2e02, - 0x4c3c2, - 0x37583, - 0x1f02, - 0x1b02, - 0x8882, - 0x12c42, - 0x1402, - 0x35c02, - 0x552c2, - 0x6202, - 0x3882, - 0x1342, - 0x14bc3, - 0x5842, - 0x1102, - 0xb0102, - 0x9602, - 0x1542, - 0x4482, - 0xe302, - 0x6f582, - 0x1e42, - 0x17c0c2, - 0x6cd82, - 0x3a3c2, - 0xec83, - 0x2002, - 0x53982, - 0x1382, - 0x6e02, - 0x29a85, - 0x8302, - 0x48602, - 0x44303, - 0x982, - 0x16f02, - 0x1282, - 0x4042, - 0xed42, - 0x2a82, - 0xc842, - 0x4182, - 0x1f8c5, - 0x582009c2, - 0x587696c3, - 0x1fbc3, - 0x58a009c2, - 0x1fbc3, - 0x179487, - 0x215383, - 0x200742, - 0x20be03, - 0x237583, - 0x203d43, - 0x201fc3, - 0x20be83, - 0x20ec83, - 0x20ae43, - 0x241d03, - 0x294a83, - 0x10ec3, - 0xcd588, - 0x20be03, - 0x237583, - 0x203d43, - 0x21f743, - 0x20ec83, - 0x20ae43, - 0xaff03, - 0x241d03, - 0x20be03, - 0x237583, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x200541, - 0x21f743, - 0x20ec83, - 0x228803, - 0x241d03, - 0x3744, - 0x2b6c03, - 0x20be03, - 0x237583, - 0x21f6c3, - 0x203d43, - 0x257e43, - 0x26b143, - 0x2a2c83, - 0x280e83, - 0x30e843, - 0x26ff84, - 0x20ec83, - 0x241d03, - 0x203f83, - 0x31e084, - 0x250b03, - 0x5583, - 0x22d443, - 0x332388, - 0x325384, - 0x2023ca, - 0x238f06, - 0x10a904, - 0x37a147, - 0x22174a, - 0x21dd89, - 0x3ade07, - 0x3b628a, - 0x2b6c03, - 0x2d750b, - 0x293609, - 0x201bc5, - 0x34e347, - 0x2c42, - 0x20be03, - 0x214447, - 0x2fb145, - 0x2f2a09, - 0x237583, - 0x306a86, - 0x2c0c03, - 0xeae83, - 0x107f06, - 0x122746, - 0x13747, - 0x2176c6, - 0x227045, - 0x39e607, - 0x2d2107, - 0x5b30e843, - 0x33c8c7, - 0x371fc3, - 0x20fb85, - 0x26ff84, - 0x26ed48, - 0x36a50c, - 0x2ad885, - 0x2a2886, - 0x214307, - 0x345847, - 0x252e47, - 0x254d08, - 0x30600f, - 0x3371c5, - 0x246e07, - 0x37c407, - 0x2a424a, - 0x2cebc9, - 0x308045, - 0x30b7ca, - 0x136686, - 0x2c0c85, - 0x36c104, - 0x2bdf86, - 0x2f1a47, - 0x382947, - 0x348748, - 0x21b545, - 0x2fb046, - 0x21d105, - 0x36dd45, - 0x289684, - 0x326bc7, - 0x2eb24a, - 0x23fc48, - 0x366446, - 0xbe83, - 0x2d8345, - 0x318a86, - 0x3b4e86, - 0x34a246, - 0x21f743, - 0x399787, - 0x37c385, - 0x20ec83, - 0x2d58cd, - 0x20ae43, - 0x348848, - 0x38c844, - 0x275b05, - 0x2a4146, - 0x23d186, - 0x355b07, - 0x204a07, - 0x289085, - 0x241d03, - 0x3268c7, - 0x3650c9, - 0x340a49, - 0x30dc0a, - 0x249a82, - 0x20fb44, - 0x2de284, - 0x349b07, - 0x398208, - 0x2e4f89, - 0x221a49, - 0x2e5dc7, - 0x35c346, - 0xe8346, - 0x2e9484, - 0x2e9a8a, - 0x2edc08, - 0x2efc09, - 0x2de106, - 0x2b1985, - 0x23fb08, - 0x2c358a, - 0x2b5d83, - 0x31e206, - 0x2e5ec7, - 0x2311c5, - 0x38c705, - 0x3a26c3, - 0x2702c4, - 0x22b985, - 0x282ac7, - 0x2ff185, - 0x337c86, - 0x14aa05, - 0x2a3203, - 0x3b8a09, - 0x2758cc, - 0x2ca74c, - 0x2cbb08, - 0x2baec7, - 0x2fbc08, - 0x2fc24a, - 0x2fcc4b, - 0x293748, - 0x23c908, - 0x23d286, - 0x201985, - 0x320fca, - 0x369705, - 0x20e982, - 0x3c0507, - 0x261146, - 0x367405, - 0x36b749, - 0x277385, - 0x36d785, - 0x35c909, - 0x3189c6, - 0x3b6988, - 0x20fc43, - 0x217806, - 0x274c46, - 0x30a485, - 0x30a489, - 0x2e56c9, - 0x251b47, - 0x10c984, - 0x30c987, - 0x221949, - 0x23d605, - 0x40788, - 0x346245, - 0x332285, - 0x3c1309, - 0x203402, - 0x250904, - 0x203c82, - 0x205842, - 0x3c0d85, - 0x30a688, - 0x2b7c05, - 0x2c2483, - 0x2c2485, - 0x2ce1c3, - 0x20ff02, - 0x208f84, - 0x2c7d03, - 0x206f02, - 0x340484, - 0x2def43, - 0x204882, - 0x21fc43, - 0x28cb84, - 0x2f01c3, - 0x256784, - 0x200a02, - 0x219443, - 0x21d403, - 0x206a42, - 0x2f4682, - 0x2e5509, - 0x20ad42, - 0x2889c4, - 0x200442, - 0x23f984, - 0x35c304, - 0x287304, - 0x204182, - 0x23c542, - 0x3295c3, - 0x23b943, - 0x24d644, - 0x2b5c04, + 0x421cec2, + 0x227843, + 0x25d24a, + 0x460a203, + 0x2568c5, + 0x323bc2, + 0x383189, + 0x4a02842, + 0x206d84, + 0x310cc6, + 0x2c4685, + 0x366104, + 0x521d3c4, + 0x2038c3, + 0x231044, + 0x5600fc2, + 0x266944, + 0x5a88844, + 0x391d8a, + 0x5e00882, + 0x2f0947, + 0x279d08, + 0x6e07982, + 0x274487, + 0x2c2404, + 0x2c2407, + 0x3cbf45, + 0x33e047, + 0x36b686, + 0x28f304, + 0x307805, + 0x28de07, + 0x7e06542, + 0x324b83, + 0x20a742, + 0x38fdc3, + 0x820aac2, + 0x20aac5, + 0x8600202, + 0x2bd4c4, + 0x277ec5, + 0x354087, + 0x39170e, + 0x23d604, + 0x232844, + 0x207083, + 0x394d89, + 0x20708b, + 0x217c48, + 0x36b148, + 0x255488, + 0x219588, + 0x32d2ca, + 0x33df47, + 0x2ad6c6, + 0x8a49282, + 0x342303, + 0x343643, + 0x343a44, + 0x3b20c3, + 0x342343, + 0x1736382, + 0x8e00bc2, + 0x27f885, + 0x290086, + 0x27c844, + 0x35bf47, + 0x31f446, + 0x37b984, + 0x37b987, + 0x200bc3, + 0x92cb342, + 0x9720f02, + 0x9a2a8c2, + 0x22a8c6, + 0x9e00282, + 0x2a8f45, + 0x3378c3, + 0x3cc584, 0x2eddc4, - 0x30b6c4, - 0x309043, - 0x335a83, - 0x336604, - 0x30eac4, - 0x30f306, - 0x2145c2, - 0x202c42, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x200742, - 0x2b6c03, - 0x20be03, - 0x237583, - 0x201b03, - 0x30e843, - 0x26ff84, - 0x2e57c4, - 0x226444, - 0x20ec83, - 0x241d03, - 0x219543, - 0x2ea0c4, - 0x31a803, - 0x2a6503, - 0x36aac4, - 0x346046, - 0x20b583, - 0x15da87, - 0x22fac3, - 0x21e903, - 0x2b0c43, - 0x20fbc3, - 0x20be83, - 0x339d45, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, + 0x2eddc5, + 0x203283, + 0xa38d8c3, + 0xa606602, + 0x207f45, + 0x207f4b, + 0x208d06, + 0x255f4b, + 0x267c44, + 0x20adc9, + 0x20bac4, + 0xaa0bd02, + 0x20c543, + 0x20cac3, + 0x160d702, + 0x3bb283, + 0x20d70a, + 0xae0a842, + 0x3d09c5, + 0x2e074a, + 0x3778c4, + 0x20ea83, + 0x210484, + 0x210b83, + 0x210b84, + 0x210b87, + 0x211285, + 0x2122c6, + 0x2125c6, + 0x213343, + 0x217708, + 0x20a843, + 0xb2020c2, + 0x246988, + 0x3c5e8b, + 0x21e548, + 0x21ef86, + 0x220007, + 0x224cc8, + 0xc2054c2, + 0xc6c25c2, + 0x312908, + 0x303c07, + 0x280805, + 0x38f548, + 0x2dee48, + 0x37ba83, + 0x229484, + 0x343a82, + 0xca29e82, + 0xce02c82, + 0xd629fc2, + 0x229fc3, + 0xda00f82, + 0x3488c3, + 0x2d3104, + 0x208f83, + 0x324f44, + 0x39424b, + 0x231783, + 0x2e6f46, + 0x231784, + 0x3516ce, + 0x248a85, + 0x3aa648, + 0x397047, + 0x39704a, + 0x207243, + 0x280b47, + 0x207245, + 0x22d9c4, + 0x2d1106, + 0x2d1107, + 0x2db144, + 0x2eef07, + 0x303644, + 0x200f84, + 0x391a46, + 0x25a344, + 0x32e046, + 0x229cc3, + 0x38f308, + 0x3ca508, + 0x232803, + 0x3bb243, + 0x3b3284, + 0x3b79c3, + 0xde48182, + 0xe28bac2, + 0x205a83, + 0x203986, + 0x2041c3, + 0x22f104, + 0xe73e842, + 0x355843, + 0x33e843, + 0x214f82, + 0xea06a82, + 0x2c5706, + 0x232d07, + 0x2f0fc7, + 0x39bec5, + 0x209e04, + 0x28dcc5, + 0x288407, + 0x302889, + 0x2d27c6, + 0x2e44c8, + 0x2ec986, + 0xee14d42, + 0x384ac8, + 0x2fb306, + 0x334c05, + 0x3cf307, + 0x3183c4, + 0x3183c5, + 0x279b44, + 0x392f88, + 0xf208002, + 0xf600482, + 0x334906, + 0x200488, + 0x352485, + 0x353406, + 0x355bc8, + 0x374b48, + 0xfa07d85, + 0xfe6dd04, + 0x381507, + 0x1020b542, + 0x10742382, + 0x11a08e02, + 0x310dc5, + 0x2a3c05, + 0x2564c6, + 0x2be307, + 0x3ae0c7, + 0x12208e03, + 0x29d007, + 0x2eac08, + 0x1b62ae49, + 0x3968c7, + 0x22bb07, + 0x22c588, + 0x22cd86, + 0x22d4c6, + 0x22e10c, + 0x22f70a, + 0x230087, + 0x231ecb, + 0x232b47, + 0x232b4e, + 0x1ba33ac4, + 0x233e84, + 0x236b87, + 0x2606c7, + 0x23df06, + 0x23df07, + 0x23e787, + 0x1be2a502, + 0x2407c6, + 0x2407ca, + 0x240d4b, + 0x2424c7, + 0x243085, + 0x243583, + 0x243c06, + 0x243c07, + 0x273203, + 0x1c200102, + 0x24448a, + 0x1c76fd82, + 0x1ca481c2, + 0x1ce46682, + 0x1d234942, + 0x2476c5, + 0x247e84, + 0x1da18982, + 0x2669c5, + 0x241343, + 0x20bbc5, + 0x219484, + 0x21fec4, + 0x30a506, + 0x31a186, + 0x208143, + 0x3b7284, + 0x328d43, + 0x1ea08a42, + 0x220384, + 0x381a86, + 0x220385, + 0x2d0006, + 0x3cf408, + 0x2358c4, + 0x230308, + 0x3a7785, + 0x2438c8, + 0x2b5386, + 0x347d87, + 0x2474c4, + 0x2474c6, + 0x29d303, + 0x3a2043, + 0x31c148, + 0x32ed84, + 0x35b5c7, + 0x1fe02186, + 0x2dda09, + 0x331908, + 0x33e8c8, + 0x39df04, + 0x210543, + 0x22eac2, + 0x2020a182, + 0x206130c2, + 0x213ac3, + 0x20a1d202, + 0x393c84, + 0x249f86, + 0x324c85, + 0x29f643, + 0x22b184, + 0x2b5d07, + 0x38a583, + 0x23bb08, + 0x221bc5, + 0x25da03, + 0x277e45, + 0x277f84, + 0x3015c6, + 0x226804, + 0x228c06, + 0x353fc6, + 0x2bda04, + 0x232f03, + 0x20e1d602, + 0x232645, + 0x200843, + 0x21202242, + 0x22e0c3, + 0x218f05, + 0x231103, + 0x231109, + 0x21600942, + 0x21e21742, + 0x28b3c5, + 0x215f46, + 0x2a52c6, + 0x2c5d08, + 0x2c5d0b, + 0x2039cb, + 0x26d545, + 0x39c0c5, + 0x2cba89, + 0x1601042, + 0x2cfc88, + 0x204544, + 0x22602bc2, + 0x2182c3, + 0x22e60886, + 0x23da48, + 0x23200c02, + 0x223dc8, + 0x23605642, + 0x2bc08a, + 0x23ad1703, + 0x205246, + 0x35c6c8, + 0x30b508, + 0x2d4146, + 0x37eb07, + 0x3adbc7, + 0x24f2ca, + 0x377944, + 0x358b04, + 0x36e249, + 0x243ad3c5, + 0x27a546, + 0x226003, + 0x24fd84, + 0x246353c4, + 0x3949c7, + 0x233cc7, + 0x2bb1c4, + 0x2d3185, + 0x256588, + 0x2484c7, + 0x248847, + 0x24a17242, + 0x312f44, + 0x290e08, + 0x24a344, + 0x24ba04, + 0x24c045, + 0x24d647, + 0x25bf49, + 0x24e2c4, + 0x24e849, + 0x24ea88, + 0x24fb04, + 0x24fb07, + 0x24e500c3, + 0x250247, + 0x1621e82, + 0x16ae902, + 0x250dc6, + 0x251407, + 0x252584, + 0x253a47, + 0x254647, + 0x254dc3, + 0x22ec42, + 0x204102, + 0x26fb03, + 0x26fb04, + 0x26fb0b, + 0x36b248, + 0x25cd84, + 0x258745, + 0x259747, + 0x25afc5, + 0x2cf5ca, + 0x25ccc3, + 0x25201482, + 0x21f584, + 0x260489, + 0x264883, + 0x264947, + 0x3cd809, + 0x218bc8, + 0x23f883, + 0x27db87, + 0x27e209, + 0x26c043, + 0x285604, + 0x2864c9, + 0x2896c6, + 0x354383, + 0x206502, + 0x238cc3, + 0x3c7a87, + 0x2df1c5, + 0x387c46, + 0x257384, + 0x2e7d45, + 0x21bd83, + 0x213586, + 0x20afc2, + 0x3aea44, + 0x25623642, + 0x25a67d03, + 0x25e028c2, + 0x24b903, + 0x212a44, + 0x212a47, + 0x3cc886, + 0x27bc42, + 0x26236142, + 0x3cf604, + 0x2662a642, + 0x26a00ac2, + 0x2b2944, + 0x2b2945, + 0x202b05, + 0x361146, + 0x26e0ccc2, + 0x20ccc5, + 0x20dac5, + 0x20e903, + 0x212bc6, + 0x21d6c5, + 0x22a842, + 0x353045, + 0x22a844, + 0x235803, + 0x235a43, + 0x27207702, + 0x2dc347, + 0x2d6244, + 0x2d6249, + 0x24fc84, + 0x285403, + 0x358409, + 0x285408, + 0x276a3a84, + 0x2a3a86, + 0x2a8bc3, + 0x21ac03, + 0x213ec3, + 0x27af9602, + 0x2fc0c2, + 0x27e00642, + 0x3394c8, + 0x275608, + 0x3b26c6, + 0x26f745, + 0x2809c5, + 0x38c3c7, + 0x236185, + 0x259a02, + 0x28294f42, + 0x28600042, + 0x238508, + 0x384a05, + 0x2f23c4, + 0x379a85, + 0x249b47, + 0x272084, + 0x244382, + 0x28a2f682, + 0x348e04, + 0x235687, + 0x3cd347, + 0x33e004, + 0x292e83, + 0x232744, + 0x232748, + 0x22d806, + 0x2d0f8a, + 0x396c44, + 0x293388, + 0x289b04, + 0x220106, + 0x294f04, + 0x3110c6, + 0x2d6509, + 0x234bc7, + 0x217b03, + 0x28e01742, + 0x39e183, + 0x20bf02, + 0x29239442, + 0x314ac6, + 0x378808, + 0x2a5447, + 0x2fe249, + 0x2928c9, + 0x2a6e05, + 0x2a7d89, + 0x2a8545, + 0x2a8689, + 0x2a9c05, + 0x2aa8c8, + 0x2960e604, + 0x29a54f07, + 0x22bec3, + 0x2aaac7, + 0x22bec6, + 0x2aaf87, + 0x2a2dc5, + 0x2ec403, + 0x29e2f4c2, + 0x20ecc4, + 0x2a22a682, + 0x2a655402, + 0x2f5086, + 0x279c85, + 0x2ae587, + 0x328c83, + 0x33bdc4, 0x210e03, - 0x2333c3, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x214bc3, - 0x20ec83, - 0x238184, - 0xaff03, - 0x241d03, - 0x209944, - 0x2bdd85, - 0x15da87, - 0x202c42, - 0x209d42, - 0x202542, - 0x2032c2, - 0xae43, - 0x200342, - 0x20be03, - 0x23d744, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x226444, - 0x20ec83, - 0xae43, - 0x241d03, - 0x207c03, - 0x28b304, - 0xcd588, - 0x20be03, - 0x20ae43, - 0x10ec3, - 0x13b5c4, - 0x253404, - 0xcd588, - 0x20be03, - 0x254a04, - 0x26ff84, - 0x20ae43, - 0x2056c2, - 0x241d03, - 0x24f3c3, - 0x702c4, - 0x2f5805, - 0x20e982, - 0x30ec03, - 0xf89, - 0xd3686, - 0xfcc8, - 0x200742, - 0xcd588, - 0x202c42, - 0x237583, - 0x30e843, - 0x201342, - 0xae43, - 0x241d03, - 0x200742, - 0x1b6447, - 0x11c889, - 0x5483, - 0xcd588, - 0x1226c3, - 0x5f33d587, - 0xbe03, - 0x1c6548, - 0x237583, - 0x30e843, - 0x178d46, - 0x214bc3, - 0x5b388, - 0xc0248, - 0x40e06, - 0x21f743, - 0xc6788, - 0x97c03, - 0xdbd45, - 0x37787, - 0xec83, - 0x6c83, - 0x41d03, - 0x4bc2, - 0x16c18a, - 0x1c0e43, - 0x30c5c4, - 0x105e0b, - 0x1063c8, - 0x8d302, - 0x200742, - 0x202c42, - 0x20be03, - 0x237583, - 0x2d2484, - 0x30e843, - 0x214bc3, - 0x21f743, - 0x20ec83, - 0x20be03, - 0x237583, - 0x30e843, - 0x20be83, - 0x20ec83, - 0x241d03, - 0x209943, - 0x207c03, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x10ec3, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x20be83, - 0x20ec83, - 0x241d03, - 0x230882, - 0x200101, - 0x200742, - 0x200301, - 0x32a782, - 0xcd588, - 0x21fe85, - 0x200781, - 0xbe03, - 0x2014c1, - 0x200041, - 0x200141, - 0x24e082, - 0x378184, - 0x24e083, - 0x201401, - 0x200901, - 0x200541, - 0x200a81, - 0x316307, - 0x337dcf, - 0x2fa486, - 0x200641, - 0x33b606, - 0x200081, - 0x2001c1, - 0x3c35ce, - 0x200341, - 0x241d03, - 0x201681, - 0x254285, - 0x204bc2, - 0x3a25c5, - 0x2002c1, - 0x200a01, - 0x200401, - 0x20e982, - 0x200441, - 0x203f81, - 0x20d601, - 0x201181, - 0x200dc1, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x21a003, - 0x20be03, - 0x30e843, - 0x8d248, - 0x21f743, - 0x20ec83, - 0x5e8c3, - 0x241d03, - 0x14e0f48, - 0x10d08, - 0xcd588, - 0xae43, - 0x24704, - 0x4cec4, - 0x14e0f4a, - 0xcd588, - 0xaff03, - 0x20be03, - 0x237583, - 0x30e843, - 0x20ec83, - 0x241d03, - 0x205583, - 0xcd588, - 0x20be03, - 0x237583, - 0x2d2484, - 0x241d03, - 0x252385, - 0x317c44, - 0x20be03, - 0x20ec83, - 0x241d03, - 0x2fc8a, - 0xfd504, - 0x112a46, - 0x202c42, - 0x20be03, - 0x234d09, - 0x237583, - 0x2a82c9, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x2e9288, - 0x397b87, - 0x2f5805, - 0x1b7888, - 0x1b6447, - 0x19848a, - 0x101d0b, - 0x13b847, - 0x45388, - 0x3a80a, - 0x13dc8, - 0x11c889, - 0x2b847, - 0x67fc7, - 0x1c28c8, - 0x1c6548, - 0x470cf, - 0x26505, - 0x1c6847, - 0x178d46, - 0x4c207, - 0x108186, - 0x5b388, - 0x9b986, - 0x1187c7, - 0x142349, - 0x1b5207, - 0xe68c9, - 0xb8209, - 0xbdb06, - 0xc0248, - 0xbeb45, - 0x77fca, - 0xc6788, - 0x97c03, - 0xcea08, - 0x37787, - 0x1ac045, - 0x4dc90, - 0x6c83, - 0xaff03, - 0x1c3147, - 0x1d5c5, - 0xe4308, - 0x605c5, - 0x1c0e43, - 0x142748, - 0x132146, - 0x199bc9, - 0xaab87, - 0x124b, - 0x137a84, - 0xfe3c4, - 0x105e0b, - 0x1063c8, - 0x107e07, - 0x117485, - 0x20be03, - 0x237583, - 0x203d43, - 0x241d03, - 0x244a03, - 0x30e843, - 0xaff03, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x7e24b, - 0x200742, - 0x202c42, - 0x241d03, - 0xcd588, - 0x200742, - 0x202c42, - 0x202542, - 0x201342, - 0x200b82, - 0x20ec83, - 0x200342, - 0x200742, - 0x2b6c03, - 0x202c42, - 0x20be03, - 0x237583, - 0x202542, - 0x30e843, - 0x214bc3, - 0x21f743, - 0x226444, - 0x20ec83, - 0x207783, - 0x241d03, - 0x30c5c4, - 0x203f83, - 0x30e843, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x20ae43, - 0x241d03, - 0x3afd87, - 0x20be03, - 0x279947, - 0x2e6686, - 0x216543, - 0x208883, - 0x30e843, - 0x204d03, - 0x26ff84, - 0x38b344, - 0x2b9906, - 0x20c743, - 0x20ec83, - 0x241d03, - 0x252385, - 0x309e84, - 0x320dc3, - 0x20d203, - 0x3c0507, - 0x23b1c5, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x98747, - 0x2149c2, - 0x26e443, - 0x20df43, - 0x2b6c03, - 0x6760be03, - 0x20b2c2, - 0x237583, - 0x206c03, - 0x30e843, - 0x26ff84, - 0x3c32c3, - 0x3371c3, - 0x21f743, - 0x226444, - 0x67a02f02, - 0x20ec83, - 0x241d03, - 0x235cc3, - 0x214c43, - 0x230882, - 0x203f83, - 0xcd588, - 0x30e843, - 0x10ec3, - 0x31b0c4, - 0x2b6c03, - 0x202c42, - 0x20be03, - 0x23d744, - 0x237583, - 0x30e843, - 0x26ff84, - 0x214bc3, - 0x39e304, - 0x21a484, - 0x2d0e06, - 0x226444, - 0x20ec83, - 0x241d03, - 0x219543, - 0x261146, - 0x4170b, - 0x29546, - 0xeb94a, - 0x10b34a, - 0xcd588, - 0x21d0c4, - 0x68e0be03, - 0x2b6bc4, - 0x237583, - 0x268984, - 0x30e843, - 0x357d43, - 0x21f743, - 0x20ec83, - 0xaff03, - 0x241d03, - 0x55a43, - 0x33840b, - 0x3b1a0a, - 0x3c580c, - 0xd80c8, - 0x200742, - 0x202c42, - 0x202542, - 0x232585, - 0x26ff84, - 0x201e42, - 0x21f743, - 0x21a484, - 0x2032c2, - 0x200342, - 0x207c02, - 0x230882, - 0xb6c03, - 0x4d5c2, - 0x386389, - 0x3a3088, - 0x310449, - 0x34e049, - 0x23e48a, - 0x24e90a, - 0x219382, - 0x2efa02, - 0x2c42, - 0x20be03, - 0x230a42, - 0x246fc6, - 0x368802, - 0x207582, - 0x30208e, - 0x21948e, - 0x27bf47, - 0x20ec07, - 0x2e89c2, - 0x237583, - 0x30e843, - 0x209182, - 0x201342, - 0x6ff83, - 0x23d94f, - 0x247302, - 0x2f9507, - 0x2ad447, - 0x314407, - 0x2b0fcc, - 0x2b9b8c, - 0x207144, - 0x27d34a, - 0x2193c2, - 0x209602, - 0x2b9844, - 0x2028c2, - 0x2c10c2, - 0x2b9dc4, - 0x217f42, - 0x201542, - 0xf003, - 0x29ba07, - 0x233805, - 0x20e302, + 0x312643, + 0x2aa00d42, + 0x2b207782, + 0x382444, + 0x22ec03, + 0x2471c5, + 0x2b60ad02, + 0x2be02ec2, + 0x2ffd86, + 0x32eec4, + 0x301b44, + 0x301b4a, + 0x2c6005c2, + 0x269f03, + 0x2094ca, + 0x2171c8, + 0x2ca20a04, + 0x2005c3, + 0x208643, + 0x2555c9, + 0x254b89, + 0x278586, + 0x2ce17383, + 0x21da05, + 0x3305cd, + 0x217386, + 0x22184b, + 0x2d2032c2, + 0x21bc08, + 0x2fa0f502, + 0x2fe01142, + 0x2df4c5, + 0x30200b02, + 0x24bc47, + 0x2b3e87, + 0x208543, + 0x325908, + 0x30602a02, + 0x2a0d04, + 0x293083, + 0x389085, + 0x3a1143, + 0x241446, + 0x21cd04, + 0x3bb203, + 0x2b1d03, + 0x30a06e82, + 0x39c044, + 0x3b9305, + 0x3bdd87, + 0x27bd03, + 0x2aed83, + 0x2b14c3, + 0x160a6c2, + 0x2b1583, + 0x2b1c83, + 0x30e05e42, + 0x30ffc4, + 0x31a386, + 0x359e03, + 0x2b1fc3, + 0x312b2d02, + 0x2b2d08, + 0x2b3b04, + 0x310586, + 0x25ddc7, + 0x362c46, + 0x2d5844, + 0x3ee01782, + 0x22bd8b, + 0x2f6d4e, + 0x216acf, + 0x300e83, + 0x3f65e242, + 0x1642ac2, + 0x3fa03cc2, + 0x258f03, + 0x203cc3, + 0x302b06, + 0x2d2ec6, + 0x274f87, + 0x300804, + 0x3fe16082, + 0x402220c2, + 0x2499c5, + 0x2f6107, + 0x3bbf86, + 0x4060e842, + 0x20e844, + 0x2b8c43, + 0x40a06f42, + 0x40f6a383, + 0x2b9504, + 0x2c1789, + 0x16c74c2, + 0x41214882, + 0x332985, + 0x416c79c2, + 0x41a00e42, + 0x357dc7, + 0x210fc9, + 0x36e94b, + 0x3ad985, + 0x26ab09, + 0x386346, + 0x208d47, + 0x41e0fa84, + 0x211d89, + 0x3428c7, + 0x211a47, + 0x223f03, + 0x2b27c6, + 0x3176c7, + 0x2450c3, + 0x3c0c46, + 0x4260d9c2, + 0x42a31382, + 0x39e2c3, + 0x33b9c5, + 0x38eac7, + 0x21df46, + 0x2df145, + 0x256984, + 0x27cf05, + 0x2fbac4, + 0x42e04382, + 0x326407, + 0x2c64c4, + 0x25b244, + 0x3cac0d, + 0x2d86c9, + 0x22a5c8, + 0x273a04, + 0x34c285, + 0x39e907, + 0x204384, + 0x31f507, + 0x214605, + 0x43215804, + 0x2b16c5, + 0x263544, + 0x3010c6, + 0x2be105, + 0x4360e802, + 0x3a3f43, + 0x2df284, + 0x2df285, + 0x343fc6, + 0x32e8c5, + 0x2b0184, + 0x259dc3, + 0x21d146, + 0x31b405, + 0x31d885, + 0x2be204, + 0x396cc3, + 0x396ccc, + 0x43afcb42, + 0x43e0cf82, + 0x44204c82, + 0x221303, + 0x221304, + 0x4460bb02, + 0x305b88, + 0x387d05, + 0x245b84, + 0x361e86, + 0x44a11602, + 0x44e27fc2, + 0x45202d82, + 0x2a5845, + 0x2bd8c6, + 0x235304, + 0x20f806, + 0x2f0706, + 0x22ef43, + 0x4572228a, + 0x26ae85, + 0x25d203, + 0x221106, + 0x383dc9, + 0x221107, + 0x2a17c8, + 0x2d5449, + 0x36bf48, + 0x2ecf86, + 0x207043, + 0x45a9d082, + 0x3a3408, + 0x45e4eb42, + 0x46202302, + 0x208e83, + 0x2e4345, + 0x26bbc4, + 0x252b89, + 0x2eff84, + 0x2141c8, + 0x20bf43, + 0x46b946c4, + 0x2a6c03, + 0x215f88, + 0x3cab47, + 0x46e0e8c2, + 0x239ec2, + 0x32b3c5, + 0x266109, + 0x270203, + 0x280584, + 0x330584, + 0x22dc03, + 0x28128a, + 0x47327182, + 0x4760eb02, + 0x2cb2c3, + 0x3865c3, + 0x161a142, + 0x3aa303, + 0x47a26942, + 0x47e01942, + 0x4820ac44, + 0x20ac46, + 0x2fffc6, + 0x241c44, + 0x279883, + 0x2069c3, + 0x2f73c3, + 0x2410c6, + 0x326b85, + 0x2cb447, + 0x2ce5c5, + 0x2cf846, + 0x2d04c8, + 0x2d06c6, + 0x201944, + 0x29988b, + 0x2d67c3, + 0x2d67c5, + 0x2d6c48, + 0x227382, + 0x3580c2, + 0x48647742, + 0x48a02382, + 0x2160c3, + 0x48e6cac2, + 0x26cac3, + 0x2d7583, + 0x49603342, + 0x49adbd86, + 0x25ae46, + 0x49edbec2, + 0x4a20cb02, + 0x4a635a82, + 0x4aa07c02, + 0x4ae21482, + 0x4b200a42, + 0x2149c3, + 0x381e05, + 0x34c406, + 0x4b6bf144, + 0x38188a, + 0x3abbc6, + 0x2e6844, + 0x239383, + 0x4c208dc2, + 0x201402, + 0x22f1c3, + 0x4c61d283, + 0x300d07, + 0x2be007, + 0x4de6fc07, + 0x3c60c7, + 0x226bc3, + 0x38808a, + 0x397244, + 0x3ae204, + 0x3ae20a, + 0x242ec5, + 0x4e217202, + 0x250d83, + 0x4e600602, + 0x24fc43, + 0x39e143, + 0x4ee00582, + 0x29cf84, + 0x21b784, + 0x210805, + 0x30e7c5, + 0x31f686, + 0x347506, + 0x4f23bbc2, + 0x4f601c02, + 0x3c8c05, + 0x25ab52, + 0x349906, + 0x289a43, + 0x365986, + 0x350705, + 0x160d742, + 0x57a0e142, + 0x3643c3, + 0x20e143, + 0x288203, + 0x57e0b142, + 0x22ab03, + 0x58201202, + 0x20ac83, + 0x310008, + 0x256503, + 0x2a6c86, + 0x237007, + 0x313786, + 0x31378b, + 0x2e6787, + 0x2fab84, + 0x58a02d42, + 0x387b85, + 0x58e1d243, + 0x2a8b43, + 0x2bc285, + 0x387f83, + 0x59387f86, + 0x2cfe0a, + 0x2418c3, + 0x20f1c4, + 0x2003c6, + 0x335006, + 0x59659003, + 0x33bc87, + 0x278487, + 0x29b3c5, + 0x3a1bc6, + 0x2a1903, + 0x5c212e03, + 0x5c60a282, + 0x267e84, + 0x213b49, + 0x238307, + 0x229585, + 0x243204, + 0x376f48, + 0x2459c5, + 0x5ca4e545, + 0x287389, + 0x3996c3, + 0x248144, + 0x5ce09302, + 0x2162c3, + 0x5d28f002, + 0x28f006, + 0x1625f42, + 0x5d607b02, + 0x2a5748, + 0x2c4003, + 0x2b1607, + 0x350b05, + 0x2c3bc5, + 0x313a0b, + 0x2e51c6, + 0x313c06, + 0x2e63c6, + 0x28ac04, + 0x2c1986, + 0x5dad9088, + 0x231843, + 0x203043, + 0x203044, + 0x30ab04, + 0x30bac7, + 0x2ea545, + 0x5deea682, + 0x5e2059c2, + 0x2059c5, + 0x2ec504, + 0x2ec50b, + 0x2edcc8, + 0x24dfc4, + 0x5ea0e882, + 0x5ee4df42, + 0x2b2f43, + 0x2ee404, + 0x2ee6c5, + 0x2ef0c7, + 0x2f1f04, + 0x26d584, + 0x5f203b02, + 0x372649, + 0x2f3505, + 0x3adc45, + 0x2f4085, + 0x5f616203, + 0x2f5dc4, + 0x2f5dcb, + 0x2f6604, + 0x2f68cb, + 0x2f7305, + 0x216c0a, + 0x2f7ac8, + 0x2f7cca, + 0x2f8283, + 0x2f828a, + 0x5fe5c8c2, + 0x6023eec2, + 0x60681ac3, + 0x60afb282, + 0x2fb283, + 0x60f73dc2, + 0x61338042, + 0x2fb944, + 0x217846, + 0x20f545, + 0x2fc503, + 0x32bb06, + 0x20f045, + 0x2afcc4, + 0x61600902, + 0x2fec84, + 0x2cb70a, + 0x23f187, + 0x36c846, + 0x318f87, + 0x202103, + 0x2b9548, + 0x3ad60b, + 0x2c1ec5, + 0x3473c5, + 0x3473c6, + 0x2e9804, + 0x3b73c8, + 0x22dfc3, + 0x266cc4, + 0x35ae07, + 0x2fa7c6, + 0x38b386, + 0x35150a, + 0x23a844, + 0x23a84a, + 0x61b28386, + 0x328387, + 0x2587c7, + 0x273844, + 0x273849, + 0x31a045, + 0x36bb4b, + 0x2ec283, + 0x228dc3, + 0x61e1bbc3, + 0x22dbc4, + 0x62200682, + 0x3226c6, + 0x627a4f05, + 0x365bc5, + 0x254206, + 0x29db44, + 0x62a01a42, + 0x2435c4, + 0x62e08782, + 0x3357c5, + 0x237804, + 0x63a24603, + 0x63e0e182, + 0x20e183, + 0x353606, + 0x64203c42, + 0x226588, + 0x220f84, + 0x220f86, + 0x386e46, + 0x208804, + 0x21d0c5, + 0x26dc08, + 0x2af987, + 0x322a47, + 0x322a4f, + 0x290d06, + 0x23e903, + 0x23e904, + 0x245ac4, + 0x20dbc3, + 0x220244, + 0x2547c4, + 0x6460ed02, + 0x28b7c3, + 0x2591c3, + 0x64a11302, + 0x273b03, + 0x393d43, + 0x21130a, + 0x38f707, + 0x25c64c, + 0x25c906, + 0x25d486, + 0x25dac7, + 0x64e2c9c7, + 0x26a249, + 0x246ac4, + 0x26c784, + 0x65216102, + 0x65601002, + 0x3518c6, + 0x33ba84, + 0x28bc46, + 0x22ce48, + 0x23d884, + 0x24bc86, + 0x2a5285, + 0x2902c8, + 0x203bc3, + 0x292ac5, + 0x293e03, + 0x3add43, + 0x3add44, + 0x21f543, + 0x65a5e142, + 0x65e01702, + 0x2ec149, + 0x29d485, + 0x29ef84, + 0x2a2f05, + 0x21a444, + 0x2c8b87, + 0x359c45, + 0x6626fdc4, + 0x26fdc8, + 0x2eba46, + 0x2ec3c4, + 0x2efe08, + 0x2f1647, + 0x66605502, + 0x2f6044, + 0x20dc84, + 0x2bbc07, + 0x66a05504, + 0x255a42, + 0x66e11782, + 0x21c883, + 0x2dfe84, + 0x29c083, + 0x29c085, + 0x6722b5c2, + 0x2fc3c5, + 0x2701c2, + 0x399e85, + 0x2bba85, + 0x67614742, + 0x33e7c4, + 0x67a00b42, + 0x25f386, + 0x31fe06, + 0x266248, + 0x2c2e08, + 0x2f5004, + 0x303105, + 0x342a49, + 0x39c144, + 0x2cfdc4, + 0x212503, + 0x67e4f405, + 0x3789c7, + 0x24ac05, + 0x2a3d04, + 0x3a8f8d, + 0x374402, + 0x39a203, + 0x3b1a43, + 0x68201dc2, + 0x3a6a45, + 0x21cf47, + 0x2baa04, + 0x3c6187, + 0x2d5649, + 0x2cb849, + 0x277107, + 0x28d7c3, + 0x31ffc8, + 0x268ac9, + 0x3ba087, + 0x2fc845, + 0x2fd786, + 0x2fdd86, + 0x2fdf05, + 0x2d87c5, + 0x68600c82, + 0x2b0c85, + 0x2b6e48, + 0x2c54c6, + 0x68a024c7, + 0x2bb104, + 0x31c587, + 0x300986, + 0x68e0fdc2, + 0x343cc6, + 0x304a0a, + 0x305285, + 0x692e6a02, + 0x69692782, + 0x317a06, + 0x2b6948, + 0x69bcd507, + 0x69e18902, + 0x219503, + 0x209246, + 0x2249c4, + 0x3c1346, + 0x202806, + 0x20184a, + 0x38b4c5, + 0x2f8dc6, + 0x388543, + 0x388544, + 0x203942, + 0x32ee43, + 0x6a221342, + 0x2f8743, + 0x209744, + 0x2b6a84, + 0x2b6a8a, + 0x219fc3, + 0x279a48, + 0x2ed04a, + 0x237a87, + 0x307f86, + 0x25f244, + 0x2917c2, + 0x2a4402, + 0x6a6007c2, + 0x232703, + 0x258587, + 0x2007c7, + 0x287244, + 0x3b0047, + 0x2ef1c6, + 0x22a9c7, + 0x303d44, + 0x385005, + 0x218745, + 0x6aa0a682, + 0x20a686, + 0x2174c3, + 0x21cb82, + 0x21cb86, + 0x6ae00e02, + 0x6b2022c2, + 0x3c4805, + 0x6b60fec2, + 0x6ba019c2, + 0x32f005, + 0x2cd585, + 0x2a6085, + 0x6be5e603, + 0x24a045, + 0x2e5287, + 0x3772c5, + 0x34ba85, + 0x3aa744, + 0x3224c6, + 0x22b3c4, + 0x6c2008c2, + 0x6cf7b0c5, + 0x2a47c7, + 0x39e5c8, + 0x250606, + 0x25060d, + 0x254949, + 0x254952, + 0x2ff345, + 0x307543, + 0x6d202442, + 0x315dc4, + 0x217403, + 0x3446c5, + 0x305f05, + 0x6d62a342, + 0x25da43, + 0x6da5a902, + 0x6e2c2742, + 0x6e600082, + 0x2e2d05, + 0x3c62c3, + 0x24a948, + 0x6ea076c2, + 0x6ee067c2, + 0x29cf46, + 0x35c20a, + 0x214b43, + 0x259d43, + 0x342cc3, + 0x6fe03642, + 0x7e21a642, + 0x7ea10602, + 0x204b02, + 0x343ac9, + 0x2c6904, + 0x2a9f08, + 0x7eefc542, + 0x7f202b82, + 0x2acd05, + 0x232308, + 0x316d48, + 0x34d2cc, + 0x2379c3, + 0x7f617b82, + 0x7fa05ec2, + 0x281c46, + 0x308e05, + 0x274683, + 0x27bb06, + 0x308f46, + 0x2c6b43, + 0x30a8c3, + 0x30af86, + 0x30c444, + 0x26cfc6, + 0x21d845, + 0x21d84a, 0x24c184, - 0x37c0c2, - 0x2d7c88, - 0x20ec83, - 0x39f108, - 0x206a82, - 0x207305, - 0x388206, - 0x241d03, - 0x208302, - 0x2e51c7, - 0x4bc2, - 0x272585, - 0x204905, - 0x212182, - 0x2030c2, - 0x293c0a, - 0x288f0a, - 0x23a382, - 0x29a184, - 0x2040c2, - 0x20fa08, - 0x200d82, - 0x39d588, - 0x302ac7, - 0x3038c9, - 0x204982, - 0x3086c5, - 0x36ba05, - 0x21b60b, - 0x2c418c, - 0x230548, - 0x31c188, - 0x2145c2, - 0x355bc2, - 0x200742, - 0xcd588, - 0x202c42, - 0x20be03, + 0x30cb04, + 0x30d24a, + 0x7fe05082, + 0x24c305, + 0x30e04a, + 0x30ea05, + 0x30f2c4, + 0x30f3c6, + 0x30f544, + 0x216586, + 0x8020fe02, + 0x2f4a06, + 0x326945, + 0x389807, + 0x3aaf06, + 0x25dcc4, + 0x2dd487, + 0x3221c6, + 0x234f05, + 0x239a87, + 0x3b8c87, + 0x3b8c8e, + 0x27b406, + 0x31f3c5, + 0x205447, + 0x20cb43, + 0x20cb47, + 0x2252c5, + 0x229ec4, + 0x233c82, + 0x2451c7, + 0x300884, + 0x240bc4, + 0x28620b, + 0x21b083, + 0x2d0807, + 0x21b084, + 0x2efc87, + 0x2921c3, + 0x3465cd, + 0x3a7608, + 0x235fc4, + 0x26fcc5, + 0x314205, + 0x314643, + 0x80620e82, + 0x316043, + 0x316a03, + 0x20a804, + 0x27e305, + 0x217547, + 0x3885c6, + 0x383c03, + 0x235acb, + 0x2740cb, + 0x2a828b, + 0x30260b, + 0x2e6a4a, + 0x33030b, + 0x3679cb, + 0x39838c, + 0x3cf08b, + 0x3d1911, + 0x31734a, + 0x317b8b, + 0x317e4c, + 0x31814b, + 0x31a8ca, + 0x31b54a, + 0x31ca8e, + 0x31dbcb, + 0x31de8a, + 0x320bd1, + 0x32100a, + 0x32150b, + 0x321a4e, + 0x32330c, + 0x32378b, + 0x323a4e, + 0x323dcc, + 0x329e4a, + 0x32ad8c, + 0x80b2b08a, + 0x32bc88, + 0x32c849, + 0x32f3ca, + 0x32f64a, + 0x32f8cb, + 0x3334ce, + 0x334511, + 0x33d349, + 0x33d58a, + 0x33dccb, + 0x34048a, + 0x340d16, + 0x34208b, + 0x34260a, + 0x34314a, + 0x3437cb, + 0x344149, + 0x346f49, + 0x34824d, + 0x348b8b, + 0x349a8b, + 0x34a44b, + 0x34a909, + 0x34af4e, + 0x34bc4a, + 0x34c98a, + 0x34cdca, + 0x34d90b, + 0x34e14b, + 0x34edcd, + 0x35218d, + 0x352cd0, + 0x35318b, + 0x354d4c, + 0x35594b, + 0x3578cb, + 0x358fce, + 0x35974b, + 0x35974d, + 0x35fc4b, + 0x3606cf, + 0x360a8b, + 0x3612ca, + 0x361809, + 0x362009, + 0x80f62dcb, + 0x36308e, + 0x36488b, + 0x36658f, + 0x3685cb, + 0x36888b, + 0x368b4b, + 0x36910a, + 0x36e549, + 0x37138f, + 0x376b4c, + 0x37740c, + 0x377a4e, + 0x377f4f, + 0x37830e, + 0x378b90, + 0x378f8f, + 0x37a68e, + 0x37b24c, + 0x37b552, + 0x37bfd1, + 0x37c7ce, + 0x37cc4e, + 0x37d18b, + 0x37d18e, + 0x37d50f, + 0x37d8ce, + 0x37dc53, + 0x37e111, + 0x37e54c, + 0x37e84e, + 0x37eccc, + 0x37f213, + 0x37f8d0, + 0x38044c, + 0x38074c, + 0x380c0b, + 0x38204e, + 0x38254b, + 0x382e0b, + 0x38400c, + 0x3977ca, + 0x397b8c, + 0x397e8c, + 0x398189, + 0x39978b, + 0x399a48, + 0x39a2c9, + 0x39a2cf, + 0x39ba4b, + 0x8139c84a, + 0x39f1cc, + 0x3a038b, + 0x3a0649, + 0x3a0f88, + 0x3a180b, + 0x3a1e0b, + 0x3a298a, + 0x3a2c0b, + 0x3a318c, + 0x3a3b48, + 0x3a7dcb, + 0x3aab4b, + 0x3ac7ce, + 0x3ade4b, + 0x3af1cb, + 0x3b880b, + 0x3b8ac9, + 0x3b900d, + 0x3c1b4a, + 0x3c4157, + 0x3c4e98, + 0x3c8509, + 0x3c9b4b, + 0x3caf54, + 0x3cb44b, + 0x3cb9ca, + 0x3cc08a, + 0x3cc30b, + 0x3ccb50, + 0x3ccf51, + 0x3cda4a, + 0x3ce68d, + 0x3ced8d, + 0x3d1d4b, + 0x3d2c06, + 0x20a783, + 0x81763703, + 0x2af846, + 0x241805, + 0x27eb07, + 0x3301c6, + 0x1660842, + 0x2aeec9, + 0x32b904, + 0x2e4708, + 0x21bb03, + 0x315d07, + 0x202cc2, + 0x2ae5c3, + 0x81a01242, + 0x2ccb46, + 0x2cde04, + 0x268244, + 0x3293c3, + 0x3293c5, + 0x822c7a02, + 0x826a8a04, + 0x273787, + 0x82a5b102, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0xe9fc8, + 0x214b83, + 0x2000c2, + 0x120648, + 0x208e02, + 0x213ec3, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x211303, + 0x33b2d6, + 0x35e713, + 0x3afec9, + 0x381408, + 0x387a09, + 0x30e1c6, + 0x348e50, + 0x2446d3, + 0x2fa888, + 0x3a79c7, + 0x2b0587, + 0x27cc4a, + 0x38d689, + 0x3a3cc9, + 0x2879cb, + 0x36b686, + 0x205bca, + 0x21ef86, + 0x32b503, + 0x2dc285, + 0x38f308, + 0x25f44d, + 0x310e8c, + 0x2eca87, + 0x31ab0d, + 0x26dd04, + 0x22de8a, + 0x22f24a, + 0x22f70a, + 0x2449c7, + 0x23cf47, + 0x241dc4, + 0x2474c6, + 0x32e644, + 0x2ff9c8, + 0x2effc9, + 0x2c5d06, + 0x2c5d08, + 0x2f8acd, + 0x2cba89, + 0x30b508, + 0x3adbc7, + 0x28be4a, + 0x251406, + 0x260247, + 0x2e2304, + 0x2274c7, + 0x213eca, + 0x24200e, + 0x236185, + 0x3c2dcb, + 0x307349, + 0x254b89, + 0x208387, + 0x20838a, + 0x2bbb47, + 0x2f6e89, + 0x2c9e88, + 0x31140b, + 0x2e4345, + 0x22a48a, + 0x235849, + 0x27460a, + 0x2ce64b, + 0x2273cb, + 0x287755, + 0x2e87c5, + 0x3adc45, + 0x2f5dca, + 0x27868a, + 0x3070c7, + 0x214c83, + 0x351848, + 0x2da04a, + 0x220f86, + 0x268909, + 0x2902c8, + 0x2ec3c4, + 0x37efc9, + 0x2c2e08, + 0x2b52c7, + 0x37b0c6, + 0x2a47c7, + 0x2b8007, + 0x240ec5, + 0x235fcc, + 0x26fcc5, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x14b83, + 0x259003, + 0x208e02, + 0x208e03, + 0x21d283, + 0x214b83, + 0x259003, + 0x208e03, + 0x21d283, + 0x14b83, + 0x256503, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x120648, + 0x208e02, + 0x202142, + 0x2fab02, + 0x202a02, + 0x20a002, + 0x2d3a42, + 0x8edc6, + 0x4e08e03, + 0x231103, + 0x3d0943, + 0x213ec3, + 0x217383, + 0x270203, + 0x2dc186, + 0x21d283, + 0x259003, + 0x2323c3, + 0x120648, + 0x394c44, + 0x394487, + 0x329d83, + 0x24f784, + 0x208203, + 0x208403, + 0x213ec3, + 0xeb207, + 0x192544, + 0x191503, + 0x192e05, + 0x2000c2, + 0x18d8c3, + 0x6208e02, + 0x648a8c9, + 0x8af4d, + 0x8b28d, + 0x2fab02, + 0x20a04, + 0x192e49, + 0x2003c2, + 0x6a20908, + 0xf5544, + 0x120648, + 0x1442f02, + 0x14005c2, + 0x1442f02, + 0x150ec46, + 0x22d083, + 0x2b9343, + 0x7208e03, + 0x22de84, + 0x7631103, + 0x7a13ec3, + 0x200d42, + 0x220a04, + 0x21d283, + 0x303303, + 0x200ec2, + 0x259003, + 0x218502, + 0x2fb883, + 0x203c42, + 0x201683, + 0x290383, + 0x206582, + 0x120648, + 0x22d083, + 0x303303, + 0x200ec2, + 0x2fb883, + 0x203c42, + 0x201683, + 0x290383, + 0x206582, + 0x2fb883, + 0x203c42, + 0x201683, + 0x290383, + 0x206582, + 0x208e03, + 0x38d8c3, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x217383, + 0x270203, + 0x2bf144, + 0x21d283, + 0x259003, 0x202542, - 0x2032c2, - 0xae43, - 0x200342, - 0x241d03, + 0x216203, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x38d8c3, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x21d283, + 0x259003, + 0x2fc845, + 0x22a342, + 0x2000c2, + 0x120648, + 0x1582ac8, + 0x15ec0a, + 0x213ec3, + 0x201ec1, + 0x201f81, + 0x201e81, + 0x201ac1, + 0x235c81, + 0x211201, + 0x207ec1, + 0x218481, + 0x203241, + 0x200001, + 0x2000c1, + 0x200201, + 0xf8945, + 0x120648, + 0x200101, + 0x200d81, + 0x200501, + 0x201481, + 0x200041, + 0x200801, + 0x200181, + 0x205e41, + 0x200701, + 0x2004c1, + 0x200d01, + 0x200581, + 0x2003c1, + 0x201b81, + 0x201301, + 0x200401, + 0x200741, + 0x2007c1, + 0x200081, + 0x202b81, + 0x201fc1, + 0x20a781, + 0x202cc1, + 0x201241, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e02, + 0x208e03, + 0x231103, + 0x2003c2, + 0x259003, + 0xeb207, + 0x7a6c7, + 0x35d46, + 0x3968a, + 0x89c48, + 0x57fc8, + 0x58487, + 0x1b8006, + 0xe1f05, + 0x129cc5, + 0xc8346, + 0x3f5c6, + 0x2879c4, + 0x274347, + 0x120648, + 0x2dd584, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x32b288, + 0x38c384, + 0x231044, + 0x267c44, + 0x281b47, + 0x2d8c47, + 0x208e03, + 0x233e8b, + 0x38d0ca, + 0x36b947, + 0x23cd48, + 0x389108, + 0x231103, + 0x328847, + 0x3d0943, + 0x3ca0c8, + 0x203589, + 0x220a04, + 0x217383, + 0x2d28c8, + 0x270203, + 0x2d690a, + 0x2dc186, + 0x3abbc7, + 0x21d283, + 0x214b86, + 0x30fa08, + 0x259003, + 0x2c8486, + 0x2edf0d, + 0x2eed88, + 0x2f660b, + 0x255e86, + 0x325847, + 0x221ac5, + 0x38cb0a, + 0x21e905, + 0x24ab0a, + 0x22a342, + 0x202043, + 0x240bc4, + 0x200006, + 0x3b2083, + 0x320543, + 0x25c183, + 0x2bb483, + 0x38cd43, + 0x201902, + 0x2d22c5, + 0x2a71c9, + 0x241543, + 0x2038c3, + 0x2063c3, + 0x200201, + 0x2cfb87, + 0x2e2a45, + 0x38f243, + 0x203283, + 0x267c44, + 0x328cc3, + 0x21d788, + 0x361a43, + 0x3038cd, + 0x27b4c8, + 0x3ca6c6, + 0x32ee83, + 0x37f683, + 0x3a1a83, + 0xb608e03, + 0x230948, + 0x233e84, + 0x2424c3, + 0x200106, + 0x2461c8, + 0x20cd83, + 0x38cb43, + 0x22e0c3, + 0x231103, + 0x227a43, + 0x228a83, + 0x267f43, + 0x32ee03, + 0x223dc3, + 0x2353c3, + 0x383085, + 0x252684, + 0x2536c7, + 0x22ec42, + 0x257c43, + 0x259e86, + 0x25ca83, + 0x25d603, + 0x278d83, + 0x203103, + 0x394943, + 0x2957c7, + 0xba13ec3, + 0x245dc3, + 0x207003, + 0x203583, + 0x2171c3, + 0x2f4d43, + 0x3637c5, + 0x36af03, + 0x24b209, + 0x215403, + 0x306203, + 0xbe4b883, + 0x2a9003, + 0x222b88, + 0x2a7106, + 0x3b8686, + 0x29af86, + 0x37ff87, + 0x20a383, + 0x208e83, + 0x270203, + 0x289d46, + 0x227382, + 0x2a4003, + 0x339905, + 0x21d283, + 0x25e7c7, + 0x1614b83, + 0x23a703, + 0x231f43, + 0x229083, + 0x259003, + 0x21f786, + 0x36be86, + 0x371c43, + 0x225f03, + 0x216203, + 0x25c8c3, + 0x30a943, + 0x2fa003, + 0x2fba43, + 0x20f045, + 0x202203, + 0x28bd46, + 0x236e48, + 0x228dc3, + 0x326609, + 0x2d6048, + 0x222e48, + 0x267dc5, + 0x23064a, + 0x239c0a, + 0x23b8cb, + 0x23c908, + 0x3bb1c3, + 0x2fba83, + 0x34bb83, + 0x3487c8, + 0x3b7a43, + 0x388544, + 0x260983, + 0x2007c3, + 0x227343, + 0x2603c3, + 0x2323c3, + 0x22a342, + 0x226f83, + 0x2379c3, + 0x30ccc3, + 0x30dc84, + 0x240bc4, + 0x21d643, + 0x120648, + 0x2000c2, + 0x204342, + 0x201902, + 0x201b42, + 0x200202, + 0x203982, + 0x232782, + 0x202bc2, + 0x200382, + 0x202d82, + 0x20e8c2, + 0x202382, + 0x26cac2, + 0x20a282, + 0x2d3a42, + 0x209302, + 0x203042, + 0x203b02, + 0x20a582, + 0x204582, + 0x200682, + 0x216b42, + 0x201a42, + 0x211302, + 0x201002, + 0x212142, + 0x2019c2, + 0xc2, + 0x4342, + 0x1902, + 0x1b42, + 0x202, + 0x3982, + 0x32782, + 0x2bc2, + 0x382, + 0x2d82, + 0xe8c2, + 0x2382, + 0x6cac2, + 0xa282, + 0xd3a42, + 0x9302, + 0x3042, + 0x3b02, + 0xa582, + 0x4582, + 0x682, + 0x16b42, + 0x1a42, + 0x11302, + 0x1002, + 0x12142, + 0x19c2, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x2042, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e02, + 0x259003, + 0xd208e03, + 0x213ec3, + 0x270203, + 0xe6143, + 0x22c942, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0xe6143, + 0x259003, + 0x1242, + 0x2001c2, + 0x15c5805, + 0x20ed42, + 0x120648, + 0x8e02, + 0x232cc2, + 0x207042, + 0x239382, + 0x217202, + 0x23bbc2, + 0x129cc5, + 0x20ce82, + 0x200ec2, + 0x20b142, + 0x205d02, + 0x209302, + 0x3a3282, + 0x211782, + 0x258ec2, + 0xeb207, + 0xbe4cd, + 0xe1f89, + 0xaa60b, + 0xe5148, + 0x73e89, + 0x106c86, + 0x213ec3, + 0x120648, + 0x192544, + 0x191503, + 0x192e05, + 0x120648, + 0xdfe07, + 0x59086, + 0x192e49, + 0x1580e, + 0x18307, + 0x2000c2, + 0x2879c4, + 0x208e02, + 0x208e03, + 0x202142, + 0x231103, + 0x200382, + 0x2dd584, + 0x217383, + 0x24eb42, + 0x21d283, + 0x2003c2, + 0x259003, + 0x3adc46, + 0x32fe8f, + 0x601fc3, + 0x120648, + 0x208e02, + 0x3d0943, + 0x213ec3, + 0x270203, + 0x14b83, + 0x15808, + 0x158c8cb, + 0x1419dca, + 0x14730c7, + 0x7cd0b, + 0xeb0c5, + 0xf8945, + 0xeb207, + 0x208e02, + 0x208e03, + 0x213ec3, + 0x21d283, + 0x2000c2, + 0x204bc2, + 0x206602, + 0x10a08e03, + 0x240042, + 0x231103, + 0x221e82, + 0x223642, + 0x213ec3, + 0x259a02, + 0x274082, + 0x2a89c2, + 0x205002, + 0x28eb02, + 0x200802, + 0x203482, + 0x201742, + 0x27c042, + 0x239442, + 0x2aed82, + 0x2c7c82, + 0x217502, + 0x249a82, + 0x270203, + 0x201942, + 0x21d283, + 0x214002, + 0x28a042, + 0x259003, + 0x2415c2, + 0x211302, + 0x216102, + 0x201702, + 0x214742, + 0x2e6a02, + 0x20a682, + 0x25a902, + 0x21ccc2, + 0x31de8a, + 0x3612ca, + 0x39d78a, + 0x3d2d82, + 0x226b82, + 0x363782, + 0x10f24a09, + 0x1134238a, + 0x142c347, + 0x11600982, + 0x140d983, + 0x1a82, + 0x14238a, + 0x245a04, + 0x11e08e03, + 0x231103, + 0x24ea84, + 0x213ec3, + 0x220a04, + 0x217383, + 0x270203, + 0xe6344, + 0x193803, + 0x21d283, + 0x1caa45, + 0x214b83, + 0x259003, + 0x1522584, + 0x202203, + 0x202043, + 0x120648, + 0x1f06, + 0x15b7f84, + 0x1291c5, + 0x180ca, + 0x12ad02, + 0x1aa546, + 0x7a91, + 0x12724a09, + 0x129248, + 0x7a888, + 0xfb547, + 0x2842, + 0xf894b, + 0x187e4b, + 0x180e8a, + 0x5b0a, + 0x67c47, + 0x120648, + 0x115888, + 0xb447, + 0x1901670b, + 0x18dc7, + 0x20c2, + 0x3dbc7, + 0x134c0a, + 0x5b58f, + 0xfd0cf, + 0x142382, + 0x8e02, + 0xa3c08, + 0xf190a, + 0xdf90a, + 0x1a158a, + 0x7b2c8, + 0x1f488, + 0x5e088, + 0xdfdc8, + 0x18e248, + 0x8a42, + 0x1c588f, + 0x9ea8b, + 0x886c8, + 0x32fc7, + 0x13278a, + 0x56a8b, + 0x7c349, + 0x132687, + 0x1f388, + 0x39dcc, + 0x1118c7, + 0x17084a, + 0x667c8, + 0xfdfce, + 0x35d4e, + 0x67a8b, + 0xafd8b, + 0xe924b, + 0xeccc9, + 0x11b18b, + 0x11ee8d, + 0x135ecb, + 0x3bccd, + 0x3c04d, + 0x3f40a, + 0x40a0b, + 0x454cb, + 0x179785, + 0x194246d0, + 0xf5cf, + 0x112a4f, + 0x1754cd, + 0xbbd50, + 0x5642, + 0x19a26388, + 0x7a548, + 0x116a8e, + 0x19f61a05, + 0x4e2cb, + 0x13aad0, + 0x552c8, + 0x1f58a, + 0xaff49, + 0x65487, + 0x657c7, + 0x65987, + 0x65d07, + 0x66b07, + 0x67107, + 0x68347, + 0x68607, + 0x68ec7, + 0x691c7, + 0x69887, + 0x69a47, + 0x69c07, + 0x69dc7, + 0x6a0c7, + 0x6a487, + 0x6ad47, + 0x6b8c7, + 0x6be87, + 0x6c147, + 0x6c307, + 0x6c607, + 0x6c987, + 0x6cb87, + 0x6e147, + 0x6e307, + 0x6e4c7, + 0x6ed47, + 0x6f247, + 0x6f887, + 0x70547, + 0x70807, + 0x70d07, + 0x70ec7, + 0x712c7, + 0x71dc7, + 0x72287, + 0x72687, + 0x72847, + 0x72a07, + 0x73587, + 0x75d07, + 0x76247, + 0x76807, + 0x769c7, + 0x76d47, + 0x772c7, + 0xafc2, + 0x5e18a, + 0xe6487, + 0x18de45, + 0xabd51, + 0xe906, + 0x1147ca, + 0xa3a8a, + 0x59086, + 0xcd78b, + 0x642, + 0x2d811, + 0xc3e09, + 0x94b49, + 0x1742, + 0x7154a, + 0xa66c9, + 0xa6e0f, + 0xa740e, + 0xa80c8, + 0x55402, + 0x1b84c9, + 0x19cd4e, + 0x10470c, + 0xe814f, + 0x1b27ce, + 0x29b8c, + 0x11eb09, + 0x15a1d1, + 0x15a788, + 0x3a1d2, + 0x4378d, + 0x4738d, + 0x49e4b, + 0x5be15, + 0x6b109, + 0x6f60a, + 0x71f49, + 0x7aa50, + 0x80ecb, + 0x16820f, + 0x1c0b0b, + 0x9140c, + 0x99c10, + 0xa15ca, + 0xa3ecd, + 0xa5ace, + 0xaa2ca, + 0xac90c, + 0xb7cd4, + 0xc3a91, + 0x1bd4cb, + 0x1513cf, + 0x1a4dcd, + 0x11fcce, + 0xb518c, + 0xb658c, + 0xb79cb, + 0xb850e, + 0xbf490, + 0xbfe0b, + 0xc34cd, + 0xc448f, + 0xc500c, + 0xc5b8e, + 0x1165d1, + 0x164d4c, + 0xd1e87, + 0xd504d, + 0xdaf4c, + 0xdc8d0, + 0xe7b4d, + 0xf9d07, + 0xff650, + 0x105488, + 0x133dcb, + 0x16ec4f, + 0x1656c8, + 0x1149cd, + 0x199e10, + 0xfcfc9, + 0x1a2b1fc6, + 0xb2f03, + 0xb8945, + 0x6f42, + 0x132c09, + 0x73a0a, + 0x1a63ddc4, + 0x10dcc6, + 0x1ba0a, + 0x1a927c89, + 0x92203, + 0x14d10a, + 0xdd811, + 0xddc49, + 0xdf887, + 0xe0607, + 0xe6548, + 0x7c0b, + 0x12d0c9, + 0xe6cd0, + 0xe718c, + 0xe8608, + 0xe8c85, + 0xca008, + 0x1b9e4a, + 0x154247, + 0x12dd47, + 0x1c02, + 0x13b7ca, + 0x112d89, + 0x70bc5, + 0x5e64a, + 0x1cd44f, + 0x13f00b, + 0x16388c, + 0x67e92, + 0x9d585, + 0xea348, + 0xd634a, + 0x1aef3f45, + 0x16348c, + 0x138043, + 0x1a3282, + 0xfbd8a, + 0x14fc10c, + 0x111c48, + 0x3be88, + 0x13f287, + 0x8782, + 0x3c42, + 0x51590, + 0x78107, + 0x2ce4f, + 0xc8346, + 0xcece, + 0x15554b, + 0x49148, + 0x7c709, + 0x1920d2, + 0x10b3cd, + 0x10b908, + 0xaa4c9, + 0xd848d, + 0x150909, + 0x19b3cb, + 0x9c48, + 0x85d88, + 0x87588, + 0x8aac9, + 0x8acca, + 0x8f64c, + 0xf708a, + 0x10a987, + 0x15abcd, + 0xfeacb, + 0x12b70c, + 0x30488, + 0x47f89, + 0x1aa750, + 0x67c2, + 0x7ec4d, + 0x3642, + 0x1a642, + 0x10a8ca, + 0x1146ca, + 0x115c0b, + 0x4568c, + 0x11518a, + 0x11560e, + 0xa8cd, + 0x1b1d2ac5, + 0x12f108, + 0x1242, + 0x12b959ce, + 0x13201b4e, + 0x13b8e6ca, + 0x14328e8e, + 0x14b911ce, + 0x1536cbcc, + 0x142c347, + 0x142c349, + 0x140d983, + 0x15b3f70c, + 0x16205649, + 0x16a07409, + 0x17217ec9, + 0x1a82, + 0x195911, + 0x1a91, + 0x18e60d, + 0x128dd1, + 0x191111, + 0x16cb0f, + 0x13f64f, + 0x14c68c, + 0x734c, + 0x17e0c, + 0x5790d, + 0x757d5, + 0xd458c, + 0x15d88c, + 0x170c10, + 0x17bccc, + 0x1c4b8c, + 0x1c6359, + 0x1d0ad9, + 0x1d2499, + 0x5394, + 0xb5d4, + 0xc054, + 0xc5d4, + 0xd214, + 0x17a0b889, + 0x1800c309, + 0x18b5d949, + 0x12f5a989, + 0x1a82, + 0x1375a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x13f5a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x1475a989, + 0x1a82, + 0x14f5a989, + 0x1a82, + 0x1575a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x15f5a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x1675a989, + 0x1a82, + 0x16f5a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x1775a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x17f5a989, + 0x1a82, + 0x1875a989, + 0x1a82, + 0x18f5a989, + 0x1a82, + 0x538a, + 0x1a82, + 0x7a85, + 0x180e84, + 0x1959ce, + 0x1b4e, + 0x1bdce, + 0x18e6ca, + 0x128e8e, + 0x1911ce, + 0x16cbcc, + 0x13f70c, + 0x5649, + 0x7409, + 0x17ec9, + 0xb889, + 0xc309, + 0x15d949, + 0x759cd, + 0xc889, + 0xd4c9, + 0x94644, + 0xf9284, + 0x12f2c4, + 0x131fc4, + 0x7cfc4, + 0x12e984, + 0x42f84, + 0x4e044, + 0xfb544, + 0x1596a83, + 0xa203, + 0x5642, + 0xa8c3, + 0xe402, + 0xe408, + 0x12d147, + 0x8a42, + 0x2000c2, + 0x208e02, + 0x202142, + 0x217242, + 0x200382, + 0x2003c2, + 0x203c42, + 0x208e03, + 0x231103, + 0x213ec3, + 0x2171c3, + 0x21d283, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x21d283, + 0x259003, + 0xec43, + 0x213ec3, + 0x20a04, + 0x2000c2, + 0x38d8c3, + 0x1d608e03, + 0x23d907, + 0x213ec3, + 0x221303, + 0x2bf144, + 0x21d283, + 0x259003, + 0x23628a, + 0x3adc45, + 0x216203, + 0x2022c2, + 0x120648, + 0x120648, + 0x8e02, + 0x134882, + 0x1df27a0b, + 0x1e21a184, + 0x3dd05, + 0x7d85, + 0xfc646, + 0x1e607d85, + 0x548c3, + 0x19a1c3, + 0x192544, + 0x191503, + 0x192e05, + 0xf8945, + 0x120648, + 0x18dc7, + 0x8e03, + 0x1ee394c7, + 0x18bc46, + 0x1f18e505, + 0x18bd07, + 0x1fd8a, + 0x1e208, + 0x1fc87, + 0x7d988, + 0xdc487, + 0xfad8f, + 0x464c7, + 0x4de46, + 0x13aad0, + 0x138b0f, + 0x6d409, + 0x10dd44, + 0x1f58bdce, + 0x6d94c, + 0x56c8a, + 0x7c4c7, + 0xe590a, + 0x190b09, + 0x1a3f8c, + 0xc254a, + 0x5988a, + 0x192e49, + 0x10dcc6, + 0x7c58a, + 0x10bf0a, + 0x9a70a, + 0x14f909, + 0xdd148, + 0xdd3c6, + 0xe34cd, + 0xb8dc5, + 0x1fb76e0c, + 0x18307, + 0x102f09, + 0x1228c7, + 0xb1794, + 0x10588b, + 0x8850a, + 0x191f4a, + 0xa428d, + 0x1515e09, + 0x10b18c, + 0x10b70b, + 0x35d43, + 0x35d43, + 0x35d46, + 0x35d43, + 0xfc648, + 0xbb1c9, + 0x18d8c3, + 0x120648, + 0x8e02, + 0x4ea84, + 0x5a483, + 0xfc845, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x2038c3, + 0x208e03, + 0x231103, + 0x3d0943, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x294d03, + 0x202043, + 0x2038c3, + 0x2879c4, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x22f583, + 0x208e03, + 0x231103, + 0x217243, + 0x3d0943, + 0x213ec3, + 0x220a04, + 0x307c83, + 0x208e83, + 0x270203, + 0x21d283, + 0x259003, + 0x216203, + 0x209283, + 0x21a08e03, + 0x231103, + 0x24a783, + 0x213ec3, + 0x2230c3, + 0x208e83, + 0x259003, + 0x203b03, + 0x35c4c4, + 0x120648, + 0x22208e03, + 0x231103, + 0x2a8183, + 0x213ec3, + 0x270203, + 0x2bf144, + 0x21d283, + 0x259003, + 0x214d03, + 0x120648, + 0x22a08e03, + 0x231103, + 0x3d0943, + 0x214b83, + 0x259003, + 0x120648, + 0x142c347, + 0x38d8c3, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x2bf144, + 0x21d283, + 0x259003, + 0xf8945, + 0xeb207, + 0xb19cb, + 0xde044, + 0xb8dc5, + 0x1582ac8, + 0xa87cd, + 0x23e4e545, + 0x9fec4, + 0x11243, + 0xfcec5, + 0x36b845, + 0x120648, + 0x1b002, + 0x40c03, + 0xf9646, + 0x32be08, + 0x3a6e47, + 0x2879c4, + 0x33ed46, + 0x34c546, + 0x120648, + 0x31cd83, + 0x312449, + 0x347a95, + 0x147a9f, + 0x208e03, + 0x2d4152, + 0x168dc6, + 0x1776c5, + 0x1f58a, + 0xaff49, + 0x2d3f0f, + 0x2dd584, + 0x238f45, + 0x305fd0, + 0x381607, + 0x214b83, + 0x23a708, + 0x15eb46, + 0x29ed0a, + 0x202584, + 0x2f3983, + 0x3adc46, + 0x2022c2, + 0x23ef4b, + 0x14b83, + 0x1842c4, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x2fa443, + 0x208e02, + 0xf0183, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x221303, + 0x235843, + 0x259003, + 0x208e02, + 0x208e03, + 0x231103, + 0x21d283, + 0x14b83, + 0x259003, + 0x2000c2, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x7d85, + 0x2879c4, + 0x208e03, + 0x231103, + 0x20ac44, + 0x21d283, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0xe6143, + 0x259003, + 0x208e03, + 0x231103, + 0x3d0943, + 0x203583, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x38d604, + 0x220a04, + 0x21d283, + 0x259003, + 0x202043, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0xe6143, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x2bd9c3, + 0x69f03, + 0x21303, + 0x21d283, + 0x259003, + 0x31de8a, + 0x340ac9, + 0x357f8b, + 0x3586ca, + 0x3612ca, + 0x36fc4b, + 0x383a0a, + 0x3977ca, + 0x39d78a, + 0x39da0b, + 0x3b9b89, + 0x3bfe4a, + 0x3c028b, + 0x3cb70b, + 0x3d16ca, + 0x208e03, + 0x231103, + 0x3d0943, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x1c5b0b, + 0x5f0c8, + 0xd85c4, + 0x7d46, + 0x3f6c9, + 0x120648, + 0x208e03, + 0x265484, + 0x221602, + 0x2bf144, + 0x393345, + 0x2038c3, + 0x2879c4, + 0x208e03, + 0x233e84, + 0x231103, + 0x24ea84, + 0x2dd584, + 0x220a04, + 0x208e83, + 0x21d283, + 0x259003, + 0x27ee05, + 0x22f583, + 0x216203, + 0x294003, + 0x26fdc4, + 0x201e84, + 0x2bb485, + 0x120648, + 0x326cc4, + 0x32e046, + 0x279b44, + 0x208e02, + 0x248947, + 0x250fc7, + 0x24ba04, + 0x25afc5, + 0x2e7d45, + 0x22bec5, + 0x220a04, + 0x380048, + 0x233506, + 0x31be48, + 0x27c085, + 0x2e4345, + 0x397244, + 0x259003, + 0x2f5544, + 0x36e886, + 0x3add43, + 0x26fdc4, + 0x24ac05, + 0x332b84, + 0x24f244, + 0x2022c2, + 0x230206, + 0x3afbc6, + 0x308e05, + 0x2000c2, + 0x38d8c3, + 0x2ae08e02, + 0x226504, + 0x200382, + 0x270203, 0x207c02, - 0x200742, - 0x6a202c42, - 0x6a70e843, - 0x20f003, - 0x201e42, - 0x20ec83, - 0x338c03, - 0x241d03, - 0x2e2d83, - 0x379586, - 0x1607c03, - 0xcd588, - 0x6e247, - 0x14a345, - 0xa7e0d, - 0xa5f4a, - 0x85047, - 0x6ae00a42, - 0x6b200602, - 0x6b600282, - 0x6ba02b82, - 0x6be12442, - 0x6c203cc2, - 0x15da87, - 0x6c602c42, - 0x6ca1b282, - 0x6ce1f9c2, - 0x6d202e02, - 0x219483, - 0x22644, - 0x282dc3, - 0x6d615902, - 0x6da039c2, - 0x55087, - 0x6de02202, - 0x6e200902, - 0x6e600542, - 0x6ea07d02, - 0x6ee03882, - 0x6f201342, - 0xc0f85, - 0x24c4c3, - 0x23a2c4, - 0x6f6028c2, - 0x6fa0dd82, - 0x6fe00682, - 0xb714b, - 0x702000c2, - 0x70a54ac2, - 0x70e01e42, - 0x71200b82, - 0x71603282, - 0x71a05a02, - 0x71e0d682, - 0x7226cd82, - 0x72602f02, - 0x72a04d42, - 0x72e032c2, - 0x7323e0c2, - 0x7362a402, - 0x73a11e82, - 0xafd44, - 0x339b43, - 0x73e0e882, - 0x742190c2, - 0x74606482, - 0x74a02882, - 0x74e00342, - 0x75206f02, - 0x7e3c7, - 0x756057c2, - 0x75a00502, - 0x75e07c02, - 0x76209f82, - 0xf778c, - 0x76627882, - 0x76a2c0c2, - 0x76e0a902, - 0x77206d02, - 0x77611d82, - 0x77a3e602, - 0x77e0fc42, - 0x78213802, - 0x78674fc2, - 0x78a4f1c2, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x11343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x707c32c3, - 0x211343, - 0x339dc4, - 0x3a2f86, - 0x2f0ec3, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x200189, - 0x24d5c2, - 0x391283, - 0x2b8503, - 0x202f85, - 0x206c03, - 0x3c32c3, - 0x211343, - 0x29ea43, - 0x233d43, - 0x3bd849, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x24d5c2, - 0x24d5c2, - 0x3c32c3, - 0x211343, - 0x7920be03, - 0x237583, - 0x332683, - 0x21f743, - 0x20ec83, - 0xae43, - 0x241d03, - 0xcd588, - 0x202c42, - 0x20be03, - 0x20ec83, - 0x241d03, - 0x20be03, - 0x237583, - 0x30e843, - 0x21f743, - 0x20ec83, - 0xae43, - 0x241d03, - 0x253404, - 0x202c42, - 0x20be03, - 0x322183, - 0x237583, - 0x254a04, - 0x203d43, - 0x30e843, - 0x26ff84, - 0x214bc3, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x24f3c3, - 0x2f5805, - 0x233d43, - 0x203f83, - 0xae43, - 0x202c42, - 0x20be03, - 0x3c32c3, - 0x20ec83, - 0x241d03, - 0x200742, - 0x2b6c03, - 0xcd588, - 0x20be03, - 0x237583, - 0x30e843, - 0x234106, - 0x26ff84, - 0x214bc3, - 0x226444, - 0x20ec83, - 0x241d03, - 0x219543, - 0x20be03, - 0x237583, - 0x20ec83, - 0x241d03, - 0x144be47, - 0x20be03, - 0x29546, - 0x237583, - 0x30e843, - 0xd91c6, - 0x20ec83, - 0x241d03, - 0x318fc8, + 0x21d283, + 0x2003c2, + 0x211303, + 0x202043, + 0xa8a04, + 0x120648, + 0x120648, + 0x213ec3, + 0xe6143, + 0x2000c2, + 0x2ba08e02, + 0x213ec3, + 0x269b83, + 0x307c83, + 0x21a184, + 0x21d283, + 0x259003, + 0x120648, + 0x2000c2, + 0x2c208e02, + 0x208e03, + 0x21d283, + 0x14b83, + 0x259003, + 0x682, + 0x202442, + 0x22a342, + 0x221303, + 0x2ec903, + 0x2000c2, + 0xf8945, + 0x120648, + 0xeb207, + 0x208e02, + 0x231103, + 0x24ea84, + 0x204a03, + 0x213ec3, + 0x203583, + 0x270203, + 0x21d283, + 0x213443, + 0x259003, + 0x214c83, + 0x93f53, + 0xd3a94, + 0xf8945, + 0xeb207, + 0x103806, + 0x73c0b, + 0x35d46, + 0x57e07, + 0x5afc6, + 0x649, + 0xb330a, + 0x89b0d, + 0xbe1cc, + 0x10c88a, + 0xf3088, + 0x129cc5, + 0x1fdc8, + 0xc8346, + 0x6fa06, + 0x3f5c6, + 0x205642, + 0x3184, + 0x823ce, + 0x5668c, + 0xf8945, + 0x181047, + 0x20dd1, + 0xfb3ca, + 0x208e03, + 0x7d905, + 0x179b48, + 0x26744, + 0x2d4240c6, + 0xabd46, + 0xd9506, + 0x8edca, + 0x18f283, + 0x2da44684, + 0x605, + 0xecc43, + 0x2de32587, + 0x1caa45, + 0xcd84c, + 0xf7f48, + 0x9d1cb, + 0x2e24aecc, + 0x14111c3, + 0xb9948, + 0x9e909, + 0x4f408, + 0x141b946, + 0x2e774c09, + 0x130d47, + 0xeb0ca, + 0xe1c8, + 0xfc648, + 0xfb544, + 0x15d405, + 0x9d307, + 0x2ea9d303, + 0x2ef9e406, + 0x2f2f5dc4, + 0x2f6fbf47, + 0xfc644, + 0xfc644, + 0xfc644, + 0xfc644, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x2000c2, + 0x208e02, + 0x213ec3, + 0x200d42, + 0x21d283, + 0x259003, + 0x211303, + 0x377f4f, + 0x37830e, + 0x120648, + 0x208e03, + 0x46007, + 0x231103, + 0x213ec3, + 0x217383, + 0x21d283, + 0x259003, + 0x18e904, + 0x191644, + 0x194cc4, + 0x21aec3, + 0x393f47, + 0x200a82, + 0x2c96c9, + 0x204342, + 0x25188b, + 0x2a1bca, + 0x2ac389, + 0x200542, + 0x326746, + 0x23ac15, + 0x2519d5, + 0x23fa13, + 0x251f53, + 0x21cec2, + 0x21cec5, + 0x3253cc, + 0x275f8b, + 0x3c2645, + 0x201b42, + 0x323bc2, + 0x386246, + 0x202842, + 0x260bc6, + 0x21f80d, + 0x3ca7cc, + 0x224744, + 0x200882, + 0x205042, + 0x23a588, + 0x200202, + 0x2231c6, + 0x33394f, + 0x2231d0, + 0x2f0c04, + 0x23add5, + 0x23fb93, + 0x20d8c3, + 0x32408a, + 0x214a47, + 0x34be89, + 0x2e5687, + 0x320f02, + 0x200282, + 0x3b4b46, + 0x207f42, + 0x120648, + 0x20d702, + 0x20a842, + 0x225387, + 0x331bc7, + 0x331bd1, + 0x2184c5, + 0x339b4e, + 0x2184cf, + 0x2020c2, + 0x214c47, + 0x21b508, + 0x2054c2, + 0x2c25c2, + 0x325a46, + 0x33cb0f, + 0x325a50, + 0x229fc2, + 0x200f82, + 0x236cc8, + 0x208f83, + 0x25a648, + 0x208f8d, + 0x231783, + 0x3131c8, + 0x23178f, + 0x231b4e, + 0x391c0a, + 0x2ed251, + 0x2ed6d0, + 0x2de80d, + 0x2deb4c, + 0x200f87, + 0x324207, + 0x33ee09, + 0x224842, + 0x203982, + 0x33fb0c, + 0x34000b, + 0x206a82, + 0x2b7b86, + 0x214d42, + 0x200482, + 0x342382, + 0x208e02, + 0x22b904, + 0x2380c7, + 0x22a502, + 0x241007, + 0x242d87, + 0x22d182, + 0x202282, + 0x245ec5, + 0x218982, + 0x37ae0e, + 0x2a454d, + 0x231103, + 0x28684e, + 0x3bcf4d, + 0x3257c3, + 0x2017c2, + 0x2852c4, + 0x2327c2, + 0x20ce02, + 0x3a0845, + 0x3a2507, + 0x249b02, + 0x217242, + 0x24e687, + 0x252f08, + 0x22ec42, + 0x29d606, + 0x34e50c, + 0x34ea0b, + 0x201482, + 0x26178f, + 0x261b50, + 0x261f4f, + 0x262315, + 0x262854, + 0x262d4e, + 0x2630ce, + 0x26344f, + 0x26380e, + 0x263b94, + 0x264093, + 0x26454d, + 0x277489, + 0x28b5c3, + 0x2028c2, + 0x219805, + 0x204a06, + 0x200382, + 0x373947, + 0x213ec3, + 0x200642, + 0x22f948, + 0x2ed491, + 0x2ed8d0, + 0x202ec2, + 0x281e87, + 0x200b02, + 0x2044c7, + 0x206f42, + 0x212089, + 0x386207, + 0x2a3008, + 0x223f06, + 0x2ec803, + 0x325e05, + 0x231382, + 0x2004c2, + 0x3b4f45, + 0x256405, + 0x204382, + 0x21db43, + 0x38c207, + 0x220c07, + 0x205c82, + 0x333084, + 0x221403, 0x31bfc9, - 0x330349, - 0x33af08, - 0x38a348, - 0x38a349, - 0x22c10d, - 0x24b00f, - 0x2ec510, - 0x354d0d, - 0x36ebcc, - 0x38bc4b, - 0xaf148, - 0xc7bc5, - 0x200742, - 0x23b005, - 0x20bfc3, - 0x7c602c42, - 0x237583, - 0x30e843, - 0x2d7ac7, - 0x20fbc3, - 0x21f743, - 0x20ec83, - 0x228803, - 0x20c0c3, - 0x20ae43, - 0x241d03, - 0x253786, - 0x20e982, - 0x203f83, - 0xcd588, - 0x200742, - 0x2b6c03, - 0x202c42, - 0x20be03, - 0x237583, - 0x30e843, - 0x26ff84, - 0x21f743, - 0x20ec83, - 0x241d03, - 0x207c03, - 0xf84, - 0x154ab06, - 0x200742, - 0x202c42, - 0x30e843, - 0x21f743, - 0x241d03, + 0x2fac08, + 0x204c82, + 0x20bb02, + 0x375987, + 0x396f85, + 0x2bc5c8, + 0x322d07, + 0x202d83, + 0x28dc06, + 0x2de68d, + 0x2dea0c, + 0x2ffe46, + 0x207042, + 0x29d082, + 0x202302, + 0x23160f, + 0x231a0e, + 0x2e7dc7, + 0x200d02, + 0x357285, + 0x357286, + 0x226942, + 0x201942, + 0x28c606, + 0x204403, + 0x204406, + 0x2cc0c5, + 0x2cc0cd, + 0x2cc695, + 0x2cd24c, + 0x2cdb4d, + 0x2cdf12, + 0x202382, + 0x26cac2, + 0x200a42, + 0x222d06, + 0x3045c6, + 0x201c02, + 0x204a86, + 0x20b142, + 0x20b145, + 0x20a002, + 0x2a4649, + 0x22788c, + 0x227bcb, + 0x2003c2, + 0x253ac8, + 0x201982, + 0x20a282, + 0x271006, + 0x35a905, + 0x327e87, + 0x2ecec5, + 0x28ddc5, + 0x207d42, + 0x2042c2, + 0x209302, + 0x2e8447, + 0x31908d, + 0x31940c, + 0x280a87, + 0x225f42, + 0x203042, + 0x233808, + 0x332d88, + 0x2e4bc8, + 0x314984, + 0x2b8787, + 0x23de43, + 0x24df42, + 0x202582, + 0x2f1cc9, + 0x2fe3c7, + 0x203b02, + 0x271405, + 0x23eec2, + 0x22c402, + 0x2c0f43, + 0x2c0f46, + 0x2fa002, + 0x2fb802, + 0x200402, + 0x3c1786, + 0x2af507, + 0x204182, + 0x200902, + 0x25a48f, + 0x28668d, + 0x39d14e, + 0x3bcdcc, + 0x202882, + 0x202e02, + 0x223d45, + 0x31b706, + 0x213242, + 0x204582, + 0x200682, + 0x286a04, + 0x2d2844, + 0x388806, + 0x203c42, + 0x2b0907, + 0x2432c3, + 0x2432c8, + 0x243d88, + 0x249cc7, + 0x254346, + 0x205502, + 0x224403, + 0x224407, + 0x292d06, + 0x2f4605, + 0x314d08, + 0x200b42, + 0x326507, + 0x212142, + 0x374402, + 0x209142, + 0x218649, + 0x20fdc2, + 0x2010c2, + 0x252383, + 0x38b547, + 0x203a42, + 0x227a0c, + 0x227d0b, + 0x2ffec6, + 0x2ecb85, + 0x20fec2, + 0x2019c2, + 0x2bea46, + 0x267343, + 0x307887, + 0x288102, + 0x2008c2, + 0x23aa95, + 0x251b95, + 0x23f8d3, + 0x2520d3, + 0x24bdc7, + 0x3bad11, + 0x3bb450, + 0x266c92, + 0x2763d1, + 0x27fcc8, + 0x27fcd0, + 0x28e68f, + 0x2a1993, + 0x2ac152, + 0x2ad890, + 0x32080f, + 0x3bba52, + 0x3bc491, + 0x334053, + 0x3b9412, + 0x2b210f, + 0x2c1ace, + 0x2c8e52, + 0x2cbc91, + 0x2d71cf, + 0x2da28e, + 0x3bd891, + 0x3be050, + 0x2db912, + 0x2e01d1, + 0x3be650, + 0x3bec4f, + 0x2e0f11, + 0x2e3cd0, + 0x2e98c6, + 0x2f4ec7, + 0x209607, + 0x200c42, + 0x282d05, + 0x377d87, + 0x22a342, + 0x207e42, + 0x226f85, + 0x21c3c3, + 0x3b83c6, + 0x31924d, + 0x31958c, + 0x204b02, + 0x32524b, + 0x275e4a, + 0x21cd8a, + 0x2ba889, + 0x2ef6cb, + 0x322e4d, + 0x30644c, + 0x272f0a, + 0x276ecc, + 0x294fcb, + 0x3c248c, + 0x3c298e, + 0x3c318b, + 0x3c364c, + 0x32aa43, + 0x350806, + 0x3075c2, + 0x2fc542, + 0x211643, + 0x202b82, + 0x205283, + 0x318846, + 0x2624c7, + 0x2aec06, + 0x2f1ac8, + 0x38c088, + 0x310146, + 0x205ec2, + 0x3087cd, + 0x308b0c, + 0x2dd647, + 0x30c6c7, + 0x232f42, + 0x216402, + 0x20fcc2, + 0x2532c2, + 0x333857, + 0x339a56, + 0x33ca17, + 0x33fa14, + 0x33ff13, + 0x34e414, + 0x34e913, + 0x3b7c10, + 0x3bac19, + 0x3bb358, + 0x3bb95a, + 0x3bc399, + 0x3bd799, + 0x3bdf58, + 0x3be558, + 0x3beb57, + 0x3c2394, + 0x3c2896, + 0x3c3093, + 0x3c3554, + 0x208e02, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x2bf144, + 0x21d283, + 0x259003, + 0x211303, + 0x2000c2, + 0x20cd42, + 0x31691185, + 0x31a88ec5, + 0x31fc1446, + 0x120648, + 0x322b2685, + 0x208e02, + 0x202142, + 0x3278df45, + 0x32a80dc5, + 0x32e82047, + 0x33283989, + 0x33753804, + 0x200382, + 0x200642, + 0x33a5d145, + 0x33e96749, + 0x34332508, + 0x346adec5, + 0x34b3a007, + 0x34e78908, + 0x352ea205, + 0x35634e06, + 0x35b74e49, + 0x35ed1448, + 0x362c42c8, + 0x36696d8a, + 0x36a77cc4, + 0x36ea4f45, + 0x372c0808, + 0x37732985, + 0x213542, + 0x37a47c43, + 0x37ea3406, + 0x3835a008, + 0x38710b06, + 0x38b62588, + 0x38f4c406, + 0x39275284, + 0x201402, + 0x39784c47, + 0x39aa9204, + 0x39e7bd87, + 0x3a237007, + 0x2003c2, + 0x3a69b3c5, + 0x3aa4f884, + 0x3aee7807, + 0x3b240607, + 0x3b685106, + 0x3ba81445, + 0x3be96847, + 0x3c2e8908, + 0x3c619b87, + 0x3cab7109, + 0x3cecd585, + 0x3d351247, + 0x3d6907c6, + 0x3dac9d08, + 0x21ea4d, + 0x27e409, + 0x28a08b, + 0x2a68cb, + 0x2b45cb, + 0x30cd4b, + 0x31b90b, + 0x31bbcb, + 0x31c749, + 0x31e10b, + 0x31e3cb, + 0x31f94b, + 0x32128a, + 0x3217ca, + 0x321dcc, + 0x32a48b, + 0x32ab0a, + 0x33d80a, + 0x3448ce, + 0x3454ce, + 0x34584a, + 0x34858a, + 0x34924b, + 0x34950b, + 0x34a18b, + 0x365d0b, + 0x36630a, + 0x366fcb, + 0x36728a, + 0x36750a, + 0x36778a, + 0x3854cb, + 0x39868b, + 0x39a9ce, + 0x39ad4b, + 0x3a26cb, + 0x3a360b, + 0x3a808a, + 0x3a8309, + 0x3a854a, + 0x3aa04a, + 0x3ba60b, + 0x3c054b, + 0x3c0dca, + 0x3c1dcb, + 0x3c7c4b, + 0x3d110b, + 0x3de833c8, + 0x3e289289, + 0x3e69e789, + 0x3eae4708, + 0x34fe85, + 0x20a343, + 0x31a304, + 0x395005, + 0x353546, + 0x234b85, + 0x288944, + 0x373848, + 0x314505, + 0x292404, + 0x2137c7, + 0x29dd0a, + 0x248c0a, + 0x2e7ec7, + 0x21a5c7, + 0x305307, + 0x27dfc7, + 0x3003c5, + 0x205f46, + 0x2b5007, + 0x240c44, + 0x2eeac6, + 0x2ee9c6, + 0x210885, + 0x39e044, + 0x297dc6, + 0x29c407, + 0x31ae06, + 0x384847, + 0x291d03, + 0x3a2ec6, + 0x22b1c5, + 0x282147, + 0x26a64a, + 0x22fa44, + 0x26d188, + 0x2b5e89, + 0x2e4f07, + 0x3460c6, + 0x253d48, + 0x311249, + 0x34c044, + 0x2d61c4, + 0x2a0885, + 0x2bf048, + 0x2ca287, + 0x350f89, + 0x236988, + 0x2e99c6, + 0x3224c6, + 0x298788, + 0x369b06, + 0x288ec5, + 0x2851c6, + 0x27c948, + 0x231506, + 0x25894b, + 0x300f06, + 0x29a00d, + 0x2082c5, + 0x2a90c6, + 0x239105, + 0x255b49, + 0x3a4887, + 0x35cbc8, + 0x365886, + 0x299009, + 0x3878c6, + 0x26a5c5, + 0x24c446, + 0x2c7846, + 0x2cefc9, + 0x38c5c6, + 0x29da07, + 0x241c85, + 0x20a003, + 0x258ac5, + 0x29a2c7, + 0x325686, + 0x2081c9, + 0x3c1446, + 0x26af46, + 0x212e89, + 0x284bc9, + 0x2a1447, + 0x3927c8, + 0x2abb89, + 0x282988, + 0x397a06, + 0x2dcf05, + 0x23870a, + 0x26afc6, + 0x23d786, + 0x2d4a45, + 0x2d0e08, + 0x396b07, + 0x22d60a, + 0x24fb86, + 0x27e845, + 0x38d446, + 0x335687, + 0x345f87, + 0x3651c5, + 0x26a785, + 0x280686, + 0x2a7746, + 0x3a11c6, + 0x2c0cc4, + 0x283f49, + 0x28a346, + 0x2fa1ca, + 0x31a6c8, + 0x34b788, + 0x248c0a, + 0x2224c5, + 0x29c345, + 0x2af6c8, + 0x2b9e88, + 0x232907, + 0x2bf346, + 0x337b88, + 0x2ace47, + 0x282288, + 0x2b83c6, + 0x285948, + 0x295d86, + 0x27c207, + 0x2d5f46, + 0x297dc6, + 0x31d50a, + 0x22b986, + 0x2dcf09, + 0x310246, + 0x2f050a, + 0x275289, + 0x25ba86, + 0x2b93c4, + 0x2198cd, + 0x289507, + 0x2bc346, + 0x2c4185, + 0x387945, + 0x386e46, + 0x2e7649, + 0x2d0987, + 0x27d3c6, + 0x2e2186, + 0x2889c9, + 0x288e04, + 0x249944, + 0x327288, + 0x318c06, + 0x270a08, + 0x2f2dc8, + 0x2a0dc7, + 0x3b6a89, + 0x3a13c7, + 0x2b254a, + 0x2f278f, + 0x3a1b8a, + 0x223b45, + 0x27cb85, + 0x354885, + 0x2f0b47, + 0x2062c3, + 0x3929c8, + 0x2287c6, + 0x2288c9, + 0x302a06, + 0x2d0307, + 0x298dc9, + 0x35cac8, + 0x2d4b07, + 0x316fc3, + 0x34ff05, + 0x3351c5, + 0x2c0b0b, + 0x332a44, + 0x23cb44, + 0x279406, + 0x317187, + 0x39fdca, + 0x247cc7, + 0x209347, + 0x280dc5, + 0x3cc5c5, + 0x271a09, + 0x297dc6, + 0x247b4d, + 0x38c805, + 0x2b5bc3, + 0x215343, + 0x3afe05, + 0x356dc5, + 0x253d48, + 0x27dc87, + 0x2496c6, + 0x29e406, + 0x229185, + 0x2313c7, + 0x274b87, + 0x2333c7, + 0x2a4fca, + 0x3a2f88, + 0x2c0cc4, + 0x27f1c7, + 0x2800c7, + 0x349786, + 0x295407, + 0x2e27c8, + 0x2b34c8, + 0x24cc86, + 0x21a808, + 0x2dab04, + 0x2b5006, + 0x252a06, + 0x3907c6, + 0x239246, + 0x29b784, + 0x27e086, + 0x2c27c6, + 0x298186, + 0x22ec86, + 0x215206, + 0x2e2606, + 0x2495c8, + 0x2b4a88, + 0x2d8908, + 0x234d88, + 0x2af646, + 0x21a3c5, + 0x38ac46, + 0x2adf45, + 0x3a6b87, + 0x236a45, + 0x206443, + 0x203305, + 0x3c8404, + 0x215345, + 0x201983, + 0x34a747, + 0x364b88, + 0x384906, + 0x2b9b0d, + 0x27cb46, + 0x297745, + 0x218643, + 0x2c01c9, + 0x288f86, + 0x293586, + 0x2714c4, + 0x3a1b07, + 0x311746, + 0x2d0c45, + 0x200cc3, + 0x209b04, + 0x280286, + 0x252b04, + 0x2d9dc8, + 0x206789, + 0x3cfd49, + 0x2a068a, + 0x2a244d, + 0x22fdc7, + 0x23d606, + 0x21fec4, + 0x283989, + 0x287008, + 0x289106, + 0x239986, + 0x295407, + 0x2d9286, + 0x2b1346, + 0x36c106, + 0x23708a, + 0x278908, + 0x31ea05, + 0x25e949, + 0x2caa0a, + 0x2fdac8, + 0x29bc08, + 0x293508, + 0x29e04c, + 0x31fbc5, + 0x29e688, + 0x2b4d86, + 0x24c986, + 0x3cde07, + 0x247bc5, + 0x285345, + 0x3cfc09, + 0x21c987, + 0x228885, + 0x2d3647, + 0x215343, + 0x2caec5, + 0x225ac8, + 0x284847, + 0x29bac9, + 0x2ec3c5, + 0x3447c4, + 0x2a2108, + 0x2d2c47, + 0x2d4cc8, + 0x39efc8, + 0x2aa1c5, + 0x2286c6, + 0x29e506, + 0x3a4bc9, + 0x330bc7, + 0x2ae386, + 0x226047, + 0x20a903, + 0x353804, + 0x2dac05, + 0x256a04, + 0x24b184, + 0x283647, + 0x269307, + 0x27d584, + 0x29b910, + 0x38ae47, + 0x3cc5c5, + 0x25110c, + 0x2117c4, + 0x2b90c8, + 0x27c109, + 0x37c646, + 0x24f548, + 0x2190c4, + 0x279708, + 0x22dc06, + 0x31d388, + 0x29a586, + 0x286b4b, + 0x32c585, + 0x2daa88, + 0x206bc4, + 0x206bca, + 0x29bac9, + 0x2d5e46, + 0x354948, + 0x2a3dc5, + 0x206404, + 0x2b8fc6, + 0x233288, + 0x2833c8, + 0x338406, + 0x388784, + 0x238686, + 0x3a1447, + 0x27bc87, + 0x29540f, + 0x203b87, + 0x25bb47, + 0x357145, + 0x364345, + 0x2a1109, + 0x2c8086, + 0x281585, + 0x284ec7, + 0x2c8a48, + 0x2d5305, + 0x2d5f46, + 0x31a508, + 0x310b0a, + 0x20ff08, + 0x28c387, + 0x2f2bc6, + 0x25e906, + 0x2003c3, + 0x213383, + 0x2cabc9, + 0x2aba09, + 0x2b7006, + 0x2ec3c5, + 0x206108, + 0x354948, + 0x369c88, + 0x36c18b, + 0x2b9d47, + 0x30f849, + 0x295688, + 0x355d44, + 0x35cf08, + 0x28e149, + 0x2ae685, + 0x2f0a47, + 0x353885, + 0x2832c8, + 0x29100b, + 0x296590, + 0x2a8b05, + 0x21280c, + 0x249885, + 0x280e43, + 0x2eb506, + 0x2c1e44, + 0x24f986, + 0x29c407, + 0x20ff04, + 0x2440c8, + 0x39288d, + 0x316285, + 0x22fe04, + 0x223884, + 0x29ff09, + 0x2b2a48, + 0x32ca47, + 0x22dc88, + 0x284008, + 0x27d6c5, + 0x27a147, + 0x27d647, + 0x312207, + 0x26a789, + 0x3289c9, + 0x26ea46, + 0x2ded46, + 0x284f86, + 0x344d05, + 0x39dc84, + 0x3c5446, + 0x3c9dc6, + 0x27d708, + 0x33534b, + 0x267287, + 0x21fec4, + 0x311686, + 0x2e2b07, + 0x324545, + 0x381b85, + 0x239bc4, + 0x328946, + 0x3c54c8, + 0x283989, + 0x261286, + 0x286e08, + 0x2d0d06, + 0x3563c8, + 0x2ccdcc, + 0x27d586, + 0x29740d, + 0x29788b, + 0x29dac5, + 0x274cc7, + 0x38c6c6, + 0x345e48, + 0x26eac9, + 0x24cf48, + 0x3cc5c5, + 0x379c87, + 0x282a88, + 0x319b49, + 0x3a78c6, + 0x26118a, + 0x345bc8, + 0x24cd8b, + 0x2d7dcc, + 0x279808, + 0x27f786, + 0x228148, + 0x310787, + 0x203cc9, + 0x33a14d, + 0x297cc6, + 0x206288, + 0x2b4949, + 0x2c0dc8, + 0x285a48, + 0x2c320c, + 0x2c4b87, + 0x2c5647, + 0x26a5c5, + 0x2b7487, + 0x2c8908, + 0x2b9046, + 0x37a38c, + 0x2f73c8, + 0x2d1648, + 0x319e46, + 0x334f47, + 0x26ec44, + 0x234d88, + 0x2b030c, + 0x38d9cc, + 0x223bc5, + 0x210907, + 0x388706, + 0x334ec6, + 0x255d08, + 0x36eb84, + 0x31ae0b, + 0x2b0a4b, + 0x2f2bc6, + 0x392707, + 0x389b05, + 0x270945, + 0x31af46, + 0x2a3d85, + 0x332a05, + 0x2cee07, + 0x281909, + 0x2a7904, + 0x25d645, + 0x30de05, + 0x2d9b48, + 0x2e5d85, + 0x2baf49, + 0x2df507, + 0x2df50b, + 0x319786, + 0x249309, + 0x39df88, + 0x299b05, + 0x312308, + 0x328a08, + 0x26b447, + 0x37a187, + 0x2836c9, + 0x2b49c7, + 0x2a6489, + 0x31c30c, + 0x2b7008, + 0x2d1289, + 0x2d1d07, + 0x2840c9, + 0x200b47, + 0x2d7ec8, + 0x3b6c45, + 0x2b4f86, + 0x2c41c8, + 0x314f88, + 0x2ca8c9, + 0x332a47, + 0x252805, + 0x35afc9, + 0x202d06, + 0x2907c4, + 0x38b686, + 0x359e88, + 0x3ae447, + 0x335548, + 0x21a8c9, + 0x3896c7, + 0x29dec6, + 0x274d84, + 0x203389, + 0x279fc8, + 0x319d07, + 0x372c06, + 0x335286, + 0x23d704, + 0x352786, + 0x2152c3, + 0x32c109, + 0x32c546, + 0x21ab45, + 0x29e406, + 0x2cf385, + 0x282f08, + 0x310647, + 0x39c4c6, + 0x38df86, + 0x34b788, + 0x2a1287, + 0x297d05, + 0x29b708, + 0x3c0948, + 0x345bc8, + 0x249745, + 0x2b5006, + 0x3cfb09, + 0x3a4a44, + 0x2cf20b, + 0x2b104b, + 0x31e909, + 0x215343, + 0x259bc5, + 0x385106, + 0x368f08, + 0x3475c4, + 0x384906, + 0x2a5109, + 0x2e0b45, + 0x2ced46, + 0x2d2c46, + 0x206104, + 0x2a928a, + 0x21aa88, + 0x314f86, + 0x2bd405, + 0x389987, + 0x357007, + 0x2286c4, + 0x2b1287, + 0x236a04, + 0x236a06, + 0x219083, + 0x26a785, + 0x36ef05, + 0x362808, + 0x27f385, + 0x27d2c9, + 0x234bc7, + 0x234bcb, + 0x2a320c, + 0x2a380a, + 0x33a007, + 0x200a03, + 0x27b5c8, + 0x249905, + 0x2d5385, + 0x34ffc4, + 0x2d7dc6, + 0x27c106, + 0x3527c7, + 0x24ed8b, + 0x29b784, + 0x2d2444, + 0x2c92c4, + 0x2ceb06, + 0x20ff04, + 0x2bf148, + 0x34fdc5, + 0x254005, + 0x369bc7, + 0x274dc9, + 0x356dc5, + 0x386e4a, + 0x241b89, + 0x2dcc8a, + 0x2371c9, + 0x36e444, + 0x2e2245, + 0x2d9388, + 0x2e78cb, + 0x2a0885, + 0x2f47c6, + 0x242e44, + 0x27d806, + 0x389549, + 0x2e2c07, + 0x3c1608, + 0x2a27c6, + 0x3a13c7, + 0x2833c8, + 0x3873c6, + 0x3ca004, + 0x378647, + 0x368105, + 0x37a9c7, + 0x218fc4, + 0x38c646, + 0x2e8a88, + 0x297a48, + 0x2f6107, + 0x24a488, + 0x295e45, + 0x215184, + 0x248b08, + 0x24a584, + 0x241985, + 0x3005c4, + 0x2acf47, + 0x28a407, + 0x284208, + 0x2d4e46, + 0x27f305, + 0x27d0c8, + 0x210108, + 0x2a05c9, + 0x2b1346, + 0x22d688, + 0x206a4a, + 0x3245c8, + 0x2ea205, + 0x222386, + 0x241a48, + 0x379d4a, + 0x235207, + 0x287d85, + 0x2909c8, + 0x2af304, + 0x2d0e86, + 0x2c59c8, + 0x215206, + 0x393988, + 0x258c87, + 0x2136c6, + 0x2b93c4, + 0x278307, + 0x2b5544, + 0x389507, + 0x2d5b8d, + 0x232985, + 0x2e744b, + 0x38dc46, + 0x253bc8, + 0x244084, + 0x2f01c6, + 0x280286, + 0x228487, + 0x2970cd, + 0x247207, + 0x2b5b08, + 0x283b45, + 0x257288, + 0x2ca206, + 0x295ec8, + 0x23a086, + 0x38a507, + 0x388bc9, + 0x359b47, + 0x2893c8, + 0x2736c5, + 0x229208, + 0x334e05, + 0x2fe545, + 0x237445, + 0x223e43, + 0x2392c4, + 0x25e945, + 0x374e49, + 0x372b06, + 0x2e28c8, + 0x379f45, + 0x2b7347, + 0x2a0aca, + 0x2cec89, + 0x2c774a, + 0x2d8988, + 0x2d348c, + 0x284f4d, + 0x306983, + 0x393888, + 0x209ac5, + 0x3108c6, + 0x35c946, + 0x355485, + 0x226149, + 0x200985, + 0x27d0c8, + 0x25aa46, + 0x358e46, + 0x2a1fc9, + 0x3ac607, + 0x2912c6, + 0x2a0a48, + 0x3906c8, + 0x2e4907, + 0x2c294e, + 0x2ca445, + 0x319a45, + 0x215108, + 0x2ebd87, + 0x206a02, + 0x2c2d84, + 0x24f88a, + 0x319dc8, + 0x328b46, + 0x298f08, + 0x29e506, + 0x390408, + 0x2ae388, + 0x2fe504, + 0x2b7705, + 0x679b44, + 0x679b44, + 0x679b44, + 0x203c83, + 0x335106, + 0x27d586, + 0x29d78c, + 0x206b03, + 0x218fc6, + 0x24c3c4, + 0x288f08, + 0x2a4f45, + 0x24f986, + 0x2c0908, + 0x2d9fc6, + 0x39c446, + 0x2d2a48, + 0x2dac87, + 0x31d089, + 0x38524a, + 0x219bc4, + 0x236a45, + 0x30f585, + 0x388946, + 0x22fe06, + 0x29cb46, + 0x2ff506, + 0x31d1c4, + 0x31d1cb, + 0x2364c4, + 0x249485, + 0x2ac805, + 0x2a0e86, + 0x208b08, + 0x284e07, + 0x32c4c4, + 0x213c03, + 0x2aee05, + 0x236847, + 0x284d0b, + 0x362707, + 0x2c0808, + 0x2b7847, + 0x26bfc6, + 0x27e6c8, + 0x29cd4b, + 0x394f46, + 0x21ed49, + 0x29cec5, + 0x316fc3, + 0x2ced46, + 0x258b88, + 0x21a943, + 0x236943, + 0x2833c6, + 0x29e506, + 0x36f60a, + 0x27f7c5, + 0x2800cb, + 0x29e34b, + 0x208a03, + 0x21b0c3, + 0x2b24c4, + 0x2a9487, + 0x258c04, + 0x24f544, + 0x2b4c04, + 0x3248c8, + 0x2bd348, + 0x214889, + 0x2cd608, + 0x2376c7, + 0x22ec86, + 0x2e250f, + 0x2ca586, + 0x2d80c4, + 0x2bd18a, + 0x236747, + 0x2b5646, + 0x290809, + 0x214805, + 0x362945, + 0x214946, + 0x229343, + 0x2af349, + 0x278a86, + 0x21a689, + 0x39fdc6, + 0x26a785, + 0x223fc5, + 0x203b83, + 0x2a95c8, + 0x32cc07, + 0x2287c4, + 0x288d88, + 0x24c704, + 0x350606, + 0x2eb506, + 0x240346, + 0x2da949, + 0x2d5305, + 0x297dc6, + 0x24bac9, + 0x2c8606, + 0x2e2606, + 0x3a5fc6, + 0x237605, + 0x3005c6, + 0x38a504, + 0x3b6c45, + 0x2c41c4, + 0x2b6446, + 0x38c7c4, + 0x200c43, + 0x2870c5, + 0x232408, + 0x2afbc7, + 0x347649, + 0x287c88, + 0x298551, + 0x2d2cca, + 0x2f2b07, + 0x2b3806, + 0x24c3c4, + 0x2c42c8, + 0x352908, + 0x29870a, + 0x2bad0d, + 0x24c446, + 0x2d2b46, + 0x2783c6, + 0x365047, + 0x2b5bc5, + 0x326807, + 0x288e45, + 0x2df644, + 0x2a7f46, + 0x352607, + 0x2af04d, + 0x241987, + 0x373748, + 0x27d3c9, + 0x222286, + 0x3a7845, + 0x23e144, + 0x359f86, + 0x2285c6, + 0x319f46, + 0x299788, + 0x227803, + 0x228483, + 0x38b905, + 0x27f446, + 0x2ae345, + 0x2a29c8, + 0x29c5ca, + 0x30ff44, + 0x288f08, + 0x293508, + 0x2a0cc7, + 0x37a009, + 0x2c0508, + 0x283a07, + 0x2b4e86, + 0x21520a, + 0x35a008, + 0x3a46c9, + 0x2b2b08, + 0x2a7b49, + 0x2b36c7, + 0x377205, + 0x36c386, + 0x2b8ec8, + 0x380248, + 0x28ff08, + 0x2f9b48, + 0x249485, + 0x212ec4, + 0x230a48, + 0x242bc4, + 0x236fc4, + 0x26a785, + 0x292447, + 0x274b89, + 0x228287, + 0x206a85, + 0x279606, + 0x362206, + 0x206d44, + 0x2a2306, + 0x27b244, + 0x29f0c6, + 0x274946, + 0x229946, + 0x3cc5c5, + 0x2a2887, + 0x200a03, + 0x224a89, + 0x34b588, + 0x283884, + 0x28388d, + 0x297b48, + 0x2ea6c8, + 0x3a4646, + 0x388cc9, + 0x2cec89, + 0x389245, + 0x29c6ca, + 0x26f3ca, + 0x28a5cc, + 0x28a746, + 0x27afc6, + 0x2cae06, + 0x3a0909, + 0x310b06, + 0x2a12c6, + 0x200a46, + 0x234d88, + 0x20ff06, + 0x2d7a4b, + 0x2925c5, + 0x254005, + 0x27bd85, + 0x327006, + 0x2151c3, + 0x2402c6, + 0x241907, + 0x2c4185, + 0x25cac5, + 0x387945, + 0x3c8a86, + 0x332404, + 0x332406, + 0x2c37c9, + 0x326e8c, + 0x2df388, + 0x233204, + 0x3002c6, + 0x38dd46, + 0x258b88, + 0x354948, + 0x326d89, + 0x389987, + 0x318949, + 0x2486c6, + 0x22a0c4, + 0x20bb44, + 0x2831c4, + 0x2833c8, + 0x2749ca, + 0x356d46, + 0x364207, + 0x37ac47, + 0x249405, + 0x350f04, + 0x28e106, + 0x2b5c06, + 0x240c03, + 0x34b3c7, + 0x39eec8, + 0x38938a, + 0x302008, + 0x362588, + 0x38c805, + 0x29dbc5, + 0x267385, + 0x2497c6, + 0x3777c6, + 0x391ac5, + 0x32c349, + 0x350d0c, + 0x267447, + 0x298788, + 0x272145, + 0x679b44, + 0x2e95c4, + 0x284984, + 0x221706, + 0x29f7ce, + 0x3629c7, + 0x365245, + 0x3a49cc, + 0x3006c7, + 0x352587, + 0x22c109, + 0x26d249, + 0x287d85, + 0x34b588, + 0x3cfb09, + 0x345a85, + 0x2c40c8, + 0x278c06, + 0x248d86, + 0x275284, + 0x28cc08, + 0x222443, + 0x27a7c4, + 0x2aee85, + 0x399c87, + 0x336145, + 0x206909, + 0x2aac4d, + 0x2b3bc6, + 0x213c44, + 0x2bf2c8, + 0x28174a, + 0x26df87, + 0x36ba85, + 0x24d583, + 0x29e50e, + 0x2a96cc, + 0x2fdbc7, + 0x29f987, + 0x219003, + 0x310b45, + 0x284985, + 0x2992c8, + 0x296bc9, + 0x233106, + 0x258c04, + 0x2f2a46, + 0x237c0b, + 0x2de40c, + 0x2d6607, + 0x2d7d05, + 0x3c0848, + 0x2e46c5, + 0x2bd187, + 0x384c47, + 0x241745, + 0x2151c3, + 0x324c04, + 0x359d45, + 0x2a7805, + 0x2a7806, + 0x2bfa48, + 0x352607, + 0x35cc46, + 0x203f46, + 0x237386, + 0x288289, + 0x27a247, + 0x31a206, + 0x2de586, + 0x277bc6, + 0x2a91c5, + 0x2105c6, + 0x3633c5, + 0x2e5e08, + 0x291d4b, + 0x28db06, + 0x37ac84, + 0x2ffc09, + 0x234bc4, + 0x278b88, + 0x38b787, + 0x285944, + 0x2bfc48, + 0x2c5444, + 0x2a9204, + 0x288cc5, + 0x3162c6, + 0x324807, + 0x2087c3, + 0x29df85, + 0x338a84, + 0x319a86, + 0x3892c8, + 0x38d8c5, + 0x291a09, + 0x33a005, + 0x218fc8, + 0x3cf847, + 0x32c648, + 0x2bee87, + 0x25bc09, + 0x27df06, + 0x33da46, + 0x200a44, + 0x2d2385, + 0x30804c, + 0x27bd87, + 0x27ca47, + 0x22fa48, + 0x2b3bc6, + 0x270b44, + 0x3b1bc4, + 0x283549, + 0x2caf06, + 0x271a87, + 0x2391c4, + 0x248346, + 0x35c185, + 0x2d4987, + 0x2d79c6, + 0x261049, + 0x2cfac7, + 0x295407, + 0x2a1e46, + 0x248285, + 0x281408, + 0x278908, + 0x22ee86, + 0x38d905, + 0x329806, + 0x210283, + 0x299149, + 0x29c8ce, + 0x2bebc8, + 0x24c808, + 0x22ec8b, + 0x291c46, + 0x34c404, + 0x284b44, + 0x29c9ca, + 0x212707, + 0x31a2c5, + 0x21ed49, + 0x2c2885, + 0x237007, + 0x24a404, + 0x2a5947, + 0x2f2cc8, + 0x2e4fc6, + 0x24c549, + 0x2c060a, + 0x212686, + 0x297686, + 0x2ac785, + 0x39b305, + 0x33eac7, + 0x247948, + 0x35c0c8, + 0x2fe506, + 0x224045, + 0x22fb8e, + 0x2c0cc4, + 0x22ee05, + 0x278f89, + 0x2c7e88, + 0x28c2c6, + 0x29b20c, + 0x29c1d0, + 0x29f40f, + 0x2a1008, + 0x33a007, + 0x3cc5c5, + 0x25e945, + 0x324689, + 0x290bc9, + 0x238786, + 0x2a0907, + 0x2d2285, + 0x232909, + 0x349806, + 0x31094d, + 0x283089, + 0x24f544, + 0x2be948, + 0x230b09, + 0x356f06, + 0x27b7c5, + 0x33da46, + 0x3c14c9, + 0x279e48, + 0x21a3c5, + 0x206b44, + 0x29b3cb, + 0x356dc5, + 0x29b506, + 0x285286, + 0x2023c6, + 0x2a030b, + 0x291b09, + 0x203e85, + 0x3a6a87, + 0x2d2c46, + 0x24f6c6, + 0x284708, + 0x3163c9, + 0x37350c, + 0x236648, + 0x30f5c6, + 0x338403, + 0x2307c6, + 0x25bac5, + 0x280408, + 0x223a46, + 0x2d4bc8, + 0x247d45, + 0x2988c5, + 0x2576c8, + 0x390587, + 0x35c887, + 0x3527c7, + 0x24f548, + 0x388a48, + 0x2e3046, + 0x2b6287, + 0x3536c7, + 0x2a000a, + 0x23ce43, + 0x327006, + 0x22fb05, + 0x24f884, + 0x27d3c9, + 0x25bb84, + 0x2afc44, + 0x29a604, + 0x29f98b, + 0x32cb47, + 0x22fdc5, + 0x295b48, + 0x279606, + 0x279608, + 0x27f706, + 0x28cb45, + 0x28ce05, + 0x28ec46, + 0x28fcc8, + 0x290748, + 0x27d586, + 0x29598f, + 0x298c10, + 0x2082c5, + 0x200a03, + 0x22a185, + 0x30f788, + 0x290ac9, + 0x345bc8, + 0x24c3c8, + 0x23d1c8, + 0x32cc07, + 0x2792c9, + 0x2d4dc8, + 0x2ab6c4, + 0x29a488, + 0x2d9c09, + 0x2b6d07, + 0x29a404, + 0x228348, + 0x2a264a, + 0x2e61c6, + 0x24c446, + 0x2b1209, + 0x29c407, + 0x2d0188, + 0x2254c8, + 0x2c9b88, + 0x24bf05, + 0x20a005, + 0x254005, + 0x284945, + 0x3bc1c7, + 0x2151c5, + 0x2c4185, + 0x201186, + 0x345b07, + 0x2e7807, + 0x2a2946, + 0x2d8ec5, + 0x29b506, + 0x24c645, + 0x2d2108, + 0x2fda44, + 0x2c8686, + 0x32f004, + 0x206408, + 0x31d94a, + 0x27dc8c, + 0x24ef85, + 0x365106, + 0x3736c6, + 0x294506, + 0x30f644, + 0x35c445, + 0x27f087, + 0x29c489, + 0x2cf0c7, + 0x679b44, + 0x679b44, + 0x32c9c5, + 0x2d6bc4, + 0x29abca, + 0x279486, + 0x27e904, + 0x210885, + 0x398905, + 0x2b5b04, + 0x284ec7, + 0x35b147, + 0x2ceb08, + 0x38b088, + 0x21a3c9, + 0x26f048, + 0x29ad8b, + 0x2db304, + 0x248605, + 0x281605, + 0x352749, + 0x3163c9, + 0x2ffb08, + 0x2364c8, + 0x2a0e84, + 0x38dd85, + 0x20a343, + 0x388905, + 0x297e46, + 0x296a0c, + 0x2390c6, + 0x27b6c6, + 0x28c545, + 0x3c8b08, + 0x3cdf46, + 0x2b3986, + 0x24c446, + 0x301d8c, + 0x31a104, + 0x2374ca, + 0x28c488, + 0x296847, + 0x338986, + 0x2331c7, + 0x2f2645, + 0x372c06, + 0x361046, + 0x36d8c7, + 0x2afc84, + 0x2ad045, + 0x278f84, + 0x2df6c7, + 0x2791c8, + 0x27ae4a, + 0x282907, + 0x21ac07, + 0x339f87, + 0x2e4809, + 0x296a0a, + 0x22a083, + 0x2afb85, + 0x229983, + 0x2b4c49, + 0x38a788, + 0x357147, + 0x345cc9, + 0x278a06, + 0x3b6d08, + 0x34a6c5, + 0x21020a, + 0x392b49, + 0x24cb49, + 0x3cde07, + 0x352a09, + 0x229848, + 0x36da86, + 0x3652c8, + 0x3d0647, + 0x2b49c7, + 0x241b87, + 0x2e8908, + 0x300146, + 0x2a2405, + 0x27f087, + 0x297188, + 0x32ef84, + 0x2fa084, + 0x2911c7, + 0x2ae707, + 0x3cf98a, + 0x36da06, + 0x25708a, + 0x2c2cc7, + 0x2c0a87, + 0x2ad104, + 0x2a6544, + 0x2d4886, + 0x3119c4, + 0x3119cc, + 0x204dc5, + 0x354789, + 0x2b28c4, + 0x2b5bc5, + 0x2816c8, + 0x27eac5, + 0x386e46, + 0x22c004, + 0x2db20a, + 0x330ac6, + 0x29368a, + 0x219b87, + 0x286c45, + 0x229345, + 0x24944a, + 0x28fe45, + 0x2a0686, + 0x242bc4, + 0x2b2646, + 0x33eb85, + 0x223b06, + 0x2f610c, + 0x354aca, + 0x26f4c4, + 0x22ec86, + 0x29c407, + 0x2d7944, + 0x234d88, + 0x2f46c6, + 0x37aac9, + 0x2dc6c9, + 0x2b7109, + 0x2cf3c6, + 0x3d0746, + 0x365407, + 0x32c288, + 0x3d0549, + 0x32cb47, + 0x295cc6, + 0x3a1447, + 0x278285, + 0x2c0cc4, + 0x364fc7, + 0x353885, + 0x288c05, + 0x3280c7, + 0x241608, + 0x3c07c6, + 0x297fcd, + 0x2994cf, + 0x29e34d, + 0x206ac4, + 0x232506, + 0x2db5c8, + 0x200a05, + 0x2a01c8, + 0x26b30a, + 0x24f544, + 0x237e46, + 0x2d8147, + 0x3bd347, + 0x2dad49, + 0x365285, + 0x2b5b04, + 0x2b764a, + 0x2c00c9, + 0x352b07, + 0x298286, + 0x356f06, + 0x38dcc6, + 0x378706, + 0x2da60f, + 0x2db489, + 0x20ff06, + 0x37fe86, + 0x32b949, + 0x2b6387, + 0x208e43, + 0x2375c6, + 0x213383, + 0x355348, + 0x3a1287, + 0x2a1209, + 0x2eb388, + 0x35c9c8, + 0x200c86, + 0x239009, + 0x384d85, + 0x230704, + 0x2f2f47, + 0x3a0985, + 0x206ac4, + 0x22fe88, + 0x2129c4, + 0x2b60c7, + 0x364b06, + 0x280745, + 0x2b2b08, + 0x356dcb, + 0x351247, + 0x2496c6, + 0x2ca604, + 0x34c386, + 0x26a785, + 0x353885, + 0x281189, + 0x284ac9, + 0x2b4a04, + 0x2b4a45, + 0x22ecc5, + 0x210086, + 0x34b688, + 0x2c2246, + 0x39ed0b, + 0x37c4ca, + 0x2bef85, + 0x28ce86, + 0x30fc45, + 0x2010c5, + 0x29bd87, + 0x327288, + 0x26f0c4, + 0x266006, + 0x2907c6, + 0x229a07, + 0x316f84, + 0x280286, + 0x2f0c45, + 0x2f0c49, + 0x3d0944, + 0x351089, + 0x27d586, + 0x2c4c48, + 0x22ecc5, + 0x37ad45, + 0x223b06, + 0x373409, + 0x26d249, + 0x27b746, + 0x2c7f88, + 0x2aad88, + 0x30fc04, + 0x2b81c4, + 0x2b81c8, + 0x2bc448, + 0x318a49, + 0x297dc6, + 0x24c446, + 0x337a4d, + 0x384906, + 0x2ccc89, + 0x38ad45, + 0x214946, + 0x2c9d08, + 0x332345, + 0x353704, + 0x26a785, + 0x284408, + 0x29a989, + 0x279044, + 0x38c646, + 0x27e98a, + 0x2fdac8, + 0x3cfb09, + 0x253eca, + 0x345c46, + 0x299688, + 0x2bcf45, + 0x289e48, + 0x2f26c5, + 0x2788c9, + 0x37cf49, + 0x204ec2, + 0x29cec5, + 0x270686, + 0x27d4c7, + 0x381345, + 0x314e86, + 0x30c4c8, + 0x2b3bc6, + 0x2d9249, + 0x27cb46, + 0x284588, + 0x26d7c5, + 0x22b786, + 0x38a608, + 0x2833c8, + 0x2b35c8, + 0x2e9a48, + 0x2105c4, + 0x228703, + 0x2d9484, + 0x282b06, + 0x2782c4, + 0x24c747, + 0x2b3889, + 0x2c92c5, + 0x2254c6, + 0x2375c6, + 0x2bf88b, + 0x2b5586, + 0x2bc746, + 0x2c8788, + 0x3224c6, + 0x2bb003, + 0x20e383, + 0x2c0cc4, + 0x22d585, + 0x2d0b47, + 0x2791c8, + 0x2791cf, + 0x27ef8b, + 0x34b488, + 0x38c6c6, + 0x34b78e, + 0x223b03, + 0x2d0ac4, + 0x2b5505, + 0x2b5986, + 0x28e20b, + 0x292506, + 0x31a589, + 0x280745, + 0x254448, + 0x217008, + 0x26d10c, + 0x29f9c6, + 0x388946, + 0x2ec3c5, + 0x289188, + 0x27dc85, + 0x355d48, + 0x29b58a, + 0x29e789, + 0x679b44, + 0x2000c2, + 0x3f208e02, + 0x200382, + 0x220a04, + 0x202302, + 0x20ac44, + 0x201402, + 0x14b83, + 0x2003c2, + 0x211302, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x38d8c3, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x21d283, + 0x259003, + 0x20db43, + 0x2879c4, + 0x208e03, + 0x233e84, + 0x231103, + 0x2dd584, + 0x213ec3, + 0x381607, + 0x270203, + 0x214b83, + 0x23a708, + 0x259003, + 0x29ed0b, + 0x2f3983, + 0x3adc46, + 0x2022c2, + 0x23ef4b, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x259003, + 0x26d503, + 0x21e403, + 0x2000c2, + 0x120648, + 0x20f385, + 0x353908, + 0x2e8d88, + 0x208e02, + 0x38d545, + 0x307a47, + 0x202bc2, + 0x2442c7, + 0x200382, + 0x25dc87, + 0x2bb789, + 0x2bcb08, + 0x2c9a09, + 0x20ecc2, + 0x269f07, + 0x22eb04, + 0x307b07, + 0x37c3c7, + 0x25e242, + 0x270203, + 0x202382, + 0x201402, + 0x2003c2, + 0x209302, + 0x200902, + 0x211302, + 0x2a9cc5, + 0x325a85, + 0x8e02, + 0x31103, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0xe6143, + 0x259003, + 0xe103, + 0x101, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x217383, + 0x21d283, + 0xe6143, + 0x259003, + 0x216cc3, + 0x420730c6, + 0x9d303, + 0xca845, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0xe6143, + 0x259003, + 0x6742, + 0x120648, + 0x14b83, + 0xe6143, + 0x47904, + 0xe4ac5, + 0x2000c2, + 0x3afcc4, + 0x208e03, + 0x231103, + 0x213ec3, + 0x245943, + 0x22bec5, + 0x217383, + 0x221303, + 0x21d283, + 0x24fc43, + 0x259003, + 0x211303, + 0x25f383, + 0x202043, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e02, + 0x259003, + 0x120648, + 0x213ec3, + 0xe6143, + 0x120648, + 0xe6143, + 0x2b9343, + 0x208e03, + 0x22de84, + 0x231103, + 0x213ec3, + 0x200d42, + 0x270203, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x200d42, + 0x208e83, + 0x21d283, + 0x259003, + 0x2ec903, + 0x211303, + 0x2000c2, + 0x208e02, + 0x213ec3, + 0x21d283, + 0x259003, + 0x3adc45, + 0x14f9c6, + 0x2879c4, + 0x2022c2, + 0x120648, + 0x2000c2, + 0xf8945, + 0x1b308, + 0x18ec43, + 0x208e02, + 0x46491f86, + 0x1f484, + 0xb19cb, + 0x395c6, + 0x7a6c7, + 0x231103, + 0x4b808, + 0x213ec3, + 0x111185, + 0x192d44, + 0x267503, + 0x526c7, + 0xe0a84, + 0x21d283, + 0x7b846, + 0xe5f84, + 0xe6143, + 0x259003, + 0x2f5544, + 0x12d147, + 0x14f5c9, + 0xb1788, + 0x1436c4, + 0x3f5c6, + 0xe1c8, + 0x131b05, + 0xe4c9, + 0xf8945, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x214b83, + 0x259003, + 0x2f3983, + 0x2022c2, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x2171c3, + 0x2bf144, + 0x21d283, + 0x14b83, + 0x259003, + 0x208e03, + 0x231103, + 0x2dd584, + 0x213ec3, + 0x21d283, + 0x259003, + 0x3adc46, + 0x231103, + 0x213ec3, + 0x2ab03, + 0xe6143, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0xf8945, + 0x7a6c7, + 0x120648, + 0x213ec3, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x49208e03, + 0x231103, + 0x21d283, + 0x259003, + 0x120648, + 0x2000c2, + 0x208e02, + 0x208e03, + 0x213ec3, + 0x21d283, + 0x2003c2, + 0x259003, + 0x33a8c7, + 0x31258b, + 0x201103, + 0x238448, + 0x32c007, + 0x2116c6, + 0x209885, + 0x38d689, + 0x27a348, + 0x3726c9, + 0x3a8bd0, + 0x3726cb, + 0x2f1789, + 0x20fac3, + 0x201389, + 0x22f406, + 0x22f40c, + 0x20f448, + 0x3cdc48, + 0x3b8209, + 0x35194e, + 0x2bb54b, + 0x36c5cc, + 0x2038c3, + 0x28f9cc, + 0x2038c9, + 0x23cc47, + 0x23104c, + 0x3c79ca, + 0x245a04, + 0x24d20d, + 0x28f888, + 0x20db4d, + 0x292c06, + 0x2879cb, + 0x34f749, + 0x380107, + 0x389c06, + 0x326289, + 0x388eca, + 0x31ac48, + 0x2f3584, + 0x33d187, + 0x246347, + 0x2393c4, + 0x2e53c4, + 0x399189, + 0x394d89, + 0x219588, + 0x20d8c5, + 0x20ec05, + 0x20b006, + 0x24d0c9, + 0x26b58d, + 0x2f48c8, + 0x20af07, + 0x209908, + 0x31f446, + 0x23b984, + 0x2a8f45, + 0x3d0446, + 0x3d1fc4, + 0x2037c7, + 0x20588a, + 0x210744, + 0x2125c6, + 0x213349, + 0x21334f, + 0x214e0d, + 0x215c86, + 0x21af10, + 0x21b306, + 0x21b887, + 0x21c147, + 0x21c14f, + 0x21dd09, + 0x2245c6, + 0x224cc7, + 0x224cc8, + 0x2258c9, + 0x280808, + 0x2e9e87, + 0x20f383, + 0x22acc6, + 0x3c49c8, + 0x351c0a, + 0x3c5dc9, + 0x222703, + 0x307946, + 0x265e4a, + 0x28df47, + 0x23ca8a, + 0x30a20e, + 0x21de46, + 0x29d0c7, + 0x354546, + 0x203986, + 0x209e0b, + 0x20ed8a, + 0x2ad44d, + 0x3d0807, + 0x2694c8, + 0x2694c9, + 0x2694cf, + 0x3477cc, + 0x26cd09, + 0x2b2fce, + 0x38170a, + 0x2bd7c6, + 0x2fc586, + 0x31cdcc, + 0x31e64c, + 0x322088, + 0x359a47, + 0x235105, + 0x292304, + 0x39520e, + 0x266744, + 0x35d187, + 0x3a6fca, + 0x3c7ed4, + 0x3cbc0f, + 0x21c308, + 0x22ab88, + 0x33e48d, + 0x33e48e, + 0x22ae49, + 0x22c588, + 0x22c58f, + 0x230d4c, + 0x230d4f, + 0x232247, + 0x2346ca, + 0x246b0b, + 0x237908, + 0x23b307, + 0x2606cd, + 0x3319c6, + 0x24d3c6, + 0x240149, + 0x278708, + 0x244c48, + 0x244c4e, + 0x312687, + 0x246dc5, + 0x2476c5, + 0x204304, + 0x211986, + 0x219488, + 0x394583, + 0x2f42ce, + 0x260a88, + 0x372dcb, + 0x309207, + 0x2fe345, + 0x2474c6, + 0x2ab4c7, + 0x2fa608, + 0x33e8c9, + 0x22e5c5, + 0x287108, + 0x2355c6, + 0x3aa44a, + 0x395109, + 0x231109, + 0x23110b, + 0x393648, + 0x239289, + 0x20d986, + 0x37510a, + 0x2088ca, + 0x2348cc, + 0x228a87, + 0x2c980a, + 0x32d60b, + 0x32d619, + 0x317888, + 0x3adcc5, + 0x260886, + 0x26bc49, + 0x381ec6, + 0x2d324a, + 0x27a546, + 0x2353c4, + 0x2cb98d, + 0x3949c7, + 0x2353c9, + 0x24a785, + 0x24ad88, + 0x24b5c9, + 0x24ba04, + 0x24d807, + 0x24d808, + 0x24dcc7, + 0x268748, + 0x253107, + 0x204105, + 0x25a00c, + 0x25a849, + 0x2cf5ca, + 0x3ac489, + 0x201489, + 0x37fc4c, + 0x25ee0b, + 0x25fac8, + 0x260e88, + 0x264944, + 0x285608, + 0x2864c9, + 0x3c7a87, + 0x213586, + 0x29a7c7, + 0x3cf649, + 0x324d4b, + 0x294387, + 0x3cc987, + 0x219cc7, + 0x20dac4, + 0x20dac5, + 0x2dd285, + 0x34f34b, + 0x3b5a04, + 0x2be748, + 0x2f774a, + 0x235687, + 0x358507, + 0x28d692, + 0x29efc6, + 0x22d806, + 0x38ff8e, + 0x2a2f46, + 0x293388, + 0x29390f, + 0x20df08, + 0x39cfc8, + 0x2c474a, + 0x2c4751, + 0x2a2bce, + 0x23b60a, + 0x23b60c, + 0x22c787, + 0x22c790, + 0x3c9e48, + 0x2a2dc5, + 0x2ab7ca, + 0x3d200c, + 0x29600d, + 0x304486, + 0x304487, + 0x30448c, + 0x383c4c, + 0x21da0c, + 0x2acacb, + 0x383104, + 0x2b1384, + 0x385749, + 0x3b1c47, + 0x384649, + 0x208709, + 0x3bdd87, + 0x3c7846, + 0x3c7849, + 0x2ae8c3, + 0x2b3cca, + 0x38cd87, + 0x335c4b, + 0x2ad2ca, + 0x22eb84, + 0x35c586, + 0x282b89, + 0x311844, + 0x204e8a, + 0x2499c5, + 0x2c10c5, + 0x2c10cd, + 0x2c140e, + 0x2d95c5, + 0x339106, + 0x3ad847, + 0x25a28a, + 0x370ac6, + 0x2ec304, + 0x303187, + 0x2700cb, + 0x31f507, + 0x27a944, + 0x3010c6, + 0x3010cd, + 0x2dfa4c, + 0x21d146, + 0x2f4aca, + 0x3472c6, + 0x23e248, + 0x22f087, + 0x245b8a, + 0x382c86, + 0x280303, + 0x2bd8c6, + 0x3c4848, + 0x3858ca, + 0x285bc7, + 0x285bc8, + 0x289fc4, + 0x33a407, + 0x202d88, + 0x298908, + 0x271bc8, + 0x2e314a, + 0x2e4345, + 0x208e87, + 0x23b453, + 0x258146, + 0x2141c8, + 0x2201c9, + 0x244188, + 0x200d0b, + 0x35cd48, + 0x270204, + 0x2577c6, + 0x31b786, + 0x316109, + 0x2c7a47, + 0x25a108, + 0x298a86, + 0x327fc4, + 0x326b85, + 0x2cf908, + 0x332e8a, + 0x2cb608, + 0x2d06c6, + 0x29988a, + 0x2a7988, + 0x2d7748, + 0x2d8308, + 0x2d8b86, + 0x2db7c6, + 0x3abf4c, + 0x2dbd90, + 0x2a4005, + 0x20dd08, + 0x330710, + 0x20dd10, + 0x3a8a4e, + 0x3abbce, + 0x3abbd4, + 0x3b144f, + 0x3b1806, + 0x390d11, + 0x389d93, + 0x38a208, + 0x3251c5, + 0x238988, + 0x2fd945, + 0x2e5b0c, + 0x226bc9, + 0x292149, + 0x227047, + 0x397249, + 0x318d07, + 0x300446, + 0x2a8d47, + 0x201c05, + 0x20e143, + 0x394749, + 0x25cf49, + 0x22ab03, + 0x381244, + 0x27548d, + 0x2f8f0f, + 0x328005, + 0x387f86, + 0x228cc7, + 0x20f1c7, + 0x2fc8c6, + 0x2fc8cb, + 0x2a39c5, + 0x25cd06, + 0x301747, + 0x253809, + 0x37cac6, + 0x392605, + 0x363b4b, + 0x20a286, + 0x229585, + 0x25b908, + 0x2a5748, + 0x32018c, + 0x320190, + 0x2b1d89, + 0x2b4447, + 0x313a0b, + 0x2e87c6, + 0x2e9d4a, + 0x2ea8cb, + 0x2eb68a, + 0x2eb906, + 0x2ec7c5, + 0x32bf06, + 0x2b0648, + 0x22710a, + 0x33e11c, + 0x2f3a4c, + 0x2f3d48, + 0x3adc45, + 0x382947, + 0x351586, + 0x281ac5, + 0x217846, + 0x2fca88, + 0x2c0347, + 0x351848, + 0x25820a, + 0x228dcc, + 0x20b209, + 0x225647, + 0x286a04, + 0x247786, + 0x39cb4a, + 0x208805, + 0x222e4c, + 0x224e88, + 0x31f748, + 0x334a4c, + 0x22d1cc, + 0x22e6c9, + 0x22e907, + 0x3a0c8c, + 0x21e984, + 0x24fe4a, + 0x303e0c, + 0x2723cb, + 0x25948b, + 0x25c906, + 0x25f907, + 0x22c9c7, + 0x22c9cf, + 0x304e51, + 0x2e1892, + 0x264acd, + 0x264ace, + 0x264e0e, + 0x3b1608, + 0x3b1612, + 0x26c788, + 0x220807, + 0x2503ca, + 0x2a62c8, + 0x2a2f05, + 0x3bc00a, + 0x21b687, + 0x2f6044, + 0x21c883, + 0x2343c5, + 0x2c49c7, + 0x3040c7, + 0x29620e, + 0x35020d, + 0x3593c9, + 0x24f405, + 0x3b01c3, + 0x218306, + 0x25d945, + 0x373008, + 0x2baa09, + 0x2608c5, + 0x2608cf, + 0x2b1ac7, + 0x209705, + 0x309bca, + 0x2f3206, + 0x268ac9, + 0x2fd48c, + 0x2ff149, + 0x209b46, + 0x2f754c, + 0x338506, + 0x302d08, + 0x303346, + 0x317a06, + 0x2b5704, + 0x316083, + 0x2b6a8a, + 0x219fd1, + 0x26ceca, + 0x26e985, + 0x271747, + 0x258587, + 0x202e84, + 0x202e8b, + 0x2bc988, + 0x2bea46, + 0x22fac5, + 0x3aa744, + 0x24ab09, + 0x2008c4, + 0x244a87, + 0x2ff345, + 0x2ff347, + 0x3901c5, + 0x24cac3, + 0x2206c8, + 0x35c20a, + 0x2087c3, + 0x20f3ca, + 0x3c1846, + 0x26064f, + 0x3bcc09, + 0x2f4250, + 0x2f9748, + 0x2d1749, + 0x296f07, + 0x30104f, + 0x346084, + 0x2dd604, + 0x21b186, + 0x280c46, + 0x3a41ca, + 0x27bb06, + 0x340907, + 0x30af88, + 0x30b187, + 0x30c287, + 0x30d24a, + 0x30eacb, + 0x326945, + 0x2e14c8, + 0x232a43, + 0x3b700c, + 0x343d0f, + 0x234f0d, + 0x25ac87, + 0x359509, + 0x241e87, + 0x25b148, + 0x3c80cc, + 0x2b8a48, + 0x26fcc8, + 0x32ea8e, + 0x341294, + 0x3417a4, + 0x358c0a, + 0x373b0b, + 0x318dc4, + 0x318dc9, + 0x237ec8, + 0x247ec5, + 0x39408a, + 0x260cc7, + 0x32be04, + 0x38d8c3, + 0x208e03, + 0x233e84, + 0x231103, + 0x213ec3, + 0x220a04, + 0x217383, + 0x270203, + 0x2dbd86, + 0x2bf144, + 0x21d283, + 0x259003, + 0x216203, + 0x2000c2, + 0x38d8c3, + 0x208e02, + 0x208e03, + 0x233e84, + 0x231103, + 0x213ec3, + 0x217383, + 0x2dbd86, + 0x21d283, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x3d0943, + 0x21d283, + 0xe6143, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x2bf144, + 0x21d283, + 0x259003, + 0x2000c2, + 0x25c183, + 0x208e02, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x200f82, + 0x248182, + 0x208e02, + 0x208e03, + 0x20f982, + 0x2005c2, + 0x220a04, + 0x20ac44, + 0x235a82, + 0x2bf144, + 0x2003c2, + 0x259003, + 0x216203, + 0x25c906, + 0x22a342, + 0x203642, + 0x220e82, + 0x4ba0df03, + 0x4be2c783, + 0x58fc6, + 0x58fc6, + 0x2879c4, + 0x214b83, + 0x1930a, + 0x112fcc, + 0xaacc, + 0xca64d, + 0xf8945, + 0x8b88c, + 0x67c47, + 0x10ec6, + 0x15e08, + 0x18dc7, + 0x1e748, + 0x17ba8a, + 0x103647, + 0x4ca8bac5, + 0xde0c9, + 0x32d0b, + 0x1c5b0b, + 0x7a748, + 0x15a49, + 0x1cd6ca, + 0x13f30e, + 0x1152cd, + 0x1443f8b, + 0xdf90a, + 0x1f484, + 0x68486, + 0x179b48, + 0x886c8, + 0x32fc7, + 0x1fd85, + 0x92f47, + 0x7c349, + 0x1118c7, + 0x667c8, + 0x107209, + 0x4a884, + 0x4db85, + 0x13bace, + 0x6d80d, + 0x7a548, + 0x4cf6af86, + 0x4d96af88, + 0xb4288, + 0x13aad0, + 0x54dcc, + 0x65b47, + 0x66607, + 0x6a9c7, + 0x70387, + 0xafc2, + 0x28c7, + 0xef4c, + 0x191b85, + 0x16bd07, + 0xa5606, + 0xa66c9, + 0xa80c8, + 0x55402, + 0x5c2, + 0x3d44b, + 0xe6007, + 0x11eb09, + 0x6b109, + 0x1656c8, + 0xb2d02, + 0x1a6cc9, + 0xd98ca, + 0x2306, + 0xce389, + 0xdf887, + 0xdffc9, + 0xe0d48, + 0xe1d47, + 0xe42c9, + 0xe6945, + 0xe6cd0, + 0x1d22c6, + 0x129cc5, + 0x8f4c7, + 0x11f04d, + 0x43105, + 0x25f46, + 0xee247, + 0xf5558, + 0x111c48, + 0x210a, + 0x8782, + 0x5924a, + 0x7324d, + 0x1002, + 0xc8346, + 0x8e4c8, + 0x49148, + 0x6e6c9, + 0x10b908, + 0x7f98e, + 0x18307, + 0x10668d, + 0xfbec5, + 0x2648, + 0x1ad488, + 0x106c86, + 0x67c2, + 0xdc586, + 0x3f5c6, + 0x1242, + 0x401, + 0x5f787, + 0x5d203, + 0x4d2f5dc4, + 0x4d694cc3, + 0xc1, + 0x12bc6, + 0xc1, + 0x201, + 0x12bc6, + 0x5d203, + 0x156ca05, + 0x245a04, + 0x208e03, + 0x24ea84, + 0x220a04, + 0x21d283, + 0x220085, + 0x216cc3, + 0x202203, + 0x2fc845, + 0x202043, + 0x4ea08e03, + 0x231103, + 0x213ec3, + 0x200181, + 0x270203, + 0x20ac44, + 0x2bf144, + 0x21d283, + 0x259003, + 0x211303, + 0x120648, + 0x2000c2, + 0x38d8c3, + 0x208e02, + 0x208e03, + 0x231103, + 0x3d0943, + 0x2005c2, + 0x220a04, + 0x217383, + 0x270203, + 0x21d283, + 0x214b83, + 0x259003, + 0x202043, + 0x120648, + 0x343a82, + 0x1018c7, + 0x8e02, + 0x126b05, + 0x54f0f, + 0x1582ac8, + 0x10bc8e, + 0x4fa23802, + 0x32b588, + 0x223c86, + 0x2c6046, + 0x223607, + 0x4fe04bc2, + 0x503bca88, + 0x22208a, + 0x2655c8, + 0x204342, + 0x335a89, + 0x326987, + 0x213506, + 0x220409, + 0x206084, + 0x2115c6, + 0x2c6444, + 0x201904, + 0x259c09, + 0x303b06, + 0x2e95c5, + 0x253585, + 0x22bc07, + 0x2c2f47, + 0x29f204, + 0x223846, + 0x2f6cc5, + 0x2acdc5, + 0x30fb85, + 0x214487, + 0x309045, + 0x32a249, + 0x38a945, + 0x2fa744, + 0x370a07, + 0x32744e, + 0x390909, + 0x38fe49, + 0x324386, + 0x2412c8, + 0x2f02cb, + 0x36230c, + 0x344d86, + 0x36c487, + 0x2b2745, + 0x2e53ca, + 0x219689, + 0x348949, + 0x38e9c6, + 0x301505, + 0x247a45, + 0x331709, + 0x30fd0b, + 0x277d46, + 0x34aa86, + 0x20af04, + 0x28d346, + 0x246e48, + 0x3c46c6, + 0x26ddc6, + 0x204688, + 0x206607, + 0x206e49, + 0x207645, + 0x120648, + 0x292ec4, + 0x30c804, + 0x20ea85, + 0x3b32c9, + 0x21f107, + 0x21f10b, + 0x22294a, + 0x226b05, + 0x5060da02, + 0x2ad187, + 0x50a26e08, + 0x211cc7, + 0x2df085, + 0x32e6ca, + 0x8e02, + 0x25324b, + 0x27b0ca, + 0x25ce46, + 0x20e243, + 0x2d58cd, + 0x39ea8c, + 0x20fc4d, + 0x24a3c5, + 0x36efc5, + 0x3945c7, + 0x215609, + 0x221f86, + 0x257cc5, + 0x30ad88, + 0x28d243, + 0x2e9088, + 0x28d248, + 0x2c7147, + 0x34c0c8, + 0x39dd09, + 0x2e2387, + 0x312107, + 0x335908, + 0x2f9e44, + 0x2f9e47, + 0x292b08, + 0x359286, + 0x35d50f, + 0x31ed07, + 0x355006, + 0x22ea45, + 0x221003, + 0x248f07, + 0x37ef43, + 0x24e146, + 0x2500c6, + 0x250946, + 0x291805, + 0x268743, + 0x3a6948, + 0x3809c9, + 0x39f44b, + 0x250ac8, + 0x252dc5, + 0x254145, + 0x50e2ec42, + 0x2a8e09, + 0x220a87, + 0x25cd85, + 0x259b07, + 0x25c506, + 0x3785c5, + 0x25d78b, + 0x25fac4, + 0x265185, + 0x2652c7, + 0x2776c6, + 0x277b05, + 0x285807, + 0x285f87, + 0x2e7784, + 0x28b68a, + 0x28c0c8, + 0x2bcfc9, + 0x238cc5, + 0x362b06, + 0x24700a, + 0x253486, + 0x269007, + 0x2bcc8d, + 0x2a3509, + 0x3a5c45, + 0x395787, + 0x327848, + 0x38a3c8, + 0x39e787, + 0x205e06, + 0x227647, + 0x24fc83, + 0x303a84, + 0x375d85, + 0x3ab2c7, + 0x3af6c9, + 0x2678c8, + 0x3c8385, + 0x245dc4, + 0x250c85, + 0x25fc8d, + 0x205002, + 0x2c1f46, + 0x2c8286, + 0x30cfca, + 0x396106, + 0x39ca85, + 0x38b185, + 0x38b187, + 0x3aa28c, + 0x2b410a, + 0x28d006, + 0x2db6c5, + 0x28d186, + 0x28d4c7, + 0x28f186, + 0x29170c, + 0x220549, + 0x51213047, + 0x293cc5, + 0x293cc6, + 0x294748, + 0x24da85, + 0x2a4205, + 0x2a4988, + 0x2a4b8a, + 0x5167c042, + 0x51a0bf02, + 0x2d24c5, + 0x2782c3, + 0x245308, + 0x204c03, + 0x2a4e04, + 0x268c0b, + 0x204c08, + 0x350b48, + 0x51f6b749, + 0x2a99c9, + 0x2aa106, + 0x2ab148, + 0x2ab349, + 0x2ac5c6, + 0x2ac745, + 0x24a1c6, + 0x2adc89, + 0x38e0c7, + 0x22b646, + 0x302b47, + 0x38e387, + 0x217b44, + 0x52311f49, + 0x281d08, + 0x3bc988, + 0x328207, + 0x2cb0c6, + 0x215409, + 0x2c6707, + 0x25594a, + 0x256ec8, + 0x212407, + 0x2131c6, + 0x28f2ca, + 0x3a0ac8, + 0x2c7d05, + 0x225e45, + 0x305d47, + 0x371e49, + 0x309e0b, + 0x32a708, + 0x38a9c9, + 0x250ec7, + 0x2b974c, + 0x2ba08c, + 0x2ba38a, + 0x2ba60c, + 0x2c5fc8, + 0x2c61c8, + 0x2c63c4, + 0x2c68c9, + 0x2c6b09, + 0x2c6d4a, + 0x2c6fc9, + 0x2c7307, + 0x3b4c4c, + 0x233706, + 0x2c9548, + 0x253546, + 0x387786, + 0x3a5b47, + 0x325008, + 0x3295cb, + 0x211b87, + 0x234489, + 0x379589, + 0x24ec07, + 0x257644, + 0x271887, + 0x39c2c6, + 0x2114c6, + 0x2f4c85, + 0x24b3c8, + 0x292044, + 0x292046, + 0x2b3fcb, + 0x32e2c9, + 0x214386, + 0x214609, + 0x20eb46, + 0x333088, + 0x221403, + 0x301685, + 0x2676c9, + 0x26df05, + 0x305b84, + 0x276bc6, + 0x23c405, + 0x255146, + 0x30ee47, + 0x32d506, + 0x22a24b, + 0x375007, + 0x2528c6, + 0x22ef46, + 0x22bcc6, + 0x29f1c9, + 0x2f638a, + 0x2bed45, + 0x20a38d, + 0x2a4c86, + 0x2f7946, + 0x2f4146, + 0x23e1c5, + 0x2e6fc7, + 0x2fe607, + 0x38f7ce, + 0x270203, + 0x2cb089, + 0x381c49, + 0x2e57c7, + 0x26c447, + 0x29cc45, + 0x372d05, + 0x527992cf, + 0x2d1987, + 0x2d1b48, + 0x2d2044, + 0x2d3806, + 0x52a47742, + 0x2d8e06, + 0x2dbd86, + 0x32990e, + 0x2e8eca, + 0x3ca246, + 0x3bd20a, + 0x2138c9, + 0x243445, + 0x373288, + 0x3504c6, + 0x29b008, + 0x275108, + 0x3a7b0b, + 0x223705, + 0x3090c8, + 0x2047cc, + 0x2def47, + 0x250306, + 0x3757c8, + 0x211848, + 0x52e3bbc2, + 0x20784b, + 0x218a89, + 0x219149, + 0x32e147, + 0x20a0c8, + 0x5321fac8, + 0x38b9cb, + 0x365549, + 0x283c0d, + 0x24a588, + 0x357388, + 0x53600ec2, + 0x33ec84, + 0x53a2c942, + 0x2fee46, + 0x53e04442, + 0x31984a, + 0x204246, + 0x31afc8, + 0x383408, + 0x39c6c6, + 0x23f2c6, + 0x2f94c6, + 0x372f85, + 0x238104, + 0x542b0104, + 0x350006, + 0x27b907, + 0x546e6687, + 0x26d5cb, + 0x211ec9, + 0x36f00a, + 0x38b2c4, + 0x2bdb48, + 0x22b40d, + 0x2f2009, + 0x2f2248, + 0x2f24c9, + 0x2f5544, + 0x2469c4, + 0x25ecc5, + 0x31034b, + 0x204b86, + 0x34fe45, + 0x21e009, + 0x223908, + 0x2a14c4, + 0x2e5549, + 0x353b85, + 0x2c2f88, + 0x3127c7, + 0x390248, + 0x282d86, + 0x38f407, + 0x2e0849, + 0x363cc9, + 0x229605, + 0x24f7c5, + 0x54a12482, + 0x2fa504, + 0x229045, + 0x223506, + 0x3c89c5, + 0x298387, + 0x350105, + 0x277704, + 0x324446, + 0x27ba07, + 0x22b286, + 0x3cf585, + 0x216e48, + 0x223e85, + 0x221287, + 0x2268c9, + 0x32e40a, + 0x2b0d87, + 0x2b0d8c, + 0x2e9586, + 0x379889, + 0x24d505, + 0x24d9c8, + 0x20cec3, + 0x20d945, + 0x39bf85, + 0x27f547, + 0x54e1f542, + 0x23eac7, + 0x2f51c6, + 0x33c486, + 0x2fb106, + 0x211786, + 0x353e08, + 0x238ac5, + 0x3550c7, + 0x3550cd, + 0x21c883, + 0x21c885, + 0x309987, + 0x23ee08, + 0x309545, + 0x216448, + 0x384546, + 0x2de287, + 0x2c9485, + 0x223786, + 0x3afd45, + 0x221d4a, + 0x377106, + 0x31d707, + 0x2e0c05, + 0x2fef87, + 0x303104, + 0x305b06, + 0x3731c5, + 0x20f8cb, + 0x39c149, + 0x25c28a, + 0x229688, + 0x30e308, + 0x313f0c, + 0x3235c7, + 0x34b288, + 0x34cbc8, + 0x34d005, + 0x3a440a, + 0x3b01c9, + 0x55201dc2, + 0x3cc786, + 0x2608c4, + 0x3230c9, + 0x2951c9, + 0x278dc7, + 0x2c5287, + 0x208589, + 0x23e3c8, + 0x23e3cf, + 0x228046, + 0x2ddd8b, + 0x3c1985, + 0x3c1987, + 0x2fcc89, + 0x268d46, + 0x2e54c7, + 0x2e1c05, + 0x22e4c4, + 0x267586, + 0x21f2c4, + 0x2eebc7, + 0x2cd048, + 0x55701408, + 0x302205, + 0x302347, + 0x3186c9, + 0x214944, + 0x242b88, + 0x55a55788, + 0x202e84, + 0x2faa48, + 0x389cc4, + 0x39e209, + 0x214105, + 0x55e022c2, + 0x228085, + 0x2d6b05, + 0x3955c8, + 0x232087, + 0x562008c2, + 0x2b0245, + 0x2d75c6, + 0x243a86, + 0x2fa4c8, + 0x2fbb88, + 0x3c8986, + 0x3b1ac6, + 0x301989, + 0x33c3c6, + 0x2930cb, + 0x36c945, + 0x2a6206, + 0x2f1108, + 0x290186, + 0x22b006, + 0x21690a, + 0x2a9d8a, + 0x25ff85, + 0x238b87, + 0x314c86, + 0x56604b02, + 0x309ac7, + 0x25f2c5, + 0x246f84, + 0x246f85, + 0x2bda46, + 0x271107, + 0x21b185, + 0x24b444, + 0x2aeac8, + 0x22b0c5, + 0x2e3b47, + 0x3b7ac5, + 0x221c85, + 0x2aa584, + 0x35a589, + 0x2f6b08, + 0x23c2c6, + 0x2d9a86, + 0x202b86, + 0x56bc0088, + 0x3c8707, + 0x306e0d, + 0x307d4c, + 0x308349, + 0x308589, + 0x56f6d342, + 0x3c7603, + 0x205ec3, + 0x39c385, + 0x3ab3ca, + 0x33c286, + 0x30cb85, + 0x30f004, + 0x30f00b, + 0x32fb8c, + 0x330f0c, + 0x331215, + 0x3320cd, + 0x33628f, + 0x336652, + 0x336acf, + 0x336e92, + 0x337313, + 0x3377cd, + 0x337d8d, + 0x33810e, + 0x33868e, + 0x338ecc, + 0x33928c, + 0x3396cb, + 0x33a5ce, + 0x33aed2, + 0x33c04c, + 0x33c610, + 0x345052, + 0x34624c, + 0x34690d, + 0x346c4c, + 0x349d51, + 0x34ac0d, + 0x34d5cd, + 0x34dbca, + 0x34de4c, + 0x34f10c, + 0x34fb4c, + 0x351e8c, + 0x355f53, + 0x3565d0, + 0x3569d0, + 0x35758d, + 0x357b8c, + 0x358949, + 0x35b78d, + 0x35bad3, + 0x35de91, + 0x35e2d3, + 0x35ee8f, + 0x35f24c, + 0x35f54f, + 0x35f90d, + 0x35ff0f, + 0x3602d0, + 0x360d4e, + 0x363f0e, + 0x364490, + 0x365fcd, + 0x36694e, + 0x366ccc, + 0x367c93, + 0x36980e, + 0x369e90, + 0x36a291, + 0x36a6cf, + 0x36aa93, + 0x36cecd, + 0x36d20f, + 0x36d5ce, + 0x36dc90, + 0x36e089, + 0x36f290, + 0x36f88f, + 0x36ff0f, + 0x3702d2, + 0x37100e, + 0x371a0d, + 0x37208d, + 0x3723cd, + 0x373e4d, + 0x37418d, + 0x3744d0, + 0x3748cb, + 0x375b4c, + 0x375ecc, + 0x3764cc, + 0x3767ce, + 0x383610, + 0x385b52, + 0x385fcb, + 0x3864ce, + 0x38684e, + 0x3870ce, + 0x38754b, + 0x5738ed16, + 0x395e0d, + 0x396294, + 0x39748d, + 0x398cd5, + 0x39a68d, + 0x39b00f, + 0x39b68f, + 0x39f70f, + 0x39face, + 0x3a004d, + 0x3a2111, + 0x3a530c, + 0x3a560c, + 0x3a590b, + 0x3a5d8c, + 0x3a614f, + 0x3a6512, + 0x3a724d, + 0x3a87cc, + 0x3a92cc, + 0x3a95cd, + 0x3a990f, + 0x3a9cce, + 0x3ab08c, + 0x3ab64d, + 0x3ab98b, + 0x3ac24c, + 0x3acb4d, + 0x3ace8e, + 0x3ad209, + 0x3ae613, + 0x3aeb4d, + 0x3aee8d, + 0x3af48c, + 0x3af90e, + 0x3b040f, + 0x3b07cc, + 0x3b0acd, + 0x3b0e0f, + 0x3b11cc, + 0x3b1e0c, + 0x3b218c, + 0x3b248c, + 0x3b2b4d, + 0x3b2e92, + 0x3b350c, + 0x3b380c, + 0x3b3b11, + 0x3b3f4f, + 0x3b430f, + 0x3b46d3, + 0x3b508e, + 0x3b540f, + 0x3b57cc, + 0x577b5b0e, + 0x3b5e8f, + 0x3b6256, + 0x3b75d2, + 0x3b988c, + 0x3ba24f, + 0x3ba8cd, + 0x3bf10f, + 0x3bf4cc, + 0x3bf7cd, + 0x3bfb0d, + 0x3c104e, + 0x3c208c, + 0x3c3a4c, + 0x3c3d50, + 0x3c6991, + 0x3c6dcb, + 0x3c720c, + 0x3c750e, + 0x3c9051, + 0x3c948e, + 0x3c980d, + 0x3ce0cb, + 0x3ce9cf, + 0x3cff94, + 0x259a02, + 0x259a02, + 0x2047c3, + 0x259a02, + 0x2047c3, + 0x259a02, + 0x206342, + 0x24a205, + 0x3c8d4c, + 0x259a02, + 0x259a02, + 0x206342, + 0x259a02, + 0x294dc5, + 0x32e405, + 0x259a02, + 0x259a02, + 0x20a842, + 0x294dc5, + 0x333289, + 0x35db8c, + 0x259a02, + 0x259a02, + 0x259a02, + 0x259a02, + 0x24a205, + 0x259a02, + 0x259a02, + 0x259a02, + 0x259a02, + 0x20a842, + 0x333289, + 0x259a02, + 0x259a02, + 0x259a02, + 0x32e405, + 0x259a02, + 0x32e405, + 0x35db8c, + 0x3c8d4c, + 0x38d8c3, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x21d283, + 0x259003, + 0x194a88, + 0x577c4, + 0x14b83, + 0xc8c88, + 0x2000c2, + 0x58608e02, + 0x2424c3, + 0x256e04, + 0x204a03, + 0x21cac4, + 0x22d806, + 0x210d83, + 0x300804, + 0x2d4f45, + 0x270203, + 0x21d283, + 0xe6143, + 0x259003, + 0x23628a, + 0x25c906, + 0x386bcc, + 0x120648, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x208e83, + 0x2dbd86, + 0x21d283, + 0x259003, + 0x216203, + 0xa5e48, + 0xf8945, + 0x1c5c49, + 0x5202, + 0x59afd885, + 0xf8945, + 0x67c47, + 0x6e808, + 0xbd0e, + 0x89792, + 0x18cecb, + 0x103746, + 0x59e8bac5, + 0x5a28bacc, + 0x38e07, + 0xeb207, + 0xbe30a, + 0x3c550, + 0x18e505, + 0xb19cb, + 0x886c8, + 0x32fc7, + 0x56a8b, + 0x7c349, + 0x132687, + 0x1118c7, + 0x77f87, + 0x32f06, + 0x667c8, + 0x5a835d46, + 0x49087, + 0x6d80d, + 0xbdcd0, + 0x5ac05642, + 0x7a548, + 0x5b2d0, + 0x17930c, + 0x5b3827cd, + 0x5df88, + 0x5e40b, + 0x6ba87, + 0x57389, + 0x59086, + 0x94948, + 0x1742, + 0x7154a, + 0xe9407, + 0x16bd07, + 0xa66c9, + 0xa80c8, + 0x111185, + 0xf360e, + 0x225ce, + 0x15440f, + 0x11eb09, + 0x6b109, + 0x87e8b, + 0x9fc0f, + 0xae04c, + 0x198a0b, + 0xcf508, + 0xebb87, + 0x100ac8, + 0x13cf8b, + 0x1406cc, + 0x15820c, + 0x16150c, + 0x16934d, + 0x1656c8, + 0xc7c82, + 0x1a6cc9, + 0xf3088, + 0x1a388b, + 0xcb2c6, + 0xd6f4b, + 0x13aa0b, + 0xe134a, + 0xe1f05, + 0xe6cd0, + 0xe9706, + 0x12dc06, + 0x129cc5, + 0x8f4c7, + 0xf99c8, + 0xee247, + 0xee507, + 0xfb647, + 0xfec8a, + 0x1204ca, + 0xc8346, + 0x9268d, + 0x49148, + 0x10b908, + 0xaa4c9, + 0xb8dc5, + 0xfe7cc, + 0x16954b, + 0x171d84, + 0x106a49, + 0x106c86, + 0x1557c6, + 0xbb946, + 0x3642, + 0x3f5c6, + 0x204b, + 0x113d87, + 0x1242, + 0xcd4c5, + 0x19fc4, + 0x101, + 0x64dc3, + 0x5a678486, + 0x94cc3, + 0x382, + 0x928c4, + 0x4342, + 0x879c4, + 0x882, + 0x7982, + 0xbc2, + 0x120f02, + 0xf82, + 0x8bac2, + 0x6a82, + 0x142382, + 0x34942, + 0x18982, + 0x8a42, + 0xa182, + 0x31103, + 0x942, + 0x2bc2, + 0x17242, + 0x1482, + 0x642, + 0x2f4c2, + 0x55402, + 0x7782, + 0xad02, + 0x5c2, + 0x17383, + 0xb02, + 0x2a02, + 0xb2d02, + 0x6f42, + 0x4c82, + 0xbb02, + 0x11602, + 0x9d082, + 0x2302, + 0x127182, + 0x6cac2, + 0x7c02, + 0x1d283, + 0x602, + 0x3bbc2, + 0x1c02, + 0x1202, + 0x29585, + 0x59c2, + 0x3eec2, + 0x40083, + 0x682, + 0x8782, + 0x1002, + 0x5502, + 0x11782, + 0x8c2, + 0x67c2, + 0x3642, + 0x7d85, + 0x5b606342, + 0x5bae3343, + 0xa203, + 0x5be06342, + 0xa203, + 0x82687, + 0x210443, + 0x2000c2, + 0x208e03, + 0x231103, + 0x3d0943, + 0x2005c3, + 0x208e83, + 0x21d283, + 0x214b83, + 0x259003, + 0x294d03, + 0x11243, + 0x120648, + 0x208e03, + 0x231103, + 0x3d0943, + 0x270203, + 0x21d283, + 0x214b83, + 0xe6143, + 0x259003, + 0x208e03, + 0x231103, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x200181, + 0x270203, + 0x21d283, + 0x24fc43, + 0x259003, + 0x195d04, + 0x38d8c3, + 0x208e03, + 0x231103, + 0x209843, + 0x3d0943, + 0x27f443, + 0x23c9c3, + 0x2a8bc3, + 0x284983, + 0x213ec3, + 0x220a04, + 0x21d283, + 0x259003, + 0x202043, + 0x38c384, + 0x25fe83, + 0x38c3, + 0x227303, + 0x38d308, + 0x28f304, + 0x20020a, + 0x23d346, + 0x117804, + 0x370707, + 0x21c44a, + 0x227f09, + 0x3aadc7, + 0x3b678a, + 0x38d8c3, + 0x2d254b, + 0x2bd549, + 0x202c85, + 0x33be87, + 0x8e02, + 0x208e03, + 0x21d4c7, + 0x375445, + 0x2c6549, + 0x231103, + 0x2bb386, + 0x2c5803, + 0xf5243, + 0x10e606, + 0x193e06, + 0x10d07, + 0x225046, + 0x31a4c5, + 0x207707, + 0x30c0c7, + 0x5e613ec3, + 0x346487, + 0x2bba43, + 0x245085, + 0x220a04, + 0x26eec8, + 0x37170c, + 0x340345, + 0x2a3686, + 0x21d387, + 0x225707, + 0x260087, + 0x266408, + 0x30d6cf, + 0x38ab85, + 0x2425c7, + 0x28c787, + 0x28c8ca, + 0x30abc9, + 0x30e745, + 0x3135ca, + 0x128546, + 0x2c5885, + 0x373d44, + 0x383346, + 0x2d2fc7, + 0x2c7b87, + 0x38fb08, + 0x221405, + 0x375346, + 0x26dd45, + 0x233b85, + 0x28b404, + 0x39c5c7, + 0x353c4a, + 0x24f108, + 0x36db06, + 0x8e83, + 0x2e4345, + 0x35d346, + 0x3b4e86, + 0x329bc6, + 0x270203, + 0x3a74c7, + 0x28c705, + 0x21d283, + 0x2e160d, + 0x214b83, + 0x38fc08, + 0x3812c4, + 0x2779c5, + 0x2a4e46, + 0x218806, + 0x2a6107, + 0x2a8c07, + 0x290605, + 0x259003, + 0x2ebc87, + 0x257509, + 0x32de49, + 0x20e64a, + 0x207d42, + 0x245044, + 0x2e9c44, + 0x329487, + 0x23e988, + 0x2ef349, + 0x21c749, + 0x2f1307, + 0x2f0846, + 0xf3386, + 0x2f5544, + 0x2f5b4a, + 0x2f8508, + 0x2f9389, + 0x30dec6, + 0x2b5c85, + 0x24efc8, + 0x2cb70a, + 0x294003, + 0x38c506, + 0x2f1407, + 0x22c005, + 0x381185, + 0x3add43, + 0x26fdc4, + 0x225e05, + 0x286087, + 0x2f6c45, + 0x342c46, + 0x1c8885, + 0x288e03, + 0x3ca309, + 0x27778c, + 0x3a500c, + 0x2d6d48, + 0x2c3907, + 0x3034c8, + 0x30428a, + 0x304c8b, + 0x2bd688, + 0x218908, + 0x233606, + 0x202a45, + 0x39344a, + 0x2e3385, + 0x2022c2, + 0x2c9347, + 0x250606, + 0x36e805, + 0x372949, + 0x279bc5, + 0x374d85, + 0x2f0e09, + 0x35d286, + 0x3b6e88, + 0x245143, + 0x225186, + 0x276b06, + 0x3133c5, + 0x3133c9, + 0x2efa89, + 0x27e5c7, + 0x115a84, + 0x315a87, + 0x21c649, + 0x233985, + 0x38208, + 0x33ef45, + 0x36b585, + 0x256289, + 0x201b42, + 0x2afb04, + 0x200f42, + 0x200b02, + 0x2d3945, + 0x317588, + 0x2b8d05, + 0x2c74c3, + 0x2c74c5, + 0x2d9003, + 0x20cb02, + 0x37a2c4, + 0x2a7903, + 0x20a282, + 0x387d84, + 0x2ea1c3, + 0x202582, + 0x2b8d83, + 0x28e444, + 0x2f9943, + 0x25dc04, + 0x203c42, + 0x216103, + 0x2343c3, + 0x200b42, + 0x374402, + 0x2ef8c9, + 0x2076c2, + 0x28a504, + 0x200cc2, + 0x24ee44, + 0x2f0804, + 0x202504, + 0x203642, + 0x233242, + 0x22e883, + 0x304a43, + 0x248204, + 0x293e84, + 0x2f1584, + 0x2f86c4, + 0x30f743, + 0x324bc3, + 0x3284c4, + 0x316f44, + 0x317086, + 0x229542, + 0x3bb43, + 0x208e02, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x2000c2, + 0x38d8c3, + 0x208e03, + 0x231103, + 0x202bc3, + 0x213ec3, + 0x220a04, + 0x2efb84, + 0x2bf144, + 0x21d283, + 0x259003, + 0x216203, + 0x2f6604, + 0x32b543, + 0x2a7143, + 0x371cc4, + 0x33ed46, + 0x207083, + 0xf8945, + 0xeb207, + 0x35a8c3, + 0x5fa1e348, + 0x228a83, + 0x2b4883, + 0x2450c3, + 0x208e83, + 0x39e4c5, + 0x13be43, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x20e2c3, + 0x22ce43, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x217383, + 0x21d283, + 0x231d04, + 0xe6143, + 0x259003, + 0x351584, + 0xf8945, + 0x2c2345, + 0xeb207, + 0x208e02, + 0x202142, + 0x200382, + 0x201402, + 0x14b83, + 0x2003c2, + 0x191584, + 0x208e03, + 0x233e84, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x2bf144, + 0x21d283, + 0x14b83, + 0x259003, + 0x211303, + 0x2879c4, + 0x120648, + 0x208e03, + 0x214b83, + 0x11243, + 0x144c04, + 0x245a04, + 0x120648, + 0x208e03, + 0x24ea84, + 0x220a04, + 0x214b83, + 0x200ec2, + 0xe6143, + 0x259003, + 0x202203, + 0x6fdc4, + 0x2fc845, + 0x2022c2, + 0x201fc3, + 0x192e49, + 0xdfd46, + 0x128688, + 0x2000c2, + 0x120648, + 0x208e02, + 0x231103, + 0x213ec3, + 0x2005c2, + 0x14b83, + 0x259003, + 0xa882, + 0x2000c2, + 0x1b6947, + 0x107609, + 0x37c3, + 0x120648, + 0x193d83, + 0x63347147, + 0x8e03, + 0x1cc688, + 0x231103, + 0x213ec3, + 0x42346, + 0x217383, + 0x58dc8, + 0xc4e48, + 0x38886, + 0x270203, + 0xce908, + 0x98403, + 0x634e3806, + 0xe8085, + 0x31307, + 0x1d283, + 0x4a83, + 0x59003, + 0x2542, + 0x18434a, + 0x17083, + 0xd3a03, + 0x2fbf44, + 0x10d4cb, + 0x10da88, + 0x8eb02, + 0x1454f07, + 0x15276c7, + 0x14c7588, + 0x1516b83, + 0x1882cb, + 0x12d147, + 0x2000c2, + 0x208e02, + 0x208e03, + 0x231103, + 0x2dd584, + 0x213ec3, + 0x217383, + 0x270203, + 0x21d283, + 0x208e03, + 0x231103, + 0x213ec3, + 0x208e83, + 0x21d283, + 0x259003, + 0x281ac3, + 0x211303, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x11243, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x208e83, + 0x21d283, + 0x259003, + 0x22a342, + 0x2000c1, + 0x2000c2, + 0x200201, + 0x336382, + 0x120648, + 0x21af05, + 0x200101, + 0x8e03, + 0x200d81, + 0x200501, + 0x201481, + 0x24a182, + 0x37ef44, + 0x24a183, + 0x200041, + 0x200801, + 0x200181, + 0x200701, + 0x2f87c7, + 0x342d8f, + 0x3024c6, + 0x2004c1, + 0x344c46, + 0x200d01, + 0x200581, + 0x3ce30e, + 0x2003c1, + 0x259003, + 0x201b81, + 0x243685, + 0x202542, + 0x3adc45, + 0x200401, + 0x200741, + 0x2007c1, + 0x2022c2, + 0x200081, + 0x201fc1, + 0x20a781, + 0x202cc1, + 0x201241, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x216cc3, + 0x208e03, + 0x213ec3, + 0x8ea48, + 0x270203, + 0x21d283, + 0x89a43, + 0x259003, + 0x14ebf48, + 0xe1c8, + 0xf8945, + 0x120648, + 0x14b83, + 0xf8945, + 0x135884, + 0x47904, + 0x14ebf4a, + 0x120648, + 0xe6143, + 0x208e03, + 0x231103, + 0x213ec3, + 0x21d283, + 0x259003, + 0x2038c3, + 0x120648, + 0x208e03, + 0x231103, + 0x2dd584, + 0x259003, + 0x27ee05, + 0x35c204, + 0x208e03, + 0x21d283, + 0x259003, + 0xa5f4a, + 0x15aa8a, + 0x115e04, + 0x11c906, + 0x208e02, + 0x208e03, + 0x22e409, + 0x231103, + 0x21acc9, + 0x213ec3, + 0x270203, + 0x21d283, + 0x7b844, + 0x14b83, + 0x259003, + 0x2f5348, + 0x23e087, + 0x2fc845, + 0x1c7008, + 0x1b6947, + 0x3ec0a, + 0x10930b, + 0x144e87, + 0x41188, + 0x111d8a, + 0x12c88, + 0x107609, + 0x25cc7, + 0x153a47, + 0x1270c8, + 0x1cc688, + 0x4288f, + 0x1541c5, + 0x1cc987, + 0x42346, + 0x4c1c7, + 0x10e886, + 0x58dc8, + 0x9bec6, + 0x15d087, + 0x126089, + 0x1b7347, + 0x105649, + 0xb92c9, + 0xc20c6, + 0xc4e48, + 0xc30c5, + 0x7beca, + 0xce908, + 0x98403, + 0xd9708, + 0x31307, + 0x74845, + 0x61390, + 0x4a83, + 0xe6143, + 0x125f07, + 0x27745, + 0xee808, + 0x69445, + 0xd3a03, + 0x1688, + 0x56146, + 0x184e49, + 0xab547, + 0x19310b, + 0x142844, + 0x1063c4, + 0x10d4cb, + 0x10da88, + 0x10e507, + 0xf8945, + 0x208e03, + 0x231103, + 0x3d0943, + 0x259003, + 0x202103, + 0x213ec3, + 0xe6143, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x87fcb, + 0x2000c2, + 0x208e02, + 0x259003, + 0x120648, + 0x2000c2, + 0x208e02, + 0x200382, + 0x2005c2, + 0x200d02, + 0x21d283, + 0x2003c2, + 0x2000c2, + 0x38d8c3, + 0x208e02, + 0x208e03, + 0x231103, + 0x200382, + 0x213ec3, + 0x217383, + 0x270203, + 0x2bf144, + 0x21d283, + 0x213443, + 0x14b83, + 0x259003, + 0x2fbf44, + 0x202043, + 0x213ec3, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x214b83, + 0x259003, + 0x3b9d47, + 0x208e03, + 0x20cd87, + 0x300c46, + 0x20ce43, + 0x217243, + 0x213ec3, + 0x203583, + 0x220a04, + 0x39cbc4, + 0x2eab46, + 0x20dd03, + 0x21d283, + 0x259003, + 0x27ee05, + 0x2ad7c4, + 0x2be803, + 0x20b103, + 0x2c9347, + 0x312745, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x259003, + 0x4df07, + 0x8f4c7, + 0x1a3e85, + 0x205042, + 0x249283, + 0x214a43, + 0x38d8c3, + 0x6c608e03, + 0x20f982, + 0x231103, + 0x204a03, + 0x213ec3, + 0x220a04, + 0x307c83, + 0x38ab83, + 0x270203, + 0x2bf144, + 0x6ca08dc2, + 0x21d283, + 0x259003, + 0x22f583, + 0x219fc3, + 0x22a342, + 0x202043, + 0x120648, + 0x213ec3, + 0x11243, + 0x32be04, + 0x38d8c3, + 0x208e02, + 0x208e03, + 0x233e84, + 0x231103, + 0x213ec3, + 0x220a04, + 0x217383, + 0x30ffc4, + 0x20ac44, + 0x2dbd86, + 0x2bf144, + 0x21d283, + 0x259003, + 0x216203, + 0x250606, + 0x3978b, + 0x35d46, + 0x11f20a, + 0x11434a, + 0x120648, + 0x26dd04, + 0x6de08e03, + 0x38d884, + 0x231103, + 0x257384, + 0x213ec3, + 0x2bd9c3, + 0x270203, + 0x21d283, + 0xe6143, + 0x259003, + 0xca143, + 0x3433cb, + 0x3bfe4a, + 0x3d13cc, + 0xe40c8, + 0x2000c2, + 0x208e02, + 0x200382, + 0x22bec5, + 0x220a04, + 0x202302, + 0x270203, + 0x20ac44, + 0x201402, + 0x2003c2, + 0x211302, + 0x22a342, + 0x18d8c3, + 0x48182, + 0x2b6789, + 0x322748, + 0x213d49, + 0x217989, + 0x22420a, + 0x3184ca, + 0x208002, + 0x342382, + 0x8e02, + 0x208e03, + 0x22a502, + 0x242786, + 0x36fd82, + 0x213ac2, + 0x30968e, + 0x21614e, + 0x280507, + 0x21d207, + 0x282742, + 0x231103, + 0x213ec3, + 0x22aa02, + 0x2005c2, + 0x171c3, + 0x23408f, + 0x242ac2, + 0x2ae947, + 0x32a8c7, + 0x2b57c7, + 0x2e2e0c, + 0x2eadcc, + 0x226304, + 0x25eb0a, + 0x216082, + 0x206f42, + 0x2bac44, + 0x200702, + 0x214502, + 0x2eb004, + 0x213542, + 0x204c82, + 0x21303, + 0x29bf47, + 0x23d0c5, + 0x211602, + 0x24c144, + 0x327182, + 0x2e3988, + 0x21d283, + 0x2050c8, + 0x204882, + 0x2264c5, + 0x399b86, + 0x259003, + 0x2059c2, + 0x2ef587, + 0x2542, + 0x25cb85, + 0x210a45, + 0x209cc2, + 0x20ed02, + 0x2bc84a, + 0x29048a, + 0x2701c2, + 0x29a684, + 0x203a42, + 0x244f08, + 0x210602, + 0x365a88, + 0x30a0c7, + 0x30a689, + 0x205082, + 0x30edc5, + 0x206045, + 0x2214cb, + 0x2cc38c, + 0x22a008, + 0x32cf88, + 0x229542, + 0x2a61c2, + 0x2000c2, + 0x120648, + 0x208e02, + 0x208e03, + 0x200382, + 0x201402, + 0x14b83, + 0x2003c2, + 0x259003, + 0x211302, + 0x2000c2, + 0xf8945, + 0x6f208e02, + 0x6f613ec3, + 0x221303, + 0x202302, + 0x21d283, + 0x3d2403, + 0x6fa59003, + 0x2ec903, + 0x282786, + 0x1611303, + 0xf8945, + 0x187e4b, + 0x120648, + 0x6ff87, + 0x6e607, + 0x129cc5, + 0xa87cd, + 0xa6b8a, + 0x902c7, + 0x2a784, + 0x2a7c3, + 0xbb9c4, + 0x70203c82, + 0x70604342, + 0x70a02842, + 0x70e00fc2, + 0x7120aac2, + 0x71600f82, + 0xeb207, + 0x71a08e02, + 0x71e02282, + 0x7221de02, + 0x72608a42, + 0x216143, + 0x26744, + 0x22ea43, + 0x72a11482, + 0x5df88, + 0x72e06502, + 0x4fc87, + 0x73200042, + 0x73603482, + 0x73a00182, + 0x73e00d42, + 0x7420ad02, + 0x746005c2, + 0x13db85, + 0x223e43, + 0x311844, + 0x74a00702, + 0x74e14882, + 0x75200e42, + 0xb074b, + 0x75602d82, + 0x75e4eb42, + 0x76202302, + 0x76600d02, + 0x76a26942, + 0x76e01942, + 0x77202382, + 0x7766cac2, + 0x77a08dc2, + 0x77e035c2, + 0x78201402, + 0x78604fc2, + 0x78a08342, + 0x78e14002, + 0xe5f84, + 0x33e783, + 0x79221882, + 0x79615d82, + 0x79a100c2, + 0x79e006c2, + 0x7a2003c2, + 0x7a60a282, + 0x88147, + 0x7aa03b02, + 0x7ae02e02, + 0x7b211302, + 0x7b616102, + 0xfe7cc, + 0x7ba0fec2, + 0x7be1ea02, + 0x7c203382, + 0x7c604b02, + 0x7ca05ec2, + 0x7ce0fcc2, + 0x7d206582, + 0x7d610dc2, + 0x7da76e82, + 0x7de77402, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x75b07c83, + 0x2230c3, + 0x39e544, + 0x322646, + 0x2fa443, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x3b8109, + 0x248182, + 0x3a2e83, + 0x2b95c3, + 0x395545, + 0x204a03, + 0x307c83, + 0x2230c3, + 0x2a4003, + 0x238043, + 0x3c5649, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x307c83, + 0x2230c3, + 0x248182, + 0x248182, + 0x307c83, + 0x2230c3, + 0x7e608e03, + 0x231103, + 0x217bc3, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x120648, + 0x208e02, + 0x208e03, + 0x21d283, + 0x259003, + 0x208e03, + 0x231103, + 0x213ec3, + 0x270203, + 0x21d283, + 0x14b83, + 0x259003, + 0x245a04, + 0x208e02, + 0x208e03, + 0x392503, + 0x231103, + 0x24ea84, + 0x3d0943, + 0x213ec3, + 0x220a04, + 0x217383, + 0x270203, + 0x21d283, + 0x259003, + 0x202203, + 0x2fc845, + 0x238043, + 0x202043, + 0x14b83, + 0x208e02, + 0x208e03, + 0x307c83, + 0x21d283, + 0x259003, + 0x2000c2, + 0x38d8c3, + 0x120648, + 0x208e03, + 0x231103, + 0x213ec3, + 0x22d806, + 0x220a04, + 0x217383, + 0x2bf144, + 0x21d283, + 0x259003, + 0x216203, + 0x208e03, + 0x231103, + 0x21d283, + 0x259003, + 0x145c187, + 0x7d87, + 0x208e03, + 0x35d46, + 0x231103, + 0x213ec3, + 0xe4dc6, + 0x21d283, + 0x259003, + 0x32a0c8, + 0x32cdc9, + 0x33d349, + 0x344508, + 0x39bd08, + 0x39bd09, + 0x31de8a, + 0x3586ca, + 0x3977ca, + 0x39d78a, + 0x3bfe4a, + 0x3cb70b, + 0x2466cd, + 0x361b0f, + 0x272b90, + 0x35b30d, + 0x3761cc, + 0x39d4cb, + 0x6e808, + 0xfa948, + 0x100ec5, + 0xcd4c5, + 0x2000c2, + 0x312585, + 0x206083, + 0x81e08e02, + 0x231103, + 0x213ec3, + 0x37f707, + 0x2450c3, + 0x270203, + 0x21d283, + 0x24fc43, + 0x2090c3, + 0x214b83, + 0x259003, + 0x25c906, + 0x2022c2, + 0x202043, + 0x120648, + 0x2000c2, + 0x38d8c3, + 0x208e02, + 0x208e03, + 0x231103, + 0x213ec3, + 0x220a04, + 0x270203, + 0x21d283, + 0x259003, + 0x211303, + 0x18fb04, + 0x14f8106, + 0x2000c2, + 0x208e02, + 0x213ec3, + 0x270203, + 0x259003, } // children is the list of nodes' children, the parent's wildcard bit and the @@ -9030,505 +9219,527 @@ var children = [...]uint32{ 0x40000000, 0x50000000, 0x60000000, - 0x185460f, - 0x1858615, - 0x187c616, - 0x19d861f, - 0x19ec676, - 0x1a0067b, - 0x1a14680, - 0x1a34685, - 0x1a3868d, - 0x1a5068e, - 0x1a78694, - 0x1a7c69e, - 0x1a9469f, + 0x1864613, + 0x1868619, + 0x186c61a, + 0x189061b, + 0x19ec624, + 0x1a0467b, + 0x1a18681, + 0x1a2c686, + 0x1a4c68b, + 0x1a50693, + 0x1a68694, + 0x1a6c69a, + 0x1a9469b, 0x1a986a5, - 0x1a9c6a6, - 0x1ad86a7, - 0x1adc6b6, - 0x21ae46b7, - 0x1b2c6b9, - 0x1b306cb, - 0x1b506cc, - 0x1b646d4, - 0x1b686d9, - 0x1b986da, - 0x1bb46e6, - 0x1bdc6ed, - 0x1bec6f7, - 0x1bf06fb, - 0x1c886fc, - 0x1c9c722, - 0x1cb0727, - 0x1ce072c, - 0x1cf0738, - 0x1d0473c, - 0x1da8741, - 0x1fa076a, - 0x1fa47e8, - 0x20107e9, - 0x207c804, - 0x209481f, - 0x20a8825, - 0x20b082a, - 0x20c482c, - 0x20c8831, - 0x20e4832, - 0x2134839, - 0x215084d, - 0x2154854, - 0x2158855, + 0x1ab06a6, + 0x1ab46ac, + 0x1ab86ad, + 0x1af46ae, + 0x1af86bd, + 0x61b006be, + 0x21b086c0, + 0x1b506c2, + 0x1b546d4, + 0x1b746d5, + 0x1b886dd, + 0x1b8c6e2, + 0x1bbc6e3, + 0x1bd86ef, + 0x1c006f6, + 0x1c10700, + 0x1c14704, + 0x1cac705, + 0x1cc072b, + 0x1cd4730, + 0x1d04735, + 0x1d14741, + 0x1d28745, + 0x1dcc74a, + 0x1fc8773, + 0x1fcc7f2, + 0x20387f3, + 0x20a480e, + 0x20bc829, + 0x20d082f, + 0x20d8834, + 0x20ec836, + 0x20f083b, + 0x210c83c, + 0x2158843, 0x2174856, - 0x21b085d, - 0x621b486c, - 0x21cc86d, - 0x21e0873, - 0x21e4878, - 0x21f4879, - 0x22a487d, - 0x22a88a9, - 0x222b88aa, - 0x222bc8ae, - 0x222c08af, - 0x22f88b0, - 0x22fc8be, - 0x278c8bf, - 0x228349e3, - 0x22838a0d, - 0x2283ca0e, - 0x22848a0f, - 0x2284ca12, - 0x22858a13, - 0x2285ca16, - 0x22860a17, - 0x22864a18, - 0x22868a19, - 0x2286ca1a, - 0x22878a1b, - 0x2287ca1e, - 0x22888a1f, - 0x2288ca22, - 0x22890a23, - 0x22894a24, - 0x228a0a25, + 0x217885d, + 0x217c85e, + 0x21a085f, + 0x21dc868, + 0x621e0877, + 0x21f8878, + 0x221087e, + 0x2218884, + 0x2228886, + 0x22d888a, + 0x22dc8b6, + 0x222ec8b7, + 0x222f08bb, + 0x222f48bc, + 0x23388bd, + 0x233c8ce, + 0x27f88cf, + 0x228a09fe, 0x228a4a28, - 0x228b0a29, - 0x228b4a2c, + 0x228a8a29, + 0x228b4a2a, 0x228b8a2d, - 0x228bca2e, - 0x28c0a2f, - 0x228c4a30, - 0x228d0a31, + 0x228c4a2e, + 0x228c8a31, + 0x228cca32, + 0x228d0a33, 0x228d4a34, - 0x28dca35, - 0x291ca37, - 0x2293ca47, + 0x228d8a35, + 0x228e4a36, + 0x228e8a39, + 0x228f4a3a, + 0x228f8a3d, + 0x228fca3e, + 0x22900a3f, + 0x2290ca40, + 0x22910a43, + 0x2291ca44, + 0x22920a47, + 0x22924a48, + 0x22928a49, + 0x292ca4a, + 0x22930a4b, + 0x2293ca4c, 0x22940a4f, - 0x22944a50, - 0x2948a51, - 0x2294ca52, - 0x2950a53, - 0x296ca54, - 0x2984a5b, - 0x2988a61, - 0x2998a62, - 0x29a4a66, - 0x29d8a69, - 0x29dca76, - 0x29f0a77, - 0x229f8a7c, - 0x2ab8a7e, - 0x22abcaae, - 0x2ac4aaf, - 0x2ac8ab1, - 0x2ae0ab2, - 0x2af4ab8, - 0x2b1cabd, - 0x2b3cac7, - 0x2b6cacf, - 0x2b94adb, - 0x2b98ae5, - 0x2bbcae6, - 0x2bc0aef, - 0x2bd4af0, - 0x2bd8af5, - 0x2bdcaf6, - 0x2bfcaf7, - 0x2c1caff, - 0x2c20b07, - 0x22c24b08, - 0x2c28b09, - 0x2c2cb0a, - 0x2c3cb0b, - 0x2c40b0f, - 0x2cb8b10, - 0x2cbcb2e, - 0x2cd8b2f, + 0x2948a50, + 0x298ca52, + 0x229aca63, + 0x229b0a6b, + 0x229b4a6c, + 0x229b8a6d, + 0x29bca6e, + 0x229c0a6f, + 0x29c8a70, + 0x29cca72, + 0x29d0a73, + 0x29eca74, + 0x2a04a7b, + 0x2a08a81, + 0x2a18a82, + 0x2a24a86, + 0x2a58a89, + 0x2a5ca96, + 0x2a74a97, + 0x22a7ca9d, + 0x22a80a9f, + 0x22a88aa0, + 0x2b60aa2, + 0x22b64ad8, + 0x2b6cad9, + 0x2b70adb, + 0x22b74adc, + 0x2b78add, + 0x2b90ade, + 0x2ba4ae4, + 0x2bccae9, + 0x2becaf3, + 0x2c1cafb, + 0x2c44b07, + 0x2c48b11, + 0x2c6cb12, + 0x2c70b1b, + 0x2c84b1c, + 0x2c88b21, + 0x2c8cb22, + 0x2cacb23, + 0x2cc8b2b, + 0x2cccb32, + 0x22cd0b33, + 0x2cd4b34, + 0x2cd8b35, 0x2ce8b36, - 0x2cfcb3a, - 0x2d14b3f, - 0x2d2cb45, - 0x2d44b4b, - 0x2d48b51, - 0x2d60b52, - 0x2d7cb58, - 0x2d9cb5f, - 0x2db4b67, - 0x2e14b6d, + 0x2cecb3a, + 0x2d64b3b, + 0x2d68b59, + 0x2d6cb5a, + 0x2d8cb5b, + 0x2d9cb63, + 0x2db0b67, + 0x2dc8b6c, + 0x2de0b72, + 0x2df8b78, + 0x2dfcb7e, + 0x2e14b7f, 0x2e30b85, - 0x2e38b8c, - 0x2e3cb8e, - 0x2e50b8f, - 0x2e94b94, - 0x2f14ba5, - 0x2f40bc5, - 0x2f44bd0, - 0x2f4cbd1, - 0x2f6cbd3, - 0x2f70bdb, - 0x2f94bdc, - 0x2f9cbe5, - 0x2fd8be7, - 0x301cbf6, - 0x3020c07, - 0x3094c08, - 0x3098c25, - 0x2309cc26, - 0x230a0c27, - 0x230a4c28, - 0x230b4c29, - 0x230b8c2d, - 0x230bcc2e, - 0x230c0c2f, - 0x230c4c30, - 0x30dcc31, - 0x3100c37, - 0x3120c40, - 0x36e4c48, - 0x36f0db9, - 0x3710dbc, - 0x38ccdc4, - 0x399ce33, - 0x3a0ce67, - 0x3a64e83, - 0x3b4ce99, - 0x3ba4ed3, - 0x3be0ee9, - 0x3cdcef8, - 0x3da8f37, - 0x3e40f6a, - 0x3ed0f90, - 0x3f34fb4, - 0x416cfcd, - 0x422505b, - 0x42f1089, - 0x433d0bc, - 0x43c50cf, - 0x44010f1, - 0x4451100, - 0x44c9114, - 0x644cd132, - 0x644d1133, - 0x644d5134, - 0x4551135, - 0x45ad154, - 0x462916b, - 0x46a118a, - 0x47211a8, - 0x478d1c8, - 0x48b91e3, - 0x491122e, - 0x64915244, - 0x49ad245, - 0x4a3526b, - 0x4a8128d, - 0x4ae92a0, - 0x4b912ba, - 0x4c592e4, - 0x4cc1316, - 0x4dd5330, - 0x64dd9375, - 0x64ddd376, - 0x4e39377, - 0x4e9538e, - 0x4f253a5, - 0x4fa13c9, - 0x4fe53e8, - 0x50c93f9, - 0x50fd432, - 0x515d43f, - 0x51d1457, - 0x5259474, - 0x5299496, - 0x53094a6, - 0x6530d4c2, - 0x53314c3, - 0x53354cc, - 0x534d4cd, - 0x53694d3, - 0x53ad4da, - 0x53bd4eb, - 0x53d54ef, - 0x544d4f5, - 0x5455513, - 0x5469515, - 0x548551a, - 0x54b1521, - 0x54b552c, - 0x54bd52d, - 0x54d152f, - 0x54ed534, - 0x54f953b, - 0x550153e, - 0x553d540, - 0x555154f, - 0x5559554, - 0x5565556, - 0x556d559, - 0x559155b, - 0x55b5564, - 0x55cd56d, - 0x55d1573, - 0x55d9574, - 0x55dd576, - 0x5645577, - 0x5649591, - 0x566d592, - 0x569159b, - 0x56ad5a4, - 0x56bd5ab, - 0x56d15af, - 0x56d55b4, - 0x56dd5b5, - 0x56f15b7, - 0x57015bc, - 0x57055c0, - 0x57215c1, - 0x5fb15c8, - 0x5fe97ec, - 0x60157fa, - 0x6031805, - 0x605180c, - 0x6071814, - 0x60b581c, - 0x60bd82d, - 0x260c182f, - 0x260c5830, - 0x60cd831, - 0x6245833, - 0x26249891, - 0x26259892, - 0x26261896, - 0x2626d898, - 0x627189b, - 0x627589c, - 0x629d89d, - 0x62c58a7, - 0x62c98b1, - 0x63018b2, - 0x63218c0, - 0x6e798c8, - 0x6e7db9e, - 0x6e81b9f, - 0x26e85ba0, - 0x6e89ba1, - 0x26e8dba2, - 0x6e91ba3, - 0x26e9dba4, - 0x6ea1ba7, - 0x6ea5ba8, - 0x26ea9ba9, - 0x6eadbaa, - 0x26eb5bab, - 0x6eb9bad, - 0x6ebdbae, - 0x26ecdbaf, - 0x6ed1bb3, - 0x6ed5bb4, - 0x6ed9bb5, - 0x6eddbb6, - 0x26ee1bb7, - 0x6ee5bb8, - 0x6ee9bb9, - 0x6eedbba, - 0x6ef1bbb, - 0x26ef9bbc, - 0x6efdbbe, - 0x6f01bbf, - 0x6f05bc0, - 0x26f09bc1, - 0x6f0dbc2, - 0x26f15bc3, - 0x26f19bc5, - 0x6f35bc6, - 0x6f45bcd, - 0x6f89bd1, - 0x6f8dbe2, - 0x6fb1be3, - 0x6fb5bec, - 0x6fb9bed, - 0x7145bee, - 0x27149c51, - 0x27151c52, - 0x27155c54, - 0x27159c55, - 0x7161c56, - 0x723dc58, - 0x27249c8f, - 0x2724dc92, - 0x27251c93, - 0x27255c94, - 0x7259c95, - 0x7285c96, - 0x7289ca1, - 0x72adca2, - 0x72b9cab, - 0x72d9cae, - 0x72ddcb6, - 0x7315cb7, - 0x75adcc5, - 0x7669d6b, - 0x767dd9a, - 0x76b1d9f, - 0x76e1dac, - 0x76fddb8, - 0x7725dbf, - 0x7745dc9, - 0x7761dd1, - 0x7789dd8, - 0x7799de2, - 0x779dde6, - 0x77a1de7, - 0x77d5de8, - 0x77e1df5, - 0x7801df8, - 0x7879e00, - 0x2787de1e, - 0x78a1e1f, - 0x78c1e28, - 0x78d5e30, - 0x78e9e35, - 0x78ede3a, - 0x790de3b, - 0x79b1e43, - 0x79cde6c, - 0x79f1e73, - 0x79f9e7c, - 0x7a05e7e, - 0x7a0de81, - 0x7a21e83, - 0x7a41e88, - 0x7a4de90, - 0x7a59e93, - 0x7a89e96, - 0x7b5dea2, - 0x7b61ed7, - 0x7b75ed8, - 0x7b7dedd, - 0x7b95edf, - 0x7b99ee5, - 0x7ba5ee6, - 0x7ba9ee9, - 0x7bc5eea, - 0x7c01ef1, - 0x7c05f00, - 0x7c25f01, - 0x7c75f09, - 0x7c91f1d, - 0x7ce5f24, - 0x7ce9f39, - 0x7cedf3a, - 0x7cf1f3b, - 0x7d35f3c, - 0x7d45f4d, - 0x7d85f51, - 0x7d89f61, - 0x7db9f62, - 0x7f01f6e, - 0x7f29fc0, - 0x7f55fca, - 0x7f65fd5, - 0x7f6dfd9, - 0x807dfdb, - 0x808a01f, - 0x8096022, - 0x80a2025, - 0x80ae028, - 0x80ba02b, - 0x80c602e, - 0x80d2031, - 0x80de034, - 0x80ea037, - 0x80f603a, - 0x810203d, - 0x810e040, - 0x811a043, - 0x8122046, - 0x812e048, - 0x813a04b, - 0x814604e, - 0x8152051, - 0x815e054, - 0x816a057, - 0x817605a, - 0x818205d, - 0x818e060, - 0x819a063, - 0x81a6066, - 0x81d2069, - 0x81de074, - 0x81ea077, - 0x81f607a, - 0x820207d, - 0x820e080, - 0x8216083, - 0x8222085, - 0x822e088, - 0x823a08b, - 0x824608e, - 0x8252091, - 0x825e094, - 0x826a097, - 0x827609a, - 0x828209d, - 0x828e0a0, - 0x829a0a3, - 0x82a60a6, - 0x82b20a9, - 0x82ba0ac, - 0x82c60ae, - 0x82d20b1, - 0x82de0b4, - 0x82ea0b7, - 0x82f60ba, - 0x83020bd, - 0x830e0c0, - 0x831a0c3, - 0x831e0c6, + 0x2e50b8c, + 0x2e70b94, + 0x2e8cb9c, + 0x2eecba3, + 0x2f08bbb, + 0x2f18bc2, + 0x2f1cbc6, + 0x2f30bc7, + 0x2f74bcc, + 0x2ff4bdd, + 0x3024bfd, + 0x3028c09, + 0x3034c0a, + 0x3054c0d, + 0x3058c15, + 0x307cc16, + 0x3084c1f, + 0x30c0c21, + 0x3110c30, + 0x3114c44, + 0x319cc45, + 0x31a0c67, + 0x231a4c68, + 0x231a8c69, + 0x231acc6a, + 0x231bcc6b, + 0x231c0c6f, + 0x231c4c70, + 0x231c8c71, + 0x231ccc72, + 0x31e4c73, + 0x3208c79, + 0x3228c82, + 0x3890c8a, + 0x389ce24, + 0x38bce27, + 0x3a78e2f, + 0x3b48e9e, + 0x3bb8ed2, + 0x3c10eee, + 0x3cf8f04, + 0x3d50f3e, + 0x3d8cf54, + 0x3e88f63, + 0x3f54fa2, + 0x3fecfd5, + 0x407cffb, + 0x40e101f, + 0x4319038, + 0x43d10c6, + 0x449d0f4, + 0x44e9127, + 0x457113a, + 0x45ad15c, + 0x45fd16b, + 0x467517f, + 0x6467919d, + 0x6467d19e, + 0x6468119f, + 0x46fd1a0, + 0x47591bf, + 0x47d51d6, + 0x484d1f5, + 0x48cd213, + 0x4939233, + 0x4a6524e, + 0x4abd299, + 0x64ac12af, + 0x4b592b0, + 0x4be12d6, + 0x4c2d2f8, + 0x4c9530b, + 0x4d3d325, + 0x4e0534f, + 0x4e6d381, + 0x4f8139b, + 0x64f853e0, + 0x64f893e1, + 0x4fe53e2, + 0x50413f9, + 0x50d1410, + 0x514d434, + 0x5191453, + 0x5275464, + 0x52a949d, + 0x53094aa, + 0x537d4c2, + 0x54054df, + 0x5445501, + 0x54b5511, + 0x654b952d, + 0x54e152e, + 0x54e5538, + 0x54fd539, + 0x551953f, + 0x555d546, + 0x556d557, + 0x558555b, + 0x55fd561, + 0x560557f, + 0x5619581, + 0x5635586, + 0x566158d, + 0x5665598, + 0x566d599, + 0x568159b, + 0x56a15a0, + 0x56ad5a8, + 0x56b55ab, + 0x56f15ad, + 0x57055bc, + 0x570d5c1, + 0x57195c3, + 0x57215c6, + 0x57455c8, + 0x57695d1, + 0x57815da, + 0x57855e0, + 0x578d5e1, + 0x57915e3, + 0x580d5e4, + 0x5811603, + 0x5815604, + 0x5839605, + 0x585d60e, + 0x5879617, + 0x588d61e, + 0x58a1623, + 0x58a9628, + 0x58b162a, + 0x58c562c, + 0x58d5631, + 0x58d9635, + 0x58f5636, + 0x618563d, + 0x61bd861, + 0x61e986f, + 0x620587a, + 0x6225881, + 0x6245889, + 0x6289891, + 0x62918a2, + 0x262958a4, + 0x262998a5, + 0x62a18a6, + 0x64498a8, + 0x2644d912, + 0x2645d913, + 0x26465917, + 0x26471919, + 0x647591c, + 0x647991d, + 0x64a191e, + 0x64c9928, + 0x64cd932, + 0x6505933, + 0x6525941, + 0x707d949, + 0x7081c1f, + 0x7085c20, + 0x27089c21, + 0x708dc22, + 0x27091c23, + 0x7095c24, + 0x270a1c25, + 0x70a5c28, + 0x70a9c29, + 0x270adc2a, + 0x70b1c2b, + 0x270b9c2c, + 0x70bdc2e, + 0x70c1c2f, + 0x270d1c30, + 0x70d5c34, + 0x70d9c35, + 0x70ddc36, + 0x70e1c37, + 0x270e5c38, + 0x70e9c39, + 0x70edc3a, + 0x70f1c3b, + 0x70f5c3c, + 0x270fdc3d, + 0x7101c3f, + 0x7105c40, + 0x7109c41, + 0x2710dc42, + 0x7111c43, + 0x27119c44, + 0x2711dc46, + 0x7139c47, + 0x7149c4e, + 0x718dc52, + 0x7191c63, + 0x71b5c64, + 0x71b9c6d, + 0x71bdc6e, + 0x7365c6f, + 0x27369cd9, + 0x27371cda, + 0x27375cdc, + 0x27379cdd, + 0x7381cde, + 0x745dce0, + 0x27469d17, + 0x2746dd1a, + 0x27471d1b, + 0x27475d1c, + 0x7479d1d, + 0x74a5d1e, + 0x74a9d29, + 0x74cdd2a, + 0x74d9d33, + 0x74f9d36, + 0x74fdd3e, + 0x7535d3f, + 0x77cdd4d, + 0x7889df3, + 0x788de22, + 0x78a1e23, + 0x78d5e28, + 0x790de35, + 0x27911e43, + 0x792de44, + 0x7955e4b, + 0x7959e55, + 0x797de56, + 0x7999e5f, + 0x79c1e66, + 0x79d1e70, + 0x79d5e74, + 0x79d9e75, + 0x7a11e76, + 0x7a1de84, + 0x7a41e87, + 0x7ac1e90, + 0x27ac5eb0, + 0x7ad5eb1, + 0x7addeb5, + 0x7b01eb7, + 0x7b21ec0, + 0x7b35ec8, + 0x7b49ecd, + 0x7b4ded2, + 0x7b6ded3, + 0x7c11edb, + 0x7c2df04, + 0x7c51f0b, + 0x7c55f14, + 0x7c5df15, + 0x7c6df17, + 0x7c75f1b, + 0x7c89f1d, + 0x7ca9f22, + 0x7cb5f2a, + 0x7cc5f2d, + 0x7cfdf31, + 0x7dd1f3f, + 0x7dd5f74, + 0x7de9f75, + 0x7df1f7a, + 0x7e09f7c, + 0x7e0df82, + 0x7e19f83, + 0x7e1df86, + 0x7e39f87, + 0x7e79f8e, + 0x7e7df9e, + 0x7e9df9f, + 0x7eedfa7, + 0x7f09fbb, + 0x7f11fc2, + 0x7f65fc4, + 0x7f69fd9, + 0x7f6dfda, + 0x7f71fdb, + 0x7fb5fdc, + 0x7fc5fed, + 0x8005ff1, + 0x800a001, + 0x803a002, + 0x818200e, + 0x81aa060, + 0x81da06a, + 0x81f6076, + 0x81fe07d, + 0x820a07f, + 0x831e082, 0x832a0c7, - 0x83460ca, - 0x834a0d1, - 0x835a0d2, - 0x83760d6, - 0x83ba0dd, - 0x83be0ee, - 0x83d20ef, - 0x84060f4, - 0x8416101, - 0x8436105, - 0x844e10d, - 0x8466113, - 0x846e119, - 0x284b211b, - 0x84b612c, - 0x84e212d, - 0x84ea138, - 0x84fe13a, + 0x83360ca, + 0x83420cd, + 0x834e0d0, + 0x835a0d3, + 0x83660d6, + 0x83720d9, + 0x837e0dc, + 0x838a0df, + 0x83960e2, + 0x83a20e5, + 0x83ae0e8, + 0x83ba0eb, + 0x83c20ee, + 0x83ce0f0, + 0x83da0f3, + 0x83e60f6, + 0x83f20f9, + 0x83fe0fc, + 0x840a0ff, + 0x8416102, + 0x8422105, + 0x842e108, + 0x843a10b, + 0x844610e, + 0x8472111, + 0x847e11c, + 0x848a11f, + 0x8496122, + 0x84a2125, + 0x84ae128, + 0x84b612b, + 0x84c212d, + 0x84ce130, + 0x84da133, + 0x84e6136, + 0x84f2139, + 0x84fe13c, + 0x850a13f, + 0x8516142, + 0x8522145, + 0x852e148, + 0x853a14b, + 0x854614e, + 0x8552151, + 0x855a154, + 0x8566156, + 0x8572159, + 0x857e15c, + 0x858a15f, + 0x8596162, + 0x85a2165, + 0x85ae168, + 0x85ba16b, + 0x85be16e, + 0x85ca16f, + 0x85e6172, + 0x85ea179, + 0x85fa17a, + 0x861617e, + 0x865a185, + 0x865e196, + 0x8672197, + 0x86a619c, + 0x86b61a9, + 0x86da1ad, + 0x86f21b6, + 0x870a1bc, + 0x87221c2, + 0x87321c8, + 0x287761cc, + 0x877a1dd, + 0x87a61de, + 0x87ae1e9, + 0x87c21eb, } -// max children 500 (capacity 1023) -// max text offset 29102 (capacity 32767) +// max children 522 (capacity 1023) +// max text offset 29878 (capacity 32767) // max text length 36 (capacity 63) -// max hi 8511 (capacity 16383) -// max lo 8506 (capacity 16383) +// max hi 8688 (capacity 16383) +// max lo 8683 (capacity 16383) diff --git a/vendor/golang.org/x/net/publicsuffix/table_test.go b/vendor/golang.org/x/net/publicsuffix/table_test.go index 228010cae..c05b8f25d 100644 --- a/vendor/golang.org/x/net/publicsuffix/table_test.go +++ b/vendor/golang.org/x/net/publicsuffix/table_test.go @@ -355,6 +355,7 @@ var rules = [...]string{ "art.br", "ato.br", "b.br", + "barueri.br", "belem.br", "bhz.br", "bio.br", @@ -850,7 +851,15 @@ var rules = [...]string{ "mil.gt", "net.gt", "org.gt", - "*.gu", + "gu", + "com.gu", + "edu.gu", + "gov.gu", + "guam.gu", + "info.gu", + "net.gu", + "org.gu", + "web.gu", "gw", "gy", "co.gy", @@ -1074,6 +1083,12 @@ var rules = [...]string{ "taa.it", "tos.it", "toscana.it", + "trentin-sud-tirol.it", + "xn--trentin-sud-tirol-tsj.it", + "trentin-sudtirol.it", + "xn--trentin-sudtirol-b9i.it", + "trentin-sued-tirol.it", + "trentin-suedtirol.it", "trentino-a-adige.it", "trentino-aadige.it", "trentino-alto-adige.it", @@ -1081,9 +1096,12 @@ var rules = [...]string{ "trentino-s-tirol.it", "trentino-stirol.it", "trentino-sud-tirol.it", + "xn--trentino-sud-tirol-dck.it", "trentino-sudtirol.it", + "xn--trentino-sudtirol-usj.it", "trentino-sued-tirol.it", "trentino-suedtirol.it", + "trentino.it", "trentinoa-adige.it", "trentinoaadige.it", "trentinoalto-adige.it", @@ -1091,9 +1109,17 @@ var rules = [...]string{ "trentinos-tirol.it", "trentinostirol.it", "trentinosud-tirol.it", + "xn--trentinosud-tirol-tsj.it", "trentinosudtirol.it", + "xn--trentinosudtirol-b9i.it", "trentinosued-tirol.it", "trentinosuedtirol.it", + "trentinsud-tirol.it", + "xn--trentinsud-tirol-98i.it", + "trentinsudtirol.it", + "xn--trentinsudtirol-rqi.it", + "trentinsued-tirol.it", + "trentinsuedtirol.it", "tuscany.it", "umb.it", "umbria.it", @@ -1108,7 +1134,13 @@ var rules = [...]string{ "valled-aosta.it", "valledaosta.it", "vallee-aoste.it", + "xn--vallee-aoste-i2g.it", + "vallee-d-aoste.it", + "xn--vallee-d-aoste-43h.it", "valleeaoste.it", + "xn--valleeaoste-6jg.it", + "valleedaoste.it", + "xn--valleedaoste-i2g.it", "vao.it", "vda.it", "ven.it", @@ -1140,6 +1172,9 @@ var rules = [...]string{ "av.it", "avellino.it", "ba.it", + "balsan-sudtirol.it", + "xn--balsan-sudtirol-rqi.it", + "balsan-suedtirol.it", "balsan.it", "bari.it", "barletta-trani-andria.it", @@ -1154,13 +1189,21 @@ var rules = [...]string{ "bn.it", "bo.it", "bologna.it", + "bolzano-altoadige.it", "bolzano.it", + "bozen-sudtirol.it", + "xn--bozen-sudtirol-76h.it", + "bozen-suedtirol.it", "bozen.it", "br.it", "brescia.it", "brindisi.it", "bs.it", "bt.it", + "bulsan-sudtirol.it", + "xn--bulsan-sudtirol-rqi.it", + "bulsan-suedtirol.it", + "bulsan.it", "bz.it", "ca.it", "cagliari.it", @@ -1178,7 +1221,9 @@ var rules = [...]string{ "cb.it", "ce.it", "cesena-forli.it", + "xn--cesena-forli-c2g.it", "cesenaforli.it", + "xn--cesenaforli-0jg.it", "ch.it", "chieti.it", "ci.it", @@ -1209,7 +1254,9 @@ var rules = [...]string{ "fm.it", "foggia.it", "forli-cesena.it", + "xn--forli-cesena-41g.it", "forlicesena.it", + "xn--forlicesena-ujg.it", "fr.it", "frosinone.it", "ge.it", @@ -1340,6 +1387,7 @@ var rules = [...]string{ "sr.it", "ss.it", "suedtirol.it", + "xn--sudtirol-y0e.it", "sv.it", "ta.it", "taranto.it", @@ -1358,7 +1406,6 @@ var rules = [...]string{ "traniandriabarletta.it", "tranibarlettaandria.it", "trapani.it", - "trentino.it", "trento.it", "treviso.it", "trieste.it", @@ -5499,7 +5546,6 @@ var rules = [...]string{ "kep.tr", "nc.tr", "gov.nc.tr", - "travel", "tt", "co.tt", "com.tt", @@ -5959,6 +6005,12 @@ var rules = [...]string{ "xn--node", "xn--qxam", "xn--j6w193g", + "xn--55qx5d.xn--j6w193g", + "xn--wcvs22d.xn--j6w193g", + "xn--mxtq1m.xn--j6w193g", + "xn--gmqw5a.xn--j6w193g", + "xn--od0alg.xn--j6w193g", + "xn--uc0atv.xn--j6w193g", "xn--2scrj9c", "xn--3hcrj9c", "xn--45br5cyl", @@ -5967,6 +6019,7 @@ var rules = [...]string{ "xn--mgbgu82a", "xn--rvc1e0am3e", "xn--h2brj9c", + "xn--mgbbh1a", "xn--mgbbh1a71e", "xn--fpcrj9c3d", "xn--gecrj9c", @@ -6192,7 +6245,6 @@ var rules = [...]string{ "boo", "book", "booking", - "boots", "bosch", "bostik", "boston", @@ -6253,6 +6305,7 @@ var rules = [...]string{ "cfd", "chanel", "channel", + "charity", "chase", "chat", "cheap", @@ -6561,6 +6614,7 @@ var rules = [...]string{ "imdb", "immo", "immobilien", + "inc", "industries", "infiniti", "ing", @@ -6663,6 +6717,7 @@ var rules = [...]string{ "live", "living", "lixil", + "llc", "loan", "loans", "locker", @@ -6739,7 +6794,6 @@ var rules = [...]string{ "movistar", "msd", "mtn", - "mtpc", "mtr", "mutual", "nab", @@ -6993,6 +7047,7 @@ var rules = [...]string{ "soy", "space", "spiegel", + "sport", "spot", "spreadbetting", "srl", @@ -7074,6 +7129,7 @@ var rules = [...]string{ "trade", "trading", "training", + "travel", "travelchannel", "travelers", "travelersinsurance", @@ -7236,6 +7292,7 @@ var rules = [...]string{ "xn--nqv7f", "xn--nqv7fs00ema", "xn--nyqy26a", + "xn--otu796d", "xn--p1acf", "xn--pbt977c", "xn--pssy2u", @@ -7278,9 +7335,6 @@ var rules = [...]string{ "cc.ua", "inf.ua", "ltd.ua", - "1password.ca", - "1password.com", - "1password.eu", "beep.pl", "*.compute.estate", "*.alces.network", @@ -7294,6 +7348,7 @@ var rules = [...]string{ "elasticbeanstalk.com", "ap-northeast-1.elasticbeanstalk.com", "ap-northeast-2.elasticbeanstalk.com", + "ap-northeast-3.elasticbeanstalk.com", "ap-south-1.elasticbeanstalk.com", "ap-southeast-1.elasticbeanstalk.com", "ap-southeast-2.elasticbeanstalk.com", @@ -7379,6 +7434,7 @@ var rules = [...]string{ "backplaneapp.io", "betainabox.com", "bnr.la", + "blackbaudcdn.net", "boomla.net", "boxfuse.io", "square7.ch", @@ -7409,7 +7465,6 @@ var rules = [...]string{ "qc.com", "ru.com", "sa.com", - "se.com", "se.net", "uk.com", "uk.net", @@ -7426,7 +7481,9 @@ var rules = [...]string{ "certmgr.org", "xenapponazure.com", "virtueeldomein.nl", + "cleverapps.io", "c66.me", + "cloud66.ws", "jdevcloud.com", "wpdevcloud.com", "cloudaccess.host", @@ -7435,6 +7492,7 @@ var rules = [...]string{ "cloudcontrolled.com", "cloudcontrolapp.com", "co.ca", + "*.otap.co", "co.cz", "c.cdn77.org", "cdn77-ssl.net", @@ -7452,6 +7510,8 @@ var rules = [...]string{ "cloudns.pro", "cloudns.pw", "cloudns.us", + "cloudeity.net", + "cnpy.gdn", "co.nl", "co.no", "webhosting.be", @@ -7472,6 +7532,12 @@ var rules = [...]string{ "cyon.site", "daplie.me", "localhost.daplie.me", + "dattolocal.com", + "dattorelay.com", + "dattoweb.com", + "mydatto.com", + "dattolocal.net", + "mydatto.net", "biz.dk", "co.dk", "firm.dk", @@ -7778,6 +7844,7 @@ var rules = [...]string{ "ddnss.org", "definima.net", "definima.io", + "bci.dnstrace.pro", "ddnsfree.com", "ddnsgeek.com", "giize.com", @@ -7952,6 +8019,8 @@ var rules = [...]string{ "a.ssl.fastly.net", "b.ssl.fastly.net", "global.ssl.fastly.net", + "fastpanel.direct", + "fastvps-server.com", "fhapp.xyz", "fedorainfracloud.org", "fedorapeople.org", @@ -7968,7 +8037,10 @@ var rules = [...]string{ "fbxos.fr", "freebox-os.fr", "freeboxos.fr", + "freedesktop.org", "*.futurecms.at", + "*.ex.futurecms.at", + "*.in.futurecms.at", "futurehosting.at", "futuremailing.at", "*.ex.ortsinfo.at", @@ -8068,10 +8140,16 @@ var rules = [...]string{ "withgoogle.com", "withyoutube.com", "hashbang.sh", + "hasura.app", "hasura-app.io", "hepforge.org", "herokuapp.com", "herokussl.com", + "myravendb.com", + "ravendb.community", + "ravendb.me", + "development.run", + "ravendb.run", "moonscale.net", "iki.fi", "biz.at", @@ -8106,6 +8184,13 @@ var rules = [...]string{ "to.leg.br", "pixolino.com", "ipifony.net", + "mein-iserv.de", + "test-iserv.de", + "myjino.ru", + "*.hosting.myjino.ru", + "*.landing.myjino.ru", + "*.spectrum.myjino.ru", + "*.vps.myjino.ru", "*.triton.zone", "*.cns.joyent.com", "js.org", @@ -8116,25 +8201,56 @@ var rules = [...]string{ "git-repos.de", "lcube-server.de", "svn-repos.de", + "app.lmpm.com", + "linkitools.space", + "linkyard.cloud", + "linkyard-cloud.ch", "we.bs", + "uklugs.org", + "glug.org.uk", + "lug.org.uk", + "lugs.org.uk", "barsy.bg", + "barsy.co.uk", + "barsyonline.co.uk", + "barsycenter.com", "barsyonline.com", + "barsy.club", "barsy.de", "barsy.eu", "barsy.in", + "barsy.info", + "barsy.io", + "barsy.me", + "barsy.menu", + "barsy.mobi", "barsy.net", "barsy.online", + "barsy.org", + "barsy.pro", + "barsy.pub", + "barsy.shop", + "barsy.site", "barsy.support", + "barsy.uk", "*.magentosite.cloud", + "mayfirst.info", + "mayfirst.org", "hb.cldmail.ru", + "miniserver.com", + "memset.net", "cloud.metacentrum.cz", "custom.metacentrum.cz", + "flt.cloud.muni.cz", + "usr.cloud.muni.cz", "meteorapp.com", "eu.meteorapp.com", "co.pl", + "azurecontainer.io", "azurewebsites.net", "azure-mobile.net", "cloudapp.net", + "mozilla-iot.org", "bmoattachments.org", "net.ru", "org.ru", @@ -8145,6 +8261,31 @@ var rules = [...]string{ "ngrok.io", "nh-serv.co.uk", "nfshost.com", + "dnsking.ch", + "mypi.co", + "n4t.co", + "001www.com", + "ddnslive.com", + "myiphost.com", + "forumz.info", + "16-b.it", + "32-b.it", + "64-b.it", + "soundcast.me", + "tcp4.me", + "dnsup.net", + "hicam.net", + "now-dns.net", + "ownip.net", + "vpndns.net", + "dynserv.org", + "now-dns.org", + "x443.pw", + "now-dns.top", + "ntdll.top", + "freeddns.us", + "crafting.xyz", + "zapto.xyz", "nsupdate.info", "nerdpol.ovh", "blogsyte.com", @@ -8235,27 +8376,35 @@ var rules = [...]string{ "stage.nodeart.io", "nodum.co", "nodum.io", + "pcloud.host", "nyc.mn", "nom.ae", + "nom.af", "nom.ai", "nom.al", "nym.by", "nym.bz", "nom.cl", "nom.gd", + "nom.ge", "nom.gl", "nym.gr", "nom.gt", + "nym.gy", "nom.hn", + "nym.ie", "nom.im", + "nom.ke", "nym.kz", "nym.la", + "nym.lc", "nom.li", "nym.li", "nym.lt", "nym.lu", "nym.me", "nom.mk", + "nym.mn", "nym.mx", "nom.nu", "nym.nz", @@ -8263,22 +8412,28 @@ var rules = [...]string{ "nym.pt", "nom.pw", "nom.qa", + "nym.ro", "nom.rs", "nom.si", "nym.sk", + "nom.st", "nym.su", "nym.sx", + "nom.tj", "nym.tw", "nom.ug", "nom.uy", "nom.vc", "nom.vg", "cya.gg", + "cloudycluster.net", "nid.io", "opencraft.hosting", "operaunite.com", "outsystemscloud.com", "ownprovider.com", + "own.pm", + "ox.rs", "oy.lc", "pgfog.com", "pagefrontapp.com", @@ -8299,6 +8454,7 @@ var rules = [...]string{ "protonet.io", "chirurgiens-dentistes-en-france.fr", "byen.site", + "ras.ru", "qa2.com", "dev-myqnapcloud.com", "alpha-myqnapcloud.com", @@ -8318,6 +8474,7 @@ var rules = [...]string{ "sandcats.io", "logoip.de", "logoip.com", + "schokokeks.net", "scrysec.com", "firewall-gateway.com", "firewall-gateway.de", @@ -8350,8 +8507,10 @@ var rules = [...]string{ "apps.lair.io", "*.stolos.io", "spacekit.io", + "customer.speedpartner.de", "stackspace.space", "storj.farm", + "utwente.io", "temp-dns.com", "diskstation.me", "dscloud.biz", @@ -8373,6 +8532,7 @@ var rules = [...]string{ "gdynia.pl", "med.pl", "sopot.pl", + "gwiddle.co.uk", "cust.dev.thingdust.io", "cust.disrec.thingdust.io", "cust.prod.thingdust.io", @@ -8416,11 +8576,15 @@ var rules = [...]string{ "synology-diskstation.de", "synology-ds.de", "uber.space", + "*.uberspace.de", "hk.com", "hk.org", "ltd.hk", "inc.hk", + "virtualuser.de", + "virtual-user.de", "lib.de.us", + "2038.io", "router.management", "v-info.info", "wedeploy.io", @@ -8428,9 +8592,14 @@ var rules = [...]string{ "wedeploy.sh", "remotewd.com", "wmflabs.org", + "half.host", + "xnbay.com", + "u2.xnbay.com", + "u2-local.xnbay.com", "cistron.nl", "demon.nl", "xs4all.space", + "official.academy", "yolasite.com", "ybo.faith", "yombo.me", @@ -8439,9 +8608,12 @@ var rules = [...]string{ "ybo.review", "ybo.science", "ybo.trade", + "nohost.me", + "noho.st", "za.net", "za.org", "now.sh", + "zone.id", } var nodeLabels = [...]string{ @@ -8611,7 +8783,6 @@ var nodeLabels = [...]string{ "boo", "book", "booking", - "boots", "bosch", "bostik", "boston", @@ -8686,6 +8857,7 @@ var nodeLabels = [...]string{ "ch", "chanel", "channel", + "charity", "chase", "chat", "cheap", @@ -9060,6 +9232,7 @@ var nodeLabels = [...]string{ "immo", "immobilien", "in", + "inc", "industries", "infiniti", "info", @@ -9190,6 +9363,7 @@ var nodeLabels = [...]string{ "living", "lixil", "lk", + "llc", "loan", "loans", "locker", @@ -9290,7 +9464,6 @@ var nodeLabels = [...]string{ "msd", "mt", "mtn", - "mtpc", "mtr", "mu", "museum", @@ -9604,6 +9777,7 @@ var nodeLabels = [...]string{ "soy", "space", "spiegel", + "sport", "spot", "spreadbetting", "sr", @@ -9911,6 +10085,7 @@ var nodeLabels = [...]string{ "xn--mgbai9azgqp6j", "xn--mgbayh7gpa", "xn--mgbb9fbpob", + "xn--mgbbh1a", "xn--mgbbh1a71e", "xn--mgbc0a9azcg", "xn--mgbca7dzdo", @@ -9939,6 +10114,7 @@ var nodeLabels = [...]string{ "xn--nyqy26a", "xn--o3cw4h", "xn--ogbpf8fl", + "xn--otu796d", "xn--p1acf", "xn--p1ai", "xn--pbt977c", @@ -10002,6 +10178,7 @@ var nodeLabels = [...]string{ "mil", "net", "org", + "official", "nom", "ac", "blogspot", @@ -10103,6 +10280,7 @@ var nodeLabels = [...]string{ "edu", "gov", "net", + "nom", "org", "co", "com", @@ -10129,6 +10307,7 @@ var nodeLabels = [...]string{ "it", "og", "pb", + "hasura", "com", "edu", "gob", @@ -10165,6 +10344,8 @@ var nodeLabels = [...]string{ "priv", "blogspot", "ex", + "in", + "ex", "kunden", "act", "asn", @@ -10356,6 +10537,7 @@ var nodeLabels = [...]string{ "art", "ato", "b", + "barueri", "belem", "bhz", "bio", @@ -10550,7 +10732,6 @@ var nodeLabels = [...]string{ "nym", "org", "za", - "1password", "ab", "awdev", "bc", @@ -10583,8 +10764,10 @@ var nodeLabels = [...]string{ "2ix", "4lima", "blogspot", + "dnsking", "gotdns", "lima-city", + "linkyard-cloud", "square7", "ac", "asso", @@ -10608,11 +10791,13 @@ var nodeLabels = [...]string{ "gov", "mil", "nom", + "linkyard", "magentosite", "sensiosite", "statics", "trafficplex", "vapor", + "barsy", "cloudns", "co", "com", @@ -10677,16 +10862,19 @@ var nodeLabels = [...]string{ "info", "int", "mil", + "mypi", + "n4t", "net", "nodum", "nom", "org", + "otap", "rec", "web", "blogspot", + "001www", "0emm", "1kapp", - "1password", "3utilities", "4u", "africa", @@ -10696,6 +10884,7 @@ var nodeLabels = [...]string{ "applinzi", "appspot", "ar", + "barsycenter", "barsyonline", "betainabox", "bitballoon", @@ -10714,9 +10903,13 @@ var nodeLabels = [...]string{ "co", "codespot", "damnserver", + "dattolocal", + "dattorelay", + "dattoweb", "ddnsfree", "ddnsgeek", "ddnsking", + "ddnslive", "de", "dev-myqnapcloud", "ditchyourip", @@ -10754,6 +10947,7 @@ var nodeLabels = [...]string{ "eu", "evennode", "familyds", + "fastvps-server", "fbsbx", "firebaseapp", "firewall-gateway", @@ -10894,14 +11088,19 @@ var nodeLabels = [...]string{ "kr", "likes-pie", "likescandy", + "lmpm", "logoip", "loseyourip", "meteorapp", "mex", + "miniserver", "myactivedirectory", "myasustor", + "mydatto", "mydrobo", + "myiphost", "myqnapcloud", + "myravendb", "mysecuritycamera", "myshopblocks", "mytuleap", @@ -10935,7 +11134,6 @@ var nodeLabels = [...]string{ "sa", "saves-the-whales", "scrysec", - "se", "securitytactics", "selfip", "sells-for-less", @@ -10974,6 +11172,7 @@ var nodeLabels = [...]string{ "wpdevcloud", "writesthisblog", "xenapponazure", + "xnbay", "yolasite", "za", "ap-northeast-1", @@ -11062,6 +11261,7 @@ var nodeLabels = [...]string{ "beta", "ap-northeast-1", "ap-northeast-2", + "ap-northeast-3", "ap-south-1", "ap-southeast-1", "ap-southeast-2", @@ -11086,8 +11286,12 @@ var nodeLabels = [...]string{ "us-4", "apps", "cns", + "app", "eu", "xen", + "u2", + "u2-local", + "ravendb", "de", "ac", "co", @@ -11128,9 +11332,13 @@ var nodeLabels = [...]string{ "co", "e4", "metacentrum", + "muni", "realm", "cloud", "custom", + "cloud", + "flt", + "usr", "12hp", "2ix", "4lima", @@ -11165,6 +11373,7 @@ var nodeLabels = [...]string{ "leitungsen", "lima-city", "logoip", + "mein-iserv", "mein-vigor", "my-gateway", "my-router", @@ -11172,17 +11381,24 @@ var nodeLabels = [...]string{ "my-wan", "myhome-server", "spdns", + "speedpartner", "square7", "svn-repos", "syno-ds", "synology-diskstation", "synology-ds", "taifun-dns", + "test-iserv", "traeumtgerade", + "uberspace", + "virtual-user", + "virtualuser", "dyn", "dyn", "dyndns", "dyn", + "customer", + "fastpanel", "biz", "blogspot", "co", @@ -11260,7 +11476,6 @@ var nodeLabels = [...]string{ "name", "net", "org", - "1password", "barsy", "cloudns", "diskstation", @@ -11308,11 +11523,13 @@ var nodeLabels = [...]string{ "tm", "veterinaire", "nom", + "cnpy", "com", "edu", "gov", "mil", "net", + "nom", "org", "pvt", "co", @@ -11364,11 +11581,20 @@ var nodeLabels = [...]string{ "net", "nom", "org", + "com", + "edu", + "gov", + "guam", + "info", + "net", + "org", + "web", "co", "com", "edu", "gov", "net", + "nym", "org", "blogspot", "com", @@ -11403,6 +11629,8 @@ var nodeLabels = [...]string{ "org", "cloudaccess", "freesite", + "half", + "pcloud", "opencraft", "blogspot", "com", @@ -11469,9 +11697,11 @@ var nodeLabels = [...]string{ "or", "sch", "web", + "zone", "blogspot", "blogspot", "gov", + "nym", "ac", "co", "gov", @@ -11509,25 +11739,32 @@ var nodeLabels = [...]string{ "res", "barrel-of-knowledge", "barrell-of-knowledge", + "barsy", "cloudns", "dvrcam", "dynamic-dns", "dyndns", "for-our", + "forumz", "groks-the", "groks-this", "here-for-more", "ilovecollege", "knowsitall", + "mayfirst", "no-ip", "nsupdate", "selfip", "v-info", "webhop", "eu", + "2038", + "azurecontainer", "backplaneapp", + "barsy", "boxfuse", "browsersafetymark", + "cleverapps", "com", "dedyn", "definima", @@ -11552,6 +11789,7 @@ var nodeLabels = [...]string{ "spacekit", "stolos", "thingdust", + "utwente", "vaporcloud", "wedeploy", "customer", @@ -11589,6 +11827,9 @@ var nodeLabels = [...]string{ "int", "net", "org", + "16-b", + "32-b", + "64-b", "abr", "abruzzo", "ag", @@ -11621,6 +11862,8 @@ var nodeLabels = [...]string{ "avellino", "ba", "balsan", + "balsan-sudtirol", + "balsan-suedtirol", "bari", "barletta-trani-andria", "barlettatraniandria", @@ -11638,12 +11881,18 @@ var nodeLabels = [...]string{ "bo", "bologna", "bolzano", + "bolzano-altoadige", "bozen", + "bozen-sudtirol", + "bozen-suedtirol", "br", "brescia", "brindisi", "bs", "bt", + "bulsan", + "bulsan-sudtirol", + "bulsan-suedtirol", "bz", "ca", "cagliari", @@ -11889,6 +12138,10 @@ var nodeLabels = [...]string{ "traniandriabarletta", "tranibarlettaandria", "trapani", + "trentin-sud-tirol", + "trentin-sudtirol", + "trentin-sued-tirol", + "trentin-suedtirol", "trentino", "trentino-a-adige", "trentino-aadige", @@ -11910,6 +12163,10 @@ var nodeLabels = [...]string{ "trentinosudtirol", "trentinosued-tirol", "trentinosuedtirol", + "trentinsud-tirol", + "trentinsudtirol", + "trentinsued-tirol", + "trentinsuedtirol", "trento", "treviso", "trieste", @@ -11935,7 +12192,9 @@ var nodeLabels = [...]string{ "valled-aosta", "valledaosta", "vallee-aoste", + "vallee-d-aoste", "valleeaoste", + "valleedaoste", "vao", "varese", "vb", @@ -11958,6 +12217,26 @@ var nodeLabels = [...]string{ "vs", "vt", "vv", + "xn--balsan-sudtirol-rqi", + "xn--bozen-sudtirol-76h", + "xn--bulsan-sudtirol-rqi", + "xn--cesena-forli-c2g", + "xn--cesenaforli-0jg", + "xn--forli-cesena-41g", + "xn--forlicesena-ujg", + "xn--sudtirol-y0e", + "xn--trentin-sud-tirol-tsj", + "xn--trentin-sudtirol-b9i", + "xn--trentino-sud-tirol-dck", + "xn--trentino-sudtirol-usj", + "xn--trentinosud-tirol-tsj", + "xn--trentinosudtirol-b9i", + "xn--trentinsud-tirol-98i", + "xn--trentinsudtirol-rqi", + "xn--vallee-aoste-i2g", + "xn--vallee-d-aoste-43h", + "xn--valleeaoste-6jg", + "xn--valleedaoste-i2g", "co", "net", "org", @@ -13767,6 +14046,7 @@ var nodeLabels = [...]string{ "me", "mobi", "ne", + "nom", "or", "sc", "blogspot", @@ -13878,6 +14158,7 @@ var nodeLabels = [...]string{ "edu", "gov", "net", + "nym", "org", "oy", "blogspot", @@ -13941,6 +14222,7 @@ var nodeLabels = [...]string{ "tm", "blogspot", "ac", + "barsy", "brasilia", "c66", "co", @@ -13958,15 +14240,20 @@ var nodeLabels = [...]string{ "loginto", "myds", "net", + "nohost", "noip", "nym", "org", "priv", + "ravendb", + "soundcast", "synology", + "tcp4", "webhop", "wedeploy", "yombo", "localhost", + "barsy", "co", "com", "edu", @@ -13995,12 +14282,14 @@ var nodeLabels = [...]string{ "edu", "gov", "nyc", + "nym", "org", "com", "edu", "gov", "net", "org", + "barsy", "dscloud", "blogspot", "gov", @@ -14645,6 +14934,7 @@ var nodeLabels = [...]string{ "azure-mobile", "azurewebsites", "barsy", + "blackbaudcdn", "blogdns", "boomla", "bounceme", @@ -14657,14 +14947,18 @@ var nodeLabels = [...]string{ "channelsdvr", "cloudaccess", "cloudapp", + "cloudeity", "cloudfront", "cloudfunctions", + "cloudycluster", "cryptonomic", + "dattolocal", "ddns", "debian", "definima", "dnsalias", "dnsdojo", + "dnsup", "does-it", "dontexist", "dsmynas", @@ -14687,6 +14981,7 @@ var nodeLabels = [...]string{ "gb", "gets-it", "ham-radio-op", + "hicam", "homeftp", "homeip", "homelinux", @@ -14701,7 +14996,9 @@ var nodeLabels = [...]string{ "jp", "kicks-ass", "knx-server", + "memset", "moonscale", + "mydatto", "mydissent", "myeffect", "myfritz", @@ -14710,13 +15007,16 @@ var nodeLabels = [...]string{ "mysecuritycamera", "nhlfan", "no-ip", + "now-dns", "office-on-the", + "ownip", "pgafan", "podzone", "privatizehealthinsurance", "rackmaze", "redirectme", "ru", + "schokokeks", "scrapper-site", "se", "selfip", @@ -14732,6 +15032,7 @@ var nodeLabels = [...]string{ "thruhere", "twmail", "uk", + "vpndns", "webhop", "za", "r", @@ -15598,6 +15899,7 @@ var nodeLabels = [...]string{ "accesscam", "ae", "amune", + "barsy", "blogdns", "blogsite", "bmoattachments", @@ -15622,6 +15924,7 @@ var nodeLabels = [...]string{ "dvrdns", "dynalias", "dyndns", + "dynserv", "endofinternet", "endoftheinternet", "eu", @@ -15630,6 +15933,7 @@ var nodeLabels = [...]string{ "fedorapeople", "fedoraproject", "freeddns", + "freedesktop", "from-me", "game-host", "gotdns", @@ -15661,8 +15965,10 @@ var nodeLabels = [...]string{ "isa-geek", "js", "kicks-ass", + "mayfirst", "misconfused", "mlbfan", + "mozilla-iot", "my-firewall", "myfirewall", "myftp", @@ -15670,6 +15976,7 @@ var nodeLabels = [...]string{ "mywire", "nflfan", "no-ip", + "now-dns", "pimienta", "podzone", "poivron", @@ -15688,6 +15995,7 @@ var nodeLabels = [...]string{ "tuxfamily", "twmail", "ufcfan", + "uklugs", "us", "webhop", "webredirect", @@ -16023,6 +16331,7 @@ var nodeLabels = [...]string{ "wuoz", "wzmiuw", "zp", + "own", "co", "edu", "gov", @@ -16046,13 +16355,16 @@ var nodeLabels = [...]string{ "acct", "avocat", "bar", + "barsy", "cloudns", "cpa", + "dnstrace", "eng", "jur", "law", "med", "recht", + "bci", "com", "edu", "gov", @@ -16070,6 +16382,7 @@ var nodeLabels = [...]string{ "nym", "org", "publ", + "barsy", "belau", "cloudns", "co", @@ -16078,6 +16391,7 @@ var nodeLabels = [...]string{ "ne", "nom", "or", + "x443", "com", "coop", "edu", @@ -16108,6 +16422,7 @@ var nodeLabels = [...]string{ "info", "nom", "nt", + "nym", "org", "rec", "shop", @@ -16125,6 +16440,7 @@ var nodeLabels = [...]string{ "in", "nom", "org", + "ox", "ac", "adygeya", "bashkiria", @@ -16144,6 +16460,7 @@ var nodeLabels = [...]string{ "mil", "mordovia", "msk", + "myjino", "mytis", "nalchik", "net", @@ -16151,11 +16468,18 @@ var nodeLabels = [...]string{ "org", "pp", "pyatigorsk", + "ras", "spb", "test", "vladikavkaz", "vladimir", "hb", + "hosting", + "landing", + "spectrum", + "vps", + "development", + "ravendb", "ac", "co", "com", @@ -16249,8 +16573,10 @@ var nodeLabels = [...]string{ "org", "platform", "wedeploy", + "barsy", "blogspot", "nom", + "barsy", "byen", "cyon", "platformsh", @@ -16272,6 +16598,7 @@ var nodeLabels = [...]string{ "com", "net", "org", + "linkitools", "stackspace", "uber", "xs4all", @@ -16283,6 +16610,8 @@ var nodeLabels = [...]string{ "gov", "mil", "net", + "noho", + "nom", "org", "principe", "saotome", @@ -16378,6 +16707,7 @@ var nodeLabels = [...]string{ "name", "net", "nic", + "nom", "org", "test", "web", @@ -16417,6 +16747,8 @@ var nodeLabels = [...]string{ "net", "org", "vpnplus", + "now-dns", + "ntdll", "av", "bbs", "bel", @@ -16584,6 +16916,7 @@ var nodeLabels = [...]string{ "org", "sc", "ac", + "barsy", "co", "gov", "ltd", @@ -16594,12 +16927,18 @@ var nodeLabels = [...]string{ "plc", "police", "sch", + "barsy", + "barsyonline", "blogspot", + "gwiddle", "nh-serv", "no-ip", "wellbeingzone", "homeoffice", "service", + "glug", + "lug", + "lugs", "ak", "al", "ar", @@ -16615,6 +16954,7 @@ var nodeLabels = [...]string{ "drud", "fed", "fl", + "freeddns", "ga", "golffan", "gu", @@ -16899,6 +17239,7 @@ var nodeLabels = [...]string{ "net", "org", "advisor", + "cloud66", "com", "dyndns", "edu", @@ -16912,6 +17253,12 @@ var nodeLabels = [...]string{ "xn--d1at", "xn--o1ac", "xn--o1ach", + "xn--55qx5d", + "xn--gmqw5a", + "xn--mxtq1m", + "xn--od0alg", + "xn--uc0atv", + "xn--wcvs22d", "xn--12c1fe0br", "xn--12cfi8ixb8l", "xn--12co0c3b4eva", @@ -16919,7 +17266,9 @@ var nodeLabels = [...]string{ "xn--m3ch0j3a", "xn--o3cyx2a", "blogsite", + "crafting", "fhapp", + "zapto", "ac", "agric", "alt", diff --git a/vendor/golang.org/x/net/webdav/prop.go b/vendor/golang.org/x/net/webdav/prop.go index e36a3b31d..4d7be3492 100644 --- a/vendor/golang.org/x/net/webdav/prop.go +++ b/vendor/golang.org/x/net/webdav/prop.go @@ -7,6 +7,7 @@ package webdav import ( "bytes" "encoding/xml" + "errors" "fmt" "io" "mime" @@ -376,10 +377,37 @@ func findContentLength(ctx context.Context, fs FileSystem, ls LockSystem, name s } func findLastModified(ctx context.Context, fs FileSystem, ls LockSystem, name string, fi os.FileInfo) (string, error) { - return fi.ModTime().Format(http.TimeFormat), nil + return fi.ModTime().UTC().Format(http.TimeFormat), nil +} + +// ErrNotImplemented should be returned by optional interfaces if they +// want the original implementation to be used. +var ErrNotImplemented = errors.New("not implemented") + +// ContentTyper is an optional interface for the os.FileInfo +// objects returned by the FileSystem. +// +// If this interface is defined then it will be used to read the +// content type from the object. +// +// If this interface is not defined the file will be opened and the +// content type will be guessed from the initial contents of the file. +type ContentTyper interface { + // ContentType returns the content type for the file. + // + // If this returns error ErrNotImplemented then the error will + // be ignored and the base implementation will be used + // instead. + ContentType(ctx context.Context) (string, error) } func findContentType(ctx context.Context, fs FileSystem, ls LockSystem, name string, fi os.FileInfo) (string, error) { + if do, ok := fi.(ContentTyper); ok { + ctype, err := do.ContentType(ctx) + if err != ErrNotImplemented { + return ctype, err + } + } f, err := fs.OpenFile(ctx, name, os.O_RDONLY, 0) if err != nil { return "", err @@ -402,7 +430,31 @@ func findContentType(ctx context.Context, fs FileSystem, ls LockSystem, name str return ctype, err } +// ETager is an optional interface for the os.FileInfo objects +// returned by the FileSystem. +// +// If this interface is defined then it will be used to read the ETag +// for the object. +// +// If this interface is not defined an ETag will be computed using the +// ModTime() and the Size() methods of the os.FileInfo object. +type ETager interface { + // ETag returns an ETag for the file. This should be of the + // form "value" or W/"value" + // + // If this returns error ErrNotImplemented then the error will + // be ignored and the base implementation will be used + // instead. + ETag(ctx context.Context) (string, error) +} + func findETag(ctx context.Context, fs FileSystem, ls LockSystem, name string, fi os.FileInfo) (string, error) { + if do, ok := fi.(ETager); ok { + etag, err := do.ETag(ctx) + if err != ErrNotImplemented { + return etag, err + } + } // The Apache http 2.4 web server by default concatenates the // modification time and size of a file. We replicate the heuristic // with nanosecond granularity. diff --git a/vendor/golang.org/x/net/webdav/prop_test.go b/vendor/golang.org/x/net/webdav/prop_test.go index 57d0e826f..bc42b9132 100644 --- a/vendor/golang.org/x/net/webdav/prop_test.go +++ b/vendor/golang.org/x/net/webdav/prop_test.go @@ -10,6 +10,7 @@ import ( "net/http" "os" "reflect" + "regexp" "sort" "testing" @@ -29,7 +30,7 @@ func TestMemPS(t *testing.T) { for i, p := range pst.Props { switch p.XMLName { case xml.Name{Space: "DAV:", Local: "getlastmodified"}: - p.InnerXML = []byte(fi.ModTime().Format(http.TimeFormat)) + p.InnerXML = []byte(fi.ModTime().UTC().Format(http.TimeFormat)) pst.Props[i] = p case xml.Name{Space: "DAV:", Local: "getetag"}: if fi.IsDir() { @@ -611,3 +612,106 @@ func (f noDeadPropsFile) Readdir(count int) ([]os.FileInfo, error) { return f.f func (f noDeadPropsFile) Seek(off int64, whence int) (int64, error) { return f.f.Seek(off, whence) } func (f noDeadPropsFile) Stat() (os.FileInfo, error) { return f.f.Stat() } func (f noDeadPropsFile) Write(p []byte) (int, error) { return f.f.Write(p) } + +type overrideContentType struct { + os.FileInfo + contentType string + err error +} + +func (o *overrideContentType) ContentType(ctx context.Context) (string, error) { + return o.contentType, o.err +} + +func TestFindContentTypeOverride(t *testing.T) { + fs, err := buildTestFS([]string{"touch /file"}) + if err != nil { + t.Fatalf("cannot create test filesystem: %v", err) + } + ctx := context.Background() + fi, err := fs.Stat(ctx, "/file") + if err != nil { + t.Fatalf("cannot Stat /file: %v", err) + } + + // Check non overridden case + originalContentType, err := findContentType(ctx, fs, nil, "/file", fi) + if err != nil { + t.Fatalf("findContentType /file failed: %v", err) + } + if originalContentType != "text/plain; charset=utf-8" { + t.Fatalf("ContentType wrong want %q got %q", "text/plain; charset=utf-8", originalContentType) + } + + // Now try overriding the ContentType + o := &overrideContentType{fi, "OverriddenContentType", nil} + ContentType, err := findContentType(ctx, fs, nil, "/file", o) + if err != nil { + t.Fatalf("findContentType /file failed: %v", err) + } + if ContentType != o.contentType { + t.Fatalf("ContentType wrong want %q got %q", o.contentType, ContentType) + } + + // Now return ErrNotImplemented and check we get the original content type + o = &overrideContentType{fi, "OverriddenContentType", ErrNotImplemented} + ContentType, err = findContentType(ctx, fs, nil, "/file", o) + if err != nil { + t.Fatalf("findContentType /file failed: %v", err) + } + if ContentType != originalContentType { + t.Fatalf("ContentType wrong want %q got %q", originalContentType, ContentType) + } +} + +type overrideETag struct { + os.FileInfo + eTag string + err error +} + +func (o *overrideETag) ETag(ctx context.Context) (string, error) { + return o.eTag, o.err +} + +func TestFindETagOverride(t *testing.T) { + fs, err := buildTestFS([]string{"touch /file"}) + if err != nil { + t.Fatalf("cannot create test filesystem: %v", err) + } + ctx := context.Background() + fi, err := fs.Stat(ctx, "/file") + if err != nil { + t.Fatalf("cannot Stat /file: %v", err) + } + + // Check non overridden case + originalETag, err := findETag(ctx, fs, nil, "/file", fi) + if err != nil { + t.Fatalf("findETag /file failed: %v", err) + } + matchETag := regexp.MustCompile(`^"-?[0-9a-f]{6,}"$`) + if !matchETag.MatchString(originalETag) { + t.Fatalf("ETag wrong, wanted something matching %v got %q", matchETag, originalETag) + } + + // Now try overriding the ETag + o := &overrideETag{fi, `"OverriddenETag"`, nil} + ETag, err := findETag(ctx, fs, nil, "/file", o) + if err != nil { + t.Fatalf("findETag /file failed: %v", err) + } + if ETag != o.eTag { + t.Fatalf("ETag wrong want %q got %q", o.eTag, ETag) + } + + // Now return ErrNotImplemented and check we get the original Etag + o = &overrideETag{fi, `"OverriddenETag"`, ErrNotImplemented} + ETag, err = findETag(ctx, fs, nil, "/file", o) + if err != nil { + t.Fatalf("findETag /file failed: %v", err) + } + if ETag != originalETag { + t.Fatalf("ETag wrong want %q got %q", originalETag, ETag) + } +} diff --git a/vendor/golang.org/x/net/websocket/websocket.go b/vendor/golang.org/x/net/websocket/websocket.go index e242c89a7..1f4f7be40 100644 --- a/vendor/golang.org/x/net/websocket/websocket.go +++ b/vendor/golang.org/x/net/websocket/websocket.go @@ -241,7 +241,10 @@ func (ws *Conn) Close() error { return err1 } +// IsClientConn reports whether ws is a client-side connection. func (ws *Conn) IsClientConn() bool { return ws.request == nil } + +// IsServerConn reports whether ws is a server-side connection. func (ws *Conn) IsServerConn() bool { return ws.request != nil } // LocalAddr returns the WebSocket Origin for the connection for client, or diff --git a/vendor/golang.org/x/sys/cpu/cpu.go b/vendor/golang.org/x/sys/cpu/cpu.go index 2d1fc12fb..3d88f8667 100644 --- a/vendor/golang.org/x/sys/cpu/cpu.go +++ b/vendor/golang.org/x/sys/cpu/cpu.go @@ -6,6 +6,9 @@ // various CPU architectures. package cpu +// CacheLinePad is used to pad structs to avoid false sharing. +type CacheLinePad struct{ _ [cacheLineSize]byte } + // X86 contains the supported CPU features of the // current X86/AMD64 platform. If the current platform // is not X86/AMD64 then all feature flags are false. @@ -14,7 +17,7 @@ package cpu // and HasAVX2 are only set if the OS supports XMM and YMM // registers in addition to the CPUID feature bit being set. var X86 struct { - _ [cacheLineSize]byte + _ CacheLinePad HasAES bool // AES hardware implementation (AES NI) HasADX bool // Multi-precision add-carry instruction extensions HasAVX bool // Advanced vector extension @@ -31,5 +34,5 @@ var X86 struct { HasSSSE3 bool // Supplemental streaming SIMD extension 3 HasSSE41 bool // Streaming SIMD extension 4 and 4.1 HasSSE42 bool // Streaming SIMD extension 4 and 4.2 - _ [cacheLineSize]byte + _ CacheLinePad } diff --git a/vendor/golang.org/x/sys/cpu/cpu_gc_x86.go b/vendor/golang.org/x/sys/cpu/cpu_gc_x86.go new file mode 100644 index 000000000..f7cb46971 --- /dev/null +++ b/vendor/golang.org/x/sys/cpu/cpu_gc_x86.go @@ -0,0 +1,16 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build 386 amd64 amd64p32 +// +build !gccgo + +package cpu + +// cpuid is implemented in cpu_x86.s for gc compiler +// and in cpu_gccgo.c for gccgo. +func cpuid(eaxArg, ecxArg uint32) (eax, ebx, ecx, edx uint32) + +// xgetbv with ecx = 0 is implemented in cpu_x86.s for gc compiler +// and in cpu_gccgo.c for gccgo. +func xgetbv() (eax, edx uint32) diff --git a/vendor/golang.org/x/sys/cpu/cpu_gccgo.c b/vendor/golang.org/x/sys/cpu/cpu_gccgo.c new file mode 100644 index 000000000..e363c7d13 --- /dev/null +++ b/vendor/golang.org/x/sys/cpu/cpu_gccgo.c @@ -0,0 +1,43 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build 386 amd64 amd64p32 +// +build gccgo + +#include +#include + +// Need to wrap __get_cpuid_count because it's declared as static. +int +gccgoGetCpuidCount(uint32_t leaf, uint32_t subleaf, + uint32_t *eax, uint32_t *ebx, + uint32_t *ecx, uint32_t *edx) +{ + return __get_cpuid_count(leaf, subleaf, eax, ebx, ecx, edx); +} + +// xgetbv reads the contents of an XCR (Extended Control Register) +// specified in the ECX register into registers EDX:EAX. +// Currently, the only supported value for XCR is 0. +// +// TODO: Replace with a better alternative: +// +// #include +// +// #pragma GCC target("xsave") +// +// void gccgoXgetbv(uint32_t *eax, uint32_t *edx) { +// unsigned long long x = _xgetbv(0); +// *eax = x & 0xffffffff; +// *edx = (x >> 32) & 0xffffffff; +// } +// +// Note that _xgetbv is defined starting with GCC 8. +void +gccgoXgetbv(uint32_t *eax, uint32_t *edx) +{ + __asm(" xorl %%ecx, %%ecx\n" + " xgetbv" + : "=a"(*eax), "=d"(*edx)); +} diff --git a/vendor/golang.org/x/sys/cpu/cpu_gccgo.go b/vendor/golang.org/x/sys/cpu/cpu_gccgo.go new file mode 100644 index 000000000..ba49b91bd --- /dev/null +++ b/vendor/golang.org/x/sys/cpu/cpu_gccgo.go @@ -0,0 +1,26 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build 386 amd64 amd64p32 +// +build gccgo + +package cpu + +//extern gccgoGetCpuidCount +func gccgoGetCpuidCount(eaxArg, ecxArg uint32, eax, ebx, ecx, edx *uint32) + +func cpuid(eaxArg, ecxArg uint32) (eax, ebx, ecx, edx uint32) { + var a, b, c, d uint32 + gccgoGetCpuidCount(eaxArg, ecxArg, &a, &b, &c, &d) + return a, b, c, d +} + +//extern gccgoXgetbv +func gccgoXgetbv(eax, edx *uint32) + +func xgetbv() (eax, edx uint32) { + var a, d uint32 + gccgoXgetbv(&a, &d) + return a, d +} diff --git a/vendor/golang.org/x/sys/cpu/cpu_x86.go b/vendor/golang.org/x/sys/cpu/cpu_x86.go index 8842b7cf7..71e288b06 100644 --- a/vendor/golang.org/x/sys/cpu/cpu_x86.go +++ b/vendor/golang.org/x/sys/cpu/cpu_x86.go @@ -8,12 +8,6 @@ package cpu const cacheLineSize = 64 -// cpuid is implemented in cpu_x86.s. -func cpuid(eaxArg, ecxArg uint32) (eax, ebx, ecx, edx uint32) - -// xgetbv with ecx = 0 is implemented in cpu_x86.s. -func xgetbv() (eax, edx uint32) - func init() { maxID, _, _, _ := cpuid(0, 0) diff --git a/vendor/golang.org/x/sys/cpu/cpu_x86.s b/vendor/golang.org/x/sys/cpu/cpu_x86.s index 4ff61e00d..47f084128 100644 --- a/vendor/golang.org/x/sys/cpu/cpu_x86.s +++ b/vendor/golang.org/x/sys/cpu/cpu_x86.s @@ -3,6 +3,7 @@ // license that can be found in the LICENSE file. // +build 386 amd64 amd64p32 +// +build !gccgo #include "textflag.h" diff --git a/vendor/golang.org/x/sys/plan9/mkerrors.sh b/vendor/golang.org/x/sys/plan9/mkerrors.sh index 052c86d94..85309c4a5 100755 --- a/vendor/golang.org/x/sys/plan9/mkerrors.sh +++ b/vendor/golang.org/x/sys/plan9/mkerrors.sh @@ -140,7 +140,7 @@ echo '#include ' | $CC -x c - -E -dM $ccflags | sort >_signal.grep echo '// mkerrors.sh' "$@" -echo '// MACHINE GENERATED BY THE COMMAND ABOVE; DO NOT EDIT' +echo '// Code generated by the command above; DO NOT EDIT.' echo go tool cgo -godefs -- "$@" _const.go >_error.out cat _error.out | grep -vf _error.grep | grep -vf _signal.grep diff --git a/vendor/golang.org/x/sys/plan9/mksyscall.pl b/vendor/golang.org/x/sys/plan9/mksyscall.pl index ce8e1e4f3..252e89e14 100755 --- a/vendor/golang.org/x/sys/plan9/mksyscall.pl +++ b/vendor/golang.org/x/sys/plan9/mksyscall.pl @@ -308,7 +308,7 @@ if($errors) { print < #include #include +#include +#include +#include #include #include #include @@ -57,7 +60,6 @@ package unix #include #include #include -#include #include #include #include @@ -70,6 +72,7 @@ package unix #include #include #include +#include // abi/abi.h generated by mkall.go. #include "abi/abi.h" @@ -138,6 +141,10 @@ struct stat { # define AT_STATX_DONT_SYNC 0x4000 // - Don't sync attributes with the server #endif +#ifndef AT_EACCESS +# define AT_EACCESS 0x200 // Test access permitted for effective IDs, not real IDs. +#endif + #ifdef TCSETS2 // On systems that have "struct termios2" use this as type Termios. typedef struct termios2 termios_t; @@ -289,6 +296,15 @@ struct perf_event_attr_go { __u32 __reserved_2; }; +// ustat is deprecated and glibc 2.28 removed ustat.h. Provide the type here for +// backwards compatibility. Copied from /usr/include/bits/ustat.h +struct ustat { + __daddr_t f_tfree; + __ino_t f_tinode; + char f_fname[6]; + char f_fpack[6]; +}; + */ import "C" @@ -646,6 +662,8 @@ const ( AT_SYMLINK_FOLLOW = C.AT_SYMLINK_FOLLOW AT_SYMLINK_NOFOLLOW = C.AT_SYMLINK_NOFOLLOW + + AT_EACCESS = C.AT_EACCESS ) type PollFd C.struct_pollfd @@ -1021,3 +1039,470 @@ const ( SizeofTpacket2Hdr = C.sizeof_struct_tpacket2_hdr SizeofTpacket3Hdr = C.sizeof_struct_tpacket3_hdr ) + +// netfilter +// generated using: +// perl -nlE '/^\s*(NF\w+)/ && say "$1 = C.$1"' /usr/include/linux/netfilter.h +const ( + NF_INET_PRE_ROUTING = C.NF_INET_PRE_ROUTING + NF_INET_LOCAL_IN = C.NF_INET_LOCAL_IN + NF_INET_FORWARD = C.NF_INET_FORWARD + NF_INET_LOCAL_OUT = C.NF_INET_LOCAL_OUT + NF_INET_POST_ROUTING = C.NF_INET_POST_ROUTING + NF_INET_NUMHOOKS = C.NF_INET_NUMHOOKS +) + +const ( + NF_NETDEV_INGRESS = C.NF_NETDEV_INGRESS + NF_NETDEV_NUMHOOKS = C.NF_NETDEV_NUMHOOKS +) + +const ( + NFPROTO_UNSPEC = C.NFPROTO_UNSPEC + NFPROTO_INET = C.NFPROTO_INET + NFPROTO_IPV4 = C.NFPROTO_IPV4 + NFPROTO_ARP = C.NFPROTO_ARP + NFPROTO_NETDEV = C.NFPROTO_NETDEV + NFPROTO_BRIDGE = C.NFPROTO_BRIDGE + NFPROTO_IPV6 = C.NFPROTO_IPV6 + NFPROTO_DECNET = C.NFPROTO_DECNET + NFPROTO_NUMPROTO = C.NFPROTO_NUMPROTO +) + +// netfilter nfnetlink +type Nfgenmsg C.struct_nfgenmsg + +const ( + NFNL_BATCH_UNSPEC = C.NFNL_BATCH_UNSPEC + NFNL_BATCH_GENID = C.NFNL_BATCH_GENID +) + +// netfilter nf_tables +// generated using: +// perl -nlE '/^\s*(NFT\w+)/ && say "$1 = C.$1"' /usr/include/linux/netfilter/nf_tables.h +const ( + NFT_REG_VERDICT = C.NFT_REG_VERDICT + NFT_REG_1 = C.NFT_REG_1 + NFT_REG_2 = C.NFT_REG_2 + NFT_REG_3 = C.NFT_REG_3 + NFT_REG_4 = C.NFT_REG_4 + NFT_REG32_00 = C.NFT_REG32_00 + NFT_REG32_01 = C.NFT_REG32_01 + NFT_REG32_02 = C.NFT_REG32_02 + NFT_REG32_03 = C.NFT_REG32_03 + NFT_REG32_04 = C.NFT_REG32_04 + NFT_REG32_05 = C.NFT_REG32_05 + NFT_REG32_06 = C.NFT_REG32_06 + NFT_REG32_07 = C.NFT_REG32_07 + NFT_REG32_08 = C.NFT_REG32_08 + NFT_REG32_09 = C.NFT_REG32_09 + NFT_REG32_10 = C.NFT_REG32_10 + NFT_REG32_11 = C.NFT_REG32_11 + NFT_REG32_12 = C.NFT_REG32_12 + NFT_REG32_13 = C.NFT_REG32_13 + NFT_REG32_14 = C.NFT_REG32_14 + NFT_REG32_15 = C.NFT_REG32_15 + NFT_CONTINUE = C.NFT_CONTINUE + NFT_BREAK = C.NFT_BREAK + NFT_JUMP = C.NFT_JUMP + NFT_GOTO = C.NFT_GOTO + NFT_RETURN = C.NFT_RETURN + NFT_MSG_NEWTABLE = C.NFT_MSG_NEWTABLE + NFT_MSG_GETTABLE = C.NFT_MSG_GETTABLE + NFT_MSG_DELTABLE = C.NFT_MSG_DELTABLE + NFT_MSG_NEWCHAIN = C.NFT_MSG_NEWCHAIN + NFT_MSG_GETCHAIN = C.NFT_MSG_GETCHAIN + NFT_MSG_DELCHAIN = C.NFT_MSG_DELCHAIN + NFT_MSG_NEWRULE = C.NFT_MSG_NEWRULE + NFT_MSG_GETRULE = C.NFT_MSG_GETRULE + NFT_MSG_DELRULE = C.NFT_MSG_DELRULE + NFT_MSG_NEWSET = C.NFT_MSG_NEWSET + NFT_MSG_GETSET = C.NFT_MSG_GETSET + NFT_MSG_DELSET = C.NFT_MSG_DELSET + NFT_MSG_NEWSETELEM = C.NFT_MSG_NEWSETELEM + NFT_MSG_GETSETELEM = C.NFT_MSG_GETSETELEM + NFT_MSG_DELSETELEM = C.NFT_MSG_DELSETELEM + NFT_MSG_NEWGEN = C.NFT_MSG_NEWGEN + NFT_MSG_GETGEN = C.NFT_MSG_GETGEN + NFT_MSG_TRACE = C.NFT_MSG_TRACE + NFT_MSG_NEWOBJ = C.NFT_MSG_NEWOBJ + NFT_MSG_GETOBJ = C.NFT_MSG_GETOBJ + NFT_MSG_DELOBJ = C.NFT_MSG_DELOBJ + NFT_MSG_GETOBJ_RESET = C.NFT_MSG_GETOBJ_RESET + NFT_MSG_MAX = C.NFT_MSG_MAX + NFTA_LIST_UNPEC = C.NFTA_LIST_UNPEC + NFTA_LIST_ELEM = C.NFTA_LIST_ELEM + NFTA_HOOK_UNSPEC = C.NFTA_HOOK_UNSPEC + NFTA_HOOK_HOOKNUM = C.NFTA_HOOK_HOOKNUM + NFTA_HOOK_PRIORITY = C.NFTA_HOOK_PRIORITY + NFTA_HOOK_DEV = C.NFTA_HOOK_DEV + NFT_TABLE_F_DORMANT = C.NFT_TABLE_F_DORMANT + NFTA_TABLE_UNSPEC = C.NFTA_TABLE_UNSPEC + NFTA_TABLE_NAME = C.NFTA_TABLE_NAME + NFTA_TABLE_FLAGS = C.NFTA_TABLE_FLAGS + NFTA_TABLE_USE = C.NFTA_TABLE_USE + NFTA_CHAIN_UNSPEC = C.NFTA_CHAIN_UNSPEC + NFTA_CHAIN_TABLE = C.NFTA_CHAIN_TABLE + NFTA_CHAIN_HANDLE = C.NFTA_CHAIN_HANDLE + NFTA_CHAIN_NAME = C.NFTA_CHAIN_NAME + NFTA_CHAIN_HOOK = C.NFTA_CHAIN_HOOK + NFTA_CHAIN_POLICY = C.NFTA_CHAIN_POLICY + NFTA_CHAIN_USE = C.NFTA_CHAIN_USE + NFTA_CHAIN_TYPE = C.NFTA_CHAIN_TYPE + NFTA_CHAIN_COUNTERS = C.NFTA_CHAIN_COUNTERS + NFTA_CHAIN_PAD = C.NFTA_CHAIN_PAD + NFTA_RULE_UNSPEC = C.NFTA_RULE_UNSPEC + NFTA_RULE_TABLE = C.NFTA_RULE_TABLE + NFTA_RULE_CHAIN = C.NFTA_RULE_CHAIN + NFTA_RULE_HANDLE = C.NFTA_RULE_HANDLE + NFTA_RULE_EXPRESSIONS = C.NFTA_RULE_EXPRESSIONS + NFTA_RULE_COMPAT = C.NFTA_RULE_COMPAT + NFTA_RULE_POSITION = C.NFTA_RULE_POSITION + NFTA_RULE_USERDATA = C.NFTA_RULE_USERDATA + NFTA_RULE_PAD = C.NFTA_RULE_PAD + NFTA_RULE_ID = C.NFTA_RULE_ID + NFT_RULE_COMPAT_F_INV = C.NFT_RULE_COMPAT_F_INV + NFT_RULE_COMPAT_F_MASK = C.NFT_RULE_COMPAT_F_MASK + NFTA_RULE_COMPAT_UNSPEC = C.NFTA_RULE_COMPAT_UNSPEC + NFTA_RULE_COMPAT_PROTO = C.NFTA_RULE_COMPAT_PROTO + NFTA_RULE_COMPAT_FLAGS = C.NFTA_RULE_COMPAT_FLAGS + NFT_SET_ANONYMOUS = C.NFT_SET_ANONYMOUS + NFT_SET_CONSTANT = C.NFT_SET_CONSTANT + NFT_SET_INTERVAL = C.NFT_SET_INTERVAL + NFT_SET_MAP = C.NFT_SET_MAP + NFT_SET_TIMEOUT = C.NFT_SET_TIMEOUT + NFT_SET_EVAL = C.NFT_SET_EVAL + NFT_SET_OBJECT = C.NFT_SET_OBJECT + NFT_SET_POL_PERFORMANCE = C.NFT_SET_POL_PERFORMANCE + NFT_SET_POL_MEMORY = C.NFT_SET_POL_MEMORY + NFTA_SET_DESC_UNSPEC = C.NFTA_SET_DESC_UNSPEC + NFTA_SET_DESC_SIZE = C.NFTA_SET_DESC_SIZE + NFTA_SET_UNSPEC = C.NFTA_SET_UNSPEC + NFTA_SET_TABLE = C.NFTA_SET_TABLE + NFTA_SET_NAME = C.NFTA_SET_NAME + NFTA_SET_FLAGS = C.NFTA_SET_FLAGS + NFTA_SET_KEY_TYPE = C.NFTA_SET_KEY_TYPE + NFTA_SET_KEY_LEN = C.NFTA_SET_KEY_LEN + NFTA_SET_DATA_TYPE = C.NFTA_SET_DATA_TYPE + NFTA_SET_DATA_LEN = C.NFTA_SET_DATA_LEN + NFTA_SET_POLICY = C.NFTA_SET_POLICY + NFTA_SET_DESC = C.NFTA_SET_DESC + NFTA_SET_ID = C.NFTA_SET_ID + NFTA_SET_TIMEOUT = C.NFTA_SET_TIMEOUT + NFTA_SET_GC_INTERVAL = C.NFTA_SET_GC_INTERVAL + NFTA_SET_USERDATA = C.NFTA_SET_USERDATA + NFTA_SET_PAD = C.NFTA_SET_PAD + NFTA_SET_OBJ_TYPE = C.NFTA_SET_OBJ_TYPE + NFT_SET_ELEM_INTERVAL_END = C.NFT_SET_ELEM_INTERVAL_END + NFTA_SET_ELEM_UNSPEC = C.NFTA_SET_ELEM_UNSPEC + NFTA_SET_ELEM_KEY = C.NFTA_SET_ELEM_KEY + NFTA_SET_ELEM_DATA = C.NFTA_SET_ELEM_DATA + NFTA_SET_ELEM_FLAGS = C.NFTA_SET_ELEM_FLAGS + NFTA_SET_ELEM_TIMEOUT = C.NFTA_SET_ELEM_TIMEOUT + NFTA_SET_ELEM_EXPIRATION = C.NFTA_SET_ELEM_EXPIRATION + NFTA_SET_ELEM_USERDATA = C.NFTA_SET_ELEM_USERDATA + NFTA_SET_ELEM_EXPR = C.NFTA_SET_ELEM_EXPR + NFTA_SET_ELEM_PAD = C.NFTA_SET_ELEM_PAD + NFTA_SET_ELEM_OBJREF = C.NFTA_SET_ELEM_OBJREF + NFTA_SET_ELEM_LIST_UNSPEC = C.NFTA_SET_ELEM_LIST_UNSPEC + NFTA_SET_ELEM_LIST_TABLE = C.NFTA_SET_ELEM_LIST_TABLE + NFTA_SET_ELEM_LIST_SET = C.NFTA_SET_ELEM_LIST_SET + NFTA_SET_ELEM_LIST_ELEMENTS = C.NFTA_SET_ELEM_LIST_ELEMENTS + NFTA_SET_ELEM_LIST_SET_ID = C.NFTA_SET_ELEM_LIST_SET_ID + NFT_DATA_VALUE = C.NFT_DATA_VALUE + NFT_DATA_VERDICT = C.NFT_DATA_VERDICT + NFTA_DATA_UNSPEC = C.NFTA_DATA_UNSPEC + NFTA_DATA_VALUE = C.NFTA_DATA_VALUE + NFTA_DATA_VERDICT = C.NFTA_DATA_VERDICT + NFTA_VERDICT_UNSPEC = C.NFTA_VERDICT_UNSPEC + NFTA_VERDICT_CODE = C.NFTA_VERDICT_CODE + NFTA_VERDICT_CHAIN = C.NFTA_VERDICT_CHAIN + NFTA_EXPR_UNSPEC = C.NFTA_EXPR_UNSPEC + NFTA_EXPR_NAME = C.NFTA_EXPR_NAME + NFTA_EXPR_DATA = C.NFTA_EXPR_DATA + NFTA_IMMEDIATE_UNSPEC = C.NFTA_IMMEDIATE_UNSPEC + NFTA_IMMEDIATE_DREG = C.NFTA_IMMEDIATE_DREG + NFTA_IMMEDIATE_DATA = C.NFTA_IMMEDIATE_DATA + NFTA_BITWISE_UNSPEC = C.NFTA_BITWISE_UNSPEC + NFTA_BITWISE_SREG = C.NFTA_BITWISE_SREG + NFTA_BITWISE_DREG = C.NFTA_BITWISE_DREG + NFTA_BITWISE_LEN = C.NFTA_BITWISE_LEN + NFTA_BITWISE_MASK = C.NFTA_BITWISE_MASK + NFTA_BITWISE_XOR = C.NFTA_BITWISE_XOR + NFT_BYTEORDER_NTOH = C.NFT_BYTEORDER_NTOH + NFT_BYTEORDER_HTON = C.NFT_BYTEORDER_HTON + NFTA_BYTEORDER_UNSPEC = C.NFTA_BYTEORDER_UNSPEC + NFTA_BYTEORDER_SREG = C.NFTA_BYTEORDER_SREG + NFTA_BYTEORDER_DREG = C.NFTA_BYTEORDER_DREG + NFTA_BYTEORDER_OP = C.NFTA_BYTEORDER_OP + NFTA_BYTEORDER_LEN = C.NFTA_BYTEORDER_LEN + NFTA_BYTEORDER_SIZE = C.NFTA_BYTEORDER_SIZE + NFT_CMP_EQ = C.NFT_CMP_EQ + NFT_CMP_NEQ = C.NFT_CMP_NEQ + NFT_CMP_LT = C.NFT_CMP_LT + NFT_CMP_LTE = C.NFT_CMP_LTE + NFT_CMP_GT = C.NFT_CMP_GT + NFT_CMP_GTE = C.NFT_CMP_GTE + NFTA_CMP_UNSPEC = C.NFTA_CMP_UNSPEC + NFTA_CMP_SREG = C.NFTA_CMP_SREG + NFTA_CMP_OP = C.NFTA_CMP_OP + NFTA_CMP_DATA = C.NFTA_CMP_DATA + NFT_RANGE_EQ = C.NFT_RANGE_EQ + NFT_RANGE_NEQ = C.NFT_RANGE_NEQ + NFTA_RANGE_UNSPEC = C.NFTA_RANGE_UNSPEC + NFTA_RANGE_SREG = C.NFTA_RANGE_SREG + NFTA_RANGE_OP = C.NFTA_RANGE_OP + NFTA_RANGE_FROM_DATA = C.NFTA_RANGE_FROM_DATA + NFTA_RANGE_TO_DATA = C.NFTA_RANGE_TO_DATA + NFT_LOOKUP_F_INV = C.NFT_LOOKUP_F_INV + NFTA_LOOKUP_UNSPEC = C.NFTA_LOOKUP_UNSPEC + NFTA_LOOKUP_SET = C.NFTA_LOOKUP_SET + NFTA_LOOKUP_SREG = C.NFTA_LOOKUP_SREG + NFTA_LOOKUP_DREG = C.NFTA_LOOKUP_DREG + NFTA_LOOKUP_SET_ID = C.NFTA_LOOKUP_SET_ID + NFTA_LOOKUP_FLAGS = C.NFTA_LOOKUP_FLAGS + NFT_DYNSET_OP_ADD = C.NFT_DYNSET_OP_ADD + NFT_DYNSET_OP_UPDATE = C.NFT_DYNSET_OP_UPDATE + NFT_DYNSET_F_INV = C.NFT_DYNSET_F_INV + NFTA_DYNSET_UNSPEC = C.NFTA_DYNSET_UNSPEC + NFTA_DYNSET_SET_NAME = C.NFTA_DYNSET_SET_NAME + NFTA_DYNSET_SET_ID = C.NFTA_DYNSET_SET_ID + NFTA_DYNSET_OP = C.NFTA_DYNSET_OP + NFTA_DYNSET_SREG_KEY = C.NFTA_DYNSET_SREG_KEY + NFTA_DYNSET_SREG_DATA = C.NFTA_DYNSET_SREG_DATA + NFTA_DYNSET_TIMEOUT = C.NFTA_DYNSET_TIMEOUT + NFTA_DYNSET_EXPR = C.NFTA_DYNSET_EXPR + NFTA_DYNSET_PAD = C.NFTA_DYNSET_PAD + NFTA_DYNSET_FLAGS = C.NFTA_DYNSET_FLAGS + NFT_PAYLOAD_LL_HEADER = C.NFT_PAYLOAD_LL_HEADER + NFT_PAYLOAD_NETWORK_HEADER = C.NFT_PAYLOAD_NETWORK_HEADER + NFT_PAYLOAD_TRANSPORT_HEADER = C.NFT_PAYLOAD_TRANSPORT_HEADER + NFT_PAYLOAD_CSUM_NONE = C.NFT_PAYLOAD_CSUM_NONE + NFT_PAYLOAD_CSUM_INET = C.NFT_PAYLOAD_CSUM_INET + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = C.NFT_PAYLOAD_L4CSUM_PSEUDOHDR + NFTA_PAYLOAD_UNSPEC = C.NFTA_PAYLOAD_UNSPEC + NFTA_PAYLOAD_DREG = C.NFTA_PAYLOAD_DREG + NFTA_PAYLOAD_BASE = C.NFTA_PAYLOAD_BASE + NFTA_PAYLOAD_OFFSET = C.NFTA_PAYLOAD_OFFSET + NFTA_PAYLOAD_LEN = C.NFTA_PAYLOAD_LEN + NFTA_PAYLOAD_SREG = C.NFTA_PAYLOAD_SREG + NFTA_PAYLOAD_CSUM_TYPE = C.NFTA_PAYLOAD_CSUM_TYPE + NFTA_PAYLOAD_CSUM_OFFSET = C.NFTA_PAYLOAD_CSUM_OFFSET + NFTA_PAYLOAD_CSUM_FLAGS = C.NFTA_PAYLOAD_CSUM_FLAGS + NFT_EXTHDR_F_PRESENT = C.NFT_EXTHDR_F_PRESENT + NFT_EXTHDR_OP_IPV6 = C.NFT_EXTHDR_OP_IPV6 + NFT_EXTHDR_OP_TCPOPT = C.NFT_EXTHDR_OP_TCPOPT + NFTA_EXTHDR_UNSPEC = C.NFTA_EXTHDR_UNSPEC + NFTA_EXTHDR_DREG = C.NFTA_EXTHDR_DREG + NFTA_EXTHDR_TYPE = C.NFTA_EXTHDR_TYPE + NFTA_EXTHDR_OFFSET = C.NFTA_EXTHDR_OFFSET + NFTA_EXTHDR_LEN = C.NFTA_EXTHDR_LEN + NFTA_EXTHDR_FLAGS = C.NFTA_EXTHDR_FLAGS + NFTA_EXTHDR_OP = C.NFTA_EXTHDR_OP + NFTA_EXTHDR_SREG = C.NFTA_EXTHDR_SREG + NFT_META_LEN = C.NFT_META_LEN + NFT_META_PROTOCOL = C.NFT_META_PROTOCOL + NFT_META_PRIORITY = C.NFT_META_PRIORITY + NFT_META_MARK = C.NFT_META_MARK + NFT_META_IIF = C.NFT_META_IIF + NFT_META_OIF = C.NFT_META_OIF + NFT_META_IIFNAME = C.NFT_META_IIFNAME + NFT_META_OIFNAME = C.NFT_META_OIFNAME + NFT_META_IIFTYPE = C.NFT_META_IIFTYPE + NFT_META_OIFTYPE = C.NFT_META_OIFTYPE + NFT_META_SKUID = C.NFT_META_SKUID + NFT_META_SKGID = C.NFT_META_SKGID + NFT_META_NFTRACE = C.NFT_META_NFTRACE + NFT_META_RTCLASSID = C.NFT_META_RTCLASSID + NFT_META_SECMARK = C.NFT_META_SECMARK + NFT_META_NFPROTO = C.NFT_META_NFPROTO + NFT_META_L4PROTO = C.NFT_META_L4PROTO + NFT_META_BRI_IIFNAME = C.NFT_META_BRI_IIFNAME + NFT_META_BRI_OIFNAME = C.NFT_META_BRI_OIFNAME + NFT_META_PKTTYPE = C.NFT_META_PKTTYPE + NFT_META_CPU = C.NFT_META_CPU + NFT_META_IIFGROUP = C.NFT_META_IIFGROUP + NFT_META_OIFGROUP = C.NFT_META_OIFGROUP + NFT_META_CGROUP = C.NFT_META_CGROUP + NFT_META_PRANDOM = C.NFT_META_PRANDOM + NFT_RT_CLASSID = C.NFT_RT_CLASSID + NFT_RT_NEXTHOP4 = C.NFT_RT_NEXTHOP4 + NFT_RT_NEXTHOP6 = C.NFT_RT_NEXTHOP6 + NFT_RT_TCPMSS = C.NFT_RT_TCPMSS + NFT_HASH_JENKINS = C.NFT_HASH_JENKINS + NFT_HASH_SYM = C.NFT_HASH_SYM + NFTA_HASH_UNSPEC = C.NFTA_HASH_UNSPEC + NFTA_HASH_SREG = C.NFTA_HASH_SREG + NFTA_HASH_DREG = C.NFTA_HASH_DREG + NFTA_HASH_LEN = C.NFTA_HASH_LEN + NFTA_HASH_MODULUS = C.NFTA_HASH_MODULUS + NFTA_HASH_SEED = C.NFTA_HASH_SEED + NFTA_HASH_OFFSET = C.NFTA_HASH_OFFSET + NFTA_HASH_TYPE = C.NFTA_HASH_TYPE + NFTA_META_UNSPEC = C.NFTA_META_UNSPEC + NFTA_META_DREG = C.NFTA_META_DREG + NFTA_META_KEY = C.NFTA_META_KEY + NFTA_META_SREG = C.NFTA_META_SREG + NFTA_RT_UNSPEC = C.NFTA_RT_UNSPEC + NFTA_RT_DREG = C.NFTA_RT_DREG + NFTA_RT_KEY = C.NFTA_RT_KEY + NFT_CT_STATE = C.NFT_CT_STATE + NFT_CT_DIRECTION = C.NFT_CT_DIRECTION + NFT_CT_STATUS = C.NFT_CT_STATUS + NFT_CT_MARK = C.NFT_CT_MARK + NFT_CT_SECMARK = C.NFT_CT_SECMARK + NFT_CT_EXPIRATION = C.NFT_CT_EXPIRATION + NFT_CT_HELPER = C.NFT_CT_HELPER + NFT_CT_L3PROTOCOL = C.NFT_CT_L3PROTOCOL + NFT_CT_SRC = C.NFT_CT_SRC + NFT_CT_DST = C.NFT_CT_DST + NFT_CT_PROTOCOL = C.NFT_CT_PROTOCOL + NFT_CT_PROTO_SRC = C.NFT_CT_PROTO_SRC + NFT_CT_PROTO_DST = C.NFT_CT_PROTO_DST + NFT_CT_LABELS = C.NFT_CT_LABELS + NFT_CT_PKTS = C.NFT_CT_PKTS + NFT_CT_BYTES = C.NFT_CT_BYTES + NFT_CT_AVGPKT = C.NFT_CT_AVGPKT + NFT_CT_ZONE = C.NFT_CT_ZONE + NFT_CT_EVENTMASK = C.NFT_CT_EVENTMASK + NFTA_CT_UNSPEC = C.NFTA_CT_UNSPEC + NFTA_CT_DREG = C.NFTA_CT_DREG + NFTA_CT_KEY = C.NFTA_CT_KEY + NFTA_CT_DIRECTION = C.NFTA_CT_DIRECTION + NFTA_CT_SREG = C.NFTA_CT_SREG + NFT_LIMIT_PKTS = C.NFT_LIMIT_PKTS + NFT_LIMIT_PKT_BYTES = C.NFT_LIMIT_PKT_BYTES + NFT_LIMIT_F_INV = C.NFT_LIMIT_F_INV + NFTA_LIMIT_UNSPEC = C.NFTA_LIMIT_UNSPEC + NFTA_LIMIT_RATE = C.NFTA_LIMIT_RATE + NFTA_LIMIT_UNIT = C.NFTA_LIMIT_UNIT + NFTA_LIMIT_BURST = C.NFTA_LIMIT_BURST + NFTA_LIMIT_TYPE = C.NFTA_LIMIT_TYPE + NFTA_LIMIT_FLAGS = C.NFTA_LIMIT_FLAGS + NFTA_LIMIT_PAD = C.NFTA_LIMIT_PAD + NFTA_COUNTER_UNSPEC = C.NFTA_COUNTER_UNSPEC + NFTA_COUNTER_BYTES = C.NFTA_COUNTER_BYTES + NFTA_COUNTER_PACKETS = C.NFTA_COUNTER_PACKETS + NFTA_COUNTER_PAD = C.NFTA_COUNTER_PAD + NFTA_LOG_UNSPEC = C.NFTA_LOG_UNSPEC + NFTA_LOG_GROUP = C.NFTA_LOG_GROUP + NFTA_LOG_PREFIX = C.NFTA_LOG_PREFIX + NFTA_LOG_SNAPLEN = C.NFTA_LOG_SNAPLEN + NFTA_LOG_QTHRESHOLD = C.NFTA_LOG_QTHRESHOLD + NFTA_LOG_LEVEL = C.NFTA_LOG_LEVEL + NFTA_LOG_FLAGS = C.NFTA_LOG_FLAGS + NFTA_QUEUE_UNSPEC = C.NFTA_QUEUE_UNSPEC + NFTA_QUEUE_NUM = C.NFTA_QUEUE_NUM + NFTA_QUEUE_TOTAL = C.NFTA_QUEUE_TOTAL + NFTA_QUEUE_FLAGS = C.NFTA_QUEUE_FLAGS + NFTA_QUEUE_SREG_QNUM = C.NFTA_QUEUE_SREG_QNUM + NFT_QUOTA_F_INV = C.NFT_QUOTA_F_INV + NFT_QUOTA_F_DEPLETED = C.NFT_QUOTA_F_DEPLETED + NFTA_QUOTA_UNSPEC = C.NFTA_QUOTA_UNSPEC + NFTA_QUOTA_BYTES = C.NFTA_QUOTA_BYTES + NFTA_QUOTA_FLAGS = C.NFTA_QUOTA_FLAGS + NFTA_QUOTA_PAD = C.NFTA_QUOTA_PAD + NFTA_QUOTA_CONSUMED = C.NFTA_QUOTA_CONSUMED + NFT_REJECT_ICMP_UNREACH = C.NFT_REJECT_ICMP_UNREACH + NFT_REJECT_TCP_RST = C.NFT_REJECT_TCP_RST + NFT_REJECT_ICMPX_UNREACH = C.NFT_REJECT_ICMPX_UNREACH + NFT_REJECT_ICMPX_NO_ROUTE = C.NFT_REJECT_ICMPX_NO_ROUTE + NFT_REJECT_ICMPX_PORT_UNREACH = C.NFT_REJECT_ICMPX_PORT_UNREACH + NFT_REJECT_ICMPX_HOST_UNREACH = C.NFT_REJECT_ICMPX_HOST_UNREACH + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = C.NFT_REJECT_ICMPX_ADMIN_PROHIBITED + NFTA_REJECT_UNSPEC = C.NFTA_REJECT_UNSPEC + NFTA_REJECT_TYPE = C.NFTA_REJECT_TYPE + NFTA_REJECT_ICMP_CODE = C.NFTA_REJECT_ICMP_CODE + NFT_NAT_SNAT = C.NFT_NAT_SNAT + NFT_NAT_DNAT = C.NFT_NAT_DNAT + NFTA_NAT_UNSPEC = C.NFTA_NAT_UNSPEC + NFTA_NAT_TYPE = C.NFTA_NAT_TYPE + NFTA_NAT_FAMILY = C.NFTA_NAT_FAMILY + NFTA_NAT_REG_ADDR_MIN = C.NFTA_NAT_REG_ADDR_MIN + NFTA_NAT_REG_ADDR_MAX = C.NFTA_NAT_REG_ADDR_MAX + NFTA_NAT_REG_PROTO_MIN = C.NFTA_NAT_REG_PROTO_MIN + NFTA_NAT_REG_PROTO_MAX = C.NFTA_NAT_REG_PROTO_MAX + NFTA_NAT_FLAGS = C.NFTA_NAT_FLAGS + NFTA_MASQ_UNSPEC = C.NFTA_MASQ_UNSPEC + NFTA_MASQ_FLAGS = C.NFTA_MASQ_FLAGS + NFTA_MASQ_REG_PROTO_MIN = C.NFTA_MASQ_REG_PROTO_MIN + NFTA_MASQ_REG_PROTO_MAX = C.NFTA_MASQ_REG_PROTO_MAX + NFTA_REDIR_UNSPEC = C.NFTA_REDIR_UNSPEC + NFTA_REDIR_REG_PROTO_MIN = C.NFTA_REDIR_REG_PROTO_MIN + NFTA_REDIR_REG_PROTO_MAX = C.NFTA_REDIR_REG_PROTO_MAX + NFTA_REDIR_FLAGS = C.NFTA_REDIR_FLAGS + NFTA_DUP_UNSPEC = C.NFTA_DUP_UNSPEC + NFTA_DUP_SREG_ADDR = C.NFTA_DUP_SREG_ADDR + NFTA_DUP_SREG_DEV = C.NFTA_DUP_SREG_DEV + NFTA_FWD_UNSPEC = C.NFTA_FWD_UNSPEC + NFTA_FWD_SREG_DEV = C.NFTA_FWD_SREG_DEV + NFTA_OBJREF_UNSPEC = C.NFTA_OBJREF_UNSPEC + NFTA_OBJREF_IMM_TYPE = C.NFTA_OBJREF_IMM_TYPE + NFTA_OBJREF_IMM_NAME = C.NFTA_OBJREF_IMM_NAME + NFTA_OBJREF_SET_SREG = C.NFTA_OBJREF_SET_SREG + NFTA_OBJREF_SET_NAME = C.NFTA_OBJREF_SET_NAME + NFTA_OBJREF_SET_ID = C.NFTA_OBJREF_SET_ID + NFTA_GEN_UNSPEC = C.NFTA_GEN_UNSPEC + NFTA_GEN_ID = C.NFTA_GEN_ID + NFTA_GEN_PROC_PID = C.NFTA_GEN_PROC_PID + NFTA_GEN_PROC_NAME = C.NFTA_GEN_PROC_NAME + NFTA_FIB_UNSPEC = C.NFTA_FIB_UNSPEC + NFTA_FIB_DREG = C.NFTA_FIB_DREG + NFTA_FIB_RESULT = C.NFTA_FIB_RESULT + NFTA_FIB_FLAGS = C.NFTA_FIB_FLAGS + NFT_FIB_RESULT_UNSPEC = C.NFT_FIB_RESULT_UNSPEC + NFT_FIB_RESULT_OIF = C.NFT_FIB_RESULT_OIF + NFT_FIB_RESULT_OIFNAME = C.NFT_FIB_RESULT_OIFNAME + NFT_FIB_RESULT_ADDRTYPE = C.NFT_FIB_RESULT_ADDRTYPE + NFTA_FIB_F_SADDR = C.NFTA_FIB_F_SADDR + NFTA_FIB_F_DADDR = C.NFTA_FIB_F_DADDR + NFTA_FIB_F_MARK = C.NFTA_FIB_F_MARK + NFTA_FIB_F_IIF = C.NFTA_FIB_F_IIF + NFTA_FIB_F_OIF = C.NFTA_FIB_F_OIF + NFTA_FIB_F_PRESENT = C.NFTA_FIB_F_PRESENT + NFTA_CT_HELPER_UNSPEC = C.NFTA_CT_HELPER_UNSPEC + NFTA_CT_HELPER_NAME = C.NFTA_CT_HELPER_NAME + NFTA_CT_HELPER_L3PROTO = C.NFTA_CT_HELPER_L3PROTO + NFTA_CT_HELPER_L4PROTO = C.NFTA_CT_HELPER_L4PROTO + NFTA_OBJ_UNSPEC = C.NFTA_OBJ_UNSPEC + NFTA_OBJ_TABLE = C.NFTA_OBJ_TABLE + NFTA_OBJ_NAME = C.NFTA_OBJ_NAME + NFTA_OBJ_TYPE = C.NFTA_OBJ_TYPE + NFTA_OBJ_DATA = C.NFTA_OBJ_DATA + NFTA_OBJ_USE = C.NFTA_OBJ_USE + NFTA_TRACE_UNSPEC = C.NFTA_TRACE_UNSPEC + NFTA_TRACE_TABLE = C.NFTA_TRACE_TABLE + NFTA_TRACE_CHAIN = C.NFTA_TRACE_CHAIN + NFTA_TRACE_RULE_HANDLE = C.NFTA_TRACE_RULE_HANDLE + NFTA_TRACE_TYPE = C.NFTA_TRACE_TYPE + NFTA_TRACE_VERDICT = C.NFTA_TRACE_VERDICT + NFTA_TRACE_ID = C.NFTA_TRACE_ID + NFTA_TRACE_LL_HEADER = C.NFTA_TRACE_LL_HEADER + NFTA_TRACE_NETWORK_HEADER = C.NFTA_TRACE_NETWORK_HEADER + NFTA_TRACE_TRANSPORT_HEADER = C.NFTA_TRACE_TRANSPORT_HEADER + NFTA_TRACE_IIF = C.NFTA_TRACE_IIF + NFTA_TRACE_IIFTYPE = C.NFTA_TRACE_IIFTYPE + NFTA_TRACE_OIF = C.NFTA_TRACE_OIF + NFTA_TRACE_OIFTYPE = C.NFTA_TRACE_OIFTYPE + NFTA_TRACE_MARK = C.NFTA_TRACE_MARK + NFTA_TRACE_NFPROTO = C.NFTA_TRACE_NFPROTO + NFTA_TRACE_POLICY = C.NFTA_TRACE_POLICY + NFTA_TRACE_PAD = C.NFTA_TRACE_PAD + NFT_TRACETYPE_UNSPEC = C.NFT_TRACETYPE_UNSPEC + NFT_TRACETYPE_POLICY = C.NFT_TRACETYPE_POLICY + NFT_TRACETYPE_RETURN = C.NFT_TRACETYPE_RETURN + NFT_TRACETYPE_RULE = C.NFT_TRACETYPE_RULE + NFTA_NG_UNSPEC = C.NFTA_NG_UNSPEC + NFTA_NG_DREG = C.NFTA_NG_DREG + NFTA_NG_MODULUS = C.NFTA_NG_MODULUS + NFTA_NG_TYPE = C.NFTA_NG_TYPE + NFTA_NG_OFFSET = C.NFTA_NG_OFFSET + NFT_NG_INCREMENTAL = C.NFT_NG_INCREMENTAL + NFT_NG_RANDOM = C.NFT_NG_RANDOM +) + +type RTCTime C.struct_rtc_time + +type RTCWkAlrm C.struct_rtc_wkalrm + +type RTCPLLInfo C.struct_rtc_pll_info diff --git a/vendor/golang.org/x/sys/unix/mkerrors.sh b/vendor/golang.org/x/sys/unix/mkerrors.sh index 374195bf8..4a2c5dc9b 100755 --- a/vendor/golang.org/x/sys/unix/mkerrors.sh +++ b/vendor/golang.org/x/sys/unix/mkerrors.sh @@ -50,6 +50,7 @@ includes_Darwin=' #include #include #include +#include #include #include #include @@ -172,6 +173,7 @@ struct ltchars { #include #include #include +#include #include #include #include @@ -191,6 +193,7 @@ struct ltchars { #include #include #include +#include #include #include @@ -402,7 +405,7 @@ ccflags="$@" $2 ~ /^LINUX_REBOOT_CMD_/ || $2 ~ /^LINUX_REBOOT_MAGIC[12]$/ || $2 !~ "NLA_TYPE_MASK" && - $2 ~ /^(NETLINK|NLM|NLMSG|NLA|IFA|IFAN|RT|RTCF|RTN|RTPROT|RTNH|ARPHRD|ETH_P)_/ || + $2 ~ /^(NETLINK|NLM|NLMSG|NLA|IFA|IFAN|RT|RTC|RTCF|RTN|RTPROT|RTNH|ARPHRD|ETH_P)_/ || $2 ~ /^SIOC/ || $2 ~ /^TIOC/ || $2 ~ /^TCGET/ || @@ -428,6 +431,7 @@ ccflags="$@" $2 ~ /^PERF_EVENT_IOC_/ || $2 ~ /^SECCOMP_MODE_/ || $2 ~ /^SPLICE_/ || + $2 !~ /^AUDIT_RECORD_MAGIC/ && $2 ~ /^[A-Z0-9_]+_MAGIC2?$/ || $2 ~ /^(VM|VMADDR)_/ || $2 ~ /^IOCTL_VM_SOCKETS_/ || @@ -436,10 +440,11 @@ ccflags="$@" $2 ~ /^GENL_/ || $2 ~ /^STATX_/ || $2 ~ /^UTIME_/ || - $2 ~ /^XATTR_(CREATE|REPLACE)/ || + $2 ~ /^XATTR_(CREATE|REPLACE|NO(DEFAULT|FOLLOW|SECURITY)|SHOWCOMPRESSION)/ || $2 ~ /^ATTR_(BIT_MAP_COUNT|(CMN|VOL|FILE)_)/ || $2 ~ /^FSOPT_/ || $2 ~ /^WDIOC_/ || + $2 ~ /^NFN/ || $2 ~ /^(HDIO|WIN|SMART)_/ || $2 !~ "WMESGLEN" && $2 ~ /^W[A-Z0-9]+$/ || diff --git a/vendor/golang.org/x/sys/unix/mksysctl_openbsd.pl b/vendor/golang.org/x/sys/unix/mksysctl_openbsd.pl index be67afa41..49f186f83 100755 --- a/vendor/golang.org/x/sys/unix/mksysctl_openbsd.pl +++ b/vendor/golang.org/x/sys/unix/mksysctl_openbsd.pl @@ -240,7 +240,7 @@ foreach my $header (@headers) { print < 0 { + destp = &dest[0] + } + return destp +} + +//sys getxattr(path string, attr string, dest *byte, size int, position uint32, options int) (sz int, err error) + +func Getxattr(path string, attr string, dest []byte) (sz int, err error) { + return getxattr(path, attr, xattrPointer(dest), len(dest), 0, 0) +} + +func Lgetxattr(link string, attr string, dest []byte) (sz int, err error) { + return getxattr(link, attr, xattrPointer(dest), len(dest), 0, XATTR_NOFOLLOW) +} + +//sys setxattr(path string, attr string, data *byte, size int, position uint32, options int) (err error) + +func Setxattr(path string, attr string, data []byte, flags int) (err error) { + // The parameters for the OS X implementation vary slightly compared to the + // linux system call, specifically the position parameter: + // + // linux: + // int setxattr( + // const char *path, + // const char *name, + // const void *value, + // size_t size, + // int flags + // ); + // + // darwin: + // int setxattr( + // const char *path, + // const char *name, + // void *value, + // size_t size, + // u_int32_t position, + // int options + // ); + // + // position specifies the offset within the extended attribute. In the + // current implementation, only the resource fork extended attribute makes + // use of this argument. For all others, position is reserved. We simply + // default to setting it to zero. + return setxattr(path, attr, xattrPointer(data), len(data), 0, flags) +} + +func Lsetxattr(link string, attr string, data []byte, flags int) (err error) { + return setxattr(link, attr, xattrPointer(data), len(data), 0, flags|XATTR_NOFOLLOW) +} + +//sys removexattr(path string, attr string, options int) (err error) + +func Removexattr(path string, attr string) (err error) { + // We wrap around and explicitly zero out the options provided to the OS X + // implementation of removexattr, we do so for interoperability with the + // linux variant. + return removexattr(path, attr, 0) +} + +func Lremovexattr(link string, attr string) (err error) { + return removexattr(link, attr, XATTR_NOFOLLOW) +} + +//sys listxattr(path string, dest *byte, size int, options int) (sz int, err error) + +func Listxattr(path string, dest []byte) (sz int, err error) { + return listxattr(path, xattrPointer(dest), len(dest), 0) +} + +func Llistxattr(link string, dest []byte) (sz int, err error) { + return listxattr(link, xattrPointer(dest), len(dest), XATTR_NOFOLLOW) +} + func setattrlistTimes(path string, times []Timespec, flags int) error { _p0, err := BytePtrFromString(path) if err != nil { @@ -447,13 +529,9 @@ func Uname(uname *Utsname) error { // Watchevent // Waitevent // Modwatch -// Getxattr // Fgetxattr -// Setxattr // Fsetxattr -// Removexattr // Fremovexattr -// Listxattr // Flistxattr // Fsctl // Initgroups diff --git a/vendor/golang.org/x/sys/unix/syscall_darwin_test.go b/vendor/golang.org/x/sys/unix/syscall_darwin_test.go new file mode 100644 index 000000000..65691d5c1 --- /dev/null +++ b/vendor/golang.org/x/sys/unix/syscall_darwin_test.go @@ -0,0 +1,19 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +package unix_test + +// stringsFromByteSlice converts a sequence of attributes to a []string. +// On Darwin, each entry is a NULL-terminated string. +func stringsFromByteSlice(buf []byte) []string { + var result []string + off := 0 + for i, b := range buf { + if b == 0 { + result = append(result, string(buf[off:i])) + off = i + 1 + } + } + return result +} diff --git a/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go b/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go index 654439e02..0fec1a827 100644 --- a/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go +++ b/vendor/golang.org/x/sys/unix/syscall_freebsd_test.go @@ -295,3 +295,18 @@ func TestCapRightsSetAndClear(t *testing.T) { t.Fatalf("Wrong rights set") } } + +// stringsFromByteSlice converts a sequence of attributes to a []string. +// On FreeBSD, each entry consists of a single byte containing the length +// of the attribute name, followed by the attribute name. +// The name is _not_ NULL-terminated. +func stringsFromByteSlice(buf []byte) []string { + var result []string + i := 0 + for i < len(buf) { + next := i + 1 + int(buf[i]) + result = append(result, string(buf[i+1:next])) + i = next + } + return result +} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux.go b/vendor/golang.org/x/sys/unix/syscall_linux.go index 04f38c53e..9908030cb 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux.go @@ -148,8 +148,6 @@ func Unlink(path string) error { //sys Unlinkat(dirfd int, path string, flags int) (err error) -//sys utimes(path string, times *[2]Timeval) (err error) - func Utimes(path string, tv []Timeval) error { if tv == nil { err := utimensat(AT_FDCWD, path, nil, 0) @@ -207,20 +205,14 @@ func UtimesNanoAt(dirfd int, path string, ts []Timespec, flags int) error { return utimensat(dirfd, path, (*[2]Timespec)(unsafe.Pointer(&ts[0])), flags) } -//sys futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) - func Futimesat(dirfd int, path string, tv []Timeval) error { - pathp, err := BytePtrFromString(path) - if err != nil { - return err - } if tv == nil { - return futimesat(dirfd, pathp, nil) + return futimesat(dirfd, path, nil) } if len(tv) != 2 { return EINVAL } - return futimesat(dirfd, pathp, (*[2]Timeval)(unsafe.Pointer(&tv[0]))) + return futimesat(dirfd, path, (*[2]Timeval)(unsafe.Pointer(&tv[0]))) } func Futimes(fd int, tv []Timeval) (err error) { @@ -1221,12 +1213,10 @@ func Mount(source string, target string, fstype string, flags uintptr, data stri //sys CopyFileRange(rfd int, roff *int64, wfd int, woff *int64, len int, flags int) (n int, err error) //sys Dup(oldfd int) (fd int, err error) //sys Dup3(oldfd int, newfd int, flags int) (err error) -//sysnb EpollCreate(size int) (fd int, err error) //sysnb EpollCreate1(flag int) (fd int, err error) //sysnb EpollCtl(epfd int, op int, fd int, event *EpollEvent) (err error) //sys Eventfd(initval uint, flags int) (fd int, err error) = SYS_EVENTFD2 //sys Exit(code int) = SYS_EXIT_GROUP -//sys Faccessat(dirfd int, path string, mode uint32, flags int) (err error) //sys Fallocate(fd int, mode uint32, off int64, len int64) (err error) //sys Fchdir(fd int) (err error) //sys Fchmod(fd int, mode uint32) (err error) @@ -1306,7 +1296,6 @@ func Setgid(uid int) (err error) { //sysnb Uname(buf *Utsname) (err error) //sys Unmount(target string, flags int) (err error) = SYS_UMOUNT2 //sys Unshare(flags int) (err error) -//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sys write(fd int, p []byte) (n int, err error) //sys exitThread(code int) (err error) = SYS_EXIT //sys readlen(fd int, p *byte, np int) (n int, err error) = SYS_READ @@ -1356,6 +1345,17 @@ func Vmsplice(fd int, iovs []Iovec, flags int) (int, error) { return int(n), nil } +//sys faccessat(dirfd int, path string, mode uint32) (err error) + +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + if flags & ^(AT_SYMLINK_NOFOLLOW|AT_EACCESS) != 0 { + return EINVAL + } else if flags&(AT_SYMLINK_NOFOLLOW|AT_EACCESS) != 0 { + return EOPNOTSUPP + } + return faccessat(dirfd, path, mode) +} + /* * Unimplemented */ diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_386.go b/vendor/golang.org/x/sys/unix/syscall_linux_386.go index bb8e4fbd8..74bc098ce 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_386.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_386.go @@ -10,7 +10,6 @@ package unix import ( - "syscall" "unsafe" ) @@ -51,6 +50,8 @@ func Pipe2(p []int, flags int) (err error) { // 64-bit file system and 32-bit uid calls // (386 default is 32-bit file system and 16-bit uid). //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) +//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64_64 //sys Fchown(fd int, uid int, gid int) (err error) = SYS_FCHOWN32 //sys Fstat(fd int, stat *Stat_t) (err error) = SYS_FSTAT64 @@ -78,12 +79,12 @@ func Pipe2(p []int, flags int) (err error) { //sys Stat(path string, stat *Stat_t) (err error) = SYS_STAT64 //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) //sys Truncate(path string, length int64) (err error) = SYS_TRUNCATE64 +//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sysnb getgroups(n int, list *_Gid_t) (nn int, err error) = SYS_GETGROUPS32 //sysnb setgroups(n int, list *_Gid_t) (err error) = SYS_SETGROUPS32 //sys Select(nfd int, r *FdSet, w *FdSet, e *FdSet, timeout *Timeval) (n int, err error) = SYS__NEWSELECT //sys mmap2(addr uintptr, length uintptr, prot int, flags int, fd int, pageOffset uintptr) (xaddr uintptr, err error) -//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Pause() (err error) func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int64) (xaddr uintptr, err error) { @@ -157,10 +158,6 @@ func Setrlimit(resource int, rlim *Rlimit) (err error) { return setrlimit(resource, &rl) } -// Underlying system call writes to newoffset via pointer. -// Implemented in assembly to avoid allocation. -func seek(fd int, offset int64, whence int) (newoffset int64, err syscall.Errno) - func Seek(fd int, offset int64, whence int) (newoffset int64, err error) { newoffset, errno := seek(fd, offset, whence) if errno != 0 { @@ -169,11 +166,11 @@ func Seek(fd int, offset int64, whence int) (newoffset int64, err error) { return newoffset, nil } -// Vsyscalls on amd64. +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) //sysnb Time(t *Time_t) (tt Time_t, err error) - //sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) // On x86 Linux, all the socket calls go through an extra indirection, // I think because the 5-register system call interface can't handle @@ -206,9 +203,6 @@ const ( _SENDMMSG = 20 ) -func socketcall(call int, a0, a1, a2, a3, a4, a5 uintptr) (n int, err syscall.Errno) -func rawsocketcall(call int, a0, a1, a2, a3, a4, a5 uintptr) (n int, err syscall.Errno) - func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { fd, e := socketcall(_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen)), 0, 0, 0) if e != 0 { diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go b/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go index d12110632..5f9b2454a 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_amd64.go @@ -7,6 +7,7 @@ package unix //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 //sys Fchown(fd int, uid int, gid int) (err error) @@ -57,6 +58,7 @@ func Stat(path string, stat *Stat_t) (err error) { //sys Statfs(path string, buf *Statfs_t) (err error) //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) //sys Truncate(path string, length int64) (err error) +//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sys accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) //sys accept4(s int, rsa *RawSockaddrAny, addrlen *_Socklen, flags int) (fd int, err error) //sys bind(s int, addr unsafe.Pointer, addrlen _Socklen) (err error) @@ -75,6 +77,8 @@ func Stat(path string, stat *Stat_t) (err error) { //sys sendmsg(s int, msg *Msghdr, flags int) (n int, err error) //sys mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int64) (xaddr uintptr, err error) +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) + func Gettimeofday(tv *Timeval) (err error) { errno := gettimeofday(tv) if errno != 0 { @@ -96,6 +100,7 @@ func Time(t *Time_t) (tt Time_t, err error) { } //sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) func setTimespec(sec, nsec int64) Timespec { return Timespec{Sec: sec, Nsec: nsec} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_arm.go b/vendor/golang.org/x/sys/unix/syscall_linux_arm.go index c59f8588f..3ec7a9329 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_arm.go @@ -75,6 +75,8 @@ func Seek(fd int, offset int64, whence int) (newoffset int64, err error) { // 64-bit file system and 32-bit uid calls // (16-bit uid calls are not always supported in newer kernels) //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) +//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fchown(fd int, uid int, gid int) (err error) = SYS_FCHOWN32 //sys Fstat(fd int, stat *Stat_t) (err error) = SYS_FSTAT64 //sys Fstatat(dirfd int, path string, stat *Stat_t, flags int) (err error) = SYS_FSTATAT64 @@ -86,6 +88,7 @@ func Seek(fd int, offset int64, whence int) (newoffset int64, err error) { //sys Lchown(path string, uid int, gid int) (err error) = SYS_LCHOWN32 //sys Listen(s int, n int) (err error) //sys Lstat(path string, stat *Stat_t) (err error) = SYS_LSTAT64 +//sys Pause() (err error) //sys sendfile(outfd int, infd int, offset *int64, count int) (written int, err error) = SYS_SENDFILE64 //sys Select(nfd int, r *FdSet, w *FdSet, e *FdSet, timeout *Timeval) (n int, err error) = SYS__NEWSELECT //sys Setfsgid(gid int) (err error) = SYS_SETFSGID32 @@ -97,11 +100,10 @@ func Seek(fd int, offset int64, whence int) (newoffset int64, err error) { //sys Shutdown(fd int, how int) (err error) //sys Splice(rfd int, roff *int64, wfd int, woff *int64, len int, flags int) (n int, err error) //sys Stat(path string, stat *Stat_t) (err error) = SYS_STAT64 +//sys Ustat(dev int, ubuf *Ustat_t) (err error) -// Vsyscalls on amd64. +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) -//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) -//sys Pause() (err error) func Time(t *Time_t) (Time_t, error) { var tv Timeval @@ -123,6 +125,8 @@ func Utime(path string, buf *Utimbuf) error { return Utimes(path, tv) } +//sys utimes(path string, times *[2]Timeval) (err error) + //sys Pread(fd int, p []byte, offset int64) (n int, err error) = SYS_PREAD64 //sys Pwrite(fd int, p []byte, offset int64) (n int, err error) = SYS_PWRITE64 //sys Truncate(path string, length int64) (err error) = SYS_TRUNCATE64 diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go b/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go index a1e8a609b..646f295ad 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_arm64.go @@ -6,6 +6,15 @@ package unix +import "unsafe" + +func EpollCreate(size int) (fd int, err error) { + if size <= 0 { + return -1, EINVAL + } + return EpollCreate1(0) +} + //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) = SYS_EPOLL_PWAIT //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 //sys Fchown(fd int, uid int, gid int) (err error) @@ -57,6 +66,11 @@ func Lstat(path string, stat *Stat_t) (err error) { //sys Statfs(path string, buf *Statfs_t) (err error) //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) //sys Truncate(path string, length int64) (err error) + +func Ustat(dev int, ubuf *Ustat_t) (err error) { + return ENOSYS +} + //sys accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) //sys accept4(s int, rsa *RawSockaddrAny, addrlen *_Socklen, flags int) (fd int, err error) //sys bind(s int, addr unsafe.Pointer, addrlen _Socklen) (err error) @@ -85,6 +99,18 @@ func setTimeval(sec, usec int64) Timeval { return Timeval{Sec: sec, Usec: usec} } +func futimesat(dirfd int, path string, tv *[2]Timeval) (err error) { + if tv == nil { + return utimensat(dirfd, path, nil, 0) + } + + ts := []Timespec{ + NsecToTimespec(TimevalToNsec(tv[0])), + NsecToTimespec(TimevalToNsec(tv[1])), + } + return utimensat(dirfd, path, (*[2]Timespec)(unsafe.Pointer(&ts[0])), 0) +} + func Time(t *Time_t) (Time_t, error) { var tv Timeval err := Gettimeofday(&tv) @@ -105,6 +131,18 @@ func Utime(path string, buf *Utimbuf) error { return Utimes(path, tv) } +func utimes(path string, tv *[2]Timeval) (err error) { + if tv == nil { + return utimensat(AT_FDCWD, path, nil, 0) + } + + ts := []Timespec{ + NsecToTimespec(TimevalToNsec(tv[0])), + NsecToTimespec(TimevalToNsec(tv[1])), + } + return utimensat(AT_FDCWD, path, (*[2]Timespec)(unsafe.Pointer(&ts[0])), 0) +} + func Pipe(p []int) (err error) { if len(p) != 2 { return EINVAL @@ -161,22 +199,6 @@ func Pause() (err error) { return } -// TODO(dfc): constants that should be in zsysnum_linux_arm64.go, remove -// these when the deprecated syscalls that the syscall package relies on -// are removed. -const ( - SYS_GETPGRP = 1060 - SYS_UTIMES = 1037 - SYS_FUTIMESAT = 1066 - SYS_PAUSE = 1061 - SYS_USTAT = 1070 - SYS_UTIME = 1063 - SYS_LCHOWN = 1032 - SYS_TIME = 1062 - SYS_EPOLL_CREATE = 1042 - SYS_EPOLL_WAIT = 1069 -) - func Poll(fds []PollFd, timeout int) (n int, err error) { var ts *Timespec if timeout >= 0 { diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_gc_386.go b/vendor/golang.org/x/sys/unix/syscall_linux_gc_386.go new file mode 100644 index 000000000..070bd3899 --- /dev/null +++ b/vendor/golang.org/x/sys/unix/syscall_linux_gc_386.go @@ -0,0 +1,16 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build linux,!gccgo,386 + +package unix + +import "syscall" + +// Underlying system call writes to newoffset via pointer. +// Implemented in assembly to avoid allocation. +func seek(fd int, offset int64, whence int) (newoffset int64, err syscall.Errno) + +func socketcall(call int, a0, a1, a2, a3, a4, a5 uintptr) (n int, err syscall.Errno) +func rawsocketcall(call int, a0, a1, a2, a3, a4, a5 uintptr) (n int, err syscall.Errno) diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_gccgo_386.go b/vendor/golang.org/x/sys/unix/syscall_linux_gccgo_386.go new file mode 100644 index 000000000..308eb7aec --- /dev/null +++ b/vendor/golang.org/x/sys/unix/syscall_linux_gccgo_386.go @@ -0,0 +1,30 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build linux,gccgo,386 + +package unix + +import ( + "syscall" + "unsafe" +) + +func seek(fd int, offset int64, whence int) (int64, syscall.Errno) { + var newoffset int64 + offsetLow := uint32(offset & 0xffffffff) + offsetHigh := uint32((offset >> 32) & 0xffffffff) + _, _, err := Syscall6(SYS__LLSEEK, uintptr(fd), uintptr(offsetHigh), uintptr(offsetLow), uintptr(unsafe.Pointer(&newoffset)), uintptr(whence), 0) + return newoffset, err +} + +func socketcall(call int, a0, a1, a2, a3, a4, a5 uintptr) (int, syscall.Errno) { + fd, _, err := Syscall(SYS_SOCKETCALL, uintptr(call), uintptr(unsafe.Pointer(&a0)), 0) + return int(fd), err +} + +func rawsocketcall(call int, a0, a1, a2, a3, a4, a5 uintptr) (int, syscall.Errno) { + fd, _, err := RawSyscall(SYS_SOCKETCALL, uintptr(call), uintptr(unsafe.Pointer(&a0)), 0) + return int(fd), err +} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_gccgo.go b/vendor/golang.org/x/sys/unix/syscall_linux_gccgo_arm.go similarity index 53% rename from vendor/golang.org/x/sys/unix/syscall_linux_gccgo.go rename to vendor/golang.org/x/sys/unix/syscall_linux_gccgo_arm.go index df9c12371..aa7fc9e19 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_gccgo.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_gccgo_arm.go @@ -2,9 +2,7 @@ // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. -// +build linux -// +build gccgo -// +build 386 arm +// +build linux,gccgo,arm package unix @@ -13,9 +11,10 @@ import ( "unsafe" ) -func seek(fd int, offset int64, whence int) (newoffset int64, err syscall.Errno) { +func seek(fd int, offset int64, whence int) (int64, syscall.Errno) { + var newoffset int64 offsetLow := uint32(offset & 0xffffffff) offsetHigh := uint32((offset >> 32) & 0xffffffff) - _, _, err = Syscall6(SYS__LLSEEK, uintptr(fd), uintptr(offsetHigh), uintptr(offsetLow), uintptr(unsafe.Pointer(&newoffset)), uintptr(whence), 0) + _, _, err := Syscall6(SYS__LLSEEK, uintptr(fd), uintptr(offsetHigh), uintptr(offsetLow), uintptr(unsafe.Pointer(&newoffset)), uintptr(whence), 0) return newoffset, err } diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go b/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go index 090ed404a..ad991031c 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_mips64x.go @@ -8,6 +8,7 @@ package unix //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 //sys Fchown(fd int, uid int, gid int) (err error) @@ -47,6 +48,7 @@ func Select(nfd int, r *FdSet, w *FdSet, e *FdSet, timeout *Timeval) (n int, err //sys Statfs(path string, buf *Statfs_t) (err error) //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) //sys Truncate(path string, length int64) (err error) +//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sys accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) //sys accept4(s int, rsa *RawSockaddrAny, addrlen *_Socklen, flags int) (fd int, err error) //sys bind(s int, addr unsafe.Pointer, addrlen _Socklen) (err error) @@ -65,6 +67,7 @@ func Select(nfd int, r *FdSet, w *FdSet, e *FdSet, timeout *Timeval) (n int, err //sys sendmsg(s int, msg *Msghdr, flags int) (n int, err error) //sys mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int64) (xaddr uintptr, err error) +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) func Time(t *Time_t) (tt Time_t, err error) { @@ -80,6 +83,7 @@ func Time(t *Time_t) (tt Time_t, err error) { } //sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) func setTimespec(sec, nsec int64) Timespec { return Timespec{Sec: sec, Nsec: nsec} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go b/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go index 9e16cc9d1..0e0592482 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_mipsx.go @@ -15,6 +15,8 @@ import ( func Syscall9(trap, a1, a2, a3, a4, a5, a6, a7, a8, a9 uintptr) (r1, r2 uintptr, err syscall.Errno) //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) +//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 //sys Fchown(fd int, uid int, gid int) (err error) //sys Ftruncate(fd int, length int64) (err error) = SYS_FTRUNCATE64 @@ -33,13 +35,12 @@ func Syscall9(trap, a1, a2, a3, a4, a5, a6, a7, a8, a9 uintptr) (r1, r2 uintptr, //sysnb Setregid(rgid int, egid int) (err error) //sysnb Setresgid(rgid int, egid int, sgid int) (err error) //sysnb Setresuid(ruid int, euid int, suid int) (err error) - //sysnb Setreuid(ruid int, euid int) (err error) //sys Shutdown(fd int, how int) (err error) //sys Splice(rfd int, roff *int64, wfd int, woff *int64, len int, flags int) (n int, err error) - //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) //sys Truncate(path string, length int64) (err error) = SYS_TRUNCATE64 +//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sys accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) //sys accept4(s int, rsa *RawSockaddrAny, addrlen *_Socklen, flags int) (fd int, err error) //sys bind(s int, addr unsafe.Pointer, addrlen _Socklen) (err error) @@ -61,16 +62,17 @@ func Syscall9(trap, a1, a2, a3, a4, a5, a6, a7, a8, a9 uintptr) (r1, r2 uintptr, //sys Ioperm(from int, num int, on int) (err error) //sys Iopl(level int) (err error) +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) //sysnb Time(t *Time_t) (tt Time_t, err error) +//sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) //sys Lstat(path string, stat *Stat_t) (err error) = SYS_LSTAT64 //sys Fstat(fd int, stat *Stat_t) (err error) = SYS_FSTAT64 //sys Fstatat(dirfd int, path string, stat *Stat_t, flags int) (err error) = SYS_FSTATAT64 //sys Stat(path string, stat *Stat_t) (err error) = SYS_STAT64 -//sys Utime(path string, buf *Utimbuf) (err error) -//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Pause() (err error) func Fstatfs(fd int, buf *Statfs_t) (err error) { diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go b/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go index 6fb8733d6..8c6720f7f 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_ppc64x.go @@ -7,8 +7,9 @@ package unix -//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) +//sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 //sys Fchown(fd int, uid int, gid int) (err error) //sys Fstat(fd int, stat *Stat_t) (err error) @@ -45,6 +46,7 @@ package unix //sys Statfs(path string, buf *Statfs_t) (err error) //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) = SYS_SYNC_FILE_RANGE2 //sys Truncate(path string, length int64) (err error) +//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sys accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) //sys accept4(s int, rsa *RawSockaddrAny, addrlen *_Socklen, flags int) (fd int, err error) //sys bind(s int, addr unsafe.Pointer, addrlen _Socklen) (err error) @@ -63,10 +65,11 @@ package unix //sys sendmsg(s int, msg *Msghdr, flags int) (n int, err error) //sys mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int64) (xaddr uintptr, err error) +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) //sysnb Time(t *Time_t) (tt Time_t, err error) - //sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) func setTimespec(sec, nsec int64) Timespec { return Timespec{Sec: sec, Nsec: nsec} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go b/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go index c0d86e722..6e4ee0cf2 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_s390x.go @@ -11,6 +11,7 @@ import ( ) //sys Dup2(oldfd int, newfd int) (err error) +//sysnb EpollCreate(size int) (fd int, err error) //sys EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_FADVISE64 //sys Fchown(fd int, uid int, gid int) (err error) @@ -44,9 +45,11 @@ import ( //sys Statfs(path string, buf *Statfs_t) (err error) //sys SyncFileRange(fd int, off int64, n int64, flags int) (err error) //sys Truncate(path string, length int64) (err error) +//sys Ustat(dev int, ubuf *Ustat_t) (err error) //sysnb getgroups(n int, list *_Gid_t) (nn int, err error) //sysnb setgroups(n int, list *_Gid_t) (err error) +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) func Time(t *Time_t) (tt Time_t, err error) { @@ -62,6 +65,7 @@ func Time(t *Time_t) (tt Time_t, err error) { } //sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) func setTimespec(sec, nsec int64) Timespec { return Timespec{Sec: sec, Nsec: nsec} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go b/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go index 78c1e0df1..72e64187d 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_sparc64.go @@ -68,6 +68,7 @@ func Iopl(level int) (err error) { return ENOSYS } +//sys futimesat(dirfd int, path string, times *[2]Timeval) (err error) //sysnb Gettimeofday(tv *Timeval) (err error) func Time(t *Time_t) (tt Time_t, err error) { @@ -83,6 +84,7 @@ func Time(t *Time_t) (tt Time_t, err error) { } //sys Utime(path string, buf *Utimbuf) (err error) +//sys utimes(path string, times *[2]Timeval) (err error) func setTimespec(sec, nsec int64) Timespec { return Timespec{Sec: sec, Nsec: nsec} diff --git a/vendor/golang.org/x/sys/unix/syscall_linux_test.go b/vendor/golang.org/x/sys/unix/syscall_linux_test.go index a2bc44015..eed17268b 100644 --- a/vendor/golang.org/x/sys/unix/syscall_linux_test.go +++ b/vendor/golang.org/x/sys/unix/syscall_linux_test.go @@ -32,6 +32,10 @@ func TestIoctlGetInt(t *testing.T) { } func TestPpoll(t *testing.T) { + if runtime.GOOS == "android" { + t.Skip("mkfifo syscall is not available on android, skipping test") + } + f, cleanup := mktmpfifo(t) defer cleanup() @@ -136,11 +140,16 @@ func TestUtimesNanoAt(t *testing.T) { if err != nil { t.Fatalf("Lstat: %v", err) } - if st.Atim != ts[0] { - t.Errorf("UtimesNanoAt: wrong atime: %v", st.Atim) + + // Only check Mtim, Atim might not be supported by the underlying filesystem + expected := ts[1] + if st.Mtim.Nsec == 0 { + // Some filesystems only support 1-second time stamp resolution + // and will always set Nsec to 0. + expected.Nsec = 0 } - if st.Mtim != ts[1] { - t.Errorf("UtimesNanoAt: wrong mtime: %v", st.Mtim) + if st.Mtim != expected { + t.Errorf("UtimesNanoAt: wrong mtime: expected %v, got %v", expected, st.Mtim) } } @@ -259,6 +268,9 @@ func TestSchedSetaffinity(t *testing.T) { if runtime.NumCPU() < 2 { t.Skip("skipping setaffinity tests on single CPU system") } + if runtime.GOOS == "android" { + t.Skip("skipping setaffinity tests on android") + } err = unix.SchedSetaffinity(0, &newMask) if err != nil { @@ -285,7 +297,7 @@ func TestSchedSetaffinity(t *testing.T) { func TestStatx(t *testing.T) { var stx unix.Statx_t err := unix.Statx(unix.AT_FDCWD, ".", 0, 0, &stx) - if err == unix.ENOSYS { + if err == unix.ENOSYS || err == unix.EPERM { t.Skip("statx syscall is not available, skipping test") } else if err != nil { t.Fatalf("Statx: %v", err) @@ -310,13 +322,9 @@ func TestStatx(t *testing.T) { t.Errorf("Statx: returned stat mode does not match Stat") } - atime := unix.StatxTimestamp{Sec: int64(st.Atim.Sec), Nsec: uint32(st.Atim.Nsec)} ctime := unix.StatxTimestamp{Sec: int64(st.Ctim.Sec), Nsec: uint32(st.Ctim.Nsec)} mtime := unix.StatxTimestamp{Sec: int64(st.Mtim.Sec), Nsec: uint32(st.Mtim.Nsec)} - if stx.Atime != atime { - t.Errorf("Statx: returned stat atime does not match Stat") - } if stx.Ctime != ctime { t.Errorf("Statx: returned stat ctime does not match Stat") } @@ -357,13 +365,9 @@ func TestStatx(t *testing.T) { t.Errorf("Statx: returned stat mode does not match Lstat") } - atime = unix.StatxTimestamp{Sec: int64(st.Atim.Sec), Nsec: uint32(st.Atim.Nsec)} ctime = unix.StatxTimestamp{Sec: int64(st.Ctim.Sec), Nsec: uint32(st.Ctim.Nsec)} mtime = unix.StatxTimestamp{Sec: int64(st.Mtim.Sec), Nsec: uint32(st.Mtim.Nsec)} - if stx.Atime != atime { - t.Errorf("Statx: returned stat atime does not match Lstat") - } if stx.Ctime != ctime { t.Errorf("Statx: returned stat ctime does not match Lstat") } @@ -371,3 +375,47 @@ func TestStatx(t *testing.T) { t.Errorf("Statx: returned stat mtime does not match Lstat") } } + +// stringsFromByteSlice converts a sequence of attributes to a []string. +// On Linux, each entry is a NULL-terminated string. +func stringsFromByteSlice(buf []byte) []string { + var result []string + off := 0 + for i, b := range buf { + if b == 0 { + result = append(result, string(buf[off:i])) + off = i + 1 + } + } + return result +} + +func TestFaccessat(t *testing.T) { + defer chtmpdir(t)() + touch(t, "file1") + + err := unix.Faccessat(unix.AT_FDCWD, "file1", unix.O_RDONLY, 0) + if err != nil { + t.Errorf("Faccessat: unexpected error: %v", err) + } + + err = unix.Faccessat(unix.AT_FDCWD, "file1", unix.O_RDONLY, 2) + if err != unix.EINVAL { + t.Errorf("Faccessat: unexpected error: %v, want EINVAL", err) + } + + err = unix.Faccessat(unix.AT_FDCWD, "file1", unix.O_RDONLY, unix.AT_EACCESS) + if err != unix.EOPNOTSUPP { + t.Errorf("Faccessat: unexpected error: %v, want EOPNOTSUPP", err) + } + + err = os.Symlink("file1", "symlink1") + if err != nil { + t.Fatal(err) + } + + err = unix.Faccessat(unix.AT_FDCWD, "symlink1", unix.O_RDONLY, unix.AT_SYMLINK_NOFOLLOW) + if err != unix.EOPNOTSUPP { + t.Errorf("Faccessat: unexpected error: %v, want EOPNOTSUPP", err) + } +} diff --git a/vendor/golang.org/x/sys/unix/syscall_netbsd.go b/vendor/golang.org/x/sys/unix/syscall_netbsd.go index e1a3baa23..369a2be2d 100644 --- a/vendor/golang.org/x/sys/unix/syscall_netbsd.go +++ b/vendor/golang.org/x/sys/unix/syscall_netbsd.go @@ -233,6 +233,7 @@ func Uname(uname *Utsname) error { //sys Dup(fd int) (nfd int, err error) //sys Dup2(from int, to int) (err error) //sys Exit(code int) +//sys Faccessat(dirfd int, path string, mode uint32, flags int) (err error) //sys Fadvise(fd int, offset int64, length int64, advice int) (err error) = SYS_POSIX_FADVISE //sys Fchdir(fd int) (err error) //sys Fchflags(fd int, flags int) (err error) diff --git a/vendor/golang.org/x/sys/unix/syscall_openbsd.go b/vendor/golang.org/x/sys/unix/syscall_openbsd.go index 614fcf049..9fc9c06a0 100644 --- a/vendor/golang.org/x/sys/unix/syscall_openbsd.go +++ b/vendor/golang.org/x/sys/unix/syscall_openbsd.go @@ -201,6 +201,7 @@ func Uname(uname *Utsname) error { //sys Dup(fd int) (nfd int, err error) //sys Dup2(from int, to int) (err error) //sys Exit(code int) +//sys Faccessat(dirfd int, path string, mode uint32, flags int) (err error) //sys Fchdir(fd int) (err error) //sys Fchflags(fd int, flags int) (err error) //sys Fchmod(fd int, mode uint32) (err error) diff --git a/vendor/golang.org/x/sys/unix/syscall_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/syscall_openbsd_amd64.go index 649e67fcc..9a35334cb 100644 --- a/vendor/golang.org/x/sys/unix/syscall_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/syscall_openbsd_amd64.go @@ -31,3 +31,7 @@ func (msghdr *Msghdr) SetControllen(length int) { func (cmsg *Cmsghdr) SetLen(length int) { cmsg.Len = uint32(length) } + +// SYS___SYSCTL is used by syscall_bsd.go for all BSDs, but in modern versions +// of openbsd/amd64 the syscall is called sysctl instead of __sysctl. +const SYS___SYSCTL = SYS_SYSCTL diff --git a/vendor/golang.org/x/sys/unix/syscall_solaris.go b/vendor/golang.org/x/sys/unix/syscall_solaris.go index b7629529b..820ef77af 100644 --- a/vendor/golang.org/x/sys/unix/syscall_solaris.go +++ b/vendor/golang.org/x/sys/unix/syscall_solaris.go @@ -314,7 +314,11 @@ func UtimesNanoAt(dirfd int, path string, ts []Timespec, flags int) error { // FcntlInt performs a fcntl syscall on fd with the provided command and argument. func FcntlInt(fd uintptr, cmd, arg int) (int, error) { - valptr, _, err := sysvicall6(uintptr(unsafe.Pointer(&procfcntl)), 3, uintptr(fd), uintptr(cmd), uintptr(arg), 0, 0, 0) + valptr, _, errno := sysvicall6(uintptr(unsafe.Pointer(&procfcntl)), 3, uintptr(fd), uintptr(cmd), uintptr(arg), 0, 0, 0) + var err error + if errno != 0 { + err = errno + } return int(valptr), err } @@ -595,6 +599,7 @@ func Poll(fds []PollFd, timeout int) (n int, err error) { //sys Dup(fd int) (nfd int, err error) //sys Dup2(oldfd int, newfd int) (err error) //sys Exit(code int) +//sys Faccessat(dirfd int, path string, mode uint32, flags int) (err error) //sys Fchdir(fd int) (err error) //sys Fchmod(fd int, mode uint32) (err error) //sys Fchmodat(dirfd int, path string, mode uint32, flags int) (err error) diff --git a/vendor/golang.org/x/sys/unix/syscall_unix_test.go b/vendor/golang.org/x/sys/unix/syscall_unix_test.go index 7372b3ea3..d694990db 100644 --- a/vendor/golang.org/x/sys/unix/syscall_unix_test.go +++ b/vendor/golang.org/x/sys/unix/syscall_unix_test.go @@ -98,6 +98,24 @@ func TestErrnoSignalName(t *testing.T) { } } +func TestFcntlInt(t *testing.T) { + t.Parallel() + file, err := ioutil.TempFile("", "TestFnctlInt") + if err != nil { + t.Fatal(err) + } + defer os.Remove(file.Name()) + defer file.Close() + f := file.Fd() + flags, err := unix.FcntlInt(f, unix.F_GETFD, 0) + if err != nil { + t.Fatal(err) + } + if flags&unix.FD_CLOEXEC == 0 { + t.Errorf("flags %#x do not include FD_CLOEXEC", flags) + } +} + // TestFcntlFlock tests whether the file locking structure matches // the calling convention of each kernel. func TestFcntlFlock(t *testing.T) { @@ -125,6 +143,10 @@ func TestFcntlFlock(t *testing.T) { // "-test.run=^TestPassFD$" and an environment variable used to signal // that the test should become the child process instead. func TestPassFD(t *testing.T) { + if runtime.GOOS == "darwin" && (runtime.GOARCH == "arm" || runtime.GOARCH == "arm64") { + t.Skip("cannot exec subprocess on iOS, skipping test") + } + if os.Getenv("GO_WANT_HELPER_PROCESS") == "1" { passFDChild() return @@ -390,6 +412,11 @@ func TestDup(t *testing.T) { } func TestPoll(t *testing.T) { + if runtime.GOOS == "android" || + (runtime.GOOS == "darwin" && (runtime.GOARCH == "arm" || runtime.GOARCH == "arm64")) { + t.Skip("mkfifo syscall is not available on android and iOS, skipping test") + } + f, cleanup := mktmpfifo(t) defer cleanup() @@ -426,7 +453,10 @@ func TestGetwd(t *testing.T) { // These are chosen carefully not to be symlinks on a Mac // (unlike, say, /var, /etc) dirs := []string{"/", "/usr/bin"} - if runtime.GOOS == "darwin" { + switch runtime.GOOS { + case "android": + dirs = []string{"/", "/system/bin"} + case "darwin": switch runtime.GOARCH { case "arm", "arm64": d1, err := ioutil.TempDir("", "d1") @@ -534,7 +564,7 @@ func TestFchmodat(t *testing.T) { didChmodSymlink := true err = unix.Fchmodat(unix.AT_FDCWD, "symlink1", uint32(mode), unix.AT_SYMLINK_NOFOLLOW) if err != nil { - if (runtime.GOOS == "linux" || runtime.GOOS == "solaris") && err == unix.EOPNOTSUPP { + if (runtime.GOOS == "android" || runtime.GOOS == "linux" || runtime.GOOS == "solaris") && err == unix.EOPNOTSUPP { // Linux and Illumos don't support flags != 0 didChmodSymlink = false } else { @@ -560,6 +590,19 @@ func TestFchmodat(t *testing.T) { } } +func TestMkdev(t *testing.T) { + major := uint32(42) + minor := uint32(7) + dev := unix.Mkdev(major, minor) + + if unix.Major(dev) != major { + t.Errorf("Major(%#x) == %d, want %d", dev, unix.Major(dev), major) + } + if unix.Minor(dev) != minor { + t.Errorf("Minor(%#x) == %d, want %d", dev, unix.Minor(dev), minor) + } +} + // mktmpfifo creates a temporary FIFO and provides a cleanup function. func mktmpfifo(t *testing.T) (*os.File, func()) { err := unix.Mkfifo("fifo", 0666) diff --git a/vendor/golang.org/x/sys/unix/xattr_test.go b/vendor/golang.org/x/sys/unix/xattr_test.go new file mode 100644 index 000000000..b8b28d0c1 --- /dev/null +++ b/vendor/golang.org/x/sys/unix/xattr_test.go @@ -0,0 +1,119 @@ +// Copyright 2018 The Go Authors. All rights reserved. +// Use of this source code is governed by a BSD-style +// license that can be found in the LICENSE file. + +// +build darwin freebsd linux + +package unix_test + +import ( + "os" + "runtime" + "strings" + "testing" + + "golang.org/x/sys/unix" +) + +func TestXattr(t *testing.T) { + defer chtmpdir(t)() + + f := "xattr1" + touch(t, f) + + xattrName := "user.test" + xattrDataSet := "gopher" + err := unix.Setxattr(f, xattrName, []byte(xattrDataSet), 0) + if err == unix.ENOTSUP || err == unix.EOPNOTSUPP { + t.Skip("filesystem does not support extended attributes, skipping test") + } else if err != nil { + t.Fatalf("Setxattr: %v", err) + } + + // find size + size, err := unix.Listxattr(f, nil) + if err != nil { + t.Fatalf("Listxattr: %v", err) + } + + if size <= 0 { + t.Fatalf("Listxattr returned an empty list of attributes") + } + + buf := make([]byte, size) + read, err := unix.Listxattr(f, buf) + if err != nil { + t.Fatalf("Listxattr: %v", err) + } + + xattrs := stringsFromByteSlice(buf[:read]) + + xattrWant := xattrName + if runtime.GOOS == "freebsd" { + // On FreeBSD, the namespace is stored separately from the xattr + // name and Listxattr doesn't return the namespace prefix. + xattrWant = strings.TrimPrefix(xattrWant, "user.") + } + found := false + for _, name := range xattrs { + if name == xattrWant { + found = true + } + } + + if !found { + t.Errorf("Listxattr did not return previously set attribute '%s'", xattrName) + } + + // find size + size, err = unix.Getxattr(f, xattrName, nil) + if err != nil { + t.Fatalf("Getxattr: %v", err) + } + + if size <= 0 { + t.Fatalf("Getxattr returned an empty attribute") + } + + xattrDataGet := make([]byte, size) + _, err = unix.Getxattr(f, xattrName, xattrDataGet) + if err != nil { + t.Fatalf("Getxattr: %v", err) + } + + got := string(xattrDataGet) + if got != xattrDataSet { + t.Errorf("Getxattr: expected attribute value %s, got %s", xattrDataSet, got) + } + + err = unix.Removexattr(f, xattrName) + if err != nil { + t.Fatalf("Removexattr: %v", err) + } + + n := "nonexistent" + err = unix.Lsetxattr(n, xattrName, []byte(xattrDataSet), 0) + if err != unix.ENOENT { + t.Errorf("Lsetxattr: expected %v on non-existent file, got %v", unix.ENOENT, err) + } + + _, err = unix.Lgetxattr(n, xattrName, nil) + if err != unix.ENOENT { + t.Errorf("Lgetxattr: %v", err) + } + + s := "symlink1" + err = os.Symlink(n, s) + if err != nil { + t.Fatal(err) + } + + err = unix.Lsetxattr(s, xattrName, []byte(xattrDataSet), 0) + if err != nil { + // Linux and Android doen't support xattrs on symlinks according + // to xattr(7), so just test that we get the proper error. + if (runtime.GOOS != "linux" && runtime.GOOS != "android") || err != unix.EPERM { + t.Fatalf("Lsetxattr: %v", err) + } + } +} diff --git a/vendor/golang.org/x/sys/unix/zerrors_darwin_386.go b/vendor/golang.org/x/sys/unix/zerrors_darwin_386.go index c5a9b73dc..3b39d7408 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_darwin_386.go +++ b/vendor/golang.org/x/sys/unix/zerrors_darwin_386.go @@ -1473,6 +1473,12 @@ const ( WORDSIZE = 0x20 WSTOPPED = 0x8 WUNTRACED = 0x2 + XATTR_CREATE = 0x2 + XATTR_NODEFAULT = 0x10 + XATTR_NOFOLLOW = 0x1 + XATTR_NOSECURITY = 0x8 + XATTR_REPLACE = 0x4 + XATTR_SHOWCOMPRESSION = 0x20 ) // Errors diff --git a/vendor/golang.org/x/sys/unix/zerrors_darwin_amd64.go b/vendor/golang.org/x/sys/unix/zerrors_darwin_amd64.go index 7de222b65..8fe554777 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_darwin_amd64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_darwin_amd64.go @@ -1473,6 +1473,12 @@ const ( WORDSIZE = 0x40 WSTOPPED = 0x8 WUNTRACED = 0x2 + XATTR_CREATE = 0x2 + XATTR_NODEFAULT = 0x10 + XATTR_NOFOLLOW = 0x1 + XATTR_NOSECURITY = 0x8 + XATTR_REPLACE = 0x4 + XATTR_SHOWCOMPRESSION = 0x20 ) // Errors diff --git a/vendor/golang.org/x/sys/unix/zerrors_darwin_arm.go b/vendor/golang.org/x/sys/unix/zerrors_darwin_arm.go index 33a42e78a..7a977770d 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_darwin_arm.go +++ b/vendor/golang.org/x/sys/unix/zerrors_darwin_arm.go @@ -1473,6 +1473,12 @@ const ( WORDSIZE = 0x40 WSTOPPED = 0x8 WUNTRACED = 0x2 + XATTR_CREATE = 0x2 + XATTR_NODEFAULT = 0x10 + XATTR_NOFOLLOW = 0x1 + XATTR_NOSECURITY = 0x8 + XATTR_REPLACE = 0x4 + XATTR_SHOWCOMPRESSION = 0x20 ) // Errors diff --git a/vendor/golang.org/x/sys/unix/zerrors_darwin_arm64.go b/vendor/golang.org/x/sys/unix/zerrors_darwin_arm64.go index 71805155b..6d56d8a05 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_darwin_arm64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_darwin_arm64.go @@ -1473,6 +1473,12 @@ const ( WORDSIZE = 0x40 WSTOPPED = 0x8 WUNTRACED = 0x2 + XATTR_CREATE = 0x2 + XATTR_NODEFAULT = 0x10 + XATTR_NOFOLLOW = 0x1 + XATTR_NOSECURITY = 0x8 + XATTR_REPLACE = 0x4 + XATTR_SHOWCOMPRESSION = 0x20 ) // Errors diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_386.go b/vendor/golang.org/x/sys/unix/zerrors_linux_386.go index 9dbd0eff0..2f0091bbc 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_386.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_386.go @@ -3,7 +3,7 @@ // +build 386,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include -m32 _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1075,6 +1076,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1205,6 +1235,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x2401 PERF_EVENT_IOC_ENABLE = 0x2400 PERF_EVENT_IOC_ID = 0x80042407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x4004240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x40042409 PERF_EVENT_IOC_PERIOD = 0x40082404 PERF_EVENT_IOC_QUERY_BPF = 0xc004240a @@ -1260,6 +1291,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1305,11 +1337,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1437,6 +1476,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x7002 + RTC_AIE_ON = 0x7001 + RTC_ALM_READ = 0x80247008 + RTC_ALM_SET = 0x40247007 + RTC_EPOCH_READ = 0x8004700d + RTC_EPOCH_SET = 0x4004700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x8004700b + RTC_IRQP_SET = 0x4004700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x7006 + RTC_PIE_ON = 0x7005 + RTC_PLL_GET = 0x801c7011 + RTC_PLL_SET = 0x401c7012 + RTC_RD_TIME = 0x80247009 + RTC_SET_TIME = 0x4024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x7004 + RTC_UIE_ON = 0x7003 + RTC_VL_CLR = 0x7014 + RTC_VL_READ = 0x80047013 + RTC_WIE_OFF = 0x7010 + RTC_WIE_ON = 0x700f + RTC_WKALM_RD = 0x80287010 + RTC_WKALM_SET = 0x4028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go index a9ace38e0..a80c7ae5f 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_amd64.go @@ -3,7 +3,7 @@ // +build amd64,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include -m64 _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1075,6 +1076,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1205,6 +1235,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x2401 PERF_EVENT_IOC_ENABLE = 0x2400 PERF_EVENT_IOC_ID = 0x80082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x4008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x40042409 PERF_EVENT_IOC_PERIOD = 0x40082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1260,6 +1291,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1305,11 +1337,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1438,6 +1477,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x7002 + RTC_AIE_ON = 0x7001 + RTC_ALM_READ = 0x80247008 + RTC_ALM_SET = 0x40247007 + RTC_EPOCH_READ = 0x8008700d + RTC_EPOCH_SET = 0x4008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x8008700b + RTC_IRQP_SET = 0x4008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x7006 + RTC_PIE_ON = 0x7005 + RTC_PLL_GET = 0x80207011 + RTC_PLL_SET = 0x40207012 + RTC_RD_TIME = 0x80247009 + RTC_SET_TIME = 0x4024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x7004 + RTC_UIE_ON = 0x7003 + RTC_VL_CLR = 0x7014 + RTC_VL_READ = 0x80047013 + RTC_WIE_OFF = 0x7010 + RTC_WIE_ON = 0x700f + RTC_WKALM_RD = 0x80287010 + RTC_WKALM_SET = 0x4028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go b/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go index fe394471b..49a9b0133 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_arm.go @@ -3,7 +3,7 @@ // +build arm,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1073,6 +1074,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1203,6 +1233,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x2401 PERF_EVENT_IOC_ENABLE = 0x2400 PERF_EVENT_IOC_ID = 0x80042407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x4004240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x40042409 PERF_EVENT_IOC_PERIOD = 0x40082404 PERF_EVENT_IOC_QUERY_BPF = 0xc004240a @@ -1258,6 +1289,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1303,11 +1335,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1444,6 +1483,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x7002 + RTC_AIE_ON = 0x7001 + RTC_ALM_READ = 0x80247008 + RTC_ALM_SET = 0x40247007 + RTC_EPOCH_READ = 0x8004700d + RTC_EPOCH_SET = 0x4004700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x8004700b + RTC_IRQP_SET = 0x4004700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x7006 + RTC_PIE_ON = 0x7005 + RTC_PLL_GET = 0x801c7011 + RTC_PLL_SET = 0x401c7012 + RTC_RD_TIME = 0x80247009 + RTC_SET_TIME = 0x4024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x7004 + RTC_UIE_ON = 0x7003 + RTC_VL_CLR = 0x7014 + RTC_VL_READ = 0x80047013 + RTC_WIE_OFF = 0x7010 + RTC_WIE_ON = 0x700f + RTC_WKALM_RD = 0x80287010 + RTC_WKALM_SET = 0x4028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go index dbe9a5192..8d70233b2 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_arm64.go @@ -3,7 +3,7 @@ // +build arm64,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include -fsigned-char _const.go package unix @@ -450,6 +450,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1076,6 +1077,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1206,6 +1236,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x2401 PERF_EVENT_IOC_ENABLE = 0x2400 PERF_EVENT_IOC_ID = 0x80082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x4008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x40042409 PERF_EVENT_IOC_PERIOD = 0x40082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1261,6 +1292,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1306,11 +1338,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1428,6 +1467,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x7002 + RTC_AIE_ON = 0x7001 + RTC_ALM_READ = 0x80247008 + RTC_ALM_SET = 0x40247007 + RTC_EPOCH_READ = 0x8008700d + RTC_EPOCH_SET = 0x4008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x8008700b + RTC_IRQP_SET = 0x4008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x7006 + RTC_PIE_ON = 0x7005 + RTC_PLL_GET = 0x80207011 + RTC_PLL_SET = 0x40207012 + RTC_RD_TIME = 0x80247009 + RTC_SET_TIME = 0x4024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x7004 + RTC_UIE_ON = 0x7003 + RTC_VL_CLR = 0x7014 + RTC_VL_READ = 0x80047013 + RTC_WIE_OFF = 0x7010 + RTC_WIE_ON = 0x700f + RTC_WKALM_RD = 0x80287010 + RTC_WKALM_SET = 0x4028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go index fb9075317..410ab56b0 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mips.go @@ -3,7 +3,7 @@ // +build mips,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1074,6 +1075,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1204,6 +1234,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x20002401 PERF_EVENT_IOC_ENABLE = 0x20002400 PERF_EVENT_IOC_ID = 0x40042407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x8004240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x80042409 PERF_EVENT_IOC_PERIOD = 0x80082404 PERF_EVENT_IOC_QUERY_BPF = 0xc004240a @@ -1259,6 +1290,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1304,11 +1336,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1438,6 +1477,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x20007002 + RTC_AIE_ON = 0x20007001 + RTC_ALM_READ = 0x40247008 + RTC_ALM_SET = 0x80247007 + RTC_EPOCH_READ = 0x4004700d + RTC_EPOCH_SET = 0x8004700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x4004700b + RTC_IRQP_SET = 0x8004700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x20007006 + RTC_PIE_ON = 0x20007005 + RTC_PLL_GET = 0x401c7011 + RTC_PLL_SET = 0x801c7012 + RTC_RD_TIME = 0x40247009 + RTC_SET_TIME = 0x8024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x20007004 + RTC_UIE_ON = 0x20007003 + RTC_VL_CLR = 0x20007014 + RTC_VL_READ = 0x40047013 + RTC_WIE_OFF = 0x20007010 + RTC_WIE_ON = 0x2000700f + RTC_WKALM_RD = 0x40287010 + RTC_WKALM_SET = 0x8028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go index aa6f31993..dac4d90fa 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64.go @@ -3,7 +3,7 @@ // +build mips64,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1074,6 +1075,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1204,6 +1234,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x20002401 PERF_EVENT_IOC_ENABLE = 0x20002400 PERF_EVENT_IOC_ID = 0x40082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x8008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x80042409 PERF_EVENT_IOC_PERIOD = 0x80082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1259,6 +1290,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1304,11 +1336,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1438,6 +1477,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x20007002 + RTC_AIE_ON = 0x20007001 + RTC_ALM_READ = 0x40247008 + RTC_ALM_SET = 0x80247007 + RTC_EPOCH_READ = 0x4008700d + RTC_EPOCH_SET = 0x8008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x4008700b + RTC_IRQP_SET = 0x8008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x20007006 + RTC_PIE_ON = 0x20007005 + RTC_PLL_GET = 0x40207011 + RTC_PLL_SET = 0x80207012 + RTC_RD_TIME = 0x40247009 + RTC_SET_TIME = 0x8024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x20007004 + RTC_UIE_ON = 0x20007003 + RTC_VL_CLR = 0x20007014 + RTC_VL_READ = 0x40047013 + RTC_WIE_OFF = 0x20007010 + RTC_WIE_ON = 0x2000700f + RTC_WKALM_RD = 0x40287010 + RTC_WKALM_SET = 0x8028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go index 9d716e2dd..1d2f0e638 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mips64le.go @@ -3,7 +3,7 @@ // +build mips64le,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1074,6 +1075,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1204,6 +1234,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x20002401 PERF_EVENT_IOC_ENABLE = 0x20002400 PERF_EVENT_IOC_ID = 0x40082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x8008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x80042409 PERF_EVENT_IOC_PERIOD = 0x80082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1259,6 +1290,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1304,11 +1336,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1438,6 +1477,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x20007002 + RTC_AIE_ON = 0x20007001 + RTC_ALM_READ = 0x40247008 + RTC_ALM_SET = 0x80247007 + RTC_EPOCH_READ = 0x4008700d + RTC_EPOCH_SET = 0x8008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x4008700b + RTC_IRQP_SET = 0x8008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x20007006 + RTC_PIE_ON = 0x20007005 + RTC_PLL_GET = 0x40207011 + RTC_PLL_SET = 0x80207012 + RTC_RD_TIME = 0x40247009 + RTC_SET_TIME = 0x8024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x20007004 + RTC_UIE_ON = 0x20007003 + RTC_VL_CLR = 0x20007014 + RTC_VL_READ = 0x40047013 + RTC_WIE_OFF = 0x20007010 + RTC_WIE_ON = 0x2000700f + RTC_WKALM_RD = 0x40287010 + RTC_WKALM_SET = 0x8028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go b/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go index 98aa7ad24..33b994023 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_mipsle.go @@ -3,7 +3,7 @@ // +build mipsle,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1074,6 +1075,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1204,6 +1234,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x20002401 PERF_EVENT_IOC_ENABLE = 0x20002400 PERF_EVENT_IOC_ID = 0x40042407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x8004240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x80042409 PERF_EVENT_IOC_PERIOD = 0x80082404 PERF_EVENT_IOC_QUERY_BPF = 0xc004240a @@ -1259,6 +1290,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1304,11 +1336,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1438,6 +1477,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x20007002 + RTC_AIE_ON = 0x20007001 + RTC_ALM_READ = 0x40247008 + RTC_ALM_SET = 0x80247007 + RTC_EPOCH_READ = 0x4004700d + RTC_EPOCH_SET = 0x8004700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x4004700b + RTC_IRQP_SET = 0x8004700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x20007006 + RTC_PIE_ON = 0x20007005 + RTC_PLL_GET = 0x401c7011 + RTC_PLL_SET = 0x801c7012 + RTC_RD_TIME = 0x40247009 + RTC_SET_TIME = 0x8024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x20007004 + RTC_UIE_ON = 0x20007003 + RTC_VL_CLR = 0x20007014 + RTC_VL_READ = 0x40047013 + RTC_WIE_OFF = 0x20007010 + RTC_WIE_ON = 0x2000700f + RTC_WKALM_RD = 0x40287010 + RTC_WKALM_SET = 0x8028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go index d545e0d7c..c78d66976 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64.go @@ -3,7 +3,7 @@ // +build ppc64,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1073,6 +1074,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1205,6 +1235,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x20002401 PERF_EVENT_IOC_ENABLE = 0x20002400 PERF_EVENT_IOC_ID = 0x40082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x8008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x80042409 PERF_EVENT_IOC_PERIOD = 0x80082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1261,6 +1292,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1306,11 +1338,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1494,6 +1533,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x20007002 + RTC_AIE_ON = 0x20007001 + RTC_ALM_READ = 0x40247008 + RTC_ALM_SET = 0x80247007 + RTC_EPOCH_READ = 0x4008700d + RTC_EPOCH_SET = 0x8008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x4008700b + RTC_IRQP_SET = 0x8008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x20007006 + RTC_PIE_ON = 0x20007005 + RTC_PLL_GET = 0x40207011 + RTC_PLL_SET = 0x80207012 + RTC_RD_TIME = 0x40247009 + RTC_SET_TIME = 0x8024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x20007004 + RTC_UIE_ON = 0x20007003 + RTC_VL_CLR = 0x20007014 + RTC_VL_READ = 0x40047013 + RTC_WIE_OFF = 0x20007010 + RTC_WIE_ON = 0x2000700f + RTC_WKALM_RD = 0x40287010 + RTC_WKALM_SET = 0x8028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go index 102f09ee1..63493756d 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_ppc64le.go @@ -3,7 +3,7 @@ // +build ppc64le,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1073,6 +1074,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1205,6 +1235,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x20002401 PERF_EVENT_IOC_ENABLE = 0x20002400 PERF_EVENT_IOC_ID = 0x40082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x8008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x80042409 PERF_EVENT_IOC_PERIOD = 0x80082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1261,6 +1292,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1306,11 +1338,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1494,6 +1533,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x20007002 + RTC_AIE_ON = 0x20007001 + RTC_ALM_READ = 0x40247008 + RTC_ALM_SET = 0x80247007 + RTC_EPOCH_READ = 0x4008700d + RTC_EPOCH_SET = 0x8008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x4008700b + RTC_IRQP_SET = 0x8008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x20007006 + RTC_PIE_ON = 0x20007005 + RTC_PLL_GET = 0x40207011 + RTC_PLL_SET = 0x80207012 + RTC_RD_TIME = 0x40247009 + RTC_SET_TIME = 0x8024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x20007004 + RTC_UIE_ON = 0x20007003 + RTC_VL_CLR = 0x20007014 + RTC_VL_READ = 0x40047013 + RTC_WIE_OFF = 0x20007010 + RTC_WIE_ON = 0x2000700f + RTC_WKALM_RD = 0x40287010 + RTC_WKALM_SET = 0x8028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go b/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go index e8326c108..3814df854 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_s390x.go @@ -3,7 +3,7 @@ // +build s390x,linux -// Created by cgo -godefs - DO NOT EDIT +// Code generated by cmd/cgo -godefs; DO NOT EDIT. // cgo -godefs -- -Wall -Werror -static -I/tmp/include -fsigned-char _const.go package unix @@ -449,6 +449,7 @@ const ( ETH_P_PPP_DISC = 0x8863 ETH_P_PPP_MP = 0x8 ETH_P_PPP_SES = 0x8864 + ETH_P_PREAUTH = 0x88c7 ETH_P_PRP = 0x88fb ETH_P_PUP = 0x200 ETH_P_PUPAT = 0x201 @@ -1073,6 +1074,35 @@ const ( NETLINK_UNUSED = 0x1 NETLINK_USERSOCK = 0x2 NETLINK_XFRM = 0x6 + NFNETLINK_V0 = 0x0 + NFNLGRP_ACCT_QUOTA = 0x8 + NFNLGRP_CONNTRACK_DESTROY = 0x3 + NFNLGRP_CONNTRACK_EXP_DESTROY = 0x6 + NFNLGRP_CONNTRACK_EXP_NEW = 0x4 + NFNLGRP_CONNTRACK_EXP_UPDATE = 0x5 + NFNLGRP_CONNTRACK_NEW = 0x1 + NFNLGRP_CONNTRACK_UPDATE = 0x2 + NFNLGRP_MAX = 0x9 + NFNLGRP_NFTABLES = 0x7 + NFNLGRP_NFTRACE = 0x9 + NFNLGRP_NONE = 0x0 + NFNL_BATCH_MAX = 0x1 + NFNL_MSG_BATCH_BEGIN = 0x10 + NFNL_MSG_BATCH_END = 0x11 + NFNL_NFA_NEST = 0x8000 + NFNL_SUBSYS_ACCT = 0x7 + NFNL_SUBSYS_COUNT = 0xc + NFNL_SUBSYS_CTHELPER = 0x9 + NFNL_SUBSYS_CTNETLINK = 0x1 + NFNL_SUBSYS_CTNETLINK_EXP = 0x2 + NFNL_SUBSYS_CTNETLINK_TIMEOUT = 0x8 + NFNL_SUBSYS_IPSET = 0x6 + NFNL_SUBSYS_NFTABLES = 0xa + NFNL_SUBSYS_NFT_COMPAT = 0xb + NFNL_SUBSYS_NONE = 0x0 + NFNL_SUBSYS_OSF = 0x5 + NFNL_SUBSYS_QUEUE = 0x3 + NFNL_SUBSYS_ULOG = 0x4 NFS_SUPER_MAGIC = 0x6969 NILFS_SUPER_MAGIC = 0x3434 NL0 = 0x0 @@ -1203,6 +1233,7 @@ const ( PERF_EVENT_IOC_DISABLE = 0x2401 PERF_EVENT_IOC_ENABLE = 0x2400 PERF_EVENT_IOC_ID = 0x80082407 + PERF_EVENT_IOC_MODIFY_ATTRIBUTES = 0x4008240b PERF_EVENT_IOC_PAUSE_OUTPUT = 0x40042409 PERF_EVENT_IOC_PERIOD = 0x40082404 PERF_EVENT_IOC_QUERY_BPF = 0xc008240a @@ -1258,6 +1289,7 @@ const ( PR_GET_PDEATHSIG = 0x2 PR_GET_SECCOMP = 0x15 PR_GET_SECUREBITS = 0x1b + PR_GET_SPECULATION_CTRL = 0x34 PR_GET_THP_DISABLE = 0x2a PR_GET_TID_ADDRESS = 0x28 PR_GET_TIMERSLACK = 0x1e @@ -1303,11 +1335,18 @@ const ( PR_SET_PTRACER_ANY = 0xffffffffffffffff PR_SET_SECCOMP = 0x16 PR_SET_SECUREBITS = 0x1c + PR_SET_SPECULATION_CTRL = 0x35 PR_SET_THP_DISABLE = 0x29 PR_SET_TIMERSLACK = 0x1d PR_SET_TIMING = 0xe PR_SET_TSC = 0x1a PR_SET_UNALIGN = 0x6 + PR_SPEC_DISABLE = 0x4 + PR_SPEC_ENABLE = 0x2 + PR_SPEC_FORCE_DISABLE = 0x8 + PR_SPEC_NOT_AFFECTED = 0x0 + PR_SPEC_PRCTL = 0x1 + PR_SPEC_STORE_BYPASS = 0x0 PR_SVE_GET_VL = 0x33 PR_SVE_SET_VL = 0x32 PR_SVE_SET_VL_ONEXEC = 0x40000 @@ -1498,6 +1537,33 @@ const ( RTCF_MASQ = 0x400000 RTCF_NAT = 0x800000 RTCF_VALVE = 0x200000 + RTC_AF = 0x20 + RTC_AIE_OFF = 0x7002 + RTC_AIE_ON = 0x7001 + RTC_ALM_READ = 0x80247008 + RTC_ALM_SET = 0x40247007 + RTC_EPOCH_READ = 0x8008700d + RTC_EPOCH_SET = 0x4008700e + RTC_IRQF = 0x80 + RTC_IRQP_READ = 0x8008700b + RTC_IRQP_SET = 0x4008700c + RTC_MAX_FREQ = 0x2000 + RTC_PF = 0x40 + RTC_PIE_OFF = 0x7006 + RTC_PIE_ON = 0x7005 + RTC_PLL_GET = 0x80207011 + RTC_PLL_SET = 0x40207012 + RTC_RD_TIME = 0x80247009 + RTC_SET_TIME = 0x4024700a + RTC_UF = 0x10 + RTC_UIE_OFF = 0x7004 + RTC_UIE_ON = 0x7003 + RTC_VL_CLR = 0x7014 + RTC_VL_READ = 0x80047013 + RTC_WIE_OFF = 0x7010 + RTC_WIE_ON = 0x700f + RTC_WKALM_RD = 0x80287010 + RTC_WKALM_SET = 0x4028700f RTF_ADDRCLASSMASK = 0xf8000000 RTF_ADDRCONF = 0x40000 RTF_ALLONLINK = 0x20000 diff --git a/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go b/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go index 95de199fc..7fdc85b17 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_linux_sparc64.go @@ -1,5 +1,5 @@ // mkerrors.sh -m64 -// MACHINE GENERATED BY THE COMMAND ABOVE; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build sparc64,linux diff --git a/vendor/golang.org/x/sys/unix/zerrors_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/zerrors_openbsd_amd64.go index 0d50f6b09..acfc66469 100644 --- a/vendor/golang.org/x/sys/unix/zerrors_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zerrors_openbsd_amd64.go @@ -45,6 +45,7 @@ const ( AF_SNA = 0xb AF_UNIX = 0x1 AF_UNSPEC = 0x0 + ALTWERASE = 0x200 ARPHRD_ETHER = 0x1 ARPHRD_FRELAY = 0xf ARPHRD_IEEE1394 = 0x18 @@ -146,6 +147,12 @@ const ( BRKINT = 0x2 CFLUSH = 0xf CLOCAL = 0x8000 + CLOCK_BOOTTIME = 0x6 + CLOCK_MONOTONIC = 0x3 + CLOCK_PROCESS_CPUTIME_ID = 0x2 + CLOCK_REALTIME = 0x0 + CLOCK_THREAD_CPUTIME_ID = 0x4 + CLOCK_UPTIME = 0x5 CREAD = 0x800 CRTSCTS = 0x10000 CS5 = 0x0 @@ -178,6 +185,7 @@ const ( DLT_LOOP = 0xc DLT_MPLS = 0xdb DLT_NULL = 0x0 + DLT_OPENFLOW = 0x10b DLT_PFLOG = 0x75 DLT_PFSYNC = 0x12 DLT_PPP = 0x9 @@ -188,6 +196,23 @@ const ( DLT_RAW = 0xe DLT_SLIP = 0x8 DLT_SLIP_BSDOS = 0xf + DLT_USBPCAP = 0xf9 + DLT_USER0 = 0x93 + DLT_USER1 = 0x94 + DLT_USER10 = 0x9d + DLT_USER11 = 0x9e + DLT_USER12 = 0x9f + DLT_USER13 = 0xa0 + DLT_USER14 = 0xa1 + DLT_USER15 = 0xa2 + DLT_USER2 = 0x95 + DLT_USER3 = 0x96 + DLT_USER4 = 0x97 + DLT_USER5 = 0x98 + DLT_USER6 = 0x99 + DLT_USER7 = 0x9a + DLT_USER8 = 0x9b + DLT_USER9 = 0x9c DT_BLK = 0x6 DT_CHR = 0x2 DT_DIR = 0x4 @@ -401,27 +426,38 @@ const ( ETHER_CRC_POLY_LE = 0xedb88320 ETHER_HDR_LEN = 0xe ETHER_MAX_DIX_LEN = 0x600 + ETHER_MAX_HARDMTU_LEN = 0xff9b ETHER_MAX_LEN = 0x5ee ETHER_MIN_LEN = 0x40 ETHER_TYPE_LEN = 0x2 ETHER_VLAN_ENCAP_LEN = 0x4 EVFILT_AIO = -0x3 + EVFILT_DEVICE = -0x8 EVFILT_PROC = -0x5 EVFILT_READ = -0x1 EVFILT_SIGNAL = -0x6 - EVFILT_SYSCOUNT = 0x7 + EVFILT_SYSCOUNT = 0x8 EVFILT_TIMER = -0x7 EVFILT_VNODE = -0x4 EVFILT_WRITE = -0x2 + EVL_ENCAPLEN = 0x4 + EVL_PRIO_BITS = 0xd + EVL_PRIO_MAX = 0x7 + EVL_VLID_MASK = 0xfff + EVL_VLID_MAX = 0xffe + EVL_VLID_MIN = 0x1 + EVL_VLID_NULL = 0x0 EV_ADD = 0x1 EV_CLEAR = 0x20 EV_DELETE = 0x2 EV_DISABLE = 0x8 + EV_DISPATCH = 0x80 EV_ENABLE = 0x4 EV_EOF = 0x8000 EV_ERROR = 0x4000 EV_FLAG1 = 0x2000 EV_ONESHOT = 0x10 + EV_RECEIPT = 0x40 EV_SYSFLAGS = 0xf000 EXTA = 0x4b00 EXTB = 0x9600 @@ -435,7 +471,7 @@ const ( F_GETFL = 0x3 F_GETLK = 0x7 F_GETOWN = 0x5 - F_OK = 0x0 + F_ISATTY = 0xb F_RDLCK = 0x1 F_SETFD = 0x2 F_SETFL = 0x4 @@ -452,7 +488,6 @@ const ( IEXTEN = 0x400 IFAN_ARRIVAL = 0x0 IFAN_DEPARTURE = 0x1 - IFA_ROUTE = 0x1 IFF_ALLMULTI = 0x200 IFF_BROADCAST = 0x2 IFF_CANTCHANGE = 0x8e52 @@ -463,12 +498,12 @@ const ( IFF_LOOPBACK = 0x8 IFF_MULTICAST = 0x8000 IFF_NOARP = 0x80 - IFF_NOTRAILERS = 0x20 IFF_OACTIVE = 0x400 IFF_POINTOPOINT = 0x10 IFF_PROMISC = 0x100 IFF_RUNNING = 0x40 IFF_SIMPLEX = 0x800 + IFF_STATICARP = 0x20 IFF_UP = 0x1 IFNAMSIZ = 0x10 IFT_1822 = 0x2 @@ -597,6 +632,7 @@ const ( IFT_LINEGROUP = 0xd2 IFT_LOCALTALK = 0x2a IFT_LOOP = 0x18 + IFT_MBIM = 0xfa IFT_MEDIAMAILOVERIP = 0x8b IFT_MFSIGLINK = 0xa7 IFT_MIOX25 = 0x26 @@ -721,8 +757,6 @@ const ( IPPROTO_AH = 0x33 IPPROTO_CARP = 0x70 IPPROTO_DIVERT = 0x102 - IPPROTO_DIVERT_INIT = 0x2 - IPPROTO_DIVERT_RESP = 0x1 IPPROTO_DONE = 0x101 IPPROTO_DSTOPTS = 0x3c IPPROTO_EGP = 0x8 @@ -779,6 +813,7 @@ const ( IPV6_LEAVE_GROUP = 0xd IPV6_MAXHLIM = 0xff IPV6_MAXPACKET = 0xffff + IPV6_MINHOPCOUNT = 0x41 IPV6_MMTU = 0x500 IPV6_MULTICAST_HOPS = 0xa IPV6_MULTICAST_IF = 0x9 @@ -818,12 +853,12 @@ const ( IP_DEFAULT_MULTICAST_LOOP = 0x1 IP_DEFAULT_MULTICAST_TTL = 0x1 IP_DF = 0x4000 - IP_DIVERTFL = 0x1022 IP_DROP_MEMBERSHIP = 0xd IP_ESP_NETWORK_LEVEL = 0x16 IP_ESP_TRANS_LEVEL = 0x15 IP_HDRINCL = 0x2 IP_IPCOMP_LEVEL = 0x1d + IP_IPDEFTTL = 0x25 IP_IPSECFLOWINFO = 0x24 IP_IPSEC_LOCAL_AUTH = 0x1b IP_IPSEC_LOCAL_CRED = 0x19 @@ -857,10 +892,12 @@ const ( IP_RETOPTS = 0x8 IP_RF = 0x8000 IP_RTABLE = 0x1021 + IP_SENDSRCADDR = 0x7 IP_TOS = 0x3 IP_TTL = 0x4 ISIG = 0x80 ISTRIP = 0x20 + IUCLC = 0x1000 IXANY = 0x800 IXOFF = 0x400 IXON = 0x200 @@ -881,25 +918,28 @@ const ( MADV_SPACEAVAIL = 0x5 MADV_WILLNEED = 0x3 MAP_ANON = 0x1000 - MAP_COPY = 0x4 + MAP_ANONYMOUS = 0x1000 + MAP_COPY = 0x2 MAP_FILE = 0x0 MAP_FIXED = 0x10 - MAP_FLAGMASK = 0x1ff7 - MAP_HASSEMAPHORE = 0x200 - MAP_INHERIT = 0x80 + MAP_FLAGMASK = 0x7ff7 + MAP_HASSEMAPHORE = 0x0 + MAP_INHERIT = 0x0 MAP_INHERIT_COPY = 0x1 - MAP_INHERIT_DONATE_COPY = 0x3 MAP_INHERIT_NONE = 0x2 MAP_INHERIT_SHARE = 0x0 - MAP_NOEXTEND = 0x100 - MAP_NORESERVE = 0x40 + MAP_INHERIT_ZERO = 0x3 + MAP_NOEXTEND = 0x0 + MAP_NORESERVE = 0x0 MAP_PRIVATE = 0x2 - MAP_RENAME = 0x20 + MAP_RENAME = 0x0 MAP_SHARED = 0x1 - MAP_TRYFIXED = 0x400 + MAP_STACK = 0x4000 + MAP_TRYFIXED = 0x0 MCL_CURRENT = 0x1 MCL_FUTURE = 0x2 MSG_BCAST = 0x100 + MSG_CMSG_CLOEXEC = 0x800 MSG_CTRUNC = 0x20 MSG_DONTROUTE = 0x4 MSG_DONTWAIT = 0x80 @@ -917,11 +957,14 @@ const ( NET_RT_DUMP = 0x1 NET_RT_FLAGS = 0x2 NET_RT_IFLIST = 0x3 - NET_RT_MAXID = 0x6 + NET_RT_IFNAMES = 0x6 + NET_RT_MAXID = 0x7 NET_RT_STATS = 0x4 NET_RT_TABLE = 0x5 NOFLSH = 0x80000000 + NOKERNINFO = 0x2000000 NOTE_ATTRIB = 0x8 + NOTE_CHANGE = 0x1 NOTE_CHILD = 0x4 NOTE_DELETE = 0x1 NOTE_EOF = 0x2 @@ -940,11 +983,13 @@ const ( NOTE_TRUNCATE = 0x80 NOTE_WRITE = 0x2 OCRNL = 0x10 + OLCUC = 0x20 ONLCR = 0x2 ONLRET = 0x80 ONOCR = 0x40 ONOEOT = 0x8 OPOST = 0x1 + OXTABS = 0x4 O_ACCMODE = 0x3 O_APPEND = 0x8 O_ASYNC = 0x40 @@ -982,23 +1027,32 @@ const ( RLIMIT_CPU = 0x0 RLIMIT_DATA = 0x2 RLIMIT_FSIZE = 0x1 + RLIMIT_MEMLOCK = 0x6 RLIMIT_NOFILE = 0x8 + RLIMIT_NPROC = 0x7 + RLIMIT_RSS = 0x5 RLIMIT_STACK = 0x3 RLIM_INFINITY = 0x7fffffffffffffff RTAX_AUTHOR = 0x6 + RTAX_BFD = 0xb RTAX_BRD = 0x7 + RTAX_DNS = 0xc RTAX_DST = 0x0 RTAX_GATEWAY = 0x1 RTAX_GENMASK = 0x3 RTAX_IFA = 0x5 RTAX_IFP = 0x4 RTAX_LABEL = 0xa - RTAX_MAX = 0xb + RTAX_MAX = 0xf RTAX_NETMASK = 0x2 + RTAX_SEARCH = 0xe RTAX_SRC = 0x8 RTAX_SRCMASK = 0x9 + RTAX_STATIC = 0xd RTA_AUTHOR = 0x40 + RTA_BFD = 0x800 RTA_BRD = 0x80 + RTA_DNS = 0x1000 RTA_DST = 0x1 RTA_GATEWAY = 0x2 RTA_GENMASK = 0x8 @@ -1006,34 +1060,39 @@ const ( RTA_IFP = 0x10 RTA_LABEL = 0x400 RTA_NETMASK = 0x4 + RTA_SEARCH = 0x4000 RTA_SRC = 0x100 RTA_SRCMASK = 0x200 + RTA_STATIC = 0x2000 RTF_ANNOUNCE = 0x4000 + RTF_BFD = 0x1000000 RTF_BLACKHOLE = 0x1000 + RTF_BROADCAST = 0x400000 + RTF_CACHED = 0x20000 RTF_CLONED = 0x10000 RTF_CLONING = 0x100 + RTF_CONNECTED = 0x800000 RTF_DONE = 0x40 RTF_DYNAMIC = 0x10 - RTF_FMASK = 0x10f808 + RTF_FMASK = 0x110fc08 RTF_GATEWAY = 0x2 RTF_HOST = 0x4 RTF_LLINFO = 0x400 - RTF_MASK = 0x80 + RTF_LOCAL = 0x200000 RTF_MODIFIED = 0x20 RTF_MPATH = 0x40000 RTF_MPLS = 0x100000 + RTF_MULTICAST = 0x200 RTF_PERMANENT_ARP = 0x2000 RTF_PROTO1 = 0x8000 RTF_PROTO2 = 0x4000 RTF_PROTO3 = 0x2000 RTF_REJECT = 0x8 - RTF_SOURCE = 0x20000 RTF_STATIC = 0x800 - RTF_TUNNEL = 0x100000 RTF_UP = 0x1 RTF_USETRAILERS = 0x8000 - RTF_XRESOLVE = 0x200 RTM_ADD = 0x1 + RTM_BFD = 0x12 RTM_CHANGE = 0x3 RTM_DELADDR = 0xd RTM_DELETE = 0x2 @@ -1041,11 +1100,13 @@ const ( RTM_GET = 0x4 RTM_IFANNOUNCE = 0xf RTM_IFINFO = 0xe + RTM_INVALIDATE = 0x11 RTM_LOCK = 0x8 RTM_LOSING = 0x5 RTM_MAXSIZE = 0x800 RTM_MISS = 0x7 RTM_NEWADDR = 0xc + RTM_PROPOSAL = 0x13 RTM_REDIRECT = 0x6 RTM_RESOLVE = 0xb RTM_RTTUNIT = 0xf4240 @@ -1058,6 +1119,8 @@ const ( RTV_RTTVAR = 0x80 RTV_SPIPE = 0x10 RTV_SSTHRESH = 0x20 + RT_TABLEID_BITS = 0x8 + RT_TABLEID_MASK = 0xff RT_TABLEID_MAX = 0xff RUSAGE_CHILDREN = -0x1 RUSAGE_SELF = 0x0 @@ -1070,55 +1133,55 @@ const ( SIOCADDMULTI = 0x80206931 SIOCAIFADDR = 0x8040691a SIOCAIFGROUP = 0x80286987 - SIOCALIFADDR = 0x8218691c SIOCATMARK = 0x40047307 - SIOCBRDGADD = 0x8058693c - SIOCBRDGADDS = 0x80586941 - SIOCBRDGARL = 0x806e694d + SIOCBRDGADD = 0x8060693c + SIOCBRDGADDL = 0x80606949 + SIOCBRDGADDS = 0x80606941 + SIOCBRDGARL = 0x808c694d SIOCBRDGDADDR = 0x81286947 - SIOCBRDGDEL = 0x8058693d - SIOCBRDGDELS = 0x80586942 - SIOCBRDGFLUSH = 0x80586948 - SIOCBRDGFRL = 0x806e694e - SIOCBRDGGCACHE = 0xc0146941 - SIOCBRDGGFD = 0xc0146952 - SIOCBRDGGHT = 0xc0146951 - SIOCBRDGGIFFLGS = 0xc058693e - SIOCBRDGGMA = 0xc0146953 + SIOCBRDGDEL = 0x8060693d + SIOCBRDGDELS = 0x80606942 + SIOCBRDGFLUSH = 0x80606948 + SIOCBRDGFRL = 0x808c694e + SIOCBRDGGCACHE = 0xc0186941 + SIOCBRDGGFD = 0xc0186952 + SIOCBRDGGHT = 0xc0186951 + SIOCBRDGGIFFLGS = 0xc060693e + SIOCBRDGGMA = 0xc0186953 SIOCBRDGGPARAM = 0xc0406958 - SIOCBRDGGPRI = 0xc0146950 + SIOCBRDGGPRI = 0xc0186950 SIOCBRDGGRL = 0xc030694f - SIOCBRDGGSIFS = 0xc058693c - SIOCBRDGGTO = 0xc0146946 - SIOCBRDGIFS = 0xc0586942 + SIOCBRDGGTO = 0xc0186946 + SIOCBRDGIFS = 0xc0606942 SIOCBRDGRTS = 0xc0206943 SIOCBRDGSADDR = 0xc1286944 - SIOCBRDGSCACHE = 0x80146940 - SIOCBRDGSFD = 0x80146952 - SIOCBRDGSHT = 0x80146951 - SIOCBRDGSIFCOST = 0x80586955 - SIOCBRDGSIFFLGS = 0x8058693f - SIOCBRDGSIFPRIO = 0x80586954 - SIOCBRDGSMA = 0x80146953 - SIOCBRDGSPRI = 0x80146950 - SIOCBRDGSPROTO = 0x8014695a - SIOCBRDGSTO = 0x80146945 - SIOCBRDGSTXHC = 0x80146959 + SIOCBRDGSCACHE = 0x80186940 + SIOCBRDGSFD = 0x80186952 + SIOCBRDGSHT = 0x80186951 + SIOCBRDGSIFCOST = 0x80606955 + SIOCBRDGSIFFLGS = 0x8060693f + SIOCBRDGSIFPRIO = 0x80606954 + SIOCBRDGSIFPROT = 0x8060694a + SIOCBRDGSMA = 0x80186953 + SIOCBRDGSPRI = 0x80186950 + SIOCBRDGSPROTO = 0x8018695a + SIOCBRDGSTO = 0x80186945 + SIOCBRDGSTXHC = 0x80186959 SIOCDELMULTI = 0x80206932 SIOCDIFADDR = 0x80206919 SIOCDIFGROUP = 0x80286989 + SIOCDIFPARENT = 0x802069b4 SIOCDIFPHYADDR = 0x80206949 - SIOCDLIFADDR = 0x8218691e + SIOCDVNETID = 0x802069af SIOCGETKALIVE = 0xc01869a4 SIOCGETLABEL = 0x8020699a + SIOCGETMPWCFG = 0xc02069ae SIOCGETPFLOW = 0xc02069fe SIOCGETPFSYNC = 0xc02069f8 SIOCGETSGCNT = 0xc0207534 SIOCGETVIFCNT = 0xc0287533 SIOCGETVLAN = 0xc0206990 - SIOCGHIWAT = 0x40047301 SIOCGIFADDR = 0xc0206921 - SIOCGIFASYNCMAP = 0xc020697c SIOCGIFBRDADDR = 0xc0206923 SIOCGIFCONF = 0xc0106924 SIOCGIFDATA = 0xc020691b @@ -1130,37 +1193,41 @@ const ( SIOCGIFGMEMB = 0xc028698a SIOCGIFGROUP = 0xc0286988 SIOCGIFHARDMTU = 0xc02069a5 - SIOCGIFMEDIA = 0xc0306936 + SIOCGIFLLPRIO = 0xc02069b6 + SIOCGIFMEDIA = 0xc0406938 SIOCGIFMETRIC = 0xc0206917 SIOCGIFMTU = 0xc020697e SIOCGIFNETMASK = 0xc0206925 - SIOCGIFPDSTADDR = 0xc0206948 + SIOCGIFPAIR = 0xc02069b1 + SIOCGIFPARENT = 0xc02069b3 SIOCGIFPRIORITY = 0xc020699c - SIOCGIFPSRCADDR = 0xc0206947 SIOCGIFRDOMAIN = 0xc02069a0 SIOCGIFRTLABEL = 0xc0206983 - SIOCGIFTIMESLOT = 0xc0206986 + SIOCGIFRXR = 0x802069aa SIOCGIFXFLAGS = 0xc020699e - SIOCGLIFADDR = 0xc218691d SIOCGLIFPHYADDR = 0xc218694b + SIOCGLIFPHYDF = 0xc02069c2 SIOCGLIFPHYRTABLE = 0xc02069a2 SIOCGLIFPHYTTL = 0xc02069a9 - SIOCGLOWAT = 0x40047303 SIOCGPGRP = 0x40047309 SIOCGSPPPPARAMS = 0xc0206994 + SIOCGUMBINFO = 0xc02069be + SIOCGUMBPARAM = 0xc02069c0 SIOCGVH = 0xc02069f6 + SIOCGVNETFLOWID = 0xc02069c4 SIOCGVNETID = 0xc02069a7 + SIOCIFAFATTACH = 0x801169ab + SIOCIFAFDETACH = 0x801169ac SIOCIFCREATE = 0x8020697a SIOCIFDESTROY = 0x80206979 SIOCIFGCLONERS = 0xc0106978 SIOCSETKALIVE = 0x801869a3 SIOCSETLABEL = 0x80206999 + SIOCSETMPWCFG = 0x802069ad SIOCSETPFLOW = 0x802069fd SIOCSETPFSYNC = 0x802069f7 SIOCSETVLAN = 0x8020698f - SIOCSHIWAT = 0x80047300 SIOCSIFADDR = 0x8020690c - SIOCSIFASYNCMAP = 0x8020697d SIOCSIFBRDADDR = 0x80206913 SIOCSIFDESCR = 0x80206980 SIOCSIFDSTADDR = 0x8020690e @@ -1168,25 +1235,36 @@ const ( SIOCSIFGATTR = 0x8028698c SIOCSIFGENERIC = 0x80206939 SIOCSIFLLADDR = 0x8020691f - SIOCSIFMEDIA = 0xc0206935 + SIOCSIFLLPRIO = 0x802069b5 + SIOCSIFMEDIA = 0xc0206937 SIOCSIFMETRIC = 0x80206918 SIOCSIFMTU = 0x8020697f SIOCSIFNETMASK = 0x80206916 - SIOCSIFPHYADDR = 0x80406946 + SIOCSIFPAIR = 0x802069b0 + SIOCSIFPARENT = 0x802069b2 SIOCSIFPRIORITY = 0x8020699b SIOCSIFRDOMAIN = 0x8020699f SIOCSIFRTLABEL = 0x80206982 - SIOCSIFTIMESLOT = 0x80206985 SIOCSIFXFLAGS = 0x8020699d SIOCSLIFPHYADDR = 0x8218694a + SIOCSLIFPHYDF = 0x802069c1 SIOCSLIFPHYRTABLE = 0x802069a1 SIOCSLIFPHYTTL = 0x802069a8 - SIOCSLOWAT = 0x80047302 SIOCSPGRP = 0x80047308 SIOCSSPPPPARAMS = 0x80206993 + SIOCSUMBPARAM = 0x802069bf SIOCSVH = 0xc02069f5 + SIOCSVNETFLOWID = 0x802069c3 SIOCSVNETID = 0x802069a6 + SIOCSWGDPID = 0xc018695b + SIOCSWGMAXFLOW = 0xc0186960 + SIOCSWGMAXGROUP = 0xc018695d + SIOCSWSDPID = 0x8018695c + SIOCSWSPORTNO = 0xc060695f + SOCK_CLOEXEC = 0x8000 SOCK_DGRAM = 0x2 + SOCK_DNS = 0x1000 + SOCK_NONBLOCK = 0x4000 SOCK_RAW = 0x3 SOCK_RDM = 0x4 SOCK_SEQPACKET = 0x5 @@ -1217,9 +1295,14 @@ const ( SO_TIMESTAMP = 0x800 SO_TYPE = 0x1008 SO_USELOOPBACK = 0x40 + SO_ZEROIZE = 0x2000 TCIFLUSH = 0x1 + TCIOFF = 0x3 TCIOFLUSH = 0x3 + TCION = 0x4 TCOFLUSH = 0x2 + TCOOFF = 0x1 + TCOON = 0x2 TCP_MAXBURST = 0x4 TCP_MAXSEG = 0x2 TCP_MAXWIN = 0xffff @@ -1229,11 +1312,12 @@ const ( TCP_MSS = 0x200 TCP_NODELAY = 0x1 TCP_NOPUSH = 0x10 - TCP_NSTATES = 0xb TCP_SACK_ENABLE = 0x8 TCSAFLUSH = 0x2 TIOCCBRK = 0x2000747a TIOCCDTR = 0x20007478 + TIOCCHKVERAUTH = 0x2000741e + TIOCCLRVERAUTH = 0x2000741d TIOCCONS = 0x80047462 TIOCDRAIN = 0x2000745e TIOCEXCL = 0x2000740d @@ -1288,16 +1372,19 @@ const ( TIOCSETAF = 0x802c7416 TIOCSETAW = 0x802c7415 TIOCSETD = 0x8004741b + TIOCSETVERAUTH = 0x8004741c TIOCSFLAGS = 0x8004745c TIOCSIG = 0x8004745f TIOCSPGRP = 0x80047476 TIOCSTART = 0x2000746e - TIOCSTAT = 0x80047465 + TIOCSTAT = 0x20007465 TIOCSTI = 0x80017472 TIOCSTOP = 0x2000746f TIOCSTSTAMP = 0x8008745a TIOCSWINSZ = 0x80087467 TIOCUCNTL = 0x80047466 + TIOCUCNTL_CBRK = 0x7a + TIOCUCNTL_SBRK = 0x7b TOSTOP = 0x400000 VDISCARD = 0xf VDSUSP = 0xb @@ -1309,6 +1396,18 @@ const ( VKILL = 0x5 VLNEXT = 0xe VMIN = 0x10 + VM_ANONMIN = 0x7 + VM_LOADAVG = 0x2 + VM_MAXID = 0xc + VM_MAXSLP = 0xa + VM_METER = 0x1 + VM_NKMEMPAGES = 0x6 + VM_PSSTRINGS = 0x3 + VM_SWAPENCRYPT = 0x5 + VM_USPACE = 0xb + VM_UVMEXP = 0x4 + VM_VNODEMIN = 0x9 + VM_VTEXTMIN = 0x8 VQUIT = 0x9 VREPRINT = 0x6 VSTART = 0xc @@ -1321,8 +1420,8 @@ const ( WCONTINUED = 0x8 WCOREFLAG = 0x80 WNOHANG = 0x1 - WSTOPPED = 0x7f WUNTRACED = 0x2 + XCASE = 0x1000000 ) // Errors @@ -1336,6 +1435,7 @@ const ( EALREADY = syscall.Errno(0x25) EAUTH = syscall.Errno(0x50) EBADF = syscall.Errno(0x9) + EBADMSG = syscall.Errno(0x5c) EBADRPC = syscall.Errno(0x48) EBUSY = syscall.Errno(0x10) ECANCELED = syscall.Errno(0x58) @@ -1362,7 +1462,7 @@ const ( EIPSEC = syscall.Errno(0x52) EISCONN = syscall.Errno(0x38) EISDIR = syscall.Errno(0x15) - ELAST = syscall.Errno(0x5b) + ELAST = syscall.Errno(0x5f) ELOOP = syscall.Errno(0x3e) EMEDIUMTYPE = syscall.Errno(0x56) EMFILE = syscall.Errno(0x18) @@ -1390,12 +1490,14 @@ const ( ENOTCONN = syscall.Errno(0x39) ENOTDIR = syscall.Errno(0x14) ENOTEMPTY = syscall.Errno(0x42) + ENOTRECOVERABLE = syscall.Errno(0x5d) ENOTSOCK = syscall.Errno(0x26) ENOTSUP = syscall.Errno(0x5b) ENOTTY = syscall.Errno(0x19) ENXIO = syscall.Errno(0x6) EOPNOTSUPP = syscall.Errno(0x2d) EOVERFLOW = syscall.Errno(0x57) + EOWNERDEAD = syscall.Errno(0x5e) EPERM = syscall.Errno(0x1) EPFNOSUPPORT = syscall.Errno(0x2e) EPIPE = syscall.Errno(0x20) @@ -1403,6 +1505,7 @@ const ( EPROCUNAVAIL = syscall.Errno(0x4c) EPROGMISMATCH = syscall.Errno(0x4b) EPROGUNAVAIL = syscall.Errno(0x4a) + EPROTO = syscall.Errno(0x5f) EPROTONOSUPPORT = syscall.Errno(0x2b) EPROTOTYPE = syscall.Errno(0x29) ERANGE = syscall.Errno(0x22) @@ -1499,7 +1602,7 @@ var errorList = [...]struct { {32, "EPIPE", "broken pipe"}, {33, "EDOM", "numerical argument out of domain"}, {34, "ERANGE", "result too large"}, - {35, "EWOULDBLOCK", "resource temporarily unavailable"}, + {35, "EAGAIN", "resource temporarily unavailable"}, {36, "EINPROGRESS", "operation now in progress"}, {37, "EALREADY", "operation already in progress"}, {38, "ENOTSOCK", "socket operation on non-socket"}, @@ -1555,7 +1658,11 @@ var errorList = [...]struct { {88, "ECANCELED", "operation canceled"}, {89, "EIDRM", "identifier removed"}, {90, "ENOMSG", "no message of desired type"}, - {91, "ELAST", "not supported"}, + {91, "ENOTSUP", "not supported"}, + {92, "EBADMSG", "bad message"}, + {93, "ENOTRECOVERABLE", "state not recoverable"}, + {94, "EOWNERDEAD", "previous owner died"}, + {95, "ELAST", "protocol error"}, } // Signal table diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go index 4c9f72756..ac02d4d84 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_386.go @@ -399,6 +399,83 @@ func pipe() (r int, w int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func getxattr(path string, attr string, dest *byte, size int, position uint32, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_GETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(position), uintptr(options)) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func setxattr(path string, attr string, data *byte, size int, position uint32, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_SETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(data)), uintptr(size), uintptr(position), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func removexattr(path string, attr string, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_REMOVEXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func listxattr(path string, dest *byte, size int, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_LISTXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(options), 0, 0) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func kill(pid int, signum int, posix int) (err error) { _, _, e1 := Syscall(SYS_KILL, uintptr(pid), uintptr(signum), uintptr(posix)) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go index 256237773..1dd3cfa0e 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_amd64.go @@ -399,6 +399,83 @@ func pipe() (r int, w int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func getxattr(path string, attr string, dest *byte, size int, position uint32, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_GETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(position), uintptr(options)) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func setxattr(path string, attr string, data *byte, size int, position uint32, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_SETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(data)), uintptr(size), uintptr(position), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func removexattr(path string, attr string, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_REMOVEXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func listxattr(path string, dest *byte, size int, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_LISTXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(options), 0, 0) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func kill(pid int, signum int, posix int) (err error) { _, _, e1 := Syscall(SYS_KILL, uintptr(pid), uintptr(signum), uintptr(posix)) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go index 4ae787e49..cab46e74b 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm.go @@ -1,4 +1,4 @@ -// mksyscall.pl -tags darwin,arm syscall_bsd.go syscall_darwin.go syscall_darwin_arm.go +// mksyscall.pl -l32 -tags darwin,arm syscall_bsd.go syscall_darwin.go syscall_darwin_arm.go // Code generated by the command above; see README.md. DO NOT EDIT. // +build darwin,arm @@ -399,6 +399,83 @@ func pipe() (r int, w int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func getxattr(path string, attr string, dest *byte, size int, position uint32, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_GETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(position), uintptr(options)) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func setxattr(path string, attr string, data *byte, size int, position uint32, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_SETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(data)), uintptr(size), uintptr(position), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func removexattr(path string, attr string, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_REMOVEXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func listxattr(path string, dest *byte, size int, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_LISTXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(options), 0, 0) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func kill(pid int, signum int, posix int) (err error) { _, _, e1 := Syscall(SYS_KILL, uintptr(pid), uintptr(signum), uintptr(posix)) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go index 14ed6886c..13478dd0b 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_darwin_arm64.go @@ -399,6 +399,83 @@ func pipe() (r int, w int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func getxattr(path string, attr string, dest *byte, size int, position uint32, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_GETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(position), uintptr(options)) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func setxattr(path string, attr string, data *byte, size int, position uint32, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_SETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(data)), uintptr(size), uintptr(position), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func removexattr(path string, attr string, options int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_REMOVEXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(options)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func listxattr(path string, dest *byte, size int, options int) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_LISTXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(dest)), uintptr(size), uintptr(options), 0, 0) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func kill(pid int, signum int, posix int) (err error) { _, _, e1 := Syscall(SYS_KILL, uintptr(pid), uintptr(signum), uintptr(posix)) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go index 433becfd0..237e96095 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_386.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(off>>32), uintptr(len), uintptr(len>>32)) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe(p *[2]_C_int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE, uintptr(unsafe.Pointer(p)), 0, 0) if e1 != 0 { @@ -1565,6 +1519,34 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { + var _p0 unsafe.Pointer + if len(events) > 0 { + _p0 = unsafe.Pointer(&events[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) + n = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall6(SYS_FADVISE64_64, uintptr(fd), uintptr(offset), uintptr(offset>>32), uintptr(length), uintptr(length>>32), uintptr(advice)) if e1 != 0 { @@ -1869,6 +1851,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func getgroups(n int, list *_Gid_t) (nn int, err error) { r0, _, e1 := RawSyscall(SYS_GETGROUPS32, uintptr(n), uintptr(unsafe.Pointer(list)), 0) nn = int(r0) @@ -1912,23 +1904,6 @@ func mmap2(addr uintptr, length uintptr, prot int, flags int, fd int, pageOffset // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { - var _p0 unsafe.Pointer - if len(events) > 0 { - _p0 = unsafe.Pointer(&events[0]) - } else { - _p0 = unsafe.Pointer(&_zero) - } - r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) - n = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Pause() (err error) { _, _, e1 := Syscall(SYS_PAUSE, 0, 0, 0) if e1 != 0 { @@ -1959,6 +1934,21 @@ func setrlimit(resource int, rlim *rlimit32) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -1995,6 +1985,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func poll(fds *PollFd, nfds int, timeout int) (n int, err error) { r0, _, e1 := Syscall(SYS_POLL, uintptr(unsafe.Pointer(fds)), uintptr(nfds), uintptr(timeout)) n = int(r0) diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go index 33c02b269..9b40f580b 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_amd64.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { @@ -1545,6 +1499,17 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { @@ -1937,6 +1902,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2127,6 +2102,21 @@ func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int6 // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Utime(path string, buf *Utimbuf) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -2142,6 +2132,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe(p *[2]_C_int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE, uintptr(unsafe.Pointer(p)), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go index f91b56c21..d93cc2f3b 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(off>>32), uintptr(len), uintptr(len>>32)) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe2(p *[2]_C_int, flags int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE2, uintptr(unsafe.Pointer(p)), uintptr(flags), 0) if e1 != 0 { @@ -1734,6 +1688,34 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { + var _p0 unsafe.Pointer + if len(events) > 0 { + _p0 = unsafe.Pointer(&events[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) + n = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fchown(fd int, uid int, gid int) (err error) { _, _, e1 := Syscall(SYS_FCHOWN32, uintptr(fd), uintptr(uid), uintptr(gid)) if e1 != 0 { @@ -1852,6 +1834,16 @@ func Lstat(path string, stat *Stat_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Pause() (err error) { + _, _, e1 := Syscall(SYS_PAUSE, 0, 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func sendfile(outfd int, infd int, offset *int64, count int) (written int, err error) { r0, _, e1 := Syscall6(SYS_SENDFILE64, uintptr(outfd), uintptr(infd), uintptr(unsafe.Pointer(offset)), uintptr(count), 0, 0) written = int(r0) @@ -1970,6 +1962,31 @@ func Stat(path string, stat *Stat_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -1980,25 +1997,13 @@ func Gettimeofday(tv *Timeval) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { - var _p0 unsafe.Pointer - if len(events) > 0 { - _p0 = unsafe.Pointer(&events[0]) - } else { - _p0 = unsafe.Pointer(&_zero) +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return } - r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) - n = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - -func Pause() (err error) { - _, _, e1 := Syscall(SYS_PAUSE, 0, 0, 0) + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) if e1 != 0 { err = errnoErr(e1) } diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go index 52d759525..5f7d0213b 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_arm64.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go index 970a5c132..e76170528 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off>>32), uintptr(off), uintptr(len>>32), uintptr(len)) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { @@ -1545,6 +1499,34 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { + var _p0 unsafe.Pointer + if len(events) > 0 { + _p0 = unsafe.Pointer(&events[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) + n = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall9(SYS_FADVISE64, uintptr(fd), 0, uintptr(offset>>32), uintptr(offset), uintptr(length>>32), uintptr(length), uintptr(advice), 0, 0) if e1 != 0 { @@ -1794,6 +1776,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2004,6 +1996,21 @@ func Iopl(level int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -2025,6 +2032,36 @@ func Time(t *Time_t) (tt Time_t, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Utime(path string, buf *Utimbuf) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIME, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(buf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Lstat(path string, stat *Stat_t) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -2080,38 +2117,6 @@ func Stat(path string, stat *Stat_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Utime(path string, buf *Utimbuf) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIME, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(buf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - -func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { - var _p0 unsafe.Pointer - if len(events) > 0 { - _p0 = unsafe.Pointer(&events[0]) - } else { - _p0 = unsafe.Pointer(&_zero) - } - r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) - n = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Pause() (err error) { _, _, e1 := Syscall(SYS_PAUSE, 0, 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go index b989d0f28..382e072f5 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { @@ -1545,6 +1499,17 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { @@ -1881,6 +1846,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2071,6 +2046,21 @@ func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int6 // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -2096,6 +2086,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe2(p *[2]_C_int, flags int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE2, uintptr(unsafe.Pointer(p)), uintptr(flags), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go index 1f8d14cac..da33eb12b 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mips64le.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { @@ -1545,6 +1499,17 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { @@ -1881,6 +1846,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2071,6 +2046,21 @@ func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int6 // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -2096,6 +2086,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe2(p *[2]_C_int, flags int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE2, uintptr(unsafe.Pointer(p)), uintptr(flags), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go index a9c7e520e..f6e45189a 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_mipsle.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(off>>32), uintptr(len), uintptr(len>>32)) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { @@ -1545,6 +1499,34 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { + var _p0 unsafe.Pointer + if len(events) > 0 { + _p0 = unsafe.Pointer(&events[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) + n = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall9(SYS_FADVISE64, uintptr(fd), 0, uintptr(offset), uintptr(offset>>32), uintptr(length), uintptr(length>>32), uintptr(advice), 0, 0) if e1 != 0 { @@ -1794,6 +1776,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2004,6 +1996,21 @@ func Iopl(level int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -2025,6 +2032,36 @@ func Time(t *Time_t) (tt Time_t, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Utime(path string, buf *Utimbuf) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIME, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(buf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Lstat(path string, stat *Stat_t) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -2080,38 +2117,6 @@ func Stat(path string, stat *Stat_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Utime(path string, buf *Utimbuf) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIME, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(buf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - -func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { - var _p0 unsafe.Pointer - if len(events) > 0 { - _p0 = unsafe.Pointer(&events[0]) - } else { - _p0 = unsafe.Pointer(&_zero) - } - r0, _, e1 := Syscall6(SYS_EPOLL_WAIT, uintptr(epfd), uintptr(_p0), uintptr(len(events)), uintptr(msec), 0, 0) - n = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Pause() (err error) { _, _, e1 := Syscall(SYS_PAUSE, 0, 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go index 3bb9a2099..0ae2aa842 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,42 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Dup2(oldfd int, newfd int) (err error) { + _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { @@ -1552,16 +1527,6 @@ func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { - _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall6(SYS_FADVISE64, uintptr(fd), uintptr(offset), uintptr(length), uintptr(advice), 0, 0) if e1 != 0 { @@ -1963,6 +1928,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2153,6 +2128,21 @@ func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int6 // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -2189,6 +2179,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe(p *[2]_C_int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE, uintptr(unsafe.Pointer(p)), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go index 56116623d..fa16c165e 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_ppc64le.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,42 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Dup2(oldfd int, newfd int) (err error) { + _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { @@ -1552,16 +1527,6 @@ func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Dup2(oldfd int, newfd int) (err error) { - _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall6(SYS_FADVISE64, uintptr(fd), uintptr(offset), uintptr(length), uintptr(advice), 0, 0) if e1 != 0 { @@ -1963,6 +1928,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func accept(s int, rsa *RawSockaddrAny, addrlen *_Socklen) (fd int, err error) { r0, _, e1 := Syscall(SYS_ACCEPT, uintptr(s), uintptr(unsafe.Pointer(rsa)), uintptr(unsafe.Pointer(addrlen))) fd = int(r0) @@ -2153,6 +2128,21 @@ func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int6 // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -2189,6 +2179,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe(p *[2]_C_int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE, uintptr(unsafe.Pointer(p)), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go index 9696a0199..38f903cd3 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_s390x.go @@ -143,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -173,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -494,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -544,21 +508,6 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fallocate(fd int, mode uint32, off int64, len int64) (err error) { _, _, e1 := Syscall6(SYS_FALLOCATE, uintptr(fd), uintptr(mode), uintptr(off), uintptr(len), 0, 0) if e1 != 0 { @@ -1366,16 +1315,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1535,6 +1474,21 @@ func Munlockall() (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func faccessat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Dup2(oldfd int, newfd int) (err error) { _, _, e1 := Syscall(SYS_DUP2, uintptr(oldfd), uintptr(newfd), 0) if e1 != 0 { @@ -1545,6 +1499,17 @@ func Dup2(oldfd int, newfd int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func EpollCreate(size int) (fd int, err error) { + r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func EpollWait(epfd int, events []EpollEvent, msec int) (n int, err error) { var _p0 unsafe.Pointer if len(events) > 0 { @@ -1923,6 +1888,16 @@ func Truncate(path string, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Ustat(dev int, ubuf *Ustat_t) (err error) { + _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func getgroups(n int, list *_Gid_t) (nn int, err error) { r0, _, e1 := RawSyscall(SYS_GETGROUPS, uintptr(n), uintptr(unsafe.Pointer(list)), 0) nn = int(r0) @@ -1944,6 +1919,21 @@ func setgroups(n int, list *_Gid_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -1969,6 +1959,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe2(p *[2]_C_int, flags int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE2, uintptr(unsafe.Pointer(p)), uintptr(flags), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go b/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go index c01b3b6ba..b26aee957 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_linux_sparc64.go @@ -14,6 +14,31 @@ var _ syscall.Errno // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func fchmodat(dirfd int, path string, mode uint32) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FCHMODAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func ioctl(fd int, req uint, arg uintptr) (err error) { + _, _, e1 := Syscall(SYS_IOCTL, uintptr(fd), uintptr(req), uintptr(arg)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Linkat(olddirfd int, oldpath string, newdirfd int, newpath string, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(oldpath) @@ -118,21 +143,6 @@ func Unlinkat(dirfd int, path string, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func utimes(path string, times *[2]Timeval) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -148,16 +158,6 @@ func utimensat(dirfd int, path string, times *[2]Timespec, flags int) (err error // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func futimesat(dirfd int, path *byte, times *[2]Timeval) (err error) { - _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(path)), uintptr(unsafe.Pointer(times))) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Getcwd(buf []byte) (n int, err error) { var _p0 unsafe.Pointer if len(buf) > 0 { @@ -186,6 +186,104 @@ func wait4(pid int, wstatus *_C_int, options int, rusage *Rusage) (wpid int, err // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func KeyctlInt(cmd int, arg2 int, arg3 int, arg4 int, arg5 int) (ret int, err error) { + r0, _, e1 := Syscall6(SYS_KEYCTL, uintptr(cmd), uintptr(arg2), uintptr(arg3), uintptr(arg4), uintptr(arg5), 0) + ret = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func KeyctlBuffer(cmd int, arg2 int, buf []byte, arg5 int) (ret int, err error) { + var _p0 unsafe.Pointer + if len(buf) > 0 { + _p0 = unsafe.Pointer(&buf[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_KEYCTL, uintptr(cmd), uintptr(arg2), uintptr(_p0), uintptr(len(buf)), uintptr(arg5), 0) + ret = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func keyctlJoin(cmd int, arg2 string) (ret int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(arg2) + if err != nil { + return + } + r0, _, e1 := Syscall(SYS_KEYCTL, uintptr(cmd), uintptr(unsafe.Pointer(_p0)), 0) + ret = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func keyctlSearch(cmd int, arg2 int, arg3 string, arg4 string, arg5 int) (ret int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(arg3) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(arg4) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_KEYCTL, uintptr(cmd), uintptr(arg2), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(arg5), 0) + ret = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func keyctlIOV(cmd int, arg2 int, payload []Iovec, arg5 int) (err error) { + var _p0 unsafe.Pointer + if len(payload) > 0 { + _p0 = unsafe.Pointer(&payload[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + _, _, e1 := Syscall6(SYS_KEYCTL, uintptr(cmd), uintptr(arg2), uintptr(_p0), uintptr(len(payload)), uintptr(arg5), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func keyctlDH(cmd int, arg2 *KeyctlDHParams, buf []byte) (ret int, err error) { + var _p0 unsafe.Pointer + if len(buf) > 0 { + _p0 = unsafe.Pointer(&buf[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_KEYCTL, uintptr(cmd), uintptr(unsafe.Pointer(arg2)), uintptr(_p0), uintptr(len(buf)), 0, 0) + ret = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func ptrace(request int, pid int, addr uintptr, data uintptr) (err error) { _, _, e1 := Syscall6(SYS_PTRACE, uintptr(request), uintptr(pid), uintptr(addr), uintptr(data), 0, 0) if e1 != 0 { @@ -251,6 +349,33 @@ func Acct(path string) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func AddKey(keyType string, description string, payload []byte, ringid int) (id int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(keyType) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(description) + if err != nil { + return + } + var _p2 unsafe.Pointer + if len(payload) > 0 { + _p2 = unsafe.Pointer(&payload[0]) + } else { + _p2 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_ADD_KEY, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(_p2), uintptr(len(payload)), uintptr(ringid), 0) + id = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Adjtimex(buf *Timex) (state int, err error) { r0, _, e1 := Syscall(SYS_ADJTIMEX, uintptr(unsafe.Pointer(buf)), 0, 0) state = int(r0) @@ -344,17 +469,6 @@ func Dup3(oldfd int, newfd int, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func EpollCreate(size int) (fd int, err error) { - r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE, uintptr(size), 0, 0) - fd = int(r0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func EpollCreate1(flag int) (fd int, err error) { r0, _, e1 := RawSyscall(SYS_EPOLL_CREATE1, uintptr(flag), 0, 0) fd = int(r0) @@ -376,8 +490,19 @@ func EpollCtl(epfd int, op int, fd int, event *EpollEvent) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Eventfd(initval uint, flags int) (fd int, err error) { + r0, _, e1 := Syscall(SYS_EVENTFD2, uintptr(initval), uintptr(flags), 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Exit(code int) { - Syscall(SYS_EXIT_GROUP, uintptr(code), 0, 0) + SyscallNoError(SYS_EXIT_GROUP, uintptr(code), 0, 0) return } @@ -428,21 +553,6 @@ func Fchmod(fd int, mode uint32) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Fchmodat(dirfd int, path string, mode uint32, flags int) (err error) { - var _p0 *byte - _p0, err = BytePtrFromString(path) - if err != nil { - return - } - _, _, e1 := Syscall6(SYS_FCHMODAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func Fchownat(dirfd int, path string, uid int, gid int, flags int) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -528,7 +638,7 @@ func Getpgid(pid int) (pgid int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Getpid() (pid int) { - r0, _, _ := RawSyscall(SYS_GETPID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETPID, 0, 0, 0) pid = int(r0) return } @@ -536,7 +646,7 @@ func Getpid() (pid int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Getppid() (ppid int) { - r0, _, _ := RawSyscall(SYS_GETPPID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETPPID, 0, 0, 0) ppid = int(r0) return } @@ -593,7 +703,7 @@ func Getsid(pid int) (sid int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Gettid() (tid int) { - r0, _, _ := RawSyscall(SYS_GETTID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETTID, 0, 0, 0) tid = int(r0) return } @@ -692,6 +802,33 @@ func Klogctl(typ int, buf []byte) (n int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Lgetxattr(path string, attr string, dest []byte) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + var _p2 unsafe.Pointer + if len(dest) > 0 { + _p2 = unsafe.Pointer(&dest[0]) + } else { + _p2 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall6(SYS_LGETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(_p2), uintptr(len(dest)), 0, 0) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Listxattr(path string, dest []byte) (sz int, err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -714,6 +851,74 @@ func Listxattr(path string, dest []byte) (sz int, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Llistxattr(path string, dest []byte) (sz int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 unsafe.Pointer + if len(dest) > 0 { + _p1 = unsafe.Pointer(&dest[0]) + } else { + _p1 = unsafe.Pointer(&_zero) + } + r0, _, e1 := Syscall(SYS_LLISTXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(_p1), uintptr(len(dest))) + sz = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func Lremovexattr(path string, attr string) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_LREMOVEXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func Lsetxattr(path string, attr string, data []byte, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(attr) + if err != nil { + return + } + var _p2 unsafe.Pointer + if len(data) > 0 { + _p2 = unsafe.Pointer(&data[0]) + } else { + _p2 = unsafe.Pointer(&_zero) + } + _, _, e1 := Syscall6(SYS_LSETXATTR, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(_p2), uintptr(len(data)), uintptr(flags), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Mkdirat(dirfd int, path string, mode uint32) (err error) { var _p0 *byte _p0, err = BytePtrFromString(path) @@ -754,6 +959,17 @@ func Nanosleep(time *Timespec, leftover *Timespec) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func PerfEventOpen(attr *PerfEventAttr, pid int, cpu int, groupFd int, flags int) (fd int, err error) { + r0, _, e1 := Syscall6(SYS_PERF_EVENT_OPEN, uintptr(unsafe.Pointer(attr)), uintptr(pid), uintptr(cpu), uintptr(groupFd), uintptr(flags), 0) + fd = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func PivotRoot(newroot string, putold string) (err error) { var _p0 *byte _p0, err = BytePtrFromString(newroot) @@ -794,6 +1010,17 @@ func Prctl(option int, arg2 uintptr, arg3 uintptr, arg4 uintptr, arg5 uintptr) ( // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Pselect(nfd int, r *FdSet, w *FdSet, e *FdSet, timeout *Timespec, sigmask *Sigset_t) (n int, err error) { + r0, _, e1 := Syscall6(SYS_PSELECT6, uintptr(nfd), uintptr(unsafe.Pointer(r)), uintptr(unsafe.Pointer(w)), uintptr(unsafe.Pointer(e)), uintptr(unsafe.Pointer(timeout)), uintptr(unsafe.Pointer(sigmask))) + n = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func read(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -851,6 +1078,32 @@ func Renameat(olddirfd int, oldpath string, newdirfd int, newpath string) (err e // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func RequestKey(keyType string, description string, callback string, destRingid int) (id int, err error) { + var _p0 *byte + _p0, err = BytePtrFromString(keyType) + if err != nil { + return + } + var _p1 *byte + _p1, err = BytePtrFromString(description) + if err != nil { + return + } + var _p2 *byte + _p2, err = BytePtrFromString(callback) + if err != nil { + return + } + r0, _, e1 := Syscall6(SYS_REQUEST_KEY, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(_p1)), uintptr(unsafe.Pointer(_p2)), uintptr(destRingid), 0, 0) + id = int(r0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Setdomainname(p []byte) (err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -960,8 +1213,33 @@ func Setxattr(path string, attr string, data []byte, flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Statx(dirfd int, path string, flags int, mask int, stat *Statx_t) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_STATX, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(flags), uintptr(mask), uintptr(unsafe.Pointer(stat)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Sync() { - Syscall(SYS_SYNC, 0, 0, 0) + SyscallNoError(SYS_SYNC, 0, 0, 0) + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func Syncfs(fd int) (err error) { + _, _, e1 := Syscall(SYS_SYNCFS, uintptr(fd), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } return } @@ -1010,7 +1288,7 @@ func Times(tms *Tms) (ticks uintptr, err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Umask(mask int) (oldmask int) { - r0, _, _ := RawSyscall(SYS_UMASK, uintptr(mask), 0, 0) + r0, _ := RawSyscallNoError(SYS_UMASK, uintptr(mask), 0, 0) oldmask = int(r0) return } @@ -1052,16 +1330,6 @@ func Unshare(flags int) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Ustat(dev int, ubuf *Ustat_t) (err error) { - _, _, e1 := Syscall(SYS_USTAT, uintptr(dev), uintptr(unsafe.Pointer(ubuf)), 0) - if e1 != 0 { - err = errnoErr(e1) - } - return -} - -// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT - func write(fd int, p []byte) (n int, err error) { var _p0 unsafe.Pointer if len(p) > 0 { @@ -1169,14 +1437,8 @@ func Mlock(b []byte) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Munlock(b []byte) (err error) { - var _p0 unsafe.Pointer - if len(b) > 0 { - _p0 = unsafe.Pointer(&b[0]) - } else { - _p0 = unsafe.Pointer(&_zero) - } - _, _, e1 := Syscall(SYS_MUNLOCK, uintptr(_p0), uintptr(len(b)), 0) +func Mlockall(flags int) (err error) { + _, _, e1 := Syscall(SYS_MLOCKALL, uintptr(flags), 0, 0) if e1 != 0 { err = errnoErr(e1) } @@ -1185,8 +1447,30 @@ func Munlock(b []byte) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT -func Mlockall(flags int) (err error) { - _, _, e1 := Syscall(SYS_MLOCKALL, uintptr(flags), 0, 0) +func Msync(b []byte, flags int) (err error) { + var _p0 unsafe.Pointer + if len(b) > 0 { + _p0 = unsafe.Pointer(&b[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + _, _, e1 := Syscall(SYS_MSYNC, uintptr(_p0), uintptr(len(b)), uintptr(flags)) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + +func Munlock(b []byte) (err error) { + var _p0 unsafe.Pointer + if len(b) > 0 { + _p0 = unsafe.Pointer(&b[0]) + } else { + _p0 = unsafe.Pointer(&_zero) + } + _, _, e1 := Syscall(SYS_MUNLOCK, uintptr(_p0), uintptr(len(b)), 0) if e1 != 0 { err = errnoErr(e1) } @@ -1262,6 +1546,21 @@ func Fstat(fd int, stat *Stat_t) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Fstatat(dirfd int, path string, stat *Stat_t, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FSTATAT64, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(stat)), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fstatfs(fd int, buf *Statfs_t) (err error) { _, _, e1 := Syscall(SYS_FSTATFS, uintptr(fd), uintptr(unsafe.Pointer(buf)), 0) if e1 != 0 { @@ -1283,7 +1582,7 @@ func Ftruncate(fd int, length int64) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Getegid() (egid int) { - r0, _, _ := RawSyscall(SYS_GETEGID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETEGID, 0, 0, 0) egid = int(r0) return } @@ -1291,7 +1590,7 @@ func Getegid() (egid int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Geteuid() (euid int) { - r0, _, _ := RawSyscall(SYS_GETEUID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETEUID, 0, 0, 0) euid = int(r0) return } @@ -1299,7 +1598,7 @@ func Geteuid() (euid int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Getgid() (gid int) { - r0, _, _ := RawSyscall(SYS_GETGID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETGID, 0, 0, 0) gid = int(r0) return } @@ -1317,7 +1616,7 @@ func Getrlimit(resource int, rlim *Rlimit) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT func Getuid() (uid int) { - r0, _, _ := RawSyscall(SYS_GETUID, 0, 0, 0) + r0, _ := RawSyscallNoError(SYS_GETUID, 0, 0, 0) uid = int(r0) return } @@ -1788,6 +2087,21 @@ func mmap(addr uintptr, length uintptr, prot int, flags int, fd int, offset int6 // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func futimesat(dirfd int, path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_FUTIMESAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times))) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Gettimeofday(tv *Timeval) (err error) { _, _, e1 := RawSyscall(SYS_GETTIMEOFDAY, uintptr(unsafe.Pointer(tv)), 0, 0) if e1 != 0 { @@ -1813,6 +2127,21 @@ func Utime(path string, buf *Utimbuf) (err error) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func utimes(path string, times *[2]Timeval) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall(SYS_UTIMES, uintptr(unsafe.Pointer(_p0)), uintptr(unsafe.Pointer(times)), 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func pipe(p *[2]_C_int) (err error) { _, _, e1 := RawSyscall(SYS_PIPE, uintptr(unsafe.Pointer(p)), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_netbsd_386.go b/vendor/golang.org/x/sys/unix/zsyscall_netbsd_386.go index fb4b96278..b3298930e 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_netbsd_386.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_netbsd_386.go @@ -571,6 +571,21 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall9(SYS_POSIX_FADVISE, uintptr(fd), 0, uintptr(offset), uintptr(offset>>32), 0, uintptr(length), uintptr(length>>32), uintptr(advice), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_netbsd_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_netbsd_amd64.go index beac82ef8..5096ef07e 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_netbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_netbsd_amd64.go @@ -571,6 +571,21 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall6(SYS_POSIX_FADVISE, uintptr(fd), 0, uintptr(offset), 0, uintptr(length), uintptr(advice)) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_netbsd_arm.go b/vendor/golang.org/x/sys/unix/zsyscall_netbsd_arm.go index 7bd5f60b0..b7141c63f 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_netbsd_arm.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_netbsd_arm.go @@ -571,6 +571,21 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fadvise(fd int, offset int64, length int64, advice int) (err error) { _, _, e1 := Syscall9(SYS_POSIX_FADVISE, uintptr(fd), 0, uintptr(offset), uintptr(offset>>32), 0, uintptr(length), uintptr(length>>32), uintptr(advice), 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_openbsd_386.go b/vendor/golang.org/x/sys/unix/zsyscall_openbsd_386.go index 49b3b5e8a..1942049b0 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_openbsd_386.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_openbsd_386.go @@ -569,6 +569,21 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fchdir(fd int) (err error) { _, _, e1 := Syscall(SYS_FCHDIR, uintptr(fd), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_openbsd_amd64.go index c4c7d8540..d351c72cb 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_openbsd_amd64.go @@ -569,6 +569,21 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fchdir(fd int) (err error) { _, _, e1 := Syscall(SYS_FCHDIR, uintptr(fd), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_openbsd_arm.go b/vendor/golang.org/x/sys/unix/zsyscall_openbsd_arm.go index 210285b0b..617d47f0f 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_openbsd_arm.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_openbsd_arm.go @@ -569,6 +569,21 @@ func Exit(code int) { // THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := Syscall6(SYS_FACCESSAT, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = errnoErr(e1) + } + return +} + +// THIS FILE IS GENERATED BY THE COMMAND AT THE TOP; DO NOT EDIT + func Fchdir(fd int) (err error) { _, _, e1 := Syscall(SYS_FCHDIR, uintptr(fd), 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsyscall_solaris_amd64.go b/vendor/golang.org/x/sys/unix/zsyscall_solaris_amd64.go index 397896300..e2e5fc5e0 100644 --- a/vendor/golang.org/x/sys/unix/zsyscall_solaris_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsyscall_solaris_amd64.go @@ -41,6 +41,7 @@ import ( //go:cgo_import_dynamic libc_dup dup "libc.so" //go:cgo_import_dynamic libc_dup2 dup2 "libc.so" //go:cgo_import_dynamic libc_exit exit "libc.so" +//go:cgo_import_dynamic libc_faccessat faccessat "libc.so" //go:cgo_import_dynamic libc_fchdir fchdir "libc.so" //go:cgo_import_dynamic libc_fchmod fchmod "libc.so" //go:cgo_import_dynamic libc_fchmodat fchmodat "libc.so" @@ -169,6 +170,7 @@ import ( //go:linkname procDup libc_dup //go:linkname procDup2 libc_dup2 //go:linkname procExit libc_exit +//go:linkname procFaccessat libc_faccessat //go:linkname procFchdir libc_fchdir //go:linkname procFchmod libc_fchmod //go:linkname procFchmodat libc_fchmodat @@ -298,6 +300,7 @@ var ( procDup, procDup2, procExit, + procFaccessat, procFchdir, procFchmod, procFchmodat, @@ -695,6 +698,19 @@ func Exit(code int) { return } +func Faccessat(dirfd int, path string, mode uint32, flags int) (err error) { + var _p0 *byte + _p0, err = BytePtrFromString(path) + if err != nil { + return + } + _, _, e1 := sysvicall6(uintptr(unsafe.Pointer(&procFaccessat)), 4, uintptr(dirfd), uintptr(unsafe.Pointer(_p0)), uintptr(mode), uintptr(flags), 0, 0) + if e1 != 0 { + err = e1 + } + return +} + func Fchdir(fd int) (err error) { _, _, e1 := sysvicall6(uintptr(unsafe.Pointer(&procFchdir)), 1, uintptr(fd), 0, 0, 0, 0, 0) if e1 != 0 { diff --git a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go index 83bb935b9..b005031ab 100644 --- a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go +++ b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_386.go @@ -1,5 +1,5 @@ // mksysctl_openbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. package unix diff --git a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go index 83bb935b9..90c95c2c7 100644 --- a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_amd64.go @@ -1,5 +1,7 @@ // mksysctl_openbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. + +// +build amd64,openbsd package unix @@ -14,6 +16,7 @@ var sysctlMib = []mibentry{ {"ddb.max_line", []_C_int{9, 3}}, {"ddb.max_width", []_C_int{9, 2}}, {"ddb.panic", []_C_int{9, 5}}, + {"ddb.profile", []_C_int{9, 9}}, {"ddb.radix", []_C_int{9, 1}}, {"ddb.tab_stop_width", []_C_int{9, 4}}, {"ddb.trigger", []_C_int{9, 8}}, @@ -29,6 +32,7 @@ var sysctlMib = []mibentry{ {"hw.ncpu", []_C_int{6, 3}}, {"hw.ncpufound", []_C_int{6, 21}}, {"hw.pagesize", []_C_int{6, 7}}, + {"hw.perfpolicy", []_C_int{6, 23}}, {"hw.physmem", []_C_int{6, 19}}, {"hw.product", []_C_int{6, 15}}, {"hw.serialno", []_C_int{6, 17}}, @@ -37,7 +41,7 @@ var sysctlMib = []mibentry{ {"hw.uuid", []_C_int{6, 18}}, {"hw.vendor", []_C_int{6, 14}}, {"hw.version", []_C_int{6, 16}}, - {"kern.arandom", []_C_int{1, 37}}, + {"kern.allowkmem", []_C_int{1, 52}}, {"kern.argmax", []_C_int{1, 8}}, {"kern.boottime", []_C_int{1, 21}}, {"kern.bufcachepercent", []_C_int{1, 72}}, @@ -46,12 +50,13 @@ var sysctlMib = []mibentry{ {"kern.consdev", []_C_int{1, 75}}, {"kern.cp_time", []_C_int{1, 40}}, {"kern.cp_time2", []_C_int{1, 71}}, - {"kern.cryptodevallowsoft", []_C_int{1, 53}}, + {"kern.dnsjackport", []_C_int{1, 13}}, {"kern.domainname", []_C_int{1, 22}}, {"kern.file", []_C_int{1, 73}}, {"kern.forkstat", []_C_int{1, 42}}, {"kern.fscale", []_C_int{1, 46}}, {"kern.fsync", []_C_int{1, 33}}, + {"kern.global_ptrace", []_C_int{1, 81}}, {"kern.hostid", []_C_int{1, 11}}, {"kern.hostname", []_C_int{1, 10}}, {"kern.intrcnt.nintrcnt", []_C_int{1, 63, 1}}, @@ -84,7 +89,6 @@ var sysctlMib = []mibentry{ {"kern.pool_debug", []_C_int{1, 77}}, {"kern.posix1version", []_C_int{1, 17}}, {"kern.proc", []_C_int{1, 66}}, - {"kern.random", []_C_int{1, 31}}, {"kern.rawpartition", []_C_int{1, 24}}, {"kern.saved_ids", []_C_int{1, 20}}, {"kern.securelevel", []_C_int{1, 9}}, @@ -102,21 +106,16 @@ var sysctlMib = []mibentry{ {"kern.timecounter.hardware", []_C_int{1, 69, 3}}, {"kern.timecounter.tick", []_C_int{1, 69, 1}}, {"kern.timecounter.timestepwarnings", []_C_int{1, 69, 2}}, - {"kern.tty.maxptys", []_C_int{1, 44, 6}}, - {"kern.tty.nptys", []_C_int{1, 44, 7}}, {"kern.tty.tk_cancc", []_C_int{1, 44, 4}}, {"kern.tty.tk_nin", []_C_int{1, 44, 1}}, {"kern.tty.tk_nout", []_C_int{1, 44, 2}}, {"kern.tty.tk_rawcc", []_C_int{1, 44, 3}}, {"kern.tty.ttyinfo", []_C_int{1, 44, 5}}, {"kern.ttycount", []_C_int{1, 57}}, - {"kern.userasymcrypto", []_C_int{1, 60}}, - {"kern.usercrypto", []_C_int{1, 52}}, - {"kern.usermount", []_C_int{1, 30}}, {"kern.version", []_C_int{1, 4}}, - {"kern.vnode", []_C_int{1, 13}}, {"kern.watchdog.auto", []_C_int{1, 64, 2}}, {"kern.watchdog.period", []_C_int{1, 64, 1}}, + {"kern.wxabort", []_C_int{1, 74}}, {"net.bpf.bufsize", []_C_int{4, 31, 1}}, {"net.bpf.maxbufsize", []_C_int{4, 31, 2}}, {"net.inet.ah.enable", []_C_int{4, 2, 51, 1}}, @@ -144,7 +143,9 @@ var sysctlMib = []mibentry{ {"net.inet.icmp.stats", []_C_int{4, 2, 1, 7}}, {"net.inet.icmp.tstamprepl", []_C_int{4, 2, 1, 6}}, {"net.inet.igmp.stats", []_C_int{4, 2, 2, 1}}, + {"net.inet.ip.arpdown", []_C_int{4, 2, 0, 40}}, {"net.inet.ip.arpqueued", []_C_int{4, 2, 0, 36}}, + {"net.inet.ip.arptimeout", []_C_int{4, 2, 0, 39}}, {"net.inet.ip.encdebug", []_C_int{4, 2, 0, 12}}, {"net.inet.ip.forwarding", []_C_int{4, 2, 0, 1}}, {"net.inet.ip.ifq.congestion", []_C_int{4, 2, 0, 30, 4}}, @@ -153,8 +154,10 @@ var sysctlMib = []mibentry{ {"net.inet.ip.ifq.maxlen", []_C_int{4, 2, 0, 30, 2}}, {"net.inet.ip.maxqueue", []_C_int{4, 2, 0, 11}}, {"net.inet.ip.mforwarding", []_C_int{4, 2, 0, 31}}, + {"net.inet.ip.mrtmfc", []_C_int{4, 2, 0, 37}}, {"net.inet.ip.mrtproto", []_C_int{4, 2, 0, 34}}, {"net.inet.ip.mrtstats", []_C_int{4, 2, 0, 35}}, + {"net.inet.ip.mrtvif", []_C_int{4, 2, 0, 38}}, {"net.inet.ip.mtu", []_C_int{4, 2, 0, 4}}, {"net.inet.ip.mtudisc", []_C_int{4, 2, 0, 27}}, {"net.inet.ip.mtudisctimeout", []_C_int{4, 2, 0, 28}}, @@ -173,7 +176,6 @@ var sysctlMib = []mibentry{ {"net.inet.ipip.stats", []_C_int{4, 2, 4, 2}}, {"net.inet.mobileip.allow", []_C_int{4, 2, 55, 1}}, {"net.inet.pfsync.stats", []_C_int{4, 2, 240, 1}}, - {"net.inet.pim.stats", []_C_int{4, 2, 103, 1}}, {"net.inet.tcp.ackonpush", []_C_int{4, 2, 6, 13}}, {"net.inet.tcp.always_keepalive", []_C_int{4, 2, 6, 22}}, {"net.inet.tcp.baddynamic", []_C_int{4, 2, 6, 6}}, @@ -187,6 +189,7 @@ var sysctlMib = []mibentry{ {"net.inet.tcp.reasslimit", []_C_int{4, 2, 6, 18}}, {"net.inet.tcp.rfc1323", []_C_int{4, 2, 6, 1}}, {"net.inet.tcp.rfc3390", []_C_int{4, 2, 6, 17}}, + {"net.inet.tcp.rootonly", []_C_int{4, 2, 6, 24}}, {"net.inet.tcp.rstppslimit", []_C_int{4, 2, 6, 12}}, {"net.inet.tcp.sack", []_C_int{4, 2, 6, 10}}, {"net.inet.tcp.sackholelimit", []_C_int{4, 2, 6, 20}}, @@ -194,9 +197,12 @@ var sysctlMib = []mibentry{ {"net.inet.tcp.stats", []_C_int{4, 2, 6, 21}}, {"net.inet.tcp.synbucketlimit", []_C_int{4, 2, 6, 16}}, {"net.inet.tcp.syncachelimit", []_C_int{4, 2, 6, 15}}, + {"net.inet.tcp.synhashsize", []_C_int{4, 2, 6, 25}}, + {"net.inet.tcp.synuselimit", []_C_int{4, 2, 6, 23}}, {"net.inet.udp.baddynamic", []_C_int{4, 2, 17, 2}}, {"net.inet.udp.checksum", []_C_int{4, 2, 17, 1}}, {"net.inet.udp.recvspace", []_C_int{4, 2, 17, 3}}, + {"net.inet.udp.rootonly", []_C_int{4, 2, 17, 6}}, {"net.inet.udp.sendspace", []_C_int{4, 2, 17, 4}}, {"net.inet.udp.stats", []_C_int{4, 2, 17, 5}}, {"net.inet6.divert.recvspace", []_C_int{4, 24, 86, 1}}, @@ -209,13 +215,8 @@ var sysctlMib = []mibentry{ {"net.inet6.icmp6.nd6_delay", []_C_int{4, 24, 30, 8}}, {"net.inet6.icmp6.nd6_maxnudhint", []_C_int{4, 24, 30, 15}}, {"net.inet6.icmp6.nd6_mmaxtries", []_C_int{4, 24, 30, 10}}, - {"net.inet6.icmp6.nd6_prune", []_C_int{4, 24, 30, 6}}, {"net.inet6.icmp6.nd6_umaxtries", []_C_int{4, 24, 30, 9}}, - {"net.inet6.icmp6.nd6_useloopback", []_C_int{4, 24, 30, 11}}, - {"net.inet6.icmp6.nodeinfo", []_C_int{4, 24, 30, 13}}, - {"net.inet6.icmp6.rediraccept", []_C_int{4, 24, 30, 2}}, {"net.inet6.icmp6.redirtimeout", []_C_int{4, 24, 30, 3}}, - {"net.inet6.ip6.accept_rtadv", []_C_int{4, 24, 17, 12}}, {"net.inet6.ip6.auto_flowlabel", []_C_int{4, 24, 17, 17}}, {"net.inet6.ip6.dad_count", []_C_int{4, 24, 17, 16}}, {"net.inet6.ip6.dad_pending", []_C_int{4, 24, 17, 49}}, @@ -228,20 +229,19 @@ var sysctlMib = []mibentry{ {"net.inet6.ip6.maxdynroutes", []_C_int{4, 24, 17, 48}}, {"net.inet6.ip6.maxfragpackets", []_C_int{4, 24, 17, 9}}, {"net.inet6.ip6.maxfrags", []_C_int{4, 24, 17, 41}}, - {"net.inet6.ip6.maxifdefrouters", []_C_int{4, 24, 17, 47}}, - {"net.inet6.ip6.maxifprefixes", []_C_int{4, 24, 17, 46}}, {"net.inet6.ip6.mforwarding", []_C_int{4, 24, 17, 42}}, + {"net.inet6.ip6.mrtmfc", []_C_int{4, 24, 17, 53}}, + {"net.inet6.ip6.mrtmif", []_C_int{4, 24, 17, 52}}, {"net.inet6.ip6.mrtproto", []_C_int{4, 24, 17, 8}}, {"net.inet6.ip6.mtudisctimeout", []_C_int{4, 24, 17, 50}}, {"net.inet6.ip6.multicast_mtudisc", []_C_int{4, 24, 17, 44}}, {"net.inet6.ip6.multipath", []_C_int{4, 24, 17, 43}}, {"net.inet6.ip6.neighborgcthresh", []_C_int{4, 24, 17, 45}}, {"net.inet6.ip6.redirect", []_C_int{4, 24, 17, 2}}, - {"net.inet6.ip6.rr_prune", []_C_int{4, 24, 17, 22}}, + {"net.inet6.ip6.soiikey", []_C_int{4, 24, 17, 54}}, {"net.inet6.ip6.sourcecheck", []_C_int{4, 24, 17, 10}}, {"net.inet6.ip6.sourcecheck_logint", []_C_int{4, 24, 17, 11}}, {"net.inet6.ip6.use_deprecated", []_C_int{4, 24, 17, 21}}, - {"net.inet6.ip6.v6only", []_C_int{4, 24, 17, 24}}, {"net.key.sadb_dump", []_C_int{4, 30, 1}}, {"net.key.spd_dump", []_C_int{4, 30, 2}}, {"net.mpls.ifq.congestion", []_C_int{4, 33, 3, 4}}, @@ -254,17 +254,4 @@ var sysctlMib = []mibentry{ {"net.mpls.ttl", []_C_int{4, 33, 2}}, {"net.pflow.stats", []_C_int{4, 34, 1}}, {"net.pipex.enable", []_C_int{4, 35, 1}}, - {"vm.anonmin", []_C_int{2, 7}}, - {"vm.loadavg", []_C_int{2, 2}}, - {"vm.maxslp", []_C_int{2, 10}}, - {"vm.nkmempages", []_C_int{2, 6}}, - {"vm.psstrings", []_C_int{2, 3}}, - {"vm.swapencrypt.enable", []_C_int{2, 5, 0}}, - {"vm.swapencrypt.keyscreated", []_C_int{2, 5, 1}}, - {"vm.swapencrypt.keysdeleted", []_C_int{2, 5, 2}}, - {"vm.uspace", []_C_int{2, 11}}, - {"vm.uvmexp", []_C_int{2, 4}}, - {"vm.vmmeter", []_C_int{2, 1}}, - {"vm.vnodemin", []_C_int{2, 9}}, - {"vm.vtextmin", []_C_int{2, 8}}, } diff --git a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go index 83bb935b9..b005031ab 100644 --- a/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go +++ b/vendor/golang.org/x/sys/unix/zsysctl_openbsd_arm.go @@ -1,5 +1,5 @@ // mksysctl_openbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. package unix diff --git a/vendor/golang.org/x/sys/unix/zsysnum_linux_s390x.go b/vendor/golang.org/x/sys/unix/zsysnum_linux_s390x.go index ed92409d9..c061d6f1d 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_linux_s390x.go @@ -331,4 +331,5 @@ const ( SYS_S390_GUARDED_STORAGE = 378 SYS_STATX = 379 SYS_S390_STHYI = 380 + SYS_KEXEC_FILE_LOAD = 381 ) diff --git a/vendor/golang.org/x/sys/unix/zsysnum_linux_sparc64.go b/vendor/golang.org/x/sys/unix/zsysnum_linux_sparc64.go index c9c129dc4..2d0993672 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_linux_sparc64.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_linux_sparc64.go @@ -1,5 +1,5 @@ // mksysnum_linux.pl -Ilinux/usr/include -m64 -D__arch64__ linux/usr/include/asm/unistd.h -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build sparc64,linux diff --git a/vendor/golang.org/x/sys/unix/zsysnum_netbsd_386.go b/vendor/golang.org/x/sys/unix/zsysnum_netbsd_386.go index 8afda9c45..f0daa05a9 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_netbsd_386.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_netbsd_386.go @@ -1,5 +1,5 @@ // mksysnum_netbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build 386,netbsd diff --git a/vendor/golang.org/x/sys/unix/zsysnum_netbsd_amd64.go b/vendor/golang.org/x/sys/unix/zsysnum_netbsd_amd64.go index aea8dbec4..ddb25b94f 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_netbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_netbsd_amd64.go @@ -1,5 +1,5 @@ // mksysnum_netbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build amd64,netbsd diff --git a/vendor/golang.org/x/sys/unix/zsysnum_netbsd_arm.go b/vendor/golang.org/x/sys/unix/zsysnum_netbsd_arm.go index c6158a7ef..315bd63f8 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_netbsd_arm.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_netbsd_arm.go @@ -1,5 +1,5 @@ // mksysnum_netbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build arm,netbsd diff --git a/vendor/golang.org/x/sys/unix/zsysnum_openbsd_386.go b/vendor/golang.org/x/sys/unix/zsysnum_openbsd_386.go index 3e8ce2a1d..07787301f 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_openbsd_386.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_openbsd_386.go @@ -1,5 +1,5 @@ // mksysnum_openbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build 386,openbsd diff --git a/vendor/golang.org/x/sys/unix/zsysnum_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/zsysnum_openbsd_amd64.go index bd28146dd..10edff07d 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_openbsd_amd64.go @@ -1,5 +1,5 @@ // mksysnum_openbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; see README.md. DO NOT EDIT. // +build amd64,openbsd @@ -12,6 +12,7 @@ const ( SYS_WRITE = 4 // { ssize_t sys_write(int fd, const void *buf, \ SYS_OPEN = 5 // { int sys_open(const char *path, \ SYS_CLOSE = 6 // { int sys_close(int fd); } + SYS_GETENTROPY = 7 // { int sys_getentropy(void *buf, size_t nbyte); } SYS___TFORK = 8 // { int sys___tfork(const struct __tfork *param, \ SYS_LINK = 9 // { int sys_link(const char *path, const char *link); } SYS_UNLINK = 10 // { int sys_unlink(const char *path); } @@ -37,11 +38,10 @@ const ( SYS_ACCEPT = 30 // { int sys_accept(int s, struct sockaddr *name, \ SYS_GETPEERNAME = 31 // { int sys_getpeername(int fdes, struct sockaddr *asa, \ SYS_GETSOCKNAME = 32 // { int sys_getsockname(int fdes, struct sockaddr *asa, \ - SYS_ACCESS = 33 // { int sys_access(const char *path, int flags); } + SYS_ACCESS = 33 // { int sys_access(const char *path, int amode); } SYS_CHFLAGS = 34 // { int sys_chflags(const char *path, u_int flags); } SYS_FCHFLAGS = 35 // { int sys_fchflags(int fd, u_int flags); } SYS_SYNC = 36 // { void sys_sync(void); } - SYS_KILL = 37 // { int sys_kill(int pid, int signum); } SYS_STAT = 38 // { int sys_stat(const char *path, struct stat *ub); } SYS_GETPPID = 39 // { pid_t sys_getppid(void); } SYS_LSTAT = 40 // { int sys_lstat(const char *path, struct stat *ub); } @@ -53,7 +53,6 @@ const ( SYS_SIGACTION = 46 // { int sys_sigaction(int signum, \ SYS_GETGID = 47 // { gid_t sys_getgid(void); } SYS_SIGPROCMASK = 48 // { int sys_sigprocmask(int how, sigset_t mask); } - SYS_GETLOGIN = 49 // { int sys_getlogin(char *namebuf, u_int namelen); } SYS_SETLOGIN = 50 // { int sys_setlogin(const char *namebuf); } SYS_ACCT = 51 // { int sys_acct(const char *path); } SYS_SIGPENDING = 52 // { int sys_sigpending(void); } @@ -62,7 +61,7 @@ const ( SYS_REBOOT = 55 // { int sys_reboot(int opt); } SYS_REVOKE = 56 // { int sys_revoke(const char *path); } SYS_SYMLINK = 57 // { int sys_symlink(const char *path, \ - SYS_READLINK = 58 // { int sys_readlink(const char *path, char *buf, \ + SYS_READLINK = 58 // { ssize_t sys_readlink(const char *path, \ SYS_EXECVE = 59 // { int sys_execve(const char *path, \ SYS_UMASK = 60 // { mode_t sys_umask(mode_t newmask); } SYS_CHROOT = 61 // { int sys_chroot(const char *path); } @@ -86,15 +85,18 @@ const ( SYS_GETGROUPS = 79 // { int sys_getgroups(int gidsetsize, \ SYS_SETGROUPS = 80 // { int sys_setgroups(int gidsetsize, \ SYS_GETPGRP = 81 // { int sys_getpgrp(void); } - SYS_SETPGID = 82 // { int sys_setpgid(pid_t pid, int pgid); } + SYS_SETPGID = 82 // { int sys_setpgid(pid_t pid, pid_t pgid); } + SYS_FUTEX = 83 // { int sys_futex(uint32_t *f, int op, int val, \ SYS_UTIMENSAT = 84 // { int sys_utimensat(int fd, const char *path, \ SYS_FUTIMENS = 85 // { int sys_futimens(int fd, \ + SYS_KBIND = 86 // { int sys_kbind(const struct __kbind *param, \ SYS_CLOCK_GETTIME = 87 // { int sys_clock_gettime(clockid_t clock_id, \ SYS_CLOCK_SETTIME = 88 // { int sys_clock_settime(clockid_t clock_id, \ SYS_CLOCK_GETRES = 89 // { int sys_clock_getres(clockid_t clock_id, \ SYS_DUP2 = 90 // { int sys_dup2(int from, int to); } SYS_NANOSLEEP = 91 // { int sys_nanosleep(const struct timespec *rqtp, \ SYS_FCNTL = 92 // { int sys_fcntl(int fd, int cmd, ... void *arg); } + SYS_ACCEPT4 = 93 // { int sys_accept4(int s, struct sockaddr *name, \ SYS___THRSLEEP = 94 // { int sys___thrsleep(const volatile void *ident, \ SYS_FSYNC = 95 // { int sys_fsync(int fd); } SYS_SETPRIORITY = 96 // { int sys_setpriority(int which, id_t who, int prio); } @@ -102,16 +104,23 @@ const ( SYS_CONNECT = 98 // { int sys_connect(int s, const struct sockaddr *name, \ SYS_GETDENTS = 99 // { int sys_getdents(int fd, void *buf, size_t buflen); } SYS_GETPRIORITY = 100 // { int sys_getpriority(int which, id_t who); } + SYS_PIPE2 = 101 // { int sys_pipe2(int *fdp, int flags); } + SYS_DUP3 = 102 // { int sys_dup3(int from, int to, int flags); } SYS_SIGRETURN = 103 // { int sys_sigreturn(struct sigcontext *sigcntxp); } SYS_BIND = 104 // { int sys_bind(int s, const struct sockaddr *name, \ SYS_SETSOCKOPT = 105 // { int sys_setsockopt(int s, int level, int name, \ SYS_LISTEN = 106 // { int sys_listen(int s, int backlog); } + SYS_CHFLAGSAT = 107 // { int sys_chflagsat(int fd, const char *path, \ + SYS_PLEDGE = 108 // { int sys_pledge(const char *promises, \ SYS_PPOLL = 109 // { int sys_ppoll(struct pollfd *fds, \ SYS_PSELECT = 110 // { int sys_pselect(int nd, fd_set *in, fd_set *ou, \ SYS_SIGSUSPEND = 111 // { int sys_sigsuspend(int mask); } + SYS_SENDSYSLOG = 112 // { int sys_sendsyslog(const char *buf, size_t nbyte, \ SYS_GETSOCKOPT = 118 // { int sys_getsockopt(int s, int level, int name, \ + SYS_THRKILL = 119 // { int sys_thrkill(pid_t tid, int signum, void *tcb); } SYS_READV = 120 // { ssize_t sys_readv(int fd, \ SYS_WRITEV = 121 // { ssize_t sys_writev(int fd, \ + SYS_KILL = 122 // { int sys_kill(int pid, int signum); } SYS_FCHOWN = 123 // { int sys_fchown(int fd, uid_t uid, gid_t gid); } SYS_FCHMOD = 124 // { int sys_fchmod(int fd, mode_t mode); } SYS_SETREUID = 126 // { int sys_setreuid(uid_t ruid, uid_t euid); } @@ -125,6 +134,7 @@ const ( SYS_MKDIR = 136 // { int sys_mkdir(const char *path, mode_t mode); } SYS_RMDIR = 137 // { int sys_rmdir(const char *path); } SYS_ADJTIME = 140 // { int sys_adjtime(const struct timeval *delta, \ + SYS_GETLOGIN_R = 141 // { int sys_getlogin_r(char *namebuf, u_int namelen); } SYS_SETSID = 147 // { int sys_setsid(void); } SYS_QUOTACTL = 148 // { int sys_quotactl(const char *path, int cmd, \ SYS_NFSSVC = 155 // { int sys_nfssvc(int flag, void *argp); } @@ -144,7 +154,7 @@ const ( SYS_LSEEK = 199 // { off_t sys_lseek(int fd, int pad, off_t offset, \ SYS_TRUNCATE = 200 // { int sys_truncate(const char *path, int pad, \ SYS_FTRUNCATE = 201 // { int sys_ftruncate(int fd, int pad, off_t length); } - SYS___SYSCTL = 202 // { int sys___sysctl(const int *name, u_int namelen, \ + SYS_SYSCTL = 202 // { int sys_sysctl(const int *name, u_int namelen, \ SYS_MLOCK = 203 // { int sys_mlock(const void *addr, size_t len); } SYS_MUNLOCK = 204 // { int sys_munlock(const void *addr, size_t len); } SYS_GETPGID = 207 // { pid_t sys_getpgid(pid_t pid); } diff --git a/vendor/golang.org/x/sys/unix/zsysnum_openbsd_arm.go b/vendor/golang.org/x/sys/unix/zsysnum_openbsd_arm.go index 32653e53c..7a1693acb 100644 --- a/vendor/golang.org/x/sys/unix/zsysnum_openbsd_arm.go +++ b/vendor/golang.org/x/sys/unix/zsysnum_openbsd_arm.go @@ -1,5 +1,5 @@ // mksysnum_openbsd.pl -// MACHINE GENERATED BY THE ABOVE COMMAND; DO NOT EDIT +// Code generated by the command above; DO NOT EDIT. // +build arm,openbsd diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_386.go b/vendor/golang.org/x/sys/unix/ztypes_linux_386.go index 84a90fa79..e89bc6b36 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_386.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_386.go @@ -705,6 +705,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1355,3 +1357,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int32 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_amd64.go b/vendor/golang.org/x/sys/unix/ztypes_linux_amd64.go index 1a81adc59..d95372bae 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_amd64.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_amd64.go @@ -723,6 +723,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1375,3 +1377,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_arm.go b/vendor/golang.org/x/sys/unix/ztypes_linux_arm.go index 9746e3160..77875ba01 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_arm.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_arm.go @@ -693,6 +693,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1344,3 +1346,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int32 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_arm64.go b/vendor/golang.org/x/sys/unix/ztypes_linux_arm64.go index 6e588f7ae..5a9df694a 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_arm64.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_arm64.go @@ -702,6 +702,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1354,3 +1356,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_mips.go b/vendor/golang.org/x/sys/unix/ztypes_linux_mips.go index 27f776856..dcb239de8 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_mips.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_mips.go @@ -697,6 +697,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1349,3 +1351,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int32 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_mips64.go b/vendor/golang.org/x/sys/unix/ztypes_linux_mips64.go index 069bcdd8a..9cf85f721 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_mips64.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_mips64.go @@ -704,6 +704,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1356,3 +1358,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_mips64le.go b/vendor/golang.org/x/sys/unix/ztypes_linux_mips64le.go index f8bbca8f5..6fd66e751 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_mips64le.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_mips64le.go @@ -704,6 +704,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1356,3 +1358,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_mipsle.go b/vendor/golang.org/x/sys/unix/ztypes_linux_mipsle.go index d35bb0e67..faa5b3ef1 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_mipsle.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_mipsle.go @@ -697,6 +697,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1349,3 +1351,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int32 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64.go b/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64.go index b709a519e..ad4c45246 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64.go @@ -712,6 +712,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1364,3 +1366,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64le.go b/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64le.go index 698eb2603..1fdb2f216 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64le.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_ppc64le.go @@ -712,6 +712,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1364,3 +1366,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_linux_s390x.go b/vendor/golang.org/x/sys/unix/ztypes_linux_s390x.go index c2088bf91..d32079d1a 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_linux_s390x.go +++ b/vendor/golang.org/x/sys/unix/ztypes_linux_s390x.go @@ -728,6 +728,8 @@ const ( AT_SYMLINK_FOLLOW = 0x400 AT_SYMLINK_NOFOLLOW = 0x100 + + AT_EACCESS = 0x200 ) type PollFd struct { @@ -1381,3 +1383,490 @@ const ( SizeofTpacket2Hdr = 0x20 SizeofTpacket3Hdr = 0x30 ) + +const ( + NF_INET_PRE_ROUTING = 0x0 + NF_INET_LOCAL_IN = 0x1 + NF_INET_FORWARD = 0x2 + NF_INET_LOCAL_OUT = 0x3 + NF_INET_POST_ROUTING = 0x4 + NF_INET_NUMHOOKS = 0x5 +) + +const ( + NF_NETDEV_INGRESS = 0x0 + NF_NETDEV_NUMHOOKS = 0x1 +) + +const ( + NFPROTO_UNSPEC = 0x0 + NFPROTO_INET = 0x1 + NFPROTO_IPV4 = 0x2 + NFPROTO_ARP = 0x3 + NFPROTO_NETDEV = 0x5 + NFPROTO_BRIDGE = 0x7 + NFPROTO_IPV6 = 0xa + NFPROTO_DECNET = 0xc + NFPROTO_NUMPROTO = 0xd +) + +type Nfgenmsg struct { + Nfgen_family uint8 + Version uint8 + Res_id uint16 +} + +const ( + NFNL_BATCH_UNSPEC = 0x0 + NFNL_BATCH_GENID = 0x1 +) + +const ( + NFT_REG_VERDICT = 0x0 + NFT_REG_1 = 0x1 + NFT_REG_2 = 0x2 + NFT_REG_3 = 0x3 + NFT_REG_4 = 0x4 + NFT_REG32_00 = 0x8 + NFT_REG32_01 = 0x9 + NFT_REG32_02 = 0xa + NFT_REG32_03 = 0xb + NFT_REG32_04 = 0xc + NFT_REG32_05 = 0xd + NFT_REG32_06 = 0xe + NFT_REG32_07 = 0xf + NFT_REG32_08 = 0x10 + NFT_REG32_09 = 0x11 + NFT_REG32_10 = 0x12 + NFT_REG32_11 = 0x13 + NFT_REG32_12 = 0x14 + NFT_REG32_13 = 0x15 + NFT_REG32_14 = 0x16 + NFT_REG32_15 = 0x17 + NFT_CONTINUE = -0x1 + NFT_BREAK = -0x2 + NFT_JUMP = -0x3 + NFT_GOTO = -0x4 + NFT_RETURN = -0x5 + NFT_MSG_NEWTABLE = 0x0 + NFT_MSG_GETTABLE = 0x1 + NFT_MSG_DELTABLE = 0x2 + NFT_MSG_NEWCHAIN = 0x3 + NFT_MSG_GETCHAIN = 0x4 + NFT_MSG_DELCHAIN = 0x5 + NFT_MSG_NEWRULE = 0x6 + NFT_MSG_GETRULE = 0x7 + NFT_MSG_DELRULE = 0x8 + NFT_MSG_NEWSET = 0x9 + NFT_MSG_GETSET = 0xa + NFT_MSG_DELSET = 0xb + NFT_MSG_NEWSETELEM = 0xc + NFT_MSG_GETSETELEM = 0xd + NFT_MSG_DELSETELEM = 0xe + NFT_MSG_NEWGEN = 0xf + NFT_MSG_GETGEN = 0x10 + NFT_MSG_TRACE = 0x11 + NFT_MSG_NEWOBJ = 0x12 + NFT_MSG_GETOBJ = 0x13 + NFT_MSG_DELOBJ = 0x14 + NFT_MSG_GETOBJ_RESET = 0x15 + NFT_MSG_MAX = 0x19 + NFTA_LIST_UNPEC = 0x0 + NFTA_LIST_ELEM = 0x1 + NFTA_HOOK_UNSPEC = 0x0 + NFTA_HOOK_HOOKNUM = 0x1 + NFTA_HOOK_PRIORITY = 0x2 + NFTA_HOOK_DEV = 0x3 + NFT_TABLE_F_DORMANT = 0x1 + NFTA_TABLE_UNSPEC = 0x0 + NFTA_TABLE_NAME = 0x1 + NFTA_TABLE_FLAGS = 0x2 + NFTA_TABLE_USE = 0x3 + NFTA_CHAIN_UNSPEC = 0x0 + NFTA_CHAIN_TABLE = 0x1 + NFTA_CHAIN_HANDLE = 0x2 + NFTA_CHAIN_NAME = 0x3 + NFTA_CHAIN_HOOK = 0x4 + NFTA_CHAIN_POLICY = 0x5 + NFTA_CHAIN_USE = 0x6 + NFTA_CHAIN_TYPE = 0x7 + NFTA_CHAIN_COUNTERS = 0x8 + NFTA_CHAIN_PAD = 0x9 + NFTA_RULE_UNSPEC = 0x0 + NFTA_RULE_TABLE = 0x1 + NFTA_RULE_CHAIN = 0x2 + NFTA_RULE_HANDLE = 0x3 + NFTA_RULE_EXPRESSIONS = 0x4 + NFTA_RULE_COMPAT = 0x5 + NFTA_RULE_POSITION = 0x6 + NFTA_RULE_USERDATA = 0x7 + NFTA_RULE_PAD = 0x8 + NFTA_RULE_ID = 0x9 + NFT_RULE_COMPAT_F_INV = 0x2 + NFT_RULE_COMPAT_F_MASK = 0x2 + NFTA_RULE_COMPAT_UNSPEC = 0x0 + NFTA_RULE_COMPAT_PROTO = 0x1 + NFTA_RULE_COMPAT_FLAGS = 0x2 + NFT_SET_ANONYMOUS = 0x1 + NFT_SET_CONSTANT = 0x2 + NFT_SET_INTERVAL = 0x4 + NFT_SET_MAP = 0x8 + NFT_SET_TIMEOUT = 0x10 + NFT_SET_EVAL = 0x20 + NFT_SET_OBJECT = 0x40 + NFT_SET_POL_PERFORMANCE = 0x0 + NFT_SET_POL_MEMORY = 0x1 + NFTA_SET_DESC_UNSPEC = 0x0 + NFTA_SET_DESC_SIZE = 0x1 + NFTA_SET_UNSPEC = 0x0 + NFTA_SET_TABLE = 0x1 + NFTA_SET_NAME = 0x2 + NFTA_SET_FLAGS = 0x3 + NFTA_SET_KEY_TYPE = 0x4 + NFTA_SET_KEY_LEN = 0x5 + NFTA_SET_DATA_TYPE = 0x6 + NFTA_SET_DATA_LEN = 0x7 + NFTA_SET_POLICY = 0x8 + NFTA_SET_DESC = 0x9 + NFTA_SET_ID = 0xa + NFTA_SET_TIMEOUT = 0xb + NFTA_SET_GC_INTERVAL = 0xc + NFTA_SET_USERDATA = 0xd + NFTA_SET_PAD = 0xe + NFTA_SET_OBJ_TYPE = 0xf + NFT_SET_ELEM_INTERVAL_END = 0x1 + NFTA_SET_ELEM_UNSPEC = 0x0 + NFTA_SET_ELEM_KEY = 0x1 + NFTA_SET_ELEM_DATA = 0x2 + NFTA_SET_ELEM_FLAGS = 0x3 + NFTA_SET_ELEM_TIMEOUT = 0x4 + NFTA_SET_ELEM_EXPIRATION = 0x5 + NFTA_SET_ELEM_USERDATA = 0x6 + NFTA_SET_ELEM_EXPR = 0x7 + NFTA_SET_ELEM_PAD = 0x8 + NFTA_SET_ELEM_OBJREF = 0x9 + NFTA_SET_ELEM_LIST_UNSPEC = 0x0 + NFTA_SET_ELEM_LIST_TABLE = 0x1 + NFTA_SET_ELEM_LIST_SET = 0x2 + NFTA_SET_ELEM_LIST_ELEMENTS = 0x3 + NFTA_SET_ELEM_LIST_SET_ID = 0x4 + NFT_DATA_VALUE = 0x0 + NFT_DATA_VERDICT = 0xffffff00 + NFTA_DATA_UNSPEC = 0x0 + NFTA_DATA_VALUE = 0x1 + NFTA_DATA_VERDICT = 0x2 + NFTA_VERDICT_UNSPEC = 0x0 + NFTA_VERDICT_CODE = 0x1 + NFTA_VERDICT_CHAIN = 0x2 + NFTA_EXPR_UNSPEC = 0x0 + NFTA_EXPR_NAME = 0x1 + NFTA_EXPR_DATA = 0x2 + NFTA_IMMEDIATE_UNSPEC = 0x0 + NFTA_IMMEDIATE_DREG = 0x1 + NFTA_IMMEDIATE_DATA = 0x2 + NFTA_BITWISE_UNSPEC = 0x0 + NFTA_BITWISE_SREG = 0x1 + NFTA_BITWISE_DREG = 0x2 + NFTA_BITWISE_LEN = 0x3 + NFTA_BITWISE_MASK = 0x4 + NFTA_BITWISE_XOR = 0x5 + NFT_BYTEORDER_NTOH = 0x0 + NFT_BYTEORDER_HTON = 0x1 + NFTA_BYTEORDER_UNSPEC = 0x0 + NFTA_BYTEORDER_SREG = 0x1 + NFTA_BYTEORDER_DREG = 0x2 + NFTA_BYTEORDER_OP = 0x3 + NFTA_BYTEORDER_LEN = 0x4 + NFTA_BYTEORDER_SIZE = 0x5 + NFT_CMP_EQ = 0x0 + NFT_CMP_NEQ = 0x1 + NFT_CMP_LT = 0x2 + NFT_CMP_LTE = 0x3 + NFT_CMP_GT = 0x4 + NFT_CMP_GTE = 0x5 + NFTA_CMP_UNSPEC = 0x0 + NFTA_CMP_SREG = 0x1 + NFTA_CMP_OP = 0x2 + NFTA_CMP_DATA = 0x3 + NFT_RANGE_EQ = 0x0 + NFT_RANGE_NEQ = 0x1 + NFTA_RANGE_UNSPEC = 0x0 + NFTA_RANGE_SREG = 0x1 + NFTA_RANGE_OP = 0x2 + NFTA_RANGE_FROM_DATA = 0x3 + NFTA_RANGE_TO_DATA = 0x4 + NFT_LOOKUP_F_INV = 0x1 + NFTA_LOOKUP_UNSPEC = 0x0 + NFTA_LOOKUP_SET = 0x1 + NFTA_LOOKUP_SREG = 0x2 + NFTA_LOOKUP_DREG = 0x3 + NFTA_LOOKUP_SET_ID = 0x4 + NFTA_LOOKUP_FLAGS = 0x5 + NFT_DYNSET_OP_ADD = 0x0 + NFT_DYNSET_OP_UPDATE = 0x1 + NFT_DYNSET_F_INV = 0x1 + NFTA_DYNSET_UNSPEC = 0x0 + NFTA_DYNSET_SET_NAME = 0x1 + NFTA_DYNSET_SET_ID = 0x2 + NFTA_DYNSET_OP = 0x3 + NFTA_DYNSET_SREG_KEY = 0x4 + NFTA_DYNSET_SREG_DATA = 0x5 + NFTA_DYNSET_TIMEOUT = 0x6 + NFTA_DYNSET_EXPR = 0x7 + NFTA_DYNSET_PAD = 0x8 + NFTA_DYNSET_FLAGS = 0x9 + NFT_PAYLOAD_LL_HEADER = 0x0 + NFT_PAYLOAD_NETWORK_HEADER = 0x1 + NFT_PAYLOAD_TRANSPORT_HEADER = 0x2 + NFT_PAYLOAD_CSUM_NONE = 0x0 + NFT_PAYLOAD_CSUM_INET = 0x1 + NFT_PAYLOAD_L4CSUM_PSEUDOHDR = 0x1 + NFTA_PAYLOAD_UNSPEC = 0x0 + NFTA_PAYLOAD_DREG = 0x1 + NFTA_PAYLOAD_BASE = 0x2 + NFTA_PAYLOAD_OFFSET = 0x3 + NFTA_PAYLOAD_LEN = 0x4 + NFTA_PAYLOAD_SREG = 0x5 + NFTA_PAYLOAD_CSUM_TYPE = 0x6 + NFTA_PAYLOAD_CSUM_OFFSET = 0x7 + NFTA_PAYLOAD_CSUM_FLAGS = 0x8 + NFT_EXTHDR_F_PRESENT = 0x1 + NFT_EXTHDR_OP_IPV6 = 0x0 + NFT_EXTHDR_OP_TCPOPT = 0x1 + NFTA_EXTHDR_UNSPEC = 0x0 + NFTA_EXTHDR_DREG = 0x1 + NFTA_EXTHDR_TYPE = 0x2 + NFTA_EXTHDR_OFFSET = 0x3 + NFTA_EXTHDR_LEN = 0x4 + NFTA_EXTHDR_FLAGS = 0x5 + NFTA_EXTHDR_OP = 0x6 + NFTA_EXTHDR_SREG = 0x7 + NFT_META_LEN = 0x0 + NFT_META_PROTOCOL = 0x1 + NFT_META_PRIORITY = 0x2 + NFT_META_MARK = 0x3 + NFT_META_IIF = 0x4 + NFT_META_OIF = 0x5 + NFT_META_IIFNAME = 0x6 + NFT_META_OIFNAME = 0x7 + NFT_META_IIFTYPE = 0x8 + NFT_META_OIFTYPE = 0x9 + NFT_META_SKUID = 0xa + NFT_META_SKGID = 0xb + NFT_META_NFTRACE = 0xc + NFT_META_RTCLASSID = 0xd + NFT_META_SECMARK = 0xe + NFT_META_NFPROTO = 0xf + NFT_META_L4PROTO = 0x10 + NFT_META_BRI_IIFNAME = 0x11 + NFT_META_BRI_OIFNAME = 0x12 + NFT_META_PKTTYPE = 0x13 + NFT_META_CPU = 0x14 + NFT_META_IIFGROUP = 0x15 + NFT_META_OIFGROUP = 0x16 + NFT_META_CGROUP = 0x17 + NFT_META_PRANDOM = 0x18 + NFT_RT_CLASSID = 0x0 + NFT_RT_NEXTHOP4 = 0x1 + NFT_RT_NEXTHOP6 = 0x2 + NFT_RT_TCPMSS = 0x3 + NFT_HASH_JENKINS = 0x0 + NFT_HASH_SYM = 0x1 + NFTA_HASH_UNSPEC = 0x0 + NFTA_HASH_SREG = 0x1 + NFTA_HASH_DREG = 0x2 + NFTA_HASH_LEN = 0x3 + NFTA_HASH_MODULUS = 0x4 + NFTA_HASH_SEED = 0x5 + NFTA_HASH_OFFSET = 0x6 + NFTA_HASH_TYPE = 0x7 + NFTA_META_UNSPEC = 0x0 + NFTA_META_DREG = 0x1 + NFTA_META_KEY = 0x2 + NFTA_META_SREG = 0x3 + NFTA_RT_UNSPEC = 0x0 + NFTA_RT_DREG = 0x1 + NFTA_RT_KEY = 0x2 + NFT_CT_STATE = 0x0 + NFT_CT_DIRECTION = 0x1 + NFT_CT_STATUS = 0x2 + NFT_CT_MARK = 0x3 + NFT_CT_SECMARK = 0x4 + NFT_CT_EXPIRATION = 0x5 + NFT_CT_HELPER = 0x6 + NFT_CT_L3PROTOCOL = 0x7 + NFT_CT_SRC = 0x8 + NFT_CT_DST = 0x9 + NFT_CT_PROTOCOL = 0xa + NFT_CT_PROTO_SRC = 0xb + NFT_CT_PROTO_DST = 0xc + NFT_CT_LABELS = 0xd + NFT_CT_PKTS = 0xe + NFT_CT_BYTES = 0xf + NFT_CT_AVGPKT = 0x10 + NFT_CT_ZONE = 0x11 + NFT_CT_EVENTMASK = 0x12 + NFTA_CT_UNSPEC = 0x0 + NFTA_CT_DREG = 0x1 + NFTA_CT_KEY = 0x2 + NFTA_CT_DIRECTION = 0x3 + NFTA_CT_SREG = 0x4 + NFT_LIMIT_PKTS = 0x0 + NFT_LIMIT_PKT_BYTES = 0x1 + NFT_LIMIT_F_INV = 0x1 + NFTA_LIMIT_UNSPEC = 0x0 + NFTA_LIMIT_RATE = 0x1 + NFTA_LIMIT_UNIT = 0x2 + NFTA_LIMIT_BURST = 0x3 + NFTA_LIMIT_TYPE = 0x4 + NFTA_LIMIT_FLAGS = 0x5 + NFTA_LIMIT_PAD = 0x6 + NFTA_COUNTER_UNSPEC = 0x0 + NFTA_COUNTER_BYTES = 0x1 + NFTA_COUNTER_PACKETS = 0x2 + NFTA_COUNTER_PAD = 0x3 + NFTA_LOG_UNSPEC = 0x0 + NFTA_LOG_GROUP = 0x1 + NFTA_LOG_PREFIX = 0x2 + NFTA_LOG_SNAPLEN = 0x3 + NFTA_LOG_QTHRESHOLD = 0x4 + NFTA_LOG_LEVEL = 0x5 + NFTA_LOG_FLAGS = 0x6 + NFTA_QUEUE_UNSPEC = 0x0 + NFTA_QUEUE_NUM = 0x1 + NFTA_QUEUE_TOTAL = 0x2 + NFTA_QUEUE_FLAGS = 0x3 + NFTA_QUEUE_SREG_QNUM = 0x4 + NFT_QUOTA_F_INV = 0x1 + NFT_QUOTA_F_DEPLETED = 0x2 + NFTA_QUOTA_UNSPEC = 0x0 + NFTA_QUOTA_BYTES = 0x1 + NFTA_QUOTA_FLAGS = 0x2 + NFTA_QUOTA_PAD = 0x3 + NFTA_QUOTA_CONSUMED = 0x4 + NFT_REJECT_ICMP_UNREACH = 0x0 + NFT_REJECT_TCP_RST = 0x1 + NFT_REJECT_ICMPX_UNREACH = 0x2 + NFT_REJECT_ICMPX_NO_ROUTE = 0x0 + NFT_REJECT_ICMPX_PORT_UNREACH = 0x1 + NFT_REJECT_ICMPX_HOST_UNREACH = 0x2 + NFT_REJECT_ICMPX_ADMIN_PROHIBITED = 0x3 + NFTA_REJECT_UNSPEC = 0x0 + NFTA_REJECT_TYPE = 0x1 + NFTA_REJECT_ICMP_CODE = 0x2 + NFT_NAT_SNAT = 0x0 + NFT_NAT_DNAT = 0x1 + NFTA_NAT_UNSPEC = 0x0 + NFTA_NAT_TYPE = 0x1 + NFTA_NAT_FAMILY = 0x2 + NFTA_NAT_REG_ADDR_MIN = 0x3 + NFTA_NAT_REG_ADDR_MAX = 0x4 + NFTA_NAT_REG_PROTO_MIN = 0x5 + NFTA_NAT_REG_PROTO_MAX = 0x6 + NFTA_NAT_FLAGS = 0x7 + NFTA_MASQ_UNSPEC = 0x0 + NFTA_MASQ_FLAGS = 0x1 + NFTA_MASQ_REG_PROTO_MIN = 0x2 + NFTA_MASQ_REG_PROTO_MAX = 0x3 + NFTA_REDIR_UNSPEC = 0x0 + NFTA_REDIR_REG_PROTO_MIN = 0x1 + NFTA_REDIR_REG_PROTO_MAX = 0x2 + NFTA_REDIR_FLAGS = 0x3 + NFTA_DUP_UNSPEC = 0x0 + NFTA_DUP_SREG_ADDR = 0x1 + NFTA_DUP_SREG_DEV = 0x2 + NFTA_FWD_UNSPEC = 0x0 + NFTA_FWD_SREG_DEV = 0x1 + NFTA_OBJREF_UNSPEC = 0x0 + NFTA_OBJREF_IMM_TYPE = 0x1 + NFTA_OBJREF_IMM_NAME = 0x2 + NFTA_OBJREF_SET_SREG = 0x3 + NFTA_OBJREF_SET_NAME = 0x4 + NFTA_OBJREF_SET_ID = 0x5 + NFTA_GEN_UNSPEC = 0x0 + NFTA_GEN_ID = 0x1 + NFTA_GEN_PROC_PID = 0x2 + NFTA_GEN_PROC_NAME = 0x3 + NFTA_FIB_UNSPEC = 0x0 + NFTA_FIB_DREG = 0x1 + NFTA_FIB_RESULT = 0x2 + NFTA_FIB_FLAGS = 0x3 + NFT_FIB_RESULT_UNSPEC = 0x0 + NFT_FIB_RESULT_OIF = 0x1 + NFT_FIB_RESULT_OIFNAME = 0x2 + NFT_FIB_RESULT_ADDRTYPE = 0x3 + NFTA_FIB_F_SADDR = 0x1 + NFTA_FIB_F_DADDR = 0x2 + NFTA_FIB_F_MARK = 0x4 + NFTA_FIB_F_IIF = 0x8 + NFTA_FIB_F_OIF = 0x10 + NFTA_FIB_F_PRESENT = 0x20 + NFTA_CT_HELPER_UNSPEC = 0x0 + NFTA_CT_HELPER_NAME = 0x1 + NFTA_CT_HELPER_L3PROTO = 0x2 + NFTA_CT_HELPER_L4PROTO = 0x3 + NFTA_OBJ_UNSPEC = 0x0 + NFTA_OBJ_TABLE = 0x1 + NFTA_OBJ_NAME = 0x2 + NFTA_OBJ_TYPE = 0x3 + NFTA_OBJ_DATA = 0x4 + NFTA_OBJ_USE = 0x5 + NFTA_TRACE_UNSPEC = 0x0 + NFTA_TRACE_TABLE = 0x1 + NFTA_TRACE_CHAIN = 0x2 + NFTA_TRACE_RULE_HANDLE = 0x3 + NFTA_TRACE_TYPE = 0x4 + NFTA_TRACE_VERDICT = 0x5 + NFTA_TRACE_ID = 0x6 + NFTA_TRACE_LL_HEADER = 0x7 + NFTA_TRACE_NETWORK_HEADER = 0x8 + NFTA_TRACE_TRANSPORT_HEADER = 0x9 + NFTA_TRACE_IIF = 0xa + NFTA_TRACE_IIFTYPE = 0xb + NFTA_TRACE_OIF = 0xc + NFTA_TRACE_OIFTYPE = 0xd + NFTA_TRACE_MARK = 0xe + NFTA_TRACE_NFPROTO = 0xf + NFTA_TRACE_POLICY = 0x10 + NFTA_TRACE_PAD = 0x11 + NFT_TRACETYPE_UNSPEC = 0x0 + NFT_TRACETYPE_POLICY = 0x1 + NFT_TRACETYPE_RETURN = 0x2 + NFT_TRACETYPE_RULE = 0x3 + NFTA_NG_UNSPEC = 0x0 + NFTA_NG_DREG = 0x1 + NFTA_NG_MODULUS = 0x2 + NFTA_NG_TYPE = 0x3 + NFTA_NG_OFFSET = 0x4 + NFT_NG_INCREMENTAL = 0x0 + NFT_NG_RANDOM = 0x1 +) + +type RTCTime struct { + Sec int32 + Min int32 + Hour int32 + Mday int32 + Mon int32 + Year int32 + Wday int32 + Yday int32 + Isdst int32 +} + +type RTCWkAlrm struct { + Enabled uint8 + Pending uint8 + _ [2]byte + Time RTCTime +} + +type RTCPLLInfo struct { + Ctrl int32 + Value int32 + Max int32 + Min int32 + Posmult int32 + Negmult int32 + Clock int64 +} diff --git a/vendor/golang.org/x/sys/unix/ztypes_openbsd_amd64.go b/vendor/golang.org/x/sys/unix/ztypes_openbsd_amd64.go index d53141085..5a9c81848 100644 --- a/vendor/golang.org/x/sys/unix/ztypes_openbsd_amd64.go +++ b/vendor/golang.org/x/sys/unix/ztypes_openbsd_amd64.go @@ -74,30 +74,30 @@ const ( ) type Stat_t struct { - Mode uint32 - Dev int32 - Ino uint64 - Nlink uint32 - Uid uint32 - Gid uint32 - Rdev int32 - Atim Timespec - Mtim Timespec - Ctim Timespec - Size int64 - Blocks int64 - Blksize uint32 - Flags uint32 - Gen uint32 - Pad_cgo_0 [4]byte - X__st_birthtim Timespec + Mode uint32 + Dev int32 + Ino uint64 + Nlink uint32 + Uid uint32 + Gid uint32 + Rdev int32 + Atim Timespec + Mtim Timespec + Ctim Timespec + Size int64 + Blocks int64 + Blksize int32 + Flags uint32 + Gen uint32 + _ [4]byte + _ Timespec } type Statfs_t struct { F_flags uint32 F_bsize uint32 F_iosize uint32 - Pad_cgo_0 [4]byte + _ [4]byte F_blocks uint64 F_bfree uint64 F_bavail int64 @@ -116,7 +116,7 @@ type Statfs_t struct { F_mntonname [90]int8 F_mntfromname [90]int8 F_mntfromspec [90]int8 - Pad_cgo_1 [2]byte + _ [2]byte Mount_info [160]byte } @@ -129,13 +129,13 @@ type Flock_t struct { } type Dirent struct { - Fileno uint64 - Off int64 - Reclen uint16 - Type uint8 - Namlen uint8 - X__d_padding [4]uint8 - Name [256]int8 + Fileno uint64 + Off int64 + Reclen uint16 + Type uint8 + Namlen uint8 + _ [4]uint8 + Name [256]int8 } type Fsid struct { @@ -216,10 +216,10 @@ type IPv6Mreq struct { type Msghdr struct { Name *byte Namelen uint32 - Pad_cgo_0 [4]byte + _ [4]byte Iov *Iovec Iovlen uint32 - Pad_cgo_1 [4]byte + _ [4]byte Control *byte Controllen uint32 Flags int32 @@ -281,8 +281,8 @@ type FdSet struct { } const ( - SizeofIfMsghdr = 0xf8 - SizeofIfData = 0xe0 + SizeofIfMsghdr = 0xa8 + SizeofIfData = 0x90 SizeofIfaMsghdr = 0x18 SizeofIfAnnounceMsghdr = 0x1a SizeofRtMsghdr = 0x60 @@ -311,7 +311,7 @@ type IfData struct { Link_state uint8 Mtu uint32 Metric uint32 - Pad uint32 + Rdomain uint32 Baudrate uint64 Ipackets uint64 Ierrors uint64 @@ -323,12 +323,11 @@ type IfData struct { Imcasts uint64 Omcasts uint64 Iqdrops uint64 + Oqdrops uint64 Noproto uint64 Capabilities uint32 - Pad_cgo_0 [4]byte + _ [4]byte Lastchange Timeval - Mclpool [7]Mclpool - Pad_cgo_1 [4]byte } type IfaMsghdr struct { @@ -389,13 +388,7 @@ type RtMetrics struct { Pad uint32 } -type Mclpool struct { - Grown int32 - Alive uint16 - Hwm uint16 - Cwm uint16 - Lwm uint16 -} +type Mclpool struct{} const ( SizeofBpfVersion = 0x4 @@ -416,9 +409,9 @@ type BpfStat struct { } type BpfProgram struct { - Len uint32 - Pad_cgo_0 [4]byte - Insns *BpfInsn + Len uint32 + _ [4]byte + Insns *BpfInsn } type BpfInsn struct { @@ -429,11 +422,11 @@ type BpfInsn struct { } type BpfHdr struct { - Tstamp BpfTimeval - Caplen uint32 - Datalen uint32 - Hdrlen uint16 - Pad_cgo_0 [2]byte + Tstamp BpfTimeval + Caplen uint32 + Datalen uint32 + Hdrlen uint16 + _ [2]byte } type BpfTimeval struct { diff --git a/vendor/golang.org/x/sys/windows/asm_windows_amd64.s b/vendor/golang.org/x/sys/windows/asm_windows_amd64.s index 4d025ab55..5bfdf7974 100644 --- a/vendor/golang.org/x/sys/windows/asm_windows_amd64.s +++ b/vendor/golang.org/x/sys/windows/asm_windows_amd64.s @@ -9,5 +9,5 @@ TEXT ·getprocaddress(SB), 7, $0-32 JMP syscall·getprocaddress(SB) -TEXT ·loadlibrary(SB), 7, $0-8 +TEXT ·loadlibrary(SB), 7, $0-24 JMP syscall·loadlibrary(SB) diff --git a/vendor/golang.org/x/sys/windows/registry/zsyscall_windows.go b/vendor/golang.org/x/sys/windows/registry/zsyscall_windows.go index ceebdd772..3778075da 100644 --- a/vendor/golang.org/x/sys/windows/registry/zsyscall_windows.go +++ b/vendor/golang.org/x/sys/windows/registry/zsyscall_windows.go @@ -1,4 +1,4 @@ -// MACHINE GENERATED BY 'go generate' COMMAND; DO NOT EDIT +// Code generated by 'go generate'; DO NOT EDIT. package registry diff --git a/vendor/golang.org/x/sys/windows/security_windows.go b/vendor/golang.org/x/sys/windows/security_windows.go index f1ec5dc4e..4f17a3331 100644 --- a/vendor/golang.org/x/sys/windows/security_windows.go +++ b/vendor/golang.org/x/sys/windows/security_windows.go @@ -296,6 +296,7 @@ const ( TOKEN_ADJUST_PRIVILEGES TOKEN_ADJUST_GROUPS TOKEN_ADJUST_DEFAULT + TOKEN_ADJUST_SESSIONID TOKEN_ALL_ACCESS = STANDARD_RIGHTS_REQUIRED | TOKEN_ASSIGN_PRIMARY | @@ -305,7 +306,8 @@ const ( TOKEN_QUERY_SOURCE | TOKEN_ADJUST_PRIVILEGES | TOKEN_ADJUST_GROUPS | - TOKEN_ADJUST_DEFAULT + TOKEN_ADJUST_DEFAULT | + TOKEN_ADJUST_SESSIONID TOKEN_READ = STANDARD_RIGHTS_READ | TOKEN_QUERY TOKEN_WRITE = STANDARD_RIGHTS_WRITE | TOKEN_ADJUST_PRIVILEGES | diff --git a/vendor/golang.org/x/sys/windows/service.go b/vendor/golang.org/x/sys/windows/service.go index a500dd7df..24aa90bbb 100644 --- a/vendor/golang.org/x/sys/windows/service.go +++ b/vendor/golang.org/x/sys/windows/service.go @@ -162,3 +162,4 @@ type ENUM_SERVICE_STATUS_PROCESS struct { //sys ChangeServiceConfig2(service Handle, infoLevel uint32, info *byte) (err error) = advapi32.ChangeServiceConfig2W //sys QueryServiceConfig2(service Handle, infoLevel uint32, buff *byte, buffSize uint32, bytesNeeded *uint32) (err error) = advapi32.QueryServiceConfig2W //sys EnumServicesStatusEx(mgr Handle, infoLevel uint32, serviceType uint32, serviceState uint32, services *byte, bufSize uint32, bytesNeeded *uint32, servicesReturned *uint32, resumeHandle *uint32, groupName *uint16) (err error) = advapi32.EnumServicesStatusExW +//sys QueryServiceStatusEx(service Handle, infoLevel uint32, buff *byte, buffSize uint32, bytesNeeded *uint32) (err error) = advapi32.QueryServiceStatusEx diff --git a/vendor/golang.org/x/sys/windows/syscall_windows_test.go b/vendor/golang.org/x/sys/windows/syscall_windows_test.go index 9c7133cc4..0e27464e8 100644 --- a/vendor/golang.org/x/sys/windows/syscall_windows_test.go +++ b/vendor/golang.org/x/sys/windows/syscall_windows_test.go @@ -105,3 +105,9 @@ func ExampleLoadLibrary() { build := uint16(r >> 16) print("windows version ", major, ".", minor, " (Build ", build, ")\n") } + +func TestTOKEN_ALL_ACCESS(t *testing.T) { + if windows.TOKEN_ALL_ACCESS != 0xF01FF { + t.Errorf("TOKEN_ALL_ACCESS = %x, want 0xF01FF", windows.TOKEN_ALL_ACCESS) + } +} diff --git a/vendor/golang.org/x/sys/windows/types_windows.go b/vendor/golang.org/x/sys/windows/types_windows.go index 52c2037b6..b4e424788 100644 --- a/vendor/golang.org/x/sys/windows/types_windows.go +++ b/vendor/golang.org/x/sys/windows/types_windows.go @@ -312,6 +312,14 @@ var ( OID_SGC_NETSCAPE = []byte("2.16.840.1.113730.4.1\x00") ) +// Pointer represents a pointer to an arbitrary Windows type. +// +// Pointer-typed fields may point to one of many different types. It's +// up to the caller to provide a pointer to the appropriate type, cast +// to Pointer. The caller must obey the unsafe.Pointer rules while +// doing so. +type Pointer *struct{} + // Invented values to support what package os expects. type Timeval struct { Sec int32 @@ -880,11 +888,15 @@ type MibIfRow struct { Descr [MAXLEN_IFDESCR]byte } +type CertInfo struct { + // Not implemented +} + type CertContext struct { EncodingType uint32 EncodedCert *byte Length uint32 - CertInfo uintptr + CertInfo *CertInfo Store Handle } @@ -899,12 +911,16 @@ type CertChainContext struct { RevocationFreshnessTime uint32 } +type CertTrustListInfo struct { + // Not implemented +} + type CertSimpleChain struct { Size uint32 TrustStatus CertTrustStatus NumElements uint32 Elements **CertChainElement - TrustListInfo uintptr + TrustListInfo *CertTrustListInfo HasRevocationFreshnessTime uint32 RevocationFreshnessTime uint32 } @@ -919,14 +935,18 @@ type CertChainElement struct { ExtendedErrorInfo *uint16 } +type CertRevocationCrlInfo struct { + // Not implemented +} + type CertRevocationInfo struct { Size uint32 RevocationResult uint32 RevocationOid *byte - OidSpecificInfo uintptr + OidSpecificInfo Pointer HasFreshnessTime uint32 FreshnessTime uint32 - CrlInfo uintptr // *CertRevocationCrlInfo + CrlInfo *CertRevocationCrlInfo } type CertTrustStatus struct { @@ -957,7 +977,7 @@ type CertChainPara struct { type CertChainPolicyPara struct { Size uint32 Flags uint32 - ExtraPolicyPara uintptr + ExtraPolicyPara Pointer } type SSLExtraCertChainPolicyPara struct { @@ -972,7 +992,7 @@ type CertChainPolicyStatus struct { Error uint32 ChainIndex uint32 ElementIndex uint32 - ExtraPolicyStatus uintptr + ExtraPolicyStatus Pointer } const ( diff --git a/vendor/golang.org/x/sys/windows/zsyscall_windows.go b/vendor/golang.org/x/sys/windows/zsyscall_windows.go index c7b3b15ea..fc56aec03 100644 --- a/vendor/golang.org/x/sys/windows/zsyscall_windows.go +++ b/vendor/golang.org/x/sys/windows/zsyscall_windows.go @@ -1,4 +1,4 @@ -// MACHINE GENERATED BY 'go generate' COMMAND; DO NOT EDIT +// Code generated by 'go generate'; DO NOT EDIT. package windows @@ -65,6 +65,7 @@ var ( procChangeServiceConfig2W = modadvapi32.NewProc("ChangeServiceConfig2W") procQueryServiceConfig2W = modadvapi32.NewProc("QueryServiceConfig2W") procEnumServicesStatusExW = modadvapi32.NewProc("EnumServicesStatusExW") + procQueryServiceStatusEx = modadvapi32.NewProc("QueryServiceStatusEx") procGetLastError = modkernel32.NewProc("GetLastError") procLoadLibraryW = modkernel32.NewProc("LoadLibraryW") procLoadLibraryExW = modkernel32.NewProc("LoadLibraryExW") @@ -472,6 +473,18 @@ func EnumServicesStatusEx(mgr Handle, infoLevel uint32, serviceType uint32, serv return } +func QueryServiceStatusEx(service Handle, infoLevel uint32, buff *byte, buffSize uint32, bytesNeeded *uint32) (err error) { + r1, _, e1 := syscall.Syscall6(procQueryServiceStatusEx.Addr(), 5, uintptr(service), uintptr(infoLevel), uintptr(unsafe.Pointer(buff)), uintptr(buffSize), uintptr(unsafe.Pointer(bytesNeeded)), 0) + if r1 == 0 { + if e1 != 0 { + err = errnoErr(e1) + } else { + err = syscall.EINVAL + } + } + return +} + func GetLastError() (lasterr error) { r0, _, _ := syscall.Syscall(procGetLastError.Addr(), 0, 0, 0, 0) if r0 != 0 { diff --git a/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go b/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go index e32d3b4f3..9d57e0584 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/annotations/http.pb.go @@ -25,14 +25,14 @@ type Http struct { // A list of HTTP configuration rules that apply to individual API methods. // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*HttpRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*HttpRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` // When set to true, URL path parmeters will be fully URI-decoded except in // cases of single segment matches in reserved expansion, where "%2F" will be // left encoded. // // The default behavior is to not decode RFC 6570 reserved characters in multi // segment matches. - FullyDecodeReservedExpansion bool `protobuf:"varint,2,opt,name=fully_decode_reserved_expansion,json=fullyDecodeReservedExpansion" json:"fully_decode_reserved_expansion,omitempty"` + FullyDecodeReservedExpansion bool `protobuf:"varint,2,opt,name=fully_decode_reserved_expansion,json=fullyDecodeReservedExpansion,proto3" json:"fully_decode_reserved_expansion,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -297,7 +297,7 @@ type HttpRule struct { // Selects methods to which this rule applies. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // Determines the URL pattern is matched by this rules. This pattern can be // used with any of the {get|put|post|delete|patch} methods. A custom method // can be defined using the 'custom' field. @@ -314,11 +314,11 @@ type HttpRule struct { // `*` for mapping all fields not captured by the path pattern to the HTTP // body. NOTE: the referred field must not be a repeated field and must be // present at the top-level of request message type. - Body string `protobuf:"bytes,7,opt,name=body" json:"body,omitempty"` + Body string `protobuf:"bytes,7,opt,name=body,proto3" json:"body,omitempty"` // Additional HTTP bindings for the selector. Nested bindings must // not contain an `additional_bindings` field themselves (that is, // the nesting may only be one level deep). - AdditionalBindings []*HttpRule `protobuf:"bytes,11,rep,name=additional_bindings,json=additionalBindings" json:"additional_bindings,omitempty"` + AdditionalBindings []*HttpRule `protobuf:"bytes,11,rep,name=additional_bindings,json=additionalBindings,proto3" json:"additional_bindings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -353,22 +353,22 @@ type isHttpRule_Pattern interface { } type HttpRule_Get struct { - Get string `protobuf:"bytes,2,opt,name=get,oneof"` + Get string `protobuf:"bytes,2,opt,name=get,proto3,oneof"` } type HttpRule_Put struct { - Put string `protobuf:"bytes,3,opt,name=put,oneof"` + Put string `protobuf:"bytes,3,opt,name=put,proto3,oneof"` } type HttpRule_Post struct { - Post string `protobuf:"bytes,4,opt,name=post,oneof"` + Post string `protobuf:"bytes,4,opt,name=post,proto3,oneof"` } type HttpRule_Delete struct { - Delete string `protobuf:"bytes,5,opt,name=delete,oneof"` + Delete string `protobuf:"bytes,5,opt,name=delete,proto3,oneof"` } type HttpRule_Patch struct { - Patch string `protobuf:"bytes,6,opt,name=patch,oneof"` + Patch string `protobuf:"bytes,6,opt,name=patch,proto3,oneof"` } type HttpRule_Custom struct { - Custom *CustomHttpPattern `protobuf:"bytes,8,opt,name=custom,oneof"` + Custom *CustomHttpPattern `protobuf:"bytes,8,opt,name=custom,proto3,oneof"` } func (*HttpRule_Get) isHttpRule_Pattern() {} @@ -581,9 +581,9 @@ func _HttpRule_OneofSizer(msg proto.Message) (n int) { // A custom pattern is used for defining custom HTTP verb. type CustomHttpPattern struct { // The name of this custom HTTP verb. - Kind string `protobuf:"bytes,1,opt,name=kind" json:"kind,omitempty"` + Kind string `protobuf:"bytes,1,opt,name=kind,proto3" json:"kind,omitempty"` // The path matched by this custom verb. - Path string `protobuf:"bytes,2,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,2,opt,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/authorization_config.pb.go b/vendor/google.golang.org/genproto/googleapis/api/authorization_config.pb.go index 1d713fa94..419629e44 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/authorization_config.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/authorization_config.pb.go @@ -31,7 +31,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type AuthorizationConfig struct { // The name of the authorization provider, such as // firebaserules.googleapis.com. - Provider string `protobuf:"bytes,1,opt,name=provider" json:"provider,omitempty"` + Provider string `protobuf:"bytes,1,opt,name=provider,proto3" json:"provider,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -41,7 +41,7 @@ func (m *AuthorizationConfig) Reset() { *m = AuthorizationConfig{} } func (m *AuthorizationConfig) String() string { return proto.CompactTextString(m) } func (*AuthorizationConfig) ProtoMessage() {} func (*AuthorizationConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_authorization_config_8049bd02b6508303, []int{0} + return fileDescriptor_authorization_config_aed97ac7a3a4cb16, []int{0} } func (m *AuthorizationConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AuthorizationConfig.Unmarshal(m, b) @@ -73,10 +73,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/experimental/authorization_config.proto", fileDescriptor_authorization_config_8049bd02b6508303) + proto.RegisterFile("google/api/experimental/authorization_config.proto", fileDescriptor_authorization_config_aed97ac7a3a4cb16) } -var fileDescriptor_authorization_config_8049bd02b6508303 = []byte{ +var fileDescriptor_authorization_config_aed97ac7a3a4cb16 = []byte{ // 180 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x32, 0x4a, 0xcf, 0xcf, 0x4f, 0xcf, 0x49, 0xd5, 0x4f, 0x2c, 0xc8, 0xd4, 0x4f, 0xad, 0x28, 0x48, 0x2d, 0xca, 0xcc, 0x4d, 0xcd, diff --git a/vendor/google.golang.org/genproto/googleapis/api/configchange/config_change.pb.go b/vendor/google.golang.org/genproto/googleapis/api/configchange/config_change.pb.go index ce5e58432..367846ec6 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/configchange/config_change.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/configchange/config_change.pb.go @@ -53,7 +53,7 @@ func (x ChangeType) String() string { return proto.EnumName(ChangeType_name, int32(x)) } func (ChangeType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_config_change_4f72b3734b1e4109, []int{0} + return fileDescriptor_config_change_0bcca47df127ce62, []int{0} } // Output generated from semantically comparing two versions of a service @@ -72,18 +72,18 @@ type ConfigChange struct { // - visibility.rules[selector=="google.LibraryService.CreateBook"].restriction // - quota.metric_rules[selector=="google"].metric_costs[key=="reads"].value // - logging.producer_destinations[0] - Element string `protobuf:"bytes,1,opt,name=element" json:"element,omitempty"` + Element string `protobuf:"bytes,1,opt,name=element,proto3" json:"element,omitempty"` // Value of the changed object in the old Service configuration, // in JSON format. This field will not be populated if ChangeType == ADDED. - OldValue string `protobuf:"bytes,2,opt,name=old_value,json=oldValue" json:"old_value,omitempty"` + OldValue string `protobuf:"bytes,2,opt,name=old_value,json=oldValue,proto3" json:"old_value,omitempty"` // Value of the changed object in the new Service configuration, // in JSON format. This field will not be populated if ChangeType == REMOVED. - NewValue string `protobuf:"bytes,3,opt,name=new_value,json=newValue" json:"new_value,omitempty"` + NewValue string `protobuf:"bytes,3,opt,name=new_value,json=newValue,proto3" json:"new_value,omitempty"` // The type for this change, either ADDED, REMOVED, or MODIFIED. - ChangeType ChangeType `protobuf:"varint,4,opt,name=change_type,json=changeType,enum=google.api.ChangeType" json:"change_type,omitempty"` + ChangeType ChangeType `protobuf:"varint,4,opt,name=change_type,json=changeType,proto3,enum=google.api.ChangeType" json:"change_type,omitempty"` // Collection of advice provided for this change, useful for determining the // possible impact of this change. - Advices []*Advice `protobuf:"bytes,5,rep,name=advices" json:"advices,omitempty"` + Advices []*Advice `protobuf:"bytes,5,rep,name=advices,proto3" json:"advices,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -93,7 +93,7 @@ func (m *ConfigChange) Reset() { *m = ConfigChange{} } func (m *ConfigChange) String() string { return proto.CompactTextString(m) } func (*ConfigChange) ProtoMessage() {} func (*ConfigChange) Descriptor() ([]byte, []int) { - return fileDescriptor_config_change_4f72b3734b1e4109, []int{0} + return fileDescriptor_config_change_0bcca47df127ce62, []int{0} } func (m *ConfigChange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ConfigChange.Unmarshal(m, b) @@ -153,7 +153,7 @@ func (m *ConfigChange) GetAdvices() []*Advice { type Advice struct { // Useful description for why this advice was applied and what actions should // be taken to mitigate any implied risks. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -163,7 +163,7 @@ func (m *Advice) Reset() { *m = Advice{} } func (m *Advice) String() string { return proto.CompactTextString(m) } func (*Advice) ProtoMessage() {} func (*Advice) Descriptor() ([]byte, []int) { - return fileDescriptor_config_change_4f72b3734b1e4109, []int{1} + return fileDescriptor_config_change_0bcca47df127ce62, []int{1} } func (m *Advice) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Advice.Unmarshal(m, b) @@ -197,10 +197,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/config_change.proto", fileDescriptor_config_change_4f72b3734b1e4109) + proto.RegisterFile("google/api/config_change.proto", fileDescriptor_config_change_0bcca47df127ce62) } -var fileDescriptor_config_change_4f72b3734b1e4109 = []byte{ +var fileDescriptor_config_change_0bcca47df127ce62 = []byte{ // 338 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x4c, 0x91, 0xcd, 0x4e, 0xc2, 0x40, 0x14, 0x85, 0x2d, 0xff, 0xdc, 0x12, 0x82, 0xb3, 0xd0, 0x26, 0x24, 0xa6, 0x61, 0x45, 0x88, 0x69, diff --git a/vendor/google.golang.org/genproto/googleapis/api/distribution/distribution.pb.go b/vendor/google.golang.org/genproto/googleapis/api/distribution/distribution.pb.go index 2933b1417..ea2ca8aad 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/distribution/distribution.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/distribution/distribution.pb.go @@ -38,10 +38,10 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // will render the `mean` and `sum_of_squared_deviation` fields meaningless. type Distribution struct { // The number of values in the population. Must be non-negative. - Count int64 `protobuf:"varint,1,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,1,opt,name=count,proto3" json:"count,omitempty"` // The arithmetic mean of the values in the population. If `count` is zero // then this field must be zero. - Mean float64 `protobuf:"fixed64,2,opt,name=mean" json:"mean,omitempty"` + Mean float64 `protobuf:"fixed64,2,opt,name=mean,proto3" json:"mean,omitempty"` // The sum of squared deviations from the mean of the values in the // population. For values x_i this is: // @@ -51,12 +51,12 @@ type Distribution struct { // describes Welford's method for accumulating this sum in one pass. // // If `count` is zero then this field must be zero. - SumOfSquaredDeviation float64 `protobuf:"fixed64,3,opt,name=sum_of_squared_deviation,json=sumOfSquaredDeviation" json:"sum_of_squared_deviation,omitempty"` + SumOfSquaredDeviation float64 `protobuf:"fixed64,3,opt,name=sum_of_squared_deviation,json=sumOfSquaredDeviation,proto3" json:"sum_of_squared_deviation,omitempty"` // If specified, contains the range of the population values. The field // must not be present if the `count` is zero. - Range *Distribution_Range `protobuf:"bytes,4,opt,name=range" json:"range,omitempty"` + Range *Distribution_Range `protobuf:"bytes,4,opt,name=range,proto3" json:"range,omitempty"` // Defines the histogram bucket boundaries. - BucketOptions *Distribution_BucketOptions `protobuf:"bytes,6,opt,name=bucket_options,json=bucketOptions" json:"bucket_options,omitempty"` + BucketOptions *Distribution_BucketOptions `protobuf:"bytes,6,opt,name=bucket_options,json=bucketOptions,proto3" json:"bucket_options,omitempty"` // If `bucket_options` is given, then the sum of the values in `bucket_counts` // must equal the value in `count`. If `bucket_options` is not given, no // `bucket_counts` fields may be given. @@ -69,7 +69,7 @@ type Distribution struct { // `bucket_options`. // // Any suffix of trailing zero bucket_count fields may be omitted. - BucketCounts []int64 `protobuf:"varint,7,rep,packed,name=bucket_counts,json=bucketCounts" json:"bucket_counts,omitempty"` + BucketCounts []int64 `protobuf:"varint,7,rep,packed,name=bucket_counts,json=bucketCounts,proto3" json:"bucket_counts,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -144,9 +144,9 @@ func (m *Distribution) GetBucketCounts() []int64 { // The range of the population values. type Distribution_Range struct { // The minimum of the population values. - Min float64 `protobuf:"fixed64,1,opt,name=min" json:"min,omitempty"` + Min float64 `protobuf:"fixed64,1,opt,name=min,proto3" json:"min,omitempty"` // The maximum of the population values. - Max float64 `protobuf:"fixed64,2,opt,name=max" json:"max,omitempty"` + Max float64 `protobuf:"fixed64,2,opt,name=max,proto3" json:"max,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -255,13 +255,13 @@ type isDistribution_BucketOptions_Options interface { } type Distribution_BucketOptions_LinearBuckets struct { - LinearBuckets *Distribution_BucketOptions_Linear `protobuf:"bytes,1,opt,name=linear_buckets,json=linearBuckets,oneof"` + LinearBuckets *Distribution_BucketOptions_Linear `protobuf:"bytes,1,opt,name=linear_buckets,json=linearBuckets,proto3,oneof"` } type Distribution_BucketOptions_ExponentialBuckets struct { - ExponentialBuckets *Distribution_BucketOptions_Exponential `protobuf:"bytes,2,opt,name=exponential_buckets,json=exponentialBuckets,oneof"` + ExponentialBuckets *Distribution_BucketOptions_Exponential `protobuf:"bytes,2,opt,name=exponential_buckets,json=exponentialBuckets,proto3,oneof"` } type Distribution_BucketOptions_ExplicitBuckets struct { - ExplicitBuckets *Distribution_BucketOptions_Explicit `protobuf:"bytes,3,opt,name=explicit_buckets,json=explicitBuckets,oneof"` + ExplicitBuckets *Distribution_BucketOptions_Explicit `protobuf:"bytes,3,opt,name=explicit_buckets,json=explicitBuckets,proto3,oneof"` } func (*Distribution_BucketOptions_LinearBuckets) isDistribution_BucketOptions_Options() {} @@ -400,11 +400,11 @@ func _Distribution_BucketOptions_OneofSizer(msg proto.Message) (n int) { // Lower bound (1 <= i < N): offset + (width * (i - 1)). type Distribution_BucketOptions_Linear struct { // Must be greater than 0. - NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets" json:"num_finite_buckets,omitempty"` + NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets,proto3" json:"num_finite_buckets,omitempty"` // Must be greater than 0. - Width float64 `protobuf:"fixed64,2,opt,name=width" json:"width,omitempty"` + Width float64 `protobuf:"fixed64,2,opt,name=width,proto3" json:"width,omitempty"` // Lower bound of the first bucket. - Offset float64 `protobuf:"fixed64,3,opt,name=offset" json:"offset,omitempty"` + Offset float64 `protobuf:"fixed64,3,opt,name=offset,proto3" json:"offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -466,11 +466,11 @@ func (m *Distribution_BucketOptions_Linear) GetOffset() float64 { // Lower bound (1 <= i < N): scale * (growth_factor ^ (i - 1)). type Distribution_BucketOptions_Exponential struct { // Must be greater than 0. - NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets" json:"num_finite_buckets,omitempty"` + NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets,proto3" json:"num_finite_buckets,omitempty"` // Must be greater than 1. - GrowthFactor float64 `protobuf:"fixed64,2,opt,name=growth_factor,json=growthFactor" json:"growth_factor,omitempty"` + GrowthFactor float64 `protobuf:"fixed64,2,opt,name=growth_factor,json=growthFactor,proto3" json:"growth_factor,omitempty"` // Must be greater than 0. - Scale float64 `protobuf:"fixed64,3,opt,name=scale" json:"scale,omitempty"` + Scale float64 `protobuf:"fixed64,3,opt,name=scale,proto3" json:"scale,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -536,7 +536,7 @@ func (m *Distribution_BucketOptions_Exponential) GetScale() float64 { // common boundary of the overflow and underflow buckets. type Distribution_BucketOptions_Explicit struct { // The values must be monotonically increasing. - Bounds []float64 `protobuf:"fixed64,1,rep,packed,name=bounds" json:"bounds,omitempty"` + Bounds []float64 `protobuf:"fixed64,1,rep,packed,name=bounds,proto3" json:"bounds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/experimental.pb.go b/vendor/google.golang.org/genproto/googleapis/api/experimental.pb.go index cef0fa102..4b34111b8 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/experimental.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/experimental.pb.go @@ -23,7 +23,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // only be used by whitelisted users. type Experimental struct { // Authorization configuration. - Authorization *AuthorizationConfig `protobuf:"bytes,8,opt,name=authorization" json:"authorization,omitempty"` + Authorization *AuthorizationConfig `protobuf:"bytes,8,opt,name=authorization,proto3" json:"authorization,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -33,7 +33,7 @@ func (m *Experimental) Reset() { *m = Experimental{} } func (m *Experimental) String() string { return proto.CompactTextString(m) } func (*Experimental) ProtoMessage() {} func (*Experimental) Descriptor() ([]byte, []int) { - return fileDescriptor_experimental_8d531d4162396b31, []int{0} + return fileDescriptor_experimental_bcdc809cb0848cf0, []int{0} } func (m *Experimental) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Experimental.Unmarshal(m, b) @@ -65,10 +65,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/experimental/experimental.proto", fileDescriptor_experimental_8d531d4162396b31) + proto.RegisterFile("google/api/experimental/experimental.proto", fileDescriptor_experimental_bcdc809cb0848cf0) } -var fileDescriptor_experimental_8d531d4162396b31 = []byte{ +var fileDescriptor_experimental_bcdc809cb0848cf0 = []byte{ // 204 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0xd2, 0x4a, 0xcf, 0xcf, 0x4f, 0xcf, 0x49, 0xd5, 0x4f, 0x2c, 0xc8, 0xd4, 0x4f, 0xad, 0x28, 0x48, 0x2d, 0xca, 0xcc, 0x4d, 0xcd, diff --git a/vendor/google.golang.org/genproto/googleapis/api/httpbody/httpbody.pb.go b/vendor/google.golang.org/genproto/googleapis/api/httpbody/httpbody.pb.go index ff0c92a30..c986cddfa 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/httpbody/httpbody.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/httpbody/httpbody.pb.go @@ -6,6 +6,7 @@ package httpbody // import "google.golang.org/genproto/googleapis/api/httpbody" import proto "github.com/golang/protobuf/proto" import fmt "fmt" import math "math" +import any "github.com/golang/protobuf/ptypes/any" // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal @@ -58,19 +59,22 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // handled, all other features will continue to work unchanged. type HttpBody struct { // The HTTP Content-Type string representing the content type of the body. - ContentType string `protobuf:"bytes,1,opt,name=content_type,json=contentType" json:"content_type,omitempty"` + ContentType string `protobuf:"bytes,1,opt,name=content_type,json=contentType,proto3" json:"content_type,omitempty"` // HTTP body binary data. - Data []byte `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Data []byte `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` + // Application specific response metadata. Must be set in the first response + // for streaming APIs. + Extensions []*any.Any `protobuf:"bytes,3,rep,name=extensions,proto3" json:"extensions,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *HttpBody) Reset() { *m = HttpBody{} } func (m *HttpBody) String() string { return proto.CompactTextString(m) } func (*HttpBody) ProtoMessage() {} func (*HttpBody) Descriptor() ([]byte, []int) { - return fileDescriptor_httpbody_555eceaa83d53411, []int{0} + return fileDescriptor_httpbody_24933373b9bfd305, []int{0} } func (m *HttpBody) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_HttpBody.Unmarshal(m, b) @@ -104,24 +108,34 @@ func (m *HttpBody) GetData() []byte { return nil } +func (m *HttpBody) GetExtensions() []*any.Any { + if m != nil { + return m.Extensions + } + return nil +} + func init() { proto.RegisterType((*HttpBody)(nil), "google.api.HttpBody") } -func init() { proto.RegisterFile("google/api/httpbody.proto", fileDescriptor_httpbody_555eceaa83d53411) } +func init() { proto.RegisterFile("google/api/httpbody.proto", fileDescriptor_httpbody_24933373b9bfd305) } -var fileDescriptor_httpbody_555eceaa83d53411 = []byte{ - // 181 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x4c, 0xcf, 0xcf, 0x4f, - 0xcf, 0x49, 0xd5, 0x4f, 0x2c, 0xc8, 0xd4, 0xcf, 0x28, 0x29, 0x29, 0x48, 0xca, 0x4f, 0xa9, 0xd4, - 0x2b, 0x28, 0xca, 0x2f, 0xc9, 0x17, 0xe2, 0x82, 0x48, 0xe9, 0x25, 0x16, 0x64, 0x2a, 0x39, 0x72, - 0x71, 0x78, 0x94, 0x94, 0x14, 0x38, 0xe5, 0xa7, 0x54, 0x0a, 0x29, 0x72, 0xf1, 0x24, 0xe7, 0xe7, - 0x95, 0xa4, 0xe6, 0x95, 0xc4, 0x97, 0x54, 0x16, 0xa4, 0x4a, 0x30, 0x2a, 0x30, 0x6a, 0x70, 0x06, - 0x71, 0x43, 0xc5, 0x42, 0x2a, 0x0b, 0x52, 0x85, 0x84, 0xb8, 0x58, 0x52, 0x12, 0x4b, 0x12, 0x25, - 0x98, 0x14, 0x18, 0x35, 0x78, 0x82, 0xc0, 0x6c, 0xa7, 0x54, 0x2e, 0xbe, 0xe4, 0xfc, 0x5c, 0x3d, - 0x84, 0xa1, 0x4e, 0xbc, 0x30, 0x23, 0x03, 0x40, 0xf6, 0x05, 0x30, 0x46, 0x59, 0x43, 0x25, 0xd3, - 0xf3, 0x73, 0x12, 0xf3, 0xd2, 0xf5, 0xf2, 0x8b, 0xd2, 0xf5, 0xd3, 0x53, 0xf3, 0xc0, 0xae, 0xd1, - 0x87, 0x48, 0x25, 0x16, 0x64, 0x16, 0xa3, 0xb8, 0xd5, 0x1a, 0xc6, 0x58, 0xc4, 0xc4, 0xe2, 0xee, - 0x18, 0xe0, 0x99, 0xc4, 0x06, 0x56, 0x6e, 0x0c, 0x08, 0x00, 0x00, 0xff, 0xff, 0x0e, 0x3a, 0xdf, - 0x30, 0xd9, 0x00, 0x00, 0x00, +var fileDescriptor_httpbody_24933373b9bfd305 = []byte{ + // 226 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x8f, 0x3f, 0x4f, 0xc3, 0x30, + 0x10, 0xc5, 0x95, 0xb6, 0x42, 0x70, 0x2d, 0x0c, 0x16, 0x43, 0x60, 0x0a, 0x4c, 0x99, 0x6c, 0x09, + 0xd8, 0x3a, 0x35, 0x0b, 0xb0, 0x45, 0x11, 0x13, 0x0b, 0x72, 0x9a, 0xc3, 0x44, 0x2a, 0x77, 0xa7, + 0xe6, 0x10, 0xf8, 0xeb, 0xf0, 0x49, 0x11, 0xf9, 0x23, 0xe8, 0xf6, 0xe4, 0xdf, 0xcf, 0x7e, 0xcf, + 0x70, 0x11, 0x98, 0xc3, 0x0e, 0x9d, 0x97, 0xd6, 0xbd, 0xa9, 0x4a, 0xcd, 0x4d, 0xb4, 0xb2, 0x67, + 0x65, 0x03, 0x03, 0xb2, 0x5e, 0xda, 0xcb, 0x49, 0xeb, 0x49, 0xfd, 0xf1, 0xea, 0x3c, 0x8d, 0xda, + 0xf5, 0x27, 0x1c, 0x3f, 0xa8, 0x4a, 0xc1, 0x4d, 0x34, 0x57, 0xb0, 0xda, 0x32, 0x29, 0x92, 0xbe, + 0x68, 0x14, 0x4c, 0x93, 0x2c, 0xc9, 0x4f, 0xaa, 0xe5, 0x78, 0xf6, 0x14, 0x05, 0x8d, 0x81, 0x45, + 0xe3, 0xd5, 0xa7, 0xb3, 0x2c, 0xc9, 0x57, 0x55, 0x9f, 0xcd, 0x1d, 0x00, 0x7e, 0x29, 0x52, 0xd7, + 0x32, 0x75, 0xe9, 0x3c, 0x9b, 0xe7, 0xcb, 0x9b, 0x73, 0x3b, 0xd6, 0x4f, 0x95, 0x76, 0x43, 0xb1, + 0xfa, 0xe7, 0x15, 0x08, 0x67, 0x5b, 0x7e, 0xb7, 0x7f, 0x2b, 0x8b, 0xd3, 0x69, 0x48, 0xf9, 0x7b, + 0xa7, 0x4c, 0x9e, 0xd7, 0x23, 0x0c, 0xbc, 0xf3, 0x14, 0x2c, 0xef, 0x83, 0x0b, 0x48, 0xfd, 0x8b, + 0x6e, 0x40, 0x5e, 0xda, 0xee, 0xe0, 0xf3, 0xeb, 0x29, 0x7c, 0xcf, 0x16, 0xf7, 0x9b, 0xf2, 0xb1, + 0x3e, 0xea, 0xf5, 0xdb, 0x9f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x3c, 0x11, 0x70, 0x02, 0x2a, 0x01, + 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/api/label/label.pb.go b/vendor/google.golang.org/genproto/googleapis/api/label/label.pb.go index 33c652f6f..ef1af9cb3 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/label/label.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/label/label.pb.go @@ -45,17 +45,17 @@ func (x LabelDescriptor_ValueType) String() string { return proto.EnumName(LabelDescriptor_ValueType_name, int32(x)) } func (LabelDescriptor_ValueType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_label_1d347e5545d4fc31, []int{0, 0} + return fileDescriptor_label_ae588de548a3b6a9, []int{0, 0} } // A description of a label. type LabelDescriptor struct { // The label key. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // The type of data that can be assigned to the label. - ValueType LabelDescriptor_ValueType `protobuf:"varint,2,opt,name=value_type,json=valueType,enum=google.api.LabelDescriptor_ValueType" json:"value_type,omitempty"` + ValueType LabelDescriptor_ValueType `protobuf:"varint,2,opt,name=value_type,json=valueType,proto3,enum=google.api.LabelDescriptor_ValueType" json:"value_type,omitempty"` // A human-readable description for the label. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -65,7 +65,7 @@ func (m *LabelDescriptor) Reset() { *m = LabelDescriptor{} } func (m *LabelDescriptor) String() string { return proto.CompactTextString(m) } func (*LabelDescriptor) ProtoMessage() {} func (*LabelDescriptor) Descriptor() ([]byte, []int) { - return fileDescriptor_label_1d347e5545d4fc31, []int{0} + return fileDescriptor_label_ae588de548a3b6a9, []int{0} } func (m *LabelDescriptor) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LabelDescriptor.Unmarshal(m, b) @@ -111,9 +111,9 @@ func init() { proto.RegisterEnum("google.api.LabelDescriptor_ValueType", LabelDescriptor_ValueType_name, LabelDescriptor_ValueType_value) } -func init() { proto.RegisterFile("google/api/label.proto", fileDescriptor_label_1d347e5545d4fc31) } +func init() { proto.RegisterFile("google/api/label.proto", fileDescriptor_label_ae588de548a3b6a9) } -var fileDescriptor_label_1d347e5545d4fc31 = []byte{ +var fileDescriptor_label_ae588de548a3b6a9 = []byte{ // 252 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x12, 0x4b, 0xcf, 0xcf, 0x4f, 0xcf, 0x49, 0xd5, 0x4f, 0x2c, 0xc8, 0xd4, 0xcf, 0x49, 0x4c, 0x4a, 0xcd, 0xd1, 0x2b, 0x28, 0xca, diff --git a/vendor/google.golang.org/genproto/googleapis/api/metric/metric.pb.go b/vendor/google.golang.org/genproto/googleapis/api/metric/metric.pb.go index 775f179d3..54f65179a 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/metric/metric.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/metric/metric.pb.go @@ -54,7 +54,7 @@ func (x MetricDescriptor_MetricKind) String() string { return proto.EnumName(MetricDescriptor_MetricKind_name, int32(x)) } func (MetricDescriptor_MetricKind) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_metric_14b8b448d25edd31, []int{0, 0} + return fileDescriptor_metric_d2ae7c695bda16a9, []int{0, 0} } // The value type of a metric. @@ -102,22 +102,15 @@ func (x MetricDescriptor_ValueType) String() string { return proto.EnumName(MetricDescriptor_ValueType_name, int32(x)) } func (MetricDescriptor_ValueType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_metric_14b8b448d25edd31, []int{0, 1} + return fileDescriptor_metric_d2ae7c695bda16a9, []int{0, 1} } // Defines a metric type and its schema. Once a metric descriptor is created, // deleting or altering it stops data collection and makes the metric type's // existing data unusable. type MetricDescriptor struct { - // The resource name of the metric descriptor. Depending on the - // implementation, the name typically includes: (1) the parent resource name - // that defines the scope of the metric type or of its data; and (2) the - // metric's URL-encoded type, which also appears in the `type` field of this - // descriptor. For example, following is the resource name of a custom - // metric within the GCP project `my-project-id`: - // - // "projects/my-project-id/metricDescriptors/custom.googleapis.com%2Finvoice%2Fpaid%2Famount" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + // The resource name of the metric descriptor. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The metric type, including its DNS name prefix. The type is not // URL-encoded. All user-defined custom metric types have the DNS name // `custom.googleapis.com`. Metric types should use a natural hierarchical @@ -125,20 +118,20 @@ type MetricDescriptor struct { // // "custom.googleapis.com/invoice/paid/amount" // "appengine.googleapis.com/http/server/response_latencies" - Type string `protobuf:"bytes,8,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,8,opt,name=type,proto3" json:"type,omitempty"` // The set of labels that can be used to describe a specific // instance of this metric type. For example, the // `appengine.googleapis.com/http/server/response_latencies` metric // type has a label for the HTTP response code, `response_code`, so // you can look at latencies for successful responses or just // for responses that failed. - Labels []*label.LabelDescriptor `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty"` + Labels []*label.LabelDescriptor `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty"` // Whether the metric records instantaneous values, changes to a value, etc. // Some combinations of `metric_kind` and `value_type` might not be supported. - MetricKind MetricDescriptor_MetricKind `protobuf:"varint,3,opt,name=metric_kind,json=metricKind,enum=google.api.MetricDescriptor_MetricKind" json:"metric_kind,omitempty"` + MetricKind MetricDescriptor_MetricKind `protobuf:"varint,3,opt,name=metric_kind,json=metricKind,proto3,enum=google.api.MetricDescriptor_MetricKind" json:"metric_kind,omitempty"` // Whether the measurement is an integer, a floating-point number, etc. // Some combinations of `metric_kind` and `value_type` might not be supported. - ValueType MetricDescriptor_ValueType `protobuf:"varint,4,opt,name=value_type,json=valueType,enum=google.api.MetricDescriptor_ValueType" json:"value_type,omitempty"` + ValueType MetricDescriptor_ValueType `protobuf:"varint,4,opt,name=value_type,json=valueType,proto3,enum=google.api.MetricDescriptor_ValueType" json:"value_type,omitempty"` // The unit in which the metric value is reported. It is only applicable // if the `value_type` is `INT64`, `DOUBLE`, or `DISTRIBUTION`. The // supported units are a subset of [The Unified Code for Units of @@ -178,8 +171,6 @@ type MetricDescriptor struct { // // **Grammar** // - // The grammar includes the dimensionless unit `1`, such as `1/s`. - // // The grammar also includes these connectors: // // * `/` division (as an infix operator, e.g. `1/s`). @@ -189,7 +180,7 @@ type MetricDescriptor struct { // // Expression = Component { "." Component } { "/" Component } ; // - // Component = [ PREFIX ] UNIT [ Annotation ] + // Component = ( [ PREFIX ] UNIT | "%" ) [ Annotation ] // | Annotation // | "1" // ; @@ -203,12 +194,17 @@ type MetricDescriptor struct { // `{requests}/s == 1/s`, `By{transmitted}/s == By/s`. // * `NAME` is a sequence of non-blank printable ASCII characters not // containing '{' or '}'. - Unit string `protobuf:"bytes,5,opt,name=unit" json:"unit,omitempty"` + // * `1` represents dimensionless value 1, such as in `1/s`. + // * `%` represents dimensionless value 1/100, and annotates values giving + // a percentage. + Unit string `protobuf:"bytes,5,opt,name=unit,proto3" json:"unit,omitempty"` // A detailed description of the metric, which can be used in documentation. - Description string `protobuf:"bytes,6,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,6,opt,name=description,proto3" json:"description,omitempty"` // A concise name for the metric, which can be displayed in user interfaces. // Use sentence case without an ending period, for example "Request count". - DisplayName string `protobuf:"bytes,7,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + // This field is optional but it is recommended to be set for any metrics + // associated with user-visible concepts, such as Quota. + DisplayName string `protobuf:"bytes,7,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -218,7 +214,7 @@ func (m *MetricDescriptor) Reset() { *m = MetricDescriptor{} } func (m *MetricDescriptor) String() string { return proto.CompactTextString(m) } func (*MetricDescriptor) ProtoMessage() {} func (*MetricDescriptor) Descriptor() ([]byte, []int) { - return fileDescriptor_metric_14b8b448d25edd31, []int{0} + return fileDescriptor_metric_d2ae7c695bda16a9, []int{0} } func (m *MetricDescriptor) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MetricDescriptor.Unmarshal(m, b) @@ -299,10 +295,10 @@ func (m *MetricDescriptor) GetDisplayName() string { type Metric struct { // An existing metric type, see [google.api.MetricDescriptor][google.api.MetricDescriptor]. // For example, `custom.googleapis.com/invoice/paid/amount`. - Type string `protobuf:"bytes,3,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,3,opt,name=type,proto3" json:"type,omitempty"` // The set of label values that uniquely identify this metric. All // labels listed in the `MetricDescriptor` must be assigned values. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -312,7 +308,7 @@ func (m *Metric) Reset() { *m = Metric{} } func (m *Metric) String() string { return proto.CompactTextString(m) } func (*Metric) ProtoMessage() {} func (*Metric) Descriptor() ([]byte, []int) { - return fileDescriptor_metric_14b8b448d25edd31, []int{1} + return fileDescriptor_metric_d2ae7c695bda16a9, []int{1} } func (m *Metric) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Metric.Unmarshal(m, b) @@ -354,9 +350,9 @@ func init() { proto.RegisterEnum("google.api.MetricDescriptor_ValueType", MetricDescriptor_ValueType_name, MetricDescriptor_ValueType_value) } -func init() { proto.RegisterFile("google/api/metric.proto", fileDescriptor_metric_14b8b448d25edd31) } +func init() { proto.RegisterFile("google/api/metric.proto", fileDescriptor_metric_d2ae7c695bda16a9) } -var fileDescriptor_metric_14b8b448d25edd31 = []byte{ +var fileDescriptor_metric_d2ae7c695bda16a9 = []byte{ // 506 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x7c, 0x53, 0x4d, 0x6f, 0xda, 0x40, 0x10, 0xad, 0x3f, 0x70, 0xc3, 0x10, 0xa1, 0xd5, 0xaa, 0x4a, 0x2c, 0x22, 0x55, 0x94, 0x43, 0xcb, diff --git a/vendor/google.golang.org/genproto/googleapis/api/monitoredres/monitored_resource.pb.go b/vendor/google.golang.org/genproto/googleapis/api/monitoredres/monitored_resource.pb.go index bc6ec8fc7..a9a8bb7cf 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/monitoredres/monitored_resource.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/monitoredres/monitored_resource.pb.go @@ -6,6 +6,7 @@ package monitoredres // import "google.golang.org/genproto/googleapis/api/monito import proto "github.com/golang/protobuf/proto" import fmt "fmt" import math "math" +import _struct "github.com/golang/protobuf/ptypes/struct" import label "google.golang.org/genproto/googleapis/api/label" // Reference imports to suppress errors if they are not otherwise used. @@ -35,23 +36,23 @@ type MonitoredResourceDescriptor struct { // {project_id} is a project ID that provides API-specific context for // accessing the type. APIs that do not use project information can use the // resource name format `"monitoredResourceDescriptors/{type}"`. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` // Required. The monitored resource type. For example, the type // `"cloudsql_database"` represents databases in Google Cloud SQL. // The maximum length of this value is 256 characters. - Type string `protobuf:"bytes,1,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` // Optional. A concise name for the monitored resource type that might be // displayed in user interfaces. It should be a Title Cased Noun Phrase, // without any article or other determiners. For example, // `"Google Cloud SQL Database"`. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Optional. A detailed description of the monitored resource type that might // be used in documentation. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Required. A set of labels used to describe instances of this monitored // resource type. For example, an individual Google Cloud SQL database is // identified by values for the labels `"database_id"` and `"zone"`. - Labels []*label.LabelDescriptor `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty"` + Labels []*label.LabelDescriptor `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -61,7 +62,7 @@ func (m *MonitoredResourceDescriptor) Reset() { *m = MonitoredResourceDe func (m *MonitoredResourceDescriptor) String() string { return proto.CompactTextString(m) } func (*MonitoredResourceDescriptor) ProtoMessage() {} func (*MonitoredResourceDescriptor) Descriptor() ([]byte, []int) { - return fileDescriptor_monitored_resource_c7b4d8e4dff11cf8, []int{0} + return fileDescriptor_monitored_resource_cdc52d43c1ee23b3, []int{0} } func (m *MonitoredResourceDescriptor) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MonitoredResourceDescriptor.Unmarshal(m, b) @@ -132,12 +133,12 @@ func (m *MonitoredResourceDescriptor) GetLabels() []*label.LabelDescriptor { type MonitoredResource struct { // Required. The monitored resource type. This field must match // the `type` field of a [MonitoredResourceDescriptor][google.api.MonitoredResourceDescriptor] object. For - // example, the type of a Cloud SQL database is `"cloudsql_database"`. - Type string `protobuf:"bytes,1,opt,name=type" json:"type,omitempty"` + // example, the type of a Compute Engine VM instance is `gce_instance`. + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` // Required. Values for all of the labels listed in the associated monitored - // resource descriptor. For example, Cloud SQL databases use the labels - // `"database_id"` and `"zone"`. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + // resource descriptor. For example, Compute Engine VM instances use the + // labels `"project_id"`, `"instance_id"`, and `"zone"`. + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -147,7 +148,7 @@ func (m *MonitoredResource) Reset() { *m = MonitoredResource{} } func (m *MonitoredResource) String() string { return proto.CompactTextString(m) } func (*MonitoredResource) ProtoMessage() {} func (*MonitoredResource) Descriptor() ([]byte, []int) { - return fileDescriptor_monitored_resource_c7b4d8e4dff11cf8, []int{1} + return fileDescriptor_monitored_resource_cdc52d43c1ee23b3, []int{1} } func (m *MonitoredResource) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MonitoredResource.Unmarshal(m, b) @@ -181,37 +182,108 @@ func (m *MonitoredResource) GetLabels() map[string]string { return nil } +// Auxiliary metadata for a [MonitoredResource][google.api.MonitoredResource] object. +// [MonitoredResource][google.api.MonitoredResource] objects contain the minimum set of information to +// uniquely identify a monitored resource instance. There is some other useful +// auxiliary metadata. Google Stackdriver Monitoring & Logging uses an ingestion +// pipeline to extract metadata for cloud resources of all types , and stores +// the metadata in this message. +type MonitoredResourceMetadata struct { + // Output only. Values for predefined system metadata labels. + // System labels are a kind of metadata extracted by Google Stackdriver. + // Stackdriver determines what system labels are useful and how to obtain + // their values. Some examples: "machine_image", "vpc", "subnet_id", + // "security_group", "name", etc. + // System label values can be only strings, Boolean values, or a list of + // strings. For example: + // + // { "name": "my-test-instance", + // "security_group": ["a", "b", "c"], + // "spot_instance": false } + SystemLabels *_struct.Struct `protobuf:"bytes,1,opt,name=system_labels,json=systemLabels,proto3" json:"system_labels,omitempty"` + // Output only. A map of user-defined metadata labels. + UserLabels map[string]string `protobuf:"bytes,2,rep,name=user_labels,json=userLabels,proto3" json:"user_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *MonitoredResourceMetadata) Reset() { *m = MonitoredResourceMetadata{} } +func (m *MonitoredResourceMetadata) String() string { return proto.CompactTextString(m) } +func (*MonitoredResourceMetadata) ProtoMessage() {} +func (*MonitoredResourceMetadata) Descriptor() ([]byte, []int) { + return fileDescriptor_monitored_resource_cdc52d43c1ee23b3, []int{2} +} +func (m *MonitoredResourceMetadata) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_MonitoredResourceMetadata.Unmarshal(m, b) +} +func (m *MonitoredResourceMetadata) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_MonitoredResourceMetadata.Marshal(b, m, deterministic) +} +func (dst *MonitoredResourceMetadata) XXX_Merge(src proto.Message) { + xxx_messageInfo_MonitoredResourceMetadata.Merge(dst, src) +} +func (m *MonitoredResourceMetadata) XXX_Size() int { + return xxx_messageInfo_MonitoredResourceMetadata.Size(m) +} +func (m *MonitoredResourceMetadata) XXX_DiscardUnknown() { + xxx_messageInfo_MonitoredResourceMetadata.DiscardUnknown(m) +} + +var xxx_messageInfo_MonitoredResourceMetadata proto.InternalMessageInfo + +func (m *MonitoredResourceMetadata) GetSystemLabels() *_struct.Struct { + if m != nil { + return m.SystemLabels + } + return nil +} + +func (m *MonitoredResourceMetadata) GetUserLabels() map[string]string { + if m != nil { + return m.UserLabels + } + return nil +} + func init() { proto.RegisterType((*MonitoredResourceDescriptor)(nil), "google.api.MonitoredResourceDescriptor") proto.RegisterType((*MonitoredResource)(nil), "google.api.MonitoredResource") proto.RegisterMapType((map[string]string)(nil), "google.api.MonitoredResource.LabelsEntry") + proto.RegisterType((*MonitoredResourceMetadata)(nil), "google.api.MonitoredResourceMetadata") + proto.RegisterMapType((map[string]string)(nil), "google.api.MonitoredResourceMetadata.UserLabelsEntry") } func init() { - proto.RegisterFile("google/api/monitored_resource.proto", fileDescriptor_monitored_resource_c7b4d8e4dff11cf8) + proto.RegisterFile("google/api/monitored_resource.proto", fileDescriptor_monitored_resource_cdc52d43c1ee23b3) } -var fileDescriptor_monitored_resource_c7b4d8e4dff11cf8 = []byte{ - // 321 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x6c, 0x51, 0x4b, 0x4b, 0x3b, 0x31, - 0x10, 0x27, 0xdb, 0x07, 0xfc, 0x67, 0xff, 0x88, 0x06, 0x29, 0x4b, 0x7b, 0xa9, 0xf5, 0x52, 0x2f, - 0xbb, 0x60, 0x2f, 0x3e, 0x4e, 0xad, 0x8a, 0x08, 0x2a, 0xa5, 0x47, 0x2f, 0x25, 0x6d, 0xc3, 0x12, - 0xdc, 0x66, 0x42, 0xb2, 0x15, 0xf6, 0xeb, 0x08, 0x7e, 0x0e, 0xbf, 0x96, 0x47, 0xc9, 0xa3, 0x76, - 0xa5, 0xde, 0x26, 0xbf, 0xf9, 0x3d, 0x66, 0x32, 0x70, 0x9a, 0x23, 0xe6, 0x05, 0xcf, 0x98, 0x12, - 0xd9, 0x1a, 0xa5, 0x28, 0x51, 0xf3, 0xd5, 0x5c, 0x73, 0x83, 0x1b, 0xbd, 0xe4, 0xa9, 0xd2, 0x58, - 0x22, 0x05, 0x4f, 0x4a, 0x99, 0x12, 0xdd, 0x4e, 0x4d, 0x50, 0xb0, 0x05, 0x2f, 0x3c, 0x67, 0xf0, - 0x49, 0xa0, 0xf7, 0xb4, 0x35, 0x98, 0x05, 0xfd, 0x2d, 0x37, 0x4b, 0x2d, 0x54, 0x89, 0x9a, 0x52, - 0x68, 0x4a, 0xb6, 0xe6, 0x49, 0xab, 0x4f, 0x86, 0xff, 0x66, 0xae, 0xb6, 0x58, 0x59, 0x29, 0x9e, - 0x10, 0x8f, 0xd9, 0x9a, 0x9e, 0xc0, 0xff, 0x95, 0x30, 0xaa, 0x60, 0xd5, 0xdc, 0xf1, 0x23, 0xd7, - 0x8b, 0x03, 0xf6, 0x6c, 0x65, 0x7d, 0x88, 0x57, 0xc1, 0x58, 0xa0, 0x4c, 0x1a, 0x81, 0xb1, 0x83, - 0xe8, 0x08, 0xda, 0x6e, 0x36, 0x93, 0x34, 0xfb, 0x8d, 0x61, 0x7c, 0xde, 0x4b, 0x77, 0x1b, 0xa4, - 0x8f, 0xb6, 0xb3, 0x9b, 0x6c, 0x16, 0xa8, 0x83, 0x0f, 0x02, 0x47, 0x7b, 0x1b, 0xfc, 0x39, 0xe3, - 0xf8, 0xc7, 0x3e, 0x72, 0xf6, 0x67, 0x75, 0xfb, 0x3d, 0x0b, 0x1f, 0x68, 0xee, 0x64, 0xa9, 0xab, - 0x6d, 0x58, 0xf7, 0x12, 0xe2, 0x1a, 0x4c, 0x0f, 0xa1, 0xf1, 0xca, 0xab, 0x10, 0x62, 0x4b, 0x7a, - 0x0c, 0xad, 0x37, 0x56, 0x6c, 0xb6, 0x1f, 0xe0, 0x1f, 0x57, 0xd1, 0x05, 0x99, 0x54, 0x70, 0xb0, - 0xc4, 0x75, 0x2d, 0x72, 0xd2, 0xd9, 0xcb, 0x9c, 0xda, 0x9b, 0x4c, 0xc9, 0xcb, 0x4d, 0x60, 0xe5, - 0x58, 0x30, 0x99, 0xa7, 0xa8, 0xf3, 0x2c, 0xe7, 0xd2, 0x5d, 0x2c, 0xf3, 0x2d, 0xa6, 0x84, 0xf9, - 0x7d, 0x7d, 0xcd, 0xcd, 0x75, 0xfd, 0xf1, 0x45, 0xc8, 0x7b, 0xd4, 0xbc, 0x1f, 0x4f, 0x1f, 0x16, - 0x6d, 0xa7, 0x1c, 0x7d, 0x07, 0x00, 0x00, 0xff, 0xff, 0xf8, 0xfb, 0xfb, 0x11, 0x36, 0x02, 0x00, - 0x00, +var fileDescriptor_monitored_resource_cdc52d43c1ee23b3 = []byte{ + // 415 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x53, 0x4d, 0xab, 0xd3, 0x40, + 0x14, 0x65, 0xd2, 0x0f, 0xf0, 0xa6, 0x7e, 0x0d, 0x52, 0x63, 0xea, 0xa2, 0xd6, 0x4d, 0xdd, 0x24, + 0xd0, 0x22, 0xf8, 0xb9, 0x68, 0x55, 0x44, 0xb0, 0x52, 0x22, 0xba, 0x70, 0x13, 0xa6, 0xc9, 0x18, + 0x82, 0x49, 0x26, 0xcc, 0x4c, 0x84, 0xfc, 0x1d, 0xc1, 0xdf, 0xe1, 0x5f, 0x72, 0xe9, 0x52, 0x32, + 0x33, 0x69, 0xd3, 0x97, 0xc7, 0x83, 0xb7, 0xbb, 0xf7, 0xdc, 0x73, 0xcf, 0x3d, 0x27, 0x43, 0xe0, + 0x71, 0xc2, 0x58, 0x92, 0x51, 0x9f, 0x94, 0xa9, 0x9f, 0xb3, 0x22, 0x95, 0x8c, 0xd3, 0x38, 0xe4, + 0x54, 0xb0, 0x8a, 0x47, 0xd4, 0x2b, 0x39, 0x93, 0x0c, 0x83, 0x26, 0x79, 0xa4, 0x4c, 0xdd, 0x69, + 0x67, 0x21, 0x23, 0x07, 0x9a, 0x69, 0x8e, 0xfb, 0xd0, 0xe0, 0xaa, 0x3b, 0x54, 0xdf, 0x7d, 0x21, + 0x79, 0x15, 0x49, 0x3d, 0x5d, 0xfc, 0x41, 0x30, 0xdb, 0xb5, 0xf2, 0x81, 0x51, 0x7f, 0x4b, 0x45, + 0xc4, 0xd3, 0x52, 0x32, 0x8e, 0x31, 0x0c, 0x0b, 0x92, 0x53, 0x67, 0x34, 0x47, 0xcb, 0x1b, 0x81, + 0xaa, 0x1b, 0x4c, 0xd6, 0x25, 0x75, 0x90, 0xc6, 0x9a, 0x1a, 0x3f, 0x82, 0x49, 0x9c, 0x8a, 0x32, + 0x23, 0x75, 0xa8, 0xf8, 0x96, 0x9a, 0xd9, 0x06, 0xfb, 0xd4, 0xac, 0xcd, 0xc1, 0x8e, 0x8d, 0x70, + 0xca, 0x0a, 0x67, 0x60, 0x18, 0x27, 0x08, 0xaf, 0x61, 0xac, 0x9c, 0x0b, 0x67, 0x38, 0x1f, 0x2c, + 0xed, 0xd5, 0xcc, 0x3b, 0xe5, 0xf3, 0x3e, 0x36, 0x93, 0x93, 0xb3, 0xc0, 0x50, 0x17, 0xbf, 0x11, + 0xdc, 0xed, 0x25, 0xb8, 0xd4, 0xe3, 0xe6, 0x28, 0x6f, 0x29, 0xf9, 0x27, 0x5d, 0xf9, 0x9e, 0x84, + 0x3e, 0x28, 0xde, 0x15, 0x92, 0xd7, 0xed, 0x31, 0xf7, 0x39, 0xd8, 0x1d, 0x18, 0xdf, 0x81, 0xc1, + 0x0f, 0x5a, 0x9b, 0x23, 0x4d, 0x89, 0xef, 0xc1, 0xe8, 0x27, 0xc9, 0xaa, 0xf6, 0x03, 0xe8, 0xe6, + 0x85, 0xf5, 0x0c, 0x2d, 0xfe, 0x22, 0x78, 0xd0, 0x3b, 0xb2, 0xa3, 0x92, 0xc4, 0x44, 0x12, 0xfc, + 0x0a, 0x6e, 0x8a, 0x5a, 0x48, 0x9a, 0x87, 0xc6, 0x62, 0xa3, 0x69, 0xaf, 0xee, 0xb7, 0x16, 0xdb, + 0xd7, 0xf3, 0x3e, 0xab, 0xd7, 0x0b, 0x26, 0x9a, 0xad, 0xcd, 0xe0, 0xaf, 0x60, 0x57, 0x82, 0xf2, + 0xf0, 0x2c, 0xde, 0xd3, 0x2b, 0xe3, 0xb5, 0x97, 0xbd, 0x2f, 0x82, 0xf2, 0x6e, 0x54, 0xa8, 0x8e, + 0x80, 0xfb, 0x1a, 0x6e, 0x5f, 0x18, 0x5f, 0x27, 0xf2, 0xb6, 0x86, 0x5b, 0x11, 0xcb, 0x3b, 0x36, + 0xb6, 0xd3, 0x9e, 0x8f, 0x7d, 0x13, 0x6c, 0x8f, 0xbe, 0xbd, 0x31, 0xac, 0x84, 0x65, 0xa4, 0x48, + 0x3c, 0xc6, 0x13, 0x3f, 0xa1, 0x85, 0x8a, 0xed, 0xeb, 0x11, 0x29, 0x53, 0x71, 0xfe, 0x3b, 0x70, + 0x2a, 0x5e, 0x76, 0x9b, 0x7f, 0x08, 0xfd, 0xb2, 0x86, 0xef, 0x37, 0xfb, 0x0f, 0x87, 0xb1, 0xda, + 0x5c, 0xff, 0x0f, 0x00, 0x00, 0xff, 0xff, 0x10, 0x16, 0x7c, 0xe9, 0x47, 0x03, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/auth.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/auth.pb.go index b38fae337..bba004a40 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/auth.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/auth.pb.go @@ -37,9 +37,9 @@ type Authentication struct { // A list of authentication rules that apply to individual API methods. // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*AuthenticationRule `protobuf:"bytes,3,rep,name=rules" json:"rules,omitempty"` + Rules []*AuthenticationRule `protobuf:"bytes,3,rep,name=rules,proto3" json:"rules,omitempty"` // Defines a set of authentication providers that a service supports. - Providers []*AuthProvider `protobuf:"bytes,4,rep,name=providers" json:"providers,omitempty"` + Providers []*AuthProvider `protobuf:"bytes,4,rep,name=providers,proto3" json:"providers,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -49,7 +49,7 @@ func (m *Authentication) Reset() { *m = Authentication{} } func (m *Authentication) String() string { return proto.CompactTextString(m) } func (*Authentication) ProtoMessage() {} func (*Authentication) Descriptor() ([]byte, []int) { - return fileDescriptor_auth_6939d4e8a8cd2956, []int{0} + return fileDescriptor_auth_52a59fcac3533a16, []int{0} } func (m *Authentication) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Authentication.Unmarshal(m, b) @@ -96,19 +96,13 @@ type AuthenticationRule struct { // Selects the methods to which this rule applies. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // The requirements for OAuth credentials. - Oauth *OAuthRequirements `protobuf:"bytes,2,opt,name=oauth" json:"oauth,omitempty"` - // Whether to allow requests without a credential. The credential can be - // an OAuth token, Google cookies (first-party auth) or EndUserCreds. - // - // For requests without credentials, if the service control environment is - // specified, each incoming request **must** be associated with a service - // consumer. This can be done by passing an API key that belongs to a consumer - // project. - AllowWithoutCredential bool `protobuf:"varint,5,opt,name=allow_without_credential,json=allowWithoutCredential" json:"allow_without_credential,omitempty"` + Oauth *OAuthRequirements `protobuf:"bytes,2,opt,name=oauth,proto3" json:"oauth,omitempty"` + // If true, the service accepts API keys without any other credential. + AllowWithoutCredential bool `protobuf:"varint,5,opt,name=allow_without_credential,json=allowWithoutCredential,proto3" json:"allow_without_credential,omitempty"` // Requirements for additional authentication providers. - Requirements []*AuthRequirement `protobuf:"bytes,7,rep,name=requirements" json:"requirements,omitempty"` + Requirements []*AuthRequirement `protobuf:"bytes,7,rep,name=requirements,proto3" json:"requirements,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -118,7 +112,7 @@ func (m *AuthenticationRule) Reset() { *m = AuthenticationRule{} } func (m *AuthenticationRule) String() string { return proto.CompactTextString(m) } func (*AuthenticationRule) ProtoMessage() {} func (*AuthenticationRule) Descriptor() ([]byte, []int) { - return fileDescriptor_auth_6939d4e8a8cd2956, []int{1} + return fileDescriptor_auth_52a59fcac3533a16, []int{1} } func (m *AuthenticationRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AuthenticationRule.Unmarshal(m, b) @@ -173,14 +167,14 @@ type AuthProvider struct { // `AuthRequirement.provider_id`. // // Example: "bookstore_auth". - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // Identifies the principal that issued the JWT. See // https://tools.ietf.org/html/draft-ietf-oauth-json-web-token-32#section-4.1.1 // Usually a URL or an email address. // // Example: https://securetoken.google.com // Example: 1234567-compute@developer.gserviceaccount.com - Issuer string `protobuf:"bytes,2,opt,name=issuer" json:"issuer,omitempty"` + Issuer string `protobuf:"bytes,2,opt,name=issuer,proto3" json:"issuer,omitempty"` // URL of the provider's public key set to validate signature of the JWT. See // [OpenID Discovery](https://openid.net/specs/openid-connect-discovery-1_0.html#ProviderMetadata). // Optional if the key set document: @@ -190,7 +184,7 @@ type AuthProvider struct { // - can be inferred from the email domain of the issuer (e.g. a Google service account). // // Example: https://www.googleapis.com/oauth2/v1/certs - JwksUri string `protobuf:"bytes,3,opt,name=jwks_uri,json=jwksUri" json:"jwks_uri,omitempty"` + JwksUri string `protobuf:"bytes,3,opt,name=jwks_uri,json=jwksUri,proto3" json:"jwks_uri,omitempty"` // The list of JWT // [audiences](https://tools.ietf.org/html/draft-ietf-oauth-json-web-token-32#section-4.1.3). // that are allowed to access. A JWT containing any of these audiences will @@ -204,10 +198,10 @@ type AuthProvider struct { // // audiences: bookstore_android.apps.googleusercontent.com, // bookstore_web.apps.googleusercontent.com - Audiences string `protobuf:"bytes,4,opt,name=audiences" json:"audiences,omitempty"` + Audiences string `protobuf:"bytes,4,opt,name=audiences,proto3" json:"audiences,omitempty"` // Redirect URL if JWT token is required but no present or is expired. // Implement authorizationUrl of securityDefinitions in OpenAPI spec. - AuthorizationUrl string `protobuf:"bytes,5,opt,name=authorization_url,json=authorizationUrl" json:"authorization_url,omitempty"` + AuthorizationUrl string `protobuf:"bytes,5,opt,name=authorization_url,json=authorizationUrl,proto3" json:"authorization_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -217,7 +211,7 @@ func (m *AuthProvider) Reset() { *m = AuthProvider{} } func (m *AuthProvider) String() string { return proto.CompactTextString(m) } func (*AuthProvider) ProtoMessage() {} func (*AuthProvider) Descriptor() ([]byte, []int) { - return fileDescriptor_auth_6939d4e8a8cd2956, []int{2} + return fileDescriptor_auth_52a59fcac3533a16, []int{2} } func (m *AuthProvider) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AuthProvider.Unmarshal(m, b) @@ -298,7 +292,7 @@ type OAuthRequirements struct { // // canonical_scopes: https://www.googleapis.com/auth/calendar, // https://www.googleapis.com/auth/calendar.read - CanonicalScopes string `protobuf:"bytes,1,opt,name=canonical_scopes,json=canonicalScopes" json:"canonical_scopes,omitempty"` + CanonicalScopes string `protobuf:"bytes,1,opt,name=canonical_scopes,json=canonicalScopes,proto3" json:"canonical_scopes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -308,7 +302,7 @@ func (m *OAuthRequirements) Reset() { *m = OAuthRequirements{} } func (m *OAuthRequirements) String() string { return proto.CompactTextString(m) } func (*OAuthRequirements) ProtoMessage() {} func (*OAuthRequirements) Descriptor() ([]byte, []int) { - return fileDescriptor_auth_6939d4e8a8cd2956, []int{3} + return fileDescriptor_auth_52a59fcac3533a16, []int{3} } func (m *OAuthRequirements) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_OAuthRequirements.Unmarshal(m, b) @@ -343,7 +337,7 @@ type AuthRequirement struct { // Example: // // provider_id: bookstore_auth - ProviderId string `protobuf:"bytes,1,opt,name=provider_id,json=providerId" json:"provider_id,omitempty"` + ProviderId string `protobuf:"bytes,1,opt,name=provider_id,json=providerId,proto3" json:"provider_id,omitempty"` // NOTE: This will be deprecated soon, once AuthProvider.audiences is // implemented and accepted in all the runtime components. // @@ -360,7 +354,7 @@ type AuthRequirement struct { // // audiences: bookstore_android.apps.googleusercontent.com, // bookstore_web.apps.googleusercontent.com - Audiences string `protobuf:"bytes,2,opt,name=audiences" json:"audiences,omitempty"` + Audiences string `protobuf:"bytes,2,opt,name=audiences,proto3" json:"audiences,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -370,7 +364,7 @@ func (m *AuthRequirement) Reset() { *m = AuthRequirement{} } func (m *AuthRequirement) String() string { return proto.CompactTextString(m) } func (*AuthRequirement) ProtoMessage() {} func (*AuthRequirement) Descriptor() ([]byte, []int) { - return fileDescriptor_auth_6939d4e8a8cd2956, []int{4} + return fileDescriptor_auth_52a59fcac3533a16, []int{4} } func (m *AuthRequirement) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AuthRequirement.Unmarshal(m, b) @@ -412,9 +406,9 @@ func init() { proto.RegisterType((*AuthRequirement)(nil), "google.api.AuthRequirement") } -func init() { proto.RegisterFile("google/api/auth.proto", fileDescriptor_auth_6939d4e8a8cd2956) } +func init() { proto.RegisterFile("google/api/auth.proto", fileDescriptor_auth_52a59fcac3533a16) } -var fileDescriptor_auth_6939d4e8a8cd2956 = []byte{ +var fileDescriptor_auth_52a59fcac3533a16 = []byte{ // 465 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x64, 0x52, 0x5f, 0x6b, 0x13, 0x4f, 0x14, 0x65, 0x93, 0xa6, 0xcd, 0xde, 0x94, 0xb4, 0x1d, 0xf8, 0x95, 0xfd, 0xd5, 0xaa, 0x21, 0x4f, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/backend.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/backend.pb.go index 604f3252a..3652109d1 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/backend.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/backend.pb.go @@ -23,7 +23,7 @@ type Backend struct { // A list of API backend rules that apply to individual API methods. // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*BackendRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*BackendRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -33,7 +33,7 @@ func (m *Backend) Reset() { *m = Backend{} } func (m *Backend) String() string { return proto.CompactTextString(m) } func (*Backend) ProtoMessage() {} func (*Backend) Descriptor() ([]byte, []int) { - return fileDescriptor_backend_6a3ac2605e71ddff, []int{0} + return fileDescriptor_backend_98db65a00dbdbb78, []int{0} } func (m *Backend) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Backend.Unmarshal(m, b) @@ -65,12 +65,15 @@ type BackendRule struct { // Selects the methods to which this rule applies. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // The address of the API backend. - Address string `protobuf:"bytes,2,opt,name=address" json:"address,omitempty"` - // The number of seconds to wait for a response from a request. The - // default depends on the deployment context. - Deadline float64 `protobuf:"fixed64,3,opt,name=deadline" json:"deadline,omitempty"` + Address string `protobuf:"bytes,2,opt,name=address,proto3" json:"address,omitempty"` + // The number of seconds to wait for a response from a request. The default + // deadline for gRPC is infinite (no deadline) and HTTP requests is 5 seconds. + Deadline float64 `protobuf:"fixed64,3,opt,name=deadline,proto3" json:"deadline,omitempty"` + // Minimum deadline in seconds needed for this method. Calls having deadline + // value lower than this will be rejected. + MinDeadline float64 `protobuf:"fixed64,4,opt,name=min_deadline,json=minDeadline,proto3" json:"min_deadline,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -80,7 +83,7 @@ func (m *BackendRule) Reset() { *m = BackendRule{} } func (m *BackendRule) String() string { return proto.CompactTextString(m) } func (*BackendRule) ProtoMessage() {} func (*BackendRule) Descriptor() ([]byte, []int) { - return fileDescriptor_backend_6a3ac2605e71ddff, []int{1} + return fileDescriptor_backend_98db65a00dbdbb78, []int{1} } func (m *BackendRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BackendRule.Unmarshal(m, b) @@ -121,28 +124,36 @@ func (m *BackendRule) GetDeadline() float64 { return 0 } +func (m *BackendRule) GetMinDeadline() float64 { + if m != nil { + return m.MinDeadline + } + return 0 +} + func init() { proto.RegisterType((*Backend)(nil), "google.api.Backend") proto.RegisterType((*BackendRule)(nil), "google.api.BackendRule") } -func init() { proto.RegisterFile("google/api/backend.proto", fileDescriptor_backend_6a3ac2605e71ddff) } +func init() { proto.RegisterFile("google/api/backend.proto", fileDescriptor_backend_98db65a00dbdbb78) } -var fileDescriptor_backend_6a3ac2605e71ddff = []byte{ - // 227 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x90, 0x31, 0x4f, 0x03, 0x21, - 0x18, 0x86, 0x43, 0xab, 0x56, 0xbf, 0x1a, 0x87, 0x5b, 0x24, 0x4e, 0x97, 0x4e, 0xb7, 0xc8, 0x25, - 0xba, 0x98, 0x38, 0x79, 0x89, 0x31, 0x6e, 0x17, 0x46, 0x17, 0x43, 0xe1, 0x93, 0x10, 0x91, 0x8f, - 0x40, 0xeb, 0x0f, 0xf2, 0x97, 0x9a, 0xc2, 0x59, 0xdb, 0xf1, 0xe1, 0x79, 0xdf, 0xc0, 0x0b, 0x70, - 0x4b, 0x64, 0x3d, 0xf6, 0x2a, 0xba, 0x7e, 0xad, 0xf4, 0x27, 0x06, 0x23, 0x62, 0xa2, 0x0d, 0x35, - 0x50, 0x8d, 0x50, 0xd1, 0xad, 0x1e, 0x60, 0x31, 0x54, 0xd9, 0xdc, 0xc2, 0x69, 0xda, 0x7a, 0xcc, - 0x9c, 0xb5, 0xf3, 0x6e, 0x79, 0x77, 0x2d, 0xfe, 0x63, 0x62, 0xca, 0xc8, 0xad, 0x47, 0x59, 0x53, - 0xab, 0x77, 0x58, 0x1e, 0x9c, 0x36, 0x37, 0x70, 0x9e, 0xd1, 0xa3, 0xde, 0x50, 0xe2, 0xac, 0x65, - 0xdd, 0x85, 0xdc, 0x73, 0xc3, 0x61, 0xa1, 0x8c, 0x49, 0x98, 0x33, 0x9f, 0x15, 0xf5, 0x87, 0xbb, - 0x96, 0x41, 0x65, 0xbc, 0x0b, 0xc8, 0xe7, 0x2d, 0xeb, 0x98, 0xdc, 0xf3, 0x10, 0xe0, 0x4a, 0xd3, - 0xd7, 0xc1, 0x2b, 0x86, 0xcb, 0xe9, 0xc2, 0x71, 0x37, 0x63, 0x64, 0x6f, 0xcf, 0x93, 0xb3, 0xe4, - 0x55, 0xb0, 0x82, 0x92, 0xed, 0x2d, 0x86, 0x32, 0xb2, 0xaf, 0x4a, 0x45, 0x97, 0xcb, 0x0f, 0x64, - 0x4c, 0xdf, 0x4e, 0xa3, 0xa6, 0xf0, 0xe1, 0xec, 0xe3, 0x11, 0xfd, 0xcc, 0x4e, 0x5e, 0x9e, 0xc6, - 0xd7, 0xf5, 0x59, 0x29, 0xde, 0xff, 0x06, 0x00, 0x00, 0xff, 0xff, 0x7e, 0x93, 0x9e, 0x00, 0x39, - 0x01, 0x00, 0x00, +var fileDescriptor_backend_98db65a00dbdbb78 = []byte{ + // 246 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x90, 0xb1, 0x4e, 0xc3, 0x30, + 0x10, 0x86, 0xe5, 0xb6, 0x50, 0xb8, 0x54, 0x0c, 0x59, 0xb0, 0x98, 0x42, 0xa7, 0x2c, 0x38, 0x12, + 0x2c, 0x48, 0x4c, 0x44, 0x20, 0xc4, 0x16, 0x65, 0x64, 0x41, 0xae, 0x7d, 0x58, 0x16, 0x8e, 0x2f, + 0xb2, 0x5b, 0x1e, 0x80, 0x47, 0xe1, 0x49, 0x51, 0x9d, 0x10, 0xda, 0xf1, 0xbb, 0xef, 0x3f, 0xe9, + 0xee, 0x07, 0x6e, 0x88, 0x8c, 0xc3, 0x4a, 0xf6, 0xb6, 0xda, 0x48, 0xf5, 0x89, 0x5e, 0x8b, 0x3e, + 0xd0, 0x96, 0x72, 0x18, 0x8c, 0x90, 0xbd, 0x5d, 0xdf, 0xc3, 0xb2, 0x1e, 0x64, 0x7e, 0x03, 0x27, + 0x61, 0xe7, 0x30, 0x72, 0x56, 0xcc, 0xcb, 0xec, 0xf6, 0x52, 0xfc, 0xc7, 0xc4, 0x98, 0x69, 0x77, + 0x0e, 0xdb, 0x21, 0xb5, 0xfe, 0x66, 0x90, 0x1d, 0x8c, 0xf3, 0x2b, 0x38, 0x8b, 0xe8, 0x50, 0x6d, + 0x29, 0x70, 0x56, 0xb0, 0xf2, 0xbc, 0x9d, 0x38, 0xe7, 0xb0, 0x94, 0x5a, 0x07, 0x8c, 0x91, 0xcf, + 0x92, 0xfa, 0xc3, 0xfd, 0x96, 0x46, 0xa9, 0x9d, 0xf5, 0xc8, 0xe7, 0x05, 0x2b, 0x59, 0x3b, 0x71, + 0x7e, 0x0d, 0xab, 0xce, 0xfa, 0xf7, 0xc9, 0x2f, 0x92, 0xcf, 0x3a, 0xeb, 0x9f, 0xc6, 0x51, 0xed, + 0xe1, 0x42, 0x51, 0x77, 0x70, 0x69, 0xbd, 0x1a, 0x6f, 0x6a, 0xf6, 0xaf, 0x36, 0xec, 0xed, 0x79, + 0x74, 0x86, 0x9c, 0xf4, 0x46, 0x50, 0x30, 0x95, 0x41, 0x9f, 0x8a, 0xa8, 0x06, 0x25, 0x7b, 0x1b, + 0x53, 0x4b, 0x11, 0xc3, 0x97, 0x55, 0xa8, 0xc8, 0x7f, 0x58, 0xf3, 0x70, 0x44, 0x3f, 0xb3, 0xc5, + 0xcb, 0x63, 0xf3, 0xba, 0x39, 0x4d, 0x8b, 0x77, 0xbf, 0x01, 0x00, 0x00, 0xff, 0xff, 0x7d, 0xc3, + 0xd8, 0x52, 0x5d, 0x01, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/billing.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/billing.pb.go index 41d071b07..53a335085 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/billing.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/billing.pb.go @@ -24,6 +24,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // // The following example shows how to configure monitored resources and metrics // for billing: +// // monitored_resources: // - type: library.googleapis.com/branch // labels: @@ -45,7 +46,7 @@ type Billing struct { // There can be multiple consumer destinations per service, each one must have // a different monitored resource type. A metric can be used in at most // one consumer destination. - ConsumerDestinations []*Billing_BillingDestination `protobuf:"bytes,8,rep,name=consumer_destinations,json=consumerDestinations" json:"consumer_destinations,omitempty"` + ConsumerDestinations []*Billing_BillingDestination `protobuf:"bytes,8,rep,name=consumer_destinations,json=consumerDestinations,proto3" json:"consumer_destinations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -55,7 +56,7 @@ func (m *Billing) Reset() { *m = Billing{} } func (m *Billing) String() string { return proto.CompactTextString(m) } func (*Billing) ProtoMessage() {} func (*Billing) Descriptor() ([]byte, []int) { - return fileDescriptor_billing_85e96a8576dd991e, []int{0} + return fileDescriptor_billing_19b9055d537fbaf3, []int{0} } func (m *Billing) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Billing.Unmarshal(m, b) @@ -87,10 +88,10 @@ func (m *Billing) GetConsumerDestinations() []*Billing_BillingDestination { type Billing_BillingDestination struct { // The monitored resource type. The type must be defined in // [Service.monitored_resources][google.api.Service.monitored_resources] section. - MonitoredResource string `protobuf:"bytes,1,opt,name=monitored_resource,json=monitoredResource" json:"monitored_resource,omitempty"` + MonitoredResource string `protobuf:"bytes,1,opt,name=monitored_resource,json=monitoredResource,proto3" json:"monitored_resource,omitempty"` // Names of the metrics to report to this billing destination. // Each name must be defined in [Service.metrics][google.api.Service.metrics] section. - Metrics []string `protobuf:"bytes,2,rep,name=metrics" json:"metrics,omitempty"` + Metrics []string `protobuf:"bytes,2,rep,name=metrics,proto3" json:"metrics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -100,7 +101,7 @@ func (m *Billing_BillingDestination) Reset() { *m = Billing_BillingDesti func (m *Billing_BillingDestination) String() string { return proto.CompactTextString(m) } func (*Billing_BillingDestination) ProtoMessage() {} func (*Billing_BillingDestination) Descriptor() ([]byte, []int) { - return fileDescriptor_billing_85e96a8576dd991e, []int{0, 0} + return fileDescriptor_billing_19b9055d537fbaf3, []int{0, 0} } func (m *Billing_BillingDestination) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Billing_BillingDestination.Unmarshal(m, b) @@ -139,9 +140,9 @@ func init() { proto.RegisterType((*Billing_BillingDestination)(nil), "google.api.Billing.BillingDestination") } -func init() { proto.RegisterFile("google/api/billing.proto", fileDescriptor_billing_85e96a8576dd991e) } +func init() { proto.RegisterFile("google/api/billing.proto", fileDescriptor_billing_19b9055d537fbaf3) } -var fileDescriptor_billing_85e96a8576dd991e = []byte{ +var fileDescriptor_billing_19b9055d537fbaf3 = []byte{ // 265 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x64, 0x90, 0xc1, 0x4a, 0xc4, 0x30, 0x10, 0x86, 0xe9, 0xae, 0xb8, 0x6e, 0x14, 0xc1, 0xa0, 0x58, 0x8a, 0x87, 0xe2, 0x41, 0x7a, 0xb1, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/consumer.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/consumer.pb.go index 8b75dd564..f5532022d 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/consumer.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/consumer.pb.go @@ -74,7 +74,7 @@ func (Property_PropertyType) EnumDescriptor() ([]byte, []int) { // type: INT64 type ProjectProperties struct { // List of per consumer project-specific properties. - Properties []*Property `protobuf:"bytes,1,rep,name=properties" json:"properties,omitempty"` + Properties []*Property `protobuf:"bytes,1,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -123,11 +123,11 @@ func (m *ProjectProperties) GetProperties() []*Property { // define and set these properties. type Property struct { // The name of the property (a.k.a key). - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The type of this property. - Type Property_PropertyType `protobuf:"varint,2,opt,name=type,enum=google.api.Property_PropertyType" json:"type,omitempty"` + Type Property_PropertyType `protobuf:"varint,2,opt,name=type,proto3,enum=google.api.Property_PropertyType" json:"type,omitempty"` // The description of the property - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/context.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/context.pb.go index b28a81b9a..8c8515082 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/context.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/context.pb.go @@ -39,7 +39,7 @@ type Context struct { // A list of RPC context rules that apply to individual API methods. // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*ContextRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*ContextRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -82,11 +82,11 @@ type ContextRule struct { // Selects the methods to which this rule applies. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // A list of full type names of requested contexts. - Requested []string `protobuf:"bytes,2,rep,name=requested" json:"requested,omitempty"` + Requested []string `protobuf:"bytes,2,rep,name=requested,proto3" json:"requested,omitempty"` // A list of full type names of provided contexts. - Provided []string `protobuf:"bytes,3,rep,name=provided" json:"provided,omitempty"` + Provided []string `protobuf:"bytes,3,rep,name=provided,proto3" json:"provided,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/control.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/control.pb.go index 70302f035..62dc24590 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/control.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/control.pb.go @@ -24,7 +24,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Control struct { // The service control environment to use. If empty, no control plane // feature (like quota and billing) will be enabled. - Environment string `protobuf:"bytes,1,opt,name=environment" json:"environment,omitempty"` + Environment string `protobuf:"bytes,1,opt,name=environment,proto3" json:"environment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -34,7 +34,7 @@ func (m *Control) Reset() { *m = Control{} } func (m *Control) String() string { return proto.CompactTextString(m) } func (*Control) ProtoMessage() {} func (*Control) Descriptor() ([]byte, []int) { - return fileDescriptor_control_76c8f7249dd124b5, []int{0} + return fileDescriptor_control_91b68496cee1317a, []int{0} } func (m *Control) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Control.Unmarshal(m, b) @@ -65,9 +65,9 @@ func init() { proto.RegisterType((*Control)(nil), "google.api.Control") } -func init() { proto.RegisterFile("google/api/control.proto", fileDescriptor_control_76c8f7249dd124b5) } +func init() { proto.RegisterFile("google/api/control.proto", fileDescriptor_control_91b68496cee1317a) } -var fileDescriptor_control_76c8f7249dd124b5 = []byte{ +var fileDescriptor_control_91b68496cee1317a = []byte{ // 165 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x48, 0xcf, 0xcf, 0x4f, 0xcf, 0x49, 0xd5, 0x4f, 0x2c, 0xc8, 0xd4, 0x4f, 0xce, 0xcf, 0x2b, 0x29, 0xca, 0xcf, 0xd1, 0x2b, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/documentation.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/documentation.pb.go index 9d6a87bac..b43c9598a 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/documentation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/documentation.pb.go @@ -61,9 +61,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package //
[display text][fully.qualified.proto.name]
// Text can be excluded from doc using the following notation: //
(-- internal comment --)
-// Comments can be made conditional using a visibility label. The below -// text will be only rendered if the `BETA` label is available: -//
(--BETA: comment for BETA users --)
+// // A few directives are available in documentation. Note that // directives must appear on a single line to be properly // identified. The `include` directive includes a markdown file from @@ -78,15 +76,15 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Documentation struct { // A short summary of what the service does. Can only be provided by // plain text. - Summary string `protobuf:"bytes,1,opt,name=summary" json:"summary,omitempty"` + Summary string `protobuf:"bytes,1,opt,name=summary,proto3" json:"summary,omitempty"` // The top level pages for the documentation set. - Pages []*Page `protobuf:"bytes,5,rep,name=pages" json:"pages,omitempty"` + Pages []*Page `protobuf:"bytes,5,rep,name=pages,proto3" json:"pages,omitempty"` // A list of documentation rules that apply to individual API elements. // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*DocumentationRule `protobuf:"bytes,3,rep,name=rules" json:"rules,omitempty"` + Rules []*DocumentationRule `protobuf:"bytes,3,rep,name=rules,proto3" json:"rules,omitempty"` // The URL to the root of documentation. - DocumentationRootUrl string `protobuf:"bytes,4,opt,name=documentation_root_url,json=documentationRootUrl" json:"documentation_root_url,omitempty"` + DocumentationRootUrl string `protobuf:"bytes,4,opt,name=documentation_root_url,json=documentationRootUrl,proto3" json:"documentation_root_url,omitempty"` // Declares a single overview page. For example: //
documentation:
 	//   summary: ...
@@ -100,7 +98,7 @@ type Documentation struct {
 	//     content: (== include overview.md ==)
 	// 
// Note: you cannot specify both `overview` field and `pages` field. - Overview string `protobuf:"bytes,2,opt,name=overview" json:"overview,omitempty"` + Overview string `protobuf:"bytes,2,opt,name=overview,proto3" json:"overview,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -110,7 +108,7 @@ func (m *Documentation) Reset() { *m = Documentation{} } func (m *Documentation) String() string { return proto.CompactTextString(m) } func (*Documentation) ProtoMessage() {} func (*Documentation) Descriptor() ([]byte, []int) { - return fileDescriptor_documentation_5c51e57b0ed8cac0, []int{0} + return fileDescriptor_documentation_fffa126753bc1b4d, []int{0} } func (m *Documentation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Documentation.Unmarshal(m, b) @@ -173,12 +171,12 @@ type DocumentationRule struct { // qualified name, i.e. "foo.*" is ok, but not "foo.b*" or "foo.*.bar". To // specify a default for all applicable elements, the whole pattern "*" // is used. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // Description of the selected API(s). - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Deprecation description of the selected element(s). It can be provided if an // element is marked as `deprecated`. - DeprecationDescription string `protobuf:"bytes,3,opt,name=deprecation_description,json=deprecationDescription" json:"deprecation_description,omitempty"` + DeprecationDescription string `protobuf:"bytes,3,opt,name=deprecation_description,json=deprecationDescription,proto3" json:"deprecation_description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -188,7 +186,7 @@ func (m *DocumentationRule) Reset() { *m = DocumentationRule{} } func (m *DocumentationRule) String() string { return proto.CompactTextString(m) } func (*DocumentationRule) ProtoMessage() {} func (*DocumentationRule) Descriptor() ([]byte, []int) { - return fileDescriptor_documentation_5c51e57b0ed8cac0, []int{1} + return fileDescriptor_documentation_fffa126753bc1b4d, []int{1} } func (m *DocumentationRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DocumentationRule.Unmarshal(m, b) @@ -246,13 +244,13 @@ type Page struct { //
// You can reference `Java` page using Markdown reference link syntax: // `[Java][Tutorial.Java]`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The Markdown content of the page. You can use (== include {path} ==) // to include content from a Markdown file. - Content string `protobuf:"bytes,2,opt,name=content" json:"content,omitempty"` + Content string `protobuf:"bytes,2,opt,name=content,proto3" json:"content,omitempty"` // Subpages of this page. The order of subpages specified here will be // honored in the generated docset. - Subpages []*Page `protobuf:"bytes,3,rep,name=subpages" json:"subpages,omitempty"` + Subpages []*Page `protobuf:"bytes,3,rep,name=subpages,proto3" json:"subpages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,7 +260,7 @@ func (m *Page) Reset() { *m = Page{} } func (m *Page) String() string { return proto.CompactTextString(m) } func (*Page) ProtoMessage() {} func (*Page) Descriptor() ([]byte, []int) { - return fileDescriptor_documentation_5c51e57b0ed8cac0, []int{2} + return fileDescriptor_documentation_fffa126753bc1b4d, []int{2} } func (m *Page) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Page.Unmarshal(m, b) @@ -310,10 +308,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/documentation.proto", fileDescriptor_documentation_5c51e57b0ed8cac0) + proto.RegisterFile("google/api/documentation.proto", fileDescriptor_documentation_fffa126753bc1b4d) } -var fileDescriptor_documentation_5c51e57b0ed8cac0 = []byte{ +var fileDescriptor_documentation_fffa126753bc1b4d = []byte{ // 356 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x74, 0x92, 0xc1, 0x6a, 0xe3, 0x30, 0x14, 0x45, 0x71, 0xec, 0xcc, 0x64, 0x5e, 0x98, 0x61, 0x46, 0x0c, 0x19, 0x33, 0xd0, 0x12, 0xb2, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/endpoint.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/endpoint.pb.go index 164307d57..0ce2e78ea 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/endpoint.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/endpoint.pb.go @@ -37,29 +37,27 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // allow_cors: true type Endpoint struct { // The canonical name of this endpoint. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // DEPRECATED: This field is no longer supported. Instead of using aliases, - // please specify multiple [google.api.Endpoint][google.api.Endpoint] for each of the intented - // alias. + // please specify multiple [google.api.Endpoint][google.api.Endpoint] for each of the intended + // aliases. // // Additional names that this endpoint will be hosted on. - Aliases []string `protobuf:"bytes,2,rep,name=aliases" json:"aliases,omitempty"` - // The list of APIs served by this endpoint. - Apis []string `protobuf:"bytes,3,rep,name=apis" json:"apis,omitempty"` + Aliases []string `protobuf:"bytes,2,rep,name=aliases,proto3" json:"aliases,omitempty"` // The list of features enabled on this endpoint. - Features []string `protobuf:"bytes,4,rep,name=features" json:"features,omitempty"` + Features []string `protobuf:"bytes,4,rep,name=features,proto3" json:"features,omitempty"` // The specification of an Internet routable address of API frontend that will // handle requests to this [API Endpoint](https://cloud.google.com/apis/design/glossary). // It should be either a valid IPv4 address or a fully-qualified domain name. // For example, "8.8.8.8" or "myservice.appspot.com". - Target string `protobuf:"bytes,101,opt,name=target" json:"target,omitempty"` + Target string `protobuf:"bytes,101,opt,name=target,proto3" json:"target,omitempty"` // Allowing // [CORS](https://en.wikipedia.org/wiki/Cross-origin_resource_sharing), aka // cross-domain traffic, would allow the backends served from this endpoint to // receive and respond to HTTP OPTIONS requests. The response will be used by // the browser to determine whether the subsequent cross-origin request is // allowed to proceed. - AllowCors bool `protobuf:"varint,5,opt,name=allow_cors,json=allowCors" json:"allow_cors,omitempty"` + AllowCors bool `protobuf:"varint,5,opt,name=allow_cors,json=allowCors,proto3" json:"allow_cors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -69,7 +67,7 @@ func (m *Endpoint) Reset() { *m = Endpoint{} } func (m *Endpoint) String() string { return proto.CompactTextString(m) } func (*Endpoint) ProtoMessage() {} func (*Endpoint) Descriptor() ([]byte, []int) { - return fileDescriptor_endpoint_3cb40552097f500d, []int{0} + return fileDescriptor_endpoint_4418f5874a831641, []int{0} } func (m *Endpoint) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Endpoint.Unmarshal(m, b) @@ -103,13 +101,6 @@ func (m *Endpoint) GetAliases() []string { return nil } -func (m *Endpoint) GetApis() []string { - if m != nil { - return m.Apis - } - return nil -} - func (m *Endpoint) GetFeatures() []string { if m != nil { return m.Features @@ -135,24 +126,24 @@ func init() { proto.RegisterType((*Endpoint)(nil), "google.api.Endpoint") } -func init() { proto.RegisterFile("google/api/endpoint.proto", fileDescriptor_endpoint_3cb40552097f500d) } +func init() { proto.RegisterFile("google/api/endpoint.proto", fileDescriptor_endpoint_4418f5874a831641) } -var fileDescriptor_endpoint_3cb40552097f500d = []byte{ - // 253 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x90, 0x41, 0x4b, 0xc4, 0x30, - 0x10, 0x85, 0xe9, 0x6e, 0x5d, 0xdb, 0x01, 0x3d, 0xe4, 0x20, 0x71, 0x51, 0x28, 0x9e, 0x7a, 0x6a, - 0x0f, 0x1e, 0x3d, 0xb9, 0xb2, 0x88, 0xb7, 0xd2, 0xa3, 0x17, 0x19, 0xeb, 0x6c, 0x08, 0x64, 0x33, - 0x21, 0x89, 0xfa, 0x73, 0x04, 0x7f, 0xa9, 0x34, 0xed, 0xaa, 0x7b, 0x9b, 0xef, 0xbd, 0x37, 0x61, - 0x5e, 0xe0, 0x52, 0x31, 0x2b, 0x43, 0x2d, 0x3a, 0xdd, 0x92, 0x7d, 0x73, 0xac, 0x6d, 0x6c, 0x9c, - 0xe7, 0xc8, 0x02, 0x26, 0xab, 0x41, 0xa7, 0xd7, 0x57, 0xff, 0x62, 0x68, 0x2d, 0x47, 0x8c, 0x9a, - 0x6d, 0x98, 0x92, 0x37, 0x5f, 0x19, 0x14, 0xdb, 0x79, 0x59, 0x08, 0xc8, 0x2d, 0xee, 0x49, 0x66, - 0x55, 0x56, 0x97, 0x7d, 0x9a, 0x85, 0x84, 0x53, 0x34, 0x1a, 0x03, 0x05, 0xb9, 0xa8, 0x96, 0x75, - 0xd9, 0x1f, 0x70, 0x4c, 0xa3, 0xd3, 0x41, 0x2e, 0x93, 0x9c, 0x66, 0xb1, 0x86, 0x62, 0x47, 0x18, - 0xdf, 0x3d, 0x05, 0x99, 0x27, 0xfd, 0x97, 0xc5, 0x05, 0xac, 0x22, 0x7a, 0x45, 0x51, 0x52, 0x7a, - 0x7f, 0x26, 0x71, 0x0d, 0x80, 0xc6, 0xf0, 0xe7, 0xcb, 0xc0, 0x3e, 0xc8, 0x93, 0x2a, 0xab, 0x8b, - 0xbe, 0x4c, 0xca, 0x03, 0xfb, 0xb0, 0x61, 0x38, 0x1f, 0x78, 0xdf, 0xfc, 0x35, 0xda, 0x9c, 0x1d, - 0x0e, 0xee, 0xc6, 0x0a, 0x5d, 0xf6, 0xbc, 0x9d, 0x4d, 0xc5, 0x06, 0xad, 0x6a, 0xd8, 0xab, 0x56, - 0x91, 0x4d, 0x05, 0xdb, 0xc9, 0x1a, 0x8f, 0x4b, 0x3f, 0x10, 0xc8, 0x7f, 0xe8, 0x81, 0x06, 0xb6, - 0x3b, 0xad, 0xee, 0x8e, 0xe8, 0x7b, 0x91, 0x3f, 0xde, 0x77, 0x4f, 0xaf, 0xab, 0xb4, 0x78, 0xfb, - 0x13, 0x00, 0x00, 0xff, 0xff, 0x34, 0x0e, 0xdd, 0x70, 0x60, 0x01, 0x00, 0x00, +var fileDescriptor_endpoint_4418f5874a831641 = []byte{ + // 245 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x8f, 0xb1, 0x4e, 0xc3, 0x30, + 0x10, 0x86, 0xe5, 0x12, 0x4a, 0x72, 0x12, 0x0c, 0x1e, 0x90, 0xa9, 0x40, 0x8a, 0x98, 0x32, 0x25, + 0x03, 0x23, 0x13, 0x45, 0x15, 0x62, 0x8b, 0x32, 0xb2, 0xa0, 0x23, 0x5c, 0x2d, 0x4b, 0xae, 0xcf, + 0xb2, 0x0d, 0x3c, 0x04, 0x6f, 0xc1, 0x93, 0x22, 0x9c, 0x14, 0xe8, 0xe6, 0xcf, 0xdf, 0xdd, 0xe9, + 0xff, 0xe1, 0x42, 0x33, 0x6b, 0x4b, 0x1d, 0x7a, 0xd3, 0x91, 0x7b, 0xf5, 0x6c, 0x5c, 0x6a, 0x7d, + 0xe0, 0xc4, 0x12, 0x26, 0xd5, 0xa2, 0x37, 0xab, 0xcb, 0x7f, 0x63, 0xe8, 0x1c, 0x27, 0x4c, 0x86, + 0x5d, 0x9c, 0x26, 0xaf, 0x3f, 0x05, 0x94, 0x9b, 0x79, 0x59, 0x4a, 0x28, 0x1c, 0xee, 0x48, 0x89, + 0x5a, 0x34, 0xd5, 0x90, 0xdf, 0x52, 0xc1, 0x09, 0x5a, 0x83, 0x91, 0xa2, 0x5a, 0xd4, 0x47, 0x4d, + 0x35, 0xec, 0x51, 0xae, 0xa0, 0xdc, 0x12, 0xa6, 0xb7, 0x40, 0x51, 0x15, 0x59, 0xfd, 0xb2, 0x3c, + 0x87, 0x65, 0xc2, 0xa0, 0x29, 0x29, 0xca, 0xb7, 0x66, 0x92, 0x57, 0x00, 0x68, 0x2d, 0x7f, 0x3c, + 0x8f, 0x1c, 0xa2, 0x3a, 0xae, 0x45, 0x53, 0x0e, 0x55, 0xfe, 0xb9, 0xe7, 0x10, 0xd7, 0x0c, 0x67, + 0x23, 0xef, 0xda, 0xbf, 0xf4, 0xeb, 0xd3, 0x7d, 0xb8, 0xfe, 0x27, 0x6e, 0x2f, 0x9e, 0x36, 0xb3, + 0xd4, 0x6c, 0xd1, 0xe9, 0x96, 0x83, 0xee, 0x34, 0xb9, 0x5c, 0xa6, 0x9b, 0x14, 0x7a, 0x13, 0x73, + 0xdb, 0x48, 0xe1, 0xdd, 0x8c, 0x34, 0xb2, 0xdb, 0x1a, 0x7d, 0x7b, 0x40, 0x5f, 0x8b, 0xe2, 0xe1, + 0xae, 0x7f, 0x7c, 0x59, 0xe6, 0xc5, 0x9b, 0xef, 0x00, 0x00, 0x00, 0xff, 0xff, 0xc8, 0x32, 0xa4, + 0x36, 0x4c, 0x01, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/log.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/log.pb.go index e51ae5faa..9627510ba 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/log.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/log.pb.go @@ -32,17 +32,17 @@ type LogDescriptor struct { // include the following characters: upper- and lower-case alphanumeric // characters [A-Za-z0-9], and punctuation characters including // slash, underscore, hyphen, period [/_-.]. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The set of labels that are available to describe a specific log entry. // Runtime requests that contain labels not specified here are // considered invalid. - Labels []*label.LabelDescriptor `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty"` + Labels []*label.LabelDescriptor `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty"` // A human-readable description of this log. This information appears in // the documentation and can contain details. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The human-readable name for this log. This information appears on // the user interface and should be concise. - DisplayName string `protobuf:"bytes,4,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,4,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -52,7 +52,7 @@ func (m *LogDescriptor) Reset() { *m = LogDescriptor{} } func (m *LogDescriptor) String() string { return proto.CompactTextString(m) } func (*LogDescriptor) ProtoMessage() {} func (*LogDescriptor) Descriptor() ([]byte, []int) { - return fileDescriptor_log_7121ad8ed535f35a, []int{0} + return fileDescriptor_log_de92a25ed5c88332, []int{0} } func (m *LogDescriptor) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LogDescriptor.Unmarshal(m, b) @@ -104,9 +104,9 @@ func init() { proto.RegisterType((*LogDescriptor)(nil), "google.api.LogDescriptor") } -func init() { proto.RegisterFile("google/api/log.proto", fileDescriptor_log_7121ad8ed535f35a) } +func init() { proto.RegisterFile("google/api/log.proto", fileDescriptor_log_de92a25ed5c88332) } -var fileDescriptor_log_7121ad8ed535f35a = []byte{ +var fileDescriptor_log_de92a25ed5c88332 = []byte{ // 238 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x8f, 0xc1, 0x4a, 0xc3, 0x40, 0x10, 0x86, 0x49, 0x1b, 0x8a, 0x6e, 0xd5, 0xc3, 0x22, 0x12, 0xf4, 0x12, 0x3d, 0xf5, 0xb4, 0x01, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/logging.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/logging.pb.go index b12fe752a..5ecd5fa60 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/logging.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/logging.pb.go @@ -53,12 +53,12 @@ type Logging struct { // There can be multiple producer destinations, each one must have a // different monitored resource type. A log can be used in at most // one producer destination. - ProducerDestinations []*Logging_LoggingDestination `protobuf:"bytes,1,rep,name=producer_destinations,json=producerDestinations" json:"producer_destinations,omitempty"` + ProducerDestinations []*Logging_LoggingDestination `protobuf:"bytes,1,rep,name=producer_destinations,json=producerDestinations,proto3" json:"producer_destinations,omitempty"` // Logging configurations for sending logs to the consumer project. // There can be multiple consumer destinations, each one must have a // different monitored resource type. A log can be used in at most // one consumer destination. - ConsumerDestinations []*Logging_LoggingDestination `protobuf:"bytes,2,rep,name=consumer_destinations,json=consumerDestinations" json:"consumer_destinations,omitempty"` + ConsumerDestinations []*Logging_LoggingDestination `protobuf:"bytes,2,rep,name=consumer_destinations,json=consumerDestinations,proto3" json:"consumer_destinations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -68,7 +68,7 @@ func (m *Logging) Reset() { *m = Logging{} } func (m *Logging) String() string { return proto.CompactTextString(m) } func (*Logging) ProtoMessage() {} func (*Logging) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_d70ddf346064a6bb, []int{0} + return fileDescriptor_logging_e829a2132d45e683, []int{0} } func (m *Logging) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Logging.Unmarshal(m, b) @@ -107,12 +107,12 @@ func (m *Logging) GetConsumerDestinations() []*Logging_LoggingDestination { type Logging_LoggingDestination struct { // The monitored resource type. The type must be defined in the // [Service.monitored_resources][google.api.Service.monitored_resources] section. - MonitoredResource string `protobuf:"bytes,3,opt,name=monitored_resource,json=monitoredResource" json:"monitored_resource,omitempty"` + MonitoredResource string `protobuf:"bytes,3,opt,name=monitored_resource,json=monitoredResource,proto3" json:"monitored_resource,omitempty"` // Names of the logs to be sent to this destination. Each name must // be defined in the [Service.logs][google.api.Service.logs] section. If the log name is // not a domain scoped name, it will be automatically prefixed with // the service name followed by "/". - Logs []string `protobuf:"bytes,1,rep,name=logs" json:"logs,omitempty"` + Logs []string `protobuf:"bytes,1,rep,name=logs,proto3" json:"logs,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -122,7 +122,7 @@ func (m *Logging_LoggingDestination) Reset() { *m = Logging_LoggingDesti func (m *Logging_LoggingDestination) String() string { return proto.CompactTextString(m) } func (*Logging_LoggingDestination) ProtoMessage() {} func (*Logging_LoggingDestination) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_d70ddf346064a6bb, []int{0, 0} + return fileDescriptor_logging_e829a2132d45e683, []int{0, 0} } func (m *Logging_LoggingDestination) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Logging_LoggingDestination.Unmarshal(m, b) @@ -161,9 +161,9 @@ func init() { proto.RegisterType((*Logging_LoggingDestination)(nil), "google.api.Logging.LoggingDestination") } -func init() { proto.RegisterFile("google/api/logging.proto", fileDescriptor_logging_d70ddf346064a6bb) } +func init() { proto.RegisterFile("google/api/logging.proto", fileDescriptor_logging_e829a2132d45e683) } -var fileDescriptor_logging_d70ddf346064a6bb = []byte{ +var fileDescriptor_logging_e829a2132d45e683 = []byte{ // 270 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x90, 0x4f, 0x4b, 0xc4, 0x30, 0x10, 0xc5, 0x69, 0x77, 0x51, 0x36, 0x8a, 0x60, 0x50, 0x28, 0x8b, 0x87, 0xc5, 0x83, 0xec, 0xc5, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/monitoring.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/monitoring.pb.go index 9106ddbce..f14702285 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/monitoring.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/monitoring.pb.go @@ -61,12 +61,12 @@ type Monitoring struct { // There can be multiple producer destinations, each one must have a // different monitored resource type. A metric can be used in at most // one producer destination. - ProducerDestinations []*Monitoring_MonitoringDestination `protobuf:"bytes,1,rep,name=producer_destinations,json=producerDestinations" json:"producer_destinations,omitempty"` + ProducerDestinations []*Monitoring_MonitoringDestination `protobuf:"bytes,1,rep,name=producer_destinations,json=producerDestinations,proto3" json:"producer_destinations,omitempty"` // Monitoring configurations for sending metrics to the consumer project. // There can be multiple consumer destinations, each one must have a // different monitored resource type. A metric can be used in at most // one consumer destination. - ConsumerDestinations []*Monitoring_MonitoringDestination `protobuf:"bytes,2,rep,name=consumer_destinations,json=consumerDestinations" json:"consumer_destinations,omitempty"` + ConsumerDestinations []*Monitoring_MonitoringDestination `protobuf:"bytes,2,rep,name=consumer_destinations,json=consumerDestinations,proto3" json:"consumer_destinations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -76,7 +76,7 @@ func (m *Monitoring) Reset() { *m = Monitoring{} } func (m *Monitoring) String() string { return proto.CompactTextString(m) } func (*Monitoring) ProtoMessage() {} func (*Monitoring) Descriptor() ([]byte, []int) { - return fileDescriptor_monitoring_0c8882cc470ba7b3, []int{0} + return fileDescriptor_monitoring_4e73ce5561a2dc9c, []int{0} } func (m *Monitoring) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Monitoring.Unmarshal(m, b) @@ -115,10 +115,10 @@ func (m *Monitoring) GetConsumerDestinations() []*Monitoring_MonitoringDestinati type Monitoring_MonitoringDestination struct { // The monitored resource type. The type must be defined in // [Service.monitored_resources][google.api.Service.monitored_resources] section. - MonitoredResource string `protobuf:"bytes,1,opt,name=monitored_resource,json=monitoredResource" json:"monitored_resource,omitempty"` + MonitoredResource string `protobuf:"bytes,1,opt,name=monitored_resource,json=monitoredResource,proto3" json:"monitored_resource,omitempty"` // Names of the metrics to report to this monitoring destination. // Each name must be defined in [Service.metrics][google.api.Service.metrics] section. - Metrics []string `protobuf:"bytes,2,rep,name=metrics" json:"metrics,omitempty"` + Metrics []string `protobuf:"bytes,2,rep,name=metrics,proto3" json:"metrics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -128,7 +128,7 @@ func (m *Monitoring_MonitoringDestination) Reset() { *m = Monitoring_Mon func (m *Monitoring_MonitoringDestination) String() string { return proto.CompactTextString(m) } func (*Monitoring_MonitoringDestination) ProtoMessage() {} func (*Monitoring_MonitoringDestination) Descriptor() ([]byte, []int) { - return fileDescriptor_monitoring_0c8882cc470ba7b3, []int{0, 0} + return fileDescriptor_monitoring_4e73ce5561a2dc9c, []int{0, 0} } func (m *Monitoring_MonitoringDestination) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Monitoring_MonitoringDestination.Unmarshal(m, b) @@ -168,10 +168,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/monitoring.proto", fileDescriptor_monitoring_0c8882cc470ba7b3) + proto.RegisterFile("google/api/monitoring.proto", fileDescriptor_monitoring_4e73ce5561a2dc9c) } -var fileDescriptor_monitoring_0c8882cc470ba7b3 = []byte{ +var fileDescriptor_monitoring_4e73ce5561a2dc9c = []byte{ // 271 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x51, 0xcd, 0x4a, 0xc4, 0x30, 0x10, 0xa6, 0x55, 0x94, 0x8d, 0xa0, 0x58, 0x5c, 0x28, 0xab, 0x87, 0xc5, 0xd3, 0x1e, 0xb4, 0x05, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/quota.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/quota.pb.go index f3fb6cb50..101e3a78c 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/quota.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/quota.pb.go @@ -72,12 +72,12 @@ type Quota struct { // List of `QuotaLimit` definitions for the service. // // Used by metric-based quotas only. - Limits []*QuotaLimit `protobuf:"bytes,3,rep,name=limits" json:"limits,omitempty"` + Limits []*QuotaLimit `protobuf:"bytes,3,rep,name=limits,proto3" json:"limits,omitempty"` // List of `MetricRule` definitions, each one mapping a selected method to one // or more metrics. // // Used by metric-based quotas only. - MetricRules []*MetricRule `protobuf:"bytes,4,rep,name=metric_rules,json=metricRules" json:"metric_rules,omitempty"` + MetricRules []*MetricRule `protobuf:"bytes,4,rep,name=metric_rules,json=metricRules,proto3" json:"metric_rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -130,14 +130,14 @@ type MetricRule struct { // Selects the methods to which this rule applies. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // Metrics to update when the selected methods are called, and the associated // cost applied to each metric. // // The key of the map is the metric name, and the values are the amount // increased for the metric against which the quota limits are defined. // The value must not be negative. - MetricCosts map[string]int64 `protobuf:"bytes,2,rep,name=metric_costs,json=metricCosts" json:"metric_costs,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + MetricCosts map[string]int64 `protobuf:"bytes,2,rep,name=metric_costs,json=metricCosts,proto3" json:"metric_costs,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -203,11 +203,11 @@ type QuotaLimit struct { // immutable. You can use the display_name field to provide a user-friendly // name for the limit. The display name can be evolved over time without // affecting the identity of the limit. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` // Optional. User-visible, extended description for this quota limit. // Should be used only when more context is needed to understand this limit // than provided by the limit's display name (see: `display_name`). - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Default number of tokens that can be consumed during the specified // duration. This is the number of tokens assigned when a client // application developer activates the service for his/her project. @@ -218,7 +218,7 @@ type QuotaLimit struct { // negative values are allowed. // // Used by group-based quotas only. - DefaultLimit int64 `protobuf:"varint,3,opt,name=default_limit,json=defaultLimit" json:"default_limit,omitempty"` + DefaultLimit int64 `protobuf:"varint,3,opt,name=default_limit,json=defaultLimit,proto3" json:"default_limit,omitempty"` // Maximum number of tokens that can be consumed during the specified // duration. Client application developers can override the default limit up // to this maximum. If specified, this value cannot be set to a value less @@ -228,7 +228,7 @@ type QuotaLimit struct { // indicating unlimited maximum quota. // // Used by group-based quotas only. - MaxLimit int64 `protobuf:"varint,4,opt,name=max_limit,json=maxLimit" json:"max_limit,omitempty"` + MaxLimit int64 `protobuf:"varint,4,opt,name=max_limit,json=maxLimit,proto3" json:"max_limit,omitempty"` // Free tier value displayed in the Developers Console for this limit. // The free tier is the number of tokens that will be subtracted from the // billed amount when billing is enabled. @@ -237,20 +237,20 @@ type QuotaLimit struct { // defaults to 0, indicating that there is no free tier for this service. // // Used by group-based quotas only. - FreeTier int64 `protobuf:"varint,7,opt,name=free_tier,json=freeTier" json:"free_tier,omitempty"` + FreeTier int64 `protobuf:"varint,7,opt,name=free_tier,json=freeTier,proto3" json:"free_tier,omitempty"` // Duration of this limit in textual notation. Example: "100s", "24h", "1d". // For duration longer than a day, only multiple of days is supported. We // support only "100s" and "1d" for now. Additional support will be added in // the future. "0" indicates indefinite duration. // // Used by group-based quotas only. - Duration string `protobuf:"bytes,5,opt,name=duration" json:"duration,omitempty"` + Duration string `protobuf:"bytes,5,opt,name=duration,proto3" json:"duration,omitempty"` // The name of the metric this quota limit applies to. The quota limits with // the same metric will be checked together during runtime. The metric must be // defined within the service config. // // Used by metric-based quotas only. - Metric string `protobuf:"bytes,8,opt,name=metric" json:"metric,omitempty"` + Metric string `protobuf:"bytes,8,opt,name=metric,proto3" json:"metric,omitempty"` // Specify the unit of the quota limit. It uses the same syntax as // [Metric.unit][]. The supported unit kinds are determined by the quota // backend system. @@ -282,7 +282,7 @@ type QuotaLimit struct { // The "1" at the beginning is required to follow the metric unit syntax. // // Used by metric-based quotas only. - Unit string `protobuf:"bytes,9,opt,name=unit" json:"unit,omitempty"` + Unit string `protobuf:"bytes,9,opt,name=unit,proto3" json:"unit,omitempty"` // Tiered limit values. Also allows for regional or zone overrides for these // values if "/{region}" or "/{zone}" is specified in the unit field. // @@ -315,12 +315,12 @@ type QuotaLimit struct { // tier as well. // // Used by metric-based quotas only. - Values map[string]int64 `protobuf:"bytes,10,rep,name=values" json:"values,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + Values map[string]int64 `protobuf:"bytes,10,rep,name=values,proto3" json:"values,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` // User-visible display name for this limit. // Optional. If not set, the UI will provide a default display name based on // the quota configuration. This field can be used to override the default // display name generated from the configuration. - DisplayName string `protobuf:"bytes,12,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,12,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/service.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/service.pb.go index e86f4bd44..bf9536e5c 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/service.pb.go @@ -55,24 +55,24 @@ type Service struct { // affects the interpretation of the service configuration. For example, // certain features are enabled by default for certain config versions. // The latest config version is `3`. - ConfigVersion *wrappers.UInt32Value `protobuf:"bytes,20,opt,name=config_version,json=configVersion" json:"config_version,omitempty"` + ConfigVersion *wrappers.UInt32Value `protobuf:"bytes,20,opt,name=config_version,json=configVersion,proto3" json:"config_version,omitempty"` // The DNS address at which this service is available, // e.g. `calendar.googleapis.com`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A unique ID for a specific instance of this message, typically assigned // by the client for tracking purpose. If empty, the server may choose to // generate one instead. - Id string `protobuf:"bytes,33,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,33,opt,name=id,proto3" json:"id,omitempty"` // The product title for this service. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // The Google project that owns this service. - ProducerProjectId string `protobuf:"bytes,22,opt,name=producer_project_id,json=producerProjectId" json:"producer_project_id,omitempty"` + ProducerProjectId string `protobuf:"bytes,22,opt,name=producer_project_id,json=producerProjectId,proto3" json:"producer_project_id,omitempty"` // A list of API interfaces exported by this service. Only the `name` field // of the [google.protobuf.Api][google.protobuf.Api] needs to be provided by the configuration // author, as the remaining fields will be derived from the IDL during the // normalization process. It is an error to specify an API interface here // which cannot be resolved against the associated IDL files. - Apis []*api.Api `protobuf:"bytes,3,rep,name=apis" json:"apis,omitempty"` + Apis []*api.Api `protobuf:"bytes,3,rep,name=apis,proto3" json:"apis,omitempty"` // A list of all proto message types included in this API service. // Types referenced directly or indirectly by the `apis` are // automatically included. Messages which are not referenced but @@ -81,7 +81,7 @@ type Service struct { // // types: // - name: google.protobuf.Int32 - Types []*ptype.Type `protobuf:"bytes,4,rep,name=types" json:"types,omitempty"` + Types []*ptype.Type `protobuf:"bytes,4,rep,name=types,proto3" json:"types,omitempty"` // A list of all enum types included in this API service. Enums // referenced directly or indirectly by the `apis` are automatically // included. Enums which are not referenced but shall be included @@ -89,46 +89,46 @@ type Service struct { // // enums: // - name: google.someapi.v1.SomeEnum - Enums []*ptype.Enum `protobuf:"bytes,5,rep,name=enums" json:"enums,omitempty"` + Enums []*ptype.Enum `protobuf:"bytes,5,rep,name=enums,proto3" json:"enums,omitempty"` // Additional API documentation. - Documentation *Documentation `protobuf:"bytes,6,opt,name=documentation" json:"documentation,omitempty"` + Documentation *Documentation `protobuf:"bytes,6,opt,name=documentation,proto3" json:"documentation,omitempty"` // API backend configuration. - Backend *Backend `protobuf:"bytes,8,opt,name=backend" json:"backend,omitempty"` + Backend *Backend `protobuf:"bytes,8,opt,name=backend,proto3" json:"backend,omitempty"` // HTTP configuration. - Http *annotations.Http `protobuf:"bytes,9,opt,name=http" json:"http,omitempty"` + Http *annotations.Http `protobuf:"bytes,9,opt,name=http,proto3" json:"http,omitempty"` // Quota configuration. - Quota *Quota `protobuf:"bytes,10,opt,name=quota" json:"quota,omitempty"` + Quota *Quota `protobuf:"bytes,10,opt,name=quota,proto3" json:"quota,omitempty"` // Auth configuration. - Authentication *Authentication `protobuf:"bytes,11,opt,name=authentication" json:"authentication,omitempty"` + Authentication *Authentication `protobuf:"bytes,11,opt,name=authentication,proto3" json:"authentication,omitempty"` // Context configuration. - Context *Context `protobuf:"bytes,12,opt,name=context" json:"context,omitempty"` + Context *Context `protobuf:"bytes,12,opt,name=context,proto3" json:"context,omitempty"` // Configuration controlling usage of this service. - Usage *Usage `protobuf:"bytes,15,opt,name=usage" json:"usage,omitempty"` + Usage *Usage `protobuf:"bytes,15,opt,name=usage,proto3" json:"usage,omitempty"` // Configuration for network endpoints. If this is empty, then an endpoint // with the same name as the service is automatically generated to service all // defined APIs. - Endpoints []*Endpoint `protobuf:"bytes,18,rep,name=endpoints" json:"endpoints,omitempty"` + Endpoints []*Endpoint `protobuf:"bytes,18,rep,name=endpoints,proto3" json:"endpoints,omitempty"` // Configuration for the service control plane. - Control *Control `protobuf:"bytes,21,opt,name=control" json:"control,omitempty"` + Control *Control `protobuf:"bytes,21,opt,name=control,proto3" json:"control,omitempty"` // Defines the logs used by this service. - Logs []*LogDescriptor `protobuf:"bytes,23,rep,name=logs" json:"logs,omitempty"` + Logs []*LogDescriptor `protobuf:"bytes,23,rep,name=logs,proto3" json:"logs,omitempty"` // Defines the metrics used by this service. - Metrics []*metric.MetricDescriptor `protobuf:"bytes,24,rep,name=metrics" json:"metrics,omitempty"` + Metrics []*metric.MetricDescriptor `protobuf:"bytes,24,rep,name=metrics,proto3" json:"metrics,omitempty"` // Defines the monitored resources used by this service. This is required // by the [Service.monitoring][google.api.Service.monitoring] and [Service.logging][google.api.Service.logging] configurations. - MonitoredResources []*monitoredres.MonitoredResourceDescriptor `protobuf:"bytes,25,rep,name=monitored_resources,json=monitoredResources" json:"monitored_resources,omitempty"` + MonitoredResources []*monitoredres.MonitoredResourceDescriptor `protobuf:"bytes,25,rep,name=monitored_resources,json=monitoredResources,proto3" json:"monitored_resources,omitempty"` // Billing configuration. - Billing *Billing `protobuf:"bytes,26,opt,name=billing" json:"billing,omitempty"` + Billing *Billing `protobuf:"bytes,26,opt,name=billing,proto3" json:"billing,omitempty"` // Logging configuration. - Logging *Logging `protobuf:"bytes,27,opt,name=logging" json:"logging,omitempty"` + Logging *Logging `protobuf:"bytes,27,opt,name=logging,proto3" json:"logging,omitempty"` // Monitoring configuration. - Monitoring *Monitoring `protobuf:"bytes,28,opt,name=monitoring" json:"monitoring,omitempty"` + Monitoring *Monitoring `protobuf:"bytes,28,opt,name=monitoring,proto3" json:"monitoring,omitempty"` // System parameter configuration. - SystemParameters *SystemParameters `protobuf:"bytes,29,opt,name=system_parameters,json=systemParameters" json:"system_parameters,omitempty"` + SystemParameters *SystemParameters `protobuf:"bytes,29,opt,name=system_parameters,json=systemParameters,proto3" json:"system_parameters,omitempty"` // Output only. The source information for this configuration if available. - SourceInfo *SourceInfo `protobuf:"bytes,37,opt,name=source_info,json=sourceInfo" json:"source_info,omitempty"` + SourceInfo *SourceInfo `protobuf:"bytes,37,opt,name=source_info,json=sourceInfo,proto3" json:"source_info,omitempty"` // Experimental configuration. - Experimental *api1.Experimental `protobuf:"bytes,101,opt,name=experimental" json:"experimental,omitempty"` + Experimental *api1.Experimental `protobuf:"bytes,101,opt,name=experimental,proto3" json:"experimental,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/source_info.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/source_info.pb.go index 056f83875..911714d2f 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/source_info.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/source_info.pb.go @@ -22,7 +22,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Source information used to create a Service Config type SourceInfo struct { // All files used during config generation. - SourceFiles []*any.Any `protobuf:"bytes,1,rep,name=source_files,json=sourceFiles" json:"source_files,omitempty"` + SourceFiles []*any.Any `protobuf:"bytes,1,rep,name=source_files,json=sourceFiles,proto3" json:"source_files,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -32,7 +32,7 @@ func (m *SourceInfo) Reset() { *m = SourceInfo{} } func (m *SourceInfo) String() string { return proto.CompactTextString(m) } func (*SourceInfo) ProtoMessage() {} func (*SourceInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_source_info_5bff73c4b21f6f4f, []int{0} + return fileDescriptor_source_info_7a3c15eb7da8e642, []int{0} } func (m *SourceInfo) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SourceInfo.Unmarshal(m, b) @@ -64,10 +64,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/source_info.proto", fileDescriptor_source_info_5bff73c4b21f6f4f) + proto.RegisterFile("google/api/source_info.proto", fileDescriptor_source_info_7a3c15eb7da8e642) } -var fileDescriptor_source_info_5bff73c4b21f6f4f = []byte{ +var fileDescriptor_source_info_7a3c15eb7da8e642 = []byte{ // 198 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x49, 0xcf, 0xcf, 0x4f, 0xcf, 0x49, 0xd5, 0x4f, 0x2c, 0xc8, 0xd4, 0x2f, 0xce, 0x2f, 0x2d, 0x4a, 0x4e, 0x8d, 0xcf, 0xcc, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/system_parameter.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/system_parameter.pb.go index 9a27bb2ee..c8b418ec3 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/system_parameter.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/system_parameter.pb.go @@ -54,7 +54,7 @@ type SystemParameters struct { // http_header: Api-Key2 // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*SystemParameterRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*SystemParameterRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -64,7 +64,7 @@ func (m *SystemParameters) Reset() { *m = SystemParameters{} } func (m *SystemParameters) String() string { return proto.CompactTextString(m) } func (*SystemParameters) ProtoMessage() {} func (*SystemParameters) Descriptor() ([]byte, []int) { - return fileDescriptor_system_parameter_11f19a50f5ca3984, []int{0} + return fileDescriptor_system_parameter_128d6f6e610b00a0, []int{0} } func (m *SystemParameters) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SystemParameters.Unmarshal(m, b) @@ -98,13 +98,13 @@ type SystemParameterRule struct { // methods in all APIs. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` // Define parameters. Multiple names may be defined for a parameter. // For a given method call, only one of them should be used. If multiple // names are used the behavior is implementation-dependent. // If none of the specified names are present the behavior is // parameter-dependent. - Parameters []*SystemParameter `protobuf:"bytes,2,rep,name=parameters" json:"parameters,omitempty"` + Parameters []*SystemParameter `protobuf:"bytes,2,rep,name=parameters,proto3" json:"parameters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -114,7 +114,7 @@ func (m *SystemParameterRule) Reset() { *m = SystemParameterRule{} } func (m *SystemParameterRule) String() string { return proto.CompactTextString(m) } func (*SystemParameterRule) ProtoMessage() {} func (*SystemParameterRule) Descriptor() ([]byte, []int) { - return fileDescriptor_system_parameter_11f19a50f5ca3984, []int{1} + return fileDescriptor_system_parameter_128d6f6e610b00a0, []int{1} } func (m *SystemParameterRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SystemParameterRule.Unmarshal(m, b) @@ -153,13 +153,13 @@ func (m *SystemParameterRule) GetParameters() []*SystemParameter { // is implementation-dependent. type SystemParameter struct { // Define the name of the parameter, such as "api_key" . It is case sensitive. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Define the HTTP header name to use for the parameter. It is case // insensitive. - HttpHeader string `protobuf:"bytes,2,opt,name=http_header,json=httpHeader" json:"http_header,omitempty"` + HttpHeader string `protobuf:"bytes,2,opt,name=http_header,json=httpHeader,proto3" json:"http_header,omitempty"` // Define the URL query parameter name to use for the parameter. It is case // sensitive. - UrlQueryParameter string `protobuf:"bytes,3,opt,name=url_query_parameter,json=urlQueryParameter" json:"url_query_parameter,omitempty"` + UrlQueryParameter string `protobuf:"bytes,3,opt,name=url_query_parameter,json=urlQueryParameter,proto3" json:"url_query_parameter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,7 +169,7 @@ func (m *SystemParameter) Reset() { *m = SystemParameter{} } func (m *SystemParameter) String() string { return proto.CompactTextString(m) } func (*SystemParameter) ProtoMessage() {} func (*SystemParameter) Descriptor() ([]byte, []int) { - return fileDescriptor_system_parameter_11f19a50f5ca3984, []int{2} + return fileDescriptor_system_parameter_128d6f6e610b00a0, []int{2} } func (m *SystemParameter) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SystemParameter.Unmarshal(m, b) @@ -217,10 +217,10 @@ func init() { } func init() { - proto.RegisterFile("google/api/system_parameter.proto", fileDescriptor_system_parameter_11f19a50f5ca3984) + proto.RegisterFile("google/api/system_parameter.proto", fileDescriptor_system_parameter_128d6f6e610b00a0) } -var fileDescriptor_system_parameter_11f19a50f5ca3984 = []byte{ +var fileDescriptor_system_parameter_128d6f6e610b00a0 = []byte{ // 286 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x7c, 0x91, 0xbf, 0x4e, 0xc3, 0x30, 0x10, 0x87, 0x95, 0xb6, 0x20, 0xb8, 0x4a, 0xfc, 0x71, 0x19, 0x22, 0x18, 0x5a, 0x3a, 0x75, 0x72, diff --git a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/usage.pb.go b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/usage.pb.go index 9643b6c13..4eb7f4bbb 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/usage.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/serviceconfig/usage.pb.go @@ -24,11 +24,11 @@ type Usage struct { // Requirements that must be satisfied before a consumer project can use the // service. Each requirement is of the form /; // for example 'serviceusage.googleapis.com/billing-enabled'. - Requirements []string `protobuf:"bytes,1,rep,name=requirements" json:"requirements,omitempty"` + Requirements []string `protobuf:"bytes,1,rep,name=requirements,proto3" json:"requirements,omitempty"` // A list of usage rules that apply to individual API methods. // // **NOTE:** All service configuration rules follow "last one wins" order. - Rules []*UsageRule `protobuf:"bytes,6,rep,name=rules" json:"rules,omitempty"` + Rules []*UsageRule `protobuf:"bytes,6,rep,name=rules,proto3" json:"rules,omitempty"` // The full resource name of a channel used for sending notifications to the // service producer. // @@ -37,7 +37,7 @@ type Usage struct { // channel. To use Google Cloud Pub/Sub as the channel, this must be the name // of a Cloud Pub/Sub topic that uses the Cloud Pub/Sub topic name format // documented in https://cloud.google.com/pubsub/docs/overview. - ProducerNotificationChannel string `protobuf:"bytes,7,opt,name=producer_notification_channel,json=producerNotificationChannel" json:"producer_notification_channel,omitempty"` + ProducerNotificationChannel string `protobuf:"bytes,7,opt,name=producer_notification_channel,json=producerNotificationChannel,proto3" json:"producer_notification_channel,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -47,7 +47,7 @@ func (m *Usage) Reset() { *m = Usage{} } func (m *Usage) String() string { return proto.CompactTextString(m) } func (*Usage) ProtoMessage() {} func (*Usage) Descriptor() ([]byte, []int) { - return fileDescriptor_usage_d16b4039e1638770, []int{0} + return fileDescriptor_usage_d5fdf432dee45e22, []int{0} } func (m *Usage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Usage.Unmarshal(m, b) @@ -118,12 +118,15 @@ type UsageRule struct { // methods in all APIs. // // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. - Selector string `protobuf:"bytes,1,opt,name=selector" json:"selector,omitempty"` - // True, if the method allows unregistered calls; false otherwise. - AllowUnregisteredCalls bool `protobuf:"varint,2,opt,name=allow_unregistered_calls,json=allowUnregisteredCalls" json:"allow_unregistered_calls,omitempty"` - // True, if the method should skip service control. If so, no control plane - // feature (like quota and billing) will be enabled. - SkipServiceControl bool `protobuf:"varint,3,opt,name=skip_service_control,json=skipServiceControl" json:"skip_service_control,omitempty"` + Selector string `protobuf:"bytes,1,opt,name=selector,proto3" json:"selector,omitempty"` + // If true, the selected method allows unregistered calls, e.g. calls + // that don't identify any user or application. + AllowUnregisteredCalls bool `protobuf:"varint,2,opt,name=allow_unregistered_calls,json=allowUnregisteredCalls,proto3" json:"allow_unregistered_calls,omitempty"` + // If true, the selected method should skip service control and the control + // plane features, such as quota and billing, will not be available. + // This flag is used by Google Cloud Endpoints to bypass checks for internal + // methods, such as service health check methods. + SkipServiceControl bool `protobuf:"varint,3,opt,name=skip_service_control,json=skipServiceControl,proto3" json:"skip_service_control,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -133,7 +136,7 @@ func (m *UsageRule) Reset() { *m = UsageRule{} } func (m *UsageRule) String() string { return proto.CompactTextString(m) } func (*UsageRule) ProtoMessage() {} func (*UsageRule) Descriptor() ([]byte, []int) { - return fileDescriptor_usage_d16b4039e1638770, []int{1} + return fileDescriptor_usage_d5fdf432dee45e22, []int{1} } func (m *UsageRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UsageRule.Unmarshal(m, b) @@ -179,9 +182,9 @@ func init() { proto.RegisterType((*UsageRule)(nil), "google.api.UsageRule") } -func init() { proto.RegisterFile("google/api/usage.proto", fileDescriptor_usage_d16b4039e1638770) } +func init() { proto.RegisterFile("google/api/usage.proto", fileDescriptor_usage_d5fdf432dee45e22) } -var fileDescriptor_usage_d16b4039e1638770 = []byte{ +var fileDescriptor_usage_d5fdf432dee45e22 = []byte{ // 331 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x54, 0x91, 0xc1, 0x4b, 0xfb, 0x30, 0x14, 0xc7, 0xe9, 0xf6, 0xdb, 0x7e, 0x5b, 0x14, 0x0f, 0x41, 0x47, 0x99, 0x0a, 0x65, 0xa7, 0x82, diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/check_error.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/check_error.pb.go index e71d7e30a..661fe019e 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/check_error.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/check_error.pb.go @@ -114,9 +114,9 @@ func (CheckError_Code) EnumDescriptor() ([]byte, []int) { // [google.api.servicecontrol.v1.CheckResponse.check_errors][google.api.servicecontrol.v1.CheckResponse.check_errors]. type CheckError struct { // The error code. - Code CheckError_Code `protobuf:"varint,1,opt,name=code,enum=google.api.servicecontrol.v1.CheckError_Code" json:"code,omitempty"` + Code CheckError_Code `protobuf:"varint,1,opt,name=code,proto3,enum=google.api.servicecontrol.v1.CheckError_Code" json:"code,omitempty"` // Free-form text providing details on the error cause of the error. - Detail string `protobuf:"bytes,2,opt,name=detail" json:"detail,omitempty"` + Detail string `protobuf:"bytes,2,opt,name=detail,proto3" json:"detail,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/distribution.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/distribution.pb.go index 32191c621..bbc13da08 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/distribution.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/distribution.pb.go @@ -28,19 +28,19 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // - a histogram of the values of the sample points type Distribution struct { // The total number of samples in the distribution. Must be >= 0. - Count int64 `protobuf:"varint,1,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,1,opt,name=count,proto3" json:"count,omitempty"` // The arithmetic mean of the samples in the distribution. If `count` is // zero then this field must be zero. - Mean float64 `protobuf:"fixed64,2,opt,name=mean" json:"mean,omitempty"` + Mean float64 `protobuf:"fixed64,2,opt,name=mean,proto3" json:"mean,omitempty"` // The minimum of the population of values. Ignored if `count` is zero. - Minimum float64 `protobuf:"fixed64,3,opt,name=minimum" json:"minimum,omitempty"` + Minimum float64 `protobuf:"fixed64,3,opt,name=minimum,proto3" json:"minimum,omitempty"` // The maximum of the population of values. Ignored if `count` is zero. - Maximum float64 `protobuf:"fixed64,4,opt,name=maximum" json:"maximum,omitempty"` + Maximum float64 `protobuf:"fixed64,4,opt,name=maximum,proto3" json:"maximum,omitempty"` // The sum of squared deviations from the mean: // Sum[i=1..count]((x_i - mean)^2) // where each x_i is a sample values. If `count` is zero then this field // must be zero, otherwise validation of the request fails. - SumOfSquaredDeviation float64 `protobuf:"fixed64,5,opt,name=sum_of_squared_deviation,json=sumOfSquaredDeviation" json:"sum_of_squared_deviation,omitempty"` + SumOfSquaredDeviation float64 `protobuf:"fixed64,5,opt,name=sum_of_squared_deviation,json=sumOfSquaredDeviation,proto3" json:"sum_of_squared_deviation,omitempty"` // The number of samples in each histogram bucket. `bucket_counts` are // optional. If present, they must sum to the `count` value. // @@ -52,7 +52,7 @@ type Distribution struct { // below for more details. // // Any suffix of trailing zeros may be omitted. - BucketCounts []int64 `protobuf:"varint,6,rep,packed,name=bucket_counts,json=bucketCounts" json:"bucket_counts,omitempty"` + BucketCounts []int64 `protobuf:"varint,6,rep,packed,name=bucket_counts,json=bucketCounts,proto3" json:"bucket_counts,omitempty"` // Defines the buckets in the histogram. `bucket_option` and `bucket_counts` // must be both set, or both unset. // @@ -116,13 +116,13 @@ type isDistribution_BucketOption interface { } type Distribution_LinearBuckets_ struct { - LinearBuckets *Distribution_LinearBuckets `protobuf:"bytes,7,opt,name=linear_buckets,json=linearBuckets,oneof"` + LinearBuckets *Distribution_LinearBuckets `protobuf:"bytes,7,opt,name=linear_buckets,json=linearBuckets,proto3,oneof"` } type Distribution_ExponentialBuckets_ struct { - ExponentialBuckets *Distribution_ExponentialBuckets `protobuf:"bytes,8,opt,name=exponential_buckets,json=exponentialBuckets,oneof"` + ExponentialBuckets *Distribution_ExponentialBuckets `protobuf:"bytes,8,opt,name=exponential_buckets,json=exponentialBuckets,proto3,oneof"` } type Distribution_ExplicitBuckets_ struct { - ExplicitBuckets *Distribution_ExplicitBuckets `protobuf:"bytes,9,opt,name=explicit_buckets,json=explicitBuckets,oneof"` + ExplicitBuckets *Distribution_ExplicitBuckets `protobuf:"bytes,9,opt,name=explicit_buckets,json=explicitBuckets,proto3,oneof"` } func (*Distribution_LinearBuckets_) isDistribution_BucketOption() {} @@ -297,16 +297,16 @@ type Distribution_LinearBuckets struct { // The number of finite buckets. With the underflow and overflow buckets, // the total number of buckets is `num_finite_buckets` + 2. // See comments on `bucket_options` for details. - NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets" json:"num_finite_buckets,omitempty"` + NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets,proto3" json:"num_finite_buckets,omitempty"` // The i'th linear bucket covers the interval // [offset + (i-1) * width, offset + i * width) // where i ranges from 1 to num_finite_buckets, inclusive. // Must be strictly positive. - Width float64 `protobuf:"fixed64,2,opt,name=width" json:"width,omitempty"` + Width float64 `protobuf:"fixed64,2,opt,name=width,proto3" json:"width,omitempty"` // The i'th linear bucket covers the interval // [offset + (i-1) * width, offset + i * width) // where i ranges from 1 to num_finite_buckets, inclusive. - Offset float64 `protobuf:"fixed64,3,opt,name=offset" json:"offset,omitempty"` + Offset float64 `protobuf:"fixed64,3,opt,name=offset,proto3" json:"offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -362,17 +362,17 @@ type Distribution_ExponentialBuckets struct { // The number of finite buckets. With the underflow and overflow buckets, // the total number of buckets is `num_finite_buckets` + 2. // See comments on `bucket_options` for details. - NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets" json:"num_finite_buckets,omitempty"` + NumFiniteBuckets int32 `protobuf:"varint,1,opt,name=num_finite_buckets,json=numFiniteBuckets,proto3" json:"num_finite_buckets,omitempty"` // The i'th exponential bucket covers the interval // [scale * growth_factor^(i-1), scale * growth_factor^i) // where i ranges from 1 to num_finite_buckets inclusive. // Must be larger than 1.0. - GrowthFactor float64 `protobuf:"fixed64,2,opt,name=growth_factor,json=growthFactor" json:"growth_factor,omitempty"` + GrowthFactor float64 `protobuf:"fixed64,2,opt,name=growth_factor,json=growthFactor,proto3" json:"growth_factor,omitempty"` // The i'th exponential bucket covers the interval // [scale * growth_factor^(i-1), scale * growth_factor^i) // where i ranges from 1 to num_finite_buckets inclusive. // Must be > 0. - Scale float64 `protobuf:"fixed64,3,opt,name=scale" json:"scale,omitempty"` + Scale float64 `protobuf:"fixed64,3,opt,name=scale,proto3" json:"scale,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -440,7 +440,7 @@ type Distribution_ExplicitBuckets struct { // i == 0 (underflow) -inf bound[i] // 0 < i < bound_size() bound[i-1] bound[i] // i == bound_size() (overflow) bound[i-1] +inf - Bounds []float64 `protobuf:"fixed64,1,rep,packed,name=bounds" json:"bounds,omitempty"` + Bounds []float64 `protobuf:"fixed64,1,rep,packed,name=bounds,proto3" json:"bounds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/log_entry.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/log_entry.pb.go index 9d439dd18..6817098f7 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/log_entry.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/log_entry.pb.go @@ -27,19 +27,19 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type LogEntry struct { // Required. The log to which this log entry belongs. Examples: `"syslog"`, // `"book_log"`. - Name string `protobuf:"bytes,10,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,10,opt,name=name,proto3" json:"name,omitempty"` // The time the event described by the log entry occurred. If // omitted, defaults to operation start time. - Timestamp *timestamp.Timestamp `protobuf:"bytes,11,opt,name=timestamp" json:"timestamp,omitempty"` + Timestamp *timestamp.Timestamp `protobuf:"bytes,11,opt,name=timestamp,proto3" json:"timestamp,omitempty"` // The severity of the log entry. The default value is // `LogSeverity.DEFAULT`. - Severity _type.LogSeverity `protobuf:"varint,12,opt,name=severity,enum=google.logging.type.LogSeverity" json:"severity,omitempty"` + Severity _type.LogSeverity `protobuf:"varint,12,opt,name=severity,proto3,enum=google.logging.type.LogSeverity" json:"severity,omitempty"` // A unique ID for the log entry used for deduplication. If omitted, // the implementation will generate one based on operation_id. - InsertId string `protobuf:"bytes,4,opt,name=insert_id,json=insertId" json:"insert_id,omitempty"` + InsertId string `protobuf:"bytes,4,opt,name=insert_id,json=insertId,proto3" json:"insert_id,omitempty"` // A set of user-defined (key, value) data that provides additional // information about the log entry. - Labels map[string]string `protobuf:"bytes,13,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,13,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The log entry payload, which can be one of multiple types. // // Types that are valid to be assigned to Payload: @@ -81,13 +81,13 @@ type isLogEntry_Payload interface { } type LogEntry_ProtoPayload struct { - ProtoPayload *any.Any `protobuf:"bytes,2,opt,name=proto_payload,json=protoPayload,oneof"` + ProtoPayload *any.Any `protobuf:"bytes,2,opt,name=proto_payload,json=protoPayload,proto3,oneof"` } type LogEntry_TextPayload struct { - TextPayload string `protobuf:"bytes,3,opt,name=text_payload,json=textPayload,oneof"` + TextPayload string `protobuf:"bytes,3,opt,name=text_payload,json=textPayload,proto3,oneof"` } type LogEntry_StructPayload struct { - StructPayload *_struct.Struct `protobuf:"bytes,6,opt,name=struct_payload,json=structPayload,oneof"` + StructPayload *_struct.Struct `protobuf:"bytes,6,opt,name=struct_payload,json=structPayload,proto3,oneof"` } func (*LogEntry_ProtoPayload) isLogEntry_Payload() {} diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/metric_value.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/metric_value.pb.go index 73c73d830..0945c20fe 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/metric_value.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/metric_value.pb.go @@ -26,15 +26,15 @@ type MetricValue struct { // The labels describing the metric value. // See comments on [google.api.servicecontrol.v1.Operation.labels][google.api.servicecontrol.v1.Operation.labels] for // the overriding relationship. - Labels map[string]string `protobuf:"bytes,1,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,1,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The start of the time period over which this metric value's measurement // applies. The time period has different semantics for different metric // types (cumulative, delta, and gauge). See the metric definition // documentation in the service configuration for details. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The end of the time period over which this metric value's measurement // applies. - EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // The value. The type of value used in the request must // agree with the metric definition in the service configuration, otherwise // the MetricValue is rejected. @@ -80,19 +80,19 @@ type isMetricValue_Value interface { } type MetricValue_BoolValue struct { - BoolValue bool `protobuf:"varint,4,opt,name=bool_value,json=boolValue,oneof"` + BoolValue bool `protobuf:"varint,4,opt,name=bool_value,json=boolValue,proto3,oneof"` } type MetricValue_Int64Value struct { - Int64Value int64 `protobuf:"varint,5,opt,name=int64_value,json=int64Value,oneof"` + Int64Value int64 `protobuf:"varint,5,opt,name=int64_value,json=int64Value,proto3,oneof"` } type MetricValue_DoubleValue struct { - DoubleValue float64 `protobuf:"fixed64,6,opt,name=double_value,json=doubleValue,oneof"` + DoubleValue float64 `protobuf:"fixed64,6,opt,name=double_value,json=doubleValue,proto3,oneof"` } type MetricValue_StringValue struct { - StringValue string `protobuf:"bytes,7,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,7,opt,name=string_value,json=stringValue,proto3,oneof"` } type MetricValue_DistributionValue struct { - DistributionValue *Distribution `protobuf:"bytes,8,opt,name=distribution_value,json=distributionValue,oneof"` + DistributionValue *Distribution `protobuf:"bytes,8,opt,name=distribution_value,json=distributionValue,proto3,oneof"` } func (*MetricValue_BoolValue) isMetricValue_Value() {} @@ -285,9 +285,9 @@ func _MetricValue_OneofSizer(msg proto.Message) (n int) { // end time, and label values. type MetricValueSet struct { // The metric name defined in the service configuration. - MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName" json:"metric_name,omitempty"` + MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName,proto3" json:"metric_name,omitempty"` // The values in this metric. - MetricValues []*MetricValue `protobuf:"bytes,2,rep,name=metric_values,json=metricValues" json:"metric_values,omitempty"` + MetricValues []*MetricValue `protobuf:"bytes,2,rep,name=metric_values,json=metricValues,proto3" json:"metric_values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/operation.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/operation.pb.go index c19d1889f..9a260f290 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/operation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/operation.pb.go @@ -60,9 +60,9 @@ type Operation struct { // In scenarios where an operation is computed from existing information // and an idempotent id is desirable for deduplication purpose, UUID version 5 // is recommended. See RFC 4122 for details. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // Fully qualified name of the operation. Reserved for future use. - OperationName string `protobuf:"bytes,2,opt,name=operation_name,json=operationName" json:"operation_name,omitempty"` + OperationName string `protobuf:"bytes,2,opt,name=operation_name,json=operationName,proto3" json:"operation_name,omitempty"` // Identity of the consumer who is using the service. // This field should be filled in for the operations initiated by a // consumer, but not for service-initiated operations that are @@ -72,13 +72,13 @@ type Operation struct { // project:, // project_number:, // api_key:. - ConsumerId string `protobuf:"bytes,3,opt,name=consumer_id,json=consumerId" json:"consumer_id,omitempty"` + ConsumerId string `protobuf:"bytes,3,opt,name=consumer_id,json=consumerId,proto3" json:"consumer_id,omitempty"` // Required. Start time of the operation. - StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End time of the operation. // Required when the operation is used in [ServiceController.Report][google.api.servicecontrol.v1.ServiceController.Report], // but optional when the operation is used in [ServiceController.Check][google.api.servicecontrol.v1.ServiceController.Check]. - EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Labels describing the operation. Only the following labels are allowed: // // - Labels describing monitored resources as defined in @@ -94,7 +94,7 @@ type Operation struct { // used to handle the API request (e.g. ESP), // - `servicecontrol.googleapis.com/platform` describing the platform // where the API is served (e.g. GAE, GCE, GKE). - Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,6,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Represents information about this operation. Each MetricValueSet // corresponds to a metric defined in the service configuration. // The data type used in the MetricValueSet must agree with @@ -105,11 +105,11 @@ type Operation struct { // label value combinations. If a request has such duplicated MetricValue // instances, the entire request is rejected with // an invalid argument error. - MetricValueSets []*MetricValueSet `protobuf:"bytes,7,rep,name=metric_value_sets,json=metricValueSets" json:"metric_value_sets,omitempty"` + MetricValueSets []*MetricValueSet `protobuf:"bytes,7,rep,name=metric_value_sets,json=metricValueSets,proto3" json:"metric_value_sets,omitempty"` // Represents information to be logged. - LogEntries []*LogEntry `protobuf:"bytes,8,rep,name=log_entries,json=logEntries" json:"log_entries,omitempty"` + LogEntries []*LogEntry `protobuf:"bytes,8,rep,name=log_entries,json=logEntries,proto3" json:"log_entries,omitempty"` // DO NOT USE. This is an experimental field. - Importance Operation_Importance `protobuf:"varint,11,opt,name=importance,enum=google.api.servicecontrol.v1.Operation_Importance" json:"importance,omitempty"` + Importance Operation_Importance `protobuf:"varint,11,opt,name=importance,proto3,enum=google.api.servicecontrol.v1.Operation_Importance" json:"importance,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/quota_controller.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/quota_controller.pb.go index 3391bd8a8..94f361441 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/quota_controller.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/quota_controller.pb.go @@ -120,13 +120,13 @@ type AllocateQuotaRequest struct { // `"pubsub.googleapis.com"`. // // See [google.api.Service][google.api.Service] for the definition of a service name. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // Operation that describes the quota allocation. - AllocateOperation *QuotaOperation `protobuf:"bytes,2,opt,name=allocate_operation,json=allocateOperation" json:"allocate_operation,omitempty"` + AllocateOperation *QuotaOperation `protobuf:"bytes,2,opt,name=allocate_operation,json=allocateOperation,proto3" json:"allocate_operation,omitempty"` // Specifies which version of service configuration should be used to process // the request. If unspecified or no matching version can be found, the latest // one will be used. - ServiceConfigId string `protobuf:"bytes,4,opt,name=service_config_id,json=serviceConfigId" json:"service_config_id,omitempty"` + ServiceConfigId string `protobuf:"bytes,4,opt,name=service_config_id,json=serviceConfigId,proto3" json:"service_config_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -187,7 +187,7 @@ type QuotaOperation struct { // operation is computed from existing information and an idempotent id is // desirable for deduplication purpose, UUID version 5 is recommended. See // RFC 4122 for details. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // Fully qualified name of the API method for which this quota operation is // requested. This name is used for matching quota rules or metric rules and // billing status rules defined in service configuration. This field is not @@ -195,16 +195,16 @@ type QuotaOperation struct { // // Example of an RPC method name: // google.example.library.v1.LibraryService.CreateShelf - MethodName string `protobuf:"bytes,2,opt,name=method_name,json=methodName" json:"method_name,omitempty"` + MethodName string `protobuf:"bytes,2,opt,name=method_name,json=methodName,proto3" json:"method_name,omitempty"` // Identity of the consumer for whom this quota operation is being performed. // // This can be in one of the following formats: // project:, // project_number:, // api_key:. - ConsumerId string `protobuf:"bytes,3,opt,name=consumer_id,json=consumerId" json:"consumer_id,omitempty"` + ConsumerId string `protobuf:"bytes,3,opt,name=consumer_id,json=consumerId,proto3" json:"consumer_id,omitempty"` // Labels describing the operation. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Represents information about this operation. Each MetricValueSet // corresponds to a metric defined in the service configuration. // The data type used in the MetricValueSet must agree with @@ -215,9 +215,9 @@ type QuotaOperation struct { // label value combinations. If a request has such duplicated MetricValue // instances, the entire request is rejected with // an invalid argument error. - QuotaMetrics []*MetricValueSet `protobuf:"bytes,5,rep,name=quota_metrics,json=quotaMetrics" json:"quota_metrics,omitempty"` + QuotaMetrics []*MetricValueSet `protobuf:"bytes,5,rep,name=quota_metrics,json=quotaMetrics,proto3" json:"quota_metrics,omitempty"` // Quota mode for this operation. - QuotaMode QuotaOperation_QuotaMode `protobuf:"varint,6,opt,name=quota_mode,json=quotaMode,enum=google.api.servicecontrol.v1.QuotaOperation_QuotaMode" json:"quota_mode,omitempty"` + QuotaMode QuotaOperation_QuotaMode `protobuf:"varint,6,opt,name=quota_mode,json=quotaMode,proto3,enum=google.api.servicecontrol.v1.QuotaOperation_QuotaMode" json:"quota_mode,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -293,9 +293,9 @@ func (m *QuotaOperation) GetQuotaMode() QuotaOperation_QuotaMode { type AllocateQuotaResponse struct { // The same operation_id value used in the AllocateQuotaRequest. Used for // logging and diagnostics purposes. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // Indicates the decision of the allocate. - AllocateErrors []*QuotaError `protobuf:"bytes,2,rep,name=allocate_errors,json=allocateErrors" json:"allocate_errors,omitempty"` + AllocateErrors []*QuotaError `protobuf:"bytes,2,rep,name=allocate_errors,json=allocateErrors,proto3" json:"allocate_errors,omitempty"` // Quota metrics to indicate the result of allocation. Depending on the // request, one or more of the following metrics will be included: // @@ -306,9 +306,9 @@ type AllocateQuotaResponse struct { // 2. The quota limit reached condition will be specified using the following // boolean metric : // "serviceruntime.googleapis.com/quota/exceeded" - QuotaMetrics []*MetricValueSet `protobuf:"bytes,3,rep,name=quota_metrics,json=quotaMetrics" json:"quota_metrics,omitempty"` + QuotaMetrics []*MetricValueSet `protobuf:"bytes,3,rep,name=quota_metrics,json=quotaMetrics,proto3" json:"quota_metrics,omitempty"` // ID of the actual config used to process the request. - ServiceConfigId string `protobuf:"bytes,4,opt,name=service_config_id,json=serviceConfigId" json:"service_config_id,omitempty"` + ServiceConfigId string `protobuf:"bytes,4,opt,name=service_config_id,json=serviceConfigId,proto3" json:"service_config_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -369,13 +369,13 @@ func (m *AllocateQuotaResponse) GetServiceConfigId() string { // Represents error information for [QuotaOperation][google.api.servicecontrol.v1.QuotaOperation]. type QuotaError struct { // Error code. - Code QuotaError_Code `protobuf:"varint,1,opt,name=code,enum=google.api.servicecontrol.v1.QuotaError_Code" json:"code,omitempty"` + Code QuotaError_Code `protobuf:"varint,1,opt,name=code,proto3,enum=google.api.servicecontrol.v1.QuotaError_Code" json:"code,omitempty"` // Subject to whom this error applies. See the specific enum for more details // on this field. For example, "clientip:" or // "project:". - Subject string `protobuf:"bytes,2,opt,name=subject" json:"subject,omitempty"` + Subject string `protobuf:"bytes,2,opt,name=subject,proto3" json:"subject,omitempty"` // Free-form text that provides details on the cause of the error. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -479,8 +479,7 @@ func (c *quotaControllerClient) AllocateQuota(ctx context.Context, in *AllocateQ return out, nil } -// Server API for QuotaController service - +// QuotaControllerServer is the server API for QuotaController service. type QuotaControllerServer interface { // Attempts to allocate quota for the specified consumer. It should be called // before the operation is executed. diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/service_controller.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/service_controller.pb.go index 3091f3053..57cbd37d2 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/service_controller.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicecontrol/v1/service_controller.pb.go @@ -33,15 +33,15 @@ type CheckRequest struct { // See // [google.api.Service](https://cloud.google.com/service-management/reference/rpc/google.api#google.api.Service) // for the definition of a service name. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The operation to be checked. - Operation *Operation `protobuf:"bytes,2,opt,name=operation" json:"operation,omitempty"` + Operation *Operation `protobuf:"bytes,2,opt,name=operation,proto3" json:"operation,omitempty"` // Specifies which version of service configuration should be used to process // the request. // // If unspecified or no matching version can be found, the // latest one will be used. - ServiceConfigId string `protobuf:"bytes,4,opt,name=service_config_id,json=serviceConfigId" json:"service_config_id,omitempty"` + ServiceConfigId string `protobuf:"bytes,4,opt,name=service_config_id,json=serviceConfigId,proto3" json:"service_config_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -96,17 +96,17 @@ func (m *CheckRequest) GetServiceConfigId() string { type CheckResponse struct { // The same operation_id value used in the [CheckRequest][google.api.servicecontrol.v1.CheckRequest]. // Used for logging and diagnostics purposes. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // Indicate the decision of the check. // // If no check errors are present, the service should process the operation. // Otherwise the service should use the list of errors to determine the // appropriate action. - CheckErrors []*CheckError `protobuf:"bytes,2,rep,name=check_errors,json=checkErrors" json:"check_errors,omitempty"` + CheckErrors []*CheckError `protobuf:"bytes,2,rep,name=check_errors,json=checkErrors,proto3" json:"check_errors,omitempty"` // The actual config id used to process the request. - ServiceConfigId string `protobuf:"bytes,5,opt,name=service_config_id,json=serviceConfigId" json:"service_config_id,omitempty"` + ServiceConfigId string `protobuf:"bytes,5,opt,name=service_config_id,json=serviceConfigId,proto3" json:"service_config_id,omitempty"` // Feedback data returned from the server during processing a Check request. - CheckInfo *CheckResponse_CheckInfo `protobuf:"bytes,6,opt,name=check_info,json=checkInfo" json:"check_info,omitempty"` + CheckInfo *CheckResponse_CheckInfo `protobuf:"bytes,6,opt,name=check_info,json=checkInfo,proto3" json:"check_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -166,7 +166,7 @@ func (m *CheckResponse) GetCheckInfo() *CheckResponse_CheckInfo { type CheckResponse_CheckInfo struct { // Consumer info of this check. - ConsumerInfo *CheckResponse_ConsumerInfo `protobuf:"bytes,2,opt,name=consumer_info,json=consumerInfo" json:"consumer_info,omitempty"` + ConsumerInfo *CheckResponse_ConsumerInfo `protobuf:"bytes,2,opt,name=consumer_info,json=consumerInfo,proto3" json:"consumer_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -207,7 +207,7 @@ func (m *CheckResponse_CheckInfo) GetConsumerInfo() *CheckResponse_ConsumerInfo type CheckResponse_ConsumerInfo struct { // The Google cloud project number, e.g. 1234567890. A value of 0 indicates // no project number is found. - ProjectNumber int64 `protobuf:"varint,1,opt,name=project_number,json=projectNumber" json:"project_number,omitempty"` + ProjectNumber int64 `protobuf:"varint,1,opt,name=project_number,json=projectNumber,proto3" json:"project_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -252,7 +252,7 @@ type ReportRequest struct { // See // [google.api.Service](https://cloud.google.com/service-management/reference/rpc/google.api#google.api.Service) // for the definition of a service name. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // Operations to be reported. // // Typically the service should report one operation per request. @@ -263,13 +263,13 @@ type ReportRequest struct { // If multiple operations are in a single request, the total request size // should be no larger than 1MB. See [ReportResponse.report_errors][google.api.servicecontrol.v1.ReportResponse.report_errors] for // partial failure behavior. - Operations []*Operation `protobuf:"bytes,2,rep,name=operations" json:"operations,omitempty"` + Operations []*Operation `protobuf:"bytes,2,rep,name=operations,proto3" json:"operations,omitempty"` // Specifies which version of service config should be used to process the // request. // // If unspecified or no matching version can be found, the // latest one will be used. - ServiceConfigId string `protobuf:"bytes,3,opt,name=service_config_id,json=serviceConfigId" json:"service_config_id,omitempty"` + ServiceConfigId string `protobuf:"bytes,3,opt,name=service_config_id,json=serviceConfigId,proto3" json:"service_config_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -336,9 +336,9 @@ type ReportResponse struct { // 3. A failed RPC status indicates a general non-deterministic failure. // When this happens, it's impossible to know which of the // 'Operations' in the request succeeded or failed. - ReportErrors []*ReportResponse_ReportError `protobuf:"bytes,1,rep,name=report_errors,json=reportErrors" json:"report_errors,omitempty"` + ReportErrors []*ReportResponse_ReportError `protobuf:"bytes,1,rep,name=report_errors,json=reportErrors,proto3" json:"report_errors,omitempty"` // The actual config id used to process the request. - ServiceConfigId string `protobuf:"bytes,2,opt,name=service_config_id,json=serviceConfigId" json:"service_config_id,omitempty"` + ServiceConfigId string `protobuf:"bytes,2,opt,name=service_config_id,json=serviceConfigId,proto3" json:"service_config_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -385,9 +385,9 @@ func (m *ReportResponse) GetServiceConfigId() string { // Represents the processing error of one [Operation][google.api.servicecontrol.v1.Operation] in the request. type ReportResponse_ReportError struct { // The [Operation.operation_id][google.api.servicecontrol.v1.Operation.operation_id] value from the request. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // Details of the error when processing the [Operation][google.api.servicecontrol.v1.Operation]. - Status *status.Status `protobuf:"bytes,2,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -510,8 +510,7 @@ func (c *serviceControllerClient) Report(ctx context.Context, in *ReportRequest, return out, nil } -// Server API for ServiceController service - +// ServiceControllerServer is the server API for ServiceController service. type ServiceControllerServer interface { // Checks an operation with Google Service Control to decide whether // the given operation should proceed. It should be called before the diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/resources.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/resources.pb.go index 158bc8295..7032f4c46 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/resources.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/resources.pb.go @@ -201,9 +201,9 @@ func (Rollout_RolloutStatus) EnumDescriptor() ([]byte, []int) { type ManagedService struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. - ServiceName string `protobuf:"bytes,2,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,2,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // ID of the project that produces and owns this service. - ProducerProjectId string `protobuf:"bytes,3,opt,name=producer_project_id,json=producerProjectId" json:"producer_project_id,omitempty"` + ProducerProjectId string `protobuf:"bytes,3,opt,name=producer_project_id,json=producerProjectId,proto3" json:"producer_project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -251,13 +251,13 @@ func (m *ManagedService) GetProducerProjectId() string { type OperationMetadata struct { // The full name of the resources that this operation is directly // associated with. - ResourceNames []string `protobuf:"bytes,1,rep,name=resource_names,json=resourceNames" json:"resource_names,omitempty"` + ResourceNames []string `protobuf:"bytes,1,rep,name=resource_names,json=resourceNames,proto3" json:"resource_names,omitempty"` // Detailed status information for each step. The order is undetermined. - Steps []*OperationMetadata_Step `protobuf:"bytes,2,rep,name=steps" json:"steps,omitempty"` + Steps []*OperationMetadata_Step `protobuf:"bytes,2,rep,name=steps,proto3" json:"steps,omitempty"` // Percentage of completion of this operation, ranging from 0 to 100. - ProgressPercentage int32 `protobuf:"varint,3,opt,name=progress_percentage,json=progressPercentage" json:"progress_percentage,omitempty"` + ProgressPercentage int32 `protobuf:"varint,3,opt,name=progress_percentage,json=progressPercentage,proto3" json:"progress_percentage,omitempty"` // The start time of the operation. - StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -318,9 +318,9 @@ func (m *OperationMetadata) GetStartTime() *timestamp.Timestamp { // Represents the status of one operation step. type OperationMetadata_Step struct { // The short description of the step. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // The status code. - Status OperationMetadata_Status `protobuf:"varint,4,opt,name=status,enum=google.api.servicemanagement.v1.OperationMetadata_Status" json:"status,omitempty"` + Status OperationMetadata_Status `protobuf:"varint,4,opt,name=status,proto3,enum=google.api.servicemanagement.v1.OperationMetadata_Status" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -367,11 +367,11 @@ func (m *OperationMetadata_Step) GetStatus() OperationMetadata_Status { // Represents a diagnostic message (error or warning) type Diagnostic struct { // File name and line number of the error or warning. - Location string `protobuf:"bytes,1,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,1,opt,name=location,proto3" json:"location,omitempty"` // The kind of diagnostic information provided. - Kind Diagnostic_Kind `protobuf:"varint,2,opt,name=kind,enum=google.api.servicemanagement.v1.Diagnostic_Kind" json:"kind,omitempty"` + Kind Diagnostic_Kind `protobuf:"varint,2,opt,name=kind,proto3,enum=google.api.servicemanagement.v1.Diagnostic_Kind" json:"kind,omitempty"` // Message describing the error or warning. - Message string `protobuf:"bytes,3,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,3,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -428,10 +428,10 @@ type ConfigSource struct { // A unique ID for a specific instance of this message, typically assigned // by the client for tracking purpose. If empty, the server may choose to // generate one instead. - Id string `protobuf:"bytes,5,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,5,opt,name=id,proto3" json:"id,omitempty"` // Set of source configuration files that are used to generate a service // configuration (`google.api.Service`). - Files []*ConfigFile `protobuf:"bytes,2,rep,name=files" json:"files,omitempty"` + Files []*ConfigFile `protobuf:"bytes,2,rep,name=files,proto3" json:"files,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -478,11 +478,11 @@ func (m *ConfigSource) GetFiles() []*ConfigFile { // Generic specification of a source configuration file type ConfigFile struct { // The file name of the configuration file (full or relative path). - FilePath string `protobuf:"bytes,1,opt,name=file_path,json=filePath" json:"file_path,omitempty"` + FilePath string `protobuf:"bytes,1,opt,name=file_path,json=filePath,proto3" json:"file_path,omitempty"` // The bytes that constitute the file. FileContents []byte `protobuf:"bytes,3,opt,name=file_contents,json=fileContents,proto3" json:"file_contents,omitempty"` // The type of configuration file this represents. - FileType ConfigFile_FileType `protobuf:"varint,4,opt,name=file_type,json=fileType,enum=google.api.servicemanagement.v1.ConfigFile_FileType" json:"file_type,omitempty"` + FileType ConfigFile_FileType `protobuf:"varint,4,opt,name=file_type,json=fileType,proto3,enum=google.api.servicemanagement.v1.ConfigFile_FileType" json:"file_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -537,7 +537,7 @@ func (m *ConfigFile) GetFileType() ConfigFile_FileType { type ConfigRef struct { // Resource name of a service config. It must have the following // format: "services/{service name}/configs/{config id}". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -584,7 +584,7 @@ type ChangeReport struct { // of each change. // A ConfigChange identifier is a dot separated path to the configuration. // Example: visibility.rules[selector='LibraryService.CreateBook'].restriction - ConfigChanges []*configchange.ConfigChange `protobuf:"bytes,1,rep,name=config_changes,json=configChanges" json:"config_changes,omitempty"` + ConfigChanges []*configchange.ConfigChange `protobuf:"bytes,1,rep,name=config_changes,json=configChanges,proto3" json:"config_changes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -633,15 +633,15 @@ type Rollout struct { // date in ISO 8601 format. "revision number" is a monotonically increasing // positive number that is reset every day for each service. // An example of the generated rollout_id is '2016-02-16r1' - RolloutId string `protobuf:"bytes,1,opt,name=rollout_id,json=rolloutId" json:"rollout_id,omitempty"` + RolloutId string `protobuf:"bytes,1,opt,name=rollout_id,json=rolloutId,proto3" json:"rollout_id,omitempty"` // Creation time of the rollout. Readonly. - CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The user who created the Rollout. Readonly. - CreatedBy string `protobuf:"bytes,3,opt,name=created_by,json=createdBy" json:"created_by,omitempty"` + CreatedBy string `protobuf:"bytes,3,opt,name=created_by,json=createdBy,proto3" json:"created_by,omitempty"` // The status of this rollout. Readonly. In case of a failed rollout, // the system will automatically rollback to the current Rollout // version. Readonly. - Status Rollout_RolloutStatus `protobuf:"varint,4,opt,name=status,enum=google.api.servicemanagement.v1.Rollout_RolloutStatus" json:"status,omitempty"` + Status Rollout_RolloutStatus `protobuf:"varint,4,opt,name=status,proto3,enum=google.api.servicemanagement.v1.Rollout_RolloutStatus" json:"status,omitempty"` // Strategy that defines which versions of service configurations should be // pushed // and how they should be used at runtime. @@ -651,7 +651,7 @@ type Rollout struct { // *Rollout_DeleteServiceStrategy_ Strategy isRollout_Strategy `protobuf_oneof:"strategy"` // The name of the service associated with this Rollout. - ServiceName string `protobuf:"bytes,8,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,8,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -686,10 +686,10 @@ type isRollout_Strategy interface { } type Rollout_TrafficPercentStrategy_ struct { - TrafficPercentStrategy *Rollout_TrafficPercentStrategy `protobuf:"bytes,5,opt,name=traffic_percent_strategy,json=trafficPercentStrategy,oneof"` + TrafficPercentStrategy *Rollout_TrafficPercentStrategy `protobuf:"bytes,5,opt,name=traffic_percent_strategy,json=trafficPercentStrategy,proto3,oneof"` } type Rollout_DeleteServiceStrategy_ struct { - DeleteServiceStrategy *Rollout_DeleteServiceStrategy `protobuf:"bytes,200,opt,name=delete_service_strategy,json=deleteServiceStrategy,oneof"` + DeleteServiceStrategy *Rollout_DeleteServiceStrategy `protobuf:"bytes,200,opt,name=delete_service_strategy,json=deleteServiceStrategy,proto3,oneof"` } func (*Rollout_TrafficPercentStrategy_) isRollout_Strategy() {} @@ -859,7 +859,7 @@ type Rollout_TrafficPercentStrategy struct { // Maps service configuration IDs to their corresponding traffic percentage. // Key is the service configuration ID, Value is the traffic percentage // which must be greater than 0.0 and the sum must equal to 100.0. - Percentages map[string]float64 `protobuf:"bytes,1,rep,name=percentages" json:"percentages,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"fixed64,2,opt,name=value"` + Percentages map[string]float64 `protobuf:"bytes,1,rep,name=percentages,proto3" json:"percentages,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"fixed64,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/servicemanager.pb.go b/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/servicemanager.pb.go index e70ed9a2e..101b8ecbc 100644 --- a/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/servicemanager.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/api/servicemanagement/v1/servicemanager.pb.go @@ -61,18 +61,18 @@ func (GetServiceConfigRequest_ConfigView) EnumDescriptor() ([]byte, []int) { // Request message for `ListServices` method. type ListServicesRequest struct { // Include services produced by the specified project. - ProducerProjectId string `protobuf:"bytes,1,opt,name=producer_project_id,json=producerProjectId" json:"producer_project_id,omitempty"` + ProducerProjectId string `protobuf:"bytes,1,opt,name=producer_project_id,json=producerProjectId,proto3" json:"producer_project_id,omitempty"` // Requested size of the next page of data. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token identifying which result to start with; returned by a previous list // call. - PageToken string `protobuf:"bytes,6,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,6,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Include services consumed by the specified consumer. // // The Google Service Management implementation accepts the following // forms: // - project: - ConsumerId string `protobuf:"bytes,7,opt,name=consumer_id,json=consumerId" json:"consumer_id,omitempty"` + ConsumerId string `protobuf:"bytes,7,opt,name=consumer_id,json=consumerId,proto3" json:"consumer_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -133,9 +133,9 @@ func (m *ListServicesRequest) GetConsumerId() string { // Response message for `ListServices` method. type ListServicesResponse struct { // The returned services will only have the name field set. - Services []*ManagedService `protobuf:"bytes,1,rep,name=services" json:"services,omitempty"` + Services []*ManagedService `protobuf:"bytes,1,rep,name=services,proto3" json:"services,omitempty"` // Token that can be passed to `ListServices` to resume a paginated query. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -183,7 +183,7 @@ func (m *ListServicesResponse) GetNextPageToken() string { type GetServiceRequest struct { // The name of the service. See the `ServiceManager` overview for naming // requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -223,7 +223,7 @@ func (m *GetServiceRequest) GetServiceName() string { // Request message for CreateService method. type CreateServiceRequest struct { // Initial values for the service resource. - Service *ManagedService `protobuf:"bytes,1,opt,name=service" json:"service,omitempty"` + Service *ManagedService `protobuf:"bytes,1,opt,name=service,proto3" json:"service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -264,7 +264,7 @@ func (m *CreateServiceRequest) GetService() *ManagedService { type DeleteServiceRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -305,7 +305,7 @@ func (m *DeleteServiceRequest) GetServiceName() string { type UndeleteServiceRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -345,7 +345,7 @@ func (m *UndeleteServiceRequest) GetServiceName() string { // Response message for UndeleteService method. type UndeleteServiceResponse struct { // Revived service resource. - Service *ManagedService `protobuf:"bytes,1,opt,name=service" json:"service,omitempty"` + Service *ManagedService `protobuf:"bytes,1,opt,name=service,proto3" json:"service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -386,12 +386,12 @@ func (m *UndeleteServiceResponse) GetService() *ManagedService { type GetServiceConfigRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The id of the service configuration resource. - ConfigId string `protobuf:"bytes,2,opt,name=config_id,json=configId" json:"config_id,omitempty"` + ConfigId string `protobuf:"bytes,2,opt,name=config_id,json=configId,proto3" json:"config_id,omitempty"` // Specifies which parts of the Service Config should be returned in the // response. - View GetServiceConfigRequest_ConfigView `protobuf:"varint,3,opt,name=view,enum=google.api.servicemanagement.v1.GetServiceConfigRequest_ConfigView" json:"view,omitempty"` + View GetServiceConfigRequest_ConfigView `protobuf:"varint,3,opt,name=view,proto3,enum=google.api.servicemanagement.v1.GetServiceConfigRequest_ConfigView" json:"view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -446,11 +446,11 @@ func (m *GetServiceConfigRequest) GetView() GetServiceConfigRequest_ConfigView { type ListServiceConfigsRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The token of the page to retrieve. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The max number of items to include in the response list. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -504,9 +504,9 @@ func (m *ListServiceConfigsRequest) GetPageSize() int32 { // Response message for ListServiceConfigs method. type ListServiceConfigsResponse struct { // The list of service configuration resources. - ServiceConfigs []*serviceconfig.Service `protobuf:"bytes,1,rep,name=service_configs,json=serviceConfigs" json:"service_configs,omitempty"` + ServiceConfigs []*serviceconfig.Service `protobuf:"bytes,1,rep,name=service_configs,json=serviceConfigs,proto3" json:"service_configs,omitempty"` // The token of the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -554,9 +554,9 @@ func (m *ListServiceConfigsResponse) GetNextPageToken() string { type CreateServiceConfigRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The service configuration resource. - ServiceConfig *serviceconfig.Service `protobuf:"bytes,2,opt,name=service_config,json=serviceConfig" json:"service_config,omitempty"` + ServiceConfig *serviceconfig.Service `protobuf:"bytes,2,opt,name=service_config,json=serviceConfig,proto3" json:"service_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -604,13 +604,13 @@ func (m *CreateServiceConfigRequest) GetServiceConfig() *serviceconfig.Service { type SubmitConfigSourceRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The source configuration for the service. - ConfigSource *ConfigSource `protobuf:"bytes,2,opt,name=config_source,json=configSource" json:"config_source,omitempty"` + ConfigSource *ConfigSource `protobuf:"bytes,2,opt,name=config_source,json=configSource,proto3" json:"config_source,omitempty"` // Optional. If set, this will result in the generation of a // `google.api.Service` configuration based on the `ConfigSource` provided, // but the generated config and the sources will NOT be persisted. - ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly" json:"validate_only,omitempty"` + ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly,proto3" json:"validate_only,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -664,7 +664,7 @@ func (m *SubmitConfigSourceRequest) GetValidateOnly() bool { // Response message for SubmitConfigSource method. type SubmitConfigSourceResponse struct { // The generated service configuration. - ServiceConfig *serviceconfig.Service `protobuf:"bytes,1,opt,name=service_config,json=serviceConfig" json:"service_config,omitempty"` + ServiceConfig *serviceconfig.Service `protobuf:"bytes,1,opt,name=service_config,json=serviceConfig,proto3" json:"service_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -705,9 +705,9 @@ func (m *SubmitConfigSourceResponse) GetServiceConfig() *serviceconfig.Service { type CreateServiceRolloutRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The rollout resource. The `service_name` field is output only. - Rollout *Rollout `protobuf:"bytes,2,opt,name=rollout" json:"rollout,omitempty"` + Rollout *Rollout `protobuf:"bytes,2,opt,name=rollout,proto3" json:"rollout,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -755,11 +755,11 @@ func (m *CreateServiceRolloutRequest) GetRollout() *Rollout { type ListServiceRolloutsRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The token of the page to retrieve. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The max number of items to include in the response list. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Use `filter` to return subset of rollouts. // The following filters are supported: // -- To limit the results to only those in @@ -768,7 +768,7 @@ type ListServiceRolloutsRequest struct { // -- To limit the results to those in // [status](google.api.servicemanagement.v1.RolloutStatus) 'CANCELLED' // or 'FAILED', use filter='status=CANCELLED OR status=FAILED' - Filter string `protobuf:"bytes,4,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,4,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -829,9 +829,9 @@ func (m *ListServiceRolloutsRequest) GetFilter() string { // Response message for ListServiceRollouts method. type ListServiceRolloutsResponse struct { // The list of rollout resources. - Rollouts []*Rollout `protobuf:"bytes,1,rep,name=rollouts" json:"rollouts,omitempty"` + Rollouts []*Rollout `protobuf:"bytes,1,rep,name=rollouts,proto3" json:"rollouts,omitempty"` // The token of the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -879,9 +879,9 @@ func (m *ListServiceRolloutsResponse) GetNextPageToken() string { type GetServiceRolloutRequest struct { // The name of the service. See the [overview](/service-management/overview) // for naming requirements. For example: `example.googleapis.com`. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The id of the rollout resource. - RolloutId string `protobuf:"bytes,2,opt,name=rollout_id,json=rolloutId" json:"rollout_id,omitempty"` + RolloutId string `protobuf:"bytes,2,opt,name=rollout_id,json=rolloutId,proto3" json:"rollout_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -929,7 +929,7 @@ func (m *GetServiceRolloutRequest) GetRolloutId() string { type EnableServiceRequest struct { // Name of the service to enable. Specifying an unknown service name will // cause the request to fail. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The identity of consumer resource which service enablement will be // applied to. // @@ -939,7 +939,7 @@ type EnableServiceRequest struct { // // Note: this is made compatible with // google.api.servicecontrol.v1.Operation.consumer_id. - ConsumerId string `protobuf:"bytes,2,opt,name=consumer_id,json=consumerId" json:"consumer_id,omitempty"` + ConsumerId string `protobuf:"bytes,2,opt,name=consumer_id,json=consumerId,proto3" json:"consumer_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -987,7 +987,7 @@ func (m *EnableServiceRequest) GetConsumerId() string { type DisableServiceRequest struct { // Name of the service to disable. Specifying an unknown service name // will cause the request to fail. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The identity of consumer resource which service disablement will be // applied to. // @@ -997,7 +997,7 @@ type DisableServiceRequest struct { // // Note: this is made compatible with // google.api.servicecontrol.v1.Operation.consumer_id. - ConsumerId string `protobuf:"bytes,2,opt,name=consumer_id,json=consumerId" json:"consumer_id,omitempty"` + ConsumerId string `protobuf:"bytes,2,opt,name=consumer_id,json=consumerId,proto3" json:"consumer_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1048,13 +1048,13 @@ type GenerateConfigReportRequest struct { // [google.api.servicemanagement.v1.ConfigRef][google.api.servicemanagement.v1.ConfigRef], // [google.api.servicemanagement.v1.ConfigSource][google.api.servicemanagement.v1.ConfigSource], // and [google.api.Service][google.api.Service] - NewConfig *any.Any `protobuf:"bytes,1,opt,name=new_config,json=newConfig" json:"new_config,omitempty"` + NewConfig *any.Any `protobuf:"bytes,1,opt,name=new_config,json=newConfig,proto3" json:"new_config,omitempty"` // Service configuration against which the comparison will be done. // For this version of API, the supported types are // [google.api.servicemanagement.v1.ConfigRef][google.api.servicemanagement.v1.ConfigRef], // [google.api.servicemanagement.v1.ConfigSource][google.api.servicemanagement.v1.ConfigSource], // and [google.api.Service][google.api.Service] - OldConfig *any.Any `protobuf:"bytes,2,opt,name=old_config,json=oldConfig" json:"old_config,omitempty"` + OldConfig *any.Any `protobuf:"bytes,2,opt,name=old_config,json=oldConfig,proto3" json:"old_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1101,16 +1101,16 @@ func (m *GenerateConfigReportRequest) GetOldConfig() *any.Any { // Response message for GenerateConfigReport method. type GenerateConfigReportResponse struct { // Name of the service this report belongs to. - ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,1,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // ID of the service configuration this report belongs to. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // list of ChangeReport, each corresponding to comparison between two // service configurations. - ChangeReports []*ChangeReport `protobuf:"bytes,3,rep,name=change_reports,json=changeReports" json:"change_reports,omitempty"` + ChangeReports []*ChangeReport `protobuf:"bytes,3,rep,name=change_reports,json=changeReports,proto3" json:"change_reports,omitempty"` // Errors / Linter warnings associated with the service definition this // report // belongs to. - Diagnostics []*Diagnostic `protobuf:"bytes,4,rep,name=diagnostics" json:"diagnostics,omitempty"` + Diagnostics []*Diagnostic `protobuf:"bytes,4,rep,name=diagnostics,proto3" json:"diagnostics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1455,8 +1455,7 @@ func (c *serviceManagerClient) DisableService(ctx context.Context, in *DisableSe return out, nil } -// Server API for ServiceManager service - +// ServiceManagerServer is the server API for ServiceManager service. type ServiceManagerServer interface { // Lists managed services. // diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/legacy/audit_data.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/legacy/audit_data.pb.go index 38397d653..c110f1668 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/legacy/audit_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/legacy/audit_data.pb.go @@ -22,10 +22,10 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type AuditData struct { // Text description of the admin event. // This is the "Event" column in Admin Console's Admin Logs. - EventMessage string `protobuf:"bytes,1,opt,name=event_message,json=eventMessage" json:"event_message,omitempty"` + EventMessage string `protobuf:"bytes,1,opt,name=event_message,json=eventMessage,proto3" json:"event_message,omitempty"` // Arbitrary event data. // This is the "Result" column in Admin Console's Admin Logs. - EventData map[string]string `protobuf:"bytes,2,rep,name=event_data,json=eventData" json:"event_data,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + EventData map[string]string `protobuf:"bytes,2,rep,name=event_data,json=eventData,proto3" json:"event_data,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/logging/v1/request_log.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/logging/v1/request_log.pb.go index 464bec39b..b52186c51 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/logging/v1/request_log.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/logging/v1/request_log.pb.go @@ -24,13 +24,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Application log line emitted while processing a request. type LogLine struct { // Approximate time when this log entry was made. - Time *timestamp.Timestamp `protobuf:"bytes,1,opt,name=time" json:"time,omitempty"` + Time *timestamp.Timestamp `protobuf:"bytes,1,opt,name=time,proto3" json:"time,omitempty"` // Severity of this log entry. - Severity _type.LogSeverity `protobuf:"varint,2,opt,name=severity,enum=google.logging.type.LogSeverity" json:"severity,omitempty"` + Severity _type.LogSeverity `protobuf:"varint,2,opt,name=severity,proto3,enum=google.logging.type.LogSeverity" json:"severity,omitempty"` // App-provided log message. - LogMessage string `protobuf:"bytes,3,opt,name=log_message,json=logMessage" json:"log_message,omitempty"` + LogMessage string `protobuf:"bytes,3,opt,name=log_message,json=logMessage,proto3" json:"log_message,omitempty"` // Where in the source code this log message was written. - SourceLocation *SourceLocation `protobuf:"bytes,4,opt,name=source_location,json=sourceLocation" json:"source_location,omitempty"` + SourceLocation *SourceLocation `protobuf:"bytes,4,opt,name=source_location,json=sourceLocation,proto3" json:"source_location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -92,16 +92,16 @@ func (m *LogLine) GetSourceLocation() *SourceLocation { type SourceLocation struct { // Source file name. Depending on the runtime environment, this might be a // simple name or a fully-qualified name. - File string `protobuf:"bytes,1,opt,name=file" json:"file,omitempty"` + File string `protobuf:"bytes,1,opt,name=file,proto3" json:"file,omitempty"` // Line within the source file. - Line int64 `protobuf:"varint,2,opt,name=line" json:"line,omitempty"` + Line int64 `protobuf:"varint,2,opt,name=line,proto3" json:"line,omitempty"` // Human-readable name of the function or method being invoked, with optional // context such as the class or package name. This information is used in // contexts such as the logs viewer, where a file and line number are less // meaningful. The format can vary by language. For example: // `qual.if.ied.Class.method` (Java), `dir/package.func` (Go), `function` // (Python). - FunctionName string `protobuf:"bytes,3,opt,name=function_name,json=functionName" json:"function_name,omitempty"` + FunctionName string `protobuf:"bytes,3,opt,name=function_name,json=functionName,proto3" json:"function_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -157,10 +157,10 @@ func (m *SourceLocation) GetFunctionName() string { type SourceReference struct { // Optional. A URI string identifying the repository. // Example: "https://github.com/GoogleCloudPlatform/kubernetes.git" - Repository string `protobuf:"bytes,1,opt,name=repository" json:"repository,omitempty"` + Repository string `protobuf:"bytes,1,opt,name=repository,proto3" json:"repository,omitempty"` // The canonical and persistent identifier of the deployed revision. // Example (git): "0035781c50ec7aa23385dc841529ce8a4b70db1b" - RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId" json:"revision_id,omitempty"` + RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,proto3" json:"revision_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -208,42 +208,42 @@ func (m *SourceReference) GetRevisionId() string { // application. type RequestLog struct { // Application that handled this request. - AppId string `protobuf:"bytes,1,opt,name=app_id,json=appId" json:"app_id,omitempty"` + AppId string `protobuf:"bytes,1,opt,name=app_id,json=appId,proto3" json:"app_id,omitempty"` // Module of the application that handled this request. - ModuleId string `protobuf:"bytes,37,opt,name=module_id,json=moduleId" json:"module_id,omitempty"` + ModuleId string `protobuf:"bytes,37,opt,name=module_id,json=moduleId,proto3" json:"module_id,omitempty"` // Version of the application that handled this request. - VersionId string `protobuf:"bytes,2,opt,name=version_id,json=versionId" json:"version_id,omitempty"` + VersionId string `protobuf:"bytes,2,opt,name=version_id,json=versionId,proto3" json:"version_id,omitempty"` // Globally unique identifier for a request, which is based on the request // start time. Request IDs for requests which started later will compare // greater as strings than those for requests which started earlier. - RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` // Origin IP address. - Ip string `protobuf:"bytes,4,opt,name=ip" json:"ip,omitempty"` + Ip string `protobuf:"bytes,4,opt,name=ip,proto3" json:"ip,omitempty"` // Time when the request started. - StartTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time when the request finished. - EndTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Latency of the request. - Latency *duration.Duration `protobuf:"bytes,8,opt,name=latency" json:"latency,omitempty"` + Latency *duration.Duration `protobuf:"bytes,8,opt,name=latency,proto3" json:"latency,omitempty"` // Number of CPU megacycles used to process request. - MegaCycles int64 `protobuf:"varint,9,opt,name=mega_cycles,json=megaCycles" json:"mega_cycles,omitempty"` + MegaCycles int64 `protobuf:"varint,9,opt,name=mega_cycles,json=megaCycles,proto3" json:"mega_cycles,omitempty"` // Request method. Example: `"GET"`, `"HEAD"`, `"PUT"`, `"POST"`, `"DELETE"`. - Method string `protobuf:"bytes,10,opt,name=method" json:"method,omitempty"` + Method string `protobuf:"bytes,10,opt,name=method,proto3" json:"method,omitempty"` // Contains the path and query portion of the URL that was requested. For // example, if the URL was "http://example.com/app?name=val", the resource // would be "/app?name=val". The fragment identifier, which is identified by // the `#` character, is not included. - Resource string `protobuf:"bytes,11,opt,name=resource" json:"resource,omitempty"` + Resource string `protobuf:"bytes,11,opt,name=resource,proto3" json:"resource,omitempty"` // HTTP version of request. Example: `"HTTP/1.1"`. - HttpVersion string `protobuf:"bytes,12,opt,name=http_version,json=httpVersion" json:"http_version,omitempty"` + HttpVersion string `protobuf:"bytes,12,opt,name=http_version,json=httpVersion,proto3" json:"http_version,omitempty"` // HTTP response status code. Example: 200, 404. - Status int32 `protobuf:"varint,13,opt,name=status" json:"status,omitempty"` + Status int32 `protobuf:"varint,13,opt,name=status,proto3" json:"status,omitempty"` // Size in bytes sent back to client by request. - ResponseSize int64 `protobuf:"varint,14,opt,name=response_size,json=responseSize" json:"response_size,omitempty"` + ResponseSize int64 `protobuf:"varint,14,opt,name=response_size,json=responseSize,proto3" json:"response_size,omitempty"` // Referrer URL of request. - Referrer string `protobuf:"bytes,15,opt,name=referrer" json:"referrer,omitempty"` + Referrer string `protobuf:"bytes,15,opt,name=referrer,proto3" json:"referrer,omitempty"` // User agent that made the request. - UserAgent string `protobuf:"bytes,16,opt,name=user_agent,json=userAgent" json:"user_agent,omitempty"` + UserAgent string `protobuf:"bytes,16,opt,name=user_agent,json=userAgent,proto3" json:"user_agent,omitempty"` // The logged-in user who made the request. // // Most likely, this is the part of the user's email before the `@` sign. The @@ -252,43 +252,43 @@ type RequestLog struct { // available to the application via the App Engine Users API. // // This field will be populated starting with App Engine 1.9.21. - Nickname string `protobuf:"bytes,40,opt,name=nickname" json:"nickname,omitempty"` + Nickname string `protobuf:"bytes,40,opt,name=nickname,proto3" json:"nickname,omitempty"` // File or class that handled the request. - UrlMapEntry string `protobuf:"bytes,17,opt,name=url_map_entry,json=urlMapEntry" json:"url_map_entry,omitempty"` + UrlMapEntry string `protobuf:"bytes,17,opt,name=url_map_entry,json=urlMapEntry,proto3" json:"url_map_entry,omitempty"` // Internet host and port number of the resource being requested. - Host string `protobuf:"bytes,20,opt,name=host" json:"host,omitempty"` + Host string `protobuf:"bytes,20,opt,name=host,proto3" json:"host,omitempty"` // An indication of the relative cost of serving this request. - Cost float64 `protobuf:"fixed64,21,opt,name=cost" json:"cost,omitempty"` + Cost float64 `protobuf:"fixed64,21,opt,name=cost,proto3" json:"cost,omitempty"` // Queue name of the request, in the case of an offline request. - TaskQueueName string `protobuf:"bytes,22,opt,name=task_queue_name,json=taskQueueName" json:"task_queue_name,omitempty"` + TaskQueueName string `protobuf:"bytes,22,opt,name=task_queue_name,json=taskQueueName,proto3" json:"task_queue_name,omitempty"` // Task name of the request, in the case of an offline request. - TaskName string `protobuf:"bytes,23,opt,name=task_name,json=taskName" json:"task_name,omitempty"` + TaskName string `protobuf:"bytes,23,opt,name=task_name,json=taskName,proto3" json:"task_name,omitempty"` // Whether this was a loading request for the instance. - WasLoadingRequest bool `protobuf:"varint,24,opt,name=was_loading_request,json=wasLoadingRequest" json:"was_loading_request,omitempty"` + WasLoadingRequest bool `protobuf:"varint,24,opt,name=was_loading_request,json=wasLoadingRequest,proto3" json:"was_loading_request,omitempty"` // Time this request spent in the pending request queue. - PendingTime *duration.Duration `protobuf:"bytes,25,opt,name=pending_time,json=pendingTime" json:"pending_time,omitempty"` + PendingTime *duration.Duration `protobuf:"bytes,25,opt,name=pending_time,json=pendingTime,proto3" json:"pending_time,omitempty"` // If the instance processing this request belongs to a manually scaled // module, then this is the 0-based index of the instance. Otherwise, this // value is -1. - InstanceIndex int32 `protobuf:"varint,26,opt,name=instance_index,json=instanceIndex" json:"instance_index,omitempty"` + InstanceIndex int32 `protobuf:"varint,26,opt,name=instance_index,json=instanceIndex,proto3" json:"instance_index,omitempty"` // Whether this request is finished or active. - Finished bool `protobuf:"varint,27,opt,name=finished" json:"finished,omitempty"` + Finished bool `protobuf:"varint,27,opt,name=finished,proto3" json:"finished,omitempty"` // Whether this is the first `RequestLog` entry for this request. If an // active request has several `RequestLog` entries written to Stackdriver // Logging, then this field will be set for one of them. - First bool `protobuf:"varint,42,opt,name=first" json:"first,omitempty"` + First bool `protobuf:"varint,42,opt,name=first,proto3" json:"first,omitempty"` // An identifier for the instance that handled the request. - InstanceId string `protobuf:"bytes,28,opt,name=instance_id,json=instanceId" json:"instance_id,omitempty"` + InstanceId string `protobuf:"bytes,28,opt,name=instance_id,json=instanceId,proto3" json:"instance_id,omitempty"` // A list of log lines emitted by the application while serving this request. - Line []*LogLine `protobuf:"bytes,29,rep,name=line" json:"line,omitempty"` + Line []*LogLine `protobuf:"bytes,29,rep,name=line,proto3" json:"line,omitempty"` // App Engine release version. - AppEngineRelease string `protobuf:"bytes,38,opt,name=app_engine_release,json=appEngineRelease" json:"app_engine_release,omitempty"` + AppEngineRelease string `protobuf:"bytes,38,opt,name=app_engine_release,json=appEngineRelease,proto3" json:"app_engine_release,omitempty"` // Stackdriver Trace identifier for this request. - TraceId string `protobuf:"bytes,39,opt,name=trace_id,json=traceId" json:"trace_id,omitempty"` + TraceId string `protobuf:"bytes,39,opt,name=trace_id,json=traceId,proto3" json:"trace_id,omitempty"` // Source code for the application that handled this request. There can be // more than one source reference per deployed application if source code is // distributed among multiple repositories. - SourceReference []*SourceReference `protobuf:"bytes,41,rep,name=source_reference,json=sourceReference" json:"source_reference,omitempty"` + SourceReference []*SourceReference `protobuf:"bytes,41,rep,name=source_reference,json=sourceReference,proto3" json:"source_reference,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/app_yaml.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/app_yaml.pb.go index 32a39e6b4..320d5df8d 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/app_yaml.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/app_yaml.pb.go @@ -218,16 +218,16 @@ func (UrlMap_RedirectHttpResponseCode) EnumDescriptor() ([]byte, []int) { type ApiConfigHandler struct { // Action to take when users access resources that require // authentication. Defaults to `redirect`. - AuthFailAction AuthFailAction `protobuf:"varint,1,opt,name=auth_fail_action,json=authFailAction,enum=google.appengine.v1.AuthFailAction" json:"auth_fail_action,omitempty"` + AuthFailAction AuthFailAction `protobuf:"varint,1,opt,name=auth_fail_action,json=authFailAction,proto3,enum=google.appengine.v1.AuthFailAction" json:"auth_fail_action,omitempty"` // Level of login required to access this resource. Defaults to // `optional`. - Login LoginRequirement `protobuf:"varint,2,opt,name=login,enum=google.appengine.v1.LoginRequirement" json:"login,omitempty"` + Login LoginRequirement `protobuf:"varint,2,opt,name=login,proto3,enum=google.appengine.v1.LoginRequirement" json:"login,omitempty"` // Path to the script from the application root directory. - Script string `protobuf:"bytes,3,opt,name=script" json:"script,omitempty"` + Script string `protobuf:"bytes,3,opt,name=script,proto3" json:"script,omitempty"` // Security (HTTPS) enforcement for this URL. - SecurityLevel SecurityLevel `protobuf:"varint,4,opt,name=security_level,json=securityLevel,enum=google.appengine.v1.SecurityLevel" json:"security_level,omitempty"` + SecurityLevel SecurityLevel `protobuf:"varint,4,opt,name=security_level,json=securityLevel,proto3,enum=google.appengine.v1.SecurityLevel" json:"security_level,omitempty"` // URL to serve the endpoint at. - Url string `protobuf:"bytes,5,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,5,opt,name=url,proto3" json:"url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -295,11 +295,11 @@ func (m *ApiConfigHandler) GetUrl() string { // Custom static error page to be served when an error occurs. type ErrorHandler struct { // Error condition this handler applies to. - ErrorCode ErrorHandler_ErrorCode `protobuf:"varint,1,opt,name=error_code,json=errorCode,enum=google.appengine.v1.ErrorHandler_ErrorCode" json:"error_code,omitempty"` + ErrorCode ErrorHandler_ErrorCode `protobuf:"varint,1,opt,name=error_code,json=errorCode,proto3,enum=google.appengine.v1.ErrorHandler_ErrorCode" json:"error_code,omitempty"` // Static file content to be served for this error. - StaticFile string `protobuf:"bytes,2,opt,name=static_file,json=staticFile" json:"static_file,omitempty"` + StaticFile string `protobuf:"bytes,2,opt,name=static_file,json=staticFile,proto3" json:"static_file,omitempty"` // MIME type of file. Defaults to `text/html`. - MimeType string `protobuf:"bytes,3,opt,name=mime_type,json=mimeType" json:"mime_type,omitempty"` + MimeType string `protobuf:"bytes,3,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -358,7 +358,7 @@ type UrlMap struct { // special characters must be escaped, but should not contain groupings. // All URLs that begin with this prefix are handled by this handler, using the // portion of the URL after the prefix as part of the file path. - UrlRegex string `protobuf:"bytes,1,opt,name=url_regex,json=urlRegex" json:"url_regex,omitempty"` + UrlRegex string `protobuf:"bytes,1,opt,name=url_regex,json=urlRegex,proto3" json:"url_regex,omitempty"` // Type of handler for this URL pattern. // // Types that are valid to be assigned to HandlerType: @@ -367,15 +367,15 @@ type UrlMap struct { // *UrlMap_ApiEndpoint HandlerType isUrlMap_HandlerType `protobuf_oneof:"handler_type"` // Security (HTTPS) enforcement for this URL. - SecurityLevel SecurityLevel `protobuf:"varint,5,opt,name=security_level,json=securityLevel,enum=google.appengine.v1.SecurityLevel" json:"security_level,omitempty"` + SecurityLevel SecurityLevel `protobuf:"varint,5,opt,name=security_level,json=securityLevel,proto3,enum=google.appengine.v1.SecurityLevel" json:"security_level,omitempty"` // Level of login required to access this resource. - Login LoginRequirement `protobuf:"varint,6,opt,name=login,enum=google.appengine.v1.LoginRequirement" json:"login,omitempty"` + Login LoginRequirement `protobuf:"varint,6,opt,name=login,proto3,enum=google.appengine.v1.LoginRequirement" json:"login,omitempty"` // Action to take when users access resources that require // authentication. Defaults to `redirect`. - AuthFailAction AuthFailAction `protobuf:"varint,7,opt,name=auth_fail_action,json=authFailAction,enum=google.appengine.v1.AuthFailAction" json:"auth_fail_action,omitempty"` + AuthFailAction AuthFailAction `protobuf:"varint,7,opt,name=auth_fail_action,json=authFailAction,proto3,enum=google.appengine.v1.AuthFailAction" json:"auth_fail_action,omitempty"` // `30x` code to use when performing redirects for the `secure` field. // Defaults to `302`. - RedirectHttpResponseCode UrlMap_RedirectHttpResponseCode `protobuf:"varint,8,opt,name=redirect_http_response_code,json=redirectHttpResponseCode,enum=google.appengine.v1.UrlMap_RedirectHttpResponseCode" json:"redirect_http_response_code,omitempty"` + RedirectHttpResponseCode UrlMap_RedirectHttpResponseCode `protobuf:"varint,8,opt,name=redirect_http_response_code,json=redirectHttpResponseCode,proto3,enum=google.appengine.v1.UrlMap_RedirectHttpResponseCode" json:"redirect_http_response_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -410,13 +410,13 @@ type isUrlMap_HandlerType interface { } type UrlMap_StaticFiles struct { - StaticFiles *StaticFilesHandler `protobuf:"bytes,2,opt,name=static_files,json=staticFiles,oneof"` + StaticFiles *StaticFilesHandler `protobuf:"bytes,2,opt,name=static_files,json=staticFiles,proto3,oneof"` } type UrlMap_Script struct { - Script *ScriptHandler `protobuf:"bytes,3,opt,name=script,oneof"` + Script *ScriptHandler `protobuf:"bytes,3,opt,name=script,proto3,oneof"` } type UrlMap_ApiEndpoint struct { - ApiEndpoint *ApiEndpointHandler `protobuf:"bytes,4,opt,name=api_endpoint,json=apiEndpoint,oneof"` + ApiEndpoint *ApiEndpointHandler `protobuf:"bytes,4,opt,name=api_endpoint,json=apiEndpoint,proto3,oneof"` } func (*UrlMap_StaticFiles) isUrlMap_HandlerType() {} @@ -587,29 +587,29 @@ type StaticFilesHandler struct { // Path to the static files matched by the URL pattern, from the // application root directory. The path can refer to text matched in groupings // in the URL pattern. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // Regular expression that matches the file paths for all files that should be // referenced by this handler. - UploadPathRegex string `protobuf:"bytes,2,opt,name=upload_path_regex,json=uploadPathRegex" json:"upload_path_regex,omitempty"` + UploadPathRegex string `protobuf:"bytes,2,opt,name=upload_path_regex,json=uploadPathRegex,proto3" json:"upload_path_regex,omitempty"` // HTTP headers to use for all responses from these URLs. - HttpHeaders map[string]string `protobuf:"bytes,3,rep,name=http_headers,json=httpHeaders" json:"http_headers,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + HttpHeaders map[string]string `protobuf:"bytes,3,rep,name=http_headers,json=httpHeaders,proto3" json:"http_headers,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // MIME type used to serve all files served by this handler. // // Defaults to file-specific MIME types, which are derived from each file's // filename extension. - MimeType string `protobuf:"bytes,4,opt,name=mime_type,json=mimeType" json:"mime_type,omitempty"` + MimeType string `protobuf:"bytes,4,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"` // Time a static file served by this handler should be cached // by web proxies and browsers. - Expiration *duration.Duration `protobuf:"bytes,5,opt,name=expiration" json:"expiration,omitempty"` + Expiration *duration.Duration `protobuf:"bytes,5,opt,name=expiration,proto3" json:"expiration,omitempty"` // Whether this handler should match the request if the file // referenced by the handler does not exist. - RequireMatchingFile bool `protobuf:"varint,6,opt,name=require_matching_file,json=requireMatchingFile" json:"require_matching_file,omitempty"` + RequireMatchingFile bool `protobuf:"varint,6,opt,name=require_matching_file,json=requireMatchingFile,proto3" json:"require_matching_file,omitempty"` // Whether files should also be uploaded as code data. By default, files // declared in static file handlers are uploaded as static // data and are only served to end users; they cannot be read by the // application. If enabled, uploads are charged against both your code and // static data storage resource quotas. - ApplicationReadable bool `protobuf:"varint,7,opt,name=application_readable,json=applicationReadable" json:"application_readable,omitempty"` + ApplicationReadable bool `protobuf:"varint,7,opt,name=application_readable,json=applicationReadable,proto3" json:"application_readable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -691,7 +691,7 @@ func (m *StaticFilesHandler) GetApplicationReadable() bool { // Executes a script to handle the request that matches the URL pattern. type ScriptHandler struct { // Path to the script from the application root directory. - ScriptPath string `protobuf:"bytes,1,opt,name=script_path,json=scriptPath" json:"script_path,omitempty"` + ScriptPath string `protobuf:"bytes,1,opt,name=script_path,json=scriptPath,proto3" json:"script_path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -731,7 +731,7 @@ func (m *ScriptHandler) GetScriptPath() string { // Uses Google Cloud Endpoints to handle requests. type ApiEndpointHandler struct { // Path to the script from the application root directory. - ScriptPath string `protobuf:"bytes,1,opt,name=script_path,json=scriptPath" json:"script_path,omitempty"` + ScriptPath string `protobuf:"bytes,1,opt,name=script_path,json=scriptPath,proto3" json:"script_path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -773,23 +773,23 @@ func (m *ApiEndpointHandler) GetScriptPath() string { // instances in App Engine flexible environment. type HealthCheck struct { // Whether to explicitly disable health checks for this instance. - DisableHealthCheck bool `protobuf:"varint,1,opt,name=disable_health_check,json=disableHealthCheck" json:"disable_health_check,omitempty"` + DisableHealthCheck bool `protobuf:"varint,1,opt,name=disable_health_check,json=disableHealthCheck,proto3" json:"disable_health_check,omitempty"` // Host header to send when performing an HTTP health check. // Example: "myapp.appspot.com" - Host string `protobuf:"bytes,2,opt,name=host" json:"host,omitempty"` + Host string `protobuf:"bytes,2,opt,name=host,proto3" json:"host,omitempty"` // Number of consecutive successful health checks required before receiving // traffic. - HealthyThreshold uint32 `protobuf:"varint,3,opt,name=healthy_threshold,json=healthyThreshold" json:"healthy_threshold,omitempty"` + HealthyThreshold uint32 `protobuf:"varint,3,opt,name=healthy_threshold,json=healthyThreshold,proto3" json:"healthy_threshold,omitempty"` // Number of consecutive failed health checks required before removing // traffic. - UnhealthyThreshold uint32 `protobuf:"varint,4,opt,name=unhealthy_threshold,json=unhealthyThreshold" json:"unhealthy_threshold,omitempty"` + UnhealthyThreshold uint32 `protobuf:"varint,4,opt,name=unhealthy_threshold,json=unhealthyThreshold,proto3" json:"unhealthy_threshold,omitempty"` // Number of consecutive failed health checks required before an instance is // restarted. - RestartThreshold uint32 `protobuf:"varint,5,opt,name=restart_threshold,json=restartThreshold" json:"restart_threshold,omitempty"` + RestartThreshold uint32 `protobuf:"varint,5,opt,name=restart_threshold,json=restartThreshold,proto3" json:"restart_threshold,omitempty"` // Interval between health checks. - CheckInterval *duration.Duration `protobuf:"bytes,6,opt,name=check_interval,json=checkInterval" json:"check_interval,omitempty"` + CheckInterval *duration.Duration `protobuf:"bytes,6,opt,name=check_interval,json=checkInterval,proto3" json:"check_interval,omitempty"` // Time before the health check is considered failed. - Timeout *duration.Duration `protobuf:"bytes,7,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,7,opt,name=timeout,proto3" json:"timeout,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -871,9 +871,9 @@ func (m *HealthCheck) GetTimeout() *duration.Duration { // Third-party Python runtime library that is required by the application. type Library struct { // Name of the library. Example: "django". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Version of the library to select, or "latest". - Version string `protobuf:"bytes,2,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/appengine.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/appengine.pb.go index 362750a5a..739783a0f 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/appengine.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/appengine.pb.go @@ -61,7 +61,7 @@ func (VersionView) EnumDescriptor() ([]byte, []int) { // Request message for `Applications.GetApplication`. type GetApplicationRequest struct { // Name of the Application resource to get. Example: `apps/myapp`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -101,7 +101,7 @@ func (m *GetApplicationRequest) GetName() string { // Request message for 'Applications.RepairApplication'. type RepairApplicationRequest struct { // Name of the application to repair. Example: `apps/myapp` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -141,11 +141,11 @@ func (m *RepairApplicationRequest) GetName() string { // Request message for `Services.ListServices`. type ListServicesRequest struct { // Name of the parent Application resource. Example: `apps/myapp`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Maximum results to return per page. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Continuation token for fetching the next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -199,9 +199,9 @@ func (m *ListServicesRequest) GetPageToken() string { // Response message for `Services.ListServices`. type ListServicesResponse struct { // The services belonging to the requested application. - Services []*Service `protobuf:"bytes,1,rep,name=services" json:"services,omitempty"` + Services []*Service `protobuf:"bytes,1,rep,name=services,proto3" json:"services,omitempty"` // Continuation token for fetching the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -248,7 +248,7 @@ func (m *ListServicesResponse) GetNextPageToken() string { // Request message for `Services.GetService`. type GetServiceRequest struct { // Name of the resource requested. Example: `apps/myapp/services/default`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -288,12 +288,12 @@ func (m *GetServiceRequest) GetName() string { // Request message for `Services.UpdateService`. type UpdateServiceRequest struct { // Name of the resource to update. Example: `apps/myapp/services/default`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A Service resource containing the updated service. Only fields set in the // field mask will be updated. - Service *Service `protobuf:"bytes,2,opt,name=service" json:"service,omitempty"` + Service *Service `protobuf:"bytes,2,opt,name=service,proto3" json:"service,omitempty"` // Standard field mask for the set of fields to be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Set to `true` to gradually shift traffic from one version to another // single version. By default, traffic is shifted immediately. // For gradual traffic migration, the target version @@ -306,7 +306,7 @@ type UpdateServiceRequest struct { // field in the Service resource. Gradual traffic migration is not // supported in the App Engine flexible environment. For examples, see // [Migrating and Splitting Traffic](https://cloud.google.com/appengine/docs/admin-api/migrating-splitting-traffic). - MigrateTraffic bool `protobuf:"varint,4,opt,name=migrate_traffic,json=migrateTraffic" json:"migrate_traffic,omitempty"` + MigrateTraffic bool `protobuf:"varint,4,opt,name=migrate_traffic,json=migrateTraffic,proto3" json:"migrate_traffic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -367,7 +367,7 @@ func (m *UpdateServiceRequest) GetMigrateTraffic() bool { // Request message for `Services.DeleteService`. type DeleteServiceRequest struct { // Name of the resource requested. Example: `apps/myapp/services/default`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -408,13 +408,13 @@ func (m *DeleteServiceRequest) GetName() string { type ListVersionsRequest struct { // Name of the parent Service resource. Example: // `apps/myapp/services/default`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Controls the set of fields returned in the `List` response. - View VersionView `protobuf:"varint,2,opt,name=view,enum=google.appengine.v1.VersionView" json:"view,omitempty"` + View VersionView `protobuf:"varint,2,opt,name=view,proto3,enum=google.appengine.v1.VersionView" json:"view,omitempty"` // Maximum results to return per page. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Continuation token for fetching the next page of results. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -475,9 +475,9 @@ func (m *ListVersionsRequest) GetPageToken() string { // Response message for `Versions.ListVersions`. type ListVersionsResponse struct { // The versions belonging to the requested service. - Versions []*Version `protobuf:"bytes,1,rep,name=versions" json:"versions,omitempty"` + Versions []*Version `protobuf:"bytes,1,rep,name=versions,proto3" json:"versions,omitempty"` // Continuation token for fetching the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -525,9 +525,9 @@ func (m *ListVersionsResponse) GetNextPageToken() string { type GetVersionRequest struct { // Name of the resource requested. Example: // `apps/myapp/services/default/versions/v1`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Controls the set of fields returned in the `Get` response. - View VersionView `protobuf:"varint,2,opt,name=view,enum=google.appengine.v1.VersionView" json:"view,omitempty"` + View VersionView `protobuf:"varint,2,opt,name=view,proto3,enum=google.appengine.v1.VersionView" json:"view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -575,9 +575,9 @@ func (m *GetVersionRequest) GetView() VersionView { type CreateVersionRequest struct { // Name of the parent resource to create this version under. Example: // `apps/myapp/services/default`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Application deployment configuration. - Version *Version `protobuf:"bytes,2,opt,name=version" json:"version,omitempty"` + Version *Version `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -625,12 +625,12 @@ func (m *CreateVersionRequest) GetVersion() *Version { type UpdateVersionRequest struct { // Name of the resource to update. Example: // `apps/myapp/services/default/versions/1`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A Version containing the updated resource. Only fields set in the field // mask will be updated. - Version *Version `protobuf:"bytes,2,opt,name=version" json:"version,omitempty"` + Version *Version `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` // Standard field mask for the set of fields to be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -685,7 +685,7 @@ func (m *UpdateVersionRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteVersionRequest struct { // Name of the resource requested. Example: // `apps/myapp/services/default/versions/v1`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -726,11 +726,11 @@ func (m *DeleteVersionRequest) GetName() string { type ListInstancesRequest struct { // Name of the parent Version resource. Example: // `apps/myapp/services/default/versions/v1`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Maximum results to return per page. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Continuation token for fetching the next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -784,9 +784,9 @@ func (m *ListInstancesRequest) GetPageToken() string { // Response message for `Instances.ListInstances`. type ListInstancesResponse struct { // The instances belonging to the requested version. - Instances []*Instance `protobuf:"bytes,1,rep,name=instances" json:"instances,omitempty"` + Instances []*Instance `protobuf:"bytes,1,rep,name=instances,proto3" json:"instances,omitempty"` // Continuation token for fetching the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -834,7 +834,7 @@ func (m *ListInstancesResponse) GetNextPageToken() string { type GetInstanceRequest struct { // Name of the resource requested. Example: // `apps/myapp/services/default/versions/v1/instances/instance-1`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -875,7 +875,7 @@ func (m *GetInstanceRequest) GetName() string { type DeleteInstanceRequest struct { // Name of the resource requested. Example: // `apps/myapp/services/default/versions/v1/instances/instance-1`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -916,7 +916,7 @@ func (m *DeleteInstanceRequest) GetName() string { type DebugInstanceRequest struct { // Name of the resource requested. Example: // `apps/myapp/services/default/versions/v1/instances/instance-1`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1048,8 +1048,7 @@ func (c *instancesClient) DebugInstance(ctx context.Context, in *DebugInstanceRe return out, nil } -// Server API for Instances service - +// InstancesServer is the server API for Instances service. type InstancesServer interface { // Lists the instances of a version. ListInstances(context.Context, *ListInstancesRequest) (*ListInstancesResponse, error) @@ -1254,8 +1253,7 @@ func (c *versionsClient) DeleteVersion(ctx context.Context, in *DeleteVersionReq return out, nil } -// Server API for Versions service - +// VersionsServer is the server API for Versions service. type VersionsServer interface { // Lists the versions of a service. ListVersions(context.Context, *ListVersionsRequest) (*ListVersionsResponse, error) @@ -1466,8 +1464,7 @@ func (c *servicesClient) DeleteService(ctx context.Context, in *DeleteServiceReq return out, nil } -// Server API for Services service - +// ServicesServer is the server API for Services service. type ServicesServer interface { // Lists all the services in the application. ListServices(context.Context, *ListServicesRequest) (*ListServicesResponse, error) @@ -1619,8 +1616,7 @@ func (c *applicationsClient) RepairApplication(ctx context.Context, in *RepairAp return out, nil } -// Server API for Applications service - +// ApplicationsServer is the server API for Applications service. type ApplicationsServer interface { // Gets information about an application. GetApplication(context.Context, *GetApplicationRequest) (*Application, error) diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/application.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/application.pb.go index 37c7974fa..b953ce83e 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/application.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/application.pb.go @@ -27,22 +27,22 @@ type Application struct { // Example: `apps/myapp`. // // @OutputOnly - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Identifier of the Application resource. This identifier is equivalent // to the project ID of the Google Cloud Platform project where you want to // deploy your application. // Example: `myapp`. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // HTTP path dispatch rules for requests to the application that do not // explicitly target a service or version. Rules are order-dependent. // // @OutputOnly - DispatchRules []*UrlDispatchRule `protobuf:"bytes,3,rep,name=dispatch_rules,json=dispatchRules" json:"dispatch_rules,omitempty"` + DispatchRules []*UrlDispatchRule `protobuf:"bytes,3,rep,name=dispatch_rules,json=dispatchRules,proto3" json:"dispatch_rules,omitempty"` // Google Apps authentication domain that controls which users can access // this application. // // Defaults to open access for any Google Account. - AuthDomain string `protobuf:"bytes,6,opt,name=auth_domain,json=authDomain" json:"auth_domain,omitempty"` + AuthDomain string `protobuf:"bytes,6,opt,name=auth_domain,json=authDomain,proto3" json:"auth_domain,omitempty"` // Location from which this application will be run. Application instances // will run out of data centers in the chosen location, which is also where // all of the application's end user content is stored. @@ -56,26 +56,26 @@ type Application struct { // `europe-west` - Western Europe // // `us-east1` - Eastern US - LocationId string `protobuf:"bytes,7,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,7,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` // Google Cloud Storage bucket that can be used for storing files // associated with this application. This bucket is associated with the // application and can be used by the gcloud deployment commands. // // @OutputOnly - CodeBucket string `protobuf:"bytes,8,opt,name=code_bucket,json=codeBucket" json:"code_bucket,omitempty"` + CodeBucket string `protobuf:"bytes,8,opt,name=code_bucket,json=codeBucket,proto3" json:"code_bucket,omitempty"` // Cookie expiration policy for this application. // // @OutputOnly - DefaultCookieExpiration *duration.Duration `protobuf:"bytes,9,opt,name=default_cookie_expiration,json=defaultCookieExpiration" json:"default_cookie_expiration,omitempty"` + DefaultCookieExpiration *duration.Duration `protobuf:"bytes,9,opt,name=default_cookie_expiration,json=defaultCookieExpiration,proto3" json:"default_cookie_expiration,omitempty"` // Hostname used to reach this application, as resolved by App Engine. // // @OutputOnly - DefaultHostname string `protobuf:"bytes,11,opt,name=default_hostname,json=defaultHostname" json:"default_hostname,omitempty"` + DefaultHostname string `protobuf:"bytes,11,opt,name=default_hostname,json=defaultHostname,proto3" json:"default_hostname,omitempty"` // Google Cloud Storage bucket that can be used by this application to store // content. // // @OutputOnly - DefaultBucket string `protobuf:"bytes,12,opt,name=default_bucket,json=defaultBucket" json:"default_bucket,omitempty"` + DefaultBucket string `protobuf:"bytes,12,opt,name=default_bucket,json=defaultBucket,proto3" json:"default_bucket,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -174,16 +174,16 @@ type UrlDispatchRule struct { // specified before a period: "`*.`". // // Defaults to matching all domains: "`*`". - Domain string `protobuf:"bytes,1,opt,name=domain" json:"domain,omitempty"` + Domain string `protobuf:"bytes,1,opt,name=domain,proto3" json:"domain,omitempty"` // Pathname within the host. Must start with a "`/`". A // single "`*`" can be included at the end of the path. The sum // of the lengths of the domain and path may not exceed 100 // characters. - Path string `protobuf:"bytes,2,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,2,opt,name=path,proto3" json:"path,omitempty"` // Resource ID of a service in this application that should // serve the matched request. The service must already // exist. Example: `default`. - Service string `protobuf:"bytes,3,opt,name=service" json:"service,omitempty"` + Service string `protobuf:"bytes,3,opt,name=service,proto3" json:"service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/audit_data.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/audit_data.pb.go index f37934cd6..e9d7c649f 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/audit_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/audit_data.pb.go @@ -64,10 +64,10 @@ type isAuditData_Method interface { } type AuditData_UpdateService struct { - UpdateService *UpdateServiceMethod `protobuf:"bytes,1,opt,name=update_service,json=updateService,oneof"` + UpdateService *UpdateServiceMethod `protobuf:"bytes,1,opt,name=update_service,json=updateService,proto3,oneof"` } type AuditData_CreateVersion struct { - CreateVersion *CreateVersionMethod `protobuf:"bytes,2,opt,name=create_version,json=createVersion,oneof"` + CreateVersion *CreateVersionMethod `protobuf:"bytes,2,opt,name=create_version,json=createVersion,proto3,oneof"` } func (*AuditData_UpdateService) isAuditData_Method() {} @@ -171,7 +171,7 @@ func _AuditData_OneofSizer(msg proto.Message) (n int) { // Detailed information about UpdateService call. type UpdateServiceMethod struct { // Update service request. - Request *UpdateServiceRequest `protobuf:"bytes,1,opt,name=request" json:"request,omitempty"` + Request *UpdateServiceRequest `protobuf:"bytes,1,opt,name=request,proto3" json:"request,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -211,7 +211,7 @@ func (m *UpdateServiceMethod) GetRequest() *UpdateServiceRequest { // Detailed information about CreateVersion call. type CreateVersionMethod struct { // Create version request. - Request *CreateVersionRequest `protobuf:"bytes,1,opt,name=request" json:"request,omitempty"` + Request *CreateVersionRequest `protobuf:"bytes,1,opt,name=request,proto3" json:"request,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/deploy.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/deploy.pb.go index 84bd2fe39..2c0babc24 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/deploy.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/deploy.pb.go @@ -24,12 +24,12 @@ type Deployment struct { // Manifest of the files stored in Google Cloud Storage that are included // as part of this version. All files must be readable using the // credentials supplied with this call. - Files map[string]*FileInfo `protobuf:"bytes,1,rep,name=files" json:"files,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Files map[string]*FileInfo `protobuf:"bytes,1,rep,name=files,proto3" json:"files,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // A Docker image that App Engine uses to run the version. // Only applicable for instances in App Engine flexible environment. - Container *ContainerInfo `protobuf:"bytes,2,opt,name=container" json:"container,omitempty"` + Container *ContainerInfo `protobuf:"bytes,2,opt,name=container,proto3" json:"container,omitempty"` // The zip file for this deployment, if this is a zip deployment. - Zip *ZipInfo `protobuf:"bytes,3,opt,name=zip" json:"zip,omitempty"` + Zip *ZipInfo `protobuf:"bytes,3,opt,name=zip,proto3" json:"zip,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -86,13 +86,13 @@ type FileInfo struct { // URL source to use to fetch this file. Must be a URL to a resource in // Google Cloud Storage in the form // 'http(s)://storage.googleapis.com/\/\'. - SourceUrl string `protobuf:"bytes,1,opt,name=source_url,json=sourceUrl" json:"source_url,omitempty"` + SourceUrl string `protobuf:"bytes,1,opt,name=source_url,json=sourceUrl,proto3" json:"source_url,omitempty"` // The SHA1 hash of the file, in hex. - Sha1Sum string `protobuf:"bytes,2,opt,name=sha1_sum,json=sha1Sum" json:"sha1_sum,omitempty"` + Sha1Sum string `protobuf:"bytes,2,opt,name=sha1_sum,json=sha1Sum,proto3" json:"sha1_sum,omitempty"` // The MIME type of the file. // // Defaults to the value from Google Cloud Storage. - MimeType string `protobuf:"bytes,3,opt,name=mime_type,json=mimeType" json:"mime_type,omitempty"` + MimeType string `protobuf:"bytes,3,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -149,7 +149,7 @@ type ContainerInfo struct { // URI to the hosted container image in a Docker repository. The URI must be // fully qualified and include a tag or digest. // Examples: "gcr.io/my-project/image:tag" or "gcr.io/my-project/image@digest" - Image string `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"` + Image string `protobuf:"bytes,1,opt,name=image,proto3" json:"image,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -190,11 +190,11 @@ type ZipInfo struct { // URL of the zip file to deploy from. Must be a URL to a resource in // Google Cloud Storage in the form // 'http(s)://storage.googleapis.com/\/\'. - SourceUrl string `protobuf:"bytes,3,opt,name=source_url,json=sourceUrl" json:"source_url,omitempty"` + SourceUrl string `protobuf:"bytes,3,opt,name=source_url,json=sourceUrl,proto3" json:"source_url,omitempty"` // An estimate of the number of files in a zip for a zip deployment. // If set, must be greater than or equal to the actual number of files. // Used for optimizing performance; if not provided, deployment may be slow. - FilesCount int32 `protobuf:"varint,4,opt,name=files_count,json=filesCount" json:"files_count,omitempty"` + FilesCount int32 `protobuf:"varint,4,opt,name=files_count,json=filesCount,proto3" json:"files_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/instance.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/instance.pb.go index a5d2a55c4..f49ecb060 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/instance.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/instance.pb.go @@ -54,69 +54,69 @@ type Instance struct { // Example: `apps/myapp/services/default/versions/v1/instances/instance-1`. // // @OutputOnly - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Relative name of the instance within the version. // Example: `instance-1`. // // @OutputOnly - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // App Engine release this instance is running on. // // @OutputOnly - AppEngineRelease string `protobuf:"bytes,3,opt,name=app_engine_release,json=appEngineRelease" json:"app_engine_release,omitempty"` + AppEngineRelease string `protobuf:"bytes,3,opt,name=app_engine_release,json=appEngineRelease,proto3" json:"app_engine_release,omitempty"` // Availability of the instance. // // @OutputOnly - Availability Instance_Availability `protobuf:"varint,4,opt,name=availability,enum=google.appengine.v1.Instance_Availability" json:"availability,omitempty"` + Availability Instance_Availability `protobuf:"varint,4,opt,name=availability,proto3,enum=google.appengine.v1.Instance_Availability" json:"availability,omitempty"` // Name of the virtual machine where this instance lives. Only applicable // for instances in App Engine flexible environment. // // @OutputOnly - VmName string `protobuf:"bytes,5,opt,name=vm_name,json=vmName" json:"vm_name,omitempty"` + VmName string `protobuf:"bytes,5,opt,name=vm_name,json=vmName,proto3" json:"vm_name,omitempty"` // Zone where the virtual machine is located. Only applicable for instances // in App Engine flexible environment. // // @OutputOnly - VmZoneName string `protobuf:"bytes,6,opt,name=vm_zone_name,json=vmZoneName" json:"vm_zone_name,omitempty"` + VmZoneName string `protobuf:"bytes,6,opt,name=vm_zone_name,json=vmZoneName,proto3" json:"vm_zone_name,omitempty"` // Virtual machine ID of this instance. Only applicable for instances in // App Engine flexible environment. // // @OutputOnly - VmId string `protobuf:"bytes,7,opt,name=vm_id,json=vmId" json:"vm_id,omitempty"` + VmId string `protobuf:"bytes,7,opt,name=vm_id,json=vmId,proto3" json:"vm_id,omitempty"` // Time that this instance was started. // // @OutputOnly - StartTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Number of requests since this instance was started. // // @OutputOnly - Requests int32 `protobuf:"varint,9,opt,name=requests" json:"requests,omitempty"` + Requests int32 `protobuf:"varint,9,opt,name=requests,proto3" json:"requests,omitempty"` // Number of errors since this instance was started. // // @OutputOnly - Errors int32 `protobuf:"varint,10,opt,name=errors" json:"errors,omitempty"` + Errors int32 `protobuf:"varint,10,opt,name=errors,proto3" json:"errors,omitempty"` // Average queries per second (QPS) over the last minute. // // @OutputOnly - Qps float32 `protobuf:"fixed32,11,opt,name=qps" json:"qps,omitempty"` + Qps float32 `protobuf:"fixed32,11,opt,name=qps,proto3" json:"qps,omitempty"` // Average latency (ms) over the last minute. // // @OutputOnly - AverageLatency int32 `protobuf:"varint,12,opt,name=average_latency,json=averageLatency" json:"average_latency,omitempty"` + AverageLatency int32 `protobuf:"varint,12,opt,name=average_latency,json=averageLatency,proto3" json:"average_latency,omitempty"` // Total memory in use (bytes). // // @OutputOnly - MemoryUsage int64 `protobuf:"varint,13,opt,name=memory_usage,json=memoryUsage" json:"memory_usage,omitempty"` + MemoryUsage int64 `protobuf:"varint,13,opt,name=memory_usage,json=memoryUsage,proto3" json:"memory_usage,omitempty"` // Status of the virtual machine where this instance lives. Only applicable // for instances in App Engine flexible environment. // // @OutputOnly - VmStatus string `protobuf:"bytes,14,opt,name=vm_status,json=vmStatus" json:"vm_status,omitempty"` + VmStatus string `protobuf:"bytes,14,opt,name=vm_status,json=vmStatus,proto3" json:"vm_status,omitempty"` // Whether this instance is in debug mode. Only applicable for instances in // App Engine flexible environment. // // @OutputOnly - VmDebugEnabled bool `protobuf:"varint,15,opt,name=vm_debug_enabled,json=vmDebugEnabled" json:"vm_debug_enabled,omitempty"` + VmDebugEnabled bool `protobuf:"varint,15,opt,name=vm_debug_enabled,json=vmDebugEnabled,proto3" json:"vm_debug_enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/location.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/location.pb.go index b3fae5ad1..bd533eac5 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/location.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/location.pb.go @@ -25,11 +25,11 @@ type LocationMetadata struct { // App Engine Standard Environment is available in the given location. // // @OutputOnly - StandardEnvironmentAvailable bool `protobuf:"varint,2,opt,name=standard_environment_available,json=standardEnvironmentAvailable" json:"standard_environment_available,omitempty"` + StandardEnvironmentAvailable bool `protobuf:"varint,2,opt,name=standard_environment_available,json=standardEnvironmentAvailable,proto3" json:"standard_environment_available,omitempty"` // App Engine Flexible Environment is available in the given location. // // @OutputOnly - FlexibleEnvironmentAvailable bool `protobuf:"varint,4,opt,name=flexible_environment_available,json=flexibleEnvironmentAvailable" json:"flexible_environment_available,omitempty"` + FlexibleEnvironmentAvailable bool `protobuf:"varint,4,opt,name=flexible_environment_available,json=flexibleEnvironmentAvailable,proto3" json:"flexible_environment_available,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/operation.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/operation.pb.go index 44091ea75..35c9f93d6 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/operation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/operation.pb.go @@ -26,24 +26,24 @@ type OperationMetadataV1 struct { // `google.appengine.v1.Versions.CreateVersion`. // // @OutputOnly - Method string `protobuf:"bytes,1,opt,name=method" json:"method,omitempty"` + Method string `protobuf:"bytes,1,opt,name=method,proto3" json:"method,omitempty"` // Time that this operation was created. // // @OutputOnly - InsertTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=insert_time,json=insertTime" json:"insert_time,omitempty"` + InsertTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=insert_time,json=insertTime,proto3" json:"insert_time,omitempty"` // Time that this operation completed. // // @OutputOnly - EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // User who requested this operation. // // @OutputOnly - User string `protobuf:"bytes,4,opt,name=user" json:"user,omitempty"` + User string `protobuf:"bytes,4,opt,name=user,proto3" json:"user,omitempty"` // Name of the resource that this operation is acting on. Example: // `apps/myapp/services/default`. // // @OutputOnly - Target string `protobuf:"bytes,5,opt,name=target" json:"target,omitempty"` + Target string `protobuf:"bytes,5,opt,name=target,proto3" json:"target,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/service.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/service.pb.go index dccb1d27e..6decdae3b 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/service.pb.go @@ -63,15 +63,15 @@ type Service struct { // Example: `apps/myapp/services/default`. // // @OutputOnly - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Relative name of the service within the application. // Example: `default`. // // @OutputOnly - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // Mapping that defines fractional HTTP traffic diversion to // different versions within the service. - Split *TrafficSplit `protobuf:"bytes,3,opt,name=split" json:"split,omitempty"` + Split *TrafficSplit `protobuf:"bytes,3,opt,name=split,proto3" json:"split,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -128,7 +128,7 @@ type TrafficSplit struct { // Mechanism used to determine which version a request is sent to. // The traffic selection algorithm will // be stable for either type until allocations are changed. - ShardBy TrafficSplit_ShardBy `protobuf:"varint,1,opt,name=shard_by,json=shardBy,enum=google.appengine.v1.TrafficSplit_ShardBy" json:"shard_by,omitempty"` + ShardBy TrafficSplit_ShardBy `protobuf:"varint,1,opt,name=shard_by,json=shardBy,proto3,enum=google.appengine.v1.TrafficSplit_ShardBy" json:"shard_by,omitempty"` // Mapping from version IDs within the service to fractional // (0.000, 1] allocations of traffic for that version. Each version can // be specified only once, but some versions in the service may not @@ -137,7 +137,7 @@ type TrafficSplit struct { // their traffic allocation is removed. Allocations must sum to 1. // Up to two decimal place precision is supported for IP-based splits and // up to three decimal places is supported for cookie-based splits. - Allocations map[string]float64 `protobuf:"bytes,2,rep,name=allocations" json:"allocations,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"fixed64,2,opt,name=value"` + Allocations map[string]float64 `protobuf:"bytes,2,rep,name=allocations,proto3" json:"allocations,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"fixed64,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/appengine/v1/version.pb.go b/vendor/google.golang.org/genproto/googleapis/appengine/v1/version.pb.go index ab507f439..8c04ac570 100644 --- a/vendor/google.golang.org/genproto/googleapis/appengine/v1/version.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/appengine/v1/version.pb.go @@ -116,11 +116,11 @@ type Version struct { // `apps/myapp/services/default/versions/v1`. // // @OutputOnly - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Relative name of the version within the service. Example: `v1`. // Version names can contain only lowercase letters, numbers, or hyphens. // Reserved names: "default", "latest", and any name with the prefix "ah-". - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // Controls how instances are created. // // Defaults to `AutomaticScaling`. @@ -132,100 +132,100 @@ type Version struct { Scaling isVersion_Scaling `protobuf_oneof:"scaling"` // Before an application can receive email or XMPP messages, the application // must be configured to enable the service. - InboundServices []InboundServiceType `protobuf:"varint,6,rep,packed,name=inbound_services,json=inboundServices,enum=google.appengine.v1.InboundServiceType" json:"inbound_services,omitempty"` + InboundServices []InboundServiceType `protobuf:"varint,6,rep,packed,name=inbound_services,json=inboundServices,proto3,enum=google.appengine.v1.InboundServiceType" json:"inbound_services,omitempty"` // Instance class that is used to run this version. Valid values are: // * AutomaticScaling: `F1`, `F2`, `F4`, `F4_1G` // * ManualScaling or BasicScaling: `B1`, `B2`, `B4`, `B8`, `B4_1G` // // Defaults to `F1` for AutomaticScaling and `B1` for ManualScaling or // BasicScaling. - InstanceClass string `protobuf:"bytes,7,opt,name=instance_class,json=instanceClass" json:"instance_class,omitempty"` + InstanceClass string `protobuf:"bytes,7,opt,name=instance_class,json=instanceClass,proto3" json:"instance_class,omitempty"` // Extra network settings. Only applicable for VM runtimes. - Network *Network `protobuf:"bytes,8,opt,name=network" json:"network,omitempty"` + Network *Network `protobuf:"bytes,8,opt,name=network,proto3" json:"network,omitempty"` // Machine resources for this version. Only applicable for VM runtimes. - Resources *Resources `protobuf:"bytes,9,opt,name=resources" json:"resources,omitempty"` + Resources *Resources `protobuf:"bytes,9,opt,name=resources,proto3" json:"resources,omitempty"` // Desired runtime. Example: `python27`. - Runtime string `protobuf:"bytes,10,opt,name=runtime" json:"runtime,omitempty"` + Runtime string `protobuf:"bytes,10,opt,name=runtime,proto3" json:"runtime,omitempty"` // Whether multiple requests can be dispatched to this version at once. - Threadsafe bool `protobuf:"varint,11,opt,name=threadsafe" json:"threadsafe,omitempty"` + Threadsafe bool `protobuf:"varint,11,opt,name=threadsafe,proto3" json:"threadsafe,omitempty"` // Whether to deploy this version in a container on a virtual machine. - Vm bool `protobuf:"varint,12,opt,name=vm" json:"vm,omitempty"` + Vm bool `protobuf:"varint,12,opt,name=vm,proto3" json:"vm,omitempty"` // Metadata settings that are supplied to this version to enable // beta runtime features. - BetaSettings map[string]string `protobuf:"bytes,13,rep,name=beta_settings,json=betaSettings" json:"beta_settings,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + BetaSettings map[string]string `protobuf:"bytes,13,rep,name=beta_settings,json=betaSettings,proto3" json:"beta_settings,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // App Engine execution environment for this version. // // Defaults to `standard`. - Env string `protobuf:"bytes,14,opt,name=env" json:"env,omitempty"` + Env string `protobuf:"bytes,14,opt,name=env,proto3" json:"env,omitempty"` // Current serving status of this version. Only the versions with a // `SERVING` status create instances and can be billed. // // `SERVING_STATUS_UNSPECIFIED` is an invalid value. Defaults to `SERVING`. - ServingStatus ServingStatus `protobuf:"varint,15,opt,name=serving_status,json=servingStatus,enum=google.appengine.v1.ServingStatus" json:"serving_status,omitempty"` + ServingStatus ServingStatus `protobuf:"varint,15,opt,name=serving_status,json=servingStatus,proto3,enum=google.appengine.v1.ServingStatus" json:"serving_status,omitempty"` // Email address of the user who created this version. // // @OutputOnly - CreatedBy string `protobuf:"bytes,16,opt,name=created_by,json=createdBy" json:"created_by,omitempty"` + CreatedBy string `protobuf:"bytes,16,opt,name=created_by,json=createdBy,proto3" json:"created_by,omitempty"` // Time that this version was created. // // @OutputOnly - CreateTime *timestamp.Timestamp `protobuf:"bytes,17,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,17,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Total size in bytes of all the files that are included in this version // and curerntly hosted on the App Engine disk. // // @OutputOnly - DiskUsageBytes int64 `protobuf:"varint,18,opt,name=disk_usage_bytes,json=diskUsageBytes" json:"disk_usage_bytes,omitempty"` + DiskUsageBytes int64 `protobuf:"varint,18,opt,name=disk_usage_bytes,json=diskUsageBytes,proto3" json:"disk_usage_bytes,omitempty"` // An ordered list of URL-matching patterns that should be applied to incoming // requests. The first matching URL handles the request and other request // handlers are not attempted. // // Only returned in `GET` requests if `view=FULL` is set. - Handlers []*UrlMap `protobuf:"bytes,100,rep,name=handlers" json:"handlers,omitempty"` + Handlers []*UrlMap `protobuf:"bytes,100,rep,name=handlers,proto3" json:"handlers,omitempty"` // Custom static error pages. Limited to 10KB per page. // // Only returned in `GET` requests if `view=FULL` is set. - ErrorHandlers []*ErrorHandler `protobuf:"bytes,101,rep,name=error_handlers,json=errorHandlers" json:"error_handlers,omitempty"` + ErrorHandlers []*ErrorHandler `protobuf:"bytes,101,rep,name=error_handlers,json=errorHandlers,proto3" json:"error_handlers,omitempty"` // Configuration for third-party Python runtime libraries that are required // by the application. // // Only returned in `GET` requests if `view=FULL` is set. - Libraries []*Library `protobuf:"bytes,102,rep,name=libraries" json:"libraries,omitempty"` + Libraries []*Library `protobuf:"bytes,102,rep,name=libraries,proto3" json:"libraries,omitempty"` // Serving configuration for // [Google Cloud Endpoints](https://cloud.google.com/appengine/docs/python/endpoints/). // // Only returned in `GET` requests if `view=FULL` is set. - ApiConfig *ApiConfigHandler `protobuf:"bytes,103,opt,name=api_config,json=apiConfig" json:"api_config,omitempty"` + ApiConfig *ApiConfigHandler `protobuf:"bytes,103,opt,name=api_config,json=apiConfig,proto3" json:"api_config,omitempty"` // Environment variables available to the application. // // Only returned in `GET` requests if `view=FULL` is set. - EnvVariables map[string]string `protobuf:"bytes,104,rep,name=env_variables,json=envVariables" json:"env_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + EnvVariables map[string]string `protobuf:"bytes,104,rep,name=env_variables,json=envVariables,proto3" json:"env_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Duration that static files should be cached by web proxies and browsers. // Only applicable if the corresponding // [StaticFilesHandler](https://cloud.google.com/appengine/docs/admin-api/reference/rest/v1/apps.services.versions#staticfileshandler) // does not specify its own expiration time. // // Only returned in `GET` requests if `view=FULL` is set. - DefaultExpiration *duration.Duration `protobuf:"bytes,105,opt,name=default_expiration,json=defaultExpiration" json:"default_expiration,omitempty"` + DefaultExpiration *duration.Duration `protobuf:"bytes,105,opt,name=default_expiration,json=defaultExpiration,proto3" json:"default_expiration,omitempty"` // Configures health checking for VM instances. Unhealthy instances are // stopped and replaced with new instances. Only applicable for VM // runtimes. // // Only returned in `GET` requests if `view=FULL` is set. - HealthCheck *HealthCheck `protobuf:"bytes,106,opt,name=health_check,json=healthCheck" json:"health_check,omitempty"` + HealthCheck *HealthCheck `protobuf:"bytes,106,opt,name=health_check,json=healthCheck,proto3" json:"health_check,omitempty"` // Files that match this pattern will not be built into this version. // Only applicable for Go runtimes. // // Only returned in `GET` requests if `view=FULL` is set. - NobuildFilesRegex string `protobuf:"bytes,107,opt,name=nobuild_files_regex,json=nobuildFilesRegex" json:"nobuild_files_regex,omitempty"` + NobuildFilesRegex string `protobuf:"bytes,107,opt,name=nobuild_files_regex,json=nobuildFilesRegex,proto3" json:"nobuild_files_regex,omitempty"` // Code and application artifacts that make up this version. // // Only returned in `GET` requests if `view=FULL` is set. - Deployment *Deployment `protobuf:"bytes,108,opt,name=deployment" json:"deployment,omitempty"` + Deployment *Deployment `protobuf:"bytes,108,opt,name=deployment,proto3" json:"deployment,omitempty"` // Serving URL for this version. Example: // "https://myversion-dot-myservice-dot-myapp.appspot.com" // // @OutputOnly - VersionUrl string `protobuf:"bytes,109,opt,name=version_url,json=versionUrl" json:"version_url,omitempty"` + VersionUrl string `protobuf:"bytes,109,opt,name=version_url,json=versionUrl,proto3" json:"version_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -260,13 +260,13 @@ type isVersion_Scaling interface { } type Version_AutomaticScaling struct { - AutomaticScaling *AutomaticScaling `protobuf:"bytes,3,opt,name=automatic_scaling,json=automaticScaling,oneof"` + AutomaticScaling *AutomaticScaling `protobuf:"bytes,3,opt,name=automatic_scaling,json=automaticScaling,proto3,oneof"` } type Version_BasicScaling struct { - BasicScaling *BasicScaling `protobuf:"bytes,4,opt,name=basic_scaling,json=basicScaling,oneof"` + BasicScaling *BasicScaling `protobuf:"bytes,4,opt,name=basic_scaling,json=basicScaling,proto3,oneof"` } type Version_ManualScaling struct { - ManualScaling *ManualScaling `protobuf:"bytes,5,opt,name=manual_scaling,json=manualScaling,oneof"` + ManualScaling *ManualScaling `protobuf:"bytes,5,opt,name=manual_scaling,json=manualScaling,proto3,oneof"` } func (*Version_AutomaticScaling) isVersion_Scaling() {} @@ -576,36 +576,36 @@ type AutomaticScaling struct { // [Autoscaler](https://cloud.google.com/compute/docs/autoscaler/) // should wait between changes to the number of virtual machines. // Only applicable for VM runtimes. - CoolDownPeriod *duration.Duration `protobuf:"bytes,1,opt,name=cool_down_period,json=coolDownPeriod" json:"cool_down_period,omitempty"` + CoolDownPeriod *duration.Duration `protobuf:"bytes,1,opt,name=cool_down_period,json=coolDownPeriod,proto3" json:"cool_down_period,omitempty"` // Target scaling by CPU usage. - CpuUtilization *CpuUtilization `protobuf:"bytes,2,opt,name=cpu_utilization,json=cpuUtilization" json:"cpu_utilization,omitempty"` + CpuUtilization *CpuUtilization `protobuf:"bytes,2,opt,name=cpu_utilization,json=cpuUtilization,proto3" json:"cpu_utilization,omitempty"` // Number of concurrent requests an automatic scaling instance can accept // before the scheduler spawns a new instance. // // Defaults to a runtime-specific value. - MaxConcurrentRequests int32 `protobuf:"varint,3,opt,name=max_concurrent_requests,json=maxConcurrentRequests" json:"max_concurrent_requests,omitempty"` + MaxConcurrentRequests int32 `protobuf:"varint,3,opt,name=max_concurrent_requests,json=maxConcurrentRequests,proto3" json:"max_concurrent_requests,omitempty"` // Maximum number of idle instances that should be maintained for this // version. - MaxIdleInstances int32 `protobuf:"varint,4,opt,name=max_idle_instances,json=maxIdleInstances" json:"max_idle_instances,omitempty"` + MaxIdleInstances int32 `protobuf:"varint,4,opt,name=max_idle_instances,json=maxIdleInstances,proto3" json:"max_idle_instances,omitempty"` // Maximum number of instances that should be started to handle requests. - MaxTotalInstances int32 `protobuf:"varint,5,opt,name=max_total_instances,json=maxTotalInstances" json:"max_total_instances,omitempty"` + MaxTotalInstances int32 `protobuf:"varint,5,opt,name=max_total_instances,json=maxTotalInstances,proto3" json:"max_total_instances,omitempty"` // Maximum amount of time that a request should wait in the pending queue // before starting a new instance to handle it. - MaxPendingLatency *duration.Duration `protobuf:"bytes,6,opt,name=max_pending_latency,json=maxPendingLatency" json:"max_pending_latency,omitempty"` + MaxPendingLatency *duration.Duration `protobuf:"bytes,6,opt,name=max_pending_latency,json=maxPendingLatency,proto3" json:"max_pending_latency,omitempty"` // Minimum number of idle instances that should be maintained for // this version. Only applicable for the default version of a service. - MinIdleInstances int32 `protobuf:"varint,7,opt,name=min_idle_instances,json=minIdleInstances" json:"min_idle_instances,omitempty"` + MinIdleInstances int32 `protobuf:"varint,7,opt,name=min_idle_instances,json=minIdleInstances,proto3" json:"min_idle_instances,omitempty"` // Minimum number of instances that should be maintained for this version. - MinTotalInstances int32 `protobuf:"varint,8,opt,name=min_total_instances,json=minTotalInstances" json:"min_total_instances,omitempty"` + MinTotalInstances int32 `protobuf:"varint,8,opt,name=min_total_instances,json=minTotalInstances,proto3" json:"min_total_instances,omitempty"` // Minimum amount of time a request should wait in the pending queue before // starting a new instance to handle it. - MinPendingLatency *duration.Duration `protobuf:"bytes,9,opt,name=min_pending_latency,json=minPendingLatency" json:"min_pending_latency,omitempty"` + MinPendingLatency *duration.Duration `protobuf:"bytes,9,opt,name=min_pending_latency,json=minPendingLatency,proto3" json:"min_pending_latency,omitempty"` // Target scaling by request utilization. - RequestUtilization *RequestUtilization `protobuf:"bytes,10,opt,name=request_utilization,json=requestUtilization" json:"request_utilization,omitempty"` + RequestUtilization *RequestUtilization `protobuf:"bytes,10,opt,name=request_utilization,json=requestUtilization,proto3" json:"request_utilization,omitempty"` // Target scaling by disk usage. - DiskUtilization *DiskUtilization `protobuf:"bytes,11,opt,name=disk_utilization,json=diskUtilization" json:"disk_utilization,omitempty"` + DiskUtilization *DiskUtilization `protobuf:"bytes,11,opt,name=disk_utilization,json=diskUtilization,proto3" json:"disk_utilization,omitempty"` // Target scaling by network usage. - NetworkUtilization *NetworkUtilization `protobuf:"bytes,12,opt,name=network_utilization,json=networkUtilization" json:"network_utilization,omitempty"` + NetworkUtilization *NetworkUtilization `protobuf:"bytes,12,opt,name=network_utilization,json=networkUtilization,proto3" json:"network_utilization,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -726,9 +726,9 @@ func (m *AutomaticScaling) GetNetworkUtilization() *NetworkUtilization { type BasicScaling struct { // Duration of time after the last request that an instance must wait before // the instance is shut down. - IdleTimeout *duration.Duration `protobuf:"bytes,1,opt,name=idle_timeout,json=idleTimeout" json:"idle_timeout,omitempty"` + IdleTimeout *duration.Duration `protobuf:"bytes,1,opt,name=idle_timeout,json=idleTimeout,proto3" json:"idle_timeout,omitempty"` // Maximum number of instances to create for this version. - MaxInstances int32 `protobuf:"varint,2,opt,name=max_instances,json=maxInstances" json:"max_instances,omitempty"` + MaxInstances int32 `protobuf:"varint,2,opt,name=max_instances,json=maxInstances,proto3" json:"max_instances,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -779,7 +779,7 @@ type ManualScaling struct { // can later be altered by using the // [Modules API](https://cloud.google.com/appengine/docs/python/modules/functions) // `set_num_instances()` function. - Instances int32 `protobuf:"varint,1,opt,name=instances" json:"instances,omitempty"` + Instances int32 `protobuf:"varint,1,opt,name=instances,proto3" json:"instances,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -819,10 +819,10 @@ func (m *ManualScaling) GetInstances() int32 { // Target scaling by CPU usage. type CpuUtilization struct { // Period of time over which CPU utilization is calculated. - AggregationWindowLength *duration.Duration `protobuf:"bytes,1,opt,name=aggregation_window_length,json=aggregationWindowLength" json:"aggregation_window_length,omitempty"` + AggregationWindowLength *duration.Duration `protobuf:"bytes,1,opt,name=aggregation_window_length,json=aggregationWindowLength,proto3" json:"aggregation_window_length,omitempty"` // Target CPU utilization ratio to maintain when scaling. Must be between 0 // and 1. - TargetUtilization float64 `protobuf:"fixed64,2,opt,name=target_utilization,json=targetUtilization" json:"target_utilization,omitempty"` + TargetUtilization float64 `protobuf:"fixed64,2,opt,name=target_utilization,json=targetUtilization,proto3" json:"target_utilization,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -869,9 +869,9 @@ func (m *CpuUtilization) GetTargetUtilization() float64 { // Target scaling by request utilization. Only applicable for VM runtimes. type RequestUtilization struct { // Target requests per second. - TargetRequestCountPerSecond int32 `protobuf:"varint,1,opt,name=target_request_count_per_second,json=targetRequestCountPerSecond" json:"target_request_count_per_second,omitempty"` + TargetRequestCountPerSecond int32 `protobuf:"varint,1,opt,name=target_request_count_per_second,json=targetRequestCountPerSecond,proto3" json:"target_request_count_per_second,omitempty"` // Target number of concurrent requests. - TargetConcurrentRequests int32 `protobuf:"varint,2,opt,name=target_concurrent_requests,json=targetConcurrentRequests" json:"target_concurrent_requests,omitempty"` + TargetConcurrentRequests int32 `protobuf:"varint,2,opt,name=target_concurrent_requests,json=targetConcurrentRequests,proto3" json:"target_concurrent_requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -918,13 +918,13 @@ func (m *RequestUtilization) GetTargetConcurrentRequests() int32 { // Target scaling by disk usage. Only applicable for VM runtimes. type DiskUtilization struct { // Target bytes written per second. - TargetWriteBytesPerSecond int32 `protobuf:"varint,14,opt,name=target_write_bytes_per_second,json=targetWriteBytesPerSecond" json:"target_write_bytes_per_second,omitempty"` + TargetWriteBytesPerSecond int32 `protobuf:"varint,14,opt,name=target_write_bytes_per_second,json=targetWriteBytesPerSecond,proto3" json:"target_write_bytes_per_second,omitempty"` // Target ops written per second. - TargetWriteOpsPerSecond int32 `protobuf:"varint,15,opt,name=target_write_ops_per_second,json=targetWriteOpsPerSecond" json:"target_write_ops_per_second,omitempty"` + TargetWriteOpsPerSecond int32 `protobuf:"varint,15,opt,name=target_write_ops_per_second,json=targetWriteOpsPerSecond,proto3" json:"target_write_ops_per_second,omitempty"` // Target bytes read per second. - TargetReadBytesPerSecond int32 `protobuf:"varint,16,opt,name=target_read_bytes_per_second,json=targetReadBytesPerSecond" json:"target_read_bytes_per_second,omitempty"` + TargetReadBytesPerSecond int32 `protobuf:"varint,16,opt,name=target_read_bytes_per_second,json=targetReadBytesPerSecond,proto3" json:"target_read_bytes_per_second,omitempty"` // Target ops read per seconds. - TargetReadOpsPerSecond int32 `protobuf:"varint,17,opt,name=target_read_ops_per_second,json=targetReadOpsPerSecond" json:"target_read_ops_per_second,omitempty"` + TargetReadOpsPerSecond int32 `protobuf:"varint,17,opt,name=target_read_ops_per_second,json=targetReadOpsPerSecond,proto3" json:"target_read_ops_per_second,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -985,13 +985,13 @@ func (m *DiskUtilization) GetTargetReadOpsPerSecond() int32 { // Target scaling by network usage. Only applicable for VM runtimes. type NetworkUtilization struct { // Target bytes sent per second. - TargetSentBytesPerSecond int32 `protobuf:"varint,1,opt,name=target_sent_bytes_per_second,json=targetSentBytesPerSecond" json:"target_sent_bytes_per_second,omitempty"` + TargetSentBytesPerSecond int32 `protobuf:"varint,1,opt,name=target_sent_bytes_per_second,json=targetSentBytesPerSecond,proto3" json:"target_sent_bytes_per_second,omitempty"` // Target packets sent per second. - TargetSentPacketsPerSecond int32 `protobuf:"varint,11,opt,name=target_sent_packets_per_second,json=targetSentPacketsPerSecond" json:"target_sent_packets_per_second,omitempty"` + TargetSentPacketsPerSecond int32 `protobuf:"varint,11,opt,name=target_sent_packets_per_second,json=targetSentPacketsPerSecond,proto3" json:"target_sent_packets_per_second,omitempty"` // Target bytes received per second. - TargetReceivedBytesPerSecond int32 `protobuf:"varint,12,opt,name=target_received_bytes_per_second,json=targetReceivedBytesPerSecond" json:"target_received_bytes_per_second,omitempty"` + TargetReceivedBytesPerSecond int32 `protobuf:"varint,12,opt,name=target_received_bytes_per_second,json=targetReceivedBytesPerSecond,proto3" json:"target_received_bytes_per_second,omitempty"` // Target packets received per second. - TargetReceivedPacketsPerSecond int32 `protobuf:"varint,13,opt,name=target_received_packets_per_second,json=targetReceivedPacketsPerSecond" json:"target_received_packets_per_second,omitempty"` + TargetReceivedPacketsPerSecond int32 `protobuf:"varint,13,opt,name=target_received_packets_per_second,json=targetReceivedPacketsPerSecond,proto3" json:"target_received_packets_per_second,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1053,14 +1053,14 @@ func (m *NetworkUtilization) GetTargetReceivedPacketsPerSecond() int32 { type Network struct { // List of ports, or port pairs, to forward from the virtual machine to the // application container. - ForwardedPorts []string `protobuf:"bytes,1,rep,name=forwarded_ports,json=forwardedPorts" json:"forwarded_ports,omitempty"` + ForwardedPorts []string `protobuf:"bytes,1,rep,name=forwarded_ports,json=forwardedPorts,proto3" json:"forwarded_ports,omitempty"` // Tag to apply to the VM instance during creation. - InstanceTag string `protobuf:"bytes,2,opt,name=instance_tag,json=instanceTag" json:"instance_tag,omitempty"` + InstanceTag string `protobuf:"bytes,2,opt,name=instance_tag,json=instanceTag,proto3" json:"instance_tag,omitempty"` // Google Cloud Platform network where the virtual machines are created. // Specify the short name, not the resource path. // // Defaults to `default`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1114,11 +1114,11 @@ func (m *Network) GetName() string { // Machine resources for a version. type Resources struct { // Number of CPU cores needed. - Cpu float64 `protobuf:"fixed64,1,opt,name=cpu" json:"cpu,omitempty"` + Cpu float64 `protobuf:"fixed64,1,opt,name=cpu,proto3" json:"cpu,omitempty"` // Disk size (GB) needed. - DiskGb float64 `protobuf:"fixed64,2,opt,name=disk_gb,json=diskGb" json:"disk_gb,omitempty"` + DiskGb float64 `protobuf:"fixed64,2,opt,name=disk_gb,json=diskGb,proto3" json:"disk_gb,omitempty"` // Memory (GB) needed. - MemoryGb float64 `protobuf:"fixed64,3,opt,name=memory_gb,json=memoryGb" json:"memory_gb,omitempty"` + MemoryGb float64 `protobuf:"fixed64,3,opt,name=memory_gb,json=memoryGb,proto3" json:"memory_gb,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha1/embedded_assistant.pb.go b/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha1/embedded_assistant.pb.go index 39c26f4dd..3f8e8f2cb 100644 --- a/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha1/embedded_assistant.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha1/embedded_assistant.pb.go @@ -168,11 +168,11 @@ func (ConverseResponse_EventType) EnumDescriptor() ([]byte, []int) { // Specifies how to process the `ConverseRequest` messages. type ConverseConfig struct { // *Required* Specifies how to process the subsequent incoming audio. - AudioInConfig *AudioInConfig `protobuf:"bytes,1,opt,name=audio_in_config,json=audioInConfig" json:"audio_in_config,omitempty"` + AudioInConfig *AudioInConfig `protobuf:"bytes,1,opt,name=audio_in_config,json=audioInConfig,proto3" json:"audio_in_config,omitempty"` // *Required* Specifies how to format the audio that will be returned. - AudioOutConfig *AudioOutConfig `protobuf:"bytes,2,opt,name=audio_out_config,json=audioOutConfig" json:"audio_out_config,omitempty"` + AudioOutConfig *AudioOutConfig `protobuf:"bytes,2,opt,name=audio_out_config,json=audioOutConfig,proto3" json:"audio_out_config,omitempty"` // *Required* Represents the current dialog state. - ConverseState *ConverseState `protobuf:"bytes,3,opt,name=converse_state,json=converseState" json:"converse_state,omitempty"` + ConverseState *ConverseState `protobuf:"bytes,3,opt,name=converse_state,json=converseState,proto3" json:"converse_state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -228,13 +228,13 @@ func (m *ConverseConfig) GetConverseState() *ConverseState { // [best practices](https://developers.google.com/assistant/sdk/develop/grpc/best-practices/audio). type AudioInConfig struct { // *Required* Encoding of audio data sent in all `audio_in` messages. - Encoding AudioInConfig_Encoding `protobuf:"varint,1,opt,name=encoding,enum=google.assistant.embedded.v1alpha1.AudioInConfig_Encoding" json:"encoding,omitempty"` + Encoding AudioInConfig_Encoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.assistant.embedded.v1alpha1.AudioInConfig_Encoding" json:"encoding,omitempty"` // *Required* Sample rate (in Hertz) of the audio data sent in all `audio_in` // messages. Valid values are from 16000-24000, but 16000 is optimal. // For best results, set the sampling rate of the audio source to 16000 Hz. // If that's not possible, use the native sample rate of the audio source // (instead of re-sampling). - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -283,13 +283,13 @@ func (m *AudioInConfig) GetSampleRateHertz() int32 { type AudioOutConfig struct { // *Required* The encoding of audio data to be returned in all `audio_out` // messages. - Encoding AudioOutConfig_Encoding `protobuf:"varint,1,opt,name=encoding,enum=google.assistant.embedded.v1alpha1.AudioOutConfig_Encoding" json:"encoding,omitempty"` + Encoding AudioOutConfig_Encoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.assistant.embedded.v1alpha1.AudioOutConfig_Encoding" json:"encoding,omitempty"` // *Required* The sample rate in Hertz of the audio data returned in // `audio_out` messages. Valid values are: 16000-24000. - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` // *Required* Current volume setting of the device's audio output. // Valid values are 1 to 100 (corresponding to 1% to 100%). - VolumePercentage int32 `protobuf:"varint,3,opt,name=volume_percentage,json=volumePercentage" json:"volume_percentage,omitempty"` + VolumePercentage int32 `protobuf:"varint,3,opt,name=volume_percentage,json=volumePercentage,proto3" json:"volume_percentage,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -430,10 +430,10 @@ func (m *AudioOut) GetAudioData() []byte { // The semantic result for the user's spoken query. type ConverseResult struct { // *Output-only* The recognized transcript of what the user said. - SpokenRequestText string `protobuf:"bytes,1,opt,name=spoken_request_text,json=spokenRequestText" json:"spoken_request_text,omitempty"` + SpokenRequestText string `protobuf:"bytes,1,opt,name=spoken_request_text,json=spokenRequestText,proto3" json:"spoken_request_text,omitempty"` // *Output-only* The text of the assistant's spoken response. This is only // returned for an IFTTT action. - SpokenResponseText string `protobuf:"bytes,2,opt,name=spoken_response_text,json=spokenResponseText" json:"spoken_response_text,omitempty"` + SpokenResponseText string `protobuf:"bytes,2,opt,name=spoken_response_text,json=spokenResponseText,proto3" json:"spoken_response_text,omitempty"` // *Output-only* State information for subsequent `ConverseRequest`. This // value should be saved in the client and returned in the // `conversation_state` with the next `ConverseRequest`. (The client does not @@ -442,7 +442,7 @@ type ConverseResult struct { ConversationState []byte `protobuf:"bytes,3,opt,name=conversation_state,json=conversationState,proto3" json:"conversation_state,omitempty"` // *Output-only* Specifies the mode of the microphone after this `Converse` // RPC is processed. - MicrophoneMode ConverseResult_MicrophoneMode `protobuf:"varint,4,opt,name=microphone_mode,json=microphoneMode,enum=google.assistant.embedded.v1alpha1.ConverseResult_MicrophoneMode" json:"microphone_mode,omitempty"` + MicrophoneMode ConverseResult_MicrophoneMode `protobuf:"varint,4,opt,name=microphone_mode,json=microphoneMode,proto3,enum=google.assistant.embedded.v1alpha1.ConverseResult_MicrophoneMode" json:"microphone_mode,omitempty"` // *Output-only* Updated volume level. The value will be 0 or omitted // (indicating no change) unless a voice command such as "Increase the volume" // or "Set volume level 4" was recognized, in which case the value will be @@ -452,7 +452,7 @@ type ConverseResult struct { // supply it in the `AudioOutConfig` of the next `ConverseRequest`. (Some // clients may also implement other ways to allow the current volume level to // be changed, for example, by providing a knob that the user can turn.) - VolumePercentage int32 `protobuf:"varint,5,opt,name=volume_percentage,json=volumePercentage" json:"volume_percentage,omitempty"` + VolumePercentage int32 `protobuf:"varint,5,opt,name=volume_percentage,json=volumePercentage,proto3" json:"volume_percentage,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -563,7 +563,7 @@ type isConverseRequest_ConverseRequest interface { } type ConverseRequest_Config struct { - Config *ConverseConfig `protobuf:"bytes,1,opt,name=config,oneof"` + Config *ConverseConfig `protobuf:"bytes,1,opt,name=config,proto3,oneof"` } type ConverseRequest_AudioIn struct { AudioIn []byte `protobuf:"bytes,2,opt,name=audio_in,json=audioIn,proto3,oneof"` @@ -708,16 +708,16 @@ type isConverseResponse_ConverseResponse interface { } type ConverseResponse_Error struct { - Error *status.Status `protobuf:"bytes,1,opt,name=error,oneof"` + Error *status.Status `protobuf:"bytes,1,opt,name=error,proto3,oneof"` } type ConverseResponse_EventType_ struct { - EventType ConverseResponse_EventType `protobuf:"varint,2,opt,name=event_type,json=eventType,enum=google.assistant.embedded.v1alpha1.ConverseResponse_EventType,oneof"` + EventType ConverseResponse_EventType `protobuf:"varint,2,opt,name=event_type,json=eventType,proto3,enum=google.assistant.embedded.v1alpha1.ConverseResponse_EventType,oneof"` } type ConverseResponse_AudioOut struct { - AudioOut *AudioOut `protobuf:"bytes,3,opt,name=audio_out,json=audioOut,oneof"` + AudioOut *AudioOut `protobuf:"bytes,3,opt,name=audio_out,json=audioOut,proto3,oneof"` } type ConverseResponse_Result struct { - Result *ConverseResult `protobuf:"bytes,5,opt,name=result,oneof"` + Result *ConverseResult `protobuf:"bytes,5,opt,name=result,proto3,oneof"` } func (*ConverseResponse_Error) isConverseResponse_ConverseResponse() {} @@ -977,8 +977,7 @@ func (x *embeddedAssistantConverseClient) Recv() (*ConverseResponse, error) { return m, nil } -// Server API for EmbeddedAssistant service - +// EmbeddedAssistantServer is the server API for EmbeddedAssistant service. type EmbeddedAssistantServer interface { // Initiates or continues a conversation with the embedded assistant service. // Each call performs one round-trip, sending an audio request to the service diff --git a/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha2/embedded_assistant.pb.go b/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha2/embedded_assistant.pb.go index bd4067154..d6f628ef1 100644 --- a/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha2/embedded_assistant.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/assistant/embedded/v1alpha2/embedded_assistant.pb.go @@ -53,7 +53,7 @@ func (x AssistResponse_EventType) String() string { return proto.EnumName(AssistResponse_EventType_name, int32(x)) } func (AssistResponse_EventType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{1, 0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{1, 0} } // Audio encoding of the data sent in the audio message. @@ -91,7 +91,7 @@ func (x AudioInConfig_Encoding) String() string { return proto.EnumName(AudioInConfig_Encoding_name, int32(x)) } func (AudioInConfig_Encoding) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{3, 0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{4, 0} } // Audio encoding of the data returned in the audio message. All encodings are @@ -129,7 +129,7 @@ func (x AudioOutConfig_Encoding) String() string { return proto.EnumName(AudioOutConfig_Encoding_name, int32(x)) } func (AudioOutConfig_Encoding) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{4, 0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{5, 0} } // Possible modes for visual screen-output on the device. @@ -163,7 +163,7 @@ func (x ScreenOutConfig_ScreenMode) String() string { return proto.EnumName(ScreenOutConfig_ScreenMode_name, int32(x)) } func (ScreenOutConfig_ScreenMode) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{5, 0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{6, 0} } // Possible formats of the screen data. @@ -173,7 +173,7 @@ const ( // No format specified. ScreenOut_FORMAT_UNSPECIFIED ScreenOut_Format = 0 // Data will contain a fully-formed HTML5 layout encoded in UTF-8, e.g. - // "
...
". It is intended to be rendered + // `
...
`. It is intended to be rendered // along with the audio response. Note that HTML5 doctype should be included // in the actual HTML data. ScreenOut_HTML ScreenOut_Format = 1 @@ -192,7 +192,7 @@ func (x ScreenOut_Format) String() string { return proto.EnumName(ScreenOut_Format_name, int32(x)) } func (ScreenOut_Format) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{9, 0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{10, 0} } // Possible states of the microphone after a `Assist` RPC completes. @@ -225,7 +225,7 @@ func (x DialogStateOut_MicrophoneMode) String() string { return proto.EnumName(DialogStateOut_MicrophoneMode_name, int32(x)) } func (DialogStateOut_MicrophoneMode) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{12, 0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{13, 0} } // The top-level message sent by the client. Clients must send at least two, and @@ -249,7 +249,7 @@ func (m *AssistRequest) Reset() { *m = AssistRequest{} } func (m *AssistRequest) String() string { return proto.CompactTextString(m) } func (*AssistRequest) ProtoMessage() {} func (*AssistRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{0} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{0} } func (m *AssistRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AssistRequest.Unmarshal(m, b) @@ -274,7 +274,7 @@ type isAssistRequest_Type interface { } type AssistRequest_Config struct { - Config *AssistConfig `protobuf:"bytes,1,opt,name=config,oneof"` + Config *AssistConfig `protobuf:"bytes,1,opt,name=config,proto3,oneof"` } type AssistRequest_AudioIn struct { AudioIn []byte `protobuf:"bytes,2,opt,name=audio_in,json=audioIn,proto3,oneof"` @@ -378,14 +378,14 @@ func _AssistRequest_OneofSizer(msg proto.Message) (n int) { // `AssistResponse` messages are streamed back to the client. type AssistResponse struct { // *Output-only* Indicates the type of event. - EventType AssistResponse_EventType `protobuf:"varint,1,opt,name=event_type,json=eventType,enum=google.assistant.embedded.v1alpha2.AssistResponse_EventType" json:"event_type,omitempty"` + EventType AssistResponse_EventType `protobuf:"varint,1,opt,name=event_type,json=eventType,proto3,enum=google.assistant.embedded.v1alpha2.AssistResponse_EventType" json:"event_type,omitempty"` // *Output-only* The audio containing the Assistant's response to the query. - AudioOut *AudioOut `protobuf:"bytes,3,opt,name=audio_out,json=audioOut" json:"audio_out,omitempty"` + AudioOut *AudioOut `protobuf:"bytes,3,opt,name=audio_out,json=audioOut,proto3" json:"audio_out,omitempty"` // *Output-only* Contains the Assistant's visual response to the query. - ScreenOut *ScreenOut `protobuf:"bytes,4,opt,name=screen_out,json=screenOut" json:"screen_out,omitempty"` + ScreenOut *ScreenOut `protobuf:"bytes,4,opt,name=screen_out,json=screenOut,proto3" json:"screen_out,omitempty"` // *Output-only* Contains the action triggered by the query with the // appropriate payloads and semantic parsing. - DeviceAction *DeviceAction `protobuf:"bytes,6,opt,name=device_action,json=deviceAction" json:"device_action,omitempty"` + DeviceAction *DeviceAction `protobuf:"bytes,6,opt,name=device_action,json=deviceAction,proto3" json:"device_action,omitempty"` // *Output-only* This repeated list contains zero or more speech recognition // results that correspond to consecutive portions of the audio currently // being processed, starting with the portion corresponding to the earliest @@ -393,19 +393,22 @@ type AssistResponse struct { // recent audio. The strings can be concatenated to view the full // in-progress response. When the speech recognition completes, this list // will contain one item with `stability` of `1.0`. - SpeechResults []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=speech_results,json=speechResults" json:"speech_results,omitempty"` + SpeechResults []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=speech_results,json=speechResults,proto3" json:"speech_results,omitempty"` // *Output-only* Contains output related to the user's query. - DialogStateOut *DialogStateOut `protobuf:"bytes,5,opt,name=dialog_state_out,json=dialogStateOut" json:"dialog_state_out,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + DialogStateOut *DialogStateOut `protobuf:"bytes,5,opt,name=dialog_state_out,json=dialogStateOut,proto3" json:"dialog_state_out,omitempty"` + // *Output-only* Debugging info for developer. Only returned if request set + // `return_debug_info` to true. + DebugInfo *DebugInfo `protobuf:"bytes,8,opt,name=debug_info,json=debugInfo,proto3" json:"debug_info,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *AssistResponse) Reset() { *m = AssistResponse{} } func (m *AssistResponse) String() string { return proto.CompactTextString(m) } func (*AssistResponse) ProtoMessage() {} func (*AssistResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{1} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{1} } func (m *AssistResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AssistResponse.Unmarshal(m, b) @@ -467,6 +470,58 @@ func (m *AssistResponse) GetDialogStateOut() *DialogStateOut { return nil } +func (m *AssistResponse) GetDebugInfo() *DebugInfo { + if m != nil { + return m.DebugInfo + } + return nil +} + +// Debug info for developer. Only returned if request set `return_debug_info` +// to true. +type DebugInfo struct { + // The original JSON response from an Action-on-Google agent to Google server. + // See + // https://developers.google.com/actions/reference/rest/Shared.Types/AppResponse. + // It will only be populated if the request maker owns the AoG project and the + // AoG project is in preview mode. + AogAgentToAssistantJson string `protobuf:"bytes,1,opt,name=aog_agent_to_assistant_json,json=aogAgentToAssistantJson,proto3" json:"aog_agent_to_assistant_json,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DebugInfo) Reset() { *m = DebugInfo{} } +func (m *DebugInfo) String() string { return proto.CompactTextString(m) } +func (*DebugInfo) ProtoMessage() {} +func (*DebugInfo) Descriptor() ([]byte, []int) { + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{2} +} +func (m *DebugInfo) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DebugInfo.Unmarshal(m, b) +} +func (m *DebugInfo) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DebugInfo.Marshal(b, m, deterministic) +} +func (dst *DebugInfo) XXX_Merge(src proto.Message) { + xxx_messageInfo_DebugInfo.Merge(dst, src) +} +func (m *DebugInfo) XXX_Size() int { + return xxx_messageInfo_DebugInfo.Size(m) +} +func (m *DebugInfo) XXX_DiscardUnknown() { + xxx_messageInfo_DebugInfo.DiscardUnknown(m) +} + +var xxx_messageInfo_DebugInfo proto.InternalMessageInfo + +func (m *DebugInfo) GetAogAgentToAssistantJson() string { + if m != nil { + return m.AogAgentToAssistantJson + } + return "" +} + // Specifies how to process the `AssistRequest` messages. type AssistConfig struct { // Types that are valid to be assigned to Type: @@ -474,24 +529,26 @@ type AssistConfig struct { // *AssistConfig_TextQuery Type isAssistConfig_Type `protobuf_oneof:"type"` // *Required* Specifies how to format the audio that will be returned. - AudioOutConfig *AudioOutConfig `protobuf:"bytes,2,opt,name=audio_out_config,json=audioOutConfig" json:"audio_out_config,omitempty"` + AudioOutConfig *AudioOutConfig `protobuf:"bytes,2,opt,name=audio_out_config,json=audioOutConfig,proto3" json:"audio_out_config,omitempty"` // *Optional* Specifies the desired format to use when server returns a // visual screen response. - ScreenOutConfig *ScreenOutConfig `protobuf:"bytes,8,opt,name=screen_out_config,json=screenOutConfig" json:"screen_out_config,omitempty"` + ScreenOutConfig *ScreenOutConfig `protobuf:"bytes,8,opt,name=screen_out_config,json=screenOutConfig,proto3" json:"screen_out_config,omitempty"` // *Required* Represents the current dialog state. - DialogStateIn *DialogStateIn `protobuf:"bytes,3,opt,name=dialog_state_in,json=dialogStateIn" json:"dialog_state_in,omitempty"` + DialogStateIn *DialogStateIn `protobuf:"bytes,3,opt,name=dialog_state_in,json=dialogStateIn,proto3" json:"dialog_state_in,omitempty"` // Device configuration that uniquely identifies a specific device. - DeviceConfig *DeviceConfig `protobuf:"bytes,4,opt,name=device_config,json=deviceConfig" json:"device_config,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + DeviceConfig *DeviceConfig `protobuf:"bytes,4,opt,name=device_config,json=deviceConfig,proto3" json:"device_config,omitempty"` + // *Optional* Debugging parameters for the whole `Assist` RPC. + DebugConfig *DebugConfig `protobuf:"bytes,5,opt,name=debug_config,json=debugConfig,proto3" json:"debug_config,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *AssistConfig) Reset() { *m = AssistConfig{} } func (m *AssistConfig) String() string { return proto.CompactTextString(m) } func (*AssistConfig) ProtoMessage() {} func (*AssistConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{2} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{3} } func (m *AssistConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AssistConfig.Unmarshal(m, b) @@ -516,10 +573,10 @@ type isAssistConfig_Type interface { } type AssistConfig_AudioInConfig struct { - AudioInConfig *AudioInConfig `protobuf:"bytes,1,opt,name=audio_in_config,json=audioInConfig,oneof"` + AudioInConfig *AudioInConfig `protobuf:"bytes,1,opt,name=audio_in_config,json=audioInConfig,proto3,oneof"` } type AssistConfig_TextQuery struct { - TextQuery string `protobuf:"bytes,6,opt,name=text_query,json=textQuery,oneof"` + TextQuery string `protobuf:"bytes,6,opt,name=text_query,json=textQuery,proto3,oneof"` } func (*AssistConfig_AudioInConfig) isAssistConfig_Type() {} @@ -574,6 +631,13 @@ func (m *AssistConfig) GetDeviceConfig() *DeviceConfig { return nil } +func (m *AssistConfig) GetDebugConfig() *DebugConfig { + if m != nil { + return m.DebugConfig + } + return nil +} + // XXX_OneofFuncs is for the internal use of the proto package. func (*AssistConfig) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { return _AssistConfig_OneofMarshaler, _AssistConfig_OneofUnmarshaler, _AssistConfig_OneofSizer, []interface{}{ @@ -649,13 +713,13 @@ func _AssistConfig_OneofSizer(msg proto.Message) (n int) { // [best practices](https://developers.google.com/assistant/sdk/guides/service/python/best-practices/audio). type AudioInConfig struct { // *Required* Encoding of audio data sent in all `audio_in` messages. - Encoding AudioInConfig_Encoding `protobuf:"varint,1,opt,name=encoding,enum=google.assistant.embedded.v1alpha2.AudioInConfig_Encoding" json:"encoding,omitempty"` + Encoding AudioInConfig_Encoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.assistant.embedded.v1alpha2.AudioInConfig_Encoding" json:"encoding,omitempty"` // *Required* Sample rate (in Hertz) of the audio data sent in all `audio_in` // messages. Valid values are from 16000-24000, but 16000 is optimal. // For best results, set the sampling rate of the audio source to 16000 Hz. // If that's not possible, use the native sample rate of the audio source // (instead of re-sampling). - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -665,7 +729,7 @@ func (m *AudioInConfig) Reset() { *m = AudioInConfig{} } func (m *AudioInConfig) String() string { return proto.CompactTextString(m) } func (*AudioInConfig) ProtoMessage() {} func (*AudioInConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{3} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{4} } func (m *AudioInConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AudioInConfig.Unmarshal(m, b) @@ -704,13 +768,13 @@ func (m *AudioInConfig) GetSampleRateHertz() int32 { type AudioOutConfig struct { // *Required* The encoding of audio data to be returned in all `audio_out` // messages. - Encoding AudioOutConfig_Encoding `protobuf:"varint,1,opt,name=encoding,enum=google.assistant.embedded.v1alpha2.AudioOutConfig_Encoding" json:"encoding,omitempty"` + Encoding AudioOutConfig_Encoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.assistant.embedded.v1alpha2.AudioOutConfig_Encoding" json:"encoding,omitempty"` // *Required* The sample rate in Hertz of the audio data returned in // `audio_out` messages. Valid values are: 16000-24000. - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` // *Required* Current volume setting of the device's audio output. // Valid values are 1 to 100 (corresponding to 1% to 100%). - VolumePercentage int32 `protobuf:"varint,3,opt,name=volume_percentage,json=volumePercentage" json:"volume_percentage,omitempty"` + VolumePercentage int32 `protobuf:"varint,3,opt,name=volume_percentage,json=volumePercentage,proto3" json:"volume_percentage,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -720,7 +784,7 @@ func (m *AudioOutConfig) Reset() { *m = AudioOutConfig{} } func (m *AudioOutConfig) String() string { return proto.CompactTextString(m) } func (*AudioOutConfig) ProtoMessage() {} func (*AudioOutConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{4} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{5} } func (m *AudioOutConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AudioOutConfig.Unmarshal(m, b) @@ -765,7 +829,7 @@ func (m *AudioOutConfig) GetVolumePercentage() int32 { // `screen_out` response. type ScreenOutConfig struct { // Current visual screen-mode for the device while issuing the query. - ScreenMode ScreenOutConfig_ScreenMode `protobuf:"varint,1,opt,name=screen_mode,json=screenMode,enum=google.assistant.embedded.v1alpha2.ScreenOutConfig_ScreenMode" json:"screen_mode,omitempty"` + ScreenMode ScreenOutConfig_ScreenMode `protobuf:"varint,1,opt,name=screen_mode,json=screenMode,proto3,enum=google.assistant.embedded.v1alpha2.ScreenOutConfig_ScreenMode" json:"screen_mode,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -775,7 +839,7 @@ func (m *ScreenOutConfig) Reset() { *m = ScreenOutConfig{} } func (m *ScreenOutConfig) String() string { return proto.CompactTextString(m) } func (*ScreenOutConfig) ProtoMessage() {} func (*ScreenOutConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{5} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{6} } func (m *ScreenOutConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ScreenOutConfig.Unmarshal(m, b) @@ -811,19 +875,20 @@ type DialogStateIn struct { // device after it was first setup and/or a factory-default reset. ConversationState []byte `protobuf:"bytes,1,opt,name=conversation_state,json=conversationState,proto3" json:"conversation_state,omitempty"` // *Required* Language of the request in - // [IETF BCP 47 syntax](https://tools.ietf.org/html/bcp47). For example: - // "en-US". If you have selected a language for this `device_id` using the - // [Settings](https://developers.google.com/assistant/sdk/guides/assistant-settings) + // [IETF BCP 47 syntax](https://tools.ietf.org/html/bcp47) (for example, + // "en-US"). See [Language Support](https://developers.google.com/assistant/sdk/reference/rpc/languages) + // for more information. If you have selected a language for this `device_id` + // using the [Settings](https://developers.google.com/assistant/sdk/reference/assistant-app/assistant-settings) // menu in your phone's Google Assistant app, that selection will override // this value. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // *Optional* Location of the device where the query originated. - DeviceLocation *DeviceLocation `protobuf:"bytes,5,opt,name=device_location,json=deviceLocation" json:"device_location,omitempty"` + DeviceLocation *DeviceLocation `protobuf:"bytes,5,opt,name=device_location,json=deviceLocation,proto3" json:"device_location,omitempty"` // *Optional* If true, the server will treat the request as a new conversation // and not use state from the prior request. Set this field to true when the // conversation should be restarted, such as after a device reboot, or after a // significant lapse of time since the prior query. - IsNewConversation bool `protobuf:"varint,7,opt,name=is_new_conversation,json=isNewConversation" json:"is_new_conversation,omitempty"` + IsNewConversation bool `protobuf:"varint,7,opt,name=is_new_conversation,json=isNewConversation,proto3" json:"is_new_conversation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -833,7 +898,7 @@ func (m *DialogStateIn) Reset() { *m = DialogStateIn{} } func (m *DialogStateIn) String() string { return proto.CompactTextString(m) } func (*DialogStateIn) ProtoMessage() {} func (*DialogStateIn) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{6} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{7} } func (m *DialogStateIn) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DialogStateIn.Unmarshal(m, b) @@ -899,11 +964,11 @@ type DeviceConfig struct { // capabilities of this device. This information should not change across // device reboots. However, it should not be saved across // factory-default resets. - DeviceId string `protobuf:"bytes,1,opt,name=device_id,json=deviceId" json:"device_id,omitempty"` + DeviceId string `protobuf:"bytes,1,opt,name=device_id,json=deviceId,proto3" json:"device_id,omitempty"` // *Required* Unique identifier for the device model. The combination of // device_model_id and device_id must have been previously associated through // device registration. - DeviceModelId string `protobuf:"bytes,3,opt,name=device_model_id,json=deviceModelId" json:"device_model_id,omitempty"` + DeviceModelId string `protobuf:"bytes,3,opt,name=device_model_id,json=deviceModelId,proto3" json:"device_model_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -913,7 +978,7 @@ func (m *DeviceConfig) Reset() { *m = DeviceConfig{} } func (m *DeviceConfig) String() string { return proto.CompactTextString(m) } func (*DeviceConfig) ProtoMessage() {} func (*DeviceConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{7} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{8} } func (m *DeviceConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeviceConfig.Unmarshal(m, b) @@ -963,7 +1028,7 @@ func (m *AudioOut) Reset() { *m = AudioOut{} } func (m *AudioOut) String() string { return proto.CompactTextString(m) } func (*AudioOut) ProtoMessage() {} func (*AudioOut) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{8} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{9} } func (m *AudioOut) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AudioOut.Unmarshal(m, b) @@ -990,10 +1055,11 @@ func (m *AudioOut) GetAudioData() []byte { return nil } -// The Assistant's visual response to query. Contains the entire visual output. +// The Assistant's visual output response to query. Enabled by +// `screen_out_config`. type ScreenOut struct { // *Output-only* The format of the provided screen data. - Format ScreenOut_Format `protobuf:"varint,1,opt,name=format,enum=google.assistant.embedded.v1alpha2.ScreenOut_Format" json:"format,omitempty"` + Format ScreenOut_Format `protobuf:"varint,1,opt,name=format,proto3,enum=google.assistant.embedded.v1alpha2.ScreenOut_Format" json:"format,omitempty"` // *Output-only* The raw screen data to be displayed as the result of the // Assistant query. Data []byte `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` @@ -1006,7 +1072,7 @@ func (m *ScreenOut) Reset() { *m = ScreenOut{} } func (m *ScreenOut) String() string { return proto.CompactTextString(m) } func (*ScreenOut) ProtoMessage() {} func (*ScreenOut) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{9} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{10} } func (m *ScreenOut) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ScreenOut.Unmarshal(m, b) @@ -1049,7 +1115,7 @@ type DeviceAction struct { // Device Action grammar. The format is given by the // `action.devices.EXECUTE` intent for a given // [trait](https://developers.google.com/assistant/sdk/reference/traits/). - DeviceRequestJson string `protobuf:"bytes,1,opt,name=device_request_json,json=deviceRequestJson" json:"device_request_json,omitempty"` + DeviceRequestJson string `protobuf:"bytes,1,opt,name=device_request_json,json=deviceRequestJson,proto3" json:"device_request_json,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1059,7 +1125,7 @@ func (m *DeviceAction) Reset() { *m = DeviceAction{} } func (m *DeviceAction) String() string { return proto.CompactTextString(m) } func (*DeviceAction) ProtoMessage() {} func (*DeviceAction) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{10} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{11} } func (m *DeviceAction) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeviceAction.Unmarshal(m, b) @@ -1090,12 +1156,12 @@ func (m *DeviceAction) GetDeviceRequestJson() string { // a single segment or the full guess of the user's spoken query. type SpeechRecognitionResult struct { // *Output-only* Transcript text representing the words that the user spoke. - Transcript string `protobuf:"bytes,1,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,1,opt,name=transcript,proto3" json:"transcript,omitempty"` // *Output-only* An estimate of the likelihood that the Assistant will not // change its guess about this result. Values range from 0.0 (completely // unstable) to 1.0 (completely stable and final). The default of 0.0 is a // sentinel value indicating `stability` was not set. - Stability float32 `protobuf:"fixed32,2,opt,name=stability" json:"stability,omitempty"` + Stability float32 `protobuf:"fixed32,2,opt,name=stability,proto3" json:"stability,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1105,7 +1171,7 @@ func (m *SpeechRecognitionResult) Reset() { *m = SpeechRecognitionResult func (m *SpeechRecognitionResult) String() string { return proto.CompactTextString(m) } func (*SpeechRecognitionResult) ProtoMessage() {} func (*SpeechRecognitionResult) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{11} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{12} } func (m *SpeechRecognitionResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SpeechRecognitionResult.Unmarshal(m, b) @@ -1145,7 +1211,7 @@ type DialogStateOut struct { // *Output-only* Supplemental display text from the Assistant. This could be // the same as the speech spoken in `AssistResponse.audio_out` or it could // be some additional information which aids the user's understanding. - SupplementalDisplayText string `protobuf:"bytes,1,opt,name=supplemental_display_text,json=supplementalDisplayText" json:"supplemental_display_text,omitempty"` + SupplementalDisplayText string `protobuf:"bytes,1,opt,name=supplemental_display_text,json=supplementalDisplayText,proto3" json:"supplemental_display_text,omitempty"` // *Output-only* State information for the subsequent `Assist` RPC. This // value should be saved in the client and returned in the // [`DialogStateIn.conversation_state`](#dialogstatein) field with the next @@ -1156,7 +1222,7 @@ type DialogStateOut struct { ConversationState []byte `protobuf:"bytes,2,opt,name=conversation_state,json=conversationState,proto3" json:"conversation_state,omitempty"` // *Output-only* Specifies the mode of the microphone after this `Assist` // RPC is processed. - MicrophoneMode DialogStateOut_MicrophoneMode `protobuf:"varint,3,opt,name=microphone_mode,json=microphoneMode,enum=google.assistant.embedded.v1alpha2.DialogStateOut_MicrophoneMode" json:"microphone_mode,omitempty"` + MicrophoneMode DialogStateOut_MicrophoneMode `protobuf:"varint,3,opt,name=microphone_mode,json=microphoneMode,proto3,enum=google.assistant.embedded.v1alpha2.DialogStateOut_MicrophoneMode" json:"microphone_mode,omitempty"` // *Output-only* Updated volume level. The value will be 0 or omitted // (indicating no change) unless a voice command such as *Increase the volume* // or *Set volume level 4* was recognized, in which case the value will be @@ -1166,7 +1232,7 @@ type DialogStateOut struct { // supply it in the `AudioOutConfig` of the next `AssistRequest`. (Some // clients may also implement other ways to allow the current volume level to // be changed, for example, by providing a knob that the user can turn.) - VolumePercentage int32 `protobuf:"varint,4,opt,name=volume_percentage,json=volumePercentage" json:"volume_percentage,omitempty"` + VolumePercentage int32 `protobuf:"varint,4,opt,name=volume_percentage,json=volumePercentage,proto3" json:"volume_percentage,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1176,7 +1242,7 @@ func (m *DialogStateOut) Reset() { *m = DialogStateOut{} } func (m *DialogStateOut) String() string { return proto.CompactTextString(m) } func (*DialogStateOut) ProtoMessage() {} func (*DialogStateOut) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{12} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{13} } func (m *DialogStateOut) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DialogStateOut.Unmarshal(m, b) @@ -1224,6 +1290,48 @@ func (m *DialogStateOut) GetVolumePercentage() int32 { return 0 } +// Debugging parameters for the current request. +type DebugConfig struct { + // When this field is set to true, the `debug_info` field in `AssistResponse` + // may be populated. However it will significantly increase latency of + // responses. Do not set this field true in production code. + ReturnDebugInfo bool `protobuf:"varint,6,opt,name=return_debug_info,json=returnDebugInfo,proto3" json:"return_debug_info,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DebugConfig) Reset() { *m = DebugConfig{} } +func (m *DebugConfig) String() string { return proto.CompactTextString(m) } +func (*DebugConfig) ProtoMessage() {} +func (*DebugConfig) Descriptor() ([]byte, []int) { + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{14} +} +func (m *DebugConfig) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DebugConfig.Unmarshal(m, b) +} +func (m *DebugConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DebugConfig.Marshal(b, m, deterministic) +} +func (dst *DebugConfig) XXX_Merge(src proto.Message) { + xxx_messageInfo_DebugConfig.Merge(dst, src) +} +func (m *DebugConfig) XXX_Size() int { + return xxx_messageInfo_DebugConfig.Size(m) +} +func (m *DebugConfig) XXX_DiscardUnknown() { + xxx_messageInfo_DebugConfig.DiscardUnknown(m) +} + +var xxx_messageInfo_DebugConfig proto.InternalMessageInfo + +func (m *DebugConfig) GetReturnDebugInfo() bool { + if m != nil { + return m.ReturnDebugInfo + } + return false +} + // There are three sources of locations. They are used with this precedence: // // 1. This `DeviceLocation`, which is primarily used for mobile devices with @@ -1245,7 +1353,7 @@ func (m *DeviceLocation) Reset() { *m = DeviceLocation{} } func (m *DeviceLocation) String() string { return proto.CompactTextString(m) } func (*DeviceLocation) ProtoMessage() {} func (*DeviceLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_embedded_assistant_3cbaf53428f54208, []int{13} + return fileDescriptor_embedded_assistant_05ddb96ecb73e493, []int{15} } func (m *DeviceLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeviceLocation.Unmarshal(m, b) @@ -1270,7 +1378,7 @@ type isDeviceLocation_Type interface { } type DeviceLocation_Coordinates struct { - Coordinates *latlng.LatLng `protobuf:"bytes,1,opt,name=coordinates,oneof"` + Coordinates *latlng.LatLng `protobuf:"bytes,1,opt,name=coordinates,proto3,oneof"` } func (*DeviceLocation_Coordinates) isDeviceLocation_Type() {} @@ -1347,6 +1455,7 @@ func _DeviceLocation_OneofSizer(msg proto.Message) (n int) { func init() { proto.RegisterType((*AssistRequest)(nil), "google.assistant.embedded.v1alpha2.AssistRequest") proto.RegisterType((*AssistResponse)(nil), "google.assistant.embedded.v1alpha2.AssistResponse") + proto.RegisterType((*DebugInfo)(nil), "google.assistant.embedded.v1alpha2.DebugInfo") proto.RegisterType((*AssistConfig)(nil), "google.assistant.embedded.v1alpha2.AssistConfig") proto.RegisterType((*AudioInConfig)(nil), "google.assistant.embedded.v1alpha2.AudioInConfig") proto.RegisterType((*AudioOutConfig)(nil), "google.assistant.embedded.v1alpha2.AudioOutConfig") @@ -1358,6 +1467,7 @@ func init() { proto.RegisterType((*DeviceAction)(nil), "google.assistant.embedded.v1alpha2.DeviceAction") proto.RegisterType((*SpeechRecognitionResult)(nil), "google.assistant.embedded.v1alpha2.SpeechRecognitionResult") proto.RegisterType((*DialogStateOut)(nil), "google.assistant.embedded.v1alpha2.DialogStateOut") + proto.RegisterType((*DebugConfig)(nil), "google.assistant.embedded.v1alpha2.DebugConfig") proto.RegisterType((*DeviceLocation)(nil), "google.assistant.embedded.v1alpha2.DeviceLocation") proto.RegisterEnum("google.assistant.embedded.v1alpha2.AssistResponse_EventType", AssistResponse_EventType_name, AssistResponse_EventType_value) proto.RegisterEnum("google.assistant.embedded.v1alpha2.AudioInConfig_Encoding", AudioInConfig_Encoding_name, AudioInConfig_Encoding_value) @@ -1464,8 +1574,7 @@ func (x *embeddedAssistantAssistClient) Recv() (*AssistResponse, error) { return m, nil } -// Server API for EmbeddedAssistant service - +// EmbeddedAssistantServer is the server API for EmbeddedAssistant service. type EmbeddedAssistantServer interface { // Initiates or continues a conversation with the embedded Assistant Service. // Each call performs one round-trip, sending an audio request to the service @@ -1559,94 +1668,100 @@ var _EmbeddedAssistant_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/assistant/embedded/v1alpha2/embedded_assistant.proto", fileDescriptor_embedded_assistant_3cbaf53428f54208) + proto.RegisterFile("google/assistant/embedded/v1alpha2/embedded_assistant.proto", fileDescriptor_embedded_assistant_05ddb96ecb73e493) } -var fileDescriptor_embedded_assistant_3cbaf53428f54208 = []byte{ - // 1347 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x57, 0x51, 0x6f, 0x1b, 0x45, - 0x10, 0xce, 0x39, 0xae, 0x63, 0x4f, 0x12, 0xdb, 0xd9, 0x56, 0xad, 0x49, 0x0a, 0x8d, 0x0e, 0xa9, - 0x0a, 0x85, 0x3a, 0x8d, 0x8b, 0x40, 0x6a, 0x4a, 0x91, 0x6b, 0x9f, 0x13, 0x17, 0xc7, 0xe7, 0xae, - 0x9d, 0x56, 0xa5, 0xa0, 0xd3, 0xc6, 0xb7, 0x75, 0xae, 0x3a, 0xdf, 0x5e, 0x6f, 0xd7, 0x69, 0xc3, - 0x13, 0x2f, 0x48, 0xbc, 0xc1, 0x2b, 0xaf, 0xfc, 0x05, 0x7e, 0x09, 0x7f, 0x80, 0xdf, 0x80, 0xc4, - 0x1f, 0x40, 0xbb, 0xb7, 0xe7, 0xd8, 0xa1, 0x01, 0xbb, 0xbc, 0xdd, 0xce, 0xec, 0x7c, 0x3b, 0x3b, - 0xf3, 0xcd, 0xcc, 0x1e, 0xec, 0x0e, 0x18, 0x1b, 0xf8, 0x74, 0x9b, 0x70, 0xee, 0x71, 0x41, 0x02, - 0xb1, 0x4d, 0x87, 0x47, 0xd4, 0x75, 0xa9, 0xbb, 0x7d, 0xb2, 0x43, 0xfc, 0xf0, 0x98, 0x54, 0xc6, - 0x12, 0x67, 0xbc, 0xa9, 0x1c, 0x46, 0x4c, 0x30, 0x64, 0xc6, 0xc6, 0xe5, 0x33, 0x79, 0xb2, 0xb5, - 0x9c, 0x18, 0xaf, 0x5f, 0x4f, 0x0e, 0x08, 0xbd, 0x6d, 0x12, 0x04, 0x4c, 0x10, 0xe1, 0xb1, 0x80, - 0xc7, 0x08, 0xeb, 0x25, 0xad, 0x15, 0xa7, 0x21, 0xdd, 0xf6, 0x89, 0xf0, 0x83, 0x41, 0xac, 0x31, - 0xbf, 0x37, 0x60, 0xb5, 0xaa, 0x70, 0x31, 0x7d, 0x35, 0xa2, 0x5c, 0xa0, 0x47, 0x90, 0xe9, 0xb3, - 0xe0, 0x85, 0x37, 0x28, 0x19, 0x9b, 0xc6, 0xd6, 0x72, 0xe5, 0x4e, 0xf9, 0xbf, 0x8f, 0x2f, 0xc7, - 0x10, 0x35, 0x65, 0xb7, 0xbf, 0x80, 0x35, 0x02, 0xda, 0x80, 0x2c, 0x19, 0xb9, 0x1e, 0x73, 0xbc, - 0xa0, 0x94, 0xda, 0x34, 0xb6, 0x56, 0xf6, 0x17, 0xf0, 0x92, 0x92, 0x34, 0x83, 0x87, 0x19, 0x48, - 0x4b, 0x7f, 0xcc, 0x3f, 0xd2, 0x90, 0x4f, 0x5c, 0xe0, 0x21, 0x0b, 0x38, 0x45, 0xcf, 0x01, 0xe8, - 0x09, 0x0d, 0x84, 0x23, 0x37, 0x28, 0x3f, 0xf2, 0x95, 0xfb, 0xb3, 0xfb, 0x91, 0xe0, 0x94, 0x2d, - 0x09, 0xd2, 0x3b, 0x0d, 0x29, 0xce, 0xd1, 0xe4, 0x13, 0x35, 0x21, 0x17, 0x3b, 0xc5, 0x46, 0xa2, - 0xb4, 0xa8, 0xee, 0xf8, 0xc9, 0x4c, 0xd8, 0xd2, 0xc8, 0x1e, 0x09, 0x1c, 0xdf, 0xc9, 0x1e, 0x09, - 0xd4, 0x02, 0xe0, 0xfd, 0x88, 0xd2, 0x40, 0x61, 0xa5, 0x15, 0xd6, 0xed, 0x59, 0xb0, 0xba, 0xca, - 0x4a, 0x82, 0xe5, 0x78, 0xf2, 0x89, 0x0e, 0x61, 0xd5, 0xa5, 0x27, 0x5e, 0x9f, 0x3a, 0xa4, 0x2f, - 0xb3, 0x57, 0xca, 0xcc, 0x9e, 0x80, 0xba, 0x32, 0xac, 0x2a, 0x3b, 0xbc, 0xe2, 0x4e, 0xac, 0xd0, - 0x11, 0xe4, 0x79, 0x48, 0x69, 0xff, 0xd8, 0x89, 0x28, 0x1f, 0xf9, 0x82, 0x97, 0x52, 0x9b, 0x8b, - 0x5b, 0xcb, 0x95, 0xdd, 0x99, 0x1c, 0x55, 0x96, 0x98, 0xf6, 0xd9, 0x20, 0xf0, 0x14, 0xb8, 0xc2, - 0xc0, 0xab, 0x5c, 0x2b, 0x14, 0x22, 0xfa, 0x06, 0x8a, 0xae, 0x47, 0x7c, 0x36, 0x70, 0xb8, 0x20, - 0x82, 0xaa, 0x70, 0x5c, 0x52, 0xde, 0x57, 0x66, 0xf2, 0x5e, 0xd9, 0x76, 0xa5, 0xa9, 0x8c, 0x49, - 0xde, 0x9d, 0x5a, 0x9b, 0x5f, 0x40, 0x6e, 0x9c, 0x49, 0xb4, 0x0e, 0x57, 0xad, 0x27, 0x56, 0xbb, - 0xe7, 0xf4, 0x9e, 0x75, 0x2c, 0xe7, 0xb0, 0xdd, 0xed, 0x58, 0xb5, 0x66, 0xa3, 0x69, 0xd5, 0x8b, - 0x0b, 0xe8, 0x0a, 0x14, 0xad, 0x76, 0xdd, 0xb1, 0x1b, 0xce, 0x61, 0xaf, 0x67, 0xe1, 0x6a, 0xbb, - 0x66, 0x15, 0x0d, 0xf3, 0x87, 0x34, 0xac, 0x4c, 0x12, 0x14, 0x3d, 0x87, 0x42, 0x42, 0x4b, 0x67, - 0x8a, 0xeb, 0x3b, 0x33, 0xf3, 0xa0, 0x19, 0x8c, 0xc9, 0xbe, 0x4a, 0x26, 0x05, 0xe8, 0x06, 0x80, - 0xa0, 0x6f, 0x84, 0xf3, 0x6a, 0x44, 0xa3, 0x53, 0x95, 0xc2, 0xdc, 0xfe, 0x02, 0xce, 0x49, 0xd9, - 0x63, 0x29, 0x92, 0xb1, 0x1a, 0xf3, 0x2f, 0x39, 0x3e, 0x35, 0x7b, 0xac, 0x12, 0x1a, 0xc6, 0xc7, - 0xe1, 0x3c, 0x99, 0x5a, 0x23, 0x07, 0xd6, 0xce, 0x28, 0x99, 0xc0, 0x67, 0x15, 0xfc, 0xdd, 0xb9, - 0x98, 0xa9, 0xf1, 0x0b, 0x7c, 0x5a, 0x80, 0x9e, 0x41, 0x61, 0x2a, 0xd5, 0x5e, 0xa0, 0x8b, 0x68, - 0x67, 0xce, 0x4c, 0x37, 0x03, 0xbc, 0xea, 0x4e, 0x2e, 0x27, 0x0a, 0x40, 0xfb, 0x9d, 0x9e, 0xb7, - 0x00, 0xb4, 0xd3, 0xba, 0x00, 0xe2, 0xd5, 0xb8, 0xd1, 0xfc, 0x2e, 0x7b, 0xdd, 0x54, 0xae, 0x9e, - 0x40, 0x96, 0x06, 0x7d, 0xe6, 0x7a, 0xc1, 0x40, 0x77, 0x99, 0x7b, 0x73, 0x33, 0xa0, 0x6c, 0x69, - 0x04, 0x3c, 0xc6, 0x42, 0xb7, 0x60, 0x8d, 0x93, 0x61, 0xe8, 0x53, 0x27, 0x92, 0x21, 0x3a, 0xa6, - 0x91, 0xf8, 0x4e, 0xe5, 0xf8, 0x12, 0x2e, 0xc4, 0x0a, 0x4c, 0x04, 0xdd, 0x97, 0x62, 0xf3, 0x3e, - 0x64, 0x13, 0x04, 0x54, 0x82, 0x2b, 0x56, 0xbb, 0x66, 0xd7, 0x9b, 0xed, 0xbd, 0x73, 0xcc, 0x5e, - 0x81, 0x6c, 0xab, 0xd9, 0xb6, 0xaa, 0x78, 0xe7, 0xb3, 0xa2, 0x81, 0xb2, 0x90, 0x6e, 0xb4, 0xaa, - 0xb5, 0x62, 0xca, 0xfc, 0x39, 0x05, 0xf9, 0x69, 0x46, 0xa0, 0xa7, 0xff, 0xb8, 0xd4, 0xee, 0xfc, - 0xbc, 0xfa, 0x9f, 0xb7, 0x42, 0x1f, 0xc3, 0xda, 0x09, 0xf3, 0x47, 0x43, 0xea, 0x84, 0x34, 0xea, - 0xd3, 0x40, 0x90, 0x01, 0x55, 0x3c, 0xb9, 0x84, 0x8b, 0xb1, 0xa2, 0x33, 0x96, 0x9b, 0xad, 0x77, - 0x08, 0xc1, 0x12, 0x2c, 0x1e, 0x74, 0xee, 0x16, 0x53, 0xa8, 0x00, 0xcb, 0x76, 0xe7, 0xb0, 0xeb, - 0x34, 0xdb, 0x8e, 0xbd, 0xb7, 0x57, 0x5c, 0x34, 0x7f, 0x33, 0xa0, 0x70, 0x8e, 0xc5, 0xc8, 0x81, - 0x65, 0x5d, 0x15, 0x43, 0xe6, 0x26, 0x13, 0xe5, 0xc1, 0x3b, 0xd4, 0x83, 0x5e, 0x1f, 0x30, 0x97, - 0x62, 0xdd, 0xfb, 0xe5, 0xb7, 0xf9, 0x25, 0xc0, 0x99, 0x06, 0x6d, 0xc0, 0xb5, 0x6e, 0x0d, 0x5b, - 0x56, 0xdb, 0x39, 0xb0, 0xeb, 0xe7, 0x9b, 0xd4, 0x12, 0x2c, 0xda, 0x8d, 0x46, 0xd1, 0x40, 0xcb, - 0xb0, 0xd4, 0x69, 0x55, 0x9f, 0x35, 0xdb, 0xd2, 0xeb, 0x3f, 0x0d, 0x58, 0x9d, 0x2a, 0x0e, 0x74, - 0x1b, 0x50, 0x9f, 0x05, 0x27, 0x34, 0xe2, 0x6a, 0x96, 0xc7, 0xe5, 0xa6, 0x5c, 0x5f, 0xc1, 0x6b, - 0x93, 0x1a, 0x65, 0x80, 0x3e, 0x84, 0x55, 0x9f, 0x04, 0x83, 0x11, 0x19, 0xc8, 0xf2, 0x71, 0xa9, - 0xca, 0x4c, 0x0e, 0xaf, 0x24, 0xc2, 0x9a, 0x74, 0xec, 0x39, 0x14, 0x74, 0x85, 0xf9, 0xac, 0xaf, - 0x8c, 0xe7, 0x6a, 0xd3, 0xca, 0xb4, 0xa5, 0x2d, 0x71, 0xde, 0x9d, 0x5a, 0xa3, 0x32, 0x5c, 0xf6, - 0xb8, 0x13, 0xd0, 0xd7, 0xce, 0xa4, 0x77, 0xa5, 0xa5, 0x4d, 0x63, 0x2b, 0x8b, 0xd7, 0x3c, 0xde, - 0xa6, 0xaf, 0x6b, 0x13, 0x0a, 0xb3, 0x0b, 0x2b, 0x93, 0x55, 0x8b, 0x36, 0x20, 0xa7, 0x9d, 0xf3, - 0x5c, 0x75, 0xcf, 0x1c, 0xce, 0xc6, 0x82, 0xa6, 0x8b, 0x6e, 0x8e, 0x3d, 0x97, 0x19, 0xf4, 0xe5, - 0x96, 0x45, 0xb5, 0x45, 0xb7, 0x0c, 0x19, 0x77, 0xbf, 0xe9, 0x9a, 0x1f, 0x41, 0x36, 0x61, 0x32, - 0x7a, 0x1f, 0x20, 0xee, 0xb4, 0x2e, 0x11, 0x44, 0x47, 0x2e, 0x9e, 0xfd, 0x75, 0x22, 0x88, 0xf9, - 0x8b, 0x01, 0xb9, 0x71, 0x7a, 0x51, 0x0b, 0x32, 0x2f, 0x58, 0x34, 0x24, 0x42, 0xb3, 0xe3, 0xd3, - 0xb9, 0xd8, 0x51, 0x6e, 0x28, 0x5b, 0xac, 0x31, 0x10, 0x82, 0xb4, 0x3a, 0x54, 0xbd, 0x7a, 0xb0, - 0xfa, 0x36, 0x6f, 0x41, 0x26, 0xde, 0x85, 0xae, 0x02, 0x6a, 0xd8, 0xf8, 0xa0, 0xda, 0x3b, 0x47, - 0x8d, 0x2c, 0xa4, 0xf7, 0x7b, 0x07, 0xad, 0xa2, 0x61, 0x3e, 0x48, 0x62, 0xa3, 0x87, 0x78, 0x19, - 0x2e, 0xeb, 0xeb, 0x47, 0xf1, 0x3b, 0xcd, 0x79, 0xc9, 0x59, 0xa0, 0xa3, 0xb4, 0x16, 0xab, 0xf4, - 0x0b, 0xee, 0x11, 0x67, 0x81, 0xf9, 0x14, 0xae, 0x5d, 0x30, 0xba, 0xd1, 0x07, 0x00, 0x22, 0x22, - 0x01, 0xef, 0x47, 0x5e, 0x28, 0x34, 0xc2, 0x84, 0x04, 0x5d, 0x87, 0x1c, 0x17, 0xe4, 0xc8, 0xf3, - 0x3d, 0x71, 0xaa, 0xfc, 0x4f, 0xe1, 0x33, 0x81, 0xf9, 0x57, 0x0a, 0xf2, 0xd3, 0xe3, 0x1a, 0xdd, - 0x83, 0xf7, 0xf8, 0x28, 0x0c, 0x7d, 0x3a, 0x94, 0xf5, 0xec, 0x3b, 0xae, 0xc7, 0x43, 0x9f, 0x9c, - 0x3a, 0x72, 0xe4, 0x69, 0xfc, 0x6b, 0x93, 0x1b, 0xea, 0xb1, 0xbe, 0x47, 0xdf, 0x88, 0x0b, 0x48, - 0x9e, 0xba, 0x88, 0xe4, 0x2f, 0xa1, 0x30, 0xf4, 0xfa, 0x11, 0x0b, 0x8f, 0x59, 0x10, 0x33, 0x41, - 0xb1, 0x20, 0x5f, 0xa9, 0xce, 0xff, 0xcc, 0x28, 0x1f, 0x8c, 0x91, 0x54, 0x39, 0xe7, 0x87, 0x53, - 0xeb, 0xb7, 0xb7, 0xb0, 0xf4, 0x05, 0x2d, 0xec, 0x5b, 0xc8, 0x4f, 0xc3, 0xa1, 0x1b, 0xb0, 0x71, - 0xd0, 0xac, 0x61, 0xbb, 0xb3, 0x6f, 0xb7, 0xad, 0xb7, 0xf5, 0x81, 0x2b, 0x50, 0xac, 0xb5, 0xec, - 0xae, 0xe5, 0x9c, 0x6d, 0x2b, 0x1a, 0x52, 0x5a, 0x6f, 0x56, 0x5b, 0xf6, 0x9e, 0xd3, 0xb0, 0x5b, - 0x2d, 0xfb, 0xa9, 0x63, 0xb7, 0x8b, 0x29, 0xf3, 0x31, 0xe4, 0xa7, 0x8b, 0x0f, 0x7d, 0x0e, 0xcb, - 0x7d, 0xc6, 0x22, 0xd7, 0x0b, 0x88, 0xa0, 0x5c, 0xbf, 0x5f, 0x2e, 0x27, 0x51, 0x90, 0xf3, 0xae, - 0xdc, 0x22, 0xa2, 0x15, 0xc8, 0x17, 0xca, 0xe4, 0xce, 0x64, 0x1a, 0x56, 0x7e, 0x34, 0x60, 0xcd, - 0xd2, 0x21, 0xaa, 0x26, 0x41, 0x43, 0x1c, 0x32, 0xf1, 0x02, 0xed, 0xcc, 0xf3, 0xde, 0x56, 0xc4, - 0x5b, 0xaf, 0xcc, 0xff, 0x44, 0xdf, 0x32, 0xee, 0x18, 0x0f, 0x7f, 0x32, 0xe0, 0x66, 0x9f, 0x0d, - 0x67, 0xb0, 0x7e, 0x98, 0x1f, 0xbb, 0xda, 0x91, 0xff, 0x2f, 0x1d, 0xe3, 0xeb, 0x47, 0xda, 0x6a, - 0xc0, 0x64, 0xa7, 0x2b, 0xb3, 0x68, 0xb0, 0x3d, 0xa0, 0x81, 0xfa, 0xbb, 0xd9, 0x8e, 0x55, 0x24, - 0xf4, 0xf8, 0xbf, 0xfd, 0x79, 0xed, 0x26, 0x92, 0x5f, 0x53, 0x99, 0x6a, 0xb7, 0xd7, 0xad, 0x7f, - 0x75, 0x94, 0x51, 0xf6, 0x77, 0xff, 0x0e, 0x00, 0x00, 0xff, 0xff, 0x03, 0x3d, 0xeb, 0x1f, 0xb8, - 0x0d, 0x00, 0x00, +var fileDescriptor_embedded_assistant_05ddb96ecb73e493 = []byte{ + // 1449 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x57, 0x41, 0x73, 0xdb, 0xc6, + 0x15, 0x16, 0x28, 0x8a, 0x22, 0x1f, 0x29, 0x92, 0x5a, 0x7b, 0x6c, 0x56, 0x72, 0x6b, 0x0d, 0x3a, + 0xe3, 0x51, 0xdd, 0x9a, 0xb4, 0xe8, 0x4e, 0x3b, 0xb5, 0x5c, 0x77, 0x68, 0x12, 0x94, 0xe0, 0x52, + 0x04, 0x0d, 0x52, 0xf6, 0xb8, 0x6e, 0x07, 0xb3, 0x22, 0x56, 0x10, 0x3c, 0xe0, 0x2e, 0x0d, 0x2c, + 0x65, 0xab, 0xa7, 0x1c, 0x73, 0x4b, 0xae, 0xb9, 0xe6, 0x9c, 0x5b, 0x7e, 0x49, 0x8e, 0xf9, 0x17, + 0x99, 0xc9, 0x1f, 0xc8, 0xec, 0x62, 0x41, 0x91, 0x8a, 0x95, 0x90, 0xce, 0x0d, 0xfb, 0xde, 0xbe, + 0x0f, 0x0f, 0xbb, 0xdf, 0xf7, 0xde, 0x03, 0xec, 0x7b, 0x8c, 0x79, 0x01, 0xa9, 0xe1, 0x28, 0xf2, + 0x23, 0x8e, 0x29, 0xaf, 0x91, 0xd1, 0x09, 0x71, 0x5d, 0xe2, 0xd6, 0xce, 0xf7, 0x70, 0x30, 0x3e, + 0xc3, 0xf5, 0xa9, 0xc5, 0x99, 0x6e, 0xaa, 0x8e, 0x43, 0xc6, 0x19, 0xd2, 0xe3, 0xe0, 0xea, 0xa5, + 0x3d, 0xd9, 0x5a, 0x4d, 0x82, 0xb7, 0xee, 0x24, 0x2f, 0x18, 0xfb, 0x35, 0x4c, 0x29, 0xe3, 0x98, + 0xfb, 0x8c, 0x46, 0x31, 0xc2, 0x56, 0x45, 0x79, 0xf9, 0xc5, 0x98, 0xd4, 0x02, 0xcc, 0x03, 0xea, + 0xc5, 0x1e, 0xfd, 0x33, 0x0d, 0x36, 0x1a, 0x12, 0xd7, 0x26, 0xef, 0x26, 0x24, 0xe2, 0xe8, 0x39, + 0x64, 0x86, 0x8c, 0x9e, 0xfa, 0x5e, 0x45, 0xdb, 0xd1, 0x76, 0xf3, 0xf5, 0x87, 0xd5, 0x5f, 0x7f, + 0x7d, 0x35, 0x86, 0x68, 0xca, 0xb8, 0xc3, 0x15, 0x5b, 0x21, 0xa0, 0x6d, 0xc8, 0xe2, 0x89, 0xeb, + 0x33, 0xc7, 0xa7, 0x95, 0xd4, 0x8e, 0xb6, 0x5b, 0x38, 0x5c, 0xb1, 0xd7, 0xa5, 0xc5, 0xa4, 0xcf, + 0x32, 0x90, 0x16, 0xf9, 0xe8, 0xdf, 0xac, 0x41, 0x31, 0x49, 0x21, 0x1a, 0x33, 0x1a, 0x11, 0xf4, + 0x06, 0x80, 0x9c, 0x13, 0xca, 0x1d, 0xb1, 0x41, 0xe6, 0x51, 0xac, 0x3f, 0x59, 0x3c, 0x8f, 0x04, + 0xa7, 0x6a, 0x08, 0x90, 0xc1, 0xc5, 0x98, 0xd8, 0x39, 0x92, 0x3c, 0x22, 0x13, 0x72, 0x71, 0x52, + 0x6c, 0xc2, 0x2b, 0xab, 0xf2, 0x1b, 0xff, 0xb2, 0x10, 0xb6, 0x08, 0xb2, 0x26, 0xdc, 0x8e, 0xbf, + 0xc9, 0x9a, 0x70, 0xd4, 0x01, 0x88, 0x86, 0x21, 0x21, 0x54, 0x62, 0xa5, 0x25, 0xd6, 0x83, 0x45, + 0xb0, 0xfa, 0x32, 0x4a, 0x80, 0xe5, 0xa2, 0xe4, 0x11, 0x1d, 0xc3, 0x86, 0x4b, 0xce, 0xfd, 0x21, + 0x71, 0xf0, 0x50, 0xdc, 0x5e, 0x25, 0xb3, 0xf8, 0x05, 0xb4, 0x64, 0x60, 0x43, 0xc6, 0xd9, 0x05, + 0x77, 0x66, 0x85, 0x4e, 0xa0, 0x18, 0x8d, 0x09, 0x19, 0x9e, 0x39, 0x21, 0x89, 0x26, 0x01, 0x8f, + 0x2a, 0xa9, 0x9d, 0xd5, 0xdd, 0x7c, 0x7d, 0x7f, 0xa1, 0x44, 0x65, 0xa4, 0x4d, 0x86, 0xcc, 0xa3, + 0xbe, 0x04, 0x97, 0x18, 0xf6, 0x46, 0xa4, 0x1c, 0x12, 0x11, 0xfd, 0x17, 0xca, 0xae, 0x8f, 0x03, + 0xe6, 0x39, 0x11, 0xc7, 0x9c, 0xc8, 0xe3, 0x58, 0x93, 0xd9, 0xd7, 0x17, 0xca, 0x5e, 0xc6, 0xf6, + 0x45, 0xa8, 0x38, 0x93, 0xa2, 0x3b, 0xb7, 0x16, 0xc7, 0xec, 0x92, 0x93, 0x89, 0xe7, 0xf8, 0xf4, + 0x94, 0x55, 0xb2, 0x8b, 0x1f, 0x73, 0x4b, 0x44, 0x99, 0xf4, 0x94, 0xd9, 0x39, 0x37, 0x79, 0xd4, + 0xff, 0x09, 0xb9, 0x29, 0x2f, 0xd0, 0x16, 0xdc, 0x32, 0x5e, 0x1a, 0xdd, 0x81, 0x33, 0x78, 0xdd, + 0x33, 0x9c, 0xe3, 0x6e, 0xbf, 0x67, 0x34, 0xcd, 0xb6, 0x69, 0xb4, 0xca, 0x2b, 0xe8, 0x26, 0x94, + 0x8d, 0x6e, 0xcb, 0xb1, 0xda, 0xce, 0xf1, 0x60, 0x60, 0xd8, 0x8d, 0x6e, 0xd3, 0x28, 0x6b, 0xba, + 0x09, 0xb9, 0x29, 0x2c, 0x7a, 0x02, 0xdb, 0x98, 0x79, 0x0e, 0xf6, 0x24, 0x59, 0xd9, 0xa5, 0x74, + 0x9d, 0xb7, 0x11, 0xa3, 0x92, 0xb9, 0x39, 0xfb, 0x36, 0x66, 0x5e, 0x43, 0xec, 0x18, 0xb0, 0x46, + 0xe2, 0x7f, 0x1e, 0x31, 0xaa, 0x7f, 0x9f, 0x86, 0xc2, 0xac, 0x72, 0xd0, 0x1b, 0x28, 0x25, 0x7a, + 0x71, 0xe6, 0x44, 0xb8, 0xb7, 0x30, 0x41, 0x4d, 0x3a, 0x55, 0xe1, 0x06, 0x9e, 0x35, 0xa0, 0xbb, + 0x00, 0x9c, 0x7c, 0xe0, 0xce, 0xbb, 0x09, 0x09, 0x2f, 0x24, 0xb7, 0x72, 0x87, 0x2b, 0x76, 0x4e, + 0xd8, 0x5e, 0x08, 0x93, 0xb8, 0xc4, 0xa9, 0x30, 0x92, 0xd7, 0xa7, 0x16, 0xbf, 0xc4, 0x44, 0x1f, + 0xf1, 0xeb, 0xec, 0x22, 0x9e, 0x5b, 0x23, 0x07, 0x36, 0x2f, 0xb5, 0x92, 0xc0, 0xc7, 0x77, 0xf9, + 0x68, 0x29, 0xc9, 0x28, 0xfc, 0x52, 0x34, 0x6f, 0x40, 0xaf, 0xa1, 0x34, 0xc7, 0x41, 0x9f, 0x2a, + 0x75, 0xef, 0x2d, 0x49, 0x41, 0x93, 0xda, 0x1b, 0xee, 0xec, 0x72, 0x46, 0x99, 0x2a, 0xef, 0xf4, + 0xb2, 0xca, 0x54, 0x49, 0x2b, 0x65, 0xaa, 0x8c, 0x6d, 0x28, 0xc4, 0xbc, 0x56, 0xa8, 0xb1, 0x62, + 0x6a, 0x0b, 0x33, 0x5b, 0x81, 0xe6, 0xdd, 0xcb, 0xc5, 0xb4, 0xaa, 0x7e, 0x27, 0x0a, 0xfb, 0xdc, + 0xfd, 0xbf, 0x84, 0x2c, 0xa1, 0x43, 0xe6, 0xfa, 0xd4, 0x53, 0x25, 0xf5, 0xf1, 0xd2, 0xac, 0xaa, + 0x1a, 0x0a, 0xc1, 0x9e, 0x62, 0xa1, 0xfb, 0xb0, 0x19, 0xe1, 0xd1, 0x38, 0x20, 0x4e, 0x28, 0x8e, + 0xfd, 0x8c, 0x84, 0xfc, 0xff, 0x92, 0x37, 0x6b, 0x76, 0x29, 0x76, 0xd8, 0x98, 0x93, 0x43, 0x61, + 0xd6, 0x9f, 0x40, 0x36, 0x41, 0x40, 0x15, 0xb8, 0x69, 0x74, 0x9b, 0x56, 0xcb, 0xec, 0x1e, 0x5c, + 0x11, 0x5e, 0x01, 0xb2, 0x1d, 0xb3, 0x6b, 0x34, 0xec, 0xbd, 0xbf, 0x95, 0x35, 0x94, 0x85, 0x74, + 0xbb, 0xd3, 0x68, 0x96, 0x53, 0xfa, 0x97, 0x29, 0x28, 0xce, 0xb3, 0x0c, 0xbd, 0xfa, 0xd9, 0x47, + 0xed, 0x2f, 0xcf, 0xd5, 0xdf, 0xf8, 0x55, 0xe8, 0xcf, 0xb0, 0x79, 0xce, 0x82, 0xc9, 0x88, 0x38, + 0x63, 0x12, 0x0e, 0x09, 0xe5, 0xd8, 0x23, 0x92, 0x7b, 0x6b, 0x76, 0x39, 0x76, 0xf4, 0xa6, 0x76, + 0xbd, 0xf3, 0x09, 0x47, 0xb0, 0x0e, 0xab, 0x47, 0xbd, 0x47, 0xe5, 0x14, 0x2a, 0x41, 0xde, 0xea, + 0x1d, 0xf7, 0x1d, 0xb3, 0xeb, 0x58, 0x07, 0x07, 0xe5, 0x55, 0xfd, 0x5b, 0x0d, 0x4a, 0x57, 0x94, + 0x81, 0x1c, 0xc8, 0x2b, 0xa5, 0x8d, 0x98, 0x9b, 0xb4, 0xcf, 0xa7, 0x9f, 0xa0, 0x31, 0xb5, 0x3e, + 0x62, 0x2e, 0xb1, 0x55, 0xa3, 0x13, 0xcf, 0xfa, 0xbf, 0x00, 0x2e, 0x3d, 0x68, 0x1b, 0x6e, 0xf7, + 0x9b, 0xb6, 0x61, 0x74, 0x9d, 0x23, 0xab, 0x75, 0xb5, 0x86, 0xae, 0xc3, 0xaa, 0xd5, 0x6e, 0x97, + 0x35, 0x94, 0x87, 0xf5, 0x5e, 0xa7, 0xf1, 0xda, 0xec, 0x8a, 0xac, 0x7f, 0xd0, 0x60, 0x63, 0x4e, + 0x70, 0xe8, 0x01, 0xa0, 0x21, 0xa3, 0xe7, 0x24, 0x8c, 0xe4, 0xe0, 0x12, 0x4b, 0x58, 0xa6, 0x5e, + 0xb0, 0x37, 0x67, 0x3d, 0x32, 0x00, 0xfd, 0x11, 0x36, 0x02, 0x4c, 0xbd, 0x09, 0xf6, 0x84, 0x24, + 0x5d, 0x22, 0x6f, 0x26, 0x67, 0x17, 0x12, 0x63, 0x53, 0x24, 0xf6, 0x06, 0x4a, 0x4a, 0xb5, 0x01, + 0x1b, 0xca, 0xe0, 0xa5, 0x7a, 0x92, 0x0c, 0xed, 0xa8, 0x48, 0xbb, 0xe8, 0xce, 0xad, 0x51, 0x15, + 0x6e, 0xf8, 0x91, 0x43, 0xc9, 0x7b, 0x67, 0x36, 0xbb, 0xca, 0xfa, 0x8e, 0xb6, 0x9b, 0xb5, 0x37, + 0xfd, 0xa8, 0x4b, 0xde, 0x37, 0x67, 0x1c, 0x7a, 0x1f, 0x0a, 0xb3, 0x95, 0x00, 0x6d, 0x43, 0x4e, + 0x25, 0xe7, 0xbb, 0xaa, 0x4f, 0x64, 0x63, 0x83, 0xe9, 0xa2, 0x7b, 0xd3, 0xcc, 0xc5, 0x0d, 0x06, + 0x62, 0xcb, 0xaa, 0xdc, 0xa2, 0xca, 0x90, 0x38, 0xf7, 0xc0, 0x74, 0xf5, 0x3f, 0x41, 0x36, 0x61, + 0x32, 0xfa, 0x3d, 0x40, 0x5c, 0xbd, 0x5d, 0xcc, 0xb1, 0x3a, 0xb9, 0x78, 0xd0, 0x69, 0x61, 0x8e, + 0xf5, 0xaf, 0x34, 0xc8, 0x4d, 0xaf, 0x17, 0x75, 0x20, 0x73, 0xca, 0xc2, 0x11, 0xe6, 0x8a, 0x1d, + 0x7f, 0x5d, 0x8a, 0x1d, 0xd5, 0xb6, 0x8c, 0xb5, 0x15, 0x06, 0x42, 0x90, 0x96, 0x2f, 0x95, 0x23, + 0x9e, 0x2d, 0x9f, 0xf5, 0xfb, 0x90, 0x89, 0x77, 0xa1, 0x5b, 0x80, 0xda, 0x96, 0x7d, 0xd4, 0x18, + 0x5c, 0xa1, 0x46, 0x16, 0xd2, 0x87, 0x83, 0xa3, 0x4e, 0x59, 0xd3, 0x9f, 0x26, 0x67, 0xa3, 0x26, + 0x96, 0x2a, 0xdc, 0x50, 0x9f, 0x1f, 0xc6, 0x43, 0xe9, 0x6c, 0x37, 0xdd, 0x8c, 0x5d, 0x6a, 0x5c, + 0x95, 0x7d, 0xf4, 0x15, 0xdc, 0xbe, 0x66, 0x4e, 0x41, 0x7f, 0x00, 0xe0, 0x21, 0xa6, 0xd1, 0x30, + 0xf4, 0xc7, 0x5c, 0x21, 0xcc, 0x58, 0xd0, 0x1d, 0xc8, 0x45, 0x1c, 0x9f, 0xf8, 0x81, 0xcf, 0x2f, + 0x64, 0xfe, 0x29, 0xfb, 0xd2, 0xa0, 0xff, 0x98, 0x82, 0xe2, 0xfc, 0x6c, 0x82, 0x1e, 0xc3, 0xef, + 0xa2, 0xc9, 0x78, 0x1c, 0x90, 0x91, 0xd0, 0x73, 0xe0, 0xb8, 0x7e, 0x34, 0x0e, 0xf0, 0x85, 0x23, + 0xda, 0x68, 0xd2, 0xef, 0x67, 0x37, 0xb4, 0x62, 0xff, 0x80, 0x7c, 0xe0, 0xd7, 0x90, 0x3c, 0x75, + 0x1d, 0xc9, 0xdf, 0x42, 0x69, 0xe4, 0x0f, 0x43, 0x36, 0x3e, 0x63, 0x34, 0x66, 0x82, 0x64, 0x41, + 0xb1, 0xde, 0x58, 0x7e, 0xa6, 0xaa, 0x1e, 0x4d, 0x91, 0xa4, 0x9c, 0x8b, 0xa3, 0xb9, 0xf5, 0xc7, + 0x4b, 0x58, 0xfa, 0x9a, 0x12, 0xf6, 0x3f, 0x28, 0xce, 0xc3, 0xa1, 0xbb, 0xb0, 0x7d, 0x64, 0x36, + 0x6d, 0xab, 0x77, 0x68, 0x75, 0x8d, 0x8f, 0xd5, 0x81, 0x9b, 0x50, 0x6e, 0x76, 0xac, 0xbe, 0xe1, + 0x5c, 0x6e, 0x2b, 0x6b, 0xc2, 0xda, 0x32, 0x1b, 0x1d, 0xeb, 0xc0, 0x69, 0x5b, 0x9d, 0x8e, 0xf5, + 0xca, 0xb1, 0xba, 0xe5, 0x94, 0xfe, 0x0f, 0xc8, 0xcf, 0xb4, 0x37, 0x51, 0x89, 0x43, 0xc2, 0x27, + 0x21, 0x75, 0x66, 0x86, 0xc0, 0x8c, 0xd4, 0x59, 0x29, 0x76, 0x4c, 0xe7, 0x31, 0xfd, 0x05, 0x14, + 0xe7, 0x75, 0x8b, 0xfe, 0x0e, 0xf9, 0x21, 0x63, 0xa1, 0xeb, 0x53, 0xcc, 0x49, 0xa4, 0xc6, 0xa9, + 0x1b, 0xc9, 0x01, 0x8a, 0x56, 0x59, 0xed, 0x60, 0xde, 0xa1, 0x62, 0x60, 0x9a, 0xdd, 0x99, 0x34, + 0xd2, 0xfa, 0xe7, 0x1a, 0x6c, 0x1a, 0xea, 0x74, 0xa7, 0xe3, 0x1b, 0x8a, 0x20, 0x13, 0x2f, 0xd0, + 0xde, 0x32, 0xff, 0x25, 0x92, 0xb3, 0x5b, 0xf5, 0xe5, 0x7f, 0x65, 0x76, 0xb5, 0x87, 0xda, 0xb3, + 0x2f, 0x34, 0xb8, 0x37, 0x64, 0xa3, 0x05, 0xa2, 0x9f, 0x15, 0xa7, 0xa9, 0xf6, 0xc4, 0x7f, 0x5e, + 0x4f, 0xfb, 0xcf, 0x73, 0x15, 0xe5, 0x31, 0x51, 0x24, 0xab, 0x2c, 0xf4, 0x6a, 0x1e, 0xa1, 0xf2, + 0x2f, 0xb0, 0x16, 0xbb, 0xf0, 0xd8, 0x8f, 0x7e, 0xe9, 0x0f, 0x75, 0x3f, 0xb1, 0x7c, 0x9d, 0xca, + 0x34, 0xfa, 0x83, 0x7e, 0xeb, 0xdf, 0x27, 0x19, 0x19, 0xff, 0xe8, 0xa7, 0x00, 0x00, 0x00, 0xff, + 0xff, 0xb9, 0x08, 0x29, 0x0f, 0xe0, 0x0e, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_data.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_data.pb.go index 994f1dfe8..f6feb845c 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_data.pb.go @@ -90,11 +90,11 @@ func (Zone_Status) EnumDescriptor() ([]byte, []int) { type Zone struct { // A permanent unique identifier for the zone. // Values are of the form projects//zones/[a-z][-a-z0-9]* - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of this zone as it appears in UIs. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The current state of this zone. - Status Zone_Status `protobuf:"varint,3,opt,name=status,enum=google.bigtable.admin.cluster.v1.Zone_Status" json:"status,omitempty"` + Status Zone_Status `protobuf:"varint,3,opt,name=status,proto3,enum=google.bigtable.admin.cluster.v1.Zone_Status" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -151,20 +151,20 @@ type Cluster struct { // zone in which the cluster resides is included here. // Values are of the form // projects//zones//clusters/[a-z][-a-z0-9]* - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The operation currently running on the cluster, if any. // This cannot be set directly, only through CreateCluster, UpdateCluster, // or UndeleteCluster. Calls to these methods will be rejected if // "current_operation" is already set. - CurrentOperation *longrunning.Operation `protobuf:"bytes,3,opt,name=current_operation,json=currentOperation" json:"current_operation,omitempty"` + CurrentOperation *longrunning.Operation `protobuf:"bytes,3,opt,name=current_operation,json=currentOperation,proto3" json:"current_operation,omitempty"` // The descriptive name for this cluster as it appears in UIs. // Must be unique per zone. - DisplayName string `protobuf:"bytes,4,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,4,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The number of serve nodes allocated to this cluster. - ServeNodes int32 `protobuf:"varint,5,opt,name=serve_nodes,json=serveNodes" json:"serve_nodes,omitempty"` + ServeNodes int32 `protobuf:"varint,5,opt,name=serve_nodes,json=serveNodes,proto3" json:"serve_nodes,omitempty"` // What storage type to use for tables in this cluster. Only configurable at // cluster creation time. If unspecified, STORAGE_SSD will be used. - DefaultStorageType StorageType `protobuf:"varint,8,opt,name=default_storage_type,json=defaultStorageType,enum=google.bigtable.admin.cluster.v1.StorageType" json:"default_storage_type,omitempty"` + DefaultStorageType StorageType `protobuf:"varint,8,opt,name=default_storage_type,json=defaultStorageType,proto3,enum=google.bigtable.admin.cluster.v1.StorageType" json:"default_storage_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service.pb.go index fc4fbf853..63b9e78dc 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service.pb.go @@ -188,8 +188,7 @@ func (c *bigtableClusterServiceClient) UndeleteCluster(ctx context.Context, in * return out, nil } -// Server API for BigtableClusterService service - +// BigtableClusterServiceServer is the server API for BigtableClusterService service. type BigtableClusterServiceServer interface { // Lists the supported zones for the given project. ListZones(context.Context, *ListZonesRequest) (*ListZonesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service_messages.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service_messages.pb.go index 0d55c3e67..79bd5491c 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service_messages.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/cluster/v1/bigtable_cluster_service_messages.pb.go @@ -24,7 +24,7 @@ type ListZonesRequest struct { // The unique name of the project for which a list of supported zones is // requested. // Values are of the form projects/ - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -64,7 +64,7 @@ func (m *ListZonesRequest) GetName() string { // Response message for BigtableClusterService.ListZones. type ListZonesResponse struct { // The list of requested zones. - Zones []*Zone `protobuf:"bytes,1,rep,name=zones" json:"zones,omitempty"` + Zones []*Zone `protobuf:"bytes,1,rep,name=zones,proto3" json:"zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -105,7 +105,7 @@ func (m *ListZonesResponse) GetZones() []*Zone { type GetClusterRequest struct { // The unique name of the requested cluster. // Values are of the form projects//zones//clusters/ - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -146,7 +146,7 @@ func (m *GetClusterRequest) GetName() string { type ListClustersRequest struct { // The unique name of the project for which a list of clusters is requested. // Values are of the form projects/ - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -186,9 +186,9 @@ func (m *ListClustersRequest) GetName() string { // Response message for BigtableClusterService.ListClusters. type ListClustersResponse struct { // The list of requested Clusters. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` // The zones for which clusters could not be retrieved. - FailedZones []*Zone `protobuf:"bytes,2,rep,name=failed_zones,json=failedZones" json:"failed_zones,omitempty"` + FailedZones []*Zone `protobuf:"bytes,2,rep,name=failed_zones,json=failedZones,proto3" json:"failed_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -236,15 +236,15 @@ func (m *ListClustersResponse) GetFailedZones() []*Zone { type CreateClusterRequest struct { // The unique name of the zone in which to create the cluster. // Values are of the form projects//zones/ - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The id to be used when referring to the new cluster within its zone, // e.g. just the "test-cluster" section of the full name // "projects//zones//clusters/test-cluster". - ClusterId string `protobuf:"bytes,2,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,2,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The cluster to create. // The "name", "delete_time", and "current_operation" fields must be left // blank. - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -299,11 +299,11 @@ func (m *CreateClusterRequest) GetCluster() *Cluster { // BigtableClusterService.CreateCluster. type CreateClusterMetadata struct { // The request which prompted the creation of this operation. - OriginalRequest *CreateClusterRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *CreateClusterRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which original_request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which this operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -358,15 +358,15 @@ func (m *CreateClusterMetadata) GetFinishTime() *timestamp.Timestamp { // BigtableClusterService.UpdateCluster. type UpdateClusterMetadata struct { // The request which prompted the creation of this operation. - OriginalRequest *Cluster `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *Cluster `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which original_request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which this operation was cancelled. If set, this operation is // in the process of undoing itself (which is guaranteed to succeed) and // cannot be cancelled again. - CancelTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=cancel_time,json=cancelTime" json:"cancel_time,omitempty"` + CancelTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=cancel_time,json=cancelTime,proto3" json:"cancel_time,omitempty"` // The time at which this operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -428,7 +428,7 @@ func (m *UpdateClusterMetadata) GetFinishTime() *timestamp.Timestamp { type DeleteClusterRequest struct { // The unique name of the cluster to be deleted. // Values are of the form projects//zones//clusters/ - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -469,7 +469,7 @@ func (m *DeleteClusterRequest) GetName() string { type UndeleteClusterRequest struct { // The unique name of the cluster to be un-deleted. // Values are of the form projects//zones//clusters/ - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -510,9 +510,9 @@ func (m *UndeleteClusterRequest) GetName() string { // BigtableClusterService.UndeleteCluster. type UndeleteClusterMetadata struct { // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which this operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_data.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_data.pb.go index 4e3e45847..518b8a619 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_data.pb.go @@ -45,17 +45,17 @@ func (Table_TimestampGranularity) EnumDescriptor() ([]byte, []int) { type Table struct { // A unique identifier of the form // /tables/[_a-zA-Z0-9][-_.a-zA-Z0-9]* - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // If this Table is in the process of being created, the Operation used to // track its progress. As long as this operation is present, the Table will // not accept any Table Admin or Read/Write requests. - CurrentOperation *longrunning.Operation `protobuf:"bytes,2,opt,name=current_operation,json=currentOperation" json:"current_operation,omitempty"` + CurrentOperation *longrunning.Operation `protobuf:"bytes,2,opt,name=current_operation,json=currentOperation,proto3" json:"current_operation,omitempty"` // The column families configured for this table, mapped by column family id. - ColumnFamilies map[string]*ColumnFamily `protobuf:"bytes,3,rep,name=column_families,json=columnFamilies" json:"column_families,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ColumnFamilies map[string]*ColumnFamily `protobuf:"bytes,3,rep,name=column_families,json=columnFamilies,proto3" json:"column_families,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The granularity (e.g. MILLIS, MICROS) at which timestamps are stored in // this table. Timestamps not matching the granularity will be rejected. // Cannot be changed once the table is created. - Granularity Table_TimestampGranularity `protobuf:"varint,4,opt,name=granularity,enum=google.bigtable.admin.table.v1.Table_TimestampGranularity" json:"granularity,omitempty"` + Granularity Table_TimestampGranularity `protobuf:"varint,4,opt,name=granularity,proto3,enum=google.bigtable.admin.table.v1.Table_TimestampGranularity" json:"granularity,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -118,7 +118,7 @@ type ColumnFamily struct { // A unique identifier of the form /columnFamilies/[-_.a-zA-Z0-9]+ // The last segment is the same as the "name" field in // google.bigtable.v1.Family. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Garbage collection expression specified by the following grammar: // GC = EXPR // | "" ; @@ -144,7 +144,7 @@ type ColumnFamily struct { // Garbage collection executes opportunistically in the background, and so // it's possible for reads to return a cell even if it matches the active GC // expression for its family. - GcExpression string `protobuf:"bytes,2,opt,name=gc_expression,json=gcExpression" json:"gc_expression,omitempty"` + GcExpression string `protobuf:"bytes,2,opt,name=gc_expression,json=gcExpression,proto3" json:"gc_expression,omitempty"` // Garbage collection rule specified as a protobuf. // Supersedes `gc_expression`. // Must serialize to at most 500 bytes. @@ -152,7 +152,7 @@ type ColumnFamily struct { // NOTE: Garbage collection executes opportunistically in the background, and // so it's possible for reads to return a cell even if it matches the active // GC expression for its family. - GcRule *GcRule `protobuf:"bytes,3,opt,name=gc_rule,json=gcRule" json:"gc_rule,omitempty"` + GcRule *GcRule `protobuf:"bytes,3,opt,name=gc_rule,json=gcRule,proto3" json:"gc_rule,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -245,16 +245,16 @@ type isGcRule_Rule interface { } type GcRule_MaxNumVersions struct { - MaxNumVersions int32 `protobuf:"varint,1,opt,name=max_num_versions,json=maxNumVersions,oneof"` + MaxNumVersions int32 `protobuf:"varint,1,opt,name=max_num_versions,json=maxNumVersions,proto3,oneof"` } type GcRule_MaxAge struct { - MaxAge *duration.Duration `protobuf:"bytes,2,opt,name=max_age,json=maxAge,oneof"` + MaxAge *duration.Duration `protobuf:"bytes,2,opt,name=max_age,json=maxAge,proto3,oneof"` } type GcRule_Intersection_ struct { - Intersection *GcRule_Intersection `protobuf:"bytes,3,opt,name=intersection,oneof"` + Intersection *GcRule_Intersection `protobuf:"bytes,3,opt,name=intersection,proto3,oneof"` } type GcRule_Union_ struct { - Union *GcRule_Union `protobuf:"bytes,4,opt,name=union,oneof"` + Union *GcRule_Union `protobuf:"bytes,4,opt,name=union,proto3,oneof"` } func (*GcRule_MaxNumVersions) isGcRule_Rule() {} @@ -407,7 +407,7 @@ func _GcRule_OneofSizer(msg proto.Message) (n int) { // A GcRule which deletes cells matching all of the given rules. type GcRule_Intersection struct { // Only delete cells which would be deleted by every element of `rules`. - Rules []*GcRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*GcRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -447,7 +447,7 @@ func (m *GcRule_Intersection) GetRules() []*GcRule { // A GcRule which deletes cells matching any of the given rules. type GcRule_Union struct { // Delete cells which would be deleted by any element of `rules`. - Rules []*GcRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*GcRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service.pb.go index 071066590..6f9519bae 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service.pb.go @@ -149,8 +149,7 @@ func (c *bigtableTableServiceClient) BulkDeleteRows(ctx context.Context, in *Bul return out, nil } -// Server API for BigtableTableService service - +// BigtableTableServiceServer is the server API for BigtableTableService service. type BigtableTableServiceServer interface { // Creates a new table, to be served from a specified cluster. // The table can be created with a full set of initial column families, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service_messages.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service_messages.pb.go index fd59676b2..b3bb6d141 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service_messages.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/table/v1/bigtable_table_service_messages.pb.go @@ -20,13 +20,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type CreateTableRequest struct { // The unique name of the cluster in which to create the new table. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name by which the new table should be referred to within the cluster, // e.g. "foobar" rather than "/tables/foobar". - TableId string `protobuf:"bytes,2,opt,name=table_id,json=tableId" json:"table_id,omitempty"` + TableId string `protobuf:"bytes,2,opt,name=table_id,json=tableId,proto3" json:"table_id,omitempty"` // The Table to create. The `name` field of the Table and all of its // ColumnFamilies must be left blank, and will be populated in the response. - Table *Table `protobuf:"bytes,3,opt,name=table" json:"table,omitempty"` + Table *Table `protobuf:"bytes,3,opt,name=table,proto3" json:"table,omitempty"` // The optional list of row keys that will be used to initially split the // table into several tablets (Tablets are similar to HBase regions). // Given two split keys, "s1" and "s2", three tablets will be created, @@ -42,7 +42,7 @@ type CreateTableRequest struct { // - Tablet 3 [customer_1, customer_2) => {"customer_1"}. // - Tablet 4 [customer_2, other) => {"customer_2"}. // - Tablet 5 [other, ) => {"other", "zz"}. - InitialSplitKeys []string `protobuf:"bytes,4,rep,name=initial_split_keys,json=initialSplitKeys" json:"initial_split_keys,omitempty"` + InitialSplitKeys []string `protobuf:"bytes,4,rep,name=initial_split_keys,json=initialSplitKeys,proto3" json:"initial_split_keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -102,7 +102,7 @@ func (m *CreateTableRequest) GetInitialSplitKeys() []string { type ListTablesRequest struct { // The unique name of the cluster for which tables should be listed. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -142,7 +142,7 @@ func (m *ListTablesRequest) GetName() string { type ListTablesResponse struct { // The tables present in the requested cluster. // At present, only the names of the tables are populated. - Tables []*Table `protobuf:"bytes,1,rep,name=tables" json:"tables,omitempty"` + Tables []*Table `protobuf:"bytes,1,rep,name=tables,proto3" json:"tables,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -181,7 +181,7 @@ func (m *ListTablesResponse) GetTables() []*Table { type GetTableRequest struct { // The unique name of the requested table. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -220,7 +220,7 @@ func (m *GetTableRequest) GetName() string { type DeleteTableRequest struct { // The unique name of the table to be deleted. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -259,10 +259,10 @@ func (m *DeleteTableRequest) GetName() string { type RenameTableRequest struct { // The current unique name of the table. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The new name by which the table should be referred to within its containing // cluster, e.g. "foobar" rather than "/tables/foobar". - NewId string `protobuf:"bytes,2,opt,name=new_id,json=newId" json:"new_id,omitempty"` + NewId string `protobuf:"bytes,2,opt,name=new_id,json=newId,proto3" json:"new_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -308,12 +308,12 @@ func (m *RenameTableRequest) GetNewId() string { type CreateColumnFamilyRequest struct { // The unique name of the table in which to create the new column family. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name by which the new column family should be referred to within the // table, e.g. "foobar" rather than "/columnFamilies/foobar". - ColumnFamilyId string `protobuf:"bytes,2,opt,name=column_family_id,json=columnFamilyId" json:"column_family_id,omitempty"` + ColumnFamilyId string `protobuf:"bytes,2,opt,name=column_family_id,json=columnFamilyId,proto3" json:"column_family_id,omitempty"` // The column family to create. The `name` field must be left blank. - ColumnFamily *ColumnFamily `protobuf:"bytes,3,opt,name=column_family,json=columnFamily" json:"column_family,omitempty"` + ColumnFamily *ColumnFamily `protobuf:"bytes,3,opt,name=column_family,json=columnFamily,proto3" json:"column_family,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -366,7 +366,7 @@ func (m *CreateColumnFamilyRequest) GetColumnFamily() *ColumnFamily { type DeleteColumnFamilyRequest struct { // The unique name of the column family to be deleted. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -405,7 +405,7 @@ func (m *DeleteColumnFamilyRequest) GetName() string { type BulkDeleteRowsRequest struct { // The unique name of the table on which to perform the bulk delete - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // Types that are valid to be assigned to Target: // *BulkDeleteRowsRequest_RowKeyPrefix // *BulkDeleteRowsRequest_DeleteAllDataFromTable @@ -447,7 +447,7 @@ type BulkDeleteRowsRequest_RowKeyPrefix struct { RowKeyPrefix []byte `protobuf:"bytes,2,opt,name=row_key_prefix,json=rowKeyPrefix,proto3,oneof"` } type BulkDeleteRowsRequest_DeleteAllDataFromTable struct { - DeleteAllDataFromTable bool `protobuf:"varint,3,opt,name=delete_all_data_from_table,json=deleteAllDataFromTable,oneof"` + DeleteAllDataFromTable bool `protobuf:"varint,3,opt,name=delete_all_data_from_table,json=deleteAllDataFromTable,proto3,oneof"` } func (*BulkDeleteRowsRequest_RowKeyPrefix) isBulkDeleteRowsRequest_Target() {} diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_instance_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_instance_admin.pb.go index 2c684c386..28768345a 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_instance_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_instance_admin.pb.go @@ -33,20 +33,20 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type CreateInstanceRequest struct { // The unique name of the project in which to create the new instance. // Values are of the form `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The ID to be used when referring to the new instance within its project, // e.g., just `myinstance` rather than // `projects/myproject/instances/myinstance`. - InstanceId string `protobuf:"bytes,2,opt,name=instance_id,json=instanceId" json:"instance_id,omitempty"` + InstanceId string `protobuf:"bytes,2,opt,name=instance_id,json=instanceId,proto3" json:"instance_id,omitempty"` // The instance to create. // Fields marked `OutputOnly` must be left blank. - Instance *Instance `protobuf:"bytes,3,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,3,opt,name=instance,proto3" json:"instance,omitempty"` // The clusters to be created within the instance, mapped by desired // cluster ID, e.g., just `mycluster` rather than // `projects/myproject/instances/myinstance/clusters/mycluster`. // Fields marked `OutputOnly` must be left blank. // Currently exactly one cluster must be specified. - Clusters map[string]*Cluster `protobuf:"bytes,4,rep,name=clusters" json:"clusters,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Clusters map[string]*Cluster `protobuf:"bytes,4,rep,name=clusters,proto3" json:"clusters,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -56,7 +56,7 @@ func (m *CreateInstanceRequest) Reset() { *m = CreateInstanceRequest{} } func (m *CreateInstanceRequest) String() string { return proto.CompactTextString(m) } func (*CreateInstanceRequest) ProtoMessage() {} func (*CreateInstanceRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{0} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{0} } func (m *CreateInstanceRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateInstanceRequest.Unmarshal(m, b) @@ -108,7 +108,7 @@ func (m *CreateInstanceRequest) GetClusters() map[string]*Cluster { type GetInstanceRequest struct { // The unique name of the requested instance. Values are of the form // `projects//instances/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -118,7 +118,7 @@ func (m *GetInstanceRequest) Reset() { *m = GetInstanceRequest{} } func (m *GetInstanceRequest) String() string { return proto.CompactTextString(m) } func (*GetInstanceRequest) ProtoMessage() {} func (*GetInstanceRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{1} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{1} } func (m *GetInstanceRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetInstanceRequest.Unmarshal(m, b) @@ -149,9 +149,9 @@ func (m *GetInstanceRequest) GetName() string { type ListInstancesRequest struct { // The unique name of the project for which a list of instances is requested. // Values are of the form `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The value of `next_page_token` returned by a previous call. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -161,7 +161,7 @@ func (m *ListInstancesRequest) Reset() { *m = ListInstancesRequest{} } func (m *ListInstancesRequest) String() string { return proto.CompactTextString(m) } func (*ListInstancesRequest) ProtoMessage() {} func (*ListInstancesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{2} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{2} } func (m *ListInstancesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListInstancesRequest.Unmarshal(m, b) @@ -198,17 +198,17 @@ func (m *ListInstancesRequest) GetPageToken() string { // Response message for BigtableInstanceAdmin.ListInstances. type ListInstancesResponse struct { // The list of requested instances. - Instances []*Instance `protobuf:"bytes,1,rep,name=instances" json:"instances,omitempty"` + Instances []*Instance `protobuf:"bytes,1,rep,name=instances,proto3" json:"instances,omitempty"` // Locations from which Instance information could not be retrieved, // due to an outage or some other transient condition. // Instances whose Clusters are all in one of the failed locations // may be missing from `instances`, and Instances with at least one // Cluster in a failed location may only have partial information returned. - FailedLocations []string `protobuf:"bytes,2,rep,name=failed_locations,json=failedLocations" json:"failed_locations,omitempty"` + FailedLocations []string `protobuf:"bytes,2,rep,name=failed_locations,json=failedLocations,proto3" json:"failed_locations,omitempty"` // Set if not all instances could be returned in a single response. // Pass this value to `page_token` in another request to get the next // page of results. - NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -218,7 +218,7 @@ func (m *ListInstancesResponse) Reset() { *m = ListInstancesResponse{} } func (m *ListInstancesResponse) String() string { return proto.CompactTextString(m) } func (*ListInstancesResponse) ProtoMessage() {} func (*ListInstancesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{3} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{3} } func (m *ListInstancesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListInstancesResponse.Unmarshal(m, b) @@ -262,10 +262,10 @@ func (m *ListInstancesResponse) GetNextPageToken() string { // Request message for BigtableInstanceAdmin.PartialUpdateInstance. type PartialUpdateInstanceRequest struct { // The Instance which will (partially) replace the current value. - Instance *Instance `protobuf:"bytes,1,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,1,opt,name=instance,proto3" json:"instance,omitempty"` // The subset of Instance fields which should be replaced. // Must be explicitly set. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -275,7 +275,7 @@ func (m *PartialUpdateInstanceRequest) Reset() { *m = PartialUpdateInsta func (m *PartialUpdateInstanceRequest) String() string { return proto.CompactTextString(m) } func (*PartialUpdateInstanceRequest) ProtoMessage() {} func (*PartialUpdateInstanceRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{4} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{4} } func (m *PartialUpdateInstanceRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PartialUpdateInstanceRequest.Unmarshal(m, b) @@ -313,7 +313,7 @@ func (m *PartialUpdateInstanceRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteInstanceRequest struct { // The unique name of the instance to be deleted. // Values are of the form `projects//instances/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -323,7 +323,7 @@ func (m *DeleteInstanceRequest) Reset() { *m = DeleteInstanceRequest{} } func (m *DeleteInstanceRequest) String() string { return proto.CompactTextString(m) } func (*DeleteInstanceRequest) ProtoMessage() {} func (*DeleteInstanceRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{5} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{5} } func (m *DeleteInstanceRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteInstanceRequest.Unmarshal(m, b) @@ -355,14 +355,14 @@ type CreateClusterRequest struct { // The unique name of the instance in which to create the new cluster. // Values are of the form // `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The ID to be used when referring to the new cluster within its instance, // e.g., just `mycluster` rather than // `projects/myproject/instances/myinstance/clusters/mycluster`. - ClusterId string `protobuf:"bytes,2,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,2,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The cluster to be created. // Fields marked `OutputOnly` must be left blank. - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -372,7 +372,7 @@ func (m *CreateClusterRequest) Reset() { *m = CreateClusterRequest{} } func (m *CreateClusterRequest) String() string { return proto.CompactTextString(m) } func (*CreateClusterRequest) ProtoMessage() {} func (*CreateClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{6} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{6} } func (m *CreateClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateClusterRequest.Unmarshal(m, b) @@ -417,7 +417,7 @@ func (m *CreateClusterRequest) GetCluster() *Cluster { type GetClusterRequest struct { // The unique name of the requested cluster. Values are of the form // `projects//instances//clusters/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -427,7 +427,7 @@ func (m *GetClusterRequest) Reset() { *m = GetClusterRequest{} } func (m *GetClusterRequest) String() string { return proto.CompactTextString(m) } func (*GetClusterRequest) ProtoMessage() {} func (*GetClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{7} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{7} } func (m *GetClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetClusterRequest.Unmarshal(m, b) @@ -460,9 +460,9 @@ type ListClustersRequest struct { // Values are of the form `projects//instances/`. // Use ` = '-'` to list Clusters for all Instances in a project, // e.g., `projects/myproject/instances/-`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The value of `next_page_token` returned by a previous call. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -472,7 +472,7 @@ func (m *ListClustersRequest) Reset() { *m = ListClustersRequest{} } func (m *ListClustersRequest) String() string { return proto.CompactTextString(m) } func (*ListClustersRequest) ProtoMessage() {} func (*ListClustersRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{8} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{8} } func (m *ListClustersRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListClustersRequest.Unmarshal(m, b) @@ -509,16 +509,16 @@ func (m *ListClustersRequest) GetPageToken() string { // Response message for BigtableInstanceAdmin.ListClusters. type ListClustersResponse struct { // The list of requested clusters. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` // Locations from which Cluster information could not be retrieved, // due to an outage or some other transient condition. // Clusters from these locations may be missing from `clusters`, // or may only have partial information returned. - FailedLocations []string `protobuf:"bytes,2,rep,name=failed_locations,json=failedLocations" json:"failed_locations,omitempty"` + FailedLocations []string `protobuf:"bytes,2,rep,name=failed_locations,json=failedLocations,proto3" json:"failed_locations,omitempty"` // Set if not all clusters could be returned in a single response. // Pass this value to `page_token` in another request to get the next // page of results. - NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -528,7 +528,7 @@ func (m *ListClustersResponse) Reset() { *m = ListClustersResponse{} } func (m *ListClustersResponse) String() string { return proto.CompactTextString(m) } func (*ListClustersResponse) ProtoMessage() {} func (*ListClustersResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{9} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{9} } func (m *ListClustersResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListClustersResponse.Unmarshal(m, b) @@ -573,7 +573,7 @@ func (m *ListClustersResponse) GetNextPageToken() string { type DeleteClusterRequest struct { // The unique name of the cluster to be deleted. Values are of the form // `projects//instances//clusters/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -583,7 +583,7 @@ func (m *DeleteClusterRequest) Reset() { *m = DeleteClusterRequest{} } func (m *DeleteClusterRequest) String() string { return proto.CompactTextString(m) } func (*DeleteClusterRequest) ProtoMessage() {} func (*DeleteClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{10} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{10} } func (m *DeleteClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteClusterRequest.Unmarshal(m, b) @@ -613,11 +613,11 @@ func (m *DeleteClusterRequest) GetName() string { // The metadata for the Operation returned by CreateInstance. type CreateInstanceMetadata struct { // The request that prompted the initiation of this CreateInstance operation. - OriginalRequest *CreateInstanceRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *CreateInstanceRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which the operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -627,7 +627,7 @@ func (m *CreateInstanceMetadata) Reset() { *m = CreateInstanceMetadata{} func (m *CreateInstanceMetadata) String() string { return proto.CompactTextString(m) } func (*CreateInstanceMetadata) ProtoMessage() {} func (*CreateInstanceMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{11} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{11} } func (m *CreateInstanceMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateInstanceMetadata.Unmarshal(m, b) @@ -671,11 +671,11 @@ func (m *CreateInstanceMetadata) GetFinishTime() *timestamp.Timestamp { // The metadata for the Operation returned by UpdateInstance. type UpdateInstanceMetadata struct { // The request that prompted the initiation of this UpdateInstance operation. - OriginalRequest *PartialUpdateInstanceRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *PartialUpdateInstanceRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which the operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -685,7 +685,7 @@ func (m *UpdateInstanceMetadata) Reset() { *m = UpdateInstanceMetadata{} func (m *UpdateInstanceMetadata) String() string { return proto.CompactTextString(m) } func (*UpdateInstanceMetadata) ProtoMessage() {} func (*UpdateInstanceMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{12} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{12} } func (m *UpdateInstanceMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateInstanceMetadata.Unmarshal(m, b) @@ -729,11 +729,11 @@ func (m *UpdateInstanceMetadata) GetFinishTime() *timestamp.Timestamp { // The metadata for the Operation returned by CreateCluster. type CreateClusterMetadata struct { // The request that prompted the initiation of this CreateCluster operation. - OriginalRequest *CreateClusterRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *CreateClusterRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which the operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -743,7 +743,7 @@ func (m *CreateClusterMetadata) Reset() { *m = CreateClusterMetadata{} } func (m *CreateClusterMetadata) String() string { return proto.CompactTextString(m) } func (*CreateClusterMetadata) ProtoMessage() {} func (*CreateClusterMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{13} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{13} } func (m *CreateClusterMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateClusterMetadata.Unmarshal(m, b) @@ -787,11 +787,11 @@ func (m *CreateClusterMetadata) GetFinishTime() *timestamp.Timestamp { // The metadata for the Operation returned by UpdateCluster. type UpdateClusterMetadata struct { // The request that prompted the initiation of this UpdateCluster operation. - OriginalRequest *Cluster `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *Cluster `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which the operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -801,7 +801,7 @@ func (m *UpdateClusterMetadata) Reset() { *m = UpdateClusterMetadata{} } func (m *UpdateClusterMetadata) String() string { return proto.CompactTextString(m) } func (*UpdateClusterMetadata) ProtoMessage() {} func (*UpdateClusterMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{14} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{14} } func (m *UpdateClusterMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateClusterMetadata.Unmarshal(m, b) @@ -842,26 +842,21 @@ func (m *UpdateClusterMetadata) GetFinishTime() *timestamp.Timestamp { return nil } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for BigtableInstanceAdmin.CreateAppProfile. type CreateAppProfileRequest struct { // The unique name of the instance in which to create the new app profile. // Values are of the form // `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The ID to be used when referring to the new app profile within its // instance, e.g., just `myprofile` rather than // `projects/myproject/instances/myinstance/appProfiles/myprofile`. - AppProfileId string `protobuf:"bytes,2,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,2,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` // The app profile to be created. // Fields marked `OutputOnly` will be ignored. - AppProfile *AppProfile `protobuf:"bytes,3,opt,name=app_profile,json=appProfile" json:"app_profile,omitempty"` + AppProfile *AppProfile `protobuf:"bytes,3,opt,name=app_profile,json=appProfile,proto3" json:"app_profile,omitempty"` // If true, ignore safety checks when creating the app profile. - IgnoreWarnings bool `protobuf:"varint,4,opt,name=ignore_warnings,json=ignoreWarnings" json:"ignore_warnings,omitempty"` + IgnoreWarnings bool `protobuf:"varint,4,opt,name=ignore_warnings,json=ignoreWarnings,proto3" json:"ignore_warnings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -871,7 +866,7 @@ func (m *CreateAppProfileRequest) Reset() { *m = CreateAppProfileRequest func (m *CreateAppProfileRequest) String() string { return proto.CompactTextString(m) } func (*CreateAppProfileRequest) ProtoMessage() {} func (*CreateAppProfileRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{15} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{15} } func (m *CreateAppProfileRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateAppProfileRequest.Unmarshal(m, b) @@ -919,16 +914,11 @@ func (m *CreateAppProfileRequest) GetIgnoreWarnings() bool { return false } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for BigtableInstanceAdmin.GetAppProfile. type GetAppProfileRequest struct { // The unique name of the requested app profile. Values are of the form // `projects//instances//appProfiles/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -938,7 +928,7 @@ func (m *GetAppProfileRequest) Reset() { *m = GetAppProfileRequest{} } func (m *GetAppProfileRequest) String() string { return proto.CompactTextString(m) } func (*GetAppProfileRequest) ProtoMessage() {} func (*GetAppProfileRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{16} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{16} } func (m *GetAppProfileRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetAppProfileRequest.Unmarshal(m, b) @@ -965,19 +955,14 @@ func (m *GetAppProfileRequest) GetName() string { return "" } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for BigtableInstanceAdmin.ListAppProfiles. type ListAppProfilesRequest struct { // The unique name of the instance for which a list of app profiles is // requested. Values are of the form // `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The value of `next_page_token` returned by a previous call. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -987,7 +972,7 @@ func (m *ListAppProfilesRequest) Reset() { *m = ListAppProfilesRequest{} func (m *ListAppProfilesRequest) String() string { return proto.CompactTextString(m) } func (*ListAppProfilesRequest) ProtoMessage() {} func (*ListAppProfilesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{17} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{17} } func (m *ListAppProfilesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListAppProfilesRequest.Unmarshal(m, b) @@ -1021,19 +1006,14 @@ func (m *ListAppProfilesRequest) GetPageToken() string { return "" } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Response message for BigtableInstanceAdmin.ListAppProfiles. type ListAppProfilesResponse struct { // The list of requested app profiles. - AppProfiles []*AppProfile `protobuf:"bytes,1,rep,name=app_profiles,json=appProfiles" json:"app_profiles,omitempty"` + AppProfiles []*AppProfile `protobuf:"bytes,1,rep,name=app_profiles,json=appProfiles,proto3" json:"app_profiles,omitempty"` // Set if not all app profiles could be returned in a single response. // Pass this value to `page_token` in another request to get the next // page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1043,7 +1023,7 @@ func (m *ListAppProfilesResponse) Reset() { *m = ListAppProfilesResponse func (m *ListAppProfilesResponse) String() string { return proto.CompactTextString(m) } func (*ListAppProfilesResponse) ProtoMessage() {} func (*ListAppProfilesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{18} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{18} } func (m *ListAppProfilesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListAppProfilesResponse.Unmarshal(m, b) @@ -1077,20 +1057,15 @@ func (m *ListAppProfilesResponse) GetNextPageToken() string { return "" } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for BigtableInstanceAdmin.UpdateAppProfile. type UpdateAppProfileRequest struct { // The app profile which will (partially) replace the current value. - AppProfile *AppProfile `protobuf:"bytes,1,opt,name=app_profile,json=appProfile" json:"app_profile,omitempty"` + AppProfile *AppProfile `protobuf:"bytes,1,opt,name=app_profile,json=appProfile,proto3" json:"app_profile,omitempty"` // The subset of app profile fields which should be replaced. // If unset, all fields will be replaced. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // If true, ignore safety checks when updating the app profile. - IgnoreWarnings bool `protobuf:"varint,3,opt,name=ignore_warnings,json=ignoreWarnings" json:"ignore_warnings,omitempty"` + IgnoreWarnings bool `protobuf:"varint,3,opt,name=ignore_warnings,json=ignoreWarnings,proto3" json:"ignore_warnings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1100,7 +1075,7 @@ func (m *UpdateAppProfileRequest) Reset() { *m = UpdateAppProfileRequest func (m *UpdateAppProfileRequest) String() string { return proto.CompactTextString(m) } func (*UpdateAppProfileRequest) ProtoMessage() {} func (*UpdateAppProfileRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{19} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{19} } func (m *UpdateAppProfileRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateAppProfileRequest.Unmarshal(m, b) @@ -1141,18 +1116,13 @@ func (m *UpdateAppProfileRequest) GetIgnoreWarnings() bool { return false } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for BigtableInstanceAdmin.DeleteAppProfile. type DeleteAppProfileRequest struct { // The unique name of the app profile to be deleted. Values are of the form // `projects//instances//appProfiles/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // If true, ignore safety checks when deleting the app profile. - IgnoreWarnings bool `protobuf:"varint,2,opt,name=ignore_warnings,json=ignoreWarnings" json:"ignore_warnings,omitempty"` + IgnoreWarnings bool `protobuf:"varint,2,opt,name=ignore_warnings,json=ignoreWarnings,proto3" json:"ignore_warnings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1162,7 +1132,7 @@ func (m *DeleteAppProfileRequest) Reset() { *m = DeleteAppProfileRequest func (m *DeleteAppProfileRequest) String() string { return proto.CompactTextString(m) } func (*DeleteAppProfileRequest) ProtoMessage() {} func (*DeleteAppProfileRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{20} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{20} } func (m *DeleteAppProfileRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteAppProfileRequest.Unmarshal(m, b) @@ -1196,11 +1166,6 @@ func (m *DeleteAppProfileRequest) GetIgnoreWarnings() bool { return false } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // The metadata for the Operation returned by UpdateAppProfile. type UpdateAppProfileMetadata struct { XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1212,7 +1177,7 @@ func (m *UpdateAppProfileMetadata) Reset() { *m = UpdateAppProfileMetada func (m *UpdateAppProfileMetadata) String() string { return proto.CompactTextString(m) } func (*UpdateAppProfileMetadata) ProtoMessage() {} func (*UpdateAppProfileMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f, []int{21} + return fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa, []int{21} } func (m *UpdateAppProfileMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateAppProfileMetadata.Unmarshal(m, b) @@ -1292,65 +1257,22 @@ type BigtableInstanceAdminClient interface { UpdateCluster(ctx context.Context, in *Cluster, opts ...grpc.CallOption) (*longrunning.Operation, error) // Deletes a cluster from an instance. DeleteCluster(ctx context.Context, in *DeleteClusterRequest, opts ...grpc.CallOption) (*empty.Empty, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Creates an app profile within an instance. CreateAppProfile(ctx context.Context, in *CreateAppProfileRequest, opts ...grpc.CallOption) (*AppProfile, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Gets information about an app profile. GetAppProfile(ctx context.Context, in *GetAppProfileRequest, opts ...grpc.CallOption) (*AppProfile, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Lists information about app profiles in an instance. ListAppProfiles(ctx context.Context, in *ListAppProfilesRequest, opts ...grpc.CallOption) (*ListAppProfilesResponse, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Updates an app profile within an instance. UpdateAppProfile(ctx context.Context, in *UpdateAppProfileRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Deletes an app profile from an instance. DeleteAppProfile(ctx context.Context, in *DeleteAppProfileRequest, opts ...grpc.CallOption) (*empty.Empty, error) - // This is a private alpha release of Cloud Bigtable instance level - // permissions. This feature is not currently available to most Cloud Bigtable - // customers. This feature might be changed in backward-incompatible ways and - // is not recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Gets the access control policy for an instance resource. Returns an empty // policy if an instance exists but does not have a policy set. GetIamPolicy(ctx context.Context, in *v1.GetIamPolicyRequest, opts ...grpc.CallOption) (*v1.Policy, error) - // This is a private alpha release of Cloud Bigtable instance level - // permissions. This feature is not currently available to most Cloud Bigtable - // customers. This feature might be changed in backward-incompatible ways and - // is not recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Sets the access control policy on an instance resource. Replaces any // existing policy. SetIamPolicy(ctx context.Context, in *v1.SetIamPolicyRequest, opts ...grpc.CallOption) (*v1.Policy, error) - // This is a private alpha release of Cloud Bigtable instance level - // permissions. This feature is not currently available to most Cloud Bigtable - // customers. This feature might be changed in backward-incompatible ways and - // is not recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Returns permissions that the caller has on the specified instance resource. TestIamPermissions(ctx context.Context, in *v1.TestIamPermissionsRequest, opts ...grpc.CallOption) (*v1.TestIamPermissionsResponse, error) } @@ -1534,8 +1456,7 @@ func (c *bigtableInstanceAdminClient) TestIamPermissions(ctx context.Context, in return out, nil } -// Server API for BigtableInstanceAdmin service - +// BigtableInstanceAdminServer is the server API for BigtableInstanceAdmin service. type BigtableInstanceAdminServer interface { // Create an instance within a project. CreateInstance(context.Context, *CreateInstanceRequest) (*longrunning.Operation, error) @@ -1559,65 +1480,22 @@ type BigtableInstanceAdminServer interface { UpdateCluster(context.Context, *Cluster) (*longrunning.Operation, error) // Deletes a cluster from an instance. DeleteCluster(context.Context, *DeleteClusterRequest) (*empty.Empty, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Creates an app profile within an instance. CreateAppProfile(context.Context, *CreateAppProfileRequest) (*AppProfile, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Gets information about an app profile. GetAppProfile(context.Context, *GetAppProfileRequest) (*AppProfile, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Lists information about app profiles in an instance. ListAppProfiles(context.Context, *ListAppProfilesRequest) (*ListAppProfilesResponse, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Updates an app profile within an instance. UpdateAppProfile(context.Context, *UpdateAppProfileRequest) (*longrunning.Operation, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Deletes an app profile from an instance. DeleteAppProfile(context.Context, *DeleteAppProfileRequest) (*empty.Empty, error) - // This is a private alpha release of Cloud Bigtable instance level - // permissions. This feature is not currently available to most Cloud Bigtable - // customers. This feature might be changed in backward-incompatible ways and - // is not recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Gets the access control policy for an instance resource. Returns an empty // policy if an instance exists but does not have a policy set. GetIamPolicy(context.Context, *v1.GetIamPolicyRequest) (*v1.Policy, error) - // This is a private alpha release of Cloud Bigtable instance level - // permissions. This feature is not currently available to most Cloud Bigtable - // customers. This feature might be changed in backward-incompatible ways and - // is not recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Sets the access control policy on an instance resource. Replaces any // existing policy. SetIamPolicy(context.Context, *v1.SetIamPolicyRequest) (*v1.Policy, error) - // This is a private alpha release of Cloud Bigtable instance level - // permissions. This feature is not currently available to most Cloud Bigtable - // customers. This feature might be changed in backward-incompatible ways and - // is not recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Returns permissions that the caller has on the specified instance resource. TestIamPermissions(context.Context, *v1.TestIamPermissionsRequest) (*v1.TestIamPermissionsResponse, error) } @@ -2054,10 +1932,10 @@ var _BigtableInstanceAdmin_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/bigtable/admin/v2/bigtable_instance_admin.proto", fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f) + proto.RegisterFile("google/bigtable/admin/v2/bigtable_instance_admin.proto", fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa) } -var fileDescriptor_bigtable_instance_admin_eff75d51f8bccd2f = []byte{ +var fileDescriptor_bigtable_instance_admin_4a25aca1138ca2aa = []byte{ // 1566 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x59, 0xcf, 0x6f, 0xdc, 0xc4, 0x17, 0xd7, 0x6c, 0xfa, 0xed, 0xb7, 0x79, 0x9b, 0x5f, 0xdf, 0xf9, 0x36, 0xc9, 0xca, 0xf4, 0x47, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_table_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_table_admin.pb.go index 355765a50..164baac25 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_table_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/bigtable_table_admin.pb.go @@ -33,12 +33,12 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type CreateTableRequest struct { // The unique name of the instance in which to create the table. // Values are of the form `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The name by which the new table should be referred to within the parent // instance, e.g., `foobar` rather than `/tables/foobar`. - TableId string `protobuf:"bytes,2,opt,name=table_id,json=tableId" json:"table_id,omitempty"` + TableId string `protobuf:"bytes,2,opt,name=table_id,json=tableId,proto3" json:"table_id,omitempty"` // The Table to create. - Table *Table `protobuf:"bytes,3,opt,name=table" json:"table,omitempty"` + Table *Table `protobuf:"bytes,3,opt,name=table,proto3" json:"table,omitempty"` // The optional list of row keys that will be used to initially split the // table into several tablets (tablets are similar to HBase regions). // Given two split keys, `s1` and `s2`, three tablets will be created, @@ -55,7 +55,7 @@ type CreateTableRequest struct { // - Tablet 3 `[customer_1, customer_2) => {"customer_1"}.` // - Tablet 4 `[customer_2, other) => {"customer_2"}.` // - Tablet 5 `[other, ) => {"other", "zz"}.` - InitialSplits []*CreateTableRequest_Split `protobuf:"bytes,4,rep,name=initial_splits,json=initialSplits" json:"initial_splits,omitempty"` + InitialSplits []*CreateTableRequest_Split `protobuf:"bytes,4,rep,name=initial_splits,json=initialSplits,proto3" json:"initial_splits,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -65,7 +65,7 @@ func (m *CreateTableRequest) Reset() { *m = CreateTableRequest{} } func (m *CreateTableRequest) String() string { return proto.CompactTextString(m) } func (*CreateTableRequest) ProtoMessage() {} func (*CreateTableRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{0} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{0} } func (m *CreateTableRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateTableRequest.Unmarshal(m, b) @@ -126,7 +126,7 @@ func (m *CreateTableRequest_Split) Reset() { *m = CreateTableRequest_Spl func (m *CreateTableRequest_Split) String() string { return proto.CompactTextString(m) } func (*CreateTableRequest_Split) ProtoMessage() {} func (*CreateTableRequest_Split) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{0, 0} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{0, 0} } func (m *CreateTableRequest_Split) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateTableRequest_Split.Unmarshal(m, b) @@ -153,25 +153,25 @@ func (m *CreateTableRequest_Split) GetKey() []byte { return nil } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.CreateTableFromSnapshot][google.bigtable.admin.v2.BigtableTableAdmin.CreateTableFromSnapshot] +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type CreateTableFromSnapshotRequest struct { // The unique name of the instance in which to create the table. // Values are of the form `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The name by which the new table should be referred to within the parent // instance, e.g., `foobar` rather than `/tables/foobar`. - TableId string `protobuf:"bytes,2,opt,name=table_id,json=tableId" json:"table_id,omitempty"` + TableId string `protobuf:"bytes,2,opt,name=table_id,json=tableId,proto3" json:"table_id,omitempty"` // The unique name of the snapshot from which to restore the table. The // snapshot and the table must be in the same instance. // Values are of the form // `projects//instances//clusters//snapshots/`. - SourceSnapshot string `protobuf:"bytes,3,opt,name=source_snapshot,json=sourceSnapshot" json:"source_snapshot,omitempty"` + SourceSnapshot string `protobuf:"bytes,3,opt,name=source_snapshot,json=sourceSnapshot,proto3" json:"source_snapshot,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -181,7 +181,7 @@ func (m *CreateTableFromSnapshotRequest) Reset() { *m = CreateTableFromS func (m *CreateTableFromSnapshotRequest) String() string { return proto.CompactTextString(m) } func (*CreateTableFromSnapshotRequest) ProtoMessage() {} func (*CreateTableFromSnapshotRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{1} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{1} } func (m *CreateTableFromSnapshotRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateTableFromSnapshotRequest.Unmarshal(m, b) @@ -228,7 +228,7 @@ type DropRowRangeRequest struct { // The unique name of the table on which to drop a range of rows. // Values are of the form // `projects//instances//tables/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Delete all rows or by prefix. // // Types that are valid to be assigned to Target: @@ -244,7 +244,7 @@ func (m *DropRowRangeRequest) Reset() { *m = DropRowRangeRequest{} } func (m *DropRowRangeRequest) String() string { return proto.CompactTextString(m) } func (*DropRowRangeRequest) ProtoMessage() {} func (*DropRowRangeRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{2} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{2} } func (m *DropRowRangeRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DropRowRangeRequest.Unmarshal(m, b) @@ -272,7 +272,7 @@ type DropRowRangeRequest_RowKeyPrefix struct { RowKeyPrefix []byte `protobuf:"bytes,2,opt,name=row_key_prefix,json=rowKeyPrefix,proto3,oneof"` } type DropRowRangeRequest_DeleteAllDataFromTable struct { - DeleteAllDataFromTable bool `protobuf:"varint,3,opt,name=delete_all_data_from_table,json=deleteAllDataFromTable,oneof"` + DeleteAllDataFromTable bool `protobuf:"varint,3,opt,name=delete_all_data_from_table,json=deleteAllDataFromTable,proto3,oneof"` } func (*DropRowRangeRequest_RowKeyPrefix) isDropRowRangeRequest_Target() {} @@ -380,12 +380,12 @@ func _DropRowRangeRequest_OneofSizer(msg proto.Message) (n int) { type ListTablesRequest struct { // The unique name of the instance for which tables should be listed. // Values are of the form `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The view to be applied to the returned tables' fields. // Defaults to `NAME_ONLY` if unspecified; no others are currently supported. - View Table_View `protobuf:"varint,2,opt,name=view,enum=google.bigtable.admin.v2.Table_View" json:"view,omitempty"` + View Table_View `protobuf:"varint,2,opt,name=view,proto3,enum=google.bigtable.admin.v2.Table_View" json:"view,omitempty"` // The value of `next_page_token` returned by a previous call. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -395,7 +395,7 @@ func (m *ListTablesRequest) Reset() { *m = ListTablesRequest{} } func (m *ListTablesRequest) String() string { return proto.CompactTextString(m) } func (*ListTablesRequest) ProtoMessage() {} func (*ListTablesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{3} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{3} } func (m *ListTablesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListTablesRequest.Unmarshal(m, b) @@ -440,11 +440,11 @@ func (m *ListTablesRequest) GetPageToken() string { // [google.bigtable.admin.v2.BigtableTableAdmin.ListTables][google.bigtable.admin.v2.BigtableTableAdmin.ListTables] type ListTablesResponse struct { // The tables present in the requested instance. - Tables []*Table `protobuf:"bytes,1,rep,name=tables" json:"tables,omitempty"` + Tables []*Table `protobuf:"bytes,1,rep,name=tables,proto3" json:"tables,omitempty"` // Set if not all tables could be returned in a single response. // Pass this value to `page_token` in another request to get the next // page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -454,7 +454,7 @@ func (m *ListTablesResponse) Reset() { *m = ListTablesResponse{} } func (m *ListTablesResponse) String() string { return proto.CompactTextString(m) } func (*ListTablesResponse) ProtoMessage() {} func (*ListTablesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{4} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{4} } func (m *ListTablesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListTablesResponse.Unmarshal(m, b) @@ -494,10 +494,10 @@ type GetTableRequest struct { // The unique name of the requested table. // Values are of the form // `projects//instances//tables/
`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The view to be applied to the returned table's fields. // Defaults to `SCHEMA_VIEW` if unspecified. - View Table_View `protobuf:"varint,2,opt,name=view,enum=google.bigtable.admin.v2.Table_View" json:"view,omitempty"` + View Table_View `protobuf:"varint,2,opt,name=view,proto3,enum=google.bigtable.admin.v2.Table_View" json:"view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -507,7 +507,7 @@ func (m *GetTableRequest) Reset() { *m = GetTableRequest{} } func (m *GetTableRequest) String() string { return proto.CompactTextString(m) } func (*GetTableRequest) ProtoMessage() {} func (*GetTableRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{5} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{5} } func (m *GetTableRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetTableRequest.Unmarshal(m, b) @@ -547,7 +547,7 @@ type DeleteTableRequest struct { // The unique name of the table to be deleted. // Values are of the form // `projects//instances//tables/
`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -557,7 +557,7 @@ func (m *DeleteTableRequest) Reset() { *m = DeleteTableRequest{} } func (m *DeleteTableRequest) String() string { return proto.CompactTextString(m) } func (*DeleteTableRequest) ProtoMessage() {} func (*DeleteTableRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{6} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{6} } func (m *DeleteTableRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteTableRequest.Unmarshal(m, b) @@ -590,12 +590,12 @@ type ModifyColumnFamiliesRequest struct { // The unique name of the table whose families should be modified. // Values are of the form // `projects//instances//tables/
`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Modifications to be atomically applied to the specified table's families. // Entries are applied in order, meaning that earlier modifications can be // masked by later ones (in the case of repeated updates to the same family, // for example). - Modifications []*ModifyColumnFamiliesRequest_Modification `protobuf:"bytes,2,rep,name=modifications" json:"modifications,omitempty"` + Modifications []*ModifyColumnFamiliesRequest_Modification `protobuf:"bytes,2,rep,name=modifications,proto3" json:"modifications,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -605,7 +605,7 @@ func (m *ModifyColumnFamiliesRequest) Reset() { *m = ModifyColumnFamilie func (m *ModifyColumnFamiliesRequest) String() string { return proto.CompactTextString(m) } func (*ModifyColumnFamiliesRequest) ProtoMessage() {} func (*ModifyColumnFamiliesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{7} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{7} } func (m *ModifyColumnFamiliesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ModifyColumnFamiliesRequest.Unmarshal(m, b) @@ -642,7 +642,7 @@ func (m *ModifyColumnFamiliesRequest) GetModifications() []*ModifyColumnFamilies // A create, update, or delete of a particular column family. type ModifyColumnFamiliesRequest_Modification struct { // The ID of the column family to be modified. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // Column familiy modifications. // // Types that are valid to be assigned to Mod: @@ -661,7 +661,7 @@ func (m *ModifyColumnFamiliesRequest_Modification) Reset() { func (m *ModifyColumnFamiliesRequest_Modification) String() string { return proto.CompactTextString(m) } func (*ModifyColumnFamiliesRequest_Modification) ProtoMessage() {} func (*ModifyColumnFamiliesRequest_Modification) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{7, 0} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{7, 0} } func (m *ModifyColumnFamiliesRequest_Modification) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ModifyColumnFamiliesRequest_Modification.Unmarshal(m, b) @@ -686,13 +686,13 @@ type isModifyColumnFamiliesRequest_Modification_Mod interface { } type ModifyColumnFamiliesRequest_Modification_Create struct { - Create *ColumnFamily `protobuf:"bytes,2,opt,name=create,oneof"` + Create *ColumnFamily `protobuf:"bytes,2,opt,name=create,proto3,oneof"` } type ModifyColumnFamiliesRequest_Modification_Update struct { - Update *ColumnFamily `protobuf:"bytes,3,opt,name=update,oneof"` + Update *ColumnFamily `protobuf:"bytes,3,opt,name=update,proto3,oneof"` } type ModifyColumnFamiliesRequest_Modification_Drop struct { - Drop bool `protobuf:"varint,4,opt,name=drop,oneof"` + Drop bool `protobuf:"varint,4,opt,name=drop,proto3,oneof"` } func (*ModifyColumnFamiliesRequest_Modification_Create) isModifyColumnFamiliesRequest_Modification_Mod() { @@ -829,18 +829,13 @@ func _ModifyColumnFamiliesRequest_Modification_OneofSizer(msg proto.Message) (n return n } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.GenerateConsistencyToken][google.bigtable.admin.v2.BigtableTableAdmin.GenerateConsistencyToken] type GenerateConsistencyTokenRequest struct { // The unique name of the Table for which to create a consistency token. // Values are of the form // `projects//instances//tables/
`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -850,7 +845,7 @@ func (m *GenerateConsistencyTokenRequest) Reset() { *m = GenerateConsist func (m *GenerateConsistencyTokenRequest) String() string { return proto.CompactTextString(m) } func (*GenerateConsistencyTokenRequest) ProtoMessage() {} func (*GenerateConsistencyTokenRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{8} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{8} } func (m *GenerateConsistencyTokenRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GenerateConsistencyTokenRequest.Unmarshal(m, b) @@ -877,16 +872,11 @@ func (m *GenerateConsistencyTokenRequest) GetName() string { return "" } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Response message for // [google.bigtable.admin.v2.BigtableTableAdmin.GenerateConsistencyToken][google.bigtable.admin.v2.BigtableTableAdmin.GenerateConsistencyToken] type GenerateConsistencyTokenResponse struct { // The generated consistency token. - ConsistencyToken string `protobuf:"bytes,1,opt,name=consistency_token,json=consistencyToken" json:"consistency_token,omitempty"` + ConsistencyToken string `protobuf:"bytes,1,opt,name=consistency_token,json=consistencyToken,proto3" json:"consistency_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -896,7 +886,7 @@ func (m *GenerateConsistencyTokenResponse) Reset() { *m = GenerateConsis func (m *GenerateConsistencyTokenResponse) String() string { return proto.CompactTextString(m) } func (*GenerateConsistencyTokenResponse) ProtoMessage() {} func (*GenerateConsistencyTokenResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{9} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{9} } func (m *GenerateConsistencyTokenResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GenerateConsistencyTokenResponse.Unmarshal(m, b) @@ -923,20 +913,15 @@ func (m *GenerateConsistencyTokenResponse) GetConsistencyToken() string { return "" } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.CheckConsistency][google.bigtable.admin.v2.BigtableTableAdmin.CheckConsistency] type CheckConsistencyRequest struct { // The unique name of the Table for which to check replication consistency. // Values are of the form // `projects//instances//tables/
`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The token created using GenerateConsistencyToken for the Table. - ConsistencyToken string `protobuf:"bytes,2,opt,name=consistency_token,json=consistencyToken" json:"consistency_token,omitempty"` + ConsistencyToken string `protobuf:"bytes,2,opt,name=consistency_token,json=consistencyToken,proto3" json:"consistency_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -946,7 +931,7 @@ func (m *CheckConsistencyRequest) Reset() { *m = CheckConsistencyRequest func (m *CheckConsistencyRequest) String() string { return proto.CompactTextString(m) } func (*CheckConsistencyRequest) ProtoMessage() {} func (*CheckConsistencyRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{10} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{10} } func (m *CheckConsistencyRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CheckConsistencyRequest.Unmarshal(m, b) @@ -980,17 +965,12 @@ func (m *CheckConsistencyRequest) GetConsistencyToken() string { return "" } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Response message for // [google.bigtable.admin.v2.BigtableTableAdmin.CheckConsistency][google.bigtable.admin.v2.BigtableTableAdmin.CheckConsistency] type CheckConsistencyResponse struct { // True only if the token is consistent. A token is consistent if replication // has caught up with the restrictions specified in the request. - Consistent bool `protobuf:"varint,1,opt,name=consistent" json:"consistent,omitempty"` + Consistent bool `protobuf:"varint,1,opt,name=consistent,proto3" json:"consistent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1000,7 +980,7 @@ func (m *CheckConsistencyResponse) Reset() { *m = CheckConsistencyRespon func (m *CheckConsistencyResponse) String() string { return proto.CompactTextString(m) } func (*CheckConsistencyResponse) ProtoMessage() {} func (*CheckConsistencyResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{11} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{11} } func (m *CheckConsistencyResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CheckConsistencyResponse.Unmarshal(m, b) @@ -1027,34 +1007,34 @@ func (m *CheckConsistencyResponse) GetConsistent() bool { return false } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.SnapshotTable][google.bigtable.admin.v2.BigtableTableAdmin.SnapshotTable] +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type SnapshotTableRequest struct { // The unique name of the table to have the snapshot taken. // Values are of the form // `projects//instances//tables/
`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the cluster where the snapshot will be created in. // Values are of the form // `projects//instances//clusters/`. - Cluster string `protobuf:"bytes,2,opt,name=cluster" json:"cluster,omitempty"` + Cluster string `protobuf:"bytes,2,opt,name=cluster,proto3" json:"cluster,omitempty"` // The ID by which the new snapshot should be referred to within the parent // cluster, e.g., `mysnapshot` of the form: `[_a-zA-Z0-9][-_.a-zA-Z0-9]*` // rather than // `projects//instances//clusters//snapshots/mysnapshot`. - SnapshotId string `protobuf:"bytes,3,opt,name=snapshot_id,json=snapshotId" json:"snapshot_id,omitempty"` + SnapshotId string `protobuf:"bytes,3,opt,name=snapshot_id,json=snapshotId,proto3" json:"snapshot_id,omitempty"` // The amount of time that the new snapshot can stay active after it is // created. Once 'ttl' expires, the snapshot will get deleted. The maximum // amount of time a snapshot can stay active is 7 days. If 'ttl' is not // specified, the default value of 24 hours will be used. - Ttl *duration.Duration `protobuf:"bytes,4,opt,name=ttl" json:"ttl,omitempty"` + Ttl *duration.Duration `protobuf:"bytes,4,opt,name=ttl,proto3" json:"ttl,omitempty"` // Description of the snapshot. - Description string `protobuf:"bytes,5,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1064,7 +1044,7 @@ func (m *SnapshotTableRequest) Reset() { *m = SnapshotTableRequest{} } func (m *SnapshotTableRequest) String() string { return proto.CompactTextString(m) } func (*SnapshotTableRequest) ProtoMessage() {} func (*SnapshotTableRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{12} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{12} } func (m *SnapshotTableRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SnapshotTableRequest.Unmarshal(m, b) @@ -1119,18 +1099,18 @@ func (m *SnapshotTableRequest) GetDescription() string { return "" } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.GetSnapshot][google.bigtable.admin.v2.BigtableTableAdmin.GetSnapshot] +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type GetSnapshotRequest struct { // The unique name of the requested snapshot. // Values are of the form // `projects//instances//clusters//snapshots/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1140,7 +1120,7 @@ func (m *GetSnapshotRequest) Reset() { *m = GetSnapshotRequest{} } func (m *GetSnapshotRequest) String() string { return proto.CompactTextString(m) } func (*GetSnapshotRequest) ProtoMessage() {} func (*GetSnapshotRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{13} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{13} } func (m *GetSnapshotRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSnapshotRequest.Unmarshal(m, b) @@ -1167,24 +1147,24 @@ func (m *GetSnapshotRequest) GetName() string { return "" } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.ListSnapshots][google.bigtable.admin.v2.BigtableTableAdmin.ListSnapshots] +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type ListSnapshotsRequest struct { // The unique name of the cluster for which snapshots should be listed. // Values are of the form // `projects//instances//clusters/`. // Use ` = '-'` to list snapshots for all clusters in an instance, // e.g., `projects//instances//clusters/-`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The maximum number of snapshots to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value of `next_page_token` returned by a previous call. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1194,7 +1174,7 @@ func (m *ListSnapshotsRequest) Reset() { *m = ListSnapshotsRequest{} } func (m *ListSnapshotsRequest) String() string { return proto.CompactTextString(m) } func (*ListSnapshotsRequest) ProtoMessage() {} func (*ListSnapshotsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{14} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{14} } func (m *ListSnapshotsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSnapshotsRequest.Unmarshal(m, b) @@ -1235,20 +1215,20 @@ func (m *ListSnapshotsRequest) GetPageToken() string { return "" } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Response message for // [google.bigtable.admin.v2.BigtableTableAdmin.ListSnapshots][google.bigtable.admin.v2.BigtableTableAdmin.ListSnapshots] +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type ListSnapshotsResponse struct { // The snapshots present in the requested cluster. - Snapshots []*Snapshot `protobuf:"bytes,1,rep,name=snapshots" json:"snapshots,omitempty"` + Snapshots []*Snapshot `protobuf:"bytes,1,rep,name=snapshots,proto3" json:"snapshots,omitempty"` // Set if not all snapshots could be returned in a single response. // Pass this value to `page_token` in another request to get the next // page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1258,7 +1238,7 @@ func (m *ListSnapshotsResponse) Reset() { *m = ListSnapshotsResponse{} } func (m *ListSnapshotsResponse) String() string { return proto.CompactTextString(m) } func (*ListSnapshotsResponse) ProtoMessage() {} func (*ListSnapshotsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{15} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{15} } func (m *ListSnapshotsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSnapshotsResponse.Unmarshal(m, b) @@ -1292,18 +1272,18 @@ func (m *ListSnapshotsResponse) GetNextPageToken() string { return "" } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // Request message for // [google.bigtable.admin.v2.BigtableTableAdmin.DeleteSnapshot][google.bigtable.admin.v2.BigtableTableAdmin.DeleteSnapshot] +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type DeleteSnapshotRequest struct { // The unique name of the snapshot to be deleted. // Values are of the form // `projects//instances//clusters//snapshots/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1313,7 +1293,7 @@ func (m *DeleteSnapshotRequest) Reset() { *m = DeleteSnapshotRequest{} } func (m *DeleteSnapshotRequest) String() string { return proto.CompactTextString(m) } func (*DeleteSnapshotRequest) ProtoMessage() {} func (*DeleteSnapshotRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{16} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{16} } func (m *DeleteSnapshotRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteSnapshotRequest.Unmarshal(m, b) @@ -1340,19 +1320,19 @@ func (m *DeleteSnapshotRequest) GetName() string { return "" } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // The metadata for the Operation returned by SnapshotTable. +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type SnapshotTableMetadata struct { // The request that prompted the initiation of this SnapshotTable operation. - OriginalRequest *SnapshotTableRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *SnapshotTableRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which the operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1362,7 +1342,7 @@ func (m *SnapshotTableMetadata) Reset() { *m = SnapshotTableMetadata{} } func (m *SnapshotTableMetadata) String() string { return proto.CompactTextString(m) } func (*SnapshotTableMetadata) ProtoMessage() {} func (*SnapshotTableMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{17} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{17} } func (m *SnapshotTableMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SnapshotTableMetadata.Unmarshal(m, b) @@ -1403,20 +1383,20 @@ func (m *SnapshotTableMetadata) GetFinishTime() *timestamp.Timestamp { return nil } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // The metadata for the Operation returned by CreateTableFromSnapshot. +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type CreateTableFromSnapshotMetadata struct { // The request that prompted the initiation of this CreateTableFromSnapshot // operation. - OriginalRequest *CreateTableFromSnapshotRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + OriginalRequest *CreateTableFromSnapshotRequest `protobuf:"bytes,1,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The time at which the original request was received. - RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime" json:"request_time,omitempty"` + RequestTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=request_time,json=requestTime,proto3" json:"request_time,omitempty"` // The time at which the operation failed or was completed successfully. - FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1426,7 +1406,7 @@ func (m *CreateTableFromSnapshotMetadata) Reset() { *m = CreateTableFrom func (m *CreateTableFromSnapshotMetadata) String() string { return proto.CompactTextString(m) } func (*CreateTableFromSnapshotMetadata) ProtoMessage() {} func (*CreateTableFromSnapshotMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_table_admin_67b5438a794c6471, []int{18} + return fileDescriptor_bigtable_table_admin_5562daab61d38018, []int{18} } func (m *CreateTableFromSnapshotMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateTableFromSnapshotMetadata.Unmarshal(m, b) @@ -1507,13 +1487,14 @@ type BigtableTableAdminClient interface { // The table can be created with a full set of initial column families, // specified in the request. CreateTable(ctx context.Context, in *CreateTableRequest, opts ...grpc.CallOption) (*Table, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Creates a new table from the specified snapshot. The target table must // not exist. The snapshot and the table must be in the same instance. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. CreateTableFromSnapshot(ctx context.Context, in *CreateTableFromSnapshotRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) // Lists all tables served from a specified instance. ListTables(ctx context.Context, in *ListTablesRequest, opts ...grpc.CallOption) (*ListTablesResponse, error) @@ -1530,53 +1511,47 @@ type BigtableTableAdminClient interface { // specify whether to delete all rows in a table, or only those that match a // particular prefix. DropRowRange(ctx context.Context, in *DropRowRangeRequest, opts ...grpc.CallOption) (*empty.Empty, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Generates a consistency token for a Table, which can be used in // CheckConsistency to check whether mutations to the table that finished // before this call started have been replicated. The tokens will be available // for 90 days. GenerateConsistencyToken(ctx context.Context, in *GenerateConsistencyTokenRequest, opts ...grpc.CallOption) (*GenerateConsistencyTokenResponse, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Checks replication consistency based on a consistency token, that is, if // replication has caught up based on the conditions specified in the token // and the check request. CheckConsistency(ctx context.Context, in *CheckConsistencyRequest, opts ...grpc.CallOption) (*CheckConsistencyResponse, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Creates a new snapshot in the specified cluster from the specified // source table. The cluster and the table must be in the same instance. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. SnapshotTable(ctx context.Context, in *SnapshotTableRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Gets metadata information about the specified snapshot. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. GetSnapshot(ctx context.Context, in *GetSnapshotRequest, opts ...grpc.CallOption) (*Snapshot, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Lists all snapshots associated with the specified cluster. - ListSnapshots(ctx context.Context, in *ListSnapshotsRequest, opts ...grpc.CallOption) (*ListSnapshotsResponse, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. + ListSnapshots(ctx context.Context, in *ListSnapshotsRequest, opts ...grpc.CallOption) (*ListSnapshotsResponse, error) // Permanently deletes the specified snapshot. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. DeleteSnapshot(ctx context.Context, in *DeleteSnapshotRequest, opts ...grpc.CallOption) (*empty.Empty, error) } @@ -1705,20 +1680,20 @@ func (c *bigtableTableAdminClient) DeleteSnapshot(ctx context.Context, in *Delet return out, nil } -// Server API for BigtableTableAdmin service - +// BigtableTableAdminServer is the server API for BigtableTableAdmin service. type BigtableTableAdminServer interface { // Creates a new table in the specified instance. // The table can be created with a full set of initial column families, // specified in the request. CreateTable(context.Context, *CreateTableRequest) (*Table, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Creates a new table from the specified snapshot. The target table must // not exist. The snapshot and the table must be in the same instance. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. CreateTableFromSnapshot(context.Context, *CreateTableFromSnapshotRequest) (*longrunning.Operation, error) // Lists all tables served from a specified instance. ListTables(context.Context, *ListTablesRequest) (*ListTablesResponse, error) @@ -1735,53 +1710,47 @@ type BigtableTableAdminServer interface { // specify whether to delete all rows in a table, or only those that match a // particular prefix. DropRowRange(context.Context, *DropRowRangeRequest) (*empty.Empty, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Generates a consistency token for a Table, which can be used in // CheckConsistency to check whether mutations to the table that finished // before this call started have been replicated. The tokens will be available // for 90 days. GenerateConsistencyToken(context.Context, *GenerateConsistencyTokenRequest) (*GenerateConsistencyTokenResponse, error) - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Checks replication consistency based on a consistency token, that is, if // replication has caught up based on the conditions specified in the token // and the check request. CheckConsistency(context.Context, *CheckConsistencyRequest) (*CheckConsistencyResponse, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Creates a new snapshot in the specified cluster from the specified // source table. The cluster and the table must be in the same instance. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. SnapshotTable(context.Context, *SnapshotTableRequest) (*longrunning.Operation, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Gets metadata information about the specified snapshot. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. GetSnapshot(context.Context, *GetSnapshotRequest) (*Snapshot, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // // Lists all snapshots associated with the specified cluster. - ListSnapshots(context.Context, *ListSnapshotsRequest) (*ListSnapshotsResponse, error) - // This is a private alpha release of Cloud Bigtable snapshots. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. + ListSnapshots(context.Context, *ListSnapshotsRequest) (*ListSnapshotsResponse, error) // Permanently deletes the specified snapshot. + // + // Note: This is a private alpha release of Cloud Bigtable snapshots. This + // feature is not currently available to most Cloud Bigtable customers. This + // feature might be changed in backward-incompatible ways and is not + // recommended for production use. It is not subject to any SLA or deprecation + // policy. DeleteSnapshot(context.Context, *DeleteSnapshotRequest) (*empty.Empty, error) } @@ -2085,10 +2054,10 @@ var _BigtableTableAdmin_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/bigtable/admin/v2/bigtable_table_admin.proto", fileDescriptor_bigtable_table_admin_67b5438a794c6471) + proto.RegisterFile("google/bigtable/admin/v2/bigtable_table_admin.proto", fileDescriptor_bigtable_table_admin_5562daab61d38018) } -var fileDescriptor_bigtable_table_admin_67b5438a794c6471 = []byte{ +var fileDescriptor_bigtable_table_admin_5562daab61d38018 = []byte{ // 1514 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x58, 0xcb, 0x6f, 0xdc, 0x54, 0x17, 0xaf, 0x27, 0x8f, 0x26, 0x67, 0xf2, 0xea, 0xfd, 0xd2, 0x76, 0x3a, 0x6d, 0x93, 0xc8, 0x5f, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/common.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/common.pb.go index b58146405..d6477716a 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/common.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/common.pb.go @@ -47,7 +47,7 @@ func (x StorageType) String() string { return proto.EnumName(StorageType_name, int32(x)) } func (StorageType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_common_a615f4766e498da1, []int{0} + return fileDescriptor_common_193eca34a1716695, []int{0} } func init() { @@ -55,10 +55,10 @@ func init() { } func init() { - proto.RegisterFile("google/bigtable/admin/v2/common.proto", fileDescriptor_common_a615f4766e498da1) + proto.RegisterFile("google/bigtable/admin/v2/common.proto", fileDescriptor_common_193eca34a1716695) } -var fileDescriptor_common_a615f4766e498da1 = []byte{ +var fileDescriptor_common_193eca34a1716695 = []byte{ // 270 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x74, 0xd0, 0xcf, 0x4b, 0xc3, 0x30, 0x14, 0x07, 0x70, 0x3b, 0x41, 0x21, 0xbb, 0x94, 0x9e, 0xc6, 0x28, 0x7a, 0xf2, 0xe2, 0x21, 0x81, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/instance.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/instance.pb.go index 0da060588..57bf20ab3 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/instance.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/instance.pb.go @@ -48,7 +48,7 @@ func (x Instance_State) String() string { return proto.EnumName(Instance_State_name, int32(x)) } func (Instance_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{0, 0} + return fileDescriptor_instance_fa63b73e18194980, []int{0, 0} } // The type of the instance. @@ -87,7 +87,7 @@ func (x Instance_Type) String() string { return proto.EnumName(Instance_Type_name, int32(x)) } func (Instance_Type) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{0, 1} + return fileDescriptor_instance_fa63b73e18194980, []int{0, 1} } // Possible states of a cluster. @@ -132,7 +132,7 @@ func (x Cluster_State) String() string { return proto.EnumName(Cluster_State_name, int32(x)) } func (Cluster_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{1, 0} + return fileDescriptor_instance_fa63b73e18194980, []int{1, 0} } // A collection of Bigtable [Tables][google.bigtable.admin.v2.Table] and @@ -143,16 +143,16 @@ type Instance struct { // (`OutputOnly`) // The unique name of the instance. Values are of the form // `projects//instances/[a-z][a-z0-9\\-]+[a-z0-9]`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The descriptive name for this instance as it appears in UIs. // Can be changed at any time, but should be kept globally unique // to avoid confusion. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // (`OutputOnly`) // The current state of the instance. - State Instance_State `protobuf:"varint,3,opt,name=state,enum=google.bigtable.admin.v2.Instance_State" json:"state,omitempty"` + State Instance_State `protobuf:"varint,3,opt,name=state,proto3,enum=google.bigtable.admin.v2.Instance_State" json:"state,omitempty"` // The type of the instance. Defaults to `PRODUCTION`. - Type Instance_Type `protobuf:"varint,4,opt,name=type,enum=google.bigtable.admin.v2.Instance_Type" json:"type,omitempty"` + Type Instance_Type `protobuf:"varint,4,opt,name=type,proto3,enum=google.bigtable.admin.v2.Instance_Type" json:"type,omitempty"` // Labels are a flexible and lightweight mechanism for organizing cloud // resources into groups that reflect a customer's organizational needs and // deployment strategies. They can be used to filter resources and aggregate @@ -164,7 +164,7 @@ type Instance struct { // the regular expression: `[\p{Ll}\p{Lo}\p{N}_-]{0,63}`. // * No more than 64 labels can be associated with a given resource. // * Keys and values must both be under 128 bytes. - Labels map[string]string `protobuf:"bytes,5,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,5,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -174,7 +174,7 @@ func (m *Instance) Reset() { *m = Instance{} } func (m *Instance) String() string { return proto.CompactTextString(m) } func (*Instance) ProtoMessage() {} func (*Instance) Descriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{0} + return fileDescriptor_instance_fa63b73e18194980, []int{0} } func (m *Instance) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Instance.Unmarshal(m, b) @@ -236,23 +236,23 @@ type Cluster struct { // (`OutputOnly`) // The unique name of the cluster. Values are of the form // `projects//instances//clusters/[a-z][-a-z0-9]*`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // (`CreationOnly`) // The location where this cluster's nodes and storage reside. For best // performance, clients should be located as close as possible to this // cluster. Currently only zones are supported, so values should be of the // form `projects//locations/`. - Location string `protobuf:"bytes,2,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,2,opt,name=location,proto3" json:"location,omitempty"` // (`OutputOnly`) // The current state of the cluster. - State Cluster_State `protobuf:"varint,3,opt,name=state,enum=google.bigtable.admin.v2.Cluster_State" json:"state,omitempty"` + State Cluster_State `protobuf:"varint,3,opt,name=state,proto3,enum=google.bigtable.admin.v2.Cluster_State" json:"state,omitempty"` // The number of nodes allocated to this cluster. More nodes enable higher // throughput and more consistent performance. - ServeNodes int32 `protobuf:"varint,4,opt,name=serve_nodes,json=serveNodes" json:"serve_nodes,omitempty"` + ServeNodes int32 `protobuf:"varint,4,opt,name=serve_nodes,json=serveNodes,proto3" json:"serve_nodes,omitempty"` // (`CreationOnly`) // The type of storage used by this cluster to serve its // parent instance's tables, unless explicitly overridden. - DefaultStorageType StorageType `protobuf:"varint,5,opt,name=default_storage_type,json=defaultStorageType,enum=google.bigtable.admin.v2.StorageType" json:"default_storage_type,omitempty"` + DefaultStorageType StorageType `protobuf:"varint,5,opt,name=default_storage_type,json=defaultStorageType,proto3,enum=google.bigtable.admin.v2.StorageType" json:"default_storage_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,7 +262,7 @@ func (m *Cluster) Reset() { *m = Cluster{} } func (m *Cluster) String() string { return proto.CompactTextString(m) } func (*Cluster) ProtoMessage() {} func (*Cluster) Descriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{1} + return fileDescriptor_instance_fa63b73e18194980, []int{1} } func (m *Cluster) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Cluster.Unmarshal(m, b) @@ -317,18 +317,13 @@ func (m *Cluster) GetDefaultStorageType() StorageType { return StorageType_STORAGE_TYPE_UNSPECIFIED } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // A configuration object describing how Cloud Bigtable should treat traffic // from a particular end user application. type AppProfile struct { // (`OutputOnly`) // The unique name of the app profile. Values are of the form // `projects//instances//appProfiles/[_a-zA-Z0-9][-_.a-zA-Z0-9]*`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Strongly validated etag for optimistic concurrency control. Preserve the // value returned from `GetAppProfile` when calling `UpdateAppProfile` to // fail the request if there has been a modification in the mean time. The @@ -337,9 +332,9 @@ type AppProfile struct { // See [Wikipedia](https://en.wikipedia.org/wiki/HTTP_ETag) and // [RFC 7232](https://tools.ietf.org/html/rfc7232#section-2.3) for more // details. - Etag string `protobuf:"bytes,2,opt,name=etag" json:"etag,omitempty"` + Etag string `protobuf:"bytes,2,opt,name=etag,proto3" json:"etag,omitempty"` // Optional long form description of the use case for this AppProfile. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The routing policy for all read/write requests which use this app profile. // A value must be explicitly set. // @@ -356,7 +351,7 @@ func (m *AppProfile) Reset() { *m = AppProfile{} } func (m *AppProfile) String() string { return proto.CompactTextString(m) } func (*AppProfile) ProtoMessage() {} func (*AppProfile) Descriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{2} + return fileDescriptor_instance_fa63b73e18194980, []int{2} } func (m *AppProfile) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AppProfile.Unmarshal(m, b) @@ -381,10 +376,10 @@ type isAppProfile_RoutingPolicy interface { } type AppProfile_MultiClusterRoutingUseAny_ struct { - MultiClusterRoutingUseAny *AppProfile_MultiClusterRoutingUseAny `protobuf:"bytes,5,opt,name=multi_cluster_routing_use_any,json=multiClusterRoutingUseAny,oneof"` + MultiClusterRoutingUseAny *AppProfile_MultiClusterRoutingUseAny `protobuf:"bytes,5,opt,name=multi_cluster_routing_use_any,json=multiClusterRoutingUseAny,proto3,oneof"` } type AppProfile_SingleClusterRouting_ struct { - SingleClusterRouting *AppProfile_SingleClusterRouting `protobuf:"bytes,6,opt,name=single_cluster_routing,json=singleClusterRouting,oneof"` + SingleClusterRouting *AppProfile_SingleClusterRouting `protobuf:"bytes,6,opt,name=single_cluster_routing,json=singleClusterRouting,proto3,oneof"` } func (*AppProfile_MultiClusterRoutingUseAny_) isAppProfile_RoutingPolicy() {} @@ -520,7 +515,7 @@ func (m *AppProfile_MultiClusterRoutingUseAny) Reset() { *m = AppProfile func (m *AppProfile_MultiClusterRoutingUseAny) String() string { return proto.CompactTextString(m) } func (*AppProfile_MultiClusterRoutingUseAny) ProtoMessage() {} func (*AppProfile_MultiClusterRoutingUseAny) Descriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{2, 0} + return fileDescriptor_instance_fa63b73e18194980, []int{2, 0} } func (m *AppProfile_MultiClusterRoutingUseAny) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AppProfile_MultiClusterRoutingUseAny.Unmarshal(m, b) @@ -545,11 +540,11 @@ var xxx_messageInfo_AppProfile_MultiClusterRoutingUseAny proto.InternalMessageIn // availability. type AppProfile_SingleClusterRouting struct { // The cluster to which read/write requests should be routed. - ClusterId string `protobuf:"bytes,1,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,1,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Whether or not `CheckAndMutateRow` and `ReadModifyWriteRow` requests are // allowed by this app profile. It is unsafe to send these requests to // the same table/row/column in multiple clusters. - AllowTransactionalWrites bool `protobuf:"varint,2,opt,name=allow_transactional_writes,json=allowTransactionalWrites" json:"allow_transactional_writes,omitempty"` + AllowTransactionalWrites bool `protobuf:"varint,2,opt,name=allow_transactional_writes,json=allowTransactionalWrites,proto3" json:"allow_transactional_writes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -559,7 +554,7 @@ func (m *AppProfile_SingleClusterRouting) Reset() { *m = AppProfile_Sing func (m *AppProfile_SingleClusterRouting) String() string { return proto.CompactTextString(m) } func (*AppProfile_SingleClusterRouting) ProtoMessage() {} func (*AppProfile_SingleClusterRouting) Descriptor() ([]byte, []int) { - return fileDescriptor_instance_acf7ca11f8e8d868, []int{2, 1} + return fileDescriptor_instance_fa63b73e18194980, []int{2, 1} } func (m *AppProfile_SingleClusterRouting) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AppProfile_SingleClusterRouting.Unmarshal(m, b) @@ -606,10 +601,10 @@ func init() { } func init() { - proto.RegisterFile("google/bigtable/admin/v2/instance.proto", fileDescriptor_instance_acf7ca11f8e8d868) + proto.RegisterFile("google/bigtable/admin/v2/instance.proto", fileDescriptor_instance_fa63b73e18194980) } -var fileDescriptor_instance_acf7ca11f8e8d868 = []byte{ +var fileDescriptor_instance_fa63b73e18194980 = []byte{ // 765 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x54, 0xdd, 0x8e, 0xdb, 0x44, 0x14, 0x8e, 0xf3, 0xb3, 0x64, 0x4f, 0xca, 0xd6, 0x1a, 0x22, 0x94, 0x86, 0x16, 0x42, 0xa4, 0xaa, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/table.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/table.pb.go index f9d36875e..7948b07d6 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/table.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/admin/v2/table.pb.go @@ -46,7 +46,7 @@ func (x Table_TimestampGranularity) String() string { return proto.EnumName(Table_TimestampGranularity_name, int32(x)) } func (Table_TimestampGranularity) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{0, 0} + return fileDescriptor_table_dac9f683d605f5c8, []int{0, 0} } // Defines a view over a table's fields. @@ -59,12 +59,6 @@ const ( Table_NAME_ONLY Table_View = 1 // Only populates `name` and fields related to the table's schema. Table_SCHEMA_VIEW Table_View = 2 - // This is a private alpha release of Cloud Bigtable replication. This - // feature is not currently available to most Cloud Bigtable customers. This - // feature might be changed in backward-incompatible ways and is not - // recommended for production use. It is not subject to any SLA or - // deprecation policy. - // // Only populates `name` and fields related to the table's // replication state. Table_REPLICATION_VIEW Table_View = 3 @@ -91,7 +85,7 @@ func (x Table_View) String() string { return proto.EnumName(Table_View_name, int32(x)) } func (Table_View) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{0, 1} + return fileDescriptor_table_dac9f683d605f5c8, []int{0, 1} } // Table replication states. @@ -102,21 +96,17 @@ const ( Table_ClusterState_STATE_NOT_KNOWN Table_ClusterState_ReplicationState = 0 // The cluster was recently created, and the table must finish copying // over pre-existing data from other clusters before it can begin - // receiving live replication updates and serving - // [Data API][google.bigtable.v2.Bigtable] requests. + // receiving live replication updates and serving Data API requests. Table_ClusterState_INITIALIZING Table_ClusterState_ReplicationState = 1 - // The table is temporarily unable to serve - // [Data API][google.bigtable.v2.Bigtable] requests from this + // The table is temporarily unable to serve Data API requests from this // cluster due to planned internal maintenance. Table_ClusterState_PLANNED_MAINTENANCE Table_ClusterState_ReplicationState = 2 - // The table is temporarily unable to serve - // [Data API][google.bigtable.v2.Bigtable] requests from this + // The table is temporarily unable to serve Data API requests from this // cluster due to unplanned or emergency maintenance. Table_ClusterState_UNPLANNED_MAINTENANCE Table_ClusterState_ReplicationState = 3 - // The table can serve - // [Data API][google.bigtable.v2.Bigtable] requests from this - // cluster. Depending on replication delay, reads may not immediately - // reflect the state of the table in other clusters. + // The table can serve Data API requests from this cluster. Depending on + // replication delay, reads may not immediately reflect the state of the + // table in other clusters. Table_ClusterState_READY Table_ClusterState_ReplicationState = 4 ) @@ -139,7 +129,7 @@ func (x Table_ClusterState_ReplicationState) String() string { return proto.EnumName(Table_ClusterState_ReplicationState_name, int32(x)) } func (Table_ClusterState_ReplicationState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{0, 0, 0} + return fileDescriptor_table_dac9f683d605f5c8, []int{0, 0, 0} } // Possible states of a snapshot. @@ -171,7 +161,7 @@ func (x Snapshot_State) String() string { return proto.EnumName(Snapshot_State_name, int32(x)) } func (Snapshot_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{3, 0} + return fileDescriptor_table_dac9f683d605f5c8, []int{3, 0} } // A collection of user data indexed by row, column, and timestamp. @@ -180,30 +170,25 @@ type Table struct { // (`OutputOnly`) // The unique name of the table. Values are of the form // `projects//instances//tables/[_a-zA-Z0-9][-_.a-zA-Z0-9]*`. - // Views: `NAME_ONLY`, `SCHEMA_VIEW`, `FULL` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + // Views: `NAME_ONLY`, `SCHEMA_VIEW`, `REPLICATION_VIEW`, `FULL` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // (`OutputOnly`) // Map from cluster ID to per-cluster table state. // If it could not be determined whether or not the table has data in a // particular cluster (for example, if its zone is unavailable), then // there will be an entry for the cluster with UNKNOWN `replication_status`. - // Views: `FULL` - ClusterStates map[string]*Table_ClusterState `protobuf:"bytes,2,rep,name=cluster_states,json=clusterStates" json:"cluster_states,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + // Views: `REPLICATION_VIEW`, `FULL` + ClusterStates map[string]*Table_ClusterState `protobuf:"bytes,2,rep,name=cluster_states,json=clusterStates,proto3" json:"cluster_states,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // (`CreationOnly`) // The column families configured for this table, mapped by column family ID. // Views: `SCHEMA_VIEW`, `FULL` - ColumnFamilies map[string]*ColumnFamily `protobuf:"bytes,3,rep,name=column_families,json=columnFamilies" json:"column_families,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ColumnFamilies map[string]*ColumnFamily `protobuf:"bytes,3,rep,name=column_families,json=columnFamilies,proto3" json:"column_families,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // (`CreationOnly`) - // The granularity (e.g. `MILLIS`, `MICROS`) at which timestamps are stored in + // The granularity (i.e. `MILLIS`) at which timestamps are stored in // this table. Timestamps not matching the granularity will be rejected. // If unspecified at creation time, the value will be set to `MILLIS`. // Views: `SCHEMA_VIEW`, `FULL` - Granularity Table_TimestampGranularity `protobuf:"varint,4,opt,name=granularity,enum=google.bigtable.admin.v2.Table_TimestampGranularity" json:"granularity,omitempty"` + Granularity Table_TimestampGranularity `protobuf:"varint,4,opt,name=granularity,proto3,enum=google.bigtable.admin.v2.Table_TimestampGranularity" json:"granularity,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -213,7 +198,7 @@ func (m *Table) Reset() { *m = Table{} } func (m *Table) String() string { return proto.CompactTextString(m) } func (*Table) ProtoMessage() {} func (*Table) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{0} + return fileDescriptor_table_dac9f683d605f5c8, []int{0} } func (m *Table) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Table.Unmarshal(m, b) @@ -261,16 +246,11 @@ func (m *Table) GetGranularity() Table_TimestampGranularity { return Table_TIMESTAMP_GRANULARITY_UNSPECIFIED } -// This is a private alpha release of Cloud Bigtable replication. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // The state of a table's data in a particular cluster. type Table_ClusterState struct { // (`OutputOnly`) // The state of replication for the table in this cluster. - ReplicationState Table_ClusterState_ReplicationState `protobuf:"varint,1,opt,name=replication_state,json=replicationState,enum=google.bigtable.admin.v2.Table_ClusterState_ReplicationState" json:"replication_state,omitempty"` + ReplicationState Table_ClusterState_ReplicationState `protobuf:"varint,1,opt,name=replication_state,json=replicationState,proto3,enum=google.bigtable.admin.v2.Table_ClusterState_ReplicationState" json:"replication_state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -280,7 +260,7 @@ func (m *Table_ClusterState) Reset() { *m = Table_ClusterState{} } func (m *Table_ClusterState) String() string { return proto.CompactTextString(m) } func (*Table_ClusterState) ProtoMessage() {} func (*Table_ClusterState) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{0, 0} + return fileDescriptor_table_dac9f683d605f5c8, []int{0, 0} } func (m *Table_ClusterState) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Table_ClusterState.Unmarshal(m, b) @@ -315,7 +295,7 @@ type ColumnFamily struct { // NOTE: Garbage collection executes opportunistically in the background, and // so it's possible for reads to return a cell even if it matches the active // GC expression for its family. - GcRule *GcRule `protobuf:"bytes,1,opt,name=gc_rule,json=gcRule" json:"gc_rule,omitempty"` + GcRule *GcRule `protobuf:"bytes,1,opt,name=gc_rule,json=gcRule,proto3" json:"gc_rule,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -325,7 +305,7 @@ func (m *ColumnFamily) Reset() { *m = ColumnFamily{} } func (m *ColumnFamily) String() string { return proto.CompactTextString(m) } func (*ColumnFamily) ProtoMessage() {} func (*ColumnFamily) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{1} + return fileDescriptor_table_dac9f683d605f5c8, []int{1} } func (m *ColumnFamily) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ColumnFamily.Unmarshal(m, b) @@ -371,7 +351,7 @@ func (m *GcRule) Reset() { *m = GcRule{} } func (m *GcRule) String() string { return proto.CompactTextString(m) } func (*GcRule) ProtoMessage() {} func (*GcRule) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{2} + return fileDescriptor_table_dac9f683d605f5c8, []int{2} } func (m *GcRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GcRule.Unmarshal(m, b) @@ -396,16 +376,16 @@ type isGcRule_Rule interface { } type GcRule_MaxNumVersions struct { - MaxNumVersions int32 `protobuf:"varint,1,opt,name=max_num_versions,json=maxNumVersions,oneof"` + MaxNumVersions int32 `protobuf:"varint,1,opt,name=max_num_versions,json=maxNumVersions,proto3,oneof"` } type GcRule_MaxAge struct { - MaxAge *duration.Duration `protobuf:"bytes,2,opt,name=max_age,json=maxAge,oneof"` + MaxAge *duration.Duration `protobuf:"bytes,2,opt,name=max_age,json=maxAge,proto3,oneof"` } type GcRule_Intersection_ struct { - Intersection *GcRule_Intersection `protobuf:"bytes,3,opt,name=intersection,oneof"` + Intersection *GcRule_Intersection `protobuf:"bytes,3,opt,name=intersection,proto3,oneof"` } type GcRule_Union_ struct { - Union *GcRule_Union `protobuf:"bytes,4,opt,name=union,oneof"` + Union *GcRule_Union `protobuf:"bytes,4,opt,name=union,proto3,oneof"` } func (*GcRule_MaxNumVersions) isGcRule_Rule() {} @@ -558,7 +538,7 @@ func _GcRule_OneofSizer(msg proto.Message) (n int) { // A GcRule which deletes cells matching all of the given rules. type GcRule_Intersection struct { // Only delete cells which would be deleted by every element of `rules`. - Rules []*GcRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*GcRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -568,7 +548,7 @@ func (m *GcRule_Intersection) Reset() { *m = GcRule_Intersection{} } func (m *GcRule_Intersection) String() string { return proto.CompactTextString(m) } func (*GcRule_Intersection) ProtoMessage() {} func (*GcRule_Intersection) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{2, 0} + return fileDescriptor_table_dac9f683d605f5c8, []int{2, 0} } func (m *GcRule_Intersection) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GcRule_Intersection.Unmarshal(m, b) @@ -598,7 +578,7 @@ func (m *GcRule_Intersection) GetRules() []*GcRule { // A GcRule which deletes cells matching any of the given rules. type GcRule_Union struct { // Delete cells which would be deleted by any element of `rules`. - Rules []*GcRule `protobuf:"bytes,1,rep,name=rules" json:"rules,omitempty"` + Rules []*GcRule `protobuf:"bytes,1,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -608,7 +588,7 @@ func (m *GcRule_Union) Reset() { *m = GcRule_Union{} } func (m *GcRule_Union) String() string { return proto.CompactTextString(m) } func (*GcRule_Union) ProtoMessage() {} func (*GcRule_Union) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{2, 1} + return fileDescriptor_table_dac9f683d605f5c8, []int{2, 1} } func (m *GcRule_Union) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GcRule_Union.Unmarshal(m, b) @@ -635,41 +615,41 @@ func (m *GcRule_Union) GetRules() []*GcRule { return nil } -// This is a private alpha release of Cloud Bigtable snapshots. This feature -// is not currently available to most Cloud Bigtable customers. This feature -// might be changed in backward-incompatible ways and is not recommended for -// production use. It is not subject to any SLA or deprecation policy. -// // A snapshot of a table at a particular time. A snapshot can be used as a // checkpoint for data restoration or a data source for a new table. +// +// Note: This is a private alpha release of Cloud Bigtable snapshots. This +// feature is not currently available to most Cloud Bigtable customers. This +// feature might be changed in backward-incompatible ways and is not recommended +// for production use. It is not subject to any SLA or deprecation policy. type Snapshot struct { // (`OutputOnly`) // The unique name of the snapshot. // Values are of the form // `projects//instances//clusters//snapshots/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // (`OutputOnly`) // The source table at the time the snapshot was taken. - SourceTable *Table `protobuf:"bytes,2,opt,name=source_table,json=sourceTable" json:"source_table,omitempty"` + SourceTable *Table `protobuf:"bytes,2,opt,name=source_table,json=sourceTable,proto3" json:"source_table,omitempty"` // (`OutputOnly`) // The size of the data in the source table at the time the snapshot was // taken. In some cases, this value may be computed asynchronously via a // background process and a placeholder of 0 will be used in the meantime. - DataSizeBytes int64 `protobuf:"varint,3,opt,name=data_size_bytes,json=dataSizeBytes" json:"data_size_bytes,omitempty"` + DataSizeBytes int64 `protobuf:"varint,3,opt,name=data_size_bytes,json=dataSizeBytes,proto3" json:"data_size_bytes,omitempty"` // (`OutputOnly`) // The time when the snapshot is created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // (`OutputOnly`) // The time when the snapshot will be deleted. The maximum amount of time a // snapshot can stay active is 365 days. If 'ttl' is not specified, // the default maximum of 365 days will be used. - DeleteTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=delete_time,json=deleteTime" json:"delete_time,omitempty"` + DeleteTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=delete_time,json=deleteTime,proto3" json:"delete_time,omitempty"` // (`OutputOnly`) // The current state of the snapshot. - State Snapshot_State `protobuf:"varint,6,opt,name=state,enum=google.bigtable.admin.v2.Snapshot_State" json:"state,omitempty"` + State Snapshot_State `protobuf:"varint,6,opt,name=state,proto3,enum=google.bigtable.admin.v2.Snapshot_State" json:"state,omitempty"` // (`OutputOnly`) // Description of the snapshot. - Description string `protobuf:"bytes,7,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,7,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -679,7 +659,7 @@ func (m *Snapshot) Reset() { *m = Snapshot{} } func (m *Snapshot) String() string { return proto.CompactTextString(m) } func (*Snapshot) ProtoMessage() {} func (*Snapshot) Descriptor() ([]byte, []int) { - return fileDescriptor_table_5c3da779706b03a9, []int{3} + return fileDescriptor_table_dac9f683d605f5c8, []int{3} } func (m *Snapshot) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Snapshot.Unmarshal(m, b) @@ -765,10 +745,10 @@ func init() { } func init() { - proto.RegisterFile("google/bigtable/admin/v2/table.proto", fileDescriptor_table_5c3da779706b03a9) + proto.RegisterFile("google/bigtable/admin/v2/table.proto", fileDescriptor_table_dac9f683d605f5c8) } -var fileDescriptor_table_5c3da779706b03a9 = []byte{ +var fileDescriptor_table_dac9f683d605f5c8 = []byte{ // 965 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x56, 0xff, 0x6e, 0xdb, 0x54, 0x18, 0xad, 0xe3, 0x38, 0x6d, 0xbf, 0xa4, 0xad, 0xb9, 0x1d, 0x22, 0x8b, 0xa6, 0x2d, 0x44, 0x30, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_data.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_data.pb.go index 321b450c9..67e48f931 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_data.pb.go @@ -27,7 +27,7 @@ type Row struct { Key []byte `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // May be empty, but only if the entire row is empty. // The mutual ordering of column families is not specified. - Families []*Family `protobuf:"bytes,2,rep,name=families" json:"families,omitempty"` + Families []*Family `protobuf:"bytes,2,rep,name=families,proto3" json:"families,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -37,7 +37,7 @@ func (m *Row) Reset() { *m = Row{} } func (m *Row) String() string { return proto.CompactTextString(m) } func (*Row) ProtoMessage() {} func (*Row) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{0} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{0} } func (m *Row) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Row.Unmarshal(m, b) @@ -79,9 +79,9 @@ type Family struct { // Must match [-_.a-zA-Z0-9]+, except that AggregatingRowProcessors may // produce cells in a sentinel family with an empty name. // Must be no greater than 64 characters in length. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Must not be empty. Sorted in order of increasing "qualifier". - Columns []*Column `protobuf:"bytes,2,rep,name=columns" json:"columns,omitempty"` + Columns []*Column `protobuf:"bytes,2,rep,name=columns,proto3" json:"columns,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -91,7 +91,7 @@ func (m *Family) Reset() { *m = Family{} } func (m *Family) String() string { return proto.CompactTextString(m) } func (*Family) ProtoMessage() {} func (*Family) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{1} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{1} } func (m *Family) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Family.Unmarshal(m, b) @@ -134,7 +134,7 @@ type Column struct { // length. Qualifier []byte `protobuf:"bytes,1,opt,name=qualifier,proto3" json:"qualifier,omitempty"` // Must not be empty. Sorted in order of decreasing "timestamp_micros". - Cells []*Cell `protobuf:"bytes,2,rep,name=cells" json:"cells,omitempty"` + Cells []*Cell `protobuf:"bytes,2,rep,name=cells,proto3" json:"cells,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -144,7 +144,7 @@ func (m *Column) Reset() { *m = Column{} } func (m *Column) String() string { return proto.CompactTextString(m) } func (*Column) ProtoMessage() {} func (*Column) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{2} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{2} } func (m *Column) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Column.Unmarshal(m, b) @@ -186,13 +186,13 @@ type Cell struct { // a coarser "granularity" to further restrict the allowed values. For // example, a table which specifies millisecond granularity will only allow // values of "timestamp_micros" which are multiples of 1000. - TimestampMicros int64 `protobuf:"varint,1,opt,name=timestamp_micros,json=timestampMicros" json:"timestamp_micros,omitempty"` + TimestampMicros int64 `protobuf:"varint,1,opt,name=timestamp_micros,json=timestampMicros,proto3" json:"timestamp_micros,omitempty"` // The value stored in the cell. // May contain any byte string, including the empty string, up to 100MiB in // length. Value []byte `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Labels applied to the cell by a [RowFilter][google.bigtable.v1.RowFilter]. - Labels []string `protobuf:"bytes,3,rep,name=labels" json:"labels,omitempty"` + Labels []string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -202,7 +202,7 @@ func (m *Cell) Reset() { *m = Cell{} } func (m *Cell) String() string { return proto.CompactTextString(m) } func (*Cell) ProtoMessage() {} func (*Cell) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{3} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{3} } func (m *Cell) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Cell.Unmarshal(m, b) @@ -258,7 +258,7 @@ func (m *RowRange) Reset() { *m = RowRange{} } func (m *RowRange) String() string { return proto.CompactTextString(m) } func (*RowRange) ProtoMessage() {} func (*RowRange) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{4} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{4} } func (m *RowRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowRange.Unmarshal(m, b) @@ -297,7 +297,7 @@ type RowSet struct { // Single rows included in the set. RowKeys [][]byte `protobuf:"bytes,1,rep,name=row_keys,json=rowKeys,proto3" json:"row_keys,omitempty"` // Contiguous row ranges included in the set. - RowRanges []*RowRange `protobuf:"bytes,2,rep,name=row_ranges,json=rowRanges" json:"row_ranges,omitempty"` + RowRanges []*RowRange `protobuf:"bytes,2,rep,name=row_ranges,json=rowRanges,proto3" json:"row_ranges,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -307,7 +307,7 @@ func (m *RowSet) Reset() { *m = RowSet{} } func (m *RowSet) String() string { return proto.CompactTextString(m) } func (*RowSet) ProtoMessage() {} func (*RowSet) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{5} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{5} } func (m *RowSet) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowSet.Unmarshal(m, b) @@ -347,7 +347,7 @@ func (m *RowSet) GetRowRanges() []*RowRange { // exclusive. type ColumnRange struct { // The name of the column family within which this range falls. - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The column qualifier at which to start the range (within 'column_family'). // If neither field is set, interpreted as the empty string, inclusive. // @@ -371,7 +371,7 @@ func (m *ColumnRange) Reset() { *m = ColumnRange{} } func (m *ColumnRange) String() string { return proto.CompactTextString(m) } func (*ColumnRange) ProtoMessage() {} func (*ColumnRange) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{6} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{6} } func (m *ColumnRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ColumnRange.Unmarshal(m, b) @@ -575,9 +575,9 @@ func _ColumnRange_OneofSizer(msg proto.Message) (n int) { // Specified a contiguous range of microsecond timestamps. type TimestampRange struct { // Inclusive lower bound. If left empty, interpreted as 0. - StartTimestampMicros int64 `protobuf:"varint,1,opt,name=start_timestamp_micros,json=startTimestampMicros" json:"start_timestamp_micros,omitempty"` + StartTimestampMicros int64 `protobuf:"varint,1,opt,name=start_timestamp_micros,json=startTimestampMicros,proto3" json:"start_timestamp_micros,omitempty"` // Exclusive upper bound. If left empty, interpreted as infinity. - EndTimestampMicros int64 `protobuf:"varint,2,opt,name=end_timestamp_micros,json=endTimestampMicros" json:"end_timestamp_micros,omitempty"` + EndTimestampMicros int64 `protobuf:"varint,2,opt,name=end_timestamp_micros,json=endTimestampMicros,proto3" json:"end_timestamp_micros,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -587,7 +587,7 @@ func (m *TimestampRange) Reset() { *m = TimestampRange{} } func (m *TimestampRange) String() string { return proto.CompactTextString(m) } func (*TimestampRange) ProtoMessage() {} func (*TimestampRange) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{7} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{7} } func (m *TimestampRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TimestampRange.Unmarshal(m, b) @@ -646,7 +646,7 @@ func (m *ValueRange) Reset() { *m = ValueRange{} } func (m *ValueRange) String() string { return proto.CompactTextString(m) } func (*ValueRange) ProtoMessage() {} func (*ValueRange) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{8} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{8} } func (m *ValueRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ValueRange.Unmarshal(m, b) @@ -907,7 +907,7 @@ func (m *RowFilter) Reset() { *m = RowFilter{} } func (m *RowFilter) String() string { return proto.CompactTextString(m) } func (*RowFilter) ProtoMessage() {} func (*RowFilter) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{9} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{9} } func (m *RowFilter) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter.Unmarshal(m, b) @@ -932,61 +932,61 @@ type isRowFilter_Filter interface { } type RowFilter_Chain_ struct { - Chain *RowFilter_Chain `protobuf:"bytes,1,opt,name=chain,oneof"` + Chain *RowFilter_Chain `protobuf:"bytes,1,opt,name=chain,proto3,oneof"` } type RowFilter_Interleave_ struct { - Interleave *RowFilter_Interleave `protobuf:"bytes,2,opt,name=interleave,oneof"` + Interleave *RowFilter_Interleave `protobuf:"bytes,2,opt,name=interleave,proto3,oneof"` } type RowFilter_Condition_ struct { - Condition *RowFilter_Condition `protobuf:"bytes,3,opt,name=condition,oneof"` + Condition *RowFilter_Condition `protobuf:"bytes,3,opt,name=condition,proto3,oneof"` } type RowFilter_Sink struct { - Sink bool `protobuf:"varint,16,opt,name=sink,oneof"` + Sink bool `protobuf:"varint,16,opt,name=sink,proto3,oneof"` } type RowFilter_PassAllFilter struct { - PassAllFilter bool `protobuf:"varint,17,opt,name=pass_all_filter,json=passAllFilter,oneof"` + PassAllFilter bool `protobuf:"varint,17,opt,name=pass_all_filter,json=passAllFilter,proto3,oneof"` } type RowFilter_BlockAllFilter struct { - BlockAllFilter bool `protobuf:"varint,18,opt,name=block_all_filter,json=blockAllFilter,oneof"` + BlockAllFilter bool `protobuf:"varint,18,opt,name=block_all_filter,json=blockAllFilter,proto3,oneof"` } type RowFilter_RowKeyRegexFilter struct { RowKeyRegexFilter []byte `protobuf:"bytes,4,opt,name=row_key_regex_filter,json=rowKeyRegexFilter,proto3,oneof"` } type RowFilter_RowSampleFilter struct { - RowSampleFilter float64 `protobuf:"fixed64,14,opt,name=row_sample_filter,json=rowSampleFilter,oneof"` + RowSampleFilter float64 `protobuf:"fixed64,14,opt,name=row_sample_filter,json=rowSampleFilter,proto3,oneof"` } type RowFilter_FamilyNameRegexFilter struct { - FamilyNameRegexFilter string `protobuf:"bytes,5,opt,name=family_name_regex_filter,json=familyNameRegexFilter,oneof"` + FamilyNameRegexFilter string `protobuf:"bytes,5,opt,name=family_name_regex_filter,json=familyNameRegexFilter,proto3,oneof"` } type RowFilter_ColumnQualifierRegexFilter struct { ColumnQualifierRegexFilter []byte `protobuf:"bytes,6,opt,name=column_qualifier_regex_filter,json=columnQualifierRegexFilter,proto3,oneof"` } type RowFilter_ColumnRangeFilter struct { - ColumnRangeFilter *ColumnRange `protobuf:"bytes,7,opt,name=column_range_filter,json=columnRangeFilter,oneof"` + ColumnRangeFilter *ColumnRange `protobuf:"bytes,7,opt,name=column_range_filter,json=columnRangeFilter,proto3,oneof"` } type RowFilter_TimestampRangeFilter struct { - TimestampRangeFilter *TimestampRange `protobuf:"bytes,8,opt,name=timestamp_range_filter,json=timestampRangeFilter,oneof"` + TimestampRangeFilter *TimestampRange `protobuf:"bytes,8,opt,name=timestamp_range_filter,json=timestampRangeFilter,proto3,oneof"` } type RowFilter_ValueRegexFilter struct { ValueRegexFilter []byte `protobuf:"bytes,9,opt,name=value_regex_filter,json=valueRegexFilter,proto3,oneof"` } type RowFilter_ValueRangeFilter struct { - ValueRangeFilter *ValueRange `protobuf:"bytes,15,opt,name=value_range_filter,json=valueRangeFilter,oneof"` + ValueRangeFilter *ValueRange `protobuf:"bytes,15,opt,name=value_range_filter,json=valueRangeFilter,proto3,oneof"` } type RowFilter_CellsPerRowOffsetFilter struct { - CellsPerRowOffsetFilter int32 `protobuf:"varint,10,opt,name=cells_per_row_offset_filter,json=cellsPerRowOffsetFilter,oneof"` + CellsPerRowOffsetFilter int32 `protobuf:"varint,10,opt,name=cells_per_row_offset_filter,json=cellsPerRowOffsetFilter,proto3,oneof"` } type RowFilter_CellsPerRowLimitFilter struct { - CellsPerRowLimitFilter int32 `protobuf:"varint,11,opt,name=cells_per_row_limit_filter,json=cellsPerRowLimitFilter,oneof"` + CellsPerRowLimitFilter int32 `protobuf:"varint,11,opt,name=cells_per_row_limit_filter,json=cellsPerRowLimitFilter,proto3,oneof"` } type RowFilter_CellsPerColumnLimitFilter struct { - CellsPerColumnLimitFilter int32 `protobuf:"varint,12,opt,name=cells_per_column_limit_filter,json=cellsPerColumnLimitFilter,oneof"` + CellsPerColumnLimitFilter int32 `protobuf:"varint,12,opt,name=cells_per_column_limit_filter,json=cellsPerColumnLimitFilter,proto3,oneof"` } type RowFilter_StripValueTransformer struct { - StripValueTransformer bool `protobuf:"varint,13,opt,name=strip_value_transformer,json=stripValueTransformer,oneof"` + StripValueTransformer bool `protobuf:"varint,13,opt,name=strip_value_transformer,json=stripValueTransformer,proto3,oneof"` } type RowFilter_ApplyLabelTransformer struct { - ApplyLabelTransformer string `protobuf:"bytes,19,opt,name=apply_label_transformer,json=applyLabelTransformer,oneof"` + ApplyLabelTransformer string `protobuf:"bytes,19,opt,name=apply_label_transformer,json=applyLabelTransformer,proto3,oneof"` } func (*RowFilter_Chain_) isRowFilter_Filter() {} @@ -1507,7 +1507,7 @@ type RowFilter_Chain struct { // The elements of "filters" are chained together to process the input row: // in row -> f(0) -> intermediate row -> f(1) -> ... -> f(N) -> out row // The full chain is executed atomically. - Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters" json:"filters,omitempty"` + Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1517,7 +1517,7 @@ func (m *RowFilter_Chain) Reset() { *m = RowFilter_Chain{} } func (m *RowFilter_Chain) String() string { return proto.CompactTextString(m) } func (*RowFilter_Chain) ProtoMessage() {} func (*RowFilter_Chain) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{9, 0} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{9, 0} } func (m *RowFilter_Chain) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter_Chain.Unmarshal(m, b) @@ -1571,7 +1571,7 @@ type RowFilter_Interleave struct { // 5: far,blah,5,x // identical to #6 // 6: far,blah,5,x // identical to #5 // All interleaved filters are executed atomically. - Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters" json:"filters,omitempty"` + Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1581,7 +1581,7 @@ func (m *RowFilter_Interleave) Reset() { *m = RowFilter_Interleave{} } func (m *RowFilter_Interleave) String() string { return proto.CompactTextString(m) } func (*RowFilter_Interleave) ProtoMessage() {} func (*RowFilter_Interleave) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{9, 1} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{9, 1} } func (m *RowFilter_Interleave) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter_Interleave.Unmarshal(m, b) @@ -1618,14 +1618,14 @@ func (m *RowFilter_Interleave) GetFilters() []*RowFilter { type RowFilter_Condition struct { // If "predicate_filter" outputs any cells, then "true_filter" will be // evaluated on the input row. Otherwise, "false_filter" will be evaluated. - PredicateFilter *RowFilter `protobuf:"bytes,1,opt,name=predicate_filter,json=predicateFilter" json:"predicate_filter,omitempty"` + PredicateFilter *RowFilter `protobuf:"bytes,1,opt,name=predicate_filter,json=predicateFilter,proto3" json:"predicate_filter,omitempty"` // The filter to apply to the input row if "predicate_filter" returns any // results. If not provided, no results will be returned in the true case. - TrueFilter *RowFilter `protobuf:"bytes,2,opt,name=true_filter,json=trueFilter" json:"true_filter,omitempty"` + TrueFilter *RowFilter `protobuf:"bytes,2,opt,name=true_filter,json=trueFilter,proto3" json:"true_filter,omitempty"` // The filter to apply to the input row if "predicate_filter" does not // return any results. If not provided, no results will be returned in the // false case. - FalseFilter *RowFilter `protobuf:"bytes,3,opt,name=false_filter,json=falseFilter" json:"false_filter,omitempty"` + FalseFilter *RowFilter `protobuf:"bytes,3,opt,name=false_filter,json=falseFilter,proto3" json:"false_filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1635,7 +1635,7 @@ func (m *RowFilter_Condition) Reset() { *m = RowFilter_Condition{} } func (m *RowFilter_Condition) String() string { return proto.CompactTextString(m) } func (*RowFilter_Condition) ProtoMessage() {} func (*RowFilter_Condition) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{9, 2} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{9, 2} } func (m *RowFilter_Condition) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter_Condition.Unmarshal(m, b) @@ -1695,7 +1695,7 @@ func (m *Mutation) Reset() { *m = Mutation{} } func (m *Mutation) String() string { return proto.CompactTextString(m) } func (*Mutation) ProtoMessage() {} func (*Mutation) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{10} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{10} } func (m *Mutation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation.Unmarshal(m, b) @@ -1720,16 +1720,16 @@ type isMutation_Mutation interface { } type Mutation_SetCell_ struct { - SetCell *Mutation_SetCell `protobuf:"bytes,1,opt,name=set_cell,json=setCell,oneof"` + SetCell *Mutation_SetCell `protobuf:"bytes,1,opt,name=set_cell,json=setCell,proto3,oneof"` } type Mutation_DeleteFromColumn_ struct { - DeleteFromColumn *Mutation_DeleteFromColumn `protobuf:"bytes,2,opt,name=delete_from_column,json=deleteFromColumn,oneof"` + DeleteFromColumn *Mutation_DeleteFromColumn `protobuf:"bytes,2,opt,name=delete_from_column,json=deleteFromColumn,proto3,oneof"` } type Mutation_DeleteFromFamily_ struct { - DeleteFromFamily *Mutation_DeleteFromFamily `protobuf:"bytes,3,opt,name=delete_from_family,json=deleteFromFamily,oneof"` + DeleteFromFamily *Mutation_DeleteFromFamily `protobuf:"bytes,3,opt,name=delete_from_family,json=deleteFromFamily,proto3,oneof"` } type Mutation_DeleteFromRow_ struct { - DeleteFromRow *Mutation_DeleteFromRow `protobuf:"bytes,4,opt,name=delete_from_row,json=deleteFromRow,oneof"` + DeleteFromRow *Mutation_DeleteFromRow `protobuf:"bytes,4,opt,name=delete_from_row,json=deleteFromRow,proto3,oneof"` } func (*Mutation_SetCell_) isMutation_Mutation() {} @@ -1888,7 +1888,7 @@ func _Mutation_OneofSizer(msg proto.Message) (n int) { type Mutation_SetCell struct { // The name of the family into which new data should be written. // Must match [-_.a-zA-Z0-9]+ - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The qualifier of the column into which new data should be written. // Can be any byte string, including the empty string. ColumnQualifier []byte `protobuf:"bytes,2,opt,name=column_qualifier,json=columnQualifier,proto3" json:"column_qualifier,omitempty"` @@ -1897,7 +1897,7 @@ type Mutation_SetCell struct { // Otherwise, the client should set this value itself, noting that the // default value is a timestamp of zero if the field is left unspecified. // Values must match the "granularity" of the table (e.g. micros, millis). - TimestampMicros int64 `protobuf:"varint,3,opt,name=timestamp_micros,json=timestampMicros" json:"timestamp_micros,omitempty"` + TimestampMicros int64 `protobuf:"varint,3,opt,name=timestamp_micros,json=timestampMicros,proto3" json:"timestamp_micros,omitempty"` // The value to be written into the specified cell. Value []byte `protobuf:"bytes,4,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1909,7 +1909,7 @@ func (m *Mutation_SetCell) Reset() { *m = Mutation_SetCell{} } func (m *Mutation_SetCell) String() string { return proto.CompactTextString(m) } func (*Mutation_SetCell) ProtoMessage() {} func (*Mutation_SetCell) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{10, 0} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{10, 0} } func (m *Mutation_SetCell) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_SetCell.Unmarshal(m, b) @@ -1962,12 +1962,12 @@ func (m *Mutation_SetCell) GetValue() []byte { type Mutation_DeleteFromColumn struct { // The name of the family from which cells should be deleted. // Must match [-_.a-zA-Z0-9]+ - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The qualifier of the column from which cells should be deleted. // Can be any byte string, including the empty string. ColumnQualifier []byte `protobuf:"bytes,2,opt,name=column_qualifier,json=columnQualifier,proto3" json:"column_qualifier,omitempty"` // The range of timestamps within which cells should be deleted. - TimeRange *TimestampRange `protobuf:"bytes,3,opt,name=time_range,json=timeRange" json:"time_range,omitempty"` + TimeRange *TimestampRange `protobuf:"bytes,3,opt,name=time_range,json=timeRange,proto3" json:"time_range,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1977,7 +1977,7 @@ func (m *Mutation_DeleteFromColumn) Reset() { *m = Mutation_DeleteFromCo func (m *Mutation_DeleteFromColumn) String() string { return proto.CompactTextString(m) } func (*Mutation_DeleteFromColumn) ProtoMessage() {} func (*Mutation_DeleteFromColumn) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{10, 1} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{10, 1} } func (m *Mutation_DeleteFromColumn) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_DeleteFromColumn.Unmarshal(m, b) @@ -2022,7 +2022,7 @@ func (m *Mutation_DeleteFromColumn) GetTimeRange() *TimestampRange { type Mutation_DeleteFromFamily struct { // The name of the family from which cells should be deleted. // Must match [-_.a-zA-Z0-9]+ - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2032,7 +2032,7 @@ func (m *Mutation_DeleteFromFamily) Reset() { *m = Mutation_DeleteFromFa func (m *Mutation_DeleteFromFamily) String() string { return proto.CompactTextString(m) } func (*Mutation_DeleteFromFamily) ProtoMessage() {} func (*Mutation_DeleteFromFamily) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{10, 2} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{10, 2} } func (m *Mutation_DeleteFromFamily) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_DeleteFromFamily.Unmarshal(m, b) @@ -2070,7 +2070,7 @@ func (m *Mutation_DeleteFromRow) Reset() { *m = Mutation_DeleteFromRow{} func (m *Mutation_DeleteFromRow) String() string { return proto.CompactTextString(m) } func (*Mutation_DeleteFromRow) ProtoMessage() {} func (*Mutation_DeleteFromRow) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{10, 3} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{10, 3} } func (m *Mutation_DeleteFromRow) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_DeleteFromRow.Unmarshal(m, b) @@ -2095,7 +2095,7 @@ var xxx_messageInfo_Mutation_DeleteFromRow proto.InternalMessageInfo type ReadModifyWriteRule struct { // The name of the family to which the read/modify/write should be applied. // Must match [-_.a-zA-Z0-9]+ - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The qualifier of the column to which the read/modify/write should be // applied. // Can be any byte string, including the empty string. @@ -2116,7 +2116,7 @@ func (m *ReadModifyWriteRule) Reset() { *m = ReadModifyWriteRule{} } func (m *ReadModifyWriteRule) String() string { return proto.CompactTextString(m) } func (*ReadModifyWriteRule) ProtoMessage() {} func (*ReadModifyWriteRule) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_data_d602162366ce13f1, []int{11} + return fileDescriptor_bigtable_data_11f19b94ec945589, []int{11} } func (m *ReadModifyWriteRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadModifyWriteRule.Unmarshal(m, b) @@ -2144,7 +2144,7 @@ type ReadModifyWriteRule_AppendValue struct { AppendValue []byte `protobuf:"bytes,3,opt,name=append_value,json=appendValue,proto3,oneof"` } type ReadModifyWriteRule_IncrementAmount struct { - IncrementAmount int64 `protobuf:"varint,4,opt,name=increment_amount,json=incrementAmount,oneof"` + IncrementAmount int64 `protobuf:"varint,4,opt,name=increment_amount,json=incrementAmount,proto3,oneof"` } func (*ReadModifyWriteRule_AppendValue) isReadModifyWriteRule_Rule() {} @@ -2273,10 +2273,10 @@ func init() { } func init() { - proto.RegisterFile("google/bigtable/v1/bigtable_data.proto", fileDescriptor_bigtable_data_d602162366ce13f1) + proto.RegisterFile("google/bigtable/v1/bigtable_data.proto", fileDescriptor_bigtable_data_11f19b94ec945589) } -var fileDescriptor_bigtable_data_d602162366ce13f1 = []byte{ +var fileDescriptor_bigtable_data_11f19b94ec945589 = []byte{ // 1378 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x57, 0xdb, 0x6e, 0x1b, 0x37, 0x13, 0xd6, 0x5a, 0xb2, 0x0e, 0xb3, 0x76, 0x24, 0xd3, 0x27, 0x45, 0x49, 0xfe, 0x18, 0xca, 0x8f, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service.pb.go index 395b54d76..de642e347 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service.pb.go @@ -171,8 +171,7 @@ func (c *bigtableServiceClient) ReadModifyWriteRow(ctx context.Context, in *Read return out, nil } -// Server API for BigtableService service - +// BigtableServiceServer is the server API for BigtableService service. type BigtableServiceServer interface { // Streams back the contents of all requested rows, optionally applying // the same Reader filter to each. Depending on their size, rows may be @@ -355,10 +354,10 @@ var _BigtableService_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/bigtable/v1/bigtable_service.proto", fileDescriptor_bigtable_service_e46f53b56ecd7201) + proto.RegisterFile("google/bigtable/v1/bigtable_service.proto", fileDescriptor_bigtable_service_4233bee4d874237b) } -var fileDescriptor_bigtable_service_e46f53b56ecd7201 = []byte{ +var fileDescriptor_bigtable_service_4233bee4d874237b = []byte{ // 521 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x94, 0xcd, 0x6e, 0xd4, 0x30, 0x10, 0xc7, 0x65, 0x0e, 0xa8, 0x58, 0x42, 0x08, 0x4b, 0x14, 0x69, 0xe1, 0x14, 0xa0, 0xa2, 0x11, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service_messages.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service_messages.pb.go index 208683fb9..43888719f 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service_messages.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/v1/bigtable_service_messages.pb.go @@ -22,7 +22,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Request message for BigtableServer.ReadRows. type ReadRowsRequest struct { // The unique name of the table from which to read. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // If neither row_key nor row_range is set, reads from all rows. // // Types that are valid to be assigned to Target: @@ -32,18 +32,18 @@ type ReadRowsRequest struct { Target isReadRowsRequest_Target `protobuf_oneof:"target"` // The filter to apply to the contents of the specified row(s). If unset, // reads the entire table. - Filter *RowFilter `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter *RowFilter `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // By default, rows are read sequentially, producing results which are // guaranteed to arrive in increasing row order. Setting // "allow_row_interleaving" to true allows multiple rows to be interleaved in // the response stream, which increases throughput but breaks this guarantee, // and may force the client to use more memory to buffer partially-received // rows. Cannot be set to true when specifying "num_rows_limit". - AllowRowInterleaving bool `protobuf:"varint,6,opt,name=allow_row_interleaving,json=allowRowInterleaving" json:"allow_row_interleaving,omitempty"` + AllowRowInterleaving bool `protobuf:"varint,6,opt,name=allow_row_interleaving,json=allowRowInterleaving,proto3" json:"allow_row_interleaving,omitempty"` // The read will terminate after committing to N rows' worth of results. The // default (zero) is to return all results. // Note that "allow_row_interleaving" cannot be set to true when this is set. - NumRowsLimit int64 `protobuf:"varint,7,opt,name=num_rows_limit,json=numRowsLimit" json:"num_rows_limit,omitempty"` + NumRowsLimit int64 `protobuf:"varint,7,opt,name=num_rows_limit,json=numRowsLimit,proto3" json:"num_rows_limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -53,7 +53,7 @@ func (m *ReadRowsRequest) Reset() { *m = ReadRowsRequest{} } func (m *ReadRowsRequest) String() string { return proto.CompactTextString(m) } func (*ReadRowsRequest) ProtoMessage() {} func (*ReadRowsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{0} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{0} } func (m *ReadRowsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadRowsRequest.Unmarshal(m, b) @@ -81,10 +81,10 @@ type ReadRowsRequest_RowKey struct { RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3,oneof"` } type ReadRowsRequest_RowRange struct { - RowRange *RowRange `protobuf:"bytes,3,opt,name=row_range,json=rowRange,oneof"` + RowRange *RowRange `protobuf:"bytes,3,opt,name=row_range,json=rowRange,proto3,oneof"` } type ReadRowsRequest_RowSet struct { - RowSet *RowSet `protobuf:"bytes,8,opt,name=row_set,json=rowSet,oneof"` + RowSet *RowSet `protobuf:"bytes,8,opt,name=row_set,json=rowSet,proto3,oneof"` } func (*ReadRowsRequest_RowKey) isReadRowsRequest_Target() {} @@ -243,7 +243,7 @@ type ReadRowsResponse struct { // "allow_row_interleaving" was specified in the request. RowKey []byte `protobuf:"bytes,1,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // One or more chunks of the row specified by "row_key". - Chunks []*ReadRowsResponse_Chunk `protobuf:"bytes,2,rep,name=chunks" json:"chunks,omitempty"` + Chunks []*ReadRowsResponse_Chunk `protobuf:"bytes,2,rep,name=chunks,proto3" json:"chunks,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -253,7 +253,7 @@ func (m *ReadRowsResponse) Reset() { *m = ReadRowsResponse{} } func (m *ReadRowsResponse) String() string { return proto.CompactTextString(m) } func (*ReadRowsResponse) ProtoMessage() {} func (*ReadRowsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{1} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{1} } func (m *ReadRowsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadRowsResponse.Unmarshal(m, b) @@ -304,7 +304,7 @@ func (m *ReadRowsResponse_Chunk) Reset() { *m = ReadRowsResponse_Chunk{} func (m *ReadRowsResponse_Chunk) String() string { return proto.CompactTextString(m) } func (*ReadRowsResponse_Chunk) ProtoMessage() {} func (*ReadRowsResponse_Chunk) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{1, 0} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{1, 0} } func (m *ReadRowsResponse_Chunk) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadRowsResponse_Chunk.Unmarshal(m, b) @@ -329,13 +329,13 @@ type isReadRowsResponse_Chunk_Chunk interface { } type ReadRowsResponse_Chunk_RowContents struct { - RowContents *Family `protobuf:"bytes,1,opt,name=row_contents,json=rowContents,oneof"` + RowContents *Family `protobuf:"bytes,1,opt,name=row_contents,json=rowContents,proto3,oneof"` } type ReadRowsResponse_Chunk_ResetRow struct { - ResetRow bool `protobuf:"varint,2,opt,name=reset_row,json=resetRow,oneof"` + ResetRow bool `protobuf:"varint,2,opt,name=reset_row,json=resetRow,proto3,oneof"` } type ReadRowsResponse_Chunk_CommitRow struct { - CommitRow bool `protobuf:"varint,3,opt,name=commit_row,json=commitRow,oneof"` + CommitRow bool `protobuf:"varint,3,opt,name=commit_row,json=commitRow,proto3,oneof"` } func (*ReadRowsResponse_Chunk_RowContents) isReadRowsResponse_Chunk_Chunk() {} @@ -464,7 +464,7 @@ func _ReadRowsResponse_Chunk_OneofSizer(msg proto.Message) (n int) { // Request message for BigtableService.SampleRowKeys. type SampleRowKeysRequest struct { // The unique name of the table from which to sample row keys. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -474,7 +474,7 @@ func (m *SampleRowKeysRequest) Reset() { *m = SampleRowKeysRequest{} } func (m *SampleRowKeysRequest) String() string { return proto.CompactTextString(m) } func (*SampleRowKeysRequest) ProtoMessage() {} func (*SampleRowKeysRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{2} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{2} } func (m *SampleRowKeysRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SampleRowKeysRequest.Unmarshal(m, b) @@ -515,7 +515,7 @@ type SampleRowKeysResponse struct { // "row_key". Buffering the contents of all rows between two subsequent // samples would require space roughly equal to the difference in their // "offset_bytes" fields. - OffsetBytes int64 `protobuf:"varint,2,opt,name=offset_bytes,json=offsetBytes" json:"offset_bytes,omitempty"` + OffsetBytes int64 `protobuf:"varint,2,opt,name=offset_bytes,json=offsetBytes,proto3" json:"offset_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -525,7 +525,7 @@ func (m *SampleRowKeysResponse) Reset() { *m = SampleRowKeysResponse{} } func (m *SampleRowKeysResponse) String() string { return proto.CompactTextString(m) } func (*SampleRowKeysResponse) ProtoMessage() {} func (*SampleRowKeysResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{3} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{3} } func (m *SampleRowKeysResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SampleRowKeysResponse.Unmarshal(m, b) @@ -562,13 +562,13 @@ func (m *SampleRowKeysResponse) GetOffsetBytes() int64 { // Request message for BigtableService.MutateRow. type MutateRowRequest struct { // The unique name of the table to which the mutation should be applied. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // The key of the row to which the mutation should be applied. RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // Changes to be atomically applied to the specified row. Entries are applied // in order, meaning that earlier mutations can be masked by later ones. // Must contain at least one entry and at most 100000. - Mutations []*Mutation `protobuf:"bytes,3,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,3,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -578,7 +578,7 @@ func (m *MutateRowRequest) Reset() { *m = MutateRowRequest{} } func (m *MutateRowRequest) String() string { return proto.CompactTextString(m) } func (*MutateRowRequest) ProtoMessage() {} func (*MutateRowRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{4} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{4} } func (m *MutateRowRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowRequest.Unmarshal(m, b) @@ -622,13 +622,13 @@ func (m *MutateRowRequest) GetMutations() []*Mutation { // Request message for BigtableService.MutateRows. type MutateRowsRequest struct { // The unique name of the table to which the mutations should be applied. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // The row keys/mutations to be applied in bulk. // Each entry is applied as an atomic mutation, but the entries may be // applied in arbitrary order (even between entries for the same row). // At least one entry must be specified, and in total the entries may // contain at most 100000 mutations. - Entries []*MutateRowsRequest_Entry `protobuf:"bytes,2,rep,name=entries" json:"entries,omitempty"` + Entries []*MutateRowsRequest_Entry `protobuf:"bytes,2,rep,name=entries,proto3" json:"entries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -638,7 +638,7 @@ func (m *MutateRowsRequest) Reset() { *m = MutateRowsRequest{} } func (m *MutateRowsRequest) String() string { return proto.CompactTextString(m) } func (*MutateRowsRequest) ProtoMessage() {} func (*MutateRowsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{5} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{5} } func (m *MutateRowsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsRequest.Unmarshal(m, b) @@ -679,7 +679,7 @@ type MutateRowsRequest_Entry struct { // applied in order, meaning that earlier mutations can be masked by // later ones. // At least one mutation must be specified. - Mutations []*Mutation `protobuf:"bytes,2,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,2,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -689,7 +689,7 @@ func (m *MutateRowsRequest_Entry) Reset() { *m = MutateRowsRequest_Entry func (m *MutateRowsRequest_Entry) String() string { return proto.CompactTextString(m) } func (*MutateRowsRequest_Entry) ProtoMessage() {} func (*MutateRowsRequest_Entry) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{5, 0} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{5, 0} } func (m *MutateRowsRequest_Entry) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsRequest_Entry.Unmarshal(m, b) @@ -730,7 +730,7 @@ type MutateRowsResponse struct { // Depending on how requests are batched during execution, it is possible // for one Entry to fail due to an error with another Entry. In the event // that this occurs, the same error will be reported for both entries. - Statuses []*status.Status `protobuf:"bytes,1,rep,name=statuses" json:"statuses,omitempty"` + Statuses []*status.Status `protobuf:"bytes,1,rep,name=statuses,proto3" json:"statuses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -740,7 +740,7 @@ func (m *MutateRowsResponse) Reset() { *m = MutateRowsResponse{} } func (m *MutateRowsResponse) String() string { return proto.CompactTextString(m) } func (*MutateRowsResponse) ProtoMessage() {} func (*MutateRowsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{6} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{6} } func (m *MutateRowsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsResponse.Unmarshal(m, b) @@ -771,26 +771,26 @@ func (m *MutateRowsResponse) GetStatuses() []*status.Status { type CheckAndMutateRowRequest struct { // The unique name of the table to which the conditional mutation should be // applied. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // The key of the row to which the conditional mutation should be applied. RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // The filter to be applied to the contents of the specified row. Depending // on whether or not any results are yielded, either "true_mutations" or // "false_mutations" will be executed. If unset, checks that the row contains // any values at all. - PredicateFilter *RowFilter `protobuf:"bytes,6,opt,name=predicate_filter,json=predicateFilter" json:"predicate_filter,omitempty"` + PredicateFilter *RowFilter `protobuf:"bytes,6,opt,name=predicate_filter,json=predicateFilter,proto3" json:"predicate_filter,omitempty"` // Changes to be atomically applied to the specified row if "predicate_filter" // yields at least one cell when applied to "row_key". Entries are applied in // order, meaning that earlier mutations can be masked by later ones. // Must contain at least one entry if "false_mutations" is empty, and at most // 100000. - TrueMutations []*Mutation `protobuf:"bytes,4,rep,name=true_mutations,json=trueMutations" json:"true_mutations,omitempty"` + TrueMutations []*Mutation `protobuf:"bytes,4,rep,name=true_mutations,json=trueMutations,proto3" json:"true_mutations,omitempty"` // Changes to be atomically applied to the specified row if "predicate_filter" // does not yield any cells when applied to "row_key". Entries are applied in // order, meaning that earlier mutations can be masked by later ones. // Must contain at least one entry if "true_mutations" is empty, and at most // 100000. - FalseMutations []*Mutation `protobuf:"bytes,5,rep,name=false_mutations,json=falseMutations" json:"false_mutations,omitempty"` + FalseMutations []*Mutation `protobuf:"bytes,5,rep,name=false_mutations,json=falseMutations,proto3" json:"false_mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -800,7 +800,7 @@ func (m *CheckAndMutateRowRequest) Reset() { *m = CheckAndMutateRowReque func (m *CheckAndMutateRowRequest) String() string { return proto.CompactTextString(m) } func (*CheckAndMutateRowRequest) ProtoMessage() {} func (*CheckAndMutateRowRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{7} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{7} } func (m *CheckAndMutateRowRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CheckAndMutateRowRequest.Unmarshal(m, b) @@ -859,7 +859,7 @@ func (m *CheckAndMutateRowRequest) GetFalseMutations() []*Mutation { type CheckAndMutateRowResponse struct { // Whether or not the request's "predicate_filter" yielded any results for // the specified row. - PredicateMatched bool `protobuf:"varint,1,opt,name=predicate_matched,json=predicateMatched" json:"predicate_matched,omitempty"` + PredicateMatched bool `protobuf:"varint,1,opt,name=predicate_matched,json=predicateMatched,proto3" json:"predicate_matched,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -869,7 +869,7 @@ func (m *CheckAndMutateRowResponse) Reset() { *m = CheckAndMutateRowResp func (m *CheckAndMutateRowResponse) String() string { return proto.CompactTextString(m) } func (*CheckAndMutateRowResponse) ProtoMessage() {} func (*CheckAndMutateRowResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{8} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{8} } func (m *CheckAndMutateRowResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CheckAndMutateRowResponse.Unmarshal(m, b) @@ -900,13 +900,13 @@ func (m *CheckAndMutateRowResponse) GetPredicateMatched() bool { type ReadModifyWriteRowRequest struct { // The unique name of the table to which the read/modify/write rules should be // applied. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // The key of the row to which the read/modify/write rules should be applied. RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // Rules specifying how the specified row's contents are to be transformed // into writes. Entries are applied in order, meaning that earlier rules will // affect the results of later ones. - Rules []*ReadModifyWriteRule `protobuf:"bytes,3,rep,name=rules" json:"rules,omitempty"` + Rules []*ReadModifyWriteRule `protobuf:"bytes,3,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -916,7 +916,7 @@ func (m *ReadModifyWriteRowRequest) Reset() { *m = ReadModifyWriteRowReq func (m *ReadModifyWriteRowRequest) String() string { return proto.CompactTextString(m) } func (*ReadModifyWriteRowRequest) ProtoMessage() {} func (*ReadModifyWriteRowRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25, []int{9} + return fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b, []int{9} } func (m *ReadModifyWriteRowRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadModifyWriteRowRequest.Unmarshal(m, b) @@ -973,10 +973,10 @@ func init() { } func init() { - proto.RegisterFile("google/bigtable/v1/bigtable_service_messages.proto", fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25) + proto.RegisterFile("google/bigtable/v1/bigtable_service_messages.proto", fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b) } -var fileDescriptor_bigtable_service_messages_c4e1e5accd76ba25 = []byte{ +var fileDescriptor_bigtable_service_messages_b1638ccc1a7cde9b = []byte{ // 788 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x55, 0x5f, 0x8b, 0x23, 0x45, 0x10, 0xdf, 0x49, 0xcc, 0xbf, 0x4a, 0xdc, 0xdd, 0x6b, 0xce, 0xdb, 0xd9, 0x70, 0x8b, 0x71, 0x10, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/v2/bigtable.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/v2/bigtable.pb.go index b94d29dea..51fce7bea 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/v2/bigtable.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/v2/bigtable.pb.go @@ -31,23 +31,18 @@ type ReadRowsRequest struct { // The unique name of the table from which to read. // Values are of the form // `projects//instances//tables/
`. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // This value specifies routing for replication. If not specified, the // "default" application profile will be used. - AppProfileId string `protobuf:"bytes,5,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,5,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` // The row keys and/or ranges to read. If not specified, reads from all rows. - Rows *RowSet `protobuf:"bytes,2,opt,name=rows" json:"rows,omitempty"` + Rows *RowSet `protobuf:"bytes,2,opt,name=rows,proto3" json:"rows,omitempty"` // The filter to apply to the contents of the specified row(s). If unset, // reads the entirety of each row. - Filter *RowFilter `protobuf:"bytes,3,opt,name=filter" json:"filter,omitempty"` + Filter *RowFilter `protobuf:"bytes,3,opt,name=filter,proto3" json:"filter,omitempty"` // The read will terminate after committing to N rows' worth of results. The // default (zero) is to return all results. - RowsLimit int64 `protobuf:"varint,4,opt,name=rows_limit,json=rowsLimit" json:"rows_limit,omitempty"` + RowsLimit int64 `protobuf:"varint,4,opt,name=rows_limit,json=rowsLimit,proto3" json:"rows_limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -57,7 +52,7 @@ func (m *ReadRowsRequest) Reset() { *m = ReadRowsRequest{} } func (m *ReadRowsRequest) String() string { return proto.CompactTextString(m) } func (*ReadRowsRequest) ProtoMessage() {} func (*ReadRowsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{0} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{0} } func (m *ReadRowsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadRowsRequest.Unmarshal(m, b) @@ -114,7 +109,7 @@ func (m *ReadRowsRequest) GetRowsLimit() int64 { // Response message for Bigtable.ReadRows. type ReadRowsResponse struct { - Chunks []*ReadRowsResponse_CellChunk `protobuf:"bytes,1,rep,name=chunks" json:"chunks,omitempty"` + Chunks []*ReadRowsResponse_CellChunk `protobuf:"bytes,1,rep,name=chunks,proto3" json:"chunks,omitempty"` // Optionally the server might return the row key of the last row it // has scanned. The client can use this to construct a more // efficient retry request if needed: any row keys or portions of @@ -132,7 +127,7 @@ func (m *ReadRowsResponse) Reset() { *m = ReadRowsResponse{} } func (m *ReadRowsResponse) String() string { return proto.CompactTextString(m) } func (*ReadRowsResponse) ProtoMessage() {} func (*ReadRowsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{1} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{1} } func (m *ReadRowsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadRowsResponse.Unmarshal(m, b) @@ -180,13 +175,13 @@ type ReadRowsResponse_CellChunk struct { // column family name in a response so clients must check // explicitly for the presence of this message, not just for // `family_name.value` being non-empty. - FamilyName *wrappers.StringValue `protobuf:"bytes,2,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName *wrappers.StringValue `protobuf:"bytes,2,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The column qualifier for this chunk of data. If this message // is not present, this CellChunk is a continuation of the same column // as the previous CellChunk. Column qualifiers may be empty so // clients must check for the presence of this message, not just // for `qualifier.value` being non-empty. - Qualifier *wrappers.BytesValue `protobuf:"bytes,3,opt,name=qualifier" json:"qualifier,omitempty"` + Qualifier *wrappers.BytesValue `protobuf:"bytes,3,opt,name=qualifier,proto3" json:"qualifier,omitempty"` // The cell's stored timestamp, which also uniquely identifies it // within its column. Values are always expressed in // microseconds, but individual tables may set a coarser @@ -195,11 +190,11 @@ type ReadRowsResponse_CellChunk struct { // only allow values of `timestamp_micros` which are multiples of // 1000. Timestamps are only set in the first CellChunk per cell // (for cells split into multiple chunks). - TimestampMicros int64 `protobuf:"varint,4,opt,name=timestamp_micros,json=timestampMicros" json:"timestamp_micros,omitempty"` + TimestampMicros int64 `protobuf:"varint,4,opt,name=timestamp_micros,json=timestampMicros,proto3" json:"timestamp_micros,omitempty"` // Labels applied to the cell by a // [RowFilter][google.bigtable.v2.RowFilter]. Labels are only set // on the first CellChunk per cell. - Labels []string `protobuf:"bytes,5,rep,name=labels" json:"labels,omitempty"` + Labels []string `protobuf:"bytes,5,rep,name=labels,proto3" json:"labels,omitempty"` // The value stored in the cell. Cell values can be split across // multiple CellChunks. In that case only the value field will be // set in CellChunks after the first: the timestamp and labels @@ -210,7 +205,7 @@ type ReadRowsResponse_CellChunk struct { // not the final chunk of that cell, value_size will be set to the // total length of the cell value. The client can use this size // to pre-allocate memory to hold the full cell value. - ValueSize int32 `protobuf:"varint,7,opt,name=value_size,json=valueSize" json:"value_size,omitempty"` + ValueSize int32 `protobuf:"varint,7,opt,name=value_size,json=valueSize,proto3" json:"value_size,omitempty"` // Types that are valid to be assigned to RowStatus: // *ReadRowsResponse_CellChunk_ResetRow // *ReadRowsResponse_CellChunk_CommitRow @@ -224,7 +219,7 @@ func (m *ReadRowsResponse_CellChunk) Reset() { *m = ReadRowsResponse_Cel func (m *ReadRowsResponse_CellChunk) String() string { return proto.CompactTextString(m) } func (*ReadRowsResponse_CellChunk) ProtoMessage() {} func (*ReadRowsResponse_CellChunk) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{1, 0} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{1, 0} } func (m *ReadRowsResponse_CellChunk) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadRowsResponse_CellChunk.Unmarshal(m, b) @@ -249,10 +244,10 @@ type isReadRowsResponse_CellChunk_RowStatus interface { } type ReadRowsResponse_CellChunk_ResetRow struct { - ResetRow bool `protobuf:"varint,8,opt,name=reset_row,json=resetRow,oneof"` + ResetRow bool `protobuf:"varint,8,opt,name=reset_row,json=resetRow,proto3,oneof"` } type ReadRowsResponse_CellChunk_CommitRow struct { - CommitRow bool `protobuf:"varint,9,opt,name=commit_row,json=commitRow,oneof"` + CommitRow bool `protobuf:"varint,9,opt,name=commit_row,json=commitRow,proto3,oneof"` } func (*ReadRowsResponse_CellChunk_ResetRow) isReadRowsResponse_CellChunk_RowStatus() {} @@ -405,15 +400,10 @@ type SampleRowKeysRequest struct { // The unique name of the table from which to sample row keys. // Values are of the form // `projects//instances//tables/
`. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // This value specifies routing for replication. If not specified, the // "default" application profile will be used. - AppProfileId string `protobuf:"bytes,2,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,2,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -423,7 +413,7 @@ func (m *SampleRowKeysRequest) Reset() { *m = SampleRowKeysRequest{} } func (m *SampleRowKeysRequest) String() string { return proto.CompactTextString(m) } func (*SampleRowKeysRequest) ProtoMessage() {} func (*SampleRowKeysRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{2} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{2} } func (m *SampleRowKeysRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SampleRowKeysRequest.Unmarshal(m, b) @@ -471,7 +461,7 @@ type SampleRowKeysResponse struct { // `row_key`. Buffering the contents of all rows between two subsequent // samples would require space roughly equal to the difference in their // `offset_bytes` fields. - OffsetBytes int64 `protobuf:"varint,2,opt,name=offset_bytes,json=offsetBytes" json:"offset_bytes,omitempty"` + OffsetBytes int64 `protobuf:"varint,2,opt,name=offset_bytes,json=offsetBytes,proto3" json:"offset_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -481,7 +471,7 @@ func (m *SampleRowKeysResponse) Reset() { *m = SampleRowKeysResponse{} } func (m *SampleRowKeysResponse) String() string { return proto.CompactTextString(m) } func (*SampleRowKeysResponse) ProtoMessage() {} func (*SampleRowKeysResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{3} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{3} } func (m *SampleRowKeysResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SampleRowKeysResponse.Unmarshal(m, b) @@ -520,21 +510,16 @@ type MutateRowRequest struct { // The unique name of the table to which the mutation should be applied. // Values are of the form // `projects//instances//tables/
`. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // This value specifies routing for replication. If not specified, the // "default" application profile will be used. - AppProfileId string `protobuf:"bytes,4,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,4,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` // The key of the row to which the mutation should be applied. RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // Changes to be atomically applied to the specified row. Entries are applied // in order, meaning that earlier mutations can be masked by later ones. // Must contain at least one entry and at most 100000. - Mutations []*Mutation `protobuf:"bytes,3,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,3,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -544,7 +529,7 @@ func (m *MutateRowRequest) Reset() { *m = MutateRowRequest{} } func (m *MutateRowRequest) String() string { return proto.CompactTextString(m) } func (*MutateRowRequest) ProtoMessage() {} func (*MutateRowRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{4} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{4} } func (m *MutateRowRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowRequest.Unmarshal(m, b) @@ -603,7 +588,7 @@ func (m *MutateRowResponse) Reset() { *m = MutateRowResponse{} } func (m *MutateRowResponse) String() string { return proto.CompactTextString(m) } func (*MutateRowResponse) ProtoMessage() {} func (*MutateRowResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{5} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{5} } func (m *MutateRowResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowResponse.Unmarshal(m, b) @@ -626,21 +611,16 @@ var xxx_messageInfo_MutateRowResponse proto.InternalMessageInfo // Request message for BigtableService.MutateRows. type MutateRowsRequest struct { // The unique name of the table to which the mutations should be applied. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // This value specifies routing for replication. If not specified, the // "default" application profile will be used. - AppProfileId string `protobuf:"bytes,3,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,3,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` // The row keys and corresponding mutations to be applied in bulk. // Each entry is applied as an atomic mutation, but the entries may be // applied in arbitrary order (even between entries for the same row). // At least one entry must be specified, and in total the entries can // contain at most 100000 mutations. - Entries []*MutateRowsRequest_Entry `protobuf:"bytes,2,rep,name=entries" json:"entries,omitempty"` + Entries []*MutateRowsRequest_Entry `protobuf:"bytes,2,rep,name=entries,proto3" json:"entries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -650,7 +630,7 @@ func (m *MutateRowsRequest) Reset() { *m = MutateRowsRequest{} } func (m *MutateRowsRequest) String() string { return proto.CompactTextString(m) } func (*MutateRowsRequest) ProtoMessage() {} func (*MutateRowsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{6} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{6} } func (m *MutateRowsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsRequest.Unmarshal(m, b) @@ -698,7 +678,7 @@ type MutateRowsRequest_Entry struct { // applied in order, meaning that earlier mutations can be masked by // later ones. // You must specify at least one mutation. - Mutations []*Mutation `protobuf:"bytes,2,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,2,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -708,7 +688,7 @@ func (m *MutateRowsRequest_Entry) Reset() { *m = MutateRowsRequest_Entry func (m *MutateRowsRequest_Entry) String() string { return proto.CompactTextString(m) } func (*MutateRowsRequest_Entry) ProtoMessage() {} func (*MutateRowsRequest_Entry) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{6, 0} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{6, 0} } func (m *MutateRowsRequest_Entry) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsRequest_Entry.Unmarshal(m, b) @@ -745,7 +725,7 @@ func (m *MutateRowsRequest_Entry) GetMutations() []*Mutation { // Response message for BigtableService.MutateRows. type MutateRowsResponse struct { // One or more results for Entries from the batch request. - Entries []*MutateRowsResponse_Entry `protobuf:"bytes,1,rep,name=entries" json:"entries,omitempty"` + Entries []*MutateRowsResponse_Entry `protobuf:"bytes,1,rep,name=entries,proto3" json:"entries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -755,7 +735,7 @@ func (m *MutateRowsResponse) Reset() { *m = MutateRowsResponse{} } func (m *MutateRowsResponse) String() string { return proto.CompactTextString(m) } func (*MutateRowsResponse) ProtoMessage() {} func (*MutateRowsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{7} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{7} } func (m *MutateRowsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsResponse.Unmarshal(m, b) @@ -785,12 +765,12 @@ func (m *MutateRowsResponse) GetEntries() []*MutateRowsResponse_Entry { type MutateRowsResponse_Entry struct { // The index into the original request's `entries` list of the Entry // for which a result is being reported. - Index int64 `protobuf:"varint,1,opt,name=index" json:"index,omitempty"` + Index int64 `protobuf:"varint,1,opt,name=index,proto3" json:"index,omitempty"` // The result of the request Entry identified by `index`. // Depending on how requests are batched during execution, it is possible // for one Entry to fail due to an error with another Entry. In the event // that this occurs, the same error will be reported for both entries. - Status *status.Status `protobuf:"bytes,2,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -800,7 +780,7 @@ func (m *MutateRowsResponse_Entry) Reset() { *m = MutateRowsResponse_Ent func (m *MutateRowsResponse_Entry) String() string { return proto.CompactTextString(m) } func (*MutateRowsResponse_Entry) ProtoMessage() {} func (*MutateRowsResponse_Entry) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{7, 0} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{7, 0} } func (m *MutateRowsResponse_Entry) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_MutateRowsResponse_Entry.Unmarshal(m, b) @@ -840,34 +820,29 @@ type CheckAndMutateRowRequest struct { // applied. // Values are of the form // `projects//instances//tables/
`. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // This value specifies routing for replication. If not specified, the // "default" application profile will be used. - AppProfileId string `protobuf:"bytes,7,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,7,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` // The key of the row to which the conditional mutation should be applied. RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // The filter to be applied to the contents of the specified row. Depending // on whether or not any results are yielded, either `true_mutations` or // `false_mutations` will be executed. If unset, checks that the row contains // any values at all. - PredicateFilter *RowFilter `protobuf:"bytes,6,opt,name=predicate_filter,json=predicateFilter" json:"predicate_filter,omitempty"` + PredicateFilter *RowFilter `protobuf:"bytes,6,opt,name=predicate_filter,json=predicateFilter,proto3" json:"predicate_filter,omitempty"` // Changes to be atomically applied to the specified row if `predicate_filter` // yields at least one cell when applied to `row_key`. Entries are applied in // order, meaning that earlier mutations can be masked by later ones. // Must contain at least one entry if `false_mutations` is empty, and at most // 100000. - TrueMutations []*Mutation `protobuf:"bytes,4,rep,name=true_mutations,json=trueMutations" json:"true_mutations,omitempty"` + TrueMutations []*Mutation `protobuf:"bytes,4,rep,name=true_mutations,json=trueMutations,proto3" json:"true_mutations,omitempty"` // Changes to be atomically applied to the specified row if `predicate_filter` // does not yield any cells when applied to `row_key`. Entries are applied in // order, meaning that earlier mutations can be masked by later ones. // Must contain at least one entry if `true_mutations` is empty, and at most // 100000. - FalseMutations []*Mutation `protobuf:"bytes,5,rep,name=false_mutations,json=falseMutations" json:"false_mutations,omitempty"` + FalseMutations []*Mutation `protobuf:"bytes,5,rep,name=false_mutations,json=falseMutations,proto3" json:"false_mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -877,7 +852,7 @@ func (m *CheckAndMutateRowRequest) Reset() { *m = CheckAndMutateRowReque func (m *CheckAndMutateRowRequest) String() string { return proto.CompactTextString(m) } func (*CheckAndMutateRowRequest) ProtoMessage() {} func (*CheckAndMutateRowRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{8} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{8} } func (m *CheckAndMutateRowRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CheckAndMutateRowRequest.Unmarshal(m, b) @@ -943,7 +918,7 @@ func (m *CheckAndMutateRowRequest) GetFalseMutations() []*Mutation { type CheckAndMutateRowResponse struct { // Whether or not the request's `predicate_filter` yielded any results for // the specified row. - PredicateMatched bool `protobuf:"varint,1,opt,name=predicate_matched,json=predicateMatched" json:"predicate_matched,omitempty"` + PredicateMatched bool `protobuf:"varint,1,opt,name=predicate_matched,json=predicateMatched,proto3" json:"predicate_matched,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -953,7 +928,7 @@ func (m *CheckAndMutateRowResponse) Reset() { *m = CheckAndMutateRowResp func (m *CheckAndMutateRowResponse) String() string { return proto.CompactTextString(m) } func (*CheckAndMutateRowResponse) ProtoMessage() {} func (*CheckAndMutateRowResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{9} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{9} } func (m *CheckAndMutateRowResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CheckAndMutateRowResponse.Unmarshal(m, b) @@ -986,21 +961,16 @@ type ReadModifyWriteRowRequest struct { // applied. // Values are of the form // `projects//instances//tables/
`. - TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` - // This is a private alpha release of Cloud Bigtable replication. This feature - // is not currently available to most Cloud Bigtable customers. This feature - // might be changed in backward-incompatible ways and is not recommended for - // production use. It is not subject to any SLA or deprecation policy. - // + TableName string `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // This value specifies routing for replication. If not specified, the // "default" application profile will be used. - AppProfileId string `protobuf:"bytes,4,opt,name=app_profile_id,json=appProfileId" json:"app_profile_id,omitempty"` + AppProfileId string `protobuf:"bytes,4,opt,name=app_profile_id,json=appProfileId,proto3" json:"app_profile_id,omitempty"` // The key of the row to which the read/modify/write rules should be applied. RowKey []byte `protobuf:"bytes,2,opt,name=row_key,json=rowKey,proto3" json:"row_key,omitempty"` // Rules specifying how the specified row's contents are to be transformed // into writes. Entries are applied in order, meaning that earlier rules will // affect the results of later ones. - Rules []*ReadModifyWriteRule `protobuf:"bytes,3,rep,name=rules" json:"rules,omitempty"` + Rules []*ReadModifyWriteRule `protobuf:"bytes,3,rep,name=rules,proto3" json:"rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1010,7 +980,7 @@ func (m *ReadModifyWriteRowRequest) Reset() { *m = ReadModifyWriteRowReq func (m *ReadModifyWriteRowRequest) String() string { return proto.CompactTextString(m) } func (*ReadModifyWriteRowRequest) ProtoMessage() {} func (*ReadModifyWriteRowRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{10} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{10} } func (m *ReadModifyWriteRowRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadModifyWriteRowRequest.Unmarshal(m, b) @@ -1061,7 +1031,7 @@ func (m *ReadModifyWriteRowRequest) GetRules() []*ReadModifyWriteRule { // Response message for Bigtable.ReadModifyWriteRow. type ReadModifyWriteRowResponse struct { // A Row containing the new contents of all cells modified by the request. - Row *Row `protobuf:"bytes,1,opt,name=row" json:"row,omitempty"` + Row *Row `protobuf:"bytes,1,opt,name=row,proto3" json:"row,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1071,7 +1041,7 @@ func (m *ReadModifyWriteRowResponse) Reset() { *m = ReadModifyWriteRowRe func (m *ReadModifyWriteRowResponse) String() string { return proto.CompactTextString(m) } func (*ReadModifyWriteRowResponse) ProtoMessage() {} func (*ReadModifyWriteRowResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_bigtable_5207bf1d2c849e53, []int{11} + return fileDescriptor_bigtable_8d6dd9ae311e02dd, []int{11} } func (m *ReadModifyWriteRowResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadModifyWriteRowResponse.Unmarshal(m, b) @@ -1287,8 +1257,7 @@ func (c *bigtableClient) ReadModifyWriteRow(ctx context.Context, in *ReadModifyW return out, nil } -// Server API for Bigtable service - +// BigtableServer is the server API for Bigtable service. type BigtableServer interface { // Streams back the contents of all requested rows in key order, optionally // applying the same Reader filter to each. Depending on their size, @@ -1477,10 +1446,10 @@ var _Bigtable_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/bigtable/v2/bigtable.proto", fileDescriptor_bigtable_5207bf1d2c849e53) + proto.RegisterFile("google/bigtable/v2/bigtable.proto", fileDescriptor_bigtable_8d6dd9ae311e02dd) } -var fileDescriptor_bigtable_5207bf1d2c849e53 = []byte{ +var fileDescriptor_bigtable_8d6dd9ae311e02dd = []byte{ // 1210 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x57, 0x41, 0x6f, 0x1b, 0x45, 0x14, 0x66, 0xec, 0xd8, 0xf1, 0xbe, 0xa4, 0x4d, 0x32, 0x84, 0x76, 0x6b, 0x5a, 0x70, 0x97, 0x16, diff --git a/vendor/google.golang.org/genproto/googleapis/bigtable/v2/data.pb.go b/vendor/google.golang.org/genproto/googleapis/bigtable/v2/data.pb.go index 28924c94f..3554517f5 100644 --- a/vendor/google.golang.org/genproto/googleapis/bigtable/v2/data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bigtable/v2/data.pb.go @@ -27,7 +27,7 @@ type Row struct { Key []byte `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // May be empty, but only if the entire row is empty. // The mutual ordering of column families is not specified. - Families []*Family `protobuf:"bytes,2,rep,name=families" json:"families,omitempty"` + Families []*Family `protobuf:"bytes,2,rep,name=families,proto3" json:"families,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -37,7 +37,7 @@ func (m *Row) Reset() { *m = Row{} } func (m *Row) String() string { return proto.CompactTextString(m) } func (*Row) ProtoMessage() {} func (*Row) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{0} + return fileDescriptor_data_b0430eb990758496, []int{0} } func (m *Row) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Row.Unmarshal(m, b) @@ -80,9 +80,9 @@ type Family struct { // Must match `[-_.a-zA-Z0-9]+`, except that AggregatingRowProcessors may // produce cells in a sentinel family with an empty name. // Must be no greater than 64 characters in length. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Must not be empty. Sorted in order of increasing "qualifier". - Columns []*Column `protobuf:"bytes,2,rep,name=columns" json:"columns,omitempty"` + Columns []*Column `protobuf:"bytes,2,rep,name=columns,proto3" json:"columns,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -92,7 +92,7 @@ func (m *Family) Reset() { *m = Family{} } func (m *Family) String() string { return proto.CompactTextString(m) } func (*Family) ProtoMessage() {} func (*Family) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{1} + return fileDescriptor_data_b0430eb990758496, []int{1} } func (m *Family) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Family.Unmarshal(m, b) @@ -136,7 +136,7 @@ type Column struct { // length. Qualifier []byte `protobuf:"bytes,1,opt,name=qualifier,proto3" json:"qualifier,omitempty"` // Must not be empty. Sorted in order of decreasing "timestamp_micros". - Cells []*Cell `protobuf:"bytes,2,rep,name=cells" json:"cells,omitempty"` + Cells []*Cell `protobuf:"bytes,2,rep,name=cells,proto3" json:"cells,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -146,7 +146,7 @@ func (m *Column) Reset() { *m = Column{} } func (m *Column) String() string { return proto.CompactTextString(m) } func (*Column) ProtoMessage() {} func (*Column) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{2} + return fileDescriptor_data_b0430eb990758496, []int{2} } func (m *Column) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Column.Unmarshal(m, b) @@ -188,13 +188,13 @@ type Cell struct { // a coarser granularity to further restrict the allowed values. For // example, a table which specifies millisecond granularity will only allow // values of `timestamp_micros` which are multiples of 1000. - TimestampMicros int64 `protobuf:"varint,1,opt,name=timestamp_micros,json=timestampMicros" json:"timestamp_micros,omitempty"` + TimestampMicros int64 `protobuf:"varint,1,opt,name=timestamp_micros,json=timestampMicros,proto3" json:"timestamp_micros,omitempty"` // The value stored in the cell. // May contain any byte string, including the empty string, up to 100MiB in // length. Value []byte `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Labels applied to the cell by a [RowFilter][google.bigtable.v2.RowFilter]. - Labels []string `protobuf:"bytes,3,rep,name=labels" json:"labels,omitempty"` + Labels []string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -204,7 +204,7 @@ func (m *Cell) Reset() { *m = Cell{} } func (m *Cell) String() string { return proto.CompactTextString(m) } func (*Cell) ProtoMessage() {} func (*Cell) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{3} + return fileDescriptor_data_b0430eb990758496, []int{3} } func (m *Cell) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Cell.Unmarshal(m, b) @@ -270,7 +270,7 @@ func (m *RowRange) Reset() { *m = RowRange{} } func (m *RowRange) String() string { return proto.CompactTextString(m) } func (*RowRange) ProtoMessage() {} func (*RowRange) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{4} + return fileDescriptor_data_b0430eb990758496, []int{4} } func (m *RowRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowRange.Unmarshal(m, b) @@ -469,7 +469,7 @@ type RowSet struct { // Single rows included in the set. RowKeys [][]byte `protobuf:"bytes,1,rep,name=row_keys,json=rowKeys,proto3" json:"row_keys,omitempty"` // Contiguous row ranges included in the set. - RowRanges []*RowRange `protobuf:"bytes,2,rep,name=row_ranges,json=rowRanges" json:"row_ranges,omitempty"` + RowRanges []*RowRange `protobuf:"bytes,2,rep,name=row_ranges,json=rowRanges,proto3" json:"row_ranges,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -479,7 +479,7 @@ func (m *RowSet) Reset() { *m = RowSet{} } func (m *RowSet) String() string { return proto.CompactTextString(m) } func (*RowSet) ProtoMessage() {} func (*RowSet) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{5} + return fileDescriptor_data_b0430eb990758496, []int{5} } func (m *RowSet) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowSet.Unmarshal(m, b) @@ -519,7 +519,7 @@ func (m *RowSet) GetRowRanges() []*RowRange { // inclusive or exclusive. type ColumnRange struct { // The name of the column family within which this range falls. - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The column qualifier at which to start the range (within `column_family`). // If neither field is set, interpreted as the empty string, inclusive. // @@ -543,7 +543,7 @@ func (m *ColumnRange) Reset() { *m = ColumnRange{} } func (m *ColumnRange) String() string { return proto.CompactTextString(m) } func (*ColumnRange) ProtoMessage() {} func (*ColumnRange) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{6} + return fileDescriptor_data_b0430eb990758496, []int{6} } func (m *ColumnRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ColumnRange.Unmarshal(m, b) @@ -747,9 +747,9 @@ func _ColumnRange_OneofSizer(msg proto.Message) (n int) { // Specified a contiguous range of microsecond timestamps. type TimestampRange struct { // Inclusive lower bound. If left empty, interpreted as 0. - StartTimestampMicros int64 `protobuf:"varint,1,opt,name=start_timestamp_micros,json=startTimestampMicros" json:"start_timestamp_micros,omitempty"` + StartTimestampMicros int64 `protobuf:"varint,1,opt,name=start_timestamp_micros,json=startTimestampMicros,proto3" json:"start_timestamp_micros,omitempty"` // Exclusive upper bound. If left empty, interpreted as infinity. - EndTimestampMicros int64 `protobuf:"varint,2,opt,name=end_timestamp_micros,json=endTimestampMicros" json:"end_timestamp_micros,omitempty"` + EndTimestampMicros int64 `protobuf:"varint,2,opt,name=end_timestamp_micros,json=endTimestampMicros,proto3" json:"end_timestamp_micros,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -759,7 +759,7 @@ func (m *TimestampRange) Reset() { *m = TimestampRange{} } func (m *TimestampRange) String() string { return proto.CompactTextString(m) } func (*TimestampRange) ProtoMessage() {} func (*TimestampRange) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{7} + return fileDescriptor_data_b0430eb990758496, []int{7} } func (m *TimestampRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TimestampRange.Unmarshal(m, b) @@ -818,7 +818,7 @@ func (m *ValueRange) Reset() { *m = ValueRange{} } func (m *ValueRange) String() string { return proto.CompactTextString(m) } func (*ValueRange) ProtoMessage() {} func (*ValueRange) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{8} + return fileDescriptor_data_b0430eb990758496, []int{8} } func (m *ValueRange) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ValueRange.Unmarshal(m, b) @@ -1079,7 +1079,7 @@ func (m *RowFilter) Reset() { *m = RowFilter{} } func (m *RowFilter) String() string { return proto.CompactTextString(m) } func (*RowFilter) ProtoMessage() {} func (*RowFilter) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{9} + return fileDescriptor_data_b0430eb990758496, []int{9} } func (m *RowFilter) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter.Unmarshal(m, b) @@ -1104,61 +1104,61 @@ type isRowFilter_Filter interface { } type RowFilter_Chain_ struct { - Chain *RowFilter_Chain `protobuf:"bytes,1,opt,name=chain,oneof"` + Chain *RowFilter_Chain `protobuf:"bytes,1,opt,name=chain,proto3,oneof"` } type RowFilter_Interleave_ struct { - Interleave *RowFilter_Interleave `protobuf:"bytes,2,opt,name=interleave,oneof"` + Interleave *RowFilter_Interleave `protobuf:"bytes,2,opt,name=interleave,proto3,oneof"` } type RowFilter_Condition_ struct { - Condition *RowFilter_Condition `protobuf:"bytes,3,opt,name=condition,oneof"` + Condition *RowFilter_Condition `protobuf:"bytes,3,opt,name=condition,proto3,oneof"` } type RowFilter_Sink struct { - Sink bool `protobuf:"varint,16,opt,name=sink,oneof"` + Sink bool `protobuf:"varint,16,opt,name=sink,proto3,oneof"` } type RowFilter_PassAllFilter struct { - PassAllFilter bool `protobuf:"varint,17,opt,name=pass_all_filter,json=passAllFilter,oneof"` + PassAllFilter bool `protobuf:"varint,17,opt,name=pass_all_filter,json=passAllFilter,proto3,oneof"` } type RowFilter_BlockAllFilter struct { - BlockAllFilter bool `protobuf:"varint,18,opt,name=block_all_filter,json=blockAllFilter,oneof"` + BlockAllFilter bool `protobuf:"varint,18,opt,name=block_all_filter,json=blockAllFilter,proto3,oneof"` } type RowFilter_RowKeyRegexFilter struct { RowKeyRegexFilter []byte `protobuf:"bytes,4,opt,name=row_key_regex_filter,json=rowKeyRegexFilter,proto3,oneof"` } type RowFilter_RowSampleFilter struct { - RowSampleFilter float64 `protobuf:"fixed64,14,opt,name=row_sample_filter,json=rowSampleFilter,oneof"` + RowSampleFilter float64 `protobuf:"fixed64,14,opt,name=row_sample_filter,json=rowSampleFilter,proto3,oneof"` } type RowFilter_FamilyNameRegexFilter struct { - FamilyNameRegexFilter string `protobuf:"bytes,5,opt,name=family_name_regex_filter,json=familyNameRegexFilter,oneof"` + FamilyNameRegexFilter string `protobuf:"bytes,5,opt,name=family_name_regex_filter,json=familyNameRegexFilter,proto3,oneof"` } type RowFilter_ColumnQualifierRegexFilter struct { ColumnQualifierRegexFilter []byte `protobuf:"bytes,6,opt,name=column_qualifier_regex_filter,json=columnQualifierRegexFilter,proto3,oneof"` } type RowFilter_ColumnRangeFilter struct { - ColumnRangeFilter *ColumnRange `protobuf:"bytes,7,opt,name=column_range_filter,json=columnRangeFilter,oneof"` + ColumnRangeFilter *ColumnRange `protobuf:"bytes,7,opt,name=column_range_filter,json=columnRangeFilter,proto3,oneof"` } type RowFilter_TimestampRangeFilter struct { - TimestampRangeFilter *TimestampRange `protobuf:"bytes,8,opt,name=timestamp_range_filter,json=timestampRangeFilter,oneof"` + TimestampRangeFilter *TimestampRange `protobuf:"bytes,8,opt,name=timestamp_range_filter,json=timestampRangeFilter,proto3,oneof"` } type RowFilter_ValueRegexFilter struct { ValueRegexFilter []byte `protobuf:"bytes,9,opt,name=value_regex_filter,json=valueRegexFilter,proto3,oneof"` } type RowFilter_ValueRangeFilter struct { - ValueRangeFilter *ValueRange `protobuf:"bytes,15,opt,name=value_range_filter,json=valueRangeFilter,oneof"` + ValueRangeFilter *ValueRange `protobuf:"bytes,15,opt,name=value_range_filter,json=valueRangeFilter,proto3,oneof"` } type RowFilter_CellsPerRowOffsetFilter struct { - CellsPerRowOffsetFilter int32 `protobuf:"varint,10,opt,name=cells_per_row_offset_filter,json=cellsPerRowOffsetFilter,oneof"` + CellsPerRowOffsetFilter int32 `protobuf:"varint,10,opt,name=cells_per_row_offset_filter,json=cellsPerRowOffsetFilter,proto3,oneof"` } type RowFilter_CellsPerRowLimitFilter struct { - CellsPerRowLimitFilter int32 `protobuf:"varint,11,opt,name=cells_per_row_limit_filter,json=cellsPerRowLimitFilter,oneof"` + CellsPerRowLimitFilter int32 `protobuf:"varint,11,opt,name=cells_per_row_limit_filter,json=cellsPerRowLimitFilter,proto3,oneof"` } type RowFilter_CellsPerColumnLimitFilter struct { - CellsPerColumnLimitFilter int32 `protobuf:"varint,12,opt,name=cells_per_column_limit_filter,json=cellsPerColumnLimitFilter,oneof"` + CellsPerColumnLimitFilter int32 `protobuf:"varint,12,opt,name=cells_per_column_limit_filter,json=cellsPerColumnLimitFilter,proto3,oneof"` } type RowFilter_StripValueTransformer struct { - StripValueTransformer bool `protobuf:"varint,13,opt,name=strip_value_transformer,json=stripValueTransformer,oneof"` + StripValueTransformer bool `protobuf:"varint,13,opt,name=strip_value_transformer,json=stripValueTransformer,proto3,oneof"` } type RowFilter_ApplyLabelTransformer struct { - ApplyLabelTransformer string `protobuf:"bytes,19,opt,name=apply_label_transformer,json=applyLabelTransformer,oneof"` + ApplyLabelTransformer string `protobuf:"bytes,19,opt,name=apply_label_transformer,json=applyLabelTransformer,proto3,oneof"` } func (*RowFilter_Chain_) isRowFilter_Filter() {} @@ -1679,7 +1679,7 @@ type RowFilter_Chain struct { // The elements of "filters" are chained together to process the input row: // in row -> f(0) -> intermediate row -> f(1) -> ... -> f(N) -> out row // The full chain is executed atomically. - Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters" json:"filters,omitempty"` + Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1689,7 +1689,7 @@ func (m *RowFilter_Chain) Reset() { *m = RowFilter_Chain{} } func (m *RowFilter_Chain) String() string { return proto.CompactTextString(m) } func (*RowFilter_Chain) ProtoMessage() {} func (*RowFilter_Chain) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{9, 0} + return fileDescriptor_data_b0430eb990758496, []int{9, 0} } func (m *RowFilter_Chain) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter_Chain.Unmarshal(m, b) @@ -1744,7 +1744,7 @@ type RowFilter_Interleave struct { // 6: far,blah,5,x // identical to #5 // // All interleaved filters are executed atomically. - Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters" json:"filters,omitempty"` + Filters []*RowFilter `protobuf:"bytes,1,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1754,7 +1754,7 @@ func (m *RowFilter_Interleave) Reset() { *m = RowFilter_Interleave{} } func (m *RowFilter_Interleave) String() string { return proto.CompactTextString(m) } func (*RowFilter_Interleave) ProtoMessage() {} func (*RowFilter_Interleave) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{9, 1} + return fileDescriptor_data_b0430eb990758496, []int{9, 1} } func (m *RowFilter_Interleave) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter_Interleave.Unmarshal(m, b) @@ -1791,14 +1791,14 @@ func (m *RowFilter_Interleave) GetFilters() []*RowFilter { type RowFilter_Condition struct { // If `predicate_filter` outputs any cells, then `true_filter` will be // evaluated on the input row. Otherwise, `false_filter` will be evaluated. - PredicateFilter *RowFilter `protobuf:"bytes,1,opt,name=predicate_filter,json=predicateFilter" json:"predicate_filter,omitempty"` + PredicateFilter *RowFilter `protobuf:"bytes,1,opt,name=predicate_filter,json=predicateFilter,proto3" json:"predicate_filter,omitempty"` // The filter to apply to the input row if `predicate_filter` returns any // results. If not provided, no results will be returned in the true case. - TrueFilter *RowFilter `protobuf:"bytes,2,opt,name=true_filter,json=trueFilter" json:"true_filter,omitempty"` + TrueFilter *RowFilter `protobuf:"bytes,2,opt,name=true_filter,json=trueFilter,proto3" json:"true_filter,omitempty"` // The filter to apply to the input row if `predicate_filter` does not // return any results. If not provided, no results will be returned in the // false case. - FalseFilter *RowFilter `protobuf:"bytes,3,opt,name=false_filter,json=falseFilter" json:"false_filter,omitempty"` + FalseFilter *RowFilter `protobuf:"bytes,3,opt,name=false_filter,json=falseFilter,proto3" json:"false_filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1808,7 +1808,7 @@ func (m *RowFilter_Condition) Reset() { *m = RowFilter_Condition{} } func (m *RowFilter_Condition) String() string { return proto.CompactTextString(m) } func (*RowFilter_Condition) ProtoMessage() {} func (*RowFilter_Condition) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{9, 2} + return fileDescriptor_data_b0430eb990758496, []int{9, 2} } func (m *RowFilter_Condition) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RowFilter_Condition.Unmarshal(m, b) @@ -1868,7 +1868,7 @@ func (m *Mutation) Reset() { *m = Mutation{} } func (m *Mutation) String() string { return proto.CompactTextString(m) } func (*Mutation) ProtoMessage() {} func (*Mutation) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{10} + return fileDescriptor_data_b0430eb990758496, []int{10} } func (m *Mutation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation.Unmarshal(m, b) @@ -1893,16 +1893,16 @@ type isMutation_Mutation interface { } type Mutation_SetCell_ struct { - SetCell *Mutation_SetCell `protobuf:"bytes,1,opt,name=set_cell,json=setCell,oneof"` + SetCell *Mutation_SetCell `protobuf:"bytes,1,opt,name=set_cell,json=setCell,proto3,oneof"` } type Mutation_DeleteFromColumn_ struct { - DeleteFromColumn *Mutation_DeleteFromColumn `protobuf:"bytes,2,opt,name=delete_from_column,json=deleteFromColumn,oneof"` + DeleteFromColumn *Mutation_DeleteFromColumn `protobuf:"bytes,2,opt,name=delete_from_column,json=deleteFromColumn,proto3,oneof"` } type Mutation_DeleteFromFamily_ struct { - DeleteFromFamily *Mutation_DeleteFromFamily `protobuf:"bytes,3,opt,name=delete_from_family,json=deleteFromFamily,oneof"` + DeleteFromFamily *Mutation_DeleteFromFamily `protobuf:"bytes,3,opt,name=delete_from_family,json=deleteFromFamily,proto3,oneof"` } type Mutation_DeleteFromRow_ struct { - DeleteFromRow *Mutation_DeleteFromRow `protobuf:"bytes,4,opt,name=delete_from_row,json=deleteFromRow,oneof"` + DeleteFromRow *Mutation_DeleteFromRow `protobuf:"bytes,4,opt,name=delete_from_row,json=deleteFromRow,proto3,oneof"` } func (*Mutation_SetCell_) isMutation_Mutation() {} @@ -2061,7 +2061,7 @@ func _Mutation_OneofSizer(msg proto.Message) (n int) { type Mutation_SetCell struct { // The name of the family into which new data should be written. // Must match `[-_.a-zA-Z0-9]+` - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The qualifier of the column into which new data should be written. // Can be any byte string, including the empty string. ColumnQualifier []byte `protobuf:"bytes,2,opt,name=column_qualifier,json=columnQualifier,proto3" json:"column_qualifier,omitempty"` @@ -2070,7 +2070,7 @@ type Mutation_SetCell struct { // Otherwise, the client should set this value itself, noting that the // default value is a timestamp of zero if the field is left unspecified. // Values must match the granularity of the table (e.g. micros, millis). - TimestampMicros int64 `protobuf:"varint,3,opt,name=timestamp_micros,json=timestampMicros" json:"timestamp_micros,omitempty"` + TimestampMicros int64 `protobuf:"varint,3,opt,name=timestamp_micros,json=timestampMicros,proto3" json:"timestamp_micros,omitempty"` // The value to be written into the specified cell. Value []byte `protobuf:"bytes,4,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -2082,7 +2082,7 @@ func (m *Mutation_SetCell) Reset() { *m = Mutation_SetCell{} } func (m *Mutation_SetCell) String() string { return proto.CompactTextString(m) } func (*Mutation_SetCell) ProtoMessage() {} func (*Mutation_SetCell) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{10, 0} + return fileDescriptor_data_b0430eb990758496, []int{10, 0} } func (m *Mutation_SetCell) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_SetCell.Unmarshal(m, b) @@ -2135,12 +2135,12 @@ func (m *Mutation_SetCell) GetValue() []byte { type Mutation_DeleteFromColumn struct { // The name of the family from which cells should be deleted. // Must match `[-_.a-zA-Z0-9]+` - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The qualifier of the column from which cells should be deleted. // Can be any byte string, including the empty string. ColumnQualifier []byte `protobuf:"bytes,2,opt,name=column_qualifier,json=columnQualifier,proto3" json:"column_qualifier,omitempty"` // The range of timestamps within which cells should be deleted. - TimeRange *TimestampRange `protobuf:"bytes,3,opt,name=time_range,json=timeRange" json:"time_range,omitempty"` + TimeRange *TimestampRange `protobuf:"bytes,3,opt,name=time_range,json=timeRange,proto3" json:"time_range,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2150,7 +2150,7 @@ func (m *Mutation_DeleteFromColumn) Reset() { *m = Mutation_DeleteFromCo func (m *Mutation_DeleteFromColumn) String() string { return proto.CompactTextString(m) } func (*Mutation_DeleteFromColumn) ProtoMessage() {} func (*Mutation_DeleteFromColumn) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{10, 1} + return fileDescriptor_data_b0430eb990758496, []int{10, 1} } func (m *Mutation_DeleteFromColumn) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_DeleteFromColumn.Unmarshal(m, b) @@ -2195,7 +2195,7 @@ func (m *Mutation_DeleteFromColumn) GetTimeRange() *TimestampRange { type Mutation_DeleteFromFamily struct { // The name of the family from which cells should be deleted. // Must match `[-_.a-zA-Z0-9]+` - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2205,7 +2205,7 @@ func (m *Mutation_DeleteFromFamily) Reset() { *m = Mutation_DeleteFromFa func (m *Mutation_DeleteFromFamily) String() string { return proto.CompactTextString(m) } func (*Mutation_DeleteFromFamily) ProtoMessage() {} func (*Mutation_DeleteFromFamily) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{10, 2} + return fileDescriptor_data_b0430eb990758496, []int{10, 2} } func (m *Mutation_DeleteFromFamily) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_DeleteFromFamily.Unmarshal(m, b) @@ -2243,7 +2243,7 @@ func (m *Mutation_DeleteFromRow) Reset() { *m = Mutation_DeleteFromRow{} func (m *Mutation_DeleteFromRow) String() string { return proto.CompactTextString(m) } func (*Mutation_DeleteFromRow) ProtoMessage() {} func (*Mutation_DeleteFromRow) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{10, 3} + return fileDescriptor_data_b0430eb990758496, []int{10, 3} } func (m *Mutation_DeleteFromRow) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Mutation_DeleteFromRow.Unmarshal(m, b) @@ -2268,7 +2268,7 @@ var xxx_messageInfo_Mutation_DeleteFromRow proto.InternalMessageInfo type ReadModifyWriteRule struct { // The name of the family to which the read/modify/write should be applied. // Must match `[-_.a-zA-Z0-9]+` - FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName" json:"family_name,omitempty"` + FamilyName string `protobuf:"bytes,1,opt,name=family_name,json=familyName,proto3" json:"family_name,omitempty"` // The qualifier of the column to which the read/modify/write should be // applied. // Can be any byte string, including the empty string. @@ -2289,7 +2289,7 @@ func (m *ReadModifyWriteRule) Reset() { *m = ReadModifyWriteRule{} } func (m *ReadModifyWriteRule) String() string { return proto.CompactTextString(m) } func (*ReadModifyWriteRule) ProtoMessage() {} func (*ReadModifyWriteRule) Descriptor() ([]byte, []int) { - return fileDescriptor_data_e6079216d37f14fa, []int{11} + return fileDescriptor_data_b0430eb990758496, []int{11} } func (m *ReadModifyWriteRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReadModifyWriteRule.Unmarshal(m, b) @@ -2317,7 +2317,7 @@ type ReadModifyWriteRule_AppendValue struct { AppendValue []byte `protobuf:"bytes,3,opt,name=append_value,json=appendValue,proto3,oneof"` } type ReadModifyWriteRule_IncrementAmount struct { - IncrementAmount int64 `protobuf:"varint,4,opt,name=increment_amount,json=incrementAmount,oneof"` + IncrementAmount int64 `protobuf:"varint,4,opt,name=increment_amount,json=incrementAmount,proto3,oneof"` } func (*ReadModifyWriteRule_AppendValue) isReadModifyWriteRule_Rule() {} @@ -2445,9 +2445,9 @@ func init() { proto.RegisterType((*ReadModifyWriteRule)(nil), "google.bigtable.v2.ReadModifyWriteRule") } -func init() { proto.RegisterFile("google/bigtable/v2/data.proto", fileDescriptor_data_e6079216d37f14fa) } +func init() { proto.RegisterFile("google/bigtable/v2/data.proto", fileDescriptor_data_b0430eb990758496) } -var fileDescriptor_data_e6079216d37f14fa = []byte{ +var fileDescriptor_data_b0430eb990758496 = []byte{ // 1444 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x57, 0xdb, 0x72, 0x1b, 0x45, 0x13, 0xd6, 0x5a, 0xb6, 0x2c, 0xf5, 0xca, 0x96, 0x32, 0x71, 0x1c, 0x45, 0x7f, 0xfc, 0xc7, 0xa5, diff --git a/vendor/google.golang.org/genproto/googleapis/bytestream/bytestream.pb.go b/vendor/google.golang.org/genproto/googleapis/bytestream/bytestream.pb.go index 026a67cae..acf2c230f 100644 --- a/vendor/google.golang.org/genproto/googleapis/bytestream/bytestream.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/bytestream/bytestream.pb.go @@ -28,13 +28,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Request object for ByteStream.Read. type ReadRequest struct { // The name of the resource to read. - ResourceName string `protobuf:"bytes,1,opt,name=resource_name,json=resourceName" json:"resource_name,omitempty"` + ResourceName string `protobuf:"bytes,1,opt,name=resource_name,json=resourceName,proto3" json:"resource_name,omitempty"` // The offset for the first byte to return in the read, relative to the start // of the resource. // // A `read_offset` that is negative or greater than the size of the resource // will cause an `OUT_OF_RANGE` error. - ReadOffset int64 `protobuf:"varint,2,opt,name=read_offset,json=readOffset" json:"read_offset,omitempty"` + ReadOffset int64 `protobuf:"varint,2,opt,name=read_offset,json=readOffset,proto3" json:"read_offset,omitempty"` // The maximum number of `data` bytes the server is allowed to return in the // sum of all `ReadResponse` messages. A `read_limit` of zero indicates that // there is no limit, and a negative `read_limit` will cause an error. @@ -42,7 +42,7 @@ type ReadRequest struct { // If the stream returns fewer bytes than allowed by the `read_limit` and no // error occurred, the stream includes all data from the `read_offset` to the // end of the resource. - ReadLimit int64 `protobuf:"varint,3,opt,name=read_limit,json=readLimit" json:"read_limit,omitempty"` + ReadLimit int64 `protobuf:"varint,3,opt,name=read_limit,json=readLimit,proto3" json:"read_limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -141,7 +141,7 @@ type WriteRequest struct { // The name of the resource to write. This **must** be set on the first // `WriteRequest` of each `Write()` action. If it is set on subsequent calls, // it **must** match the value of the first request. - ResourceName string `protobuf:"bytes,1,opt,name=resource_name,json=resourceName" json:"resource_name,omitempty"` + ResourceName string `protobuf:"bytes,1,opt,name=resource_name,json=resourceName,proto3" json:"resource_name,omitempty"` // The offset from the beginning of the resource at which the data should be // written. It is required on all `WriteRequest`s. // @@ -154,11 +154,11 @@ type WriteRequest struct { // sent previously on this stream. // // An incorrect value will cause an error. - WriteOffset int64 `protobuf:"varint,2,opt,name=write_offset,json=writeOffset" json:"write_offset,omitempty"` + WriteOffset int64 `protobuf:"varint,2,opt,name=write_offset,json=writeOffset,proto3" json:"write_offset,omitempty"` // If `true`, this indicates that the write is complete. Sending any // `WriteRequest`s subsequent to one in which `finish_write` is `true` will // cause an error. - FinishWrite bool `protobuf:"varint,3,opt,name=finish_write,json=finishWrite" json:"finish_write,omitempty"` + FinishWrite bool `protobuf:"varint,3,opt,name=finish_write,json=finishWrite,proto3" json:"finish_write,omitempty"` // A portion of the data for the resource. The client **may** leave `data` // empty for any given `WriteRequest`. This enables the client to inform the // service that the request is still live while it is running an operation to @@ -224,7 +224,7 @@ func (m *WriteRequest) GetData() []byte { // Response object for ByteStream.Write. type WriteResponse struct { // The number of bytes that have been processed for the given resource. - CommittedSize int64 `protobuf:"varint,1,opt,name=committed_size,json=committedSize" json:"committed_size,omitempty"` + CommittedSize int64 `protobuf:"varint,1,opt,name=committed_size,json=committedSize,proto3" json:"committed_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -264,7 +264,7 @@ func (m *WriteResponse) GetCommittedSize() int64 { // Request object for ByteStream.QueryWriteStatus. type QueryWriteStatusRequest struct { // The name of the resource whose write status is being requested. - ResourceName string `protobuf:"bytes,1,opt,name=resource_name,json=resourceName" json:"resource_name,omitempty"` + ResourceName string `protobuf:"bytes,1,opt,name=resource_name,json=resourceName,proto3" json:"resource_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -304,10 +304,10 @@ func (m *QueryWriteStatusRequest) GetResourceName() string { // Response object for ByteStream.QueryWriteStatus. type QueryWriteStatusResponse struct { // The number of bytes that have been processed for the given resource. - CommittedSize int64 `protobuf:"varint,1,opt,name=committed_size,json=committedSize" json:"committed_size,omitempty"` + CommittedSize int64 `protobuf:"varint,1,opt,name=committed_size,json=committedSize,proto3" json:"committed_size,omitempty"` // `complete` is `true` only if the client has sent a `WriteRequest` with // `finish_write` set to true, and the server has processed that request. - Complete bool `protobuf:"varint,2,opt,name=complete" json:"complete,omitempty"` + Complete bool `protobuf:"varint,2,opt,name=complete,proto3" json:"complete,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -499,8 +499,7 @@ func (c *byteStreamClient) QueryWriteStatus(ctx context.Context, in *QueryWriteS return out, nil } -// Server API for ByteStream service - +// ByteStreamServer is the server API for ByteStream service. type ByteStreamServer interface { // `Read()` is used to retrieve the contents of a resource as a sequence // of bytes. The bytes are returned in a sequence of responses, and the diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/audit/audit_log.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/audit/audit_log.pb.go index 8b4fd2624..9317dd533 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/audit/audit_log.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/audit/audit_log.pb.go @@ -26,51 +26,51 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type AuditLog struct { // The name of the API service performing the operation. For example, // `"datastore.googleapis.com"`. - ServiceName string `protobuf:"bytes,7,opt,name=service_name,json=serviceName" json:"service_name,omitempty"` + ServiceName string `protobuf:"bytes,7,opt,name=service_name,json=serviceName,proto3" json:"service_name,omitempty"` // The name of the service method or operation. // For API calls, this should be the name of the API method. // For example, // // "google.datastore.v1.Datastore.RunQuery" // "google.logging.v1.LoggingService.DeleteLog" - MethodName string `protobuf:"bytes,8,opt,name=method_name,json=methodName" json:"method_name,omitempty"` + MethodName string `protobuf:"bytes,8,opt,name=method_name,json=methodName,proto3" json:"method_name,omitempty"` // The resource or collection that is the target of the operation. // The name is a scheme-less URI, not including the API service name. // For example: // // "shelves/SHELF_ID/books" // "shelves/SHELF_ID/books/BOOK_ID" - ResourceName string `protobuf:"bytes,11,opt,name=resource_name,json=resourceName" json:"resource_name,omitempty"` + ResourceName string `protobuf:"bytes,11,opt,name=resource_name,json=resourceName,proto3" json:"resource_name,omitempty"` // The number of items returned from a List or Query API method, // if applicable. - NumResponseItems int64 `protobuf:"varint,12,opt,name=num_response_items,json=numResponseItems" json:"num_response_items,omitempty"` + NumResponseItems int64 `protobuf:"varint,12,opt,name=num_response_items,json=numResponseItems,proto3" json:"num_response_items,omitempty"` // The status of the overall operation. - Status *status.Status `protobuf:"bytes,2,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"` // Authentication information. - AuthenticationInfo *AuthenticationInfo `protobuf:"bytes,3,opt,name=authentication_info,json=authenticationInfo" json:"authentication_info,omitempty"` + AuthenticationInfo *AuthenticationInfo `protobuf:"bytes,3,opt,name=authentication_info,json=authenticationInfo,proto3" json:"authentication_info,omitempty"` // Authorization information. If there are multiple // resources or permissions involved, then there is // one AuthorizationInfo element for each {resource, permission} tuple. - AuthorizationInfo []*AuthorizationInfo `protobuf:"bytes,9,rep,name=authorization_info,json=authorizationInfo" json:"authorization_info,omitempty"` + AuthorizationInfo []*AuthorizationInfo `protobuf:"bytes,9,rep,name=authorization_info,json=authorizationInfo,proto3" json:"authorization_info,omitempty"` // Metadata about the operation. - RequestMetadata *RequestMetadata `protobuf:"bytes,4,opt,name=request_metadata,json=requestMetadata" json:"request_metadata,omitempty"` + RequestMetadata *RequestMetadata `protobuf:"bytes,4,opt,name=request_metadata,json=requestMetadata,proto3" json:"request_metadata,omitempty"` // The operation request. This may not include all request parameters, // such as those that are too large, privacy-sensitive, or duplicated // elsewhere in the log record. // It should never include user-generated data, such as file contents. // When the JSON object represented here has a proto equivalent, the proto // name will be indicated in the `@type` property. - Request *_struct.Struct `protobuf:"bytes,16,opt,name=request" json:"request,omitempty"` + Request *_struct.Struct `protobuf:"bytes,16,opt,name=request,proto3" json:"request,omitempty"` // The operation response. This may not include all response elements, // such as those that are too large, privacy-sensitive, or duplicated // elsewhere in the log record. // It should never include user-generated data, such as file contents. // When the JSON object represented here has a proto equivalent, the proto // name will be indicated in the `@type` property. - Response *_struct.Struct `protobuf:"bytes,17,opt,name=response" json:"response,omitempty"` + Response *_struct.Struct `protobuf:"bytes,17,opt,name=response,proto3" json:"response,omitempty"` // Other service-specific data about the request, response, and other // activities. - ServiceData *any.Any `protobuf:"bytes,15,opt,name=service_data,json=serviceData" json:"service_data,omitempty"` + ServiceData *any.Any `protobuf:"bytes,15,opt,name=service_data,json=serviceData,proto3" json:"service_data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -180,7 +180,7 @@ func (m *AuditLog) GetServiceData() *any.Any { // Authentication information for the operation. type AuthenticationInfo struct { // The email address of the authenticated user making the request. - PrincipalEmail string `protobuf:"bytes,1,opt,name=principal_email,json=principalEmail" json:"principal_email,omitempty"` + PrincipalEmail string `protobuf:"bytes,1,opt,name=principal_email,json=principalEmail,proto3" json:"principal_email,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -222,12 +222,12 @@ type AuthorizationInfo struct { // The resource being accessed, as a REST-style string. For example: // // bigquery.googlapis.com/projects/PROJECTID/datasets/DATASETID - Resource string `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource string `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` // The required IAM permission. - Permission string `protobuf:"bytes,2,opt,name=permission" json:"permission,omitempty"` + Permission string `protobuf:"bytes,2,opt,name=permission,proto3" json:"permission,omitempty"` // Whether or not authorization for `resource` and `permission` // was granted. - Granted bool `protobuf:"varint,3,opt,name=granted" json:"granted,omitempty"` + Granted bool `protobuf:"varint,3,opt,name=granted,proto3" json:"granted,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -281,7 +281,7 @@ func (m *AuthorizationInfo) GetGranted() bool { // Metadata about the request. type RequestMetadata struct { // The IP address of the caller. - CallerIp string `protobuf:"bytes,1,opt,name=caller_ip,json=callerIp" json:"caller_ip,omitempty"` + CallerIp string `protobuf:"bytes,1,opt,name=caller_ip,json=callerIp,proto3" json:"caller_ip,omitempty"` // The user agent of the caller. // This information is not authenticated and should be treated accordingly. // For example: @@ -292,7 +292,7 @@ type RequestMetadata struct { // The request was made by the Google Cloud SDK CLI (gcloud). // + `AppEngine-Google; (+http://code.google.com/appengine; appid: s~my-project`: // The request was made from the `my-project` App Engine app. - CallerSuppliedUserAgent string `protobuf:"bytes,2,opt,name=caller_supplied_user_agent,json=callerSuppliedUserAgent" json:"caller_supplied_user_agent,omitempty"` + CallerSuppliedUserAgent string `protobuf:"bytes,2,opt,name=caller_supplied_user_agent,json=callerSuppliedUserAgent,proto3" json:"caller_supplied_user_agent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/datatransfer.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/datatransfer.pb.go index 35cb14069..385eecf58 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/datatransfer.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/datatransfer.pb.go @@ -176,37 +176,37 @@ func (ListTransferRunsRequest_RunAttempt) EnumDescriptor() ([]byte, []int) { // or custom data source specific forms. type DataSourceParameter struct { // Parameter identifier. - ParamId string `protobuf:"bytes,1,opt,name=param_id,json=paramId" json:"param_id,omitempty"` + ParamId string `protobuf:"bytes,1,opt,name=param_id,json=paramId,proto3" json:"param_id,omitempty"` // Parameter display name in the user interface. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Parameter description. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Parameter type. - Type DataSourceParameter_Type `protobuf:"varint,4,opt,name=type,enum=google.cloud.bigquery.datatransfer.v1.DataSourceParameter_Type" json:"type,omitempty"` + Type DataSourceParameter_Type `protobuf:"varint,4,opt,name=type,proto3,enum=google.cloud.bigquery.datatransfer.v1.DataSourceParameter_Type" json:"type,omitempty"` // Is parameter required. - Required bool `protobuf:"varint,5,opt,name=required" json:"required,omitempty"` + Required bool `protobuf:"varint,5,opt,name=required,proto3" json:"required,omitempty"` // Can parameter have multiple values. - Repeated bool `protobuf:"varint,6,opt,name=repeated" json:"repeated,omitempty"` + Repeated bool `protobuf:"varint,6,opt,name=repeated,proto3" json:"repeated,omitempty"` // Regular expression which can be used for parameter validation. - ValidationRegex string `protobuf:"bytes,7,opt,name=validation_regex,json=validationRegex" json:"validation_regex,omitempty"` + ValidationRegex string `protobuf:"bytes,7,opt,name=validation_regex,json=validationRegex,proto3" json:"validation_regex,omitempty"` // All possible values for the parameter. - AllowedValues []string `protobuf:"bytes,8,rep,name=allowed_values,json=allowedValues" json:"allowed_values,omitempty"` + AllowedValues []string `protobuf:"bytes,8,rep,name=allowed_values,json=allowedValues,proto3" json:"allowed_values,omitempty"` // For integer and double values specifies minimum allowed value. - MinValue *wrappers.DoubleValue `protobuf:"bytes,9,opt,name=min_value,json=minValue" json:"min_value,omitempty"` + MinValue *wrappers.DoubleValue `protobuf:"bytes,9,opt,name=min_value,json=minValue,proto3" json:"min_value,omitempty"` // For integer and double values specifies maxminum allowed value. - MaxValue *wrappers.DoubleValue `protobuf:"bytes,10,opt,name=max_value,json=maxValue" json:"max_value,omitempty"` + MaxValue *wrappers.DoubleValue `protobuf:"bytes,10,opt,name=max_value,json=maxValue,proto3" json:"max_value,omitempty"` // When parameter is a record, describes child fields. - Fields []*DataSourceParameter `protobuf:"bytes,11,rep,name=fields" json:"fields,omitempty"` + Fields []*DataSourceParameter `protobuf:"bytes,11,rep,name=fields,proto3" json:"fields,omitempty"` // Description of the requirements for this field, in case the user input does // not fulfill the regex pattern or min/max values. - ValidationDescription string `protobuf:"bytes,12,opt,name=validation_description,json=validationDescription" json:"validation_description,omitempty"` + ValidationDescription string `protobuf:"bytes,12,opt,name=validation_description,json=validationDescription,proto3" json:"validation_description,omitempty"` // URL to a help document to further explain the naming requirements. - ValidationHelpUrl string `protobuf:"bytes,13,opt,name=validation_help_url,json=validationHelpUrl" json:"validation_help_url,omitempty"` + ValidationHelpUrl string `protobuf:"bytes,13,opt,name=validation_help_url,json=validationHelpUrl,proto3" json:"validation_help_url,omitempty"` // Cannot be changed after initial creation. - Immutable bool `protobuf:"varint,14,opt,name=immutable" json:"immutable,omitempty"` + Immutable bool `protobuf:"varint,14,opt,name=immutable,proto3" json:"immutable,omitempty"` // If set to true, schema should be taken from the parent with the same // parameter_id. Only applicable when parameter type is RECORD. - Recurse bool `protobuf:"varint,15,opt,name=recurse" json:"recurse,omitempty"` + Recurse bool `protobuf:"varint,15,opt,name=recurse,proto3" json:"recurse,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -345,58 +345,58 @@ func (m *DataSourceParameter) GetRecurse() bool { // render UI and request proper OAuth tokens. type DataSource struct { // Output only. Data source resource name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Data source id. - DataSourceId string `protobuf:"bytes,2,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"` + DataSourceId string `protobuf:"bytes,2,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"` // User friendly data source name. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // User friendly data source description string. - Description string `protobuf:"bytes,4,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` // Data source client id which should be used to receive refresh token. // When not supplied, no offline credentials are populated for data transfer. - ClientId string `protobuf:"bytes,5,opt,name=client_id,json=clientId" json:"client_id,omitempty"` + ClientId string `protobuf:"bytes,5,opt,name=client_id,json=clientId,proto3" json:"client_id,omitempty"` // Api auth scopes for which refresh token needs to be obtained. Only valid // when `client_id` is specified. Ignored otherwise. These are scopes needed // by a data source to prepare data and ingest them into BigQuery, // e.g., https://www.googleapis.com/auth/bigquery - Scopes []string `protobuf:"bytes,6,rep,name=scopes" json:"scopes,omitempty"` + Scopes []string `protobuf:"bytes,6,rep,name=scopes,proto3" json:"scopes,omitempty"` // Deprecated. This field has no effect. - TransferType TransferType `protobuf:"varint,7,opt,name=transfer_type,json=transferType,enum=google.cloud.bigquery.datatransfer.v1.TransferType" json:"transfer_type,omitempty"` + TransferType TransferType `protobuf:"varint,7,opt,name=transfer_type,json=transferType,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferType" json:"transfer_type,omitempty"` // Indicates whether the data source supports multiple transfers // to different BigQuery targets. - SupportsMultipleTransfers bool `protobuf:"varint,8,opt,name=supports_multiple_transfers,json=supportsMultipleTransfers" json:"supports_multiple_transfers,omitempty"` + SupportsMultipleTransfers bool `protobuf:"varint,8,opt,name=supports_multiple_transfers,json=supportsMultipleTransfers,proto3" json:"supports_multiple_transfers,omitempty"` // The number of seconds to wait for an update from the data source // before BigQuery marks the transfer as failed. - UpdateDeadlineSeconds int32 `protobuf:"varint,9,opt,name=update_deadline_seconds,json=updateDeadlineSeconds" json:"update_deadline_seconds,omitempty"` + UpdateDeadlineSeconds int32 `protobuf:"varint,9,opt,name=update_deadline_seconds,json=updateDeadlineSeconds,proto3" json:"update_deadline_seconds,omitempty"` // Default data transfer schedule. // Examples of valid schedules include: // `1st,3rd monday of month 15:30`, // `every wed,fri of jan,jun 13:15`, and // `first sunday of quarter 00:00`. - DefaultSchedule string `protobuf:"bytes,10,opt,name=default_schedule,json=defaultSchedule" json:"default_schedule,omitempty"` + DefaultSchedule string `protobuf:"bytes,10,opt,name=default_schedule,json=defaultSchedule,proto3" json:"default_schedule,omitempty"` // Specifies whether the data source supports a user defined schedule, or // operates on the default schedule. // When set to `true`, user can override default schedule. - SupportsCustomSchedule bool `protobuf:"varint,11,opt,name=supports_custom_schedule,json=supportsCustomSchedule" json:"supports_custom_schedule,omitempty"` + SupportsCustomSchedule bool `protobuf:"varint,11,opt,name=supports_custom_schedule,json=supportsCustomSchedule,proto3" json:"supports_custom_schedule,omitempty"` // Data source parameters. - Parameters []*DataSourceParameter `protobuf:"bytes,12,rep,name=parameters" json:"parameters,omitempty"` + Parameters []*DataSourceParameter `protobuf:"bytes,12,rep,name=parameters,proto3" json:"parameters,omitempty"` // Url for the help document for this data source. - HelpUrl string `protobuf:"bytes,13,opt,name=help_url,json=helpUrl" json:"help_url,omitempty"` + HelpUrl string `protobuf:"bytes,13,opt,name=help_url,json=helpUrl,proto3" json:"help_url,omitempty"` // Indicates the type of authorization. - AuthorizationType DataSource_AuthorizationType `protobuf:"varint,14,opt,name=authorization_type,json=authorizationType,enum=google.cloud.bigquery.datatransfer.v1.DataSource_AuthorizationType" json:"authorization_type,omitempty"` + AuthorizationType DataSource_AuthorizationType `protobuf:"varint,14,opt,name=authorization_type,json=authorizationType,proto3,enum=google.cloud.bigquery.datatransfer.v1.DataSource_AuthorizationType" json:"authorization_type,omitempty"` // Specifies whether the data source supports automatic data refresh for the // past few days, and how it's supported. // For some data sources, data might not be complete until a few days later, // so it's useful to refresh data automatically. - DataRefreshType DataSource_DataRefreshType `protobuf:"varint,15,opt,name=data_refresh_type,json=dataRefreshType,enum=google.cloud.bigquery.datatransfer.v1.DataSource_DataRefreshType" json:"data_refresh_type,omitempty"` + DataRefreshType DataSource_DataRefreshType `protobuf:"varint,15,opt,name=data_refresh_type,json=dataRefreshType,proto3,enum=google.cloud.bigquery.datatransfer.v1.DataSource_DataRefreshType" json:"data_refresh_type,omitempty"` // Default data refresh window on days. // Only meaningful when `data_refresh_type` = `SLIDING_WINDOW`. - DefaultDataRefreshWindowDays int32 `protobuf:"varint,16,opt,name=default_data_refresh_window_days,json=defaultDataRefreshWindowDays" json:"default_data_refresh_window_days,omitempty"` + DefaultDataRefreshWindowDays int32 `protobuf:"varint,16,opt,name=default_data_refresh_window_days,json=defaultDataRefreshWindowDays,proto3" json:"default_data_refresh_window_days,omitempty"` // Disables backfilling and manual run scheduling // for the data source. - ManualRunsDisabled bool `protobuf:"varint,17,opt,name=manual_runs_disabled,json=manualRunsDisabled" json:"manual_runs_disabled,omitempty"` + ManualRunsDisabled bool `protobuf:"varint,17,opt,name=manual_runs_disabled,json=manualRunsDisabled,proto3" json:"manual_runs_disabled,omitempty"` // The minimum interval for scheduler to schedule runs. - MinimumScheduleInterval *duration.Duration `protobuf:"bytes,18,opt,name=minimum_schedule_interval,json=minimumScheduleInterval" json:"minimum_schedule_interval,omitempty"` + MinimumScheduleInterval *duration.Duration `protobuf:"bytes,18,opt,name=minimum_schedule_interval,json=minimumScheduleInterval,proto3" json:"minimum_schedule_interval,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -556,7 +556,7 @@ func (m *DataSource) GetMinimumScheduleInterval() *duration.Duration { type GetDataSourceRequest struct { // The field will contain name of the resource requested, for example: // `projects/{project_id}/dataSources/{data_source_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -597,15 +597,15 @@ func (m *GetDataSourceRequest) GetName() string { type ListDataSourcesRequest struct { // The BigQuery project id for which data sources should be returned. // Must be in the form: `projects/{project_id}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Pagination token, which can be used to request a specific page // of `ListDataSourcesRequest` list results. For multiple-page // results, `ListDataSourcesResponse` outputs // a `next_page` token, which can be used as the // `page_token` value to request the next page of list results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Page size. The default page size is the maximum value of 1000 results. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -659,12 +659,12 @@ func (m *ListDataSourcesRequest) GetPageSize() int32 { // Returns list of supported data sources and their metadata. type ListDataSourcesResponse struct { // List of supported data sources and their transfer settings. - DataSources []*DataSource `protobuf:"bytes,1,rep,name=data_sources,json=dataSources" json:"data_sources,omitempty"` + DataSources []*DataSource `protobuf:"bytes,1,rep,name=data_sources,json=dataSources,proto3" json:"data_sources,omitempty"` // Output only. The next-pagination token. For multiple-page list results, // this token can be used as the // `ListDataSourcesRequest.page_token` // to request the next page of list results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -719,9 +719,9 @@ type CreateTransferConfigRequest struct { // Must be in the format /projects/{project_id}/locations/{location_id} // If specified location and location of the destination bigquery dataset // do not match - the request will fail. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Data transfer configuration to create. - TransferConfig *TransferConfig `protobuf:"bytes,2,opt,name=transfer_config,json=transferConfig" json:"transfer_config,omitempty"` + TransferConfig *TransferConfig `protobuf:"bytes,2,opt,name=transfer_config,json=transferConfig,proto3" json:"transfer_config,omitempty"` // Optional OAuth2 authorization code to use with this transfer configuration. // This is required if new credentials are needed, as indicated by // `CheckValidCreds`. @@ -738,7 +738,7 @@ type CreateTransferConfigRequest struct { // urn:ietf:wg:oauth:2.0:oob means that authorization code should be // returned in the title bar of the browser, with the page text prompting // the user to copy the code and paste it in the application. - AuthorizationCode string `protobuf:"bytes,3,opt,name=authorization_code,json=authorizationCode" json:"authorization_code,omitempty"` + AuthorizationCode string `protobuf:"bytes,3,opt,name=authorization_code,json=authorizationCode,proto3" json:"authorization_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -793,7 +793,7 @@ func (m *CreateTransferConfigRequest) GetAuthorizationCode() string { // transfer configuration, an authorization code needs to be provided. type UpdateTransferConfigRequest struct { // Data transfer configuration to create. - TransferConfig *TransferConfig `protobuf:"bytes,1,opt,name=transfer_config,json=transferConfig" json:"transfer_config,omitempty"` + TransferConfig *TransferConfig `protobuf:"bytes,1,opt,name=transfer_config,json=transferConfig,proto3" json:"transfer_config,omitempty"` // Optional OAuth2 authorization code to use with this transfer configuration. // If it is provided, the transfer configuration will be associated with the // authorizing user. @@ -810,9 +810,9 @@ type UpdateTransferConfigRequest struct { // urn:ietf:wg:oauth:2.0:oob means that authorization code should be // returned in the title bar of the browser, with the page text prompting // the user to copy the code and paste it in the application. - AuthorizationCode string `protobuf:"bytes,3,opt,name=authorization_code,json=authorizationCode" json:"authorization_code,omitempty"` + AuthorizationCode string `protobuf:"bytes,3,opt,name=authorization_code,json=authorizationCode,proto3" json:"authorization_code,omitempty"` // Required list of fields to be updated in this request. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -867,7 +867,7 @@ func (m *UpdateTransferConfigRequest) GetUpdateMask() *field_mask.FieldMask { type GetTransferConfigRequest struct { // The field will contain name of the resource requested, for example: // `projects/{project_id}/transferConfigs/{config_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -909,7 +909,7 @@ func (m *GetTransferConfigRequest) GetName() string { type DeleteTransferConfigRequest struct { // The field will contain name of the resource requested, for example: // `projects/{project_id}/transferConfigs/{config_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -950,7 +950,7 @@ func (m *DeleteTransferConfigRequest) GetName() string { type GetTransferRunRequest struct { // The field will contain name of the resource requested, for example: // `projects/{project_id}/transferConfigs/{config_id}/runs/{run_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -991,7 +991,7 @@ func (m *GetTransferRunRequest) GetName() string { type DeleteTransferRunRequest struct { // The field will contain name of the resource requested, for example: // `projects/{project_id}/transferConfigs/{config_id}/runs/{run_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1032,17 +1032,17 @@ func (m *DeleteTransferRunRequest) GetName() string { type ListTransferConfigsRequest struct { // The BigQuery project id for which data sources // should be returned: `projects/{project_id}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // When specified, only configurations of requested data sources are returned. - DataSourceIds []string `protobuf:"bytes,2,rep,name=data_source_ids,json=dataSourceIds" json:"data_source_ids,omitempty"` + DataSourceIds []string `protobuf:"bytes,2,rep,name=data_source_ids,json=dataSourceIds,proto3" json:"data_source_ids,omitempty"` // Pagination token, which can be used to request a specific page // of `ListTransfersRequest` list results. For multiple-page // results, `ListTransfersResponse` outputs // a `next_page` token, which can be used as the // `page_token` value to request the next page of list results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Page size. The default page size is the maximum value of 1000 results. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1103,12 +1103,12 @@ func (m *ListTransferConfigsRequest) GetPageSize() int32 { // The returned list of pipelines in the project. type ListTransferConfigsResponse struct { // Output only. The stored pipeline transfer configurations. - TransferConfigs []*TransferConfig `protobuf:"bytes,1,rep,name=transfer_configs,json=transferConfigs" json:"transfer_configs,omitempty"` + TransferConfigs []*TransferConfig `protobuf:"bytes,1,rep,name=transfer_configs,json=transferConfigs,proto3" json:"transfer_configs,omitempty"` // Output only. The next-pagination token. For multiple-page list results, // this token can be used as the // `ListTransferConfigsRequest.page_token` // to request the next page of list results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1159,19 +1159,19 @@ type ListTransferRunsRequest struct { // Name of transfer configuration for which transfer runs should be retrieved. // Format of transfer configuration resource name is: // `projects/{project_id}/transferConfigs/{config_id}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // When specified, only transfer runs with requested states are returned. - States []TransferState `protobuf:"varint,2,rep,packed,name=states,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"states,omitempty"` + States []TransferState `protobuf:"varint,2,rep,packed,name=states,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"states,omitempty"` // Pagination token, which can be used to request a specific page // of `ListTransferRunsRequest` list results. For multiple-page // results, `ListTransferRunsResponse` outputs // a `next_page` token, which can be used as the // `page_token` value to request the next page of list results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Page size. The default page size is the maximum value of 1000 results. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Indicates how run attempts are to be pulled. - RunAttempt ListTransferRunsRequest_RunAttempt `protobuf:"varint,5,opt,name=run_attempt,json=runAttempt,enum=google.cloud.bigquery.datatransfer.v1.ListTransferRunsRequest_RunAttempt" json:"run_attempt,omitempty"` + RunAttempt ListTransferRunsRequest_RunAttempt `protobuf:"varint,5,opt,name=run_attempt,json=runAttempt,proto3,enum=google.cloud.bigquery.datatransfer.v1.ListTransferRunsRequest_RunAttempt" json:"run_attempt,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1239,12 +1239,12 @@ func (m *ListTransferRunsRequest) GetRunAttempt() ListTransferRunsRequest_RunAtt // The returned list of pipelines in the project. type ListTransferRunsResponse struct { // Output only. The stored pipeline transfer runs. - TransferRuns []*TransferRun `protobuf:"bytes,1,rep,name=transfer_runs,json=transferRuns" json:"transfer_runs,omitempty"` + TransferRuns []*TransferRun `protobuf:"bytes,1,rep,name=transfer_runs,json=transferRuns,proto3" json:"transfer_runs,omitempty"` // Output only. The next-pagination token. For multiple-page list results, // this token can be used as the // `ListTransferRunsRequest.page_token` // to request the next page of list results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1292,18 +1292,18 @@ func (m *ListTransferRunsResponse) GetNextPageToken() string { type ListTransferLogsRequest struct { // Transfer run name in the form: // `projects/{project_id}/transferConfigs/{config_Id}/runs/{run_id}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Pagination token, which can be used to request a specific page // of `ListTransferLogsRequest` list results. For multiple-page // results, `ListTransferLogsResponse` outputs // a `next_page` token, which can be used as the // `page_token` value to request the next page of list results. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Page size. The default page size is the maximum value of 1000 results. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Message types to return. If not populated - INFO, WARNING and ERROR // messages are returned. - MessageTypes []TransferMessage_MessageSeverity `protobuf:"varint,6,rep,packed,name=message_types,json=messageTypes,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"message_types,omitempty"` + MessageTypes []TransferMessage_MessageSeverity `protobuf:"varint,6,rep,packed,name=message_types,json=messageTypes,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"message_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1364,12 +1364,12 @@ func (m *ListTransferLogsRequest) GetMessageTypes() []TransferMessage_MessageSev // The returned list transfer run messages. type ListTransferLogsResponse struct { // Output only. The stored pipeline transfer messages. - TransferMessages []*TransferMessage `protobuf:"bytes,1,rep,name=transfer_messages,json=transferMessages" json:"transfer_messages,omitempty"` + TransferMessages []*TransferMessage `protobuf:"bytes,1,rep,name=transfer_messages,json=transferMessages,proto3" json:"transfer_messages,omitempty"` // Output only. The next-pagination token. For multiple-page list results, // this token can be used as the // `GetTransferRunLogRequest.page_token` // to request the next page of list results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1422,7 +1422,7 @@ func (m *ListTransferLogsResponse) GetNextPageToken() string { type CheckValidCredsRequest struct { // The data source in the form: // `projects/{project_id}/dataSources/{data_source_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1462,7 +1462,7 @@ func (m *CheckValidCredsRequest) GetName() string { // A response indicating whether the credentials exist and are valid. type CheckValidCredsResponse struct { // If set to `true`, the credentials exist and are valid. - HasValidCreds bool `protobuf:"varint,1,opt,name=has_valid_creds,json=hasValidCreds" json:"has_valid_creds,omitempty"` + HasValidCreds bool `protobuf:"varint,1,opt,name=has_valid_creds,json=hasValidCreds,proto3" json:"has_valid_creds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1503,13 +1503,13 @@ func (m *CheckValidCredsResponse) GetHasValidCreds() bool { type ScheduleTransferRunsRequest struct { // Transfer configuration name in the form: // `projects/{project_id}/transferConfigs/{config_id}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Start time of the range of transfer runs. For example, // `"2017-05-25T00:00:00+00:00"`. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End time of the range of transfer runs. For example, // `"2017-05-30T00:00:00+00:00"`. - EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1563,7 +1563,7 @@ func (m *ScheduleTransferRunsRequest) GetEndTime() *timestamp.Timestamp { // A response to schedule transfer runs for a time range. type ScheduleTransferRunsResponse struct { // The transfer runs that were scheduled. - Runs []*TransferRun `protobuf:"bytes,1,rep,name=runs" json:"runs,omitempty"` + Runs []*TransferRun `protobuf:"bytes,1,rep,name=runs,proto3" json:"runs,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1805,8 +1805,7 @@ func (c *dataTransferServiceClient) CheckValidCreds(ctx context.Context, in *Che return out, nil } -// Server API for DataTransferService service - +// DataTransferServiceServer is the server API for DataTransferService service. type DataTransferServiceServer interface { // Retrieves a supported data source and returns its settings, // which can be used for UI rendering. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/transfer.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/transfer.pb.go index 67542b6c5..ee6d57341 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/transfer.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/transfer.pb.go @@ -143,15 +143,15 @@ type TransferConfig struct { // Where `config_id` is usually a uuid, even though it is not // guaranteed or required. The name is ignored when creating a transfer // config. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The BigQuery target dataset id. - DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"` + DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"` // User specified display name for the data transfer. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Data source id. Cannot be changed once data transfer is created. - DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"` + DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"` // Data transfer specific parameters. - Params *_struct.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"` + Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"` // Data transfer schedule. // If the data source does not support a custom schedule, this should be // empty. If it is empty, the default value for the data source will be @@ -164,31 +164,31 @@ type TransferConfig struct { // See more explanation about the format here: // https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format // NOTE: the granularity should be at least 8 hours, or less frequent. - Schedule string `protobuf:"bytes,7,opt,name=schedule" json:"schedule,omitempty"` + Schedule string `protobuf:"bytes,7,opt,name=schedule,proto3" json:"schedule,omitempty"` // The number of days to look back to automatically refresh the data. // For example, if `data_refresh_window_days = 10`, then every day // BigQuery reingests data for [today-10, today-1], rather than ingesting data // for just [today-1]. // Only valid if the data source supports the feature. Set the value to 0 // to use the default value. - DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays" json:"data_refresh_window_days,omitempty"` + DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays,proto3" json:"data_refresh_window_days,omitempty"` // Is this config disabled. When set to true, no runs are scheduled // for a given transfer. - Disabled bool `protobuf:"varint,13,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,13,opt,name=disabled,proto3" json:"disabled,omitempty"` // Output only. Data transfer modification time. Ignored by server on input. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // Output only. Next time when data transfer will run. - NextRunTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime" json:"next_run_time,omitempty"` + NextRunTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime,proto3" json:"next_run_time,omitempty"` // Output only. State of the most recently updated transfer run. - State TransferState `protobuf:"varint,10,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"` + State TransferState `protobuf:"varint,10,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"` // Output only. Unique ID of the user on whose behalf transfer is done. // Applicable only to data sources that do not support service accounts. // When set to 0, the data source service account credentials are used. // May be negative. Note, that this identifier is not stable. // It may change over time even for the same user. - UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"` + UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"` // Output only. Region in which BigQuery dataset is located. - DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion" json:"dataset_region,omitempty"` + DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion,proto3" json:"dataset_region,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -315,42 +315,42 @@ type TransferRun struct { // Transfer run names have the form // `projects/{project_id}/locations/{location}/transferConfigs/{config_id}/runs/{run_id}`. // The name is ignored when creating a transfer run. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Minimum time after which a transfer run can be started. - ScheduleTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime" json:"schedule_time,omitempty"` + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` // For batch transfer runs, specifies the date and time that // data should be ingested. - RunTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime" json:"run_time,omitempty"` + RunTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime,proto3" json:"run_time,omitempty"` // Status of the transfer run. - ErrorStatus *status.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus" json:"error_status,omitempty"` + ErrorStatus *status.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus,proto3" json:"error_status,omitempty"` // Output only. Time when transfer run was started. // Parameter ignored by server for input requests. - StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Output only. Time when transfer run ended. // Parameter ignored by server for input requests. - EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Output only. Last time the data transfer run state was updated. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // Output only. Data transfer specific parameters. - Params *_struct.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"` + Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"` // Output only. The BigQuery target dataset id. - DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"` + DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"` // Output only. Data source id. - DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"` + DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"` // Data transfer run state. Ignored for input requests. - State TransferState `protobuf:"varint,8,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"` + State TransferState `protobuf:"varint,8,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"` // Output only. Unique ID of the user on whose behalf transfer is done. // Applicable only to data sources that do not support service accounts. // When set to 0, the data source service account credentials are used. // May be negative. Note, that this identifier is not stable. // It may change over time even for the same user. - UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"` + UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"` // Output only. Describes the schedule of this transfer run if it was // created as part of a regular schedule. For batch transfer runs that are // scheduled manually, this is empty. // NOTE: the system might choose to delay the schedule depending on the // current load, so `schedule_time` doesn't always matches this. - Schedule string `protobuf:"bytes,12,opt,name=schedule" json:"schedule,omitempty"` + Schedule string `protobuf:"bytes,12,opt,name=schedule,proto3" json:"schedule,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -474,11 +474,11 @@ func (m *TransferRun) GetSchedule() string { // Represents a user facing message for a particular data transfer run. type TransferMessage struct { // Time when message was logged. - MessageTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime" json:"message_time,omitempty"` + MessageTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime,proto3" json:"message_time,omitempty"` // Message severity. - Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"` + Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"` // Message text. - MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText" json:"message_text,omitempty"` + MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText,proto3" json:"message_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/logging/v1/audit_data.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/logging/v1/audit_data.pb.go index 013312f32..39320dbc7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/logging/v1/audit_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/bigquery/logging/v1/audit_data.pb.go @@ -53,7 +53,7 @@ type AuditData struct { // *AuditData_JobQueryDoneResponse Response isAuditData_Response `protobuf_oneof:"response"` // A job completion event. - JobCompletedEvent *JobCompletedEvent `protobuf:"bytes,17,opt,name=job_completed_event,json=jobCompletedEvent" json:"job_completed_event,omitempty"` + JobCompletedEvent *JobCompletedEvent `protobuf:"bytes,17,opt,name=job_completed_event,json=jobCompletedEvent,proto3" json:"job_completed_event,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -91,55 +91,55 @@ type isAuditData_Response interface { } type AuditData_TableInsertRequest struct { - TableInsertRequest *TableInsertRequest `protobuf:"bytes,1,opt,name=table_insert_request,json=tableInsertRequest,oneof"` + TableInsertRequest *TableInsertRequest `protobuf:"bytes,1,opt,name=table_insert_request,json=tableInsertRequest,proto3,oneof"` } type AuditData_TableUpdateRequest struct { - TableUpdateRequest *TableUpdateRequest `protobuf:"bytes,16,opt,name=table_update_request,json=tableUpdateRequest,oneof"` + TableUpdateRequest *TableUpdateRequest `protobuf:"bytes,16,opt,name=table_update_request,json=tableUpdateRequest,proto3,oneof"` } type AuditData_DatasetListRequest struct { - DatasetListRequest *DatasetListRequest `protobuf:"bytes,2,opt,name=dataset_list_request,json=datasetListRequest,oneof"` + DatasetListRequest *DatasetListRequest `protobuf:"bytes,2,opt,name=dataset_list_request,json=datasetListRequest,proto3,oneof"` } type AuditData_DatasetInsertRequest struct { - DatasetInsertRequest *DatasetInsertRequest `protobuf:"bytes,3,opt,name=dataset_insert_request,json=datasetInsertRequest,oneof"` + DatasetInsertRequest *DatasetInsertRequest `protobuf:"bytes,3,opt,name=dataset_insert_request,json=datasetInsertRequest,proto3,oneof"` } type AuditData_DatasetUpdateRequest struct { - DatasetUpdateRequest *DatasetUpdateRequest `protobuf:"bytes,4,opt,name=dataset_update_request,json=datasetUpdateRequest,oneof"` + DatasetUpdateRequest *DatasetUpdateRequest `protobuf:"bytes,4,opt,name=dataset_update_request,json=datasetUpdateRequest,proto3,oneof"` } type AuditData_JobInsertRequest struct { - JobInsertRequest *JobInsertRequest `protobuf:"bytes,5,opt,name=job_insert_request,json=jobInsertRequest,oneof"` + JobInsertRequest *JobInsertRequest `protobuf:"bytes,5,opt,name=job_insert_request,json=jobInsertRequest,proto3,oneof"` } type AuditData_JobQueryRequest struct { - JobQueryRequest *JobQueryRequest `protobuf:"bytes,6,opt,name=job_query_request,json=jobQueryRequest,oneof"` + JobQueryRequest *JobQueryRequest `protobuf:"bytes,6,opt,name=job_query_request,json=jobQueryRequest,proto3,oneof"` } type AuditData_JobGetQueryResultsRequest struct { - JobGetQueryResultsRequest *JobGetQueryResultsRequest `protobuf:"bytes,7,opt,name=job_get_query_results_request,json=jobGetQueryResultsRequest,oneof"` + JobGetQueryResultsRequest *JobGetQueryResultsRequest `protobuf:"bytes,7,opt,name=job_get_query_results_request,json=jobGetQueryResultsRequest,proto3,oneof"` } type AuditData_TableDataListRequest struct { - TableDataListRequest *TableDataListRequest `protobuf:"bytes,8,opt,name=table_data_list_request,json=tableDataListRequest,oneof"` + TableDataListRequest *TableDataListRequest `protobuf:"bytes,8,opt,name=table_data_list_request,json=tableDataListRequest,proto3,oneof"` } type AuditData_TableInsertResponse struct { - TableInsertResponse *TableInsertResponse `protobuf:"bytes,9,opt,name=table_insert_response,json=tableInsertResponse,oneof"` + TableInsertResponse *TableInsertResponse `protobuf:"bytes,9,opt,name=table_insert_response,json=tableInsertResponse,proto3,oneof"` } type AuditData_TableUpdateResponse struct { - TableUpdateResponse *TableUpdateResponse `protobuf:"bytes,10,opt,name=table_update_response,json=tableUpdateResponse,oneof"` + TableUpdateResponse *TableUpdateResponse `protobuf:"bytes,10,opt,name=table_update_response,json=tableUpdateResponse,proto3,oneof"` } type AuditData_DatasetInsertResponse struct { - DatasetInsertResponse *DatasetInsertResponse `protobuf:"bytes,11,opt,name=dataset_insert_response,json=datasetInsertResponse,oneof"` + DatasetInsertResponse *DatasetInsertResponse `protobuf:"bytes,11,opt,name=dataset_insert_response,json=datasetInsertResponse,proto3,oneof"` } type AuditData_DatasetUpdateResponse struct { - DatasetUpdateResponse *DatasetUpdateResponse `protobuf:"bytes,12,opt,name=dataset_update_response,json=datasetUpdateResponse,oneof"` + DatasetUpdateResponse *DatasetUpdateResponse `protobuf:"bytes,12,opt,name=dataset_update_response,json=datasetUpdateResponse,proto3,oneof"` } type AuditData_JobInsertResponse struct { - JobInsertResponse *JobInsertResponse `protobuf:"bytes,18,opt,name=job_insert_response,json=jobInsertResponse,oneof"` + JobInsertResponse *JobInsertResponse `protobuf:"bytes,18,opt,name=job_insert_response,json=jobInsertResponse,proto3,oneof"` } type AuditData_JobQueryResponse struct { - JobQueryResponse *JobQueryResponse `protobuf:"bytes,13,opt,name=job_query_response,json=jobQueryResponse,oneof"` + JobQueryResponse *JobQueryResponse `protobuf:"bytes,13,opt,name=job_query_response,json=jobQueryResponse,proto3,oneof"` } type AuditData_JobGetQueryResultsResponse struct { - JobGetQueryResultsResponse *JobGetQueryResultsResponse `protobuf:"bytes,14,opt,name=job_get_query_results_response,json=jobGetQueryResultsResponse,oneof"` + JobGetQueryResultsResponse *JobGetQueryResultsResponse `protobuf:"bytes,14,opt,name=job_get_query_results_response,json=jobGetQueryResultsResponse,proto3,oneof"` } type AuditData_JobQueryDoneResponse struct { - JobQueryDoneResponse *JobQueryDoneResponse `protobuf:"bytes,15,opt,name=job_query_done_response,json=jobQueryDoneResponse,oneof"` + JobQueryDoneResponse *JobQueryDoneResponse `protobuf:"bytes,15,opt,name=job_query_done_response,json=jobQueryDoneResponse,proto3,oneof"` } func (*AuditData_TableInsertRequest) isAuditData_Request() {} @@ -673,7 +673,7 @@ func _AuditData_OneofSizer(msg proto.Message) (n int) { // Table insert request. type TableInsertRequest struct { // The new table. - Resource *Table `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Table `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -713,7 +713,7 @@ func (m *TableInsertRequest) GetResource() *Table { // Table update request. type TableUpdateRequest struct { // The table to be updated. - Resource *Table `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Table `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -753,7 +753,7 @@ func (m *TableUpdateRequest) GetResource() *Table { // Table insert response. type TableInsertResponse struct { // Final state of the inserted table. - Resource *Table `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Table `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -793,7 +793,7 @@ func (m *TableInsertResponse) GetResource() *Table { // Table update response. type TableUpdateResponse struct { // Final state of the updated table. - Resource *Table `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Table `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -833,7 +833,7 @@ func (m *TableUpdateResponse) GetResource() *Table { // Dataset list request. type DatasetListRequest struct { // Whether to list all datasets, including hidden ones. - ListAll bool `protobuf:"varint,1,opt,name=list_all,json=listAll" json:"list_all,omitempty"` + ListAll bool `protobuf:"varint,1,opt,name=list_all,json=listAll,proto3" json:"list_all,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -873,7 +873,7 @@ func (m *DatasetListRequest) GetListAll() bool { // Dataset insert request. type DatasetInsertRequest struct { // The dataset to be inserted. - Resource *Dataset `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Dataset `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -913,7 +913,7 @@ func (m *DatasetInsertRequest) GetResource() *Dataset { // Dataset insert response. type DatasetInsertResponse struct { // Final state of the inserted dataset. - Resource *Dataset `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Dataset `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -953,7 +953,7 @@ func (m *DatasetInsertResponse) GetResource() *Dataset { // Dataset update request. type DatasetUpdateRequest struct { // The dataset to be updated. - Resource *Dataset `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Dataset `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -993,7 +993,7 @@ func (m *DatasetUpdateRequest) GetResource() *Dataset { // Dataset update response. type DatasetUpdateResponse struct { // Final state of the updated dataset. - Resource *Dataset `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Dataset `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1033,7 +1033,7 @@ func (m *DatasetUpdateResponse) GetResource() *Dataset { // Job insert request. type JobInsertRequest struct { // Job insert request. - Resource *Job `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Job `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1073,7 +1073,7 @@ func (m *JobInsertRequest) GetResource() *Job { // Job insert response. type JobInsertResponse struct { // Job insert response. - Resource *Job `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource *Job `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1113,15 +1113,15 @@ func (m *JobInsertResponse) GetResource() *Job { // Job query request. type JobQueryRequest struct { // The query. - Query string `protobuf:"bytes,1,opt,name=query" json:"query,omitempty"` + Query string `protobuf:"bytes,1,opt,name=query,proto3" json:"query,omitempty"` // The maximum number of results. - MaxResults uint32 `protobuf:"varint,2,opt,name=max_results,json=maxResults" json:"max_results,omitempty"` + MaxResults uint32 `protobuf:"varint,2,opt,name=max_results,json=maxResults,proto3" json:"max_results,omitempty"` // The default dataset for tables that do not have a dataset specified. - DefaultDataset *DatasetName `protobuf:"bytes,3,opt,name=default_dataset,json=defaultDataset" json:"default_dataset,omitempty"` + DefaultDataset *DatasetName `protobuf:"bytes,3,opt,name=default_dataset,json=defaultDataset,proto3" json:"default_dataset,omitempty"` // Project that the query should be charged to. - ProjectId string `protobuf:"bytes,4,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,4,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // If true, don't actually run the job. Just check that it would run. - DryRun bool `protobuf:"varint,5,opt,name=dry_run,json=dryRun" json:"dry_run,omitempty"` + DryRun bool `protobuf:"varint,5,opt,name=dry_run,json=dryRun,proto3" json:"dry_run,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1189,9 +1189,9 @@ func (m *JobQueryRequest) GetDryRun() bool { // Job query response. type JobQueryResponse struct { // The total number of rows in the full query result set. - TotalResults uint64 `protobuf:"varint,1,opt,name=total_results,json=totalResults" json:"total_results,omitempty"` + TotalResults uint64 `protobuf:"varint,1,opt,name=total_results,json=totalResults,proto3" json:"total_results,omitempty"` // Information about the queried job. - Job *Job `protobuf:"bytes,2,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1238,9 +1238,9 @@ func (m *JobQueryResponse) GetJob() *Job { // Job getQueryResults request. type JobGetQueryResultsRequest struct { // Maximum number of results to return. - MaxResults uint32 `protobuf:"varint,1,opt,name=max_results,json=maxResults" json:"max_results,omitempty"` + MaxResults uint32 `protobuf:"varint,1,opt,name=max_results,json=maxResults,proto3" json:"max_results,omitempty"` // Zero-based row number at which to start. - StartRow uint64 `protobuf:"varint,2,opt,name=start_row,json=startRow" json:"start_row,omitempty"` + StartRow uint64 `protobuf:"varint,2,opt,name=start_row,json=startRow,proto3" json:"start_row,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1287,11 +1287,11 @@ func (m *JobGetQueryResultsRequest) GetStartRow() uint64 { // Job getQueryResults response. type JobGetQueryResultsResponse struct { // Total number of results in query results. - TotalResults uint64 `protobuf:"varint,1,opt,name=total_results,json=totalResults" json:"total_results,omitempty"` + TotalResults uint64 `protobuf:"varint,1,opt,name=total_results,json=totalResults,proto3" json:"total_results,omitempty"` // The job that was created to run the query. // It completed if `job.status.state` is `DONE`. // It failed if `job.status.errorResult` is also present. - Job *Job `protobuf:"bytes,2,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1339,7 +1339,7 @@ func (m *JobGetQueryResultsResponse) GetJob() *Job { type JobQueryDoneResponse struct { // The job and status information. // The job completed if `job.status.state` is `DONE`. - Job *Job `protobuf:"bytes,1,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,1,opt,name=job,proto3" json:"job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1379,9 +1379,9 @@ func (m *JobQueryDoneResponse) GetJob() *Job { // Query job completed event. type JobCompletedEvent struct { // Name of the event. - EventName string `protobuf:"bytes,1,opt,name=event_name,json=eventName" json:"event_name,omitempty"` + EventName string `protobuf:"bytes,1,opt,name=event_name,json=eventName,proto3" json:"event_name,omitempty"` // Job information. - Job *Job `protobuf:"bytes,2,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1428,9 +1428,9 @@ func (m *JobCompletedEvent) GetJob() *Job { // Table data-list request. type TableDataListRequest struct { // Starting row offset. - StartRow uint64 `protobuf:"varint,1,opt,name=start_row,json=startRow" json:"start_row,omitempty"` + StartRow uint64 `protobuf:"varint,1,opt,name=start_row,json=startRow,proto3" json:"start_row,omitempty"` // Maximum number of results to return. - MaxResults uint32 `protobuf:"varint,2,opt,name=max_results,json=maxResults" json:"max_results,omitempty"` + MaxResults uint32 `protobuf:"varint,2,opt,name=max_results,json=maxResults,proto3" json:"max_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1482,22 +1482,22 @@ func (m *TableDataListRequest) GetMaxResults() uint32 { // transition. type Table struct { // The name of the table. - TableName *TableName `protobuf:"bytes,1,opt,name=table_name,json=tableName" json:"table_name,omitempty"` + TableName *TableName `protobuf:"bytes,1,opt,name=table_name,json=tableName,proto3" json:"table_name,omitempty"` // User-provided metadata for the table. - Info *TableInfo `protobuf:"bytes,2,opt,name=info" json:"info,omitempty"` + Info *TableInfo `protobuf:"bytes,2,opt,name=info,proto3" json:"info,omitempty"` // A JSON representation of the table's schema. - SchemaJson string `protobuf:"bytes,8,opt,name=schema_json,json=schemaJson" json:"schema_json,omitempty"` + SchemaJson string `protobuf:"bytes,8,opt,name=schema_json,json=schemaJson,proto3" json:"schema_json,omitempty"` // If present, this is a virtual table defined by a SQL query. - View *TableViewDefinition `protobuf:"bytes,4,opt,name=view" json:"view,omitempty"` + View *TableViewDefinition `protobuf:"bytes,4,opt,name=view,proto3" json:"view,omitempty"` // The expiration date for the table, after which the table // is deleted and the storage reclaimed. // If not present, the table persists indefinitely. - ExpireTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` // The time the table was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The time the table was last truncated // by an operation with a `writeDisposition` of `WRITE_TRUNCATE`. - TruncateTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=truncate_time,json=truncateTime" json:"truncate_time,omitempty"` + TruncateTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=truncate_time,json=truncateTime,proto3" json:"truncate_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1579,10 +1579,10 @@ func (m *Table) GetTruncateTime() *timestamp.Timestamp { // User-provided metadata for a table. type TableInfo struct { // A short name for the table, such as`"Analytics Data - Jan 2011"`. - FriendlyName string `protobuf:"bytes,1,opt,name=friendly_name,json=friendlyName" json:"friendly_name,omitempty"` + FriendlyName string `protobuf:"bytes,1,opt,name=friendly_name,json=friendlyName,proto3" json:"friendly_name,omitempty"` // A long description, perhaps several paragraphs, // describing the table contents in detail. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1629,7 +1629,7 @@ func (m *TableInfo) GetDescription() string { // Describes a virtual table defined by a SQL query. type TableViewDefinition struct { // SQL query defining the view. - Query string `protobuf:"bytes,1,opt,name=query" json:"query,omitempty"` + Query string `protobuf:"bytes,1,opt,name=query,proto3" json:"query,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1671,20 +1671,20 @@ func (m *TableViewDefinition) GetQuery() string { // for more details on individual fields. type Dataset struct { // The name of the dataset. - DatasetName *DatasetName `protobuf:"bytes,1,opt,name=dataset_name,json=datasetName" json:"dataset_name,omitempty"` + DatasetName *DatasetName `protobuf:"bytes,1,opt,name=dataset_name,json=datasetName,proto3" json:"dataset_name,omitempty"` // User-provided metadata for the dataset. - Info *DatasetInfo `protobuf:"bytes,2,opt,name=info" json:"info,omitempty"` + Info *DatasetInfo `protobuf:"bytes,2,opt,name=info,proto3" json:"info,omitempty"` // The time the dataset was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The time the dataset was last modified. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // The access control list for the dataset. - Acl *BigQueryAcl `protobuf:"bytes,6,opt,name=acl" json:"acl,omitempty"` + Acl *BigQueryAcl `protobuf:"bytes,6,opt,name=acl,proto3" json:"acl,omitempty"` // If this field is present, each table that does not specify an // expiration time is assigned an expiration time by adding this // duration to the table's `createTime`. If this field is empty, // there is no default table expiration time. - DefaultTableExpireDuration *duration.Duration `protobuf:"bytes,8,opt,name=default_table_expire_duration,json=defaultTableExpireDuration" json:"default_table_expire_duration,omitempty"` + DefaultTableExpireDuration *duration.Duration `protobuf:"bytes,8,opt,name=default_table_expire_duration,json=defaultTableExpireDuration,proto3" json:"default_table_expire_duration,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1759,10 +1759,10 @@ func (m *Dataset) GetDefaultTableExpireDuration() *duration.Duration { // User-provided metadata for a dataset. type DatasetInfo struct { // A short name for the dataset, such as`"Analytics Data 2011"`. - FriendlyName string `protobuf:"bytes,1,opt,name=friendly_name,json=friendlyName" json:"friendly_name,omitempty"` + FriendlyName string `protobuf:"bytes,1,opt,name=friendly_name,json=friendlyName,proto3" json:"friendly_name,omitempty"` // A long description, perhaps several paragraphs, // describing the dataset contents in detail. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1809,7 +1809,7 @@ func (m *DatasetInfo) GetDescription() string { // An access control list. type BigQueryAcl struct { // Access control entry list. - Entries []*BigQueryAcl_Entry `protobuf:"bytes,1,rep,name=entries" json:"entries,omitempty"` + Entries []*BigQueryAcl_Entry `protobuf:"bytes,1,rep,name=entries,proto3" json:"entries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1849,18 +1849,18 @@ func (m *BigQueryAcl) GetEntries() []*BigQueryAcl_Entry { // Access control entry. type BigQueryAcl_Entry struct { // The granted role, which can be `READER`, `WRITER`, or `OWNER`. - Role string `protobuf:"bytes,1,opt,name=role" json:"role,omitempty"` + Role string `protobuf:"bytes,1,opt,name=role,proto3" json:"role,omitempty"` // Grants access to a group identified by an email address. - GroupEmail string `protobuf:"bytes,2,opt,name=group_email,json=groupEmail" json:"group_email,omitempty"` + GroupEmail string `protobuf:"bytes,2,opt,name=group_email,json=groupEmail,proto3" json:"group_email,omitempty"` // Grants access to a user identified by an email address. - UserEmail string `protobuf:"bytes,3,opt,name=user_email,json=userEmail" json:"user_email,omitempty"` + UserEmail string `protobuf:"bytes,3,opt,name=user_email,json=userEmail,proto3" json:"user_email,omitempty"` // Grants access to all members of a domain. - Domain string `protobuf:"bytes,4,opt,name=domain" json:"domain,omitempty"` + Domain string `protobuf:"bytes,4,opt,name=domain,proto3" json:"domain,omitempty"` // Grants access to special groups. Valid groups are `PROJECT_OWNERS`, // `PROJECT_READERS`, `PROJECT_WRITERS` and `ALL_AUTHENTICATED_USERS`. - SpecialGroup string `protobuf:"bytes,5,opt,name=special_group,json=specialGroup" json:"special_group,omitempty"` + SpecialGroup string `protobuf:"bytes,5,opt,name=special_group,json=specialGroup,proto3" json:"special_group,omitempty"` // Grants access to a BigQuery View. - ViewName *TableName `protobuf:"bytes,6,opt,name=view_name,json=viewName" json:"view_name,omitempty"` + ViewName *TableName `protobuf:"bytes,6,opt,name=view_name,json=viewName,proto3" json:"view_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1935,13 +1935,13 @@ func (m *BigQueryAcl_Entry) GetViewName() *TableName { // Describes a job. type Job struct { // Job name. - JobName *JobName `protobuf:"bytes,1,opt,name=job_name,json=jobName" json:"job_name,omitempty"` + JobName *JobName `protobuf:"bytes,1,opt,name=job_name,json=jobName,proto3" json:"job_name,omitempty"` // Job configuration. - JobConfiguration *JobConfiguration `protobuf:"bytes,2,opt,name=job_configuration,json=jobConfiguration" json:"job_configuration,omitempty"` + JobConfiguration *JobConfiguration `protobuf:"bytes,2,opt,name=job_configuration,json=jobConfiguration,proto3" json:"job_configuration,omitempty"` // Job status. - JobStatus *JobStatus `protobuf:"bytes,3,opt,name=job_status,json=jobStatus" json:"job_status,omitempty"` + JobStatus *JobStatus `protobuf:"bytes,3,opt,name=job_status,json=jobStatus,proto3" json:"job_status,omitempty"` // Job statistics. - JobStatistics *JobStatistics `protobuf:"bytes,4,opt,name=job_statistics,json=jobStatistics" json:"job_statistics,omitempty"` + JobStatistics *JobStatistics `protobuf:"bytes,4,opt,name=job_statistics,json=jobStatistics,proto3" json:"job_statistics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2012,7 +2012,7 @@ type JobConfiguration struct { // *JobConfiguration_TableCopy_ Configuration isJobConfiguration_Configuration `protobuf_oneof:"configuration"` // If true, don't actually run the job. Just check that it would run. - DryRun bool `protobuf:"varint,9,opt,name=dry_run,json=dryRun" json:"dry_run,omitempty"` + DryRun bool `protobuf:"varint,9,opt,name=dry_run,json=dryRun,proto3" json:"dry_run,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2047,16 +2047,16 @@ type isJobConfiguration_Configuration interface { } type JobConfiguration_Query_ struct { - Query *JobConfiguration_Query `protobuf:"bytes,5,opt,name=query,oneof"` + Query *JobConfiguration_Query `protobuf:"bytes,5,opt,name=query,proto3,oneof"` } type JobConfiguration_Load_ struct { - Load *JobConfiguration_Load `protobuf:"bytes,6,opt,name=load,oneof"` + Load *JobConfiguration_Load `protobuf:"bytes,6,opt,name=load,proto3,oneof"` } type JobConfiguration_Extract_ struct { - Extract *JobConfiguration_Extract `protobuf:"bytes,7,opt,name=extract,oneof"` + Extract *JobConfiguration_Extract `protobuf:"bytes,7,opt,name=extract,proto3,oneof"` } type JobConfiguration_TableCopy_ struct { - TableCopy *JobConfiguration_TableCopy `protobuf:"bytes,8,opt,name=table_copy,json=tableCopy,oneof"` + TableCopy *JobConfiguration_TableCopy `protobuf:"bytes,8,opt,name=table_copy,json=tableCopy,proto3,oneof"` } func (*JobConfiguration_Query_) isJobConfiguration_Configuration() {} @@ -2221,20 +2221,20 @@ func _JobConfiguration_OneofSizer(msg proto.Message) (n int) { // Describes a query job, which executes a SQL-like query. type JobConfiguration_Query struct { // The SQL query to run. - Query string `protobuf:"bytes,1,opt,name=query" json:"query,omitempty"` + Query string `protobuf:"bytes,1,opt,name=query,proto3" json:"query,omitempty"` // The table where results are written. - DestinationTable *TableName `protobuf:"bytes,2,opt,name=destination_table,json=destinationTable" json:"destination_table,omitempty"` + DestinationTable *TableName `protobuf:"bytes,2,opt,name=destination_table,json=destinationTable,proto3" json:"destination_table,omitempty"` // Describes when a job is allowed to create a table: // `CREATE_IF_NEEDED`, `CREATE_NEVER`. - CreateDisposition string `protobuf:"bytes,3,opt,name=create_disposition,json=createDisposition" json:"create_disposition,omitempty"` + CreateDisposition string `protobuf:"bytes,3,opt,name=create_disposition,json=createDisposition,proto3" json:"create_disposition,omitempty"` // Describes how writes affect existing tables: // `WRITE_TRUNCATE`, `WRITE_APPEND`, `WRITE_EMPTY`. - WriteDisposition string `protobuf:"bytes,4,opt,name=write_disposition,json=writeDisposition" json:"write_disposition,omitempty"` + WriteDisposition string `protobuf:"bytes,4,opt,name=write_disposition,json=writeDisposition,proto3" json:"write_disposition,omitempty"` // If a table name is specified without a dataset in a query, // this dataset will be added to table name. - DefaultDataset *DatasetName `protobuf:"bytes,5,opt,name=default_dataset,json=defaultDataset" json:"default_dataset,omitempty"` + DefaultDataset *DatasetName `protobuf:"bytes,5,opt,name=default_dataset,json=defaultDataset,proto3" json:"default_dataset,omitempty"` // Describes data sources outside BigQuery, if needed. - TableDefinitions []*TableDefinition `protobuf:"bytes,6,rep,name=table_definitions,json=tableDefinitions" json:"table_definitions,omitempty"` + TableDefinitions []*TableDefinition `protobuf:"bytes,6,rep,name=table_definitions,json=tableDefinitions,proto3" json:"table_definitions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2311,17 +2311,17 @@ func (m *JobConfiguration_Query) GetTableDefinitions() []*TableDefinition { type JobConfiguration_Load struct { // URIs for the data to be imported. Only Google Cloud Storage URIs are // supported. - SourceUris []string `protobuf:"bytes,1,rep,name=source_uris,json=sourceUris" json:"source_uris,omitempty"` + SourceUris []string `protobuf:"bytes,1,rep,name=source_uris,json=sourceUris,proto3" json:"source_uris,omitempty"` // The table schema in JSON format representation of a TableSchema. - SchemaJson string `protobuf:"bytes,6,opt,name=schema_json,json=schemaJson" json:"schema_json,omitempty"` + SchemaJson string `protobuf:"bytes,6,opt,name=schema_json,json=schemaJson,proto3" json:"schema_json,omitempty"` // The table where the imported data is written. - DestinationTable *TableName `protobuf:"bytes,3,opt,name=destination_table,json=destinationTable" json:"destination_table,omitempty"` + DestinationTable *TableName `protobuf:"bytes,3,opt,name=destination_table,json=destinationTable,proto3" json:"destination_table,omitempty"` // Describes when a job is allowed to create a table: // `CREATE_IF_NEEDED`, `CREATE_NEVER`. - CreateDisposition string `protobuf:"bytes,4,opt,name=create_disposition,json=createDisposition" json:"create_disposition,omitempty"` + CreateDisposition string `protobuf:"bytes,4,opt,name=create_disposition,json=createDisposition,proto3" json:"create_disposition,omitempty"` // Describes how writes affect existing tables: // `WRITE_TRUNCATE`, `WRITE_APPEND`, `WRITE_EMPTY`. - WriteDisposition string `protobuf:"bytes,5,opt,name=write_disposition,json=writeDisposition" json:"write_disposition,omitempty"` + WriteDisposition string `protobuf:"bytes,5,opt,name=write_disposition,json=writeDisposition,proto3" json:"write_disposition,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2390,9 +2390,9 @@ func (m *JobConfiguration_Load) GetWriteDisposition() string { // via the export pipeline. type JobConfiguration_Extract struct { // Google Cloud Storage URIs where extracted data should be written. - DestinationUris []string `protobuf:"bytes,1,rep,name=destination_uris,json=destinationUris" json:"destination_uris,omitempty"` + DestinationUris []string `protobuf:"bytes,1,rep,name=destination_uris,json=destinationUris,proto3" json:"destination_uris,omitempty"` // The source table. - SourceTable *TableName `protobuf:"bytes,2,opt,name=source_table,json=sourceTable" json:"source_table,omitempty"` + SourceTable *TableName `protobuf:"bytes,2,opt,name=source_table,json=sourceTable,proto3" json:"source_table,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2439,15 +2439,15 @@ func (m *JobConfiguration_Extract) GetSourceTable() *TableName { // Describes a copy job, which copies an existing table to another table. type JobConfiguration_TableCopy struct { // Source tables. - SourceTables []*TableName `protobuf:"bytes,1,rep,name=source_tables,json=sourceTables" json:"source_tables,omitempty"` + SourceTables []*TableName `protobuf:"bytes,1,rep,name=source_tables,json=sourceTables,proto3" json:"source_tables,omitempty"` // Destination table. - DestinationTable *TableName `protobuf:"bytes,2,opt,name=destination_table,json=destinationTable" json:"destination_table,omitempty"` + DestinationTable *TableName `protobuf:"bytes,2,opt,name=destination_table,json=destinationTable,proto3" json:"destination_table,omitempty"` // Describes when a job is allowed to create a table: // `CREATE_IF_NEEDED`, `CREATE_NEVER`. - CreateDisposition string `protobuf:"bytes,3,opt,name=create_disposition,json=createDisposition" json:"create_disposition,omitempty"` + CreateDisposition string `protobuf:"bytes,3,opt,name=create_disposition,json=createDisposition,proto3" json:"create_disposition,omitempty"` // Describes how writes affect existing tables: // `WRITE_TRUNCATE`, `WRITE_APPEND`, `WRITE_EMPTY`. - WriteDisposition string `protobuf:"bytes,4,opt,name=write_disposition,json=writeDisposition" json:"write_disposition,omitempty"` + WriteDisposition string `protobuf:"bytes,4,opt,name=write_disposition,json=writeDisposition,proto3" json:"write_disposition,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2508,9 +2508,9 @@ func (m *JobConfiguration_TableCopy) GetWriteDisposition() string { // Describes an external data source used in a query. type TableDefinition struct { // Name of the table, used in queries. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Google Cloud Storage URIs for the data to be imported. - SourceUris []string `protobuf:"bytes,2,rep,name=source_uris,json=sourceUris" json:"source_uris,omitempty"` + SourceUris []string `protobuf:"bytes,2,rep,name=source_uris,json=sourceUris,proto3" json:"source_uris,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2557,9 +2557,9 @@ func (m *TableDefinition) GetSourceUris() []string { // Running state of a job. type JobStatus struct { // State of a job: `PENDING`, `RUNNING`, or `DONE`. - State string `protobuf:"bytes,1,opt,name=state" json:"state,omitempty"` + State string `protobuf:"bytes,1,opt,name=state,proto3" json:"state,omitempty"` // If the job did not complete successfully, this field describes why. - Error *status.Status `protobuf:"bytes,2,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2606,17 +2606,17 @@ func (m *JobStatus) GetError() *status.Status { // Job statistics that may change after a job starts. type JobStatistics struct { // Time when the job was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time when the job started. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time when the job ended. - EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Total bytes processed for a job. - TotalProcessedBytes int64 `protobuf:"varint,4,opt,name=total_processed_bytes,json=totalProcessedBytes" json:"total_processed_bytes,omitempty"` + TotalProcessedBytes int64 `protobuf:"varint,4,opt,name=total_processed_bytes,json=totalProcessedBytes,proto3" json:"total_processed_bytes,omitempty"` // Processed bytes, adjusted by the job's CPU usage. - TotalBilledBytes int64 `protobuf:"varint,5,opt,name=total_billed_bytes,json=totalBilledBytes" json:"total_billed_bytes,omitempty"` + TotalBilledBytes int64 `protobuf:"varint,5,opt,name=total_billed_bytes,json=totalBilledBytes,proto3" json:"total_billed_bytes,omitempty"` // The tier assigned by CPU-based billing. - BillingTier int32 `protobuf:"varint,7,opt,name=billing_tier,json=billingTier" json:"billing_tier,omitempty"` + BillingTier int32 `protobuf:"varint,7,opt,name=billing_tier,json=billingTier,proto3" json:"billing_tier,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2691,9 +2691,9 @@ func (m *JobStatistics) GetBillingTier() int32 { // The fully-qualified name for a dataset. type DatasetName struct { // The project ID. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The dataset ID within the project. - DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2740,11 +2740,11 @@ func (m *DatasetName) GetDatasetId() string { // The fully-qualified name for a table. type TableName struct { // The project ID. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The dataset ID within the project. - DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The table ID of the table within the dataset. - TableId string `protobuf:"bytes,3,opt,name=table_id,json=tableId" json:"table_id,omitempty"` + TableId string `protobuf:"bytes,3,opt,name=table_id,json=tableId,proto3" json:"table_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2798,9 +2798,9 @@ func (m *TableName) GetTableId() string { // The fully-qualified name for a job. type JobName struct { // The project ID. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The job ID within the project. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/billing/v1/cloud_billing.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/billing/v1/cloud_billing.pb.go index 8453248f7..d6ff15b7e 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/billing/v1/cloud_billing.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/billing/v1/cloud_billing.pb.go @@ -32,14 +32,14 @@ type BillingAccount struct { // `billingAccounts/{billing_account_id}`. For example, // `billingAccounts/012345-567890-ABCDEF` would be the resource name for // billing account `012345-567890-ABCDEF`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // True if the billing account is open, and will therefore be charged for any // usage on associated projects. False if the billing account is closed, and // therefore projects associated with it will be unable to use paid services. - Open bool `protobuf:"varint,2,opt,name=open" json:"open,omitempty"` + Open bool `protobuf:"varint,2,opt,name=open,proto3" json:"open,omitempty"` // The display name given to the billing account, such as `My Billing // Account`. This name is displayed in the Google Cloud Console. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -98,19 +98,19 @@ type ProjectBillingInfo struct { // `projects/{project_id}/billingInfo`. For example, the resource name for the // billing information for project `tokyo-rain-123` would be // `projects/tokyo-rain-123/billingInfo`. This field is read-only. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The ID of the project that this `ProjectBillingInfo` represents, such as // `tokyo-rain-123`. This is a convenience field so that you don't need to // parse the `name` field to obtain a project ID. This field is read-only. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The resource name of the billing account associated with the project, if // any. For example, `billingAccounts/012345-567890-ABCDEF`. - BillingAccountName string `protobuf:"bytes,3,opt,name=billing_account_name,json=billingAccountName" json:"billing_account_name,omitempty"` + BillingAccountName string `protobuf:"bytes,3,opt,name=billing_account_name,json=billingAccountName,proto3" json:"billing_account_name,omitempty"` // True if the project is associated with an open billing account, to which // usage on the project is charged. False if the project is associated with a // closed billing account, or no billing account at all, and therefore cannot // use paid services. This field is read-only. - BillingEnabled bool `protobuf:"varint,4,opt,name=billing_enabled,json=billingEnabled" json:"billing_enabled,omitempty"` + BillingEnabled bool `protobuf:"varint,4,opt,name=billing_enabled,json=billingEnabled,proto3" json:"billing_enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -172,7 +172,7 @@ func (m *ProjectBillingInfo) GetBillingEnabled() bool { type GetBillingAccountRequest struct { // The resource name of the billing account to retrieve. For example, // `billingAccounts/012345-567890-ABCDEF`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -213,11 +213,11 @@ func (m *GetBillingAccountRequest) GetName() string { type ListBillingAccountsRequest struct { // Requested page size. The maximum page size is 100; this is also the // default. - PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A token identifying a page of results to return. This should be a // `next_page_token` value returned from a previous `ListBillingAccounts` // call. If unspecified, the first page of results is returned. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -264,11 +264,11 @@ func (m *ListBillingAccountsRequest) GetPageToken() string { // Response message for `ListBillingAccounts`. type ListBillingAccountsResponse struct { // A list of billing accounts. - BillingAccounts []*BillingAccount `protobuf:"bytes,1,rep,name=billing_accounts,json=billingAccounts" json:"billing_accounts,omitempty"` + BillingAccounts []*BillingAccount `protobuf:"bytes,1,rep,name=billing_accounts,json=billingAccounts,proto3" json:"billing_accounts,omitempty"` // A token to retrieve the next page of results. To retrieve the next page, // call `ListBillingAccounts` again with the `page_token` field set to this // value. This field is empty if there are no more results to retrieve. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -316,14 +316,14 @@ func (m *ListBillingAccountsResponse) GetNextPageToken() string { type ListProjectBillingInfoRequest struct { // The resource name of the billing account associated with the projects that // you want to list. For example, `billingAccounts/012345-567890-ABCDEF`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Requested page size. The maximum page size is 100; this is also the // default. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A token identifying a page of results to be returned. This should be a // `next_page_token` value returned from a previous `ListProjectBillingInfo` // call. If unspecified, the first page of results is returned. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -378,11 +378,11 @@ func (m *ListProjectBillingInfoRequest) GetPageToken() string { type ListProjectBillingInfoResponse struct { // A list of `ProjectBillingInfo` resources representing the projects // associated with the billing account. - ProjectBillingInfo []*ProjectBillingInfo `protobuf:"bytes,1,rep,name=project_billing_info,json=projectBillingInfo" json:"project_billing_info,omitempty"` + ProjectBillingInfo []*ProjectBillingInfo `protobuf:"bytes,1,rep,name=project_billing_info,json=projectBillingInfo,proto3" json:"project_billing_info,omitempty"` // A token to retrieve the next page of results. To retrieve the next page, // call `ListProjectBillingInfo` again with the `page_token` field set to this // value. This field is empty if there are no more results to retrieve. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -430,7 +430,7 @@ func (m *ListProjectBillingInfoResponse) GetNextPageToken() string { type GetProjectBillingInfoRequest struct { // The resource name of the project for which billing information is // retrieved. For example, `projects/tokyo-rain-123`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -471,10 +471,10 @@ func (m *GetProjectBillingInfoRequest) GetName() string { type UpdateProjectBillingInfoRequest struct { // The resource name of the project associated with the billing information // that you want to update. For example, `projects/tokyo-rain-123`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The new billing information for the project. Read-only fields are ignored; // thus, you may leave empty all fields except `billing_account_name`. - ProjectBillingInfo *ProjectBillingInfo `protobuf:"bytes,2,opt,name=project_billing_info,json=projectBillingInfo" json:"project_billing_info,omitempty"` + ProjectBillingInfo *ProjectBillingInfo `protobuf:"bytes,2,opt,name=project_billing_info,json=projectBillingInfo,proto3" json:"project_billing_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -645,8 +645,7 @@ func (c *cloudBillingClient) UpdateProjectBillingInfo(ctx context.Context, in *U return out, nil } -// Server API for CloudBilling service - +// CloudBillingServer is the server API for CloudBilling service. type CloudBillingServer interface { // Gets information about a billing account. The current authenticated user // must be an [owner of the billing diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/clusters.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/clusters.pb.go index d7a803ccf..66c87b5f6 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/clusters.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/clusters.pb.go @@ -109,32 +109,32 @@ func (ClusterStatus_Substate) EnumDescriptor() ([]byte, []int) { // a cluster of Google Compute Engine instances. type Cluster struct { // Required. The Google Cloud Platform project ID that the cluster belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The cluster name. Cluster names within a project must be // unique. Names of deleted clusters can be reused. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Required. The cluster config. Note that Cloud Dataproc may set // default values, and values may change when clusters are updated. - Config *ClusterConfig `protobuf:"bytes,3,opt,name=config" json:"config,omitempty"` + Config *ClusterConfig `protobuf:"bytes,3,opt,name=config,proto3" json:"config,omitempty"` // Optional. The labels to associate with this cluster. // Label **keys** must contain 1 to 63 characters, and must conform to // [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // Label **values** may be empty, but, if present, must contain 1 to 63 // characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // No more than 32 labels can be associated with a cluster. - Labels map[string]string `protobuf:"bytes,8,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,8,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Output-only. Cluster status. - Status *ClusterStatus `protobuf:"bytes,4,opt,name=status" json:"status,omitempty"` + Status *ClusterStatus `protobuf:"bytes,4,opt,name=status,proto3" json:"status,omitempty"` // Output-only. The previous cluster status. - StatusHistory []*ClusterStatus `protobuf:"bytes,7,rep,name=status_history,json=statusHistory" json:"status_history,omitempty"` + StatusHistory []*ClusterStatus `protobuf:"bytes,7,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"` // Output-only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc // generates this value when it creates the cluster. - ClusterUuid string `protobuf:"bytes,6,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` + ClusterUuid string `protobuf:"bytes,6,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` // Contains cluster daemon metrics such as HDFS and YARN stats. // // **Beta Feature**: This report is available for testing purposes only. It may // be changed before final release. - Metrics *ClusterMetrics `protobuf:"bytes,9,opt,name=metrics" json:"metrics,omitempty"` + Metrics *ClusterMetrics `protobuf:"bytes,9,opt,name=metrics,proto3" json:"metrics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -228,21 +228,21 @@ type ClusterConfig struct { // ASIA, or EU) for your cluster's staging bucket according to the Google // Compute Engine zone where your cluster is deployed, and then it will create // and manage this project-level, per-location bucket for you. - ConfigBucket string `protobuf:"bytes,1,opt,name=config_bucket,json=configBucket" json:"config_bucket,omitempty"` + ConfigBucket string `protobuf:"bytes,1,opt,name=config_bucket,json=configBucket,proto3" json:"config_bucket,omitempty"` // Required. The shared Google Compute Engine config settings for // all instances in a cluster. - GceClusterConfig *GceClusterConfig `protobuf:"bytes,8,opt,name=gce_cluster_config,json=gceClusterConfig" json:"gce_cluster_config,omitempty"` + GceClusterConfig *GceClusterConfig `protobuf:"bytes,8,opt,name=gce_cluster_config,json=gceClusterConfig,proto3" json:"gce_cluster_config,omitempty"` // Optional. The Google Compute Engine config settings for // the master instance in a cluster. - MasterConfig *InstanceGroupConfig `protobuf:"bytes,9,opt,name=master_config,json=masterConfig" json:"master_config,omitempty"` + MasterConfig *InstanceGroupConfig `protobuf:"bytes,9,opt,name=master_config,json=masterConfig,proto3" json:"master_config,omitempty"` // Optional. The Google Compute Engine config settings for // worker instances in a cluster. - WorkerConfig *InstanceGroupConfig `protobuf:"bytes,10,opt,name=worker_config,json=workerConfig" json:"worker_config,omitempty"` + WorkerConfig *InstanceGroupConfig `protobuf:"bytes,10,opt,name=worker_config,json=workerConfig,proto3" json:"worker_config,omitempty"` // Optional. The Google Compute Engine config settings for // additional worker instances in a cluster. - SecondaryWorkerConfig *InstanceGroupConfig `protobuf:"bytes,12,opt,name=secondary_worker_config,json=secondaryWorkerConfig" json:"secondary_worker_config,omitempty"` + SecondaryWorkerConfig *InstanceGroupConfig `protobuf:"bytes,12,opt,name=secondary_worker_config,json=secondaryWorkerConfig,proto3" json:"secondary_worker_config,omitempty"` // Optional. The config settings for software inside the cluster. - SoftwareConfig *SoftwareConfig `protobuf:"bytes,13,opt,name=software_config,json=softwareConfig" json:"software_config,omitempty"` + SoftwareConfig *SoftwareConfig `protobuf:"bytes,13,opt,name=software_config,json=softwareConfig,proto3" json:"software_config,omitempty"` // Optional. Commands to execute on each node after config is // completed. By default, executables are run on master and all worker nodes. // You can test a node's `role` metadata to run an executable on @@ -254,7 +254,7 @@ type ClusterConfig struct { // else // ... worker specific actions ... // fi - InitializationActions []*NodeInitializationAction `protobuf:"bytes,11,rep,name=initialization_actions,json=initializationActions" json:"initialization_actions,omitempty"` + InitializationActions []*NodeInitializationAction `protobuf:"bytes,11,rep,name=initialization_actions,json=initializationActions,proto3" json:"initialization_actions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -347,7 +347,7 @@ type GceClusterConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]` // * `projects/[project_id]/zones/[zone]` // * `us-central1-f` - ZoneUri string `protobuf:"bytes,1,opt,name=zone_uri,json=zoneUri" json:"zone_uri,omitempty"` + ZoneUri string `protobuf:"bytes,1,opt,name=zone_uri,json=zoneUri,proto3" json:"zone_uri,omitempty"` // Optional. The Google Compute Engine network to be used for machine // communications. Cannot be specified with subnetwork_uri. If neither // `network_uri` nor `subnetwork_uri` is specified, the "default" network of @@ -359,7 +359,7 @@ type GceClusterConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` // * `projects/[project_id]/regions/global/default` // * `default` - NetworkUri string `protobuf:"bytes,2,opt,name=network_uri,json=networkUri" json:"network_uri,omitempty"` + NetworkUri string `protobuf:"bytes,2,opt,name=network_uri,json=networkUri,proto3" json:"network_uri,omitempty"` // Optional. The Google Compute Engine subnetwork to be used for machine // communications. Cannot be specified with network_uri. // @@ -368,14 +368,14 @@ type GceClusterConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/us-east1/sub0` // * `projects/[project_id]/regions/us-east1/sub0` // * `sub0` - SubnetworkUri string `protobuf:"bytes,6,opt,name=subnetwork_uri,json=subnetworkUri" json:"subnetwork_uri,omitempty"` + SubnetworkUri string `protobuf:"bytes,6,opt,name=subnetwork_uri,json=subnetworkUri,proto3" json:"subnetwork_uri,omitempty"` // Optional. If true, all instances in the cluster will only have internal IP // addresses. By default, clusters are not restricted to internal IP addresses, // and will have ephemeral external IP addresses assigned to each instance. // This `internal_ip_only` restriction can only be enabled for subnetwork // enabled networks, and all off-cluster dependencies must be configured to be // accessible without external IP addresses. - InternalIpOnly bool `protobuf:"varint,7,opt,name=internal_ip_only,json=internalIpOnly" json:"internal_ip_only,omitempty"` + InternalIpOnly bool `protobuf:"varint,7,opt,name=internal_ip_only,json=internalIpOnly,proto3" json:"internal_ip_only,omitempty"` // Optional. The service account of the instances. Defaults to the default // Google Compute Engine service account. Custom service accounts need // permissions equivalent to the folloing IAM roles: @@ -386,7 +386,7 @@ type GceClusterConfig struct { // (see https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts // for more information). // Example: `[account_id]@[project_id].iam.gserviceaccount.com` - ServiceAccount string `protobuf:"bytes,8,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount string `protobuf:"bytes,8,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // Optional. The URIs of service account scopes to be included in Google // Compute Engine instances. The following base set of scopes is always // included: @@ -401,13 +401,13 @@ type GceClusterConfig struct { // * https://www.googleapis.com/auth/bigtable.admin.table // * https://www.googleapis.com/auth/bigtable.data // * https://www.googleapis.com/auth/devstorage.full_control - ServiceAccountScopes []string `protobuf:"bytes,3,rep,name=service_account_scopes,json=serviceAccountScopes" json:"service_account_scopes,omitempty"` + ServiceAccountScopes []string `protobuf:"bytes,3,rep,name=service_account_scopes,json=serviceAccountScopes,proto3" json:"service_account_scopes,omitempty"` // The Google Compute Engine tags to add to all instances (see // [Tagging instances](/compute/docs/label-or-tag-resources#tags)). - Tags []string `protobuf:"bytes,4,rep,name=tags" json:"tags,omitempty"` + Tags []string `protobuf:"bytes,4,rep,name=tags,proto3" json:"tags,omitempty"` // The Google Compute Engine metadata entries to add to all instances (see // [Project and instance metadata](https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)). - Metadata map[string]string `protobuf:"bytes,5,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,5,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -498,14 +498,14 @@ func (m *GceClusterConfig) GetMetadata() map[string]string { type InstanceGroupConfig struct { // Optional. The number of VM instances in the instance group. // For master instance groups, must be set to 1. - NumInstances int32 `protobuf:"varint,1,opt,name=num_instances,json=numInstances" json:"num_instances,omitempty"` + NumInstances int32 `protobuf:"varint,1,opt,name=num_instances,json=numInstances,proto3" json:"num_instances,omitempty"` // Optional. The list of instance names. Cloud Dataproc derives the names from // `cluster_name`, `num_instances`, and the instance group if not set by user // (recommended practice is to let Cloud Dataproc derive the name). - InstanceNames []string `protobuf:"bytes,2,rep,name=instance_names,json=instanceNames" json:"instance_names,omitempty"` + InstanceNames []string `protobuf:"bytes,2,rep,name=instance_names,json=instanceNames,proto3" json:"instance_names,omitempty"` // Output-only. The Google Compute Engine image resource used for cluster // instances. Inferred from `SoftwareConfig.image_version`. - ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` // Optional. The Google Compute Engine machine type used for cluster instances. // // A full URL, partial URI, or short name are valid. Examples: @@ -513,21 +513,21 @@ type InstanceGroupConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` // * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` // * `n1-standard-2` - MachineTypeUri string `protobuf:"bytes,4,opt,name=machine_type_uri,json=machineTypeUri" json:"machine_type_uri,omitempty"` + MachineTypeUri string `protobuf:"bytes,4,opt,name=machine_type_uri,json=machineTypeUri,proto3" json:"machine_type_uri,omitempty"` // Optional. Disk option config settings. - DiskConfig *DiskConfig `protobuf:"bytes,5,opt,name=disk_config,json=diskConfig" json:"disk_config,omitempty"` + DiskConfig *DiskConfig `protobuf:"bytes,5,opt,name=disk_config,json=diskConfig,proto3" json:"disk_config,omitempty"` // Optional. Specifies that this instance group contains preemptible instances. - IsPreemptible bool `protobuf:"varint,6,opt,name=is_preemptible,json=isPreemptible" json:"is_preemptible,omitempty"` + IsPreemptible bool `protobuf:"varint,6,opt,name=is_preemptible,json=isPreemptible,proto3" json:"is_preemptible,omitempty"` // Output-only. The config for Google Compute Engine Instance Group // Manager that manages this group. // This is only used for preemptible instance groups. - ManagedGroupConfig *ManagedGroupConfig `protobuf:"bytes,7,opt,name=managed_group_config,json=managedGroupConfig" json:"managed_group_config,omitempty"` + ManagedGroupConfig *ManagedGroupConfig `protobuf:"bytes,7,opt,name=managed_group_config,json=managedGroupConfig,proto3" json:"managed_group_config,omitempty"` // Optional. The Google Compute Engine accelerator configuration for these // instances. // // **Beta Feature**: This feature is still under development. It may be // changed before final release. - Accelerators []*AcceleratorConfig `protobuf:"bytes,8,rep,name=accelerators" json:"accelerators,omitempty"` + Accelerators []*AcceleratorConfig `protobuf:"bytes,8,rep,name=accelerators,proto3" json:"accelerators,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -617,9 +617,9 @@ func (m *InstanceGroupConfig) GetAccelerators() []*AcceleratorConfig { type ManagedGroupConfig struct { // Output-only. The name of the Instance Template used for the Managed // Instance Group. - InstanceTemplateName string `protobuf:"bytes,1,opt,name=instance_template_name,json=instanceTemplateName" json:"instance_template_name,omitempty"` + InstanceTemplateName string `protobuf:"bytes,1,opt,name=instance_template_name,json=instanceTemplateName,proto3" json:"instance_template_name,omitempty"` // Output-only. The name of the Instance Group Manager for this group. - InstanceGroupManagerName string `protobuf:"bytes,2,opt,name=instance_group_manager_name,json=instanceGroupManagerName" json:"instance_group_manager_name,omitempty"` + InstanceGroupManagerName string `protobuf:"bytes,2,opt,name=instance_group_manager_name,json=instanceGroupManagerName,proto3" json:"instance_group_manager_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -674,9 +674,9 @@ type AcceleratorConfig struct { // * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` // * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` // * `nvidia-tesla-k80` - AcceleratorTypeUri string `protobuf:"bytes,1,opt,name=accelerator_type_uri,json=acceleratorTypeUri" json:"accelerator_type_uri,omitempty"` + AcceleratorTypeUri string `protobuf:"bytes,1,opt,name=accelerator_type_uri,json=acceleratorTypeUri,proto3" json:"accelerator_type_uri,omitempty"` // The number of the accelerator cards of this type exposed to this instance. - AcceleratorCount int32 `protobuf:"varint,2,opt,name=accelerator_count,json=acceleratorCount" json:"accelerator_count,omitempty"` + AcceleratorCount int32 `protobuf:"varint,2,opt,name=accelerator_count,json=acceleratorCount,proto3" json:"accelerator_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -723,14 +723,14 @@ func (m *AcceleratorConfig) GetAcceleratorCount() int32 { // Specifies the config of disk options for a group of VM instances. type DiskConfig struct { // Optional. Size in GB of the boot disk (default is 500GB). - BootDiskSizeGb int32 `protobuf:"varint,1,opt,name=boot_disk_size_gb,json=bootDiskSizeGb" json:"boot_disk_size_gb,omitempty"` + BootDiskSizeGb int32 `protobuf:"varint,1,opt,name=boot_disk_size_gb,json=bootDiskSizeGb,proto3" json:"boot_disk_size_gb,omitempty"` // Optional. Number of attached SSDs, from 0 to 4 (default is 0). // If SSDs are not attached, the boot disk is used to store runtime logs and // [HDFS](https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. // If one or more SSDs are attached, this runtime bulk // data is spread across them, and the boot disk contains only basic // config and installed binaries. - NumLocalSsds int32 `protobuf:"varint,2,opt,name=num_local_ssds,json=numLocalSsds" json:"num_local_ssds,omitempty"` + NumLocalSsds int32 `protobuf:"varint,2,opt,name=num_local_ssds,json=numLocalSsds,proto3" json:"num_local_ssds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -778,12 +778,12 @@ func (m *DiskConfig) GetNumLocalSsds() int32 { // timeout period for executable completion. type NodeInitializationAction struct { // Required. Google Cloud Storage URI of executable file. - ExecutableFile string `protobuf:"bytes,1,opt,name=executable_file,json=executableFile" json:"executable_file,omitempty"` + ExecutableFile string `protobuf:"bytes,1,opt,name=executable_file,json=executableFile,proto3" json:"executable_file,omitempty"` // Optional. Amount of time executable has to complete. Default is // 10 minutes. Cluster creation fails with an explanatory error message (the // name of the executable that caused the error and the exceeded timeout // period) if the executable is not completed at end of the timeout period. - ExecutionTimeout *duration.Duration `protobuf:"bytes,2,opt,name=execution_timeout,json=executionTimeout" json:"execution_timeout,omitempty"` + ExecutionTimeout *duration.Duration `protobuf:"bytes,2,opt,name=execution_timeout,json=executionTimeout,proto3" json:"execution_timeout,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -830,14 +830,14 @@ func (m *NodeInitializationAction) GetExecutionTimeout() *duration.Duration { // The status of a cluster and its instances. type ClusterStatus struct { // Output-only. The cluster's state. - State ClusterStatus_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.dataproc.v1.ClusterStatus_State" json:"state,omitempty"` + State ClusterStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1.ClusterStatus_State" json:"state,omitempty"` // Output-only. Optional details of cluster's state. - Detail string `protobuf:"bytes,2,opt,name=detail" json:"detail,omitempty"` + Detail string `protobuf:"bytes,2,opt,name=detail,proto3" json:"detail,omitempty"` // Output-only. Time when this state was entered. - StateStartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=state_start_time,json=stateStartTime" json:"state_start_time,omitempty"` + StateStartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"` // Output-only. Additional state information that includes // status reported by the agent. - Substate ClusterStatus_Substate `protobuf:"varint,4,opt,name=substate,enum=google.cloud.dataproc.v1.ClusterStatus_Substate" json:"substate,omitempty"` + Substate ClusterStatus_Substate `protobuf:"varint,4,opt,name=substate,proto3,enum=google.cloud.dataproc.v1.ClusterStatus_Substate" json:"substate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -900,7 +900,7 @@ type SoftwareConfig struct { // Optional. The version of software inside the cluster. It must match the // regular expression `[0-9]+\.[0-9]+`. If unspecified, it defaults to the // latest version (see [Cloud Dataproc Versioning](/dataproc/versioning)). - ImageVersion string `protobuf:"bytes,1,opt,name=image_version,json=imageVersion" json:"image_version,omitempty"` + ImageVersion string `protobuf:"bytes,1,opt,name=image_version,json=imageVersion,proto3" json:"image_version,omitempty"` // Optional. The properties to set on daemon config files. // // Property keys are specified in `prefix:property` format, such as @@ -919,7 +919,7 @@ type SoftwareConfig struct { // // For more information, see // [Cluster properties](/dataproc/docs/concepts/cluster-properties). - Properties map[string]string `protobuf:"bytes,2,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,2,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -969,9 +969,9 @@ func (m *SoftwareConfig) GetProperties() map[string]string { // be changed before final release. type ClusterMetrics struct { // The HDFS metrics. - HdfsMetrics map[string]int64 `protobuf:"bytes,1,rep,name=hdfs_metrics,json=hdfsMetrics" json:"hdfs_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + HdfsMetrics map[string]int64 `protobuf:"bytes,1,rep,name=hdfs_metrics,json=hdfsMetrics,proto3" json:"hdfs_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` // The YARN metrics. - YarnMetrics map[string]int64 `protobuf:"bytes,2,rep,name=yarn_metrics,json=yarnMetrics" json:"yarn_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + YarnMetrics map[string]int64 `protobuf:"bytes,2,rep,name=yarn_metrics,json=yarnMetrics,proto3" json:"yarn_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1019,11 +1019,11 @@ func (m *ClusterMetrics) GetYarnMetrics() map[string]int64 { type CreateClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster to create. - Cluster *Cluster `protobuf:"bytes,2,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,2,opt,name=cluster,proto3" json:"cluster,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1078,13 +1078,13 @@ func (m *CreateClusterRequest) GetCluster() *Cluster { type UpdateClusterRequest struct { // Required. The ID of the Google Cloud Platform project the // cluster belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,5,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,5,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Required. The changes to the cluster. - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` // Required. Specifies the path, relative to `Cluster`, of // the field to update. For example, to change the number of workers // in a cluster to 5, the `update_mask` parameter would be @@ -1132,7 +1132,7 @@ type UpdateClusterRequest struct { // // //
- UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1201,11 +1201,11 @@ func (m *UpdateClusterRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1260,11 +1260,11 @@ func (m *DeleteClusterRequest) GetClusterName() string { type GetClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1319,9 +1319,9 @@ func (m *GetClusterRequest) GetClusterName() string { type ListClustersRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,4,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,4,opt,name=region,proto3" json:"region,omitempty"` // Optional. A filter constraining the clusters to list. Filters are // case-sensitive and have the following syntax: // @@ -1341,11 +1341,11 @@ type ListClustersRequest struct { // // status.state = ACTIVE AND clusterName = mycluster // AND labels.env = staging AND labels.starred = * - Filter string `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // Optional. The standard List page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The standard List page token. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1413,11 +1413,11 @@ func (m *ListClustersRequest) GetPageToken() string { // The list of all clusters in a project. type ListClustersResponse struct { // Output-only. The clusters in the project. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` // Output-only. This token is included in the response if there are more // results to fetch. To fetch additional results, provide this value as the // `page_token` in a subsequent `ListClustersRequest`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1465,11 +1465,11 @@ func (m *ListClustersResponse) GetNextPageToken() string { type DiagnoseClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1525,7 +1525,7 @@ type DiagnoseClusterResults struct { // Output-only. The Google Cloud Storage URI of the diagnostic output. // The output report is a plain text file with a summary of collected // diagnostics. - OutputUri string `protobuf:"bytes,1,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"` + OutputUri string `protobuf:"bytes,1,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1681,8 +1681,7 @@ func (c *clusterControllerClient) DiagnoseCluster(ctx context.Context, in *Diagn return out, nil } -// Server API for ClusterController service - +// ClusterControllerServer is the server API for ClusterController service. type ClusterControllerServer interface { // Creates a cluster in a project. CreateCluster(context.Context, *CreateClusterRequest) (*longrunning.Operation, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/jobs.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/jobs.pb.go index 316e4b10d..414c146f7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/jobs.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/jobs.pb.go @@ -280,7 +280,7 @@ type LoggingConfig struct { // "root" package name to configure rootLogger. // Examples: // 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' - DriverLogLevels map[string]LoggingConfig_Level `protobuf:"bytes,2,rep,name=driver_log_levels,json=driverLogLevels" json:"driver_log_levels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value,enum=google.cloud.dataproc.v1.LoggingConfig_Level"` + DriverLogLevels map[string]LoggingConfig_Level `protobuf:"bytes,2,rep,name=driver_log_levels,json=driverLogLevels,proto3" json:"driver_log_levels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3,enum=google.cloud.dataproc.v1.LoggingConfig_Level"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -334,25 +334,25 @@ type HadoopJob struct { // include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job // properties, since a collision may occur that causes an incorrect job // submission. - Args []string `protobuf:"bytes,3,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` // Optional. Jar file URIs to add to the CLASSPATHs of the // Hadoop driver and tasks. - JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. HCFS (Hadoop Compatible Filesystem) URIs of files to be copied // to the working directory of Hadoop drivers and distributed tasks. Useful // for naively parallel tasks. - FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris" json:"file_uris,omitempty"` + FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"` // Optional. HCFS URIs of archives to be extracted in the working directory of // Hadoop drivers and tasks. Supported file types: // .jar, .tar, .tar.gz, .tgz, or .zip. - ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris" json:"archive_uris,omitempty"` + ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"` // Optional. A mapping of property names to values, used to configure Hadoop. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site and // classes in user code. - Properties map[string]string `protobuf:"bytes,7,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -387,10 +387,10 @@ type isHadoopJob_Driver interface { } type HadoopJob_MainJarFileUri struct { - MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,oneof"` + MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,proto3,oneof"` } type HadoopJob_MainClass struct { - MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,oneof"` + MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,proto3,oneof"` } func (*HadoopJob_MainJarFileUri) isHadoopJob_Driver() {} @@ -540,24 +540,24 @@ type SparkJob struct { // Optional. The arguments to pass to the driver. Do not include arguments, // such as `--conf`, that can be set as job properties, since a collision may // occur that causes an incorrect job submission. - Args []string `protobuf:"bytes,3,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` // Optional. HCFS URIs of jar files to add to the CLASSPATHs of the // Spark driver and tasks. - JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. HCFS URIs of files to be copied to the working directory of // Spark drivers and distributed tasks. Useful for naively parallel tasks. - FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris" json:"file_uris,omitempty"` + FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"` // Optional. HCFS URIs of archives to be extracted in the working directory // of Spark drivers and tasks. Supported file types: // .jar, .tar, .tar.gz, .tgz, and .zip. - ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris" json:"archive_uris,omitempty"` + ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"` // Optional. A mapping of property names to values, used to configure Spark. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in // /etc/spark/conf/spark-defaults.conf and classes in user code. - Properties map[string]string `protobuf:"bytes,7,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -592,10 +592,10 @@ type isSparkJob_Driver interface { } type SparkJob_MainJarFileUri struct { - MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,oneof"` + MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,proto3,oneof"` } type SparkJob_MainClass struct { - MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,oneof"` + MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,proto3,oneof"` } func (*SparkJob_MainJarFileUri) isSparkJob_Driver() {} @@ -736,30 +736,30 @@ func _SparkJob_OneofSizer(msg proto.Message) (n int) { type PySparkJob struct { // Required. The HCFS URI of the main Python file to use as the driver. Must // be a .py file. - MainPythonFileUri string `protobuf:"bytes,1,opt,name=main_python_file_uri,json=mainPythonFileUri" json:"main_python_file_uri,omitempty"` + MainPythonFileUri string `protobuf:"bytes,1,opt,name=main_python_file_uri,json=mainPythonFileUri,proto3" json:"main_python_file_uri,omitempty"` // Optional. The arguments to pass to the driver. Do not include arguments, // such as `--conf`, that can be set as job properties, since a collision may // occur that causes an incorrect job submission. - Args []string `protobuf:"bytes,2,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,2,rep,name=args,proto3" json:"args,omitempty"` // Optional. HCFS file URIs of Python files to pass to the PySpark // framework. Supported file types: .py, .egg, and .zip. - PythonFileUris []string `protobuf:"bytes,3,rep,name=python_file_uris,json=pythonFileUris" json:"python_file_uris,omitempty"` + PythonFileUris []string `protobuf:"bytes,3,rep,name=python_file_uris,json=pythonFileUris,proto3" json:"python_file_uris,omitempty"` // Optional. HCFS URIs of jar files to add to the CLASSPATHs of the // Python driver and tasks. - JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. HCFS URIs of files to be copied to the working directory of // Python drivers and distributed tasks. Useful for naively parallel tasks. - FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris" json:"file_uris,omitempty"` + FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"` // Optional. HCFS URIs of archives to be extracted in the working directory of // .jar, .tar, .tar.gz, .tgz, and .zip. - ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris" json:"archive_uris,omitempty"` + ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"` // Optional. A mapping of property names to values, used to configure PySpark. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in // /etc/spark/conf/spark-defaults.conf and classes in user code. - Properties map[string]string `protobuf:"bytes,7,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -861,7 +861,7 @@ type QueryList struct { // ] // } // } - Queries []string `protobuf:"bytes,1,rep,name=queries" json:"queries,omitempty"` + Queries []string `protobuf:"bytes,1,rep,name=queries,proto3" json:"queries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -911,19 +911,19 @@ type HiveJob struct { // Optional. Whether to continue executing queries if a query fails. // The default value is `false`. Setting to `true` can be useful when executing // independent parallel queries. - ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure" json:"continue_on_failure,omitempty"` + ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"` // Optional. Mapping of query variable names to values (equivalent to the // Hive command: `SET name="value";`). - ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables,proto3" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. A mapping of property names and values, used to configure Hive. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, // /etc/hive/conf/hive-site.xml, and classes in user code. - Properties map[string]string `protobuf:"bytes,5,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,5,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. HCFS URIs of jar files to add to the CLASSPATH of the // Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes // and UDFs. - JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -958,10 +958,10 @@ type isHiveJob_Queries interface { } type HiveJob_QueryFileUri struct { - QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,oneof"` + QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"` } type HiveJob_QueryList struct { - QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,oneof"` + QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"` } func (*HiveJob_QueryFileUri) isHiveJob_Queries() {} @@ -1098,15 +1098,15 @@ type SparkSqlJob struct { Queries isSparkSqlJob_Queries `protobuf_oneof:"queries"` // Optional. Mapping of query variable names to values (equivalent to the // Spark SQL command: SET `name="value";`). - ScriptVariables map[string]string `protobuf:"bytes,3,rep,name=script_variables,json=scriptVariables" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ScriptVariables map[string]string `protobuf:"bytes,3,rep,name=script_variables,json=scriptVariables,proto3" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. A mapping of property names to values, used to configure // Spark SQL's SparkConf. Properties that conflict with values set by the // Cloud Dataproc API may be overwritten. - Properties map[string]string `protobuf:"bytes,4,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,4,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. - JarFileUris []string `protobuf:"bytes,56,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,56,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,6,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,6,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1141,10 +1141,10 @@ type isSparkSqlJob_Queries interface { } type SparkSqlJob_QueryFileUri struct { - QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,oneof"` + QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"` } type SparkSqlJob_QueryList struct { - QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,oneof"` + QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"` } func (*SparkSqlJob_QueryFileUri) isSparkSqlJob_Queries() {} @@ -1282,20 +1282,20 @@ type PigJob struct { // Optional. Whether to continue executing queries if a query fails. // The default value is `false`. Setting to `true` can be useful when executing // independent parallel queries. - ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure" json:"continue_on_failure,omitempty"` + ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"` // Optional. Mapping of query variable names to values (equivalent to the Pig // command: `name=[value]`). - ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables,proto3" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. A mapping of property names to values, used to configure Pig. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, // /etc/pig/conf/pig.properties, and classes in user code. - Properties map[string]string `protobuf:"bytes,5,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,5,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. HCFS URIs of jar files to add to the CLASSPATH of // the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs. - JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,7,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,7,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1330,10 +1330,10 @@ type isPigJob_Queries interface { } type PigJob_QueryFileUri struct { - QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,oneof"` + QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"` } type PigJob_QueryList struct { - QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,oneof"` + QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"` } func (*PigJob_QueryFileUri) isPigJob_Queries() {} @@ -1468,10 +1468,10 @@ func _PigJob_OneofSizer(msg proto.Message) (n int) { // Cloud Dataproc job config. type JobPlacement struct { // Required. The name of the cluster where the job will be submitted. - ClusterName string `protobuf:"bytes,1,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,1,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Output-only. A cluster UUID generated by the Cloud Dataproc service when // the job is submitted. - ClusterUuid string `protobuf:"bytes,2,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` + ClusterUuid string `protobuf:"bytes,2,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1518,15 +1518,15 @@ func (m *JobPlacement) GetClusterUuid() string { // Cloud Dataproc job status. type JobStatus struct { // Output-only. A state message specifying the overall job state. - State JobStatus_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.dataproc.v1.JobStatus_State" json:"state,omitempty"` + State JobStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1.JobStatus_State" json:"state,omitempty"` // Output-only. Optional job state details, such as an error // description if the state is ERROR. - Details string `protobuf:"bytes,2,opt,name=details" json:"details,omitempty"` + Details string `protobuf:"bytes,2,opt,name=details,proto3" json:"details,omitempty"` // Output-only. The time when this state was entered. - StateStartTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=state_start_time,json=stateStartTime" json:"state_start_time,omitempty"` + StateStartTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"` // Output-only. Additional state information, which includes // status reported by the agent. - Substate JobStatus_Substate `protobuf:"varint,7,opt,name=substate,enum=google.cloud.dataproc.v1.JobStatus_Substate" json:"substate,omitempty"` + Substate JobStatus_Substate `protobuf:"varint,7,opt,name=substate,proto3,enum=google.cloud.dataproc.v1.JobStatus_Substate" json:"substate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1588,13 +1588,13 @@ func (m *JobStatus) GetSubstate() JobStatus_Substate { type JobReference struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Optional. The job ID, which must be unique within the project. The job ID // is generated by the server upon job submission or provided by the user as a // means to perform retries without creating duplicate jobs. The ID must // contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or // hyphens (-). The maximum length is 100 characters. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1645,16 +1645,16 @@ func (m *JobReference) GetJobId() string { // be changed before final release. type YarnApplication struct { // Required. The application name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The application state. - State YarnApplication_State `protobuf:"varint,2,opt,name=state,enum=google.cloud.dataproc.v1.YarnApplication_State" json:"state,omitempty"` + State YarnApplication_State `protobuf:"varint,2,opt,name=state,proto3,enum=google.cloud.dataproc.v1.YarnApplication_State" json:"state,omitempty"` // Required. The numerical progress of the application, from 1 to 100. - Progress float32 `protobuf:"fixed32,3,opt,name=progress" json:"progress,omitempty"` + Progress float32 `protobuf:"fixed32,3,opt,name=progress,proto3" json:"progress,omitempty"` // Optional. The HTTP URL of the ApplicationMaster, HistoryServer, or // TimelineServer that provides application-specific information. The URL uses // the internal hostname, and requires a proxy server for resolution and, // possibly, access. - TrackingUrl string `protobuf:"bytes,4,opt,name=tracking_url,json=trackingUrl" json:"tracking_url,omitempty"` + TrackingUrl string `protobuf:"bytes,4,opt,name=tracking_url,json=trackingUrl,proto3" json:"tracking_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1718,10 +1718,10 @@ type Job struct { // obtain the equivalent REST path of the job resource. If this property // is not specified when a job is created, the server generates a // job_id. - Reference *JobReference `protobuf:"bytes,1,opt,name=reference" json:"reference,omitempty"` + Reference *JobReference `protobuf:"bytes,1,opt,name=reference,proto3" json:"reference,omitempty"` // Required. Job information, including how, when, and where to // run the job. - Placement *JobPlacement `protobuf:"bytes,2,opt,name=placement" json:"placement,omitempty"` + Placement *JobPlacement `protobuf:"bytes,2,opt,name=placement,proto3" json:"placement,omitempty"` // Required. The application/framework-specific portion of the job. // // Types that are valid to be assigned to TypeJob: @@ -1735,30 +1735,30 @@ type Job struct { // Output-only. The job status. Additional application-specific // status information may be contained in the type_job // and yarn_applications fields. - Status *JobStatus `protobuf:"bytes,8,opt,name=status" json:"status,omitempty"` + Status *JobStatus `protobuf:"bytes,8,opt,name=status,proto3" json:"status,omitempty"` // Output-only. The previous job status. - StatusHistory []*JobStatus `protobuf:"bytes,13,rep,name=status_history,json=statusHistory" json:"status_history,omitempty"` + StatusHistory []*JobStatus `protobuf:"bytes,13,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"` // Output-only. The collection of YARN applications spun up by this job. // // **Beta** Feature: This report is available for testing purposes only. It may // be changed before final release. - YarnApplications []*YarnApplication `protobuf:"bytes,9,rep,name=yarn_applications,json=yarnApplications" json:"yarn_applications,omitempty"` + YarnApplications []*YarnApplication `protobuf:"bytes,9,rep,name=yarn_applications,json=yarnApplications,proto3" json:"yarn_applications,omitempty"` // Output-only. A URI pointing to the location of the stdout of the job's // driver program. - DriverOutputResourceUri string `protobuf:"bytes,17,opt,name=driver_output_resource_uri,json=driverOutputResourceUri" json:"driver_output_resource_uri,omitempty"` + DriverOutputResourceUri string `protobuf:"bytes,17,opt,name=driver_output_resource_uri,json=driverOutputResourceUri,proto3" json:"driver_output_resource_uri,omitempty"` // Output-only. If present, the location of miscellaneous control files // which may be used as part of job setup and handling. If not present, // control files may be placed in the same location as `driver_output_uri`. - DriverControlFilesUri string `protobuf:"bytes,15,opt,name=driver_control_files_uri,json=driverControlFilesUri" json:"driver_control_files_uri,omitempty"` + DriverControlFilesUri string `protobuf:"bytes,15,opt,name=driver_control_files_uri,json=driverControlFilesUri,proto3" json:"driver_control_files_uri,omitempty"` // Optional. The labels to associate with this job. // Label **keys** must contain 1 to 63 characters, and must conform to // [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // Label **values** may be empty, but, if present, must contain 1 to 63 // characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // No more than 32 labels can be associated with a job. - Labels map[string]string `protobuf:"bytes,18,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,18,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. Job scheduling configuration. - Scheduling *JobScheduling `protobuf:"bytes,20,opt,name=scheduling" json:"scheduling,omitempty"` + Scheduling *JobScheduling `protobuf:"bytes,20,opt,name=scheduling,proto3" json:"scheduling,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1793,22 +1793,22 @@ type isJob_TypeJob interface { } type Job_HadoopJob struct { - HadoopJob *HadoopJob `protobuf:"bytes,3,opt,name=hadoop_job,json=hadoopJob,oneof"` + HadoopJob *HadoopJob `protobuf:"bytes,3,opt,name=hadoop_job,json=hadoopJob,proto3,oneof"` } type Job_SparkJob struct { - SparkJob *SparkJob `protobuf:"bytes,4,opt,name=spark_job,json=sparkJob,oneof"` + SparkJob *SparkJob `protobuf:"bytes,4,opt,name=spark_job,json=sparkJob,proto3,oneof"` } type Job_PysparkJob struct { - PysparkJob *PySparkJob `protobuf:"bytes,5,opt,name=pyspark_job,json=pysparkJob,oneof"` + PysparkJob *PySparkJob `protobuf:"bytes,5,opt,name=pyspark_job,json=pysparkJob,proto3,oneof"` } type Job_HiveJob struct { - HiveJob *HiveJob `protobuf:"bytes,6,opt,name=hive_job,json=hiveJob,oneof"` + HiveJob *HiveJob `protobuf:"bytes,6,opt,name=hive_job,json=hiveJob,proto3,oneof"` } type Job_PigJob struct { - PigJob *PigJob `protobuf:"bytes,7,opt,name=pig_job,json=pigJob,oneof"` + PigJob *PigJob `protobuf:"bytes,7,opt,name=pig_job,json=pigJob,proto3,oneof"` } type Job_SparkSqlJob struct { - SparkSqlJob *SparkSqlJob `protobuf:"bytes,12,opt,name=spark_sql_job,json=sparkSqlJob,oneof"` + SparkSqlJob *SparkSqlJob `protobuf:"bytes,12,opt,name=spark_sql_job,json=sparkSqlJob,proto3,oneof"` } func (*Job_HadoopJob) isJob_TypeJob() {} @@ -2093,7 +2093,7 @@ type JobScheduling struct { // 4 times within 10 minute window. // // Maximum value is 10. - MaxFailuresPerHour int32 `protobuf:"varint,1,opt,name=max_failures_per_hour,json=maxFailuresPerHour" json:"max_failures_per_hour,omitempty"` + MaxFailuresPerHour int32 `protobuf:"varint,1,opt,name=max_failures_per_hour,json=maxFailuresPerHour,proto3" json:"max_failures_per_hour,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2134,11 +2134,11 @@ func (m *JobScheduling) GetMaxFailuresPerHour() int32 { type SubmitJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job resource. - Job *Job `protobuf:"bytes,2,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2193,11 +2193,11 @@ func (m *SubmitJobRequest) GetJob() *Job { type GetJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2252,22 +2252,22 @@ func (m *GetJobRequest) GetJobId() string { type ListJobsRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,6,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,6,opt,name=region,proto3" json:"region,omitempty"` // Optional. The number of results to return in each response. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The page token, returned by a previous call, to request the // next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. If set, the returned jobs list includes only jobs that were // submitted to the named cluster. - ClusterName string `protobuf:"bytes,4,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,4,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Optional. Specifies enumerated categories of jobs to list. // (default = match ALL jobs). // // If `filter` is provided, `jobStateMatcher` will be ignored. - JobStateMatcher ListJobsRequest_JobStateMatcher `protobuf:"varint,5,opt,name=job_state_matcher,json=jobStateMatcher,enum=google.cloud.dataproc.v1.ListJobsRequest_JobStateMatcher" json:"job_state_matcher,omitempty"` + JobStateMatcher ListJobsRequest_JobStateMatcher `protobuf:"varint,5,opt,name=job_state_matcher,json=jobStateMatcher,proto3,enum=google.cloud.dataproc.v1.ListJobsRequest_JobStateMatcher" json:"job_state_matcher,omitempty"` // Optional. A filter constraining the jobs to list. Filters are // case-sensitive and have the following syntax: // @@ -2282,7 +2282,7 @@ type ListJobsRequest struct { // Example filter: // // status.state = ACTIVE AND labels.env = staging AND labels.starred = * - Filter string `protobuf:"bytes,7,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,7,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2365,20 +2365,20 @@ func (m *ListJobsRequest) GetFilter() string { type UpdateJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,2,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,2,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` // Required. The changes to the job. - Job *Job `protobuf:"bytes,4,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,4,opt,name=job,proto3" json:"job,omitempty"` // Required. Specifies the path, relative to Job, of // the field to update. For example, to update the labels of a Job the // update_mask parameter would be specified as // labels, and the `PATCH` request body would specify the new // value. Note: Currently, labels is the only // field that can be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2446,11 +2446,11 @@ func (m *UpdateJobRequest) GetUpdateMask() *field_mask.FieldMask { // A list of jobs in a project. type ListJobsResponse struct { // Output-only. Jobs list. - Jobs []*Job `protobuf:"bytes,1,rep,name=jobs" json:"jobs,omitempty"` + Jobs []*Job `protobuf:"bytes,1,rep,name=jobs,proto3" json:"jobs,omitempty"` // Optional. This token is included in the response if there are more results // to fetch. To fetch additional results, provide this value as the // `page_token` in a subsequent ListJobsRequest. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2498,11 +2498,11 @@ func (m *ListJobsResponse) GetNextPageToken() string { type CancelJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2557,11 +2557,11 @@ func (m *CancelJobRequest) GetJobId() string { type DeleteJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2744,8 +2744,7 @@ func (c *jobControllerClient) DeleteJob(ctx context.Context, in *DeleteJobReques return out, nil } -// Server API for JobController service - +// JobControllerServer is the server API for JobController service. type JobControllerServer interface { // Submits a job to a cluster. SubmitJob(context.Context, *SubmitJobRequest) (*Job, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/operations.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/operations.pb.go index e90f2fff0..b64cec485 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/operations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1/operations.pb.go @@ -59,13 +59,13 @@ func (ClusterOperationStatus_State) EnumDescriptor() ([]byte, []int) { // The status of the operation. type ClusterOperationStatus struct { // Output-only. A message containing the operation state. - State ClusterOperationStatus_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.dataproc.v1.ClusterOperationStatus_State" json:"state,omitempty"` + State ClusterOperationStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1.ClusterOperationStatus_State" json:"state,omitempty"` // Output-only. A message containing the detailed operation state. - InnerState string `protobuf:"bytes,2,opt,name=inner_state,json=innerState" json:"inner_state,omitempty"` + InnerState string `protobuf:"bytes,2,opt,name=inner_state,json=innerState,proto3" json:"inner_state,omitempty"` // Output-only.A message containing any operation metadata details. - Details string `protobuf:"bytes,3,opt,name=details" json:"details,omitempty"` + Details string `protobuf:"bytes,3,opt,name=details,proto3" json:"details,omitempty"` // Output-only. The time this state was entered. - StateStartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=state_start_time,json=stateStartTime" json:"state_start_time,omitempty"` + StateStartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -126,21 +126,21 @@ func (m *ClusterOperationStatus) GetStateStartTime() *timestamp.Timestamp { // Metadata describing the operation. type ClusterOperationMetadata struct { // Output-only. Name of the cluster for the operation. - ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Output-only. Cluster UUID for the operation. - ClusterUuid string `protobuf:"bytes,8,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` + ClusterUuid string `protobuf:"bytes,8,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` // Output-only. Current operation status. - Status *ClusterOperationStatus `protobuf:"bytes,9,opt,name=status" json:"status,omitempty"` + Status *ClusterOperationStatus `protobuf:"bytes,9,opt,name=status,proto3" json:"status,omitempty"` // Output-only. The previous operation status. - StatusHistory []*ClusterOperationStatus `protobuf:"bytes,10,rep,name=status_history,json=statusHistory" json:"status_history,omitempty"` + StatusHistory []*ClusterOperationStatus `protobuf:"bytes,10,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"` // Output-only. The operation type. - OperationType string `protobuf:"bytes,11,opt,name=operation_type,json=operationType" json:"operation_type,omitempty"` + OperationType string `protobuf:"bytes,11,opt,name=operation_type,json=operationType,proto3" json:"operation_type,omitempty"` // Output-only. Short description of operation. - Description string `protobuf:"bytes,12,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,12,opt,name=description,proto3" json:"description,omitempty"` // Output-only. Labels associated with the operation - Labels map[string]string `protobuf:"bytes,13,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,13,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Output-only. Errors encountered during operation execution. - Warnings []string `protobuf:"bytes,14,rep,name=warnings" json:"warnings,omitempty"` + Warnings []string `protobuf:"bytes,14,rep,name=warnings,proto3" json:"warnings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/clusters.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/clusters.pb.go index 88420cdfb..d36ce3845 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/clusters.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/clusters.pb.go @@ -67,12 +67,14 @@ func (x ClusterStatus_State) String() string { return proto.EnumName(ClusterStatus_State_name, int32(x)) } func (ClusterStatus_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{9, 0} + return fileDescriptor_clusters_92d23f68998e82a3, []int{9, 0} } +// The cluster substate. type ClusterStatus_Substate int32 const ( + // The cluster substate is unknown. ClusterStatus_UNSPECIFIED ClusterStatus_Substate = 0 // The cluster is known to be in an unhealthy state // (for example, critical daemons are not running or HDFS capacity is @@ -102,39 +104,39 @@ func (x ClusterStatus_Substate) String() string { return proto.EnumName(ClusterStatus_Substate_name, int32(x)) } func (ClusterStatus_Substate) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{9, 1} + return fileDescriptor_clusters_92d23f68998e82a3, []int{9, 1} } // Describes the identifying information, config, and status of -// a cluster of Google Compute Engine instances. +// a cluster of Compute Engine instances. type Cluster struct { // Required. The Google Cloud Platform project ID that the cluster belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The cluster name. Cluster names within a project must be // unique. Names of deleted clusters can be reused. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Required. The cluster config. Note that Cloud Dataproc may set // default values, and values may change when clusters are updated. - Config *ClusterConfig `protobuf:"bytes,3,opt,name=config" json:"config,omitempty"` + Config *ClusterConfig `protobuf:"bytes,3,opt,name=config,proto3" json:"config,omitempty"` // Optional. The labels to associate with this cluster. // Label **keys** must contain 1 to 63 characters, and must conform to // [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // Label **values** may be empty, but, if present, must contain 1 to 63 // characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // No more than 32 labels can be associated with a cluster. - Labels map[string]string `protobuf:"bytes,8,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` - // Output-only. Cluster status. - Status *ClusterStatus `protobuf:"bytes,4,opt,name=status" json:"status,omitempty"` - // Output-only. The previous cluster status. - StatusHistory []*ClusterStatus `protobuf:"bytes,7,rep,name=status_history,json=statusHistory" json:"status_history,omitempty"` - // Output-only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc + Labels map[string]string `protobuf:"bytes,8,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + // Output only. Cluster status. + Status *ClusterStatus `protobuf:"bytes,4,opt,name=status,proto3" json:"status,omitempty"` + // Output only. The previous cluster status. + StatusHistory []*ClusterStatus `protobuf:"bytes,7,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"` + // Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc // generates this value when it creates the cluster. - ClusterUuid string `protobuf:"bytes,6,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` + ClusterUuid string `protobuf:"bytes,6,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` // Contains cluster daemon metrics such as HDFS and YARN stats. // // **Beta Feature**: This report is available for testing purposes only. It may // be changed before final release. - Metrics *ClusterMetrics `protobuf:"bytes,9,opt,name=metrics" json:"metrics,omitempty"` + Metrics *ClusterMetrics `protobuf:"bytes,9,opt,name=metrics,proto3" json:"metrics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -144,7 +146,7 @@ func (m *Cluster) Reset() { *m = Cluster{} } func (m *Cluster) String() string { return proto.CompactTextString(m) } func (*Cluster) ProtoMessage() {} func (*Cluster) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{0} + return fileDescriptor_clusters_92d23f68998e82a3, []int{0} } func (m *Cluster) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Cluster.Unmarshal(m, b) @@ -222,29 +224,29 @@ func (m *Cluster) GetMetrics() *ClusterMetrics { // The cluster config. type ClusterConfig struct { - // Optional. A Google Cloud Storage staging bucket used for sharing generated + // Optional. A Cloud Storage staging bucket used for sharing generated // SSH keys and config. If you do not specify a staging bucket, Cloud // Dataproc will determine an appropriate Cloud Storage location (US, // ASIA, or EU) for your cluster's staging bucket according to the Google // Compute Engine zone where your cluster is deployed, and then it will create // and manage this project-level, per-location bucket for you. - ConfigBucket string `protobuf:"bytes,1,opt,name=config_bucket,json=configBucket" json:"config_bucket,omitempty"` - // Required. The shared Google Compute Engine config settings for + ConfigBucket string `protobuf:"bytes,1,opt,name=config_bucket,json=configBucket,proto3" json:"config_bucket,omitempty"` + // Required. The shared Compute Engine config settings for // all instances in a cluster. - GceClusterConfig *GceClusterConfig `protobuf:"bytes,8,opt,name=gce_cluster_config,json=gceClusterConfig" json:"gce_cluster_config,omitempty"` - // Optional. The Google Compute Engine config settings for + GceClusterConfig *GceClusterConfig `protobuf:"bytes,8,opt,name=gce_cluster_config,json=gceClusterConfig,proto3" json:"gce_cluster_config,omitempty"` + // Optional. The Compute Engine config settings for // the master instance in a cluster. - MasterConfig *InstanceGroupConfig `protobuf:"bytes,9,opt,name=master_config,json=masterConfig" json:"master_config,omitempty"` - // Optional. The Google Compute Engine config settings for + MasterConfig *InstanceGroupConfig `protobuf:"bytes,9,opt,name=master_config,json=masterConfig,proto3" json:"master_config,omitempty"` + // Optional. The Compute Engine config settings for // worker instances in a cluster. - WorkerConfig *InstanceGroupConfig `protobuf:"bytes,10,opt,name=worker_config,json=workerConfig" json:"worker_config,omitempty"` - // Optional. The Google Compute Engine config settings for + WorkerConfig *InstanceGroupConfig `protobuf:"bytes,10,opt,name=worker_config,json=workerConfig,proto3" json:"worker_config,omitempty"` + // Optional. The Compute Engine config settings for // additional worker instances in a cluster. - SecondaryWorkerConfig *InstanceGroupConfig `protobuf:"bytes,12,opt,name=secondary_worker_config,json=secondaryWorkerConfig" json:"secondary_worker_config,omitempty"` + SecondaryWorkerConfig *InstanceGroupConfig `protobuf:"bytes,12,opt,name=secondary_worker_config,json=secondaryWorkerConfig,proto3" json:"secondary_worker_config,omitempty"` // Optional. The config settings for software inside the cluster. - SoftwareConfig *SoftwareConfig `protobuf:"bytes,13,opt,name=software_config,json=softwareConfig" json:"software_config,omitempty"` + SoftwareConfig *SoftwareConfig `protobuf:"bytes,13,opt,name=software_config,json=softwareConfig,proto3" json:"software_config,omitempty"` // Optional. The config setting for auto delete cluster schedule. - LifecycleConfig *LifecycleConfig `protobuf:"bytes,14,opt,name=lifecycle_config,json=lifecycleConfig" json:"lifecycle_config,omitempty"` + LifecycleConfig *LifecycleConfig `protobuf:"bytes,14,opt,name=lifecycle_config,json=lifecycleConfig,proto3" json:"lifecycle_config,omitempty"` // Optional. Commands to execute on each node after config is // completed. By default, executables are run on master and all worker nodes. // You can test a node's role metadata to run an executable on @@ -256,7 +258,7 @@ type ClusterConfig struct { // else // ... worker specific actions ... // fi - InitializationActions []*NodeInitializationAction `protobuf:"bytes,11,rep,name=initialization_actions,json=initializationActions" json:"initialization_actions,omitempty"` + InitializationActions []*NodeInitializationAction `protobuf:"bytes,11,rep,name=initialization_actions,json=initializationActions,proto3" json:"initialization_actions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -266,7 +268,7 @@ func (m *ClusterConfig) Reset() { *m = ClusterConfig{} } func (m *ClusterConfig) String() string { return proto.CompactTextString(m) } func (*ClusterConfig) ProtoMessage() {} func (*ClusterConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{1} + return fileDescriptor_clusters_92d23f68998e82a3, []int{1} } func (m *ClusterConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterConfig.Unmarshal(m, b) @@ -342,10 +344,10 @@ func (m *ClusterConfig) GetInitializationActions() []*NodeInitializationAction { return nil } -// Common config settings for resources of Google Compute Engine cluster +// Common config settings for resources of Compute Engine cluster // instances, applicable to all instances in the cluster. type GceClusterConfig struct { - // Optional. The zone where the Google Compute Engine cluster will be located. + // Optional. The zone where the Compute Engine cluster will be located. // On a create request, it is required in the "global" region. If omitted // in a non-global Cloud Dataproc region, the service will pick a zone in the // corresponding Compute Engine region. On a get request, zone will always be @@ -356,8 +358,8 @@ type GceClusterConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]` // * `projects/[project_id]/zones/[zone]` // * `us-central1-f` - ZoneUri string `protobuf:"bytes,1,opt,name=zone_uri,json=zoneUri" json:"zone_uri,omitempty"` - // Optional. The Google Compute Engine network to be used for machine + ZoneUri string `protobuf:"bytes,1,opt,name=zone_uri,json=zoneUri,proto3" json:"zone_uri,omitempty"` + // Optional. The Compute Engine network to be used for machine // communications. Cannot be specified with subnetwork_uri. If neither // `network_uri` nor `subnetwork_uri` is specified, the "default" network of // the project is used, if it exists. Cannot be a "Custom Subnet Network" (see @@ -368,8 +370,8 @@ type GceClusterConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/default` // * `projects/[project_id]/regions/global/default` // * `default` - NetworkUri string `protobuf:"bytes,2,opt,name=network_uri,json=networkUri" json:"network_uri,omitempty"` - // Optional. The Google Compute Engine subnetwork to be used for machine + NetworkUri string `protobuf:"bytes,2,opt,name=network_uri,json=networkUri,proto3" json:"network_uri,omitempty"` + // Optional. The Compute Engine subnetwork to be used for machine // communications. Cannot be specified with network_uri. // // A full URL, partial URI, or short name are valid. Examples: @@ -377,17 +379,17 @@ type GceClusterConfig struct { // * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/us-east1/sub0` // * `projects/[project_id]/regions/us-east1/sub0` // * `sub0` - SubnetworkUri string `protobuf:"bytes,6,opt,name=subnetwork_uri,json=subnetworkUri" json:"subnetwork_uri,omitempty"` + SubnetworkUri string `protobuf:"bytes,6,opt,name=subnetwork_uri,json=subnetworkUri,proto3" json:"subnetwork_uri,omitempty"` // Optional. If true, all instances in the cluster will only have internal IP // addresses. By default, clusters are not restricted to internal IP addresses, // and will have ephemeral external IP addresses assigned to each instance. // This `internal_ip_only` restriction can only be enabled for subnetwork // enabled networks, and all off-cluster dependencies must be configured to be // accessible without external IP addresses. - InternalIpOnly bool `protobuf:"varint,7,opt,name=internal_ip_only,json=internalIpOnly" json:"internal_ip_only,omitempty"` + InternalIpOnly bool `protobuf:"varint,7,opt,name=internal_ip_only,json=internalIpOnly,proto3" json:"internal_ip_only,omitempty"` // Optional. The service account of the instances. Defaults to the default - // Google Compute Engine service account. Custom service accounts need - // permissions equivalent to the folloing IAM roles: + // Compute Engine service account. Custom service accounts need + // permissions equivalent to the following IAM roles: // // * roles/logging.logWriter // * roles/storage.objectAdmin @@ -395,8 +397,8 @@ type GceClusterConfig struct { // (see https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts // for more information). // Example: `[account_id]@[project_id].iam.gserviceaccount.com` - ServiceAccount string `protobuf:"bytes,8,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` - // Optional. The URIs of service account scopes to be included in Google + ServiceAccount string `protobuf:"bytes,8,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` + // Optional. The URIs of service account scopes to be included in // Compute Engine instances. The following base set of scopes is always // included: // @@ -410,13 +412,13 @@ type GceClusterConfig struct { // * https://www.googleapis.com/auth/bigtable.admin.table // * https://www.googleapis.com/auth/bigtable.data // * https://www.googleapis.com/auth/devstorage.full_control - ServiceAccountScopes []string `protobuf:"bytes,3,rep,name=service_account_scopes,json=serviceAccountScopes" json:"service_account_scopes,omitempty"` - // The Google Compute Engine tags to add to all instances (see + ServiceAccountScopes []string `protobuf:"bytes,3,rep,name=service_account_scopes,json=serviceAccountScopes,proto3" json:"service_account_scopes,omitempty"` + // The Compute Engine tags to add to all instances (see // [Tagging instances](/compute/docs/label-or-tag-resources#tags)). - Tags []string `protobuf:"bytes,4,rep,name=tags" json:"tags,omitempty"` - // The Google Compute Engine metadata entries to add to all instances (see + Tags []string `protobuf:"bytes,4,rep,name=tags,proto3" json:"tags,omitempty"` + // The Compute Engine metadata entries to add to all instances (see // [Project and instance metadata](https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)). - Metadata map[string]string `protobuf:"bytes,5,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,5,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -426,7 +428,7 @@ func (m *GceClusterConfig) Reset() { *m = GceClusterConfig{} } func (m *GceClusterConfig) String() string { return proto.CompactTextString(m) } func (*GceClusterConfig) ProtoMessage() {} func (*GceClusterConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{2} + return fileDescriptor_clusters_92d23f68998e82a3, []int{2} } func (m *GceClusterConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GceClusterConfig.Unmarshal(m, b) @@ -502,51 +504,59 @@ func (m *GceClusterConfig) GetMetadata() map[string]string { return nil } -// Optional. The config settings for Google Compute Engine resources in +// Optional. The config settings for Compute Engine resources in // an instance group, such as a master or worker group. type InstanceGroupConfig struct { // Optional. The number of VM instances in the instance group. // For master instance groups, must be set to 1. - NumInstances int32 `protobuf:"varint,1,opt,name=num_instances,json=numInstances" json:"num_instances,omitempty"` - // Optional. The list of instance names. Cloud Dataproc derives the names from - // `cluster_name`, `num_instances`, and the instance group if not set by user - // (recommended practice is to let Cloud Dataproc derive the name). - InstanceNames []string `protobuf:"bytes,2,rep,name=instance_names,json=instanceNames" json:"instance_names,omitempty"` - // Output-only. The Google Compute Engine image resource used for cluster + NumInstances int32 `protobuf:"varint,1,opt,name=num_instances,json=numInstances,proto3" json:"num_instances,omitempty"` + // Output only. The list of instance names. Cloud Dataproc derives the names + // from `cluster_name`, `num_instances`, and the instance group. + InstanceNames []string `protobuf:"bytes,2,rep,name=instance_names,json=instanceNames,proto3" json:"instance_names,omitempty"` + // Output only. The Compute Engine image resource used for cluster // instances. Inferred from `SoftwareConfig.image_version`. - ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` - // Optional. The Google Compute Engine machine type used for cluster instances. + ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` + // Optional. The Compute Engine machine type used for cluster instances. // // A full URL, partial URI, or short name are valid. Examples: // // * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` // * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2` // * `n1-standard-2` - MachineTypeUri string `protobuf:"bytes,4,opt,name=machine_type_uri,json=machineTypeUri" json:"machine_type_uri,omitempty"` + // + // **Auto Zone Exception**: If you are using the Cloud Dataproc + // [Auto Zone Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + // feature, you must use the short name of the machine type + // resource, for example, `n1-standard-2`. + MachineTypeUri string `protobuf:"bytes,4,opt,name=machine_type_uri,json=machineTypeUri,proto3" json:"machine_type_uri,omitempty"` // Optional. Disk option config settings. - DiskConfig *DiskConfig `protobuf:"bytes,5,opt,name=disk_config,json=diskConfig" json:"disk_config,omitempty"` + DiskConfig *DiskConfig `protobuf:"bytes,5,opt,name=disk_config,json=diskConfig,proto3" json:"disk_config,omitempty"` // Optional. Specifies that this instance group contains preemptible instances. - IsPreemptible bool `protobuf:"varint,6,opt,name=is_preemptible,json=isPreemptible" json:"is_preemptible,omitempty"` - // Output-only. The config for Google Compute Engine Instance Group + IsPreemptible bool `protobuf:"varint,6,opt,name=is_preemptible,json=isPreemptible,proto3" json:"is_preemptible,omitempty"` + // Output only. The config for Compute Engine Instance Group // Manager that manages this group. // This is only used for preemptible instance groups. - ManagedGroupConfig *ManagedGroupConfig `protobuf:"bytes,7,opt,name=managed_group_config,json=managedGroupConfig" json:"managed_group_config,omitempty"` - // Optional. The Google Compute Engine accelerator configuration for these + ManagedGroupConfig *ManagedGroupConfig `protobuf:"bytes,7,opt,name=managed_group_config,json=managedGroupConfig,proto3" json:"managed_group_config,omitempty"` + // Optional. The Compute Engine accelerator configuration for these // instances. // // **Beta Feature**: This feature is still under development. It may be // changed before final release. - Accelerators []*AcceleratorConfig `protobuf:"bytes,8,rep,name=accelerators" json:"accelerators,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Accelerators []*AcceleratorConfig `protobuf:"bytes,8,rep,name=accelerators,proto3" json:"accelerators,omitempty"` + // Optional. Specifies the minimum cpu platform for the Instance Group. + // See [Cloud Dataproc→Minimum CPU Platform] + // (/dataproc/docs/concepts/compute/dataproc-min-cpu). + MinCpuPlatform string `protobuf:"bytes,9,opt,name=min_cpu_platform,json=minCpuPlatform,proto3" json:"min_cpu_platform,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *InstanceGroupConfig) Reset() { *m = InstanceGroupConfig{} } func (m *InstanceGroupConfig) String() string { return proto.CompactTextString(m) } func (*InstanceGroupConfig) ProtoMessage() {} func (*InstanceGroupConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{3} + return fileDescriptor_clusters_92d23f68998e82a3, []int{3} } func (m *InstanceGroupConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InstanceGroupConfig.Unmarshal(m, b) @@ -622,13 +632,20 @@ func (m *InstanceGroupConfig) GetAccelerators() []*AcceleratorConfig { return nil } +func (m *InstanceGroupConfig) GetMinCpuPlatform() string { + if m != nil { + return m.MinCpuPlatform + } + return "" +} + // Specifies the resources used to actively manage an instance group. type ManagedGroupConfig struct { - // Output-only. The name of the Instance Template used for the Managed + // Output only. The name of the Instance Template used for the Managed // Instance Group. - InstanceTemplateName string `protobuf:"bytes,1,opt,name=instance_template_name,json=instanceTemplateName" json:"instance_template_name,omitempty"` - // Output-only. The name of the Instance Group Manager for this group. - InstanceGroupManagerName string `protobuf:"bytes,2,opt,name=instance_group_manager_name,json=instanceGroupManagerName" json:"instance_group_manager_name,omitempty"` + InstanceTemplateName string `protobuf:"bytes,1,opt,name=instance_template_name,json=instanceTemplateName,proto3" json:"instance_template_name,omitempty"` + // Output only. The name of the Instance Group Manager for this group. + InstanceGroupManagerName string `protobuf:"bytes,2,opt,name=instance_group_manager_name,json=instanceGroupManagerName,proto3" json:"instance_group_manager_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -638,7 +655,7 @@ func (m *ManagedGroupConfig) Reset() { *m = ManagedGroupConfig{} } func (m *ManagedGroupConfig) String() string { return proto.CompactTextString(m) } func (*ManagedGroupConfig) ProtoMessage() {} func (*ManagedGroupConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{4} + return fileDescriptor_clusters_92d23f68998e82a3, []int{4} } func (m *ManagedGroupConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ManagedGroupConfig.Unmarshal(m, b) @@ -676,16 +693,21 @@ func (m *ManagedGroupConfig) GetInstanceGroupManagerName() string { // of an instance group (see [GPUs on Compute Engine](/compute/docs/gpus/)). type AcceleratorConfig struct { // Full URL, partial URI, or short name of the accelerator type resource to - // expose to this instance. See [Google Compute Engine AcceleratorTypes]( + // expose to this instance. See [Compute Engine AcceleratorTypes]( // /compute/docs/reference/beta/acceleratorTypes) // // Examples // * `https://www.googleapis.com/compute/beta/projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` // * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80` // * `nvidia-tesla-k80` - AcceleratorTypeUri string `protobuf:"bytes,1,opt,name=accelerator_type_uri,json=acceleratorTypeUri" json:"accelerator_type_uri,omitempty"` + // + // **Auto Zone Exception**: If you are using the Cloud Dataproc + // [Auto Zone Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement) + // feature, you must use the short name of the accelerator type + // resource, for example, `nvidia-tesla-k80`. + AcceleratorTypeUri string `protobuf:"bytes,1,opt,name=accelerator_type_uri,json=acceleratorTypeUri,proto3" json:"accelerator_type_uri,omitempty"` // The number of the accelerator cards of this type exposed to this instance. - AcceleratorCount int32 `protobuf:"varint,2,opt,name=accelerator_count,json=acceleratorCount" json:"accelerator_count,omitempty"` + AcceleratorCount int32 `protobuf:"varint,2,opt,name=accelerator_count,json=acceleratorCount,proto3" json:"accelerator_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -695,7 +717,7 @@ func (m *AcceleratorConfig) Reset() { *m = AcceleratorConfig{} } func (m *AcceleratorConfig) String() string { return proto.CompactTextString(m) } func (*AcceleratorConfig) ProtoMessage() {} func (*AcceleratorConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{5} + return fileDescriptor_clusters_92d23f68998e82a3, []int{5} } func (m *AcceleratorConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AcceleratorConfig.Unmarshal(m, b) @@ -731,15 +753,19 @@ func (m *AcceleratorConfig) GetAcceleratorCount() int32 { // Specifies the config of disk options for a group of VM instances. type DiskConfig struct { + // Optional. Type of the boot disk (default is "pd-standard"). + // Valid values: "pd-ssd" (Persistent Disk Solid State Drive) or + // "pd-standard" (Persistent Disk Hard Disk Drive). + BootDiskType string `protobuf:"bytes,3,opt,name=boot_disk_type,json=bootDiskType,proto3" json:"boot_disk_type,omitempty"` // Optional. Size in GB of the boot disk (default is 500GB). - BootDiskSizeGb int32 `protobuf:"varint,1,opt,name=boot_disk_size_gb,json=bootDiskSizeGb" json:"boot_disk_size_gb,omitempty"` + BootDiskSizeGb int32 `protobuf:"varint,1,opt,name=boot_disk_size_gb,json=bootDiskSizeGb,proto3" json:"boot_disk_size_gb,omitempty"` // Optional. Number of attached SSDs, from 0 to 4 (default is 0). // If SSDs are not attached, the boot disk is used to store runtime logs and // [HDFS](https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. // If one or more SSDs are attached, this runtime bulk // data is spread across them, and the boot disk contains only basic // config and installed binaries. - NumLocalSsds int32 `protobuf:"varint,2,opt,name=num_local_ssds,json=numLocalSsds" json:"num_local_ssds,omitempty"` + NumLocalSsds int32 `protobuf:"varint,2,opt,name=num_local_ssds,json=numLocalSsds,proto3" json:"num_local_ssds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -749,7 +775,7 @@ func (m *DiskConfig) Reset() { *m = DiskConfig{} } func (m *DiskConfig) String() string { return proto.CompactTextString(m) } func (*DiskConfig) ProtoMessage() {} func (*DiskConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{6} + return fileDescriptor_clusters_92d23f68998e82a3, []int{6} } func (m *DiskConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DiskConfig.Unmarshal(m, b) @@ -769,6 +795,13 @@ func (m *DiskConfig) XXX_DiscardUnknown() { var xxx_messageInfo_DiskConfig proto.InternalMessageInfo +func (m *DiskConfig) GetBootDiskType() string { + if m != nil { + return m.BootDiskType + } + return "" +} + func (m *DiskConfig) GetBootDiskSizeGb() int32 { if m != nil { return m.BootDiskSizeGb @@ -787,7 +820,10 @@ func (m *DiskConfig) GetNumLocalSsds() int32 { type LifecycleConfig struct { // Optional. The longest duration that cluster would keep alive while staying // idle; passing this threshold will cause cluster to be auto-deleted. - IdleDeleteTtl *duration.Duration `protobuf:"bytes,1,opt,name=idle_delete_ttl,json=idleDeleteTtl" json:"idle_delete_ttl,omitempty"` + IdleDeleteTtl *duration.Duration `protobuf:"bytes,1,opt,name=idle_delete_ttl,json=idleDeleteTtl,proto3" json:"idle_delete_ttl,omitempty"` + // Optional. Either the exact time the cluster should be deleted at or + // the cluster maximum age. + // // Types that are valid to be assigned to Ttl: // *LifecycleConfig_AutoDeleteTime // *LifecycleConfig_AutoDeleteTtl @@ -801,7 +837,7 @@ func (m *LifecycleConfig) Reset() { *m = LifecycleConfig{} } func (m *LifecycleConfig) String() string { return proto.CompactTextString(m) } func (*LifecycleConfig) ProtoMessage() {} func (*LifecycleConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{7} + return fileDescriptor_clusters_92d23f68998e82a3, []int{7} } func (m *LifecycleConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LifecycleConfig.Unmarshal(m, b) @@ -826,10 +862,10 @@ type isLifecycleConfig_Ttl interface { } type LifecycleConfig_AutoDeleteTime struct { - AutoDeleteTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=auto_delete_time,json=autoDeleteTime,oneof"` + AutoDeleteTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=auto_delete_time,json=autoDeleteTime,proto3,oneof"` } type LifecycleConfig_AutoDeleteTtl struct { - AutoDeleteTtl *duration.Duration `protobuf:"bytes,3,opt,name=auto_delete_ttl,json=autoDeleteTtl,oneof"` + AutoDeleteTtl *duration.Duration `protobuf:"bytes,3,opt,name=auto_delete_ttl,json=autoDeleteTtl,proto3,oneof"` } func (*LifecycleConfig_AutoDeleteTime) isLifecycleConfig_Ttl() {} @@ -940,13 +976,13 @@ func _LifecycleConfig_OneofSizer(msg proto.Message) (n int) { // Specifies an executable to run on a fully configured node and a // timeout period for executable completion. type NodeInitializationAction struct { - // Required. Google Cloud Storage URI of executable file. - ExecutableFile string `protobuf:"bytes,1,opt,name=executable_file,json=executableFile" json:"executable_file,omitempty"` + // Required. Cloud Storage URI of executable file. + ExecutableFile string `protobuf:"bytes,1,opt,name=executable_file,json=executableFile,proto3" json:"executable_file,omitempty"` // Optional. Amount of time executable has to complete. Default is // 10 minutes. Cluster creation fails with an explanatory error message (the // name of the executable that caused the error and the exceeded timeout // period) if the executable is not completed at end of the timeout period. - ExecutionTimeout *duration.Duration `protobuf:"bytes,2,opt,name=execution_timeout,json=executionTimeout" json:"execution_timeout,omitempty"` + ExecutionTimeout *duration.Duration `protobuf:"bytes,2,opt,name=execution_timeout,json=executionTimeout,proto3" json:"execution_timeout,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -956,7 +992,7 @@ func (m *NodeInitializationAction) Reset() { *m = NodeInitializationActi func (m *NodeInitializationAction) String() string { return proto.CompactTextString(m) } func (*NodeInitializationAction) ProtoMessage() {} func (*NodeInitializationAction) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{8} + return fileDescriptor_clusters_92d23f68998e82a3, []int{8} } func (m *NodeInitializationAction) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NodeInitializationAction.Unmarshal(m, b) @@ -992,15 +1028,15 @@ func (m *NodeInitializationAction) GetExecutionTimeout() *duration.Duration { // The status of a cluster and its instances. type ClusterStatus struct { - // Output-only. The cluster's state. - State ClusterStatus_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.dataproc.v1beta2.ClusterStatus_State" json:"state,omitempty"` - // Output-only. Optional details of cluster's state. - Detail string `protobuf:"bytes,2,opt,name=detail" json:"detail,omitempty"` - // Output-only. Time when this state was entered. - StateStartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=state_start_time,json=stateStartTime" json:"state_start_time,omitempty"` - // Output-only. Additional state information that includes + // Output only. The cluster's state. + State ClusterStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1beta2.ClusterStatus_State" json:"state,omitempty"` + // Output only. Optional details of cluster's state. + Detail string `protobuf:"bytes,2,opt,name=detail,proto3" json:"detail,omitempty"` + // Output only. Time when this state was entered. + StateStartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"` + // Output only. Additional state information that includes // status reported by the agent. - Substate ClusterStatus_Substate `protobuf:"varint,4,opt,name=substate,enum=google.cloud.dataproc.v1beta2.ClusterStatus_Substate" json:"substate,omitempty"` + Substate ClusterStatus_Substate `protobuf:"varint,4,opt,name=substate,proto3,enum=google.cloud.dataproc.v1beta2.ClusterStatus_Substate" json:"substate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1010,7 +1046,7 @@ func (m *ClusterStatus) Reset() { *m = ClusterStatus{} } func (m *ClusterStatus) String() string { return proto.CompactTextString(m) } func (*ClusterStatus) ProtoMessage() {} func (*ClusterStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{9} + return fileDescriptor_clusters_92d23f68998e82a3, []int{9} } func (m *ClusterStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterStatus.Unmarshal(m, b) @@ -1060,10 +1096,12 @@ func (m *ClusterStatus) GetSubstate() ClusterStatus_Substate { // Specifies the selection and config of software inside the cluster. type SoftwareConfig struct { - // Optional. The version of software inside the cluster. It must match the - // regular expression `[0-9]+\.[0-9]+`. If unspecified, it defaults to the - // latest version (see [Cloud Dataproc Versioning](/dataproc/versioning)). - ImageVersion string `protobuf:"bytes,1,opt,name=image_version,json=imageVersion" json:"image_version,omitempty"` + // Optional. The version of software inside the cluster. It must be one of the supported + // [Cloud Dataproc Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions), + // such as "1.2" (including a subminor version, such as "1.2.29"), or the + // ["preview" version](/dataproc/docs/concepts/versioning/dataproc-versions#other_versions). + // If unspecified, it defaults to the latest version. + ImageVersion string `protobuf:"bytes,1,opt,name=image_version,json=imageVersion,proto3" json:"image_version,omitempty"` // Optional. The properties to set on daemon config files. // // Property keys are specified in `prefix:property` format, such as @@ -1082,7 +1120,7 @@ type SoftwareConfig struct { // // For more information, see // [Cluster properties](/dataproc/docs/concepts/cluster-properties). - Properties map[string]string `protobuf:"bytes,2,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,2,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1092,7 +1130,7 @@ func (m *SoftwareConfig) Reset() { *m = SoftwareConfig{} } func (m *SoftwareConfig) String() string { return proto.CompactTextString(m) } func (*SoftwareConfig) ProtoMessage() {} func (*SoftwareConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{10} + return fileDescriptor_clusters_92d23f68998e82a3, []int{10} } func (m *SoftwareConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SoftwareConfig.Unmarshal(m, b) @@ -1132,9 +1170,9 @@ func (m *SoftwareConfig) GetProperties() map[string]string { // be changed before final release. type ClusterMetrics struct { // The HDFS metrics. - HdfsMetrics map[string]int64 `protobuf:"bytes,1,rep,name=hdfs_metrics,json=hdfsMetrics" json:"hdfs_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + HdfsMetrics map[string]int64 `protobuf:"bytes,1,rep,name=hdfs_metrics,json=hdfsMetrics,proto3" json:"hdfs_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` // The YARN metrics. - YarnMetrics map[string]int64 `protobuf:"bytes,2,rep,name=yarn_metrics,json=yarnMetrics" json:"yarn_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + YarnMetrics map[string]int64 `protobuf:"bytes,2,rep,name=yarn_metrics,json=yarnMetrics,proto3" json:"yarn_metrics,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1144,7 +1182,7 @@ func (m *ClusterMetrics) Reset() { *m = ClusterMetrics{} } func (m *ClusterMetrics) String() string { return proto.CompactTextString(m) } func (*ClusterMetrics) ProtoMessage() {} func (*ClusterMetrics) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{11} + return fileDescriptor_clusters_92d23f68998e82a3, []int{11} } func (m *ClusterMetrics) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterMetrics.Unmarshal(m, b) @@ -1182,11 +1220,23 @@ func (m *ClusterMetrics) GetYarnMetrics() map[string]int64 { type CreateClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster to create. - Cluster *Cluster `protobuf:"bytes,2,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,2,opt,name=cluster,proto3" json:"cluster,omitempty"` + // Optional. A unique id used to identify the request. If the server + // receives two [CreateClusterRequest][google.cloud.dataproc.v1beta2.CreateClusterRequest] requests with the same + // id, then the second request will be ignored and the + // first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the backend + // is returned. + // + // It is recommended to always set this value to a + // [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier). + // + // The id must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). The maximum length is 40 characters. + RequestId string `protobuf:"bytes,4,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1196,7 +1246,7 @@ func (m *CreateClusterRequest) Reset() { *m = CreateClusterRequest{} } func (m *CreateClusterRequest) String() string { return proto.CompactTextString(m) } func (*CreateClusterRequest) ProtoMessage() {} func (*CreateClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{12} + return fileDescriptor_clusters_92d23f68998e82a3, []int{12} } func (m *CreateClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateClusterRequest.Unmarshal(m, b) @@ -1237,17 +1287,24 @@ func (m *CreateClusterRequest) GetCluster() *Cluster { return nil } +func (m *CreateClusterRequest) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + // A request to update a cluster. type UpdateClusterRequest struct { // Required. The ID of the Google Cloud Platform project the // cluster belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,5,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,5,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Required. The changes to the cluster. - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` // Optional. Timeout for graceful YARN decomissioning. Graceful // decommissioning allows removing nodes from the cluster without // interrupting jobs in progress. Timeout specifies how long to wait for jobs @@ -1256,11 +1313,11 @@ type UpdateClusterRequest struct { // the maximum allowed timeout is 1 day. // // Only supported on Dataproc image versions 1.2 and higher. - GracefulDecommissionTimeout *duration.Duration `protobuf:"bytes,6,opt,name=graceful_decommission_timeout,json=gracefulDecommissionTimeout" json:"graceful_decommission_timeout,omitempty"` - // Required. Specifies the path, relative to Cluster, of + GracefulDecommissionTimeout *duration.Duration `protobuf:"bytes,6,opt,name=graceful_decommission_timeout,json=gracefulDecommissionTimeout,proto3" json:"graceful_decommission_timeout,omitempty"` + // Required. Specifies the path, relative to `Cluster`, of // the field to update. For example, to change the number of workers - // in a cluster to 5, the update_mask parameter would be - // specified as config.worker_config.num_instances, + // in a cluster to 5, the `update_mask` parameter would be + // specified as `config.worker_config.num_instances`, // and the `PATCH` request body would specify the new value, as follows: // // { @@ -1270,8 +1327,9 @@ type UpdateClusterRequest struct { // } // } // } + // // Similarly, to change the number of preemptible workers in a cluster to 5, the - // update_mask parameter would be config.secondary_worker_config.num_instances, + // `update_mask` parameter would be `config.secondary_worker_config.num_instances`, // and the `PATCH` request body would be set as follows: // // { @@ -1281,22 +1339,54 @@ type UpdateClusterRequest struct { // } // } // } - // Note: currently only some fields can be updated: - // |Mask|Purpose| - // |`labels`|Updates labels| - // |`config.worker_config.num_instances`|Resize primary worker group| - // |`config.secondary_worker_config.num_instances`|Resize secondary worker group| - UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + // Note: currently only the following fields can be updated: + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + // + //
MaskPurpose
labelsUpdates labels
config.worker_config.num_instancesResize primary worker group
config.secondary_worker_config.num_instancesResize secondary worker group
config.lifecycle_config.auto_delete_ttlReset MAX TTL duration
config.lifecycle_config.auto_delete_timeUpdate MAX TTL deletion timestamp
config.lifecycle_config.idle_delete_ttlUpdate Idle TTL duration
+ UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` + // Optional. A unique id used to identify the request. If the server + // receives two [UpdateClusterRequest][google.cloud.dataproc.v1beta2.UpdateClusterRequest] requests with the same + // id, then the second request will be ignored and the + // first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the + // backend is returned. + // + // It is recommended to always set this value to a + // [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier). + // + // The id must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). The maximum length is 40 characters. + RequestId string `protobuf:"bytes,7,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *UpdateClusterRequest) Reset() { *m = UpdateClusterRequest{} } func (m *UpdateClusterRequest) String() string { return proto.CompactTextString(m) } func (*UpdateClusterRequest) ProtoMessage() {} func (*UpdateClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{13} + return fileDescriptor_clusters_92d23f68998e82a3, []int{13} } func (m *UpdateClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateClusterRequest.Unmarshal(m, b) @@ -1358,18 +1448,37 @@ func (m *UpdateClusterRequest) GetUpdateMask() *field_mask.FieldMask { return nil } +func (m *UpdateClusterRequest) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + // A request to delete a cluster. type DeleteClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Optional. Specifying the `cluster_uuid` means the RPC should fail // (with error NOT_FOUND) if cluster with specified UUID does not exist. - ClusterUuid string `protobuf:"bytes,4,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` + ClusterUuid string `protobuf:"bytes,4,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` + // Optional. A unique id used to identify the request. If the server + // receives two [DeleteClusterRequest][google.cloud.dataproc.v1beta2.DeleteClusterRequest] requests with the same + // id, then the second request will be ignored and the + // first [google.longrunning.Operation][google.longrunning.Operation] created and stored in the + // backend is returned. + // + // It is recommended to always set this value to a + // [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier). + // + // The id must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). The maximum length is 40 characters. + RequestId string `protobuf:"bytes,5,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1379,7 +1488,7 @@ func (m *DeleteClusterRequest) Reset() { *m = DeleteClusterRequest{} } func (m *DeleteClusterRequest) String() string { return proto.CompactTextString(m) } func (*DeleteClusterRequest) ProtoMessage() {} func (*DeleteClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{14} + return fileDescriptor_clusters_92d23f68998e82a3, []int{14} } func (m *DeleteClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteClusterRequest.Unmarshal(m, b) @@ -1427,15 +1536,22 @@ func (m *DeleteClusterRequest) GetClusterUuid() string { return "" } +func (m *DeleteClusterRequest) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + // Request to get the resource representation for a cluster in a project. type GetClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1445,7 +1561,7 @@ func (m *GetClusterRequest) Reset() { *m = GetClusterRequest{} } func (m *GetClusterRequest) String() string { return proto.CompactTextString(m) } func (*GetClusterRequest) ProtoMessage() {} func (*GetClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{15} + return fileDescriptor_clusters_92d23f68998e82a3, []int{15} } func (m *GetClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetClusterRequest.Unmarshal(m, b) @@ -1490,9 +1606,9 @@ func (m *GetClusterRequest) GetClusterName() string { type ListClustersRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,4,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,4,opt,name=region,proto3" json:"region,omitempty"` // Optional. A filter constraining the clusters to list. Filters are // case-sensitive and have the following syntax: // @@ -1512,11 +1628,11 @@ type ListClustersRequest struct { // // status.state = ACTIVE AND clusterName = mycluster // AND labels.env = staging AND labels.starred = * - Filter string `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // Optional. The standard List page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The standard List page token. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1526,7 +1642,7 @@ func (m *ListClustersRequest) Reset() { *m = ListClustersRequest{} } func (m *ListClustersRequest) String() string { return proto.CompactTextString(m) } func (*ListClustersRequest) ProtoMessage() {} func (*ListClustersRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{16} + return fileDescriptor_clusters_92d23f68998e82a3, []int{16} } func (m *ListClustersRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListClustersRequest.Unmarshal(m, b) @@ -1583,12 +1699,12 @@ func (m *ListClustersRequest) GetPageToken() string { // The list of all clusters in a project. type ListClustersResponse struct { - // Output-only. The clusters in the project. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` - // Output-only. This token is included in the response if there are more + // Output only. The clusters in the project. + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` + // Output only. This token is included in the response if there are more // results to fetch. To fetch additional results, provide this value as the // `page_token` in a subsequent ListClustersRequest. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1598,7 +1714,7 @@ func (m *ListClustersResponse) Reset() { *m = ListClustersResponse{} } func (m *ListClustersResponse) String() string { return proto.CompactTextString(m) } func (*ListClustersResponse) ProtoMessage() {} func (*ListClustersResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{17} + return fileDescriptor_clusters_92d23f68998e82a3, []int{17} } func (m *ListClustersResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListClustersResponse.Unmarshal(m, b) @@ -1636,11 +1752,11 @@ func (m *ListClustersResponse) GetNextPageToken() string { type DiagnoseClusterRequest struct { // Required. The ID of the Google Cloud Platform project that the cluster // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The cluster name. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1650,7 +1766,7 @@ func (m *DiagnoseClusterRequest) Reset() { *m = DiagnoseClusterRequest{} func (m *DiagnoseClusterRequest) String() string { return proto.CompactTextString(m) } func (*DiagnoseClusterRequest) ProtoMessage() {} func (*DiagnoseClusterRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{18} + return fileDescriptor_clusters_92d23f68998e82a3, []int{18} } func (m *DiagnoseClusterRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DiagnoseClusterRequest.Unmarshal(m, b) @@ -1693,10 +1809,10 @@ func (m *DiagnoseClusterRequest) GetClusterName() string { // The location of diagnostic output. type DiagnoseClusterResults struct { - // Output-only. The Google Cloud Storage URI of the diagnostic output. + // Output only. The Cloud Storage URI of the diagnostic output. // The output report is a plain text file with a summary of collected // diagnostics. - OutputUri string `protobuf:"bytes,1,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"` + OutputUri string `protobuf:"bytes,1,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1706,7 +1822,7 @@ func (m *DiagnoseClusterResults) Reset() { *m = DiagnoseClusterResults{} func (m *DiagnoseClusterResults) String() string { return proto.CompactTextString(m) } func (*DiagnoseClusterResults) ProtoMessage() {} func (*DiagnoseClusterResults) Descriptor() ([]byte, []int) { - return fileDescriptor_clusters_11ad3d3844eb379a, []int{19} + return fileDescriptor_clusters_92d23f68998e82a3, []int{19} } func (m *DiagnoseClusterResults) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DiagnoseClusterResults.Unmarshal(m, b) @@ -1853,8 +1969,7 @@ func (c *clusterControllerClient) DiagnoseCluster(ctx context.Context, in *Diagn return out, nil } -// Server API for ClusterController service - +// ClusterControllerServer is the server API for ClusterController service. type ClusterControllerServer interface { // Creates a cluster in a project. CreateCluster(context.Context, *CreateClusterRequest) (*longrunning.Operation, error) @@ -2018,140 +2133,145 @@ var _ClusterController_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/cloud/dataproc/v1beta2/clusters.proto", fileDescriptor_clusters_11ad3d3844eb379a) + proto.RegisterFile("google/cloud/dataproc/v1beta2/clusters.proto", fileDescriptor_clusters_92d23f68998e82a3) } -var fileDescriptor_clusters_11ad3d3844eb379a = []byte{ - // 2093 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x59, 0xcd, 0x72, 0x1b, 0xc7, - 0x11, 0xd6, 0x12, 0x04, 0x09, 0x36, 0x08, 0x10, 0x1c, 0xd3, 0x0c, 0x4c, 0x45, 0xb1, 0xbc, 0x4e, - 0x1c, 0xda, 0x71, 0x00, 0x9b, 0x8a, 0xcb, 0x8e, 0x14, 0xb9, 0x4c, 0x91, 0x14, 0x49, 0x87, 0xa2, - 0x98, 0x05, 0x20, 0x45, 0x49, 0xa9, 0xb6, 0x06, 0xbb, 0x43, 0x68, 0xcc, 0xfd, 0xcb, 0xce, 0xac, - 0x6c, 0x48, 0xa5, 0x8b, 0x6f, 0xa9, 0x1c, 0x72, 0xf0, 0x03, 0xc4, 0xe7, 0x1c, 0x72, 0x4a, 0x55, - 0x2a, 0x87, 0xdc, 0x72, 0xce, 0xc5, 0xa9, 0x3c, 0x41, 0x0e, 0x79, 0x84, 0x1c, 0x53, 0xf3, 0xb3, - 0xc0, 0x2e, 0x48, 0x69, 0x09, 0x46, 0xe5, 0x13, 0x76, 0x7a, 0xfa, 0xe7, 0x9b, 0xee, 0x9e, 0x9e, - 0x9e, 0x01, 0xbc, 0x3b, 0x08, 0xc3, 0x81, 0x47, 0xda, 0x8e, 0x17, 0x26, 0x6e, 0xdb, 0xc5, 0x1c, - 0x47, 0x71, 0xe8, 0xb4, 0x1f, 0xbf, 0xdf, 0x27, 0x1c, 0x6f, 0xb4, 0x1d, 0x2f, 0x61, 0x9c, 0xc4, - 0xac, 0x15, 0xc5, 0x21, 0x0f, 0xd1, 0x15, 0xc5, 0xdd, 0x92, 0xdc, 0xad, 0x94, 0xbb, 0xa5, 0xb9, - 0xd7, 0xbe, 0xab, 0x95, 0xe1, 0x88, 0xb6, 0x71, 0x10, 0x84, 0x1c, 0x73, 0x1a, 0x06, 0x5a, 0x78, - 0xed, 0x4d, 0x3d, 0xeb, 0x85, 0xc1, 0x20, 0x4e, 0x82, 0x80, 0x06, 0x83, 0x76, 0x18, 0x91, 0x38, - 0xc7, 0xf4, 0x3d, 0xcd, 0x24, 0x47, 0xfd, 0xe4, 0xb8, 0xed, 0x26, 0x8a, 0x41, 0xcf, 0x5f, 0x9d, - 0x9c, 0x3f, 0xa6, 0xc4, 0x73, 0x6d, 0x1f, 0xb3, 0x13, 0xcd, 0xf1, 0xfa, 0x24, 0x07, 0xa7, 0x3e, - 0x61, 0x1c, 0xfb, 0x91, 0x62, 0x30, 0xff, 0x30, 0x0b, 0xf3, 0x5b, 0x6a, 0x5d, 0xe8, 0x0a, 0x40, - 0x14, 0x87, 0x9f, 0x11, 0x87, 0xdb, 0xd4, 0x6d, 0x1a, 0x57, 0x8d, 0xf5, 0x05, 0x6b, 0x41, 0x53, - 0xf6, 0x5d, 0xf4, 0x06, 0x2c, 0x6a, 0x0f, 0xd8, 0x01, 0xf6, 0x49, 0x73, 0x46, 0x32, 0x54, 0x35, - 0xed, 0x10, 0xfb, 0x04, 0x6d, 0xc3, 0x9c, 0x13, 0x06, 0xc7, 0x74, 0xd0, 0x2c, 0x5d, 0x35, 0xd6, - 0xab, 0x1b, 0xef, 0xb6, 0x5e, 0xe8, 0xa3, 0x96, 0xb6, 0xbc, 0x25, 0x65, 0x2c, 0x2d, 0x8b, 0x3e, - 0x85, 0x39, 0x0f, 0xf7, 0x89, 0xc7, 0x9a, 0x95, 0xab, 0xa5, 0xf5, 0xea, 0xc6, 0xc6, 0xf9, 0xb4, - 0xb4, 0x0e, 0xa4, 0xd0, 0x4e, 0xc0, 0xe3, 0xa1, 0xa5, 0x35, 0x08, 0x44, 0x8c, 0x63, 0x9e, 0xb0, - 0xe6, 0xec, 0x34, 0x88, 0x3a, 0x52, 0xc6, 0xd2, 0xb2, 0xa8, 0x03, 0x75, 0xf5, 0x65, 0x3f, 0xa2, - 0x8c, 0x87, 0xf1, 0xb0, 0x39, 0x2f, 0x91, 0x4d, 0xa7, 0xad, 0xa6, 0x74, 0xec, 0x29, 0x15, 0x59, - 0x7f, 0x26, 0x09, 0x75, 0x9b, 0x73, 0x39, 0x7f, 0xf6, 0x12, 0xea, 0xa2, 0x5d, 0x98, 0xf7, 0x09, - 0x8f, 0xa9, 0xc3, 0x9a, 0x0b, 0x12, 0xfe, 0x8f, 0xcf, 0x67, 0xf0, 0x8e, 0x12, 0xb2, 0x52, 0xe9, - 0xb5, 0x9f, 0x42, 0x35, 0xe3, 0x1d, 0xd4, 0x80, 0xd2, 0x09, 0x19, 0xea, 0x10, 0x8b, 0x4f, 0xb4, - 0x02, 0xe5, 0xc7, 0xd8, 0x4b, 0xd2, 0xa8, 0xaa, 0xc1, 0xf5, 0x99, 0x8f, 0x0c, 0xf3, 0x9f, 0x65, - 0xa8, 0xe5, 0xe2, 0x84, 0xde, 0x84, 0x9a, 0x8a, 0x94, 0xdd, 0x4f, 0x9c, 0x13, 0xc2, 0xb5, 0x9e, - 0x45, 0x45, 0xbc, 0x25, 0x69, 0xe8, 0x21, 0xa0, 0x81, 0x43, 0xec, 0x74, 0x85, 0x3a, 0x2d, 0x2a, - 0x72, 0x15, 0xed, 0x82, 0x55, 0xec, 0x3a, 0x24, 0x9f, 0x19, 0x8d, 0xc1, 0x04, 0x05, 0xdd, 0x87, - 0x9a, 0x8f, 0xb3, 0x9a, 0x95, 0x7f, 0x8a, 0x52, 0x65, 0x3f, 0x60, 0x1c, 0x07, 0x0e, 0xd9, 0x8d, - 0xc3, 0x24, 0xd2, 0xca, 0x17, 0x95, 0xa2, 0xb1, 0xe2, 0xcf, 0xc3, 0xf8, 0x64, 0xac, 0x18, 0x2e, - 0xae, 0x58, 0x29, 0xd2, 0x8a, 0x3f, 0x83, 0xef, 0x30, 0xe2, 0x84, 0x81, 0x8b, 0xe3, 0xa1, 0x9d, - 0x37, 0xb1, 0x78, 0x61, 0x13, 0xaf, 0x8e, 0x54, 0xde, 0xcf, 0xda, 0xba, 0x07, 0x4b, 0x2c, 0x3c, - 0xe6, 0x9f, 0xe3, 0x98, 0xa4, 0x36, 0x6a, 0xe7, 0xca, 0x9f, 0x8e, 0x96, 0xd2, 0xea, 0xeb, 0x2c, - 0x37, 0x46, 0x0f, 0xa0, 0xe1, 0xd1, 0x63, 0xe2, 0x0c, 0x1d, 0x6f, 0xa4, 0xb8, 0x2e, 0x15, 0xb7, - 0x0a, 0x14, 0x1f, 0xa4, 0x62, 0x5a, 0xf3, 0x92, 0x97, 0x27, 0xa0, 0x00, 0x56, 0x69, 0x40, 0x39, - 0xc5, 0x1e, 0x7d, 0x22, 0x6b, 0x9c, 0x8d, 0x1d, 0x59, 0x0b, 0x9b, 0x55, 0xb9, 0xd5, 0x3e, 0x2c, - 0x30, 0x70, 0x18, 0xba, 0x64, 0x3f, 0xa7, 0x60, 0x53, 0xca, 0x5b, 0xaf, 0xd2, 0x33, 0xa8, 0xcc, - 0xfc, 0x63, 0x09, 0x1a, 0x93, 0x79, 0x86, 0x5e, 0x83, 0xca, 0x93, 0x30, 0x20, 0x76, 0x12, 0x53, - 0x9d, 0xd4, 0xf3, 0x62, 0xdc, 0x8b, 0x29, 0x7a, 0x1d, 0xaa, 0x01, 0xe1, 0x22, 0x6e, 0x72, 0x56, - 0x6d, 0x13, 0xd0, 0x24, 0xc1, 0xf0, 0x03, 0xa8, 0xb3, 0xa4, 0x9f, 0xe5, 0x51, 0x1b, 0xba, 0x36, - 0xa6, 0x0a, 0xb6, 0x75, 0x68, 0xd0, 0x80, 0x93, 0x38, 0xc0, 0x9e, 0x4d, 0x23, 0x3b, 0x0c, 0x3c, - 0x51, 0x4c, 0x8c, 0xf5, 0x8a, 0x55, 0x4f, 0xe9, 0xfb, 0xd1, 0xdd, 0xc0, 0x1b, 0xa2, 0x1f, 0xc2, - 0x12, 0x23, 0xf1, 0x63, 0xea, 0x10, 0x1b, 0x3b, 0x4e, 0x98, 0x04, 0x5c, 0x6e, 0x9f, 0x05, 0xab, - 0xae, 0xc9, 0x9b, 0x8a, 0x8a, 0x7e, 0x02, 0xab, 0x13, 0x8c, 0x36, 0x73, 0xc2, 0x88, 0xb0, 0x66, - 0xe9, 0x6a, 0x69, 0x7d, 0xc1, 0x5a, 0xc9, 0xf3, 0x77, 0xe4, 0x1c, 0x42, 0x30, 0xcb, 0xf1, 0x40, - 0xd4, 0x45, 0xc1, 0x23, 0xbf, 0xd1, 0x03, 0xa8, 0xf8, 0x84, 0x63, 0xe1, 0xdc, 0x66, 0x59, 0xba, - 0xfd, 0xe6, 0x94, 0x5b, 0xb5, 0x75, 0x47, 0xcb, 0xab, 0x32, 0x3c, 0x52, 0xb7, 0x76, 0x03, 0x6a, - 0xb9, 0xa9, 0xa9, 0x6a, 0xd0, 0xbf, 0x4a, 0xf0, 0xca, 0x19, 0xe9, 0x2f, 0x2a, 0x51, 0x90, 0xf8, - 0x36, 0xd5, 0x53, 0x4c, 0x6a, 0x2b, 0x5b, 0x8b, 0x41, 0xe2, 0xa7, 0xec, 0x4c, 0x04, 0x26, 0x65, - 0x90, 0x07, 0x17, 0x6b, 0xce, 0xc8, 0x25, 0xd7, 0x52, 0xaa, 0x38, 0xba, 0x18, 0xba, 0x0c, 0x0b, - 0xd4, 0xc7, 0x03, 0x15, 0xfc, 0x92, 0x44, 0x50, 0x91, 0x04, 0x1d, 0x35, 0x1f, 0x3b, 0x8f, 0x68, - 0x40, 0x6c, 0x3e, 0x8c, 0x14, 0xcf, 0xac, 0x0a, 0x86, 0xa6, 0x77, 0x87, 0x91, 0xe4, 0xfc, 0x14, - 0xaa, 0x2e, 0x65, 0x27, 0xe9, 0xee, 0x28, 0xcb, 0xdd, 0xf1, 0x76, 0x81, 0x17, 0xb7, 0x29, 0x3b, - 0xd1, 0x1b, 0x03, 0xdc, 0xd1, 0xb7, 0x44, 0xce, 0xec, 0x28, 0x26, 0xc4, 0x8f, 0x38, 0xed, 0x7b, - 0x44, 0xa6, 0x54, 0xc5, 0xaa, 0x51, 0x76, 0x34, 0x26, 0x22, 0x07, 0x56, 0x7c, 0x1c, 0xe0, 0x01, - 0x71, 0xed, 0x81, 0x70, 0x4e, 0x6a, 0x7b, 0x5e, 0xda, 0x7e, 0xbf, 0xc0, 0xf6, 0x1d, 0x25, 0x9a, - 0xad, 0x2a, 0xc8, 0x3f, 0x45, 0x43, 0x5d, 0x58, 0xc4, 0x8e, 0x43, 0x3c, 0xd1, 0xa2, 0x84, 0x71, - 0x7a, 0x34, 0xbf, 0x57, 0xa0, 0x7c, 0x73, 0x2c, 0x92, 0x16, 0xc5, 0xac, 0x16, 0xf3, 0xb7, 0x06, - 0xa0, 0xd3, 0x00, 0x44, 0x46, 0x8f, 0x42, 0xc6, 0x89, 0x1f, 0x79, 0x98, 0xab, 0xd8, 0xe9, 0x74, - 0x59, 0x49, 0x67, 0xbb, 0x7a, 0x52, 0x76, 0x1f, 0x37, 0xe1, 0xf2, 0x48, 0x4a, 0x39, 0x42, 0xad, - 0x23, 0xd7, 0xaf, 0x34, 0x69, 0x36, 0x8f, 0x94, 0x6d, 0xd9, 0xbc, 0x98, 0x31, 0x2c, 0x9f, 0x82, - 0x8b, 0xde, 0x83, 0x95, 0x0c, 0xe0, 0x71, 0xf0, 0x15, 0x0e, 0x94, 0x99, 0x4b, 0x13, 0xe0, 0x47, - 0xb0, 0x9c, 0x95, 0x50, 0x1b, 0x77, 0x46, 0xe6, 0x65, 0x03, 0x67, 0xf5, 0x27, 0x01, 0x37, 0x1f, - 0x02, 0x8c, 0x63, 0x8f, 0xde, 0x86, 0xe5, 0x7e, 0x18, 0x72, 0x5b, 0x26, 0x10, 0xa3, 0x4f, 0x88, - 0x3d, 0xe8, 0xeb, 0x94, 0xae, 0x8b, 0x09, 0xc1, 0xda, 0xa1, 0x4f, 0xc8, 0x6e, 0x1f, 0x7d, 0x1f, - 0xea, 0x22, 0xf3, 0xbd, 0xd0, 0xc1, 0x9e, 0xcd, 0x98, 0xcb, 0xb4, 0x09, 0x91, 0xfa, 0x07, 0x82, - 0xd8, 0x61, 0x2e, 0x33, 0xff, 0x63, 0xc0, 0xd2, 0x44, 0xe5, 0x45, 0x9b, 0xb0, 0x44, 0x5d, 0x8f, - 0xd8, 0x2e, 0xf1, 0x08, 0x27, 0x36, 0xe7, 0x9e, 0x34, 0x51, 0xdd, 0x78, 0x2d, 0x8d, 0x65, 0xda, - 0x2c, 0xb6, 0xb6, 0x75, 0xbb, 0x69, 0xd5, 0x84, 0xc4, 0xb6, 0x14, 0xe8, 0x72, 0x0f, 0xdd, 0x86, - 0x06, 0x4e, 0x78, 0x38, 0x52, 0x41, 0xb5, 0x77, 0xab, 0x1b, 0x6b, 0xa7, 0x74, 0x74, 0xd3, 0x86, - 0x73, 0xef, 0x92, 0x55, 0x17, 0x52, 0x5a, 0x0d, 0xf5, 0x09, 0xda, 0x82, 0xa5, 0x9c, 0x1e, 0xee, - 0xe9, 0xbe, 0xf1, 0xf9, 0x50, 0xf6, 0x2e, 0x59, 0xb5, 0x8c, 0x16, 0xee, 0xdd, 0x2a, 0x43, 0x89, - 0x73, 0xcf, 0xfc, 0x9d, 0x01, 0xcd, 0xe7, 0x9d, 0x01, 0xa2, 0x94, 0x92, 0x2f, 0x88, 0x93, 0x70, - 0xdc, 0xf7, 0x88, 0x7d, 0x4c, 0xbd, 0x34, 0x91, 0xea, 0x63, 0xf2, 0x6d, 0xea, 0x11, 0x74, 0x1b, - 0x96, 0x15, 0x45, 0x1c, 0x40, 0x62, 0x5d, 0x61, 0xc2, 0xf5, 0xd2, 0x5e, 0xe0, 0x9e, 0xc6, 0x48, - 0xa6, 0xab, 0x44, 0xcc, 0xaf, 0x4b, 0xa3, 0xa6, 0x49, 0x35, 0x7f, 0x68, 0x0f, 0xca, 0xa2, 0xfd, - 0x53, 0x86, 0xeb, 0xe7, 0xed, 0x69, 0x95, 0x70, 0x4b, 0xfc, 0x10, 0x4b, 0x29, 0x40, 0xab, 0x30, - 0xe7, 0x12, 0x8e, 0xa9, 0xa7, 0x33, 0x5a, 0x8f, 0xd0, 0x36, 0x34, 0x24, 0x83, 0xcd, 0x38, 0x8e, - 0xb9, 0x8a, 0x4a, 0xa9, 0x28, 0x2a, 0x96, 0x6c, 0x6c, 0x49, 0x47, 0x88, 0xc8, 0x98, 0xfc, 0x02, - 0x2a, 0x2c, 0xe9, 0x2b, 0xa8, 0xb3, 0x12, 0xea, 0x07, 0xd3, 0x41, 0xd5, 0xc2, 0xd6, 0x48, 0x8d, - 0x79, 0x0f, 0xca, 0x72, 0x01, 0xa8, 0x0a, 0xf3, 0xbd, 0xc3, 0x9f, 0x1f, 0xde, 0xbd, 0x7f, 0xd8, - 0xb8, 0x84, 0x16, 0xa1, 0xb2, 0x65, 0xed, 0x6c, 0x76, 0xf7, 0x0f, 0x77, 0x1b, 0x86, 0x98, 0xb2, - 0x7a, 0x87, 0x87, 0x62, 0x30, 0x83, 0x16, 0xa0, 0xbc, 0x63, 0x59, 0x77, 0xad, 0x46, 0x49, 0x70, - 0x6d, 0xef, 0x1c, 0xec, 0x48, 0xae, 0x59, 0x31, 0xea, 0x1d, 0x6d, 0x2b, 0x99, 0xb2, 0xf9, 0x33, - 0xa8, 0xa4, 0xd6, 0xd0, 0x12, 0x54, 0x7b, 0x87, 0x9d, 0xa3, 0x9d, 0xad, 0xfd, 0xdb, 0xfb, 0x3b, - 0xdb, 0x8d, 0x4b, 0xa8, 0x06, 0x0b, 0xbd, 0xc3, 0xbd, 0x9d, 0xcd, 0x83, 0xee, 0xde, 0x83, 0x86, - 0x81, 0x1a, 0xb0, 0xd8, 0xe9, 0x6e, 0x1e, 0xec, 0xd8, 0x9d, 0xee, 0x66, 0xb7, 0xd7, 0x69, 0xcc, - 0x98, 0xdf, 0x18, 0x50, 0xcf, 0xb7, 0x3b, 0xe2, 0x38, 0x51, 0x47, 0xc0, 0x63, 0x12, 0x33, 0x1a, - 0x06, 0x69, 0x63, 0x2b, 0x89, 0xf7, 0x14, 0x0d, 0x3d, 0x94, 0xb7, 0xa4, 0x88, 0xc4, 0x9c, 0xea, - 0xa3, 0xa4, 0xf8, 0x94, 0xcc, 0xdb, 0x69, 0x1d, 0x8d, 0xe4, 0xd5, 0x29, 0x99, 0x51, 0xb8, 0x76, - 0x13, 0x96, 0x26, 0xa6, 0xa7, 0x3b, 0x29, 0x67, 0xa0, 0x9e, 0xbf, 0x04, 0x20, 0x0c, 0x8b, 0x8f, - 0xdc, 0x63, 0x66, 0xa7, 0x37, 0x09, 0x43, 0x42, 0xfe, 0x78, 0xaa, 0x9b, 0x44, 0x6b, 0xcf, 0x3d, - 0x66, 0xfa, 0x5b, 0x61, 0xae, 0x3e, 0x1a, 0x53, 0x84, 0x89, 0x21, 0x8e, 0x83, 0x91, 0x89, 0x99, - 0x8b, 0x98, 0x78, 0x80, 0xe3, 0x20, 0x6f, 0x62, 0x38, 0xa6, 0xac, 0x7d, 0x0c, 0x8d, 0x49, 0x0c, - 0x45, 0x8e, 0x29, 0x65, 0x1c, 0x23, 0xe4, 0x27, 0x0d, 0x4c, 0x23, 0x6f, 0xfe, 0xde, 0x80, 0x95, - 0xad, 0x98, 0x60, 0x9e, 0xf6, 0x3b, 0x16, 0xf9, 0x4d, 0x42, 0x18, 0x2f, 0xba, 0x35, 0xaf, 0xc2, - 0x5c, 0x4c, 0x06, 0x22, 0x99, 0x54, 0x4f, 0xa1, 0x47, 0xe8, 0x13, 0x98, 0xd7, 0x77, 0x23, 0x5d, - 0x5f, 0xde, 0x3a, 0x9f, 0xb7, 0xac, 0x54, 0xcc, 0xfc, 0xc7, 0x0c, 0xac, 0xf4, 0x22, 0xf7, 0xff, - 0x40, 0x54, 0xce, 0x21, 0x3a, 0xc7, 0xfd, 0x3e, 0x03, 0xba, 0x74, 0x21, 0xd0, 0xe8, 0x21, 0x5c, - 0x19, 0xc4, 0xd8, 0x21, 0xc7, 0x89, 0x67, 0xbb, 0xc4, 0x09, 0x7d, 0x9f, 0x32, 0x96, 0x2d, 0xb6, - 0x73, 0x45, 0xc5, 0xf6, 0x72, 0x2a, 0xbf, 0x9d, 0x11, 0xd7, 0x75, 0x17, 0xdd, 0x80, 0x6a, 0x22, - 0x5d, 0x22, 0x1f, 0x41, 0xf4, 0x9d, 0xff, 0x74, 0xf9, 0xbb, 0x4d, 0x89, 0xe7, 0xde, 0xc1, 0xec, - 0xc4, 0x02, 0xc5, 0x2e, 0xbe, 0xcd, 0xaf, 0x0c, 0x58, 0x51, 0xe7, 0xca, 0xcb, 0x09, 0xf1, 0x39, - 0x1c, 0x3a, 0xf9, 0x06, 0x30, 0x7b, 0xea, 0x0d, 0xc0, 0xf4, 0x61, 0x79, 0x97, 0xf0, 0x6f, 0x0b, - 0x91, 0xf9, 0xb5, 0x01, 0xaf, 0x1c, 0x50, 0x96, 0x1a, 0x64, 0x53, 0x5b, 0x9c, 0xcd, 0x59, 0x5c, - 0x85, 0xb9, 0x63, 0xea, 0x89, 0x84, 0xd1, 0xc9, 0xa6, 0x46, 0xa2, 0xdb, 0x8e, 0x44, 0xa5, 0x15, - 0x5d, 0x8e, 0x6e, 0x5d, 0x2a, 0x82, 0x20, 0xda, 0x1b, 0x69, 0x4b, 0x4c, 0xf2, 0xf0, 0x84, 0xa4, - 0x4b, 0x90, 0xec, 0x5d, 0x41, 0x30, 0xbf, 0x34, 0x60, 0x25, 0x0f, 0x91, 0x45, 0x61, 0xc0, 0x08, - 0xba, 0x05, 0x95, 0xf4, 0x8d, 0x4e, 0x57, 0xb9, 0xf3, 0xe6, 0xe7, 0x48, 0x0e, 0xbd, 0x05, 0x4b, - 0x01, 0xf9, 0x82, 0xdb, 0x19, 0x00, 0xca, 0x4b, 0x35, 0x41, 0x3e, 0x1a, 0x81, 0x88, 0x61, 0x75, - 0x9b, 0xe2, 0x41, 0x10, 0xb2, 0x6f, 0x2d, 0x5b, 0xcc, 0x0f, 0xcf, 0xb0, 0xc9, 0x12, 0x8f, 0x33, - 0x61, 0x33, 0x4c, 0x78, 0x94, 0xf0, 0x4c, 0x73, 0xba, 0xa0, 0x28, 0xbd, 0x98, 0x6e, 0xfc, 0xb7, - 0x02, 0xcb, 0xe3, 0x6b, 0x1a, 0x8f, 0x43, 0xcf, 0x23, 0x31, 0xfa, 0x93, 0x01, 0xb5, 0x5c, 0x49, - 0x43, 0xd7, 0x8a, 0xdc, 0x75, 0x46, 0x01, 0x5c, 0xbb, 0x92, 0x0a, 0x65, 0xde, 0x32, 0x5b, 0x77, - 0xd3, 0xb7, 0x4c, 0x73, 0xff, 0xcb, 0x6f, 0xfe, 0xfd, 0xd5, 0xcc, 0x96, 0xf9, 0xd1, 0xe8, 0x1d, - 0x55, 0xfb, 0x82, 0xb5, 0x9f, 0x8e, 0xfd, 0xf4, 0xac, 0xad, 0xdc, 0xc0, 0xda, 0x4f, 0xd5, 0xc7, - 0xb3, 0xd1, 0x73, 0xeb, 0xf5, 0x51, 0xf1, 0xf8, 0x9b, 0x01, 0xb5, 0x5c, 0xc5, 0x2b, 0x04, 0x7c, - 0x56, 0x7d, 0x2c, 0x02, 0xfc, 0x4b, 0x09, 0xd8, 0xda, 0xd8, 0xbd, 0x28, 0xe0, 0xf6, 0xd3, 0x6c, - 0x20, 0x9f, 0x8d, 0xf1, 0xff, 0xd9, 0x80, 0x5a, 0xae, 0xc0, 0x14, 0xe2, 0x3f, 0xab, 0x1c, 0x15, - 0xe1, 0xbf, 0x2b, 0xf1, 0xef, 0xbf, 0xf3, 0xb2, 0xf0, 0xa3, 0xbf, 0x18, 0x00, 0xe3, 0x12, 0x84, - 0x8a, 0xee, 0x7c, 0xa7, 0xaa, 0xd5, 0xda, 0x39, 0x77, 0x61, 0x8a, 0x1c, 0xbd, 0x34, 0xe4, 0x7f, - 0x35, 0x60, 0x31, 0x5b, 0x29, 0xd0, 0x46, 0xe1, 0x33, 0xd5, 0xa9, 0xca, 0xb7, 0x76, 0x6d, 0x2a, - 0x19, 0x55, 0x8a, 0xcc, 0x4f, 0xe4, 0x52, 0xae, 0xa3, 0x0b, 0x67, 0x3d, 0xfa, 0xbb, 0x01, 0x4b, - 0x13, 0xbb, 0x1d, 0x7d, 0x50, 0xf8, 0x8e, 0x70, 0x56, 0x45, 0x2a, 0x4a, 0x98, 0x5f, 0x4b, 0xac, - 0x3d, 0xf3, 0xe8, 0x65, 0x25, 0xbc, 0xab, 0x61, 0x5c, 0x37, 0xde, 0xb9, 0xf5, 0x14, 0xde, 0x70, - 0x42, 0xff, 0xc5, 0xb8, 0x6f, 0xa5, 0x77, 0x25, 0x76, 0x24, 0x4e, 0xe8, 0x23, 0xe3, 0x57, 0x3b, - 0x9a, 0x7f, 0x10, 0x7a, 0x38, 0x18, 0xb4, 0xc2, 0x78, 0xd0, 0x1e, 0x90, 0x40, 0x9e, 0xdf, 0x6d, - 0x35, 0x85, 0x23, 0xca, 0x9e, 0xf3, 0x47, 0xcd, 0x8d, 0x94, 0xd0, 0x9f, 0x93, 0x12, 0xd7, 0xfe, - 0x17, 0x00, 0x00, 0xff, 0xff, 0x82, 0x2f, 0x0d, 0xbd, 0xd9, 0x19, 0x00, 0x00, +var fileDescriptor_clusters_92d23f68998e82a3 = []byte{ + // 2165 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x59, 0x4f, 0x73, 0x1c, 0x47, + 0x15, 0xf7, 0x68, 0xb5, 0xd2, 0xea, 0xed, 0x1f, 0xad, 0x3a, 0x8a, 0xd8, 0x28, 0x98, 0x38, 0x13, + 0x08, 0x4a, 0x08, 0xbb, 0x89, 0x4c, 0x2a, 0xc1, 0xc6, 0xa9, 0xc8, 0x92, 0x2c, 0x29, 0xc8, 0xb2, + 0x98, 0xdd, 0xb5, 0x31, 0x94, 0x6b, 0xaa, 0x77, 0xa6, 0x77, 0xdd, 0xd1, 0xfc, 0x63, 0xba, 0xc7, + 0x89, 0xec, 0x72, 0x15, 0x15, 0x4e, 0x14, 0x47, 0x3e, 0x00, 0x39, 0x73, 0xc8, 0x81, 0xa2, 0x8a, + 0xe2, 0xc0, 0x8d, 0x6f, 0x10, 0x8a, 0x4f, 0xc0, 0x81, 0x0b, 0x77, 0x8e, 0x54, 0xff, 0x99, 0xdd, + 0x99, 0x95, 0xec, 0x91, 0x84, 0x2b, 0x27, 0xed, 0xbc, 0x7e, 0xef, 0xf7, 0x7e, 0xfd, 0xfa, 0xf5, + 0xeb, 0xd7, 0x2d, 0x78, 0x67, 0x14, 0x86, 0x23, 0x8f, 0x74, 0x1c, 0x2f, 0x4c, 0xdc, 0x8e, 0x8b, + 0x39, 0x8e, 0xe2, 0xd0, 0xe9, 0x3c, 0x7a, 0x6f, 0x40, 0x38, 0x5e, 0xef, 0x38, 0x5e, 0xc2, 0x38, + 0x89, 0x59, 0x3b, 0x8a, 0x43, 0x1e, 0xa2, 0xcb, 0x4a, 0xbb, 0x2d, 0xb5, 0xdb, 0xa9, 0x76, 0x5b, + 0x6b, 0xaf, 0x7e, 0x5b, 0x83, 0xe1, 0x88, 0x76, 0x70, 0x10, 0x84, 0x1c, 0x73, 0x1a, 0x06, 0xda, + 0x78, 0xf5, 0xed, 0xe7, 0xbb, 0x62, 0x0f, 0x71, 0x4c, 0x5c, 0xad, 0xfb, 0x86, 0xd6, 0xf5, 0xc2, + 0x60, 0x14, 0x27, 0x41, 0x40, 0x83, 0x51, 0x27, 0x8c, 0x48, 0x9c, 0x03, 0xfc, 0x8e, 0x56, 0x92, + 0x5f, 0x83, 0x64, 0xd8, 0x71, 0x13, 0xa5, 0xa0, 0xc7, 0xaf, 0x4c, 0x8f, 0x0f, 0x29, 0xf1, 0x5c, + 0xdb, 0xc7, 0xec, 0x48, 0x6b, 0xbc, 0x36, 0xad, 0xc1, 0xa9, 0x4f, 0x18, 0xc7, 0x7e, 0xa4, 0x14, + 0xcc, 0x3f, 0xcc, 0xc2, 0xfc, 0xa6, 0x8a, 0x01, 0xba, 0x0c, 0x10, 0xc5, 0xe1, 0xa7, 0xc4, 0xe1, + 0x36, 0x75, 0x5b, 0xc6, 0x15, 0x63, 0x6d, 0xc1, 0x5a, 0xd0, 0x92, 0x3d, 0x17, 0xbd, 0x0e, 0x35, + 0x1d, 0x2d, 0x3b, 0xc0, 0x3e, 0x69, 0xcd, 0x48, 0x85, 0xaa, 0x96, 0x1d, 0x60, 0x9f, 0xa0, 0x2d, + 0x98, 0x73, 0xc2, 0x60, 0x48, 0x47, 0xad, 0xd2, 0x15, 0x63, 0xad, 0xba, 0xfe, 0x4e, 0xfb, 0xb9, + 0xf1, 0x6c, 0x6b, 0xcf, 0x9b, 0xd2, 0xc6, 0xd2, 0xb6, 0xe8, 0x13, 0x98, 0xf3, 0xf0, 0x80, 0x78, + 0xac, 0x55, 0xb9, 0x52, 0x5a, 0xab, 0xae, 0xaf, 0x9f, 0x0d, 0xa5, 0xbd, 0x2f, 0x8d, 0xb6, 0x03, + 0x1e, 0x1f, 0x5b, 0x1a, 0x41, 0x30, 0x62, 0x1c, 0xf3, 0x84, 0xb5, 0x66, 0xcf, 0xc3, 0xa8, 0x2b, + 0x6d, 0x2c, 0x6d, 0x8b, 0xba, 0xd0, 0x50, 0xbf, 0xec, 0x87, 0x94, 0xf1, 0x30, 0x3e, 0x6e, 0xcd, + 0x4b, 0x66, 0xe7, 0x43, 0xab, 0x2b, 0x8c, 0x5d, 0x05, 0x91, 0x8d, 0x67, 0x92, 0x50, 0xb7, 0x35, + 0x97, 0x8b, 0x67, 0x3f, 0xa1, 0x2e, 0xda, 0x81, 0x79, 0x9f, 0xf0, 0x98, 0x3a, 0xac, 0xb5, 0x20, + 0xe9, 0xff, 0xf0, 0x6c, 0x0e, 0x6f, 0x2b, 0x23, 0x2b, 0xb5, 0x5e, 0xfd, 0x31, 0x54, 0x33, 0xd1, + 0x41, 0x4d, 0x28, 0x1d, 0x91, 0x63, 0xbd, 0xc4, 0xe2, 0x27, 0x5a, 0x86, 0xf2, 0x23, 0xec, 0x25, + 0xe9, 0xaa, 0xaa, 0x8f, 0x6b, 0x33, 0x1f, 0x1a, 0xe6, 0x3f, 0xca, 0x50, 0xcf, 0xad, 0x13, 0x7a, + 0x03, 0xea, 0x6a, 0xa5, 0xec, 0x41, 0xe2, 0x1c, 0x11, 0xae, 0x71, 0x6a, 0x4a, 0x78, 0x53, 0xca, + 0xd0, 0x03, 0x40, 0x23, 0x87, 0xd8, 0xe9, 0x0c, 0x75, 0x5a, 0x54, 0xe4, 0x2c, 0x3a, 0x05, 0xb3, + 0xd8, 0x71, 0x48, 0x3e, 0x33, 0x9a, 0xa3, 0x29, 0x09, 0xba, 0x07, 0x75, 0x1f, 0x67, 0x91, 0x55, + 0x7c, 0x8a, 0x52, 0x65, 0x2f, 0x60, 0x1c, 0x07, 0x0e, 0xd9, 0x89, 0xc3, 0x24, 0xd2, 0xe0, 0x35, + 0x05, 0x34, 0x01, 0xfe, 0x2c, 0x8c, 0x8f, 0x26, 0xc0, 0x70, 0x71, 0x60, 0x05, 0xa4, 0x81, 0x3f, + 0x85, 0x6f, 0x31, 0xe2, 0x84, 0x81, 0x8b, 0xe3, 0x63, 0x3b, 0xef, 0xa2, 0x76, 0x61, 0x17, 0x2f, + 0x8f, 0x21, 0xef, 0x65, 0x7d, 0xdd, 0x85, 0x45, 0x16, 0x0e, 0xf9, 0x67, 0x38, 0x26, 0xa9, 0x8f, + 0xfa, 0x99, 0xf2, 0xa7, 0xab, 0xad, 0x34, 0x7c, 0x83, 0xe5, 0xbe, 0xd1, 0x7d, 0x68, 0x7a, 0x74, + 0x48, 0x9c, 0x63, 0xc7, 0x1b, 0x03, 0x37, 0x24, 0x70, 0xbb, 0x00, 0x78, 0x3f, 0x35, 0xd3, 0xc8, + 0x8b, 0x5e, 0x5e, 0x80, 0x02, 0x58, 0xa1, 0x01, 0xe5, 0x14, 0x7b, 0xf4, 0xb1, 0xac, 0x71, 0x36, + 0x76, 0x64, 0x2d, 0x6c, 0x55, 0xe5, 0x56, 0xfb, 0xa0, 0xc0, 0xc1, 0x41, 0xe8, 0x92, 0xbd, 0x1c, + 0xc0, 0x86, 0xb4, 0xb7, 0x5e, 0xa6, 0xa7, 0x48, 0x99, 0xf9, 0xc7, 0x12, 0x34, 0xa7, 0xf3, 0x0c, + 0xbd, 0x02, 0x95, 0xc7, 0x61, 0x40, 0xec, 0x24, 0xa6, 0x3a, 0xa9, 0xe7, 0xc5, 0x77, 0x3f, 0xa6, + 0xe8, 0x35, 0xa8, 0x06, 0x84, 0x8b, 0x75, 0x93, 0xa3, 0x6a, 0x9b, 0x80, 0x16, 0x09, 0x85, 0xef, + 0x41, 0x83, 0x25, 0x83, 0xac, 0x8e, 0xda, 0xd0, 0xf5, 0x89, 0x54, 0xa8, 0xad, 0x41, 0x93, 0x06, + 0x9c, 0xc4, 0x01, 0xf6, 0x6c, 0x1a, 0xd9, 0x61, 0xe0, 0x89, 0x62, 0x62, 0xac, 0x55, 0xac, 0x46, + 0x2a, 0xdf, 0x8b, 0xee, 0x04, 0xde, 0x31, 0xfa, 0x3e, 0x2c, 0x32, 0x12, 0x3f, 0xa2, 0x0e, 0xb1, + 0xb1, 0xe3, 0x84, 0x49, 0xc0, 0xe5, 0xf6, 0x59, 0xb0, 0x1a, 0x5a, 0xbc, 0xa1, 0xa4, 0xe8, 0x47, + 0xb0, 0x32, 0xa5, 0x68, 0x33, 0x27, 0x8c, 0x08, 0x6b, 0x95, 0xae, 0x94, 0xd6, 0x16, 0xac, 0xe5, + 0xbc, 0x7e, 0x57, 0x8e, 0x21, 0x04, 0xb3, 0x1c, 0x8f, 0x44, 0x5d, 0x14, 0x3a, 0xf2, 0x37, 0xba, + 0x0f, 0x15, 0x9f, 0x70, 0x2c, 0x82, 0xdb, 0x2a, 0xcb, 0xb0, 0xdf, 0x38, 0xe7, 0x56, 0x6d, 0xdf, + 0xd6, 0xf6, 0xaa, 0x0c, 0x8f, 0xe1, 0x56, 0xaf, 0x43, 0x3d, 0x37, 0x74, 0xae, 0x1a, 0xf4, 0xeb, + 0x59, 0x78, 0xe9, 0x94, 0xf4, 0x17, 0x95, 0x28, 0x48, 0x7c, 0x9b, 0xea, 0x21, 0x26, 0xd1, 0xca, + 0x56, 0x2d, 0x48, 0xfc, 0x54, 0x9d, 0x89, 0x85, 0x49, 0x15, 0xe4, 0xc1, 0xc5, 0x5a, 0x33, 0x72, + 0xca, 0xf5, 0x54, 0x2a, 0x8e, 0x2e, 0x86, 0x5e, 0x85, 0x05, 0xea, 0xe3, 0x91, 0x5a, 0xfc, 0x92, + 0x64, 0x50, 0x91, 0x02, 0xbd, 0x6a, 0x3e, 0x76, 0x1e, 0xd2, 0x80, 0xd8, 0xfc, 0x38, 0x52, 0x3a, + 0xb3, 0x6a, 0x31, 0xb4, 0xbc, 0x77, 0x1c, 0x49, 0xcd, 0x4f, 0xa0, 0xea, 0x52, 0x76, 0x94, 0xee, + 0x8e, 0xb2, 0xdc, 0x1d, 0x6f, 0x15, 0x44, 0x71, 0x8b, 0xb2, 0x23, 0xbd, 0x31, 0xc0, 0x1d, 0xff, + 0x96, 0xcc, 0x99, 0x1d, 0xc5, 0x84, 0xf8, 0x11, 0xa7, 0x03, 0x8f, 0xc8, 0x94, 0xaa, 0x58, 0x75, + 0xca, 0x0e, 0x27, 0x42, 0xe4, 0xc0, 0xb2, 0x8f, 0x03, 0x3c, 0x22, 0xae, 0x3d, 0x12, 0xc1, 0x49, + 0x7d, 0xcf, 0x4b, 0xdf, 0xef, 0x15, 0xf8, 0xbe, 0xad, 0x4c, 0xb3, 0x55, 0x05, 0xf9, 0x27, 0x64, + 0xa8, 0x07, 0x35, 0xec, 0x38, 0xc4, 0x13, 0x2d, 0x4a, 0x18, 0xa7, 0x47, 0xf3, 0xbb, 0x05, 0xe0, + 0x1b, 0x13, 0x93, 0xb4, 0x28, 0x66, 0x51, 0x64, 0x5c, 0x69, 0x60, 0x3b, 0x51, 0x62, 0x47, 0x1e, + 0xe6, 0xc3, 0x30, 0xf6, 0x65, 0x25, 0x17, 0x71, 0xa5, 0xc1, 0x66, 0x94, 0x1c, 0x6a, 0xa9, 0xf9, + 0x5b, 0x03, 0xd0, 0x49, 0xaa, 0x22, 0xf7, 0xc7, 0x8b, 0xcb, 0x89, 0x2f, 0x40, 0xd4, 0x2a, 0xeb, + 0xc4, 0x5a, 0x4e, 0x47, 0x7b, 0x7a, 0x50, 0xf6, 0x29, 0x37, 0xe0, 0xd5, 0xb1, 0x95, 0x0a, 0x99, + 0x9a, 0x71, 0xae, 0xb3, 0x69, 0xd1, 0x6c, 0xc6, 0x29, 0xdf, 0xb2, 0xcd, 0x31, 0x63, 0x58, 0x3a, + 0x31, 0x31, 0xf4, 0x2e, 0x2c, 0x67, 0xa6, 0x36, 0x49, 0x13, 0xc5, 0x03, 0x65, 0xc6, 0xd2, 0x54, + 0xf9, 0x01, 0x2c, 0x65, 0x2d, 0xd4, 0x16, 0x9f, 0x91, 0x19, 0xdc, 0xc4, 0x59, 0xfc, 0x24, 0xe0, + 0xe6, 0x6f, 0x0c, 0x80, 0x49, 0x9a, 0xa0, 0xef, 0x42, 0x63, 0x10, 0x86, 0xdc, 0x96, 0xb9, 0x26, + 0x7c, 0xe9, 0x94, 0xad, 0x09, 0xa9, 0xd0, 0x13, 0x4e, 0xd0, 0x5b, 0xb0, 0x34, 0xd1, 0x62, 0xf4, + 0x31, 0xb1, 0x47, 0x03, 0xbd, 0x47, 0x1a, 0xa9, 0x62, 0x97, 0x3e, 0x26, 0x3b, 0x03, 0x01, 0x28, + 0xb6, 0x92, 0x17, 0x3a, 0xd8, 0xb3, 0x19, 0x73, 0x99, 0x66, 0x22, 0xf6, 0xd2, 0xbe, 0x10, 0x76, + 0x99, 0xcb, 0xcc, 0x7f, 0x1b, 0xb0, 0x38, 0x55, 0xca, 0xd1, 0x06, 0x2c, 0x52, 0xd7, 0x23, 0xb6, + 0x4b, 0x3c, 0xc2, 0x89, 0xcd, 0xb9, 0x27, 0x5d, 0x54, 0xd7, 0x5f, 0x49, 0x93, 0x23, 0xed, 0x3e, + 0xdb, 0x5b, 0xba, 0x7f, 0xb5, 0xea, 0xc2, 0x62, 0x4b, 0x1a, 0xf4, 0xb8, 0x87, 0x6e, 0x41, 0x13, + 0x27, 0x3c, 0x1c, 0x43, 0x50, 0xbd, 0x08, 0xd5, 0xf5, 0xd5, 0x13, 0x18, 0xbd, 0xb4, 0x83, 0xdd, + 0xbd, 0x64, 0x35, 0x84, 0x95, 0x86, 0xa1, 0x3e, 0x41, 0x9b, 0xb0, 0x98, 0xc3, 0xe1, 0x9e, 0x6e, + 0x44, 0x9f, 0x4d, 0x65, 0xf7, 0x92, 0x55, 0xcf, 0xa0, 0x70, 0xef, 0x66, 0x19, 0x4a, 0x9c, 0x7b, + 0xe6, 0xef, 0x0c, 0x68, 0x3d, 0xeb, 0x50, 0x11, 0xb5, 0x99, 0x7c, 0x4e, 0x9c, 0x84, 0xe3, 0x81, + 0x47, 0xec, 0x21, 0xf5, 0xd2, 0x7c, 0x6b, 0x4c, 0xc4, 0xb7, 0xa8, 0x47, 0xd0, 0x2d, 0x58, 0x52, + 0x12, 0x71, 0xa2, 0x89, 0x79, 0x85, 0x09, 0xd7, 0x53, 0x7b, 0x4e, 0x78, 0x9a, 0x63, 0x9b, 0x9e, + 0x32, 0x31, 0xbf, 0x2c, 0x8d, 0xbb, 0x30, 0xd5, 0x4d, 0xa2, 0x5d, 0x28, 0x8b, 0x7e, 0x52, 0x39, + 0x6e, 0x9c, 0xb5, 0x49, 0x56, 0xc6, 0x6d, 0xf1, 0x87, 0x58, 0x0a, 0x00, 0xad, 0xc0, 0x9c, 0x4b, + 0x38, 0xa6, 0x9e, 0x4e, 0x7c, 0xfd, 0x85, 0xb6, 0xa0, 0x29, 0x15, 0x6c, 0xc6, 0x71, 0xcc, 0xd5, + 0xaa, 0x94, 0x8a, 0x56, 0xc5, 0x92, 0x9d, 0x32, 0xe9, 0x0a, 0x13, 0xb9, 0x26, 0x3f, 0x83, 0x0a, + 0x4b, 0x06, 0x8a, 0xea, 0xac, 0xa4, 0xfa, 0xfe, 0xf9, 0xa8, 0x6a, 0x63, 0x6b, 0x0c, 0x63, 0xde, + 0x85, 0xb2, 0x9c, 0x00, 0xaa, 0xc2, 0x7c, 0xff, 0xe0, 0xa7, 0x07, 0x77, 0xee, 0x1d, 0x34, 0x2f, + 0xa1, 0x1a, 0x54, 0x36, 0xad, 0xed, 0x8d, 0xde, 0xde, 0xc1, 0x4e, 0xd3, 0x10, 0x43, 0x56, 0xff, + 0xe0, 0x40, 0x7c, 0xcc, 0xa0, 0x05, 0x28, 0x6f, 0x5b, 0xd6, 0x1d, 0xab, 0x59, 0x12, 0x5a, 0x5b, + 0xdb, 0xfb, 0xdb, 0x52, 0x6b, 0x56, 0x7c, 0xf5, 0x0f, 0xb7, 0x94, 0x4d, 0xd9, 0xfc, 0x09, 0x54, + 0x52, 0x6f, 0x68, 0x11, 0xaa, 0xfd, 0x83, 0xee, 0xe1, 0xf6, 0xe6, 0xde, 0xad, 0xbd, 0xed, 0xad, + 0xe6, 0x25, 0x54, 0x87, 0x85, 0xfe, 0xc1, 0xee, 0xf6, 0xc6, 0x7e, 0x6f, 0xf7, 0x7e, 0xd3, 0x40, + 0x4d, 0xa8, 0x75, 0x7b, 0x1b, 0xfb, 0xdb, 0x76, 0xb7, 0xb7, 0xd1, 0xeb, 0x77, 0x9b, 0x33, 0xe6, + 0xd7, 0x06, 0x34, 0xf2, 0xfd, 0x93, 0x38, 0x9f, 0xd4, 0x99, 0xf2, 0x88, 0xc4, 0x8c, 0x86, 0x41, + 0xda, 0x29, 0x4b, 0xe1, 0x5d, 0x25, 0x43, 0x0f, 0xe4, 0xb5, 0x2b, 0x22, 0x31, 0xa7, 0xfa, 0x6c, + 0x2a, 0x3e, 0x76, 0xf3, 0x7e, 0xda, 0x87, 0x63, 0x7b, 0x75, 0xec, 0x66, 0x00, 0x57, 0x6f, 0xc0, + 0xe2, 0xd4, 0xf0, 0xb9, 0x8e, 0xde, 0x7f, 0xce, 0x40, 0x23, 0x7f, 0xab, 0x40, 0x18, 0x6a, 0x0f, + 0xdd, 0x21, 0xb3, 0xd3, 0xab, 0x89, 0x21, 0x29, 0x7f, 0x74, 0xae, 0xab, 0x49, 0x7b, 0xd7, 0x1d, + 0x32, 0xfd, 0x5b, 0x71, 0xae, 0x3e, 0x9c, 0x48, 0x84, 0x8b, 0x63, 0x1c, 0x07, 0x63, 0x17, 0x33, + 0x17, 0x71, 0x71, 0x1f, 0xc7, 0x41, 0xde, 0xc5, 0xf1, 0x44, 0xb2, 0xfa, 0x11, 0x34, 0xa7, 0x39, + 0x14, 0x05, 0xa6, 0x94, 0x09, 0x8c, 0xb0, 0x9f, 0x76, 0x70, 0x1e, 0x7b, 0xf3, 0x2b, 0x03, 0x96, + 0x37, 0x63, 0x82, 0x79, 0xda, 0x40, 0x59, 0xe4, 0x57, 0x09, 0x61, 0xbc, 0xe8, 0x1a, 0xbe, 0x02, + 0x73, 0x31, 0x19, 0x89, 0x64, 0x52, 0x15, 0x5f, 0x7f, 0xa1, 0x8f, 0x61, 0x5e, 0x5f, 0xb6, 0x74, + 0x7d, 0x79, 0xf3, 0x6c, 0xd1, 0xb2, 0x52, 0x33, 0xe1, 0x38, 0x56, 0x1c, 0x84, 0x63, 0xd5, 0xde, + 0x2c, 0x68, 0xc9, 0x9e, 0x6b, 0xfe, 0x67, 0x06, 0x96, 0xfb, 0x91, 0xfb, 0x7f, 0x10, 0x2e, 0xe7, + 0x08, 0x9f, 0xe1, 0x3d, 0x21, 0x33, 0xa7, 0xd2, 0xc5, 0xe6, 0xf4, 0x00, 0x2e, 0x8f, 0x62, 0xec, + 0x90, 0x61, 0xe2, 0xd9, 0x2e, 0x71, 0x42, 0xdf, 0xa7, 0x8c, 0x65, 0x6b, 0xf1, 0x5c, 0x51, 0x2d, + 0x7e, 0x35, 0xb5, 0xdf, 0xca, 0x98, 0xeb, 0xb2, 0x8c, 0xae, 0x43, 0x35, 0x91, 0x21, 0x91, 0x8f, + 0x2e, 0xfa, 0x8d, 0xe1, 0x64, 0x75, 0xbc, 0x45, 0x89, 0xe7, 0xde, 0xc6, 0xec, 0xc8, 0x02, 0xa5, + 0x2e, 0x7e, 0x4f, 0xc5, 0x7b, 0x7e, 0x3a, 0xde, 0x7f, 0x32, 0x60, 0x59, 0x9d, 0x4a, 0x2f, 0x26, + 0x41, 0xce, 0x10, 0xef, 0xe9, 0x27, 0x89, 0xd9, 0x93, 0x4f, 0x12, 0x79, 0xd2, 0xe5, 0x69, 0xd2, + 0x3e, 0x2c, 0xed, 0x10, 0xfe, 0x4d, 0x11, 0x36, 0xbf, 0x34, 0xe0, 0xa5, 0x7d, 0xca, 0x52, 0x87, + 0xec, 0xdc, 0x1e, 0x67, 0x73, 0x1e, 0x57, 0x60, 0x6e, 0x48, 0x3d, 0x91, 0x6e, 0x3a, 0x55, 0xd5, + 0x97, 0xb8, 0x1b, 0x44, 0xa2, 0x8c, 0x8b, 0x16, 0x4a, 0xf7, 0x45, 0x15, 0x21, 0x10, 0xbd, 0x93, + 0xf4, 0x25, 0x06, 0x79, 0x78, 0x44, 0xd2, 0x29, 0x48, 0xf5, 0x9e, 0x10, 0x98, 0x5f, 0x18, 0xb0, + 0x9c, 0xa7, 0xc8, 0xa2, 0x30, 0x60, 0x04, 0xdd, 0x84, 0x4a, 0xfa, 0xfa, 0xa8, 0x4b, 0xe8, 0x59, + 0xb3, 0x7b, 0x6c, 0x87, 0xde, 0x84, 0xc5, 0x80, 0x7c, 0xce, 0xed, 0x0c, 0x01, 0x15, 0xa5, 0xba, + 0x10, 0x1f, 0x8e, 0x49, 0xc4, 0xb0, 0xb2, 0x45, 0xf1, 0x28, 0x08, 0xd9, 0x37, 0x96, 0x4c, 0xe6, + 0x07, 0xa7, 0xf8, 0x64, 0x89, 0xc7, 0x99, 0xf0, 0x19, 0x26, 0x3c, 0x4a, 0x78, 0xa6, 0x41, 0x5e, + 0x50, 0x92, 0x7e, 0x4c, 0xd7, 0xff, 0x5b, 0x81, 0xa5, 0xc9, 0xa5, 0x92, 0xc7, 0xa1, 0xe7, 0x91, + 0x18, 0x7d, 0x65, 0x40, 0x3d, 0x57, 0x2f, 0xd1, 0xd5, 0xa2, 0x70, 0x9d, 0x52, 0x5d, 0x57, 0x2f, + 0xa7, 0x46, 0x99, 0x97, 0xd7, 0xf6, 0x9d, 0xf4, 0xe5, 0xd5, 0xdc, 0xfb, 0xe2, 0xeb, 0x7f, 0xfd, + 0x7e, 0x66, 0xd3, 0xfc, 0x70, 0xfc, 0x6c, 0xab, 0x63, 0xc1, 0x3a, 0x4f, 0x26, 0x71, 0x7a, 0xda, + 0x51, 0x61, 0x60, 0x9d, 0x27, 0xea, 0xc7, 0xd3, 0xf1, 0x43, 0xf2, 0xb5, 0x71, 0xe9, 0xf9, 0x9b, + 0x01, 0xf5, 0x5c, 0xbd, 0x2c, 0x24, 0x7c, 0x5a, 0x75, 0x2d, 0x22, 0xfc, 0x73, 0x49, 0xd8, 0x5a, + 0xdf, 0xb9, 0x28, 0xe1, 0xce, 0x93, 0xec, 0x42, 0x3e, 0x9d, 0xf0, 0xff, 0xb3, 0x01, 0xf5, 0x5c, + 0xfd, 0x29, 0xe4, 0x7f, 0x5a, 0xb5, 0x2a, 0xe2, 0x7f, 0x47, 0xf2, 0xdf, 0x7b, 0xfb, 0x45, 0xf1, + 0x47, 0x7f, 0x31, 0x00, 0x26, 0x25, 0x08, 0x15, 0xdd, 0x50, 0x4f, 0x54, 0xab, 0xd5, 0x33, 0xee, + 0xc2, 0x94, 0x39, 0x7a, 0x61, 0xcc, 0xff, 0x6a, 0x40, 0x2d, 0x5b, 0x29, 0xd0, 0x7a, 0xe1, 0xa3, + 0xda, 0x89, 0xca, 0xb7, 0x7a, 0xf5, 0x5c, 0x36, 0xaa, 0x14, 0x99, 0x1f, 0xcb, 0xa9, 0x5c, 0x43, + 0x17, 0xce, 0x7a, 0xf4, 0x77, 0x03, 0x16, 0xa7, 0x76, 0x3b, 0x7a, 0xbf, 0xf0, 0xd5, 0xe3, 0xb4, + 0x8a, 0x54, 0x94, 0x30, 0xbf, 0x94, 0x5c, 0xfb, 0xe6, 0xe1, 0x8b, 0x4a, 0x78, 0x57, 0xd3, 0xb8, + 0x66, 0xbc, 0x7d, 0xf3, 0x09, 0xbc, 0xee, 0x84, 0xfe, 0xf3, 0x79, 0xdf, 0x4c, 0x2f, 0x62, 0xec, + 0x50, 0x9c, 0xef, 0x87, 0xc6, 0x2f, 0xb6, 0xb5, 0xfe, 0x28, 0xf4, 0x70, 0x30, 0x6a, 0x87, 0xf1, + 0xa8, 0x33, 0x22, 0x81, 0x3c, 0xfd, 0x3b, 0x6a, 0x08, 0x47, 0x94, 0x3d, 0xe3, 0xff, 0x42, 0xd7, + 0x53, 0xc1, 0x60, 0x4e, 0x5a, 0x5c, 0xfd, 0x5f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x4a, 0xea, 0xcc, + 0x54, 0xb3, 0x1a, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/jobs.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/jobs.pb.go index bf12bc89f..76fa3d42c 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/jobs.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/jobs.pb.go @@ -80,7 +80,7 @@ func (x LoggingConfig_Level) String() string { return proto.EnumName(LoggingConfig_Level_name, int32(x)) } func (LoggingConfig_Level) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{0, 0} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{0, 0} } // The job state. @@ -143,12 +143,14 @@ func (x JobStatus_State) String() string { return proto.EnumName(JobStatus_State_name, int32(x)) } func (JobStatus_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{9, 0} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{9, 0} } +// The job substate. type JobStatus_Substate int32 const ( + // The job substate is unknown. JobStatus_UNSPECIFIED JobStatus_Substate = 0 // The Job is submitted to the agent. // @@ -185,7 +187,7 @@ func (x JobStatus_Substate) String() string { return proto.EnumName(JobStatus_Substate_name, int32(x)) } func (JobStatus_Substate) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{9, 1} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{9, 1} } // The application state, corresponding to @@ -240,7 +242,7 @@ func (x YarnApplication_State) String() string { return proto.EnumName(YarnApplication_State_name, int32(x)) } func (YarnApplication_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{11, 0} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{11, 0} } // A matcher that specifies categories of job states. @@ -271,7 +273,7 @@ func (x ListJobsRequest_JobStateMatcher) String() string { return proto.EnumName(ListJobsRequest_JobStateMatcher_name, int32(x)) } func (ListJobsRequest_JobStateMatcher) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{16, 0} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{16, 0} } // The runtime logging config of the job. @@ -280,7 +282,7 @@ type LoggingConfig struct { // "root" package name to configure rootLogger. // Examples: // 'com.google = FATAL', 'root = INFO', 'org.apache = DEBUG' - DriverLogLevels map[string]LoggingConfig_Level `protobuf:"bytes,2,rep,name=driver_log_levels,json=driverLogLevels" json:"driver_log_levels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value,enum=google.cloud.dataproc.v1beta2.LoggingConfig_Level"` + DriverLogLevels map[string]LoggingConfig_Level `protobuf:"bytes,2,rep,name=driver_log_levels,json=driverLogLevels,proto3" json:"driver_log_levels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3,enum=google.cloud.dataproc.v1beta2.LoggingConfig_Level"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -290,7 +292,7 @@ func (m *LoggingConfig) Reset() { *m = LoggingConfig{} } func (m *LoggingConfig) String() string { return proto.CompactTextString(m) } func (*LoggingConfig) ProtoMessage() {} func (*LoggingConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{0} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{0} } func (m *LoggingConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LoggingConfig.Unmarshal(m, b) @@ -334,25 +336,25 @@ type HadoopJob struct { // include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as job // properties, since a collision may occur that causes an incorrect job // submission. - Args []string `protobuf:"bytes,3,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` // Optional. Jar file URIs to add to the CLASSPATHs of the // Hadoop driver and tasks. - JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. HCFS (Hadoop Compatible Filesystem) URIs of files to be copied // to the working directory of Hadoop drivers and distributed tasks. Useful // for naively parallel tasks. - FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris" json:"file_uris,omitempty"` + FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"` // Optional. HCFS URIs of archives to be extracted in the working directory of // Hadoop drivers and tasks. Supported file types: // .jar, .tar, .tar.gz, .tgz, or .zip. - ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris" json:"archive_uris,omitempty"` + ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"` // Optional. A mapping of property names to values, used to configure Hadoop. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site and // classes in user code. - Properties map[string]string `protobuf:"bytes,7,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -362,7 +364,7 @@ func (m *HadoopJob) Reset() { *m = HadoopJob{} } func (m *HadoopJob) String() string { return proto.CompactTextString(m) } func (*HadoopJob) ProtoMessage() {} func (*HadoopJob) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{1} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{1} } func (m *HadoopJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_HadoopJob.Unmarshal(m, b) @@ -387,10 +389,10 @@ type isHadoopJob_Driver interface { } type HadoopJob_MainJarFileUri struct { - MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,oneof"` + MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,proto3,oneof"` } type HadoopJob_MainClass struct { - MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,oneof"` + MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,proto3,oneof"` } func (*HadoopJob_MainJarFileUri) isHadoopJob_Driver() {} @@ -540,24 +542,24 @@ type SparkJob struct { // Optional. The arguments to pass to the driver. Do not include arguments, // such as `--conf`, that can be set as job properties, since a collision may // occur that causes an incorrect job submission. - Args []string `protobuf:"bytes,3,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` // Optional. HCFS URIs of jar files to add to the CLASSPATHs of the // Spark driver and tasks. - JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. HCFS URIs of files to be copied to the working directory of // Spark drivers and distributed tasks. Useful for naively parallel tasks. - FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris" json:"file_uris,omitempty"` + FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"` // Optional. HCFS URIs of archives to be extracted in the working directory // of Spark drivers and tasks. Supported file types: // .jar, .tar, .tar.gz, .tgz, and .zip. - ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris" json:"archive_uris,omitempty"` + ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"` // Optional. A mapping of property names to values, used to configure Spark. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in // /etc/spark/conf/spark-defaults.conf and classes in user code. - Properties map[string]string `protobuf:"bytes,7,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -567,7 +569,7 @@ func (m *SparkJob) Reset() { *m = SparkJob{} } func (m *SparkJob) String() string { return proto.CompactTextString(m) } func (*SparkJob) ProtoMessage() {} func (*SparkJob) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{2} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{2} } func (m *SparkJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SparkJob.Unmarshal(m, b) @@ -592,10 +594,10 @@ type isSparkJob_Driver interface { } type SparkJob_MainJarFileUri struct { - MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,oneof"` + MainJarFileUri string `protobuf:"bytes,1,opt,name=main_jar_file_uri,json=mainJarFileUri,proto3,oneof"` } type SparkJob_MainClass struct { - MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,oneof"` + MainClass string `protobuf:"bytes,2,opt,name=main_class,json=mainClass,proto3,oneof"` } func (*SparkJob_MainJarFileUri) isSparkJob_Driver() {} @@ -736,30 +738,30 @@ func _SparkJob_OneofSizer(msg proto.Message) (n int) { type PySparkJob struct { // Required. The HCFS URI of the main Python file to use as the driver. Must // be a .py file. - MainPythonFileUri string `protobuf:"bytes,1,opt,name=main_python_file_uri,json=mainPythonFileUri" json:"main_python_file_uri,omitempty"` + MainPythonFileUri string `protobuf:"bytes,1,opt,name=main_python_file_uri,json=mainPythonFileUri,proto3" json:"main_python_file_uri,omitempty"` // Optional. The arguments to pass to the driver. Do not include arguments, // such as `--conf`, that can be set as job properties, since a collision may // occur that causes an incorrect job submission. - Args []string `protobuf:"bytes,2,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,2,rep,name=args,proto3" json:"args,omitempty"` // Optional. HCFS file URIs of Python files to pass to the PySpark // framework. Supported file types: .py, .egg, and .zip. - PythonFileUris []string `protobuf:"bytes,3,rep,name=python_file_uris,json=pythonFileUris" json:"python_file_uris,omitempty"` + PythonFileUris []string `protobuf:"bytes,3,rep,name=python_file_uris,json=pythonFileUris,proto3" json:"python_file_uris,omitempty"` // Optional. HCFS URIs of jar files to add to the CLASSPATHs of the // Python driver and tasks. - JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,4,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. HCFS URIs of files to be copied to the working directory of // Python drivers and distributed tasks. Useful for naively parallel tasks. - FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris" json:"file_uris,omitempty"` + FileUris []string `protobuf:"bytes,5,rep,name=file_uris,json=fileUris,proto3" json:"file_uris,omitempty"` // Optional. HCFS URIs of archives to be extracted in the working directory of // .jar, .tar, .tar.gz, .tgz, and .zip. - ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris" json:"archive_uris,omitempty"` + ArchiveUris []string `protobuf:"bytes,6,rep,name=archive_uris,json=archiveUris,proto3" json:"archive_uris,omitempty"` // Optional. A mapping of property names to values, used to configure PySpark. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in // /etc/spark/conf/spark-defaults.conf and classes in user code. - Properties map[string]string `protobuf:"bytes,7,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,8,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -769,7 +771,7 @@ func (m *PySparkJob) Reset() { *m = PySparkJob{} } func (m *PySparkJob) String() string { return proto.CompactTextString(m) } func (*PySparkJob) ProtoMessage() {} func (*PySparkJob) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{3} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{3} } func (m *PySparkJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PySparkJob.Unmarshal(m, b) @@ -861,7 +863,7 @@ type QueryList struct { // ] // } // } - Queries []string `protobuf:"bytes,1,rep,name=queries" json:"queries,omitempty"` + Queries []string `protobuf:"bytes,1,rep,name=queries,proto3" json:"queries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -871,7 +873,7 @@ func (m *QueryList) Reset() { *m = QueryList{} } func (m *QueryList) String() string { return proto.CompactTextString(m) } func (*QueryList) ProtoMessage() {} func (*QueryList) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{4} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{4} } func (m *QueryList) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_QueryList.Unmarshal(m, b) @@ -911,19 +913,19 @@ type HiveJob struct { // Optional. Whether to continue executing queries if a query fails. // The default value is `false`. Setting to `true` can be useful when executing // independent parallel queries. - ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure" json:"continue_on_failure,omitempty"` + ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"` // Optional. Mapping of query variable names to values (equivalent to the // Hive command: `SET name="value";`). - ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables,proto3" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. A mapping of property names and values, used to configure Hive. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, // /etc/hive/conf/hive-site.xml, and classes in user code. - Properties map[string]string `protobuf:"bytes,5,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,5,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. HCFS URIs of jar files to add to the CLASSPATH of the // Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes // and UDFs. - JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -933,7 +935,7 @@ func (m *HiveJob) Reset() { *m = HiveJob{} } func (m *HiveJob) String() string { return proto.CompactTextString(m) } func (*HiveJob) ProtoMessage() {} func (*HiveJob) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{5} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{5} } func (m *HiveJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_HiveJob.Unmarshal(m, b) @@ -958,10 +960,10 @@ type isHiveJob_Queries interface { } type HiveJob_QueryFileUri struct { - QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,oneof"` + QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"` } type HiveJob_QueryList struct { - QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,oneof"` + QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"` } func (*HiveJob_QueryFileUri) isHiveJob_Queries() {} @@ -1098,15 +1100,15 @@ type SparkSqlJob struct { Queries isSparkSqlJob_Queries `protobuf_oneof:"queries"` // Optional. Mapping of query variable names to values (equivalent to the // Spark SQL command: SET `name="value";`). - ScriptVariables map[string]string `protobuf:"bytes,3,rep,name=script_variables,json=scriptVariables" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ScriptVariables map[string]string `protobuf:"bytes,3,rep,name=script_variables,json=scriptVariables,proto3" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. A mapping of property names to values, used to configure // Spark SQL's SparkConf. Properties that conflict with values set by the // Cloud Dataproc API may be overwritten. - Properties map[string]string `protobuf:"bytes,4,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,4,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. - JarFileUris []string `protobuf:"bytes,56,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,56,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,6,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,6,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1116,7 +1118,7 @@ func (m *SparkSqlJob) Reset() { *m = SparkSqlJob{} } func (m *SparkSqlJob) String() string { return proto.CompactTextString(m) } func (*SparkSqlJob) ProtoMessage() {} func (*SparkSqlJob) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{6} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{6} } func (m *SparkSqlJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SparkSqlJob.Unmarshal(m, b) @@ -1141,10 +1143,10 @@ type isSparkSqlJob_Queries interface { } type SparkSqlJob_QueryFileUri struct { - QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,oneof"` + QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"` } type SparkSqlJob_QueryList struct { - QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,oneof"` + QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"` } func (*SparkSqlJob_QueryFileUri) isSparkSqlJob_Queries() {} @@ -1282,20 +1284,20 @@ type PigJob struct { // Optional. Whether to continue executing queries if a query fails. // The default value is `false`. Setting to `true` can be useful when executing // independent parallel queries. - ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure" json:"continue_on_failure,omitempty"` + ContinueOnFailure bool `protobuf:"varint,3,opt,name=continue_on_failure,json=continueOnFailure,proto3" json:"continue_on_failure,omitempty"` // Optional. Mapping of query variable names to values (equivalent to the Pig // command: `name=[value]`). - ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ScriptVariables map[string]string `protobuf:"bytes,4,rep,name=script_variables,json=scriptVariables,proto3" json:"script_variables,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. A mapping of property names to values, used to configure Pig. // Properties that conflict with values set by the Cloud Dataproc API may be // overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml, // /etc/pig/conf/pig.properties, and classes in user code. - Properties map[string]string `protobuf:"bytes,5,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]string `protobuf:"bytes,5,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. HCFS URIs of jar files to add to the CLASSPATH of // the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs. - JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris" json:"jar_file_uris,omitempty"` + JarFileUris []string `protobuf:"bytes,6,rep,name=jar_file_uris,json=jarFileUris,proto3" json:"jar_file_uris,omitempty"` // Optional. The runtime log config for job execution. - LoggingConfig *LoggingConfig `protobuf:"bytes,7,opt,name=logging_config,json=loggingConfig" json:"logging_config,omitempty"` + LoggingConfig *LoggingConfig `protobuf:"bytes,7,opt,name=logging_config,json=loggingConfig,proto3" json:"logging_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1305,7 +1307,7 @@ func (m *PigJob) Reset() { *m = PigJob{} } func (m *PigJob) String() string { return proto.CompactTextString(m) } func (*PigJob) ProtoMessage() {} func (*PigJob) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{7} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{7} } func (m *PigJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PigJob.Unmarshal(m, b) @@ -1330,10 +1332,10 @@ type isPigJob_Queries interface { } type PigJob_QueryFileUri struct { - QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,oneof"` + QueryFileUri string `protobuf:"bytes,1,opt,name=query_file_uri,json=queryFileUri,proto3,oneof"` } type PigJob_QueryList struct { - QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,oneof"` + QueryList *QueryList `protobuf:"bytes,2,opt,name=query_list,json=queryList,proto3,oneof"` } func (*PigJob_QueryFileUri) isPigJob_Queries() {} @@ -1468,10 +1470,10 @@ func _PigJob_OneofSizer(msg proto.Message) (n int) { // Cloud Dataproc job config. type JobPlacement struct { // Required. The name of the cluster where the job will be submitted. - ClusterName string `protobuf:"bytes,1,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` - // Output-only. A cluster UUID generated by the Cloud Dataproc service when + ClusterName string `protobuf:"bytes,1,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` + // Output only. A cluster UUID generated by the Cloud Dataproc service when // the job is submitted. - ClusterUuid string `protobuf:"bytes,2,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` + ClusterUuid string `protobuf:"bytes,2,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1481,7 +1483,7 @@ func (m *JobPlacement) Reset() { *m = JobPlacement{} } func (m *JobPlacement) String() string { return proto.CompactTextString(m) } func (*JobPlacement) ProtoMessage() {} func (*JobPlacement) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{8} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{8} } func (m *JobPlacement) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_JobPlacement.Unmarshal(m, b) @@ -1517,16 +1519,16 @@ func (m *JobPlacement) GetClusterUuid() string { // Cloud Dataproc job status. type JobStatus struct { - // Output-only. A state message specifying the overall job state. - State JobStatus_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.dataproc.v1beta2.JobStatus_State" json:"state,omitempty"` - // Output-only. Optional job state details, such as an error + // Output only. A state message specifying the overall job state. + State JobStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1beta2.JobStatus_State" json:"state,omitempty"` + // Output only. Optional job state details, such as an error // description if the state is ERROR. - Details string `protobuf:"bytes,2,opt,name=details" json:"details,omitempty"` - // Output-only. The time when this state was entered. - StateStartTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=state_start_time,json=stateStartTime" json:"state_start_time,omitempty"` - // Output-only. Additional state information, which includes + Details string `protobuf:"bytes,2,opt,name=details,proto3" json:"details,omitempty"` + // Output only. The time when this state was entered. + StateStartTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"` + // Output only. Additional state information, which includes // status reported by the agent. - Substate JobStatus_Substate `protobuf:"varint,7,opt,name=substate,enum=google.cloud.dataproc.v1beta2.JobStatus_Substate" json:"substate,omitempty"` + Substate JobStatus_Substate `protobuf:"varint,7,opt,name=substate,proto3,enum=google.cloud.dataproc.v1beta2.JobStatus_Substate" json:"substate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1536,7 +1538,7 @@ func (m *JobStatus) Reset() { *m = JobStatus{} } func (m *JobStatus) String() string { return proto.CompactTextString(m) } func (*JobStatus) ProtoMessage() {} func (*JobStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{9} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{9} } func (m *JobStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_JobStatus.Unmarshal(m, b) @@ -1588,13 +1590,13 @@ func (m *JobStatus) GetSubstate() JobStatus_Substate { type JobReference struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Optional. The job ID, which must be unique within the project. The job ID // is generated by the server upon job submission or provided by the user as a // means to perform retries without creating duplicate jobs. The ID must // contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or // hyphens (-). The maximum length is 100 characters. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1604,7 +1606,7 @@ func (m *JobReference) Reset() { *m = JobReference{} } func (m *JobReference) String() string { return proto.CompactTextString(m) } func (*JobReference) ProtoMessage() {} func (*JobReference) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{10} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{10} } func (m *JobReference) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_JobReference.Unmarshal(m, b) @@ -1645,16 +1647,16 @@ func (m *JobReference) GetJobId() string { // be changed before final release. type YarnApplication struct { // Required. The application name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The application state. - State YarnApplication_State `protobuf:"varint,2,opt,name=state,enum=google.cloud.dataproc.v1beta2.YarnApplication_State" json:"state,omitempty"` + State YarnApplication_State `protobuf:"varint,2,opt,name=state,proto3,enum=google.cloud.dataproc.v1beta2.YarnApplication_State" json:"state,omitempty"` // Required. The numerical progress of the application, from 1 to 100. - Progress float32 `protobuf:"fixed32,3,opt,name=progress" json:"progress,omitempty"` + Progress float32 `protobuf:"fixed32,3,opt,name=progress,proto3" json:"progress,omitempty"` // Optional. The HTTP URL of the ApplicationMaster, HistoryServer, or // TimelineServer that provides application-specific information. The URL uses // the internal hostname, and requires a proxy server for resolution and, // possibly, access. - TrackingUrl string `protobuf:"bytes,4,opt,name=tracking_url,json=trackingUrl" json:"tracking_url,omitempty"` + TrackingUrl string `protobuf:"bytes,4,opt,name=tracking_url,json=trackingUrl,proto3" json:"tracking_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1664,7 +1666,7 @@ func (m *YarnApplication) Reset() { *m = YarnApplication{} } func (m *YarnApplication) String() string { return proto.CompactTextString(m) } func (*YarnApplication) ProtoMessage() {} func (*YarnApplication) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{11} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{11} } func (m *YarnApplication) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_YarnApplication.Unmarshal(m, b) @@ -1718,10 +1720,10 @@ type Job struct { // obtain the equivalent REST path of the job resource. If this property // is not specified when a job is created, the server generates a // job_id. - Reference *JobReference `protobuf:"bytes,1,opt,name=reference" json:"reference,omitempty"` + Reference *JobReference `protobuf:"bytes,1,opt,name=reference,proto3" json:"reference,omitempty"` // Required. Job information, including how, when, and where to // run the job. - Placement *JobPlacement `protobuf:"bytes,2,opt,name=placement" json:"placement,omitempty"` + Placement *JobPlacement `protobuf:"bytes,2,opt,name=placement,proto3" json:"placement,omitempty"` // Required. The application/framework-specific portion of the job. // // Types that are valid to be assigned to TypeJob: @@ -1732,33 +1734,33 @@ type Job struct { // *Job_PigJob // *Job_SparkSqlJob TypeJob isJob_TypeJob `protobuf_oneof:"type_job"` - // Output-only. The job status. Additional application-specific + // Output only. The job status. Additional application-specific // status information may be contained in the type_job // and yarn_applications fields. - Status *JobStatus `protobuf:"bytes,8,opt,name=status" json:"status,omitempty"` - // Output-only. The previous job status. - StatusHistory []*JobStatus `protobuf:"bytes,13,rep,name=status_history,json=statusHistory" json:"status_history,omitempty"` - // Output-only. The collection of YARN applications spun up by this job. + Status *JobStatus `protobuf:"bytes,8,opt,name=status,proto3" json:"status,omitempty"` + // Output only. The previous job status. + StatusHistory []*JobStatus `protobuf:"bytes,13,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"` + // Output only. The collection of YARN applications spun up by this job. // // **Beta** Feature: This report is available for testing purposes only. It may // be changed before final release. - YarnApplications []*YarnApplication `protobuf:"bytes,9,rep,name=yarn_applications,json=yarnApplications" json:"yarn_applications,omitempty"` - // Output-only. A URI pointing to the location of the stdout of the job's + YarnApplications []*YarnApplication `protobuf:"bytes,9,rep,name=yarn_applications,json=yarnApplications,proto3" json:"yarn_applications,omitempty"` + // Output only. A URI pointing to the location of the stdout of the job's // driver program. - DriverOutputResourceUri string `protobuf:"bytes,17,opt,name=driver_output_resource_uri,json=driverOutputResourceUri" json:"driver_output_resource_uri,omitempty"` - // Output-only. If present, the location of miscellaneous control files + DriverOutputResourceUri string `protobuf:"bytes,17,opt,name=driver_output_resource_uri,json=driverOutputResourceUri,proto3" json:"driver_output_resource_uri,omitempty"` + // Output only. If present, the location of miscellaneous control files // which may be used as part of job setup and handling. If not present, // control files may be placed in the same location as `driver_output_uri`. - DriverControlFilesUri string `protobuf:"bytes,15,opt,name=driver_control_files_uri,json=driverControlFilesUri" json:"driver_control_files_uri,omitempty"` + DriverControlFilesUri string `protobuf:"bytes,15,opt,name=driver_control_files_uri,json=driverControlFilesUri,proto3" json:"driver_control_files_uri,omitempty"` // Optional. The labels to associate with this job. // Label **keys** must contain 1 to 63 characters, and must conform to // [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // Label **values** may be empty, but, if present, must contain 1 to 63 // characters, and must conform to [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // No more than 32 labels can be associated with a job. - Labels map[string]string `protobuf:"bytes,18,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,18,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. Job scheduling configuration. - Scheduling *JobScheduling `protobuf:"bytes,20,opt,name=scheduling" json:"scheduling,omitempty"` + Scheduling *JobScheduling `protobuf:"bytes,20,opt,name=scheduling,proto3" json:"scheduling,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1768,7 +1770,7 @@ func (m *Job) Reset() { *m = Job{} } func (m *Job) String() string { return proto.CompactTextString(m) } func (*Job) ProtoMessage() {} func (*Job) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{12} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{12} } func (m *Job) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Job.Unmarshal(m, b) @@ -1793,22 +1795,22 @@ type isJob_TypeJob interface { } type Job_HadoopJob struct { - HadoopJob *HadoopJob `protobuf:"bytes,3,opt,name=hadoop_job,json=hadoopJob,oneof"` + HadoopJob *HadoopJob `protobuf:"bytes,3,opt,name=hadoop_job,json=hadoopJob,proto3,oneof"` } type Job_SparkJob struct { - SparkJob *SparkJob `protobuf:"bytes,4,opt,name=spark_job,json=sparkJob,oneof"` + SparkJob *SparkJob `protobuf:"bytes,4,opt,name=spark_job,json=sparkJob,proto3,oneof"` } type Job_PysparkJob struct { - PysparkJob *PySparkJob `protobuf:"bytes,5,opt,name=pyspark_job,json=pysparkJob,oneof"` + PysparkJob *PySparkJob `protobuf:"bytes,5,opt,name=pyspark_job,json=pysparkJob,proto3,oneof"` } type Job_HiveJob struct { - HiveJob *HiveJob `protobuf:"bytes,6,opt,name=hive_job,json=hiveJob,oneof"` + HiveJob *HiveJob `protobuf:"bytes,6,opt,name=hive_job,json=hiveJob,proto3,oneof"` } type Job_PigJob struct { - PigJob *PigJob `protobuf:"bytes,7,opt,name=pig_job,json=pigJob,oneof"` + PigJob *PigJob `protobuf:"bytes,7,opt,name=pig_job,json=pigJob,proto3,oneof"` } type Job_SparkSqlJob struct { - SparkSqlJob *SparkSqlJob `protobuf:"bytes,12,opt,name=spark_sql_job,json=sparkSqlJob,oneof"` + SparkSqlJob *SparkSqlJob `protobuf:"bytes,12,opt,name=spark_sql_job,json=sparkSqlJob,proto3,oneof"` } func (*Job_HadoopJob) isJob_TypeJob() {} @@ -2081,9 +2083,6 @@ func _Job_OneofSizer(msg proto.Message) (n int) { } // Job scheduling options. -// -// **Beta Feature**: These options are available for testing purposes only. -// They may be changed before final release. type JobScheduling struct { // Optional. Maximum number of times per hour a driver may be restarted as // a result of driver terminating with non-zero code before job is @@ -2093,7 +2092,7 @@ type JobScheduling struct { // 4 times within 10 minute window. // // Maximum value is 10. - MaxFailuresPerHour int32 `protobuf:"varint,1,opt,name=max_failures_per_hour,json=maxFailuresPerHour" json:"max_failures_per_hour,omitempty"` + MaxFailuresPerHour int32 `protobuf:"varint,1,opt,name=max_failures_per_hour,json=maxFailuresPerHour,proto3" json:"max_failures_per_hour,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2103,7 +2102,7 @@ func (m *JobScheduling) Reset() { *m = JobScheduling{} } func (m *JobScheduling) String() string { return proto.CompactTextString(m) } func (*JobScheduling) ProtoMessage() {} func (*JobScheduling) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{13} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{13} } func (m *JobScheduling) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_JobScheduling.Unmarshal(m, b) @@ -2134,11 +2133,23 @@ func (m *JobScheduling) GetMaxFailuresPerHour() int32 { type SubmitJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job resource. - Job *Job `protobuf:"bytes,2,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"` + // Optional. A unique id used to identify the request. If the server + // receives two [SubmitJobRequest][google.cloud.dataproc.v1beta2.SubmitJobRequest] requests with the same + // id, then the second request will be ignored and the + // first [Job][google.cloud.dataproc.v1beta2.Job] created and stored in the backend + // is returned. + // + // It is recommended to always set this value to a + // [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier). + // + // The id must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). The maximum length is 40 characters. + RequestId string `protobuf:"bytes,4,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2148,7 +2159,7 @@ func (m *SubmitJobRequest) Reset() { *m = SubmitJobRequest{} } func (m *SubmitJobRequest) String() string { return proto.CompactTextString(m) } func (*SubmitJobRequest) ProtoMessage() {} func (*SubmitJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{14} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{14} } func (m *SubmitJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SubmitJobRequest.Unmarshal(m, b) @@ -2189,15 +2200,22 @@ func (m *SubmitJobRequest) GetJob() *Job { return nil } +func (m *SubmitJobRequest) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + // A request to get the resource representation for a job in a project. type GetJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2207,7 +2225,7 @@ func (m *GetJobRequest) Reset() { *m = GetJobRequest{} } func (m *GetJobRequest) String() string { return proto.CompactTextString(m) } func (*GetJobRequest) ProtoMessage() {} func (*GetJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{15} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{15} } func (m *GetJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetJobRequest.Unmarshal(m, b) @@ -2252,22 +2270,22 @@ func (m *GetJobRequest) GetJobId() string { type ListJobsRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,6,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,6,opt,name=region,proto3" json:"region,omitempty"` // Optional. The number of results to return in each response. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The page token, returned by a previous call, to request the // next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. If set, the returned jobs list includes only jobs that were // submitted to the named cluster. - ClusterName string `protobuf:"bytes,4,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + ClusterName string `protobuf:"bytes,4,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Optional. Specifies enumerated categories of jobs to list. // (default = match ALL jobs). // // If `filter` is provided, `jobStateMatcher` will be ignored. - JobStateMatcher ListJobsRequest_JobStateMatcher `protobuf:"varint,5,opt,name=job_state_matcher,json=jobStateMatcher,enum=google.cloud.dataproc.v1beta2.ListJobsRequest_JobStateMatcher" json:"job_state_matcher,omitempty"` + JobStateMatcher ListJobsRequest_JobStateMatcher `protobuf:"varint,5,opt,name=job_state_matcher,json=jobStateMatcher,proto3,enum=google.cloud.dataproc.v1beta2.ListJobsRequest_JobStateMatcher" json:"job_state_matcher,omitempty"` // Optional. A filter constraining the jobs to list. Filters are // case-sensitive and have the following syntax: // @@ -2282,7 +2300,7 @@ type ListJobsRequest struct { // Example filter: // // status.state = ACTIVE AND labels.env = staging AND labels.starred = * - Filter string `protobuf:"bytes,7,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,7,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2292,7 +2310,7 @@ func (m *ListJobsRequest) Reset() { *m = ListJobsRequest{} } func (m *ListJobsRequest) String() string { return proto.CompactTextString(m) } func (*ListJobsRequest) ProtoMessage() {} func (*ListJobsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{16} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{16} } func (m *ListJobsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListJobsRequest.Unmarshal(m, b) @@ -2365,20 +2383,20 @@ func (m *ListJobsRequest) GetFilter() string { type UpdateJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,2,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,2,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` // Required. The changes to the job. - Job *Job `protobuf:"bytes,4,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,4,opt,name=job,proto3" json:"job,omitempty"` // Required. Specifies the path, relative to Job, of // the field to update. For example, to update the labels of a Job the // update_mask parameter would be specified as // labels, and the `PATCH` request body would specify the new // value. Note: Currently, labels is the only // field that can be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2388,7 +2406,7 @@ func (m *UpdateJobRequest) Reset() { *m = UpdateJobRequest{} } func (m *UpdateJobRequest) String() string { return proto.CompactTextString(m) } func (*UpdateJobRequest) ProtoMessage() {} func (*UpdateJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{17} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{17} } func (m *UpdateJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateJobRequest.Unmarshal(m, b) @@ -2445,12 +2463,12 @@ func (m *UpdateJobRequest) GetUpdateMask() *field_mask.FieldMask { // A list of jobs in a project. type ListJobsResponse struct { - // Output-only. Jobs list. - Jobs []*Job `protobuf:"bytes,1,rep,name=jobs" json:"jobs,omitempty"` + // Output only. Jobs list. + Jobs []*Job `protobuf:"bytes,1,rep,name=jobs,proto3" json:"jobs,omitempty"` // Optional. This token is included in the response if there are more results // to fetch. To fetch additional results, provide this value as the // `page_token` in a subsequent ListJobsRequest. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2460,7 +2478,7 @@ func (m *ListJobsResponse) Reset() { *m = ListJobsResponse{} } func (m *ListJobsResponse) String() string { return proto.CompactTextString(m) } func (*ListJobsResponse) ProtoMessage() {} func (*ListJobsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{18} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{18} } func (m *ListJobsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListJobsResponse.Unmarshal(m, b) @@ -2498,11 +2516,11 @@ func (m *ListJobsResponse) GetNextPageToken() string { type CancelJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2512,7 +2530,7 @@ func (m *CancelJobRequest) Reset() { *m = CancelJobRequest{} } func (m *CancelJobRequest) String() string { return proto.CompactTextString(m) } func (*CancelJobRequest) ProtoMessage() {} func (*CancelJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{19} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{19} } func (m *CancelJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CancelJobRequest.Unmarshal(m, b) @@ -2557,11 +2575,11 @@ func (m *CancelJobRequest) GetJobId() string { type DeleteJobRequest struct { // Required. The ID of the Google Cloud Platform project that the job // belongs to. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. The Cloud Dataproc region in which to handle the request. - Region string `protobuf:"bytes,3,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,3,opt,name=region,proto3" json:"region,omitempty"` // Required. The job ID. - JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,2,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2571,7 +2589,7 @@ func (m *DeleteJobRequest) Reset() { *m = DeleteJobRequest{} } func (m *DeleteJobRequest) String() string { return proto.CompactTextString(m) } func (*DeleteJobRequest) ProtoMessage() {} func (*DeleteJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_jobs_ce1ff03ae91a7e10, []int{20} + return fileDescriptor_jobs_0f6ff1bb746c0b59, []int{20} } func (m *DeleteJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteJobRequest.Unmarshal(m, b) @@ -2744,8 +2762,7 @@ func (c *jobControllerClient) DeleteJob(ctx context.Context, in *DeleteJobReques return out, nil } -// Server API for JobController service - +// JobControllerServer is the server API for JobController service. type JobControllerServer interface { // Submits a job to a cluster. SubmitJob(context.Context, *SubmitJobRequest) (*Job, error) @@ -2911,153 +2928,154 @@ var _JobController_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/cloud/dataproc/v1beta2/jobs.proto", fileDescriptor_jobs_ce1ff03ae91a7e10) + proto.RegisterFile("google/cloud/dataproc/v1beta2/jobs.proto", fileDescriptor_jobs_0f6ff1bb746c0b59) } -var fileDescriptor_jobs_ce1ff03ae91a7e10 = []byte{ - // 2294 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xdc, 0x5a, 0xcd, 0x6f, 0xdb, 0xc8, - 0x15, 0xb7, 0xbe, 0xc5, 0x27, 0x4b, 0xa6, 0x67, 0x93, 0xad, 0xa0, 0x74, 0xb1, 0x5e, 0x02, 0x9b, - 0xba, 0xd9, 0x42, 0x42, 0xd4, 0x34, 0x9b, 0x4d, 0xba, 0xdd, 0xc8, 0x12, 0x15, 0xc9, 0x55, 0x64, - 0x2d, 0x25, 0x25, 0xdd, 0x2d, 0x0a, 0x2e, 0x25, 0x8d, 0x65, 0xca, 0x14, 0xc9, 0x70, 0x48, 0x37, - 0xda, 0x20, 0x28, 0xd0, 0x4b, 0x0f, 0x3d, 0xf6, 0x52, 0xa0, 0x40, 0x81, 0xde, 0xba, 0x40, 0x2f, - 0xbd, 0xf6, 0x1f, 0x28, 0x7a, 0x69, 0x0f, 0xfb, 0x27, 0xb4, 0x87, 0x1e, 0x7b, 0xea, 0xb9, 0x98, - 0x19, 0x52, 0x96, 0x64, 0x27, 0xa2, 0xe3, 0x7e, 0x65, 0x4f, 0x26, 0xdf, 0xd7, 0xbc, 0x99, 0xdf, - 0x6f, 0xde, 0xbc, 0xa1, 0x0c, 0xbb, 0x63, 0xcb, 0x1a, 0x1b, 0xb8, 0x34, 0x34, 0x2c, 0x6f, 0x54, - 0x1a, 0x69, 0xae, 0x66, 0x3b, 0xd6, 0xb0, 0x74, 0x72, 0x73, 0x80, 0x5d, 0xad, 0x5c, 0x9a, 0x58, - 0x03, 0x52, 0xb4, 0x1d, 0xcb, 0xb5, 0xd0, 0x5b, 0xdc, 0xb2, 0xc8, 0x2c, 0x8b, 0x81, 0x65, 0xd1, - 0xb7, 0x2c, 0x7c, 0xdd, 0x0f, 0xa4, 0xd9, 0x7a, 0x49, 0x33, 0x4d, 0xcb, 0xd5, 0x5c, 0xdd, 0x32, - 0x7d, 0xe7, 0xc2, 0x35, 0x5f, 0xcb, 0xde, 0x06, 0xde, 0x61, 0x09, 0x4f, 0x6d, 0x77, 0xe6, 0x2b, - 0x77, 0x56, 0x95, 0x87, 0x3a, 0x36, 0x46, 0xea, 0x54, 0x23, 0xc7, 0xbe, 0xc5, 0xdb, 0xab, 0x16, - 0xae, 0x3e, 0xc5, 0xc4, 0xd5, 0xa6, 0x36, 0x37, 0x90, 0xfe, 0x1e, 0x85, 0x6c, 0xcb, 0x1a, 0x8f, - 0x75, 0x73, 0x5c, 0xb5, 0xcc, 0x43, 0x7d, 0x8c, 0xa6, 0xb0, 0x3d, 0x72, 0xf4, 0x13, 0xec, 0xa8, - 0x86, 0x35, 0x56, 0x0d, 0x7c, 0x82, 0x0d, 0x92, 0x8f, 0xee, 0xc4, 0x76, 0x33, 0xe5, 0x4a, 0xf1, - 0xa5, 0x53, 0x29, 0x2e, 0x05, 0x2a, 0xd6, 0x58, 0x94, 0x96, 0x35, 0x6e, 0xb1, 0x18, 0xb2, 0xe9, - 0x3a, 0x33, 0x65, 0x6b, 0xb4, 0x2c, 0x2d, 0x9c, 0xc0, 0x95, 0xf3, 0x0c, 0x91, 0x08, 0xb1, 0x63, - 0x3c, 0xcb, 0x47, 0x76, 0x22, 0xbb, 0x82, 0x42, 0x1f, 0x51, 0x03, 0x12, 0x27, 0x9a, 0xe1, 0xe1, - 0x7c, 0x74, 0x27, 0xb2, 0x9b, 0x2b, 0x97, 0x2f, 0x94, 0x0c, 0x0b, 0xad, 0xf0, 0x00, 0x77, 0xa3, - 0x77, 0x22, 0x92, 0x0d, 0x09, 0x26, 0x43, 0x57, 0x61, 0xbb, 0x25, 0x3f, 0x92, 0x5b, 0x6a, 0xbf, - 0xdd, 0xed, 0xc8, 0xd5, 0x66, 0xbd, 0x29, 0xd7, 0xc4, 0x0d, 0x94, 0x82, 0x58, 0xa5, 0xd5, 0x12, - 0x23, 0x48, 0x80, 0x44, 0x4f, 0xa9, 0x54, 0x65, 0x31, 0x4a, 0x1f, 0x6b, 0xf2, 0x5e, 0xff, 0x81, - 0x18, 0x43, 0x69, 0x88, 0x37, 0xdb, 0xf5, 0x03, 0x31, 0x4e, 0x9f, 0x1e, 0x57, 0x94, 0xb6, 0x98, - 0xa0, 0x6a, 0x59, 0x51, 0x0e, 0x14, 0x31, 0x49, 0x1f, 0xeb, 0x95, 0x5e, 0xa5, 0x25, 0xa6, 0x68, - 0xa0, 0x83, 0x7a, 0x5d, 0x4c, 0x4b, 0x7f, 0x8a, 0x81, 0xd0, 0xd0, 0x46, 0x96, 0x65, 0xef, 0x5b, - 0x03, 0xf4, 0x1e, 0x6c, 0x4f, 0x35, 0xdd, 0x54, 0x27, 0x9a, 0xa3, 0x1e, 0xea, 0x06, 0x56, 0x3d, - 0x47, 0xe7, 0xb3, 0x6d, 0x6c, 0x28, 0x39, 0xaa, 0xda, 0xd7, 0x9c, 0xba, 0x6e, 0xe0, 0xbe, 0xa3, - 0xa3, 0xb7, 0x01, 0x98, 0xf1, 0xd0, 0xd0, 0x08, 0x61, 0xf3, 0xa7, 0x56, 0x02, 0x95, 0x55, 0xa9, - 0x08, 0x21, 0x88, 0x6b, 0xce, 0x98, 0xe4, 0x63, 0x3b, 0xb1, 0x5d, 0x41, 0x61, 0xcf, 0x48, 0x82, - 0xec, 0x62, 0x70, 0x92, 0x8f, 0x33, 0x65, 0x66, 0x32, 0x8f, 0x4b, 0xd0, 0x35, 0x10, 0x4e, 0xf5, - 0x09, 0xa6, 0x4f, 0x1f, 0x06, 0xca, 0x77, 0x60, 0x53, 0x73, 0x86, 0x47, 0xfa, 0x89, 0xaf, 0x4f, - 0x72, 0x7f, 0x5f, 0xc6, 0x4c, 0x7e, 0x00, 0x60, 0x3b, 0x96, 0x8d, 0x1d, 0x57, 0xc7, 0x24, 0x9f, - 0x62, 0x2c, 0xb9, 0xb3, 0x06, 0x98, 0xf9, 0x1a, 0x14, 0x3b, 0x73, 0x57, 0x4e, 0x8e, 0x85, 0x58, - 0xa8, 0x0b, 0x39, 0x83, 0x23, 0xa8, 0x0e, 0x19, 0x84, 0xf9, 0xf4, 0x4e, 0x64, 0x37, 0x53, 0xfe, - 0xd6, 0x45, 0x60, 0x57, 0xb2, 0xc6, 0xe2, 0x6b, 0xe1, 0x43, 0xd8, 0x5a, 0x19, 0xf3, 0x1c, 0x9e, - 0x5d, 0x59, 0xe4, 0x99, 0xb0, 0xc0, 0x99, 0xbd, 0x34, 0x24, 0x39, 0x7d, 0xa5, 0x3f, 0xc6, 0x20, - 0xdd, 0xb5, 0x35, 0xe7, 0xf8, 0xab, 0x03, 0xe5, 0xe3, 0x73, 0xa0, 0x7c, 0x7f, 0xcd, 0x62, 0x07, - 0x4b, 0xf0, 0x1a, 0x23, 0xf9, 0xe7, 0x18, 0x40, 0x67, 0x36, 0xc7, 0xb2, 0x04, 0x57, 0x18, 0x3c, - 0xf6, 0xcc, 0x3d, 0xb2, 0xcc, 0x15, 0x38, 0x15, 0x86, 0x73, 0x87, 0xa9, 0x02, 0x3c, 0x03, 0xb8, - 0xa2, 0x0b, 0x70, 0xed, 0x82, 0xb8, 0xe2, 0x1f, 0xc0, 0x99, 0xb3, 0x17, 0x9d, 0xff, 0x3b, 0xc0, - 0x7e, 0x72, 0x0e, 0xb0, 0x1f, 0xac, 0x59, 0xfb, 0xd3, 0x15, 0x79, 0xdd, 0xa0, 0x95, 0xde, 0x05, - 0xe1, 0x63, 0x0f, 0x3b, 0xb3, 0x96, 0x4e, 0x5c, 0x94, 0x87, 0xd4, 0x13, 0x0f, 0x3b, 0x74, 0xe2, - 0x11, 0xb6, 0x32, 0xc1, 0xab, 0xf4, 0xeb, 0x38, 0xa4, 0x1a, 0xfa, 0x09, 0xa6, 0xa0, 0x5f, 0x87, - 0x1c, 0x15, 0xcf, 0xce, 0xee, 0xde, 0x4d, 0x26, 0x0f, 0xb0, 0x6e, 0x02, 0x70, 0x3b, 0x43, 0x27, - 0x2e, 0x1b, 0x39, 0x53, 0xde, 0x5d, 0x33, 0xd5, 0x79, 0x2e, 0x74, 0x97, 0x3f, 0x99, 0x27, 0x56, - 0x84, 0x37, 0x86, 0x96, 0xe9, 0xea, 0xa6, 0x87, 0x55, 0xca, 0x13, 0x4d, 0x37, 0x3c, 0x07, 0xe7, - 0x63, 0x3b, 0x91, 0xdd, 0xb4, 0xb2, 0x1d, 0xa8, 0x0e, 0xcc, 0x3a, 0x57, 0xa0, 0x43, 0x10, 0xc9, - 0xd0, 0xd1, 0x6d, 0x57, 0x3d, 0xd1, 0x1c, 0x5d, 0x1b, 0x18, 0x98, 0x73, 0x25, 0x53, 0xbe, 0xb7, - 0xae, 0xdc, 0xf2, 0x49, 0x16, 0xbb, 0xcc, 0xfd, 0x51, 0xe0, 0xed, 0x1f, 0xc7, 0x64, 0x59, 0x8a, - 0x1e, 0x2d, 0x91, 0x25, 0xc1, 0x46, 0xb8, 0x1d, 0x72, 0x84, 0x97, 0x31, 0xe5, 0x0c, 0xd1, 0x93, - 0x67, 0x88, 0x5e, 0xd8, 0x83, 0x2b, 0xe7, 0x25, 0x79, 0x11, 0xf4, 0x2f, 0x5b, 0x17, 0x84, 0x39, - 0x5f, 0xa4, 0xbf, 0xc4, 0x21, 0xc3, 0x36, 0x41, 0xf7, 0x89, 0xf1, 0x3f, 0x22, 0xc9, 0xe4, 0x1c, - 0xd0, 0x63, 0x0c, 0x92, 0x8f, 0xc2, 0x14, 0x66, 0x9e, 0x78, 0x48, 0xe0, 0x3f, 0x5d, 0x02, 0x9e, - 0x53, 0xeb, 0xee, 0x05, 0x46, 0xb9, 0x10, 0xf8, 0x77, 0xce, 0x56, 0xb9, 0xb3, 0xa5, 0x24, 0x79, - 0xf9, 0x52, 0xf2, 0xff, 0xc5, 0xa8, 0x7f, 0xc4, 0x21, 0xd9, 0xd1, 0xc7, 0xaf, 0x49, 0xc5, 0xc1, - 0x2f, 0xac, 0x38, 0xeb, 0x68, 0xc1, 0xe7, 0x18, 0x92, 0x77, 0xfd, 0x73, 0x0a, 0xce, 0x77, 0xc2, - 0x0d, 0x70, 0xc9, 0x7a, 0x73, 0x0e, 0xe5, 0x52, 0x5f, 0x35, 0xca, 0xf5, 0x60, 0x73, 0xdf, 0x1a, - 0x74, 0x0c, 0x6d, 0x88, 0xa7, 0xd8, 0x74, 0x69, 0xbb, 0x30, 0x34, 0x3c, 0xe2, 0x62, 0x47, 0x35, - 0xb5, 0x29, 0xf6, 0xe3, 0x65, 0x7c, 0x59, 0x5b, 0x9b, 0xe2, 0x45, 0x13, 0xcf, 0xd3, 0x47, 0x7e, - 0xf8, 0xc0, 0xa4, 0xef, 0xe9, 0x23, 0xe9, 0x9f, 0x31, 0x10, 0xf6, 0xad, 0x41, 0xd7, 0xd5, 0x5c, - 0x8f, 0xa0, 0x1a, 0x24, 0x88, 0xab, 0xb9, 0x3c, 0x58, 0xae, 0x5c, 0x5c, 0xb3, 0x7a, 0x73, 0xc7, - 0x22, 0xfd, 0x83, 0x15, 0xee, 0x4c, 0x4f, 0xea, 0x11, 0x76, 0x35, 0xdd, 0xf0, 0x9b, 0x62, 0x25, - 0x78, 0x45, 0x35, 0x10, 0x99, 0x89, 0x4a, 0x5c, 0xcd, 0x71, 0x55, 0x7a, 0x83, 0xf5, 0x6b, 0x43, - 0x21, 0x18, 0x2a, 0xb8, 0xde, 0x16, 0x7b, 0xc1, 0xf5, 0x56, 0xc9, 0x31, 0x9f, 0x2e, 0x75, 0xa1, - 0x42, 0xf4, 0x10, 0xd2, 0xc4, 0x1b, 0xf0, 0x44, 0x53, 0x2c, 0xd1, 0x9b, 0xe1, 0x13, 0xf5, 0x1d, - 0x95, 0x79, 0x08, 0xe9, 0x8b, 0x08, 0x24, 0x58, 0xfe, 0xf4, 0xfe, 0xd8, 0xed, 0x55, 0x7a, 0xf2, - 0xca, 0xfd, 0x31, 0x03, 0xa9, 0x8e, 0xdc, 0xae, 0x35, 0xdb, 0x0f, 0xc4, 0x08, 0xca, 0x01, 0x74, - 0xe5, 0x5e, 0xbf, 0xa3, 0xd6, 0x0e, 0xda, 0xb2, 0x98, 0xa6, 0x4a, 0xa5, 0xdf, 0x6e, 0x53, 0x65, - 0x14, 0x21, 0xc8, 0x55, 0x2b, 0xed, 0xaa, 0xdc, 0x52, 0x03, 0x87, 0xd8, 0x82, 0xac, 0xdb, 0xab, - 0x28, 0x3d, 0xb9, 0x26, 0xa6, 0x50, 0x16, 0x04, 0x2e, 0x6b, 0xc9, 0x35, 0x7e, 0xef, 0x64, 0xd1, - 0x96, 0xee, 0x9d, 0x6f, 0xc0, 0x56, 0xa5, 0xd7, 0x93, 0x1f, 0x76, 0x7a, 0x6a, 0xbd, 0xd2, 0x6c, - 0xf5, 0x15, 0x59, 0x14, 0xa4, 0x06, 0xa4, 0x83, 0x19, 0xa0, 0x2d, 0xc8, 0x2c, 0xe7, 0x99, 0x05, - 0xa1, 0xdb, 0xdf, 0x7b, 0xd8, 0xec, 0xd1, 0x41, 0x22, 0x08, 0x20, 0xf9, 0x71, 0x5f, 0xee, 0xcb, - 0x35, 0x31, 0x8a, 0x44, 0xd8, 0xec, 0xf6, 0x2a, 0x2d, 0x99, 0xe6, 0xd0, 0xeb, 0x77, 0xc5, 0x98, - 0x54, 0x63, 0x74, 0x52, 0xf0, 0x21, 0x76, 0xb0, 0x39, 0xc4, 0xe8, 0x2d, 0xb6, 0x79, 0x27, 0x78, - 0xe8, 0xaa, 0xfa, 0xc8, 0x27, 0x93, 0xe0, 0x4b, 0x9a, 0x23, 0x74, 0x15, 0x92, 0x13, 0x6b, 0xa0, - 0xce, 0x49, 0x94, 0x98, 0x58, 0x83, 0xe6, 0x48, 0xfa, 0x43, 0x14, 0xb6, 0x3e, 0xd1, 0x1c, 0xb3, - 0x62, 0xdb, 0x86, 0x3e, 0x64, 0x9f, 0x3b, 0x68, 0x1b, 0xbd, 0x40, 0x48, 0xf6, 0x8c, 0xf6, 0x03, - 0x62, 0xf1, 0x0b, 0xff, 0xad, 0x35, 0x78, 0xad, 0x84, 0x5c, 0xa6, 0x57, 0x01, 0xd2, 0xb6, 0x63, - 0x8d, 0x1d, 0x4c, 0x08, 0x2b, 0x79, 0x51, 0x65, 0xfe, 0x4e, 0x19, 0xef, 0x3a, 0xda, 0xf0, 0x98, - 0x16, 0x02, 0xcf, 0x31, 0xf2, 0x71, 0xce, 0xf8, 0x40, 0xd6, 0x77, 0x0c, 0xe9, 0x67, 0xeb, 0xe0, - 0x4e, 0x41, 0xac, 0x2d, 0x3f, 0xe6, 0x50, 0xb7, 0xe5, 0xc7, 0x6a, 0xb7, 0xf2, 0x88, 0xa3, 0xbb, - 0xb4, 0xbe, 0x31, 0xb4, 0x09, 0xe9, 0x4a, 0xb5, 0x2a, 0x77, 0x7a, 0x0c, 0xc3, 0x05, 0x1e, 0x24, - 0xa8, 0xaa, 0xde, 0x6c, 0x37, 0xbb, 0x0d, 0xb9, 0x26, 0x26, 0x29, 0x10, 0x14, 0x41, 0x86, 0x3c, - 0x40, 0xf2, 0xfb, 0x4d, 0x06, 0x7b, 0x5a, 0xfa, 0xa5, 0x00, 0x31, 0x7a, 0x82, 0x34, 0x41, 0x70, - 0x02, 0x1c, 0xd8, 0xaa, 0x65, 0xca, 0xef, 0xad, 0x27, 0xf4, 0x1c, 0x3a, 0xe5, 0xd4, 0x9b, 0x86, - 0xb2, 0x83, 0x0a, 0xe1, 0x9f, 0x31, 0x21, 0x42, 0xcd, 0x8b, 0x8a, 0x72, 0xea, 0x4d, 0xcf, 0xab, - 0x23, 0x76, 0xbd, 0x57, 0x27, 0xd6, 0x80, 0x2d, 0xf4, 0xfa, 0xf3, 0x6a, 0xfe, 0x3d, 0x80, 0x9e, - 0x57, 0x47, 0xf3, 0x0f, 0x24, 0x75, 0x10, 0x08, 0xed, 0x2f, 0x58, 0xa4, 0x38, 0x8b, 0xf4, 0x8d, - 0x90, 0xd7, 0xd1, 0xc6, 0x86, 0x92, 0x26, 0xc1, 0x8d, 0xae, 0x05, 0x19, 0x7b, 0x76, 0x1a, 0x29, - 0xc1, 0x22, 0x7d, 0x33, 0xf4, 0xfd, 0xa7, 0xb1, 0xa1, 0x80, 0xef, 0x4f, 0xa3, 0x55, 0x21, 0xcd, - 0x2e, 0x5b, 0x34, 0x14, 0x2f, 0x42, 0xd7, 0xc3, 0x75, 0xc7, 0x8d, 0x0d, 0x25, 0x75, 0xe4, 0xdf, - 0x37, 0xee, 0x43, 0xca, 0xd6, 0xc7, 0x2c, 0x06, 0x3f, 0x71, 0xde, 0x0d, 0x75, 0xe0, 0x35, 0x36, - 0x94, 0xa4, 0xcd, 0xfb, 0x87, 0x0e, 0x64, 0xf9, 0x94, 0xc8, 0x13, 0x83, 0xc5, 0xd9, 0x64, 0x71, - 0x6e, 0x84, 0x6f, 0xd8, 0x1a, 0x1b, 0x4a, 0x86, 0x2c, 0xb4, 0xb7, 0xf7, 0x21, 0x49, 0x58, 0xb5, - 0xf3, 0xaf, 0x70, 0xbb, 0x61, 0xab, 0xa3, 0xe2, 0xfb, 0xa1, 0x03, 0xc8, 0xf1, 0x27, 0xf5, 0x48, - 0x27, 0xae, 0xe5, 0xcc, 0xf2, 0x59, 0x76, 0x9a, 0x87, 0x8f, 0x94, 0xe5, 0xfe, 0x0d, 0xee, 0x8e, - 0x7e, 0x08, 0xdb, 0x33, 0xcd, 0x31, 0x55, 0xed, 0x74, 0x53, 0x93, 0xbc, 0xc0, 0x62, 0x16, 0x2f, - 0x56, 0x0b, 0x14, 0x71, 0xb6, 0x2c, 0x20, 0xe8, 0x1e, 0x14, 0xfc, 0xcf, 0x9c, 0x96, 0xe7, 0xda, - 0x9e, 0xab, 0x3a, 0x98, 0x58, 0x9e, 0x33, 0xe4, 0xdd, 0xd8, 0x36, 0x2b, 0x01, 0x5f, 0xe3, 0x16, - 0x07, 0xcc, 0x40, 0xf1, 0xf5, 0xb4, 0x2d, 0x7b, 0x1f, 0xf2, 0xbe, 0x33, 0xed, 0x9b, 0x1c, 0xcb, - 0x60, 0x8d, 0x06, 0x61, 0xae, 0x5b, 0xcc, 0xf5, 0x2a, 0xd7, 0x57, 0xb9, 0x9a, 0xb6, 0x1c, 0x84, - 0x3a, 0xd6, 0x21, 0x69, 0x68, 0x03, 0x6c, 0x90, 0x3c, 0x0a, 0x35, 0x0f, 0xda, 0xe6, 0xb4, 0x98, - 0x03, 0x6f, 0x71, 0x7c, 0x6f, 0xd4, 0x02, 0x20, 0xc3, 0x23, 0x3c, 0xf2, 0x0c, 0xdd, 0x1c, 0xe7, - 0xaf, 0x84, 0x6a, 0x5b, 0xe8, 0x3a, 0xcf, 0x7d, 0x94, 0x05, 0xff, 0xc2, 0x07, 0x90, 0x59, 0x18, - 0xe4, 0x42, 0xbd, 0x06, 0x40, 0xda, 0x9d, 0xd9, 0x6c, 0x3f, 0x48, 0x7b, 0x90, 0x5d, 0x1a, 0x03, - 0xdd, 0x84, 0xab, 0x53, 0xed, 0x69, 0xd0, 0x6a, 0x12, 0xd5, 0xc6, 0x8e, 0x7a, 0x64, 0x79, 0x0e, - 0x0b, 0x9d, 0x50, 0xd0, 0x54, 0x7b, 0xea, 0x77, 0x9b, 0xa4, 0x83, 0x9d, 0x86, 0xe5, 0x39, 0xd2, - 0x4f, 0x40, 0xec, 0x7a, 0x83, 0xa9, 0xee, 0xb2, 0x62, 0xf5, 0xc4, 0xc3, 0xc4, 0x5d, 0x77, 0xca, - 0xbc, 0x09, 0x49, 0x07, 0x8f, 0x75, 0xcb, 0x64, 0xf5, 0x46, 0x50, 0xfc, 0x37, 0x74, 0x0b, 0x62, - 0x74, 0x67, 0xf0, 0x82, 0x26, 0x85, 0xa8, 0x8d, 0xd4, 0x5c, 0xfa, 0x11, 0x64, 0x1f, 0xe0, 0x7f, - 0xc3, 0xe8, 0x2f, 0x38, 0xfb, 0xfe, 0x1a, 0x85, 0x2d, 0xda, 0x8e, 0xef, 0x5b, 0x03, 0x72, 0xe1, - 0x11, 0x92, 0x4b, 0x23, 0x5c, 0x03, 0xc1, 0xd6, 0xc6, 0x58, 0x25, 0xfa, 0xe7, 0x1c, 0x98, 0x84, - 0x92, 0xa6, 0x82, 0xae, 0xfe, 0x39, 0x3f, 0x99, 0xa9, 0xd2, 0xb5, 0x8e, 0x71, 0x90, 0x1a, 0x33, - 0xef, 0x51, 0xc1, 0x99, 0x3e, 0x30, 0x7e, 0xb6, 0x0f, 0x9c, 0xc0, 0x36, 0x9d, 0x00, 0x6f, 0xbd, - 0xa6, 0x9a, 0x3b, 0x3c, 0xc2, 0x0e, 0xab, 0x9e, 0xb9, 0xf2, 0xf7, 0xd6, 0x35, 0xc8, 0xcb, 0x13, - 0x0c, 0x76, 0x38, 0x7e, 0xc8, 0xa3, 0x28, 0x5b, 0x93, 0x65, 0x01, 0x9d, 0xe2, 0xa1, 0x6e, 0xb8, - 0xd8, 0x61, 0xf5, 0x50, 0x50, 0xfc, 0x37, 0xe9, 0x36, 0x6c, 0xad, 0xf8, 0x06, 0xdf, 0xe5, 0x37, - 0xe8, 0xa1, 0x58, 0xa9, 0xf6, 0x9a, 0x8f, 0x64, 0xff, 0xd0, 0x3d, 0x68, 0xab, 0xfe, 0x7b, 0x54, - 0xfa, 0x32, 0x02, 0x62, 0xdf, 0x1e, 0x69, 0x2e, 0x7e, 0x15, 0x20, 0xa3, 0x2f, 0x00, 0x32, 0xb6, - 0x00, 0x64, 0xc0, 0xae, 0xf8, 0x85, 0xd8, 0x85, 0xee, 0x41, 0xc6, 0x63, 0x79, 0xb1, 0x1f, 0x69, - 0xfc, 0xc3, 0xe8, 0x6c, 0x1b, 0x5b, 0xd7, 0xb1, 0x31, 0x7a, 0xa8, 0x91, 0x63, 0x05, 0xb8, 0x39, - 0x7d, 0x96, 0x1c, 0x10, 0x4f, 0x57, 0x96, 0xd8, 0x96, 0x49, 0x30, 0xba, 0x0d, 0xf1, 0x89, 0x35, - 0xe0, 0x5f, 0xb7, 0xc2, 0xe5, 0xc1, 0xec, 0xd1, 0x75, 0xd8, 0x32, 0xf1, 0x53, 0x57, 0x5d, 0x20, - 0x09, 0x9f, 0x76, 0x96, 0x8a, 0x3b, 0x01, 0x51, 0xa4, 0xcf, 0x40, 0xac, 0x6a, 0xe6, 0x10, 0x1b, - 0xff, 0xb1, 0x1d, 0xf1, 0x19, 0x88, 0x35, 0x6c, 0xe0, 0x57, 0x83, 0x2a, 0xcc, 0x08, 0xe5, 0x9f, - 0xa7, 0x59, 0x61, 0xf2, 0x6b, 0xb1, 0x81, 0x1d, 0xf4, 0xdb, 0x08, 0x08, 0xf3, 0x32, 0x83, 0x4a, - 0xeb, 0x4e, 0xcd, 0x95, 0x82, 0x54, 0x08, 0xb1, 0xcc, 0x52, 0xfd, 0xa7, 0x5f, 0xfe, 0xed, 0x17, - 0xd1, 0xfb, 0xd2, 0xbd, 0xf9, 0x4f, 0x82, 0x7e, 0xfe, 0xa4, 0xf4, 0xec, 0x74, 0x6e, 0xcf, 0x4b, - 0x3c, 0x75, 0x52, 0x7a, 0xc6, 0x1f, 0x9e, 0xb3, 0x5f, 0x0e, 0xef, 0x12, 0x36, 0xe4, 0xdd, 0xc8, - 0x0d, 0xf4, 0x9b, 0x08, 0x24, 0x79, 0x41, 0x42, 0xeb, 0x0a, 0xfc, 0x52, 0xdd, 0x0a, 0x95, 0xa4, - 0xcc, 0x92, 0xfc, 0x08, 0x7d, 0xf8, 0x2a, 0x49, 0x96, 0x9e, 0xf1, 0xc5, 0x7e, 0x8e, 0xbe, 0x88, - 0x40, 0x3a, 0x60, 0x26, 0x2a, 0x5e, 0xac, 0x38, 0x14, 0x4a, 0xa1, 0xed, 0x39, 0xe5, 0xa5, 0xef, - 0xb2, 0xa4, 0x6f, 0xa3, 0x5b, 0xaf, 0x92, 0x34, 0xfa, 0x5d, 0x04, 0x84, 0x79, 0x69, 0x58, 0x0b, - 0xfd, 0x6a, 0x11, 0x09, 0xb5, 0xaa, 0xfb, 0x2c, 0xc1, 0x5a, 0xf9, 0x72, 0xab, 0x7a, 0x97, 0x15, - 0x8c, 0xdf, 0x47, 0x40, 0x98, 0x6f, 0xc0, 0xb5, 0xe9, 0xae, 0x6e, 0xd5, 0x50, 0xe9, 0x1e, 0xb0, - 0x74, 0x9b, 0x52, 0xed, 0x72, 0xe9, 0x0e, 0xd9, 0xd8, 0x94, 0xb2, 0xbf, 0x8a, 0x80, 0x30, 0xdf, - 0xd2, 0x6b, 0x73, 0x5e, 0xdd, 0xfc, 0x85, 0x37, 0xcf, 0x94, 0x43, 0x79, 0x6a, 0xbb, 0xb3, 0x80, - 0xac, 0x37, 0x2e, 0xb7, 0xac, 0x7b, 0x3f, 0x86, 0x77, 0x86, 0xd6, 0xf4, 0xe5, 0x49, 0xed, 0x09, - 0x94, 0x71, 0x1d, 0x3a, 0x7e, 0x27, 0xf2, 0xa9, 0xec, 0xdb, 0x8e, 0x2d, 0x43, 0x33, 0xc7, 0x45, - 0xcb, 0x19, 0x97, 0xc6, 0xd8, 0x64, 0xd9, 0x95, 0xb8, 0x4a, 0xb3, 0x75, 0xf2, 0x82, 0xff, 0x04, - 0xb8, 0x17, 0x08, 0x06, 0x49, 0xe6, 0xf1, 0xed, 0x7f, 0x05, 0x00, 0x00, 0xff, 0xff, 0xc8, 0x83, - 0x17, 0x31, 0x3a, 0x20, 0x00, 0x00, +var fileDescriptor_jobs_0f6ff1bb746c0b59 = []byte{ + // 2306 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xdc, 0x5a, 0xcd, 0x73, 0xdb, 0xc6, + 0x15, 0x17, 0xbf, 0x89, 0x47, 0x7d, 0x40, 0x1b, 0x3b, 0xe5, 0x30, 0xcd, 0x44, 0xc1, 0x4c, 0x52, + 0x35, 0xe9, 0x90, 0x13, 0xd6, 0x75, 0x1c, 0xbb, 0x69, 0x4c, 0x91, 0xa0, 0x49, 0x95, 0xa6, 0x18, + 0x90, 0xb4, 0x9b, 0x74, 0x3a, 0x08, 0x48, 0xae, 0x28, 0x48, 0x20, 0x00, 0xef, 0x02, 0xaa, 0x19, + 0x8f, 0x2f, 0xbd, 0xf4, 0xd0, 0x63, 0x2f, 0x9d, 0xe9, 0x4c, 0x66, 0x7a, 0x6b, 0x66, 0x7a, 0xe9, + 0xb5, 0xff, 0x40, 0xa7, 0x97, 0xf6, 0x90, 0x3f, 0xa1, 0x3d, 0xf4, 0xd8, 0x53, 0xcf, 0x9d, 0xdd, + 0x05, 0x28, 0x92, 0x92, 0x4d, 0xc8, 0xea, 0x97, 0x73, 0x12, 0xf0, 0xbe, 0xf6, 0xed, 0xfe, 0x7e, + 0xfb, 0xf6, 0x2d, 0x28, 0xd8, 0x1d, 0x3b, 0xce, 0xd8, 0xc2, 0xa5, 0xa1, 0xe5, 0xf8, 0xa3, 0xd2, + 0xc8, 0xf0, 0x0c, 0x97, 0x38, 0xc3, 0xd2, 0xe9, 0x7b, 0x03, 0xec, 0x19, 0xe5, 0xd2, 0xb1, 0x33, + 0xa0, 0x45, 0x97, 0x38, 0x9e, 0x83, 0x5e, 0x17, 0x96, 0x45, 0x6e, 0x59, 0x0c, 0x2d, 0x8b, 0x81, + 0x65, 0xe1, 0x9b, 0x41, 0x20, 0xc3, 0x35, 0x4b, 0x86, 0x6d, 0x3b, 0x9e, 0xe1, 0x99, 0x8e, 0x1d, + 0x38, 0x17, 0x5e, 0x0b, 0xb4, 0xfc, 0x6d, 0xe0, 0x1f, 0x96, 0xf0, 0xc4, 0xf5, 0xa6, 0x81, 0x72, + 0x67, 0x59, 0x79, 0x68, 0x62, 0x6b, 0xa4, 0x4f, 0x0c, 0x7a, 0x12, 0x58, 0xbc, 0xb1, 0x6c, 0xe1, + 0x99, 0x13, 0x4c, 0x3d, 0x63, 0xe2, 0x0a, 0x03, 0xe5, 0xef, 0x71, 0xd8, 0x68, 0x39, 0xe3, 0xb1, + 0x69, 0x8f, 0xab, 0x8e, 0x7d, 0x68, 0x8e, 0xd1, 0x04, 0xb6, 0x47, 0xc4, 0x3c, 0xc5, 0x44, 0xb7, + 0x9c, 0xb1, 0x6e, 0xe1, 0x53, 0x6c, 0xd1, 0x7c, 0x7c, 0x27, 0xb1, 0x9b, 0x2b, 0x57, 0x8a, 0xcf, + 0x9d, 0x4a, 0x71, 0x21, 0x50, 0xb1, 0xc6, 0xa3, 0xb4, 0x9c, 0x71, 0x8b, 0xc7, 0x50, 0x6d, 0x8f, + 0x4c, 0xb5, 0xad, 0xd1, 0xa2, 0xb4, 0x70, 0x0a, 0xd7, 0x2e, 0x32, 0x44, 0x32, 0x24, 0x4e, 0xf0, + 0x34, 0x1f, 0xdb, 0x89, 0xed, 0x4a, 0x1a, 0x7b, 0x44, 0x0d, 0x48, 0x9d, 0x1a, 0x96, 0x8f, 0xf3, + 0xf1, 0x9d, 0xd8, 0xee, 0x66, 0xb9, 0x7c, 0xa9, 0x64, 0x78, 0x68, 0x4d, 0x04, 0xb8, 0x1d, 0xbf, + 0x15, 0x53, 0x5c, 0x48, 0x71, 0x19, 0xba, 0x0e, 0xdb, 0x2d, 0xf5, 0x81, 0xda, 0xd2, 0xfb, 0xed, + 0x6e, 0x47, 0xad, 0x36, 0xeb, 0x4d, 0xb5, 0x26, 0xaf, 0xa1, 0x0c, 0x24, 0x2a, 0xad, 0x96, 0x1c, + 0x43, 0x12, 0xa4, 0x7a, 0x5a, 0xa5, 0xaa, 0xca, 0x71, 0xf6, 0x58, 0x53, 0xf7, 0xfa, 0xf7, 0xe4, + 0x04, 0xca, 0x42, 0xb2, 0xd9, 0xae, 0x1f, 0xc8, 0x49, 0xf6, 0xf4, 0xb0, 0xa2, 0xb5, 0xe5, 0x14, + 0x53, 0xab, 0x9a, 0x76, 0xa0, 0xc9, 0x69, 0xf6, 0x58, 0xaf, 0xf4, 0x2a, 0x2d, 0x39, 0xc3, 0x02, + 0x1d, 0xd4, 0xeb, 0x72, 0x56, 0xf9, 0x53, 0x02, 0xa4, 0x86, 0x31, 0x72, 0x1c, 0x77, 0xdf, 0x19, + 0xa0, 0x77, 0x61, 0x7b, 0x62, 0x98, 0xb6, 0x7e, 0x6c, 0x10, 0xfd, 0xd0, 0xb4, 0xb0, 0xee, 0x13, + 0x53, 0xcc, 0xb6, 0xb1, 0xa6, 0x6d, 0x32, 0xd5, 0xbe, 0x41, 0xea, 0xa6, 0x85, 0xfb, 0xc4, 0x44, + 0x6f, 0x00, 0x70, 0xe3, 0xa1, 0x65, 0x50, 0xca, 0xe7, 0xcf, 0xac, 0x24, 0x26, 0xab, 0x32, 0x11, + 0x42, 0x90, 0x34, 0xc8, 0x98, 0xe6, 0x13, 0x3b, 0x89, 0x5d, 0x49, 0xe3, 0xcf, 0x48, 0x81, 0x8d, + 0xf9, 0xe0, 0x34, 0x9f, 0xe4, 0xca, 0xdc, 0xf1, 0x2c, 0x2e, 0x45, 0xaf, 0x81, 0x74, 0xa6, 0x4f, + 0x71, 0x7d, 0xf6, 0x30, 0x54, 0xbe, 0x09, 0xeb, 0x06, 0x19, 0x1e, 0x99, 0xa7, 0x81, 0x3e, 0x2d, + 0xfc, 0x03, 0x19, 0x37, 0xf9, 0x11, 0x80, 0x4b, 0x1c, 0x17, 0x13, 0xcf, 0xc4, 0x34, 0x9f, 0xe1, + 0x2c, 0xb9, 0xb5, 0x02, 0x98, 0xd9, 0x1a, 0x14, 0x3b, 0x33, 0x57, 0x41, 0x8e, 0xb9, 0x58, 0xa8, + 0x0b, 0x9b, 0x96, 0x40, 0x50, 0x1f, 0x72, 0x08, 0xf3, 0xd9, 0x9d, 0xd8, 0x6e, 0xae, 0xfc, 0x9d, + 0xcb, 0xc0, 0xae, 0x6d, 0x58, 0xf3, 0xaf, 0x85, 0x0f, 0x61, 0x6b, 0x69, 0xcc, 0x0b, 0x78, 0x76, + 0x6d, 0x9e, 0x67, 0xd2, 0x1c, 0x67, 0xf6, 0xb2, 0x90, 0x16, 0xf4, 0x55, 0xfe, 0x98, 0x80, 0x6c, + 0xd7, 0x35, 0xc8, 0xc9, 0xd7, 0x07, 0xca, 0x87, 0x17, 0x40, 0xf9, 0xfe, 0x8a, 0xc5, 0x0e, 0x97, + 0xe0, 0x25, 0x46, 0xf2, 0xcf, 0x09, 0x80, 0xce, 0x74, 0x86, 0x65, 0x09, 0xae, 0x71, 0x78, 0xdc, + 0xa9, 0x77, 0xe4, 0xd8, 0x4b, 0x70, 0x6a, 0x1c, 0xe7, 0x0e, 0x57, 0x85, 0x78, 0x86, 0x70, 0xc5, + 0xe7, 0xe0, 0xda, 0x05, 0x79, 0xc9, 0x3f, 0x84, 0x73, 0xd3, 0x9d, 0x77, 0xfe, 0xef, 0x00, 0xfb, + 0xc9, 0x05, 0xc0, 0x7e, 0xb0, 0x62, 0xed, 0xcf, 0x56, 0xe4, 0x65, 0x83, 0x56, 0x79, 0x0b, 0xa4, + 0x8f, 0x7d, 0x4c, 0xa6, 0x2d, 0x93, 0x7a, 0x28, 0x0f, 0x99, 0x47, 0x3e, 0x26, 0x6c, 0xe2, 0x31, + 0xbe, 0x32, 0xe1, 0xab, 0xf2, 0x45, 0x12, 0x32, 0x0d, 0xf3, 0x14, 0x33, 0xd0, 0xdf, 0x86, 0x4d, + 0x26, 0x9e, 0x9e, 0xdf, 0xbd, 0xeb, 0x5c, 0x1e, 0x62, 0xdd, 0x04, 0x10, 0x76, 0x96, 0x49, 0x3d, + 0x3e, 0x72, 0xae, 0xbc, 0xbb, 0x62, 0xaa, 0xb3, 0x5c, 0xd8, 0x2e, 0x7f, 0x34, 0x4b, 0xac, 0x08, + 0xaf, 0x0c, 0x1d, 0xdb, 0x33, 0x6d, 0x1f, 0xeb, 0x8c, 0x27, 0x86, 0x69, 0xf9, 0x04, 0xe7, 0x13, + 0x3b, 0xb1, 0xdd, 0xac, 0xb6, 0x1d, 0xaa, 0x0e, 0xec, 0xba, 0x50, 0xa0, 0x43, 0x90, 0xe9, 0x90, + 0x98, 0xae, 0xa7, 0x9f, 0x1a, 0xc4, 0x34, 0x06, 0x16, 0x16, 0x5c, 0xc9, 0x95, 0xef, 0xac, 0x2a, + 0xb7, 0x62, 0x92, 0xc5, 0x2e, 0x77, 0x7f, 0x10, 0x7a, 0x07, 0xc7, 0x31, 0x5d, 0x94, 0xa2, 0x07, + 0x0b, 0x64, 0x49, 0xf1, 0x11, 0x6e, 0x46, 0x1c, 0xe1, 0x79, 0x4c, 0x39, 0x47, 0xf4, 0xf4, 0x39, + 0xa2, 0x17, 0xf6, 0xe0, 0xda, 0x45, 0x49, 0x5e, 0x06, 0xfd, 0xab, 0xd6, 0x05, 0x69, 0xc6, 0x17, + 0xe5, 0x2f, 0x49, 0xc8, 0xf1, 0x4d, 0xd0, 0x7d, 0x64, 0xfd, 0x8f, 0x48, 0x72, 0x7c, 0x01, 0xe8, + 0x09, 0x0e, 0xc9, 0x47, 0x51, 0x0a, 0xb3, 0x48, 0x3c, 0x22, 0xf0, 0x9f, 0x2e, 0x00, 0x2f, 0xa8, + 0x75, 0xfb, 0x12, 0xa3, 0x5c, 0x0a, 0xfc, 0x5b, 0xe7, 0xab, 0xdc, 0xf9, 0x52, 0x92, 0xbe, 0x7a, + 0x29, 0xf9, 0xff, 0x62, 0xd4, 0x3f, 0x92, 0x90, 0xee, 0x98, 0xe3, 0x97, 0xa4, 0xe2, 0xe0, 0x67, + 0x56, 0x9c, 0x55, 0xb4, 0x10, 0x73, 0x8c, 0xc8, 0xbb, 0xfe, 0x05, 0x05, 0xe7, 0x7b, 0xd1, 0x06, + 0xb8, 0x62, 0xbd, 0xb9, 0x80, 0x72, 0x99, 0xaf, 0x1b, 0xe5, 0x7a, 0xb0, 0xbe, 0xef, 0x0c, 0x3a, + 0x96, 0x31, 0xc4, 0x13, 0x6c, 0x7b, 0xac, 0x5d, 0x18, 0x5a, 0x3e, 0xf5, 0x30, 0xd1, 0x6d, 0x63, + 0x82, 0x83, 0x78, 0xb9, 0x40, 0xd6, 0x36, 0x26, 0x78, 0xde, 0xc4, 0xf7, 0xcd, 0x51, 0x10, 0x3e, + 0x34, 0xe9, 0xfb, 0xe6, 0x48, 0xf9, 0x67, 0x02, 0xa4, 0x7d, 0x67, 0xd0, 0xf5, 0x0c, 0xcf, 0xa7, + 0xa8, 0x06, 0x29, 0xea, 0x19, 0x9e, 0x08, 0xb6, 0x59, 0x2e, 0xae, 0x58, 0xbd, 0x99, 0x63, 0x91, + 0xfd, 0xc1, 0x9a, 0x70, 0x66, 0x27, 0xf5, 0x08, 0x7b, 0x86, 0x69, 0x05, 0x4d, 0xb1, 0x16, 0xbe, + 0xa2, 0x1a, 0xc8, 0xdc, 0x44, 0xa7, 0x9e, 0x41, 0x3c, 0x9d, 0xdd, 0x60, 0x83, 0xda, 0x50, 0x08, + 0x87, 0x0a, 0xaf, 0xb7, 0xc5, 0x5e, 0x78, 0xbd, 0xd5, 0x36, 0xb9, 0x4f, 0x97, 0xb9, 0x30, 0x21, + 0xba, 0x0f, 0x59, 0xea, 0x0f, 0x44, 0xa2, 0x19, 0x9e, 0xe8, 0x7b, 0xd1, 0x13, 0x0d, 0x1c, 0xb5, + 0x59, 0x08, 0xe5, 0xcb, 0x18, 0xa4, 0x78, 0xfe, 0xec, 0xfe, 0xd8, 0xed, 0x55, 0x7a, 0xea, 0xd2, + 0xfd, 0x31, 0x07, 0x99, 0x8e, 0xda, 0xae, 0x35, 0xdb, 0xf7, 0xe4, 0x18, 0xda, 0x04, 0xe8, 0xaa, + 0xbd, 0x7e, 0x47, 0xaf, 0x1d, 0xb4, 0x55, 0x39, 0xcb, 0x94, 0x5a, 0xbf, 0xdd, 0x66, 0xca, 0x38, + 0x42, 0xb0, 0x59, 0xad, 0xb4, 0xab, 0x6a, 0x4b, 0x0f, 0x1d, 0x12, 0x73, 0xb2, 0x6e, 0xaf, 0xa2, + 0xf5, 0xd4, 0x9a, 0x9c, 0x41, 0x1b, 0x20, 0x09, 0x59, 0x4b, 0xad, 0x89, 0x7b, 0x27, 0x8f, 0xb6, + 0x70, 0xef, 0x7c, 0x05, 0xb6, 0x2a, 0xbd, 0x9e, 0x7a, 0xbf, 0xd3, 0xd3, 0xeb, 0x95, 0x66, 0xab, + 0xaf, 0xa9, 0xb2, 0xa4, 0x34, 0x20, 0x1b, 0xce, 0x00, 0x6d, 0x41, 0x6e, 0x31, 0xcf, 0x0d, 0x90, + 0xba, 0xfd, 0xbd, 0xfb, 0xcd, 0x1e, 0x1b, 0x24, 0x86, 0x00, 0xd2, 0x1f, 0xf7, 0xd5, 0xbe, 0x5a, + 0x93, 0xe3, 0x48, 0x86, 0xf5, 0x6e, 0xaf, 0xd2, 0x52, 0x59, 0x0e, 0xbd, 0x7e, 0x57, 0x4e, 0x28, + 0x35, 0x4e, 0x27, 0x0d, 0x1f, 0x62, 0x82, 0xed, 0x21, 0x46, 0xaf, 0xf3, 0xcd, 0x7b, 0x8c, 0x87, + 0x9e, 0x6e, 0x8e, 0x02, 0x32, 0x49, 0x81, 0xa4, 0x39, 0x42, 0xd7, 0x21, 0x7d, 0xec, 0x0c, 0xf4, + 0x19, 0x89, 0x52, 0xc7, 0xce, 0xa0, 0x39, 0x52, 0xfe, 0x10, 0x87, 0xad, 0x4f, 0x0c, 0x62, 0x57, + 0x5c, 0xd7, 0x32, 0x87, 0xfc, 0x73, 0x07, 0x6b, 0xa3, 0xe7, 0x08, 0xc9, 0x9f, 0xd1, 0x7e, 0x48, + 0x2c, 0x71, 0xe1, 0xbf, 0xb1, 0x02, 0xaf, 0xa5, 0x90, 0x8b, 0xf4, 0x2a, 0x40, 0xd6, 0x25, 0xce, + 0x98, 0x60, 0x4a, 0x79, 0xc9, 0x8b, 0x6b, 0xb3, 0x77, 0xc6, 0x78, 0x8f, 0x18, 0xc3, 0x13, 0x56, + 0x08, 0x7c, 0x62, 0xe5, 0x93, 0x82, 0xf1, 0xa1, 0xac, 0x4f, 0x2c, 0xe5, 0xe7, 0xab, 0xe0, 0xce, + 0x40, 0xa2, 0xad, 0x3e, 0x14, 0x50, 0xb7, 0xd5, 0x87, 0x7a, 0xb7, 0xf2, 0x40, 0xa0, 0xbb, 0xb0, + 0xbe, 0x09, 0xb4, 0x0e, 0xd9, 0x4a, 0xb5, 0xaa, 0x76, 0x7a, 0x1c, 0xc3, 0x39, 0x1e, 0xa4, 0x98, + 0xaa, 0xde, 0x6c, 0x37, 0xbb, 0x0d, 0xb5, 0x26, 0xa7, 0x19, 0x10, 0x0c, 0x41, 0x8e, 0x3c, 0x40, + 0xfa, 0x87, 0x4d, 0x0e, 0x7b, 0x56, 0xf9, 0x95, 0x04, 0x09, 0x76, 0x82, 0x34, 0x41, 0x22, 0x21, + 0x0e, 0x7c, 0xd5, 0x72, 0xe5, 0x77, 0x57, 0x13, 0x7a, 0x06, 0x9d, 0x76, 0xe6, 0xcd, 0x42, 0xb9, + 0x61, 0x85, 0x08, 0xce, 0x98, 0x08, 0xa1, 0x66, 0x45, 0x45, 0x3b, 0xf3, 0x66, 0xe7, 0xd5, 0x11, + 0xbf, 0xde, 0xeb, 0xc7, 0xce, 0x80, 0x2f, 0xf4, 0xea, 0xf3, 0x6a, 0xf6, 0x3d, 0x80, 0x9d, 0x57, + 0x47, 0xb3, 0x0f, 0x24, 0x75, 0x90, 0x28, 0xeb, 0x2f, 0x78, 0xa4, 0x24, 0x8f, 0xf4, 0xad, 0x88, + 0xd7, 0xd1, 0xc6, 0x9a, 0x96, 0xa5, 0xe1, 0x8d, 0xae, 0x05, 0x39, 0x77, 0x7a, 0x16, 0x29, 0xc5, + 0x23, 0x7d, 0x3b, 0xf2, 0xfd, 0xa7, 0xb1, 0xa6, 0x41, 0xe0, 0xcf, 0xa2, 0x55, 0x21, 0xcb, 0x2f, + 0x5b, 0x2c, 0x94, 0x28, 0x42, 0x6f, 0x47, 0xeb, 0x8e, 0x1b, 0x6b, 0x5a, 0xe6, 0x28, 0xb8, 0x6f, + 0xdc, 0x85, 0x8c, 0x6b, 0x8e, 0x79, 0x0c, 0x71, 0xe2, 0xbc, 0x15, 0xe9, 0xc0, 0x6b, 0xac, 0x69, + 0x69, 0x57, 0xf4, 0x0f, 0x1d, 0xd8, 0x10, 0x53, 0xa2, 0x8f, 0x2c, 0x1e, 0x67, 0x9d, 0xc7, 0x79, + 0x27, 0x7a, 0xc3, 0xd6, 0x58, 0xd3, 0x72, 0x74, 0xae, 0xbd, 0xbd, 0x0b, 0x69, 0xca, 0xab, 0x5d, + 0x70, 0x85, 0xdb, 0x8d, 0x5a, 0x1d, 0xb5, 0xc0, 0x0f, 0x1d, 0xc0, 0xa6, 0x78, 0xd2, 0x8f, 0x4c, + 0xea, 0x39, 0x64, 0x9a, 0xdf, 0xe0, 0xa7, 0x79, 0xf4, 0x48, 0x1b, 0xc2, 0xbf, 0x21, 0xdc, 0xd1, + 0x8f, 0x61, 0x7b, 0x6a, 0x10, 0x5b, 0x37, 0xce, 0x36, 0x35, 0xcd, 0x4b, 0x3c, 0x66, 0xf1, 0x72, + 0xb5, 0x40, 0x93, 0xa7, 0x8b, 0x02, 0x8a, 0xee, 0x40, 0x21, 0xf8, 0xcc, 0xe9, 0xf8, 0x9e, 0xeb, + 0x7b, 0x3a, 0xc1, 0xd4, 0xf1, 0xc9, 0x50, 0x74, 0x63, 0xdb, 0xbc, 0x04, 0x7c, 0x43, 0x58, 0x1c, + 0x70, 0x03, 0x2d, 0xd0, 0xb3, 0xb6, 0xec, 0x7d, 0xc8, 0x07, 0xce, 0xac, 0x6f, 0x22, 0x8e, 0xc5, + 0x1b, 0x0d, 0xca, 0x5d, 0xb7, 0xb8, 0xeb, 0x75, 0xa1, 0xaf, 0x0a, 0x35, 0x6b, 0x39, 0x28, 0x73, + 0xac, 0x43, 0xda, 0x32, 0x06, 0xd8, 0xa2, 0x79, 0x14, 0x69, 0x1e, 0xac, 0xcd, 0x69, 0x71, 0x07, + 0xd1, 0xe2, 0x04, 0xde, 0xa8, 0x05, 0x40, 0x87, 0x47, 0x78, 0xe4, 0x5b, 0xa6, 0x3d, 0xce, 0x5f, + 0x8b, 0xd4, 0xb6, 0xb0, 0x75, 0x9e, 0xf9, 0x68, 0x73, 0xfe, 0x85, 0x0f, 0x20, 0x37, 0x37, 0xc8, + 0xa5, 0x7a, 0x0d, 0x80, 0xac, 0x37, 0x75, 0xf9, 0x7e, 0x50, 0xf6, 0x60, 0x63, 0x61, 0x0c, 0xf4, + 0x1e, 0x5c, 0x9f, 0x18, 0x8f, 0xc3, 0x56, 0x93, 0xea, 0x2e, 0x26, 0xfa, 0x91, 0xe3, 0x13, 0x1e, + 0x3a, 0xa5, 0xa1, 0x89, 0xf1, 0x38, 0xe8, 0x36, 0x69, 0x07, 0x93, 0x86, 0xe3, 0x13, 0xe5, 0x8b, + 0x18, 0xc8, 0x5d, 0x7f, 0x30, 0x31, 0x3d, 0x5e, 0xad, 0x1e, 0xf9, 0x98, 0x7a, 0xab, 0x8e, 0x99, + 0x57, 0x21, 0x4d, 0xf0, 0xd8, 0x74, 0x6c, 0x5e, 0x70, 0x24, 0x2d, 0x78, 0x43, 0x37, 0x20, 0xc1, + 0xb6, 0x86, 0xa8, 0x68, 0x4a, 0x84, 0xe2, 0xc8, 0xcc, 0xd9, 0x60, 0x44, 0x8c, 0xcb, 0x06, 0x13, + 0x67, 0x81, 0x14, 0x48, 0x9a, 0x23, 0xe5, 0x27, 0xb0, 0x71, 0x0f, 0xff, 0x1b, 0x92, 0x7b, 0xc6, + 0xd9, 0xf8, 0xd7, 0x38, 0x6c, 0xb1, 0x76, 0x7d, 0xdf, 0x19, 0xd0, 0x4b, 0x8f, 0x90, 0x5e, 0x18, + 0xe1, 0x35, 0x90, 0x5c, 0x63, 0x8c, 0x75, 0x6a, 0x7e, 0x2e, 0x80, 0x4b, 0x69, 0x59, 0x26, 0xe8, + 0x9a, 0x9f, 0x8b, 0x93, 0x9b, 0x29, 0x3d, 0xe7, 0x04, 0x87, 0xa9, 0x71, 0xf3, 0x1e, 0x13, 0x9c, + 0xeb, 0x13, 0x93, 0xe7, 0xfb, 0xc4, 0x63, 0xd8, 0x66, 0x13, 0x10, 0xad, 0xd9, 0xc4, 0xf0, 0x86, + 0x47, 0x98, 0xf0, 0xea, 0xba, 0x59, 0xfe, 0xc1, 0xaa, 0x06, 0x7a, 0x71, 0x82, 0x61, 0x05, 0xc0, + 0xf7, 0x45, 0x14, 0x6d, 0xeb, 0x78, 0x51, 0xc0, 0xa6, 0x78, 0x68, 0x5a, 0x1e, 0x26, 0xbc, 0x5e, + 0x4a, 0x5a, 0xf0, 0xa6, 0xdc, 0x84, 0xad, 0x25, 0xdf, 0xf0, 0xbb, 0xfd, 0x1a, 0x3b, 0x34, 0x2b, + 0xd5, 0x5e, 0xf3, 0x81, 0x1a, 0x1c, 0xca, 0x07, 0x6d, 0x3d, 0x78, 0x8f, 0x2b, 0x5f, 0xc5, 0x40, + 0xee, 0xbb, 0x23, 0xc3, 0xc3, 0x2f, 0x02, 0x64, 0xfc, 0x19, 0x40, 0x26, 0xe6, 0x80, 0x0c, 0xc9, + 0x97, 0xbc, 0x1c, 0xf9, 0xee, 0x40, 0xce, 0xe7, 0x79, 0xf1, 0x1f, 0x71, 0x82, 0xc3, 0xea, 0x7c, + 0x9b, 0x5b, 0x37, 0xb1, 0x35, 0xba, 0x6f, 0xd0, 0x13, 0x0d, 0x84, 0x39, 0x7b, 0x56, 0x08, 0xc8, + 0x67, 0x2b, 0x4b, 0x5d, 0xc7, 0xa6, 0x18, 0xdd, 0x84, 0xe4, 0xb1, 0x33, 0x10, 0x5f, 0xbf, 0xa2, + 0xe5, 0xc1, 0xed, 0xd1, 0xdb, 0xb0, 0x65, 0xe3, 0xc7, 0x9e, 0x3e, 0x47, 0x12, 0x31, 0xed, 0x0d, + 0x26, 0xee, 0x84, 0x44, 0x51, 0x3e, 0x03, 0xb9, 0x6a, 0xd8, 0x43, 0x6c, 0xfd, 0xc7, 0x76, 0xc4, + 0x67, 0x20, 0xd7, 0xb0, 0x85, 0x5f, 0x0c, 0xaa, 0x28, 0x23, 0x94, 0x7f, 0x91, 0xe5, 0x85, 0x2b, + 0xa8, 0xd5, 0x16, 0x26, 0xe8, 0xb7, 0x31, 0x90, 0x66, 0x55, 0x08, 0x95, 0x56, 0x9d, 0xaa, 0x4b, + 0xf5, 0xaa, 0x10, 0x61, 0x99, 0x95, 0xfa, 0xcf, 0xbe, 0xfa, 0xdb, 0x2f, 0xe3, 0x77, 0x95, 0x3b, + 0xb3, 0x9f, 0x0c, 0x83, 0xfc, 0x69, 0xe9, 0xc9, 0xd9, 0xdc, 0x9e, 0x96, 0x44, 0xea, 0xb4, 0xf4, + 0x44, 0x3c, 0x3c, 0xe5, 0xbf, 0x2c, 0xde, 0xa6, 0x7c, 0xc8, 0xdb, 0xb1, 0x77, 0xd0, 0x6f, 0x62, + 0x90, 0x16, 0x05, 0x09, 0xad, 0x3a, 0x00, 0x16, 0xea, 0x56, 0xa4, 0x24, 0x55, 0x9e, 0xe4, 0x47, + 0xe8, 0xc3, 0x17, 0x49, 0xb2, 0xf4, 0x44, 0x2c, 0xf6, 0x53, 0xf4, 0x65, 0x0c, 0xb2, 0x21, 0x33, + 0x51, 0xf1, 0x72, 0xc5, 0xa1, 0x50, 0x8a, 0x6c, 0x2f, 0x28, 0xaf, 0x7c, 0x9f, 0x27, 0x7d, 0x13, + 0xdd, 0x78, 0x91, 0xa4, 0xd1, 0xef, 0x62, 0x20, 0xcd, 0x4a, 0xc3, 0x4a, 0xe8, 0x97, 0x8b, 0x48, + 0xa4, 0x55, 0xdd, 0xe7, 0x09, 0xd6, 0xca, 0x57, 0x5b, 0xd5, 0xdb, 0xbc, 0x60, 0xfc, 0x3e, 0x06, + 0xd2, 0x6c, 0x03, 0xae, 0x4c, 0x77, 0x79, 0xab, 0x46, 0x4a, 0xf7, 0x80, 0xa7, 0xdb, 0x54, 0x6a, + 0x57, 0x4b, 0x77, 0xc8, 0xc7, 0x66, 0x94, 0xfd, 0x75, 0x0c, 0xa4, 0xd9, 0x96, 0x5e, 0x99, 0xf3, + 0xf2, 0xe6, 0x2f, 0xbc, 0x7a, 0xae, 0x1c, 0xaa, 0x13, 0xd7, 0x9b, 0x86, 0x64, 0x7d, 0xe7, 0x6a, + 0xcb, 0xba, 0xf7, 0x53, 0x78, 0x73, 0xe8, 0x4c, 0x9e, 0x9f, 0xd4, 0x9e, 0xc4, 0x18, 0xd7, 0x61, + 0xe3, 0x77, 0x62, 0x9f, 0xaa, 0x81, 0xed, 0xd8, 0xb1, 0x0c, 0x7b, 0x5c, 0x74, 0xc8, 0xb8, 0x34, + 0xc6, 0x36, 0xcf, 0xae, 0x24, 0x54, 0x86, 0x6b, 0xd2, 0x67, 0xfc, 0xa7, 0xc0, 0x9d, 0x50, 0x30, + 0x48, 0x73, 0x8f, 0xef, 0xfe, 0x2b, 0x00, 0x00, 0xff, 0xff, 0x8e, 0x78, 0x17, 0x92, 0x5a, 0x20, + 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/operations.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/operations.pb.go index bc8091725..18eef5f56 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/operations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/operations.pb.go @@ -51,19 +51,19 @@ func (x ClusterOperationStatus_State) String() string { return proto.EnumName(ClusterOperationStatus_State_name, int32(x)) } func (ClusterOperationStatus_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_operations_06b6b1222ad2529d, []int{0, 0} + return fileDescriptor_operations_54f6cccb1783f6c3, []int{0, 0} } // The status of the operation. type ClusterOperationStatus struct { - // Output-only. A message containing the operation state. - State ClusterOperationStatus_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.dataproc.v1beta2.ClusterOperationStatus_State" json:"state,omitempty"` - // Output-only. A message containing the detailed operation state. - InnerState string `protobuf:"bytes,2,opt,name=inner_state,json=innerState" json:"inner_state,omitempty"` - // Output-only.A message containing any operation metadata details. - Details string `protobuf:"bytes,3,opt,name=details" json:"details,omitempty"` - // Output-only. The time this state was entered. - StateStartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=state_start_time,json=stateStartTime" json:"state_start_time,omitempty"` + // Output only. A message containing the operation state. + State ClusterOperationStatus_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.dataproc.v1beta2.ClusterOperationStatus_State" json:"state,omitempty"` + // Output only. A message containing the detailed operation state. + InnerState string `protobuf:"bytes,2,opt,name=inner_state,json=innerState,proto3" json:"inner_state,omitempty"` + // Output only. A message containing any operation metadata details. + Details string `protobuf:"bytes,3,opt,name=details,proto3" json:"details,omitempty"` + // Output only. The time this state was entered. + StateStartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=state_start_time,json=stateStartTime,proto3" json:"state_start_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -73,7 +73,7 @@ func (m *ClusterOperationStatus) Reset() { *m = ClusterOperationStatus{} func (m *ClusterOperationStatus) String() string { return proto.CompactTextString(m) } func (*ClusterOperationStatus) ProtoMessage() {} func (*ClusterOperationStatus) Descriptor() ([]byte, []int) { - return fileDescriptor_operations_06b6b1222ad2529d, []int{0} + return fileDescriptor_operations_54f6cccb1783f6c3, []int{0} } func (m *ClusterOperationStatus) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterOperationStatus.Unmarshal(m, b) @@ -123,22 +123,22 @@ func (m *ClusterOperationStatus) GetStateStartTime() *timestamp.Timestamp { // Metadata describing the operation. type ClusterOperationMetadata struct { - // Output-only. Name of the cluster for the operation. - ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` - // Output-only. Cluster UUID for the operation. - ClusterUuid string `protobuf:"bytes,8,opt,name=cluster_uuid,json=clusterUuid" json:"cluster_uuid,omitempty"` - // Output-only. Current operation status. - Status *ClusterOperationStatus `protobuf:"bytes,9,opt,name=status" json:"status,omitempty"` - // Output-only. The previous operation status. - StatusHistory []*ClusterOperationStatus `protobuf:"bytes,10,rep,name=status_history,json=statusHistory" json:"status_history,omitempty"` - // Output-only. The operation type. - OperationType string `protobuf:"bytes,11,opt,name=operation_type,json=operationType" json:"operation_type,omitempty"` - // Output-only. Short description of operation. - Description string `protobuf:"bytes,12,opt,name=description" json:"description,omitempty"` - // Output-only. Labels associated with the operation - Labels map[string]string `protobuf:"bytes,13,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` - // Output-only. Errors encountered during operation execution. - Warnings []string `protobuf:"bytes,14,rep,name=warnings" json:"warnings,omitempty"` + // Output only. Name of the cluster for the operation. + ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` + // Output only. Cluster UUID for the operation. + ClusterUuid string `protobuf:"bytes,8,opt,name=cluster_uuid,json=clusterUuid,proto3" json:"cluster_uuid,omitempty"` + // Output only. Current operation status. + Status *ClusterOperationStatus `protobuf:"bytes,9,opt,name=status,proto3" json:"status,omitempty"` + // Output only. The previous operation status. + StatusHistory []*ClusterOperationStatus `protobuf:"bytes,10,rep,name=status_history,json=statusHistory,proto3" json:"status_history,omitempty"` + // Output only. The operation type. + OperationType string `protobuf:"bytes,11,opt,name=operation_type,json=operationType,proto3" json:"operation_type,omitempty"` + // Output only. Short description of operation. + Description string `protobuf:"bytes,12,opt,name=description,proto3" json:"description,omitempty"` + // Output only. Labels associated with the operation + Labels map[string]string `protobuf:"bytes,13,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + // Output only. Errors encountered during operation execution. + Warnings []string `protobuf:"bytes,14,rep,name=warnings,proto3" json:"warnings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -148,7 +148,7 @@ func (m *ClusterOperationMetadata) Reset() { *m = ClusterOperationMetada func (m *ClusterOperationMetadata) String() string { return proto.CompactTextString(m) } func (*ClusterOperationMetadata) ProtoMessage() {} func (*ClusterOperationMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_operations_06b6b1222ad2529d, []int{1} + return fileDescriptor_operations_54f6cccb1783f6c3, []int{1} } func (m *ClusterOperationMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterOperationMetadata.Unmarshal(m, b) @@ -232,10 +232,10 @@ func init() { } func init() { - proto.RegisterFile("google/cloud/dataproc/v1beta2/operations.proto", fileDescriptor_operations_06b6b1222ad2529d) + proto.RegisterFile("google/cloud/dataproc/v1beta2/operations.proto", fileDescriptor_operations_54f6cccb1783f6c3) } -var fileDescriptor_operations_06b6b1222ad2529d = []byte{ +var fileDescriptor_operations_54f6cccb1783f6c3 = []byte{ // 537 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x54, 0x5f, 0x8b, 0xd3, 0x4e, 0x14, 0xfd, 0xa5, 0xff, 0x7b, 0xb3, 0xed, 0xaf, 0x0c, 0x22, 0x43, 0x51, 0x36, 0x5b, 0x10, 0xfa, diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/shared.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/shared.pb.go new file mode 100644 index 000000000..8576e5067 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/shared.pb.go @@ -0,0 +1,38 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/dataproc/v1beta2/shared.proto + +package dataproc // import "google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +func init() { + proto.RegisterFile("google/cloud/dataproc/v1beta2/shared.proto", fileDescriptor_shared_e1c029f42217b36b) +} + +var fileDescriptor_shared_e1c029f42217b36b = []byte{ + // 157 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x7c, 0x8e, 0xa1, 0x0e, 0x02, 0x31, + 0x0c, 0x40, 0x83, 0x41, 0x1c, 0x0e, 0x49, 0x40, 0x60, 0x11, 0x6d, 0x00, 0x89, 0xbb, 0x04, 0x7f, + 0x09, 0x0e, 0xd7, 0xdb, 0x96, 0xb2, 0x64, 0xac, 0xcb, 0x36, 0x48, 0xf8, 0x7b, 0xc2, 0xb6, 0x93, + 0x9c, 0x6d, 0xdf, 0x7b, 0x6d, 0x77, 0x60, 0x11, 0x76, 0x06, 0x95, 0x93, 0x97, 0x46, 0x4d, 0x99, + 0x42, 0x14, 0x85, 0xef, 0xe3, 0x68, 0x32, 0x9d, 0x30, 0x3d, 0x28, 0x1a, 0x0d, 0x21, 0x4a, 0x96, + 0xf5, 0xae, 0xb2, 0x50, 0x58, 0x98, 0x58, 0x68, 0xec, 0x66, 0xdb, 0x52, 0x14, 0x2c, 0x92, 0xf7, + 0x92, 0x29, 0x5b, 0xf1, 0xa9, 0xca, 0xfd, 0xa7, 0xdb, 0x2b, 0x79, 0xc2, 0x6c, 0xa2, 0x5f, 0xdd, + 0xca, 0xbd, 0xe1, 0x67, 0x0c, 0x8b, 0xfb, 0xb5, 0xd1, 0x2c, 0x8e, 0x3c, 0x83, 0x44, 0x46, 0x36, + 0xbe, 0xf4, 0xb0, 0xae, 0x28, 0xd8, 0xf4, 0xe7, 0xf7, 0xcb, 0x34, 0x18, 0x97, 0xc5, 0x38, 0x7f, + 0x03, 0x00, 0x00, 0xff, 0xff, 0x63, 0xea, 0xb3, 0xa5, 0xec, 0x00, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/workflow_templates.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/workflow_templates.pb.go index c101a22d7..b345e58fc 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/workflow_templates.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dataproc/v1beta2/workflow_templates.pb.go @@ -58,12 +58,14 @@ func (x WorkflowMetadata_State) String() string { return proto.EnumName(WorkflowMetadata_State_name, int32(x)) } func (WorkflowMetadata_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{5, 0} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{5, 0} } +// The workflow node state. type WorkflowNode_NodeState int32 const ( + // State is unspecified. WorkflowNode_NODE_STATUS_UNSPECIFIED WorkflowNode_NodeState = 0 // The node is awaiting prerequisite node to finish. WorkflowNode_BLOCKED WorkflowNode_NodeState = 1 @@ -99,17 +101,21 @@ func (x WorkflowNode_NodeState) String() string { return proto.EnumName(WorkflowNode_NodeState_name, int32(x)) } func (WorkflowNode_NodeState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{8, 0} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{8, 0} } // A Cloud Dataproc workflow template resource. type WorkflowTemplate struct { // Required. The template id. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + // + // The id must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). Cannot begin or end with underscore + // or hyphen. Must consist of between 3 and 50 characters. + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // Output only. The "resource name" of the template, as described // in https://cloud.google.com/apis/design/resource_names of the form // `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. Used to perform a consistent read-modify-write. // // This field should be left blank for a `CreateWorkflowTemplate` request. It @@ -119,11 +125,11 @@ type WorkflowTemplate struct { // the current template with the `version` field filled in with the // current server version. The user updates other fields in the template, // then returns it as part of the `UpdateWorkflowTemplate` request. - Version int32 `protobuf:"varint,3,opt,name=version" json:"version,omitempty"` + Version int32 `protobuf:"varint,3,opt,name=version,proto3" json:"version,omitempty"` // Output only. The time template was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The time template was last updated. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // Optional. The labels to associate with this template. These labels // will be propagated to all jobs and clusters created by the workflow // instance. @@ -136,11 +142,11 @@ type WorkflowTemplate struct { // [RFC 1035](https://www.ietf.org/rfc/rfc1035.txt). // // No more than 32 labels can be associated with a template. - Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,6,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Required. WorkflowTemplate scheduling information. - Placement *WorkflowTemplatePlacement `protobuf:"bytes,7,opt,name=placement" json:"placement,omitempty"` + Placement *WorkflowTemplatePlacement `protobuf:"bytes,7,opt,name=placement,proto3" json:"placement,omitempty"` // Required. The Directed Acyclic Graph of Jobs to submit. - Jobs []*OrderedJob `protobuf:"bytes,8,rep,name=jobs" json:"jobs,omitempty"` + Jobs []*OrderedJob `protobuf:"bytes,8,rep,name=jobs,proto3" json:"jobs,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -150,7 +156,7 @@ func (m *WorkflowTemplate) Reset() { *m = WorkflowTemplate{} } func (m *WorkflowTemplate) String() string { return proto.CompactTextString(m) } func (*WorkflowTemplate) ProtoMessage() {} func (*WorkflowTemplate) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{0} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{0} } func (m *WorkflowTemplate) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WorkflowTemplate.Unmarshal(m, b) @@ -230,6 +236,9 @@ func (m *WorkflowTemplate) GetJobs() []*OrderedJob { // // Either `managed_cluster` or `cluster_selector` is required. type WorkflowTemplatePlacement struct { + // Required. Specifies where workflow executes; either on a managed + // cluster or an existing cluster chosen by labels. + // // Types that are valid to be assigned to Placement: // *WorkflowTemplatePlacement_ManagedCluster // *WorkflowTemplatePlacement_ClusterSelector @@ -243,7 +252,7 @@ func (m *WorkflowTemplatePlacement) Reset() { *m = WorkflowTemplatePlace func (m *WorkflowTemplatePlacement) String() string { return proto.CompactTextString(m) } func (*WorkflowTemplatePlacement) ProtoMessage() {} func (*WorkflowTemplatePlacement) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{1} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{1} } func (m *WorkflowTemplatePlacement) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WorkflowTemplatePlacement.Unmarshal(m, b) @@ -268,10 +277,10 @@ type isWorkflowTemplatePlacement_Placement interface { } type WorkflowTemplatePlacement_ManagedCluster struct { - ManagedCluster *ManagedCluster `protobuf:"bytes,1,opt,name=managed_cluster,json=managedCluster,oneof"` + ManagedCluster *ManagedCluster `protobuf:"bytes,1,opt,name=managed_cluster,json=managedCluster,proto3,oneof"` } type WorkflowTemplatePlacement_ClusterSelector struct { - ClusterSelector *ClusterSelector `protobuf:"bytes,2,opt,name=cluster_selector,json=clusterSelector,oneof"` + ClusterSelector *ClusterSelector `protobuf:"bytes,2,opt,name=cluster_selector,json=clusterSelector,proto3,oneof"` } func (*WorkflowTemplatePlacement_ManagedCluster) isWorkflowTemplatePlacement_Placement() {} @@ -374,11 +383,15 @@ func _WorkflowTemplatePlacement_OneofSizer(msg proto.Message) (n int) { // Cluster that is managed by the workflow. type ManagedCluster struct { - // Required. The cluster name. Cluster names within a project must be - // unique. Names from deleted clusters can be reused. - ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` + // Required. The cluster name prefix. A unique cluster name will be formed by + // appending a random suffix. + // + // The name must contain only lower-case letters (a-z), numbers (0-9), + // and hyphens (-). Must begin with a letter. Cannot begin or end with + // hyphen. Must consist of between 2 and 35 characters. + ClusterName string `protobuf:"bytes,2,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` // Required. The cluster configuration. - Config *ClusterConfig `protobuf:"bytes,3,opt,name=config" json:"config,omitempty"` + Config *ClusterConfig `protobuf:"bytes,3,opt,name=config,proto3" json:"config,omitempty"` // Optional. The labels to associate with this cluster. // // Label keys must be between 1 and 63 characters long, and must conform to @@ -388,8 +401,8 @@ type ManagedCluster struct { // Label values must be between 1 and 63 characters long, and must conform to // the following PCRE regular expression: [\p{Ll}\p{Lo}\p{N}_-]{0,63} // - // No more than 64 labels can be associated with a given cluster. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + // No more than 32 labels can be associated with a given cluster. + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -399,7 +412,7 @@ func (m *ManagedCluster) Reset() { *m = ManagedCluster{} } func (m *ManagedCluster) String() string { return proto.CompactTextString(m) } func (*ManagedCluster) ProtoMessage() {} func (*ManagedCluster) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{2} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{2} } func (m *ManagedCluster) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ManagedCluster.Unmarshal(m, b) @@ -447,10 +460,10 @@ type ClusterSelector struct { // // If unspecified, the zone of the first cluster matching the selector // is used. - Zone string `protobuf:"bytes,1,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,1,opt,name=zone,proto3" json:"zone,omitempty"` // Required. The cluster labels. Cluster must have all labels // to match. - ClusterLabels map[string]string `protobuf:"bytes,2,rep,name=cluster_labels,json=clusterLabels" json:"cluster_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ClusterLabels map[string]string `protobuf:"bytes,2,rep,name=cluster_labels,json=clusterLabels,proto3" json:"cluster_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -460,7 +473,7 @@ func (m *ClusterSelector) Reset() { *m = ClusterSelector{} } func (m *ClusterSelector) String() string { return proto.CompactTextString(m) } func (*ClusterSelector) ProtoMessage() {} func (*ClusterSelector) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{3} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{3} } func (m *ClusterSelector) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterSelector.Unmarshal(m, b) @@ -494,13 +507,20 @@ func (m *ClusterSelector) GetClusterLabels() map[string]string { return nil } +// A job executed by the workflow. type OrderedJob struct { // Required. The step id. The id must be unique among all jobs // within the template. // - // The step id is used as prefix for job id, as job `workflow-step-id` label, - // and in prerequisite_step_ids field from other steps. - StepId string `protobuf:"bytes,1,opt,name=step_id,json=stepId" json:"step_id,omitempty"` + // The step id is used as prefix for job id, as job + // `goog-dataproc-workflow-step-id` label, and in + // [prerequisiteStepIds][google.cloud.dataproc.v1beta2.OrderedJob.prerequisite_step_ids] field from other + // steps. + // + // The id must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). Cannot begin or end with underscore + // or hyphen. Must consist of between 3 and 50 characters. + StepId string `protobuf:"bytes,1,opt,name=step_id,json=stepId,proto3" json:"step_id,omitempty"` // Required. The job definition. // // Types that are valid to be assigned to JobType: @@ -520,13 +540,13 @@ type OrderedJob struct { // Label values must be between 1 and 63 characters long, and must conform to // the following regular expression: [\p{Ll}\p{Lo}\p{N}_-]{0,63} // - // No more than 64 labels can be associated with a given job. - Labels map[string]string `protobuf:"bytes,8,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + // No more than 32 labels can be associated with a given job. + Labels map[string]string `protobuf:"bytes,8,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. Job scheduling configuration. - Scheduling *JobScheduling `protobuf:"bytes,9,opt,name=scheduling" json:"scheduling,omitempty"` + Scheduling *JobScheduling `protobuf:"bytes,9,opt,name=scheduling,proto3" json:"scheduling,omitempty"` // Optional. The optional list of prerequisite job step_ids. // If not specified, the job will start at the beginning of workflow. - PrerequisiteStepIds []string `protobuf:"bytes,10,rep,name=prerequisite_step_ids,json=prerequisiteStepIds" json:"prerequisite_step_ids,omitempty"` + PrerequisiteStepIds []string `protobuf:"bytes,10,rep,name=prerequisite_step_ids,json=prerequisiteStepIds,proto3" json:"prerequisite_step_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -536,7 +556,7 @@ func (m *OrderedJob) Reset() { *m = OrderedJob{} } func (m *OrderedJob) String() string { return proto.CompactTextString(m) } func (*OrderedJob) ProtoMessage() {} func (*OrderedJob) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{4} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{4} } func (m *OrderedJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_OrderedJob.Unmarshal(m, b) @@ -561,22 +581,22 @@ type isOrderedJob_JobType interface { } type OrderedJob_HadoopJob struct { - HadoopJob *HadoopJob `protobuf:"bytes,2,opt,name=hadoop_job,json=hadoopJob,oneof"` + HadoopJob *HadoopJob `protobuf:"bytes,2,opt,name=hadoop_job,json=hadoopJob,proto3,oneof"` } type OrderedJob_SparkJob struct { - SparkJob *SparkJob `protobuf:"bytes,3,opt,name=spark_job,json=sparkJob,oneof"` + SparkJob *SparkJob `protobuf:"bytes,3,opt,name=spark_job,json=sparkJob,proto3,oneof"` } type OrderedJob_PysparkJob struct { - PysparkJob *PySparkJob `protobuf:"bytes,4,opt,name=pyspark_job,json=pysparkJob,oneof"` + PysparkJob *PySparkJob `protobuf:"bytes,4,opt,name=pyspark_job,json=pysparkJob,proto3,oneof"` } type OrderedJob_HiveJob struct { - HiveJob *HiveJob `protobuf:"bytes,5,opt,name=hive_job,json=hiveJob,oneof"` + HiveJob *HiveJob `protobuf:"bytes,5,opt,name=hive_job,json=hiveJob,proto3,oneof"` } type OrderedJob_PigJob struct { - PigJob *PigJob `protobuf:"bytes,6,opt,name=pig_job,json=pigJob,oneof"` + PigJob *PigJob `protobuf:"bytes,6,opt,name=pig_job,json=pigJob,proto3,oneof"` } type OrderedJob_SparkSqlJob struct { - SparkSqlJob *SparkSqlJob `protobuf:"bytes,7,opt,name=spark_sql_job,json=sparkSqlJob,oneof"` + SparkSqlJob *SparkSqlJob `protobuf:"bytes,7,opt,name=spark_sql_job,json=sparkSqlJob,proto3,oneof"` } func (*OrderedJob_HadoopJob) isOrderedJob_JobType() {} @@ -816,30 +836,32 @@ func _OrderedJob_OneofSizer(msg proto.Message) (n int) { // A Cloud Dataproc workflow template resource. type WorkflowMetadata struct { // Output only. The "resource name" of the template. - Template string `protobuf:"bytes,1,opt,name=template" json:"template,omitempty"` + Template string `protobuf:"bytes,1,opt,name=template,proto3" json:"template,omitempty"` // Output only. The version of template at the time of // workflow instantiation. - Version int32 `protobuf:"varint,2,opt,name=version" json:"version,omitempty"` + Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"` // Output only. The create cluster operation metadata. - CreateCluster *ClusterOperation `protobuf:"bytes,3,opt,name=create_cluster,json=createCluster" json:"create_cluster,omitempty"` + CreateCluster *ClusterOperation `protobuf:"bytes,3,opt,name=create_cluster,json=createCluster,proto3" json:"create_cluster,omitempty"` // Output only. The workflow graph. - Graph *WorkflowGraph `protobuf:"bytes,4,opt,name=graph" json:"graph,omitempty"` + Graph *WorkflowGraph `protobuf:"bytes,4,opt,name=graph,proto3" json:"graph,omitempty"` // Output only. The delete cluster operation metadata. - DeleteCluster *ClusterOperation `protobuf:"bytes,5,opt,name=delete_cluster,json=deleteCluster" json:"delete_cluster,omitempty"` + DeleteCluster *ClusterOperation `protobuf:"bytes,5,opt,name=delete_cluster,json=deleteCluster,proto3" json:"delete_cluster,omitempty"` // Output only. The workflow state. - State WorkflowMetadata_State `protobuf:"varint,6,opt,name=state,enum=google.cloud.dataproc.v1beta2.WorkflowMetadata_State" json:"state,omitempty"` + State WorkflowMetadata_State `protobuf:"varint,6,opt,name=state,proto3,enum=google.cloud.dataproc.v1beta2.WorkflowMetadata_State" json:"state,omitempty"` // Output only. The name of the managed cluster. - ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName" json:"cluster_name,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + ClusterName string `protobuf:"bytes,7,opt,name=cluster_name,json=clusterName,proto3" json:"cluster_name,omitempty"` + // Map from parameter names to values that were used for those parameters. + Parameters map[string]string `protobuf:"bytes,8,rep,name=parameters,proto3" json:"parameters,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *WorkflowMetadata) Reset() { *m = WorkflowMetadata{} } func (m *WorkflowMetadata) String() string { return proto.CompactTextString(m) } func (*WorkflowMetadata) ProtoMessage() {} func (*WorkflowMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{5} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{5} } func (m *WorkflowMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WorkflowMetadata.Unmarshal(m, b) @@ -908,13 +930,21 @@ func (m *WorkflowMetadata) GetClusterName() string { return "" } +func (m *WorkflowMetadata) GetParameters() map[string]string { + if m != nil { + return m.Parameters + } + return nil +} + +// The cluster operation triggered by a workflow. type ClusterOperation struct { // Output only. The id of the cluster operation. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // Output only. Error, if operation failed. - Error string `protobuf:"bytes,2,opt,name=error" json:"error,omitempty"` + Error string `protobuf:"bytes,2,opt,name=error,proto3" json:"error,omitempty"` // Output only. Indicates the operation is done. - Done bool `protobuf:"varint,3,opt,name=done" json:"done,omitempty"` + Done bool `protobuf:"varint,3,opt,name=done,proto3" json:"done,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -924,7 +954,7 @@ func (m *ClusterOperation) Reset() { *m = ClusterOperation{} } func (m *ClusterOperation) String() string { return proto.CompactTextString(m) } func (*ClusterOperation) ProtoMessage() {} func (*ClusterOperation) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{6} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{6} } func (m *ClusterOperation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClusterOperation.Unmarshal(m, b) @@ -968,7 +998,7 @@ func (m *ClusterOperation) GetDone() bool { // The workflow graph. type WorkflowGraph struct { // Output only. The workflow nodes. - Nodes []*WorkflowNode `protobuf:"bytes,1,rep,name=nodes" json:"nodes,omitempty"` + Nodes []*WorkflowNode `protobuf:"bytes,1,rep,name=nodes,proto3" json:"nodes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -978,7 +1008,7 @@ func (m *WorkflowGraph) Reset() { *m = WorkflowGraph{} } func (m *WorkflowGraph) String() string { return proto.CompactTextString(m) } func (*WorkflowGraph) ProtoMessage() {} func (*WorkflowGraph) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{7} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{7} } func (m *WorkflowGraph) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WorkflowGraph.Unmarshal(m, b) @@ -1008,15 +1038,15 @@ func (m *WorkflowGraph) GetNodes() []*WorkflowNode { // The workflow node. type WorkflowNode struct { // Output only. The name of the node. - StepId string `protobuf:"bytes,1,opt,name=step_id,json=stepId" json:"step_id,omitempty"` + StepId string `protobuf:"bytes,1,opt,name=step_id,json=stepId,proto3" json:"step_id,omitempty"` // Output only. Node's prerequisite nodes. - PrerequisiteStepIds []string `protobuf:"bytes,2,rep,name=prerequisite_step_ids,json=prerequisiteStepIds" json:"prerequisite_step_ids,omitempty"` + PrerequisiteStepIds []string `protobuf:"bytes,2,rep,name=prerequisite_step_ids,json=prerequisiteStepIds,proto3" json:"prerequisite_step_ids,omitempty"` // Output only. The job id; populated after the node enters RUNNING state. - JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,3,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` // Output only. The node state. - State WorkflowNode_NodeState `protobuf:"varint,5,opt,name=state,enum=google.cloud.dataproc.v1beta2.WorkflowNode_NodeState" json:"state,omitempty"` + State WorkflowNode_NodeState `protobuf:"varint,5,opt,name=state,proto3,enum=google.cloud.dataproc.v1beta2.WorkflowNode_NodeState" json:"state,omitempty"` // Output only. The error detail. - Error string `protobuf:"bytes,6,opt,name=error" json:"error,omitempty"` + Error string `protobuf:"bytes,6,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1026,7 +1056,7 @@ func (m *WorkflowNode) Reset() { *m = WorkflowNode{} } func (m *WorkflowNode) String() string { return proto.CompactTextString(m) } func (*WorkflowNode) ProtoMessage() {} func (*WorkflowNode) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{8} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{8} } func (m *WorkflowNode) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WorkflowNode.Unmarshal(m, b) @@ -1086,9 +1116,9 @@ type CreateWorkflowTemplateRequest struct { // Required. The "resource name" of the region, as described // in https://cloud.google.com/apis/design/resource_names of the form // `projects/{project_id}/regions/{region}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The Dataproc workflow template to create. - Template *WorkflowTemplate `protobuf:"bytes,2,opt,name=template" json:"template,omitempty"` + Template *WorkflowTemplate `protobuf:"bytes,2,opt,name=template,proto3" json:"template,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1098,7 +1128,7 @@ func (m *CreateWorkflowTemplateRequest) Reset() { *m = CreateWorkflowTem func (m *CreateWorkflowTemplateRequest) String() string { return proto.CompactTextString(m) } func (*CreateWorkflowTemplateRequest) ProtoMessage() {} func (*CreateWorkflowTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{9} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{9} } func (m *CreateWorkflowTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateWorkflowTemplateRequest.Unmarshal(m, b) @@ -1137,12 +1167,12 @@ type GetWorkflowTemplateRequest struct { // Required. The "resource name" of the workflow template, as described // in https://cloud.google.com/apis/design/resource_names of the form // `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The version of workflow template to retrieve. Only previously // instatiated versions can be retrieved. // // If unspecified, retrieves the current version. - Version int32 `protobuf:"varint,2,opt,name=version" json:"version,omitempty"` + Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1152,7 +1182,7 @@ func (m *GetWorkflowTemplateRequest) Reset() { *m = GetWorkflowTemplateR func (m *GetWorkflowTemplateRequest) String() string { return proto.CompactTextString(m) } func (*GetWorkflowTemplateRequest) ProtoMessage() {} func (*GetWorkflowTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{10} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{10} } func (m *GetWorkflowTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetWorkflowTemplateRequest.Unmarshal(m, b) @@ -1191,14 +1221,14 @@ type InstantiateWorkflowTemplateRequest struct { // Required. The "resource name" of the workflow template, as described // in https://cloud.google.com/apis/design/resource_names of the form // `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The version of workflow template to instantiate. If specified, // the workflow will be instantiated only if the current version of // the workflow template has the supplied version. // // This option cannot be used to instantiate a previous version of // workflow template. - Version int32 `protobuf:"varint,2,opt,name=version" json:"version,omitempty"` + Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"` // Optional. A tag that prevents multiple concurrent workflow // instances with the same tag from running. This mitigates risk of // concurrent instances started due to retries. @@ -1208,7 +1238,7 @@ type InstantiateWorkflowTemplateRequest struct { // // The tag must contain only letters (a-z, A-Z), numbers (0-9), // underscores (_), and hyphens (-). The maximum length is 40 characters. - InstanceId string `protobuf:"bytes,3,opt,name=instance_id,json=instanceId" json:"instance_id,omitempty"` + InstanceId string `protobuf:"bytes,3,opt,name=instance_id,json=instanceId,proto3" json:"instance_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1218,7 +1248,7 @@ func (m *InstantiateWorkflowTemplateRequest) Reset() { *m = InstantiateW func (m *InstantiateWorkflowTemplateRequest) String() string { return proto.CompactTextString(m) } func (*InstantiateWorkflowTemplateRequest) ProtoMessage() {} func (*InstantiateWorkflowTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{11} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{11} } func (m *InstantiateWorkflowTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InstantiateWorkflowTemplateRequest.Unmarshal(m, b) @@ -1259,12 +1289,82 @@ func (m *InstantiateWorkflowTemplateRequest) GetInstanceId() string { return "" } +// A request to instantiate an inline workflow template. +type InstantiateInlineWorkflowTemplateRequest struct { + // Required. The "resource name" of the workflow template region, as described + // in https://cloud.google.com/apis/design/resource_names of the form + // `projects/{project_id}/regions/{region}` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Required. The workflow template to instantiate. + Template *WorkflowTemplate `protobuf:"bytes,2,opt,name=template,proto3" json:"template,omitempty"` + // Optional. A tag that prevents multiple concurrent workflow + // instances with the same tag from running. This mitigates risk of + // concurrent instances started due to retries. + // + // It is recommended to always set this value to a + // [UUID](https://en.wikipedia.org/wiki/Universally_unique_identifier). + // + // The tag must contain only letters (a-z, A-Z), numbers (0-9), + // underscores (_), and hyphens (-). The maximum length is 40 characters. + InstanceId string `protobuf:"bytes,3,opt,name=instance_id,json=instanceId,proto3" json:"instance_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *InstantiateInlineWorkflowTemplateRequest) Reset() { + *m = InstantiateInlineWorkflowTemplateRequest{} +} +func (m *InstantiateInlineWorkflowTemplateRequest) String() string { return proto.CompactTextString(m) } +func (*InstantiateInlineWorkflowTemplateRequest) ProtoMessage() {} +func (*InstantiateInlineWorkflowTemplateRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{12} +} +func (m *InstantiateInlineWorkflowTemplateRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_InstantiateInlineWorkflowTemplateRequest.Unmarshal(m, b) +} +func (m *InstantiateInlineWorkflowTemplateRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_InstantiateInlineWorkflowTemplateRequest.Marshal(b, m, deterministic) +} +func (dst *InstantiateInlineWorkflowTemplateRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_InstantiateInlineWorkflowTemplateRequest.Merge(dst, src) +} +func (m *InstantiateInlineWorkflowTemplateRequest) XXX_Size() int { + return xxx_messageInfo_InstantiateInlineWorkflowTemplateRequest.Size(m) +} +func (m *InstantiateInlineWorkflowTemplateRequest) XXX_DiscardUnknown() { + xxx_messageInfo_InstantiateInlineWorkflowTemplateRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_InstantiateInlineWorkflowTemplateRequest proto.InternalMessageInfo + +func (m *InstantiateInlineWorkflowTemplateRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *InstantiateInlineWorkflowTemplateRequest) GetTemplate() *WorkflowTemplate { + if m != nil { + return m.Template + } + return nil +} + +func (m *InstantiateInlineWorkflowTemplateRequest) GetInstanceId() string { + if m != nil { + return m.InstanceId + } + return "" +} + // A request to update a workflow template. type UpdateWorkflowTemplateRequest struct { // Required. The updated workflow template. // // The `template.version` field must match the current version. - Template *WorkflowTemplate `protobuf:"bytes,1,opt,name=template" json:"template,omitempty"` + Template *WorkflowTemplate `protobuf:"bytes,1,opt,name=template,proto3" json:"template,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1274,7 +1374,7 @@ func (m *UpdateWorkflowTemplateRequest) Reset() { *m = UpdateWorkflowTem func (m *UpdateWorkflowTemplateRequest) String() string { return proto.CompactTextString(m) } func (*UpdateWorkflowTemplateRequest) ProtoMessage() {} func (*UpdateWorkflowTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{12} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{13} } func (m *UpdateWorkflowTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateWorkflowTemplateRequest.Unmarshal(m, b) @@ -1306,12 +1406,12 @@ type ListWorkflowTemplatesRequest struct { // Required. The "resource name" of the region, as described // in https://cloud.google.com/apis/design/resource_names of the form // `projects/{project_id}/regions/{region}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of results to return in each response. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The page token, returned by a previous call, to request the // next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1321,7 +1421,7 @@ func (m *ListWorkflowTemplatesRequest) Reset() { *m = ListWorkflowTempla func (m *ListWorkflowTemplatesRequest) String() string { return proto.CompactTextString(m) } func (*ListWorkflowTemplatesRequest) ProtoMessage() {} func (*ListWorkflowTemplatesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{13} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{14} } func (m *ListWorkflowTemplatesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListWorkflowTemplatesRequest.Unmarshal(m, b) @@ -1365,11 +1465,11 @@ func (m *ListWorkflowTemplatesRequest) GetPageToken() string { // A response to a request to list workflow templates in a project. type ListWorkflowTemplatesResponse struct { // Output only. WorkflowTemplates list. - Templates []*WorkflowTemplate `protobuf:"bytes,1,rep,name=templates" json:"templates,omitempty"` + Templates []*WorkflowTemplate `protobuf:"bytes,1,rep,name=templates,proto3" json:"templates,omitempty"` // Output only. This token is included in the response if there are more results // to fetch. To fetch additional results, provide this value as the // page_token in a subsequent ListWorkflowTemplatesRequest. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1379,7 +1479,7 @@ func (m *ListWorkflowTemplatesResponse) Reset() { *m = ListWorkflowTempl func (m *ListWorkflowTemplatesResponse) String() string { return proto.CompactTextString(m) } func (*ListWorkflowTemplatesResponse) ProtoMessage() {} func (*ListWorkflowTemplatesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{14} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{15} } func (m *ListWorkflowTemplatesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListWorkflowTemplatesResponse.Unmarshal(m, b) @@ -1420,11 +1520,11 @@ type DeleteWorkflowTemplateRequest struct { // Required. The "resource name" of the workflow template, as described // in https://cloud.google.com/apis/design/resource_names of the form // `projects/{project_id}/regions/{region}/workflowTemplates/{template_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The version of workflow template to delete. If specified, // will only delete the template if the current server version matches // specified version. - Version int32 `protobuf:"varint,2,opt,name=version" json:"version,omitempty"` + Version int32 `protobuf:"varint,2,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1434,7 +1534,7 @@ func (m *DeleteWorkflowTemplateRequest) Reset() { *m = DeleteWorkflowTem func (m *DeleteWorkflowTemplateRequest) String() string { return proto.CompactTextString(m) } func (*DeleteWorkflowTemplateRequest) ProtoMessage() {} func (*DeleteWorkflowTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_workflow_templates_7ee424539012b1eb, []int{15} + return fileDescriptor_workflow_templates_fe5253a1eb957933, []int{16} } func (m *DeleteWorkflowTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteWorkflowTemplateRequest.Unmarshal(m, b) @@ -1479,12 +1579,14 @@ func init() { proto.RegisterType((*OrderedJob)(nil), "google.cloud.dataproc.v1beta2.OrderedJob") proto.RegisterMapType((map[string]string)(nil), "google.cloud.dataproc.v1beta2.OrderedJob.LabelsEntry") proto.RegisterType((*WorkflowMetadata)(nil), "google.cloud.dataproc.v1beta2.WorkflowMetadata") + proto.RegisterMapType((map[string]string)(nil), "google.cloud.dataproc.v1beta2.WorkflowMetadata.ParametersEntry") proto.RegisterType((*ClusterOperation)(nil), "google.cloud.dataproc.v1beta2.ClusterOperation") proto.RegisterType((*WorkflowGraph)(nil), "google.cloud.dataproc.v1beta2.WorkflowGraph") proto.RegisterType((*WorkflowNode)(nil), "google.cloud.dataproc.v1beta2.WorkflowNode") proto.RegisterType((*CreateWorkflowTemplateRequest)(nil), "google.cloud.dataproc.v1beta2.CreateWorkflowTemplateRequest") proto.RegisterType((*GetWorkflowTemplateRequest)(nil), "google.cloud.dataproc.v1beta2.GetWorkflowTemplateRequest") proto.RegisterType((*InstantiateWorkflowTemplateRequest)(nil), "google.cloud.dataproc.v1beta2.InstantiateWorkflowTemplateRequest") + proto.RegisterType((*InstantiateInlineWorkflowTemplateRequest)(nil), "google.cloud.dataproc.v1beta2.InstantiateInlineWorkflowTemplateRequest") proto.RegisterType((*UpdateWorkflowTemplateRequest)(nil), "google.cloud.dataproc.v1beta2.UpdateWorkflowTemplateRequest") proto.RegisterType((*ListWorkflowTemplatesRequest)(nil), "google.cloud.dataproc.v1beta2.ListWorkflowTemplatesRequest") proto.RegisterType((*ListWorkflowTemplatesResponse)(nil), "google.cloud.dataproc.v1beta2.ListWorkflowTemplatesResponse") @@ -1516,18 +1618,44 @@ type WorkflowTemplateServiceClient interface { // // The returned Operation can be used to track execution of // workflow by polling - // [google.cloud.dataproc.v1beta2.OperationService.GetOperation][]. + // [operations.get][google.longrunning.Operations.GetOperation]. // The Operation will complete when entire workflow is finished. // // The running workflow can be aborted via - // [google.cloud.dataproc.v1beta2.OperationService.CancelOperation][]. + // [operations.cancel][google.longrunning.Operations.CancelOperation]. + // This will cause any inflight jobs to be cancelled and workflow-owned + // clusters to be deleted. // - // The [google.cloud.dataproc.v1beta2.Operation.metadata][] will always be - // [google.cloud.dataproc.v1beta2.WorkflowMetadata][google.cloud.dataproc.v1beta2.WorkflowMetadata]. + // The [Operation.metadata][google.longrunning.Operation.metadata] will be + // [WorkflowMetadata][google.cloud.dataproc.v1beta2.WorkflowMetadata]. // - // The [google.cloud.dataproc.v1beta2.Operation.result][] will always be - // [google.protobuf.Empty][google.protobuf.Empty]. + // On successful completion, + // [Operation.response][google.longrunning.Operation.response] will be + // [Empty][google.protobuf.Empty]. InstantiateWorkflowTemplate(ctx context.Context, in *InstantiateWorkflowTemplateRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) + // Instantiates a template and begins execution. + // + // This method is equivalent to executing the sequence + // [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate], + // [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate]. + // + // The returned Operation can be used to track execution of + // workflow by polling + // [operations.get][google.longrunning.Operations.GetOperation]. + // The Operation will complete when entire workflow is finished. + // + // The running workflow can be aborted via + // [operations.cancel][google.longrunning.Operations.CancelOperation]. + // This will cause any inflight jobs to be cancelled and workflow-owned + // clusters to be deleted. + // + // The [Operation.metadata][google.longrunning.Operation.metadata] will be + // [WorkflowMetadata][google.cloud.dataproc.v1beta2.WorkflowMetadata]. + // + // On successful completion, + // [Operation.response][google.longrunning.Operation.response] will be + // [Empty][google.protobuf.Empty]. + InstantiateInlineWorkflowTemplate(ctx context.Context, in *InstantiateInlineWorkflowTemplateRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) // Updates (replaces) workflow template. The updated template // must contain version that matches the current server version. UpdateWorkflowTemplate(ctx context.Context, in *UpdateWorkflowTemplateRequest, opts ...grpc.CallOption) (*WorkflowTemplate, error) @@ -1572,6 +1700,15 @@ func (c *workflowTemplateServiceClient) InstantiateWorkflowTemplate(ctx context. return out, nil } +func (c *workflowTemplateServiceClient) InstantiateInlineWorkflowTemplate(ctx context.Context, in *InstantiateInlineWorkflowTemplateRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) { + out := new(longrunning.Operation) + err := c.cc.Invoke(ctx, "/google.cloud.dataproc.v1beta2.WorkflowTemplateService/InstantiateInlineWorkflowTemplate", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + func (c *workflowTemplateServiceClient) UpdateWorkflowTemplate(ctx context.Context, in *UpdateWorkflowTemplateRequest, opts ...grpc.CallOption) (*WorkflowTemplate, error) { out := new(WorkflowTemplate) err := c.cc.Invoke(ctx, "/google.cloud.dataproc.v1beta2.WorkflowTemplateService/UpdateWorkflowTemplate", in, out, opts...) @@ -1599,8 +1736,7 @@ func (c *workflowTemplateServiceClient) DeleteWorkflowTemplate(ctx context.Conte return out, nil } -// Server API for WorkflowTemplateService service - +// WorkflowTemplateServiceServer is the server API for WorkflowTemplateService service. type WorkflowTemplateServiceServer interface { // Creates new workflow template. CreateWorkflowTemplate(context.Context, *CreateWorkflowTemplateRequest) (*WorkflowTemplate, error) @@ -1613,18 +1749,44 @@ type WorkflowTemplateServiceServer interface { // // The returned Operation can be used to track execution of // workflow by polling - // [google.cloud.dataproc.v1beta2.OperationService.GetOperation][]. + // [operations.get][google.longrunning.Operations.GetOperation]. // The Operation will complete when entire workflow is finished. // // The running workflow can be aborted via - // [google.cloud.dataproc.v1beta2.OperationService.CancelOperation][]. + // [operations.cancel][google.longrunning.Operations.CancelOperation]. + // This will cause any inflight jobs to be cancelled and workflow-owned + // clusters to be deleted. // - // The [google.cloud.dataproc.v1beta2.Operation.metadata][] will always be - // [google.cloud.dataproc.v1beta2.WorkflowMetadata][google.cloud.dataproc.v1beta2.WorkflowMetadata]. + // The [Operation.metadata][google.longrunning.Operation.metadata] will be + // [WorkflowMetadata][google.cloud.dataproc.v1beta2.WorkflowMetadata]. // - // The [google.cloud.dataproc.v1beta2.Operation.result][] will always be - // [google.protobuf.Empty][google.protobuf.Empty]. + // On successful completion, + // [Operation.response][google.longrunning.Operation.response] will be + // [Empty][google.protobuf.Empty]. InstantiateWorkflowTemplate(context.Context, *InstantiateWorkflowTemplateRequest) (*longrunning.Operation, error) + // Instantiates a template and begins execution. + // + // This method is equivalent to executing the sequence + // [CreateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.CreateWorkflowTemplate], [InstantiateWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.InstantiateWorkflowTemplate], + // [DeleteWorkflowTemplate][google.cloud.dataproc.v1beta2.WorkflowTemplateService.DeleteWorkflowTemplate]. + // + // The returned Operation can be used to track execution of + // workflow by polling + // [operations.get][google.longrunning.Operations.GetOperation]. + // The Operation will complete when entire workflow is finished. + // + // The running workflow can be aborted via + // [operations.cancel][google.longrunning.Operations.CancelOperation]. + // This will cause any inflight jobs to be cancelled and workflow-owned + // clusters to be deleted. + // + // The [Operation.metadata][google.longrunning.Operation.metadata] will be + // [WorkflowMetadata][google.cloud.dataproc.v1beta2.WorkflowMetadata]. + // + // On successful completion, + // [Operation.response][google.longrunning.Operation.response] will be + // [Empty][google.protobuf.Empty]. + InstantiateInlineWorkflowTemplate(context.Context, *InstantiateInlineWorkflowTemplateRequest) (*longrunning.Operation, error) // Updates (replaces) workflow template. The updated template // must contain version that matches the current server version. UpdateWorkflowTemplate(context.Context, *UpdateWorkflowTemplateRequest) (*WorkflowTemplate, error) @@ -1692,6 +1854,24 @@ func _WorkflowTemplateService_InstantiateWorkflowTemplate_Handler(srv interface{ return interceptor(ctx, in, info, handler) } +func _WorkflowTemplateService_InstantiateInlineWorkflowTemplate_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(InstantiateInlineWorkflowTemplateRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(WorkflowTemplateServiceServer).InstantiateInlineWorkflowTemplate(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.dataproc.v1beta2.WorkflowTemplateService/InstantiateInlineWorkflowTemplate", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(WorkflowTemplateServiceServer).InstantiateInlineWorkflowTemplate(ctx, req.(*InstantiateInlineWorkflowTemplateRequest)) + } + return interceptor(ctx, in, info, handler) +} + func _WorkflowTemplateService_UpdateWorkflowTemplate_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { in := new(UpdateWorkflowTemplateRequest) if err := dec(in); err != nil { @@ -1762,6 +1942,10 @@ var _WorkflowTemplateService_serviceDesc = grpc.ServiceDesc{ MethodName: "InstantiateWorkflowTemplate", Handler: _WorkflowTemplateService_InstantiateWorkflowTemplate_Handler, }, + { + MethodName: "InstantiateInlineWorkflowTemplate", + Handler: _WorkflowTemplateService_InstantiateInlineWorkflowTemplate_Handler, + }, { MethodName: "UpdateWorkflowTemplate", Handler: _WorkflowTemplateService_UpdateWorkflowTemplate_Handler, @@ -1780,112 +1964,122 @@ var _WorkflowTemplateService_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/cloud/dataproc/v1beta2/workflow_templates.proto", fileDescriptor_workflow_templates_7ee424539012b1eb) + proto.RegisterFile("google/cloud/dataproc/v1beta2/workflow_templates.proto", fileDescriptor_workflow_templates_fe5253a1eb957933) } -var fileDescriptor_workflow_templates_7ee424539012b1eb = []byte{ - // 1634 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x58, 0x5d, 0x6f, 0xdb, 0x5c, - 0x1d, 0xaf, 0xd3, 0x26, 0x4d, 0xfe, 0x59, 0xdb, 0xe8, 0x3c, 0xac, 0x0b, 0xe9, 0x53, 0xad, 0x33, - 0x62, 0x94, 0x32, 0x12, 0x11, 0x04, 0xea, 0xba, 0x4d, 0x5a, 0xdb, 0x64, 0x6b, 0xda, 0x34, 0x09, - 0x4e, 0xbb, 0x21, 0xb8, 0x88, 0x9c, 0xf8, 0xcc, 0x75, 0xeb, 0xf8, 0xb8, 0xb6, 0xd3, 0xd1, 0xa1, - 0xdd, 0xa0, 0x49, 0x7c, 0x00, 0x24, 0x3e, 0x07, 0x9f, 0x01, 0xbe, 0x00, 0x68, 0x5c, 0x22, 0x71, - 0x83, 0xb8, 0xdf, 0x15, 0xb7, 0xe8, 0xbc, 0xd8, 0x71, 0x5e, 0x9d, 0x74, 0xcf, 0x4d, 0xe4, 0x73, - 0x72, 0x7e, 0xbf, 0xf3, 0xfb, 0xbf, 0x1e, 0x1f, 0xc3, 0x2f, 0x75, 0x42, 0x74, 0x13, 0x17, 0x3a, - 0x26, 0xe9, 0x69, 0x05, 0x4d, 0xf5, 0x54, 0xdb, 0x21, 0x9d, 0xc2, 0xcd, 0xcf, 0xda, 0xd8, 0x53, - 0x8b, 0x85, 0xf7, 0xc4, 0xb9, 0x7a, 0x67, 0x92, 0xf7, 0x2d, 0x0f, 0x77, 0x6d, 0x53, 0xf5, 0xb0, - 0x9b, 0xb7, 0x1d, 0xe2, 0x11, 0xb4, 0xc9, 0x71, 0x79, 0x86, 0xcb, 0xfb, 0xb8, 0xbc, 0xc0, 0xe5, - 0xbe, 0x15, 0xb4, 0xaa, 0x6d, 0x14, 0x54, 0xcb, 0x22, 0x9e, 0xea, 0x19, 0xc4, 0x12, 0xe0, 0xdc, - 0x93, 0xe9, 0x9b, 0x76, 0xcc, 0x9e, 0xeb, 0x61, 0xc7, 0x5f, 0xbd, 0x3d, 0x7d, 0xf5, 0x25, 0x69, - 0xfb, 0x2b, 0x7f, 0x20, 0x56, 0x9a, 0xc4, 0xd2, 0x9d, 0x9e, 0x65, 0x19, 0x96, 0x5e, 0x20, 0x36, - 0x76, 0x06, 0x36, 0xdf, 0x10, 0x8b, 0xd8, 0xa8, 0xdd, 0x7b, 0x57, 0xc0, 0x5d, 0xdb, 0xbb, 0x15, - 0x7f, 0x3e, 0x1c, 0xfe, 0xd3, 0x33, 0xba, 0xd8, 0xf5, 0xd4, 0xae, 0xcd, 0x17, 0xc8, 0x5f, 0x16, - 0x21, 0xf3, 0x56, 0x38, 0xe5, 0x4c, 0xf8, 0x04, 0xad, 0x42, 0xcc, 0xd0, 0xb2, 0xb1, 0x2d, 0x69, - 0x3b, 0xa5, 0xc4, 0x0c, 0x0d, 0x21, 0x58, 0xb2, 0xd4, 0x2e, 0xce, 0x4a, 0x6c, 0x86, 0x3d, 0xa3, - 0x2c, 0x2c, 0xdf, 0x60, 0xc7, 0x35, 0x88, 0x95, 0x5d, 0xdc, 0x92, 0xb6, 0xe3, 0x8a, 0x3f, 0x44, - 0xcf, 0x20, 0xdd, 0x71, 0xb0, 0xea, 0xe1, 0x16, 0xdd, 0x2c, 0xbb, 0xb4, 0x25, 0x6d, 0xa7, 0x8b, - 0xb9, 0xbc, 0x70, 0xb0, 0xaf, 0x24, 0x7f, 0xe6, 0x2b, 0x51, 0x80, 0x2f, 0xa7, 0x13, 0x14, 0xdc, - 0xb3, 0xb5, 0x00, 0x1c, 0x8f, 0x06, 0xf3, 0xe5, 0x0c, 0xdc, 0x84, 0x84, 0xa9, 0xb6, 0xb1, 0xe9, - 0x66, 0x13, 0x5b, 0x8b, 0xdb, 0xe9, 0xe2, 0xb3, 0xfc, 0xd4, 0xa8, 0xe6, 0x87, 0x0d, 0xcf, 0x57, - 0x19, 0xba, 0x6c, 0x79, 0xce, 0xad, 0x22, 0xa8, 0xd0, 0x1b, 0x48, 0xd9, 0xa6, 0xda, 0xc1, 0x5d, - 0x6c, 0x79, 0xd9, 0x65, 0xa6, 0x67, 0x77, 0x4e, 0xde, 0x86, 0x8f, 0x57, 0xfa, 0x54, 0xe8, 0x05, - 0x2c, 0xd1, 0x50, 0x67, 0x93, 0x4c, 0xea, 0x8f, 0x23, 0x28, 0xeb, 0x8e, 0x86, 0x1d, 0xac, 0x1d, - 0x93, 0xb6, 0xc2, 0x60, 0xb9, 0xa7, 0x90, 0x0e, 0xa9, 0x45, 0x19, 0x58, 0xbc, 0xc2, 0xb7, 0x22, - 0x42, 0xf4, 0x11, 0x7d, 0x0f, 0xe2, 0x37, 0xaa, 0xd9, 0xc3, 0x22, 0x8e, 0x7c, 0xb0, 0x17, 0xdb, - 0x95, 0xe4, 0x7f, 0x4b, 0xf0, 0xfd, 0x89, 0x12, 0xd1, 0xaf, 0x61, 0xad, 0xab, 0x5a, 0xaa, 0x8e, - 0xb5, 0x96, 0x48, 0x5c, 0xc6, 0x9a, 0x2e, 0xfe, 0x34, 0x42, 0xe2, 0x29, 0x47, 0x1d, 0x72, 0xd0, - 0xd1, 0x82, 0xb2, 0xda, 0x1d, 0x98, 0x41, 0xbf, 0x85, 0x8c, 0x60, 0x6c, 0xb9, 0xd8, 0xc4, 0x1d, - 0x8f, 0x38, 0x4c, 0x5c, 0xba, 0x98, 0x8f, 0xa0, 0x16, 0x0c, 0x4d, 0x81, 0x3a, 0x5a, 0x50, 0xd6, - 0x3a, 0x83, 0x53, 0x07, 0xe9, 0x50, 0x98, 0xe4, 0x3f, 0xc6, 0x60, 0x75, 0x50, 0x0e, 0x7a, 0x04, - 0xf7, 0xfc, 0xcd, 0x59, 0x2e, 0x73, 0xaf, 0xa4, 0xc5, 0x5c, 0x8d, 0xa6, 0x74, 0x09, 0x12, 0x1d, - 0x62, 0xbd, 0x33, 0x74, 0x96, 0xd1, 0xe9, 0xe2, 0x93, 0xd9, 0x54, 0x1d, 0x32, 0x8c, 0x22, 0xb0, - 0xe8, 0x57, 0x41, 0x12, 0x2e, 0xb1, 0xc8, 0x3e, 0x9d, 0xcb, 0x6d, 0xe3, 0x52, 0xf0, 0x6b, 0x62, - 0xfd, 0x0f, 0x09, 0xd6, 0x86, 0xbc, 0x47, 0xcb, 0xf9, 0x03, 0xb1, 0x82, 0x72, 0xa6, 0xcf, 0xe8, - 0x02, 0x56, 0x7d, 0xf7, 0x08, 0xf5, 0x31, 0xa6, 0x7e, 0x7f, 0xbe, 0xc8, 0xf8, 0xe3, 0xb0, 0x15, - 0x2b, 0x9d, 0xf0, 0x5c, 0xee, 0x25, 0xa0, 0xd1, 0x45, 0x73, 0xd9, 0xf4, 0xbf, 0x38, 0x40, 0xbf, - 0x1e, 0xd0, 0x03, 0x58, 0x76, 0x3d, 0x6c, 0xb7, 0x0c, 0x4d, 0xc0, 0x13, 0x74, 0x58, 0xd1, 0x50, - 0x05, 0xe0, 0x42, 0xd5, 0x08, 0xb1, 0x5b, 0x97, 0xa4, 0x2d, 0x32, 0x6d, 0x3b, 0xc2, 0x9e, 0x23, - 0x06, 0x38, 0x26, 0xed, 0xa3, 0x05, 0x25, 0x75, 0xe1, 0x0f, 0xd0, 0x2b, 0x48, 0xb9, 0xb6, 0xea, - 0x5c, 0x31, 0x26, 0x9e, 0x1d, 0x3f, 0x8a, 0x60, 0x6a, 0xd2, 0xf5, 0x9c, 0x28, 0xe9, 0x8a, 0x67, - 0x54, 0x85, 0xb4, 0x7d, 0xdb, 0x67, 0xe2, 0xbd, 0x31, 0xaa, 0xf6, 0x1b, 0xb7, 0x21, 0x2e, 0x10, - 0x78, 0xca, 0x76, 0x08, 0xc9, 0x0b, 0xe3, 0x06, 0x33, 0x2a, 0xde, 0x29, 0x1f, 0x47, 0x99, 0x67, - 0xdc, 0x60, 0xce, 0xb3, 0x7c, 0xc1, 0x1f, 0xd1, 0x4b, 0x58, 0xb6, 0x0d, 0x9d, 0x71, 0x24, 0x18, - 0xc7, 0x0f, 0xa3, 0xe4, 0x18, 0x3a, 0xa7, 0x48, 0xd8, 0xec, 0x09, 0x35, 0x60, 0x85, 0x9b, 0xe4, - 0x5e, 0x9b, 0x8c, 0x87, 0x77, 0xc9, 0x9d, 0x59, 0x1c, 0xd4, 0xbc, 0x36, 0x39, 0x59, 0xda, 0xed, - 0x0f, 0xd1, 0x69, 0x50, 0x43, 0xbc, 0x3b, 0xfe, 0x62, 0xe6, 0xee, 0x38, 0xb6, 0x85, 0x57, 0x01, - 0xdc, 0xce, 0x05, 0xd6, 0x7a, 0xa6, 0x61, 0xe9, 0xd9, 0xd4, 0x4c, 0xc5, 0x7d, 0x4c, 0xda, 0xcd, - 0x00, 0xa3, 0x84, 0xf0, 0xa8, 0x08, 0xf7, 0x6d, 0x07, 0x3b, 0xf8, 0xba, 0x67, 0xb8, 0x86, 0x87, - 0x5b, 0x22, 0xf9, 0xdc, 0x2c, 0x6c, 0x2d, 0x6e, 0xa7, 0x94, 0x6f, 0xc2, 0x7f, 0x36, 0x59, 0x26, - 0x7e, 0x4d, 0x05, 0x1f, 0x00, 0x24, 0x2f, 0x49, 0xbb, 0xe5, 0xdd, 0xda, 0x58, 0xfe, 0x6f, 0xe8, - 0xb4, 0x3e, 0xc5, 0x9e, 0x4a, 0x55, 0xa3, 0x1c, 0x24, 0xfd, 0xb7, 0x19, 0xc1, 0x18, 0x8c, 0xc3, - 0xa7, 0x74, 0x6c, 0xf0, 0x94, 0x7e, 0x03, 0xab, 0xe2, 0x94, 0xf6, 0xbb, 0x3c, 0x4f, 0xeb, 0xc2, - 0x6c, 0x05, 0x5f, 0xf7, 0x5f, 0x43, 0x94, 0x15, 0x4e, 0xe3, 0xf7, 0xd9, 0x03, 0x88, 0xeb, 0x8e, - 0x6a, 0x5f, 0x88, 0xdc, 0x7e, 0x32, 0xe3, 0x51, 0xf9, 0x9a, 0x62, 0x14, 0x0e, 0xa5, 0xda, 0x34, - 0x6c, 0xe2, 0x90, 0xb6, 0xf8, 0x1d, 0xb5, 0x71, 0x1a, 0x5f, 0xdb, 0x09, 0xc4, 0x5d, 0x8f, 0xba, - 0x89, 0x26, 0xfa, 0x6a, 0x64, 0x56, 0x0d, 0x7b, 0x3a, 0xdf, 0xa4, 0x60, 0x85, 0x73, 0x8c, 0x1c, - 0x28, 0xcb, 0x23, 0x07, 0x8a, 0xbc, 0x0b, 0x71, 0x06, 0x41, 0x69, 0x58, 0x3e, 0xaf, 0x9d, 0xd4, - 0xea, 0x6f, 0x6b, 0x99, 0x05, 0x3a, 0x68, 0x94, 0x6b, 0xa5, 0x4a, 0xed, 0x75, 0x46, 0xa2, 0x03, - 0xe5, 0xbc, 0x56, 0xa3, 0x83, 0x18, 0x4a, 0xc2, 0x52, 0xa9, 0x5e, 0x2b, 0x67, 0x16, 0xe5, 0x16, - 0x64, 0x86, 0x8d, 0xa1, 0x1b, 0x06, 0x2f, 0x7f, 0xfd, 0x66, 0x97, 0x0e, 0xe6, 0x2a, 0x1a, 0xcd, - 0x22, 0xec, 0x38, 0xe2, 0x58, 0x4d, 0x29, 0x7c, 0x40, 0xfb, 0xbd, 0x46, 0xfb, 0x3d, 0x0d, 0x70, - 0x52, 0x61, 0xcf, 0xb2, 0x02, 0x2b, 0x03, 0xae, 0x47, 0xfb, 0x10, 0xb7, 0x88, 0x86, 0xdd, 0xac, - 0xc4, 0x2a, 0xee, 0x27, 0x33, 0xfa, 0xa6, 0x46, 0x34, 0xac, 0x70, 0xa4, 0xfc, 0xd7, 0x18, 0xdc, - 0x0b, 0xcf, 0x4f, 0xee, 0xcc, 0x13, 0x4b, 0x28, 0x36, 0xb1, 0x84, 0xd0, 0x7d, 0x48, 0xd0, 0x3a, - 0x30, 0x34, 0x66, 0x47, 0x4a, 0x89, 0x5f, 0x92, 0x76, 0x45, 0xeb, 0xc7, 0x34, 0x3e, 0x57, 0x4c, - 0xa9, 0xbe, 0x3c, 0xfd, 0x19, 0x88, 0x69, 0xe0, 0xbf, 0x44, 0xc8, 0x7f, 0xf2, 0x15, 0xa4, 0x82, - 0x95, 0x68, 0x03, 0x1e, 0xd4, 0xea, 0xa5, 0x72, 0xab, 0x79, 0xb6, 0x7f, 0x76, 0xde, 0x6c, 0x9d, - 0xd7, 0x9a, 0x8d, 0xf2, 0x61, 0xe5, 0x55, 0xa5, 0x5c, 0xe2, 0xa1, 0x3d, 0xa8, 0xd6, 0x0f, 0x4f, - 0xca, 0xa5, 0x8c, 0x84, 0xee, 0x41, 0x92, 0x86, 0x76, 0xff, 0xa0, 0x5a, 0xce, 0xc4, 0xc2, 0x81, - 0x5e, 0x44, 0x2b, 0x90, 0x3a, 0xac, 0x9f, 0x36, 0xaa, 0xe5, 0xb3, 0x72, 0x29, 0xb3, 0x84, 0x00, - 0x12, 0xaf, 0xf6, 0x2b, 0xd5, 0x72, 0x29, 0x13, 0x97, 0x3f, 0x49, 0xb0, 0x79, 0xc8, 0x2a, 0x6a, - 0xf8, 0x15, 0x4d, 0xc1, 0xd7, 0x3d, 0xec, 0x7a, 0x68, 0x1d, 0x12, 0xb6, 0xea, 0xd0, 0xb7, 0x51, - 0xe1, 0x54, 0x3e, 0x42, 0x27, 0xa1, 0x3e, 0x10, 0x9b, 0xa9, 0x5e, 0x46, 0x76, 0x08, 0x08, 0xe4, - 0x63, 0xc8, 0xbd, 0xc6, 0xde, 0x24, 0x09, 0x11, 0x17, 0x82, 0xc1, 0x56, 0x23, 0xbb, 0x20, 0x57, - 0x2c, 0xd7, 0x53, 0x2d, 0xcf, 0x98, 0x62, 0xd6, 0x5c, 0x9c, 0xe8, 0x21, 0xa4, 0x0d, 0xc6, 0xd9, - 0xc1, 0xfd, 0x94, 0x00, 0x7f, 0xaa, 0xa2, 0xc9, 0x26, 0x6c, 0x9e, 0xb3, 0x9b, 0xc1, 0xa4, 0xfd, - 0x4e, 0x86, 0xda, 0xe6, 0x57, 0xb9, 0xcb, 0x81, 0x6f, 0xab, 0x86, 0x3b, 0xe2, 0x2f, 0x37, 0x2a, - 0x66, 0x1b, 0x90, 0xb2, 0x55, 0x1d, 0xb7, 0x5c, 0xe3, 0x03, 0x16, 0x26, 0x26, 0xe9, 0x44, 0xd3, - 0xf8, 0x80, 0xd1, 0x26, 0x00, 0xfb, 0xd3, 0x23, 0x57, 0xd8, 0x12, 0x26, 0xb2, 0xe5, 0x67, 0x74, - 0x42, 0xfe, 0xb3, 0x04, 0x9b, 0x13, 0x36, 0x75, 0x6d, 0x62, 0xb9, 0x18, 0x9d, 0x42, 0x2a, 0xb8, - 0xe7, 0x8a, 0xba, 0x9e, 0xdb, 0xc6, 0x3e, 0x03, 0x7a, 0x0c, 0x6b, 0x16, 0xfe, 0x9d, 0xd7, 0x0a, - 0x89, 0xe2, 0x7d, 0x66, 0x85, 0x4e, 0x37, 0x02, 0x61, 0xa7, 0xb0, 0x59, 0x62, 0x7d, 0xf7, 0x3b, - 0x09, 0x75, 0xf1, 0x4b, 0x0a, 0x1e, 0x0c, 0x33, 0x35, 0xb1, 0x73, 0x63, 0x74, 0x30, 0xfa, 0x2c, - 0xc1, 0xfa, 0xf8, 0x6a, 0x41, 0xcf, 0xa3, 0x0e, 0x8b, 0x69, 0x45, 0x96, 0x9b, 0xd7, 0x4f, 0xf2, - 0xf1, 0x1f, 0x3e, 0xff, 0xe7, 0x4f, 0xb1, 0x92, 0xbc, 0x1b, 0x5c, 0xe4, 0x7f, 0xcf, 0x63, 0xfc, - 0xc2, 0x76, 0xc8, 0x25, 0xee, 0x78, 0x6e, 0x61, 0xa7, 0xe0, 0x60, 0x9d, 0xde, 0xdc, 0x0b, 0x3b, - 0x1f, 0x83, 0x2f, 0x11, 0x41, 0xd4, 0xf6, 0xfa, 0xa7, 0xf6, 0xdf, 0x24, 0xf8, 0x66, 0x4c, 0xf5, - 0xa1, 0xa8, 0xab, 0xc4, 0xe4, 0x8a, 0x9d, 0xdf, 0x9e, 0x97, 0xcc, 0x9e, 0x3d, 0x14, 0xb2, 0x87, - 0xc6, 0x69, 0xac, 0x35, 0xa3, 0xc6, 0x14, 0x76, 0x3e, 0xa2, 0xbf, 0x4b, 0xb0, 0x31, 0xa5, 0xee, - 0x51, 0xd4, 0xd5, 0x22, 0xba, 0x67, 0xe4, 0x36, 0x7d, 0x8a, 0xd0, 0x17, 0x92, 0x7c, 0x70, 0x62, - 0xca, 0x75, 0x66, 0x43, 0x45, 0x2e, 0xdd, 0xd5, 0x86, 0x3d, 0xa3, 0xaf, 0x61, 0x4f, 0xda, 0x41, - 0xff, 0x92, 0x60, 0x7d, 0x7c, 0x57, 0x89, 0x4c, 0xb7, 0xa9, 0xcd, 0x68, 0xfe, 0xf0, 0x34, 0x99, - 0x69, 0xa7, 0xb9, 0xfd, 0xbe, 0x69, 0x7e, 0xfa, 0xe4, 0xe7, 0xb4, 0xb1, 0x9f, 0x77, 0xff, 0x94, - 0xe0, 0xfe, 0xd8, 0x8e, 0x82, 0xa2, 0xbe, 0xa4, 0x4c, 0x6b, 0x7e, 0xb9, 0xe7, 0x77, 0x03, 0xf3, - 0x26, 0x36, 0x2e, 0x11, 0xe7, 0x2b, 0x2c, 0xf4, 0x17, 0x09, 0xd6, 0xc7, 0x37, 0xa4, 0xc8, 0xa8, - 0x4d, 0xed, 0x63, 0xb9, 0xf5, 0x91, 0xef, 0x52, 0xe5, 0xae, 0xed, 0xdd, 0xfa, 0x92, 0x77, 0xee, - 0x5c, 0x3b, 0x07, 0x9f, 0x24, 0x78, 0xd4, 0x21, 0xdd, 0xe9, 0xea, 0x0e, 0xd6, 0x47, 0xbc, 0xd6, - 0xa0, 0x42, 0x1a, 0xd2, 0x6f, 0xca, 0x02, 0xa8, 0x13, 0x53, 0xb5, 0xf4, 0x3c, 0x71, 0xf4, 0x82, - 0x8e, 0x2d, 0x26, 0xb3, 0xc0, 0xff, 0x52, 0x6d, 0xc3, 0x9d, 0xf0, 0x09, 0xf2, 0x99, 0x3f, 0xd1, - 0x4e, 0x30, 0xc4, 0xcf, 0xff, 0x1f, 0x00, 0x00, 0xff, 0xff, 0x4a, 0x21, 0xc2, 0x3b, 0x56, 0x15, +var fileDescriptor_workflow_templates_fe5253a1eb957933 = []byte{ + // 1794 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xc4, 0x59, 0xcf, 0x6f, 0xe3, 0x58, + 0x1d, 0x1f, 0xbb, 0x4d, 0x9a, 0x7c, 0x33, 0x6d, 0xa3, 0xb7, 0x4c, 0x27, 0x64, 0xb6, 0xda, 0x8e, + 0x11, 0x4b, 0x29, 0x43, 0x22, 0x8a, 0x40, 0xb3, 0x9d, 0x19, 0xb1, 0x6d, 0x92, 0xce, 0xa4, 0x4d, + 0x93, 0xe0, 0xb4, 0xbb, 0xab, 0xe1, 0x10, 0x39, 0xf1, 0x1b, 0xd7, 0xad, 0xe3, 0xe7, 0xb1, 0x9d, + 0xee, 0x76, 0xd0, 0x22, 0x81, 0x56, 0x70, 0xe1, 0xc4, 0x4a, 0x7b, 0x5b, 0x38, 0x71, 0x66, 0x6f, + 0xdc, 0x91, 0xf8, 0x03, 0x10, 0x7f, 0xc0, 0x5e, 0x38, 0x23, 0x71, 0xe2, 0x8a, 0xde, 0x0f, 0x3b, + 0xce, 0x4f, 0x27, 0x2d, 0xd2, 0x5e, 0x2a, 0xbf, 0xe7, 0xf7, 0xf9, 0xf8, 0xfb, 0xfd, 0x7c, 0x7f, + 0x39, 0x2e, 0xfc, 0xd4, 0x20, 0xc4, 0xb0, 0x70, 0xb1, 0x6b, 0x91, 0xbe, 0x5e, 0xd4, 0x35, 0x5f, + 0x73, 0x5c, 0xd2, 0x2d, 0x5e, 0xfd, 0xa8, 0x83, 0x7d, 0x6d, 0xb7, 0xf8, 0x31, 0x71, 0x2f, 0x5f, + 0x59, 0xe4, 0xe3, 0xb6, 0x8f, 0x7b, 0x8e, 0xa5, 0xf9, 0xd8, 0x2b, 0x38, 0x2e, 0xf1, 0x09, 0xda, + 0xe4, 0xb8, 0x02, 0xc3, 0x15, 0x02, 0x5c, 0x41, 0xe0, 0xf2, 0x6f, 0x0b, 0x5a, 0xcd, 0x31, 0x8b, + 0x9a, 0x6d, 0x13, 0x5f, 0xf3, 0x4d, 0x62, 0x0b, 0x70, 0xfe, 0xd1, 0xec, 0x87, 0x76, 0xad, 0xbe, + 0xe7, 0x63, 0x37, 0x38, 0xbd, 0x3d, 0xfb, 0xf4, 0x05, 0xe9, 0x04, 0x27, 0xbf, 0x23, 0x4e, 0x5a, + 0xc4, 0x36, 0xdc, 0xbe, 0x6d, 0x9b, 0xb6, 0x51, 0x24, 0x0e, 0x76, 0x87, 0x1e, 0xfe, 0x40, 0x1c, + 0x62, 0xab, 0x4e, 0xff, 0x55, 0x11, 0xf7, 0x1c, 0xff, 0x5a, 0xdc, 0x7c, 0x67, 0xf4, 0xa6, 0x6f, + 0xf6, 0xb0, 0xe7, 0x6b, 0x3d, 0x87, 0x1f, 0x50, 0xfe, 0xb3, 0x04, 0xd9, 0x0f, 0x85, 0x28, 0xa7, + 0x42, 0x13, 0xb4, 0x06, 0xb2, 0xa9, 0xe7, 0xe4, 0x2d, 0x69, 0x3b, 0xad, 0xca, 0xa6, 0x8e, 0x10, + 0x2c, 0xdb, 0x5a, 0x0f, 0xe7, 0x24, 0xb6, 0xc3, 0xae, 0x51, 0x0e, 0x56, 0xae, 0xb0, 0xeb, 0x99, + 0xc4, 0xce, 0x2d, 0x6d, 0x49, 0xdb, 0x09, 0x35, 0x58, 0xa2, 0x27, 0x90, 0xe9, 0xba, 0x58, 0xf3, + 0x71, 0x9b, 0x3e, 0x2c, 0xb7, 0xbc, 0x25, 0x6d, 0x67, 0x76, 0xf3, 0x05, 0x21, 0x70, 0x60, 0x49, + 0xe1, 0x34, 0xb0, 0x44, 0x05, 0x7e, 0x9c, 0x6e, 0x50, 0x70, 0xdf, 0xd1, 0x43, 0x70, 0x22, 0x1e, + 0xcc, 0x8f, 0x33, 0x70, 0x0b, 0x92, 0x96, 0xd6, 0xc1, 0x96, 0x97, 0x4b, 0x6e, 0x2d, 0x6d, 0x67, + 0x76, 0x9f, 0x14, 0x66, 0x46, 0xb5, 0x30, 0xea, 0x78, 0xa1, 0xc6, 0xd0, 0x15, 0xdb, 0x77, 0xaf, + 0x55, 0x41, 0x85, 0x3e, 0x80, 0xb4, 0x63, 0x69, 0x5d, 0xdc, 0xc3, 0xb6, 0x9f, 0x5b, 0x61, 0xf6, + 0x3c, 0x5e, 0x90, 0xb7, 0x19, 0xe0, 0xd5, 0x01, 0x15, 0x7a, 0x06, 0xcb, 0x34, 0xd4, 0xb9, 0x14, + 0x33, 0xf5, 0xfb, 0x31, 0x94, 0x0d, 0x57, 0xc7, 0x2e, 0xd6, 0x8f, 0x48, 0x47, 0x65, 0xb0, 0xfc, + 0x7b, 0x90, 0x89, 0x58, 0x8b, 0xb2, 0xb0, 0x74, 0x89, 0xaf, 0x45, 0x84, 0xe8, 0x25, 0xfa, 0x16, + 0x24, 0xae, 0x34, 0xab, 0x8f, 0x45, 0x1c, 0xf9, 0x62, 0x4f, 0x7e, 0x2c, 0x29, 0x5f, 0x4b, 0xf0, + 0xed, 0xa9, 0x26, 0xa2, 0x8f, 0x60, 0xbd, 0xa7, 0xd9, 0x9a, 0x81, 0xf5, 0xb6, 0x48, 0x5c, 0xc6, + 0x9a, 0xd9, 0xfd, 0x61, 0x8c, 0x89, 0x27, 0x1c, 0x55, 0xe2, 0xa0, 0x17, 0x77, 0xd4, 0xb5, 0xde, + 0xd0, 0x0e, 0xfa, 0x05, 0x64, 0x05, 0x63, 0xdb, 0xc3, 0x16, 0xee, 0xfa, 0xc4, 0x65, 0xc6, 0x65, + 0x76, 0x0b, 0x31, 0xd4, 0x82, 0xa1, 0x25, 0x50, 0x2f, 0xee, 0xa8, 0xeb, 0xdd, 0xe1, 0xad, 0x83, + 0x4c, 0x24, 0x4c, 0xca, 0xef, 0x64, 0x58, 0x1b, 0x36, 0x07, 0x3d, 0x84, 0xbb, 0xc1, 0xc3, 0x59, + 0x2e, 0x73, 0x55, 0x32, 0x62, 0xaf, 0x4e, 0x53, 0xba, 0x0c, 0xc9, 0x2e, 0xb1, 0x5f, 0x99, 0x06, + 0xcb, 0xe8, 0xcc, 0xee, 0xa3, 0xf9, 0xac, 0x2a, 0x31, 0x8c, 0x2a, 0xb0, 0xe8, 0xe7, 0x61, 0x12, + 0x2e, 0xb3, 0xc8, 0xbe, 0xb7, 0x90, 0x6c, 0x93, 0x52, 0xf0, 0x36, 0xb1, 0xfe, 0x87, 0x04, 0xeb, + 0x23, 0xea, 0xd1, 0x72, 0x7e, 0x43, 0xec, 0xb0, 0x9c, 0xe9, 0x35, 0x3a, 0x87, 0xb5, 0x40, 0x1e, + 0x61, 0xbd, 0xcc, 0xac, 0xdf, 0x5f, 0x2c, 0x32, 0xc1, 0x3a, 0xea, 0xc5, 0x6a, 0x37, 0xba, 0x97, + 0x7f, 0x1f, 0xd0, 0xf8, 0xa1, 0x85, 0x7c, 0xfa, 0x6f, 0x02, 0x60, 0x50, 0x0f, 0xe8, 0x3e, 0xac, + 0x78, 0x3e, 0x76, 0xda, 0xa6, 0x2e, 0xe0, 0x49, 0xba, 0xac, 0xea, 0xa8, 0x0a, 0x70, 0xae, 0xe9, + 0x84, 0x38, 0xed, 0x0b, 0xd2, 0x11, 0x99, 0xb6, 0x1d, 0xe3, 0xcf, 0x0b, 0x06, 0x38, 0x22, 0x9d, + 0x17, 0x77, 0xd4, 0xf4, 0x79, 0xb0, 0x40, 0x87, 0x90, 0xf6, 0x1c, 0xcd, 0xbd, 0x64, 0x4c, 0x3c, + 0x3b, 0xbe, 0x17, 0xc3, 0xd4, 0xa2, 0xe7, 0x39, 0x51, 0xca, 0x13, 0xd7, 0xa8, 0x06, 0x19, 0xe7, + 0x7a, 0xc0, 0xc4, 0x7b, 0x63, 0x5c, 0xed, 0x37, 0xaf, 0x23, 0x5c, 0x20, 0xf0, 0x94, 0xad, 0x04, + 0xa9, 0x73, 0xf3, 0x0a, 0x33, 0x2a, 0xde, 0x29, 0xdf, 0x8d, 0x73, 0xcf, 0xbc, 0xc2, 0x9c, 0x67, + 0xe5, 0x9c, 0x5f, 0xa2, 0xf7, 0x61, 0xc5, 0x31, 0x0d, 0xc6, 0x91, 0x64, 0x1c, 0xdf, 0x8d, 0x33, + 0xc7, 0x34, 0x38, 0x45, 0xd2, 0x61, 0x57, 0xa8, 0x09, 0xab, 0xdc, 0x25, 0xef, 0xb5, 0xc5, 0x78, + 0x78, 0x97, 0xdc, 0x99, 0x47, 0xa0, 0xd6, 0x6b, 0x8b, 0x93, 0x65, 0xbc, 0xc1, 0x12, 0x9d, 0x84, + 0x35, 0xc4, 0xbb, 0xe3, 0x4f, 0xe6, 0xee, 0x8e, 0x13, 0x5b, 0x78, 0x0d, 0xc0, 0xeb, 0x9e, 0x63, + 0xbd, 0x6f, 0x99, 0xb6, 0x91, 0x4b, 0xcf, 0x55, 0xdc, 0x47, 0xa4, 0xd3, 0x0a, 0x31, 0x6a, 0x04, + 0x8f, 0x76, 0xe1, 0x9e, 0xe3, 0x62, 0x17, 0xbf, 0xee, 0x9b, 0x9e, 0xe9, 0xe3, 0xb6, 0x48, 0x3e, + 0x2f, 0x07, 0x5b, 0x4b, 0xdb, 0x69, 0xf5, 0xad, 0xe8, 0xcd, 0x16, 0xcb, 0xc4, 0xdb, 0x54, 0xf0, + 0x01, 0x40, 0xea, 0x82, 0x74, 0xda, 0xfe, 0xb5, 0x83, 0x95, 0xdf, 0x26, 0x06, 0xd3, 0xfa, 0x04, + 0xfb, 0x1a, 0xb5, 0x1a, 0xe5, 0x21, 0x15, 0xbc, 0xcd, 0x08, 0xc6, 0x70, 0x1d, 0x9d, 0xd2, 0xf2, + 0xf0, 0x94, 0xfe, 0x00, 0xd6, 0xc4, 0x94, 0x0e, 0xba, 0x3c, 0x4f, 0xeb, 0xe2, 0x7c, 0x05, 0xdf, + 0x08, 0x5e, 0x43, 0xd4, 0x55, 0x4e, 0x13, 0xf4, 0xd9, 0x03, 0x48, 0x18, 0xae, 0xe6, 0x9c, 0x8b, + 0xdc, 0x7e, 0x34, 0xe7, 0xa8, 0x7c, 0x4e, 0x31, 0x2a, 0x87, 0x52, 0xdb, 0x74, 0x6c, 0xe1, 0x88, + 0x6d, 0x89, 0x1b, 0xda, 0xc6, 0x69, 0x02, 0xdb, 0x8e, 0x21, 0xe1, 0xf9, 0x54, 0x26, 0x9a, 0xe8, + 0x6b, 0xb1, 0x59, 0x35, 0xaa, 0x74, 0xa1, 0x45, 0xc1, 0x2a, 0xe7, 0x18, 0x1b, 0x28, 0x2b, 0xe3, + 0x03, 0xa5, 0x0d, 0xe0, 0x68, 0xae, 0xd6, 0xc3, 0xf4, 0xed, 0x4f, 0xa4, 0xf2, 0xcf, 0x16, 0x7d, + 0x68, 0x33, 0x64, 0xe0, 0x49, 0x1d, 0xa1, 0xcc, 0x3f, 0x83, 0xf5, 0x91, 0xdb, 0x0b, 0x35, 0xd2, + 0xc7, 0x90, 0x60, 0x2e, 0xa1, 0x0c, 0xac, 0x9c, 0xd5, 0x8f, 0xeb, 0x8d, 0x0f, 0xeb, 0xd9, 0x3b, + 0x74, 0xd1, 0xac, 0xd4, 0xcb, 0xd5, 0xfa, 0xf3, 0xac, 0x44, 0x17, 0xea, 0x59, 0xbd, 0x4e, 0x17, + 0x32, 0x4a, 0xc1, 0x72, 0xb9, 0x51, 0xaf, 0x64, 0x97, 0x94, 0x36, 0x64, 0x47, 0xc5, 0xa6, 0x82, + 0x84, 0x2f, 0xa7, 0x83, 0x66, 0x9c, 0x09, 0xf7, 0xaa, 0x3a, 0x35, 0x05, 0xbb, 0xae, 0x18, 0xfb, + 0x69, 0x95, 0x2f, 0xe8, 0x3c, 0xd2, 0xe9, 0x3c, 0xa2, 0x09, 0x98, 0x52, 0xd9, 0xb5, 0xa2, 0xc2, + 0xea, 0x50, 0x6a, 0xa0, 0x7d, 0x48, 0xd8, 0x44, 0xc7, 0x5e, 0x4e, 0x62, 0x32, 0xfe, 0x60, 0x4e, + 0x19, 0xeb, 0x44, 0xc7, 0x2a, 0x47, 0x2a, 0x7f, 0x93, 0xe1, 0x6e, 0x74, 0x7f, 0xfa, 0xe4, 0x98, + 0x5a, 0xe2, 0xf2, 0xd4, 0x12, 0x47, 0xf7, 0x20, 0x49, 0xeb, 0xd4, 0xd4, 0x99, 0x1f, 0x69, 0x35, + 0x71, 0x41, 0x3a, 0x55, 0x7d, 0x90, 0x73, 0x89, 0x85, 0x72, 0x8e, 0xda, 0x57, 0xa0, 0x7f, 0x86, + 0x72, 0x2e, 0xd4, 0x2f, 0x19, 0xd1, 0x4f, 0xb9, 0x84, 0x74, 0x78, 0x12, 0x3d, 0x80, 0xfb, 0xf5, + 0x46, 0xb9, 0xd2, 0x6e, 0x9d, 0xee, 0x9f, 0x9e, 0xb5, 0xda, 0x67, 0xf5, 0x56, 0xb3, 0x52, 0xaa, + 0x1e, 0x56, 0x2b, 0x65, 0x1e, 0xda, 0x83, 0x5a, 0xa3, 0x74, 0x5c, 0x29, 0x67, 0x25, 0x74, 0x17, + 0x52, 0x34, 0xb4, 0xfb, 0x07, 0xb5, 0x4a, 0x56, 0x8e, 0x06, 0x7a, 0x09, 0xad, 0x42, 0xba, 0xd4, + 0x38, 0x69, 0xd6, 0x2a, 0xa7, 0x95, 0x72, 0x76, 0x19, 0x01, 0x24, 0x0f, 0xf7, 0xab, 0xb5, 0x4a, + 0x39, 0x9b, 0x50, 0x3e, 0x93, 0x60, 0xb3, 0xc4, 0x2a, 0x7e, 0xf4, 0x15, 0x52, 0xc5, 0xaf, 0xfb, + 0xd8, 0xf3, 0xd1, 0x06, 0x24, 0x1d, 0xcd, 0xa5, 0x6f, 0xcb, 0x42, 0x54, 0xbe, 0x42, 0xc7, 0x91, + 0x3e, 0x25, 0xcf, 0x55, 0xcf, 0x63, 0x4f, 0x08, 0x09, 0x94, 0x23, 0xc8, 0x3f, 0xc7, 0xfe, 0x34, + 0x13, 0x62, 0x7e, 0xb0, 0x0c, 0xb7, 0x42, 0xc5, 0x03, 0xa5, 0x6a, 0x7b, 0xbe, 0x66, 0xfb, 0xe6, + 0x0c, 0xb7, 0x16, 0xe2, 0x44, 0xef, 0x40, 0xc6, 0x64, 0x9c, 0x5d, 0x3c, 0x48, 0x09, 0x08, 0xb6, + 0xaa, 0xba, 0xf2, 0x95, 0x04, 0xdb, 0x91, 0xa7, 0x56, 0x6d, 0xcb, 0xb4, 0xbf, 0x51, 0x49, 0xe3, + 0x4d, 0xb6, 0x60, 0xf3, 0x8c, 0xfd, 0xd8, 0x9a, 0x66, 0xe6, 0xf1, 0xc8, 0x24, 0xba, 0x55, 0x84, + 0x5d, 0x78, 0xbb, 0x66, 0x7a, 0x63, 0x21, 0xf6, 0xe2, 0x34, 0x79, 0x00, 0x69, 0x47, 0x33, 0x70, + 0xdb, 0x33, 0xdf, 0x60, 0x11, 0x95, 0x14, 0xdd, 0x68, 0x99, 0x6f, 0x30, 0xda, 0xa4, 0x1d, 0xd9, + 0xc0, 0x6d, 0x9f, 0x5c, 0x62, 0x5b, 0xb8, 0xc8, 0x8e, 0x9f, 0xd2, 0x0d, 0xe5, 0x0b, 0x09, 0x36, + 0xa7, 0x3c, 0xd4, 0x73, 0x88, 0xed, 0x61, 0x74, 0x02, 0xe9, 0xf0, 0xd3, 0x81, 0x68, 0x45, 0x0b, + 0xfb, 0x38, 0x60, 0x40, 0xef, 0xc2, 0xba, 0x8d, 0x3f, 0xf1, 0xdb, 0x11, 0xa3, 0x78, 0x6b, 0x5c, + 0xa5, 0xdb, 0xcd, 0xd0, 0xb0, 0x13, 0xd8, 0x2c, 0xb3, 0x51, 0xf6, 0x7f, 0xc9, 0xce, 0xdd, 0xaf, + 0xd7, 0xe1, 0xfe, 0x28, 0x53, 0x0b, 0xbb, 0x57, 0x66, 0x17, 0xa3, 0x2f, 0x65, 0xd8, 0x98, 0x5c, + 0xe0, 0xe8, 0x69, 0xdc, 0xfc, 0x9d, 0xd5, 0x17, 0xf2, 0x8b, 0xea, 0xa4, 0x7c, 0x2e, 0xfd, 0xe6, + 0x9f, 0xff, 0xfa, 0x5c, 0xfe, 0xbd, 0xa4, 0x3c, 0x0e, 0xbf, 0x8e, 0xfc, 0x92, 0x47, 0xf9, 0x99, + 0xe3, 0x92, 0x0b, 0xdc, 0xf5, 0xbd, 0xe2, 0x4e, 0xd1, 0xc5, 0x86, 0x49, 0x6c, 0xaf, 0xb8, 0xf3, + 0x69, 0xf8, 0x79, 0x27, 0x8c, 0xdb, 0x5e, 0x98, 0x4f, 0x2f, 0x0f, 0x95, 0xbd, 0x59, 0x2c, 0x16, + 0xe9, 0xf2, 0xcf, 0x2a, 0x31, 0x3c, 0xe8, 0xd7, 0x32, 0xbc, 0x35, 0xa1, 0xf5, 0xa0, 0xb8, 0xdf, + 0x79, 0xd3, 0xdb, 0xd5, 0xe2, 0xca, 0xfc, 0x8a, 0x09, 0xf3, 0x09, 0x8a, 0xe8, 0x42, 0x23, 0x3e, + 0x51, 0x95, 0x71, 0x67, 0x8a, 0x3b, 0x9f, 0xbe, 0x7c, 0x8a, 0xf6, 0xa6, 0x63, 0x23, 0x5a, 0x4c, + 0x44, 0xa3, 0x3f, 0xc9, 0xf0, 0x60, 0x46, 0xcb, 0x44, 0x71, 0xbf, 0x1a, 0xe3, 0xdb, 0x6d, 0x7e, + 0x33, 0xa0, 0x88, 0x7c, 0xfc, 0x2a, 0x84, 0x2f, 0x1b, 0xca, 0x97, 0x3c, 0x37, 0xbe, 0x90, 0x94, + 0xf2, 0x4d, 0x35, 0xd8, 0x33, 0x07, 0x56, 0xec, 0x49, 0x3b, 0x2f, 0x8f, 0x95, 0xc3, 0x9b, 0x4b, + 0x32, 0x42, 0x86, 0xfe, 0x2e, 0xc3, 0xc3, 0xd8, 0xf6, 0x8e, 0x9e, 0xcf, 0xaf, 0xd3, 0xcc, 0x01, + 0x11, 0xa7, 0xd6, 0x5f, 0xb9, 0x5a, 0x5f, 0x49, 0xca, 0xd1, 0x8d, 0x2b, 0xc9, 0x1c, 0xb5, 0x25, + 0x52, 0x5b, 0x1f, 0x29, 0xb5, 0x5b, 0xd4, 0xd6, 0x0c, 0x66, 0xf4, 0x17, 0x19, 0x36, 0x26, 0x0f, + 0x9d, 0xd8, 0x6e, 0x34, 0x73, 0x56, 0x2d, 0x5e, 0x73, 0x7f, 0xe6, 0x1a, 0xfe, 0x51, 0xca, 0xef, + 0x0f, 0x7c, 0x0d, 0xec, 0x2c, 0x2c, 0x98, 0x7a, 0x03, 0xe9, 0x1a, 0xf9, 0xd2, 0x1c, 0x74, 0xb1, + 0xe9, 0x37, 0x50, 0xec, 0x0f, 0x32, 0xdc, 0x9b, 0x38, 0xc3, 0x50, 0xdc, 0xe7, 0xd0, 0x59, 0xe3, + 0x36, 0xff, 0xf4, 0x66, 0x60, 0x3e, 0x36, 0x27, 0x35, 0xac, 0xc5, 0xd2, 0x6f, 0xb8, 0x61, 0x2d, + 0x9a, 0x62, 0xe8, 0xdf, 0x12, 0x6c, 0x4c, 0x1e, 0xa0, 0xb1, 0x69, 0x34, 0x73, 0xee, 0xe6, 0x37, + 0xc6, 0x3e, 0x4d, 0x57, 0x7a, 0x8e, 0x7f, 0x1d, 0x38, 0xbc, 0x73, 0x8b, 0x0e, 0xbd, 0x73, 0x8b, + 0x0e, 0x7d, 0xf0, 0x99, 0x04, 0x0f, 0xbb, 0xa4, 0x37, 0xdb, 0xb7, 0x83, 0x8d, 0xb1, 0x88, 0x35, + 0xa9, 0x1b, 0x4d, 0xe9, 0x65, 0x45, 0x00, 0x0d, 0x62, 0x69, 0xb6, 0x51, 0x20, 0xae, 0x51, 0x34, + 0xb0, 0xcd, 0x9c, 0x2c, 0xf2, 0x5b, 0x9a, 0x63, 0x7a, 0x53, 0xfe, 0x87, 0xf1, 0x24, 0xd8, 0xe8, + 0x24, 0x19, 0xe2, 0xc7, 0xff, 0x0b, 0x00, 0x00, 0xff, 0xff, 0x10, 0x10, 0x40, 0xc7, 0x97, 0x19, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/agent.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/agent.pb.go index f3cf6cf99..01dd774ed 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/agent.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/agent.pb.go @@ -64,32 +64,32 @@ func (Agent_MatchMode) EnumDescriptor() ([]byte, []int) { type Agent struct { // Required. The project of this agent. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The name of this agent. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. The default language of the agent as a language tag. See // [Language Support](https://dialogflow.com/docs/reference/language) for a // list of the currently supported language codes. // This field cannot be set by the `Update` method. - DefaultLanguageCode string `protobuf:"bytes,3,opt,name=default_language_code,json=defaultLanguageCode" json:"default_language_code,omitempty"` + DefaultLanguageCode string `protobuf:"bytes,3,opt,name=default_language_code,json=defaultLanguageCode,proto3" json:"default_language_code,omitempty"` // Optional. The list of all languages supported by this agent (except for the // `default_language_code`). - SupportedLanguageCodes []string `protobuf:"bytes,4,rep,name=supported_language_codes,json=supportedLanguageCodes" json:"supported_language_codes,omitempty"` + SupportedLanguageCodes []string `protobuf:"bytes,4,rep,name=supported_language_codes,json=supportedLanguageCodes,proto3" json:"supported_language_codes,omitempty"` // Required. The time zone of this agent from the // [time zone database](https://www.iana.org/time-zones), e.g., // America/New_York, Europe/Paris. - TimeZone string `protobuf:"bytes,5,opt,name=time_zone,json=timeZone" json:"time_zone,omitempty"` + TimeZone string `protobuf:"bytes,5,opt,name=time_zone,json=timeZone,proto3" json:"time_zone,omitempty"` // Optional. The description of this agent. // The maximum length is 500 characters. If exceeded, the request is rejected. - Description string `protobuf:"bytes,6,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,6,opt,name=description,proto3" json:"description,omitempty"` // Optional. The URI of the agent's avatar. // Avatars are used throughout the Dialogflow console and in the self-hosted // [Web Demo](https://dialogflow.com/docs/integrations/web-demo) integration. - AvatarUri string `protobuf:"bytes,7,opt,name=avatar_uri,json=avatarUri" json:"avatar_uri,omitempty"` + AvatarUri string `protobuf:"bytes,7,opt,name=avatar_uri,json=avatarUri,proto3" json:"avatar_uri,omitempty"` // Optional. Determines whether this agent should log conversation queries. - EnableLogging bool `protobuf:"varint,8,opt,name=enable_logging,json=enableLogging" json:"enable_logging,omitempty"` + EnableLogging bool `protobuf:"varint,8,opt,name=enable_logging,json=enableLogging,proto3" json:"enable_logging,omitempty"` // Optional. Determines how intents are detected from user queries. - MatchMode Agent_MatchMode `protobuf:"varint,9,opt,name=match_mode,json=matchMode,enum=google.cloud.dialogflow.v2.Agent_MatchMode" json:"match_mode,omitempty"` + MatchMode Agent_MatchMode `protobuf:"varint,9,opt,name=match_mode,json=matchMode,proto3,enum=google.cloud.dialogflow.v2.Agent_MatchMode" json:"match_mode,omitempty"` // Optional. To filter out false positive results and still get variety in // matched natural language inputs for your agent, you can tune the machine // learning classification threshold. If the returned score value is less than @@ -97,7 +97,7 @@ type Agent struct { // are no fallback intents defined, no intent will be triggered. The score // values range from 0.0 (completely uncertain) to 1.0 (completely certain). // If set to 0.0, the default of 0.3 is used. - ClassificationThreshold float32 `protobuf:"fixed32,10,opt,name=classification_threshold,json=classificationThreshold" json:"classification_threshold,omitempty"` + ClassificationThreshold float32 `protobuf:"fixed32,10,opt,name=classification_threshold,json=classificationThreshold,proto3" json:"classification_threshold,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -201,7 +201,7 @@ func (m *Agent) GetClassificationThreshold() float32 { type GetAgentRequest struct { // Required. The project that the agent to fetch is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -242,12 +242,12 @@ func (m *GetAgentRequest) GetParent() string { type SearchAgentsRequest struct { // Required. The project to list agents from. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -302,10 +302,10 @@ func (m *SearchAgentsRequest) GetPageToken() string { type SearchAgentsResponse struct { // The list of agents. There will be a maximum number of items returned based // on the page_size field in the request. - Agents []*Agent `protobuf:"bytes,1,rep,name=agents" json:"agents,omitempty"` + Agents []*Agent `protobuf:"bytes,1,rep,name=agents,proto3" json:"agents,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -353,7 +353,7 @@ func (m *SearchAgentsResponse) GetNextPageToken() string { type TrainAgentRequest struct { // Required. The project that the agent to train is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -394,11 +394,11 @@ func (m *TrainAgentRequest) GetParent() string { type ExportAgentRequest struct { // Required. The project that the agent to export is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The Google Cloud Storage URI to export the agent to. // Note: The URI must start with // "gs://". If left unspecified, the serialized agent is returned inline. - AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri" json:"agent_uri,omitempty"` + AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,proto3" json:"agent_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -484,7 +484,7 @@ type isExportAgentResponse_Agent interface { } type ExportAgentResponse_AgentUri struct { - AgentUri string `protobuf:"bytes,1,opt,name=agent_uri,json=agentUri,oneof"` + AgentUri string `protobuf:"bytes,1,opt,name=agent_uri,json=agentUri,proto3,oneof"` } type ExportAgentResponse_AgentContent struct { AgentContent []byte `protobuf:"bytes,2,opt,name=agent_content,json=agentContent,proto3,oneof"` @@ -584,7 +584,7 @@ func _ExportAgentResponse_OneofSizer(msg proto.Message) (n int) { type ImportAgentRequest struct { // Required. The project that the agent to import is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The agent to import. // // Types that are valid to be assigned to Agent: @@ -625,7 +625,7 @@ type isImportAgentRequest_Agent interface { } type ImportAgentRequest_AgentUri struct { - AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,oneof"` + AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,proto3,oneof"` } type ImportAgentRequest_AgentContent struct { AgentContent []byte `protobuf:"bytes,3,opt,name=agent_content,json=agentContent,proto3,oneof"` @@ -732,7 +732,7 @@ func _ImportAgentRequest_OneofSizer(msg proto.Message) (n int) { type RestoreAgentRequest struct { // Required. The project that the agent to restore is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The agent to restore. // // Types that are valid to be assigned to Agent: @@ -773,7 +773,7 @@ type isRestoreAgentRequest_Agent interface { } type RestoreAgentRequest_AgentUri struct { - AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,oneof"` + AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,proto3,oneof"` } type RestoreAgentRequest_AgentContent struct { AgentContent []byte `protobuf:"bytes,3,opt,name=agent_content,json=agentContent,proto3,oneof"` @@ -1002,8 +1002,7 @@ func (c *agentsClient) RestoreAgent(ctx context.Context, in *RestoreAgentRequest return out, nil } -// Server API for Agents service - +// AgentsServer is the server API for Agents service. type AgentsServer interface { // Retrieves the specified agent. GetAgent(context.Context, *GetAgentRequest) (*Agent, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/context.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/context.pb.go index d25ba1108..78ce2e75f 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/context.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/context.pb.go @@ -31,16 +31,16 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Context struct { // Required. The unique identifier of the context. Format: // `projects//agent/sessions//contexts/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The number of conversational query requests after which the // context expires. If set to `0` (the default) the context expires // immediately. Contexts expire automatically after 10 minutes even if there // are no matching queries. - LifespanCount int32 `protobuf:"varint,2,opt,name=lifespan_count,json=lifespanCount" json:"lifespan_count,omitempty"` + LifespanCount int32 `protobuf:"varint,2,opt,name=lifespan_count,json=lifespanCount,proto3" json:"lifespan_count,omitempty"` // Optional. The collection of parameters associated with this context. // Refer to [this doc](https://dialogflow.com/docs/actions-and-parameters) for // syntax. - Parameters *_struct.Struct `protobuf:"bytes,3,opt,name=parameters" json:"parameters,omitempty"` + Parameters *_struct.Struct `protobuf:"bytes,3,opt,name=parameters,proto3" json:"parameters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -95,12 +95,12 @@ func (m *Context) GetParameters() *_struct.Struct { type ListContextsRequest struct { // Required. The session to list all contexts from. // Format: `projects//agent/sessions/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -155,10 +155,10 @@ func (m *ListContextsRequest) GetPageToken() string { type ListContextsResponse struct { // The list of contexts. There will be a maximum number of items // returned based on the page_size field in the request. - Contexts []*Context `protobuf:"bytes,1,rep,name=contexts" json:"contexts,omitempty"` + Contexts []*Context `protobuf:"bytes,1,rep,name=contexts,proto3" json:"contexts,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -206,7 +206,7 @@ func (m *ListContextsResponse) GetNextPageToken() string { type GetContextRequest struct { // Required. The name of the context. Format: // `projects//agent/sessions//contexts/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -247,9 +247,9 @@ func (m *GetContextRequest) GetName() string { type CreateContextRequest struct { // Required. The session to create a context for. // Format: `projects//agent/sessions/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The context to create. - Context *Context `protobuf:"bytes,2,opt,name=context" json:"context,omitempty"` + Context *Context `protobuf:"bytes,2,opt,name=context,proto3" json:"context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -296,9 +296,9 @@ func (m *CreateContextRequest) GetContext() *Context { // The request message for [Contexts.UpdateContext][google.cloud.dialogflow.v2.Contexts.UpdateContext]. type UpdateContextRequest struct { // Required. The context to update. - Context *Context `protobuf:"bytes,1,opt,name=context" json:"context,omitempty"` + Context *Context `protobuf:"bytes,1,opt,name=context,proto3" json:"context,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -346,7 +346,7 @@ func (m *UpdateContextRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteContextRequest struct { // Required. The name of the context to delete. Format: // `projects//agent/sessions//contexts/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -387,7 +387,7 @@ func (m *DeleteContextRequest) GetName() string { type DeleteAllContextsRequest struct { // Required. The name of the session to delete all contexts from. Format: // `projects//agent/sessions/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -523,8 +523,7 @@ func (c *contextsClient) DeleteAllContexts(ctx context.Context, in *DeleteAllCon return out, nil } -// Server API for Contexts service - +// ContextsServer is the server API for Contexts service. type ContextsServer interface { // Returns the list of all contexts in the specified session. ListContexts(context.Context, *ListContextsRequest) (*ListContextsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/entity_type.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/entity_type.pb.go index 0dd2950bc..2f52867f4 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/entity_type.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/entity_type.pb.go @@ -98,16 +98,16 @@ type EntityType struct { // automatically. // The unique identifier of the entity type. Format: // `projects//agent/entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of the entity. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. Indicates the kind of entity type. - Kind EntityType_Kind `protobuf:"varint,3,opt,name=kind,enum=google.cloud.dialogflow.v2.EntityType_Kind" json:"kind,omitempty"` + Kind EntityType_Kind `protobuf:"varint,3,opt,name=kind,proto3,enum=google.cloud.dialogflow.v2.EntityType_Kind" json:"kind,omitempty"` // Optional. Indicates whether the entity type can be automatically // expanded. - AutoExpansionMode EntityType_AutoExpansionMode `protobuf:"varint,4,opt,name=auto_expansion_mode,json=autoExpansionMode,enum=google.cloud.dialogflow.v2.EntityType_AutoExpansionMode" json:"auto_expansion_mode,omitempty"` + AutoExpansionMode EntityType_AutoExpansionMode `protobuf:"varint,4,opt,name=auto_expansion_mode,json=autoExpansionMode,proto3,enum=google.cloud.dialogflow.v2.EntityType_AutoExpansionMode" json:"auto_expansion_mode,omitempty"` // Optional. The collection of entities associated with the entity type. - Entities []*EntityType_Entity `protobuf:"bytes,6,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,6,rep,name=entities,proto3" json:"entities,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -180,10 +180,10 @@ type EntityType_Entity struct { // For `KIND_LIST` entity types: // A string that can contain references to other entity types (with or // without aliases). - Value string `protobuf:"bytes,1,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` // Required. A collection of synonyms. For `KIND_LIST` entity types this // must contain exactly one synonym equal to `value`. - Synonyms []string `protobuf:"bytes,2,rep,name=synonyms" json:"synonyms,omitempty"` + Synonyms []string `protobuf:"bytes,2,rep,name=synonyms,proto3" json:"synonyms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -231,18 +231,18 @@ func (m *EntityType_Entity) GetSynonyms() []string { type ListEntityTypesRequest struct { // Required. The agent to list all entity types from. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The language to list entity synonyms for. If not specified, // the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -304,10 +304,10 @@ func (m *ListEntityTypesRequest) GetPageToken() string { type ListEntityTypesResponse struct { // The list of agent entity types. There will be a maximum number of items // returned based on the page_size field in the request. - EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes" json:"entity_types,omitempty"` + EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes,proto3" json:"entity_types,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -355,13 +355,13 @@ func (m *ListEntityTypesResponse) GetNextPageToken() string { type GetEntityTypeRequest struct { // Required. The name of the entity type. // Format: `projects//agent/entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The language to retrieve entity synonyms for. If not specified, // the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,15 +409,15 @@ func (m *GetEntityTypeRequest) GetLanguageCode() string { type CreateEntityTypeRequest struct { // Required. The agent to create a entity type for. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The entity type to create. - EntityType *EntityType `protobuf:"bytes,2,opt,name=entity_type,json=entityType" json:"entity_type,omitempty"` + EntityType *EntityType `protobuf:"bytes,2,opt,name=entity_type,json=entityType,proto3" json:"entity_type,omitempty"` // Optional. The language of entity synonyms defined in `entity_type`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -472,15 +472,15 @@ func (m *CreateEntityTypeRequest) GetLanguageCode() string { type UpdateEntityTypeRequest struct { // Required. The entity type to update. // Format: `projects//agent/entityTypes/`. - EntityType *EntityType `protobuf:"bytes,1,opt,name=entity_type,json=entityType" json:"entity_type,omitempty"` + EntityType *EntityType `protobuf:"bytes,1,opt,name=entity_type,json=entityType,proto3" json:"entity_type,omitempty"` // Optional. The language of entity synonyms defined in `entity_type`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -535,7 +535,7 @@ func (m *UpdateEntityTypeRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteEntityTypeRequest struct { // Required. The name of the entity type to delete. // Format: `projects//agent/entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -576,7 +576,7 @@ func (m *DeleteEntityTypeRequest) GetName() string { type BatchUpdateEntityTypesRequest struct { // Required. The name of the agent to update or create entity types in. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The source of the entity type batch. // // For each entity type in the batch: @@ -592,9 +592,9 @@ type BatchUpdateEntityTypesRequest struct { // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -629,10 +629,10 @@ type isBatchUpdateEntityTypesRequest_EntityTypeBatch interface { } type BatchUpdateEntityTypesRequest_EntityTypeBatchUri struct { - EntityTypeBatchUri string `protobuf:"bytes,2,opt,name=entity_type_batch_uri,json=entityTypeBatchUri,oneof"` + EntityTypeBatchUri string `protobuf:"bytes,2,opt,name=entity_type_batch_uri,json=entityTypeBatchUri,proto3,oneof"` } type BatchUpdateEntityTypesRequest_EntityTypeBatchInline struct { - EntityTypeBatchInline *EntityTypeBatch `protobuf:"bytes,3,opt,name=entity_type_batch_inline,json=entityTypeBatchInline,oneof"` + EntityTypeBatchInline *EntityTypeBatch `protobuf:"bytes,3,opt,name=entity_type_batch_inline,json=entityTypeBatchInline,proto3,oneof"` } func (*BatchUpdateEntityTypesRequest_EntityTypeBatchUri) isBatchUpdateEntityTypesRequest_EntityTypeBatch() { @@ -755,7 +755,7 @@ func _BatchUpdateEntityTypesRequest_OneofSizer(msg proto.Message) (n int) { // The response message for [EntityTypes.BatchUpdateEntityTypes][google.cloud.dialogflow.v2.EntityTypes.BatchUpdateEntityTypes]. type BatchUpdateEntityTypesResponse struct { // The collection of updated or created entity types. - EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes" json:"entity_types,omitempty"` + EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes,proto3" json:"entity_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -796,10 +796,10 @@ func (m *BatchUpdateEntityTypesResponse) GetEntityTypes() []*EntityType { type BatchDeleteEntityTypesRequest struct { // Required. The name of the agent to delete all entities types for. Format: // `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The names entity types to delete. All names must point to the // same agent as `parent`. - EntityTypeNames []string `protobuf:"bytes,2,rep,name=entity_type_names,json=entityTypeNames" json:"entity_type_names,omitempty"` + EntityTypeNames []string `protobuf:"bytes,2,rep,name=entity_type_names,json=entityTypeNames,proto3" json:"entity_type_names,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -847,15 +847,15 @@ func (m *BatchDeleteEntityTypesRequest) GetEntityTypeNames() []string { type BatchCreateEntitiesRequest struct { // Required. The name of the entity type to create entities in. Format: // `projects//agent/entityTypes/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The collection of entities to create. - Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities,proto3" json:"entities,omitempty"` // Optional. The language of entity synonyms defined in `entities`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -910,17 +910,17 @@ func (m *BatchCreateEntitiesRequest) GetLanguageCode() string { type BatchUpdateEntitiesRequest struct { // Required. The name of the entity type to update the entities in. Format: // `projects//agent/entityTypes/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The collection of new entities to replace the existing entities. - Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities,proto3" json:"entities,omitempty"` // Optional. The language of entity synonyms defined in `entities`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -982,17 +982,17 @@ func (m *BatchUpdateEntitiesRequest) GetUpdateMask() *field_mask.FieldMask { type BatchDeleteEntitiesRequest struct { // Required. The name of the entity type to delete entries for. Format: // `projects//agent/entityTypes/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The canonical `values` of the entities to delete. Note that // these are not fully-qualified names, i.e. they don't start with // `projects/`. - EntityValues []string `protobuf:"bytes,2,rep,name=entity_values,json=entityValues" json:"entity_values,omitempty"` + EntityValues []string `protobuf:"bytes,2,rep,name=entity_values,json=entityValues,proto3" json:"entity_values,omitempty"` // Optional. The language of entity synonyms defined in `entities`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1046,7 +1046,7 @@ func (m *BatchDeleteEntitiesRequest) GetLanguageCode() string { // This message is a wrapper around a collection of entity types. type EntityTypeBatch struct { // A collection of entity types. - EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes" json:"entity_types,omitempty"` + EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes,proto3" json:"entity_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1251,8 +1251,7 @@ func (c *entityTypesClient) BatchDeleteEntities(ctx context.Context, in *BatchDe return out, nil } -// Server API for EntityTypes service - +// EntityTypesServer is the server API for EntityTypes service. type EntityTypesServer interface { // Returns the list of all entity types in the specified agent. ListEntityTypes(context.Context, *ListEntityTypesRequest) (*ListEntityTypesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/intent.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/intent.pb.go index 7be8bce61..a53ab16c7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/intent.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/intent.pb.go @@ -241,63 +241,63 @@ type Intent struct { // automatically. // The unique identifier of this intent. // Format: `projects//agent/intents/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of this intent. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. Indicates whether webhooks are enabled for the intent. - WebhookState Intent_WebhookState `protobuf:"varint,6,opt,name=webhook_state,json=webhookState,enum=google.cloud.dialogflow.v2.Intent_WebhookState" json:"webhook_state,omitempty"` + WebhookState Intent_WebhookState `protobuf:"varint,6,opt,name=webhook_state,json=webhookState,proto3,enum=google.cloud.dialogflow.v2.Intent_WebhookState" json:"webhook_state,omitempty"` // Optional. The priority of this intent. Higher numbers represent higher // priorities. Zero or negative numbers mean that the intent is disabled. - Priority int32 `protobuf:"varint,3,opt,name=priority" json:"priority,omitempty"` + Priority int32 `protobuf:"varint,3,opt,name=priority,proto3" json:"priority,omitempty"` // Optional. Indicates whether this is a fallback intent. - IsFallback bool `protobuf:"varint,4,opt,name=is_fallback,json=isFallback" json:"is_fallback,omitempty"` + IsFallback bool `protobuf:"varint,4,opt,name=is_fallback,json=isFallback,proto3" json:"is_fallback,omitempty"` // Optional. Indicates whether Machine Learning is disabled for the intent. // Note: If `ml_diabled` setting is set to true, then this intent is not // taken into account during inference in `ML ONLY` match mode. Also, // auto-markup in the UI is turned off. - MlDisabled bool `protobuf:"varint,19,opt,name=ml_disabled,json=mlDisabled" json:"ml_disabled,omitempty"` + MlDisabled bool `protobuf:"varint,19,opt,name=ml_disabled,json=mlDisabled,proto3" json:"ml_disabled,omitempty"` // Optional. The list of context names required for this intent to be // triggered. // Format: `projects//agent/sessions/-/contexts/`. - InputContextNames []string `protobuf:"bytes,7,rep,name=input_context_names,json=inputContextNames" json:"input_context_names,omitempty"` + InputContextNames []string `protobuf:"bytes,7,rep,name=input_context_names,json=inputContextNames,proto3" json:"input_context_names,omitempty"` // Optional. The collection of event names that trigger the intent. // If the collection of input contexts is not empty, all of the contexts must // be present in the active user session for an event to trigger this intent. - Events []string `protobuf:"bytes,8,rep,name=events" json:"events,omitempty"` + Events []string `protobuf:"bytes,8,rep,name=events,proto3" json:"events,omitempty"` // Optional. The collection of examples/templates that the agent is // trained on. - TrainingPhrases []*Intent_TrainingPhrase `protobuf:"bytes,9,rep,name=training_phrases,json=trainingPhrases" json:"training_phrases,omitempty"` + TrainingPhrases []*Intent_TrainingPhrase `protobuf:"bytes,9,rep,name=training_phrases,json=trainingPhrases,proto3" json:"training_phrases,omitempty"` // Optional. The name of the action associated with the intent. - Action string `protobuf:"bytes,10,opt,name=action" json:"action,omitempty"` + Action string `protobuf:"bytes,10,opt,name=action,proto3" json:"action,omitempty"` // Optional. The collection of contexts that are activated when the intent // is matched. Context messages in this collection should not set the // parameters field. Setting the `lifespan_count` to 0 will reset the context // when the intent is matched. // Format: `projects//agent/sessions/-/contexts/`. - OutputContexts []*Context `protobuf:"bytes,11,rep,name=output_contexts,json=outputContexts" json:"output_contexts,omitempty"` + OutputContexts []*Context `protobuf:"bytes,11,rep,name=output_contexts,json=outputContexts,proto3" json:"output_contexts,omitempty"` // Optional. Indicates whether to delete all contexts in the current // session when this intent is matched. - ResetContexts bool `protobuf:"varint,12,opt,name=reset_contexts,json=resetContexts" json:"reset_contexts,omitempty"` + ResetContexts bool `protobuf:"varint,12,opt,name=reset_contexts,json=resetContexts,proto3" json:"reset_contexts,omitempty"` // Optional. The collection of parameters associated with the intent. - Parameters []*Intent_Parameter `protobuf:"bytes,13,rep,name=parameters" json:"parameters,omitempty"` + Parameters []*Intent_Parameter `protobuf:"bytes,13,rep,name=parameters,proto3" json:"parameters,omitempty"` // Optional. The collection of rich messages corresponding to the // `Response` field in the Dialogflow console. - Messages []*Intent_Message `protobuf:"bytes,14,rep,name=messages" json:"messages,omitempty"` + Messages []*Intent_Message `protobuf:"bytes,14,rep,name=messages,proto3" json:"messages,omitempty"` // Optional. The list of platforms for which the first response will be // taken from among the messages assigned to the DEFAULT_PLATFORM. - DefaultResponsePlatforms []Intent_Message_Platform `protobuf:"varint,15,rep,packed,name=default_response_platforms,json=defaultResponsePlatforms,enum=google.cloud.dialogflow.v2.Intent_Message_Platform" json:"default_response_platforms,omitempty"` + DefaultResponsePlatforms []Intent_Message_Platform `protobuf:"varint,15,rep,packed,name=default_response_platforms,json=defaultResponsePlatforms,proto3,enum=google.cloud.dialogflow.v2.Intent_Message_Platform" json:"default_response_platforms,omitempty"` // The unique identifier of the root intent in the chain of followup intents. // It identifies the correct followup intents chain for this intent. // Format: `projects//agent/intents/`. - RootFollowupIntentName string `protobuf:"bytes,16,opt,name=root_followup_intent_name,json=rootFollowupIntentName" json:"root_followup_intent_name,omitempty"` + RootFollowupIntentName string `protobuf:"bytes,16,opt,name=root_followup_intent_name,json=rootFollowupIntentName,proto3" json:"root_followup_intent_name,omitempty"` // The unique identifier of the parent intent in the chain of followup // intents. // It identifies the parent followup intent. // Format: `projects//agent/intents/`. - ParentFollowupIntentName string `protobuf:"bytes,17,opt,name=parent_followup_intent_name,json=parentFollowupIntentName" json:"parent_followup_intent_name,omitempty"` + ParentFollowupIntentName string `protobuf:"bytes,17,opt,name=parent_followup_intent_name,json=parentFollowupIntentName,proto3" json:"parent_followup_intent_name,omitempty"` // Optional. Collection of information about all followup intents that have // name of this intent as a root_name. - FollowupIntentInfo []*Intent_FollowupIntentInfo `protobuf:"bytes,18,rep,name=followup_intent_info,json=followupIntentInfo" json:"followup_intent_info,omitempty"` + FollowupIntentInfo []*Intent_FollowupIntentInfo `protobuf:"bytes,18,rep,name=followup_intent_info,json=followupIntentInfo,proto3" json:"followup_intent_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -456,17 +456,17 @@ func (m *Intent) GetFollowupIntentInfo() []*Intent_FollowupIntentInfo { // Represents an example or template that the agent is trained on. type Intent_TrainingPhrase struct { // Required. The unique identifier of this training phrase. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The type of the training phrase. - Type Intent_TrainingPhrase_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.dialogflow.v2.Intent_TrainingPhrase_Type" json:"type,omitempty"` + Type Intent_TrainingPhrase_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.dialogflow.v2.Intent_TrainingPhrase_Type" json:"type,omitempty"` // Required. The collection of training phrase parts (can be annotated). // Fields: `entity_type`, `alias` and `user_defined` should be populated // only for the annotated parts of the training phrase. - Parts []*Intent_TrainingPhrase_Part `protobuf:"bytes,3,rep,name=parts" json:"parts,omitempty"` + Parts []*Intent_TrainingPhrase_Part `protobuf:"bytes,3,rep,name=parts,proto3" json:"parts,omitempty"` // Optional. Indicates how many times this example or template was added to // the intent. Each time a developer adds an existing sample by editing an // intent or training, this counter is increased. - TimesAddedCount int32 `protobuf:"varint,4,opt,name=times_added_count,json=timesAddedCount" json:"times_added_count,omitempty"` + TimesAddedCount int32 `protobuf:"varint,4,opt,name=times_added_count,json=timesAddedCount,proto3" json:"times_added_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -529,17 +529,17 @@ type Intent_TrainingPhrase_Part struct { // Required. The text corresponding to the example or template, // if there are no annotations. For // annotated examples, it is the text for one of the example's parts. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Optional. The entity type name prefixed with `@`. This field is // required for the annotated part of the text and applies only to // examples. - EntityType string `protobuf:"bytes,2,opt,name=entity_type,json=entityType" json:"entity_type,omitempty"` + EntityType string `protobuf:"bytes,2,opt,name=entity_type,json=entityType,proto3" json:"entity_type,omitempty"` // Optional. The parameter name for the value extracted from the // annotated part of the example. - Alias string `protobuf:"bytes,3,opt,name=alias" json:"alias,omitempty"` + Alias string `protobuf:"bytes,3,opt,name=alias,proto3" json:"alias,omitempty"` // Optional. Indicates whether the text was manually annotated by the // developer. - UserDefined bool `protobuf:"varint,4,opt,name=user_defined,json=userDefined" json:"user_defined,omitempty"` + UserDefined bool `protobuf:"varint,4,opt,name=user_defined,json=userDefined,proto3" json:"user_defined,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -600,34 +600,34 @@ func (m *Intent_TrainingPhrase_Part) GetUserDefined() bool { // Represents intent parameters. type Intent_Parameter struct { // The unique identifier of this parameter. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of the parameter. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Optional. The definition of the parameter value. It can be: // - a constant string, // - a parameter value defined as `$parameter_name`, // - an original parameter value defined as `$parameter_name.original`, // - a parameter value from some context defined as // `#context_name.parameter_name`. - Value string `protobuf:"bytes,3,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` // Optional. The default value to use when the `value` yields an empty // result. // Default values can be extracted from contexts by using the following // syntax: `#context_name.parameter_name`. - DefaultValue string `protobuf:"bytes,4,opt,name=default_value,json=defaultValue" json:"default_value,omitempty"` + DefaultValue string `protobuf:"bytes,4,opt,name=default_value,json=defaultValue,proto3" json:"default_value,omitempty"` // Optional. The name of the entity type, prefixed with `@`, that // describes values of the parameter. If the parameter is // required, this must be provided. - EntityTypeDisplayName string `protobuf:"bytes,5,opt,name=entity_type_display_name,json=entityTypeDisplayName" json:"entity_type_display_name,omitempty"` + EntityTypeDisplayName string `protobuf:"bytes,5,opt,name=entity_type_display_name,json=entityTypeDisplayName,proto3" json:"entity_type_display_name,omitempty"` // Optional. Indicates whether the parameter is required. That is, // whether the intent cannot be completed without collecting the parameter // value. - Mandatory bool `protobuf:"varint,6,opt,name=mandatory" json:"mandatory,omitempty"` + Mandatory bool `protobuf:"varint,6,opt,name=mandatory,proto3" json:"mandatory,omitempty"` // Optional. The collection of prompts that the agent can present to the // user in order to collect value for the parameter. - Prompts []string `protobuf:"bytes,7,rep,name=prompts" json:"prompts,omitempty"` + Prompts []string `protobuf:"bytes,7,rep,name=prompts,proto3" json:"prompts,omitempty"` // Optional. Indicates whether the parameter represents a list of values. - IsList bool `protobuf:"varint,8,opt,name=is_list,json=isList" json:"is_list,omitempty"` + IsList bool `protobuf:"varint,8,opt,name=is_list,json=isList,proto3" json:"is_list,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -731,7 +731,7 @@ type Intent_Message struct { // *Intent_Message_CarouselSelect_ Message isIntent_Message_Message `protobuf_oneof:"message"` // Optional. The platform that this message is intended for. - Platform Intent_Message_Platform `protobuf:"varint,6,opt,name=platform,enum=google.cloud.dialogflow.v2.Intent_Message_Platform" json:"platform,omitempty"` + Platform Intent_Message_Platform `protobuf:"varint,6,opt,name=platform,proto3,enum=google.cloud.dialogflow.v2.Intent_Message_Platform" json:"platform,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -766,37 +766,37 @@ type isIntent_Message_Message interface { } type Intent_Message_Text_ struct { - Text *Intent_Message_Text `protobuf:"bytes,1,opt,name=text,oneof"` + Text *Intent_Message_Text `protobuf:"bytes,1,opt,name=text,proto3,oneof"` } type Intent_Message_Image_ struct { - Image *Intent_Message_Image `protobuf:"bytes,2,opt,name=image,oneof"` + Image *Intent_Message_Image `protobuf:"bytes,2,opt,name=image,proto3,oneof"` } type Intent_Message_QuickReplies_ struct { - QuickReplies *Intent_Message_QuickReplies `protobuf:"bytes,3,opt,name=quick_replies,json=quickReplies,oneof"` + QuickReplies *Intent_Message_QuickReplies `protobuf:"bytes,3,opt,name=quick_replies,json=quickReplies,proto3,oneof"` } type Intent_Message_Card_ struct { - Card *Intent_Message_Card `protobuf:"bytes,4,opt,name=card,oneof"` + Card *Intent_Message_Card `protobuf:"bytes,4,opt,name=card,proto3,oneof"` } type Intent_Message_Payload struct { - Payload *_struct.Struct `protobuf:"bytes,5,opt,name=payload,oneof"` + Payload *_struct.Struct `protobuf:"bytes,5,opt,name=payload,proto3,oneof"` } type Intent_Message_SimpleResponses_ struct { - SimpleResponses *Intent_Message_SimpleResponses `protobuf:"bytes,7,opt,name=simple_responses,json=simpleResponses,oneof"` + SimpleResponses *Intent_Message_SimpleResponses `protobuf:"bytes,7,opt,name=simple_responses,json=simpleResponses,proto3,oneof"` } type Intent_Message_BasicCard_ struct { - BasicCard *Intent_Message_BasicCard `protobuf:"bytes,8,opt,name=basic_card,json=basicCard,oneof"` + BasicCard *Intent_Message_BasicCard `protobuf:"bytes,8,opt,name=basic_card,json=basicCard,proto3,oneof"` } type Intent_Message_Suggestions_ struct { - Suggestions *Intent_Message_Suggestions `protobuf:"bytes,9,opt,name=suggestions,oneof"` + Suggestions *Intent_Message_Suggestions `protobuf:"bytes,9,opt,name=suggestions,proto3,oneof"` } type Intent_Message_LinkOutSuggestion_ struct { - LinkOutSuggestion *Intent_Message_LinkOutSuggestion `protobuf:"bytes,10,opt,name=link_out_suggestion,json=linkOutSuggestion,oneof"` + LinkOutSuggestion *Intent_Message_LinkOutSuggestion `protobuf:"bytes,10,opt,name=link_out_suggestion,json=linkOutSuggestion,proto3,oneof"` } type Intent_Message_ListSelect_ struct { - ListSelect *Intent_Message_ListSelect `protobuf:"bytes,11,opt,name=list_select,json=listSelect,oneof"` + ListSelect *Intent_Message_ListSelect `protobuf:"bytes,11,opt,name=list_select,json=listSelect,proto3,oneof"` } type Intent_Message_CarouselSelect_ struct { - CarouselSelect *Intent_Message_CarouselSelect `protobuf:"bytes,12,opt,name=carousel_select,json=carouselSelect,oneof"` + CarouselSelect *Intent_Message_CarouselSelect `protobuf:"bytes,12,opt,name=carousel_select,json=carouselSelect,proto3,oneof"` } func (*Intent_Message_Text_) isIntent_Message_Message() {} @@ -1150,7 +1150,7 @@ func _Intent_Message_OneofSizer(msg proto.Message) (n int) { // The text response message. type Intent_Message_Text struct { // Optional. The collection of the agent's responses. - Text []string `protobuf:"bytes,1,rep,name=text" json:"text,omitempty"` + Text []string `protobuf:"bytes,1,rep,name=text,proto3" json:"text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1190,10 +1190,10 @@ func (m *Intent_Message_Text) GetText() []string { // The image response message. type Intent_Message_Image struct { // Optional. The public URI to an image file. - ImageUri string `protobuf:"bytes,1,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,1,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` // Optional. A text description of the image to be used for accessibility, // e.g., screen readers. - AccessibilityText string `protobuf:"bytes,2,opt,name=accessibility_text,json=accessibilityText" json:"accessibility_text,omitempty"` + AccessibilityText string `protobuf:"bytes,2,opt,name=accessibility_text,json=accessibilityText,proto3" json:"accessibility_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1240,9 +1240,9 @@ func (m *Intent_Message_Image) GetAccessibilityText() string { // The quick replies response message. type Intent_Message_QuickReplies struct { // Optional. The title of the collection of quick replies. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Optional. The collection of quick replies. - QuickReplies []string `protobuf:"bytes,2,rep,name=quick_replies,json=quickReplies" json:"quick_replies,omitempty"` + QuickReplies []string `protobuf:"bytes,2,rep,name=quick_replies,json=quickReplies,proto3" json:"quick_replies,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1289,13 +1289,13 @@ func (m *Intent_Message_QuickReplies) GetQuickReplies() []string { // The card response message. type Intent_Message_Card struct { // Optional. The title of the card. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Optional. The subtitle of the card. - Subtitle string `protobuf:"bytes,2,opt,name=subtitle" json:"subtitle,omitempty"` + Subtitle string `protobuf:"bytes,2,opt,name=subtitle,proto3" json:"subtitle,omitempty"` // Optional. The public URI to an image file for the card. - ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` // Optional. The collection of card buttons. - Buttons []*Intent_Message_Card_Button `protobuf:"bytes,4,rep,name=buttons" json:"buttons,omitempty"` + Buttons []*Intent_Message_Card_Button `protobuf:"bytes,4,rep,name=buttons,proto3" json:"buttons,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1356,10 +1356,10 @@ func (m *Intent_Message_Card) GetButtons() []*Intent_Message_Card_Button { // Optional. Contains information about a button. type Intent_Message_Card_Button struct { // Optional. The text to show on the button. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Optional. The text to send back to the Dialogflow API or a URI to // open. - Postback string `protobuf:"bytes,2,opt,name=postback" json:"postback,omitempty"` + Postback string `protobuf:"bytes,2,opt,name=postback,proto3" json:"postback,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1407,13 +1407,13 @@ func (m *Intent_Message_Card_Button) GetPostback() string { type Intent_Message_SimpleResponse struct { // One of text_to_speech or ssml must be provided. The plain text of the // speech output. Mutually exclusive with ssml. - TextToSpeech string `protobuf:"bytes,1,opt,name=text_to_speech,json=textToSpeech" json:"text_to_speech,omitempty"` + TextToSpeech string `protobuf:"bytes,1,opt,name=text_to_speech,json=textToSpeech,proto3" json:"text_to_speech,omitempty"` // One of text_to_speech or ssml must be provided. Structured spoken // response to the user in the SSML format. Mutually exclusive with // text_to_speech. - Ssml string `protobuf:"bytes,2,opt,name=ssml" json:"ssml,omitempty"` + Ssml string `protobuf:"bytes,2,opt,name=ssml,proto3" json:"ssml,omitempty"` // Optional. The text to display. - DisplayText string `protobuf:"bytes,3,opt,name=display_text,json=displayText" json:"display_text,omitempty"` + DisplayText string `protobuf:"bytes,3,opt,name=display_text,json=displayText,proto3" json:"display_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1470,7 +1470,7 @@ func (m *Intent_Message_SimpleResponse) GetDisplayText() string { // `SimpleResponse`. type Intent_Message_SimpleResponses struct { // Required. The list of simple responses. - SimpleResponses []*Intent_Message_SimpleResponse `protobuf:"bytes,1,rep,name=simple_responses,json=simpleResponses" json:"simple_responses,omitempty"` + SimpleResponses []*Intent_Message_SimpleResponse `protobuf:"bytes,1,rep,name=simple_responses,json=simpleResponses,proto3" json:"simple_responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1510,15 +1510,15 @@ func (m *Intent_Message_SimpleResponses) GetSimpleResponses() []*Intent_Message_ // The basic card message. Useful for displaying information. type Intent_Message_BasicCard struct { // Optional. The title of the card. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Optional. The subtitle of the card. - Subtitle string `protobuf:"bytes,2,opt,name=subtitle" json:"subtitle,omitempty"` + Subtitle string `protobuf:"bytes,2,opt,name=subtitle,proto3" json:"subtitle,omitempty"` // Required, unless image is present. The body text of the card. - FormattedText string `protobuf:"bytes,3,opt,name=formatted_text,json=formattedText" json:"formatted_text,omitempty"` + FormattedText string `protobuf:"bytes,3,opt,name=formatted_text,json=formattedText,proto3" json:"formatted_text,omitempty"` // Optional. The image for the card. - Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"` + Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image,proto3" json:"image,omitempty"` // Optional. The collection of card buttons. - Buttons []*Intent_Message_BasicCard_Button `protobuf:"bytes,5,rep,name=buttons" json:"buttons,omitempty"` + Buttons []*Intent_Message_BasicCard_Button `protobuf:"bytes,5,rep,name=buttons,proto3" json:"buttons,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1586,9 +1586,9 @@ func (m *Intent_Message_BasicCard) GetButtons() []*Intent_Message_BasicCard_Butt // The button object that appears at the bottom of a card. type Intent_Message_BasicCard_Button struct { // Required. The title of the button. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Required. Action to take when a user taps on the button. - OpenUriAction *Intent_Message_BasicCard_Button_OpenUriAction `protobuf:"bytes,2,opt,name=open_uri_action,json=openUriAction" json:"open_uri_action,omitempty"` + OpenUriAction *Intent_Message_BasicCard_Button_OpenUriAction `protobuf:"bytes,2,opt,name=open_uri_action,json=openUriAction,proto3" json:"open_uri_action,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1635,7 +1635,7 @@ func (m *Intent_Message_BasicCard_Button) GetOpenUriAction() *Intent_Message_Bas // Opens the given URI. type Intent_Message_BasicCard_Button_OpenUriAction struct { // Required. The HTTP or HTTPS scheme URI. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1680,7 +1680,7 @@ func (m *Intent_Message_BasicCard_Button_OpenUriAction) GetUri() string { // to the conversation. type Intent_Message_Suggestion struct { // Required. The text shown the in the suggestion chip. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1720,7 +1720,7 @@ func (m *Intent_Message_Suggestion) GetTitle() string { // The collection of suggestions. type Intent_Message_Suggestions struct { // Required. The list of suggested replies. - Suggestions []*Intent_Message_Suggestion `protobuf:"bytes,1,rep,name=suggestions" json:"suggestions,omitempty"` + Suggestions []*Intent_Message_Suggestion `protobuf:"bytes,1,rep,name=suggestions,proto3" json:"suggestions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1761,10 +1761,10 @@ func (m *Intent_Message_Suggestions) GetSuggestions() []*Intent_Message_Suggesti // or website associated with this agent. type Intent_Message_LinkOutSuggestion struct { // Required. The name of the app or site this chip is linking to. - DestinationName string `protobuf:"bytes,1,opt,name=destination_name,json=destinationName" json:"destination_name,omitempty"` + DestinationName string `protobuf:"bytes,1,opt,name=destination_name,json=destinationName,proto3" json:"destination_name,omitempty"` // Required. The URI of the app or site to open when the user taps the // suggestion chip. - Uri string `protobuf:"bytes,2,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,2,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1811,9 +1811,9 @@ func (m *Intent_Message_LinkOutSuggestion) GetUri() string { // The card for presenting a list of options to select from. type Intent_Message_ListSelect struct { // Optional. The overall title of the list. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Required. List items. - Items []*Intent_Message_ListSelect_Item `protobuf:"bytes,2,rep,name=items" json:"items,omitempty"` + Items []*Intent_Message_ListSelect_Item `protobuf:"bytes,2,rep,name=items,proto3" json:"items,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1860,13 +1860,13 @@ func (m *Intent_Message_ListSelect) GetItems() []*Intent_Message_ListSelect_Item // An item in the list. type Intent_Message_ListSelect_Item struct { // Required. Additional information about this option. - Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info" json:"info,omitempty"` + Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info,proto3" json:"info,omitempty"` // Required. The title of the list item. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // Optional. The main text describing the item. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Optional. The image to display. - Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"` + Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image,proto3" json:"image,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1927,7 +1927,7 @@ func (m *Intent_Message_ListSelect_Item) GetImage() *Intent_Message_Image { // The card for presenting a carousel of options to select from. type Intent_Message_CarouselSelect struct { // Required. Carousel items. - Items []*Intent_Message_CarouselSelect_Item `protobuf:"bytes,1,rep,name=items" json:"items,omitempty"` + Items []*Intent_Message_CarouselSelect_Item `protobuf:"bytes,1,rep,name=items,proto3" json:"items,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1967,13 +1967,13 @@ func (m *Intent_Message_CarouselSelect) GetItems() []*Intent_Message_CarouselSel // An item in the carousel. type Intent_Message_CarouselSelect_Item struct { // Required. Additional info about the option item. - Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info" json:"info,omitempty"` + Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info,proto3" json:"info,omitempty"` // Required. Title of the carousel item. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // Optional. The body text of the card. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Optional. The image to display. - Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"` + Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image,proto3" json:"image,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2036,10 +2036,10 @@ func (m *Intent_Message_CarouselSelect_Item) GetImage() *Intent_Message_Image { type Intent_Message_SelectItemInfo struct { // Required. A unique key that will be sent back to the agent if this // response is given. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // Optional. A list of synonyms that can also be used to trigger this // item in dialog. - Synonyms []string `protobuf:"bytes,2,rep,name=synonyms" json:"synonyms,omitempty"` + Synonyms []string `protobuf:"bytes,2,rep,name=synonyms,proto3" json:"synonyms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2087,10 +2087,10 @@ func (m *Intent_Message_SelectItemInfo) GetSynonyms() []string { type Intent_FollowupIntentInfo struct { // The unique identifier of the followup intent. // Format: `projects//agent/intents/`. - FollowupIntentName string `protobuf:"bytes,1,opt,name=followup_intent_name,json=followupIntentName" json:"followup_intent_name,omitempty"` + FollowupIntentName string `protobuf:"bytes,1,opt,name=followup_intent_name,json=followupIntentName,proto3" json:"followup_intent_name,omitempty"` // The unique identifier of the followup intent parent. // Format: `projects//agent/intents/`. - ParentFollowupIntentName string `protobuf:"bytes,2,opt,name=parent_followup_intent_name,json=parentFollowupIntentName" json:"parent_followup_intent_name,omitempty"` + ParentFollowupIntentName string `protobuf:"bytes,2,opt,name=parent_followup_intent_name,json=parentFollowupIntentName,proto3" json:"parent_followup_intent_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2138,20 +2138,20 @@ func (m *Intent_FollowupIntentInfo) GetParentFollowupIntentName() string { type ListIntentsRequest struct { // Required. The agent to list all intents from. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The language to list training phrases, parameters and rich // messages for. If not specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2220,10 +2220,10 @@ func (m *ListIntentsRequest) GetPageToken() string { type ListIntentsResponse struct { // The list of agent intents. There will be a maximum number of items // returned based on the page_size field in the request. - Intents []*Intent `protobuf:"bytes,1,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,1,rep,name=intents,proto3" json:"intents,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2271,15 +2271,15 @@ func (m *ListIntentsResponse) GetNextPageToken() string { type GetIntentRequest struct { // Required. The name of the intent. // Format: `projects//agent/intents/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The language to retrieve training phrases, parameters and rich // messages for. If not specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2334,17 +2334,17 @@ func (m *GetIntentRequest) GetIntentView() IntentView { type CreateIntentRequest struct { // Required. The agent to create a intent for. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The intent to create. - Intent *Intent `protobuf:"bytes,2,opt,name=intent" json:"intent,omitempty"` + Intent *Intent `protobuf:"bytes,2,opt,name=intent,proto3" json:"intent,omitempty"` // Optional. The language of training phrases, parameters and rich messages // defined in `intent`. If not specified, the agent's default language is // used. [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2406,17 +2406,17 @@ func (m *CreateIntentRequest) GetIntentView() IntentView { type UpdateIntentRequest struct { // Required. The intent to update. // Format: `projects//agent/intents/`. - Intent *Intent `protobuf:"bytes,1,opt,name=intent" json:"intent,omitempty"` + Intent *Intent `protobuf:"bytes,1,opt,name=intent,proto3" json:"intent,omitempty"` // Optional. The language of training phrases, parameters and rich messages // defined in `intent`. If not specified, the agent's default language is // used. [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2478,7 +2478,7 @@ func (m *UpdateIntentRequest) GetIntentView() IntentView { type DeleteIntentRequest struct { // Required. The name of the intent to delete. // Format: `projects//agent/intents/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2519,7 +2519,7 @@ func (m *DeleteIntentRequest) GetName() string { type BatchUpdateIntentsRequest struct { // Required. The name of the agent to update or create intents in. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The source of the intent batch. // // Types that are valid to be assigned to IntentBatch: @@ -2531,11 +2531,11 @@ type BatchUpdateIntentsRequest struct { // used. [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,6,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,6,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2570,10 +2570,10 @@ type isBatchUpdateIntentsRequest_IntentBatch interface { } type BatchUpdateIntentsRequest_IntentBatchUri struct { - IntentBatchUri string `protobuf:"bytes,2,opt,name=intent_batch_uri,json=intentBatchUri,oneof"` + IntentBatchUri string `protobuf:"bytes,2,opt,name=intent_batch_uri,json=intentBatchUri,proto3,oneof"` } type BatchUpdateIntentsRequest_IntentBatchInline struct { - IntentBatchInline *IntentBatch `protobuf:"bytes,3,opt,name=intent_batch_inline,json=intentBatchInline,oneof"` + IntentBatchInline *IntentBatch `protobuf:"bytes,3,opt,name=intent_batch_inline,json=intentBatchInline,proto3,oneof"` } func (*BatchUpdateIntentsRequest_IntentBatchUri) isBatchUpdateIntentsRequest_IntentBatch() {} @@ -2701,7 +2701,7 @@ func _BatchUpdateIntentsRequest_OneofSizer(msg proto.Message) (n int) { // The response message for [Intents.BatchUpdateIntents][google.cloud.dialogflow.v2.Intents.BatchUpdateIntents]. type BatchUpdateIntentsResponse struct { // The collection of updated or created intents. - Intents []*Intent `protobuf:"bytes,1,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,1,rep,name=intents,proto3" json:"intents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2742,10 +2742,10 @@ func (m *BatchUpdateIntentsResponse) GetIntents() []*Intent { type BatchDeleteIntentsRequest struct { // Required. The name of the agent to delete all entities types for. Format: // `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The collection of intents to delete. Only intent `name` must be // filled in. - Intents []*Intent `protobuf:"bytes,2,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,2,rep,name=intents,proto3" json:"intents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2792,7 +2792,7 @@ func (m *BatchDeleteIntentsRequest) GetIntents() []*Intent { // This message is a wrapper around a collection of intents. type IntentBatch struct { // A collection of intents. - Intents []*Intent `protobuf:"bytes,1,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,1,rep,name=intents,proto3" json:"intents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2973,8 +2973,7 @@ func (c *intentsClient) BatchDeleteIntents(ctx context.Context, in *BatchDeleteI return out, nil } -// Server API for Intents service - +// IntentsServer is the server API for Intents service. type IntentsServer interface { // Returns the list of all intents in the specified agent. ListIntents(context.Context, *ListIntentsRequest) (*ListIntentsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session.pb.go index f6fc5fef7..07e8e18a2 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session.pb.go @@ -140,9 +140,9 @@ type DetectIntentRequest struct { // caller to choose an appropriate session ID. It can be a random number or // some type of user identifier (preferably hashed). The length of the session // ID must not exceed 36 bytes. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Optional. The parameters of this query. - QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams" json:"query_params,omitempty"` + QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams,proto3" json:"query_params,omitempty"` // Required. The input specification. It can be set to: // // 1. an audio config @@ -151,7 +151,7 @@ type DetectIntentRequest struct { // 2. a conversational query in the form of text, or // // 3. an event that specifies which intent to trigger. - QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput" json:"query_input,omitempty"` + QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput,proto3" json:"query_input,omitempty"` // Optional. The natural language speech audio to be processed. This field // should be populated iff `query_input` is set to an input audio config. // A single request can contain up to 1 minute of speech audio data. @@ -217,12 +217,12 @@ func (m *DetectIntentRequest) GetInputAudio() []byte { type DetectIntentResponse struct { // The unique identifier of the response. It can be used to // locate a response in the training example set or for reporting issues. - ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId" json:"response_id,omitempty"` + ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId,proto3" json:"response_id,omitempty"` // The results of the conversational query or event processing. - QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult" json:"query_result,omitempty"` + QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult,proto3" json:"query_result,omitempty"` // Specifies the status of the webhook request. `webhook_status` // is never populated in webhook requests. - WebhookStatus *status.Status `protobuf:"bytes,3,opt,name=webhook_status,json=webhookStatus" json:"webhook_status,omitempty"` + WebhookStatus *status.Status `protobuf:"bytes,3,opt,name=webhook_status,json=webhookStatus,proto3" json:"webhook_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -279,22 +279,22 @@ type QueryParameters struct { // [time zone database](https://www.iana.org/time-zones), e.g., // America/New_York, Europe/Paris. If not provided, the time zone specified in // agent settings is used. - TimeZone string `protobuf:"bytes,1,opt,name=time_zone,json=timeZone" json:"time_zone,omitempty"` + TimeZone string `protobuf:"bytes,1,opt,name=time_zone,json=timeZone,proto3" json:"time_zone,omitempty"` // Optional. The geo location of this conversational query. - GeoLocation *latlng.LatLng `protobuf:"bytes,2,opt,name=geo_location,json=geoLocation" json:"geo_location,omitempty"` + GeoLocation *latlng.LatLng `protobuf:"bytes,2,opt,name=geo_location,json=geoLocation,proto3" json:"geo_location,omitempty"` // Optional. The collection of contexts to be activated before this query is // executed. - Contexts []*Context `protobuf:"bytes,3,rep,name=contexts" json:"contexts,omitempty"` + Contexts []*Context `protobuf:"bytes,3,rep,name=contexts,proto3" json:"contexts,omitempty"` // Optional. Specifies whether to delete all contexts in the current session // before the new ones are activated. - ResetContexts bool `protobuf:"varint,4,opt,name=reset_contexts,json=resetContexts" json:"reset_contexts,omitempty"` + ResetContexts bool `protobuf:"varint,4,opt,name=reset_contexts,json=resetContexts,proto3" json:"reset_contexts,omitempty"` // Optional. The collection of session entity types to replace or extend // developer entities with for this query only. The entity synonyms apply // to all languages. - SessionEntityTypes []*SessionEntityType `protobuf:"bytes,5,rep,name=session_entity_types,json=sessionEntityTypes" json:"session_entity_types,omitempty"` + SessionEntityTypes []*SessionEntityType `protobuf:"bytes,5,rep,name=session_entity_types,json=sessionEntityTypes,proto3" json:"session_entity_types,omitempty"` // Optional. This field can be used to pass custom data into the webhook // associated with the agent. Arbitrary JSON objects are supported. - Payload *_struct.Struct `protobuf:"bytes,6,opt,name=payload" json:"payload,omitempty"` + Payload *_struct.Struct `protobuf:"bytes,6,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -416,13 +416,13 @@ type isQueryInput_Input interface { } type QueryInput_AudioConfig struct { - AudioConfig *InputAudioConfig `protobuf:"bytes,1,opt,name=audio_config,json=audioConfig,oneof"` + AudioConfig *InputAudioConfig `protobuf:"bytes,1,opt,name=audio_config,json=audioConfig,proto3,oneof"` } type QueryInput_Text struct { - Text *TextInput `protobuf:"bytes,2,opt,name=text,oneof"` + Text *TextInput `protobuf:"bytes,2,opt,name=text,proto3,oneof"` } type QueryInput_Event struct { - Event *EventInput `protobuf:"bytes,3,opt,name=event,oneof"` + Event *EventInput `protobuf:"bytes,3,opt,name=event,proto3,oneof"` } func (*QueryInput_AudioConfig) isQueryInput_Input() {} @@ -559,11 +559,11 @@ type QueryResult struct { // contains the speech recognition result. If speech recognizer produced // multiple alternatives, a particular one is picked. // - If an event was provided as input, `query_text` is not set. - QueryText string `protobuf:"bytes,1,opt,name=query_text,json=queryText" json:"query_text,omitempty"` + QueryText string `protobuf:"bytes,1,opt,name=query_text,json=queryText,proto3" json:"query_text,omitempty"` // The language that was triggered during intent detection. // See [Language Support](https://dialogflow.com/docs/reference/language) // for a list of the currently supported language codes. - LanguageCode string `protobuf:"bytes,15,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,15,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // The Speech recognition confidence between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. The default of 0.0 is a sentinel value indicating that confidence @@ -573,42 +573,42 @@ type QueryResult struct { // even set. In particular this field isn't set in Webhook calls and for // StreamingDetectIntent since the streaming endpoint has separate confidence // estimates per portion of the audio in StreamingRecognitionResult. - SpeechRecognitionConfidence float32 `protobuf:"fixed32,2,opt,name=speech_recognition_confidence,json=speechRecognitionConfidence" json:"speech_recognition_confidence,omitempty"` + SpeechRecognitionConfidence float32 `protobuf:"fixed32,2,opt,name=speech_recognition_confidence,json=speechRecognitionConfidence,proto3" json:"speech_recognition_confidence,omitempty"` // The action name from the matched intent. - Action string `protobuf:"bytes,3,opt,name=action" json:"action,omitempty"` + Action string `protobuf:"bytes,3,opt,name=action,proto3" json:"action,omitempty"` // The collection of extracted parameters. - Parameters *_struct.Struct `protobuf:"bytes,4,opt,name=parameters" json:"parameters,omitempty"` + Parameters *_struct.Struct `protobuf:"bytes,4,opt,name=parameters,proto3" json:"parameters,omitempty"` // This field is set to: // - `false` if the matched intent has required parameters and not all of // the required parameter values have been collected. // - `true` if all required parameter values have been collected, or if the // matched intent doesn't contain any required parameters. - AllRequiredParamsPresent bool `protobuf:"varint,5,opt,name=all_required_params_present,json=allRequiredParamsPresent" json:"all_required_params_present,omitempty"` + AllRequiredParamsPresent bool `protobuf:"varint,5,opt,name=all_required_params_present,json=allRequiredParamsPresent,proto3" json:"all_required_params_present,omitempty"` // The text to be pronounced to the user or shown on the screen. - FulfillmentText string `protobuf:"bytes,6,opt,name=fulfillment_text,json=fulfillmentText" json:"fulfillment_text,omitempty"` + FulfillmentText string `protobuf:"bytes,6,opt,name=fulfillment_text,json=fulfillmentText,proto3" json:"fulfillment_text,omitempty"` // The collection of rich messages to present to the user. - FulfillmentMessages []*Intent_Message `protobuf:"bytes,7,rep,name=fulfillment_messages,json=fulfillmentMessages" json:"fulfillment_messages,omitempty"` + FulfillmentMessages []*Intent_Message `protobuf:"bytes,7,rep,name=fulfillment_messages,json=fulfillmentMessages,proto3" json:"fulfillment_messages,omitempty"` // If the query was fulfilled by a webhook call, this field is set to the // value of the `source` field returned in the webhook response. - WebhookSource string `protobuf:"bytes,8,opt,name=webhook_source,json=webhookSource" json:"webhook_source,omitempty"` + WebhookSource string `protobuf:"bytes,8,opt,name=webhook_source,json=webhookSource,proto3" json:"webhook_source,omitempty"` // If the query was fulfilled by a webhook call, this field is set to the // value of the `payload` field returned in the webhook response. - WebhookPayload *_struct.Struct `protobuf:"bytes,9,opt,name=webhook_payload,json=webhookPayload" json:"webhook_payload,omitempty"` + WebhookPayload *_struct.Struct `protobuf:"bytes,9,opt,name=webhook_payload,json=webhookPayload,proto3" json:"webhook_payload,omitempty"` // The collection of output contexts. If applicable, // `output_contexts.parameters` contains entries with name // `.original` containing the original parameter values // before the query. - OutputContexts []*Context `protobuf:"bytes,10,rep,name=output_contexts,json=outputContexts" json:"output_contexts,omitempty"` + OutputContexts []*Context `protobuf:"bytes,10,rep,name=output_contexts,json=outputContexts,proto3" json:"output_contexts,omitempty"` // The intent that matched the conversational query. Some, not // all fields are filled in this message, including but not limited to: // `name`, `display_name` and `webhook_state`. - Intent *Intent `protobuf:"bytes,11,opt,name=intent" json:"intent,omitempty"` + Intent *Intent `protobuf:"bytes,11,opt,name=intent,proto3" json:"intent,omitempty"` // The intent detection confidence. Values range from 0.0 // (completely uncertain) to 1.0 (completely certain). - IntentDetectionConfidence float32 `protobuf:"fixed32,12,opt,name=intent_detection_confidence,json=intentDetectionConfidence" json:"intent_detection_confidence,omitempty"` + IntentDetectionConfidence float32 `protobuf:"fixed32,12,opt,name=intent_detection_confidence,json=intentDetectionConfidence,proto3" json:"intent_detection_confidence,omitempty"` // The free-form diagnostic info. For example, this field // could contain webhook call latency. - DiagnosticInfo *_struct.Struct `protobuf:"bytes,14,opt,name=diagnostic_info,json=diagnosticInfo" json:"diagnostic_info,omitempty"` + DiagnosticInfo *_struct.Struct `protobuf:"bytes,14,opt,name=diagnostic_info,json=diagnosticInfo,proto3" json:"diagnostic_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -754,9 +754,9 @@ type StreamingDetectIntentRequest struct { // caller to choose an appropriate . It can be a random number or // some type of user identifier (preferably hashed). The length of the session // ID must not exceed 36 characters. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Optional. The parameters of this query. - QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams" json:"query_params,omitempty"` + QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams,proto3" json:"query_params,omitempty"` // Required. The input specification. It can be set to: // // 1. an audio config which instructs the speech recognizer how to process @@ -765,7 +765,7 @@ type StreamingDetectIntentRequest struct { // 2. a conversational query in the form of text, or // // 3. an event that specifies which intent to trigger. - QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput" json:"query_input,omitempty"` + QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput,proto3" json:"query_input,omitempty"` // Optional. If `false` (default), recognition does not cease until the // client closes the stream. // If `true`, the recognizer will detect a single spoken utterance in input @@ -774,7 +774,7 @@ type StreamingDetectIntentRequest struct { // client should close the stream and start a new request with a new stream as // needed. // This setting is ignored when `query_input` is a piece of text or an event. - SingleUtterance bool `protobuf:"varint,4,opt,name=single_utterance,json=singleUtterance" json:"single_utterance,omitempty"` + SingleUtterance bool `protobuf:"varint,4,opt,name=single_utterance,json=singleUtterance,proto3" json:"single_utterance,omitempty"` // Optional. The input audio content to be recognized. Must be sent if // `query_input` was set to a streaming input audio config. The complete audio // over all streaming messages must not exceed 1 minute. @@ -858,13 +858,13 @@ func (m *StreamingDetectIntentRequest) GetInputAudio() []byte { type StreamingDetectIntentResponse struct { // The unique identifier of the response. It can be used to // locate a response in the training example set or for reporting issues. - ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId" json:"response_id,omitempty"` + ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId,proto3" json:"response_id,omitempty"` // The result of speech recognition. - RecognitionResult *StreamingRecognitionResult `protobuf:"bytes,2,opt,name=recognition_result,json=recognitionResult" json:"recognition_result,omitempty"` + RecognitionResult *StreamingRecognitionResult `protobuf:"bytes,2,opt,name=recognition_result,json=recognitionResult,proto3" json:"recognition_result,omitempty"` // The result of the conversational query or event processing. - QueryResult *QueryResult `protobuf:"bytes,3,opt,name=query_result,json=queryResult" json:"query_result,omitempty"` + QueryResult *QueryResult `protobuf:"bytes,3,opt,name=query_result,json=queryResult,proto3" json:"query_result,omitempty"` // Specifies the status of the webhook request. - WebhookStatus *status.Status `protobuf:"bytes,4,opt,name=webhook_status,json=webhookStatus" json:"webhook_status,omitempty"` + WebhookStatus *status.Status `protobuf:"bytes,4,opt,name=webhook_status,json=webhookStatus,proto3" json:"webhook_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -957,16 +957,16 @@ func (m *StreamingDetectIntentResponse) GetWebhookStatus() *status.Status { // * for `MESSAGE_TYPE_END_OF_SINGLE_UTTERANCE`: only `event_type`. type StreamingRecognitionResult struct { // Type of the result message. - MessageType StreamingRecognitionResult_MessageType `protobuf:"varint,1,opt,name=message_type,json=messageType,enum=google.cloud.dialogflow.v2.StreamingRecognitionResult_MessageType" json:"message_type,omitempty"` + MessageType StreamingRecognitionResult_MessageType `protobuf:"varint,1,opt,name=message_type,json=messageType,proto3,enum=google.cloud.dialogflow.v2.StreamingRecognitionResult_MessageType" json:"message_type,omitempty"` // Transcript text representing the words that the user spoke. // Populated if and only if `event_type` = `RECOGNITION_EVENT_TRANSCRIPT`. - Transcript string `protobuf:"bytes,2,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,2,opt,name=transcript,proto3" json:"transcript,omitempty"` // The default of 0.0 is a sentinel value indicating `confidence` was not set. // If `false`, the `StreamingRecognitionResult` represents an // interim result that may change. If `true`, the recognizer will not return // any further hypotheses about this piece of the audio. May only be populated // for `event_type` = `RECOGNITION_EVENT_TRANSCRIPT`. - IsFinal bool `protobuf:"varint,3,opt,name=is_final,json=isFinal" json:"is_final,omitempty"` + IsFinal bool `protobuf:"varint,3,opt,name=is_final,json=isFinal,proto3" json:"is_final,omitempty"` // The Speech confidence between 0.0 and 1.0 for the current portion of audio. // A higher number indicates an estimated greater likelihood that the // recognized words are correct. The default of 0.0 is a sentinel value @@ -974,7 +974,7 @@ type StreamingRecognitionResult struct { // // This field is typically only provided if `is_final` is true and you should // not rely on it being accurate or even set. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1035,22 +1035,22 @@ func (m *StreamingRecognitionResult) GetConfidence() float32 { // Instructs the speech recognizer how to process the audio content. type InputAudioConfig struct { // Required. Audio encoding of the audio content to process. - AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,enum=google.cloud.dialogflow.v2.AudioEncoding" json:"audio_encoding,omitempty"` + AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,proto3,enum=google.cloud.dialogflow.v2.AudioEncoding" json:"audio_encoding,omitempty"` // Required. Sample rate (in Hertz) of the audio content sent in the query. // Refer to [Cloud Speech API documentation](/speech/docs/basics) for more // details. - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` // Required. The language of the supplied audio. Dialogflow does not do // translations. See [Language // Support](https://dialogflow.com/docs/languages) for a list of the // currently supported language codes. Note that queries in the same session // do not necessarily need to specify the same language. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The collection of phrase hints which are used to boost accuracy // of speech recognition. // Refer to [Cloud Speech API documentation](/speech/docs/basics#phrase-hints) // for more details. - PhraseHints []string `protobuf:"bytes,4,rep,name=phrase_hints,json=phraseHints" json:"phrase_hints,omitempty"` + PhraseHints []string `protobuf:"bytes,4,rep,name=phrase_hints,json=phraseHints,proto3" json:"phrase_hints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1112,12 +1112,12 @@ func (m *InputAudioConfig) GetPhraseHints() []string { type TextInput struct { // Required. The UTF-8 encoded natural language text to be processed. // Text length must not exceed 256 bytes. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Required. The language of this conversational query. See [Language // Support](https://dialogflow.com/docs/languages) for a list of the // currently supported language codes. Note that queries in the same session // do not necessarily need to specify the same language. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1168,14 +1168,14 @@ func (m *TextInput) GetLanguageCode() string { // `“Hello #welcome_event.name! What can I do for you today?”`. type EventInput struct { // Required. The unique identifier of the event. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The collection of parameters associated with the event. - Parameters *_struct.Struct `protobuf:"bytes,2,opt,name=parameters" json:"parameters,omitempty"` + Parameters *_struct.Struct `protobuf:"bytes,2,opt,name=parameters,proto3" json:"parameters,omitempty"` // Required. The language of this query. See [Language // Support](https://dialogflow.com/docs/languages) for a list of the // currently supported language codes. Note that queries in the same session // do not necessarily need to specify the same language. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1313,8 +1313,7 @@ func (x *sessionsStreamingDetectIntentClient) Recv() (*StreamingDetectIntentResp return m, nil } -// Server API for Sessions service - +// SessionsServer is the server API for Sessions service. type SessionsServer interface { // Processes a natural language query and returns structured, actionable data // as a result. This method is not idempotent, because it may cause contexts diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session_entity_type.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session_entity_type.pb.go index 275b0c442..5aaac80a5 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session_entity_type.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/session_entity_type.pb.go @@ -73,13 +73,13 @@ type SessionEntityType struct { // Required. The unique identifier of this session entity type. Format: // `projects//agent/sessions//entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. Indicates whether the additional data should override or // supplement the developer entity type definition. - EntityOverrideMode SessionEntityType_EntityOverrideMode `protobuf:"varint,2,opt,name=entity_override_mode,json=entityOverrideMode,enum=google.cloud.dialogflow.v2.SessionEntityType_EntityOverrideMode" json:"entity_override_mode,omitempty"` + EntityOverrideMode SessionEntityType_EntityOverrideMode `protobuf:"varint,2,opt,name=entity_override_mode,json=entityOverrideMode,proto3,enum=google.cloud.dialogflow.v2.SessionEntityType_EntityOverrideMode" json:"entity_override_mode,omitempty"` // Required. The collection of entities associated with this session entity // type. - Entities []*EntityType_Entity `protobuf:"bytes,3,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,3,rep,name=entities,proto3" json:"entities,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -134,12 +134,12 @@ func (m *SessionEntityType) GetEntities() []*EntityType_Entity { type ListSessionEntityTypesRequest struct { // Required. The session to list all session entity types from. // Format: `projects//agent/sessions/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -194,10 +194,10 @@ func (m *ListSessionEntityTypesRequest) GetPageToken() string { type ListSessionEntityTypesResponse struct { // The list of session entity types. There will be a maximum number of items // returned based on the page_size field in the request. - SessionEntityTypes []*SessionEntityType `protobuf:"bytes,1,rep,name=session_entity_types,json=sessionEntityTypes" json:"session_entity_types,omitempty"` + SessionEntityTypes []*SessionEntityType `protobuf:"bytes,1,rep,name=session_entity_types,json=sessionEntityTypes,proto3" json:"session_entity_types,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -246,7 +246,7 @@ type GetSessionEntityTypeRequest struct { // Required. The name of the session entity type. Format: // `projects//agent/sessions//entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -287,9 +287,9 @@ func (m *GetSessionEntityTypeRequest) GetName() string { type CreateSessionEntityTypeRequest struct { // Required. The session to create a session entity type for. // Format: `projects//agent/sessions/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The session entity type to create. - SessionEntityType *SessionEntityType `protobuf:"bytes,2,opt,name=session_entity_type,json=sessionEntityType" json:"session_entity_type,omitempty"` + SessionEntityType *SessionEntityType `protobuf:"bytes,2,opt,name=session_entity_type,json=sessionEntityType,proto3" json:"session_entity_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -338,9 +338,9 @@ type UpdateSessionEntityTypeRequest struct { // Required. The entity type to update. Format: // `projects//agent/sessions//entityTypes/`. - SessionEntityType *SessionEntityType `protobuf:"bytes,1,opt,name=session_entity_type,json=sessionEntityType" json:"session_entity_type,omitempty"` + SessionEntityType *SessionEntityType `protobuf:"bytes,1,opt,name=session_entity_type,json=sessionEntityType,proto3" json:"session_entity_type,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -389,7 +389,7 @@ type DeleteSessionEntityTypeRequest struct { // Required. The name of the entity type to delete. Format: // `projects//agent/sessions//entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -514,8 +514,7 @@ func (c *sessionEntityTypesClient) DeleteSessionEntityType(ctx context.Context, return out, nil } -// Server API for SessionEntityTypes service - +// SessionEntityTypesServer is the server API for SessionEntityTypes service. type SessionEntityTypesServer interface { // Returns the list of all session entity types in the specified session. ListSessionEntityTypes(context.Context, *ListSessionEntityTypesRequest) (*ListSessionEntityTypesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/webhook.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/webhook.pb.go index b72969859..275fe40db 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/webhook.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2/webhook.pb.go @@ -25,16 +25,16 @@ type WebhookRequest struct { // The unique identifier of detectIntent request session. // Can be used to identify end-user inside webhook implementation. // Format: `projects//agent/sessions/`. - Session string `protobuf:"bytes,4,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,4,opt,name=session,proto3" json:"session,omitempty"` // The unique identifier of the response. Contains the same value as // `[Streaming]DetectIntentResponse.response_id`. - ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId" json:"response_id,omitempty"` + ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId,proto3" json:"response_id,omitempty"` // The result of the conversational query or event processing. Contains the // same value as `[Streaming]DetectIntentResponse.query_result`. - QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult" json:"query_result,omitempty"` + QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult,proto3" json:"query_result,omitempty"` // Optional. The contents of the original request that was passed to // `[Streaming]DetectIntent` call. - OriginalDetectIntentRequest *OriginalDetectIntentRequest `protobuf:"bytes,3,opt,name=original_detect_intent_request,json=originalDetectIntentRequest" json:"original_detect_intent_request,omitempty"` + OriginalDetectIntentRequest *OriginalDetectIntentRequest `protobuf:"bytes,3,opt,name=original_detect_intent_request,json=originalDetectIntentRequest,proto3" json:"original_detect_intent_request,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -96,12 +96,12 @@ func (m *WebhookRequest) GetOriginalDetectIntentRequest() *OriginalDetectIntentR type WebhookResponse struct { // Optional. The text to be shown on the screen. This value is passed directly // to `QueryResult.fulfillment_text`. - FulfillmentText string `protobuf:"bytes,1,opt,name=fulfillment_text,json=fulfillmentText" json:"fulfillment_text,omitempty"` + FulfillmentText string `protobuf:"bytes,1,opt,name=fulfillment_text,json=fulfillmentText,proto3" json:"fulfillment_text,omitempty"` // Optional. The collection of rich messages to present to the user. This // value is passed directly to `QueryResult.fulfillment_messages`. - FulfillmentMessages []*Intent_Message `protobuf:"bytes,2,rep,name=fulfillment_messages,json=fulfillmentMessages" json:"fulfillment_messages,omitempty"` + FulfillmentMessages []*Intent_Message `protobuf:"bytes,2,rep,name=fulfillment_messages,json=fulfillmentMessages,proto3" json:"fulfillment_messages,omitempty"` // Optional. This value is passed directly to `QueryResult.webhook_source`. - Source string `protobuf:"bytes,3,opt,name=source" json:"source,omitempty"` + Source string `protobuf:"bytes,3,opt,name=source,proto3" json:"source,omitempty"` // Optional. This value is passed directly to `QueryResult.webhook_payload`. // See the related `fulfillment_messages[i].payload field`, which may be used // as an alternative to this field. @@ -125,13 +125,13 @@ type WebhookResponse struct { // } // } // } - Payload *_struct.Struct `protobuf:"bytes,4,opt,name=payload" json:"payload,omitempty"` + Payload *_struct.Struct `protobuf:"bytes,4,opt,name=payload,proto3" json:"payload,omitempty"` // Optional. The collection of output contexts. This value is passed directly // to `QueryResult.output_contexts`. - OutputContexts []*Context `protobuf:"bytes,5,rep,name=output_contexts,json=outputContexts" json:"output_contexts,omitempty"` + OutputContexts []*Context `protobuf:"bytes,5,rep,name=output_contexts,json=outputContexts,proto3" json:"output_contexts,omitempty"` // Optional. Makes the platform immediately invoke another `DetectIntent` call // internally with the specified event as input. - FollowupEventInput *EventInput `protobuf:"bytes,6,opt,name=followup_event_input,json=followupEventInput" json:"followup_event_input,omitempty"` + FollowupEventInput *EventInput `protobuf:"bytes,6,opt,name=followup_event_input,json=followupEventInput,proto3" json:"followup_event_input,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -208,10 +208,10 @@ func (m *WebhookResponse) GetFollowupEventInput() *EventInput { type OriginalDetectIntentRequest struct { // The source of this request, e.g., `google`, `facebook`, `slack`. It is set // by Dialogflow-owned servers. - Source string `protobuf:"bytes,1,opt,name=source" json:"source,omitempty"` + Source string `protobuf:"bytes,1,opt,name=source,proto3" json:"source,omitempty"` // Optional. This field is set to the value of `QueryParameters.payload` field // passed in the request. - Payload *_struct.Struct `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *_struct.Struct `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/agent.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/agent.pb.go index d5573bda1..d9ca3970a 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/agent.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/agent.pb.go @@ -64,32 +64,32 @@ func (Agent_MatchMode) EnumDescriptor() ([]byte, []int) { type Agent struct { // Required. The project of this agent. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The name of this agent. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. The default language of the agent as a language tag. See // [Language Support](https://dialogflow.com/docs/reference/language) for a // list of the currently supported language codes. // This field cannot be set by the `Update` method. - DefaultLanguageCode string `protobuf:"bytes,3,opt,name=default_language_code,json=defaultLanguageCode" json:"default_language_code,omitempty"` + DefaultLanguageCode string `protobuf:"bytes,3,opt,name=default_language_code,json=defaultLanguageCode,proto3" json:"default_language_code,omitempty"` // Optional. The list of all languages supported by this agent (except for the // `default_language_code`). - SupportedLanguageCodes []string `protobuf:"bytes,4,rep,name=supported_language_codes,json=supportedLanguageCodes" json:"supported_language_codes,omitempty"` + SupportedLanguageCodes []string `protobuf:"bytes,4,rep,name=supported_language_codes,json=supportedLanguageCodes,proto3" json:"supported_language_codes,omitempty"` // Required. The time zone of this agent from the // [time zone database](https://www.iana.org/time-zones), e.g., // America/New_York, Europe/Paris. - TimeZone string `protobuf:"bytes,5,opt,name=time_zone,json=timeZone" json:"time_zone,omitempty"` + TimeZone string `protobuf:"bytes,5,opt,name=time_zone,json=timeZone,proto3" json:"time_zone,omitempty"` // Optional. The description of this agent. // The maximum length is 500 characters. If exceeded, the request is rejected. - Description string `protobuf:"bytes,6,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,6,opt,name=description,proto3" json:"description,omitempty"` // Optional. The URI of the agent's avatar. // Avatars are used throughout the Dialogflow console and in the self-hosted // [Web Demo](https://dialogflow.com/docs/integrations/web-demo) integration. - AvatarUri string `protobuf:"bytes,7,opt,name=avatar_uri,json=avatarUri" json:"avatar_uri,omitempty"` + AvatarUri string `protobuf:"bytes,7,opt,name=avatar_uri,json=avatarUri,proto3" json:"avatar_uri,omitempty"` // Optional. Determines whether this agent should log conversation queries. - EnableLogging bool `protobuf:"varint,8,opt,name=enable_logging,json=enableLogging" json:"enable_logging,omitempty"` + EnableLogging bool `protobuf:"varint,8,opt,name=enable_logging,json=enableLogging,proto3" json:"enable_logging,omitempty"` // Optional. Determines how intents are detected from user queries. - MatchMode Agent_MatchMode `protobuf:"varint,9,opt,name=match_mode,json=matchMode,enum=google.cloud.dialogflow.v2beta1.Agent_MatchMode" json:"match_mode,omitempty"` + MatchMode Agent_MatchMode `protobuf:"varint,9,opt,name=match_mode,json=matchMode,proto3,enum=google.cloud.dialogflow.v2beta1.Agent_MatchMode" json:"match_mode,omitempty"` // Optional. To filter out false positive results and still get variety in // matched natural language inputs for your agent, you can tune the machine // learning classification threshold. If the returned score value is less than @@ -97,7 +97,7 @@ type Agent struct { // are no fallback intents defined, no intent will be triggered. The score // values range from 0.0 (completely uncertain) to 1.0 (completely certain). // If set to 0.0, the default of 0.3 is used. - ClassificationThreshold float32 `protobuf:"fixed32,10,opt,name=classification_threshold,json=classificationThreshold" json:"classification_threshold,omitempty"` + ClassificationThreshold float32 `protobuf:"fixed32,10,opt,name=classification_threshold,json=classificationThreshold,proto3" json:"classification_threshold,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -201,7 +201,7 @@ func (m *Agent) GetClassificationThreshold() float32 { type GetAgentRequest struct { // Required. The project that the agent to fetch is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -242,12 +242,12 @@ func (m *GetAgentRequest) GetParent() string { type SearchAgentsRequest struct { // Required. The project to list agents from. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -302,10 +302,10 @@ func (m *SearchAgentsRequest) GetPageToken() string { type SearchAgentsResponse struct { // The list of agents. There will be a maximum number of items returned based // on the page_size field in the request. - Agents []*Agent `protobuf:"bytes,1,rep,name=agents" json:"agents,omitempty"` + Agents []*Agent `protobuf:"bytes,1,rep,name=agents,proto3" json:"agents,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -353,7 +353,7 @@ func (m *SearchAgentsResponse) GetNextPageToken() string { type TrainAgentRequest struct { // Required. The project that the agent to train is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -394,11 +394,11 @@ func (m *TrainAgentRequest) GetParent() string { type ExportAgentRequest struct { // Required. The project that the agent to export is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The Google Cloud Storage URI to export the agent to. // Note: The URI must start with // "gs://". If left unspecified, the serialized agent is returned inline. - AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri" json:"agent_uri,omitempty"` + AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,proto3" json:"agent_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -484,7 +484,7 @@ type isExportAgentResponse_Agent interface { } type ExportAgentResponse_AgentUri struct { - AgentUri string `protobuf:"bytes,1,opt,name=agent_uri,json=agentUri,oneof"` + AgentUri string `protobuf:"bytes,1,opt,name=agent_uri,json=agentUri,proto3,oneof"` } type ExportAgentResponse_AgentContent struct { AgentContent []byte `protobuf:"bytes,2,opt,name=agent_content,json=agentContent,proto3,oneof"` @@ -584,7 +584,7 @@ func _ExportAgentResponse_OneofSizer(msg proto.Message) (n int) { type ImportAgentRequest struct { // Required. The project that the agent to import is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The agent to import. // // Types that are valid to be assigned to Agent: @@ -625,7 +625,7 @@ type isImportAgentRequest_Agent interface { } type ImportAgentRequest_AgentUri struct { - AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,oneof"` + AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,proto3,oneof"` } type ImportAgentRequest_AgentContent struct { AgentContent []byte `protobuf:"bytes,3,opt,name=agent_content,json=agentContent,proto3,oneof"` @@ -732,7 +732,7 @@ func _ImportAgentRequest_OneofSizer(msg proto.Message) (n int) { type RestoreAgentRequest struct { // Required. The project that the agent to restore is associated with. // Format: `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The agent to restore. // // Types that are valid to be assigned to Agent: @@ -773,7 +773,7 @@ type isRestoreAgentRequest_Agent interface { } type RestoreAgentRequest_AgentUri struct { - AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,oneof"` + AgentUri string `protobuf:"bytes,2,opt,name=agent_uri,json=agentUri,proto3,oneof"` } type RestoreAgentRequest_AgentContent struct { AgentContent []byte `protobuf:"bytes,3,opt,name=agent_content,json=agentContent,proto3,oneof"` @@ -1006,8 +1006,7 @@ func (c *agentsClient) RestoreAgent(ctx context.Context, in *RestoreAgentRequest return out, nil } -// Server API for Agents service - +// AgentsServer is the server API for Agents service. type AgentsServer interface { // Retrieves the specified agent. GetAgent(context.Context, *GetAgentRequest) (*Agent, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/context.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/context.pb.go index 0b88be6c0..a6535b0f9 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/context.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/context.pb.go @@ -38,16 +38,16 @@ type Context struct { // always converted to lowercase. If is not specified, we // assume default 'draft' environment. If is not specified, we // assume default '-' user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The number of conversational query requests after which the // context expires. If set to `0` (the default) the context expires // immediately. Contexts expire automatically after 10 minutes even if there // are no matching queries. - LifespanCount int32 `protobuf:"varint,2,opt,name=lifespan_count,json=lifespanCount" json:"lifespan_count,omitempty"` + LifespanCount int32 `protobuf:"varint,2,opt,name=lifespan_count,json=lifespanCount,proto3" json:"lifespan_count,omitempty"` // Optional. The collection of parameters associated with this context. // Refer to [this doc](https://dialogflow.com/docs/actions-and-parameters) for // syntax. - Parameters *_struct.Struct `protobuf:"bytes,3,opt,name=parameters" json:"parameters,omitempty"` + Parameters *_struct.Struct `protobuf:"bytes,3,opt,name=parameters,proto3" json:"parameters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -107,12 +107,12 @@ type ListContextsRequest struct { // construction and will be available soon. If is not // specified, we assume default 'draft' environment. If is not // specified, we assume default '-' user. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -167,10 +167,10 @@ func (m *ListContextsRequest) GetPageToken() string { type ListContextsResponse struct { // The list of contexts. There will be a maximum number of items // returned based on the page_size field in the request. - Contexts []*Context `protobuf:"bytes,1,rep,name=contexts" json:"contexts,omitempty"` + Contexts []*Context `protobuf:"bytes,1,rep,name=contexts,proto3" json:"contexts,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -223,7 +223,7 @@ type GetContextRequest struct { // users are under construction and will be available soon. If is not specified, we assume default 'draft' environment. If // is not specified, we assume default '-' user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -269,9 +269,9 @@ type CreateContextRequest struct { // construction and will be available soon. If is not // specified, we assume default 'draft' environment. If is not // specified, we assume default '-' user. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The context to create. - Context *Context `protobuf:"bytes,2,opt,name=context" json:"context,omitempty"` + Context *Context `protobuf:"bytes,2,opt,name=context,proto3" json:"context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -318,9 +318,9 @@ func (m *CreateContextRequest) GetContext() *Context { // The request message for [Contexts.UpdateContext][google.cloud.dialogflow.v2beta1.Contexts.UpdateContext]. type UpdateContextRequest struct { // Required. The context to update. - Context *Context `protobuf:"bytes,1,opt,name=context" json:"context,omitempty"` + Context *Context `protobuf:"bytes,1,opt,name=context,proto3" json:"context,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -374,7 +374,7 @@ type DeleteContextRequest struct { // ID> is not specified, we assume default 'draft' environment. If // is not specified, we assume default // '-' user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -420,7 +420,7 @@ type DeleteAllContextsRequest struct { // available soon. If is not specified we assume default // 'draft' environment. If is not specified, we assume default // '-' user. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -556,8 +556,7 @@ func (c *contextsClient) DeleteAllContexts(ctx context.Context, in *DeleteAllCon return out, nil } -// Server API for Contexts service - +// ContextsServer is the server API for Contexts service. type ContextsServer interface { // Returns the list of all contexts in the specified session. ListContexts(context.Context, *ListContextsRequest) (*ListContextsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/entity_type.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/entity_type.pb.go index 82d09bfa0..d417e88e6 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/entity_type.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/entity_type.pb.go @@ -98,16 +98,16 @@ type EntityType struct { // automatically. // The unique identifier of the entity type. Format: // `projects//agent/entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of the entity. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. Indicates the kind of entity type. - Kind EntityType_Kind `protobuf:"varint,3,opt,name=kind,enum=google.cloud.dialogflow.v2beta1.EntityType_Kind" json:"kind,omitempty"` + Kind EntityType_Kind `protobuf:"varint,3,opt,name=kind,proto3,enum=google.cloud.dialogflow.v2beta1.EntityType_Kind" json:"kind,omitempty"` // Optional. Indicates whether the entity type can be automatically // expanded. - AutoExpansionMode EntityType_AutoExpansionMode `protobuf:"varint,4,opt,name=auto_expansion_mode,json=autoExpansionMode,enum=google.cloud.dialogflow.v2beta1.EntityType_AutoExpansionMode" json:"auto_expansion_mode,omitempty"` + AutoExpansionMode EntityType_AutoExpansionMode `protobuf:"varint,4,opt,name=auto_expansion_mode,json=autoExpansionMode,proto3,enum=google.cloud.dialogflow.v2beta1.EntityType_AutoExpansionMode" json:"auto_expansion_mode,omitempty"` // Optional. The collection of entities associated with the entity type. - Entities []*EntityType_Entity `protobuf:"bytes,6,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,6,rep,name=entities,proto3" json:"entities,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -180,10 +180,10 @@ type EntityType_Entity struct { // For `KIND_LIST` entity types: // A string that can contain references to other entity types (with or // without aliases). - Value string `protobuf:"bytes,1,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` // Required. A collection of synonyms. For `KIND_LIST` entity types this // must contain exactly one synonym equal to `value`. - Synonyms []string `protobuf:"bytes,2,rep,name=synonyms" json:"synonyms,omitempty"` + Synonyms []string `protobuf:"bytes,2,rep,name=synonyms,proto3" json:"synonyms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -231,18 +231,18 @@ func (m *EntityType_Entity) GetSynonyms() []string { type ListEntityTypesRequest struct { // Required. The agent to list all entity types from. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The language to list entity synonyms for. If not specified, // the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -304,10 +304,10 @@ func (m *ListEntityTypesRequest) GetPageToken() string { type ListEntityTypesResponse struct { // The list of agent entity types. There will be a maximum number of items // returned based on the page_size field in the request. - EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes" json:"entity_types,omitempty"` + EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes,proto3" json:"entity_types,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -355,13 +355,13 @@ func (m *ListEntityTypesResponse) GetNextPageToken() string { type GetEntityTypeRequest struct { // Required. The name of the entity type. // Format: `projects//agent/entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The language to retrieve entity synonyms for. If not specified, // the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,15 +409,15 @@ func (m *GetEntityTypeRequest) GetLanguageCode() string { type CreateEntityTypeRequest struct { // Required. The agent to create a entity type for. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The entity type to create. - EntityType *EntityType `protobuf:"bytes,2,opt,name=entity_type,json=entityType" json:"entity_type,omitempty"` + EntityType *EntityType `protobuf:"bytes,2,opt,name=entity_type,json=entityType,proto3" json:"entity_type,omitempty"` // Optional. The language of entity synonyms defined in `entity_type`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -472,15 +472,15 @@ func (m *CreateEntityTypeRequest) GetLanguageCode() string { type UpdateEntityTypeRequest struct { // Required. The entity type to update. // Format: `projects//agent/entityTypes/`. - EntityType *EntityType `protobuf:"bytes,1,opt,name=entity_type,json=entityType" json:"entity_type,omitempty"` + EntityType *EntityType `protobuf:"bytes,1,opt,name=entity_type,json=entityType,proto3" json:"entity_type,omitempty"` // Optional. The language of entity synonyms defined in `entity_type`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -535,7 +535,7 @@ func (m *UpdateEntityTypeRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteEntityTypeRequest struct { // Required. The name of the entity type to delete. // Format: `projects//agent/entityTypes/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -576,7 +576,7 @@ func (m *DeleteEntityTypeRequest) GetName() string { type BatchUpdateEntityTypesRequest struct { // Required. The name of the agent to update or create entity types in. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The source of the entity type batch. // // For each entity type in the batch: @@ -592,9 +592,9 @@ type BatchUpdateEntityTypesRequest struct { // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -629,10 +629,10 @@ type isBatchUpdateEntityTypesRequest_EntityTypeBatch interface { } type BatchUpdateEntityTypesRequest_EntityTypeBatchUri struct { - EntityTypeBatchUri string `protobuf:"bytes,2,opt,name=entity_type_batch_uri,json=entityTypeBatchUri,oneof"` + EntityTypeBatchUri string `protobuf:"bytes,2,opt,name=entity_type_batch_uri,json=entityTypeBatchUri,proto3,oneof"` } type BatchUpdateEntityTypesRequest_EntityTypeBatchInline struct { - EntityTypeBatchInline *EntityTypeBatch `protobuf:"bytes,3,opt,name=entity_type_batch_inline,json=entityTypeBatchInline,oneof"` + EntityTypeBatchInline *EntityTypeBatch `protobuf:"bytes,3,opt,name=entity_type_batch_inline,json=entityTypeBatchInline,proto3,oneof"` } func (*BatchUpdateEntityTypesRequest_EntityTypeBatchUri) isBatchUpdateEntityTypesRequest_EntityTypeBatch() { @@ -755,7 +755,7 @@ func _BatchUpdateEntityTypesRequest_OneofSizer(msg proto.Message) (n int) { // The response message for [EntityTypes.BatchUpdateEntityTypes][google.cloud.dialogflow.v2beta1.EntityTypes.BatchUpdateEntityTypes]. type BatchUpdateEntityTypesResponse struct { // The collection of updated or created entity types. - EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes" json:"entity_types,omitempty"` + EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes,proto3" json:"entity_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -796,10 +796,10 @@ func (m *BatchUpdateEntityTypesResponse) GetEntityTypes() []*EntityType { type BatchDeleteEntityTypesRequest struct { // Required. The name of the agent to delete all entities types for. Format: // `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The names entity types to delete. All names must point to the // same agent as `parent`. - EntityTypeNames []string `protobuf:"bytes,2,rep,name=entity_type_names,json=entityTypeNames" json:"entity_type_names,omitempty"` + EntityTypeNames []string `protobuf:"bytes,2,rep,name=entity_type_names,json=entityTypeNames,proto3" json:"entity_type_names,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -847,15 +847,15 @@ func (m *BatchDeleteEntityTypesRequest) GetEntityTypeNames() []string { type BatchCreateEntitiesRequest struct { // Required. The name of the entity type to create entities in. Format: // `projects//agent/entityTypes/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The collection of entities to create. - Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities,proto3" json:"entities,omitempty"` // Optional. The language of entity synonyms defined in `entities`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -910,17 +910,17 @@ func (m *BatchCreateEntitiesRequest) GetLanguageCode() string { type BatchUpdateEntitiesRequest struct { // Required. The name of the entity type to update the entities in. Format: // `projects//agent/entityTypes/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The collection of new entities to replace the existing entities. - Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,2,rep,name=entities,proto3" json:"entities,omitempty"` // Optional. The language of entity synonyms defined in `entities`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -982,17 +982,17 @@ func (m *BatchUpdateEntitiesRequest) GetUpdateMask() *field_mask.FieldMask { type BatchDeleteEntitiesRequest struct { // Required. The name of the entity type to delete entries for. Format: // `projects//agent/entityTypes/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The canonical `values` of the entities to delete. Note that // these are not fully-qualified names, i.e. they don't start with // `projects/`. - EntityValues []string `protobuf:"bytes,2,rep,name=entity_values,json=entityValues" json:"entity_values,omitempty"` + EntityValues []string `protobuf:"bytes,2,rep,name=entity_values,json=entityValues,proto3" json:"entity_values,omitempty"` // Optional. The language of entity synonyms defined in `entities`. If not // specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1046,7 +1046,7 @@ func (m *BatchDeleteEntitiesRequest) GetLanguageCode() string { // This message is a wrapper around a collection of entity types. type EntityTypeBatch struct { // A collection of entity types. - EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes" json:"entity_types,omitempty"` + EntityTypes []*EntityType `protobuf:"bytes,1,rep,name=entity_types,json=entityTypes,proto3" json:"entity_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1251,8 +1251,7 @@ func (c *entityTypesClient) BatchDeleteEntities(ctx context.Context, in *BatchDe return out, nil } -// Server API for EntityTypes service - +// EntityTypesServer is the server API for EntityTypes service. type EntityTypesServer interface { // Returns the list of all entity types in the specified agent. ListEntityTypes(context.Context, *ListEntityTypesRequest) (*ListEntityTypesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/intent.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/intent.pb.go index 28fd678ae..be7484a05 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/intent.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/intent.pb.go @@ -241,16 +241,16 @@ type Intent struct { // automatically. // The unique identifier of this intent. // Format: `projects//agent/intents/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of this intent. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. Indicates whether webhooks are enabled for the intent. - WebhookState Intent_WebhookState `protobuf:"varint,6,opt,name=webhook_state,json=webhookState,enum=google.cloud.dialogflow.v2beta1.Intent_WebhookState" json:"webhook_state,omitempty"` + WebhookState Intent_WebhookState `protobuf:"varint,6,opt,name=webhook_state,json=webhookState,proto3,enum=google.cloud.dialogflow.v2beta1.Intent_WebhookState" json:"webhook_state,omitempty"` // Optional. The priority of this intent. Higher numbers represent higher // priorities. Zero or negative numbers mean that the intent is disabled. - Priority int32 `protobuf:"varint,3,opt,name=priority" json:"priority,omitempty"` + Priority int32 `protobuf:"varint,3,opt,name=priority,proto3" json:"priority,omitempty"` // Optional. Indicates whether this is a fallback intent. - IsFallback bool `protobuf:"varint,4,opt,name=is_fallback,json=isFallback" json:"is_fallback,omitempty"` + IsFallback bool `protobuf:"varint,4,opt,name=is_fallback,json=isFallback,proto3" json:"is_fallback,omitempty"` // Optional. Indicates whether Machine Learning is enabled for the intent. // Note: If `ml_enabled` setting is set to false, then this intent is not // taken into account during inference in `ML ONLY` match mode. Also, @@ -262,54 +262,54 @@ type Intent struct { // ml_enabled = false / ml_disabled = true. // - After April 15th, 2018 the default is: // ml_enabled = true / ml_disabled = false. - MlEnabled bool `protobuf:"varint,5,opt,name=ml_enabled,json=mlEnabled" json:"ml_enabled,omitempty"` + MlEnabled bool `protobuf:"varint,5,opt,name=ml_enabled,json=mlEnabled,proto3" json:"ml_enabled,omitempty"` // Optional. Indicates whether Machine Learning is disabled for the intent. // Note: If `ml_disabled` setting is set to true, then this intent is not // taken into account during inference in `ML ONLY` match mode. Also, // auto-markup in the UI is turned off. - MlDisabled bool `protobuf:"varint,19,opt,name=ml_disabled,json=mlDisabled" json:"ml_disabled,omitempty"` + MlDisabled bool `protobuf:"varint,19,opt,name=ml_disabled,json=mlDisabled,proto3" json:"ml_disabled,omitempty"` // Optional. The list of context names required for this intent to be // triggered. // Format: `projects//agent/sessions/-/contexts/`. - InputContextNames []string `protobuf:"bytes,7,rep,name=input_context_names,json=inputContextNames" json:"input_context_names,omitempty"` + InputContextNames []string `protobuf:"bytes,7,rep,name=input_context_names,json=inputContextNames,proto3" json:"input_context_names,omitempty"` // Optional. The collection of event names that trigger the intent. // If the collection of input contexts is not empty, all of the contexts must // be present in the active user session for an event to trigger this intent. - Events []string `protobuf:"bytes,8,rep,name=events" json:"events,omitempty"` + Events []string `protobuf:"bytes,8,rep,name=events,proto3" json:"events,omitempty"` // Optional. The collection of examples/templates that the agent is // trained on. - TrainingPhrases []*Intent_TrainingPhrase `protobuf:"bytes,9,rep,name=training_phrases,json=trainingPhrases" json:"training_phrases,omitempty"` + TrainingPhrases []*Intent_TrainingPhrase `protobuf:"bytes,9,rep,name=training_phrases,json=trainingPhrases,proto3" json:"training_phrases,omitempty"` // Optional. The name of the action associated with the intent. - Action string `protobuf:"bytes,10,opt,name=action" json:"action,omitempty"` + Action string `protobuf:"bytes,10,opt,name=action,proto3" json:"action,omitempty"` // Optional. The collection of contexts that are activated when the intent // is matched. Context messages in this collection should not set the // parameters field. Setting the `lifespan_count` to 0 will reset the context // when the intent is matched. // Format: `projects//agent/sessions/-/contexts/`. - OutputContexts []*Context `protobuf:"bytes,11,rep,name=output_contexts,json=outputContexts" json:"output_contexts,omitempty"` + OutputContexts []*Context `protobuf:"bytes,11,rep,name=output_contexts,json=outputContexts,proto3" json:"output_contexts,omitempty"` // Optional. Indicates whether to delete all contexts in the current // session when this intent is matched. - ResetContexts bool `protobuf:"varint,12,opt,name=reset_contexts,json=resetContexts" json:"reset_contexts,omitempty"` + ResetContexts bool `protobuf:"varint,12,opt,name=reset_contexts,json=resetContexts,proto3" json:"reset_contexts,omitempty"` // Optional. The collection of parameters associated with the intent. - Parameters []*Intent_Parameter `protobuf:"bytes,13,rep,name=parameters" json:"parameters,omitempty"` + Parameters []*Intent_Parameter `protobuf:"bytes,13,rep,name=parameters,proto3" json:"parameters,omitempty"` // Optional. The collection of rich messages corresponding to the // `Response` field in the Dialogflow console. - Messages []*Intent_Message `protobuf:"bytes,14,rep,name=messages" json:"messages,omitempty"` + Messages []*Intent_Message `protobuf:"bytes,14,rep,name=messages,proto3" json:"messages,omitempty"` // Optional. The list of platforms for which the first response will be // taken from among the messages assigned to the DEFAULT_PLATFORM. - DefaultResponsePlatforms []Intent_Message_Platform `protobuf:"varint,15,rep,packed,name=default_response_platforms,json=defaultResponsePlatforms,enum=google.cloud.dialogflow.v2beta1.Intent_Message_Platform" json:"default_response_platforms,omitempty"` + DefaultResponsePlatforms []Intent_Message_Platform `protobuf:"varint,15,rep,packed,name=default_response_platforms,json=defaultResponsePlatforms,proto3,enum=google.cloud.dialogflow.v2beta1.Intent_Message_Platform" json:"default_response_platforms,omitempty"` // The unique identifier of the root intent in the chain of followup intents. // It identifies the correct followup intents chain for this intent. // Format: `projects//agent/intents/`. - RootFollowupIntentName string `protobuf:"bytes,16,opt,name=root_followup_intent_name,json=rootFollowupIntentName" json:"root_followup_intent_name,omitempty"` + RootFollowupIntentName string `protobuf:"bytes,16,opt,name=root_followup_intent_name,json=rootFollowupIntentName,proto3" json:"root_followup_intent_name,omitempty"` // The unique identifier of the parent intent in the chain of followup // intents. // It identifies the parent followup intent. // Format: `projects//agent/intents/`. - ParentFollowupIntentName string `protobuf:"bytes,17,opt,name=parent_followup_intent_name,json=parentFollowupIntentName" json:"parent_followup_intent_name,omitempty"` + ParentFollowupIntentName string `protobuf:"bytes,17,opt,name=parent_followup_intent_name,json=parentFollowupIntentName,proto3" json:"parent_followup_intent_name,omitempty"` // Optional. Collection of information about all followup intents that have // name of this intent as a root_name. - FollowupIntentInfo []*Intent_FollowupIntentInfo `protobuf:"bytes,18,rep,name=followup_intent_info,json=followupIntentInfo" json:"followup_intent_info,omitempty"` + FollowupIntentInfo []*Intent_FollowupIntentInfo `protobuf:"bytes,18,rep,name=followup_intent_info,json=followupIntentInfo,proto3" json:"followup_intent_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -475,17 +475,17 @@ func (m *Intent) GetFollowupIntentInfo() []*Intent_FollowupIntentInfo { // Represents an example or template that the agent is trained on. type Intent_TrainingPhrase struct { // Required. The unique identifier of this training phrase. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The type of the training phrase. - Type Intent_TrainingPhrase_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.dialogflow.v2beta1.Intent_TrainingPhrase_Type" json:"type,omitempty"` + Type Intent_TrainingPhrase_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.dialogflow.v2beta1.Intent_TrainingPhrase_Type" json:"type,omitempty"` // Required. The collection of training phrase parts (can be annotated). // Fields: `entity_type`, `alias` and `user_defined` should be populated // only for the annotated parts of the training phrase. - Parts []*Intent_TrainingPhrase_Part `protobuf:"bytes,3,rep,name=parts" json:"parts,omitempty"` + Parts []*Intent_TrainingPhrase_Part `protobuf:"bytes,3,rep,name=parts,proto3" json:"parts,omitempty"` // Optional. Indicates how many times this example or template was added to // the intent. Each time a developer adds an existing sample by editing an // intent or training, this counter is increased. - TimesAddedCount int32 `protobuf:"varint,4,opt,name=times_added_count,json=timesAddedCount" json:"times_added_count,omitempty"` + TimesAddedCount int32 `protobuf:"varint,4,opt,name=times_added_count,json=timesAddedCount,proto3" json:"times_added_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -548,17 +548,17 @@ type Intent_TrainingPhrase_Part struct { // Required. The text corresponding to the example or template, // if there are no annotations. For // annotated examples, it is the text for one of the example's parts. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Optional. The entity type name prefixed with `@`. This field is // required for the annotated part of the text and applies only to // examples. - EntityType string `protobuf:"bytes,2,opt,name=entity_type,json=entityType" json:"entity_type,omitempty"` + EntityType string `protobuf:"bytes,2,opt,name=entity_type,json=entityType,proto3" json:"entity_type,omitempty"` // Optional. The parameter name for the value extracted from the // annotated part of the example. - Alias string `protobuf:"bytes,3,opt,name=alias" json:"alias,omitempty"` + Alias string `protobuf:"bytes,3,opt,name=alias,proto3" json:"alias,omitempty"` // Optional. Indicates whether the text was manually annotated by the // developer. - UserDefined bool `protobuf:"varint,4,opt,name=user_defined,json=userDefined" json:"user_defined,omitempty"` + UserDefined bool `protobuf:"varint,4,opt,name=user_defined,json=userDefined,proto3" json:"user_defined,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -619,34 +619,34 @@ func (m *Intent_TrainingPhrase_Part) GetUserDefined() bool { // Represents intent parameters. type Intent_Parameter struct { // The unique identifier of this parameter. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of the parameter. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Optional. The definition of the parameter value. It can be: // - a constant string, // - a parameter value defined as `$parameter_name`, // - an original parameter value defined as `$parameter_name.original`, // - a parameter value from some context defined as // `#context_name.parameter_name`. - Value string `protobuf:"bytes,3,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` // Optional. The default value to use when the `value` yields an empty // result. // Default values can be extracted from contexts by using the following // syntax: `#context_name.parameter_name`. - DefaultValue string `protobuf:"bytes,4,opt,name=default_value,json=defaultValue" json:"default_value,omitempty"` + DefaultValue string `protobuf:"bytes,4,opt,name=default_value,json=defaultValue,proto3" json:"default_value,omitempty"` // Optional. The name of the entity type, prefixed with `@`, that // describes values of the parameter. If the parameter is // required, this must be provided. - EntityTypeDisplayName string `protobuf:"bytes,5,opt,name=entity_type_display_name,json=entityTypeDisplayName" json:"entity_type_display_name,omitempty"` + EntityTypeDisplayName string `protobuf:"bytes,5,opt,name=entity_type_display_name,json=entityTypeDisplayName,proto3" json:"entity_type_display_name,omitempty"` // Optional. Indicates whether the parameter is required. That is, // whether the intent cannot be completed without collecting the parameter // value. - Mandatory bool `protobuf:"varint,6,opt,name=mandatory" json:"mandatory,omitempty"` + Mandatory bool `protobuf:"varint,6,opt,name=mandatory,proto3" json:"mandatory,omitempty"` // Optional. The collection of prompts that the agent can present to the // user in order to collect value for the parameter. - Prompts []string `protobuf:"bytes,7,rep,name=prompts" json:"prompts,omitempty"` + Prompts []string `protobuf:"bytes,7,rep,name=prompts,proto3" json:"prompts,omitempty"` // Optional. Indicates whether the parameter represents a list of values. - IsList bool `protobuf:"varint,8,opt,name=is_list,json=isList" json:"is_list,omitempty"` + IsList bool `protobuf:"varint,8,opt,name=is_list,json=isList,proto3" json:"is_list,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -750,7 +750,7 @@ type Intent_Message struct { // *Intent_Message_CarouselSelect_ Message isIntent_Message_Message `protobuf_oneof:"message"` // Optional. The platform that this message is intended for. - Platform Intent_Message_Platform `protobuf:"varint,6,opt,name=platform,enum=google.cloud.dialogflow.v2beta1.Intent_Message_Platform" json:"platform,omitempty"` + Platform Intent_Message_Platform `protobuf:"varint,6,opt,name=platform,proto3,enum=google.cloud.dialogflow.v2beta1.Intent_Message_Platform" json:"platform,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -785,37 +785,37 @@ type isIntent_Message_Message interface { } type Intent_Message_Text_ struct { - Text *Intent_Message_Text `protobuf:"bytes,1,opt,name=text,oneof"` + Text *Intent_Message_Text `protobuf:"bytes,1,opt,name=text,proto3,oneof"` } type Intent_Message_Image_ struct { - Image *Intent_Message_Image `protobuf:"bytes,2,opt,name=image,oneof"` + Image *Intent_Message_Image `protobuf:"bytes,2,opt,name=image,proto3,oneof"` } type Intent_Message_QuickReplies_ struct { - QuickReplies *Intent_Message_QuickReplies `protobuf:"bytes,3,opt,name=quick_replies,json=quickReplies,oneof"` + QuickReplies *Intent_Message_QuickReplies `protobuf:"bytes,3,opt,name=quick_replies,json=quickReplies,proto3,oneof"` } type Intent_Message_Card_ struct { - Card *Intent_Message_Card `protobuf:"bytes,4,opt,name=card,oneof"` + Card *Intent_Message_Card `protobuf:"bytes,4,opt,name=card,proto3,oneof"` } type Intent_Message_Payload struct { - Payload *_struct.Struct `protobuf:"bytes,5,opt,name=payload,oneof"` + Payload *_struct.Struct `protobuf:"bytes,5,opt,name=payload,proto3,oneof"` } type Intent_Message_SimpleResponses_ struct { - SimpleResponses *Intent_Message_SimpleResponses `protobuf:"bytes,7,opt,name=simple_responses,json=simpleResponses,oneof"` + SimpleResponses *Intent_Message_SimpleResponses `protobuf:"bytes,7,opt,name=simple_responses,json=simpleResponses,proto3,oneof"` } type Intent_Message_BasicCard_ struct { - BasicCard *Intent_Message_BasicCard `protobuf:"bytes,8,opt,name=basic_card,json=basicCard,oneof"` + BasicCard *Intent_Message_BasicCard `protobuf:"bytes,8,opt,name=basic_card,json=basicCard,proto3,oneof"` } type Intent_Message_Suggestions_ struct { - Suggestions *Intent_Message_Suggestions `protobuf:"bytes,9,opt,name=suggestions,oneof"` + Suggestions *Intent_Message_Suggestions `protobuf:"bytes,9,opt,name=suggestions,proto3,oneof"` } type Intent_Message_LinkOutSuggestion_ struct { - LinkOutSuggestion *Intent_Message_LinkOutSuggestion `protobuf:"bytes,10,opt,name=link_out_suggestion,json=linkOutSuggestion,oneof"` + LinkOutSuggestion *Intent_Message_LinkOutSuggestion `protobuf:"bytes,10,opt,name=link_out_suggestion,json=linkOutSuggestion,proto3,oneof"` } type Intent_Message_ListSelect_ struct { - ListSelect *Intent_Message_ListSelect `protobuf:"bytes,11,opt,name=list_select,json=listSelect,oneof"` + ListSelect *Intent_Message_ListSelect `protobuf:"bytes,11,opt,name=list_select,json=listSelect,proto3,oneof"` } type Intent_Message_CarouselSelect_ struct { - CarouselSelect *Intent_Message_CarouselSelect `protobuf:"bytes,12,opt,name=carousel_select,json=carouselSelect,oneof"` + CarouselSelect *Intent_Message_CarouselSelect `protobuf:"bytes,12,opt,name=carousel_select,json=carouselSelect,proto3,oneof"` } func (*Intent_Message_Text_) isIntent_Message_Message() {} @@ -1169,7 +1169,7 @@ func _Intent_Message_OneofSizer(msg proto.Message) (n int) { // The text response message. type Intent_Message_Text struct { // Optional. The collection of the agent's responses. - Text []string `protobuf:"bytes,1,rep,name=text" json:"text,omitempty"` + Text []string `protobuf:"bytes,1,rep,name=text,proto3" json:"text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1209,10 +1209,10 @@ func (m *Intent_Message_Text) GetText() []string { // The image response message. type Intent_Message_Image struct { // Optional. The public URI to an image file. - ImageUri string `protobuf:"bytes,1,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,1,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` // Optional. A text description of the image to be used for accessibility, // e.g., screen readers. - AccessibilityText string `protobuf:"bytes,2,opt,name=accessibility_text,json=accessibilityText" json:"accessibility_text,omitempty"` + AccessibilityText string `protobuf:"bytes,2,opt,name=accessibility_text,json=accessibilityText,proto3" json:"accessibility_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1259,9 +1259,9 @@ func (m *Intent_Message_Image) GetAccessibilityText() string { // The quick replies response message. type Intent_Message_QuickReplies struct { // Optional. The title of the collection of quick replies. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Optional. The collection of quick replies. - QuickReplies []string `protobuf:"bytes,2,rep,name=quick_replies,json=quickReplies" json:"quick_replies,omitempty"` + QuickReplies []string `protobuf:"bytes,2,rep,name=quick_replies,json=quickReplies,proto3" json:"quick_replies,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1308,13 +1308,13 @@ func (m *Intent_Message_QuickReplies) GetQuickReplies() []string { // The card response message. type Intent_Message_Card struct { // Optional. The title of the card. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Optional. The subtitle of the card. - Subtitle string `protobuf:"bytes,2,opt,name=subtitle" json:"subtitle,omitempty"` + Subtitle string `protobuf:"bytes,2,opt,name=subtitle,proto3" json:"subtitle,omitempty"` // Optional. The public URI to an image file for the card. - ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,3,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` // Optional. The collection of card buttons. - Buttons []*Intent_Message_Card_Button `protobuf:"bytes,4,rep,name=buttons" json:"buttons,omitempty"` + Buttons []*Intent_Message_Card_Button `protobuf:"bytes,4,rep,name=buttons,proto3" json:"buttons,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1375,10 +1375,10 @@ func (m *Intent_Message_Card) GetButtons() []*Intent_Message_Card_Button { // Optional. Contains information about a button. type Intent_Message_Card_Button struct { // Optional. The text to show on the button. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Optional. The text to send back to the Dialogflow API or a URI to // open. - Postback string `protobuf:"bytes,2,opt,name=postback" json:"postback,omitempty"` + Postback string `protobuf:"bytes,2,opt,name=postback,proto3" json:"postback,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1426,13 +1426,13 @@ func (m *Intent_Message_Card_Button) GetPostback() string { type Intent_Message_SimpleResponse struct { // One of text_to_speech or ssml must be provided. The plain text of the // speech output. Mutually exclusive with ssml. - TextToSpeech string `protobuf:"bytes,1,opt,name=text_to_speech,json=textToSpeech" json:"text_to_speech,omitempty"` + TextToSpeech string `protobuf:"bytes,1,opt,name=text_to_speech,json=textToSpeech,proto3" json:"text_to_speech,omitempty"` // One of text_to_speech or ssml must be provided. Structured spoken // response to the user in the SSML format. Mutually exclusive with // text_to_speech. - Ssml string `protobuf:"bytes,2,opt,name=ssml" json:"ssml,omitempty"` + Ssml string `protobuf:"bytes,2,opt,name=ssml,proto3" json:"ssml,omitempty"` // Optional. The text to display. - DisplayText string `protobuf:"bytes,3,opt,name=display_text,json=displayText" json:"display_text,omitempty"` + DisplayText string `protobuf:"bytes,3,opt,name=display_text,json=displayText,proto3" json:"display_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1489,7 +1489,7 @@ func (m *Intent_Message_SimpleResponse) GetDisplayText() string { // `SimpleResponse`. type Intent_Message_SimpleResponses struct { // Required. The list of simple responses. - SimpleResponses []*Intent_Message_SimpleResponse `protobuf:"bytes,1,rep,name=simple_responses,json=simpleResponses" json:"simple_responses,omitempty"` + SimpleResponses []*Intent_Message_SimpleResponse `protobuf:"bytes,1,rep,name=simple_responses,json=simpleResponses,proto3" json:"simple_responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1529,15 +1529,15 @@ func (m *Intent_Message_SimpleResponses) GetSimpleResponses() []*Intent_Message_ // The basic card message. Useful for displaying information. type Intent_Message_BasicCard struct { // Optional. The title of the card. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Optional. The subtitle of the card. - Subtitle string `protobuf:"bytes,2,opt,name=subtitle" json:"subtitle,omitempty"` + Subtitle string `protobuf:"bytes,2,opt,name=subtitle,proto3" json:"subtitle,omitempty"` // Required, unless image is present. The body text of the card. - FormattedText string `protobuf:"bytes,3,opt,name=formatted_text,json=formattedText" json:"formatted_text,omitempty"` + FormattedText string `protobuf:"bytes,3,opt,name=formatted_text,json=formattedText,proto3" json:"formatted_text,omitempty"` // Optional. The image for the card. - Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"` + Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image,proto3" json:"image,omitempty"` // Optional. The collection of card buttons. - Buttons []*Intent_Message_BasicCard_Button `protobuf:"bytes,5,rep,name=buttons" json:"buttons,omitempty"` + Buttons []*Intent_Message_BasicCard_Button `protobuf:"bytes,5,rep,name=buttons,proto3" json:"buttons,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1605,9 +1605,9 @@ func (m *Intent_Message_BasicCard) GetButtons() []*Intent_Message_BasicCard_Butt // The button object that appears at the bottom of a card. type Intent_Message_BasicCard_Button struct { // Required. The title of the button. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Required. Action to take when a user taps on the button. - OpenUriAction *Intent_Message_BasicCard_Button_OpenUriAction `protobuf:"bytes,2,opt,name=open_uri_action,json=openUriAction" json:"open_uri_action,omitempty"` + OpenUriAction *Intent_Message_BasicCard_Button_OpenUriAction `protobuf:"bytes,2,opt,name=open_uri_action,json=openUriAction,proto3" json:"open_uri_action,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1654,7 +1654,7 @@ func (m *Intent_Message_BasicCard_Button) GetOpenUriAction() *Intent_Message_Bas // Opens the given URI. type Intent_Message_BasicCard_Button_OpenUriAction struct { // Required. The HTTP or HTTPS scheme URI. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1699,7 +1699,7 @@ func (m *Intent_Message_BasicCard_Button_OpenUriAction) GetUri() string { // to the conversation. type Intent_Message_Suggestion struct { // Required. The text shown the in the suggestion chip. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1739,7 +1739,7 @@ func (m *Intent_Message_Suggestion) GetTitle() string { // The collection of suggestions. type Intent_Message_Suggestions struct { // Required. The list of suggested replies. - Suggestions []*Intent_Message_Suggestion `protobuf:"bytes,1,rep,name=suggestions" json:"suggestions,omitempty"` + Suggestions []*Intent_Message_Suggestion `protobuf:"bytes,1,rep,name=suggestions,proto3" json:"suggestions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1780,10 +1780,10 @@ func (m *Intent_Message_Suggestions) GetSuggestions() []*Intent_Message_Suggesti // or website associated with this agent. type Intent_Message_LinkOutSuggestion struct { // Required. The name of the app or site this chip is linking to. - DestinationName string `protobuf:"bytes,1,opt,name=destination_name,json=destinationName" json:"destination_name,omitempty"` + DestinationName string `protobuf:"bytes,1,opt,name=destination_name,json=destinationName,proto3" json:"destination_name,omitempty"` // Required. The URI of the app or site to open when the user taps the // suggestion chip. - Uri string `protobuf:"bytes,2,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,2,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1830,9 +1830,9 @@ func (m *Intent_Message_LinkOutSuggestion) GetUri() string { // The card for presenting a list of options to select from. type Intent_Message_ListSelect struct { // Optional. The overall title of the list. - Title string `protobuf:"bytes,1,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,1,opt,name=title,proto3" json:"title,omitempty"` // Required. List items. - Items []*Intent_Message_ListSelect_Item `protobuf:"bytes,2,rep,name=items" json:"items,omitempty"` + Items []*Intent_Message_ListSelect_Item `protobuf:"bytes,2,rep,name=items,proto3" json:"items,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1879,13 +1879,13 @@ func (m *Intent_Message_ListSelect) GetItems() []*Intent_Message_ListSelect_Item // An item in the list. type Intent_Message_ListSelect_Item struct { // Required. Additional information about this option. - Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info" json:"info,omitempty"` + Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info,proto3" json:"info,omitempty"` // Required. The title of the list item. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // Optional. The main text describing the item. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Optional. The image to display. - Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"` + Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image,proto3" json:"image,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1946,7 +1946,7 @@ func (m *Intent_Message_ListSelect_Item) GetImage() *Intent_Message_Image { // The card for presenting a carousel of options to select from. type Intent_Message_CarouselSelect struct { // Required. Carousel items. - Items []*Intent_Message_CarouselSelect_Item `protobuf:"bytes,1,rep,name=items" json:"items,omitempty"` + Items []*Intent_Message_CarouselSelect_Item `protobuf:"bytes,1,rep,name=items,proto3" json:"items,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1986,13 +1986,13 @@ func (m *Intent_Message_CarouselSelect) GetItems() []*Intent_Message_CarouselSel // An item in the carousel. type Intent_Message_CarouselSelect_Item struct { // Required. Additional info about the option item. - Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info" json:"info,omitempty"` + Info *Intent_Message_SelectItemInfo `protobuf:"bytes,1,opt,name=info,proto3" json:"info,omitempty"` // Required. Title of the carousel item. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // Optional. The body text of the card. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Optional. The image to display. - Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image" json:"image,omitempty"` + Image *Intent_Message_Image `protobuf:"bytes,4,opt,name=image,proto3" json:"image,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2055,10 +2055,10 @@ func (m *Intent_Message_CarouselSelect_Item) GetImage() *Intent_Message_Image { type Intent_Message_SelectItemInfo struct { // Required. A unique key that will be sent back to the agent if this // response is given. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // Optional. A list of synonyms that can also be used to trigger this // item in dialog. - Synonyms []string `protobuf:"bytes,2,rep,name=synonyms" json:"synonyms,omitempty"` + Synonyms []string `protobuf:"bytes,2,rep,name=synonyms,proto3" json:"synonyms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2106,10 +2106,10 @@ func (m *Intent_Message_SelectItemInfo) GetSynonyms() []string { type Intent_FollowupIntentInfo struct { // The unique identifier of the followup intent. // Format: `projects//agent/intents/`. - FollowupIntentName string `protobuf:"bytes,1,opt,name=followup_intent_name,json=followupIntentName" json:"followup_intent_name,omitempty"` + FollowupIntentName string `protobuf:"bytes,1,opt,name=followup_intent_name,json=followupIntentName,proto3" json:"followup_intent_name,omitempty"` // The unique identifier of the followup intent parent. // Format: `projects//agent/intents/`. - ParentFollowupIntentName string `protobuf:"bytes,2,opt,name=parent_followup_intent_name,json=parentFollowupIntentName" json:"parent_followup_intent_name,omitempty"` + ParentFollowupIntentName string `protobuf:"bytes,2,opt,name=parent_followup_intent_name,json=parentFollowupIntentName,proto3" json:"parent_followup_intent_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2157,20 +2157,20 @@ func (m *Intent_FollowupIntentInfo) GetParentFollowupIntentName() string { type ListIntentsRequest struct { // Required. The agent to list all intents from. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The language to list training phrases, parameters and rich // messages for. If not specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2239,10 +2239,10 @@ func (m *ListIntentsRequest) GetPageToken() string { type ListIntentsResponse struct { // The list of agent intents. There will be a maximum number of items // returned based on the page_size field in the request. - Intents []*Intent `protobuf:"bytes,1,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,1,rep,name=intents,proto3" json:"intents,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2290,15 +2290,15 @@ func (m *ListIntentsResponse) GetNextPageToken() string { type GetIntentRequest struct { // Required. The name of the intent. // Format: `projects//agent/intents/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The language to retrieve training phrases, parameters and rich // messages for. If not specified, the agent's default language is used. // [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,3,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2353,17 +2353,17 @@ func (m *GetIntentRequest) GetIntentView() IntentView { type CreateIntentRequest struct { // Required. The agent to create a intent for. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The intent to create. - Intent *Intent `protobuf:"bytes,2,opt,name=intent" json:"intent,omitempty"` + Intent *Intent `protobuf:"bytes,2,opt,name=intent,proto3" json:"intent,omitempty"` // Optional. The language of training phrases, parameters and rich messages // defined in `intent`. If not specified, the agent's default language is // used. [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2425,17 +2425,17 @@ func (m *CreateIntentRequest) GetIntentView() IntentView { type UpdateIntentRequest struct { // Required. The intent to update. // Format: `projects//agent/intents/`. - Intent *Intent `protobuf:"bytes,1,opt,name=intent" json:"intent,omitempty"` + Intent *Intent `protobuf:"bytes,1,opt,name=intent,proto3" json:"intent,omitempty"` // Optional. The language of training phrases, parameters and rich messages // defined in `intent`. If not specified, the agent's default language is // used. [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,4,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2497,7 +2497,7 @@ func (m *UpdateIntentRequest) GetIntentView() IntentView { type DeleteIntentRequest struct { // Required. The name of the intent to delete. // Format: `projects//agent/intents/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2538,7 +2538,7 @@ func (m *DeleteIntentRequest) GetName() string { type BatchUpdateIntentsRequest struct { // Required. The name of the agent to update or create intents in. // Format: `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The source of the intent batch. // // Types that are valid to be assigned to IntentBatch: @@ -2550,11 +2550,11 @@ type BatchUpdateIntentsRequest struct { // used. [More than a dozen // languages](https://dialogflow.com/docs/reference/language) are supported. // Note: languages must be enabled in the agent, before they can be used. - LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,4,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Optional. The resource view to apply to the returned intent. - IntentView IntentView `protobuf:"varint,6,opt,name=intent_view,json=intentView,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` + IntentView IntentView `protobuf:"varint,6,opt,name=intent_view,json=intentView,proto3,enum=google.cloud.dialogflow.v2beta1.IntentView" json:"intent_view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2589,10 +2589,10 @@ type isBatchUpdateIntentsRequest_IntentBatch interface { } type BatchUpdateIntentsRequest_IntentBatchUri struct { - IntentBatchUri string `protobuf:"bytes,2,opt,name=intent_batch_uri,json=intentBatchUri,oneof"` + IntentBatchUri string `protobuf:"bytes,2,opt,name=intent_batch_uri,json=intentBatchUri,proto3,oneof"` } type BatchUpdateIntentsRequest_IntentBatchInline struct { - IntentBatchInline *IntentBatch `protobuf:"bytes,3,opt,name=intent_batch_inline,json=intentBatchInline,oneof"` + IntentBatchInline *IntentBatch `protobuf:"bytes,3,opt,name=intent_batch_inline,json=intentBatchInline,proto3,oneof"` } func (*BatchUpdateIntentsRequest_IntentBatchUri) isBatchUpdateIntentsRequest_IntentBatch() {} @@ -2720,7 +2720,7 @@ func _BatchUpdateIntentsRequest_OneofSizer(msg proto.Message) (n int) { // The response message for [Intents.BatchUpdateIntents][google.cloud.dialogflow.v2beta1.Intents.BatchUpdateIntents]. type BatchUpdateIntentsResponse struct { // The collection of updated or created intents. - Intents []*Intent `protobuf:"bytes,1,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,1,rep,name=intents,proto3" json:"intents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2761,10 +2761,10 @@ func (m *BatchUpdateIntentsResponse) GetIntents() []*Intent { type BatchDeleteIntentsRequest struct { // Required. The name of the agent to delete all entities types for. Format: // `projects//agent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The collection of intents to delete. Only intent `name` must be // filled in. - Intents []*Intent `protobuf:"bytes,2,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,2,rep,name=intents,proto3" json:"intents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2811,7 +2811,7 @@ func (m *BatchDeleteIntentsRequest) GetIntents() []*Intent { // This message is a wrapper around a collection of intents. type IntentBatch struct { // A collection of intents. - Intents []*Intent `protobuf:"bytes,1,rep,name=intents" json:"intents,omitempty"` + Intents []*Intent `protobuf:"bytes,1,rep,name=intents,proto3" json:"intents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2992,8 +2992,7 @@ func (c *intentsClient) BatchDeleteIntents(ctx context.Context, in *BatchDeleteI return out, nil } -// Server API for Intents service - +// IntentsServer is the server API for Intents service. type IntentsServer interface { // Returns the list of all intents in the specified agent. ListIntents(context.Context, *ListIntentsRequest) (*ListIntentsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session.pb.go index 02937e3c1..bbc870bfd 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session.pb.go @@ -145,9 +145,9 @@ type DetectIntentRequest struct { // appropriate . and . They can be a random numbers or // some type of user and session identifiers (preferably hashed). The length // of the and must not exceed 36 characters. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Optional. The parameters of this query. - QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams" json:"query_params,omitempty"` + QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams,proto3" json:"query_params,omitempty"` // Required. The input specification. It can be set to: // // 1. an audio config @@ -156,7 +156,7 @@ type DetectIntentRequest struct { // 2. a conversational query in the form of text, or // // 3. an event that specifies which intent to trigger. - QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput" json:"query_input,omitempty"` + QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput,proto3" json:"query_input,omitempty"` // Optional. The natural language speech audio to be processed. This field // should be populated iff `query_input` is set to an input audio config. // A single request can contain up to 1 minute of speech audio data. @@ -222,12 +222,12 @@ func (m *DetectIntentRequest) GetInputAudio() []byte { type DetectIntentResponse struct { // The unique identifier of the response. It can be used to // locate a response in the training example set or for reporting issues. - ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId" json:"response_id,omitempty"` + ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId,proto3" json:"response_id,omitempty"` // The results of the conversational query or event processing. - QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult" json:"query_result,omitempty"` + QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult,proto3" json:"query_result,omitempty"` // Specifies the status of the webhook request. `webhook_status` // is never populated in webhook requests. - WebhookStatus *status.Status `protobuf:"bytes,3,opt,name=webhook_status,json=webhookStatus" json:"webhook_status,omitempty"` + WebhookStatus *status.Status `protobuf:"bytes,3,opt,name=webhook_status,json=webhookStatus,proto3" json:"webhook_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -284,22 +284,22 @@ type QueryParameters struct { // [time zone database](https://www.iana.org/time-zones), e.g., // America/New_York, Europe/Paris. If not provided, the time zone specified in // agent settings is used. - TimeZone string `protobuf:"bytes,1,opt,name=time_zone,json=timeZone" json:"time_zone,omitempty"` + TimeZone string `protobuf:"bytes,1,opt,name=time_zone,json=timeZone,proto3" json:"time_zone,omitempty"` // Optional. The geo location of this conversational query. - GeoLocation *latlng.LatLng `protobuf:"bytes,2,opt,name=geo_location,json=geoLocation" json:"geo_location,omitempty"` + GeoLocation *latlng.LatLng `protobuf:"bytes,2,opt,name=geo_location,json=geoLocation,proto3" json:"geo_location,omitempty"` // Optional. The collection of contexts to be activated before this query is // executed. - Contexts []*Context `protobuf:"bytes,3,rep,name=contexts" json:"contexts,omitempty"` + Contexts []*Context `protobuf:"bytes,3,rep,name=contexts,proto3" json:"contexts,omitempty"` // Optional. Specifies whether to delete all contexts in the current session // before the new ones are activated. - ResetContexts bool `protobuf:"varint,4,opt,name=reset_contexts,json=resetContexts" json:"reset_contexts,omitempty"` + ResetContexts bool `protobuf:"varint,4,opt,name=reset_contexts,json=resetContexts,proto3" json:"reset_contexts,omitempty"` // Optional. The collection of session entity types to replace or extend // developer entities with for this query only. The entity synonyms apply // to all languages. - SessionEntityTypes []*SessionEntityType `protobuf:"bytes,5,rep,name=session_entity_types,json=sessionEntityTypes" json:"session_entity_types,omitempty"` + SessionEntityTypes []*SessionEntityType `protobuf:"bytes,5,rep,name=session_entity_types,json=sessionEntityTypes,proto3" json:"session_entity_types,omitempty"` // Optional. This field can be used to pass custom data into the webhook // associated with the agent. Arbitrary JSON objects are supported. - Payload *_struct.Struct `protobuf:"bytes,6,opt,name=payload" json:"payload,omitempty"` + Payload *_struct.Struct `protobuf:"bytes,6,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -421,13 +421,13 @@ type isQueryInput_Input interface { } type QueryInput_AudioConfig struct { - AudioConfig *InputAudioConfig `protobuf:"bytes,1,opt,name=audio_config,json=audioConfig,oneof"` + AudioConfig *InputAudioConfig `protobuf:"bytes,1,opt,name=audio_config,json=audioConfig,proto3,oneof"` } type QueryInput_Text struct { - Text *TextInput `protobuf:"bytes,2,opt,name=text,oneof"` + Text *TextInput `protobuf:"bytes,2,opt,name=text,proto3,oneof"` } type QueryInput_Event struct { - Event *EventInput `protobuf:"bytes,3,opt,name=event,oneof"` + Event *EventInput `protobuf:"bytes,3,opt,name=event,proto3,oneof"` } func (*QueryInput_AudioConfig) isQueryInput_Input() {} @@ -564,11 +564,11 @@ type QueryResult struct { // contains the speech recognition result. If speech recognizer produced // multiple alternatives, a particular one is picked. // - If an event was provided as input, `query_text` is not set. - QueryText string `protobuf:"bytes,1,opt,name=query_text,json=queryText" json:"query_text,omitempty"` + QueryText string `protobuf:"bytes,1,opt,name=query_text,json=queryText,proto3" json:"query_text,omitempty"` // The language that was triggered during intent detection. // See [Language Support](https://dialogflow.com/docs/reference/language) // for a list of the currently supported language codes. - LanguageCode string `protobuf:"bytes,15,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,15,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // The Speech recognition confidence between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. The default of 0.0 is a sentinel value indicating that confidence @@ -578,42 +578,42 @@ type QueryResult struct { // even set. In particular this field isn't set in Webhook calls and for // StreamingDetectIntent since the streaming endpoint has separate confidence // estimates per portion of the audio in StreamingRecognitionResult. - SpeechRecognitionConfidence float32 `protobuf:"fixed32,2,opt,name=speech_recognition_confidence,json=speechRecognitionConfidence" json:"speech_recognition_confidence,omitempty"` + SpeechRecognitionConfidence float32 `protobuf:"fixed32,2,opt,name=speech_recognition_confidence,json=speechRecognitionConfidence,proto3" json:"speech_recognition_confidence,omitempty"` // The action name from the matched intent. - Action string `protobuf:"bytes,3,opt,name=action" json:"action,omitempty"` + Action string `protobuf:"bytes,3,opt,name=action,proto3" json:"action,omitempty"` // The collection of extracted parameters. - Parameters *_struct.Struct `protobuf:"bytes,4,opt,name=parameters" json:"parameters,omitempty"` + Parameters *_struct.Struct `protobuf:"bytes,4,opt,name=parameters,proto3" json:"parameters,omitempty"` // This field is set to: // - `false` if the matched intent has required parameters and not all of // the required parameter values have been collected. // - `true` if all required parameter values have been collected, or if the // matched intent doesn't contain any required parameters. - AllRequiredParamsPresent bool `protobuf:"varint,5,opt,name=all_required_params_present,json=allRequiredParamsPresent" json:"all_required_params_present,omitempty"` + AllRequiredParamsPresent bool `protobuf:"varint,5,opt,name=all_required_params_present,json=allRequiredParamsPresent,proto3" json:"all_required_params_present,omitempty"` // The text to be pronounced to the user or shown on the screen. - FulfillmentText string `protobuf:"bytes,6,opt,name=fulfillment_text,json=fulfillmentText" json:"fulfillment_text,omitempty"` + FulfillmentText string `protobuf:"bytes,6,opt,name=fulfillment_text,json=fulfillmentText,proto3" json:"fulfillment_text,omitempty"` // The collection of rich messages to present to the user. - FulfillmentMessages []*Intent_Message `protobuf:"bytes,7,rep,name=fulfillment_messages,json=fulfillmentMessages" json:"fulfillment_messages,omitempty"` + FulfillmentMessages []*Intent_Message `protobuf:"bytes,7,rep,name=fulfillment_messages,json=fulfillmentMessages,proto3" json:"fulfillment_messages,omitempty"` // If the query was fulfilled by a webhook call, this field is set to the // value of the `source` field returned in the webhook response. - WebhookSource string `protobuf:"bytes,8,opt,name=webhook_source,json=webhookSource" json:"webhook_source,omitempty"` + WebhookSource string `protobuf:"bytes,8,opt,name=webhook_source,json=webhookSource,proto3" json:"webhook_source,omitempty"` // If the query was fulfilled by a webhook call, this field is set to the // value of the `payload` field returned in the webhook response. - WebhookPayload *_struct.Struct `protobuf:"bytes,9,opt,name=webhook_payload,json=webhookPayload" json:"webhook_payload,omitempty"` + WebhookPayload *_struct.Struct `protobuf:"bytes,9,opt,name=webhook_payload,json=webhookPayload,proto3" json:"webhook_payload,omitempty"` // The collection of output contexts. If applicable, // `output_contexts.parameters` contains entries with name // `.original` containing the original parameter values // before the query. - OutputContexts []*Context `protobuf:"bytes,10,rep,name=output_contexts,json=outputContexts" json:"output_contexts,omitempty"` + OutputContexts []*Context `protobuf:"bytes,10,rep,name=output_contexts,json=outputContexts,proto3" json:"output_contexts,omitempty"` // The intent that matched the conversational query. Some, not // all fields are filled in this message, including but not limited to: // `name`, `display_name` and `webhook_state`. - Intent *Intent `protobuf:"bytes,11,opt,name=intent" json:"intent,omitempty"` + Intent *Intent `protobuf:"bytes,11,opt,name=intent,proto3" json:"intent,omitempty"` // The intent detection confidence. Values range from 0.0 // (completely uncertain) to 1.0 (completely certain). - IntentDetectionConfidence float32 `protobuf:"fixed32,12,opt,name=intent_detection_confidence,json=intentDetectionConfidence" json:"intent_detection_confidence,omitempty"` + IntentDetectionConfidence float32 `protobuf:"fixed32,12,opt,name=intent_detection_confidence,json=intentDetectionConfidence,proto3" json:"intent_detection_confidence,omitempty"` // The free-form diagnostic info. For example, this field // could contain webhook call latency. - DiagnosticInfo *_struct.Struct `protobuf:"bytes,14,opt,name=diagnostic_info,json=diagnosticInfo" json:"diagnostic_info,omitempty"` + DiagnosticInfo *_struct.Struct `protobuf:"bytes,14,opt,name=diagnostic_info,json=diagnosticInfo,proto3" json:"diagnostic_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -764,9 +764,9 @@ type StreamingDetectIntentRequest struct { // appropriate . and . They can be a random numbers or // some type of user and session identifiers (preferably hashed). The length // of the and must not exceed 36 characters. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Optional. The parameters of this query. - QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams" json:"query_params,omitempty"` + QueryParams *QueryParameters `protobuf:"bytes,2,opt,name=query_params,json=queryParams,proto3" json:"query_params,omitempty"` // Required. The input specification. It can be set to: // // 1. an audio config which instructs the speech recognizer how to process @@ -775,7 +775,7 @@ type StreamingDetectIntentRequest struct { // 2. a conversational query in the form of text, or // // 3. an event that specifies which intent to trigger. - QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput" json:"query_input,omitempty"` + QueryInput *QueryInput `protobuf:"bytes,3,opt,name=query_input,json=queryInput,proto3" json:"query_input,omitempty"` // Optional. If `false` (default), recognition does not cease until the // client closes the stream. // If `true`, the recognizer will detect a single spoken utterance in input @@ -784,7 +784,7 @@ type StreamingDetectIntentRequest struct { // client should close the stream and start a new request with a new stream as // needed. // This setting is ignored when `query_input` is a piece of text or an event. - SingleUtterance bool `protobuf:"varint,4,opt,name=single_utterance,json=singleUtterance" json:"single_utterance,omitempty"` + SingleUtterance bool `protobuf:"varint,4,opt,name=single_utterance,json=singleUtterance,proto3" json:"single_utterance,omitempty"` // Optional. The input audio content to be recognized. Must be sent if // `query_input` was set to a streaming input audio config. The complete audio // over all streaming messages must not exceed 1 minute. @@ -868,13 +868,13 @@ func (m *StreamingDetectIntentRequest) GetInputAudio() []byte { type StreamingDetectIntentResponse struct { // The unique identifier of the response. It can be used to // locate a response in the training example set or for reporting issues. - ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId" json:"response_id,omitempty"` + ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId,proto3" json:"response_id,omitempty"` // The result of speech recognition. - RecognitionResult *StreamingRecognitionResult `protobuf:"bytes,2,opt,name=recognition_result,json=recognitionResult" json:"recognition_result,omitempty"` + RecognitionResult *StreamingRecognitionResult `protobuf:"bytes,2,opt,name=recognition_result,json=recognitionResult,proto3" json:"recognition_result,omitempty"` // The result of the conversational query or event processing. - QueryResult *QueryResult `protobuf:"bytes,3,opt,name=query_result,json=queryResult" json:"query_result,omitempty"` + QueryResult *QueryResult `protobuf:"bytes,3,opt,name=query_result,json=queryResult,proto3" json:"query_result,omitempty"` // Specifies the status of the webhook request. - WebhookStatus *status.Status `protobuf:"bytes,4,opt,name=webhook_status,json=webhookStatus" json:"webhook_status,omitempty"` + WebhookStatus *status.Status `protobuf:"bytes,4,opt,name=webhook_status,json=webhookStatus,proto3" json:"webhook_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -967,16 +967,16 @@ func (m *StreamingDetectIntentResponse) GetWebhookStatus() *status.Status { // * for `MESSAGE_TYPE_END_OF_SINGLE_UTTERANCE`: only `event_type`. type StreamingRecognitionResult struct { // Type of the result message. - MessageType StreamingRecognitionResult_MessageType `protobuf:"varint,1,opt,name=message_type,json=messageType,enum=google.cloud.dialogflow.v2beta1.StreamingRecognitionResult_MessageType" json:"message_type,omitempty"` + MessageType StreamingRecognitionResult_MessageType `protobuf:"varint,1,opt,name=message_type,json=messageType,proto3,enum=google.cloud.dialogflow.v2beta1.StreamingRecognitionResult_MessageType" json:"message_type,omitempty"` // Transcript text representing the words that the user spoke. // Populated if and only if `event_type` = `RECOGNITION_EVENT_TRANSCRIPT`. - Transcript string `protobuf:"bytes,2,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,2,opt,name=transcript,proto3" json:"transcript,omitempty"` // The default of 0.0 is a sentinel value indicating `confidence` was not set. // If `false`, the `StreamingRecognitionResult` represents an // interim result that may change. If `true`, the recognizer will not return // any further hypotheses about this piece of the audio. May only be populated // for `event_type` = `RECOGNITION_EVENT_TRANSCRIPT`. - IsFinal bool `protobuf:"varint,3,opt,name=is_final,json=isFinal" json:"is_final,omitempty"` + IsFinal bool `protobuf:"varint,3,opt,name=is_final,json=isFinal,proto3" json:"is_final,omitempty"` // The Speech confidence between 0.0 and 1.0 for the current portion of audio. // A higher number indicates an estimated greater likelihood that the // recognized words are correct. The default of 0.0 is a sentinel value @@ -984,7 +984,7 @@ type StreamingRecognitionResult struct { // // This field is typically only provided if `is_final` is true and you should // not rely on it being accurate or even set. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1045,22 +1045,22 @@ func (m *StreamingRecognitionResult) GetConfidence() float32 { // Instructs the speech recognizer how to process the audio content. type InputAudioConfig struct { // Required. Audio encoding of the audio content to process. - AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,enum=google.cloud.dialogflow.v2beta1.AudioEncoding" json:"audio_encoding,omitempty"` + AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,proto3,enum=google.cloud.dialogflow.v2beta1.AudioEncoding" json:"audio_encoding,omitempty"` // Required. Sample rate (in Hertz) of the audio content sent in the query. // Refer to [Cloud Speech API documentation](/speech/docs/basics) for more // details. - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` // Required. The language of the supplied audio. Dialogflow does not do // translations. See [Language // Support](https://dialogflow.com/docs/languages) for a list of the // currently supported language codes. Note that queries in the same session // do not necessarily need to specify the same language. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. The collection of phrase hints which are used to boost accuracy // of speech recognition. // Refer to [Cloud Speech API documentation](/speech/docs/basics#phrase-hints) // for more details. - PhraseHints []string `protobuf:"bytes,4,rep,name=phrase_hints,json=phraseHints" json:"phrase_hints,omitempty"` + PhraseHints []string `protobuf:"bytes,4,rep,name=phrase_hints,json=phraseHints,proto3" json:"phrase_hints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1122,12 +1122,12 @@ func (m *InputAudioConfig) GetPhraseHints() []string { type TextInput struct { // Required. The UTF-8 encoded natural language text to be processed. // Text length must not exceed 256 bytes. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Required. The language of this conversational query. See [Language // Support](https://dialogflow.com/docs/languages) for a list of the // currently supported language codes. Note that queries in the same session // do not necessarily need to specify the same language. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1178,14 +1178,14 @@ func (m *TextInput) GetLanguageCode() string { // `“Hello #welcome_event.name! What can I do for you today?”`. type EventInput struct { // Required. The unique identifier of the event. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The collection of parameters associated with the event. - Parameters *_struct.Struct `protobuf:"bytes,2,opt,name=parameters" json:"parameters,omitempty"` + Parameters *_struct.Struct `protobuf:"bytes,2,opt,name=parameters,proto3" json:"parameters,omitempty"` // Required. The language of this query. See [Language // Support](https://dialogflow.com/docs/languages) for a list of the // currently supported language codes. Note that queries in the same session // do not necessarily need to specify the same language. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1323,8 +1323,7 @@ func (x *sessionsStreamingDetectIntentClient) Recv() (*StreamingDetectIntentResp return m, nil } -// Server API for Sessions service - +// SessionsServer is the server API for Sessions service. type SessionsServer interface { // Processes a natural language query and returns structured, actionable data // as a result. This method is not idempotent, because it may cause contexts diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session_entity_type.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session_entity_type.pb.go index 38681eb53..8e4857281 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session_entity_type.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/session_entity_type.pb.go @@ -79,13 +79,13 @@ type SessionEntityType struct { // Note: Environments and users are under construction and will be available // soon. If is not specified, we assume default 'draft' // environment. If is not specified, we assume default '-' user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. Indicates whether the additional data should override or // supplement the developer entity type definition. - EntityOverrideMode SessionEntityType_EntityOverrideMode `protobuf:"varint,2,opt,name=entity_override_mode,json=entityOverrideMode,enum=google.cloud.dialogflow.v2beta1.SessionEntityType_EntityOverrideMode" json:"entity_override_mode,omitempty"` + EntityOverrideMode SessionEntityType_EntityOverrideMode `protobuf:"varint,2,opt,name=entity_override_mode,json=entityOverrideMode,proto3,enum=google.cloud.dialogflow.v2beta1.SessionEntityType_EntityOverrideMode" json:"entity_override_mode,omitempty"` // Required. The collection of entities associated with this session entity // type. - Entities []*EntityType_Entity `protobuf:"bytes,3,rep,name=entities" json:"entities,omitempty"` + Entities []*EntityType_Entity `protobuf:"bytes,3,rep,name=entities,proto3" json:"entities,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -145,12 +145,12 @@ type ListSessionEntityTypesRequest struct { // Note: Environments and users are under construction and will be available // soon. If is not specified, we assume default 'draft' // environment. If is not specified, we assume default '-' user. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of items to return in a single page. By // default 100 and at most 1000. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. The next_page_token value returned from a previous list request. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -205,10 +205,10 @@ func (m *ListSessionEntityTypesRequest) GetPageToken() string { type ListSessionEntityTypesResponse struct { // The list of session entity types. There will be a maximum number of items // returned based on the page_size field in the request. - SessionEntityTypes []*SessionEntityType `protobuf:"bytes,1,rep,name=session_entity_types,json=sessionEntityTypes" json:"session_entity_types,omitempty"` + SessionEntityTypes []*SessionEntityType `protobuf:"bytes,1,rep,name=session_entity_types,json=sessionEntityTypes,proto3" json:"session_entity_types,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,7 +262,7 @@ type GetSessionEntityTypeRequest struct { // Note: Environments and users re under construction and will be available // soon. If is not specified, we assume default 'draft' // environment. If is not specified, we assume default '-' user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -308,9 +308,9 @@ type CreateSessionEntityTypeRequest struct { // Note: Environments and users are under construction and will be available // soon. If is not specified, we assume default 'draft' // environment. If is not specified, we assume default '-' user. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The session entity type to create. - SessionEntityType *SessionEntityType `protobuf:"bytes,2,opt,name=session_entity_type,json=sessionEntityType" json:"session_entity_type,omitempty"` + SessionEntityType *SessionEntityType `protobuf:"bytes,2,opt,name=session_entity_type,json=sessionEntityType,proto3" json:"session_entity_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -364,9 +364,9 @@ type UpdateSessionEntityTypeRequest struct { // Note: Environments and users are under construction and will be available // soon. If is not specified, we assume default 'draft' // environment. If is not specified, we assume default '-' user. - SessionEntityType *SessionEntityType `protobuf:"bytes,1,opt,name=session_entity_type,json=sessionEntityType" json:"session_entity_type,omitempty"` + SessionEntityType *SessionEntityType `protobuf:"bytes,1,opt,name=session_entity_type,json=sessionEntityType,proto3" json:"session_entity_type,omitempty"` // Optional. The mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -420,7 +420,7 @@ type DeleteSessionEntityTypeRequest struct { // Note: Environments and users are under construction and will be available // soon. If is not specified, we assume default 'draft' // environment. If is not specified, we assume default '-' user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -545,8 +545,7 @@ func (c *sessionEntityTypesClient) DeleteSessionEntityType(ctx context.Context, return out, nil } -// Server API for SessionEntityTypes service - +// SessionEntityTypesServer is the server API for SessionEntityTypes service. type SessionEntityTypesServer interface { // Returns the list of all session entity types in the specified session. ListSessionEntityTypes(context.Context, *ListSessionEntityTypesRequest) (*ListSessionEntityTypesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/webhook.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/webhook.pb.go index 75ff09c51..dfd7c5d9b 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/webhook.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/dialogflow/v2beta1/webhook.pb.go @@ -25,16 +25,16 @@ type WebhookRequest struct { // The unique identifier of detectIntent request session. // Can be used to identify end-user inside webhook implementation. // Format: `projects//agent/sessions/`. - Session string `protobuf:"bytes,4,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,4,opt,name=session,proto3" json:"session,omitempty"` // The unique identifier of the response. Contains the same value as // `[Streaming]DetectIntentResponse.response_id`. - ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId" json:"response_id,omitempty"` + ResponseId string `protobuf:"bytes,1,opt,name=response_id,json=responseId,proto3" json:"response_id,omitempty"` // The result of the conversational query or event processing. Contains the // same value as `[Streaming]DetectIntentResponse.query_result`. - QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult" json:"query_result,omitempty"` + QueryResult *QueryResult `protobuf:"bytes,2,opt,name=query_result,json=queryResult,proto3" json:"query_result,omitempty"` // Optional. The contents of the original request that was passed to // `[Streaming]DetectIntent` call. - OriginalDetectIntentRequest *OriginalDetectIntentRequest `protobuf:"bytes,3,opt,name=original_detect_intent_request,json=originalDetectIntentRequest" json:"original_detect_intent_request,omitempty"` + OriginalDetectIntentRequest *OriginalDetectIntentRequest `protobuf:"bytes,3,opt,name=original_detect_intent_request,json=originalDetectIntentRequest,proto3" json:"original_detect_intent_request,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -96,12 +96,12 @@ func (m *WebhookRequest) GetOriginalDetectIntentRequest() *OriginalDetectIntentR type WebhookResponse struct { // Optional. The text to be shown on the screen. This value is passed directly // to `QueryResult.fulfillment_text`. - FulfillmentText string `protobuf:"bytes,1,opt,name=fulfillment_text,json=fulfillmentText" json:"fulfillment_text,omitempty"` + FulfillmentText string `protobuf:"bytes,1,opt,name=fulfillment_text,json=fulfillmentText,proto3" json:"fulfillment_text,omitempty"` // Optional. The collection of rich messages to present to the user. This // value is passed directly to `QueryResult.fulfillment_messages`. - FulfillmentMessages []*Intent_Message `protobuf:"bytes,2,rep,name=fulfillment_messages,json=fulfillmentMessages" json:"fulfillment_messages,omitempty"` + FulfillmentMessages []*Intent_Message `protobuf:"bytes,2,rep,name=fulfillment_messages,json=fulfillmentMessages,proto3" json:"fulfillment_messages,omitempty"` // Optional. This value is passed directly to `QueryResult.webhook_source`. - Source string `protobuf:"bytes,3,opt,name=source" json:"source,omitempty"` + Source string `protobuf:"bytes,3,opt,name=source,proto3" json:"source,omitempty"` // Optional. This value is passed directly to `QueryResult.webhook_payload`. // See the related `fulfillment_messages[i].payload field`, which may be used // as an alternative to this field. @@ -125,13 +125,13 @@ type WebhookResponse struct { // } // } // } - Payload *_struct.Struct `protobuf:"bytes,4,opt,name=payload" json:"payload,omitempty"` + Payload *_struct.Struct `protobuf:"bytes,4,opt,name=payload,proto3" json:"payload,omitempty"` // Optional. The collection of output contexts. This value is passed directly // to `QueryResult.output_contexts`. - OutputContexts []*Context `protobuf:"bytes,5,rep,name=output_contexts,json=outputContexts" json:"output_contexts,omitempty"` + OutputContexts []*Context `protobuf:"bytes,5,rep,name=output_contexts,json=outputContexts,proto3" json:"output_contexts,omitempty"` // Optional. Makes the platform immediately invoke another `DetectIntent` call // internally with the specified event as input. - FollowupEventInput *EventInput `protobuf:"bytes,6,opt,name=followup_event_input,json=followupEventInput" json:"followup_event_input,omitempty"` + FollowupEventInput *EventInput `protobuf:"bytes,6,opt,name=followup_event_input,json=followupEventInput,proto3" json:"followup_event_input,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -208,10 +208,10 @@ func (m *WebhookResponse) GetFollowupEventInput() *EventInput { type OriginalDetectIntentRequest struct { // The source of this request, e.g., `google`, `facebook`, `slack`. It is set // by Dialogflow-owned servers. - Source string `protobuf:"bytes,1,opt,name=source" json:"source,omitempty"` + Source string `protobuf:"bytes,1,opt,name=source,proto3" json:"source,omitempty"` // Optional. This field is set to the value of `QueryParameters.payload` field // passed in the request. - Payload *_struct.Struct `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *_struct.Struct `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/functions.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/functions.pb.go index fb596e385..90e79f940 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/functions.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/functions.pb.go @@ -72,7 +72,7 @@ func (CloudFunctionStatus) EnumDescriptor() ([]byte, []int) { type CloudFunction struct { // A user-defined name of the function. Function names must be unique // globally and match pattern `projects/*/locations/*/functions/*` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The location of the function source code. // // Types that are valid to be assigned to SourceCode: @@ -86,29 +86,29 @@ type CloudFunction struct { // *CloudFunction_EventTrigger Trigger isCloudFunction_Trigger `protobuf_oneof:"trigger"` // Output only. Status of the function deployment. - Status CloudFunctionStatus `protobuf:"varint,7,opt,name=status,enum=google.cloud.functions.v1beta2.CloudFunctionStatus" json:"status,omitempty"` + Status CloudFunctionStatus `protobuf:"varint,7,opt,name=status,proto3,enum=google.cloud.functions.v1beta2.CloudFunctionStatus" json:"status,omitempty"` // Output only. Name of the most recent operation modifying the function. If // the function status is `DEPLOYING` or `DELETING`, then it points to the // active operation. - LatestOperation string `protobuf:"bytes,8,opt,name=latest_operation,json=latestOperation" json:"latest_operation,omitempty"` + LatestOperation string `protobuf:"bytes,8,opt,name=latest_operation,json=latestOperation,proto3" json:"latest_operation,omitempty"` // The name of the function (as defined in source code) that will be // executed. Defaults to the resource name suffix, if not specified. For // backward compatibility, if function with given name is not found, then the // system will try to use function named "function". // For Node.js this is name of a function exported by the module specified // in `source_location`. - EntryPoint string `protobuf:"bytes,9,opt,name=entry_point,json=entryPoint" json:"entry_point,omitempty"` + EntryPoint string `protobuf:"bytes,9,opt,name=entry_point,json=entryPoint,proto3" json:"entry_point,omitempty"` // The function execution timeout. Execution is considered failed and // can be terminated if the function is not completed at the end of the // timeout period. Defaults to 60 seconds. - Timeout *duration.Duration `protobuf:"bytes,10,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,10,opt,name=timeout,proto3" json:"timeout,omitempty"` // The amount of memory in MB available for a function. // Defaults to 256MB. - AvailableMemoryMb int32 `protobuf:"varint,11,opt,name=available_memory_mb,json=availableMemoryMb" json:"available_memory_mb,omitempty"` + AvailableMemoryMb int32 `protobuf:"varint,11,opt,name=available_memory_mb,json=availableMemoryMb,proto3" json:"available_memory_mb,omitempty"` // Output only. The service account of the function. - ServiceAccount string `protobuf:"bytes,13,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount string `protobuf:"bytes,13,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // Output only. The last update timestamp of a Cloud Function. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,15,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,15,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -146,16 +146,16 @@ type isCloudFunction_Trigger interface { } type CloudFunction_SourceArchiveUrl struct { - SourceArchiveUrl string `protobuf:"bytes,14,opt,name=source_archive_url,json=sourceArchiveUrl,oneof"` + SourceArchiveUrl string `protobuf:"bytes,14,opt,name=source_archive_url,json=sourceArchiveUrl,proto3,oneof"` } type CloudFunction_SourceRepository struct { - SourceRepository *SourceRepository `protobuf:"bytes,3,opt,name=source_repository,json=sourceRepository,oneof"` + SourceRepository *SourceRepository `protobuf:"bytes,3,opt,name=source_repository,json=sourceRepository,proto3,oneof"` } type CloudFunction_HttpsTrigger struct { - HttpsTrigger *HTTPSTrigger `protobuf:"bytes,6,opt,name=https_trigger,json=httpsTrigger,oneof"` + HttpsTrigger *HTTPSTrigger `protobuf:"bytes,6,opt,name=https_trigger,json=httpsTrigger,proto3,oneof"` } type CloudFunction_EventTrigger struct { - EventTrigger *EventTrigger `protobuf:"bytes,12,opt,name=event_trigger,json=eventTrigger,oneof"` + EventTrigger *EventTrigger `protobuf:"bytes,12,opt,name=event_trigger,json=eventTrigger,proto3,oneof"` } func (*CloudFunction_SourceArchiveUrl) isCloudFunction_SourceCode() {} @@ -383,7 +383,7 @@ func _CloudFunction_OneofSizer(msg proto.Message) (n int) { // Describes HTTPSTrigger, could be used to connect web hooks to function. type HTTPSTrigger struct { // Output only. The deployed url for the function. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -433,13 +433,13 @@ type EventTrigger struct { // // Handle a write to the Firebase Realtime Database // `providers/firebase.database/eventTypes/data.write` - EventType string `protobuf:"bytes,1,opt,name=event_type,json=eventType" json:"event_type,omitempty"` + EventType string `protobuf:"bytes,1,opt,name=event_type,json=eventType,proto3" json:"event_type,omitempty"` // Which instance of the source's service should send events. E.g. for Pub/Sub // this would be a Pub/Sub topic at `projects/*/topics/*`. For Google Cloud // Storage this would be a bucket at `projects/*/buckets/*`. For any source // that only supports one instance per-project, this should be the name of the // project (`projects/*`) - Resource string `protobuf:"bytes,2,opt,name=resource" json:"resource,omitempty"` + Resource string `protobuf:"bytes,2,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -488,12 +488,12 @@ type SourceRepository struct { // URL to the hosted repository where the function is defined. Only paths in // https://source.developers.google.com domain are supported. The path should // contain the name of the repository. - RepositoryUrl string `protobuf:"bytes,1,opt,name=repository_url,json=repositoryUrl" json:"repository_url,omitempty"` + RepositoryUrl string `protobuf:"bytes,1,opt,name=repository_url,json=repositoryUrl,proto3" json:"repository_url,omitempty"` // The path within the repository where the function is defined. The path // should point to the directory where Cloud Functions files are located. Use // "/" if the function is defined directly in the root directory of a // repository. - SourcePath string `protobuf:"bytes,2,opt,name=source_path,json=sourcePath" json:"source_path,omitempty"` + SourcePath string `protobuf:"bytes,2,opt,name=source_path,json=sourcePath,proto3" json:"source_path,omitempty"` // The version of a function. Defaults to the latest version of the master // branch. // @@ -507,7 +507,7 @@ type SourceRepository struct { // branch, it will be the revision id of the latest change on this branch at // that time. If user deployed from revision then this value will be always // equal to the revision specified by the user. - DeployedRevision string `protobuf:"bytes,6,opt,name=deployed_revision,json=deployedRevision" json:"deployed_revision,omitempty"` + DeployedRevision string `protobuf:"bytes,6,opt,name=deployed_revision,json=deployedRevision,proto3" json:"deployed_revision,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -542,13 +542,13 @@ type isSourceRepository_Version interface { } type SourceRepository_Branch struct { - Branch string `protobuf:"bytes,3,opt,name=branch,oneof"` + Branch string `protobuf:"bytes,3,opt,name=branch,proto3,oneof"` } type SourceRepository_Tag struct { - Tag string `protobuf:"bytes,4,opt,name=tag,oneof"` + Tag string `protobuf:"bytes,4,opt,name=tag,proto3,oneof"` } type SourceRepository_Revision struct { - Revision string `protobuf:"bytes,5,opt,name=revision,oneof"` + Revision string `protobuf:"bytes,5,opt,name=revision,proto3,oneof"` } func (*SourceRepository_Branch) isSourceRepository_Version() {} @@ -689,9 +689,9 @@ func _SourceRepository_OneofSizer(msg proto.Message) (n int) { type CreateFunctionRequest struct { // The project and location in which the function should be created, specified // in the format `projects/*/locations/*` - Location string `protobuf:"bytes,1,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,1,opt,name=location,proto3" json:"location,omitempty"` // Function to be created. - Function *CloudFunction `protobuf:"bytes,2,opt,name=function" json:"function,omitempty"` + Function *CloudFunction `protobuf:"bytes,2,opt,name=function,proto3" json:"function,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -738,9 +738,9 @@ func (m *CreateFunctionRequest) GetFunction() *CloudFunction { // Request for the `UpdateFunction` method. type UpdateFunctionRequest struct { // The name of the function to be updated. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // New version of the function. - Function *CloudFunction `protobuf:"bytes,2,opt,name=function" json:"function,omitempty"` + Function *CloudFunction `protobuf:"bytes,2,opt,name=function,proto3" json:"function,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -787,7 +787,7 @@ func (m *UpdateFunctionRequest) GetFunction() *CloudFunction { // Request for the `GetFunction` method. type GetFunctionRequest struct { // The name of the function which details should be obtained. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -830,14 +830,14 @@ type ListFunctionsRequest struct { // specified in the format `projects/*/locations/*` // If you want to list functions in all locations, use "-" in place of a // location. - Location string `protobuf:"bytes,1,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,1,opt,name=location,proto3" json:"location,omitempty"` // Maximum number of functions to return per call. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last // `ListFunctionsResponse`; indicates that // this is a continuation of a prior `ListFunctions` call, and that the // system should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -891,12 +891,12 @@ func (m *ListFunctionsRequest) GetPageToken() string { // Response for the `ListFunctions` method. type ListFunctionsResponse struct { // The functions that match the request. - Functions []*CloudFunction `protobuf:"bytes,1,rep,name=functions" json:"functions,omitempty"` + Functions []*CloudFunction `protobuf:"bytes,1,rep,name=functions,proto3" json:"functions,omitempty"` // If not empty, indicates that there may be more functions that match // the request; this value should be passed in a new // [google.cloud.functions.v1beta2.ListFunctionsRequest][] // to get more functions. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -943,7 +943,7 @@ func (m *ListFunctionsResponse) GetNextPageToken() string { // Request for the `DeleteFunction` method. type DeleteFunctionRequest struct { // The name of the function which should be deleted. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -983,9 +983,9 @@ func (m *DeleteFunctionRequest) GetName() string { // Request for the `CallFunction` method. type CallFunctionRequest struct { // The name of the function to be called. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Input to be passed to the function. - Data string `protobuf:"bytes,2,opt,name=data" json:"data,omitempty"` + Data string `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1032,13 +1032,13 @@ func (m *CallFunctionRequest) GetData() string { // Response of `CallFunction` method. type CallFunctionResponse struct { // Execution id of function invocation. - ExecutionId string `protobuf:"bytes,1,opt,name=execution_id,json=executionId" json:"execution_id,omitempty"` + ExecutionId string `protobuf:"bytes,1,opt,name=execution_id,json=executionId,proto3" json:"execution_id,omitempty"` // Result populated for successful execution of synchronous function. Will // not be populated if function does not return a result through context. - Result string `protobuf:"bytes,2,opt,name=result" json:"result,omitempty"` + Result string `protobuf:"bytes,2,opt,name=result,proto3" json:"result,omitempty"` // Either system or user-function generated error. Set if execution // was not successful. - Error string `protobuf:"bytes,3,opt,name=error" json:"error,omitempty"` + Error string `protobuf:"bytes,3,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1198,8 +1198,7 @@ func (c *cloudFunctionsServiceClient) CallFunction(ctx context.Context, in *Call return out, nil } -// Server API for CloudFunctionsService service - +// CloudFunctionsServiceServer is the server API for CloudFunctionsService service. type CloudFunctionsServiceServer interface { // Returns a list of functions that belong to the requested project. ListFunctions(context.Context, *ListFunctionsRequest) (*ListFunctionsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/operations.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/operations.pb.go index 5cf56d40a..56990f47b 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/operations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/functions/v1beta2/operations.pb.go @@ -58,11 +58,11 @@ func (OperationType) EnumDescriptor() ([]byte, []int) { type OperationMetadataV1Beta2 struct { // Target of the operation - for example // projects/project-1/locations/region-1/functions/function-1 - Target string `protobuf:"bytes,1,opt,name=target" json:"target,omitempty"` + Target string `protobuf:"bytes,1,opt,name=target,proto3" json:"target,omitempty"` // Type of operation. - Type OperationType `protobuf:"varint,2,opt,name=type,enum=google.cloud.functions.v1beta2.OperationType" json:"type,omitempty"` + Type OperationType `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.functions.v1beta2.OperationType" json:"type,omitempty"` // The original request that started the operation. - Request *any.Any `protobuf:"bytes,3,opt,name=request" json:"request,omitempty"` + Request *any.Any `protobuf:"bytes,3,opt,name=request,proto3" json:"request,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/device_manager.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/device_manager.pb.go index 70ee97bac..58891f700 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/device_manager.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/device_manager.pb.go @@ -33,11 +33,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type CreateDeviceRegistryRequest struct { // The project and cloud region where this device registry must be created. // For example, `projects/example-project/locations/us-central1`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The device registry. The field `name` must be empty. The server will // generate that field from the device registry `id` provided and the // `parent` field. - DeviceRegistry *DeviceRegistry `protobuf:"bytes,2,opt,name=device_registry,json=deviceRegistry" json:"device_registry,omitempty"` + DeviceRegistry *DeviceRegistry `protobuf:"bytes,2,opt,name=device_registry,json=deviceRegistry,proto3" json:"device_registry,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -85,7 +85,7 @@ func (m *CreateDeviceRegistryRequest) GetDeviceRegistry() *DeviceRegistry { type GetDeviceRegistryRequest struct { // The name of the device registry. For example, // `projects/example-project/locations/us-central1/registries/my-registry`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -126,7 +126,7 @@ func (m *GetDeviceRegistryRequest) GetName() string { type DeleteDeviceRegistryRequest struct { // The name of the device registry. For example, // `projects/example-project/locations/us-central1/registries/my-registry`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -168,13 +168,13 @@ type UpdateDeviceRegistryRequest struct { // The new values for the device registry. The `id` field must be empty, and // the `name` field must indicate the path of the resource. For example, // `projects/example-project/locations/us-central1/registries/my-registry`. - DeviceRegistry *DeviceRegistry `protobuf:"bytes,1,opt,name=device_registry,json=deviceRegistry" json:"device_registry,omitempty"` + DeviceRegistry *DeviceRegistry `protobuf:"bytes,1,opt,name=device_registry,json=deviceRegistry,proto3" json:"device_registry,omitempty"` // Only updates the `device_registry` fields indicated by this mask. // The field mask must not be empty, and it must not contain fields that // are immutable or only set by the server. // Mutable top-level fields: `event_notification_config`, `http_config`, // `mqtt_config`, and `state_notification_config`. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -222,16 +222,16 @@ func (m *UpdateDeviceRegistryRequest) GetUpdateMask() *field_mask.FieldMask { type ListDeviceRegistriesRequest struct { // The project and cloud region path. For example, // `projects/example-project/locations/us-central1`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The maximum number of registries to return in the response. If this value // is zero, the service will select a default size. A call may return fewer // objects than requested, but if there is a non-empty `page_token`, it // indicates that more entries are available. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last `ListDeviceRegistriesResponse`; indicates // that this is a continuation of a prior `ListDeviceRegistries` call, and // that the system should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -285,11 +285,11 @@ func (m *ListDeviceRegistriesRequest) GetPageToken() string { // Response for `ListDeviceRegistries`. type ListDeviceRegistriesResponse struct { // The registries that matched the query. - DeviceRegistries []*DeviceRegistry `protobuf:"bytes,1,rep,name=device_registries,json=deviceRegistries" json:"device_registries,omitempty"` + DeviceRegistries []*DeviceRegistry `protobuf:"bytes,1,rep,name=device_registries,json=deviceRegistries,proto3" json:"device_registries,omitempty"` // If not empty, indicates that there may be more registries that match the // request; this value should be passed in a new // `ListDeviceRegistriesRequest`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -338,11 +338,11 @@ type CreateDeviceRequest struct { // The name of the device registry where this device should be created. // For example, // `projects/example-project/locations/us-central1/registries/my-registry`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The device registration details. The field `name` must be empty. The server // will generate that field from the device registry `id` provided and the // `parent` field. - Device *Device `protobuf:"bytes,2,opt,name=device" json:"device,omitempty"` + Device *Device `protobuf:"bytes,2,opt,name=device,proto3" json:"device,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -391,10 +391,10 @@ type GetDeviceRequest struct { // The name of the device. For example, // `projects/p0/locations/us-central1/registries/registry0/devices/device0` or // `projects/p0/locations/us-central1/registries/registry0/devices/{num_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The fields of the `Device` resource to be returned in the response. If the // field mask is unset or empty, all fields are returned. - FieldMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=field_mask,json=fieldMask" json:"field_mask,omitempty"` + FieldMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=field_mask,json=fieldMask,proto3" json:"field_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -444,12 +444,12 @@ type UpdateDeviceRequest struct { // be empty, and the field `name` must specify the name path. For example, // `projects/p0/locations/us-central1/registries/registry0/devices/device0`or // `projects/p0/locations/us-central1/registries/registry0/devices/{num_id}`. - Device *Device `protobuf:"bytes,2,opt,name=device" json:"device,omitempty"` + Device *Device `protobuf:"bytes,2,opt,name=device,proto3" json:"device,omitempty"` // Only updates the `device` fields indicated by this mask. // The field mask must not be empty, and it must not contain fields that // are immutable or only set by the server. // Mutable top-level fields: `credentials`, `blocked`, and `metadata` - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -498,7 +498,7 @@ type DeleteDeviceRequest struct { // The name of the device. For example, // `projects/p0/locations/us-central1/registries/registry0/devices/device0` or // `projects/p0/locations/us-central1/registries/registry0/devices/{num_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -539,27 +539,27 @@ func (m *DeleteDeviceRequest) GetName() string { type ListDevicesRequest struct { // The device registry path. Required. For example, // `projects/my-project/locations/us-central1/registries/my-registry`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // A list of device numerical ids. If empty, it will ignore this field. This // field cannot hold more than 10,000 entries. - DeviceNumIds []uint64 `protobuf:"varint,2,rep,packed,name=device_num_ids,json=deviceNumIds" json:"device_num_ids,omitempty"` + DeviceNumIds []uint64 `protobuf:"varint,2,rep,packed,name=device_num_ids,json=deviceNumIds,proto3" json:"device_num_ids,omitempty"` // A list of device string identifiers. If empty, it will ignore this field. // For example, `['device0', 'device12']`. This field cannot hold more than // 10,000 entries. - DeviceIds []string `protobuf:"bytes,3,rep,name=device_ids,json=deviceIds" json:"device_ids,omitempty"` + DeviceIds []string `protobuf:"bytes,3,rep,name=device_ids,json=deviceIds,proto3" json:"device_ids,omitempty"` // The fields of the `Device` resource to be returned in the response. The // fields `id`, and `num_id` are always returned by default, along with any // other fields specified. - FieldMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=field_mask,json=fieldMask" json:"field_mask,omitempty"` + FieldMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=field_mask,json=fieldMask,proto3" json:"field_mask,omitempty"` // The maximum number of devices to return in the response. If this value // is zero, the service will select a default size. A call may return fewer // objects than requested, but if there is a non-empty `page_token`, it // indicates that more entries are available. - PageSize int32 `protobuf:"varint,100,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,100,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last `ListDevicesResponse`; indicates // that this is a continuation of a prior `ListDevices` call, and // that the system should return the next page of data. - PageToken string `protobuf:"bytes,101,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,101,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -634,10 +634,10 @@ func (m *ListDevicesRequest) GetPageToken() string { // Response for `ListDevices`. type ListDevicesResponse struct { // The devices that match the request. - Devices []*Device `protobuf:"bytes,1,rep,name=devices" json:"devices,omitempty"` + Devices []*Device `protobuf:"bytes,1,rep,name=devices,proto3" json:"devices,omitempty"` // If not empty, indicates that there may be more devices that match the // request; this value should be passed in a new `ListDevicesRequest`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -686,13 +686,13 @@ type ModifyCloudToDeviceConfigRequest struct { // The name of the device. For example, // `projects/p0/locations/us-central1/registries/registry0/devices/device0` or // `projects/p0/locations/us-central1/registries/registry0/devices/{num_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The version number to update. If this value is zero, it will not check the // version number of the server and will always update the current version; // otherwise, this update will fail if the version number found on the server // does not match this version number. This is used to support multiple // simultaneous updates without losing data. - VersionToUpdate int64 `protobuf:"varint,2,opt,name=version_to_update,json=versionToUpdate" json:"version_to_update,omitempty"` + VersionToUpdate int64 `protobuf:"varint,2,opt,name=version_to_update,json=versionToUpdate,proto3" json:"version_to_update,omitempty"` // The configuration data for the device. BinaryData []byte `protobuf:"bytes,3,opt,name=binary_data,json=binaryData,proto3" json:"binary_data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -750,11 +750,11 @@ type ListDeviceConfigVersionsRequest struct { // The name of the device. For example, // `projects/p0/locations/us-central1/registries/registry0/devices/device0` or // `projects/p0/locations/us-central1/registries/registry0/devices/{num_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The number of versions to list. Versions are listed in decreasing order of // the version number. The maximum number of versions retained is 10. If this // value is zero, it will return all the versions available. - NumVersions int32 `protobuf:"varint,2,opt,name=num_versions,json=numVersions" json:"num_versions,omitempty"` + NumVersions int32 `protobuf:"varint,2,opt,name=num_versions,json=numVersions,proto3" json:"num_versions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -802,7 +802,7 @@ func (m *ListDeviceConfigVersionsRequest) GetNumVersions() int32 { type ListDeviceConfigVersionsResponse struct { // The device configuration for the last few versions. Versions are listed // in decreasing order, starting from the most recent one. - DeviceConfigs []*DeviceConfig `protobuf:"bytes,1,rep,name=device_configs,json=deviceConfigs" json:"device_configs,omitempty"` + DeviceConfigs []*DeviceConfig `protobuf:"bytes,1,rep,name=device_configs,json=deviceConfigs,proto3" json:"device_configs,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -844,11 +844,11 @@ type ListDeviceStatesRequest struct { // The name of the device. For example, // `projects/p0/locations/us-central1/registries/registry0/devices/device0` or // `projects/p0/locations/us-central1/registries/registry0/devices/{num_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The number of states to list. States are listed in descending order of // update time. The maximum number of states retained is 10. If this // value is zero, it will return all the states available. - NumStates int32 `protobuf:"varint,2,opt,name=num_states,json=numStates" json:"num_states,omitempty"` + NumStates int32 `protobuf:"varint,2,opt,name=num_states,json=numStates,proto3" json:"num_states,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -896,7 +896,7 @@ func (m *ListDeviceStatesRequest) GetNumStates() int32 { type ListDeviceStatesResponse struct { // The last few device states. States are listed in descending order of server // update time, starting from the most recent one. - DeviceStates []*DeviceState `protobuf:"bytes,1,rep,name=device_states,json=deviceStates" json:"device_states,omitempty"` + DeviceStates []*DeviceState `protobuf:"bytes,1,rep,name=device_states,json=deviceStates,proto3" json:"device_states,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1160,8 +1160,7 @@ func (c *deviceManagerClient) TestIamPermissions(ctx context.Context, in *v1.Tes return out, nil } -// Server API for DeviceManager service - +// DeviceManagerServer is the server API for DeviceManager service. type DeviceManagerServer interface { // Creates a device registry that contains devices. CreateDeviceRegistry(context.Context, *CreateDeviceRegistryRequest) (*DeviceRegistry, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/resources.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/resources.pb.go index e7ff06bfb..ee440bb85 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/resources.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/iot/v1/resources.pb.go @@ -167,67 +167,67 @@ func (PublicKeyFormat) EnumDescriptor() ([]byte, []int) { type Device struct { // The user-defined device identifier. The device ID must be unique // within a device registry. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The resource path name. For example, // `projects/p1/locations/us-central1/registries/registry0/devices/dev0` or // `projects/p1/locations/us-central1/registries/registry0/devices/{num_id}`. // When `name` is populated as a response from the service, it always ends // in the device numeric ID. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // [Output only] A server-defined unique numeric ID for the device. This is a // more compact way to identify devices, and it is globally unique. - NumId uint64 `protobuf:"varint,3,opt,name=num_id,json=numId" json:"num_id,omitempty"` + NumId uint64 `protobuf:"varint,3,opt,name=num_id,json=numId,proto3" json:"num_id,omitempty"` // The credentials used to authenticate this device. To allow credential // rotation without interruption, multiple device credentials can be bound to // this device. No more than 3 credentials can be bound to a single device at // a time. When new credentials are added to a device, they are verified // against the registry credentials. For details, see the description of the // `DeviceRegistry.credentials` field. - Credentials []*DeviceCredential `protobuf:"bytes,12,rep,name=credentials" json:"credentials,omitempty"` + Credentials []*DeviceCredential `protobuf:"bytes,12,rep,name=credentials,proto3" json:"credentials,omitempty"` // [Output only] The last time an MQTT `PINGREQ` was received. This field // applies only to devices connecting through MQTT. MQTT clients usually only // send `PINGREQ` messages if the connection is idle, and no other messages // have been sent. Timestamps are periodically collected and written to // storage; they may be stale by a few minutes. - LastHeartbeatTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=last_heartbeat_time,json=lastHeartbeatTime" json:"last_heartbeat_time,omitempty"` + LastHeartbeatTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=last_heartbeat_time,json=lastHeartbeatTime,proto3" json:"last_heartbeat_time,omitempty"` // [Output only] The last time a telemetry event was received. Timestamps are // periodically collected and written to storage; they may be stale by a few // minutes. - LastEventTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=last_event_time,json=lastEventTime" json:"last_event_time,omitempty"` + LastEventTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=last_event_time,json=lastEventTime,proto3" json:"last_event_time,omitempty"` // [Output only] The last time a state event was received. Timestamps are // periodically collected and written to storage; they may be stale by a few // minutes. - LastStateTime *timestamp.Timestamp `protobuf:"bytes,20,opt,name=last_state_time,json=lastStateTime" json:"last_state_time,omitempty"` + LastStateTime *timestamp.Timestamp `protobuf:"bytes,20,opt,name=last_state_time,json=lastStateTime,proto3" json:"last_state_time,omitempty"` // [Output only] The last time a cloud-to-device config version acknowledgment // was received from the device. This field is only for configurations // sent through MQTT. - LastConfigAckTime *timestamp.Timestamp `protobuf:"bytes,14,opt,name=last_config_ack_time,json=lastConfigAckTime" json:"last_config_ack_time,omitempty"` + LastConfigAckTime *timestamp.Timestamp `protobuf:"bytes,14,opt,name=last_config_ack_time,json=lastConfigAckTime,proto3" json:"last_config_ack_time,omitempty"` // [Output only] The last time a cloud-to-device config version was sent to // the device. - LastConfigSendTime *timestamp.Timestamp `protobuf:"bytes,18,opt,name=last_config_send_time,json=lastConfigSendTime" json:"last_config_send_time,omitempty"` + LastConfigSendTime *timestamp.Timestamp `protobuf:"bytes,18,opt,name=last_config_send_time,json=lastConfigSendTime,proto3" json:"last_config_send_time,omitempty"` // If a device is blocked, connections or requests from this device will fail. // Can be used to temporarily prevent the device from connecting if, for // example, the sensor is generating bad data and needs maintenance. - Blocked bool `protobuf:"varint,19,opt,name=blocked" json:"blocked,omitempty"` + Blocked bool `protobuf:"varint,19,opt,name=blocked,proto3" json:"blocked,omitempty"` // [Output only] The time the most recent error occurred, such as a failure to // publish to Cloud Pub/Sub. This field is the timestamp of // 'last_error_status'. - LastErrorTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=last_error_time,json=lastErrorTime" json:"last_error_time,omitempty"` + LastErrorTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=last_error_time,json=lastErrorTime,proto3" json:"last_error_time,omitempty"` // [Output only] The error message of the most recent error, such as a failure // to publish to Cloud Pub/Sub. 'last_error_time' is the timestamp of this // field. If no errors have occurred, this field has an empty message // and the status code 0 == OK. Otherwise, this field is expected to have a // status code other than OK. - LastErrorStatus *status.Status `protobuf:"bytes,11,opt,name=last_error_status,json=lastErrorStatus" json:"last_error_status,omitempty"` + LastErrorStatus *status.Status `protobuf:"bytes,11,opt,name=last_error_status,json=lastErrorStatus,proto3" json:"last_error_status,omitempty"` // The most recent device configuration, which is eventually sent from // Cloud IoT Core to the device. If not present on creation, the // configuration will be initialized with an empty payload and version value // of `1`. To update this field after creation, use the // `DeviceManager.ModifyCloudToDeviceConfig` method. - Config *DeviceConfig `protobuf:"bytes,13,opt,name=config" json:"config,omitempty"` + Config *DeviceConfig `protobuf:"bytes,13,opt,name=config,proto3" json:"config,omitempty"` // [Output only] The state most recently received from the device. If no state // has been reported, this field is not present. - State *DeviceState `protobuf:"bytes,16,opt,name=state" json:"state,omitempty"` + State *DeviceState `protobuf:"bytes,16,opt,name=state,proto3" json:"state,omitempty"` // The metadata key-value pairs assigned to the device. This metadata is not // interpreted or indexed by Cloud IoT Core. It can be used to add contextual // information for the device. @@ -240,7 +240,7 @@ type Device struct { // // The total size of all keys and values must be less than 256 KB, and the // maximum number of key-value pairs is 500. - Metadata map[string]string `protobuf:"bytes,17,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,17,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -378,10 +378,10 @@ func (m *Device) GetMetadata() map[string]string { // A container for a group of devices. type DeviceRegistry struct { // The identifier of this device registry. For example, `myRegistry`. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The resource path name. For example, // `projects/example-project/locations/us-central1/registries/my-registry`. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // The configuration for notification of telemetry events received from the // device. All telemetry events that were successfully published by the // device and acknowledged by Cloud IoT Core are guaranteed to be @@ -391,18 +391,18 @@ type DeviceRegistry struct { // for the device's registry, the connection closes automatically. If you try // to do so using an HTTP connection, an error is returned. Up to 10 // configurations may be provided. - EventNotificationConfigs []*EventNotificationConfig `protobuf:"bytes,10,rep,name=event_notification_configs,json=eventNotificationConfigs" json:"event_notification_configs,omitempty"` + EventNotificationConfigs []*EventNotificationConfig `protobuf:"bytes,10,rep,name=event_notification_configs,json=eventNotificationConfigs,proto3" json:"event_notification_configs,omitempty"` // The configuration for notification of new states received from the device. // State updates are guaranteed to be stored in the state history, but // notifications to Cloud Pub/Sub are not guaranteed. For example, if // permissions are misconfigured or the specified topic doesn't exist, no // notification will be published but the state will still be stored in Cloud // IoT Core. - StateNotificationConfig *StateNotificationConfig `protobuf:"bytes,7,opt,name=state_notification_config,json=stateNotificationConfig" json:"state_notification_config,omitempty"` + StateNotificationConfig *StateNotificationConfig `protobuf:"bytes,7,opt,name=state_notification_config,json=stateNotificationConfig,proto3" json:"state_notification_config,omitempty"` // The MQTT configuration for this device registry. - MqttConfig *MqttConfig `protobuf:"bytes,4,opt,name=mqtt_config,json=mqttConfig" json:"mqtt_config,omitempty"` + MqttConfig *MqttConfig `protobuf:"bytes,4,opt,name=mqtt_config,json=mqttConfig,proto3" json:"mqtt_config,omitempty"` // The DeviceService (HTTP) configuration for this device registry. - HttpConfig *HttpConfig `protobuf:"bytes,9,opt,name=http_config,json=httpConfig" json:"http_config,omitempty"` + HttpConfig *HttpConfig `protobuf:"bytes,9,opt,name=http_config,json=httpConfig,proto3" json:"http_config,omitempty"` // The credentials used to verify the device credentials. No more than 10 // credentials can be bound to a single registry at a time. The verification // process occurs at the time of device creation or update. If this field is @@ -414,7 +414,7 @@ type DeviceRegistry struct { // modifications to this list of credentials: after a device has been // successfully created in a registry, it should be able to connect even if // its registry credentials are revoked, deleted, or modified. - Credentials []*RegistryCredential `protobuf:"bytes,8,rep,name=credentials" json:"credentials,omitempty"` + Credentials []*RegistryCredential `protobuf:"bytes,8,rep,name=credentials,proto3" json:"credentials,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -497,7 +497,7 @@ func (m *DeviceRegistry) GetCredentials() []*RegistryCredential { type MqttConfig struct { // If enabled, allows connections using the MQTT protocol. Otherwise, MQTT // connections to this registry will fail. - MqttEnabledState MqttState `protobuf:"varint,1,opt,name=mqtt_enabled_state,json=mqttEnabledState,enum=google.cloud.iot.v1.MqttState" json:"mqtt_enabled_state,omitempty"` + MqttEnabledState MqttState `protobuf:"varint,1,opt,name=mqtt_enabled_state,json=mqttEnabledState,proto3,enum=google.cloud.iot.v1.MqttState" json:"mqtt_enabled_state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -538,7 +538,7 @@ func (m *MqttConfig) GetMqttEnabledState() MqttState { type HttpConfig struct { // If enabled, allows devices to use DeviceService via the HTTP protocol. // Otherwise, any requests to DeviceService will fail for this registry. - HttpEnabledState HttpState `protobuf:"varint,1,opt,name=http_enabled_state,json=httpEnabledState,enum=google.cloud.iot.v1.HttpState" json:"http_enabled_state,omitempty"` + HttpEnabledState HttpState `protobuf:"varint,1,opt,name=http_enabled_state,json=httpEnabledState,proto3,enum=google.cloud.iot.v1.HttpState" json:"http_enabled_state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -581,10 +581,10 @@ type EventNotificationConfig struct { // be used. The string must not include the leading '/' character. If empty, // all strings are matched. This field is used only for telemetry events; // subfolders are not supported for state changes. - SubfolderMatches string `protobuf:"bytes,2,opt,name=subfolder_matches,json=subfolderMatches" json:"subfolder_matches,omitempty"` + SubfolderMatches string `protobuf:"bytes,2,opt,name=subfolder_matches,json=subfolderMatches,proto3" json:"subfolder_matches,omitempty"` // A Cloud Pub/Sub topic name. For example, // `projects/myProject/topics/deviceEvents`. - PubsubTopicName string `protobuf:"bytes,1,opt,name=pubsub_topic_name,json=pubsubTopicName" json:"pubsub_topic_name,omitempty"` + PubsubTopicName string `protobuf:"bytes,1,opt,name=pubsub_topic_name,json=pubsubTopicName,proto3" json:"pubsub_topic_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -632,7 +632,7 @@ func (m *EventNotificationConfig) GetPubsubTopicName() string { type StateNotificationConfig struct { // A Cloud Pub/Sub topic name. For example, // `projects/myProject/topics/deviceEvents`. - PubsubTopicName string `protobuf:"bytes,1,opt,name=pubsub_topic_name,json=pubsubTopicName" json:"pubsub_topic_name,omitempty"` + PubsubTopicName string `protobuf:"bytes,1,opt,name=pubsub_topic_name,json=pubsubTopicName,proto3" json:"pubsub_topic_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -710,7 +710,7 @@ type isRegistryCredential_Credential interface { } type RegistryCredential_PublicKeyCertificate struct { - PublicKeyCertificate *PublicKeyCertificate `protobuf:"bytes,1,opt,name=public_key_certificate,json=publicKeyCertificate,oneof"` + PublicKeyCertificate *PublicKeyCertificate `protobuf:"bytes,1,opt,name=public_key_certificate,json=publicKeyCertificate,proto3,oneof"` } func (*RegistryCredential_PublicKeyCertificate) isRegistryCredential_Credential() {} @@ -787,17 +787,17 @@ func _RegistryCredential_OneofSizer(msg proto.Message) (n int) { // Details of an X.509 certificate. For informational purposes only. type X509CertificateDetails struct { // The entity that signed the certificate. - Issuer string `protobuf:"bytes,1,opt,name=issuer" json:"issuer,omitempty"` + Issuer string `protobuf:"bytes,1,opt,name=issuer,proto3" json:"issuer,omitempty"` // The entity the certificate and public key belong to. - Subject string `protobuf:"bytes,2,opt,name=subject" json:"subject,omitempty"` + Subject string `protobuf:"bytes,2,opt,name=subject,proto3" json:"subject,omitempty"` // The time the certificate becomes valid. - StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time the certificate becomes invalid. - ExpiryTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=expiry_time,json=expiryTime" json:"expiry_time,omitempty"` + ExpiryTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=expiry_time,json=expiryTime,proto3" json:"expiry_time,omitempty"` // The algorithm used to sign the certificate. - SignatureAlgorithm string `protobuf:"bytes,5,opt,name=signature_algorithm,json=signatureAlgorithm" json:"signature_algorithm,omitempty"` + SignatureAlgorithm string `protobuf:"bytes,5,opt,name=signature_algorithm,json=signatureAlgorithm,proto3" json:"signature_algorithm,omitempty"` // The type of public key in the certificate. - PublicKeyType string `protobuf:"bytes,6,opt,name=public_key_type,json=publicKeyType" json:"public_key_type,omitempty"` + PublicKeyType string `protobuf:"bytes,6,opt,name=public_key_type,json=publicKeyType,proto3" json:"public_key_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -872,11 +872,11 @@ func (m *X509CertificateDetails) GetPublicKeyType() string { // A public key certificate format and data. type PublicKeyCertificate struct { // The certificate format. - Format PublicKeyCertificateFormat `protobuf:"varint,1,opt,name=format,enum=google.cloud.iot.v1.PublicKeyCertificateFormat" json:"format,omitempty"` + Format PublicKeyCertificateFormat `protobuf:"varint,1,opt,name=format,proto3,enum=google.cloud.iot.v1.PublicKeyCertificateFormat" json:"format,omitempty"` // The certificate data. - Certificate string `protobuf:"bytes,2,opt,name=certificate" json:"certificate,omitempty"` + Certificate string `protobuf:"bytes,2,opt,name=certificate,proto3" json:"certificate,omitempty"` // [Output only] The certificate details. Used only for X.509 certificates. - X509Details *X509CertificateDetails `protobuf:"bytes,3,opt,name=x509_details,json=x509Details" json:"x509_details,omitempty"` + X509Details *X509CertificateDetails `protobuf:"bytes,3,opt,name=x509_details,json=x509Details,proto3" json:"x509_details,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -937,7 +937,7 @@ type DeviceCredential struct { // [Optional] The time at which this credential becomes invalid. This // credential will be ignored for new client authentication requests after // this timestamp; however, it will not be automatically deleted. - ExpirationTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=expiration_time,json=expirationTime" json:"expiration_time,omitempty"` + ExpirationTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=expiration_time,json=expirationTime,proto3" json:"expiration_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -972,7 +972,7 @@ type isDeviceCredential_Credential interface { } type DeviceCredential_PublicKey struct { - PublicKey *PublicKeyCredential `protobuf:"bytes,2,opt,name=public_key,json=publicKey,oneof"` + PublicKey *PublicKeyCredential `protobuf:"bytes,2,opt,name=public_key,json=publicKey,proto3,oneof"` } func (*DeviceCredential_PublicKey) isDeviceCredential_Credential() {} @@ -1056,9 +1056,9 @@ func _DeviceCredential_OneofSizer(msg proto.Message) (n int) { // A public key format and data. type PublicKeyCredential struct { // The format of the key. - Format PublicKeyFormat `protobuf:"varint,1,opt,name=format,enum=google.cloud.iot.v1.PublicKeyFormat" json:"format,omitempty"` + Format PublicKeyFormat `protobuf:"varint,1,opt,name=format,proto3,enum=google.cloud.iot.v1.PublicKeyFormat" json:"format,omitempty"` // The key data. - Key string `protobuf:"bytes,2,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,2,opt,name=key,proto3" json:"key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1108,10 +1108,10 @@ type DeviceConfig struct { // the server, and is always greater than 0 after device creation. The // version must be 0 on the `CreateDevice` request if a `config` is // specified; the response of `CreateDevice` will always have a value of 1. - Version int64 `protobuf:"varint,1,opt,name=version" json:"version,omitempty"` + Version int64 `protobuf:"varint,1,opt,name=version,proto3" json:"version,omitempty"` // [Output only] The time at which this configuration version was updated in // Cloud IoT Core. This timestamp is set by the server. - CloudUpdateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=cloud_update_time,json=cloudUpdateTime" json:"cloud_update_time,omitempty"` + CloudUpdateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=cloud_update_time,json=cloudUpdateTime,proto3" json:"cloud_update_time,omitempty"` // [Output only] The time at which Cloud IoT Core received the // acknowledgment from the device, indicating that the device has received // this configuration version. If this field is not present, the device has @@ -1121,7 +1121,7 @@ type DeviceConfig struct { // connection, only the latest version is sent to the device. Some // versions may never be sent to the device, and therefore are never // acknowledged. This timestamp is set by Cloud IoT Core. - DeviceAckTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=device_ack_time,json=deviceAckTime" json:"device_ack_time,omitempty"` + DeviceAckTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=device_ack_time,json=deviceAckTime,proto3" json:"device_ack_time,omitempty"` // The device configuration data. BinaryData []byte `protobuf:"bytes,4,opt,name=binary_data,json=binaryData,proto3" json:"binary_data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1185,7 +1185,7 @@ func (m *DeviceConfig) GetBinaryData() []byte { type DeviceState struct { // [Output only] The time at which this state version was updated in Cloud // IoT Core. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // The device state data. BinaryData []byte `protobuf:"bytes,2,opt,name=binary_data,json=binaryData,proto3" json:"binary_data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/kms/v1/resources.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/kms/v1/resources.pb.go new file mode 100644 index 000000000..bc3483a1a --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/kms/v1/resources.pb.go @@ -0,0 +1,486 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/kms/v1/resources.proto + +package kms // import "google.golang.org/genproto/googleapis/cloud/kms/v1" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// [CryptoKeyPurpose][google.cloud.kms.v1.CryptoKey.CryptoKeyPurpose] describes the capabilities of a [CryptoKey][google.cloud.kms.v1.CryptoKey]. Two +// keys with the same purpose may use different underlying algorithms, but +// must support the same set of operations. +type CryptoKey_CryptoKeyPurpose int32 + +const ( + // Not specified. + CryptoKey_CRYPTO_KEY_PURPOSE_UNSPECIFIED CryptoKey_CryptoKeyPurpose = 0 + // [CryptoKeys][google.cloud.kms.v1.CryptoKey] with this purpose may be used with + // [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt] and + // [Decrypt][google.cloud.kms.v1.KeyManagementService.Decrypt]. + CryptoKey_ENCRYPT_DECRYPT CryptoKey_CryptoKeyPurpose = 1 +) + +var CryptoKey_CryptoKeyPurpose_name = map[int32]string{ + 0: "CRYPTO_KEY_PURPOSE_UNSPECIFIED", + 1: "ENCRYPT_DECRYPT", +} +var CryptoKey_CryptoKeyPurpose_value = map[string]int32{ + "CRYPTO_KEY_PURPOSE_UNSPECIFIED": 0, + "ENCRYPT_DECRYPT": 1, +} + +func (x CryptoKey_CryptoKeyPurpose) String() string { + return proto.EnumName(CryptoKey_CryptoKeyPurpose_name, int32(x)) +} +func (CryptoKey_CryptoKeyPurpose) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_resources_3fd296b2ee28c3bd, []int{1, 0} +} + +// The state of a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion], indicating if it can be used. +type CryptoKeyVersion_CryptoKeyVersionState int32 + +const ( + // Not specified. + CryptoKeyVersion_CRYPTO_KEY_VERSION_STATE_UNSPECIFIED CryptoKeyVersion_CryptoKeyVersionState = 0 + // This version may be used in [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt] and + // [Decrypt][google.cloud.kms.v1.KeyManagementService.Decrypt] requests. + CryptoKeyVersion_ENABLED CryptoKeyVersion_CryptoKeyVersionState = 1 + // This version may not be used, but the key material is still available, + // and the version can be placed back into the [ENABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.ENABLED] state. + CryptoKeyVersion_DISABLED CryptoKeyVersion_CryptoKeyVersionState = 2 + // This version is destroyed, and the key material is no longer stored. + // A version may not leave this state once entered. + CryptoKeyVersion_DESTROYED CryptoKeyVersion_CryptoKeyVersionState = 3 + // This version is scheduled for destruction, and will be destroyed soon. + // Call + // [RestoreCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.RestoreCryptoKeyVersion] + // to put it back into the [DISABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DISABLED] state. + CryptoKeyVersion_DESTROY_SCHEDULED CryptoKeyVersion_CryptoKeyVersionState = 4 +) + +var CryptoKeyVersion_CryptoKeyVersionState_name = map[int32]string{ + 0: "CRYPTO_KEY_VERSION_STATE_UNSPECIFIED", + 1: "ENABLED", + 2: "DISABLED", + 3: "DESTROYED", + 4: "DESTROY_SCHEDULED", +} +var CryptoKeyVersion_CryptoKeyVersionState_value = map[string]int32{ + "CRYPTO_KEY_VERSION_STATE_UNSPECIFIED": 0, + "ENABLED": 1, + "DISABLED": 2, + "DESTROYED": 3, + "DESTROY_SCHEDULED": 4, +} + +func (x CryptoKeyVersion_CryptoKeyVersionState) String() string { + return proto.EnumName(CryptoKeyVersion_CryptoKeyVersionState_name, int32(x)) +} +func (CryptoKeyVersion_CryptoKeyVersionState) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_resources_3fd296b2ee28c3bd, []int{2, 0} +} + +// A [KeyRing][google.cloud.kms.v1.KeyRing] is a toplevel logical grouping of [CryptoKeys][google.cloud.kms.v1.CryptoKey]. +type KeyRing struct { + // Output only. The resource name for the [KeyRing][google.cloud.kms.v1.KeyRing] in the format + // `projects/*/locations/*/keyRings/*`. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Output only. The time at which this [KeyRing][google.cloud.kms.v1.KeyRing] was created. + CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *KeyRing) Reset() { *m = KeyRing{} } +func (m *KeyRing) String() string { return proto.CompactTextString(m) } +func (*KeyRing) ProtoMessage() {} +func (*KeyRing) Descriptor() ([]byte, []int) { + return fileDescriptor_resources_3fd296b2ee28c3bd, []int{0} +} +func (m *KeyRing) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_KeyRing.Unmarshal(m, b) +} +func (m *KeyRing) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_KeyRing.Marshal(b, m, deterministic) +} +func (dst *KeyRing) XXX_Merge(src proto.Message) { + xxx_messageInfo_KeyRing.Merge(dst, src) +} +func (m *KeyRing) XXX_Size() int { + return xxx_messageInfo_KeyRing.Size(m) +} +func (m *KeyRing) XXX_DiscardUnknown() { + xxx_messageInfo_KeyRing.DiscardUnknown(m) +} + +var xxx_messageInfo_KeyRing proto.InternalMessageInfo + +func (m *KeyRing) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *KeyRing) GetCreateTime() *timestamp.Timestamp { + if m != nil { + return m.CreateTime + } + return nil +} + +// A [CryptoKey][google.cloud.kms.v1.CryptoKey] represents a logical key that can be used for cryptographic +// operations. +// +// A [CryptoKey][google.cloud.kms.v1.CryptoKey] is made up of one or more [versions][google.cloud.kms.v1.CryptoKeyVersion], which +// represent the actual key material used in cryptographic operations. +type CryptoKey struct { + // Output only. The resource name for this [CryptoKey][google.cloud.kms.v1.CryptoKey] in the format + // `projects/*/locations/*/keyRings/*/cryptoKeys/*`. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Output only. A copy of the "primary" [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] that will be used + // by [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt] when this [CryptoKey][google.cloud.kms.v1.CryptoKey] is given + // in [EncryptRequest.name][google.cloud.kms.v1.EncryptRequest.name]. + // + // The [CryptoKey][google.cloud.kms.v1.CryptoKey]'s primary version can be updated via + // [UpdateCryptoKeyPrimaryVersion][google.cloud.kms.v1.KeyManagementService.UpdateCryptoKeyPrimaryVersion]. + Primary *CryptoKeyVersion `protobuf:"bytes,2,opt,name=primary,proto3" json:"primary,omitempty"` + // The immutable purpose of this [CryptoKey][google.cloud.kms.v1.CryptoKey]. Currently, the only acceptable + // purpose is [ENCRYPT_DECRYPT][google.cloud.kms.v1.CryptoKey.CryptoKeyPurpose.ENCRYPT_DECRYPT]. + Purpose CryptoKey_CryptoKeyPurpose `protobuf:"varint,3,opt,name=purpose,proto3,enum=google.cloud.kms.v1.CryptoKey_CryptoKeyPurpose" json:"purpose,omitempty"` + // Output only. The time at which this [CryptoKey][google.cloud.kms.v1.CryptoKey] was created. + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` + // At [next_rotation_time][google.cloud.kms.v1.CryptoKey.next_rotation_time], the Key Management Service will automatically: + // + // 1. Create a new version of this [CryptoKey][google.cloud.kms.v1.CryptoKey]. + // 2. Mark the new version as primary. + // + // Key rotations performed manually via + // [CreateCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.CreateCryptoKeyVersion] and + // [UpdateCryptoKeyPrimaryVersion][google.cloud.kms.v1.KeyManagementService.UpdateCryptoKeyPrimaryVersion] + // do not affect [next_rotation_time][google.cloud.kms.v1.CryptoKey.next_rotation_time]. + NextRotationTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=next_rotation_time,json=nextRotationTime,proto3" json:"next_rotation_time,omitempty"` + // Controls the rate of automatic rotation. + // + // Types that are valid to be assigned to RotationSchedule: + // *CryptoKey_RotationPeriod + RotationSchedule isCryptoKey_RotationSchedule `protobuf_oneof:"rotation_schedule"` + // Labels with user-defined metadata. For more information, see + // [Labeling Keys](/kms/docs/labeling-keys). + Labels map[string]string `protobuf:"bytes,10,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CryptoKey) Reset() { *m = CryptoKey{} } +func (m *CryptoKey) String() string { return proto.CompactTextString(m) } +func (*CryptoKey) ProtoMessage() {} +func (*CryptoKey) Descriptor() ([]byte, []int) { + return fileDescriptor_resources_3fd296b2ee28c3bd, []int{1} +} +func (m *CryptoKey) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CryptoKey.Unmarshal(m, b) +} +func (m *CryptoKey) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CryptoKey.Marshal(b, m, deterministic) +} +func (dst *CryptoKey) XXX_Merge(src proto.Message) { + xxx_messageInfo_CryptoKey.Merge(dst, src) +} +func (m *CryptoKey) XXX_Size() int { + return xxx_messageInfo_CryptoKey.Size(m) +} +func (m *CryptoKey) XXX_DiscardUnknown() { + xxx_messageInfo_CryptoKey.DiscardUnknown(m) +} + +var xxx_messageInfo_CryptoKey proto.InternalMessageInfo + +type isCryptoKey_RotationSchedule interface { + isCryptoKey_RotationSchedule() +} + +type CryptoKey_RotationPeriod struct { + RotationPeriod *duration.Duration `protobuf:"bytes,8,opt,name=rotation_period,json=rotationPeriod,proto3,oneof"` +} + +func (*CryptoKey_RotationPeriod) isCryptoKey_RotationSchedule() {} + +func (m *CryptoKey) GetRotationSchedule() isCryptoKey_RotationSchedule { + if m != nil { + return m.RotationSchedule + } + return nil +} + +func (m *CryptoKey) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *CryptoKey) GetPrimary() *CryptoKeyVersion { + if m != nil { + return m.Primary + } + return nil +} + +func (m *CryptoKey) GetPurpose() CryptoKey_CryptoKeyPurpose { + if m != nil { + return m.Purpose + } + return CryptoKey_CRYPTO_KEY_PURPOSE_UNSPECIFIED +} + +func (m *CryptoKey) GetCreateTime() *timestamp.Timestamp { + if m != nil { + return m.CreateTime + } + return nil +} + +func (m *CryptoKey) GetNextRotationTime() *timestamp.Timestamp { + if m != nil { + return m.NextRotationTime + } + return nil +} + +func (m *CryptoKey) GetRotationPeriod() *duration.Duration { + if x, ok := m.GetRotationSchedule().(*CryptoKey_RotationPeriod); ok { + return x.RotationPeriod + } + return nil +} + +func (m *CryptoKey) GetLabels() map[string]string { + if m != nil { + return m.Labels + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*CryptoKey) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _CryptoKey_OneofMarshaler, _CryptoKey_OneofUnmarshaler, _CryptoKey_OneofSizer, []interface{}{ + (*CryptoKey_RotationPeriod)(nil), + } +} + +func _CryptoKey_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*CryptoKey) + // rotation_schedule + switch x := m.RotationSchedule.(type) { + case *CryptoKey_RotationPeriod: + b.EncodeVarint(8<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.RotationPeriod); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("CryptoKey.RotationSchedule has unexpected type %T", x) + } + return nil +} + +func _CryptoKey_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*CryptoKey) + switch tag { + case 8: // rotation_schedule.rotation_period + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(duration.Duration) + err := b.DecodeMessage(msg) + m.RotationSchedule = &CryptoKey_RotationPeriod{msg} + return true, err + default: + return false, nil + } +} + +func _CryptoKey_OneofSizer(msg proto.Message) (n int) { + m := msg.(*CryptoKey) + // rotation_schedule + switch x := m.RotationSchedule.(type) { + case *CryptoKey_RotationPeriod: + s := proto.Size(x.RotationPeriod) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// A [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] represents an individual cryptographic key, and the +// associated key material. +// +// It can be used for cryptographic operations either directly, or via its +// parent [CryptoKey][google.cloud.kms.v1.CryptoKey], in which case the server will choose the appropriate +// version for the operation. +// +// For security reasons, the raw cryptographic key material represented by a +// [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] can never be viewed or exported. It can only be used to +// encrypt or decrypt data when an authorized user or application invokes Cloud +// KMS. +type CryptoKeyVersion struct { + // Output only. The resource name for this [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] in the format + // `projects/*/locations/*/keyRings/*/cryptoKeys/*/cryptoKeyVersions/*`. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The current state of the [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]. + State CryptoKeyVersion_CryptoKeyVersionState `protobuf:"varint,3,opt,name=state,proto3,enum=google.cloud.kms.v1.CryptoKeyVersion_CryptoKeyVersionState" json:"state,omitempty"` + // Output only. The time at which this [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] was created. + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` + // Output only. The time this [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]'s key material is scheduled + // for destruction. Only present if [state][google.cloud.kms.v1.CryptoKeyVersion.state] is + // [DESTROY_SCHEDULED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROY_SCHEDULED]. + DestroyTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=destroy_time,json=destroyTime,proto3" json:"destroy_time,omitempty"` + // Output only. The time this CryptoKeyVersion's key material was + // destroyed. Only present if [state][google.cloud.kms.v1.CryptoKeyVersion.state] is + // [DESTROYED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROYED]. + DestroyEventTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=destroy_event_time,json=destroyEventTime,proto3" json:"destroy_event_time,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CryptoKeyVersion) Reset() { *m = CryptoKeyVersion{} } +func (m *CryptoKeyVersion) String() string { return proto.CompactTextString(m) } +func (*CryptoKeyVersion) ProtoMessage() {} +func (*CryptoKeyVersion) Descriptor() ([]byte, []int) { + return fileDescriptor_resources_3fd296b2ee28c3bd, []int{2} +} +func (m *CryptoKeyVersion) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CryptoKeyVersion.Unmarshal(m, b) +} +func (m *CryptoKeyVersion) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CryptoKeyVersion.Marshal(b, m, deterministic) +} +func (dst *CryptoKeyVersion) XXX_Merge(src proto.Message) { + xxx_messageInfo_CryptoKeyVersion.Merge(dst, src) +} +func (m *CryptoKeyVersion) XXX_Size() int { + return xxx_messageInfo_CryptoKeyVersion.Size(m) +} +func (m *CryptoKeyVersion) XXX_DiscardUnknown() { + xxx_messageInfo_CryptoKeyVersion.DiscardUnknown(m) +} + +var xxx_messageInfo_CryptoKeyVersion proto.InternalMessageInfo + +func (m *CryptoKeyVersion) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *CryptoKeyVersion) GetState() CryptoKeyVersion_CryptoKeyVersionState { + if m != nil { + return m.State + } + return CryptoKeyVersion_CRYPTO_KEY_VERSION_STATE_UNSPECIFIED +} + +func (m *CryptoKeyVersion) GetCreateTime() *timestamp.Timestamp { + if m != nil { + return m.CreateTime + } + return nil +} + +func (m *CryptoKeyVersion) GetDestroyTime() *timestamp.Timestamp { + if m != nil { + return m.DestroyTime + } + return nil +} + +func (m *CryptoKeyVersion) GetDestroyEventTime() *timestamp.Timestamp { + if m != nil { + return m.DestroyEventTime + } + return nil +} + +func init() { + proto.RegisterType((*KeyRing)(nil), "google.cloud.kms.v1.KeyRing") + proto.RegisterType((*CryptoKey)(nil), "google.cloud.kms.v1.CryptoKey") + proto.RegisterMapType((map[string]string)(nil), "google.cloud.kms.v1.CryptoKey.LabelsEntry") + proto.RegisterType((*CryptoKeyVersion)(nil), "google.cloud.kms.v1.CryptoKeyVersion") + proto.RegisterEnum("google.cloud.kms.v1.CryptoKey_CryptoKeyPurpose", CryptoKey_CryptoKeyPurpose_name, CryptoKey_CryptoKeyPurpose_value) + proto.RegisterEnum("google.cloud.kms.v1.CryptoKeyVersion_CryptoKeyVersionState", CryptoKeyVersion_CryptoKeyVersionState_name, CryptoKeyVersion_CryptoKeyVersionState_value) +} + +func init() { + proto.RegisterFile("google/cloud/kms/v1/resources.proto", fileDescriptor_resources_3fd296b2ee28c3bd) +} + +var fileDescriptor_resources_3fd296b2ee28c3bd = []byte{ + // 673 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x94, 0x51, 0x6e, 0xda, 0x4a, + 0x14, 0x86, 0x63, 0x20, 0x21, 0x1c, 0x72, 0x13, 0x32, 0xdc, 0xe8, 0x72, 0x51, 0x95, 0x22, 0xda, + 0x4a, 0xa8, 0x0f, 0xb6, 0x48, 0xa5, 0xaa, 0x6d, 0x54, 0x55, 0x01, 0xbb, 0x0d, 0x22, 0x02, 0x77, + 0x4c, 0x90, 0x12, 0x45, 0xb2, 0x1c, 0x98, 0x52, 0x0b, 0xec, 0xb1, 0x66, 0x0c, 0xaa, 0x5f, 0xbb, + 0x87, 0x6e, 0xa2, 0xeb, 0xe8, 0x53, 0x77, 0xd0, 0x9d, 0xf4, 0xb1, 0xf2, 0x78, 0x9c, 0xa6, 0x04, + 0x35, 0xc9, 0x13, 0x33, 0x67, 0xfe, 0xff, 0x33, 0x67, 0xce, 0x6f, 0xc3, 0xa3, 0x09, 0xa5, 0x93, + 0x19, 0xd1, 0x46, 0x33, 0x3a, 0x1f, 0x6b, 0x53, 0x8f, 0x6b, 0x8b, 0xa6, 0xc6, 0x08, 0xa7, 0x73, + 0x36, 0x22, 0x5c, 0x0d, 0x18, 0x0d, 0x29, 0x2a, 0x27, 0x22, 0x55, 0x88, 0xd4, 0xa9, 0xc7, 0xd5, + 0x45, 0xb3, 0xfa, 0x40, 0x3a, 0x9d, 0xc0, 0xd5, 0x1c, 0xdf, 0xa7, 0xa1, 0x13, 0xba, 0xd4, 0x97, + 0x96, 0xea, 0xbe, 0x3c, 0x15, 0xbb, 0xcb, 0xf9, 0x07, 0x6d, 0x3c, 0x67, 0x42, 0x20, 0xcf, 0x1f, + 0x2e, 0x9f, 0x87, 0xae, 0x47, 0x78, 0xe8, 0x78, 0x41, 0x22, 0xa8, 0x9f, 0x43, 0xbe, 0x4b, 0x22, + 0xec, 0xfa, 0x13, 0x84, 0x20, 0xe7, 0x3b, 0x1e, 0xa9, 0x28, 0x35, 0xa5, 0x51, 0xc0, 0x62, 0x8d, + 0x0e, 0xa1, 0x38, 0x62, 0xc4, 0x09, 0x89, 0x1d, 0x1b, 0x2b, 0x99, 0x9a, 0xd2, 0x28, 0x1e, 0x54, + 0x55, 0xf9, 0x47, 0x53, 0xaa, 0x3a, 0x48, 0xa9, 0x18, 0x12, 0x79, 0x5c, 0xa8, 0xff, 0xc8, 0x41, + 0xa1, 0xcd, 0xa2, 0x20, 0xa4, 0x5d, 0x12, 0xad, 0xc4, 0xbf, 0x81, 0x7c, 0xc0, 0x5c, 0xcf, 0x61, + 0x91, 0x44, 0x3f, 0x51, 0x57, 0xdc, 0x81, 0x7a, 0x05, 0x19, 0x12, 0xc6, 0x5d, 0xea, 0xe3, 0xd4, + 0x85, 0x3a, 0x90, 0x0f, 0xe6, 0x2c, 0xa0, 0x9c, 0x54, 0xb2, 0x35, 0xa5, 0xb1, 0x7d, 0xa0, 0xfd, + 0x1d, 0xf0, 0x7b, 0x65, 0x26, 0x36, 0x9c, 0xfa, 0x97, 0x5b, 0x5d, 0xbf, 0x4f, 0xab, 0xe8, 0x18, + 0x90, 0x4f, 0x3e, 0x85, 0x36, 0x93, 0xf3, 0x49, 0x18, 0xf9, 0x5b, 0x19, 0xa5, 0xd8, 0x85, 0xa5, + 0x49, 0x90, 0x74, 0xd8, 0xb9, 0x82, 0x04, 0x84, 0xb9, 0x74, 0x5c, 0xd9, 0x14, 0x98, 0xff, 0x6f, + 0x60, 0x74, 0x39, 0xeb, 0xe3, 0x35, 0xbc, 0x9d, 0x7a, 0x4c, 0x61, 0x41, 0x2d, 0xd8, 0x98, 0x39, + 0x97, 0x64, 0xc6, 0x2b, 0x50, 0xcb, 0x36, 0x8a, 0x07, 0x4f, 0x6f, 0xb9, 0x96, 0x13, 0x21, 0x36, + 0xfc, 0x90, 0x45, 0x58, 0x3a, 0xab, 0x2f, 0xa1, 0x78, 0xad, 0x8c, 0x4a, 0x90, 0x9d, 0x92, 0x48, + 0x8e, 0x2f, 0x5e, 0xa2, 0x7f, 0x61, 0x7d, 0xe1, 0xcc, 0xe6, 0x49, 0x2c, 0x0a, 0x38, 0xd9, 0xbc, + 0xca, 0xbc, 0x50, 0xea, 0x5d, 0x28, 0x2d, 0x5f, 0x34, 0xaa, 0xc3, 0x7e, 0x1b, 0x9f, 0x99, 0x83, + 0xbe, 0xdd, 0x35, 0xce, 0x6c, 0xf3, 0x14, 0x9b, 0x7d, 0xcb, 0xb0, 0x4f, 0x7b, 0x96, 0x69, 0xb4, + 0x3b, 0x6f, 0x3b, 0x86, 0x5e, 0x5a, 0x43, 0x65, 0xd8, 0x31, 0x7a, 0x42, 0x65, 0xeb, 0x86, 0xf8, + 0x2d, 0x29, 0xad, 0x32, 0xec, 0x5e, 0xdd, 0x08, 0x1f, 0x7d, 0x24, 0xe3, 0xf9, 0x8c, 0xd4, 0xbf, + 0x65, 0xaf, 0x3d, 0x42, 0xc6, 0x62, 0x65, 0xc4, 0xde, 0xc3, 0x3a, 0x0f, 0x9d, 0x30, 0xcd, 0xc7, + 0xe1, 0x9d, 0x02, 0x76, 0xa3, 0x60, 0xc5, 0x08, 0x9c, 0x90, 0x96, 0x93, 0x92, 0xbb, 0x57, 0x52, + 0x5e, 0xc3, 0xd6, 0x98, 0xf0, 0x90, 0xd1, 0xe8, 0xae, 0x39, 0x2b, 0x4a, 0x7d, 0x1a, 0xb4, 0xd4, + 0x4e, 0x16, 0xc4, 0x0f, 0x13, 0xc8, 0xc6, 0xed, 0x41, 0x93, 0x2e, 0x23, 0x36, 0x89, 0xb7, 0xf3, + 0xb3, 0x02, 0x7b, 0x2b, 0xdb, 0x44, 0x0d, 0x78, 0x7c, 0x6d, 0x52, 0x43, 0x03, 0x5b, 0x9d, 0x7e, + 0xcf, 0xb6, 0x06, 0x47, 0x83, 0xe5, 0x79, 0x15, 0x21, 0x6f, 0xf4, 0x8e, 0x5a, 0x27, 0x86, 0x5e, + 0x52, 0xd0, 0x16, 0x6c, 0xea, 0x1d, 0x2b, 0xd9, 0x65, 0xd0, 0x3f, 0x50, 0xd0, 0x0d, 0x6b, 0x80, + 0xfb, 0x67, 0x86, 0x5e, 0xca, 0xa2, 0x3d, 0xd8, 0x95, 0x5b, 0xdb, 0x6a, 0x1f, 0x1b, 0xfa, 0x69, + 0xac, 0xca, 0xb5, 0xbe, 0x28, 0xf0, 0xdf, 0x88, 0x7a, 0xab, 0x86, 0xd2, 0xda, 0xed, 0x7a, 0x1c, + 0xa7, 0x9f, 0x48, 0x33, 0x6e, 0xc9, 0x54, 0xce, 0x9f, 0x4b, 0xe5, 0x84, 0xce, 0x1c, 0x7f, 0xa2, + 0x52, 0x36, 0xd1, 0x26, 0xc4, 0x17, 0x0d, 0x6b, 0xc9, 0x91, 0x13, 0xb8, 0xfc, 0x8f, 0xef, 0xec, + 0xe1, 0xd4, 0xe3, 0x3f, 0x15, 0xe5, 0x6b, 0xa6, 0xfc, 0x2e, 0xf1, 0xb6, 0xc5, 0x53, 0xba, 0x1e, + 0x57, 0x87, 0xcd, 0xef, 0x69, 0xf5, 0x42, 0x54, 0x2f, 0xba, 0x1e, 0xbf, 0x18, 0x36, 0x2f, 0x37, + 0x04, 0xf1, 0xd9, 0xaf, 0x00, 0x00, 0x00, 0xff, 0xff, 0xe2, 0xc0, 0x0a, 0x0e, 0xb8, 0x05, 0x00, + 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/kms/v1/service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/kms/v1/service.pb.go new file mode 100644 index 000000000..825fb1aed --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/kms/v1/service.pb.go @@ -0,0 +1,1905 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/kms/v1/service.proto + +package kms // import "google.golang.org/genproto/googleapis/cloud/kms/v1" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _ "github.com/golang/protobuf/ptypes/struct" +import _ "github.com/golang/protobuf/ptypes/wrappers" +import _ "google.golang.org/genproto/googleapis/api/annotations" +import field_mask "google.golang.org/genproto/protobuf/field_mask" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Request message for [KeyManagementService.ListKeyRings][google.cloud.kms.v1.KeyManagementService.ListKeyRings]. +type ListKeyRingsRequest struct { + // Required. The resource name of the location associated with the + // [KeyRings][google.cloud.kms.v1.KeyRing], in the format `projects/*/locations/*`. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Optional limit on the number of [KeyRings][google.cloud.kms.v1.KeyRing] to include in the + // response. Further [KeyRings][google.cloud.kms.v1.KeyRing] can subsequently be obtained by + // including the [ListKeyRingsResponse.next_page_token][google.cloud.kms.v1.ListKeyRingsResponse.next_page_token] in a subsequent + // request. If unspecified, the server will pick an appropriate default. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Optional pagination token, returned earlier via + // [ListKeyRingsResponse.next_page_token][google.cloud.kms.v1.ListKeyRingsResponse.next_page_token]. + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListKeyRingsRequest) Reset() { *m = ListKeyRingsRequest{} } +func (m *ListKeyRingsRequest) String() string { return proto.CompactTextString(m) } +func (*ListKeyRingsRequest) ProtoMessage() {} +func (*ListKeyRingsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{0} +} +func (m *ListKeyRingsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListKeyRingsRequest.Unmarshal(m, b) +} +func (m *ListKeyRingsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListKeyRingsRequest.Marshal(b, m, deterministic) +} +func (dst *ListKeyRingsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListKeyRingsRequest.Merge(dst, src) +} +func (m *ListKeyRingsRequest) XXX_Size() int { + return xxx_messageInfo_ListKeyRingsRequest.Size(m) +} +func (m *ListKeyRingsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListKeyRingsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListKeyRingsRequest proto.InternalMessageInfo + +func (m *ListKeyRingsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListKeyRingsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListKeyRingsRequest) GetPageToken() string { + if m != nil { + return m.PageToken + } + return "" +} + +// Request message for [KeyManagementService.ListCryptoKeys][google.cloud.kms.v1.KeyManagementService.ListCryptoKeys]. +type ListCryptoKeysRequest struct { + // Required. The resource name of the [KeyRing][google.cloud.kms.v1.KeyRing] to list, in the format + // `projects/*/locations/*/keyRings/*`. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Optional limit on the number of [CryptoKeys][google.cloud.kms.v1.CryptoKey] to include in the + // response. Further [CryptoKeys][google.cloud.kms.v1.CryptoKey] can subsequently be obtained by + // including the [ListCryptoKeysResponse.next_page_token][google.cloud.kms.v1.ListCryptoKeysResponse.next_page_token] in a subsequent + // request. If unspecified, the server will pick an appropriate default. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Optional pagination token, returned earlier via + // [ListCryptoKeysResponse.next_page_token][google.cloud.kms.v1.ListCryptoKeysResponse.next_page_token]. + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListCryptoKeysRequest) Reset() { *m = ListCryptoKeysRequest{} } +func (m *ListCryptoKeysRequest) String() string { return proto.CompactTextString(m) } +func (*ListCryptoKeysRequest) ProtoMessage() {} +func (*ListCryptoKeysRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{1} +} +func (m *ListCryptoKeysRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListCryptoKeysRequest.Unmarshal(m, b) +} +func (m *ListCryptoKeysRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListCryptoKeysRequest.Marshal(b, m, deterministic) +} +func (dst *ListCryptoKeysRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListCryptoKeysRequest.Merge(dst, src) +} +func (m *ListCryptoKeysRequest) XXX_Size() int { + return xxx_messageInfo_ListCryptoKeysRequest.Size(m) +} +func (m *ListCryptoKeysRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListCryptoKeysRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListCryptoKeysRequest proto.InternalMessageInfo + +func (m *ListCryptoKeysRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListCryptoKeysRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListCryptoKeysRequest) GetPageToken() string { + if m != nil { + return m.PageToken + } + return "" +} + +// Request message for [KeyManagementService.ListCryptoKeyVersions][google.cloud.kms.v1.KeyManagementService.ListCryptoKeyVersions]. +type ListCryptoKeyVersionsRequest struct { + // Required. The resource name of the [CryptoKey][google.cloud.kms.v1.CryptoKey] to list, in the format + // `projects/*/locations/*/keyRings/*/cryptoKeys/*`. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Optional limit on the number of [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion] to + // include in the response. Further [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion] can + // subsequently be obtained by including the + // [ListCryptoKeyVersionsResponse.next_page_token][google.cloud.kms.v1.ListCryptoKeyVersionsResponse.next_page_token] in a subsequent request. + // If unspecified, the server will pick an appropriate default. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Optional pagination token, returned earlier via + // [ListCryptoKeyVersionsResponse.next_page_token][google.cloud.kms.v1.ListCryptoKeyVersionsResponse.next_page_token]. + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListCryptoKeyVersionsRequest) Reset() { *m = ListCryptoKeyVersionsRequest{} } +func (m *ListCryptoKeyVersionsRequest) String() string { return proto.CompactTextString(m) } +func (*ListCryptoKeyVersionsRequest) ProtoMessage() {} +func (*ListCryptoKeyVersionsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{2} +} +func (m *ListCryptoKeyVersionsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListCryptoKeyVersionsRequest.Unmarshal(m, b) +} +func (m *ListCryptoKeyVersionsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListCryptoKeyVersionsRequest.Marshal(b, m, deterministic) +} +func (dst *ListCryptoKeyVersionsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListCryptoKeyVersionsRequest.Merge(dst, src) +} +func (m *ListCryptoKeyVersionsRequest) XXX_Size() int { + return xxx_messageInfo_ListCryptoKeyVersionsRequest.Size(m) +} +func (m *ListCryptoKeyVersionsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListCryptoKeyVersionsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListCryptoKeyVersionsRequest proto.InternalMessageInfo + +func (m *ListCryptoKeyVersionsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListCryptoKeyVersionsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListCryptoKeyVersionsRequest) GetPageToken() string { + if m != nil { + return m.PageToken + } + return "" +} + +// Response message for [KeyManagementService.ListKeyRings][google.cloud.kms.v1.KeyManagementService.ListKeyRings]. +type ListKeyRingsResponse struct { + // The list of [KeyRings][google.cloud.kms.v1.KeyRing]. + KeyRings []*KeyRing `protobuf:"bytes,1,rep,name=key_rings,json=keyRings,proto3" json:"key_rings,omitempty"` + // A token to retrieve next page of results. Pass this value in + // [ListKeyRingsRequest.page_token][google.cloud.kms.v1.ListKeyRingsRequest.page_token] to retrieve the next page of results. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + // The total number of [KeyRings][google.cloud.kms.v1.KeyRing] that matched the query. + TotalSize int32 `protobuf:"varint,3,opt,name=total_size,json=totalSize,proto3" json:"total_size,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListKeyRingsResponse) Reset() { *m = ListKeyRingsResponse{} } +func (m *ListKeyRingsResponse) String() string { return proto.CompactTextString(m) } +func (*ListKeyRingsResponse) ProtoMessage() {} +func (*ListKeyRingsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{3} +} +func (m *ListKeyRingsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListKeyRingsResponse.Unmarshal(m, b) +} +func (m *ListKeyRingsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListKeyRingsResponse.Marshal(b, m, deterministic) +} +func (dst *ListKeyRingsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListKeyRingsResponse.Merge(dst, src) +} +func (m *ListKeyRingsResponse) XXX_Size() int { + return xxx_messageInfo_ListKeyRingsResponse.Size(m) +} +func (m *ListKeyRingsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListKeyRingsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListKeyRingsResponse proto.InternalMessageInfo + +func (m *ListKeyRingsResponse) GetKeyRings() []*KeyRing { + if m != nil { + return m.KeyRings + } + return nil +} + +func (m *ListKeyRingsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +func (m *ListKeyRingsResponse) GetTotalSize() int32 { + if m != nil { + return m.TotalSize + } + return 0 +} + +// Response message for [KeyManagementService.ListCryptoKeys][google.cloud.kms.v1.KeyManagementService.ListCryptoKeys]. +type ListCryptoKeysResponse struct { + // The list of [CryptoKeys][google.cloud.kms.v1.CryptoKey]. + CryptoKeys []*CryptoKey `protobuf:"bytes,1,rep,name=crypto_keys,json=cryptoKeys,proto3" json:"crypto_keys,omitempty"` + // A token to retrieve next page of results. Pass this value in + // [ListCryptoKeysRequest.page_token][google.cloud.kms.v1.ListCryptoKeysRequest.page_token] to retrieve the next page of results. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + // The total number of [CryptoKeys][google.cloud.kms.v1.CryptoKey] that matched the query. + TotalSize int32 `protobuf:"varint,3,opt,name=total_size,json=totalSize,proto3" json:"total_size,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListCryptoKeysResponse) Reset() { *m = ListCryptoKeysResponse{} } +func (m *ListCryptoKeysResponse) String() string { return proto.CompactTextString(m) } +func (*ListCryptoKeysResponse) ProtoMessage() {} +func (*ListCryptoKeysResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{4} +} +func (m *ListCryptoKeysResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListCryptoKeysResponse.Unmarshal(m, b) +} +func (m *ListCryptoKeysResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListCryptoKeysResponse.Marshal(b, m, deterministic) +} +func (dst *ListCryptoKeysResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListCryptoKeysResponse.Merge(dst, src) +} +func (m *ListCryptoKeysResponse) XXX_Size() int { + return xxx_messageInfo_ListCryptoKeysResponse.Size(m) +} +func (m *ListCryptoKeysResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListCryptoKeysResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListCryptoKeysResponse proto.InternalMessageInfo + +func (m *ListCryptoKeysResponse) GetCryptoKeys() []*CryptoKey { + if m != nil { + return m.CryptoKeys + } + return nil +} + +func (m *ListCryptoKeysResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +func (m *ListCryptoKeysResponse) GetTotalSize() int32 { + if m != nil { + return m.TotalSize + } + return 0 +} + +// Response message for [KeyManagementService.ListCryptoKeyVersions][google.cloud.kms.v1.KeyManagementService.ListCryptoKeyVersions]. +type ListCryptoKeyVersionsResponse struct { + // The list of [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion]. + CryptoKeyVersions []*CryptoKeyVersion `protobuf:"bytes,1,rep,name=crypto_key_versions,json=cryptoKeyVersions,proto3" json:"crypto_key_versions,omitempty"` + // A token to retrieve next page of results. Pass this value in + // [ListCryptoKeyVersionsRequest.page_token][google.cloud.kms.v1.ListCryptoKeyVersionsRequest.page_token] to retrieve the next page of + // results. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + // The total number of [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion] that matched the + // query. + TotalSize int32 `protobuf:"varint,3,opt,name=total_size,json=totalSize,proto3" json:"total_size,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListCryptoKeyVersionsResponse) Reset() { *m = ListCryptoKeyVersionsResponse{} } +func (m *ListCryptoKeyVersionsResponse) String() string { return proto.CompactTextString(m) } +func (*ListCryptoKeyVersionsResponse) ProtoMessage() {} +func (*ListCryptoKeyVersionsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{5} +} +func (m *ListCryptoKeyVersionsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListCryptoKeyVersionsResponse.Unmarshal(m, b) +} +func (m *ListCryptoKeyVersionsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListCryptoKeyVersionsResponse.Marshal(b, m, deterministic) +} +func (dst *ListCryptoKeyVersionsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListCryptoKeyVersionsResponse.Merge(dst, src) +} +func (m *ListCryptoKeyVersionsResponse) XXX_Size() int { + return xxx_messageInfo_ListCryptoKeyVersionsResponse.Size(m) +} +func (m *ListCryptoKeyVersionsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListCryptoKeyVersionsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListCryptoKeyVersionsResponse proto.InternalMessageInfo + +func (m *ListCryptoKeyVersionsResponse) GetCryptoKeyVersions() []*CryptoKeyVersion { + if m != nil { + return m.CryptoKeyVersions + } + return nil +} + +func (m *ListCryptoKeyVersionsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +func (m *ListCryptoKeyVersionsResponse) GetTotalSize() int32 { + if m != nil { + return m.TotalSize + } + return 0 +} + +// Request message for [KeyManagementService.GetKeyRing][google.cloud.kms.v1.KeyManagementService.GetKeyRing]. +type GetKeyRingRequest struct { + // The [name][google.cloud.kms.v1.KeyRing.name] of the [KeyRing][google.cloud.kms.v1.KeyRing] to get. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetKeyRingRequest) Reset() { *m = GetKeyRingRequest{} } +func (m *GetKeyRingRequest) String() string { return proto.CompactTextString(m) } +func (*GetKeyRingRequest) ProtoMessage() {} +func (*GetKeyRingRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{6} +} +func (m *GetKeyRingRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetKeyRingRequest.Unmarshal(m, b) +} +func (m *GetKeyRingRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetKeyRingRequest.Marshal(b, m, deterministic) +} +func (dst *GetKeyRingRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetKeyRingRequest.Merge(dst, src) +} +func (m *GetKeyRingRequest) XXX_Size() int { + return xxx_messageInfo_GetKeyRingRequest.Size(m) +} +func (m *GetKeyRingRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetKeyRingRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetKeyRingRequest proto.InternalMessageInfo + +func (m *GetKeyRingRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [KeyManagementService.GetCryptoKey][google.cloud.kms.v1.KeyManagementService.GetCryptoKey]. +type GetCryptoKeyRequest struct { + // The [name][google.cloud.kms.v1.CryptoKey.name] of the [CryptoKey][google.cloud.kms.v1.CryptoKey] to get. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetCryptoKeyRequest) Reset() { *m = GetCryptoKeyRequest{} } +func (m *GetCryptoKeyRequest) String() string { return proto.CompactTextString(m) } +func (*GetCryptoKeyRequest) ProtoMessage() {} +func (*GetCryptoKeyRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{7} +} +func (m *GetCryptoKeyRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetCryptoKeyRequest.Unmarshal(m, b) +} +func (m *GetCryptoKeyRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetCryptoKeyRequest.Marshal(b, m, deterministic) +} +func (dst *GetCryptoKeyRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetCryptoKeyRequest.Merge(dst, src) +} +func (m *GetCryptoKeyRequest) XXX_Size() int { + return xxx_messageInfo_GetCryptoKeyRequest.Size(m) +} +func (m *GetCryptoKeyRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetCryptoKeyRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetCryptoKeyRequest proto.InternalMessageInfo + +func (m *GetCryptoKeyRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [KeyManagementService.GetCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.GetCryptoKeyVersion]. +type GetCryptoKeyVersionRequest struct { + // The [name][google.cloud.kms.v1.CryptoKeyVersion.name] of the [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] to get. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetCryptoKeyVersionRequest) Reset() { *m = GetCryptoKeyVersionRequest{} } +func (m *GetCryptoKeyVersionRequest) String() string { return proto.CompactTextString(m) } +func (*GetCryptoKeyVersionRequest) ProtoMessage() {} +func (*GetCryptoKeyVersionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{8} +} +func (m *GetCryptoKeyVersionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetCryptoKeyVersionRequest.Unmarshal(m, b) +} +func (m *GetCryptoKeyVersionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetCryptoKeyVersionRequest.Marshal(b, m, deterministic) +} +func (dst *GetCryptoKeyVersionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetCryptoKeyVersionRequest.Merge(dst, src) +} +func (m *GetCryptoKeyVersionRequest) XXX_Size() int { + return xxx_messageInfo_GetCryptoKeyVersionRequest.Size(m) +} +func (m *GetCryptoKeyVersionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetCryptoKeyVersionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetCryptoKeyVersionRequest proto.InternalMessageInfo + +func (m *GetCryptoKeyVersionRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [KeyManagementService.CreateKeyRing][google.cloud.kms.v1.KeyManagementService.CreateKeyRing]. +type CreateKeyRingRequest struct { + // Required. The resource name of the location associated with the + // [KeyRings][google.cloud.kms.v1.KeyRing], in the format `projects/*/locations/*`. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Required. It must be unique within a location and match the regular + // expression `[a-zA-Z0-9_-]{1,63}` + KeyRingId string `protobuf:"bytes,2,opt,name=key_ring_id,json=keyRingId,proto3" json:"key_ring_id,omitempty"` + // A [KeyRing][google.cloud.kms.v1.KeyRing] with initial field values. + KeyRing *KeyRing `protobuf:"bytes,3,opt,name=key_ring,json=keyRing,proto3" json:"key_ring,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CreateKeyRingRequest) Reset() { *m = CreateKeyRingRequest{} } +func (m *CreateKeyRingRequest) String() string { return proto.CompactTextString(m) } +func (*CreateKeyRingRequest) ProtoMessage() {} +func (*CreateKeyRingRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{9} +} +func (m *CreateKeyRingRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CreateKeyRingRequest.Unmarshal(m, b) +} +func (m *CreateKeyRingRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CreateKeyRingRequest.Marshal(b, m, deterministic) +} +func (dst *CreateKeyRingRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CreateKeyRingRequest.Merge(dst, src) +} +func (m *CreateKeyRingRequest) XXX_Size() int { + return xxx_messageInfo_CreateKeyRingRequest.Size(m) +} +func (m *CreateKeyRingRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CreateKeyRingRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CreateKeyRingRequest proto.InternalMessageInfo + +func (m *CreateKeyRingRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *CreateKeyRingRequest) GetKeyRingId() string { + if m != nil { + return m.KeyRingId + } + return "" +} + +func (m *CreateKeyRingRequest) GetKeyRing() *KeyRing { + if m != nil { + return m.KeyRing + } + return nil +} + +// Request message for [KeyManagementService.CreateCryptoKey][google.cloud.kms.v1.KeyManagementService.CreateCryptoKey]. +type CreateCryptoKeyRequest struct { + // Required. The [name][google.cloud.kms.v1.KeyRing.name] of the KeyRing associated with the + // [CryptoKeys][google.cloud.kms.v1.CryptoKey]. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Required. It must be unique within a KeyRing and match the regular + // expression `[a-zA-Z0-9_-]{1,63}` + CryptoKeyId string `protobuf:"bytes,2,opt,name=crypto_key_id,json=cryptoKeyId,proto3" json:"crypto_key_id,omitempty"` + // A [CryptoKey][google.cloud.kms.v1.CryptoKey] with initial field values. + CryptoKey *CryptoKey `protobuf:"bytes,3,opt,name=crypto_key,json=cryptoKey,proto3" json:"crypto_key,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CreateCryptoKeyRequest) Reset() { *m = CreateCryptoKeyRequest{} } +func (m *CreateCryptoKeyRequest) String() string { return proto.CompactTextString(m) } +func (*CreateCryptoKeyRequest) ProtoMessage() {} +func (*CreateCryptoKeyRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{10} +} +func (m *CreateCryptoKeyRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CreateCryptoKeyRequest.Unmarshal(m, b) +} +func (m *CreateCryptoKeyRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CreateCryptoKeyRequest.Marshal(b, m, deterministic) +} +func (dst *CreateCryptoKeyRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CreateCryptoKeyRequest.Merge(dst, src) +} +func (m *CreateCryptoKeyRequest) XXX_Size() int { + return xxx_messageInfo_CreateCryptoKeyRequest.Size(m) +} +func (m *CreateCryptoKeyRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CreateCryptoKeyRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CreateCryptoKeyRequest proto.InternalMessageInfo + +func (m *CreateCryptoKeyRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *CreateCryptoKeyRequest) GetCryptoKeyId() string { + if m != nil { + return m.CryptoKeyId + } + return "" +} + +func (m *CreateCryptoKeyRequest) GetCryptoKey() *CryptoKey { + if m != nil { + return m.CryptoKey + } + return nil +} + +// Request message for [KeyManagementService.CreateCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.CreateCryptoKeyVersion]. +type CreateCryptoKeyVersionRequest struct { + // Required. The [name][google.cloud.kms.v1.CryptoKey.name] of the [CryptoKey][google.cloud.kms.v1.CryptoKey] associated with + // the [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion]. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // A [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] with initial field values. + CryptoKeyVersion *CryptoKeyVersion `protobuf:"bytes,2,opt,name=crypto_key_version,json=cryptoKeyVersion,proto3" json:"crypto_key_version,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CreateCryptoKeyVersionRequest) Reset() { *m = CreateCryptoKeyVersionRequest{} } +func (m *CreateCryptoKeyVersionRequest) String() string { return proto.CompactTextString(m) } +func (*CreateCryptoKeyVersionRequest) ProtoMessage() {} +func (*CreateCryptoKeyVersionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{11} +} +func (m *CreateCryptoKeyVersionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CreateCryptoKeyVersionRequest.Unmarshal(m, b) +} +func (m *CreateCryptoKeyVersionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CreateCryptoKeyVersionRequest.Marshal(b, m, deterministic) +} +func (dst *CreateCryptoKeyVersionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CreateCryptoKeyVersionRequest.Merge(dst, src) +} +func (m *CreateCryptoKeyVersionRequest) XXX_Size() int { + return xxx_messageInfo_CreateCryptoKeyVersionRequest.Size(m) +} +func (m *CreateCryptoKeyVersionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CreateCryptoKeyVersionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CreateCryptoKeyVersionRequest proto.InternalMessageInfo + +func (m *CreateCryptoKeyVersionRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *CreateCryptoKeyVersionRequest) GetCryptoKeyVersion() *CryptoKeyVersion { + if m != nil { + return m.CryptoKeyVersion + } + return nil +} + +// Request message for [KeyManagementService.UpdateCryptoKey][google.cloud.kms.v1.KeyManagementService.UpdateCryptoKey]. +type UpdateCryptoKeyRequest struct { + // [CryptoKey][google.cloud.kms.v1.CryptoKey] with updated values. + CryptoKey *CryptoKey `protobuf:"bytes,1,opt,name=crypto_key,json=cryptoKey,proto3" json:"crypto_key,omitempty"` + // Required list of fields to be updated in this request. + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *UpdateCryptoKeyRequest) Reset() { *m = UpdateCryptoKeyRequest{} } +func (m *UpdateCryptoKeyRequest) String() string { return proto.CompactTextString(m) } +func (*UpdateCryptoKeyRequest) ProtoMessage() {} +func (*UpdateCryptoKeyRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{12} +} +func (m *UpdateCryptoKeyRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_UpdateCryptoKeyRequest.Unmarshal(m, b) +} +func (m *UpdateCryptoKeyRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_UpdateCryptoKeyRequest.Marshal(b, m, deterministic) +} +func (dst *UpdateCryptoKeyRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_UpdateCryptoKeyRequest.Merge(dst, src) +} +func (m *UpdateCryptoKeyRequest) XXX_Size() int { + return xxx_messageInfo_UpdateCryptoKeyRequest.Size(m) +} +func (m *UpdateCryptoKeyRequest) XXX_DiscardUnknown() { + xxx_messageInfo_UpdateCryptoKeyRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_UpdateCryptoKeyRequest proto.InternalMessageInfo + +func (m *UpdateCryptoKeyRequest) GetCryptoKey() *CryptoKey { + if m != nil { + return m.CryptoKey + } + return nil +} + +func (m *UpdateCryptoKeyRequest) GetUpdateMask() *field_mask.FieldMask { + if m != nil { + return m.UpdateMask + } + return nil +} + +// Request message for [KeyManagementService.UpdateCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.UpdateCryptoKeyVersion]. +type UpdateCryptoKeyVersionRequest struct { + // [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] with updated values. + CryptoKeyVersion *CryptoKeyVersion `protobuf:"bytes,1,opt,name=crypto_key_version,json=cryptoKeyVersion,proto3" json:"crypto_key_version,omitempty"` + // Required list of fields to be updated in this request. + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *UpdateCryptoKeyVersionRequest) Reset() { *m = UpdateCryptoKeyVersionRequest{} } +func (m *UpdateCryptoKeyVersionRequest) String() string { return proto.CompactTextString(m) } +func (*UpdateCryptoKeyVersionRequest) ProtoMessage() {} +func (*UpdateCryptoKeyVersionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{13} +} +func (m *UpdateCryptoKeyVersionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_UpdateCryptoKeyVersionRequest.Unmarshal(m, b) +} +func (m *UpdateCryptoKeyVersionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_UpdateCryptoKeyVersionRequest.Marshal(b, m, deterministic) +} +func (dst *UpdateCryptoKeyVersionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_UpdateCryptoKeyVersionRequest.Merge(dst, src) +} +func (m *UpdateCryptoKeyVersionRequest) XXX_Size() int { + return xxx_messageInfo_UpdateCryptoKeyVersionRequest.Size(m) +} +func (m *UpdateCryptoKeyVersionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_UpdateCryptoKeyVersionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_UpdateCryptoKeyVersionRequest proto.InternalMessageInfo + +func (m *UpdateCryptoKeyVersionRequest) GetCryptoKeyVersion() *CryptoKeyVersion { + if m != nil { + return m.CryptoKeyVersion + } + return nil +} + +func (m *UpdateCryptoKeyVersionRequest) GetUpdateMask() *field_mask.FieldMask { + if m != nil { + return m.UpdateMask + } + return nil +} + +// Request message for [KeyManagementService.Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt]. +type EncryptRequest struct { + // Required. The resource name of the [CryptoKey][google.cloud.kms.v1.CryptoKey] or [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] + // to use for encryption. + // + // If a [CryptoKey][google.cloud.kms.v1.CryptoKey] is specified, the server will use its + // [primary version][google.cloud.kms.v1.CryptoKey.primary]. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Required. The data to encrypt. Must be no larger than 64KiB. + Plaintext []byte `protobuf:"bytes,2,opt,name=plaintext,proto3" json:"plaintext,omitempty"` + // Optional data that, if specified, must also be provided during decryption + // through [DecryptRequest.additional_authenticated_data][google.cloud.kms.v1.DecryptRequest.additional_authenticated_data]. Must be no + // larger than 64KiB. + AdditionalAuthenticatedData []byte `protobuf:"bytes,3,opt,name=additional_authenticated_data,json=additionalAuthenticatedData,proto3" json:"additional_authenticated_data,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *EncryptRequest) Reset() { *m = EncryptRequest{} } +func (m *EncryptRequest) String() string { return proto.CompactTextString(m) } +func (*EncryptRequest) ProtoMessage() {} +func (*EncryptRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{14} +} +func (m *EncryptRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_EncryptRequest.Unmarshal(m, b) +} +func (m *EncryptRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_EncryptRequest.Marshal(b, m, deterministic) +} +func (dst *EncryptRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_EncryptRequest.Merge(dst, src) +} +func (m *EncryptRequest) XXX_Size() int { + return xxx_messageInfo_EncryptRequest.Size(m) +} +func (m *EncryptRequest) XXX_DiscardUnknown() { + xxx_messageInfo_EncryptRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_EncryptRequest proto.InternalMessageInfo + +func (m *EncryptRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *EncryptRequest) GetPlaintext() []byte { + if m != nil { + return m.Plaintext + } + return nil +} + +func (m *EncryptRequest) GetAdditionalAuthenticatedData() []byte { + if m != nil { + return m.AdditionalAuthenticatedData + } + return nil +} + +// Request message for [KeyManagementService.Decrypt][google.cloud.kms.v1.KeyManagementService.Decrypt]. +type DecryptRequest struct { + // Required. The resource name of the [CryptoKey][google.cloud.kms.v1.CryptoKey] to use for decryption. + // The server will choose the appropriate version. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Required. The encrypted data originally returned in + // [EncryptResponse.ciphertext][google.cloud.kms.v1.EncryptResponse.ciphertext]. + Ciphertext []byte `protobuf:"bytes,2,opt,name=ciphertext,proto3" json:"ciphertext,omitempty"` + // Optional data that must match the data originally supplied in + // [EncryptRequest.additional_authenticated_data][google.cloud.kms.v1.EncryptRequest.additional_authenticated_data]. + AdditionalAuthenticatedData []byte `protobuf:"bytes,3,opt,name=additional_authenticated_data,json=additionalAuthenticatedData,proto3" json:"additional_authenticated_data,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DecryptRequest) Reset() { *m = DecryptRequest{} } +func (m *DecryptRequest) String() string { return proto.CompactTextString(m) } +func (*DecryptRequest) ProtoMessage() {} +func (*DecryptRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{15} +} +func (m *DecryptRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DecryptRequest.Unmarshal(m, b) +} +func (m *DecryptRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DecryptRequest.Marshal(b, m, deterministic) +} +func (dst *DecryptRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_DecryptRequest.Merge(dst, src) +} +func (m *DecryptRequest) XXX_Size() int { + return xxx_messageInfo_DecryptRequest.Size(m) +} +func (m *DecryptRequest) XXX_DiscardUnknown() { + xxx_messageInfo_DecryptRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_DecryptRequest proto.InternalMessageInfo + +func (m *DecryptRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *DecryptRequest) GetCiphertext() []byte { + if m != nil { + return m.Ciphertext + } + return nil +} + +func (m *DecryptRequest) GetAdditionalAuthenticatedData() []byte { + if m != nil { + return m.AdditionalAuthenticatedData + } + return nil +} + +// Response message for [KeyManagementService.Decrypt][google.cloud.kms.v1.KeyManagementService.Decrypt]. +type DecryptResponse struct { + // The decrypted data originally supplied in [EncryptRequest.plaintext][google.cloud.kms.v1.EncryptRequest.plaintext]. + Plaintext []byte `protobuf:"bytes,1,opt,name=plaintext,proto3" json:"plaintext,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DecryptResponse) Reset() { *m = DecryptResponse{} } +func (m *DecryptResponse) String() string { return proto.CompactTextString(m) } +func (*DecryptResponse) ProtoMessage() {} +func (*DecryptResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{16} +} +func (m *DecryptResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DecryptResponse.Unmarshal(m, b) +} +func (m *DecryptResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DecryptResponse.Marshal(b, m, deterministic) +} +func (dst *DecryptResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_DecryptResponse.Merge(dst, src) +} +func (m *DecryptResponse) XXX_Size() int { + return xxx_messageInfo_DecryptResponse.Size(m) +} +func (m *DecryptResponse) XXX_DiscardUnknown() { + xxx_messageInfo_DecryptResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_DecryptResponse proto.InternalMessageInfo + +func (m *DecryptResponse) GetPlaintext() []byte { + if m != nil { + return m.Plaintext + } + return nil +} + +// Response message for [KeyManagementService.Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt]. +type EncryptResponse struct { + // The resource name of the [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] used in encryption. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The encrypted data. + Ciphertext []byte `protobuf:"bytes,2,opt,name=ciphertext,proto3" json:"ciphertext,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *EncryptResponse) Reset() { *m = EncryptResponse{} } +func (m *EncryptResponse) String() string { return proto.CompactTextString(m) } +func (*EncryptResponse) ProtoMessage() {} +func (*EncryptResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{17} +} +func (m *EncryptResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_EncryptResponse.Unmarshal(m, b) +} +func (m *EncryptResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_EncryptResponse.Marshal(b, m, deterministic) +} +func (dst *EncryptResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_EncryptResponse.Merge(dst, src) +} +func (m *EncryptResponse) XXX_Size() int { + return xxx_messageInfo_EncryptResponse.Size(m) +} +func (m *EncryptResponse) XXX_DiscardUnknown() { + xxx_messageInfo_EncryptResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_EncryptResponse proto.InternalMessageInfo + +func (m *EncryptResponse) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *EncryptResponse) GetCiphertext() []byte { + if m != nil { + return m.Ciphertext + } + return nil +} + +// Request message for [KeyManagementService.UpdateCryptoKeyPrimaryVersion][google.cloud.kms.v1.KeyManagementService.UpdateCryptoKeyPrimaryVersion]. +type UpdateCryptoKeyPrimaryVersionRequest struct { + // The resource name of the [CryptoKey][google.cloud.kms.v1.CryptoKey] to update. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The id of the child [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] to use as primary. + CryptoKeyVersionId string `protobuf:"bytes,2,opt,name=crypto_key_version_id,json=cryptoKeyVersionId,proto3" json:"crypto_key_version_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *UpdateCryptoKeyPrimaryVersionRequest) Reset() { *m = UpdateCryptoKeyPrimaryVersionRequest{} } +func (m *UpdateCryptoKeyPrimaryVersionRequest) String() string { return proto.CompactTextString(m) } +func (*UpdateCryptoKeyPrimaryVersionRequest) ProtoMessage() {} +func (*UpdateCryptoKeyPrimaryVersionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{18} +} +func (m *UpdateCryptoKeyPrimaryVersionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_UpdateCryptoKeyPrimaryVersionRequest.Unmarshal(m, b) +} +func (m *UpdateCryptoKeyPrimaryVersionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_UpdateCryptoKeyPrimaryVersionRequest.Marshal(b, m, deterministic) +} +func (dst *UpdateCryptoKeyPrimaryVersionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_UpdateCryptoKeyPrimaryVersionRequest.Merge(dst, src) +} +func (m *UpdateCryptoKeyPrimaryVersionRequest) XXX_Size() int { + return xxx_messageInfo_UpdateCryptoKeyPrimaryVersionRequest.Size(m) +} +func (m *UpdateCryptoKeyPrimaryVersionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_UpdateCryptoKeyPrimaryVersionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_UpdateCryptoKeyPrimaryVersionRequest proto.InternalMessageInfo + +func (m *UpdateCryptoKeyPrimaryVersionRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *UpdateCryptoKeyPrimaryVersionRequest) GetCryptoKeyVersionId() string { + if m != nil { + return m.CryptoKeyVersionId + } + return "" +} + +// Request message for [KeyManagementService.DestroyCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.DestroyCryptoKeyVersion]. +type DestroyCryptoKeyVersionRequest struct { + // The resource name of the [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] to destroy. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DestroyCryptoKeyVersionRequest) Reset() { *m = DestroyCryptoKeyVersionRequest{} } +func (m *DestroyCryptoKeyVersionRequest) String() string { return proto.CompactTextString(m) } +func (*DestroyCryptoKeyVersionRequest) ProtoMessage() {} +func (*DestroyCryptoKeyVersionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{19} +} +func (m *DestroyCryptoKeyVersionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DestroyCryptoKeyVersionRequest.Unmarshal(m, b) +} +func (m *DestroyCryptoKeyVersionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DestroyCryptoKeyVersionRequest.Marshal(b, m, deterministic) +} +func (dst *DestroyCryptoKeyVersionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_DestroyCryptoKeyVersionRequest.Merge(dst, src) +} +func (m *DestroyCryptoKeyVersionRequest) XXX_Size() int { + return xxx_messageInfo_DestroyCryptoKeyVersionRequest.Size(m) +} +func (m *DestroyCryptoKeyVersionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_DestroyCryptoKeyVersionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_DestroyCryptoKeyVersionRequest proto.InternalMessageInfo + +func (m *DestroyCryptoKeyVersionRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [KeyManagementService.RestoreCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.RestoreCryptoKeyVersion]. +type RestoreCryptoKeyVersionRequest struct { + // The resource name of the [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] to restore. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RestoreCryptoKeyVersionRequest) Reset() { *m = RestoreCryptoKeyVersionRequest{} } +func (m *RestoreCryptoKeyVersionRequest) String() string { return proto.CompactTextString(m) } +func (*RestoreCryptoKeyVersionRequest) ProtoMessage() {} +func (*RestoreCryptoKeyVersionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_service_d4f59c45586ec2a6, []int{20} +} +func (m *RestoreCryptoKeyVersionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RestoreCryptoKeyVersionRequest.Unmarshal(m, b) +} +func (m *RestoreCryptoKeyVersionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RestoreCryptoKeyVersionRequest.Marshal(b, m, deterministic) +} +func (dst *RestoreCryptoKeyVersionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_RestoreCryptoKeyVersionRequest.Merge(dst, src) +} +func (m *RestoreCryptoKeyVersionRequest) XXX_Size() int { + return xxx_messageInfo_RestoreCryptoKeyVersionRequest.Size(m) +} +func (m *RestoreCryptoKeyVersionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_RestoreCryptoKeyVersionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_RestoreCryptoKeyVersionRequest proto.InternalMessageInfo + +func (m *RestoreCryptoKeyVersionRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func init() { + proto.RegisterType((*ListKeyRingsRequest)(nil), "google.cloud.kms.v1.ListKeyRingsRequest") + proto.RegisterType((*ListCryptoKeysRequest)(nil), "google.cloud.kms.v1.ListCryptoKeysRequest") + proto.RegisterType((*ListCryptoKeyVersionsRequest)(nil), "google.cloud.kms.v1.ListCryptoKeyVersionsRequest") + proto.RegisterType((*ListKeyRingsResponse)(nil), "google.cloud.kms.v1.ListKeyRingsResponse") + proto.RegisterType((*ListCryptoKeysResponse)(nil), "google.cloud.kms.v1.ListCryptoKeysResponse") + proto.RegisterType((*ListCryptoKeyVersionsResponse)(nil), "google.cloud.kms.v1.ListCryptoKeyVersionsResponse") + proto.RegisterType((*GetKeyRingRequest)(nil), "google.cloud.kms.v1.GetKeyRingRequest") + proto.RegisterType((*GetCryptoKeyRequest)(nil), "google.cloud.kms.v1.GetCryptoKeyRequest") + proto.RegisterType((*GetCryptoKeyVersionRequest)(nil), "google.cloud.kms.v1.GetCryptoKeyVersionRequest") + proto.RegisterType((*CreateKeyRingRequest)(nil), "google.cloud.kms.v1.CreateKeyRingRequest") + proto.RegisterType((*CreateCryptoKeyRequest)(nil), "google.cloud.kms.v1.CreateCryptoKeyRequest") + proto.RegisterType((*CreateCryptoKeyVersionRequest)(nil), "google.cloud.kms.v1.CreateCryptoKeyVersionRequest") + proto.RegisterType((*UpdateCryptoKeyRequest)(nil), "google.cloud.kms.v1.UpdateCryptoKeyRequest") + proto.RegisterType((*UpdateCryptoKeyVersionRequest)(nil), "google.cloud.kms.v1.UpdateCryptoKeyVersionRequest") + proto.RegisterType((*EncryptRequest)(nil), "google.cloud.kms.v1.EncryptRequest") + proto.RegisterType((*DecryptRequest)(nil), "google.cloud.kms.v1.DecryptRequest") + proto.RegisterType((*DecryptResponse)(nil), "google.cloud.kms.v1.DecryptResponse") + proto.RegisterType((*EncryptResponse)(nil), "google.cloud.kms.v1.EncryptResponse") + proto.RegisterType((*UpdateCryptoKeyPrimaryVersionRequest)(nil), "google.cloud.kms.v1.UpdateCryptoKeyPrimaryVersionRequest") + proto.RegisterType((*DestroyCryptoKeyVersionRequest)(nil), "google.cloud.kms.v1.DestroyCryptoKeyVersionRequest") + proto.RegisterType((*RestoreCryptoKeyVersionRequest)(nil), "google.cloud.kms.v1.RestoreCryptoKeyVersionRequest") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// KeyManagementServiceClient is the client API for KeyManagementService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type KeyManagementServiceClient interface { + // Lists [KeyRings][google.cloud.kms.v1.KeyRing]. + ListKeyRings(ctx context.Context, in *ListKeyRingsRequest, opts ...grpc.CallOption) (*ListKeyRingsResponse, error) + // Lists [CryptoKeys][google.cloud.kms.v1.CryptoKey]. + ListCryptoKeys(ctx context.Context, in *ListCryptoKeysRequest, opts ...grpc.CallOption) (*ListCryptoKeysResponse, error) + // Lists [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion]. + ListCryptoKeyVersions(ctx context.Context, in *ListCryptoKeyVersionsRequest, opts ...grpc.CallOption) (*ListCryptoKeyVersionsResponse, error) + // Returns metadata for a given [KeyRing][google.cloud.kms.v1.KeyRing]. + GetKeyRing(ctx context.Context, in *GetKeyRingRequest, opts ...grpc.CallOption) (*KeyRing, error) + // Returns metadata for a given [CryptoKey][google.cloud.kms.v1.CryptoKey], as well as its + // [primary][google.cloud.kms.v1.CryptoKey.primary] [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]. + GetCryptoKey(ctx context.Context, in *GetCryptoKeyRequest, opts ...grpc.CallOption) (*CryptoKey, error) + // Returns metadata for a given [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]. + GetCryptoKeyVersion(ctx context.Context, in *GetCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) + // Create a new [KeyRing][google.cloud.kms.v1.KeyRing] in a given Project and Location. + CreateKeyRing(ctx context.Context, in *CreateKeyRingRequest, opts ...grpc.CallOption) (*KeyRing, error) + // Create a new [CryptoKey][google.cloud.kms.v1.CryptoKey] within a [KeyRing][google.cloud.kms.v1.KeyRing]. + // + // [CryptoKey.purpose][google.cloud.kms.v1.CryptoKey.purpose] is required. + CreateCryptoKey(ctx context.Context, in *CreateCryptoKeyRequest, opts ...grpc.CallOption) (*CryptoKey, error) + // Create a new [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] in a [CryptoKey][google.cloud.kms.v1.CryptoKey]. + // + // The server will assign the next sequential id. If unset, + // [state][google.cloud.kms.v1.CryptoKeyVersion.state] will be set to + // [ENABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.ENABLED]. + CreateCryptoKeyVersion(ctx context.Context, in *CreateCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) + // Update a [CryptoKey][google.cloud.kms.v1.CryptoKey]. + UpdateCryptoKey(ctx context.Context, in *UpdateCryptoKeyRequest, opts ...grpc.CallOption) (*CryptoKey, error) + // Update a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]'s metadata. + // + // [state][google.cloud.kms.v1.CryptoKeyVersion.state] may be changed between + // [ENABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.ENABLED] and + // [DISABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DISABLED] using this + // method. See [DestroyCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.DestroyCryptoKeyVersion] and [RestoreCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.RestoreCryptoKeyVersion] to + // move between other states. + UpdateCryptoKeyVersion(ctx context.Context, in *UpdateCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) + // Encrypts data, so that it can only be recovered by a call to [Decrypt][google.cloud.kms.v1.KeyManagementService.Decrypt]. + Encrypt(ctx context.Context, in *EncryptRequest, opts ...grpc.CallOption) (*EncryptResponse, error) + // Decrypts data that was protected by [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt]. + Decrypt(ctx context.Context, in *DecryptRequest, opts ...grpc.CallOption) (*DecryptResponse, error) + // Update the version of a [CryptoKey][google.cloud.kms.v1.CryptoKey] that will be used in [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt] + UpdateCryptoKeyPrimaryVersion(ctx context.Context, in *UpdateCryptoKeyPrimaryVersionRequest, opts ...grpc.CallOption) (*CryptoKey, error) + // Schedule a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] for destruction. + // + // Upon calling this method, [CryptoKeyVersion.state][google.cloud.kms.v1.CryptoKeyVersion.state] will be set to + // [DESTROY_SCHEDULED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROY_SCHEDULED] + // and [destroy_time][google.cloud.kms.v1.CryptoKeyVersion.destroy_time] will be set to a time 24 + // hours in the future, at which point the [state][google.cloud.kms.v1.CryptoKeyVersion.state] + // will be changed to + // [DESTROYED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROYED], and the key + // material will be irrevocably destroyed. + // + // Before the [destroy_time][google.cloud.kms.v1.CryptoKeyVersion.destroy_time] is reached, + // [RestoreCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.RestoreCryptoKeyVersion] may be called to reverse the process. + DestroyCryptoKeyVersion(ctx context.Context, in *DestroyCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) + // Restore a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] in the + // [DESTROY_SCHEDULED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROY_SCHEDULED], + // state. + // + // Upon restoration of the CryptoKeyVersion, [state][google.cloud.kms.v1.CryptoKeyVersion.state] + // will be set to [DISABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DISABLED], + // and [destroy_time][google.cloud.kms.v1.CryptoKeyVersion.destroy_time] will be cleared. + RestoreCryptoKeyVersion(ctx context.Context, in *RestoreCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) +} + +type keyManagementServiceClient struct { + cc *grpc.ClientConn +} + +func NewKeyManagementServiceClient(cc *grpc.ClientConn) KeyManagementServiceClient { + return &keyManagementServiceClient{cc} +} + +func (c *keyManagementServiceClient) ListKeyRings(ctx context.Context, in *ListKeyRingsRequest, opts ...grpc.CallOption) (*ListKeyRingsResponse, error) { + out := new(ListKeyRingsResponse) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/ListKeyRings", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) ListCryptoKeys(ctx context.Context, in *ListCryptoKeysRequest, opts ...grpc.CallOption) (*ListCryptoKeysResponse, error) { + out := new(ListCryptoKeysResponse) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/ListCryptoKeys", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) ListCryptoKeyVersions(ctx context.Context, in *ListCryptoKeyVersionsRequest, opts ...grpc.CallOption) (*ListCryptoKeyVersionsResponse, error) { + out := new(ListCryptoKeyVersionsResponse) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/ListCryptoKeyVersions", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) GetKeyRing(ctx context.Context, in *GetKeyRingRequest, opts ...grpc.CallOption) (*KeyRing, error) { + out := new(KeyRing) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/GetKeyRing", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) GetCryptoKey(ctx context.Context, in *GetCryptoKeyRequest, opts ...grpc.CallOption) (*CryptoKey, error) { + out := new(CryptoKey) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/GetCryptoKey", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) GetCryptoKeyVersion(ctx context.Context, in *GetCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) { + out := new(CryptoKeyVersion) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/GetCryptoKeyVersion", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) CreateKeyRing(ctx context.Context, in *CreateKeyRingRequest, opts ...grpc.CallOption) (*KeyRing, error) { + out := new(KeyRing) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/CreateKeyRing", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) CreateCryptoKey(ctx context.Context, in *CreateCryptoKeyRequest, opts ...grpc.CallOption) (*CryptoKey, error) { + out := new(CryptoKey) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/CreateCryptoKey", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) CreateCryptoKeyVersion(ctx context.Context, in *CreateCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) { + out := new(CryptoKeyVersion) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/CreateCryptoKeyVersion", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) UpdateCryptoKey(ctx context.Context, in *UpdateCryptoKeyRequest, opts ...grpc.CallOption) (*CryptoKey, error) { + out := new(CryptoKey) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/UpdateCryptoKey", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) UpdateCryptoKeyVersion(ctx context.Context, in *UpdateCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) { + out := new(CryptoKeyVersion) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/UpdateCryptoKeyVersion", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) Encrypt(ctx context.Context, in *EncryptRequest, opts ...grpc.CallOption) (*EncryptResponse, error) { + out := new(EncryptResponse) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/Encrypt", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) Decrypt(ctx context.Context, in *DecryptRequest, opts ...grpc.CallOption) (*DecryptResponse, error) { + out := new(DecryptResponse) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/Decrypt", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) UpdateCryptoKeyPrimaryVersion(ctx context.Context, in *UpdateCryptoKeyPrimaryVersionRequest, opts ...grpc.CallOption) (*CryptoKey, error) { + out := new(CryptoKey) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/UpdateCryptoKeyPrimaryVersion", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) DestroyCryptoKeyVersion(ctx context.Context, in *DestroyCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) { + out := new(CryptoKeyVersion) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/DestroyCryptoKeyVersion", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *keyManagementServiceClient) RestoreCryptoKeyVersion(ctx context.Context, in *RestoreCryptoKeyVersionRequest, opts ...grpc.CallOption) (*CryptoKeyVersion, error) { + out := new(CryptoKeyVersion) + err := c.cc.Invoke(ctx, "/google.cloud.kms.v1.KeyManagementService/RestoreCryptoKeyVersion", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// KeyManagementServiceServer is the server API for KeyManagementService service. +type KeyManagementServiceServer interface { + // Lists [KeyRings][google.cloud.kms.v1.KeyRing]. + ListKeyRings(context.Context, *ListKeyRingsRequest) (*ListKeyRingsResponse, error) + // Lists [CryptoKeys][google.cloud.kms.v1.CryptoKey]. + ListCryptoKeys(context.Context, *ListCryptoKeysRequest) (*ListCryptoKeysResponse, error) + // Lists [CryptoKeyVersions][google.cloud.kms.v1.CryptoKeyVersion]. + ListCryptoKeyVersions(context.Context, *ListCryptoKeyVersionsRequest) (*ListCryptoKeyVersionsResponse, error) + // Returns metadata for a given [KeyRing][google.cloud.kms.v1.KeyRing]. + GetKeyRing(context.Context, *GetKeyRingRequest) (*KeyRing, error) + // Returns metadata for a given [CryptoKey][google.cloud.kms.v1.CryptoKey], as well as its + // [primary][google.cloud.kms.v1.CryptoKey.primary] [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]. + GetCryptoKey(context.Context, *GetCryptoKeyRequest) (*CryptoKey, error) + // Returns metadata for a given [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]. + GetCryptoKeyVersion(context.Context, *GetCryptoKeyVersionRequest) (*CryptoKeyVersion, error) + // Create a new [KeyRing][google.cloud.kms.v1.KeyRing] in a given Project and Location. + CreateKeyRing(context.Context, *CreateKeyRingRequest) (*KeyRing, error) + // Create a new [CryptoKey][google.cloud.kms.v1.CryptoKey] within a [KeyRing][google.cloud.kms.v1.KeyRing]. + // + // [CryptoKey.purpose][google.cloud.kms.v1.CryptoKey.purpose] is required. + CreateCryptoKey(context.Context, *CreateCryptoKeyRequest) (*CryptoKey, error) + // Create a new [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] in a [CryptoKey][google.cloud.kms.v1.CryptoKey]. + // + // The server will assign the next sequential id. If unset, + // [state][google.cloud.kms.v1.CryptoKeyVersion.state] will be set to + // [ENABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.ENABLED]. + CreateCryptoKeyVersion(context.Context, *CreateCryptoKeyVersionRequest) (*CryptoKeyVersion, error) + // Update a [CryptoKey][google.cloud.kms.v1.CryptoKey]. + UpdateCryptoKey(context.Context, *UpdateCryptoKeyRequest) (*CryptoKey, error) + // Update a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion]'s metadata. + // + // [state][google.cloud.kms.v1.CryptoKeyVersion.state] may be changed between + // [ENABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.ENABLED] and + // [DISABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DISABLED] using this + // method. See [DestroyCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.DestroyCryptoKeyVersion] and [RestoreCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.RestoreCryptoKeyVersion] to + // move between other states. + UpdateCryptoKeyVersion(context.Context, *UpdateCryptoKeyVersionRequest) (*CryptoKeyVersion, error) + // Encrypts data, so that it can only be recovered by a call to [Decrypt][google.cloud.kms.v1.KeyManagementService.Decrypt]. + Encrypt(context.Context, *EncryptRequest) (*EncryptResponse, error) + // Decrypts data that was protected by [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt]. + Decrypt(context.Context, *DecryptRequest) (*DecryptResponse, error) + // Update the version of a [CryptoKey][google.cloud.kms.v1.CryptoKey] that will be used in [Encrypt][google.cloud.kms.v1.KeyManagementService.Encrypt] + UpdateCryptoKeyPrimaryVersion(context.Context, *UpdateCryptoKeyPrimaryVersionRequest) (*CryptoKey, error) + // Schedule a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] for destruction. + // + // Upon calling this method, [CryptoKeyVersion.state][google.cloud.kms.v1.CryptoKeyVersion.state] will be set to + // [DESTROY_SCHEDULED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROY_SCHEDULED] + // and [destroy_time][google.cloud.kms.v1.CryptoKeyVersion.destroy_time] will be set to a time 24 + // hours in the future, at which point the [state][google.cloud.kms.v1.CryptoKeyVersion.state] + // will be changed to + // [DESTROYED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROYED], and the key + // material will be irrevocably destroyed. + // + // Before the [destroy_time][google.cloud.kms.v1.CryptoKeyVersion.destroy_time] is reached, + // [RestoreCryptoKeyVersion][google.cloud.kms.v1.KeyManagementService.RestoreCryptoKeyVersion] may be called to reverse the process. + DestroyCryptoKeyVersion(context.Context, *DestroyCryptoKeyVersionRequest) (*CryptoKeyVersion, error) + // Restore a [CryptoKeyVersion][google.cloud.kms.v1.CryptoKeyVersion] in the + // [DESTROY_SCHEDULED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DESTROY_SCHEDULED], + // state. + // + // Upon restoration of the CryptoKeyVersion, [state][google.cloud.kms.v1.CryptoKeyVersion.state] + // will be set to [DISABLED][google.cloud.kms.v1.CryptoKeyVersion.CryptoKeyVersionState.DISABLED], + // and [destroy_time][google.cloud.kms.v1.CryptoKeyVersion.destroy_time] will be cleared. + RestoreCryptoKeyVersion(context.Context, *RestoreCryptoKeyVersionRequest) (*CryptoKeyVersion, error) +} + +func RegisterKeyManagementServiceServer(s *grpc.Server, srv KeyManagementServiceServer) { + s.RegisterService(&_KeyManagementService_serviceDesc, srv) +} + +func _KeyManagementService_ListKeyRings_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListKeyRingsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).ListKeyRings(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/ListKeyRings", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).ListKeyRings(ctx, req.(*ListKeyRingsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_ListCryptoKeys_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListCryptoKeysRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).ListCryptoKeys(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/ListCryptoKeys", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).ListCryptoKeys(ctx, req.(*ListCryptoKeysRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_ListCryptoKeyVersions_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListCryptoKeyVersionsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).ListCryptoKeyVersions(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/ListCryptoKeyVersions", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).ListCryptoKeyVersions(ctx, req.(*ListCryptoKeyVersionsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_GetKeyRing_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetKeyRingRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).GetKeyRing(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/GetKeyRing", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).GetKeyRing(ctx, req.(*GetKeyRingRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_GetCryptoKey_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetCryptoKeyRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).GetCryptoKey(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/GetCryptoKey", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).GetCryptoKey(ctx, req.(*GetCryptoKeyRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_GetCryptoKeyVersion_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetCryptoKeyVersionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).GetCryptoKeyVersion(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/GetCryptoKeyVersion", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).GetCryptoKeyVersion(ctx, req.(*GetCryptoKeyVersionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_CreateKeyRing_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(CreateKeyRingRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).CreateKeyRing(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/CreateKeyRing", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).CreateKeyRing(ctx, req.(*CreateKeyRingRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_CreateCryptoKey_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(CreateCryptoKeyRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).CreateCryptoKey(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/CreateCryptoKey", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).CreateCryptoKey(ctx, req.(*CreateCryptoKeyRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_CreateCryptoKeyVersion_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(CreateCryptoKeyVersionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).CreateCryptoKeyVersion(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/CreateCryptoKeyVersion", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).CreateCryptoKeyVersion(ctx, req.(*CreateCryptoKeyVersionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_UpdateCryptoKey_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(UpdateCryptoKeyRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).UpdateCryptoKey(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/UpdateCryptoKey", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).UpdateCryptoKey(ctx, req.(*UpdateCryptoKeyRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_UpdateCryptoKeyVersion_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(UpdateCryptoKeyVersionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).UpdateCryptoKeyVersion(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/UpdateCryptoKeyVersion", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).UpdateCryptoKeyVersion(ctx, req.(*UpdateCryptoKeyVersionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_Encrypt_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(EncryptRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).Encrypt(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/Encrypt", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).Encrypt(ctx, req.(*EncryptRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_Decrypt_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(DecryptRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).Decrypt(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/Decrypt", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).Decrypt(ctx, req.(*DecryptRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_UpdateCryptoKeyPrimaryVersion_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(UpdateCryptoKeyPrimaryVersionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).UpdateCryptoKeyPrimaryVersion(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/UpdateCryptoKeyPrimaryVersion", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).UpdateCryptoKeyPrimaryVersion(ctx, req.(*UpdateCryptoKeyPrimaryVersionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_DestroyCryptoKeyVersion_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(DestroyCryptoKeyVersionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).DestroyCryptoKeyVersion(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/DestroyCryptoKeyVersion", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).DestroyCryptoKeyVersion(ctx, req.(*DestroyCryptoKeyVersionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _KeyManagementService_RestoreCryptoKeyVersion_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(RestoreCryptoKeyVersionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(KeyManagementServiceServer).RestoreCryptoKeyVersion(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.kms.v1.KeyManagementService/RestoreCryptoKeyVersion", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(KeyManagementServiceServer).RestoreCryptoKeyVersion(ctx, req.(*RestoreCryptoKeyVersionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _KeyManagementService_serviceDesc = grpc.ServiceDesc{ + ServiceName: "google.cloud.kms.v1.KeyManagementService", + HandlerType: (*KeyManagementServiceServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "ListKeyRings", + Handler: _KeyManagementService_ListKeyRings_Handler, + }, + { + MethodName: "ListCryptoKeys", + Handler: _KeyManagementService_ListCryptoKeys_Handler, + }, + { + MethodName: "ListCryptoKeyVersions", + Handler: _KeyManagementService_ListCryptoKeyVersions_Handler, + }, + { + MethodName: "GetKeyRing", + Handler: _KeyManagementService_GetKeyRing_Handler, + }, + { + MethodName: "GetCryptoKey", + Handler: _KeyManagementService_GetCryptoKey_Handler, + }, + { + MethodName: "GetCryptoKeyVersion", + Handler: _KeyManagementService_GetCryptoKeyVersion_Handler, + }, + { + MethodName: "CreateKeyRing", + Handler: _KeyManagementService_CreateKeyRing_Handler, + }, + { + MethodName: "CreateCryptoKey", + Handler: _KeyManagementService_CreateCryptoKey_Handler, + }, + { + MethodName: "CreateCryptoKeyVersion", + Handler: _KeyManagementService_CreateCryptoKeyVersion_Handler, + }, + { + MethodName: "UpdateCryptoKey", + Handler: _KeyManagementService_UpdateCryptoKey_Handler, + }, + { + MethodName: "UpdateCryptoKeyVersion", + Handler: _KeyManagementService_UpdateCryptoKeyVersion_Handler, + }, + { + MethodName: "Encrypt", + Handler: _KeyManagementService_Encrypt_Handler, + }, + { + MethodName: "Decrypt", + Handler: _KeyManagementService_Decrypt_Handler, + }, + { + MethodName: "UpdateCryptoKeyPrimaryVersion", + Handler: _KeyManagementService_UpdateCryptoKeyPrimaryVersion_Handler, + }, + { + MethodName: "DestroyCryptoKeyVersion", + Handler: _KeyManagementService_DestroyCryptoKeyVersion_Handler, + }, + { + MethodName: "RestoreCryptoKeyVersion", + Handler: _KeyManagementService_RestoreCryptoKeyVersion_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "google/cloud/kms/v1/service.proto", +} + +func init() { + proto.RegisterFile("google/cloud/kms/v1/service.proto", fileDescriptor_service_d4f59c45586ec2a6) +} + +var fileDescriptor_service_d4f59c45586ec2a6 = []byte{ + // 1362 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xc4, 0x98, 0xcd, 0x6f, 0xdc, 0xc4, + 0x1b, 0xc7, 0x35, 0xe9, 0xef, 0xd7, 0x66, 0x9f, 0xa4, 0x0d, 0x9d, 0xb4, 0x69, 0xb4, 0x4d, 0x56, + 0xc1, 0x2d, 0x25, 0xdd, 0x22, 0x9b, 0xdd, 0x56, 0x40, 0xb7, 0x2a, 0xa8, 0x69, 0xda, 0xa8, 0xda, + 0x06, 0x22, 0x87, 0x54, 0x50, 0x45, 0xda, 0x4e, 0xed, 0xe9, 0xd6, 0xec, 0xfa, 0x05, 0x7b, 0x36, + 0x74, 0x0b, 0xbd, 0x70, 0x80, 0x1e, 0x7a, 0x2b, 0x87, 0x22, 0x2e, 0x08, 0x6e, 0x08, 0x0e, 0xdc, + 0xb8, 0x70, 0xe3, 0xc4, 0x15, 0x55, 0x1c, 0x90, 0x38, 0xf1, 0x0f, 0x70, 0xe3, 0x88, 0x3c, 0x1e, + 0x7b, 0xd7, 0x5e, 0x7b, 0xdf, 0xba, 0x15, 0xb7, 0xec, 0xf8, 0x99, 0x99, 0xcf, 0x77, 0x9e, 0xc7, + 0xf3, 0x7c, 0x1d, 0x78, 0xb1, 0x6e, 0xdb, 0xf5, 0x26, 0x55, 0xb4, 0xa6, 0xdd, 0xd2, 0x95, 0x86, + 0xe9, 0x29, 0x7b, 0x25, 0xc5, 0xa3, 0xee, 0x9e, 0xa1, 0x51, 0xd9, 0x71, 0x6d, 0x66, 0xe3, 0xf9, + 0x20, 0x44, 0xe6, 0x21, 0x72, 0xc3, 0xf4, 0xe4, 0xbd, 0x52, 0x7e, 0x49, 0xcc, 0x23, 0x8e, 0xa1, + 0x10, 0xcb, 0xb2, 0x19, 0x61, 0x86, 0x6d, 0x79, 0xc1, 0x94, 0xfc, 0x89, 0xb4, 0x55, 0x5d, 0xea, + 0xd9, 0x2d, 0x57, 0xa3, 0x61, 0xd0, 0x8a, 0x08, 0xe2, 0xbf, 0x6e, 0xb7, 0xee, 0x28, 0x77, 0x0c, + 0xda, 0xd4, 0x6b, 0x26, 0xf1, 0x1a, 0x22, 0x62, 0x29, 0x19, 0xe1, 0x31, 0xb7, 0xa5, 0x31, 0xf1, + 0xb4, 0x90, 0x7c, 0xfa, 0x91, 0x4b, 0x1c, 0x87, 0xba, 0x62, 0x7d, 0xc9, 0x80, 0xf9, 0xeb, 0x86, + 0xc7, 0xaa, 0xb4, 0xad, 0x1a, 0x56, 0xdd, 0x53, 0xe9, 0x87, 0x2d, 0xea, 0x31, 0xbc, 0x00, 0xfb, + 0x1d, 0xe2, 0x52, 0x8b, 0x2d, 0xa2, 0x15, 0xb4, 0x9a, 0x53, 0xc5, 0x2f, 0x7c, 0x1c, 0x72, 0x0e, + 0xa9, 0xd3, 0x9a, 0x67, 0xdc, 0xa7, 0x8b, 0x53, 0x2b, 0x68, 0xf5, 0xff, 0xea, 0xb4, 0x3f, 0xb0, + 0x6d, 0xdc, 0xa7, 0x78, 0x19, 0x80, 0x3f, 0x64, 0x76, 0x83, 0x5a, 0x8b, 0xfb, 0xf8, 0x44, 0x1e, + 0xfe, 0xae, 0x3f, 0x20, 0x35, 0xe0, 0xa8, 0xbf, 0xd5, 0x65, 0xb7, 0xed, 0x30, 0xbb, 0x4a, 0xdb, + 0xcf, 0x75, 0x33, 0x17, 0x96, 0x62, 0x9b, 0xdd, 0xa0, 0xae, 0xe7, 0x9f, 0xfd, 0xf3, 0xdc, 0xf3, + 0x09, 0x82, 0x23, 0xf1, 0xc3, 0xf4, 0x1c, 0xdb, 0xf2, 0x28, 0x3e, 0x0f, 0xb9, 0x06, 0x6d, 0xd7, + 0x5c, 0x7f, 0x70, 0x11, 0xad, 0xec, 0x5b, 0x9d, 0x29, 0x2f, 0xc9, 0x29, 0x05, 0x23, 0x8b, 0x99, + 0xea, 0x74, 0x43, 0x2c, 0x81, 0x4f, 0xc1, 0x9c, 0x45, 0xef, 0xb1, 0x5a, 0xd7, 0xbe, 0x53, 0x7c, + 0xdf, 0x83, 0xfe, 0xf0, 0x56, 0xb8, 0xb7, 0x8f, 0xc6, 0x6c, 0x46, 0x9a, 0x01, 0xf8, 0x3e, 0x0e, + 0x9e, 0xe3, 0x23, 0x3e, 0xb9, 0xf4, 0x35, 0x82, 0x85, 0xe4, 0xe1, 0x0b, 0xb8, 0xb7, 0x60, 0x46, + 0xe3, 0xa3, 0xb5, 0x06, 0x6d, 0x87, 0x78, 0x85, 0x54, 0xbc, 0x68, 0xb6, 0x0a, 0x5a, 0xb4, 0xd0, + 0xa4, 0x10, 0x7f, 0x46, 0xb0, 0x9c, 0x91, 0x32, 0x41, 0xba, 0x03, 0xf3, 0x1d, 0xd2, 0xda, 0x9e, + 0x78, 0x2c, 0x88, 0x5f, 0xea, 0x4f, 0x2c, 0x16, 0x53, 0x0f, 0x6b, 0xc9, 0xe5, 0x27, 0xc5, 0xff, + 0x32, 0x1c, 0xde, 0xa0, 0x61, 0xee, 0xc3, 0x32, 0xc3, 0xf0, 0x3f, 0x8b, 0x98, 0x54, 0x14, 0x19, + 0xff, 0x5b, 0x3a, 0x0d, 0xf3, 0x1b, 0xb4, 0x23, 0xb3, 0x5f, 0xe8, 0xab, 0x90, 0xef, 0x0e, 0x0d, + 0x45, 0xf4, 0x99, 0xf1, 0x39, 0x82, 0x23, 0x97, 0x5d, 0x4a, 0x18, 0x4d, 0x90, 0x64, 0x15, 0x7c, + 0x01, 0x66, 0xc2, 0xda, 0xac, 0x19, 0xba, 0x50, 0x9e, 0x13, 0xf5, 0x77, 0x4d, 0xc7, 0xaf, 0xc3, + 0x74, 0xf8, 0x9c, 0x6b, 0x1e, 0x54, 0xba, 0x07, 0xc4, 0x54, 0xe9, 0x31, 0x82, 0x85, 0x80, 0xa4, + 0x47, 0x6a, 0x16, 0x8b, 0x04, 0x07, 0xbb, 0x12, 0x1c, 0xd1, 0xcc, 0x44, 0x39, 0xbb, 0xa6, 0xe3, + 0x8b, 0x00, 0x9d, 0x18, 0x41, 0x34, 0xa8, 0x5a, 0x73, 0xd1, 0x02, 0xd2, 0x23, 0x04, 0xcb, 0x09, + 0xaa, 0xc4, 0xa9, 0x66, 0xc1, 0x6d, 0x03, 0xee, 0xad, 0x3e, 0x4e, 0x38, 0x74, 0xf1, 0xbd, 0x90, + 0x2c, 0x3e, 0xe9, 0x0b, 0x04, 0x0b, 0x3b, 0x8e, 0x9e, 0x76, 0x48, 0x71, 0xa1, 0x68, 0x44, 0xa1, + 0xf8, 0x02, 0xcc, 0xb4, 0xf8, 0xc2, 0xbc, 0x57, 0x08, 0xce, 0x7c, 0x38, 0x3f, 0x6c, 0x07, 0xf2, + 0x55, 0xbf, 0x9d, 0x6c, 0x12, 0xaf, 0xa1, 0x42, 0x10, 0xee, 0xff, 0x2d, 0xfd, 0x88, 0x60, 0x39, + 0x81, 0x95, 0x38, 0xa5, 0xf4, 0xd3, 0x40, 0xcf, 0x74, 0x1a, 0xcf, 0xc6, 0xfc, 0x19, 0x82, 0x43, + 0x57, 0x2c, 0xbe, 0x66, 0x9f, 0x17, 0x04, 0x2f, 0x41, 0xce, 0x69, 0x12, 0xc3, 0x62, 0xf4, 0x1e, + 0xe3, 0x3b, 0xcc, 0xaa, 0x9d, 0x01, 0xbc, 0x06, 0xcb, 0x44, 0xd7, 0x0d, 0xbf, 0x4d, 0x93, 0x66, + 0x8d, 0xb4, 0xd8, 0x5d, 0x6a, 0x31, 0x43, 0x23, 0x8c, 0xea, 0x35, 0x9d, 0x30, 0xc2, 0x0b, 0x6e, + 0x56, 0x3d, 0xde, 0x09, 0xba, 0xd4, 0x1d, 0xb3, 0x4e, 0x18, 0x91, 0x1e, 0x22, 0x38, 0xb4, 0x4e, + 0x07, 0x82, 0x14, 0x00, 0x34, 0xc3, 0xb9, 0x4b, 0xdd, 0x2e, 0x92, 0xae, 0x91, 0x89, 0xa0, 0x28, + 0x30, 0x17, 0x91, 0x88, 0x4b, 0x34, 0xa6, 0x1f, 0x25, 0xf4, 0x4b, 0x57, 0x60, 0x2e, 0x3a, 0x43, + 0x31, 0x61, 0x0c, 0x76, 0xc9, 0x84, 0x93, 0x89, 0xf2, 0xd9, 0x72, 0x0d, 0x93, 0xb8, 0x43, 0xdc, + 0x60, 0xb8, 0x04, 0x47, 0x7b, 0x2b, 0xab, 0x73, 0x19, 0xe0, 0x64, 0xd5, 0x5c, 0xd3, 0xa5, 0x73, + 0x50, 0x58, 0xa7, 0x1e, 0x73, 0xed, 0xf6, 0x28, 0x57, 0xe5, 0x39, 0x28, 0xa8, 0xd4, 0x63, 0xb6, + 0x4b, 0x47, 0x98, 0x55, 0xfe, 0xe3, 0x18, 0x1c, 0xa9, 0xd2, 0xf6, 0x26, 0xb1, 0x48, 0x9d, 0x9a, + 0xd4, 0x62, 0xdb, 0x81, 0x0f, 0xc4, 0x5f, 0x22, 0x98, 0xed, 0xee, 0xfe, 0x78, 0x35, 0xf5, 0x35, + 0x48, 0x71, 0x5b, 0xf9, 0xd3, 0x43, 0x44, 0x06, 0xd9, 0x90, 0xce, 0x7d, 0xfa, 0xdb, 0x5f, 0x8f, + 0xa7, 0x64, 0xfc, 0x8a, 0x6f, 0x16, 0x3f, 0x0e, 0xae, 0xa6, 0x8b, 0x8e, 0x6b, 0x7f, 0x40, 0x35, + 0xe6, 0x29, 0x45, 0xa5, 0x69, 0x6b, 0x81, 0xc9, 0x54, 0x8a, 0x0f, 0x94, 0xc8, 0x45, 0x7c, 0x8f, + 0xe0, 0x50, 0xbc, 0xfd, 0xe3, 0x62, 0xe6, 0x9e, 0x3d, 0x06, 0x2d, 0x7f, 0x66, 0xa8, 0x58, 0x41, + 0x78, 0x89, 0x13, 0x5e, 0xc0, 0xe7, 0x07, 0x13, 0x46, 0x80, 0x3e, 0x6c, 0x97, 0xa3, 0x78, 0x8a, + 0x12, 0x56, 0x31, 0xea, 0xd5, 0xa5, 0xc1, 0x24, 0x09, 0xa7, 0x97, 0x2f, 0x8f, 0x32, 0x45, 0x68, + 0xd8, 0xe1, 0x1a, 0xde, 0xc1, 0x9b, 0x23, 0x69, 0xe8, 0x92, 0x10, 0x13, 0x14, 0xd1, 0x3f, 0x44, + 0x00, 0x1d, 0x8f, 0x80, 0x4f, 0xa5, 0x92, 0xf5, 0x98, 0x88, 0x7c, 0xdf, 0x86, 0x9b, 0xa8, 0x08, + 0xbf, 0x46, 0x87, 0x38, 0x6d, 0xfc, 0x15, 0x82, 0xd9, 0x6e, 0x6b, 0x91, 0x51, 0xad, 0x29, 0x46, + 0x25, 0x3f, 0xa0, 0x09, 0x25, 0x0a, 0x60, 0x38, 0xa0, 0xf8, 0xd1, 0xe1, 0x5f, 0x50, 0xdc, 0x23, + 0x85, 0x0d, 0x42, 0x19, 0x08, 0x19, 0x7f, 0x83, 0xf3, 0xc3, 0xb5, 0xa2, 0x44, 0xbe, 0xc7, 0x40, + 0xee, 0x4d, 0xb6, 0x2f, 0xe3, 0x09, 0x82, 0x83, 0x31, 0x33, 0x86, 0x4f, 0x67, 0xf0, 0xf4, 0x1a, + 0xb6, 0x01, 0x59, 0x7f, 0x93, 0x13, 0xbf, 0x21, 0x8d, 0x74, 0x0f, 0x54, 0x22, 0x2b, 0x87, 0x7f, + 0x40, 0x30, 0x97, 0xf0, 0x41, 0xf8, 0x4c, 0x1f, 0xb8, 0x91, 0xab, 0x60, 0x93, 0x03, 0x6e, 0x48, + 0xe3, 0x5f, 0x03, 0x95, 0x2e, 0xff, 0x83, 0xff, 0xec, 0x75, 0x93, 0x61, 0x51, 0x94, 0x87, 0xc1, + 0x1e, 0xaf, 0x2e, 0x0c, 0x2e, 0x42, 0x93, 0x26, 0x7b, 0x0f, 0x54, 0x52, 0xac, 0x13, 0xfe, 0x09, + 0xc1, 0x5c, 0xa2, 0x67, 0x66, 0x24, 0x24, 0xdd, 0x2f, 0x0e, 0x4c, 0xc8, 0x7b, 0x5c, 0x8b, 0x5a, + 0x5e, 0xe7, 0x5a, 0x3a, 0x04, 0xf2, 0x38, 0x6f, 0x68, 0x2c, 0x37, 0x7f, 0xf7, 0x9a, 0xd8, 0xfe, + 0xb9, 0xe9, 0x6b, 0x2d, 0x87, 0xcd, 0xcd, 0x27, 0x5c, 0xcf, 0x5e, 0xf9, 0x56, 0x42, 0x4f, 0x78, + 0xa2, 0xf2, 0xa4, 0x5e, 0xe3, 0xd4, 0x74, 0x7d, 0x8b, 0xe0, 0x80, 0x70, 0x4a, 0xf8, 0x44, 0x2a, + 0x70, 0xdc, 0x8b, 0xe6, 0x4f, 0xf6, 0x0f, 0x12, 0x8d, 0x27, 0x7c, 0x6b, 0xd6, 0xc6, 0xbd, 0x88, + 0x8a, 0x0f, 0x2a, 0x34, 0x58, 0xb3, 0x82, 0x8a, 0xf8, 0x1b, 0x04, 0x07, 0x84, 0x01, 0xcc, 0xa0, + 0x8c, 0x1b, 0xd5, 0x0c, 0xca, 0x84, 0x87, 0x94, 0xae, 0x73, 0xca, 0xab, 0xd2, 0xa5, 0xb1, 0x6f, + 0xf8, 0x8a, 0x4e, 0x23, 0xc8, 0xa7, 0xbd, 0x1f, 0x1b, 0x71, 0xb7, 0x88, 0xcf, 0x0f, 0x53, 0x45, + 0xa9, 0x0e, 0x73, 0xe0, 0x5b, 0xf1, 0x3e, 0x97, 0xb2, 0x2d, 0xbd, 0x3d, 0xbe, 0x94, 0xe0, 0x1b, + 0x24, 0xbe, 0xbd, 0xaf, 0xeb, 0x77, 0x04, 0xc7, 0x32, 0x6c, 0x29, 0x3e, 0x9b, 0x71, 0xce, 0xfd, + 0x4c, 0xec, 0xb0, 0x2f, 0xc6, 0x2d, 0x2e, 0xe9, 0xa6, 0xb4, 0x33, 0xd1, 0x66, 0x56, 0xd1, 0x03, + 0xb8, 0x50, 0x59, 0x86, 0x75, 0xce, 0x50, 0xd6, 0xdf, 0x68, 0xff, 0xc7, 0xca, 0xdc, 0x00, 0xae, + 0x82, 0x8a, 0x6b, 0x8f, 0x10, 0x1c, 0xd3, 0x6c, 0x33, 0x0d, 0x67, 0x6d, 0xba, 0x6a, 0x7a, 0x5b, + 0xfe, 0x37, 0xe8, 0x16, 0xba, 0xf9, 0x9a, 0x08, 0xa8, 0xdb, 0x4d, 0x62, 0xd5, 0x65, 0xdb, 0xad, + 0x2b, 0x75, 0x6a, 0xf1, 0x2f, 0x54, 0x25, 0x78, 0x44, 0x1c, 0xc3, 0x8b, 0xfd, 0x6b, 0xf7, 0x42, + 0xc3, 0xf4, 0xfe, 0x41, 0xe8, 0xbb, 0xa9, 0xf9, 0x8d, 0x60, 0xee, 0x65, 0xbe, 0x78, 0xd5, 0xf4, + 0xe4, 0x1b, 0xa5, 0x5f, 0xc3, 0xd1, 0x5d, 0x3e, 0xba, 0x5b, 0x35, 0xbd, 0xdd, 0x1b, 0xa5, 0xdb, + 0xfb, 0xf9, 0x8a, 0x67, 0xff, 0x0d, 0x00, 0x00, 0xff, 0xff, 0x63, 0xc5, 0xfd, 0xff, 0x81, 0x16, + 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/language/v1/language_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/language/v1/language_service.pb.go index 8276bc57b..ac110e7c3 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/language/v1/language_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/language/v1/language_service.pb.go @@ -1086,7 +1086,7 @@ func (EntityMention_Type) EnumDescriptor() ([]byte, []int) { type Document struct { // Required. If the type is not set or is `TYPE_UNSPECIFIED`, // returns an `INVALID_ARGUMENT` error. - Type Document_Type `protobuf:"varint,1,opt,name=type,enum=google.cloud.language.v1.Document_Type" json:"type,omitempty"` + Type Document_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.language.v1.Document_Type" json:"type,omitempty"` // The source of the document: a string containing the content or a // Google Cloud Storage URI. // @@ -1102,7 +1102,7 @@ type Document struct { // If the language (either specified by the caller or automatically detected) // is not supported by the called API method, an `INVALID_ARGUMENT` error // is returned. - Language string `protobuf:"bytes,4,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,4,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1137,10 +1137,10 @@ type isDocument_Source interface { } type Document_Content struct { - Content string `protobuf:"bytes,2,opt,name=content,oneof"` + Content string `protobuf:"bytes,2,opt,name=content,proto3,oneof"` } type Document_GcsContentUri struct { - GcsContentUri string `protobuf:"bytes,3,opt,name=gcs_content_uri,json=gcsContentUri,oneof"` + GcsContentUri string `protobuf:"bytes,3,opt,name=gcs_content_uri,json=gcsContentUri,proto3,oneof"` } func (*Document_Content) isDocument_Source() {} @@ -1250,11 +1250,11 @@ func _Document_OneofSizer(msg proto.Message) (n int) { // Represents a sentence in the input document. type Sentence struct { // The sentence text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // For calls to [AnalyzeSentiment][] or if // [AnnotateTextRequest.Features.extract_document_sentiment][google.cloud.language.v1.AnnotateTextRequest.Features.extract_document_sentiment] is set to // true, this field will contain the sentiment for the sentence. - Sentiment *Sentiment `protobuf:"bytes,2,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,2,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1303,29 +1303,29 @@ func (m *Sentence) GetSentiment() *Sentiment { // as salience and mentions, with entities. type Entity struct { // The representative name for the entity. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The entity type. - Type Entity_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.language.v1.Entity_Type" json:"type,omitempty"` + Type Entity_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.language.v1.Entity_Type" json:"type,omitempty"` // Metadata associated with the entity. // // Currently, Wikipedia URLs and Knowledge Graph MIDs are provided, if // available. The associated keys are "wikipedia_url" and "mid", respectively. - Metadata map[string]string `protobuf:"bytes,3,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,3,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The salience score associated with the entity in the [0, 1.0] range. // // The salience score for an entity provides information about the // importance or centrality of that entity to the entire document text. // Scores closer to 0 are less salient, while scores closer to 1.0 are highly // salient. - Salience float32 `protobuf:"fixed32,4,opt,name=salience" json:"salience,omitempty"` + Salience float32 `protobuf:"fixed32,4,opt,name=salience,proto3" json:"salience,omitempty"` // The mentions of this entity in the input document. The API currently // supports proper noun mentions. - Mentions []*EntityMention `protobuf:"bytes,5,rep,name=mentions" json:"mentions,omitempty"` + Mentions []*EntityMention `protobuf:"bytes,5,rep,name=mentions,proto3" json:"mentions,omitempty"` // For calls to [AnalyzeEntitySentiment][] or if // [AnnotateTextRequest.Features.extract_entity_sentiment][google.cloud.language.v1.AnnotateTextRequest.Features.extract_entity_sentiment] is set to // true, this field will contain the aggregate sentiment expressed for this // entity in the provided document. - Sentiment *Sentiment `protobuf:"bytes,6,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,6,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1400,13 +1400,13 @@ func (m *Entity) GetSentiment() *Sentiment { // Represents the smallest syntactic building block of the text. type Token struct { // The token text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Parts of speech tag for this token. - PartOfSpeech *PartOfSpeech `protobuf:"bytes,2,opt,name=part_of_speech,json=partOfSpeech" json:"part_of_speech,omitempty"` + PartOfSpeech *PartOfSpeech `protobuf:"bytes,2,opt,name=part_of_speech,json=partOfSpeech,proto3" json:"part_of_speech,omitempty"` // Dependency tree parse for this token. - DependencyEdge *DependencyEdge `protobuf:"bytes,3,opt,name=dependency_edge,json=dependencyEdge" json:"dependency_edge,omitempty"` + DependencyEdge *DependencyEdge `protobuf:"bytes,3,opt,name=dependency_edge,json=dependencyEdge,proto3" json:"dependency_edge,omitempty"` // [Lemma](https://en.wikipedia.org/wiki/Lemma_%28morphology%29) of the token. - Lemma string `protobuf:"bytes,4,opt,name=lemma" json:"lemma,omitempty"` + Lemma string `protobuf:"bytes,4,opt,name=lemma,proto3" json:"lemma,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1470,10 +1470,10 @@ type Sentiment struct { // A non-negative number in the [0, +inf) range, which represents // the absolute magnitude of sentiment regardless of score (positive or // negative). - Magnitude float32 `protobuf:"fixed32,2,opt,name=magnitude" json:"magnitude,omitempty"` + Magnitude float32 `protobuf:"fixed32,2,opt,name=magnitude,proto3" json:"magnitude,omitempty"` // Sentiment score between -1.0 (negative sentiment) and 1.0 // (positive sentiment). - Score float32 `protobuf:"fixed32,3,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,3,opt,name=score,proto3" json:"score,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1522,29 +1522,29 @@ func (m *Sentiment) GetScore() float32 { // http://www.lrec-conf.org/proceedings/lrec2012/pdf/274_Paper.pdf type PartOfSpeech struct { // The part of speech tag. - Tag PartOfSpeech_Tag `protobuf:"varint,1,opt,name=tag,enum=google.cloud.language.v1.PartOfSpeech_Tag" json:"tag,omitempty"` + Tag PartOfSpeech_Tag `protobuf:"varint,1,opt,name=tag,proto3,enum=google.cloud.language.v1.PartOfSpeech_Tag" json:"tag,omitempty"` // The grammatical aspect. - Aspect PartOfSpeech_Aspect `protobuf:"varint,2,opt,name=aspect,enum=google.cloud.language.v1.PartOfSpeech_Aspect" json:"aspect,omitempty"` + Aspect PartOfSpeech_Aspect `protobuf:"varint,2,opt,name=aspect,proto3,enum=google.cloud.language.v1.PartOfSpeech_Aspect" json:"aspect,omitempty"` // The grammatical case. - Case PartOfSpeech_Case `protobuf:"varint,3,opt,name=case,enum=google.cloud.language.v1.PartOfSpeech_Case" json:"case,omitempty"` + Case PartOfSpeech_Case `protobuf:"varint,3,opt,name=case,proto3,enum=google.cloud.language.v1.PartOfSpeech_Case" json:"case,omitempty"` // The grammatical form. - Form PartOfSpeech_Form `protobuf:"varint,4,opt,name=form,enum=google.cloud.language.v1.PartOfSpeech_Form" json:"form,omitempty"` + Form PartOfSpeech_Form `protobuf:"varint,4,opt,name=form,proto3,enum=google.cloud.language.v1.PartOfSpeech_Form" json:"form,omitempty"` // The grammatical gender. - Gender PartOfSpeech_Gender `protobuf:"varint,5,opt,name=gender,enum=google.cloud.language.v1.PartOfSpeech_Gender" json:"gender,omitempty"` + Gender PartOfSpeech_Gender `protobuf:"varint,5,opt,name=gender,proto3,enum=google.cloud.language.v1.PartOfSpeech_Gender" json:"gender,omitempty"` // The grammatical mood. - Mood PartOfSpeech_Mood `protobuf:"varint,6,opt,name=mood,enum=google.cloud.language.v1.PartOfSpeech_Mood" json:"mood,omitempty"` + Mood PartOfSpeech_Mood `protobuf:"varint,6,opt,name=mood,proto3,enum=google.cloud.language.v1.PartOfSpeech_Mood" json:"mood,omitempty"` // The grammatical number. - Number PartOfSpeech_Number `protobuf:"varint,7,opt,name=number,enum=google.cloud.language.v1.PartOfSpeech_Number" json:"number,omitempty"` + Number PartOfSpeech_Number `protobuf:"varint,7,opt,name=number,proto3,enum=google.cloud.language.v1.PartOfSpeech_Number" json:"number,omitempty"` // The grammatical person. - Person PartOfSpeech_Person `protobuf:"varint,8,opt,name=person,enum=google.cloud.language.v1.PartOfSpeech_Person" json:"person,omitempty"` + Person PartOfSpeech_Person `protobuf:"varint,8,opt,name=person,proto3,enum=google.cloud.language.v1.PartOfSpeech_Person" json:"person,omitempty"` // The grammatical properness. - Proper PartOfSpeech_Proper `protobuf:"varint,9,opt,name=proper,enum=google.cloud.language.v1.PartOfSpeech_Proper" json:"proper,omitempty"` + Proper PartOfSpeech_Proper `protobuf:"varint,9,opt,name=proper,proto3,enum=google.cloud.language.v1.PartOfSpeech_Proper" json:"proper,omitempty"` // The grammatical reciprocity. - Reciprocity PartOfSpeech_Reciprocity `protobuf:"varint,10,opt,name=reciprocity,enum=google.cloud.language.v1.PartOfSpeech_Reciprocity" json:"reciprocity,omitempty"` + Reciprocity PartOfSpeech_Reciprocity `protobuf:"varint,10,opt,name=reciprocity,proto3,enum=google.cloud.language.v1.PartOfSpeech_Reciprocity" json:"reciprocity,omitempty"` // The grammatical tense. - Tense PartOfSpeech_Tense `protobuf:"varint,11,opt,name=tense,enum=google.cloud.language.v1.PartOfSpeech_Tense" json:"tense,omitempty"` + Tense PartOfSpeech_Tense `protobuf:"varint,11,opt,name=tense,proto3,enum=google.cloud.language.v1.PartOfSpeech_Tense" json:"tense,omitempty"` // The grammatical voice. - Voice PartOfSpeech_Voice `protobuf:"varint,12,opt,name=voice,enum=google.cloud.language.v1.PartOfSpeech_Voice" json:"voice,omitempty"` + Voice PartOfSpeech_Voice `protobuf:"varint,12,opt,name=voice,proto3,enum=google.cloud.language.v1.PartOfSpeech_Voice" json:"voice,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1667,9 +1667,9 @@ type DependencyEdge struct { // The index is the position of the token in the array of tokens returned // by the API method. If this token is a root token, then the // `head_token_index` is its own index. - HeadTokenIndex int32 `protobuf:"varint,1,opt,name=head_token_index,json=headTokenIndex" json:"head_token_index,omitempty"` + HeadTokenIndex int32 `protobuf:"varint,1,opt,name=head_token_index,json=headTokenIndex,proto3" json:"head_token_index,omitempty"` // The parse label for the token. - Label DependencyEdge_Label `protobuf:"varint,2,opt,name=label,enum=google.cloud.language.v1.DependencyEdge_Label" json:"label,omitempty"` + Label DependencyEdge_Label `protobuf:"varint,2,opt,name=label,proto3,enum=google.cloud.language.v1.DependencyEdge_Label" json:"label,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1717,14 +1717,14 @@ func (m *DependencyEdge) GetLabel() DependencyEdge_Label { // mentions are supported. type EntityMention struct { // The mention text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // The type of the entity mention. - Type EntityMention_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.language.v1.EntityMention_Type" json:"type,omitempty"` + Type EntityMention_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.language.v1.EntityMention_Type" json:"type,omitempty"` // For calls to [AnalyzeEntitySentiment][] or if // [AnnotateTextRequest.Features.extract_entity_sentiment][google.cloud.language.v1.AnnotateTextRequest.Features.extract_entity_sentiment] is set to // true, this field will contain the sentiment expressed for this mention of // the entity in the provided document. - Sentiment *Sentiment `protobuf:"bytes,3,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,3,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1778,10 +1778,10 @@ func (m *EntityMention) GetSentiment() *Sentiment { // Represents an output piece of text. type TextSpan struct { // The content of the output text. - Content string `protobuf:"bytes,1,opt,name=content" json:"content,omitempty"` + Content string `protobuf:"bytes,1,opt,name=content,proto3" json:"content,omitempty"` // The API calculates the beginning offset of the content in the original // document according to the [EncodingType][google.cloud.language.v1.EncodingType] specified in the API request. - BeginOffset int32 `protobuf:"varint,2,opt,name=begin_offset,json=beginOffset" json:"begin_offset,omitempty"` + BeginOffset int32 `protobuf:"varint,2,opt,name=begin_offset,json=beginOffset,proto3" json:"begin_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1828,10 +1828,10 @@ func (m *TextSpan) GetBeginOffset() int32 { // Represents a category returned from the text classifier. type ClassificationCategory struct { // The name of the category representing the document. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The classifier's confidence of the category. Number represents how certain // the classifier is that this category represents the given text. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1878,9 +1878,9 @@ func (m *ClassificationCategory) GetConfidence() float32 { // The sentiment analysis request message. type AnalyzeSentimentRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate sentence offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1927,13 +1927,13 @@ func (m *AnalyzeSentimentRequest) GetEncodingType() EncodingType { // The sentiment analysis response message. type AnalyzeSentimentResponse struct { // The overall sentiment of the input document. - DocumentSentiment *Sentiment `protobuf:"bytes,1,opt,name=document_sentiment,json=documentSentiment" json:"document_sentiment,omitempty"` + DocumentSentiment *Sentiment `protobuf:"bytes,1,opt,name=document_sentiment,json=documentSentiment,proto3" json:"document_sentiment,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` // The sentiment for all the sentences in the document. - Sentences []*Sentence `protobuf:"bytes,3,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,3,rep,name=sentences,proto3" json:"sentences,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1987,9 +1987,9 @@ func (m *AnalyzeSentimentResponse) GetSentences() []*Sentence { // The entity-level sentiment analysis request message. type AnalyzeEntitySentimentRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2036,11 +2036,11 @@ func (m *AnalyzeEntitySentimentRequest) GetEncodingType() EncodingType { // The entity-level sentiment analysis response message. type AnalyzeEntitySentimentResponse struct { // The recognized entities in the input document with associated sentiments. - Entities []*Entity `protobuf:"bytes,1,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,1,rep,name=entities,proto3" json:"entities,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2087,9 +2087,9 @@ func (m *AnalyzeEntitySentimentResponse) GetLanguage() string { // The entity analysis request message. type AnalyzeEntitiesRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2136,11 +2136,11 @@ func (m *AnalyzeEntitiesRequest) GetEncodingType() EncodingType { // The entity analysis response message. type AnalyzeEntitiesResponse struct { // The recognized entities in the input document. - Entities []*Entity `protobuf:"bytes,1,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,1,rep,name=entities,proto3" json:"entities,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2187,9 +2187,9 @@ func (m *AnalyzeEntitiesResponse) GetLanguage() string { // The syntax analysis request message. type AnalyzeSyntaxRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2236,13 +2236,13 @@ func (m *AnalyzeSyntaxRequest) GetEncodingType() EncodingType { // The syntax analysis response message. type AnalyzeSyntaxResponse struct { // Sentences in the input document. - Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences,proto3" json:"sentences,omitempty"` // Tokens, along with their syntactic information, in the input document. - Tokens []*Token `protobuf:"bytes,2,rep,name=tokens" json:"tokens,omitempty"` + Tokens []*Token `protobuf:"bytes,2,rep,name=tokens,proto3" json:"tokens,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1.Document.language] field for more details. - Language string `protobuf:"bytes,3,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,3,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2296,7 +2296,7 @@ func (m *AnalyzeSyntaxResponse) GetLanguage() string { // The document classification request message. type ClassifyTextRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2336,7 +2336,7 @@ func (m *ClassifyTextRequest) GetDocument() *Document { // The document classification response message. type ClassifyTextResponse struct { // Categories representing the input document. - Categories []*ClassificationCategory `protobuf:"bytes,1,rep,name=categories" json:"categories,omitempty"` + Categories []*ClassificationCategory `protobuf:"bytes,1,rep,name=categories,proto3" json:"categories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2377,11 +2377,11 @@ func (m *ClassifyTextResponse) GetCategories() []*ClassificationCategory { // analysis types (sentiment, entities, and syntax) in one call. type AnnotateTextRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The enabled features. - Features *AnnotateTextRequest_Features `protobuf:"bytes,2,opt,name=features" json:"features,omitempty"` + Features *AnnotateTextRequest_Features `protobuf:"bytes,2,opt,name=features,proto3" json:"features,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,3,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,3,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2436,15 +2436,15 @@ func (m *AnnotateTextRequest) GetEncodingType() EncodingType { // Setting each one to true will enable that specific analysis for the input. type AnnotateTextRequest_Features struct { // Extract syntax information. - ExtractSyntax bool `protobuf:"varint,1,opt,name=extract_syntax,json=extractSyntax" json:"extract_syntax,omitempty"` + ExtractSyntax bool `protobuf:"varint,1,opt,name=extract_syntax,json=extractSyntax,proto3" json:"extract_syntax,omitempty"` // Extract entities. - ExtractEntities bool `protobuf:"varint,2,opt,name=extract_entities,json=extractEntities" json:"extract_entities,omitempty"` + ExtractEntities bool `protobuf:"varint,2,opt,name=extract_entities,json=extractEntities,proto3" json:"extract_entities,omitempty"` // Extract document-level sentiment. - ExtractDocumentSentiment bool `protobuf:"varint,3,opt,name=extract_document_sentiment,json=extractDocumentSentiment" json:"extract_document_sentiment,omitempty"` + ExtractDocumentSentiment bool `protobuf:"varint,3,opt,name=extract_document_sentiment,json=extractDocumentSentiment,proto3" json:"extract_document_sentiment,omitempty"` // Extract entities and their associated sentiment. - ExtractEntitySentiment bool `protobuf:"varint,4,opt,name=extract_entity_sentiment,json=extractEntitySentiment" json:"extract_entity_sentiment,omitempty"` + ExtractEntitySentiment bool `protobuf:"varint,4,opt,name=extract_entity_sentiment,json=extractEntitySentiment,proto3" json:"extract_entity_sentiment,omitempty"` // Classify the full document into categories. - ClassifyText bool `protobuf:"varint,6,opt,name=classify_text,json=classifyText" json:"classify_text,omitempty"` + ClassifyText bool `protobuf:"varint,6,opt,name=classify_text,json=classifyText,proto3" json:"classify_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2513,24 +2513,24 @@ func (m *AnnotateTextRequest_Features) GetClassifyText() bool { type AnnotateTextResponse struct { // Sentences in the input document. Populated if the user enables // [AnnotateTextRequest.Features.extract_syntax][google.cloud.language.v1.AnnotateTextRequest.Features.extract_syntax]. - Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences,proto3" json:"sentences,omitempty"` // Tokens, along with their syntactic information, in the input document. // Populated if the user enables // [AnnotateTextRequest.Features.extract_syntax][google.cloud.language.v1.AnnotateTextRequest.Features.extract_syntax]. - Tokens []*Token `protobuf:"bytes,2,rep,name=tokens" json:"tokens,omitempty"` + Tokens []*Token `protobuf:"bytes,2,rep,name=tokens,proto3" json:"tokens,omitempty"` // Entities, along with their semantic information, in the input document. // Populated if the user enables // [AnnotateTextRequest.Features.extract_entities][google.cloud.language.v1.AnnotateTextRequest.Features.extract_entities]. - Entities []*Entity `protobuf:"bytes,3,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,3,rep,name=entities,proto3" json:"entities,omitempty"` // The overall sentiment for the document. Populated if the user enables // [AnnotateTextRequest.Features.extract_document_sentiment][google.cloud.language.v1.AnnotateTextRequest.Features.extract_document_sentiment]. - DocumentSentiment *Sentiment `protobuf:"bytes,4,opt,name=document_sentiment,json=documentSentiment" json:"document_sentiment,omitempty"` + DocumentSentiment *Sentiment `protobuf:"bytes,4,opt,name=document_sentiment,json=documentSentiment,proto3" json:"document_sentiment,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1.Document.language] field for more details. - Language string `protobuf:"bytes,5,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,5,opt,name=language,proto3" json:"language,omitempty"` // Categories identified in the input document. - Categories []*ClassificationCategory `protobuf:"bytes,6,rep,name=categories" json:"categories,omitempty"` + Categories []*ClassificationCategory `protobuf:"bytes,6,rep,name=categories,proto3" json:"categories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2740,8 +2740,7 @@ func (c *languageServiceClient) AnnotateText(ctx context.Context, in *AnnotateTe return out, nil } -// Server API for LanguageService service - +// LanguageServiceServer is the server API for LanguageService service. type LanguageServiceServer interface { // Analyzes the sentiment of the provided text. AnalyzeSentiment(context.Context, *AnalyzeSentimentRequest) (*AnalyzeSentimentResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta1/language_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta1/language_service.pb.go index cdeb9ab6a..d639509eb 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta1/language_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta1/language_service.pb.go @@ -1062,7 +1062,7 @@ func (EntityMention_Type) EnumDescriptor() ([]byte, []int) { type Document struct { // Required. If the type is not set or is `TYPE_UNSPECIFIED`, // returns an `INVALID_ARGUMENT` error. - Type Document_Type `protobuf:"varint,1,opt,name=type,enum=google.cloud.language.v1beta1.Document_Type" json:"type,omitempty"` + Type Document_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.language.v1beta1.Document_Type" json:"type,omitempty"` // The source of the document: a string containing the content or a // Google Cloud Storage URI. // @@ -1078,7 +1078,7 @@ type Document struct { // If the language (either specified by the caller or automatically detected) // is not supported by the called API method, an `INVALID_ARGUMENT` error // is returned. - Language string `protobuf:"bytes,4,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,4,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1113,10 +1113,10 @@ type isDocument_Source interface { } type Document_Content struct { - Content string `protobuf:"bytes,2,opt,name=content,oneof"` + Content string `protobuf:"bytes,2,opt,name=content,proto3,oneof"` } type Document_GcsContentUri struct { - GcsContentUri string `protobuf:"bytes,3,opt,name=gcs_content_uri,json=gcsContentUri,oneof"` + GcsContentUri string `protobuf:"bytes,3,opt,name=gcs_content_uri,json=gcsContentUri,proto3,oneof"` } func (*Document_Content) isDocument_Source() {} @@ -1226,11 +1226,11 @@ func _Document_OneofSizer(msg proto.Message) (n int) { // Represents a sentence in the input document. type Sentence struct { // The sentence text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // For calls to [AnalyzeSentiment][] or if // [AnnotateTextRequest.Features.extract_document_sentiment][google.cloud.language.v1beta1.AnnotateTextRequest.Features.extract_document_sentiment] is set to // true, this field will contain the sentiment for the sentence. - Sentiment *Sentiment `protobuf:"bytes,2,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,2,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1279,24 +1279,24 @@ func (m *Sentence) GetSentiment() *Sentiment { // as salience and mentions, with entities. type Entity struct { // The representative name for the entity. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The entity type. - Type Entity_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.language.v1beta1.Entity_Type" json:"type,omitempty"` + Type Entity_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.language.v1beta1.Entity_Type" json:"type,omitempty"` // Metadata associated with the entity. // // Currently, Wikipedia URLs and Knowledge Graph MIDs are provided, if // available. The associated keys are "wikipedia_url" and "mid", respectively. - Metadata map[string]string `protobuf:"bytes,3,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,3,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The salience score associated with the entity in the [0, 1.0] range. // // The salience score for an entity provides information about the // importance or centrality of that entity to the entire document text. // Scores closer to 0 are less salient, while scores closer to 1.0 are highly // salient. - Salience float32 `protobuf:"fixed32,4,opt,name=salience" json:"salience,omitempty"` + Salience float32 `protobuf:"fixed32,4,opt,name=salience,proto3" json:"salience,omitempty"` // The mentions of this entity in the input document. The API currently // supports proper noun mentions. - Mentions []*EntityMention `protobuf:"bytes,5,rep,name=mentions" json:"mentions,omitempty"` + Mentions []*EntityMention `protobuf:"bytes,5,rep,name=mentions,proto3" json:"mentions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1364,13 +1364,13 @@ func (m *Entity) GetMentions() []*EntityMention { // Represents the smallest syntactic building block of the text. type Token struct { // The token text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Parts of speech tag for this token. - PartOfSpeech *PartOfSpeech `protobuf:"bytes,2,opt,name=part_of_speech,json=partOfSpeech" json:"part_of_speech,omitempty"` + PartOfSpeech *PartOfSpeech `protobuf:"bytes,2,opt,name=part_of_speech,json=partOfSpeech,proto3" json:"part_of_speech,omitempty"` // Dependency tree parse for this token. - DependencyEdge *DependencyEdge `protobuf:"bytes,3,opt,name=dependency_edge,json=dependencyEdge" json:"dependency_edge,omitempty"` + DependencyEdge *DependencyEdge `protobuf:"bytes,3,opt,name=dependency_edge,json=dependencyEdge,proto3" json:"dependency_edge,omitempty"` // [Lemma](https://en.wikipedia.org/wiki/Lemma_%28morphology%29) of the token. - Lemma string `protobuf:"bytes,4,opt,name=lemma" json:"lemma,omitempty"` + Lemma string `protobuf:"bytes,4,opt,name=lemma,proto3" json:"lemma,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1434,14 +1434,14 @@ type Sentiment struct { // DEPRECATED FIELD - This field is being deprecated in // favor of score. Please refer to our documentation at // https://cloud.google.com/natural-language/docs for more information. - Polarity float32 `protobuf:"fixed32,1,opt,name=polarity" json:"polarity,omitempty"` + Polarity float32 `protobuf:"fixed32,1,opt,name=polarity,proto3" json:"polarity,omitempty"` // A non-negative number in the [0, +inf) range, which represents // the absolute magnitude of sentiment regardless of score (positive or // negative). - Magnitude float32 `protobuf:"fixed32,2,opt,name=magnitude" json:"magnitude,omitempty"` + Magnitude float32 `protobuf:"fixed32,2,opt,name=magnitude,proto3" json:"magnitude,omitempty"` // Sentiment score between -1.0 (negative sentiment) and 1.0 // (positive sentiment). - Score float32 `protobuf:"fixed32,3,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,3,opt,name=score,proto3" json:"score,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1495,29 +1495,29 @@ func (m *Sentiment) GetScore() float32 { // Represents part of speech information for a token. type PartOfSpeech struct { // The part of speech tag. - Tag PartOfSpeech_Tag `protobuf:"varint,1,opt,name=tag,enum=google.cloud.language.v1beta1.PartOfSpeech_Tag" json:"tag,omitempty"` + Tag PartOfSpeech_Tag `protobuf:"varint,1,opt,name=tag,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Tag" json:"tag,omitempty"` // The grammatical aspect. - Aspect PartOfSpeech_Aspect `protobuf:"varint,2,opt,name=aspect,enum=google.cloud.language.v1beta1.PartOfSpeech_Aspect" json:"aspect,omitempty"` + Aspect PartOfSpeech_Aspect `protobuf:"varint,2,opt,name=aspect,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Aspect" json:"aspect,omitempty"` // The grammatical case. - Case PartOfSpeech_Case `protobuf:"varint,3,opt,name=case,enum=google.cloud.language.v1beta1.PartOfSpeech_Case" json:"case,omitempty"` + Case PartOfSpeech_Case `protobuf:"varint,3,opt,name=case,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Case" json:"case,omitempty"` // The grammatical form. - Form PartOfSpeech_Form `protobuf:"varint,4,opt,name=form,enum=google.cloud.language.v1beta1.PartOfSpeech_Form" json:"form,omitempty"` + Form PartOfSpeech_Form `protobuf:"varint,4,opt,name=form,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Form" json:"form,omitempty"` // The grammatical gender. - Gender PartOfSpeech_Gender `protobuf:"varint,5,opt,name=gender,enum=google.cloud.language.v1beta1.PartOfSpeech_Gender" json:"gender,omitempty"` + Gender PartOfSpeech_Gender `protobuf:"varint,5,opt,name=gender,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Gender" json:"gender,omitempty"` // The grammatical mood. - Mood PartOfSpeech_Mood `protobuf:"varint,6,opt,name=mood,enum=google.cloud.language.v1beta1.PartOfSpeech_Mood" json:"mood,omitempty"` + Mood PartOfSpeech_Mood `protobuf:"varint,6,opt,name=mood,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Mood" json:"mood,omitempty"` // The grammatical number. - Number PartOfSpeech_Number `protobuf:"varint,7,opt,name=number,enum=google.cloud.language.v1beta1.PartOfSpeech_Number" json:"number,omitempty"` + Number PartOfSpeech_Number `protobuf:"varint,7,opt,name=number,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Number" json:"number,omitempty"` // The grammatical person. - Person PartOfSpeech_Person `protobuf:"varint,8,opt,name=person,enum=google.cloud.language.v1beta1.PartOfSpeech_Person" json:"person,omitempty"` + Person PartOfSpeech_Person `protobuf:"varint,8,opt,name=person,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Person" json:"person,omitempty"` // The grammatical properness. - Proper PartOfSpeech_Proper `protobuf:"varint,9,opt,name=proper,enum=google.cloud.language.v1beta1.PartOfSpeech_Proper" json:"proper,omitempty"` + Proper PartOfSpeech_Proper `protobuf:"varint,9,opt,name=proper,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Proper" json:"proper,omitempty"` // The grammatical reciprocity. - Reciprocity PartOfSpeech_Reciprocity `protobuf:"varint,10,opt,name=reciprocity,enum=google.cloud.language.v1beta1.PartOfSpeech_Reciprocity" json:"reciprocity,omitempty"` + Reciprocity PartOfSpeech_Reciprocity `protobuf:"varint,10,opt,name=reciprocity,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Reciprocity" json:"reciprocity,omitempty"` // The grammatical tense. - Tense PartOfSpeech_Tense `protobuf:"varint,11,opt,name=tense,enum=google.cloud.language.v1beta1.PartOfSpeech_Tense" json:"tense,omitempty"` + Tense PartOfSpeech_Tense `protobuf:"varint,11,opt,name=tense,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Tense" json:"tense,omitempty"` // The grammatical voice. - Voice PartOfSpeech_Voice `protobuf:"varint,12,opt,name=voice,enum=google.cloud.language.v1beta1.PartOfSpeech_Voice" json:"voice,omitempty"` + Voice PartOfSpeech_Voice `protobuf:"varint,12,opt,name=voice,proto3,enum=google.cloud.language.v1beta1.PartOfSpeech_Voice" json:"voice,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1638,9 +1638,9 @@ type DependencyEdge struct { // The index is the position of the token in the array of tokens returned // by the API method. If this token is a root token, then the // `head_token_index` is its own index. - HeadTokenIndex int32 `protobuf:"varint,1,opt,name=head_token_index,json=headTokenIndex" json:"head_token_index,omitempty"` + HeadTokenIndex int32 `protobuf:"varint,1,opt,name=head_token_index,json=headTokenIndex,proto3" json:"head_token_index,omitempty"` // The parse label for the token. - Label DependencyEdge_Label `protobuf:"varint,2,opt,name=label,enum=google.cloud.language.v1beta1.DependencyEdge_Label" json:"label,omitempty"` + Label DependencyEdge_Label `protobuf:"varint,2,opt,name=label,proto3,enum=google.cloud.language.v1beta1.DependencyEdge_Label" json:"label,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1688,9 +1688,9 @@ func (m *DependencyEdge) GetLabel() DependencyEdge_Label { // mentions are supported. type EntityMention struct { // The mention text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // The type of the entity mention. - Type EntityMention_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.language.v1beta1.EntityMention_Type" json:"type,omitempty"` + Type EntityMention_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.language.v1beta1.EntityMention_Type" json:"type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1737,10 +1737,10 @@ func (m *EntityMention) GetType() EntityMention_Type { // Represents an output piece of text. type TextSpan struct { // The content of the output text. - Content string `protobuf:"bytes,1,opt,name=content" json:"content,omitempty"` + Content string `protobuf:"bytes,1,opt,name=content,proto3" json:"content,omitempty"` // The API calculates the beginning offset of the content in the original // document according to the [EncodingType][google.cloud.language.v1beta1.EncodingType] specified in the API request. - BeginOffset int32 `protobuf:"varint,2,opt,name=begin_offset,json=beginOffset" json:"begin_offset,omitempty"` + BeginOffset int32 `protobuf:"varint,2,opt,name=begin_offset,json=beginOffset,proto3" json:"begin_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1787,10 +1787,10 @@ func (m *TextSpan) GetBeginOffset() int32 { // The sentiment analysis request message. type AnalyzeSentimentRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate sentence offsets for the // sentence sentiment. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1837,13 +1837,13 @@ func (m *AnalyzeSentimentRequest) GetEncodingType() EncodingType { // The sentiment analysis response message. type AnalyzeSentimentResponse struct { // The overall sentiment of the input document. - DocumentSentiment *Sentiment `protobuf:"bytes,1,opt,name=document_sentiment,json=documentSentiment" json:"document_sentiment,omitempty"` + DocumentSentiment *Sentiment `protobuf:"bytes,1,opt,name=document_sentiment,json=documentSentiment,proto3" json:"document_sentiment,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta1.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` // The sentiment for all the sentences in the document. - Sentences []*Sentence `protobuf:"bytes,3,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,3,rep,name=sentences,proto3" json:"sentences,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1897,9 +1897,9 @@ func (m *AnalyzeSentimentResponse) GetSentences() []*Sentence { // The entity analysis request message. type AnalyzeEntitiesRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1946,11 +1946,11 @@ func (m *AnalyzeEntitiesRequest) GetEncodingType() EncodingType { // The entity analysis response message. type AnalyzeEntitiesResponse struct { // The recognized entities in the input document. - Entities []*Entity `protobuf:"bytes,1,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,1,rep,name=entities,proto3" json:"entities,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta1.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1997,9 +1997,9 @@ func (m *AnalyzeEntitiesResponse) GetLanguage() string { // The syntax analysis request message. type AnalyzeSyntaxRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2046,13 +2046,13 @@ func (m *AnalyzeSyntaxRequest) GetEncodingType() EncodingType { // The syntax analysis response message. type AnalyzeSyntaxResponse struct { // Sentences in the input document. - Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences,proto3" json:"sentences,omitempty"` // Tokens, along with their syntactic information, in the input document. - Tokens []*Token `protobuf:"bytes,2,rep,name=tokens" json:"tokens,omitempty"` + Tokens []*Token `protobuf:"bytes,2,rep,name=tokens,proto3" json:"tokens,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta1.Document.language] field for more details. - Language string `protobuf:"bytes,3,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,3,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2107,11 +2107,11 @@ func (m *AnalyzeSyntaxResponse) GetLanguage() string { // analysis types (sentiment, entities, and syntax) in one call. type AnnotateTextRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The enabled features. - Features *AnnotateTextRequest_Features `protobuf:"bytes,2,opt,name=features" json:"features,omitempty"` + Features *AnnotateTextRequest_Features `protobuf:"bytes,2,opt,name=features,proto3" json:"features,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,3,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,3,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta1.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2166,11 +2166,11 @@ func (m *AnnotateTextRequest) GetEncodingType() EncodingType { // Setting each one to true will enable that specific analysis for the input. type AnnotateTextRequest_Features struct { // Extract syntax information. - ExtractSyntax bool `protobuf:"varint,1,opt,name=extract_syntax,json=extractSyntax" json:"extract_syntax,omitempty"` + ExtractSyntax bool `protobuf:"varint,1,opt,name=extract_syntax,json=extractSyntax,proto3" json:"extract_syntax,omitempty"` // Extract entities. - ExtractEntities bool `protobuf:"varint,2,opt,name=extract_entities,json=extractEntities" json:"extract_entities,omitempty"` + ExtractEntities bool `protobuf:"varint,2,opt,name=extract_entities,json=extractEntities,proto3" json:"extract_entities,omitempty"` // Extract document-level sentiment. - ExtractDocumentSentiment bool `protobuf:"varint,3,opt,name=extract_document_sentiment,json=extractDocumentSentiment" json:"extract_document_sentiment,omitempty"` + ExtractDocumentSentiment bool `protobuf:"varint,3,opt,name=extract_document_sentiment,json=extractDocumentSentiment,proto3" json:"extract_document_sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2225,22 +2225,22 @@ func (m *AnnotateTextRequest_Features) GetExtractDocumentSentiment() bool { type AnnotateTextResponse struct { // Sentences in the input document. Populated if the user enables // [AnnotateTextRequest.Features.extract_syntax][google.cloud.language.v1beta1.AnnotateTextRequest.Features.extract_syntax]. - Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences,proto3" json:"sentences,omitempty"` // Tokens, along with their syntactic information, in the input document. // Populated if the user enables // [AnnotateTextRequest.Features.extract_syntax][google.cloud.language.v1beta1.AnnotateTextRequest.Features.extract_syntax]. - Tokens []*Token `protobuf:"bytes,2,rep,name=tokens" json:"tokens,omitempty"` + Tokens []*Token `protobuf:"bytes,2,rep,name=tokens,proto3" json:"tokens,omitempty"` // Entities, along with their semantic information, in the input document. // Populated if the user enables // [AnnotateTextRequest.Features.extract_entities][google.cloud.language.v1beta1.AnnotateTextRequest.Features.extract_entities]. - Entities []*Entity `protobuf:"bytes,3,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,3,rep,name=entities,proto3" json:"entities,omitempty"` // The overall sentiment for the document. Populated if the user enables // [AnnotateTextRequest.Features.extract_document_sentiment][google.cloud.language.v1beta1.AnnotateTextRequest.Features.extract_document_sentiment]. - DocumentSentiment *Sentiment `protobuf:"bytes,4,opt,name=document_sentiment,json=documentSentiment" json:"document_sentiment,omitempty"` + DocumentSentiment *Sentiment `protobuf:"bytes,4,opt,name=document_sentiment,json=documentSentiment,proto3" json:"document_sentiment,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta1.Document.language] field for more details. - Language string `protobuf:"bytes,5,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,5,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2415,8 +2415,7 @@ func (c *languageServiceClient) AnnotateText(ctx context.Context, in *AnnotateTe return out, nil } -// Server API for LanguageService service - +// LanguageServiceServer is the server API for LanguageService service. type LanguageServiceServer interface { // Analyzes the sentiment of the provided text. AnalyzeSentiment(context.Context, *AnalyzeSentimentRequest) (*AnalyzeSentimentResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta2/language_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta2/language_service.pb.go index 934bde3ec..a1d685cbe 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta2/language_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/language/v1beta2/language_service.pb.go @@ -1089,7 +1089,7 @@ func (EntityMention_Type) EnumDescriptor() ([]byte, []int) { type Document struct { // Required. If the type is not set or is `TYPE_UNSPECIFIED`, // returns an `INVALID_ARGUMENT` error. - Type Document_Type `protobuf:"varint,1,opt,name=type,enum=google.cloud.language.v1beta2.Document_Type" json:"type,omitempty"` + Type Document_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.language.v1beta2.Document_Type" json:"type,omitempty"` // The source of the document: a string containing the content or a // Google Cloud Storage URI. // @@ -1105,7 +1105,7 @@ type Document struct { // If the language (either specified by the caller or automatically detected) // is not supported by the called API method, an `INVALID_ARGUMENT` error // is returned. - Language string `protobuf:"bytes,4,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,4,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1140,10 +1140,10 @@ type isDocument_Source interface { } type Document_Content struct { - Content string `protobuf:"bytes,2,opt,name=content,oneof"` + Content string `protobuf:"bytes,2,opt,name=content,proto3,oneof"` } type Document_GcsContentUri struct { - GcsContentUri string `protobuf:"bytes,3,opt,name=gcs_content_uri,json=gcsContentUri,oneof"` + GcsContentUri string `protobuf:"bytes,3,opt,name=gcs_content_uri,json=gcsContentUri,proto3,oneof"` } func (*Document_Content) isDocument_Source() {} @@ -1253,11 +1253,11 @@ func _Document_OneofSizer(msg proto.Message) (n int) { // Represents a sentence in the input document. type Sentence struct { // The sentence text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // For calls to [AnalyzeSentiment][] or if // [AnnotateTextRequest.Features.extract_document_sentiment][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_document_sentiment] is set to // true, this field will contain the sentiment for the sentence. - Sentiment *Sentiment `protobuf:"bytes,2,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,2,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1306,29 +1306,29 @@ func (m *Sentence) GetSentiment() *Sentiment { // as salience and mentions, with entities. type Entity struct { // The representative name for the entity. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The entity type. - Type Entity_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.language.v1beta2.Entity_Type" json:"type,omitempty"` + Type Entity_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.language.v1beta2.Entity_Type" json:"type,omitempty"` // Metadata associated with the entity. // // Currently, Wikipedia URLs and Knowledge Graph MIDs are provided, if // available. The associated keys are "wikipedia_url" and "mid", respectively. - Metadata map[string]string `protobuf:"bytes,3,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,3,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The salience score associated with the entity in the [0, 1.0] range. // // The salience score for an entity provides information about the // importance or centrality of that entity to the entire document text. // Scores closer to 0 are less salient, while scores closer to 1.0 are highly // salient. - Salience float32 `protobuf:"fixed32,4,opt,name=salience" json:"salience,omitempty"` + Salience float32 `protobuf:"fixed32,4,opt,name=salience,proto3" json:"salience,omitempty"` // The mentions of this entity in the input document. The API currently // supports proper noun mentions. - Mentions []*EntityMention `protobuf:"bytes,5,rep,name=mentions" json:"mentions,omitempty"` + Mentions []*EntityMention `protobuf:"bytes,5,rep,name=mentions,proto3" json:"mentions,omitempty"` // For calls to [AnalyzeEntitySentiment][] or if // [AnnotateTextRequest.Features.extract_entity_sentiment][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_entity_sentiment] is set to // true, this field will contain the aggregate sentiment expressed for this // entity in the provided document. - Sentiment *Sentiment `protobuf:"bytes,6,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,6,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1403,13 +1403,13 @@ func (m *Entity) GetSentiment() *Sentiment { // Represents the smallest syntactic building block of the text. type Token struct { // The token text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Parts of speech tag for this token. - PartOfSpeech *PartOfSpeech `protobuf:"bytes,2,opt,name=part_of_speech,json=partOfSpeech" json:"part_of_speech,omitempty"` + PartOfSpeech *PartOfSpeech `protobuf:"bytes,2,opt,name=part_of_speech,json=partOfSpeech,proto3" json:"part_of_speech,omitempty"` // Dependency tree parse for this token. - DependencyEdge *DependencyEdge `protobuf:"bytes,3,opt,name=dependency_edge,json=dependencyEdge" json:"dependency_edge,omitempty"` + DependencyEdge *DependencyEdge `protobuf:"bytes,3,opt,name=dependency_edge,json=dependencyEdge,proto3" json:"dependency_edge,omitempty"` // [Lemma](https://en.wikipedia.org/wiki/Lemma_%28morphology%29) of the token. - Lemma string `protobuf:"bytes,4,opt,name=lemma" json:"lemma,omitempty"` + Lemma string `protobuf:"bytes,4,opt,name=lemma,proto3" json:"lemma,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1473,10 +1473,10 @@ type Sentiment struct { // A non-negative number in the [0, +inf) range, which represents // the absolute magnitude of sentiment regardless of score (positive or // negative). - Magnitude float32 `protobuf:"fixed32,2,opt,name=magnitude" json:"magnitude,omitempty"` + Magnitude float32 `protobuf:"fixed32,2,opt,name=magnitude,proto3" json:"magnitude,omitempty"` // Sentiment score between -1.0 (negative sentiment) and 1.0 // (positive sentiment). - Score float32 `protobuf:"fixed32,3,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,3,opt,name=score,proto3" json:"score,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1523,29 +1523,29 @@ func (m *Sentiment) GetScore() float32 { // Represents part of speech information for a token. type PartOfSpeech struct { // The part of speech tag. - Tag PartOfSpeech_Tag `protobuf:"varint,1,opt,name=tag,enum=google.cloud.language.v1beta2.PartOfSpeech_Tag" json:"tag,omitempty"` + Tag PartOfSpeech_Tag `protobuf:"varint,1,opt,name=tag,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Tag" json:"tag,omitempty"` // The grammatical aspect. - Aspect PartOfSpeech_Aspect `protobuf:"varint,2,opt,name=aspect,enum=google.cloud.language.v1beta2.PartOfSpeech_Aspect" json:"aspect,omitempty"` + Aspect PartOfSpeech_Aspect `protobuf:"varint,2,opt,name=aspect,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Aspect" json:"aspect,omitempty"` // The grammatical case. - Case PartOfSpeech_Case `protobuf:"varint,3,opt,name=case,enum=google.cloud.language.v1beta2.PartOfSpeech_Case" json:"case,omitempty"` + Case PartOfSpeech_Case `protobuf:"varint,3,opt,name=case,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Case" json:"case,omitempty"` // The grammatical form. - Form PartOfSpeech_Form `protobuf:"varint,4,opt,name=form,enum=google.cloud.language.v1beta2.PartOfSpeech_Form" json:"form,omitempty"` + Form PartOfSpeech_Form `protobuf:"varint,4,opt,name=form,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Form" json:"form,omitempty"` // The grammatical gender. - Gender PartOfSpeech_Gender `protobuf:"varint,5,opt,name=gender,enum=google.cloud.language.v1beta2.PartOfSpeech_Gender" json:"gender,omitempty"` + Gender PartOfSpeech_Gender `protobuf:"varint,5,opt,name=gender,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Gender" json:"gender,omitempty"` // The grammatical mood. - Mood PartOfSpeech_Mood `protobuf:"varint,6,opt,name=mood,enum=google.cloud.language.v1beta2.PartOfSpeech_Mood" json:"mood,omitempty"` + Mood PartOfSpeech_Mood `protobuf:"varint,6,opt,name=mood,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Mood" json:"mood,omitempty"` // The grammatical number. - Number PartOfSpeech_Number `protobuf:"varint,7,opt,name=number,enum=google.cloud.language.v1beta2.PartOfSpeech_Number" json:"number,omitempty"` + Number PartOfSpeech_Number `protobuf:"varint,7,opt,name=number,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Number" json:"number,omitempty"` // The grammatical person. - Person PartOfSpeech_Person `protobuf:"varint,8,opt,name=person,enum=google.cloud.language.v1beta2.PartOfSpeech_Person" json:"person,omitempty"` + Person PartOfSpeech_Person `protobuf:"varint,8,opt,name=person,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Person" json:"person,omitempty"` // The grammatical properness. - Proper PartOfSpeech_Proper `protobuf:"varint,9,opt,name=proper,enum=google.cloud.language.v1beta2.PartOfSpeech_Proper" json:"proper,omitempty"` + Proper PartOfSpeech_Proper `protobuf:"varint,9,opt,name=proper,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Proper" json:"proper,omitempty"` // The grammatical reciprocity. - Reciprocity PartOfSpeech_Reciprocity `protobuf:"varint,10,opt,name=reciprocity,enum=google.cloud.language.v1beta2.PartOfSpeech_Reciprocity" json:"reciprocity,omitempty"` + Reciprocity PartOfSpeech_Reciprocity `protobuf:"varint,10,opt,name=reciprocity,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Reciprocity" json:"reciprocity,omitempty"` // The grammatical tense. - Tense PartOfSpeech_Tense `protobuf:"varint,11,opt,name=tense,enum=google.cloud.language.v1beta2.PartOfSpeech_Tense" json:"tense,omitempty"` + Tense PartOfSpeech_Tense `protobuf:"varint,11,opt,name=tense,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Tense" json:"tense,omitempty"` // The grammatical voice. - Voice PartOfSpeech_Voice `protobuf:"varint,12,opt,name=voice,enum=google.cloud.language.v1beta2.PartOfSpeech_Voice" json:"voice,omitempty"` + Voice PartOfSpeech_Voice `protobuf:"varint,12,opt,name=voice,proto3,enum=google.cloud.language.v1beta2.PartOfSpeech_Voice" json:"voice,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1666,9 +1666,9 @@ type DependencyEdge struct { // The index is the position of the token in the array of tokens returned // by the API method. If this token is a root token, then the // `head_token_index` is its own index. - HeadTokenIndex int32 `protobuf:"varint,1,opt,name=head_token_index,json=headTokenIndex" json:"head_token_index,omitempty"` + HeadTokenIndex int32 `protobuf:"varint,1,opt,name=head_token_index,json=headTokenIndex,proto3" json:"head_token_index,omitempty"` // The parse label for the token. - Label DependencyEdge_Label `protobuf:"varint,2,opt,name=label,enum=google.cloud.language.v1beta2.DependencyEdge_Label" json:"label,omitempty"` + Label DependencyEdge_Label `protobuf:"varint,2,opt,name=label,proto3,enum=google.cloud.language.v1beta2.DependencyEdge_Label" json:"label,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1716,14 +1716,14 @@ func (m *DependencyEdge) GetLabel() DependencyEdge_Label { // mentions are supported. type EntityMention struct { // The mention text. - Text *TextSpan `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text *TextSpan `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // The type of the entity mention. - Type EntityMention_Type `protobuf:"varint,2,opt,name=type,enum=google.cloud.language.v1beta2.EntityMention_Type" json:"type,omitempty"` + Type EntityMention_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.cloud.language.v1beta2.EntityMention_Type" json:"type,omitempty"` // For calls to [AnalyzeEntitySentiment][] or if // [AnnotateTextRequest.Features.extract_entity_sentiment][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_entity_sentiment] is set to // true, this field will contain the sentiment expressed for this mention of // the entity in the provided document. - Sentiment *Sentiment `protobuf:"bytes,3,opt,name=sentiment" json:"sentiment,omitempty"` + Sentiment *Sentiment `protobuf:"bytes,3,opt,name=sentiment,proto3" json:"sentiment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1777,10 +1777,10 @@ func (m *EntityMention) GetSentiment() *Sentiment { // Represents an output piece of text. type TextSpan struct { // The content of the output text. - Content string `protobuf:"bytes,1,opt,name=content" json:"content,omitempty"` + Content string `protobuf:"bytes,1,opt,name=content,proto3" json:"content,omitempty"` // The API calculates the beginning offset of the content in the original // document according to the [EncodingType][google.cloud.language.v1beta2.EncodingType] specified in the API request. - BeginOffset int32 `protobuf:"varint,2,opt,name=begin_offset,json=beginOffset" json:"begin_offset,omitempty"` + BeginOffset int32 `protobuf:"varint,2,opt,name=begin_offset,json=beginOffset,proto3" json:"begin_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1827,10 +1827,10 @@ func (m *TextSpan) GetBeginOffset() int32 { // Represents a category returned from the text classifier. type ClassificationCategory struct { // The name of the category representing the document. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The classifier's confidence of the category. Number represents how certain // the classifier is that this category represents the given text. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1877,10 +1877,10 @@ func (m *ClassificationCategory) GetConfidence() float32 { // The sentiment analysis request message. type AnalyzeSentimentRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate sentence offsets for the // sentence sentiment. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1927,13 +1927,13 @@ func (m *AnalyzeSentimentRequest) GetEncodingType() EncodingType { // The sentiment analysis response message. type AnalyzeSentimentResponse struct { // The overall sentiment of the input document. - DocumentSentiment *Sentiment `protobuf:"bytes,1,opt,name=document_sentiment,json=documentSentiment" json:"document_sentiment,omitempty"` + DocumentSentiment *Sentiment `protobuf:"bytes,1,opt,name=document_sentiment,json=documentSentiment,proto3" json:"document_sentiment,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta2.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` // The sentiment for all the sentences in the document. - Sentences []*Sentence `protobuf:"bytes,3,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,3,rep,name=sentences,proto3" json:"sentences,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1987,9 +1987,9 @@ func (m *AnalyzeSentimentResponse) GetSentences() []*Sentence { // The entity-level sentiment analysis request message. type AnalyzeEntitySentimentRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2036,11 +2036,11 @@ func (m *AnalyzeEntitySentimentRequest) GetEncodingType() EncodingType { // The entity-level sentiment analysis response message. type AnalyzeEntitySentimentResponse struct { // The recognized entities in the input document with associated sentiments. - Entities []*Entity `protobuf:"bytes,1,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,1,rep,name=entities,proto3" json:"entities,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta2.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2087,9 +2087,9 @@ func (m *AnalyzeEntitySentimentResponse) GetLanguage() string { // The entity analysis request message. type AnalyzeEntitiesRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2136,11 +2136,11 @@ func (m *AnalyzeEntitiesRequest) GetEncodingType() EncodingType { // The entity analysis response message. type AnalyzeEntitiesResponse struct { // The recognized entities in the input document. - Entities []*Entity `protobuf:"bytes,1,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,1,rep,name=entities,proto3" json:"entities,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta2.Document.language] field for more details. - Language string `protobuf:"bytes,2,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,2,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2187,9 +2187,9 @@ func (m *AnalyzeEntitiesResponse) GetLanguage() string { // The syntax analysis request message. type AnalyzeSyntaxRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,2,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2236,13 +2236,13 @@ func (m *AnalyzeSyntaxRequest) GetEncodingType() EncodingType { // The syntax analysis response message. type AnalyzeSyntaxResponse struct { // Sentences in the input document. - Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences,proto3" json:"sentences,omitempty"` // Tokens, along with their syntactic information, in the input document. - Tokens []*Token `protobuf:"bytes,2,rep,name=tokens" json:"tokens,omitempty"` + Tokens []*Token `protobuf:"bytes,2,rep,name=tokens,proto3" json:"tokens,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta2.Document.language] field for more details. - Language string `protobuf:"bytes,3,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,3,opt,name=language,proto3" json:"language,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2296,7 +2296,7 @@ func (m *AnalyzeSyntaxResponse) GetLanguage() string { // The document classification request message. type ClassifyTextRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2336,7 +2336,7 @@ func (m *ClassifyTextRequest) GetDocument() *Document { // The document classification response message. type ClassifyTextResponse struct { // Categories representing the input document. - Categories []*ClassificationCategory `protobuf:"bytes,1,rep,name=categories" json:"categories,omitempty"` + Categories []*ClassificationCategory `protobuf:"bytes,1,rep,name=categories,proto3" json:"categories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2377,11 +2377,11 @@ func (m *ClassifyTextResponse) GetCategories() []*ClassificationCategory { // analysis types (sentiment, entities, and syntax) in one call. type AnnotateTextRequest struct { // Input document. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The enabled features. - Features *AnnotateTextRequest_Features `protobuf:"bytes,2,opt,name=features" json:"features,omitempty"` + Features *AnnotateTextRequest_Features `protobuf:"bytes,2,opt,name=features,proto3" json:"features,omitempty"` // The encoding type used by the API to calculate offsets. - EncodingType EncodingType `protobuf:"varint,3,opt,name=encoding_type,json=encodingType,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` + EncodingType EncodingType `protobuf:"varint,3,opt,name=encoding_type,json=encodingType,proto3,enum=google.cloud.language.v1beta2.EncodingType" json:"encoding_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2436,15 +2436,15 @@ func (m *AnnotateTextRequest) GetEncodingType() EncodingType { // Setting each one to true will enable that specific analysis for the input. type AnnotateTextRequest_Features struct { // Extract syntax information. - ExtractSyntax bool `protobuf:"varint,1,opt,name=extract_syntax,json=extractSyntax" json:"extract_syntax,omitempty"` + ExtractSyntax bool `protobuf:"varint,1,opt,name=extract_syntax,json=extractSyntax,proto3" json:"extract_syntax,omitempty"` // Extract entities. - ExtractEntities bool `protobuf:"varint,2,opt,name=extract_entities,json=extractEntities" json:"extract_entities,omitempty"` + ExtractEntities bool `protobuf:"varint,2,opt,name=extract_entities,json=extractEntities,proto3" json:"extract_entities,omitempty"` // Extract document-level sentiment. - ExtractDocumentSentiment bool `protobuf:"varint,3,opt,name=extract_document_sentiment,json=extractDocumentSentiment" json:"extract_document_sentiment,omitempty"` + ExtractDocumentSentiment bool `protobuf:"varint,3,opt,name=extract_document_sentiment,json=extractDocumentSentiment,proto3" json:"extract_document_sentiment,omitempty"` // Extract entities and their associated sentiment. - ExtractEntitySentiment bool `protobuf:"varint,4,opt,name=extract_entity_sentiment,json=extractEntitySentiment" json:"extract_entity_sentiment,omitempty"` + ExtractEntitySentiment bool `protobuf:"varint,4,opt,name=extract_entity_sentiment,json=extractEntitySentiment,proto3" json:"extract_entity_sentiment,omitempty"` // Classify the full document into categories. - ClassifyText bool `protobuf:"varint,6,opt,name=classify_text,json=classifyText" json:"classify_text,omitempty"` + ClassifyText bool `protobuf:"varint,6,opt,name=classify_text,json=classifyText,proto3" json:"classify_text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2513,24 +2513,24 @@ func (m *AnnotateTextRequest_Features) GetClassifyText() bool { type AnnotateTextResponse struct { // Sentences in the input document. Populated if the user enables // [AnnotateTextRequest.Features.extract_syntax][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_syntax]. - Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences" json:"sentences,omitempty"` + Sentences []*Sentence `protobuf:"bytes,1,rep,name=sentences,proto3" json:"sentences,omitempty"` // Tokens, along with their syntactic information, in the input document. // Populated if the user enables // [AnnotateTextRequest.Features.extract_syntax][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_syntax]. - Tokens []*Token `protobuf:"bytes,2,rep,name=tokens" json:"tokens,omitempty"` + Tokens []*Token `protobuf:"bytes,2,rep,name=tokens,proto3" json:"tokens,omitempty"` // Entities, along with their semantic information, in the input document. // Populated if the user enables // [AnnotateTextRequest.Features.extract_entities][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_entities]. - Entities []*Entity `protobuf:"bytes,3,rep,name=entities" json:"entities,omitempty"` + Entities []*Entity `protobuf:"bytes,3,rep,name=entities,proto3" json:"entities,omitempty"` // The overall sentiment for the document. Populated if the user enables // [AnnotateTextRequest.Features.extract_document_sentiment][google.cloud.language.v1beta2.AnnotateTextRequest.Features.extract_document_sentiment]. - DocumentSentiment *Sentiment `protobuf:"bytes,4,opt,name=document_sentiment,json=documentSentiment" json:"document_sentiment,omitempty"` + DocumentSentiment *Sentiment `protobuf:"bytes,4,opt,name=document_sentiment,json=documentSentiment,proto3" json:"document_sentiment,omitempty"` // The language of the text, which will be the same as the language specified // in the request or, if not specified, the automatically-detected language. // See [Document.language][google.cloud.language.v1beta2.Document.language] field for more details. - Language string `protobuf:"bytes,5,opt,name=language" json:"language,omitempty"` + Language string `protobuf:"bytes,5,opt,name=language,proto3" json:"language,omitempty"` // Categories identified in the input document. - Categories []*ClassificationCategory `protobuf:"bytes,6,rep,name=categories" json:"categories,omitempty"` + Categories []*ClassificationCategory `protobuf:"bytes,6,rep,name=categories,proto3" json:"categories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2740,8 +2740,7 @@ func (c *languageServiceClient) AnnotateText(ctx context.Context, in *AnnotateTe return out, nil } -// Server API for LanguageService service - +// LanguageServiceServer is the server API for LanguageService service. type LanguageServiceServer interface { // Analyzes the sentiment of the provided text. AnalyzeSentiment(context.Context, *AnalyzeSentimentRequest) (*AnalyzeSentimentResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/location/locations.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/location/locations.pb.go index ce0f66fd8..69702e083 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/location/locations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/location/locations.pb.go @@ -28,13 +28,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // The request message for [Locations.ListLocations][google.cloud.location.Locations.ListLocations]. type ListLocationsRequest struct { // The resource that owns the locations collection, if applicable. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The standard list filter. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // The standard list page size. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The standard list page token. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -95,9 +95,9 @@ func (m *ListLocationsRequest) GetPageToken() string { // The response message for [Locations.ListLocations][google.cloud.location.Locations.ListLocations]. type ListLocationsResponse struct { // A list of locations that matches the specified filter in the request. - Locations []*Location `protobuf:"bytes,1,rep,name=locations" json:"locations,omitempty"` + Locations []*Location `protobuf:"bytes,1,rep,name=locations,proto3" json:"locations,omitempty"` // The standard List next-page token. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -144,7 +144,7 @@ func (m *ListLocationsResponse) GetNextPageToken() string { // The request message for [Locations.GetLocation][google.cloud.location.Locations.GetLocation]. type GetLocationRequest struct { // Resource name for the location. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -185,19 +185,19 @@ func (m *GetLocationRequest) GetName() string { type Location struct { // Resource name for the location, which may vary between implementations. // For example: `"projects/example-project/locations/us-east1"` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The canonical id for this location. For example: `"us-east1"`. - LocationId string `protobuf:"bytes,4,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,4,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` // The friendly name for this location, typically a nearby city name. // For example, "Tokyo". - DisplayName string `protobuf:"bytes,5,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,5,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Cross-service attributes for the location. For example // // {"cloud.googleapis.com/region": "us-east1"} - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Service-specific metadata. For example the available capacity at the given // location. - Metadata *any.Any `protobuf:"bytes,3,opt,name=metadata" json:"metadata,omitempty"` + Metadata *any.Any `protobuf:"bytes,3,opt,name=metadata,proto3" json:"metadata,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -314,8 +314,7 @@ func (c *locationsClient) GetLocation(ctx context.Context, in *GetLocationReques return out, nil } -// Server API for Locations service - +// LocationsServer is the server API for Locations service. type LocationsServer interface { // Lists information about the supported locations for this service. ListLocations(context.Context, *ListLocationsRequest) (*ListLocationsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/job_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/job_service.pb.go index e9529de94..4f86bb8c7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/job_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/job_service.pb.go @@ -300,7 +300,7 @@ func (Job_State) EnumDescriptor() ([]byte, []int) { type TrainingInput struct { // Required. Specifies the machine types, the number of replicas for workers // and parameter servers. - ScaleTier TrainingInput_ScaleTier `protobuf:"varint,1,opt,name=scale_tier,json=scaleTier,enum=google.cloud.ml.v1.TrainingInput_ScaleTier" json:"scale_tier,omitempty"` + ScaleTier TrainingInput_ScaleTier `protobuf:"varint,1,opt,name=scale_tier,json=scaleTier,proto3,enum=google.cloud.ml.v1.TrainingInput_ScaleTier" json:"scale_tier,omitempty"` // Optional. Specifies the type of virtual machine to use for your training // job's master worker. // @@ -350,7 +350,7 @@ type TrainingInput struct { // // // You must set this value when `scaleTier` is set to `CUSTOM`. - MasterType string `protobuf:"bytes,2,opt,name=master_type,json=masterType" json:"master_type,omitempty"` + MasterType string `protobuf:"bytes,2,opt,name=master_type,json=masterType,proto3" json:"master_type,omitempty"` // Optional. Specifies the type of virtual machine to use for your training // job's worker nodes. // @@ -359,7 +359,7 @@ type TrainingInput struct { // // This value must be present when `scaleTier` is set to `CUSTOM` and // `workerCount` is greater than zero. - WorkerType string `protobuf:"bytes,3,opt,name=worker_type,json=workerType" json:"worker_type,omitempty"` + WorkerType string `protobuf:"bytes,3,opt,name=worker_type,json=workerType,proto3" json:"worker_type,omitempty"` // Optional. Specifies the type of virtual machine to use for your training // job's parameter server. // @@ -368,39 +368,39 @@ type TrainingInput struct { // // This value must be present when `scaleTier` is set to `CUSTOM` and // `parameter_server_count` is greater than zero. - ParameterServerType string `protobuf:"bytes,4,opt,name=parameter_server_type,json=parameterServerType" json:"parameter_server_type,omitempty"` + ParameterServerType string `protobuf:"bytes,4,opt,name=parameter_server_type,json=parameterServerType,proto3" json:"parameter_server_type,omitempty"` // Optional. The number of worker replicas to use for the training job. Each // replica in the cluster will be of the type specified in `worker_type`. // // This value can only be used when `scale_tier` is set to `CUSTOM`. If you // set this value, you must also set `worker_type`. - WorkerCount int64 `protobuf:"varint,5,opt,name=worker_count,json=workerCount" json:"worker_count,omitempty"` + WorkerCount int64 `protobuf:"varint,5,opt,name=worker_count,json=workerCount,proto3" json:"worker_count,omitempty"` // Optional. The number of parameter server replicas to use for the training // job. Each replica in the cluster will be of the type specified in // `parameter_server_type`. // // This value can only be used when `scale_tier` is set to `CUSTOM`.If you // set this value, you must also set `parameter_server_type`. - ParameterServerCount int64 `protobuf:"varint,6,opt,name=parameter_server_count,json=parameterServerCount" json:"parameter_server_count,omitempty"` + ParameterServerCount int64 `protobuf:"varint,6,opt,name=parameter_server_count,json=parameterServerCount,proto3" json:"parameter_server_count,omitempty"` // Required. The Google Cloud Storage location of the packages with // the training program and any additional dependencies. - PackageUris []string `protobuf:"bytes,7,rep,name=package_uris,json=packageUris" json:"package_uris,omitempty"` + PackageUris []string `protobuf:"bytes,7,rep,name=package_uris,json=packageUris,proto3" json:"package_uris,omitempty"` // Required. The Python module name to run after installing the packages. - PythonModule string `protobuf:"bytes,8,opt,name=python_module,json=pythonModule" json:"python_module,omitempty"` + PythonModule string `protobuf:"bytes,8,opt,name=python_module,json=pythonModule,proto3" json:"python_module,omitempty"` // Optional. Command line arguments to pass to the program. - Args []string `protobuf:"bytes,10,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,10,rep,name=args,proto3" json:"args,omitempty"` // Optional. The set of Hyperparameters to tune. - Hyperparameters *HyperparameterSpec `protobuf:"bytes,12,opt,name=hyperparameters" json:"hyperparameters,omitempty"` + Hyperparameters *HyperparameterSpec `protobuf:"bytes,12,opt,name=hyperparameters,proto3" json:"hyperparameters,omitempty"` // Required. The Google Compute Engine region to run the training job in. - Region string `protobuf:"bytes,14,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,14,opt,name=region,proto3" json:"region,omitempty"` // Optional. A Google Cloud Storage path in which to store training outputs // and other data needed for training. This path is passed to your TensorFlow // program as the 'job_dir' command-line argument. The benefit of specifying // this field is that Cloud ML validates the path for use in training. - JobDir string `protobuf:"bytes,16,opt,name=job_dir,json=jobDir" json:"job_dir,omitempty"` + JobDir string `protobuf:"bytes,16,opt,name=job_dir,json=jobDir,proto3" json:"job_dir,omitempty"` // Optional. The Google Cloud ML runtime version to use for training. If not // set, Google Cloud ML will choose the latest stable version. - RuntimeVersion string `protobuf:"bytes,15,opt,name=runtime_version,json=runtimeVersion" json:"runtime_version,omitempty"` + RuntimeVersion string `protobuf:"bytes,15,opt,name=runtime_version,json=runtimeVersion,proto3" json:"runtime_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -527,14 +527,14 @@ type HyperparameterSpec struct { // `MAXIMIZE` and `MINIMIZE`. // // Defaults to `MAXIMIZE`. - Goal HyperparameterSpec_GoalType `protobuf:"varint,1,opt,name=goal,enum=google.cloud.ml.v1.HyperparameterSpec_GoalType" json:"goal,omitempty"` + Goal HyperparameterSpec_GoalType `protobuf:"varint,1,opt,name=goal,proto3,enum=google.cloud.ml.v1.HyperparameterSpec_GoalType" json:"goal,omitempty"` // Required. The set of parameters to tune. - Params []*ParameterSpec `protobuf:"bytes,2,rep,name=params" json:"params,omitempty"` + Params []*ParameterSpec `protobuf:"bytes,2,rep,name=params,proto3" json:"params,omitempty"` // Optional. How many training trials should be attempted to optimize // the specified hyperparameters. // // Defaults to one. - MaxTrials int32 `protobuf:"varint,3,opt,name=max_trials,json=maxTrials" json:"max_trials,omitempty"` + MaxTrials int32 `protobuf:"varint,3,opt,name=max_trials,json=maxTrials,proto3" json:"max_trials,omitempty"` // Optional. The number of training trials to run concurrently. // You can reduce the time it takes to perform hyperparameter tuning by adding // trials in parallel. However, each trail only benefits from the information @@ -545,13 +545,13 @@ type HyperparameterSpec struct { // Each trial will use the same scale tier and machine types. // // Defaults to one. - MaxParallelTrials int32 `protobuf:"varint,4,opt,name=max_parallel_trials,json=maxParallelTrials" json:"max_parallel_trials,omitempty"` + MaxParallelTrials int32 `protobuf:"varint,4,opt,name=max_parallel_trials,json=maxParallelTrials,proto3" json:"max_parallel_trials,omitempty"` // Optional. The Tensorflow summary tag name to use for optimizing trials. For // current versions of Tensorflow, this tag name should exactly match what is // shown in Tensorboard, including all scopes. For versions of Tensorflow // prior to 0.12, this should be only the tag passed to tf.Summary. // By default, "training/hptuning/metric" will be used. - HyperparameterMetricTag string `protobuf:"bytes,5,opt,name=hyperparameter_metric_tag,json=hyperparameterMetricTag" json:"hyperparameter_metric_tag,omitempty"` + HyperparameterMetricTag string `protobuf:"bytes,5,opt,name=hyperparameter_metric_tag,json=hyperparameterMetricTag,proto3" json:"hyperparameter_metric_tag,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -620,30 +620,30 @@ func (m *HyperparameterSpec) GetHyperparameterMetricTag() string { type ParameterSpec struct { // Required. The parameter name must be unique amongst all ParameterConfigs in // a HyperparameterSpec message. E.g., "learning_rate". - ParameterName string `protobuf:"bytes,1,opt,name=parameter_name,json=parameterName" json:"parameter_name,omitempty"` + ParameterName string `protobuf:"bytes,1,opt,name=parameter_name,json=parameterName,proto3" json:"parameter_name,omitempty"` // Required. The type of the parameter. - Type ParameterSpec_ParameterType `protobuf:"varint,4,opt,name=type,enum=google.cloud.ml.v1.ParameterSpec_ParameterType" json:"type,omitempty"` + Type ParameterSpec_ParameterType `protobuf:"varint,4,opt,name=type,proto3,enum=google.cloud.ml.v1.ParameterSpec_ParameterType" json:"type,omitempty"` // Required if type is `DOUBLE` or `INTEGER`. This field // should be unset if type is `CATEGORICAL`. This value should be integers if // type is INTEGER. - MinValue float64 `protobuf:"fixed64,2,opt,name=min_value,json=minValue" json:"min_value,omitempty"` + MinValue float64 `protobuf:"fixed64,2,opt,name=min_value,json=minValue,proto3" json:"min_value,omitempty"` // Required if typeis `DOUBLE` or `INTEGER`. This field // should be unset if type is `CATEGORICAL`. This value should be integers if // type is `INTEGER`. - MaxValue float64 `protobuf:"fixed64,3,opt,name=max_value,json=maxValue" json:"max_value,omitempty"` + MaxValue float64 `protobuf:"fixed64,3,opt,name=max_value,json=maxValue,proto3" json:"max_value,omitempty"` // Required if type is `CATEGORICAL`. The list of possible categories. - CategoricalValues []string `protobuf:"bytes,5,rep,name=categorical_values,json=categoricalValues" json:"categorical_values,omitempty"` + CategoricalValues []string `protobuf:"bytes,5,rep,name=categorical_values,json=categoricalValues,proto3" json:"categorical_values,omitempty"` // Required if type is `DISCRETE`. // A list of feasible points. // The list should be in strictly increasing order. For instance, this // parameter might have possible settings of 1.5, 2.5, and 4.0. This list // should not contain more than 1,000 values. - DiscreteValues []float64 `protobuf:"fixed64,6,rep,packed,name=discrete_values,json=discreteValues" json:"discrete_values,omitempty"` + DiscreteValues []float64 `protobuf:"fixed64,6,rep,packed,name=discrete_values,json=discreteValues,proto3" json:"discrete_values,omitempty"` // Optional. How the parameter should be scaled to the hypercube. // Leave unset for categorical parameters. // Some kind of scaling is strongly recommended for real or integral // parameters (e.g., `UNIT_LINEAR_SCALE`). - ScaleType ParameterSpec_ScaleType `protobuf:"varint,7,opt,name=scale_type,json=scaleType,enum=google.cloud.ml.v1.ParameterSpec_ScaleType" json:"scale_type,omitempty"` + ScaleType ParameterSpec_ScaleType `protobuf:"varint,7,opt,name=scale_type,json=scaleType,proto3,enum=google.cloud.ml.v1.ParameterSpec_ScaleType" json:"scale_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -728,13 +728,13 @@ func (m *ParameterSpec) GetScaleType() ParameterSpec_ScaleType { // of HyperparameterOutput objects, one for each successful trial. type HyperparameterOutput struct { // The trial id for these results. - TrialId string `protobuf:"bytes,1,opt,name=trial_id,json=trialId" json:"trial_id,omitempty"` + TrialId string `protobuf:"bytes,1,opt,name=trial_id,json=trialId,proto3" json:"trial_id,omitempty"` // The hyperparameters given to this trial. - Hyperparameters map[string]string `protobuf:"bytes,2,rep,name=hyperparameters" json:"hyperparameters,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Hyperparameters map[string]string `protobuf:"bytes,2,rep,name=hyperparameters,proto3" json:"hyperparameters,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The final objective metric seen for this trial. - FinalMetric *HyperparameterOutput_HyperparameterMetric `protobuf:"bytes,3,opt,name=final_metric,json=finalMetric" json:"final_metric,omitempty"` + FinalMetric *HyperparameterOutput_HyperparameterMetric `protobuf:"bytes,3,opt,name=final_metric,json=finalMetric,proto3" json:"final_metric,omitempty"` // All recorded object metrics for this trial. - AllMetrics []*HyperparameterOutput_HyperparameterMetric `protobuf:"bytes,4,rep,name=all_metrics,json=allMetrics" json:"all_metrics,omitempty"` + AllMetrics []*HyperparameterOutput_HyperparameterMetric `protobuf:"bytes,4,rep,name=all_metrics,json=allMetrics,proto3" json:"all_metrics,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -795,9 +795,9 @@ func (m *HyperparameterOutput) GetAllMetrics() []*HyperparameterOutput_Hyperpara // An observed value of a metric. type HyperparameterOutput_HyperparameterMetric struct { // The global training step for this metric. - TrainingStep int64 `protobuf:"varint,1,opt,name=training_step,json=trainingStep" json:"training_step,omitempty"` + TrainingStep int64 `protobuf:"varint,1,opt,name=training_step,json=trainingStep,proto3" json:"training_step,omitempty"` // The objective value at this training step. - ObjectiveValue float64 `protobuf:"fixed64,2,opt,name=objective_value,json=objectiveValue" json:"objective_value,omitempty"` + ObjectiveValue float64 `protobuf:"fixed64,2,opt,name=objective_value,json=objectiveValue,proto3" json:"objective_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -847,14 +847,14 @@ func (m *HyperparameterOutput_HyperparameterMetric) GetObjectiveValue() float64 type TrainingOutput struct { // The number of hyperparameter tuning trials that completed successfully. // Only set for hyperparameter tuning jobs. - CompletedTrialCount int64 `protobuf:"varint,1,opt,name=completed_trial_count,json=completedTrialCount" json:"completed_trial_count,omitempty"` + CompletedTrialCount int64 `protobuf:"varint,1,opt,name=completed_trial_count,json=completedTrialCount,proto3" json:"completed_trial_count,omitempty"` // Results for individual Hyperparameter trials. // Only set for hyperparameter tuning jobs. - Trials []*HyperparameterOutput `protobuf:"bytes,2,rep,name=trials" json:"trials,omitempty"` + Trials []*HyperparameterOutput `protobuf:"bytes,2,rep,name=trials,proto3" json:"trials,omitempty"` // The amount of ML units consumed by the job. - ConsumedMlUnits float64 `protobuf:"fixed64,3,opt,name=consumed_ml_units,json=consumedMlUnits" json:"consumed_ml_units,omitempty"` + ConsumedMlUnits float64 `protobuf:"fixed64,3,opt,name=consumed_ml_units,json=consumedMlUnits,proto3" json:"consumed_ml_units,omitempty"` // Whether this job is a hyperparameter tuning job. - IsHyperparameterTuningJob bool `protobuf:"varint,4,opt,name=is_hyperparameter_tuning_job,json=isHyperparameterTuningJob" json:"is_hyperparameter_tuning_job,omitempty"` + IsHyperparameterTuningJob bool `protobuf:"varint,4,opt,name=is_hyperparameter_tuning_job,json=isHyperparameterTuningJob,proto3" json:"is_hyperparameter_tuning_job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -922,23 +922,23 @@ type PredictionInput struct { // *PredictionInput_Uri ModelVersion isPredictionInput_ModelVersion `protobuf_oneof:"model_version"` // Required. The format of the input data files. - DataFormat PredictionInput_DataFormat `protobuf:"varint,3,opt,name=data_format,json=dataFormat,enum=google.cloud.ml.v1.PredictionInput_DataFormat" json:"data_format,omitempty"` + DataFormat PredictionInput_DataFormat `protobuf:"varint,3,opt,name=data_format,json=dataFormat,proto3,enum=google.cloud.ml.v1.PredictionInput_DataFormat" json:"data_format,omitempty"` // Required. The Google Cloud Storage location of the input data files. // May contain wildcards. - InputPaths []string `protobuf:"bytes,4,rep,name=input_paths,json=inputPaths" json:"input_paths,omitempty"` + InputPaths []string `protobuf:"bytes,4,rep,name=input_paths,json=inputPaths,proto3" json:"input_paths,omitempty"` // Required. The output Google Cloud Storage location. - OutputPath string `protobuf:"bytes,5,opt,name=output_path,json=outputPath" json:"output_path,omitempty"` + OutputPath string `protobuf:"bytes,5,opt,name=output_path,json=outputPath,proto3" json:"output_path,omitempty"` // Optional. The maximum number of workers to be used for parallel processing. // Defaults to 10 if not specified. - MaxWorkerCount int64 `protobuf:"varint,6,opt,name=max_worker_count,json=maxWorkerCount" json:"max_worker_count,omitempty"` + MaxWorkerCount int64 `protobuf:"varint,6,opt,name=max_worker_count,json=maxWorkerCount,proto3" json:"max_worker_count,omitempty"` // Required. The Google Compute Engine region to run the prediction job in. - Region string `protobuf:"bytes,7,opt,name=region" json:"region,omitempty"` + Region string `protobuf:"bytes,7,opt,name=region,proto3" json:"region,omitempty"` // Optional. The Google Cloud ML runtime version to use for this batch // prediction. If not set, Google Cloud ML will pick the runtime version used // during the CreateVersion request for this model version, or choose the // latest stable version when model version information is not available // such as when the model is specified by uri. - RuntimeVersion string `protobuf:"bytes,8,opt,name=runtime_version,json=runtimeVersion" json:"runtime_version,omitempty"` + RuntimeVersion string `protobuf:"bytes,8,opt,name=runtime_version,json=runtimeVersion,proto3" json:"runtime_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -973,13 +973,13 @@ type isPredictionInput_ModelVersion interface { } type PredictionInput_ModelName struct { - ModelName string `protobuf:"bytes,1,opt,name=model_name,json=modelName,oneof"` + ModelName string `protobuf:"bytes,1,opt,name=model_name,json=modelName,proto3,oneof"` } type PredictionInput_VersionName struct { - VersionName string `protobuf:"bytes,2,opt,name=version_name,json=versionName,oneof"` + VersionName string `protobuf:"bytes,2,opt,name=version_name,json=versionName,proto3,oneof"` } type PredictionInput_Uri struct { - Uri string `protobuf:"bytes,9,opt,name=uri,oneof"` + Uri string `protobuf:"bytes,9,opt,name=uri,proto3,oneof"` } func (*PredictionInput_ModelName) isPredictionInput_ModelVersion() {} @@ -1140,13 +1140,13 @@ func _PredictionInput_OneofSizer(msg proto.Message) (n int) { // Represents results of a prediction job. type PredictionOutput struct { // The output Google Cloud Storage location provided at the job creation time. - OutputPath string `protobuf:"bytes,1,opt,name=output_path,json=outputPath" json:"output_path,omitempty"` + OutputPath string `protobuf:"bytes,1,opt,name=output_path,json=outputPath,proto3" json:"output_path,omitempty"` // The number of generated predictions. - PredictionCount int64 `protobuf:"varint,2,opt,name=prediction_count,json=predictionCount" json:"prediction_count,omitempty"` + PredictionCount int64 `protobuf:"varint,2,opt,name=prediction_count,json=predictionCount,proto3" json:"prediction_count,omitempty"` // The number of data instances which resulted in errors. - ErrorCount int64 `protobuf:"varint,3,opt,name=error_count,json=errorCount" json:"error_count,omitempty"` + ErrorCount int64 `protobuf:"varint,3,opt,name=error_count,json=errorCount,proto3" json:"error_count,omitempty"` // Node hours used by the batch prediction job. - NodeHours float64 `protobuf:"fixed64,4,opt,name=node_hours,json=nodeHours" json:"node_hours,omitempty"` + NodeHours float64 `protobuf:"fixed64,4,opt,name=node_hours,json=nodeHours,proto3" json:"node_hours,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1207,7 +1207,7 @@ func (m *PredictionOutput) GetNodeHours() float64 { // Represents a training or prediction job. type Job struct { // Required. The user-specified id of the job. - JobId string `protobuf:"bytes,1,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,1,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` // Required. Parameters to create a job. // // Types that are valid to be assigned to Input: @@ -1215,15 +1215,15 @@ type Job struct { // *Job_PredictionInput Input isJob_Input `protobuf_oneof:"input"` // Output only. When the job was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. When the job processing was started. - StartTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Output only. When the job processing was completed. - EndTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Output only. The detailed state of a job. - State Job_State `protobuf:"varint,7,opt,name=state,enum=google.cloud.ml.v1.Job_State" json:"state,omitempty"` + State Job_State `protobuf:"varint,7,opt,name=state,proto3,enum=google.cloud.ml.v1.Job_State" json:"state,omitempty"` // Output only. The details of a failure or a cancellation. - ErrorMessage string `protobuf:"bytes,8,opt,name=error_message,json=errorMessage" json:"error_message,omitempty"` + ErrorMessage string `protobuf:"bytes,8,opt,name=error_message,json=errorMessage,proto3" json:"error_message,omitempty"` // Output only. The current result of the job. // // Types that are valid to be assigned to Output: @@ -1267,16 +1267,16 @@ type isJob_Output interface { } type Job_TrainingInput struct { - TrainingInput *TrainingInput `protobuf:"bytes,2,opt,name=training_input,json=trainingInput,oneof"` + TrainingInput *TrainingInput `protobuf:"bytes,2,opt,name=training_input,json=trainingInput,proto3,oneof"` } type Job_PredictionInput struct { - PredictionInput *PredictionInput `protobuf:"bytes,3,opt,name=prediction_input,json=predictionInput,oneof"` + PredictionInput *PredictionInput `protobuf:"bytes,3,opt,name=prediction_input,json=predictionInput,proto3,oneof"` } type Job_TrainingOutput struct { - TrainingOutput *TrainingOutput `protobuf:"bytes,9,opt,name=training_output,json=trainingOutput,oneof"` + TrainingOutput *TrainingOutput `protobuf:"bytes,9,opt,name=training_output,json=trainingOutput,proto3,oneof"` } type Job_PredictionOutput struct { - PredictionOutput *PredictionOutput `protobuf:"bytes,10,opt,name=prediction_output,json=predictionOutput,oneof"` + PredictionOutput *PredictionOutput `protobuf:"bytes,10,opt,name=prediction_output,json=predictionOutput,proto3,oneof"` } func (*Job_TrainingInput) isJob_Input() {} @@ -1496,9 +1496,9 @@ type CreateJobRequest struct { // Required. The project name. // // Authorization: requires `Editor` role on the specified project. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The job to create. - Job *Job `protobuf:"bytes,2,opt,name=job" json:"job,omitempty"` + Job *Job `protobuf:"bytes,2,opt,name=job,proto3" json:"job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1547,20 +1547,20 @@ type ListJobsRequest struct { // Required. The name of the project for which to list jobs. // // Authorization: requires `Viewer` role on the specified project. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. Specifies the subset of jobs to retrieve. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Optional. A page token to request the next page of results. // // You get the token from the `next_page_token` field of the response from // the previous call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. The number of jobs to retrieve per "page" of results. If there // are more remaining results than this number, the response message will // contain a valid value in the `next_page_token` field. // // The default value is 20, and the maximum page size is 100. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1621,10 +1621,10 @@ func (m *ListJobsRequest) GetPageSize() int32 { // Response message for the ListJobs method. type ListJobsResponse struct { // The list of jobs. - Jobs []*Job `protobuf:"bytes,1,rep,name=jobs" json:"jobs,omitempty"` + Jobs []*Job `protobuf:"bytes,1,rep,name=jobs,proto3" json:"jobs,omitempty"` // Optional. Pass this token as the `page_token` field of the request for a // subsequent call. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1673,7 +1673,7 @@ type GetJobRequest struct { // Required. The name of the job to get the description of. // // Authorization: requires `Viewer` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1715,7 +1715,7 @@ type CancelJobRequest struct { // Required. The name of the job to cancel. // // Authorization: requires `Editor` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1842,8 +1842,7 @@ func (c *jobServiceClient) CancelJob(ctx context.Context, in *CancelJobRequest, return out, nil } -// Server API for JobService service - +// JobServiceServer is the server API for JobService service. type JobServiceServer interface { // Creates a training or a batch prediction job. CreateJob(context.Context, *CreateJobRequest) (*Job, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/model_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/model_service.pb.go index f5882aeb2..f24943c1d 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/model_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/model_service.pb.go @@ -36,22 +36,22 @@ type Model struct { // Required. The name specified for the model when it was created. // // The model name must be unique within the project it is created in. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The description specified for the model when it was created. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Output only. The default version of the model. This version will be used to // handle prediction requests that do not specify a version. // // You can change the default version by calling // [projects.methods.versions.setDefault](/ml/reference/rest/v1/projects.models.versions/setDefault). - DefaultVersion *Version `protobuf:"bytes,3,opt,name=default_version,json=defaultVersion" json:"default_version,omitempty"` + DefaultVersion *Version `protobuf:"bytes,3,opt,name=default_version,json=defaultVersion,proto3" json:"default_version,omitempty"` // Optional. The list of regions where the model is going to be deployed. // Currently only one region per model is supported. // Defaults to 'us-central1' if nothing is set. - Regions []string `protobuf:"bytes,4,rep,name=regions" json:"regions,omitempty"` + Regions []string `protobuf:"bytes,4,rep,name=regions,proto3" json:"regions,omitempty"` // Optional. If true, enables StackDriver Logging for online prediction. // Default is false. - OnlinePredictionLogging bool `protobuf:"varint,5,opt,name=online_prediction_logging,json=onlinePredictionLogging" json:"online_prediction_logging,omitempty"` + OnlinePredictionLogging bool `protobuf:"varint,5,opt,name=online_prediction_logging,json=onlinePredictionLogging,proto3" json:"online_prediction_logging,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -126,15 +126,15 @@ type Version struct { // Required.The name specified for the version when it was created. // // The version name must be unique within the model it is created in. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. The description specified for the version when it was created. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Output only. If true, this version will be used to handle prediction // requests that do not specify a version. // // You can change the default version by calling // [projects.methods.versions.setDefault](/ml/reference/rest/v1/projects.models.versions/setDefault). - IsDefault bool `protobuf:"varint,3,opt,name=is_default,json=isDefault" json:"is_default,omitempty"` + IsDefault bool `protobuf:"varint,3,opt,name=is_default,json=isDefault,proto3" json:"is_default,omitempty"` // Required. The Google Cloud Storage location of the trained model used to // create the version. See the // [overview of model deployment](/ml/docs/concepts/deployment-overview) for @@ -145,21 +145,21 @@ type Version struct { // the model service uses the specified location as the source of the model. // Once deployed, the model version is hosted by the prediction service, so // this location is useful only as a historical record. - DeploymentUri string `protobuf:"bytes,4,opt,name=deployment_uri,json=deploymentUri" json:"deployment_uri,omitempty"` + DeploymentUri string `protobuf:"bytes,4,opt,name=deployment_uri,json=deploymentUri,proto3" json:"deployment_uri,omitempty"` // Output only. The time the version was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The time the version was last used for prediction. - LastUseTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=last_use_time,json=lastUseTime" json:"last_use_time,omitempty"` + LastUseTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=last_use_time,json=lastUseTime,proto3" json:"last_use_time,omitempty"` // Optional. The Google Cloud ML runtime version to use for this deployment. // If not set, Google Cloud ML will choose a version. - RuntimeVersion string `protobuf:"bytes,8,opt,name=runtime_version,json=runtimeVersion" json:"runtime_version,omitempty"` + RuntimeVersion string `protobuf:"bytes,8,opt,name=runtime_version,json=runtimeVersion,proto3" json:"runtime_version,omitempty"` // Optional. Manually select the number of nodes to use for serving the // model. If unset (i.e., by default), the number of nodes used to serve // the model automatically scales with traffic. However, care should be // taken to ramp up traffic according to the model's ability to scale. If // your model needs to handle bursts of traffic beyond it's ability to // scale, it is recommended you set this field appropriately. - ManualScaling *ManualScaling `protobuf:"bytes,9,opt,name=manual_scaling,json=manualScaling" json:"manual_scaling,omitempty"` + ManualScaling *ManualScaling `protobuf:"bytes,9,opt,name=manual_scaling,json=manualScaling,proto3" json:"manual_scaling,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -251,7 +251,7 @@ type ManualScaling struct { // starting from the time the model is deployed, so the cost of operating // this model will be proportional to nodes * number of hours since // deployment. - Nodes int32 `protobuf:"varint,1,opt,name=nodes" json:"nodes,omitempty"` + Nodes int32 `protobuf:"varint,1,opt,name=nodes,proto3" json:"nodes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -293,9 +293,9 @@ type CreateModelRequest struct { // Required. The project name. // // Authorization: requires `Editor` role on the specified project. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The model to create. - Model *Model `protobuf:"bytes,2,opt,name=model" json:"model,omitempty"` + Model *Model `protobuf:"bytes,2,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -344,18 +344,18 @@ type ListModelsRequest struct { // Required. The name of the project whose models are to be listed. // // Authorization: requires `Viewer` role on the specified project. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. A page token to request the next page of results. // // You get the token from the `next_page_token` field of the response from // the previous call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. The number of models to retrieve per "page" of results. If there // are more remaining results than this number, the response message will // contain a valid value in the `next_page_token` field. // // The default value is 20, and the maximum page size is 100. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,10 +409,10 @@ func (m *ListModelsRequest) GetPageSize() int32 { // Response message for the ListModels method. type ListModelsResponse struct { // The list of models. - Models []*Model `protobuf:"bytes,1,rep,name=models" json:"models,omitempty"` + Models []*Model `protobuf:"bytes,1,rep,name=models,proto3" json:"models,omitempty"` // Optional. Pass this token as the `page_token` field of the request for a // subsequent call. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -461,7 +461,7 @@ type GetModelRequest struct { // Required. The name of the model. // // Authorization: requires `Viewer` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -503,7 +503,7 @@ type DeleteModelRequest struct { // Required. The name of the model. // // Authorization: requires `Editor` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -545,9 +545,9 @@ type CreateVersionRequest struct { // Required. The name of the model. // // Authorization: requires `Editor` role on the parent project. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The version details. - Version *Version `protobuf:"bytes,2,opt,name=version" json:"version,omitempty"` + Version *Version `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -596,18 +596,18 @@ type ListVersionsRequest struct { // Required. The name of the model for which to list the version. // // Authorization: requires `Viewer` role on the parent project. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. A page token to request the next page of results. // // You get the token from the `next_page_token` field of the response from // the previous call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. The number of versions to retrieve per "page" of results. If // there are more remaining results than this number, the response message // will contain a valid value in the `next_page_token` field. // // The default value is 20, and the maximum page size is 100. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -661,10 +661,10 @@ func (m *ListVersionsRequest) GetPageSize() int32 { // Response message for the ListVersions method. type ListVersionsResponse struct { // The list of versions. - Versions []*Version `protobuf:"bytes,1,rep,name=versions" json:"versions,omitempty"` + Versions []*Version `protobuf:"bytes,1,rep,name=versions,proto3" json:"versions,omitempty"` // Optional. Pass this token as the `page_token` field of the request for a // subsequent call. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -713,7 +713,7 @@ type GetVersionRequest struct { // Required. The name of the version. // // Authorization: requires `Viewer` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -757,7 +757,7 @@ type DeleteVersionRequest struct { // [projects.models.versions.list](/ml/reference/rest/v1/projects.models.versions/list). // // Authorization: requires `Editor` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -801,7 +801,7 @@ type SetDefaultVersionRequest struct { // [projects.models.versions.list](/ml/reference/rest/v1/projects.models.versions/list). // // Authorization: requires `Editor` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1018,8 +1018,7 @@ func (c *modelServiceClient) SetDefaultVersion(ctx context.Context, in *SetDefau return out, nil } -// Server API for ModelService service - +// ModelServiceServer is the server API for ModelService service. type ModelServiceServer interface { // Creates a model which will later contain one or more versions. // diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/operation_metadata.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/operation_metadata.pb.go index 1c677ed5e..878276313 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/operation_metadata.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/operation_metadata.pb.go @@ -57,19 +57,19 @@ func (OperationMetadata_OperationType) EnumDescriptor() ([]byte, []int) { // Represents the metadata of the long-running operation. type OperationMetadata struct { // The time the operation was submitted. - CreateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The time operation processing started. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time operation processing completed. - EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Indicates whether a request to cancel this operation has been made. - IsCancellationRequested bool `protobuf:"varint,4,opt,name=is_cancellation_requested,json=isCancellationRequested" json:"is_cancellation_requested,omitempty"` + IsCancellationRequested bool `protobuf:"varint,4,opt,name=is_cancellation_requested,json=isCancellationRequested,proto3" json:"is_cancellation_requested,omitempty"` // The operation type. - OperationType OperationMetadata_OperationType `protobuf:"varint,5,opt,name=operation_type,json=operationType,enum=google.cloud.ml.v1.OperationMetadata_OperationType" json:"operation_type,omitempty"` + OperationType OperationMetadata_OperationType `protobuf:"varint,5,opt,name=operation_type,json=operationType,proto3,enum=google.cloud.ml.v1.OperationMetadata_OperationType" json:"operation_type,omitempty"` // Contains the name of the model associated with the operation. - ModelName string `protobuf:"bytes,6,opt,name=model_name,json=modelName" json:"model_name,omitempty"` + ModelName string `protobuf:"bytes,6,opt,name=model_name,json=modelName,proto3" json:"model_name,omitempty"` // Contains the version associated with the operation. - Version *Version `protobuf:"bytes,7,opt,name=version" json:"version,omitempty"` + Version *Version `protobuf:"bytes,7,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/prediction_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/prediction_service.pb.go index a6f416108..7d0d9b880 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/prediction_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/prediction_service.pb.go @@ -215,10 +215,10 @@ type PredictRequest struct { // Required. The resource name of a model or a version. // // Authorization: requires `Viewer` role on the parent project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // // Required. The prediction request body. - HttpBody *httpbody.HttpBody `protobuf:"bytes,2,opt,name=http_body,json=httpBody" json:"http_body,omitempty"` + HttpBody *httpbody.HttpBody `protobuf:"bytes,2,opt,name=http_body,json=httpBody,proto3" json:"http_body,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -301,8 +301,7 @@ func (c *onlinePredictionServiceClient) Predict(ctx context.Context, in *Predict return out, nil } -// Server API for OnlinePredictionService service - +// OnlinePredictionServiceServer is the server API for OnlinePredictionService service. type OnlinePredictionServiceServer interface { // Performs prediction on the data in the request. // diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/project_service.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/project_service.pb.go index 73bf2059b..5f34e8fad 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/project_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/ml/v1/project_service.pb.go @@ -29,7 +29,7 @@ type GetConfigRequest struct { // Required. The project name. // // Authorization: requires `Viewer` role on the specified project. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -69,9 +69,9 @@ func (m *GetConfigRequest) GetName() string { // Returns service account information associated with a project. type GetConfigResponse struct { // The service account Cloud ML uses to access resources in the project. - ServiceAccount string `protobuf:"bytes,1,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount string `protobuf:"bytes,1,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // The project number for `service_account`. - ServiceAccountProject int64 `protobuf:"varint,2,opt,name=service_account_project,json=serviceAccountProject" json:"service_account_project,omitempty"` + ServiceAccountProject int64 `protobuf:"varint,2,opt,name=service_account_project,json=serviceAccountProject,proto3" json:"service_account_project,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -156,8 +156,7 @@ func (c *projectManagementServiceClient) GetConfig(ctx context.Context, in *GetC return out, nil } -// Server API for ProjectManagementService service - +// ProjectManagementServiceServer is the server API for ProjectManagementService service. type ProjectManagementServiceServer interface { // Get the service account information associated with your project. You need // this information in order to grant the service account persmissions for diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/common/common.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/common/common.pb.go index d82b96cd1..cbe9c1201 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/common/common.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/common/common.pb.go @@ -22,24 +22,24 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // The POSIX account information associated with a Google account. type PosixAccount struct { // Only one POSIX account can be marked as primary. - Primary bool `protobuf:"varint,1,opt,name=primary" json:"primary,omitempty"` + Primary bool `protobuf:"varint,1,opt,name=primary,proto3" json:"primary,omitempty"` // The username of the POSIX account. - Username string `protobuf:"bytes,2,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,2,opt,name=username,proto3" json:"username,omitempty"` // The user ID. - Uid int64 `protobuf:"varint,3,opt,name=uid" json:"uid,omitempty"` + Uid int64 `protobuf:"varint,3,opt,name=uid,proto3" json:"uid,omitempty"` // The default group ID. - Gid int64 `protobuf:"varint,4,opt,name=gid" json:"gid,omitempty"` + Gid int64 `protobuf:"varint,4,opt,name=gid,proto3" json:"gid,omitempty"` // The path to the home directory for this account. - HomeDirectory string `protobuf:"bytes,5,opt,name=home_directory,json=homeDirectory" json:"home_directory,omitempty"` + HomeDirectory string `protobuf:"bytes,5,opt,name=home_directory,json=homeDirectory,proto3" json:"home_directory,omitempty"` // The path to the logic shell for this account. - Shell string `protobuf:"bytes,6,opt,name=shell" json:"shell,omitempty"` + Shell string `protobuf:"bytes,6,opt,name=shell,proto3" json:"shell,omitempty"` // The GECOS (user information) entry for this account. - Gecos string `protobuf:"bytes,7,opt,name=gecos" json:"gecos,omitempty"` + Gecos string `protobuf:"bytes,7,opt,name=gecos,proto3" json:"gecos,omitempty"` // System identifier for which account the username or uid applies to. // By default, the empty value is used. - SystemId string `protobuf:"bytes,8,opt,name=system_id,json=systemId" json:"system_id,omitempty"` + SystemId string `protobuf:"bytes,8,opt,name=system_id,json=systemId,proto3" json:"system_id,omitempty"` // Output only. A POSIX account identifier. - AccountId string `protobuf:"bytes,9,opt,name=account_id,json=accountId" json:"account_id,omitempty"` + AccountId string `protobuf:"bytes,9,opt,name=account_id,json=accountId,proto3" json:"account_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -137,11 +137,11 @@ type SshPublicKey struct { // Public key text in SSH format, defined by // RFC4253 // section 6.6. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // An expiration time in microseconds since epoch. - ExpirationTimeUsec int64 `protobuf:"varint,2,opt,name=expiration_time_usec,json=expirationTimeUsec" json:"expiration_time_usec,omitempty"` + ExpirationTimeUsec int64 `protobuf:"varint,2,opt,name=expiration_time_usec,json=expirationTimeUsec,proto3" json:"expiration_time_usec,omitempty"` // Output only. The SHA-256 fingerprint of the SSH public key. - Fingerprint string `protobuf:"bytes,3,opt,name=fingerprint" json:"fingerprint,omitempty"` + Fingerprint string `protobuf:"bytes,3,opt,name=fingerprint,proto3" json:"fingerprint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1/oslogin.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1/oslogin.pb.go index a39f8fa40..34cfe7963 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1/oslogin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1/oslogin.pb.go @@ -31,14 +31,14 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Google Compute Engine. type LoginProfile struct { // The primary email address that uniquely identifies the user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The list of POSIX accounts associated with the user. - PosixAccounts []*common.PosixAccount `protobuf:"bytes,2,rep,name=posix_accounts,json=posixAccounts" json:"posix_accounts,omitempty"` + PosixAccounts []*common.PosixAccount `protobuf:"bytes,2,rep,name=posix_accounts,json=posixAccounts,proto3" json:"posix_accounts,omitempty"` // A map from SSH public key fingerprint to the associated key object. - SshPublicKeys map[string]*common.SshPublicKey `protobuf:"bytes,3,rep,name=ssh_public_keys,json=sshPublicKeys" json:"ssh_public_keys,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + SshPublicKeys map[string]*common.SshPublicKey `protobuf:"bytes,3,rep,name=ssh_public_keys,json=sshPublicKeys,proto3" json:"ssh_public_keys,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Indicates if the user is suspended. A suspended user cannot log in but // their profile information is retained. - Suspended bool `protobuf:"varint,4,opt,name=suspended" json:"suspended,omitempty"` + Suspended bool `protobuf:"varint,4,opt,name=suspended,proto3" json:"suspended,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -101,7 +101,7 @@ type DeletePosixAccountRequest struct { // A reference to the POSIX account to update. POSIX accounts are identified // by the project ID they are associated with. A reference to the POSIX // account is in format `users/{user}/projects/{project}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -143,7 +143,7 @@ type DeleteSshPublicKeyRequest struct { // The fingerprint of the public key to update. Public keys are identified by // their SHA-256 fingerprint. The fingerprint of the public key is in format // `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -183,7 +183,7 @@ func (m *DeleteSshPublicKeyRequest) GetName() string { // A request message for retrieving the login profile information for a user. type GetLoginProfileRequest struct { // The unique ID for the user in format `users/{user}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -225,7 +225,7 @@ type GetSshPublicKeyRequest struct { // The fingerprint of the public key to retrieve. Public keys are identified // by their SHA-256 fingerprint. The fingerprint of the public key is in // format `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -265,11 +265,11 @@ func (m *GetSshPublicKeyRequest) GetName() string { // A request message for importing an SSH public key. type ImportSshPublicKeyRequest struct { // The unique ID for the user in format `users/{user}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The SSH public key and expiration time. - SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey" json:"ssh_public_key,omitempty"` + SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey,proto3" json:"ssh_public_key,omitempty"` // The project ID of the Google Cloud Platform project. - ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -323,7 +323,7 @@ func (m *ImportSshPublicKeyRequest) GetProjectId() string { // A response message for importing an SSH public key. type ImportSshPublicKeyResponse struct { // The login profile information for the user. - LoginProfile *LoginProfile `protobuf:"bytes,1,opt,name=login_profile,json=loginProfile" json:"login_profile,omitempty"` + LoginProfile *LoginProfile `protobuf:"bytes,1,opt,name=login_profile,json=loginProfile,proto3" json:"login_profile,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -365,11 +365,11 @@ type UpdateSshPublicKeyRequest struct { // The fingerprint of the public key to update. Public keys are identified by // their SHA-256 fingerprint. The fingerprint of the public key is in format // `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The SSH public key and expiration time. - SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey" json:"ssh_public_key,omitempty"` + SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey,proto3" json:"ssh_public_key,omitempty"` // Mask to control which fields get updated. Updates all if not present. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -524,8 +524,7 @@ func (c *osLoginServiceClient) UpdateSshPublicKey(ctx context.Context, in *Updat return out, nil } -// Server API for OsLoginService service - +// OsLoginServiceServer is the server API for OsLoginService service. type OsLoginServiceServer interface { // Deletes a POSIX account. DeletePosixAccount(context.Context, *DeletePosixAccountRequest) (*empty.Empty, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1alpha/oslogin.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1alpha/oslogin.pb.go index 1d0b7afee..1fb92423c 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1alpha/oslogin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1alpha/oslogin.pb.go @@ -31,13 +31,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Google Compute Engine. type LoginProfile struct { // A unique user ID for identifying the user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The list of POSIX accounts associated with the Directory API user. - PosixAccounts []*common.PosixAccount `protobuf:"bytes,2,rep,name=posix_accounts,json=posixAccounts" json:"posix_accounts,omitempty"` + PosixAccounts []*common.PosixAccount `protobuf:"bytes,2,rep,name=posix_accounts,json=posixAccounts,proto3" json:"posix_accounts,omitempty"` // A map from SSH public key fingerprint to the associated key object. - SshPublicKeys map[string]*common.SshPublicKey `protobuf:"bytes,3,rep,name=ssh_public_keys,json=sshPublicKeys" json:"ssh_public_keys,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + SshPublicKeys map[string]*common.SshPublicKey `protobuf:"bytes,3,rep,name=ssh_public_keys,json=sshPublicKeys,proto3" json:"ssh_public_keys,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Indicates if the user is suspended. - Suspended bool `protobuf:"varint,4,opt,name=suspended" json:"suspended,omitempty"` + Suspended bool `protobuf:"varint,4,opt,name=suspended,proto3" json:"suspended,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -100,7 +100,7 @@ type DeletePosixAccountRequest struct { // A reference to the POSIX account to update. POSIX accounts are identified // by the project ID they are associated with. A reference to the POSIX // account is in format `users/{user}/projects/{project}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -142,7 +142,7 @@ type DeleteSshPublicKeyRequest struct { // The fingerprint of the public key to update. Public keys are identified by // their SHA-256 fingerprint. The fingerprint of the public key is in format // `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -182,7 +182,7 @@ func (m *DeleteSshPublicKeyRequest) GetName() string { // A request message for retrieving the login profile information for a user. type GetLoginProfileRequest struct { // The unique ID for the user in format `users/{user}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -224,7 +224,7 @@ type GetSshPublicKeyRequest struct { // The fingerprint of the public key to retrieve. Public keys are identified // by their SHA-256 fingerprint. The fingerprint of the public key is in // format `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -264,11 +264,11 @@ func (m *GetSshPublicKeyRequest) GetName() string { // A request message for importing an SSH public key. type ImportSshPublicKeyRequest struct { // The unique ID for the user in format `users/{user}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The SSH public key and expiration time. - SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey" json:"ssh_public_key,omitempty"` + SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey,proto3" json:"ssh_public_key,omitempty"` // The project ID of the Google Cloud Platform project. - ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -322,7 +322,7 @@ func (m *ImportSshPublicKeyRequest) GetProjectId() string { // A response message for importing an SSH public key. type ImportSshPublicKeyResponse struct { // The login profile information for the user. - LoginProfile *LoginProfile `protobuf:"bytes,1,opt,name=login_profile,json=loginProfile" json:"login_profile,omitempty"` + LoginProfile *LoginProfile `protobuf:"bytes,1,opt,name=login_profile,json=loginProfile,proto3" json:"login_profile,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -364,11 +364,11 @@ type UpdateSshPublicKeyRequest struct { // The fingerprint of the public key to update. Public keys are identified by // their SHA-256 fingerprint. The fingerprint of the public key is in format // `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The SSH public key and expiration time. - SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey" json:"ssh_public_key,omitempty"` + SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey,proto3" json:"ssh_public_key,omitempty"` // Mask to control which fields get updated. Updates all if not present. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -523,8 +523,7 @@ func (c *osLoginServiceClient) UpdateSshPublicKey(ctx context.Context, in *Updat return out, nil } -// Server API for OsLoginService service - +// OsLoginServiceServer is the server API for OsLoginService service. type OsLoginServiceServer interface { // Deletes a POSIX account. DeletePosixAccount(context.Context, *DeletePosixAccountRequest) (*empty.Empty, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1beta/oslogin.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1beta/oslogin.pb.go index b072f6326..998f7569d 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1beta/oslogin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/oslogin/v1beta/oslogin.pb.go @@ -31,14 +31,14 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Google Compute Engine. type LoginProfile struct { // The primary email address that uniquely identifies the user. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The list of POSIX accounts associated with the user. - PosixAccounts []*common.PosixAccount `protobuf:"bytes,2,rep,name=posix_accounts,json=posixAccounts" json:"posix_accounts,omitempty"` + PosixAccounts []*common.PosixAccount `protobuf:"bytes,2,rep,name=posix_accounts,json=posixAccounts,proto3" json:"posix_accounts,omitempty"` // A map from SSH public key fingerprint to the associated key object. - SshPublicKeys map[string]*common.SshPublicKey `protobuf:"bytes,3,rep,name=ssh_public_keys,json=sshPublicKeys" json:"ssh_public_keys,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + SshPublicKeys map[string]*common.SshPublicKey `protobuf:"bytes,3,rep,name=ssh_public_keys,json=sshPublicKeys,proto3" json:"ssh_public_keys,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Indicates if the user is suspended. A suspended user cannot log in but // their profile information is retained. - Suspended bool `protobuf:"varint,4,opt,name=suspended" json:"suspended,omitempty"` + Suspended bool `protobuf:"varint,4,opt,name=suspended,proto3" json:"suspended,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -101,7 +101,7 @@ type DeletePosixAccountRequest struct { // A reference to the POSIX account to update. POSIX accounts are identified // by the project ID they are associated with. A reference to the POSIX // account is in format `users/{user}/projects/{project}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -143,7 +143,7 @@ type DeleteSshPublicKeyRequest struct { // The fingerprint of the public key to update. Public keys are identified by // their SHA-256 fingerprint. The fingerprint of the public key is in format // `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -183,7 +183,7 @@ func (m *DeleteSshPublicKeyRequest) GetName() string { // A request message for retrieving the login profile information for a user. type GetLoginProfileRequest struct { // The unique ID for the user in format `users/{user}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -225,7 +225,7 @@ type GetSshPublicKeyRequest struct { // The fingerprint of the public key to retrieve. Public keys are identified // by their SHA-256 fingerprint. The fingerprint of the public key is in // format `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -265,11 +265,11 @@ func (m *GetSshPublicKeyRequest) GetName() string { // A request message for importing an SSH public key. type ImportSshPublicKeyRequest struct { // The unique ID for the user in format `users/{user}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The SSH public key and expiration time. - SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey" json:"ssh_public_key,omitempty"` + SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey,proto3" json:"ssh_public_key,omitempty"` // The project ID of the Google Cloud Platform project. - ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -323,7 +323,7 @@ func (m *ImportSshPublicKeyRequest) GetProjectId() string { // A response message for importing an SSH public key. type ImportSshPublicKeyResponse struct { // The login profile information for the user. - LoginProfile *LoginProfile `protobuf:"bytes,1,opt,name=login_profile,json=loginProfile" json:"login_profile,omitempty"` + LoginProfile *LoginProfile `protobuf:"bytes,1,opt,name=login_profile,json=loginProfile,proto3" json:"login_profile,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -365,11 +365,11 @@ type UpdateSshPublicKeyRequest struct { // The fingerprint of the public key to update. Public keys are identified by // their SHA-256 fingerprint. The fingerprint of the public key is in format // `users/{user}/sshPublicKeys/{fingerprint}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The SSH public key and expiration time. - SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey" json:"ssh_public_key,omitempty"` + SshPublicKey *common.SshPublicKey `protobuf:"bytes,2,opt,name=ssh_public_key,json=sshPublicKey,proto3" json:"ssh_public_key,omitempty"` // Mask to control which fields get updated. Updates all if not present. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -524,8 +524,7 @@ func (c *osLoginServiceClient) UpdateSshPublicKey(ctx context.Context, in *Updat return out, nil } -// Server API for OsLoginService service - +// OsLoginServiceServer is the server API for OsLoginService service. type OsLoginServiceServer interface { // Deletes a POSIX account. DeletePosixAccount(context.Context, *DeletePosixAccountRequest) (*empty.Empty, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/redis/v1beta1/cloud_redis.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/redis/v1beta1/cloud_redis.pb.go index df758d3b3..1b2108c41 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/redis/v1beta1/cloud_redis.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/redis/v1beta1/cloud_redis.pb.go @@ -117,62 +117,62 @@ type Instance struct { // specific zone (or collection of zones for cross-zone instances) an instance // should be provisioned in. Refer to [location_id] and // [alternative_location_id] fields for more details. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // An arbitrary and optional user-provided name for the instance. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Resource labels to represent user provided metadata - Labels map[string]string `protobuf:"bytes,3,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The zone where the instance will be provisioned. If not provided, // the service will choose a zone for the instance. For STANDARD_HA tier, // instances will be created across two zones for protection against zonal // failures. if [alternative_location_id] is also provided, it must be // different from [location_id]. - LocationId string `protobuf:"bytes,4,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,4,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` // Optional. Only applicable to STANDARD_HA tier which protects the instance // against zonal failures by provisioning it across two zones. If provided, it // must be a different zone from the one provided in [location_id]. - AlternativeLocationId string `protobuf:"bytes,5,opt,name=alternative_location_id,json=alternativeLocationId" json:"alternative_location_id,omitempty"` + AlternativeLocationId string `protobuf:"bytes,5,opt,name=alternative_location_id,json=alternativeLocationId,proto3" json:"alternative_location_id,omitempty"` // Optional. The version of Redis software. // If not provided, latest supported version will be used. - RedisVersion string `protobuf:"bytes,7,opt,name=redis_version,json=redisVersion" json:"redis_version,omitempty"` + RedisVersion string `protobuf:"bytes,7,opt,name=redis_version,json=redisVersion,proto3" json:"redis_version,omitempty"` // Optional. The CIDR range of internal addresses that are reserved for this // instance. If not provided, the service will choose an unused /29 block, // for example, 10.0.0.0/29 or 192.168.0.0/29. Ranges must be unique // and non-overlapping with existing subnets in a network. - ReservedIpRange string `protobuf:"bytes,9,opt,name=reserved_ip_range,json=reservedIpRange" json:"reserved_ip_range,omitempty"` + ReservedIpRange string `protobuf:"bytes,9,opt,name=reserved_ip_range,json=reservedIpRange,proto3" json:"reserved_ip_range,omitempty"` // Output only. Hostname or IP address of the exposed redis endpoint used by // clients to connect to the service. - Host string `protobuf:"bytes,10,opt,name=host" json:"host,omitempty"` + Host string `protobuf:"bytes,10,opt,name=host,proto3" json:"host,omitempty"` // Output only. The port number of the exposed redis endpoint. - Port int32 `protobuf:"varint,11,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,11,opt,name=port,proto3" json:"port,omitempty"` // Output only. The current zone where the Redis endpoint is placed. In // single zone deployments, this will always be the same as [location_id] // provided by the user at creation time. In cross-zone instances (only // applicable in STANDARD_HA tier), this can be either [location_id] or // [alternative_location_id] and can change on a failover event. - CurrentLocationId string `protobuf:"bytes,12,opt,name=current_location_id,json=currentLocationId" json:"current_location_id,omitempty"` + CurrentLocationId string `protobuf:"bytes,12,opt,name=current_location_id,json=currentLocationId,proto3" json:"current_location_id,omitempty"` // Output only. The time the instance was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,13,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,13,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The current state of this instance. - State Instance_State `protobuf:"varint,14,opt,name=state,enum=google.cloud.redis.v1beta1.Instance_State" json:"state,omitempty"` + State Instance_State `protobuf:"varint,14,opt,name=state,proto3,enum=google.cloud.redis.v1beta1.Instance_State" json:"state,omitempty"` // Output only. Additional information about the current status of this // instance, if available. - StatusMessage string `protobuf:"bytes,15,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,15,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // Optional. Redis configuration parameters, according to // http://redis.io/topics/config. Currently, the only supported parameters // are: // * maxmemory-policy // * notify-keyspace-events - RedisConfigs map[string]string `protobuf:"bytes,16,rep,name=redis_configs,json=redisConfigs" json:"redis_configs,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + RedisConfigs map[string]string `protobuf:"bytes,16,rep,name=redis_configs,json=redisConfigs,proto3" json:"redis_configs,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Required. The service tier of the instance. - Tier Instance_Tier `protobuf:"varint,17,opt,name=tier,enum=google.cloud.redis.v1beta1.Instance_Tier" json:"tier,omitempty"` + Tier Instance_Tier `protobuf:"varint,17,opt,name=tier,proto3,enum=google.cloud.redis.v1beta1.Instance_Tier" json:"tier,omitempty"` // Required. Redis memory size in GB. - MemorySizeGb int32 `protobuf:"varint,18,opt,name=memory_size_gb,json=memorySizeGb" json:"memory_size_gb,omitempty"` + MemorySizeGb int32 `protobuf:"varint,18,opt,name=memory_size_gb,json=memorySizeGb,proto3" json:"memory_size_gb,omitempty"` // Optional. The full name of the Google Compute Engine // [network](/compute/docs/networks-and-firewalls#networks) to which the // instance is connected. If left unspecified, the `default` network // will be used. - AuthorizedNetwork string `protobuf:"bytes,20,opt,name=authorized_network,json=authorizedNetwork" json:"authorized_network,omitempty"` + AuthorizedNetwork string `protobuf:"bytes,20,opt,name=authorized_network,json=authorizedNetwork,proto3" json:"authorized_network,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -326,7 +326,7 @@ type ListInstancesRequest struct { // Required. The resource name of the instance location using the form: // `projects/{project_id}/locations/{location_id}` // where `location_id` refers to a GCP region - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The maximum number of items to return. // // If not specified, a default value of 1000 will be used by the service. @@ -334,10 +334,10 @@ type ListInstancesRequest struct { // and a caller should only rely on response's // [next_page_token][CloudRedis.ListInstancesResponse.next_page_token] // to determine if there are more instances left to be queried. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The next_page_token value returned from a previous List request, // if any. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -400,10 +400,10 @@ type ListInstancesResponse struct { // the form projects/{project_id}/locations/{location_id}/instances/- and the // "status" field set to ERROR and "status_message" field set to "location not // available for ListInstances". - Instances []*Instance `protobuf:"bytes,1,rep,name=instances" json:"instances,omitempty"` + Instances []*Instance `protobuf:"bytes,1,rep,name=instances,proto3" json:"instances,omitempty"` // Token to retrieve the next page of results, or empty if there are no more // results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -452,7 +452,7 @@ type GetInstanceRequest struct { // Required. Redis instance resource name using the form: // `projects/{project_id}/locations/{location_id}/instances/{instance_id}` // where `location_id` refers to a GCP region - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -494,7 +494,7 @@ type CreateInstanceRequest struct { // Required. The resource name of the instance location using the form: // `projects/{project_id}/locations/{location_id}` // where `location_id` refers to a GCP region - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The logical name of the Redis instance in the customer project // with the following restrictions: // @@ -503,9 +503,9 @@ type CreateInstanceRequest struct { // * Must be between 1-40 characters. // * Must end with a number or a letter. // * Must be unique within the customer project / location - InstanceId string `protobuf:"bytes,2,opt,name=instance_id,json=instanceId" json:"instance_id,omitempty"` + InstanceId string `protobuf:"bytes,2,opt,name=instance_id,json=instanceId,proto3" json:"instance_id,omitempty"` // Required. A Redis [Instance] resource - Instance *Instance `protobuf:"bytes,3,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,3,opt,name=instance,proto3" json:"instance,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -565,10 +565,10 @@ type UpdateInstanceRequest struct { // * `labels` // * `memory_size_gb` // * `redis_config` - UpdateMask *field_mask.FieldMask `protobuf:"bytes,1,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,1,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Required. Update description. // Only fields specified in update_mask are updated. - Instance *Instance `protobuf:"bytes,2,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,2,opt,name=instance,proto3" json:"instance,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -617,7 +617,7 @@ type DeleteInstanceRequest struct { // Required. Redis instance resource name using the form: // `projects/{project_id}/locations/{location_id}/instances/{instance_id}` // where `location_id` refers to a GCP region - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -663,7 +663,7 @@ type LocationMetadata struct { // by the lowercase ID of each zone, as defined by GCE. These keys can be // specified in `location_id` or `alternative_location_id` fields when // creating a Redis instance. - AvailableZones map[string]*ZoneMetadata `protobuf:"bytes,1,rep,name=available_zones,json=availableZones" json:"available_zones,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + AvailableZones map[string]*ZoneMetadata `protobuf:"bytes,1,rep,name=available_zones,json=availableZones,proto3" json:"available_zones,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -849,8 +849,7 @@ func (c *cloudRedisClient) DeleteInstance(ctx context.Context, in *DeleteInstanc return out, nil } -// Server API for CloudRedis service - +// CloudRedisServer is the server API for CloudRedis service. type CloudRedisServer interface { // Lists all Redis instances owned by a project in either the specified // location (region) or all locations. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/resourcemanager/v2/folders.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/resourcemanager/v2/folders.pb.go index 8044e7abf..559035f78 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/resourcemanager/v2/folders.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/resourcemanager/v2/folders.pb.go @@ -93,10 +93,10 @@ func (FolderOperation_OperationType) EnumDescriptor() ([]byte, []int) { type Folder struct { // Output only. The resource name of the Folder. // Its format is `folders/{folder_id}`, for example: "folders/1234". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The Folder’s parent's resource name. // Updates to the folder's parent must be performed via [MoveFolders]. - Parent string `protobuf:"bytes,2,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,2,opt,name=parent,proto3" json:"parent,omitempty"` // The folder’s display name. // A folder’s display name must be unique amongst its siblings, e.g. // no two folders with the same parent can share the same display name. @@ -104,15 +104,15 @@ type Folder struct { // letters, digits, spaces, hyphens and underscores and can be no longer // than 30 characters. This is captured by the regular expression: // [\p{L}\p{N}]({\p{L}\p{N}_- ]{0,28}[\p{L}\p{N}])?. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Output only. The lifecycle state of the folder. // Updates to the lifecycle_state must be performed via // [DeleteFolder] and [UndeleteFolder]. - LifecycleState Folder_LifecycleState `protobuf:"varint,4,opt,name=lifecycle_state,json=lifecycleState,enum=google.cloud.resourcemanager.v2.Folder_LifecycleState" json:"lifecycle_state,omitempty"` + LifecycleState Folder_LifecycleState `protobuf:"varint,4,opt,name=lifecycle_state,json=lifecycleState,proto3,enum=google.cloud.resourcemanager.v2.Folder_LifecycleState" json:"lifecycle_state,omitempty"` // Output only. Timestamp when the Folder was created. Assigned by the server. - CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. Timestamp when the Folder was last modified. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -191,17 +191,17 @@ type ListFoldersRequest struct { // Must be of the form `folders/{folder_id}` or `organizations/{org_id}`. // Access to this method is controlled by checking the // `resourcemanager.folders.list` permission on the `parent`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The maximum number of Folders to return in the response. // This field is optional. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A pagination token returned from a previous call to `ListFolders` // that indicates where this listing should continue from. // This field is optional. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Controls whether Folders in the [DELETE_REQUESTED} state should // be returned. - ShowDeleted bool `protobuf:"varint,4,opt,name=show_deleted,json=showDeleted" json:"show_deleted,omitempty"` + ShowDeleted bool `protobuf:"varint,4,opt,name=show_deleted,json=showDeleted,proto3" json:"show_deleted,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -263,11 +263,11 @@ func (m *ListFoldersRequest) GetShowDeleted() bool { type ListFoldersResponse struct { // A possibly paginated list of Folders that are direct descendants of // the specified parent resource. - Folders []*Folder `protobuf:"bytes,1,rep,name=folders" json:"folders,omitempty"` + Folders []*Folder `protobuf:"bytes,1,rep,name=folders,proto3" json:"folders,omitempty"` // A pagination token returned from a previous call to `ListFolders` // that indicates from where listing should continue. // This field is optional. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -315,11 +315,11 @@ func (m *ListFoldersResponse) GetNextPageToken() string { type SearchFoldersRequest struct { // The maximum number of folders to return in the response. // This field is optional. - PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A pagination token returned from a previous call to `SearchFolders` // that indicates from where search should continue. // This field is optional. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Search criteria used to select the Folders to return. // If no search criteria is specified then all accessible folders will be // returned. @@ -336,7 +336,7 @@ type SearchFoldersRequest struct { // |parent=folders/123|Folders whose parent is "folders/123".| // |parent=folders/123 AND lifecycleState=ACTIVE|Active folders whose // parent is "folders/123".| - Query string `protobuf:"bytes,3,opt,name=query" json:"query,omitempty"` + Query string `protobuf:"bytes,3,opt,name=query,proto3" json:"query,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -391,11 +391,11 @@ func (m *SearchFoldersRequest) GetQuery() string { type SearchFoldersResponse struct { // A possibly paginated folder search results. // the specified parent resource. - Folders []*Folder `protobuf:"bytes,1,rep,name=folders" json:"folders,omitempty"` + Folders []*Folder `protobuf:"bytes,1,rep,name=folders,proto3" json:"folders,omitempty"` // A pagination token returned from a previous call to `SearchFolders` // that indicates from where searching should continue. // This field is optional. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -443,7 +443,7 @@ func (m *SearchFoldersResponse) GetNextPageToken() string { type GetFolderRequest struct { // The resource name of the Folder to retrieve. // Must be of the form `folders/{folder_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -484,10 +484,10 @@ func (m *GetFolderRequest) GetName() string { type CreateFolderRequest struct { // The resource name of the new Folder's parent. // Must be of the form `folders/{folder_id}` or `organizations/{org_id}`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The Folder being created, only the display name will be consulted. // All other fields will be ignored. - Folder *Folder `protobuf:"bytes,2,opt,name=folder" json:"folder,omitempty"` + Folder *Folder `protobuf:"bytes,2,opt,name=folder,proto3" json:"folder,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -535,11 +535,11 @@ func (m *CreateFolderRequest) GetFolder() *Folder { type MoveFolderRequest struct { // The resource name of the Folder to move. // Must be of the form folders/{folder_id} - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The resource name of the Folder or Organization to reparent // the folder under. // Must be of the form `folders/{folder_id}` or `organizations/{org_id}`. - DestinationParent string `protobuf:"bytes,2,opt,name=destination_parent,json=destinationParent" json:"destination_parent,omitempty"` + DestinationParent string `protobuf:"bytes,2,opt,name=destination_parent,json=destinationParent,proto3" json:"destination_parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -588,10 +588,10 @@ type UpdateFolderRequest struct { // The new definition of the Folder. It must include a // a `name` and `display_name` field. The other fields // will be ignored. - Folder *Folder `protobuf:"bytes,1,opt,name=folder" json:"folder,omitempty"` + Folder *Folder `protobuf:"bytes,1,opt,name=folder,proto3" json:"folder,omitempty"` // Fields to be updated. // Only the `display_name` can be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -639,10 +639,10 @@ func (m *UpdateFolderRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteFolderRequest struct { // the resource name of the Folder to be deleted. // Must be of the form `folders/{folder_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Instructs DeleteFolderAction to delete a folder even when the folder is not // empty. - RecursiveDelete bool `protobuf:"varint,2,opt,name=recursive_delete,json=recursiveDelete" json:"recursive_delete,omitempty"` + RecursiveDelete bool `protobuf:"varint,2,opt,name=recursive_delete,json=recursiveDelete,proto3" json:"recursive_delete,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -690,7 +690,7 @@ func (m *DeleteFolderRequest) GetRecursiveDelete() bool { type UndeleteFolderRequest struct { // The resource name of the Folder to undelete. // Must be of the form `folders/{folder_id}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -730,15 +730,15 @@ func (m *UndeleteFolderRequest) GetName() string { // Metadata describing a long running folder operation type FolderOperation struct { // The display name of the folder. - DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The type of this operation. - OperationType FolderOperation_OperationType `protobuf:"varint,2,opt,name=operation_type,json=operationType,enum=google.cloud.resourcemanager.v2.FolderOperation_OperationType" json:"operation_type,omitempty"` + OperationType FolderOperation_OperationType `protobuf:"varint,2,opt,name=operation_type,json=operationType,proto3,enum=google.cloud.resourcemanager.v2.FolderOperation_OperationType" json:"operation_type,omitempty"` // The resource name of the folder's parent. // Only applicable when the operation_type is MOVE. - SourceParent string `protobuf:"bytes,3,opt,name=source_parent,json=sourceParent" json:"source_parent,omitempty"` + SourceParent string `protobuf:"bytes,3,opt,name=source_parent,json=sourceParent,proto3" json:"source_parent,omitempty"` // The resource name of the folder or organization we are either creating // the folder under or moving the folder to. - DestinationParent string `protobuf:"bytes,4,opt,name=destination_parent,json=destinationParent" json:"destination_parent,omitempty"` + DestinationParent string `protobuf:"bytes,4,opt,name=destination_parent,json=destinationParent,proto3" json:"destination_parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1049,8 +1049,7 @@ func (c *foldersClient) TestIamPermissions(ctx context.Context, in *v1.TestIamPe return out, nil } -// Server API for Folders service - +// FoldersServer is the server API for Folders service. type FoldersServer interface { // Lists the Folders that are direct descendants of supplied parent resource. // List provides a strongly consistent view of the Folders underneath diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/resources.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/resources.pb.go index 2def256e6..8e5647b98 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/resources.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/resources.pb.go @@ -69,9 +69,9 @@ type RuntimeConfig struct { // You pick the RuntimeConfig resource name, but the server will validate that // the name adheres to this format. After you create the resource, you cannot // change the resource's name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // An optional description of the RuntimeConfig object. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -137,7 +137,7 @@ type Variable struct { // The length of a `[VARIABLE_NAME]` must be less than 256 bytes. // // Once you create a variable, you cannot change the variable name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The the value of the variable. It can be either a binary or a string // value. You must specify one of either `value` or `text`. Specifying both // will cause the server to return an error. @@ -147,11 +147,11 @@ type Variable struct { // *Variable_Text Contents isVariable_Contents `protobuf_oneof:"contents"` // [Output Only] The time of the last variable update. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // [Ouput only] The current state of the variable. The variable state indicates // the outcome of the `variables().watch` call and is visible through the // `get` and `list` calls. - State VariableState `protobuf:"varint,4,opt,name=state,enum=google.cloud.runtimeconfig.v1beta1.VariableState" json:"state,omitempty"` + State VariableState `protobuf:"varint,4,opt,name=state,proto3,enum=google.cloud.runtimeconfig.v1beta1.VariableState" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -189,7 +189,7 @@ type Variable_Value struct { Value []byte `protobuf:"bytes,2,opt,name=value,proto3,oneof"` } type Variable_Text struct { - Text string `protobuf:"bytes,5,opt,name=text,oneof"` + Text string `protobuf:"bytes,5,opt,name=text,proto3,oneof"` } func (*Variable_Value) isVariable_Contents() {} @@ -345,7 +345,7 @@ type isEndCondition_Condition interface { } type EndCondition_Cardinality_ struct { - Cardinality *EndCondition_Cardinality `protobuf:"bytes,1,opt,name=cardinality,oneof"` + Cardinality *EndCondition_Cardinality `protobuf:"bytes,1,opt,name=cardinality,proto3,oneof"` } func (*EndCondition_Cardinality_) isEndCondition_Condition() {} @@ -435,10 +435,10 @@ func _EndCondition_OneofSizer(msg proto.Message) (n int) { // path prefix are counted. type EndCondition_Cardinality struct { // The root of the variable subtree to monitor. For example, `/foo`. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // The number variables under the `path` that must exist to meet this // condition. Defaults to 1 if not specified. - Number int32 `protobuf:"varint,2,opt,name=number" json:"number,omitempty"` + Number int32 `protobuf:"varint,2,opt,name=number,proto3" json:"number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -508,36 +508,36 @@ type Waiter struct { // of `[WAITER_NAME]` must be less than 64 bytes. // // After you create a Waiter resource, you cannot change the resource name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // [Required] Specifies the timeout of the waiter in seconds, beginning from // the instant that `waiters().create` method is called. If this time elapses // before the success or failure conditions are met, the waiter fails and sets // the `error` code to `DEADLINE_EXCEEDED`. - Timeout *duration.Duration `protobuf:"bytes,2,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,2,opt,name=timeout,proto3" json:"timeout,omitempty"` // [Optional] The failure condition of this waiter. If this condition is met, // `done` will be set to `true` and the `error` code will be set to `ABORTED`. // The failure condition takes precedence over the success condition. If both // conditions are met, a failure will be indicated. This value is optional; if // no failure condition is set, the only failure scenario will be a timeout. - Failure *EndCondition `protobuf:"bytes,3,opt,name=failure" json:"failure,omitempty"` + Failure *EndCondition `protobuf:"bytes,3,opt,name=failure,proto3" json:"failure,omitempty"` // [Required] The success condition. If this condition is met, `done` will be // set to `true` and the `error` value will remain unset. The failure condition // takes precedence over the success condition. If both conditions are met, a // failure will be indicated. - Success *EndCondition `protobuf:"bytes,4,opt,name=success" json:"success,omitempty"` + Success *EndCondition `protobuf:"bytes,4,opt,name=success,proto3" json:"success,omitempty"` // [Output Only] The instant at which this Waiter resource was created. Adding // the value of `timeout` to this instant yields the timeout deadline for the // waiter. - CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // [Output Only] If the value is `false`, it means the waiter is still waiting // for one of its conditions to be met. // // If true, the waiter has finished. If the waiter finished due to a timeout // or failure, `error` will be set. - Done bool `protobuf:"varint,6,opt,name=done" json:"done,omitempty"` + Done bool `protobuf:"varint,6,opt,name=done,proto3" json:"done,omitempty"` // [Output Only] If the waiter ended due to a failure or timeout, this value // will be set. - Error *status.Status `protobuf:"bytes,7,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,7,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/runtimeconfig.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/runtimeconfig.pb.go index 07037b3aa..13d71d4ad 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/runtimeconfig.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/runtimeconfig/v1beta1/runtimeconfig.pb.go @@ -31,13 +31,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type ListConfigsRequest struct { // The [project ID](https://support.google.com/cloud/answer/6158840?hl=en&ref_topic=6158848) // for this request, in the format `projects/[PROJECT_ID]`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Specifies the number of results to return per page. If there are fewer // elements than the specified number, returns all elements. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Specifies a page token to use. Set `pageToken` to a `nextPageToken` // returned by a previous list request to get the next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -93,13 +93,13 @@ func (m *ListConfigsRequest) GetPageToken() string { type ListConfigsResponse struct { // A list of the configurations in the project. The order of returned // objects is arbitrary; that is, it is not ordered in any particular way. - Configs []*RuntimeConfig `protobuf:"bytes,1,rep,name=configs" json:"configs,omitempty"` + Configs []*RuntimeConfig `protobuf:"bytes,1,rep,name=configs,proto3" json:"configs,omitempty"` // This token allows you to get the next page of results for list requests. // If the number of results is larger than `pageSize`, use the `nextPageToken` // as a value for the query parameter `pageToken` in the next list request. // Subsequent list requests will have their own `nextPageToken` to continue // paging through the results - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -148,7 +148,7 @@ type GetConfigRequest struct { // The name of the RuntimeConfig resource to retrieve, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -189,9 +189,9 @@ func (m *GetConfigRequest) GetName() string { type CreateConfigRequest struct { // The [project ID](https://support.google.com/cloud/answer/6158840?hl=en&ref_topic=6158848) // for this request, in the format `projects/[PROJECT_ID]`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The RuntimeConfig to create. - Config *RuntimeConfig `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"` + Config *RuntimeConfig `protobuf:"bytes,2,opt,name=config,proto3" json:"config,omitempty"` // An optional but recommended unique `request_id`. If the server // receives two `create()` requests with the same // `request_id`, then the second request will be ignored and the @@ -202,7 +202,7 @@ type CreateConfigRequest struct { // `request_id` strings. // // `request_id` strings are limited to 64 characters. - RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -258,9 +258,9 @@ type UpdateConfigRequest struct { // The name of the RuntimeConfig resource to update, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The config resource to update. - Config *RuntimeConfig `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"` + Config *RuntimeConfig `protobuf:"bytes,2,opt,name=config,proto3" json:"config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -309,7 +309,7 @@ type DeleteConfigRequest struct { // The RuntimeConfig resource to delete, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -352,21 +352,21 @@ type ListVariablesRequest struct { // The configuration must exist beforehand; the path must by in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Filters variables by matching the specified filter. For example: // // `projects/example-project/config/[CONFIG_NAME]/variables/example-variable`. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Specifies the number of results to return per page. If there are fewer // elements than the specified number, returns all elements. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Specifies a page token to use. Set `pageToken` to a `nextPageToken` // returned by a previous list request to get the next page of results. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The flag indicates whether the user wants to return values of variables. // If true, then only those variables that user has IAM GetVariable permission // will be returned along with their values. - ReturnValues bool `protobuf:"varint,5,opt,name=return_values,json=returnValues" json:"return_values,omitempty"` + ReturnValues bool `protobuf:"varint,5,opt,name=return_values,json=returnValues,proto3" json:"return_values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -435,13 +435,13 @@ func (m *ListVariablesRequest) GetReturnValues() bool { type ListVariablesResponse struct { // A list of variables and their values. The order of returned variable // objects is arbitrary. - Variables []*Variable `protobuf:"bytes,1,rep,name=variables" json:"variables,omitempty"` + Variables []*Variable `protobuf:"bytes,1,rep,name=variables,proto3" json:"variables,omitempty"` // This token allows you to get the next page of results for list requests. // If the number of results is larger than `pageSize`, use the `nextPageToken` // as a value for the query parameter `pageToken` in the next list request. // Subsequent list requests will have their own `nextPageToken` to continue // paging through the results - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -490,14 +490,14 @@ type WatchVariableRequest struct { // The name of the variable to watch, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // If specified, checks the current timestamp of the variable and if the // current timestamp is newer than `newerThan` timestamp, the method returns // immediately. // // If not specified or the variable has an older timestamp, the watcher waits // for a the value to change before returning. - NewerThan *timestamp.Timestamp `protobuf:"bytes,4,opt,name=newer_than,json=newerThan" json:"newer_than,omitempty"` + NewerThan *timestamp.Timestamp `protobuf:"bytes,4,opt,name=newer_than,json=newerThan,proto3" json:"newer_than,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -546,7 +546,7 @@ type GetVariableRequest struct { // The name of the variable to return, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]/variables/[VARIBLE_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -589,9 +589,9 @@ type CreateVariableRequest struct { // The configuration must exist beforehand; the path must by in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The variable to create. - Variable *Variable `protobuf:"bytes,2,opt,name=variable" json:"variable,omitempty"` + Variable *Variable `protobuf:"bytes,2,opt,name=variable,proto3" json:"variable,omitempty"` // An optional but recommended unique `request_id`. If the server // receives two `create()` requests with the same // `request_id`, then the second request will be ignored and the @@ -602,7 +602,7 @@ type CreateVariableRequest struct { // `request_id` strings. // // `request_id` strings are limited to 64 characters. - RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -658,9 +658,9 @@ type UpdateVariableRequest struct { // The name of the variable to update, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]/variables/[VARIABLE_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The variable to update. - Variable *Variable `protobuf:"bytes,2,opt,name=variable" json:"variable,omitempty"` + Variable *Variable `protobuf:"bytes,2,opt,name=variable,proto3" json:"variable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -709,10 +709,10 @@ type DeleteVariableRequest struct { // The name of the variable to delete, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]/variables/[VARIABLE_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Set to `true` to recursively delete multiple variables with the same // prefix. - Recursive bool `protobuf:"varint,2,opt,name=recursive" json:"recursive,omitempty"` + Recursive bool `protobuf:"varint,2,opt,name=recursive,proto3" json:"recursive,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -762,13 +762,13 @@ type ListWaitersRequest struct { // The configuration must exist beforehand; the path must by in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Specifies the number of results to return per page. If there are fewer // elements than the specified number, returns all elements. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Specifies a page token to use. Set `pageToken` to a `nextPageToken` // returned by a previous list request to get the next page of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -823,13 +823,13 @@ func (m *ListWaitersRequest) GetPageToken() string { // Order of returned waiter objects is arbitrary. type ListWaitersResponse struct { // Found waiters in the project. - Waiters []*Waiter `protobuf:"bytes,1,rep,name=waiters" json:"waiters,omitempty"` + Waiters []*Waiter `protobuf:"bytes,1,rep,name=waiters,proto3" json:"waiters,omitempty"` // This token allows you to get the next page of results for list requests. // If the number of results is larger than `pageSize`, use the `nextPageToken` // as a value for the query parameter `pageToken` in the next list request. // Subsequent list requests will have their own `nextPageToken` to continue // paging through the results - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -879,7 +879,7 @@ type GetWaiterRequest struct { // format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]/waiters/[WAITER_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -922,9 +922,9 @@ type CreateWaiterRequest struct { // The configuration must exist beforehand; the path must by in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The Waiter resource to create. - Waiter *Waiter `protobuf:"bytes,2,opt,name=waiter" json:"waiter,omitempty"` + Waiter *Waiter `protobuf:"bytes,2,opt,name=waiter,proto3" json:"waiter,omitempty"` // An optional but recommended unique `request_id`. If the server // receives two `create()` requests with the same // `request_id`, then the second request will be ignored and the @@ -935,7 +935,7 @@ type CreateWaiterRequest struct { // `request_id` strings. // // `request_id` strings are limited to 64 characters. - RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,3,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -991,7 +991,7 @@ type DeleteWaiterRequest struct { // The Waiter resource to delete, in the format: // // `projects/[PROJECT_ID]/configs/[CONFIG_NAME]/waiters/[WAITER_NAME]` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1267,8 +1267,7 @@ func (c *runtimeConfigManagerClient) DeleteWaiter(ctx context.Context, in *Delet return out, nil } -// Server API for RuntimeConfigManager service - +// RuntimeConfigManagerServer is the server API for RuntimeConfigManager service. type RuntimeConfigManagerServer interface { // Lists all the RuntimeConfig resources within project. ListConfigs(context.Context, *ListConfigsRequest) (*ListConfigsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1/cloud_speech.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1/cloud_speech.pb.go index e169b4895..0140ecf69 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1/cloud_speech.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1/cloud_speech.pb.go @@ -143,9 +143,9 @@ func (StreamingRecognizeResponse_SpeechEventType) EnumDescriptor() ([]byte, []in type RecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Required* The audio data to be recognized. - Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` + Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio,proto3" json:"audio,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -194,9 +194,9 @@ func (m *RecognizeRequest) GetAudio() *RecognitionAudio { type LongRunningRecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Required* The audio data to be recognized. - Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` + Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio,proto3" json:"audio,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -286,7 +286,7 @@ type isStreamingRecognizeRequest_StreamingRequest interface { } type StreamingRecognizeRequest_StreamingConfig struct { - StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,oneof"` + StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,proto3,oneof"` } type StreamingRecognizeRequest_AudioContent struct { AudioContent []byte `protobuf:"bytes,2,opt,name=audio_content,json=audioContent,proto3,oneof"` @@ -391,7 +391,7 @@ func _StreamingRecognizeRequest_OneofSizer(msg proto.Message) (n int) { type StreamingRecognitionConfig struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Optional* If `false` or omitted, the recognizer will perform continuous // recognition (continuing to wait for and process audio even if the user // pauses speaking) until the client closes the input stream (gRPC API) or @@ -403,12 +403,12 @@ type StreamingRecognitionConfig struct { // `END_OF_SINGLE_UTTERANCE` event and cease recognition. It will return no // more than one `StreamingRecognitionResult` with the `is_final` flag set to // `true`. - SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance" json:"single_utterance,omitempty"` + SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance,proto3" json:"single_utterance,omitempty"` // *Optional* If `true`, interim results (tentative hypotheses) may be // returned as they become available (these interim results are indicated with // the `is_final=false` flag). // If `false` or omitted, only `is_final=true` result(s) are returned. - InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults" json:"interim_results,omitempty"` + InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults,proto3" json:"interim_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -463,38 +463,38 @@ func (m *StreamingRecognitionConfig) GetInterimResults() bool { // request. type RecognitionConfig struct { // *Required* Encoding of audio data sent in all `RecognitionAudio` messages. - Encoding RecognitionConfig_AudioEncoding `protobuf:"varint,1,opt,name=encoding,enum=google.cloud.speech.v1.RecognitionConfig_AudioEncoding" json:"encoding,omitempty"` + Encoding RecognitionConfig_AudioEncoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.cloud.speech.v1.RecognitionConfig_AudioEncoding" json:"encoding,omitempty"` // *Required* Sample rate in Hertz of the audio data sent in all // `RecognitionAudio` messages. Valid values are: 8000-48000. // 16000 is optimal. For best results, set the sampling rate of the audio // source to 16000 Hz. If that's not possible, use the native sample rate of // the audio source (instead of re-sampling). - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` // *Required* The language of the supplied audio as a // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. // Example: "en-US". // See [Language Support](https://cloud.google.com/speech/docs/languages) // for a list of the currently supported language codes. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // *Optional* Maximum number of recognition hypotheses to be returned. // Specifically, the maximum number of `SpeechRecognitionAlternative` messages // within each `SpeechRecognitionResult`. // The server may return fewer than `max_alternatives`. // Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of // one. If omitted, will return a maximum of one. - MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives" json:"max_alternatives,omitempty"` + MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives,proto3" json:"max_alternatives,omitempty"` // *Optional* If set to `true`, the server will attempt to filter out // profanities, replacing all but the initial character in each filtered word // with asterisks, e.g. "f***". If set to `false` or omitted, profanities // won't be filtered out. - ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter" json:"profanity_filter,omitempty"` + ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter,proto3" json:"profanity_filter,omitempty"` // *Optional* A means to provide context to assist the speech recognition. - SpeechContexts []*SpeechContext `protobuf:"bytes,6,rep,name=speech_contexts,json=speechContexts" json:"speech_contexts,omitempty"` + SpeechContexts []*SpeechContext `protobuf:"bytes,6,rep,name=speech_contexts,json=speechContexts,proto3" json:"speech_contexts,omitempty"` // *Optional* If `true`, the top result includes a list of words and // the start and end time offsets (timestamps) for those words. If // `false`, no word-level time offset information is returned. The default is // `false`. - EnableWordTimeOffsets bool `protobuf:"varint,8,opt,name=enable_word_time_offsets,json=enableWordTimeOffsets" json:"enable_word_time_offsets,omitempty"` + EnableWordTimeOffsets bool `protobuf:"varint,8,opt,name=enable_word_time_offsets,json=enableWordTimeOffsets,proto3" json:"enable_word_time_offsets,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -582,7 +582,7 @@ type SpeechContext struct { // specific commands are typically spoken by the user. This can also be used // to add additional words to the vocabulary of the recognizer. See // [usage limits](https://cloud.google.com/speech/limits#content). - Phrases []string `protobuf:"bytes,1,rep,name=phrases" json:"phrases,omitempty"` + Phrases []string `protobuf:"bytes,1,rep,name=phrases,proto3" json:"phrases,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -668,7 +668,7 @@ type RecognitionAudio_Content struct { Content []byte `protobuf:"bytes,1,opt,name=content,proto3,oneof"` } type RecognitionAudio_Uri struct { - Uri string `protobuf:"bytes,2,opt,name=uri,oneof"` + Uri string `protobuf:"bytes,2,opt,name=uri,proto3,oneof"` } func (*RecognitionAudio_Content) isRecognitionAudio_AudioSource() {} @@ -767,7 +767,7 @@ func _RecognitionAudio_OneofSizer(msg proto.Message) (n int) { type RecognizeResponse struct { // *Output-only* Sequential list of transcription results corresponding to // sequential portions of audio. - Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -812,7 +812,7 @@ func (m *RecognizeResponse) GetResults() []*SpeechRecognitionResult { type LongRunningRecognizeResponse struct { // *Output-only* Sequential list of transcription results corresponding to // sequential portions of audio. - Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -855,11 +855,11 @@ func (m *LongRunningRecognizeResponse) GetResults() []*SpeechRecognitionResult { type LongRunningRecognizeMetadata struct { // Approximate percentage of audio processed thus far. Guaranteed to be 100 // when the audio is fully processed and the results are available. - ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent processing update. - LastUpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime" json:"last_update_time,omitempty"` + LastUpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime,proto3" json:"last_update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -962,14 +962,14 @@ func (m *LongRunningRecognizeMetadata) GetLastUpdateTime() *timestamp.Timestamp type StreamingRecognizeResponse struct { // *Output-only* If set, returns a [google.rpc.Status][google.rpc.Status] message that // specifies the error for the operation. - Error *status.Status `protobuf:"bytes,1,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,1,opt,name=error,proto3" json:"error,omitempty"` // *Output-only* This repeated list contains zero or more results that // correspond to consecutive portions of the audio currently being processed. // It contains zero or more `is_final=false` results followed by zero or one // `is_final=true` result (the newly settled portion). - Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` // *Output-only* Indicates the type of speech event. - SpeechEventType StreamingRecognizeResponse_SpeechEventType `protobuf:"varint,4,opt,name=speech_event_type,json=speechEventType,enum=google.cloud.speech.v1.StreamingRecognizeResponse_SpeechEventType" json:"speech_event_type,omitempty"` + SpeechEventType StreamingRecognizeResponse_SpeechEventType `protobuf:"varint,4,opt,name=speech_event_type,json=speechEventType,proto3,enum=google.cloud.speech.v1.StreamingRecognizeResponse_SpeechEventType" json:"speech_event_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1025,19 +1025,19 @@ func (m *StreamingRecognizeResponse) GetSpeechEventType() StreamingRecognizeResp type StreamingRecognitionResult struct { // *Output-only* May contain one or more recognition hypotheses (up to the // maximum specified in `max_alternatives`). - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` // *Output-only* If `false`, this `StreamingRecognitionResult` represents an // interim result that may change. If `true`, this is the final time the // speech service will return this particular `StreamingRecognitionResult`, // the recognizer will not return any further hypotheses for this portion of // the transcript and corresponding audio. - IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal" json:"is_final,omitempty"` + IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal,proto3" json:"is_final,omitempty"` // *Output-only* An estimate of the likelihood that the recognizer will not // change its guess about this interim result. Values range from 0.0 // (completely unstable) to 1.0 (completely stable). // This field is only provided for interim results (`is_final=false`). // The default of 0.0 is a sentinel value indicating `stability` was not set. - Stability float32 `protobuf:"fixed32,3,opt,name=stability" json:"stability,omitempty"` + Stability float32 `protobuf:"fixed32,3,opt,name=stability,proto3" json:"stability,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1094,7 +1094,7 @@ type SpeechRecognitionResult struct { // maximum specified in `max_alternatives`). // These alternatives are ordered in terms of accuracy, with the top (first) // alternative being the most probable, as ranked by the recognizer. - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1134,16 +1134,16 @@ func (m *SpeechRecognitionResult) GetAlternatives() []*SpeechRecognitionAlternat // Alternative hypotheses (a.k.a. n-best list). type SpeechRecognitionAlternative struct { // *Output-only* Transcript text representing the words that the user spoke. - Transcript string `protobuf:"bytes,1,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,1,opt,name=transcript,proto3" json:"transcript,omitempty"` // *Output-only* The confidence estimate between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. This field is typically provided only for the top hypothesis, and // only for `is_final=true` results. Clients should not rely on the // `confidence` field as it is not guaranteed to be accurate or consistent. // The default of 0.0 is a sentinel value indicating `confidence` was not set. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // *Output-only* A list of word-specific information for each recognized word. - Words []*WordInfo `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` + Words []*WordInfo `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1204,16 +1204,16 @@ type WordInfo struct { // in the top hypothesis. // This is an experimental feature and the accuracy of the time offset can // vary. - StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // *Output-only* Time offset relative to the beginning of the audio, // and corresponding to the end of the spoken word. // This field is only set if `enable_word_time_offsets=true` and only // in the top hypothesis. // This is an experimental feature and the accuracy of the time offset can // vary. - EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // *Output-only* The word corresponding to this set of information. - Word string `protobuf:"bytes,3,opt,name=word" json:"word,omitempty"` + Word string `protobuf:"bytes,3,opt,name=word,proto3" json:"word,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1366,8 +1366,7 @@ func (x *speechStreamingRecognizeClient) Recv() (*StreamingRecognizeResponse, er return m, nil } -// Server API for Speech service - +// SpeechServer is the server API for Speech service. type SpeechServer interface { // Performs synchronous speech recognition: receive results after all audio // has been sent and processed. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1beta1/cloud_speech.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1beta1/cloud_speech.pb.go index db6a62cd1..81bd0ede4 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1beta1/cloud_speech.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1beta1/cloud_speech.pb.go @@ -141,9 +141,9 @@ func (StreamingRecognizeResponse_EndpointerType) EnumDescriptor() ([]byte, []int type SyncRecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Required* The audio data to be recognized. - Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` + Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio,proto3" json:"audio,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -191,9 +191,9 @@ func (m *SyncRecognizeRequest) GetAudio() *RecognitionAudio { type AsyncRecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Required* The audio data to be recognized. - Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` + Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio,proto3" json:"audio,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -283,7 +283,7 @@ type isStreamingRecognizeRequest_StreamingRequest interface { } type StreamingRecognizeRequest_StreamingConfig struct { - StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,oneof"` + StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,proto3,oneof"` } type StreamingRecognizeRequest_AudioContent struct { AudioContent []byte `protobuf:"bytes,2,opt,name=audio_content,json=audioContent,proto3,oneof"` @@ -388,7 +388,7 @@ func _StreamingRecognizeRequest_OneofSizer(msg proto.Message) (n int) { type StreamingRecognitionConfig struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Optional* If `false` or omitted, the recognizer will perform continuous // recognition (continuing to wait for and process audio even if the user // pauses speaking) until the client closes the input stream (gRPC API) or @@ -399,12 +399,12 @@ type StreamingRecognitionConfig struct { // detects that the user has paused or stopped speaking, it will return an // `END_OF_UTTERANCE` event and cease recognition. It will return no more than // one `StreamingRecognitionResult` with the `is_final` flag set to `true`. - SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance" json:"single_utterance,omitempty"` + SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance,proto3" json:"single_utterance,omitempty"` // *Optional* If `true`, interim results (tentative hypotheses) may be // returned as they become available (these interim results are indicated with // the `is_final=false` flag). // If `false` or omitted, only `is_final=true` result(s) are returned. - InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults" json:"interim_results,omitempty"` + InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults,proto3" json:"interim_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -459,33 +459,33 @@ func (m *StreamingRecognitionConfig) GetInterimResults() bool { // request. type RecognitionConfig struct { // *Required* Encoding of audio data sent in all `RecognitionAudio` messages. - Encoding RecognitionConfig_AudioEncoding `protobuf:"varint,1,opt,name=encoding,enum=google.cloud.speech.v1beta1.RecognitionConfig_AudioEncoding" json:"encoding,omitempty"` + Encoding RecognitionConfig_AudioEncoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.cloud.speech.v1beta1.RecognitionConfig_AudioEncoding" json:"encoding,omitempty"` // *Required* Sample rate in Hertz of the audio data sent in all // `RecognitionAudio` messages. Valid values are: 8000-48000. // 16000 is optimal. For best results, set the sampling rate of the audio // source to 16000 Hz. If that's not possible, use the native sample rate of // the audio source (instead of re-sampling). - SampleRate int32 `protobuf:"varint,2,opt,name=sample_rate,json=sampleRate" json:"sample_rate,omitempty"` + SampleRate int32 `protobuf:"varint,2,opt,name=sample_rate,json=sampleRate,proto3" json:"sample_rate,omitempty"` // *Optional* The language of the supplied audio as a BCP-47 language tag. // Example: "en-GB" https://www.rfc-editor.org/rfc/bcp/bcp47.txt // If omitted, defaults to "en-US". See // [Language Support](https://cloud.google.com/speech/docs/languages) // for a list of the currently supported language codes. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // *Optional* Maximum number of recognition hypotheses to be returned. // Specifically, the maximum number of `SpeechRecognitionAlternative` messages // within each `SpeechRecognitionResult`. // The server may return fewer than `max_alternatives`. // Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of // one. If omitted, will return a maximum of one. - MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives" json:"max_alternatives,omitempty"` + MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives,proto3" json:"max_alternatives,omitempty"` // *Optional* If set to `true`, the server will attempt to filter out // profanities, replacing all but the initial character in each filtered word // with asterisks, e.g. "f***". If set to `false` or omitted, profanities // won't be filtered out. - ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter" json:"profanity_filter,omitempty"` + ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter,proto3" json:"profanity_filter,omitempty"` // *Optional* A means to provide context to assist the speech recognition. - SpeechContext *SpeechContext `protobuf:"bytes,6,opt,name=speech_context,json=speechContext" json:"speech_context,omitempty"` + SpeechContext *SpeechContext `protobuf:"bytes,6,opt,name=speech_context,json=speechContext,proto3" json:"speech_context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -566,7 +566,7 @@ type SpeechContext struct { // specific commands are typically spoken by the user. This can also be used // to add additional words to the vocabulary of the recognizer. See // [usage limits](https://cloud.google.com/speech/limits#content). - Phrases []string `protobuf:"bytes,1,rep,name=phrases" json:"phrases,omitempty"` + Phrases []string `protobuf:"bytes,1,rep,name=phrases,proto3" json:"phrases,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -651,7 +651,7 @@ type RecognitionAudio_Content struct { Content []byte `protobuf:"bytes,1,opt,name=content,proto3,oneof"` } type RecognitionAudio_Uri struct { - Uri string `protobuf:"bytes,2,opt,name=uri,oneof"` + Uri string `protobuf:"bytes,2,opt,name=uri,proto3,oneof"` } func (*RecognitionAudio_Content) isRecognitionAudio_AudioSource() {} @@ -750,7 +750,7 @@ func _RecognitionAudio_OneofSizer(msg proto.Message) (n int) { type SyncRecognizeResponse struct { // *Output-only* Sequential list of transcription results corresponding to // sequential portions of audio. - Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -794,7 +794,7 @@ func (m *SyncRecognizeResponse) GetResults() []*SpeechRecognitionResult { type AsyncRecognizeResponse struct { // *Output-only* Sequential list of transcription results corresponding to // sequential portions of audio. - Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -837,11 +837,11 @@ func (m *AsyncRecognizeResponse) GetResults() []*SpeechRecognitionResult { type AsyncRecognizeMetadata struct { // Approximate percentage of audio processed thus far. Guaranteed to be 100 // when the audio is fully processed and the results are available. - ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent processing update. - LastUpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime" json:"last_update_time,omitempty"` + LastUpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime,proto3" json:"last_update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -955,18 +955,18 @@ func (m *AsyncRecognizeMetadata) GetLastUpdateTime() *timestamp.Timestamp { type StreamingRecognizeResponse struct { // *Output-only* If set, returns a [google.rpc.Status][google.rpc.Status] message that // specifies the error for the operation. - Error *status.Status `protobuf:"bytes,1,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,1,opt,name=error,proto3" json:"error,omitempty"` // *Output-only* This repeated list contains zero or more results that // correspond to consecutive portions of the audio currently being processed. // It contains zero or one `is_final=true` result (the newly settled portion), // followed by zero or more `is_final=false` results. - Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` // *Output-only* Indicates the lowest index in the `results` array that has // changed. The repeated `StreamingRecognitionResult` results overwrite past // results at this index and higher. - ResultIndex int32 `protobuf:"varint,3,opt,name=result_index,json=resultIndex" json:"result_index,omitempty"` + ResultIndex int32 `protobuf:"varint,3,opt,name=result_index,json=resultIndex,proto3" json:"result_index,omitempty"` // *Output-only* Indicates the type of endpointer event. - EndpointerType StreamingRecognizeResponse_EndpointerType `protobuf:"varint,4,opt,name=endpointer_type,json=endpointerType,enum=google.cloud.speech.v1beta1.StreamingRecognizeResponse_EndpointerType" json:"endpointer_type,omitempty"` + EndpointerType StreamingRecognizeResponse_EndpointerType `protobuf:"varint,4,opt,name=endpointer_type,json=endpointerType,proto3,enum=google.cloud.speech.v1beta1.StreamingRecognizeResponse_EndpointerType" json:"endpointer_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1029,19 +1029,19 @@ func (m *StreamingRecognizeResponse) GetEndpointerType() StreamingRecognizeRespo type StreamingRecognitionResult struct { // *Output-only* May contain one or more recognition hypotheses (up to the // maximum specified in `max_alternatives`). - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` // *Output-only* If `false`, this `StreamingRecognitionResult` represents an // interim result that may change. If `true`, this is the final time the // speech service will return this particular `StreamingRecognitionResult`, // the recognizer will not return any further hypotheses for this portion of // the transcript and corresponding audio. - IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal" json:"is_final,omitempty"` + IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal,proto3" json:"is_final,omitempty"` // *Output-only* An estimate of the likelihood that the recognizer will not // change its guess about this interim result. Values range from 0.0 // (completely unstable) to 1.0 (completely stable). // This field is only provided for interim results (`is_final=false`). // The default of 0.0 is a sentinel value indicating `stability` was not set. - Stability float32 `protobuf:"fixed32,3,opt,name=stability" json:"stability,omitempty"` + Stability float32 `protobuf:"fixed32,3,opt,name=stability,proto3" json:"stability,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1096,7 +1096,7 @@ func (m *StreamingRecognitionResult) GetStability() float32 { type SpeechRecognitionResult struct { // *Output-only* May contain one or more recognition hypotheses (up to the // maximum specified in `max_alternatives`). - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1136,7 +1136,7 @@ func (m *SpeechRecognitionResult) GetAlternatives() []*SpeechRecognitionAlternat // Alternative hypotheses (a.k.a. n-best list). type SpeechRecognitionAlternative struct { // *Output-only* Transcript text representing the words that the user spoke. - Transcript string `protobuf:"bytes,1,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,1,opt,name=transcript,proto3" json:"transcript,omitempty"` // *Output-only* The confidence estimate between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. This field is typically provided only for the top hypothesis, and @@ -1144,7 +1144,7 @@ type SpeechRecognitionAlternative struct { // `confidence` field as it is not guaranteed to be accurate, or even set, in // any of the results. // The default of 0.0 is a sentinel value indicating `confidence` was not set. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1291,8 +1291,7 @@ func (x *speechStreamingRecognizeClient) Recv() (*StreamingRecognizeResponse, er return m, nil } -// Server API for Speech service - +// SpeechServer is the server API for Speech service. type SpeechServer interface { // Performs synchronous speech recognition: receive results after all audio // has been sent and processed. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1p1beta1/cloud_speech.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1p1beta1/cloud_speech.pb.go index 0535088bb..c1cd937db 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1p1beta1/cloud_speech.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/speech/v1p1beta1/cloud_speech.pb.go @@ -327,9 +327,9 @@ func (StreamingRecognizeResponse_SpeechEventType) EnumDescriptor() ([]byte, []in type RecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Required* The audio data to be recognized. - Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` + Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio,proto3" json:"audio,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -378,9 +378,9 @@ func (m *RecognizeRequest) GetAudio() *RecognitionAudio { type LongRunningRecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Required* The audio data to be recognized. - Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` + Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio,proto3" json:"audio,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -470,7 +470,7 @@ type isStreamingRecognizeRequest_StreamingRequest interface { } type StreamingRecognizeRequest_StreamingConfig struct { - StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,oneof"` + StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,proto3,oneof"` } type StreamingRecognizeRequest_AudioContent struct { AudioContent []byte `protobuf:"bytes,2,opt,name=audio_content,json=audioContent,proto3,oneof"` @@ -575,7 +575,7 @@ func _StreamingRecognizeRequest_OneofSizer(msg proto.Message) (n int) { type StreamingRecognitionConfig struct { // *Required* Provides information to the recognizer that specifies how to // process the request. - Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` + Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config,proto3" json:"config,omitempty"` // *Optional* If `false` or omitted, the recognizer will perform continuous // recognition (continuing to wait for and process audio even if the user // pauses speaking) until the client closes the input stream (gRPC API) or @@ -587,12 +587,12 @@ type StreamingRecognitionConfig struct { // `END_OF_SINGLE_UTTERANCE` event and cease recognition. It will return no // more than one `StreamingRecognitionResult` with the `is_final` flag set to // `true`. - SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance" json:"single_utterance,omitempty"` + SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance,proto3" json:"single_utterance,omitempty"` // *Optional* If `true`, interim results (tentative hypotheses) may be // returned as they become available (these interim results are indicated with // the `is_final=false` flag). // If `false` or omitted, only `is_final=true` result(s) are returned. - InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults" json:"interim_results,omitempty"` + InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults,proto3" json:"interim_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -649,7 +649,7 @@ type RecognitionConfig struct { // Encoding of audio data sent in all `RecognitionAudio` messages. // This field is optional for `FLAC` and `WAV` audio files and required // for all other audio formats. For details, see [AudioEncoding][google.cloud.speech.v1p1beta1.RecognitionConfig.AudioEncoding]. - Encoding RecognitionConfig_AudioEncoding `protobuf:"varint,1,opt,name=encoding,enum=google.cloud.speech.v1p1beta1.RecognitionConfig_AudioEncoding" json:"encoding,omitempty"` + Encoding RecognitionConfig_AudioEncoding `protobuf:"varint,1,opt,name=encoding,proto3,enum=google.cloud.speech.v1p1beta1.RecognitionConfig_AudioEncoding" json:"encoding,omitempty"` // Sample rate in Hertz of the audio data sent in all // `RecognitionAudio` messages. Valid values are: 8000-48000. // 16000 is optimal. For best results, set the sampling rate of the audio @@ -657,32 +657,32 @@ type RecognitionConfig struct { // the audio source (instead of re-sampling). // This field is optional for `FLAC` and `WAV` audio files and required // for all other audio formats. For details, see [AudioEncoding][google.cloud.speech.v1p1beta1.RecognitionConfig.AudioEncoding]. - SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` // *Required* The language of the supplied audio as a // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. // Example: "en-US". // See [Language Support](https://cloud.google.com/speech/docs/languages) // for a list of the currently supported language codes. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // *Optional* Maximum number of recognition hypotheses to be returned. // Specifically, the maximum number of `SpeechRecognitionAlternative` messages // within each `SpeechRecognitionResult`. // The server may return fewer than `max_alternatives`. // Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of // one. If omitted, will return a maximum of one. - MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives" json:"max_alternatives,omitempty"` + MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives,proto3" json:"max_alternatives,omitempty"` // *Optional* If set to `true`, the server will attempt to filter out // profanities, replacing all but the initial character in each filtered word // with asterisks, e.g. "f***". If set to `false` or omitted, profanities // won't be filtered out. - ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter" json:"profanity_filter,omitempty"` + ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter,proto3" json:"profanity_filter,omitempty"` // *Optional* A means to provide context to assist the speech recognition. - SpeechContexts []*SpeechContext `protobuf:"bytes,6,rep,name=speech_contexts,json=speechContexts" json:"speech_contexts,omitempty"` + SpeechContexts []*SpeechContext `protobuf:"bytes,6,rep,name=speech_contexts,json=speechContexts,proto3" json:"speech_contexts,omitempty"` // *Optional* If `true`, the top result includes a list of words and // the start and end time offsets (timestamps) for those words. If // `false`, no word-level time offset information is returned. The default is // `false`. - EnableWordTimeOffsets bool `protobuf:"varint,8,opt,name=enable_word_time_offsets,json=enableWordTimeOffsets" json:"enable_word_time_offsets,omitempty"` + EnableWordTimeOffsets bool `protobuf:"varint,8,opt,name=enable_word_time_offsets,json=enableWordTimeOffsets,proto3" json:"enable_word_time_offsets,omitempty"` // *Optional* If 'true', adds punctuation to recognition result hypotheses. // This feature is only available in select languages. Setting this for // requests in other languages has no effect at all. @@ -690,9 +690,9 @@ type RecognitionConfig struct { // NOTE: "This is currently offered as an experimental service, complimentary // to all users. In the future this may be exclusively available as a // premium feature." - EnableAutomaticPunctuation bool `protobuf:"varint,11,opt,name=enable_automatic_punctuation,json=enableAutomaticPunctuation" json:"enable_automatic_punctuation,omitempty"` + EnableAutomaticPunctuation bool `protobuf:"varint,11,opt,name=enable_automatic_punctuation,json=enableAutomaticPunctuation,proto3" json:"enable_automatic_punctuation,omitempty"` // *Optional* Metadata regarding this request. - Metadata *RecognitionMetadata `protobuf:"bytes,9,opt,name=metadata" json:"metadata,omitempty"` + Metadata *RecognitionMetadata `protobuf:"bytes,9,opt,name=metadata,proto3" json:"metadata,omitempty"` // *Optional* Which model to select for the given request. Select the model // best suited to your domain to get best results. If a model is not // explicitly specified, then we auto-select a model based on the parameters @@ -725,7 +725,7 @@ type RecognitionConfig struct { // recorded at a 16khz or greater sampling rate. // // - Model string `protobuf:"bytes,13,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,13,opt,name=model,proto3" json:"model,omitempty"` // *Optional* Set to true to use an enhanced model for speech recognition. // You must also set the `model` field to a valid, enhanced model. If // `use_enhanced` is set to true and the `model` field is not set, then @@ -737,7 +737,7 @@ type RecognitionConfig struct { // instructions in the [alpha documentation](/speech/data-sharing). If you set // `use_enhanced` to true and you have not enabled audio logging, then you // will receive an error. - UseEnhanced bool `protobuf:"varint,14,opt,name=use_enhanced,json=useEnhanced" json:"use_enhanced,omitempty"` + UseEnhanced bool `protobuf:"varint,14,opt,name=use_enhanced,json=useEnhanced,proto3" json:"use_enhanced,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -847,33 +847,33 @@ func (m *RecognitionConfig) GetUseEnhanced() bool { // Description of audio data to be recognized. type RecognitionMetadata struct { // The use case most closely describing the audio content to be recognized. - InteractionType RecognitionMetadata_InteractionType `protobuf:"varint,1,opt,name=interaction_type,json=interactionType,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_InteractionType" json:"interaction_type,omitempty"` + InteractionType RecognitionMetadata_InteractionType `protobuf:"varint,1,opt,name=interaction_type,json=interactionType,proto3,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_InteractionType" json:"interaction_type,omitempty"` // The industry vertical to which this speech recognition request most // closely applies. This is most indicative of the topics contained // in the audio. Use the 6-digit NAICS code to identify the industry // vertical - see https://www.naics.com/search/. - IndustryNaicsCodeOfAudio uint32 `protobuf:"varint,3,opt,name=industry_naics_code_of_audio,json=industryNaicsCodeOfAudio" json:"industry_naics_code_of_audio,omitempty"` + IndustryNaicsCodeOfAudio uint32 `protobuf:"varint,3,opt,name=industry_naics_code_of_audio,json=industryNaicsCodeOfAudio,proto3" json:"industry_naics_code_of_audio,omitempty"` // The audio type that most closely describes the audio being recognized. - MicrophoneDistance RecognitionMetadata_MicrophoneDistance `protobuf:"varint,4,opt,name=microphone_distance,json=microphoneDistance,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_MicrophoneDistance" json:"microphone_distance,omitempty"` + MicrophoneDistance RecognitionMetadata_MicrophoneDistance `protobuf:"varint,4,opt,name=microphone_distance,json=microphoneDistance,proto3,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_MicrophoneDistance" json:"microphone_distance,omitempty"` // The original media the speech was recorded on. - OriginalMediaType RecognitionMetadata_OriginalMediaType `protobuf:"varint,5,opt,name=original_media_type,json=originalMediaType,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_OriginalMediaType" json:"original_media_type,omitempty"` + OriginalMediaType RecognitionMetadata_OriginalMediaType `protobuf:"varint,5,opt,name=original_media_type,json=originalMediaType,proto3,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_OriginalMediaType" json:"original_media_type,omitempty"` // The type of device the speech was recorded with. - RecordingDeviceType RecognitionMetadata_RecordingDeviceType `protobuf:"varint,6,opt,name=recording_device_type,json=recordingDeviceType,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_RecordingDeviceType" json:"recording_device_type,omitempty"` + RecordingDeviceType RecognitionMetadata_RecordingDeviceType `protobuf:"varint,6,opt,name=recording_device_type,json=recordingDeviceType,proto3,enum=google.cloud.speech.v1p1beta1.RecognitionMetadata_RecordingDeviceType" json:"recording_device_type,omitempty"` // The device used to make the recording. Examples 'Nexus 5X' or // 'Polycom SoundStation IP 6000' or 'POTS' or 'VoIP' or // 'Cardioid Microphone'. - RecordingDeviceName string `protobuf:"bytes,7,opt,name=recording_device_name,json=recordingDeviceName" json:"recording_device_name,omitempty"` + RecordingDeviceName string `protobuf:"bytes,7,opt,name=recording_device_name,json=recordingDeviceName,proto3" json:"recording_device_name,omitempty"` // Mime type of the original audio file. For example `audio/m4a`, // `audio/x-alaw-basic`, `audio/mp3`, `audio/3gpp`. // A list of possible audio mime types is maintained at // http://www.iana.org/assignments/media-types/media-types.xhtml#audio - OriginalMimeType string `protobuf:"bytes,8,opt,name=original_mime_type,json=originalMimeType" json:"original_mime_type,omitempty"` + OriginalMimeType string `protobuf:"bytes,8,opt,name=original_mime_type,json=originalMimeType,proto3" json:"original_mime_type,omitempty"` // Obfuscated (privacy-protected) ID of the user, to identify number of // unique users using the service. - ObfuscatedId int64 `protobuf:"varint,9,opt,name=obfuscated_id,json=obfuscatedId" json:"obfuscated_id,omitempty"` + ObfuscatedId int64 `protobuf:"varint,9,opt,name=obfuscated_id,json=obfuscatedId,proto3" json:"obfuscated_id,omitempty"` // Description of the content. Eg. "Recordings of federal supreme court // hearings from 2012". - AudioTopic string `protobuf:"bytes,10,opt,name=audio_topic,json=audioTopic" json:"audio_topic,omitempty"` + AudioTopic string `protobuf:"bytes,10,opt,name=audio_topic,json=audioTopic,proto3" json:"audio_topic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -975,7 +975,7 @@ type SpeechContext struct { // specific commands are typically spoken by the user. This can also be used // to add additional words to the vocabulary of the recognizer. See // [usage limits](https://cloud.google.com/speech/limits#content). - Phrases []string `protobuf:"bytes,1,rep,name=phrases" json:"phrases,omitempty"` + Phrases []string `protobuf:"bytes,1,rep,name=phrases,proto3" json:"phrases,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1061,7 +1061,7 @@ type RecognitionAudio_Content struct { Content []byte `protobuf:"bytes,1,opt,name=content,proto3,oneof"` } type RecognitionAudio_Uri struct { - Uri string `protobuf:"bytes,2,opt,name=uri,oneof"` + Uri string `protobuf:"bytes,2,opt,name=uri,proto3,oneof"` } func (*RecognitionAudio_Content) isRecognitionAudio_AudioSource() {} @@ -1160,7 +1160,7 @@ func _RecognitionAudio_OneofSizer(msg proto.Message) (n int) { type RecognizeResponse struct { // Output only. Sequential list of transcription results corresponding to // sequential portions of audio. - Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1205,7 +1205,7 @@ func (m *RecognizeResponse) GetResults() []*SpeechRecognitionResult { type LongRunningRecognizeResponse struct { // Output only. Sequential list of transcription results corresponding to // sequential portions of audio. - Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1248,11 +1248,11 @@ func (m *LongRunningRecognizeResponse) GetResults() []*SpeechRecognitionResult { type LongRunningRecognizeMetadata struct { // Approximate percentage of audio processed thus far. Guaranteed to be 100 // when the audio is fully processed and the results are available. - ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent processing update. - LastUpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime" json:"last_update_time,omitempty"` + LastUpdateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime,proto3" json:"last_update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1355,14 +1355,14 @@ func (m *LongRunningRecognizeMetadata) GetLastUpdateTime() *timestamp.Timestamp type StreamingRecognizeResponse struct { // Output only. If set, returns a [google.rpc.Status][google.rpc.Status] message that // specifies the error for the operation. - Error *status.Status `protobuf:"bytes,1,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,1,opt,name=error,proto3" json:"error,omitempty"` // Output only. This repeated list contains zero or more results that // correspond to consecutive portions of the audio currently being processed. // It contains zero or one `is_final=true` result (the newly settled portion), // followed by zero or more `is_final=false` results (the interim results). - Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` + Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results,proto3" json:"results,omitempty"` // Output only. Indicates the type of speech event. - SpeechEventType StreamingRecognizeResponse_SpeechEventType `protobuf:"varint,4,opt,name=speech_event_type,json=speechEventType,enum=google.cloud.speech.v1p1beta1.StreamingRecognizeResponse_SpeechEventType" json:"speech_event_type,omitempty"` + SpeechEventType StreamingRecognizeResponse_SpeechEventType `protobuf:"varint,4,opt,name=speech_event_type,json=speechEventType,proto3,enum=google.cloud.speech.v1p1beta1.StreamingRecognizeResponse_SpeechEventType" json:"speech_event_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1420,19 +1420,19 @@ type StreamingRecognitionResult struct { // maximum specified in `max_alternatives`). // These alternatives are ordered in terms of accuracy, with the top (first) // alternative being the most probable, as ranked by the recognizer. - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` // Output only. If `false`, this `StreamingRecognitionResult` represents an // interim result that may change. If `true`, this is the final time the // speech service will return this particular `StreamingRecognitionResult`, // the recognizer will not return any further hypotheses for this portion of // the transcript and corresponding audio. - IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal" json:"is_final,omitempty"` + IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal,proto3" json:"is_final,omitempty"` // Output only. An estimate of the likelihood that the recognizer will not // change its guess about this interim result. Values range from 0.0 // (completely unstable) to 1.0 (completely stable). // This field is only provided for interim results (`is_final=false`). // The default of 0.0 is a sentinel value indicating `stability` was not set. - Stability float32 `protobuf:"fixed32,3,opt,name=stability" json:"stability,omitempty"` + Stability float32 `protobuf:"fixed32,3,opt,name=stability,proto3" json:"stability,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1489,7 +1489,7 @@ type SpeechRecognitionResult struct { // maximum specified in `max_alternatives`). // These alternatives are ordered in terms of accuracy, with the top (first) // alternative being the most probable, as ranked by the recognizer. - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1529,7 +1529,7 @@ func (m *SpeechRecognitionResult) GetAlternatives() []*SpeechRecognitionAlternat // Alternative hypotheses (a.k.a. n-best list). type SpeechRecognitionAlternative struct { // Output only. Transcript text representing the words that the user spoke. - Transcript string `protobuf:"bytes,1,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,1,opt,name=transcript,proto3" json:"transcript,omitempty"` // Output only. The confidence estimate between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. This field is set only for the top alternative of a non-streaming @@ -1537,11 +1537,11 @@ type SpeechRecognitionAlternative struct { // This field is not guaranteed to be accurate and users should not rely on it // to be always provided. // The default of 0.0 is a sentinel value indicating `confidence` was not set. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // Output only. A list of word-specific information for each recognized word. // Note: When enable_speaker_diarization is true, you will see all the words // from the beginning of the audio. - Words []*WordInfo `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` + Words []*WordInfo `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1600,16 +1600,16 @@ type WordInfo struct { // in the top hypothesis. // This is an experimental feature and the accuracy of the time offset can // vary. - StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Output only. Time offset relative to the beginning of the audio, // and corresponding to the end of the spoken word. // This field is only set if `enable_word_time_offsets=true` and only // in the top hypothesis. // This is an experimental feature and the accuracy of the time offset can // vary. - EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Output only. The word corresponding to this set of information. - Word string `protobuf:"bytes,3,opt,name=word" json:"word,omitempty"` + Word string `protobuf:"bytes,3,opt,name=word,proto3" json:"word,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1767,8 +1767,7 @@ func (x *speechStreamingRecognizeClient) Recv() (*StreamingRecognizeResponse, er return m, nil } -// Server API for Speech service - +// SpeechServer is the server API for Speech service. type SpeechServer interface { // Performs synchronous speech recognition: receive results after all audio // has been sent and processed. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/support/common/common.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/support/common/common.pb.go index c3ef734b4..65817f283 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/support/common/common.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/support/common/common.pb.go @@ -272,26 +272,26 @@ type SupportAccount struct { // The resource name for a support account in format // `supportAccounts/{account_id}`. // Output only. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Identifier for this entity that gets persisted in storage system. The // resource name is populated using this field in format // `supportAccounts/{account_id}`. - AccountId string `protobuf:"bytes,2,opt,name=account_id,json=accountId" json:"account_id,omitempty"` + AccountId string `protobuf:"bytes,2,opt,name=account_id,json=accountId,proto3" json:"account_id,omitempty"` // The Cloud resource with which this support account is associated. - CloudResource string `protobuf:"bytes,3,opt,name=cloud_resource,json=cloudResource" json:"cloud_resource,omitempty"` + CloudResource string `protobuf:"bytes,3,opt,name=cloud_resource,json=cloudResource,proto3" json:"cloud_resource,omitempty"` // A user friendly display name assigned to this support account. - DisplayName string `protobuf:"bytes,4,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,4,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Indicates the current state of an account. - State SupportAccount_State `protobuf:"varint,5,opt,name=state,enum=google.cloud.support.common.SupportAccount_State" json:"state,omitempty"` + State SupportAccount_State `protobuf:"varint,5,opt,name=state,proto3,enum=google.cloud.support.common.SupportAccount_State" json:"state,omitempty"` // Time when this account was created. // Output only. - CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The resource name of a billing account associated with this support // account. For example, `billingAccounts/ABCDEF-012345-567890`. - BillingAccountName string `protobuf:"bytes,7,opt,name=billing_account_name,json=billingAccountName" json:"billing_account_name,omitempty"` - UnifyAccountId string `protobuf:"bytes,8,opt,name=unify_account_id,json=unifyAccountId" json:"unify_account_id,omitempty"` + BillingAccountName string `protobuf:"bytes,7,opt,name=billing_account_name,json=billingAccountName,proto3" json:"billing_account_name,omitempty"` + UnifyAccountId string `protobuf:"bytes,8,opt,name=unify_account_id,json=unifyAccountId,proto3" json:"unify_account_id,omitempty"` // The PricingModel applicable to this support account. - PricingModel SupportAccount_PricingModel `protobuf:"varint,9,opt,name=pricing_model,json=pricingModel,enum=google.cloud.support.common.SupportAccount_PricingModel" json:"pricing_model,omitempty"` + PricingModel SupportAccount_PricingModel `protobuf:"varint,9,opt,name=pricing_model,json=pricingModel,proto3,enum=google.cloud.support.common.SupportAccount_PricingModel" json:"pricing_model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -388,41 +388,41 @@ func (m *SupportAccount) GetPricingModel() SupportAccount_PricingModel { type Case struct { // The resource name for the Case in format // `supportAccounts/{account_id}/cases/{case_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The short summary of the issue reported in this case. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The board description of issue provided with initial summary. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The product component for which this Case is reported. - Component string `protobuf:"bytes,4,opt,name=component" json:"component,omitempty"` + Component string `protobuf:"bytes,4,opt,name=component,proto3" json:"component,omitempty"` // The product subcomponent for which this Case is reported. - Subcomponent string `protobuf:"bytes,5,opt,name=subcomponent" json:"subcomponent,omitempty"` + Subcomponent string `protobuf:"bytes,5,opt,name=subcomponent,proto3" json:"subcomponent,omitempty"` // Timezone the client sending this request is in. // It should be in a format IANA recognizes: https://www.iana.org/time-zone // There is no additional validation done by the API. - ClientTimezone string `protobuf:"bytes,6,opt,name=client_timezone,json=clientTimezone" json:"client_timezone,omitempty"` + ClientTimezone string `protobuf:"bytes,6,opt,name=client_timezone,json=clientTimezone,proto3" json:"client_timezone,omitempty"` // The email addresses that can be copied to receive updates on this case. // Users can specify a maximum of 10 email addresses. - CcAddresses []string `protobuf:"bytes,7,rep,name=cc_addresses,json=ccAddresses" json:"cc_addresses,omitempty"` + CcAddresses []string `protobuf:"bytes,7,rep,name=cc_addresses,json=ccAddresses,proto3" json:"cc_addresses,omitempty"` // The Google Cloud Platform project ID for which this case is created. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // List of customer issues associated with this case. - Issues []*CustomerIssue `protobuf:"bytes,10,rep,name=issues" json:"issues,omitempty"` + Issues []*CustomerIssue `protobuf:"bytes,10,rep,name=issues,proto3" json:"issues,omitempty"` // The current priority of this case. - Priority Case_Priority `protobuf:"varint,11,opt,name=priority,enum=google.cloud.support.common.Case_Priority" json:"priority,omitempty"` + Priority Case_Priority `protobuf:"varint,11,opt,name=priority,proto3,enum=google.cloud.support.common.Case_Priority" json:"priority,omitempty"` // The current state of this case. - State Case_State `protobuf:"varint,12,opt,name=state,enum=google.cloud.support.common.Case_State" json:"state,omitempty"` + State Case_State `protobuf:"varint,12,opt,name=state,proto3,enum=google.cloud.support.common.Case_State" json:"state,omitempty"` // Time when this case was created. // Output only. - CreateTime *timestamp.Timestamp `protobuf:"bytes,13,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,13,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time when this case was last updated. // Output only. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,14,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,14,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // Email address of user who created this case. // Output only. It is inferred from credentials supplied during case creation. - CreatorEmail string `protobuf:"bytes,15,opt,name=creator_email,json=creatorEmail" json:"creator_email,omitempty"` + CreatorEmail string `protobuf:"bytes,15,opt,name=creator_email,json=creatorEmail,proto3" json:"creator_email,omitempty"` // The issue category applicable to this case. - Category string `protobuf:"bytes,16,opt,name=category" json:"category,omitempty"` + Category string `protobuf:"bytes,16,opt,name=category,proto3" json:"category,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -563,19 +563,19 @@ func (m *Case) GetCategory() string { type CustomerIssue struct { // Unique identifier for the internal issue. // Output only. - IssueId string `protobuf:"bytes,1,opt,name=issue_id,json=issueId" json:"issue_id,omitempty"` + IssueId string `protobuf:"bytes,1,opt,name=issue_id,json=issueId,proto3" json:"issue_id,omitempty"` // Represents current status of the internal ticket. // Output only. - State CustomerIssue_IssueState `protobuf:"varint,2,opt,name=state,enum=google.cloud.support.common.CustomerIssue_IssueState" json:"state,omitempty"` + State CustomerIssue_IssueState `protobuf:"varint,2,opt,name=state,proto3,enum=google.cloud.support.common.CustomerIssue_IssueState" json:"state,omitempty"` // Time when the internal issue was created. // Output only. - CreateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time when the internal issue was marked as resolved. // Output only. - ResolveTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=resolve_time,json=resolveTime" json:"resolve_time,omitempty"` + ResolveTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=resolve_time,json=resolveTime,proto3" json:"resolve_time,omitempty"` // Time when the internal issue was last updated. // Output only. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -644,9 +644,9 @@ func (m *CustomerIssue) GetUpdateTime() *timestamp.Timestamp { // account. type SupportRole struct { // Email address of user being added through this Role. - Email string `protobuf:"bytes,1,opt,name=email" json:"email,omitempty"` + Email string `protobuf:"bytes,1,opt,name=email,proto3" json:"email,omitempty"` // The type of role assigned to user. - Role SupportRole_Role `protobuf:"varint,2,opt,name=role,enum=google.cloud.support.common.SupportRole_Role" json:"role,omitempty"` + Role SupportRole_Role `protobuf:"varint,2,opt,name=role,proto3,enum=google.cloud.support.common.SupportRole_Role" json:"role,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -693,17 +693,17 @@ func (m *SupportRole) GetRole() SupportRole_Role { // The comment text associated with a `Case`. type Comment struct { // Text containing a maximum of 3000 characters. - Text string `protobuf:"bytes,1,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,1,opt,name=text,proto3" json:"text,omitempty"` // Time when this update was created. // Output only. - CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The email address/name of user who created this comment. // Output only. - Author string `protobuf:"bytes,3,opt,name=author" json:"author,omitempty"` + Author string `protobuf:"bytes,3,opt,name=author,proto3" json:"author,omitempty"` // The resource name for this comment in format // `supportAccounts/{account_id}/cases/{case_id}/{comment_id}`. // Output only. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -767,7 +767,7 @@ func (m *Comment) GetName() string { // `Case.component` and `Case.subcomponent` fields respectively. type IssueTaxonomy struct { // Map of available categories. - Categories map[string]*IssueTaxonomy_Category `protobuf:"bytes,1,rep,name=categories" json:"categories,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Categories map[string]*IssueTaxonomy_Category `protobuf:"bytes,1,rep,name=categories,proto3" json:"categories,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -813,14 +813,14 @@ func (m *IssueTaxonomy) GetCategories() map[string]*IssueTaxonomy_Category { // (e.g., Google App Engine > Memcache). type IssueTaxonomy_Component struct { // User friendly name of this component. - DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // List of languages in which a support case can be created under this // component. Represented by language codes in ISO_639-1 standard. - Languages []string `protobuf:"bytes,2,rep,name=languages" json:"languages,omitempty"` + Languages []string `protobuf:"bytes,2,rep,name=languages,proto3" json:"languages,omitempty"` // Template to be used while filling the description of a support case. - Template string `protobuf:"bytes,3,opt,name=template" json:"template,omitempty"` + Template string `protobuf:"bytes,3,opt,name=template,proto3" json:"template,omitempty"` // List of subcomponents under this component. - Subcomponents []*IssueTaxonomy_Component `protobuf:"bytes,4,rep,name=subcomponents" json:"subcomponents,omitempty"` + Subcomponents []*IssueTaxonomy_Component `protobuf:"bytes,4,rep,name=subcomponents,proto3" json:"subcomponents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -882,9 +882,9 @@ func (m *IssueTaxonomy_Component) GetSubcomponents() []*IssueTaxonomy_Component // reported through a support case. type IssueTaxonomy_Category struct { // User friendly name of this category. - DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Map of product components under this category. - Components map[string]*IssueTaxonomy_Component `protobuf:"bytes,2,rep,name=components" json:"components,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Components map[string]*IssueTaxonomy_Component `protobuf:"bytes,2,rep,name=components,proto3" json:"components,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/support/v1alpha1/cloud_support.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/support/v1alpha1/cloud_support.pb.go index 329510718..bad4ddbc9 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/support/v1alpha1/cloud_support.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/support/v1alpha1/cloud_support.pb.go @@ -31,7 +31,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type GetSupportAccountRequest struct { // The resource name of the support accounts. For example: // `supportAccounts/accountA`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -74,12 +74,12 @@ type ListSupportAccountsRequest struct { // account list by a cloud_resource. For example, to filter results by support // accounts associated with an Organization, its value should be: // "cloud_resource:organizations/" - Filter string `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,1,opt,name=filter,proto3" json:"filter,omitempty"` // Maximum number of accounts fetched with each request. - PageSize int64 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int64 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A token identifying the page of results to return. If unspecified, the // first page is retrieved. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -133,11 +133,11 @@ func (m *ListSupportAccountsRequest) GetPageToken() string { // The response message for `ListSupportAccount`. type ListSupportAccountsResponse struct { // A list of support accounts. - Accounts []*common.SupportAccount `protobuf:"bytes,1,rep,name=accounts" json:"accounts,omitempty"` + Accounts []*common.SupportAccount `protobuf:"bytes,1,rep,name=accounts,proto3" json:"accounts,omitempty"` // A token to retrieve the next page of results. This should be passed on in // `page_token` field of `ListSupportAccountRequest` for next request. If // unspecified, there are no more results to retrieve. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -185,7 +185,7 @@ func (m *ListSupportAccountsResponse) GetNextPageToken() string { type GetCaseRequest struct { // Name of case resource requested. // For example: "supportAccounts/accountA/cases/123" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -226,15 +226,15 @@ func (m *GetCaseRequest) GetName() string { type ListCasesRequest struct { // Name of the account resource for which cases are requested. For example: // "supportAccounts/accountA" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The filter applied to the search results. Currently it only accepts "OPEN" // or "CLOSED" strings, filtering out cases that are open or resolved. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Maximum number of cases fetched with each request. - PageSize int64 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int64 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A token identifying the page of results to return. If unspecified, the // first page is retrieved. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -295,11 +295,11 @@ func (m *ListCasesRequest) GetPageToken() string { // The response message for `ListCase` method. type ListCasesResponse struct { // A list of cases. - Cases []*common.Case `protobuf:"bytes,1,rep,name=cases" json:"cases,omitempty"` + Cases []*common.Case `protobuf:"bytes,1,rep,name=cases,proto3" json:"cases,omitempty"` // A token to retrieve the next page of results. This should be passed on in // `page_token` field of `ListCaseRequest` for next request. If unspecified, // there are no more results to retrieve. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -346,7 +346,7 @@ func (m *ListCasesResponse) GetNextPageToken() string { // The request message for `ListComments` method. type ListCommentsRequest struct { // The resource name of case for which comments should be listed. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -386,7 +386,7 @@ func (m *ListCommentsRequest) GetName() string { // The response message for `ListComments` method. type ListCommentsResponse struct { // A list of comments. - Comments []*common.Comment `protobuf:"bytes,1,rep,name=comments" json:"comments,omitempty"` + Comments []*common.Comment `protobuf:"bytes,1,rep,name=comments,proto3" json:"comments,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -426,9 +426,9 @@ func (m *ListCommentsResponse) GetComments() []*common.Comment { // The request message for `CreateCase` method. type CreateCaseRequest struct { // The resource name for `SupportAccount` under which this case is created. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The case resource to create. - Case *common.Case `protobuf:"bytes,2,opt,name=case" json:"case,omitempty"` + Case *common.Case `protobuf:"bytes,2,opt,name=case,proto3" json:"case,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -475,10 +475,10 @@ func (m *CreateCaseRequest) GetCase() *common.Case { // The request message for `UpdateCase` method. type UpdateCaseRequest struct { // The case resource to update. - Case *common.Case `protobuf:"bytes,1,opt,name=case" json:"case,omitempty"` + Case *common.Case `protobuf:"bytes,1,opt,name=case,proto3" json:"case,omitempty"` // A field that represents attributes of a Case object that should be updated // as part of this request. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -525,9 +525,9 @@ func (m *UpdateCaseRequest) GetUpdateMask() *field_mask.FieldMask { // The request message for `CreateComment` method. type CreateCommentRequest struct { // The resource name of case to which this comment should be added. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The `Comment` to be added to this case. - Comment *common.Comment `protobuf:"bytes,2,opt,name=comment" json:"comment,omitempty"` + Comment *common.Comment `protobuf:"bytes,2,opt,name=comment,proto3" json:"comment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -747,8 +747,7 @@ func (c *cloudSupportClient) GetIssueTaxonomy(ctx context.Context, in *GetIssueT return out, nil } -// Server API for CloudSupport service - +// CloudSupportServer is the server API for CloudSupport service. type CloudSupportServer interface { // Retrieves the support account details given an account identifier. // The authenticated user calling this method must be the account owner. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/cloudtasks.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/cloudtasks.pb.go new file mode 100644 index 000000000..e6f4728cf --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/cloudtasks.pb.go @@ -0,0 +1,2545 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/tasks/v2beta2/cloudtasks.proto + +package tasks // import "google.golang.org/genproto/googleapis/cloud/tasks/v2beta2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" +import empty "github.com/golang/protobuf/ptypes/empty" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" +import _ "google.golang.org/genproto/googleapis/api/annotations" +import v1 "google.golang.org/genproto/googleapis/iam/v1" +import _ "google.golang.org/genproto/googleapis/rpc/code" +import field_mask "google.golang.org/genproto/protobuf/field_mask" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Request message for [ListQueues][google.cloud.tasks.v2beta2.CloudTasks.ListQueues]. +type ListQueuesRequest struct { + // Required. + // + // The location name. + // For example: `projects/PROJECT_ID/locations/LOCATION_ID` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // `filter` can be used to specify a subset of queues. Any [Queue][google.cloud.tasks.v2beta2.Queue] + // field can be used as a filter and several operators as supported. + // For example: `<=, <, >=, >, !=, =, :`. The filter syntax is the same as + // described in + // [Stackdriver's Advanced Logs Filters](/logging/docs/view/advanced_filters). + // + // Sample filter "app_engine_http_target: *". + // + // Note that using filters might cause fewer queues than the + // requested_page size to be returned. + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` + // Requested page size. + // + // The maximum page size is 9800. If unspecified, the page size will + // be the maximum. Fewer queues than requested might be returned, + // even if more queues exist; use the + // [next_page_token][google.cloud.tasks.v2beta2.ListQueuesResponse.next_page_token] in the + // response to determine if more queues exist. + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // A token identifying the page of results to return. + // + // To request the first page results, page_token must be empty. To + // request the next page of results, page_token must be the value of + // [next_page_token][google.cloud.tasks.v2beta2.ListQueuesResponse.next_page_token] returned + // from the previous call to [ListQueues][google.cloud.tasks.v2beta2.CloudTasks.ListQueues] + // method. It is an error to switch the value of the + // [filter][google.cloud.tasks.v2beta2.ListQueuesRequest.filter] while iterating through pages. + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListQueuesRequest) Reset() { *m = ListQueuesRequest{} } +func (m *ListQueuesRequest) String() string { return proto.CompactTextString(m) } +func (*ListQueuesRequest) ProtoMessage() {} +func (*ListQueuesRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{0} +} +func (m *ListQueuesRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListQueuesRequest.Unmarshal(m, b) +} +func (m *ListQueuesRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListQueuesRequest.Marshal(b, m, deterministic) +} +func (dst *ListQueuesRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListQueuesRequest.Merge(dst, src) +} +func (m *ListQueuesRequest) XXX_Size() int { + return xxx_messageInfo_ListQueuesRequest.Size(m) +} +func (m *ListQueuesRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListQueuesRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListQueuesRequest proto.InternalMessageInfo + +func (m *ListQueuesRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListQueuesRequest) GetFilter() string { + if m != nil { + return m.Filter + } + return "" +} + +func (m *ListQueuesRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListQueuesRequest) GetPageToken() string { + if m != nil { + return m.PageToken + } + return "" +} + +// Response message for [ListQueues][google.cloud.tasks.v2beta2.CloudTasks.ListQueues]. +type ListQueuesResponse struct { + // The list of queues. + Queues []*Queue `protobuf:"bytes,1,rep,name=queues,proto3" json:"queues,omitempty"` + // A token to retrieve next page of results. + // + // To return the next page of results, call + // [ListQueues][google.cloud.tasks.v2beta2.CloudTasks.ListQueues] with this value as the + // [page_token][google.cloud.tasks.v2beta2.ListQueuesRequest.page_token]. + // + // If the next_page_token is empty, there are no more results. + // + // The page token is valid for only 2 hours. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListQueuesResponse) Reset() { *m = ListQueuesResponse{} } +func (m *ListQueuesResponse) String() string { return proto.CompactTextString(m) } +func (*ListQueuesResponse) ProtoMessage() {} +func (*ListQueuesResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{1} +} +func (m *ListQueuesResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListQueuesResponse.Unmarshal(m, b) +} +func (m *ListQueuesResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListQueuesResponse.Marshal(b, m, deterministic) +} +func (dst *ListQueuesResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListQueuesResponse.Merge(dst, src) +} +func (m *ListQueuesResponse) XXX_Size() int { + return xxx_messageInfo_ListQueuesResponse.Size(m) +} +func (m *ListQueuesResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListQueuesResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListQueuesResponse proto.InternalMessageInfo + +func (m *ListQueuesResponse) GetQueues() []*Queue { + if m != nil { + return m.Queues + } + return nil +} + +func (m *ListQueuesResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request message for [GetQueue][google.cloud.tasks.v2beta2.CloudTasks.GetQueue]. +type GetQueueRequest struct { + // Required. + // + // The resource name of the queue. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetQueueRequest) Reset() { *m = GetQueueRequest{} } +func (m *GetQueueRequest) String() string { return proto.CompactTextString(m) } +func (*GetQueueRequest) ProtoMessage() {} +func (*GetQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{2} +} +func (m *GetQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetQueueRequest.Unmarshal(m, b) +} +func (m *GetQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetQueueRequest.Marshal(b, m, deterministic) +} +func (dst *GetQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetQueueRequest.Merge(dst, src) +} +func (m *GetQueueRequest) XXX_Size() int { + return xxx_messageInfo_GetQueueRequest.Size(m) +} +func (m *GetQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetQueueRequest proto.InternalMessageInfo + +func (m *GetQueueRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [CreateQueue][google.cloud.tasks.v2beta2.CloudTasks.CreateQueue]. +type CreateQueueRequest struct { + // Required. + // + // The location name in which the queue will be created. + // For example: `projects/PROJECT_ID/locations/LOCATION_ID` + // + // The list of allowed locations can be obtained by calling Cloud + // Tasks' implementation of + // [ListLocations][google.cloud.location.Locations.ListLocations]. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Required. + // + // The queue to create. + // + // [Queue's name][google.cloud.tasks.v2beta2.Queue.name] cannot be the same as an existing queue. + Queue *Queue `protobuf:"bytes,2,opt,name=queue,proto3" json:"queue,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CreateQueueRequest) Reset() { *m = CreateQueueRequest{} } +func (m *CreateQueueRequest) String() string { return proto.CompactTextString(m) } +func (*CreateQueueRequest) ProtoMessage() {} +func (*CreateQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{3} +} +func (m *CreateQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CreateQueueRequest.Unmarshal(m, b) +} +func (m *CreateQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CreateQueueRequest.Marshal(b, m, deterministic) +} +func (dst *CreateQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CreateQueueRequest.Merge(dst, src) +} +func (m *CreateQueueRequest) XXX_Size() int { + return xxx_messageInfo_CreateQueueRequest.Size(m) +} +func (m *CreateQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CreateQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CreateQueueRequest proto.InternalMessageInfo + +func (m *CreateQueueRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *CreateQueueRequest) GetQueue() *Queue { + if m != nil { + return m.Queue + } + return nil +} + +// Request message for [UpdateQueue][google.cloud.tasks.v2beta2.CloudTasks.UpdateQueue]. +type UpdateQueueRequest struct { + // Required. + // + // The queue to create or update. + // + // The queue's [name][google.cloud.tasks.v2beta2.Queue.name] must be specified. + // + // Output only fields cannot be modified using UpdateQueue. + // Any value specified for an output only field will be ignored. + // The queue's [name][google.cloud.tasks.v2beta2.Queue.name] cannot be changed. + Queue *Queue `protobuf:"bytes,1,opt,name=queue,proto3" json:"queue,omitempty"` + // A mask used to specify which fields of the queue are being updated. + // + // If empty, then all fields will be updated. + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *UpdateQueueRequest) Reset() { *m = UpdateQueueRequest{} } +func (m *UpdateQueueRequest) String() string { return proto.CompactTextString(m) } +func (*UpdateQueueRequest) ProtoMessage() {} +func (*UpdateQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{4} +} +func (m *UpdateQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_UpdateQueueRequest.Unmarshal(m, b) +} +func (m *UpdateQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_UpdateQueueRequest.Marshal(b, m, deterministic) +} +func (dst *UpdateQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_UpdateQueueRequest.Merge(dst, src) +} +func (m *UpdateQueueRequest) XXX_Size() int { + return xxx_messageInfo_UpdateQueueRequest.Size(m) +} +func (m *UpdateQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_UpdateQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_UpdateQueueRequest proto.InternalMessageInfo + +func (m *UpdateQueueRequest) GetQueue() *Queue { + if m != nil { + return m.Queue + } + return nil +} + +func (m *UpdateQueueRequest) GetUpdateMask() *field_mask.FieldMask { + if m != nil { + return m.UpdateMask + } + return nil +} + +// Request message for [DeleteQueue][google.cloud.tasks.v2beta2.CloudTasks.DeleteQueue]. +type DeleteQueueRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DeleteQueueRequest) Reset() { *m = DeleteQueueRequest{} } +func (m *DeleteQueueRequest) String() string { return proto.CompactTextString(m) } +func (*DeleteQueueRequest) ProtoMessage() {} +func (*DeleteQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{5} +} +func (m *DeleteQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DeleteQueueRequest.Unmarshal(m, b) +} +func (m *DeleteQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DeleteQueueRequest.Marshal(b, m, deterministic) +} +func (dst *DeleteQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_DeleteQueueRequest.Merge(dst, src) +} +func (m *DeleteQueueRequest) XXX_Size() int { + return xxx_messageInfo_DeleteQueueRequest.Size(m) +} +func (m *DeleteQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_DeleteQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_DeleteQueueRequest proto.InternalMessageInfo + +func (m *DeleteQueueRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [PurgeQueue][google.cloud.tasks.v2beta2.CloudTasks.PurgeQueue]. +type PurgeQueueRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/location/LOCATION_ID/queues/QUEUE_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *PurgeQueueRequest) Reset() { *m = PurgeQueueRequest{} } +func (m *PurgeQueueRequest) String() string { return proto.CompactTextString(m) } +func (*PurgeQueueRequest) ProtoMessage() {} +func (*PurgeQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{6} +} +func (m *PurgeQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_PurgeQueueRequest.Unmarshal(m, b) +} +func (m *PurgeQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_PurgeQueueRequest.Marshal(b, m, deterministic) +} +func (dst *PurgeQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_PurgeQueueRequest.Merge(dst, src) +} +func (m *PurgeQueueRequest) XXX_Size() int { + return xxx_messageInfo_PurgeQueueRequest.Size(m) +} +func (m *PurgeQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_PurgeQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_PurgeQueueRequest proto.InternalMessageInfo + +func (m *PurgeQueueRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [PauseQueue][google.cloud.tasks.v2beta2.CloudTasks.PauseQueue]. +type PauseQueueRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/location/LOCATION_ID/queues/QUEUE_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *PauseQueueRequest) Reset() { *m = PauseQueueRequest{} } +func (m *PauseQueueRequest) String() string { return proto.CompactTextString(m) } +func (*PauseQueueRequest) ProtoMessage() {} +func (*PauseQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{7} +} +func (m *PauseQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_PauseQueueRequest.Unmarshal(m, b) +} +func (m *PauseQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_PauseQueueRequest.Marshal(b, m, deterministic) +} +func (dst *PauseQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_PauseQueueRequest.Merge(dst, src) +} +func (m *PauseQueueRequest) XXX_Size() int { + return xxx_messageInfo_PauseQueueRequest.Size(m) +} +func (m *PauseQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_PauseQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_PauseQueueRequest proto.InternalMessageInfo + +func (m *PauseQueueRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for [ResumeQueue][google.cloud.tasks.v2beta2.CloudTasks.ResumeQueue]. +type ResumeQueueRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/location/LOCATION_ID/queues/QUEUE_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ResumeQueueRequest) Reset() { *m = ResumeQueueRequest{} } +func (m *ResumeQueueRequest) String() string { return proto.CompactTextString(m) } +func (*ResumeQueueRequest) ProtoMessage() {} +func (*ResumeQueueRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{8} +} +func (m *ResumeQueueRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ResumeQueueRequest.Unmarshal(m, b) +} +func (m *ResumeQueueRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ResumeQueueRequest.Marshal(b, m, deterministic) +} +func (dst *ResumeQueueRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ResumeQueueRequest.Merge(dst, src) +} +func (m *ResumeQueueRequest) XXX_Size() int { + return xxx_messageInfo_ResumeQueueRequest.Size(m) +} +func (m *ResumeQueueRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ResumeQueueRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ResumeQueueRequest proto.InternalMessageInfo + +func (m *ResumeQueueRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for listing tasks using [ListTasks][google.cloud.tasks.v2beta2.CloudTasks.ListTasks]. +type ListTasksRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,2,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + // Sort order used for the query. The only fields supported for sorting + // are `schedule_time` and `pull_message.tag`. All results will be + // returned in approximately ascending order. The default ordering is by + // `schedule_time`. + OrderBy string `protobuf:"bytes,3,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` + // Requested page size. Fewer tasks than requested might be returned. + // + // The maximum page size is 1000. If unspecified, the page size will + // be the maximum. Fewer tasks than requested might be returned, + // even if more tasks exist; use + // [next_page_token][google.cloud.tasks.v2beta2.ListTasksResponse.next_page_token] in the + // response to determine if more tasks exist. + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // A token identifying the page of results to return. + // + // To request the first page results, page_token must be empty. To + // request the next page of results, page_token must be the value of + // [next_page_token][google.cloud.tasks.v2beta2.ListTasksResponse.next_page_token] returned + // from the previous call to [ListTasks][google.cloud.tasks.v2beta2.CloudTasks.ListTasks] + // method. + // + // The page token is valid for only 2 hours. + PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListTasksRequest) Reset() { *m = ListTasksRequest{} } +func (m *ListTasksRequest) String() string { return proto.CompactTextString(m) } +func (*ListTasksRequest) ProtoMessage() {} +func (*ListTasksRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{9} +} +func (m *ListTasksRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListTasksRequest.Unmarshal(m, b) +} +func (m *ListTasksRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListTasksRequest.Marshal(b, m, deterministic) +} +func (dst *ListTasksRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListTasksRequest.Merge(dst, src) +} +func (m *ListTasksRequest) XXX_Size() int { + return xxx_messageInfo_ListTasksRequest.Size(m) +} +func (m *ListTasksRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListTasksRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListTasksRequest proto.InternalMessageInfo + +func (m *ListTasksRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListTasksRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +func (m *ListTasksRequest) GetOrderBy() string { + if m != nil { + return m.OrderBy + } + return "" +} + +func (m *ListTasksRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListTasksRequest) GetPageToken() string { + if m != nil { + return m.PageToken + } + return "" +} + +// Response message for listing tasks using [ListTasks][google.cloud.tasks.v2beta2.CloudTasks.ListTasks]. +type ListTasksResponse struct { + // The list of tasks. + Tasks []*Task `protobuf:"bytes,1,rep,name=tasks,proto3" json:"tasks,omitempty"` + // A token to retrieve next page of results. + // + // To return the next page of results, call + // [ListTasks][google.cloud.tasks.v2beta2.CloudTasks.ListTasks] with this value as the + // [page_token][google.cloud.tasks.v2beta2.ListTasksRequest.page_token]. + // + // If the next_page_token is empty, there are no more results. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListTasksResponse) Reset() { *m = ListTasksResponse{} } +func (m *ListTasksResponse) String() string { return proto.CompactTextString(m) } +func (*ListTasksResponse) ProtoMessage() {} +func (*ListTasksResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{10} +} +func (m *ListTasksResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListTasksResponse.Unmarshal(m, b) +} +func (m *ListTasksResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListTasksResponse.Marshal(b, m, deterministic) +} +func (dst *ListTasksResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListTasksResponse.Merge(dst, src) +} +func (m *ListTasksResponse) XXX_Size() int { + return xxx_messageInfo_ListTasksResponse.Size(m) +} +func (m *ListTasksResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListTasksResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListTasksResponse proto.InternalMessageInfo + +func (m *ListTasksResponse) GetTasks() []*Task { + if m != nil { + return m.Tasks + } + return nil +} + +func (m *ListTasksResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request message for getting a task using [GetTask][google.cloud.tasks.v2beta2.CloudTasks.GetTask]. +type GetTaskRequest struct { + // Required. + // + // The task name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,2,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetTaskRequest) Reset() { *m = GetTaskRequest{} } +func (m *GetTaskRequest) String() string { return proto.CompactTextString(m) } +func (*GetTaskRequest) ProtoMessage() {} +func (*GetTaskRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{11} +} +func (m *GetTaskRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetTaskRequest.Unmarshal(m, b) +} +func (m *GetTaskRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetTaskRequest.Marshal(b, m, deterministic) +} +func (dst *GetTaskRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetTaskRequest.Merge(dst, src) +} +func (m *GetTaskRequest) XXX_Size() int { + return xxx_messageInfo_GetTaskRequest.Size(m) +} +func (m *GetTaskRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetTaskRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetTaskRequest proto.InternalMessageInfo + +func (m *GetTaskRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *GetTaskRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +// Request message for [CreateTask][google.cloud.tasks.v2beta2.CloudTasks.CreateTask]. +type CreateTaskRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID` + // + // The queue must already exist. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // Required. + // + // The task to add. + // + // Task names have the following format: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID`. + // The user can optionally specify a task [name][google.cloud.tasks.v2beta2.Task.name]. If a + // name is not specified then the system will generate a random + // unique task id, which will be set in the task returned in the + // [response][google.cloud.tasks.v2beta2.Task.name]. + // + // If [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] is not set or is in the + // past then Cloud Tasks will set it to the current time. + // + // Task De-duplication: + // + // Explicitly specifying a task ID enables task de-duplication. If + // a task's ID is identical to that of an existing task or a task + // that was deleted or completed recently then the call will fail + // with [ALREADY_EXISTS][google.rpc.Code.ALREADY_EXISTS]. + // If the task's queue was created using Cloud Tasks, then another task with + // the same name can't be created for ~1hour after the original task was + // deleted or completed. If the task's queue was created using queue.yaml or + // queue.xml, then another task with the same name can't be created + // for ~9days after the original task was deleted or completed. + // + // Because there is an extra lookup cost to identify duplicate task + // names, these [CreateTask][google.cloud.tasks.v2beta2.CloudTasks.CreateTask] calls have significantly + // increased latency. Using hashed strings for the task id or for + // the prefix of the task id is recommended. Choosing task ids that + // are sequential or have sequential prefixes, for example using a + // timestamp, causes an increase in latency and error rates in all + // task commands. The infrastructure relies on an approximately + // uniform distribution of task ids to store and serve tasks + // efficiently. + Task *Task `protobuf:"bytes,2,opt,name=task,proto3" json:"task,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,3,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CreateTaskRequest) Reset() { *m = CreateTaskRequest{} } +func (m *CreateTaskRequest) String() string { return proto.CompactTextString(m) } +func (*CreateTaskRequest) ProtoMessage() {} +func (*CreateTaskRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{12} +} +func (m *CreateTaskRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CreateTaskRequest.Unmarshal(m, b) +} +func (m *CreateTaskRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CreateTaskRequest.Marshal(b, m, deterministic) +} +func (dst *CreateTaskRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CreateTaskRequest.Merge(dst, src) +} +func (m *CreateTaskRequest) XXX_Size() int { + return xxx_messageInfo_CreateTaskRequest.Size(m) +} +func (m *CreateTaskRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CreateTaskRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CreateTaskRequest proto.InternalMessageInfo + +func (m *CreateTaskRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *CreateTaskRequest) GetTask() *Task { + if m != nil { + return m.Task + } + return nil +} + +func (m *CreateTaskRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +// Request message for deleting a task using +// [DeleteTask][google.cloud.tasks.v2beta2.CloudTasks.DeleteTask]. +type DeleteTaskRequest struct { + // Required. + // + // The task name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DeleteTaskRequest) Reset() { *m = DeleteTaskRequest{} } +func (m *DeleteTaskRequest) String() string { return proto.CompactTextString(m) } +func (*DeleteTaskRequest) ProtoMessage() {} +func (*DeleteTaskRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{13} +} +func (m *DeleteTaskRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DeleteTaskRequest.Unmarshal(m, b) +} +func (m *DeleteTaskRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DeleteTaskRequest.Marshal(b, m, deterministic) +} +func (dst *DeleteTaskRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_DeleteTaskRequest.Merge(dst, src) +} +func (m *DeleteTaskRequest) XXX_Size() int { + return xxx_messageInfo_DeleteTaskRequest.Size(m) +} +func (m *DeleteTaskRequest) XXX_DiscardUnknown() { + xxx_messageInfo_DeleteTaskRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_DeleteTaskRequest proto.InternalMessageInfo + +func (m *DeleteTaskRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request message for leasing tasks using [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks]. +type LeaseTasksRequest struct { + // Required. + // + // The queue name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The maximum number of tasks to lease. + // + // The system will make a best effort to return as close to as + // `max_tasks` as possible. + // + // The largest that `max_tasks` can be is 1000. + MaxTasks int32 `protobuf:"varint,2,opt,name=max_tasks,json=maxTasks,proto3" json:"max_tasks,omitempty"` + // + // After the worker has successfully finished the work associated + // with the task, the worker must call via + // [AcknowledgeTask][google.cloud.tasks.v2beta2.CloudTasks.AcknowledgeTask] before the + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time]. Otherwise the task will be + // returned to a later [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] call so + // that another worker can retry it. + // + // The maximum lease duration is 1 week. + // `lease_duration` will be truncated to the nearest second. + LeaseDuration *duration.Duration `protobuf:"bytes,3,opt,name=lease_duration,json=leaseDuration,proto3" json:"lease_duration,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,4,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + // `filter` can be used to specify a subset of tasks to lease. + // + // When `filter` is set to `tag=` then the + // [response][google.cloud.tasks.v2beta2.LeaseTasksResponse] will contain only tasks whose + // [tag][google.cloud.tasks.v2beta2.PullMessage.tag] is equal to ``. `` must be + // less than 500 characters. + // + // When `filter` is set to `tag_function=oldest_tag()`, only tasks which have + // the same tag as the task with the oldest + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] will be returned. + // + // Grammar Syntax: + // + // * `filter = "tag=" tag | "tag_function=" function` + // + // * `tag = string` + // + // * `function = "oldest_tag()"` + // + // The `oldest_tag()` function returns tasks which have the same tag as the + // oldest task (ordered by schedule time). + // + // SDK compatibility: Although the SDK allows tags to be either + // string or + // [bytes](/appengine/docs/standard/java/javadoc/com/google/appengine/api/taskqueue/TaskOptions.html#tag-byte:A-), + // only UTF-8 encoded tags can be used in Cloud Tasks. Tag which + // aren't UTF-8 encoded can't be used in the + // [filter][google.cloud.tasks.v2beta2.LeaseTasksRequest.filter] and the task's + // [tag][google.cloud.tasks.v2beta2.PullMessage.tag] will be displayed as empty in Cloud Tasks. + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LeaseTasksRequest) Reset() { *m = LeaseTasksRequest{} } +func (m *LeaseTasksRequest) String() string { return proto.CompactTextString(m) } +func (*LeaseTasksRequest) ProtoMessage() {} +func (*LeaseTasksRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{14} +} +func (m *LeaseTasksRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LeaseTasksRequest.Unmarshal(m, b) +} +func (m *LeaseTasksRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LeaseTasksRequest.Marshal(b, m, deterministic) +} +func (dst *LeaseTasksRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_LeaseTasksRequest.Merge(dst, src) +} +func (m *LeaseTasksRequest) XXX_Size() int { + return xxx_messageInfo_LeaseTasksRequest.Size(m) +} +func (m *LeaseTasksRequest) XXX_DiscardUnknown() { + xxx_messageInfo_LeaseTasksRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_LeaseTasksRequest proto.InternalMessageInfo + +func (m *LeaseTasksRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *LeaseTasksRequest) GetMaxTasks() int32 { + if m != nil { + return m.MaxTasks + } + return 0 +} + +func (m *LeaseTasksRequest) GetLeaseDuration() *duration.Duration { + if m != nil { + return m.LeaseDuration + } + return nil +} + +func (m *LeaseTasksRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +func (m *LeaseTasksRequest) GetFilter() string { + if m != nil { + return m.Filter + } + return "" +} + +// Response message for leasing tasks using [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks]. +type LeaseTasksResponse struct { + // The leased tasks. + Tasks []*Task `protobuf:"bytes,1,rep,name=tasks,proto3" json:"tasks,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LeaseTasksResponse) Reset() { *m = LeaseTasksResponse{} } +func (m *LeaseTasksResponse) String() string { return proto.CompactTextString(m) } +func (*LeaseTasksResponse) ProtoMessage() {} +func (*LeaseTasksResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{15} +} +func (m *LeaseTasksResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LeaseTasksResponse.Unmarshal(m, b) +} +func (m *LeaseTasksResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LeaseTasksResponse.Marshal(b, m, deterministic) +} +func (dst *LeaseTasksResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_LeaseTasksResponse.Merge(dst, src) +} +func (m *LeaseTasksResponse) XXX_Size() int { + return xxx_messageInfo_LeaseTasksResponse.Size(m) +} +func (m *LeaseTasksResponse) XXX_DiscardUnknown() { + xxx_messageInfo_LeaseTasksResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_LeaseTasksResponse proto.InternalMessageInfo + +func (m *LeaseTasksResponse) GetTasks() []*Task { + if m != nil { + return m.Tasks + } + return nil +} + +// Request message for acknowledging a task using +// [AcknowledgeTask][google.cloud.tasks.v2beta2.CloudTasks.AcknowledgeTask]. +type AcknowledgeTaskRequest struct { + // Required. + // + // The task name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Required. + // + // The task's current schedule time, available in the + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] returned by + // [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] response or + // [RenewLease][google.cloud.tasks.v2beta2.CloudTasks.RenewLease] response. This restriction is + // to ensure that your worker currently holds the lease. + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AcknowledgeTaskRequest) Reset() { *m = AcknowledgeTaskRequest{} } +func (m *AcknowledgeTaskRequest) String() string { return proto.CompactTextString(m) } +func (*AcknowledgeTaskRequest) ProtoMessage() {} +func (*AcknowledgeTaskRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{16} +} +func (m *AcknowledgeTaskRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AcknowledgeTaskRequest.Unmarshal(m, b) +} +func (m *AcknowledgeTaskRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AcknowledgeTaskRequest.Marshal(b, m, deterministic) +} +func (dst *AcknowledgeTaskRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_AcknowledgeTaskRequest.Merge(dst, src) +} +func (m *AcknowledgeTaskRequest) XXX_Size() int { + return xxx_messageInfo_AcknowledgeTaskRequest.Size(m) +} +func (m *AcknowledgeTaskRequest) XXX_DiscardUnknown() { + xxx_messageInfo_AcknowledgeTaskRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_AcknowledgeTaskRequest proto.InternalMessageInfo + +func (m *AcknowledgeTaskRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *AcknowledgeTaskRequest) GetScheduleTime() *timestamp.Timestamp { + if m != nil { + return m.ScheduleTime + } + return nil +} + +// Request message for renewing a lease using +// [RenewLease][google.cloud.tasks.v2beta2.CloudTasks.RenewLease]. +type RenewLeaseRequest struct { + // Required. + // + // The task name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Required. + // + // The task's current schedule time, available in the + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] returned by + // [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] response or + // [RenewLease][google.cloud.tasks.v2beta2.CloudTasks.RenewLease] response. This restriction is + // to ensure that your worker currently holds the lease. + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` + // Required. + // + // The desired new lease duration, starting from now. + // + // + // The maximum lease duration is 1 week. + // `lease_duration` will be truncated to the nearest second. + LeaseDuration *duration.Duration `protobuf:"bytes,3,opt,name=lease_duration,json=leaseDuration,proto3" json:"lease_duration,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,4,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RenewLeaseRequest) Reset() { *m = RenewLeaseRequest{} } +func (m *RenewLeaseRequest) String() string { return proto.CompactTextString(m) } +func (*RenewLeaseRequest) ProtoMessage() {} +func (*RenewLeaseRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{17} +} +func (m *RenewLeaseRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RenewLeaseRequest.Unmarshal(m, b) +} +func (m *RenewLeaseRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RenewLeaseRequest.Marshal(b, m, deterministic) +} +func (dst *RenewLeaseRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_RenewLeaseRequest.Merge(dst, src) +} +func (m *RenewLeaseRequest) XXX_Size() int { + return xxx_messageInfo_RenewLeaseRequest.Size(m) +} +func (m *RenewLeaseRequest) XXX_DiscardUnknown() { + xxx_messageInfo_RenewLeaseRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_RenewLeaseRequest proto.InternalMessageInfo + +func (m *RenewLeaseRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *RenewLeaseRequest) GetScheduleTime() *timestamp.Timestamp { + if m != nil { + return m.ScheduleTime + } + return nil +} + +func (m *RenewLeaseRequest) GetLeaseDuration() *duration.Duration { + if m != nil { + return m.LeaseDuration + } + return nil +} + +func (m *RenewLeaseRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +// Request message for canceling a lease using +// [CancelLease][google.cloud.tasks.v2beta2.CloudTasks.CancelLease]. +type CancelLeaseRequest struct { + // Required. + // + // The task name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Required. + // + // The task's current schedule time, available in the + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] returned by + // [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] response or + // [RenewLease][google.cloud.tasks.v2beta2.CloudTasks.RenewLease] response. This restriction is + // to ensure that your worker currently holds the lease. + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,3,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CancelLeaseRequest) Reset() { *m = CancelLeaseRequest{} } +func (m *CancelLeaseRequest) String() string { return proto.CompactTextString(m) } +func (*CancelLeaseRequest) ProtoMessage() {} +func (*CancelLeaseRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{18} +} +func (m *CancelLeaseRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CancelLeaseRequest.Unmarshal(m, b) +} +func (m *CancelLeaseRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CancelLeaseRequest.Marshal(b, m, deterministic) +} +func (dst *CancelLeaseRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_CancelLeaseRequest.Merge(dst, src) +} +func (m *CancelLeaseRequest) XXX_Size() int { + return xxx_messageInfo_CancelLeaseRequest.Size(m) +} +func (m *CancelLeaseRequest) XXX_DiscardUnknown() { + xxx_messageInfo_CancelLeaseRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_CancelLeaseRequest proto.InternalMessageInfo + +func (m *CancelLeaseRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *CancelLeaseRequest) GetScheduleTime() *timestamp.Timestamp { + if m != nil { + return m.ScheduleTime + } + return nil +} + +func (m *CancelLeaseRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +// Request message for forcing a task to run now using +// [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask]. +type RunTaskRequest struct { + // Required. + // + // The task name. For example: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The response_view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] will be + // returned. + // + // By default response_view is [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC]; not all + // information is retrieved by default because some data, such as + // payloads, might be desirable to return only when needed because + // of its large size or because of the sensitivity of data that it + // contains. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](/iam/) permission on the + // [Task][google.cloud.tasks.v2beta2.Task] resource. + ResponseView Task_View `protobuf:"varint,2,opt,name=response_view,json=responseView,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"response_view,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RunTaskRequest) Reset() { *m = RunTaskRequest{} } +func (m *RunTaskRequest) String() string { return proto.CompactTextString(m) } +func (*RunTaskRequest) ProtoMessage() {} +func (*RunTaskRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloudtasks_45e173e2f58580c4, []int{19} +} +func (m *RunTaskRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RunTaskRequest.Unmarshal(m, b) +} +func (m *RunTaskRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RunTaskRequest.Marshal(b, m, deterministic) +} +func (dst *RunTaskRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_RunTaskRequest.Merge(dst, src) +} +func (m *RunTaskRequest) XXX_Size() int { + return xxx_messageInfo_RunTaskRequest.Size(m) +} +func (m *RunTaskRequest) XXX_DiscardUnknown() { + xxx_messageInfo_RunTaskRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_RunTaskRequest proto.InternalMessageInfo + +func (m *RunTaskRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *RunTaskRequest) GetResponseView() Task_View { + if m != nil { + return m.ResponseView + } + return Task_VIEW_UNSPECIFIED +} + +func init() { + proto.RegisterType((*ListQueuesRequest)(nil), "google.cloud.tasks.v2beta2.ListQueuesRequest") + proto.RegisterType((*ListQueuesResponse)(nil), "google.cloud.tasks.v2beta2.ListQueuesResponse") + proto.RegisterType((*GetQueueRequest)(nil), "google.cloud.tasks.v2beta2.GetQueueRequest") + proto.RegisterType((*CreateQueueRequest)(nil), "google.cloud.tasks.v2beta2.CreateQueueRequest") + proto.RegisterType((*UpdateQueueRequest)(nil), "google.cloud.tasks.v2beta2.UpdateQueueRequest") + proto.RegisterType((*DeleteQueueRequest)(nil), "google.cloud.tasks.v2beta2.DeleteQueueRequest") + proto.RegisterType((*PurgeQueueRequest)(nil), "google.cloud.tasks.v2beta2.PurgeQueueRequest") + proto.RegisterType((*PauseQueueRequest)(nil), "google.cloud.tasks.v2beta2.PauseQueueRequest") + proto.RegisterType((*ResumeQueueRequest)(nil), "google.cloud.tasks.v2beta2.ResumeQueueRequest") + proto.RegisterType((*ListTasksRequest)(nil), "google.cloud.tasks.v2beta2.ListTasksRequest") + proto.RegisterType((*ListTasksResponse)(nil), "google.cloud.tasks.v2beta2.ListTasksResponse") + proto.RegisterType((*GetTaskRequest)(nil), "google.cloud.tasks.v2beta2.GetTaskRequest") + proto.RegisterType((*CreateTaskRequest)(nil), "google.cloud.tasks.v2beta2.CreateTaskRequest") + proto.RegisterType((*DeleteTaskRequest)(nil), "google.cloud.tasks.v2beta2.DeleteTaskRequest") + proto.RegisterType((*LeaseTasksRequest)(nil), "google.cloud.tasks.v2beta2.LeaseTasksRequest") + proto.RegisterType((*LeaseTasksResponse)(nil), "google.cloud.tasks.v2beta2.LeaseTasksResponse") + proto.RegisterType((*AcknowledgeTaskRequest)(nil), "google.cloud.tasks.v2beta2.AcknowledgeTaskRequest") + proto.RegisterType((*RenewLeaseRequest)(nil), "google.cloud.tasks.v2beta2.RenewLeaseRequest") + proto.RegisterType((*CancelLeaseRequest)(nil), "google.cloud.tasks.v2beta2.CancelLeaseRequest") + proto.RegisterType((*RunTaskRequest)(nil), "google.cloud.tasks.v2beta2.RunTaskRequest") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// CloudTasksClient is the client API for CloudTasks service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type CloudTasksClient interface { + // Lists queues. + // + // Queues are returned in lexicographical order. + ListQueues(ctx context.Context, in *ListQueuesRequest, opts ...grpc.CallOption) (*ListQueuesResponse, error) + // Gets a queue. + GetQueue(ctx context.Context, in *GetQueueRequest, opts ...grpc.CallOption) (*Queue, error) + // Creates a queue. + // + // Queues created with this method allow tasks to live for a maximum of 31 + // days. After a task is 31 days old, the task will be deleted regardless of whether + // it was dispatched or not. + // + // WARNING: Using this method may have unintended side effects if you are + // using an App Engine `queue.yaml` or `queue.xml` file to manage your queues. + // Read + // [Overview of Queue Management and queue.yaml](/cloud-tasks/docs/queue-yaml) + // before using this method. + CreateQueue(ctx context.Context, in *CreateQueueRequest, opts ...grpc.CallOption) (*Queue, error) + // Updates a queue. + // + // This method creates the queue if it does not exist and updates + // the queue if it does exist. + // + // Queues created with this method allow tasks to live for a maximum of 31 + // days. After a task is 31 days old, the task will be deleted regardless of whether + // it was dispatched or not. + // + // WARNING: Using this method may have unintended side effects if you are + // using an App Engine `queue.yaml` or `queue.xml` file to manage your queues. + // Read + // [Overview of Queue Management and queue.yaml](/cloud-tasks/docs/queue-yaml) + // before using this method. + UpdateQueue(ctx context.Context, in *UpdateQueueRequest, opts ...grpc.CallOption) (*Queue, error) + // Deletes a queue. + // + // This command will delete the queue even if it has tasks in it. + // + // Note: If you delete a queue, a queue with the same name can't be created + // for 7 days. + // + // WARNING: Using this method may have unintended side effects if you are + // using an App Engine `queue.yaml` or `queue.xml` file to manage your queues. + // Read + // [Overview of Queue Management and queue.yaml](/cloud-tasks/docs/queue-yaml) + // before using this method. + DeleteQueue(ctx context.Context, in *DeleteQueueRequest, opts ...grpc.CallOption) (*empty.Empty, error) + // Purges a queue by deleting all of its tasks. + // + // All tasks created before this method is called are permanently deleted. + // + // Purge operations can take up to one minute to take effect. Tasks + // might be dispatched before the purge takes effect. A purge is irreversible. + PurgeQueue(ctx context.Context, in *PurgeQueueRequest, opts ...grpc.CallOption) (*Queue, error) + // Pauses the queue. + // + // If a queue is paused then the system will stop dispatching tasks + // until the queue is resumed via + // [ResumeQueue][google.cloud.tasks.v2beta2.CloudTasks.ResumeQueue]. Tasks can still be added + // when the queue is paused. A queue is paused if its + // [state][google.cloud.tasks.v2beta2.Queue.state] is [PAUSED][google.cloud.tasks.v2beta2.Queue.State.PAUSED]. + PauseQueue(ctx context.Context, in *PauseQueueRequest, opts ...grpc.CallOption) (*Queue, error) + // Resume a queue. + // + // This method resumes a queue after it has been + // [PAUSED][google.cloud.tasks.v2beta2.Queue.State.PAUSED] or + // [DISABLED][google.cloud.tasks.v2beta2.Queue.State.DISABLED]. The state of a queue is stored + // in the queue's [state][google.cloud.tasks.v2beta2.Queue.state]; after calling this method it + // will be set to [RUNNING][google.cloud.tasks.v2beta2.Queue.State.RUNNING]. + // + // WARNING: Resuming many high-QPS queues at the same time can + // lead to target overloading. If you are resuming high-QPS + // queues, follow the 500/50/5 pattern described in + // [Managing Cloud Tasks Scaling Risks](/cloud-tasks/pdfs/managing-cloud-tasks-scaling-risks-2017-06-05.pdf). + ResumeQueue(ctx context.Context, in *ResumeQueueRequest, opts ...grpc.CallOption) (*Queue, error) + // Gets the access control policy for a [Queue][google.cloud.tasks.v2beta2.Queue]. + // Returns an empty policy if the resource exists and does not have a policy + // set. + // + // Authorization requires the following [Google IAM](/iam) permission on the + // specified resource parent: + // + // * `cloudtasks.queues.getIamPolicy` + GetIamPolicy(ctx context.Context, in *v1.GetIamPolicyRequest, opts ...grpc.CallOption) (*v1.Policy, error) + // Sets the access control policy for a [Queue][google.cloud.tasks.v2beta2.Queue]. Replaces any existing + // policy. + // + // Note: The Cloud Console does not check queue-level IAM permissions yet. + // Project-level permissions are required to use the Cloud Console. + // + // Authorization requires the following [Google IAM](/iam) permission on the + // specified resource parent: + // + // * `cloudtasks.queues.setIamPolicy` + SetIamPolicy(ctx context.Context, in *v1.SetIamPolicyRequest, opts ...grpc.CallOption) (*v1.Policy, error) + // Returns permissions that a caller has on a [Queue][google.cloud.tasks.v2beta2.Queue]. + // If the resource does not exist, this will return an empty set of + // permissions, not a [NOT_FOUND][google.rpc.Code.NOT_FOUND] error. + // + // Note: This operation is designed to be used for building permission-aware + // UIs and command-line tools, not for authorization checking. This operation + // may "fail open" without warning. + TestIamPermissions(ctx context.Context, in *v1.TestIamPermissionsRequest, opts ...grpc.CallOption) (*v1.TestIamPermissionsResponse, error) + // Lists the tasks in a queue. + // + // By default, only the [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC] view is retrieved + // due to performance considerations; + // [response_view][google.cloud.tasks.v2beta2.ListTasksRequest.response_view] controls the + // subset of information which is returned. + ListTasks(ctx context.Context, in *ListTasksRequest, opts ...grpc.CallOption) (*ListTasksResponse, error) + // Gets a task. + GetTask(ctx context.Context, in *GetTaskRequest, opts ...grpc.CallOption) (*Task, error) + // Creates a task and adds it to a queue. + // + // Tasks cannot be updated after creation; there is no UpdateTask command. + // + // * For [App Engine queues](google.cloud.tasks.v2beta2.AppEngineHttpTarget), + // the maximum task size is 100KB. + // * For [pull queues](google.cloud.tasks.v2beta2.PullTarget), this + // the maximum task size is 1MB. + CreateTask(ctx context.Context, in *CreateTaskRequest, opts ...grpc.CallOption) (*Task, error) + // Deletes a task. + // + // A task can be deleted if it is scheduled or dispatched. A task + // cannot be deleted if it has completed successfully or permanently + // failed. + DeleteTask(ctx context.Context, in *DeleteTaskRequest, opts ...grpc.CallOption) (*empty.Empty, error) + // Leases tasks from a pull queue for + // [lease_duration][google.cloud.tasks.v2beta2.LeaseTasksRequest.lease_duration]. + // + // This method is invoked by the worker to obtain a lease. The + // worker must acknowledge the task via + // [AcknowledgeTask][google.cloud.tasks.v2beta2.CloudTasks.AcknowledgeTask] after they have + // performed the work associated with the task. + // + // The [payload][google.cloud.tasks.v2beta2.PullMessage.payload] is intended to store data that + // the worker needs to perform the work associated with the task. To + // return the payloads in the [response][google.cloud.tasks.v2beta2.LeaseTasksResponse], set + // [response_view][google.cloud.tasks.v2beta2.LeaseTasksRequest.response_view] to + // [FULL][google.cloud.tasks.v2beta2.Task.View.FULL]. + // + // A maximum of 10 qps of [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] + // requests are allowed per + // queue. [RESOURCE_EXHAUSTED][google.rpc.Code.RESOURCE_EXHAUSTED] + // is returned when this limit is + // exceeded. [RESOURCE_EXHAUSTED][google.rpc.Code.RESOURCE_EXHAUSTED] + // is also returned when + // [max_tasks_dispatched_per_second][google.cloud.tasks.v2beta2.RateLimits.max_tasks_dispatched_per_second] + // is exceeded. + LeaseTasks(ctx context.Context, in *LeaseTasksRequest, opts ...grpc.CallOption) (*LeaseTasksResponse, error) + // Acknowledges a pull task. + // + // The worker, that is, the entity that + // [leased][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] this task must call this method + // to indicate that the work associated with the task has finished. + // + // The worker must acknowledge a task within the + // [lease_duration][google.cloud.tasks.v2beta2.LeaseTasksRequest.lease_duration] or the lease + // will expire and the task will become available to be leased + // again. After the task is acknowledged, it will not be returned + // by a later [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks], + // [GetTask][google.cloud.tasks.v2beta2.CloudTasks.GetTask], or + // [ListTasks][google.cloud.tasks.v2beta2.CloudTasks.ListTasks]. + AcknowledgeTask(ctx context.Context, in *AcknowledgeTaskRequest, opts ...grpc.CallOption) (*empty.Empty, error) + // Renew the current lease of a pull task. + // + // The worker can use this method to extend the lease by a new + // duration, starting from now. The new task lease will be + // returned in the task's [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time]. + RenewLease(ctx context.Context, in *RenewLeaseRequest, opts ...grpc.CallOption) (*Task, error) + // Cancel a pull task's lease. + // + // The worker can use this method to cancel a task's lease by + // setting its [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] to now. This will + // make the task available to be leased to the next caller of + // [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks]. + CancelLease(ctx context.Context, in *CancelLeaseRequest, opts ...grpc.CallOption) (*Task, error) + // Forces a task to run now. + // + // When this method is called, Cloud Tasks will dispatch the task, even if + // the task is already running, the queue has reached its [RateLimits][google.cloud.tasks.v2beta2.RateLimits] or + // is [PAUSED][google.cloud.tasks.v2beta2.Queue.State.PAUSED]. + // + // This command is meant to be used for manual debugging. For + // example, [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] can be used to retry a failed + // task after a fix has been made or to manually force a task to be + // dispatched now. + // + // The dispatched task is returned. That is, the task that is returned + // contains the [status][google.cloud.tasks.v2beta2.Task.status] after the task is dispatched but + // before the task is received by its target. + // + // If Cloud Tasks receives a successful response from the task's + // target, then the task will be deleted; otherwise the task's + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] will be reset to the time that + // [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] was called plus the retry delay specified + // in the queue's [RetryConfig][google.cloud.tasks.v2beta2.RetryConfig]. + // + // [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] returns + // [NOT_FOUND][google.rpc.Code.NOT_FOUND] when it is called on a + // task that has already succeeded or permanently failed. + // + // [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] cannot be called on a + // [pull task][google.cloud.tasks.v2beta2.PullMessage]. + RunTask(ctx context.Context, in *RunTaskRequest, opts ...grpc.CallOption) (*Task, error) +} + +type cloudTasksClient struct { + cc *grpc.ClientConn +} + +func NewCloudTasksClient(cc *grpc.ClientConn) CloudTasksClient { + return &cloudTasksClient{cc} +} + +func (c *cloudTasksClient) ListQueues(ctx context.Context, in *ListQueuesRequest, opts ...grpc.CallOption) (*ListQueuesResponse, error) { + out := new(ListQueuesResponse) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/ListQueues", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) GetQueue(ctx context.Context, in *GetQueueRequest, opts ...grpc.CallOption) (*Queue, error) { + out := new(Queue) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/GetQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) CreateQueue(ctx context.Context, in *CreateQueueRequest, opts ...grpc.CallOption) (*Queue, error) { + out := new(Queue) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/CreateQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) UpdateQueue(ctx context.Context, in *UpdateQueueRequest, opts ...grpc.CallOption) (*Queue, error) { + out := new(Queue) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/UpdateQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) DeleteQueue(ctx context.Context, in *DeleteQueueRequest, opts ...grpc.CallOption) (*empty.Empty, error) { + out := new(empty.Empty) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/DeleteQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) PurgeQueue(ctx context.Context, in *PurgeQueueRequest, opts ...grpc.CallOption) (*Queue, error) { + out := new(Queue) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/PurgeQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) PauseQueue(ctx context.Context, in *PauseQueueRequest, opts ...grpc.CallOption) (*Queue, error) { + out := new(Queue) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/PauseQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) ResumeQueue(ctx context.Context, in *ResumeQueueRequest, opts ...grpc.CallOption) (*Queue, error) { + out := new(Queue) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/ResumeQueue", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) GetIamPolicy(ctx context.Context, in *v1.GetIamPolicyRequest, opts ...grpc.CallOption) (*v1.Policy, error) { + out := new(v1.Policy) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/GetIamPolicy", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) SetIamPolicy(ctx context.Context, in *v1.SetIamPolicyRequest, opts ...grpc.CallOption) (*v1.Policy, error) { + out := new(v1.Policy) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/SetIamPolicy", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) TestIamPermissions(ctx context.Context, in *v1.TestIamPermissionsRequest, opts ...grpc.CallOption) (*v1.TestIamPermissionsResponse, error) { + out := new(v1.TestIamPermissionsResponse) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/TestIamPermissions", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) ListTasks(ctx context.Context, in *ListTasksRequest, opts ...grpc.CallOption) (*ListTasksResponse, error) { + out := new(ListTasksResponse) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/ListTasks", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) GetTask(ctx context.Context, in *GetTaskRequest, opts ...grpc.CallOption) (*Task, error) { + out := new(Task) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/GetTask", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) CreateTask(ctx context.Context, in *CreateTaskRequest, opts ...grpc.CallOption) (*Task, error) { + out := new(Task) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/CreateTask", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) DeleteTask(ctx context.Context, in *DeleteTaskRequest, opts ...grpc.CallOption) (*empty.Empty, error) { + out := new(empty.Empty) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/DeleteTask", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) LeaseTasks(ctx context.Context, in *LeaseTasksRequest, opts ...grpc.CallOption) (*LeaseTasksResponse, error) { + out := new(LeaseTasksResponse) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/LeaseTasks", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) AcknowledgeTask(ctx context.Context, in *AcknowledgeTaskRequest, opts ...grpc.CallOption) (*empty.Empty, error) { + out := new(empty.Empty) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/AcknowledgeTask", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) RenewLease(ctx context.Context, in *RenewLeaseRequest, opts ...grpc.CallOption) (*Task, error) { + out := new(Task) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/RenewLease", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) CancelLease(ctx context.Context, in *CancelLeaseRequest, opts ...grpc.CallOption) (*Task, error) { + out := new(Task) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/CancelLease", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *cloudTasksClient) RunTask(ctx context.Context, in *RunTaskRequest, opts ...grpc.CallOption) (*Task, error) { + out := new(Task) + err := c.cc.Invoke(ctx, "/google.cloud.tasks.v2beta2.CloudTasks/RunTask", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// CloudTasksServer is the server API for CloudTasks service. +type CloudTasksServer interface { + // Lists queues. + // + // Queues are returned in lexicographical order. + ListQueues(context.Context, *ListQueuesRequest) (*ListQueuesResponse, error) + // Gets a queue. + GetQueue(context.Context, *GetQueueRequest) (*Queue, error) + // Creates a queue. + // + // Queues created with this method allow tasks to live for a maximum of 31 + // days. After a task is 31 days old, the task will be deleted regardless of whether + // it was dispatched or not. + // + // WARNING: Using this method may have unintended side effects if you are + // using an App Engine `queue.yaml` or `queue.xml` file to manage your queues. + // Read + // [Overview of Queue Management and queue.yaml](/cloud-tasks/docs/queue-yaml) + // before using this method. + CreateQueue(context.Context, *CreateQueueRequest) (*Queue, error) + // Updates a queue. + // + // This method creates the queue if it does not exist and updates + // the queue if it does exist. + // + // Queues created with this method allow tasks to live for a maximum of 31 + // days. After a task is 31 days old, the task will be deleted regardless of whether + // it was dispatched or not. + // + // WARNING: Using this method may have unintended side effects if you are + // using an App Engine `queue.yaml` or `queue.xml` file to manage your queues. + // Read + // [Overview of Queue Management and queue.yaml](/cloud-tasks/docs/queue-yaml) + // before using this method. + UpdateQueue(context.Context, *UpdateQueueRequest) (*Queue, error) + // Deletes a queue. + // + // This command will delete the queue even if it has tasks in it. + // + // Note: If you delete a queue, a queue with the same name can't be created + // for 7 days. + // + // WARNING: Using this method may have unintended side effects if you are + // using an App Engine `queue.yaml` or `queue.xml` file to manage your queues. + // Read + // [Overview of Queue Management and queue.yaml](/cloud-tasks/docs/queue-yaml) + // before using this method. + DeleteQueue(context.Context, *DeleteQueueRequest) (*empty.Empty, error) + // Purges a queue by deleting all of its tasks. + // + // All tasks created before this method is called are permanently deleted. + // + // Purge operations can take up to one minute to take effect. Tasks + // might be dispatched before the purge takes effect. A purge is irreversible. + PurgeQueue(context.Context, *PurgeQueueRequest) (*Queue, error) + // Pauses the queue. + // + // If a queue is paused then the system will stop dispatching tasks + // until the queue is resumed via + // [ResumeQueue][google.cloud.tasks.v2beta2.CloudTasks.ResumeQueue]. Tasks can still be added + // when the queue is paused. A queue is paused if its + // [state][google.cloud.tasks.v2beta2.Queue.state] is [PAUSED][google.cloud.tasks.v2beta2.Queue.State.PAUSED]. + PauseQueue(context.Context, *PauseQueueRequest) (*Queue, error) + // Resume a queue. + // + // This method resumes a queue after it has been + // [PAUSED][google.cloud.tasks.v2beta2.Queue.State.PAUSED] or + // [DISABLED][google.cloud.tasks.v2beta2.Queue.State.DISABLED]. The state of a queue is stored + // in the queue's [state][google.cloud.tasks.v2beta2.Queue.state]; after calling this method it + // will be set to [RUNNING][google.cloud.tasks.v2beta2.Queue.State.RUNNING]. + // + // WARNING: Resuming many high-QPS queues at the same time can + // lead to target overloading. If you are resuming high-QPS + // queues, follow the 500/50/5 pattern described in + // [Managing Cloud Tasks Scaling Risks](/cloud-tasks/pdfs/managing-cloud-tasks-scaling-risks-2017-06-05.pdf). + ResumeQueue(context.Context, *ResumeQueueRequest) (*Queue, error) + // Gets the access control policy for a [Queue][google.cloud.tasks.v2beta2.Queue]. + // Returns an empty policy if the resource exists and does not have a policy + // set. + // + // Authorization requires the following [Google IAM](/iam) permission on the + // specified resource parent: + // + // * `cloudtasks.queues.getIamPolicy` + GetIamPolicy(context.Context, *v1.GetIamPolicyRequest) (*v1.Policy, error) + // Sets the access control policy for a [Queue][google.cloud.tasks.v2beta2.Queue]. Replaces any existing + // policy. + // + // Note: The Cloud Console does not check queue-level IAM permissions yet. + // Project-level permissions are required to use the Cloud Console. + // + // Authorization requires the following [Google IAM](/iam) permission on the + // specified resource parent: + // + // * `cloudtasks.queues.setIamPolicy` + SetIamPolicy(context.Context, *v1.SetIamPolicyRequest) (*v1.Policy, error) + // Returns permissions that a caller has on a [Queue][google.cloud.tasks.v2beta2.Queue]. + // If the resource does not exist, this will return an empty set of + // permissions, not a [NOT_FOUND][google.rpc.Code.NOT_FOUND] error. + // + // Note: This operation is designed to be used for building permission-aware + // UIs and command-line tools, not for authorization checking. This operation + // may "fail open" without warning. + TestIamPermissions(context.Context, *v1.TestIamPermissionsRequest) (*v1.TestIamPermissionsResponse, error) + // Lists the tasks in a queue. + // + // By default, only the [BASIC][google.cloud.tasks.v2beta2.Task.View.BASIC] view is retrieved + // due to performance considerations; + // [response_view][google.cloud.tasks.v2beta2.ListTasksRequest.response_view] controls the + // subset of information which is returned. + ListTasks(context.Context, *ListTasksRequest) (*ListTasksResponse, error) + // Gets a task. + GetTask(context.Context, *GetTaskRequest) (*Task, error) + // Creates a task and adds it to a queue. + // + // Tasks cannot be updated after creation; there is no UpdateTask command. + // + // * For [App Engine queues](google.cloud.tasks.v2beta2.AppEngineHttpTarget), + // the maximum task size is 100KB. + // * For [pull queues](google.cloud.tasks.v2beta2.PullTarget), this + // the maximum task size is 1MB. + CreateTask(context.Context, *CreateTaskRequest) (*Task, error) + // Deletes a task. + // + // A task can be deleted if it is scheduled or dispatched. A task + // cannot be deleted if it has completed successfully or permanently + // failed. + DeleteTask(context.Context, *DeleteTaskRequest) (*empty.Empty, error) + // Leases tasks from a pull queue for + // [lease_duration][google.cloud.tasks.v2beta2.LeaseTasksRequest.lease_duration]. + // + // This method is invoked by the worker to obtain a lease. The + // worker must acknowledge the task via + // [AcknowledgeTask][google.cloud.tasks.v2beta2.CloudTasks.AcknowledgeTask] after they have + // performed the work associated with the task. + // + // The [payload][google.cloud.tasks.v2beta2.PullMessage.payload] is intended to store data that + // the worker needs to perform the work associated with the task. To + // return the payloads in the [response][google.cloud.tasks.v2beta2.LeaseTasksResponse], set + // [response_view][google.cloud.tasks.v2beta2.LeaseTasksRequest.response_view] to + // [FULL][google.cloud.tasks.v2beta2.Task.View.FULL]. + // + // A maximum of 10 qps of [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] + // requests are allowed per + // queue. [RESOURCE_EXHAUSTED][google.rpc.Code.RESOURCE_EXHAUSTED] + // is returned when this limit is + // exceeded. [RESOURCE_EXHAUSTED][google.rpc.Code.RESOURCE_EXHAUSTED] + // is also returned when + // [max_tasks_dispatched_per_second][google.cloud.tasks.v2beta2.RateLimits.max_tasks_dispatched_per_second] + // is exceeded. + LeaseTasks(context.Context, *LeaseTasksRequest) (*LeaseTasksResponse, error) + // Acknowledges a pull task. + // + // The worker, that is, the entity that + // [leased][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] this task must call this method + // to indicate that the work associated with the task has finished. + // + // The worker must acknowledge a task within the + // [lease_duration][google.cloud.tasks.v2beta2.LeaseTasksRequest.lease_duration] or the lease + // will expire and the task will become available to be leased + // again. After the task is acknowledged, it will not be returned + // by a later [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks], + // [GetTask][google.cloud.tasks.v2beta2.CloudTasks.GetTask], or + // [ListTasks][google.cloud.tasks.v2beta2.CloudTasks.ListTasks]. + AcknowledgeTask(context.Context, *AcknowledgeTaskRequest) (*empty.Empty, error) + // Renew the current lease of a pull task. + // + // The worker can use this method to extend the lease by a new + // duration, starting from now. The new task lease will be + // returned in the task's [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time]. + RenewLease(context.Context, *RenewLeaseRequest) (*Task, error) + // Cancel a pull task's lease. + // + // The worker can use this method to cancel a task's lease by + // setting its [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] to now. This will + // make the task available to be leased to the next caller of + // [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks]. + CancelLease(context.Context, *CancelLeaseRequest) (*Task, error) + // Forces a task to run now. + // + // When this method is called, Cloud Tasks will dispatch the task, even if + // the task is already running, the queue has reached its [RateLimits][google.cloud.tasks.v2beta2.RateLimits] or + // is [PAUSED][google.cloud.tasks.v2beta2.Queue.State.PAUSED]. + // + // This command is meant to be used for manual debugging. For + // example, [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] can be used to retry a failed + // task after a fix has been made or to manually force a task to be + // dispatched now. + // + // The dispatched task is returned. That is, the task that is returned + // contains the [status][google.cloud.tasks.v2beta2.Task.status] after the task is dispatched but + // before the task is received by its target. + // + // If Cloud Tasks receives a successful response from the task's + // target, then the task will be deleted; otherwise the task's + // [schedule_time][google.cloud.tasks.v2beta2.Task.schedule_time] will be reset to the time that + // [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] was called plus the retry delay specified + // in the queue's [RetryConfig][google.cloud.tasks.v2beta2.RetryConfig]. + // + // [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] returns + // [NOT_FOUND][google.rpc.Code.NOT_FOUND] when it is called on a + // task that has already succeeded or permanently failed. + // + // [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask] cannot be called on a + // [pull task][google.cloud.tasks.v2beta2.PullMessage]. + RunTask(context.Context, *RunTaskRequest) (*Task, error) +} + +func RegisterCloudTasksServer(s *grpc.Server, srv CloudTasksServer) { + s.RegisterService(&_CloudTasks_serviceDesc, srv) +} + +func _CloudTasks_ListQueues_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListQueuesRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).ListQueues(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/ListQueues", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).ListQueues(ctx, req.(*ListQueuesRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_GetQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).GetQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/GetQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).GetQueue(ctx, req.(*GetQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_CreateQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(CreateQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).CreateQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/CreateQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).CreateQueue(ctx, req.(*CreateQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_UpdateQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(UpdateQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).UpdateQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/UpdateQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).UpdateQueue(ctx, req.(*UpdateQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_DeleteQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(DeleteQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).DeleteQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/DeleteQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).DeleteQueue(ctx, req.(*DeleteQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_PurgeQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(PurgeQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).PurgeQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/PurgeQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).PurgeQueue(ctx, req.(*PurgeQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_PauseQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(PauseQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).PauseQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/PauseQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).PauseQueue(ctx, req.(*PauseQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_ResumeQueue_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ResumeQueueRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).ResumeQueue(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/ResumeQueue", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).ResumeQueue(ctx, req.(*ResumeQueueRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_GetIamPolicy_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(v1.GetIamPolicyRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).GetIamPolicy(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/GetIamPolicy", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).GetIamPolicy(ctx, req.(*v1.GetIamPolicyRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_SetIamPolicy_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(v1.SetIamPolicyRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).SetIamPolicy(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/SetIamPolicy", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).SetIamPolicy(ctx, req.(*v1.SetIamPolicyRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_TestIamPermissions_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(v1.TestIamPermissionsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).TestIamPermissions(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/TestIamPermissions", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).TestIamPermissions(ctx, req.(*v1.TestIamPermissionsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_ListTasks_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListTasksRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).ListTasks(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/ListTasks", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).ListTasks(ctx, req.(*ListTasksRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_GetTask_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetTaskRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).GetTask(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/GetTask", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).GetTask(ctx, req.(*GetTaskRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_CreateTask_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(CreateTaskRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).CreateTask(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/CreateTask", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).CreateTask(ctx, req.(*CreateTaskRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_DeleteTask_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(DeleteTaskRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).DeleteTask(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/DeleteTask", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).DeleteTask(ctx, req.(*DeleteTaskRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_LeaseTasks_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(LeaseTasksRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).LeaseTasks(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/LeaseTasks", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).LeaseTasks(ctx, req.(*LeaseTasksRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_AcknowledgeTask_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(AcknowledgeTaskRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).AcknowledgeTask(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/AcknowledgeTask", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).AcknowledgeTask(ctx, req.(*AcknowledgeTaskRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_RenewLease_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(RenewLeaseRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).RenewLease(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/RenewLease", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).RenewLease(ctx, req.(*RenewLeaseRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_CancelLease_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(CancelLeaseRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).CancelLease(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/CancelLease", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).CancelLease(ctx, req.(*CancelLeaseRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _CloudTasks_RunTask_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(RunTaskRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(CloudTasksServer).RunTask(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.tasks.v2beta2.CloudTasks/RunTask", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(CloudTasksServer).RunTask(ctx, req.(*RunTaskRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _CloudTasks_serviceDesc = grpc.ServiceDesc{ + ServiceName: "google.cloud.tasks.v2beta2.CloudTasks", + HandlerType: (*CloudTasksServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "ListQueues", + Handler: _CloudTasks_ListQueues_Handler, + }, + { + MethodName: "GetQueue", + Handler: _CloudTasks_GetQueue_Handler, + }, + { + MethodName: "CreateQueue", + Handler: _CloudTasks_CreateQueue_Handler, + }, + { + MethodName: "UpdateQueue", + Handler: _CloudTasks_UpdateQueue_Handler, + }, + { + MethodName: "DeleteQueue", + Handler: _CloudTasks_DeleteQueue_Handler, + }, + { + MethodName: "PurgeQueue", + Handler: _CloudTasks_PurgeQueue_Handler, + }, + { + MethodName: "PauseQueue", + Handler: _CloudTasks_PauseQueue_Handler, + }, + { + MethodName: "ResumeQueue", + Handler: _CloudTasks_ResumeQueue_Handler, + }, + { + MethodName: "GetIamPolicy", + Handler: _CloudTasks_GetIamPolicy_Handler, + }, + { + MethodName: "SetIamPolicy", + Handler: _CloudTasks_SetIamPolicy_Handler, + }, + { + MethodName: "TestIamPermissions", + Handler: _CloudTasks_TestIamPermissions_Handler, + }, + { + MethodName: "ListTasks", + Handler: _CloudTasks_ListTasks_Handler, + }, + { + MethodName: "GetTask", + Handler: _CloudTasks_GetTask_Handler, + }, + { + MethodName: "CreateTask", + Handler: _CloudTasks_CreateTask_Handler, + }, + { + MethodName: "DeleteTask", + Handler: _CloudTasks_DeleteTask_Handler, + }, + { + MethodName: "LeaseTasks", + Handler: _CloudTasks_LeaseTasks_Handler, + }, + { + MethodName: "AcknowledgeTask", + Handler: _CloudTasks_AcknowledgeTask_Handler, + }, + { + MethodName: "RenewLease", + Handler: _CloudTasks_RenewLease_Handler, + }, + { + MethodName: "CancelLease", + Handler: _CloudTasks_CancelLease_Handler, + }, + { + MethodName: "RunTask", + Handler: _CloudTasks_RunTask_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "google/cloud/tasks/v2beta2/cloudtasks.proto", +} + +func init() { + proto.RegisterFile("google/cloud/tasks/v2beta2/cloudtasks.proto", fileDescriptor_cloudtasks_45e173e2f58580c4) +} + +var fileDescriptor_cloudtasks_45e173e2f58580c4 = []byte{ + // 1404 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x98, 0xcd, 0x73, 0x14, 0xc5, + 0x1b, 0xc7, 0xab, 0x43, 0x16, 0xc8, 0xb3, 0x24, 0xf9, 0x6d, 0x57, 0x41, 0x85, 0xe5, 0x27, 0xc6, + 0xae, 0x02, 0xd6, 0x45, 0x76, 0x8a, 0x55, 0x40, 0x36, 0x02, 0x61, 0x03, 0x04, 0x05, 0xac, 0x30, + 0x89, 0x1e, 0xbc, 0x6c, 0x4d, 0x66, 0x9b, 0x75, 0xcc, 0xbc, 0x31, 0x2f, 0x09, 0xa0, 0x94, 0x55, + 0x1e, 0x3d, 0x78, 0xc1, 0xb2, 0xbc, 0xf9, 0x5a, 0x56, 0x59, 0x25, 0xe5, 0x89, 0x9b, 0xff, 0x82, + 0x27, 0xff, 0x05, 0xcf, 0x96, 0x27, 0xcf, 0x56, 0xbf, 0xcc, 0xce, 0x64, 0x26, 0x3b, 0x33, 0x59, + 0x5e, 0xca, 0x53, 0xd2, 0xdd, 0x4f, 0xf7, 0xf3, 0xe9, 0xa7, 0xfb, 0xe9, 0xf9, 0x3e, 0x0b, 0x27, + 0x07, 0x8e, 0x33, 0x30, 0xa9, 0xa2, 0x9b, 0x4e, 0xd8, 0x57, 0x02, 0xcd, 0xdf, 0xf0, 0x95, 0xcd, + 0xf6, 0x3a, 0x0d, 0xb4, 0xb6, 0xe8, 0xe3, 0x5d, 0x2d, 0xd7, 0x73, 0x02, 0x07, 0xd7, 0x85, 0x71, + 0x8b, 0x0f, 0xb4, 0xc4, 0x88, 0x34, 0xae, 0xff, 0x5f, 0x2e, 0xa4, 0xb9, 0x86, 0xa2, 0xd9, 0xb6, + 0x13, 0x68, 0x81, 0xe1, 0xd8, 0x72, 0x66, 0xfd, 0x78, 0x8e, 0x9b, 0xbb, 0x21, 0x0d, 0xa9, 0xb4, + 0x3b, 0x96, 0x63, 0xc7, 0x5a, 0xd2, 0xec, 0xa8, 0x34, 0x33, 0x34, 0x4b, 0xd9, 0x3c, 0xcd, 0xfe, + 0xf4, 0x5c, 0xc7, 0x34, 0xf4, 0xfb, 0x72, 0xbc, 0xbe, 0x7d, 0x7c, 0xdb, 0x58, 0x34, 0x97, 0xb7, + 0xd6, 0xc3, 0x3b, 0x4a, 0x3f, 0xf4, 0x38, 0xab, 0x1c, 0x3f, 0x92, 0x1e, 0xa7, 0x96, 0x1b, 0x44, + 0x93, 0xe7, 0xd3, 0x83, 0x77, 0x0c, 0x6a, 0xf6, 0x7b, 0x56, 0x8c, 0xf6, 0x72, 0xda, 0x22, 0x30, + 0x2c, 0xea, 0x07, 0x9a, 0xe5, 0x4a, 0x83, 0x83, 0xd2, 0xc0, 0x73, 0x75, 0x45, 0x77, 0xfa, 0x72, + 0xe7, 0xe4, 0x53, 0xa8, 0xdd, 0x34, 0xfc, 0xe0, 0x36, 0x0b, 0x86, 0xaf, 0xd2, 0xbb, 0x21, 0xf5, + 0x03, 0x7c, 0x08, 0xf6, 0xba, 0x9a, 0x47, 0xed, 0x60, 0x0e, 0xcd, 0xa3, 0xc6, 0x94, 0x2a, 0x5b, + 0xac, 0xff, 0x8e, 0x61, 0x06, 0xd4, 0x9b, 0x9b, 0x10, 0xfd, 0xa2, 0x85, 0x8f, 0xc0, 0x94, 0xab, + 0x0d, 0x68, 0xcf, 0x37, 0x1e, 0xd0, 0xb9, 0x3d, 0xf3, 0xa8, 0x51, 0x51, 0xf7, 0xb3, 0x8e, 0x55, + 0xe3, 0x01, 0xc5, 0x2f, 0x01, 0xf0, 0xc1, 0xc0, 0xd9, 0xa0, 0xf6, 0xdc, 0x24, 0x9f, 0xc8, 0xcd, + 0xd7, 0x58, 0x07, 0xd9, 0x02, 0x9c, 0x04, 0xf0, 0x5d, 0xc7, 0xf6, 0x29, 0x3e, 0x0f, 0x7b, 0xf9, + 0xf9, 0xf8, 0x73, 0x68, 0x7e, 0x4f, 0xa3, 0xda, 0x7e, 0xa5, 0x35, 0xfa, 0x0e, 0xb4, 0xf8, 0x5c, + 0x55, 0x4e, 0xc0, 0xc7, 0x61, 0xd6, 0xa6, 0xf7, 0x82, 0x5e, 0xc2, 0xa9, 0xa0, 0x9d, 0x66, 0xdd, + 0x2b, 0x43, 0xc7, 0xc7, 0x60, 0x76, 0x99, 0x0a, 0xbf, 0xd1, 0xbe, 0x31, 0x4c, 0xda, 0x9a, 0x45, + 0xe5, 0xae, 0xf9, 0xff, 0x84, 0x02, 0x5e, 0xf2, 0xa8, 0x16, 0xd0, 0x6d, 0x96, 0xa3, 0x22, 0x74, + 0x0e, 0x2a, 0x1c, 0x83, 0xbb, 0x2c, 0x85, 0x2d, 0xec, 0xc9, 0xe7, 0x08, 0xf0, 0x7b, 0x6e, 0x3f, + 0xed, 0x67, 0xb8, 0x1e, 0xda, 0xdd, 0x7a, 0x78, 0x01, 0xaa, 0x21, 0x5f, 0x8e, 0x5f, 0x12, 0x89, + 0x53, 0x8f, 0xa6, 0x47, 0xb7, 0xa4, 0x75, 0x8d, 0xdd, 0xa3, 0x5b, 0x9a, 0xbf, 0xa1, 0x82, 0x30, + 0x67, 0xff, 0x93, 0x06, 0xe0, 0x2b, 0xd4, 0xa4, 0x29, 0x96, 0x9d, 0xa2, 0x73, 0x02, 0x6a, 0x2b, + 0xa1, 0x37, 0x28, 0x67, 0xa8, 0x85, 0x7e, 0xb1, 0x61, 0x03, 0xb0, 0x4a, 0xfd, 0xd0, 0x2a, 0xb6, + 0xfc, 0x1d, 0xc1, 0xff, 0xd8, 0xd5, 0x59, 0x63, 0x51, 0x28, 0x3a, 0x98, 0x77, 0x60, 0xda, 0x93, + 0x97, 0xab, 0xb7, 0x69, 0xd0, 0x2d, 0x1e, 0x91, 0x99, 0xf6, 0xb1, 0xbc, 0x80, 0xb2, 0x85, 0x5b, + 0xef, 0x1b, 0x74, 0x4b, 0x3d, 0x10, 0xcd, 0x65, 0x2d, 0x7c, 0x18, 0xf6, 0x3b, 0x5e, 0x9f, 0x7a, + 0xbd, 0xf5, 0xfb, 0xfc, 0xb6, 0x4f, 0xa9, 0xfb, 0x78, 0xbb, 0x7b, 0x7f, 0x7b, 0x26, 0x4c, 0xe6, + 0x66, 0x42, 0x25, 0x9d, 0x09, 0xbe, 0x48, 0x45, 0xb9, 0x1d, 0x99, 0x08, 0x67, 0xa1, 0xc2, 0xa1, + 0x64, 0x1e, 0xcc, 0x17, 0xf1, 0xaa, 0xc2, 0xbc, 0x74, 0x16, 0xb8, 0x30, 0xb3, 0x4c, 0xb9, 0xcf, + 0x9c, 0x50, 0x3f, 0xcb, 0xe8, 0x91, 0xc7, 0x08, 0x6a, 0x22, 0xa3, 0x92, 0x5e, 0x47, 0x9d, 0xdb, + 0x1b, 0x30, 0x19, 0xc4, 0x17, 0xb8, 0x78, 0xfb, 0xdc, 0x3a, 0xcb, 0xbb, 0x67, 0x7c, 0xde, 0x13, + 0x50, 0x13, 0xc9, 0x50, 0x10, 0x24, 0xf2, 0x17, 0x82, 0xda, 0x4d, 0xaa, 0xf9, 0xb4, 0xd4, 0x85, + 0x3c, 0x02, 0x53, 0x96, 0x76, 0xaf, 0x27, 0x0e, 0x77, 0x42, 0xdc, 0x14, 0x4b, 0xbb, 0xc7, 0xe7, + 0xe2, 0x45, 0x98, 0x31, 0xd9, 0x4a, 0xbd, 0xe8, 0x23, 0xc1, 0x37, 0x50, 0x6d, 0x1f, 0xce, 0x24, + 0xf0, 0x15, 0x69, 0xa0, 0x4e, 0xf3, 0x09, 0x51, 0x33, 0x1b, 0x81, 0xc9, 0xf1, 0xef, 0x7b, 0xfc, + 0xec, 0x57, 0x92, 0xcf, 0x3e, 0xb9, 0x09, 0x38, 0xb9, 0xdf, 0xa7, 0xbb, 0xb1, 0xc4, 0x82, 0x43, + 0x97, 0xf5, 0x0d, 0xdb, 0xd9, 0x32, 0x69, 0x7f, 0x50, 0x14, 0x6c, 0x7c, 0x09, 0xa6, 0x7d, 0xfd, + 0x43, 0xda, 0x0f, 0x4d, 0xda, 0x63, 0x9f, 0xba, 0x91, 0x2f, 0xdc, 0x5a, 0xf4, 0x1d, 0x54, 0x0f, + 0x44, 0x13, 0x58, 0x17, 0xf9, 0x07, 0x41, 0x4d, 0xa5, 0x36, 0xdd, 0xe2, 0x5b, 0x78, 0x9e, 0xae, + 0xfe, 0x5b, 0xa7, 0x49, 0x9e, 0x20, 0xc0, 0x4b, 0x9a, 0xad, 0x53, 0xf3, 0xf9, 0xef, 0xfc, 0x59, + 0xe6, 0xa1, 0x0b, 0x33, 0x6a, 0x68, 0xbf, 0xc0, 0x97, 0xaa, 0xfd, 0xf7, 0x61, 0x80, 0x25, 0x66, + 0x2f, 0x92, 0xf2, 0x47, 0x04, 0x10, 0x4b, 0x15, 0x7c, 0x2a, 0x6f, 0xc9, 0x8c, 0xa6, 0xaa, 0xb7, + 0xca, 0x9a, 0x0b, 0x00, 0x72, 0xee, 0xb3, 0x3f, 0xfe, 0x7c, 0x34, 0x71, 0x1a, 0x2b, 0x43, 0x21, + 0xfa, 0xb1, 0x78, 0x39, 0x2e, 0xb8, 0x9e, 0xf3, 0x11, 0xd5, 0x03, 0x5f, 0x69, 0x2a, 0xa6, 0xa3, + 0x0b, 0xbd, 0xab, 0x34, 0x1f, 0x2a, 0x52, 0xff, 0x3c, 0x42, 0xb0, 0x3f, 0x12, 0x36, 0xf8, 0x64, + 0x9e, 0xd7, 0x94, 0xfc, 0xa9, 0x17, 0xab, 0x8b, 0x9d, 0xa8, 0x58, 0xd8, 0x47, 0x30, 0x49, 0x24, + 0xa5, 0xf9, 0x10, 0x7f, 0x8b, 0xa0, 0x9a, 0xd0, 0x51, 0x38, 0x37, 0x1c, 0x59, 0xc1, 0x55, 0x86, + 0xed, 0x22, 0x67, 0x7b, 0x93, 0xec, 0x36, 0x62, 0x1d, 0x29, 0x99, 0x7e, 0x42, 0x50, 0x4d, 0x48, + 0xb0, 0x7c, 0xc4, 0xac, 0x56, 0x2b, 0x83, 0x78, 0x85, 0x23, 0x5e, 0x6c, 0x9f, 0x89, 0x11, 0x45, + 0x19, 0x52, 0x2a, 0x88, 0x11, 0xe8, 0x17, 0x08, 0xaa, 0x09, 0x7d, 0x96, 0x0f, 0x9a, 0x15, 0x72, + 0xf5, 0x43, 0x99, 0xfc, 0xbd, 0xca, 0x6a, 0x8d, 0xe8, 0x70, 0x9b, 0xe3, 0x1c, 0x2e, 0xc4, 0x32, + 0x30, 0x3f, 0x33, 0x32, 0x72, 0xb1, 0x4c, 0xdc, 0x16, 0x39, 0x59, 0x87, 0x9c, 0xd9, 0x25, 0x59, + 0xc7, 0x65, 0xde, 0x3a, 0xa8, 0x29, 0x10, 0x87, 0x02, 0xb4, 0x00, 0x31, 0x2d, 0x54, 0x9f, 0x2f, + 0x22, 0xf3, 0xc6, 0x10, 0x7f, 0x40, 0x50, 0x4d, 0x48, 0xdf, 0xfc, 0x63, 0xcd, 0x6a, 0xe4, 0x32, + 0x90, 0x97, 0x39, 0xe4, 0x02, 0x39, 0xbb, 0x5b, 0x48, 0x8f, 0xbb, 0x63, 0x94, 0x5f, 0x21, 0x38, + 0xb0, 0x4c, 0x83, 0xb7, 0x35, 0x6b, 0x85, 0x97, 0xb7, 0x98, 0x44, 0x6e, 0x0d, 0xcd, 0x6a, 0x6d, + 0x9e, 0x6e, 0x25, 0x07, 0x23, 0xb4, 0x83, 0x29, 0x1b, 0x31, 0x4a, 0x6e, 0x70, 0x9c, 0xab, 0x64, + 0x31, 0xc6, 0xf1, 0xa8, 0xef, 0x84, 0x9e, 0x5e, 0x02, 0x69, 0x90, 0xf0, 0x13, 0x81, 0xad, 0xe6, + 0x81, 0xad, 0xbe, 0x20, 0x30, 0x3f, 0x05, 0xf6, 0x1b, 0x02, 0xbc, 0x46, 0x7d, 0xde, 0x49, 0x3d, + 0xcb, 0xf0, 0x7d, 0x36, 0x05, 0x37, 0x52, 0xae, 0xb3, 0x26, 0x11, 0xe4, 0xab, 0x25, 0x2c, 0xe5, + 0x57, 0xe3, 0x36, 0x07, 0xbf, 0x41, 0xae, 0x8d, 0x03, 0x1e, 0x64, 0xd6, 0x65, 0xf8, 0x3f, 0x23, + 0x98, 0x1a, 0xd6, 0x25, 0xf8, 0xb5, 0xa2, 0xcf, 0x58, 0x52, 0xfc, 0xd6, 0x4f, 0x95, 0xb4, 0x96, + 0xf4, 0x97, 0x38, 0xfd, 0x79, 0x7c, 0xae, 0xe4, 0x0b, 0x1e, 0xb3, 0x8b, 0xdf, 0x6c, 0xf0, 0xd7, + 0x08, 0xf6, 0xc9, 0x72, 0x06, 0x37, 0x0b, 0x3e, 0x7d, 0x09, 0x25, 0x51, 0x2f, 0x14, 0xa9, 0x3b, + 0xa1, 0x95, 0xc9, 0x1c, 0xf9, 0x5b, 0x52, 0xf3, 0x21, 0xfe, 0x0e, 0x01, 0xc4, 0x65, 0x4f, 0xfe, + 0x03, 0x94, 0x29, 0x8f, 0x4a, 0x00, 0x76, 0x39, 0xe0, 0x5b, 0x64, 0xdc, 0xd8, 0xb1, 0xa3, 0xfe, + 0x12, 0x01, 0xc4, 0xb5, 0x4e, 0x3e, 0x63, 0xa6, 0x26, 0x1a, 0xf9, 0x59, 0x91, 0xa1, 0x6b, 0x8e, + 0x1d, 0xba, 0x27, 0x4c, 0x78, 0x0d, 0x0b, 0x8d, 0x02, 0xe1, 0x95, 0x2e, 0xc0, 0x0a, 0x84, 0x57, + 0xa6, 0x7e, 0x21, 0xd7, 0x39, 0x6e, 0x97, 0x5c, 0x18, 0x37, 0x90, 0x5c, 0xba, 0xb3, 0x70, 0x3e, + 0x46, 0x30, 0x9b, 0x2a, 0x69, 0x70, 0x3b, 0x8f, 0x66, 0xe7, 0xfa, 0x67, 0x64, 0x60, 0xdf, 0xe5, + 0xa4, 0xd7, 0xc9, 0xd2, 0x98, 0x81, 0xed, 0x68, 0xb1, 0x3f, 0xc6, 0xfb, 0x0b, 0x02, 0x88, 0x4b, + 0xa2, 0xfc, 0x38, 0x67, 0x4a, 0xa7, 0x12, 0x57, 0xf4, 0x16, 0xe7, 0x5d, 0x26, 0xdd, 0x71, 0x79, + 0xbd, 0xa1, 0x53, 0x86, 0xfb, 0x2b, 0x93, 0x94, 0x71, 0x21, 0x53, 0x20, 0x29, 0x33, 0x15, 0x4f, + 0x09, 0xe0, 0xa7, 0x0e, 0xb0, 0x1e, 0x7b, 0x65, 0xc4, 0xdf, 0x20, 0xd8, 0x27, 0x6b, 0x98, 0xfc, + 0xe7, 0x69, 0x7b, 0xa1, 0x53, 0x82, 0xf4, 0x1a, 0x27, 0x5d, 0x24, 0x0b, 0x63, 0x87, 0x36, 0xb4, + 0x3b, 0xa8, 0xd9, 0xfd, 0x04, 0x8e, 0xea, 0x8e, 0x95, 0xe3, 0xae, 0x3b, 0x1b, 0x57, 0x44, 0x2b, + 0xec, 0x3a, 0xae, 0xa0, 0x0f, 0x2e, 0x49, 0xf3, 0x81, 0x63, 0x6a, 0xf6, 0xa0, 0xe5, 0x78, 0x03, + 0x65, 0x40, 0x6d, 0x7e, 0x59, 0x15, 0x31, 0xa4, 0xb9, 0x86, 0xbf, 0xd3, 0x8f, 0xeb, 0x0b, 0xbc, + 0xf5, 0xfd, 0x44, 0x65, 0xed, 0xf2, 0xea, 0x8d, 0xd5, 0xf5, 0xbd, 0x7c, 0xce, 0xeb, 0xff, 0x06, + 0x00, 0x00, 0xff, 0xff, 0x41, 0xab, 0xa9, 0x48, 0x1e, 0x18, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/queue.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/queue.pb.go new file mode 100644 index 000000000..6e5bf704e --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/queue.pb.go @@ -0,0 +1,783 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/tasks/v2beta2/queue.proto + +package tasks // import "google.golang.org/genproto/googleapis/cloud/tasks/v2beta2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// State of the queue. +type Queue_State int32 + +const ( + // Unspecified state. + Queue_STATE_UNSPECIFIED Queue_State = 0 + // The queue is running. Tasks can be dispatched. + // + // If the queue was created using Cloud Tasks and the queue has + // had no activity (method calls or task dispatches) for 30 days, + // the queue may take a few minutes to re-activate. Some method + // calls may return [NOT_FOUND][google.rpc.Code.NOT_FOUND] and + // tasks may not be dispatched for a few minutes until the queue + // has been re-activated. + Queue_RUNNING Queue_State = 1 + // Tasks are paused by the user. If the queue is paused then Cloud + // Tasks will stop delivering tasks from it, but more tasks can + // still be added to it by the user. When a pull queue is paused, + // all [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] calls will return a + // [FAILED_PRECONDITION][google.rpc.Code.FAILED_PRECONDITION]. + Queue_PAUSED Queue_State = 2 + // The queue is disabled. + // + // A queue becomes `DISABLED` when + // [queue.yaml](/appengine/docs/python/config/queueref) or + // [queue.xml](appengine/docs/standard/java/config/queueref) is uploaded + // which does not contain the queue. You cannot directly disable a queue. + // + // When a queue is disabled, tasks can still be added to a queue + // but the tasks are not dispatched and + // [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] calls return a + // `FAILED_PRECONDITION` error. + // + // To permanently delete this queue and all of its tasks, call + // [DeleteQueue][google.cloud.tasks.v2beta2.CloudTasks.DeleteQueue]. + Queue_DISABLED Queue_State = 3 +) + +var Queue_State_name = map[int32]string{ + 0: "STATE_UNSPECIFIED", + 1: "RUNNING", + 2: "PAUSED", + 3: "DISABLED", +} +var Queue_State_value = map[string]int32{ + "STATE_UNSPECIFIED": 0, + "RUNNING": 1, + "PAUSED": 2, + "DISABLED": 3, +} + +func (x Queue_State) String() string { + return proto.EnumName(Queue_State_name, int32(x)) +} +func (Queue_State) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_queue_2041daffc3c0cfe8, []int{0, 0} +} + +// A queue is a container of related tasks. Queues are configured to manage +// how those tasks are dispatched. Configurable properties include rate limits, +// retry options, target types, and others. +type Queue struct { + // Caller-specified and required in [CreateQueue][google.cloud.tasks.v2beta2.CloudTasks.CreateQueue], + // after which it becomes output only. + // + // The queue name. + // + // The queue name must have the following format: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID` + // + // * `PROJECT_ID` can contain letters ([A-Za-z]), numbers ([0-9]), + // hyphens (-), colons (:), or periods (.). + // For more information, see + // [Identifying projects](/resource-manager/docs/creating-managing-projects#identifying_projects) + // * `LOCATION_ID` is the canonical ID for the queue's location. + // The list of available locations can be obtained by calling + // [ListLocations][google.cloud.location.Locations.ListLocations]. + // For more information, see https://cloud.google.com/about/locations/. + // * `QUEUE_ID` can contain letters ([A-Za-z]), numbers ([0-9]), or + // hyphens (-). The maximum length is 100 characters. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Caller-specified and required in [CreateQueue][google.cloud.tasks.v2beta2.CloudTasks.CreateQueue][], + // after which the queue config type becomes output only, though fields within + // the config are mutable. + // + // The queue's target. + // + // The target applies to all tasks in the queue. + // + // Types that are valid to be assigned to TargetType: + // *Queue_AppEngineHttpTarget + // *Queue_PullTarget + TargetType isQueue_TargetType `protobuf_oneof:"target_type"` + // Rate limits for task dispatches. + // + // [rate_limits][google.cloud.tasks.v2beta2.Queue.rate_limits] and + // [retry_config][google.cloud.tasks.v2beta2.Queue.retry_config] are related because they both + // control task attempts however they control how tasks are + // attempted in different ways: + // + // * [rate_limits][google.cloud.tasks.v2beta2.Queue.rate_limits] controls the total rate of + // dispatches from a queue (i.e. all traffic dispatched from the + // queue, regardless of whether the dispatch is from a first + // attempt or a retry). + // * [retry_config][google.cloud.tasks.v2beta2.Queue.retry_config] controls what happens to + // particular a task after its first attempt fails. That is, + // [retry_config][google.cloud.tasks.v2beta2.Queue.retry_config] controls task retries (the + // second attempt, third attempt, etc). + RateLimits *RateLimits `protobuf:"bytes,5,opt,name=rate_limits,json=rateLimits,proto3" json:"rate_limits,omitempty"` + // Settings that determine the retry behavior. + // + // * For tasks created using Cloud Tasks: the queue-level retry settings + // apply to all tasks in the queue that were created using Cloud Tasks. + // Retry settings cannot be set on individual tasks. + // * For tasks created using the App Engine SDK: the queue-level retry + // settings apply to all tasks in the queue which do not have retry settings + // explicitly set on the task and were created by the App Engine SDK. See + // [App Engine documentation](/appengine/docs/standard/python/taskqueue/push/retrying-tasks). + RetryConfig *RetryConfig `protobuf:"bytes,6,opt,name=retry_config,json=retryConfig,proto3" json:"retry_config,omitempty"` + // Output only. The state of the queue. + // + // `state` can only be changed by called + // [PauseQueue][google.cloud.tasks.v2beta2.CloudTasks.PauseQueue], + // [ResumeQueue][google.cloud.tasks.v2beta2.CloudTasks.ResumeQueue], or uploading + // [queue.yaml/xml](/appengine/docs/python/config/queueref). + // [UpdateQueue][google.cloud.tasks.v2beta2.CloudTasks.UpdateQueue] cannot be used to change `state`. + State Queue_State `protobuf:"varint,7,opt,name=state,proto3,enum=google.cloud.tasks.v2beta2.Queue_State" json:"state,omitempty"` + // Output only. The last time this queue was purged. + // + // All tasks that were [created][google.cloud.tasks.v2beta2.Task.create_time] before this time + // were purged. + // + // A queue can be purged using [PurgeQueue][google.cloud.tasks.v2beta2.CloudTasks.PurgeQueue], the + // [App Engine Task Queue SDK, or the Cloud Console](/appengine/docs/standard/python/taskqueue/push/deleting-tasks-and-queues#purging_all_tasks_from_a_queue). + // + // Purge time will be truncated to the nearest microsecond. Purge + // time will be unset if the queue has never been purged. + PurgeTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=purge_time,json=purgeTime,proto3" json:"purge_time,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Queue) Reset() { *m = Queue{} } +func (m *Queue) String() string { return proto.CompactTextString(m) } +func (*Queue) ProtoMessage() {} +func (*Queue) Descriptor() ([]byte, []int) { + return fileDescriptor_queue_2041daffc3c0cfe8, []int{0} +} +func (m *Queue) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Queue.Unmarshal(m, b) +} +func (m *Queue) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Queue.Marshal(b, m, deterministic) +} +func (dst *Queue) XXX_Merge(src proto.Message) { + xxx_messageInfo_Queue.Merge(dst, src) +} +func (m *Queue) XXX_Size() int { + return xxx_messageInfo_Queue.Size(m) +} +func (m *Queue) XXX_DiscardUnknown() { + xxx_messageInfo_Queue.DiscardUnknown(m) +} + +var xxx_messageInfo_Queue proto.InternalMessageInfo + +type isQueue_TargetType interface { + isQueue_TargetType() +} + +type Queue_AppEngineHttpTarget struct { + AppEngineHttpTarget *AppEngineHttpTarget `protobuf:"bytes,3,opt,name=app_engine_http_target,json=appEngineHttpTarget,proto3,oneof"` +} +type Queue_PullTarget struct { + PullTarget *PullTarget `protobuf:"bytes,4,opt,name=pull_target,json=pullTarget,proto3,oneof"` +} + +func (*Queue_AppEngineHttpTarget) isQueue_TargetType() {} +func (*Queue_PullTarget) isQueue_TargetType() {} + +func (m *Queue) GetTargetType() isQueue_TargetType { + if m != nil { + return m.TargetType + } + return nil +} + +func (m *Queue) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Queue) GetAppEngineHttpTarget() *AppEngineHttpTarget { + if x, ok := m.GetTargetType().(*Queue_AppEngineHttpTarget); ok { + return x.AppEngineHttpTarget + } + return nil +} + +func (m *Queue) GetPullTarget() *PullTarget { + if x, ok := m.GetTargetType().(*Queue_PullTarget); ok { + return x.PullTarget + } + return nil +} + +func (m *Queue) GetRateLimits() *RateLimits { + if m != nil { + return m.RateLimits + } + return nil +} + +func (m *Queue) GetRetryConfig() *RetryConfig { + if m != nil { + return m.RetryConfig + } + return nil +} + +func (m *Queue) GetState() Queue_State { + if m != nil { + return m.State + } + return Queue_STATE_UNSPECIFIED +} + +func (m *Queue) GetPurgeTime() *timestamp.Timestamp { + if m != nil { + return m.PurgeTime + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Queue) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Queue_OneofMarshaler, _Queue_OneofUnmarshaler, _Queue_OneofSizer, []interface{}{ + (*Queue_AppEngineHttpTarget)(nil), + (*Queue_PullTarget)(nil), + } +} + +func _Queue_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Queue) + // target_type + switch x := m.TargetType.(type) { + case *Queue_AppEngineHttpTarget: + b.EncodeVarint(3<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.AppEngineHttpTarget); err != nil { + return err + } + case *Queue_PullTarget: + b.EncodeVarint(4<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.PullTarget); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("Queue.TargetType has unexpected type %T", x) + } + return nil +} + +func _Queue_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Queue) + switch tag { + case 3: // target_type.app_engine_http_target + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(AppEngineHttpTarget) + err := b.DecodeMessage(msg) + m.TargetType = &Queue_AppEngineHttpTarget{msg} + return true, err + case 4: // target_type.pull_target + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(PullTarget) + err := b.DecodeMessage(msg) + m.TargetType = &Queue_PullTarget{msg} + return true, err + default: + return false, nil + } +} + +func _Queue_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Queue) + // target_type + switch x := m.TargetType.(type) { + case *Queue_AppEngineHttpTarget: + s := proto.Size(x.AppEngineHttpTarget) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *Queue_PullTarget: + s := proto.Size(x.PullTarget) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Rate limits. +// +// This message determines the maximum rate that tasks can be dispatched by a +// queue, regardless of whether the dispatch is a first task attempt or a retry. +// +// Note: The debugging command, [RunTask][google.cloud.tasks.v2beta2.CloudTasks.RunTask], will run a task +// even if the queue has reached its [RateLimits][google.cloud.tasks.v2beta2.RateLimits]. +type RateLimits struct { + // The maximum rate at which tasks are dispatched from this queue. + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // * For App Engine queues, the maximum allowed value is 500. + // * This field is output only for [pull queues](google.cloud.tasks.v2beta2.PullTarget). In + // addition to the `max_tasks_dispatched_per_second` limit, a + // maximum of 10 QPS of [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] + // requests are allowed per pull queue. + // + // + // This field has the same meaning as + // [rate in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#rate). + MaxTasksDispatchedPerSecond float64 `protobuf:"fixed64,1,opt,name=max_tasks_dispatched_per_second,json=maxTasksDispatchedPerSecond,proto3" json:"max_tasks_dispatched_per_second,omitempty"` + // Output only. The max burst size. + // + // Max burst size limits how fast tasks in queue are processed when + // many tasks are in the queue and the rate is high. This field + // allows the queue to have a high rate so processing starts shortly + // after a task is enqueued, but still limits resource usage when + // many tasks are enqueued in a short period of time. + // + // The [token bucket](https://wikipedia.org/wiki/Token_Bucket) + // algorithm is used to control the rate of task dispatches. Each + // queue has a token bucket that holds tokens, up to the maximum + // specified by `max_burst_size`. Each time a task is dispatched, a + // token is removed from the bucket. Tasks will be dispatched until + // the queue's bucket runs out of tokens. The bucket will be + // continuously refilled with new tokens based on + // [max_tasks_dispatched_per_second][google.cloud.tasks.v2beta2.RateLimits.max_tasks_dispatched_per_second]. + // + // Cloud Tasks will pick the value of `max_burst_size` based on the + // value of + // [max_tasks_dispatched_per_second][google.cloud.tasks.v2beta2.RateLimits.max_tasks_dispatched_per_second]. + // + // For App Engine queues that were created or updated using + // `queue.yaml/xml`, `max_burst_size` is equal to + // [bucket_size](/appengine/docs/standard/python/config/queueref#bucket_size). + // Since `max_burst_size` is output only, if + // [UpdateQueue][google.cloud.tasks.v2beta2.CloudTasks.UpdateQueue] is called on a queue + // created by `queue.yaml/xml`, `max_burst_size` will be reset based + // on the value of + // [max_tasks_dispatched_per_second][google.cloud.tasks.v2beta2.RateLimits.max_tasks_dispatched_per_second], + // regardless of whether + // [max_tasks_dispatched_per_second][google.cloud.tasks.v2beta2.RateLimits.max_tasks_dispatched_per_second] + // is updated. + // + MaxBurstSize int32 `protobuf:"varint,2,opt,name=max_burst_size,json=maxBurstSize,proto3" json:"max_burst_size,omitempty"` + // The maximum number of concurrent tasks that Cloud Tasks allows + // to be dispatched for this queue. After this threshold has been + // reached, Cloud Tasks stops dispatching tasks until the number of + // concurrent requests decreases. + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // + // The maximum allowed value is 5,000. + // + // This field is output only for + // [pull queues](google.cloud.tasks.v2beta2.PullTarget) and always -1, which + // indicates no limit. No other queue types can have `max_concurrent_tasks` + // set to -1. + // + // + // This field has the same meaning as + // [max_concurrent_requests in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#max_concurrent_requests). + MaxConcurrentTasks int32 `protobuf:"varint,3,opt,name=max_concurrent_tasks,json=maxConcurrentTasks,proto3" json:"max_concurrent_tasks,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RateLimits) Reset() { *m = RateLimits{} } +func (m *RateLimits) String() string { return proto.CompactTextString(m) } +func (*RateLimits) ProtoMessage() {} +func (*RateLimits) Descriptor() ([]byte, []int) { + return fileDescriptor_queue_2041daffc3c0cfe8, []int{1} +} +func (m *RateLimits) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RateLimits.Unmarshal(m, b) +} +func (m *RateLimits) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RateLimits.Marshal(b, m, deterministic) +} +func (dst *RateLimits) XXX_Merge(src proto.Message) { + xxx_messageInfo_RateLimits.Merge(dst, src) +} +func (m *RateLimits) XXX_Size() int { + return xxx_messageInfo_RateLimits.Size(m) +} +func (m *RateLimits) XXX_DiscardUnknown() { + xxx_messageInfo_RateLimits.DiscardUnknown(m) +} + +var xxx_messageInfo_RateLimits proto.InternalMessageInfo + +func (m *RateLimits) GetMaxTasksDispatchedPerSecond() float64 { + if m != nil { + return m.MaxTasksDispatchedPerSecond + } + return 0 +} + +func (m *RateLimits) GetMaxBurstSize() int32 { + if m != nil { + return m.MaxBurstSize + } + return 0 +} + +func (m *RateLimits) GetMaxConcurrentTasks() int32 { + if m != nil { + return m.MaxConcurrentTasks + } + return 0 +} + +// Retry config. +// +// These settings determine how a failed task attempt is retried. +type RetryConfig struct { + // Number of attempts per task. + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // + // + // This field has the same meaning as + // [task_retry_limit in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#retry_parameters). + // + // Types that are valid to be assigned to NumAttempts: + // *RetryConfig_MaxAttempts + // *RetryConfig_UnlimitedAttempts + NumAttempts isRetryConfig_NumAttempts `protobuf_oneof:"num_attempts"` + // If positive, `max_retry_duration` specifies the time limit for + // retrying a failed task, measured from when the task was first + // attempted. Once `max_retry_duration` time has passed *and* the + // task has been attempted [max_attempts][google.cloud.tasks.v2beta2.RetryConfig.max_attempts] + // times, no further attempts will be made and the task will be + // deleted. + // + // If zero, then the task age is unlimited. + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // This field is output only for + // [pull queues](google.cloud.tasks.v2beta2.PullTarget). + // + // + // `max_retry_duration` will be truncated to the nearest second. + // + // This field has the same meaning as + // [task_age_limit in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#retry_parameters). + MaxRetryDuration *duration.Duration `protobuf:"bytes,3,opt,name=max_retry_duration,json=maxRetryDuration,proto3" json:"max_retry_duration,omitempty"` + // A task will be [scheduled](Task.schedule_time) for retry between + // [min_backoff][google.cloud.tasks.v2beta2.RetryConfig.min_backoff] and + // [max_backoff][google.cloud.tasks.v2beta2.RetryConfig.max_backoff] duration after it fails, + // if the queue's [RetryConfig][google.cloud.tasks.v2beta2.RetryConfig] specifies that the task should be + // retried. + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // This field is output only for + // [pull queues](google.cloud.tasks.v2beta2.PullTarget). + // + // + // `min_backoff` will be truncated to the nearest second. + // + // This field has the same meaning as + // [min_backoff_seconds in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#retry_parameters). + MinBackoff *duration.Duration `protobuf:"bytes,4,opt,name=min_backoff,json=minBackoff,proto3" json:"min_backoff,omitempty"` + // A task will be [scheduled](Task.schedule_time) for retry between + // [min_backoff][google.cloud.tasks.v2beta2.RetryConfig.min_backoff] and + // [max_backoff][google.cloud.tasks.v2beta2.RetryConfig.max_backoff] duration after it fails, + // if the queue's [RetryConfig][google.cloud.tasks.v2beta2.RetryConfig] specifies that the task should be + // retried. + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // This field is output only for + // [pull queues](google.cloud.tasks.v2beta2.PullTarget). + // + // + // `max_backoff` will be truncated to the nearest second. + // + // This field has the same meaning as + // [max_backoff_seconds in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#retry_parameters). + MaxBackoff *duration.Duration `protobuf:"bytes,5,opt,name=max_backoff,json=maxBackoff,proto3" json:"max_backoff,omitempty"` + // The time between retries will double `max_doublings` times. + // + // A task's retry interval starts at + // [min_backoff][google.cloud.tasks.v2beta2.RetryConfig.min_backoff], then doubles + // `max_doublings` times, then increases linearly, and finally + // retries retries at intervals of + // [max_backoff][google.cloud.tasks.v2beta2.RetryConfig.max_backoff] up to + // [max_attempts][google.cloud.tasks.v2beta2.RetryConfig.max_attempts] times. + // + // For example, if [min_backoff][google.cloud.tasks.v2beta2.RetryConfig.min_backoff] is 10s, + // [max_backoff][google.cloud.tasks.v2beta2.RetryConfig.max_backoff] is 300s, and + // `max_doublings` is 3, then the a task will first be retried in + // 10s. The retry interval will double three times, and then + // increase linearly by 2^3 * 10s. Finally, the task will retry at + // intervals of [max_backoff][google.cloud.tasks.v2beta2.RetryConfig.max_backoff] until the + // task has been attempted [max_attempts][google.cloud.tasks.v2beta2.RetryConfig.max_attempts] + // times. Thus, the requests will retry at 10s, 20s, 40s, 80s, 160s, + // 240s, 300s, 300s, .... + // + // If unspecified when the queue is created, Cloud Tasks will pick the + // default. + // + // This field is output only for + // [pull queues](google.cloud.tasks.v2beta2.PullTarget). + // + // + // This field has the same meaning as + // [max_doublings in queue.yaml/xml](/appengine/docs/standard/python/config/queueref#retry_parameters). + MaxDoublings int32 `protobuf:"varint,6,opt,name=max_doublings,json=maxDoublings,proto3" json:"max_doublings,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RetryConfig) Reset() { *m = RetryConfig{} } +func (m *RetryConfig) String() string { return proto.CompactTextString(m) } +func (*RetryConfig) ProtoMessage() {} +func (*RetryConfig) Descriptor() ([]byte, []int) { + return fileDescriptor_queue_2041daffc3c0cfe8, []int{2} +} +func (m *RetryConfig) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RetryConfig.Unmarshal(m, b) +} +func (m *RetryConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RetryConfig.Marshal(b, m, deterministic) +} +func (dst *RetryConfig) XXX_Merge(src proto.Message) { + xxx_messageInfo_RetryConfig.Merge(dst, src) +} +func (m *RetryConfig) XXX_Size() int { + return xxx_messageInfo_RetryConfig.Size(m) +} +func (m *RetryConfig) XXX_DiscardUnknown() { + xxx_messageInfo_RetryConfig.DiscardUnknown(m) +} + +var xxx_messageInfo_RetryConfig proto.InternalMessageInfo + +type isRetryConfig_NumAttempts interface { + isRetryConfig_NumAttempts() +} + +type RetryConfig_MaxAttempts struct { + MaxAttempts int32 `protobuf:"varint,1,opt,name=max_attempts,json=maxAttempts,proto3,oneof"` +} +type RetryConfig_UnlimitedAttempts struct { + UnlimitedAttempts bool `protobuf:"varint,2,opt,name=unlimited_attempts,json=unlimitedAttempts,proto3,oneof"` +} + +func (*RetryConfig_MaxAttempts) isRetryConfig_NumAttempts() {} +func (*RetryConfig_UnlimitedAttempts) isRetryConfig_NumAttempts() {} + +func (m *RetryConfig) GetNumAttempts() isRetryConfig_NumAttempts { + if m != nil { + return m.NumAttempts + } + return nil +} + +func (m *RetryConfig) GetMaxAttempts() int32 { + if x, ok := m.GetNumAttempts().(*RetryConfig_MaxAttempts); ok { + return x.MaxAttempts + } + return 0 +} + +func (m *RetryConfig) GetUnlimitedAttempts() bool { + if x, ok := m.GetNumAttempts().(*RetryConfig_UnlimitedAttempts); ok { + return x.UnlimitedAttempts + } + return false +} + +func (m *RetryConfig) GetMaxRetryDuration() *duration.Duration { + if m != nil { + return m.MaxRetryDuration + } + return nil +} + +func (m *RetryConfig) GetMinBackoff() *duration.Duration { + if m != nil { + return m.MinBackoff + } + return nil +} + +func (m *RetryConfig) GetMaxBackoff() *duration.Duration { + if m != nil { + return m.MaxBackoff + } + return nil +} + +func (m *RetryConfig) GetMaxDoublings() int32 { + if m != nil { + return m.MaxDoublings + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*RetryConfig) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _RetryConfig_OneofMarshaler, _RetryConfig_OneofUnmarshaler, _RetryConfig_OneofSizer, []interface{}{ + (*RetryConfig_MaxAttempts)(nil), + (*RetryConfig_UnlimitedAttempts)(nil), + } +} + +func _RetryConfig_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*RetryConfig) + // num_attempts + switch x := m.NumAttempts.(type) { + case *RetryConfig_MaxAttempts: + b.EncodeVarint(1<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.MaxAttempts)) + case *RetryConfig_UnlimitedAttempts: + t := uint64(0) + if x.UnlimitedAttempts { + t = 1 + } + b.EncodeVarint(2<<3 | proto.WireVarint) + b.EncodeVarint(t) + case nil: + default: + return fmt.Errorf("RetryConfig.NumAttempts has unexpected type %T", x) + } + return nil +} + +func _RetryConfig_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*RetryConfig) + switch tag { + case 1: // num_attempts.max_attempts + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.NumAttempts = &RetryConfig_MaxAttempts{int32(x)} + return true, err + case 2: // num_attempts.unlimited_attempts + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.NumAttempts = &RetryConfig_UnlimitedAttempts{x != 0} + return true, err + default: + return false, nil + } +} + +func _RetryConfig_OneofSizer(msg proto.Message) (n int) { + m := msg.(*RetryConfig) + // num_attempts + switch x := m.NumAttempts.(type) { + case *RetryConfig_MaxAttempts: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.MaxAttempts)) + case *RetryConfig_UnlimitedAttempts: + n += 1 // tag and wire + n += 1 + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +func init() { + proto.RegisterType((*Queue)(nil), "google.cloud.tasks.v2beta2.Queue") + proto.RegisterType((*RateLimits)(nil), "google.cloud.tasks.v2beta2.RateLimits") + proto.RegisterType((*RetryConfig)(nil), "google.cloud.tasks.v2beta2.RetryConfig") + proto.RegisterEnum("google.cloud.tasks.v2beta2.Queue_State", Queue_State_name, Queue_State_value) +} + +func init() { + proto.RegisterFile("google/cloud/tasks/v2beta2/queue.proto", fileDescriptor_queue_2041daffc3c0cfe8) +} + +var fileDescriptor_queue_2041daffc3c0cfe8 = []byte{ + // 702 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x94, 0xcb, 0x6e, 0xdb, 0x3a, + 0x10, 0x86, 0xad, 0x24, 0xca, 0x65, 0xe4, 0x04, 0x0e, 0xcf, 0x05, 0x3e, 0x3e, 0x45, 0x62, 0x38, + 0x45, 0xe2, 0x95, 0x54, 0xa4, 0xab, 0xb6, 0x28, 0x0a, 0x3b, 0x72, 0x63, 0x17, 0x81, 0xe1, 0xca, + 0xce, 0xa6, 0x1b, 0x82, 0x96, 0x69, 0x45, 0x88, 0x44, 0xb2, 0x12, 0x55, 0x38, 0x79, 0xa4, 0x6e, + 0xfb, 0x34, 0x7d, 0x9b, 0x82, 0xd4, 0x25, 0x45, 0xd3, 0x3a, 0x3b, 0x73, 0xe6, 0xfb, 0xff, 0x19, + 0x71, 0x86, 0x86, 0xd3, 0x80, 0xf3, 0x20, 0xa2, 0x8e, 0x1f, 0xf1, 0x6c, 0xe1, 0x48, 0x92, 0xde, + 0xa6, 0xce, 0x97, 0xf3, 0x39, 0x95, 0xe4, 0xdc, 0xf9, 0x9c, 0xd1, 0x8c, 0xda, 0x22, 0xe1, 0x92, + 0xa3, 0x56, 0xce, 0xd9, 0x9a, 0xb3, 0x35, 0x67, 0x17, 0x5c, 0xeb, 0x59, 0xe1, 0x41, 0x44, 0xe8, + 0x10, 0xc6, 0xb8, 0x24, 0x32, 0xe4, 0x2c, 0xcd, 0x95, 0xad, 0xb3, 0x35, 0x15, 0x24, 0x49, 0x02, + 0x2a, 0x0b, 0xf0, 0xa8, 0x00, 0xf5, 0x69, 0x9e, 0x2d, 0x9d, 0x45, 0x96, 0x68, 0xa7, 0x22, 0x7f, + 0xfc, 0x6b, 0x5e, 0x86, 0x31, 0x4d, 0x25, 0x89, 0x45, 0x0e, 0x74, 0xbe, 0x6d, 0x81, 0xf9, 0x51, + 0xf5, 0x8c, 0x10, 0x6c, 0x31, 0x12, 0xd3, 0xa6, 0xd1, 0x36, 0xba, 0x7b, 0x9e, 0xfe, 0x8d, 0x96, + 0xf0, 0x2f, 0x11, 0x02, 0x53, 0x16, 0x84, 0x8c, 0xe2, 0x1b, 0x29, 0x05, 0xce, 0xcb, 0x37, 0x37, + 0xdb, 0x46, 0xd7, 0x3a, 0x77, 0xec, 0x3f, 0x7f, 0xa2, 0xdd, 0x13, 0x62, 0xa0, 0x85, 0x43, 0x29, + 0xc5, 0x4c, 0xcb, 0x86, 0x35, 0xef, 0x2f, 0xf2, 0x38, 0x8c, 0x46, 0x60, 0x89, 0x2c, 0x8a, 0x4a, + 0xf3, 0x2d, 0x6d, 0x7e, 0xba, 0xce, 0x7c, 0x92, 0x45, 0x51, 0xe5, 0x09, 0xa2, 0x3a, 0xa1, 0x4b, + 0xb0, 0x12, 0x22, 0x29, 0x8e, 0xc2, 0x38, 0x94, 0x69, 0xd3, 0x7c, 0xda, 0xca, 0x23, 0x92, 0x5e, + 0x69, 0xda, 0x83, 0xa4, 0xfa, 0x8d, 0x3e, 0x40, 0x3d, 0xa1, 0x32, 0xb9, 0xc3, 0x3e, 0x67, 0xcb, + 0x30, 0x68, 0x6e, 0x6b, 0xa7, 0xb3, 0xb5, 0x4e, 0x8a, 0xbf, 0xd0, 0xb8, 0x67, 0x25, 0x0f, 0x07, + 0xf4, 0x16, 0xcc, 0x54, 0x12, 0x49, 0x9b, 0x3b, 0x6d, 0xa3, 0x7b, 0xb0, 0xde, 0x44, 0x4f, 0xc3, + 0x9e, 0x2a, 0xdc, 0xcb, 0x55, 0xe8, 0x15, 0x80, 0xc8, 0x92, 0x80, 0x62, 0x35, 0xbd, 0xe6, 0xae, + 0x6e, 0xa4, 0x55, 0x7a, 0x94, 0xa3, 0xb5, 0x67, 0xe5, 0x68, 0xbd, 0x3d, 0x4d, 0xab, 0x73, 0x67, + 0x00, 0xa6, 0xb6, 0x42, 0xff, 0xc0, 0xe1, 0x74, 0xd6, 0x9b, 0x0d, 0xf0, 0xf5, 0x78, 0x3a, 0x19, + 0x5c, 0x8c, 0xde, 0x8f, 0x06, 0x6e, 0xa3, 0x86, 0x2c, 0xd8, 0xf1, 0xae, 0xc7, 0xe3, 0xd1, 0xf8, + 0xb2, 0x61, 0x20, 0x80, 0xed, 0x49, 0xef, 0x7a, 0x3a, 0x70, 0x1b, 0x1b, 0xa8, 0x0e, 0xbb, 0xee, + 0x68, 0xda, 0xeb, 0x5f, 0x0d, 0xdc, 0xc6, 0x66, 0x7f, 0x1f, 0xac, 0x7c, 0x36, 0x58, 0xde, 0x09, + 0xda, 0xf9, 0x6a, 0x00, 0x3c, 0x5c, 0x1b, 0x72, 0xe1, 0x38, 0x26, 0x2b, 0xac, 0xbf, 0x03, 0x2f, + 0xc2, 0x54, 0x10, 0xe9, 0xdf, 0xd0, 0x05, 0x16, 0x34, 0xc1, 0x29, 0xf5, 0x39, 0x5b, 0xe8, 0xad, + 0x32, 0xbc, 0xff, 0x63, 0xb2, 0x9a, 0x29, 0xca, 0xad, 0xa0, 0x09, 0x4d, 0xa6, 0x1a, 0x41, 0xcf, + 0xe1, 0x40, 0xb9, 0xcc, 0xb3, 0x24, 0x95, 0x38, 0x0d, 0xef, 0x69, 0x73, 0xa3, 0x6d, 0x74, 0x4d, + 0xaf, 0x1e, 0x93, 0x55, 0x5f, 0x05, 0xa7, 0xe1, 0x3d, 0x45, 0x2f, 0xe0, 0x6f, 0x45, 0xf9, 0x9c, + 0xf9, 0x59, 0x92, 0x50, 0x26, 0xf3, 0xb2, 0x7a, 0x21, 0x4d, 0x0f, 0xc5, 0x64, 0x75, 0x51, 0xa5, + 0x74, 0xa9, 0xce, 0xf7, 0x0d, 0xb0, 0x7e, 0x9a, 0x0c, 0x3a, 0x01, 0xe5, 0x88, 0x89, 0x94, 0x34, + 0x16, 0x32, 0xd5, 0xad, 0x99, 0xc3, 0x9a, 0x67, 0xc5, 0x64, 0xd5, 0x2b, 0x82, 0xc8, 0x01, 0x94, + 0x31, 0xbd, 0x43, 0x74, 0xf1, 0x80, 0xaa, 0x86, 0x76, 0x87, 0x35, 0xef, 0xb0, 0xca, 0x55, 0x82, + 0x4b, 0x50, 0xb5, 0x71, 0xbe, 0x32, 0xe5, 0x2b, 0x2c, 0x9e, 0xc9, 0x7f, 0x8f, 0x66, 0xe5, 0x16, + 0x80, 0xd7, 0x88, 0xc9, 0x4a, 0x37, 0x57, 0x46, 0xd0, 0x6b, 0xb0, 0xe2, 0x90, 0xe1, 0x39, 0xf1, + 0x6f, 0xf9, 0x72, 0x59, 0xbc, 0x85, 0x35, 0x0e, 0x10, 0x87, 0xac, 0x9f, 0xc3, 0x5a, 0xab, 0xae, + 0xb0, 0xd0, 0x9a, 0x4f, 0x6b, 0xc9, 0xaa, 0xd4, 0x9e, 0xc0, 0xbe, 0xd2, 0x2e, 0x78, 0x36, 0x8f, + 0x42, 0x16, 0xa4, 0x7a, 0xe1, 0xf3, 0xdb, 0x77, 0xcb, 0x58, 0xff, 0x00, 0xea, 0x2c, 0x8b, 0xab, + 0x0b, 0xe9, 0x73, 0x38, 0xf2, 0x79, 0xbc, 0x66, 0x9d, 0xfb, 0xa0, 0xf7, 0x79, 0xa2, 0x4a, 0x4f, + 0x8c, 0x4f, 0xef, 0x0a, 0x32, 0xe0, 0x11, 0x61, 0x81, 0xcd, 0x93, 0xc0, 0x09, 0x28, 0xd3, 0x8d, + 0x39, 0x79, 0x8a, 0x88, 0x30, 0xfd, 0xdd, 0xff, 0xde, 0x1b, 0x7d, 0x9a, 0x6f, 0x6b, 0xf6, 0xe5, + 0x8f, 0x00, 0x00, 0x00, 0xff, 0xff, 0xd6, 0xd1, 0xaa, 0xde, 0x84, 0x05, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/target.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/target.pb.go new file mode 100644 index 000000000..0505ce9df --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/target.pb.go @@ -0,0 +1,626 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/tasks/v2beta2/target.proto + +package tasks // import "google.golang.org/genproto/googleapis/cloud/tasks/v2beta2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _ "github.com/golang/protobuf/ptypes/duration" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// The HTTP method used to execute the task. +type HttpMethod int32 + +const ( + // HTTP method unspecified + HttpMethod_HTTP_METHOD_UNSPECIFIED HttpMethod = 0 + // HTTP Post + HttpMethod_POST HttpMethod = 1 + // HTTP Get + HttpMethod_GET HttpMethod = 2 + // HTTP Head + HttpMethod_HEAD HttpMethod = 3 + // HTTP Put + HttpMethod_PUT HttpMethod = 4 + // HTTP Delete + HttpMethod_DELETE HttpMethod = 5 +) + +var HttpMethod_name = map[int32]string{ + 0: "HTTP_METHOD_UNSPECIFIED", + 1: "POST", + 2: "GET", + 3: "HEAD", + 4: "PUT", + 5: "DELETE", +} +var HttpMethod_value = map[string]int32{ + "HTTP_METHOD_UNSPECIFIED": 0, + "POST": 1, + "GET": 2, + "HEAD": 3, + "PUT": 4, + "DELETE": 5, +} + +func (x HttpMethod) String() string { + return proto.EnumName(HttpMethod_name, int32(x)) +} +func (HttpMethod) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_target_b871819819dd67b9, []int{0} +} + +// Pull target. +type PullTarget struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *PullTarget) Reset() { *m = PullTarget{} } +func (m *PullTarget) String() string { return proto.CompactTextString(m) } +func (*PullTarget) ProtoMessage() {} +func (*PullTarget) Descriptor() ([]byte, []int) { + return fileDescriptor_target_b871819819dd67b9, []int{0} +} +func (m *PullTarget) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_PullTarget.Unmarshal(m, b) +} +func (m *PullTarget) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_PullTarget.Marshal(b, m, deterministic) +} +func (dst *PullTarget) XXX_Merge(src proto.Message) { + xxx_messageInfo_PullTarget.Merge(dst, src) +} +func (m *PullTarget) XXX_Size() int { + return xxx_messageInfo_PullTarget.Size(m) +} +func (m *PullTarget) XXX_DiscardUnknown() { + xxx_messageInfo_PullTarget.DiscardUnknown(m) +} + +var xxx_messageInfo_PullTarget proto.InternalMessageInfo + +// The pull message contains data that can be used by the caller of +// [LeaseTasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] to process the task. +// +// This proto can only be used for tasks in a queue which has +// [pull_target][google.cloud.tasks.v2beta2.Queue.pull_target] set. +type PullMessage struct { + // A data payload consumed by the worker to execute the task. + Payload []byte `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` + // The task's tag. + // + // Tags allow similar tasks to be processed in a batch. If you label + // tasks with a tag, your worker can + // [lease tasks][google.cloud.tasks.v2beta2.CloudTasks.LeaseTasks] with the same tag using + // [filter][google.cloud.tasks.v2beta2.LeaseTasksRequest.filter]. For example, if you want to + // aggregate the events associated with a specific user once a day, + // you could tag tasks with the user ID. + // + // The task's tag can only be set when the + // [task is created][google.cloud.tasks.v2beta2.CloudTasks.CreateTask]. + // + // The tag must be less than 500 characters. + // + // SDK compatibility: Although the SDK allows tags to be either + // string or [bytes](/appengine/docs/standard/java/javadoc/com/google/appengine/api/taskqueue/TaskOptions.html#tag-byte:A-), + // only UTF-8 encoded tags can be used in Cloud Tasks. If a tag isn't UTF-8 + // encoded, the tag will be empty when the task is returned by Cloud Tasks. + Tag string `protobuf:"bytes,2,opt,name=tag,proto3" json:"tag,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *PullMessage) Reset() { *m = PullMessage{} } +func (m *PullMessage) String() string { return proto.CompactTextString(m) } +func (*PullMessage) ProtoMessage() {} +func (*PullMessage) Descriptor() ([]byte, []int) { + return fileDescriptor_target_b871819819dd67b9, []int{1} +} +func (m *PullMessage) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_PullMessage.Unmarshal(m, b) +} +func (m *PullMessage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_PullMessage.Marshal(b, m, deterministic) +} +func (dst *PullMessage) XXX_Merge(src proto.Message) { + xxx_messageInfo_PullMessage.Merge(dst, src) +} +func (m *PullMessage) XXX_Size() int { + return xxx_messageInfo_PullMessage.Size(m) +} +func (m *PullMessage) XXX_DiscardUnknown() { + xxx_messageInfo_PullMessage.DiscardUnknown(m) +} + +var xxx_messageInfo_PullMessage proto.InternalMessageInfo + +func (m *PullMessage) GetPayload() []byte { + if m != nil { + return m.Payload + } + return nil +} + +func (m *PullMessage) GetTag() string { + if m != nil { + return m.Tag + } + return "" +} + +// App Engine HTTP target. +// +// The task will be delivered to the App Engine application hostname +// specified by its [AppEngineHttpTarget][google.cloud.tasks.v2beta2.AppEngineHttpTarget] and [AppEngineHttpRequest][google.cloud.tasks.v2beta2.AppEngineHttpRequest]. +// The documentation for [AppEngineHttpRequest][google.cloud.tasks.v2beta2.AppEngineHttpRequest] explains how the +// task's host URL is constructed. +// +// Using [AppEngineHttpTarget][google.cloud.tasks.v2beta2.AppEngineHttpTarget] requires +// [`appengine.applications.get`](/appengine/docs/admin-api/access-control) +// Google IAM permission for the project +// and the following scope: +// +// `https://www.googleapis.com/auth/cloud-platform` +type AppEngineHttpTarget struct { + // Overrides for the + // [task-level app_engine_routing][google.cloud.tasks.v2beta2.AppEngineHttpRequest.app_engine_routing]. + // + // If set, `app_engine_routing_override` is used for all tasks in + // the queue, no matter what the setting is for the + // [task-level app_engine_routing][google.cloud.tasks.v2beta2.AppEngineHttpRequest.app_engine_routing]. + AppEngineRoutingOverride *AppEngineRouting `protobuf:"bytes,1,opt,name=app_engine_routing_override,json=appEngineRoutingOverride,proto3" json:"app_engine_routing_override,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AppEngineHttpTarget) Reset() { *m = AppEngineHttpTarget{} } +func (m *AppEngineHttpTarget) String() string { return proto.CompactTextString(m) } +func (*AppEngineHttpTarget) ProtoMessage() {} +func (*AppEngineHttpTarget) Descriptor() ([]byte, []int) { + return fileDescriptor_target_b871819819dd67b9, []int{2} +} +func (m *AppEngineHttpTarget) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AppEngineHttpTarget.Unmarshal(m, b) +} +func (m *AppEngineHttpTarget) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AppEngineHttpTarget.Marshal(b, m, deterministic) +} +func (dst *AppEngineHttpTarget) XXX_Merge(src proto.Message) { + xxx_messageInfo_AppEngineHttpTarget.Merge(dst, src) +} +func (m *AppEngineHttpTarget) XXX_Size() int { + return xxx_messageInfo_AppEngineHttpTarget.Size(m) +} +func (m *AppEngineHttpTarget) XXX_DiscardUnknown() { + xxx_messageInfo_AppEngineHttpTarget.DiscardUnknown(m) +} + +var xxx_messageInfo_AppEngineHttpTarget proto.InternalMessageInfo + +func (m *AppEngineHttpTarget) GetAppEngineRoutingOverride() *AppEngineRouting { + if m != nil { + return m.AppEngineRoutingOverride + } + return nil +} + +// App Engine HTTP request. +// +// The message defines the HTTP request that is sent to an App Engine app when +// the task is dispatched. +// +// This proto can only be used for tasks in a queue which has +// [app_engine_http_target][google.cloud.tasks.v2beta2.Queue.app_engine_http_target] set. +// +// Using [AppEngineHttpRequest][google.cloud.tasks.v2beta2.AppEngineHttpRequest] requires +// [`appengine.applications.get`](/appengine/docs/admin-api/access-control) +// Google IAM permission for the project +// and the following scope: +// +// `https://www.googleapis.com/auth/cloud-platform` +// +// The task will be delivered to the App Engine app which belongs to the same +// project as the queue. For more information, see +// [How Requests are Routed](/appengine/docs/standard/python/how-requests-are-routed) +// and how routing is affected by +// [dispatch files](/appengine/docs/python/config/dispatchref). +// +// The [AppEngineRouting][google.cloud.tasks.v2beta2.AppEngineRouting] used to construct the URL that the task is +// delivered to can be set at the queue-level or task-level: +// +// * If set, +// [app_engine_routing_override][google.cloud.tasks.v2beta2.AppEngineHttpTarget.app_engine_routing_override] +// is used for all tasks in the queue, no matter what the setting +// is for the +// [task-level app_engine_routing][google.cloud.tasks.v2beta2.AppEngineHttpRequest.app_engine_routing]. +// +// +// The `url` that the task will be sent to is: +// +// * `url =` [host][google.cloud.tasks.v2beta2.AppEngineRouting.host] `+` +// [relative_url][google.cloud.tasks.v2beta2.AppEngineHttpRequest.relative_url] +// +// The task attempt has succeeded if the app's request handler returns +// an HTTP response code in the range [`200` - `299`]. `503` is +// considered an App Engine system error instead of an application +// error. Requests returning error `503` will be retried regardless of +// retry configuration and not counted against retry counts. +// Any other response code or a failure to receive a response before the +// deadline is a failed attempt. +type AppEngineHttpRequest struct { + // The HTTP method to use for the request. The default is POST. + // + // The app's request handler for the task's target URL must be able to handle + // HTTP requests with this http_method, otherwise the task attempt will fail + // with error code 405 (Method Not Allowed). See + // [Writing a push task request handler](/appengine/docs/java/taskqueue/push/creating-handlers#writing_a_push_task_request_handler) + // and the documentation for the request handlers in the language your app is + // written in e.g. + // [Python Request Handler](/appengine/docs/python/tools/webapp/requesthandlerclass). + HttpMethod HttpMethod `protobuf:"varint,1,opt,name=http_method,json=httpMethod,proto3,enum=google.cloud.tasks.v2beta2.HttpMethod" json:"http_method,omitempty"` + // Task-level setting for App Engine routing. + // + // If set, + // [app_engine_routing_override][google.cloud.tasks.v2beta2.AppEngineHttpTarget.app_engine_routing_override] + // is used for all tasks in the queue, no matter what the setting is for the + // [task-level app_engine_routing][google.cloud.tasks.v2beta2.AppEngineHttpRequest.app_engine_routing]. + AppEngineRouting *AppEngineRouting `protobuf:"bytes,2,opt,name=app_engine_routing,json=appEngineRouting,proto3" json:"app_engine_routing,omitempty"` + // The relative URL. + // + // The relative URL must begin with "/" and must be a valid HTTP relative URL. + // It can contain a path and query string arguments. + // If the relative URL is empty, then the root path "/" will be used. + // No spaces are allowed, and the maximum length allowed is 2083 characters. + RelativeUrl string `protobuf:"bytes,3,opt,name=relative_url,json=relativeUrl,proto3" json:"relative_url,omitempty"` + // HTTP request headers. + // + // This map contains the header field names and values. + // Headers can be set when the + // [task is created](google.cloud.tasks.v2beta2.CloudTasks.CreateTask). + // Repeated headers are not supported but a header value can contain commas. + // + // Cloud Tasks sets some headers to default values: + // + // * `User-Agent`: By default, this header is + // `"AppEngine-Google; (+http://code.google.com/appengine)"`. + // This header can be modified, but Cloud Tasks will append + // `"AppEngine-Google; (+http://code.google.com/appengine)"` to the + // modified `User-Agent`. + // + // If the task has a [payload][google.cloud.tasks.v2beta2.AppEngineHttpRequest.payload], Cloud + // Tasks sets the following headers: + // + // * `Content-Type`: By default, the `Content-Type` header is set to + // `"application/octet-stream"`. The default can be overridden by explicitly + // setting `Content-Type` to a particular media type when the + // [task is created](google.cloud.tasks.v2beta2.CloudTasks.CreateTask). + // For example, `Content-Type` can be set to `"application/json"`. + // * `Content-Length`: This is computed by Cloud Tasks. This value is + // output only. It cannot be changed. + // + // The headers below cannot be set or overridden: + // + // * `Host` + // * `X-Google-*` + // * `X-AppEngine-*` + // + // In addition, Cloud Tasks sets some headers when the task is dispatched, + // such as headers containing information about the task; see + // [request headers](/appengine/docs/python/taskqueue/push/creating-handlers#reading_request_headers). + // These headers are set only when the task is dispatched, so they are not + // visible when the task is returned in a Cloud Tasks response. + // + // Although there is no specific limit for the maximum number of headers or + // the size, there is a limit on the maximum size of the [Task][google.cloud.tasks.v2beta2.Task]. For more + // information, see the [CreateTask][google.cloud.tasks.v2beta2.CloudTasks.CreateTask] documentation. + Headers map[string]string `protobuf:"bytes,4,rep,name=headers,proto3" json:"headers,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + // Payload. + // + // The payload will be sent as the HTTP message body. A message + // body, and thus a payload, is allowed only if the HTTP method is + // POST or PUT. It is an error to set a data payload on a task with + // an incompatible [HttpMethod][google.cloud.tasks.v2beta2.HttpMethod]. + Payload []byte `protobuf:"bytes,5,opt,name=payload,proto3" json:"payload,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AppEngineHttpRequest) Reset() { *m = AppEngineHttpRequest{} } +func (m *AppEngineHttpRequest) String() string { return proto.CompactTextString(m) } +func (*AppEngineHttpRequest) ProtoMessage() {} +func (*AppEngineHttpRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_target_b871819819dd67b9, []int{3} +} +func (m *AppEngineHttpRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AppEngineHttpRequest.Unmarshal(m, b) +} +func (m *AppEngineHttpRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AppEngineHttpRequest.Marshal(b, m, deterministic) +} +func (dst *AppEngineHttpRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_AppEngineHttpRequest.Merge(dst, src) +} +func (m *AppEngineHttpRequest) XXX_Size() int { + return xxx_messageInfo_AppEngineHttpRequest.Size(m) +} +func (m *AppEngineHttpRequest) XXX_DiscardUnknown() { + xxx_messageInfo_AppEngineHttpRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_AppEngineHttpRequest proto.InternalMessageInfo + +func (m *AppEngineHttpRequest) GetHttpMethod() HttpMethod { + if m != nil { + return m.HttpMethod + } + return HttpMethod_HTTP_METHOD_UNSPECIFIED +} + +func (m *AppEngineHttpRequest) GetAppEngineRouting() *AppEngineRouting { + if m != nil { + return m.AppEngineRouting + } + return nil +} + +func (m *AppEngineHttpRequest) GetRelativeUrl() string { + if m != nil { + return m.RelativeUrl + } + return "" +} + +func (m *AppEngineHttpRequest) GetHeaders() map[string]string { + if m != nil { + return m.Headers + } + return nil +} + +func (m *AppEngineHttpRequest) GetPayload() []byte { + if m != nil { + return m.Payload + } + return nil +} + +// App Engine Routing. +// +// For more information about services, versions, and instances see +// [An Overview of App Engine](/appengine/docs/python/an-overview-of-app-engine), +// [Microservices Architecture on Google App Engine](/appengine/docs/python/microservices-on-app-engine), +// [App Engine Standard request routing](/appengine/docs/standard/python/how-requests-are-routed), +// and [App Engine Flex request routing](/appengine/docs/flexible/python/how-requests-are-routed). +type AppEngineRouting struct { + // App service. + // + // By default, the task is sent to the service which is the default + // service when the task is attempted. + // + // For some queues or tasks which were created using the App Engine + // Task Queue API, [host][google.cloud.tasks.v2beta2.AppEngineRouting.host] is not parsable + // into [service][google.cloud.tasks.v2beta2.AppEngineRouting.service], + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version], and + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance]. For example, some tasks + // which were created using the App Engine SDK use a custom domain + // name; custom domains are not parsed by Cloud Tasks. If + // [host][google.cloud.tasks.v2beta2.AppEngineRouting.host] is not parsable, then + // [service][google.cloud.tasks.v2beta2.AppEngineRouting.service], + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version], and + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] are the empty string. + Service string `protobuf:"bytes,1,opt,name=service,proto3" json:"service,omitempty"` + // App version. + // + // By default, the task is sent to the version which is the default + // version when the task is attempted. + // + // For some queues or tasks which were created using the App Engine + // Task Queue API, [host][google.cloud.tasks.v2beta2.AppEngineRouting.host] is not parsable + // into [service][google.cloud.tasks.v2beta2.AppEngineRouting.service], + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version], and + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance]. For example, some tasks + // which were created using the App Engine SDK use a custom domain + // name; custom domains are not parsed by Cloud Tasks. If + // [host][google.cloud.tasks.v2beta2.AppEngineRouting.host] is not parsable, then + // [service][google.cloud.tasks.v2beta2.AppEngineRouting.service], + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version], and + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] are the empty string. + Version string `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` + // App instance. + // + // By default, the task is sent to an instance which is available when + // the task is attempted. + // + // Requests can only be sent to a specific instance if + // [manual scaling is used in App Engine Standard](/appengine/docs/python/an-overview-of-app-engine?hl=en_US#scaling_types_and_instance_classes). + // App Engine Flex does not support instances. For more information, see + // [App Engine Standard request routing](/appengine/docs/standard/python/how-requests-are-routed) + // and [App Engine Flex request routing](/appengine/docs/flexible/python/how-requests-are-routed). + Instance string `protobuf:"bytes,3,opt,name=instance,proto3" json:"instance,omitempty"` + // Output only. The host that the task is sent to. + // + // For more information, see + // [How Requests are Routed](/appengine/docs/standard/python/how-requests-are-routed). + // + // The host is constructed as: + // + // + // * `host = [application_domain_name]`
+ // `| [service] + '.' + [application_domain_name]`
+ // `| [version] + '.' + [application_domain_name]`
+ // `| [version_dot_service]+ '.' + [application_domain_name]`
+ // `| [instance] + '.' + [application_domain_name]`
+ // `| [instance_dot_service] + '.' + [application_domain_name]`
+ // `| [instance_dot_version] + '.' + [application_domain_name]`
+ // `| [instance_dot_version_dot_service] + '.' + [application_domain_name]` + // + // * `application_domain_name` = The domain name of the app, for + // example .appspot.com, which is associated with the + // queue's project ID. Some tasks which were created using the App Engine + // SDK use a custom domain name. + // + // * `service =` [service][google.cloud.tasks.v2beta2.AppEngineRouting.service] + // + // * `version =` [version][google.cloud.tasks.v2beta2.AppEngineRouting.version] + // + // * `version_dot_service =` + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version] `+ '.' +` + // [service][google.cloud.tasks.v2beta2.AppEngineRouting.service] + // + // * `instance =` [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] + // + // * `instance_dot_service =` + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] `+ '.' +` + // [service][google.cloud.tasks.v2beta2.AppEngineRouting.service] + // + // * `instance_dot_version =` + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] `+ '.' +` + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version] + // + // * `instance_dot_version_dot_service =` + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] `+ '.' +` + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version] `+ '.' +` + // [service][google.cloud.tasks.v2beta2.AppEngineRouting.service] + // + // If [service][google.cloud.tasks.v2beta2.AppEngineRouting.service] is empty, then the task will be sent + // to the service which is the default service when the task is attempted. + // + // If [version][google.cloud.tasks.v2beta2.AppEngineRouting.version] is empty, then the task will be sent + // to the version which is the default version when the task is attempted. + // + // If [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] is empty, then the task + // will be sent to an instance which is available when the task is + // attempted. + // + // If [service][google.cloud.tasks.v2beta2.AppEngineRouting.service], + // [version][google.cloud.tasks.v2beta2.AppEngineRouting.version], or + // [instance][google.cloud.tasks.v2beta2.AppEngineRouting.instance] is invalid, then the task + // will be sent to the default version of the default service when + // the task is attempted. + Host string `protobuf:"bytes,4,opt,name=host,proto3" json:"host,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AppEngineRouting) Reset() { *m = AppEngineRouting{} } +func (m *AppEngineRouting) String() string { return proto.CompactTextString(m) } +func (*AppEngineRouting) ProtoMessage() {} +func (*AppEngineRouting) Descriptor() ([]byte, []int) { + return fileDescriptor_target_b871819819dd67b9, []int{4} +} +func (m *AppEngineRouting) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AppEngineRouting.Unmarshal(m, b) +} +func (m *AppEngineRouting) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AppEngineRouting.Marshal(b, m, deterministic) +} +func (dst *AppEngineRouting) XXX_Merge(src proto.Message) { + xxx_messageInfo_AppEngineRouting.Merge(dst, src) +} +func (m *AppEngineRouting) XXX_Size() int { + return xxx_messageInfo_AppEngineRouting.Size(m) +} +func (m *AppEngineRouting) XXX_DiscardUnknown() { + xxx_messageInfo_AppEngineRouting.DiscardUnknown(m) +} + +var xxx_messageInfo_AppEngineRouting proto.InternalMessageInfo + +func (m *AppEngineRouting) GetService() string { + if m != nil { + return m.Service + } + return "" +} + +func (m *AppEngineRouting) GetVersion() string { + if m != nil { + return m.Version + } + return "" +} + +func (m *AppEngineRouting) GetInstance() string { + if m != nil { + return m.Instance + } + return "" +} + +func (m *AppEngineRouting) GetHost() string { + if m != nil { + return m.Host + } + return "" +} + +func init() { + proto.RegisterType((*PullTarget)(nil), "google.cloud.tasks.v2beta2.PullTarget") + proto.RegisterType((*PullMessage)(nil), "google.cloud.tasks.v2beta2.PullMessage") + proto.RegisterType((*AppEngineHttpTarget)(nil), "google.cloud.tasks.v2beta2.AppEngineHttpTarget") + proto.RegisterType((*AppEngineHttpRequest)(nil), "google.cloud.tasks.v2beta2.AppEngineHttpRequest") + proto.RegisterMapType((map[string]string)(nil), "google.cloud.tasks.v2beta2.AppEngineHttpRequest.HeadersEntry") + proto.RegisterType((*AppEngineRouting)(nil), "google.cloud.tasks.v2beta2.AppEngineRouting") + proto.RegisterEnum("google.cloud.tasks.v2beta2.HttpMethod", HttpMethod_name, HttpMethod_value) +} + +func init() { + proto.RegisterFile("google/cloud/tasks/v2beta2/target.proto", fileDescriptor_target_b871819819dd67b9) +} + +var fileDescriptor_target_b871819819dd67b9 = []byte{ + // 557 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x54, 0xcf, 0x6f, 0xd3, 0x4c, + 0x10, 0xfd, 0x5c, 0xa7, 0xbf, 0xc6, 0xd1, 0x27, 0x6b, 0xa9, 0x84, 0x95, 0xa2, 0xaa, 0xe4, 0x00, + 0x15, 0x42, 0xb6, 0x14, 0x2e, 0x50, 0x84, 0x50, 0x4b, 0x4c, 0x53, 0x89, 0x12, 0xcb, 0x75, 0x84, + 0x54, 0x0e, 0xd6, 0x26, 0x19, 0x1c, 0x2b, 0xee, 0xae, 0xd9, 0x5d, 0x5b, 0xca, 0x95, 0x3b, 0xff, + 0x33, 0xf2, 0xda, 0x09, 0x69, 0x80, 0x0a, 0x6e, 0xf3, 0x66, 0xde, 0xbc, 0xc9, 0xbc, 0xf1, 0x06, + 0x9e, 0x26, 0x9c, 0x27, 0x19, 0x7a, 0x93, 0x8c, 0x17, 0x53, 0x4f, 0x51, 0x39, 0x97, 0x5e, 0xd9, + 0x1b, 0xa3, 0xa2, 0x3d, 0x4f, 0x51, 0x91, 0xa0, 0x72, 0x73, 0xc1, 0x15, 0x27, 0x9d, 0x9a, 0xe8, + 0x6a, 0xa2, 0xab, 0x89, 0x6e, 0x43, 0xec, 0x3c, 0x6a, 0x44, 0x68, 0x9e, 0x7a, 0x94, 0x31, 0xae, + 0xa8, 0x4a, 0x39, 0x93, 0x75, 0x67, 0xe7, 0xa8, 0xa9, 0x6a, 0x34, 0x2e, 0xbe, 0x78, 0xd3, 0x42, + 0x68, 0x42, 0x5d, 0xef, 0xb6, 0x01, 0x82, 0x22, 0xcb, 0x22, 0x3d, 0xad, 0xfb, 0x0a, 0xac, 0x0a, + 0x5d, 0xa1, 0x94, 0x34, 0x41, 0xe2, 0xc0, 0x6e, 0x4e, 0x17, 0x19, 0xa7, 0x53, 0xc7, 0x38, 0x36, + 0x4e, 0xda, 0xe1, 0x12, 0x12, 0x1b, 0x4c, 0x45, 0x13, 0x67, 0xeb, 0xd8, 0x38, 0xd9, 0x0f, 0xab, + 0xb0, 0xfb, 0xcd, 0x80, 0x07, 0x67, 0x79, 0xee, 0xb3, 0x24, 0x65, 0x38, 0x50, 0x2a, 0xaf, 0x25, + 0xc9, 0x1c, 0x0e, 0x69, 0x9e, 0xc7, 0xa8, 0xf3, 0xb1, 0xe0, 0x85, 0x4a, 0x59, 0x12, 0xf3, 0x12, + 0x85, 0x48, 0xa7, 0xa8, 0x75, 0xad, 0xde, 0x73, 0xf7, 0xcf, 0x0b, 0xba, 0x2b, 0xd5, 0xb0, 0x6e, + 0x0e, 0x1d, 0xba, 0x91, 0x19, 0x36, 0x6a, 0xdd, 0xef, 0x26, 0x1c, 0xdc, 0xf9, 0x11, 0x21, 0x7e, + 0x2d, 0x50, 0x2a, 0x72, 0x01, 0xd6, 0x4c, 0xa9, 0x3c, 0xbe, 0x45, 0x35, 0xe3, 0xf5, 0x36, 0xff, + 0xf7, 0x9e, 0xdc, 0x37, 0xb5, 0xea, 0xbe, 0xd2, 0xec, 0x10, 0x66, 0xab, 0x98, 0xdc, 0x00, 0xf9, + 0x75, 0x1d, 0xed, 0xc3, 0xbf, 0x6e, 0x61, 0x6f, 0x6e, 0x41, 0x1e, 0x43, 0x5b, 0x60, 0x46, 0x55, + 0x5a, 0x62, 0x5c, 0x88, 0xcc, 0x31, 0xb5, 0xbb, 0xd6, 0x32, 0x37, 0x12, 0x19, 0xf9, 0x04, 0xbb, + 0x33, 0xa4, 0x53, 0x14, 0xd2, 0x69, 0x1d, 0x9b, 0x27, 0x56, 0xef, 0xcd, 0x5f, 0xcd, 0x5c, 0xb3, + 0xc2, 0x1d, 0xd4, 0xfd, 0x3e, 0x53, 0x62, 0x11, 0x2e, 0xd5, 0xd6, 0x4f, 0xbd, 0x7d, 0xe7, 0xd4, + 0x9d, 0x53, 0x68, 0xaf, 0xb7, 0x54, 0xa7, 0x9f, 0xe3, 0x42, 0x5b, 0xb8, 0x1f, 0x56, 0x21, 0x39, + 0x80, 0xed, 0x92, 0x66, 0x05, 0x36, 0x9f, 0x43, 0x0d, 0x4e, 0xb7, 0x5e, 0x1a, 0xdd, 0x12, 0xec, + 0xcd, 0xbd, 0xab, 0x49, 0x12, 0x45, 0x99, 0x4e, 0xb0, 0xd1, 0x58, 0xc2, 0xaa, 0x52, 0xa2, 0x90, + 0x29, 0x67, 0x8d, 0xd2, 0x12, 0x92, 0x0e, 0xec, 0xa5, 0x4c, 0x2a, 0xca, 0x26, 0xd8, 0xb8, 0xb2, + 0xc2, 0x84, 0x40, 0x6b, 0xc6, 0xa5, 0x72, 0x5a, 0x3a, 0xaf, 0xe3, 0x67, 0x9f, 0x01, 0x7e, 0xde, + 0x8f, 0x1c, 0xc2, 0xc3, 0x41, 0x14, 0x05, 0xf1, 0x95, 0x1f, 0x0d, 0x86, 0xfd, 0x78, 0xf4, 0xf1, + 0x3a, 0xf0, 0xdf, 0x5d, 0xbe, 0xbf, 0xf4, 0xfb, 0xf6, 0x7f, 0x64, 0x0f, 0x5a, 0xc1, 0xf0, 0x3a, + 0xb2, 0x0d, 0xb2, 0x0b, 0xe6, 0x85, 0x1f, 0xd9, 0x5b, 0x55, 0x6a, 0xe0, 0x9f, 0xf5, 0x6d, 0xb3, + 0x4a, 0x05, 0xa3, 0xc8, 0x6e, 0x11, 0x80, 0x9d, 0xbe, 0xff, 0xc1, 0x8f, 0x7c, 0x7b, 0xfb, 0x3c, + 0x87, 0xa3, 0x09, 0xbf, 0xbd, 0xc7, 0xf7, 0x73, 0xab, 0xfe, 0xf6, 0x83, 0xea, 0x85, 0x05, 0xc6, + 0xcd, 0xdb, 0x86, 0x9a, 0xf0, 0x8c, 0xb2, 0xc4, 0xe5, 0x22, 0xf1, 0x12, 0x64, 0xfa, 0xfd, 0x79, + 0x75, 0x89, 0xe6, 0xa9, 0xfc, 0xdd, 0xbf, 0xc0, 0x6b, 0x8d, 0xc6, 0x3b, 0x9a, 0xfb, 0xe2, 0x47, + 0x00, 0x00, 0x00, 0xff, 0xff, 0xbb, 0x5d, 0x40, 0x26, 0x30, 0x04, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/task.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/task.pb.go new file mode 100644 index 000000000..5e0ea9cac --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/tasks/v2beta2/task.pb.go @@ -0,0 +1,505 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/tasks/v2beta2/task.proto + +package tasks // import "google.golang.org/genproto/googleapis/cloud/tasks/v2beta2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" +import _ "google.golang.org/genproto/googleapis/api/annotations" +import status "google.golang.org/genproto/googleapis/rpc/status" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// The view specifies a subset of [Task][google.cloud.tasks.v2beta2.Task] data. +// +// When a task is returned in a response, not all +// information is retrieved by default because some data, such as +// payloads, might be desirable to return only when needed because +// of its large size or because of the sensitivity of data that it +// contains. +type Task_View int32 + +const ( + // Unspecified. Defaults to BASIC. + Task_VIEW_UNSPECIFIED Task_View = 0 + // The basic view omits fields which can be large or can contain + // sensitive data. + // + // This view does not include the + // ([payload in AppEngineHttpRequest][google.cloud.tasks.v2beta2.AppEngineHttpRequest] and + // [payload in PullMessage][google.cloud.tasks.v2beta2.PullMessage.payload]). These payloads are + // desirable to return only when needed, because they can be large + // and because of the sensitivity of the data that you choose to + // store in it. + Task_BASIC Task_View = 1 + // All information is returned. + // + // Authorization for [FULL][google.cloud.tasks.v2beta2.Task.View.FULL] requires + // `cloudtasks.tasks.fullView` [Google IAM](https://cloud.google.com/iam/) + // permission on the [Queue][google.cloud.tasks.v2beta2.Queue] resource. + Task_FULL Task_View = 2 +) + +var Task_View_name = map[int32]string{ + 0: "VIEW_UNSPECIFIED", + 1: "BASIC", + 2: "FULL", +} +var Task_View_value = map[string]int32{ + "VIEW_UNSPECIFIED": 0, + "BASIC": 1, + "FULL": 2, +} + +func (x Task_View) String() string { + return proto.EnumName(Task_View_name, int32(x)) +} +func (Task_View) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_task_3b902b34f1c5392b, []int{0, 0} +} + +// A unit of scheduled work. +type Task struct { + // Optionally caller-specified in [CreateTask][google.cloud.tasks.v2beta2.CloudTasks.CreateTask]. + // + // The task name. + // + // The task name must have the following format: + // `projects/PROJECT_ID/locations/LOCATION_ID/queues/QUEUE_ID/tasks/TASK_ID` + // + // * `PROJECT_ID` can contain letters ([A-Za-z]), numbers ([0-9]), + // hyphens (-), colons (:), or periods (.). + // For more information, see + // [Identifying projects](/resource-manager/docs/creating-managing-projects#identifying_projects) + // * `LOCATION_ID` is the canonical ID for the task's location. + // The list of available locations can be obtained by calling + // [ListLocations][google.cloud.location.Locations.ListLocations]. + // For more information, see https://cloud.google.com/about/locations/. + // * `QUEUE_ID` can contain letters ([A-Za-z]), numbers ([0-9]), or + // hyphens (-). The maximum length is 100 characters. + // * `TASK_ID` can contain only letters ([A-Za-z]), numbers ([0-9]), + // hyphens (-), or underscores (_). The maximum length is 500 characters. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Required. + // + // The task's payload is used by the task's target to process the task. + // A payload is valid only if it is compatible with the queue's target. + // + // Types that are valid to be assigned to PayloadType: + // *Task_AppEngineHttpRequest + // *Task_PullMessage + PayloadType isTask_PayloadType `protobuf_oneof:"payload_type"` + // The time when the task is scheduled to be attempted. + // + // For App Engine queues, this is when the task will be attempted or retried. + // + // For pull queues, this is the time when the task is available to + // be leased; if a task is currently leased, this is the time when + // the current lease expires, that is, the time that the task was + // leased plus the [lease_duration][google.cloud.tasks.v2beta2.LeaseTasksRequest.lease_duration]. + // + // `schedule_time` will be truncated to the nearest microsecond. + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` + // Output only. The time that the task was created. + // + // `create_time` will be truncated to the nearest second. + CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` + // Output only. The task status. + Status *TaskStatus `protobuf:"bytes,7,opt,name=status,proto3" json:"status,omitempty"` + // Output only. The view specifies which subset of the [Task][google.cloud.tasks.v2beta2.Task] has + // been returned. + View Task_View `protobuf:"varint,8,opt,name=view,proto3,enum=google.cloud.tasks.v2beta2.Task_View" json:"view,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Task) Reset() { *m = Task{} } +func (m *Task) String() string { return proto.CompactTextString(m) } +func (*Task) ProtoMessage() {} +func (*Task) Descriptor() ([]byte, []int) { + return fileDescriptor_task_3b902b34f1c5392b, []int{0} +} +func (m *Task) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Task.Unmarshal(m, b) +} +func (m *Task) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Task.Marshal(b, m, deterministic) +} +func (dst *Task) XXX_Merge(src proto.Message) { + xxx_messageInfo_Task.Merge(dst, src) +} +func (m *Task) XXX_Size() int { + return xxx_messageInfo_Task.Size(m) +} +func (m *Task) XXX_DiscardUnknown() { + xxx_messageInfo_Task.DiscardUnknown(m) +} + +var xxx_messageInfo_Task proto.InternalMessageInfo + +type isTask_PayloadType interface { + isTask_PayloadType() +} + +type Task_AppEngineHttpRequest struct { + AppEngineHttpRequest *AppEngineHttpRequest `protobuf:"bytes,3,opt,name=app_engine_http_request,json=appEngineHttpRequest,proto3,oneof"` +} +type Task_PullMessage struct { + PullMessage *PullMessage `protobuf:"bytes,4,opt,name=pull_message,json=pullMessage,proto3,oneof"` +} + +func (*Task_AppEngineHttpRequest) isTask_PayloadType() {} +func (*Task_PullMessage) isTask_PayloadType() {} + +func (m *Task) GetPayloadType() isTask_PayloadType { + if m != nil { + return m.PayloadType + } + return nil +} + +func (m *Task) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Task) GetAppEngineHttpRequest() *AppEngineHttpRequest { + if x, ok := m.GetPayloadType().(*Task_AppEngineHttpRequest); ok { + return x.AppEngineHttpRequest + } + return nil +} + +func (m *Task) GetPullMessage() *PullMessage { + if x, ok := m.GetPayloadType().(*Task_PullMessage); ok { + return x.PullMessage + } + return nil +} + +func (m *Task) GetScheduleTime() *timestamp.Timestamp { + if m != nil { + return m.ScheduleTime + } + return nil +} + +func (m *Task) GetCreateTime() *timestamp.Timestamp { + if m != nil { + return m.CreateTime + } + return nil +} + +func (m *Task) GetStatus() *TaskStatus { + if m != nil { + return m.Status + } + return nil +} + +func (m *Task) GetView() Task_View { + if m != nil { + return m.View + } + return Task_VIEW_UNSPECIFIED +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Task) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Task_OneofMarshaler, _Task_OneofUnmarshaler, _Task_OneofSizer, []interface{}{ + (*Task_AppEngineHttpRequest)(nil), + (*Task_PullMessage)(nil), + } +} + +func _Task_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Task) + // payload_type + switch x := m.PayloadType.(type) { + case *Task_AppEngineHttpRequest: + b.EncodeVarint(3<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.AppEngineHttpRequest); err != nil { + return err + } + case *Task_PullMessage: + b.EncodeVarint(4<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.PullMessage); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("Task.PayloadType has unexpected type %T", x) + } + return nil +} + +func _Task_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Task) + switch tag { + case 3: // payload_type.app_engine_http_request + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(AppEngineHttpRequest) + err := b.DecodeMessage(msg) + m.PayloadType = &Task_AppEngineHttpRequest{msg} + return true, err + case 4: // payload_type.pull_message + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(PullMessage) + err := b.DecodeMessage(msg) + m.PayloadType = &Task_PullMessage{msg} + return true, err + default: + return false, nil + } +} + +func _Task_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Task) + // payload_type + switch x := m.PayloadType.(type) { + case *Task_AppEngineHttpRequest: + s := proto.Size(x.AppEngineHttpRequest) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *Task_PullMessage: + s := proto.Size(x.PullMessage) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Status of the task. +type TaskStatus struct { + // Output only. The number of attempts dispatched. + // + // This count includes tasks which have been dispatched but haven't + // received a response. + AttemptDispatchCount int32 `protobuf:"varint,1,opt,name=attempt_dispatch_count,json=attemptDispatchCount,proto3" json:"attempt_dispatch_count,omitempty"` + // Output only. The number of attempts which have received a response. + // + // This field is not calculated for + // [pull tasks](google.cloud.tasks.v2beta2.PullTaskTarget). + AttemptResponseCount int32 `protobuf:"varint,2,opt,name=attempt_response_count,json=attemptResponseCount,proto3" json:"attempt_response_count,omitempty"` + // Output only. The status of the task's first attempt. + // + // Only [dispatch_time][google.cloud.tasks.v2beta2.AttemptStatus.dispatch_time] will be set. + // The other [AttemptStatus][google.cloud.tasks.v2beta2.AttemptStatus] information is not retained by Cloud Tasks. + // + // This field is not calculated for + // [pull tasks](google.cloud.tasks.v2beta2.PullTaskTarget). + FirstAttemptStatus *AttemptStatus `protobuf:"bytes,3,opt,name=first_attempt_status,json=firstAttemptStatus,proto3" json:"first_attempt_status,omitempty"` + // Output only. The status of the task's last attempt. + // + // This field is not calculated for + // [pull tasks](google.cloud.tasks.v2beta2.PullTaskTarget). + LastAttemptStatus *AttemptStatus `protobuf:"bytes,4,opt,name=last_attempt_status,json=lastAttemptStatus,proto3" json:"last_attempt_status,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TaskStatus) Reset() { *m = TaskStatus{} } +func (m *TaskStatus) String() string { return proto.CompactTextString(m) } +func (*TaskStatus) ProtoMessage() {} +func (*TaskStatus) Descriptor() ([]byte, []int) { + return fileDescriptor_task_3b902b34f1c5392b, []int{1} +} +func (m *TaskStatus) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TaskStatus.Unmarshal(m, b) +} +func (m *TaskStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TaskStatus.Marshal(b, m, deterministic) +} +func (dst *TaskStatus) XXX_Merge(src proto.Message) { + xxx_messageInfo_TaskStatus.Merge(dst, src) +} +func (m *TaskStatus) XXX_Size() int { + return xxx_messageInfo_TaskStatus.Size(m) +} +func (m *TaskStatus) XXX_DiscardUnknown() { + xxx_messageInfo_TaskStatus.DiscardUnknown(m) +} + +var xxx_messageInfo_TaskStatus proto.InternalMessageInfo + +func (m *TaskStatus) GetAttemptDispatchCount() int32 { + if m != nil { + return m.AttemptDispatchCount + } + return 0 +} + +func (m *TaskStatus) GetAttemptResponseCount() int32 { + if m != nil { + return m.AttemptResponseCount + } + return 0 +} + +func (m *TaskStatus) GetFirstAttemptStatus() *AttemptStatus { + if m != nil { + return m.FirstAttemptStatus + } + return nil +} + +func (m *TaskStatus) GetLastAttemptStatus() *AttemptStatus { + if m != nil { + return m.LastAttemptStatus + } + return nil +} + +// The status of a task attempt. +type AttemptStatus struct { + // Output only. The time that this attempt was scheduled. + // + // `schedule_time` will be truncated to the nearest microsecond. + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` + // Output only. The time that this attempt was dispatched. + // + // `dispatch_time` will be truncated to the nearest microsecond. + DispatchTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=dispatch_time,json=dispatchTime,proto3" json:"dispatch_time,omitempty"` + // Output only. The time that this attempt response was received. + // + // `response_time` will be truncated to the nearest microsecond. + ResponseTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=response_time,json=responseTime,proto3" json:"response_time,omitempty"` + // Output only. The response from the target for this attempt. + // + // If the task has not been attempted or the task is currently running + // then the response status is unset. + ResponseStatus *status.Status `protobuf:"bytes,4,opt,name=response_status,json=responseStatus,proto3" json:"response_status,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AttemptStatus) Reset() { *m = AttemptStatus{} } +func (m *AttemptStatus) String() string { return proto.CompactTextString(m) } +func (*AttemptStatus) ProtoMessage() {} +func (*AttemptStatus) Descriptor() ([]byte, []int) { + return fileDescriptor_task_3b902b34f1c5392b, []int{2} +} +func (m *AttemptStatus) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AttemptStatus.Unmarshal(m, b) +} +func (m *AttemptStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AttemptStatus.Marshal(b, m, deterministic) +} +func (dst *AttemptStatus) XXX_Merge(src proto.Message) { + xxx_messageInfo_AttemptStatus.Merge(dst, src) +} +func (m *AttemptStatus) XXX_Size() int { + return xxx_messageInfo_AttemptStatus.Size(m) +} +func (m *AttemptStatus) XXX_DiscardUnknown() { + xxx_messageInfo_AttemptStatus.DiscardUnknown(m) +} + +var xxx_messageInfo_AttemptStatus proto.InternalMessageInfo + +func (m *AttemptStatus) GetScheduleTime() *timestamp.Timestamp { + if m != nil { + return m.ScheduleTime + } + return nil +} + +func (m *AttemptStatus) GetDispatchTime() *timestamp.Timestamp { + if m != nil { + return m.DispatchTime + } + return nil +} + +func (m *AttemptStatus) GetResponseTime() *timestamp.Timestamp { + if m != nil { + return m.ResponseTime + } + return nil +} + +func (m *AttemptStatus) GetResponseStatus() *status.Status { + if m != nil { + return m.ResponseStatus + } + return nil +} + +func init() { + proto.RegisterType((*Task)(nil), "google.cloud.tasks.v2beta2.Task") + proto.RegisterType((*TaskStatus)(nil), "google.cloud.tasks.v2beta2.TaskStatus") + proto.RegisterType((*AttemptStatus)(nil), "google.cloud.tasks.v2beta2.AttemptStatus") + proto.RegisterEnum("google.cloud.tasks.v2beta2.Task_View", Task_View_name, Task_View_value) +} + +func init() { + proto.RegisterFile("google/cloud/tasks/v2beta2/task.proto", fileDescriptor_task_3b902b34f1c5392b) +} + +var fileDescriptor_task_3b902b34f1c5392b = []byte{ + // 601 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x94, 0x5f, 0x6f, 0xda, 0x3c, + 0x14, 0xc6, 0x1b, 0x9a, 0xf6, 0x2d, 0xe6, 0xcf, 0xcb, 0x3c, 0xb4, 0x46, 0x68, 0xda, 0x10, 0x52, + 0x57, 0x76, 0x93, 0x6c, 0x6c, 0x37, 0x53, 0xa5, 0xa1, 0x42, 0xa9, 0x40, 0x62, 0x13, 0x0a, 0x6d, + 0xa7, 0x6d, 0x17, 0x91, 0x09, 0x6e, 0x88, 0x9a, 0xc4, 0x5e, 0x7c, 0xd2, 0xaa, 0x9f, 0xa2, 0x5f, + 0x65, 0x1f, 0x71, 0x8a, 0xe3, 0xd0, 0xa1, 0x76, 0xa0, 0xde, 0x71, 0xce, 0x79, 0x7e, 0x8f, 0xed, + 0x47, 0x87, 0xa0, 0x03, 0x8f, 0x31, 0x2f, 0xa0, 0x96, 0x1b, 0xb0, 0x64, 0x6e, 0x01, 0x11, 0x57, + 0xc2, 0xba, 0xee, 0xcc, 0x28, 0x90, 0x8e, 0xac, 0x4c, 0x1e, 0x33, 0x60, 0xb8, 0x91, 0xc9, 0x4c, + 0x29, 0x33, 0xa5, 0xcc, 0x54, 0xb2, 0xc6, 0x4b, 0x65, 0x41, 0xb8, 0x6f, 0x91, 0x28, 0x62, 0x40, + 0xc0, 0x67, 0x91, 0xc8, 0xc8, 0xc6, 0xe1, 0xda, 0x03, 0x62, 0x8f, 0x82, 0x12, 0xbe, 0x56, 0x42, + 0x59, 0xcd, 0x92, 0x4b, 0x0b, 0xfc, 0x90, 0x0a, 0x20, 0x21, 0x57, 0x82, 0x7d, 0x25, 0x88, 0xb9, + 0x6b, 0x09, 0x20, 0x90, 0xa8, 0x23, 0x5a, 0x77, 0x3a, 0xd2, 0xcf, 0x88, 0xb8, 0xc2, 0x18, 0xe9, + 0x11, 0x09, 0xa9, 0xa1, 0x35, 0xb5, 0x76, 0xd1, 0x96, 0xbf, 0xb1, 0x8f, 0xf6, 0x09, 0xe7, 0x0e, + 0x8d, 0x3c, 0x3f, 0xa2, 0xce, 0x02, 0x80, 0x3b, 0x31, 0xfd, 0x95, 0x50, 0x01, 0xc6, 0x76, 0x53, + 0x6b, 0x97, 0x3a, 0xef, 0xcc, 0x7f, 0xbf, 0xcd, 0x3c, 0xe6, 0x7c, 0x20, 0xc9, 0x21, 0x00, 0xb7, + 0x33, 0x6e, 0xb8, 0x65, 0xd7, 0xc9, 0x23, 0x7d, 0x3c, 0x46, 0x65, 0x9e, 0x04, 0x81, 0x13, 0x52, + 0x21, 0x88, 0x47, 0x0d, 0x5d, 0xfa, 0x1f, 0xae, 0xf3, 0x9f, 0x24, 0x41, 0xf0, 0x25, 0x93, 0x0f, + 0xb7, 0xec, 0x12, 0xbf, 0x2f, 0x71, 0x17, 0x55, 0x84, 0xbb, 0xa0, 0xf3, 0x24, 0xa0, 0x4e, 0x1a, + 0x85, 0xb1, 0x23, 0xed, 0x1a, 0xb9, 0x5d, 0x9e, 0x93, 0x79, 0x96, 0xe7, 0x64, 0x97, 0x73, 0x20, + 0x6d, 0xe1, 0x23, 0x54, 0x72, 0x63, 0x4a, 0x40, 0xe1, 0xbb, 0x1b, 0x71, 0x94, 0xc9, 0x25, 0xfc, + 0x19, 0xed, 0x66, 0x19, 0x1b, 0xff, 0x49, 0xee, 0xcd, 0xba, 0x57, 0xa4, 0xe1, 0x4f, 0xa5, 0xda, + 0x56, 0x14, 0xfe, 0x84, 0xf4, 0x6b, 0x9f, 0xde, 0x18, 0x7b, 0x4d, 0xad, 0x5d, 0xed, 0x1c, 0x6c, + 0xa2, 0xcd, 0x0b, 0x9f, 0xde, 0xd8, 0x12, 0x69, 0xbd, 0x47, 0x7a, 0x5a, 0xe1, 0x3a, 0xaa, 0x5d, + 0x8c, 0x06, 0xdf, 0x9c, 0xf3, 0xaf, 0xd3, 0xc9, 0xa0, 0x3f, 0x3a, 0x1d, 0x0d, 0x4e, 0x6a, 0x5b, + 0xb8, 0x88, 0x76, 0x7a, 0xc7, 0xd3, 0x51, 0xbf, 0xa6, 0xe1, 0x3d, 0xa4, 0x9f, 0x9e, 0x8f, 0xc7, + 0xb5, 0x42, 0xaf, 0x8a, 0xca, 0x9c, 0xdc, 0x06, 0x8c, 0xcc, 0x1d, 0xb8, 0xe5, 0xb4, 0xf5, 0xbb, + 0x80, 0xd0, 0xfd, 0xa5, 0xf0, 0x47, 0xf4, 0x82, 0x00, 0xd0, 0x90, 0x83, 0x33, 0xf7, 0x05, 0x27, + 0xe0, 0x2e, 0x1c, 0x97, 0x25, 0x11, 0xc8, 0x4d, 0xd9, 0xb1, 0xeb, 0x6a, 0x7a, 0xa2, 0x86, 0xfd, + 0x74, 0xf6, 0x37, 0x15, 0x53, 0xc1, 0x59, 0x24, 0xa8, 0xa2, 0x0a, 0x2b, 0x94, 0xad, 0x86, 0x19, + 0xf5, 0x13, 0xd5, 0x2f, 0xfd, 0x58, 0x80, 0x93, 0xb3, 0x2a, 0xc6, 0x6c, 0xd9, 0xde, 0xae, 0x5d, + 0xb6, 0x8c, 0x50, 0x49, 0x62, 0x69, 0xb3, 0xd2, 0xc3, 0xdf, 0xd1, 0xf3, 0x80, 0x3c, 0xf4, 0xd6, + 0x9f, 0xea, 0xfd, 0x2c, 0x75, 0x59, 0x69, 0xb5, 0xee, 0x0a, 0xa8, 0xb2, 0x7a, 0xd8, 0x83, 0x05, + 0xd4, 0x9e, 0xb8, 0x80, 0x5d, 0x54, 0x59, 0xc6, 0x2d, 0x0d, 0x0a, 0x9b, 0x0d, 0x72, 0x20, 0x37, + 0x58, 0x26, 0x2f, 0x0d, 0xb6, 0x37, 0x1b, 0xe4, 0x80, 0xfa, 0x0b, 0xfc, 0xbf, 0x34, 0x58, 0xc9, + 0x0a, 0xe7, 0x16, 0x31, 0x77, 0x4d, 0x15, 0x4a, 0x35, 0x97, 0x66, 0x75, 0x2f, 0x42, 0xaf, 0x5c, + 0x16, 0xae, 0x09, 0xb5, 0x57, 0x4c, 0x77, 0x6c, 0x92, 0x5e, 0x62, 0xa2, 0xfd, 0xe8, 0x2a, 0xa1, + 0xc7, 0x02, 0x12, 0x79, 0x26, 0x8b, 0x3d, 0xcb, 0xa3, 0x91, 0xbc, 0xa2, 0x95, 0x8d, 0x08, 0xf7, + 0xc5, 0x63, 0xdf, 0xc1, 0x23, 0x59, 0xcd, 0x76, 0xa5, 0xf6, 0xc3, 0x9f, 0x00, 0x00, 0x00, 0xff, + 0xff, 0x67, 0x07, 0xb1, 0x59, 0x93, 0x05, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1/cloud_tts.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1/cloud_tts.pb.go new file mode 100644 index 000000000..431f3d955 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1/cloud_tts.pb.go @@ -0,0 +1,856 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/cloud/texttospeech/v1/cloud_tts.proto + +package texttospeech // import "google.golang.org/genproto/googleapis/cloud/texttospeech/v1" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Gender of the voice as described in +// [SSML voice element](https://www.w3.org/TR/speech-synthesis11/#edef_voice). +type SsmlVoiceGender int32 + +const ( + // An unspecified gender. + // In VoiceSelectionParams, this means that the client doesn't care which + // gender the selected voice will have. In the Voice field of + // ListVoicesResponse, this may mean that the voice doesn't fit any of the + // other categories in this enum, or that the gender of the voice isn't known. + SsmlVoiceGender_SSML_VOICE_GENDER_UNSPECIFIED SsmlVoiceGender = 0 + // A male voice. + SsmlVoiceGender_MALE SsmlVoiceGender = 1 + // A female voice. + SsmlVoiceGender_FEMALE SsmlVoiceGender = 2 + // A gender-neutral voice. + SsmlVoiceGender_NEUTRAL SsmlVoiceGender = 3 +) + +var SsmlVoiceGender_name = map[int32]string{ + 0: "SSML_VOICE_GENDER_UNSPECIFIED", + 1: "MALE", + 2: "FEMALE", + 3: "NEUTRAL", +} +var SsmlVoiceGender_value = map[string]int32{ + "SSML_VOICE_GENDER_UNSPECIFIED": 0, + "MALE": 1, + "FEMALE": 2, + "NEUTRAL": 3, +} + +func (x SsmlVoiceGender) String() string { + return proto.EnumName(SsmlVoiceGender_name, int32(x)) +} +func (SsmlVoiceGender) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{0} +} + +// Configuration to set up audio encoder. The encoding determines the output +// audio format that we'd like. +type AudioEncoding int32 + +const ( + // Not specified. Will return result [google.rpc.Code.INVALID_ARGUMENT][]. + AudioEncoding_AUDIO_ENCODING_UNSPECIFIED AudioEncoding = 0 + // Uncompressed 16-bit signed little-endian samples (Linear PCM). + // Audio content returned as LINEAR16 also contains a WAV header. + AudioEncoding_LINEAR16 AudioEncoding = 1 + // MP3 audio. + AudioEncoding_MP3 AudioEncoding = 2 + // Opus encoded audio wrapped in an ogg container. The result will be a + // file which can be played natively on Android, and in browsers (at least + // Chrome and Firefox). The quality of the encoding is considerably higher + // than MP3 while using approximately the same bitrate. + AudioEncoding_OGG_OPUS AudioEncoding = 3 +) + +var AudioEncoding_name = map[int32]string{ + 0: "AUDIO_ENCODING_UNSPECIFIED", + 1: "LINEAR16", + 2: "MP3", + 3: "OGG_OPUS", +} +var AudioEncoding_value = map[string]int32{ + "AUDIO_ENCODING_UNSPECIFIED": 0, + "LINEAR16": 1, + "MP3": 2, + "OGG_OPUS": 3, +} + +func (x AudioEncoding) String() string { + return proto.EnumName(AudioEncoding_name, int32(x)) +} +func (AudioEncoding) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{1} +} + +// The top-level message sent by the client for the `ListVoices` method. +type ListVoicesRequest struct { + // Optional (but recommended) + // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. If + // specified, the ListVoices call will only return voices that can be used to + // synthesize this language_code. E.g. when specifying "en-NZ", you will get + // supported "en-*" voices; when specifying "no", you will get supported + // "no-*" (Norwegian) and "nb-*" (Norwegian Bokmal) voices; specifying "zh" + // will also get supported "cmn-*" voices; specifying "zh-hk" will also get + // supported "yue-*" voices. + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListVoicesRequest) Reset() { *m = ListVoicesRequest{} } +func (m *ListVoicesRequest) String() string { return proto.CompactTextString(m) } +func (*ListVoicesRequest) ProtoMessage() {} +func (*ListVoicesRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{0} +} +func (m *ListVoicesRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListVoicesRequest.Unmarshal(m, b) +} +func (m *ListVoicesRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListVoicesRequest.Marshal(b, m, deterministic) +} +func (dst *ListVoicesRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListVoicesRequest.Merge(dst, src) +} +func (m *ListVoicesRequest) XXX_Size() int { + return xxx_messageInfo_ListVoicesRequest.Size(m) +} +func (m *ListVoicesRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListVoicesRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListVoicesRequest proto.InternalMessageInfo + +func (m *ListVoicesRequest) GetLanguageCode() string { + if m != nil { + return m.LanguageCode + } + return "" +} + +// The message returned to the client by the `ListVoices` method. +type ListVoicesResponse struct { + // The list of voices. + Voices []*Voice `protobuf:"bytes,1,rep,name=voices,proto3" json:"voices,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListVoicesResponse) Reset() { *m = ListVoicesResponse{} } +func (m *ListVoicesResponse) String() string { return proto.CompactTextString(m) } +func (*ListVoicesResponse) ProtoMessage() {} +func (*ListVoicesResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{1} +} +func (m *ListVoicesResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListVoicesResponse.Unmarshal(m, b) +} +func (m *ListVoicesResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListVoicesResponse.Marshal(b, m, deterministic) +} +func (dst *ListVoicesResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListVoicesResponse.Merge(dst, src) +} +func (m *ListVoicesResponse) XXX_Size() int { + return xxx_messageInfo_ListVoicesResponse.Size(m) +} +func (m *ListVoicesResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListVoicesResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListVoicesResponse proto.InternalMessageInfo + +func (m *ListVoicesResponse) GetVoices() []*Voice { + if m != nil { + return m.Voices + } + return nil +} + +// Description of a voice supported by the TTS service. +type Voice struct { + // The languages that this voice supports, expressed as + // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags (e.g. + // "en-US", "es-419", "cmn-tw"). + LanguageCodes []string `protobuf:"bytes,1,rep,name=language_codes,json=languageCodes,proto3" json:"language_codes,omitempty"` + // The name of this voice. Each distinct voice has a unique name. + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` + // The gender of this voice. + SsmlGender SsmlVoiceGender `protobuf:"varint,3,opt,name=ssml_gender,json=ssmlGender,proto3,enum=google.cloud.texttospeech.v1.SsmlVoiceGender" json:"ssml_gender,omitempty"` + // The natural sample rate (in hertz) for this voice. + NaturalSampleRateHertz int32 `protobuf:"varint,4,opt,name=natural_sample_rate_hertz,json=naturalSampleRateHertz,proto3" json:"natural_sample_rate_hertz,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Voice) Reset() { *m = Voice{} } +func (m *Voice) String() string { return proto.CompactTextString(m) } +func (*Voice) ProtoMessage() {} +func (*Voice) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{2} +} +func (m *Voice) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Voice.Unmarshal(m, b) +} +func (m *Voice) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Voice.Marshal(b, m, deterministic) +} +func (dst *Voice) XXX_Merge(src proto.Message) { + xxx_messageInfo_Voice.Merge(dst, src) +} +func (m *Voice) XXX_Size() int { + return xxx_messageInfo_Voice.Size(m) +} +func (m *Voice) XXX_DiscardUnknown() { + xxx_messageInfo_Voice.DiscardUnknown(m) +} + +var xxx_messageInfo_Voice proto.InternalMessageInfo + +func (m *Voice) GetLanguageCodes() []string { + if m != nil { + return m.LanguageCodes + } + return nil +} + +func (m *Voice) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Voice) GetSsmlGender() SsmlVoiceGender { + if m != nil { + return m.SsmlGender + } + return SsmlVoiceGender_SSML_VOICE_GENDER_UNSPECIFIED +} + +func (m *Voice) GetNaturalSampleRateHertz() int32 { + if m != nil { + return m.NaturalSampleRateHertz + } + return 0 +} + +// The top-level message sent by the client for the `SynthesizeSpeech` method. +type SynthesizeSpeechRequest struct { + // Required. The Synthesizer requires either plain text or SSML as input. + Input *SynthesisInput `protobuf:"bytes,1,opt,name=input,proto3" json:"input,omitempty"` + // Required. The desired voice of the synthesized audio. + Voice *VoiceSelectionParams `protobuf:"bytes,2,opt,name=voice,proto3" json:"voice,omitempty"` + // Required. The configuration of the synthesized audio. + AudioConfig *AudioConfig `protobuf:"bytes,3,opt,name=audio_config,json=audioConfig,proto3" json:"audio_config,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SynthesizeSpeechRequest) Reset() { *m = SynthesizeSpeechRequest{} } +func (m *SynthesizeSpeechRequest) String() string { return proto.CompactTextString(m) } +func (*SynthesizeSpeechRequest) ProtoMessage() {} +func (*SynthesizeSpeechRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{3} +} +func (m *SynthesizeSpeechRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SynthesizeSpeechRequest.Unmarshal(m, b) +} +func (m *SynthesizeSpeechRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SynthesizeSpeechRequest.Marshal(b, m, deterministic) +} +func (dst *SynthesizeSpeechRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_SynthesizeSpeechRequest.Merge(dst, src) +} +func (m *SynthesizeSpeechRequest) XXX_Size() int { + return xxx_messageInfo_SynthesizeSpeechRequest.Size(m) +} +func (m *SynthesizeSpeechRequest) XXX_DiscardUnknown() { + xxx_messageInfo_SynthesizeSpeechRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_SynthesizeSpeechRequest proto.InternalMessageInfo + +func (m *SynthesizeSpeechRequest) GetInput() *SynthesisInput { + if m != nil { + return m.Input + } + return nil +} + +func (m *SynthesizeSpeechRequest) GetVoice() *VoiceSelectionParams { + if m != nil { + return m.Voice + } + return nil +} + +func (m *SynthesizeSpeechRequest) GetAudioConfig() *AudioConfig { + if m != nil { + return m.AudioConfig + } + return nil +} + +// Contains text input to be synthesized. Either `text` or `ssml` must be +// supplied. Supplying both or neither returns +// [google.rpc.Code.INVALID_ARGUMENT][]. The input size is limited to 5000 +// characters. +type SynthesisInput struct { + // The input source, which is either plain text or SSML. + // + // Types that are valid to be assigned to InputSource: + // *SynthesisInput_Text + // *SynthesisInput_Ssml + InputSource isSynthesisInput_InputSource `protobuf_oneof:"input_source"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SynthesisInput) Reset() { *m = SynthesisInput{} } +func (m *SynthesisInput) String() string { return proto.CompactTextString(m) } +func (*SynthesisInput) ProtoMessage() {} +func (*SynthesisInput) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{4} +} +func (m *SynthesisInput) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SynthesisInput.Unmarshal(m, b) +} +func (m *SynthesisInput) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SynthesisInput.Marshal(b, m, deterministic) +} +func (dst *SynthesisInput) XXX_Merge(src proto.Message) { + xxx_messageInfo_SynthesisInput.Merge(dst, src) +} +func (m *SynthesisInput) XXX_Size() int { + return xxx_messageInfo_SynthesisInput.Size(m) +} +func (m *SynthesisInput) XXX_DiscardUnknown() { + xxx_messageInfo_SynthesisInput.DiscardUnknown(m) +} + +var xxx_messageInfo_SynthesisInput proto.InternalMessageInfo + +type isSynthesisInput_InputSource interface { + isSynthesisInput_InputSource() +} + +type SynthesisInput_Text struct { + Text string `protobuf:"bytes,1,opt,name=text,proto3,oneof"` +} +type SynthesisInput_Ssml struct { + Ssml string `protobuf:"bytes,2,opt,name=ssml,proto3,oneof"` +} + +func (*SynthesisInput_Text) isSynthesisInput_InputSource() {} +func (*SynthesisInput_Ssml) isSynthesisInput_InputSource() {} + +func (m *SynthesisInput) GetInputSource() isSynthesisInput_InputSource { + if m != nil { + return m.InputSource + } + return nil +} + +func (m *SynthesisInput) GetText() string { + if x, ok := m.GetInputSource().(*SynthesisInput_Text); ok { + return x.Text + } + return "" +} + +func (m *SynthesisInput) GetSsml() string { + if x, ok := m.GetInputSource().(*SynthesisInput_Ssml); ok { + return x.Ssml + } + return "" +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*SynthesisInput) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _SynthesisInput_OneofMarshaler, _SynthesisInput_OneofUnmarshaler, _SynthesisInput_OneofSizer, []interface{}{ + (*SynthesisInput_Text)(nil), + (*SynthesisInput_Ssml)(nil), + } +} + +func _SynthesisInput_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*SynthesisInput) + // input_source + switch x := m.InputSource.(type) { + case *SynthesisInput_Text: + b.EncodeVarint(1<<3 | proto.WireBytes) + b.EncodeStringBytes(x.Text) + case *SynthesisInput_Ssml: + b.EncodeVarint(2<<3 | proto.WireBytes) + b.EncodeStringBytes(x.Ssml) + case nil: + default: + return fmt.Errorf("SynthesisInput.InputSource has unexpected type %T", x) + } + return nil +} + +func _SynthesisInput_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*SynthesisInput) + switch tag { + case 1: // input_source.text + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.InputSource = &SynthesisInput_Text{x} + return true, err + case 2: // input_source.ssml + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.InputSource = &SynthesisInput_Ssml{x} + return true, err + default: + return false, nil + } +} + +func _SynthesisInput_OneofSizer(msg proto.Message) (n int) { + m := msg.(*SynthesisInput) + // input_source + switch x := m.InputSource.(type) { + case *SynthesisInput_Text: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.Text))) + n += len(x.Text) + case *SynthesisInput_Ssml: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.Ssml))) + n += len(x.Ssml) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Description of which voice to use for a synthesis request. +type VoiceSelectionParams struct { + // The language (and optionally also the region) of the voice expressed as a + // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag, e.g. + // "en-US". Required. This should not include a script tag (e.g. use + // "cmn-cn" rather than "cmn-Hant-cn"), because the script will be inferred + // from the input provided in the SynthesisInput. The TTS service + // will use this parameter to help choose an appropriate voice. Note that + // the TTS service may choose a voice with a slightly different language code + // than the one selected; it may substitute a different region + // (e.g. using en-US rather than en-CA if there isn't a Canadian voice + // available), or even a different language, e.g. using "nb" (Norwegian + // Bokmal) instead of "no" (Norwegian)". + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` + // The name of the voice. Optional; if not set, the service will choose a + // voice based on the other parameters such as language_code and gender. + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` + // The preferred gender of the voice. Optional; if not set, the service will + // choose a voice based on the other parameters such as language_code and + // name. Note that this is only a preference, not requirement; if a + // voice of the appropriate gender is not available, the synthesizer should + // substitute a voice with a different gender rather than failing the request. + SsmlGender SsmlVoiceGender `protobuf:"varint,3,opt,name=ssml_gender,json=ssmlGender,proto3,enum=google.cloud.texttospeech.v1.SsmlVoiceGender" json:"ssml_gender,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *VoiceSelectionParams) Reset() { *m = VoiceSelectionParams{} } +func (m *VoiceSelectionParams) String() string { return proto.CompactTextString(m) } +func (*VoiceSelectionParams) ProtoMessage() {} +func (*VoiceSelectionParams) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{5} +} +func (m *VoiceSelectionParams) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_VoiceSelectionParams.Unmarshal(m, b) +} +func (m *VoiceSelectionParams) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_VoiceSelectionParams.Marshal(b, m, deterministic) +} +func (dst *VoiceSelectionParams) XXX_Merge(src proto.Message) { + xxx_messageInfo_VoiceSelectionParams.Merge(dst, src) +} +func (m *VoiceSelectionParams) XXX_Size() int { + return xxx_messageInfo_VoiceSelectionParams.Size(m) +} +func (m *VoiceSelectionParams) XXX_DiscardUnknown() { + xxx_messageInfo_VoiceSelectionParams.DiscardUnknown(m) +} + +var xxx_messageInfo_VoiceSelectionParams proto.InternalMessageInfo + +func (m *VoiceSelectionParams) GetLanguageCode() string { + if m != nil { + return m.LanguageCode + } + return "" +} + +func (m *VoiceSelectionParams) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *VoiceSelectionParams) GetSsmlGender() SsmlVoiceGender { + if m != nil { + return m.SsmlGender + } + return SsmlVoiceGender_SSML_VOICE_GENDER_UNSPECIFIED +} + +// Description of audio data to be synthesized. +type AudioConfig struct { + // Required. The format of the requested audio byte stream. + AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,proto3,enum=google.cloud.texttospeech.v1.AudioEncoding" json:"audio_encoding,omitempty"` + // Optional speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal + // native speed supported by the specific voice. 2.0 is twice as fast, and + // 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any + // other values < 0.25 or > 4.0 will return an error. + SpeakingRate float64 `protobuf:"fixed64,2,opt,name=speaking_rate,json=speakingRate,proto3" json:"speaking_rate,omitempty"` + // Optional speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 + // semitones from the original pitch. -20 means decrease 20 semitones from the + // original pitch. + Pitch float64 `protobuf:"fixed64,3,opt,name=pitch,proto3" json:"pitch,omitempty"` + // Optional volume gain (in dB) of the normal native volume supported by the + // specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of + // 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) + // will play at approximately half the amplitude of the normal native signal + // amplitude. A value of +6.0 (dB) will play at approximately twice the + // amplitude of the normal native signal amplitude. Strongly recommend not to + // exceed +10 (dB) as there's usually no effective increase in loudness for + // any value greater than that. + VolumeGainDb float64 `protobuf:"fixed64,4,opt,name=volume_gain_db,json=volumeGainDb,proto3" json:"volume_gain_db,omitempty"` + // The synthesis sample rate (in hertz) for this audio. Optional. If this is + // different from the voice's natural sample rate, then the synthesizer will + // honor this request by converting to the desired sample rate (which might + // result in worse audio quality), unless the specified sample rate is not + // supported for the encoding chosen, in which case it will fail the request + // and return [google.rpc.Code.INVALID_ARGUMENT][]. + SampleRateHertz int32 `protobuf:"varint,5,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AudioConfig) Reset() { *m = AudioConfig{} } +func (m *AudioConfig) String() string { return proto.CompactTextString(m) } +func (*AudioConfig) ProtoMessage() {} +func (*AudioConfig) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{6} +} +func (m *AudioConfig) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AudioConfig.Unmarshal(m, b) +} +func (m *AudioConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AudioConfig.Marshal(b, m, deterministic) +} +func (dst *AudioConfig) XXX_Merge(src proto.Message) { + xxx_messageInfo_AudioConfig.Merge(dst, src) +} +func (m *AudioConfig) XXX_Size() int { + return xxx_messageInfo_AudioConfig.Size(m) +} +func (m *AudioConfig) XXX_DiscardUnknown() { + xxx_messageInfo_AudioConfig.DiscardUnknown(m) +} + +var xxx_messageInfo_AudioConfig proto.InternalMessageInfo + +func (m *AudioConfig) GetAudioEncoding() AudioEncoding { + if m != nil { + return m.AudioEncoding + } + return AudioEncoding_AUDIO_ENCODING_UNSPECIFIED +} + +func (m *AudioConfig) GetSpeakingRate() float64 { + if m != nil { + return m.SpeakingRate + } + return 0 +} + +func (m *AudioConfig) GetPitch() float64 { + if m != nil { + return m.Pitch + } + return 0 +} + +func (m *AudioConfig) GetVolumeGainDb() float64 { + if m != nil { + return m.VolumeGainDb + } + return 0 +} + +func (m *AudioConfig) GetSampleRateHertz() int32 { + if m != nil { + return m.SampleRateHertz + } + return 0 +} + +// The message returned to the client by the `SynthesizeSpeech` method. +type SynthesizeSpeechResponse struct { + // The audio data bytes encoded as specified in the request, including the + // header (For LINEAR16 audio, we include the WAV header). Note: as + // with all bytes fields, protobuffers use a pure binary representation, + // whereas JSON representations use base64. + AudioContent []byte `protobuf:"bytes,1,opt,name=audio_content,json=audioContent,proto3" json:"audio_content,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SynthesizeSpeechResponse) Reset() { *m = SynthesizeSpeechResponse{} } +func (m *SynthesizeSpeechResponse) String() string { return proto.CompactTextString(m) } +func (*SynthesizeSpeechResponse) ProtoMessage() {} +func (*SynthesizeSpeechResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_cloud_tts_bc3ee81beb174f4f, []int{7} +} +func (m *SynthesizeSpeechResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SynthesizeSpeechResponse.Unmarshal(m, b) +} +func (m *SynthesizeSpeechResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SynthesizeSpeechResponse.Marshal(b, m, deterministic) +} +func (dst *SynthesizeSpeechResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_SynthesizeSpeechResponse.Merge(dst, src) +} +func (m *SynthesizeSpeechResponse) XXX_Size() int { + return xxx_messageInfo_SynthesizeSpeechResponse.Size(m) +} +func (m *SynthesizeSpeechResponse) XXX_DiscardUnknown() { + xxx_messageInfo_SynthesizeSpeechResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_SynthesizeSpeechResponse proto.InternalMessageInfo + +func (m *SynthesizeSpeechResponse) GetAudioContent() []byte { + if m != nil { + return m.AudioContent + } + return nil +} + +func init() { + proto.RegisterType((*ListVoicesRequest)(nil), "google.cloud.texttospeech.v1.ListVoicesRequest") + proto.RegisterType((*ListVoicesResponse)(nil), "google.cloud.texttospeech.v1.ListVoicesResponse") + proto.RegisterType((*Voice)(nil), "google.cloud.texttospeech.v1.Voice") + proto.RegisterType((*SynthesizeSpeechRequest)(nil), "google.cloud.texttospeech.v1.SynthesizeSpeechRequest") + proto.RegisterType((*SynthesisInput)(nil), "google.cloud.texttospeech.v1.SynthesisInput") + proto.RegisterType((*VoiceSelectionParams)(nil), "google.cloud.texttospeech.v1.VoiceSelectionParams") + proto.RegisterType((*AudioConfig)(nil), "google.cloud.texttospeech.v1.AudioConfig") + proto.RegisterType((*SynthesizeSpeechResponse)(nil), "google.cloud.texttospeech.v1.SynthesizeSpeechResponse") + proto.RegisterEnum("google.cloud.texttospeech.v1.SsmlVoiceGender", SsmlVoiceGender_name, SsmlVoiceGender_value) + proto.RegisterEnum("google.cloud.texttospeech.v1.AudioEncoding", AudioEncoding_name, AudioEncoding_value) +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// TextToSpeechClient is the client API for TextToSpeech service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type TextToSpeechClient interface { + // Returns a list of [Voice][google.cloud.texttospeech.v1.Voice] + // supported for synthesis. + ListVoices(ctx context.Context, in *ListVoicesRequest, opts ...grpc.CallOption) (*ListVoicesResponse, error) + // Synthesizes speech synchronously: receive results after all text input + // has been processed. + SynthesizeSpeech(ctx context.Context, in *SynthesizeSpeechRequest, opts ...grpc.CallOption) (*SynthesizeSpeechResponse, error) +} + +type textToSpeechClient struct { + cc *grpc.ClientConn +} + +func NewTextToSpeechClient(cc *grpc.ClientConn) TextToSpeechClient { + return &textToSpeechClient{cc} +} + +func (c *textToSpeechClient) ListVoices(ctx context.Context, in *ListVoicesRequest, opts ...grpc.CallOption) (*ListVoicesResponse, error) { + out := new(ListVoicesResponse) + err := c.cc.Invoke(ctx, "/google.cloud.texttospeech.v1.TextToSpeech/ListVoices", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *textToSpeechClient) SynthesizeSpeech(ctx context.Context, in *SynthesizeSpeechRequest, opts ...grpc.CallOption) (*SynthesizeSpeechResponse, error) { + out := new(SynthesizeSpeechResponse) + err := c.cc.Invoke(ctx, "/google.cloud.texttospeech.v1.TextToSpeech/SynthesizeSpeech", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// TextToSpeechServer is the server API for TextToSpeech service. +type TextToSpeechServer interface { + // Returns a list of [Voice][google.cloud.texttospeech.v1.Voice] + // supported for synthesis. + ListVoices(context.Context, *ListVoicesRequest) (*ListVoicesResponse, error) + // Synthesizes speech synchronously: receive results after all text input + // has been processed. + SynthesizeSpeech(context.Context, *SynthesizeSpeechRequest) (*SynthesizeSpeechResponse, error) +} + +func RegisterTextToSpeechServer(s *grpc.Server, srv TextToSpeechServer) { + s.RegisterService(&_TextToSpeech_serviceDesc, srv) +} + +func _TextToSpeech_ListVoices_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListVoicesRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(TextToSpeechServer).ListVoices(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.texttospeech.v1.TextToSpeech/ListVoices", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(TextToSpeechServer).ListVoices(ctx, req.(*ListVoicesRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _TextToSpeech_SynthesizeSpeech_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(SynthesizeSpeechRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(TextToSpeechServer).SynthesizeSpeech(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.cloud.texttospeech.v1.TextToSpeech/SynthesizeSpeech", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(TextToSpeechServer).SynthesizeSpeech(ctx, req.(*SynthesizeSpeechRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _TextToSpeech_serviceDesc = grpc.ServiceDesc{ + ServiceName: "google.cloud.texttospeech.v1.TextToSpeech", + HandlerType: (*TextToSpeechServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "ListVoices", + Handler: _TextToSpeech_ListVoices_Handler, + }, + { + MethodName: "SynthesizeSpeech", + Handler: _TextToSpeech_SynthesizeSpeech_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "google/cloud/texttospeech/v1/cloud_tts.proto", +} + +func init() { + proto.RegisterFile("google/cloud/texttospeech/v1/cloud_tts.proto", fileDescriptor_cloud_tts_bc3ee81beb174f4f) +} + +var fileDescriptor_cloud_tts_bc3ee81beb174f4f = []byte{ + // 868 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x55, 0x41, 0x6f, 0x1b, 0x45, + 0x14, 0xee, 0xd8, 0x71, 0xda, 0x3e, 0xaf, 0x5d, 0x67, 0x88, 0xc0, 0x44, 0xa5, 0x32, 0x1b, 0x90, + 0x42, 0x28, 0x36, 0x76, 0x45, 0x05, 0xed, 0x01, 0x39, 0xf6, 0xd6, 0xb1, 0xe4, 0xd8, 0x66, 0x36, + 0x09, 0x12, 0x8a, 0xb4, 0x9a, 0xac, 0x87, 0xcd, 0x8a, 0xf5, 0xcc, 0xe2, 0x19, 0x47, 0xa5, 0x47, + 0xf8, 0x09, 0xfc, 0x00, 0xc4, 0x95, 0x5f, 0xc0, 0x99, 0x23, 0x12, 0x27, 0xfe, 0x02, 0x67, 0x6e, + 0x48, 0x1c, 0xd1, 0xcc, 0xae, 0xdb, 0x8d, 0x13, 0x19, 0x73, 0xe1, 0xb6, 0xf3, 0xde, 0x7c, 0x6f, + 0xbe, 0xf7, 0xcd, 0xb7, 0x6f, 0xe0, 0x61, 0x20, 0x44, 0x10, 0xb1, 0x86, 0x1f, 0x89, 0xf9, 0xa4, + 0xa1, 0xd8, 0x73, 0xa5, 0x84, 0x8c, 0x19, 0xf3, 0x2f, 0x1a, 0x97, 0xcd, 0x24, 0xea, 0x29, 0x25, + 0xeb, 0xf1, 0x4c, 0x28, 0x81, 0xef, 0x27, 0xbb, 0xeb, 0x26, 0x5e, 0xcf, 0xee, 0xae, 0x5f, 0x36, + 0x77, 0xd2, 0x6c, 0x83, 0xc6, 0x61, 0x83, 0x72, 0x2e, 0x14, 0x55, 0xa1, 0xe0, 0x29, 0xd6, 0xfe, + 0x18, 0xb6, 0x06, 0xa1, 0x54, 0xa7, 0x22, 0xf4, 0x99, 0x24, 0xec, 0xeb, 0x39, 0x93, 0x0a, 0xef, + 0x42, 0x29, 0xa2, 0x3c, 0x98, 0xd3, 0x80, 0x79, 0xbe, 0x98, 0xb0, 0x2a, 0xaa, 0xa1, 0xbd, 0xbb, + 0xc4, 0x5a, 0x04, 0x3b, 0x62, 0xc2, 0xec, 0xcf, 0x00, 0x67, 0x91, 0x32, 0x16, 0x5c, 0x32, 0xfc, + 0x14, 0x36, 0x2f, 0x4d, 0xa4, 0x8a, 0x6a, 0xf9, 0xbd, 0x62, 0x6b, 0xb7, 0xbe, 0x8a, 0x5c, 0xdd, + 0xa0, 0x49, 0x0a, 0xb1, 0x7f, 0x43, 0x50, 0x30, 0x11, 0xfc, 0x2e, 0x94, 0xaf, 0x30, 0x48, 0xca, + 0xdd, 0x25, 0xa5, 0x2c, 0x05, 0x89, 0x31, 0x6c, 0x70, 0x3a, 0x65, 0xd5, 0x9c, 0xe1, 0x67, 0xbe, + 0xf1, 0x10, 0x8a, 0x52, 0x4e, 0x23, 0x2f, 0x60, 0x7c, 0xc2, 0x66, 0xd5, 0x7c, 0x0d, 0xed, 0x95, + 0x5b, 0x1f, 0xac, 0xa6, 0xe1, 0xca, 0x69, 0x64, 0x0e, 0xee, 0x19, 0x10, 0x01, 0x5d, 0x21, 0xf9, + 0xc6, 0x9f, 0xc0, 0x9b, 0x9c, 0xaa, 0xf9, 0x8c, 0x46, 0x9e, 0xa4, 0xd3, 0x38, 0x62, 0xde, 0x8c, + 0x2a, 0xe6, 0x5d, 0xb0, 0x99, 0x7a, 0x51, 0xdd, 0xa8, 0xa1, 0xbd, 0x02, 0x79, 0x3d, 0xdd, 0xe0, + 0x9a, 0x3c, 0xa1, 0x8a, 0x1d, 0xea, 0xac, 0xfd, 0x17, 0x82, 0x37, 0xdc, 0x6f, 0xb8, 0xba, 0x60, + 0x32, 0x7c, 0xc1, 0x5c, 0x73, 0xdc, 0x42, 0xe3, 0x03, 0x28, 0x84, 0x3c, 0x9e, 0x2b, 0xa3, 0x6d, + 0xb1, 0xf5, 0xf0, 0x5f, 0x08, 0xa6, 0x55, 0x64, 0x5f, 0x63, 0x48, 0x02, 0xc5, 0x87, 0x50, 0x30, + 0xca, 0x99, 0xfe, 0x8b, 0xad, 0xd6, 0x1a, 0x5a, 0xbb, 0x2c, 0x62, 0xbe, 0x36, 0xc0, 0x98, 0xce, + 0xe8, 0x54, 0x92, 0xa4, 0x00, 0x1e, 0x80, 0x45, 0xe7, 0x93, 0x50, 0x78, 0xbe, 0xe0, 0x5f, 0x86, + 0x81, 0x51, 0xad, 0xd8, 0x7a, 0x6f, 0x75, 0xc1, 0xb6, 0x46, 0x74, 0x0c, 0x80, 0x14, 0xe9, 0xab, + 0x85, 0x3d, 0x80, 0xf2, 0x55, 0xc2, 0x78, 0x1b, 0x36, 0x34, 0x3a, 0x31, 0xd2, 0xe1, 0x2d, 0x62, + 0x56, 0x3a, 0xaa, 0x85, 0x4e, 0xae, 0x4f, 0x47, 0xf5, 0xea, 0xa0, 0x0c, 0x96, 0x69, 0xcf, 0x93, + 0x62, 0x3e, 0xf3, 0x99, 0xfd, 0x03, 0x82, 0xed, 0x9b, 0xb8, 0xaf, 0x65, 0xd3, 0xff, 0xc3, 0x22, + 0xf6, 0x9f, 0x08, 0x8a, 0x19, 0x31, 0x30, 0x81, 0x72, 0xa2, 0x26, 0xe3, 0xbe, 0x98, 0x84, 0x3c, + 0x30, 0xcc, 0xca, 0xad, 0xf7, 0xd7, 0xd0, 0xd3, 0x49, 0x21, 0xa4, 0x44, 0xb3, 0x4b, 0xdd, 0xac, + 0x8c, 0x19, 0xfd, 0x2a, 0xe4, 0x81, 0x31, 0xa0, 0x69, 0x08, 0x11, 0x6b, 0x11, 0xd4, 0xae, 0xc3, + 0xdb, 0x50, 0x88, 0x43, 0xe5, 0x5f, 0x98, 0x96, 0x10, 0x49, 0x16, 0xf8, 0x1d, 0x28, 0x5f, 0x8a, + 0x68, 0x3e, 0x65, 0x5e, 0x40, 0x43, 0xee, 0x4d, 0xce, 0x8d, 0x6d, 0x11, 0xb1, 0x92, 0x68, 0x8f, + 0x86, 0xbc, 0x7b, 0x8e, 0xf7, 0x61, 0xeb, 0xba, 0xbf, 0x0b, 0xc6, 0xdf, 0xf7, 0xe4, 0x92, 0xb1, + 0x3f, 0x85, 0xea, 0x75, 0x5f, 0xa7, 0x13, 0x60, 0x17, 0x4a, 0x2f, 0xad, 0xa4, 0x18, 0x4f, 0xee, + 0xdc, 0x22, 0xd6, 0xc2, 0x20, 0x3a, 0xb6, 0xff, 0x39, 0xdc, 0x5b, 0x12, 0x14, 0xbf, 0x0d, 0x6f, + 0xb9, 0xee, 0xd1, 0xc0, 0x3b, 0x1d, 0xf5, 0x3b, 0x8e, 0xd7, 0x73, 0x86, 0x5d, 0x87, 0x78, 0x27, + 0x43, 0x77, 0xec, 0x74, 0xfa, 0xcf, 0xfa, 0x4e, 0xb7, 0x72, 0x0b, 0xdf, 0x81, 0x8d, 0xa3, 0xf6, + 0xc0, 0xa9, 0x20, 0x0c, 0xb0, 0xf9, 0xcc, 0x31, 0xdf, 0x39, 0x5c, 0x84, 0xdb, 0x43, 0xe7, 0xe4, + 0x98, 0xb4, 0x07, 0x95, 0xfc, 0xfe, 0x31, 0x94, 0xae, 0xc8, 0x88, 0x1f, 0xc0, 0x4e, 0xfb, 0xa4, + 0xdb, 0x1f, 0x79, 0xce, 0xb0, 0x33, 0xea, 0xf6, 0x87, 0xbd, 0xa5, 0x9a, 0x16, 0xdc, 0x19, 0xf4, + 0x87, 0x4e, 0x9b, 0x34, 0x1f, 0x57, 0x10, 0xbe, 0x0d, 0xf9, 0xa3, 0xf1, 0xa3, 0x4a, 0x4e, 0x87, + 0x47, 0xbd, 0x9e, 0x37, 0x1a, 0x9f, 0xb8, 0x95, 0x7c, 0xeb, 0xe7, 0x1c, 0x58, 0xc7, 0xec, 0xb9, + 0x3a, 0x16, 0x49, 0xb3, 0xf8, 0x3b, 0x04, 0xf0, 0x6a, 0xfa, 0xe1, 0xc6, 0xea, 0x8b, 0xbd, 0x36, + 0x61, 0x77, 0x3e, 0x5c, 0x1f, 0x90, 0xc8, 0x6a, 0xe3, 0x6f, 0x7f, 0xff, 0xe3, 0xfb, 0x9c, 0x85, + 0x41, 0xbf, 0x00, 0xc9, 0xbc, 0xc4, 0x3f, 0x22, 0xa8, 0x2c, 0xdf, 0x03, 0xfe, 0x68, 0xbd, 0x49, + 0xb2, 0x34, 0x8f, 0x76, 0x1e, 0xff, 0x57, 0x58, 0xca, 0xeb, 0x81, 0xe1, 0x55, 0xb5, 0x5f, 0xd3, + 0xbc, 0x34, 0xea, 0x89, 0x7c, 0xb9, 0xf5, 0x09, 0xda, 0x3f, 0xf8, 0x05, 0x41, 0xcd, 0x17, 0xd3, + 0x95, 0xd5, 0x0f, 0xb6, 0xb2, 0xe2, 0x8e, 0xf5, 0xc3, 0x34, 0x46, 0x5f, 0x1c, 0xa6, 0x90, 0x40, + 0xe8, 0x1f, 0xba, 0x2e, 0x66, 0x41, 0x23, 0x60, 0xdc, 0x3c, 0x5b, 0x8d, 0x24, 0x45, 0xe3, 0x50, + 0xde, 0xfc, 0x46, 0x3e, 0xcd, 0xae, 0xff, 0x46, 0xe8, 0xa7, 0xdc, 0xfd, 0x5e, 0x52, 0xad, 0x63, + 0x08, 0x64, 0xcf, 0xab, 0x9f, 0x36, 0x7f, 0x5d, 0xa4, 0xcf, 0x4c, 0xfa, 0x2c, 0x9b, 0x3e, 0x3b, + 0x6d, 0x9e, 0x6f, 0x9a, 0x53, 0x1f, 0xfd, 0x13, 0x00, 0x00, 0xff, 0xff, 0x38, 0xd7, 0xcf, 0x2c, + 0x98, 0x07, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1beta1/cloud_tts.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1beta1/cloud_tts.pb.go index bdfb5b678..72c97b1de 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1beta1/cloud_tts.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/texttospeech/v1beta1/cloud_tts.pb.go @@ -60,7 +60,7 @@ func (x SsmlVoiceGender) String() string { return proto.EnumName(SsmlVoiceGender_name, int32(x)) } func (SsmlVoiceGender) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{0} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{0} } // Configuration to set up audio encoder. The encoding determines the output @@ -99,7 +99,7 @@ func (x AudioEncoding) String() string { return proto.EnumName(AudioEncoding_name, int32(x)) } func (AudioEncoding) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{1} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{1} } // The top-level message sent by the client for the `ListVoices` method. @@ -112,7 +112,7 @@ type ListVoicesRequest struct { // "no-*" (Norwegian) and "nb-*" (Norwegian Bokmal) voices; specifying "zh" // will also get supported "cmn-*" voices; specifying "zh-hk" will also get // supported "yue-*" voices. - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -122,7 +122,7 @@ func (m *ListVoicesRequest) Reset() { *m = ListVoicesRequest{} } func (m *ListVoicesRequest) String() string { return proto.CompactTextString(m) } func (*ListVoicesRequest) ProtoMessage() {} func (*ListVoicesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{0} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{0} } func (m *ListVoicesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListVoicesRequest.Unmarshal(m, b) @@ -152,7 +152,7 @@ func (m *ListVoicesRequest) GetLanguageCode() string { // The message returned to the client by the `ListVoices` method. type ListVoicesResponse struct { // The list of voices. - Voices []*Voice `protobuf:"bytes,1,rep,name=voices" json:"voices,omitempty"` + Voices []*Voice `protobuf:"bytes,1,rep,name=voices,proto3" json:"voices,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -162,7 +162,7 @@ func (m *ListVoicesResponse) Reset() { *m = ListVoicesResponse{} } func (m *ListVoicesResponse) String() string { return proto.CompactTextString(m) } func (*ListVoicesResponse) ProtoMessage() {} func (*ListVoicesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{1} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{1} } func (m *ListVoicesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListVoicesResponse.Unmarshal(m, b) @@ -194,13 +194,13 @@ type Voice struct { // The languages that this voice supports, expressed as // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags (e.g. // "en-US", "es-419", "cmn-tw"). - LanguageCodes []string `protobuf:"bytes,1,rep,name=language_codes,json=languageCodes" json:"language_codes,omitempty"` + LanguageCodes []string `protobuf:"bytes,1,rep,name=language_codes,json=languageCodes,proto3" json:"language_codes,omitempty"` // The name of this voice. Each distinct voice has a unique name. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // The gender of this voice. - SsmlGender SsmlVoiceGender `protobuf:"varint,3,opt,name=ssml_gender,json=ssmlGender,enum=google.cloud.texttospeech.v1beta1.SsmlVoiceGender" json:"ssml_gender,omitempty"` + SsmlGender SsmlVoiceGender `protobuf:"varint,3,opt,name=ssml_gender,json=ssmlGender,proto3,enum=google.cloud.texttospeech.v1beta1.SsmlVoiceGender" json:"ssml_gender,omitempty"` // The natural sample rate (in hertz) for this voice. - NaturalSampleRateHertz int32 `protobuf:"varint,4,opt,name=natural_sample_rate_hertz,json=naturalSampleRateHertz" json:"natural_sample_rate_hertz,omitempty"` + NaturalSampleRateHertz int32 `protobuf:"varint,4,opt,name=natural_sample_rate_hertz,json=naturalSampleRateHertz,proto3" json:"natural_sample_rate_hertz,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -210,7 +210,7 @@ func (m *Voice) Reset() { *m = Voice{} } func (m *Voice) String() string { return proto.CompactTextString(m) } func (*Voice) ProtoMessage() {} func (*Voice) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{2} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{2} } func (m *Voice) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Voice.Unmarshal(m, b) @@ -261,11 +261,11 @@ func (m *Voice) GetNaturalSampleRateHertz() int32 { // The top-level message sent by the client for the `SynthesizeSpeech` method. type SynthesizeSpeechRequest struct { // Required. The Synthesizer requires either plain text or SSML as input. - Input *SynthesisInput `protobuf:"bytes,1,opt,name=input" json:"input,omitempty"` + Input *SynthesisInput `protobuf:"bytes,1,opt,name=input,proto3" json:"input,omitempty"` // Required. The desired voice of the synthesized audio. - Voice *VoiceSelectionParams `protobuf:"bytes,2,opt,name=voice" json:"voice,omitempty"` + Voice *VoiceSelectionParams `protobuf:"bytes,2,opt,name=voice,proto3" json:"voice,omitempty"` // Required. The configuration of the synthesized audio. - AudioConfig *AudioConfig `protobuf:"bytes,3,opt,name=audio_config,json=audioConfig" json:"audio_config,omitempty"` + AudioConfig *AudioConfig `protobuf:"bytes,3,opt,name=audio_config,json=audioConfig,proto3" json:"audio_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -275,7 +275,7 @@ func (m *SynthesizeSpeechRequest) Reset() { *m = SynthesizeSpeechRequest func (m *SynthesizeSpeechRequest) String() string { return proto.CompactTextString(m) } func (*SynthesizeSpeechRequest) ProtoMessage() {} func (*SynthesizeSpeechRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{3} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{3} } func (m *SynthesizeSpeechRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SynthesizeSpeechRequest.Unmarshal(m, b) @@ -336,7 +336,7 @@ func (m *SynthesisInput) Reset() { *m = SynthesisInput{} } func (m *SynthesisInput) String() string { return proto.CompactTextString(m) } func (*SynthesisInput) ProtoMessage() {} func (*SynthesisInput) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{4} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{4} } func (m *SynthesisInput) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SynthesisInput.Unmarshal(m, b) @@ -361,10 +361,10 @@ type isSynthesisInput_InputSource interface { } type SynthesisInput_Text struct { - Text string `protobuf:"bytes,1,opt,name=text,oneof"` + Text string `protobuf:"bytes,1,opt,name=text,proto3,oneof"` } type SynthesisInput_Ssml struct { - Ssml string `protobuf:"bytes,2,opt,name=ssml,oneof"` + Ssml string `protobuf:"bytes,2,opt,name=ssml,proto3,oneof"` } func (*SynthesisInput_Text) isSynthesisInput_InputSource() {} @@ -470,16 +470,16 @@ type VoiceSelectionParams struct { // (e.g. using en-US rather than en-CA if there isn't a Canadian voice // available), or even a different language, e.g. using "nb" (Norwegian // Bokmal) instead of "no" (Norwegian)". - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // The name of the voice. Optional; if not set, the service will choose a // voice based on the other parameters such as language_code and gender. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // The preferred gender of the voice. Optional; if not set, the service will // choose a voice based on the other parameters such as language_code and // name. Note that this is only a preference, not requirement; if a // voice of the appropriate gender is not available, the synthesizer should // substitute a voice with a different gender rather than failing the request. - SsmlGender SsmlVoiceGender `protobuf:"varint,3,opt,name=ssml_gender,json=ssmlGender,enum=google.cloud.texttospeech.v1beta1.SsmlVoiceGender" json:"ssml_gender,omitempty"` + SsmlGender SsmlVoiceGender `protobuf:"varint,3,opt,name=ssml_gender,json=ssmlGender,proto3,enum=google.cloud.texttospeech.v1beta1.SsmlVoiceGender" json:"ssml_gender,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -489,7 +489,7 @@ func (m *VoiceSelectionParams) Reset() { *m = VoiceSelectionParams{} } func (m *VoiceSelectionParams) String() string { return proto.CompactTextString(m) } func (*VoiceSelectionParams) ProtoMessage() {} func (*VoiceSelectionParams) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{5} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{5} } func (m *VoiceSelectionParams) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_VoiceSelectionParams.Unmarshal(m, b) @@ -533,16 +533,16 @@ func (m *VoiceSelectionParams) GetSsmlGender() SsmlVoiceGender { // Description of audio data to be synthesized. type AudioConfig struct { // Required. The format of the requested audio byte stream. - AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,enum=google.cloud.texttospeech.v1beta1.AudioEncoding" json:"audio_encoding,omitempty"` + AudioEncoding AudioEncoding `protobuf:"varint,1,opt,name=audio_encoding,json=audioEncoding,proto3,enum=google.cloud.texttospeech.v1beta1.AudioEncoding" json:"audio_encoding,omitempty"` // Optional speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal // native speed supported by the specific voice. 2.0 is twice as fast, and // 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any // other values < 0.25 or > 4.0 will return an error. - SpeakingRate float64 `protobuf:"fixed64,2,opt,name=speaking_rate,json=speakingRate" json:"speaking_rate,omitempty"` + SpeakingRate float64 `protobuf:"fixed64,2,opt,name=speaking_rate,json=speakingRate,proto3" json:"speaking_rate,omitempty"` // Optional speaking pitch, in the range [-20.0, 20.0]. 20 means increase 20 // semitones from the original pitch. -20 means decrease 20 semitones from the // original pitch. - Pitch float64 `protobuf:"fixed64,3,opt,name=pitch" json:"pitch,omitempty"` + Pitch float64 `protobuf:"fixed64,3,opt,name=pitch,proto3" json:"pitch,omitempty"` // Optional volume gain (in dB) of the normal native volume supported by the // specific voice, in the range [-96.0, 16.0]. If unset, or set to a value of // 0.0 (dB), will play at normal native signal amplitude. A value of -6.0 (dB) @@ -551,14 +551,18 @@ type AudioConfig struct { // amplitude of the normal native signal amplitude. Strongly recommend not to // exceed +10 (dB) as there's usually no effective increase in loudness for // any value greater than that. - VolumeGainDb float64 `protobuf:"fixed64,4,opt,name=volume_gain_db,json=volumeGainDb" json:"volume_gain_db,omitempty"` + VolumeGainDb float64 `protobuf:"fixed64,4,opt,name=volume_gain_db,json=volumeGainDb,proto3" json:"volume_gain_db,omitempty"` // The synthesis sample rate (in hertz) for this audio. Optional. If this is // different from the voice's natural sample rate, then the synthesizer will // honor this request by converting to the desired sample rate (which might // result in worse audio quality), unless the specified sample rate is not // supported for the encoding chosen, in which case it will fail the request // and return [google.rpc.Code.INVALID_ARGUMENT][]. - SampleRateHertz int32 `protobuf:"varint,5,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` + SampleRateHertz int32 `protobuf:"varint,5,opt,name=sample_rate_hertz,json=sampleRateHertz,proto3" json:"sample_rate_hertz,omitempty"` + // An identifier which selects 'audio effects' profiles that are applied on + // (post synthesized) text to speech. + // Effects are applied on top of each other in the order they are given. + EffectsProfileId []string `protobuf:"bytes,6,rep,name=effects_profile_id,json=effectsProfileId,proto3" json:"effects_profile_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -568,7 +572,7 @@ func (m *AudioConfig) Reset() { *m = AudioConfig{} } func (m *AudioConfig) String() string { return proto.CompactTextString(m) } func (*AudioConfig) ProtoMessage() {} func (*AudioConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{6} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{6} } func (m *AudioConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AudioConfig.Unmarshal(m, b) @@ -623,6 +627,13 @@ func (m *AudioConfig) GetSampleRateHertz() int32 { return 0 } +func (m *AudioConfig) GetEffectsProfileId() []string { + if m != nil { + return m.EffectsProfileId + } + return nil +} + // The message returned to the client by the `SynthesizeSpeech` method. type SynthesizeSpeechResponse struct { // The audio data bytes encoded as specified in the request, including the @@ -639,7 +650,7 @@ func (m *SynthesizeSpeechResponse) Reset() { *m = SynthesizeSpeechRespon func (m *SynthesizeSpeechResponse) String() string { return proto.CompactTextString(m) } func (*SynthesizeSpeechResponse) ProtoMessage() {} func (*SynthesizeSpeechResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_cloud_tts_48487c666911cd00, []int{7} + return fileDescriptor_cloud_tts_b74274d69cd9534d, []int{7} } func (m *SynthesizeSpeechResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SynthesizeSpeechResponse.Unmarshal(m, b) @@ -725,8 +736,7 @@ func (c *textToSpeechClient) SynthesizeSpeech(ctx context.Context, in *Synthesiz return out, nil } -// Server API for TextToSpeech service - +// TextToSpeechServer is the server API for TextToSpeech service. type TextToSpeechServer interface { // Returns a list of [Voice][google.cloud.texttospeech.v1beta1.Voice] // supported for synthesis. @@ -794,62 +804,64 @@ var _TextToSpeech_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/cloud/texttospeech/v1beta1/cloud_tts.proto", fileDescriptor_cloud_tts_48487c666911cd00) + proto.RegisterFile("google/cloud/texttospeech/v1beta1/cloud_tts.proto", fileDescriptor_cloud_tts_b74274d69cd9534d) } -var fileDescriptor_cloud_tts_48487c666911cd00 = []byte{ - // 846 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x55, 0x4d, 0x6f, 0x1b, 0x45, - 0x18, 0xee, 0xd8, 0x71, 0xda, 0xbe, 0x5e, 0x3b, 0xce, 0x28, 0xa2, 0x26, 0xa2, 0x28, 0xdd, 0x50, - 0xc9, 0xca, 0xc1, 0xc6, 0x2e, 0x9f, 0xe9, 0x01, 0x1c, 0x7b, 0xeb, 0x5a, 0xf2, 0x17, 0xb3, 0x49, - 0x2a, 0x71, 0x59, 0x4d, 0xd6, 0xc3, 0x66, 0xc5, 0x7a, 0x66, 0xf1, 0x8c, 0xa3, 0xd2, 0x23, 0xe2, - 0xcc, 0x01, 0xfe, 0x02, 0x3f, 0x80, 0xdf, 0x02, 0x12, 0xbf, 0x80, 0x7f, 0xc0, 0x85, 0x23, 0x9a, - 0x99, 0x4d, 0xea, 0x38, 0x88, 0x3a, 0x1c, 0xb8, 0xed, 0x3c, 0xe3, 0xe7, 0x9d, 0xe7, 0x7d, 0xe6, - 0xf1, 0x3b, 0xd0, 0x8c, 0x84, 0x88, 0x12, 0xd6, 0x08, 0x13, 0xb1, 0x98, 0x36, 0x14, 0x7b, 0xa9, - 0x94, 0x90, 0x29, 0x63, 0xe1, 0x79, 0xe3, 0xa2, 0x79, 0xc6, 0x14, 0x6d, 0xda, 0xad, 0x40, 0x29, - 0x59, 0x4f, 0xe7, 0x42, 0x09, 0xfc, 0xc8, 0x52, 0xea, 0x06, 0xaf, 0x2f, 0x53, 0xea, 0x19, 0x65, - 0xf7, 0x9d, 0xac, 0x2a, 0x4d, 0xe3, 0x06, 0xe5, 0x5c, 0x28, 0xaa, 0x62, 0xc1, 0xb3, 0x02, 0xee, - 0x27, 0xb0, 0x3d, 0x88, 0xa5, 0x3a, 0x15, 0x71, 0xc8, 0x24, 0x61, 0xdf, 0x2c, 0x98, 0x54, 0x78, - 0x1f, 0x4a, 0x09, 0xe5, 0xd1, 0x82, 0x46, 0x2c, 0x08, 0xc5, 0x94, 0x55, 0xd1, 0x1e, 0xaa, 0xdd, - 0x27, 0xce, 0x25, 0xd8, 0x11, 0x53, 0xe6, 0x9e, 0x02, 0x5e, 0x66, 0xca, 0x54, 0x70, 0xc9, 0xf0, - 0xe7, 0xb0, 0x79, 0x61, 0x90, 0x2a, 0xda, 0xcb, 0xd7, 0x8a, 0xad, 0x5a, 0xfd, 0x8d, 0x0a, 0xeb, - 0xa6, 0x04, 0xc9, 0x78, 0xee, 0xaf, 0x08, 0x0a, 0x06, 0xc1, 0x8f, 0xa1, 0x7c, 0x4d, 0x86, 0xad, - 0x79, 0x9f, 0x94, 0x96, 0x75, 0x48, 0x8c, 0x61, 0x83, 0xd3, 0x19, 0xab, 0xe6, 0x8c, 0x48, 0xf3, - 0x8d, 0x7d, 0x28, 0x4a, 0x39, 0x4b, 0x82, 0x88, 0xf1, 0x29, 0x9b, 0x57, 0xf3, 0x7b, 0xa8, 0x56, - 0x6e, 0xb5, 0xd6, 0xd0, 0xe2, 0xcb, 0x59, 0x62, 0x4e, 0xef, 0x19, 0x26, 0x01, 0x5d, 0xc6, 0x7e, - 0xe3, 0x4f, 0xe1, 0x6d, 0x4e, 0xd5, 0x62, 0x4e, 0x93, 0x40, 0xd2, 0x59, 0x9a, 0xb0, 0x60, 0x4e, - 0x15, 0x0b, 0xce, 0xd9, 0x5c, 0xbd, 0xaa, 0x6e, 0xec, 0xa1, 0x5a, 0x81, 0xbc, 0x95, 0xfd, 0xc0, - 0x37, 0xfb, 0x84, 0x2a, 0xf6, 0x5c, 0xef, 0xba, 0xdf, 0xe7, 0xe0, 0x81, 0xff, 0x2d, 0x57, 0xe7, - 0x4c, 0xc6, 0xaf, 0x98, 0x6f, 0xce, 0xbc, 0x74, 0xbb, 0x07, 0x85, 0x98, 0xa7, 0x0b, 0x65, 0x5c, - 0x2e, 0xb6, 0x9a, 0xeb, 0xa8, 0xcc, 0x4a, 0xc9, 0xbe, 0x26, 0x12, 0xcb, 0xc7, 0x43, 0x28, 0x18, - 0x0f, 0x8d, 0x13, 0xc5, 0xd6, 0xc7, 0xeb, 0x5a, 0xef, 0xb3, 0x84, 0x85, 0x3a, 0x14, 0x13, 0x3a, - 0xa7, 0x33, 0x49, 0x6c, 0x15, 0xfc, 0x05, 0x38, 0x74, 0x31, 0x8d, 0x45, 0x10, 0x0a, 0xfe, 0x55, - 0x1c, 0x19, 0x13, 0x8b, 0xad, 0xfa, 0x1a, 0x55, 0xdb, 0x9a, 0xd6, 0x31, 0x2c, 0x52, 0xa4, 0xaf, - 0x17, 0xee, 0x00, 0xca, 0xd7, 0xa5, 0xe3, 0x1d, 0xd8, 0xd0, 0x25, 0x6c, 0xc2, 0x9e, 0xdf, 0x21, - 0x66, 0xa5, 0x51, 0xed, 0xbb, 0xbd, 0x52, 0x8d, 0xea, 0xd5, 0x51, 0x19, 0x1c, 0xd3, 0x68, 0x20, - 0xc5, 0x62, 0x1e, 0x32, 0xf7, 0x67, 0x04, 0x3b, 0xff, 0xd4, 0xc0, 0x5a, 0xf9, 0xfd, 0xdf, 0x62, - 0xe3, 0xfe, 0x89, 0xa0, 0xb8, 0xe4, 0x08, 0x7e, 0x01, 0x65, 0xeb, 0x2b, 0xe3, 0xa1, 0x98, 0xc6, - 0x3c, 0x32, 0xf2, 0xca, 0xad, 0xf7, 0xd7, 0x75, 0xd6, 0xcb, 0x78, 0xa4, 0x44, 0x97, 0x97, 0xba, - 0x6d, 0x99, 0x32, 0xfa, 0x75, 0xcc, 0x23, 0x93, 0x4c, 0xd3, 0x1a, 0x22, 0xce, 0x25, 0xa8, 0xe3, - 0x88, 0x77, 0xa0, 0x90, 0xc6, 0x2a, 0x3c, 0x37, 0xcd, 0x21, 0x62, 0x17, 0xf8, 0x3d, 0x28, 0x5f, - 0x88, 0x64, 0x31, 0x63, 0x41, 0x44, 0x63, 0x1e, 0x4c, 0xcf, 0x4c, 0x9e, 0x11, 0x71, 0x2c, 0xda, - 0xa3, 0x31, 0xef, 0x9e, 0xe1, 0x03, 0xd8, 0xbe, 0x19, 0xfc, 0x82, 0x09, 0xfe, 0x96, 0x5c, 0x49, - 0xfc, 0x67, 0x50, 0xbd, 0x19, 0xf8, 0x6c, 0x48, 0xec, 0x43, 0xe9, 0x2a, 0x59, 0x8a, 0x71, 0x7b, - 0xfb, 0x0e, 0x71, 0x2e, 0xa3, 0xa2, 0xb1, 0x83, 0x17, 0xb0, 0xb5, 0xe2, 0x2a, 0x7e, 0x04, 0x0f, - 0x7d, 0x7f, 0x38, 0x08, 0x4e, 0xc7, 0xfd, 0x8e, 0x17, 0xf4, 0xbc, 0x51, 0xd7, 0x23, 0xc1, 0xc9, - 0xc8, 0x9f, 0x78, 0x9d, 0xfe, 0xb3, 0xbe, 0xd7, 0xad, 0xdc, 0xc1, 0xf7, 0x60, 0x63, 0xd8, 0x1e, - 0x78, 0x15, 0x84, 0x01, 0x36, 0x9f, 0x79, 0xe6, 0x3b, 0x87, 0x8b, 0x70, 0x77, 0xe4, 0x9d, 0x1c, - 0x93, 0xf6, 0xa0, 0x92, 0x3f, 0x38, 0x86, 0xd2, 0x35, 0x1b, 0xf1, 0xbb, 0xb0, 0xdb, 0x3e, 0xe9, - 0xf6, 0xc7, 0x81, 0x37, 0xea, 0x8c, 0xbb, 0xfd, 0x51, 0x6f, 0xa5, 0xa6, 0x03, 0xf7, 0x06, 0xfd, - 0x91, 0xd7, 0x26, 0xcd, 0x8f, 0x2a, 0x08, 0xdf, 0x85, 0xfc, 0x70, 0xf2, 0xa4, 0x92, 0xd3, 0xf0, - 0xb8, 0xd7, 0x0b, 0xc6, 0x93, 0x13, 0xbf, 0x92, 0x6f, 0xfd, 0x9e, 0x03, 0xe7, 0x98, 0xbd, 0x54, - 0xc7, 0xc2, 0x36, 0x8b, 0x7f, 0x44, 0x00, 0xaf, 0x07, 0x24, 0xfe, 0x60, 0x8d, 0xdb, 0xbd, 0x31, - 0x89, 0x77, 0x3f, 0xbc, 0x25, 0xcb, 0x1a, 0xec, 0x3e, 0xf8, 0xee, 0xb7, 0x3f, 0x7e, 0xca, 0x6d, - 0xe3, 0xad, 0xab, 0x87, 0xc3, 0x0e, 0x57, 0xfc, 0x0b, 0x82, 0xca, 0xea, 0xb5, 0xe0, 0xc3, 0x5b, - 0x4c, 0x9c, 0x95, 0xe1, 0xb5, 0xfb, 0xf4, 0x3f, 0x71, 0x33, 0x99, 0xfb, 0x46, 0xe6, 0x43, 0xb7, - 0x7a, 0x25, 0x53, 0xf3, 0x0f, 0xe5, 0xd5, 0xef, 0x0f, 0xd1, 0xc1, 0xd1, 0x0f, 0x08, 0x1e, 0x87, - 0x62, 0xf6, 0xe6, 0x73, 0x8e, 0xb6, 0x97, 0xfd, 0x9f, 0xe8, 0xe7, 0x6d, 0x82, 0xbe, 0x1c, 0x66, - 0xbc, 0x48, 0xe8, 0x7f, 0x7f, 0x5d, 0xcc, 0xa3, 0x46, 0xc4, 0xb8, 0x79, 0xfc, 0x1a, 0x76, 0x8b, - 0xa6, 0xb1, 0xfc, 0x97, 0x37, 0xf7, 0xe9, 0x32, 0xf8, 0x17, 0x42, 0x67, 0x9b, 0x86, 0xfc, 0xe4, - 0xef, 0x00, 0x00, 0x00, 0xff, 0xff, 0x09, 0xb8, 0x07, 0xd3, 0xaf, 0x07, 0x00, 0x00, +var fileDescriptor_cloud_tts_b74274d69cd9534d = []byte{ + // 875 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x55, 0xdd, 0x6e, 0x1b, 0x45, + 0x14, 0xee, 0xd8, 0x71, 0xda, 0x1e, 0xaf, 0x1d, 0x67, 0x14, 0x51, 0x13, 0x51, 0x94, 0x6e, 0xa8, + 0x64, 0x45, 0xc8, 0xc6, 0x2e, 0xbf, 0xe9, 0x05, 0x38, 0xf6, 0xd6, 0xb5, 0xe4, 0x3f, 0x66, 0x93, + 0x54, 0xe2, 0x66, 0x35, 0xd9, 0x9d, 0x6c, 0x56, 0xac, 0x67, 0x96, 0x9d, 0x71, 0x54, 0x7a, 0x89, + 0xb8, 0xe6, 0x02, 0x9e, 0x00, 0x89, 0x07, 0xe0, 0x59, 0x40, 0xe2, 0x09, 0x78, 0x08, 0x2e, 0xd1, + 0xce, 0x6c, 0x52, 0x27, 0x41, 0xd4, 0xe1, 0x82, 0xbb, 0x99, 0x73, 0xf6, 0x3b, 0xf3, 0x9d, 0x6f, + 0xbe, 0x3d, 0x03, 0xed, 0x50, 0x88, 0x30, 0x66, 0x2d, 0x3f, 0x16, 0x8b, 0xa0, 0xa5, 0xd8, 0x4b, + 0xa5, 0x84, 0x4c, 0x18, 0xf3, 0xcf, 0x5a, 0xe7, 0xed, 0x13, 0xa6, 0x68, 0xdb, 0xa4, 0x3c, 0xa5, + 0x64, 0x33, 0x49, 0x85, 0x12, 0xf8, 0x91, 0x81, 0x34, 0x75, 0xbc, 0xb9, 0x0c, 0x69, 0xe6, 0x90, + 0xed, 0x77, 0xf2, 0xaa, 0x34, 0x89, 0x5a, 0x94, 0x73, 0xa1, 0xa8, 0x8a, 0x04, 0xcf, 0x0b, 0xd8, + 0x9f, 0xc2, 0xe6, 0x28, 0x92, 0xea, 0x58, 0x44, 0x3e, 0x93, 0x84, 0x7d, 0xb3, 0x60, 0x52, 0xe1, + 0x5d, 0xa8, 0xc4, 0x94, 0x87, 0x0b, 0x1a, 0x32, 0xcf, 0x17, 0x01, 0xab, 0xa3, 0x1d, 0xd4, 0xb8, + 0x4f, 0xac, 0x8b, 0x60, 0x4f, 0x04, 0xcc, 0x3e, 0x06, 0xbc, 0x8c, 0x94, 0x89, 0xe0, 0x92, 0xe1, + 0x2f, 0x60, 0xfd, 0x5c, 0x47, 0xea, 0x68, 0xa7, 0xd8, 0x28, 0x77, 0x1a, 0xcd, 0x37, 0x32, 0x6c, + 0xea, 0x12, 0x24, 0xc7, 0xd9, 0xbf, 0x21, 0x28, 0xe9, 0x08, 0x7e, 0x0c, 0xd5, 0x2b, 0x34, 0x4c, + 0xcd, 0xfb, 0xa4, 0xb2, 0xcc, 0x43, 0x62, 0x0c, 0x6b, 0x9c, 0xce, 0x59, 0xbd, 0xa0, 0x49, 0xea, + 0x35, 0x76, 0xa1, 0x2c, 0xe5, 0x3c, 0xf6, 0x42, 0xc6, 0x03, 0x96, 0xd6, 0x8b, 0x3b, 0xa8, 0x51, + 0xed, 0x74, 0x56, 0xe0, 0xe2, 0xca, 0x79, 0xac, 0x4f, 0x1f, 0x68, 0x24, 0x81, 0xac, 0x8c, 0x59, + 0xe3, 0xcf, 0xe0, 0x6d, 0x4e, 0xd5, 0x22, 0xa5, 0xb1, 0x27, 0xe9, 0x3c, 0x89, 0x99, 0x97, 0x52, + 0xc5, 0xbc, 0x33, 0x96, 0xaa, 0x57, 0xf5, 0xb5, 0x1d, 0xd4, 0x28, 0x91, 0xb7, 0xf2, 0x0f, 0x5c, + 0x9d, 0x27, 0x54, 0xb1, 0xe7, 0x59, 0xd6, 0xfe, 0xbe, 0x00, 0x0f, 0xdc, 0x6f, 0xb9, 0x3a, 0x63, + 0x32, 0x7a, 0xc5, 0x5c, 0x7d, 0xe6, 0x85, 0xda, 0x03, 0x28, 0x45, 0x3c, 0x59, 0x28, 0xad, 0x72, + 0xb9, 0xd3, 0x5e, 0x85, 0x65, 0x5e, 0x4a, 0x0e, 0x33, 0x20, 0x31, 0x78, 0x3c, 0x86, 0x92, 0xd6, + 0x50, 0x2b, 0x51, 0xee, 0x7c, 0xb2, 0xaa, 0xf4, 0x2e, 0x8b, 0x99, 0x9f, 0x99, 0x62, 0x46, 0x53, + 0x3a, 0x97, 0xc4, 0x54, 0xc1, 0x5f, 0x82, 0x45, 0x17, 0x41, 0x24, 0x3c, 0x5f, 0xf0, 0xd3, 0x28, + 0xd4, 0x22, 0x96, 0x3b, 0xcd, 0x15, 0xaa, 0x76, 0x33, 0x58, 0x4f, 0xa3, 0x48, 0x99, 0xbe, 0xde, + 0xd8, 0x23, 0xa8, 0x5e, 0xa5, 0x8e, 0xb7, 0x60, 0x2d, 0x2b, 0x61, 0x1c, 0xf6, 0xfc, 0x0e, 0xd1, + 0xbb, 0x2c, 0x9a, 0xe9, 0x6e, 0xae, 0x34, 0x8b, 0x66, 0xbb, 0x83, 0x2a, 0x58, 0xba, 0x51, 0x4f, + 0x8a, 0x45, 0xea, 0x33, 0xfb, 0x17, 0x04, 0x5b, 0xff, 0xd4, 0xc0, 0x4a, 0xfe, 0xfd, 0xdf, 0x6c, + 0x63, 0xff, 0x5c, 0x80, 0xf2, 0x92, 0x22, 0xf8, 0x05, 0x54, 0x8d, 0xae, 0x8c, 0xfb, 0x22, 0x88, + 0x78, 0xa8, 0xe9, 0x55, 0x3b, 0x1f, 0xac, 0xaa, 0xac, 0x93, 0xe3, 0x48, 0x85, 0x2e, 0x6f, 0xb3, + 0xb6, 0x65, 0xc2, 0xe8, 0xd7, 0x11, 0x0f, 0xb5, 0x33, 0x75, 0x6b, 0x88, 0x58, 0x17, 0xc1, 0xcc, + 0x8e, 0x78, 0x0b, 0x4a, 0x49, 0xa4, 0xfc, 0x33, 0xdd, 0x1c, 0x22, 0x66, 0x83, 0xdf, 0x83, 0xea, + 0xb9, 0x88, 0x17, 0x73, 0xe6, 0x85, 0x34, 0xe2, 0x5e, 0x70, 0xa2, 0xfd, 0x8c, 0x88, 0x65, 0xa2, + 0x03, 0x1a, 0xf1, 0xfe, 0x09, 0xde, 0x83, 0xcd, 0x9b, 0xc6, 0x2f, 0x69, 0xe3, 0x6f, 0xc8, 0xab, + 0x8e, 0xc7, 0xef, 0x03, 0x66, 0xa7, 0xa7, 0xcc, 0x57, 0xd2, 0x4b, 0x52, 0x71, 0x1a, 0xc5, 0xcc, + 0x8b, 0x82, 0xfa, 0xba, 0xfe, 0x81, 0x6b, 0x79, 0x66, 0x66, 0x12, 0xc3, 0xc0, 0xfe, 0x1c, 0xea, + 0x37, 0x7f, 0x8f, 0x7c, 0xa4, 0xec, 0x42, 0xe5, 0xd2, 0x87, 0x8a, 0x71, 0xe3, 0x15, 0x8b, 0x58, + 0x17, 0xc6, 0xca, 0x62, 0x7b, 0x2f, 0x60, 0xe3, 0xda, 0x1d, 0xe0, 0x47, 0xf0, 0xd0, 0x75, 0xc7, + 0x23, 0xef, 0x78, 0x3a, 0xec, 0x39, 0xde, 0xc0, 0x99, 0xf4, 0x1d, 0xe2, 0x1d, 0x4d, 0xdc, 0x99, + 0xd3, 0x1b, 0x3e, 0x1b, 0x3a, 0xfd, 0xda, 0x1d, 0x7c, 0x0f, 0xd6, 0xc6, 0xdd, 0x91, 0x53, 0x43, + 0x18, 0x60, 0xfd, 0x99, 0xa3, 0xd7, 0x05, 0x5c, 0x86, 0xbb, 0x13, 0xe7, 0xe8, 0x90, 0x74, 0x47, + 0xb5, 0xe2, 0xde, 0x21, 0x54, 0xae, 0x88, 0x8e, 0xdf, 0x85, 0xed, 0xee, 0x51, 0x7f, 0x38, 0xf5, + 0x9c, 0x49, 0x6f, 0xda, 0x1f, 0x4e, 0x06, 0xd7, 0x6a, 0x5a, 0x70, 0x6f, 0x34, 0x9c, 0x38, 0x5d, + 0xd2, 0xfe, 0xb8, 0x86, 0xf0, 0x5d, 0x28, 0x8e, 0x67, 0x4f, 0x6a, 0x85, 0x2c, 0x3c, 0x1d, 0x0c, + 0xbc, 0xe9, 0xec, 0xc8, 0xad, 0x15, 0x3b, 0x7f, 0x14, 0xc0, 0x3a, 0x64, 0x2f, 0xd5, 0xa1, 0x30, + 0xcd, 0xe2, 0x1f, 0x11, 0xc0, 0xeb, 0x71, 0x8a, 0x3f, 0x5c, 0xc1, 0x0b, 0x37, 0xe6, 0xf6, 0xf6, + 0x47, 0xb7, 0x44, 0x19, 0x81, 0xed, 0x07, 0xdf, 0xfd, 0xfe, 0xe7, 0x4f, 0x85, 0x4d, 0xbc, 0x71, + 0xf9, 0xcc, 0x98, 0x51, 0x8c, 0x7f, 0x45, 0x50, 0xbb, 0x7e, 0x2d, 0x78, 0xff, 0x16, 0xf3, 0xe9, + 0xda, 0xa8, 0xdb, 0x7e, 0xfa, 0x9f, 0xb0, 0x39, 0xcd, 0x5d, 0x4d, 0xf3, 0xa1, 0x5d, 0xbf, 0xa4, + 0x99, 0xe1, 0xf7, 0xe5, 0xe5, 0xf7, 0xfb, 0x68, 0xef, 0xe0, 0x07, 0x04, 0x8f, 0x7d, 0x31, 0x7f, + 0xf3, 0x39, 0x07, 0x9b, 0xcb, 0xfa, 0xcf, 0xb2, 0xc7, 0x70, 0x86, 0xbe, 0x1a, 0xe7, 0xb8, 0x50, + 0x64, 0xb3, 0xa2, 0x29, 0xd2, 0xb0, 0x15, 0x32, 0xae, 0x9f, 0xca, 0x96, 0x49, 0xd1, 0x24, 0x92, + 0xff, 0xf2, 0x42, 0x3f, 0x5d, 0x0e, 0xfe, 0x85, 0xd0, 0xc9, 0xba, 0x06, 0x3f, 0xf9, 0x3b, 0x00, + 0x00, 0xff, 0xff, 0x26, 0x28, 0xff, 0x42, 0xdd, 0x07, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1/video_intelligence.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1/video_intelligence.pb.go index 0b0294165..df963ce5f 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1/video_intelligence.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1/video_intelligence.pb.go @@ -154,26 +154,26 @@ type AnnotateVideoRequest struct { // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded // in the request as `input_content`. If set, `input_content` should be unset. - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // The video data bytes. // If unset, the input video(s) should be specified via `input_uri`. // If set, `input_uri` should be unset. InputContent []byte `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"` // Requested video annotation features. - Features []Feature `protobuf:"varint,2,rep,packed,name=features,enum=google.cloud.videointelligence.v1.Feature" json:"features,omitempty"` + Features []Feature `protobuf:"varint,2,rep,packed,name=features,proto3,enum=google.cloud.videointelligence.v1.Feature" json:"features,omitempty"` // Additional video context and/or feature-specific parameters. - VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext" json:"video_context,omitempty"` + VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext,proto3" json:"video_context,omitempty"` // Optional location where the output (in JSON format) should be stored. // Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see // [Request URIs](/storage/docs/reference-uris). - OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"` + OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` // Optional cloud region where annotation should take place. Supported cloud // regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region // is specified, a region will be determined based on video file location. - LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -250,15 +250,15 @@ type VideoContext struct { // Video segments to annotate. The segments may overlap and are not required // to be contiguous or span the whole video. If unspecified, each video // is treated as a single segment. - Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments" json:"segments,omitempty"` + Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments,proto3" json:"segments,omitempty"` // Config for LABEL_DETECTION. - LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig" json:"label_detection_config,omitempty"` + LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig,proto3" json:"label_detection_config,omitempty"` // Config for SHOT_CHANGE_DETECTION. - ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig" json:"shot_change_detection_config,omitempty"` + ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig,proto3" json:"shot_change_detection_config,omitempty"` // Config for EXPLICIT_CONTENT_DETECTION. - ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig" json:"explicit_content_detection_config,omitempty"` + ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig,proto3" json:"explicit_content_detection_config,omitempty"` // Config for FACE_DETECTION. - FaceDetectionConfig *FaceDetectionConfig `protobuf:"bytes,5,opt,name=face_detection_config,json=faceDetectionConfig" json:"face_detection_config,omitempty"` + FaceDetectionConfig *FaceDetectionConfig `protobuf:"bytes,5,opt,name=face_detection_config,json=faceDetectionConfig,proto3" json:"face_detection_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -328,15 +328,15 @@ type LabelDetectionConfig struct { // What labels should be detected with LABEL_DETECTION, in addition to // video-level labels or segment-level labels. // If unspecified, defaults to `SHOT_MODE`. - LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,enum=google.cloud.videointelligence.v1.LabelDetectionMode" json:"label_detection_mode,omitempty"` + LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,proto3,enum=google.cloud.videointelligence.v1.LabelDetectionMode" json:"label_detection_mode,omitempty"` // Whether the video has been shot from a stationary (i.e. non-moving) camera. // When set to true, might improve detection accuracy for moving objects. // Should be used with `SHOT_AND_FRAME_MODE` enabled. - StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera" json:"stationary_camera,omitempty"` + StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera,proto3" json:"stationary_camera,omitempty"` // Model to use for label detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -392,7 +392,7 @@ type ShotChangeDetectionConfig struct { // Model to use for shot change detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -434,7 +434,7 @@ type ExplicitContentDetectionConfig struct { // Model to use for explicit content detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -476,9 +476,9 @@ type FaceDetectionConfig struct { // Model to use for face detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` // Whether bounding boxes be included in the face annotation output. - IncludeBoundingBoxes bool `protobuf:"varint,2,opt,name=include_bounding_boxes,json=includeBoundingBoxes" json:"include_bounding_boxes,omitempty"` + IncludeBoundingBoxes bool `protobuf:"varint,2,opt,name=include_bounding_boxes,json=includeBoundingBoxes,proto3" json:"include_bounding_boxes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -526,10 +526,10 @@ func (m *FaceDetectionConfig) GetIncludeBoundingBoxes() bool { type VideoSegment struct { // Time-offset, relative to the beginning of the video, // corresponding to the start of the segment (inclusive). - StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset" json:"start_time_offset,omitempty"` + StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset,proto3" json:"start_time_offset,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the end of the segment (inclusive). - EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset" json:"end_time_offset,omitempty"` + EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset,proto3" json:"end_time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -576,9 +576,9 @@ func (m *VideoSegment) GetEndTimeOffset() *duration.Duration { // Video segment level annotation results for label detection. type LabelSegment struct { // Video segment where a label was detected. - Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"` + Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -626,9 +626,9 @@ func (m *LabelSegment) GetConfidence() float32 { type LabelFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -677,11 +677,11 @@ type Entity struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search // API](https://developers.google.com/knowledge-graph/). - EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Textual description, e.g. `Fixed-gear bicycle`. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Language code for `description` in BCP-47 format. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -735,16 +735,16 @@ func (m *Entity) GetLanguageCode() string { // Label annotation. type LabelAnnotation struct { // Detected entity. - Entity *Entity `protobuf:"bytes,1,opt,name=entity" json:"entity,omitempty"` + Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"` // Common categories for the detected entity. // E.g. when the label is `Terrier` the category is likely `dog`. And in some // cases there might be more than one categories e.g. `Terrier` could also be // a `pet`. - CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities" json:"category_entities,omitempty"` + CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities,proto3" json:"category_entities,omitempty"` // All video segments where a label was detected. - Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments" json:"segments,omitempty"` + Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a label was detected. - Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames" json:"frames,omitempty"` + Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -806,9 +806,9 @@ func (m *LabelAnnotation) GetFrames() []*LabelFrame { type ExplicitContentFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Likelihood of the pornography content.. - PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,enum=google.cloud.videointelligence.v1.Likelihood" json:"pornography_likelihood,omitempty"` + PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,proto3,enum=google.cloud.videointelligence.v1.Likelihood" json:"pornography_likelihood,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -857,7 +857,7 @@ func (m *ExplicitContentFrame) GetPornographyLikelihood() Likelihood { // present for that frame. type ExplicitContentAnnotation struct { // All video frames where explicit content was detected. - Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames" json:"frames,omitempty"` + Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -899,13 +899,13 @@ func (m *ExplicitContentAnnotation) GetFrames() []*ExplicitContentFrame { // Range: [0, 1]. type NormalizedBoundingBox struct { // Left X coordinate. - Left float32 `protobuf:"fixed32,1,opt,name=left" json:"left,omitempty"` + Left float32 `protobuf:"fixed32,1,opt,name=left,proto3" json:"left,omitempty"` // Top Y coordinate. - Top float32 `protobuf:"fixed32,2,opt,name=top" json:"top,omitempty"` + Top float32 `protobuf:"fixed32,2,opt,name=top,proto3" json:"top,omitempty"` // Right X coordinate. - Right float32 `protobuf:"fixed32,3,opt,name=right" json:"right,omitempty"` + Right float32 `protobuf:"fixed32,3,opt,name=right,proto3" json:"right,omitempty"` // Bottom Y coordinate. - Bottom float32 `protobuf:"fixed32,4,opt,name=bottom" json:"bottom,omitempty"` + Bottom float32 `protobuf:"fixed32,4,opt,name=bottom,proto3" json:"bottom,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -966,7 +966,7 @@ func (m *NormalizedBoundingBox) GetBottom() float32 { // Video segment level annotation results for face detection. type FaceSegment struct { // Video segment where a face was detected. - Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"` + Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1008,10 +1008,10 @@ type FaceFrame struct { // Normalized Bounding boxes in a frame. // There can be more than one boxes if the same face is detected in multiple // locations within the current frame. - NormalizedBoundingBoxes []*NormalizedBoundingBox `protobuf:"bytes,1,rep,name=normalized_bounding_boxes,json=normalizedBoundingBoxes" json:"normalized_bounding_boxes,omitempty"` + NormalizedBoundingBoxes []*NormalizedBoundingBox `protobuf:"bytes,1,rep,name=normalized_bounding_boxes,json=normalizedBoundingBoxes,proto3" json:"normalized_bounding_boxes,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1060,9 +1060,9 @@ type FaceAnnotation struct { // Thumbnail of a representative face view (in JPEG format). Thumbnail []byte `protobuf:"bytes,1,opt,name=thumbnail,proto3" json:"thumbnail,omitempty"` // All video segments where a face was detected. - Segments []*FaceSegment `protobuf:"bytes,2,rep,name=segments" json:"segments,omitempty"` + Segments []*FaceSegment `protobuf:"bytes,2,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a face was detected. - Frames []*FaceFrame `protobuf:"bytes,3,rep,name=frames" json:"frames,omitempty"` + Frames []*FaceFrame `protobuf:"bytes,3,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1117,25 +1117,25 @@ func (m *FaceAnnotation) GetFrames() []*FaceFrame { type VideoAnnotationResults struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Label annotations on video level or user specified segment level. // There is exactly one element for each unique label. - SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations" json:"segment_label_annotations,omitempty"` + SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations,proto3" json:"segment_label_annotations,omitempty"` // Label annotations on shot level. // There is exactly one element for each unique label. - ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations" json:"shot_label_annotations,omitempty"` + ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations,proto3" json:"shot_label_annotations,omitempty"` // Label annotations on frame level. // There is exactly one element for each unique label. - FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations" json:"frame_label_annotations,omitempty"` + FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations,proto3" json:"frame_label_annotations,omitempty"` // Face annotations. There is exactly one element for each unique face. - FaceAnnotations []*FaceAnnotation `protobuf:"bytes,5,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"` + FaceAnnotations []*FaceAnnotation `protobuf:"bytes,5,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // Shot annotations. Each shot is represented as a video segment. - ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations" json:"shot_annotations,omitempty"` + ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations,proto3" json:"shot_annotations,omitempty"` // Explicit content annotation. - ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation" json:"explicit_annotation,omitempty"` + ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation,proto3" json:"explicit_annotation,omitempty"` // If set, indicates an error. Note that for a single `AnnotateVideoRequest` // some videos may succeed and some may fail. - Error *status.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1226,7 +1226,7 @@ func (m *VideoAnnotationResults) GetError() *status.Status { // call of the `google::longrunning::Operations` service. type AnnotateVideoResponse struct { // Annotation results for all videos specified in `AnnotateVideoRequest`. - AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults" json:"annotation_results,omitempty"` + AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults,proto3" json:"annotation_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1267,14 +1267,14 @@ func (m *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults type VideoAnnotationProgress struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Approximate percentage processed thus far. // Guaranteed to be 100 when fully processed. - ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent update. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1337,7 +1337,7 @@ func (m *VideoAnnotationProgress) GetUpdateTime() *timestamp.Timestamp { // call of the `google::longrunning::Operations` service. type AnnotateVideoProgress struct { // Progress metadata for all videos specified in `AnnotateVideoRequest`. - AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress" json:"annotation_progress,omitempty"` + AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress,proto3" json:"annotation_progress,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1437,8 +1437,7 @@ func (c *videoIntelligenceServiceClient) AnnotateVideo(ctx context.Context, in * return out, nil } -// Server API for VideoIntelligenceService service - +// VideoIntelligenceServiceServer is the server API for VideoIntelligenceService service. type VideoIntelligenceServiceServer interface { // Performs asynchronous video annotation. Progress and results can be // retrieved through the `google.longrunning.Operations` interface. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta1/video_intelligence.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta1/video_intelligence.pb.go index dc11605a9..7a0595b9a 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta1/video_intelligence.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta1/video_intelligence.pb.go @@ -192,25 +192,25 @@ type AnnotateVideoRequest struct { // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded // in the request as `input_content`. If set, `input_content` should be unset. - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // The video data bytes. Encoding: base64. If unset, the input video(s) // should be specified via `input_uri`. If set, `input_uri` should be unset. - InputContent string `protobuf:"bytes,6,opt,name=input_content,json=inputContent" json:"input_content,omitempty"` + InputContent string `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"` // Requested video annotation features. - Features []Feature `protobuf:"varint,2,rep,packed,name=features,enum=google.cloud.videointelligence.v1beta1.Feature" json:"features,omitempty"` + Features []Feature `protobuf:"varint,2,rep,packed,name=features,proto3,enum=google.cloud.videointelligence.v1beta1.Feature" json:"features,omitempty"` // Additional video context and/or feature-specific parameters. - VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext" json:"video_context,omitempty"` + VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext,proto3" json:"video_context,omitempty"` // Optional location where the output (in JSON format) should be stored. // Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see // [Request URIs](/storage/docs/reference-uris). - OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"` + OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` // Optional cloud region where annotation should take place. Supported cloud // regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region // is specified, a region will be determined based on video file location. - LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -287,26 +287,26 @@ type VideoContext struct { // Video segments to annotate. The segments may overlap and are not required // to be contiguous or span the whole video. If unspecified, each video // is treated as a single segment. - Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments" json:"segments,omitempty"` + Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments,proto3" json:"segments,omitempty"` // If label detection has been requested, what labels should be detected // in addition to video-level labels or segment-level labels. If unspecified, // defaults to `SHOT_MODE`. - LabelDetectionMode LabelDetectionMode `protobuf:"varint,2,opt,name=label_detection_mode,json=labelDetectionMode,enum=google.cloud.videointelligence.v1beta1.LabelDetectionMode" json:"label_detection_mode,omitempty"` + LabelDetectionMode LabelDetectionMode `protobuf:"varint,2,opt,name=label_detection_mode,json=labelDetectionMode,proto3,enum=google.cloud.videointelligence.v1beta1.LabelDetectionMode" json:"label_detection_mode,omitempty"` // Whether the video has been shot from a stationary (i.e. non-moving) camera. // When set to true, might improve detection accuracy for moving objects. - StationaryCamera bool `protobuf:"varint,3,opt,name=stationary_camera,json=stationaryCamera" json:"stationary_camera,omitempty"` + StationaryCamera bool `protobuf:"varint,3,opt,name=stationary_camera,json=stationaryCamera,proto3" json:"stationary_camera,omitempty"` // Model to use for label detection. // Supported values: "latest" and "stable" (the default). - LabelDetectionModel string `protobuf:"bytes,4,opt,name=label_detection_model,json=labelDetectionModel" json:"label_detection_model,omitempty"` + LabelDetectionModel string `protobuf:"bytes,4,opt,name=label_detection_model,json=labelDetectionModel,proto3" json:"label_detection_model,omitempty"` // Model to use for face detection. // Supported values: "latest" and "stable" (the default). - FaceDetectionModel string `protobuf:"bytes,5,opt,name=face_detection_model,json=faceDetectionModel" json:"face_detection_model,omitempty"` + FaceDetectionModel string `protobuf:"bytes,5,opt,name=face_detection_model,json=faceDetectionModel,proto3" json:"face_detection_model,omitempty"` // Model to use for shot change detection. // Supported values: "latest" and "stable" (the default). - ShotChangeDetectionModel string `protobuf:"bytes,6,opt,name=shot_change_detection_model,json=shotChangeDetectionModel" json:"shot_change_detection_model,omitempty"` + ShotChangeDetectionModel string `protobuf:"bytes,6,opt,name=shot_change_detection_model,json=shotChangeDetectionModel,proto3" json:"shot_change_detection_model,omitempty"` // Model to use for safe search detection. // Supported values: "latest" and "stable" (the default). - SafeSearchDetectionModel string `protobuf:"bytes,7,opt,name=safe_search_detection_model,json=safeSearchDetectionModel" json:"safe_search_detection_model,omitempty"` + SafeSearchDetectionModel string `protobuf:"bytes,7,opt,name=safe_search_detection_model,json=safeSearchDetectionModel,proto3" json:"safe_search_detection_model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -388,9 +388,9 @@ func (m *VideoContext) GetSafeSearchDetectionModel() string { // Video segment. type VideoSegment struct { // Start offset in microseconds (inclusive). Unset means 0. - StartTimeOffset int64 `protobuf:"varint,1,opt,name=start_time_offset,json=startTimeOffset" json:"start_time_offset,omitempty"` + StartTimeOffset int64 `protobuf:"varint,1,opt,name=start_time_offset,json=startTimeOffset,proto3" json:"start_time_offset,omitempty"` // End offset in microseconds (inclusive). Unset means 0. - EndTimeOffset int64 `protobuf:"varint,2,opt,name=end_time_offset,json=endTimeOffset" json:"end_time_offset,omitempty"` + EndTimeOffset int64 `protobuf:"varint,2,opt,name=end_time_offset,json=endTimeOffset,proto3" json:"end_time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -440,11 +440,11 @@ type LabelLocation struct { // Set to [timestamp, timestamp] for frame-level labels. // Otherwise, corresponds to one of `AnnotateSpec.segments` // (if specified) or to shot boundaries (if requested). - Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"` + Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // Label level. - Level LabelLevel `protobuf:"varint,3,opt,name=level,enum=google.cloud.videointelligence.v1beta1.LabelLevel" json:"level,omitempty"` + Level LabelLevel `protobuf:"varint,3,opt,name=level,proto3,enum=google.cloud.videointelligence.v1beta1.LabelLevel" json:"level,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -498,11 +498,11 @@ func (m *LabelLocation) GetLevel() LabelLevel { // Label annotation. type LabelAnnotation struct { // Textual description, e.g. `Fixed-gear bicycle`. - Description string `protobuf:"bytes,1,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` // Language code for `description` in BCP-47 format. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Where the label was detected and with what confidence. - Locations []*LabelLocation `protobuf:"bytes,3,rep,name=locations" json:"locations,omitempty"` + Locations []*LabelLocation `protobuf:"bytes,3,rep,name=locations,proto3" json:"locations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -560,18 +560,18 @@ func (m *LabelAnnotation) GetLocations() []*LabelLocation { // for all other types of unsafe content. type SafeSearchAnnotation struct { // Likelihood of adult content. - Adult Likelihood `protobuf:"varint,1,opt,name=adult,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"adult,omitempty"` + Adult Likelihood `protobuf:"varint,1,opt,name=adult,proto3,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"adult,omitempty"` // Likelihood that an obvious modification was made to the original // version to make it appear funny or offensive. - Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"spoof,omitempty"` + Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,proto3,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"spoof,omitempty"` // Likelihood of medical content. - Medical Likelihood `protobuf:"varint,3,opt,name=medical,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"medical,omitempty"` + Medical Likelihood `protobuf:"varint,3,opt,name=medical,proto3,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"medical,omitempty"` // Likelihood of violent content. - Violent Likelihood `protobuf:"varint,4,opt,name=violent,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"violent,omitempty"` + Violent Likelihood `protobuf:"varint,4,opt,name=violent,proto3,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"violent,omitempty"` // Likelihood of racy content. - Racy Likelihood `protobuf:"varint,5,opt,name=racy,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"racy,omitempty"` + Racy Likelihood `protobuf:"varint,5,opt,name=racy,proto3,enum=google.cloud.videointelligence.v1beta1.Likelihood" json:"racy,omitempty"` // Video time offset in microseconds. - TimeOffset int64 `protobuf:"varint,6,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset int64 `protobuf:"varint,6,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -646,13 +646,13 @@ func (m *SafeSearchAnnotation) GetTimeOffset() int64 { // Bounding box. type BoundingBox struct { // Left X coordinate. - Left int32 `protobuf:"varint,1,opt,name=left" json:"left,omitempty"` + Left int32 `protobuf:"varint,1,opt,name=left,proto3" json:"left,omitempty"` // Right X coordinate. - Right int32 `protobuf:"varint,2,opt,name=right" json:"right,omitempty"` + Right int32 `protobuf:"varint,2,opt,name=right,proto3" json:"right,omitempty"` // Bottom Y coordinate. - Bottom int32 `protobuf:"varint,3,opt,name=bottom" json:"bottom,omitempty"` + Bottom int32 `protobuf:"varint,3,opt,name=bottom,proto3" json:"bottom,omitempty"` // Top Y coordinate. - Top int32 `protobuf:"varint,4,opt,name=top" json:"top,omitempty"` + Top int32 `protobuf:"varint,4,opt,name=top,proto3" json:"top,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -713,9 +713,9 @@ func (m *BoundingBox) GetTop() int32 { // Face location. type FaceLocation struct { // Bounding box in a frame. - BoundingBox *BoundingBox `protobuf:"bytes,1,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingBox `protobuf:"bytes,1,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // Video time offset in microseconds. - TimeOffset int64 `protobuf:"varint,2,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset int64 `protobuf:"varint,2,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -762,13 +762,13 @@ func (m *FaceLocation) GetTimeOffset() int64 { // Face annotation. type FaceAnnotation struct { // Thumbnail of a representative face view (in JPEG format). Encoding: base64. - Thumbnail string `protobuf:"bytes,1,opt,name=thumbnail" json:"thumbnail,omitempty"` + Thumbnail string `protobuf:"bytes,1,opt,name=thumbnail,proto3" json:"thumbnail,omitempty"` // All locations where a face was detected. // Faces are detected and tracked on a per-video basis // (as opposed to across multiple videos). - Segments []*VideoSegment `protobuf:"bytes,2,rep,name=segments" json:"segments,omitempty"` + Segments []*VideoSegment `protobuf:"bytes,2,rep,name=segments,proto3" json:"segments,omitempty"` // Face locations at one frame per second. - Locations []*FaceLocation `protobuf:"bytes,3,rep,name=locations" json:"locations,omitempty"` + Locations []*FaceLocation `protobuf:"bytes,3,rep,name=locations,proto3" json:"locations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -823,18 +823,18 @@ func (m *FaceAnnotation) GetLocations() []*FaceLocation { type VideoAnnotationResults struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Label annotations. There is exactly one element for each unique label. - LabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=label_annotations,json=labelAnnotations" json:"label_annotations,omitempty"` + LabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=label_annotations,json=labelAnnotations,proto3" json:"label_annotations,omitempty"` // Face annotations. There is exactly one element for each unique face. - FaceAnnotations []*FaceAnnotation `protobuf:"bytes,3,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"` + FaceAnnotations []*FaceAnnotation `protobuf:"bytes,3,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // Shot annotations. Each shot is represented as a video segment. - ShotAnnotations []*VideoSegment `protobuf:"bytes,4,rep,name=shot_annotations,json=shotAnnotations" json:"shot_annotations,omitempty"` + ShotAnnotations []*VideoSegment `protobuf:"bytes,4,rep,name=shot_annotations,json=shotAnnotations,proto3" json:"shot_annotations,omitempty"` // Safe search annotations. - SafeSearchAnnotations []*SafeSearchAnnotation `protobuf:"bytes,6,rep,name=safe_search_annotations,json=safeSearchAnnotations" json:"safe_search_annotations,omitempty"` + SafeSearchAnnotations []*SafeSearchAnnotation `protobuf:"bytes,6,rep,name=safe_search_annotations,json=safeSearchAnnotations,proto3" json:"safe_search_annotations,omitempty"` // If set, indicates an error. Note that for a single `AnnotateVideoRequest` // some videos may succeed and some may fail. - Error *status.Status `protobuf:"bytes,5,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,5,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -911,7 +911,7 @@ func (m *VideoAnnotationResults) GetError() *status.Status { // call of the `google::longrunning::Operations` service. type AnnotateVideoResponse struct { // Annotation results for all videos specified in `AnnotateVideoRequest`. - AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults" json:"annotation_results,omitempty"` + AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults,proto3" json:"annotation_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -952,14 +952,14 @@ func (m *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults type VideoAnnotationProgress struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Approximate percentage processed thus far. // Guaranteed to be 100 when fully processed. - ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent update. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1022,7 +1022,7 @@ func (m *VideoAnnotationProgress) GetUpdateTime() *timestamp.Timestamp { // call of the `google::longrunning::Operations` service. type AnnotateVideoProgress struct { // Progress metadata for all videos specified in `AnnotateVideoRequest`. - AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress" json:"annotation_progress,omitempty"` + AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress,proto3" json:"annotation_progress,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1115,8 +1115,7 @@ func (c *videoIntelligenceServiceClient) AnnotateVideo(ctx context.Context, in * return out, nil } -// Server API for VideoIntelligenceService service - +// VideoIntelligenceServiceServer is the server API for VideoIntelligenceService service. type VideoIntelligenceServiceServer interface { // Performs asynchronous video annotation. Progress and results can be // retrieved through the `google.longrunning.Operations` interface. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta2/video_intelligence.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta2/video_intelligence.pb.go index f6cc1fa03..ef4d619cf 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta2/video_intelligence.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta2/video_intelligence.pb.go @@ -154,26 +154,26 @@ type AnnotateVideoRequest struct { // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded // in the request as `input_content`. If set, `input_content` should be unset. - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // The video data bytes. // If unset, the input video(s) should be specified via `input_uri`. // If set, `input_uri` should be unset. InputContent []byte `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"` // Requested video annotation features. - Features []Feature `protobuf:"varint,2,rep,packed,name=features,enum=google.cloud.videointelligence.v1beta2.Feature" json:"features,omitempty"` + Features []Feature `protobuf:"varint,2,rep,packed,name=features,proto3,enum=google.cloud.videointelligence.v1beta2.Feature" json:"features,omitempty"` // Additional video context and/or feature-specific parameters. - VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext" json:"video_context,omitempty"` + VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext,proto3" json:"video_context,omitempty"` // Optional location where the output (in JSON format) should be stored. // Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see // [Request URIs](/storage/docs/reference-uris). - OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"` + OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` // Optional cloud region where annotation should take place. Supported cloud // regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region // is specified, a region will be determined based on video file location. - LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -250,15 +250,15 @@ type VideoContext struct { // Video segments to annotate. The segments may overlap and are not required // to be contiguous or span the whole video. If unspecified, each video // is treated as a single segment. - Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments" json:"segments,omitempty"` + Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments,proto3" json:"segments,omitempty"` // Config for LABEL_DETECTION. - LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig" json:"label_detection_config,omitempty"` + LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig,proto3" json:"label_detection_config,omitempty"` // Config for SHOT_CHANGE_DETECTION. - ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig" json:"shot_change_detection_config,omitempty"` + ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig,proto3" json:"shot_change_detection_config,omitempty"` // Config for EXPLICIT_CONTENT_DETECTION. - ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig" json:"explicit_content_detection_config,omitempty"` + ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig,proto3" json:"explicit_content_detection_config,omitempty"` // Config for FACE_DETECTION. - FaceDetectionConfig *FaceDetectionConfig `protobuf:"bytes,5,opt,name=face_detection_config,json=faceDetectionConfig" json:"face_detection_config,omitempty"` + FaceDetectionConfig *FaceDetectionConfig `protobuf:"bytes,5,opt,name=face_detection_config,json=faceDetectionConfig,proto3" json:"face_detection_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -328,15 +328,15 @@ type LabelDetectionConfig struct { // What labels should be detected with LABEL_DETECTION, in addition to // video-level labels or segment-level labels. // If unspecified, defaults to `SHOT_MODE`. - LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,enum=google.cloud.videointelligence.v1beta2.LabelDetectionMode" json:"label_detection_mode,omitempty"` + LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,proto3,enum=google.cloud.videointelligence.v1beta2.LabelDetectionMode" json:"label_detection_mode,omitempty"` // Whether the video has been shot from a stationary (i.e. non-moving) camera. // When set to true, might improve detection accuracy for moving objects. // Should be used with `SHOT_AND_FRAME_MODE` enabled. - StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera" json:"stationary_camera,omitempty"` + StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera,proto3" json:"stationary_camera,omitempty"` // Model to use for label detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -392,7 +392,7 @@ type ShotChangeDetectionConfig struct { // Model to use for shot change detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -434,7 +434,7 @@ type ExplicitContentDetectionConfig struct { // Model to use for explicit content detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -476,9 +476,9 @@ type FaceDetectionConfig struct { // Model to use for face detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` // Whether bounding boxes be included in the face annotation output. - IncludeBoundingBoxes bool `protobuf:"varint,2,opt,name=include_bounding_boxes,json=includeBoundingBoxes" json:"include_bounding_boxes,omitempty"` + IncludeBoundingBoxes bool `protobuf:"varint,2,opt,name=include_bounding_boxes,json=includeBoundingBoxes,proto3" json:"include_bounding_boxes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -526,10 +526,10 @@ func (m *FaceDetectionConfig) GetIncludeBoundingBoxes() bool { type VideoSegment struct { // Time-offset, relative to the beginning of the video, // corresponding to the start of the segment (inclusive). - StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset" json:"start_time_offset,omitempty"` + StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset,proto3" json:"start_time_offset,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the end of the segment (inclusive). - EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset" json:"end_time_offset,omitempty"` + EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset,proto3" json:"end_time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -576,9 +576,9 @@ func (m *VideoSegment) GetEndTimeOffset() *duration.Duration { // Video segment level annotation results for label detection. type LabelSegment struct { // Video segment where a label was detected. - Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"` + Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -626,9 +626,9 @@ func (m *LabelSegment) GetConfidence() float32 { type LabelFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -677,11 +677,11 @@ type Entity struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search // API](https://developers.google.com/knowledge-graph/). - EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Textual description, e.g. `Fixed-gear bicycle`. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Language code for `description` in BCP-47 format. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -735,16 +735,16 @@ func (m *Entity) GetLanguageCode() string { // Label annotation. type LabelAnnotation struct { // Detected entity. - Entity *Entity `protobuf:"bytes,1,opt,name=entity" json:"entity,omitempty"` + Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"` // Common categories for the detected entity. // E.g. when the label is `Terrier` the category is likely `dog`. And in some // cases there might be more than one categories e.g. `Terrier` could also be // a `pet`. - CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities" json:"category_entities,omitempty"` + CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities,proto3" json:"category_entities,omitempty"` // All video segments where a label was detected. - Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments" json:"segments,omitempty"` + Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a label was detected. - Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames" json:"frames,omitempty"` + Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -806,9 +806,9 @@ func (m *LabelAnnotation) GetFrames() []*LabelFrame { type ExplicitContentFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Likelihood of the pornography content.. - PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,enum=google.cloud.videointelligence.v1beta2.Likelihood" json:"pornography_likelihood,omitempty"` + PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,proto3,enum=google.cloud.videointelligence.v1beta2.Likelihood" json:"pornography_likelihood,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -857,7 +857,7 @@ func (m *ExplicitContentFrame) GetPornographyLikelihood() Likelihood { // present for that frame. type ExplicitContentAnnotation struct { // All video frames where explicit content was detected. - Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames" json:"frames,omitempty"` + Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -899,13 +899,13 @@ func (m *ExplicitContentAnnotation) GetFrames() []*ExplicitContentFrame { // Range: [0, 1]. type NormalizedBoundingBox struct { // Left X coordinate. - Left float32 `protobuf:"fixed32,1,opt,name=left" json:"left,omitempty"` + Left float32 `protobuf:"fixed32,1,opt,name=left,proto3" json:"left,omitempty"` // Top Y coordinate. - Top float32 `protobuf:"fixed32,2,opt,name=top" json:"top,omitempty"` + Top float32 `protobuf:"fixed32,2,opt,name=top,proto3" json:"top,omitempty"` // Right X coordinate. - Right float32 `protobuf:"fixed32,3,opt,name=right" json:"right,omitempty"` + Right float32 `protobuf:"fixed32,3,opt,name=right,proto3" json:"right,omitempty"` // Bottom Y coordinate. - Bottom float32 `protobuf:"fixed32,4,opt,name=bottom" json:"bottom,omitempty"` + Bottom float32 `protobuf:"fixed32,4,opt,name=bottom,proto3" json:"bottom,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -966,7 +966,7 @@ func (m *NormalizedBoundingBox) GetBottom() float32 { // Video segment level annotation results for face detection. type FaceSegment struct { // Video segment where a face was detected. - Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"` + Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1008,10 +1008,10 @@ type FaceFrame struct { // Normalized Bounding boxes in a frame. // There can be more than one boxes if the same face is detected in multiple // locations within the current frame. - NormalizedBoundingBoxes []*NormalizedBoundingBox `protobuf:"bytes,1,rep,name=normalized_bounding_boxes,json=normalizedBoundingBoxes" json:"normalized_bounding_boxes,omitempty"` + NormalizedBoundingBoxes []*NormalizedBoundingBox `protobuf:"bytes,1,rep,name=normalized_bounding_boxes,json=normalizedBoundingBoxes,proto3" json:"normalized_bounding_boxes,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1060,9 +1060,9 @@ type FaceAnnotation struct { // Thumbnail of a representative face view (in JPEG format). Thumbnail []byte `protobuf:"bytes,1,opt,name=thumbnail,proto3" json:"thumbnail,omitempty"` // All video segments where a face was detected. - Segments []*FaceSegment `protobuf:"bytes,2,rep,name=segments" json:"segments,omitempty"` + Segments []*FaceSegment `protobuf:"bytes,2,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a face was detected. - Frames []*FaceFrame `protobuf:"bytes,3,rep,name=frames" json:"frames,omitempty"` + Frames []*FaceFrame `protobuf:"bytes,3,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1117,25 +1117,25 @@ func (m *FaceAnnotation) GetFrames() []*FaceFrame { type VideoAnnotationResults struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Label annotations on video level or user specified segment level. // There is exactly one element for each unique label. - SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations" json:"segment_label_annotations,omitempty"` + SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations,proto3" json:"segment_label_annotations,omitempty"` // Label annotations on shot level. // There is exactly one element for each unique label. - ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations" json:"shot_label_annotations,omitempty"` + ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations,proto3" json:"shot_label_annotations,omitempty"` // Label annotations on frame level. // There is exactly one element for each unique label. - FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations" json:"frame_label_annotations,omitempty"` + FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations,proto3" json:"frame_label_annotations,omitempty"` // Face annotations. There is exactly one element for each unique face. - FaceAnnotations []*FaceAnnotation `protobuf:"bytes,5,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"` + FaceAnnotations []*FaceAnnotation `protobuf:"bytes,5,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // Shot annotations. Each shot is represented as a video segment. - ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations" json:"shot_annotations,omitempty"` + ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations,proto3" json:"shot_annotations,omitempty"` // Explicit content annotation. - ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation" json:"explicit_annotation,omitempty"` + ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation,proto3" json:"explicit_annotation,omitempty"` // If set, indicates an error. Note that for a single `AnnotateVideoRequest` // some videos may succeed and some may fail. - Error *status.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1226,7 +1226,7 @@ func (m *VideoAnnotationResults) GetError() *status.Status { // call of the `google::longrunning::Operations` service. type AnnotateVideoResponse struct { // Annotation results for all videos specified in `AnnotateVideoRequest`. - AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults" json:"annotation_results,omitempty"` + AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults,proto3" json:"annotation_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1267,14 +1267,14 @@ func (m *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults type VideoAnnotationProgress struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Approximate percentage processed thus far. // Guaranteed to be 100 when fully processed. - ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent update. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1337,7 +1337,7 @@ func (m *VideoAnnotationProgress) GetUpdateTime() *timestamp.Timestamp { // call of the `google::longrunning::Operations` service. type AnnotateVideoProgress struct { // Progress metadata for all videos specified in `AnnotateVideoRequest`. - AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress" json:"annotation_progress,omitempty"` + AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress,proto3" json:"annotation_progress,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1437,8 +1437,7 @@ func (c *videoIntelligenceServiceClient) AnnotateVideo(ctx context.Context, in * return out, nil } -// Server API for VideoIntelligenceService service - +// VideoIntelligenceServiceServer is the server API for VideoIntelligenceService service. type VideoIntelligenceServiceServer interface { // Performs asynchronous video annotation. Progress and results can be // retrieved through the `google.longrunning.Operations` interface. diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1p1beta1/video_intelligence.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1p1beta1/video_intelligence.pb.go index eee7bd146..8e972a6a4 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1p1beta1/video_intelligence.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/videointelligence/v1p1beta1/video_intelligence.pb.go @@ -63,7 +63,7 @@ func (x Feature) String() string { return proto.EnumName(Feature_name, int32(x)) } func (Feature) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{0} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{0} } // Label detection mode. @@ -97,7 +97,7 @@ func (x LabelDetectionMode) String() string { return proto.EnumName(LabelDetectionMode_name, int32(x)) } func (LabelDetectionMode) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{1} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{1} } // Bucketized representation of likelihood. @@ -139,7 +139,7 @@ func (x Likelihood) String() string { return proto.EnumName(Likelihood_name, int32(x)) } func (Likelihood) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{2} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{2} } // Video annotation request. @@ -154,26 +154,26 @@ type AnnotateVideoRequest struct { // multiple videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded // in the request as `input_content`. If set, `input_content` should be unset. - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // The video data bytes. // If unset, the input video(s) should be specified via `input_uri`. // If set, `input_uri` should be unset. InputContent []byte `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"` // Requested video annotation features. - Features []Feature `protobuf:"varint,2,rep,packed,name=features,enum=google.cloud.videointelligence.v1p1beta1.Feature" json:"features,omitempty"` + Features []Feature `protobuf:"varint,2,rep,packed,name=features,proto3,enum=google.cloud.videointelligence.v1p1beta1.Feature" json:"features,omitempty"` // Additional video context and/or feature-specific parameters. - VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext" json:"video_context,omitempty"` + VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext,proto3" json:"video_context,omitempty"` // Optional location where the output (in JSON format) should be stored. // Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see // [Request URIs](/storage/docs/reference-uris). - OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"` + OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` // Optional cloud region where annotation should take place. Supported cloud // regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region // is specified, a region will be determined based on video file location. - LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId" json:"location_id,omitempty"` + LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -183,7 +183,7 @@ func (m *AnnotateVideoRequest) Reset() { *m = AnnotateVideoRequest{} } func (m *AnnotateVideoRequest) String() string { return proto.CompactTextString(m) } func (*AnnotateVideoRequest) ProtoMessage() {} func (*AnnotateVideoRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{0} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{0} } func (m *AnnotateVideoRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnnotateVideoRequest.Unmarshal(m, b) @@ -248,17 +248,17 @@ func (m *AnnotateVideoRequest) GetLocationId() string { // Video context and/or feature-specific parameters. type VideoContext struct { // Video segments to annotate. The segments may overlap and are not required - // to be contiguous or span the whole video. If unspecified, each video - // is treated as a single segment. - Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments" json:"segments,omitempty"` + // to be contiguous or span the whole video. If unspecified, each video is + // treated as a single segment. + Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments,proto3" json:"segments,omitempty"` // Config for LABEL_DETECTION. - LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig" json:"label_detection_config,omitempty"` + LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig,proto3" json:"label_detection_config,omitempty"` // Config for SHOT_CHANGE_DETECTION. - ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig" json:"shot_change_detection_config,omitempty"` + ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig,proto3" json:"shot_change_detection_config,omitempty"` // Config for EXPLICIT_CONTENT_DETECTION. - ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig" json:"explicit_content_detection_config,omitempty"` + ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig,proto3" json:"explicit_content_detection_config,omitempty"` // Config for SPEECH_TRANSCRIPTION. - SpeechTranscriptionConfig *SpeechTranscriptionConfig `protobuf:"bytes,6,opt,name=speech_transcription_config,json=speechTranscriptionConfig" json:"speech_transcription_config,omitempty"` + SpeechTranscriptionConfig *SpeechTranscriptionConfig `protobuf:"bytes,6,opt,name=speech_transcription_config,json=speechTranscriptionConfig,proto3" json:"speech_transcription_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -268,7 +268,7 @@ func (m *VideoContext) Reset() { *m = VideoContext{} } func (m *VideoContext) String() string { return proto.CompactTextString(m) } func (*VideoContext) ProtoMessage() {} func (*VideoContext) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{1} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{1} } func (m *VideoContext) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_VideoContext.Unmarshal(m, b) @@ -328,15 +328,15 @@ type LabelDetectionConfig struct { // What labels should be detected with LABEL_DETECTION, in addition to // video-level labels or segment-level labels. // If unspecified, defaults to `SHOT_MODE`. - LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,enum=google.cloud.videointelligence.v1p1beta1.LabelDetectionMode" json:"label_detection_mode,omitempty"` + LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,proto3,enum=google.cloud.videointelligence.v1p1beta1.LabelDetectionMode" json:"label_detection_mode,omitempty"` // Whether the video has been shot from a stationary (i.e. non-moving) camera. // When set to true, might improve detection accuracy for moving objects. // Should be used with `SHOT_AND_FRAME_MODE` enabled. - StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera" json:"stationary_camera,omitempty"` + StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera,proto3" json:"stationary_camera,omitempty"` // Model to use for label detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -346,7 +346,7 @@ func (m *LabelDetectionConfig) Reset() { *m = LabelDetectionConfig{} } func (m *LabelDetectionConfig) String() string { return proto.CompactTextString(m) } func (*LabelDetectionConfig) ProtoMessage() {} func (*LabelDetectionConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{2} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{2} } func (m *LabelDetectionConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LabelDetectionConfig.Unmarshal(m, b) @@ -392,7 +392,7 @@ type ShotChangeDetectionConfig struct { // Model to use for shot change detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -402,7 +402,7 @@ func (m *ShotChangeDetectionConfig) Reset() { *m = ShotChangeDetectionCo func (m *ShotChangeDetectionConfig) String() string { return proto.CompactTextString(m) } func (*ShotChangeDetectionConfig) ProtoMessage() {} func (*ShotChangeDetectionConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{3} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{3} } func (m *ShotChangeDetectionConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ShotChangeDetectionConfig.Unmarshal(m, b) @@ -434,7 +434,7 @@ type ExplicitContentDetectionConfig struct { // Model to use for explicit content detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -444,7 +444,7 @@ func (m *ExplicitContentDetectionConfig) Reset() { *m = ExplicitContentD func (m *ExplicitContentDetectionConfig) String() string { return proto.CompactTextString(m) } func (*ExplicitContentDetectionConfig) ProtoMessage() {} func (*ExplicitContentDetectionConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{4} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{4} } func (m *ExplicitContentDetectionConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExplicitContentDetectionConfig.Unmarshal(m, b) @@ -475,10 +475,10 @@ func (m *ExplicitContentDetectionConfig) GetModel() string { type VideoSegment struct { // Time-offset, relative to the beginning of the video, // corresponding to the start of the segment (inclusive). - StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset" json:"start_time_offset,omitempty"` + StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset,proto3" json:"start_time_offset,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the end of the segment (inclusive). - EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset" json:"end_time_offset,omitempty"` + EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset,proto3" json:"end_time_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -488,7 +488,7 @@ func (m *VideoSegment) Reset() { *m = VideoSegment{} } func (m *VideoSegment) String() string { return proto.CompactTextString(m) } func (*VideoSegment) ProtoMessage() {} func (*VideoSegment) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{5} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{5} } func (m *VideoSegment) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_VideoSegment.Unmarshal(m, b) @@ -525,9 +525,9 @@ func (m *VideoSegment) GetEndTimeOffset() *duration.Duration { // Video segment level annotation results for label detection. type LabelSegment struct { // Video segment where a label was detected. - Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"` + Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -537,7 +537,7 @@ func (m *LabelSegment) Reset() { *m = LabelSegment{} } func (m *LabelSegment) String() string { return proto.CompactTextString(m) } func (*LabelSegment) ProtoMessage() {} func (*LabelSegment) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{6} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{6} } func (m *LabelSegment) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LabelSegment.Unmarshal(m, b) @@ -575,9 +575,9 @@ func (m *LabelSegment) GetConfidence() float32 { type LabelFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -587,7 +587,7 @@ func (m *LabelFrame) Reset() { *m = LabelFrame{} } func (m *LabelFrame) String() string { return proto.CompactTextString(m) } func (*LabelFrame) ProtoMessage() {} func (*LabelFrame) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{7} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{7} } func (m *LabelFrame) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LabelFrame.Unmarshal(m, b) @@ -626,11 +626,11 @@ type Entity struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search // API](https://developers.google.com/knowledge-graph/). - EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Textual description, e.g. `Fixed-gear bicycle`. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Language code for `description` in BCP-47 format. - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -640,7 +640,7 @@ func (m *Entity) Reset() { *m = Entity{} } func (m *Entity) String() string { return proto.CompactTextString(m) } func (*Entity) ProtoMessage() {} func (*Entity) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{8} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{8} } func (m *Entity) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Entity.Unmarshal(m, b) @@ -684,16 +684,16 @@ func (m *Entity) GetLanguageCode() string { // Label annotation. type LabelAnnotation struct { // Detected entity. - Entity *Entity `protobuf:"bytes,1,opt,name=entity" json:"entity,omitempty"` + Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"` // Common categories for the detected entity. // E.g. when the label is `Terrier` the category is likely `dog`. And in some // cases there might be more than one categories e.g. `Terrier` could also be // a `pet`. - CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities" json:"category_entities,omitempty"` + CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities,proto3" json:"category_entities,omitempty"` // All video segments where a label was detected. - Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments" json:"segments,omitempty"` + Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a label was detected. - Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames" json:"frames,omitempty"` + Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -703,7 +703,7 @@ func (m *LabelAnnotation) Reset() { *m = LabelAnnotation{} } func (m *LabelAnnotation) String() string { return proto.CompactTextString(m) } func (*LabelAnnotation) ProtoMessage() {} func (*LabelAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{9} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{9} } func (m *LabelAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LabelAnnotation.Unmarshal(m, b) @@ -755,9 +755,9 @@ func (m *LabelAnnotation) GetFrames() []*LabelFrame { type ExplicitContentFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. - TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"` + TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Likelihood of the pornography content.. - PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,enum=google.cloud.videointelligence.v1p1beta1.Likelihood" json:"pornography_likelihood,omitempty"` + PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,proto3,enum=google.cloud.videointelligence.v1p1beta1.Likelihood" json:"pornography_likelihood,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -767,7 +767,7 @@ func (m *ExplicitContentFrame) Reset() { *m = ExplicitContentFrame{} } func (m *ExplicitContentFrame) String() string { return proto.CompactTextString(m) } func (*ExplicitContentFrame) ProtoMessage() {} func (*ExplicitContentFrame) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{10} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{10} } func (m *ExplicitContentFrame) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExplicitContentFrame.Unmarshal(m, b) @@ -806,7 +806,7 @@ func (m *ExplicitContentFrame) GetPornographyLikelihood() Likelihood { // present for that frame. type ExplicitContentAnnotation struct { // All video frames where explicit content was detected. - Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames" json:"frames,omitempty"` + Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames,proto3" json:"frames,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -816,7 +816,7 @@ func (m *ExplicitContentAnnotation) Reset() { *m = ExplicitContentAnnota func (m *ExplicitContentAnnotation) String() string { return proto.CompactTextString(m) } func (*ExplicitContentAnnotation) ProtoMessage() {} func (*ExplicitContentAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{11} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{11} } func (m *ExplicitContentAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ExplicitContentAnnotation.Unmarshal(m, b) @@ -843,98 +843,29 @@ func (m *ExplicitContentAnnotation) GetFrames() []*ExplicitContentFrame { return nil } -// Normalized bounding box. -// The normalized vertex coordinates are relative to the original image. -// Range: [0, 1]. -type NormalizedBoundingBox struct { - // Left X coordinate. - Left float32 `protobuf:"fixed32,1,opt,name=left" json:"left,omitempty"` - // Top Y coordinate. - Top float32 `protobuf:"fixed32,2,opt,name=top" json:"top,omitempty"` - // Right X coordinate. - Right float32 `protobuf:"fixed32,3,opt,name=right" json:"right,omitempty"` - // Bottom Y coordinate. - Bottom float32 `protobuf:"fixed32,4,opt,name=bottom" json:"bottom,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *NormalizedBoundingBox) Reset() { *m = NormalizedBoundingBox{} } -func (m *NormalizedBoundingBox) String() string { return proto.CompactTextString(m) } -func (*NormalizedBoundingBox) ProtoMessage() {} -func (*NormalizedBoundingBox) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{12} -} -func (m *NormalizedBoundingBox) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_NormalizedBoundingBox.Unmarshal(m, b) -} -func (m *NormalizedBoundingBox) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_NormalizedBoundingBox.Marshal(b, m, deterministic) -} -func (dst *NormalizedBoundingBox) XXX_Merge(src proto.Message) { - xxx_messageInfo_NormalizedBoundingBox.Merge(dst, src) -} -func (m *NormalizedBoundingBox) XXX_Size() int { - return xxx_messageInfo_NormalizedBoundingBox.Size(m) -} -func (m *NormalizedBoundingBox) XXX_DiscardUnknown() { - xxx_messageInfo_NormalizedBoundingBox.DiscardUnknown(m) -} - -var xxx_messageInfo_NormalizedBoundingBox proto.InternalMessageInfo - -func (m *NormalizedBoundingBox) GetLeft() float32 { - if m != nil { - return m.Left - } - return 0 -} - -func (m *NormalizedBoundingBox) GetTop() float32 { - if m != nil { - return m.Top - } - return 0 -} - -func (m *NormalizedBoundingBox) GetRight() float32 { - if m != nil { - return m.Right - } - return 0 -} - -func (m *NormalizedBoundingBox) GetBottom() float32 { - if m != nil { - return m.Bottom - } - return 0 -} - // Annotation results for a single video. type VideoAnnotationResults struct { - // Video file location in + // Output only. Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Label annotations on video level or user specified segment level. // There is exactly one element for each unique label. - SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations" json:"segment_label_annotations,omitempty"` + SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations,proto3" json:"segment_label_annotations,omitempty"` // Label annotations on shot level. // There is exactly one element for each unique label. - ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations" json:"shot_label_annotations,omitempty"` + ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations,proto3" json:"shot_label_annotations,omitempty"` // Label annotations on frame level. // There is exactly one element for each unique label. - FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations" json:"frame_label_annotations,omitempty"` + FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations,proto3" json:"frame_label_annotations,omitempty"` // Shot annotations. Each shot is represented as a video segment. - ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations" json:"shot_annotations,omitempty"` + ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations,proto3" json:"shot_annotations,omitempty"` // Explicit content annotation. - ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation" json:"explicit_annotation,omitempty"` + ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation,proto3" json:"explicit_annotation,omitempty"` // Speech transcription. - SpeechTranscriptions []*SpeechTranscription `protobuf:"bytes,11,rep,name=speech_transcriptions,json=speechTranscriptions" json:"speech_transcriptions,omitempty"` - // If set, indicates an error. Note that for a single `AnnotateVideoRequest` - // some videos may succeed and some may fail. - Error *status.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"` + SpeechTranscriptions []*SpeechTranscription `protobuf:"bytes,11,rep,name=speech_transcriptions,json=speechTranscriptions,proto3" json:"speech_transcriptions,omitempty"` + // Output only. If set, indicates an error. Note that for a single + // `AnnotateVideoRequest` some videos may succeed and some may fail. + Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -944,7 +875,7 @@ func (m *VideoAnnotationResults) Reset() { *m = VideoAnnotationResults{} func (m *VideoAnnotationResults) String() string { return proto.CompactTextString(m) } func (*VideoAnnotationResults) ProtoMessage() {} func (*VideoAnnotationResults) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{13} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{12} } func (m *VideoAnnotationResults) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_VideoAnnotationResults.Unmarshal(m, b) @@ -1025,7 +956,7 @@ func (m *VideoAnnotationResults) GetError() *status.Status { // call of the `google::longrunning::Operations` service. type AnnotateVideoResponse struct { // Annotation results for all videos specified in `AnnotateVideoRequest`. - AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults" json:"annotation_results,omitempty"` + AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults,proto3" json:"annotation_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1035,7 +966,7 @@ func (m *AnnotateVideoResponse) Reset() { *m = AnnotateVideoResponse{} } func (m *AnnotateVideoResponse) String() string { return proto.CompactTextString(m) } func (*AnnotateVideoResponse) ProtoMessage() {} func (*AnnotateVideoResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{14} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{13} } func (m *AnnotateVideoResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnnotateVideoResponse.Unmarshal(m, b) @@ -1064,16 +995,16 @@ func (m *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults // Annotation progress for a single video. type VideoAnnotationProgress struct { - // Video file location in + // Output only. Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). - InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"` - // Approximate percentage processed thus far. - // Guaranteed to be 100 when fully processed. - ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` - // Time when the request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"` - // Time of the most recent update. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` + // Output only. Approximate percentage processed thus far. Guaranteed to be + // 100 when fully processed. + ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` + // Output only. Time when the request was received. + StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` + // Output only. Time of the most recent update. + UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1083,7 +1014,7 @@ func (m *VideoAnnotationProgress) Reset() { *m = VideoAnnotationProgress func (m *VideoAnnotationProgress) String() string { return proto.CompactTextString(m) } func (*VideoAnnotationProgress) ProtoMessage() {} func (*VideoAnnotationProgress) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{15} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{14} } func (m *VideoAnnotationProgress) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_VideoAnnotationProgress.Unmarshal(m, b) @@ -1136,7 +1067,7 @@ func (m *VideoAnnotationProgress) GetUpdateTime() *timestamp.Timestamp { // call of the `google::longrunning::Operations` service. type AnnotateVideoProgress struct { // Progress metadata for all videos specified in `AnnotateVideoRequest`. - AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress" json:"annotation_progress,omitempty"` + AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress,proto3" json:"annotation_progress,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1146,7 +1077,7 @@ func (m *AnnotateVideoProgress) Reset() { *m = AnnotateVideoProgress{} } func (m *AnnotateVideoProgress) String() string { return proto.CompactTextString(m) } func (*AnnotateVideoProgress) ProtoMessage() {} func (*AnnotateVideoProgress) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{16} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{15} } func (m *AnnotateVideoProgress) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnnotateVideoProgress.Unmarshal(m, b) @@ -1180,23 +1111,30 @@ type SpeechTranscriptionConfig struct { // Example: "en-US". // See [Language Support](https://cloud.google.com/speech/docs/languages) // for a list of the currently supported language codes. - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // *Optional* Maximum number of recognition hypotheses to be returned. // Specifically, the maximum number of `SpeechRecognitionAlternative` messages // within each `SpeechRecognitionResult`. The server may return fewer than // `max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will // return a maximum of one. If omitted, will return a maximum of one. - MaxAlternatives int32 `protobuf:"varint,2,opt,name=max_alternatives,json=maxAlternatives" json:"max_alternatives,omitempty"` + MaxAlternatives int32 `protobuf:"varint,2,opt,name=max_alternatives,json=maxAlternatives,proto3" json:"max_alternatives,omitempty"` // *Optional* If set to `true`, the server will attempt to filter out // profanities, replacing all but the initial character in each filtered word // with asterisks, e.g. "f***". If set to `false` or omitted, profanities // won't be filtered out. - FilterProfanity bool `protobuf:"varint,3,opt,name=filter_profanity,json=filterProfanity" json:"filter_profanity,omitempty"` + FilterProfanity bool `protobuf:"varint,3,opt,name=filter_profanity,json=filterProfanity,proto3" json:"filter_profanity,omitempty"` // *Optional* A means to provide context to assist the speech recognition. - SpeechContexts []*SpeechContext `protobuf:"bytes,4,rep,name=speech_contexts,json=speechContexts" json:"speech_contexts,omitempty"` + SpeechContexts []*SpeechContext `protobuf:"bytes,4,rep,name=speech_contexts,json=speechContexts,proto3" json:"speech_contexts,omitempty"` + // *Optional* If 'true', adds punctuation to recognition result hypotheses. + // This feature is only available in select languages. Setting this for + // requests in other languages has no effect at all. The default 'false' value + // does not add punctuation to result hypotheses. NOTE: "This is currently + // offered as an experimental service, complimentary to all users. In the + // future this may be exclusively available as a premium feature." + EnableAutomaticPunctuation bool `protobuf:"varint,5,opt,name=enable_automatic_punctuation,json=enableAutomaticPunctuation,proto3" json:"enable_automatic_punctuation,omitempty"` // *Optional* For file formats, such as MXF or MKV, supporting multiple audio // tracks, specify up to two tracks. Default: track 0. - AudioTracks []int32 `protobuf:"varint,6,rep,packed,name=audio_tracks,json=audioTracks" json:"audio_tracks,omitempty"` + AudioTracks []int32 `protobuf:"varint,6,rep,packed,name=audio_tracks,json=audioTracks,proto3" json:"audio_tracks,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1206,7 +1144,7 @@ func (m *SpeechTranscriptionConfig) Reset() { *m = SpeechTranscriptionCo func (m *SpeechTranscriptionConfig) String() string { return proto.CompactTextString(m) } func (*SpeechTranscriptionConfig) ProtoMessage() {} func (*SpeechTranscriptionConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{17} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{16} } func (m *SpeechTranscriptionConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SpeechTranscriptionConfig.Unmarshal(m, b) @@ -1254,6 +1192,13 @@ func (m *SpeechTranscriptionConfig) GetSpeechContexts() []*SpeechContext { return nil } +func (m *SpeechTranscriptionConfig) GetEnableAutomaticPunctuation() bool { + if m != nil { + return m.EnableAutomaticPunctuation + } + return false +} + func (m *SpeechTranscriptionConfig) GetAudioTracks() []int32 { if m != nil { return m.AudioTracks @@ -1270,7 +1215,7 @@ type SpeechContext struct { // specific commands are typically spoken by the user. This can also be used // to add additional words to the vocabulary of the recognizer. See // [usage limits](https://cloud.google.com/speech/limits#content). - Phrases []string `protobuf:"bytes,1,rep,name=phrases" json:"phrases,omitempty"` + Phrases []string `protobuf:"bytes,1,rep,name=phrases,proto3" json:"phrases,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1280,7 +1225,7 @@ func (m *SpeechContext) Reset() { *m = SpeechContext{} } func (m *SpeechContext) String() string { return proto.CompactTextString(m) } func (*SpeechContext) ProtoMessage() {} func (*SpeechContext) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{18} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{17} } func (m *SpeechContext) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SpeechContext.Unmarshal(m, b) @@ -1313,7 +1258,7 @@ type SpeechTranscription struct { // maximum specified in `max_alternatives`). // These alternatives are ordered in terms of accuracy, with the top (first) // alternative being the most probable, as ranked by the recognizer. - Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` + Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1323,7 +1268,7 @@ func (m *SpeechTranscription) Reset() { *m = SpeechTranscription{} } func (m *SpeechTranscription) String() string { return proto.CompactTextString(m) } func (*SpeechTranscription) ProtoMessage() {} func (*SpeechTranscription) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{19} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{18} } func (m *SpeechTranscription) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SpeechTranscription.Unmarshal(m, b) @@ -1353,16 +1298,16 @@ func (m *SpeechTranscription) GetAlternatives() []*SpeechRecognitionAlternative // Alternative hypotheses (a.k.a. n-best list). type SpeechRecognitionAlternative struct { // Output only. Transcript text representing the words that the user spoke. - Transcript string `protobuf:"bytes,1,opt,name=transcript" json:"transcript,omitempty"` + Transcript string `protobuf:"bytes,1,opt,name=transcript,proto3" json:"transcript,omitempty"` // Output only. The confidence estimate between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. This field is typically provided only for the top hypothesis, and // only for `is_final=true` results. Clients should not rely on the // `confidence` field as it is not guaranteed to be accurate or consistent. // The default of 0.0 is a sentinel value indicating `confidence` was not set. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // Output only. A list of word-specific information for each recognized word. - Words []*WordInfo `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` + Words []*WordInfo `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1372,7 +1317,7 @@ func (m *SpeechRecognitionAlternative) Reset() { *m = SpeechRecognitionA func (m *SpeechRecognitionAlternative) String() string { return proto.CompactTextString(m) } func (*SpeechRecognitionAlternative) ProtoMessage() {} func (*SpeechRecognitionAlternative) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{20} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{19} } func (m *SpeechRecognitionAlternative) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SpeechRecognitionAlternative.Unmarshal(m, b) @@ -1421,14 +1366,14 @@ type WordInfo struct { // corresponding to the start of the spoken word. This field is only set if // `enable_word_time_offsets=true` and only in the top hypothesis. This is an // experimental feature and the accuracy of the time offset can vary. - StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Output only. Time offset relative to the beginning of the audio, and // corresponding to the end of the spoken word. This field is only set if // `enable_word_time_offsets=true` and only in the top hypothesis. This is an // experimental feature and the accuracy of the time offset can vary. - EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Output only. The word corresponding to this set of information. - Word string `protobuf:"bytes,3,opt,name=word" json:"word,omitempty"` + Word string `protobuf:"bytes,3,opt,name=word,proto3" json:"word,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1438,7 +1383,7 @@ func (m *WordInfo) Reset() { *m = WordInfo{} } func (m *WordInfo) String() string { return proto.CompactTextString(m) } func (*WordInfo) ProtoMessage() {} func (*WordInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_video_intelligence_de0b3deab89e8118, []int{21} + return fileDescriptor_video_intelligence_de2dfee67bcd43cb, []int{20} } func (m *WordInfo) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WordInfo.Unmarshal(m, b) @@ -1492,7 +1437,6 @@ func init() { proto.RegisterType((*LabelAnnotation)(nil), "google.cloud.videointelligence.v1p1beta1.LabelAnnotation") proto.RegisterType((*ExplicitContentFrame)(nil), "google.cloud.videointelligence.v1p1beta1.ExplicitContentFrame") proto.RegisterType((*ExplicitContentAnnotation)(nil), "google.cloud.videointelligence.v1p1beta1.ExplicitContentAnnotation") - proto.RegisterType((*NormalizedBoundingBox)(nil), "google.cloud.videointelligence.v1p1beta1.NormalizedBoundingBox") proto.RegisterType((*VideoAnnotationResults)(nil), "google.cloud.videointelligence.v1p1beta1.VideoAnnotationResults") proto.RegisterType((*AnnotateVideoResponse)(nil), "google.cloud.videointelligence.v1p1beta1.AnnotateVideoResponse") proto.RegisterType((*VideoAnnotationProgress)(nil), "google.cloud.videointelligence.v1p1beta1.VideoAnnotationProgress") @@ -1543,8 +1487,7 @@ func (c *videoIntelligenceServiceClient) AnnotateVideo(ctx context.Context, in * return out, nil } -// Server API for VideoIntelligenceService service - +// VideoIntelligenceServiceServer is the server API for VideoIntelligenceService service. type VideoIntelligenceServiceServer interface { // Performs asynchronous video annotation. Progress and results can be // retrieved through the `google.longrunning.Operations` interface. @@ -1589,124 +1532,122 @@ var _VideoIntelligenceService_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/cloud/videointelligence/v1p1beta1/video_intelligence.proto", fileDescriptor_video_intelligence_de0b3deab89e8118) + proto.RegisterFile("google/cloud/videointelligence/v1p1beta1/video_intelligence.proto", fileDescriptor_video_intelligence_de2dfee67bcd43cb) } -var fileDescriptor_video_intelligence_de0b3deab89e8118 = []byte{ - // 1838 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x58, 0x4f, 0x6f, 0x23, 0x49, +var fileDescriptor_video_intelligence_de2dfee67bcd43cb = []byte{ + // 1807 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x58, 0xcd, 0x6f, 0x23, 0x49, 0x15, 0xa7, 0xed, 0xc4, 0x89, 0x9f, 0x9d, 0xd8, 0x53, 0x71, 0x12, 0x27, 0x3b, 0x93, 0xc9, 0xf6, - 0x82, 0x94, 0x1d, 0x90, 0x4d, 0xc2, 0x6a, 0x61, 0x67, 0x61, 0x85, 0xe3, 0x74, 0x36, 0x16, 0x19, - 0xdb, 0x2a, 0x7b, 0x02, 0x03, 0x83, 0x9a, 0x4e, 0x77, 0xb9, 0xdd, 0x4c, 0xbb, 0xab, 0xb7, 0xbb, - 0x1c, 0x32, 0x9c, 0x96, 0x3f, 0x12, 0x48, 0x20, 0x2e, 0x23, 0x24, 0x24, 0xbe, 0x00, 0xe2, 0xc8, - 0x17, 0xe0, 0xc2, 0x05, 0xae, 0x9c, 0x90, 0x38, 0xf2, 0x2d, 0xb8, 0xac, 0xba, 0xaa, 0xda, 0x6e, - 0xdb, 0xc9, 0x24, 0x4e, 0x6e, 0x5d, 0xef, 0xd5, 0xfb, 0xbd, 0x3f, 0xf5, 0xde, 0xab, 0x57, 0x0d, - 0x35, 0x9b, 0x52, 0xdb, 0x25, 0x55, 0xd3, 0xa5, 0x43, 0xab, 0x7a, 0xe1, 0x58, 0x84, 0x3a, 0x1e, - 0x23, 0xae, 0xeb, 0xd8, 0xc4, 0x33, 0x49, 0xf5, 0x62, 0xdf, 0xdf, 0x3f, 0x27, 0xcc, 0xd8, 0x17, - 0x3c, 0x3d, 0xc9, 0xac, 0xf8, 0x01, 0x65, 0x14, 0xed, 0x09, 0x88, 0x0a, 0x87, 0xa8, 0xcc, 0x40, - 0x54, 0x46, 0x10, 0xdb, 0x0f, 0xa5, 0x32, 0xc3, 0x77, 0xaa, 0x86, 0xe7, 0x51, 0x66, 0x30, 0x87, - 0x7a, 0xa1, 0xc0, 0xd9, 0x7e, 0x4f, 0x72, 0x5d, 0xea, 0xd9, 0xc1, 0xd0, 0xf3, 0x1c, 0xcf, 0xae, - 0x52, 0x9f, 0x04, 0x13, 0x9b, 0x76, 0xe4, 0x26, 0xbe, 0x3a, 0x1f, 0xf6, 0xaa, 0xd6, 0x50, 0x6c, - 0x90, 0xfc, 0xc7, 0xd3, 0x7c, 0xe6, 0x0c, 0x48, 0xc8, 0x8c, 0x81, 0x2f, 0x37, 0x6c, 0xca, 0x0d, - 0x81, 0x6f, 0x56, 0x43, 0x66, 0xb0, 0xa1, 0x44, 0x56, 0xff, 0x91, 0x82, 0x52, 0x4d, 0x18, 0x45, - 0xce, 0x22, 0x27, 0x30, 0xf9, 0x6c, 0x48, 0x42, 0x86, 0xde, 0x81, 0xac, 0xe3, 0xf9, 0x43, 0xa6, - 0x0f, 0x03, 0xa7, 0xac, 0xec, 0x2a, 0x7b, 0x59, 0xbc, 0xcc, 0x09, 0xcf, 0x03, 0x07, 0xbd, 0x07, - 0x2b, 0x82, 0x69, 0x52, 0x8f, 0x11, 0x8f, 0x95, 0x33, 0xbb, 0xca, 0x5e, 0x1e, 0xe7, 0x39, 0xb1, - 0x2e, 0x68, 0xe8, 0x19, 0x2c, 0xf7, 0x88, 0xc1, 0x86, 0x01, 0x09, 0xcb, 0xa9, 0xdd, 0xf4, 0xde, - 0xea, 0xc1, 0x7e, 0xe5, 0xb6, 0x41, 0xab, 0x1c, 0x0b, 0x49, 0x3c, 0x82, 0x40, 0x3f, 0x82, 0x15, - 0x71, 0x18, 0x5c, 0xe7, 0x25, 0x2b, 0xa7, 0x77, 0x95, 0xbd, 0xdc, 0xc1, 0x87, 0xb7, 0xc7, 0xe4, - 0xfe, 0xd5, 0x85, 0x34, 0xce, 0x5f, 0x24, 0x56, 0xe8, 0x11, 0x00, 0x1d, 0xb2, 0xd8, 0xdd, 0x05, - 0xee, 0x6e, 0x56, 0x50, 0x22, 0x7f, 0x1f, 0x43, 0xce, 0xa5, 0x26, 0x8f, 0xb8, 0xee, 0x58, 0xe5, - 0x45, 0xce, 0x87, 0x98, 0xd4, 0xb0, 0xd4, 0xff, 0x2f, 0x40, 0x3e, 0x09, 0x8f, 0x30, 0x2c, 0x87, - 0xc4, 0x1e, 0x10, 0x8f, 0x85, 0x65, 0x65, 0x37, 0x7d, 0x07, 0x43, 0x3b, 0x42, 0x1c, 0x8f, 0x70, - 0x10, 0x83, 0x0d, 0xd7, 0x38, 0x27, 0xae, 0x6e, 0x11, 0x46, 0x4c, 0x6e, 0x8c, 0x49, 0xbd, 0x9e, - 0x63, 0x97, 0x53, 0x3c, 0x14, 0x9f, 0xdc, 0x5e, 0xc3, 0x69, 0x84, 0x73, 0x14, 0xc3, 0xd4, 0x39, - 0x0a, 0x2e, 0xb9, 0x57, 0x50, 0xd1, 0xaf, 0x15, 0x78, 0x18, 0xf6, 0x29, 0xd3, 0xcd, 0xbe, 0xe1, - 0xd9, 0x64, 0x56, 0xb9, 0x38, 0x87, 0xfa, 0xed, 0x95, 0x77, 0xfa, 0x94, 0xd5, 0x39, 0xd8, 0xb4, - 0x05, 0x5b, 0xe1, 0x75, 0x2c, 0xf4, 0x46, 0x81, 0x77, 0xc9, 0xa5, 0xef, 0x3a, 0xa6, 0x33, 0x4a, - 0xbb, 0x59, 0x5b, 0x16, 0xb8, 0x2d, 0x27, 0xb7, 0xb7, 0x45, 0x93, 0x90, 0x32, 0x69, 0xa7, 0x0d, - 0xda, 0x21, 0x6f, 0xe5, 0xa3, 0x5f, 0x29, 0xf0, 0x4e, 0xe8, 0x13, 0x62, 0xf6, 0x75, 0x16, 0x18, - 0x5e, 0x68, 0x06, 0x8e, 0x9f, 0xb4, 0x27, 0x33, 0x77, 0x6c, 0x38, 0x58, 0x37, 0x89, 0x35, 0x8a, - 0xcd, 0x75, 0x2c, 0xf5, 0x9f, 0x0a, 0x94, 0xae, 0x3a, 0x51, 0xe4, 0x41, 0x69, 0x3a, 0x63, 0x06, - 0xd4, 0x22, 0xbc, 0x9e, 0x57, 0x0f, 0xbe, 0x7d, 0xd7, 0x7c, 0x79, 0x46, 0x2d, 0x82, 0x91, 0x3b, - 0x43, 0x43, 0x5f, 0x85, 0x07, 0xa1, 0x68, 0x6f, 0x46, 0xf0, 0x5a, 0x37, 0x8d, 0x01, 0x09, 0x0c, - 0x9e, 0x9c, 0xcb, 0xb8, 0x38, 0x66, 0xd4, 0x39, 0x1d, 0x95, 0x60, 0x31, 0x32, 0xc6, 0xe5, 0x09, - 0x94, 0xc5, 0x62, 0xa1, 0xee, 0xc3, 0xd6, 0xb5, 0xf9, 0x31, 0x16, 0x51, 0x92, 0x22, 0x1f, 0xc2, - 0xce, 0xdb, 0x8f, 0xf1, 0x1a, 0xb9, 0x3f, 0x29, 0xb2, 0x68, 0x65, 0xa9, 0x21, 0x8d, 0x9b, 0x1f, - 0x30, 0x3d, 0x6a, 0x9f, 0x3a, 0xed, 0xf5, 0x42, 0xc2, 0xb8, 0x48, 0xee, 0x60, 0x2b, 0x8e, 0x55, - 0xdc, 0x62, 0x2b, 0x47, 0xb2, 0x05, 0xe3, 0x02, 0x97, 0xe9, 0x3a, 0x03, 0xd2, 0xe2, 0x12, 0xa8, - 0x06, 0x05, 0xe2, 0x59, 0x13, 0x20, 0xa9, 0x9b, 0x40, 0x56, 0x88, 0x67, 0x8d, 0x21, 0xd4, 0xcf, - 0x15, 0xc8, 0xf3, 0x98, 0xc7, 0xa6, 0xb5, 0x61, 0x49, 0xf6, 0x01, 0x69, 0xd0, 0x5d, 0xdb, 0x49, - 0x0c, 0x83, 0x76, 0x00, 0x78, 0x92, 0x5a, 0xd1, 0x6e, 0x6e, 0x60, 0x0a, 0x27, 0x28, 0x6a, 0x1f, - 0x80, 0x5b, 0x70, 0x1c, 0x18, 0x03, 0x82, 0x9e, 0x42, 0x6e, 0xae, 0xa0, 0x00, 0x1b, 0xc7, 0xe3, - 0x26, 0x4d, 0x2e, 0x64, 0x34, 0x8f, 0x39, 0xec, 0x75, 0x74, 0xe9, 0x10, 0xfe, 0x15, 0x75, 0x59, - 0x79, 0xe9, 0x08, 0x42, 0xc3, 0x42, 0xbb, 0x90, 0xb3, 0xc8, 0x28, 0xf5, 0x39, 0x4e, 0x16, 0x27, - 0x49, 0xd1, 0xb5, 0xe4, 0x1a, 0x9e, 0x3d, 0x34, 0x6c, 0xa2, 0x9b, 0x51, 0x9e, 0x8b, 0xcc, 0xca, - 0xc7, 0xc4, 0x3a, 0xb5, 0x88, 0xfa, 0xdf, 0x14, 0x14, 0xb8, 0x63, 0xb5, 0xd1, 0x5d, 0x8c, 0x4e, - 0x20, 0x23, 0xd4, 0x48, 0xc7, 0xbe, 0x3e, 0x47, 0x03, 0xe1, 0x72, 0x58, 0xca, 0xa3, 0x1f, 0xc3, - 0x03, 0xd3, 0x60, 0xc4, 0xa6, 0xc1, 0x6b, 0x9d, 0x93, 0x1c, 0x79, 0xfb, 0xdd, 0x05, 0xb4, 0x18, - 0x43, 0x69, 0x12, 0x69, 0xe2, 0x5a, 0x49, 0xcf, 0x7b, 0xad, 0x24, 0x13, 0x2a, 0x71, 0xad, 0x9c, - 0x42, 0xa6, 0x17, 0x9d, 0x71, 0x58, 0x5e, 0xe0, 0x88, 0x1f, 0xcc, 0x89, 0xc8, 0x13, 0x04, 0x4b, - 0x0c, 0xf5, 0xef, 0x0a, 0x94, 0xa6, 0xaa, 0xf1, 0xfe, 0x19, 0xf4, 0x0a, 0x36, 0x7c, 0x1a, 0x78, - 0xd4, 0x0e, 0x0c, 0xbf, 0xff, 0x5a, 0x77, 0x9d, 0x57, 0xc4, 0x75, 0xfa, 0x94, 0x5a, 0x3c, 0x0b, - 0x56, 0xe7, 0x32, 0x79, 0x24, 0x8b, 0xd7, 0x13, 0x98, 0x63, 0xb2, 0x1a, 0xc2, 0xd6, 0x94, 0x03, - 0x89, 0x4c, 0x39, 0x1b, 0x05, 0x4b, 0xdc, 0xea, 0x9f, 0xdc, 0xf9, 0xaa, 0x99, 0x0c, 0xdb, 0x2b, - 0x58, 0x6f, 0xd2, 0x60, 0x60, 0xb8, 0xce, 0xcf, 0x89, 0x75, 0x48, 0x87, 0x9e, 0xe5, 0x78, 0xf6, - 0x21, 0xbd, 0x44, 0x08, 0x16, 0x5c, 0xd2, 0x13, 0xf1, 0x4a, 0x61, 0xfe, 0x8d, 0x8a, 0x90, 0x66, - 0xd4, 0x97, 0x95, 0x14, 0x7d, 0x46, 0x0d, 0x2e, 0x70, 0xec, 0xbe, 0x18, 0x8a, 0x52, 0x58, 0x2c, - 0xd0, 0x06, 0x64, 0xce, 0x29, 0x63, 0x74, 0xc0, 0xef, 0xc5, 0x14, 0x96, 0x2b, 0xf5, 0x37, 0x19, - 0xd8, 0xe0, 0x4d, 0x61, 0xec, 0x18, 0x26, 0xe1, 0xd0, 0x65, 0xe1, 0xdb, 0xc7, 0xbe, 0x21, 0x6c, - 0xc9, 0xac, 0xd1, 0xc5, 0xb5, 0x92, 0x18, 0x67, 0x65, 0x92, 0x7f, 0x34, 0x67, 0xf2, 0x24, 0x2c, - 0xd8, 0x94, 0xd8, 0x53, 0xf4, 0x10, 0x51, 0xd8, 0xe0, 0x03, 0xc8, 0xac, 0xce, 0xf4, 0x7d, 0x75, - 0x96, 0x22, 0xe0, 0x19, 0x85, 0x9f, 0xc1, 0x26, 0x3f, 0x96, 0x2b, 0x34, 0x2e, 0xdc, 0x57, 0xe3, - 0x3a, 0x47, 0x9e, 0x51, 0x69, 0x40, 0x91, 0xfb, 0x98, 0xd4, 0x95, 0xb9, 0xd7, 0xdc, 0x58, 0x88, - 0xf0, 0x92, 0x2a, 0x18, 0xac, 0x8d, 0x06, 0xa8, 0xb1, 0x9a, 0xf2, 0xd2, 0xbc, 0x23, 0xca, 0xb5, - 0xc5, 0x81, 0x51, 0x8c, 0x9f, 0x28, 0x98, 0x00, 0xd6, 0xaf, 0x1a, 0x90, 0xc2, 0x72, 0x8e, 0x7b, - 0xf7, 0x9d, 0x7b, 0x8d, 0x46, 0xb8, 0x74, 0xc5, 0x50, 0x14, 0xa2, 0x3d, 0x58, 0x24, 0x41, 0x40, - 0x83, 0x72, 0x96, 0xfb, 0x86, 0x62, 0x1d, 0x81, 0x6f, 0x56, 0x3a, 0xfc, 0xf5, 0x83, 0xc5, 0x06, - 0xf5, 0xb7, 0x0a, 0xac, 0x4f, 0x3d, 0x7f, 0x42, 0x9f, 0x7a, 0x21, 0x41, 0x14, 0xd0, 0x38, 0x48, - 0x7a, 0x20, 0xca, 0x43, 0x16, 0xfd, 0x77, 0xe7, 0x3c, 0x92, 0x99, 0x32, 0xc3, 0x0f, 0x8c, 0x69, - 0x92, 0xfa, 0x1f, 0x05, 0x36, 0xa7, 0x76, 0xb7, 0x03, 0x6a, 0x07, 0x24, 0xbc, 0xa1, 0x2a, 0xdf, - 0x87, 0xa2, 0x2f, 0x37, 0xea, 0x3e, 0x09, 0xcc, 0x68, 0x46, 0x88, 0x5a, 0xc3, 0x22, 0x2e, 0xc4, - 0xf4, 0xb6, 0x20, 0xa3, 0x8f, 0x00, 0xc6, 0x03, 0x8e, 0x1c, 0xdc, 0xb7, 0x67, 0x5a, 0x70, 0x37, - 0x7e, 0x3c, 0xe2, 0xec, 0x68, 0xb4, 0x41, 0x1f, 0x43, 0x6e, 0xe8, 0x5b, 0x06, 0x23, 0x42, 0x76, - 0xe1, 0x46, 0x59, 0x10, 0xdb, 0x23, 0x82, 0xfa, 0xbb, 0xe9, 0x30, 0x8f, 0x3c, 0x0b, 0x60, 0x2d, - 0x11, 0xe6, 0xd8, 0x5e, 0x19, 0xe7, 0xda, 0x9d, 0xe3, 0x1c, 0xe3, 0xe3, 0xc4, 0x21, 0xc6, 0x34, - 0xf5, 0xcf, 0x29, 0xd8, 0xba, 0x76, 0xce, 0x9e, 0x1d, 0x22, 0x94, 0xd9, 0x21, 0x22, 0x8a, 0xf9, - 0xc0, 0xb8, 0xd4, 0x0d, 0x97, 0x91, 0xc0, 0x33, 0x98, 0x73, 0xc1, 0x6f, 0x79, 0x1e, 0xf3, 0x81, - 0x71, 0x59, 0x4b, 0x90, 0xa3, 0xad, 0x3d, 0x27, 0x22, 0x44, 0xde, 0xf5, 0x0c, 0x2f, 0x9a, 0x32, - 0xd2, 0x7c, 0x24, 0x2e, 0x08, 0x7a, 0x3b, 0x26, 0xa3, 0x9f, 0x40, 0x41, 0xd6, 0x8a, 0x7c, 0xe3, - 0xc6, 0xfd, 0xe6, 0x9b, 0xf3, 0x56, 0x49, 0xfc, 0xca, 0x5d, 0x0d, 0x93, 0xcb, 0x10, 0xbd, 0x0b, - 0x79, 0x63, 0x68, 0x39, 0x34, 0x2a, 0x46, 0xf3, 0x95, 0x68, 0x31, 0x8b, 0x38, 0xc7, 0x69, 0x5d, - 0x4e, 0x52, 0xdf, 0x87, 0x95, 0x09, 0x0c, 0x54, 0x86, 0x25, 0xbf, 0x1f, 0x18, 0xa1, 0xbc, 0xf3, - 0xb2, 0x38, 0x5e, 0xaa, 0xbf, 0x50, 0x60, 0xed, 0x8a, 0x40, 0xa2, 0x9f, 0x42, 0x7e, 0x22, 0x32, - 0xe2, 0x34, 0x8f, 0xe7, 0x75, 0x02, 0x13, 0x93, 0xda, 0x9e, 0x13, 0x41, 0x26, 0x22, 0x8a, 0x27, - 0xb0, 0xd5, 0xbf, 0x28, 0xf0, 0xf0, 0x6d, 0xdb, 0xa3, 0xe9, 0x73, 0xdc, 0x79, 0xe4, 0x61, 0x26, - 0x28, 0x37, 0x4d, 0xa7, 0xe8, 0x04, 0x16, 0x7f, 0x46, 0x03, 0x2b, 0xbe, 0x6c, 0x0e, 0x6e, 0xef, - 0xc5, 0xf7, 0x69, 0x60, 0x35, 0xbc, 0x1e, 0xc5, 0x02, 0x40, 0xfd, 0x83, 0x02, 0xcb, 0x31, 0x0d, - 0x7d, 0x6b, 0xa2, 0x14, 0x6f, 0x9c, 0x86, 0x12, 0x95, 0xf8, 0x01, 0x2c, 0xc7, 0xcf, 0x8b, 0x9b, - 0xdf, 0x15, 0x4b, 0xf2, 0x5d, 0x11, 0xcd, 0x11, 0x91, 0x15, 0x72, 0x24, 0xe6, 0xdf, 0x4f, 0x7e, - 0xaf, 0xc0, 0x92, 0xfc, 0xd1, 0x82, 0x36, 0x61, 0xed, 0x58, 0xab, 0x75, 0x9f, 0x63, 0x4d, 0x7f, - 0xde, 0xec, 0xb4, 0xb5, 0x7a, 0xe3, 0xb8, 0xa1, 0x1d, 0x15, 0xbf, 0x84, 0xd6, 0xa0, 0x70, 0x5a, - 0x3b, 0xd4, 0x4e, 0xf5, 0x23, 0xad, 0xab, 0xd5, 0xbb, 0x8d, 0x56, 0xb3, 0xa8, 0xa0, 0x2d, 0x58, - 0xef, 0x9c, 0xb4, 0xba, 0x7a, 0xfd, 0xa4, 0xd6, 0xfc, 0x54, 0x4b, 0xb0, 0x52, 0x68, 0x07, 0xb6, - 0xb5, 0x1f, 0xb4, 0x4f, 0x1b, 0xf5, 0x46, 0x57, 0xaf, 0xb7, 0x9a, 0x5d, 0xad, 0xd9, 0x4d, 0xf0, - 0xd3, 0xa8, 0x0c, 0xa5, 0x4e, 0x5b, 0xd3, 0xea, 0x27, 0x7a, 0x17, 0xd7, 0x9a, 0x9d, 0x3a, 0x6e, - 0xb4, 0x39, 0x27, 0xf3, 0x24, 0x00, 0x34, 0xfb, 0xce, 0x44, 0x5f, 0x86, 0xdd, 0x29, 0xfd, 0xfa, - 0xb3, 0xd6, 0xd1, 0xb4, 0x95, 0x2b, 0x90, 0xe5, 0x06, 0x45, 0xac, 0xa2, 0x82, 0x56, 0x01, 0x8e, - 0x71, 0xed, 0x99, 0x26, 0xd6, 0xa9, 0xc8, 0x3b, 0xce, 0xae, 0x35, 0x8f, 0xf4, 0x04, 0x23, 0xfd, - 0x84, 0x01, 0x8c, 0x47, 0x3f, 0xb4, 0x0d, 0x1b, 0xa7, 0x8d, 0xef, 0x69, 0xa7, 0x8d, 0x93, 0x56, - 0xeb, 0x68, 0x4a, 0xc3, 0x03, 0x58, 0x39, 0xd3, 0xf0, 0x0b, 0xfd, 0x79, 0x93, 0x6f, 0x79, 0x51, - 0x54, 0x50, 0x1e, 0x96, 0x47, 0xab, 0x54, 0xb4, 0x6a, 0xb7, 0x3a, 0x9d, 0xc6, 0xe1, 0xa9, 0x56, - 0x4c, 0x23, 0x80, 0x8c, 0xe4, 0x2c, 0xa0, 0x02, 0xe4, 0xb8, 0xa8, 0x24, 0x2c, 0x1e, 0xfc, 0x4d, - 0x81, 0x32, 0xef, 0x58, 0x8d, 0x44, 0xe6, 0x74, 0x48, 0x70, 0xe1, 0x98, 0x04, 0xfd, 0x51, 0x81, - 0x95, 0x89, 0x66, 0x89, 0xe6, 0x18, 0x32, 0xaf, 0xfa, 0x97, 0xb7, 0xfd, 0x28, 0x96, 0x4f, 0xfc, - 0x64, 0xac, 0xb4, 0xe2, 0x9f, 0x8c, 0xea, 0x57, 0x7e, 0xf9, 0xef, 0xff, 0xbd, 0x49, 0x3d, 0x56, - 0xb7, 0xa7, 0xff, 0x7b, 0x86, 0x4f, 0x65, 0xf7, 0x24, 0x4f, 0x95, 0x27, 0x87, 0x9f, 0xa7, 0xe0, - 0x6b, 0x26, 0x1d, 0xdc, 0xda, 0x96, 0xc3, 0x47, 0xd7, 0xb9, 0xd8, 0x8e, 0xd2, 0xb4, 0xad, 0xfc, - 0xf0, 0x85, 0x84, 0xb2, 0x69, 0xd4, 0x5c, 0x2b, 0x34, 0xb0, 0xab, 0x36, 0xf1, 0x78, 0x12, 0x57, - 0x05, 0xcb, 0xf0, 0x9d, 0xf0, 0xe6, 0xbf, 0xb4, 0x1f, 0xcf, 0xf0, 0xfe, 0x9a, 0xda, 0xfb, 0x54, - 0x60, 0xd7, 0xb9, 0x99, 0x33, 0x96, 0x54, 0xce, 0xf6, 0xdb, 0xfb, 0x87, 0x91, 0xf0, 0xbf, 0xe2, - 0xad, 0x2f, 0xf9, 0xd6, 0x97, 0x33, 0x5b, 0x5f, 0x9e, 0xc5, 0x7a, 0xce, 0x33, 0xdc, 0xb6, 0x6f, - 0x7c, 0x11, 0x00, 0x00, 0xff, 0xff, 0xf4, 0xad, 0xf2, 0x59, 0x40, 0x16, 0x00, 0x00, + 0x82, 0x94, 0x1d, 0x90, 0x4d, 0xc2, 0x6a, 0x61, 0x67, 0x61, 0xb5, 0x8e, 0xd3, 0xd9, 0x58, 0x64, + 0x62, 0xab, 0xec, 0x09, 0x0c, 0x0c, 0x6a, 0x3a, 0xdd, 0xe5, 0x4e, 0x33, 0xed, 0xae, 0xde, 0xee, + 0xea, 0x30, 0x73, 0x5b, 0x3e, 0x24, 0x90, 0x40, 0x5c, 0x46, 0x48, 0xfc, 0x09, 0x88, 0x23, 0xff, + 0x00, 0x17, 0x2e, 0x70, 0xe5, 0x84, 0x04, 0x37, 0xfe, 0x0b, 0x2e, 0xab, 0xae, 0xaa, 0xb6, 0xdb, + 0x76, 0xbe, 0x9c, 0xdc, 0xdc, 0xef, 0xe3, 0xf7, 0x3e, 0xea, 0xd5, 0x7b, 0xaf, 0x0c, 0x0d, 0x9b, + 0x52, 0xdb, 0x25, 0x75, 0xd3, 0xa5, 0x91, 0x55, 0xbf, 0x70, 0x2c, 0x42, 0x1d, 0x8f, 0x11, 0xd7, + 0x75, 0x6c, 0xe2, 0x99, 0xa4, 0x7e, 0xb1, 0xeb, 0xef, 0x9e, 0x11, 0x66, 0xec, 0x0a, 0x9e, 0x9e, + 0x66, 0xd6, 0xfc, 0x80, 0x32, 0x8a, 0x76, 0x04, 0x44, 0x8d, 0x43, 0xd4, 0xa6, 0x20, 0x6a, 0x43, + 0x88, 0xcd, 0x87, 0xd2, 0x98, 0xe1, 0x3b, 0x75, 0xc3, 0xf3, 0x28, 0x33, 0x98, 0x43, 0xbd, 0x50, + 0xe0, 0x6c, 0xbe, 0x27, 0xb9, 0x2e, 0xf5, 0xec, 0x20, 0xf2, 0x3c, 0xc7, 0xb3, 0xeb, 0xd4, 0x27, + 0xc1, 0x98, 0xd0, 0x96, 0x14, 0xe2, 0x5f, 0x67, 0x51, 0xbf, 0x6e, 0x45, 0x42, 0x40, 0xf2, 0x1f, + 0x4f, 0xf2, 0x99, 0x33, 0x20, 0x21, 0x33, 0x06, 0xbe, 0x14, 0x58, 0x97, 0x02, 0x81, 0x6f, 0xd6, + 0x43, 0x66, 0xb0, 0x48, 0x22, 0xab, 0x7f, 0xcf, 0x40, 0xa5, 0x21, 0x9c, 0x22, 0xa7, 0x71, 0x10, + 0x98, 0x7c, 0x1e, 0x91, 0x90, 0xa1, 0x77, 0x20, 0xef, 0x78, 0x7e, 0xc4, 0xf4, 0x28, 0x70, 0xaa, + 0xca, 0xb6, 0xb2, 0x93, 0xc7, 0x8b, 0x9c, 0xf0, 0x3c, 0x70, 0xd0, 0x7b, 0xb0, 0x24, 0x98, 0x26, + 0xf5, 0x18, 0xf1, 0x58, 0x35, 0xb7, 0xad, 0xec, 0x14, 0x71, 0x91, 0x13, 0x9b, 0x82, 0x86, 0x9e, + 0xc1, 0x62, 0x9f, 0x18, 0x2c, 0x0a, 0x48, 0x58, 0xcd, 0x6c, 0x67, 0x77, 0x96, 0xf7, 0x76, 0x6b, + 0xb7, 0x4d, 0x5a, 0xed, 0x50, 0x68, 0xe2, 0x21, 0x04, 0xfa, 0x31, 0x2c, 0x89, 0xc3, 0xe0, 0x36, + 0x5f, 0xb3, 0x6a, 0x76, 0x5b, 0xd9, 0x29, 0xec, 0x7d, 0x78, 0x7b, 0x4c, 0x1e, 0x5f, 0x53, 0x68, + 0xe3, 0xe2, 0x45, 0xea, 0x0b, 0x3d, 0x02, 0xa0, 0x11, 0x4b, 0xc2, 0x9d, 0xe3, 0xe1, 0xe6, 0x05, + 0x25, 0x8e, 0xf7, 0x31, 0x14, 0x5c, 0x6a, 0xf2, 0x8c, 0xeb, 0x8e, 0x55, 0x9d, 0xe7, 0x7c, 0x48, + 0x48, 0x2d, 0x4b, 0xfd, 0xff, 0x1c, 0x14, 0xd3, 0xf0, 0x08, 0xc3, 0x62, 0x48, 0xec, 0x01, 0xf1, + 0x58, 0x58, 0x55, 0xb6, 0xb3, 0x77, 0x70, 0xb4, 0x2b, 0xd4, 0xf1, 0x10, 0x07, 0x31, 0x58, 0x73, + 0x8d, 0x33, 0xe2, 0xea, 0x16, 0x61, 0xc4, 0xe4, 0xce, 0x98, 0xd4, 0xeb, 0x3b, 0x76, 0x35, 0xc3, + 0x53, 0xf1, 0xc9, 0xed, 0x2d, 0x1c, 0xc7, 0x38, 0x07, 0x09, 0x4c, 0x93, 0xa3, 0xe0, 0x8a, 0x7b, + 0x09, 0x15, 0xfd, 0x5a, 0x81, 0x87, 0xe1, 0x39, 0x65, 0xba, 0x79, 0x6e, 0x78, 0x36, 0x99, 0x36, + 0x2e, 0xce, 0xa1, 0x79, 0x7b, 0xe3, 0xdd, 0x73, 0xca, 0x9a, 0x1c, 0x6c, 0xd2, 0x83, 0x8d, 0xf0, + 0x2a, 0x16, 0x7a, 0xab, 0xc0, 0xbb, 0xe4, 0xb5, 0xef, 0x3a, 0xa6, 0x33, 0x2c, 0xbb, 0x69, 0x5f, + 0xe6, 0xb8, 0x2f, 0x47, 0xb7, 0xf7, 0x45, 0x93, 0x90, 0xb2, 0x68, 0x27, 0x1d, 0xda, 0x22, 0xd7, + 0xf2, 0xd1, 0xaf, 0x14, 0x78, 0x27, 0xf4, 0x09, 0x31, 0xcf, 0x75, 0x16, 0x18, 0x5e, 0x68, 0x06, + 0x8e, 0x9f, 0xf6, 0x27, 0x37, 0x73, 0x6e, 0x38, 0x58, 0x2f, 0x8d, 0x35, 0xcc, 0xcd, 0x55, 0x2c, + 0xf5, 0x1f, 0x0a, 0x54, 0x2e, 0x3b, 0x51, 0xe4, 0x41, 0x65, 0xb2, 0x62, 0x06, 0xd4, 0x22, 0xfc, + 0x3e, 0x2f, 0xef, 0x7d, 0xf7, 0xae, 0xf5, 0xf2, 0x8c, 0x5a, 0x04, 0x23, 0x77, 0x8a, 0x86, 0xbe, + 0x0e, 0x0f, 0x42, 0xd1, 0xde, 0x8c, 0xe0, 0x8d, 0x6e, 0x1a, 0x03, 0x12, 0x18, 0xbc, 0x38, 0x17, + 0x71, 0x79, 0xc4, 0x68, 0x72, 0x3a, 0xaa, 0xc0, 0x7c, 0xec, 0x8c, 0xcb, 0x0b, 0x28, 0x8f, 0xc5, + 0x87, 0xba, 0x0b, 0x1b, 0x57, 0xd6, 0xc7, 0x48, 0x45, 0x49, 0xab, 0x7c, 0x08, 0x5b, 0xd7, 0x1f, + 0xe3, 0x15, 0x7a, 0x7f, 0x52, 0xe4, 0xa5, 0x95, 0x57, 0x0d, 0x69, 0xdc, 0xfd, 0x80, 0xe9, 0x71, + 0xfb, 0xd4, 0x69, 0xbf, 0x1f, 0x12, 0xc6, 0x55, 0x0a, 0x7b, 0x1b, 0x49, 0xae, 0x92, 0x16, 0x5b, + 0x3b, 0x90, 0x2d, 0x18, 0x97, 0xb8, 0x4e, 0xcf, 0x19, 0x90, 0x36, 0xd7, 0x40, 0x0d, 0x28, 0x11, + 0xcf, 0x1a, 0x03, 0xc9, 0xdc, 0x04, 0xb2, 0x44, 0x3c, 0x6b, 0x04, 0xa1, 0x7e, 0xa1, 0x40, 0x91, + 0xe7, 0x3c, 0x71, 0xad, 0x03, 0x0b, 0xb2, 0x0f, 0x48, 0x87, 0xee, 0xda, 0x4e, 0x12, 0x18, 0xb4, + 0x05, 0xc0, 0x8b, 0xd4, 0x8a, 0xa5, 0xb9, 0x83, 0x19, 0x9c, 0xa2, 0xa8, 0xe7, 0x00, 0xdc, 0x83, + 0xc3, 0xc0, 0x18, 0x10, 0xf4, 0x14, 0x0a, 0x33, 0x25, 0x05, 0xd8, 0x28, 0x1f, 0x37, 0x59, 0x72, + 0x21, 0xa7, 0x79, 0xcc, 0x61, 0x6f, 0xe2, 0xa1, 0x43, 0xf8, 0xaf, 0xb8, 0xcb, 0xca, 0xa1, 0x23, + 0x08, 0x2d, 0x0b, 0x6d, 0x43, 0xc1, 0x22, 0xc3, 0xd2, 0xe7, 0x38, 0x79, 0x9c, 0x26, 0xc5, 0x63, + 0xc9, 0x35, 0x3c, 0x3b, 0x32, 0x6c, 0xa2, 0x9b, 0x71, 0x9d, 0x8b, 0xca, 0x2a, 0x26, 0xc4, 0x26, + 0xb5, 0x88, 0xfa, 0x9f, 0x0c, 0x94, 0x78, 0x60, 0x8d, 0xe1, 0x2c, 0x46, 0x47, 0x90, 0x13, 0x66, + 0x64, 0x60, 0xdf, 0x9c, 0xa1, 0x81, 0x70, 0x3d, 0x2c, 0xf5, 0xd1, 0x4f, 0xe0, 0x81, 0x69, 0x30, + 0x62, 0xd3, 0xe0, 0x8d, 0xce, 0x49, 0x8e, 0x9c, 0x7e, 0x77, 0x01, 0x2d, 0x27, 0x50, 0x9a, 0x44, + 0x1a, 0x1b, 0x2b, 0xd9, 0x59, 0xc7, 0x4a, 0xba, 0xa0, 0x52, 0x63, 0xe5, 0x18, 0x72, 0xfd, 0xf8, + 0x8c, 0xc3, 0xea, 0x1c, 0x47, 0xfc, 0x60, 0x46, 0x44, 0x5e, 0x20, 0x58, 0x62, 0xa8, 0x7f, 0x53, + 0xa0, 0x32, 0x71, 0x1b, 0xef, 0x5f, 0x41, 0xaf, 0x60, 0xcd, 0xa7, 0x81, 0x47, 0xed, 0xc0, 0xf0, + 0xcf, 0xdf, 0xe8, 0xae, 0xf3, 0x8a, 0xb8, 0xce, 0x39, 0xa5, 0x16, 0xaf, 0x82, 0xe5, 0x99, 0x5c, + 0x1e, 0xea, 0xe2, 0xd5, 0x14, 0xe6, 0x88, 0xac, 0x86, 0xb0, 0x31, 0x11, 0x40, 0xaa, 0x52, 0x4e, + 0x87, 0xc9, 0x12, 0x53, 0xfd, 0x93, 0x3b, 0x8f, 0x9a, 0xf1, 0xb4, 0xfd, 0x26, 0x07, 0x6b, 0xfc, + 0x9e, 0x8e, 0x6c, 0x61, 0x12, 0x46, 0x2e, 0x0b, 0xaf, 0xdf, 0xc4, 0x22, 0xd8, 0x90, 0x07, 0xa9, + 0x8b, 0x4e, 0x9f, 0xda, 0x30, 0x65, 0xdd, 0x7d, 0x34, 0xe3, 0x79, 0xa6, 0x3c, 0x58, 0x97, 0xd8, + 0x13, 0xf4, 0x10, 0x51, 0x58, 0xe3, 0x3b, 0xc1, 0xb4, 0xcd, 0xec, 0x7d, 0x6d, 0x56, 0x62, 0xe0, + 0x29, 0x83, 0x9f, 0xc3, 0x3a, 0xcf, 0xd4, 0x25, 0x16, 0xe7, 0xee, 0x6b, 0x71, 0x95, 0x23, 0x4f, + 0x99, 0x34, 0xa0, 0xcc, 0x63, 0x4c, 0xdb, 0xca, 0xdd, 0x6b, 0x95, 0x2b, 0xc5, 0x78, 0x69, 0x13, + 0x0c, 0x56, 0x86, 0x3b, 0xcd, 0xc8, 0x4c, 0x75, 0x61, 0xd6, 0xad, 0xe1, 0xca, 0x7a, 0xc5, 0x28, + 0xc1, 0x4f, 0xd5, 0x70, 0x00, 0xab, 0x97, 0xed, 0x2c, 0x61, 0xb5, 0xc0, 0xa3, 0xfb, 0xde, 0xbd, + 0xb6, 0x15, 0x5c, 0xb9, 0x64, 0x4f, 0x09, 0xd1, 0x0e, 0xcc, 0x93, 0x20, 0xa0, 0x41, 0x35, 0xcf, + 0x63, 0x43, 0x89, 0x8d, 0xc0, 0x37, 0x6b, 0x5d, 0xfe, 0x20, 0xc1, 0x42, 0x40, 0xfd, 0xad, 0x02, + 0xab, 0x13, 0x2f, 0x92, 0xd0, 0xa7, 0x5e, 0x48, 0x10, 0x05, 0x34, 0x4a, 0x92, 0x1e, 0x88, 0xeb, + 0x21, 0xef, 0xe1, 0xa7, 0x33, 0x1e, 0xc9, 0xd4, 0x35, 0xc3, 0x0f, 0x8c, 0x49, 0x92, 0xfa, 0x6f, + 0x05, 0xd6, 0x27, 0xa4, 0x3b, 0x01, 0xb5, 0x03, 0x12, 0xde, 0x70, 0x2b, 0xdf, 0x87, 0xb2, 0x2f, + 0x05, 0x75, 0x9f, 0x04, 0x66, 0x3c, 0xb6, 0xe3, 0x4e, 0x35, 0x8f, 0x4b, 0x09, 0xbd, 0x23, 0xc8, + 0xe8, 0x23, 0x80, 0xd1, 0xce, 0x21, 0x77, 0xe9, 0xcd, 0xa9, 0xae, 0xd8, 0x4b, 0xde, 0x73, 0x38, + 0x3f, 0xdc, 0x36, 0xd0, 0xc7, 0x50, 0x88, 0x7c, 0xcb, 0x60, 0x44, 0xe8, 0xce, 0xdd, 0xa8, 0x0b, + 0x42, 0x3c, 0x26, 0xa8, 0xbf, 0x9b, 0x4c, 0xf3, 0x30, 0xb2, 0x00, 0x56, 0x52, 0x69, 0x4e, 0xfc, + 0x95, 0x79, 0x6e, 0xdc, 0x39, 0xcf, 0x09, 0x3e, 0x4e, 0x1d, 0x62, 0x42, 0x53, 0xff, 0x9b, 0x81, + 0x8d, 0x2b, 0x57, 0xdf, 0xe9, 0xb9, 0xae, 0x4c, 0xcf, 0xf5, 0x38, 0xe7, 0x03, 0xe3, 0xb5, 0x6e, + 0xb8, 0x8c, 0x04, 0x9e, 0xc1, 0x9c, 0x0b, 0x3e, 0x78, 0x79, 0xce, 0x07, 0xc6, 0xeb, 0x46, 0x8a, + 0x1c, 0x8b, 0xf6, 0x9d, 0x98, 0x10, 0x47, 0xd7, 0x37, 0xbc, 0x78, 0xf0, 0x67, 0xf9, 0x96, 0x5a, + 0x12, 0xf4, 0x4e, 0x42, 0x46, 0x3f, 0x85, 0x92, 0xbc, 0x2b, 0xf2, 0xd9, 0x99, 0xf4, 0x9b, 0x6f, + 0xcf, 0x7a, 0x4b, 0x92, 0x87, 0xe7, 0x72, 0x98, 0xfe, 0x0c, 0xd1, 0xa7, 0xf0, 0x90, 0x78, 0xc6, + 0x99, 0x4b, 0x74, 0x23, 0x62, 0x74, 0x60, 0x30, 0xc7, 0xd4, 0xfd, 0xc8, 0x33, 0x59, 0x24, 0x9a, + 0xc1, 0x3c, 0x77, 0x6c, 0x53, 0xc8, 0x34, 0x12, 0x91, 0xce, 0x48, 0x02, 0xbd, 0x0b, 0x45, 0x23, + 0xb2, 0x1c, 0x1a, 0x5f, 0x67, 0xf3, 0x95, 0x68, 0x52, 0xf3, 0xb8, 0xc0, 0x69, 0x3d, 0x4e, 0x52, + 0xdf, 0x87, 0xa5, 0x31, 0x2f, 0x50, 0x15, 0x16, 0xfc, 0xf3, 0xc0, 0x08, 0xe5, 0x20, 0xcb, 0xe3, + 0xe4, 0x53, 0xfd, 0x85, 0x02, 0x2b, 0x97, 0x1c, 0x05, 0xfa, 0x19, 0x14, 0xc7, 0x72, 0x2b, 0xea, + 0xe1, 0x70, 0xd6, 0x34, 0x60, 0x62, 0x52, 0xdb, 0x73, 0x62, 0xc8, 0xd4, 0x99, 0xe0, 0x31, 0x6c, + 0xf5, 0xcf, 0x0a, 0x3c, 0xbc, 0x4e, 0x3c, 0x5e, 0x29, 0x47, 0xbd, 0x4b, 0x96, 0x43, 0x8a, 0x72, + 0xd3, 0xca, 0x89, 0x8e, 0x60, 0xfe, 0xe7, 0x34, 0xb0, 0x92, 0x71, 0xb5, 0x77, 0xfb, 0x28, 0x7e, + 0x40, 0x03, 0xab, 0xe5, 0xf5, 0x29, 0x16, 0x00, 0xea, 0x1f, 0x14, 0x58, 0x4c, 0x68, 0xe8, 0x3b, + 0x63, 0x97, 0xf9, 0xc6, 0x15, 0x27, 0x75, 0x97, 0x3f, 0x80, 0xc5, 0xe4, 0xcd, 0x70, 0xf3, 0x63, + 0x61, 0x41, 0x3e, 0x16, 0x10, 0x82, 0xb9, 0xd8, 0x0b, 0xb9, 0xe7, 0xf2, 0xdf, 0x4f, 0x7e, 0xaf, + 0xc0, 0x82, 0xfc, 0xf7, 0x04, 0xad, 0xc3, 0xca, 0xa1, 0xd6, 0xe8, 0x3d, 0xc7, 0x9a, 0xfe, 0xfc, + 0xa4, 0xdb, 0xd1, 0x9a, 0xad, 0xc3, 0x96, 0x76, 0x50, 0xfe, 0x0a, 0x5a, 0x81, 0xd2, 0x71, 0x63, + 0x5f, 0x3b, 0xd6, 0x0f, 0xb4, 0x9e, 0xd6, 0xec, 0xb5, 0xda, 0x27, 0x65, 0x05, 0x6d, 0xc0, 0x6a, + 0xf7, 0xa8, 0xdd, 0xd3, 0x9b, 0x47, 0x8d, 0x93, 0xcf, 0xb4, 0x14, 0x2b, 0x83, 0xb6, 0x60, 0x53, + 0xfb, 0x61, 0xe7, 0xb8, 0xd5, 0x6c, 0xf5, 0xf4, 0x66, 0xfb, 0xa4, 0xa7, 0x9d, 0xf4, 0x52, 0xfc, + 0x2c, 0xaa, 0x42, 0xa5, 0xdb, 0xd1, 0xb4, 0xe6, 0x91, 0xde, 0xc3, 0x8d, 0x93, 0x6e, 0x13, 0xb7, + 0x3a, 0x9c, 0x93, 0x7b, 0x12, 0x00, 0x9a, 0x7e, 0x3c, 0xa2, 0xaf, 0xc2, 0xf6, 0x84, 0x7d, 0xfd, + 0x59, 0xfb, 0x60, 0xd2, 0xcb, 0x25, 0xc8, 0x73, 0x87, 0x62, 0x56, 0x59, 0x41, 0xcb, 0x00, 0x87, + 0xb8, 0xf1, 0x4c, 0x13, 0xdf, 0x99, 0x38, 0x3a, 0xce, 0x6e, 0x9c, 0x1c, 0xe8, 0x29, 0x46, 0xf6, + 0x09, 0x03, 0x18, 0xed, 0x73, 0x68, 0x13, 0xd6, 0x8e, 0x5b, 0xdf, 0xd7, 0x8e, 0x5b, 0x47, 0xed, + 0xf6, 0xc1, 0x84, 0x85, 0x07, 0xb0, 0x74, 0xaa, 0xe1, 0x17, 0xfa, 0xf3, 0x13, 0x2e, 0xf2, 0xa2, + 0xac, 0xa0, 0x22, 0x2c, 0x0e, 0xbf, 0x32, 0xf1, 0x57, 0xa7, 0xdd, 0xed, 0xb6, 0xf6, 0x8f, 0xb5, + 0x72, 0x16, 0x01, 0xe4, 0x24, 0x67, 0x0e, 0x95, 0xa0, 0xc0, 0x55, 0x25, 0x61, 0x7e, 0xef, 0xaf, + 0x0a, 0x54, 0x79, 0xcf, 0x6b, 0xa5, 0x2a, 0xa7, 0x4b, 0x82, 0x0b, 0xc7, 0x24, 0xe8, 0x8f, 0x0a, + 0x2c, 0x8d, 0xb5, 0x5b, 0x34, 0xc3, 0xe6, 0x78, 0xd9, 0x1f, 0x74, 0x9b, 0x8f, 0x12, 0xfd, 0xd4, + 0x3f, 0x87, 0xb5, 0x76, 0xf2, 0xcf, 0xa1, 0xfa, 0xb5, 0x5f, 0xfe, 0xeb, 0x7f, 0x6f, 0x33, 0x8f, + 0xd5, 0xcd, 0xc9, 0x3f, 0x33, 0xc3, 0xa7, 0xb2, 0xff, 0x92, 0xa7, 0xca, 0x93, 0xfd, 0x2f, 0x32, + 0xf0, 0x0d, 0x93, 0x0e, 0x6e, 0xed, 0xcb, 0xfe, 0xa3, 0xab, 0x42, 0xec, 0xc4, 0x65, 0xda, 0x51, + 0x7e, 0xf4, 0x42, 0x42, 0xd9, 0x34, 0x6e, 0xcf, 0x35, 0x1a, 0xd8, 0x75, 0x9b, 0x78, 0xbc, 0x88, + 0xeb, 0x82, 0x65, 0xf8, 0x4e, 0x78, 0xf3, 0x5f, 0xaf, 0x1f, 0x4f, 0xf1, 0xfe, 0x92, 0xd9, 0xf9, + 0x4c, 0x60, 0x37, 0xb9, 0x9b, 0x53, 0x9e, 0xd4, 0x4e, 0x77, 0x3b, 0xbb, 0xfb, 0xb1, 0xf2, 0x3f, + 0x13, 0xd1, 0x97, 0x5c, 0xf4, 0xe5, 0x94, 0xe8, 0xcb, 0xd3, 0xc4, 0xce, 0x59, 0x8e, 0xfb, 0xf6, + 0xad, 0x2f, 0x03, 0x00, 0x00, 0xff, 0xff, 0x82, 0x99, 0xe4, 0x9b, 0x15, 0x16, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/geometry.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/geometry.pb.go index 9dcccfc39..17b13c17c 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/geometry.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/geometry.pb.go @@ -22,9 +22,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // NOTE: the vertex coordinates are in the same scale as the original image. type Vertex struct { // X coordinate. - X int32 `protobuf:"varint,1,opt,name=x" json:"x,omitempty"` + X int32 `protobuf:"varint,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y int32 `protobuf:"varint,2,opt,name=y" json:"y,omitempty"` + Y int32 `protobuf:"varint,2,opt,name=y,proto3" json:"y,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -34,7 +34,7 @@ func (m *Vertex) Reset() { *m = Vertex{} } func (m *Vertex) String() string { return proto.CompactTextString(m) } func (*Vertex) ProtoMessage() {} func (*Vertex) Descriptor() ([]byte, []int) { - return fileDescriptor_geometry_8551ce1bb04014e1, []int{0} + return fileDescriptor_geometry_3423d75926342ff3, []int{0} } func (m *Vertex) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Vertex.Unmarshal(m, b) @@ -73,9 +73,9 @@ func (m *Vertex) GetY() int32 { // and range from 0 to 1. type NormalizedVertex struct { // X coordinate. - X float32 `protobuf:"fixed32,1,opt,name=x" json:"x,omitempty"` + X float32 `protobuf:"fixed32,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y float32 `protobuf:"fixed32,2,opt,name=y" json:"y,omitempty"` + Y float32 `protobuf:"fixed32,2,opt,name=y,proto3" json:"y,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -85,7 +85,7 @@ func (m *NormalizedVertex) Reset() { *m = NormalizedVertex{} } func (m *NormalizedVertex) String() string { return proto.CompactTextString(m) } func (*NormalizedVertex) ProtoMessage() {} func (*NormalizedVertex) Descriptor() ([]byte, []int) { - return fileDescriptor_geometry_8551ce1bb04014e1, []int{1} + return fileDescriptor_geometry_3423d75926342ff3, []int{1} } func (m *NormalizedVertex) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NormalizedVertex.Unmarshal(m, b) @@ -122,9 +122,9 @@ func (m *NormalizedVertex) GetY() float32 { // A bounding polygon for the detected image annotation. type BoundingPoly struct { // The bounding polygon vertices. - Vertices []*Vertex `protobuf:"bytes,1,rep,name=vertices" json:"vertices,omitempty"` + Vertices []*Vertex `protobuf:"bytes,1,rep,name=vertices,proto3" json:"vertices,omitempty"` // The bounding polygon normalized vertices. - NormalizedVertices []*NormalizedVertex `protobuf:"bytes,2,rep,name=normalized_vertices,json=normalizedVertices" json:"normalized_vertices,omitempty"` + NormalizedVertices []*NormalizedVertex `protobuf:"bytes,2,rep,name=normalized_vertices,json=normalizedVertices,proto3" json:"normalized_vertices,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -134,7 +134,7 @@ func (m *BoundingPoly) Reset() { *m = BoundingPoly{} } func (m *BoundingPoly) String() string { return proto.CompactTextString(m) } func (*BoundingPoly) ProtoMessage() {} func (*BoundingPoly) Descriptor() ([]byte, []int) { - return fileDescriptor_geometry_8551ce1bb04014e1, []int{2} + return fileDescriptor_geometry_3423d75926342ff3, []int{2} } func (m *BoundingPoly) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BoundingPoly.Unmarshal(m, b) @@ -173,11 +173,11 @@ func (m *BoundingPoly) GetNormalizedVertices() []*NormalizedVertex { // The position coordinates are in the same scale as the original image. type Position struct { // X coordinate. - X float32 `protobuf:"fixed32,1,opt,name=x" json:"x,omitempty"` + X float32 `protobuf:"fixed32,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y float32 `protobuf:"fixed32,2,opt,name=y" json:"y,omitempty"` + Y float32 `protobuf:"fixed32,2,opt,name=y,proto3" json:"y,omitempty"` // Z coordinate (or depth). - Z float32 `protobuf:"fixed32,3,opt,name=z" json:"z,omitempty"` + Z float32 `protobuf:"fixed32,3,opt,name=z,proto3" json:"z,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -187,7 +187,7 @@ func (m *Position) Reset() { *m = Position{} } func (m *Position) String() string { return proto.CompactTextString(m) } func (*Position) ProtoMessage() {} func (*Position) Descriptor() ([]byte, []int) { - return fileDescriptor_geometry_8551ce1bb04014e1, []int{3} + return fileDescriptor_geometry_3423d75926342ff3, []int{3} } func (m *Position) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Position.Unmarshal(m, b) @@ -236,27 +236,27 @@ func init() { } func init() { - proto.RegisterFile("google/cloud/vision/v1/geometry.proto", fileDescriptor_geometry_8551ce1bb04014e1) + proto.RegisterFile("google/cloud/vision/v1/geometry.proto", fileDescriptor_geometry_3423d75926342ff3) } -var fileDescriptor_geometry_8551ce1bb04014e1 = []byte{ - // 276 bytes of a gzipped FileDescriptorProto +var fileDescriptor_geometry_3423d75926342ff3 = []byte{ + // 285 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x7c, 0x91, 0x41, 0x4b, 0xc3, 0x30, - 0x14, 0xc7, 0x49, 0x87, 0x63, 0xc4, 0x09, 0x52, 0x41, 0x8a, 0x07, 0x19, 0x45, 0xa1, 0xa7, 0x84, - 0xa9, 0x27, 0xf5, 0xd4, 0x8b, 0x37, 0x29, 0x3d, 0x08, 0x7a, 0x91, 0xda, 0x86, 0x10, 0x48, 0xdf, - 0x1b, 0x69, 0x56, 0xd6, 0x7e, 0x1f, 0xbf, 0xa3, 0x47, 0x69, 0x32, 0x0a, 0x2b, 0xce, 0x5b, 0xfe, - 0xc9, 0x8f, 0xdf, 0x7b, 0x79, 0x8f, 0xde, 0x4a, 0x44, 0xa9, 0x05, 0x2f, 0x35, 0x6e, 0x2b, 0xde, - 0xaa, 0x46, 0x21, 0xf0, 0x76, 0xcd, 0xa5, 0xc0, 0x5a, 0x58, 0xd3, 0xb1, 0x8d, 0x41, 0x8b, 0xe1, - 0xa5, 0xc7, 0x98, 0xc3, 0x98, 0xc7, 0x58, 0xbb, 0x8e, 0x6f, 0xe8, 0xfc, 0x4d, 0x18, 0x2b, 0x76, - 0xe1, 0x92, 0x92, 0x5d, 0x44, 0x56, 0x24, 0x39, 0xc9, 0x89, 0x4b, 0x5d, 0x14, 0xf8, 0xd4, 0xc5, - 0x8c, 0x9e, 0xbf, 0xa2, 0xa9, 0x0b, 0xad, 0x7a, 0x51, 0x4d, 0xf9, 0xe0, 0x80, 0x0f, 0x06, 0xfe, - 0x9b, 0xd0, 0x65, 0x8a, 0x5b, 0xa8, 0x14, 0xc8, 0x0c, 0x75, 0x17, 0x3e, 0xd2, 0x45, 0x2b, 0x8c, - 0x55, 0xa5, 0x68, 0x22, 0xb2, 0x9a, 0x25, 0xa7, 0x77, 0xd7, 0xec, 0xef, 0x8e, 0x98, 0xd7, 0xe7, - 0x23, 0x1f, 0xbe, 0xd3, 0x0b, 0x18, 0x8b, 0x7f, 0x8e, 0x9a, 0xc0, 0x69, 0x92, 0x63, 0x9a, 0x69, - 0xbf, 0x79, 0x08, 0x07, 0x37, 0x83, 0x23, 0x7e, 0xa0, 0x8b, 0x0c, 0x1b, 0x65, 0x15, 0xc2, 0x7f, - 0xff, 0x19, 0x52, 0x1f, 0xcd, 0x7c, 0xea, 0x53, 0xa0, 0x57, 0x25, 0xd6, 0x47, 0x0a, 0xa7, 0x67, - 0x2f, 0xfb, 0xc9, 0x67, 0xc3, 0xe0, 0x33, 0xf2, 0xf1, 0xbc, 0x07, 0x25, 0xea, 0x02, 0x24, 0x43, - 0x23, 0xb9, 0x14, 0xe0, 0xd6, 0xc2, 0xfd, 0x53, 0xb1, 0x51, 0xcd, 0x74, 0x81, 0x4f, 0xfe, 0xf4, - 0x43, 0xc8, 0xd7, 0xdc, 0xb1, 0xf7, 0xbf, 0x01, 0x00, 0x00, 0xff, 0xff, 0x6f, 0xd6, 0x4f, 0x3a, - 0xeb, 0x01, 0x00, 0x00, + 0x14, 0xc7, 0x49, 0xa7, 0x63, 0xc4, 0x09, 0x52, 0x41, 0x8a, 0x07, 0x19, 0x45, 0xa1, 0xa7, 0x84, + 0xa9, 0x27, 0xf5, 0x54, 0x0f, 0xbb, 0x8d, 0xd2, 0xc3, 0x40, 0x2f, 0x52, 0xdb, 0x10, 0x02, 0x69, + 0xde, 0x48, 0xd3, 0xb2, 0xf6, 0xe3, 0x08, 0x7e, 0x47, 0x8f, 0xd2, 0x64, 0x14, 0x56, 0x9c, 0xb7, + 0xfe, 0x9b, 0x5f, 0x7e, 0xef, 0xe5, 0x3d, 0x7c, 0xc7, 0x01, 0xb8, 0x64, 0x34, 0x97, 0x50, 0x17, + 0xb4, 0x11, 0x95, 0x00, 0x45, 0x9b, 0x25, 0xe5, 0x0c, 0x4a, 0x66, 0x74, 0x4b, 0xb6, 0x1a, 0x0c, + 0xf8, 0x57, 0x0e, 0x23, 0x16, 0x23, 0x0e, 0x23, 0xcd, 0x32, 0xbc, 0xc5, 0xd3, 0x0d, 0xd3, 0x86, + 0xed, 0xfc, 0x39, 0x46, 0xbb, 0x00, 0x2d, 0x50, 0x74, 0x9a, 0x22, 0x9b, 0xda, 0xc0, 0x73, 0xa9, + 0x0d, 0x09, 0xbe, 0x58, 0x83, 0x2e, 0x33, 0x29, 0x3a, 0x56, 0x8c, 0x79, 0xef, 0x80, 0xf7, 0x7a, + 0xfe, 0x1b, 0xe1, 0x79, 0x0c, 0xb5, 0x2a, 0x84, 0xe2, 0x09, 0xc8, 0xd6, 0x7f, 0xc2, 0xb3, 0x86, + 0x69, 0x23, 0x72, 0x56, 0x05, 0x68, 0x31, 0x89, 0xce, 0xee, 0x6f, 0xc8, 0xdf, 0x1d, 0x11, 0xa7, + 0x4f, 0x07, 0xde, 0x7f, 0xc3, 0x97, 0x6a, 0x28, 0xfe, 0x31, 0x68, 0x3c, 0xab, 0x89, 0x8e, 0x69, + 0xc6, 0xfd, 0xa6, 0xbe, 0x3a, 0xf8, 0xd3, 0x3b, 0xc2, 0x47, 0x3c, 0x4b, 0xa0, 0x12, 0x46, 0x80, + 0xfa, 0xef, 0x3d, 0x7d, 0xea, 0x82, 0x89, 0x4b, 0x5d, 0x5c, 0xe3, 0xeb, 0x1c, 0xca, 0x23, 0x85, + 0xe3, 0xf3, 0xd5, 0x7e, 0xf2, 0x49, 0x3f, 0xf8, 0x04, 0xbd, 0xbf, 0xec, 0x41, 0x0e, 0x32, 0x53, + 0x9c, 0x80, 0xe6, 0x94, 0x33, 0x65, 0xd7, 0x42, 0xdd, 0x51, 0xb6, 0x15, 0xd5, 0x78, 0x81, 0xcf, + 0xee, 0xeb, 0x07, 0xa1, 0x2f, 0xef, 0x64, 0xf5, 0xba, 0x59, 0x7f, 0x4e, 0xed, 0x95, 0x87, 0xdf, + 0x00, 0x00, 0x00, 0xff, 0xff, 0xf1, 0x3c, 0x20, 0x46, 0xf2, 0x01, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/image_annotator.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/image_annotator.pb.go index 56c0e2589..00cb906e4 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/image_annotator.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/image_annotator.pb.go @@ -69,7 +69,7 @@ func (x Likelihood) String() string { return proto.EnumName(Likelihood_name, int32(x)) } func (Likelihood) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{0} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{0} } // Type of Google Cloud Vision API feature to be extracted. @@ -136,7 +136,7 @@ func (x Feature_Type) String() string { return proto.EnumName(Feature_Type_name, int32(x)) } func (Feature_Type) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{0, 0} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{0, 0} } // Face landmark (feature) type. @@ -297,7 +297,7 @@ func (x FaceAnnotation_Landmark_Type) String() string { return proto.EnumName(FaceAnnotation_Landmark_Type_name, int32(x)) } func (FaceAnnotation_Landmark_Type) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{3, 0, 0} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{3, 0, 0} } // Batch operation states. @@ -335,7 +335,7 @@ func (x OperationMetadata_State) String() string { return proto.EnumName(OperationMetadata_State_name, int32(x)) } func (OperationMetadata_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{31, 0} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{31, 0} } // The type of Google Cloud Vision API detection to perform, and the maximum @@ -343,14 +343,14 @@ func (OperationMetadata_State) EnumDescriptor() ([]byte, []int) { // be specified in the `features` list. type Feature struct { // The feature type. - Type Feature_Type `protobuf:"varint,1,opt,name=type,enum=google.cloud.vision.v1.Feature_Type" json:"type,omitempty"` + Type Feature_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.vision.v1.Feature_Type" json:"type,omitempty"` // Maximum number of results of this type. Does not apply to // `TEXT_DETECTION`, `DOCUMENT_TEXT_DETECTION`, or `CROP_HINTS`. - MaxResults int32 `protobuf:"varint,2,opt,name=max_results,json=maxResults" json:"max_results,omitempty"` + MaxResults int32 `protobuf:"varint,2,opt,name=max_results,json=maxResults,proto3" json:"max_results,omitempty"` // Model to use for the feature. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -360,7 +360,7 @@ func (m *Feature) Reset() { *m = Feature{} } func (m *Feature) String() string { return proto.CompactTextString(m) } func (*Feature) ProtoMessage() {} func (*Feature) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{0} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{0} } func (m *Feature) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Feature.Unmarshal(m, b) @@ -409,7 +409,7 @@ type ImageSource struct { // `gs://bucket_name/object_name`. Object versioning is not supported. See // [Google Cloud Storage Request // URIs](https://cloud.google.com/storage/docs/reference-uris) for more info. - GcsImageUri string `protobuf:"bytes,1,opt,name=gcs_image_uri,json=gcsImageUri" json:"gcs_image_uri,omitempty"` + GcsImageUri string `protobuf:"bytes,1,opt,name=gcs_image_uri,json=gcsImageUri,proto3" json:"gcs_image_uri,omitempty"` // The URI of the source image. Can be either: // // 1. A Google Cloud Storage URI of the form @@ -427,7 +427,7 @@ type ImageSource struct { // // When both `gcs_image_uri` and `image_uri` are specified, `image_uri` takes // precedence. - ImageUri string `protobuf:"bytes,2,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,2,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -437,7 +437,7 @@ func (m *ImageSource) Reset() { *m = ImageSource{} } func (m *ImageSource) String() string { return proto.CompactTextString(m) } func (*ImageSource) ProtoMessage() {} func (*ImageSource) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{1} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{1} } func (m *ImageSource) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ImageSource.Unmarshal(m, b) @@ -480,7 +480,7 @@ type Image struct { // Google Cloud Storage image location, or publicly-accessible image // URL. If both `content` and `source` are provided for an image, `content` // takes precedence and is used to perform the image annotation request. - Source *ImageSource `protobuf:"bytes,2,opt,name=source" json:"source,omitempty"` + Source *ImageSource `protobuf:"bytes,2,opt,name=source,proto3" json:"source,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -490,7 +490,7 @@ func (m *Image) Reset() { *m = Image{} } func (m *Image) String() string { return proto.CompactTextString(m) } func (*Image) ProtoMessage() {} func (*Image) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{2} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{2} } func (m *Image) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Image.Unmarshal(m, b) @@ -533,45 +533,45 @@ type FaceAnnotation struct { // Note that one or more x and/or y coordinates may not be generated in the // `BoundingPoly` (the polygon will be unbounded) if only a partial face // appears in the image to be annotated. - BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // The `fd_bounding_poly` bounding polygon is tighter than the // `boundingPoly`, and encloses only the skin part of the face. Typically, it // is used to eliminate the face from any image analysis that detects the // "amount of skin" visible in an image. It is not based on the // landmarker results, only on the initial face detection, hence // the fd (face detection) prefix. - FdBoundingPoly *BoundingPoly `protobuf:"bytes,2,opt,name=fd_bounding_poly,json=fdBoundingPoly" json:"fd_bounding_poly,omitempty"` + FdBoundingPoly *BoundingPoly `protobuf:"bytes,2,opt,name=fd_bounding_poly,json=fdBoundingPoly,proto3" json:"fd_bounding_poly,omitempty"` // Detected face landmarks. - Landmarks []*FaceAnnotation_Landmark `protobuf:"bytes,3,rep,name=landmarks" json:"landmarks,omitempty"` + Landmarks []*FaceAnnotation_Landmark `protobuf:"bytes,3,rep,name=landmarks,proto3" json:"landmarks,omitempty"` // Roll angle, which indicates the amount of clockwise/anti-clockwise rotation // of the face relative to the image vertical about the axis perpendicular to // the face. Range [-180,180]. - RollAngle float32 `protobuf:"fixed32,4,opt,name=roll_angle,json=rollAngle" json:"roll_angle,omitempty"` + RollAngle float32 `protobuf:"fixed32,4,opt,name=roll_angle,json=rollAngle,proto3" json:"roll_angle,omitempty"` // Yaw angle, which indicates the leftward/rightward angle that the face is // pointing relative to the vertical plane perpendicular to the image. Range // [-180,180]. - PanAngle float32 `protobuf:"fixed32,5,opt,name=pan_angle,json=panAngle" json:"pan_angle,omitempty"` + PanAngle float32 `protobuf:"fixed32,5,opt,name=pan_angle,json=panAngle,proto3" json:"pan_angle,omitempty"` // Pitch angle, which indicates the upwards/downwards angle that the face is // pointing relative to the image's horizontal plane. Range [-180,180]. - TiltAngle float32 `protobuf:"fixed32,6,opt,name=tilt_angle,json=tiltAngle" json:"tilt_angle,omitempty"` + TiltAngle float32 `protobuf:"fixed32,6,opt,name=tilt_angle,json=tiltAngle,proto3" json:"tilt_angle,omitempty"` // Detection confidence. Range [0, 1]. - DetectionConfidence float32 `protobuf:"fixed32,7,opt,name=detection_confidence,json=detectionConfidence" json:"detection_confidence,omitempty"` + DetectionConfidence float32 `protobuf:"fixed32,7,opt,name=detection_confidence,json=detectionConfidence,proto3" json:"detection_confidence,omitempty"` // Face landmarking confidence. Range [0, 1]. - LandmarkingConfidence float32 `protobuf:"fixed32,8,opt,name=landmarking_confidence,json=landmarkingConfidence" json:"landmarking_confidence,omitempty"` + LandmarkingConfidence float32 `protobuf:"fixed32,8,opt,name=landmarking_confidence,json=landmarkingConfidence,proto3" json:"landmarking_confidence,omitempty"` // Joy likelihood. - JoyLikelihood Likelihood `protobuf:"varint,9,opt,name=joy_likelihood,json=joyLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"joy_likelihood,omitempty"` + JoyLikelihood Likelihood `protobuf:"varint,9,opt,name=joy_likelihood,json=joyLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"joy_likelihood,omitempty"` // Sorrow likelihood. - SorrowLikelihood Likelihood `protobuf:"varint,10,opt,name=sorrow_likelihood,json=sorrowLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"sorrow_likelihood,omitempty"` + SorrowLikelihood Likelihood `protobuf:"varint,10,opt,name=sorrow_likelihood,json=sorrowLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"sorrow_likelihood,omitempty"` // Anger likelihood. - AngerLikelihood Likelihood `protobuf:"varint,11,opt,name=anger_likelihood,json=angerLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"anger_likelihood,omitempty"` + AngerLikelihood Likelihood `protobuf:"varint,11,opt,name=anger_likelihood,json=angerLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"anger_likelihood,omitempty"` // Surprise likelihood. - SurpriseLikelihood Likelihood `protobuf:"varint,12,opt,name=surprise_likelihood,json=surpriseLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"surprise_likelihood,omitempty"` + SurpriseLikelihood Likelihood `protobuf:"varint,12,opt,name=surprise_likelihood,json=surpriseLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"surprise_likelihood,omitempty"` // Under-exposed likelihood. - UnderExposedLikelihood Likelihood `protobuf:"varint,13,opt,name=under_exposed_likelihood,json=underExposedLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"under_exposed_likelihood,omitempty"` + UnderExposedLikelihood Likelihood `protobuf:"varint,13,opt,name=under_exposed_likelihood,json=underExposedLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"under_exposed_likelihood,omitempty"` // Blurred likelihood. - BlurredLikelihood Likelihood `protobuf:"varint,14,opt,name=blurred_likelihood,json=blurredLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"blurred_likelihood,omitempty"` + BlurredLikelihood Likelihood `protobuf:"varint,14,opt,name=blurred_likelihood,json=blurredLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"blurred_likelihood,omitempty"` // Headwear likelihood. - HeadwearLikelihood Likelihood `protobuf:"varint,15,opt,name=headwear_likelihood,json=headwearLikelihood,enum=google.cloud.vision.v1.Likelihood" json:"headwear_likelihood,omitempty"` + HeadwearLikelihood Likelihood `protobuf:"varint,15,opt,name=headwear_likelihood,json=headwearLikelihood,proto3,enum=google.cloud.vision.v1.Likelihood" json:"headwear_likelihood,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -581,7 +581,7 @@ func (m *FaceAnnotation) Reset() { *m = FaceAnnotation{} } func (m *FaceAnnotation) String() string { return proto.CompactTextString(m) } func (*FaceAnnotation) ProtoMessage() {} func (*FaceAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{3} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{3} } func (m *FaceAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FaceAnnotation.Unmarshal(m, b) @@ -709,9 +709,9 @@ func (m *FaceAnnotation) GetHeadwearLikelihood() Likelihood { // A face-specific landmark (for example, a face feature). type FaceAnnotation_Landmark struct { // Face landmark type. - Type FaceAnnotation_Landmark_Type `protobuf:"varint,3,opt,name=type,enum=google.cloud.vision.v1.FaceAnnotation_Landmark_Type" json:"type,omitempty"` + Type FaceAnnotation_Landmark_Type `protobuf:"varint,3,opt,name=type,proto3,enum=google.cloud.vision.v1.FaceAnnotation_Landmark_Type" json:"type,omitempty"` // Face landmark position. - Position *Position `protobuf:"bytes,4,opt,name=position" json:"position,omitempty"` + Position *Position `protobuf:"bytes,4,opt,name=position,proto3" json:"position,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -721,7 +721,7 @@ func (m *FaceAnnotation_Landmark) Reset() { *m = FaceAnnotation_Landmark func (m *FaceAnnotation_Landmark) String() string { return proto.CompactTextString(m) } func (*FaceAnnotation_Landmark) ProtoMessage() {} func (*FaceAnnotation_Landmark) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{3, 0} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{3, 0} } func (m *FaceAnnotation_Landmark) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FaceAnnotation_Landmark.Unmarshal(m, b) @@ -758,7 +758,7 @@ func (m *FaceAnnotation_Landmark) GetPosition() *Position { // Detected entity location information. type LocationInfo struct { // lat/long location coordinates. - LatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng,json=latLng" json:"lat_lng,omitempty"` + LatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng,json=latLng,proto3" json:"lat_lng,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -768,7 +768,7 @@ func (m *LocationInfo) Reset() { *m = LocationInfo{} } func (m *LocationInfo) String() string { return proto.CompactTextString(m) } func (*LocationInfo) ProtoMessage() {} func (*LocationInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{4} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{4} } func (m *LocationInfo) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LocationInfo.Unmarshal(m, b) @@ -798,11 +798,11 @@ func (m *LocationInfo) GetLatLng() *latlng.LatLng { // A `Property` consists of a user-supplied name/value pair. type Property struct { // Name of the property. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Value of the property. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Value of numeric properties. - Uint64Value uint64 `protobuf:"varint,3,opt,name=uint64_value,json=uint64Value" json:"uint64_value,omitempty"` + Uint64Value uint64 `protobuf:"varint,3,opt,name=uint64_value,json=uint64Value,proto3" json:"uint64_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -812,7 +812,7 @@ func (m *Property) Reset() { *m = Property{} } func (m *Property) String() string { return proto.CompactTextString(m) } func (*Property) ProtoMessage() {} func (*Property) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{5} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{5} } func (m *Property) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Property.Unmarshal(m, b) @@ -858,38 +858,38 @@ type EntityAnnotation struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search // API](https://developers.google.com/knowledge-graph/). - Mid string `protobuf:"bytes,1,opt,name=mid" json:"mid,omitempty"` + Mid string `protobuf:"bytes,1,opt,name=mid,proto3" json:"mid,omitempty"` // The language code for the locale in which the entity textual // `description` is expressed. - Locale string `protobuf:"bytes,2,opt,name=locale" json:"locale,omitempty"` + Locale string `protobuf:"bytes,2,opt,name=locale,proto3" json:"locale,omitempty"` // Entity textual description, expressed in its `locale` language. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Overall score of the result. Range [0, 1]. - Score float32 `protobuf:"fixed32,4,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,4,opt,name=score,proto3" json:"score,omitempty"` // **Deprecated. Use `score` instead.** // The accuracy of the entity detection in an image. // For example, for an image in which the "Eiffel Tower" entity is detected, // this field represents the confidence that there is a tower in the query // image. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` // The relevancy of the ICA (Image Content Annotation) label to the // image. For example, the relevancy of "tower" is likely higher to an image // containing the detected "Eiffel Tower" than to an image containing a // detected distant towering building, even though the confidence that // there is a tower in each image may be the same. Range [0, 1]. - Topicality float32 `protobuf:"fixed32,6,opt,name=topicality" json:"topicality,omitempty"` + Topicality float32 `protobuf:"fixed32,6,opt,name=topicality,proto3" json:"topicality,omitempty"` // Image region to which this entity belongs. Not produced // for `LABEL_DETECTION` features. - BoundingPoly *BoundingPoly `protobuf:"bytes,7,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,7,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // The location information for the detected entity. Multiple // `LocationInfo` elements can be present because one location may // indicate the location of the scene in the image, and another location // may indicate the location of the place where the image was taken. // Location information is usually present for landmarks. - Locations []*LocationInfo `protobuf:"bytes,8,rep,name=locations" json:"locations,omitempty"` + Locations []*LocationInfo `protobuf:"bytes,8,rep,name=locations,proto3" json:"locations,omitempty"` // Some entities may have optional user-supplied `Property` (name/value) // fields, such a score or string that qualifies the entity. - Properties []*Property `protobuf:"bytes,9,rep,name=properties" json:"properties,omitempty"` + Properties []*Property `protobuf:"bytes,9,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -899,7 +899,7 @@ func (m *EntityAnnotation) Reset() { *m = EntityAnnotation{} } func (m *EntityAnnotation) String() string { return proto.CompactTextString(m) } func (*EntityAnnotation) ProtoMessage() {} func (*EntityAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{6} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{6} } func (m *EntityAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_EntityAnnotation.Unmarshal(m, b) @@ -989,20 +989,20 @@ type SafeSearchAnnotation struct { // Represents the adult content likelihood for the image. Adult content may // contain elements such as nudity, pornographic images or cartoons, or // sexual activities. - Adult Likelihood `protobuf:"varint,1,opt,name=adult,enum=google.cloud.vision.v1.Likelihood" json:"adult,omitempty"` + Adult Likelihood `protobuf:"varint,1,opt,name=adult,proto3,enum=google.cloud.vision.v1.Likelihood" json:"adult,omitempty"` // Spoof likelihood. The likelihood that an modification // was made to the image's canonical version to make it appear // funny or offensive. - Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,enum=google.cloud.vision.v1.Likelihood" json:"spoof,omitempty"` + Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,proto3,enum=google.cloud.vision.v1.Likelihood" json:"spoof,omitempty"` // Likelihood that this is a medical image. - Medical Likelihood `protobuf:"varint,3,opt,name=medical,enum=google.cloud.vision.v1.Likelihood" json:"medical,omitempty"` + Medical Likelihood `protobuf:"varint,3,opt,name=medical,proto3,enum=google.cloud.vision.v1.Likelihood" json:"medical,omitempty"` // Likelihood that this image contains violent content. - Violence Likelihood `protobuf:"varint,4,opt,name=violence,enum=google.cloud.vision.v1.Likelihood" json:"violence,omitempty"` + Violence Likelihood `protobuf:"varint,4,opt,name=violence,proto3,enum=google.cloud.vision.v1.Likelihood" json:"violence,omitempty"` // Likelihood that the request image contains racy content. Racy content may // include (but is not limited to) skimpy or sheer clothing, strategically // covered nudity, lewd or provocative poses, or close-ups of sensitive // body areas. - Racy Likelihood `protobuf:"varint,9,opt,name=racy,enum=google.cloud.vision.v1.Likelihood" json:"racy,omitempty"` + Racy Likelihood `protobuf:"varint,9,opt,name=racy,proto3,enum=google.cloud.vision.v1.Likelihood" json:"racy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1012,7 +1012,7 @@ func (m *SafeSearchAnnotation) Reset() { *m = SafeSearchAnnotation{} } func (m *SafeSearchAnnotation) String() string { return proto.CompactTextString(m) } func (*SafeSearchAnnotation) ProtoMessage() {} func (*SafeSearchAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{7} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{7} } func (m *SafeSearchAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SafeSearchAnnotation.Unmarshal(m, b) @@ -1070,9 +1070,9 @@ func (m *SafeSearchAnnotation) GetRacy() Likelihood { // Rectangle determined by min and max `LatLng` pairs. type LatLongRect struct { // Min lat/long pair. - MinLatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=min_lat_lng,json=minLatLng" json:"min_lat_lng,omitempty"` + MinLatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=min_lat_lng,json=minLatLng,proto3" json:"min_lat_lng,omitempty"` // Max lat/long pair. - MaxLatLng *latlng.LatLng `protobuf:"bytes,2,opt,name=max_lat_lng,json=maxLatLng" json:"max_lat_lng,omitempty"` + MaxLatLng *latlng.LatLng `protobuf:"bytes,2,opt,name=max_lat_lng,json=maxLatLng,proto3" json:"max_lat_lng,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1082,7 +1082,7 @@ func (m *LatLongRect) Reset() { *m = LatLongRect{} } func (m *LatLongRect) String() string { return proto.CompactTextString(m) } func (*LatLongRect) ProtoMessage() {} func (*LatLongRect) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{8} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{8} } func (m *LatLongRect) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LatLongRect.Unmarshal(m, b) @@ -1120,12 +1120,12 @@ func (m *LatLongRect) GetMaxLatLng() *latlng.LatLng { // the image that the color occupies in the image. type ColorInfo struct { // RGB components of the color. - Color *color.Color `protobuf:"bytes,1,opt,name=color" json:"color,omitempty"` + Color *color.Color `protobuf:"bytes,1,opt,name=color,proto3" json:"color,omitempty"` // Image-specific score for this color. Value in range [0, 1]. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // The fraction of pixels the color occupies in the image. // Value in range [0, 1]. - PixelFraction float32 `protobuf:"fixed32,3,opt,name=pixel_fraction,json=pixelFraction" json:"pixel_fraction,omitempty"` + PixelFraction float32 `protobuf:"fixed32,3,opt,name=pixel_fraction,json=pixelFraction,proto3" json:"pixel_fraction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1135,7 +1135,7 @@ func (m *ColorInfo) Reset() { *m = ColorInfo{} } func (m *ColorInfo) String() string { return proto.CompactTextString(m) } func (*ColorInfo) ProtoMessage() {} func (*ColorInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{9} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{9} } func (m *ColorInfo) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ColorInfo.Unmarshal(m, b) @@ -1179,7 +1179,7 @@ func (m *ColorInfo) GetPixelFraction() float32 { // Set of dominant colors and their corresponding scores. type DominantColorsAnnotation struct { // RGB color values with their score and pixel fraction. - Colors []*ColorInfo `protobuf:"bytes,1,rep,name=colors" json:"colors,omitempty"` + Colors []*ColorInfo `protobuf:"bytes,1,rep,name=colors,proto3" json:"colors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1189,7 +1189,7 @@ func (m *DominantColorsAnnotation) Reset() { *m = DominantColorsAnnotati func (m *DominantColorsAnnotation) String() string { return proto.CompactTextString(m) } func (*DominantColorsAnnotation) ProtoMessage() {} func (*DominantColorsAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{10} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{10} } func (m *DominantColorsAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DominantColorsAnnotation.Unmarshal(m, b) @@ -1219,7 +1219,7 @@ func (m *DominantColorsAnnotation) GetColors() []*ColorInfo { // Stores image properties, such as dominant colors. type ImageProperties struct { // If present, dominant colors completed successfully. - DominantColors *DominantColorsAnnotation `protobuf:"bytes,1,opt,name=dominant_colors,json=dominantColors" json:"dominant_colors,omitempty"` + DominantColors *DominantColorsAnnotation `protobuf:"bytes,1,opt,name=dominant_colors,json=dominantColors,proto3" json:"dominant_colors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1229,7 +1229,7 @@ func (m *ImageProperties) Reset() { *m = ImageProperties{} } func (m *ImageProperties) String() string { return proto.CompactTextString(m) } func (*ImageProperties) ProtoMessage() {} func (*ImageProperties) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{11} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{11} } func (m *ImageProperties) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ImageProperties.Unmarshal(m, b) @@ -1260,12 +1260,12 @@ func (m *ImageProperties) GetDominantColors() *DominantColorsAnnotation { type CropHint struct { // The bounding polygon for the crop region. The coordinates of the bounding // box are in the original image's scale, as returned in `ImageParams`. - BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // Confidence of this being a salient region. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // Fraction of importance of this salient region with respect to the original // image. - ImportanceFraction float32 `protobuf:"fixed32,3,opt,name=importance_fraction,json=importanceFraction" json:"importance_fraction,omitempty"` + ImportanceFraction float32 `protobuf:"fixed32,3,opt,name=importance_fraction,json=importanceFraction,proto3" json:"importance_fraction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1275,7 +1275,7 @@ func (m *CropHint) Reset() { *m = CropHint{} } func (m *CropHint) String() string { return proto.CompactTextString(m) } func (*CropHint) ProtoMessage() {} func (*CropHint) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{12} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{12} } func (m *CropHint) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CropHint.Unmarshal(m, b) @@ -1319,7 +1319,7 @@ func (m *CropHint) GetImportanceFraction() float32 { // Set of crop hints that are used to generate new crops when serving images. type CropHintsAnnotation struct { // Crop hint results. - CropHints []*CropHint `protobuf:"bytes,1,rep,name=crop_hints,json=cropHints" json:"crop_hints,omitempty"` + CropHints []*CropHint `protobuf:"bytes,1,rep,name=crop_hints,json=cropHints,proto3" json:"crop_hints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1329,7 +1329,7 @@ func (m *CropHintsAnnotation) Reset() { *m = CropHintsAnnotation{} } func (m *CropHintsAnnotation) String() string { return proto.CompactTextString(m) } func (*CropHintsAnnotation) ProtoMessage() {} func (*CropHintsAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{13} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{13} } func (m *CropHintsAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CropHintsAnnotation.Unmarshal(m, b) @@ -1364,7 +1364,7 @@ type CropHintsParams struct { // best possible crop is returned. The number of provided aspect ratios is // limited to a maximum of 16; any aspect ratios provided after the 16th are // ignored. - AspectRatios []float32 `protobuf:"fixed32,1,rep,packed,name=aspect_ratios,json=aspectRatios" json:"aspect_ratios,omitempty"` + AspectRatios []float32 `protobuf:"fixed32,1,rep,packed,name=aspect_ratios,json=aspectRatios,proto3" json:"aspect_ratios,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1374,7 +1374,7 @@ func (m *CropHintsParams) Reset() { *m = CropHintsParams{} } func (m *CropHintsParams) String() string { return proto.CompactTextString(m) } func (*CropHintsParams) ProtoMessage() {} func (*CropHintsParams) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{14} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{14} } func (m *CropHintsParams) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CropHintsParams.Unmarshal(m, b) @@ -1404,7 +1404,7 @@ func (m *CropHintsParams) GetAspectRatios() []float32 { // Parameters for web detection request. type WebDetectionParams struct { // Whether to include results derived from the geo information in the image. - IncludeGeoResults bool `protobuf:"varint,2,opt,name=include_geo_results,json=includeGeoResults" json:"include_geo_results,omitempty"` + IncludeGeoResults bool `protobuf:"varint,2,opt,name=include_geo_results,json=includeGeoResults,proto3" json:"include_geo_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1414,7 +1414,7 @@ func (m *WebDetectionParams) Reset() { *m = WebDetectionParams{} } func (m *WebDetectionParams) String() string { return proto.CompactTextString(m) } func (*WebDetectionParams) ProtoMessage() {} func (*WebDetectionParams) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{15} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{15} } func (m *WebDetectionParams) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WebDetectionParams.Unmarshal(m, b) @@ -1444,7 +1444,7 @@ func (m *WebDetectionParams) GetIncludeGeoResults() bool { // Image context and/or feature-specific parameters. type ImageContext struct { // Not used. - LatLongRect *LatLongRect `protobuf:"bytes,1,opt,name=lat_long_rect,json=latLongRect" json:"lat_long_rect,omitempty"` + LatLongRect *LatLongRect `protobuf:"bytes,1,opt,name=lat_long_rect,json=latLongRect,proto3" json:"lat_long_rect,omitempty"` // List of languages to use for TEXT_DETECTION. In most cases, an empty value // yields the best results since it enables automatic language detection. For // languages based on the Latin alphabet, setting `language_hints` is not @@ -1453,11 +1453,11 @@ type ImageContext struct { // significant hindrance if the hint is wrong). Text detection returns an // error if one or more of the specified languages is not one of the // [supported languages](/vision/docs/languages). - LanguageHints []string `protobuf:"bytes,2,rep,name=language_hints,json=languageHints" json:"language_hints,omitempty"` + LanguageHints []string `protobuf:"bytes,2,rep,name=language_hints,json=languageHints,proto3" json:"language_hints,omitempty"` // Parameters for crop hints annotation request. - CropHintsParams *CropHintsParams `protobuf:"bytes,4,opt,name=crop_hints_params,json=cropHintsParams" json:"crop_hints_params,omitempty"` + CropHintsParams *CropHintsParams `protobuf:"bytes,4,opt,name=crop_hints_params,json=cropHintsParams,proto3" json:"crop_hints_params,omitempty"` // Parameters for web detection. - WebDetectionParams *WebDetectionParams `protobuf:"bytes,6,opt,name=web_detection_params,json=webDetectionParams" json:"web_detection_params,omitempty"` + WebDetectionParams *WebDetectionParams `protobuf:"bytes,6,opt,name=web_detection_params,json=webDetectionParams,proto3" json:"web_detection_params,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1467,7 +1467,7 @@ func (m *ImageContext) Reset() { *m = ImageContext{} } func (m *ImageContext) String() string { return proto.CompactTextString(m) } func (*ImageContext) ProtoMessage() {} func (*ImageContext) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{16} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{16} } func (m *ImageContext) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ImageContext.Unmarshal(m, b) @@ -1519,11 +1519,11 @@ func (m *ImageContext) GetWebDetectionParams() *WebDetectionParams { // image, with user-requested features. type AnnotateImageRequest struct { // The image to be processed. - Image *Image `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"` + Image *Image `protobuf:"bytes,1,opt,name=image,proto3" json:"image,omitempty"` // Requested features. - Features []*Feature `protobuf:"bytes,2,rep,name=features" json:"features,omitempty"` + Features []*Feature `protobuf:"bytes,2,rep,name=features,proto3" json:"features,omitempty"` // Additional context that may accompany the image. - ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext" json:"image_context,omitempty"` + ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext,proto3" json:"image_context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1533,7 +1533,7 @@ func (m *AnnotateImageRequest) Reset() { *m = AnnotateImageRequest{} } func (m *AnnotateImageRequest) String() string { return proto.CompactTextString(m) } func (*AnnotateImageRequest) ProtoMessage() {} func (*AnnotateImageRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{17} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{17} } func (m *AnnotateImageRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnnotateImageRequest.Unmarshal(m, b) @@ -1578,10 +1578,10 @@ func (m *AnnotateImageRequest) GetImageContext() *ImageContext { // information about the source of that image. type ImageAnnotationContext struct { // The URI of the file used to produce the image. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` // If the file was a PDF or TIFF, this field gives the page number within // the file used to produce the image. - PageNumber int32 `protobuf:"varint,2,opt,name=page_number,json=pageNumber" json:"page_number,omitempty"` + PageNumber int32 `protobuf:"varint,2,opt,name=page_number,json=pageNumber,proto3" json:"page_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1591,7 +1591,7 @@ func (m *ImageAnnotationContext) Reset() { *m = ImageAnnotationContext{} func (m *ImageAnnotationContext) String() string { return proto.CompactTextString(m) } func (*ImageAnnotationContext) ProtoMessage() {} func (*ImageAnnotationContext) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{18} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{18} } func (m *ImageAnnotationContext) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ImageAnnotationContext.Unmarshal(m, b) @@ -1628,35 +1628,35 @@ func (m *ImageAnnotationContext) GetPageNumber() int32 { // Response to an image annotation request. type AnnotateImageResponse struct { // If present, face detection has completed successfully. - FaceAnnotations []*FaceAnnotation `protobuf:"bytes,1,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"` + FaceAnnotations []*FaceAnnotation `protobuf:"bytes,1,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // If present, landmark detection has completed successfully. - LandmarkAnnotations []*EntityAnnotation `protobuf:"bytes,2,rep,name=landmark_annotations,json=landmarkAnnotations" json:"landmark_annotations,omitempty"` + LandmarkAnnotations []*EntityAnnotation `protobuf:"bytes,2,rep,name=landmark_annotations,json=landmarkAnnotations,proto3" json:"landmark_annotations,omitempty"` // If present, logo detection has completed successfully. - LogoAnnotations []*EntityAnnotation `protobuf:"bytes,3,rep,name=logo_annotations,json=logoAnnotations" json:"logo_annotations,omitempty"` + LogoAnnotations []*EntityAnnotation `protobuf:"bytes,3,rep,name=logo_annotations,json=logoAnnotations,proto3" json:"logo_annotations,omitempty"` // If present, label detection has completed successfully. - LabelAnnotations []*EntityAnnotation `protobuf:"bytes,4,rep,name=label_annotations,json=labelAnnotations" json:"label_annotations,omitempty"` + LabelAnnotations []*EntityAnnotation `protobuf:"bytes,4,rep,name=label_annotations,json=labelAnnotations,proto3" json:"label_annotations,omitempty"` // If present, text (OCR) detection has completed successfully. - TextAnnotations []*EntityAnnotation `protobuf:"bytes,5,rep,name=text_annotations,json=textAnnotations" json:"text_annotations,omitempty"` + TextAnnotations []*EntityAnnotation `protobuf:"bytes,5,rep,name=text_annotations,json=textAnnotations,proto3" json:"text_annotations,omitempty"` // If present, text (OCR) detection or document (OCR) text detection has // completed successfully. // This annotation provides the structural hierarchy for the OCR detected // text. - FullTextAnnotation *TextAnnotation `protobuf:"bytes,12,opt,name=full_text_annotation,json=fullTextAnnotation" json:"full_text_annotation,omitempty"` + FullTextAnnotation *TextAnnotation `protobuf:"bytes,12,opt,name=full_text_annotation,json=fullTextAnnotation,proto3" json:"full_text_annotation,omitempty"` // If present, safe-search annotation has completed successfully. - SafeSearchAnnotation *SafeSearchAnnotation `protobuf:"bytes,6,opt,name=safe_search_annotation,json=safeSearchAnnotation" json:"safe_search_annotation,omitempty"` + SafeSearchAnnotation *SafeSearchAnnotation `protobuf:"bytes,6,opt,name=safe_search_annotation,json=safeSearchAnnotation,proto3" json:"safe_search_annotation,omitempty"` // If present, image properties were extracted successfully. - ImagePropertiesAnnotation *ImageProperties `protobuf:"bytes,8,opt,name=image_properties_annotation,json=imagePropertiesAnnotation" json:"image_properties_annotation,omitempty"` + ImagePropertiesAnnotation *ImageProperties `protobuf:"bytes,8,opt,name=image_properties_annotation,json=imagePropertiesAnnotation,proto3" json:"image_properties_annotation,omitempty"` // If present, crop hints have completed successfully. - CropHintsAnnotation *CropHintsAnnotation `protobuf:"bytes,11,opt,name=crop_hints_annotation,json=cropHintsAnnotation" json:"crop_hints_annotation,omitempty"` + CropHintsAnnotation *CropHintsAnnotation `protobuf:"bytes,11,opt,name=crop_hints_annotation,json=cropHintsAnnotation,proto3" json:"crop_hints_annotation,omitempty"` // If present, web detection has completed successfully. - WebDetection *WebDetection `protobuf:"bytes,13,opt,name=web_detection,json=webDetection" json:"web_detection,omitempty"` + WebDetection *WebDetection `protobuf:"bytes,13,opt,name=web_detection,json=webDetection,proto3" json:"web_detection,omitempty"` // If set, represents the error message for the operation. // Note that filled-in image annotations are guaranteed to be // correct, even when `error` is set. - Error *status.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` // If present, contextual information is needed to understand where this image // comes from. - Context *ImageAnnotationContext `protobuf:"bytes,21,opt,name=context" json:"context,omitempty"` + Context *ImageAnnotationContext `protobuf:"bytes,21,opt,name=context,proto3" json:"context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1666,7 +1666,7 @@ func (m *AnnotateImageResponse) Reset() { *m = AnnotateImageResponse{} } func (m *AnnotateImageResponse) String() string { return proto.CompactTextString(m) } func (*AnnotateImageResponse) ProtoMessage() {} func (*AnnotateImageResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{19} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{19} } func (m *AnnotateImageResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnnotateImageResponse.Unmarshal(m, b) @@ -1774,9 +1774,9 @@ func (m *AnnotateImageResponse) GetContext() *ImageAnnotationContext { // images, which individually have their own responses. type AnnotateFileResponse struct { // Information about the file for which this response is generated. - InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig" json:"input_config,omitempty"` + InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig,proto3" json:"input_config,omitempty"` // Individual responses to images found within the file. - Responses []*AnnotateImageResponse `protobuf:"bytes,2,rep,name=responses" json:"responses,omitempty"` + Responses []*AnnotateImageResponse `protobuf:"bytes,2,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1786,7 +1786,7 @@ func (m *AnnotateFileResponse) Reset() { *m = AnnotateFileResponse{} } func (m *AnnotateFileResponse) String() string { return proto.CompactTextString(m) } func (*AnnotateFileResponse) ProtoMessage() {} func (*AnnotateFileResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{20} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{20} } func (m *AnnotateFileResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnnotateFileResponse.Unmarshal(m, b) @@ -1823,7 +1823,7 @@ func (m *AnnotateFileResponse) GetResponses() []*AnnotateImageResponse { // Multiple image annotation requests are batched into a single service call. type BatchAnnotateImagesRequest struct { // Individual image annotation requests for this batch. - Requests []*AnnotateImageRequest `protobuf:"bytes,1,rep,name=requests" json:"requests,omitempty"` + Requests []*AnnotateImageRequest `protobuf:"bytes,1,rep,name=requests,proto3" json:"requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1833,7 +1833,7 @@ func (m *BatchAnnotateImagesRequest) Reset() { *m = BatchAnnotateImagesR func (m *BatchAnnotateImagesRequest) String() string { return proto.CompactTextString(m) } func (*BatchAnnotateImagesRequest) ProtoMessage() {} func (*BatchAnnotateImagesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{21} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{21} } func (m *BatchAnnotateImagesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BatchAnnotateImagesRequest.Unmarshal(m, b) @@ -1863,7 +1863,7 @@ func (m *BatchAnnotateImagesRequest) GetRequests() []*AnnotateImageRequest { // Response to a batch image annotation request. type BatchAnnotateImagesResponse struct { // Individual responses to image annotation requests within the batch. - Responses []*AnnotateImageResponse `protobuf:"bytes,1,rep,name=responses" json:"responses,omitempty"` + Responses []*AnnotateImageResponse `protobuf:"bytes,1,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1873,7 +1873,7 @@ func (m *BatchAnnotateImagesResponse) Reset() { *m = BatchAnnotateImages func (m *BatchAnnotateImagesResponse) String() string { return proto.CompactTextString(m) } func (*BatchAnnotateImagesResponse) ProtoMessage() {} func (*BatchAnnotateImagesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{22} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{22} } func (m *BatchAnnotateImagesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BatchAnnotateImagesResponse.Unmarshal(m, b) @@ -1903,13 +1903,13 @@ func (m *BatchAnnotateImagesResponse) GetResponses() []*AnnotateImageResponse { // An offline file annotation request. type AsyncAnnotateFileRequest struct { // Required. Information about the input file. - InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig" json:"input_config,omitempty"` + InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig,proto3" json:"input_config,omitempty"` // Required. Requested features. - Features []*Feature `protobuf:"bytes,2,rep,name=features" json:"features,omitempty"` + Features []*Feature `protobuf:"bytes,2,rep,name=features,proto3" json:"features,omitempty"` // Additional context that may accompany the image(s) in the file. - ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext" json:"image_context,omitempty"` + ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext,proto3" json:"image_context,omitempty"` // Required. The desired output location and metadata (e.g. format). - OutputConfig *OutputConfig `protobuf:"bytes,4,opt,name=output_config,json=outputConfig" json:"output_config,omitempty"` + OutputConfig *OutputConfig `protobuf:"bytes,4,opt,name=output_config,json=outputConfig,proto3" json:"output_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1919,7 +1919,7 @@ func (m *AsyncAnnotateFileRequest) Reset() { *m = AsyncAnnotateFileReque func (m *AsyncAnnotateFileRequest) String() string { return proto.CompactTextString(m) } func (*AsyncAnnotateFileRequest) ProtoMessage() {} func (*AsyncAnnotateFileRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{23} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{23} } func (m *AsyncAnnotateFileRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AsyncAnnotateFileRequest.Unmarshal(m, b) @@ -1970,7 +1970,7 @@ func (m *AsyncAnnotateFileRequest) GetOutputConfig() *OutputConfig { // The response for a single offline file annotation request. type AsyncAnnotateFileResponse struct { // The output location and metadata from AsyncAnnotateFileRequest. - OutputConfig *OutputConfig `protobuf:"bytes,1,opt,name=output_config,json=outputConfig" json:"output_config,omitempty"` + OutputConfig *OutputConfig `protobuf:"bytes,1,opt,name=output_config,json=outputConfig,proto3" json:"output_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1980,7 +1980,7 @@ func (m *AsyncAnnotateFileResponse) Reset() { *m = AsyncAnnotateFileResp func (m *AsyncAnnotateFileResponse) String() string { return proto.CompactTextString(m) } func (*AsyncAnnotateFileResponse) ProtoMessage() {} func (*AsyncAnnotateFileResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{24} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{24} } func (m *AsyncAnnotateFileResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AsyncAnnotateFileResponse.Unmarshal(m, b) @@ -2011,7 +2011,7 @@ func (m *AsyncAnnotateFileResponse) GetOutputConfig() *OutputConfig { // call. type AsyncBatchAnnotateFilesRequest struct { // Individual async file annotation requests for this batch. - Requests []*AsyncAnnotateFileRequest `protobuf:"bytes,1,rep,name=requests" json:"requests,omitempty"` + Requests []*AsyncAnnotateFileRequest `protobuf:"bytes,1,rep,name=requests,proto3" json:"requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2021,7 +2021,7 @@ func (m *AsyncBatchAnnotateFilesRequest) Reset() { *m = AsyncBatchAnnota func (m *AsyncBatchAnnotateFilesRequest) String() string { return proto.CompactTextString(m) } func (*AsyncBatchAnnotateFilesRequest) ProtoMessage() {} func (*AsyncBatchAnnotateFilesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{25} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{25} } func (m *AsyncBatchAnnotateFilesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AsyncBatchAnnotateFilesRequest.Unmarshal(m, b) @@ -2052,7 +2052,7 @@ func (m *AsyncBatchAnnotateFilesRequest) GetRequests() []*AsyncAnnotateFileReque type AsyncBatchAnnotateFilesResponse struct { // The list of file annotation responses, one for each request in // AsyncBatchAnnotateFilesRequest. - Responses []*AsyncAnnotateFileResponse `protobuf:"bytes,1,rep,name=responses" json:"responses,omitempty"` + Responses []*AsyncAnnotateFileResponse `protobuf:"bytes,1,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2062,7 +2062,7 @@ func (m *AsyncBatchAnnotateFilesResponse) Reset() { *m = AsyncBatchAnnot func (m *AsyncBatchAnnotateFilesResponse) String() string { return proto.CompactTextString(m) } func (*AsyncBatchAnnotateFilesResponse) ProtoMessage() {} func (*AsyncBatchAnnotateFilesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{26} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{26} } func (m *AsyncBatchAnnotateFilesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AsyncBatchAnnotateFilesResponse.Unmarshal(m, b) @@ -2092,10 +2092,10 @@ func (m *AsyncBatchAnnotateFilesResponse) GetResponses() []*AsyncAnnotateFileRes // The desired input location and metadata. type InputConfig struct { // The Google Cloud Storage location to read the input from. - GcsSource *GcsSource `protobuf:"bytes,1,opt,name=gcs_source,json=gcsSource" json:"gcs_source,omitempty"` + GcsSource *GcsSource `protobuf:"bytes,1,opt,name=gcs_source,json=gcsSource,proto3" json:"gcs_source,omitempty"` // The type of the file. Currently only "application/pdf" and "image/tiff" // are supported. Wildcards are not supported. - MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType" json:"mime_type,omitempty"` + MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2105,7 +2105,7 @@ func (m *InputConfig) Reset() { *m = InputConfig{} } func (m *InputConfig) String() string { return proto.CompactTextString(m) } func (*InputConfig) ProtoMessage() {} func (*InputConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{27} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{27} } func (m *InputConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InputConfig.Unmarshal(m, b) @@ -2142,7 +2142,7 @@ func (m *InputConfig) GetMimeType() string { // The desired output location and metadata. type OutputConfig struct { // The Google Cloud Storage location to write the output(s) to. - GcsDestination *GcsDestination `protobuf:"bytes,1,opt,name=gcs_destination,json=gcsDestination" json:"gcs_destination,omitempty"` + GcsDestination *GcsDestination `protobuf:"bytes,1,opt,name=gcs_destination,json=gcsDestination,proto3" json:"gcs_destination,omitempty"` // The max number of response protos to put into each output JSON file on // Google Cloud Storage. // The valid range is [1, 100]. If not specified, the default value is 20. @@ -2154,7 +2154,7 @@ type OutputConfig struct { // // Currently, batch_size only applies to GcsDestination, with potential future // support for other output configurations. - BatchSize int32 `protobuf:"varint,2,opt,name=batch_size,json=batchSize" json:"batch_size,omitempty"` + BatchSize int32 `protobuf:"varint,2,opt,name=batch_size,json=batchSize,proto3" json:"batch_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2164,7 +2164,7 @@ func (m *OutputConfig) Reset() { *m = OutputConfig{} } func (m *OutputConfig) String() string { return proto.CompactTextString(m) } func (*OutputConfig) ProtoMessage() {} func (*OutputConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{28} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{28} } func (m *OutputConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_OutputConfig.Unmarshal(m, b) @@ -2202,7 +2202,7 @@ func (m *OutputConfig) GetBatchSize() int32 { type GcsSource struct { // Google Cloud Storage URI for the input file. This must only be a // Google Cloud Storage object. Wildcards are not currently supported. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2212,7 +2212,7 @@ func (m *GcsSource) Reset() { *m = GcsSource{} } func (m *GcsSource) String() string { return proto.CompactTextString(m) } func (*GcsSource) ProtoMessage() {} func (*GcsSource) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{29} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{29} } func (m *GcsSource) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GcsSource.Unmarshal(m, b) @@ -2256,7 +2256,7 @@ type GcsDestination struct { // which contains some subset of the full list of AnnotateImageResponse. // Multiple outputs can happen if, for example, the output JSON is too large // and overflows into multiple sharded files. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2266,7 +2266,7 @@ func (m *GcsDestination) Reset() { *m = GcsDestination{} } func (m *GcsDestination) String() string { return proto.CompactTextString(m) } func (*GcsDestination) ProtoMessage() {} func (*GcsDestination) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{30} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{30} } func (m *GcsDestination) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GcsDestination.Unmarshal(m, b) @@ -2296,11 +2296,11 @@ func (m *GcsDestination) GetUri() string { // Contains metadata for the BatchAnnotateImages operation. type OperationMetadata struct { // Current state of the batch operation. - State OperationMetadata_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.vision.v1.OperationMetadata_State" json:"state,omitempty"` + State OperationMetadata_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.vision.v1.OperationMetadata_State" json:"state,omitempty"` // The time when the batch request was received. - CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The time when the operation result was last updated. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2310,7 +2310,7 @@ func (m *OperationMetadata) Reset() { *m = OperationMetadata{} } func (m *OperationMetadata) String() string { return proto.CompactTextString(m) } func (*OperationMetadata) ProtoMessage() {} func (*OperationMetadata) Descriptor() ([]byte, []int) { - return fileDescriptor_image_annotator_255f02956495b90e, []int{31} + return fileDescriptor_image_annotator_aa4dfcf881de3ecf, []int{31} } func (m *OperationMetadata) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_OperationMetadata.Unmarshal(m, b) @@ -2440,8 +2440,7 @@ func (c *imageAnnotatorClient) AsyncBatchAnnotateFiles(ctx context.Context, in * return out, nil } -// Server API for ImageAnnotator service - +// ImageAnnotatorServer is the server API for ImageAnnotator service. type ImageAnnotatorServer interface { // Run image detection and annotation for a batch of images. BatchAnnotateImages(context.Context, *BatchAnnotateImagesRequest) (*BatchAnnotateImagesResponse, error) @@ -2512,190 +2511,190 @@ var _ImageAnnotator_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/cloud/vision/v1/image_annotator.proto", fileDescriptor_image_annotator_255f02956495b90e) + proto.RegisterFile("google/cloud/vision/v1/image_annotator.proto", fileDescriptor_image_annotator_aa4dfcf881de3ecf) } -var fileDescriptor_image_annotator_255f02956495b90e = []byte{ - // 2882 bytes of a gzipped FileDescriptorProto +var fileDescriptor_image_annotator_aa4dfcf881de3ecf = []byte{ + // 2892 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x5a, 0xcd, 0x73, 0xe3, 0x48, - 0x15, 0x5f, 0x3b, 0x5f, 0xf6, 0xb3, 0x93, 0x28, 0x9d, 0x8f, 0xf1, 0x64, 0xbe, 0x32, 0x5a, 0x96, - 0x4d, 0x0d, 0x83, 0xc3, 0x64, 0x97, 0xa9, 0x65, 0x66, 0x0a, 0x70, 0x6c, 0x25, 0x71, 0x8d, 0x63, - 0x7b, 0xdb, 0xce, 0xce, 0x0e, 0x6c, 0xa1, 0x52, 0xe4, 0xb6, 0x46, 0xbb, 0xb2, 0x24, 0x24, 0x79, - 0x26, 0xd9, 0x03, 0x07, 0xaa, 0x38, 0x73, 0xe0, 0xc4, 0x85, 0x23, 0x1c, 0x38, 0x51, 0x05, 0x77, - 0xae, 0x1c, 0x38, 0x50, 0xfb, 0x2f, 0xf0, 0x1f, 0x50, 0x45, 0x71, 0xa4, 0xfa, 0x43, 0x72, 0xcb, - 0x89, 0x32, 0xc9, 0xb2, 0x07, 0x4e, 0x71, 0xbf, 0xf7, 0x7e, 0xbf, 0xd7, 0xfd, 0xba, 0x5f, 0xf7, - 0xeb, 0x56, 0xe0, 0xa1, 0xe5, 0x79, 0x96, 0x43, 0x76, 0x4c, 0xc7, 0x1b, 0x0f, 0x76, 0x5e, 0xdb, - 0xa1, 0xed, 0xb9, 0x3b, 0xaf, 0x1f, 0xed, 0xd8, 0x23, 0xc3, 0x22, 0xba, 0xe1, 0xba, 0x5e, 0x64, - 0x44, 0x5e, 0x50, 0xf5, 0x03, 0x2f, 0xf2, 0xd0, 0x06, 0xb7, 0xae, 0x32, 0xeb, 0x2a, 0xb7, 0xae, - 0xbe, 0x7e, 0xb4, 0x79, 0x5b, 0xb0, 0x18, 0xbe, 0xbd, 0x23, 0x30, 0xb6, 0xe7, 0x86, 0x1c, 0xb5, - 0xf9, 0x5e, 0x86, 0x0f, 0x8b, 0x78, 0x23, 0x12, 0x05, 0x67, 0xc2, 0x2c, 0xab, 0x2b, 0x11, 0x39, - 0x8d, 0xf4, 0x09, 0xab, 0xb0, 0x7e, 0x90, 0x61, 0xfd, 0x86, 0x9c, 0xe8, 0x03, 0x12, 0x11, 0x53, - 0xb2, 0x7d, 0x57, 0xd8, 0x3a, 0x9e, 0x6b, 0x05, 0x63, 0xd7, 0xb5, 0x5d, 0x6b, 0xc7, 0xf3, 0x49, - 0x90, 0xea, 0xe5, 0x3d, 0x61, 0xc4, 0x5a, 0x27, 0xe3, 0xe1, 0x4e, 0x64, 0x8f, 0x48, 0x18, 0x19, - 0x23, 0x5f, 0x18, 0xdc, 0x10, 0x06, 0x81, 0x6f, 0xee, 0x84, 0x91, 0x11, 0x8d, 0xc3, 0x29, 0x45, - 0x74, 0xe6, 0x93, 0x1d, 0xd3, 0x73, 0xe2, 0x70, 0x6d, 0x56, 0x64, 0x85, 0x63, 0x44, 0x8e, 0x6b, - 0x71, 0x8d, 0xfa, 0xaf, 0x3c, 0x2c, 0xec, 0x13, 0x23, 0x1a, 0x07, 0x04, 0x7d, 0x04, 0xb3, 0xd4, - 0xa0, 0x92, 0xdb, 0xca, 0x6d, 0x2f, 0xed, 0x7e, 0xab, 0x7a, 0x71, 0x8c, 0xab, 0xc2, 0xbc, 0xda, - 0x3f, 0xf3, 0x09, 0x66, 0x08, 0x74, 0x0f, 0x4a, 0x23, 0xe3, 0x54, 0x0f, 0x48, 0x38, 0x76, 0xa2, - 0xb0, 0x92, 0xdf, 0xca, 0x6d, 0xcf, 0x61, 0x18, 0x19, 0xa7, 0x98, 0x4b, 0xd0, 0x1a, 0xcc, 0x8d, - 0xbc, 0x01, 0x71, 0x2a, 0x33, 0x5b, 0xb9, 0xed, 0x22, 0xe6, 0x0d, 0xf5, 0xdf, 0x39, 0x98, 0xa5, - 0x2c, 0x68, 0x0d, 0x94, 0xfe, 0xcb, 0xae, 0xa6, 0x1f, 0xb7, 0x7b, 0x5d, 0xad, 0xde, 0xdc, 0x6f, - 0x6a, 0x0d, 0xe5, 0x1d, 0x84, 0x60, 0x69, 0xbf, 0x56, 0xd7, 0xf4, 0x86, 0xd6, 0xd7, 0xea, 0xfd, - 0x66, 0xa7, 0xad, 0xe4, 0xd0, 0x06, 0xa0, 0x56, 0xad, 0xdd, 0x38, 0xaa, 0xe1, 0xe7, 0x92, 0x3c, - 0x4f, 0x6d, 0x5b, 0x9d, 0x83, 0x8e, 0x24, 0x9b, 0x41, 0xab, 0xb0, 0xdc, 0xaa, 0xed, 0x69, 0x2d, - 0x49, 0x38, 0x4b, 0x0d, 0xfb, 0xda, 0xa7, 0x7d, 0x49, 0x36, 0x87, 0x6e, 0xc1, 0x8d, 0x46, 0xa7, - 0x7e, 0x7c, 0xa4, 0xb5, 0xfb, 0xfa, 0x94, 0xb2, 0x84, 0x6e, 0xc2, 0x7a, 0xaf, 0xb6, 0xaf, 0xe9, - 0x3d, 0xad, 0x86, 0xeb, 0x87, 0x92, 0x6a, 0x9e, 0x76, 0xbb, 0x79, 0x54, 0x3b, 0xd0, 0xf4, 0x2e, - 0xee, 0x74, 0x35, 0xdc, 0x6f, 0x6a, 0x3d, 0x65, 0x01, 0x2d, 0x01, 0xd4, 0x71, 0xa7, 0xab, 0x1f, - 0x36, 0xdb, 0xfd, 0x9e, 0x52, 0x44, 0x2b, 0xb0, 0xf8, 0x42, 0xdb, 0x93, 0x80, 0xa0, 0xb6, 0xa1, - 0xd4, 0xa4, 0xeb, 0xba, 0xe7, 0x8d, 0x03, 0x93, 0x20, 0x15, 0x16, 0x2d, 0x33, 0xd4, 0xf9, 0x52, - 0x1f, 0x07, 0x36, 0x9b, 0x81, 0x22, 0x2e, 0x59, 0x66, 0xc8, 0xcc, 0x8e, 0x03, 0x1b, 0xdd, 0x82, - 0xe2, 0x44, 0x9f, 0x67, 0xfa, 0x82, 0x2d, 0x94, 0xea, 0xcf, 0x60, 0x8e, 0x19, 0xa2, 0x0a, 0x2c, - 0x98, 0x9e, 0x1b, 0x11, 0x37, 0x62, 0x1c, 0x65, 0x1c, 0x37, 0xd1, 0x53, 0x98, 0x0f, 0x99, 0x37, - 0x06, 0x2e, 0xed, 0xbe, 0x9b, 0x35, 0xbd, 0x52, 0xc7, 0xb0, 0x80, 0xa8, 0xff, 0x58, 0x86, 0xa5, - 0x7d, 0xc3, 0x24, 0xb5, 0x64, 0xf1, 0xa3, 0x26, 0x2c, 0x9e, 0x78, 0x63, 0x77, 0x60, 0xbb, 0x96, - 0xee, 0x7b, 0xce, 0x19, 0xf3, 0x57, 0xca, 0x5e, 0x35, 0x7b, 0xc2, 0xb8, 0xeb, 0x39, 0x67, 0xb8, - 0x7c, 0x22, 0xb5, 0x50, 0x1b, 0x94, 0xe1, 0x40, 0x4f, 0xb3, 0xe5, 0xaf, 0xc1, 0xb6, 0x34, 0x1c, - 0xc8, 0x6d, 0x74, 0x04, 0x45, 0xc7, 0x70, 0x07, 0x23, 0x23, 0xf8, 0x22, 0xac, 0xcc, 0x6c, 0xcd, - 0x6c, 0x97, 0x76, 0x77, 0x32, 0x17, 0x73, 0x6a, 0x54, 0xd5, 0x96, 0xc0, 0xe1, 0x09, 0x03, 0xba, - 0x03, 0x10, 0x78, 0x8e, 0xa3, 0x1b, 0xae, 0xe5, 0x90, 0xca, 0xec, 0x56, 0x6e, 0x3b, 0x8f, 0x8b, - 0x54, 0x52, 0xa3, 0x02, 0x3a, 0x31, 0xbe, 0xe1, 0x0a, 0xed, 0x1c, 0xd3, 0x16, 0x7c, 0xc3, 0xe5, - 0xca, 0x3b, 0x00, 0x91, 0xed, 0x44, 0x42, 0x3b, 0xcf, 0xb1, 0x54, 0xc2, 0xd5, 0x8f, 0x60, 0x2d, - 0xd9, 0x22, 0x74, 0xd3, 0x73, 0x87, 0xf6, 0x80, 0xb8, 0x26, 0xa9, 0x2c, 0x30, 0xc3, 0xd5, 0x44, - 0x57, 0x4f, 0x54, 0xe8, 0xfb, 0xb0, 0x11, 0x77, 0x8d, 0x06, 0x4b, 0x02, 0x15, 0x18, 0x68, 0x5d, - 0xd2, 0x4a, 0xb0, 0x26, 0x2c, 0x7d, 0xee, 0x9d, 0xe9, 0x8e, 0xfd, 0x05, 0x71, 0xec, 0x57, 0x9e, - 0x37, 0xa8, 0x14, 0x59, 0x96, 0xab, 0x59, 0x81, 0x69, 0x25, 0x96, 0x78, 0xf1, 0x73, 0xef, 0x6c, - 0xd2, 0x44, 0x1d, 0x58, 0x09, 0xbd, 0x20, 0xf0, 0xde, 0xc8, 0x6c, 0x70, 0x65, 0x36, 0x85, 0x83, - 0x25, 0xc2, 0x23, 0x50, 0x0c, 0xd7, 0x22, 0x81, 0xcc, 0x57, 0xba, 0x32, 0xdf, 0x32, 0xc3, 0x4a, - 0x74, 0x3d, 0x58, 0x0d, 0xc7, 0x81, 0x1f, 0xd8, 0x21, 0x91, 0x19, 0xcb, 0x57, 0x66, 0x44, 0x31, - 0x5c, 0x22, 0xfd, 0x0c, 0x2a, 0x63, 0x77, 0x40, 0x02, 0x9d, 0x9c, 0xfa, 0x5e, 0x48, 0x06, 0x32, - 0xf3, 0xe2, 0x95, 0x99, 0x37, 0x18, 0x87, 0xc6, 0x29, 0x24, 0xf6, 0x8f, 0x01, 0x9d, 0x38, 0xe3, - 0x20, 0x48, 0xf3, 0x2e, 0x5d, 0x99, 0x77, 0x45, 0xa0, 0xd3, 0x51, 0x78, 0x45, 0x8c, 0xc1, 0x1b, - 0x62, 0xa4, 0xe2, 0xba, 0x7c, 0xf5, 0x28, 0xc4, 0xf0, 0x89, 0x6c, 0xf3, 0x6f, 0x0b, 0x50, 0x88, - 0x53, 0x04, 0x1d, 0x8a, 0xe3, 0x62, 0x86, 0x51, 0x7e, 0x78, 0xcd, 0x0c, 0x93, 0x8f, 0x8f, 0x67, - 0x50, 0xf0, 0xbd, 0xd0, 0xa6, 0x7a, 0x96, 0x5f, 0xa5, 0xdd, 0xad, 0x2c, 0xb6, 0xae, 0xb0, 0xc3, - 0x09, 0x42, 0xfd, 0xf3, 0xfc, 0xe4, 0x14, 0x39, 0x6e, 0x3f, 0x6f, 0x77, 0x5e, 0xb4, 0xf5, 0xf8, - 0x8c, 0x50, 0xde, 0x41, 0x65, 0x28, 0xb4, 0xb4, 0xfd, 0xbe, 0xae, 0xbd, 0xd4, 0x94, 0x1c, 0x5a, - 0x84, 0x22, 0x6e, 0x1e, 0x1c, 0xf2, 0x66, 0x1e, 0x55, 0x60, 0x8d, 0x29, 0x3b, 0xfb, 0x7a, 0x6c, - 0xb4, 0x87, 0x3b, 0x2f, 0x94, 0x19, 0xba, 0xed, 0x73, 0xc3, 0x69, 0xd5, 0x2c, 0x55, 0xc5, 0xa0, - 0x84, 0x8b, 0xa9, 0xe6, 0xd0, 0x26, 0x6c, 0x24, 0xa8, 0xb4, 0x6e, 0x9e, 0xc2, 0x8e, 0x9a, 0x8d, - 0x6e, 0xa7, 0xd9, 0xee, 0xeb, 0x7b, 0x5a, 0xff, 0x85, 0xa6, 0xb5, 0xa9, 0x96, 0x1e, 0x19, 0x65, - 0x28, 0xb4, 0x3b, 0x3d, 0x4d, 0xef, 0x37, 0xbb, 0x4a, 0x81, 0xf6, 0xf1, 0xb8, 0xdb, 0xd5, 0xb0, - 0xde, 0x6a, 0x76, 0x95, 0x22, 0x6d, 0xb6, 0x3a, 0x2f, 0x44, 0x13, 0xe8, 0xf1, 0x72, 0xd4, 0x39, - 0xee, 0x1f, 0xb2, 0x5e, 0x29, 0x25, 0xb4, 0x0c, 0x25, 0xde, 0x66, 0xfe, 0x94, 0x32, 0x52, 0xa0, - 0xcc, 0x05, 0x75, 0xad, 0xdd, 0xd7, 0xb0, 0xb2, 0x88, 0xd6, 0x61, 0x85, 0xd1, 0xef, 0x75, 0xfa, - 0xfd, 0xce, 0x91, 0x30, 0x5c, 0xa2, 0xf1, 0x92, 0xc5, 0x8c, 0x6f, 0x99, 0x9e, 0xb0, 0xb2, 0x54, - 0x90, 0x28, 0xc9, 0xa8, 0xb5, 0x97, 0x9a, 0xde, 0xef, 0x74, 0xf5, 0xbd, 0xce, 0x71, 0xbb, 0x51, - 0xc3, 0x2f, 0x95, 0x95, 0x94, 0x8a, 0x8f, 0xba, 0xde, 0xc1, 0x6d, 0x0d, 0x2b, 0x08, 0xdd, 0x86, - 0x4a, 0xa2, 0x12, 0x8c, 0x09, 0x70, 0x35, 0x09, 0x3f, 0xd5, 0xb2, 0x1f, 0x02, 0xb7, 0x36, 0x09, - 0xe4, 0x39, 0x77, 0xeb, 0x69, 0x5d, 0xca, 0xdf, 0x06, 0xba, 0x03, 0x37, 0x27, 0xba, 0x69, 0x87, - 0x37, 0x26, 0xb3, 0x3a, 0xed, 0xb1, 0x82, 0xee, 0xc1, 0x2d, 0x79, 0x9e, 0x75, 0x3e, 0x05, 0xf1, - 0x8c, 0x29, 0x37, 0xd1, 0x16, 0xdc, 0x4e, 0x4d, 0xe9, 0xb4, 0xc5, 0x26, 0x0d, 0x28, 0xa7, 0xa8, - 0x61, 0xbd, 0x8f, 0x6b, 0x07, 0xf4, 0xb0, 0xbf, 0x45, 0xa3, 0x2f, 0x70, 0x92, 0xf8, 0x36, 0xab, - 0x58, 0xe2, 0xb1, 0x77, 0x8f, 0xbb, 0xcd, 0x96, 0x72, 0x87, 0x56, 0x2c, 0x93, 0xee, 0x71, 0xe1, - 0x5d, 0x8a, 0xdf, 0xef, 0x60, 0xed, 0x50, 0xab, 0x35, 0xf4, 0x03, 0x56, 0xd0, 0xb4, 0x6a, 0xca, - 0x3d, 0x5a, 0x56, 0xd4, 0x0f, 0x9b, 0x6d, 0xfd, 0xa0, 0x5d, 0xeb, 0x1f, 0x52, 0xca, 0x2d, 0xea, - 0x9f, 0x89, 0x18, 0xef, 0x41, 0xa7, 0x4d, 0xa5, 0xf7, 0x29, 0x9e, 0x49, 0x39, 0xb3, 0x10, 0xab, - 0xea, 0x33, 0x28, 0xb7, 0x3c, 0x93, 0x25, 0x65, 0xd3, 0x1d, 0x7a, 0xe8, 0x21, 0x2c, 0x38, 0x46, - 0xa4, 0x3b, 0xae, 0x25, 0x8e, 0xf2, 0xd5, 0x38, 0x07, 0x69, 0x8e, 0x56, 0x5b, 0x46, 0xd4, 0x72, - 0x2d, 0x3c, 0xef, 0xb0, 0xbf, 0xea, 0x0b, 0x28, 0x74, 0x03, 0x5a, 0xba, 0x46, 0x67, 0x08, 0xc1, - 0xac, 0x6b, 0x8c, 0x88, 0xa8, 0x5a, 0xd8, 0x6f, 0x5a, 0xf0, 0xbd, 0x36, 0x9c, 0x31, 0x11, 0xa5, - 0x0a, 0x6f, 0xa0, 0xfb, 0x50, 0x1e, 0xdb, 0x6e, 0xf4, 0xf8, 0x43, 0x9d, 0x2b, 0xe9, 0xd6, 0x31, - 0x8b, 0x4b, 0x5c, 0xf6, 0x09, 0x15, 0xa9, 0xbf, 0x9a, 0x01, 0x45, 0x73, 0x23, 0x3b, 0x3a, 0x93, - 0x8a, 0x0d, 0x05, 0x66, 0x46, 0xf6, 0x40, 0x38, 0xa0, 0x3f, 0xd1, 0x06, 0xcc, 0x3b, 0x9e, 0x69, - 0x38, 0xb1, 0x03, 0xd1, 0x42, 0x5b, 0x50, 0x1a, 0x90, 0xd0, 0x0c, 0x6c, 0x9f, 0xed, 0x26, 0xbc, - 0xdc, 0x94, 0x45, 0xb4, 0x67, 0xa1, 0xe9, 0x05, 0xf1, 0x49, 0xce, 0x1b, 0xe8, 0x2e, 0x80, 0x74, - 0x94, 0xf2, 0x63, 0x5c, 0x92, 0x50, 0x7d, 0xe4, 0xf9, 0xb6, 0x69, 0x38, 0x76, 0x74, 0x26, 0x0e, - 0x72, 0x49, 0x72, 0xbe, 0x1c, 0x5a, 0xf8, 0xda, 0xe5, 0xd0, 0x1e, 0x14, 0x1d, 0x31, 0x31, 0x61, - 0xa5, 0xc0, 0xca, 0x97, 0x4c, 0x1a, 0x79, 0x06, 0xf1, 0x04, 0x86, 0x7e, 0x0c, 0xe0, 0xf3, 0xe9, - 0xb1, 0x49, 0x58, 0x29, 0x32, 0x92, 0xec, 0x3d, 0x55, 0x4c, 0x24, 0x96, 0x30, 0xea, 0x5f, 0xf3, - 0xb0, 0xd6, 0x33, 0x86, 0xa4, 0x47, 0x8c, 0xc0, 0x7c, 0x25, 0xcd, 0xc5, 0x47, 0x30, 0x67, 0x0c, - 0xc6, 0x4e, 0x24, 0xae, 0x09, 0x57, 0x39, 0x4a, 0x38, 0x80, 0x22, 0x43, 0xdf, 0xf3, 0x86, 0x6c, - 0xca, 0xae, 0x88, 0x64, 0x00, 0xf4, 0x0c, 0x16, 0x46, 0x64, 0x40, 0x63, 0x2d, 0x4e, 0x9b, 0xab, - 0x60, 0x63, 0x08, 0xfa, 0x21, 0x14, 0x5e, 0xdb, 0x9e, 0xc3, 0x66, 0x76, 0xf6, 0xca, 0xf0, 0x04, - 0x83, 0x1e, 0xc3, 0x6c, 0x60, 0x98, 0x67, 0xd7, 0xa8, 0x98, 0x98, 0xbd, 0xfa, 0x06, 0x4a, 0x34, - 0x6b, 0x3c, 0xd7, 0xc2, 0xc4, 0x8c, 0xd0, 0x07, 0x50, 0x1a, 0xd9, 0xae, 0x7e, 0x85, 0x24, 0x2b, - 0x8e, 0x6c, 0x97, 0xff, 0x64, 0x20, 0xe3, 0x34, 0x01, 0xe5, 0x2f, 0x03, 0x19, 0xa7, 0xfc, 0xa7, - 0x1a, 0x40, 0xb1, 0x4e, 0x6f, 0x7f, 0x2c, 0xaf, 0xb7, 0x61, 0x8e, 0x5d, 0x05, 0x85, 0x43, 0x94, - 0xc2, 0x32, 0x33, 0xcc, 0x0d, 0x26, 0x99, 0x91, 0x97, 0x33, 0xe3, 0x3d, 0x58, 0xf2, 0xed, 0x53, - 0xe2, 0xe8, 0xc3, 0xc0, 0x30, 0x93, 0xa4, 0xca, 0xe3, 0x45, 0x26, 0xdd, 0x17, 0x42, 0xf5, 0x18, - 0x2a, 0x0d, 0x6f, 0x64, 0xbb, 0x86, 0x1b, 0x31, 0xd2, 0x50, 0x5a, 0x32, 0x3f, 0x80, 0x79, 0xe6, - 0x21, 0xac, 0xe4, 0xd8, 0x4a, 0xbc, 0x9f, 0x15, 0xc2, 0xa4, 0xd7, 0x58, 0x00, 0x54, 0x07, 0x96, - 0xd9, 0x85, 0xa4, 0x9b, 0xac, 0x4c, 0xf4, 0x12, 0x96, 0x07, 0xc2, 0x93, 0x9e, 0xd0, 0xd2, 0xa1, - 0x7d, 0x2f, 0x8b, 0x36, 0xab, 0x63, 0x78, 0x69, 0x90, 0xd2, 0xa8, 0xbf, 0xcf, 0x41, 0xa1, 0x1e, - 0x78, 0xfe, 0xa1, 0xed, 0x46, 0xdf, 0xe4, 0x0d, 0x27, 0xbd, 0xbb, 0xe4, 0xcf, 0xed, 0x2e, 0x3b, - 0xb0, 0x6a, 0x8f, 0x7c, 0x2f, 0x88, 0x0c, 0xd7, 0x24, 0xd3, 0x81, 0x46, 0x13, 0x55, 0x12, 0xed, - 0x4f, 0x60, 0x35, 0xee, 0xa7, 0x1c, 0xe8, 0x1f, 0x01, 0x98, 0x81, 0xe7, 0xeb, 0xaf, 0xa8, 0x5c, - 0x04, 0x3b, 0x33, 0xed, 0x63, 0x02, 0x5c, 0x34, 0x63, 0x2a, 0xf5, 0x31, 0x2c, 0x27, 0xbc, 0x5d, - 0x23, 0x30, 0x46, 0x21, 0x7a, 0x17, 0x16, 0x8d, 0xd0, 0x27, 0x66, 0xa4, 0xb3, 0x67, 0x0a, 0x4e, - 0x9b, 0xc7, 0x65, 0x2e, 0xc4, 0x4c, 0xa6, 0x36, 0x00, 0xbd, 0x20, 0x27, 0x8d, 0xf8, 0xbe, 0x22, - 0xa0, 0x55, 0x58, 0xb5, 0x5d, 0xd3, 0x19, 0x0f, 0x88, 0x6e, 0x11, 0x2f, 0xf5, 0x3c, 0x50, 0xc0, - 0x2b, 0x42, 0x75, 0x40, 0x3c, 0xf1, 0x4a, 0xa0, 0xfe, 0x29, 0x0f, 0x65, 0x36, 0xdb, 0x75, 0x7a, - 0x69, 0x3d, 0x8d, 0xd0, 0x01, 0x2c, 0xb2, 0x95, 0xef, 0xb9, 0x96, 0x1e, 0x10, 0x33, 0x12, 0x53, - 0x90, 0x79, 0x77, 0x95, 0xd2, 0x0d, 0x97, 0x1c, 0x29, 0xf7, 0xde, 0x83, 0x25, 0xc7, 0x70, 0xad, - 0x31, 0xbd, 0x40, 0xf3, 0xe0, 0xe4, 0xb7, 0x66, 0xb6, 0x8b, 0x78, 0x31, 0x96, 0xb2, 0x11, 0xa3, - 0x1e, 0xac, 0x4c, 0xe2, 0xa7, 0xfb, 0x6c, 0x14, 0xa2, 0x22, 0x7d, 0xff, 0x6d, 0x61, 0x14, 0xf1, - 0xc2, 0xcb, 0xe6, 0x54, 0x00, 0x3f, 0x83, 0xb5, 0xd4, 0x5b, 0x50, 0xcc, 0x3b, 0xcf, 0x78, 0x1f, - 0x64, 0xf1, 0x9e, 0x8f, 0x27, 0x46, 0x6f, 0xce, 0xc9, 0xd4, 0xaf, 0x72, 0xb0, 0x26, 0x56, 0x00, - 0x61, 0xb1, 0xc3, 0xe4, 0xe7, 0x63, 0x12, 0xd2, 0xed, 0x66, 0x8e, 0xbd, 0x0f, 0x88, 0x98, 0xdd, - 0xb9, 0xf4, 0xbe, 0x8f, 0xb9, 0x2d, 0x7a, 0x0a, 0x85, 0x21, 0x7f, 0xde, 0xe1, 0x11, 0x2a, 0xed, - 0xde, 0x7b, 0xcb, 0x33, 0x10, 0x4e, 0x00, 0x34, 0x61, 0xf8, 0x13, 0x85, 0xc9, 0xa7, 0x8f, 0xad, - 0xdf, 0x4b, 0x12, 0x46, 0x9e, 0x6a, 0x5c, 0xb6, 0xa5, 0x96, 0xfa, 0x1c, 0x36, 0x98, 0x76, 0xb2, - 0xb6, 0xe3, 0x25, 0xa1, 0xc0, 0xcc, 0xe4, 0x85, 0x84, 0xfe, 0x44, 0xf7, 0xa0, 0xe4, 0x53, 0xaf, - 0xee, 0x78, 0x74, 0x42, 0x82, 0xf8, 0xf1, 0x89, 0x8a, 0xda, 0x4c, 0xa2, 0xfe, 0xba, 0x00, 0xeb, - 0x53, 0x21, 0x0a, 0x7d, 0xcf, 0x0d, 0x09, 0xfa, 0x18, 0x94, 0xa1, 0x61, 0x12, 0xe9, 0x51, 0x2f, - 0xce, 0x9a, 0x6f, 0x5f, 0xed, 0x3a, 0x83, 0x97, 0x87, 0xa9, 0x76, 0x88, 0x7e, 0x0a, 0x6b, 0xf1, - 0x0d, 0x3c, 0x45, 0xcb, 0xa3, 0xb9, 0x9d, 0x45, 0x3b, 0x5d, 0xf2, 0xe0, 0xd5, 0x98, 0x45, 0x26, - 0xef, 0x81, 0xe2, 0x78, 0x96, 0x97, 0x22, 0x9e, 0xb9, 0x26, 0xf1, 0x32, 0x65, 0x90, 0x49, 0x8f, - 0x61, 0xc5, 0x31, 0x4e, 0x88, 0x93, 0x62, 0x9d, 0xbd, 0x26, 0xab, 0xc2, 0x28, 0xa6, 0xfa, 0x3a, - 0xf5, 0x60, 0x1a, 0x56, 0xe6, 0xae, 0xdb, 0x57, 0xca, 0x20, 0x93, 0x7e, 0x0a, 0x6b, 0xc3, 0xb1, - 0xe3, 0xe8, 0x53, 0xcc, 0xec, 0x72, 0x7f, 0xc9, 0xa4, 0xf5, 0x53, 0x34, 0x18, 0x51, 0x8e, 0xb4, - 0x0c, 0x9d, 0xc0, 0x46, 0x68, 0x0c, 0x89, 0x1e, 0xb2, 0x7a, 0x47, 0xe6, 0xe6, 0x79, 0xfa, 0x30, - 0x8b, 0xfb, 0xa2, 0x22, 0x09, 0xaf, 0x85, 0x17, 0x95, 0x4e, 0x16, 0xdc, 0xe2, 0x09, 0x32, 0xa9, - 0xb3, 0x64, 0x47, 0x85, 0xcb, 0x37, 0x9a, 0xa9, 0x73, 0x10, 0xdf, 0xb4, 0xd3, 0x02, 0xc9, 0x91, - 0x0e, 0xeb, 0xd2, 0x3e, 0x26, 0xb9, 0x28, 0x31, 0x17, 0xdf, 0x79, 0xeb, 0x5e, 0x26, 0x2f, 0x44, - 0xf3, 0x82, 0x83, 0xa6, 0x09, 0x8b, 0xa9, 0x3d, 0x8d, 0xbd, 0x81, 0x5c, 0x92, 0xea, 0xf2, 0x66, - 0x86, 0xcb, 0xf2, 0x36, 0x46, 0xeb, 0x13, 0x12, 0x04, 0x5e, 0xc0, 0xca, 0x2b, 0xa9, 0x3e, 0x09, - 0x7c, 0xb3, 0xda, 0x63, 0xaf, 0xdb, 0x98, 0x1b, 0xa0, 0x43, 0xf1, 0xb8, 0x79, 0x1a, 0x55, 0xd6, - 0x99, 0x6d, 0xf5, 0xd2, 0x50, 0x9d, 0xdb, 0x3b, 0x70, 0x0c, 0x57, 0xff, 0x28, 0x6d, 0x9a, 0xfb, - 0xb6, 0x33, 0xd9, 0x10, 0xf6, 0xa1, 0x6c, 0xbb, 0xfe, 0x38, 0xe2, 0xef, 0x6a, 0xd6, 0xdb, 0xce, - 0x9b, 0x26, 0xb5, 0x65, 0xaf, 0x6c, 0x16, 0x2e, 0xd9, 0x93, 0x06, 0x7a, 0x0e, 0xc5, 0x40, 0x70, - 0xc6, 0xa9, 0xff, 0xdd, 0x2c, 0x92, 0x0b, 0xb7, 0x26, 0x3c, 0xc1, 0xab, 0x43, 0xd8, 0xdc, 0x33, - 0xa2, 0x64, 0x25, 0x71, 0xc3, 0x30, 0xde, 0xe7, 0x0f, 0xa1, 0x10, 0xf0, 0x9f, 0xf1, 0xde, 0xf5, - 0xf0, 0x8a, 0x9e, 0x18, 0x08, 0x27, 0x68, 0xf5, 0x73, 0xb8, 0x75, 0xa1, 0x1f, 0x11, 0x9b, 0xd4, - 0x98, 0x72, 0xff, 0xe3, 0x98, 0xfe, 0x92, 0x87, 0x4a, 0x2d, 0x3c, 0x73, 0xcd, 0xf4, 0x34, 0xf0, - 0x21, 0x7d, 0x53, 0xb3, 0xf0, 0x7f, 0x72, 0x9a, 0x51, 0x2a, 0x6f, 0x1c, 0x49, 0x03, 0x9a, 0xbd, - 0x9c, 0xaa, 0xc3, 0x8c, 0xc5, 0x88, 0xca, 0x9e, 0xd4, 0x52, 0x87, 0x70, 0xf3, 0x82, 0xb0, 0x89, - 0x19, 0x3a, 0xe7, 0x27, 0xf7, 0xb5, 0xfd, 0xb8, 0x70, 0x97, 0xf9, 0x49, 0x2d, 0x08, 0xea, 0x2c, - 0x59, 0x77, 0xad, 0x73, 0xeb, 0x2e, 0xb3, 0xfe, 0xce, 0x9a, 0x68, 0x69, 0xed, 0x05, 0x70, 0x2f, - 0xd3, 0x9f, 0x18, 0x5d, 0xe7, 0xfc, 0xfa, 0x7b, 0x74, 0x0d, 0x8f, 0xe7, 0xd7, 0xa0, 0x03, 0x25, - 0x69, 0xe9, 0xd0, 0x3b, 0xb3, 0x65, 0x86, 0xba, 0xf8, 0x4a, 0xc2, 0x43, 0x97, 0x79, 0x53, 0x39, - 0x30, 0x43, 0xf1, 0x8d, 0xa4, 0x68, 0xc5, 0x3f, 0xd1, 0x2d, 0x28, 0x8e, 0xec, 0x11, 0xd1, 0xd9, - 0xb3, 0xa8, 0xf8, 0x46, 0x43, 0x05, 0xfd, 0x33, 0x9f, 0xa8, 0xbf, 0x80, 0xb2, 0x1c, 0x6f, 0xd4, - 0x81, 0x65, 0xea, 0x6e, 0x40, 0xc2, 0xc8, 0x76, 0xf9, 0xee, 0x9c, 0xbb, 0xfc, 0x14, 0x3b, 0x30, - 0xc3, 0xc6, 0xc4, 0x1a, 0x2f, 0x59, 0xa9, 0x36, 0xba, 0x03, 0x70, 0x42, 0xa3, 0xa7, 0x87, 0xf6, - 0x97, 0x44, 0x94, 0x41, 0x45, 0x26, 0xe9, 0xd9, 0x5f, 0x12, 0xf5, 0x0e, 0x14, 0x93, 0x4e, 0x9f, - 0xaf, 0xa2, 0x54, 0x15, 0x96, 0xd2, 0xfc, 0x17, 0xd8, 0xfc, 0x21, 0x0f, 0x2b, 0x9d, 0xf8, 0x73, - 0xe5, 0x11, 0x89, 0x8c, 0x81, 0x11, 0x19, 0x48, 0x83, 0xb9, 0x90, 0x86, 0x59, 0x3c, 0x08, 0x64, - 0x7e, 0x6a, 0x39, 0x87, 0x64, 0xdb, 0x3b, 0xc1, 0x1c, 0x8d, 0x9e, 0x42, 0xc9, 0x0c, 0x88, 0x11, - 0x11, 0x3d, 0xb2, 0x47, 0xfc, 0x09, 0xa6, 0xb4, 0xbb, 0x19, 0x93, 0xc5, 0x1f, 0x43, 0xab, 0xfd, - 0xf8, 0x63, 0x28, 0x06, 0x6e, 0x4e, 0x05, 0x14, 0x3c, 0xf6, 0x07, 0x09, 0x78, 0xfe, 0xed, 0x60, - 0x6e, 0x4e, 0x05, 0xea, 0xc7, 0x30, 0xc7, 0x7a, 0x82, 0xd6, 0x61, 0xa5, 0xd7, 0xaf, 0xf5, 0xa7, - 0xbf, 0x43, 0x96, 0x60, 0xa1, 0x8e, 0xb5, 0x5a, 0x5f, 0x6b, 0x28, 0x39, 0xda, 0xc0, 0xc7, 0xed, - 0x76, 0xb3, 0x7d, 0xa0, 0xe4, 0x51, 0x01, 0x66, 0x1b, 0x9d, 0xb6, 0xa6, 0xcc, 0xa0, 0x45, 0x28, - 0xd6, 0x6b, 0xed, 0xba, 0xd6, 0x6a, 0x69, 0x0d, 0x65, 0xf6, 0x01, 0x01, 0x90, 0xde, 0xe2, 0x4b, - 0xb0, 0x20, 0x1e, 0xa6, 0x95, 0x77, 0xd0, 0x0a, 0x2c, 0x7e, 0xa2, 0xe1, 0x97, 0xfa, 0x71, 0xbb, - 0xd5, 0x7c, 0xae, 0xb5, 0x5e, 0x2a, 0x39, 0x54, 0x86, 0x42, 0xd2, 0xca, 0xd3, 0x56, 0xb7, 0xd3, - 0xeb, 0x35, 0xf7, 0x5a, 0x94, 0x18, 0x60, 0x5e, 0x68, 0x66, 0xd1, 0x32, 0x94, 0x18, 0x54, 0x08, - 0xe6, 0x76, 0xff, 0x9e, 0x87, 0x25, 0xf9, 0xac, 0xf3, 0x02, 0xf4, 0xbb, 0x1c, 0xac, 0x5e, 0xb0, - 0x8b, 0xa3, 0xdd, 0xcc, 0x6b, 0x6b, 0xe6, 0xd1, 0xb2, 0xf9, 0xc1, 0xb5, 0x30, 0x3c, 0xad, 0xd4, - 0xbb, 0xbf, 0xfc, 0xea, 0x9f, 0xbf, 0xc9, 0x57, 0xd4, 0xd5, 0xe4, 0xeb, 0x7d, 0xf8, 0x44, 0x54, - 0x20, 0xe4, 0x49, 0xee, 0x01, 0xfa, 0x6d, 0x0e, 0x6e, 0x64, 0xa4, 0x3a, 0x7a, 0x7c, 0x69, 0x3e, - 0x67, 0xee, 0x45, 0x9b, 0xc9, 0xe5, 0x46, 0xfa, 0xb0, 0x3e, 0x59, 0x6f, 0xea, 0xfb, 0xac, 0x4b, - 0xf7, 0xd5, 0xdb, 0xb4, 0x4b, 0x43, 0x0a, 0x7c, 0x62, 0x9c, 0x23, 0x7c, 0x92, 0x7b, 0xb0, 0x17, - 0xc1, 0xa6, 0xe9, 0x8d, 0x32, 0x3a, 0xb1, 0xb7, 0x9a, 0x0e, 0x75, 0x97, 0xae, 0xaa, 0x6e, 0xee, - 0x27, 0xcf, 0x84, 0xb9, 0xe5, 0xd1, 0xcb, 0x64, 0xd5, 0x0b, 0xac, 0x1d, 0x8b, 0xb8, 0x6c, 0xcd, - 0xed, 0x70, 0x95, 0xe1, 0xdb, 0xe1, 0xf4, 0xff, 0x07, 0x3c, 0xe5, 0xbf, 0xfe, 0x93, 0xcb, 0x9d, - 0xcc, 0x33, 0xdb, 0x0f, 0xfe, 0x1b, 0x00, 0x00, 0xff, 0xff, 0x7f, 0xbb, 0x90, 0x20, 0x03, 0x21, - 0x00, 0x00, + 0x15, 0x5f, 0x3b, 0x4e, 0x62, 0x3f, 0x3b, 0x89, 0xd2, 0xf9, 0x18, 0x4f, 0xe6, 0x2b, 0xa3, 0x65, + 0xd9, 0xd4, 0x30, 0x24, 0x4c, 0x76, 0x99, 0x5a, 0x66, 0xa6, 0x00, 0xc7, 0x51, 0x12, 0xd7, 0x38, + 0xb6, 0xb7, 0xed, 0xcc, 0x6c, 0x60, 0x0b, 0x95, 0x22, 0xb7, 0x35, 0xda, 0x95, 0x25, 0x21, 0xc9, + 0x33, 0xc9, 0x52, 0xc5, 0x81, 0x2a, 0xce, 0x1c, 0x38, 0x71, 0xe1, 0xc0, 0x01, 0x0e, 0x9c, 0xa8, + 0x82, 0x3b, 0x57, 0x0e, 0x1c, 0xa8, 0xfd, 0x17, 0xf8, 0x0f, 0xa8, 0xa2, 0x38, 0x52, 0xfd, 0x21, + 0xb9, 0xe5, 0x44, 0x99, 0x64, 0xd9, 0x03, 0xa7, 0xb8, 0xdf, 0x7b, 0xbf, 0xdf, 0xeb, 0x7e, 0xdd, + 0xaf, 0xfb, 0x75, 0x2b, 0xf0, 0xd0, 0xf2, 0x3c, 0xcb, 0x21, 0x5b, 0xa6, 0xe3, 0x8d, 0xfa, 0x5b, + 0xaf, 0xed, 0xd0, 0xf6, 0xdc, 0xad, 0xd7, 0x8f, 0xb6, 0xec, 0xa1, 0x61, 0x11, 0xdd, 0x70, 0x5d, + 0x2f, 0x32, 0x22, 0x2f, 0xd8, 0xf4, 0x03, 0x2f, 0xf2, 0xd0, 0x2a, 0xb7, 0xde, 0x64, 0xd6, 0x9b, + 0xdc, 0x7a, 0xf3, 0xf5, 0xa3, 0xb5, 0xdb, 0x82, 0xc5, 0xf0, 0xed, 0x2d, 0x81, 0xb1, 0x3d, 0x37, + 0xe4, 0xa8, 0xb5, 0xf7, 0x32, 0x7c, 0x58, 0xc4, 0x1b, 0x92, 0x28, 0x38, 0x13, 0x66, 0x59, 0x5d, + 0x89, 0xc8, 0x69, 0xa4, 0x8f, 0x59, 0x85, 0xf5, 0x83, 0x0c, 0xeb, 0x37, 0xe4, 0x44, 0xef, 0x93, + 0x88, 0x98, 0x92, 0xed, 0xbb, 0xc2, 0xd6, 0xf1, 0x5c, 0x2b, 0x18, 0xb9, 0xae, 0xed, 0x5a, 0x5b, + 0x9e, 0x4f, 0x82, 0x54, 0x2f, 0xef, 0x09, 0x23, 0xd6, 0x3a, 0x19, 0x0d, 0xb6, 0x22, 0x7b, 0x48, + 0xc2, 0xc8, 0x18, 0xfa, 0xc2, 0xe0, 0x86, 0x30, 0x08, 0x7c, 0x73, 0x2b, 0x8c, 0x8c, 0x68, 0x14, + 0x4e, 0x28, 0xa2, 0x33, 0x9f, 0x6c, 0x99, 0x9e, 0x13, 0x87, 0x6b, 0xad, 0x2a, 0x2b, 0x1c, 0x23, + 0x72, 0x5c, 0x8b, 0x6b, 0xd4, 0x7f, 0xe5, 0x61, 0x76, 0x8f, 0x18, 0xd1, 0x28, 0x20, 0xe8, 0x23, + 0x28, 0x50, 0x83, 0x6a, 0x6e, 0x3d, 0xb7, 0x31, 0xbf, 0xfd, 0x8d, 0xcd, 0x8b, 0x63, 0xbc, 0x29, + 0xcc, 0x37, 0x7b, 0x67, 0x3e, 0xc1, 0x0c, 0x81, 0xee, 0x41, 0x79, 0x68, 0x9c, 0xea, 0x01, 0x09, + 0x47, 0x4e, 0x14, 0x56, 0xf3, 0xeb, 0xb9, 0x8d, 0x69, 0x0c, 0x43, 0xe3, 0x14, 0x73, 0x09, 0x5a, + 0x86, 0xe9, 0xa1, 0xd7, 0x27, 0x4e, 0x75, 0x6a, 0x3d, 0xb7, 0x51, 0xc2, 0xbc, 0xa1, 0xfe, 0x3b, + 0x07, 0x05, 0xca, 0x82, 0x96, 0x41, 0xe9, 0x1d, 0x77, 0x34, 0xfd, 0xa8, 0xd5, 0xed, 0x68, 0xf5, + 0xc6, 0x5e, 0x43, 0xdb, 0x55, 0xde, 0x41, 0x08, 0xe6, 0xf7, 0x6a, 0x75, 0x4d, 0xdf, 0xd5, 0x7a, + 0x5a, 0xbd, 0xd7, 0x68, 0xb7, 0x94, 0x1c, 0x5a, 0x05, 0xd4, 0xac, 0xb5, 0x76, 0x0f, 0x6b, 0xf8, + 0xb9, 0x24, 0xcf, 0x53, 0xdb, 0x66, 0x7b, 0xbf, 0x2d, 0xc9, 0xa6, 0xd0, 0x12, 0x2c, 0x34, 0x6b, + 0x3b, 0x5a, 0x53, 0x12, 0x16, 0xa8, 0x61, 0x4f, 0xfb, 0xa4, 0x27, 0xc9, 0xa6, 0xd1, 0x2d, 0xb8, + 0xb1, 0xdb, 0xae, 0x1f, 0x1d, 0x6a, 0xad, 0x9e, 0x3e, 0xa1, 0x2c, 0xa3, 0x9b, 0xb0, 0xd2, 0xad, + 0xed, 0x69, 0x7a, 0x57, 0xab, 0xe1, 0xfa, 0x81, 0xa4, 0x9a, 0xa1, 0xdd, 0x6e, 0x1c, 0xd6, 0xf6, + 0x35, 0xbd, 0x83, 0xdb, 0x1d, 0x0d, 0xf7, 0x1a, 0x5a, 0x57, 0x99, 0x45, 0xf3, 0x00, 0x75, 0xdc, + 0xee, 0xe8, 0x07, 0x8d, 0x56, 0xaf, 0xab, 0x94, 0xd0, 0x22, 0xcc, 0xbd, 0xd4, 0x76, 0x24, 0x20, + 0xa8, 0x2d, 0x28, 0x37, 0xe8, 0xba, 0xee, 0x7a, 0xa3, 0xc0, 0x24, 0x48, 0x85, 0x39, 0xcb, 0x0c, + 0x75, 0xbe, 0xd4, 0x47, 0x81, 0xcd, 0x66, 0xa0, 0x84, 0xcb, 0x96, 0x19, 0x32, 0xb3, 0xa3, 0xc0, + 0x46, 0xb7, 0xa0, 0x34, 0xd6, 0xe7, 0x99, 0xbe, 0x68, 0x0b, 0xa5, 0xfa, 0x13, 0x98, 0x66, 0x86, + 0xa8, 0x0a, 0xb3, 0xa6, 0xe7, 0x46, 0xc4, 0x8d, 0x18, 0x47, 0x05, 0xc7, 0x4d, 0xf4, 0x14, 0x66, + 0x42, 0xe6, 0x8d, 0x81, 0xcb, 0xdb, 0xef, 0x66, 0x4d, 0xaf, 0xd4, 0x31, 0x2c, 0x20, 0xea, 0x3f, + 0x16, 0x60, 0x7e, 0xcf, 0x30, 0x49, 0x2d, 0x59, 0xfc, 0xa8, 0x01, 0x73, 0x27, 0xde, 0xc8, 0xed, + 0xdb, 0xae, 0xa5, 0xfb, 0x9e, 0x73, 0xc6, 0xfc, 0x95, 0xb3, 0x57, 0xcd, 0x8e, 0x30, 0xee, 0x78, + 0xce, 0x19, 0xae, 0x9c, 0x48, 0x2d, 0xd4, 0x02, 0x65, 0xd0, 0xd7, 0xd3, 0x6c, 0xf9, 0x6b, 0xb0, + 0xcd, 0x0f, 0xfa, 0x72, 0x1b, 0x1d, 0x42, 0xc9, 0x31, 0xdc, 0xfe, 0xd0, 0x08, 0x3e, 0x0f, 0xab, + 0x53, 0xeb, 0x53, 0x1b, 0xe5, 0xed, 0xad, 0xcc, 0xc5, 0x9c, 0x1a, 0xd5, 0x66, 0x53, 0xe0, 0xf0, + 0x98, 0x01, 0xdd, 0x01, 0x08, 0x3c, 0xc7, 0xd1, 0x0d, 0xd7, 0x72, 0x48, 0xb5, 0xb0, 0x9e, 0xdb, + 0xc8, 0xe3, 0x12, 0x95, 0xd4, 0xa8, 0x80, 0x4e, 0x8c, 0x6f, 0xb8, 0x42, 0x3b, 0xcd, 0xb4, 0x45, + 0xdf, 0x70, 0xb9, 0xf2, 0x0e, 0x40, 0x64, 0x3b, 0x91, 0xd0, 0xce, 0x70, 0x2c, 0x95, 0x70, 0xf5, + 0x23, 0x58, 0x4e, 0xb6, 0x08, 0xdd, 0xf4, 0xdc, 0x81, 0xdd, 0x27, 0xae, 0x49, 0xaa, 0xb3, 0xcc, + 0x70, 0x29, 0xd1, 0xd5, 0x13, 0x15, 0xfa, 0x2e, 0xac, 0xc6, 0x5d, 0xa3, 0xc1, 0x92, 0x40, 0x45, + 0x06, 0x5a, 0x91, 0xb4, 0x12, 0xac, 0x01, 0xf3, 0x9f, 0x79, 0x67, 0xba, 0x63, 0x7f, 0x4e, 0x1c, + 0xfb, 0x95, 0xe7, 0xf5, 0xab, 0x25, 0x96, 0xe5, 0x6a, 0x56, 0x60, 0x9a, 0x89, 0x25, 0x9e, 0xfb, + 0xcc, 0x3b, 0x1b, 0x37, 0x51, 0x1b, 0x16, 0x43, 0x2f, 0x08, 0xbc, 0x37, 0x32, 0x1b, 0x5c, 0x99, + 0x4d, 0xe1, 0x60, 0x89, 0xf0, 0x10, 0x14, 0xc3, 0xb5, 0x48, 0x20, 0xf3, 0x95, 0xaf, 0xcc, 0xb7, + 0xc0, 0xb0, 0x12, 0x5d, 0x17, 0x96, 0xc2, 0x51, 0xe0, 0x07, 0x76, 0x48, 0x64, 0xc6, 0xca, 0x95, + 0x19, 0x51, 0x0c, 0x97, 0x48, 0x3f, 0x85, 0xea, 0xc8, 0xed, 0x93, 0x40, 0x27, 0xa7, 0xbe, 0x17, + 0x92, 0xbe, 0xcc, 0x3c, 0x77, 0x65, 0xe6, 0x55, 0xc6, 0xa1, 0x71, 0x0a, 0x89, 0xfd, 0x63, 0x40, + 0x27, 0xce, 0x28, 0x08, 0xd2, 0xbc, 0xf3, 0x57, 0xe6, 0x5d, 0x14, 0xe8, 0x74, 0x14, 0x5e, 0x11, + 0xa3, 0xff, 0x86, 0x18, 0xa9, 0xb8, 0x2e, 0x5c, 0x3d, 0x0a, 0x31, 0x7c, 0x2c, 0x5b, 0xfb, 0xdb, + 0x2c, 0x14, 0xe3, 0x14, 0x41, 0x07, 0xe2, 0xb8, 0x98, 0x62, 0x94, 0x1f, 0x5e, 0x33, 0xc3, 0xe4, + 0xe3, 0xe3, 0x19, 0x14, 0x7d, 0x2f, 0xb4, 0xa9, 0x9e, 0xe5, 0x57, 0x79, 0x7b, 0x3d, 0x8b, 0xad, + 0x23, 0xec, 0x70, 0x82, 0x50, 0xff, 0x3c, 0x33, 0x3e, 0x45, 0x8e, 0x5a, 0xcf, 0x5b, 0xed, 0x97, + 0x2d, 0x3d, 0x3e, 0x23, 0x94, 0x77, 0x50, 0x05, 0x8a, 0x4d, 0x6d, 0xaf, 0xa7, 0x6b, 0xc7, 0x9a, + 0x92, 0x43, 0x73, 0x50, 0xc2, 0x8d, 0xfd, 0x03, 0xde, 0xcc, 0xa3, 0x2a, 0x2c, 0x33, 0x65, 0x7b, + 0x4f, 0x8f, 0x8d, 0x76, 0x70, 0xfb, 0xa5, 0x32, 0x45, 0xb7, 0x7d, 0x6e, 0x38, 0xa9, 0x2a, 0x50, + 0x55, 0x0c, 0x4a, 0xb8, 0x98, 0x6a, 0x1a, 0xad, 0xc1, 0x6a, 0x82, 0x4a, 0xeb, 0x66, 0x28, 0xec, + 0xb0, 0xb1, 0xdb, 0x69, 0x37, 0x5a, 0x3d, 0x7d, 0x47, 0xeb, 0xbd, 0xd4, 0xb4, 0x16, 0xd5, 0xd2, + 0x23, 0xa3, 0x02, 0xc5, 0x56, 0xbb, 0xab, 0xe9, 0xbd, 0x46, 0x47, 0x29, 0xd2, 0x3e, 0x1e, 0x75, + 0x3a, 0x1a, 0xd6, 0x9b, 0x8d, 0x8e, 0x52, 0xa2, 0xcd, 0x66, 0xfb, 0xa5, 0x68, 0x02, 0x3d, 0x5e, + 0x0e, 0xdb, 0x47, 0xbd, 0x03, 0xd6, 0x2b, 0xa5, 0x8c, 0x16, 0xa0, 0xcc, 0xdb, 0xcc, 0x9f, 0x52, + 0x41, 0x0a, 0x54, 0xb8, 0xa0, 0xae, 0xb5, 0x7a, 0x1a, 0x56, 0xe6, 0xd0, 0x0a, 0x2c, 0x32, 0xfa, + 0x9d, 0x76, 0xaf, 0xd7, 0x3e, 0x14, 0x86, 0xf3, 0x34, 0x5e, 0xb2, 0x98, 0xf1, 0x2d, 0xd0, 0x13, + 0x56, 0x96, 0x0a, 0x12, 0x25, 0x19, 0xb5, 0x76, 0xac, 0xe9, 0xbd, 0x76, 0x47, 0xdf, 0x69, 0x1f, + 0xb5, 0x76, 0x6b, 0xf8, 0x58, 0x59, 0x4c, 0xa9, 0xf8, 0xa8, 0xeb, 0x6d, 0xdc, 0xd2, 0xb0, 0x82, + 0xd0, 0x6d, 0xa8, 0x26, 0x2a, 0xc1, 0x98, 0x00, 0x97, 0x92, 0xf0, 0x53, 0x2d, 0xfb, 0x21, 0x70, + 0xcb, 0xe3, 0x40, 0x9e, 0x73, 0xb7, 0x92, 0xd6, 0xa5, 0xfc, 0xad, 0xa2, 0x3b, 0x70, 0x73, 0xac, + 0x9b, 0x74, 0x78, 0x63, 0x3c, 0xab, 0x93, 0x1e, 0xab, 0xe8, 0x1e, 0xdc, 0x92, 0xe7, 0x59, 0xe7, + 0x53, 0x10, 0xcf, 0x98, 0x72, 0x13, 0xad, 0xc3, 0xed, 0xd4, 0x94, 0x4e, 0x5a, 0xac, 0xd1, 0x80, + 0x72, 0x8a, 0x1a, 0xd6, 0x7b, 0xb8, 0xb6, 0x4f, 0x0f, 0xfb, 0x5b, 0x34, 0xfa, 0x02, 0x27, 0x89, + 0x6f, 0xb3, 0x8a, 0x25, 0x1e, 0x7b, 0xe7, 0xa8, 0xd3, 0x68, 0x2a, 0x77, 0x68, 0xc5, 0x32, 0xee, + 0x1e, 0x17, 0xde, 0xa5, 0xf8, 0xbd, 0x36, 0xd6, 0x0e, 0xb4, 0xda, 0xae, 0xbe, 0xcf, 0x0a, 0x9a, + 0x66, 0x4d, 0xb9, 0x47, 0xcb, 0x8a, 0xfa, 0x41, 0xa3, 0xa5, 0xef, 0xb7, 0x6a, 0xbd, 0x03, 0x4a, + 0xb9, 0x4e, 0xfd, 0x33, 0x11, 0xe3, 0xdd, 0x6f, 0xb7, 0xa8, 0xf4, 0x3e, 0xc5, 0x33, 0x29, 0x67, + 0x16, 0x62, 0x55, 0x7d, 0x06, 0x95, 0xa6, 0x67, 0xb2, 0xa4, 0x6c, 0xb8, 0x03, 0x0f, 0x3d, 0x84, + 0x59, 0xc7, 0x88, 0x74, 0xc7, 0xb5, 0xc4, 0x51, 0xbe, 0x14, 0xe7, 0x20, 0xcd, 0xd1, 0xcd, 0xa6, + 0x11, 0x35, 0x5d, 0x0b, 0xcf, 0x38, 0xec, 0xaf, 0xfa, 0x12, 0x8a, 0x9d, 0x80, 0x96, 0xae, 0xd1, + 0x19, 0x42, 0x50, 0x70, 0x8d, 0x21, 0x11, 0x55, 0x0b, 0xfb, 0x4d, 0x0b, 0xbe, 0xd7, 0x86, 0x33, + 0x22, 0xa2, 0x54, 0xe1, 0x0d, 0x74, 0x1f, 0x2a, 0x23, 0xdb, 0x8d, 0x1e, 0x7f, 0xa8, 0x73, 0x25, + 0xdd, 0x3a, 0x0a, 0xb8, 0xcc, 0x65, 0x2f, 0xa8, 0x48, 0xfd, 0xe5, 0x14, 0x28, 0x9a, 0x1b, 0xd9, + 0xd1, 0x99, 0x54, 0x6c, 0x28, 0x30, 0x35, 0xb4, 0xfb, 0xc2, 0x01, 0xfd, 0x89, 0x56, 0x61, 0xc6, + 0xf1, 0x4c, 0xc3, 0x89, 0x1d, 0x88, 0x16, 0x5a, 0x87, 0x72, 0x9f, 0x84, 0x66, 0x60, 0xfb, 0x6c, + 0x37, 0xe1, 0xe5, 0xa6, 0x2c, 0xa2, 0x3d, 0x0b, 0x4d, 0x2f, 0x88, 0x4f, 0x72, 0xde, 0x40, 0x77, + 0x01, 0xa4, 0xa3, 0x94, 0x1f, 0xe3, 0x92, 0x84, 0xea, 0x23, 0xcf, 0xb7, 0x4d, 0xc3, 0xb1, 0xa3, + 0x33, 0x71, 0x90, 0x4b, 0x92, 0xf3, 0xe5, 0xd0, 0xec, 0x57, 0x2e, 0x87, 0x76, 0xa0, 0xe4, 0x88, + 0x89, 0x09, 0xab, 0x45, 0x56, 0xbe, 0x64, 0xd2, 0xc8, 0x33, 0x88, 0xc7, 0x30, 0xf4, 0x43, 0x00, + 0x9f, 0x4f, 0x8f, 0x4d, 0xc2, 0x6a, 0x89, 0x91, 0x64, 0xef, 0xa9, 0x62, 0x22, 0xb1, 0x84, 0x51, + 0xff, 0x9a, 0x87, 0xe5, 0xae, 0x31, 0x20, 0x5d, 0x62, 0x04, 0xe6, 0x2b, 0x69, 0x2e, 0x3e, 0x82, + 0x69, 0xa3, 0x3f, 0x72, 0x22, 0x71, 0x4d, 0xb8, 0xca, 0x51, 0xc2, 0x01, 0x14, 0x19, 0xfa, 0x9e, + 0x37, 0x60, 0x53, 0x76, 0x45, 0x24, 0x03, 0xa0, 0x67, 0x30, 0x3b, 0x24, 0x7d, 0x1a, 0x6b, 0x71, + 0xda, 0x5c, 0x05, 0x1b, 0x43, 0xd0, 0xf7, 0xa1, 0xf8, 0xda, 0xf6, 0x1c, 0x36, 0xb3, 0x85, 0x2b, + 0xc3, 0x13, 0x0c, 0x7a, 0x0c, 0x85, 0xc0, 0x30, 0xcf, 0xae, 0x51, 0x31, 0x31, 0x7b, 0xf5, 0x0d, + 0x94, 0x69, 0xd6, 0x78, 0xae, 0x85, 0x89, 0x19, 0xa1, 0x0f, 0xa0, 0x3c, 0xb4, 0x5d, 0xfd, 0x0a, + 0x49, 0x56, 0x1a, 0xda, 0x2e, 0xff, 0xc9, 0x40, 0xc6, 0x69, 0x02, 0xca, 0x5f, 0x06, 0x32, 0x4e, + 0xf9, 0x4f, 0x35, 0x80, 0x52, 0x9d, 0xde, 0xfe, 0x58, 0x5e, 0x6f, 0xc0, 0x34, 0xbb, 0x0a, 0x0a, + 0x87, 0x28, 0x85, 0x65, 0x66, 0x98, 0x1b, 0x8c, 0x33, 0x23, 0x2f, 0x67, 0xc6, 0x7b, 0x30, 0xef, + 0xdb, 0xa7, 0xc4, 0xd1, 0x07, 0x81, 0x61, 0x26, 0x49, 0x95, 0xc7, 0x73, 0x4c, 0xba, 0x27, 0x84, + 0xea, 0x11, 0x54, 0x77, 0xbd, 0xa1, 0xed, 0x1a, 0x6e, 0xc4, 0x48, 0x43, 0x69, 0xc9, 0x7c, 0x0f, + 0x66, 0x98, 0x87, 0xb0, 0x9a, 0x63, 0x2b, 0xf1, 0x7e, 0x56, 0x08, 0x93, 0x5e, 0x63, 0x01, 0x50, + 0x1d, 0x58, 0x60, 0x17, 0x92, 0x4e, 0xb2, 0x32, 0xd1, 0x31, 0x2c, 0xf4, 0x85, 0x27, 0x3d, 0xa1, + 0xa5, 0x43, 0xfb, 0x4e, 0x16, 0x6d, 0x56, 0xc7, 0xf0, 0x7c, 0x3f, 0xa5, 0x51, 0x7f, 0x9f, 0x83, + 0x62, 0x3d, 0xf0, 0xfc, 0x03, 0xdb, 0x8d, 0xbe, 0xce, 0x1b, 0x4e, 0x7a, 0x77, 0xc9, 0x9f, 0xdb, + 0x5d, 0xb6, 0x60, 0xc9, 0x1e, 0xfa, 0x5e, 0x10, 0x19, 0xae, 0x49, 0x26, 0x03, 0x8d, 0xc6, 0xaa, + 0x24, 0xda, 0x2f, 0x60, 0x29, 0xee, 0xa7, 0x1c, 0xe8, 0x1f, 0x00, 0x98, 0x81, 0xe7, 0xeb, 0xaf, + 0xa8, 0x5c, 0x04, 0x3b, 0x33, 0xed, 0x63, 0x02, 0x5c, 0x32, 0x63, 0x2a, 0xf5, 0x31, 0x2c, 0x24, + 0xbc, 0x1d, 0x23, 0x30, 0x86, 0x21, 0x7a, 0x17, 0xe6, 0x8c, 0xd0, 0x27, 0x66, 0xa4, 0xb3, 0x67, + 0x0a, 0x4e, 0x9b, 0xc7, 0x15, 0x2e, 0xc4, 0x4c, 0xa6, 0xee, 0x02, 0x7a, 0x49, 0x4e, 0x76, 0xe3, + 0xfb, 0x8a, 0x80, 0x6e, 0xc2, 0x92, 0xed, 0x9a, 0xce, 0xa8, 0x4f, 0x74, 0x8b, 0x78, 0xa9, 0xe7, + 0x81, 0x22, 0x5e, 0x14, 0xaa, 0x7d, 0xe2, 0x89, 0x57, 0x02, 0xf5, 0x4f, 0x79, 0xa8, 0xb0, 0xd9, + 0xae, 0xd3, 0x4b, 0xeb, 0x69, 0x84, 0xf6, 0x61, 0x8e, 0xad, 0x7c, 0xcf, 0xb5, 0xf4, 0x80, 0x98, + 0x91, 0x98, 0x82, 0xcc, 0xbb, 0xab, 0x94, 0x6e, 0xb8, 0xec, 0x48, 0xb9, 0xf7, 0x1e, 0xcc, 0x3b, + 0x86, 0x6b, 0x8d, 0xe8, 0x05, 0x9a, 0x07, 0x27, 0xbf, 0x3e, 0xb5, 0x51, 0xc2, 0x73, 0xb1, 0x94, + 0x8d, 0x18, 0x75, 0x61, 0x71, 0x1c, 0x3f, 0xdd, 0x67, 0xa3, 0x10, 0x15, 0xe9, 0xfb, 0x6f, 0x0b, + 0xa3, 0x88, 0x17, 0x5e, 0x30, 0x27, 0x02, 0xf8, 0x29, 0x2c, 0xa7, 0xde, 0x82, 0x62, 0xde, 0x19, + 0xc6, 0xfb, 0x20, 0x8b, 0xf7, 0x7c, 0x3c, 0x31, 0x7a, 0x73, 0x4e, 0xa6, 0x7e, 0x99, 0x83, 0x65, + 0xb1, 0x02, 0x08, 0x8b, 0x1d, 0x26, 0x3f, 0x1d, 0x91, 0x90, 0x6e, 0x37, 0xd3, 0xec, 0x7d, 0x40, + 0xc4, 0xec, 0xce, 0xa5, 0xf7, 0x7d, 0xcc, 0x6d, 0xd1, 0x53, 0x28, 0x0e, 0xf8, 0xf3, 0x0e, 0x8f, + 0x50, 0x79, 0xfb, 0xde, 0x5b, 0x9e, 0x81, 0x70, 0x02, 0xa0, 0x09, 0xc3, 0x9f, 0x28, 0x4c, 0x3e, + 0x7d, 0x6c, 0xfd, 0x5e, 0x92, 0x30, 0xf2, 0x54, 0xe3, 0x8a, 0x2d, 0xb5, 0xd4, 0xe7, 0xb0, 0xca, + 0xb4, 0xe3, 0xb5, 0x1d, 0x2f, 0x09, 0x05, 0xa6, 0xc6, 0x2f, 0x24, 0xf4, 0x27, 0xba, 0x07, 0x65, + 0x9f, 0x7a, 0x75, 0x47, 0xc3, 0x13, 0x12, 0xc4, 0x8f, 0x4f, 0x54, 0xd4, 0x62, 0x12, 0xf5, 0x57, + 0x45, 0x58, 0x99, 0x08, 0x51, 0xe8, 0x7b, 0x6e, 0x48, 0xd0, 0xc7, 0xa0, 0x0c, 0x0c, 0x93, 0x48, + 0x8f, 0x7a, 0x71, 0xd6, 0x7c, 0xf3, 0x6a, 0xd7, 0x19, 0xbc, 0x30, 0x48, 0xb5, 0x43, 0xf4, 0x63, + 0x58, 0x8e, 0x6f, 0xe0, 0x29, 0x5a, 0x1e, 0xcd, 0x8d, 0x2c, 0xda, 0xc9, 0x92, 0x07, 0x2f, 0xc5, + 0x2c, 0x32, 0x79, 0x17, 0x14, 0xc7, 0xb3, 0xbc, 0x14, 0xf1, 0xd4, 0x35, 0x89, 0x17, 0x28, 0x83, + 0x4c, 0x7a, 0x04, 0x8b, 0x8e, 0x71, 0x42, 0x9c, 0x14, 0x6b, 0xe1, 0x9a, 0xac, 0x0a, 0xa3, 0x98, + 0xe8, 0xeb, 0xc4, 0x83, 0x69, 0x58, 0x9d, 0xbe, 0x6e, 0x5f, 0x29, 0x83, 0x4c, 0xfa, 0x09, 0x2c, + 0x0f, 0x46, 0x8e, 0xa3, 0x4f, 0x30, 0xb3, 0xcb, 0xfd, 0x25, 0x93, 0xd6, 0x4b, 0xd1, 0x60, 0x44, + 0x39, 0xd2, 0x32, 0x74, 0x02, 0xab, 0xa1, 0x31, 0x20, 0x7a, 0xc8, 0xea, 0x1d, 0x99, 0x9b, 0xe7, + 0xe9, 0xc3, 0x2c, 0xee, 0x8b, 0x8a, 0x24, 0xbc, 0x1c, 0x5e, 0x54, 0x3a, 0x59, 0x70, 0x8b, 0x27, + 0xc8, 0xb8, 0xce, 0x92, 0x1d, 0x15, 0x2f, 0xdf, 0x68, 0x26, 0xce, 0x41, 0x7c, 0xd3, 0x4e, 0x0b, + 0x24, 0x47, 0x3a, 0xac, 0x48, 0xfb, 0x98, 0xe4, 0xa2, 0xcc, 0x5c, 0x7c, 0xeb, 0xad, 0x7b, 0x99, + 0xbc, 0x10, 0xcd, 0x0b, 0x0e, 0x9a, 0x06, 0xcc, 0xa5, 0xf6, 0x34, 0xf6, 0x06, 0x72, 0x49, 0xaa, + 0xcb, 0x9b, 0x19, 0xae, 0xc8, 0xdb, 0x18, 0xad, 0x4f, 0x48, 0x10, 0x78, 0x01, 0x2b, 0xaf, 0xa4, + 0xfa, 0x24, 0xf0, 0xcd, 0xcd, 0x2e, 0x7b, 0xdd, 0xc6, 0xdc, 0x00, 0x1d, 0x88, 0xc7, 0xcd, 0xd3, + 0xa8, 0xba, 0xc2, 0x6c, 0x37, 0x2f, 0x0d, 0xd5, 0xb9, 0xbd, 0x03, 0xc7, 0x70, 0xf5, 0x8f, 0xd2, + 0xa6, 0xb9, 0x67, 0x3b, 0xe3, 0x0d, 0x61, 0x0f, 0x2a, 0xb6, 0xeb, 0x8f, 0x22, 0xfe, 0xae, 0x66, + 0xbd, 0xed, 0xbc, 0x69, 0x50, 0x5b, 0xf6, 0xca, 0x66, 0xe1, 0xb2, 0x3d, 0x6e, 0xa0, 0xe7, 0x50, + 0x0a, 0x04, 0x67, 0x9c, 0xfa, 0xdf, 0xce, 0x22, 0xb9, 0x70, 0x6b, 0xc2, 0x63, 0xbc, 0x3a, 0x80, + 0xb5, 0x1d, 0x23, 0x4a, 0x56, 0x12, 0x37, 0x0c, 0xe3, 0x7d, 0xfe, 0x00, 0x8a, 0x01, 0xff, 0x19, + 0xef, 0x5d, 0x0f, 0xaf, 0xe8, 0x89, 0x81, 0x70, 0x82, 0x56, 0x3f, 0x83, 0x5b, 0x17, 0xfa, 0x11, + 0xb1, 0x49, 0x8d, 0x29, 0xf7, 0x3f, 0x8e, 0xe9, 0x2f, 0x79, 0xa8, 0xd6, 0xc2, 0x33, 0xd7, 0x4c, + 0x4f, 0x03, 0x1f, 0xd2, 0xd7, 0x35, 0x0b, 0xff, 0x27, 0xa7, 0x19, 0xa5, 0xf2, 0x46, 0x91, 0x34, + 0xa0, 0xc2, 0xe5, 0x54, 0x6d, 0x66, 0x2c, 0x46, 0x54, 0xf1, 0xa4, 0x96, 0x3a, 0x80, 0x9b, 0x17, + 0x84, 0x4d, 0xcc, 0xd0, 0x39, 0x3f, 0xb9, 0xaf, 0xec, 0xc7, 0x85, 0xbb, 0xcc, 0x4f, 0x6a, 0x41, + 0x50, 0x67, 0xc9, 0xba, 0x6b, 0x9e, 0x5b, 0x77, 0x99, 0xf5, 0x77, 0xd6, 0x44, 0x4b, 0x6b, 0x2f, + 0x80, 0x7b, 0x99, 0xfe, 0xc4, 0xe8, 0xda, 0xe7, 0xd7, 0xdf, 0xa3, 0x6b, 0x78, 0x3c, 0xbf, 0x06, + 0x1d, 0x28, 0x4b, 0x4b, 0x87, 0xde, 0x99, 0x2d, 0x33, 0xd4, 0xc5, 0x57, 0x12, 0x1e, 0xba, 0xcc, + 0x9b, 0xca, 0xbe, 0x19, 0x8a, 0x6f, 0x24, 0x25, 0x2b, 0xfe, 0x89, 0x6e, 0x41, 0x69, 0x68, 0x0f, + 0x89, 0xce, 0x9e, 0x45, 0xc5, 0x37, 0x1a, 0x2a, 0xe8, 0x9d, 0xf9, 0x44, 0xfd, 0x39, 0x54, 0xe4, + 0x78, 0xa3, 0x36, 0x2c, 0x50, 0x77, 0x7d, 0x12, 0x46, 0xb6, 0xcb, 0x77, 0xe7, 0xdc, 0xe5, 0xa7, + 0xd8, 0xbe, 0x19, 0xee, 0x8e, 0xad, 0xf1, 0xbc, 0x95, 0x6a, 0xa3, 0x3b, 0x00, 0x27, 0x34, 0x7a, + 0x7a, 0x68, 0x7f, 0x41, 0x44, 0x19, 0x54, 0x62, 0x92, 0xae, 0xfd, 0x05, 0x51, 0xef, 0x40, 0x29, + 0xe9, 0xf4, 0xf9, 0x2a, 0x4a, 0x55, 0x61, 0x3e, 0xcd, 0x7f, 0x81, 0xcd, 0x1f, 0xf2, 0xb0, 0xd8, + 0x8e, 0x3f, 0x57, 0x1e, 0x92, 0xc8, 0xe8, 0x1b, 0x91, 0x81, 0x34, 0x98, 0x0e, 0x69, 0x98, 0xc5, + 0x83, 0x40, 0xe6, 0xa7, 0x96, 0x73, 0x48, 0xb6, 0xbd, 0x13, 0xcc, 0xd1, 0xe8, 0x29, 0x94, 0xcd, + 0x80, 0x18, 0x11, 0xd1, 0x23, 0x7b, 0xc8, 0x9f, 0x60, 0xca, 0xdb, 0x6b, 0x31, 0x59, 0xfc, 0x31, + 0x74, 0xb3, 0x17, 0x7f, 0x0c, 0xc5, 0xc0, 0xcd, 0xa9, 0x80, 0x82, 0x47, 0x7e, 0x3f, 0x01, 0xcf, + 0xbc, 0x1d, 0xcc, 0xcd, 0xa9, 0x40, 0xfd, 0x18, 0xa6, 0x59, 0x4f, 0xd0, 0x0a, 0x2c, 0x76, 0x7b, + 0xb5, 0xde, 0xe4, 0x77, 0xc8, 0x32, 0xcc, 0xd6, 0xb1, 0x56, 0xeb, 0x69, 0xbb, 0x4a, 0x8e, 0x36, + 0xf0, 0x51, 0xab, 0xd5, 0x68, 0xed, 0x2b, 0x79, 0x54, 0x84, 0xc2, 0x6e, 0xbb, 0xa5, 0x29, 0x53, + 0x68, 0x0e, 0x4a, 0xf5, 0x5a, 0xab, 0xae, 0x35, 0x9b, 0xda, 0xae, 0x52, 0x78, 0x40, 0x00, 0xa4, + 0xb7, 0xf8, 0x32, 0xcc, 0x8a, 0x87, 0x69, 0xe5, 0x1d, 0xb4, 0x08, 0x73, 0x2f, 0x34, 0x7c, 0xac, + 0x1f, 0xb5, 0x9a, 0x8d, 0xe7, 0x5a, 0xf3, 0x58, 0xc9, 0xa1, 0x0a, 0x14, 0x93, 0x56, 0x9e, 0xb6, + 0x3a, 0xed, 0x6e, 0xb7, 0xb1, 0xd3, 0xa4, 0xc4, 0x00, 0x33, 0x42, 0x53, 0x40, 0x0b, 0x50, 0x66, + 0x50, 0x21, 0x98, 0xde, 0xfe, 0x7b, 0x1e, 0xe6, 0xe5, 0xb3, 0xce, 0x0b, 0xd0, 0x6f, 0x73, 0xb0, + 0x74, 0xc1, 0x2e, 0x8e, 0xb6, 0x33, 0xaf, 0xad, 0x99, 0x47, 0xcb, 0xda, 0x07, 0xd7, 0xc2, 0xf0, + 0xb4, 0x52, 0xef, 0xfe, 0xe2, 0xcb, 0x7f, 0xfe, 0x3a, 0x5f, 0x55, 0x97, 0x92, 0xaf, 0xf7, 0xe1, + 0x13, 0x51, 0x81, 0x90, 0x27, 0xb9, 0x07, 0xe8, 0x37, 0x39, 0xb8, 0x91, 0x91, 0xea, 0xe8, 0xf1, + 0xa5, 0xf9, 0x9c, 0xb9, 0x17, 0xad, 0x25, 0x97, 0x1b, 0xe9, 0xc3, 0xfa, 0x78, 0xbd, 0xa9, 0xef, + 0xb3, 0x2e, 0xdd, 0x57, 0x6f, 0xd3, 0x2e, 0x0d, 0x28, 0xf0, 0x89, 0x71, 0x8e, 0xf0, 0x49, 0xee, + 0xc1, 0xce, 0xcf, 0x60, 0xcd, 0xf4, 0x86, 0x19, 0x9d, 0xd8, 0x59, 0x4a, 0x87, 0xba, 0x43, 0x57, + 0x55, 0x27, 0xf7, 0xa3, 0x67, 0xc2, 0xdc, 0xf2, 0xe8, 0x65, 0x72, 0xd3, 0x0b, 0xac, 0x2d, 0x8b, + 0xb8, 0x6c, 0xcd, 0x6d, 0x71, 0x95, 0xe1, 0xdb, 0xe1, 0xe4, 0xff, 0x07, 0x3c, 0xe5, 0xbf, 0xfe, + 0x93, 0xcb, 0xfd, 0x2e, 0x5f, 0xd8, 0xaf, 0xbf, 0x68, 0x9d, 0xcc, 0x30, 0xc8, 0x07, 0xff, 0x0d, + 0x00, 0x00, 0xff, 0xff, 0x7a, 0xea, 0x20, 0x54, 0x0a, 0x21, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/text_annotation.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/text_annotation.pb.go index 9346860f0..457b369b7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/text_annotation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/text_annotation.pb.go @@ -59,7 +59,7 @@ func (x TextAnnotation_DetectedBreak_BreakType) String() string { return proto.EnumName(TextAnnotation_DetectedBreak_BreakType_name, int32(x)) } func (TextAnnotation_DetectedBreak_BreakType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{0, 1, 0} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{0, 1, 0} } // Type of a block (text, image etc) as identified by OCR. @@ -101,7 +101,7 @@ func (x Block_BlockType) String() string { return proto.EnumName(Block_BlockType_name, int32(x)) } func (Block_BlockType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{2, 0} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{2, 0} } // TextAnnotation contains a structured representation of OCR extracted text. @@ -113,9 +113,9 @@ func (Block_BlockType) EnumDescriptor() ([]byte, []int) { // detail. type TextAnnotation struct { // List of pages detected by OCR. - Pages []*Page `protobuf:"bytes,1,rep,name=pages" json:"pages,omitempty"` + Pages []*Page `protobuf:"bytes,1,rep,name=pages,proto3" json:"pages,omitempty"` // UTF-8 text detected on the pages. - Text string `protobuf:"bytes,2,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,2,opt,name=text,proto3" json:"text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -125,7 +125,7 @@ func (m *TextAnnotation) Reset() { *m = TextAnnotation{} } func (m *TextAnnotation) String() string { return proto.CompactTextString(m) } func (*TextAnnotation) ProtoMessage() {} func (*TextAnnotation) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{0} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{0} } func (m *TextAnnotation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TextAnnotation.Unmarshal(m, b) @@ -164,9 +164,9 @@ type TextAnnotation_DetectedLanguage struct { // The BCP-47 language code, such as "en-US" or "sr-Latn". For more // information, see // http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Confidence of detected language. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -176,7 +176,7 @@ func (m *TextAnnotation_DetectedLanguage) Reset() { *m = TextAnnotation_ func (m *TextAnnotation_DetectedLanguage) String() string { return proto.CompactTextString(m) } func (*TextAnnotation_DetectedLanguage) ProtoMessage() {} func (*TextAnnotation_DetectedLanguage) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{0, 0} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{0, 0} } func (m *TextAnnotation_DetectedLanguage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TextAnnotation_DetectedLanguage.Unmarshal(m, b) @@ -213,9 +213,9 @@ func (m *TextAnnotation_DetectedLanguage) GetConfidence() float32 { // Detected start or end of a structural component. type TextAnnotation_DetectedBreak struct { // Detected break type. - Type TextAnnotation_DetectedBreak_BreakType `protobuf:"varint,1,opt,name=type,enum=google.cloud.vision.v1.TextAnnotation_DetectedBreak_BreakType" json:"type,omitempty"` + Type TextAnnotation_DetectedBreak_BreakType `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.vision.v1.TextAnnotation_DetectedBreak_BreakType" json:"type,omitempty"` // True if break prepends the element. - IsPrefix bool `protobuf:"varint,2,opt,name=is_prefix,json=isPrefix" json:"is_prefix,omitempty"` + IsPrefix bool `protobuf:"varint,2,opt,name=is_prefix,json=isPrefix,proto3" json:"is_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -225,7 +225,7 @@ func (m *TextAnnotation_DetectedBreak) Reset() { *m = TextAnnotation_Det func (m *TextAnnotation_DetectedBreak) String() string { return proto.CompactTextString(m) } func (*TextAnnotation_DetectedBreak) ProtoMessage() {} func (*TextAnnotation_DetectedBreak) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{0, 1} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{0, 1} } func (m *TextAnnotation_DetectedBreak) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TextAnnotation_DetectedBreak.Unmarshal(m, b) @@ -262,9 +262,9 @@ func (m *TextAnnotation_DetectedBreak) GetIsPrefix() bool { // Additional information detected on the structural component. type TextAnnotation_TextProperty struct { // A list of detected languages together with confidence. - DetectedLanguages []*TextAnnotation_DetectedLanguage `protobuf:"bytes,1,rep,name=detected_languages,json=detectedLanguages" json:"detected_languages,omitempty"` + DetectedLanguages []*TextAnnotation_DetectedLanguage `protobuf:"bytes,1,rep,name=detected_languages,json=detectedLanguages,proto3" json:"detected_languages,omitempty"` // Detected start or end of a text segment. - DetectedBreak *TextAnnotation_DetectedBreak `protobuf:"bytes,2,opt,name=detected_break,json=detectedBreak" json:"detected_break,omitempty"` + DetectedBreak *TextAnnotation_DetectedBreak `protobuf:"bytes,2,opt,name=detected_break,json=detectedBreak,proto3" json:"detected_break,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -274,7 +274,7 @@ func (m *TextAnnotation_TextProperty) Reset() { *m = TextAnnotation_Text func (m *TextAnnotation_TextProperty) String() string { return proto.CompactTextString(m) } func (*TextAnnotation_TextProperty) ProtoMessage() {} func (*TextAnnotation_TextProperty) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{0, 2} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{0, 2} } func (m *TextAnnotation_TextProperty) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TextAnnotation_TextProperty.Unmarshal(m, b) @@ -311,17 +311,17 @@ func (m *TextAnnotation_TextProperty) GetDetectedBreak() *TextAnnotation_Detecte // Detected page from OCR. type Page struct { // Additional information detected on the page. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // Page width. For PDFs the unit is points. For images (including // TIFFs) the unit is pixels. - Width int32 `protobuf:"varint,2,opt,name=width" json:"width,omitempty"` + Width int32 `protobuf:"varint,2,opt,name=width,proto3" json:"width,omitempty"` // Page height. For PDFs the unit is points. For images (including // TIFFs) the unit is pixels. - Height int32 `protobuf:"varint,3,opt,name=height" json:"height,omitempty"` + Height int32 `protobuf:"varint,3,opt,name=height,proto3" json:"height,omitempty"` // List of blocks of text, images etc on this page. - Blocks []*Block `protobuf:"bytes,4,rep,name=blocks" json:"blocks,omitempty"` + Blocks []*Block `protobuf:"bytes,4,rep,name=blocks,proto3" json:"blocks,omitempty"` // Confidence of the OCR results on the page. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -331,7 +331,7 @@ func (m *Page) Reset() { *m = Page{} } func (m *Page) String() string { return proto.CompactTextString(m) } func (*Page) ProtoMessage() {} func (*Page) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{1} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{1} } func (m *Page) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Page.Unmarshal(m, b) @@ -389,7 +389,7 @@ func (m *Page) GetConfidence() float32 { // Logical element on the page. type Block struct { // Additional information detected for the block. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the block. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -410,13 +410,13 @@ type Block struct { // 1----0 // // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of paragraphs in this block (if this blocks is of type text). - Paragraphs []*Paragraph `protobuf:"bytes,3,rep,name=paragraphs" json:"paragraphs,omitempty"` + Paragraphs []*Paragraph `protobuf:"bytes,3,rep,name=paragraphs,proto3" json:"paragraphs,omitempty"` // Detected block type (text, image etc) for this block. - BlockType Block_BlockType `protobuf:"varint,4,opt,name=block_type,json=blockType,enum=google.cloud.vision.v1.Block_BlockType" json:"block_type,omitempty"` + BlockType Block_BlockType `protobuf:"varint,4,opt,name=block_type,json=blockType,proto3,enum=google.cloud.vision.v1.Block_BlockType" json:"block_type,omitempty"` // Confidence of the OCR results on the block. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -426,7 +426,7 @@ func (m *Block) Reset() { *m = Block{} } func (m *Block) String() string { return proto.CompactTextString(m) } func (*Block) ProtoMessage() {} func (*Block) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{2} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{2} } func (m *Block) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Block.Unmarshal(m, b) @@ -484,7 +484,7 @@ func (m *Block) GetConfidence() float32 { // Structural unit of text representing a number of words in certain order. type Paragraph struct { // Additional information detected for the paragraph. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the paragraph. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -500,11 +500,11 @@ type Paragraph struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of words in this paragraph. - Words []*Word `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` + Words []*Word `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"` // Confidence of the OCR results for the paragraph. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -514,7 +514,7 @@ func (m *Paragraph) Reset() { *m = Paragraph{} } func (m *Paragraph) String() string { return proto.CompactTextString(m) } func (*Paragraph) ProtoMessage() {} func (*Paragraph) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{3} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{3} } func (m *Paragraph) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Paragraph.Unmarshal(m, b) @@ -565,7 +565,7 @@ func (m *Paragraph) GetConfidence() float32 { // A word representation. type Word struct { // Additional information detected for the word. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the word. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -581,12 +581,12 @@ type Word struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of symbols in the word. // The order of the symbols follows the natural reading order. - Symbols []*Symbol `protobuf:"bytes,3,rep,name=symbols" json:"symbols,omitempty"` + Symbols []*Symbol `protobuf:"bytes,3,rep,name=symbols,proto3" json:"symbols,omitempty"` // Confidence of the OCR results for the word. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -596,7 +596,7 @@ func (m *Word) Reset() { *m = Word{} } func (m *Word) String() string { return proto.CompactTextString(m) } func (*Word) ProtoMessage() {} func (*Word) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{4} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{4} } func (m *Word) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Word.Unmarshal(m, b) @@ -647,7 +647,7 @@ func (m *Word) GetConfidence() float32 { // A single symbol representation. type Symbol struct { // Additional information detected for the symbol. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the symbol. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -663,11 +663,11 @@ type Symbol struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // The actual UTF-8 representation of the symbol. - Text string `protobuf:"bytes,3,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,3,opt,name=text,proto3" json:"text,omitempty"` // Confidence of the OCR results for the symbol. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -677,7 +677,7 @@ func (m *Symbol) Reset() { *m = Symbol{} } func (m *Symbol) String() string { return proto.CompactTextString(m) } func (*Symbol) ProtoMessage() {} func (*Symbol) Descriptor() ([]byte, []int) { - return fileDescriptor_text_annotation_df7c6c14e8dd9209, []int{5} + return fileDescriptor_text_annotation_80dbf5518322be1d, []int{5} } func (m *Symbol) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Symbol.Unmarshal(m, b) @@ -740,57 +740,58 @@ func init() { } func init() { - proto.RegisterFile("google/cloud/vision/v1/text_annotation.proto", fileDescriptor_text_annotation_df7c6c14e8dd9209) + proto.RegisterFile("google/cloud/vision/v1/text_annotation.proto", fileDescriptor_text_annotation_80dbf5518322be1d) } -var fileDescriptor_text_annotation_df7c6c14e8dd9209 = []byte{ - // 763 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x56, 0xcf, 0x6f, 0xd3, 0x4a, - 0x10, 0x7e, 0x6e, 0xec, 0x34, 0x9e, 0xb4, 0x91, 0xdf, 0xbe, 0xa7, 0x2a, 0x0a, 0xa5, 0x2a, 0x01, - 0x44, 0x0f, 0xc8, 0x51, 0x53, 0x10, 0x48, 0x20, 0xa4, 0x38, 0x35, 0xb4, 0x6a, 0x94, 0x58, 0xdb, - 0x44, 0xe5, 0xc7, 0xc1, 0xf2, 0x8f, 0xad, 0x63, 0x35, 0xf5, 0x5a, 0xb6, 0xdb, 0x26, 0xff, 0x0d, - 0xff, 0x13, 0x12, 0x27, 0xae, 0x9c, 0xb9, 0x02, 0x27, 0xe4, 0xb5, 0x9d, 0x26, 0x01, 0x53, 0x40, - 0x1c, 0x7a, 0xb1, 0x76, 0x26, 0xdf, 0x7c, 0x33, 0xdf, 0x8c, 0x27, 0x6b, 0xb8, 0xef, 0x50, 0xea, - 0x8c, 0x48, 0xc3, 0x1a, 0xd1, 0x33, 0xbb, 0x71, 0xee, 0x86, 0x2e, 0xf5, 0x1a, 0xe7, 0xdb, 0x8d, - 0x88, 0x8c, 0x23, 0xdd, 0xf0, 0x3c, 0x1a, 0x19, 0x91, 0x4b, 0x3d, 0xd9, 0x0f, 0x68, 0x44, 0xd1, - 0x5a, 0x82, 0x96, 0x19, 0x5a, 0x4e, 0xd0, 0xf2, 0xf9, 0x76, 0x6d, 0x3d, 0x65, 0x31, 0x7c, 0xb7, - 0x71, 0x19, 0x14, 0x26, 0x51, 0xb5, 0xbb, 0x39, 0x39, 0x1c, 0x42, 0x4f, 0x49, 0x14, 0x4c, 0x12, - 0x58, 0xfd, 0x13, 0x0f, 0x95, 0x3e, 0x19, 0x47, 0xad, 0x29, 0x01, 0x6a, 0x82, 0xe0, 0x1b, 0x0e, - 0x09, 0xab, 0xdc, 0x66, 0x61, 0xab, 0xdc, 0x5c, 0x97, 0x7f, 0x9c, 0x5f, 0xd6, 0x0c, 0x87, 0xe0, - 0x04, 0x8a, 0x10, 0xf0, 0x71, 0xf1, 0xd5, 0xa5, 0x4d, 0x6e, 0x4b, 0xc4, 0xec, 0x5c, 0x3b, 0x02, - 0x69, 0x97, 0x44, 0xc4, 0x8a, 0x88, 0xdd, 0x31, 0x3c, 0xe7, 0xcc, 0x70, 0x08, 0xba, 0x0d, 0xab, - 0xa3, 0xf4, 0xac, 0x5b, 0xd4, 0x26, 0x55, 0x8e, 0x05, 0xac, 0x64, 0xce, 0x36, 0xb5, 0x09, 0xda, - 0x00, 0xb0, 0xa8, 0x77, 0xec, 0xda, 0xc4, 0xb3, 0x08, 0xa3, 0x5c, 0xc2, 0x33, 0x9e, 0xda, 0x47, - 0x0e, 0x56, 0x33, 0x66, 0x25, 0x20, 0xc6, 0x09, 0xc2, 0xc0, 0x47, 0x13, 0x3f, 0x61, 0xab, 0x34, - 0x9f, 0xe5, 0x55, 0x3c, 0x2f, 0x54, 0x9e, 0xe3, 0x90, 0xd9, 0xb3, 0x3f, 0xf1, 0x09, 0x66, 0x5c, - 0xe8, 0x06, 0x88, 0x6e, 0xa8, 0xfb, 0x01, 0x39, 0x76, 0xc7, 0xac, 0x88, 0x12, 0x2e, 0xb9, 0xa1, - 0xc6, 0xec, 0xba, 0x05, 0xe2, 0x14, 0x8f, 0xca, 0xb0, 0x3c, 0xe8, 0x1e, 0x74, 0x7b, 0x47, 0x5d, - 0xe9, 0x1f, 0x24, 0x82, 0x70, 0xa8, 0xb5, 0xda, 0xaa, 0xc4, 0xa1, 0x0a, 0xc0, 0xe1, 0x00, 0xab, - 0x7a, 0x62, 0x2f, 0x21, 0x04, 0x15, 0xb5, 0xd7, 0xd1, 0x67, 0x7c, 0x05, 0x04, 0x50, 0xdc, 0x7b, - 0xa5, 0xed, 0xa9, 0x5d, 0x89, 0x8f, 0xf1, 0x9d, 0xfd, 0xae, 0xaa, 0x2b, 0x58, 0x6d, 0x1d, 0x48, - 0x42, 0xed, 0x1d, 0x07, 0x2b, 0x71, 0xc9, 0x5a, 0x40, 0x7d, 0x12, 0x44, 0x13, 0x74, 0x0c, 0xc8, - 0x4e, 0x6b, 0xd6, 0xb3, 0x8e, 0x65, 0x63, 0x7a, 0xf4, 0x9b, 0xa2, 0xb3, 0x91, 0xe0, 0x7f, 0xed, - 0x05, 0x4f, 0x88, 0xde, 0x40, 0x65, 0x9a, 0xc7, 0x8c, 0x65, 0x32, 0xfd, 0xe5, 0xe6, 0x83, 0x3f, - 0x69, 0x2c, 0x5e, 0xb5, 0x67, 0xcd, 0xfa, 0x07, 0x0e, 0xf8, 0xf8, 0xd5, 0x41, 0x3d, 0x28, 0xf9, - 0xa9, 0x32, 0x36, 0xb8, 0x72, 0x73, 0xe7, 0x17, 0xf9, 0x67, 0x9b, 0x82, 0xa7, 0x24, 0xe8, 0x7f, - 0x10, 0x2e, 0x5c, 0x3b, 0x1a, 0xb2, 0x6a, 0x05, 0x9c, 0x18, 0x68, 0x0d, 0x8a, 0x43, 0xe2, 0x3a, - 0xc3, 0xa8, 0x5a, 0x60, 0xee, 0xd4, 0x42, 0x0f, 0xa1, 0x68, 0x8e, 0xa8, 0x75, 0x12, 0x56, 0x79, - 0xd6, 0xc0, 0x9b, 0x79, 0xc9, 0x95, 0x18, 0x85, 0x53, 0xf0, 0xc2, 0xcb, 0x29, 0x2c, 0xbe, 0x9c, - 0xf5, 0xb7, 0x05, 0x10, 0x58, 0xc4, 0xdf, 0xd7, 0xf7, 0x02, 0x56, 0x4c, 0x7a, 0xe6, 0xd9, 0xae, - 0xe7, 0xe8, 0x26, 0x1d, 0xa7, 0x43, 0xb9, 0x93, 0x5b, 0x77, 0x8a, 0xd5, 0xe8, 0x68, 0x82, 0xcb, - 0x59, 0xa4, 0x42, 0xc7, 0xa8, 0x05, 0xe0, 0x1b, 0x81, 0xe1, 0x04, 0x86, 0x3f, 0x0c, 0xab, 0x05, - 0x26, 0xff, 0x56, 0xfe, 0x9a, 0xa7, 0x48, 0x3c, 0x13, 0x84, 0x9e, 0x03, 0xb0, 0x86, 0xe8, 0x6c, - 0xef, 0x78, 0xb6, 0x77, 0xf7, 0x7e, 0xda, 0xc1, 0xe4, 0xc9, 0x16, 0x4c, 0x34, 0xb3, 0xe3, 0x95, - 0xed, 0xc4, 0x20, 0x4e, 0xe3, 0xe6, 0x17, 0xad, 0x04, 0x7c, 0x5f, 0x7d, 0xd9, 0x97, 0xb8, 0x78, - 0xe5, 0xfa, 0x2d, 0xa5, 0x13, 0xaf, 0x58, 0x19, 0x96, 0xb5, 0xfd, 0x76, 0x7f, 0x80, 0xe3, 0xdd, - 0x12, 0x41, 0xc0, 0x83, 0x8e, 0x8a, 0x25, 0x3e, 0xf6, 0x2b, 0x2d, 0xdc, 0xee, 0xed, 0xaa, 0x92, - 0x50, 0xff, 0xc2, 0x81, 0x38, 0x55, 0x75, 0x8d, 0xc7, 0xd4, 0x04, 0xe1, 0x82, 0x06, 0x76, 0x36, - 0xa1, 0xdc, 0x3f, 0xe2, 0x23, 0x1a, 0xd8, 0x38, 0x81, 0x2e, 0xf4, 0x93, 0xff, 0xae, 0x9f, 0x5f, - 0x39, 0xe0, 0x63, 0xfc, 0x35, 0x96, 0xfd, 0x18, 0x96, 0xc3, 0xc9, 0xa9, 0x49, 0x47, 0x99, 0xf0, - 0x8d, 0x3c, 0x8e, 0x43, 0x06, 0xc3, 0x19, 0xfc, 0x4a, 0xf1, 0xef, 0x39, 0x28, 0x26, 0x31, 0xd7, - 0x58, 0x7e, 0x76, 0x95, 0x16, 0x2e, 0xaf, 0xd2, 0xab, 0x84, 0x29, 0x11, 0xd4, 0x2c, 0x7a, 0x9a, - 0x93, 0x4b, 0xf9, 0x6f, 0x5e, 0x81, 0x16, 0x5f, 0xfc, 0x1a, 0xf7, 0xfa, 0x69, 0x0a, 0x77, 0x68, - 0x7c, 0x97, 0xc8, 0x34, 0x70, 0x1a, 0x0e, 0xf1, 0xd8, 0x67, 0x41, 0x23, 0xf9, 0xc9, 0xf0, 0xdd, - 0x70, 0xf1, 0x03, 0xe2, 0x49, 0x72, 0xfa, 0xcc, 0x71, 0x66, 0x91, 0x61, 0x77, 0xbe, 0x05, 0x00, - 0x00, 0xff, 0xff, 0x3f, 0x4a, 0xe7, 0xb0, 0xcf, 0x08, 0x00, 0x00, +var fileDescriptor_text_annotation_80dbf5518322be1d = []byte{ + // 774 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x56, 0xcf, 0x6f, 0xd3, 0x58, + 0x10, 0x5e, 0x27, 0x76, 0x1a, 0x4f, 0xda, 0xc8, 0xfb, 0x76, 0x55, 0x45, 0xd9, 0x6e, 0xd5, 0xcd, + 0x82, 0xe8, 0x01, 0x39, 0x6a, 0x0a, 0x02, 0x09, 0x84, 0x14, 0xa7, 0xa6, 0xad, 0x1a, 0x25, 0xd6, + 0x6b, 0x42, 0xf9, 0x71, 0xb0, 0x1c, 0xfb, 0xd5, 0xb1, 0x9a, 0xfa, 0x59, 0xb6, 0xdb, 0x26, 0xe2, + 0x9f, 0x41, 0xfc, 0x4b, 0x48, 0x9c, 0xb8, 0x72, 0xe6, 0x0a, 0x9c, 0x90, 0x9f, 0xed, 0x34, 0x09, + 0x98, 0x02, 0xe2, 0xd0, 0x8b, 0xf5, 0x66, 0xf2, 0xcd, 0x37, 0xf3, 0xcd, 0x78, 0xf2, 0x0c, 0xb7, + 0x6d, 0x4a, 0xed, 0x11, 0xa9, 0x9b, 0x23, 0x7a, 0x66, 0xd5, 0xcf, 0x9d, 0xc0, 0xa1, 0x6e, 0xfd, + 0x7c, 0xab, 0x1e, 0x92, 0x71, 0xa8, 0x1b, 0xae, 0x4b, 0x43, 0x23, 0x74, 0xa8, 0x2b, 0x7b, 0x3e, + 0x0d, 0x29, 0x5a, 0x8d, 0xd1, 0x32, 0x43, 0xcb, 0x31, 0x5a, 0x3e, 0xdf, 0xaa, 0xae, 0x25, 0x2c, + 0x86, 0xe7, 0xd4, 0x2f, 0x83, 0x82, 0x38, 0xaa, 0x7a, 0x33, 0x23, 0x87, 0x4d, 0xe8, 0x29, 0x09, + 0xfd, 0x49, 0x0c, 0xab, 0x7d, 0xe0, 0xa1, 0xdc, 0x23, 0xe3, 0xb0, 0x39, 0x25, 0x40, 0x0d, 0x10, + 0x3c, 0xc3, 0x26, 0x41, 0x85, 0xdb, 0xc8, 0x6f, 0x96, 0x1a, 0x6b, 0xf2, 0xb7, 0xf3, 0xcb, 0x9a, + 0x61, 0x13, 0x1c, 0x43, 0x11, 0x02, 0x3e, 0x2a, 0xbe, 0x92, 0xdb, 0xe0, 0x36, 0x45, 0xcc, 0xce, + 0xd5, 0x23, 0x90, 0x76, 0x48, 0x48, 0xcc, 0x90, 0x58, 0x6d, 0xc3, 0xb5, 0xcf, 0x0c, 0x9b, 0xa0, + 0xff, 0x61, 0x65, 0x94, 0x9c, 0x75, 0x93, 0x5a, 0xa4, 0xc2, 0xb1, 0x80, 0xe5, 0xd4, 0xd9, 0xa2, + 0x16, 0x41, 0xeb, 0x00, 0x26, 0x75, 0x8f, 0x1d, 0x8b, 0xb8, 0x26, 0x61, 0x94, 0x39, 0x3c, 0xe3, + 0xa9, 0xbe, 0xe7, 0x60, 0x25, 0x65, 0x56, 0x7c, 0x62, 0x9c, 0x20, 0x0c, 0x7c, 0x38, 0xf1, 0x62, + 0xb6, 0x72, 0xe3, 0x51, 0x56, 0xc5, 0xf3, 0x42, 0xe5, 0x39, 0x0e, 0x99, 0x3d, 0x7b, 0x13, 0x8f, + 0x60, 0xc6, 0x85, 0xfe, 0x01, 0xd1, 0x09, 0x74, 0xcf, 0x27, 0xc7, 0xce, 0x98, 0x15, 0x51, 0xc4, + 0x45, 0x27, 0xd0, 0x98, 0x5d, 0x33, 0x41, 0x9c, 0xe2, 0x51, 0x09, 0x96, 0xfa, 0x9d, 0x83, 0x4e, + 0xf7, 0xa8, 0x23, 0xfd, 0x81, 0x44, 0x10, 0x0e, 0xb5, 0x66, 0x4b, 0x95, 0x38, 0x54, 0x06, 0x38, + 0xec, 0x63, 0x55, 0x8f, 0xed, 0x1c, 0x42, 0x50, 0x56, 0xbb, 0x6d, 0x7d, 0xc6, 0x97, 0x47, 0x00, + 0x85, 0xbd, 0x67, 0xda, 0x9e, 0xda, 0x91, 0xf8, 0x08, 0xdf, 0xde, 0xef, 0xa8, 0xba, 0x82, 0xd5, + 0xe6, 0x81, 0x24, 0x54, 0xdf, 0x70, 0xb0, 0x1c, 0x95, 0xac, 0xf9, 0xd4, 0x23, 0x7e, 0x38, 0x41, + 0xc7, 0x80, 0xac, 0xa4, 0x66, 0x3d, 0xed, 0x58, 0x3a, 0xa6, 0x7b, 0x3f, 0x29, 0x3a, 0x1d, 0x09, + 0xfe, 0xd3, 0x5a, 0xf0, 0x04, 0xe8, 0x05, 0x94, 0xa7, 0x79, 0x06, 0x91, 0x4c, 0xa6, 0xbf, 0xd4, + 0xb8, 0xf3, 0x2b, 0x8d, 0xc5, 0x2b, 0xd6, 0xac, 0x59, 0x7b, 0xc7, 0x01, 0x1f, 0xbd, 0x3a, 0xa8, + 0x0b, 0x45, 0x2f, 0x51, 0xc6, 0x06, 0x57, 0x6a, 0x6c, 0xff, 0x20, 0xff, 0x6c, 0x53, 0xf0, 0x94, + 0x04, 0xfd, 0x0d, 0xc2, 0x85, 0x63, 0x85, 0x43, 0x56, 0xad, 0x80, 0x63, 0x03, 0xad, 0x42, 0x61, + 0x48, 0x1c, 0x7b, 0x18, 0x56, 0xf2, 0xcc, 0x9d, 0x58, 0xe8, 0x2e, 0x14, 0x06, 0x23, 0x6a, 0x9e, + 0x04, 0x15, 0x9e, 0x35, 0xf0, 0xdf, 0xac, 0xe4, 0x4a, 0x84, 0xc2, 0x09, 0x78, 0xe1, 0xe5, 0x14, + 0x16, 0x5f, 0xce, 0xda, 0xab, 0x3c, 0x08, 0x2c, 0xe2, 0xf7, 0xeb, 0xdb, 0x85, 0xe5, 0x01, 0x3d, + 0x73, 0x2d, 0xc7, 0xb5, 0xf5, 0x01, 0x1d, 0x27, 0x43, 0xb9, 0x91, 0x59, 0x77, 0x82, 0xd5, 0xe8, + 0x68, 0x82, 0x4b, 0x69, 0xa4, 0x42, 0xc7, 0xa8, 0x09, 0xe0, 0x19, 0xbe, 0x61, 0xfb, 0x86, 0x37, + 0x0c, 0x2a, 0x79, 0x26, 0xff, 0xbf, 0xec, 0x35, 0x4f, 0x90, 0x78, 0x26, 0x08, 0x3d, 0x06, 0x60, + 0x0d, 0xd1, 0xd9, 0xde, 0xf1, 0x6c, 0xef, 0x6e, 0x7d, 0xb7, 0x83, 0xf1, 0x93, 0x2d, 0x98, 0x38, + 0x48, 0x8f, 0x57, 0xb6, 0x13, 0x83, 0x38, 0x8d, 0x9b, 0x5f, 0xb4, 0x22, 0xf0, 0x3d, 0xf5, 0x69, + 0x4f, 0xe2, 0xa2, 0x95, 0xeb, 0x35, 0x95, 0x76, 0xb4, 0x62, 0x25, 0x58, 0xd2, 0xf6, 0x5b, 0xbd, + 0x3e, 0x8e, 0x76, 0x4b, 0x04, 0x01, 0xf7, 0xdb, 0x2a, 0x96, 0xf8, 0xc8, 0xaf, 0x34, 0x71, 0xab, + 0xbb, 0xa3, 0x4a, 0x42, 0xed, 0x13, 0x07, 0xe2, 0x54, 0xd5, 0x35, 0x1e, 0x53, 0x03, 0x84, 0x0b, + 0xea, 0x5b, 0xe9, 0x84, 0x32, 0xff, 0x88, 0x8f, 0xa8, 0x6f, 0xe1, 0x18, 0xba, 0xd0, 0x4f, 0xfe, + 0xab, 0x7e, 0x7e, 0xe6, 0x80, 0x8f, 0xf0, 0xd7, 0x58, 0xf6, 0x7d, 0x58, 0x0a, 0x26, 0xa7, 0x03, + 0x3a, 0x4a, 0x85, 0xaf, 0x67, 0x71, 0x1c, 0x32, 0x18, 0x4e, 0xe1, 0x57, 0x8a, 0x7f, 0xcb, 0x41, + 0x21, 0x8e, 0xb9, 0xc6, 0xf2, 0xd3, 0xab, 0x34, 0x7f, 0x79, 0x95, 0x5e, 0x25, 0x4c, 0x79, 0x09, + 0x55, 0x93, 0x9e, 0x66, 0xe4, 0x52, 0xfe, 0x9a, 0x57, 0xa0, 0x45, 0x17, 0xbf, 0xc6, 0x3d, 0x7f, + 0x98, 0xc0, 0x6d, 0x1a, 0xdd, 0x25, 0x32, 0xf5, 0xed, 0xba, 0x4d, 0x5c, 0xf6, 0x59, 0x50, 0x8f, + 0x7f, 0x32, 0x3c, 0x27, 0x58, 0xfc, 0x80, 0x78, 0x10, 0x9f, 0x3e, 0x72, 0xdc, 0xeb, 0x1c, 0xbf, + 0xdb, 0x7a, 0xd2, 0x19, 0x14, 0x58, 0xc8, 0xf6, 0x97, 0x00, 0x00, 0x00, 0xff, 0xff, 0xc4, 0x88, + 0x57, 0x38, 0xd6, 0x08, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/web_detection.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/web_detection.pb.go index 4da22b3c9..ce4eaef98 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/web_detection.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1/web_detection.pb.go @@ -22,20 +22,20 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Relevant information for the image from the Internet. type WebDetection struct { // Deduced entities from similar images on the Internet. - WebEntities []*WebDetection_WebEntity `protobuf:"bytes,1,rep,name=web_entities,json=webEntities" json:"web_entities,omitempty"` + WebEntities []*WebDetection_WebEntity `protobuf:"bytes,1,rep,name=web_entities,json=webEntities,proto3" json:"web_entities,omitempty"` // Fully matching images from the Internet. // Can include resized copies of the query image. - FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,2,rep,name=full_matching_images,json=fullMatchingImages" json:"full_matching_images,omitempty"` + FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,2,rep,name=full_matching_images,json=fullMatchingImages,proto3" json:"full_matching_images,omitempty"` // Partial matching images from the Internet. // Those images are similar enough to share some key-point features. For // example an original image will likely have partial matching for its crops. - PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,3,rep,name=partial_matching_images,json=partialMatchingImages" json:"partial_matching_images,omitempty"` + PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,3,rep,name=partial_matching_images,json=partialMatchingImages,proto3" json:"partial_matching_images,omitempty"` // Web pages containing the matching images from the Internet. - PagesWithMatchingImages []*WebDetection_WebPage `protobuf:"bytes,4,rep,name=pages_with_matching_images,json=pagesWithMatchingImages" json:"pages_with_matching_images,omitempty"` + PagesWithMatchingImages []*WebDetection_WebPage `protobuf:"bytes,4,rep,name=pages_with_matching_images,json=pagesWithMatchingImages,proto3" json:"pages_with_matching_images,omitempty"` // The visually similar image results. - VisuallySimilarImages []*WebDetection_WebImage `protobuf:"bytes,6,rep,name=visually_similar_images,json=visuallySimilarImages" json:"visually_similar_images,omitempty"` + VisuallySimilarImages []*WebDetection_WebImage `protobuf:"bytes,6,rep,name=visually_similar_images,json=visuallySimilarImages,proto3" json:"visually_similar_images,omitempty"` // Best guess text labels for the request image. - BestGuessLabels []*WebDetection_WebLabel `protobuf:"bytes,8,rep,name=best_guess_labels,json=bestGuessLabels" json:"best_guess_labels,omitempty"` + BestGuessLabels []*WebDetection_WebLabel `protobuf:"bytes,8,rep,name=best_guess_labels,json=bestGuessLabels,proto3" json:"best_guess_labels,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -45,7 +45,7 @@ func (m *WebDetection) Reset() { *m = WebDetection{} } func (m *WebDetection) String() string { return proto.CompactTextString(m) } func (*WebDetection) ProtoMessage() {} func (*WebDetection) Descriptor() ([]byte, []int) { - return fileDescriptor_web_detection_4e5f575acd9ec7d5, []int{0} + return fileDescriptor_web_detection_7a96d879901fa424, []int{0} } func (m *WebDetection) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WebDetection.Unmarshal(m, b) @@ -110,12 +110,12 @@ func (m *WebDetection) GetBestGuessLabels() []*WebDetection_WebLabel { // Entity deduced from similar images on the Internet. type WebDetection_WebEntity struct { // Opaque entity ID. - EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Overall relevancy score for the entity. // Not normalized and not comparable across different image queries. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // Canonical description of the entity, in English. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -125,7 +125,7 @@ func (m *WebDetection_WebEntity) Reset() { *m = WebDetection_WebEntity{} func (m *WebDetection_WebEntity) String() string { return proto.CompactTextString(m) } func (*WebDetection_WebEntity) ProtoMessage() {} func (*WebDetection_WebEntity) Descriptor() ([]byte, []int) { - return fileDescriptor_web_detection_4e5f575acd9ec7d5, []int{0, 0} + return fileDescriptor_web_detection_7a96d879901fa424, []int{0, 0} } func (m *WebDetection_WebEntity) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WebDetection_WebEntity.Unmarshal(m, b) @@ -169,9 +169,9 @@ func (m *WebDetection_WebEntity) GetDescription() string { // Metadata for online images. type WebDetection_WebImage struct { // The result image URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // (Deprecated) Overall relevancy score for the image. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -181,7 +181,7 @@ func (m *WebDetection_WebImage) Reset() { *m = WebDetection_WebImage{} } func (m *WebDetection_WebImage) String() string { return proto.CompactTextString(m) } func (*WebDetection_WebImage) ProtoMessage() {} func (*WebDetection_WebImage) Descriptor() ([]byte, []int) { - return fileDescriptor_web_detection_4e5f575acd9ec7d5, []int{0, 1} + return fileDescriptor_web_detection_7a96d879901fa424, []int{0, 1} } func (m *WebDetection_WebImage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WebDetection_WebImage.Unmarshal(m, b) @@ -218,19 +218,19 @@ func (m *WebDetection_WebImage) GetScore() float32 { // Metadata for web pages. type WebDetection_WebPage struct { // The result web page URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // (Deprecated) Overall relevancy score for the web page. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // Title for the web page, may contain HTML markups. - PageTitle string `protobuf:"bytes,3,opt,name=page_title,json=pageTitle" json:"page_title,omitempty"` + PageTitle string `protobuf:"bytes,3,opt,name=page_title,json=pageTitle,proto3" json:"page_title,omitempty"` // Fully matching images on the page. // Can include resized copies of the query image. - FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,4,rep,name=full_matching_images,json=fullMatchingImages" json:"full_matching_images,omitempty"` + FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,4,rep,name=full_matching_images,json=fullMatchingImages,proto3" json:"full_matching_images,omitempty"` // Partial matching images on the page. // Those images are similar enough to share some key-point features. For // example an original image will likely have partial matching for its // crops. - PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,5,rep,name=partial_matching_images,json=partialMatchingImages" json:"partial_matching_images,omitempty"` + PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,5,rep,name=partial_matching_images,json=partialMatchingImages,proto3" json:"partial_matching_images,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -240,7 +240,7 @@ func (m *WebDetection_WebPage) Reset() { *m = WebDetection_WebPage{} } func (m *WebDetection_WebPage) String() string { return proto.CompactTextString(m) } func (*WebDetection_WebPage) ProtoMessage() {} func (*WebDetection_WebPage) Descriptor() ([]byte, []int) { - return fileDescriptor_web_detection_4e5f575acd9ec7d5, []int{0, 2} + return fileDescriptor_web_detection_7a96d879901fa424, []int{0, 2} } func (m *WebDetection_WebPage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WebDetection_WebPage.Unmarshal(m, b) @@ -298,11 +298,11 @@ func (m *WebDetection_WebPage) GetPartialMatchingImages() []*WebDetection_WebIma // Label to provide extra metadata for the web detection. type WebDetection_WebLabel struct { // Label for extra metadata. - Label string `protobuf:"bytes,1,opt,name=label" json:"label,omitempty"` + Label string `protobuf:"bytes,1,opt,name=label,proto3" json:"label,omitempty"` // The BCP-47 language code for `label`, such as "en-US" or "sr-Latn". // For more information, see // http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -312,7 +312,7 @@ func (m *WebDetection_WebLabel) Reset() { *m = WebDetection_WebLabel{} } func (m *WebDetection_WebLabel) String() string { return proto.CompactTextString(m) } func (*WebDetection_WebLabel) ProtoMessage() {} func (*WebDetection_WebLabel) Descriptor() ([]byte, []int) { - return fileDescriptor_web_detection_4e5f575acd9ec7d5, []int{0, 3} + return fileDescriptor_web_detection_7a96d879901fa424, []int{0, 3} } func (m *WebDetection_WebLabel) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WebDetection_WebLabel.Unmarshal(m, b) @@ -355,41 +355,41 @@ func init() { } func init() { - proto.RegisterFile("google/cloud/vision/v1/web_detection.proto", fileDescriptor_web_detection_4e5f575acd9ec7d5) + proto.RegisterFile("google/cloud/vision/v1/web_detection.proto", fileDescriptor_web_detection_7a96d879901fa424) } -var fileDescriptor_web_detection_4e5f575acd9ec7d5 = []byte{ - // 505 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x94, 0xd1, 0x8b, 0xd3, 0x4e, - 0x10, 0xc7, 0x49, 0x7b, 0x77, 0xbf, 0x76, 0xdb, 0x1f, 0x7a, 0xcb, 0xe9, 0x85, 0xa8, 0x50, 0xf4, - 0xa5, 0x88, 0x26, 0xdc, 0xf9, 0xa8, 0x4f, 0xa7, 0x87, 0x1c, 0x28, 0xd4, 0x28, 0x1c, 0xfa, 0xb2, - 0x6e, 0x92, 0x35, 0x1d, 0xd8, 0x66, 0x43, 0x76, 0xd3, 0xd2, 0xff, 0xc4, 0xbf, 0xca, 0xbf, 0xc7, - 0x47, 0x99, 0xdd, 0x8d, 0x94, 0xeb, 0x1d, 0xd4, 0x43, 0x7c, 0x9b, 0x99, 0xcc, 0xf7, 0xf3, 0xdd, - 0x9d, 0x0c, 0x4b, 0x9e, 0x96, 0x4a, 0x95, 0x52, 0x24, 0xb9, 0x54, 0x6d, 0x91, 0x2c, 0x41, 0x83, - 0xaa, 0x92, 0xe5, 0x49, 0xb2, 0x12, 0x19, 0x2b, 0x84, 0x11, 0xb9, 0x01, 0x55, 0xc5, 0x75, 0xa3, - 0x8c, 0xa2, 0xf7, 0x5d, 0x6f, 0x6c, 0x7b, 0x63, 0xd7, 0x1b, 0x2f, 0x4f, 0xa2, 0x87, 0x9e, 0xc1, - 0x6b, 0x48, 0x78, 0x55, 0x29, 0xc3, 0x51, 0xa4, 0x9d, 0xea, 0xf1, 0x8f, 0x01, 0x19, 0x5f, 0x8a, - 0xec, 0x4d, 0x07, 0xa3, 0x1f, 0xc8, 0x18, 0xe9, 0xa2, 0x32, 0x60, 0x40, 0xe8, 0x30, 0x98, 0xf4, - 0xa7, 0xa3, 0xd3, 0x38, 0xbe, 0x9e, 0x1e, 0x6f, 0x6a, 0x31, 0x39, 0x47, 0xdd, 0x3a, 0x1d, 0xad, - 0x7c, 0x08, 0x42, 0x53, 0x46, 0x8e, 0xbe, 0xb5, 0x52, 0xb2, 0x05, 0x37, 0xf9, 0x1c, 0xaa, 0x92, - 0xc1, 0x82, 0x97, 0x42, 0x87, 0x3d, 0x8b, 0x7e, 0xbe, 0x2b, 0xfa, 0x02, 0x55, 0x29, 0x45, 0xd4, - 0x7b, 0x4f, 0xb2, 0x25, 0x4d, 0x05, 0x39, 0xae, 0x79, 0x63, 0x80, 0x6f, 0x7b, 0xf4, 0x6f, 0xe3, - 0x71, 0xcf, 0xd3, 0xae, 0xd8, 0x00, 0x89, 0x6a, 0x0c, 0xd8, 0x0a, 0xcc, 0x7c, 0xcb, 0x69, 0xcf, - 0x3a, 0x3d, 0xdb, 0xd5, 0x69, 0x86, 0x46, 0xc7, 0x96, 0x77, 0x09, 0x66, 0xbe, 0x7d, 0xa3, 0x25, - 0xe8, 0x96, 0x4b, 0xb9, 0x66, 0x1a, 0x16, 0x20, 0x79, 0xd3, 0xf9, 0x1c, 0xdc, 0xea, 0x46, 0x1d, - 0xed, 0xa3, 0x83, 0x79, 0x9b, 0xcf, 0xe4, 0x30, 0x13, 0xda, 0xb0, 0xb2, 0x15, 0x5a, 0x33, 0xc9, - 0x33, 0x21, 0x75, 0x38, 0xf8, 0x33, 0x83, 0x77, 0xa8, 0x4a, 0xef, 0x20, 0xe7, 0x2d, 0x62, 0x6c, - 0xae, 0xa3, 0xaf, 0x64, 0xf8, 0x7b, 0x1d, 0xe8, 0x03, 0x32, 0xb4, 0x0b, 0xb5, 0x66, 0x50, 0x84, - 0xc1, 0x24, 0x98, 0x0e, 0xd3, 0x81, 0x2b, 0x5c, 0x14, 0xf4, 0x88, 0xec, 0xeb, 0x5c, 0x35, 0x22, - 0xec, 0x4d, 0x82, 0x69, 0x2f, 0x75, 0x09, 0x9d, 0x90, 0x51, 0x21, 0x74, 0xde, 0x40, 0x8d, 0x46, - 0x61, 0xdf, 0x8a, 0x36, 0x4b, 0xd1, 0x29, 0x19, 0x74, 0xf7, 0xa3, 0x77, 0x49, 0xbf, 0x6d, 0xa4, - 0x47, 0x63, 0x78, 0x3d, 0x35, 0xfa, 0xde, 0x23, 0xff, 0xf9, 0xe1, 0xef, 0xaa, 0xa1, 0x8f, 0x08, - 0xc1, 0xdf, 0xc4, 0x0c, 0x18, 0x29, 0xfc, 0x41, 0x86, 0x58, 0xf9, 0x84, 0x85, 0x1b, 0xb7, 0x7b, - 0xef, 0x1f, 0x6c, 0xf7, 0xfe, 0xdf, 0xdb, 0xee, 0xe8, 0xdc, 0x8e, 0xd3, 0xfe, 0x3d, 0x1c, 0x84, - 0x5d, 0x06, 0x3f, 0x1c, 0x97, 0xd0, 0x27, 0xe4, 0x7f, 0xc9, 0xab, 0xb2, 0xc5, 0x61, 0xe4, 0xaa, - 0x70, 0x63, 0x1a, 0xa6, 0xe3, 0xae, 0xf8, 0x5a, 0x15, 0xe2, 0xac, 0x21, 0x51, 0xae, 0x16, 0x37, - 0x9c, 0xe8, 0xec, 0x70, 0xf3, 0x48, 0x33, 0x7c, 0x81, 0x66, 0xc1, 0x97, 0x57, 0xbe, 0xb9, 0x54, - 0x48, 0x8a, 0x55, 0x53, 0x26, 0xa5, 0xa8, 0xec, 0xfb, 0x94, 0xb8, 0x4f, 0xbc, 0x06, 0x7d, 0xf5, - 0x11, 0x7c, 0xe9, 0xa2, 0x9f, 0x41, 0x90, 0x1d, 0xd8, 0xde, 0x17, 0xbf, 0x02, 0x00, 0x00, 0xff, - 0xff, 0x1c, 0xe5, 0x3e, 0x5b, 0x2f, 0x05, 0x00, 0x00, +var fileDescriptor_web_detection_7a96d879901fa424 = []byte{ + // 512 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xbc, 0x94, 0x51, 0x6f, 0xd3, 0x30, + 0x10, 0xc7, 0x95, 0xb6, 0x1b, 0xad, 0x5b, 0x04, 0xb3, 0x06, 0x8b, 0x02, 0x48, 0x15, 0xbc, 0x54, + 0x08, 0x12, 0x6d, 0x3c, 0xc2, 0xd3, 0xc6, 0x34, 0x4d, 0x02, 0x54, 0x02, 0x62, 0x82, 0x17, 0xe3, + 0x26, 0x26, 0x3d, 0xc9, 0x8d, 0xa3, 0xd8, 0x69, 0xd5, 0x6f, 0xc2, 0x33, 0x1f, 0x88, 0xcf, 0xc3, + 0x23, 0x3a, 0xdb, 0x45, 0xd5, 0xba, 0x49, 0x65, 0x42, 0xbc, 0xdd, 0x5d, 0xee, 0xff, 0xfb, 0xdb, + 0x97, 0x93, 0xc9, 0xd3, 0x42, 0xa9, 0x42, 0x8a, 0x24, 0x93, 0xaa, 0xc9, 0x93, 0x39, 0x68, 0x50, + 0x65, 0x32, 0x3f, 0x4c, 0x16, 0x62, 0xc2, 0x72, 0x61, 0x44, 0x66, 0x40, 0x95, 0x71, 0x55, 0x2b, + 0xa3, 0xe8, 0x7d, 0xd7, 0x1b, 0xdb, 0xde, 0xd8, 0xf5, 0xc6, 0xf3, 0xc3, 0xe8, 0xa1, 0x67, 0xf0, + 0x0a, 0x12, 0x5e, 0x96, 0xca, 0x70, 0x14, 0x69, 0xa7, 0x7a, 0xfc, 0xb3, 0x4b, 0x06, 0x17, 0x62, + 0xf2, 0x7a, 0x05, 0xa3, 0xef, 0xc9, 0x00, 0xe9, 0xa2, 0x34, 0x60, 0x40, 0xe8, 0x30, 0x18, 0xb6, + 0x47, 0xfd, 0xa3, 0x38, 0xbe, 0x9a, 0x1e, 0xaf, 0x6b, 0x31, 0x39, 0x45, 0xdd, 0x32, 0xed, 0x2f, + 0x7c, 0x08, 0x42, 0x53, 0x46, 0xf6, 0xbf, 0x35, 0x52, 0xb2, 0x19, 0x37, 0xd9, 0x14, 0xca, 0x82, + 0xc1, 0x8c, 0x17, 0x42, 0x87, 0x2d, 0x8b, 0x7e, 0xbe, 0x2d, 0xfa, 0x1c, 0x55, 0x29, 0x45, 0xd4, + 0x5b, 0x4f, 0xb2, 0x25, 0x4d, 0x05, 0x39, 0xa8, 0x78, 0x6d, 0x80, 0x6f, 0x7a, 0xb4, 0x6f, 0xe2, + 0x71, 0xcf, 0xd3, 0x2e, 0xd9, 0x00, 0x89, 0x2a, 0x0c, 0xd8, 0x02, 0xcc, 0x74, 0xc3, 0xa9, 0x63, + 0x9d, 0x9e, 0x6d, 0xeb, 0x34, 0x46, 0xa3, 0x03, 0xcb, 0xbb, 0x00, 0x33, 0xdd, 0xbc, 0xd1, 0x1c, + 0x74, 0xc3, 0xa5, 0x5c, 0x32, 0x0d, 0x33, 0x90, 0xbc, 0x5e, 0xf9, 0xec, 0xde, 0xe8, 0x46, 0x2b, + 0xda, 0x07, 0x07, 0xf3, 0x36, 0x9f, 0xc9, 0xde, 0x44, 0x68, 0xc3, 0x8a, 0x46, 0x68, 0xcd, 0x24, + 0x9f, 0x08, 0xa9, 0xc3, 0xee, 0xdf, 0x19, 0xbc, 0x41, 0x55, 0x7a, 0x07, 0x39, 0x67, 0x88, 0xb1, + 0xb9, 0x8e, 0xbe, 0x92, 0xde, 0x9f, 0x75, 0xa0, 0x0f, 0x48, 0xcf, 0x2e, 0xd4, 0x92, 0x41, 0x1e, + 0x06, 0xc3, 0x60, 0xd4, 0x4b, 0xbb, 0xae, 0x70, 0x9e, 0xd3, 0x7d, 0xb2, 0xa3, 0x33, 0x55, 0x8b, + 0xb0, 0x35, 0x0c, 0x46, 0xad, 0xd4, 0x25, 0x74, 0x48, 0xfa, 0xb9, 0xd0, 0x59, 0x0d, 0x15, 0x1a, + 0x85, 0x6d, 0x2b, 0x5a, 0x2f, 0x45, 0x47, 0xa4, 0xbb, 0xba, 0x1f, 0xbd, 0x4b, 0xda, 0x4d, 0x2d, + 0x3d, 0x1a, 0xc3, 0xab, 0xa9, 0xd1, 0xf7, 0x16, 0xb9, 0xe5, 0x87, 0xbf, 0xad, 0x86, 0x3e, 0x22, + 0x04, 0x7f, 0x13, 0x33, 0x60, 0xa4, 0xf0, 0x07, 0xe9, 0x61, 0xe5, 0x23, 0x16, 0xae, 0xdd, 0xee, + 0xce, 0x7f, 0xd8, 0xee, 0x9d, 0x7f, 0xb7, 0xdd, 0xd1, 0xa9, 0x1d, 0xa7, 0xfd, 0x7b, 0x38, 0x08, + 0xbb, 0x0c, 0x7e, 0x38, 0x2e, 0xa1, 0x4f, 0xc8, 0x6d, 0xc9, 0xcb, 0xa2, 0xc1, 0x61, 0x64, 0x2a, + 0x77, 0x63, 0xea, 0xa5, 0x83, 0x55, 0xf1, 0x44, 0xe5, 0xe2, 0x78, 0x49, 0xa2, 0x4c, 0xcd, 0xae, + 0x39, 0xd1, 0xf1, 0xde, 0xfa, 0x91, 0xc6, 0xf8, 0x02, 0x8d, 0x83, 0x2f, 0xaf, 0x7c, 0x73, 0xa1, + 0x90, 0x14, 0xab, 0xba, 0x48, 0x0a, 0x51, 0xda, 0xf7, 0x29, 0x71, 0x9f, 0x78, 0x05, 0xfa, 0xf2, + 0x23, 0xf8, 0xd2, 0x45, 0xbf, 0x82, 0xe0, 0x47, 0xab, 0x73, 0x76, 0xf2, 0xe9, 0xdd, 0x64, 0xd7, + 0x4a, 0x5e, 0xfc, 0x0e, 0x00, 0x00, 0xff, 0xff, 0x99, 0x1b, 0x06, 0x03, 0x36, 0x05, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/geometry.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/geometry.pb.go index a5ddec040..ec0453117 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/geometry.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/geometry.pb.go @@ -22,9 +22,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // NOTE: the vertex coordinates are in the same scale as the original image. type Vertex struct { // X coordinate. - X int32 `protobuf:"varint,1,opt,name=x" json:"x,omitempty"` + X int32 `protobuf:"varint,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y int32 `protobuf:"varint,2,opt,name=y" json:"y,omitempty"` + Y int32 `protobuf:"varint,2,opt,name=y,proto3" json:"y,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -71,7 +71,7 @@ func (m *Vertex) GetY() int32 { // A bounding polygon for the detected image annotation. type BoundingPoly struct { // The bounding polygon vertices. - Vertices []*Vertex `protobuf:"bytes,1,rep,name=vertices" json:"vertices,omitempty"` + Vertices []*Vertex `protobuf:"bytes,1,rep,name=vertices,proto3" json:"vertices,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -113,11 +113,11 @@ func (m *BoundingPoly) GetVertices() []*Vertex { // The position coordinates are in the same scale as the original image. type Position struct { // X coordinate. - X float32 `protobuf:"fixed32,1,opt,name=x" json:"x,omitempty"` + X float32 `protobuf:"fixed32,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y float32 `protobuf:"fixed32,2,opt,name=y" json:"y,omitempty"` + Y float32 `protobuf:"fixed32,2,opt,name=y,proto3" json:"y,omitempty"` // Z coordinate (or depth). - Z float32 `protobuf:"fixed32,3,opt,name=z" json:"z,omitempty"` + Z float32 `protobuf:"fixed32,3,opt,name=z,proto3" json:"z,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/image_annotator.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/image_annotator.pb.go index f1208baef..f42b092d2 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/image_annotator.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/image_annotator.pb.go @@ -300,13 +300,13 @@ func (FaceAnnotation_Landmark_Type) EnumDescriptor() ([]byte, []int) { // vertical to operate on and the number of top-scoring results to return. type Feature struct { // The feature type. - Type Feature_Type `protobuf:"varint,1,opt,name=type,enum=google.cloud.vision.v1p1beta1.Feature_Type" json:"type,omitempty"` + Type Feature_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.vision.v1p1beta1.Feature_Type" json:"type,omitempty"` // Maximum number of results of this type. - MaxResults int32 `protobuf:"varint,2,opt,name=max_results,json=maxResults" json:"max_results,omitempty"` + MaxResults int32 `protobuf:"varint,2,opt,name=max_results,json=maxResults,proto3" json:"max_results,omitempty"` // Model to use for the feature. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -365,7 +365,7 @@ type ImageSource struct { // [Google Cloud Storage Request // URIs](https://cloud.google.com/storage/docs/reference-uris)). // NOTE: Cloud Storage object versioning is not supported. - GcsImageUri string `protobuf:"bytes,1,opt,name=gcs_image_uri,json=gcsImageUri" json:"gcs_image_uri,omitempty"` + GcsImageUri string `protobuf:"bytes,1,opt,name=gcs_image_uri,json=gcsImageUri,proto3" json:"gcs_image_uri,omitempty"` // Image URI which supports: // 1) Google Cloud Storage image URI, which must be in the following form: // `gs://bucket_name/object_name` (for details, see @@ -376,7 +376,7 @@ type ImageSource struct { // This is preferred over the legacy `gcs_image_uri` above. When both // `gcs_image_uri` and `image_uri` are specified, `image_uri` takes // precedence. - ImageUri string `protobuf:"bytes,2,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,2,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -429,7 +429,7 @@ type Image struct { // Google Cloud Storage image location. If both `content` and `source` // are provided for an image, `content` takes precedence and is // used to perform the image annotation request. - Source *ImageSource `protobuf:"bytes,2,opt,name=source" json:"source,omitempty"` + Source *ImageSource `protobuf:"bytes,2,opt,name=source,proto3" json:"source,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -482,45 +482,45 @@ type FaceAnnotation struct { // Note that one or more x and/or y coordinates may not be generated in the // `BoundingPoly` (the polygon will be unbounded) if only a partial face // appears in the image to be annotated. - BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // The `fd_bounding_poly` bounding polygon is tighter than the // `boundingPoly`, and encloses only the skin part of the face. Typically, it // is used to eliminate the face from any image analysis that detects the // "amount of skin" visible in an image. It is not based on the // landmarker results, only on the initial face detection, hence // the fd (face detection) prefix. - FdBoundingPoly *BoundingPoly `protobuf:"bytes,2,opt,name=fd_bounding_poly,json=fdBoundingPoly" json:"fd_bounding_poly,omitempty"` + FdBoundingPoly *BoundingPoly `protobuf:"bytes,2,opt,name=fd_bounding_poly,json=fdBoundingPoly,proto3" json:"fd_bounding_poly,omitempty"` // Detected face landmarks. - Landmarks []*FaceAnnotation_Landmark `protobuf:"bytes,3,rep,name=landmarks" json:"landmarks,omitempty"` + Landmarks []*FaceAnnotation_Landmark `protobuf:"bytes,3,rep,name=landmarks,proto3" json:"landmarks,omitempty"` // Roll angle, which indicates the amount of clockwise/anti-clockwise rotation // of the face relative to the image vertical about the axis perpendicular to // the face. Range [-180,180]. - RollAngle float32 `protobuf:"fixed32,4,opt,name=roll_angle,json=rollAngle" json:"roll_angle,omitempty"` + RollAngle float32 `protobuf:"fixed32,4,opt,name=roll_angle,json=rollAngle,proto3" json:"roll_angle,omitempty"` // Yaw angle, which indicates the leftward/rightward angle that the face is // pointing relative to the vertical plane perpendicular to the image. Range // [-180,180]. - PanAngle float32 `protobuf:"fixed32,5,opt,name=pan_angle,json=panAngle" json:"pan_angle,omitempty"` + PanAngle float32 `protobuf:"fixed32,5,opt,name=pan_angle,json=panAngle,proto3" json:"pan_angle,omitempty"` // Pitch angle, which indicates the upwards/downwards angle that the face is // pointing relative to the image's horizontal plane. Range [-180,180]. - TiltAngle float32 `protobuf:"fixed32,6,opt,name=tilt_angle,json=tiltAngle" json:"tilt_angle,omitempty"` + TiltAngle float32 `protobuf:"fixed32,6,opt,name=tilt_angle,json=tiltAngle,proto3" json:"tilt_angle,omitempty"` // Detection confidence. Range [0, 1]. - DetectionConfidence float32 `protobuf:"fixed32,7,opt,name=detection_confidence,json=detectionConfidence" json:"detection_confidence,omitempty"` + DetectionConfidence float32 `protobuf:"fixed32,7,opt,name=detection_confidence,json=detectionConfidence,proto3" json:"detection_confidence,omitempty"` // Face landmarking confidence. Range [0, 1]. - LandmarkingConfidence float32 `protobuf:"fixed32,8,opt,name=landmarking_confidence,json=landmarkingConfidence" json:"landmarking_confidence,omitempty"` + LandmarkingConfidence float32 `protobuf:"fixed32,8,opt,name=landmarking_confidence,json=landmarkingConfidence,proto3" json:"landmarking_confidence,omitempty"` // Joy likelihood. - JoyLikelihood Likelihood `protobuf:"varint,9,opt,name=joy_likelihood,json=joyLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"joy_likelihood,omitempty"` + JoyLikelihood Likelihood `protobuf:"varint,9,opt,name=joy_likelihood,json=joyLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"joy_likelihood,omitempty"` // Sorrow likelihood. - SorrowLikelihood Likelihood `protobuf:"varint,10,opt,name=sorrow_likelihood,json=sorrowLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"sorrow_likelihood,omitempty"` + SorrowLikelihood Likelihood `protobuf:"varint,10,opt,name=sorrow_likelihood,json=sorrowLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"sorrow_likelihood,omitempty"` // Anger likelihood. - AngerLikelihood Likelihood `protobuf:"varint,11,opt,name=anger_likelihood,json=angerLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"anger_likelihood,omitempty"` + AngerLikelihood Likelihood `protobuf:"varint,11,opt,name=anger_likelihood,json=angerLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"anger_likelihood,omitempty"` // Surprise likelihood. - SurpriseLikelihood Likelihood `protobuf:"varint,12,opt,name=surprise_likelihood,json=surpriseLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"surprise_likelihood,omitempty"` + SurpriseLikelihood Likelihood `protobuf:"varint,12,opt,name=surprise_likelihood,json=surpriseLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"surprise_likelihood,omitempty"` // Under-exposed likelihood. - UnderExposedLikelihood Likelihood `protobuf:"varint,13,opt,name=under_exposed_likelihood,json=underExposedLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"under_exposed_likelihood,omitempty"` + UnderExposedLikelihood Likelihood `protobuf:"varint,13,opt,name=under_exposed_likelihood,json=underExposedLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"under_exposed_likelihood,omitempty"` // Blurred likelihood. - BlurredLikelihood Likelihood `protobuf:"varint,14,opt,name=blurred_likelihood,json=blurredLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"blurred_likelihood,omitempty"` + BlurredLikelihood Likelihood `protobuf:"varint,14,opt,name=blurred_likelihood,json=blurredLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"blurred_likelihood,omitempty"` // Headwear likelihood. - HeadwearLikelihood Likelihood `protobuf:"varint,15,opt,name=headwear_likelihood,json=headwearLikelihood,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"headwear_likelihood,omitempty"` + HeadwearLikelihood Likelihood `protobuf:"varint,15,opt,name=headwear_likelihood,json=headwearLikelihood,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"headwear_likelihood,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -658,9 +658,9 @@ func (m *FaceAnnotation) GetHeadwearLikelihood() Likelihood { // A face-specific landmark (for example, a face feature). type FaceAnnotation_Landmark struct { // Face landmark type. - Type FaceAnnotation_Landmark_Type `protobuf:"varint,3,opt,name=type,enum=google.cloud.vision.v1p1beta1.FaceAnnotation_Landmark_Type" json:"type,omitempty"` + Type FaceAnnotation_Landmark_Type `protobuf:"varint,3,opt,name=type,proto3,enum=google.cloud.vision.v1p1beta1.FaceAnnotation_Landmark_Type" json:"type,omitempty"` // Face landmark position. - Position *Position `protobuf:"bytes,4,opt,name=position" json:"position,omitempty"` + Position *Position `protobuf:"bytes,4,opt,name=position,proto3" json:"position,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -707,7 +707,7 @@ func (m *FaceAnnotation_Landmark) GetPosition() *Position { // Detected entity location information. type LocationInfo struct { // lat/long location coordinates. - LatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng,json=latLng" json:"lat_lng,omitempty"` + LatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng,json=latLng,proto3" json:"lat_lng,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -747,11 +747,11 @@ func (m *LocationInfo) GetLatLng() *latlng.LatLng { // A `Property` consists of a user-supplied name/value pair. type Property struct { // Name of the property. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Value of the property. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Value of numeric properties. - Uint64Value uint64 `protobuf:"varint,3,opt,name=uint64_value,json=uint64Value" json:"uint64_value,omitempty"` + Uint64Value uint64 `protobuf:"varint,3,opt,name=uint64_value,json=uint64Value,proto3" json:"uint64_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -806,37 +806,37 @@ func (m *Property) GetUint64Value() uint64 { type EntityAnnotation struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/). - Mid string `protobuf:"bytes,1,opt,name=mid" json:"mid,omitempty"` + Mid string `protobuf:"bytes,1,opt,name=mid,proto3" json:"mid,omitempty"` // The language code for the locale in which the entity textual // `description` is expressed. - Locale string `protobuf:"bytes,2,opt,name=locale" json:"locale,omitempty"` + Locale string `protobuf:"bytes,2,opt,name=locale,proto3" json:"locale,omitempty"` // Entity textual description, expressed in its `locale` language. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Overall score of the result. Range [0, 1]. - Score float32 `protobuf:"fixed32,4,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,4,opt,name=score,proto3" json:"score,omitempty"` // The accuracy of the entity detection in an image. // For example, for an image in which the "Eiffel Tower" entity is detected, // this field represents the confidence that there is a tower in the query // image. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` // The relevancy of the ICA (Image Content Annotation) label to the // image. For example, the relevancy of "tower" is likely higher to an image // containing the detected "Eiffel Tower" than to an image containing a // detected distant towering building, even though the confidence that // there is a tower in each image may be the same. Range [0, 1]. - Topicality float32 `protobuf:"fixed32,6,opt,name=topicality" json:"topicality,omitempty"` + Topicality float32 `protobuf:"fixed32,6,opt,name=topicality,proto3" json:"topicality,omitempty"` // Image region to which this entity belongs. Not produced // for `LABEL_DETECTION` features. - BoundingPoly *BoundingPoly `protobuf:"bytes,7,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,7,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // The location information for the detected entity. Multiple // `LocationInfo` elements can be present because one location may // indicate the location of the scene in the image, and another location // may indicate the location of the place where the image was taken. // Location information is usually present for landmarks. - Locations []*LocationInfo `protobuf:"bytes,8,rep,name=locations" json:"locations,omitempty"` + Locations []*LocationInfo `protobuf:"bytes,8,rep,name=locations,proto3" json:"locations,omitempty"` // Some entities may have optional user-supplied `Property` (name/value) // fields, such a score or string that qualifies the entity. - Properties []*Property `protobuf:"bytes,9,rep,name=properties" json:"properties,omitempty"` + Properties []*Property `protobuf:"bytes,9,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -936,20 +936,20 @@ type SafeSearchAnnotation struct { // Represents the adult content likelihood for the image. Adult content may // contain elements such as nudity, pornographic images or cartoons, or // sexual activities. - Adult Likelihood `protobuf:"varint,1,opt,name=adult,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"adult,omitempty"` + Adult Likelihood `protobuf:"varint,1,opt,name=adult,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"adult,omitempty"` // Spoof likelihood. The likelihood that an modification // was made to the image's canonical version to make it appear // funny or offensive. - Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"spoof,omitempty"` + Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"spoof,omitempty"` // Likelihood that this is a medical image. - Medical Likelihood `protobuf:"varint,3,opt,name=medical,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"medical,omitempty"` + Medical Likelihood `protobuf:"varint,3,opt,name=medical,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"medical,omitempty"` // Likelihood that this image contains violent content. - Violence Likelihood `protobuf:"varint,4,opt,name=violence,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"violence,omitempty"` + Violence Likelihood `protobuf:"varint,4,opt,name=violence,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"violence,omitempty"` // Likelihood that the request image contains racy content. Racy content may // include (but is not limited to) skimpy or sheer clothing, strategically // covered nudity, lewd or provocative poses, or close-ups of sensitive // body areas. - Racy Likelihood `protobuf:"varint,9,opt,name=racy,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"racy,omitempty"` + Racy Likelihood `protobuf:"varint,9,opt,name=racy,proto3,enum=google.cloud.vision.v1p1beta1.Likelihood" json:"racy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1017,9 +1017,9 @@ func (m *SafeSearchAnnotation) GetRacy() Likelihood { // Rectangle determined by min and max `LatLng` pairs. type LatLongRect struct { // Min lat/long pair. - MinLatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=min_lat_lng,json=minLatLng" json:"min_lat_lng,omitempty"` + MinLatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=min_lat_lng,json=minLatLng,proto3" json:"min_lat_lng,omitempty"` // Max lat/long pair. - MaxLatLng *latlng.LatLng `protobuf:"bytes,2,opt,name=max_lat_lng,json=maxLatLng" json:"max_lat_lng,omitempty"` + MaxLatLng *latlng.LatLng `protobuf:"bytes,2,opt,name=max_lat_lng,json=maxLatLng,proto3" json:"max_lat_lng,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1067,12 +1067,12 @@ func (m *LatLongRect) GetMaxLatLng() *latlng.LatLng { // the image that the color occupies in the image. type ColorInfo struct { // RGB components of the color. - Color *color.Color `protobuf:"bytes,1,opt,name=color" json:"color,omitempty"` + Color *color.Color `protobuf:"bytes,1,opt,name=color,proto3" json:"color,omitempty"` // Image-specific score for this color. Value in range [0, 1]. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // The fraction of pixels the color occupies in the image. // Value in range [0, 1]. - PixelFraction float32 `protobuf:"fixed32,3,opt,name=pixel_fraction,json=pixelFraction" json:"pixel_fraction,omitempty"` + PixelFraction float32 `protobuf:"fixed32,3,opt,name=pixel_fraction,json=pixelFraction,proto3" json:"pixel_fraction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1126,7 +1126,7 @@ func (m *ColorInfo) GetPixelFraction() float32 { // Set of dominant colors and their corresponding scores. type DominantColorsAnnotation struct { // RGB color values with their score and pixel fraction. - Colors []*ColorInfo `protobuf:"bytes,1,rep,name=colors" json:"colors,omitempty"` + Colors []*ColorInfo `protobuf:"bytes,1,rep,name=colors,proto3" json:"colors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1166,7 +1166,7 @@ func (m *DominantColorsAnnotation) GetColors() []*ColorInfo { // Stores image properties, such as dominant colors. type ImageProperties struct { // If present, dominant colors completed successfully. - DominantColors *DominantColorsAnnotation `protobuf:"bytes,1,opt,name=dominant_colors,json=dominantColors" json:"dominant_colors,omitempty"` + DominantColors *DominantColorsAnnotation `protobuf:"bytes,1,opt,name=dominant_colors,json=dominantColors,proto3" json:"dominant_colors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1207,12 +1207,12 @@ func (m *ImageProperties) GetDominantColors() *DominantColorsAnnotation { type CropHint struct { // The bounding polygon for the crop region. The coordinates of the bounding // box are in the original image's scale, as returned in `ImageParams`. - BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // Confidence of this being a salient region. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // Fraction of importance of this salient region with respect to the original // image. - ImportanceFraction float32 `protobuf:"fixed32,3,opt,name=importance_fraction,json=importanceFraction" json:"importance_fraction,omitempty"` + ImportanceFraction float32 `protobuf:"fixed32,3,opt,name=importance_fraction,json=importanceFraction,proto3" json:"importance_fraction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1266,7 +1266,7 @@ func (m *CropHint) GetImportanceFraction() float32 { // Set of crop hints that are used to generate new crops when serving images. type CropHintsAnnotation struct { // Crop hint results. - CropHints []*CropHint `protobuf:"bytes,1,rep,name=crop_hints,json=cropHints" json:"crop_hints,omitempty"` + CropHints []*CropHint `protobuf:"bytes,1,rep,name=crop_hints,json=cropHints,proto3" json:"crop_hints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1311,7 +1311,7 @@ type CropHintsParams struct { // best possible crop is returned. The number of provided aspect ratios is // limited to a maximum of 16; any aspect ratios provided after the 16th are // ignored. - AspectRatios []float32 `protobuf:"fixed32,1,rep,packed,name=aspect_ratios,json=aspectRatios" json:"aspect_ratios,omitempty"` + AspectRatios []float32 `protobuf:"fixed32,1,rep,packed,name=aspect_ratios,json=aspectRatios,proto3" json:"aspect_ratios,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1351,7 +1351,7 @@ func (m *CropHintsParams) GetAspectRatios() []float32 { // Parameters for web detection request. type WebDetectionParams struct { // Whether to include results derived from the geo information in the image. - IncludeGeoResults bool `protobuf:"varint,2,opt,name=include_geo_results,json=includeGeoResults" json:"include_geo_results,omitempty"` + IncludeGeoResults bool `protobuf:"varint,2,opt,name=include_geo_results,json=includeGeoResults,proto3" json:"include_geo_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1391,7 +1391,7 @@ func (m *WebDetectionParams) GetIncludeGeoResults() bool { // Image context and/or feature-specific parameters. type ImageContext struct { // lat/long rectangle that specifies the location of the image. - LatLongRect *LatLongRect `protobuf:"bytes,1,opt,name=lat_long_rect,json=latLongRect" json:"lat_long_rect,omitempty"` + LatLongRect *LatLongRect `protobuf:"bytes,1,opt,name=lat_long_rect,json=latLongRect,proto3" json:"lat_long_rect,omitempty"` // List of languages to use for TEXT_DETECTION. In most cases, an empty value // yields the best results since it enables automatic language detection. For // languages based on the Latin alphabet, setting `language_hints` is not @@ -1400,11 +1400,11 @@ type ImageContext struct { // significant hindrance if the hint is wrong). Text detection returns an // error if one or more of the specified languages is not one of the // [supported languages](/vision/docs/languages). - LanguageHints []string `protobuf:"bytes,2,rep,name=language_hints,json=languageHints" json:"language_hints,omitempty"` + LanguageHints []string `protobuf:"bytes,2,rep,name=language_hints,json=languageHints,proto3" json:"language_hints,omitempty"` // Parameters for crop hints annotation request. - CropHintsParams *CropHintsParams `protobuf:"bytes,4,opt,name=crop_hints_params,json=cropHintsParams" json:"crop_hints_params,omitempty"` + CropHintsParams *CropHintsParams `protobuf:"bytes,4,opt,name=crop_hints_params,json=cropHintsParams,proto3" json:"crop_hints_params,omitempty"` // Parameters for web detection. - WebDetectionParams *WebDetectionParams `protobuf:"bytes,6,opt,name=web_detection_params,json=webDetectionParams" json:"web_detection_params,omitempty"` + WebDetectionParams *WebDetectionParams `protobuf:"bytes,6,opt,name=web_detection_params,json=webDetectionParams,proto3" json:"web_detection_params,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1466,11 +1466,11 @@ func (m *ImageContext) GetWebDetectionParams() *WebDetectionParams { // image, with user-requested features. type AnnotateImageRequest struct { // The image to be processed. - Image *Image `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"` + Image *Image `protobuf:"bytes,1,opt,name=image,proto3" json:"image,omitempty"` // Requested features. - Features []*Feature `protobuf:"bytes,2,rep,name=features" json:"features,omitempty"` + Features []*Feature `protobuf:"bytes,2,rep,name=features,proto3" json:"features,omitempty"` // Additional context that may accompany the image. - ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext" json:"image_context,omitempty"` + ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext,proto3" json:"image_context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1524,32 +1524,32 @@ func (m *AnnotateImageRequest) GetImageContext() *ImageContext { // Response to an image annotation request. type AnnotateImageResponse struct { // If present, face detection has completed successfully. - FaceAnnotations []*FaceAnnotation `protobuf:"bytes,1,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"` + FaceAnnotations []*FaceAnnotation `protobuf:"bytes,1,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // If present, landmark detection has completed successfully. - LandmarkAnnotations []*EntityAnnotation `protobuf:"bytes,2,rep,name=landmark_annotations,json=landmarkAnnotations" json:"landmark_annotations,omitempty"` + LandmarkAnnotations []*EntityAnnotation `protobuf:"bytes,2,rep,name=landmark_annotations,json=landmarkAnnotations,proto3" json:"landmark_annotations,omitempty"` // If present, logo detection has completed successfully. - LogoAnnotations []*EntityAnnotation `protobuf:"bytes,3,rep,name=logo_annotations,json=logoAnnotations" json:"logo_annotations,omitempty"` + LogoAnnotations []*EntityAnnotation `protobuf:"bytes,3,rep,name=logo_annotations,json=logoAnnotations,proto3" json:"logo_annotations,omitempty"` // If present, label detection has completed successfully. - LabelAnnotations []*EntityAnnotation `protobuf:"bytes,4,rep,name=label_annotations,json=labelAnnotations" json:"label_annotations,omitempty"` + LabelAnnotations []*EntityAnnotation `protobuf:"bytes,4,rep,name=label_annotations,json=labelAnnotations,proto3" json:"label_annotations,omitempty"` // If present, text (OCR) detection has completed successfully. - TextAnnotations []*EntityAnnotation `protobuf:"bytes,5,rep,name=text_annotations,json=textAnnotations" json:"text_annotations,omitempty"` + TextAnnotations []*EntityAnnotation `protobuf:"bytes,5,rep,name=text_annotations,json=textAnnotations,proto3" json:"text_annotations,omitempty"` // If present, text (OCR) detection or document (OCR) text detection has // completed successfully. // This annotation provides the structural hierarchy for the OCR detected // text. - FullTextAnnotation *TextAnnotation `protobuf:"bytes,12,opt,name=full_text_annotation,json=fullTextAnnotation" json:"full_text_annotation,omitempty"` + FullTextAnnotation *TextAnnotation `protobuf:"bytes,12,opt,name=full_text_annotation,json=fullTextAnnotation,proto3" json:"full_text_annotation,omitempty"` // If present, safe-search annotation has completed successfully. - SafeSearchAnnotation *SafeSearchAnnotation `protobuf:"bytes,6,opt,name=safe_search_annotation,json=safeSearchAnnotation" json:"safe_search_annotation,omitempty"` + SafeSearchAnnotation *SafeSearchAnnotation `protobuf:"bytes,6,opt,name=safe_search_annotation,json=safeSearchAnnotation,proto3" json:"safe_search_annotation,omitempty"` // If present, image properties were extracted successfully. - ImagePropertiesAnnotation *ImageProperties `protobuf:"bytes,8,opt,name=image_properties_annotation,json=imagePropertiesAnnotation" json:"image_properties_annotation,omitempty"` + ImagePropertiesAnnotation *ImageProperties `protobuf:"bytes,8,opt,name=image_properties_annotation,json=imagePropertiesAnnotation,proto3" json:"image_properties_annotation,omitempty"` // If present, crop hints have completed successfully. - CropHintsAnnotation *CropHintsAnnotation `protobuf:"bytes,11,opt,name=crop_hints_annotation,json=cropHintsAnnotation" json:"crop_hints_annotation,omitempty"` + CropHintsAnnotation *CropHintsAnnotation `protobuf:"bytes,11,opt,name=crop_hints_annotation,json=cropHintsAnnotation,proto3" json:"crop_hints_annotation,omitempty"` // If present, web detection has completed successfully. - WebDetection *WebDetection `protobuf:"bytes,13,opt,name=web_detection,json=webDetection" json:"web_detection,omitempty"` + WebDetection *WebDetection `protobuf:"bytes,13,opt,name=web_detection,json=webDetection,proto3" json:"web_detection,omitempty"` // If set, represents the error message for the operation. // Note that filled-in image annotations are guaranteed to be // correct, even when `error` is set. - Error *status.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1659,7 +1659,7 @@ func (m *AnnotateImageResponse) GetError() *status.Status { // Multiple image annotation requests are batched into a single service call. type BatchAnnotateImagesRequest struct { // Individual image annotation requests for this batch. - Requests []*AnnotateImageRequest `protobuf:"bytes,1,rep,name=requests" json:"requests,omitempty"` + Requests []*AnnotateImageRequest `protobuf:"bytes,1,rep,name=requests,proto3" json:"requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1699,7 +1699,7 @@ func (m *BatchAnnotateImagesRequest) GetRequests() []*AnnotateImageRequest { // Response to a batch image annotation request. type BatchAnnotateImagesResponse struct { // Individual responses to image annotation requests within the batch. - Responses []*AnnotateImageResponse `protobuf:"bytes,1,rep,name=responses" json:"responses,omitempty"` + Responses []*AnnotateImageResponse `protobuf:"bytes,1,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1797,8 +1797,7 @@ func (c *imageAnnotatorClient) BatchAnnotateImages(ctx context.Context, in *Batc return out, nil } -// Server API for ImageAnnotator service - +// ImageAnnotatorServer is the server API for ImageAnnotator service. type ImageAnnotatorServer interface { // Run image detection and annotation for a batch of images. BatchAnnotateImages(context.Context, *BatchAnnotateImagesRequest) (*BatchAnnotateImagesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/text_annotation.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/text_annotation.pb.go index a3261184c..9cdbb9756 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/text_annotation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/text_annotation.pb.go @@ -114,9 +114,9 @@ func (Block_BlockType) EnumDescriptor() ([]byte, []int) { // message definition below for more detail. type TextAnnotation struct { // List of pages detected by OCR. - Pages []*Page `protobuf:"bytes,1,rep,name=pages" json:"pages,omitempty"` + Pages []*Page `protobuf:"bytes,1,rep,name=pages,proto3" json:"pages,omitempty"` // UTF-8 text detected on the pages. - Text string `protobuf:"bytes,2,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,2,opt,name=text,proto3" json:"text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -165,9 +165,9 @@ type TextAnnotation_DetectedLanguage struct { // The BCP-47 language code, such as "en-US" or "sr-Latn". For more // information, see // http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Confidence of detected language. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -214,9 +214,9 @@ func (m *TextAnnotation_DetectedLanguage) GetConfidence() float32 { // Detected start or end of a structural component. type TextAnnotation_DetectedBreak struct { // Detected break type. - Type TextAnnotation_DetectedBreak_BreakType `protobuf:"varint,1,opt,name=type,enum=google.cloud.vision.v1p1beta1.TextAnnotation_DetectedBreak_BreakType" json:"type,omitempty"` + Type TextAnnotation_DetectedBreak_BreakType `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.vision.v1p1beta1.TextAnnotation_DetectedBreak_BreakType" json:"type,omitempty"` // True if break prepends the element. - IsPrefix bool `protobuf:"varint,2,opt,name=is_prefix,json=isPrefix" json:"is_prefix,omitempty"` + IsPrefix bool `protobuf:"varint,2,opt,name=is_prefix,json=isPrefix,proto3" json:"is_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -263,9 +263,9 @@ func (m *TextAnnotation_DetectedBreak) GetIsPrefix() bool { // Additional information detected on the structural component. type TextAnnotation_TextProperty struct { // A list of detected languages together with confidence. - DetectedLanguages []*TextAnnotation_DetectedLanguage `protobuf:"bytes,1,rep,name=detected_languages,json=detectedLanguages" json:"detected_languages,omitempty"` + DetectedLanguages []*TextAnnotation_DetectedLanguage `protobuf:"bytes,1,rep,name=detected_languages,json=detectedLanguages,proto3" json:"detected_languages,omitempty"` // Detected start or end of a text segment. - DetectedBreak *TextAnnotation_DetectedBreak `protobuf:"bytes,2,opt,name=detected_break,json=detectedBreak" json:"detected_break,omitempty"` + DetectedBreak *TextAnnotation_DetectedBreak `protobuf:"bytes,2,opt,name=detected_break,json=detectedBreak,proto3" json:"detected_break,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -312,15 +312,15 @@ func (m *TextAnnotation_TextProperty) GetDetectedBreak() *TextAnnotation_Detecte // Detected page from OCR. type Page struct { // Additional information detected on the page. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // Page width in pixels. - Width int32 `protobuf:"varint,2,opt,name=width" json:"width,omitempty"` + Width int32 `protobuf:"varint,2,opt,name=width,proto3" json:"width,omitempty"` // Page height in pixels. - Height int32 `protobuf:"varint,3,opt,name=height" json:"height,omitempty"` + Height int32 `protobuf:"varint,3,opt,name=height,proto3" json:"height,omitempty"` // List of blocks of text, images etc on this page. - Blocks []*Block `protobuf:"bytes,4,rep,name=blocks" json:"blocks,omitempty"` + Blocks []*Block `protobuf:"bytes,4,rep,name=blocks,proto3" json:"blocks,omitempty"` // Confidence of the OCR results on the page. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -388,7 +388,7 @@ func (m *Page) GetConfidence() float32 { // Logical element on the page. type Block struct { // Additional information detected for the block. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the block. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -404,13 +404,13 @@ type Block struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of paragraphs in this block (if this blocks is of type text). - Paragraphs []*Paragraph `protobuf:"bytes,3,rep,name=paragraphs" json:"paragraphs,omitempty"` + Paragraphs []*Paragraph `protobuf:"bytes,3,rep,name=paragraphs,proto3" json:"paragraphs,omitempty"` // Detected block type (text, image etc) for this block. - BlockType Block_BlockType `protobuf:"varint,4,opt,name=block_type,json=blockType,enum=google.cloud.vision.v1p1beta1.Block_BlockType" json:"block_type,omitempty"` + BlockType Block_BlockType `protobuf:"varint,4,opt,name=block_type,json=blockType,proto3,enum=google.cloud.vision.v1p1beta1.Block_BlockType" json:"block_type,omitempty"` // Confidence of the OCR results on the block. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -478,7 +478,7 @@ func (m *Block) GetConfidence() float32 { // Structural unit of text representing a number of words in certain order. type Paragraph struct { // Additional information detected for the paragraph. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the paragraph. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -494,11 +494,11 @@ type Paragraph struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of words in this paragraph. - Words []*Word `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` + Words []*Word `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"` // Confidence of the OCR results for the paragraph. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -559,7 +559,7 @@ func (m *Paragraph) GetConfidence() float32 { // A word representation. type Word struct { // Additional information detected for the word. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the word. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -575,12 +575,12 @@ type Word struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of symbols in the word. // The order of the symbols follows the natural reading order. - Symbols []*Symbol `protobuf:"bytes,3,rep,name=symbols" json:"symbols,omitempty"` + Symbols []*Symbol `protobuf:"bytes,3,rep,name=symbols,proto3" json:"symbols,omitempty"` // Confidence of the OCR results for the word. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -641,7 +641,7 @@ func (m *Word) GetConfidence() float32 { // A single symbol representation. type Symbol struct { // Additional information detected for the symbol. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the symbol. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -657,11 +657,11 @@ type Symbol struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // The actual UTF-8 representation of the symbol. - Text string `protobuf:"bytes,3,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,3,opt,name=text,proto3" json:"text,omitempty"` // Confidence of the OCR results for the symbol. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/web_detection.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/web_detection.pb.go index d17c77b45..faa8a7861 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/web_detection.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p1beta1/web_detection.pb.go @@ -22,20 +22,20 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Relevant information for the image from the Internet. type WebDetection struct { // Deduced entities from similar images on the Internet. - WebEntities []*WebDetection_WebEntity `protobuf:"bytes,1,rep,name=web_entities,json=webEntities" json:"web_entities,omitempty"` + WebEntities []*WebDetection_WebEntity `protobuf:"bytes,1,rep,name=web_entities,json=webEntities,proto3" json:"web_entities,omitempty"` // Fully matching images from the Internet. // Can include resized copies of the query image. - FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,2,rep,name=full_matching_images,json=fullMatchingImages" json:"full_matching_images,omitempty"` + FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,2,rep,name=full_matching_images,json=fullMatchingImages,proto3" json:"full_matching_images,omitempty"` // Partial matching images from the Internet. // Those images are similar enough to share some key-point features. For // example an original image will likely have partial matching for its crops. - PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,3,rep,name=partial_matching_images,json=partialMatchingImages" json:"partial_matching_images,omitempty"` + PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,3,rep,name=partial_matching_images,json=partialMatchingImages,proto3" json:"partial_matching_images,omitempty"` // Web pages containing the matching images from the Internet. - PagesWithMatchingImages []*WebDetection_WebPage `protobuf:"bytes,4,rep,name=pages_with_matching_images,json=pagesWithMatchingImages" json:"pages_with_matching_images,omitempty"` + PagesWithMatchingImages []*WebDetection_WebPage `protobuf:"bytes,4,rep,name=pages_with_matching_images,json=pagesWithMatchingImages,proto3" json:"pages_with_matching_images,omitempty"` // The visually similar image results. - VisuallySimilarImages []*WebDetection_WebImage `protobuf:"bytes,6,rep,name=visually_similar_images,json=visuallySimilarImages" json:"visually_similar_images,omitempty"` + VisuallySimilarImages []*WebDetection_WebImage `protobuf:"bytes,6,rep,name=visually_similar_images,json=visuallySimilarImages,proto3" json:"visually_similar_images,omitempty"` // Best guess text labels for the request image. - BestGuessLabels []*WebDetection_WebLabel `protobuf:"bytes,8,rep,name=best_guess_labels,json=bestGuessLabels" json:"best_guess_labels,omitempty"` + BestGuessLabels []*WebDetection_WebLabel `protobuf:"bytes,8,rep,name=best_guess_labels,json=bestGuessLabels,proto3" json:"best_guess_labels,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -110,12 +110,12 @@ func (m *WebDetection) GetBestGuessLabels() []*WebDetection_WebLabel { // Entity deduced from similar images on the Internet. type WebDetection_WebEntity struct { // Opaque entity ID. - EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Overall relevancy score for the entity. // Not normalized and not comparable across different image queries. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // Canonical description of the entity, in English. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,9 +169,9 @@ func (m *WebDetection_WebEntity) GetDescription() string { // Metadata for online images. type WebDetection_WebImage struct { // The result image URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // (Deprecated) Overall relevancy score for the image. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -218,19 +218,19 @@ func (m *WebDetection_WebImage) GetScore() float32 { // Metadata for web pages. type WebDetection_WebPage struct { // The result web page URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // (Deprecated) Overall relevancy score for the web page. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // Title for the web page, may contain HTML markups. - PageTitle string `protobuf:"bytes,3,opt,name=page_title,json=pageTitle" json:"page_title,omitempty"` + PageTitle string `protobuf:"bytes,3,opt,name=page_title,json=pageTitle,proto3" json:"page_title,omitempty"` // Fully matching images on the page. // Can include resized copies of the query image. - FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,4,rep,name=full_matching_images,json=fullMatchingImages" json:"full_matching_images,omitempty"` + FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,4,rep,name=full_matching_images,json=fullMatchingImages,proto3" json:"full_matching_images,omitempty"` // Partial matching images on the page. // Those images are similar enough to share some key-point features. For // example an original image will likely have partial matching for its // crops. - PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,5,rep,name=partial_matching_images,json=partialMatchingImages" json:"partial_matching_images,omitempty"` + PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,5,rep,name=partial_matching_images,json=partialMatchingImages,proto3" json:"partial_matching_images,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -298,11 +298,11 @@ func (m *WebDetection_WebPage) GetPartialMatchingImages() []*WebDetection_WebIma // Label to provide extra metadata for the web detection. type WebDetection_WebLabel struct { // Label for extra metadata. - Label string `protobuf:"bytes,1,opt,name=label" json:"label,omitempty"` + Label string `protobuf:"bytes,1,opt,name=label,proto3" json:"label,omitempty"` // The BCP-47 language code for `label`, such as "en-US" or "sr-Latn". // For more information, see // http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/geometry.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/geometry.pb.go index 75d44ad37..6f31b29d9 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/geometry.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/geometry.pb.go @@ -22,9 +22,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // NOTE: the vertex coordinates are in the same scale as the original image. type Vertex struct { // X coordinate. - X int32 `protobuf:"varint,1,opt,name=x" json:"x,omitempty"` + X int32 `protobuf:"varint,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y int32 `protobuf:"varint,2,opt,name=y" json:"y,omitempty"` + Y int32 `protobuf:"varint,2,opt,name=y,proto3" json:"y,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -73,9 +73,9 @@ func (m *Vertex) GetY() int32 { // and range from 0 to 1. type NormalizedVertex struct { // X coordinate. - X float32 `protobuf:"fixed32,1,opt,name=x" json:"x,omitempty"` + X float32 `protobuf:"fixed32,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y float32 `protobuf:"fixed32,2,opt,name=y" json:"y,omitempty"` + Y float32 `protobuf:"fixed32,2,opt,name=y,proto3" json:"y,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -122,9 +122,9 @@ func (m *NormalizedVertex) GetY() float32 { // A bounding polygon for the detected image annotation. type BoundingPoly struct { // The bounding polygon vertices. - Vertices []*Vertex `protobuf:"bytes,1,rep,name=vertices" json:"vertices,omitempty"` + Vertices []*Vertex `protobuf:"bytes,1,rep,name=vertices,proto3" json:"vertices,omitempty"` // The bounding polygon normalized vertices. - NormalizedVertices []*NormalizedVertex `protobuf:"bytes,2,rep,name=normalized_vertices,json=normalizedVertices" json:"normalized_vertices,omitempty"` + NormalizedVertices []*NormalizedVertex `protobuf:"bytes,2,rep,name=normalized_vertices,json=normalizedVertices,proto3" json:"normalized_vertices,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -173,11 +173,11 @@ func (m *BoundingPoly) GetNormalizedVertices() []*NormalizedVertex { // The position coordinates are in the same scale as the original image. type Position struct { // X coordinate. - X float32 `protobuf:"fixed32,1,opt,name=x" json:"x,omitempty"` + X float32 `protobuf:"fixed32,1,opt,name=x,proto3" json:"x,omitempty"` // Y coordinate. - Y float32 `protobuf:"fixed32,2,opt,name=y" json:"y,omitempty"` + Y float32 `protobuf:"fixed32,2,opt,name=y,proto3" json:"y,omitempty"` // Z coordinate (or depth). - Z float32 `protobuf:"fixed32,3,opt,name=z" json:"z,omitempty"` + Z float32 `protobuf:"fixed32,3,opt,name=z,proto3" json:"z,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/image_annotator.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/image_annotator.pb.go index 2d65e0c5e..7071eddc7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/image_annotator.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/image_annotator.pb.go @@ -343,14 +343,14 @@ func (OperationMetadata_State) EnumDescriptor() ([]byte, []int) { // be specified in the `features` list. type Feature struct { // The feature type. - Type Feature_Type `protobuf:"varint,1,opt,name=type,enum=google.cloud.vision.v1p2beta1.Feature_Type" json:"type,omitempty"` + Type Feature_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.vision.v1p2beta1.Feature_Type" json:"type,omitempty"` // Maximum number of results of this type. Does not apply to // `TEXT_DETECTION`, `DOCUMENT_TEXT_DETECTION`, or `CROP_HINTS`. - MaxResults int32 `protobuf:"varint,2,opt,name=max_results,json=maxResults" json:"max_results,omitempty"` + MaxResults int32 `protobuf:"varint,2,opt,name=max_results,json=maxResults,proto3" json:"max_results,omitempty"` // Model to use for the feature. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". - Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"` + Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,7 +409,7 @@ type ImageSource struct { // `gs://bucket_name/object_name`. Object versioning is not supported. See // [Google Cloud Storage Request // URIs](https://cloud.google.com/storage/docs/reference-uris) for more info. - GcsImageUri string `protobuf:"bytes,1,opt,name=gcs_image_uri,json=gcsImageUri" json:"gcs_image_uri,omitempty"` + GcsImageUri string `protobuf:"bytes,1,opt,name=gcs_image_uri,json=gcsImageUri,proto3" json:"gcs_image_uri,omitempty"` // The URI of the source image. Can be either: // // 1. A Google Cloud Storage URI of the form @@ -427,7 +427,7 @@ type ImageSource struct { // // When both `gcs_image_uri` and `image_uri` are specified, `image_uri` takes // precedence. - ImageUri string `protobuf:"bytes,2,opt,name=image_uri,json=imageUri" json:"image_uri,omitempty"` + ImageUri string `protobuf:"bytes,2,opt,name=image_uri,json=imageUri,proto3" json:"image_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -480,7 +480,7 @@ type Image struct { // Google Cloud Storage image location, or publicly-accessible image // URL. If both `content` and `source` are provided for an image, `content` // takes precedence and is used to perform the image annotation request. - Source *ImageSource `protobuf:"bytes,2,opt,name=source" json:"source,omitempty"` + Source *ImageSource `protobuf:"bytes,2,opt,name=source,proto3" json:"source,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -533,45 +533,45 @@ type FaceAnnotation struct { // Note that one or more x and/or y coordinates may not be generated in the // `BoundingPoly` (the polygon will be unbounded) if only a partial face // appears in the image to be annotated. - BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // The `fd_bounding_poly` bounding polygon is tighter than the // `boundingPoly`, and encloses only the skin part of the face. Typically, it // is used to eliminate the face from any image analysis that detects the // "amount of skin" visible in an image. It is not based on the // landmarker results, only on the initial face detection, hence // the fd (face detection) prefix. - FdBoundingPoly *BoundingPoly `protobuf:"bytes,2,opt,name=fd_bounding_poly,json=fdBoundingPoly" json:"fd_bounding_poly,omitempty"` + FdBoundingPoly *BoundingPoly `protobuf:"bytes,2,opt,name=fd_bounding_poly,json=fdBoundingPoly,proto3" json:"fd_bounding_poly,omitempty"` // Detected face landmarks. - Landmarks []*FaceAnnotation_Landmark `protobuf:"bytes,3,rep,name=landmarks" json:"landmarks,omitempty"` + Landmarks []*FaceAnnotation_Landmark `protobuf:"bytes,3,rep,name=landmarks,proto3" json:"landmarks,omitempty"` // Roll angle, which indicates the amount of clockwise/anti-clockwise rotation // of the face relative to the image vertical about the axis perpendicular to // the face. Range [-180,180]. - RollAngle float32 `protobuf:"fixed32,4,opt,name=roll_angle,json=rollAngle" json:"roll_angle,omitempty"` + RollAngle float32 `protobuf:"fixed32,4,opt,name=roll_angle,json=rollAngle,proto3" json:"roll_angle,omitempty"` // Yaw angle, which indicates the leftward/rightward angle that the face is // pointing relative to the vertical plane perpendicular to the image. Range // [-180,180]. - PanAngle float32 `protobuf:"fixed32,5,opt,name=pan_angle,json=panAngle" json:"pan_angle,omitempty"` + PanAngle float32 `protobuf:"fixed32,5,opt,name=pan_angle,json=panAngle,proto3" json:"pan_angle,omitempty"` // Pitch angle, which indicates the upwards/downwards angle that the face is // pointing relative to the image's horizontal plane. Range [-180,180]. - TiltAngle float32 `protobuf:"fixed32,6,opt,name=tilt_angle,json=tiltAngle" json:"tilt_angle,omitempty"` + TiltAngle float32 `protobuf:"fixed32,6,opt,name=tilt_angle,json=tiltAngle,proto3" json:"tilt_angle,omitempty"` // Detection confidence. Range [0, 1]. - DetectionConfidence float32 `protobuf:"fixed32,7,opt,name=detection_confidence,json=detectionConfidence" json:"detection_confidence,omitempty"` + DetectionConfidence float32 `protobuf:"fixed32,7,opt,name=detection_confidence,json=detectionConfidence,proto3" json:"detection_confidence,omitempty"` // Face landmarking confidence. Range [0, 1]. - LandmarkingConfidence float32 `protobuf:"fixed32,8,opt,name=landmarking_confidence,json=landmarkingConfidence" json:"landmarking_confidence,omitempty"` + LandmarkingConfidence float32 `protobuf:"fixed32,8,opt,name=landmarking_confidence,json=landmarkingConfidence,proto3" json:"landmarking_confidence,omitempty"` // Joy likelihood. - JoyLikelihood Likelihood `protobuf:"varint,9,opt,name=joy_likelihood,json=joyLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"joy_likelihood,omitempty"` + JoyLikelihood Likelihood `protobuf:"varint,9,opt,name=joy_likelihood,json=joyLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"joy_likelihood,omitempty"` // Sorrow likelihood. - SorrowLikelihood Likelihood `protobuf:"varint,10,opt,name=sorrow_likelihood,json=sorrowLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"sorrow_likelihood,omitempty"` + SorrowLikelihood Likelihood `protobuf:"varint,10,opt,name=sorrow_likelihood,json=sorrowLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"sorrow_likelihood,omitempty"` // Anger likelihood. - AngerLikelihood Likelihood `protobuf:"varint,11,opt,name=anger_likelihood,json=angerLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"anger_likelihood,omitempty"` + AngerLikelihood Likelihood `protobuf:"varint,11,opt,name=anger_likelihood,json=angerLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"anger_likelihood,omitempty"` // Surprise likelihood. - SurpriseLikelihood Likelihood `protobuf:"varint,12,opt,name=surprise_likelihood,json=surpriseLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"surprise_likelihood,omitempty"` + SurpriseLikelihood Likelihood `protobuf:"varint,12,opt,name=surprise_likelihood,json=surpriseLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"surprise_likelihood,omitempty"` // Under-exposed likelihood. - UnderExposedLikelihood Likelihood `protobuf:"varint,13,opt,name=under_exposed_likelihood,json=underExposedLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"under_exposed_likelihood,omitempty"` + UnderExposedLikelihood Likelihood `protobuf:"varint,13,opt,name=under_exposed_likelihood,json=underExposedLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"under_exposed_likelihood,omitempty"` // Blurred likelihood. - BlurredLikelihood Likelihood `protobuf:"varint,14,opt,name=blurred_likelihood,json=blurredLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"blurred_likelihood,omitempty"` + BlurredLikelihood Likelihood `protobuf:"varint,14,opt,name=blurred_likelihood,json=blurredLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"blurred_likelihood,omitempty"` // Headwear likelihood. - HeadwearLikelihood Likelihood `protobuf:"varint,15,opt,name=headwear_likelihood,json=headwearLikelihood,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"headwear_likelihood,omitempty"` + HeadwearLikelihood Likelihood `protobuf:"varint,15,opt,name=headwear_likelihood,json=headwearLikelihood,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"headwear_likelihood,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -709,9 +709,9 @@ func (m *FaceAnnotation) GetHeadwearLikelihood() Likelihood { // A face-specific landmark (for example, a face feature). type FaceAnnotation_Landmark struct { // Face landmark type. - Type FaceAnnotation_Landmark_Type `protobuf:"varint,3,opt,name=type,enum=google.cloud.vision.v1p2beta1.FaceAnnotation_Landmark_Type" json:"type,omitempty"` + Type FaceAnnotation_Landmark_Type `protobuf:"varint,3,opt,name=type,proto3,enum=google.cloud.vision.v1p2beta1.FaceAnnotation_Landmark_Type" json:"type,omitempty"` // Face landmark position. - Position *Position `protobuf:"bytes,4,opt,name=position" json:"position,omitempty"` + Position *Position `protobuf:"bytes,4,opt,name=position,proto3" json:"position,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -758,7 +758,7 @@ func (m *FaceAnnotation_Landmark) GetPosition() *Position { // Detected entity location information. type LocationInfo struct { // lat/long location coordinates. - LatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng,json=latLng" json:"lat_lng,omitempty"` + LatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng,json=latLng,proto3" json:"lat_lng,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -798,11 +798,11 @@ func (m *LocationInfo) GetLatLng() *latlng.LatLng { // A `Property` consists of a user-supplied name/value pair. type Property struct { // Name of the property. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Value of the property. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Value of numeric properties. - Uint64Value uint64 `protobuf:"varint,3,opt,name=uint64_value,json=uint64Value" json:"uint64_value,omitempty"` + Uint64Value uint64 `protobuf:"varint,3,opt,name=uint64_value,json=uint64Value,proto3" json:"uint64_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -858,38 +858,38 @@ type EntityAnnotation struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search // API](https://developers.google.com/knowledge-graph/). - Mid string `protobuf:"bytes,1,opt,name=mid" json:"mid,omitempty"` + Mid string `protobuf:"bytes,1,opt,name=mid,proto3" json:"mid,omitempty"` // The language code for the locale in which the entity textual // `description` is expressed. - Locale string `protobuf:"bytes,2,opt,name=locale" json:"locale,omitempty"` + Locale string `protobuf:"bytes,2,opt,name=locale,proto3" json:"locale,omitempty"` // Entity textual description, expressed in its `locale` language. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Overall score of the result. Range [0, 1]. - Score float32 `protobuf:"fixed32,4,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,4,opt,name=score,proto3" json:"score,omitempty"` // **Deprecated. Use `score` instead.** // The accuracy of the entity detection in an image. // For example, for an image in which the "Eiffel Tower" entity is detected, // this field represents the confidence that there is a tower in the query // image. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` // The relevancy of the ICA (Image Content Annotation) label to the // image. For example, the relevancy of "tower" is likely higher to an image // containing the detected "Eiffel Tower" than to an image containing a // detected distant towering building, even though the confidence that // there is a tower in each image may be the same. Range [0, 1]. - Topicality float32 `protobuf:"fixed32,6,opt,name=topicality" json:"topicality,omitempty"` + Topicality float32 `protobuf:"fixed32,6,opt,name=topicality,proto3" json:"topicality,omitempty"` // Image region to which this entity belongs. Not produced // for `LABEL_DETECTION` features. - BoundingPoly *BoundingPoly `protobuf:"bytes,7,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,7,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // The location information for the detected entity. Multiple // `LocationInfo` elements can be present because one location may // indicate the location of the scene in the image, and another location // may indicate the location of the place where the image was taken. // Location information is usually present for landmarks. - Locations []*LocationInfo `protobuf:"bytes,8,rep,name=locations" json:"locations,omitempty"` + Locations []*LocationInfo `protobuf:"bytes,8,rep,name=locations,proto3" json:"locations,omitempty"` // Some entities may have optional user-supplied `Property` (name/value) // fields, such a score or string that qualifies the entity. - Properties []*Property `protobuf:"bytes,9,rep,name=properties" json:"properties,omitempty"` + Properties []*Property `protobuf:"bytes,9,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -989,20 +989,20 @@ type SafeSearchAnnotation struct { // Represents the adult content likelihood for the image. Adult content may // contain elements such as nudity, pornographic images or cartoons, or // sexual activities. - Adult Likelihood `protobuf:"varint,1,opt,name=adult,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"adult,omitempty"` + Adult Likelihood `protobuf:"varint,1,opt,name=adult,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"adult,omitempty"` // Spoof likelihood. The likelihood that an modification // was made to the image's canonical version to make it appear // funny or offensive. - Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"spoof,omitempty"` + Spoof Likelihood `protobuf:"varint,2,opt,name=spoof,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"spoof,omitempty"` // Likelihood that this is a medical image. - Medical Likelihood `protobuf:"varint,3,opt,name=medical,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"medical,omitempty"` + Medical Likelihood `protobuf:"varint,3,opt,name=medical,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"medical,omitempty"` // Likelihood that this image contains violent content. - Violence Likelihood `protobuf:"varint,4,opt,name=violence,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"violence,omitempty"` + Violence Likelihood `protobuf:"varint,4,opt,name=violence,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"violence,omitempty"` // Likelihood that the request image contains racy content. Racy content may // include (but is not limited to) skimpy or sheer clothing, strategically // covered nudity, lewd or provocative poses, or close-ups of sensitive // body areas. - Racy Likelihood `protobuf:"varint,9,opt,name=racy,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"racy,omitempty"` + Racy Likelihood `protobuf:"varint,9,opt,name=racy,proto3,enum=google.cloud.vision.v1p2beta1.Likelihood" json:"racy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1070,9 +1070,9 @@ func (m *SafeSearchAnnotation) GetRacy() Likelihood { // Rectangle determined by min and max `LatLng` pairs. type LatLongRect struct { // Min lat/long pair. - MinLatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=min_lat_lng,json=minLatLng" json:"min_lat_lng,omitempty"` + MinLatLng *latlng.LatLng `protobuf:"bytes,1,opt,name=min_lat_lng,json=minLatLng,proto3" json:"min_lat_lng,omitempty"` // Max lat/long pair. - MaxLatLng *latlng.LatLng `protobuf:"bytes,2,opt,name=max_lat_lng,json=maxLatLng" json:"max_lat_lng,omitempty"` + MaxLatLng *latlng.LatLng `protobuf:"bytes,2,opt,name=max_lat_lng,json=maxLatLng,proto3" json:"max_lat_lng,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1120,12 +1120,12 @@ func (m *LatLongRect) GetMaxLatLng() *latlng.LatLng { // the image that the color occupies in the image. type ColorInfo struct { // RGB components of the color. - Color *color.Color `protobuf:"bytes,1,opt,name=color" json:"color,omitempty"` + Color *color.Color `protobuf:"bytes,1,opt,name=color,proto3" json:"color,omitempty"` // Image-specific score for this color. Value in range [0, 1]. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // The fraction of pixels the color occupies in the image. // Value in range [0, 1]. - PixelFraction float32 `protobuf:"fixed32,3,opt,name=pixel_fraction,json=pixelFraction" json:"pixel_fraction,omitempty"` + PixelFraction float32 `protobuf:"fixed32,3,opt,name=pixel_fraction,json=pixelFraction,proto3" json:"pixel_fraction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1179,7 +1179,7 @@ func (m *ColorInfo) GetPixelFraction() float32 { // Set of dominant colors and their corresponding scores. type DominantColorsAnnotation struct { // RGB color values with their score and pixel fraction. - Colors []*ColorInfo `protobuf:"bytes,1,rep,name=colors" json:"colors,omitempty"` + Colors []*ColorInfo `protobuf:"bytes,1,rep,name=colors,proto3" json:"colors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1219,7 +1219,7 @@ func (m *DominantColorsAnnotation) GetColors() []*ColorInfo { // Stores image properties, such as dominant colors. type ImageProperties struct { // If present, dominant colors completed successfully. - DominantColors *DominantColorsAnnotation `protobuf:"bytes,1,opt,name=dominant_colors,json=dominantColors" json:"dominant_colors,omitempty"` + DominantColors *DominantColorsAnnotation `protobuf:"bytes,1,opt,name=dominant_colors,json=dominantColors,proto3" json:"dominant_colors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1260,12 +1260,12 @@ func (m *ImageProperties) GetDominantColors() *DominantColorsAnnotation { type CropHint struct { // The bounding polygon for the crop region. The coordinates of the bounding // box are in the original image's scale, as returned in `ImageParams`. - BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly" json:"bounding_poly,omitempty"` + BoundingPoly *BoundingPoly `protobuf:"bytes,1,opt,name=bounding_poly,json=boundingPoly,proto3" json:"bounding_poly,omitempty"` // Confidence of this being a salient region. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // Fraction of importance of this salient region with respect to the original // image. - ImportanceFraction float32 `protobuf:"fixed32,3,opt,name=importance_fraction,json=importanceFraction" json:"importance_fraction,omitempty"` + ImportanceFraction float32 `protobuf:"fixed32,3,opt,name=importance_fraction,json=importanceFraction,proto3" json:"importance_fraction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1319,7 +1319,7 @@ func (m *CropHint) GetImportanceFraction() float32 { // Set of crop hints that are used to generate new crops when serving images. type CropHintsAnnotation struct { // Crop hint results. - CropHints []*CropHint `protobuf:"bytes,1,rep,name=crop_hints,json=cropHints" json:"crop_hints,omitempty"` + CropHints []*CropHint `protobuf:"bytes,1,rep,name=crop_hints,json=cropHints,proto3" json:"crop_hints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1364,7 +1364,7 @@ type CropHintsParams struct { // best possible crop is returned. The number of provided aspect ratios is // limited to a maximum of 16; any aspect ratios provided after the 16th are // ignored. - AspectRatios []float32 `protobuf:"fixed32,1,rep,packed,name=aspect_ratios,json=aspectRatios" json:"aspect_ratios,omitempty"` + AspectRatios []float32 `protobuf:"fixed32,1,rep,packed,name=aspect_ratios,json=aspectRatios,proto3" json:"aspect_ratios,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1404,7 +1404,7 @@ func (m *CropHintsParams) GetAspectRatios() []float32 { // Parameters for web detection request. type WebDetectionParams struct { // Whether to include results derived from the geo information in the image. - IncludeGeoResults bool `protobuf:"varint,2,opt,name=include_geo_results,json=includeGeoResults" json:"include_geo_results,omitempty"` + IncludeGeoResults bool `protobuf:"varint,2,opt,name=include_geo_results,json=includeGeoResults,proto3" json:"include_geo_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1444,7 +1444,7 @@ func (m *WebDetectionParams) GetIncludeGeoResults() bool { // Image context and/or feature-specific parameters. type ImageContext struct { // Not used. - LatLongRect *LatLongRect `protobuf:"bytes,1,opt,name=lat_long_rect,json=latLongRect" json:"lat_long_rect,omitempty"` + LatLongRect *LatLongRect `protobuf:"bytes,1,opt,name=lat_long_rect,json=latLongRect,proto3" json:"lat_long_rect,omitempty"` // List of languages to use for TEXT_DETECTION. In most cases, an empty value // yields the best results since it enables automatic language detection. For // languages based on the Latin alphabet, setting `language_hints` is not @@ -1453,11 +1453,11 @@ type ImageContext struct { // significant hindrance if the hint is wrong). Text detection returns an // error if one or more of the specified languages is not one of the // [supported languages](/vision/docs/languages). - LanguageHints []string `protobuf:"bytes,2,rep,name=language_hints,json=languageHints" json:"language_hints,omitempty"` + LanguageHints []string `protobuf:"bytes,2,rep,name=language_hints,json=languageHints,proto3" json:"language_hints,omitempty"` // Parameters for crop hints annotation request. - CropHintsParams *CropHintsParams `protobuf:"bytes,4,opt,name=crop_hints_params,json=cropHintsParams" json:"crop_hints_params,omitempty"` + CropHintsParams *CropHintsParams `protobuf:"bytes,4,opt,name=crop_hints_params,json=cropHintsParams,proto3" json:"crop_hints_params,omitempty"` // Parameters for web detection. - WebDetectionParams *WebDetectionParams `protobuf:"bytes,6,opt,name=web_detection_params,json=webDetectionParams" json:"web_detection_params,omitempty"` + WebDetectionParams *WebDetectionParams `protobuf:"bytes,6,opt,name=web_detection_params,json=webDetectionParams,proto3" json:"web_detection_params,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1519,11 +1519,11 @@ func (m *ImageContext) GetWebDetectionParams() *WebDetectionParams { // image, with user-requested features. type AnnotateImageRequest struct { // The image to be processed. - Image *Image `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"` + Image *Image `protobuf:"bytes,1,opt,name=image,proto3" json:"image,omitempty"` // Requested features. - Features []*Feature `protobuf:"bytes,2,rep,name=features" json:"features,omitempty"` + Features []*Feature `protobuf:"bytes,2,rep,name=features,proto3" json:"features,omitempty"` // Additional context that may accompany the image. - ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext" json:"image_context,omitempty"` + ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext,proto3" json:"image_context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1578,10 +1578,10 @@ func (m *AnnotateImageRequest) GetImageContext() *ImageContext { // information about the source of that image. type ImageAnnotationContext struct { // The URI of the file used to produce the image. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` // If the file was a PDF or TIFF, this field gives the page number within // the file used to produce the image. - PageNumber int32 `protobuf:"varint,2,opt,name=page_number,json=pageNumber" json:"page_number,omitempty"` + PageNumber int32 `protobuf:"varint,2,opt,name=page_number,json=pageNumber,proto3" json:"page_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1628,35 +1628,35 @@ func (m *ImageAnnotationContext) GetPageNumber() int32 { // Response to an image annotation request. type AnnotateImageResponse struct { // If present, face detection has completed successfully. - FaceAnnotations []*FaceAnnotation `protobuf:"bytes,1,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"` + FaceAnnotations []*FaceAnnotation `protobuf:"bytes,1,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // If present, landmark detection has completed successfully. - LandmarkAnnotations []*EntityAnnotation `protobuf:"bytes,2,rep,name=landmark_annotations,json=landmarkAnnotations" json:"landmark_annotations,omitempty"` + LandmarkAnnotations []*EntityAnnotation `protobuf:"bytes,2,rep,name=landmark_annotations,json=landmarkAnnotations,proto3" json:"landmark_annotations,omitempty"` // If present, logo detection has completed successfully. - LogoAnnotations []*EntityAnnotation `protobuf:"bytes,3,rep,name=logo_annotations,json=logoAnnotations" json:"logo_annotations,omitempty"` + LogoAnnotations []*EntityAnnotation `protobuf:"bytes,3,rep,name=logo_annotations,json=logoAnnotations,proto3" json:"logo_annotations,omitempty"` // If present, label detection has completed successfully. - LabelAnnotations []*EntityAnnotation `protobuf:"bytes,4,rep,name=label_annotations,json=labelAnnotations" json:"label_annotations,omitempty"` + LabelAnnotations []*EntityAnnotation `protobuf:"bytes,4,rep,name=label_annotations,json=labelAnnotations,proto3" json:"label_annotations,omitempty"` // If present, text (OCR) detection has completed successfully. - TextAnnotations []*EntityAnnotation `protobuf:"bytes,5,rep,name=text_annotations,json=textAnnotations" json:"text_annotations,omitempty"` + TextAnnotations []*EntityAnnotation `protobuf:"bytes,5,rep,name=text_annotations,json=textAnnotations,proto3" json:"text_annotations,omitempty"` // If present, text (OCR) detection or document (OCR) text detection has // completed successfully. // This annotation provides the structural hierarchy for the OCR detected // text. - FullTextAnnotation *TextAnnotation `protobuf:"bytes,12,opt,name=full_text_annotation,json=fullTextAnnotation" json:"full_text_annotation,omitempty"` + FullTextAnnotation *TextAnnotation `protobuf:"bytes,12,opt,name=full_text_annotation,json=fullTextAnnotation,proto3" json:"full_text_annotation,omitempty"` // If present, safe-search annotation has completed successfully. - SafeSearchAnnotation *SafeSearchAnnotation `protobuf:"bytes,6,opt,name=safe_search_annotation,json=safeSearchAnnotation" json:"safe_search_annotation,omitempty"` + SafeSearchAnnotation *SafeSearchAnnotation `protobuf:"bytes,6,opt,name=safe_search_annotation,json=safeSearchAnnotation,proto3" json:"safe_search_annotation,omitempty"` // If present, image properties were extracted successfully. - ImagePropertiesAnnotation *ImageProperties `protobuf:"bytes,8,opt,name=image_properties_annotation,json=imagePropertiesAnnotation" json:"image_properties_annotation,omitempty"` + ImagePropertiesAnnotation *ImageProperties `protobuf:"bytes,8,opt,name=image_properties_annotation,json=imagePropertiesAnnotation,proto3" json:"image_properties_annotation,omitempty"` // If present, crop hints have completed successfully. - CropHintsAnnotation *CropHintsAnnotation `protobuf:"bytes,11,opt,name=crop_hints_annotation,json=cropHintsAnnotation" json:"crop_hints_annotation,omitempty"` + CropHintsAnnotation *CropHintsAnnotation `protobuf:"bytes,11,opt,name=crop_hints_annotation,json=cropHintsAnnotation,proto3" json:"crop_hints_annotation,omitempty"` // If present, web detection has completed successfully. - WebDetection *WebDetection `protobuf:"bytes,13,opt,name=web_detection,json=webDetection" json:"web_detection,omitempty"` + WebDetection *WebDetection `protobuf:"bytes,13,opt,name=web_detection,json=webDetection,proto3" json:"web_detection,omitempty"` // If set, represents the error message for the operation. // Note that filled-in image annotations are guaranteed to be // correct, even when `error` is set. - Error *status.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"` + Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` // If present, contextual information is needed to understand where this image // comes from. - Context *ImageAnnotationContext `protobuf:"bytes,21,opt,name=context" json:"context,omitempty"` + Context *ImageAnnotationContext `protobuf:"bytes,21,opt,name=context,proto3" json:"context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1774,9 +1774,9 @@ func (m *AnnotateImageResponse) GetContext() *ImageAnnotationContext { // images, which individually have their own responses. type AnnotateFileResponse struct { // Information about the file for which this response is generated. - InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig" json:"input_config,omitempty"` + InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig,proto3" json:"input_config,omitempty"` // Individual responses to images found within the file. - Responses []*AnnotateImageResponse `protobuf:"bytes,2,rep,name=responses" json:"responses,omitempty"` + Responses []*AnnotateImageResponse `protobuf:"bytes,2,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1823,7 +1823,7 @@ func (m *AnnotateFileResponse) GetResponses() []*AnnotateImageResponse { // Multiple image annotation requests are batched into a single service call. type BatchAnnotateImagesRequest struct { // Individual image annotation requests for this batch. - Requests []*AnnotateImageRequest `protobuf:"bytes,1,rep,name=requests" json:"requests,omitempty"` + Requests []*AnnotateImageRequest `protobuf:"bytes,1,rep,name=requests,proto3" json:"requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1863,7 +1863,7 @@ func (m *BatchAnnotateImagesRequest) GetRequests() []*AnnotateImageRequest { // Response to a batch image annotation request. type BatchAnnotateImagesResponse struct { // Individual responses to image annotation requests within the batch. - Responses []*AnnotateImageResponse `protobuf:"bytes,1,rep,name=responses" json:"responses,omitempty"` + Responses []*AnnotateImageResponse `protobuf:"bytes,1,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1903,13 +1903,13 @@ func (m *BatchAnnotateImagesResponse) GetResponses() []*AnnotateImageResponse { // An offline file annotation request. type AsyncAnnotateFileRequest struct { // Required. Information about the input file. - InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig" json:"input_config,omitempty"` + InputConfig *InputConfig `protobuf:"bytes,1,opt,name=input_config,json=inputConfig,proto3" json:"input_config,omitempty"` // Required. Requested features. - Features []*Feature `protobuf:"bytes,2,rep,name=features" json:"features,omitempty"` + Features []*Feature `protobuf:"bytes,2,rep,name=features,proto3" json:"features,omitempty"` // Additional context that may accompany the image(s) in the file. - ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext" json:"image_context,omitempty"` + ImageContext *ImageContext `protobuf:"bytes,3,opt,name=image_context,json=imageContext,proto3" json:"image_context,omitempty"` // Required. The desired output location and metadata (e.g. format). - OutputConfig *OutputConfig `protobuf:"bytes,4,opt,name=output_config,json=outputConfig" json:"output_config,omitempty"` + OutputConfig *OutputConfig `protobuf:"bytes,4,opt,name=output_config,json=outputConfig,proto3" json:"output_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1970,7 +1970,7 @@ func (m *AsyncAnnotateFileRequest) GetOutputConfig() *OutputConfig { // The response for a single offline file annotation request. type AsyncAnnotateFileResponse struct { // The output location and metadata from AsyncAnnotateFileRequest. - OutputConfig *OutputConfig `protobuf:"bytes,1,opt,name=output_config,json=outputConfig" json:"output_config,omitempty"` + OutputConfig *OutputConfig `protobuf:"bytes,1,opt,name=output_config,json=outputConfig,proto3" json:"output_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2011,7 +2011,7 @@ func (m *AsyncAnnotateFileResponse) GetOutputConfig() *OutputConfig { // call. type AsyncBatchAnnotateFilesRequest struct { // Individual async file annotation requests for this batch. - Requests []*AsyncAnnotateFileRequest `protobuf:"bytes,1,rep,name=requests" json:"requests,omitempty"` + Requests []*AsyncAnnotateFileRequest `protobuf:"bytes,1,rep,name=requests,proto3" json:"requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2052,7 +2052,7 @@ func (m *AsyncBatchAnnotateFilesRequest) GetRequests() []*AsyncAnnotateFileReque type AsyncBatchAnnotateFilesResponse struct { // The list of file annotation responses, one for each request in // AsyncBatchAnnotateFilesRequest. - Responses []*AsyncAnnotateFileResponse `protobuf:"bytes,1,rep,name=responses" json:"responses,omitempty"` + Responses []*AsyncAnnotateFileResponse `protobuf:"bytes,1,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2092,10 +2092,10 @@ func (m *AsyncBatchAnnotateFilesResponse) GetResponses() []*AsyncAnnotateFileRes // The desired input location and metadata. type InputConfig struct { // The Google Cloud Storage location to read the input from. - GcsSource *GcsSource `protobuf:"bytes,1,opt,name=gcs_source,json=gcsSource" json:"gcs_source,omitempty"` + GcsSource *GcsSource `protobuf:"bytes,1,opt,name=gcs_source,json=gcsSource,proto3" json:"gcs_source,omitempty"` // The type of the file. Currently only "application/pdf" and "image/tiff" // are supported. Wildcards are not supported. - MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType" json:"mime_type,omitempty"` + MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2142,7 +2142,7 @@ func (m *InputConfig) GetMimeType() string { // The desired output location and metadata. type OutputConfig struct { // The Google Cloud Storage location to write the output(s) to. - GcsDestination *GcsDestination `protobuf:"bytes,1,opt,name=gcs_destination,json=gcsDestination" json:"gcs_destination,omitempty"` + GcsDestination *GcsDestination `protobuf:"bytes,1,opt,name=gcs_destination,json=gcsDestination,proto3" json:"gcs_destination,omitempty"` // The max number of response protos to put into each output JSON file on GCS. // The valid range is [1, 100]. If not specified, the default value is 20. // @@ -2153,7 +2153,7 @@ type OutputConfig struct { // // Currently, batch_size only applies to GcsDestination, with potential future // support for other output configurations. - BatchSize int32 `protobuf:"varint,2,opt,name=batch_size,json=batchSize" json:"batch_size,omitempty"` + BatchSize int32 `protobuf:"varint,2,opt,name=batch_size,json=batchSize,proto3" json:"batch_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2201,7 +2201,7 @@ func (m *OutputConfig) GetBatchSize() int32 { type GcsSource struct { // Google Cloud Storage URI for the input file. This must only be a GCS // object. Wildcards are not currently supported. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2255,7 +2255,7 @@ type GcsDestination struct { // which contains some subset of the full list of AnnotateImageResponse. // Multiple outputs can happen if, for example, the output JSON is too large // and overflows into multiple sharded files. - Uri string `protobuf:"bytes,1,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2295,11 +2295,11 @@ func (m *GcsDestination) GetUri() string { // Contains metadata for the BatchAnnotateImages operation. type OperationMetadata struct { // Current state of the batch operation. - State OperationMetadata_State `protobuf:"varint,1,opt,name=state,enum=google.cloud.vision.v1p2beta1.OperationMetadata_State" json:"state,omitempty"` + State OperationMetadata_State `protobuf:"varint,1,opt,name=state,proto3,enum=google.cloud.vision.v1p2beta1.OperationMetadata_State" json:"state,omitempty"` // The time when the batch request was received. - CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The time when the operation result was last updated. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2439,8 +2439,7 @@ func (c *imageAnnotatorClient) AsyncBatchAnnotateFiles(ctx context.Context, in * return out, nil } -// Server API for ImageAnnotator service - +// ImageAnnotatorServer is the server API for ImageAnnotator service. type ImageAnnotatorServer interface { // Run image detection and annotation for a batch of images. BatchAnnotateImages(context.Context, *BatchAnnotateImagesRequest) (*BatchAnnotateImagesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/text_annotation.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/text_annotation.pb.go index 2e3e9d872..0475576e5 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/text_annotation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/text_annotation.pb.go @@ -113,9 +113,9 @@ func (Block_BlockType) EnumDescriptor() ([]byte, []int) { // detail. type TextAnnotation struct { // List of pages detected by OCR. - Pages []*Page `protobuf:"bytes,1,rep,name=pages" json:"pages,omitempty"` + Pages []*Page `protobuf:"bytes,1,rep,name=pages,proto3" json:"pages,omitempty"` // UTF-8 text detected on the pages. - Text string `protobuf:"bytes,2,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,2,opt,name=text,proto3" json:"text,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -164,9 +164,9 @@ type TextAnnotation_DetectedLanguage struct { // The BCP-47 language code, such as "en-US" or "sr-Latn". For more // information, see // http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Confidence of detected language. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -213,9 +213,9 @@ func (m *TextAnnotation_DetectedLanguage) GetConfidence() float32 { // Detected start or end of a structural component. type TextAnnotation_DetectedBreak struct { // Detected break type. - Type TextAnnotation_DetectedBreak_BreakType `protobuf:"varint,1,opt,name=type,enum=google.cloud.vision.v1p2beta1.TextAnnotation_DetectedBreak_BreakType" json:"type,omitempty"` + Type TextAnnotation_DetectedBreak_BreakType `protobuf:"varint,1,opt,name=type,proto3,enum=google.cloud.vision.v1p2beta1.TextAnnotation_DetectedBreak_BreakType" json:"type,omitempty"` // True if break prepends the element. - IsPrefix bool `protobuf:"varint,2,opt,name=is_prefix,json=isPrefix" json:"is_prefix,omitempty"` + IsPrefix bool `protobuf:"varint,2,opt,name=is_prefix,json=isPrefix,proto3" json:"is_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,9 +262,9 @@ func (m *TextAnnotation_DetectedBreak) GetIsPrefix() bool { // Additional information detected on the structural component. type TextAnnotation_TextProperty struct { // A list of detected languages together with confidence. - DetectedLanguages []*TextAnnotation_DetectedLanguage `protobuf:"bytes,1,rep,name=detected_languages,json=detectedLanguages" json:"detected_languages,omitempty"` + DetectedLanguages []*TextAnnotation_DetectedLanguage `protobuf:"bytes,1,rep,name=detected_languages,json=detectedLanguages,proto3" json:"detected_languages,omitempty"` // Detected start or end of a text segment. - DetectedBreak *TextAnnotation_DetectedBreak `protobuf:"bytes,2,opt,name=detected_break,json=detectedBreak" json:"detected_break,omitempty"` + DetectedBreak *TextAnnotation_DetectedBreak `protobuf:"bytes,2,opt,name=detected_break,json=detectedBreak,proto3" json:"detected_break,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -311,17 +311,17 @@ func (m *TextAnnotation_TextProperty) GetDetectedBreak() *TextAnnotation_Detecte // Detected page from OCR. type Page struct { // Additional information detected on the page. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // Page width. For PDFs the unit is points. For images (including // TIFFs) the unit is pixels. - Width int32 `protobuf:"varint,2,opt,name=width" json:"width,omitempty"` + Width int32 `protobuf:"varint,2,opt,name=width,proto3" json:"width,omitempty"` // Page height. For PDFs the unit is points. For images (including // TIFFs) the unit is pixels. - Height int32 `protobuf:"varint,3,opt,name=height" json:"height,omitempty"` + Height int32 `protobuf:"varint,3,opt,name=height,proto3" json:"height,omitempty"` // List of blocks of text, images etc on this page. - Blocks []*Block `protobuf:"bytes,4,rep,name=blocks" json:"blocks,omitempty"` + Blocks []*Block `protobuf:"bytes,4,rep,name=blocks,proto3" json:"blocks,omitempty"` // Confidence of the OCR results on the page. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -389,7 +389,7 @@ func (m *Page) GetConfidence() float32 { // Logical element on the page. type Block struct { // Additional information detected for the block. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the block. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -410,13 +410,13 @@ type Block struct { // 1----0 // // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of paragraphs in this block (if this blocks is of type text). - Paragraphs []*Paragraph `protobuf:"bytes,3,rep,name=paragraphs" json:"paragraphs,omitempty"` + Paragraphs []*Paragraph `protobuf:"bytes,3,rep,name=paragraphs,proto3" json:"paragraphs,omitempty"` // Detected block type (text, image etc) for this block. - BlockType Block_BlockType `protobuf:"varint,4,opt,name=block_type,json=blockType,enum=google.cloud.vision.v1p2beta1.Block_BlockType" json:"block_type,omitempty"` + BlockType Block_BlockType `protobuf:"varint,4,opt,name=block_type,json=blockType,proto3,enum=google.cloud.vision.v1p2beta1.Block_BlockType" json:"block_type,omitempty"` // Confidence of the OCR results on the block. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,5,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,5,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -484,7 +484,7 @@ func (m *Block) GetConfidence() float32 { // Structural unit of text representing a number of words in certain order. type Paragraph struct { // Additional information detected for the paragraph. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the paragraph. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -500,11 +500,11 @@ type Paragraph struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of words in this paragraph. - Words []*Word `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` + Words []*Word `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"` // Confidence of the OCR results for the paragraph. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -565,7 +565,7 @@ func (m *Paragraph) GetConfidence() float32 { // A word representation. type Word struct { // Additional information detected for the word. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the word. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -581,12 +581,12 @@ type Word struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // List of symbols in the word. // The order of the symbols follows the natural reading order. - Symbols []*Symbol `protobuf:"bytes,3,rep,name=symbols" json:"symbols,omitempty"` + Symbols []*Symbol `protobuf:"bytes,3,rep,name=symbols,proto3" json:"symbols,omitempty"` // Confidence of the OCR results for the word. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -647,7 +647,7 @@ func (m *Word) GetConfidence() float32 { // A single symbol representation. type Symbol struct { // Additional information detected for the symbol. - Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *TextAnnotation_TextProperty `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The bounding box for the symbol. // The vertices are in the order of top-left, top-right, bottom-right, // bottom-left. When a rotation of the bounding box is detected the rotation @@ -663,11 +663,11 @@ type Symbol struct { // | | // 1----0 // and the vertice order will still be (0, 1, 2, 3). - BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox" json:"bounding_box,omitempty"` + BoundingBox *BoundingPoly `protobuf:"bytes,2,opt,name=bounding_box,json=boundingBox,proto3" json:"bounding_box,omitempty"` // The actual UTF-8 representation of the symbol. - Text string `protobuf:"bytes,3,opt,name=text" json:"text,omitempty"` + Text string `protobuf:"bytes,3,opt,name=text,proto3" json:"text,omitempty"` // Confidence of the OCR results for the symbol. Range [0, 1]. - Confidence float32 `protobuf:"fixed32,4,opt,name=confidence" json:"confidence,omitempty"` + Confidence float32 `protobuf:"fixed32,4,opt,name=confidence,proto3" json:"confidence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/web_detection.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/web_detection.pb.go index bc00971cd..d5bf295f7 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/web_detection.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/vision/v1p2beta1/web_detection.pb.go @@ -22,20 +22,20 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Relevant information for the image from the Internet. type WebDetection struct { // Deduced entities from similar images on the Internet. - WebEntities []*WebDetection_WebEntity `protobuf:"bytes,1,rep,name=web_entities,json=webEntities" json:"web_entities,omitempty"` + WebEntities []*WebDetection_WebEntity `protobuf:"bytes,1,rep,name=web_entities,json=webEntities,proto3" json:"web_entities,omitempty"` // Fully matching images from the Internet. // Can include resized copies of the query image. - FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,2,rep,name=full_matching_images,json=fullMatchingImages" json:"full_matching_images,omitempty"` + FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,2,rep,name=full_matching_images,json=fullMatchingImages,proto3" json:"full_matching_images,omitempty"` // Partial matching images from the Internet. // Those images are similar enough to share some key-point features. For // example an original image will likely have partial matching for its crops. - PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,3,rep,name=partial_matching_images,json=partialMatchingImages" json:"partial_matching_images,omitempty"` + PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,3,rep,name=partial_matching_images,json=partialMatchingImages,proto3" json:"partial_matching_images,omitempty"` // Web pages containing the matching images from the Internet. - PagesWithMatchingImages []*WebDetection_WebPage `protobuf:"bytes,4,rep,name=pages_with_matching_images,json=pagesWithMatchingImages" json:"pages_with_matching_images,omitempty"` + PagesWithMatchingImages []*WebDetection_WebPage `protobuf:"bytes,4,rep,name=pages_with_matching_images,json=pagesWithMatchingImages,proto3" json:"pages_with_matching_images,omitempty"` // The visually similar image results. - VisuallySimilarImages []*WebDetection_WebImage `protobuf:"bytes,6,rep,name=visually_similar_images,json=visuallySimilarImages" json:"visually_similar_images,omitempty"` + VisuallySimilarImages []*WebDetection_WebImage `protobuf:"bytes,6,rep,name=visually_similar_images,json=visuallySimilarImages,proto3" json:"visually_similar_images,omitempty"` // Best guess text labels for the request image. - BestGuessLabels []*WebDetection_WebLabel `protobuf:"bytes,8,rep,name=best_guess_labels,json=bestGuessLabels" json:"best_guess_labels,omitempty"` + BestGuessLabels []*WebDetection_WebLabel `protobuf:"bytes,8,rep,name=best_guess_labels,json=bestGuessLabels,proto3" json:"best_guess_labels,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -110,12 +110,12 @@ func (m *WebDetection) GetBestGuessLabels() []*WebDetection_WebLabel { // Entity deduced from similar images on the Internet. type WebDetection_WebEntity struct { // Opaque entity ID. - EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Overall relevancy score for the entity. // Not normalized and not comparable across different image queries. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // Canonical description of the entity, in English. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,9 +169,9 @@ func (m *WebDetection_WebEntity) GetDescription() string { // Metadata for online images. type WebDetection_WebImage struct { // The result image URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // (Deprecated) Overall relevancy score for the image. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -218,19 +218,19 @@ func (m *WebDetection_WebImage) GetScore() float32 { // Metadata for web pages. type WebDetection_WebPage struct { // The result web page URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // (Deprecated) Overall relevancy score for the web page. - Score float32 `protobuf:"fixed32,2,opt,name=score" json:"score,omitempty"` + Score float32 `protobuf:"fixed32,2,opt,name=score,proto3" json:"score,omitempty"` // Title for the web page, may contain HTML markups. - PageTitle string `protobuf:"bytes,3,opt,name=page_title,json=pageTitle" json:"page_title,omitempty"` + PageTitle string `protobuf:"bytes,3,opt,name=page_title,json=pageTitle,proto3" json:"page_title,omitempty"` // Fully matching images on the page. // Can include resized copies of the query image. - FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,4,rep,name=full_matching_images,json=fullMatchingImages" json:"full_matching_images,omitempty"` + FullMatchingImages []*WebDetection_WebImage `protobuf:"bytes,4,rep,name=full_matching_images,json=fullMatchingImages,proto3" json:"full_matching_images,omitempty"` // Partial matching images on the page. // Those images are similar enough to share some key-point features. For // example an original image will likely have partial matching for its // crops. - PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,5,rep,name=partial_matching_images,json=partialMatchingImages" json:"partial_matching_images,omitempty"` + PartialMatchingImages []*WebDetection_WebImage `protobuf:"bytes,5,rep,name=partial_matching_images,json=partialMatchingImages,proto3" json:"partial_matching_images,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -298,11 +298,11 @@ func (m *WebDetection_WebPage) GetPartialMatchingImages() []*WebDetection_WebIma // Label to provide extra metadata for the web detection. type WebDetection_WebLabel struct { // Label for extra metadata. - Label string `protobuf:"bytes,1,opt,name=label" json:"label,omitempty"` + Label string `protobuf:"bytes,1,opt,name=label,proto3" json:"label,omitempty"` // The BCP-47 language code for `label`, such as "en-US" or "sr-Latn". // For more information, see // http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. - LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,2,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/crawled_url.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/crawled_url.pb.go index 91588924c..7d8e495c4 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/crawled_url.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/crawled_url.pb.go @@ -26,13 +26,13 @@ type CrawledUrl struct { // Output only. // The http method of the request that was used to visit the URL, in // uppercase. - HttpMethod string `protobuf:"bytes,1,opt,name=http_method,json=httpMethod" json:"http_method,omitempty"` + HttpMethod string `protobuf:"bytes,1,opt,name=http_method,json=httpMethod,proto3" json:"http_method,omitempty"` // Output only. // The URL that was crawled. - Url string `protobuf:"bytes,2,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` // Output only. // The body of the request that was used to visit the URL. - Body string `protobuf:"bytes,3,opt,name=body" json:"body,omitempty"` + Body string `protobuf:"bytes,3,opt,name=body,proto3" json:"body,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding.pb.go index 1929d030f..47b53a745 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding.pb.go @@ -94,53 +94,53 @@ type Finding struct { // The resource name of the Finding. The name follows the format of // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanruns/{scanRunId}/findings/{findingId}'. // The finding IDs are generated by the system. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Output only. // The type of the Finding. - FindingType Finding_FindingType `protobuf:"varint,2,opt,name=finding_type,json=findingType,enum=google.cloud.websecurityscanner.v1alpha.Finding_FindingType" json:"finding_type,omitempty"` + FindingType Finding_FindingType `protobuf:"varint,2,opt,name=finding_type,json=findingType,proto3,enum=google.cloud.websecurityscanner.v1alpha.Finding_FindingType" json:"finding_type,omitempty"` // Output only. // The http method of the request that triggered the vulnerability, in // uppercase. - HttpMethod string `protobuf:"bytes,3,opt,name=http_method,json=httpMethod" json:"http_method,omitempty"` + HttpMethod string `protobuf:"bytes,3,opt,name=http_method,json=httpMethod,proto3" json:"http_method,omitempty"` // Output only. // The URL produced by the server-side fuzzer and used in the request that // triggered the vulnerability. - FuzzedUrl string `protobuf:"bytes,4,opt,name=fuzzed_url,json=fuzzedUrl" json:"fuzzed_url,omitempty"` + FuzzedUrl string `protobuf:"bytes,4,opt,name=fuzzed_url,json=fuzzedUrl,proto3" json:"fuzzed_url,omitempty"` // Output only. // The body of the request that triggered the vulnerability. - Body string `protobuf:"bytes,5,opt,name=body" json:"body,omitempty"` + Body string `protobuf:"bytes,5,opt,name=body,proto3" json:"body,omitempty"` // Output only. // The description of the vulnerability. - Description string `protobuf:"bytes,6,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,6,opt,name=description,proto3" json:"description,omitempty"` // Output only. // The URL containing human-readable payload that user can leverage to // reproduce the vulnerability. - ReproductionUrl string `protobuf:"bytes,7,opt,name=reproduction_url,json=reproductionUrl" json:"reproduction_url,omitempty"` + ReproductionUrl string `protobuf:"bytes,7,opt,name=reproduction_url,json=reproductionUrl,proto3" json:"reproduction_url,omitempty"` // Output only. // If the vulnerability was originated from nested IFrame, the immediate // parent IFrame is reported. - FrameUrl string `protobuf:"bytes,8,opt,name=frame_url,json=frameUrl" json:"frame_url,omitempty"` + FrameUrl string `protobuf:"bytes,8,opt,name=frame_url,json=frameUrl,proto3" json:"frame_url,omitempty"` // Output only. // The URL where the browser lands when the vulnerability is detected. - FinalUrl string `protobuf:"bytes,9,opt,name=final_url,json=finalUrl" json:"final_url,omitempty"` + FinalUrl string `protobuf:"bytes,9,opt,name=final_url,json=finalUrl,proto3" json:"final_url,omitempty"` // Output only. // The tracking ID uniquely identifies a vulnerability instance across // multiple ScanRuns. - TrackingId string `protobuf:"bytes,10,opt,name=tracking_id,json=trackingId" json:"tracking_id,omitempty"` + TrackingId string `protobuf:"bytes,10,opt,name=tracking_id,json=trackingId,proto3" json:"tracking_id,omitempty"` // Output only. // An addon containing information about outdated libraries. - OutdatedLibrary *OutdatedLibrary `protobuf:"bytes,11,opt,name=outdated_library,json=outdatedLibrary" json:"outdated_library,omitempty"` + OutdatedLibrary *OutdatedLibrary `protobuf:"bytes,11,opt,name=outdated_library,json=outdatedLibrary,proto3" json:"outdated_library,omitempty"` // Output only. // An addon containing detailed information regarding any resource causing the // vulnerability such as JavaScript sources, image, audio files, etc. - ViolatingResource *ViolatingResource `protobuf:"bytes,12,opt,name=violating_resource,json=violatingResource" json:"violating_resource,omitempty"` + ViolatingResource *ViolatingResource `protobuf:"bytes,12,opt,name=violating_resource,json=violatingResource,proto3" json:"violating_resource,omitempty"` // Output only. // An addon containing information about request parameters which were found // to be vulnerable. - VulnerableParameters *VulnerableParameters `protobuf:"bytes,13,opt,name=vulnerable_parameters,json=vulnerableParameters" json:"vulnerable_parameters,omitempty"` + VulnerableParameters *VulnerableParameters `protobuf:"bytes,13,opt,name=vulnerable_parameters,json=vulnerableParameters,proto3" json:"vulnerable_parameters,omitempty"` // Output only. // An addon containing information reported for an XSS, if any. - Xss *Xss `protobuf:"bytes,14,opt,name=xss" json:"xss,omitempty"` + Xss *Xss `protobuf:"bytes,14,opt,name=xss,proto3" json:"xss,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_addon.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_addon.pb.go index 5fff24a2a..4f8c883e0 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_addon.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_addon.pb.go @@ -22,11 +22,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Information reported for an outdated library. type OutdatedLibrary struct { // The name of the outdated library. - LibraryName string `protobuf:"bytes,1,opt,name=library_name,json=libraryName" json:"library_name,omitempty"` + LibraryName string `protobuf:"bytes,1,opt,name=library_name,json=libraryName,proto3" json:"library_name,omitempty"` // The version number. - Version string `protobuf:"bytes,2,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` // URLs to learn more information about the vulnerabilities in the library. - LearnMoreUrls []string `protobuf:"bytes,3,rep,name=learn_more_urls,json=learnMoreUrls" json:"learn_more_urls,omitempty"` + LearnMoreUrls []string `protobuf:"bytes,3,rep,name=learn_more_urls,json=learnMoreUrls,proto3" json:"learn_more_urls,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -81,9 +81,9 @@ func (m *OutdatedLibrary) GetLearnMoreUrls() []string { // as JavaScript sources, image, audio files, etc. type ViolatingResource struct { // The MIME type of this resource. - ContentType string `protobuf:"bytes,1,opt,name=content_type,json=contentType" json:"content_type,omitempty"` + ContentType string `protobuf:"bytes,1,opt,name=content_type,json=contentType,proto3" json:"content_type,omitempty"` // URL of this violating resource. - ResourceUrl string `protobuf:"bytes,2,opt,name=resource_url,json=resourceUrl" json:"resource_url,omitempty"` + ResourceUrl string `protobuf:"bytes,2,opt,name=resource_url,json=resourceUrl,proto3" json:"resource_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -130,7 +130,7 @@ func (m *ViolatingResource) GetResourceUrl() string { // Information about vulnerable request parameters. type VulnerableParameters struct { // The vulnerable parameter names. - ParameterNames []string `protobuf:"bytes,1,rep,name=parameter_names,json=parameterNames" json:"parameter_names,omitempty"` + ParameterNames []string `protobuf:"bytes,1,rep,name=parameter_names,json=parameterNames,proto3" json:"parameter_names,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -170,9 +170,9 @@ func (m *VulnerableParameters) GetParameterNames() []string { // Information reported for an XSS. type Xss struct { // Stack traces leading to the point where the XSS occurred. - StackTraces []string `protobuf:"bytes,1,rep,name=stack_traces,json=stackTraces" json:"stack_traces,omitempty"` + StackTraces []string `protobuf:"bytes,1,rep,name=stack_traces,json=stackTraces,proto3" json:"stack_traces,omitempty"` // An error message generated by a javascript breakage. - ErrorMessage string `protobuf:"bytes,2,opt,name=error_message,json=errorMessage" json:"error_message,omitempty"` + ErrorMessage string `protobuf:"bytes,2,opt,name=error_message,json=errorMessage,proto3" json:"error_message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_type_stats.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_type_stats.pb.go index b9965e9a0..5a5c81f3d 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_type_stats.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/finding_type_stats.pb.go @@ -24,10 +24,10 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type FindingTypeStats struct { // Output only. // The finding type associated with the stats. - FindingType Finding_FindingType `protobuf:"varint,1,opt,name=finding_type,json=findingType,enum=google.cloud.websecurityscanner.v1alpha.Finding_FindingType" json:"finding_type,omitempty"` + FindingType Finding_FindingType `protobuf:"varint,1,opt,name=finding_type,json=findingType,proto3,enum=google.cloud.websecurityscanner.v1alpha.Finding_FindingType" json:"finding_type,omitempty"` // Output only. // The count of findings belonging to this finding type. - FindingCount int32 `protobuf:"varint,2,opt,name=finding_count,json=findingCount" json:"finding_count,omitempty"` + FindingCount int32 `protobuf:"varint,2,opt,name=finding_count,json=findingCount,proto3" json:"finding_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_config.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_config.pb.go index 6c9af0d58..cc1eece0f 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_config.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_config.pb.go @@ -90,31 +90,31 @@ type ScanConfig struct { // The resource name of the ScanConfig. The name follows the format of // 'projects/{projectId}/scanConfigs/{scanConfigId}'. The ScanConfig IDs are // generated by the system. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. // The user provided display name of the ScanConfig. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The maximum QPS during scanning. A valid value ranges from 5 to 20 // inclusively. If the field is unspecified or its value is set 0, server will // default to 15. Other values outside of [5, 20] range will be rejected with // INVALID_ARGUMENT error. - MaxQps int32 `protobuf:"varint,3,opt,name=max_qps,json=maxQps" json:"max_qps,omitempty"` + MaxQps int32 `protobuf:"varint,3,opt,name=max_qps,json=maxQps,proto3" json:"max_qps,omitempty"` // Required. // The starting URLs from which the scanner finds site pages. - StartingUrls []string `protobuf:"bytes,4,rep,name=starting_urls,json=startingUrls" json:"starting_urls,omitempty"` + StartingUrls []string `protobuf:"bytes,4,rep,name=starting_urls,json=startingUrls,proto3" json:"starting_urls,omitempty"` // The authentication configuration. If specified, service will use the // authentication configuration during scanning. - Authentication *ScanConfig_Authentication `protobuf:"bytes,5,opt,name=authentication" json:"authentication,omitempty"` + Authentication *ScanConfig_Authentication `protobuf:"bytes,5,opt,name=authentication,proto3" json:"authentication,omitempty"` // The user agent used during scanning. - UserAgent ScanConfig_UserAgent `protobuf:"varint,6,opt,name=user_agent,json=userAgent,enum=google.cloud.websecurityscanner.v1alpha.ScanConfig_UserAgent" json:"user_agent,omitempty"` + UserAgent ScanConfig_UserAgent `protobuf:"varint,6,opt,name=user_agent,json=userAgent,proto3,enum=google.cloud.websecurityscanner.v1alpha.ScanConfig_UserAgent" json:"user_agent,omitempty"` // The blacklist URL patterns as described in // https://cloud.google.com/security-scanner/docs/excluded-urls - BlacklistPatterns []string `protobuf:"bytes,7,rep,name=blacklist_patterns,json=blacklistPatterns" json:"blacklist_patterns,omitempty"` + BlacklistPatterns []string `protobuf:"bytes,7,rep,name=blacklist_patterns,json=blacklistPatterns,proto3" json:"blacklist_patterns,omitempty"` // The schedule of the ScanConfig. - Schedule *ScanConfig_Schedule `protobuf:"bytes,8,opt,name=schedule" json:"schedule,omitempty"` + Schedule *ScanConfig_Schedule `protobuf:"bytes,8,opt,name=schedule,proto3" json:"schedule,omitempty"` // Set of Cloud Platforms targeted by the scan. If empty, APP_ENGINE will be // used as a default. - TargetPlatforms []ScanConfig_TargetPlatform `protobuf:"varint,9,rep,packed,name=target_platforms,json=targetPlatforms,enum=google.cloud.websecurityscanner.v1alpha.ScanConfig_TargetPlatform" json:"target_platforms,omitempty"` + TargetPlatforms []ScanConfig_TargetPlatform `protobuf:"varint,9,rep,packed,name=target_platforms,json=targetPlatforms,proto3,enum=google.cloud.websecurityscanner.v1alpha.ScanConfig_TargetPlatform" json:"target_platforms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -250,10 +250,10 @@ type isScanConfig_Authentication_Authentication interface { } type ScanConfig_Authentication_GoogleAccount_ struct { - GoogleAccount *ScanConfig_Authentication_GoogleAccount `protobuf:"bytes,1,opt,name=google_account,json=googleAccount,oneof"` + GoogleAccount *ScanConfig_Authentication_GoogleAccount `protobuf:"bytes,1,opt,name=google_account,json=googleAccount,proto3,oneof"` } type ScanConfig_Authentication_CustomAccount_ struct { - CustomAccount *ScanConfig_Authentication_CustomAccount `protobuf:"bytes,2,opt,name=custom_account,json=customAccount,oneof"` + CustomAccount *ScanConfig_Authentication_CustomAccount `protobuf:"bytes,2,opt,name=custom_account,json=customAccount,proto3,oneof"` } func (*ScanConfig_Authentication_GoogleAccount_) isScanConfig_Authentication_Authentication() {} @@ -358,12 +358,12 @@ func _ScanConfig_Authentication_OneofSizer(msg proto.Message) (n int) { type ScanConfig_Authentication_GoogleAccount struct { // Required. // The user name of the Google account. - Username string `protobuf:"bytes,1,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"` // Input only. // Required. // The password of the Google account. The credential is stored encrypted // and not returned in any response. - Password string `protobuf:"bytes,2,opt,name=password" json:"password,omitempty"` + Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -413,15 +413,15 @@ func (m *ScanConfig_Authentication_GoogleAccount) GetPassword() string { type ScanConfig_Authentication_CustomAccount struct { // Required. // The user name of the custom account. - Username string `protobuf:"bytes,1,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"` // Input only. // Required. // The password of the custom account. The credential is stored encrypted // and not returned in any response. - Password string `protobuf:"bytes,2,opt,name=password" json:"password,omitempty"` + Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"` // Required. // The login form URL of the website. - LoginUrl string `protobuf:"bytes,3,opt,name=login_url,json=loginUrl" json:"login_url,omitempty"` + LoginUrl string `protobuf:"bytes,3,opt,name=login_url,json=loginUrl,proto3" json:"login_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -480,10 +480,10 @@ type ScanConfig_Schedule struct { // refreshed by the server after each run. If unspecified, it will default // to current server time, which means the scan will be scheduled to start // immediately. - ScheduleTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=schedule_time,json=scheduleTime" json:"schedule_time,omitempty"` + ScheduleTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"` // Required. // The duration of time between executions in days. - IntervalDurationDays int32 `protobuf:"varint,2,opt,name=interval_duration_days,json=intervalDurationDays" json:"interval_duration_days,omitempty"` + IntervalDurationDays int32 `protobuf:"varint,2,opt,name=interval_duration_days,json=intervalDurationDays,proto3" json:"interval_duration_days,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_run.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_run.pb.go index 2f77de68f..323002594 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_run.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/scan_run.pb.go @@ -96,40 +96,40 @@ type ScanRun struct { // The resource name of the ScanRun. The name follows the format of // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}'. // The ScanRun IDs are generated by the system. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Output only. // The execution state of the ScanRun. - ExecutionState ScanRun_ExecutionState `protobuf:"varint,2,opt,name=execution_state,json=executionState,enum=google.cloud.websecurityscanner.v1alpha.ScanRun_ExecutionState" json:"execution_state,omitempty"` + ExecutionState ScanRun_ExecutionState `protobuf:"varint,2,opt,name=execution_state,json=executionState,proto3,enum=google.cloud.websecurityscanner.v1alpha.ScanRun_ExecutionState" json:"execution_state,omitempty"` // Output only. // The result state of the ScanRun. This field is only available after the // execution state reaches "FINISHED". - ResultState ScanRun_ResultState `protobuf:"varint,3,opt,name=result_state,json=resultState,enum=google.cloud.websecurityscanner.v1alpha.ScanRun_ResultState" json:"result_state,omitempty"` + ResultState ScanRun_ResultState `protobuf:"varint,3,opt,name=result_state,json=resultState,proto3,enum=google.cloud.websecurityscanner.v1alpha.ScanRun_ResultState" json:"result_state,omitempty"` // Output only. // The time at which the ScanRun started. - StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Output only. // The time at which the ScanRun reached termination state - that the ScanRun // is either finished or stopped by user. - EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Output only. // The number of URLs crawled during this ScanRun. If the scan is in progress, // the value represents the number of URLs crawled up to now. - UrlsCrawledCount int64 `protobuf:"varint,6,opt,name=urls_crawled_count,json=urlsCrawledCount" json:"urls_crawled_count,omitempty"` + UrlsCrawledCount int64 `protobuf:"varint,6,opt,name=urls_crawled_count,json=urlsCrawledCount,proto3" json:"urls_crawled_count,omitempty"` // Output only. // The number of URLs tested during this ScanRun. If the scan is in progress, // the value represents the number of URLs tested up to now. The number of // URLs tested is usually larger than the number URLS crawled because // typically a crawled URL is tested with multiple test payloads. - UrlsTestedCount int64 `protobuf:"varint,7,opt,name=urls_tested_count,json=urlsTestedCount" json:"urls_tested_count,omitempty"` + UrlsTestedCount int64 `protobuf:"varint,7,opt,name=urls_tested_count,json=urlsTestedCount,proto3" json:"urls_tested_count,omitempty"` // Output only. // Whether the scan run has found any vulnerabilities. - HasVulnerabilities bool `protobuf:"varint,8,opt,name=has_vulnerabilities,json=hasVulnerabilities" json:"has_vulnerabilities,omitempty"` + HasVulnerabilities bool `protobuf:"varint,8,opt,name=has_vulnerabilities,json=hasVulnerabilities,proto3" json:"has_vulnerabilities,omitempty"` // Output only. // The percentage of total completion ranging from 0 to 100. // If the scan is in queue, the value is 0. // If the scan is running, the value ranges from 0 to 100. // If the scan is finished, the value is 100. - ProgressPercent int32 `protobuf:"varint,9,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` + ProgressPercent int32 `protobuf:"varint,9,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/web_security_scanner.pb.go b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/web_security_scanner.pb.go index af4ca037d..ac208fa38 100644 --- a/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/web_security_scanner.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/cloud/websecurityscanner/v1alpha/web_security_scanner.pb.go @@ -31,10 +31,10 @@ type CreateScanConfigRequest struct { // Required. // The parent resource name where the scan is created, which should be a // project resource name in the format 'projects/{projectId}'. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. // The ScanConfig to be created. - ScanConfig *ScanConfig `protobuf:"bytes,2,opt,name=scan_config,json=scanConfig" json:"scan_config,omitempty"` + ScanConfig *ScanConfig `protobuf:"bytes,2,opt,name=scan_config,json=scanConfig,proto3" json:"scan_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -83,7 +83,7 @@ type DeleteScanConfigRequest struct { // Required. // The resource name of the ScanConfig to be deleted. The name follows the // format of 'projects/{projectId}/scanConfigs/{scanConfigId}'. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -125,7 +125,7 @@ type GetScanConfigRequest struct { // Required. // The resource name of the ScanConfig to be returned. The name follows the // format of 'projects/{projectId}/scanConfigs/{scanConfigId}'. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -167,15 +167,15 @@ type ListScanConfigsRequest struct { // Required. // The parent resource name, which should be a project resource name in the // format 'projects/{projectId}'. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // A token identifying a page of results to be returned. This should be a // `next_page_token` value returned from a previous List request. // If unspecified, the first page of results is returned. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of ScanConfigs to return, can be limited by server. // If not specified or not positive, the implementation will select a // reasonable value. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -232,12 +232,12 @@ type UpdateScanConfigRequest struct { // The ScanConfig to be updated. The name field must be set to identify the // resource to be updated. The values of fields not covered by the mask // will be ignored. - ScanConfig *ScanConfig `protobuf:"bytes,2,opt,name=scan_config,json=scanConfig" json:"scan_config,omitempty"` + ScanConfig *ScanConfig `protobuf:"bytes,2,opt,name=scan_config,json=scanConfig,proto3" json:"scan_config,omitempty"` // Required. // The update mask applies to the resource. For the `FieldMask` definition, // see // https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#fieldmask - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -284,10 +284,10 @@ func (m *UpdateScanConfigRequest) GetUpdateMask() *field_mask.FieldMask { // Response for the `ListScanConfigs` method. type ListScanConfigsResponse struct { // The list of ScanConfigs returned. - ScanConfigs []*ScanConfig `protobuf:"bytes,1,rep,name=scan_configs,json=scanConfigs" json:"scan_configs,omitempty"` + ScanConfigs []*ScanConfig `protobuf:"bytes,1,rep,name=scan_configs,json=scanConfigs,proto3" json:"scan_configs,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -336,7 +336,7 @@ type StartScanRunRequest struct { // Required. // The resource name of the ScanConfig to be used. The name follows the // format of 'projects/{projectId}/scanConfigs/{scanConfigId}'. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -379,7 +379,7 @@ type GetScanRunRequest struct { // The resource name of the ScanRun to be returned. The name follows the // format of // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}'. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -421,15 +421,15 @@ type ListScanRunsRequest struct { // Required. // The parent resource name, which should be a scan resource name in the // format 'projects/{projectId}/scanConfigs/{scanConfigId}'. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // A token identifying a page of results to be returned. This should be a // `next_page_token` value returned from a previous List request. // If unspecified, the first page of results is returned. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of ScanRuns to return, can be limited by server. // If not specified or not positive, the implementation will select a // reasonable value. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -483,10 +483,10 @@ func (m *ListScanRunsRequest) GetPageSize() int32 { // Response for the `ListScanRuns` method. type ListScanRunsResponse struct { // The list of ScanRuns returned. - ScanRuns []*ScanRun `protobuf:"bytes,1,rep,name=scan_runs,json=scanRuns" json:"scan_runs,omitempty"` + ScanRuns []*ScanRun `protobuf:"bytes,1,rep,name=scan_runs,json=scanRuns,proto3" json:"scan_runs,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -536,7 +536,7 @@ type StopScanRunRequest struct { // The resource name of the ScanRun to be stopped. The name follows the // format of // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}'. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -579,15 +579,15 @@ type ListCrawledUrlsRequest struct { // The parent resource name, which should be a scan run resource name in the // format // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}'. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // A token identifying a page of results to be returned. This should be a // `next_page_token` value returned from a previous List request. // If unspecified, the first page of results is returned. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of CrawledUrls to return, can be limited by server. // If not specified or not positive, the implementation will select a // reasonable value. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -641,10 +641,10 @@ func (m *ListCrawledUrlsRequest) GetPageSize() int32 { // Response for the `ListCrawledUrls` method. type ListCrawledUrlsResponse struct { // The list of CrawledUrls returned. - CrawledUrls []*CrawledUrl `protobuf:"bytes,1,rep,name=crawled_urls,json=crawledUrls" json:"crawled_urls,omitempty"` + CrawledUrls []*CrawledUrl `protobuf:"bytes,1,rep,name=crawled_urls,json=crawledUrls,proto3" json:"crawled_urls,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -694,7 +694,7 @@ type GetFindingRequest struct { // The resource name of the Finding to be returned. The name follows the // format of // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}/findings/{findingId}'. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -737,20 +737,20 @@ type ListFindingsRequest struct { // The parent resource name, which should be a scan run resource name in the // format // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}'. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The filter expression. The expression must be in the format: // . // Supported field: 'finding_type'. // Supported operator: '='. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // A token identifying a page of results to be returned. This should be a // `next_page_token` value returned from a previous List request. // If unspecified, the first page of results is returned. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of Findings to return, can be limited by server. // If not specified or not positive, the implementation will select a // reasonable value. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -811,10 +811,10 @@ func (m *ListFindingsRequest) GetPageSize() int32 { // Response for the `ListFindings` method. type ListFindingsResponse struct { // The list of Findings returned. - Findings []*Finding `protobuf:"bytes,1,rep,name=findings" json:"findings,omitempty"` + Findings []*Finding `protobuf:"bytes,1,rep,name=findings,proto3" json:"findings,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -864,7 +864,7 @@ type ListFindingTypeStatsRequest struct { // The parent resource name, which should be a scan run resource name in the // format // 'projects/{projectId}/scanConfigs/{scanConfigId}/scanRuns/{scanRunId}'. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -904,7 +904,7 @@ func (m *ListFindingTypeStatsRequest) GetParent() string { // Response for the `ListFindingTypeStats` method. type ListFindingTypeStatsResponse struct { // The list of FindingTypeStats returned. - FindingTypeStats []*FindingTypeStats `protobuf:"bytes,1,rep,name=finding_type_stats,json=findingTypeStats" json:"finding_type_stats,omitempty"` + FindingTypeStats []*FindingTypeStats `protobuf:"bytes,1,rep,name=finding_type_stats,json=findingTypeStats,proto3" json:"finding_type_stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1128,8 +1128,7 @@ func (c *webSecurityScannerClient) ListFindingTypeStats(ctx context.Context, in return out, nil } -// Server API for WebSecurityScanner service - +// WebSecurityScannerServer is the server API for WebSecurityScanner service. type WebSecurityScannerServer interface { // Creates a new ScanConfig. CreateScanConfig(context.Context, *CreateScanConfigRequest) (*ScanConfig, error) diff --git a/vendor/google.golang.org/genproto/googleapis/container/v1/cluster_service.pb.go b/vendor/google.golang.org/genproto/googleapis/container/v1/cluster_service.pb.go index cad7d1350..7fcb2380e 100644 --- a/vendor/google.golang.org/genproto/googleapis/container/v1/cluster_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/container/v1/cluster_service.pb.go @@ -319,12 +319,12 @@ type NodeConfig struct { // // If unspecified, the default machine type is // `n1-standard-1`. - MachineType string `protobuf:"bytes,1,opt,name=machine_type,json=machineType" json:"machine_type,omitempty"` + MachineType string `protobuf:"bytes,1,opt,name=machine_type,json=machineType,proto3" json:"machine_type,omitempty"` // Size of the disk attached to each node, specified in GB. // The smallest allowed disk size is 10GB. // // If unspecified, the default disk size is 100GB. - DiskSizeGb int32 `protobuf:"varint,2,opt,name=disk_size_gb,json=diskSizeGb" json:"disk_size_gb,omitempty"` + DiskSizeGb int32 `protobuf:"varint,2,opt,name=disk_size_gb,json=diskSizeGb,proto3" json:"disk_size_gb,omitempty"` // The set of Google API scopes to be made available on all of the // node VMs under the "default" service account. // @@ -339,10 +339,10 @@ type NodeConfig struct { // // If unspecified, no scopes are added, unless Cloud Logging or Cloud // Monitoring are enabled, in which case their required scopes will be added. - OauthScopes []string `protobuf:"bytes,3,rep,name=oauth_scopes,json=oauthScopes" json:"oauth_scopes,omitempty"` + OauthScopes []string `protobuf:"bytes,3,rep,name=oauth_scopes,json=oauthScopes,proto3" json:"oauth_scopes,omitempty"` // The Google Cloud Platform Service Account to be used by the node VMs. If // no Service Account is specified, the "default" service account is used. - ServiceAccount string `protobuf:"bytes,9,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount string `protobuf:"bytes,9,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // The metadata key/value pairs assigned to instances in the cluster. // // Keys must conform to the regexp [a-zA-Z0-9-_]+ and be less than 128 bytes @@ -356,10 +356,10 @@ type NodeConfig struct { // that each value's size must be less than or equal to 32 KB. // // The total size of all keys and values must be less than 512 KB. - Metadata map[string]string `protobuf:"bytes,4,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,4,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The image type to use for this node. Note that for a given image type, // the latest version of it will be used. - ImageType string `protobuf:"bytes,5,opt,name=image_type,json=imageType" json:"image_type,omitempty"` + ImageType string `protobuf:"bytes,5,opt,name=image_type,json=imageType,proto3" json:"image_type,omitempty"` // The map of Kubernetes labels (key/value pairs) to be applied to each node. // These will added in addition to any default label(s) that // Kubernetes may apply to the node. @@ -368,34 +368,34 @@ type NodeConfig struct { // and conflicts should be avoided. // For more information, including usage and the valid values, see: // https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/ - Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,6,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The number of local SSD disks to be attached to the node. // // The limit for this value is dependant upon the maximum number of // disks available on a machine per zone. See: // https://cloud.google.com/compute/docs/disks/local-ssd#local_ssd_limits // for more information. - LocalSsdCount int32 `protobuf:"varint,7,opt,name=local_ssd_count,json=localSsdCount" json:"local_ssd_count,omitempty"` + LocalSsdCount int32 `protobuf:"varint,7,opt,name=local_ssd_count,json=localSsdCount,proto3" json:"local_ssd_count,omitempty"` // The list of instance tags applied to all nodes. Tags are used to identify // valid sources or targets for network firewalls and are specified by // the client during cluster or node pool creation. Each tag within the list // must comply with RFC1035. - Tags []string `protobuf:"bytes,8,rep,name=tags" json:"tags,omitempty"` + Tags []string `protobuf:"bytes,8,rep,name=tags,proto3" json:"tags,omitempty"` // Whether the nodes are created as preemptible VM instances. See: // https://cloud.google.com/compute/docs/instances/preemptible for more // information about preemptible VM instances. - Preemptible bool `protobuf:"varint,10,opt,name=preemptible" json:"preemptible,omitempty"` + Preemptible bool `protobuf:"varint,10,opt,name=preemptible,proto3" json:"preemptible,omitempty"` // A list of hardware accelerators to be attached to each node. // See https://cloud.google.com/compute/docs/gpus for more information about // support for GPUs. - Accelerators []*AcceleratorConfig `protobuf:"bytes,11,rep,name=accelerators" json:"accelerators,omitempty"` + Accelerators []*AcceleratorConfig `protobuf:"bytes,11,rep,name=accelerators,proto3" json:"accelerators,omitempty"` // Minimum CPU platform to be used by this instance. The instance may be // scheduled on the specified or newer CPU platform. Applicable values are the // friendly names of CPU platforms, such as // minCpuPlatform: "Intel Haswell" or // minCpuPlatform: "Intel Sandy Bridge". For more // information, read [how to specify min CPU platform](https://cloud.google.com/compute/docs/instances/specify-min-cpu-platform) - MinCpuPlatform string `protobuf:"bytes,13,opt,name=min_cpu_platform,json=minCpuPlatform" json:"min_cpu_platform,omitempty"` + MinCpuPlatform string `protobuf:"bytes,13,opt,name=min_cpu_platform,json=minCpuPlatform,proto3" json:"min_cpu_platform,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -516,24 +516,24 @@ type MasterAuth struct { // The username to use for HTTP basic authentication to the master endpoint. // For clusters v1.6.0 and later, you can disable basic authentication by // providing an empty username. - Username string `protobuf:"bytes,1,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"` // The password to use for HTTP basic authentication to the master endpoint. // Because the master endpoint is open to the Internet, you should create a // strong password. If a password is provided for cluster creation, username // must be non-empty. - Password string `protobuf:"bytes,2,opt,name=password" json:"password,omitempty"` + Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"` // Configuration for client certificate authentication on the cluster. If no // configuration is specified, a client certificate is issued. - ClientCertificateConfig *ClientCertificateConfig `protobuf:"bytes,3,opt,name=client_certificate_config,json=clientCertificateConfig" json:"client_certificate_config,omitempty"` + ClientCertificateConfig *ClientCertificateConfig `protobuf:"bytes,3,opt,name=client_certificate_config,json=clientCertificateConfig,proto3" json:"client_certificate_config,omitempty"` // [Output only] Base64-encoded public certificate that is the root of // trust for the cluster. - ClusterCaCertificate string `protobuf:"bytes,100,opt,name=cluster_ca_certificate,json=clusterCaCertificate" json:"cluster_ca_certificate,omitempty"` + ClusterCaCertificate string `protobuf:"bytes,100,opt,name=cluster_ca_certificate,json=clusterCaCertificate,proto3" json:"cluster_ca_certificate,omitempty"` // [Output only] Base64-encoded public certificate used by clients to // authenticate to the cluster endpoint. - ClientCertificate string `protobuf:"bytes,101,opt,name=client_certificate,json=clientCertificate" json:"client_certificate,omitempty"` + ClientCertificate string `protobuf:"bytes,101,opt,name=client_certificate,json=clientCertificate,proto3" json:"client_certificate,omitempty"` // [Output only] Base64-encoded private key used by clients to authenticate // to the cluster endpoint. - ClientKey string `protobuf:"bytes,102,opt,name=client_key,json=clientKey" json:"client_key,omitempty"` + ClientKey string `protobuf:"bytes,102,opt,name=client_key,json=clientKey,proto3" json:"client_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -608,7 +608,7 @@ func (m *MasterAuth) GetClientKey() string { // Configuration for client certificates on the cluster. type ClientCertificateConfig struct { // Issue a client certificate. - IssueClientCertificate bool `protobuf:"varint,1,opt,name=issue_client_certificate,json=issueClientCertificate" json:"issue_client_certificate,omitempty"` + IssueClientCertificate bool `protobuf:"varint,1,opt,name=issue_client_certificate,json=issueClientCertificate,proto3" json:"issue_client_certificate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -650,17 +650,17 @@ func (m *ClientCertificateConfig) GetIssueClientCertificate() bool { type AddonsConfig struct { // Configuration for the HTTP (L7) load balancing controller addon, which // makes it easy to set up HTTP load balancers for services in a cluster. - HttpLoadBalancing *HttpLoadBalancing `protobuf:"bytes,1,opt,name=http_load_balancing,json=httpLoadBalancing" json:"http_load_balancing,omitempty"` + HttpLoadBalancing *HttpLoadBalancing `protobuf:"bytes,1,opt,name=http_load_balancing,json=httpLoadBalancing,proto3" json:"http_load_balancing,omitempty"` // Configuration for the horizontal pod autoscaling feature, which // increases or decreases the number of replica pods a replication controller // has based on the resource usage of the existing pods. - HorizontalPodAutoscaling *HorizontalPodAutoscaling `protobuf:"bytes,2,opt,name=horizontal_pod_autoscaling,json=horizontalPodAutoscaling" json:"horizontal_pod_autoscaling,omitempty"` + HorizontalPodAutoscaling *HorizontalPodAutoscaling `protobuf:"bytes,2,opt,name=horizontal_pod_autoscaling,json=horizontalPodAutoscaling,proto3" json:"horizontal_pod_autoscaling,omitempty"` // Configuration for the Kubernetes Dashboard. - KubernetesDashboard *KubernetesDashboard `protobuf:"bytes,3,opt,name=kubernetes_dashboard,json=kubernetesDashboard" json:"kubernetes_dashboard,omitempty"` + KubernetesDashboard *KubernetesDashboard `protobuf:"bytes,3,opt,name=kubernetes_dashboard,json=kubernetesDashboard,proto3" json:"kubernetes_dashboard,omitempty"` // Configuration for NetworkPolicy. This only tracks whether the addon // is enabled or not on the Master, it does not track whether network policy // is enabled for the nodes. - NetworkPolicyConfig *NetworkPolicyConfig `protobuf:"bytes,4,opt,name=network_policy_config,json=networkPolicyConfig" json:"network_policy_config,omitempty"` + NetworkPolicyConfig *NetworkPolicyConfig `protobuf:"bytes,4,opt,name=network_policy_config,json=networkPolicyConfig,proto3" json:"network_policy_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -724,7 +724,7 @@ type HttpLoadBalancing struct { // Whether the HTTP Load Balancing controller is enabled in the cluster. // When enabled, it runs a small pod in the cluster that manages the load // balancers. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -768,7 +768,7 @@ type HorizontalPodAutoscaling struct { // Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. // When enabled, it ensures that a Heapster pod is running in the cluster, // which is also used by the Cloud Monitoring service. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -808,7 +808,7 @@ func (m *HorizontalPodAutoscaling) GetDisabled() bool { // Configuration for the Kubernetes Dashboard. type KubernetesDashboard struct { // Whether the Kubernetes Dashboard is enabled for this cluster. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -850,7 +850,7 @@ func (m *KubernetesDashboard) GetDisabled() bool { // is enabled for the nodes. type NetworkPolicyConfig struct { // Whether NetworkPolicy is enabled for this cluster. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -894,10 +894,10 @@ func (m *NetworkPolicyConfig) GetDisabled() bool { // Google Compute Engine Public IPs and Google Prod IPs. type MasterAuthorizedNetworksConfig struct { // Whether or not master authorized networks is enabled. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` // cidr_blocks define up to 10 external networks that could access // Kubernetes master through HTTPS. - CidrBlocks []*MasterAuthorizedNetworksConfig_CidrBlock `protobuf:"bytes,2,rep,name=cidr_blocks,json=cidrBlocks" json:"cidr_blocks,omitempty"` + CidrBlocks []*MasterAuthorizedNetworksConfig_CidrBlock `protobuf:"bytes,2,rep,name=cidr_blocks,json=cidrBlocks,proto3" json:"cidr_blocks,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -944,9 +944,9 @@ func (m *MasterAuthorizedNetworksConfig) GetCidrBlocks() []*MasterAuthorizedNetw // CidrBlock contains an optional name and one CIDR block. type MasterAuthorizedNetworksConfig_CidrBlock struct { // display_name is an optional field for users to identify CIDR blocks. - DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // cidr_block must be specified in CIDR notation. - CidrBlock string `protobuf:"bytes,2,opt,name=cidr_block,json=cidrBlock" json:"cidr_block,omitempty"` + CidrBlock string `protobuf:"bytes,2,opt,name=cidr_block,json=cidrBlock,proto3" json:"cidr_block,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -999,7 +999,7 @@ type LegacyAbac struct { // identities in the system, including service accounts, nodes, and // controllers, will have statically granted permissions beyond those // provided by the RBAC configuration or IAM. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1040,9 +1040,9 @@ func (m *LegacyAbac) GetEnabled() bool { // https://kubernetes.io/docs/concepts/services-networking/networkpolicies/ type NetworkPolicy struct { // The selected network policy provider. - Provider NetworkPolicy_Provider `protobuf:"varint,1,opt,name=provider,enum=google.container.v1.NetworkPolicy_Provider" json:"provider,omitempty"` + Provider NetworkPolicy_Provider `protobuf:"varint,1,opt,name=provider,proto3,enum=google.container.v1.NetworkPolicy_Provider" json:"provider,omitempty"` // Whether network policy is enabled on the cluster. - Enabled bool `protobuf:"varint,2,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,2,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1089,21 +1089,21 @@ func (m *NetworkPolicy) GetEnabled() bool { // Configuration for controlling how IPs are allocated in the cluster. type IPAllocationPolicy struct { // Whether alias IPs will be used for pod IPs in the cluster. - UseIpAliases bool `protobuf:"varint,1,opt,name=use_ip_aliases,json=useIpAliases" json:"use_ip_aliases,omitempty"` + UseIpAliases bool `protobuf:"varint,1,opt,name=use_ip_aliases,json=useIpAliases,proto3" json:"use_ip_aliases,omitempty"` // Whether a new subnetwork will be created automatically for the cluster. // // This field is only applicable when `use_ip_aliases` is true. - CreateSubnetwork bool `protobuf:"varint,2,opt,name=create_subnetwork,json=createSubnetwork" json:"create_subnetwork,omitempty"` + CreateSubnetwork bool `protobuf:"varint,2,opt,name=create_subnetwork,json=createSubnetwork,proto3" json:"create_subnetwork,omitempty"` // A custom subnetwork name to be used if `create_subnetwork` is true. If // this field is empty, then an automatic name will be chosen for the new // subnetwork. - SubnetworkName string `protobuf:"bytes,3,opt,name=subnetwork_name,json=subnetworkName" json:"subnetwork_name,omitempty"` + SubnetworkName string `protobuf:"bytes,3,opt,name=subnetwork_name,json=subnetworkName,proto3" json:"subnetwork_name,omitempty"` // This field is deprecated, use cluster_ipv4_cidr_block. - ClusterIpv4Cidr string `protobuf:"bytes,4,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr" json:"cluster_ipv4_cidr,omitempty"` + ClusterIpv4Cidr string `protobuf:"bytes,4,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr,proto3" json:"cluster_ipv4_cidr,omitempty"` // This field is deprecated, use node_ipv4_cidr_block. - NodeIpv4Cidr string `protobuf:"bytes,5,opt,name=node_ipv4_cidr,json=nodeIpv4Cidr" json:"node_ipv4_cidr,omitempty"` + NodeIpv4Cidr string `protobuf:"bytes,5,opt,name=node_ipv4_cidr,json=nodeIpv4Cidr,proto3" json:"node_ipv4_cidr,omitempty"` // This field is deprecated, use services_ipv4_cidr_block. - ServicesIpv4Cidr string `protobuf:"bytes,6,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr" json:"services_ipv4_cidr,omitempty"` + ServicesIpv4Cidr string `protobuf:"bytes,6,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr,proto3" json:"services_ipv4_cidr,omitempty"` // The name of the secondary range to be used for the cluster CIDR // block. The secondary range will be used for pod IP // addresses. This must be an existing secondary range associated @@ -1111,7 +1111,7 @@ type IPAllocationPolicy struct { // // This field is only applicable with use_ip_aliases is true and // create_subnetwork is false. - ClusterSecondaryRangeName string `protobuf:"bytes,7,opt,name=cluster_secondary_range_name,json=clusterSecondaryRangeName" json:"cluster_secondary_range_name,omitempty"` + ClusterSecondaryRangeName string `protobuf:"bytes,7,opt,name=cluster_secondary_range_name,json=clusterSecondaryRangeName,proto3" json:"cluster_secondary_range_name,omitempty"` // The name of the secondary range to be used as for the services // CIDR block. The secondary range will be used for service // ClusterIPs. This must be an existing secondary range associated @@ -1119,7 +1119,7 @@ type IPAllocationPolicy struct { // // This field is only applicable with use_ip_aliases is true and // create_subnetwork is false. - ServicesSecondaryRangeName string `protobuf:"bytes,8,opt,name=services_secondary_range_name,json=servicesSecondaryRangeName" json:"services_secondary_range_name,omitempty"` + ServicesSecondaryRangeName string `protobuf:"bytes,8,opt,name=services_secondary_range_name,json=servicesSecondaryRangeName,proto3" json:"services_secondary_range_name,omitempty"` // The IP address range for the cluster pod IPs. If this field is set, then // `cluster.cluster_ipv4_cidr` must be left blank. // @@ -1135,7 +1135,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - ClusterIpv4CidrBlock string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr_block,json=clusterIpv4CidrBlock" json:"cluster_ipv4_cidr_block,omitempty"` + ClusterIpv4CidrBlock string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr_block,json=clusterIpv4CidrBlock,proto3" json:"cluster_ipv4_cidr_block,omitempty"` // The IP address range of the instance IPs in this cluster. // // This is applicable only if `create_subnetwork` is true. @@ -1150,7 +1150,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - NodeIpv4CidrBlock string `protobuf:"bytes,10,opt,name=node_ipv4_cidr_block,json=nodeIpv4CidrBlock" json:"node_ipv4_cidr_block,omitempty"` + NodeIpv4CidrBlock string `protobuf:"bytes,10,opt,name=node_ipv4_cidr_block,json=nodeIpv4CidrBlock,proto3" json:"node_ipv4_cidr_block,omitempty"` // The IP address range of the services IPs in this cluster. If blank, a range // will be automatically chosen with the default size. // @@ -1166,7 +1166,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - ServicesIpv4CidrBlock string `protobuf:"bytes,11,opt,name=services_ipv4_cidr_block,json=servicesIpv4CidrBlock" json:"services_ipv4_cidr_block,omitempty"` + ServicesIpv4CidrBlock string `protobuf:"bytes,11,opt,name=services_ipv4_cidr_block,json=servicesIpv4CidrBlock,proto3" json:"services_ipv4_cidr_block,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1281,9 +1281,9 @@ type Cluster struct { // * Lowercase letters, numbers, and hyphens only. // * Must start with a letter. // * Must end with a number or a letter. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // An optional description of this cluster. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // The number of nodes to create in this cluster. You must ensure that your // Compute Engine resource quota // is sufficient for this number of instances. You must also have available @@ -1292,7 +1292,7 @@ type Cluster struct { // "node_pool" object, since this configuration (along with the // "node_config") will be used to create a "NodePool" object with an // auto-generated name. Do not use this and a node_pool at the same time. - InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount" json:"initial_node_count,omitempty"` + InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount,proto3" json:"initial_node_count,omitempty"` // Parameters used in creating the cluster's nodes. // See `nodeConfig` for the description of its properties. // For requests, this field should only be used in lieu of a @@ -1303,119 +1303,119 @@ type Cluster struct { // the first node pool. // // If unspecified, the defaults are used. - NodeConfig *NodeConfig `protobuf:"bytes,4,opt,name=node_config,json=nodeConfig" json:"node_config,omitempty"` + NodeConfig *NodeConfig `protobuf:"bytes,4,opt,name=node_config,json=nodeConfig,proto3" json:"node_config,omitempty"` // The authentication information for accessing the master endpoint. - MasterAuth *MasterAuth `protobuf:"bytes,5,opt,name=master_auth,json=masterAuth" json:"master_auth,omitempty"` + MasterAuth *MasterAuth `protobuf:"bytes,5,opt,name=master_auth,json=masterAuth,proto3" json:"master_auth,omitempty"` // The logging service the cluster should use to write logs. // Currently available options: // // * `logging.googleapis.com` - the Google Cloud Logging service. // * `none` - no logs will be exported from the cluster. // * if left as an empty string,`logging.googleapis.com` will be used. - LoggingService string `protobuf:"bytes,6,opt,name=logging_service,json=loggingService" json:"logging_service,omitempty"` + LoggingService string `protobuf:"bytes,6,opt,name=logging_service,json=loggingService,proto3" json:"logging_service,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * `monitoring.googleapis.com` - the Google Cloud Monitoring service. // * `none` - no metrics will be exported from the cluster. // * if left as an empty string, `monitoring.googleapis.com` will be used. - MonitoringService string `protobuf:"bytes,7,opt,name=monitoring_service,json=monitoringService" json:"monitoring_service,omitempty"` + MonitoringService string `protobuf:"bytes,7,opt,name=monitoring_service,json=monitoringService,proto3" json:"monitoring_service,omitempty"` // The name of the Google Compute Engine // [network](/compute/docs/networks-and-firewalls#networks) to which the // cluster is connected. If left unspecified, the `default` network // will be used. - Network string `protobuf:"bytes,8,opt,name=network" json:"network,omitempty"` + Network string `protobuf:"bytes,8,opt,name=network,proto3" json:"network,omitempty"` // The IP address range of the container pods in this cluster, in // [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) // notation (e.g. `10.96.0.0/14`). Leave blank to have // one automatically chosen or specify a `/14` block in `10.0.0.0/8`. - ClusterIpv4Cidr string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr" json:"cluster_ipv4_cidr,omitempty"` + ClusterIpv4Cidr string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr,proto3" json:"cluster_ipv4_cidr,omitempty"` // Configurations for the various addons available to run in the cluster. - AddonsConfig *AddonsConfig `protobuf:"bytes,10,opt,name=addons_config,json=addonsConfig" json:"addons_config,omitempty"` + AddonsConfig *AddonsConfig `protobuf:"bytes,10,opt,name=addons_config,json=addonsConfig,proto3" json:"addons_config,omitempty"` // The name of the Google Compute Engine // [subnetwork](/compute/docs/subnetworks) to which the // cluster is connected. - Subnetwork string `protobuf:"bytes,11,opt,name=subnetwork" json:"subnetwork,omitempty"` + Subnetwork string `protobuf:"bytes,11,opt,name=subnetwork,proto3" json:"subnetwork,omitempty"` // The node pools associated with this cluster. // This field should not be set if "node_config" or "initial_node_count" are // specified. - NodePools []*NodePool `protobuf:"bytes,12,rep,name=node_pools,json=nodePools" json:"node_pools,omitempty"` + NodePools []*NodePool `protobuf:"bytes,12,rep,name=node_pools,json=nodePools,proto3" json:"node_pools,omitempty"` // The list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. - Locations []string `protobuf:"bytes,13,rep,name=locations" json:"locations,omitempty"` + Locations []string `protobuf:"bytes,13,rep,name=locations,proto3" json:"locations,omitempty"` // Kubernetes alpha features are enabled on this cluster. This includes alpha // API groups (e.g. v1alpha1) and features that may not be production ready in // the kubernetes version of the master and nodes. // The cluster has no SLA for uptime and master/node upgrades are disabled. // Alpha enabled clusters are automatically deleted thirty days after // creation. - EnableKubernetesAlpha bool `protobuf:"varint,14,opt,name=enable_kubernetes_alpha,json=enableKubernetesAlpha" json:"enable_kubernetes_alpha,omitempty"` + EnableKubernetesAlpha bool `protobuf:"varint,14,opt,name=enable_kubernetes_alpha,json=enableKubernetesAlpha,proto3" json:"enable_kubernetes_alpha,omitempty"` // The resource labels for the cluster to use to annotate any related // Google Compute Engine resources. - ResourceLabels map[string]string `protobuf:"bytes,15,rep,name=resource_labels,json=resourceLabels" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ResourceLabels map[string]string `protobuf:"bytes,15,rep,name=resource_labels,json=resourceLabels,proto3" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The fingerprint of the set of labels for this cluster. - LabelFingerprint string `protobuf:"bytes,16,opt,name=label_fingerprint,json=labelFingerprint" json:"label_fingerprint,omitempty"` + LabelFingerprint string `protobuf:"bytes,16,opt,name=label_fingerprint,json=labelFingerprint,proto3" json:"label_fingerprint,omitempty"` // Configuration for the legacy ABAC authorization mode. - LegacyAbac *LegacyAbac `protobuf:"bytes,18,opt,name=legacy_abac,json=legacyAbac" json:"legacy_abac,omitempty"` + LegacyAbac *LegacyAbac `protobuf:"bytes,18,opt,name=legacy_abac,json=legacyAbac,proto3" json:"legacy_abac,omitempty"` // Configuration options for the NetworkPolicy feature. - NetworkPolicy *NetworkPolicy `protobuf:"bytes,19,opt,name=network_policy,json=networkPolicy" json:"network_policy,omitempty"` + NetworkPolicy *NetworkPolicy `protobuf:"bytes,19,opt,name=network_policy,json=networkPolicy,proto3" json:"network_policy,omitempty"` // Configuration for cluster IP allocation. - IpAllocationPolicy *IPAllocationPolicy `protobuf:"bytes,20,opt,name=ip_allocation_policy,json=ipAllocationPolicy" json:"ip_allocation_policy,omitempty"` + IpAllocationPolicy *IPAllocationPolicy `protobuf:"bytes,20,opt,name=ip_allocation_policy,json=ipAllocationPolicy,proto3" json:"ip_allocation_policy,omitempty"` // Master authorized networks is a Beta feature. // The configuration options for master authorized networks feature. - MasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,22,opt,name=master_authorized_networks_config,json=masterAuthorizedNetworksConfig" json:"master_authorized_networks_config,omitempty"` + MasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,22,opt,name=master_authorized_networks_config,json=masterAuthorizedNetworksConfig,proto3" json:"master_authorized_networks_config,omitempty"` // Configure the maintenance policy for this cluster. - MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,23,opt,name=maintenance_policy,json=maintenancePolicy" json:"maintenance_policy,omitempty"` + MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,23,opt,name=maintenance_policy,json=maintenancePolicy,proto3" json:"maintenance_policy,omitempty"` // [Output only] Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,101,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,101,opt,name=zone,proto3" json:"zone,omitempty"` // [Output only] The IP address of this cluster's master endpoint. // The endpoint can be accessed from the internet at // `https://username:password@endpoint/`. // // See the `masterAuth` property of this resource for username and // password information. - Endpoint string `protobuf:"bytes,102,opt,name=endpoint" json:"endpoint,omitempty"` + Endpoint string `protobuf:"bytes,102,opt,name=endpoint,proto3" json:"endpoint,omitempty"` // The initial Kubernetes version for this cluster. Valid versions are those // found in validMasterVersions returned by getServerConfig. The version can // be upgraded over time; such upgrades are reflected in // currentMasterVersion and currentNodeVersion. - InitialClusterVersion string `protobuf:"bytes,103,opt,name=initial_cluster_version,json=initialClusterVersion" json:"initial_cluster_version,omitempty"` + InitialClusterVersion string `protobuf:"bytes,103,opt,name=initial_cluster_version,json=initialClusterVersion,proto3" json:"initial_cluster_version,omitempty"` // [Output only] The current software version of the master endpoint. - CurrentMasterVersion string `protobuf:"bytes,104,opt,name=current_master_version,json=currentMasterVersion" json:"current_master_version,omitempty"` + CurrentMasterVersion string `protobuf:"bytes,104,opt,name=current_master_version,json=currentMasterVersion,proto3" json:"current_master_version,omitempty"` // [Output only] The current version of the node software components. // If they are currently at multiple versions because they're in the process // of being upgraded, this reflects the minimum version of all nodes. - CurrentNodeVersion string `protobuf:"bytes,105,opt,name=current_node_version,json=currentNodeVersion" json:"current_node_version,omitempty"` + CurrentNodeVersion string `protobuf:"bytes,105,opt,name=current_node_version,json=currentNodeVersion,proto3" json:"current_node_version,omitempty"` // [Output only] The time the cluster was created, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - CreateTime string `protobuf:"bytes,106,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime string `protobuf:"bytes,106,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // [Output only] The current status of this cluster. - Status Cluster_Status `protobuf:"varint,107,opt,name=status,enum=google.container.v1.Cluster_Status" json:"status,omitempty"` + Status Cluster_Status `protobuf:"varint,107,opt,name=status,proto3,enum=google.container.v1.Cluster_Status" json:"status,omitempty"` // [Output only] Additional information about the current status of this // cluster, if available. - StatusMessage string `protobuf:"bytes,108,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,108,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // [Output only] The size of the address space on each node for hosting // containers. This is provisioned from within the `container_ipv4_cidr` // range. - NodeIpv4CidrSize int32 `protobuf:"varint,109,opt,name=node_ipv4_cidr_size,json=nodeIpv4CidrSize" json:"node_ipv4_cidr_size,omitempty"` + NodeIpv4CidrSize int32 `protobuf:"varint,109,opt,name=node_ipv4_cidr_size,json=nodeIpv4CidrSize,proto3" json:"node_ipv4_cidr_size,omitempty"` // [Output only] The IP address range of the Kubernetes services in // this cluster, in // [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) // notation (e.g. `1.2.3.4/29`). Service addresses are // typically put in the last `/16` from the container CIDR. - ServicesIpv4Cidr string `protobuf:"bytes,110,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr" json:"services_ipv4_cidr,omitempty"` + ServicesIpv4Cidr string `protobuf:"bytes,110,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr,proto3" json:"services_ipv4_cidr,omitempty"` // Deprecated. Use node_pools.instance_group_urls. - InstanceGroupUrls []string `protobuf:"bytes,111,rep,name=instance_group_urls,json=instanceGroupUrls" json:"instance_group_urls,omitempty"` + InstanceGroupUrls []string `protobuf:"bytes,111,rep,name=instance_group_urls,json=instanceGroupUrls,proto3" json:"instance_group_urls,omitempty"` // [Output only] The number of nodes currently in the cluster. - CurrentNodeCount int32 `protobuf:"varint,112,opt,name=current_node_count,json=currentNodeCount" json:"current_node_count,omitempty"` + CurrentNodeCount int32 `protobuf:"varint,112,opt,name=current_node_count,json=currentNodeCount,proto3" json:"current_node_count,omitempty"` // [Output only] The time the cluster will be automatically // deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - ExpireTime string `protobuf:"bytes,113,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime string `protobuf:"bytes,113,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1697,28 +1697,28 @@ type ClusterUpdate struct { // The Kubernetes version to change the nodes to (typically an // upgrade). Use `-` to upgrade to the latest version supported by // the server. - DesiredNodeVersion string `protobuf:"bytes,4,opt,name=desired_node_version,json=desiredNodeVersion" json:"desired_node_version,omitempty"` + DesiredNodeVersion string `protobuf:"bytes,4,opt,name=desired_node_version,json=desiredNodeVersion,proto3" json:"desired_node_version,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * "monitoring.googleapis.com" - the Google Cloud Monitoring service // * "none" - no metrics will be exported from the cluster - DesiredMonitoringService string `protobuf:"bytes,5,opt,name=desired_monitoring_service,json=desiredMonitoringService" json:"desired_monitoring_service,omitempty"` + DesiredMonitoringService string `protobuf:"bytes,5,opt,name=desired_monitoring_service,json=desiredMonitoringService,proto3" json:"desired_monitoring_service,omitempty"` // Configurations for the various addons available to run in the cluster. - DesiredAddonsConfig *AddonsConfig `protobuf:"bytes,6,opt,name=desired_addons_config,json=desiredAddonsConfig" json:"desired_addons_config,omitempty"` + DesiredAddonsConfig *AddonsConfig `protobuf:"bytes,6,opt,name=desired_addons_config,json=desiredAddonsConfig,proto3" json:"desired_addons_config,omitempty"` // The node pool to be upgraded. This field is mandatory if // "desired_node_version", "desired_image_family" or // "desired_node_pool_autoscaling" is specified and there is more than one // node pool on the cluster. - DesiredNodePoolId string `protobuf:"bytes,7,opt,name=desired_node_pool_id,json=desiredNodePoolId" json:"desired_node_pool_id,omitempty"` + DesiredNodePoolId string `protobuf:"bytes,7,opt,name=desired_node_pool_id,json=desiredNodePoolId,proto3" json:"desired_node_pool_id,omitempty"` // The desired image type for the node pool. // NOTE: Set the "desired_node_pool" field as well. - DesiredImageType string `protobuf:"bytes,8,opt,name=desired_image_type,json=desiredImageType" json:"desired_image_type,omitempty"` + DesiredImageType string `protobuf:"bytes,8,opt,name=desired_image_type,json=desiredImageType,proto3" json:"desired_image_type,omitempty"` // Autoscaler configuration for the node pool specified in // desired_node_pool_id. If there is only one pool in the // cluster and desired_node_pool_id is not provided then // the change applies to that single node pool. - DesiredNodePoolAutoscaling *NodePoolAutoscaling `protobuf:"bytes,9,opt,name=desired_node_pool_autoscaling,json=desiredNodePoolAutoscaling" json:"desired_node_pool_autoscaling,omitempty"` + DesiredNodePoolAutoscaling *NodePoolAutoscaling `protobuf:"bytes,9,opt,name=desired_node_pool_autoscaling,json=desiredNodePoolAutoscaling,proto3" json:"desired_node_pool_autoscaling,omitempty"` // The desired list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. Changing the locations a cluster is in will result @@ -1726,14 +1726,14 @@ type ClusterUpdate struct { // whether locations are being added or removed. // // This list must always include the cluster's primary zone. - DesiredLocations []string `protobuf:"bytes,10,rep,name=desired_locations,json=desiredLocations" json:"desired_locations,omitempty"` + DesiredLocations []string `protobuf:"bytes,10,rep,name=desired_locations,json=desiredLocations,proto3" json:"desired_locations,omitempty"` // Master authorized networks is a Beta feature. // The desired configuration options for master authorized networks feature. - DesiredMasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,12,opt,name=desired_master_authorized_networks_config,json=desiredMasterAuthorizedNetworksConfig" json:"desired_master_authorized_networks_config,omitempty"` + DesiredMasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,12,opt,name=desired_master_authorized_networks_config,json=desiredMasterAuthorizedNetworksConfig,proto3" json:"desired_master_authorized_networks_config,omitempty"` // The Kubernetes version to change the master to. The only valid value is the // latest supported version. Use "-" to have the server automatically select // the latest version. - DesiredMasterVersion string `protobuf:"bytes,100,opt,name=desired_master_version,json=desiredMasterVersion" json:"desired_master_version,omitempty"` + DesiredMasterVersion string `protobuf:"bytes,100,opt,name=desired_master_version,json=desiredMasterVersion,proto3" json:"desired_master_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1830,29 +1830,29 @@ func (m *ClusterUpdate) GetDesiredMasterVersion() string { // happening on the cluster. All fields are output only. type Operation struct { // The server-assigned ID for the operation. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the operation // is taking place. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The operation type. - OperationType Operation_Type `protobuf:"varint,3,opt,name=operation_type,json=operationType,enum=google.container.v1.Operation_Type" json:"operation_type,omitempty"` + OperationType Operation_Type `protobuf:"varint,3,opt,name=operation_type,json=operationType,proto3,enum=google.container.v1.Operation_Type" json:"operation_type,omitempty"` // The current status of the operation. - Status Operation_Status `protobuf:"varint,4,opt,name=status,enum=google.container.v1.Operation_Status" json:"status,omitempty"` + Status Operation_Status `protobuf:"varint,4,opt,name=status,proto3,enum=google.container.v1.Operation_Status" json:"status,omitempty"` // Detailed operation progress, if available. - Detail string `protobuf:"bytes,8,opt,name=detail" json:"detail,omitempty"` + Detail string `protobuf:"bytes,8,opt,name=detail,proto3" json:"detail,omitempty"` // If an error has occurred, a textual description of the error. - StatusMessage string `protobuf:"bytes,5,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,5,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,6,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,6,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // Server-defined URL for the target of the operation. - TargetLink string `protobuf:"bytes,7,opt,name=target_link,json=targetLink" json:"target_link,omitempty"` + TargetLink string `protobuf:"bytes,7,opt,name=target_link,json=targetLink,proto3" json:"target_link,omitempty"` // [Output only] The time the operation started, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - StartTime string `protobuf:"bytes,10,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime string `protobuf:"bytes,10,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // [Output only] The time the operation completed, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - EndTime string `protobuf:"bytes,11,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime string `protobuf:"bytes,11,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1956,14 +1956,14 @@ func (m *Operation) GetEndTime() string { type CreateClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // A [cluster // resource](/container-engine/reference/rest/v1/projects.zones.clusters) - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2018,13 +2018,13 @@ func (m *CreateClusterRequest) GetCluster() *Cluster { type GetClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to retrieve. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2079,15 +2079,15 @@ func (m *GetClusterRequest) GetClusterId() string { type UpdateClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // A description of the update. - Update *ClusterUpdate `protobuf:"bytes,4,opt,name=update" json:"update,omitempty"` + Update *ClusterUpdate `protobuf:"bytes,4,opt,name=update,proto3" json:"update,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2149,21 +2149,21 @@ func (m *UpdateClusterRequest) GetUpdate() *ClusterUpdate { type UpdateNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to upgrade. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The Kubernetes version to change the nodes to (typically an // upgrade). Use `-` to upgrade to the latest version supported by // the server. - NodeVersion string `protobuf:"bytes,5,opt,name=node_version,json=nodeVersion" json:"node_version,omitempty"` + NodeVersion string `protobuf:"bytes,5,opt,name=node_version,json=nodeVersion,proto3" json:"node_version,omitempty"` // The desired image type for the node pool. - ImageType string `protobuf:"bytes,6,opt,name=image_type,json=imageType" json:"image_type,omitempty"` + ImageType string `protobuf:"bytes,6,opt,name=image_type,json=imageType,proto3" json:"image_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2239,17 +2239,17 @@ func (m *UpdateNodePoolRequest) GetImageType() string { type SetNodePoolAutoscalingRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to upgrade. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // Autoscaling configuration for the node pool. - Autoscaling *NodePoolAutoscaling `protobuf:"bytes,5,opt,name=autoscaling" json:"autoscaling,omitempty"` + Autoscaling *NodePoolAutoscaling `protobuf:"bytes,5,opt,name=autoscaling,proto3" json:"autoscaling,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2318,19 +2318,19 @@ func (m *SetNodePoolAutoscalingRequest) GetAutoscaling() *NodePoolAutoscaling { type SetLoggingServiceRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The logging service the cluster should use to write metrics. // Currently available options: // // * "logging.googleapis.com" - the Google Cloud Logging service // * "none" - no metrics will be exported from the cluster - LoggingService string `protobuf:"bytes,4,opt,name=logging_service,json=loggingService" json:"logging_service,omitempty"` + LoggingService string `protobuf:"bytes,4,opt,name=logging_service,json=loggingService,proto3" json:"logging_service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2392,19 +2392,19 @@ func (m *SetLoggingServiceRequest) GetLoggingService() string { type SetMonitoringServiceRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * "monitoring.googleapis.com" - the Google Cloud Monitoring service // * "none" - no metrics will be exported from the cluster - MonitoringService string `protobuf:"bytes,4,opt,name=monitoring_service,json=monitoringService" json:"monitoring_service,omitempty"` + MonitoringService string `protobuf:"bytes,4,opt,name=monitoring_service,json=monitoringService,proto3" json:"monitoring_service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2466,16 +2466,16 @@ func (m *SetMonitoringServiceRequest) GetMonitoringService() string { type SetAddonsConfigRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The desired configurations for the various addons available to run in the // cluster. - AddonsConfig *AddonsConfig `protobuf:"bytes,4,opt,name=addons_config,json=addonsConfig" json:"addons_config,omitempty"` + AddonsConfig *AddonsConfig `protobuf:"bytes,4,opt,name=addons_config,json=addonsConfig,proto3" json:"addons_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2537,13 +2537,13 @@ func (m *SetAddonsConfigRequest) GetAddonsConfig() *AddonsConfig { type SetLocationsRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The desired list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. Changing the locations a cluster is in will result @@ -2551,7 +2551,7 @@ type SetLocationsRequest struct { // whether locations are being added or removed. // // This list must always include the cluster's primary zone. - Locations []string `protobuf:"bytes,4,rep,name=locations" json:"locations,omitempty"` + Locations []string `protobuf:"bytes,4,rep,name=locations,proto3" json:"locations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2613,17 +2613,17 @@ func (m *SetLocationsRequest) GetLocations() []string { type UpdateMasterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The Kubernetes version to change the master to. The only valid value is the // latest supported version. Use "-" to have the server automatically select // the latest version. - MasterVersion string `protobuf:"bytes,4,opt,name=master_version,json=masterVersion" json:"master_version,omitempty"` + MasterVersion string `protobuf:"bytes,4,opt,name=master_version,json=masterVersion,proto3" json:"master_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2685,17 +2685,17 @@ func (m *UpdateMasterRequest) GetMasterVersion() string { type SetMasterAuthRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The exact form of action to be taken on the master auth. - Action SetMasterAuthRequest_Action `protobuf:"varint,4,opt,name=action,enum=google.container.v1.SetMasterAuthRequest_Action" json:"action,omitempty"` + Action SetMasterAuthRequest_Action `protobuf:"varint,4,opt,name=action,proto3,enum=google.container.v1.SetMasterAuthRequest_Action" json:"action,omitempty"` // A description of the update. - Update *MasterAuth `protobuf:"bytes,5,opt,name=update" json:"update,omitempty"` + Update *MasterAuth `protobuf:"bytes,5,opt,name=update,proto3" json:"update,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2764,13 +2764,13 @@ func (m *SetMasterAuthRequest) GetUpdate() *MasterAuth { type DeleteClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to delete. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2825,11 +2825,11 @@ func (m *DeleteClusterRequest) GetClusterId() string { type ListClustersRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides, or "-" for all zones. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2877,10 +2877,10 @@ func (m *ListClustersRequest) GetZone() string { type ListClustersResponse struct { // A list of clusters in the project in the specified zone, or // across all ones. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` // If any zones are listed here, the list of clusters returned // may be missing those zones. - MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones" json:"missing_zones,omitempty"` + MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones,proto3" json:"missing_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2928,13 +2928,13 @@ func (m *ListClustersResponse) GetMissingZones() []string { type GetOperationRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The server-assigned `name` of the operation. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2989,10 +2989,10 @@ func (m *GetOperationRequest) GetOperationId() string { type ListOperationsRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine [zone](/compute/docs/zones#available) // to return operations for, or `-` for all zones. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3040,12 +3040,12 @@ func (m *ListOperationsRequest) GetZone() string { type CancelOperationRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the operation resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The server-assigned `name` of the operation. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3099,10 +3099,10 @@ func (m *CancelOperationRequest) GetOperationId() string { // ListOperationsResponse is the result of ListOperationsRequest. type ListOperationsResponse struct { // A list of operations in the project in the specified zone. - Operations []*Operation `protobuf:"bytes,1,rep,name=operations" json:"operations,omitempty"` + Operations []*Operation `protobuf:"bytes,1,rep,name=operations,proto3" json:"operations,omitempty"` // If any zones are listed here, the list of operations returned // may be missing the operations from those zones. - MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones" json:"missing_zones,omitempty"` + MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones,proto3" json:"missing_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3150,10 +3150,10 @@ func (m *ListOperationsResponse) GetMissingZones() []string { type GetServerConfigRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine [zone](/compute/docs/zones#available) // to return operations for. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3200,15 +3200,15 @@ func (m *GetServerConfigRequest) GetZone() string { // Container Engine service configuration. type ServerConfig struct { // Version of Kubernetes the service deploys by default. - DefaultClusterVersion string `protobuf:"bytes,1,opt,name=default_cluster_version,json=defaultClusterVersion" json:"default_cluster_version,omitempty"` + DefaultClusterVersion string `protobuf:"bytes,1,opt,name=default_cluster_version,json=defaultClusterVersion,proto3" json:"default_cluster_version,omitempty"` // List of valid node upgrade target versions. - ValidNodeVersions []string `protobuf:"bytes,3,rep,name=valid_node_versions,json=validNodeVersions" json:"valid_node_versions,omitempty"` + ValidNodeVersions []string `protobuf:"bytes,3,rep,name=valid_node_versions,json=validNodeVersions,proto3" json:"valid_node_versions,omitempty"` // Default image type. - DefaultImageType string `protobuf:"bytes,4,opt,name=default_image_type,json=defaultImageType" json:"default_image_type,omitempty"` + DefaultImageType string `protobuf:"bytes,4,opt,name=default_image_type,json=defaultImageType,proto3" json:"default_image_type,omitempty"` // List of valid image types. - ValidImageTypes []string `protobuf:"bytes,5,rep,name=valid_image_types,json=validImageTypes" json:"valid_image_types,omitempty"` + ValidImageTypes []string `protobuf:"bytes,5,rep,name=valid_image_types,json=validImageTypes,proto3" json:"valid_image_types,omitempty"` // List of valid master versions. - ValidMasterVersions []string `protobuf:"bytes,6,rep,name=valid_master_versions,json=validMasterVersions" json:"valid_master_versions,omitempty"` + ValidMasterVersions []string `protobuf:"bytes,6,rep,name=valid_master_versions,json=validMasterVersions,proto3" json:"valid_master_versions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3277,15 +3277,15 @@ func (m *ServerConfig) GetValidMasterVersions() []string { type CreateNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The node pool to create. - NodePool *NodePool `protobuf:"bytes,4,opt,name=node_pool,json=nodePool" json:"node_pool,omitempty"` + NodePool *NodePool `protobuf:"bytes,4,opt,name=node_pool,json=nodePool,proto3" json:"node_pool,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3347,15 +3347,15 @@ func (m *CreateNodePoolRequest) GetNodePool() *NodePool { type DeleteNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to delete. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3417,13 +3417,13 @@ func (m *DeleteNodePoolRequest) GetNodePoolId() string { type ListNodePoolsRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3478,15 +3478,15 @@ func (m *ListNodePoolsRequest) GetClusterId() string { type GetNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3552,32 +3552,32 @@ func (m *GetNodePoolRequest) GetNodePoolId() string { // the workload. type NodePool struct { // The name of the node pool. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The node configuration of the pool. - Config *NodeConfig `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"` + Config *NodeConfig `protobuf:"bytes,2,opt,name=config,proto3" json:"config,omitempty"` // The initial node count for the pool. You must ensure that your // Compute Engine resource quota // is sufficient for this number of instances. You must also have available // firewall and routes quota. - InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount" json:"initial_node_count,omitempty"` + InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount,proto3" json:"initial_node_count,omitempty"` // [Output only] Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // The version of the Kubernetes of this node. - Version string `protobuf:"bytes,101,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,101,opt,name=version,proto3" json:"version,omitempty"` // [Output only] The resource URLs of the [managed instance // groups](/compute/docs/instance-groups/creating-groups-of-managed-instances) // associated with this node pool. - InstanceGroupUrls []string `protobuf:"bytes,102,rep,name=instance_group_urls,json=instanceGroupUrls" json:"instance_group_urls,omitempty"` + InstanceGroupUrls []string `protobuf:"bytes,102,rep,name=instance_group_urls,json=instanceGroupUrls,proto3" json:"instance_group_urls,omitempty"` // [Output only] The status of the nodes in this pool instance. - Status NodePool_Status `protobuf:"varint,103,opt,name=status,enum=google.container.v1.NodePool_Status" json:"status,omitempty"` + Status NodePool_Status `protobuf:"varint,103,opt,name=status,proto3,enum=google.container.v1.NodePool_Status" json:"status,omitempty"` // [Output only] Additional information about the current status of this // node pool instance, if available. - StatusMessage string `protobuf:"bytes,104,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,104,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // Autoscaler configuration for this NodePool. Autoscaler is enabled // only if a valid configuration is present. - Autoscaling *NodePoolAutoscaling `protobuf:"bytes,4,opt,name=autoscaling" json:"autoscaling,omitempty"` + Autoscaling *NodePoolAutoscaling `protobuf:"bytes,4,opt,name=autoscaling,proto3" json:"autoscaling,omitempty"` // NodeManagement configuration for this NodePool. - Management *NodeManagement `protobuf:"bytes,5,opt,name=management" json:"management,omitempty"` + Management *NodeManagement `protobuf:"bytes,5,opt,name=management,proto3" json:"management,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3683,14 +3683,14 @@ type NodeManagement struct { // A flag that specifies whether node auto-upgrade is enabled for the node // pool. If enabled, node auto-upgrade helps keep the nodes in your node pool // up to date with the latest release version of Kubernetes. - AutoUpgrade bool `protobuf:"varint,1,opt,name=auto_upgrade,json=autoUpgrade" json:"auto_upgrade,omitempty"` + AutoUpgrade bool `protobuf:"varint,1,opt,name=auto_upgrade,json=autoUpgrade,proto3" json:"auto_upgrade,omitempty"` // A flag that specifies whether the node auto-repair is enabled for the node // pool. If enabled, the nodes in this node pool will be monitored and, if // they fail health checks too many times, an automatic repair action will be // triggered. - AutoRepair bool `protobuf:"varint,2,opt,name=auto_repair,json=autoRepair" json:"auto_repair,omitempty"` + AutoRepair bool `protobuf:"varint,2,opt,name=auto_repair,json=autoRepair,proto3" json:"auto_repair,omitempty"` // Specifies the Auto Upgrade knobs for the node pool. - UpgradeOptions *AutoUpgradeOptions `protobuf:"bytes,10,opt,name=upgrade_options,json=upgradeOptions" json:"upgrade_options,omitempty"` + UpgradeOptions *AutoUpgradeOptions `protobuf:"bytes,10,opt,name=upgrade_options,json=upgradeOptions,proto3" json:"upgrade_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3747,10 +3747,10 @@ type AutoUpgradeOptions struct { // [Output only] This field is set when upgrades are about to commence // with the approximate start time for the upgrades, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - AutoUpgradeStartTime string `protobuf:"bytes,1,opt,name=auto_upgrade_start_time,json=autoUpgradeStartTime" json:"auto_upgrade_start_time,omitempty"` + AutoUpgradeStartTime string `protobuf:"bytes,1,opt,name=auto_upgrade_start_time,json=autoUpgradeStartTime,proto3" json:"auto_upgrade_start_time,omitempty"` // [Output only] This field is set when upgrades are about to commence // with the description of the upgrade. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3797,7 +3797,7 @@ func (m *AutoUpgradeOptions) GetDescription() string { // MaintenancePolicy defines the maintenance policy to be used for the cluster. type MaintenancePolicy struct { // Specifies the maintenance window in which maintenance may be performed. - Window *MaintenanceWindow `protobuf:"bytes,1,opt,name=window" json:"window,omitempty"` + Window *MaintenanceWindow `protobuf:"bytes,1,opt,name=window,proto3" json:"window,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3873,7 +3873,7 @@ type isMaintenanceWindow_Policy interface { } type MaintenanceWindow_DailyMaintenanceWindow struct { - DailyMaintenanceWindow *DailyMaintenanceWindow `protobuf:"bytes,2,opt,name=daily_maintenance_window,json=dailyMaintenanceWindow,oneof"` + DailyMaintenanceWindow *DailyMaintenanceWindow `protobuf:"bytes,2,opt,name=daily_maintenance_window,json=dailyMaintenanceWindow,proto3,oneof"` } func (*MaintenanceWindow_DailyMaintenanceWindow) isMaintenanceWindow_Policy() {} @@ -3952,12 +3952,12 @@ type DailyMaintenanceWindow struct { // Time within the maintenance window to start the maintenance operations. // Time format should be in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) // format "HH:MM”, where HH : [00-23] and MM : [00-59] GMT. - StartTime string `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime string `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // [Output only] Duration of the time window, automatically chosen to be // smallest possible in the given scenario. // Duration will be in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) // format "PTnHnMnS". - Duration string `protobuf:"bytes,3,opt,name=duration" json:"duration,omitempty"` + Duration string `protobuf:"bytes,3,opt,name=duration,proto3" json:"duration,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4006,17 +4006,17 @@ func (m *DailyMaintenanceWindow) GetDuration() string { type SetNodePoolManagementRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to update. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // NodeManagement configuration for the node pool. - Management *NodeManagement `protobuf:"bytes,5,opt,name=management" json:"management,omitempty"` + Management *NodeManagement `protobuf:"bytes,5,opt,name=management,proto3" json:"management,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4086,17 +4086,17 @@ func (m *SetNodePoolManagementRequest) GetManagement() *NodeManagement { type SetNodePoolSizeRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to update. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The desired node count for the pool. - NodeCount int32 `protobuf:"varint,5,opt,name=node_count,json=nodeCount" json:"node_count,omitempty"` + NodeCount int32 `protobuf:"varint,5,opt,name=node_count,json=nodeCount,proto3" json:"node_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4167,15 +4167,15 @@ func (m *SetNodePoolSizeRequest) GetNodeCount() int32 { type RollbackNodePoolUpgradeRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to rollback. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to rollback. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4236,7 +4236,7 @@ func (m *RollbackNodePoolUpgradeRequest) GetNodePoolId() string { // ListNodePoolsResponse is the result of ListNodePoolsRequest. type ListNodePoolsResponse struct { // A list of node pools for a cluster. - NodePools []*NodePool `protobuf:"bytes,1,rep,name=node_pools,json=nodePools" json:"node_pools,omitempty"` + NodePools []*NodePool `protobuf:"bytes,1,rep,name=node_pools,json=nodePools,proto3" json:"node_pools,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4277,13 +4277,13 @@ func (m *ListNodePoolsResponse) GetNodePools() []*NodePool { // adjust the size of the node pool to the current cluster usage. type NodePoolAutoscaling struct { // Is autoscaling enabled for this node pool. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` // Minimum number of nodes in the NodePool. Must be >= 1 and <= // max_node_count. - MinNodeCount int32 `protobuf:"varint,2,opt,name=min_node_count,json=minNodeCount" json:"min_node_count,omitempty"` + MinNodeCount int32 `protobuf:"varint,2,opt,name=min_node_count,json=minNodeCount,proto3" json:"min_node_count,omitempty"` // Maximum number of nodes in the NodePool. Must be >= min_node_count. There // has to enough quota to scale up the cluster. - MaxNodeCount int32 `protobuf:"varint,3,opt,name=max_node_count,json=maxNodeCount" json:"max_node_count,omitempty"` + MaxNodeCount int32 `protobuf:"varint,3,opt,name=max_node_count,json=maxNodeCount,proto3" json:"max_node_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4340,22 +4340,22 @@ func (m *NodePoolAutoscaling) GetMaxNodeCount() int32 { type SetLabelsRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The labels to set for that cluster. - ResourceLabels map[string]string `protobuf:"bytes,4,rep,name=resource_labels,json=resourceLabels" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ResourceLabels map[string]string `protobuf:"bytes,4,rep,name=resource_labels,json=resourceLabels,proto3" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The fingerprint of the previous set of labels for this resource, // used to detect conflicts. The fingerprint is initially generated by // Container Engine and changes after every request to modify or update // labels. You must always provide an up-to-date fingerprint hash when // updating or changing labels. Make a get() request to the // resource to get the latest fingerprint. - LabelFingerprint string `protobuf:"bytes,5,opt,name=label_fingerprint,json=labelFingerprint" json:"label_fingerprint,omitempty"` + LabelFingerprint string `protobuf:"bytes,5,opt,name=label_fingerprint,json=labelFingerprint,proto3" json:"label_fingerprint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4425,15 +4425,15 @@ func (m *SetLabelsRequest) GetLabelFingerprint() string { type SetLegacyAbacRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Whether ABAC authorization will be enabled in the cluster. - Enabled bool `protobuf:"varint,4,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,4,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4496,13 +4496,13 @@ func (m *SetLegacyAbacRequest) GetEnabled() bool { type StartIPRotationRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4557,13 +4557,13 @@ func (m *StartIPRotationRequest) GetClusterId() string { type CompleteIPRotationRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4617,10 +4617,10 @@ func (m *CompleteIPRotationRequest) GetClusterId() string { // AcceleratorConfig represents a Hardware Accelerator request. type AcceleratorConfig struct { // The number of the accelerator cards exposed to an instance. - AcceleratorCount int64 `protobuf:"varint,1,opt,name=accelerator_count,json=acceleratorCount" json:"accelerator_count,omitempty"` + AcceleratorCount int64 `protobuf:"varint,1,opt,name=accelerator_count,json=acceleratorCount,proto3" json:"accelerator_count,omitempty"` // The accelerator type resource name. List of supported accelerators // [here](/compute/docs/gpus/#Introduction) - AcceleratorType string `protobuf:"bytes,2,opt,name=accelerator_type,json=acceleratorType" json:"accelerator_type,omitempty"` + AcceleratorType string `protobuf:"bytes,2,opt,name=accelerator_type,json=acceleratorType,proto3" json:"accelerator_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4668,15 +4668,15 @@ func (m *AcceleratorConfig) GetAcceleratorType() string { type SetNetworkPolicyRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Configuration options for the NetworkPolicy feature. - NetworkPolicy *NetworkPolicy `protobuf:"bytes,4,opt,name=network_policy,json=networkPolicy" json:"network_policy,omitempty"` + NetworkPolicy *NetworkPolicy `protobuf:"bytes,4,opt,name=network_policy,json=networkPolicy,proto3" json:"network_policy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4738,16 +4738,16 @@ func (m *SetNetworkPolicyRequest) GetNetworkPolicy() *NetworkPolicy { type SetMaintenancePolicyRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The maintenance policy to be set for the cluster. An empty field // clears the existing maintenance policy. - MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,4,opt,name=maintenance_policy,json=maintenancePolicy" json:"maintenance_policy,omitempty"` + MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,4,opt,name=maintenance_policy,json=maintenancePolicy,proto3" json:"maintenance_policy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5252,8 +5252,7 @@ func (c *clusterManagerClient) SetMaintenancePolicy(ctx context.Context, in *Set return out, nil } -// Server API for ClusterManager service - +// ClusterManagerServer is the server API for ClusterManager service. type ClusterManagerServer interface { // Lists all clusters owned by a project in either the specified zone or all // zones. diff --git a/vendor/google.golang.org/genproto/googleapis/container/v1alpha1/cluster_service.pb.go b/vendor/google.golang.org/genproto/googleapis/container/v1alpha1/cluster_service.pb.go index ee509e448..98835d459 100644 --- a/vendor/google.golang.org/genproto/googleapis/container/v1alpha1/cluster_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/container/v1alpha1/cluster_service.pb.go @@ -353,12 +353,12 @@ type NodeConfig struct { // // If unspecified, the default machine type is // `n1-standard-1`. - MachineType string `protobuf:"bytes,1,opt,name=machine_type,json=machineType" json:"machine_type,omitempty"` + MachineType string `protobuf:"bytes,1,opt,name=machine_type,json=machineType,proto3" json:"machine_type,omitempty"` // Size of the disk attached to each node, specified in GB. // The smallest allowed disk size is 10GB. // // If unspecified, the default disk size is 100GB. - DiskSizeGb int32 `protobuf:"varint,2,opt,name=disk_size_gb,json=diskSizeGb" json:"disk_size_gb,omitempty"` + DiskSizeGb int32 `protobuf:"varint,2,opt,name=disk_size_gb,json=diskSizeGb,proto3" json:"disk_size_gb,omitempty"` // The set of Google API scopes to be made available on all of the // node VMs under the "default" service account. // @@ -373,10 +373,10 @@ type NodeConfig struct { // // If unspecified, no scopes are added, unless Cloud Logging or Cloud // Monitoring are enabled, in which case their required scopes will be added. - OauthScopes []string `protobuf:"bytes,3,rep,name=oauth_scopes,json=oauthScopes" json:"oauth_scopes,omitempty"` + OauthScopes []string `protobuf:"bytes,3,rep,name=oauth_scopes,json=oauthScopes,proto3" json:"oauth_scopes,omitempty"` // The Google Cloud Platform Service Account to be used by the node VMs. If // no Service Account is specified, the "default" service account is used. - ServiceAccount string `protobuf:"bytes,9,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount string `protobuf:"bytes,9,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // The metadata key/value pairs assigned to instances in the cluster. // // Keys must conform to the regexp [a-zA-Z0-9-_]+ and be less than 128 bytes @@ -390,10 +390,10 @@ type NodeConfig struct { // that each value's size must be less than or equal to 32 KB. // // The total size of all keys and values must be less than 512 KB. - Metadata map[string]string `protobuf:"bytes,4,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,4,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The image type to use for this node. Note that for a given image type, // the latest version of it will be used. - ImageType string `protobuf:"bytes,5,opt,name=image_type,json=imageType" json:"image_type,omitempty"` + ImageType string `protobuf:"bytes,5,opt,name=image_type,json=imageType,proto3" json:"image_type,omitempty"` // The map of Kubernetes labels (key/value pairs) to be applied to each node. // These will added in addition to any default label(s) that // Kubernetes may apply to the node. @@ -402,39 +402,39 @@ type NodeConfig struct { // and conflicts should be avoided. // For more information, including usage and the valid values, see: // https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/ - Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,6,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The number of local SSD disks to be attached to the node. // // The limit for this value is dependant upon the maximum number of // disks available on a machine per zone. See: // https://cloud.google.com/compute/docs/disks/local-ssd#local_ssd_limits // for more information. - LocalSsdCount int32 `protobuf:"varint,7,opt,name=local_ssd_count,json=localSsdCount" json:"local_ssd_count,omitempty"` + LocalSsdCount int32 `protobuf:"varint,7,opt,name=local_ssd_count,json=localSsdCount,proto3" json:"local_ssd_count,omitempty"` // The list of instance tags applied to all nodes. Tags are used to identify // valid sources or targets for network firewalls and are specified by // the client during cluster or node pool creation. Each tag within the list // must comply with RFC1035. - Tags []string `protobuf:"bytes,8,rep,name=tags" json:"tags,omitempty"` + Tags []string `protobuf:"bytes,8,rep,name=tags,proto3" json:"tags,omitempty"` // Whether the nodes are created as preemptible VM instances. See: // https://cloud.google.com/compute/docs/instances/preemptible for more // inforamtion about preemptible VM instances. - Preemptible bool `protobuf:"varint,10,opt,name=preemptible" json:"preemptible,omitempty"` + Preemptible bool `protobuf:"varint,10,opt,name=preemptible,proto3" json:"preemptible,omitempty"` // A list of hardware accelerators to be attached to each node. // See https://cloud.google.com/compute/docs/gpus for more information about // support for GPUs. - Accelerators []*AcceleratorConfig `protobuf:"bytes,11,rep,name=accelerators" json:"accelerators,omitempty"` + Accelerators []*AcceleratorConfig `protobuf:"bytes,11,rep,name=accelerators,proto3" json:"accelerators,omitempty"` // Minimum CPU platform to be used by this instance. The instance may be // scheduled on the specified or newer CPU platform. Applicable values are the // friendly names of CPU platforms, such as // minCpuPlatform: "Intel Haswell" or // minCpuPlatform: "Intel Sandy Bridge". For more // information, read [how to specify min CPU platform](https://cloud.google.com/compute/docs/instances/specify-min-cpu-platform) - MinCpuPlatform string `protobuf:"bytes,13,opt,name=min_cpu_platform,json=minCpuPlatform" json:"min_cpu_platform,omitempty"` + MinCpuPlatform string `protobuf:"bytes,13,opt,name=min_cpu_platform,json=minCpuPlatform,proto3" json:"min_cpu_platform,omitempty"` // List of kubernetes taints to be applied to each node. // // For more information, including usage and the valid values, see: // https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/ - Taints []*NodeTaint `protobuf:"bytes,15,rep,name=taints" json:"taints,omitempty"` + Taints []*NodeTaint `protobuf:"bytes,15,rep,name=taints,proto3" json:"taints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -562,11 +562,11 @@ func (m *NodeConfig) GetTaints() []*NodeTaint { // https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/ type NodeTaint struct { // Key for taint. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // Value for taint. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Effect for taint. - Effect NodeTaint_Effect `protobuf:"varint,3,opt,name=effect,enum=google.container.v1alpha1.NodeTaint_Effect" json:"effect,omitempty"` + Effect NodeTaint_Effect `protobuf:"varint,3,opt,name=effect,proto3,enum=google.container.v1alpha1.NodeTaint_Effect" json:"effect,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -624,24 +624,24 @@ type MasterAuth struct { // The username to use for HTTP basic authentication to the master endpoint. // For clusters v1.6.0 and later, you can disable basic authentication by // providing an empty username. - Username string `protobuf:"bytes,1,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"` // The password to use for HTTP basic authentication to the master endpoint. // Because the master endpoint is open to the Internet, you should create a // strong password. If a password is provided for cluster creation, username // must be non-empty. - Password string `protobuf:"bytes,2,opt,name=password" json:"password,omitempty"` + Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"` // Configuration for client certificate authentication on the cluster. If no // configuration is specified, a client certificate is issued. - ClientCertificateConfig *ClientCertificateConfig `protobuf:"bytes,3,opt,name=client_certificate_config,json=clientCertificateConfig" json:"client_certificate_config,omitempty"` + ClientCertificateConfig *ClientCertificateConfig `protobuf:"bytes,3,opt,name=client_certificate_config,json=clientCertificateConfig,proto3" json:"client_certificate_config,omitempty"` // [Output only] Base64-encoded public certificate that is the root of // trust for the cluster. - ClusterCaCertificate string `protobuf:"bytes,100,opt,name=cluster_ca_certificate,json=clusterCaCertificate" json:"cluster_ca_certificate,omitempty"` + ClusterCaCertificate string `protobuf:"bytes,100,opt,name=cluster_ca_certificate,json=clusterCaCertificate,proto3" json:"cluster_ca_certificate,omitempty"` // [Output only] Base64-encoded public certificate used by clients to // authenticate to the cluster endpoint. - ClientCertificate string `protobuf:"bytes,101,opt,name=client_certificate,json=clientCertificate" json:"client_certificate,omitempty"` + ClientCertificate string `protobuf:"bytes,101,opt,name=client_certificate,json=clientCertificate,proto3" json:"client_certificate,omitempty"` // [Output only] Base64-encoded private key used by clients to authenticate // to the cluster endpoint. - ClientKey string `protobuf:"bytes,102,opt,name=client_key,json=clientKey" json:"client_key,omitempty"` + ClientKey string `protobuf:"bytes,102,opt,name=client_key,json=clientKey,proto3" json:"client_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -716,7 +716,7 @@ func (m *MasterAuth) GetClientKey() string { // Configuration for client certificates on the cluster. type ClientCertificateConfig struct { // Issue a client certificate. - IssueClientCertificate bool `protobuf:"varint,1,opt,name=issue_client_certificate,json=issueClientCertificate" json:"issue_client_certificate,omitempty"` + IssueClientCertificate bool `protobuf:"varint,1,opt,name=issue_client_certificate,json=issueClientCertificate,proto3" json:"issue_client_certificate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -758,17 +758,17 @@ func (m *ClientCertificateConfig) GetIssueClientCertificate() bool { type AddonsConfig struct { // Configuration for the HTTP (L7) load balancing controller addon, which // makes it easy to set up HTTP load balancers for services in a cluster. - HttpLoadBalancing *HttpLoadBalancing `protobuf:"bytes,1,opt,name=http_load_balancing,json=httpLoadBalancing" json:"http_load_balancing,omitempty"` + HttpLoadBalancing *HttpLoadBalancing `protobuf:"bytes,1,opt,name=http_load_balancing,json=httpLoadBalancing,proto3" json:"http_load_balancing,omitempty"` // Configuration for the horizontal pod autoscaling feature, which // increases or decreases the number of replica pods a replication controller // has based on the resource usage of the existing pods. - HorizontalPodAutoscaling *HorizontalPodAutoscaling `protobuf:"bytes,2,opt,name=horizontal_pod_autoscaling,json=horizontalPodAutoscaling" json:"horizontal_pod_autoscaling,omitempty"` + HorizontalPodAutoscaling *HorizontalPodAutoscaling `protobuf:"bytes,2,opt,name=horizontal_pod_autoscaling,json=horizontalPodAutoscaling,proto3" json:"horizontal_pod_autoscaling,omitempty"` // Configuration for the Kubernetes Dashboard. - KubernetesDashboard *KubernetesDashboard `protobuf:"bytes,3,opt,name=kubernetes_dashboard,json=kubernetesDashboard" json:"kubernetes_dashboard,omitempty"` + KubernetesDashboard *KubernetesDashboard `protobuf:"bytes,3,opt,name=kubernetes_dashboard,json=kubernetesDashboard,proto3" json:"kubernetes_dashboard,omitempty"` // Configuration for NetworkPolicy. This only tracks whether the addon // is enabled or not on the Master, it does not track whether network policy // is enabled for the nodes. - NetworkPolicyConfig *NetworkPolicyConfig `protobuf:"bytes,4,opt,name=network_policy_config,json=networkPolicyConfig" json:"network_policy_config,omitempty"` + NetworkPolicyConfig *NetworkPolicyConfig `protobuf:"bytes,4,opt,name=network_policy_config,json=networkPolicyConfig,proto3" json:"network_policy_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -832,7 +832,7 @@ type HttpLoadBalancing struct { // Whether the HTTP Load Balancing controller is enabled in the cluster. // When enabled, it runs a small pod in the cluster that manages the load // balancers. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -876,7 +876,7 @@ type HorizontalPodAutoscaling struct { // Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. // When enabled, it ensures that a Heapster pod is running in the cluster, // which is also used by the Cloud Monitoring service. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -916,7 +916,7 @@ func (m *HorizontalPodAutoscaling) GetDisabled() bool { // Configuration for the Kubernetes Dashboard. type KubernetesDashboard struct { // Whether the Kubernetes Dashboard is enabled for this cluster. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -958,7 +958,7 @@ func (m *KubernetesDashboard) GetDisabled() bool { // is enabled for the nodes. type NetworkPolicyConfig struct { // Whether NetworkPolicy is enabled for this cluster. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1001,10 +1001,10 @@ func (m *NetworkPolicyConfig) GetDisabled() bool { // Google Compute Engine Public IPs and Google Prod IPs. type MasterAuthorizedNetworksConfig struct { // Whether or not master authorized networks is enabled. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` // cidr_blocks define up to 10 external networks that could access // Kubernetes master through HTTPS. - CidrBlocks []*MasterAuthorizedNetworksConfig_CidrBlock `protobuf:"bytes,2,rep,name=cidr_blocks,json=cidrBlocks" json:"cidr_blocks,omitempty"` + CidrBlocks []*MasterAuthorizedNetworksConfig_CidrBlock `protobuf:"bytes,2,rep,name=cidr_blocks,json=cidrBlocks,proto3" json:"cidr_blocks,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1051,9 +1051,9 @@ func (m *MasterAuthorizedNetworksConfig) GetCidrBlocks() []*MasterAuthorizedNetw // CidrBlock contains an optional name and one CIDR block. type MasterAuthorizedNetworksConfig_CidrBlock struct { // display_name is an optional field for users to identify CIDR blocks. - DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // cidr_block must be specified in CIDR notation. - CidrBlock string `protobuf:"bytes,2,opt,name=cidr_block,json=cidrBlock" json:"cidr_block,omitempty"` + CidrBlock string `protobuf:"bytes,2,opt,name=cidr_block,json=cidrBlock,proto3" json:"cidr_block,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1103,9 +1103,9 @@ func (m *MasterAuthorizedNetworksConfig_CidrBlock) GetCidrBlock() string { // https://kubernetes.io/docs/concepts/services-networking/networkpolicies/ type NetworkPolicy struct { // The selected network policy provider. - Provider NetworkPolicy_Provider `protobuf:"varint,1,opt,name=provider,enum=google.container.v1alpha1.NetworkPolicy_Provider" json:"provider,omitempty"` + Provider NetworkPolicy_Provider `protobuf:"varint,1,opt,name=provider,proto3,enum=google.container.v1alpha1.NetworkPolicy_Provider" json:"provider,omitempty"` // Whether network policy is enabled on the cluster. - Enabled bool `protobuf:"varint,2,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,2,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1152,21 +1152,21 @@ func (m *NetworkPolicy) GetEnabled() bool { // Configuration for controlling how IPs are allocated in the cluster. type IPAllocationPolicy struct { // Whether alias IPs will be used for pod IPs in the cluster. - UseIpAliases bool `protobuf:"varint,1,opt,name=use_ip_aliases,json=useIpAliases" json:"use_ip_aliases,omitempty"` + UseIpAliases bool `protobuf:"varint,1,opt,name=use_ip_aliases,json=useIpAliases,proto3" json:"use_ip_aliases,omitempty"` // Whether a new subnetwork will be created automatically for the cluster. // // This field is only applicable when `use_ip_aliases` is true. - CreateSubnetwork bool `protobuf:"varint,2,opt,name=create_subnetwork,json=createSubnetwork" json:"create_subnetwork,omitempty"` + CreateSubnetwork bool `protobuf:"varint,2,opt,name=create_subnetwork,json=createSubnetwork,proto3" json:"create_subnetwork,omitempty"` // A custom subnetwork name to be used if `create_subnetwork` is true. If // this field is empty, then an automatic name will be chosen for the new // subnetwork. - SubnetworkName string `protobuf:"bytes,3,opt,name=subnetwork_name,json=subnetworkName" json:"subnetwork_name,omitempty"` + SubnetworkName string `protobuf:"bytes,3,opt,name=subnetwork_name,json=subnetworkName,proto3" json:"subnetwork_name,omitempty"` // This field is deprecated, use cluster_ipv4_cidr_block. - ClusterIpv4Cidr string `protobuf:"bytes,4,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr" json:"cluster_ipv4_cidr,omitempty"` + ClusterIpv4Cidr string `protobuf:"bytes,4,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr,proto3" json:"cluster_ipv4_cidr,omitempty"` // This field is deprecated, use node_ipv4_cidr_block. - NodeIpv4Cidr string `protobuf:"bytes,5,opt,name=node_ipv4_cidr,json=nodeIpv4Cidr" json:"node_ipv4_cidr,omitempty"` + NodeIpv4Cidr string `protobuf:"bytes,5,opt,name=node_ipv4_cidr,json=nodeIpv4Cidr,proto3" json:"node_ipv4_cidr,omitempty"` // This field is deprecated, use services_ipv4_cidr_block. - ServicesIpv4Cidr string `protobuf:"bytes,6,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr" json:"services_ipv4_cidr,omitempty"` + ServicesIpv4Cidr string `protobuf:"bytes,6,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr,proto3" json:"services_ipv4_cidr,omitempty"` // The name of the secondary range to be used for the cluster CIDR // block. The secondary range will be used for pod IP // addresses. This must be an existing secondary range associated @@ -1174,7 +1174,7 @@ type IPAllocationPolicy struct { // // This field is only applicable if use_ip_aliases is true and // create_subnetwork is false. - ClusterSecondaryRangeName string `protobuf:"bytes,7,opt,name=cluster_secondary_range_name,json=clusterSecondaryRangeName" json:"cluster_secondary_range_name,omitempty"` + ClusterSecondaryRangeName string `protobuf:"bytes,7,opt,name=cluster_secondary_range_name,json=clusterSecondaryRangeName,proto3" json:"cluster_secondary_range_name,omitempty"` // The name of the secondary range to be used as for the services // CIDR block. The secondary range will be used for service // ClusterIPs. This must be an existing secondary range associated @@ -1182,7 +1182,7 @@ type IPAllocationPolicy struct { // // This field is only applicable with use_ip_aliases is true and // create_subnetwork is false. - ServicesSecondaryRangeName string `protobuf:"bytes,8,opt,name=services_secondary_range_name,json=servicesSecondaryRangeName" json:"services_secondary_range_name,omitempty"` + ServicesSecondaryRangeName string `protobuf:"bytes,8,opt,name=services_secondary_range_name,json=servicesSecondaryRangeName,proto3" json:"services_secondary_range_name,omitempty"` // The IP address range for the cluster pod IPs. If this field is set, then // `cluster.cluster_ipv4_cidr` must be left blank. // @@ -1198,7 +1198,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - ClusterIpv4CidrBlock string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr_block,json=clusterIpv4CidrBlock" json:"cluster_ipv4_cidr_block,omitempty"` + ClusterIpv4CidrBlock string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr_block,json=clusterIpv4CidrBlock,proto3" json:"cluster_ipv4_cidr_block,omitempty"` // The IP address range of the instance IPs in this cluster. // // This is applicable only if `create_subnetwork` is true. @@ -1213,7 +1213,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - NodeIpv4CidrBlock string `protobuf:"bytes,10,opt,name=node_ipv4_cidr_block,json=nodeIpv4CidrBlock" json:"node_ipv4_cidr_block,omitempty"` + NodeIpv4CidrBlock string `protobuf:"bytes,10,opt,name=node_ipv4_cidr_block,json=nodeIpv4CidrBlock,proto3" json:"node_ipv4_cidr_block,omitempty"` // The IP address range of the services IPs in this cluster. If blank, a range // will be automatically chosen with the default size. // @@ -1229,7 +1229,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - ServicesIpv4CidrBlock string `protobuf:"bytes,11,opt,name=services_ipv4_cidr_block,json=servicesIpv4CidrBlock" json:"services_ipv4_cidr_block,omitempty"` + ServicesIpv4CidrBlock string `protobuf:"bytes,11,opt,name=services_ipv4_cidr_block,json=servicesIpv4CidrBlock,proto3" json:"services_ipv4_cidr_block,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1340,7 +1340,7 @@ func (m *IPAllocationPolicy) GetServicesIpv4CidrBlock() string { type PodSecurityPolicyConfig struct { // Enable the PodSecurityPolicy controller for this cluster. If enabled, pods // must be valid under a PodSecurityPolicy to be created. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1385,9 +1385,9 @@ type Cluster struct { // * Lowercase letters, numbers, and hyphens only. // * Must start with a letter. // * Must end with a number or a letter. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // An optional description of this cluster. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // The number of nodes to create in this cluster. You must ensure that your // Compute Engine resource quota // is sufficient for this number of instances. You must also have available @@ -1396,7 +1396,7 @@ type Cluster struct { // "node_pool" object, since this configuration (along with the // "node_config") will be used to create a "NodePool" object with an // auto-generated name. Do not use this and a node_pool at the same time. - InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount" json:"initial_node_count,omitempty"` + InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount,proto3" json:"initial_node_count,omitempty"` // Parameters used in creating the cluster's nodes. // See `nodeConfig` for the description of its properties. // For requests, this field should only be used in lieu of a @@ -1407,121 +1407,121 @@ type Cluster struct { // the first node pool. // // If unspecified, the defaults are used. - NodeConfig *NodeConfig `protobuf:"bytes,4,opt,name=node_config,json=nodeConfig" json:"node_config,omitempty"` + NodeConfig *NodeConfig `protobuf:"bytes,4,opt,name=node_config,json=nodeConfig,proto3" json:"node_config,omitempty"` // The authentication information for accessing the master endpoint. - MasterAuth *MasterAuth `protobuf:"bytes,5,opt,name=master_auth,json=masterAuth" json:"master_auth,omitempty"` + MasterAuth *MasterAuth `protobuf:"bytes,5,opt,name=master_auth,json=masterAuth,proto3" json:"master_auth,omitempty"` // The logging service the cluster should use to write logs. // Currently available options: // // * `logging.googleapis.com` - the Google Cloud Logging service. // * `none` - no logs will be exported from the cluster. // * if left as an empty string,`logging.googleapis.com` will be used. - LoggingService string `protobuf:"bytes,6,opt,name=logging_service,json=loggingService" json:"logging_service,omitempty"` + LoggingService string `protobuf:"bytes,6,opt,name=logging_service,json=loggingService,proto3" json:"logging_service,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * `monitoring.googleapis.com` - the Google Cloud Monitoring service. // * `none` - no metrics will be exported from the cluster. // * if left as an empty string, `monitoring.googleapis.com` will be used. - MonitoringService string `protobuf:"bytes,7,opt,name=monitoring_service,json=monitoringService" json:"monitoring_service,omitempty"` + MonitoringService string `protobuf:"bytes,7,opt,name=monitoring_service,json=monitoringService,proto3" json:"monitoring_service,omitempty"` // The name of the Google Compute Engine // [network](/compute/docs/networks-and-firewalls#networks) to which the // cluster is connected. If left unspecified, the `default` network // will be used. - Network string `protobuf:"bytes,8,opt,name=network" json:"network,omitempty"` + Network string `protobuf:"bytes,8,opt,name=network,proto3" json:"network,omitempty"` // The IP address range of the container pods in this cluster, in // [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) // notation (e.g. `10.96.0.0/14`). Leave blank to have // one automatically chosen or specify a `/14` block in `10.0.0.0/8`. - ClusterIpv4Cidr string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr" json:"cluster_ipv4_cidr,omitempty"` + ClusterIpv4Cidr string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr,proto3" json:"cluster_ipv4_cidr,omitempty"` // Configurations for the various addons available to run in the cluster. - AddonsConfig *AddonsConfig `protobuf:"bytes,10,opt,name=addons_config,json=addonsConfig" json:"addons_config,omitempty"` + AddonsConfig *AddonsConfig `protobuf:"bytes,10,opt,name=addons_config,json=addonsConfig,proto3" json:"addons_config,omitempty"` // The name of the Google Compute Engine // [subnetwork](/compute/docs/subnetworks) to which the // cluster is connected. - Subnetwork string `protobuf:"bytes,11,opt,name=subnetwork" json:"subnetwork,omitempty"` + Subnetwork string `protobuf:"bytes,11,opt,name=subnetwork,proto3" json:"subnetwork,omitempty"` // The node pools associated with this cluster. // This field should not be set if "node_config" or "initial_node_count" are // specified. - NodePools []*NodePool `protobuf:"bytes,12,rep,name=node_pools,json=nodePools" json:"node_pools,omitempty"` + NodePools []*NodePool `protobuf:"bytes,12,rep,name=node_pools,json=nodePools,proto3" json:"node_pools,omitempty"` // The list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. - Locations []string `protobuf:"bytes,13,rep,name=locations" json:"locations,omitempty"` + Locations []string `protobuf:"bytes,13,rep,name=locations,proto3" json:"locations,omitempty"` // Kubernetes alpha features are enabled on this cluster. This includes alpha // API groups (e.g. v1alpha1) and features that may not be production ready in // the kubernetes version of the master and nodes. // The cluster has no SLA for uptime and master/node upgrades are disabled. // Alpha enabled clusters are automatically deleted thirty days after // creation. - EnableKubernetesAlpha bool `protobuf:"varint,14,opt,name=enable_kubernetes_alpha,json=enableKubernetesAlpha" json:"enable_kubernetes_alpha,omitempty"` + EnableKubernetesAlpha bool `protobuf:"varint,14,opt,name=enable_kubernetes_alpha,json=enableKubernetesAlpha,proto3" json:"enable_kubernetes_alpha,omitempty"` // Configuration options for the NetworkPolicy feature. - NetworkPolicy *NetworkPolicy `protobuf:"bytes,19,opt,name=network_policy,json=networkPolicy" json:"network_policy,omitempty"` + NetworkPolicy *NetworkPolicy `protobuf:"bytes,19,opt,name=network_policy,json=networkPolicy,proto3" json:"network_policy,omitempty"` // Configuration for cluster IP allocation. - IpAllocationPolicy *IPAllocationPolicy `protobuf:"bytes,20,opt,name=ip_allocation_policy,json=ipAllocationPolicy" json:"ip_allocation_policy,omitempty"` + IpAllocationPolicy *IPAllocationPolicy `protobuf:"bytes,20,opt,name=ip_allocation_policy,json=ipAllocationPolicy,proto3" json:"ip_allocation_policy,omitempty"` // The configuration options for master authorized networks feature. - MasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,22,opt,name=master_authorized_networks_config,json=masterAuthorizedNetworksConfig" json:"master_authorized_networks_config,omitempty"` + MasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,22,opt,name=master_authorized_networks_config,json=masterAuthorizedNetworksConfig,proto3" json:"master_authorized_networks_config,omitempty"` // Configure the maintenance policy for this cluster. - MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,23,opt,name=maintenance_policy,json=maintenancePolicy" json:"maintenance_policy,omitempty"` + MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,23,opt,name=maintenance_policy,json=maintenancePolicy,proto3" json:"maintenance_policy,omitempty"` // Configuration for the PodSecurityPolicy feature. - PodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,25,opt,name=pod_security_policy_config,json=podSecurityPolicyConfig" json:"pod_security_policy_config,omitempty"` + PodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,25,opt,name=pod_security_policy_config,json=podSecurityPolicyConfig,proto3" json:"pod_security_policy_config,omitempty"` // [Output only] Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use location instead. - Zone string `protobuf:"bytes,101,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,101,opt,name=zone,proto3" json:"zone,omitempty"` // [Output only] The IP address of this cluster's master endpoint. // The endpoint can be accessed from the internet at // `https://username:password@endpoint/`. // // See the `masterAuth` property of this resource for username and // password information. - Endpoint string `protobuf:"bytes,102,opt,name=endpoint" json:"endpoint,omitempty"` + Endpoint string `protobuf:"bytes,102,opt,name=endpoint,proto3" json:"endpoint,omitempty"` // The initial Kubernetes version for this cluster. Valid versions are those // found in validMasterVersions returned by getServerConfig. The version can // be upgraded over time; such upgrades are reflected in // currentMasterVersion and currentNodeVersion. - InitialClusterVersion string `protobuf:"bytes,103,opt,name=initial_cluster_version,json=initialClusterVersion" json:"initial_cluster_version,omitempty"` + InitialClusterVersion string `protobuf:"bytes,103,opt,name=initial_cluster_version,json=initialClusterVersion,proto3" json:"initial_cluster_version,omitempty"` // [Output only] The current software version of the master endpoint. - CurrentMasterVersion string `protobuf:"bytes,104,opt,name=current_master_version,json=currentMasterVersion" json:"current_master_version,omitempty"` + CurrentMasterVersion string `protobuf:"bytes,104,opt,name=current_master_version,json=currentMasterVersion,proto3" json:"current_master_version,omitempty"` // [Output only] The current version of the node software components. // If they are currently at multiple versions because they're in the process // of being upgraded, this reflects the minimum version of all nodes. - CurrentNodeVersion string `protobuf:"bytes,105,opt,name=current_node_version,json=currentNodeVersion" json:"current_node_version,omitempty"` + CurrentNodeVersion string `protobuf:"bytes,105,opt,name=current_node_version,json=currentNodeVersion,proto3" json:"current_node_version,omitempty"` // [Output only] The time the cluster was created, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - CreateTime string `protobuf:"bytes,106,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime string `protobuf:"bytes,106,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // [Output only] The current status of this cluster. - Status Cluster_Status `protobuf:"varint,107,opt,name=status,enum=google.container.v1alpha1.Cluster_Status" json:"status,omitempty"` + Status Cluster_Status `protobuf:"varint,107,opt,name=status,proto3,enum=google.container.v1alpha1.Cluster_Status" json:"status,omitempty"` // [Output only] Additional information about the current status of this // cluster, if available. - StatusMessage string `protobuf:"bytes,108,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,108,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // [Output only] The size of the address space on each node for hosting // containers. This is provisioned from within the `container_ipv4_cidr` // range. - NodeIpv4CidrSize int32 `protobuf:"varint,109,opt,name=node_ipv4_cidr_size,json=nodeIpv4CidrSize" json:"node_ipv4_cidr_size,omitempty"` + NodeIpv4CidrSize int32 `protobuf:"varint,109,opt,name=node_ipv4_cidr_size,json=nodeIpv4CidrSize,proto3" json:"node_ipv4_cidr_size,omitempty"` // [Output only] The IP address range of the Kubernetes services in // this cluster, in // [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) // notation (e.g. `1.2.3.4/29`). Service addresses are // typically put in the last `/16` from the container CIDR. - ServicesIpv4Cidr string `protobuf:"bytes,110,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr" json:"services_ipv4_cidr,omitempty"` + ServicesIpv4Cidr string `protobuf:"bytes,110,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr,proto3" json:"services_ipv4_cidr,omitempty"` // [Output only] The resource URLs of [instance // groups](/compute/docs/instance-groups/) associated with this // cluster. - InstanceGroupUrls []string `protobuf:"bytes,111,rep,name=instance_group_urls,json=instanceGroupUrls" json:"instance_group_urls,omitempty"` + InstanceGroupUrls []string `protobuf:"bytes,111,rep,name=instance_group_urls,json=instanceGroupUrls,proto3" json:"instance_group_urls,omitempty"` // [Output only] The number of nodes currently in the cluster. - CurrentNodeCount int32 `protobuf:"varint,112,opt,name=current_node_count,json=currentNodeCount" json:"current_node_count,omitempty"` + CurrentNodeCount int32 `protobuf:"varint,112,opt,name=current_node_count,json=currentNodeCount,proto3" json:"current_node_count,omitempty"` // [Output only] The time the cluster will be automatically // deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - ExpireTime string `protobuf:"bytes,113,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime string `protobuf:"bytes,113,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/regions-zones/regions-zones#available) or // [region](/compute/docs/regions-zones/regions-zones#available) in which // the cluster resides. - Location string `protobuf:"bytes,114,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,114,opt,name=location,proto3" json:"location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1796,28 +1796,28 @@ type ClusterUpdate struct { // The Kubernetes version to change the nodes to (typically an // upgrade). Use `-` to upgrade to the latest version supported by // the server. - DesiredNodeVersion string `protobuf:"bytes,4,opt,name=desired_node_version,json=desiredNodeVersion" json:"desired_node_version,omitempty"` + DesiredNodeVersion string `protobuf:"bytes,4,opt,name=desired_node_version,json=desiredNodeVersion,proto3" json:"desired_node_version,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * "monitoring.googleapis.com" - the Google Cloud Monitoring service // * "none" - no metrics will be exported from the cluster - DesiredMonitoringService string `protobuf:"bytes,5,opt,name=desired_monitoring_service,json=desiredMonitoringService" json:"desired_monitoring_service,omitempty"` + DesiredMonitoringService string `protobuf:"bytes,5,opt,name=desired_monitoring_service,json=desiredMonitoringService,proto3" json:"desired_monitoring_service,omitempty"` // Configurations for the various addons available to run in the cluster. - DesiredAddonsConfig *AddonsConfig `protobuf:"bytes,6,opt,name=desired_addons_config,json=desiredAddonsConfig" json:"desired_addons_config,omitempty"` + DesiredAddonsConfig *AddonsConfig `protobuf:"bytes,6,opt,name=desired_addons_config,json=desiredAddonsConfig,proto3" json:"desired_addons_config,omitempty"` // The node pool to be upgraded. This field is mandatory if // "desired_node_version", "desired_image_family" or // "desired_node_pool_autoscaling" is specified and there is more than one // node pool on the cluster. - DesiredNodePoolId string `protobuf:"bytes,7,opt,name=desired_node_pool_id,json=desiredNodePoolId" json:"desired_node_pool_id,omitempty"` + DesiredNodePoolId string `protobuf:"bytes,7,opt,name=desired_node_pool_id,json=desiredNodePoolId,proto3" json:"desired_node_pool_id,omitempty"` // The desired image type for the node pool. // NOTE: Set the "desired_node_pool" field as well. - DesiredImageType string `protobuf:"bytes,8,opt,name=desired_image_type,json=desiredImageType" json:"desired_image_type,omitempty"` + DesiredImageType string `protobuf:"bytes,8,opt,name=desired_image_type,json=desiredImageType,proto3" json:"desired_image_type,omitempty"` // Autoscaler configuration for the node pool specified in // desired_node_pool_id. If there is only one pool in the // cluster and desired_node_pool_id is not provided then // the change applies to that single node pool. - DesiredNodePoolAutoscaling *NodePoolAutoscaling `protobuf:"bytes,9,opt,name=desired_node_pool_autoscaling,json=desiredNodePoolAutoscaling" json:"desired_node_pool_autoscaling,omitempty"` + DesiredNodePoolAutoscaling *NodePoolAutoscaling `protobuf:"bytes,9,opt,name=desired_node_pool_autoscaling,json=desiredNodePoolAutoscaling,proto3" json:"desired_node_pool_autoscaling,omitempty"` // The desired list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. Changing the locations a cluster is in will result @@ -1825,15 +1825,15 @@ type ClusterUpdate struct { // whether locations are being added or removed. // // This list must always include the cluster's primary zone. - DesiredLocations []string `protobuf:"bytes,10,rep,name=desired_locations,json=desiredLocations" json:"desired_locations,omitempty"` + DesiredLocations []string `protobuf:"bytes,10,rep,name=desired_locations,json=desiredLocations,proto3" json:"desired_locations,omitempty"` // The desired configuration options for master authorized networks feature. - DesiredMasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,12,opt,name=desired_master_authorized_networks_config,json=desiredMasterAuthorizedNetworksConfig" json:"desired_master_authorized_networks_config,omitempty"` + DesiredMasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,12,opt,name=desired_master_authorized_networks_config,json=desiredMasterAuthorizedNetworksConfig,proto3" json:"desired_master_authorized_networks_config,omitempty"` // The desired configuration options for the PodSecurityPolicy feature. - DesiredPodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,14,opt,name=desired_pod_security_policy_config,json=desiredPodSecurityPolicyConfig" json:"desired_pod_security_policy_config,omitempty"` + DesiredPodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,14,opt,name=desired_pod_security_policy_config,json=desiredPodSecurityPolicyConfig,proto3" json:"desired_pod_security_policy_config,omitempty"` // The Kubernetes version to change the master to. The only valid value is the // latest supported version. Use "-" to have the server automatically select // the latest version. - DesiredMasterVersion string `protobuf:"bytes,100,opt,name=desired_master_version,json=desiredMasterVersion" json:"desired_master_version,omitempty"` + DesiredMasterVersion string `protobuf:"bytes,100,opt,name=desired_master_version,json=desiredMasterVersion,proto3" json:"desired_master_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1937,35 +1937,35 @@ func (m *ClusterUpdate) GetDesiredMasterVersion() string { // happening on the cluster. All fields are output only. type Operation struct { // The server-assigned ID for the operation. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the operation // is taking place. // This field is deprecated, use location instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The operation type. - OperationType Operation_Type `protobuf:"varint,3,opt,name=operation_type,json=operationType,enum=google.container.v1alpha1.Operation_Type" json:"operation_type,omitempty"` + OperationType Operation_Type `protobuf:"varint,3,opt,name=operation_type,json=operationType,proto3,enum=google.container.v1alpha1.Operation_Type" json:"operation_type,omitempty"` // The current status of the operation. - Status Operation_Status `protobuf:"varint,4,opt,name=status,enum=google.container.v1alpha1.Operation_Status" json:"status,omitempty"` + Status Operation_Status `protobuf:"varint,4,opt,name=status,proto3,enum=google.container.v1alpha1.Operation_Status" json:"status,omitempty"` // Detailed operation progress, if available. - Detail string `protobuf:"bytes,8,opt,name=detail" json:"detail,omitempty"` + Detail string `protobuf:"bytes,8,opt,name=detail,proto3" json:"detail,omitempty"` // If an error has occurred, a textual description of the error. - StatusMessage string `protobuf:"bytes,5,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,5,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,6,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,6,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // Server-defined URL for the target of the operation. - TargetLink string `protobuf:"bytes,7,opt,name=target_link,json=targetLink" json:"target_link,omitempty"` + TargetLink string `protobuf:"bytes,7,opt,name=target_link,json=targetLink,proto3" json:"target_link,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/regions-zones/regions-zones#available) or // [region](/compute/docs/regions-zones/regions-zones#available) in which // the cluster resides. - Location string `protobuf:"bytes,9,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,9,opt,name=location,proto3" json:"location,omitempty"` // [Output only] The time the operation started, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - StartTime string `protobuf:"bytes,10,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime string `protobuf:"bytes,10,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // [Output only] The time the operation completed, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - EndTime string `protobuf:"bytes,11,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime string `protobuf:"bytes,11,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2077,18 +2077,18 @@ type CreateClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // A [cluster // resource](/container-engine/reference/rest/v1alpha1/projects.zones.clusters) - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` // The parent (project and location) where the cluster will be created. // Specified in the format 'projects/*/locations/*'. - Parent string `protobuf:"bytes,5,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,5,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2151,18 +2151,18 @@ type GetClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to retrieve. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster) of the cluster to retrieve. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2225,20 +2225,20 @@ type UpdateClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // A description of the update. - Update *ClusterUpdate `protobuf:"bytes,4,opt,name=update" json:"update,omitempty"` + Update *ClusterUpdate `protobuf:"bytes,4,opt,name=update,proto3" json:"update,omitempty"` // The name (project, location, cluster) of the cluster to update. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2308,27 +2308,27 @@ type UpdateNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to upgrade. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The Kubernetes version to change the nodes to (typically an // upgrade). Use `-` to upgrade to the latest version supported by // the server. - NodeVersion string `protobuf:"bytes,5,opt,name=node_version,json=nodeVersion" json:"node_version,omitempty"` + NodeVersion string `protobuf:"bytes,5,opt,name=node_version,json=nodeVersion,proto3" json:"node_version,omitempty"` // The desired image type for the node pool. - ImageType string `protobuf:"bytes,6,opt,name=image_type,json=imageType" json:"image_type,omitempty"` + ImageType string `protobuf:"bytes,6,opt,name=image_type,json=imageType,proto3" json:"image_type,omitempty"` // The name (project, location, cluster, node pool) of the node pool to update. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,8,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,8,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2412,24 +2412,24 @@ type SetNodePoolAutoscalingRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to upgrade. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // Autoscaling configuration for the node pool. - Autoscaling *NodePoolAutoscaling `protobuf:"bytes,5,opt,name=autoscaling" json:"autoscaling,omitempty"` + Autoscaling *NodePoolAutoscaling `protobuf:"bytes,5,opt,name=autoscaling,proto3" json:"autoscaling,omitempty"` // The name (project, location, cluster, node pool) of the node pool to set // autoscaler settings. Specified in the format // 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2506,23 +2506,23 @@ type SetLoggingServiceRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The logging service the cluster should use to write metrics. // Currently available options: // // * "logging.googleapis.com" - the Google Cloud Logging service // * "none" - no metrics will be exported from the cluster - LoggingService string `protobuf:"bytes,4,opt,name=logging_service,json=loggingService" json:"logging_service,omitempty"` + LoggingService string `protobuf:"bytes,4,opt,name=logging_service,json=loggingService,proto3" json:"logging_service,omitempty"` // The name (project, location, cluster) of the cluster to set logging. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2592,24 +2592,24 @@ type SetMonitoringServiceRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * "monitoring.googleapis.com" - the Google Cloud Monitoring service // * "none" - no metrics will be exported from the cluster - MonitoringService string `protobuf:"bytes,4,opt,name=monitoring_service,json=monitoringService" json:"monitoring_service,omitempty"` + MonitoringService string `protobuf:"bytes,4,opt,name=monitoring_service,json=monitoringService,proto3" json:"monitoring_service,omitempty"` // The name (project, location, cluster) of the cluster to set monitoring. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2679,21 +2679,21 @@ type SetAddonsConfigRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The desired configurations for the various addons available to run in the // cluster. - AddonsConfig *AddonsConfig `protobuf:"bytes,4,opt,name=addons_config,json=addonsConfig" json:"addons_config,omitempty"` + AddonsConfig *AddonsConfig `protobuf:"bytes,4,opt,name=addons_config,json=addonsConfig,proto3" json:"addons_config,omitempty"` // The name (project, location, cluster) of the cluster to set addons. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2763,15 +2763,15 @@ type SetLocationsRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The desired list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. Changing the locations a cluster is in will result @@ -2779,10 +2779,10 @@ type SetLocationsRequest struct { // whether locations are being added or removed. // // This list must always include the cluster's primary zone. - Locations []string `protobuf:"bytes,4,rep,name=locations" json:"locations,omitempty"` + Locations []string `protobuf:"bytes,4,rep,name=locations,proto3" json:"locations,omitempty"` // The name (project, location, cluster) of the cluster to set locations. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2851,22 +2851,22 @@ func (m *SetLocationsRequest) GetName() string { type UpdateMasterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The Kubernetes version to change the master to. The only valid value is the // latest supported version. Use "-" to have the server automatically select // the latest version. - MasterVersion string `protobuf:"bytes,4,opt,name=master_version,json=masterVersion" json:"master_version,omitempty"` + MasterVersion string `protobuf:"bytes,4,opt,name=master_version,json=masterVersion,proto3" json:"master_version,omitempty"` // The name (project, location, cluster) of the cluster to update. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2936,22 +2936,22 @@ type SetMasterAuthRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The exact form of action to be taken on the master auth. - Action SetMasterAuthRequest_Action `protobuf:"varint,4,opt,name=action,enum=google.container.v1alpha1.SetMasterAuthRequest_Action" json:"action,omitempty"` + Action SetMasterAuthRequest_Action `protobuf:"varint,4,opt,name=action,proto3,enum=google.container.v1alpha1.SetMasterAuthRequest_Action" json:"action,omitempty"` // A description of the update. - Update *MasterAuth `protobuf:"bytes,5,opt,name=update" json:"update,omitempty"` + Update *MasterAuth `protobuf:"bytes,5,opt,name=update,proto3" json:"update,omitempty"` // The name (project, location, cluster) of the cluster to set auth. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3028,18 +3028,18 @@ type DeleteClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to delete. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster) of the cluster to delete. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3102,16 +3102,16 @@ type ListClustersRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides, or "-" for all zones. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The parent (project and location) where the clusters will be listed. // Specified in the format 'projects/*/locations/*'. // Location "-" matches all zones and all regions. - Parent string `protobuf:"bytes,4,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,4,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3166,10 +3166,10 @@ func (m *ListClustersRequest) GetParent() string { type ListClustersResponse struct { // A list of clusters in the project in the specified zone, or // across all ones. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` // If any zones are listed here, the list of clusters returned // may be missing those zones. - MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones" json:"missing_zones,omitempty"` + MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones,proto3" json:"missing_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3218,18 +3218,18 @@ type GetOperationRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The server-assigned `name` of the operation. // This field is deprecated, use name instead. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // The name (project, location, operation id) of the operation to get. // Specified in the format 'projects/*/locations/*/operations/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3292,15 +3292,15 @@ type ListOperationsRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine [zone](/compute/docs/zones#available) // to return operations for, or `-` for all zones. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The parent (project and location) where the operations will be listed. // Specified in the format 'projects/*/locations/*'. // Location "-" matches all zones and all regions. - Parent string `protobuf:"bytes,4,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,4,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3356,17 +3356,17 @@ type CancelOperationRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the operation resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The server-assigned `name` of the operation. // This field is deprecated, use name instead. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // The name (project, location, operation id) of the operation to cancel. // Specified in the format 'projects/*/locations/*/operations/*'. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3427,10 +3427,10 @@ func (m *CancelOperationRequest) GetName() string { // ListOperationsResponse is the result of ListOperationsRequest. type ListOperationsResponse struct { // A list of operations in the project in the specified zone. - Operations []*Operation `protobuf:"bytes,1,rep,name=operations" json:"operations,omitempty"` + Operations []*Operation `protobuf:"bytes,1,rep,name=operations,proto3" json:"operations,omitempty"` // If any zones are listed here, the list of operations returned // may be missing the operations from those zones. - MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones" json:"missing_zones,omitempty"` + MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones,proto3" json:"missing_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3479,14 +3479,14 @@ type GetServerConfigRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine [zone](/compute/docs/zones#available) // to return operations for. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name (project and location) of the server config to get // Specified in the format 'projects/*/locations/*'. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3540,15 +3540,15 @@ func (m *GetServerConfigRequest) GetName() string { // Container Engine service configuration. type ServerConfig struct { // Version of Kubernetes the service deploys by default. - DefaultClusterVersion string `protobuf:"bytes,1,opt,name=default_cluster_version,json=defaultClusterVersion" json:"default_cluster_version,omitempty"` + DefaultClusterVersion string `protobuf:"bytes,1,opt,name=default_cluster_version,json=defaultClusterVersion,proto3" json:"default_cluster_version,omitempty"` // List of valid node upgrade target versions. - ValidNodeVersions []string `protobuf:"bytes,3,rep,name=valid_node_versions,json=validNodeVersions" json:"valid_node_versions,omitempty"` + ValidNodeVersions []string `protobuf:"bytes,3,rep,name=valid_node_versions,json=validNodeVersions,proto3" json:"valid_node_versions,omitempty"` // Default image type. - DefaultImageType string `protobuf:"bytes,4,opt,name=default_image_type,json=defaultImageType" json:"default_image_type,omitempty"` + DefaultImageType string `protobuf:"bytes,4,opt,name=default_image_type,json=defaultImageType,proto3" json:"default_image_type,omitempty"` // List of valid image types. - ValidImageTypes []string `protobuf:"bytes,5,rep,name=valid_image_types,json=validImageTypes" json:"valid_image_types,omitempty"` + ValidImageTypes []string `protobuf:"bytes,5,rep,name=valid_image_types,json=validImageTypes,proto3" json:"valid_image_types,omitempty"` // List of valid master versions. - ValidMasterVersions []string `protobuf:"bytes,6,rep,name=valid_master_versions,json=validMasterVersions" json:"valid_master_versions,omitempty"` + ValidMasterVersions []string `protobuf:"bytes,6,rep,name=valid_master_versions,json=validMasterVersions,proto3" json:"valid_master_versions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3618,20 +3618,20 @@ type CreateNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use parent instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The node pool to create. - NodePool *NodePool `protobuf:"bytes,4,opt,name=node_pool,json=nodePool" json:"node_pool,omitempty"` + NodePool *NodePool `protobuf:"bytes,4,opt,name=node_pool,json=nodePool,proto3" json:"node_pool,omitempty"` // The parent (project, location, cluster id) where the node pool will be created. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Parent string `protobuf:"bytes,6,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,6,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3701,21 +3701,21 @@ type DeleteNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to delete. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to delete. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3785,18 +3785,18 @@ type ListNodePoolsRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use parent instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The parent (project, location, cluster id) where the node pools will be listed. // Specified in the format 'projects/*/locations/*/clusters/*'. - Parent string `protobuf:"bytes,5,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,5,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3859,21 +3859,21 @@ type GetNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to get. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3946,32 +3946,32 @@ func (m *GetNodePoolRequest) GetName() string { // the workload. type NodePool struct { // The name of the node pool. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The node configuration of the pool. - Config *NodeConfig `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"` + Config *NodeConfig `protobuf:"bytes,2,opt,name=config,proto3" json:"config,omitempty"` // The initial node count for the pool. You must ensure that your // Compute Engine resource quota // is sufficient for this number of instances. You must also have available // firewall and routes quota. - InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount" json:"initial_node_count,omitempty"` + InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount,proto3" json:"initial_node_count,omitempty"` // Autoscaler configuration for this NodePool. Autoscaler is enabled // only if a valid configuration is present. - Autoscaling *NodePoolAutoscaling `protobuf:"bytes,4,opt,name=autoscaling" json:"autoscaling,omitempty"` + Autoscaling *NodePoolAutoscaling `protobuf:"bytes,4,opt,name=autoscaling,proto3" json:"autoscaling,omitempty"` // NodeManagement configuration for this NodePool. - Management *NodeManagement `protobuf:"bytes,5,opt,name=management" json:"management,omitempty"` + Management *NodeManagement `protobuf:"bytes,5,opt,name=management,proto3" json:"management,omitempty"` // [Output only] Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // [Output only] The version of the Kubernetes of this node. - Version string `protobuf:"bytes,101,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,101,opt,name=version,proto3" json:"version,omitempty"` // [Output only] The resource URLs of [instance // groups](/compute/docs/instance-groups/) associated with this // node pool. - InstanceGroupUrls []string `protobuf:"bytes,102,rep,name=instance_group_urls,json=instanceGroupUrls" json:"instance_group_urls,omitempty"` + InstanceGroupUrls []string `protobuf:"bytes,102,rep,name=instance_group_urls,json=instanceGroupUrls,proto3" json:"instance_group_urls,omitempty"` // [Output only] The status of the nodes in this pool instance. - Status NodePool_Status `protobuf:"varint,103,opt,name=status,enum=google.container.v1alpha1.NodePool_Status" json:"status,omitempty"` + Status NodePool_Status `protobuf:"varint,103,opt,name=status,proto3,enum=google.container.v1alpha1.NodePool_Status" json:"status,omitempty"` // [Output only] Additional information about the current status of this // node pool instance, if available. - StatusMessage string `protobuf:"bytes,104,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,104,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4075,11 +4075,11 @@ func (m *NodePool) GetStatusMessage() string { // node pool. type NodeManagement struct { // Whether the nodes will be automatically upgraded. - AutoUpgrade bool `protobuf:"varint,1,opt,name=auto_upgrade,json=autoUpgrade" json:"auto_upgrade,omitempty"` + AutoUpgrade bool `protobuf:"varint,1,opt,name=auto_upgrade,json=autoUpgrade,proto3" json:"auto_upgrade,omitempty"` // Whether the nodes will be automatically repaired. - AutoRepair bool `protobuf:"varint,2,opt,name=auto_repair,json=autoRepair" json:"auto_repair,omitempty"` + AutoRepair bool `protobuf:"varint,2,opt,name=auto_repair,json=autoRepair,proto3" json:"auto_repair,omitempty"` // Specifies the Auto Upgrade knobs for the node pool. - UpgradeOptions *AutoUpgradeOptions `protobuf:"bytes,10,opt,name=upgrade_options,json=upgradeOptions" json:"upgrade_options,omitempty"` + UpgradeOptions *AutoUpgradeOptions `protobuf:"bytes,10,opt,name=upgrade_options,json=upgradeOptions,proto3" json:"upgrade_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4136,10 +4136,10 @@ type AutoUpgradeOptions struct { // [Output only] This field is set when upgrades are about to commence // with the approximate start time for the upgrades, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - AutoUpgradeStartTime string `protobuf:"bytes,1,opt,name=auto_upgrade_start_time,json=autoUpgradeStartTime" json:"auto_upgrade_start_time,omitempty"` + AutoUpgradeStartTime string `protobuf:"bytes,1,opt,name=auto_upgrade_start_time,json=autoUpgradeStartTime,proto3" json:"auto_upgrade_start_time,omitempty"` // [Output only] This field is set when upgrades are about to commence // with the description of the upgrade. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4186,7 +4186,7 @@ func (m *AutoUpgradeOptions) GetDescription() string { // MaintenancePolicy defines the maintenance policy to be used for the cluster. type MaintenancePolicy struct { // Specifies the maintenance window in which maintenance may be performed. - Window *MaintenanceWindow `protobuf:"bytes,1,opt,name=window" json:"window,omitempty"` + Window *MaintenanceWindow `protobuf:"bytes,1,opt,name=window,proto3" json:"window,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4265,7 +4265,7 @@ type isMaintenanceWindow_Policy interface { } type MaintenanceWindow_DailyMaintenanceWindow struct { - DailyMaintenanceWindow *DailyMaintenanceWindow `protobuf:"bytes,2,opt,name=daily_maintenance_window,json=dailyMaintenanceWindow,oneof"` + DailyMaintenanceWindow *DailyMaintenanceWindow `protobuf:"bytes,2,opt,name=daily_maintenance_window,json=dailyMaintenanceWindow,proto3,oneof"` } func (*MaintenanceWindow_DailyMaintenanceWindow) isMaintenanceWindow_Policy() {} @@ -4343,10 +4343,10 @@ func _MaintenanceWindow_OneofSizer(msg proto.Message) (n int) { type DailyMaintenanceWindow struct { // Time within the maintenance window to start the maintenance operations. // It must be in format "HH:MM”, where HH : [00-23] and MM : [00-59] GMT. - StartTime string `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime string `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // [Output only] Duration of the time window, automatically chosen to be // smallest possible in the given scenario. - Duration string `protobuf:"bytes,3,opt,name=duration" json:"duration,omitempty"` + Duration string `protobuf:"bytes,3,opt,name=duration,proto3" json:"duration,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4396,24 +4396,24 @@ type SetNodePoolManagementRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to update. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // NodeManagement configuration for the node pool. - Management *NodeManagement `protobuf:"bytes,5,opt,name=management" json:"management,omitempty"` + Management *NodeManagement `protobuf:"bytes,5,opt,name=management,proto3" json:"management,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to set // management properties. Specified in the format // 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4490,24 +4490,24 @@ func (m *SetNodePoolManagementRequest) GetName() string { type SetNodePoolSizeRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to update. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The desired node count for the pool. - NodeCount int32 `protobuf:"varint,5,opt,name=node_count,json=nodeCount" json:"node_count,omitempty"` + NodeCount int32 `protobuf:"varint,5,opt,name=node_count,json=nodeCount,proto3" json:"node_count,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to set // size. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4586,22 +4586,22 @@ type RollbackNodePoolUpgradeRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to rollback. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to rollback. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The name (project, location, cluster, node pool id) of the node poll to // rollback upgrade. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4669,7 +4669,7 @@ func (m *RollbackNodePoolUpgradeRequest) GetName() string { // ListNodePoolsResponse is the result of ListNodePoolsRequest. type ListNodePoolsResponse struct { // A list of node pools for a cluster. - NodePools []*NodePool `protobuf:"bytes,1,rep,name=node_pools,json=nodePools" json:"node_pools,omitempty"` + NodePools []*NodePool `protobuf:"bytes,1,rep,name=node_pools,json=nodePools,proto3" json:"node_pools,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4710,13 +4710,13 @@ func (m *ListNodePoolsResponse) GetNodePools() []*NodePool { // adjust the size of the node pool to the current cluster usage. type NodePoolAutoscaling struct { // Is autoscaling enabled for this node pool. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` // Minimum number of nodes in the NodePool. Must be >= 1 and <= // max_node_count. - MinNodeCount int32 `protobuf:"varint,2,opt,name=min_node_count,json=minNodeCount" json:"min_node_count,omitempty"` + MinNodeCount int32 `protobuf:"varint,2,opt,name=min_node_count,json=minNodeCount,proto3" json:"min_node_count,omitempty"` // Maximum number of nodes in the NodePool. Must be >= min_node_count. There // has to enough quota to scale up the cluster. - MaxNodeCount int32 `protobuf:"varint,3,opt,name=max_node_count,json=maxNodeCount" json:"max_node_count,omitempty"` + MaxNodeCount int32 `protobuf:"varint,3,opt,name=max_node_count,json=maxNodeCount,proto3" json:"max_node_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4774,27 +4774,27 @@ type SetLabelsRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The labels to set for that cluster. - ResourceLabels map[string]string `protobuf:"bytes,4,rep,name=resource_labels,json=resourceLabels" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ResourceLabels map[string]string `protobuf:"bytes,4,rep,name=resource_labels,json=resourceLabels,proto3" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The fingerprint of the previous set of labels for this resource, // used to detect conflicts. The fingerprint is initially generated by // Container Engine and changes after every request to modify or update // labels. You must always provide an up-to-date fingerprint hash when // updating or changing labels. Make a get() request to the // resource to get the latest fingerprint. - LabelFingerprint string `protobuf:"bytes,5,opt,name=label_fingerprint,json=labelFingerprint" json:"label_fingerprint,omitempty"` + LabelFingerprint string `protobuf:"bytes,5,opt,name=label_fingerprint,json=labelFingerprint,proto3" json:"label_fingerprint,omitempty"` // The name (project, location, cluster id) of the cluster to set labels. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4872,20 +4872,20 @@ type SetLegacyAbacRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Whether ABAC authorization will be enabled in the cluster. - Enabled bool `protobuf:"varint,4,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,4,opt,name=enabled,proto3" json:"enabled,omitempty"` // The name (project, location, cluster id) of the cluster to set legacy abac. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4956,18 +4956,18 @@ type StartIPRotationRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster id) of the cluster to start IP rotation. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5030,18 +5030,18 @@ type CompleteIPRotationRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster id) of the cluster to complete IP rotation. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5102,10 +5102,10 @@ func (m *CompleteIPRotationRequest) GetName() string { // AcceleratorConfig represents a Hardware Accelerator request. type AcceleratorConfig struct { // The number of the accelerator cards exposed to an instance. - AcceleratorCount int64 `protobuf:"varint,1,opt,name=accelerator_count,json=acceleratorCount" json:"accelerator_count,omitempty"` + AcceleratorCount int64 `protobuf:"varint,1,opt,name=accelerator_count,json=acceleratorCount,proto3" json:"accelerator_count,omitempty"` // The accelerator type resource name. List of supported accelerators // [here](/compute/docs/gpus/#Introduction) - AcceleratorType string `protobuf:"bytes,2,opt,name=accelerator_type,json=acceleratorType" json:"accelerator_type,omitempty"` + AcceleratorType string `protobuf:"bytes,2,opt,name=accelerator_type,json=acceleratorType,proto3" json:"accelerator_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5154,21 +5154,21 @@ type SetNetworkPolicyRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Configuration options for the NetworkPolicy feature. - NetworkPolicy *NetworkPolicy `protobuf:"bytes,4,opt,name=network_policy,json=networkPolicy" json:"network_policy,omitempty"` + NetworkPolicy *NetworkPolicy `protobuf:"bytes,4,opt,name=network_policy,json=networkPolicy,proto3" json:"network_policy,omitempty"` // The name (project, location, cluster id) of the cluster to set networking // policy. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5237,20 +5237,20 @@ func (m *SetNetworkPolicyRequest) GetName() string { type SetMaintenancePolicyRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The maintenance policy to be set for the cluster. An empty field // clears the existing maintenance policy. - MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,4,opt,name=maintenance_policy,json=maintenancePolicy" json:"maintenance_policy,omitempty"` + MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,4,opt,name=maintenance_policy,json=maintenancePolicy,proto3" json:"maintenance_policy,omitempty"` // The name (project, location, cluster id) of the cluster to set maintenance // policy. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5764,8 +5764,7 @@ func (c *clusterManagerClient) SetMaintenancePolicy(ctx context.Context, in *Set return out, nil } -// Server API for ClusterManager service - +// ClusterManagerServer is the server API for ClusterManager service. type ClusterManagerServer interface { // Lists all clusters owned by a project in either the specified zone or all // zones. diff --git a/vendor/google.golang.org/genproto/googleapis/container/v1beta1/cluster_service.pb.go b/vendor/google.golang.org/genproto/googleapis/container/v1beta1/cluster_service.pb.go index f070e3400..698dd1c12 100644 --- a/vendor/google.golang.org/genproto/googleapis/container/v1beta1/cluster_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/container/v1beta1/cluster_service.pb.go @@ -353,12 +353,12 @@ type NodeConfig struct { // // If unspecified, the default machine type is // `n1-standard-1`. - MachineType string `protobuf:"bytes,1,opt,name=machine_type,json=machineType" json:"machine_type,omitempty"` + MachineType string `protobuf:"bytes,1,opt,name=machine_type,json=machineType,proto3" json:"machine_type,omitempty"` // Size of the disk attached to each node, specified in GB. // The smallest allowed disk size is 10GB. // // If unspecified, the default disk size is 100GB. - DiskSizeGb int32 `protobuf:"varint,2,opt,name=disk_size_gb,json=diskSizeGb" json:"disk_size_gb,omitempty"` + DiskSizeGb int32 `protobuf:"varint,2,opt,name=disk_size_gb,json=diskSizeGb,proto3" json:"disk_size_gb,omitempty"` // The set of Google API scopes to be made available on all of the // node VMs under the "default" service account. // @@ -373,10 +373,10 @@ type NodeConfig struct { // // If unspecified, no scopes are added, unless Cloud Logging or Cloud // Monitoring are enabled, in which case their required scopes will be added. - OauthScopes []string `protobuf:"bytes,3,rep,name=oauth_scopes,json=oauthScopes" json:"oauth_scopes,omitempty"` + OauthScopes []string `protobuf:"bytes,3,rep,name=oauth_scopes,json=oauthScopes,proto3" json:"oauth_scopes,omitempty"` // The Google Cloud Platform Service Account to be used by the node VMs. If // no Service Account is specified, the "default" service account is used. - ServiceAccount string `protobuf:"bytes,9,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount string `protobuf:"bytes,9,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // The metadata key/value pairs assigned to instances in the cluster. // // Keys must conform to the regexp [a-zA-Z0-9-_]+ and be less than 128 bytes @@ -390,10 +390,10 @@ type NodeConfig struct { // that each value's size must be less than or equal to 32 KB. // // The total size of all keys and values must be less than 512 KB. - Metadata map[string]string `protobuf:"bytes,4,rep,name=metadata" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Metadata map[string]string `protobuf:"bytes,4,rep,name=metadata,proto3" json:"metadata,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The image type to use for this node. Note that for a given image type, // the latest version of it will be used. - ImageType string `protobuf:"bytes,5,opt,name=image_type,json=imageType" json:"image_type,omitempty"` + ImageType string `protobuf:"bytes,5,opt,name=image_type,json=imageType,proto3" json:"image_type,omitempty"` // The map of Kubernetes labels (key/value pairs) to be applied to each node. // These will added in addition to any default label(s) that // Kubernetes may apply to the node. @@ -402,39 +402,39 @@ type NodeConfig struct { // and conflicts should be avoided. // For more information, including usage and the valid values, see: // https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/ - Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,6,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The number of local SSD disks to be attached to the node. // // The limit for this value is dependant upon the maximum number of // disks available on a machine per zone. See: // https://cloud.google.com/compute/docs/disks/local-ssd#local_ssd_limits // for more information. - LocalSsdCount int32 `protobuf:"varint,7,opt,name=local_ssd_count,json=localSsdCount" json:"local_ssd_count,omitempty"` + LocalSsdCount int32 `protobuf:"varint,7,opt,name=local_ssd_count,json=localSsdCount,proto3" json:"local_ssd_count,omitempty"` // The list of instance tags applied to all nodes. Tags are used to identify // valid sources or targets for network firewalls and are specified by // the client during cluster or node pool creation. Each tag within the list // must comply with RFC1035. - Tags []string `protobuf:"bytes,8,rep,name=tags" json:"tags,omitempty"` + Tags []string `protobuf:"bytes,8,rep,name=tags,proto3" json:"tags,omitempty"` // Whether the nodes are created as preemptible VM instances. See: // https://cloud.google.com/compute/docs/instances/preemptible for more // inforamtion about preemptible VM instances. - Preemptible bool `protobuf:"varint,10,opt,name=preemptible" json:"preemptible,omitempty"` + Preemptible bool `protobuf:"varint,10,opt,name=preemptible,proto3" json:"preemptible,omitempty"` // A list of hardware accelerators to be attached to each node. // See https://cloud.google.com/compute/docs/gpus for more information about // support for GPUs. - Accelerators []*AcceleratorConfig `protobuf:"bytes,11,rep,name=accelerators" json:"accelerators,omitempty"` + Accelerators []*AcceleratorConfig `protobuf:"bytes,11,rep,name=accelerators,proto3" json:"accelerators,omitempty"` // Minimum CPU platform to be used by this instance. The instance may be // scheduled on the specified or newer CPU platform. Applicable values are the // friendly names of CPU platforms, such as // minCpuPlatform: "Intel Haswell" or // minCpuPlatform: "Intel Sandy Bridge". For more // information, read [how to specify min CPU platform](https://cloud.google.com/compute/docs/instances/specify-min-cpu-platform) - MinCpuPlatform string `protobuf:"bytes,13,opt,name=min_cpu_platform,json=minCpuPlatform" json:"min_cpu_platform,omitempty"` + MinCpuPlatform string `protobuf:"bytes,13,opt,name=min_cpu_platform,json=minCpuPlatform,proto3" json:"min_cpu_platform,omitempty"` // List of kubernetes taints to be applied to each node. // // For more information, including usage and the valid values, see: // https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/ - Taints []*NodeTaint `protobuf:"bytes,15,rep,name=taints" json:"taints,omitempty"` + Taints []*NodeTaint `protobuf:"bytes,15,rep,name=taints,proto3" json:"taints,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -562,11 +562,11 @@ func (m *NodeConfig) GetTaints() []*NodeTaint { // https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/ type NodeTaint struct { // Key for taint. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // Value for taint. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Effect for taint. - Effect NodeTaint_Effect `protobuf:"varint,3,opt,name=effect,enum=google.container.v1beta1.NodeTaint_Effect" json:"effect,omitempty"` + Effect NodeTaint_Effect `protobuf:"varint,3,opt,name=effect,proto3,enum=google.container.v1beta1.NodeTaint_Effect" json:"effect,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -624,24 +624,24 @@ type MasterAuth struct { // The username to use for HTTP basic authentication to the master endpoint. // For clusters v1.6.0 and later, you can disable basic authentication by // providing an empty username. - Username string `protobuf:"bytes,1,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"` // The password to use for HTTP basic authentication to the master endpoint. // Because the master endpoint is open to the Internet, you should create a // strong password. If a password is provided for cluster creation, username // must be non-empty. - Password string `protobuf:"bytes,2,opt,name=password" json:"password,omitempty"` + Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"` // Configuration for client certificate authentication on the cluster. If no // configuration is specified, a client certificate is issued. - ClientCertificateConfig *ClientCertificateConfig `protobuf:"bytes,3,opt,name=client_certificate_config,json=clientCertificateConfig" json:"client_certificate_config,omitempty"` + ClientCertificateConfig *ClientCertificateConfig `protobuf:"bytes,3,opt,name=client_certificate_config,json=clientCertificateConfig,proto3" json:"client_certificate_config,omitempty"` // [Output only] Base64-encoded public certificate that is the root of // trust for the cluster. - ClusterCaCertificate string `protobuf:"bytes,100,opt,name=cluster_ca_certificate,json=clusterCaCertificate" json:"cluster_ca_certificate,omitempty"` + ClusterCaCertificate string `protobuf:"bytes,100,opt,name=cluster_ca_certificate,json=clusterCaCertificate,proto3" json:"cluster_ca_certificate,omitempty"` // [Output only] Base64-encoded public certificate used by clients to // authenticate to the cluster endpoint. - ClientCertificate string `protobuf:"bytes,101,opt,name=client_certificate,json=clientCertificate" json:"client_certificate,omitempty"` + ClientCertificate string `protobuf:"bytes,101,opt,name=client_certificate,json=clientCertificate,proto3" json:"client_certificate,omitempty"` // [Output only] Base64-encoded private key used by clients to authenticate // to the cluster endpoint. - ClientKey string `protobuf:"bytes,102,opt,name=client_key,json=clientKey" json:"client_key,omitempty"` + ClientKey string `protobuf:"bytes,102,opt,name=client_key,json=clientKey,proto3" json:"client_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -716,7 +716,7 @@ func (m *MasterAuth) GetClientKey() string { // Configuration for client certificates on the cluster. type ClientCertificateConfig struct { // Issue a client certificate. - IssueClientCertificate bool `protobuf:"varint,1,opt,name=issue_client_certificate,json=issueClientCertificate" json:"issue_client_certificate,omitempty"` + IssueClientCertificate bool `protobuf:"varint,1,opt,name=issue_client_certificate,json=issueClientCertificate,proto3" json:"issue_client_certificate,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -758,17 +758,17 @@ func (m *ClientCertificateConfig) GetIssueClientCertificate() bool { type AddonsConfig struct { // Configuration for the HTTP (L7) load balancing controller addon, which // makes it easy to set up HTTP load balancers for services in a cluster. - HttpLoadBalancing *HttpLoadBalancing `protobuf:"bytes,1,opt,name=http_load_balancing,json=httpLoadBalancing" json:"http_load_balancing,omitempty"` + HttpLoadBalancing *HttpLoadBalancing `protobuf:"bytes,1,opt,name=http_load_balancing,json=httpLoadBalancing,proto3" json:"http_load_balancing,omitempty"` // Configuration for the horizontal pod autoscaling feature, which // increases or decreases the number of replica pods a replication controller // has based on the resource usage of the existing pods. - HorizontalPodAutoscaling *HorizontalPodAutoscaling `protobuf:"bytes,2,opt,name=horizontal_pod_autoscaling,json=horizontalPodAutoscaling" json:"horizontal_pod_autoscaling,omitempty"` + HorizontalPodAutoscaling *HorizontalPodAutoscaling `protobuf:"bytes,2,opt,name=horizontal_pod_autoscaling,json=horizontalPodAutoscaling,proto3" json:"horizontal_pod_autoscaling,omitempty"` // Configuration for the Kubernetes Dashboard. - KubernetesDashboard *KubernetesDashboard `protobuf:"bytes,3,opt,name=kubernetes_dashboard,json=kubernetesDashboard" json:"kubernetes_dashboard,omitempty"` + KubernetesDashboard *KubernetesDashboard `protobuf:"bytes,3,opt,name=kubernetes_dashboard,json=kubernetesDashboard,proto3" json:"kubernetes_dashboard,omitempty"` // Configuration for NetworkPolicy. This only tracks whether the addon // is enabled or not on the Master, it does not track whether network policy // is enabled for the nodes. - NetworkPolicyConfig *NetworkPolicyConfig `protobuf:"bytes,4,opt,name=network_policy_config,json=networkPolicyConfig" json:"network_policy_config,omitempty"` + NetworkPolicyConfig *NetworkPolicyConfig `protobuf:"bytes,4,opt,name=network_policy_config,json=networkPolicyConfig,proto3" json:"network_policy_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -832,7 +832,7 @@ type HttpLoadBalancing struct { // Whether the HTTP Load Balancing controller is enabled in the cluster. // When enabled, it runs a small pod in the cluster that manages the load // balancers. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -876,7 +876,7 @@ type HorizontalPodAutoscaling struct { // Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. // When enabled, it ensures that a Heapster pod is running in the cluster, // which is also used by the Cloud Monitoring service. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -916,7 +916,7 @@ func (m *HorizontalPodAutoscaling) GetDisabled() bool { // Configuration for the Kubernetes Dashboard. type KubernetesDashboard struct { // Whether the Kubernetes Dashboard is enabled for this cluster. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -958,7 +958,7 @@ func (m *KubernetesDashboard) GetDisabled() bool { // is enabled for the nodes. type NetworkPolicyConfig struct { // Whether NetworkPolicy is enabled for this cluster. - Disabled bool `protobuf:"varint,1,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,1,opt,name=disabled,proto3" json:"disabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1001,10 +1001,10 @@ func (m *NetworkPolicyConfig) GetDisabled() bool { // Google Compute Engine Public IPs and Google Prod IPs. type MasterAuthorizedNetworksConfig struct { // Whether or not master authorized networks is enabled. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` // cidr_blocks define up to 10 external networks that could access // Kubernetes master through HTTPS. - CidrBlocks []*MasterAuthorizedNetworksConfig_CidrBlock `protobuf:"bytes,2,rep,name=cidr_blocks,json=cidrBlocks" json:"cidr_blocks,omitempty"` + CidrBlocks []*MasterAuthorizedNetworksConfig_CidrBlock `protobuf:"bytes,2,rep,name=cidr_blocks,json=cidrBlocks,proto3" json:"cidr_blocks,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1051,9 +1051,9 @@ func (m *MasterAuthorizedNetworksConfig) GetCidrBlocks() []*MasterAuthorizedNetw // CidrBlock contains an optional name and one CIDR block. type MasterAuthorizedNetworksConfig_CidrBlock struct { // display_name is an optional field for users to identify CIDR blocks. - DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,1,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // cidr_block must be specified in CIDR notation. - CidrBlock string `protobuf:"bytes,2,opt,name=cidr_block,json=cidrBlock" json:"cidr_block,omitempty"` + CidrBlock string `protobuf:"bytes,2,opt,name=cidr_block,json=cidrBlock,proto3" json:"cidr_block,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1103,9 +1103,9 @@ func (m *MasterAuthorizedNetworksConfig_CidrBlock) GetCidrBlock() string { // https://kubernetes.io/docs/concepts/services-networking/networkpolicies/ type NetworkPolicy struct { // The selected network policy provider. - Provider NetworkPolicy_Provider `protobuf:"varint,1,opt,name=provider,enum=google.container.v1beta1.NetworkPolicy_Provider" json:"provider,omitempty"` + Provider NetworkPolicy_Provider `protobuf:"varint,1,opt,name=provider,proto3,enum=google.container.v1beta1.NetworkPolicy_Provider" json:"provider,omitempty"` // Whether network policy is enabled on the cluster. - Enabled bool `protobuf:"varint,2,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,2,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1152,21 +1152,21 @@ func (m *NetworkPolicy) GetEnabled() bool { // Configuration for controlling how IPs are allocated in the cluster. type IPAllocationPolicy struct { // Whether alias IPs will be used for pod IPs in the cluster. - UseIpAliases bool `protobuf:"varint,1,opt,name=use_ip_aliases,json=useIpAliases" json:"use_ip_aliases,omitempty"` + UseIpAliases bool `protobuf:"varint,1,opt,name=use_ip_aliases,json=useIpAliases,proto3" json:"use_ip_aliases,omitempty"` // Whether a new subnetwork will be created automatically for the cluster. // // This field is only applicable when `use_ip_aliases` is true. - CreateSubnetwork bool `protobuf:"varint,2,opt,name=create_subnetwork,json=createSubnetwork" json:"create_subnetwork,omitempty"` + CreateSubnetwork bool `protobuf:"varint,2,opt,name=create_subnetwork,json=createSubnetwork,proto3" json:"create_subnetwork,omitempty"` // A custom subnetwork name to be used if `create_subnetwork` is true. If // this field is empty, then an automatic name will be chosen for the new // subnetwork. - SubnetworkName string `protobuf:"bytes,3,opt,name=subnetwork_name,json=subnetworkName" json:"subnetwork_name,omitempty"` + SubnetworkName string `protobuf:"bytes,3,opt,name=subnetwork_name,json=subnetworkName,proto3" json:"subnetwork_name,omitempty"` // This field is deprecated, use cluster_ipv4_cidr_block. - ClusterIpv4Cidr string `protobuf:"bytes,4,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr" json:"cluster_ipv4_cidr,omitempty"` + ClusterIpv4Cidr string `protobuf:"bytes,4,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr,proto3" json:"cluster_ipv4_cidr,omitempty"` // This field is deprecated, use node_ipv4_cidr_block. - NodeIpv4Cidr string `protobuf:"bytes,5,opt,name=node_ipv4_cidr,json=nodeIpv4Cidr" json:"node_ipv4_cidr,omitempty"` + NodeIpv4Cidr string `protobuf:"bytes,5,opt,name=node_ipv4_cidr,json=nodeIpv4Cidr,proto3" json:"node_ipv4_cidr,omitempty"` // This field is deprecated, use services_ipv4_cidr_block. - ServicesIpv4Cidr string `protobuf:"bytes,6,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr" json:"services_ipv4_cidr,omitempty"` + ServicesIpv4Cidr string `protobuf:"bytes,6,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr,proto3" json:"services_ipv4_cidr,omitempty"` // The name of the secondary range to be used for the cluster CIDR // block. The secondary range will be used for pod IP // addresses. This must be an existing secondary range associated @@ -1174,7 +1174,7 @@ type IPAllocationPolicy struct { // // This field is only applicable with use_ip_aliases and // create_subnetwork is false. - ClusterSecondaryRangeName string `protobuf:"bytes,7,opt,name=cluster_secondary_range_name,json=clusterSecondaryRangeName" json:"cluster_secondary_range_name,omitempty"` + ClusterSecondaryRangeName string `protobuf:"bytes,7,opt,name=cluster_secondary_range_name,json=clusterSecondaryRangeName,proto3" json:"cluster_secondary_range_name,omitempty"` // The name of the secondary range to be used as for the services // CIDR block. The secondary range will be used for service // ClusterIPs. This must be an existing secondary range associated @@ -1182,7 +1182,7 @@ type IPAllocationPolicy struct { // // This field is only applicable with use_ip_aliases and // create_subnetwork is false. - ServicesSecondaryRangeName string `protobuf:"bytes,8,opt,name=services_secondary_range_name,json=servicesSecondaryRangeName" json:"services_secondary_range_name,omitempty"` + ServicesSecondaryRangeName string `protobuf:"bytes,8,opt,name=services_secondary_range_name,json=servicesSecondaryRangeName,proto3" json:"services_secondary_range_name,omitempty"` // The IP address range for the cluster pod IPs. If this field is set, then // `cluster.cluster_ipv4_cidr` must be left blank. // @@ -1198,7 +1198,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - ClusterIpv4CidrBlock string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr_block,json=clusterIpv4CidrBlock" json:"cluster_ipv4_cidr_block,omitempty"` + ClusterIpv4CidrBlock string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr_block,json=clusterIpv4CidrBlock,proto3" json:"cluster_ipv4_cidr_block,omitempty"` // The IP address range of the instance IPs in this cluster. // // This is applicable only if `create_subnetwork` is true. @@ -1213,7 +1213,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - NodeIpv4CidrBlock string `protobuf:"bytes,10,opt,name=node_ipv4_cidr_block,json=nodeIpv4CidrBlock" json:"node_ipv4_cidr_block,omitempty"` + NodeIpv4CidrBlock string `protobuf:"bytes,10,opt,name=node_ipv4_cidr_block,json=nodeIpv4CidrBlock,proto3" json:"node_ipv4_cidr_block,omitempty"` // The IP address range of the services IPs in this cluster. If blank, a range // will be automatically chosen with the default size. // @@ -1229,7 +1229,7 @@ type IPAllocationPolicy struct { // notation (e.g. `10.96.0.0/14`) from the RFC-1918 private networks (e.g. // `10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`) to pick a specific range // to use. - ServicesIpv4CidrBlock string `protobuf:"bytes,11,opt,name=services_ipv4_cidr_block,json=servicesIpv4CidrBlock" json:"services_ipv4_cidr_block,omitempty"` + ServicesIpv4CidrBlock string `protobuf:"bytes,11,opt,name=services_ipv4_cidr_block,json=servicesIpv4CidrBlock,proto3" json:"services_ipv4_cidr_block,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1340,7 +1340,7 @@ func (m *IPAllocationPolicy) GetServicesIpv4CidrBlock() string { type PodSecurityPolicyConfig struct { // Enable the PodSecurityPolicy controller for this cluster. If enabled, pods // must be valid under a PodSecurityPolicy to be created. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1385,9 +1385,9 @@ type Cluster struct { // * Lowercase letters, numbers, and hyphens only. // * Must start with a letter. // * Must end with a number or a letter. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // An optional description of this cluster. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // The number of nodes to create in this cluster. You must ensure that your // Compute Engine resource quota // is sufficient for this number of instances. You must also have available @@ -1396,7 +1396,7 @@ type Cluster struct { // "node_pool" object, since this configuration (along with the // "node_config") will be used to create a "NodePool" object with an // auto-generated name. Do not use this and a node_pool at the same time. - InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount" json:"initial_node_count,omitempty"` + InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount,proto3" json:"initial_node_count,omitempty"` // Parameters used in creating the cluster's nodes. // See `nodeConfig` for the description of its properties. // For requests, this field should only be used in lieu of a @@ -1407,121 +1407,121 @@ type Cluster struct { // the first node pool. // // If unspecified, the defaults are used. - NodeConfig *NodeConfig `protobuf:"bytes,4,opt,name=node_config,json=nodeConfig" json:"node_config,omitempty"` + NodeConfig *NodeConfig `protobuf:"bytes,4,opt,name=node_config,json=nodeConfig,proto3" json:"node_config,omitempty"` // The authentication information for accessing the master endpoint. - MasterAuth *MasterAuth `protobuf:"bytes,5,opt,name=master_auth,json=masterAuth" json:"master_auth,omitempty"` + MasterAuth *MasterAuth `protobuf:"bytes,5,opt,name=master_auth,json=masterAuth,proto3" json:"master_auth,omitempty"` // The logging service the cluster should use to write logs. // Currently available options: // // * `logging.googleapis.com` - the Google Cloud Logging service. // * `none` - no logs will be exported from the cluster. // * if left as an empty string,`logging.googleapis.com` will be used. - LoggingService string `protobuf:"bytes,6,opt,name=logging_service,json=loggingService" json:"logging_service,omitempty"` + LoggingService string `protobuf:"bytes,6,opt,name=logging_service,json=loggingService,proto3" json:"logging_service,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * `monitoring.googleapis.com` - the Google Cloud Monitoring service. // * `none` - no metrics will be exported from the cluster. // * if left as an empty string, `monitoring.googleapis.com` will be used. - MonitoringService string `protobuf:"bytes,7,opt,name=monitoring_service,json=monitoringService" json:"monitoring_service,omitempty"` + MonitoringService string `protobuf:"bytes,7,opt,name=monitoring_service,json=monitoringService,proto3" json:"monitoring_service,omitempty"` // The name of the Google Compute Engine // [network](/compute/docs/networks-and-firewalls#networks) to which the // cluster is connected. If left unspecified, the `default` network // will be used. - Network string `protobuf:"bytes,8,opt,name=network" json:"network,omitempty"` + Network string `protobuf:"bytes,8,opt,name=network,proto3" json:"network,omitempty"` // The IP address range of the container pods in this cluster, in // [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) // notation (e.g. `10.96.0.0/14`). Leave blank to have // one automatically chosen or specify a `/14` block in `10.0.0.0/8`. - ClusterIpv4Cidr string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr" json:"cluster_ipv4_cidr,omitempty"` + ClusterIpv4Cidr string `protobuf:"bytes,9,opt,name=cluster_ipv4_cidr,json=clusterIpv4Cidr,proto3" json:"cluster_ipv4_cidr,omitempty"` // Configurations for the various addons available to run in the cluster. - AddonsConfig *AddonsConfig `protobuf:"bytes,10,opt,name=addons_config,json=addonsConfig" json:"addons_config,omitempty"` + AddonsConfig *AddonsConfig `protobuf:"bytes,10,opt,name=addons_config,json=addonsConfig,proto3" json:"addons_config,omitempty"` // The name of the Google Compute Engine // [subnetwork](/compute/docs/subnetworks) to which the // cluster is connected. - Subnetwork string `protobuf:"bytes,11,opt,name=subnetwork" json:"subnetwork,omitempty"` + Subnetwork string `protobuf:"bytes,11,opt,name=subnetwork,proto3" json:"subnetwork,omitempty"` // The node pools associated with this cluster. // This field should not be set if "node_config" or "initial_node_count" are // specified. - NodePools []*NodePool `protobuf:"bytes,12,rep,name=node_pools,json=nodePools" json:"node_pools,omitempty"` + NodePools []*NodePool `protobuf:"bytes,12,rep,name=node_pools,json=nodePools,proto3" json:"node_pools,omitempty"` // The list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. - Locations []string `protobuf:"bytes,13,rep,name=locations" json:"locations,omitempty"` + Locations []string `protobuf:"bytes,13,rep,name=locations,proto3" json:"locations,omitempty"` // Kubernetes alpha features are enabled on this cluster. This includes alpha // API groups (e.g. v1beta1) and features that may not be production ready in // the kubernetes version of the master and nodes. // The cluster has no SLA for uptime and master/node upgrades are disabled. // Alpha enabled clusters are automatically deleted thirty days after // creation. - EnableKubernetesAlpha bool `protobuf:"varint,14,opt,name=enable_kubernetes_alpha,json=enableKubernetesAlpha" json:"enable_kubernetes_alpha,omitempty"` + EnableKubernetesAlpha bool `protobuf:"varint,14,opt,name=enable_kubernetes_alpha,json=enableKubernetesAlpha,proto3" json:"enable_kubernetes_alpha,omitempty"` // Configuration options for the NetworkPolicy feature. - NetworkPolicy *NetworkPolicy `protobuf:"bytes,19,opt,name=network_policy,json=networkPolicy" json:"network_policy,omitempty"` + NetworkPolicy *NetworkPolicy `protobuf:"bytes,19,opt,name=network_policy,json=networkPolicy,proto3" json:"network_policy,omitempty"` // Configuration for cluster IP allocation. - IpAllocationPolicy *IPAllocationPolicy `protobuf:"bytes,20,opt,name=ip_allocation_policy,json=ipAllocationPolicy" json:"ip_allocation_policy,omitempty"` + IpAllocationPolicy *IPAllocationPolicy `protobuf:"bytes,20,opt,name=ip_allocation_policy,json=ipAllocationPolicy,proto3" json:"ip_allocation_policy,omitempty"` // The configuration options for master authorized networks feature. - MasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,22,opt,name=master_authorized_networks_config,json=masterAuthorizedNetworksConfig" json:"master_authorized_networks_config,omitempty"` + MasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,22,opt,name=master_authorized_networks_config,json=masterAuthorizedNetworksConfig,proto3" json:"master_authorized_networks_config,omitempty"` // Configure the maintenance policy for this cluster. - MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,23,opt,name=maintenance_policy,json=maintenancePolicy" json:"maintenance_policy,omitempty"` + MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,23,opt,name=maintenance_policy,json=maintenancePolicy,proto3" json:"maintenance_policy,omitempty"` // Configuration for the PodSecurityPolicy feature. - PodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,25,opt,name=pod_security_policy_config,json=podSecurityPolicyConfig" json:"pod_security_policy_config,omitempty"` + PodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,25,opt,name=pod_security_policy_config,json=podSecurityPolicyConfig,proto3" json:"pod_security_policy_config,omitempty"` // [Output only] Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use location instead. - Zone string `protobuf:"bytes,101,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,101,opt,name=zone,proto3" json:"zone,omitempty"` // [Output only] The IP address of this cluster's master endpoint. // The endpoint can be accessed from the internet at // `https://username:password@endpoint/`. // // See the `masterAuth` property of this resource for username and // password information. - Endpoint string `protobuf:"bytes,102,opt,name=endpoint" json:"endpoint,omitempty"` + Endpoint string `protobuf:"bytes,102,opt,name=endpoint,proto3" json:"endpoint,omitempty"` // The initial Kubernetes version for this cluster. Valid versions are those // found in validMasterVersions returned by getServerConfig. The version can // be upgraded over time; such upgrades are reflected in // currentMasterVersion and currentNodeVersion. - InitialClusterVersion string `protobuf:"bytes,103,opt,name=initial_cluster_version,json=initialClusterVersion" json:"initial_cluster_version,omitempty"` + InitialClusterVersion string `protobuf:"bytes,103,opt,name=initial_cluster_version,json=initialClusterVersion,proto3" json:"initial_cluster_version,omitempty"` // [Output only] The current software version of the master endpoint. - CurrentMasterVersion string `protobuf:"bytes,104,opt,name=current_master_version,json=currentMasterVersion" json:"current_master_version,omitempty"` + CurrentMasterVersion string `protobuf:"bytes,104,opt,name=current_master_version,json=currentMasterVersion,proto3" json:"current_master_version,omitempty"` // [Output only] The current version of the node software components. // If they are currently at multiple versions because they're in the process // of being upgraded, this reflects the minimum version of all nodes. - CurrentNodeVersion string `protobuf:"bytes,105,opt,name=current_node_version,json=currentNodeVersion" json:"current_node_version,omitempty"` + CurrentNodeVersion string `protobuf:"bytes,105,opt,name=current_node_version,json=currentNodeVersion,proto3" json:"current_node_version,omitempty"` // [Output only] The time the cluster was created, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - CreateTime string `protobuf:"bytes,106,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime string `protobuf:"bytes,106,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // [Output only] The current status of this cluster. - Status Cluster_Status `protobuf:"varint,107,opt,name=status,enum=google.container.v1beta1.Cluster_Status" json:"status,omitempty"` + Status Cluster_Status `protobuf:"varint,107,opt,name=status,proto3,enum=google.container.v1beta1.Cluster_Status" json:"status,omitempty"` // [Output only] Additional information about the current status of this // cluster, if available. - StatusMessage string `protobuf:"bytes,108,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,108,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // [Output only] The size of the address space on each node for hosting // containers. This is provisioned from within the `container_ipv4_cidr` // range. - NodeIpv4CidrSize int32 `protobuf:"varint,109,opt,name=node_ipv4_cidr_size,json=nodeIpv4CidrSize" json:"node_ipv4_cidr_size,omitempty"` + NodeIpv4CidrSize int32 `protobuf:"varint,109,opt,name=node_ipv4_cidr_size,json=nodeIpv4CidrSize,proto3" json:"node_ipv4_cidr_size,omitempty"` // [Output only] The IP address range of the Kubernetes services in // this cluster, in // [CIDR](http://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) // notation (e.g. `1.2.3.4/29`). Service addresses are // typically put in the last `/16` from the container CIDR. - ServicesIpv4Cidr string `protobuf:"bytes,110,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr" json:"services_ipv4_cidr,omitempty"` + ServicesIpv4Cidr string `protobuf:"bytes,110,opt,name=services_ipv4_cidr,json=servicesIpv4Cidr,proto3" json:"services_ipv4_cidr,omitempty"` // [Output only] The resource URLs of [instance // groups](/compute/docs/instance-groups/) associated with this // cluster. - InstanceGroupUrls []string `protobuf:"bytes,111,rep,name=instance_group_urls,json=instanceGroupUrls" json:"instance_group_urls,omitempty"` + InstanceGroupUrls []string `protobuf:"bytes,111,rep,name=instance_group_urls,json=instanceGroupUrls,proto3" json:"instance_group_urls,omitempty"` // [Output only] The number of nodes currently in the cluster. - CurrentNodeCount int32 `protobuf:"varint,112,opt,name=current_node_count,json=currentNodeCount" json:"current_node_count,omitempty"` + CurrentNodeCount int32 `protobuf:"varint,112,opt,name=current_node_count,json=currentNodeCount,proto3" json:"current_node_count,omitempty"` // [Output only] The time the cluster will be automatically // deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - ExpireTime string `protobuf:"bytes,113,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime string `protobuf:"bytes,113,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/regions-zones/regions-zones#available) or // [region](/compute/docs/regions-zones/regions-zones#available) in which // the cluster resides. - Location string `protobuf:"bytes,114,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,114,opt,name=location,proto3" json:"location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1796,28 +1796,28 @@ type ClusterUpdate struct { // The Kubernetes version to change the nodes to (typically an // upgrade). Use `-` to upgrade to the latest version supported by // the server. - DesiredNodeVersion string `protobuf:"bytes,4,opt,name=desired_node_version,json=desiredNodeVersion" json:"desired_node_version,omitempty"` + DesiredNodeVersion string `protobuf:"bytes,4,opt,name=desired_node_version,json=desiredNodeVersion,proto3" json:"desired_node_version,omitempty"` // The monitoring service the cluster should use to write metrics. // Currently available options: // // * "monitoring.googleapis.com" - the Google Cloud Monitoring service // * "none" - no metrics will be exported from the cluster - DesiredMonitoringService string `protobuf:"bytes,5,opt,name=desired_monitoring_service,json=desiredMonitoringService" json:"desired_monitoring_service,omitempty"` + DesiredMonitoringService string `protobuf:"bytes,5,opt,name=desired_monitoring_service,json=desiredMonitoringService,proto3" json:"desired_monitoring_service,omitempty"` // Configurations for the various addons available to run in the cluster. - DesiredAddonsConfig *AddonsConfig `protobuf:"bytes,6,opt,name=desired_addons_config,json=desiredAddonsConfig" json:"desired_addons_config,omitempty"` + DesiredAddonsConfig *AddonsConfig `protobuf:"bytes,6,opt,name=desired_addons_config,json=desiredAddonsConfig,proto3" json:"desired_addons_config,omitempty"` // The node pool to be upgraded. This field is mandatory if // "desired_node_version", "desired_image_family" or // "desired_node_pool_autoscaling" is specified and there is more than one // node pool on the cluster. - DesiredNodePoolId string `protobuf:"bytes,7,opt,name=desired_node_pool_id,json=desiredNodePoolId" json:"desired_node_pool_id,omitempty"` + DesiredNodePoolId string `protobuf:"bytes,7,opt,name=desired_node_pool_id,json=desiredNodePoolId,proto3" json:"desired_node_pool_id,omitempty"` // The desired image type for the node pool. // NOTE: Set the "desired_node_pool" field as well. - DesiredImageType string `protobuf:"bytes,8,opt,name=desired_image_type,json=desiredImageType" json:"desired_image_type,omitempty"` + DesiredImageType string `protobuf:"bytes,8,opt,name=desired_image_type,json=desiredImageType,proto3" json:"desired_image_type,omitempty"` // Autoscaler configuration for the node pool specified in // desired_node_pool_id. If there is only one pool in the // cluster and desired_node_pool_id is not provided then // the change applies to that single node pool. - DesiredNodePoolAutoscaling *NodePoolAutoscaling `protobuf:"bytes,9,opt,name=desired_node_pool_autoscaling,json=desiredNodePoolAutoscaling" json:"desired_node_pool_autoscaling,omitempty"` + DesiredNodePoolAutoscaling *NodePoolAutoscaling `protobuf:"bytes,9,opt,name=desired_node_pool_autoscaling,json=desiredNodePoolAutoscaling,proto3" json:"desired_node_pool_autoscaling,omitempty"` // The desired list of Google Compute Engine // [locations](/compute/docs/zones#available) in which the cluster's nodes // should be located. Changing the locations a cluster is in will result @@ -1825,15 +1825,15 @@ type ClusterUpdate struct { // whether locations are being added or removed. // // This list must always include the cluster's primary zone. - DesiredLocations []string `protobuf:"bytes,10,rep,name=desired_locations,json=desiredLocations" json:"desired_locations,omitempty"` + DesiredLocations []string `protobuf:"bytes,10,rep,name=desired_locations,json=desiredLocations,proto3" json:"desired_locations,omitempty"` // The desired configuration options for master authorized networks feature. - DesiredMasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,12,opt,name=desired_master_authorized_networks_config,json=desiredMasterAuthorizedNetworksConfig" json:"desired_master_authorized_networks_config,omitempty"` + DesiredMasterAuthorizedNetworksConfig *MasterAuthorizedNetworksConfig `protobuf:"bytes,12,opt,name=desired_master_authorized_networks_config,json=desiredMasterAuthorizedNetworksConfig,proto3" json:"desired_master_authorized_networks_config,omitempty"` // The desired configuration options for the PodSecurityPolicy feature. - DesiredPodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,14,opt,name=desired_pod_security_policy_config,json=desiredPodSecurityPolicyConfig" json:"desired_pod_security_policy_config,omitempty"` + DesiredPodSecurityPolicyConfig *PodSecurityPolicyConfig `protobuf:"bytes,14,opt,name=desired_pod_security_policy_config,json=desiredPodSecurityPolicyConfig,proto3" json:"desired_pod_security_policy_config,omitempty"` // The Kubernetes version to change the master to. The only valid value is the // latest supported version. Use "-" to have the server automatically select // the latest version. - DesiredMasterVersion string `protobuf:"bytes,100,opt,name=desired_master_version,json=desiredMasterVersion" json:"desired_master_version,omitempty"` + DesiredMasterVersion string `protobuf:"bytes,100,opt,name=desired_master_version,json=desiredMasterVersion,proto3" json:"desired_master_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1937,35 +1937,35 @@ func (m *ClusterUpdate) GetDesiredMasterVersion() string { // happening on the cluster. All fields are output only. type Operation struct { // The server-assigned ID for the operation. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the operation // is taking place. // This field is deprecated, use location instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The operation type. - OperationType Operation_Type `protobuf:"varint,3,opt,name=operation_type,json=operationType,enum=google.container.v1beta1.Operation_Type" json:"operation_type,omitempty"` + OperationType Operation_Type `protobuf:"varint,3,opt,name=operation_type,json=operationType,proto3,enum=google.container.v1beta1.Operation_Type" json:"operation_type,omitempty"` // The current status of the operation. - Status Operation_Status `protobuf:"varint,4,opt,name=status,enum=google.container.v1beta1.Operation_Status" json:"status,omitempty"` + Status Operation_Status `protobuf:"varint,4,opt,name=status,proto3,enum=google.container.v1beta1.Operation_Status" json:"status,omitempty"` // Detailed operation progress, if available. - Detail string `protobuf:"bytes,8,opt,name=detail" json:"detail,omitempty"` + Detail string `protobuf:"bytes,8,opt,name=detail,proto3" json:"detail,omitempty"` // If an error has occurred, a textual description of the error. - StatusMessage string `protobuf:"bytes,5,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,5,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,6,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,6,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // Server-defined URL for the target of the operation. - TargetLink string `protobuf:"bytes,7,opt,name=target_link,json=targetLink" json:"target_link,omitempty"` + TargetLink string `protobuf:"bytes,7,opt,name=target_link,json=targetLink,proto3" json:"target_link,omitempty"` // [Output only] The name of the Google Compute Engine // [zone](/compute/docs/regions-zones/regions-zones#available) or // [region](/compute/docs/regions-zones/regions-zones#available) in which // the cluster resides. - Location string `protobuf:"bytes,9,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,9,opt,name=location,proto3" json:"location,omitempty"` // [Output only] The time the operation started, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - StartTime string `protobuf:"bytes,10,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime string `protobuf:"bytes,10,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // [Output only] The time the operation completed, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - EndTime string `protobuf:"bytes,11,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime string `protobuf:"bytes,11,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2077,18 +2077,18 @@ type CreateClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // A [cluster // resource](/container-engine/reference/rest/v1beta1/projects.zones.clusters) - Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster" json:"cluster,omitempty"` + Cluster *Cluster `protobuf:"bytes,3,opt,name=cluster,proto3" json:"cluster,omitempty"` // The parent (project and location) where the cluster will be created. // Specified in the format 'projects/*/locations/*'. - Parent string `protobuf:"bytes,5,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,5,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2151,18 +2151,18 @@ type GetClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to retrieve. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster) of the cluster to retrieve. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2225,20 +2225,20 @@ type UpdateClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // A description of the update. - Update *ClusterUpdate `protobuf:"bytes,4,opt,name=update" json:"update,omitempty"` + Update *ClusterUpdate `protobuf:"bytes,4,opt,name=update,proto3" json:"update,omitempty"` // The name (project, location, cluster) of the cluster to update. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2308,22 +2308,22 @@ type SetMasterAuthRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to upgrade. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The exact form of action to be taken on the master auth. - Action SetMasterAuthRequest_Action `protobuf:"varint,4,opt,name=action,enum=google.container.v1beta1.SetMasterAuthRequest_Action" json:"action,omitempty"` + Action SetMasterAuthRequest_Action `protobuf:"varint,4,opt,name=action,proto3,enum=google.container.v1beta1.SetMasterAuthRequest_Action" json:"action,omitempty"` // A description of the update. - Update *MasterAuth `protobuf:"bytes,5,opt,name=update" json:"update,omitempty"` + Update *MasterAuth `protobuf:"bytes,5,opt,name=update,proto3" json:"update,omitempty"` // The name (project, location, cluster) of the cluster to set auth. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2400,18 +2400,18 @@ type DeleteClusterRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to delete. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster) of the cluster to delete. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2474,16 +2474,16 @@ type ListClustersRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides, or "-" for all zones. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The parent (project and location) where the clusters will be listed. // Specified in the format 'projects/*/locations/*'. // Location "-" matches all zones and all regions. - Parent string `protobuf:"bytes,4,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,4,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2538,10 +2538,10 @@ func (m *ListClustersRequest) GetParent() string { type ListClustersResponse struct { // A list of clusters in the project in the specified zone, or // across all ones. - Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters" json:"clusters,omitempty"` + Clusters []*Cluster `protobuf:"bytes,1,rep,name=clusters,proto3" json:"clusters,omitempty"` // If any zones are listed here, the list of clusters returned // may be missing those zones. - MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones" json:"missing_zones,omitempty"` + MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones,proto3" json:"missing_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2590,18 +2590,18 @@ type GetOperationRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The server-assigned `name` of the operation. // This field is deprecated, use name instead. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // The name (project, location, operation id) of the operation to get. // Specified in the format 'projects/*/locations/*/operations/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2664,15 +2664,15 @@ type ListOperationsRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine [zone](/compute/docs/zones#available) // to return operations for, or `-` for all zones. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The parent (project and location) where the operations will be listed. // Specified in the format 'projects/*/locations/*'. // Location "-" matches all zones and all regions. - Parent string `protobuf:"bytes,4,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,4,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2728,17 +2728,17 @@ type CancelOperationRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the operation resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The server-assigned `name` of the operation. // This field is deprecated, use name instead. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // The name (project, location, operation id) of the operation to cancel. // Specified in the format 'projects/*/locations/*/operations/*'. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2799,10 +2799,10 @@ func (m *CancelOperationRequest) GetName() string { // ListOperationsResponse is the result of ListOperationsRequest. type ListOperationsResponse struct { // A list of operations in the project in the specified zone. - Operations []*Operation `protobuf:"bytes,1,rep,name=operations" json:"operations,omitempty"` + Operations []*Operation `protobuf:"bytes,1,rep,name=operations,proto3" json:"operations,omitempty"` // If any zones are listed here, the list of operations returned // may be missing the operations from those zones. - MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones" json:"missing_zones,omitempty"` + MissingZones []string `protobuf:"bytes,2,rep,name=missing_zones,json=missingZones,proto3" json:"missing_zones,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2851,14 +2851,14 @@ type GetServerConfigRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine [zone](/compute/docs/zones#available) // to return operations for. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name (project and location) of the server config to get // Specified in the format 'projects/*/locations/*'. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2912,15 +2912,15 @@ func (m *GetServerConfigRequest) GetName() string { // Container Engine service configuration. type ServerConfig struct { // Version of Kubernetes the service deploys by default. - DefaultClusterVersion string `protobuf:"bytes,1,opt,name=default_cluster_version,json=defaultClusterVersion" json:"default_cluster_version,omitempty"` + DefaultClusterVersion string `protobuf:"bytes,1,opt,name=default_cluster_version,json=defaultClusterVersion,proto3" json:"default_cluster_version,omitempty"` // List of valid node upgrade target versions. - ValidNodeVersions []string `protobuf:"bytes,3,rep,name=valid_node_versions,json=validNodeVersions" json:"valid_node_versions,omitempty"` + ValidNodeVersions []string `protobuf:"bytes,3,rep,name=valid_node_versions,json=validNodeVersions,proto3" json:"valid_node_versions,omitempty"` // Default image type. - DefaultImageType string `protobuf:"bytes,4,opt,name=default_image_type,json=defaultImageType" json:"default_image_type,omitempty"` + DefaultImageType string `protobuf:"bytes,4,opt,name=default_image_type,json=defaultImageType,proto3" json:"default_image_type,omitempty"` // List of valid image types. - ValidImageTypes []string `protobuf:"bytes,5,rep,name=valid_image_types,json=validImageTypes" json:"valid_image_types,omitempty"` + ValidImageTypes []string `protobuf:"bytes,5,rep,name=valid_image_types,json=validImageTypes,proto3" json:"valid_image_types,omitempty"` // List of valid master versions. - ValidMasterVersions []string `protobuf:"bytes,6,rep,name=valid_master_versions,json=validMasterVersions" json:"valid_master_versions,omitempty"` + ValidMasterVersions []string `protobuf:"bytes,6,rep,name=valid_master_versions,json=validMasterVersions,proto3" json:"valid_master_versions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2990,20 +2990,20 @@ type CreateNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use parent instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The node pool to create. - NodePool *NodePool `protobuf:"bytes,4,opt,name=node_pool,json=nodePool" json:"node_pool,omitempty"` + NodePool *NodePool `protobuf:"bytes,4,opt,name=node_pool,json=nodePool,proto3" json:"node_pool,omitempty"` // The parent (project, location, cluster id) where the node pool will be created. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Parent string `protobuf:"bytes,6,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,6,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3073,21 +3073,21 @@ type DeleteNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to delete. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to delete. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3157,18 +3157,18 @@ type ListNodePoolsRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use parent instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use parent instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use parent instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The parent (project, location, cluster id) where the node pools will be listed. // Specified in the format 'projects/*/locations/*/clusters/*'. - Parent string `protobuf:"bytes,5,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,5,opt,name=parent,proto3" json:"parent,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3231,21 +3231,21 @@ type GetNodePoolRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to get. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3318,32 +3318,32 @@ func (m *GetNodePoolRequest) GetName() string { // the workload. type NodePool struct { // The name of the node pool. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The node configuration of the pool. - Config *NodeConfig `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"` + Config *NodeConfig `protobuf:"bytes,2,opt,name=config,proto3" json:"config,omitempty"` // The initial node count for the pool. You must ensure that your // Compute Engine resource quota // is sufficient for this number of instances. You must also have available // firewall and routes quota. - InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount" json:"initial_node_count,omitempty"` + InitialNodeCount int32 `protobuf:"varint,3,opt,name=initial_node_count,json=initialNodeCount,proto3" json:"initial_node_count,omitempty"` // [Output only] Server-defined URL for the resource. - SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink" json:"self_link,omitempty"` + SelfLink string `protobuf:"bytes,100,opt,name=self_link,json=selfLink,proto3" json:"self_link,omitempty"` // [Output only] The version of the Kubernetes of this node. - Version string `protobuf:"bytes,101,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,101,opt,name=version,proto3" json:"version,omitempty"` // [Output only] The resource URLs of [instance // groups](/compute/docs/instance-groups/) associated with this // node pool. - InstanceGroupUrls []string `protobuf:"bytes,102,rep,name=instance_group_urls,json=instanceGroupUrls" json:"instance_group_urls,omitempty"` + InstanceGroupUrls []string `protobuf:"bytes,102,rep,name=instance_group_urls,json=instanceGroupUrls,proto3" json:"instance_group_urls,omitempty"` // [Output only] The status of the nodes in this pool instance. - Status NodePool_Status `protobuf:"varint,103,opt,name=status,enum=google.container.v1beta1.NodePool_Status" json:"status,omitempty"` + Status NodePool_Status `protobuf:"varint,103,opt,name=status,proto3,enum=google.container.v1beta1.NodePool_Status" json:"status,omitempty"` // [Output only] Additional information about the current status of this // node pool instance, if available. - StatusMessage string `protobuf:"bytes,104,opt,name=status_message,json=statusMessage" json:"status_message,omitempty"` + StatusMessage string `protobuf:"bytes,104,opt,name=status_message,json=statusMessage,proto3" json:"status_message,omitempty"` // Autoscaler configuration for this NodePool. Autoscaler is enabled // only if a valid configuration is present. - Autoscaling *NodePoolAutoscaling `protobuf:"bytes,4,opt,name=autoscaling" json:"autoscaling,omitempty"` + Autoscaling *NodePoolAutoscaling `protobuf:"bytes,4,opt,name=autoscaling,proto3" json:"autoscaling,omitempty"` // NodeManagement configuration for this NodePool. - Management *NodeManagement `protobuf:"bytes,5,opt,name=management" json:"management,omitempty"` + Management *NodeManagement `protobuf:"bytes,5,opt,name=management,proto3" json:"management,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3447,11 +3447,11 @@ func (m *NodePool) GetManagement() *NodeManagement { // node pool. type NodeManagement struct { // Whether the nodes will be automatically upgraded. - AutoUpgrade bool `protobuf:"varint,1,opt,name=auto_upgrade,json=autoUpgrade" json:"auto_upgrade,omitempty"` + AutoUpgrade bool `protobuf:"varint,1,opt,name=auto_upgrade,json=autoUpgrade,proto3" json:"auto_upgrade,omitempty"` // Whether the nodes will be automatically repaired. - AutoRepair bool `protobuf:"varint,2,opt,name=auto_repair,json=autoRepair" json:"auto_repair,omitempty"` + AutoRepair bool `protobuf:"varint,2,opt,name=auto_repair,json=autoRepair,proto3" json:"auto_repair,omitempty"` // Specifies the Auto Upgrade knobs for the node pool. - UpgradeOptions *AutoUpgradeOptions `protobuf:"bytes,10,opt,name=upgrade_options,json=upgradeOptions" json:"upgrade_options,omitempty"` + UpgradeOptions *AutoUpgradeOptions `protobuf:"bytes,10,opt,name=upgrade_options,json=upgradeOptions,proto3" json:"upgrade_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3508,10 +3508,10 @@ type AutoUpgradeOptions struct { // [Output only] This field is set when upgrades are about to commence // with the approximate start time for the upgrades, in // [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. - AutoUpgradeStartTime string `protobuf:"bytes,1,opt,name=auto_upgrade_start_time,json=autoUpgradeStartTime" json:"auto_upgrade_start_time,omitempty"` + AutoUpgradeStartTime string `protobuf:"bytes,1,opt,name=auto_upgrade_start_time,json=autoUpgradeStartTime,proto3" json:"auto_upgrade_start_time,omitempty"` // [Output only] This field is set when upgrades are about to commence // with the description of the upgrade. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3558,7 +3558,7 @@ func (m *AutoUpgradeOptions) GetDescription() string { // MaintenancePolicy defines the maintenance policy to be used for the cluster. type MaintenancePolicy struct { // Specifies the maintenance window in which maintenance may be performed. - Window *MaintenanceWindow `protobuf:"bytes,1,opt,name=window" json:"window,omitempty"` + Window *MaintenanceWindow `protobuf:"bytes,1,opt,name=window,proto3" json:"window,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3637,7 +3637,7 @@ type isMaintenanceWindow_Policy interface { } type MaintenanceWindow_DailyMaintenanceWindow struct { - DailyMaintenanceWindow *DailyMaintenanceWindow `protobuf:"bytes,2,opt,name=daily_maintenance_window,json=dailyMaintenanceWindow,oneof"` + DailyMaintenanceWindow *DailyMaintenanceWindow `protobuf:"bytes,2,opt,name=daily_maintenance_window,json=dailyMaintenanceWindow,proto3,oneof"` } func (*MaintenanceWindow_DailyMaintenanceWindow) isMaintenanceWindow_Policy() {} @@ -3715,10 +3715,10 @@ func _MaintenanceWindow_OneofSizer(msg proto.Message) (n int) { type DailyMaintenanceWindow struct { // Time within the maintenance window to start the maintenance operations. // It must be in format "HH:MM”, where HH : [00-23] and MM : [00-59] GMT. - StartTime string `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime string `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // [Output only] Duration of the time window, automatically chosen to be // smallest possible in the given scenario. - Duration string `protobuf:"bytes,3,opt,name=duration" json:"duration,omitempty"` + Duration string `protobuf:"bytes,3,opt,name=duration,proto3" json:"duration,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3768,24 +3768,24 @@ type SetNodePoolManagementRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to update. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // NodeManagement configuration for the node pool. - Management *NodeManagement `protobuf:"bytes,5,opt,name=management" json:"management,omitempty"` + Management *NodeManagement `protobuf:"bytes,5,opt,name=management,proto3" json:"management,omitempty"` // The name (project, location, cluster, node pool id) of the node pool to set // management properties. Specified in the format // 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3864,22 +3864,22 @@ type RollbackNodePoolUpgradeRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to rollback. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name of the node pool to rollback. // This field is deprecated, use name instead. - NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId" json:"node_pool_id,omitempty"` + NodePoolId string `protobuf:"bytes,4,opt,name=node_pool_id,json=nodePoolId,proto3" json:"node_pool_id,omitempty"` // The name (project, location, cluster, node pool id) of the node poll to // rollback upgrade. // Specified in the format 'projects/*/locations/*/clusters/*/nodePools/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3947,7 +3947,7 @@ func (m *RollbackNodePoolUpgradeRequest) GetName() string { // ListNodePoolsResponse is the result of ListNodePoolsRequest. type ListNodePoolsResponse struct { // A list of node pools for a cluster. - NodePools []*NodePool `protobuf:"bytes,1,rep,name=node_pools,json=nodePools" json:"node_pools,omitempty"` + NodePools []*NodePool `protobuf:"bytes,1,rep,name=node_pools,json=nodePools,proto3" json:"node_pools,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3988,13 +3988,13 @@ func (m *ListNodePoolsResponse) GetNodePools() []*NodePool { // adjust the size of the node pool to the current cluster usage. type NodePoolAutoscaling struct { // Is autoscaling enabled for this node pool. - Enabled bool `protobuf:"varint,1,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,1,opt,name=enabled,proto3" json:"enabled,omitempty"` // Minimum number of nodes in the NodePool. Must be >= 1 and <= // max_node_count. - MinNodeCount int32 `protobuf:"varint,2,opt,name=min_node_count,json=minNodeCount" json:"min_node_count,omitempty"` + MinNodeCount int32 `protobuf:"varint,2,opt,name=min_node_count,json=minNodeCount,proto3" json:"min_node_count,omitempty"` // Maximum number of nodes in the NodePool. Must be >= min_node_count. There // has to enough quota to scale up the cluster. - MaxNodeCount int32 `protobuf:"varint,3,opt,name=max_node_count,json=maxNodeCount" json:"max_node_count,omitempty"` + MaxNodeCount int32 `protobuf:"varint,3,opt,name=max_node_count,json=maxNodeCount,proto3" json:"max_node_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4052,27 +4052,27 @@ type SetLabelsRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The labels to set for that cluster. - ResourceLabels map[string]string `protobuf:"bytes,4,rep,name=resource_labels,json=resourceLabels" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ResourceLabels map[string]string `protobuf:"bytes,4,rep,name=resource_labels,json=resourceLabels,proto3" json:"resource_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The fingerprint of the previous set of labels for this resource, // used to detect conflicts. The fingerprint is initially generated by // Container Engine and changes after every request to modify or update // labels. You must always provide an up-to-date fingerprint hash when // updating or changing labels. Make a get() request to the // resource to get the latest fingerprint. - LabelFingerprint string `protobuf:"bytes,5,opt,name=label_fingerprint,json=labelFingerprint" json:"label_fingerprint,omitempty"` + LabelFingerprint string `protobuf:"bytes,5,opt,name=label_fingerprint,json=labelFingerprint,proto3" json:"label_fingerprint,omitempty"` // The name (project, location, cluster id) of the cluster to set labels. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4150,20 +4150,20 @@ type SetLegacyAbacRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Whether ABAC authorization will be enabled in the cluster. - Enabled bool `protobuf:"varint,4,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,4,opt,name=enabled,proto3" json:"enabled,omitempty"` // The name (project, location, cluster id) of the cluster to set legacy abac. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4234,18 +4234,18 @@ type StartIPRotationRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster id) of the cluster to start IP rotation. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4308,18 +4308,18 @@ type CompleteIPRotationRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The name (project, location, cluster id) of the cluster to complete IP rotation. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4380,10 +4380,10 @@ func (m *CompleteIPRotationRequest) GetName() string { // AcceleratorConfig represents a Hardware Accelerator request. type AcceleratorConfig struct { // The number of the accelerator cards exposed to an instance. - AcceleratorCount int64 `protobuf:"varint,1,opt,name=accelerator_count,json=acceleratorCount" json:"accelerator_count,omitempty"` + AcceleratorCount int64 `protobuf:"varint,1,opt,name=accelerator_count,json=acceleratorCount,proto3" json:"accelerator_count,omitempty"` // The accelerator type resource name. List of supported accelerators // [here](/compute/docs/gpus/#Introduction) - AcceleratorType string `protobuf:"bytes,2,opt,name=accelerator_type,json=acceleratorType" json:"accelerator_type,omitempty"` + AcceleratorType string `protobuf:"bytes,2,opt,name=accelerator_type,json=acceleratorType,proto3" json:"accelerator_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4432,20 +4432,20 @@ type SetNetworkPolicyRequest struct { // The Google Developers Console [project ID or project // number](https://developers.google.com/console/help/new/#projectnumber). // This field is deprecated, use name instead. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. // This field is deprecated, use name instead. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster. // This field is deprecated, use name instead. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // Configuration options for the NetworkPolicy feature. - NetworkPolicy *NetworkPolicy `protobuf:"bytes,4,opt,name=network_policy,json=networkPolicy" json:"network_policy,omitempty"` + NetworkPolicy *NetworkPolicy `protobuf:"bytes,4,opt,name=network_policy,json=networkPolicy,proto3" json:"network_policy,omitempty"` // The name (project, location, cluster id) of the cluster to set networking policy. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4514,20 +4514,20 @@ func (m *SetNetworkPolicyRequest) GetName() string { type SetMaintenancePolicyRequest struct { // The Google Developers Console [project ID or project // number](https://support.google.com/cloud/answer/6158840). - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the Google Compute Engine // [zone](/compute/docs/zones#available) in which the cluster // resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The name of the cluster to update. - ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId" json:"cluster_id,omitempty"` + ClusterId string `protobuf:"bytes,3,opt,name=cluster_id,json=clusterId,proto3" json:"cluster_id,omitempty"` // The maintenance policy to be set for the cluster. An empty field // clears the existing maintenance policy. - MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,4,opt,name=maintenance_policy,json=maintenancePolicy" json:"maintenance_policy,omitempty"` + MaintenancePolicy *MaintenancePolicy `protobuf:"bytes,4,opt,name=maintenance_policy,json=maintenancePolicy,proto3" json:"maintenance_policy,omitempty"` // The name (project, location, cluster id) of the cluster to set maintenance // policy. // Specified in the format 'projects/*/locations/*/clusters/*'. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4945,8 +4945,7 @@ func (c *clusterManagerClient) SetMaintenancePolicy(ctx context.Context, in *Set return out, nil } -// Server API for ClusterManager service - +// ClusterManagerServer is the server API for ClusterManager service. type ClusterManagerServer interface { // Lists all clusters owned by a project in either the specified zone or all // zones. diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1/datastore_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1/datastore_admin.pb.go index 68b43044a..1c495d01e 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1/datastore_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1/datastore_admin.pb.go @@ -111,17 +111,17 @@ func (CommonMetadata_State) EnumDescriptor() ([]byte, []int) { // Metadata common to all Datastore Admin operations. type CommonMetadata struct { // The time that work began on the operation. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time the operation ended, either successfully or otherwise. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // The type of the operation. Can be used as a filter in // ListOperationsRequest. - OperationType OperationType `protobuf:"varint,3,opt,name=operation_type,json=operationType,enum=google.datastore.admin.v1.OperationType" json:"operation_type,omitempty"` + OperationType OperationType `protobuf:"varint,3,opt,name=operation_type,json=operationType,proto3,enum=google.datastore.admin.v1.OperationType" json:"operation_type,omitempty"` // The client-assigned labels which were provided when the operation was // created. May also include additional labels. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The current state of the Operation. - State CommonMetadata_State `protobuf:"varint,5,opt,name=state,enum=google.datastore.admin.v1.CommonMetadata_State" json:"state,omitempty"` + State CommonMetadata_State `protobuf:"varint,5,opt,name=state,proto3,enum=google.datastore.admin.v1.CommonMetadata_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -190,10 +190,10 @@ func (m *CommonMetadata) GetState() CommonMetadata_State { type Progress struct { // The amount of work that has been completed. Note that this may be greater // than work_estimated. - WorkCompleted int64 `protobuf:"varint,1,opt,name=work_completed,json=workCompleted" json:"work_completed,omitempty"` + WorkCompleted int64 `protobuf:"varint,1,opt,name=work_completed,json=workCompleted,proto3" json:"work_completed,omitempty"` // An estimate of how much work needs to be performed. May be zero if the // work estimate is unavailable. - WorkEstimated int64 `protobuf:"varint,2,opt,name=work_estimated,json=workEstimated" json:"work_estimated,omitempty"` + WorkEstimated int64 `protobuf:"varint,2,opt,name=work_estimated,json=workEstimated,proto3" json:"work_estimated,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -241,11 +241,11 @@ func (m *Progress) GetWorkEstimated() int64 { // [google.datastore.admin.v1.DatastoreAdmin.ExportEntities][google.datastore.admin.v1.DatastoreAdmin.ExportEntities]. type ExportEntitiesRequest struct { // Project ID against which to make the request. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Client-assigned labels. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Description of what data from the project is included in the export. - EntityFilter *EntityFilter `protobuf:"bytes,3,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,3,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` // Location for the export metadata and data files. // // The full resource URL of the external storage location. Currently, only @@ -264,7 +264,7 @@ type ExportEntitiesRequest struct { // // By nesting the data files deeper, the same Cloud Storage bucket can be used // in multiple ExportEntities operations without conflict. - OutputUrlPrefix string `protobuf:"bytes,4,opt,name=output_url_prefix,json=outputUrlPrefix" json:"output_url_prefix,omitempty"` + OutputUrlPrefix string `protobuf:"bytes,4,opt,name=output_url_prefix,json=outputUrlPrefix,proto3" json:"output_url_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -326,9 +326,9 @@ func (m *ExportEntitiesRequest) GetOutputUrlPrefix() string { // [google.datastore.admin.v1.DatastoreAdmin.ImportEntities][google.datastore.admin.v1.DatastoreAdmin.ImportEntities]. type ImportEntitiesRequest struct { // Project ID against which to make the request. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Client-assigned labels. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The full resource URL of the external storage location. Currently, only // Google Cloud Storage is supported. So input_url should be of the form: // `gs://BUCKET_NAME[/NAMESPACE_PATH]/OVERALL_EXPORT_METADATA_FILE`, where @@ -342,12 +342,12 @@ type ImportEntitiesRequest struct { // // For more information, see // [google.datastore.admin.v1.ExportEntitiesResponse.output_url][google.datastore.admin.v1.ExportEntitiesResponse.output_url]. - InputUrl string `protobuf:"bytes,3,opt,name=input_url,json=inputUrl" json:"input_url,omitempty"` + InputUrl string `protobuf:"bytes,3,opt,name=input_url,json=inputUrl,proto3" json:"input_url,omitempty"` // Optionally specify which kinds/namespaces are to be imported. If provided, // the list must be a subset of the EntityFilter used in creating the export, // otherwise a FAILED_PRECONDITION error will be returned. If no filter is // specified then all entities from the export are imported. - EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -412,7 +412,7 @@ type ExportEntitiesResponse struct { // into Cloud Datastore (this project or another project). See // [google.datastore.admin.v1.ImportEntitiesRequest.input_url][google.datastore.admin.v1.ImportEntitiesRequest.input_url]. // Only present if the operation completed successfully. - OutputUrl string `protobuf:"bytes,1,opt,name=output_url,json=outputUrl" json:"output_url,omitempty"` + OutputUrl string `protobuf:"bytes,1,opt,name=output_url,json=outputUrl,proto3" json:"output_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -452,19 +452,19 @@ func (m *ExportEntitiesResponse) GetOutputUrl() string { // Metadata for ExportEntities operations. type ExportEntitiesMetadata struct { // Metadata common to all Datastore Admin operations. - Common *CommonMetadata `protobuf:"bytes,1,opt,name=common" json:"common,omitempty"` + Common *CommonMetadata `protobuf:"bytes,1,opt,name=common,proto3" json:"common,omitempty"` // An estimate of the number of entities processed. - ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities" json:"progress_entities,omitempty"` + ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities,proto3" json:"progress_entities,omitempty"` // An estimate of the number of bytes processed. - ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes" json:"progress_bytes,omitempty"` + ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes,proto3" json:"progress_bytes,omitempty"` // Description of which entities are being exported. - EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` // Location for the export metadata and data files. This will be the same // value as the // [google.datastore.admin.v1.ExportEntitiesRequest.output_url_prefix][google.datastore.admin.v1.ExportEntitiesRequest.output_url_prefix] // field. The final output location is provided in // [google.datastore.admin.v1.ExportEntitiesResponse.output_url][google.datastore.admin.v1.ExportEntitiesResponse.output_url]. - OutputUrlPrefix string `protobuf:"bytes,5,opt,name=output_url_prefix,json=outputUrlPrefix" json:"output_url_prefix,omitempty"` + OutputUrlPrefix string `protobuf:"bytes,5,opt,name=output_url_prefix,json=outputUrlPrefix,proto3" json:"output_url_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -532,16 +532,16 @@ func (m *ExportEntitiesMetadata) GetOutputUrlPrefix() string { // Metadata for ImportEntities operations. type ImportEntitiesMetadata struct { // Metadata common to all Datastore Admin operations. - Common *CommonMetadata `protobuf:"bytes,1,opt,name=common" json:"common,omitempty"` + Common *CommonMetadata `protobuf:"bytes,1,opt,name=common,proto3" json:"common,omitempty"` // An estimate of the number of entities processed. - ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities" json:"progress_entities,omitempty"` + ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities,proto3" json:"progress_entities,omitempty"` // An estimate of the number of bytes processed. - ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes" json:"progress_bytes,omitempty"` + ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes,proto3" json:"progress_bytes,omitempty"` // Description of which entities are being imported. - EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` // The location of the import metadata file. This will be the same value as // the [google.datastore.admin.v1.ExportEntitiesResponse.output_url][google.datastore.admin.v1.ExportEntitiesResponse.output_url] field. - InputUrl string `protobuf:"bytes,5,opt,name=input_url,json=inputUrl" json:"input_url,omitempty"` + InputUrl string `protobuf:"bytes,5,opt,name=input_url,json=inputUrl,proto3" json:"input_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -627,7 +627,7 @@ func (m *ImportEntitiesMetadata) GetInputUrl() string { // kinds=[], namespace_ids=['Baz'] type EntityFilter struct { // If empty, then this represents all kinds. - Kinds []string `protobuf:"bytes,1,rep,name=kinds" json:"kinds,omitempty"` + Kinds []string `protobuf:"bytes,1,rep,name=kinds,proto3" json:"kinds,omitempty"` // An empty list represents all namespaces. This is the preferred // usage for projects that don't use namespaces. // @@ -635,7 +635,7 @@ type EntityFilter struct { // used if the project has data in non-default namespaces, but doesn't want to // include them. // Each namespace in this list must be unique. - NamespaceIds []string `protobuf:"bytes,2,rep,name=namespace_ids,json=namespaceIds" json:"namespace_ids,omitempty"` + NamespaceIds []string `protobuf:"bytes,2,rep,name=namespace_ids,json=namespaceIds,proto3" json:"namespace_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -750,8 +750,7 @@ func (c *datastoreAdminClient) ImportEntities(ctx context.Context, in *ImportEnt return out, nil } -// Server API for DatastoreAdmin service - +// DatastoreAdminServer is the server API for DatastoreAdmin service. type DatastoreAdminServer interface { // Exports a copy of all or a subset of entities from Google Cloud Datastore // to another storage system, such as Google Cloud Storage. Recent updates to diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1beta1/datastore_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1beta1/datastore_admin.pb.go index 44cdffede..59d56865e 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1beta1/datastore_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/admin/v1beta1/datastore_admin.pb.go @@ -111,17 +111,17 @@ func (CommonMetadata_State) EnumDescriptor() ([]byte, []int) { // Metadata common to all Datastore Admin operations. type CommonMetadata struct { // The time that work began on the operation. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time the operation ended, either successfully or otherwise. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // The type of the operation. Can be used as a filter in // ListOperationsRequest. - OperationType OperationType `protobuf:"varint,3,opt,name=operation_type,json=operationType,enum=google.datastore.admin.v1beta1.OperationType" json:"operation_type,omitempty"` + OperationType OperationType `protobuf:"varint,3,opt,name=operation_type,json=operationType,proto3,enum=google.datastore.admin.v1beta1.OperationType" json:"operation_type,omitempty"` // The client-assigned labels which were provided when the operation was // created. May also include additional labels. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The current state of the Operation. - State CommonMetadata_State `protobuf:"varint,5,opt,name=state,enum=google.datastore.admin.v1beta1.CommonMetadata_State" json:"state,omitempty"` + State CommonMetadata_State `protobuf:"varint,5,opt,name=state,proto3,enum=google.datastore.admin.v1beta1.CommonMetadata_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -190,10 +190,10 @@ func (m *CommonMetadata) GetState() CommonMetadata_State { type Progress struct { // The amount of work that has been completed. Note that this may be greater // than work_estimated. - WorkCompleted int64 `protobuf:"varint,1,opt,name=work_completed,json=workCompleted" json:"work_completed,omitempty"` + WorkCompleted int64 `protobuf:"varint,1,opt,name=work_completed,json=workCompleted,proto3" json:"work_completed,omitempty"` // An estimate of how much work needs to be performed. May be zero if the // work estimate is unavailable. - WorkEstimated int64 `protobuf:"varint,2,opt,name=work_estimated,json=workEstimated" json:"work_estimated,omitempty"` + WorkEstimated int64 `protobuf:"varint,2,opt,name=work_estimated,json=workEstimated,proto3" json:"work_estimated,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -241,11 +241,11 @@ func (m *Progress) GetWorkEstimated() int64 { // [google.datastore.admin.v1beta1.DatastoreAdmin.ExportEntities][google.datastore.admin.v1beta1.DatastoreAdmin.ExportEntities]. type ExportEntitiesRequest struct { // Project ID against which to make the request. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Client-assigned labels. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Description of what data from the project is included in the export. - EntityFilter *EntityFilter `protobuf:"bytes,3,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,3,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` // Location for the export metadata and data files. // // The full resource URL of the external storage location. Currently, only @@ -264,7 +264,7 @@ type ExportEntitiesRequest struct { // // By nesting the data files deeper, the same Cloud Storage bucket can be used // in multiple ExportEntities operations without conflict. - OutputUrlPrefix string `protobuf:"bytes,4,opt,name=output_url_prefix,json=outputUrlPrefix" json:"output_url_prefix,omitempty"` + OutputUrlPrefix string `protobuf:"bytes,4,opt,name=output_url_prefix,json=outputUrlPrefix,proto3" json:"output_url_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -326,9 +326,9 @@ func (m *ExportEntitiesRequest) GetOutputUrlPrefix() string { // [google.datastore.admin.v1beta1.DatastoreAdmin.ImportEntities][google.datastore.admin.v1beta1.DatastoreAdmin.ImportEntities]. type ImportEntitiesRequest struct { // Project ID against which to make the request. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Client-assigned labels. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The full resource URL of the external storage location. Currently, only // Google Cloud Storage is supported. So input_url should be of the form: // `gs://BUCKET_NAME[/NAMESPACE_PATH]/OVERALL_EXPORT_METADATA_FILE`, where @@ -342,12 +342,12 @@ type ImportEntitiesRequest struct { // // For more information, see // [google.datastore.admin.v1beta1.ExportEntitiesResponse.output_url][google.datastore.admin.v1beta1.ExportEntitiesResponse.output_url]. - InputUrl string `protobuf:"bytes,3,opt,name=input_url,json=inputUrl" json:"input_url,omitempty"` + InputUrl string `protobuf:"bytes,3,opt,name=input_url,json=inputUrl,proto3" json:"input_url,omitempty"` // Optionally specify which kinds/namespaces are to be imported. If provided, // the list must be a subset of the EntityFilter used in creating the export, // otherwise a FAILED_PRECONDITION error will be returned. If no filter is // specified then all entities from the export are imported. - EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -412,7 +412,7 @@ type ExportEntitiesResponse struct { // into Cloud Datastore (this project or another project). See // [google.datastore.admin.v1beta1.ImportEntitiesRequest.input_url][google.datastore.admin.v1beta1.ImportEntitiesRequest.input_url]. // Only present if the operation completed successfully. - OutputUrl string `protobuf:"bytes,1,opt,name=output_url,json=outputUrl" json:"output_url,omitempty"` + OutputUrl string `protobuf:"bytes,1,opt,name=output_url,json=outputUrl,proto3" json:"output_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -452,19 +452,19 @@ func (m *ExportEntitiesResponse) GetOutputUrl() string { // Metadata for ExportEntities operations. type ExportEntitiesMetadata struct { // Metadata common to all Datastore Admin operations. - Common *CommonMetadata `protobuf:"bytes,1,opt,name=common" json:"common,omitempty"` + Common *CommonMetadata `protobuf:"bytes,1,opt,name=common,proto3" json:"common,omitempty"` // An estimate of the number of entities processed. - ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities" json:"progress_entities,omitempty"` + ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities,proto3" json:"progress_entities,omitempty"` // An estimate of the number of bytes processed. - ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes" json:"progress_bytes,omitempty"` + ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes,proto3" json:"progress_bytes,omitempty"` // Description of which entities are being exported. - EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` // Location for the export metadata and data files. This will be the same // value as the // [google.datastore.admin.v1beta1.ExportEntitiesRequest.output_url_prefix][google.datastore.admin.v1beta1.ExportEntitiesRequest.output_url_prefix] // field. The final output location is provided in // [google.datastore.admin.v1beta1.ExportEntitiesResponse.output_url][google.datastore.admin.v1beta1.ExportEntitiesResponse.output_url]. - OutputUrlPrefix string `protobuf:"bytes,5,opt,name=output_url_prefix,json=outputUrlPrefix" json:"output_url_prefix,omitempty"` + OutputUrlPrefix string `protobuf:"bytes,5,opt,name=output_url_prefix,json=outputUrlPrefix,proto3" json:"output_url_prefix,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -532,17 +532,17 @@ func (m *ExportEntitiesMetadata) GetOutputUrlPrefix() string { // Metadata for ImportEntities operations. type ImportEntitiesMetadata struct { // Metadata common to all Datastore Admin operations. - Common *CommonMetadata `protobuf:"bytes,1,opt,name=common" json:"common,omitempty"` + Common *CommonMetadata `protobuf:"bytes,1,opt,name=common,proto3" json:"common,omitempty"` // An estimate of the number of entities processed. - ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities" json:"progress_entities,omitempty"` + ProgressEntities *Progress `protobuf:"bytes,2,opt,name=progress_entities,json=progressEntities,proto3" json:"progress_entities,omitempty"` // An estimate of the number of bytes processed. - ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes" json:"progress_bytes,omitempty"` + ProgressBytes *Progress `protobuf:"bytes,3,opt,name=progress_bytes,json=progressBytes,proto3" json:"progress_bytes,omitempty"` // Description of which entities are being imported. - EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter" json:"entity_filter,omitempty"` + EntityFilter *EntityFilter `protobuf:"bytes,4,opt,name=entity_filter,json=entityFilter,proto3" json:"entity_filter,omitempty"` // The location of the import metadata file. This will be the same value as // the [google.datastore.admin.v1beta1.ExportEntitiesResponse.output_url][google.datastore.admin.v1beta1.ExportEntitiesResponse.output_url] // field. - InputUrl string `protobuf:"bytes,5,opt,name=input_url,json=inputUrl" json:"input_url,omitempty"` + InputUrl string `protobuf:"bytes,5,opt,name=input_url,json=inputUrl,proto3" json:"input_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -628,7 +628,7 @@ func (m *ImportEntitiesMetadata) GetInputUrl() string { // kinds=[], namespace_ids=['Baz'] type EntityFilter struct { // If empty, then this represents all kinds. - Kinds []string `protobuf:"bytes,1,rep,name=kinds" json:"kinds,omitempty"` + Kinds []string `protobuf:"bytes,1,rep,name=kinds,proto3" json:"kinds,omitempty"` // An empty list represents all namespaces. This is the preferred // usage for projects that don't use namespaces. // @@ -636,7 +636,7 @@ type EntityFilter struct { // used if the project has data in non-default namespaces, but doesn't want to // include them. // Each namespace in this list must be unique. - NamespaceIds []string `protobuf:"bytes,2,rep,name=namespace_ids,json=namespaceIds" json:"namespace_ids,omitempty"` + NamespaceIds []string `protobuf:"bytes,2,rep,name=namespace_ids,json=namespaceIds,proto3" json:"namespace_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -751,8 +751,7 @@ func (c *datastoreAdminClient) ImportEntities(ctx context.Context, in *ImportEnt return out, nil } -// Server API for DatastoreAdmin service - +// DatastoreAdminServer is the server API for DatastoreAdmin service. type DatastoreAdminServer interface { // Exports a copy of all or a subset of entities from Google Cloud Datastore // to another storage system, such as Google Cloud Storage. Recent updates to diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/v1/datastore.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/v1/datastore.pb.go index 52ec22eb9..1dcf40d61 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/v1/datastore.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/v1/datastore.pb.go @@ -88,11 +88,11 @@ func (ReadOptions_ReadConsistency) EnumDescriptor() ([]byte, []int) { // The request for [Datastore.Lookup][google.datastore.v1.Datastore.Lookup]. type LookupRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The options for this lookup request. - ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions" json:"read_options,omitempty"` + ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions,proto3" json:"read_options,omitempty"` // Keys of entities to look up. - Keys []*Key `protobuf:"bytes,3,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,3,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -148,15 +148,15 @@ type LookupResponse struct { // Entities found as `ResultType.FULL` entities. The order of results in this // field is undefined and has no relation to the order of the keys in the // input. - Found []*EntityResult `protobuf:"bytes,1,rep,name=found" json:"found,omitempty"` + Found []*EntityResult `protobuf:"bytes,1,rep,name=found,proto3" json:"found,omitempty"` // Entities not found as `ResultType.KEY_ONLY` entities. The order of results // in this field is undefined and has no relation to the order of the keys // in the input. - Missing []*EntityResult `protobuf:"bytes,2,rep,name=missing" json:"missing,omitempty"` + Missing []*EntityResult `protobuf:"bytes,2,rep,name=missing,proto3" json:"missing,omitempty"` // A list of keys that were not looked up due to resource constraints. The // order of results in this field is undefined and has no relation to the // order of the keys in the input. - Deferred []*Key `protobuf:"bytes,3,rep,name=deferred" json:"deferred,omitempty"` + Deferred []*Key `protobuf:"bytes,3,rep,name=deferred,proto3" json:"deferred,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -210,14 +210,14 @@ func (m *LookupResponse) GetDeferred() []*Key { // The request for [Datastore.RunQuery][google.datastore.v1.Datastore.RunQuery]. type RunQueryRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Entities are partitioned into subsets, identified by a partition ID. // Queries are scoped to a single partition. // This partition ID is normalized with the standard default context // partition ID. - PartitionId *PartitionId `protobuf:"bytes,2,opt,name=partition_id,json=partitionId" json:"partition_id,omitempty"` + PartitionId *PartitionId `protobuf:"bytes,2,opt,name=partition_id,json=partitionId,proto3" json:"partition_id,omitempty"` // The options for this query. - ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions" json:"read_options,omitempty"` + ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions,proto3" json:"read_options,omitempty"` // The type of query. // // Types that are valid to be assigned to QueryType: @@ -258,10 +258,10 @@ type isRunQueryRequest_QueryType interface { } type RunQueryRequest_Query struct { - Query *Query `protobuf:"bytes,3,opt,name=query,oneof"` + Query *Query `protobuf:"bytes,3,opt,name=query,proto3,oneof"` } type RunQueryRequest_GqlQuery struct { - GqlQuery *GqlQuery `protobuf:"bytes,7,opt,name=gql_query,json=gqlQuery,oneof"` + GqlQuery *GqlQuery `protobuf:"bytes,7,opt,name=gql_query,json=gqlQuery,proto3,oneof"` } func (*RunQueryRequest_Query) isRunQueryRequest_QueryType() {} @@ -386,9 +386,9 @@ func _RunQueryRequest_OneofSizer(msg proto.Message) (n int) { // The response for [Datastore.RunQuery][google.datastore.v1.Datastore.RunQuery]. type RunQueryResponse struct { // A batch of query results (always present). - Batch *QueryResultBatch `protobuf:"bytes,1,opt,name=batch" json:"batch,omitempty"` + Batch *QueryResultBatch `protobuf:"bytes,1,opt,name=batch,proto3" json:"batch,omitempty"` // The parsed form of the `GqlQuery` from the request, if it was set. - Query *Query `protobuf:"bytes,2,opt,name=query" json:"query,omitempty"` + Query *Query `protobuf:"bytes,2,opt,name=query,proto3" json:"query,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -435,9 +435,9 @@ func (m *RunQueryResponse) GetQuery() *Query { // The request for [Datastore.BeginTransaction][google.datastore.v1.Datastore.BeginTransaction]. type BeginTransactionRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Options for a new transaction. - TransactionOptions *TransactionOptions `protobuf:"bytes,10,opt,name=transaction_options,json=transactionOptions" json:"transaction_options,omitempty"` + TransactionOptions *TransactionOptions `protobuf:"bytes,10,opt,name=transaction_options,json=transactionOptions,proto3" json:"transaction_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -524,7 +524,7 @@ func (m *BeginTransactionResponse) GetTransaction() []byte { // The request for [Datastore.Rollback][google.datastore.v1.Datastore.Rollback]. type RollbackRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The transaction identifier, returned by a call to // [Datastore.BeginTransaction][google.datastore.v1.Datastore.BeginTransaction]. Transaction []byte `protobuf:"bytes,1,opt,name=transaction,proto3" json:"transaction,omitempty"` @@ -606,9 +606,9 @@ var xxx_messageInfo_RollbackResponse proto.InternalMessageInfo // The request for [Datastore.Commit][google.datastore.v1.Datastore.Commit]. type CommitRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The type of commit to perform. Defaults to `TRANSACTIONAL`. - Mode CommitRequest_Mode `protobuf:"varint,5,opt,name=mode,enum=google.datastore.v1.CommitRequest_Mode" json:"mode,omitempty"` + Mode CommitRequest_Mode `protobuf:"varint,5,opt,name=mode,proto3,enum=google.datastore.v1.CommitRequest_Mode" json:"mode,omitempty"` // Must be set when mode is `TRANSACTIONAL`. // // Types that are valid to be assigned to TransactionSelector: @@ -627,7 +627,7 @@ type CommitRequest struct { // // When mode is `NON_TRANSACTIONAL`, no two mutations may affect a single // entity. - Mutations []*Mutation `protobuf:"bytes,6,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,6,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -757,10 +757,10 @@ func _CommitRequest_OneofSizer(msg proto.Message) (n int) { type CommitResponse struct { // The result of performing the mutations. // The i-th mutation result corresponds to the i-th mutation in the request. - MutationResults []*MutationResult `protobuf:"bytes,3,rep,name=mutation_results,json=mutationResults" json:"mutation_results,omitempty"` + MutationResults []*MutationResult `protobuf:"bytes,3,rep,name=mutation_results,json=mutationResults,proto3" json:"mutation_results,omitempty"` // The number of index entries updated during the commit, or zero if none were // updated. - IndexUpdates int32 `protobuf:"varint,4,opt,name=index_updates,json=indexUpdates" json:"index_updates,omitempty"` + IndexUpdates int32 `protobuf:"varint,4,opt,name=index_updates,json=indexUpdates,proto3" json:"index_updates,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -807,10 +807,10 @@ func (m *CommitResponse) GetIndexUpdates() int32 { // The request for [Datastore.AllocateIds][google.datastore.v1.Datastore.AllocateIds]. type AllocateIdsRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // A list of keys with incomplete key paths for which to allocate IDs. // No key may be reserved/read-only. - Keys []*Key `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -858,7 +858,7 @@ func (m *AllocateIdsRequest) GetKeys() []*Key { type AllocateIdsResponse struct { // The keys specified in the request (in the same order), each with // its key path completed with a newly allocated ID. - Keys []*Key `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -898,12 +898,12 @@ func (m *AllocateIdsResponse) GetKeys() []*Key { // The request for [Datastore.ReserveIds][google.datastore.v1.Datastore.ReserveIds]. type ReserveIdsRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // If not empty, the ID of the database against which to make the request. - DatabaseId string `protobuf:"bytes,9,opt,name=database_id,json=databaseId" json:"database_id,omitempty"` + DatabaseId string `protobuf:"bytes,9,opt,name=database_id,json=databaseId,proto3" json:"database_id,omitempty"` // A list of keys with complete key paths whose numeric IDs should not be // auto-allocated. - Keys []*Key `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1046,19 +1046,19 @@ type isMutation_ConflictDetectionStrategy interface { } type Mutation_Insert struct { - Insert *Entity `protobuf:"bytes,4,opt,name=insert,oneof"` + Insert *Entity `protobuf:"bytes,4,opt,name=insert,proto3,oneof"` } type Mutation_Update struct { - Update *Entity `protobuf:"bytes,5,opt,name=update,oneof"` + Update *Entity `protobuf:"bytes,5,opt,name=update,proto3,oneof"` } type Mutation_Upsert struct { - Upsert *Entity `protobuf:"bytes,6,opt,name=upsert,oneof"` + Upsert *Entity `protobuf:"bytes,6,opt,name=upsert,proto3,oneof"` } type Mutation_Delete struct { - Delete *Key `protobuf:"bytes,7,opt,name=delete,oneof"` + Delete *Key `protobuf:"bytes,7,opt,name=delete,proto3,oneof"` } type Mutation_BaseVersion struct { - BaseVersion int64 `protobuf:"varint,8,opt,name=base_version,json=baseVersion,oneof"` + BaseVersion int64 `protobuf:"varint,8,opt,name=base_version,json=baseVersion,proto3,oneof"` } func (*Mutation_Insert) isMutation_Operation() {} @@ -1257,16 +1257,16 @@ func _Mutation_OneofSizer(msg proto.Message) (n int) { type MutationResult struct { // The automatically allocated key. // Set only when the mutation allocated a key. - Key *Key `protobuf:"bytes,3,opt,name=key" json:"key,omitempty"` + Key *Key `protobuf:"bytes,3,opt,name=key,proto3" json:"key,omitempty"` // The version of the entity on the server after processing the mutation. If // the mutation doesn't change anything on the server, then the version will // be the version of the current entity or, if no entity is present, a version // that is strictly greater than the version of any previous entity and less // than the version of any possible future entity. - Version int64 `protobuf:"varint,4,opt,name=version" json:"version,omitempty"` + Version int64 `protobuf:"varint,4,opt,name=version,proto3" json:"version,omitempty"` // Whether a conflict was detected for this mutation. Always false when a // conflict detection strategy field is not set in the mutation. - ConflictDetected bool `protobuf:"varint,5,opt,name=conflict_detected,json=conflictDetected" json:"conflict_detected,omitempty"` + ConflictDetected bool `protobuf:"varint,5,opt,name=conflict_detected,json=conflictDetected,proto3" json:"conflict_detected,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1361,7 +1361,7 @@ type isReadOptions_ConsistencyType interface { } type ReadOptions_ReadConsistency_ struct { - ReadConsistency ReadOptions_ReadConsistency `protobuf:"varint,1,opt,name=read_consistency,json=readConsistency,enum=google.datastore.v1.ReadOptions_ReadConsistency,oneof"` + ReadConsistency ReadOptions_ReadConsistency `protobuf:"varint,1,opt,name=read_consistency,json=readConsistency,proto3,enum=google.datastore.v1.ReadOptions_ReadConsistency,oneof"` } type ReadOptions_Transaction struct { Transaction []byte `protobuf:"bytes,2,opt,name=transaction,proto3,oneof"` @@ -1503,10 +1503,10 @@ type isTransactionOptions_Mode interface { } type TransactionOptions_ReadWrite_ struct { - ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,1,opt,name=read_write,json=readWrite,oneof"` + ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,1,opt,name=read_write,json=readWrite,proto3,oneof"` } type TransactionOptions_ReadOnly_ struct { - ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,oneof"` + ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,proto3,oneof"` } func (*TransactionOptions_ReadWrite_) isTransactionOptions_Mode() {} @@ -1805,8 +1805,7 @@ func (c *datastoreClient) ReserveIds(ctx context.Context, in *ReserveIdsRequest, return out, nil } -// Server API for Datastore service - +// DatastoreServer is the server API for Datastore service. type DatastoreServer interface { // Looks up entities by key. Lookup(context.Context, *LookupRequest) (*LookupResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/v1/entity.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/v1/entity.pb.go index 20b7709a0..3cc8edc29 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/v1/entity.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/v1/entity.pb.go @@ -43,9 +43,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Reads and writes of foreign partition IDs may fail if the project is not in an active state. type PartitionId struct { // The ID of the project to which the entities belong. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // If not empty, the ID of the namespace to which the entities belong. - NamespaceId string `protobuf:"bytes,4,opt,name=namespace_id,json=namespaceId" json:"namespace_id,omitempty"` + NamespaceId string `protobuf:"bytes,4,opt,name=namespace_id,json=namespaceId,proto3" json:"namespace_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -97,7 +97,7 @@ type Key struct { // Entities are partitioned into subsets, currently identified by a project // ID and namespace ID. // Queries are scoped to a single partition. - PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId" json:"partition_id,omitempty"` + PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId,proto3" json:"partition_id,omitempty"` // The entity path. // An entity path consists of one or more elements composed of a kind and a // string or numerical identifier, which identify entities. The first @@ -114,7 +114,7 @@ type Key struct { // identifier. // // A path can never be empty, and a path can have at most 100 elements. - Path []*Key_PathElement `protobuf:"bytes,2,rep,name=path" json:"path,omitempty"` + Path []*Key_PathElement `protobuf:"bytes,2,rep,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -167,7 +167,7 @@ type Key_PathElement struct { // A kind matching regex `__.*__` is reserved/read-only. // A kind must not contain more than 1500 bytes when UTF-8 encoded. // Cannot be `""`. - Kind string `protobuf:"bytes,1,opt,name=kind" json:"kind,omitempty"` + Kind string `protobuf:"bytes,1,opt,name=kind,proto3" json:"kind,omitempty"` // The type of ID. // // Types that are valid to be assigned to IdType: @@ -208,10 +208,10 @@ type isKey_PathElement_IdType interface { } type Key_PathElement_Id struct { - Id int64 `protobuf:"varint,2,opt,name=id,oneof"` + Id int64 `protobuf:"varint,2,opt,name=id,proto3,oneof"` } type Key_PathElement_Name struct { - Name string `protobuf:"bytes,3,opt,name=name,oneof"` + Name string `protobuf:"bytes,3,opt,name=name,proto3,oneof"` } func (*Key_PathElement_Id) isKey_PathElement_IdType() {} @@ -315,7 +315,7 @@ type ArrayValue struct { // Values in the array. // The order of this array may not be preserved if it contains a mix of // indexed and unindexed values. - Values []*Value `protobuf:"bytes,1,rep,name=values" json:"values,omitempty"` + Values []*Value `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -371,10 +371,10 @@ type Value struct { // *Value_ArrayValue ValueType isValue_ValueType `protobuf_oneof:"value_type"` // The `meaning` field should only be populated for backwards compatibility. - Meaning int32 `protobuf:"varint,14,opt,name=meaning" json:"meaning,omitempty"` + Meaning int32 `protobuf:"varint,14,opt,name=meaning,proto3" json:"meaning,omitempty"` // If the value should be excluded from all indexes including those defined // explicitly. - ExcludeFromIndexes bool `protobuf:"varint,19,opt,name=exclude_from_indexes,json=excludeFromIndexes" json:"exclude_from_indexes,omitempty"` + ExcludeFromIndexes bool `protobuf:"varint,19,opt,name=exclude_from_indexes,json=excludeFromIndexes,proto3" json:"exclude_from_indexes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,37 +409,37 @@ type isValue_ValueType interface { } type Value_NullValue struct { - NullValue _struct.NullValue `protobuf:"varint,11,opt,name=null_value,json=nullValue,enum=google.protobuf.NullValue,oneof"` + NullValue _struct.NullValue `protobuf:"varint,11,opt,name=null_value,json=nullValue,proto3,enum=google.protobuf.NullValue,oneof"` } type Value_BooleanValue struct { - BooleanValue bool `protobuf:"varint,1,opt,name=boolean_value,json=booleanValue,oneof"` + BooleanValue bool `protobuf:"varint,1,opt,name=boolean_value,json=booleanValue,proto3,oneof"` } type Value_IntegerValue struct { - IntegerValue int64 `protobuf:"varint,2,opt,name=integer_value,json=integerValue,oneof"` + IntegerValue int64 `protobuf:"varint,2,opt,name=integer_value,json=integerValue,proto3,oneof"` } type Value_DoubleValue struct { - DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,oneof"` + DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,proto3,oneof"` } type Value_TimestampValue struct { - TimestampValue *timestamp.Timestamp `protobuf:"bytes,10,opt,name=timestamp_value,json=timestampValue,oneof"` + TimestampValue *timestamp.Timestamp `protobuf:"bytes,10,opt,name=timestamp_value,json=timestampValue,proto3,oneof"` } type Value_KeyValue struct { - KeyValue *Key `protobuf:"bytes,5,opt,name=key_value,json=keyValue,oneof"` + KeyValue *Key `protobuf:"bytes,5,opt,name=key_value,json=keyValue,proto3,oneof"` } type Value_StringValue struct { - StringValue string `protobuf:"bytes,17,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,17,opt,name=string_value,json=stringValue,proto3,oneof"` } type Value_BlobValue struct { BlobValue []byte `protobuf:"bytes,18,opt,name=blob_value,json=blobValue,proto3,oneof"` } type Value_GeoPointValue struct { - GeoPointValue *latlng.LatLng `protobuf:"bytes,8,opt,name=geo_point_value,json=geoPointValue,oneof"` + GeoPointValue *latlng.LatLng `protobuf:"bytes,8,opt,name=geo_point_value,json=geoPointValue,proto3,oneof"` } type Value_EntityValue struct { - EntityValue *Entity `protobuf:"bytes,6,opt,name=entity_value,json=entityValue,oneof"` + EntityValue *Entity `protobuf:"bytes,6,opt,name=entity_value,json=entityValue,proto3,oneof"` } type Value_ArrayValue struct { - ArrayValue *ArrayValue `protobuf:"bytes,9,opt,name=array_value,json=arrayValue,oneof"` + ArrayValue *ArrayValue `protobuf:"bytes,9,opt,name=array_value,json=arrayValue,proto3,oneof"` } func (*Value_NullValue) isValue_ValueType() {} @@ -785,14 +785,14 @@ type Entity struct { // an entity in `Value.entity_value` may have no key). // An entity's kind is its key path's last element's kind, // or null if it has no key. - Key *Key `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key *Key `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // The entity's properties. // The map's keys are property names. // A property name matching regex `__.*__` is reserved. // A reserved property name is forbidden in certain documented contexts. // The name must not contain more than 500 characters. // The name cannot be `""`. - Properties map[string]*Value `protobuf:"bytes,3,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]*Value `protobuf:"bytes,3,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/v1/query.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/v1/query.pb.go index e9edae7e6..fd19b083f 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/v1/query.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/v1/query.pb.go @@ -203,7 +203,7 @@ func (QueryResultBatch_MoreResultsType) EnumDescriptor() ([]byte, []int) { // The result of fetching an entity from Datastore. type EntityResult struct { // The resulting entity. - Entity *Entity `protobuf:"bytes,1,opt,name=entity" json:"entity,omitempty"` + Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"` // The version of the entity, a strictly positive number that monotonically // increases with changes to the entity. // @@ -213,7 +213,7 @@ type EntityResult struct { // For [missing][google.datastore.v1.LookupResponse.missing] entities in `LookupResponse`, this // is the version of the snapshot that was used to look up the entity, and it // is always set except for eventually consistent reads. - Version int64 `protobuf:"varint,4,opt,name=version" json:"version,omitempty"` + Version int64 `protobuf:"varint,4,opt,name=version,proto3" json:"version,omitempty"` // A cursor that points to the position after the result entity. // Set only when the `EntityResult` is part of a `QueryResultBatch` message. Cursor []byte `protobuf:"bytes,3,opt,name=cursor,proto3" json:"cursor,omitempty"` @@ -270,18 +270,18 @@ func (m *EntityResult) GetCursor() []byte { // A query for entities. type Query struct { // The projection to return. Defaults to returning all properties. - Projection []*Projection `protobuf:"bytes,2,rep,name=projection" json:"projection,omitempty"` + Projection []*Projection `protobuf:"bytes,2,rep,name=projection,proto3" json:"projection,omitempty"` // The kinds to query (if empty, returns entities of all kinds). // Currently at most 1 kind may be specified. - Kind []*KindExpression `protobuf:"bytes,3,rep,name=kind" json:"kind,omitempty"` + Kind []*KindExpression `protobuf:"bytes,3,rep,name=kind,proto3" json:"kind,omitempty"` // The filter to apply. - Filter *Filter `protobuf:"bytes,4,opt,name=filter" json:"filter,omitempty"` + Filter *Filter `protobuf:"bytes,4,opt,name=filter,proto3" json:"filter,omitempty"` // The order to apply to the query results (if empty, order is unspecified). - Order []*PropertyOrder `protobuf:"bytes,5,rep,name=order" json:"order,omitempty"` + Order []*PropertyOrder `protobuf:"bytes,5,rep,name=order,proto3" json:"order,omitempty"` // The properties to make distinct. The query results will contain the first // result for each distinct combination of values for the given properties // (if empty, all results are returned). - DistinctOn []*PropertyReference `protobuf:"bytes,6,rep,name=distinct_on,json=distinctOn" json:"distinct_on,omitempty"` + DistinctOn []*PropertyReference `protobuf:"bytes,6,rep,name=distinct_on,json=distinctOn,proto3" json:"distinct_on,omitempty"` // A starting point for the query results. Query cursors are // returned in query result batches and // [can only be used to continue the same query](https://cloud.google.com/datastore/docs/concepts/queries#cursors_limits_and_offsets). @@ -292,12 +292,12 @@ type Query struct { EndCursor []byte `protobuf:"bytes,8,opt,name=end_cursor,json=endCursor,proto3" json:"end_cursor,omitempty"` // The number of results to skip. Applies before limit, but after all other // constraints. Optional. Must be >= 0 if specified. - Offset int32 `protobuf:"varint,10,opt,name=offset" json:"offset,omitempty"` + Offset int32 `protobuf:"varint,10,opt,name=offset,proto3" json:"offset,omitempty"` // The maximum number of results to return. Applies after all other // constraints. Optional. // Unspecified is interpreted as no limit. // Must be >= 0 if specified. - Limit *wrappers.Int32Value `protobuf:"bytes,12,opt,name=limit" json:"limit,omitempty"` + Limit *wrappers.Int32Value `protobuf:"bytes,12,opt,name=limit,proto3" json:"limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -393,7 +393,7 @@ func (m *Query) GetLimit() *wrappers.Int32Value { // A representation of a kind. type KindExpression struct { // The name of the kind. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -434,7 +434,7 @@ func (m *KindExpression) GetName() string { type PropertyReference struct { // The name of the property. // If name includes "."s, it may be interpreted as a property name path. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -474,7 +474,7 @@ func (m *PropertyReference) GetName() string { // A representation of a property in a projection. type Projection struct { // The property to project. - Property *PropertyReference `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *PropertyReference `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -514,9 +514,9 @@ func (m *Projection) GetProperty() *PropertyReference { // The desired order for a specific property. type PropertyOrder struct { // The property to order by. - Property *PropertyReference `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *PropertyReference `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The direction to order by. Defaults to `ASCENDING`. - Direction PropertyOrder_Direction `protobuf:"varint,2,opt,name=direction,enum=google.datastore.v1.PropertyOrder_Direction" json:"direction,omitempty"` + Direction PropertyOrder_Direction `protobuf:"varint,2,opt,name=direction,proto3,enum=google.datastore.v1.PropertyOrder_Direction" json:"direction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -602,10 +602,10 @@ type isFilter_FilterType interface { } type Filter_CompositeFilter struct { - CompositeFilter *CompositeFilter `protobuf:"bytes,1,opt,name=composite_filter,json=compositeFilter,oneof"` + CompositeFilter *CompositeFilter `protobuf:"bytes,1,opt,name=composite_filter,json=compositeFilter,proto3,oneof"` } type Filter_PropertyFilter struct { - PropertyFilter *PropertyFilter `protobuf:"bytes,2,opt,name=property_filter,json=propertyFilter,oneof"` + PropertyFilter *PropertyFilter `protobuf:"bytes,2,opt,name=property_filter,json=propertyFilter,proto3,oneof"` } func (*Filter_CompositeFilter) isFilter_FilterType() {} @@ -709,10 +709,10 @@ func _Filter_OneofSizer(msg proto.Message) (n int) { // A filter that merges multiple other filters using the given operator. type CompositeFilter struct { // The operator for combining multiple filters. - Op CompositeFilter_Operator `protobuf:"varint,1,opt,name=op,enum=google.datastore.v1.CompositeFilter_Operator" json:"op,omitempty"` + Op CompositeFilter_Operator `protobuf:"varint,1,opt,name=op,proto3,enum=google.datastore.v1.CompositeFilter_Operator" json:"op,omitempty"` // The list of filters to combine. // Must contain at least one filter. - Filters []*Filter `protobuf:"bytes,2,rep,name=filters" json:"filters,omitempty"` + Filters []*Filter `protobuf:"bytes,2,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -759,11 +759,11 @@ func (m *CompositeFilter) GetFilters() []*Filter { // A filter on a specific property. type PropertyFilter struct { // The property to filter by. - Property *PropertyReference `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *PropertyReference `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The operator to filter by. - Op PropertyFilter_Operator `protobuf:"varint,2,opt,name=op,enum=google.datastore.v1.PropertyFilter_Operator" json:"op,omitempty"` + Op PropertyFilter_Operator `protobuf:"varint,2,opt,name=op,proto3,enum=google.datastore.v1.PropertyFilter_Operator" json:"op,omitempty"` // The value to compare the property to. - Value *Value `protobuf:"bytes,3,opt,name=value" json:"value,omitempty"` + Value *Value `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -818,24 +818,24 @@ func (m *PropertyFilter) GetValue() *Value { type GqlQuery struct { // A string of the format described // [here](https://cloud.google.com/datastore/docs/apis/gql/gql_reference). - QueryString string `protobuf:"bytes,1,opt,name=query_string,json=queryString" json:"query_string,omitempty"` + QueryString string `protobuf:"bytes,1,opt,name=query_string,json=queryString,proto3" json:"query_string,omitempty"` // When false, the query string must not contain any literals and instead must // bind all values. For example, // `SELECT * FROM Kind WHERE a = 'string literal'` is not allowed, while // `SELECT * FROM Kind WHERE a = @value` is. - AllowLiterals bool `protobuf:"varint,2,opt,name=allow_literals,json=allowLiterals" json:"allow_literals,omitempty"` + AllowLiterals bool `protobuf:"varint,2,opt,name=allow_literals,json=allowLiterals,proto3" json:"allow_literals,omitempty"` // For each non-reserved named binding site in the query string, there must be // a named parameter with that name, but not necessarily the inverse. // // Key must match regex `[A-Za-z_$][A-Za-z_$0-9]*`, must not match regex // `__.*__`, and must not be `""`. - NamedBindings map[string]*GqlQueryParameter `protobuf:"bytes,5,rep,name=named_bindings,json=namedBindings" json:"named_bindings,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + NamedBindings map[string]*GqlQueryParameter `protobuf:"bytes,5,rep,name=named_bindings,json=namedBindings,proto3" json:"named_bindings,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Numbered binding site @1 references the first numbered parameter, // effectively using 1-based indexing, rather than the usual 0. // // For each binding site numbered i in `query_string`, there must be an i-th // numbered parameter. The inverse must also be true. - PositionalBindings []*GqlQueryParameter `protobuf:"bytes,4,rep,name=positional_bindings,json=positionalBindings" json:"positional_bindings,omitempty"` + PositionalBindings []*GqlQueryParameter `protobuf:"bytes,4,rep,name=positional_bindings,json=positionalBindings,proto3" json:"positional_bindings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -935,7 +935,7 @@ type isGqlQueryParameter_ParameterType interface { } type GqlQueryParameter_Value struct { - Value *Value `protobuf:"bytes,2,opt,name=value,oneof"` + Value *Value `protobuf:"bytes,2,opt,name=value,proto3,oneof"` } type GqlQueryParameter_Cursor struct { Cursor []byte `protobuf:"bytes,3,opt,name=cursor,proto3,oneof"` @@ -1038,18 +1038,18 @@ func _GqlQueryParameter_OneofSizer(msg proto.Message) (n int) { // A batch of results produced by a query. type QueryResultBatch struct { // The number of results skipped, typically because of an offset. - SkippedResults int32 `protobuf:"varint,6,opt,name=skipped_results,json=skippedResults" json:"skipped_results,omitempty"` + SkippedResults int32 `protobuf:"varint,6,opt,name=skipped_results,json=skippedResults,proto3" json:"skipped_results,omitempty"` // A cursor that points to the position after the last skipped result. // Will be set when `skipped_results` != 0. SkippedCursor []byte `protobuf:"bytes,3,opt,name=skipped_cursor,json=skippedCursor,proto3" json:"skipped_cursor,omitempty"` // The result type for every entity in `entity_results`. - EntityResultType EntityResult_ResultType `protobuf:"varint,1,opt,name=entity_result_type,json=entityResultType,enum=google.datastore.v1.EntityResult_ResultType" json:"entity_result_type,omitempty"` + EntityResultType EntityResult_ResultType `protobuf:"varint,1,opt,name=entity_result_type,json=entityResultType,proto3,enum=google.datastore.v1.EntityResult_ResultType" json:"entity_result_type,omitempty"` // The results for this batch. - EntityResults []*EntityResult `protobuf:"bytes,2,rep,name=entity_results,json=entityResults" json:"entity_results,omitempty"` + EntityResults []*EntityResult `protobuf:"bytes,2,rep,name=entity_results,json=entityResults,proto3" json:"entity_results,omitempty"` // A cursor that points to the position after the last result in the batch. EndCursor []byte `protobuf:"bytes,4,opt,name=end_cursor,json=endCursor,proto3" json:"end_cursor,omitempty"` // The state of the query after the current batch. - MoreResults QueryResultBatch_MoreResultsType `protobuf:"varint,5,opt,name=more_results,json=moreResults,enum=google.datastore.v1.QueryResultBatch_MoreResultsType" json:"more_results,omitempty"` + MoreResults QueryResultBatch_MoreResultsType `protobuf:"varint,5,opt,name=more_results,json=moreResults,proto3,enum=google.datastore.v1.QueryResultBatch_MoreResultsType" json:"more_results,omitempty"` // The version number of the snapshot this batch was returned from. // This applies to the range of results from the query's `start_cursor` (or // the beginning of the query if no cursor was given) to this batch's @@ -1059,7 +1059,7 @@ type QueryResultBatch struct { // can have a greater snapshot version number. Each batch's snapshot version // is valid for all preceding batches. // The value will be zero for eventually consistent queries. - SnapshotVersion int64 `protobuf:"varint,7,opt,name=snapshot_version,json=snapshotVersion" json:"snapshot_version,omitempty"` + SnapshotVersion int64 `protobuf:"varint,7,opt,name=snapshot_version,json=snapshotVersion,proto3" json:"snapshot_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/datastore.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/datastore.pb.go index 6adce562d..f3905cb10 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/datastore.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/datastore.pb.go @@ -88,11 +88,11 @@ func (ReadOptions_ReadConsistency) EnumDescriptor() ([]byte, []int) { // The request for [Datastore.Lookup][google.datastore.v1beta3.Datastore.Lookup]. type LookupRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The options for this lookup request. - ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions" json:"read_options,omitempty"` + ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions,proto3" json:"read_options,omitempty"` // Keys of entities to look up. - Keys []*Key `protobuf:"bytes,3,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,3,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -148,15 +148,15 @@ type LookupResponse struct { // Entities found as `ResultType.FULL` entities. The order of results in this // field is undefined and has no relation to the order of the keys in the // input. - Found []*EntityResult `protobuf:"bytes,1,rep,name=found" json:"found,omitempty"` + Found []*EntityResult `protobuf:"bytes,1,rep,name=found,proto3" json:"found,omitempty"` // Entities not found as `ResultType.KEY_ONLY` entities. The order of results // in this field is undefined and has no relation to the order of the keys // in the input. - Missing []*EntityResult `protobuf:"bytes,2,rep,name=missing" json:"missing,omitempty"` + Missing []*EntityResult `protobuf:"bytes,2,rep,name=missing,proto3" json:"missing,omitempty"` // A list of keys that were not looked up due to resource constraints. The // order of results in this field is undefined and has no relation to the // order of the keys in the input. - Deferred []*Key `protobuf:"bytes,3,rep,name=deferred" json:"deferred,omitempty"` + Deferred []*Key `protobuf:"bytes,3,rep,name=deferred,proto3" json:"deferred,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -210,14 +210,14 @@ func (m *LookupResponse) GetDeferred() []*Key { // The request for [Datastore.RunQuery][google.datastore.v1beta3.Datastore.RunQuery]. type RunQueryRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Entities are partitioned into subsets, identified by a partition ID. // Queries are scoped to a single partition. // This partition ID is normalized with the standard default context // partition ID. - PartitionId *PartitionId `protobuf:"bytes,2,opt,name=partition_id,json=partitionId" json:"partition_id,omitempty"` + PartitionId *PartitionId `protobuf:"bytes,2,opt,name=partition_id,json=partitionId,proto3" json:"partition_id,omitempty"` // The options for this query. - ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions" json:"read_options,omitempty"` + ReadOptions *ReadOptions `protobuf:"bytes,1,opt,name=read_options,json=readOptions,proto3" json:"read_options,omitempty"` // The type of query. // // Types that are valid to be assigned to QueryType: @@ -258,10 +258,10 @@ type isRunQueryRequest_QueryType interface { } type RunQueryRequest_Query struct { - Query *Query `protobuf:"bytes,3,opt,name=query,oneof"` + Query *Query `protobuf:"bytes,3,opt,name=query,proto3,oneof"` } type RunQueryRequest_GqlQuery struct { - GqlQuery *GqlQuery `protobuf:"bytes,7,opt,name=gql_query,json=gqlQuery,oneof"` + GqlQuery *GqlQuery `protobuf:"bytes,7,opt,name=gql_query,json=gqlQuery,proto3,oneof"` } func (*RunQueryRequest_Query) isRunQueryRequest_QueryType() {} @@ -386,9 +386,9 @@ func _RunQueryRequest_OneofSizer(msg proto.Message) (n int) { // The response for [Datastore.RunQuery][google.datastore.v1beta3.Datastore.RunQuery]. type RunQueryResponse struct { // A batch of query results (always present). - Batch *QueryResultBatch `protobuf:"bytes,1,opt,name=batch" json:"batch,omitempty"` + Batch *QueryResultBatch `protobuf:"bytes,1,opt,name=batch,proto3" json:"batch,omitempty"` // The parsed form of the `GqlQuery` from the request, if it was set. - Query *Query `protobuf:"bytes,2,opt,name=query" json:"query,omitempty"` + Query *Query `protobuf:"bytes,2,opt,name=query,proto3" json:"query,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -435,9 +435,9 @@ func (m *RunQueryResponse) GetQuery() *Query { // The request for [Datastore.BeginTransaction][google.datastore.v1beta3.Datastore.BeginTransaction]. type BeginTransactionRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Options for a new transaction. - TransactionOptions *TransactionOptions `protobuf:"bytes,10,opt,name=transaction_options,json=transactionOptions" json:"transaction_options,omitempty"` + TransactionOptions *TransactionOptions `protobuf:"bytes,10,opt,name=transaction_options,json=transactionOptions,proto3" json:"transaction_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -524,7 +524,7 @@ func (m *BeginTransactionResponse) GetTransaction() []byte { // The request for [Datastore.Rollback][google.datastore.v1beta3.Datastore.Rollback]. type RollbackRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The transaction identifier, returned by a call to // [Datastore.BeginTransaction][google.datastore.v1beta3.Datastore.BeginTransaction]. Transaction []byte `protobuf:"bytes,1,opt,name=transaction,proto3" json:"transaction,omitempty"` @@ -606,9 +606,9 @@ var xxx_messageInfo_RollbackResponse proto.InternalMessageInfo // The request for [Datastore.Commit][google.datastore.v1beta3.Datastore.Commit]. type CommitRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The type of commit to perform. Defaults to `TRANSACTIONAL`. - Mode CommitRequest_Mode `protobuf:"varint,5,opt,name=mode,enum=google.datastore.v1beta3.CommitRequest_Mode" json:"mode,omitempty"` + Mode CommitRequest_Mode `protobuf:"varint,5,opt,name=mode,proto3,enum=google.datastore.v1beta3.CommitRequest_Mode" json:"mode,omitempty"` // Must be set when mode is `TRANSACTIONAL`. // // Types that are valid to be assigned to TransactionSelector: @@ -627,7 +627,7 @@ type CommitRequest struct { // // When mode is `NON_TRANSACTIONAL`, no two mutations may affect a single // entity. - Mutations []*Mutation `protobuf:"bytes,6,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,6,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -757,10 +757,10 @@ func _CommitRequest_OneofSizer(msg proto.Message) (n int) { type CommitResponse struct { // The result of performing the mutations. // The i-th mutation result corresponds to the i-th mutation in the request. - MutationResults []*MutationResult `protobuf:"bytes,3,rep,name=mutation_results,json=mutationResults" json:"mutation_results,omitempty"` + MutationResults []*MutationResult `protobuf:"bytes,3,rep,name=mutation_results,json=mutationResults,proto3" json:"mutation_results,omitempty"` // The number of index entries updated during the commit, or zero if none were // updated. - IndexUpdates int32 `protobuf:"varint,4,opt,name=index_updates,json=indexUpdates" json:"index_updates,omitempty"` + IndexUpdates int32 `protobuf:"varint,4,opt,name=index_updates,json=indexUpdates,proto3" json:"index_updates,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -807,10 +807,10 @@ func (m *CommitResponse) GetIndexUpdates() int32 { // The request for [Datastore.AllocateIds][google.datastore.v1beta3.Datastore.AllocateIds]. type AllocateIdsRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // A list of keys with incomplete key paths for which to allocate IDs. // No key may be reserved/read-only. - Keys []*Key `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -858,7 +858,7 @@ func (m *AllocateIdsRequest) GetKeys() []*Key { type AllocateIdsResponse struct { // The keys specified in the request (in the same order), each with // its key path completed with a newly allocated ID. - Keys []*Key `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -898,12 +898,12 @@ func (m *AllocateIdsResponse) GetKeys() []*Key { // The request for [Datastore.ReserveIds][google.datastore.v1beta3.Datastore.ReserveIds]. type ReserveIdsRequest struct { // The ID of the project against which to make the request. - ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,8,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // If not empty, the ID of the database against which to make the request. - DatabaseId string `protobuf:"bytes,9,opt,name=database_id,json=databaseId" json:"database_id,omitempty"` + DatabaseId string `protobuf:"bytes,9,opt,name=database_id,json=databaseId,proto3" json:"database_id,omitempty"` // A list of keys with complete key paths whose numeric IDs should not be // auto-allocated. - Keys []*Key `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*Key `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1046,19 +1046,19 @@ type isMutation_ConflictDetectionStrategy interface { } type Mutation_Insert struct { - Insert *Entity `protobuf:"bytes,4,opt,name=insert,oneof"` + Insert *Entity `protobuf:"bytes,4,opt,name=insert,proto3,oneof"` } type Mutation_Update struct { - Update *Entity `protobuf:"bytes,5,opt,name=update,oneof"` + Update *Entity `protobuf:"bytes,5,opt,name=update,proto3,oneof"` } type Mutation_Upsert struct { - Upsert *Entity `protobuf:"bytes,6,opt,name=upsert,oneof"` + Upsert *Entity `protobuf:"bytes,6,opt,name=upsert,proto3,oneof"` } type Mutation_Delete struct { - Delete *Key `protobuf:"bytes,7,opt,name=delete,oneof"` + Delete *Key `protobuf:"bytes,7,opt,name=delete,proto3,oneof"` } type Mutation_BaseVersion struct { - BaseVersion int64 `protobuf:"varint,8,opt,name=base_version,json=baseVersion,oneof"` + BaseVersion int64 `protobuf:"varint,8,opt,name=base_version,json=baseVersion,proto3,oneof"` } func (*Mutation_Insert) isMutation_Operation() {} @@ -1257,16 +1257,16 @@ func _Mutation_OneofSizer(msg proto.Message) (n int) { type MutationResult struct { // The automatically allocated key. // Set only when the mutation allocated a key. - Key *Key `protobuf:"bytes,3,opt,name=key" json:"key,omitempty"` + Key *Key `protobuf:"bytes,3,opt,name=key,proto3" json:"key,omitempty"` // The version of the entity on the server after processing the mutation. If // the mutation doesn't change anything on the server, then the version will // be the version of the current entity or, if no entity is present, a version // that is strictly greater than the version of any previous entity and less // than the version of any possible future entity. - Version int64 `protobuf:"varint,4,opt,name=version" json:"version,omitempty"` + Version int64 `protobuf:"varint,4,opt,name=version,proto3" json:"version,omitempty"` // Whether a conflict was detected for this mutation. Always false when a // conflict detection strategy field is not set in the mutation. - ConflictDetected bool `protobuf:"varint,5,opt,name=conflict_detected,json=conflictDetected" json:"conflict_detected,omitempty"` + ConflictDetected bool `protobuf:"varint,5,opt,name=conflict_detected,json=conflictDetected,proto3" json:"conflict_detected,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1361,7 +1361,7 @@ type isReadOptions_ConsistencyType interface { } type ReadOptions_ReadConsistency_ struct { - ReadConsistency ReadOptions_ReadConsistency `protobuf:"varint,1,opt,name=read_consistency,json=readConsistency,enum=google.datastore.v1beta3.ReadOptions_ReadConsistency,oneof"` + ReadConsistency ReadOptions_ReadConsistency `protobuf:"varint,1,opt,name=read_consistency,json=readConsistency,proto3,enum=google.datastore.v1beta3.ReadOptions_ReadConsistency,oneof"` } type ReadOptions_Transaction struct { Transaction []byte `protobuf:"bytes,2,opt,name=transaction,proto3,oneof"` @@ -1503,10 +1503,10 @@ type isTransactionOptions_Mode interface { } type TransactionOptions_ReadWrite_ struct { - ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,1,opt,name=read_write,json=readWrite,oneof"` + ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,1,opt,name=read_write,json=readWrite,proto3,oneof"` } type TransactionOptions_ReadOnly_ struct { - ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,oneof"` + ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,proto3,oneof"` } func (*TransactionOptions_ReadWrite_) isTransactionOptions_Mode() {} @@ -1805,8 +1805,7 @@ func (c *datastoreClient) ReserveIds(ctx context.Context, in *ReserveIdsRequest, return out, nil } -// Server API for Datastore service - +// DatastoreServer is the server API for Datastore service. type DatastoreServer interface { // Looks up entities by key. Lookup(context.Context, *LookupRequest) (*LookupResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/entity.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/entity.pb.go index e2539c4bc..18b8be93e 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/entity.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/entity.pb.go @@ -43,9 +43,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Reads and writes of foreign partition IDs may fail if the project is not in an active state. type PartitionId struct { // The ID of the project to which the entities belong. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // If not empty, the ID of the namespace to which the entities belong. - NamespaceId string `protobuf:"bytes,4,opt,name=namespace_id,json=namespaceId" json:"namespace_id,omitempty"` + NamespaceId string `protobuf:"bytes,4,opt,name=namespace_id,json=namespaceId,proto3" json:"namespace_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -97,7 +97,7 @@ type Key struct { // Entities are partitioned into subsets, currently identified by a project // ID and namespace ID. // Queries are scoped to a single partition. - PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId" json:"partition_id,omitempty"` + PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId,proto3" json:"partition_id,omitempty"` // The entity path. // An entity path consists of one or more elements composed of a kind and a // string or numerical identifier, which identify entities. The first @@ -114,7 +114,7 @@ type Key struct { // identifier. // // A path can never be empty, and a path can have at most 100 elements. - Path []*Key_PathElement `protobuf:"bytes,2,rep,name=path" json:"path,omitempty"` + Path []*Key_PathElement `protobuf:"bytes,2,rep,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -167,7 +167,7 @@ type Key_PathElement struct { // A kind matching regex `__.*__` is reserved/read-only. // A kind must not contain more than 1500 bytes when UTF-8 encoded. // Cannot be `""`. - Kind string `protobuf:"bytes,1,opt,name=kind" json:"kind,omitempty"` + Kind string `protobuf:"bytes,1,opt,name=kind,proto3" json:"kind,omitempty"` // The type of ID. // // Types that are valid to be assigned to IdType: @@ -208,10 +208,10 @@ type isKey_PathElement_IdType interface { } type Key_PathElement_Id struct { - Id int64 `protobuf:"varint,2,opt,name=id,oneof"` + Id int64 `protobuf:"varint,2,opt,name=id,proto3,oneof"` } type Key_PathElement_Name struct { - Name string `protobuf:"bytes,3,opt,name=name,oneof"` + Name string `protobuf:"bytes,3,opt,name=name,proto3,oneof"` } func (*Key_PathElement_Id) isKey_PathElement_IdType() {} @@ -315,7 +315,7 @@ type ArrayValue struct { // Values in the array. // The order of this array may not be preserved if it contains a mix of // indexed and unindexed values. - Values []*Value `protobuf:"bytes,1,rep,name=values" json:"values,omitempty"` + Values []*Value `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -371,10 +371,10 @@ type Value struct { // *Value_ArrayValue ValueType isValue_ValueType `protobuf_oneof:"value_type"` // The `meaning` field should only be populated for backwards compatibility. - Meaning int32 `protobuf:"varint,14,opt,name=meaning" json:"meaning,omitempty"` + Meaning int32 `protobuf:"varint,14,opt,name=meaning,proto3" json:"meaning,omitempty"` // If the value should be excluded from all indexes including those defined // explicitly. - ExcludeFromIndexes bool `protobuf:"varint,19,opt,name=exclude_from_indexes,json=excludeFromIndexes" json:"exclude_from_indexes,omitempty"` + ExcludeFromIndexes bool `protobuf:"varint,19,opt,name=exclude_from_indexes,json=excludeFromIndexes,proto3" json:"exclude_from_indexes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,37 +409,37 @@ type isValue_ValueType interface { } type Value_NullValue struct { - NullValue _struct.NullValue `protobuf:"varint,11,opt,name=null_value,json=nullValue,enum=google.protobuf.NullValue,oneof"` + NullValue _struct.NullValue `protobuf:"varint,11,opt,name=null_value,json=nullValue,proto3,enum=google.protobuf.NullValue,oneof"` } type Value_BooleanValue struct { - BooleanValue bool `protobuf:"varint,1,opt,name=boolean_value,json=booleanValue,oneof"` + BooleanValue bool `protobuf:"varint,1,opt,name=boolean_value,json=booleanValue,proto3,oneof"` } type Value_IntegerValue struct { - IntegerValue int64 `protobuf:"varint,2,opt,name=integer_value,json=integerValue,oneof"` + IntegerValue int64 `protobuf:"varint,2,opt,name=integer_value,json=integerValue,proto3,oneof"` } type Value_DoubleValue struct { - DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,oneof"` + DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,proto3,oneof"` } type Value_TimestampValue struct { - TimestampValue *timestamp.Timestamp `protobuf:"bytes,10,opt,name=timestamp_value,json=timestampValue,oneof"` + TimestampValue *timestamp.Timestamp `protobuf:"bytes,10,opt,name=timestamp_value,json=timestampValue,proto3,oneof"` } type Value_KeyValue struct { - KeyValue *Key `protobuf:"bytes,5,opt,name=key_value,json=keyValue,oneof"` + KeyValue *Key `protobuf:"bytes,5,opt,name=key_value,json=keyValue,proto3,oneof"` } type Value_StringValue struct { - StringValue string `protobuf:"bytes,17,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,17,opt,name=string_value,json=stringValue,proto3,oneof"` } type Value_BlobValue struct { BlobValue []byte `protobuf:"bytes,18,opt,name=blob_value,json=blobValue,proto3,oneof"` } type Value_GeoPointValue struct { - GeoPointValue *latlng.LatLng `protobuf:"bytes,8,opt,name=geo_point_value,json=geoPointValue,oneof"` + GeoPointValue *latlng.LatLng `protobuf:"bytes,8,opt,name=geo_point_value,json=geoPointValue,proto3,oneof"` } type Value_EntityValue struct { - EntityValue *Entity `protobuf:"bytes,6,opt,name=entity_value,json=entityValue,oneof"` + EntityValue *Entity `protobuf:"bytes,6,opt,name=entity_value,json=entityValue,proto3,oneof"` } type Value_ArrayValue struct { - ArrayValue *ArrayValue `protobuf:"bytes,9,opt,name=array_value,json=arrayValue,oneof"` + ArrayValue *ArrayValue `protobuf:"bytes,9,opt,name=array_value,json=arrayValue,proto3,oneof"` } func (*Value_NullValue) isValue_ValueType() {} @@ -785,14 +785,14 @@ type Entity struct { // an entity in `Value.entity_value` may have no key). // An entity's kind is its key path's last element's kind, // or null if it has no key. - Key *Key `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key *Key `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // The entity's properties. // The map's keys are property names. // A property name matching regex `__.*__` is reserved. // A reserved property name is forbidden in certain documented contexts. // The name must not contain more than 500 characters. // The name cannot be `""`. - Properties map[string]*Value `protobuf:"bytes,3,rep,name=properties" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Properties map[string]*Value `protobuf:"bytes,3,rep,name=properties,proto3" json:"properties,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/query.pb.go b/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/query.pb.go index 72bc549f1..559802ad4 100644 --- a/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/query.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/datastore/v1beta3/query.pb.go @@ -203,7 +203,7 @@ func (QueryResultBatch_MoreResultsType) EnumDescriptor() ([]byte, []int) { // The result of fetching an entity from Datastore. type EntityResult struct { // The resulting entity. - Entity *Entity `protobuf:"bytes,1,opt,name=entity" json:"entity,omitempty"` + Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"` // The version of the entity, a strictly positive number that monotonically // increases with changes to the entity. // @@ -213,7 +213,7 @@ type EntityResult struct { // For [missing][google.datastore.v1beta3.LookupResponse.missing] entities in `LookupResponse`, this // is the version of the snapshot that was used to look up the entity, and it // is always set except for eventually consistent reads. - Version int64 `protobuf:"varint,4,opt,name=version" json:"version,omitempty"` + Version int64 `protobuf:"varint,4,opt,name=version,proto3" json:"version,omitempty"` // A cursor that points to the position after the result entity. // Set only when the `EntityResult` is part of a `QueryResultBatch` message. Cursor []byte `protobuf:"bytes,3,opt,name=cursor,proto3" json:"cursor,omitempty"` @@ -270,18 +270,18 @@ func (m *EntityResult) GetCursor() []byte { // A query for entities. type Query struct { // The projection to return. Defaults to returning all properties. - Projection []*Projection `protobuf:"bytes,2,rep,name=projection" json:"projection,omitempty"` + Projection []*Projection `protobuf:"bytes,2,rep,name=projection,proto3" json:"projection,omitempty"` // The kinds to query (if empty, returns entities of all kinds). // Currently at most 1 kind may be specified. - Kind []*KindExpression `protobuf:"bytes,3,rep,name=kind" json:"kind,omitempty"` + Kind []*KindExpression `protobuf:"bytes,3,rep,name=kind,proto3" json:"kind,omitempty"` // The filter to apply. - Filter *Filter `protobuf:"bytes,4,opt,name=filter" json:"filter,omitempty"` + Filter *Filter `protobuf:"bytes,4,opt,name=filter,proto3" json:"filter,omitempty"` // The order to apply to the query results (if empty, order is unspecified). - Order []*PropertyOrder `protobuf:"bytes,5,rep,name=order" json:"order,omitempty"` + Order []*PropertyOrder `protobuf:"bytes,5,rep,name=order,proto3" json:"order,omitempty"` // The properties to make distinct. The query results will contain the first // result for each distinct combination of values for the given properties // (if empty, all results are returned). - DistinctOn []*PropertyReference `protobuf:"bytes,6,rep,name=distinct_on,json=distinctOn" json:"distinct_on,omitempty"` + DistinctOn []*PropertyReference `protobuf:"bytes,6,rep,name=distinct_on,json=distinctOn,proto3" json:"distinct_on,omitempty"` // A starting point for the query results. Query cursors are // returned in query result batches and // [can only be used to continue the same query](https://cloud.google.com/datastore/docs/concepts/queries#cursors_limits_and_offsets). @@ -292,12 +292,12 @@ type Query struct { EndCursor []byte `protobuf:"bytes,8,opt,name=end_cursor,json=endCursor,proto3" json:"end_cursor,omitempty"` // The number of results to skip. Applies before limit, but after all other // constraints. Optional. Must be >= 0 if specified. - Offset int32 `protobuf:"varint,10,opt,name=offset" json:"offset,omitempty"` + Offset int32 `protobuf:"varint,10,opt,name=offset,proto3" json:"offset,omitempty"` // The maximum number of results to return. Applies after all other // constraints. Optional. // Unspecified is interpreted as no limit. // Must be >= 0 if specified. - Limit *wrappers.Int32Value `protobuf:"bytes,12,opt,name=limit" json:"limit,omitempty"` + Limit *wrappers.Int32Value `protobuf:"bytes,12,opt,name=limit,proto3" json:"limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -393,7 +393,7 @@ func (m *Query) GetLimit() *wrappers.Int32Value { // A representation of a kind. type KindExpression struct { // The name of the kind. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -434,7 +434,7 @@ func (m *KindExpression) GetName() string { type PropertyReference struct { // The name of the property. // If name includes "."s, it may be interpreted as a property name path. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -474,7 +474,7 @@ func (m *PropertyReference) GetName() string { // A representation of a property in a projection. type Projection struct { // The property to project. - Property *PropertyReference `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *PropertyReference `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -514,9 +514,9 @@ func (m *Projection) GetProperty() *PropertyReference { // The desired order for a specific property. type PropertyOrder struct { // The property to order by. - Property *PropertyReference `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *PropertyReference `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The direction to order by. Defaults to `ASCENDING`. - Direction PropertyOrder_Direction `protobuf:"varint,2,opt,name=direction,enum=google.datastore.v1beta3.PropertyOrder_Direction" json:"direction,omitempty"` + Direction PropertyOrder_Direction `protobuf:"varint,2,opt,name=direction,proto3,enum=google.datastore.v1beta3.PropertyOrder_Direction" json:"direction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -602,10 +602,10 @@ type isFilter_FilterType interface { } type Filter_CompositeFilter struct { - CompositeFilter *CompositeFilter `protobuf:"bytes,1,opt,name=composite_filter,json=compositeFilter,oneof"` + CompositeFilter *CompositeFilter `protobuf:"bytes,1,opt,name=composite_filter,json=compositeFilter,proto3,oneof"` } type Filter_PropertyFilter struct { - PropertyFilter *PropertyFilter `protobuf:"bytes,2,opt,name=property_filter,json=propertyFilter,oneof"` + PropertyFilter *PropertyFilter `protobuf:"bytes,2,opt,name=property_filter,json=propertyFilter,proto3,oneof"` } func (*Filter_CompositeFilter) isFilter_FilterType() {} @@ -709,10 +709,10 @@ func _Filter_OneofSizer(msg proto.Message) (n int) { // A filter that merges multiple other filters using the given operator. type CompositeFilter struct { // The operator for combining multiple filters. - Op CompositeFilter_Operator `protobuf:"varint,1,opt,name=op,enum=google.datastore.v1beta3.CompositeFilter_Operator" json:"op,omitempty"` + Op CompositeFilter_Operator `protobuf:"varint,1,opt,name=op,proto3,enum=google.datastore.v1beta3.CompositeFilter_Operator" json:"op,omitempty"` // The list of filters to combine. // Must contain at least one filter. - Filters []*Filter `protobuf:"bytes,2,rep,name=filters" json:"filters,omitempty"` + Filters []*Filter `protobuf:"bytes,2,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -759,11 +759,11 @@ func (m *CompositeFilter) GetFilters() []*Filter { // A filter on a specific property. type PropertyFilter struct { // The property to filter by. - Property *PropertyReference `protobuf:"bytes,1,opt,name=property" json:"property,omitempty"` + Property *PropertyReference `protobuf:"bytes,1,opt,name=property,proto3" json:"property,omitempty"` // The operator to filter by. - Op PropertyFilter_Operator `protobuf:"varint,2,opt,name=op,enum=google.datastore.v1beta3.PropertyFilter_Operator" json:"op,omitempty"` + Op PropertyFilter_Operator `protobuf:"varint,2,opt,name=op,proto3,enum=google.datastore.v1beta3.PropertyFilter_Operator" json:"op,omitempty"` // The value to compare the property to. - Value *Value `protobuf:"bytes,3,opt,name=value" json:"value,omitempty"` + Value *Value `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -818,24 +818,24 @@ func (m *PropertyFilter) GetValue() *Value { type GqlQuery struct { // A string of the format described // [here](https://cloud.google.com/datastore/docs/apis/gql/gql_reference). - QueryString string `protobuf:"bytes,1,opt,name=query_string,json=queryString" json:"query_string,omitempty"` + QueryString string `protobuf:"bytes,1,opt,name=query_string,json=queryString,proto3" json:"query_string,omitempty"` // When false, the query string must not contain any literals and instead must // bind all values. For example, // `SELECT * FROM Kind WHERE a = 'string literal'` is not allowed, while // `SELECT * FROM Kind WHERE a = @value` is. - AllowLiterals bool `protobuf:"varint,2,opt,name=allow_literals,json=allowLiterals" json:"allow_literals,omitempty"` + AllowLiterals bool `protobuf:"varint,2,opt,name=allow_literals,json=allowLiterals,proto3" json:"allow_literals,omitempty"` // For each non-reserved named binding site in the query string, there must be // a named parameter with that name, but not necessarily the inverse. // // Key must match regex `[A-Za-z_$][A-Za-z_$0-9]*`, must not match regex // `__.*__`, and must not be `""`. - NamedBindings map[string]*GqlQueryParameter `protobuf:"bytes,5,rep,name=named_bindings,json=namedBindings" json:"named_bindings,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + NamedBindings map[string]*GqlQueryParameter `protobuf:"bytes,5,rep,name=named_bindings,json=namedBindings,proto3" json:"named_bindings,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Numbered binding site @1 references the first numbered parameter, // effectively using 1-based indexing, rather than the usual 0. // // For each binding site numbered i in `query_string`, there must be an i-th // numbered parameter. The inverse must also be true. - PositionalBindings []*GqlQueryParameter `protobuf:"bytes,4,rep,name=positional_bindings,json=positionalBindings" json:"positional_bindings,omitempty"` + PositionalBindings []*GqlQueryParameter `protobuf:"bytes,4,rep,name=positional_bindings,json=positionalBindings,proto3" json:"positional_bindings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -935,7 +935,7 @@ type isGqlQueryParameter_ParameterType interface { } type GqlQueryParameter_Value struct { - Value *Value `protobuf:"bytes,2,opt,name=value,oneof"` + Value *Value `protobuf:"bytes,2,opt,name=value,proto3,oneof"` } type GqlQueryParameter_Cursor struct { Cursor []byte `protobuf:"bytes,3,opt,name=cursor,proto3,oneof"` @@ -1038,18 +1038,18 @@ func _GqlQueryParameter_OneofSizer(msg proto.Message) (n int) { // A batch of results produced by a query. type QueryResultBatch struct { // The number of results skipped, typically because of an offset. - SkippedResults int32 `protobuf:"varint,6,opt,name=skipped_results,json=skippedResults" json:"skipped_results,omitempty"` + SkippedResults int32 `protobuf:"varint,6,opt,name=skipped_results,json=skippedResults,proto3" json:"skipped_results,omitempty"` // A cursor that points to the position after the last skipped result. // Will be set when `skipped_results` != 0. SkippedCursor []byte `protobuf:"bytes,3,opt,name=skipped_cursor,json=skippedCursor,proto3" json:"skipped_cursor,omitempty"` // The result type for every entity in `entity_results`. - EntityResultType EntityResult_ResultType `protobuf:"varint,1,opt,name=entity_result_type,json=entityResultType,enum=google.datastore.v1beta3.EntityResult_ResultType" json:"entity_result_type,omitempty"` + EntityResultType EntityResult_ResultType `protobuf:"varint,1,opt,name=entity_result_type,json=entityResultType,proto3,enum=google.datastore.v1beta3.EntityResult_ResultType" json:"entity_result_type,omitempty"` // The results for this batch. - EntityResults []*EntityResult `protobuf:"bytes,2,rep,name=entity_results,json=entityResults" json:"entity_results,omitempty"` + EntityResults []*EntityResult `protobuf:"bytes,2,rep,name=entity_results,json=entityResults,proto3" json:"entity_results,omitempty"` // A cursor that points to the position after the last result in the batch. EndCursor []byte `protobuf:"bytes,4,opt,name=end_cursor,json=endCursor,proto3" json:"end_cursor,omitempty"` // The state of the query after the current batch. - MoreResults QueryResultBatch_MoreResultsType `protobuf:"varint,5,opt,name=more_results,json=moreResults,enum=google.datastore.v1beta3.QueryResultBatch_MoreResultsType" json:"more_results,omitempty"` + MoreResults QueryResultBatch_MoreResultsType `protobuf:"varint,5,opt,name=more_results,json=moreResults,proto3,enum=google.datastore.v1beta3.QueryResultBatch_MoreResultsType" json:"more_results,omitempty"` // The version number of the snapshot this batch was returned from. // This applies to the range of results from the query's `start_cursor` (or // the beginning of the query if no cursor was given) to this batch's @@ -1059,7 +1059,7 @@ type QueryResultBatch struct { // can have a greater snapshot version number. Each batch's snapshot version // is valid for all preceding batches. // The value will be zero for eventually consistent queries. - SnapshotVersion int64 `protobuf:"varint,7,opt,name=snapshot_version,json=snapshotVersion" json:"snapshot_version,omitempty"` + SnapshotVersion int64 `protobuf:"varint,7,opt,name=snapshot_version,json=snapshotVersion,proto3" json:"snapshot_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_events.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_events.pb.go index eb20dd0d4..04df02c6c 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_events.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_events.pb.go @@ -125,7 +125,7 @@ func (StreamId_BuildComponent) EnumDescriptor() ([]byte, []int) { // message does not include field for uniquely identifying an event. type BuildEvent struct { // The timestamp of this event. - EventTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=event_time,json=eventTime" json:"event_time,omitempty"` + EventTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=event_time,json=eventTime,proto3" json:"event_time,omitempty"` // ////////////////////////////////////////////////////////////////////////// // Events that indicate a state change of a build request in the build // queue. @@ -175,31 +175,31 @@ type isBuildEvent_Event interface { } type BuildEvent_InvocationAttemptStarted_ struct { - InvocationAttemptStarted *BuildEvent_InvocationAttemptStarted `protobuf:"bytes,51,opt,name=invocation_attempt_started,json=invocationAttemptStarted,oneof"` + InvocationAttemptStarted *BuildEvent_InvocationAttemptStarted `protobuf:"bytes,51,opt,name=invocation_attempt_started,json=invocationAttemptStarted,proto3,oneof"` } type BuildEvent_InvocationAttemptFinished_ struct { - InvocationAttemptFinished *BuildEvent_InvocationAttemptFinished `protobuf:"bytes,52,opt,name=invocation_attempt_finished,json=invocationAttemptFinished,oneof"` + InvocationAttemptFinished *BuildEvent_InvocationAttemptFinished `protobuf:"bytes,52,opt,name=invocation_attempt_finished,json=invocationAttemptFinished,proto3,oneof"` } type BuildEvent_BuildEnqueued_ struct { - BuildEnqueued *BuildEvent_BuildEnqueued `protobuf:"bytes,53,opt,name=build_enqueued,json=buildEnqueued,oneof"` + BuildEnqueued *BuildEvent_BuildEnqueued `protobuf:"bytes,53,opt,name=build_enqueued,json=buildEnqueued,proto3,oneof"` } type BuildEvent_BuildFinished_ struct { - BuildFinished *BuildEvent_BuildFinished `protobuf:"bytes,55,opt,name=build_finished,json=buildFinished,oneof"` + BuildFinished *BuildEvent_BuildFinished `protobuf:"bytes,55,opt,name=build_finished,json=buildFinished,proto3,oneof"` } type BuildEvent_ConsoleOutput_ struct { - ConsoleOutput *BuildEvent_ConsoleOutput `protobuf:"bytes,56,opt,name=console_output,json=consoleOutput,oneof"` + ConsoleOutput *BuildEvent_ConsoleOutput `protobuf:"bytes,56,opt,name=console_output,json=consoleOutput,proto3,oneof"` } type BuildEvent_ComponentStreamFinished struct { - ComponentStreamFinished *BuildEvent_BuildComponentStreamFinished `protobuf:"bytes,59,opt,name=component_stream_finished,json=componentStreamFinished,oneof"` + ComponentStreamFinished *BuildEvent_BuildComponentStreamFinished `protobuf:"bytes,59,opt,name=component_stream_finished,json=componentStreamFinished,proto3,oneof"` } type BuildEvent_BazelEvent struct { - BazelEvent *any.Any `protobuf:"bytes,60,opt,name=bazel_event,json=bazelEvent,oneof"` + BazelEvent *any.Any `protobuf:"bytes,60,opt,name=bazel_event,json=bazelEvent,proto3,oneof"` } type BuildEvent_BuildExecutionEvent struct { - BuildExecutionEvent *any.Any `protobuf:"bytes,61,opt,name=build_execution_event,json=buildExecutionEvent,oneof"` + BuildExecutionEvent *any.Any `protobuf:"bytes,61,opt,name=build_execution_event,json=buildExecutionEvent,proto3,oneof"` } type BuildEvent_SourceFetchEvent struct { - SourceFetchEvent *any.Any `protobuf:"bytes,62,opt,name=source_fetch_event,json=sourceFetchEvent,oneof"` + SourceFetchEvent *any.Any `protobuf:"bytes,62,opt,name=source_fetch_event,json=sourceFetchEvent,proto3,oneof"` } func (*BuildEvent_InvocationAttemptStarted_) isBuildEvent_Event() {} @@ -501,7 +501,7 @@ type BuildEvent_InvocationAttemptStarted struct { // The number of the invocation attempt, starting at 1 and increasing by 1 // for each new attempt. Can be used to determine if there is a later // invocation attempt replacing the current one a client is processing. - AttemptNumber int64 `protobuf:"varint,1,opt,name=attempt_number,json=attemptNumber" json:"attempt_number,omitempty"` + AttemptNumber int64 `protobuf:"varint,1,opt,name=attempt_number,json=attemptNumber,proto3" json:"attempt_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -541,9 +541,9 @@ func (m *BuildEvent_InvocationAttemptStarted) GetAttemptNumber() int64 { // Notification that an invocation attempt has finished. type BuildEvent_InvocationAttemptFinished struct { // The exit code of the build tool. - ExitCode *wrappers.Int32Value `protobuf:"bytes,2,opt,name=exit_code,json=exitCode" json:"exit_code,omitempty"` + ExitCode *wrappers.Int32Value `protobuf:"bytes,2,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"` // Final status of the invocation. - InvocationStatus *BuildStatus `protobuf:"bytes,3,opt,name=invocation_status,json=invocationStatus" json:"invocation_status,omitempty"` + InvocationStatus *BuildStatus `protobuf:"bytes,3,opt,name=invocation_status,json=invocationStatus,proto3" json:"invocation_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -626,7 +626,7 @@ var xxx_messageInfo_BuildEvent_BuildEnqueued proto.InternalMessageInfo // Individual invocations trigger InvocationFinished when they finish. type BuildEvent_BuildFinished struct { // Final status of the build. - Status *BuildStatus `protobuf:"bytes,1,opt,name=status" json:"status,omitempty"` + Status *BuildStatus `protobuf:"bytes,1,opt,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -666,7 +666,7 @@ func (m *BuildEvent_BuildFinished) GetStatus() *BuildStatus { // Textual output written to standard output or standard error. type BuildEvent_ConsoleOutput struct { // The output stream type. - Type ConsoleOutputStream `protobuf:"varint,1,opt,name=type,enum=google.devtools.build.v1.ConsoleOutputStream" json:"type,omitempty"` + Type ConsoleOutputStream `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.build.v1.ConsoleOutputStream" json:"type,omitempty"` // The output stream content. // // Types that are valid to be assigned to Output: @@ -707,7 +707,7 @@ type isBuildEvent_ConsoleOutput_Output interface { } type BuildEvent_ConsoleOutput_TextOutput struct { - TextOutput string `protobuf:"bytes,2,opt,name=text_output,json=textOutput,oneof"` + TextOutput string `protobuf:"bytes,2,opt,name=text_output,json=textOutput,proto3,oneof"` } type BuildEvent_ConsoleOutput_BinaryOutput struct { BinaryOutput []byte `protobuf:"bytes,3,opt,name=binary_output,json=binaryOutput,proto3,oneof"` @@ -814,7 +814,7 @@ func _BuildEvent_ConsoleOutput_OneofSizer(msg proto.Message) (n int) { // component other than CONTROLLER (See StreamId.BuildComponents). type BuildEvent_BuildComponentStreamFinished struct { // How the event stream finished. - Type BuildEvent_BuildComponentStreamFinished_FinishType `protobuf:"varint,1,opt,name=type,enum=google.devtools.build.v1.BuildEvent_BuildComponentStreamFinished_FinishType" json:"type,omitempty"` + Type BuildEvent_BuildComponentStreamFinished_FinishType `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.build.v1.BuildEvent_BuildComponentStreamFinished_FinishType" json:"type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -856,12 +856,12 @@ func (m *BuildEvent_BuildComponentStreamFinished) GetType() BuildEvent_BuildComp // Unique identifier for a build event stream. type StreamId struct { // The id of a Build message. - BuildId string `protobuf:"bytes,1,opt,name=build_id,json=buildId" json:"build_id,omitempty"` + BuildId string `protobuf:"bytes,1,opt,name=build_id,json=buildId,proto3" json:"build_id,omitempty"` // The unique invocation ID within this build. // It should be the same as {invocation} (below) during the migration. - InvocationId string `protobuf:"bytes,6,opt,name=invocation_id,json=invocationId" json:"invocation_id,omitempty"` + InvocationId string `protobuf:"bytes,6,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` // The component that emitted this event. - Component StreamId_BuildComponent `protobuf:"varint,3,opt,name=component,enum=google.devtools.build.v1.StreamId_BuildComponent" json:"component,omitempty"` + Component StreamId_BuildComponent `protobuf:"varint,3,opt,name=component,proto3,enum=google.devtools.build.v1.StreamId_BuildComponent" json:"component,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_status.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_status.pb.go index c8b4216e2..a362821e7 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_status.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/build_status.pb.go @@ -77,9 +77,9 @@ func (BuildStatus_Result) EnumDescriptor() ([]byte, []int) { // Status used for both invocation attempt and overall build completion. type BuildStatus struct { // The end result. - Result BuildStatus_Result `protobuf:"varint,1,opt,name=result,enum=google.devtools.build.v1.BuildStatus_Result" json:"result,omitempty"` + Result BuildStatus_Result `protobuf:"varint,1,opt,name=result,proto3,enum=google.devtools.build.v1.BuildStatus_Result" json:"result,omitempty"` // Fine-grained diagnostic information to complement the status. - Details *any.Any `protobuf:"bytes,2,opt,name=details" json:"details,omitempty"` + Details *any.Any `protobuf:"bytes,2,opt,name=details,proto3" json:"details,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/publish_build_event.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/publish_build_event.pb.go index d6e042181..64cdd6bd9 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/publish_build_event.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/build/v1/publish_build_event.pb.go @@ -63,23 +63,23 @@ func (PublishLifecycleEventRequest_ServiceLevel) EnumDescriptor() ([]byte, []int // - BuildFinished: When a build is finished. type PublishLifecycleEventRequest struct { // The interactivity of this build. - ServiceLevel PublishLifecycleEventRequest_ServiceLevel `protobuf:"varint,1,opt,name=service_level,json=serviceLevel,enum=google.devtools.build.v1.PublishLifecycleEventRequest_ServiceLevel" json:"service_level,omitempty"` + ServiceLevel PublishLifecycleEventRequest_ServiceLevel `protobuf:"varint,1,opt,name=service_level,json=serviceLevel,proto3,enum=google.devtools.build.v1.PublishLifecycleEventRequest_ServiceLevel" json:"service_level,omitempty"` // The lifecycle build event. If this is a build tool event, the RPC will fail // with INVALID_REQUEST. - BuildEvent *OrderedBuildEvent `protobuf:"bytes,2,opt,name=build_event,json=buildEvent" json:"build_event,omitempty"` + BuildEvent *OrderedBuildEvent `protobuf:"bytes,2,opt,name=build_event,json=buildEvent,proto3" json:"build_event,omitempty"` // If the next event for this build or invocation (depending on the event // type) hasn't been published after this duration from when {build_event} // is written to BES, consider this stream expired. If this field is not set, // BES backend will use its own default value. - StreamTimeout *duration.Duration `protobuf:"bytes,3,opt,name=stream_timeout,json=streamTimeout" json:"stream_timeout,omitempty"` + StreamTimeout *duration.Duration `protobuf:"bytes,3,opt,name=stream_timeout,json=streamTimeout,proto3" json:"stream_timeout,omitempty"` // Additional information about a build request. These are define by the event // publishers, and the Build Event Service does not validate or interpret // them. They are used while notifying internal systems of new builds and // invocations if the OrderedBuildEvent.event type is // BuildEnqueued/InvocationAttemptStarted. - NotificationKeywords []string `protobuf:"bytes,4,rep,name=notification_keywords,json=notificationKeywords" json:"notification_keywords,omitempty"` + NotificationKeywords []string `protobuf:"bytes,4,rep,name=notification_keywords,json=notificationKeywords,proto3" json:"notification_keywords,omitempty"` // This field identifies which project (if any) the build is associated with. - ProjectId string `protobuf:"bytes,6,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,6,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -148,9 +148,9 @@ func (m *PublishLifecycleEventRequest) GetProjectId() string { // RPC errors, hence not recorded by this proto. type PublishBuildToolEventStreamResponse struct { // The stream that contains this event. - StreamId *StreamId `protobuf:"bytes,1,opt,name=stream_id,json=streamId" json:"stream_id,omitempty"` + StreamId *StreamId `protobuf:"bytes,1,opt,name=stream_id,json=streamId,proto3" json:"stream_id,omitempty"` // The sequence number of this event that has been committed. - SequenceNumber int64 `protobuf:"varint,2,opt,name=sequence_number,json=sequenceNumber" json:"sequence_number,omitempty"` + SequenceNumber int64 `protobuf:"varint,2,opt,name=sequence_number,json=sequenceNumber,proto3" json:"sequence_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -198,13 +198,13 @@ func (m *PublishBuildToolEventStreamResponse) GetSequenceNumber() int64 { // its position in that stream. type OrderedBuildEvent struct { // Which build event stream this event belongs to. - StreamId *StreamId `protobuf:"bytes,1,opt,name=stream_id,json=streamId" json:"stream_id,omitempty"` + StreamId *StreamId `protobuf:"bytes,1,opt,name=stream_id,json=streamId,proto3" json:"stream_id,omitempty"` // The position of this event in the stream. The sequence numbers for a build // event stream should be a sequence of consecutive natural numbers starting // from one. (1, 2, 3, ...) - SequenceNumber int64 `protobuf:"varint,2,opt,name=sequence_number,json=sequenceNumber" json:"sequence_number,omitempty"` + SequenceNumber int64 `protobuf:"varint,2,opt,name=sequence_number,json=sequenceNumber,proto3" json:"sequence_number,omitempty"` // The actual event. - Event *BuildEvent `protobuf:"bytes,3,opt,name=event" json:"event,omitempty"` + Event *BuildEvent `protobuf:"bytes,3,opt,name=event,proto3" json:"event,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -257,21 +257,21 @@ func (m *OrderedBuildEvent) GetEvent() *BuildEvent { type PublishBuildToolEventStreamRequest struct { // Which build event stream this event belongs to. - StreamId *StreamId `protobuf:"bytes,1,opt,name=stream_id,json=streamId" json:"stream_id,omitempty"` // Deprecated: Do not use. + StreamId *StreamId `protobuf:"bytes,1,opt,name=stream_id,json=streamId,proto3" json:"stream_id,omitempty"` // Deprecated: Do not use. // The position of this event in the stream. The sequence numbers for a build // event stream should be a sequence of consecutive natural numbers starting // from one. (1, 2, 3, ...) - SequenceNumber int64 `protobuf:"varint,2,opt,name=sequence_number,json=sequenceNumber" json:"sequence_number,omitempty"` // Deprecated: Do not use. + SequenceNumber int64 `protobuf:"varint,2,opt,name=sequence_number,json=sequenceNumber,proto3" json:"sequence_number,omitempty"` // Deprecated: Do not use. // The actual event. - Event *BuildEvent `protobuf:"bytes,3,opt,name=event" json:"event,omitempty"` // Deprecated: Do not use. + Event *BuildEvent `protobuf:"bytes,3,opt,name=event,proto3" json:"event,omitempty"` // Deprecated: Do not use. // The build event with position info. // New publishing clients should use this field rather than the 3 above. - OrderedBuildEvent *OrderedBuildEvent `protobuf:"bytes,4,opt,name=ordered_build_event,json=orderedBuildEvent" json:"ordered_build_event,omitempty"` + OrderedBuildEvent *OrderedBuildEvent `protobuf:"bytes,4,opt,name=ordered_build_event,json=orderedBuildEvent,proto3" json:"ordered_build_event,omitempty"` // The keywords to be attached to the notification which notifies the start // of a new build event stream. BES only reads this field when sequence_number // or ordered_build_event.sequence_number is 1 in this message. If this field // is empty, BES will not publish notification messages for this stream. - NotificationKeywords []string `protobuf:"bytes,5,rep,name=notification_keywords,json=notificationKeywords" json:"notification_keywords,omitempty"` + NotificationKeywords []string `protobuf:"bytes,5,rep,name=notification_keywords,json=notificationKeywords,proto3" json:"notification_keywords,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -423,8 +423,7 @@ func (x *publishBuildEventPublishBuildToolEventStreamClient) Recv() (*PublishBui return m, nil } -// Server API for PublishBuildEvent service - +// PublishBuildEventServer is the server API for PublishBuildEvent service. type PublishBuildEventServer interface { // Publish a build event stating the new state of a build (typically from the // build queue). If the event is a BuildEnqueued event, also register the new diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/cloudbuild/v1/cloudbuild.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/cloudbuild/v1/cloudbuild.pb.go index b45de4620..2dbf74930 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/cloudbuild/v1/cloudbuild.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/cloudbuild/v1/cloudbuild.pb.go @@ -227,9 +227,9 @@ func (BuildOptions_LogStreamingOption) EnumDescriptor() ([]byte, []int) { // Specifies a build to retry. type RetryBuildRequest struct { // ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Build ID of the original build. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -276,11 +276,11 @@ func (m *RetryBuildRequest) GetId() string { // Specifies a build trigger to run and the source to use. type RunBuildTriggerRequest struct { // ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the trigger. - TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId" json:"trigger_id,omitempty"` + TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId,proto3" json:"trigger_id,omitempty"` // Source to build against this trigger. - Source *RepoSource `protobuf:"bytes,3,opt,name=source" json:"source,omitempty"` + Source *RepoSource `protobuf:"bytes,3,opt,name=source,proto3" json:"source,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -336,15 +336,15 @@ type StorageSource struct { // Google Cloud Storage bucket containing the source (see // [Bucket Name // Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). - Bucket string `protobuf:"bytes,1,opt,name=bucket" json:"bucket,omitempty"` + Bucket string `protobuf:"bytes,1,opt,name=bucket,proto3" json:"bucket,omitempty"` // Google Cloud Storage object containing the source. // // This object must be a gzipped archive file (`.tar.gz`) containing source to // build. - Object string `protobuf:"bytes,2,opt,name=object" json:"object,omitempty"` + Object string `protobuf:"bytes,2,opt,name=object,proto3" json:"object,omitempty"` // Google Cloud Storage generation for the object. If the generation is // omitted, the latest generation will be used. - Generation int64 `protobuf:"varint,3,opt,name=generation" json:"generation,omitempty"` + Generation int64 `protobuf:"varint,3,opt,name=generation,proto3" json:"generation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -399,10 +399,10 @@ func (m *StorageSource) GetGeneration() int64 { type RepoSource struct { // ID of the project that owns the Cloud Source Repository. If omitted, the // project ID requesting the build is assumed. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Name of the Cloud Source Repository. If omitted, the name "default" is // assumed. - RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName" json:"repo_name,omitempty"` + RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName,proto3" json:"repo_name,omitempty"` // A revision within the Cloud Source Repository must be specified in // one of these ways. // @@ -415,7 +415,7 @@ type RepoSource struct { // // This must be a relative path. If a step's `dir` is specified and is an // absolute path, this value is ignored for that step's execution. - Dir string `protobuf:"bytes,7,opt,name=dir" json:"dir,omitempty"` + Dir string `protobuf:"bytes,7,opt,name=dir,proto3" json:"dir,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -450,13 +450,13 @@ type isRepoSource_Revision interface { } type RepoSource_BranchName struct { - BranchName string `protobuf:"bytes,3,opt,name=branch_name,json=branchName,oneof"` + BranchName string `protobuf:"bytes,3,opt,name=branch_name,json=branchName,proto3,oneof"` } type RepoSource_TagName struct { - TagName string `protobuf:"bytes,4,opt,name=tag_name,json=tagName,oneof"` + TagName string `protobuf:"bytes,4,opt,name=tag_name,json=tagName,proto3,oneof"` } type RepoSource_CommitSha struct { - CommitSha string `protobuf:"bytes,5,opt,name=commit_sha,json=commitSha,oneof"` + CommitSha string `protobuf:"bytes,5,opt,name=commit_sha,json=commitSha,proto3,oneof"` } func (*RepoSource_BranchName) isRepoSource_Revision() {} @@ -635,10 +635,10 @@ type isSource_Source interface { } type Source_StorageSource struct { - StorageSource *StorageSource `protobuf:"bytes,2,opt,name=storage_source,json=storageSource,oneof"` + StorageSource *StorageSource `protobuf:"bytes,2,opt,name=storage_source,json=storageSource,proto3,oneof"` } type Source_RepoSource struct { - RepoSource *RepoSource `protobuf:"bytes,3,opt,name=repo_source,json=repoSource,oneof"` + RepoSource *RepoSource `protobuf:"bytes,3,opt,name=repo_source,json=repoSource,proto3,oneof"` } func (*Source_StorageSource) isSource_Source() {} @@ -743,12 +743,12 @@ func _Source_OneofSizer(msg proto.Message) (n int) { type BuiltImage struct { // Name used to push the container image to Google Container Registry, as // presented to `docker push`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Docker Registry 2.0 digest. - Digest string `protobuf:"bytes,3,opt,name=digest" json:"digest,omitempty"` + Digest string `protobuf:"bytes,3,opt,name=digest,proto3" json:"digest,omitempty"` // Stores timing information for pushing the specified image. // @OutputOnly - PushTiming *TimeSpan `protobuf:"bytes,4,opt,name=push_timing,json=pushTiming" json:"push_timing,omitempty"` + PushTiming *TimeSpan `protobuf:"bytes,4,opt,name=push_timing,json=pushTiming,proto3" json:"push_timing,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -818,19 +818,19 @@ type BuildStep struct { // If you built an image in a previous build step, it will be stored in the // host's Docker daemon's cache and is available to use as the name for a // later build step. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A list of environment variable definitions to be used when running a step. // // The elements are of the form "KEY=VALUE" for the environment variable "KEY" // being given the value "VALUE". - Env []string `protobuf:"bytes,2,rep,name=env" json:"env,omitempty"` + Env []string `protobuf:"bytes,2,rep,name=env,proto3" json:"env,omitempty"` // A list of arguments that will be presented to the step when it is started. // // If the image used to run the step's container has an entrypoint, the `args` // are used as arguments to that entrypoint. If the image does not define // an entrypoint, the first element in args is used as the entrypoint, // and the remainder will be used as arguments. - Args []string `protobuf:"bytes,3,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` // Working directory to use when running this step's container. // // If this value is a relative path, it is relative to the build's working @@ -841,23 +841,23 @@ type BuildStep struct { // If the build specifies a `RepoSource` with `dir` and a step with a `dir`, // which specifies an absolute path, the `RepoSource` `dir` is ignored for // the step's execution. - Dir string `protobuf:"bytes,4,opt,name=dir" json:"dir,omitempty"` + Dir string `protobuf:"bytes,4,opt,name=dir,proto3" json:"dir,omitempty"` // Unique identifier for this build step, used in `wait_for` to // reference this build step as a dependency. - Id string `protobuf:"bytes,5,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,5,opt,name=id,proto3" json:"id,omitempty"` // The ID(s) of the step(s) that this build step depends on. // This build step will not start until all the build steps in `wait_for` // have completed successfully. If `wait_for` is empty, this build step will // start when all previous build steps in the `Build.Steps` list have // completed successfully. - WaitFor []string `protobuf:"bytes,6,rep,name=wait_for,json=waitFor" json:"wait_for,omitempty"` + WaitFor []string `protobuf:"bytes,6,rep,name=wait_for,json=waitFor,proto3" json:"wait_for,omitempty"` // Entrypoint to be used instead of the build step image's default entrypoint. // If unset, the image's default entrypoint is used. - Entrypoint string `protobuf:"bytes,7,opt,name=entrypoint" json:"entrypoint,omitempty"` + Entrypoint string `protobuf:"bytes,7,opt,name=entrypoint,proto3" json:"entrypoint,omitempty"` // A list of environment variables which are encrypted using a Cloud Key // Management Service crypto key. These values must be specified in the // build's `Secret`. - SecretEnv []string `protobuf:"bytes,8,rep,name=secret_env,json=secretEnv" json:"secret_env,omitempty"` + SecretEnv []string `protobuf:"bytes,8,rep,name=secret_env,json=secretEnv,proto3" json:"secret_env,omitempty"` // List of volumes to mount into the build step. // // Each volume will be created as an empty volume prior to execution of the @@ -866,19 +866,19 @@ type BuildStep struct { // // Using a named volume in only one step is not valid as it is indicative // of a mis-configured build request. - Volumes []*Volume `protobuf:"bytes,9,rep,name=volumes" json:"volumes,omitempty"` + Volumes []*Volume `protobuf:"bytes,9,rep,name=volumes,proto3" json:"volumes,omitempty"` // Stores timing information for executing this build step. // @OutputOnly - Timing *TimeSpan `protobuf:"bytes,10,opt,name=timing" json:"timing,omitempty"` + Timing *TimeSpan `protobuf:"bytes,10,opt,name=timing,proto3" json:"timing,omitempty"` // Time limit for executing this build step. If not defined, the step has no // time limit and will be allowed to continue to run until either it completes // or the build itself times out. - Timeout *duration.Duration `protobuf:"bytes,11,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,11,opt,name=timeout,proto3" json:"timeout,omitempty"` // Status of the build step. At this time, build step status is only updated // on build completion; step status is not updated in real-time as the build // progresses. // @OutputOnly - Status Build_Status `protobuf:"varint,12,opt,name=status,enum=google.devtools.cloudbuild.v1.Build_Status" json:"status,omitempty"` + Status Build_Status `protobuf:"varint,12,opt,name=status,proto3,enum=google.devtools.cloudbuild.v1.Build_Status" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -999,12 +999,12 @@ type Volume struct { // // Volume names must be unique per build step and must be valid names for // Docker volumes. Each named volume must be used by at least two build steps. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Path at which to mount the volume. // // Paths must be absolute and cannot conflict with other volume paths on the // same build step or with certain reserved volume paths. - Path string `protobuf:"bytes,2,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,2,opt,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1051,14 +1051,14 @@ func (m *Volume) GetPath() string { // Artifacts created by the build pipeline. type Results struct { // Container images that were built as a part of the build. - Images []*BuiltImage `protobuf:"bytes,2,rep,name=images" json:"images,omitempty"` + Images []*BuiltImage `protobuf:"bytes,2,rep,name=images,proto3" json:"images,omitempty"` // List of build step digests, in the order corresponding to build step // indices. - BuildStepImages []string `protobuf:"bytes,3,rep,name=build_step_images,json=buildStepImages" json:"build_step_images,omitempty"` + BuildStepImages []string `protobuf:"bytes,3,rep,name=build_step_images,json=buildStepImages,proto3" json:"build_step_images,omitempty"` // Path to the artifact manifest. Only populated when artifacts are uploaded. - ArtifactManifest string `protobuf:"bytes,4,opt,name=artifact_manifest,json=artifactManifest" json:"artifact_manifest,omitempty"` + ArtifactManifest string `protobuf:"bytes,4,opt,name=artifact_manifest,json=artifactManifest,proto3" json:"artifact_manifest,omitempty"` // Number of artifacts uploaded. Only populated when artifacts are uploaded. - NumArtifacts int64 `protobuf:"varint,5,opt,name=num_artifacts,json=numArtifacts" json:"num_artifacts,omitempty"` + NumArtifacts int64 `protobuf:"varint,5,opt,name=num_artifacts,json=numArtifacts,proto3" json:"num_artifacts,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1122,9 +1122,9 @@ type ArtifactResult struct { // The path of an artifact in a Google Cloud Storage bucket, with the // generation number. For example, // `gs://mybucket/path/to/output.jar#generation`. - Location string `protobuf:"bytes,1,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,1,opt,name=location,proto3" json:"location,omitempty"` // The file hash of the artifact. - FileHash []*FileHashes `protobuf:"bytes,2,rep,name=file_hash,json=fileHash" json:"file_hash,omitempty"` + FileHash []*FileHashes `protobuf:"bytes,2,rep,name=file_hash,json=fileHash,proto3" json:"file_hash,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1188,41 +1188,41 @@ func (m *ArtifactResult) GetFileHash() []*FileHashes { type Build struct { // Unique identifier of the build. // @OutputOnly - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // ID of the project. // @OutputOnly. - ProjectId string `protobuf:"bytes,16,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,16,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Status of the build. // @OutputOnly - Status Build_Status `protobuf:"varint,2,opt,name=status,enum=google.devtools.cloudbuild.v1.Build_Status" json:"status,omitempty"` + Status Build_Status `protobuf:"varint,2,opt,name=status,proto3,enum=google.devtools.cloudbuild.v1.Build_Status" json:"status,omitempty"` // Customer-readable message about the current status. // @OutputOnly - StatusDetail string `protobuf:"bytes,24,opt,name=status_detail,json=statusDetail" json:"status_detail,omitempty"` + StatusDetail string `protobuf:"bytes,24,opt,name=status_detail,json=statusDetail,proto3" json:"status_detail,omitempty"` // The location of the source files to build. - Source *Source `protobuf:"bytes,3,opt,name=source" json:"source,omitempty"` + Source *Source `protobuf:"bytes,3,opt,name=source,proto3" json:"source,omitempty"` // Required. The operations to be performed on the workspace. - Steps []*BuildStep `protobuf:"bytes,11,rep,name=steps" json:"steps,omitempty"` + Steps []*BuildStep `protobuf:"bytes,11,rep,name=steps,proto3" json:"steps,omitempty"` // Results of the build. // @OutputOnly - Results *Results `protobuf:"bytes,10,opt,name=results" json:"results,omitempty"` + Results *Results `protobuf:"bytes,10,opt,name=results,proto3" json:"results,omitempty"` // Time at which the request to create the build was received. // @OutputOnly - CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time at which execution of the build was started. // @OutputOnly - StartTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time at which execution of the build was finished. // // The difference between finish_time and start_time is the duration of the // build's execution. // @OutputOnly - FinishTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` // Amount of time that this build should be allowed to run, to second // granularity. If this amount of time elapses, work on the build will cease // and the build status will be `TIMEOUT`. // // Default time is ten minutes. - Timeout *duration.Duration `protobuf:"bytes,12,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,12,opt,name=timeout,proto3" json:"timeout,omitempty"` // A list of images to be pushed upon the successful completion of all build // steps. // @@ -1233,33 +1233,33 @@ type Build struct { // // If any of the images fail to be pushed, the build status is marked // `FAILURE`. - Images []string `protobuf:"bytes,13,rep,name=images" json:"images,omitempty"` + Images []string `protobuf:"bytes,13,rep,name=images,proto3" json:"images,omitempty"` // Artifacts produced by the build that should be uploaded upon // successful completion of all build steps. - Artifacts *Artifacts `protobuf:"bytes,37,opt,name=artifacts" json:"artifacts,omitempty"` + Artifacts *Artifacts `protobuf:"bytes,37,opt,name=artifacts,proto3" json:"artifacts,omitempty"` // Google Cloud Storage bucket where logs should be written (see // [Bucket Name // Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). // Logs file names will be of the format `${logs_bucket}/log-${build_id}.txt`. - LogsBucket string `protobuf:"bytes,19,opt,name=logs_bucket,json=logsBucket" json:"logs_bucket,omitempty"` + LogsBucket string `protobuf:"bytes,19,opt,name=logs_bucket,json=logsBucket,proto3" json:"logs_bucket,omitempty"` // A permanent fixed identifier for source. // @OutputOnly - SourceProvenance *SourceProvenance `protobuf:"bytes,21,opt,name=source_provenance,json=sourceProvenance" json:"source_provenance,omitempty"` + SourceProvenance *SourceProvenance `protobuf:"bytes,21,opt,name=source_provenance,json=sourceProvenance,proto3" json:"source_provenance,omitempty"` // The ID of the `BuildTrigger` that triggered this build, if it was // triggered automatically. // @OutputOnly - BuildTriggerId string `protobuf:"bytes,22,opt,name=build_trigger_id,json=buildTriggerId" json:"build_trigger_id,omitempty"` + BuildTriggerId string `protobuf:"bytes,22,opt,name=build_trigger_id,json=buildTriggerId,proto3" json:"build_trigger_id,omitempty"` // Special options for this build. - Options *BuildOptions `protobuf:"bytes,23,opt,name=options" json:"options,omitempty"` + Options *BuildOptions `protobuf:"bytes,23,opt,name=options,proto3" json:"options,omitempty"` // URL to logs for this build in Google Cloud Console. // @OutputOnly - LogUrl string `protobuf:"bytes,25,opt,name=log_url,json=logUrl" json:"log_url,omitempty"` + LogUrl string `protobuf:"bytes,25,opt,name=log_url,json=logUrl,proto3" json:"log_url,omitempty"` // Substitutions data for `Build` resource. - Substitutions map[string]string `protobuf:"bytes,29,rep,name=substitutions" json:"substitutions,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Substitutions map[string]string `protobuf:"bytes,29,rep,name=substitutions,proto3" json:"substitutions,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Tags for annotation of a `Build`. These are not docker tags. - Tags []string `protobuf:"bytes,31,rep,name=tags" json:"tags,omitempty"` + Tags []string `protobuf:"bytes,31,rep,name=tags,proto3" json:"tags,omitempty"` // Secrets to decrypt using Cloud Key Management Service. - Secrets []*Secret `protobuf:"bytes,32,rep,name=secrets" json:"secrets,omitempty"` + Secrets []*Secret `protobuf:"bytes,32,rep,name=secrets,proto3" json:"secrets,omitempty"` // Stores timing information for phases of the build. Valid keys are: // // * BUILD: time to execute all build steps @@ -1269,7 +1269,7 @@ type Build struct { // If the build does not specify source or images, // these keys will not be included. // @OutputOnly - Timing map[string]*TimeSpan `protobuf:"bytes,33,rep,name=timing" json:"timing,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Timing map[string]*TimeSpan `protobuf:"bytes,33,rep,name=timing,proto3" json:"timing,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1465,7 +1465,7 @@ type Artifacts struct { // results field. // // If any of the images fail to be pushed, the build is marked FAILURE. - Images []string `protobuf:"bytes,1,rep,name=images" json:"images,omitempty"` + Images []string `protobuf:"bytes,1,rep,name=images,proto3" json:"images,omitempty"` // A list of objects to be uploaded to Cloud Storage upon successful // completion of all build steps. // @@ -1477,7 +1477,7 @@ type Artifacts struct { // Build resource's results field. // // If any objects fail to be pushed, the build is marked FAILURE. - Objects *Artifacts_ArtifactObjects `protobuf:"bytes,2,opt,name=objects" json:"objects,omitempty"` + Objects *Artifacts_ArtifactObjects `protobuf:"bytes,2,opt,name=objects,proto3" json:"objects,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1530,12 +1530,12 @@ type Artifacts_ArtifactObjects struct { // // Files in the workspace matching any path pattern will be uploaded to // Cloud Storage with this location as a prefix. - Location string `protobuf:"bytes,1,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,1,opt,name=location,proto3" json:"location,omitempty"` // Path globs used to match files in the build's workspace. - Paths []string `protobuf:"bytes,2,rep,name=paths" json:"paths,omitempty"` + Paths []string `protobuf:"bytes,2,rep,name=paths,proto3" json:"paths,omitempty"` // Stores timing information for pushing all artifact objects. // @OutputOnly - Timing *TimeSpan `protobuf:"bytes,3,opt,name=timing" json:"timing,omitempty"` + Timing *TimeSpan `protobuf:"bytes,3,opt,name=timing,proto3" json:"timing,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1589,9 +1589,9 @@ func (m *Artifacts_ArtifactObjects) GetTiming() *TimeSpan { // Start and end times for a build execution phase. type TimeSpan struct { // Start of time span. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End of time span. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1638,7 +1638,7 @@ func (m *TimeSpan) GetEndTime() *timestamp.Timestamp { // Metadata for build operations. type BuildOperationMetadata struct { // The build that the operation is tracking. - Build *Build `protobuf:"bytes,1,opt,name=build" json:"build,omitempty"` + Build *Build `protobuf:"bytes,1,opt,name=build,proto3" json:"build,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1680,10 +1680,10 @@ func (m *BuildOperationMetadata) GetBuild() *Build { type SourceProvenance struct { // A copy of the build's `source.storage_source`, if exists, with any // generations resolved. - ResolvedStorageSource *StorageSource `protobuf:"bytes,3,opt,name=resolved_storage_source,json=resolvedStorageSource" json:"resolved_storage_source,omitempty"` + ResolvedStorageSource *StorageSource `protobuf:"bytes,3,opt,name=resolved_storage_source,json=resolvedStorageSource,proto3" json:"resolved_storage_source,omitempty"` // A copy of the build's `source.repo_source`, if exists, with any // revisions resolved. - ResolvedRepoSource *RepoSource `protobuf:"bytes,6,opt,name=resolved_repo_source,json=resolvedRepoSource" json:"resolved_repo_source,omitempty"` + ResolvedRepoSource *RepoSource `protobuf:"bytes,6,opt,name=resolved_repo_source,json=resolvedRepoSource,proto3" json:"resolved_repo_source,omitempty"` // Hash(es) of the build source, which can be used to verify that the original // source integrity was maintained in the build. Note that `FileHashes` will // only be populated if `BuildOptions` has requested a `SourceProvenanceHash`. @@ -1694,7 +1694,7 @@ type SourceProvenance struct { // If the build source came in a single package such as a gzipped tarfile // (`.tar.gz`), the `FileHash` will be for the single path to that file. // @OutputOnly - FileHashes map[string]*FileHashes `protobuf:"bytes,4,rep,name=file_hashes,json=fileHashes" json:"file_hashes,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + FileHashes map[string]*FileHashes `protobuf:"bytes,4,rep,name=file_hashes,json=fileHashes,proto3" json:"file_hashes,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1749,7 +1749,7 @@ func (m *SourceProvenance) GetFileHashes() map[string]*FileHashes { // SourceProvenance messages to verify integrity of source input to the build. type FileHashes struct { // Collection of file hashes. - FileHash []*Hash `protobuf:"bytes,1,rep,name=file_hash,json=fileHash" json:"file_hash,omitempty"` + FileHash []*Hash `protobuf:"bytes,1,rep,name=file_hash,json=fileHash,proto3" json:"file_hash,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1789,7 +1789,7 @@ func (m *FileHashes) GetFileHash() []*Hash { // Container message for hash values. type Hash struct { // The type of hash that was performed. - Type Hash_HashType `protobuf:"varint,1,opt,name=type,enum=google.devtools.cloudbuild.v1.Hash_HashType" json:"type,omitempty"` + Type Hash_HashType `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.cloudbuild.v1.Hash_HashType" json:"type,omitempty"` // The hash value. Value []byte `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1839,14 +1839,14 @@ func (m *Hash) GetValue() []byte { // values with the Cloud KMS key to use to decrypt the value. type Secret struct { // Cloud KMS key name to use to decrypt these envs. - KmsKeyName string `protobuf:"bytes,1,opt,name=kms_key_name,json=kmsKeyName" json:"kms_key_name,omitempty"` + KmsKeyName string `protobuf:"bytes,1,opt,name=kms_key_name,json=kmsKeyName,proto3" json:"kms_key_name,omitempty"` // Map of environment variable name to its encrypted value. // // Secret environment variables must be unique across all of a build's // secrets, and must be used by at least one build step. Values can be at most // 1 KB in size. There can be at most ten secret values across all of a // build's secrets. - SecretEnv map[string][]byte `protobuf:"bytes,3,rep,name=secret_env,json=secretEnv" json:"secret_env,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value,proto3"` + SecretEnv map[string][]byte `protobuf:"bytes,3,rep,name=secret_env,json=secretEnv,proto3" json:"secret_env,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1893,9 +1893,9 @@ func (m *Secret) GetSecretEnv() map[string][]byte { // Request to create a new build. type CreateBuildRequest struct { // ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Build resource to create. - Build *Build `protobuf:"bytes,2,opt,name=build" json:"build,omitempty"` + Build *Build `protobuf:"bytes,2,opt,name=build,proto3" json:"build,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1942,9 +1942,9 @@ func (m *CreateBuildRequest) GetBuild() *Build { // Request to get a build. type GetBuildRequest struct { // ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the build. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1991,13 +1991,13 @@ func (m *GetBuildRequest) GetId() string { // Request to list builds. type ListBuildsRequest struct { // ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Number of results to return in the list. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token to provide to skip to a particular spot in the list. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The raw filter text to constrain the results. - Filter string `protobuf:"bytes,8,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,8,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2058,9 +2058,9 @@ func (m *ListBuildsRequest) GetFilter() string { // Response including listed builds. type ListBuildsResponse struct { // Builds will be sorted by `create_time`, descending. - Builds []*Build `protobuf:"bytes,1,rep,name=builds" json:"builds,omitempty"` + Builds []*Build `protobuf:"bytes,1,rep,name=builds,proto3" json:"builds,omitempty"` // Token to receive the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2107,9 +2107,9 @@ func (m *ListBuildsResponse) GetNextPageToken() string { // Request to cancel an ongoing build. type CancelBuildRequest struct { // ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the build. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2159,15 +2159,15 @@ type BuildTrigger struct { // Unique identifier of the trigger. // // @OutputOnly - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // Human-readable description of this trigger. - Description string `protobuf:"bytes,10,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,10,opt,name=description,proto3" json:"description,omitempty"` // Template describing the types of source changes to trigger a build. // // Branch and tag names in trigger templates are interpreted as regular // expressions. Any branch or tag change that matches that regular expression // will trigger a build. - TriggerTemplate *RepoSource `protobuf:"bytes,7,opt,name=trigger_template,json=triggerTemplate" json:"trigger_template,omitempty"` + TriggerTemplate *RepoSource `protobuf:"bytes,7,opt,name=trigger_template,json=triggerTemplate,proto3" json:"trigger_template,omitempty"` // Template describing the Build request to make when the trigger is matched. // // Types that are valid to be assigned to BuildTemplate: @@ -2177,11 +2177,11 @@ type BuildTrigger struct { // Time when the trigger was created. // // @OutputOnly - CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // If true, the trigger will never result in a build. - Disabled bool `protobuf:"varint,9,opt,name=disabled" json:"disabled,omitempty"` + Disabled bool `protobuf:"varint,9,opt,name=disabled,proto3" json:"disabled,omitempty"` // Substitutions data for Build resource. - Substitutions map[string]string `protobuf:"bytes,11,rep,name=substitutions" json:"substitutions,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Substitutions map[string]string `protobuf:"bytes,11,rep,name=substitutions,proto3" json:"substitutions,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2216,10 +2216,10 @@ type isBuildTrigger_BuildTemplate interface { } type BuildTrigger_Build struct { - Build *Build `protobuf:"bytes,4,opt,name=build,oneof"` + Build *Build `protobuf:"bytes,4,opt,name=build,proto3,oneof"` } type BuildTrigger_Filename struct { - Filename string `protobuf:"bytes,8,opt,name=filename,oneof"` + Filename string `protobuf:"bytes,8,opt,name=filename,proto3,oneof"` } func (*BuildTrigger_Build) isBuildTrigger_BuildTemplate() {} @@ -2361,9 +2361,9 @@ func _BuildTrigger_OneofSizer(msg proto.Message) (n int) { // Request to create a new `BuildTrigger`. type CreateBuildTriggerRequest struct { // ID of the project for which to configure automatic builds. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // `BuildTrigger` to create. - Trigger *BuildTrigger `protobuf:"bytes,2,opt,name=trigger" json:"trigger,omitempty"` + Trigger *BuildTrigger `protobuf:"bytes,2,opt,name=trigger,proto3" json:"trigger,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2410,9 +2410,9 @@ func (m *CreateBuildTriggerRequest) GetTrigger() *BuildTrigger { // Returns the `BuildTrigger` with the specified ID. type GetBuildTriggerRequest struct { // ID of the project that owns the trigger. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the `BuildTrigger` to get. - TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId" json:"trigger_id,omitempty"` + TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId,proto3" json:"trigger_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2459,7 +2459,7 @@ func (m *GetBuildTriggerRequest) GetTriggerId() string { // Request to list existing `BuildTriggers`. type ListBuildTriggersRequest struct { // ID of the project for which to list BuildTriggers. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2499,7 +2499,7 @@ func (m *ListBuildTriggersRequest) GetProjectId() string { // Response containing existing `BuildTriggers`. type ListBuildTriggersResponse struct { // `BuildTriggers` for the project, sorted by `create_time` descending. - Triggers []*BuildTrigger `protobuf:"bytes,1,rep,name=triggers" json:"triggers,omitempty"` + Triggers []*BuildTrigger `protobuf:"bytes,1,rep,name=triggers,proto3" json:"triggers,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2539,9 +2539,9 @@ func (m *ListBuildTriggersResponse) GetTriggers() []*BuildTrigger { // Request to delete a `BuildTrigger`. type DeleteBuildTriggerRequest struct { // ID of the project that owns the trigger. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the `BuildTrigger` to delete. - TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId" json:"trigger_id,omitempty"` + TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId,proto3" json:"trigger_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2588,11 +2588,11 @@ func (m *DeleteBuildTriggerRequest) GetTriggerId() string { // Request to update an existing `BuildTrigger`. type UpdateBuildTriggerRequest struct { // ID of the project that owns the trigger. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the `BuildTrigger` to update. - TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId" json:"trigger_id,omitempty"` + TriggerId string `protobuf:"bytes,2,opt,name=trigger_id,json=triggerId,proto3" json:"trigger_id,omitempty"` // `BuildTrigger` to update. - Trigger *BuildTrigger `protobuf:"bytes,3,opt,name=trigger" json:"trigger,omitempty"` + Trigger *BuildTrigger `protobuf:"bytes,3,opt,name=trigger,proto3" json:"trigger,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2646,24 +2646,24 @@ func (m *UpdateBuildTriggerRequest) GetTrigger() *BuildTrigger { // Optional arguments to enable specific features of builds. type BuildOptions struct { // Requested hash for SourceProvenance. - SourceProvenanceHash []Hash_HashType `protobuf:"varint,1,rep,packed,name=source_provenance_hash,json=sourceProvenanceHash,enum=google.devtools.cloudbuild.v1.Hash_HashType" json:"source_provenance_hash,omitempty"` + SourceProvenanceHash []Hash_HashType `protobuf:"varint,1,rep,packed,name=source_provenance_hash,json=sourceProvenanceHash,proto3,enum=google.devtools.cloudbuild.v1.Hash_HashType" json:"source_provenance_hash,omitempty"` // Requested verifiability options. - RequestedVerifyOption BuildOptions_VerifyOption `protobuf:"varint,2,opt,name=requested_verify_option,json=requestedVerifyOption,enum=google.devtools.cloudbuild.v1.BuildOptions_VerifyOption" json:"requested_verify_option,omitempty"` + RequestedVerifyOption BuildOptions_VerifyOption `protobuf:"varint,2,opt,name=requested_verify_option,json=requestedVerifyOption,proto3,enum=google.devtools.cloudbuild.v1.BuildOptions_VerifyOption" json:"requested_verify_option,omitempty"` // Compute Engine machine type on which to run the build. - MachineType BuildOptions_MachineType `protobuf:"varint,3,opt,name=machine_type,json=machineType,enum=google.devtools.cloudbuild.v1.BuildOptions_MachineType" json:"machine_type,omitempty"` + MachineType BuildOptions_MachineType `protobuf:"varint,3,opt,name=machine_type,json=machineType,proto3,enum=google.devtools.cloudbuild.v1.BuildOptions_MachineType" json:"machine_type,omitempty"` // Requested disk size for the VM that runs the build. Note that this is *NOT* // "disk free"; some of the space will be used by the operating system and // build utilities. Also note that this is the minimum disk size that will be // allocated for the build -- the build may run with a larger disk than // requested. At present, the maximum disk size is 1000GB; builds that request // more than the maximum are rejected with an error. - DiskSizeGb int64 `protobuf:"varint,6,opt,name=disk_size_gb,json=diskSizeGb" json:"disk_size_gb,omitempty"` + DiskSizeGb int64 `protobuf:"varint,6,opt,name=disk_size_gb,json=diskSizeGb,proto3" json:"disk_size_gb,omitempty"` // Option to specify behavior when there is an error in the substitution // checks. - SubstitutionOption BuildOptions_SubstitutionOption `protobuf:"varint,4,opt,name=substitution_option,json=substitutionOption,enum=google.devtools.cloudbuild.v1.BuildOptions_SubstitutionOption" json:"substitution_option,omitempty"` + SubstitutionOption BuildOptions_SubstitutionOption `protobuf:"varint,4,opt,name=substitution_option,json=substitutionOption,proto3,enum=google.devtools.cloudbuild.v1.BuildOptions_SubstitutionOption" json:"substitution_option,omitempty"` // Option to define build log streaming behavior to Google Cloud // Storage. - LogStreamingOption BuildOptions_LogStreamingOption `protobuf:"varint,5,opt,name=log_streaming_option,json=logStreamingOption,enum=google.devtools.cloudbuild.v1.BuildOptions_LogStreamingOption" json:"log_streaming_option,omitempty"` + LogStreamingOption BuildOptions_LogStreamingOption `protobuf:"varint,5,opt,name=log_streaming_option,json=logStreamingOption,proto3,enum=google.devtools.cloudbuild.v1.BuildOptions_LogStreamingOption" json:"log_streaming_option,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2970,8 +2970,7 @@ func (c *cloudBuildClient) RunBuildTrigger(ctx context.Context, in *RunBuildTrig return out, nil } -// Server API for CloudBuild service - +// CloudBuildServer is the server API for CloudBuild service. type CloudBuildServer interface { // Starts a build with the specified configuration. // diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/controller.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/controller.pb.go index 4eb87ca71..d5c68b115 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/controller.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/controller.pb.go @@ -30,7 +30,7 @@ type RegisterDebuggeeRequest struct { // Debuggee information to register. // The fields `project`, `uniquifier`, `description` and `agent_version` // of the debuggee must be set. - Debuggee *Debuggee `protobuf:"bytes,1,opt,name=debuggee" json:"debuggee,omitempty"` + Debuggee *Debuggee `protobuf:"bytes,1,opt,name=debuggee,proto3" json:"debuggee,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -74,7 +74,7 @@ type RegisterDebuggeeResponse struct { // If the field `is_disabled` is set to `true`, the agent should disable // itself by removing all breakpoints and detaching from the application. // It should however continue to poll `RegisterDebuggee` until reenabled. - Debuggee *Debuggee `protobuf:"bytes,1,opt,name=debuggee" json:"debuggee,omitempty"` + Debuggee *Debuggee `protobuf:"bytes,1,opt,name=debuggee,proto3" json:"debuggee,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -114,19 +114,19 @@ func (m *RegisterDebuggeeResponse) GetDebuggee() *Debuggee { // Request to list active breakpoints. type ListActiveBreakpointsRequest struct { // Identifies the debuggee. - DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId" json:"debuggee_id,omitempty"` + DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId,proto3" json:"debuggee_id,omitempty"` // A token that, if specified, blocks the method call until the list // of active breakpoints has changed, or a server-selected timeout has // expired. The value should be set from the `next_wait_token` field in // the last response. The initial value should be set to `"init"`. - WaitToken string `protobuf:"bytes,2,opt,name=wait_token,json=waitToken" json:"wait_token,omitempty"` + WaitToken string `protobuf:"bytes,2,opt,name=wait_token,json=waitToken,proto3" json:"wait_token,omitempty"` // If set to `true` (recommended), returns `google.rpc.Code.OK` status and // sets the `wait_expired` response field to `true` when the server-selected // timeout has expired. // // If set to `false` (deprecated), returns `google.rpc.Code.ABORTED` status // when the server-selected timeout has expired. - SuccessOnTimeout bool `protobuf:"varint,3,opt,name=success_on_timeout,json=successOnTimeout" json:"success_on_timeout,omitempty"` + SuccessOnTimeout bool `protobuf:"varint,3,opt,name=success_on_timeout,json=successOnTimeout,proto3" json:"success_on_timeout,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -181,14 +181,14 @@ func (m *ListActiveBreakpointsRequest) GetSuccessOnTimeout() bool { type ListActiveBreakpointsResponse struct { // List of all active breakpoints. // The fields `id` and `location` are guaranteed to be set on each breakpoint. - Breakpoints []*Breakpoint `protobuf:"bytes,1,rep,name=breakpoints" json:"breakpoints,omitempty"` + Breakpoints []*Breakpoint `protobuf:"bytes,1,rep,name=breakpoints,proto3" json:"breakpoints,omitempty"` // A token that can be used in the next method call to block until // the list of breakpoints changes. - NextWaitToken string `protobuf:"bytes,2,opt,name=next_wait_token,json=nextWaitToken" json:"next_wait_token,omitempty"` + NextWaitToken string `protobuf:"bytes,2,opt,name=next_wait_token,json=nextWaitToken,proto3" json:"next_wait_token,omitempty"` // If set to `true`, indicates that there is no change to the // list of active breakpoints and the server-selected timeout has expired. // The `breakpoints` field would be empty and should be ignored. - WaitExpired bool `protobuf:"varint,3,opt,name=wait_expired,json=waitExpired" json:"wait_expired,omitempty"` + WaitExpired bool `protobuf:"varint,3,opt,name=wait_expired,json=waitExpired,proto3" json:"wait_expired,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -242,11 +242,11 @@ func (m *ListActiveBreakpointsResponse) GetWaitExpired() bool { // Request to update an active breakpoint. type UpdateActiveBreakpointRequest struct { // Identifies the debuggee being debugged. - DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId" json:"debuggee_id,omitempty"` + DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId,proto3" json:"debuggee_id,omitempty"` // Updated breakpoint information. // The field `id` must be set. // The agent must echo all Breakpoint specification fields in the update. - Breakpoint *Breakpoint `protobuf:"bytes,2,opt,name=breakpoint" json:"breakpoint,omitempty"` + Breakpoint *Breakpoint `protobuf:"bytes,2,opt,name=breakpoint,proto3" json:"breakpoint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -414,8 +414,7 @@ func (c *controller2Client) UpdateActiveBreakpoint(ctx context.Context, in *Upda return out, nil } -// Server API for Controller2 service - +// Controller2Server is the server API for Controller2 service. type Controller2Server interface { // Registers the debuggee with the controller service. // diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/data.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/data.pb.go index 42a2ccfbe..d2901244c 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/data.pb.go @@ -139,9 +139,9 @@ type FormatMessage struct { // * `Failed to load '$0' which helps debug $1 the first time it // is loaded. Again, $0 is very important.` // * `Please pay $$10 to use $0 instead of $1.` - Format string `protobuf:"bytes,1,opt,name=format" json:"format,omitempty"` + Format string `protobuf:"bytes,1,opt,name=format,proto3" json:"format,omitempty"` // Optional parameters to be embedded into the message. - Parameters []string `protobuf:"bytes,2,rep,name=parameters" json:"parameters,omitempty"` + Parameters []string `protobuf:"bytes,2,rep,name=parameters,proto3" json:"parameters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -192,11 +192,11 @@ func (m *FormatMessage) GetParameters() []string { // to the `BREAKPOINT_SOURCE_LOCATION` with the message `Location not found`. type StatusMessage struct { // Distinguishes errors from informational messages. - IsError bool `protobuf:"varint,1,opt,name=is_error,json=isError" json:"is_error,omitempty"` + IsError bool `protobuf:"varint,1,opt,name=is_error,json=isError,proto3" json:"is_error,omitempty"` // Reference to which the message applies. - RefersTo StatusMessage_Reference `protobuf:"varint,2,opt,name=refers_to,json=refersTo,enum=google.devtools.clouddebugger.v2.StatusMessage_Reference" json:"refers_to,omitempty"` + RefersTo StatusMessage_Reference `protobuf:"varint,2,opt,name=refers_to,json=refersTo,proto3,enum=google.devtools.clouddebugger.v2.StatusMessage_Reference" json:"refers_to,omitempty"` // Status message text. - Description *FormatMessage `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description *FormatMessage `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -250,9 +250,9 @@ func (m *StatusMessage) GetDescription() *FormatMessage { // Represents a location in the source code. type SourceLocation struct { // Path to the source file within the source context of the target binary. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // Line inside the file. The first line in the file has the value `1`. - Line int32 `protobuf:"varint,2,opt,name=line" json:"line,omitempty"` + Line int32 `protobuf:"varint,2,opt,name=line,proto3" json:"line,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -407,20 +407,20 @@ func (m *SourceLocation) GetLine() int32 { // The type field is optional. The debugger agent may or may not support it. type Variable struct { // Name of the variable, if any. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Simple value of the variable. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Variable type (e.g. `MyClass`). If the variable is split with // `var_table_index`, `type` goes next to `value`. The interpretation of // a type is agent specific. It is recommended to include the dynamic type // rather than a static type of an object. - Type string `protobuf:"bytes,6,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,6,opt,name=type,proto3" json:"type,omitempty"` // Members contained or pointed to by the variable. - Members []*Variable `protobuf:"bytes,3,rep,name=members" json:"members,omitempty"` + Members []*Variable `protobuf:"bytes,3,rep,name=members,proto3" json:"members,omitempty"` // Reference to a variable in the shared variable table. More than // one variable can reference the same variable in the table. The // `var_table_index` field is an index into `variable_table` in Breakpoint. - VarTableIndex *wrappers.Int32Value `protobuf:"bytes,4,opt,name=var_table_index,json=varTableIndex" json:"var_table_index,omitempty"` + VarTableIndex *wrappers.Int32Value `protobuf:"bytes,4,opt,name=var_table_index,json=varTableIndex,proto3" json:"var_table_index,omitempty"` // Status associated with the variable. This field will usually stay // unset. A status of a single variable only applies to that variable or // expression. The rest of breakpoint data still remains valid. Variables @@ -440,7 +440,7 @@ type Variable struct { // * `Malformed string`, // * `Field f not found in class C` // * `Null pointer dereference` - Status *StatusMessage `protobuf:"bytes,5,opt,name=status" json:"status,omitempty"` + Status *StatusMessage `protobuf:"bytes,5,opt,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -515,15 +515,15 @@ func (m *Variable) GetStatus() *StatusMessage { // Represents a stack frame context. type StackFrame struct { // Demangled function name at the call site. - Function string `protobuf:"bytes,1,opt,name=function" json:"function,omitempty"` + Function string `protobuf:"bytes,1,opt,name=function,proto3" json:"function,omitempty"` // Source location of the call site. - Location *SourceLocation `protobuf:"bytes,2,opt,name=location" json:"location,omitempty"` + Location *SourceLocation `protobuf:"bytes,2,opt,name=location,proto3" json:"location,omitempty"` // Set of arguments passed to this function. // Note that this might not be populated for all stack frames. - Arguments []*Variable `protobuf:"bytes,3,rep,name=arguments" json:"arguments,omitempty"` + Arguments []*Variable `protobuf:"bytes,3,rep,name=arguments,proto3" json:"arguments,omitempty"` // Set of local variables at the stack frame location. // Note that this might not be populated for all stack frames. - Locals []*Variable `protobuf:"bytes,4,rep,name=locals" json:"locals,omitempty"` + Locals []*Variable `protobuf:"bytes,4,rep,name=locals,proto3" json:"locals,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -584,21 +584,21 @@ func (m *StackFrame) GetLocals() []*Variable { // Represents the breakpoint specification, status and results. type Breakpoint struct { // Breakpoint identifier, unique in the scope of the debuggee. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // Action that the agent should perform when the code at the // breakpoint location is hit. - Action Breakpoint_Action `protobuf:"varint,13,opt,name=action,enum=google.devtools.clouddebugger.v2.Breakpoint_Action" json:"action,omitempty"` + Action Breakpoint_Action `protobuf:"varint,13,opt,name=action,proto3,enum=google.devtools.clouddebugger.v2.Breakpoint_Action" json:"action,omitempty"` // Breakpoint source location. - Location *SourceLocation `protobuf:"bytes,2,opt,name=location" json:"location,omitempty"` + Location *SourceLocation `protobuf:"bytes,2,opt,name=location,proto3" json:"location,omitempty"` // Condition that triggers the breakpoint. // The condition is a compound boolean expression composed using expressions // in a programming language at the source location. - Condition string `protobuf:"bytes,3,opt,name=condition" json:"condition,omitempty"` + Condition string `protobuf:"bytes,3,opt,name=condition,proto3" json:"condition,omitempty"` // List of read-only expressions to evaluate at the breakpoint location. // The expressions are composed using expressions in the programming language // at the source location. If the breakpoint action is `LOG`, the evaluated // expressions are included in log statements. - Expressions []string `protobuf:"bytes,4,rep,name=expressions" json:"expressions,omitempty"` + Expressions []string `protobuf:"bytes,4,rep,name=expressions,proto3" json:"expressions,omitempty"` // Only relevant when action is `LOG`. Defines the message to log when // the breakpoint hits. The message may include parameter placeholders `$0`, // `$1`, etc. These placeholders are replaced with the evaluated value @@ -607,19 +607,19 @@ type Breakpoint struct { // // Example: `Message received, id = $0, count = $1` with // `expressions` = `[ message.id, message.count ]`. - LogMessageFormat string `protobuf:"bytes,14,opt,name=log_message_format,json=logMessageFormat" json:"log_message_format,omitempty"` + LogMessageFormat string `protobuf:"bytes,14,opt,name=log_message_format,json=logMessageFormat,proto3" json:"log_message_format,omitempty"` // Indicates the severity of the log. Only relevant when action is `LOG`. - LogLevel Breakpoint_LogLevel `protobuf:"varint,15,opt,name=log_level,json=logLevel,enum=google.devtools.clouddebugger.v2.Breakpoint_LogLevel" json:"log_level,omitempty"` + LogLevel Breakpoint_LogLevel `protobuf:"varint,15,opt,name=log_level,json=logLevel,proto3,enum=google.devtools.clouddebugger.v2.Breakpoint_LogLevel" json:"log_level,omitempty"` // When true, indicates that this is a final result and the // breakpoint state will not change from here on. - IsFinalState bool `protobuf:"varint,5,opt,name=is_final_state,json=isFinalState" json:"is_final_state,omitempty"` + IsFinalState bool `protobuf:"varint,5,opt,name=is_final_state,json=isFinalState,proto3" json:"is_final_state,omitempty"` // Time this breakpoint was created by the server in seconds resolution. - CreateTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time this breakpoint was finalized as seen by the server in seconds // resolution. - FinalTime *timestamp.Timestamp `protobuf:"bytes,12,opt,name=final_time,json=finalTime" json:"final_time,omitempty"` + FinalTime *timestamp.Timestamp `protobuf:"bytes,12,opt,name=final_time,json=finalTime,proto3" json:"final_time,omitempty"` // E-mail address of the user that created this breakpoint - UserEmail string `protobuf:"bytes,16,opt,name=user_email,json=userEmail" json:"user_email,omitempty"` + UserEmail string `protobuf:"bytes,16,opt,name=user_email,json=userEmail,proto3" json:"user_email,omitempty"` // Breakpoint status. // // The status includes an error flag and a human readable message. @@ -635,9 +635,9 @@ type Breakpoint struct { // // * `Invalid line number` referring to location // * `Field f not found in class C` referring to condition - Status *StatusMessage `protobuf:"bytes,10,opt,name=status" json:"status,omitempty"` + Status *StatusMessage `protobuf:"bytes,10,opt,name=status,proto3" json:"status,omitempty"` // The stack at breakpoint time. - StackFrames []*StackFrame `protobuf:"bytes,7,rep,name=stack_frames,json=stackFrames" json:"stack_frames,omitempty"` + StackFrames []*StackFrame `protobuf:"bytes,7,rep,name=stack_frames,json=stackFrames,proto3" json:"stack_frames,omitempty"` // Values of evaluated expressions at breakpoint time. // The evaluated expressions appear in exactly the same order they // are listed in the `expressions` field. @@ -645,7 +645,7 @@ type Breakpoint struct { // `members` field holds the result of the evaluated expression. // If the expression cannot be evaluated, the `status` inside the `Variable` // will indicate an error and contain the error text. - EvaluatedExpressions []*Variable `protobuf:"bytes,8,rep,name=evaluated_expressions,json=evaluatedExpressions" json:"evaluated_expressions,omitempty"` + EvaluatedExpressions []*Variable `protobuf:"bytes,8,rep,name=evaluated_expressions,json=evaluatedExpressions,proto3" json:"evaluated_expressions,omitempty"` // The `variable_table` exists to aid with computation, memory and network // traffic optimization. It enables storing a variable once and reference // it from multiple variables, including variables stored in the @@ -658,10 +658,10 @@ type Breakpoint struct { // The stored objects are nameless and get their name from the referencing // variable. The effective variable is a merge of the referencing variable // and the referenced variable. - VariableTable []*Variable `protobuf:"bytes,9,rep,name=variable_table,json=variableTable" json:"variable_table,omitempty"` + VariableTable []*Variable `protobuf:"bytes,9,rep,name=variable_table,json=variableTable,proto3" json:"variable_table,omitempty"` // A set of custom breakpoint properties, populated by the agent, to be // displayed to the user. - Labels map[string]string `protobuf:"bytes,17,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,17,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -810,46 +810,46 @@ func (m *Breakpoint) GetLabels() map[string]string { // exactly the same Debuggee message value when registering. type Debuggee struct { // Unique identifier for the debuggee generated by the controller service. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // Project the debuggee is associated with. // Use project number or id when registering a Google Cloud Platform project. - Project string `protobuf:"bytes,2,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,2,opt,name=project,proto3" json:"project,omitempty"` // Uniquifier to further distiguish the application. // It is possible that different applications might have identical values in // the debuggee message, thus, incorrectly identified as a single application // by the Controller service. This field adds salt to further distiguish the // application. Agents should consider seeding this field with value that // identifies the code, binary, configuration and environment. - Uniquifier string `protobuf:"bytes,3,opt,name=uniquifier" json:"uniquifier,omitempty"` + Uniquifier string `protobuf:"bytes,3,opt,name=uniquifier,proto3" json:"uniquifier,omitempty"` // Human readable description of the debuggee. // Including a human-readable project name, environment name and version // information is recommended. - Description string `protobuf:"bytes,4,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` // If set to `true`, indicates that Controller service does not detect any // activity from the debuggee agents and the application is possibly stopped. - IsInactive bool `protobuf:"varint,5,opt,name=is_inactive,json=isInactive" json:"is_inactive,omitempty"` + IsInactive bool `protobuf:"varint,5,opt,name=is_inactive,json=isInactive,proto3" json:"is_inactive,omitempty"` // Version ID of the agent. // Schema: `domain/language-platform/vmajor.minor` (for example // `google.com/java-gcp/v1.1`). - AgentVersion string `protobuf:"bytes,6,opt,name=agent_version,json=agentVersion" json:"agent_version,omitempty"` + AgentVersion string `protobuf:"bytes,6,opt,name=agent_version,json=agentVersion,proto3" json:"agent_version,omitempty"` // If set to `true`, indicates that the agent should disable itself and // detach from the debuggee. - IsDisabled bool `protobuf:"varint,7,opt,name=is_disabled,json=isDisabled" json:"is_disabled,omitempty"` + IsDisabled bool `protobuf:"varint,7,opt,name=is_disabled,json=isDisabled,proto3" json:"is_disabled,omitempty"` // Human readable message to be displayed to the user about this debuggee. // Absence of this field indicates no status. The message can be either // informational or an error status. - Status *StatusMessage `protobuf:"bytes,8,opt,name=status" json:"status,omitempty"` + Status *StatusMessage `protobuf:"bytes,8,opt,name=status,proto3" json:"status,omitempty"` // References to the locations and revisions of the source code used in the // deployed application. - SourceContexts []*v1.SourceContext `protobuf:"bytes,9,rep,name=source_contexts,json=sourceContexts" json:"source_contexts,omitempty"` + SourceContexts []*v1.SourceContext `protobuf:"bytes,9,rep,name=source_contexts,json=sourceContexts,proto3" json:"source_contexts,omitempty"` // References to the locations and revisions of the source code used in the // deployed application. // // NOTE: this field is experimental and can be ignored. - ExtSourceContexts []*v1.ExtendedSourceContext `protobuf:"bytes,13,rep,name=ext_source_contexts,json=extSourceContexts" json:"ext_source_contexts,omitempty"` + ExtSourceContexts []*v1.ExtendedSourceContext `protobuf:"bytes,13,rep,name=ext_source_contexts,json=extSourceContexts,proto3" json:"ext_source_contexts,omitempty"` // A set of custom debuggee properties, populated by the agent, to be // displayed to the user. - Labels map[string]string `protobuf:"bytes,11,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,11,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/debugger.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/debugger.pb.go index b6a4bb276..b23cc39e4 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/debugger.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouddebugger/v2/debugger.pb.go @@ -28,13 +28,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Request to set a breakpoint type SetBreakpointRequest struct { // ID of the debuggee where the breakpoint is to be set. - DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId" json:"debuggee_id,omitempty"` + DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId,proto3" json:"debuggee_id,omitempty"` // Breakpoint specification to set. // The field `location` of the breakpoint must be set. - Breakpoint *Breakpoint `protobuf:"bytes,2,opt,name=breakpoint" json:"breakpoint,omitempty"` + Breakpoint *Breakpoint `protobuf:"bytes,2,opt,name=breakpoint,proto3" json:"breakpoint,omitempty"` // The client version making the call. // Schema: `domain/type/version` (e.g., `google.com/intellij/v1`). - ClientVersion string `protobuf:"bytes,4,opt,name=client_version,json=clientVersion" json:"client_version,omitempty"` + ClientVersion string `protobuf:"bytes,4,opt,name=client_version,json=clientVersion,proto3" json:"client_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -89,7 +89,7 @@ func (m *SetBreakpointRequest) GetClientVersion() string { type SetBreakpointResponse struct { // Breakpoint resource. // The field `id` is guaranteed to be set (in addition to the echoed fileds). - Breakpoint *Breakpoint `protobuf:"bytes,1,opt,name=breakpoint" json:"breakpoint,omitempty"` + Breakpoint *Breakpoint `protobuf:"bytes,1,opt,name=breakpoint,proto3" json:"breakpoint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -129,12 +129,12 @@ func (m *SetBreakpointResponse) GetBreakpoint() *Breakpoint { // Request to get breakpoint information. type GetBreakpointRequest struct { // ID of the debuggee whose breakpoint to get. - DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId" json:"debuggee_id,omitempty"` + DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId,proto3" json:"debuggee_id,omitempty"` // ID of the breakpoint to get. - BreakpointId string `protobuf:"bytes,2,opt,name=breakpoint_id,json=breakpointId" json:"breakpoint_id,omitempty"` + BreakpointId string `protobuf:"bytes,2,opt,name=breakpoint_id,json=breakpointId,proto3" json:"breakpoint_id,omitempty"` // The client version making the call. // Schema: `domain/type/version` (e.g., `google.com/intellij/v1`). - ClientVersion string `protobuf:"bytes,4,opt,name=client_version,json=clientVersion" json:"client_version,omitempty"` + ClientVersion string `protobuf:"bytes,4,opt,name=client_version,json=clientVersion,proto3" json:"client_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -189,7 +189,7 @@ func (m *GetBreakpointRequest) GetClientVersion() string { type GetBreakpointResponse struct { // Complete breakpoint state. // The fields `id` and `location` are guaranteed to be set. - Breakpoint *Breakpoint `protobuf:"bytes,1,opt,name=breakpoint" json:"breakpoint,omitempty"` + Breakpoint *Breakpoint `protobuf:"bytes,1,opt,name=breakpoint,proto3" json:"breakpoint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -229,12 +229,12 @@ func (m *GetBreakpointResponse) GetBreakpoint() *Breakpoint { // Request to delete a breakpoint. type DeleteBreakpointRequest struct { // ID of the debuggee whose breakpoint to delete. - DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId" json:"debuggee_id,omitempty"` + DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId,proto3" json:"debuggee_id,omitempty"` // ID of the breakpoint to delete. - BreakpointId string `protobuf:"bytes,2,opt,name=breakpoint_id,json=breakpointId" json:"breakpoint_id,omitempty"` + BreakpointId string `protobuf:"bytes,2,opt,name=breakpoint_id,json=breakpointId,proto3" json:"breakpoint_id,omitempty"` // The client version making the call. // Schema: `domain/type/version` (e.g., `google.com/intellij/v1`). - ClientVersion string `protobuf:"bytes,3,opt,name=client_version,json=clientVersion" json:"client_version,omitempty"` + ClientVersion string `protobuf:"bytes,3,opt,name=client_version,json=clientVersion,proto3" json:"client_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -288,27 +288,27 @@ func (m *DeleteBreakpointRequest) GetClientVersion() string { // Request to list breakpoints. type ListBreakpointsRequest struct { // ID of the debuggee whose breakpoints to list. - DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId" json:"debuggee_id,omitempty"` + DebuggeeId string `protobuf:"bytes,1,opt,name=debuggee_id,json=debuggeeId,proto3" json:"debuggee_id,omitempty"` // When set to `true`, the response includes the list of breakpoints set by // any user. Otherwise, it includes only breakpoints set by the caller. - IncludeAllUsers bool `protobuf:"varint,2,opt,name=include_all_users,json=includeAllUsers" json:"include_all_users,omitempty"` + IncludeAllUsers bool `protobuf:"varint,2,opt,name=include_all_users,json=includeAllUsers,proto3" json:"include_all_users,omitempty"` // When set to `true`, the response includes active and inactive // breakpoints. Otherwise, it includes only active breakpoints. - IncludeInactive bool `protobuf:"varint,3,opt,name=include_inactive,json=includeInactive" json:"include_inactive,omitempty"` + IncludeInactive bool `protobuf:"varint,3,opt,name=include_inactive,json=includeInactive,proto3" json:"include_inactive,omitempty"` // When set, the response includes only breakpoints with the specified action. - Action *ListBreakpointsRequest_BreakpointActionValue `protobuf:"bytes,4,opt,name=action" json:"action,omitempty"` + Action *ListBreakpointsRequest_BreakpointActionValue `protobuf:"bytes,4,opt,name=action,proto3" json:"action,omitempty"` // This field is deprecated. The following fields are always stripped out of // the result: `stack_frames`, `evaluated_expressions` and `variable_table`. - StripResults bool `protobuf:"varint,5,opt,name=strip_results,json=stripResults" json:"strip_results,omitempty"` + StripResults bool `protobuf:"varint,5,opt,name=strip_results,json=stripResults,proto3" json:"strip_results,omitempty"` // A wait token that, if specified, blocks the call until the breakpoints // list has changed, or a server selected timeout has expired. The value // should be set from the last response. The error code // `google.rpc.Code.ABORTED` (RPC) is returned on wait timeout, which // should be called again with the same `wait_token`. - WaitToken string `protobuf:"bytes,6,opt,name=wait_token,json=waitToken" json:"wait_token,omitempty"` + WaitToken string `protobuf:"bytes,6,opt,name=wait_token,json=waitToken,proto3" json:"wait_token,omitempty"` // The client version making the call. // Schema: `domain/type/version` (e.g., `google.com/intellij/v1`). - ClientVersion string `protobuf:"bytes,8,opt,name=client_version,json=clientVersion" json:"client_version,omitempty"` + ClientVersion string `protobuf:"bytes,8,opt,name=client_version,json=clientVersion,proto3" json:"client_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -391,7 +391,7 @@ func (m *ListBreakpointsRequest) GetClientVersion() string { // field of breakpoints. type ListBreakpointsRequest_BreakpointActionValue struct { // Only breakpoints with the specified action will pass the filter. - Value Breakpoint_Action `protobuf:"varint,1,opt,name=value,enum=google.devtools.clouddebugger.v2.Breakpoint_Action" json:"value,omitempty"` + Value Breakpoint_Action `protobuf:"varint,1,opt,name=value,proto3,enum=google.devtools.clouddebugger.v2.Breakpoint_Action" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -438,10 +438,10 @@ type ListBreakpointsResponse struct { // The fields `id` and `location` are guaranteed to be set on each breakpoint. // The fields: `stack_frames`, `evaluated_expressions` and `variable_table` // are cleared on each breakpoint regardless of its status. - Breakpoints []*Breakpoint `protobuf:"bytes,1,rep,name=breakpoints" json:"breakpoints,omitempty"` + Breakpoints []*Breakpoint `protobuf:"bytes,1,rep,name=breakpoints,proto3" json:"breakpoints,omitempty"` // A wait token that can be used in the next call to `list` (REST) or // `ListBreakpoints` (RPC) to block until the list of breakpoints has changes. - NextWaitToken string `protobuf:"bytes,2,opt,name=next_wait_token,json=nextWaitToken" json:"next_wait_token,omitempty"` + NextWaitToken string `protobuf:"bytes,2,opt,name=next_wait_token,json=nextWaitToken,proto3" json:"next_wait_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -488,13 +488,13 @@ func (m *ListBreakpointsResponse) GetNextWaitToken() string { // Request to list debuggees. type ListDebuggeesRequest struct { // Project number of a Google Cloud project whose debuggees to list. - Project string `protobuf:"bytes,2,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,2,opt,name=project,proto3" json:"project,omitempty"` // When set to `true`, the result includes all debuggees. Otherwise, the // result includes only debuggees that are active. - IncludeInactive bool `protobuf:"varint,3,opt,name=include_inactive,json=includeInactive" json:"include_inactive,omitempty"` + IncludeInactive bool `protobuf:"varint,3,opt,name=include_inactive,json=includeInactive,proto3" json:"include_inactive,omitempty"` // The client version making the call. // Schema: `domain/type/version` (e.g., `google.com/intellij/v1`). - ClientVersion string `protobuf:"bytes,4,opt,name=client_version,json=clientVersion" json:"client_version,omitempty"` + ClientVersion string `protobuf:"bytes,4,opt,name=client_version,json=clientVersion,proto3" json:"client_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -551,7 +551,7 @@ type ListDebuggeesResponse struct { // The fields `debuggee.id` and `description` are guaranteed to be set. // The `description` field is a human readable field provided by agents and // can be displayed to users. - Debuggees []*Debuggee `protobuf:"bytes,1,rep,name=debuggees" json:"debuggees,omitempty"` + Debuggees []*Debuggee `protobuf:"bytes,1,rep,name=debuggees,proto3" json:"debuggees,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -678,8 +678,7 @@ func (c *debugger2Client) ListDebuggees(ctx context.Context, in *ListDebuggeesRe return out, nil } -// Server API for Debugger2 service - +// Debugger2Server is the server API for Debugger2 service. type Debugger2Server interface { // Sets the breakpoint to the debuggee. SetBreakpoint(context.Context, *SetBreakpointRequest) (*SetBreakpointResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/common.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/common.pb.go index 9084cf7c9..0063d3339 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/common.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/common.pb.go @@ -25,12 +25,12 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type ErrorGroup struct { // The group resource name. // Example: projects/my-project-123/groups/my-groupid - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Group IDs are unique for a given project. If the same kind of error // occurs in different service contexts, it will receive the same group ID. - GroupId string `protobuf:"bytes,2,opt,name=group_id,json=groupId" json:"group_id,omitempty"` + GroupId string `protobuf:"bytes,2,opt,name=group_id,json=groupId,proto3" json:"group_id,omitempty"` // Associated tracking issues. - TrackingIssues []*TrackingIssue `protobuf:"bytes,3,rep,name=tracking_issues,json=trackingIssues" json:"tracking_issues,omitempty"` + TrackingIssues []*TrackingIssue `protobuf:"bytes,3,rep,name=tracking_issues,json=trackingIssues,proto3" json:"tracking_issues,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -85,7 +85,7 @@ func (m *ErrorGroup) GetTrackingIssues() []*TrackingIssue { type TrackingIssue struct { // A URL pointing to a related entry in an issue tracking system. // Example: https://github.com/user/project/issues/4 - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -127,13 +127,13 @@ type ErrorEvent struct { // Time when the event occurred as provided in the error report. // If the report did not contain a timestamp, the time the error was received // by the Error Reporting system is used. - EventTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=event_time,json=eventTime" json:"event_time,omitempty"` + EventTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=event_time,json=eventTime,proto3" json:"event_time,omitempty"` // The `ServiceContext` for which this error was reported. - ServiceContext *ServiceContext `protobuf:"bytes,2,opt,name=service_context,json=serviceContext" json:"service_context,omitempty"` + ServiceContext *ServiceContext `protobuf:"bytes,2,opt,name=service_context,json=serviceContext,proto3" json:"service_context,omitempty"` // The stack trace that was reported or logged by the service. - Message string `protobuf:"bytes,3,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,3,opt,name=message,proto3" json:"message,omitempty"` // Data about the context in which the error occurred. - Context *ErrorContext `protobuf:"bytes,5,opt,name=context" json:"context,omitempty"` + Context *ErrorContext `protobuf:"bytes,5,opt,name=context,proto3" json:"context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -201,16 +201,16 @@ type ServiceContext struct { // // Contains the service name for error reports extracted from Google // App Engine logs or `default` if the App Engine default service is used. - Service string `protobuf:"bytes,2,opt,name=service" json:"service,omitempty"` + Service string `protobuf:"bytes,2,opt,name=service,proto3" json:"service,omitempty"` // Represents the source code version that the developer provided, // which could represent a version label or a Git SHA-1 hash, for example. - Version string `protobuf:"bytes,3,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,3,opt,name=version,proto3" json:"version,omitempty"` // Type of the MonitoredResource. List of possible values: // https://cloud.google.com/monitoring/api/resources // // Value is set automatically for incoming errors and must not be set when // reporting errors. - ResourceType string `protobuf:"bytes,4,opt,name=resource_type,json=resourceType" json:"resource_type,omitempty"` + ResourceType string `protobuf:"bytes,4,opt,name=resource_type,json=resourceType,proto3" json:"resource_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -268,7 +268,7 @@ func (m *ServiceContext) GetResourceType() string { type ErrorContext struct { // The HTTP request which was processed when the error was // triggered. - HttpRequest *HttpRequestContext `protobuf:"bytes,1,opt,name=http_request,json=httpRequest" json:"http_request,omitempty"` + HttpRequest *HttpRequestContext `protobuf:"bytes,1,opt,name=http_request,json=httpRequest,proto3" json:"http_request,omitempty"` // The user who caused or was affected by the crash. // This can be a user ID, an email address, or an arbitrary token that // uniquely identifies the user. @@ -277,14 +277,14 @@ type ErrorContext struct { // Error Reporting system will use other data, such as remote IP address, to // distinguish affected users. See `affected_users_count` in // `ErrorGroupStats`. - User string `protobuf:"bytes,2,opt,name=user" json:"user,omitempty"` + User string `protobuf:"bytes,2,opt,name=user,proto3" json:"user,omitempty"` // The location in the source code where the decision was made to // report the error, usually the place where it was logged. // For a logged exception this would be the source line where the // exception is logged, usually close to the place where it was // caught. This value is in contrast to `Exception.cause_location`, // which describes the source line where the exception was thrown. - ReportLocation *SourceLocation `protobuf:"bytes,3,opt,name=report_location,json=reportLocation" json:"report_location,omitempty"` + ReportLocation *SourceLocation `protobuf:"bytes,3,opt,name=report_location,json=reportLocation,proto3" json:"report_location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -341,20 +341,20 @@ func (m *ErrorContext) GetReportLocation() *SourceLocation { // error report has been generated automatically from Google App Engine logs. type HttpRequestContext struct { // The type of HTTP request, such as `GET`, `POST`, etc. - Method string `protobuf:"bytes,1,opt,name=method" json:"method,omitempty"` + Method string `protobuf:"bytes,1,opt,name=method,proto3" json:"method,omitempty"` // The URL of the request. - Url string `protobuf:"bytes,2,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` // The user agent information that is provided with the request. - UserAgent string `protobuf:"bytes,3,opt,name=user_agent,json=userAgent" json:"user_agent,omitempty"` + UserAgent string `protobuf:"bytes,3,opt,name=user_agent,json=userAgent,proto3" json:"user_agent,omitempty"` // The referrer information that is provided with the request. - Referrer string `protobuf:"bytes,4,opt,name=referrer" json:"referrer,omitempty"` + Referrer string `protobuf:"bytes,4,opt,name=referrer,proto3" json:"referrer,omitempty"` // The HTTP response status code for the request. - ResponseStatusCode int32 `protobuf:"varint,5,opt,name=response_status_code,json=responseStatusCode" json:"response_status_code,omitempty"` + ResponseStatusCode int32 `protobuf:"varint,5,opt,name=response_status_code,json=responseStatusCode,proto3" json:"response_status_code,omitempty"` // The IP address from which the request originated. // This can be IPv4, IPv6, or a token which is derived from the // IP address, depending on the data that has been provided // in the error report. - RemoteIp string `protobuf:"bytes,6,opt,name=remote_ip,json=remoteIp" json:"remote_ip,omitempty"` + RemoteIp string `protobuf:"bytes,6,opt,name=remote_ip,json=remoteIp,proto3" json:"remote_ip,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -434,13 +434,13 @@ func (m *HttpRequestContext) GetRemoteIp() string { type SourceLocation struct { // The source code filename, which can include a truncated relative // path, or a full path from a production machine. - FilePath string `protobuf:"bytes,1,opt,name=file_path,json=filePath" json:"file_path,omitempty"` + FilePath string `protobuf:"bytes,1,opt,name=file_path,json=filePath,proto3" json:"file_path,omitempty"` // 1-based. 0 indicates that the line number is unknown. - LineNumber int32 `protobuf:"varint,2,opt,name=line_number,json=lineNumber" json:"line_number,omitempty"` + LineNumber int32 `protobuf:"varint,2,opt,name=line_number,json=lineNumber,proto3" json:"line_number,omitempty"` // Human-readable name of a function or method. // The value can include optional context like the class or package name. // For example, `my.package.MyClass.method` in case of Java. - FunctionName string `protobuf:"bytes,4,opt,name=function_name,json=functionName" json:"function_name,omitempty"` + FunctionName string `protobuf:"bytes,4,opt,name=function_name,json=functionName,proto3" json:"function_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_group_service.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_group_service.pb.go index fa40e370e..52d7ac10e 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_group_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_group_service.pb.go @@ -34,7 +34,7 @@ type GetGroupRequest struct { // this project. // // Example: projects/my-project-123/groups/my-group - GroupName string `protobuf:"bytes,1,opt,name=group_name,json=groupName" json:"group_name,omitempty"` + GroupName string `protobuf:"bytes,1,opt,name=group_name,json=groupName,proto3" json:"group_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -74,7 +74,7 @@ func (m *GetGroupRequest) GetGroupName() string { // A request to replace the existing data for the given group. type UpdateGroupRequest struct { // [Required] The group which replaces the resource on the server. - Group *ErrorGroup `protobuf:"bytes,1,opt,name=group" json:"group,omitempty"` + Group *ErrorGroup `protobuf:"bytes,1,opt,name=group,proto3" json:"group,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -161,8 +161,7 @@ func (c *errorGroupServiceClient) UpdateGroup(ctx context.Context, in *UpdateGro return out, nil } -// Server API for ErrorGroupService service - +// ErrorGroupServiceServer is the server API for ErrorGroupService service. type ErrorGroupServiceServer interface { // Get the specified group. GetGroup(context.Context, *GetGroupRequest) (*ErrorGroup, error) diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_stats_service.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_stats_service.pb.go index 671226258..5e2e18f4b 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_stats_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/error_stats_service.pb.go @@ -160,13 +160,13 @@ type ListGroupStatsRequest struct { // Platform project ID. // // Example: projects/my-project-123. - ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName" json:"project_name,omitempty"` + ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName,proto3" json:"project_name,omitempty"` // [Optional] List all ErrorGroupStats with these IDs. - GroupId []string `protobuf:"bytes,2,rep,name=group_id,json=groupId" json:"group_id,omitempty"` + GroupId []string `protobuf:"bytes,2,rep,name=group_id,json=groupId,proto3" json:"group_id,omitempty"` // [Optional] List only ErrorGroupStats which belong to a service // context that matches the filter. // Data for all service contexts is returned if this field is not specified. - ServiceFilter *ServiceContextFilter `protobuf:"bytes,3,opt,name=service_filter,json=serviceFilter" json:"service_filter,omitempty"` + ServiceFilter *ServiceContextFilter `protobuf:"bytes,3,opt,name=service_filter,json=serviceFilter,proto3" json:"service_filter,omitempty"` // [Optional] List data for the given time range. // If not set a default time range is used. The field time_range_begin // in the response will specify the beginning of this time range. @@ -174,26 +174,26 @@ type ListGroupStatsRequest struct { // range are returned, unless the request contains an explicit group_id list. // If a group_id list is given, also ErrorGroupStats with zero // occurrences are returned. - TimeRange *QueryTimeRange `protobuf:"bytes,5,opt,name=time_range,json=timeRange" json:"time_range,omitempty"` + TimeRange *QueryTimeRange `protobuf:"bytes,5,opt,name=time_range,json=timeRange,proto3" json:"time_range,omitempty"` // [Optional] The preferred duration for a single returned `TimedCount`. // If not set, no timed counts are returned. - TimedCountDuration *duration.Duration `protobuf:"bytes,6,opt,name=timed_count_duration,json=timedCountDuration" json:"timed_count_duration,omitempty"` + TimedCountDuration *duration.Duration `protobuf:"bytes,6,opt,name=timed_count_duration,json=timedCountDuration,proto3" json:"timed_count_duration,omitempty"` // [Optional] The alignment of the timed counts to be returned. // Default is `ALIGNMENT_EQUAL_AT_END`. - Alignment TimedCountAlignment `protobuf:"varint,7,opt,name=alignment,enum=google.devtools.clouderrorreporting.v1beta1.TimedCountAlignment" json:"alignment,omitempty"` + Alignment TimedCountAlignment `protobuf:"varint,7,opt,name=alignment,proto3,enum=google.devtools.clouderrorreporting.v1beta1.TimedCountAlignment" json:"alignment,omitempty"` // [Optional] Time where the timed counts shall be aligned if rounded // alignment is chosen. Default is 00:00 UTC. - AlignmentTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=alignment_time,json=alignmentTime" json:"alignment_time,omitempty"` + AlignmentTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=alignment_time,json=alignmentTime,proto3" json:"alignment_time,omitempty"` // [Optional] The sort order in which the results are returned. // Default is `COUNT_DESC`. - Order ErrorGroupOrder `protobuf:"varint,9,opt,name=order,enum=google.devtools.clouderrorreporting.v1beta1.ErrorGroupOrder" json:"order,omitempty"` + Order ErrorGroupOrder `protobuf:"varint,9,opt,name=order,proto3,enum=google.devtools.clouderrorreporting.v1beta1.ErrorGroupOrder" json:"order,omitempty"` // [Optional] The maximum number of results to return per response. // Default is 20. - PageSize int32 `protobuf:"varint,11,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,11,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // [Optional] A `next_page_token` provided by a previous response. To view // additional results, pass this token along with the identical query // parameters as the first request. - PageToken string `protobuf:"bytes,12,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,12,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -296,16 +296,16 @@ func (m *ListGroupStatsRequest) GetPageToken() string { // Contains a set of requested error group stats. type ListGroupStatsResponse struct { // The error group stats which match the given request. - ErrorGroupStats []*ErrorGroupStats `protobuf:"bytes,1,rep,name=error_group_stats,json=errorGroupStats" json:"error_group_stats,omitempty"` + ErrorGroupStats []*ErrorGroupStats `protobuf:"bytes,1,rep,name=error_group_stats,json=errorGroupStats,proto3" json:"error_group_stats,omitempty"` // If non-empty, more results are available. // Pass this token, along with the same query parameters as the first // request, to view the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` // The timestamp specifies the start time to which the request was restricted. // The start time is set based on the requested time range. It may be adjusted // to a later time if a project has exceeded the storage quota and older data // has been deleted. - TimeRangeBegin *timestamp.Timestamp `protobuf:"bytes,4,opt,name=time_range_begin,json=timeRangeBegin" json:"time_range_begin,omitempty"` + TimeRangeBegin *timestamp.Timestamp `protobuf:"bytes,4,opt,name=time_range_begin,json=timeRangeBegin,proto3" json:"time_range_begin,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -360,10 +360,10 @@ func (m *ListGroupStatsResponse) GetTimeRangeBegin() *timestamp.Timestamp { // such as a given time period and/or service filter. type ErrorGroupStats struct { // Group data that is independent of the filter criteria. - Group *ErrorGroup `protobuf:"bytes,1,opt,name=group" json:"group,omitempty"` + Group *ErrorGroup `protobuf:"bytes,1,opt,name=group,proto3" json:"group,omitempty"` // Approximate total number of events in the given group that match // the filter criteria. - Count int64 `protobuf:"varint,2,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,2,opt,name=count,proto3" json:"count,omitempty"` // Approximate number of affected users in the given group that // match the filter criteria. // Users are distinguished by data in the `ErrorContext` of the @@ -376,35 +376,35 @@ type ErrorGroupStats struct { // context that was provided in the error report. If more users are // implicitly affected, such as due to a crash of the whole service, // this is not reflected here. - AffectedUsersCount int64 `protobuf:"varint,3,opt,name=affected_users_count,json=affectedUsersCount" json:"affected_users_count,omitempty"` + AffectedUsersCount int64 `protobuf:"varint,3,opt,name=affected_users_count,json=affectedUsersCount,proto3" json:"affected_users_count,omitempty"` // Approximate number of occurrences over time. // Timed counts returned by ListGroups are guaranteed to be: // // - Inside the requested time interval // - Non-overlapping, and // - Ordered by ascending time. - TimedCounts []*TimedCount `protobuf:"bytes,4,rep,name=timed_counts,json=timedCounts" json:"timed_counts,omitempty"` + TimedCounts []*TimedCount `protobuf:"bytes,4,rep,name=timed_counts,json=timedCounts,proto3" json:"timed_counts,omitempty"` // Approximate first occurrence that was ever seen for this group // and which matches the given filter criteria, ignoring the // time_range that was specified in the request. - FirstSeenTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=first_seen_time,json=firstSeenTime" json:"first_seen_time,omitempty"` + FirstSeenTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=first_seen_time,json=firstSeenTime,proto3" json:"first_seen_time,omitempty"` // Approximate last occurrence that was ever seen for this group and // which matches the given filter criteria, ignoring the time_range // that was specified in the request. - LastSeenTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=last_seen_time,json=lastSeenTime" json:"last_seen_time,omitempty"` + LastSeenTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=last_seen_time,json=lastSeenTime,proto3" json:"last_seen_time,omitempty"` // Service contexts with a non-zero error count for the given filter // criteria. This list can be truncated if multiple services are affected. // Refer to `num_affected_services` for the total count. - AffectedServices []*ServiceContext `protobuf:"bytes,7,rep,name=affected_services,json=affectedServices" json:"affected_services,omitempty"` + AffectedServices []*ServiceContext `protobuf:"bytes,7,rep,name=affected_services,json=affectedServices,proto3" json:"affected_services,omitempty"` // The total number of services with a non-zero error count for the given // filter criteria. - NumAffectedServices int32 `protobuf:"varint,8,opt,name=num_affected_services,json=numAffectedServices" json:"num_affected_services,omitempty"` + NumAffectedServices int32 `protobuf:"varint,8,opt,name=num_affected_services,json=numAffectedServices,proto3" json:"num_affected_services,omitempty"` // An arbitrary event that is chosen as representative for the whole group. // The representative event is intended to be used as a quick preview for // the whole group. Events in the group are usually sufficiently similar // to each other such that showing an arbitrary representative provides // insight into the characteristics of the group as a whole. - Representative *ErrorEvent `protobuf:"bytes,9,opt,name=representative" json:"representative,omitempty"` + Representative *ErrorEvent `protobuf:"bytes,9,opt,name=representative,proto3" json:"representative,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -502,11 +502,11 @@ func (m *ErrorGroupStats) GetRepresentative() *ErrorEvent { // before counting them. type TimedCount struct { // Approximate number of occurrences in the given time period. - Count int64 `protobuf:"varint,1,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,1,opt,name=count,proto3" json:"count,omitempty"` // Start of the time period to which `count` refers (included). - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End of the time period to which `count` refers (excluded). - EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -564,21 +564,21 @@ type ListEventsRequest struct { // [Google Cloud Platform project // ID](https://support.google.com/cloud/answer/6158840). // Example: `projects/my-project-123`. - ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName" json:"project_name,omitempty"` + ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName,proto3" json:"project_name,omitempty"` // [Required] The group for which events shall be returned. - GroupId string `protobuf:"bytes,2,opt,name=group_id,json=groupId" json:"group_id,omitempty"` + GroupId string `protobuf:"bytes,2,opt,name=group_id,json=groupId,proto3" json:"group_id,omitempty"` // [Optional] List only ErrorGroups which belong to a service context that // matches the filter. // Data for all service contexts is returned if this field is not specified. - ServiceFilter *ServiceContextFilter `protobuf:"bytes,3,opt,name=service_filter,json=serviceFilter" json:"service_filter,omitempty"` + ServiceFilter *ServiceContextFilter `protobuf:"bytes,3,opt,name=service_filter,json=serviceFilter,proto3" json:"service_filter,omitempty"` // [Optional] List only data for the given time range. // If not set a default time range is used. The field time_range_begin // in the response will specify the beginning of this time range. - TimeRange *QueryTimeRange `protobuf:"bytes,4,opt,name=time_range,json=timeRange" json:"time_range,omitempty"` + TimeRange *QueryTimeRange `protobuf:"bytes,4,opt,name=time_range,json=timeRange,proto3" json:"time_range,omitempty"` // [Optional] The maximum number of results to return per response. - PageSize int32 `protobuf:"varint,6,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,6,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // [Optional] A `next_page_token` provided by a previous response. - PageToken string `protobuf:"bytes,7,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,7,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -653,13 +653,13 @@ func (m *ListEventsRequest) GetPageToken() string { // Contains a set of requested error events. type ListEventsResponse struct { // The error events which match the given request. - ErrorEvents []*ErrorEvent `protobuf:"bytes,1,rep,name=error_events,json=errorEvents" json:"error_events,omitempty"` + ErrorEvents []*ErrorEvent `protobuf:"bytes,1,rep,name=error_events,json=errorEvents,proto3" json:"error_events,omitempty"` // If non-empty, more results are available. // Pass this token, along with the same query parameters as the first // request, to view the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` // The timestamp specifies the start time to which the request was restricted. - TimeRangeBegin *timestamp.Timestamp `protobuf:"bytes,4,opt,name=time_range_begin,json=timeRangeBegin" json:"time_range_begin,omitempty"` + TimeRangeBegin *timestamp.Timestamp `protobuf:"bytes,4,opt,name=time_range_begin,json=timeRangeBegin,proto3" json:"time_range_begin,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -714,7 +714,7 @@ func (m *ListEventsResponse) GetTimeRangeBegin() *timestamp.Timestamp { // adjusted for lower durations. type QueryTimeRange struct { // Restricts the query to the specified time range. - Period QueryTimeRange_Period `protobuf:"varint,1,opt,name=period,enum=google.devtools.clouderrorreporting.v1beta1.QueryTimeRange_Period" json:"period,omitempty"` + Period QueryTimeRange_Period `protobuf:"varint,1,opt,name=period,proto3,enum=google.devtools.clouderrorreporting.v1beta1.QueryTimeRange_Period" json:"period,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -758,13 +758,13 @@ func (m *QueryTimeRange) GetPeriod() QueryTimeRange_Period { type ServiceContextFilter struct { // [Optional] The exact value to match against // [`ServiceContext.service`](/error-reporting/reference/rest/v1beta1/ServiceContext#FIELDS.service). - Service string `protobuf:"bytes,2,opt,name=service" json:"service,omitempty"` + Service string `protobuf:"bytes,2,opt,name=service,proto3" json:"service,omitempty"` // [Optional] The exact value to match against // [`ServiceContext.version`](/error-reporting/reference/rest/v1beta1/ServiceContext#FIELDS.version). - Version string `protobuf:"bytes,3,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,3,opt,name=version,proto3" json:"version,omitempty"` // [Optional] The exact value to match against // [`ServiceContext.resource_type`](/error-reporting/reference/rest/v1beta1/ServiceContext#FIELDS.resource_type). - ResourceType string `protobuf:"bytes,4,opt,name=resource_type,json=resourceType" json:"resource_type,omitempty"` + ResourceType string `protobuf:"bytes,4,opt,name=resource_type,json=resourceType,proto3" json:"resource_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -822,7 +822,7 @@ type DeleteEventsRequest struct { // [Google Cloud Platform project // ID](https://support.google.com/cloud/answer/6158840). // Example: `projects/my-project-123`. - ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName" json:"project_name,omitempty"` + ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName,proto3" json:"project_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -961,8 +961,7 @@ func (c *errorStatsServiceClient) DeleteEvents(ctx context.Context, in *DeleteEv return out, nil } -// Server API for ErrorStatsService service - +// ErrorStatsServiceServer is the server API for ErrorStatsService service. type ErrorStatsServiceServer interface { // Lists the specified groups. ListGroupStats(context.Context, *ListGroupStatsRequest) (*ListGroupStatsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/report_errors_service.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/report_errors_service.pb.go index f126c62b2..6a9b5ba4c 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/report_errors_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/clouderrorreporting/v1beta1/report_errors_service.pb.go @@ -31,9 +31,9 @@ type ReportErrorEventRequest struct { // as `projects/` plus the // [Google Cloud Platform project ID](https://support.google.com/cloud/answer/6158840). // Example: `projects/my-project-123`. - ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName" json:"project_name,omitempty"` + ProjectName string `protobuf:"bytes,1,opt,name=project_name,json=projectName,proto3" json:"project_name,omitempty"` // [Required] The error event to be reported. - Event *ReportedErrorEvent `protobuf:"bytes,2,opt,name=event" json:"event,omitempty"` + Event *ReportedErrorEvent `protobuf:"bytes,2,opt,name=event,proto3" json:"event,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -114,16 +114,16 @@ type ReportedErrorEvent struct { // [Optional] Time when the event occurred. // If not provided, the time when the event was received by the // Error Reporting system will be used. - EventTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=event_time,json=eventTime" json:"event_time,omitempty"` + EventTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=event_time,json=eventTime,proto3" json:"event_time,omitempty"` // [Required] The service context in which this error has occurred. - ServiceContext *ServiceContext `protobuf:"bytes,2,opt,name=service_context,json=serviceContext" json:"service_context,omitempty"` + ServiceContext *ServiceContext `protobuf:"bytes,2,opt,name=service_context,json=serviceContext,proto3" json:"service_context,omitempty"` // [Required] A message describing the error. The message can contain an // exception stack in one of the supported programming languages and formats. // In that case, the message is parsed and detailed exception information // is returned when retrieving the error event again. - Message string `protobuf:"bytes,3,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,3,opt,name=message,proto3" json:"message,omitempty"` // [Optional] A description of the context in which the error occurred. - Context *ErrorContext `protobuf:"bytes,4,opt,name=context" json:"context,omitempty"` + Context *ErrorContext `protobuf:"bytes,4,opt,name=context,proto3" json:"context,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -227,8 +227,7 @@ func (c *reportErrorsServiceClient) ReportErrorEvent(ctx context.Context, in *Re return out, nil } -// Server API for ReportErrorsService service - +// ReportErrorsServiceServer is the server API for ReportErrorsService service. type ReportErrorsServiceServer interface { // Report an individual error event. // diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/cloudprofiler/v2/profiler.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/cloudprofiler/v2/profiler.pb.go index 1148cadc2..53887f892 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/cloudprofiler/v2/profiler.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/cloudprofiler/v2/profiler.pb.go @@ -79,12 +79,12 @@ func (ProfileType) EnumDescriptor() ([]byte, []int) { // field must be identical to the deployment in the profile. type CreateProfileRequest struct { // Deployment details. - Deployment *Deployment `protobuf:"bytes,1,opt,name=deployment" json:"deployment,omitempty"` + Deployment *Deployment `protobuf:"bytes,1,opt,name=deployment,proto3" json:"deployment,omitempty"` // Online mode: One or more profile types that the agent is capable of // providing. - ProfileType []ProfileType `protobuf:"varint,2,rep,packed,name=profile_type,json=profileType,enum=google.devtools.cloudprofiler.v2.ProfileType" json:"profile_type,omitempty"` + ProfileType []ProfileType `protobuf:"varint,2,rep,packed,name=profile_type,json=profileType,proto3,enum=google.devtools.cloudprofiler.v2.ProfileType" json:"profile_type,omitempty"` // Offline mode: Contents of the profile to create. - Profile *Profile `protobuf:"bytes,3,opt,name=profile" json:"profile,omitempty"` + Profile *Profile `protobuf:"bytes,3,opt,name=profile,proto3" json:"profile,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -138,7 +138,7 @@ func (m *CreateProfileRequest) GetProfile() *Profile { // UpdateProfileRequest contains the profile to update. type UpdateProfileRequest struct { // Profile to update - Profile *Profile `protobuf:"bytes,1,opt,name=profile" json:"profile,omitempty"` + Profile *Profile `protobuf:"bytes,1,opt,name=profile,proto3" json:"profile,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -179,19 +179,19 @@ func (m *UpdateProfileRequest) GetProfile() *Profile { type Profile struct { // Opaque, server-assigned, unique ID for this profile. // Output only. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Type of profile. // Input (for the offline mode) or output (for the online mode). - ProfileType ProfileType `protobuf:"varint,2,opt,name=profile_type,json=profileType,enum=google.devtools.cloudprofiler.v2.ProfileType" json:"profile_type,omitempty"` + ProfileType ProfileType `protobuf:"varint,2,opt,name=profile_type,json=profileType,proto3,enum=google.devtools.cloudprofiler.v2.ProfileType" json:"profile_type,omitempty"` // Deployment this profile corresponds to. - Deployment *Deployment `protobuf:"bytes,3,opt,name=deployment" json:"deployment,omitempty"` + Deployment *Deployment `protobuf:"bytes,3,opt,name=deployment,proto3" json:"deployment,omitempty"` // Duration of the profiling session. // Input (for the offline mode) or output (for the online mode). // The field represents requested profiling duration. It may slightly differ // from the effective profiling duration, which is recorded in the profile // data, in case the profiling can't be stopped immediately (e.g. in case // stopping the profiling is handled asynchronously). - Duration *duration.Duration `protobuf:"bytes,4,opt,name=duration" json:"duration,omitempty"` + Duration *duration.Duration `protobuf:"bytes,4,opt,name=duration,proto3" json:"duration,omitempty"` // Profile bytes, as a gzip compressed serialized proto, the format is // https://github.com/google/pprof/blob/master/proto/profile.proto. ProfileBytes []byte `protobuf:"bytes,5,opt,name=profile_bytes,json=profileBytes,proto3" json:"profile_bytes,omitempty"` @@ -199,7 +199,7 @@ type Profile struct { // with the deployment labels for the final data set. // See documentation on deployment labels for validation rules and limits. // Input only, will not be populated on responses. - Labels map[string]string `protobuf:"bytes,6,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,6,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -275,14 +275,14 @@ func (m *Profile) GetLabels() map[string]string { type Deployment struct { // Project ID is the ID of a cloud project. // Validation regex: `^[a-z][-a-z0-9:.]{4,61}[a-z0-9]$`. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Target is the service name used to group related deployments: // * Service name for GAE Flex / Standard. // * Cluster and container name for GKE. // * User-specified string for direct GCE profiling (e.g. Java). // * Job name for Dataflow. // Validation regex: `^[a-z]([-a-z0-9_.]{0,253}[a-z0-9])?$`. - Target string `protobuf:"bytes,2,opt,name=target" json:"target,omitempty"` + Target string `protobuf:"bytes,2,opt,name=target,proto3" json:"target,omitempty"` // Labels identify the deployment within the user universe and same target. // Validation regex for label names: `^[a-z0-9]([a-z0-9-]{0,61}[a-z0-9])?$`. // Value for an individual label must be <= 512 bytes, the total @@ -296,7 +296,7 @@ type Deployment struct { // should be present describing the deployment location. An example of a zone // is "us-central1-a", an example of a region is "us-central1" or // "us-central". - Labels map[string]string `protobuf:"bytes,3,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -418,8 +418,7 @@ func (c *profilerServiceClient) UpdateProfile(ctx context.Context, in *UpdatePro return out, nil } -// Server API for ProfilerService service - +// ProfilerServiceServer is the server API for ProfilerService service. type ProfilerServiceServer interface { // CreateProfile creates a new profile resource. // diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v1/trace.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v1/trace.pb.go index 5f69519d9..30b3fc681 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v1/trace.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v1/trace.pb.go @@ -102,12 +102,12 @@ func (ListTracesRequest_ViewType) EnumDescriptor() ([]byte, []int) { // timed event within the operation. type Trace struct { // Project ID of the Cloud project where the trace data is stored. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Globally unique identifier for the trace. This identifier is a 128-bit // numeric value formatted as a 32-byte hex string. - TraceId string `protobuf:"bytes,2,opt,name=trace_id,json=traceId" json:"trace_id,omitempty"` + TraceId string `protobuf:"bytes,2,opt,name=trace_id,json=traceId,proto3" json:"trace_id,omitempty"` // Collection of spans in the trace. - Spans []*TraceSpan `protobuf:"bytes,3,rep,name=spans" json:"spans,omitempty"` + Spans []*TraceSpan `protobuf:"bytes,3,rep,name=spans,proto3" json:"spans,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -161,7 +161,7 @@ func (m *Trace) GetSpans() []*TraceSpan { // List of new or updated traces. type Traces struct { // List of traces. - Traces []*Trace `protobuf:"bytes,1,rep,name=traces" json:"traces,omitempty"` + Traces []*Trace `protobuf:"bytes,1,rep,name=traces,proto3" json:"traces,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -206,11 +206,11 @@ func (m *Traces) GetTraces() []*Trace { type TraceSpan struct { // Identifier for the span. Must be a 64-bit integer other than 0 and // unique within a trace. - SpanId uint64 `protobuf:"fixed64,1,opt,name=span_id,json=spanId" json:"span_id,omitempty"` + SpanId uint64 `protobuf:"fixed64,1,opt,name=span_id,json=spanId,proto3" json:"span_id,omitempty"` // Distinguishes between spans generated in a particular context. For example, // two spans with the same name may be distinguished using `RPC_CLIENT` // and `RPC_SERVER` to identify queueing latency associated with the span. - Kind TraceSpan_SpanKind `protobuf:"varint,2,opt,name=kind,enum=google.devtools.cloudtrace.v1.TraceSpan_SpanKind" json:"kind,omitempty"` + Kind TraceSpan_SpanKind `protobuf:"varint,2,opt,name=kind,proto3,enum=google.devtools.cloudtrace.v1.TraceSpan_SpanKind" json:"kind,omitempty"` // Name of the span. Must be less than 128 bytes. The span name is sanitized // and displayed in the Stackdriver Trace tool in the // {% dynamic print site_values.console_name %}. @@ -218,13 +218,13 @@ type TraceSpan struct { // For the same executable and the same call point, a best practice is // to use a consistent name, which makes it easier to correlate // cross-trace spans. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // Start time of the span in nanoseconds from the UNIX epoch. - StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End time of the span in nanoseconds from the UNIX epoch. - EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // ID of the parent span, if any. Optional. - ParentSpanId uint64 `protobuf:"fixed64,6,opt,name=parent_span_id,json=parentSpanId" json:"parent_span_id,omitempty"` + ParentSpanId uint64 `protobuf:"fixed64,6,opt,name=parent_span_id,json=parentSpanId,proto3" json:"parent_span_id,omitempty"` // Collection of labels associated with the span. Label keys must be less than // 128 bytes. Label values must be less than 16 kilobytes (10MB for // `/stacktrace` values). @@ -260,7 +260,7 @@ type TraceSpan struct { // * `/pid` // * `/stacktrace` // * `/tid` - Labels map[string]string `protobuf:"bytes,7,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,7,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -343,23 +343,23 @@ func (m *TraceSpan) GetLabels() map[string]string { // unless specified. type ListTracesRequest struct { // ID of the Cloud project where the trace data is stored. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Type of data returned for traces in the list. Optional. Default is // `MINIMAL`. - View ListTracesRequest_ViewType `protobuf:"varint,2,opt,name=view,enum=google.devtools.cloudtrace.v1.ListTracesRequest_ViewType" json:"view,omitempty"` + View ListTracesRequest_ViewType `protobuf:"varint,2,opt,name=view,proto3,enum=google.devtools.cloudtrace.v1.ListTracesRequest_ViewType" json:"view,omitempty"` // Maximum number of traces to return. If not specified or <= 0, the // implementation selects a reasonable value. The implementation may // return fewer traces than the requested page size. Optional. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token identifying the page of results to return. If provided, use the // value of the `next_page_token` field from a previous request. Optional. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Start of the time interval (inclusive) during which the trace data was // collected from the application. - StartTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End of the time interval (inclusive) during which the trace data was // collected from the application. - EndTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // An optional filter against labels for the request. // // By default, searches use prefix matching. To specify exact match, prepend @@ -390,7 +390,7 @@ type ListTracesRequest struct { // specified. // * `method:VALUE`: Equivalent to `/http/method:VALUE`. // * `url:VALUE`: Equivalent to `/http/url:VALUE`. - Filter string `protobuf:"bytes,7,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,7,opt,name=filter,proto3" json:"filter,omitempty"` // Field used to sort the returned traces. Optional. // Can be one of the following: // @@ -404,7 +404,7 @@ type ListTracesRequest struct { // (for example, `name desc`). // // Only one sort field is permitted. - OrderBy string `protobuf:"bytes,8,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy string `protobuf:"bytes,8,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -493,11 +493,11 @@ func (m *ListTracesRequest) GetOrderBy() string { // The response message for the `ListTraces` method. type ListTracesResponse struct { // List of trace records returned. - Traces []*Trace `protobuf:"bytes,1,rep,name=traces" json:"traces,omitempty"` + Traces []*Trace `protobuf:"bytes,1,rep,name=traces,proto3" json:"traces,omitempty"` // If defined, indicates that there are more traces that match the request // and that this value should be passed to the next request to continue // retrieving additional traces. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -544,9 +544,9 @@ func (m *ListTracesResponse) GetNextPageToken() string { // The request message for the `GetTrace` method. type GetTraceRequest struct { // ID of the Cloud project where the trace data is stored. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // ID of the trace to return. - TraceId string `protobuf:"bytes,2,opt,name=trace_id,json=traceId" json:"trace_id,omitempty"` + TraceId string `protobuf:"bytes,2,opt,name=trace_id,json=traceId,proto3" json:"trace_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -593,9 +593,9 @@ func (m *GetTraceRequest) GetTraceId() string { // The request message for the `PatchTraces` method. type PatchTracesRequest struct { // ID of the Cloud project where the trace data is stored. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The body of the message. - Traces *Traces `protobuf:"bytes,2,opt,name=traces" json:"traces,omitempty"` + Traces *Traces `protobuf:"bytes,2,opt,name=traces,proto3" json:"traces,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -711,8 +711,7 @@ func (c *traceServiceClient) PatchTraces(ctx context.Context, in *PatchTracesReq return out, nil } -// Server API for TraceService service - +// TraceServiceServer is the server API for TraceService service. type TraceServiceServer interface { // Returns of a list of traces that match the specified filter conditions. ListTraces(context.Context, *ListTracesRequest) (*ListTracesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/trace.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/trace.pb.go index e9d968dac..acdc2f11b 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/trace.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/trace.pb.go @@ -99,12 +99,12 @@ type Span struct { // // [SPAN_ID] is a unique identifier for a span within a trace; it // is a 16-character hexadecimal encoding of an 8-byte array. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The [SPAN_ID] portion of the span's resource name. - SpanId string `protobuf:"bytes,2,opt,name=span_id,json=spanId" json:"span_id,omitempty"` + SpanId string `protobuf:"bytes,2,opt,name=span_id,json=spanId,proto3" json:"span_id,omitempty"` // The [SPAN_ID] of this span's parent span. If this is a root span, // then this field must be empty. - ParentSpanId string `protobuf:"bytes,3,opt,name=parent_span_id,json=parentSpanId" json:"parent_span_id,omitempty"` + ParentSpanId string `protobuf:"bytes,3,opt,name=parent_span_id,json=parentSpanId,proto3" json:"parent_span_id,omitempty"` // A description of the span's operation (up to 128 bytes). // Stackdriver Trace displays the description in the // {% dynamic print site_values.console_name %}. @@ -112,35 +112,35 @@ type Span struct { // and a line number where the operation is called. A best practice is to use // the same display name within an application and at the same call point. // This makes it easier to correlate spans in different traces. - DisplayName *TruncatableString `protobuf:"bytes,4,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName *TruncatableString `protobuf:"bytes,4,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The start time of the span. On the client side, this is the time kept by // the local machine where the span execution starts. On the server side, this // is the time when the server's application handler starts running. - StartTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The end time of the span. On the client side, this is the time kept by // the local machine where the span execution ends. On the server side, this // is the time when the server application handler stops running. - EndTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // A set of attributes on the span. You can have up to 32 attributes per // span. - Attributes *Span_Attributes `protobuf:"bytes,7,opt,name=attributes" json:"attributes,omitempty"` + Attributes *Span_Attributes `protobuf:"bytes,7,opt,name=attributes,proto3" json:"attributes,omitempty"` // Stack trace captured at the start of the span. - StackTrace *StackTrace `protobuf:"bytes,8,opt,name=stack_trace,json=stackTrace" json:"stack_trace,omitempty"` + StackTrace *StackTrace `protobuf:"bytes,8,opt,name=stack_trace,json=stackTrace,proto3" json:"stack_trace,omitempty"` // A set of time events. You can have up to 32 annotations and 128 message // events per span. - TimeEvents *Span_TimeEvents `protobuf:"bytes,9,opt,name=time_events,json=timeEvents" json:"time_events,omitempty"` + TimeEvents *Span_TimeEvents `protobuf:"bytes,9,opt,name=time_events,json=timeEvents,proto3" json:"time_events,omitempty"` // Links associated with the span. You can have up to 128 links per Span. - Links *Span_Links `protobuf:"bytes,10,opt,name=links" json:"links,omitempty"` + Links *Span_Links `protobuf:"bytes,10,opt,name=links,proto3" json:"links,omitempty"` // An optional final status for this span. - Status *status.Status `protobuf:"bytes,11,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,11,opt,name=status,proto3" json:"status,omitempty"` // (Optional) Set this parameter to indicate whether this span is in // the same process as its parent. If you do not set this parameter, // Stackdriver Trace is unable to take advantage of this helpful // information. - SameProcessAsParentSpan *wrappers.BoolValue `protobuf:"bytes,12,opt,name=same_process_as_parent_span,json=sameProcessAsParentSpan" json:"same_process_as_parent_span,omitempty"` + SameProcessAsParentSpan *wrappers.BoolValue `protobuf:"bytes,12,opt,name=same_process_as_parent_span,json=sameProcessAsParentSpan,proto3" json:"same_process_as_parent_span,omitempty"` // An optional number of child spans that were generated while this span // was active. If set, allows implementation to detect missing child spans. - ChildSpanCount *wrappers.Int32Value `protobuf:"bytes,13,opt,name=child_span_count,json=childSpanCount" json:"child_span_count,omitempty"` + ChildSpanCount *wrappers.Int32Value `protobuf:"bytes,13,opt,name=child_span_count,json=childSpanCount,proto3" json:"child_span_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -271,11 +271,11 @@ type Span_Attributes struct { // "/http/user_agent": "" // "/http/request_bytes": 300 // "abc.com/myattribute": true - AttributeMap map[string]*AttributeValue `protobuf:"bytes,1,rep,name=attribute_map,json=attributeMap" json:"attribute_map,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + AttributeMap map[string]*AttributeValue `protobuf:"bytes,1,rep,name=attribute_map,json=attributeMap,proto3" json:"attribute_map,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The number of attributes that were discarded. Attributes can be discarded // because their keys are too long or because there are too many attributes. // If this value is 0 then all attributes are valid. - DroppedAttributesCount int32 `protobuf:"varint,2,opt,name=dropped_attributes_count,json=droppedAttributesCount" json:"dropped_attributes_count,omitempty"` + DroppedAttributesCount int32 `protobuf:"varint,2,opt,name=dropped_attributes_count,json=droppedAttributesCount,proto3" json:"dropped_attributes_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -322,7 +322,7 @@ func (m *Span_Attributes) GetDroppedAttributesCount() int32 { // A time-stamped annotation or message event in the Span. type Span_TimeEvent struct { // The timestamp indicating the time the event occurred. - Time *timestamp.Timestamp `protobuf:"bytes,1,opt,name=time" json:"time,omitempty"` + Time *timestamp.Timestamp `protobuf:"bytes,1,opt,name=time,proto3" json:"time,omitempty"` // A `TimeEvent` can contain either an `Annotation` object or a // `MessageEvent` object, but not both. // @@ -364,10 +364,10 @@ type isSpan_TimeEvent_Value interface { } type Span_TimeEvent_Annotation_ struct { - Annotation *Span_TimeEvent_Annotation `protobuf:"bytes,2,opt,name=annotation,oneof"` + Annotation *Span_TimeEvent_Annotation `protobuf:"bytes,2,opt,name=annotation,proto3,oneof"` } type Span_TimeEvent_MessageEvent_ struct { - MessageEvent *Span_TimeEvent_MessageEvent `protobuf:"bytes,3,opt,name=message_event,json=messageEvent,oneof"` + MessageEvent *Span_TimeEvent_MessageEvent `protobuf:"bytes,3,opt,name=message_event,json=messageEvent,proto3,oneof"` } func (*Span_TimeEvent_Annotation_) isSpan_TimeEvent_Value() {} @@ -479,10 +479,10 @@ func _Span_TimeEvent_OneofSizer(msg proto.Message) (n int) { type Span_TimeEvent_Annotation struct { // A user-supplied message describing the event. The maximum length for // the description is 256 bytes. - Description *TruncatableString `protobuf:"bytes,1,opt,name=description" json:"description,omitempty"` + Description *TruncatableString `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` // A set of attributes on the annotation. You can have up to 4 attributes // per Annotation. - Attributes *Span_Attributes `protobuf:"bytes,2,opt,name=attributes" json:"attributes,omitempty"` + Attributes *Span_Attributes `protobuf:"bytes,2,opt,name=attributes,proto3" json:"attributes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -530,16 +530,16 @@ func (m *Span_TimeEvent_Annotation) GetAttributes() *Span_Attributes { type Span_TimeEvent_MessageEvent struct { // Type of MessageEvent. Indicates whether the message was sent or // received. - Type Span_TimeEvent_MessageEvent_Type `protobuf:"varint,1,opt,name=type,enum=google.devtools.cloudtrace.v2.Span_TimeEvent_MessageEvent_Type" json:"type,omitempty"` + Type Span_TimeEvent_MessageEvent_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.cloudtrace.v2.Span_TimeEvent_MessageEvent_Type" json:"type,omitempty"` // An identifier for the MessageEvent's message that can be used to match // SENT and RECEIVED MessageEvents. It is recommended to be unique within // a Span. - Id int64 `protobuf:"varint,2,opt,name=id" json:"id,omitempty"` + Id int64 `protobuf:"varint,2,opt,name=id,proto3" json:"id,omitempty"` // The number of uncompressed bytes sent or received. - UncompressedSizeBytes int64 `protobuf:"varint,3,opt,name=uncompressed_size_bytes,json=uncompressedSizeBytes" json:"uncompressed_size_bytes,omitempty"` + UncompressedSizeBytes int64 `protobuf:"varint,3,opt,name=uncompressed_size_bytes,json=uncompressedSizeBytes,proto3" json:"uncompressed_size_bytes,omitempty"` // The number of compressed bytes sent or received. If missing assumed to // be the same size as uncompressed. - CompressedSizeBytes int64 `protobuf:"varint,4,opt,name=compressed_size_bytes,json=compressedSizeBytes" json:"compressed_size_bytes,omitempty"` + CompressedSizeBytes int64 `protobuf:"varint,4,opt,name=compressed_size_bytes,json=compressedSizeBytes,proto3" json:"compressed_size_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -602,13 +602,13 @@ func (m *Span_TimeEvent_MessageEvent) GetCompressedSizeBytes() int64 { // details of a message sent/received between Spans. type Span_TimeEvents struct { // A collection of `TimeEvent`s. - TimeEvent []*Span_TimeEvent `protobuf:"bytes,1,rep,name=time_event,json=timeEvent" json:"time_event,omitempty"` + TimeEvent []*Span_TimeEvent `protobuf:"bytes,1,rep,name=time_event,json=timeEvent,proto3" json:"time_event,omitempty"` // The number of dropped annotations in all the included time events. // If the value is 0, then no annotations were dropped. - DroppedAnnotationsCount int32 `protobuf:"varint,2,opt,name=dropped_annotations_count,json=droppedAnnotationsCount" json:"dropped_annotations_count,omitempty"` + DroppedAnnotationsCount int32 `protobuf:"varint,2,opt,name=dropped_annotations_count,json=droppedAnnotationsCount,proto3" json:"dropped_annotations_count,omitempty"` // The number of dropped message events in all the included time events. // If the value is 0, then no message events were dropped. - DroppedMessageEventsCount int32 `protobuf:"varint,3,opt,name=dropped_message_events_count,json=droppedMessageEventsCount" json:"dropped_message_events_count,omitempty"` + DroppedMessageEventsCount int32 `protobuf:"varint,3,opt,name=dropped_message_events_count,json=droppedMessageEventsCount,proto3" json:"dropped_message_events_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -665,14 +665,14 @@ func (m *Span_TimeEvents) GetDroppedMessageEventsCount() int32 { // traces or when the handler receives a request from a different project. type Span_Link struct { // The [TRACE_ID] for a trace within a project. - TraceId string `protobuf:"bytes,1,opt,name=trace_id,json=traceId" json:"trace_id,omitempty"` + TraceId string `protobuf:"bytes,1,opt,name=trace_id,json=traceId,proto3" json:"trace_id,omitempty"` // The [SPAN_ID] for a span within a trace. - SpanId string `protobuf:"bytes,2,opt,name=span_id,json=spanId" json:"span_id,omitempty"` + SpanId string `protobuf:"bytes,2,opt,name=span_id,json=spanId,proto3" json:"span_id,omitempty"` // The relationship of the current span relative to the linked span. - Type Span_Link_Type `protobuf:"varint,3,opt,name=type,enum=google.devtools.cloudtrace.v2.Span_Link_Type" json:"type,omitempty"` + Type Span_Link_Type `protobuf:"varint,3,opt,name=type,proto3,enum=google.devtools.cloudtrace.v2.Span_Link_Type" json:"type,omitempty"` // A set of attributes on the link. You have have up to 32 attributes per // link. - Attributes *Span_Attributes `protobuf:"bytes,4,opt,name=attributes" json:"attributes,omitempty"` + Attributes *Span_Attributes `protobuf:"bytes,4,opt,name=attributes,proto3" json:"attributes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -734,10 +734,10 @@ func (m *Span_Link) GetAttributes() *Span_Attributes { // in the same or different trace. type Span_Links struct { // A collection of links. - Link []*Span_Link `protobuf:"bytes,1,rep,name=link" json:"link,omitempty"` + Link []*Span_Link `protobuf:"bytes,1,rep,name=link,proto3" json:"link,omitempty"` // The number of dropped links after the maximum size was enforced. If // this value is 0, then no links were dropped. - DroppedLinksCount int32 `protobuf:"varint,2,opt,name=dropped_links_count,json=droppedLinksCount" json:"dropped_links_count,omitempty"` + DroppedLinksCount int32 `protobuf:"varint,2,opt,name=dropped_links_count,json=droppedLinksCount,proto3" json:"dropped_links_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -824,13 +824,13 @@ type isAttributeValue_Value interface { } type AttributeValue_StringValue struct { - StringValue *TruncatableString `protobuf:"bytes,1,opt,name=string_value,json=stringValue,oneof"` + StringValue *TruncatableString `protobuf:"bytes,1,opt,name=string_value,json=stringValue,proto3,oneof"` } type AttributeValue_IntValue struct { - IntValue int64 `protobuf:"varint,2,opt,name=int_value,json=intValue,oneof"` + IntValue int64 `protobuf:"varint,2,opt,name=int_value,json=intValue,proto3,oneof"` } type AttributeValue_BoolValue struct { - BoolValue bool `protobuf:"varint,3,opt,name=bool_value,json=boolValue,oneof"` + BoolValue bool `protobuf:"varint,3,opt,name=bool_value,json=boolValue,proto3,oneof"` } func (*AttributeValue_StringValue) isAttributeValue_Value() {} @@ -955,7 +955,7 @@ func _AttributeValue_OneofSizer(msg proto.Message) (n int) { // A call stack appearing in a trace. type StackTrace struct { // Stack frames in this stack trace. A maximum of 128 frames are allowed. - StackFrames *StackTrace_StackFrames `protobuf:"bytes,1,opt,name=stack_frames,json=stackFrames" json:"stack_frames,omitempty"` + StackFrames *StackTrace_StackFrames `protobuf:"bytes,1,opt,name=stack_frames,json=stackFrames,proto3" json:"stack_frames,omitempty"` // The hash ID is used to conserve network bandwidth for duplicate // stack traces within a single trace. // @@ -965,7 +965,7 @@ type StackTrace struct { // // Subsequent spans within the same request can refer // to that stack trace by only setting `stackTraceHashId`. - StackTraceHashId int64 `protobuf:"varint,2,opt,name=stack_trace_hash_id,json=stackTraceHashId" json:"stack_trace_hash_id,omitempty"` + StackTraceHashId int64 `protobuf:"varint,2,opt,name=stack_trace_hash_id,json=stackTraceHashId,proto3" json:"stack_trace_hash_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1013,23 +1013,23 @@ func (m *StackTrace) GetStackTraceHashId() int64 { type StackTrace_StackFrame struct { // The fully-qualified name that uniquely identifies the function or // method that is active in this frame (up to 1024 bytes). - FunctionName *TruncatableString `protobuf:"bytes,1,opt,name=function_name,json=functionName" json:"function_name,omitempty"` + FunctionName *TruncatableString `protobuf:"bytes,1,opt,name=function_name,json=functionName,proto3" json:"function_name,omitempty"` // An un-mangled function name, if `function_name` is // [mangled](http://www.avabodh.com/cxxin/namemangling.html). The name can // be fully-qualified (up to 1024 bytes). - OriginalFunctionName *TruncatableString `protobuf:"bytes,2,opt,name=original_function_name,json=originalFunctionName" json:"original_function_name,omitempty"` + OriginalFunctionName *TruncatableString `protobuf:"bytes,2,opt,name=original_function_name,json=originalFunctionName,proto3" json:"original_function_name,omitempty"` // The name of the source file where the function call appears (up to 256 // bytes). - FileName *TruncatableString `protobuf:"bytes,3,opt,name=file_name,json=fileName" json:"file_name,omitempty"` + FileName *TruncatableString `protobuf:"bytes,3,opt,name=file_name,json=fileName,proto3" json:"file_name,omitempty"` // The line number in `file_name` where the function call appears. - LineNumber int64 `protobuf:"varint,4,opt,name=line_number,json=lineNumber" json:"line_number,omitempty"` + LineNumber int64 `protobuf:"varint,4,opt,name=line_number,json=lineNumber,proto3" json:"line_number,omitempty"` // The column number where the function call appears, if available. // This is important in JavaScript because of its anonymous functions. - ColumnNumber int64 `protobuf:"varint,5,opt,name=column_number,json=columnNumber" json:"column_number,omitempty"` + ColumnNumber int64 `protobuf:"varint,5,opt,name=column_number,json=columnNumber,proto3" json:"column_number,omitempty"` // The binary module from where the code was loaded. - LoadModule *Module `protobuf:"bytes,6,opt,name=load_module,json=loadModule" json:"load_module,omitempty"` + LoadModule *Module `protobuf:"bytes,6,opt,name=load_module,json=loadModule,proto3" json:"load_module,omitempty"` // The version of the deployed source code (up to 128 bytes). - SourceVersion *TruncatableString `protobuf:"bytes,7,opt,name=source_version,json=sourceVersion" json:"source_version,omitempty"` + SourceVersion *TruncatableString `protobuf:"bytes,7,opt,name=source_version,json=sourceVersion,proto3" json:"source_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1111,11 +1111,11 @@ func (m *StackTrace_StackFrame) GetSourceVersion() *TruncatableString { // A collection of stack frames, which can be truncated. type StackTrace_StackFrames struct { // Stack frames in this call stack. - Frame []*StackTrace_StackFrame `protobuf:"bytes,1,rep,name=frame" json:"frame,omitempty"` + Frame []*StackTrace_StackFrame `protobuf:"bytes,1,rep,name=frame,proto3" json:"frame,omitempty"` // The number of stack frames that were dropped because there // were too many stack frames. // If this value is 0, then no stack frames were dropped. - DroppedFramesCount int32 `protobuf:"varint,2,opt,name=dropped_frames_count,json=droppedFramesCount" json:"dropped_frames_count,omitempty"` + DroppedFramesCount int32 `protobuf:"varint,2,opt,name=dropped_frames_count,json=droppedFramesCount,proto3" json:"dropped_frames_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1163,10 +1163,10 @@ func (m *StackTrace_StackFrames) GetDroppedFramesCount() int32 { type Module struct { // For example: main binary, kernel modules, and dynamic libraries // such as libc.so, sharedlib.so (up to 256 bytes). - Module *TruncatableString `protobuf:"bytes,1,opt,name=module" json:"module,omitempty"` + Module *TruncatableString `protobuf:"bytes,1,opt,name=module,proto3" json:"module,omitempty"` // A unique identifier for the module, usually a hash of its // contents (up to 128 bytes). - BuildId *TruncatableString `protobuf:"bytes,2,opt,name=build_id,json=buildId" json:"build_id,omitempty"` + BuildId *TruncatableString `protobuf:"bytes,2,opt,name=build_id,json=buildId,proto3" json:"build_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1219,10 +1219,10 @@ type TruncatableString struct { // Truncation always happens on a UTF8 character boundary. If there // are multi-byte characters in the string, then the length of the // shortened string might be less than the size limit. - Value string `protobuf:"bytes,1,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` // The number of bytes removed from the original string. If this // value is 0, then the string was not shortened. - TruncatedByteCount int32 `protobuf:"varint,2,opt,name=truncated_byte_count,json=truncatedByteCount" json:"truncated_byte_count,omitempty"` + TruncatedByteCount int32 `protobuf:"varint,2,opt,name=truncated_byte_count,json=truncatedByteCount,proto3" json:"truncated_byte_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/tracing.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/tracing.pb.go index ea53eb884..6d2dd510c 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/tracing.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/cloudtrace/v2/tracing.pb.go @@ -30,10 +30,10 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type BatchWriteSpansRequest struct { // Required. The name of the project where the spans belong. The format is // `projects/[PROJECT_ID]`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A list of new spans. The span names must not match existing // spans, or the results are undefined. - Spans []*Span `protobuf:"bytes,2,rep,name=spans" json:"spans,omitempty"` + Spans []*Span `protobuf:"bytes,2,rep,name=spans,proto3" json:"spans,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -126,8 +126,7 @@ func (c *traceServiceClient) CreateSpan(ctx context.Context, in *Span, opts ...g return out, nil } -// Server API for TraceService service - +// TraceServiceServer is the server API for TraceService service. type TraceServiceServer interface { // Sends new spans to new or existing traces. You cannot update // existing spans. diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/bill_of_materials.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/bill_of_materials.pb.go index 5bb0dbe58..794f02fae 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/bill_of_materials.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/bill_of_materials.pb.go @@ -85,19 +85,19 @@ var xxx_messageInfo_PackageManager proto.InternalMessageInfo type PackageManager_Distribution struct { // The cpe_uri in [cpe format](https://cpe.mitre.org/specification/) // denoting the package manager version distributing a package. - CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri" json:"cpe_uri,omitempty"` + CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri,proto3" json:"cpe_uri,omitempty"` // The CPU architecture for which packages in this distribution // channel were built - Architecture PackageManager_Architecture `protobuf:"varint,2,opt,name=architecture,enum=google.devtools.containeranalysis.v1alpha1.PackageManager_Architecture" json:"architecture,omitempty"` + Architecture PackageManager_Architecture `protobuf:"varint,2,opt,name=architecture,proto3,enum=google.devtools.containeranalysis.v1alpha1.PackageManager_Architecture" json:"architecture,omitempty"` // The latest available version of this package in // this distribution channel. - LatestVersion *VulnerabilityType_Version `protobuf:"bytes,3,opt,name=latest_version,json=latestVersion" json:"latest_version,omitempty"` + LatestVersion *VulnerabilityType_Version `protobuf:"bytes,3,opt,name=latest_version,json=latestVersion,proto3" json:"latest_version,omitempty"` // A freeform string denoting the maintainer of this package. - Maintainer string `protobuf:"bytes,4,opt,name=maintainer" json:"maintainer,omitempty"` + Maintainer string `protobuf:"bytes,4,opt,name=maintainer,proto3" json:"maintainer,omitempty"` // The distribution channel-specific homepage for this package. - Url string `protobuf:"bytes,6,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,6,opt,name=url,proto3" json:"url,omitempty"` // The distribution channel-specific description of this package. - Description string `protobuf:"bytes,7,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,7,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -175,11 +175,11 @@ func (m *PackageManager_Distribution) GetDescription() string { type PackageManager_Location struct { // The cpe_uri in [cpe format](https://cpe.mitre.org/specification/) // denoting the package manager version distributing a package. - CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri" json:"cpe_uri,omitempty"` + CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri,proto3" json:"cpe_uri,omitempty"` // The version installed at this location. - Version *VulnerabilityType_Version `protobuf:"bytes,2,opt,name=version" json:"version,omitempty"` + Version *VulnerabilityType_Version `protobuf:"bytes,2,opt,name=version,proto3" json:"version,omitempty"` // The path from which we gathered that this package/version is installed. - Path string `protobuf:"bytes,3,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,3,opt,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -235,9 +235,9 @@ func (m *PackageManager_Location) GetPath() string { // e.g. glibc (aka libc6) is distributed by many, at various versions. type PackageManager_Package struct { // The name of the package. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The various channels by which a package is distributed. - Distribution []*PackageManager_Distribution `protobuf:"bytes,10,rep,name=distribution" json:"distribution,omitempty"` + Distribution []*PackageManager_Distribution `protobuf:"bytes,10,rep,name=distribution,proto3" json:"distribution,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -285,10 +285,10 @@ func (m *PackageManager_Package) GetDistribution() []*PackageManager_Distributio // a system. type PackageManager_Installation struct { // Output only. The name of the installed package. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // All of the places within the filesystem versions of this package // have been found. - Location []*PackageManager_Location `protobuf:"bytes,2,rep,name=location" json:"location,omitempty"` + Location []*PackageManager_Location `protobuf:"bytes,2,rep,name=location,proto3" json:"location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/containeranalysis.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/containeranalysis.pb.go index 7b078463e..c74f41e71 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/containeranalysis.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/containeranalysis.pb.go @@ -222,20 +222,20 @@ func (PgpSignedAttestation_ContentType) EnumDescriptor() ([]byte, []int) { type Occurrence struct { // Output only. The name of the `Occurrence` in the form // "projects/{project_id}/occurrences/{OCCURRENCE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The unique URL of the image or the container for which the `Occurrence` // applies. For example, https://gcr.io/project/image@sha256:foo This field // can be used as a filter in list requests. - ResourceUrl string `protobuf:"bytes,2,opt,name=resource_url,json=resourceUrl" json:"resource_url,omitempty"` + ResourceUrl string `protobuf:"bytes,2,opt,name=resource_url,json=resourceUrl,proto3" json:"resource_url,omitempty"` // The resource for which the `Occurrence` applies. - Resource *Resource `protobuf:"bytes,17,opt,name=resource" json:"resource,omitempty"` + Resource *Resource `protobuf:"bytes,17,opt,name=resource,proto3" json:"resource,omitempty"` // An analysis note associated with this image, in the form // "providers/{provider_id}/notes/{NOTE_ID}" // This field can be used as a filter in list requests. - NoteName string `protobuf:"bytes,3,opt,name=note_name,json=noteName" json:"note_name,omitempty"` + NoteName string `protobuf:"bytes,3,opt,name=note_name,json=noteName,proto3" json:"note_name,omitempty"` // Output only. This explicitly denotes which of the `Occurrence` details are // specified. This field can be used as a filter in list requests. - Kind Note_Kind `protobuf:"varint,6,opt,name=kind,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"kind,omitempty"` + Kind Note_Kind `protobuf:"varint,6,opt,name=kind,proto3,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"kind,omitempty"` // Describes the details of the vulnerability `Note` found in this resource. // // Types that are valid to be assigned to Details: @@ -248,11 +248,11 @@ type Occurrence struct { // *Occurrence_Attestation Details isOccurrence_Details `protobuf_oneof:"details"` // A description of actions that can be taken to remedy the `Note` - Remediation string `protobuf:"bytes,5,opt,name=remediation" json:"remediation,omitempty"` + Remediation string `protobuf:"bytes,5,opt,name=remediation,proto3" json:"remediation,omitempty"` // Output only. The time this `Occurrence` was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The time this `Occurrence` was last updated. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -287,25 +287,25 @@ type isOccurrence_Details interface { } type Occurrence_VulnerabilityDetails struct { - VulnerabilityDetails *VulnerabilityType_VulnerabilityDetails `protobuf:"bytes,8,opt,name=vulnerability_details,json=vulnerabilityDetails,oneof"` + VulnerabilityDetails *VulnerabilityType_VulnerabilityDetails `protobuf:"bytes,8,opt,name=vulnerability_details,json=vulnerabilityDetails,proto3,oneof"` } type Occurrence_BuildDetails struct { - BuildDetails *BuildDetails `protobuf:"bytes,7,opt,name=build_details,json=buildDetails,oneof"` + BuildDetails *BuildDetails `protobuf:"bytes,7,opt,name=build_details,json=buildDetails,proto3,oneof"` } type Occurrence_DerivedImage struct { - DerivedImage *DockerImage_Derived `protobuf:"bytes,11,opt,name=derived_image,json=derivedImage,oneof"` + DerivedImage *DockerImage_Derived `protobuf:"bytes,11,opt,name=derived_image,json=derivedImage,proto3,oneof"` } type Occurrence_Installation struct { - Installation *PackageManager_Installation `protobuf:"bytes,12,opt,name=installation,oneof"` + Installation *PackageManager_Installation `protobuf:"bytes,12,opt,name=installation,proto3,oneof"` } type Occurrence_Deployment struct { - Deployment *Deployable_Deployment `protobuf:"bytes,14,opt,name=deployment,oneof"` + Deployment *Deployable_Deployment `protobuf:"bytes,14,opt,name=deployment,proto3,oneof"` } type Occurrence_Discovered struct { - Discovered *Discovery_Discovered `protobuf:"bytes,15,opt,name=discovered,oneof"` + Discovered *Discovery_Discovered `protobuf:"bytes,15,opt,name=discovered,proto3,oneof"` } type Occurrence_Attestation struct { - Attestation *AttestationAuthority_Attestation `protobuf:"bytes,16,opt,name=attestation,oneof"` + Attestation *AttestationAuthority_Attestation `protobuf:"bytes,16,opt,name=attestation,proto3,oneof"` } func (*Occurrence_VulnerabilityDetails) isOccurrence_Details() {} @@ -600,12 +600,12 @@ func _Occurrence_OneofSizer(msg proto.Message) (n int) { // Resource is an entity that can have metadata. E.g., a Docker image. type Resource struct { // The name of the resource. E.g., the name of a Docker image - "Debian". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The unique URI of the resource. E.g., // "https://gcr.io/project/image@sha256:foo" for a Docker image. - Uri string `protobuf:"bytes,2,opt,name=uri" json:"uri,omitempty"` + Uri string `protobuf:"bytes,2,opt,name=uri,proto3" json:"uri,omitempty"` // The hash of the resource content. E.g., the Docker digest. - ContentHash *Hash `protobuf:"bytes,3,opt,name=content_hash,json=contentHash" json:"content_hash,omitempty"` + ContentHash *Hash `protobuf:"bytes,3,opt,name=content_hash,json=contentHash,proto3" json:"content_hash,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -660,14 +660,14 @@ func (m *Resource) GetContentHash() *Hash { type Note struct { // The name of the note in the form // "providers/{provider_id}/notes/{NOTE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A one sentence description of this `Note`. - ShortDescription string `protobuf:"bytes,3,opt,name=short_description,json=shortDescription" json:"short_description,omitempty"` + ShortDescription string `protobuf:"bytes,3,opt,name=short_description,json=shortDescription,proto3" json:"short_description,omitempty"` // A detailed description of this `Note`. - LongDescription string `protobuf:"bytes,4,opt,name=long_description,json=longDescription" json:"long_description,omitempty"` + LongDescription string `protobuf:"bytes,4,opt,name=long_description,json=longDescription,proto3" json:"long_description,omitempty"` // Output only. This explicitly denotes which kind of note is specified. This // field can be used as a filter in list requests. - Kind Note_Kind `protobuf:"varint,9,opt,name=kind,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"kind,omitempty"` + Kind Note_Kind `protobuf:"varint,9,opt,name=kind,proto3,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"kind,omitempty"` // The type of note. // // Types that are valid to be assigned to NoteType: @@ -680,15 +680,15 @@ type Note struct { // *Note_AttestationAuthority NoteType isNote_NoteType `protobuf_oneof:"note_type"` // URLs associated with this note - RelatedUrl []*Note_RelatedUrl `protobuf:"bytes,7,rep,name=related_url,json=relatedUrl" json:"related_url,omitempty"` + RelatedUrl []*Note_RelatedUrl `protobuf:"bytes,7,rep,name=related_url,json=relatedUrl,proto3" json:"related_url,omitempty"` // Time of expiration for this note, null if note does not expire. - ExpirationTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=expiration_time,json=expirationTime" json:"expiration_time,omitempty"` + ExpirationTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=expiration_time,json=expirationTime,proto3" json:"expiration_time,omitempty"` // Output only. The time this note was created. This field can be used as a // filter in list requests. - CreateTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The time this note was last updated. This field can be used as // a filter in list requests. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,12,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,12,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -723,25 +723,25 @@ type isNote_NoteType interface { } type Note_VulnerabilityType struct { - VulnerabilityType *VulnerabilityType `protobuf:"bytes,6,opt,name=vulnerability_type,json=vulnerabilityType,oneof"` + VulnerabilityType *VulnerabilityType `protobuf:"bytes,6,opt,name=vulnerability_type,json=vulnerabilityType,proto3,oneof"` } type Note_BuildType struct { - BuildType *BuildType `protobuf:"bytes,8,opt,name=build_type,json=buildType,oneof"` + BuildType *BuildType `protobuf:"bytes,8,opt,name=build_type,json=buildType,proto3,oneof"` } type Note_BaseImage struct { - BaseImage *DockerImage_Basis `protobuf:"bytes,13,opt,name=base_image,json=baseImage,oneof"` + BaseImage *DockerImage_Basis `protobuf:"bytes,13,opt,name=base_image,json=baseImage,proto3,oneof"` } type Note_Package struct { - Package *PackageManager_Package `protobuf:"bytes,14,opt,name=package,oneof"` + Package *PackageManager_Package `protobuf:"bytes,14,opt,name=package,proto3,oneof"` } type Note_Deployable struct { - Deployable *Deployable `protobuf:"bytes,17,opt,name=deployable,oneof"` + Deployable *Deployable `protobuf:"bytes,17,opt,name=deployable,proto3,oneof"` } type Note_Discovery struct { - Discovery *Discovery `protobuf:"bytes,18,opt,name=discovery,oneof"` + Discovery *Discovery `protobuf:"bytes,18,opt,name=discovery,proto3,oneof"` } type Note_AttestationAuthority struct { - AttestationAuthority *AttestationAuthority `protobuf:"bytes,19,opt,name=attestation_authority,json=attestationAuthority,oneof"` + AttestationAuthority *AttestationAuthority `protobuf:"bytes,19,opt,name=attestation_authority,json=attestationAuthority,proto3,oneof"` } func (*Note_VulnerabilityType) isNote_NoteType() {} @@ -1036,9 +1036,9 @@ func _Note_OneofSizer(msg proto.Message) (n int) { // Metadata for any related URL information type Note_RelatedUrl struct { // Specific URL to associate with the note - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // Label to describe usage of the URL - Label string `protobuf:"bytes,2,opt,name=label" json:"label,omitempty"` + Label string `protobuf:"bytes,2,opt,name=label,proto3" json:"label,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1085,7 +1085,7 @@ func (m *Note_RelatedUrl) GetLabel() string { // An artifact that can be deployed in some runtime. type Deployable struct { // Resource URI for the artifact being deployed. - ResourceUri []string `protobuf:"bytes,1,rep,name=resource_uri,json=resourceUri" json:"resource_uri,omitempty"` + ResourceUri []string `protobuf:"bytes,1,rep,name=resource_uri,json=resourceUri,proto3" json:"resource_uri,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1125,20 +1125,20 @@ func (m *Deployable) GetResourceUri() []string { // The period during which some deployable was active in a runtime. type Deployable_Deployment struct { // Identity of the user that triggered this deployment. - UserEmail string `protobuf:"bytes,1,opt,name=user_email,json=userEmail" json:"user_email,omitempty"` + UserEmail string `protobuf:"bytes,1,opt,name=user_email,json=userEmail,proto3" json:"user_email,omitempty"` // Beginning of the lifetime of this deployment. - DeployTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=deploy_time,json=deployTime" json:"deploy_time,omitempty"` + DeployTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=deploy_time,json=deployTime,proto3" json:"deploy_time,omitempty"` // End of the lifetime of this deployment. - UndeployTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=undeploy_time,json=undeployTime" json:"undeploy_time,omitempty"` + UndeployTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=undeploy_time,json=undeployTime,proto3" json:"undeploy_time,omitempty"` // Configuration used to create this deployment. - Config string `protobuf:"bytes,8,opt,name=config" json:"config,omitempty"` + Config string `protobuf:"bytes,8,opt,name=config,proto3" json:"config,omitempty"` // Address of the runtime element hosting this deployment. - Address string `protobuf:"bytes,5,opt,name=address" json:"address,omitempty"` + Address string `protobuf:"bytes,5,opt,name=address,proto3" json:"address,omitempty"` // Output only. Resource URI for the artifact being deployed taken from the // deployable field with the same name. - ResourceUri []string `protobuf:"bytes,6,rep,name=resource_uri,json=resourceUri" json:"resource_uri,omitempty"` + ResourceUri []string `protobuf:"bytes,6,rep,name=resource_uri,json=resourceUri,proto3" json:"resource_uri,omitempty"` // Platform hosting this deployment. - Platform Deployable_Deployment_Platform `protobuf:"varint,7,opt,name=platform,enum=google.devtools.containeranalysis.v1alpha1.Deployable_Deployment_Platform" json:"platform,omitempty"` + Platform Deployable_Deployment_Platform `protobuf:"varint,7,opt,name=platform,proto3,enum=google.devtools.containeranalysis.v1alpha1.Deployable_Deployment_Platform" json:"platform,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1224,7 +1224,7 @@ func (m *Deployable_Deployment) GetPlatform() Deployable_Deployment_Platform { // note for a resource indicates that analysis hasn't started. type Discovery struct { // The kind of analysis that is handled by this discovery. - AnalysisKind Note_Kind `protobuf:"varint,1,opt,name=analysis_kind,json=analysisKind,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"analysis_kind,omitempty"` + AnalysisKind Note_Kind `protobuf:"varint,1,opt,name=analysis_kind,json=analysisKind,proto3,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"analysis_kind,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1264,13 +1264,13 @@ func (m *Discovery) GetAnalysisKind() Note_Kind { // Provides information about the scan status of a discovered resource. type Discovery_Discovered struct { // Output only. An operation that indicates the status of the current scan. - Operation *longrunning.Operation `protobuf:"bytes,1,opt,name=operation" json:"operation,omitempty"` + Operation *longrunning.Operation `protobuf:"bytes,1,opt,name=operation,proto3" json:"operation,omitempty"` // The status of discovery for the resource. - AnalysisStatus Discovery_Discovered_AnalysisStatus `protobuf:"varint,5,opt,name=analysis_status,json=analysisStatus,enum=google.devtools.containeranalysis.v1alpha1.Discovery_Discovered_AnalysisStatus" json:"analysis_status,omitempty"` + AnalysisStatus Discovery_Discovered_AnalysisStatus `protobuf:"varint,5,opt,name=analysis_status,json=analysisStatus,proto3,enum=google.devtools.containeranalysis.v1alpha1.Discovery_Discovered_AnalysisStatus" json:"analysis_status,omitempty"` // When an error is encountered this will contain a LocalizedMessage under // details to show to the user. The LocalizedMessage output only and // populated by the API. - AnalysisStatusError *status.Status `protobuf:"bytes,6,opt,name=analysis_status_error,json=analysisStatusError" json:"analysis_status_error,omitempty"` + AnalysisStatusError *status.Status `protobuf:"bytes,6,opt,name=analysis_status_error,json=analysisStatusError,proto3" json:"analysis_status_error,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1325,10 +1325,10 @@ func (m *Discovery_Discovered) GetAnalysisStatusError() *status.Status { // the provenance message in linked BuildDetails. type BuildType struct { // Version of the builder which produced this Note. - BuilderVersion string `protobuf:"bytes,1,opt,name=builder_version,json=builderVersion" json:"builder_version,omitempty"` + BuilderVersion string `protobuf:"bytes,1,opt,name=builder_version,json=builderVersion,proto3" json:"builder_version,omitempty"` // Signature of the build in Occurrences pointing to the Note containing this // `BuilderDetails`. - Signature *BuildSignature `protobuf:"bytes,2,opt,name=signature" json:"signature,omitempty"` + Signature *BuildSignature `protobuf:"bytes,2,opt,name=signature,proto3" json:"signature,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1387,17 +1387,17 @@ type BuildSignature struct { // from `BuildDetails` are base64-decoded into a binary representation in // signed.bin. OpenSSL can then verify the signature: // `openssl sha256 -verify public.pem -signature signature.bin signed.bin` - PublicKey string `protobuf:"bytes,1,opt,name=public_key,json=publicKey" json:"public_key,omitempty"` + PublicKey string `protobuf:"bytes,1,opt,name=public_key,json=publicKey,proto3" json:"public_key,omitempty"` // Signature of the related `BuildProvenance`, encoded in a base64 string. - Signature string `protobuf:"bytes,2,opt,name=signature" json:"signature,omitempty"` + Signature string `protobuf:"bytes,2,opt,name=signature,proto3" json:"signature,omitempty"` // An Id for the key used to sign. This could be either an Id for the key // stored in `public_key` (such as the Id or fingerprint for a PGP key, or the // CN for a cert), or a reference to an external key (such as a reference to a // key in Cloud Key Management Service). - KeyId string `protobuf:"bytes,3,opt,name=key_id,json=keyId" json:"key_id,omitempty"` + KeyId string `protobuf:"bytes,3,opt,name=key_id,json=keyId,proto3" json:"key_id,omitempty"` // The type of the key, either stored in `public_key` or referenced in // `key_id` - KeyType BuildSignature_KeyType `protobuf:"varint,4,opt,name=key_type,json=keyType,enum=google.devtools.containeranalysis.v1alpha1.BuildSignature_KeyType" json:"key_type,omitempty"` + KeyType BuildSignature_KeyType `protobuf:"varint,4,opt,name=key_type,json=keyType,proto3,enum=google.devtools.containeranalysis.v1alpha1.BuildSignature_KeyType" json:"key_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1468,12 +1468,12 @@ type PgpSignedAttestation struct { // --output=signature.gpg payload.json` will create the signature content // expected in this field in `signature.gpg` for the `payload.json` // attestation payload. - Signature string `protobuf:"bytes,1,opt,name=signature" json:"signature,omitempty"` + Signature string `protobuf:"bytes,1,opt,name=signature,proto3" json:"signature,omitempty"` // Type (for example schema) of the attestation payload that was signed. // The verifier must ensure that the provided type is one that the verifier // supports, and that the attestation payload is a valid instantiation of that // type (for example by validating a JSON schema). - ContentType PgpSignedAttestation_ContentType `protobuf:"varint,3,opt,name=content_type,json=contentType,enum=google.devtools.containeranalysis.v1alpha1.PgpSignedAttestation_ContentType" json:"content_type,omitempty"` + ContentType PgpSignedAttestation_ContentType `protobuf:"varint,3,opt,name=content_type,json=contentType,proto3,enum=google.devtools.containeranalysis.v1alpha1.PgpSignedAttestation_ContentType" json:"content_type,omitempty"` // This field is used by verifiers to select the public key used to validate // the signature. Note that the policy of the verifier ultimately determines // which public keys verify a signature based on the context of the @@ -1520,7 +1520,7 @@ type isPgpSignedAttestation_KeyId interface { } type PgpSignedAttestation_PgpKeyId struct { - PgpKeyId string `protobuf:"bytes,2,opt,name=pgp_key_id,json=pgpKeyId,oneof"` + PgpKeyId string `protobuf:"bytes,2,opt,name=pgp_key_id,json=pgpKeyId,proto3,oneof"` } func (*PgpSignedAttestation_PgpKeyId) isPgpSignedAttestation_KeyId() {} @@ -1613,7 +1613,7 @@ func _PgpSignedAttestation_OneofSizer(msg proto.Message) (n int) { // single point of lookup to find all attached Attestation Occurrences, even if // they don't all live in the same project. type AttestationAuthority struct { - Hint *AttestationAuthority_AttestationAuthorityHint `protobuf:"bytes,1,opt,name=hint" json:"hint,omitempty"` + Hint *AttestationAuthority_AttestationAuthorityHint `protobuf:"bytes,1,opt,name=hint,proto3" json:"hint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1659,7 +1659,7 @@ func (m *AttestationAuthority) GetHint() *AttestationAuthority_AttestationAuthor // looking up Attestations to verify. type AttestationAuthority_AttestationAuthorityHint struct { // The human readable name of this Attestation Authority, for example "qa". - HumanReadableName string `protobuf:"bytes,1,opt,name=human_readable_name,json=humanReadableName" json:"human_readable_name,omitempty"` + HumanReadableName string `protobuf:"bytes,1,opt,name=human_readable_name,json=humanReadableName,proto3" json:"human_readable_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1750,7 +1750,7 @@ type isAttestationAuthority_Attestation_Signature interface { } type AttestationAuthority_Attestation_PgpSignedAttestation struct { - PgpSignedAttestation *PgpSignedAttestation `protobuf:"bytes,1,opt,name=pgp_signed_attestation,json=pgpSignedAttestation,oneof"` + PgpSignedAttestation *PgpSignedAttestation `protobuf:"bytes,1,opt,name=pgp_signed_attestation,json=pgpSignedAttestation,proto3,oneof"` } func (*AttestationAuthority_Attestation_PgpSignedAttestation) isAttestationAuthority_Attestation_Signature() { @@ -1828,7 +1828,7 @@ func _AttestationAuthority_Attestation_OneofSizer(msg proto.Message) (n int) { // Message encapsulating build provenance details. type BuildDetails struct { // The actual provenance - Provenance *BuildProvenance `protobuf:"bytes,1,opt,name=provenance" json:"provenance,omitempty"` + Provenance *BuildProvenance `protobuf:"bytes,1,opt,name=provenance,proto3" json:"provenance,omitempty"` // Serialized JSON representation of the provenance, used in generating the // `BuildSignature` in the corresponding Result. After verifying the // signature, `provenance_bytes` can be unmarshalled and compared to the @@ -1840,7 +1840,7 @@ type BuildDetails struct { // The serialized form is captured both to avoid ambiguity in how the // provenance is marshalled to json as well to prevent incompatibilities with // future changes. - ProvenanceBytes string `protobuf:"bytes,2,opt,name=provenance_bytes,json=provenanceBytes" json:"provenance_bytes,omitempty"` + ProvenanceBytes string `protobuf:"bytes,2,opt,name=provenance_bytes,json=provenanceBytes,proto3" json:"provenance_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1888,11 +1888,11 @@ func (m *BuildDetails) GetProvenanceBytes() string { type ScanConfig struct { // Output only. The name of the ScanConfig in the form // “projects/{project_id}/ScanConfigs/{ScanConfig_id}". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Output only. A human-readable description of what the `ScanConfig` does. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Indicates whether the Scan is enabled. - Enabled bool `protobuf:"varint,3,opt,name=enabled" json:"enabled,omitempty"` + Enabled bool `protobuf:"varint,3,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1947,7 +1947,7 @@ func (m *ScanConfig) GetEnabled() bool { type GetOccurrenceRequest struct { // The name of the occurrence of the form // "projects/{project_id}/occurrences/{OCCURRENCE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1989,17 +1989,17 @@ type ListOccurrencesRequest struct { // The name field contains the project Id. For example: // "projects/{project_id} // @Deprecated - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // This contains the project Id for example: projects/{project_id}. - Parent string `protobuf:"bytes,5,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,5,opt,name=parent,proto3" json:"parent,omitempty"` // The filter expression. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Number of occurrences to return in the list. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token to provide to skip to a particular spot in the list. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The kind of occurrences to filter on. - Kind Note_Kind `protobuf:"varint,6,opt,name=kind,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"kind,omitempty"` + Kind Note_Kind `protobuf:"varint,6,opt,name=kind,proto3,enum=google.devtools.containeranalysis.v1alpha1.Note_Kind" json:"kind,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2074,11 +2074,11 @@ func (m *ListOccurrencesRequest) GetKind() Note_Kind { // Response including listed active occurrences. type ListOccurrencesResponse struct { // The occurrences requested. - Occurrences []*Occurrence `protobuf:"bytes,1,rep,name=occurrences" json:"occurrences,omitempty"` + Occurrences []*Occurrence `protobuf:"bytes,1,rep,name=occurrences,proto3" json:"occurrences,omitempty"` // The next pagination token in the list response. It should be used as // `page_token` for the following request. An empty value means no more // results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2126,7 +2126,7 @@ func (m *ListOccurrencesResponse) GetNextPageToken() string { type DeleteOccurrenceRequest struct { // The name of the occurrence in the form of // "projects/{project_id}/occurrences/{OCCURRENCE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2167,11 +2167,11 @@ func (m *DeleteOccurrenceRequest) GetName() string { type CreateOccurrenceRequest struct { // The name of the project. Should be of the form "projects/{project_id}". // @Deprecated - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // This field contains the project Id for example: "projects/{project_id}" - Parent string `protobuf:"bytes,3,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,3,opt,name=parent,proto3" json:"parent,omitempty"` // The occurrence to be inserted - Occurrence *Occurrence `protobuf:"bytes,2,opt,name=occurrence" json:"occurrence,omitempty"` + Occurrence *Occurrence `protobuf:"bytes,2,opt,name=occurrence,proto3" json:"occurrence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2226,11 +2226,11 @@ func (m *CreateOccurrenceRequest) GetOccurrence() *Occurrence { type UpdateOccurrenceRequest struct { // The name of the occurrence. // Should be of the form "projects/{project_id}/occurrences/{OCCURRENCE_ID}". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The updated occurrence. - Occurrence *Occurrence `protobuf:"bytes,2,opt,name=occurrence" json:"occurrence,omitempty"` + Occurrence *Occurrence `protobuf:"bytes,2,opt,name=occurrence,proto3" json:"occurrence,omitempty"` // The fields to update. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2285,7 +2285,7 @@ func (m *UpdateOccurrenceRequest) GetUpdateMask() *field_mask.FieldMask { type GetNoteRequest struct { // The name of the note in the form of // "providers/{provider_id}/notes/{NOTE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2326,7 +2326,7 @@ func (m *GetNoteRequest) GetName() string { type GetOccurrenceNoteRequest struct { // The name of the occurrence in the form // "projects/{project_id}/occurrences/{OCCURRENCE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2368,15 +2368,15 @@ type ListNotesRequest struct { // The name field will contain the project Id for example: // "providers/{provider_id} // @Deprecated - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // This field contains the project Id for example: "projects/{PROJECT_ID}". - Parent string `protobuf:"bytes,5,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,5,opt,name=parent,proto3" json:"parent,omitempty"` // The filter expression. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Number of notes to return in the list. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token to provide to skip to a particular spot in the list. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2444,10 +2444,10 @@ func (m *ListNotesRequest) GetPageToken() string { // Response including listed notes. type ListNotesResponse struct { // The occurrences requested - Notes []*Note `protobuf:"bytes,1,rep,name=notes" json:"notes,omitempty"` + Notes []*Note `protobuf:"bytes,1,rep,name=notes,proto3" json:"notes,omitempty"` // The next pagination token in the list response. It should be used as // page_token for the following request. An empty value means no more result. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2495,7 +2495,7 @@ func (m *ListNotesResponse) GetNextPageToken() string { type DeleteNoteRequest struct { // The name of the note in the form of // "providers/{provider_id}/notes/{NOTE_ID}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2537,14 +2537,14 @@ type CreateNoteRequest struct { // The name of the project. // Should be of the form "providers/{provider_id}". // @Deprecated - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // This field contains the project Id for example: // "projects/{project_id} - Parent string `protobuf:"bytes,4,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,4,opt,name=parent,proto3" json:"parent,omitempty"` // The ID to use for this note. - NoteId string `protobuf:"bytes,2,opt,name=note_id,json=noteId" json:"note_id,omitempty"` + NoteId string `protobuf:"bytes,2,opt,name=note_id,json=noteId,proto3" json:"note_id,omitempty"` // The Note to be inserted - Note *Note `protobuf:"bytes,3,opt,name=note" json:"note,omitempty"` + Note *Note `protobuf:"bytes,3,opt,name=note,proto3" json:"note,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2606,11 +2606,11 @@ func (m *CreateNoteRequest) GetNote() *Note { type UpdateNoteRequest struct { // The name of the note. // Should be of the form "projects/{provider_id}/notes/{note_id}". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The updated note. - Note *Note `protobuf:"bytes,2,opt,name=note" json:"note,omitempty"` + Note *Note `protobuf:"bytes,2,opt,name=note,proto3" json:"note,omitempty"` // The fields to update. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2665,13 +2665,13 @@ func (m *UpdateNoteRequest) GetUpdateMask() *field_mask.FieldMask { type ListNoteOccurrencesRequest struct { // The name field will contain the note name for example: // "provider/{provider_id}/notes/{note_id}" - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The filter expression. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Number of notes to return in the list. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token to provide to skip to a particular spot in the list. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2732,9 +2732,9 @@ func (m *ListNoteOccurrencesRequest) GetPageToken() string { // Response including listed occurrences for a note. type ListNoteOccurrencesResponse struct { // The occurrences attached to the specified note. - Occurrences []*Occurrence `protobuf:"bytes,1,rep,name=occurrences" json:"occurrences,omitempty"` + Occurrences []*Occurrence `protobuf:"bytes,1,rep,name=occurrences,proto3" json:"occurrences,omitempty"` // Token to receive the next page of notes. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2781,11 +2781,11 @@ func (m *ListNoteOccurrencesResponse) GetNextPageToken() string { // Request for creating an operation type CreateOperationRequest struct { // The project Id that this operation should be created under. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The ID to use for this operation. - OperationId string `protobuf:"bytes,2,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,2,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` // The operation to create. - Operation *longrunning.Operation `protobuf:"bytes,3,opt,name=operation" json:"operation,omitempty"` + Operation *longrunning.Operation `protobuf:"bytes,3,opt,name=operation,proto3" json:"operation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2840,10 +2840,10 @@ func (m *CreateOperationRequest) GetOperation() *longrunning.Operation { type UpdateOperationRequest struct { // The name of the Operation. // Should be of the form "projects/{provider_id}/operations/{operation_id}". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The operation to create. - Operation *longrunning.Operation `protobuf:"bytes,3,opt,name=operation" json:"operation,omitempty"` - UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + Operation *longrunning.Operation `protobuf:"bytes,3,opt,name=operation,proto3" json:"operation,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2898,9 +2898,9 @@ func (m *UpdateOperationRequest) GetUpdateMask() *field_mask.FieldMask { // that created by Container Analysis Providers type OperationMetadata struct { // Output only. The time this operation was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The time that this operation was marked completed or failed. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2947,9 +2947,9 @@ func (m *OperationMetadata) GetEndTime() *timestamp.Timestamp { // Request to get the vulnz summary for some set of vulnerability Occurrences. type GetVulnzOccurrencesSummaryRequest struct { // This contains the project Id for example: projects/{project_id} - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The filter expression. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2998,7 +2998,7 @@ func (m *GetVulnzOccurrencesSummaryRequest) GetFilter() string { // like this. type GetVulnzOccurrencesSummaryResponse struct { // A map of how many occurrences were found for each severity. - Counts []*GetVulnzOccurrencesSummaryResponse_SeverityCount `protobuf:"bytes,1,rep,name=counts" json:"counts,omitempty"` + Counts []*GetVulnzOccurrencesSummaryResponse_SeverityCount `protobuf:"bytes,1,rep,name=counts,proto3" json:"counts,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3038,9 +3038,9 @@ func (m *GetVulnzOccurrencesSummaryResponse) GetCounts() []*GetVulnzOccurrencesS // The number of occurrences created for a specific severity. type GetVulnzOccurrencesSummaryResponse_SeverityCount struct { // The severity of the occurrences. - Severity VulnerabilityType_Severity `protobuf:"varint,1,opt,name=severity,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Severity" json:"severity,omitempty"` + Severity VulnerabilityType_Severity `protobuf:"varint,1,opt,name=severity,proto3,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Severity" json:"severity,omitempty"` // The number of occurrences with the severity. - Count int64 `protobuf:"varint,2,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,2,opt,name=count,proto3" json:"count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3093,7 +3093,7 @@ type GetScanConfigRequest struct { // The name of the ScanConfig in the form // projects/{project_id}/scan_configs/{ScanConfig_id} // instead. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3134,13 +3134,13 @@ func (m *GetScanConfigRequest) GetName() string { type ListScanConfigsRequest struct { // This containers the project Id i.e.: projects/{project_id} // instead. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The filter expression. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // The number of items to return. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The page token to use for the next request. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3201,9 +3201,9 @@ func (m *ListScanConfigsRequest) GetPageToken() string { // A list of ScanConfigs for the project. type ListScanConfigsResponse struct { // The set of scan configs - ScanConfigs []*ScanConfig `protobuf:"bytes,1,rep,name=scan_configs,json=scanConfigs" json:"scan_configs,omitempty"` + ScanConfigs []*ScanConfig `protobuf:"bytes,1,rep,name=scan_configs,json=scanConfigs,proto3" json:"scan_configs,omitempty"` // A page token to pass in order to get more scans. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3252,10 +3252,10 @@ type UpdateScanConfigRequest struct { // The scan config to update of the form // projects/{project_id}/scan_configs/{ScanConfig_id} // instead. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The new scan configuration - ScanConfig *ScanConfig `protobuf:"bytes,2,opt,name=scan_config,json=scanConfig" json:"scan_config,omitempty"` - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + ScanConfig *ScanConfig `protobuf:"bytes,2,opt,name=scan_config,json=scanConfig,proto3" json:"scan_config,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3642,8 +3642,7 @@ func (c *containerAnalysisClient) UpdateScanConfig(ctx context.Context, in *Upda return out, nil } -// Server API for ContainerAnalysis service - +// ContainerAnalysisServer is the server API for ContainerAnalysis service. type ContainerAnalysisServer interface { // Returns the requested `Occurrence`. GetOccurrence(context.Context, *GetOccurrenceRequest) (*Occurrence, error) diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/image_basis.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/image_basis.pb.go index 341d42280..fa3ac68e1 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/image_basis.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/image_basis.pb.go @@ -144,9 +144,9 @@ var xxx_messageInfo_DockerImage proto.InternalMessageInfo // Layer holds metadata specific to a layer of a Docker image. type DockerImage_Layer struct { // The recovered Dockerfile directive used to construct this layer. - Directive DockerImage_Layer_Directive `protobuf:"varint,1,opt,name=directive,enum=google.devtools.containeranalysis.v1alpha1.DockerImage_Layer_Directive" json:"directive,omitempty"` + Directive DockerImage_Layer_Directive `protobuf:"varint,1,opt,name=directive,proto3,enum=google.devtools.containeranalysis.v1alpha1.DockerImage_Layer_Directive" json:"directive,omitempty"` // The recovered arguments to the Dockerfile directive. - Arguments string `protobuf:"bytes,2,opt,name=arguments" json:"arguments,omitempty"` + Arguments string `protobuf:"bytes,2,opt,name=arguments,proto3" json:"arguments,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -195,15 +195,15 @@ type DockerImage_Fingerprint struct { // The layer-id of the final layer in the Docker image's v1 // representation. // This field can be used as a filter in list requests. - V1Name string `protobuf:"bytes,1,opt,name=v1_name,json=v1Name" json:"v1_name,omitempty"` + V1Name string `protobuf:"bytes,1,opt,name=v1_name,json=v1Name,proto3" json:"v1_name,omitempty"` // The ordered list of v2 blobs that represent a given image. - V2Blob []string `protobuf:"bytes,2,rep,name=v2_blob,json=v2Blob" json:"v2_blob,omitempty"` + V2Blob []string `protobuf:"bytes,2,rep,name=v2_blob,json=v2Blob,proto3" json:"v2_blob,omitempty"` // Output only. The name of the image's v2 blobs computed via: // [bottom] := v2_blob[bottom] // [N] := sha256(v2_blob[N] + " " + v2_name[N+1]) // Only the name of the final blob is kept. // This field can be used as a filter in list requests. - V2Name string `protobuf:"bytes,3,opt,name=v2_name,json=v2Name" json:"v2_name,omitempty"` + V2Name string `protobuf:"bytes,3,opt,name=v2_name,json=v2Name,proto3" json:"v2_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,9 +262,9 @@ func (m *DockerImage_Fingerprint) GetV2Name() string { type DockerImage_Basis struct { // The resource_url for the resource representing the basis of // associated occurrence images. - ResourceUrl string `protobuf:"bytes,1,opt,name=resource_url,json=resourceUrl" json:"resource_url,omitempty"` + ResourceUrl string `protobuf:"bytes,1,opt,name=resource_url,json=resourceUrl,proto3" json:"resource_url,omitempty"` // The fingerprint of the base image. - Fingerprint *DockerImage_Fingerprint `protobuf:"bytes,2,opt,name=fingerprint" json:"fingerprint,omitempty"` + Fingerprint *DockerImage_Fingerprint `protobuf:"bytes,2,opt,name=fingerprint,proto3" json:"fingerprint,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -313,17 +313,17 @@ func (m *DockerImage_Basis) GetFingerprint() *DockerImage_Fingerprint { // with FROM . type DockerImage_Derived struct { // The fingerprint of the derived image. - Fingerprint *DockerImage_Fingerprint `protobuf:"bytes,1,opt,name=fingerprint" json:"fingerprint,omitempty"` + Fingerprint *DockerImage_Fingerprint `protobuf:"bytes,1,opt,name=fingerprint,proto3" json:"fingerprint,omitempty"` // Output only. The number of layers by which this image differs from the // associated image basis. - Distance uint32 `protobuf:"varint,2,opt,name=distance" json:"distance,omitempty"` + Distance uint32 `protobuf:"varint,2,opt,name=distance,proto3" json:"distance,omitempty"` // This contains layer-specific metadata, if populated it has length // "distance" and is ordered with [distance] being the layer immediately // following the base image and [1] being the final layer. - LayerInfo []*DockerImage_Layer `protobuf:"bytes,3,rep,name=layer_info,json=layerInfo" json:"layer_info,omitempty"` + LayerInfo []*DockerImage_Layer `protobuf:"bytes,3,rep,name=layer_info,json=layerInfo,proto3" json:"layer_info,omitempty"` // Output only. This contains the base image URL for the derived image // occurrence. - BaseResourceUrl string `protobuf:"bytes,4,opt,name=base_resource_url,json=baseResourceUrl" json:"base_resource_url,omitempty"` + BaseResourceUrl string `protobuf:"bytes,4,opt,name=base_resource_url,json=baseResourceUrl,proto3" json:"base_resource_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/package_vulnerability.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/package_vulnerability.pb.go index 53a07445c..c679e652c 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/package_vulnerability.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/package_vulnerability.pb.go @@ -97,13 +97,13 @@ func (VulnerabilityType_Version_VersionKind) EnumDescriptor() ([]byte, []int) { // VulnerabilityType provides metadata about a security vulnerability. type VulnerabilityType struct { // The CVSS score for this Vulnerability. - CvssScore float32 `protobuf:"fixed32,2,opt,name=cvss_score,json=cvssScore" json:"cvss_score,omitempty"` + CvssScore float32 `protobuf:"fixed32,2,opt,name=cvss_score,json=cvssScore,proto3" json:"cvss_score,omitempty"` // Note provider assigned impact of the vulnerability - Severity VulnerabilityType_Severity `protobuf:"varint,3,opt,name=severity,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Severity" json:"severity,omitempty"` + Severity VulnerabilityType_Severity `protobuf:"varint,3,opt,name=severity,proto3,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Severity" json:"severity,omitempty"` // All information about the package to specifically identify this // vulnerability. One entry per (version range and cpe_uri) the // package vulnerability has manifested in. - Details []*VulnerabilityType_Detail `protobuf:"bytes,4,rep,name=details" json:"details,omitempty"` + Details []*VulnerabilityType_Detail `protobuf:"bytes,4,rep,name=details,proto3" json:"details,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -161,14 +161,14 @@ func (m *VulnerabilityType) GetDetails() []*VulnerabilityType_Detail { // http://blog.jasonantman.com/2014/07/how-yum-and-rpm-compare-versions/ type VulnerabilityType_Version struct { // Used to correct mistakes in the version numbering scheme. - Epoch int32 `protobuf:"varint,1,opt,name=epoch" json:"epoch,omitempty"` + Epoch int32 `protobuf:"varint,1,opt,name=epoch,proto3" json:"epoch,omitempty"` // The main part of the version name. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // The iteration of the package build from the above version. - Revision string `protobuf:"bytes,3,opt,name=revision" json:"revision,omitempty"` + Revision string `protobuf:"bytes,3,opt,name=revision,proto3" json:"revision,omitempty"` // Distinguish between sentinel MIN/MAX versions and normal versions. // If kind is not NORMAL, then the other fields are ignored. - Kind VulnerabilityType_Version_VersionKind `protobuf:"varint,5,opt,name=kind,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Version_VersionKind" json:"kind,omitempty"` + Kind VulnerabilityType_Version_VersionKind `protobuf:"varint,5,opt,name=kind,proto3,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Version_VersionKind" json:"kind,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -234,27 +234,27 @@ type VulnerabilityType_Detail struct { // which the vulnerability manifests. Examples include distro or storage // location for vulnerable jar. // This field can be used as a filter in list requests. - CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri" json:"cpe_uri,omitempty"` + CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri,proto3" json:"cpe_uri,omitempty"` // The name of the package where the vulnerability was found. // This field can be used as a filter in list requests. - Package string `protobuf:"bytes,8,opt,name=package" json:"package,omitempty"` + Package string `protobuf:"bytes,8,opt,name=package,proto3" json:"package,omitempty"` // The min version of the package in which the vulnerability exists. - MinAffectedVersion *VulnerabilityType_Version `protobuf:"bytes,6,opt,name=min_affected_version,json=minAffectedVersion" json:"min_affected_version,omitempty"` + MinAffectedVersion *VulnerabilityType_Version `protobuf:"bytes,6,opt,name=min_affected_version,json=minAffectedVersion,proto3" json:"min_affected_version,omitempty"` // The max version of the package in which the vulnerability exists. // This field can be used as a filter in list requests. - MaxAffectedVersion *VulnerabilityType_Version `protobuf:"bytes,7,opt,name=max_affected_version,json=maxAffectedVersion" json:"max_affected_version,omitempty"` + MaxAffectedVersion *VulnerabilityType_Version `protobuf:"bytes,7,opt,name=max_affected_version,json=maxAffectedVersion,proto3" json:"max_affected_version,omitempty"` // The severity (eg: distro assigned severity) for this vulnerability. - SeverityName string `protobuf:"bytes,4,opt,name=severity_name,json=severityName" json:"severity_name,omitempty"` + SeverityName string `protobuf:"bytes,4,opt,name=severity_name,json=severityName,proto3" json:"severity_name,omitempty"` // A vendor-specific description of this note. - Description string `protobuf:"bytes,9,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,9,opt,name=description,proto3" json:"description,omitempty"` // The fix for this specific package version. - FixedLocation *VulnerabilityType_VulnerabilityLocation `protobuf:"bytes,5,opt,name=fixed_location,json=fixedLocation" json:"fixed_location,omitempty"` + FixedLocation *VulnerabilityType_VulnerabilityLocation `protobuf:"bytes,5,opt,name=fixed_location,json=fixedLocation,proto3" json:"fixed_location,omitempty"` // The type of package; whether native or non native(ruby gems, // node.js packages etc) - PackageType string `protobuf:"bytes,10,opt,name=package_type,json=packageType" json:"package_type,omitempty"` + PackageType string `protobuf:"bytes,10,opt,name=package_type,json=packageType,proto3" json:"package_type,omitempty"` // Whether this Detail is obsolete. Occurrences are expected not to point to // obsolete details. - IsObsolete bool `protobuf:"varint,11,opt,name=is_obsolete,json=isObsolete" json:"is_obsolete,omitempty"` + IsObsolete bool `protobuf:"varint,11,opt,name=is_obsolete,json=isObsolete,proto3" json:"is_obsolete,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -352,16 +352,16 @@ func (m *VulnerabilityType_Detail) GetIsObsolete() bool { type VulnerabilityType_VulnerabilityDetails struct { // The type of package; whether native or non native(ruby gems, // node.js packages etc) - Type string `protobuf:"bytes,3,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,3,opt,name=type,proto3" json:"type,omitempty"` // Output only. The note provider assigned Severity of the vulnerability. - Severity VulnerabilityType_Severity `protobuf:"varint,4,opt,name=severity,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Severity" json:"severity,omitempty"` + Severity VulnerabilityType_Severity `protobuf:"varint,4,opt,name=severity,proto3,enum=google.devtools.containeranalysis.v1alpha1.VulnerabilityType_Severity" json:"severity,omitempty"` // Output only. The CVSS score of this vulnerability. CVSS score is on a // scale of 0-10 where 0 indicates low severity and 10 indicates high // severity. - CvssScore float32 `protobuf:"fixed32,5,opt,name=cvss_score,json=cvssScore" json:"cvss_score,omitempty"` + CvssScore float32 `protobuf:"fixed32,5,opt,name=cvss_score,json=cvssScore,proto3" json:"cvss_score,omitempty"` // The set of affected locations and their fixes (if available) within // the associated resource. - PackageIssue []*VulnerabilityType_PackageIssue `protobuf:"bytes,6,rep,name=package_issue,json=packageIssue" json:"package_issue,omitempty"` + PackageIssue []*VulnerabilityType_PackageIssue `protobuf:"bytes,6,rep,name=package_issue,json=packageIssue,proto3" json:"package_issue,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -425,11 +425,11 @@ func (m *VulnerabilityType_VulnerabilityDetails) GetPackageIssue() []*Vulnerabil // associated fix (if one is available). type VulnerabilityType_PackageIssue struct { // The location of the vulnerability. - AffectedLocation *VulnerabilityType_VulnerabilityLocation `protobuf:"bytes,1,opt,name=affected_location,json=affectedLocation" json:"affected_location,omitempty"` + AffectedLocation *VulnerabilityType_VulnerabilityLocation `protobuf:"bytes,1,opt,name=affected_location,json=affectedLocation,proto3" json:"affected_location,omitempty"` // The location of the available fix for vulnerability. - FixedLocation *VulnerabilityType_VulnerabilityLocation `protobuf:"bytes,2,opt,name=fixed_location,json=fixedLocation" json:"fixed_location,omitempty"` + FixedLocation *VulnerabilityType_VulnerabilityLocation `protobuf:"bytes,2,opt,name=fixed_location,json=fixedLocation,proto3" json:"fixed_location,omitempty"` // The severity (eg: distro assigned severity) for this vulnerability. - SeverityName string `protobuf:"bytes,3,opt,name=severity_name,json=severityName" json:"severity_name,omitempty"` + SeverityName string `protobuf:"bytes,3,opt,name=severity_name,json=severityName,proto3" json:"severity_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -485,12 +485,12 @@ type VulnerabilityType_VulnerabilityLocation struct { // The cpe_uri in [cpe format] (https://cpe.mitre.org/specification/) // format. Examples include distro or storage location for vulnerable jar. // This field can be used as a filter in list requests. - CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri" json:"cpe_uri,omitempty"` + CpeUri string `protobuf:"bytes,1,opt,name=cpe_uri,json=cpeUri,proto3" json:"cpe_uri,omitempty"` // The package being described. - Package string `protobuf:"bytes,2,opt,name=package" json:"package,omitempty"` + Package string `protobuf:"bytes,2,opt,name=package,proto3" json:"package,omitempty"` // The version of the package being described. // This field can be used as a filter in list requests. - Version *VulnerabilityType_Version `protobuf:"bytes,4,opt,name=version" json:"version,omitempty"` + Version *VulnerabilityType_Version `protobuf:"bytes,4,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/provenance.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/provenance.pb.go index 6f2ec83af..ab27d49c5 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/provenance.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/provenance.pb.go @@ -50,34 +50,34 @@ func (Hash_HashType) EnumDescriptor() ([]byte, []int) { // details about the build from source to completion. type BuildProvenance struct { // Unique identifier of the build. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // ID of the project. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Commands requested by the build. - Commands []*Command `protobuf:"bytes,5,rep,name=commands" json:"commands,omitempty"` + Commands []*Command `protobuf:"bytes,5,rep,name=commands,proto3" json:"commands,omitempty"` // Output of the build. - BuiltArtifacts []*Artifact `protobuf:"bytes,6,rep,name=built_artifacts,json=builtArtifacts" json:"built_artifacts,omitempty"` + BuiltArtifacts []*Artifact `protobuf:"bytes,6,rep,name=built_artifacts,json=builtArtifacts,proto3" json:"built_artifacts,omitempty"` // Time at which the build was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time at which execution of the build was started. - StartTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time at which execution of the build was finished. - FinishTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=finish_time,json=finishTime" json:"finish_time,omitempty"` + FinishTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=finish_time,json=finishTime,proto3" json:"finish_time,omitempty"` // E-mail address of the user who initiated this build. Note that this was the // user's e-mail address at the time the build was initiated; this address may // not represent the same end-user for all time. - Creator string `protobuf:"bytes,11,opt,name=creator" json:"creator,omitempty"` + Creator string `protobuf:"bytes,11,opt,name=creator,proto3" json:"creator,omitempty"` // Google Cloud Storage bucket where logs were written. - LogsBucket string `protobuf:"bytes,13,opt,name=logs_bucket,json=logsBucket" json:"logs_bucket,omitempty"` + LogsBucket string `protobuf:"bytes,13,opt,name=logs_bucket,json=logsBucket,proto3" json:"logs_bucket,omitempty"` // Details of the Source input to the build. - SourceProvenance *Source `protobuf:"bytes,14,opt,name=source_provenance,json=sourceProvenance" json:"source_provenance,omitempty"` + SourceProvenance *Source `protobuf:"bytes,14,opt,name=source_provenance,json=sourceProvenance,proto3" json:"source_provenance,omitempty"` // Trigger identifier if the build was triggered automatically; empty if not. - TriggerId string `protobuf:"bytes,15,opt,name=trigger_id,json=triggerId" json:"trigger_id,omitempty"` + TriggerId string `protobuf:"bytes,15,opt,name=trigger_id,json=triggerId,proto3" json:"trigger_id,omitempty"` // Special options applied to this build. This is a catch-all field where // build providers can enter any desired additional details. - BuildOptions map[string]string `protobuf:"bytes,16,rep,name=build_options,json=buildOptions" json:"build_options,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + BuildOptions map[string]string `protobuf:"bytes,16,rep,name=build_options,json=buildOptions,proto3" json:"build_options,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Version string of the builder at the time this build was executed. - BuilderVersion string `protobuf:"bytes,17,opt,name=builder_version,json=builderVersion" json:"builder_version,omitempty"` + BuilderVersion string `protobuf:"bytes,17,opt,name=builder_version,json=builderVersion,proto3" json:"builder_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -208,7 +208,7 @@ type Source struct { Source isSource_Source `protobuf_oneof:"source"` // If provided, the input binary artifacts for the build came from this // location. - ArtifactStorageSource *StorageSource `protobuf:"bytes,4,opt,name=artifact_storage_source,json=artifactStorageSource" json:"artifact_storage_source,omitempty"` + ArtifactStorageSource *StorageSource `protobuf:"bytes,4,opt,name=artifact_storage_source,json=artifactStorageSource,proto3" json:"artifact_storage_source,omitempty"` // Hash(es) of the build source, which can be used to verify that the original // source integrity was maintained in the build. // @@ -217,14 +217,14 @@ type Source struct { // // If the build source came in a single package such as a gzipped tarfile // (.tar.gz), the FileHash will be for the single path to that file. - FileHashes map[string]*FileHashes `protobuf:"bytes,3,rep,name=file_hashes,json=fileHashes" json:"file_hashes,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + FileHashes map[string]*FileHashes `protobuf:"bytes,3,rep,name=file_hashes,json=fileHashes,proto3" json:"file_hashes,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // If provided, the source code used for the build came from this location. - Context *SourceContext `protobuf:"bytes,7,opt,name=context" json:"context,omitempty"` + Context *SourceContext `protobuf:"bytes,7,opt,name=context,proto3" json:"context,omitempty"` // If provided, some of the source code used for the build may be found in // these locations, in the case where the source repository had multiple // remotes or submodules. This list will not include the context specified in // the context field. - AdditionalContexts []*SourceContext `protobuf:"bytes,8,rep,name=additional_contexts,json=additionalContexts" json:"additional_contexts,omitempty"` + AdditionalContexts []*SourceContext `protobuf:"bytes,8,rep,name=additional_contexts,json=additionalContexts,proto3" json:"additional_contexts,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -259,10 +259,10 @@ type isSource_Source interface { } type Source_StorageSource struct { - StorageSource *StorageSource `protobuf:"bytes,1,opt,name=storage_source,json=storageSource,oneof"` + StorageSource *StorageSource `protobuf:"bytes,1,opt,name=storage_source,json=storageSource,proto3,oneof"` } type Source_RepoSource struct { - RepoSource *RepoSource `protobuf:"bytes,2,opt,name=repo_source,json=repoSource,oneof"` + RepoSource *RepoSource `protobuf:"bytes,2,opt,name=repo_source,json=repoSource,proto3,oneof"` } func (*Source_StorageSource) isSource_Source() {} @@ -395,7 +395,7 @@ func _Source_OneofSizer(msg proto.Message) (n int) { // messages to verify integrity of source input to the build. type FileHashes struct { // Collection of file hashes. - FileHash []*Hash `protobuf:"bytes,1,rep,name=file_hash,json=fileHash" json:"file_hash,omitempty"` + FileHash []*Hash `protobuf:"bytes,1,rep,name=file_hash,json=fileHash,proto3" json:"file_hash,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -435,7 +435,7 @@ func (m *FileHashes) GetFileHash() []*Hash { // Container message for hash values. type Hash struct { // The type of hash that was performed. - Type Hash_HashType `protobuf:"varint,1,opt,name=type,enum=google.devtools.containeranalysis.v1alpha1.Hash_HashType" json:"type,omitempty"` + Type Hash_HashType `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.containeranalysis.v1alpha1.Hash_HashType" json:"type,omitempty"` // The hash value. Value []byte `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -487,11 +487,11 @@ type StorageSource struct { // Google Cloud Storage bucket containing source (see [Bucket Name // Requirements] // (https://cloud.google.com/storage/docs/bucket-naming#requirements)). - Bucket string `protobuf:"bytes,1,opt,name=bucket" json:"bucket,omitempty"` + Bucket string `protobuf:"bytes,1,opt,name=bucket,proto3" json:"bucket,omitempty"` // Google Cloud Storage object containing source. - Object string `protobuf:"bytes,2,opt,name=object" json:"object,omitempty"` + Object string `protobuf:"bytes,2,opt,name=object,proto3" json:"object,omitempty"` // Google Cloud Storage generation for the object. - Generation int64 `protobuf:"varint,3,opt,name=generation" json:"generation,omitempty"` + Generation int64 `protobuf:"varint,3,opt,name=generation,proto3" json:"generation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -546,9 +546,9 @@ func (m *StorageSource) GetGeneration() int64 { // Repository. type RepoSource struct { // ID of the project that owns the repo. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Name of the repo. - RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName" json:"repo_name,omitempty"` + RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName,proto3" json:"repo_name,omitempty"` // A revision within the source repository must be specified in // one of these ways. // @@ -591,13 +591,13 @@ type isRepoSource_Revision interface { } type RepoSource_BranchName struct { - BranchName string `protobuf:"bytes,3,opt,name=branch_name,json=branchName,oneof"` + BranchName string `protobuf:"bytes,3,opt,name=branch_name,json=branchName,proto3,oneof"` } type RepoSource_TagName struct { - TagName string `protobuf:"bytes,4,opt,name=tag_name,json=tagName,oneof"` + TagName string `protobuf:"bytes,4,opt,name=tag_name,json=tagName,proto3,oneof"` } type RepoSource_CommitSha struct { - CommitSha string `protobuf:"bytes,5,opt,name=commit_sha,json=commitSha,oneof"` + CommitSha string `protobuf:"bytes,5,opt,name=commit_sha,json=commitSha,proto3,oneof"` } func (*RepoSource_BranchName) isRepoSource_Revision() {} @@ -731,19 +731,19 @@ func _RepoSource_OneofSizer(msg proto.Message) (n int) { type Command struct { // Name of the command, as presented on the command line, or if the command is // packaged as a Docker container, as presented to `docker pull`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Environment variables set before running this Command. - Env []string `protobuf:"bytes,2,rep,name=env" json:"env,omitempty"` + Env []string `protobuf:"bytes,2,rep,name=env,proto3" json:"env,omitempty"` // Command-line arguments used when executing this Command. - Args []string `protobuf:"bytes,3,rep,name=args" json:"args,omitempty"` + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` // Working directory (relative to project source root) used when running // this Command. - Dir string `protobuf:"bytes,4,opt,name=dir" json:"dir,omitempty"` + Dir string `protobuf:"bytes,4,opt,name=dir,proto3" json:"dir,omitempty"` // Optional unique identifier for this Command, used in wait_for to reference // this Command as a dependency. - Id string `protobuf:"bytes,5,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,5,opt,name=id,proto3" json:"id,omitempty"` // The ID(s) of the Command(s) that this Command depends on. - WaitFor []string `protobuf:"bytes,6,rep,name=wait_for,json=waitFor" json:"wait_for,omitempty"` + WaitFor []string `protobuf:"bytes,6,rep,name=wait_for,json=waitFor,proto3" json:"wait_for,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -825,19 +825,19 @@ type Artifact struct { // to exist here to allow existing BuildProvenance serialized to json in // google.devtools.containeranalysis.v1alpha1.BuildDetails.provenance_bytes to // deserialize back into proto. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Hash or checksum value of a binary, or Docker Registry 2.0 digest of a // container. - Checksum string `protobuf:"bytes,2,opt,name=checksum" json:"checksum,omitempty"` + Checksum string `protobuf:"bytes,2,opt,name=checksum,proto3" json:"checksum,omitempty"` // Artifact ID, if any; for container images, this will be a URL by digest // like gcr.io/projectID/imagename@sha256:123456 - Id string `protobuf:"bytes,3,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,3,opt,name=id,proto3" json:"id,omitempty"` // Related artifact names. This may be the path to a binary or jar file, or in // the case of a container build, the name used to push the container image to // Google Container Registry, as presented to `docker push`. Note that a // single Artifact ID can have multiple names, for example if two tags are // applied to one image. - Names []string `protobuf:"bytes,4,rep,name=names" json:"names,omitempty"` + Names []string `protobuf:"bytes,4,rep,name=names,proto3" json:"names,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/source_context.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/source_context.pb.go index c5592aa24..fb17d4eba 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/source_context.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/containeranalysis/v1alpha1/source_context.pb.go @@ -65,7 +65,7 @@ type SourceContext struct { // *SourceContext_Git Context isSourceContext_Context `protobuf_oneof:"context"` // Labels with user defined metadata. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -100,13 +100,13 @@ type isSourceContext_Context interface { } type SourceContext_CloudRepo struct { - CloudRepo *CloudRepoSourceContext `protobuf:"bytes,1,opt,name=cloud_repo,json=cloudRepo,oneof"` + CloudRepo *CloudRepoSourceContext `protobuf:"bytes,1,opt,name=cloud_repo,json=cloudRepo,proto3,oneof"` } type SourceContext_Gerrit struct { - Gerrit *GerritSourceContext `protobuf:"bytes,2,opt,name=gerrit,oneof"` + Gerrit *GerritSourceContext `protobuf:"bytes,2,opt,name=gerrit,proto3,oneof"` } type SourceContext_Git struct { - Git *GitSourceContext `protobuf:"bytes,3,opt,name=git,oneof"` + Git *GitSourceContext `protobuf:"bytes,3,opt,name=git,proto3,oneof"` } func (*SourceContext_CloudRepo) isSourceContext_Context() {} @@ -244,9 +244,9 @@ func _SourceContext_OneofSizer(msg proto.Message) (n int) { // An alias to a repo revision. type AliasContext struct { // The alias kind. - Kind AliasContext_Kind `protobuf:"varint,1,opt,name=kind,enum=google.devtools.containeranalysis.v1alpha1.AliasContext_Kind" json:"kind,omitempty"` + Kind AliasContext_Kind `protobuf:"varint,1,opt,name=kind,proto3,enum=google.devtools.containeranalysis.v1alpha1.AliasContext_Kind" json:"kind,omitempty"` // The alias name. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -294,7 +294,7 @@ func (m *AliasContext) GetName() string { // Source Repo. type CloudRepoSourceContext struct { // The ID of the repo. - RepoId *RepoId `protobuf:"bytes,1,opt,name=repo_id,json=repoId" json:"repo_id,omitempty"` + RepoId *RepoId `protobuf:"bytes,1,opt,name=repo_id,json=repoId,proto3" json:"repo_id,omitempty"` // A revision in a Cloud Repo can be identified by either its revision ID or // its alias. // @@ -336,10 +336,10 @@ type isCloudRepoSourceContext_Revision interface { } type CloudRepoSourceContext_RevisionId struct { - RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,oneof"` + RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,proto3,oneof"` } type CloudRepoSourceContext_AliasContext struct { - AliasContext *AliasContext `protobuf:"bytes,3,opt,name=alias_context,json=aliasContext,oneof"` + AliasContext *AliasContext `protobuf:"bytes,3,opt,name=alias_context,json=aliasContext,proto3,oneof"` } func (*CloudRepoSourceContext_RevisionId) isCloudRepoSourceContext_Revision() {} @@ -446,11 +446,11 @@ func _CloudRepoSourceContext_OneofSizer(msg proto.Message) (n int) { // A SourceContext referring to a Gerrit project. type GerritSourceContext struct { // The URI of a running Gerrit instance. - HostUri string `protobuf:"bytes,1,opt,name=host_uri,json=hostUri" json:"host_uri,omitempty"` + HostUri string `protobuf:"bytes,1,opt,name=host_uri,json=hostUri,proto3" json:"host_uri,omitempty"` // The full project name within the host. Projects may be nested, so // "project/subproject" is a valid project name. The "repo name" is // the hostURI/project. - GerritProject string `protobuf:"bytes,2,opt,name=gerrit_project,json=gerritProject" json:"gerrit_project,omitempty"` + GerritProject string `protobuf:"bytes,2,opt,name=gerrit_project,json=gerritProject,proto3" json:"gerrit_project,omitempty"` // A revision in a Gerrit project can be identified by either its revision ID // or its alias. // @@ -492,10 +492,10 @@ type isGerritSourceContext_Revision interface { } type GerritSourceContext_RevisionId struct { - RevisionId string `protobuf:"bytes,3,opt,name=revision_id,json=revisionId,oneof"` + RevisionId string `protobuf:"bytes,3,opt,name=revision_id,json=revisionId,proto3,oneof"` } type GerritSourceContext_AliasContext struct { - AliasContext *AliasContext `protobuf:"bytes,4,opt,name=alias_context,json=aliasContext,oneof"` + AliasContext *AliasContext `protobuf:"bytes,4,opt,name=alias_context,json=aliasContext,proto3,oneof"` } func (*GerritSourceContext_RevisionId) isGerritSourceContext_Revision() {} @@ -610,10 +610,10 @@ func _GerritSourceContext_OneofSizer(msg proto.Message) (n int) { // repository (e.g., GitHub). type GitSourceContext struct { // Git repository URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // Required. // Git commit hash. - RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId" json:"revision_id,omitempty"` + RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,proto3" json:"revision_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -700,10 +700,10 @@ type isRepoId_Id interface { } type RepoId_ProjectRepoId struct { - ProjectRepoId *ProjectRepoId `protobuf:"bytes,1,opt,name=project_repo_id,json=projectRepoId,oneof"` + ProjectRepoId *ProjectRepoId `protobuf:"bytes,1,opt,name=project_repo_id,json=projectRepoId,proto3,oneof"` } type RepoId_Uid struct { - Uid string `protobuf:"bytes,2,opt,name=uid,oneof"` + Uid string `protobuf:"bytes,2,opt,name=uid,proto3,oneof"` } func (*RepoId_ProjectRepoId) isRepoId_Id() {} @@ -804,9 +804,9 @@ func _RepoId_OneofSizer(msg proto.Message) (n int) { // winged-cargo-31) and a repo name within that project. type ProjectRepoId struct { // The ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the repo. Leave empty for the default repo. - RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName" json:"repo_name,omitempty"` + RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName,proto3" json:"repo_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/remoteexecution/v1test/remote_execution.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/remoteexecution/v1test/remote_execution.pb.go index f8be19149..3defc4741 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/remoteexecution/v1test/remote_execution.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/remoteexecution/v1test/remote_execution.pb.go @@ -87,7 +87,7 @@ type Action struct { // The digest of the [Command][google.devtools.remoteexecution.v1test.Command] // to run, which MUST be present in the // [ContentAddressableStorage][google.devtools.remoteexecution.v1test.ContentAddressableStorage]. - CommandDigest *Digest `protobuf:"bytes,1,opt,name=command_digest,json=commandDigest" json:"command_digest,omitempty"` + CommandDigest *Digest `protobuf:"bytes,1,opt,name=command_digest,json=commandDigest,proto3" json:"command_digest,omitempty"` // The digest of the root // [Directory][google.devtools.remoteexecution.v1test.Directory] for the input // files. The files in the directory tree are available in the correct @@ -95,7 +95,7 @@ type Action struct { // directory, as well as every subdirectory and content blob referred to, MUST // be in the // [ContentAddressableStorage][google.devtools.remoteexecution.v1test.ContentAddressableStorage]. - InputRootDigest *Digest `protobuf:"bytes,2,opt,name=input_root_digest,json=inputRootDigest" json:"input_root_digest,omitempty"` + InputRootDigest *Digest `protobuf:"bytes,2,opt,name=input_root_digest,json=inputRootDigest,proto3" json:"input_root_digest,omitempty"` // A list of the output files that the client expects to retrieve from the // action. Only the listed files, as well as directories listed in // `output_directories`, will be returned to the client as output. @@ -110,7 +110,7 @@ type Action struct { // In order to ensure consistent hashing of the same Action, the output paths // MUST be sorted lexicographically by code point (or, equivalently, by UTF-8 // bytes). - OutputFiles []string `protobuf:"bytes,3,rep,name=output_files,json=outputFiles" json:"output_files,omitempty"` + OutputFiles []string `protobuf:"bytes,3,rep,name=output_files,json=outputFiles,proto3" json:"output_files,omitempty"` // A list of the output directories that the client expects to retrieve from // the action. Only the contents of the indicated directories (recursively // including the contents of their subdirectories) will be @@ -128,12 +128,12 @@ type Action struct { // In order to ensure consistent hashing of the same Action, the output paths // MUST be sorted lexicographically by code point (or, equivalently, by UTF-8 // bytes). - OutputDirectories []string `protobuf:"bytes,4,rep,name=output_directories,json=outputDirectories" json:"output_directories,omitempty"` + OutputDirectories []string `protobuf:"bytes,4,rep,name=output_directories,json=outputDirectories,proto3" json:"output_directories,omitempty"` // The platform requirements for the execution environment. The server MAY // choose to execute the action on any worker satisfying the requirements, so // the client SHOULD ensure that running the action on any such worker will // have the same result. - Platform *Platform `protobuf:"bytes,5,opt,name=platform" json:"platform,omitempty"` + Platform *Platform `protobuf:"bytes,5,opt,name=platform,proto3" json:"platform,omitempty"` // A timeout after which the execution should be killed. If the timeout is // absent, then the client is specifying that the execution should continue // as long as the server will let it. The server SHOULD impose a timeout if @@ -150,9 +150,9 @@ type Action struct { // timeout is too short. By encoding it directly in the `Action`, a lower // timeout will result in a cache miss and the execution timeout will fail // immediately, rather than whenever the cache entry gets evicted. - Timeout *duration.Duration `protobuf:"bytes,6,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,6,opt,name=timeout,proto3" json:"timeout,omitempty"` // If true, then the `Action`'s result cannot be cached. - DoNotCache bool `protobuf:"varint,7,opt,name=do_not_cache,json=doNotCache" json:"do_not_cache,omitempty"` + DoNotCache bool `protobuf:"varint,7,opt,name=do_not_cache,json=doNotCache,proto3" json:"do_not_cache,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -243,7 +243,7 @@ type Command struct { // evaluated with respect to the input root, or an absolute path. // // The working directory will always be the input root. - Arguments []string `protobuf:"bytes,1,rep,name=arguments" json:"arguments,omitempty"` + Arguments []string `protobuf:"bytes,1,rep,name=arguments,proto3" json:"arguments,omitempty"` // The environment variables to set when running the program. The worker may // provide its own default environment variables; these defaults can be // overridden using this field. Additional variables can also be specified. @@ -251,7 +251,7 @@ type Command struct { // In order to ensure that equivalent `Command`s always hash to the same // value, the environment variables MUST be lexicographically sorted by name. // Sorting of strings is done by code point, equivalently, by the UTF-8 bytes. - EnvironmentVariables []*Command_EnvironmentVariable `protobuf:"bytes,2,rep,name=environment_variables,json=environmentVariables" json:"environment_variables,omitempty"` + EnvironmentVariables []*Command_EnvironmentVariable `protobuf:"bytes,2,rep,name=environment_variables,json=environmentVariables,proto3" json:"environment_variables,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -299,9 +299,9 @@ func (m *Command) GetEnvironmentVariables() []*Command_EnvironmentVariable { // environment. type Command_EnvironmentVariable struct { // The variable name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The variable value. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -358,7 +358,7 @@ type Platform struct { // equivalent `Platform`s always hash to the same value, the properties MUST // be lexicographically sorted by name, and then by value. Sorting of strings // is done by code point, equivalently, by the UTF-8 bytes. - Properties []*Platform_Property `protobuf:"bytes,1,rep,name=properties" json:"properties,omitempty"` + Properties []*Platform_Property `protobuf:"bytes,1,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -414,9 +414,9 @@ func (m *Platform) GetProperties() []*Platform_Property { // files available to the worker. type Platform_Property struct { // The property name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The property value. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -522,9 +522,9 @@ func (m *Platform_Property) GetValue() string { // ``` type Directory struct { // The files in the directory. - Files []*FileNode `protobuf:"bytes,1,rep,name=files" json:"files,omitempty"` + Files []*FileNode `protobuf:"bytes,1,rep,name=files,proto3" json:"files,omitempty"` // The subdirectories in the directory. - Directories []*DirectoryNode `protobuf:"bytes,2,rep,name=directories" json:"directories,omitempty"` + Directories []*DirectoryNode `protobuf:"bytes,2,rep,name=directories,proto3" json:"directories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -571,11 +571,11 @@ func (m *Directory) GetDirectories() []*DirectoryNode { // A `FileNode` represents a single file and associated metadata. type FileNode struct { // The name of the file. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The digest of the file's content. - Digest *Digest `protobuf:"bytes,2,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,2,opt,name=digest,proto3" json:"digest,omitempty"` // True if file is executable, false otherwise. - IsExecutable bool `protobuf:"varint,4,opt,name=is_executable,json=isExecutable" json:"is_executable,omitempty"` + IsExecutable bool `protobuf:"varint,4,opt,name=is_executable,json=isExecutable,proto3" json:"is_executable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -631,12 +631,12 @@ func (m *FileNode) GetIsExecutable() bool { // a `Directory` and its associated metadata. type DirectoryNode struct { // The name of the directory. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The digest of the // [Directory][google.devtools.remoteexecution.v1test.Directory] object // represented. See [Digest][google.devtools.remoteexecution.v1test.Digest] // for information about how to take the digest of a proto message. - Digest *Digest `protobuf:"bytes,2,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,2,opt,name=digest,proto3" json:"digest,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -714,9 +714,9 @@ func (m *DirectoryNode) GetDigest() *Digest { type Digest struct { // The hash. In the case of SHA-256, it will always be a lowercase hex string // exactly 64 characters long. - Hash string `protobuf:"bytes,1,opt,name=hash" json:"hash,omitempty"` + Hash string `protobuf:"bytes,1,opt,name=hash,proto3" json:"hash,omitempty"` // The size of the blob, in bytes. - SizeBytes int64 `protobuf:"varint,2,opt,name=size_bytes,json=sizeBytes" json:"size_bytes,omitempty"` + SizeBytes int64 `protobuf:"varint,2,opt,name=size_bytes,json=sizeBytes,proto3" json:"size_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -771,7 +771,7 @@ type ActionResult struct { // directory where a regular file is expected or vice versa, then that output // will be omitted from the list. The server is free to arrange the output // list as desired; clients MUST NOT assume that the output list is sorted. - OutputFiles []*OutputFile `protobuf:"bytes,2,rep,name=output_files,json=outputFiles" json:"output_files,omitempty"` + OutputFiles []*OutputFile `protobuf:"bytes,2,rep,name=output_files,json=outputFiles,proto3" json:"output_files,omitempty"` // The output directories of the action. For each output directory requested // in the `output_directories` field of the Action, if the corresponding // directory existed after the action completed, a single entry will be @@ -830,9 +830,9 @@ type ActionResult struct { // } // } // ``` - OutputDirectories []*OutputDirectory `protobuf:"bytes,3,rep,name=output_directories,json=outputDirectories" json:"output_directories,omitempty"` + OutputDirectories []*OutputDirectory `protobuf:"bytes,3,rep,name=output_directories,json=outputDirectories,proto3" json:"output_directories,omitempty"` // The exit code of the command. - ExitCode int32 `protobuf:"varint,4,opt,name=exit_code,json=exitCode" json:"exit_code,omitempty"` + ExitCode int32 `protobuf:"varint,4,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"` // The standard output buffer of the action. The server will determine, based // on the size of the buffer, whether to return it in raw form or to return // a digest in `stdout_digest` that points to the buffer. If neither is set, @@ -844,7 +844,7 @@ type ActionResult struct { // can be retrieved from the // [ContentAddressableStorage][google.devtools.remoteexecution.v1test.ContentAddressableStorage]. // See `stdout_raw` for when this will be set. - StdoutDigest *Digest `protobuf:"bytes,6,opt,name=stdout_digest,json=stdoutDigest" json:"stdout_digest,omitempty"` + StdoutDigest *Digest `protobuf:"bytes,6,opt,name=stdout_digest,json=stdoutDigest,proto3" json:"stdout_digest,omitempty"` // The standard error buffer of the action. The server will determine, based // on the size of the buffer, whether to return it in raw form or to return // a digest in `stderr_digest` that points to the buffer. If neither is set, @@ -856,7 +856,7 @@ type ActionResult struct { // can be retrieved from the // [ContentAddressableStorage][google.devtools.remoteexecution.v1test.ContentAddressableStorage]. // See `stderr_raw` for when this will be set. - StderrDigest *Digest `protobuf:"bytes,8,opt,name=stderr_digest,json=stderrDigest" json:"stderr_digest,omitempty"` + StderrDigest *Digest `protobuf:"bytes,8,opt,name=stderr_digest,json=stderrDigest,proto3" json:"stderr_digest,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -945,9 +945,9 @@ type OutputFile struct { // The full path of the file relative to the input root, including the // filename. The path separator is a forward slash `/`. Since this is a // relative path, it MUST NOT begin with a leading forward slash. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // The digest of the file's content. - Digest *Digest `protobuf:"bytes,2,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,2,opt,name=digest,proto3" json:"digest,omitempty"` // The raw content of the file. // // This field may be used by the server to provide the content of a file @@ -960,7 +960,7 @@ type OutputFile struct { // and always be prepared to retrieve it via `digest`. Content []byte `protobuf:"bytes,3,opt,name=content,proto3" json:"content,omitempty"` // True if file is executable, false otherwise. - IsExecutable bool `protobuf:"varint,4,opt,name=is_executable,json=isExecutable" json:"is_executable,omitempty"` + IsExecutable bool `protobuf:"varint,4,opt,name=is_executable,json=isExecutable,proto3" json:"is_executable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1023,12 +1023,12 @@ func (m *OutputFile) GetIsExecutable() bool { // single directory Merkle tree, compressed into one message. type Tree struct { // The root directory in the tree. - Root *Directory `protobuf:"bytes,1,opt,name=root" json:"root,omitempty"` + Root *Directory `protobuf:"bytes,1,opt,name=root,proto3" json:"root,omitempty"` // All the child directories: the directories referred to by the root and, // recursively, all its children. In order to reconstruct the directory tree, // the client must take the digests of each of the child directories and then // build up a tree starting from the `root`. - Children []*Directory `protobuf:"bytes,2,rep,name=children" json:"children,omitempty"` + Children []*Directory `protobuf:"bytes,2,rep,name=children,proto3" json:"children,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1079,13 +1079,13 @@ type OutputDirectory struct { // separator is a forward slash `/`. Since this is a relative path, it MUST // NOT begin with a leading forward slash. The empty string value is allowed, // and it denotes the entire working directory. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // DEPRECATED: This field is deprecated and should no longer be used. - Digest *Digest `protobuf:"bytes,2,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,2,opt,name=digest,proto3" json:"digest,omitempty"` // The digest of the encoded // [Tree][google.devtools.remoteexecution.v1test.Tree] proto containing the // directory's contents. - TreeDigest *Digest `protobuf:"bytes,3,opt,name=tree_digest,json=treeDigest" json:"tree_digest,omitempty"` + TreeDigest *Digest `protobuf:"bytes,3,opt,name=tree_digest,json=treeDigest,proto3" json:"tree_digest,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1144,19 +1144,19 @@ type ExecuteRequest struct { // storage, caches, etc.). The server MAY require use of this field to select // between them in an implementation-defined fashion, otherwise it can be // omitted. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // The action to be performed. - Action *Action `protobuf:"bytes,2,opt,name=action" json:"action,omitempty"` + Action *Action `protobuf:"bytes,2,opt,name=action,proto3" json:"action,omitempty"` // If true, the action will be executed anew even if its result was already // present in the cache. If false, the result may be served from the // [ActionCache][google.devtools.remoteexecution.v1test.ActionCache]. - SkipCacheLookup bool `protobuf:"varint,3,opt,name=skip_cache_lookup,json=skipCacheLookup" json:"skip_cache_lookup,omitempty"` + SkipCacheLookup bool `protobuf:"varint,3,opt,name=skip_cache_lookup,json=skipCacheLookup,proto3" json:"skip_cache_lookup,omitempty"` // DEPRECATED: This field should be ignored by clients and servers and will be // removed. - TotalInputFileCount int32 `protobuf:"varint,4,opt,name=total_input_file_count,json=totalInputFileCount" json:"total_input_file_count,omitempty"` + TotalInputFileCount int32 `protobuf:"varint,4,opt,name=total_input_file_count,json=totalInputFileCount,proto3" json:"total_input_file_count,omitempty"` // DEPRECATED: This field should be ignored by clients and servers and will be // removed. - TotalInputFileBytes int64 `protobuf:"varint,5,opt,name=total_input_file_bytes,json=totalInputFileBytes" json:"total_input_file_bytes,omitempty"` + TotalInputFileBytes int64 `protobuf:"varint,5,opt,name=total_input_file_bytes,json=totalInputFileBytes,proto3" json:"total_input_file_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1224,13 +1224,13 @@ func (m *ExecuteRequest) GetTotalInputFileBytes() int64 { // A `LogFile` is a log stored in the CAS. type LogFile struct { // The digest of the log contents. - Digest *Digest `protobuf:"bytes,1,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,1,opt,name=digest,proto3" json:"digest,omitempty"` // This is a hint as to the purpose of the log, and is set to true if the log // is human-readable text that can be usefully displayed to a user, and false // otherwise. For instance, if a command-line client wishes to print the // server logs to the terminal for a failed action, this allows it to avoid // displaying a binary file. - HumanReadable bool `protobuf:"varint,2,opt,name=human_readable,json=humanReadable" json:"human_readable,omitempty"` + HumanReadable bool `protobuf:"varint,2,opt,name=human_readable,json=humanReadable,proto3" json:"human_readable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1281,9 +1281,9 @@ func (m *LogFile) GetHumanReadable() bool { // [Operation][google.longrunning.Operation]. type ExecuteResponse struct { // The result of the action. - Result *ActionResult `protobuf:"bytes,1,opt,name=result" json:"result,omitempty"` + Result *ActionResult `protobuf:"bytes,1,opt,name=result,proto3" json:"result,omitempty"` // True if the result was served from cache, false if it was executed. - CachedResult bool `protobuf:"varint,2,opt,name=cached_result,json=cachedResult" json:"cached_result,omitempty"` + CachedResult bool `protobuf:"varint,2,opt,name=cached_result,json=cachedResult,proto3" json:"cached_result,omitempty"` // If the status has a code other than `OK`, it indicates that the action did // not finish execution. For example, if the operation times out during // execution, the status will have a `DEADLINE_EXCEEDED` code. Servers MUST @@ -1294,7 +1294,7 @@ type ExecuteResponse struct { // For an error status, the `result` field is optional; the server may // populate the output-, stdout-, and stderr-related fields if it has any // information available, such as the stdout and stderr of a timed-out action. - Status *status.Status `protobuf:"bytes,3,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,3,opt,name=status,proto3" json:"status,omitempty"` // An optional list of additional log outputs the server wishes to provide. A // server can use this to return execution-specific logs however it wishes. // This is intended primarily to make it easier for users to debug issues that @@ -1302,7 +1302,7 @@ type ExecuteResponse struct { // worker executing the action or by providing logs from the worker's setup // phase. The keys SHOULD be human readable so that a client can display them // to a user. - ServerLogs map[string]*LogFile `protobuf:"bytes,4,rep,name=server_logs,json=serverLogs" json:"server_logs,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ServerLogs map[string]*LogFile `protobuf:"bytes,4,rep,name=server_logs,json=serverLogs,proto3" json:"server_logs,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1366,18 +1366,18 @@ func (m *ExecuteResponse) GetServerLogs() map[string]*LogFile { // field][google.longrunning.Operation.response] of the // [Operation][google.longrunning.Operation]. type ExecuteOperationMetadata struct { - Stage ExecuteOperationMetadata_Stage `protobuf:"varint,1,opt,name=stage,enum=google.devtools.remoteexecution.v1test.ExecuteOperationMetadata_Stage" json:"stage,omitempty"` + Stage ExecuteOperationMetadata_Stage `protobuf:"varint,1,opt,name=stage,proto3,enum=google.devtools.remoteexecution.v1test.ExecuteOperationMetadata_Stage" json:"stage,omitempty"` // The digest of the [Action][google.devtools.remoteexecution.v1test.Action] // being executed. - ActionDigest *Digest `protobuf:"bytes,2,opt,name=action_digest,json=actionDigest" json:"action_digest,omitempty"` + ActionDigest *Digest `protobuf:"bytes,2,opt,name=action_digest,json=actionDigest,proto3" json:"action_digest,omitempty"` // If set, the client can use this name with // [ByteStream.Read][google.bytestream.ByteStream.Read] to stream the // standard output. - StdoutStreamName string `protobuf:"bytes,3,opt,name=stdout_stream_name,json=stdoutStreamName" json:"stdout_stream_name,omitempty"` + StdoutStreamName string `protobuf:"bytes,3,opt,name=stdout_stream_name,json=stdoutStreamName,proto3" json:"stdout_stream_name,omitempty"` // If set, the client can use this name with // [ByteStream.Read][google.bytestream.ByteStream.Read] to stream the // standard error. - StderrStreamName string `protobuf:"bytes,4,opt,name=stderr_stream_name,json=stderrStreamName" json:"stderr_stream_name,omitempty"` + StderrStreamName string `protobuf:"bytes,4,opt,name=stderr_stream_name,json=stderrStreamName,proto3" json:"stderr_stream_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1443,10 +1443,10 @@ type GetActionResultRequest struct { // storage, caches, etc.). The server MAY require use of this field to select // between them in an implementation-defined fashion, otherwise it can be // omitted. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // The digest of the [Action][google.devtools.remoteexecution.v1test.Action] // whose result is requested. - ActionDigest *Digest `protobuf:"bytes,2,opt,name=action_digest,json=actionDigest" json:"action_digest,omitempty"` + ActionDigest *Digest `protobuf:"bytes,2,opt,name=action_digest,json=actionDigest,proto3" json:"action_digest,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1498,13 +1498,13 @@ type UpdateActionResultRequest struct { // storage, caches, etc.). The server MAY require use of this field to select // between them in an implementation-defined fashion, otherwise it can be // omitted. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // The digest of the [Action][google.devtools.remoteexecution.v1test.Action] // whose result is being uploaded. - ActionDigest *Digest `protobuf:"bytes,2,opt,name=action_digest,json=actionDigest" json:"action_digest,omitempty"` + ActionDigest *Digest `protobuf:"bytes,2,opt,name=action_digest,json=actionDigest,proto3" json:"action_digest,omitempty"` // The [ActionResult][google.devtools.remoteexecution.v1test.ActionResult] // to store in the cache. - ActionResult *ActionResult `protobuf:"bytes,3,opt,name=action_result,json=actionResult" json:"action_result,omitempty"` + ActionResult *ActionResult `protobuf:"bytes,3,opt,name=action_result,json=actionResult,proto3" json:"action_result,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1563,9 +1563,9 @@ type FindMissingBlobsRequest struct { // storage, caches, etc.). The server MAY require use of this field to select // between them in an implementation-defined fashion, otherwise it can be // omitted. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // A list of the blobs to check. - BlobDigests []*Digest `protobuf:"bytes,2,rep,name=blob_digests,json=blobDigests" json:"blob_digests,omitempty"` + BlobDigests []*Digest `protobuf:"bytes,2,rep,name=blob_digests,json=blobDigests,proto3" json:"blob_digests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1613,7 +1613,7 @@ func (m *FindMissingBlobsRequest) GetBlobDigests() []*Digest { // [ContentAddressableStorage.FindMissingBlobs][google.devtools.remoteexecution.v1test.ContentAddressableStorage.FindMissingBlobs]. type FindMissingBlobsResponse struct { // A list of the blobs requested *not* present in the storage. - MissingBlobDigests []*Digest `protobuf:"bytes,2,rep,name=missing_blob_digests,json=missingBlobDigests" json:"missing_blob_digests,omitempty"` + MissingBlobDigests []*Digest `protobuf:"bytes,2,rep,name=missing_blob_digests,json=missingBlobDigests,proto3" json:"missing_blob_digests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1654,7 +1654,7 @@ func (m *FindMissingBlobsResponse) GetMissingBlobDigests() []*Digest { // [ContentAddressableStorage.BatchUpdateBlobs][google.devtools.remoteexecution.v1test.ContentAddressableStorage.BatchUpdateBlobs]. type UpdateBlobRequest struct { // The digest of the blob. This MUST be the digest of `data`. - ContentDigest *Digest `protobuf:"bytes,1,opt,name=content_digest,json=contentDigest" json:"content_digest,omitempty"` + ContentDigest *Digest `protobuf:"bytes,1,opt,name=content_digest,json=contentDigest,proto3" json:"content_digest,omitempty"` // The raw binary data. Data []byte `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1708,9 +1708,9 @@ type BatchUpdateBlobsRequest struct { // storage, caches, etc.). The server MAY require use of this field to select // between them in an implementation-defined fashion, otherwise it can be // omitted. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // The individual upload requests. - Requests []*UpdateBlobRequest `protobuf:"bytes,2,rep,name=requests" json:"requests,omitempty"` + Requests []*UpdateBlobRequest `protobuf:"bytes,2,rep,name=requests,proto3" json:"requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1758,7 +1758,7 @@ func (m *BatchUpdateBlobsRequest) GetRequests() []*UpdateBlobRequest { // [ContentAddressableStorage.BatchUpdateBlobs][google.devtools.remoteexecution.v1test.ContentAddressableStorage.BatchUpdateBlobs]. type BatchUpdateBlobsResponse struct { // The responses to the requests. - Responses []*BatchUpdateBlobsResponse_Response `protobuf:"bytes,1,rep,name=responses" json:"responses,omitempty"` + Responses []*BatchUpdateBlobsResponse_Response `protobuf:"bytes,1,rep,name=responses,proto3" json:"responses,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1798,9 +1798,9 @@ func (m *BatchUpdateBlobsResponse) GetResponses() []*BatchUpdateBlobsResponse_Re // A response corresponding to a single blob that the client tried to upload. type BatchUpdateBlobsResponse_Response struct { // The digest to which this response corresponds. - BlobDigest *Digest `protobuf:"bytes,1,opt,name=blob_digest,json=blobDigest" json:"blob_digest,omitempty"` + BlobDigest *Digest `protobuf:"bytes,1,opt,name=blob_digest,json=blobDigest,proto3" json:"blob_digest,omitempty"` // The result of attempting to upload that blob. - Status *status.Status `protobuf:"bytes,2,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1852,21 +1852,21 @@ type GetTreeRequest struct { // storage, caches, etc.). The server MAY require use of this field to select // between them in an implementation-defined fashion, otherwise it can be // omitted. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // The digest of the root, which must be an encoded // [Directory][google.devtools.remoteexecution.v1test.Directory] message // stored in the // [ContentAddressableStorage][google.devtools.remoteexecution.v1test.ContentAddressableStorage]. - RootDigest *Digest `protobuf:"bytes,2,opt,name=root_digest,json=rootDigest" json:"root_digest,omitempty"` + RootDigest *Digest `protobuf:"bytes,2,opt,name=root_digest,json=rootDigest,proto3" json:"root_digest,omitempty"` // A maximum page size to request. If present, the server will request no more // than this many items. Regardless of whether a page size is specified, the // server may place its own limit on the number of items to be returned and // require the client to retrieve more items using a subsequent request. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A page token, which must be a value received in a previous // [GetTreeResponse][google.devtools.remoteexecution.v1test.GetTreeResponse]. // If present, the server will use it to return the following page of results. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1928,12 +1928,12 @@ func (m *GetTreeRequest) GetPageToken() string { // [ContentAddressableStorage.GetTree][google.devtools.remoteexecution.v1test.ContentAddressableStorage.GetTree]. type GetTreeResponse struct { // The directories descended from the requested root. - Directories []*Directory `protobuf:"bytes,1,rep,name=directories" json:"directories,omitempty"` + Directories []*Directory `protobuf:"bytes,1,rep,name=directories,proto3" json:"directories,omitempty"` // If present, signifies that there are more results which the client can // retrieve by passing this as the page_token in a subsequent // [request][google.devtools.remoteexecution.v1test.GetTreeRequest]. // If empty, signifies that this is the last page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1980,9 +1980,9 @@ func (m *GetTreeResponse) GetNextPageToken() string { // Details for the tool used to call the API. type ToolDetails struct { // Name of the tool, e.g. bazel. - ToolName string `protobuf:"bytes,1,opt,name=tool_name,json=toolName" json:"tool_name,omitempty"` + ToolName string `protobuf:"bytes,1,opt,name=tool_name,json=toolName,proto3" json:"tool_name,omitempty"` // Version of the tool used for the request, e.g. 5.0.3. - ToolVersion string `protobuf:"bytes,2,opt,name=tool_version,json=toolVersion" json:"tool_version,omitempty"` + ToolVersion string `protobuf:"bytes,2,opt,name=tool_version,json=toolVersion,proto3" json:"tool_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2034,17 +2034,17 @@ func (m *ToolDetails) GetToolVersion() string { // contents: the base64 encoded binary RequestMetadata message. type RequestMetadata struct { // The details for the tool invoking the requests. - ToolDetails *ToolDetails `protobuf:"bytes,1,opt,name=tool_details,json=toolDetails" json:"tool_details,omitempty"` + ToolDetails *ToolDetails `protobuf:"bytes,1,opt,name=tool_details,json=toolDetails,proto3" json:"tool_details,omitempty"` // An identifier that ties multiple requests to the same action. // For example, multiple requests to the CAS, Action Cache, and Execution // API are used in order to compile foo.cc. - ActionId string `protobuf:"bytes,2,opt,name=action_id,json=actionId" json:"action_id,omitempty"` + ActionId string `protobuf:"bytes,2,opt,name=action_id,json=actionId,proto3" json:"action_id,omitempty"` // An identifier that ties multiple actions together to a final result. // For example, multiple actions are required to build and run foo_test. - ToolInvocationId string `protobuf:"bytes,3,opt,name=tool_invocation_id,json=toolInvocationId" json:"tool_invocation_id,omitempty"` + ToolInvocationId string `protobuf:"bytes,3,opt,name=tool_invocation_id,json=toolInvocationId,proto3" json:"tool_invocation_id,omitempty"` // An identifier to tie multiple tool invocations together. For example, // runs of foo_test, bar_test and baz_test on a post-submit of a given patch. - CorrelatedInvocationsId string `protobuf:"bytes,4,opt,name=correlated_invocations_id,json=correlatedInvocationsId" json:"correlated_invocations_id,omitempty"` + CorrelatedInvocationsId string `protobuf:"bytes,4,opt,name=correlated_invocations_id,json=correlatedInvocationsId,proto3" json:"correlated_invocations_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2231,8 +2231,7 @@ func (c *executionClient) Execute(ctx context.Context, in *ExecuteRequest, opts return out, nil } -// Server API for Execution service - +// ExecutionServer is the server API for Execution service. type ExecutionServer interface { // Execute an action remotely. // @@ -2384,8 +2383,7 @@ func (c *actionCacheClient) UpdateActionResult(ctx context.Context, in *UpdateAc return out, nil } -// Server API for ActionCache service - +// ActionCacheServer is the server API for ActionCache service. type ActionCacheServer interface { // Retrieve a cached execution result. // @@ -2547,8 +2545,7 @@ func (c *contentAddressableStorageClient) GetTree(ctx context.Context, in *GetTr return out, nil } -// Server API for ContentAddressableStorage service - +// ContentAddressableStorageServer is the server API for ContentAddressableStorage service. type ContentAddressableStorageServer interface { // Determine if blobs are present in the CAS. // diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/bots.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/bots.pb.go index f436938b1..a859a71ab 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/bots.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/bots.pb.go @@ -205,7 +205,7 @@ func (PostBotEventTempRequest_Type) EnumDescriptor() ([]byte, []int) { type BotSession struct { // The bot session name, as selected by the server. Output only during a call // to CreateBotSession. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A unique bot ID within the farm used to persistently identify this bot over // time (i.e., over multiple sessions). This ID must be unique within a // farm. Typically, the bot ID will be the same as the name of the primary @@ -220,10 +220,10 @@ type BotSession struct { // invalid session, the server must reject that request, and may also // quarantine the other bot with the same bot IDs (ie, stop sending it new // leases and alert an admin). - BotId string `protobuf:"bytes,2,opt,name=bot_id,json=botId" json:"bot_id,omitempty"` + BotId string `protobuf:"bytes,2,opt,name=bot_id,json=botId,proto3" json:"bot_id,omitempty"` // The status of the bot. This must be populated in every call to // UpdateBotSession. - Status BotStatus `protobuf:"varint,3,opt,name=status,enum=google.devtools.remoteworkers.v1test2.BotStatus" json:"status,omitempty"` + Status BotStatus `protobuf:"varint,3,opt,name=status,proto3,enum=google.devtools.remoteworkers.v1test2.BotStatus" json:"status,omitempty"` // A description of the worker hosting this bot. The Worker message is used // here in the Status context (see Worker for more information). If multiple // bots are running on the worker, this field should only describe the @@ -232,16 +232,16 @@ type BotSession struct { // During the call to CreateBotSession, the server may make arbitrary changes // to the worker's `server_properties` field (see that field for more // information). Otherwise, this field is input-only. - Worker *Worker `protobuf:"bytes,4,opt,name=worker" json:"worker,omitempty"` + Worker *Worker `protobuf:"bytes,4,opt,name=worker,proto3" json:"worker,omitempty"` // A list of all leases that are a part of this session. See the Lease message // for details. - Leases []*Lease `protobuf:"bytes,5,rep,name=leases" json:"leases,omitempty"` + Leases []*Lease `protobuf:"bytes,5,rep,name=leases,proto3" json:"leases,omitempty"` // The time at which this bot session will expire, unless the bot calls // UpdateBotSession again. Output only. - ExpireTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` // The version of the bot code currently running. The server may use this // information to issue an admin action to tell the bot to update itself. - Version string `protobuf:"bytes,7,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,7,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -358,9 +358,9 @@ type Lease struct { // [https://cloud.google.com/apis/design/resource_names#full_resource_name] if // it should be accessed through an endpoint that is not already known to the // bot. - Assignment string `protobuf:"bytes,1,opt,name=assignment" json:"assignment,omitempty"` + Assignment string `protobuf:"bytes,1,opt,name=assignment,proto3" json:"assignment,omitempty"` // The state of the lease. See LeaseState for more information. - State LeaseState `protobuf:"varint,2,opt,name=state,enum=google.devtools.remoteworkers.v1test2.LeaseState" json:"state,omitempty"` + State LeaseState `protobuf:"varint,2,opt,name=state,proto3,enum=google.devtools.remoteworkers.v1test2.LeaseState" json:"state,omitempty"` // The final status of the lease (should be populated by the bot if the state // is completed). This is the status of the lease, not of any task represented // by the lease. For example, if the bot could not accept the lease because it @@ -368,14 +368,14 @@ type Lease struct { // FAILED_PRECONDITION. But if the assignment in the lease didn't execute // correctly, this field will be `OK` while the failure of the assignment must // be tracked elsewhere (e.g., through the Tasks interface). - Status *status.Status `protobuf:"bytes,3,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,3,opt,name=status,proto3" json:"status,omitempty"` // The requirements that are being claimed by this lease. This field may be // omitted by the server if the lease is not pending. - Requirements *Worker `protobuf:"bytes,4,opt,name=requirements" json:"requirements,omitempty"` + Requirements *Worker `protobuf:"bytes,4,opt,name=requirements,proto3" json:"requirements,omitempty"` // The time at which this lease expires. The server *may* extend this over // time, but due to race conditions, the bot is not *required* to respect any // expiry date except the first one. - ExpireTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` // While the `assignment` field is a resource name that allows the bot to // get the actual assignment, the server can also optionally include the // assignment itself inline in order to save a round trip to the server. @@ -388,7 +388,7 @@ type Lease struct { // // This field may be omitted by the server if the lease is not in the // `PENDING` state. - InlineAssignment *any.Any `protobuf:"bytes,6,opt,name=inline_assignment,json=inlineAssignment" json:"inline_assignment,omitempty"` + InlineAssignment *any.Any `protobuf:"bytes,6,opt,name=inline_assignment,json=inlineAssignment,proto3" json:"inline_assignment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -474,9 +474,9 @@ func (m *Lease) GetInlineAssignment() *any.Any { // project (http://github.com/luci/luci-py/appengine/swarming). type AdminTemp struct { // The admin action; see `Command` for legal values. - Command AdminTemp_Command `protobuf:"varint,1,opt,name=command,enum=google.devtools.remoteworkers.v1test2.AdminTemp_Command" json:"command,omitempty"` + Command AdminTemp_Command `protobuf:"varint,1,opt,name=command,proto3,enum=google.devtools.remoteworkers.v1test2.AdminTemp_Command" json:"command,omitempty"` // The argument to the admin action; see `Command` for semantics. - Arg string `protobuf:"bytes,2,opt,name=arg" json:"arg,omitempty"` + Arg string `protobuf:"bytes,2,opt,name=arg,proto3" json:"arg,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -523,9 +523,9 @@ func (m *AdminTemp) GetArg() string { // Request message for CreateBotSession. type CreateBotSessionRequest struct { // The farm resource. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The bot session to create. Server-assigned fields like name must be unset. - BotSession *BotSession `protobuf:"bytes,2,opt,name=bot_session,json=botSession" json:"bot_session,omitempty"` + BotSession *BotSession `protobuf:"bytes,2,opt,name=bot_session,json=botSession,proto3" json:"bot_session,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -572,12 +572,12 @@ func (m *CreateBotSessionRequest) GetBotSession() *BotSession { // Request message for UpdateBotSession. type UpdateBotSessionRequest struct { // The bot session name. Must match bot_session.name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The bot session resource to update. - BotSession *BotSession `protobuf:"bytes,2,opt,name=bot_session,json=botSession" json:"bot_session,omitempty"` + BotSession *BotSession `protobuf:"bytes,2,opt,name=bot_session,json=botSession,proto3" json:"bot_session,omitempty"` // The fields on the bot that should be updated. See the BotSession resource // for which fields are updatable by which caller. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -631,11 +631,11 @@ func (m *UpdateBotSessionRequest) GetUpdateMask() *field_mask.FieldMask { // Request message for PostBotEventTemp type PostBotEventTempRequest struct { // The bot session name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The type of bot event. - Type PostBotEventTempRequest_Type `protobuf:"varint,2,opt,name=type,enum=google.devtools.remoteworkers.v1test2.PostBotEventTempRequest_Type" json:"type,omitempty"` + Type PostBotEventTempRequest_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.devtools.remoteworkers.v1test2.PostBotEventTempRequest_Type" json:"type,omitempty"` // A human-readable message. - Msg string `protobuf:"bytes,3,opt,name=msg" json:"msg,omitempty"` + Msg string `protobuf:"bytes,3,opt,name=msg,proto3" json:"msg,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -761,8 +761,7 @@ func (c *botsClient) PostBotEventTemp(ctx context.Context, in *PostBotEventTempR return out, nil } -// Server API for Bots service - +// BotsServer is the server API for Bots service. type BotsServer interface { // CreateBotSession is called when the bot first joins the farm, and // establishes a session ID to ensure that multiple machines do not register diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/command.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/command.pb.go index 79554110b..75ffebcdf 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/command.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/command.pb.go @@ -22,11 +22,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Describes a shell-style task to execute. type CommandTask struct { // The inputs to the task. - Inputs *CommandTask_Inputs `protobuf:"bytes,1,opt,name=inputs" json:"inputs,omitempty"` + Inputs *CommandTask_Inputs `protobuf:"bytes,1,opt,name=inputs,proto3" json:"inputs,omitempty"` // The expected outputs from the task. - ExpectedOutputs *CommandTask_Outputs `protobuf:"bytes,4,opt,name=expected_outputs,json=expectedOutputs" json:"expected_outputs,omitempty"` + ExpectedOutputs *CommandTask_Outputs `protobuf:"bytes,4,opt,name=expected_outputs,json=expectedOutputs,proto3" json:"expected_outputs,omitempty"` // The timeouts of this task. - Timeouts *CommandTask_Timeouts `protobuf:"bytes,5,opt,name=timeouts" json:"timeouts,omitempty"` + Timeouts *CommandTask_Timeouts `protobuf:"bytes,5,opt,name=timeouts,proto3" json:"timeouts,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -80,7 +80,7 @@ func (m *CommandTask) GetTimeouts() *CommandTask_Timeouts { // Describes the inputs to a shell-style task. type CommandTask_Inputs struct { // The command itself to run (e.g., argv) - Arguments []string `protobuf:"bytes,1,rep,name=arguments" json:"arguments,omitempty"` + Arguments []string `protobuf:"bytes,1,rep,name=arguments,proto3" json:"arguments,omitempty"` // The input filesystem to be set up prior to the task beginning. The // contents should be a repeated set of FileMetadata messages though other // formats are allowed if better for the implementation (eg, a LUCI-style @@ -90,9 +90,9 @@ type CommandTask_Inputs struct { // metadata, in which case it may be useful to break up portions of the // filesystem that change frequently (eg, specific input files) from those // that don't (eg, standard header files). - Files []*Digest `protobuf:"bytes,2,rep,name=files" json:"files,omitempty"` + Files []*Digest `protobuf:"bytes,2,rep,name=files,proto3" json:"files,omitempty"` // All environment variables required by the task. - EnvironmentVariables []*CommandTask_Inputs_EnvironmentVariable `protobuf:"bytes,3,rep,name=environment_variables,json=environmentVariables" json:"environment_variables,omitempty"` + EnvironmentVariables []*CommandTask_Inputs_EnvironmentVariable `protobuf:"bytes,3,rep,name=environment_variables,json=environmentVariables,proto3" json:"environment_variables,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -146,9 +146,9 @@ func (m *CommandTask_Inputs) GetEnvironmentVariables() []*CommandTask_Inputs_Env // An environment variable required by this task. type CommandTask_Inputs_EnvironmentVariable struct { // The envvar name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The envvar value. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -197,9 +197,9 @@ func (m *CommandTask_Inputs_EnvironmentVariable) GetValue() string { // Describes the expected outputs of the command. type CommandTask_Outputs struct { // A list of expected files, relative to the execution root. - Files []string `protobuf:"bytes,1,rep,name=files" json:"files,omitempty"` + Files []string `protobuf:"bytes,1,rep,name=files,proto3" json:"files,omitempty"` // A list of expected directories, relative to the execution root. - Directories []string `protobuf:"bytes,2,rep,name=directories" json:"directories,omitempty"` + Directories []string `protobuf:"bytes,2,rep,name=directories,proto3" json:"directories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -248,19 +248,19 @@ type CommandTask_Timeouts struct { // This specifies the maximum time that the task can run, excluding the // time required to download inputs or upload outputs. That is, the worker // will terminate the task if it runs longer than this. - Execution *duration.Duration `protobuf:"bytes,1,opt,name=execution" json:"execution,omitempty"` + Execution *duration.Duration `protobuf:"bytes,1,opt,name=execution,proto3" json:"execution,omitempty"` // This specifies the maximum amount of time the task can be idle - that is, // go without generating some output in either stdout or stderr. If the // process is silent for more than the specified time, the worker will // terminate the task. - Idle *duration.Duration `protobuf:"bytes,2,opt,name=idle" json:"idle,omitempty"` + Idle *duration.Duration `protobuf:"bytes,2,opt,name=idle,proto3" json:"idle,omitempty"` // If the execution or IO timeouts are exceeded, the worker will try to // gracefully terminate the task and return any existing logs. However, // tasks may be hard-frozen in which case this process will fail. This // timeout specifies how long to wait for a terminated task to shut down // gracefully (e.g. via SIGTERM) before we bring down the hammer (e.g. // SIGKILL on *nix, CTRL_BREAK_EVENT on Windows). - Shutdown *duration.Duration `protobuf:"bytes,3,opt,name=shutdown" json:"shutdown,omitempty"` + Shutdown *duration.Duration `protobuf:"bytes,3,opt,name=shutdown,proto3" json:"shutdown,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -317,12 +317,12 @@ type CommandOutputs struct { // exceeded its deadline or was cancelled, the process may still produce an // exit code as it is cancelled, and this will be populated, but a successful // (zero) is unlikely to be correct unless the status code is OK. - ExitCode int32 `protobuf:"varint,1,opt,name=exit_code,json=exitCode" json:"exit_code,omitempty"` + ExitCode int32 `protobuf:"varint,1,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"` // The output files. The blob referenced by the digest should contain // one of the following (implementation-dependent): // * A marshalled DirectoryMetadata of the returned filesystem // * A LUCI-style .isolated file - Outputs *Digest `protobuf:"bytes,2,opt,name=outputs" json:"outputs,omitempty"` + Outputs *Digest `protobuf:"bytes,2,opt,name=outputs,proto3" json:"outputs,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -372,10 +372,10 @@ type CommandOverhead struct { // The elapsed time between calling Accept and Complete. The server will also // have its own idea of what this should be, but this excludes the overhead of // the RPCs and the bot response time. - Duration *duration.Duration `protobuf:"bytes,1,opt,name=duration" json:"duration,omitempty"` + Duration *duration.Duration `protobuf:"bytes,1,opt,name=duration,proto3" json:"duration,omitempty"` // The amount of time *not* spent executing the command (ie // uploading/downloading files). - Overhead *duration.Duration `protobuf:"bytes,2,opt,name=overhead" json:"overhead,omitempty"` + Overhead *duration.Duration `protobuf:"bytes,2,opt,name=overhead,proto3" json:"overhead,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -427,15 +427,15 @@ type FileMetadata struct { // root and must correspond to an entry in CommandTask.outputs.files. If this // message is part of a Directory message, then the path is relative to the // root of that directory. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // A pointer to the contents of the file. The method by which a client // retrieves the contents from a CAS system is not defined here. - Digest *Digest `protobuf:"bytes,2,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,2,opt,name=digest,proto3" json:"digest,omitempty"` // If the file is small enough, its contents may also or alternatively be // listed here. Contents []byte `protobuf:"bytes,3,opt,name=contents,proto3" json:"contents,omitempty"` // Properties of the file - IsExecutable bool `protobuf:"varint,4,opt,name=is_executable,json=isExecutable" json:"is_executable,omitempty"` + IsExecutable bool `protobuf:"varint,4,opt,name=is_executable,json=isExecutable,proto3" json:"is_executable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -497,10 +497,10 @@ func (m *FileMetadata) GetIsExecutable() bool { // Execution API. type DirectoryMetadata struct { // The path of the directory, as in [FileMetadata.path][google.devtools.remoteworkers.v1test2.FileMetadata.path]. - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // A pointer to the contents of the directory, in the form of a marshalled // Directory message. - Digest *Digest `protobuf:"bytes,2,opt,name=digest" json:"digest,omitempty"` + Digest *Digest `protobuf:"bytes,2,opt,name=digest,proto3" json:"digest,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -550,12 +550,12 @@ func (m *DirectoryMetadata) GetDigest() *Digest { type Digest struct { // A string-encoded hash (eg "1a2b3c", not the byte array [0x1a, 0x2b, 0x3c]) // using an implementation-defined hash algorithm (eg SHA-256). - Hash string `protobuf:"bytes,1,opt,name=hash" json:"hash,omitempty"` + Hash string `protobuf:"bytes,1,opt,name=hash,proto3" json:"hash,omitempty"` // The size of the contents. While this is not strictly required as part of an // identifier (after all, any given hash will have exactly one canonical // size), it's useful in almost all cases when one might want to send or // retrieve blobs of content and is included here for this reason. - SizeBytes int64 `protobuf:"varint,2,opt,name=size_bytes,json=sizeBytes" json:"size_bytes,omitempty"` + SizeBytes int64 `protobuf:"varint,2,opt,name=size_bytes,json=sizeBytes,proto3" json:"size_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -603,9 +603,9 @@ func (m *Digest) GetSizeBytes() int64 { // Execution API. type Directory struct { // The files in this directory - Files []*FileMetadata `protobuf:"bytes,1,rep,name=files" json:"files,omitempty"` + Files []*FileMetadata `protobuf:"bytes,1,rep,name=files,proto3" json:"files,omitempty"` // Any subdirectories - Directories []*DirectoryMetadata `protobuf:"bytes,2,rep,name=directories" json:"directories,omitempty"` + Directories []*DirectoryMetadata `protobuf:"bytes,2,rep,name=directories,proto3" json:"directories,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/tasks.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/tasks.pb.go index b2827a393..f3cd9964e 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/tasks.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/tasks.pb.go @@ -38,17 +38,17 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // overhead. type Task struct { // The name of this task. Output only. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The actual task to perform. For example, this could be CommandTask to run a // command line. - Description *any.Any `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description *any.Any `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Handles to logs. The key is a human-readable name like `stdout`, and the // handle is a resource name that can be passed to ByteStream or other // accessors. // // An implementation may define some logs by default (like `stdout`), and may // allow clients to add new logs via AddTaskLog. - Logs map[string]string `protobuf:"bytes,3,rep,name=logs" json:"logs,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Logs map[string]string `protobuf:"bytes,3,rep,name=logs,proto3" json:"logs,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -103,7 +103,7 @@ func (m *Task) GetLogs() map[string]string { type TaskResult struct { // The name of the task result; must be a name of a `Task` followed by // `/result`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The result may be updated several times; the client must only set // `complete` to true to indicate that no further updates are allowed. // If this is not true, the `status` field must not be examined since its zero @@ -111,17 +111,17 @@ type TaskResult struct { // // Once a task is completed, it must not be updated with further results, // though the implementation may choose to continue to receive logs. - Complete bool `protobuf:"varint,2,opt,name=complete" json:"complete,omitempty"` + Complete bool `protobuf:"varint,2,opt,name=complete,proto3" json:"complete,omitempty"` // The final status of the task itself. For example, if task.description // included a timeout which was violated, status.code may be // DEADLINE_EXCEEDED. This field can only be read if `complete` is true. - Status *status.Status `protobuf:"bytes,3,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,3,opt,name=status,proto3" json:"status,omitempty"` // Any non-log output, such as output files and exit codes. See // CommandResult as an example. - Output *any.Any `protobuf:"bytes,4,opt,name=output" json:"output,omitempty"` + Output *any.Any `protobuf:"bytes,4,opt,name=output,proto3" json:"output,omitempty"` // Any information about how the command was executed, eg runtime. See // CommandOverhead as an example. - Meta *any.Any `protobuf:"bytes,5,opt,name=meta" json:"meta,omitempty"` + Meta *any.Any `protobuf:"bytes,5,opt,name=meta,proto3" json:"meta,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -189,7 +189,7 @@ func (m *TaskResult) GetMeta() *any.Any { // Request message for `GetTask`. type GetTaskRequest struct { // The task name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -229,15 +229,15 @@ func (m *GetTaskRequest) GetName() string { // Request message for `UpdateTaskResult`. type UpdateTaskResultRequest struct { // The task result name; must match `result.name`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The result being updated. - Result *TaskResult `protobuf:"bytes,2,opt,name=result" json:"result,omitempty"` + Result *TaskResult `protobuf:"bytes,2,opt,name=result,proto3" json:"result,omitempty"` // The fields within `result` that are specified. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // If this is being updated by a bot from BotManager, the source should be // bot.session_id. That way, if two bots accidentally get the same name, we'll // know to reject updates from the older one. - Source string `protobuf:"bytes,4,opt,name=source" json:"source,omitempty"` + Source string `protobuf:"bytes,4,opt,name=source,proto3" json:"source,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -298,9 +298,9 @@ func (m *UpdateTaskResultRequest) GetSource() string { // Request message for `AddTaskLog`. type AddTaskLogRequest struct { // The name of the task that will own the new log. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The human-readable name of the log, like `stdout` or a relative file path. - LogId string `protobuf:"bytes,2,opt,name=log_id,json=logId" json:"log_id,omitempty"` + LogId string `protobuf:"bytes,2,opt,name=log_id,json=logId,proto3" json:"log_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -347,7 +347,7 @@ func (m *AddTaskLogRequest) GetLogId() string { // Response message for `AddTaskLog`. type AddTaskLogResponse struct { // The handle for the new log, as would be returned in Task.logs. - Handle string `protobuf:"bytes,1,opt,name=handle" json:"handle,omitempty"` + Handle string `protobuf:"bytes,1,opt,name=handle,proto3" json:"handle,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -454,8 +454,7 @@ func (c *tasksClient) AddTaskLog(ctx context.Context, in *AddTaskLogRequest, opt return out, nil } -// Server API for Tasks service - +// TasksServer is the server API for Tasks service. type TasksServer interface { // GetTask reads the current state of the task. Tasks must be created through // some other interface, and should be immutable once created and exposed to diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/worker.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/worker.pb.go index 7bd386963..58ea4c197 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/worker.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/remoteworkers/v1test2/worker.pb.go @@ -56,7 +56,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Worker struct { // A list of devices; the first device is the primary device. See the `Device` // message for more information. - Devices []*Device `protobuf:"bytes,1,rep,name=devices" json:"devices,omitempty"` + Devices []*Device `protobuf:"bytes,1,rep,name=devices,proto3" json:"devices,omitempty"` // A worker may contain "global" properties. For example, certain machines // might be reserved for certain types of jobs, like short-running compilation // versus long-running integration tests. This property is known as a "pool" @@ -64,7 +64,7 @@ type Worker struct { // to the worker as a whole. // // The behaviour of repeated keys is identical to that of Device.Property. - Properties []*Worker_Property `protobuf:"bytes,2,rep,name=properties" json:"properties,omitempty"` + Properties []*Worker_Property `protobuf:"bytes,2,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -120,9 +120,9 @@ type Worker_Property struct { // throughput. To support this, the server can assign different values for // `pool` (such as "itest" and "utest") to different workers, and then have // jobs request workers from those pools. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // The property's value. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -179,7 +179,7 @@ type Device struct { // logical name expected by the task to a machine-readable name that the task // can actually use, such as a USB address. The method by which this mapping // is communicated to the task is not covered in this API. - Handle string `protobuf:"bytes,1,opt,name=handle" json:"handle,omitempty"` + Handle string `protobuf:"bytes,1,opt,name=handle,proto3" json:"handle,omitempty"` // Properties of this device that don't change based on the tasks that are // running on it, e.g. OS, CPU architecture, etc. // @@ -189,7 +189,7 @@ type Device struct { // example, an "ISA" property might include "x86", "x86-64" and "sse4". // // * Request context: the device *must* support *all* of the listed values. - Properties []*Device_Property `protobuf:"bytes,2,rep,name=properties" json:"properties,omitempty"` + Properties []*Device_Property `protobuf:"bytes,2,rep,name=properties,proto3" json:"properties,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -246,9 +246,9 @@ type Device_Property struct { // // * has-docker: "true" if the bot has Docker installed. This will be // replaced in the future by a more structured message for Docker support. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // The property's value. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/action.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/action.pb.go new file mode 100644 index 000000000..b1419277f --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/action.pb.go @@ -0,0 +1,1446 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/action.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Indicates how/where this Action was executed. +type ExecutionStrategy int32 + +const ( + // The action did not indicate how it was executed. + ExecutionStrategy_EXECUTION_STRATEGY_UNSPECIFIED ExecutionStrategy = 0 + // The action was executed in some other form. + ExecutionStrategy_OTHER_ENVIRONMENT ExecutionStrategy = 1 + // The action used a remote build service. + ExecutionStrategy_REMOTE_SERVICE ExecutionStrategy = 2 + // The action was executed locally, in parallel with other actions. + ExecutionStrategy_LOCAL_PARALLEL ExecutionStrategy = 3 + // The action was executed locally, without parallelism. + ExecutionStrategy_LOCAL_SEQUENTIAL ExecutionStrategy = 4 +) + +var ExecutionStrategy_name = map[int32]string{ + 0: "EXECUTION_STRATEGY_UNSPECIFIED", + 1: "OTHER_ENVIRONMENT", + 2: "REMOTE_SERVICE", + 3: "LOCAL_PARALLEL", + 4: "LOCAL_SEQUENTIAL", +} +var ExecutionStrategy_value = map[string]int32{ + "EXECUTION_STRATEGY_UNSPECIFIED": 0, + "OTHER_ENVIRONMENT": 1, + "REMOTE_SERVICE": 2, + "LOCAL_PARALLEL": 3, + "LOCAL_SEQUENTIAL": 4, +} + +func (x ExecutionStrategy) String() string { + return proto.EnumName(ExecutionStrategy_name, int32(x)) +} +func (ExecutionStrategy) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{0} +} + +// Most build systems cache build results to speed up incremental builds. +// Some also cache test results too. This indicates whether the test results +// were found in a cache, and where that cache was located. +type TestCaching int32 + +const ( + // The implicit default enum value. Should never be set. + TestCaching_TEST_CACHING_UNSPECIFIED TestCaching = 0 + // The test result was found in a local cache, so it wasn't run again. + TestCaching_LOCAL_CACHE_HIT TestCaching = 1 + // The test result was found in a remote cache, so it wasn't run again. + TestCaching_REMOTE_CACHE_HIT TestCaching = 2 + // The test result was not found in any cache, so it had to be run again. + TestCaching_CACHE_MISS TestCaching = 3 +) + +var TestCaching_name = map[int32]string{ + 0: "TEST_CACHING_UNSPECIFIED", + 1: "LOCAL_CACHE_HIT", + 2: "REMOTE_CACHE_HIT", + 3: "CACHE_MISS", +} +var TestCaching_value = map[string]int32{ + "TEST_CACHING_UNSPECIFIED": 0, + "LOCAL_CACHE_HIT": 1, + "REMOTE_CACHE_HIT": 2, + "CACHE_MISS": 3, +} + +func (x TestCaching) String() string { + return proto.EnumName(TestCaching_name, int32(x)) +} +func (TestCaching) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{1} +} + +// Errors in file post-processing are categorized using this enum. +type FileProcessingErrorType int32 + +const ( + // Type unspecified or not listed here. + FileProcessingErrorType_FILE_PROCESSING_ERROR_TYPE_UNSPECIFIED FileProcessingErrorType = 0 + // A read error occurred trying to read the file. + FileProcessingErrorType_GENERIC_READ_ERROR FileProcessingErrorType = 1 + // There was an error trying to parse the file. + FileProcessingErrorType_GENERIC_PARSE_ERROR FileProcessingErrorType = 2 + // File is exceeds size limit. + FileProcessingErrorType_FILE_TOO_LARGE FileProcessingErrorType = 3 + // The result of parsing the file exceeded size limit. + FileProcessingErrorType_OUTPUT_TOO_LARGE FileProcessingErrorType = 4 + // Read access to the file was denied by file system. + FileProcessingErrorType_ACCESS_DENIED FileProcessingErrorType = 5 + // Deadline exceeded trying to read the file. + FileProcessingErrorType_DEADLINE_EXCEEDED FileProcessingErrorType = 6 + // File not found. + FileProcessingErrorType_NOT_FOUND FileProcessingErrorType = 7 +) + +var FileProcessingErrorType_name = map[int32]string{ + 0: "FILE_PROCESSING_ERROR_TYPE_UNSPECIFIED", + 1: "GENERIC_READ_ERROR", + 2: "GENERIC_PARSE_ERROR", + 3: "FILE_TOO_LARGE", + 4: "OUTPUT_TOO_LARGE", + 5: "ACCESS_DENIED", + 6: "DEADLINE_EXCEEDED", + 7: "NOT_FOUND", +} +var FileProcessingErrorType_value = map[string]int32{ + "FILE_PROCESSING_ERROR_TYPE_UNSPECIFIED": 0, + "GENERIC_READ_ERROR": 1, + "GENERIC_PARSE_ERROR": 2, + "FILE_TOO_LARGE": 3, + "OUTPUT_TOO_LARGE": 4, + "ACCESS_DENIED": 5, + "DEADLINE_EXCEEDED": 6, + "NOT_FOUND": 7, +} + +func (x FileProcessingErrorType) String() string { + return proto.EnumName(FileProcessingErrorType_name, int32(x)) +} +func (FileProcessingErrorType) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{2} +} + +// An action that happened as part of a configured target. This action could be +// a build, a test, or another type of action. +// Each parent ConfiguredTarget resource should have at least one Action as its +// child resource before the invocation is finalized. ResultStore is a tool to +// store build & test results. ConfiguredTarget proto by itself does not contain +// enough fields to fully represent such results. For a simple build, at least +// one build action is required to represent the build result. +type Action struct { + // The resource name. Its format must be: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID}/configuredTargets/${CONFIG_ID}/actions/${ACTION_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The resource ID components that identify the Action. They must match the + // resource name after proper encoding. + Id *Action_Id `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` + // The status of the action. + StatusAttributes *StatusAttributes `protobuf:"bytes,3,opt,name=status_attributes,json=statusAttributes,proto3" json:"status_attributes,omitempty"` + // The timing of the whole action. For TestActions, the start time may be + // before the test actually started, and the duration may last until after the + // test actually finished. + Timing *Timing `protobuf:"bytes,4,opt,name=timing,proto3" json:"timing,omitempty"` + // The type of the action. The type of an action may not change over the + // lifetime of the invocation. If one of these fields is to be set, it must be + // set in the CreateAction method. It may be set to an empty message that is + // populated in later methods or post-processing. A generic "untyped" action + // can be created by not setting any of these fields. An untyped action will + // be untyped for the lifetime of the invocation. + // + // Types that are valid to be assigned to ActionType: + // *Action_BuildAction + // *Action_TestAction + ActionType isAction_ActionType `protobuf_oneof:"action_type"` + // General attributes of the action. + ActionAttributes *ActionAttributes `protobuf:"bytes,5,opt,name=action_attributes,json=actionAttributes,proto3" json:"action_attributes,omitempty"` + // A list of resources that this action depended upon. May be used to provide + // the cause of a build failure in the case of a failed build action. + ActionDependencies []*Dependency `protobuf:"bytes,14,rep,name=action_dependencies,json=actionDependencies,proto3" json:"action_dependencies,omitempty"` + // Arbitrary name-value pairs. + // This is implemented as a multi-map. Multiple properties are allowed with + // the same key. Properties will be returned in lexicographical order by key. + Properties []*Property `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty"` + // A list of file references for action level files. + // The file IDs must be unique within this list. Duplicate file IDs will + // result in an error. Files will be returned in lexicographical order by ID. + // Files with the following reserved file IDs cause specific post-processing + // or have special handling: + // + // For build actions: + // stdout: The stdout of the action + // stderr: The stderr of the action + // baseline.lcov: Baseline coverage file to be parsed by the server. This + // uses a stripped down implementation of the LCOV standard. + // http://ltp.sourceforge.net/coverage/lcov/geninfo.1.php + // + // For test actions: + // test.xml: The test suite / test case data in XML format. + // test.log: The combined stdout and stderr of the test process. + // test.lcov: Coverage file to be parsed by the server. This uses a stripped + // down implementation of the LCOV standard. + // http://ltp.sourceforge.net/coverage/lcov/geninfo.1.php + Files []*File `protobuf:"bytes,8,rep,name=files,proto3" json:"files,omitempty"` + // Coverage data was collected while running the build or test action. This + // usually includes line coverage, and may also include branch coverage. + // For test actions, this is usually only for the source files which were + // actually executed by that particular action. + // For build actions, this is the baseline coverage, which captures the + // instrumented files and lines, without any lines being executed. This + // ensures files that are never covered at all are included. + Coverage *ActionCoverage `protobuf:"bytes,11,opt,name=coverage,proto3" json:"coverage,omitempty"` + // Summary of aggregate coverage across all files for this action in. The + // server populates this for you in the post-processing phase. + CoverageSummaries []*LanguageCoverageSummary `protobuf:"bytes,12,rep,name=coverage_summaries,json=coverageSummaries,proto3" json:"coverage_summaries,omitempty"` + // ResultStore will read and parse Files with reserved IDs listed above. Read + // and parse errors for all these Files are reported here. + // This is implemented as a map, with one FileProcessingErrors for each file. + // Typically produced when parsing Files, but may also be provided directly + // by clients. + FileProcessingErrors []*FileProcessingErrors `protobuf:"bytes,13,rep,name=file_processing_errors,json=fileProcessingErrors,proto3" json:"file_processing_errors,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Action) Reset() { *m = Action{} } +func (m *Action) String() string { return proto.CompactTextString(m) } +func (*Action) ProtoMessage() {} +func (*Action) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{0} +} +func (m *Action) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Action.Unmarshal(m, b) +} +func (m *Action) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Action.Marshal(b, m, deterministic) +} +func (dst *Action) XXX_Merge(src proto.Message) { + xxx_messageInfo_Action.Merge(dst, src) +} +func (m *Action) XXX_Size() int { + return xxx_messageInfo_Action.Size(m) +} +func (m *Action) XXX_DiscardUnknown() { + xxx_messageInfo_Action.DiscardUnknown(m) +} + +var xxx_messageInfo_Action proto.InternalMessageInfo + +type isAction_ActionType interface { + isAction_ActionType() +} + +type Action_BuildAction struct { + BuildAction *BuildAction `protobuf:"bytes,9,opt,name=build_action,json=buildAction,proto3,oneof"` +} +type Action_TestAction struct { + TestAction *TestAction `protobuf:"bytes,10,opt,name=test_action,json=testAction,proto3,oneof"` +} + +func (*Action_BuildAction) isAction_ActionType() {} +func (*Action_TestAction) isAction_ActionType() {} + +func (m *Action) GetActionType() isAction_ActionType { + if m != nil { + return m.ActionType + } + return nil +} + +func (m *Action) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Action) GetId() *Action_Id { + if m != nil { + return m.Id + } + return nil +} + +func (m *Action) GetStatusAttributes() *StatusAttributes { + if m != nil { + return m.StatusAttributes + } + return nil +} + +func (m *Action) GetTiming() *Timing { + if m != nil { + return m.Timing + } + return nil +} + +func (m *Action) GetBuildAction() *BuildAction { + if x, ok := m.GetActionType().(*Action_BuildAction); ok { + return x.BuildAction + } + return nil +} + +func (m *Action) GetTestAction() *TestAction { + if x, ok := m.GetActionType().(*Action_TestAction); ok { + return x.TestAction + } + return nil +} + +func (m *Action) GetActionAttributes() *ActionAttributes { + if m != nil { + return m.ActionAttributes + } + return nil +} + +func (m *Action) GetActionDependencies() []*Dependency { + if m != nil { + return m.ActionDependencies + } + return nil +} + +func (m *Action) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +func (m *Action) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +func (m *Action) GetCoverage() *ActionCoverage { + if m != nil { + return m.Coverage + } + return nil +} + +func (m *Action) GetCoverageSummaries() []*LanguageCoverageSummary { + if m != nil { + return m.CoverageSummaries + } + return nil +} + +func (m *Action) GetFileProcessingErrors() []*FileProcessingErrors { + if m != nil { + return m.FileProcessingErrors + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Action) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Action_OneofMarshaler, _Action_OneofUnmarshaler, _Action_OneofSizer, []interface{}{ + (*Action_BuildAction)(nil), + (*Action_TestAction)(nil), + } +} + +func _Action_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Action) + // action_type + switch x := m.ActionType.(type) { + case *Action_BuildAction: + b.EncodeVarint(9<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.BuildAction); err != nil { + return err + } + case *Action_TestAction: + b.EncodeVarint(10<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.TestAction); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("Action.ActionType has unexpected type %T", x) + } + return nil +} + +func _Action_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Action) + switch tag { + case 9: // action_type.build_action + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(BuildAction) + err := b.DecodeMessage(msg) + m.ActionType = &Action_BuildAction{msg} + return true, err + case 10: // action_type.test_action + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(TestAction) + err := b.DecodeMessage(msg) + m.ActionType = &Action_TestAction{msg} + return true, err + default: + return false, nil + } +} + +func _Action_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Action) + // action_type + switch x := m.ActionType.(type) { + case *Action_BuildAction: + s := proto.Size(x.BuildAction) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *Action_TestAction: + s := proto.Size(x.TestAction) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// The resource ID components that identify the Action. +type Action_Id struct { + // The Invocation ID. + InvocationId string `protobuf:"bytes,1,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` + // The Target ID. + TargetId string `protobuf:"bytes,2,opt,name=target_id,json=targetId,proto3" json:"target_id,omitempty"` + // The Configuration ID. + ConfigurationId string `protobuf:"bytes,3,opt,name=configuration_id,json=configurationId,proto3" json:"configuration_id,omitempty"` + // The Action ID. + ActionId string `protobuf:"bytes,4,opt,name=action_id,json=actionId,proto3" json:"action_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Action_Id) Reset() { *m = Action_Id{} } +func (m *Action_Id) String() string { return proto.CompactTextString(m) } +func (*Action_Id) ProtoMessage() {} +func (*Action_Id) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{0, 0} +} +func (m *Action_Id) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Action_Id.Unmarshal(m, b) +} +func (m *Action_Id) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Action_Id.Marshal(b, m, deterministic) +} +func (dst *Action_Id) XXX_Merge(src proto.Message) { + xxx_messageInfo_Action_Id.Merge(dst, src) +} +func (m *Action_Id) XXX_Size() int { + return xxx_messageInfo_Action_Id.Size(m) +} +func (m *Action_Id) XXX_DiscardUnknown() { + xxx_messageInfo_Action_Id.DiscardUnknown(m) +} + +var xxx_messageInfo_Action_Id proto.InternalMessageInfo + +func (m *Action_Id) GetInvocationId() string { + if m != nil { + return m.InvocationId + } + return "" +} + +func (m *Action_Id) GetTargetId() string { + if m != nil { + return m.TargetId + } + return "" +} + +func (m *Action_Id) GetConfigurationId() string { + if m != nil { + return m.ConfigurationId + } + return "" +} + +func (m *Action_Id) GetActionId() string { + if m != nil { + return m.ActionId + } + return "" +} + +// A build action, such as building a java library. +type BuildAction struct { + // The type of the action. This is intended to be a clue as to how the output + // of the action should be parsed. For example "javac" for a Java compile + // action. + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` + // The "primary" input artifact processed by this action. E.g., the .cc file + // of a C++ compile action. Empty string ("") if the action has no input + // artifacts or no "primary" input artifact. + PrimaryInputPath string `protobuf:"bytes,2,opt,name=primary_input_path,json=primaryInputPath,proto3" json:"primary_input_path,omitempty"` + // The "primary" output artifact processed by this action. E.g., the .o file + // of a C++ compile action. Empty string ("") if the action has no output + // artifacts or no "primary" output artifact. + PrimaryOutputPath string `protobuf:"bytes,3,opt,name=primary_output_path,json=primaryOutputPath,proto3" json:"primary_output_path,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *BuildAction) Reset() { *m = BuildAction{} } +func (m *BuildAction) String() string { return proto.CompactTextString(m) } +func (*BuildAction) ProtoMessage() {} +func (*BuildAction) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{1} +} +func (m *BuildAction) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_BuildAction.Unmarshal(m, b) +} +func (m *BuildAction) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_BuildAction.Marshal(b, m, deterministic) +} +func (dst *BuildAction) XXX_Merge(src proto.Message) { + xxx_messageInfo_BuildAction.Merge(dst, src) +} +func (m *BuildAction) XXX_Size() int { + return xxx_messageInfo_BuildAction.Size(m) +} +func (m *BuildAction) XXX_DiscardUnknown() { + xxx_messageInfo_BuildAction.DiscardUnknown(m) +} + +var xxx_messageInfo_BuildAction proto.InternalMessageInfo + +func (m *BuildAction) GetType() string { + if m != nil { + return m.Type + } + return "" +} + +func (m *BuildAction) GetPrimaryInputPath() string { + if m != nil { + return m.PrimaryInputPath + } + return "" +} + +func (m *BuildAction) GetPrimaryOutputPath() string { + if m != nil { + return m.PrimaryOutputPath + } + return "" +} + +// A test action, such as running a JUnit4 test binary. +type TestAction struct { + // Timing data for execution of the test action. + TestTiming *TestTiming `protobuf:"bytes,1,opt,name=test_timing,json=testTiming,proto3" json:"test_timing,omitempty"` + // If the test is divided up into shards to improve performance, set this to + // indicate which shard this test action is for. Value must be in interval + // [0, total_shard_count). Defaults to 0, which is appropriate if all test + // cases are run in the same process. + ShardNumber int32 `protobuf:"varint,2,opt,name=shard_number,json=shardNumber,proto3" json:"shard_number,omitempty"` + // If the user requested that every test be run multiple times, as is often + // done to measure flakiness, set this to indicate which run this test action + // is for. Value must be in interval [0, total_run_count). Defaults to 0, + // which is appropriate if multiple runs were not requested. + RunNumber int32 `protobuf:"varint,3,opt,name=run_number,json=runNumber,proto3" json:"run_number,omitempty"` + // If flaky tests are automatically retried, set this to indicate which + // attempt this test action is for. (e.g. 0 for the first attempt, 1 for + // second, and so on). Defaults to 0, which is appropriate if this is only + // attempt. + AttemptNumber int32 `protobuf:"varint,4,opt,name=attempt_number,json=attemptNumber,proto3" json:"attempt_number,omitempty"` + // A tree of test suites and test cases that were run by this test action. + // Each test case has its own status information, including stack traces. + // Typically produced by parsing an XML Log, but may also be provided directly + // by clients. + TestSuite *TestSuite `protobuf:"bytes,5,opt,name=test_suite,json=testSuite,proto3" json:"test_suite,omitempty"` + // Warnings for this test action. + Warnings []*TestWarning `protobuf:"bytes,8,rep,name=warnings,proto3" json:"warnings,omitempty"` + // Estimated memory consumption of the test action, in bytes. A default value + // of 0 means there is no memory consumption estimate specified. + EstimatedMemoryBytes int64 `protobuf:"varint,10,opt,name=estimated_memory_bytes,json=estimatedMemoryBytes,proto3" json:"estimated_memory_bytes,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestAction) Reset() { *m = TestAction{} } +func (m *TestAction) String() string { return proto.CompactTextString(m) } +func (*TestAction) ProtoMessage() {} +func (*TestAction) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{2} +} +func (m *TestAction) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestAction.Unmarshal(m, b) +} +func (m *TestAction) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestAction.Marshal(b, m, deterministic) +} +func (dst *TestAction) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestAction.Merge(dst, src) +} +func (m *TestAction) XXX_Size() int { + return xxx_messageInfo_TestAction.Size(m) +} +func (m *TestAction) XXX_DiscardUnknown() { + xxx_messageInfo_TestAction.DiscardUnknown(m) +} + +var xxx_messageInfo_TestAction proto.InternalMessageInfo + +func (m *TestAction) GetTestTiming() *TestTiming { + if m != nil { + return m.TestTiming + } + return nil +} + +func (m *TestAction) GetShardNumber() int32 { + if m != nil { + return m.ShardNumber + } + return 0 +} + +func (m *TestAction) GetRunNumber() int32 { + if m != nil { + return m.RunNumber + } + return 0 +} + +func (m *TestAction) GetAttemptNumber() int32 { + if m != nil { + return m.AttemptNumber + } + return 0 +} + +func (m *TestAction) GetTestSuite() *TestSuite { + if m != nil { + return m.TestSuite + } + return nil +} + +func (m *TestAction) GetWarnings() []*TestWarning { + if m != nil { + return m.Warnings + } + return nil +} + +func (m *TestAction) GetEstimatedMemoryBytes() int64 { + if m != nil { + return m.EstimatedMemoryBytes + } + return 0 +} + +// General attributes of an action +type ActionAttributes struct { + // Strategy used for executing the action. + ExecutionStrategy ExecutionStrategy `protobuf:"varint,1,opt,name=execution_strategy,json=executionStrategy,proto3,enum=google.devtools.resultstore.v2.ExecutionStrategy" json:"execution_strategy,omitempty"` + // Exit code of the process that ran the action. A non-zero value means + // failure. + ExitCode int32 `protobuf:"varint,2,opt,name=exit_code,json=exitCode,proto3" json:"exit_code,omitempty"` + // Where the action was run. + Hostname string `protobuf:"bytes,3,opt,name=hostname,proto3" json:"hostname,omitempty"` + // Information about the input files used in all actions under this configured + // target. + InputFileInfo *InputFileInfo `protobuf:"bytes,4,opt,name=input_file_info,json=inputFileInfo,proto3" json:"input_file_info,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ActionAttributes) Reset() { *m = ActionAttributes{} } +func (m *ActionAttributes) String() string { return proto.CompactTextString(m) } +func (*ActionAttributes) ProtoMessage() {} +func (*ActionAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{3} +} +func (m *ActionAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ActionAttributes.Unmarshal(m, b) +} +func (m *ActionAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ActionAttributes.Marshal(b, m, deterministic) +} +func (dst *ActionAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_ActionAttributes.Merge(dst, src) +} +func (m *ActionAttributes) XXX_Size() int { + return xxx_messageInfo_ActionAttributes.Size(m) +} +func (m *ActionAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_ActionAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_ActionAttributes proto.InternalMessageInfo + +func (m *ActionAttributes) GetExecutionStrategy() ExecutionStrategy { + if m != nil { + return m.ExecutionStrategy + } + return ExecutionStrategy_EXECUTION_STRATEGY_UNSPECIFIED +} + +func (m *ActionAttributes) GetExitCode() int32 { + if m != nil { + return m.ExitCode + } + return 0 +} + +func (m *ActionAttributes) GetHostname() string { + if m != nil { + return m.Hostname + } + return "" +} + +func (m *ActionAttributes) GetInputFileInfo() *InputFileInfo { + if m != nil { + return m.InputFileInfo + } + return nil +} + +// File count and size information for the input files to a configured target. +type InputFileInfo struct { + // The number of input files (counting every file, even if a duplicate). + Count int64 `protobuf:"varint,1,opt,name=count,proto3" json:"count,omitempty"` + // The number of distinct input files. + DistinctCount int64 `protobuf:"varint,2,opt,name=distinct_count,json=distinctCount,proto3" json:"distinct_count,omitempty"` + // The max number of input files allowed by the build system (counting every + // file, even if a duplicate). + CountLimit int64 `protobuf:"varint,3,opt,name=count_limit,json=countLimit,proto3" json:"count_limit,omitempty"` + // The total size of the distinct input files. + DistinctBytes int64 `protobuf:"varint,4,opt,name=distinct_bytes,json=distinctBytes,proto3" json:"distinct_bytes,omitempty"` + // The max allowed total size of the distinct input files. + DistinctByteLimit int64 `protobuf:"varint,5,opt,name=distinct_byte_limit,json=distinctByteLimit,proto3" json:"distinct_byte_limit,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *InputFileInfo) Reset() { *m = InputFileInfo{} } +func (m *InputFileInfo) String() string { return proto.CompactTextString(m) } +func (*InputFileInfo) ProtoMessage() {} +func (*InputFileInfo) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{4} +} +func (m *InputFileInfo) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_InputFileInfo.Unmarshal(m, b) +} +func (m *InputFileInfo) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_InputFileInfo.Marshal(b, m, deterministic) +} +func (dst *InputFileInfo) XXX_Merge(src proto.Message) { + xxx_messageInfo_InputFileInfo.Merge(dst, src) +} +func (m *InputFileInfo) XXX_Size() int { + return xxx_messageInfo_InputFileInfo.Size(m) +} +func (m *InputFileInfo) XXX_DiscardUnknown() { + xxx_messageInfo_InputFileInfo.DiscardUnknown(m) +} + +var xxx_messageInfo_InputFileInfo proto.InternalMessageInfo + +func (m *InputFileInfo) GetCount() int64 { + if m != nil { + return m.Count + } + return 0 +} + +func (m *InputFileInfo) GetDistinctCount() int64 { + if m != nil { + return m.DistinctCount + } + return 0 +} + +func (m *InputFileInfo) GetCountLimit() int64 { + if m != nil { + return m.CountLimit + } + return 0 +} + +func (m *InputFileInfo) GetDistinctBytes() int64 { + if m != nil { + return m.DistinctBytes + } + return 0 +} + +func (m *InputFileInfo) GetDistinctByteLimit() int64 { + if m != nil { + return m.DistinctByteLimit + } + return 0 +} + +// Timing data for tests executed locally on the machine running the build. +type LocalTestTiming struct { + // Time taken by the test process, typically surrounded by a small wrapper + // script. + TestProcessDuration *duration.Duration `protobuf:"bytes,1,opt,name=test_process_duration,json=testProcessDuration,proto3" json:"test_process_duration,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LocalTestTiming) Reset() { *m = LocalTestTiming{} } +func (m *LocalTestTiming) String() string { return proto.CompactTextString(m) } +func (*LocalTestTiming) ProtoMessage() {} +func (*LocalTestTiming) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{5} +} +func (m *LocalTestTiming) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LocalTestTiming.Unmarshal(m, b) +} +func (m *LocalTestTiming) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LocalTestTiming.Marshal(b, m, deterministic) +} +func (dst *LocalTestTiming) XXX_Merge(src proto.Message) { + xxx_messageInfo_LocalTestTiming.Merge(dst, src) +} +func (m *LocalTestTiming) XXX_Size() int { + return xxx_messageInfo_LocalTestTiming.Size(m) +} +func (m *LocalTestTiming) XXX_DiscardUnknown() { + xxx_messageInfo_LocalTestTiming.DiscardUnknown(m) +} + +var xxx_messageInfo_LocalTestTiming proto.InternalMessageInfo + +func (m *LocalTestTiming) GetTestProcessDuration() *duration.Duration { + if m != nil { + return m.TestProcessDuration + } + return nil +} + +// Timing data for one attempt to execute a test action remotely. +type RemoteTestAttemptTiming struct { + // Idle period before the test process is invoked on the remote machine. + QueueDuration *duration.Duration `protobuf:"bytes,1,opt,name=queue_duration,json=queueDuration,proto3" json:"queue_duration,omitempty"` + // Time to upload data dependencies from the local machine to the remote + // machine running the test, or to the distributed cache. + UploadDuration *duration.Duration `protobuf:"bytes,2,opt,name=upload_duration,json=uploadDuration,proto3" json:"upload_duration,omitempty"` + // Time to set up the remote machine. + // Not to be confused with setup time in + // xUnit test frameworks, which falls within the test_process_time. + MachineSetupDuration *duration.Duration `protobuf:"bytes,3,opt,name=machine_setup_duration,json=machineSetupDuration,proto3" json:"machine_setup_duration,omitempty"` + // Time taken by the test process, typically surrounded by a small wrapper + // script. + // For Java tests, this includes JVM setup, flag parsing, class path setup, + // parsing files to setup the suite, and finally running your test methods. + // In many cases, only a small fraction of the test process time is spent + // running the test methods. + TestProcessDuration *duration.Duration `protobuf:"bytes,4,opt,name=test_process_duration,json=testProcessDuration,proto3" json:"test_process_duration,omitempty"` + // Time spent retrieving test logs and any other test outputs, back to the + // local machine. + DownloadDuration *duration.Duration `protobuf:"bytes,5,opt,name=download_duration,json=downloadDuration,proto3" json:"download_duration,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RemoteTestAttemptTiming) Reset() { *m = RemoteTestAttemptTiming{} } +func (m *RemoteTestAttemptTiming) String() string { return proto.CompactTextString(m) } +func (*RemoteTestAttemptTiming) ProtoMessage() {} +func (*RemoteTestAttemptTiming) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{6} +} +func (m *RemoteTestAttemptTiming) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RemoteTestAttemptTiming.Unmarshal(m, b) +} +func (m *RemoteTestAttemptTiming) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RemoteTestAttemptTiming.Marshal(b, m, deterministic) +} +func (dst *RemoteTestAttemptTiming) XXX_Merge(src proto.Message) { + xxx_messageInfo_RemoteTestAttemptTiming.Merge(dst, src) +} +func (m *RemoteTestAttemptTiming) XXX_Size() int { + return xxx_messageInfo_RemoteTestAttemptTiming.Size(m) +} +func (m *RemoteTestAttemptTiming) XXX_DiscardUnknown() { + xxx_messageInfo_RemoteTestAttemptTiming.DiscardUnknown(m) +} + +var xxx_messageInfo_RemoteTestAttemptTiming proto.InternalMessageInfo + +func (m *RemoteTestAttemptTiming) GetQueueDuration() *duration.Duration { + if m != nil { + return m.QueueDuration + } + return nil +} + +func (m *RemoteTestAttemptTiming) GetUploadDuration() *duration.Duration { + if m != nil { + return m.UploadDuration + } + return nil +} + +func (m *RemoteTestAttemptTiming) GetMachineSetupDuration() *duration.Duration { + if m != nil { + return m.MachineSetupDuration + } + return nil +} + +func (m *RemoteTestAttemptTiming) GetTestProcessDuration() *duration.Duration { + if m != nil { + return m.TestProcessDuration + } + return nil +} + +func (m *RemoteTestAttemptTiming) GetDownloadDuration() *duration.Duration { + if m != nil { + return m.DownloadDuration + } + return nil +} + +// Timing data for the part of the test execution that is done remotely. +type RemoteTestTiming struct { + // Time taken locally to determine what to do. + LocalAnalysisDuration *duration.Duration `protobuf:"bytes,1,opt,name=local_analysis_duration,json=localAnalysisDuration,proto3" json:"local_analysis_duration,omitempty"` + // Normally there is only one attempt, but the system may retry on internal + // errors, leading to multiple attempts. + Attempts []*RemoteTestAttemptTiming `protobuf:"bytes,2,rep,name=attempts,proto3" json:"attempts,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *RemoteTestTiming) Reset() { *m = RemoteTestTiming{} } +func (m *RemoteTestTiming) String() string { return proto.CompactTextString(m) } +func (*RemoteTestTiming) ProtoMessage() {} +func (*RemoteTestTiming) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{7} +} +func (m *RemoteTestTiming) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RemoteTestTiming.Unmarshal(m, b) +} +func (m *RemoteTestTiming) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RemoteTestTiming.Marshal(b, m, deterministic) +} +func (dst *RemoteTestTiming) XXX_Merge(src proto.Message) { + xxx_messageInfo_RemoteTestTiming.Merge(dst, src) +} +func (m *RemoteTestTiming) XXX_Size() int { + return xxx_messageInfo_RemoteTestTiming.Size(m) +} +func (m *RemoteTestTiming) XXX_DiscardUnknown() { + xxx_messageInfo_RemoteTestTiming.DiscardUnknown(m) +} + +var xxx_messageInfo_RemoteTestTiming proto.InternalMessageInfo + +func (m *RemoteTestTiming) GetLocalAnalysisDuration() *duration.Duration { + if m != nil { + return m.LocalAnalysisDuration + } + return nil +} + +func (m *RemoteTestTiming) GetAttempts() []*RemoteTestAttemptTiming { + if m != nil { + return m.Attempts + } + return nil +} + +// Timing data for execution of a test action. The action may be performed +// locally, on the machine running the build, or remotely. +type TestTiming struct { + // Test timing for either a local or remote execution. + // + // Types that are valid to be assigned to Location: + // *TestTiming_Local + // *TestTiming_Remote + Location isTestTiming_Location `protobuf_oneof:"location"` + // The amount of CPU time spent by the test process executing system calls + // within the kernel, as opposed to library code. Time the test process spent + // blocked does not count towards this figure. + SystemTimeDuration *duration.Duration `protobuf:"bytes,3,opt,name=system_time_duration,json=systemTimeDuration,proto3" json:"system_time_duration,omitempty"` + // The amount of CPU time spent by the test process executing user-mode code + // outside the kernel, as opposed to library code. Time the test process + // spent blocked does not count towards this figure. You can add user_time to + // system_time to get total CPU time taken by the test process. + UserTimeDuration *duration.Duration `protobuf:"bytes,4,opt,name=user_time_duration,json=userTimeDuration,proto3" json:"user_time_duration,omitempty"` + // Most build systems cache build results to speed up incremental builds. + // Some also cache test results too. This indicates whether the test results + // were found in a cache, and where that cache was located. + TestCaching TestCaching `protobuf:"varint,5,opt,name=test_caching,json=testCaching,proto3,enum=google.devtools.resultstore.v2.TestCaching" json:"test_caching,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestTiming) Reset() { *m = TestTiming{} } +func (m *TestTiming) String() string { return proto.CompactTextString(m) } +func (*TestTiming) ProtoMessage() {} +func (*TestTiming) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{8} +} +func (m *TestTiming) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestTiming.Unmarshal(m, b) +} +func (m *TestTiming) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestTiming.Marshal(b, m, deterministic) +} +func (dst *TestTiming) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestTiming.Merge(dst, src) +} +func (m *TestTiming) XXX_Size() int { + return xxx_messageInfo_TestTiming.Size(m) +} +func (m *TestTiming) XXX_DiscardUnknown() { + xxx_messageInfo_TestTiming.DiscardUnknown(m) +} + +var xxx_messageInfo_TestTiming proto.InternalMessageInfo + +type isTestTiming_Location interface { + isTestTiming_Location() +} + +type TestTiming_Local struct { + Local *LocalTestTiming `protobuf:"bytes,1,opt,name=local,proto3,oneof"` +} +type TestTiming_Remote struct { + Remote *RemoteTestTiming `protobuf:"bytes,2,opt,name=remote,proto3,oneof"` +} + +func (*TestTiming_Local) isTestTiming_Location() {} +func (*TestTiming_Remote) isTestTiming_Location() {} + +func (m *TestTiming) GetLocation() isTestTiming_Location { + if m != nil { + return m.Location + } + return nil +} + +func (m *TestTiming) GetLocal() *LocalTestTiming { + if x, ok := m.GetLocation().(*TestTiming_Local); ok { + return x.Local + } + return nil +} + +func (m *TestTiming) GetRemote() *RemoteTestTiming { + if x, ok := m.GetLocation().(*TestTiming_Remote); ok { + return x.Remote + } + return nil +} + +func (m *TestTiming) GetSystemTimeDuration() *duration.Duration { + if m != nil { + return m.SystemTimeDuration + } + return nil +} + +func (m *TestTiming) GetUserTimeDuration() *duration.Duration { + if m != nil { + return m.UserTimeDuration + } + return nil +} + +func (m *TestTiming) GetTestCaching() TestCaching { + if m != nil { + return m.TestCaching + } + return TestCaching_TEST_CACHING_UNSPECIFIED +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*TestTiming) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _TestTiming_OneofMarshaler, _TestTiming_OneofUnmarshaler, _TestTiming_OneofSizer, []interface{}{ + (*TestTiming_Local)(nil), + (*TestTiming_Remote)(nil), + } +} + +func _TestTiming_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*TestTiming) + // location + switch x := m.Location.(type) { + case *TestTiming_Local: + b.EncodeVarint(1<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.Local); err != nil { + return err + } + case *TestTiming_Remote: + b.EncodeVarint(2<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.Remote); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("TestTiming.Location has unexpected type %T", x) + } + return nil +} + +func _TestTiming_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*TestTiming) + switch tag { + case 1: // location.local + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(LocalTestTiming) + err := b.DecodeMessage(msg) + m.Location = &TestTiming_Local{msg} + return true, err + case 2: // location.remote + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(RemoteTestTiming) + err := b.DecodeMessage(msg) + m.Location = &TestTiming_Remote{msg} + return true, err + default: + return false, nil + } +} + +func _TestTiming_OneofSizer(msg proto.Message) (n int) { + m := msg.(*TestTiming) + // location + switch x := m.Location.(type) { + case *TestTiming_Local: + s := proto.Size(x.Local) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *TestTiming_Remote: + s := proto.Size(x.Remote) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// A warning from a test execution. +type TestWarning struct { + // Contains the message detailing the warning. + WarningMessage string `protobuf:"bytes,1,opt,name=warning_message,json=warningMessage,proto3" json:"warning_message,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestWarning) Reset() { *m = TestWarning{} } +func (m *TestWarning) String() string { return proto.CompactTextString(m) } +func (*TestWarning) ProtoMessage() {} +func (*TestWarning) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{9} +} +func (m *TestWarning) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestWarning.Unmarshal(m, b) +} +func (m *TestWarning) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestWarning.Marshal(b, m, deterministic) +} +func (dst *TestWarning) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestWarning.Merge(dst, src) +} +func (m *TestWarning) XXX_Size() int { + return xxx_messageInfo_TestWarning.Size(m) +} +func (m *TestWarning) XXX_DiscardUnknown() { + xxx_messageInfo_TestWarning.DiscardUnknown(m) +} + +var xxx_messageInfo_TestWarning proto.InternalMessageInfo + +func (m *TestWarning) GetWarningMessage() string { + if m != nil { + return m.WarningMessage + } + return "" +} + +// Stores errors reading or parsing a file during post-processing. +type FileProcessingErrors struct { + // The uid of the File being read or parsed. + FileUid string `protobuf:"bytes,1,opt,name=file_uid,json=fileUid,proto3" json:"file_uid,omitempty"` + // What went wrong. + FileProcessingErrors []*FileProcessingError `protobuf:"bytes,3,rep,name=file_processing_errors,json=fileProcessingErrors,proto3" json:"file_processing_errors,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *FileProcessingErrors) Reset() { *m = FileProcessingErrors{} } +func (m *FileProcessingErrors) String() string { return proto.CompactTextString(m) } +func (*FileProcessingErrors) ProtoMessage() {} +func (*FileProcessingErrors) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{10} +} +func (m *FileProcessingErrors) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_FileProcessingErrors.Unmarshal(m, b) +} +func (m *FileProcessingErrors) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_FileProcessingErrors.Marshal(b, m, deterministic) +} +func (dst *FileProcessingErrors) XXX_Merge(src proto.Message) { + xxx_messageInfo_FileProcessingErrors.Merge(dst, src) +} +func (m *FileProcessingErrors) XXX_Size() int { + return xxx_messageInfo_FileProcessingErrors.Size(m) +} +func (m *FileProcessingErrors) XXX_DiscardUnknown() { + xxx_messageInfo_FileProcessingErrors.DiscardUnknown(m) +} + +var xxx_messageInfo_FileProcessingErrors proto.InternalMessageInfo + +func (m *FileProcessingErrors) GetFileUid() string { + if m != nil { + return m.FileUid + } + return "" +} + +func (m *FileProcessingErrors) GetFileProcessingErrors() []*FileProcessingError { + if m != nil { + return m.FileProcessingErrors + } + return nil +} + +// Stores an error reading or parsing a file during post-processing. +type FileProcessingError struct { + // The type of error that occurred. + Type FileProcessingErrorType `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.resultstore.v2.FileProcessingErrorType" json:"type,omitempty"` + // Error message describing the problem. + Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *FileProcessingError) Reset() { *m = FileProcessingError{} } +func (m *FileProcessingError) String() string { return proto.CompactTextString(m) } +func (*FileProcessingError) ProtoMessage() {} +func (*FileProcessingError) Descriptor() ([]byte, []int) { + return fileDescriptor_action_d83ec9d6d82ebb41, []int{11} +} +func (m *FileProcessingError) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_FileProcessingError.Unmarshal(m, b) +} +func (m *FileProcessingError) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_FileProcessingError.Marshal(b, m, deterministic) +} +func (dst *FileProcessingError) XXX_Merge(src proto.Message) { + xxx_messageInfo_FileProcessingError.Merge(dst, src) +} +func (m *FileProcessingError) XXX_Size() int { + return xxx_messageInfo_FileProcessingError.Size(m) +} +func (m *FileProcessingError) XXX_DiscardUnknown() { + xxx_messageInfo_FileProcessingError.DiscardUnknown(m) +} + +var xxx_messageInfo_FileProcessingError proto.InternalMessageInfo + +func (m *FileProcessingError) GetType() FileProcessingErrorType { + if m != nil { + return m.Type + } + return FileProcessingErrorType_FILE_PROCESSING_ERROR_TYPE_UNSPECIFIED +} + +func (m *FileProcessingError) GetMessage() string { + if m != nil { + return m.Message + } + return "" +} + +func init() { + proto.RegisterType((*Action)(nil), "google.devtools.resultstore.v2.Action") + proto.RegisterType((*Action_Id)(nil), "google.devtools.resultstore.v2.Action.Id") + proto.RegisterType((*BuildAction)(nil), "google.devtools.resultstore.v2.BuildAction") + proto.RegisterType((*TestAction)(nil), "google.devtools.resultstore.v2.TestAction") + proto.RegisterType((*ActionAttributes)(nil), "google.devtools.resultstore.v2.ActionAttributes") + proto.RegisterType((*InputFileInfo)(nil), "google.devtools.resultstore.v2.InputFileInfo") + proto.RegisterType((*LocalTestTiming)(nil), "google.devtools.resultstore.v2.LocalTestTiming") + proto.RegisterType((*RemoteTestAttemptTiming)(nil), "google.devtools.resultstore.v2.RemoteTestAttemptTiming") + proto.RegisterType((*RemoteTestTiming)(nil), "google.devtools.resultstore.v2.RemoteTestTiming") + proto.RegisterType((*TestTiming)(nil), "google.devtools.resultstore.v2.TestTiming") + proto.RegisterType((*TestWarning)(nil), "google.devtools.resultstore.v2.TestWarning") + proto.RegisterType((*FileProcessingErrors)(nil), "google.devtools.resultstore.v2.FileProcessingErrors") + proto.RegisterType((*FileProcessingError)(nil), "google.devtools.resultstore.v2.FileProcessingError") + proto.RegisterEnum("google.devtools.resultstore.v2.ExecutionStrategy", ExecutionStrategy_name, ExecutionStrategy_value) + proto.RegisterEnum("google.devtools.resultstore.v2.TestCaching", TestCaching_name, TestCaching_value) + proto.RegisterEnum("google.devtools.resultstore.v2.FileProcessingErrorType", FileProcessingErrorType_name, FileProcessingErrorType_value) +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/action.proto", fileDescriptor_action_d83ec9d6d82ebb41) +} + +var fileDescriptor_action_d83ec9d6d82ebb41 = []byte{ + // 1658 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x58, 0x5f, 0x6f, 0x23, 0x49, + 0x11, 0x5f, 0xdb, 0x49, 0xd6, 0x2e, 0xc7, 0xce, 0xb8, 0x93, 0xdd, 0xf5, 0x2d, 0x70, 0x2c, 0x06, + 0x8e, 0xdd, 0x2c, 0x67, 0x1f, 0x7b, 0x07, 0x08, 0x90, 0x10, 0x8e, 0xdd, 0x49, 0xe6, 0xd6, 0xb1, + 0x7d, 0xed, 0xc9, 0x71, 0x07, 0x42, 0x7d, 0x13, 0x4f, 0xc7, 0x69, 0xe4, 0x99, 0xf1, 0xcd, 0xf4, + 0xe4, 0xce, 0x12, 0x12, 0x2f, 0x48, 0x3c, 0x80, 0x78, 0xe1, 0x09, 0x89, 0x4f, 0xc2, 0x33, 0x5f, + 0x82, 0x6f, 0x83, 0xfa, 0xcf, 0xf8, 0x1f, 0x09, 0xe3, 0xdd, 0x37, 0x77, 0xd5, 0xef, 0xf7, 0xeb, + 0x9a, 0xea, 0xaa, 0xea, 0x96, 0xe1, 0xe5, 0x24, 0x0c, 0x27, 0x53, 0xd6, 0xf2, 0xd8, 0xad, 0x08, + 0xc3, 0x69, 0xdc, 0x8a, 0x58, 0x9c, 0x4c, 0x45, 0x2c, 0xc2, 0x88, 0xb5, 0x6e, 0x5f, 0xb5, 0xdc, + 0xb1, 0xe0, 0x61, 0xd0, 0x9c, 0x45, 0xa1, 0x08, 0xd1, 0xbb, 0x1a, 0xdc, 0x4c, 0xc1, 0xcd, 0x15, + 0x70, 0xf3, 0xf6, 0xd5, 0xd3, 0x2c, 0xb1, 0x71, 0xe8, 0xfb, 0xa9, 0xd8, 0xd3, 0xf7, 0x33, 0xc1, + 0xb7, 0x2c, 0x72, 0x27, 0xcc, 0xc0, 0x7f, 0xbc, 0x25, 0x9c, 0xc6, 0x89, 0xef, 0xbb, 0xd1, 0xdc, + 0xd0, 0x5e, 0x64, 0xd0, 0xae, 0xf9, 0x34, 0xdd, 0xa1, 0x95, 0x01, 0x15, 0x2c, 0x16, 0x34, 0x4e, + 0xb8, 0x48, 0x09, 0x26, 0x1d, 0x2d, 0xb5, 0xba, 0x4a, 0xae, 0x5b, 0x5e, 0x12, 0xb9, 0xcb, 0x74, + 0x35, 0xfe, 0x51, 0x82, 0xbd, 0xb6, 0xca, 0x1f, 0x42, 0xb0, 0x13, 0xb8, 0x3e, 0xab, 0xe7, 0x9e, + 0xe5, 0x9e, 0x97, 0x88, 0xfa, 0x8d, 0x7e, 0x06, 0x79, 0xee, 0xd5, 0xf3, 0xcf, 0x72, 0xcf, 0xcb, + 0xaf, 0x5e, 0x34, 0xff, 0x7f, 0x6a, 0x9b, 0x5a, 0xa7, 0x69, 0x7b, 0x24, 0xcf, 0x3d, 0xf4, 0x3b, + 0xa8, 0xc5, 0xc2, 0x15, 0x49, 0x4c, 0x5d, 0x21, 0x22, 0x7e, 0x95, 0x08, 0x16, 0xd7, 0x0b, 0x4a, + 0xe9, 0x83, 0x2c, 0xa5, 0x91, 0x22, 0xb6, 0x17, 0x3c, 0x62, 0xc5, 0x1b, 0x16, 0xf4, 0x4b, 0xd8, + 0x13, 0xdc, 0xe7, 0xc1, 0xa4, 0xbe, 0xa3, 0x34, 0xdf, 0xcb, 0xd2, 0x74, 0x14, 0x9a, 0x18, 0x16, + 0x1a, 0xc2, 0xfe, 0x55, 0xc2, 0xa7, 0x1e, 0xd5, 0xd5, 0x53, 0x2f, 0x29, 0x95, 0x97, 0x59, 0x2a, + 0x27, 0x92, 0xa3, 0x3f, 0xf4, 0xfc, 0x01, 0x29, 0x5f, 0x2d, 0x97, 0xe8, 0x02, 0xca, 0x2a, 0xfd, + 0x46, 0x10, 0x94, 0xe0, 0x71, 0x66, 0x58, 0x2c, 0x16, 0x0b, 0x3d, 0x10, 0x8b, 0x95, 0xcc, 0x9f, + 0x56, 0x5a, 0xcd, 0xdf, 0xee, 0x76, 0xf9, 0xd3, 0x12, 0xab, 0xf9, 0x73, 0x37, 0x2c, 0xe8, 0xb7, + 0x70, 0x68, 0xe4, 0x3d, 0x36, 0x63, 0x81, 0xc7, 0x82, 0x31, 0x67, 0x71, 0xbd, 0xfa, 0xac, 0xb0, + 0x4d, 0xd4, 0xdd, 0x94, 0x33, 0x27, 0x48, 0xcb, 0x74, 0x57, 0x54, 0xd0, 0x39, 0xc0, 0x2c, 0x0a, + 0x67, 0x2c, 0x12, 0x52, 0xf3, 0xa1, 0xd2, 0x7c, 0x9e, 0xa5, 0x39, 0xd4, 0x8c, 0x39, 0x59, 0xe1, + 0xa2, 0x9f, 0xc3, 0xae, 0x2c, 0xff, 0xb8, 0x5e, 0x54, 0x22, 0xdf, 0xcb, 0x12, 0x39, 0xe5, 0x53, + 0x46, 0x34, 0x05, 0x7d, 0x0c, 0xc5, 0xb4, 0xe3, 0xea, 0x65, 0x95, 0xb8, 0xe6, 0x76, 0x89, 0xeb, + 0x18, 0x16, 0x59, 0xf0, 0xd1, 0x35, 0xa0, 0x8d, 0xee, 0x95, 0x5f, 0xb6, 0xaf, 0x82, 0xfa, 0x69, + 0x96, 0x6a, 0xcf, 0x0d, 0x26, 0x89, 0x3b, 0x61, 0xa9, 0xee, 0x48, 0xb7, 0x3f, 0xa9, 0x8d, 0xd7, + 0x0c, 0xf2, 0x7b, 0x7f, 0x0f, 0x8f, 0x65, 0xf0, 0x74, 0x16, 0x85, 0x63, 0x16, 0xc7, 0x3c, 0x98, + 0x50, 0x16, 0x45, 0x61, 0x14, 0xd7, 0x2b, 0x6a, 0xaf, 0x8f, 0xb6, 0x49, 0xc0, 0x70, 0x41, 0xc6, + 0x8a, 0x4b, 0x8e, 0xae, 0xef, 0xb0, 0x3e, 0xfd, 0x5b, 0x0e, 0xf2, 0xb6, 0x87, 0xbe, 0x0b, 0x15, + 0x1e, 0xdc, 0x86, 0x63, 0x35, 0x16, 0x28, 0xf7, 0xcc, 0x00, 0xd8, 0x5f, 0x1a, 0x6d, 0x0f, 0x7d, + 0x03, 0x4a, 0xc2, 0x8d, 0x26, 0x4c, 0x50, 0x33, 0x0f, 0x4a, 0xa4, 0xa8, 0x0d, 0xb6, 0x87, 0x5e, + 0x80, 0x35, 0x0e, 0x83, 0x6b, 0x3e, 0x31, 0xb3, 0x45, 0x62, 0x0a, 0x0a, 0x73, 0xb0, 0x66, 0xd7, + 0x3a, 0xa6, 0xec, 0xb8, 0xa7, 0x3a, 0xb7, 0x44, 0x8a, 0xda, 0x60, 0x7b, 0x27, 0x15, 0x28, 0x1b, + 0xa7, 0x98, 0xcf, 0x58, 0xe3, 0x8f, 0x50, 0x5e, 0x69, 0x37, 0x39, 0x9f, 0xa4, 0x39, 0x9d, 0x4f, + 0xf2, 0x37, 0xfa, 0x21, 0xa0, 0x59, 0xc4, 0x65, 0x32, 0x29, 0x0f, 0x66, 0x89, 0xa0, 0x33, 0x57, + 0xdc, 0x98, 0xf8, 0x2c, 0xe3, 0xb1, 0xa5, 0x63, 0xe8, 0x8a, 0x1b, 0xd4, 0x84, 0xc3, 0x14, 0x1d, + 0x26, 0x62, 0x01, 0xd7, 0xa1, 0xd6, 0x8c, 0x6b, 0xa0, 0x3c, 0x12, 0xdf, 0xf8, 0x4b, 0x01, 0x60, + 0xd9, 0x9f, 0xe8, 0xb5, 0x69, 0x70, 0x33, 0x77, 0x72, 0xdb, 0x37, 0xb8, 0x99, 0x3d, 0xaa, 0xbd, + 0xf5, 0x6f, 0xf4, 0x1d, 0xd8, 0x8f, 0x6f, 0xdc, 0xc8, 0xa3, 0x41, 0xe2, 0x5f, 0xb1, 0x48, 0xc5, + 0xbc, 0x4b, 0xca, 0xca, 0xd6, 0x57, 0x26, 0xf4, 0x2d, 0x80, 0x28, 0x09, 0x52, 0x40, 0x41, 0x01, + 0x4a, 0x51, 0x12, 0x18, 0xf7, 0xf7, 0xa1, 0xea, 0x0a, 0xc1, 0xfc, 0x99, 0x48, 0x21, 0x3b, 0x0a, + 0x52, 0x31, 0x56, 0x03, 0x3b, 0x07, 0x58, 0xde, 0x0a, 0x66, 0x80, 0xbc, 0xd8, 0x26, 0xe8, 0x91, + 0x24, 0x90, 0x92, 0x48, 0x7f, 0xa2, 0x33, 0x28, 0x7e, 0xe5, 0x46, 0x01, 0x0f, 0x26, 0x69, 0x3b, + 0xbe, 0xdc, 0x46, 0xe7, 0xd7, 0x9a, 0x43, 0x16, 0x64, 0xf4, 0x11, 0x3c, 0x66, 0xb1, 0xe0, 0xbe, + 0x2b, 0x98, 0x47, 0x7d, 0xe6, 0x87, 0xd1, 0x9c, 0x5e, 0xcd, 0xe5, 0x7c, 0x93, 0x43, 0xb3, 0x40, + 0x8e, 0x16, 0xde, 0x0b, 0xe5, 0x3c, 0x91, 0xbe, 0xc6, 0x9f, 0xf2, 0x60, 0x6d, 0x0e, 0x36, 0xf4, + 0x05, 0x20, 0xf6, 0x35, 0x1b, 0x27, 0xaa, 0x6a, 0x62, 0x11, 0xb9, 0x82, 0x4d, 0xe6, 0xea, 0x68, + 0xaa, 0xaf, 0x7e, 0x94, 0x15, 0x1d, 0x4e, 0x99, 0x23, 0x43, 0x24, 0x35, 0xb6, 0x69, 0x92, 0x15, + 0xcb, 0xbe, 0xe6, 0x82, 0x8e, 0x43, 0x8f, 0x99, 0x53, 0x2a, 0x4a, 0x43, 0x27, 0xf4, 0x18, 0x7a, + 0x0a, 0xc5, 0x9b, 0x30, 0x16, 0xea, 0xde, 0xd4, 0x65, 0xb4, 0x58, 0xa3, 0x4b, 0x38, 0xd0, 0x35, + 0xa9, 0x1a, 0x9a, 0x07, 0xd7, 0xa1, 0xb9, 0xaa, 0xde, 0xcf, 0x8a, 0x4b, 0x55, 0xac, 0x6c, 0x64, + 0x3b, 0xb8, 0x0e, 0x49, 0x85, 0xaf, 0x2e, 0x1b, 0xff, 0xce, 0x41, 0x65, 0x0d, 0x80, 0x8e, 0x60, + 0x77, 0x1c, 0x26, 0x81, 0x50, 0x9f, 0x5d, 0x20, 0x7a, 0x21, 0xcb, 0xc3, 0xe3, 0xb1, 0xe0, 0xc1, + 0x58, 0xc6, 0x2e, 0xdd, 0x79, 0xe5, 0xae, 0xa4, 0xd6, 0x8e, 0x82, 0x7d, 0x1b, 0xca, 0xca, 0x4b, + 0xa7, 0xdc, 0xe7, 0x42, 0x7d, 0x44, 0x81, 0x80, 0x32, 0xf5, 0xa4, 0x65, 0x4d, 0x47, 0x1f, 0xd2, + 0xce, 0xba, 0x8e, 0x3a, 0x1d, 0xd9, 0x5b, 0x6b, 0x30, 0xa3, 0xb7, 0xab, 0xb0, 0xb5, 0x55, 0xac, + 0x92, 0x6d, 0x7c, 0x01, 0x07, 0xbd, 0x70, 0xec, 0x4e, 0x97, 0xed, 0x81, 0x2e, 0xe0, 0x91, 0xaa, + 0x54, 0x33, 0xfb, 0x68, 0xfa, 0x54, 0x31, 0x9d, 0xf6, 0x4e, 0x9a, 0xb6, 0xf4, 0x2d, 0xd3, 0xec, + 0x1a, 0x00, 0x39, 0x94, 0x3c, 0x33, 0xdf, 0x52, 0x63, 0xe3, 0xcf, 0x05, 0x78, 0x42, 0x98, 0x1f, + 0x0a, 0xa6, 0x7a, 0x58, 0x37, 0x85, 0xd9, 0xea, 0x57, 0x50, 0xfd, 0x32, 0x61, 0x09, 0x7b, 0x83, + 0x3d, 0x2a, 0x8a, 0x90, 0x2e, 0xd1, 0x09, 0x1c, 0x24, 0xb3, 0x69, 0xe8, 0x7a, 0x4b, 0x89, 0x7c, + 0x96, 0x44, 0x55, 0x33, 0x16, 0x1a, 0x03, 0x78, 0xec, 0xbb, 0xe3, 0x1b, 0x1e, 0x30, 0x1a, 0x33, + 0x91, 0xcc, 0x96, 0x52, 0x85, 0x2c, 0xa9, 0x23, 0x43, 0x1c, 0x49, 0xde, 0x42, 0xf0, 0xde, 0x0c, + 0xee, 0xbc, 0x4d, 0x06, 0xd1, 0x29, 0xd4, 0xbc, 0xf0, 0xab, 0x60, 0xfd, 0x2b, 0x77, 0xb3, 0xa4, + 0xac, 0x94, 0xb3, 0x38, 0x89, 0x7f, 0xe5, 0xc0, 0x5a, 0x9e, 0x84, 0x39, 0x82, 0x4f, 0xe0, 0xc9, + 0x54, 0x16, 0x00, 0x75, 0x03, 0x77, 0x3a, 0x8f, 0xf9, 0x9b, 0x9c, 0xf7, 0x23, 0xc5, 0x6c, 0x1b, + 0xe2, 0x22, 0xde, 0x11, 0x14, 0xcd, 0xec, 0x8b, 0xeb, 0xf9, 0xed, 0xae, 0xe6, 0x7b, 0x0a, 0x84, + 0x2c, 0x84, 0x1a, 0x7f, 0x37, 0x97, 0x80, 0x09, 0xfb, 0x0c, 0x76, 0xd5, 0xe6, 0x26, 0xc8, 0x56, + 0xe6, 0xdd, 0xbf, 0x5e, 0xe4, 0xe7, 0x0f, 0x88, 0xe6, 0xa3, 0x8f, 0x61, 0x2f, 0x52, 0x9b, 0x9b, + 0xba, 0xf9, 0x60, 0xfb, 0x50, 0x17, 0x52, 0x46, 0x01, 0xbd, 0x86, 0xa3, 0x78, 0x1e, 0x0b, 0xe6, + 0xcb, 0xbb, 0x89, 0xbd, 0x41, 0x19, 0x21, 0x4d, 0x73, 0xb8, 0xbf, 0xac, 0xec, 0x33, 0x40, 0x49, + 0xcc, 0xa2, 0x0d, 0xa9, 0xcc, 0x0a, 0xb2, 0x24, 0x69, 0x4d, 0xa8, 0x0f, 0xfb, 0xaa, 0x1a, 0xc7, + 0xaa, 0x54, 0x27, 0xaa, 0x72, 0xaa, 0xdb, 0xdd, 0x19, 0x1d, 0x4d, 0x21, 0xea, 0xc2, 0x35, 0x8b, + 0x13, 0x80, 0xe2, 0xd4, 0xbc, 0x48, 0x1a, 0x3f, 0x81, 0xf2, 0xca, 0xdd, 0x82, 0x7e, 0x00, 0x07, + 0xe6, 0x76, 0xa1, 0x3e, 0x8b, 0x63, 0xf9, 0xe2, 0xd3, 0xcf, 0x84, 0xaa, 0x31, 0x5f, 0x68, 0x6b, + 0xe3, 0x9f, 0x39, 0x38, 0xba, 0xeb, 0x89, 0x84, 0xde, 0x81, 0xa2, 0x9a, 0xd3, 0xc9, 0xe2, 0x01, + 0xf4, 0x50, 0xae, 0x2f, 0xb9, 0x87, 0xf8, 0xbd, 0x6f, 0xb2, 0x82, 0x2a, 0xb2, 0x0f, 0xdf, 0xe2, + 0x4d, 0x76, 0xf7, 0x93, 0xac, 0xf1, 0x07, 0x38, 0xbc, 0x03, 0x8c, 0x5e, 0xaf, 0x3c, 0x7d, 0xaa, + 0xd9, 0x45, 0x7d, 0x87, 0x84, 0x33, 0x9f, 0x31, 0xf3, 0x66, 0xaa, 0xc3, 0xc3, 0x34, 0x47, 0xfa, + 0xa1, 0x94, 0x2e, 0x8f, 0xff, 0x9a, 0x83, 0xda, 0xff, 0xdc, 0x89, 0xa8, 0x01, 0xef, 0xe2, 0xcf, + 0x70, 0xe7, 0xd2, 0xb1, 0x07, 0x7d, 0x3a, 0x72, 0x48, 0xdb, 0xc1, 0x67, 0x9f, 0xd3, 0xcb, 0xfe, + 0x68, 0x88, 0x3b, 0xf6, 0xa9, 0x8d, 0xbb, 0xd6, 0x03, 0xf4, 0x08, 0x6a, 0x03, 0xe7, 0x1c, 0x13, + 0x8a, 0xfb, 0x9f, 0xda, 0x64, 0xd0, 0xbf, 0xc0, 0x7d, 0xc7, 0xca, 0x21, 0x04, 0x55, 0x82, 0x2f, + 0x06, 0x0e, 0xa6, 0x23, 0x4c, 0x3e, 0xb5, 0x3b, 0xd8, 0xca, 0x4b, 0x5b, 0x6f, 0xd0, 0x69, 0xf7, + 0xe8, 0xb0, 0x4d, 0xda, 0xbd, 0x1e, 0xee, 0x59, 0x05, 0x74, 0x04, 0x96, 0xb6, 0x8d, 0xf0, 0x27, + 0x97, 0xb8, 0xef, 0xd8, 0xed, 0x9e, 0xb5, 0x73, 0x7c, 0xad, 0xcf, 0xd8, 0x1c, 0x3f, 0xfa, 0x26, + 0xd4, 0x1d, 0x3c, 0x72, 0x68, 0xa7, 0xdd, 0x39, 0xb7, 0xfb, 0x67, 0x1b, 0x11, 0x1c, 0xc2, 0x81, + 0x96, 0x90, 0x6e, 0x4c, 0xcf, 0x6d, 0xb9, 0xff, 0x11, 0x58, 0x66, 0xff, 0xa5, 0x35, 0x8f, 0xaa, + 0x00, 0x7a, 0x79, 0x61, 0x8f, 0x46, 0x56, 0xe1, 0xf8, 0x3f, 0x39, 0x78, 0x72, 0x4f, 0xca, 0xd0, + 0x31, 0xbc, 0x77, 0x6a, 0xf7, 0x30, 0x1d, 0x92, 0x41, 0x07, 0x8f, 0x46, 0x72, 0x5f, 0x4c, 0xc8, + 0x80, 0x50, 0xe7, 0xf3, 0x21, 0xde, 0x08, 0xe1, 0x31, 0xa0, 0x33, 0xdc, 0xc7, 0xc4, 0xee, 0x50, + 0x82, 0xdb, 0x5d, 0x0d, 0xb4, 0x72, 0xe8, 0x09, 0x1c, 0xa6, 0xf6, 0x61, 0x9b, 0x8c, 0xb0, 0x71, + 0xa8, 0x54, 0x28, 0x71, 0x67, 0x30, 0xa0, 0xbd, 0x36, 0x39, 0xc3, 0x3a, 0x15, 0x83, 0x4b, 0x67, + 0x78, 0xe9, 0xac, 0x58, 0x77, 0x50, 0x0d, 0x2a, 0xed, 0x8e, 0x0c, 0x80, 0x76, 0x71, 0x5f, 0xee, + 0xb6, 0x2b, 0x53, 0xde, 0xc5, 0xed, 0x6e, 0xcf, 0xee, 0x63, 0x8a, 0x3f, 0xeb, 0x60, 0xdc, 0xc5, + 0x5d, 0x6b, 0x0f, 0x55, 0xa0, 0xd4, 0x1f, 0x38, 0xf4, 0x74, 0x70, 0xd9, 0xef, 0x5a, 0x0f, 0x4f, + 0xbe, 0x84, 0xc6, 0x38, 0xf4, 0x33, 0x0a, 0x66, 0x98, 0xfb, 0x8d, 0x6d, 0x10, 0x93, 0x70, 0xea, + 0x06, 0x93, 0x66, 0x18, 0x4d, 0x5a, 0x13, 0x16, 0xa8, 0x2e, 0x37, 0xff, 0x39, 0xb8, 0x33, 0x1e, + 0xdf, 0xf7, 0xbf, 0xc3, 0x2f, 0x56, 0x96, 0x57, 0x7b, 0x8a, 0xf5, 0xe1, 0x7f, 0x03, 0x00, 0x00, + 0xff, 0xff, 0xef, 0xcf, 0x51, 0x38, 0xb7, 0x11, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/common.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/common.pb.go new file mode 100644 index 000000000..a54da05e1 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/common.pb.go @@ -0,0 +1,588 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/common.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// These correspond to the prefix of the rule name. Eg cc_test has language CC. +type Language int32 + +const ( + // Language unspecified or not listed here. + Language_LANGUAGE_UNSPECIFIED Language = 0 + // Not related to any particular language + Language_NONE Language = 1 + // Android + Language_ANDROID Language = 2 + // ActionScript (Flash) + Language_AS Language = 3 + // C++ or C + Language_CC Language = 4 + // Cascading-Style-Sheets + Language_CSS Language = 5 + // Dart + Language_DART Language = 6 + // Go + Language_GO Language = 7 + // Google-Web-Toolkit + Language_GWT Language = 8 + // Haskell + Language_HASKELL Language = 9 + // Java + Language_JAVA Language = 10 + // Javascript + Language_JS Language = 11 + // Lisp + Language_LISP Language = 12 + // Objective-C + Language_OBJC Language = 13 + // Python + Language_PY Language = 14 + // Shell (Typically Bash) + Language_SH Language = 15 + // Swift + Language_SWIFT Language = 16 + // Typescript + Language_TS Language = 18 + // Webtesting + Language_WEB Language = 19 +) + +var Language_name = map[int32]string{ + 0: "LANGUAGE_UNSPECIFIED", + 1: "NONE", + 2: "ANDROID", + 3: "AS", + 4: "CC", + 5: "CSS", + 6: "DART", + 7: "GO", + 8: "GWT", + 9: "HASKELL", + 10: "JAVA", + 11: "JS", + 12: "LISP", + 13: "OBJC", + 14: "PY", + 15: "SH", + 16: "SWIFT", + 18: "TS", + 19: "WEB", +} +var Language_value = map[string]int32{ + "LANGUAGE_UNSPECIFIED": 0, + "NONE": 1, + "ANDROID": 2, + "AS": 3, + "CC": 4, + "CSS": 5, + "DART": 6, + "GO": 7, + "GWT": 8, + "HASKELL": 9, + "JAVA": 10, + "JS": 11, + "LISP": 12, + "OBJC": 13, + "PY": 14, + "SH": 15, + "SWIFT": 16, + "TS": 18, + "WEB": 19, +} + +func (x Language) String() string { + return proto.EnumName(Language_name, int32(x)) +} +func (Language) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_common_47ffbc3d0eb889a6, []int{0} +} + +// Status of a resource. +type Status int32 + +const ( + // The implicit default enum value. Should never be set. + Status_STATUS_UNSPECIFIED Status = 0 + // Displays as "Building". Means the target is compiling, linking, etc. + Status_BUILDING Status = 1 + // Displays as "Built". Means the target was built successfully. + // If testing was requested, it should never reach this status: it should go + // straight from BUILDING to TESTING. + Status_BUILT Status = 2 + // Displays as "Broken". Means build failure such as compile error. + Status_FAILED_TO_BUILD Status = 3 + // Displays as "Testing". Means the test is running. + Status_TESTING Status = 4 + // Displays as "Passed". Means the test was run and passed. + Status_PASSED Status = 5 + // Displays as "Failed". Means the test was run and failed. + Status_FAILED Status = 6 + // Displays as "Timed out". Means the test didn't finish in time. + Status_TIMED_OUT Status = 7 + // Displays as "Cancelled". Means the build or test was cancelled. + // E.g. User hit control-C. + Status_CANCELLED Status = 8 + // Displays as "Tool Failed". Means the build or test had internal tool + // failure. + Status_TOOL_FAILED Status = 9 + // Displays as "Incomplete". Means the build or test did not complete. This + // might happen when a build breakage or test failure causes the tool to stop + // trying to build anything more or run any more tests, with the default + // bazel --nokeep_going option or the --notest_keep_going option. + Status_INCOMPLETE Status = 10 + // Displays as "Flaky". Means the aggregate status contains some runs that + // were successful, and some that were not. + Status_FLAKY Status = 11 + // Displays as "Unknown". Means the tool uploading to the server died + // mid-upload or does not know the state. + Status_UNKNOWN Status = 12 + // Displays as "Skipped". Means building and testing were skipped. + // (E.g. Restricted to a different configuration.) + Status_SKIPPED Status = 13 +) + +var Status_name = map[int32]string{ + 0: "STATUS_UNSPECIFIED", + 1: "BUILDING", + 2: "BUILT", + 3: "FAILED_TO_BUILD", + 4: "TESTING", + 5: "PASSED", + 6: "FAILED", + 7: "TIMED_OUT", + 8: "CANCELLED", + 9: "TOOL_FAILED", + 10: "INCOMPLETE", + 11: "FLAKY", + 12: "UNKNOWN", + 13: "SKIPPED", +} +var Status_value = map[string]int32{ + "STATUS_UNSPECIFIED": 0, + "BUILDING": 1, + "BUILT": 2, + "FAILED_TO_BUILD": 3, + "TESTING": 4, + "PASSED": 5, + "FAILED": 6, + "TIMED_OUT": 7, + "CANCELLED": 8, + "TOOL_FAILED": 9, + "INCOMPLETE": 10, + "FLAKY": 11, + "UNKNOWN": 12, + "SKIPPED": 13, +} + +func (x Status) String() string { + return proto.EnumName(Status_name, int32(x)) +} +func (Status) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_common_47ffbc3d0eb889a6, []int{1} +} + +// Describes the status of a resource in both enum and string form. +// Only use description when conveying additional info not captured in the enum +// name. +type StatusAttributes struct { + // Enum representation of the status. + Status Status `protobuf:"varint,1,opt,name=status,proto3,enum=google.devtools.resultstore.v2.Status" json:"status,omitempty"` + // A longer description about the status. + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StatusAttributes) Reset() { *m = StatusAttributes{} } +func (m *StatusAttributes) String() string { return proto.CompactTextString(m) } +func (*StatusAttributes) ProtoMessage() {} +func (*StatusAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_common_47ffbc3d0eb889a6, []int{0} +} +func (m *StatusAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StatusAttributes.Unmarshal(m, b) +} +func (m *StatusAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StatusAttributes.Marshal(b, m, deterministic) +} +func (dst *StatusAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_StatusAttributes.Merge(dst, src) +} +func (m *StatusAttributes) XXX_Size() int { + return xxx_messageInfo_StatusAttributes.Size(m) +} +func (m *StatusAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_StatusAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_StatusAttributes proto.InternalMessageInfo + +func (m *StatusAttributes) GetStatus() Status { + if m != nil { + return m.Status + } + return Status_STATUS_UNSPECIFIED +} + +func (m *StatusAttributes) GetDescription() string { + if m != nil { + return m.Description + } + return "" +} + +// A generic key-value property definition. +type Property struct { + // The key. + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` + // The value. + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Property) Reset() { *m = Property{} } +func (m *Property) String() string { return proto.CompactTextString(m) } +func (*Property) ProtoMessage() {} +func (*Property) Descriptor() ([]byte, []int) { + return fileDescriptor_common_47ffbc3d0eb889a6, []int{1} +} +func (m *Property) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Property.Unmarshal(m, b) +} +func (m *Property) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Property.Marshal(b, m, deterministic) +} +func (dst *Property) XXX_Merge(src proto.Message) { + xxx_messageInfo_Property.Merge(dst, src) +} +func (m *Property) XXX_Size() int { + return xxx_messageInfo_Property.Size(m) +} +func (m *Property) XXX_DiscardUnknown() { + xxx_messageInfo_Property.DiscardUnknown(m) +} + +var xxx_messageInfo_Property proto.InternalMessageInfo + +func (m *Property) GetKey() string { + if m != nil { + return m.Key + } + return "" +} + +func (m *Property) GetValue() string { + if m != nil { + return m.Value + } + return "" +} + +// The timing of a particular Invocation, Action, etc. The start_time is +// specified, stop time can be calculated by adding duration to start_time. +type Timing struct { + // The time the resource started running. This is in UTC Epoch time. + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` + // The duration for which the resource ran. + Duration *duration.Duration `protobuf:"bytes,2,opt,name=duration,proto3" json:"duration,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Timing) Reset() { *m = Timing{} } +func (m *Timing) String() string { return proto.CompactTextString(m) } +func (*Timing) ProtoMessage() {} +func (*Timing) Descriptor() ([]byte, []int) { + return fileDescriptor_common_47ffbc3d0eb889a6, []int{2} +} +func (m *Timing) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Timing.Unmarshal(m, b) +} +func (m *Timing) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Timing.Marshal(b, m, deterministic) +} +func (dst *Timing) XXX_Merge(src proto.Message) { + xxx_messageInfo_Timing.Merge(dst, src) +} +func (m *Timing) XXX_Size() int { + return xxx_messageInfo_Timing.Size(m) +} +func (m *Timing) XXX_DiscardUnknown() { + xxx_messageInfo_Timing.DiscardUnknown(m) +} + +var xxx_messageInfo_Timing proto.InternalMessageInfo + +func (m *Timing) GetStartTime() *timestamp.Timestamp { + if m != nil { + return m.StartTime + } + return nil +} + +func (m *Timing) GetDuration() *duration.Duration { + if m != nil { + return m.Duration + } + return nil +} + +// Represents a dependency of a resource on another resource. This can be used +// to define a graph or a workflow paradigm through resources. +type Dependency struct { + // The resource depended upon. It may be a Target, ConfiguredTarget, or + // Action. + // + // Types that are valid to be assigned to Resource: + // *Dependency_Target + // *Dependency_ConfiguredTarget + // *Dependency_Action + Resource isDependency_Resource `protobuf_oneof:"resource"` + // A label describing this dependency. + // The label "Root Cause" is handled specially. It is used to point to the + // exact resource that caused a resource to fail. + Label string `protobuf:"bytes,4,opt,name=label,proto3" json:"label,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Dependency) Reset() { *m = Dependency{} } +func (m *Dependency) String() string { return proto.CompactTextString(m) } +func (*Dependency) ProtoMessage() {} +func (*Dependency) Descriptor() ([]byte, []int) { + return fileDescriptor_common_47ffbc3d0eb889a6, []int{3} +} +func (m *Dependency) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Dependency.Unmarshal(m, b) +} +func (m *Dependency) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Dependency.Marshal(b, m, deterministic) +} +func (dst *Dependency) XXX_Merge(src proto.Message) { + xxx_messageInfo_Dependency.Merge(dst, src) +} +func (m *Dependency) XXX_Size() int { + return xxx_messageInfo_Dependency.Size(m) +} +func (m *Dependency) XXX_DiscardUnknown() { + xxx_messageInfo_Dependency.DiscardUnknown(m) +} + +var xxx_messageInfo_Dependency proto.InternalMessageInfo + +type isDependency_Resource interface { + isDependency_Resource() +} + +type Dependency_Target struct { + Target string `protobuf:"bytes,1,opt,name=target,proto3,oneof"` +} +type Dependency_ConfiguredTarget struct { + ConfiguredTarget string `protobuf:"bytes,2,opt,name=configured_target,json=configuredTarget,proto3,oneof"` +} +type Dependency_Action struct { + Action string `protobuf:"bytes,3,opt,name=action,proto3,oneof"` +} + +func (*Dependency_Target) isDependency_Resource() {} +func (*Dependency_ConfiguredTarget) isDependency_Resource() {} +func (*Dependency_Action) isDependency_Resource() {} + +func (m *Dependency) GetResource() isDependency_Resource { + if m != nil { + return m.Resource + } + return nil +} + +func (m *Dependency) GetTarget() string { + if x, ok := m.GetResource().(*Dependency_Target); ok { + return x.Target + } + return "" +} + +func (m *Dependency) GetConfiguredTarget() string { + if x, ok := m.GetResource().(*Dependency_ConfiguredTarget); ok { + return x.ConfiguredTarget + } + return "" +} + +func (m *Dependency) GetAction() string { + if x, ok := m.GetResource().(*Dependency_Action); ok { + return x.Action + } + return "" +} + +func (m *Dependency) GetLabel() string { + if m != nil { + return m.Label + } + return "" +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Dependency) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Dependency_OneofMarshaler, _Dependency_OneofUnmarshaler, _Dependency_OneofSizer, []interface{}{ + (*Dependency_Target)(nil), + (*Dependency_ConfiguredTarget)(nil), + (*Dependency_Action)(nil), + } +} + +func _Dependency_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Dependency) + // resource + switch x := m.Resource.(type) { + case *Dependency_Target: + b.EncodeVarint(1<<3 | proto.WireBytes) + b.EncodeStringBytes(x.Target) + case *Dependency_ConfiguredTarget: + b.EncodeVarint(2<<3 | proto.WireBytes) + b.EncodeStringBytes(x.ConfiguredTarget) + case *Dependency_Action: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.Action) + case nil: + default: + return fmt.Errorf("Dependency.Resource has unexpected type %T", x) + } + return nil +} + +func _Dependency_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Dependency) + switch tag { + case 1: // resource.target + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.Resource = &Dependency_Target{x} + return true, err + case 2: // resource.configured_target + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.Resource = &Dependency_ConfiguredTarget{x} + return true, err + case 3: // resource.action + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.Resource = &Dependency_Action{x} + return true, err + default: + return false, nil + } +} + +func _Dependency_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Dependency) + // resource + switch x := m.Resource.(type) { + case *Dependency_Target: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.Target))) + n += len(x.Target) + case *Dependency_ConfiguredTarget: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.ConfiguredTarget))) + n += len(x.ConfiguredTarget) + case *Dependency_Action: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.Action))) + n += len(x.Action) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +func init() { + proto.RegisterType((*StatusAttributes)(nil), "google.devtools.resultstore.v2.StatusAttributes") + proto.RegisterType((*Property)(nil), "google.devtools.resultstore.v2.Property") + proto.RegisterType((*Timing)(nil), "google.devtools.resultstore.v2.Timing") + proto.RegisterType((*Dependency)(nil), "google.devtools.resultstore.v2.Dependency") + proto.RegisterEnum("google.devtools.resultstore.v2.Language", Language_name, Language_value) + proto.RegisterEnum("google.devtools.resultstore.v2.Status", Status_name, Status_value) +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/common.proto", fileDescriptor_common_47ffbc3d0eb889a6) +} + +var fileDescriptor_common_47ffbc3d0eb889a6 = []byte{ + // 690 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x94, 0xcd, 0x8e, 0xe3, 0x44, + 0x10, 0xc7, 0xd7, 0xf9, 0x70, 0x9c, 0xca, 0x7c, 0x14, 0xbd, 0x2b, 0x14, 0xe6, 0x00, 0xa3, 0x1c, + 0xd0, 0x6a, 0x11, 0xb6, 0x14, 0xc4, 0x01, 0x21, 0x21, 0x39, 0xb6, 0x93, 0xf1, 0xc4, 0x6b, 0x5b, + 0xee, 0x0e, 0xd1, 0x72, 0x89, 0x9c, 0xa4, 0xd7, 0xb2, 0x48, 0xec, 0x60, 0xb7, 0x23, 0x0d, 0x6f, + 0xc1, 0x8b, 0x71, 0xe5, 0x75, 0x50, 0xdb, 0x0e, 0x8c, 0x06, 0x01, 0xa7, 0xee, 0x7f, 0xfd, 0x7f, + 0x55, 0x5d, 0xd5, 0x6a, 0x1b, 0xbe, 0x4a, 0xf2, 0x3c, 0x39, 0x70, 0x63, 0xcf, 0xcf, 0x22, 0xcf, + 0x0f, 0xa5, 0x51, 0xf0, 0xb2, 0x3a, 0x88, 0x52, 0xe4, 0x05, 0x37, 0xce, 0x53, 0x63, 0x97, 0x1f, + 0x8f, 0x79, 0xa6, 0x9f, 0x8a, 0x5c, 0xe4, 0xe4, 0xf3, 0x06, 0xd6, 0x2f, 0xb0, 0xfe, 0x0c, 0xd6, + 0xcf, 0xd3, 0xbb, 0xd6, 0x37, 0x6a, 0x7a, 0x5b, 0x7d, 0x34, 0xf6, 0x55, 0x11, 0x8b, 0xf4, 0x92, + 0x7f, 0xf7, 0xc5, 0x4b, 0x5f, 0xa4, 0x47, 0x5e, 0x8a, 0xf8, 0x78, 0x6a, 0x80, 0x89, 0x00, 0xa4, + 0x22, 0x16, 0x55, 0x69, 0x0a, 0x51, 0xa4, 0xdb, 0x4a, 0xf0, 0x92, 0xfc, 0x00, 0x6a, 0x59, 0xc7, + 0xc6, 0xca, 0xbd, 0xf2, 0xf6, 0x66, 0xfa, 0xa5, 0xfe, 0xdf, 0x5d, 0xe8, 0x4d, 0x85, 0xa8, 0xcd, + 0x22, 0xf7, 0x30, 0xda, 0xf3, 0x72, 0x57, 0xa4, 0x27, 0xd9, 0xc9, 0xb8, 0x73, 0xaf, 0xbc, 0x1d, + 0x46, 0xcf, 0x43, 0x93, 0x29, 0x68, 0x61, 0x91, 0x9f, 0x78, 0x21, 0x9e, 0x08, 0x42, 0xf7, 0x67, + 0xfe, 0x54, 0x1f, 0x35, 0x8c, 0xe4, 0x96, 0xbc, 0x81, 0xfe, 0x39, 0x3e, 0x54, 0xbc, 0xcd, 0x6c, + 0xc4, 0xe4, 0x57, 0x50, 0x59, 0x7a, 0x4c, 0xb3, 0x84, 0x7c, 0x07, 0x50, 0x8a, 0xb8, 0x10, 0x1b, + 0x39, 0x4c, 0x9d, 0x38, 0x9a, 0xde, 0x5d, 0x7a, 0xbc, 0x4c, 0xaa, 0xb3, 0xcb, 0xa4, 0xd1, 0xb0, + 0xa6, 0xa5, 0x26, 0xdf, 0x82, 0x76, 0xb9, 0xa1, 0xba, 0xfa, 0x68, 0xfa, 0xd9, 0x3f, 0x12, 0xed, + 0x16, 0x88, 0xfe, 0x42, 0x27, 0xbf, 0x29, 0x00, 0x36, 0x3f, 0xf1, 0x6c, 0xcf, 0xb3, 0xdd, 0x13, + 0x19, 0x83, 0x2a, 0xe2, 0x22, 0xe1, 0xa2, 0xe9, 0xfa, 0xe1, 0x55, 0xd4, 0x6a, 0xf2, 0x35, 0x7c, + 0xb2, 0xcb, 0xb3, 0x8f, 0x69, 0x52, 0x15, 0x7c, 0xbf, 0x69, 0xa1, 0x4e, 0x0b, 0xe1, 0xdf, 0x16, + 0x6b, 0xf0, 0x31, 0xa8, 0xf1, 0xae, 0x6e, 0xa6, 0x7b, 0x29, 0xd4, 0x68, 0x79, 0x07, 0x87, 0x78, + 0xcb, 0x0f, 0xe3, 0x5e, 0x73, 0x07, 0xb5, 0x98, 0x01, 0x68, 0x05, 0x2f, 0xf3, 0xaa, 0xd8, 0xf1, + 0x77, 0xbf, 0x2b, 0xa0, 0x79, 0x71, 0x96, 0x54, 0x71, 0xc2, 0xc9, 0x18, 0xde, 0x78, 0xa6, 0xbf, + 0x58, 0x99, 0x0b, 0x67, 0xb3, 0xf2, 0x69, 0xe8, 0x58, 0xee, 0xdc, 0x75, 0x6c, 0x7c, 0x45, 0x34, + 0xe8, 0xf9, 0x81, 0xef, 0xa0, 0x42, 0x46, 0x30, 0x30, 0x7d, 0x3b, 0x0a, 0x5c, 0x1b, 0x3b, 0x44, + 0x85, 0x8e, 0x49, 0xb1, 0x2b, 0x57, 0xcb, 0xc2, 0x1e, 0x19, 0x40, 0xd7, 0xa2, 0x14, 0xfb, 0x92, + 0xb7, 0xcd, 0x88, 0xa1, 0x2a, 0xad, 0x45, 0x80, 0x03, 0x69, 0x2d, 0xd6, 0x0c, 0x35, 0x59, 0xe0, + 0xc1, 0xa4, 0x4b, 0xc7, 0xf3, 0x70, 0x28, 0xb9, 0x47, 0xf3, 0x47, 0x13, 0x41, 0x72, 0x8f, 0x14, + 0x47, 0x32, 0xe2, 0xb9, 0x34, 0xc4, 0x2b, 0xb9, 0x0b, 0x66, 0x8f, 0x16, 0x5e, 0x4b, 0x2f, 0xfc, + 0x80, 0x37, 0x72, 0xa5, 0x0f, 0x78, 0x4b, 0x86, 0xd0, 0xa7, 0x6b, 0x77, 0xce, 0x10, 0x65, 0x88, + 0x51, 0x24, 0xb2, 0xfc, 0xda, 0x99, 0xe1, 0xeb, 0x77, 0x7f, 0x28, 0xa0, 0x36, 0x2f, 0x89, 0x7c, + 0x0a, 0x84, 0x32, 0x93, 0xad, 0xe8, 0x8b, 0x61, 0xae, 0x40, 0x9b, 0xad, 0x5c, 0xcf, 0x76, 0xfd, + 0x05, 0x2a, 0xb2, 0x98, 0x54, 0x0c, 0x3b, 0xe4, 0x35, 0xdc, 0xce, 0x4d, 0xd7, 0x73, 0xec, 0x0d, + 0x0b, 0x36, 0x35, 0x82, 0x5d, 0xd9, 0x2f, 0x73, 0x28, 0x93, 0x70, 0x8f, 0x00, 0xa8, 0xa1, 0x49, + 0xa9, 0x63, 0x63, 0x5f, 0xee, 0x1b, 0x1a, 0x55, 0x72, 0x0d, 0x43, 0xe6, 0xbe, 0x77, 0xec, 0x4d, + 0xb0, 0x62, 0x38, 0x90, 0xd2, 0x32, 0x7d, 0xcb, 0xf1, 0xa4, 0xab, 0x91, 0x5b, 0x18, 0xb1, 0x20, + 0xf0, 0x36, 0x2d, 0x3e, 0x24, 0x37, 0x00, 0xae, 0x6f, 0x05, 0xef, 0x43, 0xcf, 0x61, 0x0e, 0x82, + 0xec, 0x61, 0xee, 0x99, 0xcb, 0x0f, 0x38, 0x92, 0xc7, 0xad, 0xfc, 0xa5, 0x1f, 0xac, 0x7d, 0xbc, + 0x92, 0x82, 0x2e, 0xdd, 0x30, 0x74, 0x6c, 0xbc, 0x9e, 0xfd, 0x02, 0x93, 0x5d, 0x7e, 0xfc, 0x9f, + 0xaf, 0x28, 0x54, 0x7e, 0x72, 0x5b, 0x22, 0xc9, 0x0f, 0x71, 0x96, 0xe8, 0x79, 0x91, 0x18, 0x09, + 0xcf, 0xea, 0x87, 0x69, 0x34, 0x56, 0x7c, 0x4a, 0xcb, 0x7f, 0xfb, 0x73, 0x7c, 0xff, 0x4c, 0x6e, + 0xd5, 0x3a, 0xeb, 0x9b, 0x3f, 0x03, 0x00, 0x00, 0xff, 0xff, 0xdb, 0xae, 0x2d, 0xfb, 0x6e, 0x04, + 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/configuration.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/configuration.pb.go new file mode 100644 index 000000000..149cc541b --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/configuration.pb.go @@ -0,0 +1,227 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/configuration.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Represents a configuration within an Invocation associated with one or more +// ConfiguredTargets. It captures the environment and other settings that +// were used. +type Configuration struct { + // The format of this Configuration resource name must be: + // invocations/${INVOCATION_ID}/configs/${CONFIG_ID} + // The configuration ID of "default" should be preferred for the default + // configuration in a single-config invocation. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The resource ID components that identify the Configuration. They must match + // the resource name after proper encoding. + Id *Configuration_Id `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` + // The aggregate status for this configuration. + StatusAttributes *StatusAttributes `protobuf:"bytes,3,opt,name=status_attributes,json=statusAttributes,proto3" json:"status_attributes,omitempty"` + // Attributes that apply only to this configuration. + ConfigurationAttributes *ConfigurationAttributes `protobuf:"bytes,5,opt,name=configuration_attributes,json=configurationAttributes,proto3" json:"configuration_attributes,omitempty"` + // Arbitrary name-value pairs. + // This is implemented as a multi-map. Multiple properties are allowed with + // the same key. Properties will be returned in lexicographical order by key. + Properties []*Property `protobuf:"bytes,6,rep,name=properties,proto3" json:"properties,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Configuration) Reset() { *m = Configuration{} } +func (m *Configuration) String() string { return proto.CompactTextString(m) } +func (*Configuration) ProtoMessage() {} +func (*Configuration) Descriptor() ([]byte, []int) { + return fileDescriptor_configuration_4aca4ad29b870e00, []int{0} +} +func (m *Configuration) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Configuration.Unmarshal(m, b) +} +func (m *Configuration) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Configuration.Marshal(b, m, deterministic) +} +func (dst *Configuration) XXX_Merge(src proto.Message) { + xxx_messageInfo_Configuration.Merge(dst, src) +} +func (m *Configuration) XXX_Size() int { + return xxx_messageInfo_Configuration.Size(m) +} +func (m *Configuration) XXX_DiscardUnknown() { + xxx_messageInfo_Configuration.DiscardUnknown(m) +} + +var xxx_messageInfo_Configuration proto.InternalMessageInfo + +func (m *Configuration) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Configuration) GetId() *Configuration_Id { + if m != nil { + return m.Id + } + return nil +} + +func (m *Configuration) GetStatusAttributes() *StatusAttributes { + if m != nil { + return m.StatusAttributes + } + return nil +} + +func (m *Configuration) GetConfigurationAttributes() *ConfigurationAttributes { + if m != nil { + return m.ConfigurationAttributes + } + return nil +} + +func (m *Configuration) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +// The resource ID components that identify the Configuration. +type Configuration_Id struct { + // The Invocation ID. + InvocationId string `protobuf:"bytes,1,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` + // The Configuration ID. + ConfigurationId string `protobuf:"bytes,2,opt,name=configuration_id,json=configurationId,proto3" json:"configuration_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Configuration_Id) Reset() { *m = Configuration_Id{} } +func (m *Configuration_Id) String() string { return proto.CompactTextString(m) } +func (*Configuration_Id) ProtoMessage() {} +func (*Configuration_Id) Descriptor() ([]byte, []int) { + return fileDescriptor_configuration_4aca4ad29b870e00, []int{0, 0} +} +func (m *Configuration_Id) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Configuration_Id.Unmarshal(m, b) +} +func (m *Configuration_Id) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Configuration_Id.Marshal(b, m, deterministic) +} +func (dst *Configuration_Id) XXX_Merge(src proto.Message) { + xxx_messageInfo_Configuration_Id.Merge(dst, src) +} +func (m *Configuration_Id) XXX_Size() int { + return xxx_messageInfo_Configuration_Id.Size(m) +} +func (m *Configuration_Id) XXX_DiscardUnknown() { + xxx_messageInfo_Configuration_Id.DiscardUnknown(m) +} + +var xxx_messageInfo_Configuration_Id proto.InternalMessageInfo + +func (m *Configuration_Id) GetInvocationId() string { + if m != nil { + return m.InvocationId + } + return "" +} + +func (m *Configuration_Id) GetConfigurationId() string { + if m != nil { + return m.ConfigurationId + } + return "" +} + +// Attributes that apply only to the configuration. +type ConfigurationAttributes struct { + // The type of cpu. (e.g. "x86", "powerpc") + Cpu string `protobuf:"bytes,1,opt,name=cpu,proto3" json:"cpu,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ConfigurationAttributes) Reset() { *m = ConfigurationAttributes{} } +func (m *ConfigurationAttributes) String() string { return proto.CompactTextString(m) } +func (*ConfigurationAttributes) ProtoMessage() {} +func (*ConfigurationAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_configuration_4aca4ad29b870e00, []int{1} +} +func (m *ConfigurationAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ConfigurationAttributes.Unmarshal(m, b) +} +func (m *ConfigurationAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ConfigurationAttributes.Marshal(b, m, deterministic) +} +func (dst *ConfigurationAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_ConfigurationAttributes.Merge(dst, src) +} +func (m *ConfigurationAttributes) XXX_Size() int { + return xxx_messageInfo_ConfigurationAttributes.Size(m) +} +func (m *ConfigurationAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_ConfigurationAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_ConfigurationAttributes proto.InternalMessageInfo + +func (m *ConfigurationAttributes) GetCpu() string { + if m != nil { + return m.Cpu + } + return "" +} + +func init() { + proto.RegisterType((*Configuration)(nil), "google.devtools.resultstore.v2.Configuration") + proto.RegisterType((*Configuration_Id)(nil), "google.devtools.resultstore.v2.Configuration.Id") + proto.RegisterType((*ConfigurationAttributes)(nil), "google.devtools.resultstore.v2.ConfigurationAttributes") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/configuration.proto", fileDescriptor_configuration_4aca4ad29b870e00) +} + +var fileDescriptor_configuration_4aca4ad29b870e00 = []byte{ + // 335 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x92, 0x4f, 0x4f, 0xb3, 0x40, + 0x10, 0xc6, 0x03, 0xbc, 0x6f, 0x93, 0x4e, 0x6d, 0xac, 0x7b, 0x29, 0xe9, 0xc1, 0x34, 0xf5, 0x82, + 0x69, 0xb2, 0x18, 0x3c, 0x78, 0xf0, 0xe2, 0x9f, 0x8b, 0xdc, 0x1a, 0xf4, 0x64, 0x62, 0x0c, 0x65, + 0xd7, 0xcd, 0x26, 0xc0, 0xe0, 0xee, 0x42, 0xe2, 0x37, 0xf5, 0xe3, 0x18, 0x81, 0x2a, 0x18, 0x6b, + 0xf5, 0x36, 0xcc, 0xf0, 0x7b, 0x9e, 0x99, 0x9d, 0x81, 0x40, 0x20, 0x8a, 0x94, 0xfb, 0x8c, 0x57, + 0x06, 0x31, 0xd5, 0xbe, 0xe2, 0xba, 0x4c, 0x8d, 0x36, 0xa8, 0xb8, 0x5f, 0x05, 0x7e, 0x82, 0xf9, + 0x93, 0x14, 0xa5, 0x8a, 0x8d, 0xc4, 0x9c, 0x16, 0x0a, 0x0d, 0x92, 0xc3, 0x86, 0xa1, 0x1b, 0x86, + 0x76, 0x18, 0x5a, 0x05, 0xb3, 0xe5, 0x4e, 0xcd, 0x2c, 0xdb, 0x88, 0x2d, 0x5e, 0x1d, 0x18, 0x5f, + 0x77, 0x4d, 0x08, 0x81, 0x7f, 0x79, 0x9c, 0x71, 0xd7, 0x9a, 0x5b, 0xde, 0x30, 0xaa, 0x63, 0x72, + 0x01, 0xb6, 0x64, 0xae, 0x3d, 0xb7, 0xbc, 0x51, 0x70, 0x42, 0x7f, 0xf6, 0xa7, 0x3d, 0x39, 0x1a, + 0xb2, 0xc8, 0x96, 0x8c, 0x3c, 0xc0, 0x81, 0x36, 0xb1, 0x29, 0xf5, 0x63, 0x6c, 0x8c, 0x92, 0xeb, + 0xd2, 0x70, 0xed, 0x3a, 0xbf, 0x13, 0xbc, 0xad, 0xc1, 0xcb, 0x0f, 0x2e, 0x9a, 0xe8, 0x2f, 0x19, + 0xa2, 0xc0, 0xed, 0x3d, 0x55, 0xd7, 0xe5, 0x7f, 0xed, 0x72, 0xf6, 0xa7, 0xb6, 0x3b, 0x66, 0xd3, + 0xe4, 0xfb, 0x02, 0xb9, 0x01, 0x28, 0x14, 0x16, 0x5c, 0x19, 0xc9, 0xb5, 0x3b, 0x98, 0x3b, 0xde, + 0x28, 0xf0, 0x76, 0xb9, 0xac, 0x1a, 0xe2, 0x25, 0xea, 0xb0, 0xb3, 0x3b, 0xb0, 0x43, 0x46, 0x8e, + 0x60, 0x2c, 0xf3, 0x0a, 0x93, 0x66, 0x00, 0xc9, 0xda, 0x0d, 0xec, 0x7d, 0x26, 0x43, 0x46, 0x8e, + 0x61, 0xd2, 0x1f, 0xb4, 0xdd, 0xcb, 0x30, 0xda, 0xef, 0xe5, 0x43, 0xb6, 0x58, 0xc2, 0x74, 0xcb, + 0x4c, 0x64, 0x02, 0x4e, 0x52, 0x94, 0xad, 0xc1, 0x7b, 0x78, 0xf5, 0x0c, 0x8b, 0x04, 0xb3, 0x1d, + 0xdd, 0xaf, 0xac, 0xfb, 0xb0, 0xfd, 0x43, 0x60, 0x1a, 0xe7, 0x82, 0xa2, 0x12, 0xbe, 0xe0, 0x79, + 0x7d, 0x4b, 0x7e, 0x53, 0x8a, 0x0b, 0xa9, 0xb7, 0xdd, 0xde, 0x79, 0xe7, 0x73, 0x3d, 0xa8, 0xa9, + 0xd3, 0xb7, 0x00, 0x00, 0x00, 0xff, 0xff, 0xad, 0xea, 0x27, 0xe2, 0x04, 0x03, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/configured_target.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/configured_target.pb.go new file mode 100644 index 000000000..fbf18f2bc --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/configured_target.pb.go @@ -0,0 +1,304 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/configured_target.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Each ConfiguredTarget represents data for a given configuration of a given +// target in a given Invocation. +// Every ConfiguredTarget should have at least one Action as a child resource +// before the invocation is finalized. Refer to the Action's documentation for +// more info on this. +type ConfiguredTarget struct { + // The resource name. Its format must be: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID}/configuredTargets/${CONFIG_ID} + // where ${CONFIG_ID} must match the ID of an existing Configuration under + // this Invocation. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The resource ID components that identify the ConfiguredTarget. They must + // match the resource name after proper encoding. + Id *ConfiguredTarget_Id `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` + // The aggregate status for this configuration of this target. If testing + // was not requested, set this to the build status (e.g. BUILT or + // FAILED_TO_BUILD). + StatusAttributes *StatusAttributes `protobuf:"bytes,3,opt,name=status_attributes,json=statusAttributes,proto3" json:"status_attributes,omitempty"` + // Captures the start time and duration of this configured target. + Timing *Timing `protobuf:"bytes,4,opt,name=timing,proto3" json:"timing,omitempty"` + // Test specific attributes for this ConfiguredTarget. + TestAttributes *ConfiguredTestAttributes `protobuf:"bytes,6,opt,name=test_attributes,json=testAttributes,proto3" json:"test_attributes,omitempty"` + // Arbitrary name-value pairs. + // This is implemented as a multi-map. Multiple properties are allowed with + // the same key. Properties will be returned in lexicographical order by key. + Properties []*Property `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty"` + // A list of file references for configured target level files. + // The file IDs must be unique within this list. Duplicate file IDs will + // result in an error. Files will be returned in lexicographical order by ID. + Files []*File `protobuf:"bytes,8,rep,name=files,proto3" json:"files,omitempty"` + // Summary of aggregate coverage across all test actions for this target in + // this configuration. The server populates this for you in the + // post-processing phase. + CoverageSummaries []*LanguageCoverageSummary `protobuf:"bytes,9,rep,name=coverage_summaries,json=coverageSummaries,proto3" json:"coverage_summaries,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ConfiguredTarget) Reset() { *m = ConfiguredTarget{} } +func (m *ConfiguredTarget) String() string { return proto.CompactTextString(m) } +func (*ConfiguredTarget) ProtoMessage() {} +func (*ConfiguredTarget) Descriptor() ([]byte, []int) { + return fileDescriptor_configured_target_076f4fd568a29657, []int{0} +} +func (m *ConfiguredTarget) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ConfiguredTarget.Unmarshal(m, b) +} +func (m *ConfiguredTarget) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ConfiguredTarget.Marshal(b, m, deterministic) +} +func (dst *ConfiguredTarget) XXX_Merge(src proto.Message) { + xxx_messageInfo_ConfiguredTarget.Merge(dst, src) +} +func (m *ConfiguredTarget) XXX_Size() int { + return xxx_messageInfo_ConfiguredTarget.Size(m) +} +func (m *ConfiguredTarget) XXX_DiscardUnknown() { + xxx_messageInfo_ConfiguredTarget.DiscardUnknown(m) +} + +var xxx_messageInfo_ConfiguredTarget proto.InternalMessageInfo + +func (m *ConfiguredTarget) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *ConfiguredTarget) GetId() *ConfiguredTarget_Id { + if m != nil { + return m.Id + } + return nil +} + +func (m *ConfiguredTarget) GetStatusAttributes() *StatusAttributes { + if m != nil { + return m.StatusAttributes + } + return nil +} + +func (m *ConfiguredTarget) GetTiming() *Timing { + if m != nil { + return m.Timing + } + return nil +} + +func (m *ConfiguredTarget) GetTestAttributes() *ConfiguredTestAttributes { + if m != nil { + return m.TestAttributes + } + return nil +} + +func (m *ConfiguredTarget) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +func (m *ConfiguredTarget) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +func (m *ConfiguredTarget) GetCoverageSummaries() []*LanguageCoverageSummary { + if m != nil { + return m.CoverageSummaries + } + return nil +} + +// The resource ID components that identify the ConfiguredTarget. +type ConfiguredTarget_Id struct { + // The Invocation ID. + InvocationId string `protobuf:"bytes,1,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` + // The Target ID. + TargetId string `protobuf:"bytes,2,opt,name=target_id,json=targetId,proto3" json:"target_id,omitempty"` + // The Configuration ID. + ConfigurationId string `protobuf:"bytes,3,opt,name=configuration_id,json=configurationId,proto3" json:"configuration_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ConfiguredTarget_Id) Reset() { *m = ConfiguredTarget_Id{} } +func (m *ConfiguredTarget_Id) String() string { return proto.CompactTextString(m) } +func (*ConfiguredTarget_Id) ProtoMessage() {} +func (*ConfiguredTarget_Id) Descriptor() ([]byte, []int) { + return fileDescriptor_configured_target_076f4fd568a29657, []int{0, 0} +} +func (m *ConfiguredTarget_Id) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ConfiguredTarget_Id.Unmarshal(m, b) +} +func (m *ConfiguredTarget_Id) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ConfiguredTarget_Id.Marshal(b, m, deterministic) +} +func (dst *ConfiguredTarget_Id) XXX_Merge(src proto.Message) { + xxx_messageInfo_ConfiguredTarget_Id.Merge(dst, src) +} +func (m *ConfiguredTarget_Id) XXX_Size() int { + return xxx_messageInfo_ConfiguredTarget_Id.Size(m) +} +func (m *ConfiguredTarget_Id) XXX_DiscardUnknown() { + xxx_messageInfo_ConfiguredTarget_Id.DiscardUnknown(m) +} + +var xxx_messageInfo_ConfiguredTarget_Id proto.InternalMessageInfo + +func (m *ConfiguredTarget_Id) GetInvocationId() string { + if m != nil { + return m.InvocationId + } + return "" +} + +func (m *ConfiguredTarget_Id) GetTargetId() string { + if m != nil { + return m.TargetId + } + return "" +} + +func (m *ConfiguredTarget_Id) GetConfigurationId() string { + if m != nil { + return m.ConfigurationId + } + return "" +} + +// Attributes that apply only to test actions under this configured target. +type ConfiguredTestAttributes struct { + // Total number of test runs. For example, in bazel this is specified with + // --runs_per_test. Zero if runs_per_test is not used. + TotalRunCount int32 `protobuf:"varint,2,opt,name=total_run_count,json=totalRunCount,proto3" json:"total_run_count,omitempty"` + // Total number of test shards. Zero if shard count was not specified. + TotalShardCount int32 `protobuf:"varint,3,opt,name=total_shard_count,json=totalShardCount,proto3" json:"total_shard_count,omitempty"` + // How long test is allowed to run. + TimeoutDuration *duration.Duration `protobuf:"bytes,5,opt,name=timeout_duration,json=timeoutDuration,proto3" json:"timeout_duration,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ConfiguredTestAttributes) Reset() { *m = ConfiguredTestAttributes{} } +func (m *ConfiguredTestAttributes) String() string { return proto.CompactTextString(m) } +func (*ConfiguredTestAttributes) ProtoMessage() {} +func (*ConfiguredTestAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_configured_target_076f4fd568a29657, []int{1} +} +func (m *ConfiguredTestAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ConfiguredTestAttributes.Unmarshal(m, b) +} +func (m *ConfiguredTestAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ConfiguredTestAttributes.Marshal(b, m, deterministic) +} +func (dst *ConfiguredTestAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_ConfiguredTestAttributes.Merge(dst, src) +} +func (m *ConfiguredTestAttributes) XXX_Size() int { + return xxx_messageInfo_ConfiguredTestAttributes.Size(m) +} +func (m *ConfiguredTestAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_ConfiguredTestAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_ConfiguredTestAttributes proto.InternalMessageInfo + +func (m *ConfiguredTestAttributes) GetTotalRunCount() int32 { + if m != nil { + return m.TotalRunCount + } + return 0 +} + +func (m *ConfiguredTestAttributes) GetTotalShardCount() int32 { + if m != nil { + return m.TotalShardCount + } + return 0 +} + +func (m *ConfiguredTestAttributes) GetTimeoutDuration() *duration.Duration { + if m != nil { + return m.TimeoutDuration + } + return nil +} + +func init() { + proto.RegisterType((*ConfiguredTarget)(nil), "google.devtools.resultstore.v2.ConfiguredTarget") + proto.RegisterType((*ConfiguredTarget_Id)(nil), "google.devtools.resultstore.v2.ConfiguredTarget.Id") + proto.RegisterType((*ConfiguredTestAttributes)(nil), "google.devtools.resultstore.v2.ConfiguredTestAttributes") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/configured_target.proto", fileDescriptor_configured_target_076f4fd568a29657) +} + +var fileDescriptor_configured_target_076f4fd568a29657 = []byte{ + // 536 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x94, 0xdf, 0x8a, 0x13, 0x31, + 0x14, 0xc6, 0x69, 0xbb, 0x5d, 0xb7, 0x59, 0xd7, 0xb6, 0xb9, 0x1a, 0x2b, 0x48, 0xa9, 0xb2, 0x74, + 0x15, 0x66, 0xa4, 0x8b, 0x7f, 0x50, 0x10, 0xb4, 0x8b, 0x38, 0xe0, 0xc5, 0x32, 0xdd, 0x2b, 0x41, + 0x86, 0xb4, 0x49, 0x63, 0x60, 0x26, 0xe9, 0x26, 0x27, 0x85, 0x7d, 0x2f, 0x9f, 0xc3, 0x67, 0x92, + 0x26, 0x99, 0xda, 0x16, 0xd6, 0xc1, 0xbb, 0xe6, 0xeb, 0xf7, 0xfd, 0x72, 0x92, 0x9c, 0x33, 0xe8, + 0x0d, 0x57, 0x8a, 0x17, 0x2c, 0xa1, 0x6c, 0x0d, 0x4a, 0x15, 0x26, 0xd1, 0xcc, 0xd8, 0x02, 0x0c, + 0x28, 0xcd, 0x92, 0xf5, 0x24, 0x59, 0x28, 0xb9, 0x14, 0xdc, 0x6a, 0x46, 0x73, 0x20, 0x9a, 0x33, + 0x88, 0x57, 0x5a, 0x81, 0xc2, 0x4f, 0x7d, 0x2e, 0xae, 0x72, 0xf1, 0x4e, 0x2e, 0x5e, 0x4f, 0x06, + 0x2f, 0x6b, 0xb9, 0x65, 0xa9, 0xa4, 0x87, 0x0d, 0x5e, 0xd7, 0x9a, 0xd7, 0x4c, 0x13, 0xce, 0x72, + 0x63, 0xcb, 0x92, 0xe8, 0xbb, 0x10, 0xbb, 0xa8, 0x89, 0x2d, 0x45, 0xc1, 0x82, 0x35, 0x94, 0x9b, + 0xb8, 0xd5, 0xdc, 0x2e, 0x13, 0x6a, 0x35, 0x01, 0x51, 0x55, 0x30, 0xfa, 0xdd, 0x46, 0xbd, 0xe9, + 0xf6, 0xa8, 0x37, 0xee, 0xa4, 0x18, 0xa3, 0x23, 0x49, 0x4a, 0x16, 0x35, 0x86, 0x8d, 0x71, 0x27, + 0x73, 0xbf, 0xf1, 0x14, 0x35, 0x05, 0x8d, 0x9a, 0xc3, 0xc6, 0xf8, 0x74, 0x72, 0x19, 0xff, 0xfb, + 0x12, 0xe2, 0x43, 0x62, 0x9c, 0xd2, 0xac, 0x29, 0x28, 0xfe, 0x81, 0xfa, 0x06, 0x08, 0x58, 0x93, + 0x13, 0x00, 0x2d, 0xe6, 0x16, 0x98, 0x89, 0x5a, 0x8e, 0xf9, 0xaa, 0x8e, 0x39, 0x73, 0xc1, 0x4f, + 0xdb, 0x5c, 0xd6, 0x33, 0x07, 0x0a, 0xfe, 0x88, 0x8e, 0x41, 0x94, 0x42, 0xf2, 0xe8, 0xc8, 0x31, + 0xcf, 0xeb, 0x98, 0x37, 0xce, 0x9d, 0x85, 0x14, 0x26, 0xa8, 0x0b, 0xcc, 0xc0, 0x6e, 0x71, 0xc7, + 0x0e, 0xf4, 0xee, 0x3f, 0x0e, 0xcc, 0x0c, 0xec, 0x14, 0xf9, 0x08, 0xf6, 0xd6, 0xf8, 0x2b, 0x42, + 0x2b, 0xad, 0x56, 0x4c, 0x83, 0x60, 0x26, 0x7a, 0x30, 0x6c, 0x8d, 0x4f, 0x27, 0xe3, 0x3a, 0xfa, + 0xb5, 0x4f, 0xdc, 0x65, 0x3b, 0x59, 0xfc, 0x1e, 0xb5, 0x37, 0xef, 0x6c, 0xa2, 0x13, 0x07, 0x79, + 0x5e, 0x07, 0xf9, 0x22, 0x0a, 0x96, 0xf9, 0x08, 0x5e, 0x22, 0x7c, 0xd0, 0x5a, 0x9b, 0x6a, 0x3a, + 0x0e, 0xf4, 0xb6, 0x0e, 0xf4, 0x8d, 0x48, 0x6e, 0x09, 0x67, 0xd3, 0x40, 0x98, 0xf9, 0xde, 0xcc, + 0xfa, 0x8b, 0x3d, 0x41, 0x30, 0x33, 0xb8, 0x45, 0xcd, 0x94, 0xe2, 0x67, 0xe8, 0x4c, 0xc8, 0xb5, + 0x5a, 0xb8, 0xbe, 0xcb, 0x05, 0x0d, 0x7d, 0xf5, 0xf0, 0xaf, 0x98, 0x52, 0xfc, 0x04, 0x75, 0xfc, + 0x9c, 0xe5, 0xa1, 0xcd, 0x3a, 0xd9, 0x89, 0x17, 0x52, 0x8a, 0x2f, 0x50, 0xaf, 0x9a, 0xc7, 0x2d, + 0xa4, 0xe5, 0x3c, 0xdd, 0x3d, 0x3d, 0xa5, 0xa3, 0x5f, 0x0d, 0x14, 0xdd, 0xf7, 0x1a, 0xf8, 0x1c, + 0x75, 0x41, 0x01, 0x29, 0x72, 0x6d, 0x65, 0xbe, 0x50, 0x56, 0x82, 0xdb, 0xaa, 0x9d, 0x9d, 0x39, + 0x39, 0xb3, 0x72, 0xba, 0x11, 0xf1, 0x0b, 0xd4, 0xf7, 0x3e, 0xf3, 0x93, 0x68, 0x1a, 0x9c, 0x2d, + 0xe7, 0xf4, 0x80, 0xd9, 0x46, 0xf7, 0xde, 0x2b, 0xd4, 0x03, 0x51, 0x32, 0x65, 0x21, 0xaf, 0x66, + 0x2b, 0x6a, 0xbb, 0xae, 0x79, 0x5c, 0xdd, 0x64, 0x35, 0x7c, 0xf1, 0x55, 0x30, 0x64, 0xdd, 0x10, + 0xa9, 0x84, 0xcf, 0xb7, 0x68, 0xb4, 0x50, 0x65, 0xcd, 0xd5, 0x5f, 0x37, 0xbe, 0xa7, 0xc1, 0xc1, + 0x55, 0x41, 0x24, 0x8f, 0x95, 0xe6, 0x09, 0x67, 0xd2, 0x6d, 0x90, 0xf8, 0xbf, 0xc8, 0x4a, 0x98, + 0xfb, 0xbe, 0x0c, 0x1f, 0x76, 0x96, 0xf3, 0x63, 0x97, 0xba, 0xfc, 0x13, 0x00, 0x00, 0xff, 0xff, + 0x8f, 0xb1, 0xb2, 0xc2, 0x0a, 0x05, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/coverage.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/coverage.pb.go new file mode 100644 index 000000000..f268209d0 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/coverage.pb.go @@ -0,0 +1,310 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/coverage.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Describes line coverage for a file +type LineCoverage struct { + // Which source lines in the file represent the start of a statement that was + // instrumented to detect whether it was executed by the test. + // + // This is a bitfield where i-th bit corresponds to the i-th line. Divide line + // number by 8 to get index into byte array. Mod line number by 8 to get bit + // number (0 = LSB, 7 = MSB). + // + // A 1 denotes the line was instrumented. + // A 0 denotes the line was not instrumented. + InstrumentedLines []byte `protobuf:"bytes,1,opt,name=instrumented_lines,json=instrumentedLines,proto3" json:"instrumented_lines,omitempty"` + // Which of the instrumented source lines were executed by the test. Should + // include lines that were not instrumented. + // + // This is a bitfield where i-th bit corresponds to the i-th line. Divide line + // number by 8 to get index into byte array. Mod line number by 8 to get bit + // number (0 = LSB, 7 = MSB). + // + // A 1 denotes the line was executed. + // A 0 denotes the line was not executed. + ExecutedLines []byte `protobuf:"bytes,2,opt,name=executed_lines,json=executedLines,proto3" json:"executed_lines,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LineCoverage) Reset() { *m = LineCoverage{} } +func (m *LineCoverage) String() string { return proto.CompactTextString(m) } +func (*LineCoverage) ProtoMessage() {} +func (*LineCoverage) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_0d431861e82a62cf, []int{0} +} +func (m *LineCoverage) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LineCoverage.Unmarshal(m, b) +} +func (m *LineCoverage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LineCoverage.Marshal(b, m, deterministic) +} +func (dst *LineCoverage) XXX_Merge(src proto.Message) { + xxx_messageInfo_LineCoverage.Merge(dst, src) +} +func (m *LineCoverage) XXX_Size() int { + return xxx_messageInfo_LineCoverage.Size(m) +} +func (m *LineCoverage) XXX_DiscardUnknown() { + xxx_messageInfo_LineCoverage.DiscardUnknown(m) +} + +var xxx_messageInfo_LineCoverage proto.InternalMessageInfo + +func (m *LineCoverage) GetInstrumentedLines() []byte { + if m != nil { + return m.InstrumentedLines + } + return nil +} + +func (m *LineCoverage) GetExecutedLines() []byte { + if m != nil { + return m.ExecutedLines + } + return nil +} + +// Describes branch coverage for a file +type BranchCoverage struct { + // The field branch_present denotes the lines containing at least one branch. + // + // This is a bitfield where i-th bit corresponds to the i-th line. Divide line + // number by 8 to get index into byte array. Mod line number by 8 to get bit + // number (0 = LSB, 7 = MSB). + // + // A 1 denotes the line contains at least one branch. + // A 0 denotes the line contains no branches. + BranchPresent []byte `protobuf:"bytes,1,opt,name=branch_present,json=branchPresent,proto3" json:"branch_present,omitempty"` + // Contains the number of branches present, only for the lines which have the + // corresponding bit set in branch_present, in a relative order ignoring + // lines which do not have any branches. + BranchesInLine []int32 `protobuf:"varint,2,rep,packed,name=branches_in_line,json=branchesInLine,proto3" json:"branches_in_line,omitempty"` + // As each branch can have any one of the following three states: not + // executed, executed but not taken, executed and taken. + // + // This is a bitfield where i-th bit corresponds to the i-th line. Divide line + // number by 8 to get index into byte array. Mod line number by 8 to get bit + // number (0 = LSB, 7 = MSB). + // + // i-th bit of the following two byte arrays are used to denote the above + // mentioned states. + // + // not executed: i-th bit of executed == 0 && i-th bit of taken == 0 + // executed but not taken: i-th bit of executed == 1 && i-th bit of taken == 0 + // executed and taken: i-th bit of executed == 1 && i-th bit of taken == 1 + Executed []byte `protobuf:"bytes,3,opt,name=executed,proto3" json:"executed,omitempty"` + // Described above. + Taken []byte `protobuf:"bytes,4,opt,name=taken,proto3" json:"taken,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *BranchCoverage) Reset() { *m = BranchCoverage{} } +func (m *BranchCoverage) String() string { return proto.CompactTextString(m) } +func (*BranchCoverage) ProtoMessage() {} +func (*BranchCoverage) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_0d431861e82a62cf, []int{1} +} +func (m *BranchCoverage) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_BranchCoverage.Unmarshal(m, b) +} +func (m *BranchCoverage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_BranchCoverage.Marshal(b, m, deterministic) +} +func (dst *BranchCoverage) XXX_Merge(src proto.Message) { + xxx_messageInfo_BranchCoverage.Merge(dst, src) +} +func (m *BranchCoverage) XXX_Size() int { + return xxx_messageInfo_BranchCoverage.Size(m) +} +func (m *BranchCoverage) XXX_DiscardUnknown() { + xxx_messageInfo_BranchCoverage.DiscardUnknown(m) +} + +var xxx_messageInfo_BranchCoverage proto.InternalMessageInfo + +func (m *BranchCoverage) GetBranchPresent() []byte { + if m != nil { + return m.BranchPresent + } + return nil +} + +func (m *BranchCoverage) GetBranchesInLine() []int32 { + if m != nil { + return m.BranchesInLine + } + return nil +} + +func (m *BranchCoverage) GetExecuted() []byte { + if m != nil { + return m.Executed + } + return nil +} + +func (m *BranchCoverage) GetTaken() []byte { + if m != nil { + return m.Taken + } + return nil +} + +// Describes code coverage for a particular file under test. +type FileCoverage struct { + // Path of source file within the SourceContext of this Invocation. + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` + // Details of lines in a file required to calculate line coverage. + LineCoverage *LineCoverage `protobuf:"bytes,2,opt,name=line_coverage,json=lineCoverage,proto3" json:"line_coverage,omitempty"` + // Details of branches in a file required to calculate branch coverage. + BranchCoverage *BranchCoverage `protobuf:"bytes,3,opt,name=branch_coverage,json=branchCoverage,proto3" json:"branch_coverage,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *FileCoverage) Reset() { *m = FileCoverage{} } +func (m *FileCoverage) String() string { return proto.CompactTextString(m) } +func (*FileCoverage) ProtoMessage() {} +func (*FileCoverage) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_0d431861e82a62cf, []int{2} +} +func (m *FileCoverage) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_FileCoverage.Unmarshal(m, b) +} +func (m *FileCoverage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_FileCoverage.Marshal(b, m, deterministic) +} +func (dst *FileCoverage) XXX_Merge(src proto.Message) { + xxx_messageInfo_FileCoverage.Merge(dst, src) +} +func (m *FileCoverage) XXX_Size() int { + return xxx_messageInfo_FileCoverage.Size(m) +} +func (m *FileCoverage) XXX_DiscardUnknown() { + xxx_messageInfo_FileCoverage.DiscardUnknown(m) +} + +var xxx_messageInfo_FileCoverage proto.InternalMessageInfo + +func (m *FileCoverage) GetPath() string { + if m != nil { + return m.Path + } + return "" +} + +func (m *FileCoverage) GetLineCoverage() *LineCoverage { + if m != nil { + return m.LineCoverage + } + return nil +} + +func (m *FileCoverage) GetBranchCoverage() *BranchCoverage { + if m != nil { + return m.BranchCoverage + } + return nil +} + +// Describes code coverage for a build or test Action. This is used to store +// baseline coverage for build Actions and test coverage for test Actions. +type ActionCoverage struct { + // List of coverage info for all source files that the TestResult covers. + FileCoverages []*FileCoverage `protobuf:"bytes,2,rep,name=file_coverages,json=fileCoverages,proto3" json:"file_coverages,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ActionCoverage) Reset() { *m = ActionCoverage{} } +func (m *ActionCoverage) String() string { return proto.CompactTextString(m) } +func (*ActionCoverage) ProtoMessage() {} +func (*ActionCoverage) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_0d431861e82a62cf, []int{3} +} +func (m *ActionCoverage) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ActionCoverage.Unmarshal(m, b) +} +func (m *ActionCoverage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ActionCoverage.Marshal(b, m, deterministic) +} +func (dst *ActionCoverage) XXX_Merge(src proto.Message) { + xxx_messageInfo_ActionCoverage.Merge(dst, src) +} +func (m *ActionCoverage) XXX_Size() int { + return xxx_messageInfo_ActionCoverage.Size(m) +} +func (m *ActionCoverage) XXX_DiscardUnknown() { + xxx_messageInfo_ActionCoverage.DiscardUnknown(m) +} + +var xxx_messageInfo_ActionCoverage proto.InternalMessageInfo + +func (m *ActionCoverage) GetFileCoverages() []*FileCoverage { + if m != nil { + return m.FileCoverages + } + return nil +} + +func init() { + proto.RegisterType((*LineCoverage)(nil), "google.devtools.resultstore.v2.LineCoverage") + proto.RegisterType((*BranchCoverage)(nil), "google.devtools.resultstore.v2.BranchCoverage") + proto.RegisterType((*FileCoverage)(nil), "google.devtools.resultstore.v2.FileCoverage") + proto.RegisterType((*ActionCoverage)(nil), "google.devtools.resultstore.v2.ActionCoverage") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/coverage.proto", fileDescriptor_coverage_0d431861e82a62cf) +} + +var fileDescriptor_coverage_0d431861e82a62cf = []byte{ + // 372 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x52, 0x4d, 0x6b, 0xe3, 0x30, + 0x10, 0xc5, 0xf9, 0x58, 0x76, 0x67, 0x6d, 0xef, 0xae, 0xd8, 0x83, 0xd9, 0xc3, 0x12, 0x0c, 0x81, + 0x1c, 0x1a, 0x19, 0xd2, 0x63, 0x4f, 0x4d, 0xa1, 0x10, 0xe8, 0x21, 0x75, 0x0f, 0x85, 0x5e, 0x8c, + 0xe3, 0x4c, 0x1c, 0x51, 0x45, 0x72, 0x25, 0xc5, 0xf4, 0x7f, 0xf4, 0x7f, 0xf5, 0x37, 0x95, 0xc8, + 0x1f, 0x71, 0x0f, 0x6d, 0x7a, 0xd3, 0xbc, 0x79, 0xf3, 0xde, 0xcc, 0x43, 0x30, 0xcd, 0xa5, 0xcc, + 0x39, 0x46, 0x6b, 0x2c, 0x8d, 0x94, 0x5c, 0x47, 0x0a, 0xf5, 0x9e, 0x1b, 0x6d, 0xa4, 0xc2, 0xa8, + 0x9c, 0x45, 0x99, 0x2c, 0x51, 0xa5, 0x39, 0xd2, 0x42, 0x49, 0x23, 0xc9, 0xff, 0x8a, 0x4e, 0x1b, + 0x3a, 0xed, 0xd0, 0x69, 0x39, 0x0b, 0xd7, 0xe0, 0xde, 0x30, 0x81, 0x57, 0xf5, 0x14, 0x99, 0x02, + 0x61, 0x42, 0x1b, 0xb5, 0xdf, 0xa1, 0x30, 0xb8, 0x4e, 0x38, 0x13, 0xa8, 0x03, 0x67, 0xe4, 0x4c, + 0xdc, 0xf8, 0x4f, 0xb7, 0x73, 0x98, 0xd2, 0x64, 0x0c, 0x3e, 0x3e, 0x63, 0xb6, 0x3f, 0x52, 0x7b, + 0x96, 0xea, 0x35, 0xa8, 0xa5, 0x85, 0x2f, 0x0e, 0xf8, 0x73, 0x95, 0x8a, 0x6c, 0xdb, 0x1a, 0x8d, + 0xc1, 0x5f, 0x59, 0x24, 0x29, 0x14, 0x6a, 0x14, 0xa6, 0x36, 0xf1, 0x2a, 0x74, 0x59, 0x81, 0x64, + 0x02, 0xbf, 0x2b, 0x00, 0x75, 0xc2, 0x84, 0xf5, 0x08, 0x7a, 0xa3, 0xfe, 0x64, 0x18, 0xfb, 0x0d, + 0xbe, 0x10, 0x07, 0x13, 0xf2, 0x0f, 0xbe, 0x37, 0xa6, 0x41, 0xdf, 0x4a, 0xb5, 0x35, 0xf9, 0x0b, + 0x43, 0x93, 0x3e, 0xa2, 0x08, 0x06, 0xb6, 0x51, 0x15, 0xe1, 0xab, 0x03, 0xee, 0x35, 0xe3, 0xc7, + 0xe3, 0x09, 0x0c, 0x8a, 0xd4, 0x6c, 0xed, 0x26, 0x3f, 0x62, 0xfb, 0x26, 0xb7, 0xe0, 0x1d, 0x4c, + 0x93, 0x26, 0x57, 0x7b, 0xe0, 0xcf, 0xd9, 0x19, 0xfd, 0x3c, 0x58, 0xda, 0x4d, 0x35, 0x76, 0x79, + 0x37, 0xe3, 0x7b, 0xf8, 0x55, 0x9f, 0xde, 0x8a, 0xf6, 0xad, 0x28, 0x3d, 0x25, 0xfa, 0x3e, 0xc3, + 0x26, 0x82, 0xa6, 0x0e, 0x11, 0xfc, 0xcb, 0xcc, 0x30, 0x29, 0x5a, 0xab, 0x3b, 0xf0, 0x37, 0x8c, + 0x1f, 0xb7, 0xd7, 0x36, 0xbc, 0x2f, 0xac, 0xdf, 0xcd, 0x25, 0xf6, 0x36, 0x9d, 0x4a, 0xcf, 0x9f, + 0x20, 0xcc, 0xe4, 0xee, 0x84, 0xc2, 0xd2, 0x79, 0x58, 0xd4, 0x8c, 0x5c, 0xf2, 0x54, 0xe4, 0x54, + 0xaa, 0x3c, 0xca, 0x51, 0xd8, 0x7f, 0x19, 0x55, 0xad, 0xb4, 0x60, 0xfa, 0xa3, 0x9f, 0x7c, 0xd1, + 0x29, 0x57, 0xdf, 0xec, 0xd4, 0xf9, 0x5b, 0x00, 0x00, 0x00, 0xff, 0xff, 0x3e, 0x92, 0x57, 0x34, + 0xfe, 0x02, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/coverage_summary.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/coverage_summary.pb.go new file mode 100644 index 000000000..5eb6e29d3 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/coverage_summary.pb.go @@ -0,0 +1,229 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/coverage_summary.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Summary of line coverage +type LineCoverageSummary struct { + // Number of lines instrumented for coverage. + InstrumentedLineCount int32 `protobuf:"varint,1,opt,name=instrumented_line_count,json=instrumentedLineCount,proto3" json:"instrumented_line_count,omitempty"` + // Number of instrumented lines that were executed by the test. + ExecutedLineCount int32 `protobuf:"varint,2,opt,name=executed_line_count,json=executedLineCount,proto3" json:"executed_line_count,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LineCoverageSummary) Reset() { *m = LineCoverageSummary{} } +func (m *LineCoverageSummary) String() string { return proto.CompactTextString(m) } +func (*LineCoverageSummary) ProtoMessage() {} +func (*LineCoverageSummary) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_summary_c482cb7c4cc6b938, []int{0} +} +func (m *LineCoverageSummary) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LineCoverageSummary.Unmarshal(m, b) +} +func (m *LineCoverageSummary) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LineCoverageSummary.Marshal(b, m, deterministic) +} +func (dst *LineCoverageSummary) XXX_Merge(src proto.Message) { + xxx_messageInfo_LineCoverageSummary.Merge(dst, src) +} +func (m *LineCoverageSummary) XXX_Size() int { + return xxx_messageInfo_LineCoverageSummary.Size(m) +} +func (m *LineCoverageSummary) XXX_DiscardUnknown() { + xxx_messageInfo_LineCoverageSummary.DiscardUnknown(m) +} + +var xxx_messageInfo_LineCoverageSummary proto.InternalMessageInfo + +func (m *LineCoverageSummary) GetInstrumentedLineCount() int32 { + if m != nil { + return m.InstrumentedLineCount + } + return 0 +} + +func (m *LineCoverageSummary) GetExecutedLineCount() int32 { + if m != nil { + return m.ExecutedLineCount + } + return 0 +} + +// Summary of branch coverage +// A branch may be: +// * not executed. Counted only in total. +// * executed but not taken. Appears in total and executed. +// * executed and taken. Appears in all three fields. +type BranchCoverageSummary struct { + // The number of branches present in the file. + TotalBranchCount int32 `protobuf:"varint,1,opt,name=total_branch_count,json=totalBranchCount,proto3" json:"total_branch_count,omitempty"` + // The number of branches executed out of the total branches present. + // A branch is executed when its condition is evaluated. + // This is <= total_branch_count as not all branches are executed. + ExecutedBranchCount int32 `protobuf:"varint,2,opt,name=executed_branch_count,json=executedBranchCount,proto3" json:"executed_branch_count,omitempty"` + // The number of branches taken out of the total branches executed. + // A branch is taken when its condition is satisfied. + // This is <= executed_branch_count as not all executed branches are taken. + TakenBranchCount int32 `protobuf:"varint,3,opt,name=taken_branch_count,json=takenBranchCount,proto3" json:"taken_branch_count,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *BranchCoverageSummary) Reset() { *m = BranchCoverageSummary{} } +func (m *BranchCoverageSummary) String() string { return proto.CompactTextString(m) } +func (*BranchCoverageSummary) ProtoMessage() {} +func (*BranchCoverageSummary) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_summary_c482cb7c4cc6b938, []int{1} +} +func (m *BranchCoverageSummary) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_BranchCoverageSummary.Unmarshal(m, b) +} +func (m *BranchCoverageSummary) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_BranchCoverageSummary.Marshal(b, m, deterministic) +} +func (dst *BranchCoverageSummary) XXX_Merge(src proto.Message) { + xxx_messageInfo_BranchCoverageSummary.Merge(dst, src) +} +func (m *BranchCoverageSummary) XXX_Size() int { + return xxx_messageInfo_BranchCoverageSummary.Size(m) +} +func (m *BranchCoverageSummary) XXX_DiscardUnknown() { + xxx_messageInfo_BranchCoverageSummary.DiscardUnknown(m) +} + +var xxx_messageInfo_BranchCoverageSummary proto.InternalMessageInfo + +func (m *BranchCoverageSummary) GetTotalBranchCount() int32 { + if m != nil { + return m.TotalBranchCount + } + return 0 +} + +func (m *BranchCoverageSummary) GetExecutedBranchCount() int32 { + if m != nil { + return m.ExecutedBranchCount + } + return 0 +} + +func (m *BranchCoverageSummary) GetTakenBranchCount() int32 { + if m != nil { + return m.TakenBranchCount + } + return 0 +} + +// Summary of coverage in each language +type LanguageCoverageSummary struct { + // This summary is for all files written in this programming language. + Language Language `protobuf:"varint,1,opt,name=language,proto3,enum=google.devtools.resultstore.v2.Language" json:"language,omitempty"` + // Summary of lines covered vs instrumented. + LineSummary *LineCoverageSummary `protobuf:"bytes,2,opt,name=line_summary,json=lineSummary,proto3" json:"line_summary,omitempty"` + // Summary of branch coverage. + BranchSummary *BranchCoverageSummary `protobuf:"bytes,3,opt,name=branch_summary,json=branchSummary,proto3" json:"branch_summary,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LanguageCoverageSummary) Reset() { *m = LanguageCoverageSummary{} } +func (m *LanguageCoverageSummary) String() string { return proto.CompactTextString(m) } +func (*LanguageCoverageSummary) ProtoMessage() {} +func (*LanguageCoverageSummary) Descriptor() ([]byte, []int) { + return fileDescriptor_coverage_summary_c482cb7c4cc6b938, []int{2} +} +func (m *LanguageCoverageSummary) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LanguageCoverageSummary.Unmarshal(m, b) +} +func (m *LanguageCoverageSummary) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LanguageCoverageSummary.Marshal(b, m, deterministic) +} +func (dst *LanguageCoverageSummary) XXX_Merge(src proto.Message) { + xxx_messageInfo_LanguageCoverageSummary.Merge(dst, src) +} +func (m *LanguageCoverageSummary) XXX_Size() int { + return xxx_messageInfo_LanguageCoverageSummary.Size(m) +} +func (m *LanguageCoverageSummary) XXX_DiscardUnknown() { + xxx_messageInfo_LanguageCoverageSummary.DiscardUnknown(m) +} + +var xxx_messageInfo_LanguageCoverageSummary proto.InternalMessageInfo + +func (m *LanguageCoverageSummary) GetLanguage() Language { + if m != nil { + return m.Language + } + return Language_LANGUAGE_UNSPECIFIED +} + +func (m *LanguageCoverageSummary) GetLineSummary() *LineCoverageSummary { + if m != nil { + return m.LineSummary + } + return nil +} + +func (m *LanguageCoverageSummary) GetBranchSummary() *BranchCoverageSummary { + if m != nil { + return m.BranchSummary + } + return nil +} + +func init() { + proto.RegisterType((*LineCoverageSummary)(nil), "google.devtools.resultstore.v2.LineCoverageSummary") + proto.RegisterType((*BranchCoverageSummary)(nil), "google.devtools.resultstore.v2.BranchCoverageSummary") + proto.RegisterType((*LanguageCoverageSummary)(nil), "google.devtools.resultstore.v2.LanguageCoverageSummary") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/coverage_summary.proto", fileDescriptor_coverage_summary_c482cb7c4cc6b938) +} + +var fileDescriptor_coverage_summary_c482cb7c4cc6b938 = []byte{ + // 358 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x92, 0x41, 0x4b, 0xeb, 0x40, + 0x14, 0x85, 0x49, 0xcb, 0x7b, 0x3c, 0xa6, 0xef, 0x95, 0x67, 0x4a, 0xa9, 0xb8, 0x10, 0xc9, 0xaa, + 0xa0, 0x4c, 0x20, 0xa5, 0x6e, 0xdc, 0x55, 0x37, 0x42, 0x17, 0x12, 0xc1, 0x85, 0x08, 0x61, 0x9a, + 0x5e, 0xc6, 0x60, 0x32, 0xb7, 0x4e, 0x26, 0x41, 0x17, 0xfe, 0x0c, 0x7f, 0x83, 0x7f, 0x53, 0x32, + 0x33, 0x09, 0x69, 0xab, 0x66, 0x39, 0x9c, 0x7b, 0xbe, 0x73, 0x66, 0xee, 0x90, 0x39, 0x47, 0xe4, + 0x29, 0xf8, 0x6b, 0x28, 0x15, 0x62, 0x9a, 0xfb, 0x12, 0xf2, 0x22, 0x55, 0xb9, 0x42, 0x09, 0x7e, + 0x19, 0xf8, 0x31, 0x96, 0x20, 0x19, 0x87, 0x28, 0x2f, 0xb2, 0x8c, 0xc9, 0x57, 0xba, 0x91, 0xa8, + 0xd0, 0x3d, 0x36, 0x36, 0x5a, 0xdb, 0x68, 0xcb, 0x46, 0xcb, 0xe0, 0xe8, 0xb4, 0x13, 0x9b, 0x65, + 0x28, 0x0c, 0xcc, 0x7b, 0x23, 0xa3, 0x65, 0x22, 0xe0, 0xd2, 0x46, 0xdd, 0x9a, 0x24, 0xf7, 0x9c, + 0x4c, 0x12, 0x91, 0x2b, 0x59, 0x64, 0x20, 0x14, 0xac, 0xa3, 0x34, 0x11, 0x10, 0xc5, 0x58, 0x08, + 0x75, 0xe8, 0x9c, 0x38, 0xd3, 0x5f, 0xe1, 0xb8, 0x2d, 0x1b, 0x42, 0x21, 0x94, 0x4b, 0xc9, 0x08, + 0x5e, 0x20, 0x2e, 0x76, 0x3c, 0x3d, 0xed, 0x39, 0xa8, 0xa5, 0x66, 0xde, 0xfb, 0x70, 0xc8, 0x78, + 0x21, 0x99, 0x88, 0x1f, 0x77, 0x1b, 0x9c, 0x11, 0x57, 0xa1, 0x62, 0x69, 0xb4, 0xd2, 0xf2, 0x56, + 0xf8, 0x7f, 0xad, 0xd4, 0xbe, 0x2a, 0x37, 0x20, 0xe3, 0x26, 0x77, 0xcb, 0x60, 0x92, 0x9b, 0x52, + 0x6d, 0x4f, 0x95, 0xc0, 0x9e, 0x40, 0x6c, 0x1b, 0xfa, 0x36, 0xa1, 0x52, 0x5a, 0xd3, 0xde, 0x7b, + 0x8f, 0x4c, 0x96, 0x4c, 0xf0, 0x82, 0xf1, 0xbd, 0xd7, 0xba, 0x22, 0x7f, 0x52, 0x2b, 0xe9, 0x86, + 0xc3, 0x60, 0x4a, 0x7f, 0x5e, 0x12, 0xad, 0x51, 0x61, 0xe3, 0x74, 0xef, 0xc8, 0x5f, 0xfd, 0x64, + 0x76, 0xdb, 0xba, 0xfa, 0x20, 0x98, 0x75, 0x92, 0xf6, 0xd7, 0x17, 0x0e, 0x2a, 0x50, 0xdd, 0xee, + 0x81, 0x0c, 0xed, 0x0d, 0x6b, 0x72, 0x5f, 0x93, 0xe7, 0x5d, 0xe4, 0x2f, 0x17, 0x13, 0xfe, 0x33, + 0x30, 0x7b, 0x5c, 0x3c, 0x13, 0x2f, 0xc6, 0xac, 0x03, 0x75, 0xe3, 0xdc, 0x5f, 0xdb, 0x09, 0x8e, + 0xd5, 0x85, 0x29, 0x4a, 0xee, 0x73, 0x10, 0xfa, 0x13, 0xfa, 0x46, 0x62, 0x9b, 0x24, 0xff, 0xee, + 0xd3, 0x5e, 0xb4, 0x8e, 0xab, 0xdf, 0xda, 0x35, 0xfb, 0x0c, 0x00, 0x00, 0xff, 0xff, 0x0f, 0x65, + 0x24, 0xb5, 0x40, 0x03, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/file.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/file.pb.go new file mode 100644 index 000000000..17ac84150 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/file.pb.go @@ -0,0 +1,238 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/file.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import wrappers "github.com/golang/protobuf/ptypes/wrappers" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// The metadata for a file or an archive file entry. +type File struct { + // The identifier of the file or archive entry. + // User-provided, must be unique for the repeated field it is in. When an + // Append RPC is called with a Files field populated, if a File already exists + // with this ID, that File will be overwritten with the new File proto. + Uid string `protobuf:"bytes,1,opt,name=uid,proto3" json:"uid,omitempty"` + // The URI of a file. + // This could also be the URI of an entire archive. + // Most log data doesn't need to be stored forever, so a ttl is suggested. + // Note that if you ever move or delete the file at this URI, the link from + // the server will be broken. + Uri string `protobuf:"bytes,2,opt,name=uri,proto3" json:"uri,omitempty"` + // (Optional) The length of the file in bytes. Allows the filesize to be + // shown in the UI. Omit if file is still being written or length is + // not known. This could also be the length of an entire archive. + Length *wrappers.Int64Value `protobuf:"bytes,3,opt,name=length,proto3" json:"length,omitempty"` + // (Optional) The content-type (aka MIME-type) of the file. This is sent to + // the web browser so it knows how to handle the file. (e.g. text/plain, + // image/jpeg, text/html, etc). For zip archives, use "application/zip". + ContentType string `protobuf:"bytes,4,opt,name=content_type,json=contentType,proto3" json:"content_type,omitempty"` + // (Optional) If the above path, length, and content_type are referring to an + // archive, and you wish to refer to a particular entry within that archive, + // put the particular archive entry data here. + ArchiveEntry *ArchiveEntry `protobuf:"bytes,5,opt,name=archive_entry,json=archiveEntry,proto3" json:"archive_entry,omitempty"` + // (Optional) A url to a content display app/site for this file or archive + // entry. + ContentViewer string `protobuf:"bytes,6,opt,name=content_viewer,json=contentViewer,proto3" json:"content_viewer,omitempty"` + // (Optional) Whether to hide this file or archive entry in the UI. Defaults + // to false. A checkbox lets users see hidden files, but they're hidden by + // default. + Hidden bool `protobuf:"varint,7,opt,name=hidden,proto3" json:"hidden,omitempty"` + // (Optional) A short description of what this file or archive entry + // contains. This description should help someone viewing the list of these + // files to understand the purpose of this file and what they would want to + // view it for. + Description string `protobuf:"bytes,8,opt,name=description,proto3" json:"description,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *File) Reset() { *m = File{} } +func (m *File) String() string { return proto.CompactTextString(m) } +func (*File) ProtoMessage() {} +func (*File) Descriptor() ([]byte, []int) { + return fileDescriptor_file_2d7ceba920c83983, []int{0} +} +func (m *File) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_File.Unmarshal(m, b) +} +func (m *File) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_File.Marshal(b, m, deterministic) +} +func (dst *File) XXX_Merge(src proto.Message) { + xxx_messageInfo_File.Merge(dst, src) +} +func (m *File) XXX_Size() int { + return xxx_messageInfo_File.Size(m) +} +func (m *File) XXX_DiscardUnknown() { + xxx_messageInfo_File.DiscardUnknown(m) +} + +var xxx_messageInfo_File proto.InternalMessageInfo + +func (m *File) GetUid() string { + if m != nil { + return m.Uid + } + return "" +} + +func (m *File) GetUri() string { + if m != nil { + return m.Uri + } + return "" +} + +func (m *File) GetLength() *wrappers.Int64Value { + if m != nil { + return m.Length + } + return nil +} + +func (m *File) GetContentType() string { + if m != nil { + return m.ContentType + } + return "" +} + +func (m *File) GetArchiveEntry() *ArchiveEntry { + if m != nil { + return m.ArchiveEntry + } + return nil +} + +func (m *File) GetContentViewer() string { + if m != nil { + return m.ContentViewer + } + return "" +} + +func (m *File) GetHidden() bool { + if m != nil { + return m.Hidden + } + return false +} + +func (m *File) GetDescription() string { + if m != nil { + return m.Description + } + return "" +} + +// Information specific to an entry in an archive. +type ArchiveEntry struct { + // The relative path of the entry within the archive. + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` + // (Optional) The uncompressed length of the archive entry in bytes. Allows + // the entry size to be shown in the UI. Omit if the length is not known. + Length *wrappers.Int64Value `protobuf:"bytes,2,opt,name=length,proto3" json:"length,omitempty"` + // (Optional) The content-type (aka MIME-type) of the archive entry. (e.g. + // text/plain, image/jpeg, text/html, etc). This is sent to the web browser + // so it knows how to handle the entry. + ContentType string `protobuf:"bytes,3,opt,name=content_type,json=contentType,proto3" json:"content_type,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ArchiveEntry) Reset() { *m = ArchiveEntry{} } +func (m *ArchiveEntry) String() string { return proto.CompactTextString(m) } +func (*ArchiveEntry) ProtoMessage() {} +func (*ArchiveEntry) Descriptor() ([]byte, []int) { + return fileDescriptor_file_2d7ceba920c83983, []int{1} +} +func (m *ArchiveEntry) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ArchiveEntry.Unmarshal(m, b) +} +func (m *ArchiveEntry) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ArchiveEntry.Marshal(b, m, deterministic) +} +func (dst *ArchiveEntry) XXX_Merge(src proto.Message) { + xxx_messageInfo_ArchiveEntry.Merge(dst, src) +} +func (m *ArchiveEntry) XXX_Size() int { + return xxx_messageInfo_ArchiveEntry.Size(m) +} +func (m *ArchiveEntry) XXX_DiscardUnknown() { + xxx_messageInfo_ArchiveEntry.DiscardUnknown(m) +} + +var xxx_messageInfo_ArchiveEntry proto.InternalMessageInfo + +func (m *ArchiveEntry) GetPath() string { + if m != nil { + return m.Path + } + return "" +} + +func (m *ArchiveEntry) GetLength() *wrappers.Int64Value { + if m != nil { + return m.Length + } + return nil +} + +func (m *ArchiveEntry) GetContentType() string { + if m != nil { + return m.ContentType + } + return "" +} + +func init() { + proto.RegisterType((*File)(nil), "google.devtools.resultstore.v2.File") + proto.RegisterType((*ArchiveEntry)(nil), "google.devtools.resultstore.v2.ArchiveEntry") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/file.proto", fileDescriptor_file_2d7ceba920c83983) +} + +var fileDescriptor_file_2d7ceba920c83983 = []byte{ + // 350 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x52, 0xdf, 0x4b, 0xeb, 0x30, + 0x14, 0xa6, 0xdb, 0x6e, 0xef, 0x6e, 0xb6, 0x5d, 0x24, 0x0f, 0x12, 0x14, 0xa4, 0x0e, 0x84, 0x09, + 0x92, 0xc0, 0x26, 0xbe, 0xf8, 0xa4, 0xa0, 0xb0, 0x37, 0x2d, 0xb2, 0x07, 0x5f, 0x46, 0xd7, 0x9e, + 0xb5, 0x81, 0x98, 0xc4, 0x34, 0xed, 0x98, 0x7f, 0x98, 0x7f, 0x9f, 0x2c, 0xcd, 0xa0, 0x20, 0x2a, + 0xf8, 0x76, 0xce, 0x77, 0xbe, 0x1f, 0xed, 0x47, 0xd0, 0x79, 0xae, 0x54, 0x2e, 0x80, 0x65, 0x50, + 0x5b, 0xa5, 0x44, 0xc9, 0x0c, 0x94, 0x95, 0xb0, 0xa5, 0x55, 0x06, 0x58, 0x3d, 0x65, 0x6b, 0x2e, + 0x80, 0x6a, 0xa3, 0xac, 0xc2, 0x27, 0x0d, 0x95, 0xee, 0xa9, 0xb4, 0x45, 0xa5, 0xf5, 0xf4, 0xc8, + 0xdf, 0x99, 0x63, 0xaf, 0xaa, 0x35, 0xdb, 0x98, 0x44, 0x6b, 0x30, 0x65, 0xa3, 0x1f, 0xbf, 0x77, + 0x50, 0xef, 0x9e, 0x0b, 0xc0, 0x07, 0xa8, 0x5b, 0xf1, 0x8c, 0x04, 0x51, 0x30, 0xf9, 0x17, 0xef, + 0x46, 0x87, 0x18, 0x4e, 0x3a, 0x1e, 0x31, 0x1c, 0xcf, 0x50, 0x28, 0x40, 0xe6, 0xb6, 0x20, 0xdd, + 0x28, 0x98, 0x0c, 0xa6, 0xc7, 0xd4, 0xa7, 0xef, 0xdd, 0xe9, 0x5c, 0xda, 0xab, 0xcb, 0x45, 0x22, + 0x2a, 0x88, 0x3d, 0x15, 0x9f, 0xa2, 0x61, 0xaa, 0xa4, 0x05, 0x69, 0x97, 0x76, 0xab, 0x81, 0xf4, + 0x9c, 0xdf, 0xc0, 0x63, 0x4f, 0x5b, 0x0d, 0xf8, 0x11, 0x8d, 0x12, 0x93, 0x16, 0xbc, 0x86, 0x25, + 0x48, 0x6b, 0xb6, 0xe4, 0x8f, 0xb3, 0xbf, 0xa0, 0xdf, 0xff, 0x1c, 0xbd, 0x69, 0x44, 0x77, 0x3b, + 0x4d, 0x3c, 0x4c, 0x5a, 0x1b, 0x3e, 0x43, 0xff, 0xf7, 0xa9, 0x35, 0x87, 0x0d, 0x18, 0x12, 0xba, + 0xdc, 0x91, 0x47, 0x17, 0x0e, 0xc4, 0x87, 0x28, 0x2c, 0x78, 0x96, 0x81, 0x24, 0x7f, 0xa3, 0x60, + 0xd2, 0x8f, 0xfd, 0x86, 0x23, 0x34, 0xc8, 0xa0, 0x4c, 0x0d, 0xd7, 0x96, 0x2b, 0x49, 0xfa, 0xcd, + 0x37, 0xb7, 0xa0, 0xf1, 0x1b, 0x1a, 0xb6, 0xe3, 0x31, 0x46, 0x3d, 0x9d, 0xd8, 0xc2, 0x17, 0xe8, + 0xe6, 0x56, 0x5f, 0x9d, 0xdf, 0xf7, 0xd5, 0xfd, 0xd4, 0xd7, 0xed, 0x2b, 0x1a, 0xa7, 0xea, 0xe5, + 0x87, 0x76, 0x1e, 0x82, 0xe7, 0xb9, 0x67, 0xe4, 0x4a, 0x24, 0x32, 0xa7, 0xca, 0xe4, 0x2c, 0x07, + 0xe9, 0xc2, 0x59, 0x73, 0x4a, 0x34, 0x2f, 0xbf, 0x7a, 0x66, 0xd7, 0xad, 0x75, 0x15, 0x3a, 0xd5, + 0xec, 0x23, 0x00, 0x00, 0xff, 0xff, 0x65, 0xa9, 0x29, 0x4f, 0x9b, 0x02, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/file_set.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/file_set.pb.go new file mode 100644 index 000000000..65adff00e --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/file_set.pb.go @@ -0,0 +1,169 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/file_set.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Represents a set of files within an Invocation. Can contain other file sets. +type FileSet struct { + // The format of this FileSet resource name must be: + // invocations/${INVOCATION_ID}/fileSets/${FILE_SET_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The resource ID components that identify the file set. They must match the + // resource name after proper encoding. + Id *FileSet_Id `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` + // List of names of other file sets that are referenced from this one. + // Each name must point to a file set under the same invocation. The name + // format must be: invocations/${INVOCATION_ID}/fileSets/${FILE_SET_ID} + FileSets []string `protobuf:"bytes,3,rep,name=file_sets,json=fileSets,proto3" json:"file_sets,omitempty"` + // Files that are contained within this file set. + Files []*File `protobuf:"bytes,4,rep,name=files,proto3" json:"files,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *FileSet) Reset() { *m = FileSet{} } +func (m *FileSet) String() string { return proto.CompactTextString(m) } +func (*FileSet) ProtoMessage() {} +func (*FileSet) Descriptor() ([]byte, []int) { + return fileDescriptor_file_set_7e71c7510a3af0da, []int{0} +} +func (m *FileSet) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_FileSet.Unmarshal(m, b) +} +func (m *FileSet) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_FileSet.Marshal(b, m, deterministic) +} +func (dst *FileSet) XXX_Merge(src proto.Message) { + xxx_messageInfo_FileSet.Merge(dst, src) +} +func (m *FileSet) XXX_Size() int { + return xxx_messageInfo_FileSet.Size(m) +} +func (m *FileSet) XXX_DiscardUnknown() { + xxx_messageInfo_FileSet.DiscardUnknown(m) +} + +var xxx_messageInfo_FileSet proto.InternalMessageInfo + +func (m *FileSet) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *FileSet) GetId() *FileSet_Id { + if m != nil { + return m.Id + } + return nil +} + +func (m *FileSet) GetFileSets() []string { + if m != nil { + return m.FileSets + } + return nil +} + +func (m *FileSet) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +// The resource ID components that identify the FileSet. +type FileSet_Id struct { + // The Invocation ID. + InvocationId string `protobuf:"bytes,1,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` + // The FileSet ID. + FileSetId string `protobuf:"bytes,2,opt,name=file_set_id,json=fileSetId,proto3" json:"file_set_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *FileSet_Id) Reset() { *m = FileSet_Id{} } +func (m *FileSet_Id) String() string { return proto.CompactTextString(m) } +func (*FileSet_Id) ProtoMessage() {} +func (*FileSet_Id) Descriptor() ([]byte, []int) { + return fileDescriptor_file_set_7e71c7510a3af0da, []int{0, 0} +} +func (m *FileSet_Id) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_FileSet_Id.Unmarshal(m, b) +} +func (m *FileSet_Id) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_FileSet_Id.Marshal(b, m, deterministic) +} +func (dst *FileSet_Id) XXX_Merge(src proto.Message) { + xxx_messageInfo_FileSet_Id.Merge(dst, src) +} +func (m *FileSet_Id) XXX_Size() int { + return xxx_messageInfo_FileSet_Id.Size(m) +} +func (m *FileSet_Id) XXX_DiscardUnknown() { + xxx_messageInfo_FileSet_Id.DiscardUnknown(m) +} + +var xxx_messageInfo_FileSet_Id proto.InternalMessageInfo + +func (m *FileSet_Id) GetInvocationId() string { + if m != nil { + return m.InvocationId + } + return "" +} + +func (m *FileSet_Id) GetFileSetId() string { + if m != nil { + return m.FileSetId + } + return "" +} + +func init() { + proto.RegisterType((*FileSet)(nil), "google.devtools.resultstore.v2.FileSet") + proto.RegisterType((*FileSet_Id)(nil), "google.devtools.resultstore.v2.FileSet.Id") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/file_set.proto", fileDescriptor_file_set_7e71c7510a3af0da) +} + +var fileDescriptor_file_set_7e71c7510a3af0da = []byte{ + // 271 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x90, 0xbb, 0x4f, 0xc3, 0x30, + 0x10, 0x87, 0x95, 0xa4, 0x3c, 0x72, 0x81, 0xc5, 0x53, 0x54, 0xa4, 0x2a, 0x2a, 0x0c, 0x01, 0x09, + 0x5b, 0x0a, 0x5b, 0xd9, 0x18, 0x90, 0xbc, 0xa1, 0xb0, 0xb1, 0x44, 0xa1, 0xbe, 0x5a, 0x96, 0xdc, + 0x5c, 0x89, 0x4d, 0xfe, 0x7b, 0x24, 0x44, 0x1e, 0xa2, 0x0b, 0xb4, 0xdb, 0xd9, 0xfe, 0xee, 0x77, + 0x9f, 0x0f, 0xee, 0x35, 0x91, 0xb6, 0x28, 0x14, 0x76, 0x9e, 0xc8, 0x3a, 0xd1, 0xa2, 0xfb, 0xb4, + 0xde, 0x79, 0x6a, 0x51, 0x74, 0x85, 0xd8, 0x18, 0x8b, 0x95, 0x43, 0xcf, 0x77, 0x2d, 0x79, 0x62, + 0x8b, 0x01, 0xe7, 0x13, 0xce, 0xf7, 0x70, 0xde, 0x15, 0xf3, 0xdb, 0x23, 0xe2, 0x86, 0xa8, 0xe5, + 0x57, 0x00, 0x67, 0xcf, 0xc6, 0xe2, 0x2b, 0x7a, 0xc6, 0x60, 0xd6, 0xd4, 0x5b, 0x4c, 0x83, 0x2c, + 0xc8, 0xe3, 0xb2, 0xaf, 0xd9, 0x0a, 0x42, 0xa3, 0xd2, 0x30, 0x0b, 0xf2, 0xa4, 0xb8, 0xe3, 0xff, + 0xcf, 0xe5, 0x63, 0x10, 0x97, 0xaa, 0x0c, 0x8d, 0x62, 0x57, 0x10, 0x4f, 0xe2, 0x2e, 0x8d, 0xb2, + 0x28, 0x8f, 0xcb, 0xf3, 0xcd, 0x80, 0x38, 0xb6, 0x82, 0x93, 0x9f, 0xda, 0xa5, 0xb3, 0x2c, 0xca, + 0x93, 0xe2, 0xe6, 0x98, 0xec, 0x72, 0x68, 0x99, 0x4b, 0x08, 0xa5, 0x62, 0xd7, 0x70, 0x69, 0x9a, + 0x8e, 0xd6, 0xb5, 0x37, 0xd4, 0x54, 0x46, 0x8d, 0xde, 0x17, 0xbf, 0x97, 0x52, 0xb1, 0x05, 0x24, + 0x93, 0x43, 0x35, 0x7e, 0x24, 0x2e, 0xe3, 0xd1, 0x42, 0xaa, 0xa7, 0x0f, 0x58, 0xae, 0x69, 0x7b, + 0x60, 0xf8, 0x4b, 0xf0, 0x26, 0x47, 0x42, 0x93, 0xad, 0x1b, 0xcd, 0xa9, 0xd5, 0x42, 0x63, 0xd3, + 0xef, 0x50, 0x0c, 0x4f, 0xf5, 0xce, 0xb8, 0xbf, 0x36, 0xfe, 0xb8, 0x77, 0x7c, 0x3f, 0xed, 0xbb, + 0x1e, 0xbe, 0x03, 0x00, 0x00, 0xff, 0xff, 0xc9, 0x67, 0xbb, 0x64, 0xf5, 0x01, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/invocation.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/invocation.pb.go new file mode 100644 index 000000000..caf183439 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/invocation.pb.go @@ -0,0 +1,485 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/invocation.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// An Invocation typically represents the result of running a tool. Each has a +// unique ID, typically generated by the server. Target resources under each +// Invocation contain the bulk of the data. +type Invocation struct { + // The resource name. Its format must be: + // invocations/${INVOCATION_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The resource ID components that identify the Invocation. They must match + // the resource name after proper encoding. + Id *Invocation_Id `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` + // The aggregate status of the invocation. + StatusAttributes *StatusAttributes `protobuf:"bytes,3,opt,name=status_attributes,json=statusAttributes,proto3" json:"status_attributes,omitempty"` + // When this invocation started and its duration. + Timing *Timing `protobuf:"bytes,4,opt,name=timing,proto3" json:"timing,omitempty"` + // Attributes of this invocation. + InvocationAttributes *InvocationAttributes `protobuf:"bytes,5,opt,name=invocation_attributes,json=invocationAttributes,proto3" json:"invocation_attributes,omitempty"` + // The workspace the tool was run in. + WorkspaceInfo *WorkspaceInfo `protobuf:"bytes,6,opt,name=workspace_info,json=workspaceInfo,proto3" json:"workspace_info,omitempty"` + // Arbitrary name-value pairs. + // This is implemented as a multi-map. Multiple properties are allowed with + // the same key. Properties will be returned in lexicographical order by key. + Properties []*Property `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty"` + // A list of file references for invocation level files. + // The file IDs must be unique within this list. Duplicate file IDs will + // result in an error. Files will be returned in lexicographical order by ID. + // Use this field to specify build logs, and other invocation level logs. + Files []*File `protobuf:"bytes,8,rep,name=files,proto3" json:"files,omitempty"` + // Summary of aggregate coverage across all Actions in this Invocation. + // the server populates this for you in the post-processing phase. + CoverageSummaries []*LanguageCoverageSummary `protobuf:"bytes,9,rep,name=coverage_summaries,json=coverageSummaries,proto3" json:"coverage_summaries,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Invocation) Reset() { *m = Invocation{} } +func (m *Invocation) String() string { return proto.CompactTextString(m) } +func (*Invocation) ProtoMessage() {} +func (*Invocation) Descriptor() ([]byte, []int) { + return fileDescriptor_invocation_b6bc8ed548b71603, []int{0} +} +func (m *Invocation) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Invocation.Unmarshal(m, b) +} +func (m *Invocation) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Invocation.Marshal(b, m, deterministic) +} +func (dst *Invocation) XXX_Merge(src proto.Message) { + xxx_messageInfo_Invocation.Merge(dst, src) +} +func (m *Invocation) XXX_Size() int { + return xxx_messageInfo_Invocation.Size(m) +} +func (m *Invocation) XXX_DiscardUnknown() { + xxx_messageInfo_Invocation.DiscardUnknown(m) +} + +var xxx_messageInfo_Invocation proto.InternalMessageInfo + +func (m *Invocation) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Invocation) GetId() *Invocation_Id { + if m != nil { + return m.Id + } + return nil +} + +func (m *Invocation) GetStatusAttributes() *StatusAttributes { + if m != nil { + return m.StatusAttributes + } + return nil +} + +func (m *Invocation) GetTiming() *Timing { + if m != nil { + return m.Timing + } + return nil +} + +func (m *Invocation) GetInvocationAttributes() *InvocationAttributes { + if m != nil { + return m.InvocationAttributes + } + return nil +} + +func (m *Invocation) GetWorkspaceInfo() *WorkspaceInfo { + if m != nil { + return m.WorkspaceInfo + } + return nil +} + +func (m *Invocation) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +func (m *Invocation) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +func (m *Invocation) GetCoverageSummaries() []*LanguageCoverageSummary { + if m != nil { + return m.CoverageSummaries + } + return nil +} + +// The resource ID components that identify the Invocation. +type Invocation_Id struct { + // The Invocation ID. + InvocationId string `protobuf:"bytes,1,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Invocation_Id) Reset() { *m = Invocation_Id{} } +func (m *Invocation_Id) String() string { return proto.CompactTextString(m) } +func (*Invocation_Id) ProtoMessage() {} +func (*Invocation_Id) Descriptor() ([]byte, []int) { + return fileDescriptor_invocation_b6bc8ed548b71603, []int{0, 0} +} +func (m *Invocation_Id) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Invocation_Id.Unmarshal(m, b) +} +func (m *Invocation_Id) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Invocation_Id.Marshal(b, m, deterministic) +} +func (dst *Invocation_Id) XXX_Merge(src proto.Message) { + xxx_messageInfo_Invocation_Id.Merge(dst, src) +} +func (m *Invocation_Id) XXX_Size() int { + return xxx_messageInfo_Invocation_Id.Size(m) +} +func (m *Invocation_Id) XXX_DiscardUnknown() { + xxx_messageInfo_Invocation_Id.DiscardUnknown(m) +} + +var xxx_messageInfo_Invocation_Id proto.InternalMessageInfo + +func (m *Invocation_Id) GetInvocationId() string { + if m != nil { + return m.InvocationId + } + return "" +} + +// If known, represents the state of the user/build-system workspace. +type WorkspaceContext struct { + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *WorkspaceContext) Reset() { *m = WorkspaceContext{} } +func (m *WorkspaceContext) String() string { return proto.CompactTextString(m) } +func (*WorkspaceContext) ProtoMessage() {} +func (*WorkspaceContext) Descriptor() ([]byte, []int) { + return fileDescriptor_invocation_b6bc8ed548b71603, []int{1} +} +func (m *WorkspaceContext) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_WorkspaceContext.Unmarshal(m, b) +} +func (m *WorkspaceContext) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_WorkspaceContext.Marshal(b, m, deterministic) +} +func (dst *WorkspaceContext) XXX_Merge(src proto.Message) { + xxx_messageInfo_WorkspaceContext.Merge(dst, src) +} +func (m *WorkspaceContext) XXX_Size() int { + return xxx_messageInfo_WorkspaceContext.Size(m) +} +func (m *WorkspaceContext) XXX_DiscardUnknown() { + xxx_messageInfo_WorkspaceContext.DiscardUnknown(m) +} + +var xxx_messageInfo_WorkspaceContext proto.InternalMessageInfo + +// Describes the workspace under which the tool was invoked, this includes +// information that was fed into the command, the source code referenced, and +// the tool itself. +type WorkspaceInfo struct { + // Data about the workspace that might be useful for debugging. + WorkspaceContext *WorkspaceContext `protobuf:"bytes,1,opt,name=workspace_context,json=workspaceContext,proto3" json:"workspace_context,omitempty"` + // Where the tool was invoked + Hostname string `protobuf:"bytes,3,opt,name=hostname,proto3" json:"hostname,omitempty"` + // The client's working directory where the build/test was run from. + WorkingDirectory string `protobuf:"bytes,4,opt,name=working_directory,json=workingDirectory,proto3" json:"working_directory,omitempty"` + // Tools should set tool_tag to the name of the tool or use case. + ToolTag string `protobuf:"bytes,5,opt,name=tool_tag,json=toolTag,proto3" json:"tool_tag,omitempty"` + // The command lines invoked. The first command line is the one typed by the + // user, then each one after that should be an expansion of the previous + // command line. + CommandLines []*CommandLine `protobuf:"bytes,7,rep,name=command_lines,json=commandLines,proto3" json:"command_lines,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *WorkspaceInfo) Reset() { *m = WorkspaceInfo{} } +func (m *WorkspaceInfo) String() string { return proto.CompactTextString(m) } +func (*WorkspaceInfo) ProtoMessage() {} +func (*WorkspaceInfo) Descriptor() ([]byte, []int) { + return fileDescriptor_invocation_b6bc8ed548b71603, []int{2} +} +func (m *WorkspaceInfo) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_WorkspaceInfo.Unmarshal(m, b) +} +func (m *WorkspaceInfo) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_WorkspaceInfo.Marshal(b, m, deterministic) +} +func (dst *WorkspaceInfo) XXX_Merge(src proto.Message) { + xxx_messageInfo_WorkspaceInfo.Merge(dst, src) +} +func (m *WorkspaceInfo) XXX_Size() int { + return xxx_messageInfo_WorkspaceInfo.Size(m) +} +func (m *WorkspaceInfo) XXX_DiscardUnknown() { + xxx_messageInfo_WorkspaceInfo.DiscardUnknown(m) +} + +var xxx_messageInfo_WorkspaceInfo proto.InternalMessageInfo + +func (m *WorkspaceInfo) GetWorkspaceContext() *WorkspaceContext { + if m != nil { + return m.WorkspaceContext + } + return nil +} + +func (m *WorkspaceInfo) GetHostname() string { + if m != nil { + return m.Hostname + } + return "" +} + +func (m *WorkspaceInfo) GetWorkingDirectory() string { + if m != nil { + return m.WorkingDirectory + } + return "" +} + +func (m *WorkspaceInfo) GetToolTag() string { + if m != nil { + return m.ToolTag + } + return "" +} + +func (m *WorkspaceInfo) GetCommandLines() []*CommandLine { + if m != nil { + return m.CommandLines + } + return nil +} + +// The command and arguments that produced this Invocation. +type CommandLine struct { + // A label describing this command line. + Label string `protobuf:"bytes,1,opt,name=label,proto3" json:"label,omitempty"` + // The command-line tool that is run: argv[0]. + Tool string `protobuf:"bytes,2,opt,name=tool,proto3" json:"tool,omitempty"` + // The arguments to the above tool: argv[1]...argv[N]. + Args []string `protobuf:"bytes,3,rep,name=args,proto3" json:"args,omitempty"` + // The actual command that was run with the tool. (e.g. "build", or "test") + // Omit if the tool doesn't accept a command. + // This is a duplicate of one of the fields in args. + Command string `protobuf:"bytes,4,opt,name=command,proto3" json:"command,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *CommandLine) Reset() { *m = CommandLine{} } +func (m *CommandLine) String() string { return proto.CompactTextString(m) } +func (*CommandLine) ProtoMessage() {} +func (*CommandLine) Descriptor() ([]byte, []int) { + return fileDescriptor_invocation_b6bc8ed548b71603, []int{3} +} +func (m *CommandLine) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_CommandLine.Unmarshal(m, b) +} +func (m *CommandLine) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_CommandLine.Marshal(b, m, deterministic) +} +func (dst *CommandLine) XXX_Merge(src proto.Message) { + xxx_messageInfo_CommandLine.Merge(dst, src) +} +func (m *CommandLine) XXX_Size() int { + return xxx_messageInfo_CommandLine.Size(m) +} +func (m *CommandLine) XXX_DiscardUnknown() { + xxx_messageInfo_CommandLine.DiscardUnknown(m) +} + +var xxx_messageInfo_CommandLine proto.InternalMessageInfo + +func (m *CommandLine) GetLabel() string { + if m != nil { + return m.Label + } + return "" +} + +func (m *CommandLine) GetTool() string { + if m != nil { + return m.Tool + } + return "" +} + +func (m *CommandLine) GetArgs() []string { + if m != nil { + return m.Args + } + return nil +} + +func (m *CommandLine) GetCommand() string { + if m != nil { + return m.Command + } + return "" +} + +// Attributes that apply to all invocations. +type InvocationAttributes struct { + // The project ID this invocation is associated with. This must be + // set in the CreateInvocation call, and can't be changed. + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` + // The list of users in the command chain. The first user in this sequence + // is the one who instigated the first command in the chain. + Users []string `protobuf:"bytes,2,rep,name=users,proto3" json:"users,omitempty"` + // Labels to categorize this invocation. + // This is implemented as a set. All labels will be unique. Any duplicate + // labels added will be ignored. Labels will be returned in lexicographical + // order. Labels should be short, easy to read, and you + // shouldn't have more than a handful. + // Labels should match regex \w([- \w]*\w)? + // Labels should not be used for unique properties such as unique IDs. + // Use properties in cases that don't meet these conditions. + Labels []string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *InvocationAttributes) Reset() { *m = InvocationAttributes{} } +func (m *InvocationAttributes) String() string { return proto.CompactTextString(m) } +func (*InvocationAttributes) ProtoMessage() {} +func (*InvocationAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_invocation_b6bc8ed548b71603, []int{4} +} +func (m *InvocationAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_InvocationAttributes.Unmarshal(m, b) +} +func (m *InvocationAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_InvocationAttributes.Marshal(b, m, deterministic) +} +func (dst *InvocationAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_InvocationAttributes.Merge(dst, src) +} +func (m *InvocationAttributes) XXX_Size() int { + return xxx_messageInfo_InvocationAttributes.Size(m) +} +func (m *InvocationAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_InvocationAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_InvocationAttributes proto.InternalMessageInfo + +func (m *InvocationAttributes) GetProjectId() string { + if m != nil { + return m.ProjectId + } + return "" +} + +func (m *InvocationAttributes) GetUsers() []string { + if m != nil { + return m.Users + } + return nil +} + +func (m *InvocationAttributes) GetLabels() []string { + if m != nil { + return m.Labels + } + return nil +} + +func init() { + proto.RegisterType((*Invocation)(nil), "google.devtools.resultstore.v2.Invocation") + proto.RegisterType((*Invocation_Id)(nil), "google.devtools.resultstore.v2.Invocation.Id") + proto.RegisterType((*WorkspaceContext)(nil), "google.devtools.resultstore.v2.WorkspaceContext") + proto.RegisterType((*WorkspaceInfo)(nil), "google.devtools.resultstore.v2.WorkspaceInfo") + proto.RegisterType((*CommandLine)(nil), "google.devtools.resultstore.v2.CommandLine") + proto.RegisterType((*InvocationAttributes)(nil), "google.devtools.resultstore.v2.InvocationAttributes") +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/invocation.proto", fileDescriptor_invocation_b6bc8ed548b71603) +} + +var fileDescriptor_invocation_b6bc8ed548b71603 = []byte{ + // 631 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x95, 0x71, 0x6b, 0x13, 0x3f, + 0x18, 0xc7, 0x69, 0xbb, 0x76, 0xeb, 0xb3, 0xf5, 0xc7, 0x16, 0xf6, 0x93, 0xb3, 0xa0, 0x8c, 0x2a, + 0xd2, 0x31, 0xbc, 0x93, 0xaa, 0x08, 0x8a, 0x82, 0x4e, 0xc4, 0x83, 0xfd, 0x31, 0xb2, 0x81, 0x20, + 0xc8, 0x91, 0xdd, 0xa5, 0x31, 0x7a, 0x97, 0xd4, 0x24, 0xd7, 0xb9, 0xf7, 0xe3, 0x8b, 0xf2, 0xe5, + 0x48, 0x72, 0x77, 0xbd, 0x5b, 0x99, 0xde, 0xfe, 0xbb, 0xe7, 0x69, 0xbe, 0x9f, 0x3c, 0x79, 0xf2, + 0xcd, 0x53, 0x08, 0x98, 0x94, 0x2c, 0xa5, 0x41, 0x42, 0x97, 0x46, 0xca, 0x54, 0x07, 0x8a, 0xea, + 0x3c, 0x35, 0xda, 0x48, 0x45, 0x83, 0xe5, 0x2c, 0xe0, 0x62, 0x29, 0x63, 0x62, 0xb8, 0x14, 0xfe, + 0x42, 0x49, 0x23, 0xd1, 0xfd, 0x42, 0xe0, 0x57, 0x02, 0xbf, 0x21, 0xf0, 0x97, 0xb3, 0xf1, 0x51, + 0x0b, 0x30, 0x96, 0x59, 0x56, 0xc1, 0xc6, 0xcf, 0x5b, 0x17, 0x2f, 0xa9, 0x22, 0x8c, 0x46, 0x3a, + 0xcf, 0x32, 0xa2, 0xae, 0x4a, 0xd9, 0x61, 0x8b, 0x6c, 0xce, 0x53, 0x5a, 0x2c, 0x9d, 0xfc, 0xee, + 0x03, 0x84, 0xab, 0x33, 0x20, 0x04, 0x1b, 0x82, 0x64, 0xd4, 0xeb, 0x1c, 0x74, 0xa6, 0x43, 0xec, + 0xbe, 0xd1, 0x6b, 0xe8, 0xf2, 0xc4, 0xeb, 0x1e, 0x74, 0xa6, 0xdb, 0xb3, 0xc7, 0xfe, 0xbf, 0x8f, + 0xe7, 0xd7, 0x2c, 0x3f, 0x4c, 0x70, 0x97, 0x27, 0xe8, 0x0b, 0xec, 0x69, 0x43, 0x4c, 0xae, 0x23, + 0x62, 0x8c, 0xe2, 0x17, 0xb9, 0xa1, 0xda, 0xeb, 0x39, 0xda, 0x93, 0x36, 0xda, 0x99, 0x13, 0xbe, + 0x5d, 0xe9, 0xf0, 0xae, 0x5e, 0xcb, 0xa0, 0x37, 0x30, 0x30, 0x3c, 0xe3, 0x82, 0x79, 0x1b, 0x8e, + 0xf9, 0xa8, 0x8d, 0x79, 0xee, 0x56, 0xe3, 0x52, 0x85, 0x38, 0xfc, 0x5f, 0xdf, 0x61, 0xb3, 0xc4, + 0xbe, 0xc3, 0x3d, 0xbb, 0xfd, 0x81, 0x1b, 0x65, 0xee, 0xf3, 0x1b, 0xb2, 0xe8, 0x1c, 0xfe, 0xbb, + 0x94, 0xea, 0xbb, 0x5e, 0x90, 0x98, 0x46, 0x5c, 0xcc, 0xa5, 0x37, 0xb8, 0x5d, 0x53, 0x3f, 0x55, + 0xaa, 0x50, 0xcc, 0x25, 0x1e, 0x5d, 0x36, 0x43, 0xf4, 0x11, 0x60, 0xa1, 0xe4, 0x82, 0x2a, 0xc3, + 0xa9, 0xf6, 0x36, 0x0f, 0x7a, 0xd3, 0xed, 0xd9, 0xb4, 0x8d, 0x78, 0x5a, 0x28, 0xae, 0x70, 0x43, + 0x8b, 0x5e, 0x42, 0xdf, 0x3a, 0x43, 0x7b, 0x5b, 0x0e, 0xf2, 0xb0, 0x0d, 0xf2, 0x81, 0xa7, 0x14, + 0x17, 0x12, 0x34, 0x07, 0xb4, 0x66, 0x46, 0x5b, 0xcd, 0xd0, 0x81, 0x5e, 0xb4, 0x81, 0x4e, 0x88, + 0x60, 0x39, 0x61, 0xf4, 0xb8, 0x24, 0x9c, 0x15, 0x6e, 0xc6, 0x7b, 0xf1, 0xb5, 0x04, 0xa7, 0x7a, + 0x7c, 0x08, 0xdd, 0x30, 0x41, 0x0f, 0x60, 0xd4, 0xb8, 0x34, 0x9e, 0x94, 0x7e, 0xdd, 0xa9, 0x93, + 0x61, 0x32, 0x41, 0xb0, 0xbb, 0x6a, 0xdc, 0xb1, 0x14, 0x86, 0xfe, 0x34, 0x93, 0x5f, 0x5d, 0x18, + 0x5d, 0xeb, 0xa6, 0xb5, 0x67, 0x7d, 0x29, 0x71, 0xb1, 0xcc, 0xe1, 0x6e, 0x61, 0xcf, 0x75, 0x3c, + 0xde, 0xbd, 0x5c, 0xcb, 0xa0, 0x31, 0x6c, 0x7d, 0x95, 0xda, 0xb8, 0x47, 0xd5, 0x73, 0x45, 0xae, + 0x62, 0x74, 0x54, 0x6c, 0xcd, 0x05, 0x8b, 0x12, 0xae, 0x68, 0x6c, 0xa4, 0xba, 0x72, 0x2e, 0x1e, + 0x16, 0x20, 0x2e, 0xd8, 0xfb, 0x2a, 0x8f, 0xee, 0xc2, 0x96, 0xad, 0x21, 0x32, 0x84, 0x39, 0x6b, + 0x0e, 0xf1, 0xa6, 0x8d, 0xcf, 0x09, 0x43, 0xa7, 0x30, 0xb2, 0x53, 0x83, 0x88, 0x24, 0x4a, 0xb9, + 0x58, 0x99, 0xe0, 0xa8, 0xad, 0xfc, 0xe3, 0x42, 0x74, 0xc2, 0x05, 0xc5, 0x3b, 0x71, 0x1d, 0xe8, + 0x09, 0x85, 0xed, 0xc6, 0x8f, 0x68, 0x1f, 0xfa, 0x29, 0xb9, 0xa0, 0x69, 0xd9, 0xe6, 0x22, 0xb0, + 0xb3, 0xc2, 0x62, 0xdd, 0x64, 0x18, 0x62, 0xf7, 0x6d, 0x73, 0x44, 0x31, 0xfb, 0xbe, 0x7b, 0x36, + 0x67, 0xbf, 0x91, 0x07, 0x9b, 0x25, 0xbc, 0x3c, 0x5c, 0x15, 0x4e, 0x62, 0xd8, 0xbf, 0xe9, 0xf9, + 0xa0, 0x7b, 0xce, 0xd2, 0xdf, 0x68, 0x6c, 0xea, 0xbb, 0x1d, 0x96, 0x99, 0x30, 0xb1, 0xe5, 0xe4, + 0x9a, 0x2a, 0xed, 0x75, 0xdd, 0x2e, 0x45, 0x80, 0xee, 0xc0, 0xc0, 0xd5, 0x55, 0x6d, 0x5e, 0x46, + 0xef, 0x7e, 0xc0, 0x24, 0x96, 0x59, 0x4b, 0x2f, 0x4e, 0x3b, 0x9f, 0xc3, 0x72, 0x05, 0x93, 0x29, + 0x11, 0xcc, 0x97, 0x8a, 0x05, 0x8c, 0x0a, 0x37, 0x25, 0xcb, 0x3f, 0x01, 0xb2, 0xe0, 0xfa, 0x6f, + 0x33, 0xf5, 0x55, 0x23, 0xbc, 0x18, 0x38, 0xd5, 0xd3, 0x3f, 0x01, 0x00, 0x00, 0xff, 0xff, 0xfa, + 0x5b, 0x4b, 0x30, 0x3d, 0x06, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/resultstore_download.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/resultstore_download.pb.go new file mode 100644 index 000000000..dc105e3a7 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/resultstore_download.pb.go @@ -0,0 +1,2204 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/resultstore_download.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Request passed into GetInvocation +type GetInvocationRequest struct { + // The name of the invocation to retrieve. It must match this format: + // invocations/${INVOCATION_ID} + // where INVOCATION_ID must be an RFC 4122-compliant random UUID. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetInvocationRequest) Reset() { *m = GetInvocationRequest{} } +func (m *GetInvocationRequest) String() string { return proto.CompactTextString(m) } +func (*GetInvocationRequest) ProtoMessage() {} +func (*GetInvocationRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{0} +} +func (m *GetInvocationRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetInvocationRequest.Unmarshal(m, b) +} +func (m *GetInvocationRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetInvocationRequest.Marshal(b, m, deterministic) +} +func (dst *GetInvocationRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetInvocationRequest.Merge(dst, src) +} +func (m *GetInvocationRequest) XXX_Size() int { + return xxx_messageInfo_GetInvocationRequest.Size(m) +} +func (m *GetInvocationRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetInvocationRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetInvocationRequest proto.InternalMessageInfo + +func (m *GetInvocationRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request passed into SearchInvocations +type SearchInvocationsRequest struct { + // The maximum number of items to return. Zero means all, but may be capped by + // the server. + PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Options for pagination. + // + // Types that are valid to be assigned to PageStart: + // *SearchInvocationsRequest_PageToken + // *SearchInvocationsRequest_Offset + PageStart isSearchInvocationsRequest_PageStart `protobuf_oneof:"page_start"` + // A filtering query string. + Query string `protobuf:"bytes,4,opt,name=query,proto3" json:"query,omitempty"` + // The project id to search under. + ProjectId string `protobuf:"bytes,5,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SearchInvocationsRequest) Reset() { *m = SearchInvocationsRequest{} } +func (m *SearchInvocationsRequest) String() string { return proto.CompactTextString(m) } +func (*SearchInvocationsRequest) ProtoMessage() {} +func (*SearchInvocationsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{1} +} +func (m *SearchInvocationsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SearchInvocationsRequest.Unmarshal(m, b) +} +func (m *SearchInvocationsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SearchInvocationsRequest.Marshal(b, m, deterministic) +} +func (dst *SearchInvocationsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_SearchInvocationsRequest.Merge(dst, src) +} +func (m *SearchInvocationsRequest) XXX_Size() int { + return xxx_messageInfo_SearchInvocationsRequest.Size(m) +} +func (m *SearchInvocationsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_SearchInvocationsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_SearchInvocationsRequest proto.InternalMessageInfo + +type isSearchInvocationsRequest_PageStart interface { + isSearchInvocationsRequest_PageStart() +} + +type SearchInvocationsRequest_PageToken struct { + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3,oneof"` +} +type SearchInvocationsRequest_Offset struct { + Offset int64 `protobuf:"varint,3,opt,name=offset,proto3,oneof"` +} + +func (*SearchInvocationsRequest_PageToken) isSearchInvocationsRequest_PageStart() {} +func (*SearchInvocationsRequest_Offset) isSearchInvocationsRequest_PageStart() {} + +func (m *SearchInvocationsRequest) GetPageStart() isSearchInvocationsRequest_PageStart { + if m != nil { + return m.PageStart + } + return nil +} + +func (m *SearchInvocationsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *SearchInvocationsRequest) GetPageToken() string { + if x, ok := m.GetPageStart().(*SearchInvocationsRequest_PageToken); ok { + return x.PageToken + } + return "" +} + +func (m *SearchInvocationsRequest) GetOffset() int64 { + if x, ok := m.GetPageStart().(*SearchInvocationsRequest_Offset); ok { + return x.Offset + } + return 0 +} + +func (m *SearchInvocationsRequest) GetQuery() string { + if m != nil { + return m.Query + } + return "" +} + +func (m *SearchInvocationsRequest) GetProjectId() string { + if m != nil { + return m.ProjectId + } + return "" +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*SearchInvocationsRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _SearchInvocationsRequest_OneofMarshaler, _SearchInvocationsRequest_OneofUnmarshaler, _SearchInvocationsRequest_OneofSizer, []interface{}{ + (*SearchInvocationsRequest_PageToken)(nil), + (*SearchInvocationsRequest_Offset)(nil), + } +} + +func _SearchInvocationsRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*SearchInvocationsRequest) + // page_start + switch x := m.PageStart.(type) { + case *SearchInvocationsRequest_PageToken: + b.EncodeVarint(2<<3 | proto.WireBytes) + b.EncodeStringBytes(x.PageToken) + case *SearchInvocationsRequest_Offset: + b.EncodeVarint(3<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Offset)) + case nil: + default: + return fmt.Errorf("SearchInvocationsRequest.PageStart has unexpected type %T", x) + } + return nil +} + +func _SearchInvocationsRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*SearchInvocationsRequest) + switch tag { + case 2: // page_start.page_token + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.PageStart = &SearchInvocationsRequest_PageToken{x} + return true, err + case 3: // page_start.offset + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.PageStart = &SearchInvocationsRequest_Offset{int64(x)} + return true, err + default: + return false, nil + } +} + +func _SearchInvocationsRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*SearchInvocationsRequest) + // page_start + switch x := m.PageStart.(type) { + case *SearchInvocationsRequest_PageToken: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.PageToken))) + n += len(x.PageToken) + case *SearchInvocationsRequest_Offset: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.Offset)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Response from calling SearchInvocations +type SearchInvocationsResponse struct { + // Invocations matching the search, possibly capped at request.page_size or a + // server limit. + Invocations []*Invocation `protobuf:"bytes,1,rep,name=invocations,proto3" json:"invocations,omitempty"` + // Token to retrieve the next page of results, or empty if there are no + // more results. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SearchInvocationsResponse) Reset() { *m = SearchInvocationsResponse{} } +func (m *SearchInvocationsResponse) String() string { return proto.CompactTextString(m) } +func (*SearchInvocationsResponse) ProtoMessage() {} +func (*SearchInvocationsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{2} +} +func (m *SearchInvocationsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SearchInvocationsResponse.Unmarshal(m, b) +} +func (m *SearchInvocationsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SearchInvocationsResponse.Marshal(b, m, deterministic) +} +func (dst *SearchInvocationsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_SearchInvocationsResponse.Merge(dst, src) +} +func (m *SearchInvocationsResponse) XXX_Size() int { + return xxx_messageInfo_SearchInvocationsResponse.Size(m) +} +func (m *SearchInvocationsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_SearchInvocationsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_SearchInvocationsResponse proto.InternalMessageInfo + +func (m *SearchInvocationsResponse) GetInvocations() []*Invocation { + if m != nil { + return m.Invocations + } + return nil +} + +func (m *SearchInvocationsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request passed into GetConfiguration +type GetConfigurationRequest struct { + // The name of the configuration to retrieve. It must match this format: + // invocations/${INVOCATION_ID}/configs/${CONFIGURATION_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetConfigurationRequest) Reset() { *m = GetConfigurationRequest{} } +func (m *GetConfigurationRequest) String() string { return proto.CompactTextString(m) } +func (*GetConfigurationRequest) ProtoMessage() {} +func (*GetConfigurationRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{3} +} +func (m *GetConfigurationRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetConfigurationRequest.Unmarshal(m, b) +} +func (m *GetConfigurationRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetConfigurationRequest.Marshal(b, m, deterministic) +} +func (dst *GetConfigurationRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetConfigurationRequest.Merge(dst, src) +} +func (m *GetConfigurationRequest) XXX_Size() int { + return xxx_messageInfo_GetConfigurationRequest.Size(m) +} +func (m *GetConfigurationRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetConfigurationRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetConfigurationRequest proto.InternalMessageInfo + +func (m *GetConfigurationRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request passed into ListConfigurations +type ListConfigurationsRequest struct { + // The invocation name of the configurations to retrieve. + // It must match this format: invocations/${INVOCATION_ID} + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The maximum number of items to return. + // Zero means all, but may be capped by the server. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Options for pagination. + // + // Types that are valid to be assigned to PageStart: + // *ListConfigurationsRequest_PageToken + // *ListConfigurationsRequest_Offset + PageStart isListConfigurationsRequest_PageStart `protobuf_oneof:"page_start"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListConfigurationsRequest) Reset() { *m = ListConfigurationsRequest{} } +func (m *ListConfigurationsRequest) String() string { return proto.CompactTextString(m) } +func (*ListConfigurationsRequest) ProtoMessage() {} +func (*ListConfigurationsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{4} +} +func (m *ListConfigurationsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListConfigurationsRequest.Unmarshal(m, b) +} +func (m *ListConfigurationsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListConfigurationsRequest.Marshal(b, m, deterministic) +} +func (dst *ListConfigurationsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListConfigurationsRequest.Merge(dst, src) +} +func (m *ListConfigurationsRequest) XXX_Size() int { + return xxx_messageInfo_ListConfigurationsRequest.Size(m) +} +func (m *ListConfigurationsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListConfigurationsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListConfigurationsRequest proto.InternalMessageInfo + +type isListConfigurationsRequest_PageStart interface { + isListConfigurationsRequest_PageStart() +} + +type ListConfigurationsRequest_PageToken struct { + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3,oneof"` +} +type ListConfigurationsRequest_Offset struct { + Offset int64 `protobuf:"varint,4,opt,name=offset,proto3,oneof"` +} + +func (*ListConfigurationsRequest_PageToken) isListConfigurationsRequest_PageStart() {} +func (*ListConfigurationsRequest_Offset) isListConfigurationsRequest_PageStart() {} + +func (m *ListConfigurationsRequest) GetPageStart() isListConfigurationsRequest_PageStart { + if m != nil { + return m.PageStart + } + return nil +} + +func (m *ListConfigurationsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListConfigurationsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListConfigurationsRequest) GetPageToken() string { + if x, ok := m.GetPageStart().(*ListConfigurationsRequest_PageToken); ok { + return x.PageToken + } + return "" +} + +func (m *ListConfigurationsRequest) GetOffset() int64 { + if x, ok := m.GetPageStart().(*ListConfigurationsRequest_Offset); ok { + return x.Offset + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*ListConfigurationsRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _ListConfigurationsRequest_OneofMarshaler, _ListConfigurationsRequest_OneofUnmarshaler, _ListConfigurationsRequest_OneofSizer, []interface{}{ + (*ListConfigurationsRequest_PageToken)(nil), + (*ListConfigurationsRequest_Offset)(nil), + } +} + +func _ListConfigurationsRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*ListConfigurationsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListConfigurationsRequest_PageToken: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.PageToken) + case *ListConfigurationsRequest_Offset: + b.EncodeVarint(4<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Offset)) + case nil: + default: + return fmt.Errorf("ListConfigurationsRequest.PageStart has unexpected type %T", x) + } + return nil +} + +func _ListConfigurationsRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*ListConfigurationsRequest) + switch tag { + case 3: // page_start.page_token + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.PageStart = &ListConfigurationsRequest_PageToken{x} + return true, err + case 4: // page_start.offset + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.PageStart = &ListConfigurationsRequest_Offset{int64(x)} + return true, err + default: + return false, nil + } +} + +func _ListConfigurationsRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*ListConfigurationsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListConfigurationsRequest_PageToken: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.PageToken))) + n += len(x.PageToken) + case *ListConfigurationsRequest_Offset: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.Offset)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Response from calling ListConfigurations +type ListConfigurationsResponse struct { + // Configurations matching the request invocation, + // possibly capped at request.page_size or a server limit. + Configurations []*Configuration `protobuf:"bytes,1,rep,name=configurations,proto3" json:"configurations,omitempty"` + // Token to retrieve the next page of results, or empty if there are no + // more results in the list. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListConfigurationsResponse) Reset() { *m = ListConfigurationsResponse{} } +func (m *ListConfigurationsResponse) String() string { return proto.CompactTextString(m) } +func (*ListConfigurationsResponse) ProtoMessage() {} +func (*ListConfigurationsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{5} +} +func (m *ListConfigurationsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListConfigurationsResponse.Unmarshal(m, b) +} +func (m *ListConfigurationsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListConfigurationsResponse.Marshal(b, m, deterministic) +} +func (dst *ListConfigurationsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListConfigurationsResponse.Merge(dst, src) +} +func (m *ListConfigurationsResponse) XXX_Size() int { + return xxx_messageInfo_ListConfigurationsResponse.Size(m) +} +func (m *ListConfigurationsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListConfigurationsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListConfigurationsResponse proto.InternalMessageInfo + +func (m *ListConfigurationsResponse) GetConfigurations() []*Configuration { + if m != nil { + return m.Configurations + } + return nil +} + +func (m *ListConfigurationsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request passed into GetTarget +type GetTargetRequest struct { + // The name of the target to retrieve. It must match this format: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetTargetRequest) Reset() { *m = GetTargetRequest{} } +func (m *GetTargetRequest) String() string { return proto.CompactTextString(m) } +func (*GetTargetRequest) ProtoMessage() {} +func (*GetTargetRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{6} +} +func (m *GetTargetRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetTargetRequest.Unmarshal(m, b) +} +func (m *GetTargetRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetTargetRequest.Marshal(b, m, deterministic) +} +func (dst *GetTargetRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetTargetRequest.Merge(dst, src) +} +func (m *GetTargetRequest) XXX_Size() int { + return xxx_messageInfo_GetTargetRequest.Size(m) +} +func (m *GetTargetRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetTargetRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetTargetRequest proto.InternalMessageInfo + +func (m *GetTargetRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request passed into ListTargets +type ListTargetsRequest struct { + // The invocation name of the targets to retrieve. It must match this format: + // invocations/${INVOCATION_ID} + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The maximum number of items to return. + // Zero means all, but may be capped by the server. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Options for pagination. + // + // Types that are valid to be assigned to PageStart: + // *ListTargetsRequest_PageToken + // *ListTargetsRequest_Offset + PageStart isListTargetsRequest_PageStart `protobuf_oneof:"page_start"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListTargetsRequest) Reset() { *m = ListTargetsRequest{} } +func (m *ListTargetsRequest) String() string { return proto.CompactTextString(m) } +func (*ListTargetsRequest) ProtoMessage() {} +func (*ListTargetsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{7} +} +func (m *ListTargetsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListTargetsRequest.Unmarshal(m, b) +} +func (m *ListTargetsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListTargetsRequest.Marshal(b, m, deterministic) +} +func (dst *ListTargetsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListTargetsRequest.Merge(dst, src) +} +func (m *ListTargetsRequest) XXX_Size() int { + return xxx_messageInfo_ListTargetsRequest.Size(m) +} +func (m *ListTargetsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListTargetsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListTargetsRequest proto.InternalMessageInfo + +type isListTargetsRequest_PageStart interface { + isListTargetsRequest_PageStart() +} + +type ListTargetsRequest_PageToken struct { + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3,oneof"` +} +type ListTargetsRequest_Offset struct { + Offset int64 `protobuf:"varint,4,opt,name=offset,proto3,oneof"` +} + +func (*ListTargetsRequest_PageToken) isListTargetsRequest_PageStart() {} +func (*ListTargetsRequest_Offset) isListTargetsRequest_PageStart() {} + +func (m *ListTargetsRequest) GetPageStart() isListTargetsRequest_PageStart { + if m != nil { + return m.PageStart + } + return nil +} + +func (m *ListTargetsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListTargetsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListTargetsRequest) GetPageToken() string { + if x, ok := m.GetPageStart().(*ListTargetsRequest_PageToken); ok { + return x.PageToken + } + return "" +} + +func (m *ListTargetsRequest) GetOffset() int64 { + if x, ok := m.GetPageStart().(*ListTargetsRequest_Offset); ok { + return x.Offset + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*ListTargetsRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _ListTargetsRequest_OneofMarshaler, _ListTargetsRequest_OneofUnmarshaler, _ListTargetsRequest_OneofSizer, []interface{}{ + (*ListTargetsRequest_PageToken)(nil), + (*ListTargetsRequest_Offset)(nil), + } +} + +func _ListTargetsRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*ListTargetsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListTargetsRequest_PageToken: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.PageToken) + case *ListTargetsRequest_Offset: + b.EncodeVarint(4<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Offset)) + case nil: + default: + return fmt.Errorf("ListTargetsRequest.PageStart has unexpected type %T", x) + } + return nil +} + +func _ListTargetsRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*ListTargetsRequest) + switch tag { + case 3: // page_start.page_token + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.PageStart = &ListTargetsRequest_PageToken{x} + return true, err + case 4: // page_start.offset + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.PageStart = &ListTargetsRequest_Offset{int64(x)} + return true, err + default: + return false, nil + } +} + +func _ListTargetsRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*ListTargetsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListTargetsRequest_PageToken: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.PageToken))) + n += len(x.PageToken) + case *ListTargetsRequest_Offset: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.Offset)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Response from calling ListTargetsResponse +type ListTargetsResponse struct { + // Targets matching the request invocation, + // possibly capped at request.page_size or a server limit. + Targets []*Target `protobuf:"bytes,1,rep,name=targets,proto3" json:"targets,omitempty"` + // Token to retrieve the next page of results, or empty if there are no + // more results in the list. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListTargetsResponse) Reset() { *m = ListTargetsResponse{} } +func (m *ListTargetsResponse) String() string { return proto.CompactTextString(m) } +func (*ListTargetsResponse) ProtoMessage() {} +func (*ListTargetsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{8} +} +func (m *ListTargetsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListTargetsResponse.Unmarshal(m, b) +} +func (m *ListTargetsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListTargetsResponse.Marshal(b, m, deterministic) +} +func (dst *ListTargetsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListTargetsResponse.Merge(dst, src) +} +func (m *ListTargetsResponse) XXX_Size() int { + return xxx_messageInfo_ListTargetsResponse.Size(m) +} +func (m *ListTargetsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListTargetsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListTargetsResponse proto.InternalMessageInfo + +func (m *ListTargetsResponse) GetTargets() []*Target { + if m != nil { + return m.Targets + } + return nil +} + +func (m *ListTargetsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request passed into GetConfiguredTarget +type GetConfiguredTargetRequest struct { + // The name of the configured target to retrieve. It must match this format: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID}/configuredTargets/${CONFIGURATION_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetConfiguredTargetRequest) Reset() { *m = GetConfiguredTargetRequest{} } +func (m *GetConfiguredTargetRequest) String() string { return proto.CompactTextString(m) } +func (*GetConfiguredTargetRequest) ProtoMessage() {} +func (*GetConfiguredTargetRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{9} +} +func (m *GetConfiguredTargetRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetConfiguredTargetRequest.Unmarshal(m, b) +} +func (m *GetConfiguredTargetRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetConfiguredTargetRequest.Marshal(b, m, deterministic) +} +func (dst *GetConfiguredTargetRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetConfiguredTargetRequest.Merge(dst, src) +} +func (m *GetConfiguredTargetRequest) XXX_Size() int { + return xxx_messageInfo_GetConfiguredTargetRequest.Size(m) +} +func (m *GetConfiguredTargetRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetConfiguredTargetRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetConfiguredTargetRequest proto.InternalMessageInfo + +func (m *GetConfiguredTargetRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request passed into ListConfiguredTargets +type ListConfiguredTargetsRequest struct { + // The invocation and target name of the configured targets to retrieve. + // It must match this format: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID} + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The maximum number of items to return. + // Zero means all, but may be capped by the server. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Options for pagination. + // + // Types that are valid to be assigned to PageStart: + // *ListConfiguredTargetsRequest_PageToken + // *ListConfiguredTargetsRequest_Offset + PageStart isListConfiguredTargetsRequest_PageStart `protobuf_oneof:"page_start"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListConfiguredTargetsRequest) Reset() { *m = ListConfiguredTargetsRequest{} } +func (m *ListConfiguredTargetsRequest) String() string { return proto.CompactTextString(m) } +func (*ListConfiguredTargetsRequest) ProtoMessage() {} +func (*ListConfiguredTargetsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{10} +} +func (m *ListConfiguredTargetsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListConfiguredTargetsRequest.Unmarshal(m, b) +} +func (m *ListConfiguredTargetsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListConfiguredTargetsRequest.Marshal(b, m, deterministic) +} +func (dst *ListConfiguredTargetsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListConfiguredTargetsRequest.Merge(dst, src) +} +func (m *ListConfiguredTargetsRequest) XXX_Size() int { + return xxx_messageInfo_ListConfiguredTargetsRequest.Size(m) +} +func (m *ListConfiguredTargetsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListConfiguredTargetsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListConfiguredTargetsRequest proto.InternalMessageInfo + +type isListConfiguredTargetsRequest_PageStart interface { + isListConfiguredTargetsRequest_PageStart() +} + +type ListConfiguredTargetsRequest_PageToken struct { + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3,oneof"` +} +type ListConfiguredTargetsRequest_Offset struct { + Offset int64 `protobuf:"varint,4,opt,name=offset,proto3,oneof"` +} + +func (*ListConfiguredTargetsRequest_PageToken) isListConfiguredTargetsRequest_PageStart() {} +func (*ListConfiguredTargetsRequest_Offset) isListConfiguredTargetsRequest_PageStart() {} + +func (m *ListConfiguredTargetsRequest) GetPageStart() isListConfiguredTargetsRequest_PageStart { + if m != nil { + return m.PageStart + } + return nil +} + +func (m *ListConfiguredTargetsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListConfiguredTargetsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListConfiguredTargetsRequest) GetPageToken() string { + if x, ok := m.GetPageStart().(*ListConfiguredTargetsRequest_PageToken); ok { + return x.PageToken + } + return "" +} + +func (m *ListConfiguredTargetsRequest) GetOffset() int64 { + if x, ok := m.GetPageStart().(*ListConfiguredTargetsRequest_Offset); ok { + return x.Offset + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*ListConfiguredTargetsRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _ListConfiguredTargetsRequest_OneofMarshaler, _ListConfiguredTargetsRequest_OneofUnmarshaler, _ListConfiguredTargetsRequest_OneofSizer, []interface{}{ + (*ListConfiguredTargetsRequest_PageToken)(nil), + (*ListConfiguredTargetsRequest_Offset)(nil), + } +} + +func _ListConfiguredTargetsRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*ListConfiguredTargetsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListConfiguredTargetsRequest_PageToken: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.PageToken) + case *ListConfiguredTargetsRequest_Offset: + b.EncodeVarint(4<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Offset)) + case nil: + default: + return fmt.Errorf("ListConfiguredTargetsRequest.PageStart has unexpected type %T", x) + } + return nil +} + +func _ListConfiguredTargetsRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*ListConfiguredTargetsRequest) + switch tag { + case 3: // page_start.page_token + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.PageStart = &ListConfiguredTargetsRequest_PageToken{x} + return true, err + case 4: // page_start.offset + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.PageStart = &ListConfiguredTargetsRequest_Offset{int64(x)} + return true, err + default: + return false, nil + } +} + +func _ListConfiguredTargetsRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*ListConfiguredTargetsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListConfiguredTargetsRequest_PageToken: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.PageToken))) + n += len(x.PageToken) + case *ListConfiguredTargetsRequest_Offset: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.Offset)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Response from calling ListConfiguredTargets +type ListConfiguredTargetsResponse struct { + // ConfiguredTargets matching the request, + // possibly capped at request.page_size or a server limit. + ConfiguredTargets []*ConfiguredTarget `protobuf:"bytes,1,rep,name=configured_targets,json=configuredTargets,proto3" json:"configured_targets,omitempty"` + // Token to retrieve the next page of results, or empty if there are no + // more results in the list. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListConfiguredTargetsResponse) Reset() { *m = ListConfiguredTargetsResponse{} } +func (m *ListConfiguredTargetsResponse) String() string { return proto.CompactTextString(m) } +func (*ListConfiguredTargetsResponse) ProtoMessage() {} +func (*ListConfiguredTargetsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{11} +} +func (m *ListConfiguredTargetsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListConfiguredTargetsResponse.Unmarshal(m, b) +} +func (m *ListConfiguredTargetsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListConfiguredTargetsResponse.Marshal(b, m, deterministic) +} +func (dst *ListConfiguredTargetsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListConfiguredTargetsResponse.Merge(dst, src) +} +func (m *ListConfiguredTargetsResponse) XXX_Size() int { + return xxx_messageInfo_ListConfiguredTargetsResponse.Size(m) +} +func (m *ListConfiguredTargetsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListConfiguredTargetsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListConfiguredTargetsResponse proto.InternalMessageInfo + +func (m *ListConfiguredTargetsResponse) GetConfiguredTargets() []*ConfiguredTarget { + if m != nil { + return m.ConfiguredTargets + } + return nil +} + +func (m *ListConfiguredTargetsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request passed into GetAction +type GetActionRequest struct { + // The name of the action to retrieve. It must match this format: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID}/configuredTargets/${CONFIGURATION_ID}/actions/${ACTION_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetActionRequest) Reset() { *m = GetActionRequest{} } +func (m *GetActionRequest) String() string { return proto.CompactTextString(m) } +func (*GetActionRequest) ProtoMessage() {} +func (*GetActionRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{12} +} +func (m *GetActionRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetActionRequest.Unmarshal(m, b) +} +func (m *GetActionRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetActionRequest.Marshal(b, m, deterministic) +} +func (dst *GetActionRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetActionRequest.Merge(dst, src) +} +func (m *GetActionRequest) XXX_Size() int { + return xxx_messageInfo_GetActionRequest.Size(m) +} +func (m *GetActionRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetActionRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetActionRequest proto.InternalMessageInfo + +func (m *GetActionRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request passed into ListActions +type ListActionsRequest struct { + // The invocation, target, and configuration name of the action to retrieve. + // It must match this format: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID}/configuredTargets/${CONFIGURATION_ID} + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The maximum number of items to return. + // Zero means all, but may be capped by the server. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Options for pagination. + // + // Types that are valid to be assigned to PageStart: + // *ListActionsRequest_PageToken + // *ListActionsRequest_Offset + PageStart isListActionsRequest_PageStart `protobuf_oneof:"page_start"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListActionsRequest) Reset() { *m = ListActionsRequest{} } +func (m *ListActionsRequest) String() string { return proto.CompactTextString(m) } +func (*ListActionsRequest) ProtoMessage() {} +func (*ListActionsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{13} +} +func (m *ListActionsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListActionsRequest.Unmarshal(m, b) +} +func (m *ListActionsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListActionsRequest.Marshal(b, m, deterministic) +} +func (dst *ListActionsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListActionsRequest.Merge(dst, src) +} +func (m *ListActionsRequest) XXX_Size() int { + return xxx_messageInfo_ListActionsRequest.Size(m) +} +func (m *ListActionsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListActionsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListActionsRequest proto.InternalMessageInfo + +type isListActionsRequest_PageStart interface { + isListActionsRequest_PageStart() +} + +type ListActionsRequest_PageToken struct { + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3,oneof"` +} +type ListActionsRequest_Offset struct { + Offset int64 `protobuf:"varint,4,opt,name=offset,proto3,oneof"` +} + +func (*ListActionsRequest_PageToken) isListActionsRequest_PageStart() {} +func (*ListActionsRequest_Offset) isListActionsRequest_PageStart() {} + +func (m *ListActionsRequest) GetPageStart() isListActionsRequest_PageStart { + if m != nil { + return m.PageStart + } + return nil +} + +func (m *ListActionsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListActionsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListActionsRequest) GetPageToken() string { + if x, ok := m.GetPageStart().(*ListActionsRequest_PageToken); ok { + return x.PageToken + } + return "" +} + +func (m *ListActionsRequest) GetOffset() int64 { + if x, ok := m.GetPageStart().(*ListActionsRequest_Offset); ok { + return x.Offset + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*ListActionsRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _ListActionsRequest_OneofMarshaler, _ListActionsRequest_OneofUnmarshaler, _ListActionsRequest_OneofSizer, []interface{}{ + (*ListActionsRequest_PageToken)(nil), + (*ListActionsRequest_Offset)(nil), + } +} + +func _ListActionsRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*ListActionsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListActionsRequest_PageToken: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.PageToken) + case *ListActionsRequest_Offset: + b.EncodeVarint(4<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Offset)) + case nil: + default: + return fmt.Errorf("ListActionsRequest.PageStart has unexpected type %T", x) + } + return nil +} + +func _ListActionsRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*ListActionsRequest) + switch tag { + case 3: // page_start.page_token + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.PageStart = &ListActionsRequest_PageToken{x} + return true, err + case 4: // page_start.offset + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.PageStart = &ListActionsRequest_Offset{int64(x)} + return true, err + default: + return false, nil + } +} + +func _ListActionsRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*ListActionsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListActionsRequest_PageToken: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.PageToken))) + n += len(x.PageToken) + case *ListActionsRequest_Offset: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.Offset)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Response from calling ListActions +type ListActionsResponse struct { + // Actions matching the request, + // possibly capped at request.page_size or a server limit. + Actions []*Action `protobuf:"bytes,1,rep,name=actions,proto3" json:"actions,omitempty"` + // Token to retrieve the next page of results, or empty if there are no + // more results in the list. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListActionsResponse) Reset() { *m = ListActionsResponse{} } +func (m *ListActionsResponse) String() string { return proto.CompactTextString(m) } +func (*ListActionsResponse) ProtoMessage() {} +func (*ListActionsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{14} +} +func (m *ListActionsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListActionsResponse.Unmarshal(m, b) +} +func (m *ListActionsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListActionsResponse.Marshal(b, m, deterministic) +} +func (dst *ListActionsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListActionsResponse.Merge(dst, src) +} +func (m *ListActionsResponse) XXX_Size() int { + return xxx_messageInfo_ListActionsResponse.Size(m) +} +func (m *ListActionsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListActionsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListActionsResponse proto.InternalMessageInfo + +func (m *ListActionsResponse) GetActions() []*Action { + if m != nil { + return m.Actions + } + return nil +} + +func (m *ListActionsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +// Request passed into GetFileSet +type GetFileSetRequest struct { + // The name of the file set to retrieve. It must match this format: + // invocations/${INVOCATION_ID}/fileSets/${FILE_SET_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetFileSetRequest) Reset() { *m = GetFileSetRequest{} } +func (m *GetFileSetRequest) String() string { return proto.CompactTextString(m) } +func (*GetFileSetRequest) ProtoMessage() {} +func (*GetFileSetRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{15} +} +func (m *GetFileSetRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetFileSetRequest.Unmarshal(m, b) +} +func (m *GetFileSetRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetFileSetRequest.Marshal(b, m, deterministic) +} +func (dst *GetFileSetRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetFileSetRequest.Merge(dst, src) +} +func (m *GetFileSetRequest) XXX_Size() int { + return xxx_messageInfo_GetFileSetRequest.Size(m) +} +func (m *GetFileSetRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetFileSetRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetFileSetRequest proto.InternalMessageInfo + +func (m *GetFileSetRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Request passed into ListFileSets +type ListFileSetsRequest struct { + // The invocation name of the file sets to retrieve. + // It must match this format: invocations/${INVOCATION_ID} + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + // The maximum number of items to return. + // Zero means all, but may be capped by the server. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // Options for pagination. + // + // Types that are valid to be assigned to PageStart: + // *ListFileSetsRequest_PageToken + // *ListFileSetsRequest_Offset + PageStart isListFileSetsRequest_PageStart `protobuf_oneof:"page_start"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListFileSetsRequest) Reset() { *m = ListFileSetsRequest{} } +func (m *ListFileSetsRequest) String() string { return proto.CompactTextString(m) } +func (*ListFileSetsRequest) ProtoMessage() {} +func (*ListFileSetsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{16} +} +func (m *ListFileSetsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListFileSetsRequest.Unmarshal(m, b) +} +func (m *ListFileSetsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListFileSetsRequest.Marshal(b, m, deterministic) +} +func (dst *ListFileSetsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListFileSetsRequest.Merge(dst, src) +} +func (m *ListFileSetsRequest) XXX_Size() int { + return xxx_messageInfo_ListFileSetsRequest.Size(m) +} +func (m *ListFileSetsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListFileSetsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListFileSetsRequest proto.InternalMessageInfo + +type isListFileSetsRequest_PageStart interface { + isListFileSetsRequest_PageStart() +} + +type ListFileSetsRequest_PageToken struct { + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3,oneof"` +} +type ListFileSetsRequest_Offset struct { + Offset int64 `protobuf:"varint,4,opt,name=offset,proto3,oneof"` +} + +func (*ListFileSetsRequest_PageToken) isListFileSetsRequest_PageStart() {} +func (*ListFileSetsRequest_Offset) isListFileSetsRequest_PageStart() {} + +func (m *ListFileSetsRequest) GetPageStart() isListFileSetsRequest_PageStart { + if m != nil { + return m.PageStart + } + return nil +} + +func (m *ListFileSetsRequest) GetParent() string { + if m != nil { + return m.Parent + } + return "" +} + +func (m *ListFileSetsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListFileSetsRequest) GetPageToken() string { + if x, ok := m.GetPageStart().(*ListFileSetsRequest_PageToken); ok { + return x.PageToken + } + return "" +} + +func (m *ListFileSetsRequest) GetOffset() int64 { + if x, ok := m.GetPageStart().(*ListFileSetsRequest_Offset); ok { + return x.Offset + } + return 0 +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*ListFileSetsRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _ListFileSetsRequest_OneofMarshaler, _ListFileSetsRequest_OneofUnmarshaler, _ListFileSetsRequest_OneofSizer, []interface{}{ + (*ListFileSetsRequest_PageToken)(nil), + (*ListFileSetsRequest_Offset)(nil), + } +} + +func _ListFileSetsRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*ListFileSetsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListFileSetsRequest_PageToken: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.PageToken) + case *ListFileSetsRequest_Offset: + b.EncodeVarint(4<<3 | proto.WireVarint) + b.EncodeVarint(uint64(x.Offset)) + case nil: + default: + return fmt.Errorf("ListFileSetsRequest.PageStart has unexpected type %T", x) + } + return nil +} + +func _ListFileSetsRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*ListFileSetsRequest) + switch tag { + case 3: // page_start.page_token + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.PageStart = &ListFileSetsRequest_PageToken{x} + return true, err + case 4: // page_start.offset + if wire != proto.WireVarint { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeVarint() + m.PageStart = &ListFileSetsRequest_Offset{int64(x)} + return true, err + default: + return false, nil + } +} + +func _ListFileSetsRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*ListFileSetsRequest) + // page_start + switch x := m.PageStart.(type) { + case *ListFileSetsRequest_PageToken: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.PageToken))) + n += len(x.PageToken) + case *ListFileSetsRequest_Offset: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(x.Offset)) + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// Response from calling ListFileSets +type ListFileSetsResponse struct { + // File sets matching the request, + // possibly capped at request.page_size or a server limit. + FileSets []*FileSet `protobuf:"bytes,1,rep,name=file_sets,json=fileSets,proto3" json:"file_sets,omitempty"` + // Token to retrieve the next page of results, or empty if there are no + // more results in the list. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListFileSetsResponse) Reset() { *m = ListFileSetsResponse{} } +func (m *ListFileSetsResponse) String() string { return proto.CompactTextString(m) } +func (*ListFileSetsResponse) ProtoMessage() {} +func (*ListFileSetsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_download_7c774ab49c606fb1, []int{17} +} +func (m *ListFileSetsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListFileSetsResponse.Unmarshal(m, b) +} +func (m *ListFileSetsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListFileSetsResponse.Marshal(b, m, deterministic) +} +func (dst *ListFileSetsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListFileSetsResponse.Merge(dst, src) +} +func (m *ListFileSetsResponse) XXX_Size() int { + return xxx_messageInfo_ListFileSetsResponse.Size(m) +} +func (m *ListFileSetsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListFileSetsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListFileSetsResponse proto.InternalMessageInfo + +func (m *ListFileSetsResponse) GetFileSets() []*FileSet { + if m != nil { + return m.FileSets + } + return nil +} + +func (m *ListFileSetsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + +func init() { + proto.RegisterType((*GetInvocationRequest)(nil), "google.devtools.resultstore.v2.GetInvocationRequest") + proto.RegisterType((*SearchInvocationsRequest)(nil), "google.devtools.resultstore.v2.SearchInvocationsRequest") + proto.RegisterType((*SearchInvocationsResponse)(nil), "google.devtools.resultstore.v2.SearchInvocationsResponse") + proto.RegisterType((*GetConfigurationRequest)(nil), "google.devtools.resultstore.v2.GetConfigurationRequest") + proto.RegisterType((*ListConfigurationsRequest)(nil), "google.devtools.resultstore.v2.ListConfigurationsRequest") + proto.RegisterType((*ListConfigurationsResponse)(nil), "google.devtools.resultstore.v2.ListConfigurationsResponse") + proto.RegisterType((*GetTargetRequest)(nil), "google.devtools.resultstore.v2.GetTargetRequest") + proto.RegisterType((*ListTargetsRequest)(nil), "google.devtools.resultstore.v2.ListTargetsRequest") + proto.RegisterType((*ListTargetsResponse)(nil), "google.devtools.resultstore.v2.ListTargetsResponse") + proto.RegisterType((*GetConfiguredTargetRequest)(nil), "google.devtools.resultstore.v2.GetConfiguredTargetRequest") + proto.RegisterType((*ListConfiguredTargetsRequest)(nil), "google.devtools.resultstore.v2.ListConfiguredTargetsRequest") + proto.RegisterType((*ListConfiguredTargetsResponse)(nil), "google.devtools.resultstore.v2.ListConfiguredTargetsResponse") + proto.RegisterType((*GetActionRequest)(nil), "google.devtools.resultstore.v2.GetActionRequest") + proto.RegisterType((*ListActionsRequest)(nil), "google.devtools.resultstore.v2.ListActionsRequest") + proto.RegisterType((*ListActionsResponse)(nil), "google.devtools.resultstore.v2.ListActionsResponse") + proto.RegisterType((*GetFileSetRequest)(nil), "google.devtools.resultstore.v2.GetFileSetRequest") + proto.RegisterType((*ListFileSetsRequest)(nil), "google.devtools.resultstore.v2.ListFileSetsRequest") + proto.RegisterType((*ListFileSetsResponse)(nil), "google.devtools.resultstore.v2.ListFileSetsResponse") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// ResultStoreDownloadClient is the client API for ResultStoreDownload service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type ResultStoreDownloadClient interface { + // Retrieves the invocation with the given name. + // + // An error will be reported in the following cases: + // - If the invocation is not found. + // - If the given invocation name is badly formatted. + GetInvocation(ctx context.Context, in *GetInvocationRequest, opts ...grpc.CallOption) (*Invocation, error) + // Searches for invocations matching the given query parameters. + // + // + // An error will be reported in the following cases: + // - If a query string is not provided + SearchInvocations(ctx context.Context, in *SearchInvocationsRequest, opts ...grpc.CallOption) (*SearchInvocationsResponse, error) + // Retrieves the configuration with the given name. + // + // An error will be reported in the following cases: + // - If the configuration or its parent invocation is not found. + // - If the given configuration name is badly formatted. + GetConfiguration(ctx context.Context, in *GetConfigurationRequest, opts ...grpc.CallOption) (*Configuration, error) + // Retrieves all configurations for a parent invocation. + // This might be limited by user or server, + // in which case a continuation token is provided. + // + // An error will be reported in the following cases: + // - If the parent invocation is not found. + // - If the given parent invocation name is badly formatted. + ListConfigurations(ctx context.Context, in *ListConfigurationsRequest, opts ...grpc.CallOption) (*ListConfigurationsResponse, error) + // Retrieves the target with the given name. + // + // An error will be reported in the following cases: + // - If the target or its parent invocation is not found. + // - If the given target name is badly formatted. + GetTarget(ctx context.Context, in *GetTargetRequest, opts ...grpc.CallOption) (*Target, error) + // Retrieves all targets for a parent invocation. This might be limited by + // user or server, in which case a continuation token is provided. + // + // An error will be reported in the following cases: + // - If the parent is not found. + // - If the given parent name is badly formatted. + ListTargets(ctx context.Context, in *ListTargetsRequest, opts ...grpc.CallOption) (*ListTargetsResponse, error) + // Retrieves the configured target with the given name. + // + // An error will be reported in the following cases: + // - If the configured target is not found. + // - If the given name is badly formatted. + GetConfiguredTarget(ctx context.Context, in *GetConfiguredTargetRequest, opts ...grpc.CallOption) (*ConfiguredTarget, error) + // Retrieves all configured targets for a parent invocation/target. + // This might be limited by user or server, in which case a continuation + // token is provided. Supports '-' for targetId meaning all targets. + // + // An error will be reported in the following cases: + // - If the parent is not found. + // - If the given parent name is badly formatted. + ListConfiguredTargets(ctx context.Context, in *ListConfiguredTargetsRequest, opts ...grpc.CallOption) (*ListConfiguredTargetsResponse, error) + // Retrieves the action with the given name. + // + // An error will be reported in the following cases: + // - If the action is not found. + // - If the given name is badly formatted. + GetAction(ctx context.Context, in *GetActionRequest, opts ...grpc.CallOption) (*Action, error) + // Retrieves all actions for a parent invocation/target/configuration. + // This might be limited by user or server, in which case a continuation + // token is provided. Supports '-' for configurationId to mean all + // actions for all configurations for a target, or '-' for targetId and + // configurationId to mean all actions for all configurations and all targets. + // Does not support targetId '-' with a specified configuration. + // + // An error will be reported in the following cases: + // - If the parent is not found. + // - If the given parent name is badly formatted. + ListActions(ctx context.Context, in *ListActionsRequest, opts ...grpc.CallOption) (*ListActionsResponse, error) + // Retrieves the file set with the given name. + // + // An error will be reported in the following cases: + // - If the file set or its parent invocation is not found. + // - If the given file set name is badly formatted. + GetFileSet(ctx context.Context, in *GetFileSetRequest, opts ...grpc.CallOption) (*FileSet, error) + // Retrieves all file sets for a parent invocation. + // This might be limited by user or server, + // in which case a continuation token is provided. + // + // An error will be reported in the following cases: + // - If the parent invocation is not found. + // - If the given parent invocation name is badly formatted. + ListFileSets(ctx context.Context, in *ListFileSetsRequest, opts ...grpc.CallOption) (*ListFileSetsResponse, error) +} + +type resultStoreDownloadClient struct { + cc *grpc.ClientConn +} + +func NewResultStoreDownloadClient(cc *grpc.ClientConn) ResultStoreDownloadClient { + return &resultStoreDownloadClient{cc} +} + +func (c *resultStoreDownloadClient) GetInvocation(ctx context.Context, in *GetInvocationRequest, opts ...grpc.CallOption) (*Invocation, error) { + out := new(Invocation) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/GetInvocation", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) SearchInvocations(ctx context.Context, in *SearchInvocationsRequest, opts ...grpc.CallOption) (*SearchInvocationsResponse, error) { + out := new(SearchInvocationsResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/SearchInvocations", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) GetConfiguration(ctx context.Context, in *GetConfigurationRequest, opts ...grpc.CallOption) (*Configuration, error) { + out := new(Configuration) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/GetConfiguration", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) ListConfigurations(ctx context.Context, in *ListConfigurationsRequest, opts ...grpc.CallOption) (*ListConfigurationsResponse, error) { + out := new(ListConfigurationsResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/ListConfigurations", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) GetTarget(ctx context.Context, in *GetTargetRequest, opts ...grpc.CallOption) (*Target, error) { + out := new(Target) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/GetTarget", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) ListTargets(ctx context.Context, in *ListTargetsRequest, opts ...grpc.CallOption) (*ListTargetsResponse, error) { + out := new(ListTargetsResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/ListTargets", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) GetConfiguredTarget(ctx context.Context, in *GetConfiguredTargetRequest, opts ...grpc.CallOption) (*ConfiguredTarget, error) { + out := new(ConfiguredTarget) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/GetConfiguredTarget", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) ListConfiguredTargets(ctx context.Context, in *ListConfiguredTargetsRequest, opts ...grpc.CallOption) (*ListConfiguredTargetsResponse, error) { + out := new(ListConfiguredTargetsResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/ListConfiguredTargets", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) GetAction(ctx context.Context, in *GetActionRequest, opts ...grpc.CallOption) (*Action, error) { + out := new(Action) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/GetAction", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) ListActions(ctx context.Context, in *ListActionsRequest, opts ...grpc.CallOption) (*ListActionsResponse, error) { + out := new(ListActionsResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/ListActions", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) GetFileSet(ctx context.Context, in *GetFileSetRequest, opts ...grpc.CallOption) (*FileSet, error) { + out := new(FileSet) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/GetFileSet", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *resultStoreDownloadClient) ListFileSets(ctx context.Context, in *ListFileSetsRequest, opts ...grpc.CallOption) (*ListFileSetsResponse, error) { + out := new(ListFileSetsResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreDownload/ListFileSets", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// ResultStoreDownloadServer is the server API for ResultStoreDownload service. +type ResultStoreDownloadServer interface { + // Retrieves the invocation with the given name. + // + // An error will be reported in the following cases: + // - If the invocation is not found. + // - If the given invocation name is badly formatted. + GetInvocation(context.Context, *GetInvocationRequest) (*Invocation, error) + // Searches for invocations matching the given query parameters. + // + // + // An error will be reported in the following cases: + // - If a query string is not provided + SearchInvocations(context.Context, *SearchInvocationsRequest) (*SearchInvocationsResponse, error) + // Retrieves the configuration with the given name. + // + // An error will be reported in the following cases: + // - If the configuration or its parent invocation is not found. + // - If the given configuration name is badly formatted. + GetConfiguration(context.Context, *GetConfigurationRequest) (*Configuration, error) + // Retrieves all configurations for a parent invocation. + // This might be limited by user or server, + // in which case a continuation token is provided. + // + // An error will be reported in the following cases: + // - If the parent invocation is not found. + // - If the given parent invocation name is badly formatted. + ListConfigurations(context.Context, *ListConfigurationsRequest) (*ListConfigurationsResponse, error) + // Retrieves the target with the given name. + // + // An error will be reported in the following cases: + // - If the target or its parent invocation is not found. + // - If the given target name is badly formatted. + GetTarget(context.Context, *GetTargetRequest) (*Target, error) + // Retrieves all targets for a parent invocation. This might be limited by + // user or server, in which case a continuation token is provided. + // + // An error will be reported in the following cases: + // - If the parent is not found. + // - If the given parent name is badly formatted. + ListTargets(context.Context, *ListTargetsRequest) (*ListTargetsResponse, error) + // Retrieves the configured target with the given name. + // + // An error will be reported in the following cases: + // - If the configured target is not found. + // - If the given name is badly formatted. + GetConfiguredTarget(context.Context, *GetConfiguredTargetRequest) (*ConfiguredTarget, error) + // Retrieves all configured targets for a parent invocation/target. + // This might be limited by user or server, in which case a continuation + // token is provided. Supports '-' for targetId meaning all targets. + // + // An error will be reported in the following cases: + // - If the parent is not found. + // - If the given parent name is badly formatted. + ListConfiguredTargets(context.Context, *ListConfiguredTargetsRequest) (*ListConfiguredTargetsResponse, error) + // Retrieves the action with the given name. + // + // An error will be reported in the following cases: + // - If the action is not found. + // - If the given name is badly formatted. + GetAction(context.Context, *GetActionRequest) (*Action, error) + // Retrieves all actions for a parent invocation/target/configuration. + // This might be limited by user or server, in which case a continuation + // token is provided. Supports '-' for configurationId to mean all + // actions for all configurations for a target, or '-' for targetId and + // configurationId to mean all actions for all configurations and all targets. + // Does not support targetId '-' with a specified configuration. + // + // An error will be reported in the following cases: + // - If the parent is not found. + // - If the given parent name is badly formatted. + ListActions(context.Context, *ListActionsRequest) (*ListActionsResponse, error) + // Retrieves the file set with the given name. + // + // An error will be reported in the following cases: + // - If the file set or its parent invocation is not found. + // - If the given file set name is badly formatted. + GetFileSet(context.Context, *GetFileSetRequest) (*FileSet, error) + // Retrieves all file sets for a parent invocation. + // This might be limited by user or server, + // in which case a continuation token is provided. + // + // An error will be reported in the following cases: + // - If the parent invocation is not found. + // - If the given parent invocation name is badly formatted. + ListFileSets(context.Context, *ListFileSetsRequest) (*ListFileSetsResponse, error) +} + +func RegisterResultStoreDownloadServer(s *grpc.Server, srv ResultStoreDownloadServer) { + s.RegisterService(&_ResultStoreDownload_serviceDesc, srv) +} + +func _ResultStoreDownload_GetInvocation_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetInvocationRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).GetInvocation(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/GetInvocation", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).GetInvocation(ctx, req.(*GetInvocationRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_SearchInvocations_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(SearchInvocationsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).SearchInvocations(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/SearchInvocations", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).SearchInvocations(ctx, req.(*SearchInvocationsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_GetConfiguration_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetConfigurationRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).GetConfiguration(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/GetConfiguration", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).GetConfiguration(ctx, req.(*GetConfigurationRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_ListConfigurations_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListConfigurationsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).ListConfigurations(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/ListConfigurations", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).ListConfigurations(ctx, req.(*ListConfigurationsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_GetTarget_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetTargetRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).GetTarget(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/GetTarget", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).GetTarget(ctx, req.(*GetTargetRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_ListTargets_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListTargetsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).ListTargets(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/ListTargets", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).ListTargets(ctx, req.(*ListTargetsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_GetConfiguredTarget_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetConfiguredTargetRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).GetConfiguredTarget(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/GetConfiguredTarget", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).GetConfiguredTarget(ctx, req.(*GetConfiguredTargetRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_ListConfiguredTargets_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListConfiguredTargetsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).ListConfiguredTargets(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/ListConfiguredTargets", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).ListConfiguredTargets(ctx, req.(*ListConfiguredTargetsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_GetAction_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetActionRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).GetAction(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/GetAction", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).GetAction(ctx, req.(*GetActionRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_ListActions_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListActionsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).ListActions(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/ListActions", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).ListActions(ctx, req.(*ListActionsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_GetFileSet_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetFileSetRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).GetFileSet(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/GetFileSet", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).GetFileSet(ctx, req.(*GetFileSetRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _ResultStoreDownload_ListFileSets_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListFileSetsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreDownloadServer).ListFileSets(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreDownload/ListFileSets", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreDownloadServer).ListFileSets(ctx, req.(*ListFileSetsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _ResultStoreDownload_serviceDesc = grpc.ServiceDesc{ + ServiceName: "google.devtools.resultstore.v2.ResultStoreDownload", + HandlerType: (*ResultStoreDownloadServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "GetInvocation", + Handler: _ResultStoreDownload_GetInvocation_Handler, + }, + { + MethodName: "SearchInvocations", + Handler: _ResultStoreDownload_SearchInvocations_Handler, + }, + { + MethodName: "GetConfiguration", + Handler: _ResultStoreDownload_GetConfiguration_Handler, + }, + { + MethodName: "ListConfigurations", + Handler: _ResultStoreDownload_ListConfigurations_Handler, + }, + { + MethodName: "GetTarget", + Handler: _ResultStoreDownload_GetTarget_Handler, + }, + { + MethodName: "ListTargets", + Handler: _ResultStoreDownload_ListTargets_Handler, + }, + { + MethodName: "GetConfiguredTarget", + Handler: _ResultStoreDownload_GetConfiguredTarget_Handler, + }, + { + MethodName: "ListConfiguredTargets", + Handler: _ResultStoreDownload_ListConfiguredTargets_Handler, + }, + { + MethodName: "GetAction", + Handler: _ResultStoreDownload_GetAction_Handler, + }, + { + MethodName: "ListActions", + Handler: _ResultStoreDownload_ListActions_Handler, + }, + { + MethodName: "GetFileSet", + Handler: _ResultStoreDownload_GetFileSet_Handler, + }, + { + MethodName: "ListFileSets", + Handler: _ResultStoreDownload_ListFileSets_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "google/devtools/resultstore/v2/resultstore_download.proto", +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/resultstore_download.proto", fileDescriptor_resultstore_download_7c774ab49c606fb1) +} + +var fileDescriptor_resultstore_download_7c774ab49c606fb1 = []byte{ + // 1043 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xcc, 0x98, 0xcf, 0x6f, 0xdc, 0x44, + 0x14, 0xc7, 0x3b, 0xf9, 0x45, 0xf7, 0xa5, 0x85, 0xe6, 0x25, 0x14, 0xd7, 0xb4, 0x25, 0x32, 0x90, + 0x46, 0xa9, 0xb2, 0x1b, 0xb6, 0x55, 0x69, 0x03, 0x45, 0x25, 0x54, 0x24, 0x91, 0x7a, 0xa8, 0x36, + 0xe5, 0xc2, 0x65, 0x65, 0x76, 0x67, 0x8d, 0x61, 0xe3, 0xd9, 0x78, 0x26, 0x01, 0x8a, 0x10, 0x12, + 0xe2, 0x1f, 0x28, 0x11, 0x07, 0x04, 0x27, 0x0e, 0x55, 0x6e, 0x9c, 0x10, 0x57, 0x24, 0xc4, 0x5f, + 0x00, 0x7f, 0x02, 0x7f, 0x08, 0xf2, 0x78, 0x66, 0xe3, 0x71, 0xbc, 0xeb, 0x71, 0x4e, 0xb9, 0xc5, + 0xde, 0xf7, 0x9d, 0x79, 0xef, 0x33, 0x5f, 0xfb, 0xbd, 0x18, 0xee, 0x05, 0x8c, 0x05, 0x7d, 0xda, + 0xe8, 0xd2, 0x03, 0xc1, 0x58, 0x9f, 0x37, 0x62, 0xca, 0xf7, 0xfb, 0x82, 0x0b, 0x16, 0xd3, 0xc6, + 0x41, 0x33, 0x7b, 0xd9, 0xee, 0xb2, 0x2f, 0xa2, 0x3e, 0xf3, 0xbb, 0xf5, 0x41, 0xcc, 0x04, 0xc3, + 0xeb, 0xa9, 0xb4, 0xae, 0xa5, 0xf5, 0x4c, 0x6c, 0xfd, 0xa0, 0xe9, 0x5e, 0x55, 0x4b, 0xfb, 0x83, + 0xb0, 0xe1, 0x47, 0x11, 0x13, 0xbe, 0x08, 0x59, 0xc4, 0x53, 0xb5, 0x7b, 0xb3, 0x64, 0x63, 0xbf, + 0x93, 0x44, 0x5b, 0x06, 0x77, 0xd8, 0xee, 0xee, 0x30, 0xb8, 0x59, 0x1a, 0x1c, 0xf5, 0xc2, 0x60, + 0x3f, 0xf6, 0x33, 0x1b, 0xdc, 0xb1, 0xd4, 0xd0, 0x6e, 0x5b, 0xf8, 0x71, 0x40, 0x85, 0xd2, 0xad, + 0x96, 0xe8, 0x7a, 0x61, 0x9f, 0xb6, 0xf9, 0x30, 0xbc, 0x51, 0x12, 0x1e, 0x46, 0x07, 0xac, 0xe3, + 0x57, 0x28, 0x3c, 0x9b, 0x8c, 0xb7, 0x02, 0x0b, 0x9b, 0x54, 0x6c, 0x0f, 0xd7, 0x68, 0xd1, 0xbd, + 0x7d, 0xca, 0x05, 0x22, 0x4c, 0x45, 0xfe, 0x2e, 0x75, 0xc8, 0x22, 0x59, 0xae, 0xb5, 0xe4, 0xdf, + 0xde, 0xef, 0x04, 0x9c, 0x1d, 0xea, 0xc7, 0x9d, 0x4f, 0x8f, 0xe3, 0xb9, 0x16, 0xbc, 0x0a, 0xb5, + 0x81, 0x1f, 0xd0, 0x36, 0x0f, 0x9f, 0xa6, 0xaa, 0xe9, 0xd6, 0xf9, 0xe4, 0xc6, 0x4e, 0xf8, 0x94, + 0xe2, 0x6b, 0x00, 0xf2, 0x47, 0xc1, 0x3e, 0xa7, 0x91, 0x33, 0x91, 0xac, 0xb9, 0x75, 0xae, 0x25, + 0x05, 0x4f, 0x92, 0x5b, 0xe8, 0xc0, 0x0c, 0xeb, 0xf5, 0x38, 0x15, 0xce, 0xe4, 0x22, 0x59, 0x9e, + 0xdc, 0x3a, 0xd7, 0x52, 0xd7, 0xb8, 0x00, 0xd3, 0x7b, 0xfb, 0x34, 0xfe, 0xca, 0x99, 0x92, 0x99, + 0xa4, 0x17, 0x78, 0x0d, 0x60, 0x10, 0xb3, 0xcf, 0x68, 0x47, 0xb4, 0xc3, 0xae, 0x33, 0x2d, 0x7f, + 0xaa, 0xa9, 0x3b, 0xdb, 0xdd, 0x8d, 0x0b, 0x6a, 0x3f, 0x2e, 0xfc, 0x58, 0x78, 0xcf, 0x08, 0x5c, + 0x29, 0xc8, 0x9b, 0x0f, 0x58, 0xc4, 0x29, 0x3e, 0x82, 0xd9, 0x63, 0x84, 0xdc, 0x21, 0x8b, 0x93, + 0xcb, 0xb3, 0xcd, 0x95, 0xfa, 0x78, 0xa3, 0xd6, 0x33, 0xc4, 0xb2, 0x72, 0x5c, 0x82, 0x97, 0x22, + 0xfa, 0xa5, 0x68, 0xe7, 0xcb, 0x6d, 0x5d, 0x4c, 0x6e, 0x3f, 0xd6, 0x05, 0x7b, 0xab, 0xf0, 0xca, + 0x26, 0x15, 0x1f, 0x64, 0x6d, 0x35, 0x0e, 0xfd, 0x4f, 0x04, 0xae, 0x3c, 0x0a, 0xb9, 0x29, 0x18, + 0xb2, 0xbf, 0x0c, 0x33, 0x03, 0x3f, 0xa6, 0x91, 0x50, 0x1a, 0x75, 0x65, 0x9e, 0xc9, 0xc4, 0xd8, + 0x33, 0x99, 0x1c, 0x77, 0x26, 0x53, 0xe6, 0x99, 0xe4, 0xf0, 0xfe, 0x4c, 0xc0, 0x2d, 0xca, 0x4d, + 0xf1, 0xfd, 0x08, 0x5e, 0x34, 0x9e, 0x1e, 0x8d, 0x78, 0xb5, 0x0c, 0xb1, 0x09, 0x27, 0xb7, 0x88, + 0x35, 0xe8, 0x25, 0xb8, 0xb4, 0x49, 0xc5, 0x13, 0xe9, 0xf9, 0x71, 0x84, 0x7f, 0x20, 0x80, 0x49, + 0x15, 0x69, 0xe4, 0x19, 0x41, 0xfb, 0x2d, 0xcc, 0x1b, 0x39, 0x29, 0xa4, 0x0f, 0xe0, 0x85, 0xf4, + 0x21, 0xd6, 0x2c, 0x97, 0xca, 0x58, 0xaa, 0xfa, 0xb5, 0xcc, 0x9a, 0xde, 0x1a, 0xb8, 0x19, 0x9b, + 0xd2, 0x6e, 0x39, 0xc7, 0x5f, 0x08, 0x5c, 0xcd, 0xba, 0x41, 0x6b, 0xce, 0x08, 0xd1, 0x23, 0x02, + 0xd7, 0x46, 0xa4, 0xa7, 0xe0, 0xb6, 0x01, 0x4f, 0xbc, 0xb9, 0x35, 0xe7, 0x35, 0x5b, 0xcf, 0x0e, + 0x49, 0xcd, 0x75, 0xf2, 0x1b, 0x55, 0x74, 0xee, 0xfb, 0x9d, 0xb2, 0x77, 0x83, 0x76, 0x6e, 0x1a, + 0x79, 0xb6, 0x9c, 0x3b, 0xcc, 0xe9, 0xd8, 0xb9, 0x69, 0x93, 0xb6, 0x76, 0xae, 0xaa, 0x5f, 0xcb, + 0xac, 0xe9, 0xdd, 0x80, 0xb9, 0x4d, 0x2a, 0x3e, 0x0c, 0xfb, 0x74, 0x67, 0xbc, 0x61, 0x0f, 0x49, + 0x9a, 0xaa, 0x0a, 0x3d, 0x23, 0xfc, 0xbe, 0x27, 0xb0, 0x60, 0x66, 0xa5, 0x08, 0x3e, 0x84, 0x9a, + 0x1e, 0x10, 0x34, 0xc3, 0x1b, 0x65, 0x0c, 0x35, 0x85, 0xf3, 0x3d, 0xb5, 0x9a, 0x2d, 0xc5, 0xe6, + 0xdf, 0x97, 0x60, 0xbe, 0x25, 0x17, 0xdb, 0x49, 0x16, 0x7b, 0xa8, 0xa6, 0x39, 0x3c, 0x24, 0x70, + 0xd1, 0x98, 0x1b, 0xf0, 0x76, 0x59, 0x12, 0x45, 0x63, 0x86, 0x5b, 0xa1, 0xcf, 0x7a, 0x8b, 0xdf, + 0xfd, 0xf3, 0xdf, 0xe1, 0x84, 0x8b, 0x4e, 0x32, 0xc4, 0x7c, 0x9d, 0x9c, 0xdd, 0xfd, 0x4c, 0xe7, + 0x6d, 0xac, 0x7c, 0x83, 0x47, 0x04, 0xe6, 0x4e, 0x34, 0x7a, 0xbc, 0x5b, 0xb6, 0xc7, 0xa8, 0x99, + 0xc6, 0xbd, 0x77, 0x0a, 0x65, 0x7a, 0x4c, 0xde, 0x75, 0x99, 0xac, 0x83, 0x97, 0xcd, 0x11, 0x8d, + 0xaf, 0x73, 0x29, 0xc1, 0xe7, 0x44, 0x3e, 0xdd, 0x46, 0x8f, 0xc3, 0xb7, 0x2d, 0x18, 0x16, 0x8d, + 0x0c, 0x6e, 0xb5, 0x5e, 0xea, 0xad, 0xc8, 0xe4, 0xde, 0x40, 0x6f, 0x04, 0x49, 0x35, 0xba, 0x4a, + 0xa6, 0x7f, 0xa8, 0xb7, 0x8b, 0xd9, 0xdd, 0xb1, 0x14, 0xcd, 0xc8, 0x69, 0xc5, 0x5d, 0x3f, 0x8d, + 0x54, 0x61, 0x35, 0x33, 0x4f, 0x9f, 0xc8, 0x9c, 0x0b, 0x74, 0xf2, 0xf8, 0x8c, 0x40, 0x6d, 0xd8, + 0xfa, 0x71, 0xcd, 0x82, 0xad, 0xd1, 0xdd, 0x5c, 0xcb, 0xa6, 0x5a, 0x4e, 0x53, 0xf5, 0x90, 0x84, + 0xe6, 0xaf, 0x04, 0x66, 0x33, 0x1d, 0x1d, 0x9b, 0x36, 0x2c, 0xcc, 0x06, 0xea, 0xde, 0xaa, 0xa4, + 0xa9, 0x00, 0x4e, 0x0f, 0x07, 0x7f, 0x11, 0x98, 0x2f, 0xe8, 0xfa, 0xb8, 0x5e, 0xc1, 0x9e, 0xb9, + 0x51, 0xc1, 0xad, 0xdc, 0x39, 0xbd, 0xf7, 0x64, 0xc6, 0x77, 0xf1, 0x4e, 0x29, 0xd6, 0xc6, 0x89, + 0x1e, 0x9b, 0xa0, 0xfe, 0x97, 0xc0, 0xcb, 0x85, 0x9d, 0x1e, 0xdf, 0xad, 0x62, 0xc0, 0xfc, 0xfc, + 0xe2, 0xde, 0x3f, 0xa5, 0x5a, 0x1d, 0x84, 0x59, 0x56, 0xd1, 0x41, 0x64, 0xfc, 0x72, 0xb2, 0x32, + 0xfc, 0x2d, 0x75, 0x75, 0xda, 0x16, 0xad, 0x5c, 0x6d, 0x4c, 0x10, 0xae, 0x65, 0xc3, 0xf5, 0xb6, + 0x64, 0x9e, 0x1b, 0xf8, 0xe0, 0x54, 0xf8, 0xd5, 0xbf, 0xe2, 0xf2, 0x20, 0xfe, 0x54, 0x9e, 0x57, + 0xb3, 0x80, 0x9d, 0xe7, 0xcd, 0x61, 0xc6, 0xce, 0xf3, 0xb9, 0x61, 0x23, 0x57, 0xc2, 0x78, 0xd4, + 0x85, 0x1e, 0xd2, 0x55, 0xe0, 0x8f, 0x04, 0xe0, 0x78, 0x9a, 0xc0, 0xb7, 0x2c, 0xa8, 0x9b, 0x93, + 0x87, 0x6b, 0xdb, 0xa3, 0xbd, 0x9b, 0x32, 0xe9, 0x37, 0xf1, 0xf5, 0x51, 0xdc, 0x75, 0x0f, 0x4f, + 0xd0, 0x3e, 0x27, 0x70, 0x21, 0x3b, 0x25, 0xa0, 0x15, 0xa7, 0xdc, 0xa4, 0xe3, 0xde, 0xae, 0x26, + 0x52, 0x74, 0xcd, 0x44, 0x8b, 0xdf, 0x28, 0x3a, 0xd7, 0x8d, 0x3d, 0xf0, 0x3a, 0x6c, 0xb7, 0x64, + 0x9f, 0xc7, 0xe4, 0xe3, 0x6d, 0x15, 0x11, 0xb0, 0xbe, 0x1f, 0x05, 0x75, 0x16, 0x07, 0x8d, 0x80, + 0x46, 0xf2, 0x53, 0x85, 0xfa, 0x0e, 0xe2, 0x0f, 0x42, 0x3e, 0xea, 0xd3, 0xc6, 0x3b, 0x99, 0xcb, + 0x4f, 0x66, 0xa4, 0xea, 0xd6, 0xff, 0x01, 0x00, 0x00, 0xff, 0xff, 0x69, 0xe0, 0x2a, 0x9a, 0xae, + 0x12, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/resultstore_file_download.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/resultstore_file_download.pb.go new file mode 100644 index 000000000..f0ab5a2f5 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/resultstore_file_download.pb.go @@ -0,0 +1,433 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/resultstore_file_download.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _ "google.golang.org/genproto/googleapis/api/annotations" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Request object for GetFile +type GetFileRequest struct { + // This corresponds to the uri field in the File message. + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` + // The offset for the first byte to return in the read, relative to the start + // of the resource. + // + // A `read_offset` that is negative or greater than the size of the resource + // will cause an `OUT_OF_RANGE` error. + ReadOffset int64 `protobuf:"varint,2,opt,name=read_offset,json=readOffset,proto3" json:"read_offset,omitempty"` + // The maximum number of `data` bytes the server is allowed to return in the + // sum of all `ReadResponse` messages. A `read_limit` of zero indicates that + // there is no limit, and a negative `read_limit` will cause an error. + // + // If the stream returns fewer bytes than allowed by the `read_limit` and no + // error occurred, the stream includes all data from the `read_offset` to the + // end of the resource. + ReadLimit int64 `protobuf:"varint,3,opt,name=read_limit,json=readLimit,proto3" json:"read_limit,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetFileRequest) Reset() { *m = GetFileRequest{} } +func (m *GetFileRequest) String() string { return proto.CompactTextString(m) } +func (*GetFileRequest) ProtoMessage() {} +func (*GetFileRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_file_download_02cbdf544c71a6bf, []int{0} +} +func (m *GetFileRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetFileRequest.Unmarshal(m, b) +} +func (m *GetFileRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetFileRequest.Marshal(b, m, deterministic) +} +func (dst *GetFileRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetFileRequest.Merge(dst, src) +} +func (m *GetFileRequest) XXX_Size() int { + return xxx_messageInfo_GetFileRequest.Size(m) +} +func (m *GetFileRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetFileRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetFileRequest proto.InternalMessageInfo + +func (m *GetFileRequest) GetUri() string { + if m != nil { + return m.Uri + } + return "" +} + +func (m *GetFileRequest) GetReadOffset() int64 { + if m != nil { + return m.ReadOffset + } + return 0 +} + +func (m *GetFileRequest) GetReadLimit() int64 { + if m != nil { + return m.ReadLimit + } + return 0 +} + +// Response object for GetFile +type GetFileResponse struct { + // The file data. + Data []byte `protobuf:"bytes,1,opt,name=data,proto3" json:"data,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetFileResponse) Reset() { *m = GetFileResponse{} } +func (m *GetFileResponse) String() string { return proto.CompactTextString(m) } +func (*GetFileResponse) ProtoMessage() {} +func (*GetFileResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_file_download_02cbdf544c71a6bf, []int{1} +} +func (m *GetFileResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetFileResponse.Unmarshal(m, b) +} +func (m *GetFileResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetFileResponse.Marshal(b, m, deterministic) +} +func (dst *GetFileResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetFileResponse.Merge(dst, src) +} +func (m *GetFileResponse) XXX_Size() int { + return xxx_messageInfo_GetFileResponse.Size(m) +} +func (m *GetFileResponse) XXX_DiscardUnknown() { + xxx_messageInfo_GetFileResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_GetFileResponse proto.InternalMessageInfo + +func (m *GetFileResponse) GetData() []byte { + if m != nil { + return m.Data + } + return nil +} + +// Request object for GetFileTail +type GetFileTailRequest struct { + // This corresponds to the uri field in the File message. + Uri string `protobuf:"bytes,1,opt,name=uri,proto3" json:"uri,omitempty"` + // The offset for the first byte to return in the read, relative to the end + // of the resource. + // + // A `read_offset` that is negative or greater than the size of the resource + // will cause an `OUT_OF_RANGE` error. + ReadOffset int64 `protobuf:"varint,2,opt,name=read_offset,json=readOffset,proto3" json:"read_offset,omitempty"` + // The maximum number of `data` bytes the server is allowed to return. The + // server will return bytes starting from the tail of the file. + // + // A `read_limit` of zero indicates that there is no limit, and a negative + // `read_limit` will cause an error. + ReadLimit int64 `protobuf:"varint,3,opt,name=read_limit,json=readLimit,proto3" json:"read_limit,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetFileTailRequest) Reset() { *m = GetFileTailRequest{} } +func (m *GetFileTailRequest) String() string { return proto.CompactTextString(m) } +func (*GetFileTailRequest) ProtoMessage() {} +func (*GetFileTailRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_file_download_02cbdf544c71a6bf, []int{2} +} +func (m *GetFileTailRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetFileTailRequest.Unmarshal(m, b) +} +func (m *GetFileTailRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetFileTailRequest.Marshal(b, m, deterministic) +} +func (dst *GetFileTailRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetFileTailRequest.Merge(dst, src) +} +func (m *GetFileTailRequest) XXX_Size() int { + return xxx_messageInfo_GetFileTailRequest.Size(m) +} +func (m *GetFileTailRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetFileTailRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetFileTailRequest proto.InternalMessageInfo + +func (m *GetFileTailRequest) GetUri() string { + if m != nil { + return m.Uri + } + return "" +} + +func (m *GetFileTailRequest) GetReadOffset() int64 { + if m != nil { + return m.ReadOffset + } + return 0 +} + +func (m *GetFileTailRequest) GetReadLimit() int64 { + if m != nil { + return m.ReadLimit + } + return 0 +} + +// Response object for GetFileTail +type GetFileTailResponse struct { + // The file data, encoded with UTF-8. + Data []byte `protobuf:"bytes,1,opt,name=data,proto3" json:"data,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetFileTailResponse) Reset() { *m = GetFileTailResponse{} } +func (m *GetFileTailResponse) String() string { return proto.CompactTextString(m) } +func (*GetFileTailResponse) ProtoMessage() {} +func (*GetFileTailResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_resultstore_file_download_02cbdf544c71a6bf, []int{3} +} +func (m *GetFileTailResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetFileTailResponse.Unmarshal(m, b) +} +func (m *GetFileTailResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetFileTailResponse.Marshal(b, m, deterministic) +} +func (dst *GetFileTailResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetFileTailResponse.Merge(dst, src) +} +func (m *GetFileTailResponse) XXX_Size() int { + return xxx_messageInfo_GetFileTailResponse.Size(m) +} +func (m *GetFileTailResponse) XXX_DiscardUnknown() { + xxx_messageInfo_GetFileTailResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_GetFileTailResponse proto.InternalMessageInfo + +func (m *GetFileTailResponse) GetData() []byte { + if m != nil { + return m.Data + } + return nil +} + +func init() { + proto.RegisterType((*GetFileRequest)(nil), "google.devtools.resultstore.v2.GetFileRequest") + proto.RegisterType((*GetFileResponse)(nil), "google.devtools.resultstore.v2.GetFileResponse") + proto.RegisterType((*GetFileTailRequest)(nil), "google.devtools.resultstore.v2.GetFileTailRequest") + proto.RegisterType((*GetFileTailResponse)(nil), "google.devtools.resultstore.v2.GetFileTailResponse") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// ResultStoreFileDownloadClient is the client API for ResultStoreFileDownload service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type ResultStoreFileDownloadClient interface { + // Retrieves the File with the given uri. + // returns a stream of bytes to be stitched together in order. + // + // An error will be reported in the following cases: + // - If the File is not found. + // - If the given File uri is badly formatted. + GetFile(ctx context.Context, in *GetFileRequest, opts ...grpc.CallOption) (ResultStoreFileDownload_GetFileClient, error) + // Retrieves the tail of a File with the given uri. + // + // An error will be reported in the following cases: + // - If the File is not found. + // - If the given File uri is badly formatted. + GetFileTail(ctx context.Context, in *GetFileTailRequest, opts ...grpc.CallOption) (*GetFileTailResponse, error) +} + +type resultStoreFileDownloadClient struct { + cc *grpc.ClientConn +} + +func NewResultStoreFileDownloadClient(cc *grpc.ClientConn) ResultStoreFileDownloadClient { + return &resultStoreFileDownloadClient{cc} +} + +func (c *resultStoreFileDownloadClient) GetFile(ctx context.Context, in *GetFileRequest, opts ...grpc.CallOption) (ResultStoreFileDownload_GetFileClient, error) { + stream, err := c.cc.NewStream(ctx, &_ResultStoreFileDownload_serviceDesc.Streams[0], "/google.devtools.resultstore.v2.ResultStoreFileDownload/GetFile", opts...) + if err != nil { + return nil, err + } + x := &resultStoreFileDownloadGetFileClient{stream} + if err := x.ClientStream.SendMsg(in); err != nil { + return nil, err + } + if err := x.ClientStream.CloseSend(); err != nil { + return nil, err + } + return x, nil +} + +type ResultStoreFileDownload_GetFileClient interface { + Recv() (*GetFileResponse, error) + grpc.ClientStream +} + +type resultStoreFileDownloadGetFileClient struct { + grpc.ClientStream +} + +func (x *resultStoreFileDownloadGetFileClient) Recv() (*GetFileResponse, error) { + m := new(GetFileResponse) + if err := x.ClientStream.RecvMsg(m); err != nil { + return nil, err + } + return m, nil +} + +func (c *resultStoreFileDownloadClient) GetFileTail(ctx context.Context, in *GetFileTailRequest, opts ...grpc.CallOption) (*GetFileTailResponse, error) { + out := new(GetFileTailResponse) + err := c.cc.Invoke(ctx, "/google.devtools.resultstore.v2.ResultStoreFileDownload/GetFileTail", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// ResultStoreFileDownloadServer is the server API for ResultStoreFileDownload service. +type ResultStoreFileDownloadServer interface { + // Retrieves the File with the given uri. + // returns a stream of bytes to be stitched together in order. + // + // An error will be reported in the following cases: + // - If the File is not found. + // - If the given File uri is badly formatted. + GetFile(*GetFileRequest, ResultStoreFileDownload_GetFileServer) error + // Retrieves the tail of a File with the given uri. + // + // An error will be reported in the following cases: + // - If the File is not found. + // - If the given File uri is badly formatted. + GetFileTail(context.Context, *GetFileTailRequest) (*GetFileTailResponse, error) +} + +func RegisterResultStoreFileDownloadServer(s *grpc.Server, srv ResultStoreFileDownloadServer) { + s.RegisterService(&_ResultStoreFileDownload_serviceDesc, srv) +} + +func _ResultStoreFileDownload_GetFile_Handler(srv interface{}, stream grpc.ServerStream) error { + m := new(GetFileRequest) + if err := stream.RecvMsg(m); err != nil { + return err + } + return srv.(ResultStoreFileDownloadServer).GetFile(m, &resultStoreFileDownloadGetFileServer{stream}) +} + +type ResultStoreFileDownload_GetFileServer interface { + Send(*GetFileResponse) error + grpc.ServerStream +} + +type resultStoreFileDownloadGetFileServer struct { + grpc.ServerStream +} + +func (x *resultStoreFileDownloadGetFileServer) Send(m *GetFileResponse) error { + return x.ServerStream.SendMsg(m) +} + +func _ResultStoreFileDownload_GetFileTail_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetFileTailRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(ResultStoreFileDownloadServer).GetFileTail(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.devtools.resultstore.v2.ResultStoreFileDownload/GetFileTail", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(ResultStoreFileDownloadServer).GetFileTail(ctx, req.(*GetFileTailRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _ResultStoreFileDownload_serviceDesc = grpc.ServiceDesc{ + ServiceName: "google.devtools.resultstore.v2.ResultStoreFileDownload", + HandlerType: (*ResultStoreFileDownloadServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "GetFileTail", + Handler: _ResultStoreFileDownload_GetFileTail_Handler, + }, + }, + Streams: []grpc.StreamDesc{ + { + StreamName: "GetFile", + Handler: _ResultStoreFileDownload_GetFile_Handler, + ServerStreams: true, + }, + }, + Metadata: "google/devtools/resultstore/v2/resultstore_file_download.proto", +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/resultstore_file_download.proto", fileDescriptor_resultstore_file_download_02cbdf544c71a6bf) +} + +var fileDescriptor_resultstore_file_download_02cbdf544c71a6bf = []byte{ + // 384 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x53, 0xc1, 0x4a, 0xe3, 0x40, + 0x18, 0x26, 0xed, 0xb2, 0x4b, 0xa7, 0xcb, 0x6e, 0x99, 0x65, 0x69, 0x28, 0xdb, 0xb5, 0x04, 0x84, + 0xda, 0xc3, 0x8c, 0xa4, 0x47, 0xd1, 0x83, 0x88, 0x22, 0x08, 0x4a, 0xf4, 0xe4, 0x25, 0x4c, 0xcd, + 0x24, 0x0c, 0x4c, 0xf3, 0xa7, 0x99, 0x49, 0x3d, 0x48, 0x2f, 0x1e, 0x7c, 0x01, 0xf1, 0x65, 0x7c, + 0x0d, 0x5f, 0xc1, 0x07, 0x91, 0x99, 0x44, 0x49, 0x0f, 0xd5, 0x7a, 0xf0, 0x36, 0xff, 0x37, 0xff, + 0xf7, 0xfd, 0x5f, 0xbe, 0xfc, 0x83, 0xf6, 0x12, 0x80, 0x44, 0x72, 0x1a, 0xf1, 0xb9, 0x06, 0x90, + 0x8a, 0xe6, 0x5c, 0x15, 0x52, 0x2b, 0x0d, 0x39, 0xa7, 0x73, 0xbf, 0x5e, 0x86, 0xb1, 0x90, 0x3c, + 0x8c, 0xe0, 0x3a, 0x95, 0xc0, 0x22, 0x92, 0xe5, 0xa0, 0x01, 0xff, 0x2f, 0xf9, 0xe4, 0x95, 0x4f, + 0x6a, 0x04, 0x32, 0xf7, 0x7b, 0xff, 0x2a, 0x7d, 0x96, 0x09, 0xca, 0xd2, 0x14, 0x34, 0xd3, 0x02, + 0x52, 0x55, 0xb2, 0xbd, 0x09, 0xfa, 0x75, 0xc4, 0xf5, 0xa1, 0x90, 0x3c, 0xe0, 0xb3, 0x82, 0x2b, + 0x8d, 0x3b, 0xa8, 0x59, 0xe4, 0xc2, 0x75, 0x06, 0xce, 0xb0, 0x15, 0x98, 0x23, 0xde, 0x40, 0xed, + 0x9c, 0xb3, 0x28, 0x84, 0x38, 0x56, 0x5c, 0xbb, 0x8d, 0x81, 0x33, 0x6c, 0x06, 0xc8, 0x40, 0xa7, + 0x16, 0xc1, 0x7d, 0x64, 0xab, 0x50, 0x8a, 0xa9, 0xd0, 0x6e, 0xd3, 0xde, 0xb7, 0x0c, 0x72, 0x62, + 0x00, 0x6f, 0x13, 0xfd, 0x7e, 0x9b, 0xa1, 0x32, 0x48, 0x15, 0xc7, 0x18, 0x7d, 0x8b, 0x98, 0x66, + 0x76, 0xca, 0xcf, 0xc0, 0x9e, 0xbd, 0x18, 0xe1, 0xaa, 0xed, 0x82, 0x09, 0xf9, 0x75, 0x76, 0xb6, + 0xd0, 0x9f, 0xa5, 0x39, 0xab, 0x2d, 0xf9, 0x8f, 0x0d, 0xd4, 0x0d, 0x6c, 0x9c, 0xe7, 0x26, 0x4e, + 0xc3, 0x39, 0xa8, 0xd2, 0xc7, 0x77, 0x0e, 0xfa, 0x51, 0xe9, 0x60, 0x42, 0xde, 0xff, 0x09, 0x64, + 0x39, 0xe3, 0x1e, 0x5d, 0xbb, 0xbf, 0x34, 0xe7, 0xb9, 0xb7, 0x4f, 0xcf, 0xf7, 0x0d, 0x8c, 0x3b, + 0x66, 0x23, 0x6e, 0x8a, 0x5c, 0xec, 0x9a, 0x55, 0xa0, 0xa3, 0xc5, 0xb6, 0x83, 0x1f, 0x1c, 0xd4, + 0xae, 0x7d, 0x10, 0xf6, 0xd7, 0x14, 0xaf, 0xa5, 0xdc, 0x1b, 0x7f, 0x8a, 0x53, 0x99, 0xea, 0x5b, + 0x53, 0x5d, 0xfc, 0x77, 0xd9, 0x94, 0x66, 0x42, 0xd2, 0xd1, 0x62, 0x7f, 0x86, 0xbc, 0x2b, 0x98, + 0x7e, 0x20, 0x7c, 0xe6, 0x5c, 0x1e, 0x57, 0x1d, 0x09, 0x48, 0x96, 0x26, 0x04, 0xf2, 0x84, 0x26, + 0x3c, 0xb5, 0xeb, 0x49, 0xcb, 0x2b, 0x96, 0x09, 0xb5, 0xea, 0x7d, 0xec, 0xd4, 0xca, 0xc9, 0x77, + 0xcb, 0x1a, 0xbf, 0x04, 0x00, 0x00, 0xff, 0xff, 0x0d, 0x3d, 0x87, 0x5d, 0x54, 0x03, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/target.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/target.pb.go new file mode 100644 index 000000000..f2599a305 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/target.pb.go @@ -0,0 +1,434 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/target.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// These correspond to the suffix of the rule name. Eg cc_test has type TEST. +type TargetType int32 + +const ( + // Unspecified by the build system. + TargetType_TARGET_TYPE_UNSPECIFIED TargetType = 0 + // An application e.g. ios_application. + TargetType_APPLICATION TargetType = 1 + // A binary target e.g. cc_binary. + TargetType_BINARY TargetType = 2 + // A library target e.g. java_library + TargetType_LIBRARY TargetType = 3 + // A package + TargetType_PACKAGE TargetType = 4 + // Any test target, in bazel that means a rule with a '_test' suffix. + TargetType_TEST TargetType = 5 +) + +var TargetType_name = map[int32]string{ + 0: "TARGET_TYPE_UNSPECIFIED", + 1: "APPLICATION", + 2: "BINARY", + 3: "LIBRARY", + 4: "PACKAGE", + 5: "TEST", +} +var TargetType_value = map[string]int32{ + "TARGET_TYPE_UNSPECIFIED": 0, + "APPLICATION": 1, + "BINARY": 2, + "LIBRARY": 3, + "PACKAGE": 4, + "TEST": 5, +} + +func (x TargetType) String() string { + return proto.EnumName(TargetType_name, int32(x)) +} +func (TargetType) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_target_3b072e177729a29a, []int{0} +} + +// Indicates how big the user indicated the test action was. +type TestSize int32 + +const ( + // Unspecified by the user. + TestSize_TEST_SIZE_UNSPECIFIED TestSize = 0 + // Unit test taking less than 1 minute. + TestSize_SMALL TestSize = 1 + // Integration tests taking less than 5 minutes. + TestSize_MEDIUM TestSize = 2 + // End-to-end tests taking less than 15 minutes. + TestSize_LARGE TestSize = 3 + // Even bigger than LARGE. + TestSize_ENORMOUS TestSize = 4 + // Something that doesn't fit into the above categories. + TestSize_OTHER_SIZE TestSize = 5 +) + +var TestSize_name = map[int32]string{ + 0: "TEST_SIZE_UNSPECIFIED", + 1: "SMALL", + 2: "MEDIUM", + 3: "LARGE", + 4: "ENORMOUS", + 5: "OTHER_SIZE", +} +var TestSize_value = map[string]int32{ + "TEST_SIZE_UNSPECIFIED": 0, + "SMALL": 1, + "MEDIUM": 2, + "LARGE": 3, + "ENORMOUS": 4, + "OTHER_SIZE": 5, +} + +func (x TestSize) String() string { + return proto.EnumName(TestSize_name, int32(x)) +} +func (TestSize) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_target_3b072e177729a29a, []int{1} +} + +// Each Target represents data for a given target in a given Invocation. +// ConfiguredTarget and Action resources under each Target contain the bulk of +// the data. +type Target struct { + // The resource name. Its format must be: + // invocations/${INVOCATION_ID}/targets/${TARGET_ID} + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // The resource ID components that identify the Target. They must match the + // resource name after proper encoding. + Id *Target_Id `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` + // This is the aggregate status of the target. + StatusAttributes *StatusAttributes `protobuf:"bytes,3,opt,name=status_attributes,json=statusAttributes,proto3" json:"status_attributes,omitempty"` + // When this target started and its duration. + Timing *Timing `protobuf:"bytes,4,opt,name=timing,proto3" json:"timing,omitempty"` + // Attributes that apply to all targets. + TargetAttributes *TargetAttributes `protobuf:"bytes,5,opt,name=target_attributes,json=targetAttributes,proto3" json:"target_attributes,omitempty"` + // Attributes that apply to all test actions under this target. + TestAttributes *TestAttributes `protobuf:"bytes,6,opt,name=test_attributes,json=testAttributes,proto3" json:"test_attributes,omitempty"` + // Arbitrary name-value pairs. + // This is implemented as a multi-map. Multiple properties are allowed with + // the same key. Properties will be returned in lexicographical order by key. + Properties []*Property `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty"` + // A list of file references for target level files. + // The file IDs must be unique within this list. Duplicate file IDs will + // result in an error. Files will be returned in lexicographical order by ID. + // Use this field to specify outputs not related to a configuration. + Files []*File `protobuf:"bytes,8,rep,name=files,proto3" json:"files,omitempty"` + // Provides a hint to clients as to whether to display the Target to users. + // If true then clients likely want to display the Target by default. + // Once set to true, this may not be set back to false. + IsVisible bool `protobuf:"varint,9,opt,name=is_visible,json=isVisible,proto3" json:"is_visible,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Target) Reset() { *m = Target{} } +func (m *Target) String() string { return proto.CompactTextString(m) } +func (*Target) ProtoMessage() {} +func (*Target) Descriptor() ([]byte, []int) { + return fileDescriptor_target_3b072e177729a29a, []int{0} +} +func (m *Target) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Target.Unmarshal(m, b) +} +func (m *Target) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Target.Marshal(b, m, deterministic) +} +func (dst *Target) XXX_Merge(src proto.Message) { + xxx_messageInfo_Target.Merge(dst, src) +} +func (m *Target) XXX_Size() int { + return xxx_messageInfo_Target.Size(m) +} +func (m *Target) XXX_DiscardUnknown() { + xxx_messageInfo_Target.DiscardUnknown(m) +} + +var xxx_messageInfo_Target proto.InternalMessageInfo + +func (m *Target) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *Target) GetId() *Target_Id { + if m != nil { + return m.Id + } + return nil +} + +func (m *Target) GetStatusAttributes() *StatusAttributes { + if m != nil { + return m.StatusAttributes + } + return nil +} + +func (m *Target) GetTiming() *Timing { + if m != nil { + return m.Timing + } + return nil +} + +func (m *Target) GetTargetAttributes() *TargetAttributes { + if m != nil { + return m.TargetAttributes + } + return nil +} + +func (m *Target) GetTestAttributes() *TestAttributes { + if m != nil { + return m.TestAttributes + } + return nil +} + +func (m *Target) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +func (m *Target) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +func (m *Target) GetIsVisible() bool { + if m != nil { + return m.IsVisible + } + return false +} + +// The resource ID components that identify the Target. +type Target_Id struct { + // The Invocation ID. + InvocationId string `protobuf:"bytes,1,opt,name=invocation_id,json=invocationId,proto3" json:"invocation_id,omitempty"` + // The Target ID. + TargetId string `protobuf:"bytes,2,opt,name=target_id,json=targetId,proto3" json:"target_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Target_Id) Reset() { *m = Target_Id{} } +func (m *Target_Id) String() string { return proto.CompactTextString(m) } +func (*Target_Id) ProtoMessage() {} +func (*Target_Id) Descriptor() ([]byte, []int) { + return fileDescriptor_target_3b072e177729a29a, []int{0, 0} +} +func (m *Target_Id) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Target_Id.Unmarshal(m, b) +} +func (m *Target_Id) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Target_Id.Marshal(b, m, deterministic) +} +func (dst *Target_Id) XXX_Merge(src proto.Message) { + xxx_messageInfo_Target_Id.Merge(dst, src) +} +func (m *Target_Id) XXX_Size() int { + return xxx_messageInfo_Target_Id.Size(m) +} +func (m *Target_Id) XXX_DiscardUnknown() { + xxx_messageInfo_Target_Id.DiscardUnknown(m) +} + +var xxx_messageInfo_Target_Id proto.InternalMessageInfo + +func (m *Target_Id) GetInvocationId() string { + if m != nil { + return m.InvocationId + } + return "" +} + +func (m *Target_Id) GetTargetId() string { + if m != nil { + return m.TargetId + } + return "" +} + +// Attributes that apply to all targets. +type TargetAttributes struct { + // If known, indicates the type of this target. In bazel this corresponds + // to the rule-suffix. + Type TargetType `protobuf:"varint,1,opt,name=type,proto3,enum=google.devtools.resultstore.v2.TargetType" json:"type,omitempty"` + // If known, the main language of this target, e.g. java, cc, python, etc. + Language Language `protobuf:"varint,2,opt,name=language,proto3,enum=google.devtools.resultstore.v2.Language" json:"language,omitempty"` + // The tags attribute of the build rule. These should be short, descriptive + // words, and there should only be a few of them. + // This is implemented as a set. All tags will be unique. Any duplicate tags + // will be ignored. Tags will be returned in lexicographical order. + Tags []string `protobuf:"bytes,3,rep,name=tags,proto3" json:"tags,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TargetAttributes) Reset() { *m = TargetAttributes{} } +func (m *TargetAttributes) String() string { return proto.CompactTextString(m) } +func (*TargetAttributes) ProtoMessage() {} +func (*TargetAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_target_3b072e177729a29a, []int{1} +} +func (m *TargetAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TargetAttributes.Unmarshal(m, b) +} +func (m *TargetAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TargetAttributes.Marshal(b, m, deterministic) +} +func (dst *TargetAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_TargetAttributes.Merge(dst, src) +} +func (m *TargetAttributes) XXX_Size() int { + return xxx_messageInfo_TargetAttributes.Size(m) +} +func (m *TargetAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_TargetAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_TargetAttributes proto.InternalMessageInfo + +func (m *TargetAttributes) GetType() TargetType { + if m != nil { + return m.Type + } + return TargetType_TARGET_TYPE_UNSPECIFIED +} + +func (m *TargetAttributes) GetLanguage() Language { + if m != nil { + return m.Language + } + return Language_LANGUAGE_UNSPECIFIED +} + +func (m *TargetAttributes) GetTags() []string { + if m != nil { + return m.Tags + } + return nil +} + +// Attributes that apply only to test actions under this target. +type TestAttributes struct { + // Indicates how big the user indicated the test action was. + Size TestSize `protobuf:"varint,1,opt,name=size,proto3,enum=google.devtools.resultstore.v2.TestSize" json:"size,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestAttributes) Reset() { *m = TestAttributes{} } +func (m *TestAttributes) String() string { return proto.CompactTextString(m) } +func (*TestAttributes) ProtoMessage() {} +func (*TestAttributes) Descriptor() ([]byte, []int) { + return fileDescriptor_target_3b072e177729a29a, []int{2} +} +func (m *TestAttributes) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestAttributes.Unmarshal(m, b) +} +func (m *TestAttributes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestAttributes.Marshal(b, m, deterministic) +} +func (dst *TestAttributes) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestAttributes.Merge(dst, src) +} +func (m *TestAttributes) XXX_Size() int { + return xxx_messageInfo_TestAttributes.Size(m) +} +func (m *TestAttributes) XXX_DiscardUnknown() { + xxx_messageInfo_TestAttributes.DiscardUnknown(m) +} + +var xxx_messageInfo_TestAttributes proto.InternalMessageInfo + +func (m *TestAttributes) GetSize() TestSize { + if m != nil { + return m.Size + } + return TestSize_TEST_SIZE_UNSPECIFIED +} + +func init() { + proto.RegisterType((*Target)(nil), "google.devtools.resultstore.v2.Target") + proto.RegisterType((*Target_Id)(nil), "google.devtools.resultstore.v2.Target.Id") + proto.RegisterType((*TargetAttributes)(nil), "google.devtools.resultstore.v2.TargetAttributes") + proto.RegisterType((*TestAttributes)(nil), "google.devtools.resultstore.v2.TestAttributes") + proto.RegisterEnum("google.devtools.resultstore.v2.TargetType", TargetType_name, TargetType_value) + proto.RegisterEnum("google.devtools.resultstore.v2.TestSize", TestSize_name, TestSize_value) +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/target.proto", fileDescriptor_target_3b072e177729a29a) +} + +var fileDescriptor_target_3b072e177729a29a = []byte{ + // 634 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x54, 0xed, 0x6a, 0xdb, 0x4a, + 0x10, 0xbd, 0xb2, 0x65, 0x47, 0x1e, 0xe7, 0x3a, 0xba, 0x0b, 0x97, 0xaa, 0x09, 0x2d, 0xc6, 0x2d, + 0xc5, 0x49, 0x41, 0x2e, 0xee, 0xaf, 0x7e, 0x10, 0x50, 0x12, 0x25, 0x11, 0xf5, 0x87, 0x58, 0x29, + 0x2d, 0x09, 0x14, 0xa3, 0x44, 0x5b, 0xb1, 0x45, 0xd6, 0xba, 0xda, 0xb5, 0x21, 0x79, 0x95, 0xbe, + 0x42, 0x1f, 0xb2, 0x68, 0xd7, 0x49, 0x1c, 0x43, 0xab, 0xfc, 0xdb, 0x19, 0x9d, 0x73, 0xe6, 0x68, + 0x66, 0x76, 0xe1, 0x75, 0xc2, 0x58, 0x92, 0x92, 0x5e, 0x4c, 0x16, 0x82, 0xb1, 0x94, 0xf7, 0x72, + 0xc2, 0xe7, 0xa9, 0xe0, 0x82, 0xe5, 0xa4, 0xb7, 0xe8, 0xf7, 0x44, 0x94, 0x27, 0x44, 0xd8, 0xb3, + 0x9c, 0x09, 0x86, 0x9e, 0x2b, 0xb0, 0x7d, 0x0b, 0xb6, 0x57, 0xc0, 0xf6, 0xa2, 0xbf, 0x5d, 0x26, + 0x76, 0xc5, 0xa6, 0x53, 0x96, 0x29, 0xb1, 0xed, 0xdd, 0x12, 0xf0, 0x37, 0x9a, 0x12, 0x05, 0xed, + 0xfc, 0xac, 0x41, 0x3d, 0x94, 0x46, 0x10, 0x02, 0x3d, 0x8b, 0xa6, 0xc4, 0xd2, 0xda, 0x5a, 0xb7, + 0x81, 0xe5, 0x19, 0xbd, 0x83, 0x0a, 0x8d, 0xad, 0x4a, 0x5b, 0xeb, 0x36, 0xfb, 0xbb, 0xf6, 0xdf, + 0x3d, 0xda, 0x4a, 0xc7, 0xf6, 0x62, 0x5c, 0xa1, 0x31, 0xfa, 0x0a, 0xff, 0x71, 0x11, 0x89, 0x39, + 0x9f, 0x44, 0x42, 0xe4, 0xf4, 0x72, 0x2e, 0x08, 0xb7, 0xaa, 0x52, 0xe9, 0x4d, 0x99, 0x52, 0x20, + 0x89, 0xce, 0x1d, 0x0f, 0x9b, 0x7c, 0x2d, 0x83, 0xf6, 0xa1, 0x2e, 0xe8, 0x94, 0x66, 0x89, 0xa5, + 0x4b, 0xcd, 0x57, 0xa5, 0xee, 0x24, 0x1a, 0x2f, 0x59, 0x85, 0x3d, 0x35, 0x80, 0x55, 0x7b, 0xb5, + 0xc7, 0xd9, 0x53, 0x3f, 0xba, 0x6a, 0x4f, 0xac, 0x65, 0xd0, 0x17, 0xd8, 0x12, 0x84, 0x3f, 0x10, + 0xaf, 0x4b, 0x71, 0xbb, 0x54, 0x9c, 0xf0, 0x55, 0xe9, 0x96, 0x78, 0x10, 0xa3, 0x53, 0x80, 0x59, + 0xce, 0x66, 0x24, 0x17, 0x94, 0x70, 0x6b, 0xa3, 0x5d, 0xed, 0x36, 0xfb, 0xdd, 0x32, 0x4d, 0x5f, + 0x31, 0xae, 0xf1, 0x0a, 0x17, 0xbd, 0x87, 0x5a, 0xb1, 0x08, 0xdc, 0x32, 0xa4, 0xc8, 0xcb, 0x32, + 0x91, 0x63, 0x9a, 0x12, 0xac, 0x28, 0xe8, 0x19, 0x00, 0xe5, 0x93, 0x05, 0xe5, 0xf4, 0x32, 0x25, + 0x56, 0xa3, 0xad, 0x75, 0x0d, 0xdc, 0xa0, 0xfc, 0xb3, 0x4a, 0x6c, 0x1f, 0x43, 0xc5, 0x8b, 0xd1, + 0x0b, 0xf8, 0x97, 0x66, 0x0b, 0x76, 0x15, 0x09, 0xca, 0xb2, 0x09, 0x8d, 0x97, 0x9b, 0xb5, 0x79, + 0x9f, 0xf4, 0x62, 0xb4, 0x03, 0x8d, 0xe5, 0x1c, 0x96, 0x8b, 0xd6, 0xc0, 0x86, 0x4a, 0x78, 0x71, + 0xe7, 0x97, 0x06, 0xe6, 0x7a, 0xb3, 0xd1, 0x3e, 0xe8, 0xe2, 0x7a, 0xa6, 0xf6, 0xb4, 0xd5, 0xdf, + 0x7b, 0xdc, 0xb0, 0xc2, 0xeb, 0x19, 0xc1, 0x92, 0x87, 0x8e, 0xc0, 0x48, 0xa3, 0x2c, 0x99, 0x47, + 0x09, 0x91, 0x05, 0x5b, 0xe5, 0xfd, 0x1b, 0x2c, 0xf1, 0xf8, 0x8e, 0x59, 0xdc, 0x16, 0x11, 0x25, + 0xc5, 0x46, 0x57, 0x8b, 0xdb, 0x52, 0x9c, 0x3b, 0x23, 0x68, 0x3d, 0x9c, 0x1e, 0xfa, 0x08, 0x3a, + 0xa7, 0x37, 0xb7, 0x5e, 0xbb, 0x8f, 0x99, 0x7d, 0x40, 0x6f, 0x08, 0x96, 0xac, 0xbd, 0xef, 0x00, + 0xf7, 0xee, 0xd1, 0x0e, 0x3c, 0x09, 0x1d, 0x7c, 0xe2, 0x86, 0x93, 0xf0, 0xdc, 0x77, 0x27, 0x67, + 0xa3, 0xc0, 0x77, 0x0f, 0xbd, 0x63, 0xcf, 0x3d, 0x32, 0xff, 0x41, 0x5b, 0xd0, 0x74, 0x7c, 0x7f, + 0xe0, 0x1d, 0x3a, 0xa1, 0x37, 0x1e, 0x99, 0x1a, 0x02, 0xa8, 0x1f, 0x78, 0x23, 0x07, 0x9f, 0x9b, + 0x15, 0xd4, 0x84, 0x8d, 0x81, 0x77, 0x80, 0x8b, 0xa0, 0x5a, 0x04, 0xbe, 0x73, 0xf8, 0xc9, 0x39, + 0x71, 0x4d, 0x1d, 0x19, 0xa0, 0x87, 0x6e, 0x10, 0x9a, 0xb5, 0x3d, 0x02, 0xc6, 0x6d, 0x75, 0xf4, + 0x14, 0xfe, 0x2f, 0xb2, 0x93, 0xc0, 0xbb, 0x58, 0xaf, 0xd3, 0x80, 0x5a, 0x30, 0x74, 0x06, 0x03, + 0x55, 0x61, 0xe8, 0x1e, 0x79, 0x67, 0x43, 0xb3, 0x52, 0xa4, 0x07, 0x85, 0x37, 0xb3, 0x8a, 0x36, + 0xc1, 0x70, 0x47, 0x63, 0x3c, 0x1c, 0x9f, 0x05, 0xa6, 0x8e, 0x5a, 0x00, 0xe3, 0xf0, 0xd4, 0xc5, + 0x52, 0xcb, 0xac, 0x1d, 0xfc, 0x80, 0xce, 0x15, 0x9b, 0x96, 0xf4, 0xc1, 0xd7, 0x2e, 0xbc, 0x25, + 0x22, 0x61, 0x45, 0xc7, 0x6d, 0x96, 0x27, 0xbd, 0x84, 0x64, 0xf2, 0xcd, 0xea, 0xa9, 0x4f, 0xd1, + 0x8c, 0xf2, 0x3f, 0xbd, 0x70, 0x1f, 0x56, 0xc2, 0xcb, 0xba, 0x64, 0xbd, 0xfd, 0x1d, 0x00, 0x00, + 0xff, 0xff, 0xc9, 0x5f, 0x96, 0x8a, 0x90, 0x05, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/test_suite.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/test_suite.pb.go new file mode 100644 index 000000000..81f6b9c2f --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/devtools/resultstore/v2/test_suite.pb.go @@ -0,0 +1,630 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/devtools/resultstore/v2/test_suite.proto + +package resultstore // import "google.golang.org/genproto/googleapis/devtools/resultstore/v2" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// The result of running a test case. +type TestCase_Result int32 + +const ( + // The implicit default enum value. Do not use. + TestCase_RESULT_UNSPECIFIED TestCase_Result = 0 + // Test case ran to completion. Look for failures or errors to determine + // whether it passed, failed, or errored. + TestCase_COMPLETED TestCase_Result = 1 + // Test case started but did not complete because the test harness received + // a signal and decided to stop running tests. + TestCase_INTERRUPTED TestCase_Result = 2 + // Test case was not started because the test harness received a SIGINT or + // timed out. + TestCase_CANCELLED TestCase_Result = 3 + // Test case was not run because the user or process running the test + // specified a filter that excluded this test case. + TestCase_FILTERED TestCase_Result = 4 + // Test case was not run to completion because the test case decided it + // should not be run (eg. due to a failed assumption in a JUnit4 test). + // Per-test setup or tear-down may or may not have run. + TestCase_SKIPPED TestCase_Result = 5 + // The test framework did not run the test case because it was labeled as + // suppressed. Eg. if someone temporarily disables a failing test. + TestCase_SUPPRESSED TestCase_Result = 6 +) + +var TestCase_Result_name = map[int32]string{ + 0: "RESULT_UNSPECIFIED", + 1: "COMPLETED", + 2: "INTERRUPTED", + 3: "CANCELLED", + 4: "FILTERED", + 5: "SKIPPED", + 6: "SUPPRESSED", +} +var TestCase_Result_value = map[string]int32{ + "RESULT_UNSPECIFIED": 0, + "COMPLETED": 1, + "INTERRUPTED": 2, + "CANCELLED": 3, + "FILTERED": 4, + "SKIPPED": 5, + "SUPPRESSED": 6, +} + +func (x TestCase_Result) String() string { + return proto.EnumName(TestCase_Result_name, int32(x)) +} +func (TestCase_Result) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_test_suite_ee31d75e3e5b9dce, []int{2, 0} +} + +// The result of running a test suite, as reported in a element of +// an XML log. +type TestSuite struct { + // The full name of this suite, as reported in the name attribute. For Java + // tests, this is normally the fully qualified class name. Eg. + // "com.google.common.hash.BloomFilterTest". + SuiteName string `protobuf:"bytes,1,opt,name=suite_name,json=suiteName,proto3" json:"suite_name,omitempty"` + // The results of the test cases and test suites contained in this suite, + // as reported in the and elements contained within + // this . + Tests []*Test `protobuf:"bytes,2,rep,name=tests,proto3" json:"tests,omitempty"` + // Failures reported in elements within this . + Failures []*TestFailure `protobuf:"bytes,3,rep,name=failures,proto3" json:"failures,omitempty"` + // Errors reported in elements within this . + Errors []*TestError `protobuf:"bytes,4,rep,name=errors,proto3" json:"errors,omitempty"` + // The timing for the entire TestSuite, as reported by the time attribute. + Timing *Timing `protobuf:"bytes,6,opt,name=timing,proto3" json:"timing,omitempty"` + // Arbitrary name-value pairs, as reported in custom attributes or in a + // element within this . Multiple properties are + // allowed with the same key. Properties will be returned in lexicographical + // order by key. + Properties []*Property `protobuf:"bytes,7,rep,name=properties,proto3" json:"properties,omitempty"` + // Files produced by this test suite, as reported by undeclared output + // annotations. + // The file IDs must be unique within this list. Duplicate file IDs will + // result in an error. Files will be returned in lexicographical order by ID. + Files []*File `protobuf:"bytes,8,rep,name=files,proto3" json:"files,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestSuite) Reset() { *m = TestSuite{} } +func (m *TestSuite) String() string { return proto.CompactTextString(m) } +func (*TestSuite) ProtoMessage() {} +func (*TestSuite) Descriptor() ([]byte, []int) { + return fileDescriptor_test_suite_ee31d75e3e5b9dce, []int{0} +} +func (m *TestSuite) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestSuite.Unmarshal(m, b) +} +func (m *TestSuite) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestSuite.Marshal(b, m, deterministic) +} +func (dst *TestSuite) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestSuite.Merge(dst, src) +} +func (m *TestSuite) XXX_Size() int { + return xxx_messageInfo_TestSuite.Size(m) +} +func (m *TestSuite) XXX_DiscardUnknown() { + xxx_messageInfo_TestSuite.DiscardUnknown(m) +} + +var xxx_messageInfo_TestSuite proto.InternalMessageInfo + +func (m *TestSuite) GetSuiteName() string { + if m != nil { + return m.SuiteName + } + return "" +} + +func (m *TestSuite) GetTests() []*Test { + if m != nil { + return m.Tests + } + return nil +} + +func (m *TestSuite) GetFailures() []*TestFailure { + if m != nil { + return m.Failures + } + return nil +} + +func (m *TestSuite) GetErrors() []*TestError { + if m != nil { + return m.Errors + } + return nil +} + +func (m *TestSuite) GetTiming() *Timing { + if m != nil { + return m.Timing + } + return nil +} + +func (m *TestSuite) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +func (m *TestSuite) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +// The result of running a test case or test suite. JUnit3 TestDecorators are +// represented as a TestSuite with a single test. +type Test struct { + // Either a TestCase of a TestSuite + // + // Types that are valid to be assigned to TestType: + // *Test_TestCase + // *Test_TestSuite + TestType isTest_TestType `protobuf_oneof:"test_type"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Test) Reset() { *m = Test{} } +func (m *Test) String() string { return proto.CompactTextString(m) } +func (*Test) ProtoMessage() {} +func (*Test) Descriptor() ([]byte, []int) { + return fileDescriptor_test_suite_ee31d75e3e5b9dce, []int{1} +} +func (m *Test) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Test.Unmarshal(m, b) +} +func (m *Test) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Test.Marshal(b, m, deterministic) +} +func (dst *Test) XXX_Merge(src proto.Message) { + xxx_messageInfo_Test.Merge(dst, src) +} +func (m *Test) XXX_Size() int { + return xxx_messageInfo_Test.Size(m) +} +func (m *Test) XXX_DiscardUnknown() { + xxx_messageInfo_Test.DiscardUnknown(m) +} + +var xxx_messageInfo_Test proto.InternalMessageInfo + +type isTest_TestType interface { + isTest_TestType() +} + +type Test_TestCase struct { + TestCase *TestCase `protobuf:"bytes,1,opt,name=test_case,json=testCase,proto3,oneof"` +} +type Test_TestSuite struct { + TestSuite *TestSuite `protobuf:"bytes,2,opt,name=test_suite,json=testSuite,proto3,oneof"` +} + +func (*Test_TestCase) isTest_TestType() {} +func (*Test_TestSuite) isTest_TestType() {} + +func (m *Test) GetTestType() isTest_TestType { + if m != nil { + return m.TestType + } + return nil +} + +func (m *Test) GetTestCase() *TestCase { + if x, ok := m.GetTestType().(*Test_TestCase); ok { + return x.TestCase + } + return nil +} + +func (m *Test) GetTestSuite() *TestSuite { + if x, ok := m.GetTestType().(*Test_TestSuite); ok { + return x.TestSuite + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*Test) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _Test_OneofMarshaler, _Test_OneofUnmarshaler, _Test_OneofSizer, []interface{}{ + (*Test_TestCase)(nil), + (*Test_TestSuite)(nil), + } +} + +func _Test_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*Test) + // test_type + switch x := m.TestType.(type) { + case *Test_TestCase: + b.EncodeVarint(1<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.TestCase); err != nil { + return err + } + case *Test_TestSuite: + b.EncodeVarint(2<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.TestSuite); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("Test.TestType has unexpected type %T", x) + } + return nil +} + +func _Test_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*Test) + switch tag { + case 1: // test_type.test_case + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(TestCase) + err := b.DecodeMessage(msg) + m.TestType = &Test_TestCase{msg} + return true, err + case 2: // test_type.test_suite + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(TestSuite) + err := b.DecodeMessage(msg) + m.TestType = &Test_TestSuite{msg} + return true, err + default: + return false, nil + } +} + +func _Test_OneofSizer(msg proto.Message) (n int) { + m := msg.(*Test) + // test_type + switch x := m.TestType.(type) { + case *Test_TestCase: + s := proto.Size(x.TestCase) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *Test_TestSuite: + s := proto.Size(x.TestSuite) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// The result of running a test case, as reported in a element of +// an XML log. +type TestCase struct { + // The name of the test case, as reported in the name attribute. For Java, + // this is normally the method name. Eg. "testBasic". + CaseName string `protobuf:"bytes,1,opt,name=case_name,json=caseName,proto3" json:"case_name,omitempty"` + // The name of the class in which the test case was defined, as reported in + // the classname attribute. For Java, this is normally the fully qualified + // class name. Eg. "com.google.common.hash.BloomFilterTest". + ClassName string `protobuf:"bytes,2,opt,name=class_name,json=className,proto3" json:"class_name,omitempty"` + // An enum reported in the result attribute that is used in conjunction with + // failures and errors below to report the outcome. + Result TestCase_Result `protobuf:"varint,3,opt,name=result,proto3,enum=google.devtools.resultstore.v2.TestCase_Result" json:"result,omitempty"` + // Failures reported in elements within this . + Failures []*TestFailure `protobuf:"bytes,4,rep,name=failures,proto3" json:"failures,omitempty"` + // Errors reported in elements within this . + Errors []*TestError `protobuf:"bytes,5,rep,name=errors,proto3" json:"errors,omitempty"` + // The timing for the TestCase, as reported by the time attribute. + Timing *Timing `protobuf:"bytes,7,opt,name=timing,proto3" json:"timing,omitempty"` + // Arbitrary name-value pairs, as reported in custom attributes or in a + // element within this . Multiple properties are + // allowed with the same key. Properties will be returned in lexicographical + // order by key. + Properties []*Property `protobuf:"bytes,8,rep,name=properties,proto3" json:"properties,omitempty"` + // Files produced by this test case, as reported by undeclared output + // annotations. + // The file IDs must be unique within this list. Duplicate file IDs will + // result in an error. Files will be returned in lexicographical order by ID. + Files []*File `protobuf:"bytes,9,rep,name=files,proto3" json:"files,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestCase) Reset() { *m = TestCase{} } +func (m *TestCase) String() string { return proto.CompactTextString(m) } +func (*TestCase) ProtoMessage() {} +func (*TestCase) Descriptor() ([]byte, []int) { + return fileDescriptor_test_suite_ee31d75e3e5b9dce, []int{2} +} +func (m *TestCase) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestCase.Unmarshal(m, b) +} +func (m *TestCase) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestCase.Marshal(b, m, deterministic) +} +func (dst *TestCase) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestCase.Merge(dst, src) +} +func (m *TestCase) XXX_Size() int { + return xxx_messageInfo_TestCase.Size(m) +} +func (m *TestCase) XXX_DiscardUnknown() { + xxx_messageInfo_TestCase.DiscardUnknown(m) +} + +var xxx_messageInfo_TestCase proto.InternalMessageInfo + +func (m *TestCase) GetCaseName() string { + if m != nil { + return m.CaseName + } + return "" +} + +func (m *TestCase) GetClassName() string { + if m != nil { + return m.ClassName + } + return "" +} + +func (m *TestCase) GetResult() TestCase_Result { + if m != nil { + return m.Result + } + return TestCase_RESULT_UNSPECIFIED +} + +func (m *TestCase) GetFailures() []*TestFailure { + if m != nil { + return m.Failures + } + return nil +} + +func (m *TestCase) GetErrors() []*TestError { + if m != nil { + return m.Errors + } + return nil +} + +func (m *TestCase) GetTiming() *Timing { + if m != nil { + return m.Timing + } + return nil +} + +func (m *TestCase) GetProperties() []*Property { + if m != nil { + return m.Properties + } + return nil +} + +func (m *TestCase) GetFiles() []*File { + if m != nil { + return m.Files + } + return nil +} + +// Represents a violated assertion, as reported in a element within a +// . Some languages allow assertions to be made without stopping the +// test case when they're violated, leading to multiple TestFailures. For Java, +// multiple TestFailures are used to represent a chained exception. +type TestFailure struct { + // The exception message reported in the message attribute. Typically short, + // but may be multi-line. Eg. "Expected 'foo' but was 'bar'". + FailureMessage string `protobuf:"bytes,1,opt,name=failure_message,json=failureMessage,proto3" json:"failure_message,omitempty"` + // The type of the exception being thrown, reported in the type attribute. + // Eg: "org.junit.ComparisonFailure" + ExceptionType string `protobuf:"bytes,2,opt,name=exception_type,json=exceptionType,proto3" json:"exception_type,omitempty"` + // The stack trace reported as the content of the element, often in + // a CDATA block. This contains one line for each stack frame, each including + // a method/function name, a class/file name, and a line number. Most recent + // call is usually first, but not for Python stack traces. May contain the + // exception_type and message. + StackTrace string `protobuf:"bytes,3,opt,name=stack_trace,json=stackTrace,proto3" json:"stack_trace,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestFailure) Reset() { *m = TestFailure{} } +func (m *TestFailure) String() string { return proto.CompactTextString(m) } +func (*TestFailure) ProtoMessage() {} +func (*TestFailure) Descriptor() ([]byte, []int) { + return fileDescriptor_test_suite_ee31d75e3e5b9dce, []int{3} +} +func (m *TestFailure) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestFailure.Unmarshal(m, b) +} +func (m *TestFailure) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestFailure.Marshal(b, m, deterministic) +} +func (dst *TestFailure) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestFailure.Merge(dst, src) +} +func (m *TestFailure) XXX_Size() int { + return xxx_messageInfo_TestFailure.Size(m) +} +func (m *TestFailure) XXX_DiscardUnknown() { + xxx_messageInfo_TestFailure.DiscardUnknown(m) +} + +var xxx_messageInfo_TestFailure proto.InternalMessageInfo + +func (m *TestFailure) GetFailureMessage() string { + if m != nil { + return m.FailureMessage + } + return "" +} + +func (m *TestFailure) GetExceptionType() string { + if m != nil { + return m.ExceptionType + } + return "" +} + +func (m *TestFailure) GetStackTrace() string { + if m != nil { + return m.StackTrace + } + return "" +} + +// Represents an exception that prevented a test case from completing, as +// reported in an element within a . For Java, multiple +// TestErrors are used to represent a chained exception. +type TestError struct { + // The exception message, as reported in the message attribute. Typically + // short, but may be multi-line. Eg. "argument cannot be null". + ErrorMessage string `protobuf:"bytes,1,opt,name=error_message,json=errorMessage,proto3" json:"error_message,omitempty"` + // The type of the exception being thrown, reported in the type attribute. + // For Java, this is a fully qualified Throwable class name. + // Eg: "java.lang.IllegalArgumentException" + ExceptionType string `protobuf:"bytes,2,opt,name=exception_type,json=exceptionType,proto3" json:"exception_type,omitempty"` + // The stack trace reported as the content of the element, often in + // a CDATA block. This contains one line for each stack frame, each including + // a method/function name, a class/file name, and a line number. Most recent + // call is usually first, but not for Python stack traces. May contain the + // exception_type and message. + StackTrace string `protobuf:"bytes,3,opt,name=stack_trace,json=stackTrace,proto3" json:"stack_trace,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *TestError) Reset() { *m = TestError{} } +func (m *TestError) String() string { return proto.CompactTextString(m) } +func (*TestError) ProtoMessage() {} +func (*TestError) Descriptor() ([]byte, []int) { + return fileDescriptor_test_suite_ee31d75e3e5b9dce, []int{4} +} +func (m *TestError) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_TestError.Unmarshal(m, b) +} +func (m *TestError) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_TestError.Marshal(b, m, deterministic) +} +func (dst *TestError) XXX_Merge(src proto.Message) { + xxx_messageInfo_TestError.Merge(dst, src) +} +func (m *TestError) XXX_Size() int { + return xxx_messageInfo_TestError.Size(m) +} +func (m *TestError) XXX_DiscardUnknown() { + xxx_messageInfo_TestError.DiscardUnknown(m) +} + +var xxx_messageInfo_TestError proto.InternalMessageInfo + +func (m *TestError) GetErrorMessage() string { + if m != nil { + return m.ErrorMessage + } + return "" +} + +func (m *TestError) GetExceptionType() string { + if m != nil { + return m.ExceptionType + } + return "" +} + +func (m *TestError) GetStackTrace() string { + if m != nil { + return m.StackTrace + } + return "" +} + +func init() { + proto.RegisterType((*TestSuite)(nil), "google.devtools.resultstore.v2.TestSuite") + proto.RegisterType((*Test)(nil), "google.devtools.resultstore.v2.Test") + proto.RegisterType((*TestCase)(nil), "google.devtools.resultstore.v2.TestCase") + proto.RegisterType((*TestFailure)(nil), "google.devtools.resultstore.v2.TestFailure") + proto.RegisterType((*TestError)(nil), "google.devtools.resultstore.v2.TestError") + proto.RegisterEnum("google.devtools.resultstore.v2.TestCase_Result", TestCase_Result_name, TestCase_Result_value) +} + +func init() { + proto.RegisterFile("google/devtools/resultstore/v2/test_suite.proto", fileDescriptor_test_suite_ee31d75e3e5b9dce) +} + +var fileDescriptor_test_suite_ee31d75e3e5b9dce = []byte{ + // 641 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x95, 0xdf, 0x6e, 0xd3, 0x30, + 0x14, 0xc6, 0x97, 0xb6, 0xcb, 0x92, 0x93, 0xad, 0xab, 0x7c, 0x81, 0xa2, 0xa1, 0x41, 0x55, 0xfe, + 0x75, 0x9a, 0x94, 0x48, 0xe5, 0x0e, 0x24, 0xa4, 0xad, 0x75, 0xb7, 0x42, 0x57, 0x22, 0x27, 0xbd, + 0xe1, 0xa6, 0x0a, 0xc1, 0x8b, 0x22, 0x92, 0x3a, 0xc4, 0xde, 0xb4, 0x71, 0xc1, 0xf3, 0xf0, 0x08, + 0xbc, 0x0c, 0xef, 0x82, 0x62, 0xa7, 0xa3, 0x20, 0x41, 0x8b, 0xe8, 0x5d, 0xf3, 0xd9, 0xbf, 0x73, + 0x3e, 0xfb, 0x7c, 0xaa, 0xc1, 0x8d, 0x19, 0x8b, 0x53, 0xea, 0x7e, 0xa0, 0xd7, 0x82, 0xb1, 0x94, + 0xbb, 0x05, 0xe5, 0x57, 0xa9, 0xe0, 0x82, 0x15, 0xd4, 0xbd, 0xee, 0xb9, 0x82, 0x72, 0x31, 0xe3, + 0x57, 0x89, 0xa0, 0x4e, 0x5e, 0x30, 0xc1, 0xd0, 0x03, 0x05, 0x38, 0x0b, 0xc0, 0x59, 0x02, 0x9c, + 0xeb, 0xde, 0xc1, 0xf1, 0x8a, 0x82, 0x11, 0xcb, 0x32, 0x36, 0x57, 0xc5, 0x0e, 0x8e, 0x56, 0x6c, + 0xbe, 0x4c, 0xd2, 0xaa, 0x6f, 0xe7, 0x5b, 0x1d, 0xcc, 0x80, 0x72, 0xe1, 0x97, 0x5e, 0xd0, 0x21, + 0x80, 0x34, 0x35, 0x9b, 0x87, 0x19, 0xb5, 0xb5, 0xb6, 0xd6, 0x35, 0x89, 0x29, 0x95, 0x49, 0x98, + 0x51, 0xf4, 0x02, 0xb6, 0x4b, 0xe3, 0xdc, 0xae, 0xb5, 0xeb, 0x5d, 0xab, 0xf7, 0xd8, 0xf9, 0xbb, + 0x69, 0xa7, 0x2c, 0x4c, 0x14, 0x82, 0xce, 0xc0, 0xb8, 0x0c, 0x93, 0xf4, 0xaa, 0xa0, 0xdc, 0xae, + 0x4b, 0xfc, 0x78, 0x1d, 0x7c, 0xa8, 0x18, 0x72, 0x07, 0xa3, 0x13, 0xd0, 0x69, 0x51, 0xb0, 0x82, + 0xdb, 0x0d, 0x59, 0xe6, 0x68, 0x9d, 0x32, 0xb8, 0x24, 0x48, 0x05, 0xa2, 0x57, 0xa0, 0x8b, 0x24, + 0x4b, 0xe6, 0xb1, 0xad, 0xb7, 0xb5, 0xae, 0xd5, 0x7b, 0xba, 0xb2, 0x84, 0xdc, 0x4d, 0x2a, 0x0a, + 0x9d, 0x03, 0xe4, 0x05, 0xcb, 0x69, 0x21, 0x12, 0xca, 0xed, 0x1d, 0x69, 0xa3, 0xbb, 0xaa, 0x86, + 0xa7, 0x88, 0x5b, 0xb2, 0xc4, 0x96, 0x37, 0x5a, 0x0e, 0x83, 0xdb, 0xc6, 0x7a, 0x37, 0x3a, 0x4c, + 0x52, 0x4a, 0x14, 0xd2, 0xf9, 0xaa, 0x41, 0xa3, 0x3c, 0x1b, 0x3a, 0x03, 0x53, 0xe6, 0x29, 0x0a, + 0xb9, 0x1a, 0xda, 0x1a, 0x6e, 0x4a, 0xb0, 0x1f, 0x72, 0x7a, 0xbe, 0x45, 0x0c, 0x51, 0xfd, 0x46, + 0xaf, 0x01, 0x7e, 0x06, 0xd3, 0xae, 0xc9, 0x4a, 0x6b, 0x5d, 0xaf, 0x4c, 0xcf, 0xf9, 0x16, 0x91, + 0x3e, 0xe4, 0xc7, 0xa9, 0x55, 0x99, 0x12, 0xb7, 0x39, 0xed, 0x7c, 0x6f, 0x80, 0xb1, 0xe8, 0x88, + 0xee, 0x83, 0x59, 0x3a, 0x5d, 0xce, 0x98, 0x51, 0x0a, 0x32, 0x62, 0x87, 0x00, 0x51, 0x1a, 0x72, + 0xae, 0x56, 0x6b, 0x2a, 0x81, 0x52, 0x91, 0xcb, 0x67, 0xa0, 0xab, 0xf6, 0x76, 0xbd, 0xad, 0x75, + 0x9b, 0x3d, 0x77, 0xdd, 0x73, 0x3a, 0x44, 0xea, 0xa4, 0xc2, 0x7f, 0x89, 0x63, 0x63, 0x33, 0x71, + 0xdc, 0xfe, 0xff, 0x38, 0xee, 0x6c, 0x20, 0x8e, 0xc6, 0x26, 0xe2, 0x68, 0xfe, 0x7b, 0x1c, 0x3f, + 0x83, 0xae, 0xee, 0x18, 0xdd, 0x03, 0x44, 0xb0, 0x3f, 0x1d, 0x07, 0xb3, 0xe9, 0xc4, 0xf7, 0x70, + 0x7f, 0x34, 0x1c, 0xe1, 0x41, 0x6b, 0x0b, 0xed, 0x81, 0xd9, 0x7f, 0x7b, 0xe1, 0x8d, 0x71, 0x80, + 0x07, 0x2d, 0x0d, 0xed, 0x83, 0x35, 0x9a, 0x04, 0x98, 0x90, 0xa9, 0x57, 0x0a, 0x35, 0xb9, 0x7e, + 0x32, 0xe9, 0xe3, 0xf1, 0x18, 0x0f, 0x5a, 0x75, 0xb4, 0x0b, 0xc6, 0x70, 0x34, 0x0e, 0x30, 0xc1, + 0x83, 0x56, 0x03, 0x59, 0xb0, 0xe3, 0xbf, 0x19, 0x79, 0x1e, 0x1e, 0xb4, 0xb6, 0x51, 0x13, 0xc0, + 0x9f, 0x7a, 0x1e, 0xc1, 0xbe, 0x8f, 0x07, 0x2d, 0xbd, 0xf3, 0x05, 0xac, 0xa5, 0xe9, 0xa0, 0x67, + 0xb0, 0x5f, 0xcd, 0x67, 0x96, 0x51, 0xce, 0xc3, 0x78, 0x91, 0xb3, 0x66, 0x25, 0x5f, 0x28, 0x15, + 0x3d, 0x81, 0x26, 0xbd, 0x89, 0x68, 0x2e, 0x12, 0x36, 0x97, 0x49, 0xad, 0x12, 0xb7, 0x77, 0xa7, + 0x06, 0xb7, 0x39, 0x45, 0x0f, 0xc1, 0xe2, 0x22, 0x8c, 0x3e, 0xce, 0x44, 0x11, 0x46, 0x54, 0x46, + 0xcf, 0x24, 0x20, 0xa5, 0xa0, 0x54, 0x3a, 0x37, 0xea, 0x4f, 0x54, 0x8e, 0x15, 0x3d, 0x82, 0x3d, + 0x39, 0xd8, 0xdf, 0x7a, 0xef, 0x4a, 0x71, 0xc3, 0x9d, 0x4f, 0x3f, 0x41, 0x27, 0x62, 0xd9, 0x8a, + 0x39, 0x79, 0xda, 0xbb, 0x51, 0xb5, 0x23, 0x66, 0x69, 0x38, 0x8f, 0x1d, 0x56, 0xc4, 0x6e, 0x4c, + 0xe7, 0xf2, 0x0d, 0xa8, 0xde, 0xaa, 0x30, 0x4f, 0xf8, 0x9f, 0x5e, 0x8c, 0x97, 0x4b, 0x9f, 0xef, + 0x75, 0x49, 0x3d, 0xff, 0x11, 0x00, 0x00, 0xff, 0xff, 0x38, 0x57, 0x5a, 0x20, 0xe4, 0x06, 0x00, + 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/source/v1/source_context.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/source/v1/source_context.pb.go index a3fb2ec0d..da8bccca9 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/source/v1/source_context.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/source/v1/source_context.pb.go @@ -100,16 +100,16 @@ type isSourceContext_Context interface { } type SourceContext_CloudRepo struct { - CloudRepo *CloudRepoSourceContext `protobuf:"bytes,1,opt,name=cloud_repo,json=cloudRepo,oneof"` + CloudRepo *CloudRepoSourceContext `protobuf:"bytes,1,opt,name=cloud_repo,json=cloudRepo,proto3,oneof"` } type SourceContext_CloudWorkspace struct { - CloudWorkspace *CloudWorkspaceSourceContext `protobuf:"bytes,2,opt,name=cloud_workspace,json=cloudWorkspace,oneof"` + CloudWorkspace *CloudWorkspaceSourceContext `protobuf:"bytes,2,opt,name=cloud_workspace,json=cloudWorkspace,proto3,oneof"` } type SourceContext_Gerrit struct { - Gerrit *GerritSourceContext `protobuf:"bytes,3,opt,name=gerrit,oneof"` + Gerrit *GerritSourceContext `protobuf:"bytes,3,opt,name=gerrit,proto3,oneof"` } type SourceContext_Git struct { - Git *GitSourceContext `protobuf:"bytes,6,opt,name=git,oneof"` + Git *GitSourceContext `protobuf:"bytes,6,opt,name=git,proto3,oneof"` } func (*SourceContext_CloudRepo) isSourceContext_Context() {} @@ -268,9 +268,9 @@ func _SourceContext_OneofSizer(msg proto.Message) (n int) { // details describing the context. type ExtendedSourceContext struct { // Any source context. - Context *SourceContext `protobuf:"bytes,1,opt,name=context" json:"context,omitempty"` + Context *SourceContext `protobuf:"bytes,1,opt,name=context,proto3" json:"context,omitempty"` // Labels with user defined metadata. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -317,9 +317,9 @@ func (m *ExtendedSourceContext) GetLabels() map[string]string { // An alias to a repo revision. type AliasContext struct { // The alias kind. - Kind AliasContext_Kind `protobuf:"varint,1,opt,name=kind,enum=google.devtools.source.v1.AliasContext_Kind" json:"kind,omitempty"` + Kind AliasContext_Kind `protobuf:"varint,1,opt,name=kind,proto3,enum=google.devtools.source.v1.AliasContext_Kind" json:"kind,omitempty"` // The alias name. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -367,7 +367,7 @@ func (m *AliasContext) GetName() string { // repo (a repo hosted by the Google Cloud Platform). type CloudRepoSourceContext struct { // The ID of the repo. - RepoId *RepoId `protobuf:"bytes,1,opt,name=repo_id,json=repoId" json:"repo_id,omitempty"` + RepoId *RepoId `protobuf:"bytes,1,opt,name=repo_id,json=repoId,proto3" json:"repo_id,omitempty"` // A revision in a cloud repository can be identified by either its revision // ID or its Alias. // @@ -410,13 +410,13 @@ type isCloudRepoSourceContext_Revision interface { } type CloudRepoSourceContext_RevisionId struct { - RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,oneof"` + RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,proto3,oneof"` } type CloudRepoSourceContext_AliasName struct { - AliasName string `protobuf:"bytes,3,opt,name=alias_name,json=aliasName,oneof"` + AliasName string `protobuf:"bytes,3,opt,name=alias_name,json=aliasName,proto3,oneof"` } type CloudRepoSourceContext_AliasContext struct { - AliasContext *AliasContext `protobuf:"bytes,4,opt,name=alias_context,json=aliasContext,oneof"` + AliasContext *AliasContext `protobuf:"bytes,4,opt,name=alias_context,json=aliasContext,proto3,oneof"` } func (*CloudRepoSourceContext_RevisionId) isCloudRepoSourceContext_Revision() {} @@ -546,10 +546,10 @@ func _CloudRepoSourceContext_OneofSizer(msg proto.Message) (n int) { // A CloudWorkspaceSourceContext denotes a workspace at a particular snapshot. type CloudWorkspaceSourceContext struct { // The ID of the workspace. - WorkspaceId *CloudWorkspaceId `protobuf:"bytes,1,opt,name=workspace_id,json=workspaceId" json:"workspace_id,omitempty"` + WorkspaceId *CloudWorkspaceId `protobuf:"bytes,1,opt,name=workspace_id,json=workspaceId,proto3" json:"workspace_id,omitempty"` // The ID of the snapshot. // An empty snapshot_id refers to the most recent snapshot. - SnapshotId string `protobuf:"bytes,2,opt,name=snapshot_id,json=snapshotId" json:"snapshot_id,omitempty"` + SnapshotId string `protobuf:"bytes,2,opt,name=snapshot_id,json=snapshotId,proto3" json:"snapshot_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -596,11 +596,11 @@ func (m *CloudWorkspaceSourceContext) GetSnapshotId() string { // A SourceContext referring to a Gerrit project. type GerritSourceContext struct { // The URI of a running Gerrit instance. - HostUri string `protobuf:"bytes,1,opt,name=host_uri,json=hostUri" json:"host_uri,omitempty"` + HostUri string `protobuf:"bytes,1,opt,name=host_uri,json=hostUri,proto3" json:"host_uri,omitempty"` // The full project name within the host. Projects may be nested, so // "project/subproject" is a valid project name. // The "repo name" is hostURI/project. - GerritProject string `protobuf:"bytes,2,opt,name=gerrit_project,json=gerritProject" json:"gerrit_project,omitempty"` + GerritProject string `protobuf:"bytes,2,opt,name=gerrit_project,json=gerritProject,proto3" json:"gerrit_project,omitempty"` // A revision in a Gerrit project can be identified by either its revision ID // or its alias. // @@ -643,13 +643,13 @@ type isGerritSourceContext_Revision interface { } type GerritSourceContext_RevisionId struct { - RevisionId string `protobuf:"bytes,3,opt,name=revision_id,json=revisionId,oneof"` + RevisionId string `protobuf:"bytes,3,opt,name=revision_id,json=revisionId,proto3,oneof"` } type GerritSourceContext_AliasName struct { - AliasName string `protobuf:"bytes,4,opt,name=alias_name,json=aliasName,oneof"` + AliasName string `protobuf:"bytes,4,opt,name=alias_name,json=aliasName,proto3,oneof"` } type GerritSourceContext_AliasContext struct { - AliasContext *AliasContext `protobuf:"bytes,5,opt,name=alias_context,json=aliasContext,oneof"` + AliasContext *AliasContext `protobuf:"bytes,5,opt,name=alias_context,json=aliasContext,proto3,oneof"` } func (*GerritSourceContext_RevisionId) isGerritSourceContext_Revision() {} @@ -787,10 +787,10 @@ func _GerritSourceContext_OneofSizer(msg proto.Message) (n int) { // repository (e.g. GitHub). type GitSourceContext struct { // Git repository URL. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // Git commit hash. // required. - RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId" json:"revision_id,omitempty"` + RevisionId string `protobuf:"bytes,2,opt,name=revision_id,json=revisionId,proto3" json:"revision_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -877,10 +877,10 @@ type isRepoId_Id interface { } type RepoId_ProjectRepoId struct { - ProjectRepoId *ProjectRepoId `protobuf:"bytes,1,opt,name=project_repo_id,json=projectRepoId,oneof"` + ProjectRepoId *ProjectRepoId `protobuf:"bytes,1,opt,name=project_repo_id,json=projectRepoId,proto3,oneof"` } type RepoId_Uid struct { - Uid string `protobuf:"bytes,2,opt,name=uid,oneof"` + Uid string `protobuf:"bytes,2,opt,name=uid,proto3,oneof"` } func (*RepoId_ProjectRepoId) isRepoId_Id() {} @@ -981,9 +981,9 @@ func _RepoId_OneofSizer(msg proto.Message) (n int) { // (e.g. winged-cargo-31) and a repo name within that project. type ProjectRepoId struct { // The ID of the project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The name of the repo. Leave empty for the default repo. - RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName" json:"repo_name,omitempty"` + RepoName string `protobuf:"bytes,2,opt,name=repo_name,json=repoName,proto3" json:"repo_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1032,10 +1032,10 @@ func (m *ProjectRepoId) GetRepoName() string { // can be stored before they are committed. type CloudWorkspaceId struct { // The ID of the repo containing the workspace. - RepoId *RepoId `protobuf:"bytes,1,opt,name=repo_id,json=repoId" json:"repo_id,omitempty"` + RepoId *RepoId `protobuf:"bytes,1,opt,name=repo_id,json=repoId,proto3" json:"repo_id,omitempty"` // The unique name of the workspace within the repo. This is the name // chosen by the client in the Source API's CreateWorkspace method. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/devtools/sourcerepo/v1/sourcerepo.pb.go b/vendor/google.golang.org/genproto/googleapis/devtools/sourcerepo/v1/sourcerepo.pb.go index b01fa42ff..326d0a844 100644 --- a/vendor/google.golang.org/genproto/googleapis/devtools/sourcerepo/v1/sourcerepo.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/devtools/sourcerepo/v1/sourcerepo.pb.go @@ -31,16 +31,16 @@ type Repo struct { // Resource name of the repository, of the form // `projects//repos/`. The repo name may contain slashes. // eg, `projects/myproject/repos/name/with/slash` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The disk usage of the repo, in bytes. Read-only field. Size is only // returned by GetRepo. - Size int64 `protobuf:"varint,2,opt,name=size" json:"size,omitempty"` + Size int64 `protobuf:"varint,2,opt,name=size,proto3" json:"size,omitempty"` // URL to clone the repository from Google Cloud Source Repositories. // Read-only field. - Url string `protobuf:"bytes,3,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,3,opt,name=url,proto3" json:"url,omitempty"` // How this repository mirrors a repository managed by another service. // Read-only field. - MirrorConfig *MirrorConfig `protobuf:"bytes,4,opt,name=mirror_config,json=mirrorConfig" json:"mirror_config,omitempty"` + MirrorConfig *MirrorConfig `protobuf:"bytes,4,opt,name=mirror_config,json=mirrorConfig,proto3" json:"mirror_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -102,16 +102,16 @@ func (m *Repo) GetMirrorConfig() *MirrorConfig { // hosting service, for example GitHub or BitBucket. type MirrorConfig struct { // URL of the main repository at the other hosting service. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // ID of the webhook listening to updates to trigger mirroring. // Removing this webhook from the other hosting service will stop // Google Cloud Source Repositories from receiving notifications, // and thereby disabling mirroring. - WebhookId string `protobuf:"bytes,2,opt,name=webhook_id,json=webhookId" json:"webhook_id,omitempty"` + WebhookId string `protobuf:"bytes,2,opt,name=webhook_id,json=webhookId,proto3" json:"webhook_id,omitempty"` // ID of the SSH deploy key at the other hosting service. // Removing this key from the other service would deauthorize // Google Cloud Source Repositories from mirroring. - DeployKeyId string `protobuf:"bytes,3,opt,name=deploy_key_id,json=deployKeyId" json:"deploy_key_id,omitempty"` + DeployKeyId string `protobuf:"bytes,3,opt,name=deploy_key_id,json=deployKeyId,proto3" json:"deploy_key_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -166,7 +166,7 @@ func (m *MirrorConfig) GetDeployKeyId() string { type GetRepoRequest struct { // The name of the requested repository. Values are of the form // `projects//repos/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -207,14 +207,14 @@ func (m *GetRepoRequest) GetName() string { type ListReposRequest struct { // The project ID whose repos should be listed. Values are of the form // `projects/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Maximum number of repositories to return; between 1 and 500. // If not set or zero, defaults to 100 at the server. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Resume listing repositories where a prior ListReposResponse // left off. This is an opaque token that must be obtained from // a recent, prior ListReposResponse's next_page_token field. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -268,11 +268,11 @@ func (m *ListReposRequest) GetPageToken() string { // Response for ListRepos. The size is not set in the returned repositories. type ListReposResponse struct { // The listed repos. - Repos []*Repo `protobuf:"bytes,1,rep,name=repos" json:"repos,omitempty"` + Repos []*Repo `protobuf:"bytes,1,rep,name=repos,proto3" json:"repos,omitempty"` // If non-empty, additional repositories exist within the project. These // can be retrieved by including this value in the next ListReposRequest's // page_token field. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -320,10 +320,10 @@ func (m *ListReposResponse) GetNextPageToken() string { type CreateRepoRequest struct { // The project in which to create the repo. Values are of the form // `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The repo to create. Only name should be set; setting other fields // is an error. The project in the name should match the parent field. - Repo *Repo `protobuf:"bytes,2,opt,name=repo" json:"repo,omitempty"` + Repo *Repo `protobuf:"bytes,2,opt,name=repo,proto3" json:"repo,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -371,7 +371,7 @@ func (m *CreateRepoRequest) GetRepo() *Repo { type DeleteRepoRequest struct { // The name of the repo to delete. Values are of the form // `projects//repos/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -526,8 +526,7 @@ func (c *sourceRepoClient) TestIamPermissions(ctx context.Context, in *v1.TestIa return out, nil } -// Server API for SourceRepo service - +// SourceRepoServer is the server API for SourceRepo service. type SourceRepoServer interface { // Returns all repos belonging to a project. The sizes of the repos are // not set by ListRepos. To get the size of a repo, use GetRepo. diff --git a/vendor/google.golang.org/genproto/googleapis/example/library/v1/library.pb.go b/vendor/google.golang.org/genproto/googleapis/example/library/v1/library.pb.go index 251eabed0..8039d942d 100644 --- a/vendor/google.golang.org/genproto/googleapis/example/library/v1/library.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/example/library/v1/library.pb.go @@ -30,13 +30,13 @@ type Book struct { // The resource name of the book. // Book names have the form `shelves/{shelf_id}/books/{book_id}`. // The name is ignored when creating a book. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the book author. - Author string `protobuf:"bytes,2,opt,name=author" json:"author,omitempty"` + Author string `protobuf:"bytes,2,opt,name=author,proto3" json:"author,omitempty"` // The title of the book. - Title string `protobuf:"bytes,3,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,3,opt,name=title,proto3" json:"title,omitempty"` // Value indicating whether the book has been read. - Read bool `protobuf:"varint,4,opt,name=read" json:"read,omitempty"` + Read bool `protobuf:"varint,4,opt,name=read,proto3" json:"read,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -99,9 +99,9 @@ type Shelf struct { // The resource name of the shelf. // Shelf names have the form `shelves/{shelf_id}`. // The name is ignored when creating a shelf. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The theme of the shelf - Theme string `protobuf:"bytes,2,opt,name=theme" json:"theme,omitempty"` + Theme string `protobuf:"bytes,2,opt,name=theme,proto3" json:"theme,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -148,7 +148,7 @@ func (m *Shelf) GetTheme() string { // Request message for LibraryService.CreateShelf. type CreateShelfRequest struct { // The shelf to create. - Shelf *Shelf `protobuf:"bytes,1,opt,name=shelf" json:"shelf,omitempty"` + Shelf *Shelf `protobuf:"bytes,1,opt,name=shelf,proto3" json:"shelf,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -188,7 +188,7 @@ func (m *CreateShelfRequest) GetShelf() *Shelf { // Request message for LibraryService.GetShelf. type GetShelfRequest struct { // The name of the shelf to retrieve. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -229,12 +229,12 @@ func (m *GetShelfRequest) GetName() string { type ListShelvesRequest struct { // Requested page size. Server may return fewer shelves than requested. // If unspecified, server will pick an appropriate default. - PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A token identifying a page of results the server should return. // Typically, this is the value of // [ListShelvesResponse.next_page_token][google.example.library.v1.ListShelvesResponse.next_page_token] // returned from the previous call to `ListShelves` method. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -281,13 +281,13 @@ func (m *ListShelvesRequest) GetPageToken() string { // Response message for LibraryService.ListShelves. type ListShelvesResponse struct { // The list of shelves. - Shelves []*Shelf `protobuf:"bytes,1,rep,name=shelves" json:"shelves,omitempty"` + Shelves []*Shelf `protobuf:"bytes,1,rep,name=shelves,proto3" json:"shelves,omitempty"` // A token to retrieve next page of results. // Pass this value in the // [ListShelvesRequest.page_token][google.example.library.v1.ListShelvesRequest.page_token] // field in the subsequent call to `ListShelves` method to retrieve the next // page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -334,7 +334,7 @@ func (m *ListShelvesResponse) GetNextPageToken() string { // Request message for LibraryService.DeleteShelf. type DeleteShelfRequest struct { // The name of the shelf to delete. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -375,9 +375,9 @@ func (m *DeleteShelfRequest) GetName() string { // (name) in this merge. type MergeShelvesRequest struct { // The name of the shelf we're adding books to. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the shelf we're removing books from and deleting. - OtherShelfName string `protobuf:"bytes,2,opt,name=other_shelf_name,json=otherShelfName" json:"other_shelf_name,omitempty"` + OtherShelfName string `protobuf:"bytes,2,opt,name=other_shelf_name,json=otherShelfName,proto3" json:"other_shelf_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -424,9 +424,9 @@ func (m *MergeShelvesRequest) GetOtherShelfName() string { // Request message for LibraryService.CreateBook. type CreateBookRequest struct { // The name of the shelf in which the book is created. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The book to create. - Book *Book `protobuf:"bytes,2,opt,name=book" json:"book,omitempty"` + Book *Book `protobuf:"bytes,2,opt,name=book,proto3" json:"book,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -473,7 +473,7 @@ func (m *CreateBookRequest) GetBook() *Book { // Request message for LibraryService.GetBook. type GetBookRequest struct { // The name of the book to retrieve. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -513,15 +513,15 @@ func (m *GetBookRequest) GetName() string { // Request message for LibraryService.ListBooks. type ListBooksRequest struct { // The name of the shelf whose books we'd like to list. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Requested page size. Server may return fewer books than requested. // If unspecified, server will pick an appropriate default. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A token identifying a page of results the server should return. // Typically, this is the value of // [ListBooksResponse.next_page_token][google.example.library.v1.ListBooksResponse.next_page_token]. // returned from the previous call to `ListBooks` method. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -575,13 +575,13 @@ func (m *ListBooksRequest) GetPageToken() string { // Response message for LibraryService.ListBooks. type ListBooksResponse struct { // The list of books. - Books []*Book `protobuf:"bytes,1,rep,name=books" json:"books,omitempty"` + Books []*Book `protobuf:"bytes,1,rep,name=books,proto3" json:"books,omitempty"` // A token to retrieve next page of results. // Pass this value in the // [ListBooksRequest.page_token][google.example.library.v1.ListBooksRequest.page_token] // field in the subsequent call to `ListBooks` method to retrieve the next // page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -628,9 +628,9 @@ func (m *ListBooksResponse) GetNextPageToken() string { // Request message for LibraryService.UpdateBook. type UpdateBookRequest struct { // The name of the book to update. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The book to update with. The name must match or be empty. - Book *Book `protobuf:"bytes,2,opt,name=book" json:"book,omitempty"` + Book *Book `protobuf:"bytes,2,opt,name=book,proto3" json:"book,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -677,7 +677,7 @@ func (m *UpdateBookRequest) GetBook() *Book { // Request message for LibraryService.DeleteBook. type DeleteBookRequest struct { // The name of the book to delete. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -718,9 +718,9 @@ func (m *DeleteBookRequest) GetName() string { // to (other_shelf_name). type MoveBookRequest struct { // The name of the book to move. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the destination shelf. - OtherShelfName string `protobuf:"bytes,2,opt,name=other_shelf_name,json=otherShelfName" json:"other_shelf_name,omitempty"` + OtherShelfName string `protobuf:"bytes,2,opt,name=other_shelf_name,json=otherShelfName,proto3" json:"other_shelf_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -936,8 +936,7 @@ func (c *libraryServiceClient) MoveBook(ctx context.Context, in *MoveBookRequest return out, nil } -// Server API for LibraryService service - +// LibraryServiceServer is the server API for LibraryService service. type LibraryServiceServer interface { // Creates a shelf, and returns the new Shelf. CreateShelf(context.Context, *CreateShelfRequest) (*Shelf, error) diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/firestore_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/firestore_admin.pb.go index 0029c3a8a..09c077a40 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/firestore_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/firestore_admin.pb.go @@ -57,21 +57,21 @@ func (IndexOperationMetadata_OperationType) EnumDescriptor() ([]byte, []int) { // the metadata field of [google.longrunning.Operation][google.longrunning.Operation]. type IndexOperationMetadata struct { // The time that work began on the operation. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time the operation ended, either successfully or otherwise. Unset if // the operation is still active. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // The index resource that this operation is acting on. For example: // `projects/{project_id}/databases/{database_id}/indexes/{index_id}` - Index string `protobuf:"bytes,3,opt,name=index" json:"index,omitempty"` + Index string `protobuf:"bytes,3,opt,name=index,proto3" json:"index,omitempty"` // The type of index operation. - OperationType IndexOperationMetadata_OperationType `protobuf:"varint,4,opt,name=operation_type,json=operationType,enum=google.firestore.admin.v1beta1.IndexOperationMetadata_OperationType" json:"operation_type,omitempty"` + OperationType IndexOperationMetadata_OperationType `protobuf:"varint,4,opt,name=operation_type,json=operationType,proto3,enum=google.firestore.admin.v1beta1.IndexOperationMetadata_OperationType" json:"operation_type,omitempty"` // True if the [google.longrunning.Operation] was cancelled. If the // cancellation is in progress, cancelled will be true but // [google.longrunning.Operation.done][google.longrunning.Operation.done] will be false. - Cancelled bool `protobuf:"varint,5,opt,name=cancelled" json:"cancelled,omitempty"` + Cancelled bool `protobuf:"varint,5,opt,name=cancelled,proto3" json:"cancelled,omitempty"` // Progress of the existing operation, measured in number of documents. - DocumentProgress *Progress `protobuf:"bytes,6,opt,name=document_progress,json=documentProgress" json:"document_progress,omitempty"` + DocumentProgress *Progress `protobuf:"bytes,6,opt,name=document_progress,json=documentProgress,proto3" json:"document_progress,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -147,10 +147,10 @@ func (m *IndexOperationMetadata) GetDocumentProgress() *Progress { type Progress struct { // An estimate of how much work has been completed. Note that this may be // greater than `work_estimated`. - WorkCompleted int64 `protobuf:"varint,1,opt,name=work_completed,json=workCompleted" json:"work_completed,omitempty"` + WorkCompleted int64 `protobuf:"varint,1,opt,name=work_completed,json=workCompleted,proto3" json:"work_completed,omitempty"` // An estimate of how much work needs to be performed. Zero if the // work estimate is unavailable. May change as work progresses. - WorkEstimated int64 `protobuf:"varint,2,opt,name=work_estimated,json=workEstimated" json:"work_estimated,omitempty"` + WorkEstimated int64 `protobuf:"varint,2,opt,name=work_estimated,json=workEstimated,proto3" json:"work_estimated,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -198,10 +198,10 @@ func (m *Progress) GetWorkEstimated() int64 { type CreateIndexRequest struct { // The name of the database this index will apply to. For example: // `projects/{project_id}/databases/{database_id}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The index to create. The name and state fields are output only and will be // ignored. Certain single field indexes cannot be created or deleted. - Index *Index `protobuf:"bytes,2,opt,name=index" json:"index,omitempty"` + Index *Index `protobuf:"bytes,2,opt,name=index,proto3" json:"index,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -249,7 +249,7 @@ func (m *CreateIndexRequest) GetIndex() *Index { type GetIndexRequest struct { // The name of the index. For example: // `projects/{project_id}/databases/{database_id}/indexes/{index_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -290,12 +290,12 @@ func (m *GetIndexRequest) GetName() string { type ListIndexesRequest struct { // The database name. For example: // `projects/{project_id}/databases/{database_id}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // The standard List page size. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The standard List page token. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -357,7 +357,7 @@ func (m *ListIndexesRequest) GetPageToken() string { type DeleteIndexRequest struct { // The index name. For example: // `projects/{project_id}/databases/{database_id}/indexes/{index_id}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -397,9 +397,9 @@ func (m *DeleteIndexRequest) GetName() string { // The response for [FirestoreAdmin.ListIndexes][google.firestore.admin.v1beta1.FirestoreAdmin.ListIndexes]. type ListIndexesResponse struct { // The indexes. - Indexes []*Index `protobuf:"bytes,1,rep,name=indexes" json:"indexes,omitempty"` + Indexes []*Index `protobuf:"bytes,1,rep,name=indexes,proto3" json:"indexes,omitempty"` // The standard List next-page token. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -532,8 +532,7 @@ func (c *firestoreAdminClient) DeleteIndex(ctx context.Context, in *DeleteIndexR return out, nil } -// Server API for FirestoreAdmin service - +// FirestoreAdminServer is the server API for FirestoreAdmin service. type FirestoreAdminServer interface { // Creates the specified index. // A newly created index's initial state is `CREATING`. On completion of the diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/index.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/index.pb.go index d10a33445..64d16ec0e 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/index.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/admin/v1beta1/index.pb.go @@ -103,9 +103,9 @@ type IndexField struct { // by [google.firestore.v1beta1.Document.fields][fields]. // Special field path `__name__` may be used by itself or at the end of a // path. `__type__` may be used only at the end of path. - FieldPath string `protobuf:"bytes,1,opt,name=field_path,json=fieldPath" json:"field_path,omitempty"` + FieldPath string `protobuf:"bytes,1,opt,name=field_path,json=fieldPath,proto3" json:"field_path,omitempty"` // The field's mode. - Mode IndexField_Mode `protobuf:"varint,2,opt,name=mode,enum=google.firestore.admin.v1beta1.IndexField_Mode" json:"mode,omitempty"` + Mode IndexField_Mode `protobuf:"varint,2,opt,name=mode,proto3,enum=google.firestore.admin.v1beta1.IndexField_Mode" json:"mode,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -153,14 +153,14 @@ func (m *IndexField) GetMode() IndexField_Mode { type Index struct { // The resource name of the index. // Output only. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The collection ID to which this index applies. Required. - CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId" json:"collection_id,omitempty"` + CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId,proto3" json:"collection_id,omitempty"` // The fields to index. - Fields []*IndexField `protobuf:"bytes,3,rep,name=fields" json:"fields,omitempty"` + Fields []*IndexField `protobuf:"bytes,3,rep,name=fields,proto3" json:"fields,omitempty"` // The state of the index. // Output only. - State Index_State `protobuf:"varint,6,opt,name=state,enum=google.firestore.admin.v1beta1.Index_State" json:"state,omitempty"` + State Index_State `protobuf:"varint,6,opt,name=state,proto3,enum=google.firestore.admin.v1beta1.Index_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/common.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/common.pb.go index 27eecc261..ccf88690f 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/common.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/common.pb.go @@ -28,7 +28,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type DocumentMask struct { // The list of field paths in the mask. See [Document.fields][google.firestore.v1beta1.Document.fields] for a field // path syntax reference. - FieldPaths []string `protobuf:"bytes,1,rep,name=field_paths,json=fieldPaths" json:"field_paths,omitempty"` + FieldPaths []string `protobuf:"bytes,1,rep,name=field_paths,json=fieldPaths,proto3" json:"field_paths,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -107,10 +107,10 @@ type isPrecondition_ConditionType interface { } type Precondition_Exists struct { - Exists bool `protobuf:"varint,1,opt,name=exists,oneof"` + Exists bool `protobuf:"varint,1,opt,name=exists,proto3,oneof"` } type Precondition_UpdateTime struct { - UpdateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=update_time,json=updateTime,oneof"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=update_time,json=updateTime,proto3,oneof"` } func (*Precondition_Exists) isPrecondition_ConditionType() {} @@ -252,10 +252,10 @@ type isTransactionOptions_Mode interface { } type TransactionOptions_ReadOnly_ struct { - ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,oneof"` + ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,proto3,oneof"` } type TransactionOptions_ReadWrite_ struct { - ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,3,opt,name=read_write,json=readWrite,oneof"` + ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,3,opt,name=read_write,json=readWrite,proto3,oneof"` } func (*TransactionOptions_ReadOnly_) isTransactionOptions_Mode() {} @@ -438,7 +438,7 @@ type isTransactionOptions_ReadOnly_ConsistencySelector interface { } type TransactionOptions_ReadOnly_ReadTime struct { - ReadTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=read_time,json=readTime,oneof"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=read_time,json=readTime,proto3,oneof"` } func (*TransactionOptions_ReadOnly_ReadTime) isTransactionOptions_ReadOnly_ConsistencySelector() {} diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/document.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/document.pb.go index 5fbb4a76f..e9bde698c 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/document.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/document.pb.go @@ -28,7 +28,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Document struct { // The resource name of the document, for example // `projects/{project_id}/databases/{database_id}/documents/{document_path}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The document's fields. // // The map keys represent field names. @@ -53,19 +53,19 @@ type Document struct { // may contain any character. Some characters, including `` ` ``, must be // escaped using a `\`. For example, `` `x&y` `` represents `x&y` and // `` `bak\`tik` `` represents `` bak`tik ``. - Fields map[string]*Value `protobuf:"bytes,2,rep,name=fields" json:"fields,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Fields map[string]*Value `protobuf:"bytes,2,rep,name=fields,proto3" json:"fields,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Output only. The time at which the document was created. // // This value increases monotonically when a document is deleted then // recreated. It can also be compared to values from other documents and // the `read_time` of a query. - CreateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The time at which the document was last changed. // // This value is initally set to the `create_time` then increases // monotonically with each change to the document. It can also be // compared to values from other documents and the `read_time` of a query. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -174,37 +174,37 @@ type isValue_ValueType interface { } type Value_NullValue struct { - NullValue _struct.NullValue `protobuf:"varint,11,opt,name=null_value,json=nullValue,enum=google.protobuf.NullValue,oneof"` + NullValue _struct.NullValue `protobuf:"varint,11,opt,name=null_value,json=nullValue,proto3,enum=google.protobuf.NullValue,oneof"` } type Value_BooleanValue struct { - BooleanValue bool `protobuf:"varint,1,opt,name=boolean_value,json=booleanValue,oneof"` + BooleanValue bool `protobuf:"varint,1,opt,name=boolean_value,json=booleanValue,proto3,oneof"` } type Value_IntegerValue struct { - IntegerValue int64 `protobuf:"varint,2,opt,name=integer_value,json=integerValue,oneof"` + IntegerValue int64 `protobuf:"varint,2,opt,name=integer_value,json=integerValue,proto3,oneof"` } type Value_DoubleValue struct { - DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,oneof"` + DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,proto3,oneof"` } type Value_TimestampValue struct { - TimestampValue *timestamp.Timestamp `protobuf:"bytes,10,opt,name=timestamp_value,json=timestampValue,oneof"` + TimestampValue *timestamp.Timestamp `protobuf:"bytes,10,opt,name=timestamp_value,json=timestampValue,proto3,oneof"` } type Value_StringValue struct { - StringValue string `protobuf:"bytes,17,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,17,opt,name=string_value,json=stringValue,proto3,oneof"` } type Value_BytesValue struct { BytesValue []byte `protobuf:"bytes,18,opt,name=bytes_value,json=bytesValue,proto3,oneof"` } type Value_ReferenceValue struct { - ReferenceValue string `protobuf:"bytes,5,opt,name=reference_value,json=referenceValue,oneof"` + ReferenceValue string `protobuf:"bytes,5,opt,name=reference_value,json=referenceValue,proto3,oneof"` } type Value_GeoPointValue struct { - GeoPointValue *latlng.LatLng `protobuf:"bytes,8,opt,name=geo_point_value,json=geoPointValue,oneof"` + GeoPointValue *latlng.LatLng `protobuf:"bytes,8,opt,name=geo_point_value,json=geoPointValue,proto3,oneof"` } type Value_ArrayValue struct { - ArrayValue *ArrayValue `protobuf:"bytes,9,opt,name=array_value,json=arrayValue,oneof"` + ArrayValue *ArrayValue `protobuf:"bytes,9,opt,name=array_value,json=arrayValue,proto3,oneof"` } type Value_MapValue struct { - MapValue *MapValue `protobuf:"bytes,6,opt,name=map_value,json=mapValue,oneof"` + MapValue *MapValue `protobuf:"bytes,6,opt,name=map_value,json=mapValue,proto3,oneof"` } func (*Value_NullValue) isValue_ValueType() {} @@ -523,7 +523,7 @@ func _Value_OneofSizer(msg proto.Message) (n int) { // An array value. type ArrayValue struct { // Values in the array. - Values []*Value `protobuf:"bytes,1,rep,name=values" json:"values,omitempty"` + Values []*Value `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -568,7 +568,7 @@ type MapValue struct { // expression `__.*__` are reserved. Reserved field names are forbidden except // in certain documented contexts. The map keys, represented as UTF-8, must // not exceed 1,500 bytes and cannot be empty. - Fields map[string]*Value `protobuf:"bytes,1,rep,name=fields" json:"fields,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Fields map[string]*Value `protobuf:"bytes,1,rep,name=fields,proto3" json:"fields,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/firestore.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/firestore.pb.go index 181265400..c953f05ee 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/firestore.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/firestore.pb.go @@ -80,12 +80,12 @@ func (TargetChange_TargetChangeType) EnumDescriptor() ([]byte, []int) { type GetDocumentRequest struct { // The resource name of the Document to get. In the format: // `projects/{project_id}/databases/{database_id}/documents/{document_path}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The fields to return. If not set, returns all fields. // // If the document has a field that is not present in this mask, that field // will not be returned in the response. - Mask *DocumentMask `protobuf:"bytes,2,opt,name=mask" json:"mask,omitempty"` + Mask *DocumentMask `protobuf:"bytes,2,opt,name=mask,proto3" json:"mask,omitempty"` // The consistency mode for this transaction. // If not set, defaults to strong consistency. // @@ -130,7 +130,7 @@ type GetDocumentRequest_Transaction struct { Transaction []byte `protobuf:"bytes,3,opt,name=transaction,proto3,oneof"` } type GetDocumentRequest_ReadTime struct { - ReadTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=read_time,json=readTime,oneof"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=read_time,json=readTime,proto3,oneof"` } func (*GetDocumentRequest_Transaction) isGetDocumentRequest_ConsistencySelector() {} @@ -249,21 +249,21 @@ type ListDocumentsRequest struct { // For example: // `projects/my-project/databases/my-database/documents` or // `projects/my-project/databases/my-database/documents/chatrooms/my-chatroom` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The collection ID, relative to `parent`, to list. For example: `chatrooms` // or `messages`. - CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId" json:"collection_id,omitempty"` + CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId,proto3" json:"collection_id,omitempty"` // The maximum number of documents to return. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The `next_page_token` value returned from a previous List request, if any. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The order to sort results by. For example: `priority desc, name`. - OrderBy string `protobuf:"bytes,6,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy string `protobuf:"bytes,6,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` // The fields to return. If not set, returns all fields. // // If a document has a field that is not present in this mask, that field // will not be returned in the response. - Mask *DocumentMask `protobuf:"bytes,7,opt,name=mask" json:"mask,omitempty"` + Mask *DocumentMask `protobuf:"bytes,7,opt,name=mask,proto3" json:"mask,omitempty"` // The consistency mode for this transaction. // If not set, defaults to strong consistency. // @@ -278,7 +278,7 @@ type ListDocumentsRequest struct { // // Requests with `show_missing` may not specify `where` or // `order_by`. - ShowMissing bool `protobuf:"varint,12,opt,name=show_missing,json=showMissing" json:"show_missing,omitempty"` + ShowMissing bool `protobuf:"varint,12,opt,name=show_missing,json=showMissing,proto3" json:"show_missing,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -316,7 +316,7 @@ type ListDocumentsRequest_Transaction struct { Transaction []byte `protobuf:"bytes,8,opt,name=transaction,proto3,oneof"` } type ListDocumentsRequest_ReadTime struct { - ReadTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=read_time,json=readTime,oneof"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=read_time,json=readTime,proto3,oneof"` } func (*ListDocumentsRequest_Transaction) isListDocumentsRequest_ConsistencySelector() {} @@ -465,9 +465,9 @@ func _ListDocumentsRequest_OneofSizer(msg proto.Message) (n int) { // The response for [Firestore.ListDocuments][google.firestore.v1beta1.Firestore.ListDocuments]. type ListDocumentsResponse struct { // The Documents found. - Documents []*Document `protobuf:"bytes,1,rep,name=documents" json:"documents,omitempty"` + Documents []*Document `protobuf:"bytes,1,rep,name=documents,proto3" json:"documents,omitempty"` // The next page token. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -516,20 +516,20 @@ type CreateDocumentRequest struct { // The parent resource. For example: // `projects/{project_id}/databases/{database_id}/documents` or // `projects/{project_id}/databases/{database_id}/documents/chatrooms/{chatroom_id}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The collection ID, relative to `parent`, to list. For example: `chatrooms`. - CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId" json:"collection_id,omitempty"` + CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId,proto3" json:"collection_id,omitempty"` // The client-assigned document ID to use for this document. // // Optional. If not specified, an ID will be assigned by the service. - DocumentId string `protobuf:"bytes,3,opt,name=document_id,json=documentId" json:"document_id,omitempty"` + DocumentId string `protobuf:"bytes,3,opt,name=document_id,json=documentId,proto3" json:"document_id,omitempty"` // The document to create. `name` must not be set. - Document *Document `protobuf:"bytes,4,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,4,opt,name=document,proto3" json:"document,omitempty"` // The fields to return. If not set, returns all fields. // // If the document has a field that is not present in this mask, that field // will not be returned in the response. - Mask *DocumentMask `protobuf:"bytes,5,opt,name=mask" json:"mask,omitempty"` + Mask *DocumentMask `protobuf:"bytes,5,opt,name=mask,proto3" json:"mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -598,7 +598,7 @@ func (m *CreateDocumentRequest) GetMask() *DocumentMask { type UpdateDocumentRequest struct { // The updated document. // Creates the document if it does not already exist. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The fields to update. // None of the field paths in the mask may contain a reserved name. // @@ -606,15 +606,15 @@ type UpdateDocumentRequest struct { // mask, they are left unchanged. // Fields referenced in the mask, but not present in the input document, are // deleted from the document on the server. - UpdateMask *DocumentMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *DocumentMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // The fields to return. If not set, returns all fields. // // If the document has a field that is not present in this mask, that field // will not be returned in the response. - Mask *DocumentMask `protobuf:"bytes,3,opt,name=mask" json:"mask,omitempty"` + Mask *DocumentMask `protobuf:"bytes,3,opt,name=mask,proto3" json:"mask,omitempty"` // An optional precondition on the document. // The request will fail if this is set and not met by the target document. - CurrentDocument *Precondition `protobuf:"bytes,4,opt,name=current_document,json=currentDocument" json:"current_document,omitempty"` + CurrentDocument *Precondition `protobuf:"bytes,4,opt,name=current_document,json=currentDocument,proto3" json:"current_document,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -676,10 +676,10 @@ func (m *UpdateDocumentRequest) GetCurrentDocument() *Precondition { type DeleteDocumentRequest struct { // The resource name of the Document to delete. In the format: // `projects/{project_id}/databases/{database_id}/documents/{document_path}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // An optional precondition on the document. // The request will fail if this is set and not met by the target document. - CurrentDocument *Precondition `protobuf:"bytes,2,opt,name=current_document,json=currentDocument" json:"current_document,omitempty"` + CurrentDocument *Precondition `protobuf:"bytes,2,opt,name=current_document,json=currentDocument,proto3" json:"current_document,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -727,17 +727,17 @@ func (m *DeleteDocumentRequest) GetCurrentDocument() *Precondition { type BatchGetDocumentsRequest struct { // The database name. In the format: // `projects/{project_id}/databases/{database_id}`. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The names of the documents to retrieve. In the format: // `projects/{project_id}/databases/{database_id}/documents/{document_path}`. // The request will fail if any of the document is not a child resource of the // given `database`. Duplicate names will be elided. - Documents []string `protobuf:"bytes,2,rep,name=documents" json:"documents,omitempty"` + Documents []string `protobuf:"bytes,2,rep,name=documents,proto3" json:"documents,omitempty"` // The fields to return. If not set, returns all fields. // // If a document has a field that is not present in this mask, that field will // not be returned in the response. - Mask *DocumentMask `protobuf:"bytes,3,opt,name=mask" json:"mask,omitempty"` + Mask *DocumentMask `protobuf:"bytes,3,opt,name=mask,proto3" json:"mask,omitempty"` // The consistency mode for this transaction. // If not set, defaults to strong consistency. // @@ -783,10 +783,10 @@ type BatchGetDocumentsRequest_Transaction struct { Transaction []byte `protobuf:"bytes,4,opt,name=transaction,proto3,oneof"` } type BatchGetDocumentsRequest_NewTransaction struct { - NewTransaction *TransactionOptions `protobuf:"bytes,5,opt,name=new_transaction,json=newTransaction,oneof"` + NewTransaction *TransactionOptions `protobuf:"bytes,5,opt,name=new_transaction,json=newTransaction,proto3,oneof"` } type BatchGetDocumentsRequest_ReadTime struct { - ReadTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=read_time,json=readTime,oneof"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=read_time,json=readTime,proto3,oneof"` } func (*BatchGetDocumentsRequest_Transaction) isBatchGetDocumentsRequest_ConsistencySelector() {} @@ -948,7 +948,7 @@ type BatchGetDocumentsResponse struct { // This may be monotically increasing, in this case the previous documents in // the result stream are guaranteed not to have changed between their // read_time and this one. - ReadTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_time,json=readTime" json:"read_time,omitempty"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_time,json=readTime,proto3" json:"read_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -983,10 +983,10 @@ type isBatchGetDocumentsResponse_Result interface { } type BatchGetDocumentsResponse_Found struct { - Found *Document `protobuf:"bytes,1,opt,name=found,oneof"` + Found *Document `protobuf:"bytes,1,opt,name=found,proto3,oneof"` } type BatchGetDocumentsResponse_Missing struct { - Missing string `protobuf:"bytes,2,opt,name=missing,oneof"` + Missing string `protobuf:"bytes,2,opt,name=missing,proto3,oneof"` } func (*BatchGetDocumentsResponse_Found) isBatchGetDocumentsResponse_Result() {} @@ -1101,10 +1101,10 @@ func _BatchGetDocumentsResponse_OneofSizer(msg proto.Message) (n int) { type BeginTransactionRequest struct { // The database name. In the format: // `projects/{project_id}/databases/{database_id}`. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The options for the transaction. // Defaults to a read-write transaction. - Options *TransactionOptions `protobuf:"bytes,2,opt,name=options" json:"options,omitempty"` + Options *TransactionOptions `protobuf:"bytes,2,opt,name=options,proto3" json:"options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1192,11 +1192,11 @@ func (m *BeginTransactionResponse) GetTransaction() []byte { type CommitRequest struct { // The database name. In the format: // `projects/{project_id}/databases/{database_id}`. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The writes to apply. // // Always executed atomically and in order. - Writes []*Write `protobuf:"bytes,2,rep,name=writes" json:"writes,omitempty"` + Writes []*Write `protobuf:"bytes,2,rep,name=writes,proto3" json:"writes,omitempty"` // If set, applies all writes in this transaction, and commits it. Transaction []byte `protobuf:"bytes,3,opt,name=transaction,proto3" json:"transaction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1255,9 +1255,9 @@ type CommitResponse struct { // // This i-th write result corresponds to the i-th write in the // request. - WriteResults []*WriteResult `protobuf:"bytes,1,rep,name=write_results,json=writeResults" json:"write_results,omitempty"` + WriteResults []*WriteResult `protobuf:"bytes,1,rep,name=write_results,json=writeResults,proto3" json:"write_results,omitempty"` // The time at which the commit occurred. - CommitTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=commit_time,json=commitTime" json:"commit_time,omitempty"` + CommitTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=commit_time,json=commitTime,proto3" json:"commit_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1305,7 +1305,7 @@ func (m *CommitResponse) GetCommitTime() *timestamp.Timestamp { type RollbackRequest struct { // The database name. In the format: // `projects/{project_id}/databases/{database_id}`. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The transaction to roll back. Transaction []byte `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1359,7 +1359,7 @@ type RunQueryRequest struct { // For example: // `projects/my-project/databases/my-database/documents` or // `projects/my-project/databases/my-database/documents/chatrooms/my-chatroom` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The query to run. // // Types that are valid to be assigned to QueryType: @@ -1410,16 +1410,16 @@ type isRunQueryRequest_ConsistencySelector interface { } type RunQueryRequest_StructuredQuery struct { - StructuredQuery *StructuredQuery `protobuf:"bytes,2,opt,name=structured_query,json=structuredQuery,oneof"` + StructuredQuery *StructuredQuery `protobuf:"bytes,2,opt,name=structured_query,json=structuredQuery,proto3,oneof"` } type RunQueryRequest_Transaction struct { Transaction []byte `protobuf:"bytes,5,opt,name=transaction,proto3,oneof"` } type RunQueryRequest_NewTransaction struct { - NewTransaction *TransactionOptions `protobuf:"bytes,6,opt,name=new_transaction,json=newTransaction,oneof"` + NewTransaction *TransactionOptions `protobuf:"bytes,6,opt,name=new_transaction,json=newTransaction,proto3,oneof"` } type RunQueryRequest_ReadTime struct { - ReadTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=read_time,json=readTime,oneof"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=read_time,json=readTime,proto3,oneof"` } func (*RunQueryRequest_StructuredQuery) isRunQueryRequest_QueryType() {} @@ -1604,7 +1604,7 @@ type RunQueryResponse struct { Transaction []byte `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` // A query result. // Not set when reporting partial progress. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The time at which the document was read. This may be monotonically // increasing; in this case, the previous documents in the result stream are // guaranteed not to have changed between their `read_time` and this one. @@ -1612,10 +1612,10 @@ type RunQueryResponse struct { // If the query returns no results, a response with `read_time` and no // `document` will be sent, and this represents the time at which the query // was run. - ReadTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=read_time,json=readTime" json:"read_time,omitempty"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=read_time,json=readTime,proto3" json:"read_time,omitempty"` // The number of results that have been skipped due to an offset between // the last response and the current response. - SkippedResults int32 `protobuf:"varint,4,opt,name=skipped_results,json=skippedResults" json:"skipped_results,omitempty"` + SkippedResults int32 `protobuf:"varint,4,opt,name=skipped_results,json=skippedResults,proto3" json:"skipped_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1687,18 +1687,18 @@ type WriteRequest struct { // The database name. In the format: // `projects/{project_id}/databases/{database_id}`. // This is only required in the first message. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The ID of the write stream to resume. // This may only be set in the first message. When left empty, a new write // stream will be created. - StreamId string `protobuf:"bytes,2,opt,name=stream_id,json=streamId" json:"stream_id,omitempty"` + StreamId string `protobuf:"bytes,2,opt,name=stream_id,json=streamId,proto3" json:"stream_id,omitempty"` // The writes to apply. // // Always executed atomically and in order. // This must be empty on the first request. // This may be empty on the last request. // This must not be empty on all other requests. - Writes []*Write `protobuf:"bytes,3,rep,name=writes" json:"writes,omitempty"` + Writes []*Write `protobuf:"bytes,3,rep,name=writes,proto3" json:"writes,omitempty"` // A stream token that was previously sent by the server. // // The client should set this field to the token from the most recent @@ -1715,7 +1715,7 @@ type WriteRequest struct { // Leave this field unset when creating a new stream. StreamToken []byte `protobuf:"bytes,4,opt,name=stream_token,json=streamToken,proto3" json:"stream_token,omitempty"` // Labels associated with this write request. - Labels map[string]string `protobuf:"bytes,5,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,5,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1784,7 +1784,7 @@ func (m *WriteRequest) GetLabels() map[string]string { type WriteResponse struct { // The ID of the stream. // Only set on the first message, when a new stream was created. - StreamId string `protobuf:"bytes,1,opt,name=stream_id,json=streamId" json:"stream_id,omitempty"` + StreamId string `protobuf:"bytes,1,opt,name=stream_id,json=streamId,proto3" json:"stream_id,omitempty"` // A token that represents the position of this response in the stream. // This can be used by a client to resume the stream at this point. // @@ -1794,9 +1794,9 @@ type WriteResponse struct { // // This i-th write result corresponds to the i-th write in the // request. - WriteResults []*WriteResult `protobuf:"bytes,3,rep,name=write_results,json=writeResults" json:"write_results,omitempty"` + WriteResults []*WriteResult `protobuf:"bytes,3,rep,name=write_results,json=writeResults,proto3" json:"write_results,omitempty"` // The time at which the commit occurred. - CommitTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=commit_time,json=commitTime" json:"commit_time,omitempty"` + CommitTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=commit_time,json=commitTime,proto3" json:"commit_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1858,7 +1858,7 @@ func (m *WriteResponse) GetCommitTime() *timestamp.Timestamp { type ListenRequest struct { // The database name. In the format: // `projects/{project_id}/databases/{database_id}`. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The supported target changes. // // Types that are valid to be assigned to TargetChange: @@ -1866,7 +1866,7 @@ type ListenRequest struct { // *ListenRequest_RemoveTarget TargetChange isListenRequest_TargetChange `protobuf_oneof:"target_change"` // Labels associated with this target change. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1901,10 +1901,10 @@ type isListenRequest_TargetChange interface { } type ListenRequest_AddTarget struct { - AddTarget *Target `protobuf:"bytes,2,opt,name=add_target,json=addTarget,oneof"` + AddTarget *Target `protobuf:"bytes,2,opt,name=add_target,json=addTarget,proto3,oneof"` } type ListenRequest_RemoveTarget struct { - RemoveTarget int32 `protobuf:"varint,3,opt,name=remove_target,json=removeTarget,oneof"` + RemoveTarget int32 `protobuf:"varint,3,opt,name=remove_target,json=removeTarget,proto3,oneof"` } func (*ListenRequest_AddTarget) isListenRequest_TargetChange() {} @@ -2059,19 +2059,19 @@ type isListenResponse_ResponseType interface { } type ListenResponse_TargetChange struct { - TargetChange *TargetChange `protobuf:"bytes,2,opt,name=target_change,json=targetChange,oneof"` + TargetChange *TargetChange `protobuf:"bytes,2,opt,name=target_change,json=targetChange,proto3,oneof"` } type ListenResponse_DocumentChange struct { - DocumentChange *DocumentChange `protobuf:"bytes,3,opt,name=document_change,json=documentChange,oneof"` + DocumentChange *DocumentChange `protobuf:"bytes,3,opt,name=document_change,json=documentChange,proto3,oneof"` } type ListenResponse_DocumentDelete struct { - DocumentDelete *DocumentDelete `protobuf:"bytes,4,opt,name=document_delete,json=documentDelete,oneof"` + DocumentDelete *DocumentDelete `protobuf:"bytes,4,opt,name=document_delete,json=documentDelete,proto3,oneof"` } type ListenResponse_DocumentRemove struct { - DocumentRemove *DocumentRemove `protobuf:"bytes,6,opt,name=document_remove,json=documentRemove,oneof"` + DocumentRemove *DocumentRemove `protobuf:"bytes,6,opt,name=document_remove,json=documentRemove,proto3,oneof"` } type ListenResponse_Filter struct { - Filter *ExistenceFilter `protobuf:"bytes,5,opt,name=filter,oneof"` + Filter *ExistenceFilter `protobuf:"bytes,5,opt,name=filter,proto3,oneof"` } func (*ListenResponse_TargetChange) isListenResponse_ResponseType() {} @@ -2278,9 +2278,9 @@ type Target struct { // client-assigned or be server-assigned in a previous stream. All targets // with client provided IDs must be added before adding a target that needs // a server-assigned id. - TargetId int32 `protobuf:"varint,5,opt,name=target_id,json=targetId" json:"target_id,omitempty"` + TargetId int32 `protobuf:"varint,5,opt,name=target_id,json=targetId,proto3" json:"target_id,omitempty"` // If the target should be removed once it is current and consistent. - Once bool `protobuf:"varint,6,opt,name=once" json:"once,omitempty"` + Once bool `protobuf:"varint,6,opt,name=once,proto3" json:"once,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2318,16 +2318,16 @@ type isTarget_ResumeType interface { } type Target_Query struct { - Query *Target_QueryTarget `protobuf:"bytes,2,opt,name=query,oneof"` + Query *Target_QueryTarget `protobuf:"bytes,2,opt,name=query,proto3,oneof"` } type Target_Documents struct { - Documents *Target_DocumentsTarget `protobuf:"bytes,3,opt,name=documents,oneof"` + Documents *Target_DocumentsTarget `protobuf:"bytes,3,opt,name=documents,proto3,oneof"` } type Target_ResumeToken struct { ResumeToken []byte `protobuf:"bytes,4,opt,name=resume_token,json=resumeToken,proto3,oneof"` } type Target_ReadTime struct { - ReadTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=read_time,json=readTime,oneof"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=read_time,json=readTime,proto3,oneof"` } func (*Target_Query) isTarget_TargetType() {} @@ -2516,7 +2516,7 @@ type Target_DocumentsTarget struct { // `projects/{project_id}/databases/{database_id}/documents/{document_path}`. // The request will fail if any of the document is not a child resource of // the given `database`. Duplicate names will be elided. - Documents []string `protobuf:"bytes,2,rep,name=documents" json:"documents,omitempty"` + Documents []string `protobuf:"bytes,2,rep,name=documents,proto3" json:"documents,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2561,7 +2561,7 @@ type Target_QueryTarget struct { // For example: // `projects/my-project/databases/my-database/documents` or // `projects/my-project/databases/my-database/documents/chatrooms/my-chatroom` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The query to run. // // Types that are valid to be assigned to QueryType: @@ -2601,7 +2601,7 @@ type isTarget_QueryTarget_QueryType interface { } type Target_QueryTarget_StructuredQuery struct { - StructuredQuery *StructuredQuery `protobuf:"bytes,2,opt,name=structured_query,json=structuredQuery,oneof"` + StructuredQuery *StructuredQuery `protobuf:"bytes,2,opt,name=structured_query,json=structuredQuery,proto3,oneof"` } func (*Target_QueryTarget_StructuredQuery) isTarget_QueryTarget_QueryType() {} @@ -2685,7 +2685,7 @@ func _Target_QueryTarget_OneofSizer(msg proto.Message) (n int) { // Targets being watched have changed. type TargetChange struct { // The type of change that occurred. - TargetChangeType TargetChange_TargetChangeType `protobuf:"varint,1,opt,name=target_change_type,json=targetChangeType,enum=google.firestore.v1beta1.TargetChange_TargetChangeType" json:"target_change_type,omitempty"` + TargetChangeType TargetChange_TargetChangeType `protobuf:"varint,1,opt,name=target_change_type,json=targetChangeType,proto3,enum=google.firestore.v1beta1.TargetChange_TargetChangeType" json:"target_change_type,omitempty"` // The target IDs of targets that have changed. // // If empty, the change applies to all targets. @@ -2695,9 +2695,9 @@ type TargetChange struct { // associate server-assigned target IDs with added targets. // // For other states, the order of the target IDs is not defined. - TargetIds []int32 `protobuf:"varint,2,rep,packed,name=target_ids,json=targetIds" json:"target_ids,omitempty"` + TargetIds []int32 `protobuf:"varint,2,rep,packed,name=target_ids,json=targetIds,proto3" json:"target_ids,omitempty"` // The error that resulted in this change, if applicable. - Cause *status.Status `protobuf:"bytes,3,opt,name=cause" json:"cause,omitempty"` + Cause *status.Status `protobuf:"bytes,3,opt,name=cause,proto3" json:"cause,omitempty"` // A token that can be used to resume the stream for the given `target_ids`, // or all targets if `target_ids` is empty. // @@ -2713,7 +2713,7 @@ type TargetChange struct { // // For a given stream, `read_time` is guaranteed to be monotonically // increasing. - ReadTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=read_time,json=readTime" json:"read_time,omitempty"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=read_time,json=readTime,proto3" json:"read_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2784,12 +2784,12 @@ type ListCollectionIdsRequest struct { // `projects/{project_id}/databases/{database_id}/documents/{document_path}`. // For example: // `projects/my-project/databases/my-database/documents/chatrooms/my-chatroom` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The maximum number of results to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // A page token. Must be a value from // [ListCollectionIdsResponse][google.firestore.v1beta1.ListCollectionIdsResponse]. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2843,9 +2843,9 @@ func (m *ListCollectionIdsRequest) GetPageToken() string { // The response from [Firestore.ListCollectionIds][google.firestore.v1beta1.Firestore.ListCollectionIds]. type ListCollectionIdsResponse struct { // The collection ids. - CollectionIds []string `protobuf:"bytes,1,rep,name=collection_ids,json=collectionIds" json:"collection_ids,omitempty"` + CollectionIds []string `protobuf:"bytes,1,rep,name=collection_ids,json=collectionIds,proto3" json:"collection_ids,omitempty"` // A page token that may be used to continue the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3178,8 +3178,7 @@ func (c *firestoreClient) ListCollectionIds(ctx context.Context, in *ListCollect return out, nil } -// Server API for Firestore service - +// FirestoreServer is the server API for Firestore service. type FirestoreServer interface { // Gets a single document. GetDocument(context.Context, *GetDocumentRequest) (*Document, error) diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/query.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/query.pb.go index 1423a08af..83362e74c 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/query.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/query.pb.go @@ -152,11 +152,11 @@ func (StructuredQuery_UnaryFilter_Operator) EnumDescriptor() ([]byte, []int) { // A Firestore query. type StructuredQuery struct { // The projection to return. - Select *StructuredQuery_Projection `protobuf:"bytes,1,opt,name=select" json:"select,omitempty"` + Select *StructuredQuery_Projection `protobuf:"bytes,1,opt,name=select,proto3" json:"select,omitempty"` // The collections to query. - From []*StructuredQuery_CollectionSelector `protobuf:"bytes,2,rep,name=from" json:"from,omitempty"` + From []*StructuredQuery_CollectionSelector `protobuf:"bytes,2,rep,name=from,proto3" json:"from,omitempty"` // The filter to apply. - Where *StructuredQuery_Filter `protobuf:"bytes,3,opt,name=where" json:"where,omitempty"` + Where *StructuredQuery_Filter `protobuf:"bytes,3,opt,name=where,proto3" json:"where,omitempty"` // The order to apply to the query results. // // Firestore guarantees a stable ordering through the following rules: @@ -175,21 +175,21 @@ type StructuredQuery struct { // `SELECT * FROM Foo ORDER BY A DESC, __name__ DESC` // * `SELECT * FROM Foo WHERE A > 1` becomes // `SELECT * FROM Foo WHERE A > 1 ORDER BY A, __name__` - OrderBy []*StructuredQuery_Order `protobuf:"bytes,4,rep,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy []*StructuredQuery_Order `protobuf:"bytes,4,rep,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` // A starting point for the query results. - StartAt *Cursor `protobuf:"bytes,7,opt,name=start_at,json=startAt" json:"start_at,omitempty"` + StartAt *Cursor `protobuf:"bytes,7,opt,name=start_at,json=startAt,proto3" json:"start_at,omitempty"` // A end point for the query results. - EndAt *Cursor `protobuf:"bytes,8,opt,name=end_at,json=endAt" json:"end_at,omitempty"` + EndAt *Cursor `protobuf:"bytes,8,opt,name=end_at,json=endAt,proto3" json:"end_at,omitempty"` // The number of results to skip. // // Applies before limit, but after all other constraints. Must be >= 0 if // specified. - Offset int32 `protobuf:"varint,6,opt,name=offset" json:"offset,omitempty"` + Offset int32 `protobuf:"varint,6,opt,name=offset,proto3" json:"offset,omitempty"` // The maximum number of results to return. // // Applies after all other constraints. // Must be >= 0 if specified. - Limit *wrappers.Int32Value `protobuf:"bytes,5,opt,name=limit" json:"limit,omitempty"` + Limit *wrappers.Int32Value `protobuf:"bytes,5,opt,name=limit,proto3" json:"limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -279,11 +279,11 @@ func (m *StructuredQuery) GetLimit() *wrappers.Int32Value { type StructuredQuery_CollectionSelector struct { // The collection ID. // When set, selects only collections with this ID. - CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId" json:"collection_id,omitempty"` + CollectionId string `protobuf:"bytes,2,opt,name=collection_id,json=collectionId,proto3" json:"collection_id,omitempty"` // When false, selects only collections that are immediate children of // the `parent` specified in the containing `RunQueryRequest`. // When true, selects all descendant collections. - AllDescendants bool `protobuf:"varint,3,opt,name=all_descendants,json=allDescendants" json:"all_descendants,omitempty"` + AllDescendants bool `protobuf:"varint,3,opt,name=all_descendants,json=allDescendants,proto3" json:"all_descendants,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -370,13 +370,13 @@ type isStructuredQuery_Filter_FilterType interface { } type StructuredQuery_Filter_CompositeFilter struct { - CompositeFilter *StructuredQuery_CompositeFilter `protobuf:"bytes,1,opt,name=composite_filter,json=compositeFilter,oneof"` + CompositeFilter *StructuredQuery_CompositeFilter `protobuf:"bytes,1,opt,name=composite_filter,json=compositeFilter,proto3,oneof"` } type StructuredQuery_Filter_FieldFilter struct { - FieldFilter *StructuredQuery_FieldFilter `protobuf:"bytes,2,opt,name=field_filter,json=fieldFilter,oneof"` + FieldFilter *StructuredQuery_FieldFilter `protobuf:"bytes,2,opt,name=field_filter,json=fieldFilter,proto3,oneof"` } type StructuredQuery_Filter_UnaryFilter struct { - UnaryFilter *StructuredQuery_UnaryFilter `protobuf:"bytes,3,opt,name=unary_filter,json=unaryFilter,oneof"` + UnaryFilter *StructuredQuery_UnaryFilter `protobuf:"bytes,3,opt,name=unary_filter,json=unaryFilter,proto3,oneof"` } func (*StructuredQuery_Filter_CompositeFilter) isStructuredQuery_Filter_FilterType() {} @@ -507,10 +507,10 @@ func _StructuredQuery_Filter_OneofSizer(msg proto.Message) (n int) { // A filter that merges multiple other filters using the given operator. type StructuredQuery_CompositeFilter struct { // The operator for combining multiple filters. - Op StructuredQuery_CompositeFilter_Operator `protobuf:"varint,1,opt,name=op,enum=google.firestore.v1beta1.StructuredQuery_CompositeFilter_Operator" json:"op,omitempty"` + Op StructuredQuery_CompositeFilter_Operator `protobuf:"varint,1,opt,name=op,proto3,enum=google.firestore.v1beta1.StructuredQuery_CompositeFilter_Operator" json:"op,omitempty"` // The list of filters to combine. // Must contain at least one filter. - Filters []*StructuredQuery_Filter `protobuf:"bytes,2,rep,name=filters" json:"filters,omitempty"` + Filters []*StructuredQuery_Filter `protobuf:"bytes,2,rep,name=filters,proto3" json:"filters,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -557,11 +557,11 @@ func (m *StructuredQuery_CompositeFilter) GetFilters() []*StructuredQuery_Filter // A filter on a specific field. type StructuredQuery_FieldFilter struct { // The field to filter by. - Field *StructuredQuery_FieldReference `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *StructuredQuery_FieldReference `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` // The operator to filter by. - Op StructuredQuery_FieldFilter_Operator `protobuf:"varint,2,opt,name=op,enum=google.firestore.v1beta1.StructuredQuery_FieldFilter_Operator" json:"op,omitempty"` + Op StructuredQuery_FieldFilter_Operator `protobuf:"varint,2,opt,name=op,proto3,enum=google.firestore.v1beta1.StructuredQuery_FieldFilter_Operator" json:"op,omitempty"` // The value to compare to. - Value *Value `protobuf:"bytes,3,opt,name=value" json:"value,omitempty"` + Value *Value `protobuf:"bytes,3,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -615,7 +615,7 @@ func (m *StructuredQuery_FieldFilter) GetValue() *Value { // A filter with a single operand. type StructuredQuery_UnaryFilter struct { // The unary operator to apply. - Op StructuredQuery_UnaryFilter_Operator `protobuf:"varint,1,opt,name=op,enum=google.firestore.v1beta1.StructuredQuery_UnaryFilter_Operator" json:"op,omitempty"` + Op StructuredQuery_UnaryFilter_Operator `protobuf:"varint,1,opt,name=op,proto3,enum=google.firestore.v1beta1.StructuredQuery_UnaryFilter_Operator" json:"op,omitempty"` // The argument to the filter. // // Types that are valid to be assigned to OperandType: @@ -655,7 +655,7 @@ type isStructuredQuery_UnaryFilter_OperandType interface { } type StructuredQuery_UnaryFilter_Field struct { - Field *StructuredQuery_FieldReference `protobuf:"bytes,2,opt,name=field,oneof"` + Field *StructuredQuery_FieldReference `protobuf:"bytes,2,opt,name=field,proto3,oneof"` } func (*StructuredQuery_UnaryFilter_Field) isStructuredQuery_UnaryFilter_OperandType() {} @@ -739,9 +739,9 @@ func _StructuredQuery_UnaryFilter_OneofSizer(msg proto.Message) (n int) { // An order on a field. type StructuredQuery_Order struct { // The field to order by. - Field *StructuredQuery_FieldReference `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *StructuredQuery_FieldReference `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` // The direction to order by. Defaults to `ASCENDING`. - Direction StructuredQuery_Direction `protobuf:"varint,2,opt,name=direction,enum=google.firestore.v1beta1.StructuredQuery_Direction" json:"direction,omitempty"` + Direction StructuredQuery_Direction `protobuf:"varint,2,opt,name=direction,proto3,enum=google.firestore.v1beta1.StructuredQuery_Direction" json:"direction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -787,7 +787,7 @@ func (m *StructuredQuery_Order) GetDirection() StructuredQuery_Direction { // A reference to a field, such as `max(messages.time) as max_time`. type StructuredQuery_FieldReference struct { - FieldPath string `protobuf:"bytes,2,opt,name=field_path,json=fieldPath" json:"field_path,omitempty"` + FieldPath string `protobuf:"bytes,2,opt,name=field_path,json=fieldPath,proto3" json:"field_path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -830,7 +830,7 @@ type StructuredQuery_Projection struct { // // If empty, all fields are returned. To only return the name // of the document, use `['__name__']`. - Fields []*StructuredQuery_FieldReference `protobuf:"bytes,2,rep,name=fields" json:"fields,omitempty"` + Fields []*StructuredQuery_FieldReference `protobuf:"bytes,2,rep,name=fields,proto3" json:"fields,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -873,10 +873,10 @@ type Cursor struct { // the order by clause of a query. // // Can contain fewer values than specified in the order by clause. - Values []*Value `protobuf:"bytes,1,rep,name=values" json:"values,omitempty"` + Values []*Value `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` // If the position is just before or just after the given values, relative // to the sort order defined by the query. - Before bool `protobuf:"varint,2,opt,name=before" json:"before,omitempty"` + Before bool `protobuf:"varint,2,opt,name=before,proto3" json:"before,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/write.pb.go b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/write.pb.go index 90cce09fc..49615c106 100644 --- a/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/write.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/firestore/v1beta1/write.pb.go @@ -65,11 +65,11 @@ type Write struct { // Fields referenced in the mask, but not present in the input document, are // deleted from the document on the server. // The field paths in this mask must not contain a reserved field name. - UpdateMask *DocumentMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *DocumentMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // An optional precondition on the document. // // The write will fail if this is set and not met by the target document. - CurrentDocument *Precondition `protobuf:"bytes,4,opt,name=current_document,json=currentDocument" json:"current_document,omitempty"` + CurrentDocument *Precondition `protobuf:"bytes,4,opt,name=current_document,json=currentDocument,proto3" json:"current_document,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -104,13 +104,13 @@ type isWrite_Operation interface { } type Write_Update struct { - Update *Document `protobuf:"bytes,1,opt,name=update,oneof"` + Update *Document `protobuf:"bytes,1,opt,name=update,proto3,oneof"` } type Write_Delete struct { - Delete string `protobuf:"bytes,2,opt,name=delete,oneof"` + Delete string `protobuf:"bytes,2,opt,name=delete,proto3,oneof"` } type Write_Transform struct { - Transform *DocumentTransform `protobuf:"bytes,6,opt,name=transform,oneof"` + Transform *DocumentTransform `protobuf:"bytes,6,opt,name=transform,proto3,oneof"` } func (*Write_Update) isWrite_Operation() {} @@ -251,11 +251,11 @@ func _Write_OneofSizer(msg proto.Message) (n int) { // A transformation of a document. type DocumentTransform struct { // The name of the document to transform. - Document string `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document string `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // The list of transformations to apply to the fields of the document, in // order. // This must not be empty. - FieldTransforms []*DocumentTransform_FieldTransform `protobuf:"bytes,2,rep,name=field_transforms,json=fieldTransforms" json:"field_transforms,omitempty"` + FieldTransforms []*DocumentTransform_FieldTransform `protobuf:"bytes,2,rep,name=field_transforms,json=fieldTransforms,proto3" json:"field_transforms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -303,7 +303,7 @@ func (m *DocumentTransform) GetFieldTransforms() []*DocumentTransform_FieldTrans type DocumentTransform_FieldTransform struct { // The path of the field. See [Document.fields][google.firestore.v1beta1.Document.fields] for the field path syntax // reference. - FieldPath string `protobuf:"bytes,1,opt,name=field_path,json=fieldPath" json:"field_path,omitempty"` + FieldPath string `protobuf:"bytes,1,opt,name=field_path,json=fieldPath,proto3" json:"field_path,omitempty"` // The transformation to apply on the field. // // Types that are valid to be assigned to TransformType: @@ -343,7 +343,7 @@ type isDocumentTransform_FieldTransform_TransformType interface { } type DocumentTransform_FieldTransform_SetToServerValue struct { - SetToServerValue DocumentTransform_FieldTransform_ServerValue `protobuf:"varint,2,opt,name=set_to_server_value,json=setToServerValue,enum=google.firestore.v1beta1.DocumentTransform_FieldTransform_ServerValue,oneof"` + SetToServerValue DocumentTransform_FieldTransform_ServerValue `protobuf:"varint,2,opt,name=set_to_server_value,json=setToServerValue,proto3,enum=google.firestore.v1beta1.DocumentTransform_FieldTransform_ServerValue,oneof"` } func (*DocumentTransform_FieldTransform_SetToServerValue) isDocumentTransform_FieldTransform_TransformType() { @@ -427,10 +427,10 @@ type WriteResult struct { // // If the write did not actually change the document, this will be the // previous update_time. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // The results of applying each [DocumentTransform.FieldTransform][google.firestore.v1beta1.DocumentTransform.FieldTransform], in the // same order. - TransformResults []*Value `protobuf:"bytes,2,rep,name=transform_results,json=transformResults" json:"transform_results,omitempty"` + TransformResults []*Value `protobuf:"bytes,2,rep,name=transform_results,json=transformResults,proto3" json:"transform_results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -485,11 +485,11 @@ type DocumentChange struct { // The new state of the [Document][google.firestore.v1beta1.Document]. // // If `mask` is set, contains only fields that were updated or added. - Document *Document `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document *Document `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // A set of target IDs of targets that match this document. - TargetIds []int32 `protobuf:"varint,5,rep,packed,name=target_ids,json=targetIds" json:"target_ids,omitempty"` + TargetIds []int32 `protobuf:"varint,5,rep,packed,name=target_ids,json=targetIds,proto3" json:"target_ids,omitempty"` // A set of target IDs for targets that no longer match this document. - RemovedTargetIds []int32 `protobuf:"varint,6,rep,packed,name=removed_target_ids,json=removedTargetIds" json:"removed_target_ids,omitempty"` + RemovedTargetIds []int32 `protobuf:"varint,6,rep,packed,name=removed_target_ids,json=removedTargetIds,proto3" json:"removed_target_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -549,13 +549,13 @@ func (m *DocumentChange) GetRemovedTargetIds() []int32 { // delete, if multiple targets are affected. type DocumentDelete struct { // The resource name of the [Document][google.firestore.v1beta1.Document] that was deleted. - Document string `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document string `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // A set of target IDs for targets that previously matched this entity. - RemovedTargetIds []int32 `protobuf:"varint,6,rep,packed,name=removed_target_ids,json=removedTargetIds" json:"removed_target_ids,omitempty"` + RemovedTargetIds []int32 `protobuf:"varint,6,rep,packed,name=removed_target_ids,json=removedTargetIds,proto3" json:"removed_target_ids,omitempty"` // The read timestamp at which the delete was observed. // // Greater or equal to the `commit_time` of the delete. - ReadTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_time,json=readTime" json:"read_time,omitempty"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_time,json=readTime,proto3" json:"read_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -616,13 +616,13 @@ func (m *DocumentDelete) GetReadTime() *timestamp.Timestamp { // write or delete, if multiple targets are affected. type DocumentRemove struct { // The resource name of the [Document][google.firestore.v1beta1.Document] that has gone out of view. - Document string `protobuf:"bytes,1,opt,name=document" json:"document,omitempty"` + Document string `protobuf:"bytes,1,opt,name=document,proto3" json:"document,omitempty"` // A set of target IDs for targets that previously matched this document. - RemovedTargetIds []int32 `protobuf:"varint,2,rep,packed,name=removed_target_ids,json=removedTargetIds" json:"removed_target_ids,omitempty"` + RemovedTargetIds []int32 `protobuf:"varint,2,rep,packed,name=removed_target_ids,json=removedTargetIds,proto3" json:"removed_target_ids,omitempty"` // The read timestamp at which the remove was observed. // // Greater or equal to the `commit_time` of the change/delete/remove. - ReadTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_time,json=readTime" json:"read_time,omitempty"` + ReadTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_time,json=readTime,proto3" json:"read_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -676,12 +676,12 @@ func (m *DocumentRemove) GetReadTime() *timestamp.Timestamp { // A digest of all the documents that match a given target. type ExistenceFilter struct { // The target ID to which this filter applies. - TargetId int32 `protobuf:"varint,1,opt,name=target_id,json=targetId" json:"target_id,omitempty"` + TargetId int32 `protobuf:"varint,1,opt,name=target_id,json=targetId,proto3" json:"target_id,omitempty"` // The total count of documents that match [target_id][google.firestore.v1beta1.ExistenceFilter.target_id]. // // If different from the count of documents in the client that match, the // client must manually determine which documents no longer match the target. - Count int32 `protobuf:"varint,2,opt,name=count" json:"count,omitempty"` + Count int32 `protobuf:"varint,2,opt,name=count,proto3" json:"count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/annotations.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/annotations.pb.go index c7b83b42c..cd4ff0264 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/annotations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/annotations.pb.go @@ -258,22 +258,22 @@ func (VariantAnnotation_ClinicalSignificance) EnumDescriptor() ([]byte, []int) { // genes from refseq', and 'all variant annotations from ClinVar'. type AnnotationSet struct { // The server-generated annotation set ID, unique across all annotation sets. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The dataset to which this annotation set belongs. - DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The ID of the reference set that defines the coordinate space for this // set's annotations. - ReferenceSetId string `protobuf:"bytes,3,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,3,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // The display name for this annotation set. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // The source URI describing the file from which this annotation set was // generated, if any. - SourceUri string `protobuf:"bytes,5,opt,name=source_uri,json=sourceUri" json:"source_uri,omitempty"` + SourceUri string `protobuf:"bytes,5,opt,name=source_uri,json=sourceUri,proto3" json:"source_uri,omitempty"` // The type of annotations contained within this set. - Type AnnotationType `protobuf:"varint,6,opt,name=type,enum=google.genomics.v1.AnnotationType" json:"type,omitempty"` + Type AnnotationType `protobuf:"varint,6,opt,name=type,proto3,enum=google.genomics.v1.AnnotationType" json:"type,omitempty"` // A map of additional read alignment information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,17,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,17,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -360,34 +360,34 @@ func (m *AnnotationSet) GetInfo() map[string]*_struct.ListValue { // `VARIANT`. type Annotation struct { // The server-generated annotation ID, unique across all annotations. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The annotation set to which this annotation belongs. - AnnotationSetId string `protobuf:"bytes,2,opt,name=annotation_set_id,json=annotationSetId" json:"annotation_set_id,omitempty"` + AnnotationSetId string `protobuf:"bytes,2,opt,name=annotation_set_id,json=annotationSetId,proto3" json:"annotation_set_id,omitempty"` // The display name of this annotation. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The ID of the Google Genomics reference associated with this range. - ReferenceId string `protobuf:"bytes,4,opt,name=reference_id,json=referenceId" json:"reference_id,omitempty"` + ReferenceId string `protobuf:"bytes,4,opt,name=reference_id,json=referenceId,proto3" json:"reference_id,omitempty"` // The display name corresponding to the reference specified by // `referenceId`, for example `chr1`, `1`, or `chrX`. - ReferenceName string `protobuf:"bytes,5,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,5,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The start position of the range on the reference, 0-based inclusive. - Start int64 `protobuf:"varint,6,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,6,opt,name=start,proto3" json:"start,omitempty"` // The end position of the range on the reference, 0-based exclusive. - End int64 `protobuf:"varint,7,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,7,opt,name=end,proto3" json:"end,omitempty"` // Whether this range refers to the reverse strand, as opposed to the forward // strand. Note that regardless of this field, the start/end position of the // range always refer to the forward strand. - ReverseStrand bool `protobuf:"varint,8,opt,name=reverse_strand,json=reverseStrand" json:"reverse_strand,omitempty"` + ReverseStrand bool `protobuf:"varint,8,opt,name=reverse_strand,json=reverseStrand,proto3" json:"reverse_strand,omitempty"` // The data type for this annotation. Must match the containing annotation // set's type. - Type AnnotationType `protobuf:"varint,9,opt,name=type,enum=google.genomics.v1.AnnotationType" json:"type,omitempty"` + Type AnnotationType `protobuf:"varint,9,opt,name=type,proto3,enum=google.genomics.v1.AnnotationType" json:"type,omitempty"` // Types that are valid to be assigned to Value: // *Annotation_Variant // *Annotation_Transcript Value isAnnotation_Value `protobuf_oneof:"value"` // A map of additional read alignment information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,12,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,12,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -422,10 +422,10 @@ type isAnnotation_Value interface { } type Annotation_Variant struct { - Variant *VariantAnnotation `protobuf:"bytes,10,opt,name=variant,oneof"` + Variant *VariantAnnotation `protobuf:"bytes,10,opt,name=variant,proto3,oneof"` } type Annotation_Transcript struct { - Transcript *Transcript `protobuf:"bytes,11,opt,name=transcript,oneof"` + Transcript *Transcript `protobuf:"bytes,11,opt,name=transcript,proto3,oneof"` } func (*Annotation_Variant) isAnnotation_Value() {} @@ -598,27 +598,27 @@ func _Annotation_OneofSizer(msg proto.Message) (n int) { type VariantAnnotation struct { // Type has been adapted from ClinVar's list of variant types. - Type VariantAnnotation_Type `protobuf:"varint,1,opt,name=type,enum=google.genomics.v1.VariantAnnotation_Type" json:"type,omitempty"` + Type VariantAnnotation_Type `protobuf:"varint,1,opt,name=type,proto3,enum=google.genomics.v1.VariantAnnotation_Type" json:"type,omitempty"` // Effect of the variant on the coding sequence. - Effect VariantAnnotation_Effect `protobuf:"varint,2,opt,name=effect,enum=google.genomics.v1.VariantAnnotation_Effect" json:"effect,omitempty"` + Effect VariantAnnotation_Effect `protobuf:"varint,2,opt,name=effect,proto3,enum=google.genomics.v1.VariantAnnotation_Effect" json:"effect,omitempty"` // The alternate allele for this variant. If multiple alternate alleles // exist at this location, create a separate variant for each one, as they // may represent distinct conditions. - AlternateBases string `protobuf:"bytes,3,opt,name=alternate_bases,json=alternateBases" json:"alternate_bases,omitempty"` + AlternateBases string `protobuf:"bytes,3,opt,name=alternate_bases,json=alternateBases,proto3" json:"alternate_bases,omitempty"` // Google annotation ID of the gene affected by this variant. This should // be provided when the variant is created. - GeneId string `protobuf:"bytes,4,opt,name=gene_id,json=geneId" json:"gene_id,omitempty"` + GeneId string `protobuf:"bytes,4,opt,name=gene_id,json=geneId,proto3" json:"gene_id,omitempty"` // Google annotation IDs of the transcripts affected by this variant. These // should be provided when the variant is created. - TranscriptIds []string `protobuf:"bytes,5,rep,name=transcript_ids,json=transcriptIds" json:"transcript_ids,omitempty"` + TranscriptIds []string `protobuf:"bytes,5,rep,name=transcript_ids,json=transcriptIds,proto3" json:"transcript_ids,omitempty"` // The set of conditions associated with this variant. // A condition describes the way a variant influences human health. - Conditions []*VariantAnnotation_ClinicalCondition `protobuf:"bytes,6,rep,name=conditions" json:"conditions,omitempty"` + Conditions []*VariantAnnotation_ClinicalCondition `protobuf:"bytes,6,rep,name=conditions,proto3" json:"conditions,omitempty"` // Describes the clinical significance of a variant. // It is adapted from the ClinVar controlled vocabulary for clinical // significance described at: // http://www.ncbi.nlm.nih.gov/clinvar/docs/clinsig/ - ClinicalSignificance VariantAnnotation_ClinicalSignificance `protobuf:"varint,7,opt,name=clinical_significance,json=clinicalSignificance,enum=google.genomics.v1.VariantAnnotation_ClinicalSignificance" json:"clinical_significance,omitempty"` + ClinicalSignificance VariantAnnotation_ClinicalSignificance `protobuf:"varint,7,opt,name=clinical_significance,json=clinicalSignificance,proto3,enum=google.genomics.v1.VariantAnnotation_ClinicalSignificance" json:"clinical_significance,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -699,15 +699,15 @@ func (m *VariantAnnotation) GetClinicalSignificance() VariantAnnotation_Clinical type VariantAnnotation_ClinicalCondition struct { // A set of names for the condition. - Names []string `protobuf:"bytes,1,rep,name=names" json:"names,omitempty"` + Names []string `protobuf:"bytes,1,rep,name=names,proto3" json:"names,omitempty"` // The set of external IDs for this condition. - ExternalIds []*ExternalId `protobuf:"bytes,2,rep,name=external_ids,json=externalIds" json:"external_ids,omitempty"` + ExternalIds []*ExternalId `protobuf:"bytes,2,rep,name=external_ids,json=externalIds,proto3" json:"external_ids,omitempty"` // The MedGen concept id associated with this gene. // Search for these IDs at http://www.ncbi.nlm.nih.gov/medgen/ - ConceptId string `protobuf:"bytes,3,opt,name=concept_id,json=conceptId" json:"concept_id,omitempty"` + ConceptId string `protobuf:"bytes,3,opt,name=concept_id,json=conceptId,proto3" json:"concept_id,omitempty"` // The OMIM id for this condition. // Search for these IDs at http://omim.org/ - OmimId string `protobuf:"bytes,4,opt,name=omim_id,json=omimId" json:"omim_id,omitempty"` + OmimId string `protobuf:"bytes,4,opt,name=omim_id,json=omimId,proto3" json:"omim_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -769,7 +769,7 @@ func (m *VariantAnnotation_ClinicalCondition) GetOmimId() string { // reference genome may be transcribed as RNA. type Transcript struct { // The annotation ID of the gene from which this transcript is transcribed. - GeneId string `protobuf:"bytes,1,opt,name=gene_id,json=geneId" json:"gene_id,omitempty"` + GeneId string `protobuf:"bytes,1,opt,name=gene_id,json=geneId,proto3" json:"gene_id,omitempty"` // The exons that compose // this transcript. This field should be unset for genomes where transcript // splicing does not occur, for example prokaryotes. @@ -784,7 +784,7 @@ type Transcript struct { // to coding DNA sequence. // // Exons are ordered by start position and may not overlap. - Exons []*Transcript_Exon `protobuf:"bytes,2,rep,name=exons" json:"exons,omitempty"` + Exons []*Transcript_Exon `protobuf:"bytes,2,rep,name=exons,proto3" json:"exons,omitempty"` // The range of the coding sequence for this transcript, if any. To determine // the exact ranges of coding sequence, intersect this range with those of the // [exons][google.genomics.v1.Transcript.exons], if any. If there are any @@ -798,7 +798,7 @@ type Transcript struct { // [exon.frame][google.genomics.v1.Transcript.Exon.frame] will not necessarily // match the expected reference reading frame and coding exon reference bases // cannot necessarily be concatenated to produce the original transcript mRNA. - CodingSequence *Transcript_CodingSequence `protobuf:"bytes,3,opt,name=coding_sequence,json=codingSequence" json:"coding_sequence,omitempty"` + CodingSequence *Transcript_CodingSequence `protobuf:"bytes,3,opt,name=coding_sequence,json=codingSequence,proto3" json:"coding_sequence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -853,11 +853,11 @@ type Transcript_Exon struct { // The start position of the exon on this annotation's reference sequence, // 0-based inclusive. Note that this is relative to the reference start, and // **not** the containing annotation start. - Start int64 `protobuf:"varint,1,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,1,opt,name=start,proto3" json:"start,omitempty"` // The end position of the exon on this annotation's reference sequence, // 0-based exclusive. Note that this is relative to the reference start, and // *not* the containing annotation start. - End int64 `protobuf:"varint,2,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,2,opt,name=end,proto3" json:"end,omitempty"` // The frame of this exon. Contains a value of 0, 1, or 2, which indicates // the offset of the first coding base of the exon within the reading frame // of the coding DNA sequence, if any. This field is dependent on the @@ -871,7 +871,7 @@ type Transcript_Exon struct { // Unset if this exon does not intersect the coding sequence. Upon creation // of a transcript, the frame must be populated for all or none of the // coding exons. - Frame *wrappers.Int32Value `protobuf:"bytes,3,opt,name=frame" json:"frame,omitempty"` + Frame *wrappers.Int32Value `protobuf:"bytes,3,opt,name=frame,proto3" json:"frame,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -926,11 +926,11 @@ type Transcript_CodingSequence struct { // The start of the coding sequence on this annotation's reference sequence, // 0-based inclusive. Note that this position is relative to the reference // start, and *not* the containing annotation start. - Start int64 `protobuf:"varint,1,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,1,opt,name=start,proto3" json:"start,omitempty"` // The end of the coding sequence on this annotation's reference sequence, // 0-based exclusive. Note that this position is relative to the reference // start, and *not* the containing annotation start. - End int64 `protobuf:"varint,2,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,2,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -976,9 +976,9 @@ func (m *Transcript_CodingSequence) GetEnd() int64 { type ExternalId struct { // The name of the source of this data. - SourceName string `protobuf:"bytes,1,opt,name=source_name,json=sourceName" json:"source_name,omitempty"` + SourceName string `protobuf:"bytes,1,opt,name=source_name,json=sourceName,proto3" json:"source_name,omitempty"` // The id used by the source of this data. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1024,7 +1024,7 @@ func (m *ExternalId) GetId() string { type CreateAnnotationSetRequest struct { // The annotation set to create. - AnnotationSet *AnnotationSet `protobuf:"bytes,1,opt,name=annotation_set,json=annotationSet" json:"annotation_set,omitempty"` + AnnotationSet *AnnotationSet `protobuf:"bytes,1,opt,name=annotation_set,json=annotationSet,proto3" json:"annotation_set,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1063,7 +1063,7 @@ func (m *CreateAnnotationSetRequest) GetAnnotationSet() *AnnotationSet { type GetAnnotationSetRequest struct { // The ID of the annotation set to be retrieved. - AnnotationSetId string `protobuf:"bytes,1,opt,name=annotation_set_id,json=annotationSetId" json:"annotation_set_id,omitempty"` + AnnotationSetId string `protobuf:"bytes,1,opt,name=annotation_set_id,json=annotationSetId,proto3" json:"annotation_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1102,15 +1102,15 @@ func (m *GetAnnotationSetRequest) GetAnnotationSetId() string { type UpdateAnnotationSetRequest struct { // The ID of the annotation set to be updated. - AnnotationSetId string `protobuf:"bytes,1,opt,name=annotation_set_id,json=annotationSetId" json:"annotation_set_id,omitempty"` + AnnotationSetId string `protobuf:"bytes,1,opt,name=annotation_set_id,json=annotationSetId,proto3" json:"annotation_set_id,omitempty"` // The new annotation set. - AnnotationSet *AnnotationSet `protobuf:"bytes,2,opt,name=annotation_set,json=annotationSet" json:"annotation_set,omitempty"` + AnnotationSet *AnnotationSet `protobuf:"bytes,2,opt,name=annotation_set,json=annotationSet,proto3" json:"annotation_set,omitempty"` // An optional mask specifying which fields to update. Mutable fields are // [name][google.genomics.v1.AnnotationSet.name], // [source_uri][google.genomics.v1.AnnotationSet.source_uri], and // [info][google.genomics.v1.AnnotationSet.info]. If unspecified, all // mutable fields will be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1163,7 +1163,7 @@ func (m *UpdateAnnotationSetRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteAnnotationSetRequest struct { // The ID of the annotation set to be deleted. - AnnotationSetId string `protobuf:"bytes,1,opt,name=annotation_set_id,json=annotationSetId" json:"annotation_set_id,omitempty"` + AnnotationSetId string `protobuf:"bytes,1,opt,name=annotation_set_id,json=annotationSetId,proto3" json:"annotation_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1203,23 +1203,23 @@ func (m *DeleteAnnotationSetRequest) GetAnnotationSetId() string { type SearchAnnotationSetsRequest struct { // Required. The dataset IDs to search within. Caller must have `READ` access // to these datasets. - DatasetIds []string `protobuf:"bytes,1,rep,name=dataset_ids,json=datasetIds" json:"dataset_ids,omitempty"` + DatasetIds []string `protobuf:"bytes,1,rep,name=dataset_ids,json=datasetIds,proto3" json:"dataset_ids,omitempty"` // If specified, only annotation sets associated with the given reference set // are returned. - ReferenceSetId string `protobuf:"bytes,2,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,2,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // Only return annotations sets for which a substring of the name matches this // string (case insensitive). - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // If specified, only annotation sets that have any of these types are // returned. - Types []AnnotationType `protobuf:"varint,4,rep,packed,name=types,enum=google.genomics.v1.AnnotationType" json:"types,omitempty"` + Types []AnnotationType `protobuf:"varint,4,rep,packed,name=types,proto3,enum=google.genomics.v1.AnnotationType" json:"types,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 128. The maximum value is 1024. - PageSize int32 `protobuf:"varint,6,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,6,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1293,11 +1293,11 @@ func (m *SearchAnnotationSetsRequest) GetPageSize() int32 { type SearchAnnotationSetsResponse struct { // The matching annotation sets. - AnnotationSets []*AnnotationSet `protobuf:"bytes,1,rep,name=annotation_sets,json=annotationSets" json:"annotation_sets,omitempty"` + AnnotationSets []*AnnotationSet `protobuf:"bytes,1,rep,name=annotation_sets,json=annotationSets,proto3" json:"annotation_sets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1343,7 +1343,7 @@ func (m *SearchAnnotationSetsResponse) GetNextPageToken() string { type CreateAnnotationRequest struct { // The annotation to be created. - Annotation *Annotation `protobuf:"bytes,1,opt,name=annotation" json:"annotation,omitempty"` + Annotation *Annotation `protobuf:"bytes,1,opt,name=annotation,proto3" json:"annotation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1383,7 +1383,7 @@ func (m *CreateAnnotationRequest) GetAnnotation() *Annotation { type BatchCreateAnnotationsRequest struct { // The annotations to be created. At most 4096 can be specified in a single // request. - Annotations []*Annotation `protobuf:"bytes,1,rep,name=annotations" json:"annotations,omitempty"` + Annotations []*Annotation `protobuf:"bytes,1,rep,name=annotations,proto3" json:"annotations,omitempty"` // A unique request ID which enables the server to detect duplicated requests. // If provided, duplicated requests will result in the same response; if not // provided, duplicated requests may result in duplicated data. For a given @@ -1391,7 +1391,7 @@ type BatchCreateAnnotationsRequest struct { // different batches of annotations - behavior in this case is undefined. // A common approach is to use a UUID. For batch jobs where worker crashes are // a possibility, consider using some unique variant of a worker or run ID. - RequestId string `protobuf:"bytes,2,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,2,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1438,7 +1438,7 @@ func (m *BatchCreateAnnotationsRequest) GetRequestId() string { type BatchCreateAnnotationsResponse struct { // The resulting per-annotation entries, ordered consistently with the // original request. - Entries []*BatchCreateAnnotationsResponse_Entry `protobuf:"bytes,1,rep,name=entries" json:"entries,omitempty"` + Entries []*BatchCreateAnnotationsResponse_Entry `protobuf:"bytes,1,rep,name=entries,proto3" json:"entries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1477,9 +1477,9 @@ func (m *BatchCreateAnnotationsResponse) GetEntries() []*BatchCreateAnnotationsR type BatchCreateAnnotationsResponse_Entry struct { // The creation status. - Status *status.Status `protobuf:"bytes,1,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,1,opt,name=status,proto3" json:"status,omitempty"` // The created annotation, if creation was successful. - Annotation *Annotation `protobuf:"bytes,2,opt,name=annotation" json:"annotation,omitempty"` + Annotation *Annotation `protobuf:"bytes,2,opt,name=annotation,proto3" json:"annotation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1525,7 +1525,7 @@ func (m *BatchCreateAnnotationsResponse_Entry) GetAnnotation() *Annotation { type GetAnnotationRequest struct { // The ID of the annotation to be retrieved. - AnnotationId string `protobuf:"bytes,1,opt,name=annotation_id,json=annotationId" json:"annotation_id,omitempty"` + AnnotationId string `protobuf:"bytes,1,opt,name=annotation_id,json=annotationId,proto3" json:"annotation_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1564,16 +1564,16 @@ func (m *GetAnnotationRequest) GetAnnotationId() string { type UpdateAnnotationRequest struct { // The ID of the annotation to be updated. - AnnotationId string `protobuf:"bytes,1,opt,name=annotation_id,json=annotationId" json:"annotation_id,omitempty"` + AnnotationId string `protobuf:"bytes,1,opt,name=annotation_id,json=annotationId,proto3" json:"annotation_id,omitempty"` // The new annotation. - Annotation *Annotation `protobuf:"bytes,2,opt,name=annotation" json:"annotation,omitempty"` + Annotation *Annotation `protobuf:"bytes,2,opt,name=annotation,proto3" json:"annotation,omitempty"` // An optional mask specifying which fields to update. Mutable fields are // [name][google.genomics.v1.Annotation.name], // [variant][google.genomics.v1.Annotation.variant], // [transcript][google.genomics.v1.Annotation.transcript], and // [info][google.genomics.v1.Annotation.info]. If unspecified, all mutable // fields will be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1626,7 +1626,7 @@ func (m *UpdateAnnotationRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteAnnotationRequest struct { // The ID of the annotation to be deleted. - AnnotationId string `protobuf:"bytes,1,opt,name=annotation_id,json=annotationId" json:"annotation_id,omitempty"` + AnnotationId string `protobuf:"bytes,1,opt,name=annotation_id,json=annotationId,proto3" json:"annotation_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1667,7 +1667,7 @@ type SearchAnnotationsRequest struct { // Required. The annotation sets to search within. The caller must have // `READ` access to these annotation sets. // All queried annotation sets must have the same type. - AnnotationSetIds []string `protobuf:"bytes,1,rep,name=annotation_set_ids,json=annotationSetIds" json:"annotation_set_ids,omitempty"` + AnnotationSetIds []string `protobuf:"bytes,1,rep,name=annotation_set_ids,json=annotationSetIds,proto3" json:"annotation_set_ids,omitempty"` // Required. `reference_id` or `reference_name` must be set. // // Types that are valid to be assigned to Reference: @@ -1679,19 +1679,19 @@ type SearchAnnotationsRequest struct { // [referenceId][google.genomics.v1.SearchAnnotationsRequest.reference_id] or // [referenceName][google.genomics.v1.SearchAnnotationsRequest.reference_name] // must be specified. Defaults to 0. - Start int64 `protobuf:"varint,4,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,4,opt,name=start,proto3" json:"start,omitempty"` // The end position of the range on the reference, 0-based exclusive. If // [referenceId][google.genomics.v1.SearchAnnotationsRequest.reference_id] or // [referenceName][google.genomics.v1.SearchAnnotationsRequest.reference_name] // must be specified, Defaults to the length of the reference. - End int64 `protobuf:"varint,5,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,5,opt,name=end,proto3" json:"end,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,6,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,6,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 256. The maximum value is 2048. - PageSize int32 `protobuf:"varint,7,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,7,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1726,10 +1726,10 @@ type isSearchAnnotationsRequest_Reference interface { } type SearchAnnotationsRequest_ReferenceId struct { - ReferenceId string `protobuf:"bytes,2,opt,name=reference_id,json=referenceId,oneof"` + ReferenceId string `protobuf:"bytes,2,opt,name=reference_id,json=referenceId,proto3,oneof"` } type SearchAnnotationsRequest_ReferenceName struct { - ReferenceName string `protobuf:"bytes,3,opt,name=reference_name,json=referenceName,oneof"` + ReferenceName string `protobuf:"bytes,3,opt,name=reference_name,json=referenceName,proto3,oneof"` } func (*SearchAnnotationsRequest_ReferenceId) isSearchAnnotationsRequest_Reference() {} @@ -1859,11 +1859,11 @@ func _SearchAnnotationsRequest_OneofSizer(msg proto.Message) (n int) { type SearchAnnotationsResponse struct { // The matching annotations. - Annotations []*Annotation `protobuf:"bytes,1,rep,name=annotations" json:"annotations,omitempty"` + Annotations []*Annotation `protobuf:"bytes,1,rep,name=annotations,proto3" json:"annotations,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2142,8 +2142,7 @@ func (c *annotationServiceV1Client) SearchAnnotations(ctx context.Context, in *S return out, nil } -// Server API for AnnotationServiceV1 service - +// AnnotationServiceV1Server is the server API for AnnotationServiceV1 service. type AnnotationServiceV1Server interface { // Creates a new annotation set. Caller must have WRITE permission for the // associated dataset. diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/cigar.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/cigar.pb.go index baa71743b..1f346b300 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/cigar.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/cigar.pb.go @@ -105,14 +105,14 @@ func (CigarUnit_Operation) EnumDescriptor() ([]byte, []int) { // A single CIGAR operation. type CigarUnit struct { - Operation CigarUnit_Operation `protobuf:"varint,1,opt,name=operation,enum=google.genomics.v1.CigarUnit_Operation" json:"operation,omitempty"` + Operation CigarUnit_Operation `protobuf:"varint,1,opt,name=operation,proto3,enum=google.genomics.v1.CigarUnit_Operation" json:"operation,omitempty"` // The number of genomic bases that the operation runs for. Required. - OperationLength int64 `protobuf:"varint,2,opt,name=operation_length,json=operationLength" json:"operation_length,omitempty"` + OperationLength int64 `protobuf:"varint,2,opt,name=operation_length,json=operationLength,proto3" json:"operation_length,omitempty"` // `referenceSequence` is only used at mismatches // (`SEQUENCE_MISMATCH`) and deletions (`DELETE`). // Filling this field replaces SAM's MD tag. If the relevant information is // not available, this field is unset. - ReferenceSequence string `protobuf:"bytes,3,opt,name=reference_sequence,json=referenceSequence" json:"reference_sequence,omitempty"` + ReferenceSequence string `protobuf:"bytes,3,opt,name=reference_sequence,json=referenceSequence,proto3" json:"reference_sequence,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/datasets.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/datasets.pb.go index ac042278f..b3ad79a4e 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/datasets.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/datasets.pb.go @@ -34,13 +34,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Genomics](https://cloud.google.com/genomics/fundamentals-of-google-genomics) type Dataset struct { // The server-generated dataset ID, unique across all datasets. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The Google Cloud project ID that this dataset belongs to. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The dataset name. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The time this dataset was created, in seconds from the epoch. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -101,14 +101,14 @@ func (m *Dataset) GetCreateTime() *timestamp.Timestamp { // The dataset list request. type ListDatasetsRequest struct { // Required. The Google Cloud project ID to list datasets for. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 50. The maximum value is 1024. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -162,11 +162,11 @@ func (m *ListDatasetsRequest) GetPageToken() string { // The dataset list response. type ListDatasetsResponse struct { // The list of matching Datasets. - Datasets []*Dataset `protobuf:"bytes,1,rep,name=datasets" json:"datasets,omitempty"` + Datasets []*Dataset `protobuf:"bytes,1,rep,name=datasets,proto3" json:"datasets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -212,7 +212,7 @@ func (m *ListDatasetsResponse) GetNextPageToken() string { type CreateDatasetRequest struct { // The dataset to be created. Must contain projectId and name. - Dataset *Dataset `protobuf:"bytes,1,opt,name=dataset" json:"dataset,omitempty"` + Dataset *Dataset `protobuf:"bytes,1,opt,name=dataset,proto3" json:"dataset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -251,14 +251,14 @@ func (m *CreateDatasetRequest) GetDataset() *Dataset { type UpdateDatasetRequest struct { // The ID of the dataset to be updated. - DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The new dataset data. - Dataset *Dataset `protobuf:"bytes,2,opt,name=dataset" json:"dataset,omitempty"` + Dataset *Dataset `protobuf:"bytes,2,opt,name=dataset,proto3" json:"dataset,omitempty"` // An optional mask specifying which fields to update. At this time, the only // mutable field is [name][google.genomics.v1.Dataset.name]. The only // acceptable value is "name". If unspecified, all mutable fields will be // updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -311,7 +311,7 @@ func (m *UpdateDatasetRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteDatasetRequest struct { // The ID of the dataset to be deleted. - DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -350,7 +350,7 @@ func (m *DeleteDatasetRequest) GetDatasetId() string { type UndeleteDatasetRequest struct { // The ID of the dataset to be undeleted. - DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -389,7 +389,7 @@ func (m *UndeleteDatasetRequest) GetDatasetId() string { type GetDatasetRequest struct { // The ID of the dataset. - DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -613,8 +613,7 @@ func (c *datasetServiceV1Client) TestIamPermissions(ctx context.Context, in *v1. return out, nil } -// Server API for DatasetServiceV1 service - +// DatasetServiceV1Server is the server API for DatasetServiceV1 service. type DatasetServiceV1Server interface { // Lists datasets within a project. // diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/operations.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/operations.pb.go index 78ef44cda..78c85fe4c 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/operations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/operations.pb.go @@ -24,29 +24,29 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Metadata describing an [Operation][google.longrunning.Operation]. type OperationMetadata struct { // The Google Cloud Project in which the job is scoped. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The time at which the job was submitted to the Genomics service. - CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The time at which the job began to run. - StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time at which the job stopped running. - EndTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // The original request that started the operation. Note that this will be in // current version of the API. If the operation was started with v1beta2 API // and a GetOperation is performed on v1 API, a v1 request will be returned. - Request *any.Any `protobuf:"bytes,5,opt,name=request" json:"request,omitempty"` + Request *any.Any `protobuf:"bytes,5,opt,name=request,proto3" json:"request,omitempty"` // Optional event messages that were generated during the job's execution. // This also contains any warnings that were generated during import // or export. - Events []*OperationEvent `protobuf:"bytes,6,rep,name=events" json:"events,omitempty"` + Events []*OperationEvent `protobuf:"bytes,6,rep,name=events,proto3" json:"events,omitempty"` // This field is deprecated. Use `labels` instead. Optionally provided by the // caller when submitting the request that creates the operation. - ClientId string `protobuf:"bytes,7,opt,name=client_id,json=clientId" json:"client_id,omitempty"` + ClientId string `protobuf:"bytes,7,opt,name=client_id,json=clientId,proto3" json:"client_id,omitempty"` // Runtime metadata on this Operation. - RuntimeMetadata *any.Any `protobuf:"bytes,8,opt,name=runtime_metadata,json=runtimeMetadata" json:"runtime_metadata,omitempty"` + RuntimeMetadata *any.Any `protobuf:"bytes,8,opt,name=runtime_metadata,json=runtimeMetadata,proto3" json:"runtime_metadata,omitempty"` // Optionally provided by the caller when submitting the request that creates // the operation. - Labels map[string]string `protobuf:"bytes,9,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,9,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -142,12 +142,12 @@ func (m *OperationMetadata) GetLabels() map[string]string { // An event that occurred during an [Operation][google.longrunning.Operation]. type OperationEvent struct { // Optional time of when event started. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Optional time of when event finished. An event can have a start time and no // finish time. If an event has a finish time, there must be a start time. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Required description of event. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/position.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/position.pb.go index e9ef3b300..61d024c94 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/position.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/position.pb.go @@ -25,12 +25,12 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // determination of forward or reverse strand. type Position struct { // The name of the reference in whatever reference set is being used. - ReferenceName string `protobuf:"bytes,1,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,1,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The 0-based offset from the start of the forward strand for that reference. - Position int64 `protobuf:"varint,2,opt,name=position" json:"position,omitempty"` + Position int64 `protobuf:"varint,2,opt,name=position,proto3" json:"position,omitempty"` // Whether this position is on the reverse strand, as opposed to the forward // strand. - ReverseStrand bool `protobuf:"varint,3,opt,name=reverse_strand,json=reverseStrand" json:"reverse_strand,omitempty"` + ReverseStrand bool `protobuf:"varint,3,opt,name=reverse_strand,json=reverseStrand,proto3" json:"reverse_strand,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/range.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/range.pb.go index 35d73ec17..c4d22fb95 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/range.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/range.pb.go @@ -23,11 +23,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Range struct { // The reference sequence name, for example `chr1`, // `1`, or `chrX`. - ReferenceName string `protobuf:"bytes,1,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,1,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The start position of the range on the reference, 0-based inclusive. - Start int64 `protobuf:"varint,2,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,2,opt,name=start,proto3" json:"start,omitempty"` // The end position of the range on the reference, 0-based exclusive. - End int64 `protobuf:"varint,3,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,3,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/readalignment.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/readalignment.pb.go index 78ed325e7..3f615fa48 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/readalignment.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/readalignment.pb.go @@ -24,16 +24,16 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // mapped position and local alignment of the read to the reference. type LinearAlignment struct { // The position of this alignment. - Position *Position `protobuf:"bytes,1,opt,name=position" json:"position,omitempty"` + Position *Position `protobuf:"bytes,1,opt,name=position,proto3" json:"position,omitempty"` // The mapping quality of this alignment. Represents how likely // the read maps to this position as opposed to other locations. // // Specifically, this is -10 log10 Pr(mapping position is wrong), rounded to // the nearest integer. - MappingQuality int32 `protobuf:"varint,2,opt,name=mapping_quality,json=mappingQuality" json:"mapping_quality,omitempty"` + MappingQuality int32 `protobuf:"varint,2,opt,name=mapping_quality,json=mappingQuality,proto3" json:"mapping_quality,omitempty"` // Represents the local alignment of this sequence (alignment matches, indels, // etc) against the reference. - Cigar []*CigarUnit `protobuf:"bytes,3,rep,name=cigar" json:"cigar,omitempty"` + Cigar []*CigarUnit `protobuf:"bytes,3,rep,name=cigar,proto3" json:"cigar,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,42 +169,42 @@ func (m *LinearAlignment) GetCigar() []*CigarUnit { type Read struct { // The server-generated read ID, unique across all reads. This is different // from the `fragmentName`. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The ID of the read group this read belongs to. A read belongs to exactly // one read group. This is a server-generated ID which is distinct from SAM's // RG tag (for that value, see // [ReadGroup.name][google.genomics.v1.ReadGroup.name]). - ReadGroupId string `protobuf:"bytes,2,opt,name=read_group_id,json=readGroupId" json:"read_group_id,omitempty"` + ReadGroupId string `protobuf:"bytes,2,opt,name=read_group_id,json=readGroupId,proto3" json:"read_group_id,omitempty"` // The ID of the read group set this read belongs to. A read belongs to // exactly one read group set. - ReadGroupSetId string `protobuf:"bytes,3,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,3,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` // The fragment name. Equivalent to QNAME (query template name) in SAM. - FragmentName string `protobuf:"bytes,4,opt,name=fragment_name,json=fragmentName" json:"fragment_name,omitempty"` + FragmentName string `protobuf:"bytes,4,opt,name=fragment_name,json=fragmentName,proto3" json:"fragment_name,omitempty"` // The orientation and the distance between reads from the fragment are // consistent with the sequencing protocol (SAM flag 0x2). - ProperPlacement bool `protobuf:"varint,5,opt,name=proper_placement,json=properPlacement" json:"proper_placement,omitempty"` + ProperPlacement bool `protobuf:"varint,5,opt,name=proper_placement,json=properPlacement,proto3" json:"proper_placement,omitempty"` // The fragment is a PCR or optical duplicate (SAM flag 0x400). - DuplicateFragment bool `protobuf:"varint,6,opt,name=duplicate_fragment,json=duplicateFragment" json:"duplicate_fragment,omitempty"` + DuplicateFragment bool `protobuf:"varint,6,opt,name=duplicate_fragment,json=duplicateFragment,proto3" json:"duplicate_fragment,omitempty"` // The observed length of the fragment, equivalent to TLEN in SAM. - FragmentLength int32 `protobuf:"varint,7,opt,name=fragment_length,json=fragmentLength" json:"fragment_length,omitempty"` + FragmentLength int32 `protobuf:"varint,7,opt,name=fragment_length,json=fragmentLength,proto3" json:"fragment_length,omitempty"` // The read number in sequencing. 0-based and less than numberReads. This // field replaces SAM flag 0x40 and 0x80. - ReadNumber int32 `protobuf:"varint,8,opt,name=read_number,json=readNumber" json:"read_number,omitempty"` + ReadNumber int32 `protobuf:"varint,8,opt,name=read_number,json=readNumber,proto3" json:"read_number,omitempty"` // The number of reads in the fragment (extension to SAM flag 0x1). - NumberReads int32 `protobuf:"varint,9,opt,name=number_reads,json=numberReads" json:"number_reads,omitempty"` + NumberReads int32 `protobuf:"varint,9,opt,name=number_reads,json=numberReads,proto3" json:"number_reads,omitempty"` // Whether this read did not pass filters, such as platform or vendor quality // controls (SAM flag 0x200). - FailedVendorQualityChecks bool `protobuf:"varint,10,opt,name=failed_vendor_quality_checks,json=failedVendorQualityChecks" json:"failed_vendor_quality_checks,omitempty"` + FailedVendorQualityChecks bool `protobuf:"varint,10,opt,name=failed_vendor_quality_checks,json=failedVendorQualityChecks,proto3" json:"failed_vendor_quality_checks,omitempty"` // The linear alignment for this alignment record. This field is null for // unmapped reads. - Alignment *LinearAlignment `protobuf:"bytes,11,opt,name=alignment" json:"alignment,omitempty"` + Alignment *LinearAlignment `protobuf:"bytes,11,opt,name=alignment,proto3" json:"alignment,omitempty"` // Whether this alignment is secondary. Equivalent to SAM flag 0x100. // A secondary alignment represents an alternative to the primary alignment // for this read. Aligners may return secondary alignments if a read can map // ambiguously to multiple coordinates in the genome. By convention, each read // has one and only one alignment where both `secondaryAlignment` // and `supplementaryAlignment` are false. - SecondaryAlignment bool `protobuf:"varint,12,opt,name=secondary_alignment,json=secondaryAlignment" json:"secondary_alignment,omitempty"` + SecondaryAlignment bool `protobuf:"varint,12,opt,name=secondary_alignment,json=secondaryAlignment,proto3" json:"secondary_alignment,omitempty"` // Whether this alignment is supplementary. Equivalent to SAM flag 0x800. // Supplementary alignments are used in the representation of a chimeric // alignment. In a chimeric alignment, a read is split into multiple @@ -216,7 +216,7 @@ type Read struct { // will be hard clipped. The `alignedSequence` and // `alignedQuality` fields in the alignment record will only // represent the bases for its respective linear alignment. - SupplementaryAlignment bool `protobuf:"varint,13,opt,name=supplementary_alignment,json=supplementaryAlignment" json:"supplementary_alignment,omitempty"` + SupplementaryAlignment bool `protobuf:"varint,13,opt,name=supplementary_alignment,json=supplementaryAlignment,proto3" json:"supplementary_alignment,omitempty"` // The bases of the read sequence contained in this alignment record, // **without CIGAR operations applied** (equivalent to SEQ in SAM). // `alignedSequence` and `alignedQuality` may be @@ -224,7 +224,7 @@ type Read struct { // alignment is part of a chimeric alignment, or if the read was trimmed. When // this occurs, the CIGAR for this read will begin/end with a hard clip // operator that will indicate the length of the excised sequence. - AlignedSequence string `protobuf:"bytes,14,opt,name=aligned_sequence,json=alignedSequence" json:"aligned_sequence,omitempty"` + AlignedSequence string `protobuf:"bytes,14,opt,name=aligned_sequence,json=alignedSequence,proto3" json:"aligned_sequence,omitempty"` // The quality of the read sequence contained in this alignment record // (equivalent to QUAL in SAM). // `alignedSequence` and `alignedQuality` may be shorter than the full read @@ -232,14 +232,14 @@ type Read struct { // chimeric alignment, or if the read was trimmed. When this occurs, the CIGAR // for this read will begin/end with a hard clip operator that will indicate // the length of the excised sequence. - AlignedQuality []int32 `protobuf:"varint,15,rep,packed,name=aligned_quality,json=alignedQuality" json:"aligned_quality,omitempty"` + AlignedQuality []int32 `protobuf:"varint,15,rep,packed,name=aligned_quality,json=alignedQuality,proto3" json:"aligned_quality,omitempty"` // The mapping of the primary alignment of the // `(readNumber+1)%numberReads` read in the fragment. It replaces // mate position and mate strand in SAM. - NextMatePosition *Position `protobuf:"bytes,16,opt,name=next_mate_position,json=nextMatePosition" json:"next_mate_position,omitempty"` + NextMatePosition *Position `protobuf:"bytes,16,opt,name=next_mate_position,json=nextMatePosition,proto3" json:"next_mate_position,omitempty"` // A map of additional read alignment information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,17,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,17,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroup.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroup.pb.go index 22db0f91e..a047f6980 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroup.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroup.pb.go @@ -25,30 +25,30 @@ type ReadGroup struct { // The server-generated read group ID, unique for all read groups. // Note: This is different than the @RG ID field in the SAM spec. For that // value, see [name][google.genomics.v1.ReadGroup.name]. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The dataset to which this read group belongs. - DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The read group name. This corresponds to the @RG ID field in the SAM spec. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // A free-form text description of this read group. - Description string `protobuf:"bytes,4,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` // A client-supplied sample identifier for the reads in this read group. - SampleId string `protobuf:"bytes,5,opt,name=sample_id,json=sampleId" json:"sample_id,omitempty"` + SampleId string `protobuf:"bytes,5,opt,name=sample_id,json=sampleId,proto3" json:"sample_id,omitempty"` // The experiment used to generate this read group. - Experiment *ReadGroup_Experiment `protobuf:"bytes,6,opt,name=experiment" json:"experiment,omitempty"` + Experiment *ReadGroup_Experiment `protobuf:"bytes,6,opt,name=experiment,proto3" json:"experiment,omitempty"` // The predicted insert size of this read group. The insert size is the length // the sequenced DNA fragment from end-to-end, not including the adapters. - PredictedInsertSize int32 `protobuf:"varint,7,opt,name=predicted_insert_size,json=predictedInsertSize" json:"predicted_insert_size,omitempty"` + PredictedInsertSize int32 `protobuf:"varint,7,opt,name=predicted_insert_size,json=predictedInsertSize,proto3" json:"predicted_insert_size,omitempty"` // The programs used to generate this read group. Programs are always // identical for all read groups within a read group set. For this reason, // only the first read group in a returned set will have this field // populated. - Programs []*ReadGroup_Program `protobuf:"bytes,10,rep,name=programs" json:"programs,omitempty"` + Programs []*ReadGroup_Program `protobuf:"bytes,10,rep,name=programs,proto3" json:"programs,omitempty"` // The reference set the reads in this read group are aligned to. - ReferenceSetId string `protobuf:"bytes,11,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,11,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // A map of additional read group information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,12,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,12,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -153,16 +153,16 @@ type ReadGroup_Experiment struct { // fragments which have been prepared for sequencing from a sample. This // field is important for quality control as error or bias can be introduced // during sample preparation. - LibraryId string `protobuf:"bytes,1,opt,name=library_id,json=libraryId" json:"library_id,omitempty"` + LibraryId string `protobuf:"bytes,1,opt,name=library_id,json=libraryId,proto3" json:"library_id,omitempty"` // The platform unit used as part of this experiment, for example // flowcell-barcode.lane for Illumina or slide for SOLiD. Corresponds to the // @RG PU field in the SAM spec. - PlatformUnit string `protobuf:"bytes,2,opt,name=platform_unit,json=platformUnit" json:"platform_unit,omitempty"` + PlatformUnit string `protobuf:"bytes,2,opt,name=platform_unit,json=platformUnit,proto3" json:"platform_unit,omitempty"` // The sequencing center used as part of this experiment. - SequencingCenter string `protobuf:"bytes,3,opt,name=sequencing_center,json=sequencingCenter" json:"sequencing_center,omitempty"` + SequencingCenter string `protobuf:"bytes,3,opt,name=sequencing_center,json=sequencingCenter,proto3" json:"sequencing_center,omitempty"` // The instrument model used as part of this experiment. This maps to // sequencing technology in the SAM spec. - InstrumentModel string `protobuf:"bytes,4,opt,name=instrument_model,json=instrumentModel" json:"instrument_model,omitempty"` + InstrumentModel string `protobuf:"bytes,4,opt,name=instrument_model,json=instrumentModel,proto3" json:"instrument_model,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -222,17 +222,17 @@ func (m *ReadGroup_Experiment) GetInstrumentModel() string { type ReadGroup_Program struct { // The command line used to run this program. - CommandLine string `protobuf:"bytes,1,opt,name=command_line,json=commandLine" json:"command_line,omitempty"` + CommandLine string `protobuf:"bytes,1,opt,name=command_line,json=commandLine,proto3" json:"command_line,omitempty"` // The user specified locally unique ID of the program. Used along with // `prevProgramId` to define an ordering between programs. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // The display name of the program. This is typically the colloquial name of // the tool used, for example 'bwa' or 'picard'. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The ID of the program run before this one. - PrevProgramId string `protobuf:"bytes,4,opt,name=prev_program_id,json=prevProgramId" json:"prev_program_id,omitempty"` + PrevProgramId string `protobuf:"bytes,4,opt,name=prev_program_id,json=prevProgramId,proto3" json:"prev_program_id,omitempty"` // The version of the program run. - Version string `protobuf:"bytes,5,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,5,opt,name=version,proto3" json:"version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroupset.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroupset.pb.go index 3843c10d3..ca22d42cd 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroupset.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/readgroupset.pb.go @@ -33,21 +33,21 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Genomics](https://cloud.google.com/genomics/fundamentals-of-google-genomics) type ReadGroupSet struct { // The server-generated read group set ID, unique for all read group sets. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The dataset to which this read group set belongs. - DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The reference set to which the reads in this read group set are aligned. - ReferenceSetId string `protobuf:"bytes,3,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,3,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // The read group set name. By default this will be initialized to the sample // name of the sequenced data contained in this set. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // The filename of the original source file for this read group set, if any. - Filename string `protobuf:"bytes,5,opt,name=filename" json:"filename,omitempty"` + Filename string `protobuf:"bytes,5,opt,name=filename,proto3" json:"filename,omitempty"` // The read groups in this set. There are typically 1-10 read groups in a read // group set. - ReadGroups []*ReadGroup `protobuf:"bytes,6,rep,name=read_groups,json=readGroups" json:"read_groups,omitempty"` + ReadGroups []*ReadGroup `protobuf:"bytes,6,rep,name=read_groups,json=readGroups,proto3" json:"read_groups,omitempty"` // A map of additional read group set information. - Info map[string]*_struct.ListValue `protobuf:"bytes,7,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,7,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/reads.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/reads.pb.go index 1bbb94807..81c78752c 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/reads.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/reads.pb.go @@ -68,17 +68,17 @@ func (ImportReadGroupSetsRequest_PartitionStrategy) EnumDescriptor() ([]byte, [] type SearchReadGroupSetsRequest struct { // Restricts this query to read group sets within the given datasets. At least // one ID must be provided. - DatasetIds []string `protobuf:"bytes,1,rep,name=dataset_ids,json=datasetIds" json:"dataset_ids,omitempty"` + DatasetIds []string `protobuf:"bytes,1,rep,name=dataset_ids,json=datasetIds,proto3" json:"dataset_ids,omitempty"` // Only return read group sets for which a substring of the name matches this // string. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 256. The maximum value is 1024. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -139,11 +139,11 @@ func (m *SearchReadGroupSetsRequest) GetPageSize() int32 { // The read group set search response. type SearchReadGroupSetsResponse struct { // The list of matching read group sets. - ReadGroupSets []*ReadGroupSet `protobuf:"bytes,1,rep,name=read_group_sets,json=readGroupSets" json:"read_group_sets,omitempty"` + ReadGroupSets []*ReadGroupSet `protobuf:"bytes,1,rep,name=read_group_sets,json=readGroupSets,proto3" json:"read_group_sets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -191,12 +191,12 @@ func (m *SearchReadGroupSetsResponse) GetNextPageToken() string { type ImportReadGroupSetsRequest struct { // Required. The ID of the dataset these read group sets will belong to. The // caller must have WRITE permissions to this dataset. - DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The reference set to which the imported read group sets are aligned to, if // any. The reference names of this reference set must be a superset of those // found in the imported file headers. If no reference set id is provided, a // best effort is made to associate with a matching reference set. - ReferenceSetId string `protobuf:"bytes,4,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,4,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // A list of URIs pointing at [BAM // files](https://samtools.github.io/hts-specs/SAMv1.pdf) // in Google Cloud Storage. @@ -207,10 +207,10 @@ type ImportReadGroupSetsRequest struct { // consistent: files added may be not be immediately visible to // everyone. Thus, if using a wildcard it is preferable not to start // the import immediately after the files are created. - SourceUris []string `protobuf:"bytes,2,rep,name=source_uris,json=sourceUris" json:"source_uris,omitempty"` + SourceUris []string `protobuf:"bytes,2,rep,name=source_uris,json=sourceUris,proto3" json:"source_uris,omitempty"` // The partition strategy describes how read groups are partitioned into read // group sets. - PartitionStrategy ImportReadGroupSetsRequest_PartitionStrategy `protobuf:"varint,5,opt,name=partition_strategy,json=partitionStrategy,enum=google.genomics.v1.ImportReadGroupSetsRequest_PartitionStrategy" json:"partition_strategy,omitempty"` + PartitionStrategy ImportReadGroupSetsRequest_PartitionStrategy `protobuf:"varint,5,opt,name=partition_strategy,json=partitionStrategy,proto3,enum=google.genomics.v1.ImportReadGroupSetsRequest_PartitionStrategy" json:"partition_strategy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -271,7 +271,7 @@ func (m *ImportReadGroupSetsRequest) GetPartitionStrategy() ImportReadGroupSetsR // The read group set import response. type ImportReadGroupSetsResponse struct { // IDs of the read group sets that were created. - ReadGroupSetIds []string `protobuf:"bytes,1,rep,name=read_group_set_ids,json=readGroupSetIds" json:"read_group_set_ids,omitempty"` + ReadGroupSetIds []string `protobuf:"bytes,1,rep,name=read_group_set_ids,json=readGroupSetIds,proto3" json:"read_group_set_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -312,18 +312,18 @@ func (m *ImportReadGroupSetsResponse) GetReadGroupSetIds() []string { type ExportReadGroupSetRequest struct { // Required. The Google Cloud project ID that owns this // export. The caller must have WRITE access to this project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. A Google Cloud Storage URI for the exported BAM file. // The currently authenticated user must have write access to the new file. // An error will be returned if the URI already contains data. - ExportUri string `protobuf:"bytes,2,opt,name=export_uri,json=exportUri" json:"export_uri,omitempty"` + ExportUri string `protobuf:"bytes,2,opt,name=export_uri,json=exportUri,proto3" json:"export_uri,omitempty"` // Required. The ID of the read group set to export. The caller must have // READ access to this read group set. - ReadGroupSetId string `protobuf:"bytes,3,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,3,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` // The reference names to export. If this is not specified, all reference // sequences, including unmapped reads, are exported. // Use `*` to export only unmapped reads. - ReferenceNames []string `protobuf:"bytes,4,rep,name=reference_names,json=referenceNames" json:"reference_names,omitempty"` + ReferenceNames []string `protobuf:"bytes,4,rep,name=reference_names,json=referenceNames,proto3" json:"reference_names,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -384,10 +384,10 @@ func (m *ExportReadGroupSetRequest) GetReferenceNames() []string { type UpdateReadGroupSetRequest struct { // The ID of the read group set to be updated. The caller must have WRITE // permissions to the dataset associated with this read group set. - ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` // The new read group set data. See `updateMask` for details on mutability of // fields. - ReadGroupSet *ReadGroupSet `protobuf:"bytes,2,opt,name=read_group_set,json=readGroupSet" json:"read_group_set,omitempty"` + ReadGroupSet *ReadGroupSet `protobuf:"bytes,2,opt,name=read_group_set,json=readGroupSet,proto3" json:"read_group_set,omitempty"` // An optional mask specifying which fields to update. Supported fields: // // * [name][google.genomics.v1.ReadGroupSet.name]. @@ -395,7 +395,7 @@ type UpdateReadGroupSetRequest struct { // // Leaving `updateMask` unset is equivalent to specifying all mutable // fields. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -449,7 +449,7 @@ func (m *UpdateReadGroupSetRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteReadGroupSetRequest struct { // The ID of the read group set to be deleted. The caller must have WRITE // permissions to the dataset associated with this read group set. - ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -488,7 +488,7 @@ func (m *DeleteReadGroupSetRequest) GetReadGroupSetId() string { type GetReadGroupSetRequest struct { // The ID of the read group set. - ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -527,17 +527,17 @@ func (m *GetReadGroupSetRequest) GetReadGroupSetId() string { type ListCoverageBucketsRequest struct { // Required. The ID of the read group set over which coverage is requested. - ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,1,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` // The name of the reference to query, within the reference set associated // with this query. Optional. - ReferenceName string `protobuf:"bytes,3,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,3,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The start position of the range on the reference, 0-based inclusive. If // specified, `referenceName` must also be specified. Defaults to 0. - Start int64 `protobuf:"varint,4,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,4,opt,name=start,proto3" json:"start,omitempty"` // The end position of the range on the reference, 0-based exclusive. If // specified, `referenceName` must also be specified. If unset or 0, defaults // to the length of the reference. - End int64 `protobuf:"varint,5,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,5,opt,name=end,proto3" json:"end,omitempty"` // The desired width of each reported coverage bucket in base pairs. This // will be rounded down to the nearest precomputed bucket width; the value // of which is returned as `bucketWidth` in the response. Defaults @@ -545,14 +545,14 @@ type ListCoverageBucketsRequest struct { // of the target range, if specified. The smallest precomputed // `bucketWidth` is currently 2048 base pairs; this is subject to // change. - TargetBucketWidth int64 `protobuf:"varint,6,opt,name=target_bucket_width,json=targetBucketWidth" json:"target_bucket_width,omitempty"` + TargetBucketWidth int64 `protobuf:"varint,6,opt,name=target_bucket_width,json=targetBucketWidth,proto3" json:"target_bucket_width,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,7,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,7,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 1024. The maximum value is 2048. - PageSize int32 `protobuf:"varint,8,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,8,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -635,10 +635,10 @@ func (m *ListCoverageBucketsRequest) GetPageSize() int32 { // to a specific range of the reference sequence. type CoverageBucket struct { // The genomic coordinate range spanned by this bucket. - Range *Range `protobuf:"bytes,1,opt,name=range" json:"range,omitempty"` + Range *Range `protobuf:"bytes,1,opt,name=range,proto3" json:"range,omitempty"` // The average number of reads which are aligned to each individual // reference base in this bucket. - MeanCoverage float32 `protobuf:"fixed32,2,opt,name=mean_coverage,json=meanCoverage" json:"mean_coverage,omitempty"` + MeanCoverage float32 `protobuf:"fixed32,2,opt,name=mean_coverage,json=meanCoverage,proto3" json:"mean_coverage,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -687,16 +687,16 @@ type ListCoverageBucketsResponse struct { // end of a reference sequence may be shorter. This value is omitted if the // bucket width is infinity (the default behaviour, with no range or // `targetBucketWidth`). - BucketWidth int64 `protobuf:"varint,1,opt,name=bucket_width,json=bucketWidth" json:"bucket_width,omitempty"` + BucketWidth int64 `protobuf:"varint,1,opt,name=bucket_width,json=bucketWidth,proto3" json:"bucket_width,omitempty"` // The coverage buckets. The list of buckets is sparse; a bucket with 0 // overlapping reads is not returned. A bucket never crosses more than one // reference sequence. Each bucket has width `bucketWidth`, unless // its end is the end of the reference sequence. - CoverageBuckets []*CoverageBucket `protobuf:"bytes,2,rep,name=coverage_buckets,json=coverageBuckets" json:"coverage_buckets,omitempty"` + CoverageBuckets []*CoverageBucket `protobuf:"bytes,2,rep,name=coverage_buckets,json=coverageBuckets,proto3" json:"coverage_buckets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -753,28 +753,28 @@ type SearchReadsRequest struct { // specified read group sets must be aligned against a common set of reference // sequences; this defines the genomic coordinates for the query. Must specify // one of `readGroupSetIds` or `readGroupIds`. - ReadGroupSetIds []string `protobuf:"bytes,1,rep,name=read_group_set_ids,json=readGroupSetIds" json:"read_group_set_ids,omitempty"` + ReadGroupSetIds []string `protobuf:"bytes,1,rep,name=read_group_set_ids,json=readGroupSetIds,proto3" json:"read_group_set_ids,omitempty"` // The IDs of the read groups within which to search for reads. All specified // read groups must belong to the same read group sets. Must specify one of // `readGroupSetIds` or `readGroupIds`. - ReadGroupIds []string `protobuf:"bytes,5,rep,name=read_group_ids,json=readGroupIds" json:"read_group_ids,omitempty"` + ReadGroupIds []string `protobuf:"bytes,5,rep,name=read_group_ids,json=readGroupIds,proto3" json:"read_group_ids,omitempty"` // The reference sequence name, for example `chr1`, `1`, or `chrX`. If set to // `*`, only unmapped reads are returned. If unspecified, all reads (mapped // and unmapped) are returned. - ReferenceName string `protobuf:"bytes,7,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,7,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The start position of the range on the reference, 0-based inclusive. If // specified, `referenceName` must also be specified. - Start int64 `protobuf:"varint,8,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,8,opt,name=start,proto3" json:"start,omitempty"` // The end position of the range on the reference, 0-based exclusive. If // specified, `referenceName` must also be specified. - End int64 `protobuf:"varint,9,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,9,opt,name=end,proto3" json:"end,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 256. The maximum value is 2048. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -859,11 +859,11 @@ type SearchReadsResponse struct { // if any, ascending in position within the same reference. Unmapped reads, // which have no position, are returned contiguously and are sorted in // ascending lexicographic order by fragment name. - Alignments []*Read `protobuf:"bytes,1,rep,name=alignments" json:"alignments,omitempty"` + Alignments []*Read `protobuf:"bytes,1,rep,name=alignments,proto3" json:"alignments,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -912,19 +912,19 @@ type StreamReadsRequest struct { // The Google Cloud project ID which will be billed // for this access. The caller must have WRITE access to this project. // Required. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The ID of the read group set from which to stream reads. - ReadGroupSetId string `protobuf:"bytes,2,opt,name=read_group_set_id,json=readGroupSetId" json:"read_group_set_id,omitempty"` + ReadGroupSetId string `protobuf:"bytes,2,opt,name=read_group_set_id,json=readGroupSetId,proto3" json:"read_group_set_id,omitempty"` // The reference sequence name, for example `chr1`, // `1`, or `chrX`. If set to *, only unmapped reads are // returned. - ReferenceName string `protobuf:"bytes,3,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,3,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The start position of the range on the reference, 0-based inclusive. If // specified, `referenceName` must also be specified. - Start int64 `protobuf:"varint,4,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,4,opt,name=start,proto3" json:"start,omitempty"` // The end position of the range on the reference, 0-based exclusive. If // specified, `referenceName` must also be specified. - End int64 `protobuf:"varint,5,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,5,opt,name=end,proto3" json:"end,omitempty"` // Restricts results to a shard containing approximately `1/totalShards` // of the normal response payload for this query. Results from a sharded // request are disjoint from those returned by all queries which differ only @@ -932,7 +932,7 @@ type StreamReadsRequest struct { // likely for large values of `totalShards`. // // Valid values are `[0, totalShards)`. - Shard int32 `protobuf:"varint,6,opt,name=shard" json:"shard,omitempty"` + Shard int32 `protobuf:"varint,6,opt,name=shard,proto3" json:"shard,omitempty"` // Specifying `totalShards` causes a disjoint subset of the normal response // payload to be returned for each query with a unique `shard` parameter // specified. A best effort is made to yield equally sized shards. Sharding @@ -945,7 +945,7 @@ type StreamReadsRequest struct { // share shard boundaries. For example, streaming `shard` 2 of 5 // `totalShards` yields the same results as streaming `shard`s 4 and 5 of 10 // `totalShards`. This property can be leveraged for adaptive retries. - TotalShards int32 `protobuf:"varint,7,opt,name=total_shards,json=totalShards" json:"total_shards,omitempty"` + TotalShards int32 `protobuf:"varint,7,opt,name=total_shards,json=totalShards,proto3" json:"total_shards,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1025,7 +1025,7 @@ func (m *StreamReadsRequest) GetTotalShards() int32 { } type StreamReadsResponse struct { - Alignments []*Read `protobuf:"bytes,1,rep,name=alignments" json:"alignments,omitempty"` + Alignments []*Read `protobuf:"bytes,1,rep,name=alignments,proto3" json:"alignments,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1138,8 +1138,7 @@ func (x *streamingReadServiceStreamReadsClient) Recv() (*StreamReadsResponse, er return m, nil } -// Server API for StreamingReadService service - +// StreamingReadServiceServer is the server API for StreamingReadService service. type StreamingReadServiceServer interface { // Returns a stream of all the reads matching the search request, ordered // by reference name, position, and ID. @@ -1369,8 +1368,7 @@ func (c *readServiceV1Client) SearchReads(ctx context.Context, in *SearchReadsRe return out, nil } -// Server API for ReadServiceV1 service - +// ReadServiceV1Server is the server API for ReadServiceV1 service. type ReadServiceV1Server interface { // Creates read group sets by asynchronously importing the provided // information. diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/references.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/references.pb.go index 14441e81d..bfc5a96ed 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/references.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/references.pb.go @@ -33,23 +33,23 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Genomics](https://cloud.google.com/genomics/fundamentals-of-google-genomics) type Reference struct { // The server-generated reference ID, unique across all references. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The length of this reference's sequence. - Length int64 `protobuf:"varint,2,opt,name=length" json:"length,omitempty"` + Length int64 `protobuf:"varint,2,opt,name=length,proto3" json:"length,omitempty"` // MD5 of the upper-case sequence excluding all whitespace characters (this // is equivalent to SQ:M5 in SAM). This value is represented in lower case // hexadecimal format. - Md5Checksum string `protobuf:"bytes,3,opt,name=md5checksum" json:"md5checksum,omitempty"` + Md5Checksum string `protobuf:"bytes,3,opt,name=md5checksum,proto3" json:"md5checksum,omitempty"` // The name of this reference, for example `22`. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // The URI from which the sequence was obtained. Typically specifies a FASTA // format file. - SourceUri string `protobuf:"bytes,5,opt,name=source_uri,json=sourceUri" json:"source_uri,omitempty"` + SourceUri string `protobuf:"bytes,5,opt,name=source_uri,json=sourceUri,proto3" json:"source_uri,omitempty"` // All known corresponding accession IDs in INSDC (GenBank/ENA/DDBJ) ideally // with a version number, for example `GCF_000001405.26`. - SourceAccessions []string `protobuf:"bytes,6,rep,name=source_accessions,json=sourceAccessions" json:"source_accessions,omitempty"` + SourceAccessions []string `protobuf:"bytes,6,rep,name=source_accessions,json=sourceAccessions,proto3" json:"source_accessions,omitempty"` // ID from http://www.ncbi.nlm.nih.gov/taxonomy. For example, 9606 for human. - NcbiTaxonId int32 `protobuf:"varint,7,opt,name=ncbi_taxon_id,json=ncbiTaxonId" json:"ncbi_taxon_id,omitempty"` + NcbiTaxonId int32 `protobuf:"varint,7,opt,name=ncbi_taxon_id,json=ncbiTaxonId,proto3" json:"ncbi_taxon_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -138,31 +138,31 @@ func (m *Reference) GetNcbiTaxonId() int32 { // Genomics](https://cloud.google.com/genomics/fundamentals-of-google-genomics) type ReferenceSet struct { // The server-generated reference set ID, unique across all reference sets. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The IDs of the reference objects that are part of this set. // `Reference.md5checksum` must be unique within this set. - ReferenceIds []string `protobuf:"bytes,2,rep,name=reference_ids,json=referenceIds" json:"reference_ids,omitempty"` + ReferenceIds []string `protobuf:"bytes,2,rep,name=reference_ids,json=referenceIds,proto3" json:"reference_ids,omitempty"` // Order-independent MD5 checksum which identifies this reference set. The // checksum is computed by sorting all lower case hexidecimal string // `reference.md5checksum` (for all reference in this set) in // ascending lexicographic order, concatenating, and taking the MD5 of that // value. The resulting value is represented in lower case hexadecimal format. - Md5Checksum string `protobuf:"bytes,3,opt,name=md5checksum" json:"md5checksum,omitempty"` + Md5Checksum string `protobuf:"bytes,3,opt,name=md5checksum,proto3" json:"md5checksum,omitempty"` // ID from http://www.ncbi.nlm.nih.gov/taxonomy (for example, 9606 for human) // indicating the species which this reference set is intended to model. Note // that contained references may specify a different `ncbiTaxonId`, as // assemblies may contain reference sequences which do not belong to the // modeled species, for example EBV in a human reference genome. - NcbiTaxonId int32 `protobuf:"varint,4,opt,name=ncbi_taxon_id,json=ncbiTaxonId" json:"ncbi_taxon_id,omitempty"` + NcbiTaxonId int32 `protobuf:"varint,4,opt,name=ncbi_taxon_id,json=ncbiTaxonId,proto3" json:"ncbi_taxon_id,omitempty"` // Free text description of this reference set. - Description string `protobuf:"bytes,5,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,5,opt,name=description,proto3" json:"description,omitempty"` // Public id of this reference set, such as `GRCh37`. - AssemblyId string `protobuf:"bytes,6,opt,name=assembly_id,json=assemblyId" json:"assembly_id,omitempty"` + AssemblyId string `protobuf:"bytes,6,opt,name=assembly_id,json=assemblyId,proto3" json:"assembly_id,omitempty"` // The URI from which the references were obtained. - SourceUri string `protobuf:"bytes,7,opt,name=source_uri,json=sourceUri" json:"source_uri,omitempty"` + SourceUri string `protobuf:"bytes,7,opt,name=source_uri,json=sourceUri,proto3" json:"source_uri,omitempty"` // All known corresponding accession IDs in INSDC (GenBank/ENA/DDBJ) ideally // with a version number, for example `NC_000001.11`. - SourceAccessions []string `protobuf:"bytes,8,rep,name=source_accessions,json=sourceAccessions" json:"source_accessions,omitempty"` + SourceAccessions []string `protobuf:"bytes,8,rep,name=source_accessions,json=sourceAccessions,proto3" json:"source_accessions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -251,22 +251,22 @@ func (m *ReferenceSet) GetSourceAccessions() []string { type SearchReferenceSetsRequest struct { // If present, return reference sets for which the // [md5checksum][google.genomics.v1.ReferenceSet.md5checksum] matches exactly. - Md5Checksums []string `protobuf:"bytes,1,rep,name=md5checksums" json:"md5checksums,omitempty"` + Md5Checksums []string `protobuf:"bytes,1,rep,name=md5checksums,proto3" json:"md5checksums,omitempty"` // If present, return reference sets for which a prefix of any of // [sourceAccessions][google.genomics.v1.ReferenceSet.source_accessions] // match any of these strings. Accession numbers typically have a main number // and a version, for example `NC_000001.11`. - Accessions []string `protobuf:"bytes,2,rep,name=accessions" json:"accessions,omitempty"` + Accessions []string `protobuf:"bytes,2,rep,name=accessions,proto3" json:"accessions,omitempty"` // If present, return reference sets for which a substring of their // `assemblyId` matches this string (case insensitive). - AssemblyId string `protobuf:"bytes,3,opt,name=assembly_id,json=assemblyId" json:"assembly_id,omitempty"` + AssemblyId string `protobuf:"bytes,3,opt,name=assembly_id,json=assemblyId,proto3" json:"assembly_id,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 1024. The maximum value is 4096. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -333,11 +333,11 @@ func (m *SearchReferenceSetsRequest) GetPageSize() int32 { type SearchReferenceSetsResponse struct { // The matching references sets. - ReferenceSets []*ReferenceSet `protobuf:"bytes,1,rep,name=reference_sets,json=referenceSets" json:"reference_sets,omitempty"` + ReferenceSets []*ReferenceSet `protobuf:"bytes,1,rep,name=reference_sets,json=referenceSets,proto3" json:"reference_sets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -383,7 +383,7 @@ func (m *SearchReferenceSetsResponse) GetNextPageToken() string { type GetReferenceSetRequest struct { // The ID of the reference set. - ReferenceSetId string `protobuf:"bytes,1,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,1,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -423,21 +423,21 @@ func (m *GetReferenceSetRequest) GetReferenceSetId() string { type SearchReferencesRequest struct { // If present, return references for which the // [md5checksum][google.genomics.v1.Reference.md5checksum] matches exactly. - Md5Checksums []string `protobuf:"bytes,1,rep,name=md5checksums" json:"md5checksums,omitempty"` + Md5Checksums []string `protobuf:"bytes,1,rep,name=md5checksums,proto3" json:"md5checksums,omitempty"` // If present, return references for which a prefix of any of // [sourceAccessions][google.genomics.v1.Reference.source_accessions] match // any of these strings. Accession numbers typically have a main number and a // version, for example `GCF_000001405.26`. - Accessions []string `protobuf:"bytes,2,rep,name=accessions" json:"accessions,omitempty"` + Accessions []string `protobuf:"bytes,2,rep,name=accessions,proto3" json:"accessions,omitempty"` // If present, return only references which belong to this reference set. - ReferenceSetId string `protobuf:"bytes,3,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,3,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 1024. The maximum value is 4096. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -504,11 +504,11 @@ func (m *SearchReferencesRequest) GetPageSize() int32 { type SearchReferencesResponse struct { // The matching references. - References []*Reference `protobuf:"bytes,1,rep,name=references" json:"references,omitempty"` + References []*Reference `protobuf:"bytes,1,rep,name=references,proto3" json:"references,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -554,7 +554,7 @@ func (m *SearchReferencesResponse) GetNextPageToken() string { type GetReferenceRequest struct { // The ID of the reference. - ReferenceId string `protobuf:"bytes,1,opt,name=reference_id,json=referenceId" json:"reference_id,omitempty"` + ReferenceId string `protobuf:"bytes,1,opt,name=reference_id,json=referenceId,proto3" json:"reference_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -593,20 +593,20 @@ func (m *GetReferenceRequest) GetReferenceId() string { type ListBasesRequest struct { // The ID of the reference. - ReferenceId string `protobuf:"bytes,1,opt,name=reference_id,json=referenceId" json:"reference_id,omitempty"` + ReferenceId string `protobuf:"bytes,1,opt,name=reference_id,json=referenceId,proto3" json:"reference_id,omitempty"` // The start position (0-based) of this query. Defaults to 0. - Start int64 `protobuf:"varint,2,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,2,opt,name=start,proto3" json:"start,omitempty"` // The end position (0-based, exclusive) of this query. Defaults to the length // of this reference. - End int64 `protobuf:"varint,3,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,3,opt,name=end,proto3" json:"end,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of bases to return in a single page. If unspecified, // defaults to 200Kbp (kilo base pairs). The maximum value is 10Mbp (mega base // pairs). - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -675,13 +675,13 @@ type ListBasesResponse struct { // The offset position (0-based) of the given `sequence` from the // start of this `Reference`. This value will differ for each page // in a paginated request. - Offset int64 `protobuf:"varint,1,opt,name=offset" json:"offset,omitempty"` + Offset int64 `protobuf:"varint,1,opt,name=offset,proto3" json:"offset,omitempty"` // A substring of the bases that make up this reference. - Sequence string `protobuf:"bytes,2,opt,name=sequence" json:"sequence,omitempty"` + Sequence string `protobuf:"bytes,2,opt,name=sequence,proto3" json:"sequence,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,3,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -857,8 +857,7 @@ func (c *referenceServiceV1Client) ListBases(ctx context.Context, in *ListBasesR return out, nil } -// Server API for ReferenceServiceV1 service - +// ReferenceServiceV1Server is the server API for ReferenceServiceV1 service. type ReferenceServiceV1Server interface { // Searches for reference sets which match the given criteria. // diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1/variants.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1/variants.pb.go index 92032eafd..3b338b3bd 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1/variants.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1/variants.pb.go @@ -159,24 +159,24 @@ func (ExportVariantSetRequest_Format) EnumDescriptor() ([]byte, []int) { // Value and info are mutually exclusive. type VariantSetMetadata struct { // The top-level key. - Key string `protobuf:"bytes,1,opt,name=key" json:"key,omitempty"` + Key string `protobuf:"bytes,1,opt,name=key,proto3" json:"key,omitempty"` // The value field for simple metadata - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // User-provided ID field, not enforced by this API. // Two or more pieces of structured metadata with identical // id and key fields are considered equivalent. - Id string `protobuf:"bytes,4,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,4,opt,name=id,proto3" json:"id,omitempty"` // The type of data. Possible types include: Integer, Float, // Flag, Character, and String. - Type VariantSetMetadata_Type `protobuf:"varint,5,opt,name=type,enum=google.genomics.v1.VariantSetMetadata_Type" json:"type,omitempty"` + Type VariantSetMetadata_Type `protobuf:"varint,5,opt,name=type,proto3,enum=google.genomics.v1.VariantSetMetadata_Type" json:"type,omitempty"` // The number of values that can be included in a field described by this // metadata. - Number string `protobuf:"bytes,8,opt,name=number" json:"number,omitempty"` + Number string `protobuf:"bytes,8,opt,name=number,proto3" json:"number,omitempty"` // A textual description of this metadata. - Description string `protobuf:"bytes,7,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,7,opt,name=description,proto3" json:"description,omitempty"` // Remaining structured metadata key-value pairs. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,3,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,3,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,9 +262,9 @@ func (m *VariantSetMetadata) GetInfo() map[string]*_struct.ListValue { // Genomics](https://cloud.google.com/genomics/fundamentals-of-google-genomics) type VariantSet struct { // The dataset to which this variant set belongs. - DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,1,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // The server-generated variant set ID, unique across all variant sets. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // The reference set to which the variant set is mapped. The reference set // describes the alignment provenance of the variant set, while the // `referenceBounds` describe the shape of the actual variant data. The @@ -275,16 +275,16 @@ type VariantSet struct { // and contains a single variant on reference 'X', `referenceBounds` would // contain only an entry for 'X', while the associated reference set // enumerates all possible references: '1', '2', 'X', 'Y', 'MT', etc. - ReferenceSetId string `protobuf:"bytes,6,opt,name=reference_set_id,json=referenceSetId" json:"reference_set_id,omitempty"` + ReferenceSetId string `protobuf:"bytes,6,opt,name=reference_set_id,json=referenceSetId,proto3" json:"reference_set_id,omitempty"` // A list of all references used by the variants in a variant set // with associated coordinate upper bounds for each one. - ReferenceBounds []*ReferenceBound `protobuf:"bytes,5,rep,name=reference_bounds,json=referenceBounds" json:"reference_bounds,omitempty"` + ReferenceBounds []*ReferenceBound `protobuf:"bytes,5,rep,name=reference_bounds,json=referenceBounds,proto3" json:"reference_bounds,omitempty"` // The metadata associated with this variant set. - Metadata []*VariantSetMetadata `protobuf:"bytes,4,rep,name=metadata" json:"metadata,omitempty"` + Metadata []*VariantSetMetadata `protobuf:"bytes,4,rep,name=metadata,proto3" json:"metadata,omitempty"` // User-specified, mutable name. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` // A textual description of this variant set. - Description string `protobuf:"bytes,8,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,8,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -377,41 +377,41 @@ func (m *VariantSet) GetDescription() string { // sample. type Variant struct { // The ID of the variant set this variant belongs to. - VariantSetId string `protobuf:"bytes,15,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,15,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` // The server-generated variant ID, unique across all variants. - Id string `protobuf:"bytes,2,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,2,opt,name=id,proto3" json:"id,omitempty"` // Names for the variant, for example a RefSNP ID. - Names []string `protobuf:"bytes,3,rep,name=names" json:"names,omitempty"` + Names []string `protobuf:"bytes,3,rep,name=names,proto3" json:"names,omitempty"` // The date this variant was created, in milliseconds from the epoch. - Created int64 `protobuf:"varint,12,opt,name=created" json:"created,omitempty"` + Created int64 `protobuf:"varint,12,opt,name=created,proto3" json:"created,omitempty"` // The reference on which this variant occurs. // (such as `chr20` or `X`) - ReferenceName string `protobuf:"bytes,14,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,14,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The position at which this variant occurs (0-based). // This corresponds to the first base of the string of reference bases. - Start int64 `protobuf:"varint,16,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,16,opt,name=start,proto3" json:"start,omitempty"` // The end position (0-based) of this variant. This corresponds to the first // base after the last base in the reference allele. So, the length of // the reference allele is (end - start). This is useful for variants // that don't explicitly give alternate bases, for example large deletions. - End int64 `protobuf:"varint,13,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,13,opt,name=end,proto3" json:"end,omitempty"` // The reference bases for this variant. They start at the given // position. - ReferenceBases string `protobuf:"bytes,6,opt,name=reference_bases,json=referenceBases" json:"reference_bases,omitempty"` + ReferenceBases string `protobuf:"bytes,6,opt,name=reference_bases,json=referenceBases,proto3" json:"reference_bases,omitempty"` // The bases that appear instead of the reference bases. - AlternateBases []string `protobuf:"bytes,7,rep,name=alternate_bases,json=alternateBases" json:"alternate_bases,omitempty"` + AlternateBases []string `protobuf:"bytes,7,rep,name=alternate_bases,json=alternateBases,proto3" json:"alternate_bases,omitempty"` // A measure of how likely this variant is to be real. // A higher value is better. - Quality float64 `protobuf:"fixed64,8,opt,name=quality" json:"quality,omitempty"` + Quality float64 `protobuf:"fixed64,8,opt,name=quality,proto3" json:"quality,omitempty"` // A list of filters (normally quality filters) this variant has failed. // `PASS` indicates this variant has passed all filters. - Filter []string `protobuf:"bytes,9,rep,name=filter" json:"filter,omitempty"` + Filter []string `protobuf:"bytes,9,rep,name=filter,proto3" json:"filter,omitempty"` // A map of additional variant information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,10,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,10,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The variant calls for this particular variant. Each one represents the // determination of genotype with respect to this variant. - Calls []*VariantCall `protobuf:"bytes,11,rep,name=calls" json:"calls,omitempty"` + Calls []*VariantCall `protobuf:"bytes,11,rep,name=calls,proto3" json:"calls,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -538,9 +538,9 @@ func (m *Variant) GetCalls() []*VariantCall { // a SNP named rs1234 in a call set with the name NA12345. type VariantCall struct { // The ID of the call set this variant call belongs to. - CallSetId string `protobuf:"bytes,8,opt,name=call_set_id,json=callSetId" json:"call_set_id,omitempty"` + CallSetId string `protobuf:"bytes,8,opt,name=call_set_id,json=callSetId,proto3" json:"call_set_id,omitempty"` // The name of the call set this variant call belongs to. - CallSetName string `protobuf:"bytes,9,opt,name=call_set_name,json=callSetName" json:"call_set_name,omitempty"` + CallSetName string `protobuf:"bytes,9,opt,name=call_set_name,json=callSetName,proto3" json:"call_set_name,omitempty"` // The genotype of this variant call. Each value represents either the value // of the `referenceBases` field or a 1-based index into // `alternateBases`. If a variant had a `referenceBases` @@ -553,23 +553,23 @@ type VariantCall struct { // genotype values is important if the `phaseset` is present. // If a genotype is not called (that is, a `.` is present in the // GT string) -1 is returned. - Genotype []int32 `protobuf:"varint,7,rep,packed,name=genotype" json:"genotype,omitempty"` + Genotype []int32 `protobuf:"varint,7,rep,packed,name=genotype,proto3" json:"genotype,omitempty"` // If this field is present, this variant call's genotype ordering implies // the phase of the bases and is consistent with any other variant calls in // the same reference sequence which have the same phaseset value. // When importing data from VCF, if the genotype data was phased but no // phase set was specified this field will be set to `*`. - Phaseset string `protobuf:"bytes,5,opt,name=phaseset" json:"phaseset,omitempty"` + Phaseset string `protobuf:"bytes,5,opt,name=phaseset,proto3" json:"phaseset,omitempty"` // The genotype likelihoods for this variant call. Each array entry // represents how likely a specific genotype is for this call. The value // ordering is defined by the GL tag in the VCF spec. // If Phred-scaled genotype likelihood scores (PL) are available and // log10(P) genotype likelihood scores (GL) are not, PL scores are converted // to GL scores. If both are available, PL scores are stored in `info`. - GenotypeLikelihood []float64 `protobuf:"fixed64,6,rep,packed,name=genotype_likelihood,json=genotypeLikelihood" json:"genotype_likelihood,omitempty"` + GenotypeLikelihood []float64 `protobuf:"fixed64,6,rep,packed,name=genotype_likelihood,json=genotypeLikelihood,proto3" json:"genotype_likelihood,omitempty"` // A map of additional variant call information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,2,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,2,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -648,22 +648,22 @@ func (m *VariantCall) GetInfo() map[string]*_struct.ListValue { // Genomics](https://cloud.google.com/genomics/fundamentals-of-google-genomics) type CallSet struct { // The server-generated call set ID, unique across all call sets. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // The call set name. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // The sample ID this call set corresponds to. - SampleId string `protobuf:"bytes,7,opt,name=sample_id,json=sampleId" json:"sample_id,omitempty"` + SampleId string `protobuf:"bytes,7,opt,name=sample_id,json=sampleId,proto3" json:"sample_id,omitempty"` // The IDs of the variant sets this call set belongs to. This field must // have exactly length one, as a call set belongs to a single variant set. // This field is repeated for compatibility with the // [GA4GH 0.5.1 // API](https://github.com/ga4gh/schemas/blob/v0.5.1/src/main/resources/avro/variants.avdl#L76). - VariantSetIds []string `protobuf:"bytes,6,rep,name=variant_set_ids,json=variantSetIds" json:"variant_set_ids,omitempty"` + VariantSetIds []string `protobuf:"bytes,6,rep,name=variant_set_ids,json=variantSetIds,proto3" json:"variant_set_ids,omitempty"` // The date this call set was created in milliseconds from the epoch. - Created int64 `protobuf:"varint,5,opt,name=created" json:"created,omitempty"` + Created int64 `protobuf:"varint,5,opt,name=created,proto3" json:"created,omitempty"` // A map of additional call set information. This must be of the form // map (string key mapping to a list of string values). - Info map[string]*_struct.ListValue `protobuf:"bytes,4,rep,name=info" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Info map[string]*_struct.ListValue `protobuf:"bytes,4,rep,name=info,proto3" json:"info,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -739,10 +739,10 @@ func (m *CallSet) GetInfo() map[string]*_struct.ListValue { // variants in a particular reference. type ReferenceBound struct { // The name of the reference associated with this reference bound. - ReferenceName string `protobuf:"bytes,1,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,1,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // An upper bound (inclusive) on the starting coordinate of any // variant in the reference sequence. - UpperBound int64 `protobuf:"varint,2,opt,name=upper_bound,json=upperBound" json:"upper_bound,omitempty"` + UpperBound int64 `protobuf:"varint,2,opt,name=upper_bound,json=upperBound,proto3" json:"upper_bound,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -789,15 +789,15 @@ func (m *ReferenceBound) GetUpperBound() int64 { // The variant data import request. type ImportVariantsRequest struct { // Required. The variant set to which variant data should be imported. - VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` // A list of URIs referencing variant files in Google Cloud Storage. URIs can // include wildcards [as described // here](https://cloud.google.com/storage/docs/gsutil/addlhelp/WildcardNames). // Note that recursive wildcards ('**') are not supported. - SourceUris []string `protobuf:"bytes,2,rep,name=source_uris,json=sourceUris" json:"source_uris,omitempty"` + SourceUris []string `protobuf:"bytes,2,rep,name=source_uris,json=sourceUris,proto3" json:"source_uris,omitempty"` // The format of the variant data being imported. If unspecified, defaults to // to `VCF`. - Format ImportVariantsRequest_Format `protobuf:"varint,3,opt,name=format,enum=google.genomics.v1.ImportVariantsRequest_Format" json:"format,omitempty"` + Format ImportVariantsRequest_Format `protobuf:"varint,3,opt,name=format,proto3,enum=google.genomics.v1.ImportVariantsRequest_Format" json:"format,omitempty"` // Convert reference names to the canonical representation. // hg19 haploytypes (those reference names containing "_hap") // are not modified in any way. @@ -806,11 +806,11 @@ type ImportVariantsRequest struct { // The "chr" prefix is dropped for all autosomes and sex chromsomes. // For example "chr17" becomes "17" and "chrX" becomes "X". // All mitochondrial chromosomes ("chrM", "chrMT", etc) become "MT". - NormalizeReferenceNames bool `protobuf:"varint,5,opt,name=normalize_reference_names,json=normalizeReferenceNames" json:"normalize_reference_names,omitempty"` + NormalizeReferenceNames bool `protobuf:"varint,5,opt,name=normalize_reference_names,json=normalizeReferenceNames,proto3" json:"normalize_reference_names,omitempty"` // A mapping between info field keys and the InfoMergeOperations to // be performed on them. This is plumbed down to the MergeVariantRequests // generated by the resulting import job. - InfoMergeConfig map[string]InfoMergeOperation `protobuf:"bytes,6,rep,name=info_merge_config,json=infoMergeConfig" json:"info_merge_config,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value,enum=google.genomics.v1.InfoMergeOperation"` + InfoMergeConfig map[string]InfoMergeOperation `protobuf:"bytes,6,rep,name=info_merge_config,json=infoMergeConfig,proto3" json:"info_merge_config,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3,enum=google.genomics.v1.InfoMergeOperation"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -878,7 +878,7 @@ func (m *ImportVariantsRequest) GetInfoMergeConfig() map[string]InfoMergeOperati // The variant data import response. type ImportVariantsResponse struct { // IDs of the call sets created during the import. - CallSetIds []string `protobuf:"bytes,1,rep,name=call_set_ids,json=callSetIds" json:"call_set_ids,omitempty"` + CallSetIds []string `protobuf:"bytes,1,rep,name=call_set_ids,json=callSetIds,proto3" json:"call_set_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -918,7 +918,7 @@ func (m *ImportVariantsResponse) GetCallSetIds() []string { // The CreateVariantSet request type CreateVariantSetRequest struct { // Required. The variant set to be created. Must have a valid `datasetId`. - VariantSet *VariantSet `protobuf:"bytes,1,opt,name=variant_set,json=variantSet" json:"variant_set,omitempty"` + VariantSet *VariantSet `protobuf:"bytes,1,opt,name=variant_set,json=variantSet,proto3" json:"variant_set,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -959,23 +959,23 @@ func (m *CreateVariantSetRequest) GetVariantSet() *VariantSet { type ExportVariantSetRequest struct { // Required. The ID of the variant set that contains variant data which // should be exported. The caller must have READ access to this variant set. - VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` // If provided, only variant call information from the specified call sets // will be exported. By default all variant calls are exported. - CallSetIds []string `protobuf:"bytes,2,rep,name=call_set_ids,json=callSetIds" json:"call_set_ids,omitempty"` + CallSetIds []string `protobuf:"bytes,2,rep,name=call_set_ids,json=callSetIds,proto3" json:"call_set_ids,omitempty"` // Required. The Google Cloud project ID that owns the destination // BigQuery dataset. The caller must have WRITE access to this project. This // project will also own the resulting export job. - ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The format for the exported data. - Format ExportVariantSetRequest_Format `protobuf:"varint,4,opt,name=format,enum=google.genomics.v1.ExportVariantSetRequest_Format" json:"format,omitempty"` + Format ExportVariantSetRequest_Format `protobuf:"varint,4,opt,name=format,proto3,enum=google.genomics.v1.ExportVariantSetRequest_Format" json:"format,omitempty"` // Required. The BigQuery dataset to export data to. This dataset must already // exist. Note that this is distinct from the Genomics concept of "dataset". - BigqueryDataset string `protobuf:"bytes,5,opt,name=bigquery_dataset,json=bigqueryDataset" json:"bigquery_dataset,omitempty"` + BigqueryDataset string `protobuf:"bytes,5,opt,name=bigquery_dataset,json=bigqueryDataset,proto3" json:"bigquery_dataset,omitempty"` // Required. The BigQuery table to export data to. // If the table doesn't exist, it will be created. If it already exists, it // will be overwritten. - BigqueryTable string `protobuf:"bytes,6,opt,name=bigquery_table,json=bigqueryTable" json:"bigquery_table,omitempty"` + BigqueryTable string `protobuf:"bytes,6,opt,name=bigquery_table,json=bigqueryTable,proto3" json:"bigquery_table,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1050,7 +1050,7 @@ func (m *ExportVariantSetRequest) GetBigqueryTable() string { // The variant set request. type GetVariantSetRequest struct { // Required. The ID of the variant set. - VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1091,14 +1091,14 @@ func (m *GetVariantSetRequest) GetVariantSetId() string { type SearchVariantSetsRequest struct { // Exactly one dataset ID must be provided here. Only variant sets which // belong to this dataset will be returned. - DatasetIds []string `protobuf:"bytes,1,rep,name=dataset_ids,json=datasetIds" json:"dataset_ids,omitempty"` + DatasetIds []string `protobuf:"bytes,1,rep,name=dataset_ids,json=datasetIds,proto3" json:"dataset_ids,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 1024. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1152,11 +1152,11 @@ func (m *SearchVariantSetsRequest) GetPageSize() int32 { // The search variant sets response. type SearchVariantSetsResponse struct { // The variant sets belonging to the requested dataset. - VariantSets []*VariantSet `protobuf:"bytes,1,rep,name=variant_sets,json=variantSets" json:"variant_sets,omitempty"` + VariantSets []*VariantSet `protobuf:"bytes,1,rep,name=variant_sets,json=variantSets,proto3" json:"variant_sets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1203,7 +1203,7 @@ func (m *SearchVariantSetsResponse) GetNextPageToken() string { // The delete variant set request. type DeleteVariantSetRequest struct { // The ID of the variant set to be deleted. - VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1242,10 +1242,10 @@ func (m *DeleteVariantSetRequest) GetVariantSetId() string { type UpdateVariantSetRequest struct { // The ID of the variant to be updated (must already exist). - VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` // The new variant data. Only the variant_set.metadata will be considered // for update. - VariantSet *VariantSet `protobuf:"bytes,2,opt,name=variant_set,json=variantSet" json:"variant_set,omitempty"` + VariantSet *VariantSet `protobuf:"bytes,2,opt,name=variant_set,json=variantSet,proto3" json:"variant_set,omitempty"` // An optional mask specifying which fields to update. Supported fields: // // * [metadata][google.genomics.v1.VariantSet.metadata]. @@ -1254,7 +1254,7 @@ type UpdateVariantSetRequest struct { // // Leaving `updateMask` unset is equivalent to specifying all mutable // fields. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,5,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1310,33 +1310,33 @@ type SearchVariantsRequest struct { // At most one variant set ID must be provided. Only variants from this // variant set will be returned. If omitted, a call set id must be included in // the request. - VariantSetIds []string `protobuf:"bytes,1,rep,name=variant_set_ids,json=variantSetIds" json:"variant_set_ids,omitempty"` + VariantSetIds []string `protobuf:"bytes,1,rep,name=variant_set_ids,json=variantSetIds,proto3" json:"variant_set_ids,omitempty"` // Only return variants which have exactly this name. - VariantName string `protobuf:"bytes,2,opt,name=variant_name,json=variantName" json:"variant_name,omitempty"` + VariantName string `protobuf:"bytes,2,opt,name=variant_name,json=variantName,proto3" json:"variant_name,omitempty"` // Only return variant calls which belong to call sets with these ids. // Leaving this blank returns all variant calls. If a variant has no // calls belonging to any of these call sets, it won't be returned at all. - CallSetIds []string `protobuf:"bytes,3,rep,name=call_set_ids,json=callSetIds" json:"call_set_ids,omitempty"` + CallSetIds []string `protobuf:"bytes,3,rep,name=call_set_ids,json=callSetIds,proto3" json:"call_set_ids,omitempty"` // Required. Only return variants in this reference sequence. - ReferenceName string `protobuf:"bytes,4,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,4,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The beginning of the window (0-based, inclusive) for which // overlapping variants should be returned. If unspecified, defaults to 0. - Start int64 `protobuf:"varint,5,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,5,opt,name=start,proto3" json:"start,omitempty"` // The end of the window, 0-based exclusive. If unspecified or 0, defaults to // the length of the reference. - End int64 `protobuf:"varint,6,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,6,opt,name=end,proto3" json:"end,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,7,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,7,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of variants to return in a single page. If unspecified, // defaults to 5000. The maximum value is 10000. - PageSize int32 `protobuf:"varint,8,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,8,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The maximum number of calls to return in a single page. Note that this // limit may be exceeded in the event that a matching variant contains more // calls than the requested maximum. If unspecified, defaults to 5000. The // maximum value is 10000. - MaxCalls int32 `protobuf:"varint,9,opt,name=max_calls,json=maxCalls" json:"max_calls,omitempty"` + MaxCalls int32 `protobuf:"varint,9,opt,name=max_calls,json=maxCalls,proto3" json:"max_calls,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1432,11 +1432,11 @@ func (m *SearchVariantsRequest) GetMaxCalls() int32 { // The variant search response. type SearchVariantsResponse struct { // The list of matching Variants. - Variants []*Variant `protobuf:"bytes,1,rep,name=variants" json:"variants,omitempty"` + Variants []*Variant `protobuf:"bytes,1,rep,name=variants,proto3" json:"variants,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1482,7 +1482,7 @@ func (m *SearchVariantsResponse) GetNextPageToken() string { type CreateVariantRequest struct { // The variant to be created. - Variant *Variant `protobuf:"bytes,1,opt,name=variant" json:"variant,omitempty"` + Variant *Variant `protobuf:"bytes,1,opt,name=variant,proto3" json:"variant,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1521,14 +1521,14 @@ func (m *CreateVariantRequest) GetVariant() *Variant { type UpdateVariantRequest struct { // The ID of the variant to be updated. - VariantId string `protobuf:"bytes,1,opt,name=variant_id,json=variantId" json:"variant_id,omitempty"` + VariantId string `protobuf:"bytes,1,opt,name=variant_id,json=variantId,proto3" json:"variant_id,omitempty"` // The new variant data. - Variant *Variant `protobuf:"bytes,2,opt,name=variant" json:"variant,omitempty"` + Variant *Variant `protobuf:"bytes,2,opt,name=variant,proto3" json:"variant,omitempty"` // An optional mask specifying which fields to update. At this time, mutable // fields are [names][google.genomics.v1.Variant.names] and // [info][google.genomics.v1.Variant.info]. Acceptable values are "names" and // "info". If unspecified, all mutable fields will be updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1581,7 +1581,7 @@ func (m *UpdateVariantRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteVariantRequest struct { // The ID of the variant to be deleted. - VariantId string `protobuf:"bytes,1,opt,name=variant_id,json=variantId" json:"variant_id,omitempty"` + VariantId string `protobuf:"bytes,1,opt,name=variant_id,json=variantId,proto3" json:"variant_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1620,7 +1620,7 @@ func (m *DeleteVariantRequest) GetVariantId() string { type GetVariantRequest struct { // The ID of the variant. - VariantId string `protobuf:"bytes,1,opt,name=variant_id,json=variantId" json:"variant_id,omitempty"` + VariantId string `protobuf:"bytes,1,opt,name=variant_id,json=variantId,proto3" json:"variant_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1659,12 +1659,12 @@ func (m *GetVariantRequest) GetVariantId() string { type MergeVariantsRequest struct { // The destination variant set. - VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,1,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` // The variants to be merged with existing variants. - Variants []*Variant `protobuf:"bytes,2,rep,name=variants" json:"variants,omitempty"` + Variants []*Variant `protobuf:"bytes,2,rep,name=variants,proto3" json:"variants,omitempty"` // A mapping between info field keys and the InfoMergeOperations to // be performed on them. - InfoMergeConfig map[string]InfoMergeOperation `protobuf:"bytes,3,rep,name=info_merge_config,json=infoMergeConfig" json:"info_merge_config,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value,enum=google.genomics.v1.InfoMergeOperation"` + InfoMergeConfig map[string]InfoMergeOperation `protobuf:"bytes,3,rep,name=info_merge_config,json=infoMergeConfig,proto3" json:"info_merge_config,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3,enum=google.genomics.v1.InfoMergeOperation"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1719,17 +1719,17 @@ func (m *MergeVariantsRequest) GetInfoMergeConfig() map[string]InfoMergeOperatio type SearchCallSetsRequest struct { // Restrict the query to call sets within the given variant sets. At least one // ID must be provided. - VariantSetIds []string `protobuf:"bytes,1,rep,name=variant_set_ids,json=variantSetIds" json:"variant_set_ids,omitempty"` + VariantSetIds []string `protobuf:"bytes,1,rep,name=variant_set_ids,json=variantSetIds,proto3" json:"variant_set_ids,omitempty"` // Only return call sets for which a substring of the name matches this // string. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // The continuation token, which is used to page through large result sets. // To get the next page of results, set this parameter to the value of // `nextPageToken` from the previous response. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The maximum number of results to return in a single page. If unspecified, // defaults to 1024. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1790,11 +1790,11 @@ func (m *SearchCallSetsRequest) GetPageSize() int32 { // The call set search response. type SearchCallSetsResponse struct { // The list of matching call sets. - CallSets []*CallSet `protobuf:"bytes,1,rep,name=call_sets,json=callSets" json:"call_sets,omitempty"` + CallSets []*CallSet `protobuf:"bytes,1,rep,name=call_sets,json=callSets,proto3" json:"call_sets,omitempty"` // The continuation token, which is used to page through large result sets. // Provide this value in a subsequent request to return the next page of // results. This field will be empty if there aren't any additional results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1840,7 +1840,7 @@ func (m *SearchCallSetsResponse) GetNextPageToken() string { type CreateCallSetRequest struct { // The call set to be created. - CallSet *CallSet `protobuf:"bytes,1,opt,name=call_set,json=callSet" json:"call_set,omitempty"` + CallSet *CallSet `protobuf:"bytes,1,opt,name=call_set,json=callSet,proto3" json:"call_set,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1879,14 +1879,14 @@ func (m *CreateCallSetRequest) GetCallSet() *CallSet { type UpdateCallSetRequest struct { // The ID of the call set to be updated. - CallSetId string `protobuf:"bytes,1,opt,name=call_set_id,json=callSetId" json:"call_set_id,omitempty"` + CallSetId string `protobuf:"bytes,1,opt,name=call_set_id,json=callSetId,proto3" json:"call_set_id,omitempty"` // The new call set data. - CallSet *CallSet `protobuf:"bytes,2,opt,name=call_set,json=callSet" json:"call_set,omitempty"` + CallSet *CallSet `protobuf:"bytes,2,opt,name=call_set,json=callSet,proto3" json:"call_set,omitempty"` // An optional mask specifying which fields to update. At this time, the only // mutable field is [name][google.genomics.v1.CallSet.name]. The only // acceptable value is "name". If unspecified, all mutable fields will be // updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1939,7 +1939,7 @@ func (m *UpdateCallSetRequest) GetUpdateMask() *field_mask.FieldMask { type DeleteCallSetRequest struct { // The ID of the call set to be deleted. - CallSetId string `protobuf:"bytes,1,opt,name=call_set_id,json=callSetId" json:"call_set_id,omitempty"` + CallSetId string `protobuf:"bytes,1,opt,name=call_set_id,json=callSetId,proto3" json:"call_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1978,7 +1978,7 @@ func (m *DeleteCallSetRequest) GetCallSetId() string { type GetCallSetRequest struct { // The ID of the call set. - CallSetId string `protobuf:"bytes,1,opt,name=call_set_id,json=callSetId" json:"call_set_id,omitempty"` + CallSetId string `protobuf:"bytes,1,opt,name=call_set_id,json=callSetId,proto3" json:"call_set_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2020,20 +2020,20 @@ type StreamVariantsRequest struct { // The Google Cloud project ID which will be billed // for this access. The caller must have WRITE access to this project. // Required. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The variant set ID from which to stream variants. - VariantSetId string `protobuf:"bytes,2,opt,name=variant_set_id,json=variantSetId" json:"variant_set_id,omitempty"` + VariantSetId string `protobuf:"bytes,2,opt,name=variant_set_id,json=variantSetId,proto3" json:"variant_set_id,omitempty"` // Only return variant calls which belong to call sets with these IDs. // Leaving this blank returns all variant calls. - CallSetIds []string `protobuf:"bytes,3,rep,name=call_set_ids,json=callSetIds" json:"call_set_ids,omitempty"` + CallSetIds []string `protobuf:"bytes,3,rep,name=call_set_ids,json=callSetIds,proto3" json:"call_set_ids,omitempty"` // Required. Only return variants in this reference sequence. - ReferenceName string `protobuf:"bytes,4,opt,name=reference_name,json=referenceName" json:"reference_name,omitempty"` + ReferenceName string `protobuf:"bytes,4,opt,name=reference_name,json=referenceName,proto3" json:"reference_name,omitempty"` // The beginning of the window (0-based, inclusive) for which // overlapping variants should be returned. - Start int64 `protobuf:"varint,5,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,5,opt,name=start,proto3" json:"start,omitempty"` // The end of the window (0-based, exclusive) for which overlapping // variants should be returned. - End int64 `protobuf:"varint,6,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,6,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2106,7 +2106,7 @@ func (m *StreamVariantsRequest) GetEnd() int64 { } type StreamVariantsResponse struct { - Variants []*Variant `protobuf:"bytes,1,rep,name=variants" json:"variants,omitempty"` + Variants []*Variant `protobuf:"bytes,1,rep,name=variants,proto3" json:"variants,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2243,8 +2243,7 @@ func (x *streamingVariantServiceStreamVariantsClient) Recv() (*StreamVariantsRes return m, nil } -// Server API for StreamingVariantService service - +// StreamingVariantServiceServer is the server API for StreamingVariantService service. type StreamingVariantServiceServer interface { // Returns a stream of all the variants matching the search request, ordered // by reference name, position, and ID. @@ -2692,8 +2691,7 @@ func (c *variantServiceV1Client) GetCallSet(ctx context.Context, in *GetCallSetR return out, nil } -// Server API for VariantServiceV1 service - +// VariantServiceV1Server is the server API for VariantServiceV1 service. type VariantServiceV1Server interface { // Creates variant data by asynchronously importing the provided information. // diff --git a/vendor/google.golang.org/genproto/googleapis/genomics/v1alpha2/pipelines.pb.go b/vendor/google.golang.org/genproto/googleapis/genomics/v1alpha2/pipelines.pb.go index 3a687c4b9..5ebad6071 100644 --- a/vendor/google.golang.org/genproto/googleapis/genomics/v1alpha2/pipelines.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/genomics/v1alpha2/pipelines.pb.go @@ -70,13 +70,13 @@ func (PipelineResources_Disk_Type) EnumDescriptor() ([]byte, []int) { // [pipeline][google.genomics.v1alpha2.Pipeline]. type ComputeEngine struct { // The instance on which the operation is running. - InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName" json:"instance_name,omitempty"` + InstanceName string `protobuf:"bytes,1,opt,name=instance_name,json=instanceName,proto3" json:"instance_name,omitempty"` // The availability zone in which the instance resides. - Zone string `protobuf:"bytes,2,opt,name=zone" json:"zone,omitempty"` + Zone string `protobuf:"bytes,2,opt,name=zone,proto3" json:"zone,omitempty"` // The machine type of the instance. - MachineType string `protobuf:"bytes,3,opt,name=machine_type,json=machineType" json:"machine_type,omitempty"` + MachineType string `protobuf:"bytes,3,opt,name=machine_type,json=machineType,proto3" json:"machine_type,omitempty"` // The names of the disks that were created for this pipeline. - DiskNames []string `protobuf:"bytes,4,rep,name=disk_names,json=diskNames" json:"disk_names,omitempty"` + DiskNames []string `protobuf:"bytes,4,rep,name=disk_names,json=diskNames,proto3" json:"disk_names,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -139,7 +139,7 @@ func (m *ComputeEngine) GetDiskNames() []string { // field of the Operation associated with a RunPipeline execution. type RuntimeMetadata struct { // Execution information specific to Google Compute Engine. - ComputeEngine *ComputeEngine `protobuf:"bytes,1,opt,name=compute_engine,json=computeEngine" json:"compute_engine,omitempty"` + ComputeEngine *ComputeEngine `protobuf:"bytes,1,opt,name=compute_engine,json=computeEngine,proto3" json:"compute_engine,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -185,16 +185,16 @@ func (m *RuntimeMetadata) GetComputeEngine() *ComputeEngine { type Pipeline struct { // Required. The project in which to create the pipeline. The caller must have // WRITE access. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Required. A user specified pipeline name that does not have to be unique. // This name can be used for filtering Pipelines in ListPipelines. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` // User-specified description. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // Input parameters of the pipeline. - InputParameters []*PipelineParameter `protobuf:"bytes,8,rep,name=input_parameters,json=inputParameters" json:"input_parameters,omitempty"` + InputParameters []*PipelineParameter `protobuf:"bytes,8,rep,name=input_parameters,json=inputParameters,proto3" json:"input_parameters,omitempty"` // Output parameters of the pipeline. - OutputParameters []*PipelineParameter `protobuf:"bytes,9,rep,name=output_parameters,json=outputParameters" json:"output_parameters,omitempty"` + OutputParameters []*PipelineParameter `protobuf:"bytes,9,rep,name=output_parameters,json=outputParameters,proto3" json:"output_parameters,omitempty"` // Required. The executor indicates in which environment the pipeline runs. // // Types that are valid to be assigned to Executor: @@ -208,13 +208,13 @@ type Pipeline struct { // // * // [minimumRamGb][google.genomics.v1alpha2.PipelineResources.minimum_ram_gb] - Resources *PipelineResources `protobuf:"bytes,6,opt,name=resources" json:"resources,omitempty"` + Resources *PipelineResources `protobuf:"bytes,6,opt,name=resources,proto3" json:"resources,omitempty"` // Unique pipeline id that is generated by the service when CreatePipeline // is called. Cannot be specified in the Pipeline used in the // CreatePipelineRequest, and will be populated in the response to // CreatePipeline and all subsequent Get and List calls. Indicates that the // service has registered this pipeline. - PipelineId string `protobuf:"bytes,7,opt,name=pipeline_id,json=pipelineId" json:"pipeline_id,omitempty"` + PipelineId string `protobuf:"bytes,7,opt,name=pipeline_id,json=pipelineId,proto3" json:"pipeline_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -249,7 +249,7 @@ type isPipeline_Executor interface { } type Pipeline_Docker struct { - Docker *DockerExecutor `protobuf:"bytes,5,opt,name=docker,oneof"` + Docker *DockerExecutor `protobuf:"bytes,5,opt,name=docker,proto3,oneof"` } func (*Pipeline_Docker) isPipeline_Executor() {} @@ -376,7 +376,7 @@ func _Pipeline_OneofSizer(msg proto.Message) (n int) { // `pipelineId` populated, as that will be populated by the server. type CreatePipelineRequest struct { // The pipeline to create. Should not have `pipelineId` populated. - Pipeline *Pipeline `protobuf:"bytes,1,opt,name=pipeline" json:"pipeline,omitempty"` + Pipeline *Pipeline `protobuf:"bytes,1,opt,name=pipeline,proto3" json:"pipeline,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -418,33 +418,33 @@ type RunPipelineArgs struct { // Required. The project in which to run the pipeline. The caller must have // WRITER access to all Google Cloud services and resources (e.g. Google // Compute Engine) will be used. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Pipeline input arguments; keys are defined in the pipeline documentation. // All input parameters that do not have default values must be specified. // If parameters with defaults are specified here, the defaults will be // overridden. - Inputs map[string]string `protobuf:"bytes,2,rep,name=inputs" json:"inputs,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Inputs map[string]string `protobuf:"bytes,2,rep,name=inputs,proto3" json:"inputs,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Pipeline output arguments; keys are defined in the pipeline // documentation. All output parameters of without default values // must be specified. If parameters with defaults are specified // here, the defaults will be overridden. - Outputs map[string]string `protobuf:"bytes,3,rep,name=outputs" json:"outputs,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Outputs map[string]string `protobuf:"bytes,3,rep,name=outputs,proto3" json:"outputs,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // The Google Cloud Service Account that will be used to access data and // services. By default, the compute service account associated with // `projectId` is used. - ServiceAccount *ServiceAccount `protobuf:"bytes,4,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount *ServiceAccount `protobuf:"bytes,4,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` // This field is deprecated. Use `labels` instead. Client-specified pipeline // operation identifier. - ClientId string `protobuf:"bytes,5,opt,name=client_id,json=clientId" json:"client_id,omitempty"` + ClientId string `protobuf:"bytes,5,opt,name=client_id,json=clientId,proto3" json:"client_id,omitempty"` // Specifies resource requirements/overrides for the pipeline run. - Resources *PipelineResources `protobuf:"bytes,6,opt,name=resources" json:"resources,omitempty"` + Resources *PipelineResources `protobuf:"bytes,6,opt,name=resources,proto3" json:"resources,omitempty"` // Required. Logging options. Used by the service to communicate results // to the user. - Logging *LoggingOptions `protobuf:"bytes,7,opt,name=logging" json:"logging,omitempty"` + Logging *LoggingOptions `protobuf:"bytes,7,opt,name=logging,proto3" json:"logging,omitempty"` // How long to keep the VM up after a failure (for example docker command // failed, copying input or output files failed, etc). While the VM is up, one // can ssh into the VM to debug. Default is 0; maximum allowed value is 1 day. - KeepVmAliveOnFailureDuration *duration.Duration `protobuf:"bytes,8,opt,name=keep_vm_alive_on_failure_duration,json=keepVmAliveOnFailureDuration" json:"keep_vm_alive_on_failure_duration,omitempty"` + KeepVmAliveOnFailureDuration *duration.Duration `protobuf:"bytes,8,opt,name=keep_vm_alive_on_failure_duration,json=keepVmAliveOnFailureDuration,proto3" json:"keep_vm_alive_on_failure_duration,omitempty"` // Labels to apply to this pipeline run. Labels will also be applied to // compute resources (VM, disks) created by this pipeline run. When listing // operations, operations can [filtered by labels] @@ -457,7 +457,7 @@ type RunPipelineArgs struct { // character must be a lowercase letter, and all following characters must be // a dash, lowercase letter, or digit, except the last character, which cannot // be a dash. - Labels map[string]string `protobuf:"bytes,9,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,9,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -563,7 +563,7 @@ type RunPipelineRequest struct { // *RunPipelineRequest_EphemeralPipeline Pipeline isRunPipelineRequest_Pipeline `protobuf_oneof:"pipeline"` // The arguments to use when running this pipeline. - PipelineArgs *RunPipelineArgs `protobuf:"bytes,3,opt,name=pipeline_args,json=pipelineArgs" json:"pipeline_args,omitempty"` + PipelineArgs *RunPipelineArgs `protobuf:"bytes,3,opt,name=pipeline_args,json=pipelineArgs,proto3" json:"pipeline_args,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -598,10 +598,10 @@ type isRunPipelineRequest_Pipeline interface { } type RunPipelineRequest_PipelineId struct { - PipelineId string `protobuf:"bytes,1,opt,name=pipeline_id,json=pipelineId,oneof"` + PipelineId string `protobuf:"bytes,1,opt,name=pipeline_id,json=pipelineId,proto3,oneof"` } type RunPipelineRequest_EphemeralPipeline struct { - EphemeralPipeline *Pipeline `protobuf:"bytes,2,opt,name=ephemeral_pipeline,json=ephemeralPipeline,oneof"` + EphemeralPipeline *Pipeline `protobuf:"bytes,2,opt,name=ephemeral_pipeline,json=ephemeralPipeline,proto3,oneof"` } func (*RunPipelineRequest_PipelineId) isRunPipelineRequest_Pipeline() {} @@ -709,7 +709,7 @@ func _RunPipelineRequest_OneofSizer(msg proto.Message) (n int) { type GetPipelineRequest struct { // Caller must have READ access to the project in which this pipeline // is defined. - PipelineId string `protobuf:"bytes,1,opt,name=pipeline_id,json=pipelineId" json:"pipeline_id,omitempty"` + PipelineId string `protobuf:"bytes,1,opt,name=pipeline_id,json=pipelineId,proto3" json:"pipeline_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -755,17 +755,17 @@ func (m *GetPipelineRequest) GetPipelineId() string { type ListPipelinesRequest struct { // Required. The name of the project to search for pipelines. Caller // must have READ access to this project. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Pipelines with names that match this prefix should be // returned. If unspecified, all pipelines in the project, up to // `pageSize`, will be returned. - NamePrefix string `protobuf:"bytes,2,opt,name=name_prefix,json=namePrefix" json:"name_prefix,omitempty"` + NamePrefix string `protobuf:"bytes,2,opt,name=name_prefix,json=namePrefix,proto3" json:"name_prefix,omitempty"` // Number of pipelines to return at once. Defaults to 256, and max // is 2048. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Token to use to indicate where to start getting results. // If unspecified, returns the first page of results. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -830,9 +830,9 @@ func (m *ListPipelinesRequest) GetPageToken() string { // request. type ListPipelinesResponse struct { // The matched pipelines. - Pipelines []*Pipeline `protobuf:"bytes,1,rep,name=pipelines" json:"pipelines,omitempty"` + Pipelines []*Pipeline `protobuf:"bytes,1,rep,name=pipelines,proto3" json:"pipelines,omitempty"` // The token to use to get the next page of results. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -880,7 +880,7 @@ func (m *ListPipelinesResponse) GetNextPageToken() string { type DeletePipelineRequest struct { // Caller must have WRITE access to the project in which this pipeline // is defined. - PipelineId string `protobuf:"bytes,1,opt,name=pipeline_id,json=pipelineId" json:"pipeline_id,omitempty"` + PipelineId string `protobuf:"bytes,1,opt,name=pipeline_id,json=pipelineId,proto3" json:"pipeline_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -921,8 +921,8 @@ func (m *DeletePipelineRequest) GetPipelineId() string { // by VMs created by the Pipelines Service and not by end users. type GetControllerConfigRequest struct { // The operation to retrieve controller configuration for. - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` - ValidationToken uint64 `protobuf:"varint,2,opt,name=validation_token,json=validationToken" json:"validation_token,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` + ValidationToken uint64 `protobuf:"varint,2,opt,name=validation_token,json=validationToken,proto3" json:"validation_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -970,14 +970,14 @@ func (m *GetControllerConfigRequest) GetValidationToken() uint64 { // server in order to run. Should only be used by VMs created by the // Pipelines Service and not by end users. type ControllerConfig struct { - Image string `protobuf:"bytes,1,opt,name=image" json:"image,omitempty"` - Cmd string `protobuf:"bytes,2,opt,name=cmd" json:"cmd,omitempty"` - GcsLogPath string `protobuf:"bytes,3,opt,name=gcs_log_path,json=gcsLogPath" json:"gcs_log_path,omitempty"` - MachineType string `protobuf:"bytes,4,opt,name=machine_type,json=machineType" json:"machine_type,omitempty"` - Vars map[string]string `protobuf:"bytes,5,rep,name=vars" json:"vars,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` - Disks map[string]string `protobuf:"bytes,6,rep,name=disks" json:"disks,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` - GcsSources map[string]*ControllerConfig_RepeatedString `protobuf:"bytes,7,rep,name=gcs_sources,json=gcsSources" json:"gcs_sources,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` - GcsSinks map[string]*ControllerConfig_RepeatedString `protobuf:"bytes,8,rep,name=gcs_sinks,json=gcsSinks" json:"gcs_sinks,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Image string `protobuf:"bytes,1,opt,name=image,proto3" json:"image,omitempty"` + Cmd string `protobuf:"bytes,2,opt,name=cmd,proto3" json:"cmd,omitempty"` + GcsLogPath string `protobuf:"bytes,3,opt,name=gcs_log_path,json=gcsLogPath,proto3" json:"gcs_log_path,omitempty"` + MachineType string `protobuf:"bytes,4,opt,name=machine_type,json=machineType,proto3" json:"machine_type,omitempty"` + Vars map[string]string `protobuf:"bytes,5,rep,name=vars,proto3" json:"vars,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + Disks map[string]string `protobuf:"bytes,6,rep,name=disks,proto3" json:"disks,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + GcsSources map[string]*ControllerConfig_RepeatedString `protobuf:"bytes,7,rep,name=gcs_sources,json=gcsSources,proto3" json:"gcs_sources,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + GcsSinks map[string]*ControllerConfig_RepeatedString `protobuf:"bytes,8,rep,name=gcs_sinks,json=gcsSinks,proto3" json:"gcs_sinks,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1064,7 +1064,7 @@ func (m *ControllerConfig) GetGcsSinks() map[string]*ControllerConfig_RepeatedSt } type ControllerConfig_RepeatedString struct { - Values []string `protobuf:"bytes,1,rep,name=values" json:"values,omitempty"` + Values []string `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1105,9 +1105,9 @@ func (m *ControllerConfig_RepeatedString) GetValues() []string { // execution. type TimestampEvent struct { // String indicating the type of event - Description string `protobuf:"bytes,1,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` // The time this event occured. - Timestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=timestamp" json:"timestamp,omitempty"` + Timestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=timestamp,proto3" json:"timestamp,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1154,11 +1154,11 @@ func (m *TimestampEvent) GetTimestamp() *timestamp.Timestamp { // Request to set operation status. Should only be used by VMs // created by the Pipelines Service and not by end users. type SetOperationStatusRequest struct { - OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` - TimestampEvents []*TimestampEvent `protobuf:"bytes,2,rep,name=timestamp_events,json=timestampEvents" json:"timestamp_events,omitempty"` - ErrorCode code.Code `protobuf:"varint,3,opt,name=error_code,json=errorCode,enum=google.rpc.Code" json:"error_code,omitempty"` - ErrorMessage string `protobuf:"bytes,4,opt,name=error_message,json=errorMessage" json:"error_message,omitempty"` - ValidationToken uint64 `protobuf:"varint,5,opt,name=validation_token,json=validationToken" json:"validation_token,omitempty"` + OperationId string `protobuf:"bytes,1,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` + TimestampEvents []*TimestampEvent `protobuf:"bytes,2,rep,name=timestamp_events,json=timestampEvents,proto3" json:"timestamp_events,omitempty"` + ErrorCode code.Code `protobuf:"varint,3,opt,name=error_code,json=errorCode,proto3,enum=google.rpc.Code" json:"error_code,omitempty"` + ErrorMessage string `protobuf:"bytes,4,opt,name=error_message,json=errorMessage,proto3" json:"error_message,omitempty"` + ValidationToken uint64 `protobuf:"varint,5,opt,name=validation_token,json=validationToken,proto3" json:"validation_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1227,7 +1227,7 @@ func (m *SetOperationStatusRequest) GetValidationToken() uint64 { type ServiceAccount struct { // Email address of the service account. Defaults to `default`, // which uses the compute service account associated with the project. - Email string `protobuf:"bytes,1,opt,name=email" json:"email,omitempty"` + Email string `protobuf:"bytes,1,opt,name=email,proto3" json:"email,omitempty"` // List of scopes to be enabled for this service account on the VM. // The following scopes are automatically included: // @@ -1236,7 +1236,7 @@ type ServiceAccount struct { // * https://www.googleapis.com/auth/genomics // * https://www.googleapis.com/auth/logging.write // * https://www.googleapis.com/auth/monitoring.write - Scopes []string `protobuf:"bytes,2,rep,name=scopes" json:"scopes,omitempty"` + Scopes []string `protobuf:"bytes,2,rep,name=scopes,proto3" json:"scopes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1290,7 +1290,7 @@ type LoggingOptions struct { // used, and the user must ensure that logs are not // overwritten. Stdout and stderr logs from the run are also // generated and output as `-stdout.log` and `-stderr.log`. - GcsPath string `protobuf:"bytes,1,opt,name=gcs_path,json=gcsPath" json:"gcs_path,omitempty"` + GcsPath string `protobuf:"bytes,1,opt,name=gcs_path,json=gcsPath,proto3" json:"gcs_path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1330,20 +1330,20 @@ func (m *LoggingOptions) GetGcsPath() string { // The system resources for the pipeline run. type PipelineResources struct { // The minimum number of cores to use. Defaults to 1. - MinimumCpuCores int32 `protobuf:"varint,1,opt,name=minimum_cpu_cores,json=minimumCpuCores" json:"minimum_cpu_cores,omitempty"` + MinimumCpuCores int32 `protobuf:"varint,1,opt,name=minimum_cpu_cores,json=minimumCpuCores,proto3" json:"minimum_cpu_cores,omitempty"` // Whether to use preemptible VMs. Defaults to `false`. In order to use this, // must be true for both create time and run time. Cannot be true at run time // if false at create time. - Preemptible bool `protobuf:"varint,2,opt,name=preemptible" json:"preemptible,omitempty"` + Preemptible bool `protobuf:"varint,2,opt,name=preemptible,proto3" json:"preemptible,omitempty"` // The minimum amount of RAM to use. Defaults to 3.75 (GB) - MinimumRamGb float64 `protobuf:"fixed64,3,opt,name=minimum_ram_gb,json=minimumRamGb" json:"minimum_ram_gb,omitempty"` + MinimumRamGb float64 `protobuf:"fixed64,3,opt,name=minimum_ram_gb,json=minimumRamGb,proto3" json:"minimum_ram_gb,omitempty"` // Disks to attach. - Disks []*PipelineResources_Disk `protobuf:"bytes,4,rep,name=disks" json:"disks,omitempty"` + Disks []*PipelineResources_Disk `protobuf:"bytes,4,rep,name=disks,proto3" json:"disks,omitempty"` // List of Google Compute Engine availability zones to which resource // creation will restricted. If empty, any zone may be chosen. - Zones []string `protobuf:"bytes,5,rep,name=zones" json:"zones,omitempty"` + Zones []string `protobuf:"bytes,5,rep,name=zones,proto3" json:"zones,omitempty"` // The size of the boot disk. Defaults to 10 (GB). - BootDiskSizeGb int32 `protobuf:"varint,6,opt,name=boot_disk_size_gb,json=bootDiskSizeGb" json:"boot_disk_size_gb,omitempty"` + BootDiskSizeGb int32 `protobuf:"varint,6,opt,name=boot_disk_size_gb,json=bootDiskSizeGb,proto3" json:"boot_disk_size_gb,omitempty"` // Whether to assign an external IP to the instance. This is an experimental // feature that may go away. Defaults to false. // Corresponds to `--no_address` flag for [gcloud compute instances create] @@ -1355,7 +1355,7 @@ type PipelineResources struct { // only load docker images from Google Container Registry and not Docker Hub. // ** Note: To use this option, your project must be in Google Access for // Private IPs Early Access Program.** - NoAddress bool `protobuf:"varint,7,opt,name=no_address,json=noAddress" json:"no_address,omitempty"` + NoAddress bool `protobuf:"varint,7,opt,name=no_address,json=noAddress,proto3" json:"no_address,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1439,28 +1439,28 @@ type PipelineResources_Disk struct { // Required. The name of the disk that can be used in the pipeline // parameters. Must be 1 - 63 characters. // The name "boot" is reserved for system use. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The type of the disk to create. - Type PipelineResources_Disk_Type `protobuf:"varint,2,opt,name=type,enum=google.genomics.v1alpha2.PipelineResources_Disk_Type" json:"type,omitempty"` + Type PipelineResources_Disk_Type `protobuf:"varint,2,opt,name=type,proto3,enum=google.genomics.v1alpha2.PipelineResources_Disk_Type" json:"type,omitempty"` // The size of the disk. Defaults to 500 (GB). // This field is not applicable for local SSD. - SizeGb int32 `protobuf:"varint,3,opt,name=size_gb,json=sizeGb" json:"size_gb,omitempty"` + SizeGb int32 `protobuf:"varint,3,opt,name=size_gb,json=sizeGb,proto3" json:"size_gb,omitempty"` // The full or partial URL of the persistent disk to attach. See // https://cloud.google.com/compute/docs/reference/latest/instances#resource // and // https://cloud.google.com/compute/docs/disks/persistent-disks#snapshots // for more details. - Source string `protobuf:"bytes,4,opt,name=source" json:"source,omitempty"` + Source string `protobuf:"bytes,4,opt,name=source,proto3" json:"source,omitempty"` // Deprecated. Disks created by the Pipelines API will be deleted at the end // of the pipeline run, regardless of what this field is set to. - AutoDelete bool `protobuf:"varint,6,opt,name=auto_delete,json=autoDelete" json:"auto_delete,omitempty"` + AutoDelete bool `protobuf:"varint,6,opt,name=auto_delete,json=autoDelete,proto3" json:"auto_delete,omitempty"` // Required at create time and cannot be overridden at run time. // Specifies the path in the docker container where files on // this disk should be located. For example, if `mountPoint` // is `/mnt/disk`, and the parameter has `localPath` // `inputs/file.txt`, the docker container can access the data at // `/mnt/disk/inputs/file.txt`. - MountPoint string `protobuf:"bytes,8,opt,name=mount_point,json=mountPoint" json:"mount_point,omitempty"` + MountPoint string `protobuf:"bytes,8,opt,name=mount_point,json=mountPoint,proto3" json:"mount_point,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1626,18 +1626,18 @@ func (m *PipelineResources_Disk) GetMountPoint() string { type PipelineParameter struct { // Required. Name of the parameter - the pipeline runner uses this string // as the key to the input and output maps in RunPipeline. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Human-readable description. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // The default value for this parameter. Can be overridden at runtime. // If `localCopy` is present, then this must be a Google Cloud Storage path // beginning with `gs://`. - DefaultValue string `protobuf:"bytes,5,opt,name=default_value,json=defaultValue" json:"default_value,omitempty"` + DefaultValue string `protobuf:"bytes,5,opt,name=default_value,json=defaultValue,proto3" json:"default_value,omitempty"` // If present, this parameter is marked for copying to and from the VM. // `LocalCopy` indicates where on the VM the file should be. The value // given to this parameter (either at runtime or using `defaultValue`) // must be the remote path where the file should be. - LocalCopy *PipelineParameter_LocalCopy `protobuf:"bytes,6,opt,name=local_copy,json=localCopy" json:"local_copy,omitempty"` + LocalCopy *PipelineParameter_LocalCopy `protobuf:"bytes,6,opt,name=local_copy,json=localCopy,proto3" json:"local_copy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1700,12 +1700,12 @@ type PipelineParameter_LocalCopy struct { // Required. The path within the user's docker container where // this input should be localized to and from, relative to the specified // disk's mount point. For example: file.txt, - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` // Required. The name of the disk where this parameter is // located. Can be the name of one of the disks specified in the // Resources field, or "boot", which represents the Docker // instance's boot disk and has a mount point of `/`. - Disk string `protobuf:"bytes,2,opt,name=disk" json:"disk,omitempty"` + Disk string `protobuf:"bytes,2,opt,name=disk,proto3" json:"disk,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1753,7 +1753,7 @@ func (m *PipelineParameter_LocalCopy) GetDisk() string { type DockerExecutor struct { // Required. Image name from either Docker Hub or Google Container Registry. // Users that run pipelines must have READ access to the image. - ImageName string `protobuf:"bytes,1,opt,name=image_name,json=imageName" json:"image_name,omitempty"` + ImageName string `protobuf:"bytes,1,opt,name=image_name,json=imageName,proto3" json:"image_name,omitempty"` // Required. The command or newline delimited script to run. The command // string will be executed within a bash shell. // @@ -1762,7 +1762,7 @@ type DockerExecutor struct { // [`error`][google.longrunning.Operation.error] field will be populated. // // Maximum command string length is 16384. - Cmd string `protobuf:"bytes,2,opt,name=cmd" json:"cmd,omitempty"` + Cmd string `protobuf:"bytes,2,opt,name=cmd,proto3" json:"cmd,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1958,8 +1958,7 @@ func (c *pipelinesV1Alpha2Client) SetOperationStatus(ctx context.Context, in *Se return out, nil } -// Server API for PipelinesV1Alpha2 service - +// PipelinesV1Alpha2Server is the server API for PipelinesV1Alpha2 service. type PipelinesV1Alpha2Server interface { // Creates a pipeline that can be run later. Create takes a Pipeline that // has all fields other than `pipelineId` populated, and then returns diff --git a/vendor/google.golang.org/genproto/googleapis/home/graph/v1/device.pb.go b/vendor/google.golang.org/genproto/googleapis/home/graph/v1/device.pb.go new file mode 100644 index 000000000..17c8cf838 --- /dev/null +++ b/vendor/google.golang.org/genproto/googleapis/home/graph/v1/device.pb.go @@ -0,0 +1,316 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: google/home/graph/v1/device.proto + +package graph // import "google.golang.org/genproto/googleapis/home/graph/v1" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import _struct "github.com/golang/protobuf/ptypes/struct" + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +// Third-party partner's device definition. +type Device struct { + // Third-party partner's device ID. + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` + // Hardware type of the device (e.g. light, outlet, etc). + Type string `protobuf:"bytes,2,opt,name=type,proto3" json:"type,omitempty"` + // Traits supported by the device. + Traits []string `protobuf:"bytes,3,rep,name=traits,proto3" json:"traits,omitempty"` + // Name of the device given by the third party. This includes names given to + // the device via third party device manufacturer's app, model names for the + // device, etc. + Name *DeviceNames `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` + // Indicates whether the state of this device is being reported to Google + // through ReportStateAndNotification call. + WillReportState bool `protobuf:"varint,5,opt,name=will_report_state,json=willReportState,proto3" json:"will_report_state,omitempty"` + // If the third-party partner's cloud configuration includes placing devices + // in rooms, the name of the room can be provided here. + RoomHint string `protobuf:"bytes,6,opt,name=room_hint,json=roomHint,proto3" json:"room_hint,omitempty"` + // As in roomHint, for structures that users set up in the partner's system. + StructureHint string `protobuf:"bytes,7,opt,name=structure_hint,json=structureHint,proto3" json:"structure_hint,omitempty"` + // Device manufacturer, model, hardware version, and software version. + DeviceInfo *DeviceInfo `protobuf:"bytes,8,opt,name=device_info,json=deviceInfo,proto3" json:"device_info,omitempty"` + // Attributes for the traits supported by the device. + Attributes *_struct.Struct `protobuf:"bytes,9,opt,name=attributes,proto3" json:"attributes,omitempty"` + // Custom JSON data provided by the manufacturer and attached to QUERY and + // EXECUTE requests in AoG. + CustomData string `protobuf:"bytes,10,opt,name=custom_data,json=customData,proto3" json:"custom_data,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Device) Reset() { *m = Device{} } +func (m *Device) String() string { return proto.CompactTextString(m) } +func (*Device) ProtoMessage() {} +func (*Device) Descriptor() ([]byte, []int) { + return fileDescriptor_device_310b64582597de9b, []int{0} +} +func (m *Device) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Device.Unmarshal(m, b) +} +func (m *Device) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Device.Marshal(b, m, deterministic) +} +func (dst *Device) XXX_Merge(src proto.Message) { + xxx_messageInfo_Device.Merge(dst, src) +} +func (m *Device) XXX_Size() int { + return xxx_messageInfo_Device.Size(m) +} +func (m *Device) XXX_DiscardUnknown() { + xxx_messageInfo_Device.DiscardUnknown(m) +} + +var xxx_messageInfo_Device proto.InternalMessageInfo + +func (m *Device) GetId() string { + if m != nil { + return m.Id + } + return "" +} + +func (m *Device) GetType() string { + if m != nil { + return m.Type + } + return "" +} + +func (m *Device) GetTraits() []string { + if m != nil { + return m.Traits + } + return nil +} + +func (m *Device) GetName() *DeviceNames { + if m != nil { + return m.Name + } + return nil +} + +func (m *Device) GetWillReportState() bool { + if m != nil { + return m.WillReportState + } + return false +} + +func (m *Device) GetRoomHint() string { + if m != nil { + return m.RoomHint + } + return "" +} + +func (m *Device) GetStructureHint() string { + if m != nil { + return m.StructureHint + } + return "" +} + +func (m *Device) GetDeviceInfo() *DeviceInfo { + if m != nil { + return m.DeviceInfo + } + return nil +} + +func (m *Device) GetAttributes() *_struct.Struct { + if m != nil { + return m.Attributes + } + return nil +} + +func (m *Device) GetCustomData() string { + if m != nil { + return m.CustomData + } + return "" +} + +// Different names for the device. +type DeviceNames struct { + // Primary name of the device, generally provided by the user. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + // Additional names provided by the user for the device. + Nicknames []string `protobuf:"bytes,2,rep,name=nicknames,proto3" json:"nicknames,omitempty"` + // List of names provided by the partner rather than the user, often + // manufacturer names, SKUs, etc. + DefaultNames []string `protobuf:"bytes,3,rep,name=default_names,json=defaultNames,proto3" json:"default_names,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DeviceNames) Reset() { *m = DeviceNames{} } +func (m *DeviceNames) String() string { return proto.CompactTextString(m) } +func (*DeviceNames) ProtoMessage() {} +func (*DeviceNames) Descriptor() ([]byte, []int) { + return fileDescriptor_device_310b64582597de9b, []int{1} +} +func (m *DeviceNames) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DeviceNames.Unmarshal(m, b) +} +func (m *DeviceNames) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DeviceNames.Marshal(b, m, deterministic) +} +func (dst *DeviceNames) XXX_Merge(src proto.Message) { + xxx_messageInfo_DeviceNames.Merge(dst, src) +} +func (m *DeviceNames) XXX_Size() int { + return xxx_messageInfo_DeviceNames.Size(m) +} +func (m *DeviceNames) XXX_DiscardUnknown() { + xxx_messageInfo_DeviceNames.DiscardUnknown(m) +} + +var xxx_messageInfo_DeviceNames proto.InternalMessageInfo + +func (m *DeviceNames) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +func (m *DeviceNames) GetNicknames() []string { + if m != nil { + return m.Nicknames + } + return nil +} + +func (m *DeviceNames) GetDefaultNames() []string { + if m != nil { + return m.DefaultNames + } + return nil +} + +// Device information. +type DeviceInfo struct { + // Device manufacturer. + Manufacturer string `protobuf:"bytes,1,opt,name=manufacturer,proto3" json:"manufacturer,omitempty"` + // Device model. + Model string `protobuf:"bytes,2,opt,name=model,proto3" json:"model,omitempty"` + // Device hardware version. + HwVersion string `protobuf:"bytes,3,opt,name=hw_version,json=hwVersion,proto3" json:"hw_version,omitempty"` + // Device software version. + SwVersion string `protobuf:"bytes,4,opt,name=sw_version,json=swVersion,proto3" json:"sw_version,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *DeviceInfo) Reset() { *m = DeviceInfo{} } +func (m *DeviceInfo) String() string { return proto.CompactTextString(m) } +func (*DeviceInfo) ProtoMessage() {} +func (*DeviceInfo) Descriptor() ([]byte, []int) { + return fileDescriptor_device_310b64582597de9b, []int{2} +} +func (m *DeviceInfo) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_DeviceInfo.Unmarshal(m, b) +} +func (m *DeviceInfo) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_DeviceInfo.Marshal(b, m, deterministic) +} +func (dst *DeviceInfo) XXX_Merge(src proto.Message) { + xxx_messageInfo_DeviceInfo.Merge(dst, src) +} +func (m *DeviceInfo) XXX_Size() int { + return xxx_messageInfo_DeviceInfo.Size(m) +} +func (m *DeviceInfo) XXX_DiscardUnknown() { + xxx_messageInfo_DeviceInfo.DiscardUnknown(m) +} + +var xxx_messageInfo_DeviceInfo proto.InternalMessageInfo + +func (m *DeviceInfo) GetManufacturer() string { + if m != nil { + return m.Manufacturer + } + return "" +} + +func (m *DeviceInfo) GetModel() string { + if m != nil { + return m.Model + } + return "" +} + +func (m *DeviceInfo) GetHwVersion() string { + if m != nil { + return m.HwVersion + } + return "" +} + +func (m *DeviceInfo) GetSwVersion() string { + if m != nil { + return m.SwVersion + } + return "" +} + +func init() { + proto.RegisterType((*Device)(nil), "google.home.graph.v1.Device") + proto.RegisterType((*DeviceNames)(nil), "google.home.graph.v1.DeviceNames") + proto.RegisterType((*DeviceInfo)(nil), "google.home.graph.v1.DeviceInfo") +} + +func init() { + proto.RegisterFile("google/home/graph/v1/device.proto", fileDescriptor_device_310b64582597de9b) +} + +var fileDescriptor_device_310b64582597de9b = []byte{ + // 470 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x7c, 0x93, 0xc1, 0x6f, 0xd4, 0x3c, + 0x10, 0xc5, 0x95, 0xdd, 0xed, 0x7e, 0x9b, 0xd9, 0xb6, 0x9f, 0xb0, 0x2a, 0xb0, 0xa0, 0x88, 0x74, + 0x11, 0xd2, 0x8a, 0x43, 0xa2, 0x82, 0x10, 0x42, 0x9c, 0xa8, 0x7a, 0x80, 0x0b, 0x42, 0xa9, 0xc4, + 0x81, 0x4b, 0xe4, 0x4d, 0x9c, 0xc4, 0x22, 0xb1, 0x23, 0x7b, 0xb2, 0x2b, 0xee, 0x1c, 0xf8, 0xb3, + 0x51, 0xc6, 0xd9, 0x6e, 0x2b, 0x55, 0xdc, 0xc6, 0xef, 0xfd, 0x3c, 0x1e, 0x3f, 0x27, 0x70, 0x51, + 0x19, 0x53, 0x35, 0x32, 0xa9, 0x4d, 0x2b, 0x93, 0xca, 0x8a, 0xae, 0x4e, 0xb6, 0x97, 0x49, 0x21, + 0xb7, 0x2a, 0x97, 0x71, 0x67, 0x0d, 0x1a, 0x76, 0xe6, 0x91, 0x78, 0x40, 0x62, 0x42, 0xe2, 0xed, + 0xe5, 0xd3, 0xf3, 0x71, 0x23, 0x31, 0x9b, 0xbe, 0x4c, 0x1c, 0xda, 0x3e, 0x47, 0xbf, 0x67, 0xf5, + 0x67, 0x0a, 0xf3, 0x6b, 0x6a, 0xc2, 0x4e, 0x61, 0xa2, 0x0a, 0x1e, 0x44, 0xc1, 0x3a, 0x4c, 0x27, + 0xaa, 0x60, 0x0c, 0x66, 0xf8, 0xab, 0x93, 0x7c, 0x42, 0x0a, 0xd5, 0xec, 0x31, 0xcc, 0xd1, 0x0a, + 0x85, 0x8e, 0x4f, 0xa3, 0xe9, 0x3a, 0x4c, 0xc7, 0x15, 0x7b, 0x07, 0x33, 0x2d, 0x5a, 0xc9, 0x67, + 0x51, 0xb0, 0x5e, 0xbe, 0xb9, 0x88, 0x1f, 0x9a, 0x24, 0xf6, 0xe7, 0x7c, 0x15, 0xad, 0x74, 0x29, + 0xe1, 0xec, 0x35, 0x3c, 0xda, 0xa9, 0xa6, 0xc9, 0xac, 0xec, 0x8c, 0xc5, 0xcc, 0xa1, 0x40, 0xc9, + 0x8f, 0xa2, 0x60, 0xbd, 0x48, 0xff, 0x1f, 0x8c, 0x94, 0xf4, 0x9b, 0x41, 0x66, 0xcf, 0x20, 0xb4, + 0xc6, 0xb4, 0x59, 0xad, 0x34, 0xf2, 0x39, 0xcd, 0xb4, 0x18, 0x84, 0xcf, 0x4a, 0x23, 0x7b, 0x05, + 0xa7, 0xfe, 0x5a, 0xbd, 0x95, 0x9e, 0xf8, 0x8f, 0x88, 0x93, 0x5b, 0x95, 0xb0, 0x4f, 0xb0, 0xf4, + 0x89, 0x65, 0x4a, 0x97, 0x86, 0x2f, 0x68, 0xda, 0xe8, 0x5f, 0xd3, 0x7e, 0xd1, 0xa5, 0x49, 0xa1, + 0xb8, 0xad, 0xd9, 0x7b, 0x00, 0x81, 0x68, 0xd5, 0xa6, 0x47, 0xe9, 0x78, 0x48, 0x1d, 0x9e, 0xec, + 0x3b, 0xec, 0x33, 0x8e, 0x6f, 0xe8, 0xd8, 0xf4, 0x0e, 0xca, 0x5e, 0xc0, 0x32, 0xef, 0x1d, 0x9a, + 0x36, 0x2b, 0x04, 0x0a, 0x0e, 0x34, 0x1f, 0x78, 0xe9, 0x5a, 0xa0, 0x58, 0x15, 0xb0, 0xbc, 0x93, + 0xd0, 0x10, 0x3f, 0x45, 0xea, 0x1f, 0xc4, 0xe7, 0x75, 0x0e, 0xa1, 0x56, 0xf9, 0xcf, 0xa1, 0x76, + 0x7c, 0x42, 0x2f, 0x70, 0x10, 0xd8, 0x4b, 0x38, 0x29, 0x64, 0x29, 0xfa, 0x06, 0x33, 0x4f, 0xf8, + 0x37, 0x3a, 0x1e, 0x45, 0x6a, 0xbb, 0xfa, 0x1d, 0x00, 0x1c, 0xae, 0xc6, 0x56, 0x70, 0xdc, 0x0a, + 0xdd, 0x97, 0x82, 0x42, 0xb2, 0xe3, 0x69, 0xf7, 0x34, 0x76, 0x06, 0x47, 0xad, 0x29, 0x64, 0x33, + 0x7e, 0x09, 0x7e, 0xc1, 0x9e, 0x03, 0xd4, 0xbb, 0x6c, 0x2b, 0xad, 0x53, 0x46, 0xf3, 0x29, 0x59, + 0x61, 0xbd, 0xfb, 0xee, 0x85, 0xc1, 0x76, 0x07, 0x7b, 0xe6, 0x6d, 0xb7, 0xb7, 0xaf, 0x36, 0xc0, + 0x73, 0xd3, 0x3e, 0x98, 0xfc, 0xd5, 0x18, 0xc3, 0xb7, 0x21, 0xcc, 0x1f, 0x1f, 0x46, 0xa4, 0x32, + 0x8d, 0xd0, 0x55, 0x6c, 0x6c, 0x95, 0x54, 0x52, 0x53, 0xd0, 0x89, 0xb7, 0x44, 0xa7, 0xdc, 0xfd, + 0xdf, 0xe2, 0x23, 0x15, 0x9b, 0x39, 0x51, 0x6f, 0xff, 0x06, 0x00, 0x00, 0xff, 0xff, 0xb7, 0x2a, + 0xc2, 0xaf, 0x3b, 0x03, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/genproto/googleapis/home/graph/v1/homegraph.pb.go b/vendor/google.golang.org/genproto/googleapis/home/graph/v1/homegraph.pb.go index 6695dfeed..f52896188 100644 --- a/vendor/google.golang.org/genproto/googleapis/home/graph/v1/homegraph.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/home/graph/v1/homegraph.pb.go @@ -30,11 +30,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type RequestSyncDevicesRequest struct { // Required. Third-party user id issued by agent's third-party identity // provider. - AgentUserId string `protobuf:"bytes,1,opt,name=agent_user_id,json=agentUserId" json:"agent_user_id,omitempty"` + AgentUserId string `protobuf:"bytes,1,opt,name=agent_user_id,json=agentUserId,proto3" json:"agent_user_id,omitempty"` // Optional. If set, the request will be added to a queue and a response will // be returned immediately. The queue allows for de-duplication of // simultaneous requests. - Async bool `protobuf:"varint,2,opt,name=async" json:"async,omitempty"` + Async bool `protobuf:"varint,2,opt,name=async,proto3" json:"async,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -44,7 +44,7 @@ func (m *RequestSyncDevicesRequest) Reset() { *m = RequestSyncDevicesReq func (m *RequestSyncDevicesRequest) String() string { return proto.CompactTextString(m) } func (*RequestSyncDevicesRequest) ProtoMessage() {} func (*RequestSyncDevicesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{0} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{0} } func (m *RequestSyncDevicesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RequestSyncDevicesRequest.Unmarshal(m, b) @@ -90,7 +90,7 @@ func (m *RequestSyncDevicesResponse) Reset() { *m = RequestSyncDevicesRe func (m *RequestSyncDevicesResponse) String() string { return proto.CompactTextString(m) } func (*RequestSyncDevicesResponse) ProtoMessage() {} func (*RequestSyncDevicesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{1} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{1} } func (m *RequestSyncDevicesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RequestSyncDevicesResponse.Unmarshal(m, b) @@ -158,11 +158,11 @@ var xxx_messageInfo_RequestSyncDevicesResponse proto.InternalMessageInfo // Next tag: 5. type ReportStateAndNotificationRequest struct { // Request id used for debugging. - RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` // Unique identifier per event (eg: doorbell press). - EventId string `protobuf:"bytes,4,opt,name=event_id,json=eventId" json:"event_id,omitempty"` + EventId string `protobuf:"bytes,4,opt,name=event_id,json=eventId,proto3" json:"event_id,omitempty"` // Required. Third-party user id. - AgentUserId string `protobuf:"bytes,2,opt,name=agent_user_id,json=agentUserId" json:"agent_user_id,omitempty"` + AgentUserId string `protobuf:"bytes,2,opt,name=agent_user_id,json=agentUserId,proto3" json:"agent_user_id,omitempty"` // State of devices to update and notification metadata for devices. For // example, if a user turns a light on manually, a State update should be // sent so that the information is always the current status of the device. @@ -173,7 +173,7 @@ type ReportStateAndNotificationRequest struct { // include both "on" and "70 degrees" but the 3p may choose not to send any // notification for that, or to only say that the "the room is heating up", // keeping state and notification independent. - Payload *StateAndNotificationPayload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *StateAndNotificationPayload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -183,7 +183,7 @@ func (m *ReportStateAndNotificationRequest) Reset() { *m = ReportStateAn func (m *ReportStateAndNotificationRequest) String() string { return proto.CompactTextString(m) } func (*ReportStateAndNotificationRequest) ProtoMessage() {} func (*ReportStateAndNotificationRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{2} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{2} } func (m *ReportStateAndNotificationRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReportStateAndNotificationRequest.Unmarshal(m, b) @@ -234,7 +234,7 @@ func (m *ReportStateAndNotificationRequest) GetPayload() *StateAndNotificationPa // Response type for ReportStateAndNotification call. type ReportStateAndNotificationResponse struct { // Request id copied from ReportStateAndNotificationRequest. - RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -244,7 +244,7 @@ func (m *ReportStateAndNotificationResponse) Reset() { *m = ReportStateA func (m *ReportStateAndNotificationResponse) String() string { return proto.CompactTextString(m) } func (*ReportStateAndNotificationResponse) ProtoMessage() {} func (*ReportStateAndNotificationResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{3} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{3} } func (m *ReportStateAndNotificationResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReportStateAndNotificationResponse.Unmarshal(m, b) @@ -274,7 +274,7 @@ func (m *ReportStateAndNotificationResponse) GetRequestId() string { // Payload containing the State and Notification information for devices. type StateAndNotificationPayload struct { // The devices for updating State and sending Notifications. - Devices *ReportStateAndNotificationDevice `protobuf:"bytes,1,opt,name=devices" json:"devices,omitempty"` + Devices *ReportStateAndNotificationDevice `protobuf:"bytes,1,opt,name=devices,proto3" json:"devices,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -284,7 +284,7 @@ func (m *StateAndNotificationPayload) Reset() { *m = StateAndNotificatio func (m *StateAndNotificationPayload) String() string { return proto.CompactTextString(m) } func (*StateAndNotificationPayload) ProtoMessage() {} func (*StateAndNotificationPayload) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{4} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{4} } func (m *StateAndNotificationPayload) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StateAndNotificationPayload.Unmarshal(m, b) @@ -314,9 +314,9 @@ func (m *StateAndNotificationPayload) GetDevices() *ReportStateAndNotificationDe // The States and Notifications specific to a device. type ReportStateAndNotificationDevice struct { // States of devices to update. - States *_struct.Struct `protobuf:"bytes,1,opt,name=states" json:"states,omitempty"` + States *_struct.Struct `protobuf:"bytes,1,opt,name=states,proto3" json:"states,omitempty"` // Notifications metadata for devices. - Notifications *_struct.Struct `protobuf:"bytes,2,opt,name=notifications" json:"notifications,omitempty"` + Notifications *_struct.Struct `protobuf:"bytes,2,opt,name=notifications,proto3" json:"notifications,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -326,7 +326,7 @@ func (m *ReportStateAndNotificationDevice) Reset() { *m = ReportStateAnd func (m *ReportStateAndNotificationDevice) String() string { return proto.CompactTextString(m) } func (*ReportStateAndNotificationDevice) ProtoMessage() {} func (*ReportStateAndNotificationDevice) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{5} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{5} } func (m *ReportStateAndNotificationDevice) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReportStateAndNotificationDevice.Unmarshal(m, b) @@ -363,9 +363,9 @@ func (m *ReportStateAndNotificationDevice) GetNotifications() *_struct.Struct { // Request type for DeleteAgentUser call. type DeleteAgentUserRequest struct { // Request id used for debugging. - RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` // Required. Third-party user id. - AgentUserId string `protobuf:"bytes,2,opt,name=agent_user_id,json=agentUserId" json:"agent_user_id,omitempty"` + AgentUserId string `protobuf:"bytes,2,opt,name=agent_user_id,json=agentUserId,proto3" json:"agent_user_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -375,7 +375,7 @@ func (m *DeleteAgentUserRequest) Reset() { *m = DeleteAgentUserRequest{} func (m *DeleteAgentUserRequest) String() string { return proto.CompactTextString(m) } func (*DeleteAgentUserRequest) ProtoMessage() {} func (*DeleteAgentUserRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_homegraph_5bfcd56e20a4e9c3, []int{6} + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{6} } func (m *DeleteAgentUserRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteAgentUserRequest.Unmarshal(m, b) @@ -409,6 +409,465 @@ func (m *DeleteAgentUserRequest) GetAgentUserId() string { return "" } +// Request type for Query call. This should be the same format as the AoG +// action.devices.QUERY request +// (https://developers.google.com/actions/smarthome/create-app#actiondevicesquery) +// with the exception of the extra "agent_user_id" and no "intent" and +// "customData" field. +type QueryRequest struct { + // Request ID used for debugging. + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` + // Required. Third-party user ID. + AgentUserId string `protobuf:"bytes,2,opt,name=agent_user_id,json=agentUserId,proto3" json:"agent_user_id,omitempty"` + // Required. Inputs containing third-party partner's device IDs for which to + // get the device states. + Inputs []*QueryRequestInput `protobuf:"bytes,3,rep,name=inputs,proto3" json:"inputs,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *QueryRequest) Reset() { *m = QueryRequest{} } +func (m *QueryRequest) String() string { return proto.CompactTextString(m) } +func (*QueryRequest) ProtoMessage() {} +func (*QueryRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{7} +} +func (m *QueryRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_QueryRequest.Unmarshal(m, b) +} +func (m *QueryRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_QueryRequest.Marshal(b, m, deterministic) +} +func (dst *QueryRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_QueryRequest.Merge(dst, src) +} +func (m *QueryRequest) XXX_Size() int { + return xxx_messageInfo_QueryRequest.Size(m) +} +func (m *QueryRequest) XXX_DiscardUnknown() { + xxx_messageInfo_QueryRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_QueryRequest proto.InternalMessageInfo + +func (m *QueryRequest) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + +func (m *QueryRequest) GetAgentUserId() string { + if m != nil { + return m.AgentUserId + } + return "" +} + +func (m *QueryRequest) GetInputs() []*QueryRequestInput { + if m != nil { + return m.Inputs + } + return nil +} + +// Device ID inputs to QueryRequest. +type QueryRequestInput struct { + // Payload containing third-party partner's device IDs. + Payload *QueryRequestPayload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *QueryRequestInput) Reset() { *m = QueryRequestInput{} } +func (m *QueryRequestInput) String() string { return proto.CompactTextString(m) } +func (*QueryRequestInput) ProtoMessage() {} +func (*QueryRequestInput) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{8} +} +func (m *QueryRequestInput) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_QueryRequestInput.Unmarshal(m, b) +} +func (m *QueryRequestInput) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_QueryRequestInput.Marshal(b, m, deterministic) +} +func (dst *QueryRequestInput) XXX_Merge(src proto.Message) { + xxx_messageInfo_QueryRequestInput.Merge(dst, src) +} +func (m *QueryRequestInput) XXX_Size() int { + return xxx_messageInfo_QueryRequestInput.Size(m) +} +func (m *QueryRequestInput) XXX_DiscardUnknown() { + xxx_messageInfo_QueryRequestInput.DiscardUnknown(m) +} + +var xxx_messageInfo_QueryRequestInput proto.InternalMessageInfo + +func (m *QueryRequestInput) GetPayload() *QueryRequestPayload { + if m != nil { + return m.Payload + } + return nil +} + +// Payload containing device IDs. +type QueryRequestPayload struct { + // Third-party partner's device IDs to get device states for. + Devices []*AgentDeviceId `protobuf:"bytes,1,rep,name=devices,proto3" json:"devices,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *QueryRequestPayload) Reset() { *m = QueryRequestPayload{} } +func (m *QueryRequestPayload) String() string { return proto.CompactTextString(m) } +func (*QueryRequestPayload) ProtoMessage() {} +func (*QueryRequestPayload) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{9} +} +func (m *QueryRequestPayload) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_QueryRequestPayload.Unmarshal(m, b) +} +func (m *QueryRequestPayload) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_QueryRequestPayload.Marshal(b, m, deterministic) +} +func (dst *QueryRequestPayload) XXX_Merge(src proto.Message) { + xxx_messageInfo_QueryRequestPayload.Merge(dst, src) +} +func (m *QueryRequestPayload) XXX_Size() int { + return xxx_messageInfo_QueryRequestPayload.Size(m) +} +func (m *QueryRequestPayload) XXX_DiscardUnknown() { + xxx_messageInfo_QueryRequestPayload.DiscardUnknown(m) +} + +var xxx_messageInfo_QueryRequestPayload proto.InternalMessageInfo + +func (m *QueryRequestPayload) GetDevices() []*AgentDeviceId { + if m != nil { + return m.Devices + } + return nil +} + +// Third-party partner's device ID for one device. +type AgentDeviceId struct { + // Third-party partner's device ID. + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AgentDeviceId) Reset() { *m = AgentDeviceId{} } +func (m *AgentDeviceId) String() string { return proto.CompactTextString(m) } +func (*AgentDeviceId) ProtoMessage() {} +func (*AgentDeviceId) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{10} +} +func (m *AgentDeviceId) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AgentDeviceId.Unmarshal(m, b) +} +func (m *AgentDeviceId) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AgentDeviceId.Marshal(b, m, deterministic) +} +func (dst *AgentDeviceId) XXX_Merge(src proto.Message) { + xxx_messageInfo_AgentDeviceId.Merge(dst, src) +} +func (m *AgentDeviceId) XXX_Size() int { + return xxx_messageInfo_AgentDeviceId.Size(m) +} +func (m *AgentDeviceId) XXX_DiscardUnknown() { + xxx_messageInfo_AgentDeviceId.DiscardUnknown(m) +} + +var xxx_messageInfo_AgentDeviceId proto.InternalMessageInfo + +func (m *AgentDeviceId) GetId() string { + if m != nil { + return m.Id + } + return "" +} + +// Response type for Query call. This should follow the same format as AoG +// action.devices.QUERY response +// (https://developers.google.com/actions/smarthome/create-app#actiondevicesquery). +type QueryResponse struct { + // Request ID used for debugging. Copied from the request. + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` + // Device states for the devices given in the request. + Payload *QueryResponsePayload `protobuf:"bytes,2,opt,name=payload,proto3" json:"payload,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *QueryResponse) Reset() { *m = QueryResponse{} } +func (m *QueryResponse) String() string { return proto.CompactTextString(m) } +func (*QueryResponse) ProtoMessage() {} +func (*QueryResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{11} +} +func (m *QueryResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_QueryResponse.Unmarshal(m, b) +} +func (m *QueryResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_QueryResponse.Marshal(b, m, deterministic) +} +func (dst *QueryResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_QueryResponse.Merge(dst, src) +} +func (m *QueryResponse) XXX_Size() int { + return xxx_messageInfo_QueryResponse.Size(m) +} +func (m *QueryResponse) XXX_DiscardUnknown() { + xxx_messageInfo_QueryResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_QueryResponse proto.InternalMessageInfo + +func (m *QueryResponse) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + +func (m *QueryResponse) GetPayload() *QueryResponsePayload { + if m != nil { + return m.Payload + } + return nil +} + +// Payload containing device states information. +type QueryResponsePayload struct { + // States of the devices. Map of third-party device ID to struct of device + // states. + Devices map[string]*_struct.Struct `protobuf:"bytes,1,rep,name=devices,proto3" json:"devices,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *QueryResponsePayload) Reset() { *m = QueryResponsePayload{} } +func (m *QueryResponsePayload) String() string { return proto.CompactTextString(m) } +func (*QueryResponsePayload) ProtoMessage() {} +func (*QueryResponsePayload) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{12} +} +func (m *QueryResponsePayload) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_QueryResponsePayload.Unmarshal(m, b) +} +func (m *QueryResponsePayload) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_QueryResponsePayload.Marshal(b, m, deterministic) +} +func (dst *QueryResponsePayload) XXX_Merge(src proto.Message) { + xxx_messageInfo_QueryResponsePayload.Merge(dst, src) +} +func (m *QueryResponsePayload) XXX_Size() int { + return xxx_messageInfo_QueryResponsePayload.Size(m) +} +func (m *QueryResponsePayload) XXX_DiscardUnknown() { + xxx_messageInfo_QueryResponsePayload.DiscardUnknown(m) +} + +var xxx_messageInfo_QueryResponsePayload proto.InternalMessageInfo + +func (m *QueryResponsePayload) GetDevices() map[string]*_struct.Struct { + if m != nil { + return m.Devices + } + return nil +} + +// Request type for Sync call. This should follow the same format as AoG +// action.devices.SYNC request +// (https://developers.google.com/actions/smarthome/create-app#actiondevicessync) +// with the exception of the extra "agent_user_id" and no "intent" field. +type SyncRequest struct { + // Request ID used for debugging. + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` + // Required. Third-party user ID. + AgentUserId string `protobuf:"bytes,2,opt,name=agent_user_id,json=agentUserId,proto3" json:"agent_user_id,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SyncRequest) Reset() { *m = SyncRequest{} } +func (m *SyncRequest) String() string { return proto.CompactTextString(m) } +func (*SyncRequest) ProtoMessage() {} +func (*SyncRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{13} +} +func (m *SyncRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SyncRequest.Unmarshal(m, b) +} +func (m *SyncRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SyncRequest.Marshal(b, m, deterministic) +} +func (dst *SyncRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_SyncRequest.Merge(dst, src) +} +func (m *SyncRequest) XXX_Size() int { + return xxx_messageInfo_SyncRequest.Size(m) +} +func (m *SyncRequest) XXX_DiscardUnknown() { + xxx_messageInfo_SyncRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_SyncRequest proto.InternalMessageInfo + +func (m *SyncRequest) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + +func (m *SyncRequest) GetAgentUserId() string { + if m != nil { + return m.AgentUserId + } + return "" +} + +// Example SyncResponse: +// { +// "requestId": "ff36a3cc-ec34-11e6-b1a0-64510650abcf", +// "payload": { +// "agentUserId": "1836.15267389", +// "devices": [{ +// "id": "123", +// "type": "action.devices.types.OUTLET", +// "traits": [ +// "action.devices.traits.OnOff" +// ], +// "name": { +// "defaultNames": ["My Outlet 1234"], +// "name": "Night light", +// "nicknames": ["wall plug"] +// }, +// "willReportState": false, +// "deviceInfo": { +// "manufacturer": "lights-out-inc", +// "model": "hs1234", +// "hwVersion": "3.2", +// "swVersion": "11.4" +// }, +// "customData": { +// "fooValue": 74, +// "barValue": true, +// "bazValue": "foo" +// } +// }] +// } +// } +// +// Response type for Sync call. This should follow the same format as AoG +// action.devices.SYNC response +// (https://developers.google.com/actions/smarthome/create-app#actiondevicessync). +type SyncResponse struct { + // Request ID used for debugging. Copied from the request. + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` + // Devices associated with the third-party user. + Payload *SyncResponsePayload `protobuf:"bytes,2,opt,name=payload,proto3" json:"payload,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SyncResponse) Reset() { *m = SyncResponse{} } +func (m *SyncResponse) String() string { return proto.CompactTextString(m) } +func (*SyncResponse) ProtoMessage() {} +func (*SyncResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{14} +} +func (m *SyncResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SyncResponse.Unmarshal(m, b) +} +func (m *SyncResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SyncResponse.Marshal(b, m, deterministic) +} +func (dst *SyncResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_SyncResponse.Merge(dst, src) +} +func (m *SyncResponse) XXX_Size() int { + return xxx_messageInfo_SyncResponse.Size(m) +} +func (m *SyncResponse) XXX_DiscardUnknown() { + xxx_messageInfo_SyncResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_SyncResponse proto.InternalMessageInfo + +func (m *SyncResponse) GetRequestId() string { + if m != nil { + return m.RequestId + } + return "" +} + +func (m *SyncResponse) GetPayload() *SyncResponsePayload { + if m != nil { + return m.Payload + } + return nil +} + +// Payload containing device information. +type SyncResponsePayload struct { + // Third-party user ID + AgentUserId string `protobuf:"bytes,1,opt,name=agent_user_id,json=agentUserId,proto3" json:"agent_user_id,omitempty"` + // Devices associated with the third-party user. + Devices []*Device `protobuf:"bytes,2,rep,name=devices,proto3" json:"devices,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *SyncResponsePayload) Reset() { *m = SyncResponsePayload{} } +func (m *SyncResponsePayload) String() string { return proto.CompactTextString(m) } +func (*SyncResponsePayload) ProtoMessage() {} +func (*SyncResponsePayload) Descriptor() ([]byte, []int) { + return fileDescriptor_homegraph_5a60696c46a9dcf0, []int{15} +} +func (m *SyncResponsePayload) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_SyncResponsePayload.Unmarshal(m, b) +} +func (m *SyncResponsePayload) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_SyncResponsePayload.Marshal(b, m, deterministic) +} +func (dst *SyncResponsePayload) XXX_Merge(src proto.Message) { + xxx_messageInfo_SyncResponsePayload.Merge(dst, src) +} +func (m *SyncResponsePayload) XXX_Size() int { + return xxx_messageInfo_SyncResponsePayload.Size(m) +} +func (m *SyncResponsePayload) XXX_DiscardUnknown() { + xxx_messageInfo_SyncResponsePayload.DiscardUnknown(m) +} + +var xxx_messageInfo_SyncResponsePayload proto.InternalMessageInfo + +func (m *SyncResponsePayload) GetAgentUserId() string { + if m != nil { + return m.AgentUserId + } + return "" +} + +func (m *SyncResponsePayload) GetDevices() []*Device { + if m != nil { + return m.Devices + } + return nil +} + func init() { proto.RegisterType((*RequestSyncDevicesRequest)(nil), "google.home.graph.v1.RequestSyncDevicesRequest") proto.RegisterType((*RequestSyncDevicesResponse)(nil), "google.home.graph.v1.RequestSyncDevicesResponse") @@ -417,6 +876,16 @@ func init() { proto.RegisterType((*StateAndNotificationPayload)(nil), "google.home.graph.v1.StateAndNotificationPayload") proto.RegisterType((*ReportStateAndNotificationDevice)(nil), "google.home.graph.v1.ReportStateAndNotificationDevice") proto.RegisterType((*DeleteAgentUserRequest)(nil), "google.home.graph.v1.DeleteAgentUserRequest") + proto.RegisterType((*QueryRequest)(nil), "google.home.graph.v1.QueryRequest") + proto.RegisterType((*QueryRequestInput)(nil), "google.home.graph.v1.QueryRequestInput") + proto.RegisterType((*QueryRequestPayload)(nil), "google.home.graph.v1.QueryRequestPayload") + proto.RegisterType((*AgentDeviceId)(nil), "google.home.graph.v1.AgentDeviceId") + proto.RegisterType((*QueryResponse)(nil), "google.home.graph.v1.QueryResponse") + proto.RegisterType((*QueryResponsePayload)(nil), "google.home.graph.v1.QueryResponsePayload") + proto.RegisterMapType((map[string]*_struct.Struct)(nil), "google.home.graph.v1.QueryResponsePayload.DevicesEntry") + proto.RegisterType((*SyncRequest)(nil), "google.home.graph.v1.SyncRequest") + proto.RegisterType((*SyncResponse)(nil), "google.home.graph.v1.SyncResponse") + proto.RegisterType((*SyncResponsePayload)(nil), "google.home.graph.v1.SyncResponsePayload") } // Reference imports to suppress errors if they are not otherwise used. @@ -455,11 +924,33 @@ type HomeGraphApiServiceClient interface { // Unlink an agent user from Google. As result, all data related to this user // will be deleted. // - // Third-party user's identity is passed in as agent_user_id. + // Here is how the agent user is created in Google: + // When users open their Google Home App, they can begin linking a 3p + // partner. User is guided through the OAuth process. After entering the 3p + // credentials, Google gets the 3p OAuth token, and uses it to make a + // Sync call to the 3p partner and gets back all the user's data, including + // agent_user_id and devices. + // Google then creates the agent user and stores a mapping from the + // agent_user_id -> Google ID mapping. Google also stores all user's devices + // under that Google ID. + // The mapping from agent_user_id -> Google ID is many to many, since one + // Google user can have multiple 3p accounts, and multiple Google users can + // map to one agent_user_id (e.g. husband and wife share one Nest account + // username/password). + // + // Third-party user's identity is passed in as agent_user_id // Agent is identified by the JWT signed by the partner's service account. // // Note: Special characters (except "/") in agent_user_id must be URL encoded. DeleteAgentUser(ctx context.Context, in *DeleteAgentUserRequest, opts ...grpc.CallOption) (*empty.Empty, error) + // Gets the device states for the devices in QueryRequest. + // Third-party user's identity is passed in as agent_user_id. Agent is + // identified by the JWT signed by the third-party partner's service account. + Query(ctx context.Context, in *QueryRequest, opts ...grpc.CallOption) (*QueryResponse, error) + // Gets all the devices associated with the given third-party user. + // Third-party user's identity is passed in as agent_user_id. Agent is + // identified by the JWT signed by the third-party partner's service account. + Sync(ctx context.Context, in *SyncRequest, opts ...grpc.CallOption) (*SyncResponse, error) } type homeGraphApiServiceClient struct { @@ -497,8 +988,25 @@ func (c *homeGraphApiServiceClient) DeleteAgentUser(ctx context.Context, in *Del return out, nil } -// Server API for HomeGraphApiService service +func (c *homeGraphApiServiceClient) Query(ctx context.Context, in *QueryRequest, opts ...grpc.CallOption) (*QueryResponse, error) { + out := new(QueryResponse) + err := c.cc.Invoke(ctx, "/google.home.graph.v1.HomeGraphApiService/Query", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} +func (c *homeGraphApiServiceClient) Sync(ctx context.Context, in *SyncRequest, opts ...grpc.CallOption) (*SyncResponse, error) { + out := new(SyncResponse) + err := c.cc.Invoke(ctx, "/google.home.graph.v1.HomeGraphApiService/Sync", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// HomeGraphApiServiceServer is the server API for HomeGraphApiService service. type HomeGraphApiServiceServer interface { // Requests a Sync call from Google to a 3p partner's home control agent for // a user. @@ -524,11 +1032,33 @@ type HomeGraphApiServiceServer interface { // Unlink an agent user from Google. As result, all data related to this user // will be deleted. // - // Third-party user's identity is passed in as agent_user_id. + // Here is how the agent user is created in Google: + // When users open their Google Home App, they can begin linking a 3p + // partner. User is guided through the OAuth process. After entering the 3p + // credentials, Google gets the 3p OAuth token, and uses it to make a + // Sync call to the 3p partner and gets back all the user's data, including + // agent_user_id and devices. + // Google then creates the agent user and stores a mapping from the + // agent_user_id -> Google ID mapping. Google also stores all user's devices + // under that Google ID. + // The mapping from agent_user_id -> Google ID is many to many, since one + // Google user can have multiple 3p accounts, and multiple Google users can + // map to one agent_user_id (e.g. husband and wife share one Nest account + // username/password). + // + // Third-party user's identity is passed in as agent_user_id // Agent is identified by the JWT signed by the partner's service account. // // Note: Special characters (except "/") in agent_user_id must be URL encoded. DeleteAgentUser(context.Context, *DeleteAgentUserRequest) (*empty.Empty, error) + // Gets the device states for the devices in QueryRequest. + // Third-party user's identity is passed in as agent_user_id. Agent is + // identified by the JWT signed by the third-party partner's service account. + Query(context.Context, *QueryRequest) (*QueryResponse, error) + // Gets all the devices associated with the given third-party user. + // Third-party user's identity is passed in as agent_user_id. Agent is + // identified by the JWT signed by the third-party partner's service account. + Sync(context.Context, *SyncRequest) (*SyncResponse, error) } func RegisterHomeGraphApiServiceServer(s *grpc.Server, srv HomeGraphApiServiceServer) { @@ -589,6 +1119,42 @@ func _HomeGraphApiService_DeleteAgentUser_Handler(srv interface{}, ctx context.C return interceptor(ctx, in, info, handler) } +func _HomeGraphApiService_Query_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(QueryRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(HomeGraphApiServiceServer).Query(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.home.graph.v1.HomeGraphApiService/Query", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(HomeGraphApiServiceServer).Query(ctx, req.(*QueryRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _HomeGraphApiService_Sync_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(SyncRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(HomeGraphApiServiceServer).Sync(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.home.graph.v1.HomeGraphApiService/Sync", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(HomeGraphApiServiceServer).Sync(ctx, req.(*SyncRequest)) + } + return interceptor(ctx, in, info, handler) +} + var _HomeGraphApiService_serviceDesc = grpc.ServiceDesc{ ServiceName: "google.home.graph.v1.HomeGraphApiService", HandlerType: (*HomeGraphApiServiceServer)(nil), @@ -605,51 +1171,77 @@ var _HomeGraphApiService_serviceDesc = grpc.ServiceDesc{ MethodName: "DeleteAgentUser", Handler: _HomeGraphApiService_DeleteAgentUser_Handler, }, + { + MethodName: "Query", + Handler: _HomeGraphApiService_Query_Handler, + }, + { + MethodName: "Sync", + Handler: _HomeGraphApiService_Sync_Handler, + }, }, Streams: []grpc.StreamDesc{}, Metadata: "google/home/graph/v1/homegraph.proto", } func init() { - proto.RegisterFile("google/home/graph/v1/homegraph.proto", fileDescriptor_homegraph_5bfcd56e20a4e9c3) + proto.RegisterFile("google/home/graph/v1/homegraph.proto", fileDescriptor_homegraph_5a60696c46a9dcf0) } -var fileDescriptor_homegraph_5bfcd56e20a4e9c3 = []byte{ - // 575 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x54, 0xdd, 0x6a, 0x13, 0x41, - 0x14, 0x66, 0x5b, 0x6d, 0xda, 0x13, 0x8a, 0x30, 0x96, 0x76, 0x9b, 0x46, 0x48, 0x47, 0x91, 0x18, - 0x65, 0xd7, 0x44, 0xf0, 0xa7, 0xd2, 0x8b, 0xd4, 0x8a, 0x06, 0x41, 0xc2, 0x86, 0xde, 0xe8, 0x45, - 0x99, 0xee, 0x9e, 0x6e, 0x17, 0x92, 0x99, 0x75, 0x67, 0x12, 0x08, 0xe2, 0x4d, 0x1f, 0x41, 0x2f, - 0x7d, 0x1b, 0x1f, 0xc0, 0x1b, 0x5f, 0xc1, 0xd7, 0x10, 0x64, 0x67, 0x76, 0xad, 0x69, 0x36, 0x69, - 0x7a, 0x37, 0x67, 0xce, 0xdf, 0x77, 0xce, 0xf7, 0xcd, 0xc0, 0xbd, 0x50, 0x88, 0xb0, 0x8f, 0xee, - 0x99, 0x18, 0xa0, 0x1b, 0x26, 0x2c, 0x3e, 0x73, 0x47, 0x4d, 0x6d, 0x69, 0xc3, 0x89, 0x13, 0xa1, - 0x04, 0xd9, 0x30, 0x51, 0x4e, 0x7a, 0xef, 0x18, 0xc7, 0xa8, 0x59, 0xa9, 0x66, 0xb9, 0x2c, 0x8e, - 0x5c, 0xc6, 0xb9, 0x50, 0x4c, 0x45, 0x82, 0x4b, 0x93, 0x53, 0xd9, 0xc9, 0xbc, 0xda, 0x3a, 0x19, - 0x9e, 0xba, 0x38, 0x88, 0xd5, 0x38, 0x73, 0x56, 0x2f, 0x3b, 0xa5, 0x4a, 0x86, 0xbe, 0x32, 0x5e, - 0x7a, 0x04, 0xdb, 0x1e, 0x7e, 0x1a, 0xa2, 0x54, 0xbd, 0x31, 0xf7, 0x0f, 0x71, 0x14, 0xf9, 0x28, - 0xb3, 0x1b, 0x42, 0x61, 0x9d, 0x85, 0xc8, 0xd5, 0xf1, 0x50, 0x62, 0x72, 0x1c, 0x05, 0xb6, 0x55, - 0xb3, 0xea, 0x6b, 0x5e, 0x59, 0x5f, 0x1e, 0x49, 0x4c, 0x3a, 0x01, 0xd9, 0x80, 0x9b, 0x4c, 0x8e, - 0xb9, 0x6f, 0x2f, 0xd5, 0xac, 0xfa, 0xaa, 0x67, 0x0c, 0x5a, 0x85, 0x4a, 0x51, 0x59, 0x19, 0x0b, - 0x2e, 0x91, 0xfe, 0xb4, 0x60, 0xd7, 0xc3, 0x58, 0x24, 0xaa, 0xa7, 0x98, 0xc2, 0x36, 0x0f, 0xde, - 0x0b, 0x15, 0x9d, 0x46, 0xbe, 0x9e, 0x2a, 0xef, 0x7e, 0x07, 0x20, 0x31, 0xc7, 0x8b, 0xd6, 0x6b, - 0xd9, 0x4d, 0x27, 0x20, 0xdb, 0xb0, 0x8a, 0xa3, 0x14, 0x5c, 0x14, 0xd8, 0x37, 0xb4, 0xb3, 0xa4, - 0xed, 0x4e, 0x30, 0x8d, 0x7b, 0x69, 0x1a, 0xf7, 0x3b, 0x28, 0xc5, 0x6c, 0xdc, 0x17, 0x2c, 0xb0, - 0x97, 0x6b, 0x56, 0xbd, 0xdc, 0x6a, 0x3a, 0x45, 0x9b, 0x77, 0x8a, 0x10, 0x76, 0x4d, 0xa2, 0x97, - 0x57, 0xa0, 0xaf, 0x80, 0xce, 0x9b, 0xc7, 0x8c, 0x7d, 0xc5, 0x40, 0x54, 0xc0, 0xce, 0x9c, 0x66, - 0xa4, 0x0b, 0xa5, 0xc0, 0xec, 0x51, 0xa7, 0x96, 0x5b, 0x4f, 0x8b, 0x01, 0xcf, 0x06, 0x62, 0x68, - 0xf0, 0xf2, 0x32, 0xf4, 0xab, 0x05, 0xb5, 0xab, 0xa2, 0x89, 0x0b, 0x2b, 0x32, 0xf5, 0xe6, 0x5d, - 0xb7, 0xf2, 0xae, 0xb9, 0x9e, 0x9c, 0x9e, 0xd6, 0x93, 0x97, 0x85, 0x91, 0x7d, 0x58, 0xe7, 0xff, - 0x95, 0x91, 0x7a, 0xf9, 0x73, 0xf2, 0x26, 0xa3, 0xe9, 0x47, 0xd8, 0x3c, 0xc4, 0x3e, 0x2a, 0x6c, - 0xe7, 0x64, 0x2d, 0xa8, 0x87, 0x05, 0x48, 0x6f, 0xfd, 0x59, 0x86, 0xdb, 0x6f, 0xc5, 0x00, 0xdf, - 0xa4, 0xcb, 0x6a, 0xc7, 0x51, 0x0f, 0x13, 0x3d, 0xe4, 0x77, 0x0b, 0xc8, 0xb4, 0x5e, 0x89, 0x3b, - 0x6b, 0xc3, 0x33, 0x1e, 0x4c, 0xe5, 0xf1, 0xe2, 0x09, 0xd9, 0x53, 0xa0, 0xe7, 0xbf, 0x7e, 0x7f, - 0x5b, 0xaa, 0xd2, 0xad, 0xf4, 0x2f, 0xc8, 0x88, 0xd9, 0x4b, 0x2e, 0xe2, 0xf7, 0xac, 0x06, 0xf9, - 0x61, 0xa5, 0xaf, 0x69, 0x16, 0x4f, 0xe4, 0xd9, 0x75, 0x75, 0x90, 0xa3, 0x7d, 0x7e, 0xfd, 0xc4, - 0x0c, 0x75, 0x53, 0xa3, 0x7e, 0x48, 0xef, 0x4f, 0xa2, 0x9e, 0x95, 0x97, 0x0e, 0x71, 0x6e, 0xc1, - 0xad, 0x4b, 0xc4, 0x92, 0x47, 0xc5, 0x00, 0x8a, 0xf9, 0xaf, 0x6c, 0x4e, 0x29, 0xe8, 0x75, 0xfa, - 0xcd, 0xd1, 0x07, 0x1a, 0xcc, 0xdd, 0xc6, 0x6e, 0x0a, 0xe6, 0xf3, 0x84, 0x06, 0xf6, 0xff, 0x91, - 0x2f, 0xdd, 0x46, 0xe3, 0xcb, 0x81, 0x00, 0xdb, 0x17, 0x83, 0xc2, 0xae, 0x07, 0x76, 0x81, 0x30, - 0xba, 0x69, 0xa7, 0x0f, 0x2f, 0xb2, 0xf8, 0x50, 0xf4, 0x19, 0x0f, 0x1d, 0x91, 0x84, 0x6e, 0x88, - 0x5c, 0xa3, 0x70, 0x8d, 0x8b, 0xc5, 0x91, 0x9c, 0xfc, 0xd7, 0x5f, 0xea, 0xc3, 0xc9, 0x8a, 0x8e, - 0x7a, 0xf2, 0x37, 0x00, 0x00, 0xff, 0xff, 0x92, 0xb2, 0x5b, 0xe2, 0xfc, 0x05, 0x00, 0x00, +var fileDescriptor_homegraph_5a60696c46a9dcf0 = []byte{ + // 850 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x56, 0xcd, 0x6e, 0xeb, 0x44, + 0x14, 0x96, 0x93, 0x9b, 0xf6, 0xde, 0x93, 0x06, 0xee, 0x9d, 0x96, 0x36, 0x4d, 0x83, 0x48, 0xa6, + 0x08, 0xd2, 0x00, 0x36, 0x09, 0x52, 0x5b, 0x8a, 0x2a, 0xd4, 0x3f, 0x41, 0x84, 0x84, 0x52, 0x87, + 0x4a, 0x08, 0x16, 0x95, 0x1b, 0x4f, 0x5d, 0x43, 0xe2, 0x71, 0x3d, 0x93, 0x48, 0x11, 0x62, 0xd3, + 0x47, 0x80, 0x25, 0xef, 0xc2, 0x82, 0x07, 0x60, 0xc3, 0x8a, 0x3d, 0x0f, 0x82, 0x66, 0xc6, 0x6e, + 0x7e, 0xec, 0x24, 0xae, 0xd4, 0x9d, 0x67, 0xce, 0x39, 0xdf, 0x39, 0x67, 0xbe, 0x6f, 0xce, 0x18, + 0xde, 0x77, 0x28, 0x75, 0x7a, 0xc4, 0xb8, 0xa3, 0x7d, 0x62, 0x38, 0x81, 0xe5, 0xdf, 0x19, 0xc3, + 0x86, 0x5c, 0xc9, 0x85, 0xee, 0x07, 0x94, 0x53, 0xb4, 0xa1, 0xbc, 0x74, 0xb1, 0xaf, 0x2b, 0xc3, + 0xb0, 0x51, 0x2a, 0x87, 0xb1, 0x96, 0xef, 0x1a, 0x96, 0xe7, 0x51, 0x6e, 0x71, 0x97, 0x7a, 0x4c, + 0xc5, 0x94, 0xaa, 0x89, 0xc8, 0x36, 0x19, 0xba, 0x5d, 0x12, 0xba, 0xec, 0x84, 0x2e, 0x72, 0x75, + 0x33, 0xb8, 0x35, 0x48, 0xdf, 0xe7, 0xa3, 0xd0, 0x58, 0x9e, 0x35, 0x32, 0x1e, 0x0c, 0xba, 0x5c, + 0x59, 0xf1, 0x15, 0x6c, 0x9b, 0xe4, 0x7e, 0x40, 0x18, 0xef, 0x8c, 0xbc, 0xee, 0xb9, 0x44, 0x65, + 0xe1, 0x0e, 0xc2, 0x50, 0xb0, 0x1c, 0xe2, 0xf1, 0xeb, 0x01, 0x23, 0xc1, 0xb5, 0x6b, 0x17, 0xb5, + 0x8a, 0x56, 0x7b, 0x65, 0xe6, 0xe5, 0xe6, 0x15, 0x23, 0x41, 0xcb, 0x46, 0x1b, 0x90, 0xb3, 0xd8, + 0xc8, 0xeb, 0x16, 0x33, 0x15, 0xad, 0xf6, 0xd2, 0x54, 0x0b, 0x5c, 0x86, 0x52, 0x12, 0x2c, 0xf3, + 0xa9, 0xc7, 0x08, 0xfe, 0x5b, 0x83, 0xaa, 0x49, 0x7c, 0x1a, 0xf0, 0x0e, 0xb7, 0x38, 0x39, 0xf1, + 0xec, 0x6f, 0x29, 0x77, 0x6f, 0xdd, 0xae, 0x6c, 0x3c, 0xca, 0xfe, 0x2e, 0x40, 0xa0, 0x3e, 0xc7, + 0xa9, 0x5f, 0x85, 0x3b, 0x2d, 0x1b, 0x6d, 0xc3, 0x4b, 0x32, 0x14, 0xc5, 0xb9, 0x76, 0xf1, 0x85, + 0x34, 0xae, 0xca, 0x75, 0xcb, 0x8e, 0xd7, 0x9d, 0x89, 0xd7, 0xfd, 0x0d, 0xac, 0xfa, 0xd6, 0xa8, + 0x47, 0x2d, 0xbb, 0x98, 0xad, 0x68, 0xb5, 0x7c, 0xb3, 0xa1, 0x27, 0x91, 0xa3, 0x27, 0x55, 0xd8, + 0x56, 0x81, 0x66, 0x84, 0x80, 0xcf, 0x00, 0x2f, 0xea, 0x47, 0xb5, 0xbd, 0xa4, 0x21, 0x4c, 0x61, + 0x67, 0x41, 0x32, 0xd4, 0x86, 0x55, 0x45, 0x3a, 0x93, 0xa1, 0xf9, 0xe6, 0x7e, 0x72, 0xc1, 0xf3, + 0x0b, 0x51, 0x34, 0x98, 0x11, 0x0c, 0xfe, 0x4d, 0x83, 0xca, 0x32, 0x6f, 0x64, 0xc0, 0x0a, 0x13, + 0xd6, 0x28, 0xeb, 0x56, 0x94, 0x35, 0xd2, 0x93, 0xde, 0x91, 0x7a, 0x32, 0x43, 0x37, 0x74, 0x0c, + 0x05, 0x6f, 0x02, 0x86, 0xc9, 0xc3, 0x5f, 0x10, 0x37, 0xed, 0x8d, 0x7f, 0x84, 0xcd, 0x73, 0xd2, + 0x23, 0x9c, 0x9c, 0x44, 0x64, 0xa5, 0xd4, 0x43, 0x0a, 0xd2, 0x45, 0xc7, 0x6b, 0x97, 0x03, 0x12, + 0x8c, 0x9e, 0x0f, 0x13, 0x7d, 0x09, 0x2b, 0xae, 0xe7, 0x0f, 0x38, 0x2b, 0x66, 0x2b, 0xd9, 0x5a, + 0xbe, 0xf9, 0x61, 0x32, 0x2d, 0x93, 0x69, 0x5b, 0xc2, 0xdf, 0x0c, 0xc3, 0xf0, 0xf7, 0xf0, 0x26, + 0x66, 0x44, 0x67, 0x63, 0x79, 0xaa, 0x73, 0xdf, 0x5b, 0x0e, 0x1b, 0x93, 0xe5, 0x77, 0xb0, 0x9e, + 0x60, 0x47, 0xc7, 0x93, 0x4a, 0x12, 0x25, 0xef, 0x26, 0x63, 0x4b, 0x06, 0x94, 0x0c, 0x5a, 0xf6, + 0x58, 0x36, 0xef, 0x41, 0x61, 0xca, 0x82, 0xde, 0x82, 0xcc, 0xe3, 0xe1, 0x65, 0x5c, 0x1b, 0x73, + 0x28, 0x84, 0x69, 0x53, 0x09, 0x1f, 0x9d, 0x8f, 0x7b, 0x55, 0x5a, 0xa9, 0x2f, 0xec, 0x55, 0x81, + 0xc6, 0x9a, 0xfd, 0x53, 0x83, 0x8d, 0x24, 0x0f, 0x74, 0x39, 0xdb, 0xee, 0x41, 0x7a, 0x78, 0x3d, + 0x1c, 0x5d, 0x17, 0x1e, 0x0f, 0x46, 0x8f, 0x47, 0x50, 0xea, 0xc0, 0xda, 0xa4, 0x01, 0xbd, 0x86, + 0xec, 0xcf, 0x64, 0x14, 0x76, 0x26, 0x3e, 0xd1, 0x27, 0x90, 0x1b, 0x5a, 0xbd, 0x01, 0x59, 0xa6, + 0x7e, 0xe5, 0x75, 0x94, 0x39, 0xd4, 0x70, 0x1b, 0xf2, 0x62, 0x58, 0x3e, 0xa3, 0xdc, 0x03, 0x58, + 0x53, 0x88, 0xe9, 0x78, 0x38, 0x9b, 0xe5, 0x61, 0x8e, 0xe6, 0x26, 0x31, 0x63, 0x34, 0xdc, 0xc3, + 0x7a, 0x82, 0x3d, 0xd5, 0x53, 0xb2, 0x3f, 0x26, 0x2a, 0x23, 0x89, 0x2a, 0x27, 0xe7, 0x9f, 0x99, + 0x63, 0xcd, 0x7f, 0x73, 0xb0, 0xfe, 0x35, 0xed, 0x93, 0xaf, 0x84, 0xc3, 0x89, 0xef, 0x76, 0x48, + 0x20, 0x47, 0xd7, 0x1f, 0x1a, 0xa0, 0xf8, 0x2b, 0x84, 0x8c, 0x79, 0x73, 0x73, 0xce, 0x33, 0x58, + 0xfa, 0x34, 0x7d, 0x40, 0xf8, 0xc0, 0xe1, 0x87, 0x7f, 0xfe, 0xfb, 0x3d, 0x53, 0xc6, 0x5b, 0xe3, + 0xa7, 0x9a, 0x1d, 0x05, 0x63, 0xff, 0x23, 0xad, 0x8e, 0xfe, 0xd2, 0xc4, 0x1b, 0x39, 0x6f, 0xfa, + 0xa2, 0x83, 0xa7, 0x4e, 0xf7, 0xa8, 0xda, 0xc3, 0xa7, 0x07, 0x86, 0x55, 0x37, 0x64, 0xd5, 0x1f, + 0xe1, 0x0f, 0xa6, 0xab, 0x9e, 0x17, 0x27, 0x9a, 0x78, 0xd0, 0xe0, 0xed, 0x99, 0x71, 0x8d, 0x3e, + 0x9e, 0xc7, 0x5a, 0xd2, 0x54, 0x2f, 0x6d, 0xc6, 0x6e, 0xc6, 0x85, 0xf8, 0x79, 0xc1, 0x7b, 0xb2, + 0x98, 0xdd, 0x7a, 0x55, 0x14, 0xf3, 0xcb, 0x94, 0x76, 0x8e, 0x1f, 0x45, 0xc3, 0x8c, 0x7a, 0xfd, + 0x57, 0xe4, 0x41, 0x4e, 0xde, 0x5d, 0x84, 0x97, 0xcf, 0xc8, 0xd2, 0x6e, 0x8a, 0xcb, 0x8f, 0xcb, + 0x32, 0xf9, 0x26, 0x7e, 0x33, 0x79, 0x12, 0xf7, 0xc2, 0x45, 0x34, 0xfd, 0x13, 0xbc, 0x10, 0x24, + 0xa2, 0xea, 0xa2, 0xeb, 0xa1, 0xb2, 0xe1, 0xe5, 0x37, 0x08, 0xef, 0xc8, 0x64, 0xef, 0xe0, 0xd7, + 0x93, 0xc9, 0x98, 0x52, 0xc9, 0x29, 0x85, 0x62, 0x97, 0xf6, 0x13, 0x51, 0x4e, 0x8b, 0x09, 0xa2, + 0x6f, 0x8b, 0x53, 0xfc, 0xe1, 0xf3, 0xd0, 0xdf, 0xa1, 0x3d, 0xcb, 0x73, 0x74, 0x1a, 0x38, 0x86, + 0x43, 0x3c, 0x79, 0xc2, 0x86, 0x32, 0x59, 0xbe, 0xcb, 0xa6, 0x7f, 0x29, 0xbf, 0x90, 0x1f, 0x37, + 0x2b, 0xd2, 0xeb, 0xb3, 0xff, 0x03, 0x00, 0x00, 0xff, 0xff, 0x8d, 0x03, 0xb1, 0x23, 0xd1, 0x0a, + 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/iam/admin/v1/iam.pb.go b/vendor/google.golang.org/genproto/googleapis/iam/admin/v1/iam.pb.go index 099da8e82..fec056089 100644 --- a/vendor/google.golang.org/genproto/googleapis/iam/admin/v1/iam.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/iam/admin/v1/iam.pb.go @@ -313,22 +313,22 @@ type ServiceAccount struct { // // In responses the resource name will always be in the format // `projects/{PROJECT_ID}/serviceAccounts/{SERVICE_ACCOUNT_EMAIL}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // @OutputOnly The id of the project that owns the service account. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // @OutputOnly The unique and stable id of the service account. - UniqueId string `protobuf:"bytes,4,opt,name=unique_id,json=uniqueId" json:"unique_id,omitempty"` + UniqueId string `protobuf:"bytes,4,opt,name=unique_id,json=uniqueId,proto3" json:"unique_id,omitempty"` // @OutputOnly The email address of the service account. - Email string `protobuf:"bytes,5,opt,name=email" json:"email,omitempty"` + Email string `protobuf:"bytes,5,opt,name=email,proto3" json:"email,omitempty"` // Optional. A user-specified description of the service account. Must be // fewer than 100 UTF-8 bytes. - DisplayName string `protobuf:"bytes,6,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,6,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Used to perform a consistent read-modify-write. Etag []byte `protobuf:"bytes,7,opt,name=etag,proto3" json:"etag,omitempty"` // @OutputOnly. The OAuth2 client id for the service account. // This is used in conjunction with the OAuth2 clientconfig API to make // three legged OAuth2 (3LO) flows to access the data of Google users. - Oauth2ClientId string `protobuf:"bytes,9,opt,name=oauth2_client_id,json=oauth2ClientId" json:"oauth2_client_id,omitempty"` + Oauth2ClientId string `protobuf:"bytes,9,opt,name=oauth2_client_id,json=oauth2ClientId,proto3" json:"oauth2_client_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -411,16 +411,16 @@ func (m *ServiceAccount) GetOauth2ClientId() string { type CreateServiceAccountRequest struct { // Required. The resource name of the project associated with the service // accounts, such as `projects/my-project-123`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The account id that is used to generate the service account // email address and a stable unique id. It is unique within a project, // must be 6-30 characters long, and match the regular expression // `[a-z]([-a-z0-9]*[a-z0-9])` to comply with RFC1035. - AccountId string `protobuf:"bytes,2,opt,name=account_id,json=accountId" json:"account_id,omitempty"` + AccountId string `protobuf:"bytes,2,opt,name=account_id,json=accountId,proto3" json:"account_id,omitempty"` // The [ServiceAccount][google.iam.admin.v1.ServiceAccount] resource to create. // Currently, only the following values are user assignable: // `display_name` . - ServiceAccount *ServiceAccount `protobuf:"bytes,3,opt,name=service_account,json=serviceAccount" json:"service_account,omitempty"` + ServiceAccount *ServiceAccount `protobuf:"bytes,3,opt,name=service_account,json=serviceAccount,proto3" json:"service_account,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -475,15 +475,15 @@ func (m *CreateServiceAccountRequest) GetServiceAccount() *ServiceAccount { type ListServiceAccountsRequest struct { // Required. The resource name of the project associated with the service // accounts, such as `projects/my-project-123`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional limit on the number of service accounts to include in the // response. Further accounts can subsequently be obtained by including the // [ListServiceAccountsResponse.next_page_token][google.iam.admin.v1.ListServiceAccountsResponse.next_page_token] // in a subsequent request. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional pagination token returned in an earlier // [ListServiceAccountsResponse.next_page_token][google.iam.admin.v1.ListServiceAccountsResponse.next_page_token]. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -537,11 +537,11 @@ func (m *ListServiceAccountsRequest) GetPageToken() string { // The service account list response. type ListServiceAccountsResponse struct { // The list of matching service accounts. - Accounts []*ServiceAccount `protobuf:"bytes,1,rep,name=accounts" json:"accounts,omitempty"` + Accounts []*ServiceAccount `protobuf:"bytes,1,rep,name=accounts,proto3" json:"accounts,omitempty"` // To retrieve the next page of results, set // [ListServiceAccountsRequest.page_token][google.iam.admin.v1.ListServiceAccountsRequest.page_token] // to this value. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -592,7 +592,7 @@ type GetServiceAccountRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -636,7 +636,7 @@ type DeleteServiceAccountRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -681,11 +681,11 @@ type ListServiceAccountKeysRequest struct { // Using `-` as a wildcard for the project, will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Filters the types of keys the user wants to include in the list // response. Duplicate key types are not allowed. If no key type // is provided, all keys are returned. - KeyTypes []ListServiceAccountKeysRequest_KeyType `protobuf:"varint,2,rep,packed,name=key_types,json=keyTypes,enum=google.iam.admin.v1.ListServiceAccountKeysRequest_KeyType" json:"key_types,omitempty"` + KeyTypes []ListServiceAccountKeysRequest_KeyType `protobuf:"varint,2,rep,packed,name=key_types,json=keyTypes,proto3,enum=google.iam.admin.v1.ListServiceAccountKeysRequest_KeyType" json:"key_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -732,7 +732,7 @@ func (m *ListServiceAccountKeysRequest) GetKeyTypes() []ListServiceAccountKeysRe // The service account keys list response. type ListServiceAccountKeysResponse struct { // The public keys for the service account. - Keys []*ServiceAccountKey `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*ServiceAccountKey `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -777,10 +777,10 @@ type GetServiceAccountKeyRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The output format of the public key requested. // X509_PEM is the default output format. - PublicKeyType ServiceAccountPublicKeyType `protobuf:"varint,2,opt,name=public_key_type,json=publicKeyType,enum=google.iam.admin.v1.ServiceAccountPublicKeyType" json:"public_key_type,omitempty"` + PublicKeyType ServiceAccountPublicKeyType `protobuf:"varint,2,opt,name=public_key_type,json=publicKeyType,proto3,enum=google.iam.admin.v1.ServiceAccountPublicKeyType" json:"public_key_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -843,16 +843,16 @@ func (m *GetServiceAccountKeyRequest) GetPublicKeyType() ServiceAccountPublicKey type ServiceAccountKey struct { // The resource name of the service account key in the following format // `projects/{PROJECT_ID}/serviceAccounts/{SERVICE_ACCOUNT_EMAIL}/keys/{key}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The output format for the private key. // Only provided in `CreateServiceAccountKey` responses, not // in `GetServiceAccountKey` or `ListServiceAccountKey` responses. // // Google never exposes system-managed private keys, and never retains // user-managed private keys. - PrivateKeyType ServiceAccountPrivateKeyType `protobuf:"varint,2,opt,name=private_key_type,json=privateKeyType,enum=google.iam.admin.v1.ServiceAccountPrivateKeyType" json:"private_key_type,omitempty"` + PrivateKeyType ServiceAccountPrivateKeyType `protobuf:"varint,2,opt,name=private_key_type,json=privateKeyType,proto3,enum=google.iam.admin.v1.ServiceAccountPrivateKeyType" json:"private_key_type,omitempty"` // Specifies the algorithm (and possibly key size) for the key. - KeyAlgorithm ServiceAccountKeyAlgorithm `protobuf:"varint,8,opt,name=key_algorithm,json=keyAlgorithm,enum=google.iam.admin.v1.ServiceAccountKeyAlgorithm" json:"key_algorithm,omitempty"` + KeyAlgorithm ServiceAccountKeyAlgorithm `protobuf:"varint,8,opt,name=key_algorithm,json=keyAlgorithm,proto3,enum=google.iam.admin.v1.ServiceAccountKeyAlgorithm" json:"key_algorithm,omitempty"` // The private key data. Only provided in `CreateServiceAccountKey` // responses. Make sure to keep the private key data secure because it // allows for the assertion of the service account identity. @@ -864,9 +864,9 @@ type ServiceAccountKey struct { // The public key data. Only provided in `GetServiceAccountKey` responses. PublicKeyData []byte `protobuf:"bytes,7,opt,name=public_key_data,json=publicKeyData,proto3" json:"public_key_data,omitempty"` // The key can be used after this timestamp. - ValidAfterTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=valid_after_time,json=validAfterTime" json:"valid_after_time,omitempty"` + ValidAfterTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=valid_after_time,json=validAfterTime,proto3" json:"valid_after_time,omitempty"` // The key can be used before this timestamp. - ValidBeforeTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=valid_before_time,json=validBeforeTime" json:"valid_before_time,omitempty"` + ValidBeforeTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=valid_before_time,json=validBeforeTime,proto3" json:"valid_before_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -952,14 +952,14 @@ type CreateServiceAccountKeyRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The output format of the private key. `GOOGLE_CREDENTIALS_FILE` is the // default output format. - PrivateKeyType ServiceAccountPrivateKeyType `protobuf:"varint,2,opt,name=private_key_type,json=privateKeyType,enum=google.iam.admin.v1.ServiceAccountPrivateKeyType" json:"private_key_type,omitempty"` + PrivateKeyType ServiceAccountPrivateKeyType `protobuf:"varint,2,opt,name=private_key_type,json=privateKeyType,proto3,enum=google.iam.admin.v1.ServiceAccountPrivateKeyType" json:"private_key_type,omitempty"` // Which type of key and algorithm to use for the key. // The default is currently a 2K RSA key. However this may change in the // future. - KeyAlgorithm ServiceAccountKeyAlgorithm `protobuf:"varint,3,opt,name=key_algorithm,json=keyAlgorithm,enum=google.iam.admin.v1.ServiceAccountKeyAlgorithm" json:"key_algorithm,omitempty"` + KeyAlgorithm ServiceAccountKeyAlgorithm `protobuf:"varint,3,opt,name=key_algorithm,json=keyAlgorithm,proto3,enum=google.iam.admin.v1.ServiceAccountKeyAlgorithm" json:"key_algorithm,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1017,7 +1017,7 @@ type DeleteServiceAccountKeyRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1061,7 +1061,7 @@ type SignBlobRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The bytes to sign. BytesToSign []byte `protobuf:"bytes,2,opt,name=bytes_to_sign,json=bytesToSign,proto3" json:"bytes_to_sign,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1110,7 +1110,7 @@ func (m *SignBlobRequest) GetBytesToSign() []byte { // The service account sign blob response. type SignBlobResponse struct { // The id of the key used to sign the blob. - KeyId string `protobuf:"bytes,1,opt,name=key_id,json=keyId" json:"key_id,omitempty"` + KeyId string `protobuf:"bytes,1,opt,name=key_id,json=keyId,proto3" json:"key_id,omitempty"` // The signed blob. Signature []byte `protobuf:"bytes,2,opt,name=signature,proto3" json:"signature,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1163,9 +1163,9 @@ type SignJwtRequest struct { // Using `-` as a wildcard for the project will infer the project from // the account. The `account` value can be the `email` address or the // `unique_id` of the service account. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The JWT payload to sign, a JSON JWT Claim set. - Payload string `protobuf:"bytes,2,opt,name=payload" json:"payload,omitempty"` + Payload string `protobuf:"bytes,2,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1212,9 +1212,9 @@ func (m *SignJwtRequest) GetPayload() string { // The service account sign JWT response. type SignJwtResponse struct { // The id of the key used to sign the JWT. - KeyId string `protobuf:"bytes,1,opt,name=key_id,json=keyId" json:"key_id,omitempty"` + KeyId string `protobuf:"bytes,1,opt,name=key_id,json=keyId,proto3" json:"key_id,omitempty"` // The signed JWT. - SignedJwt string `protobuf:"bytes,2,opt,name=signed_jwt,json=signedJwt" json:"signed_jwt,omitempty"` + SignedJwt string `protobuf:"bytes,2,opt,name=signed_jwt,json=signedJwt,proto3" json:"signed_jwt,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1267,21 +1267,21 @@ type Role struct { // When Role is used in output and other input such as UpdateRole, the role // name is the complete path, e.g., roles/logging.viewer for curated roles // and organizations/{ORGANIZATION_ID}/roles/logging.viewer for custom roles. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. A human-readable title for the role. Typically this // is limited to 100 UTF-8 bytes. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // Optional. A human-readable description for the role. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The names of the permissions this role grants when bound in an IAM policy. - IncludedPermissions []string `protobuf:"bytes,7,rep,name=included_permissions,json=includedPermissions" json:"included_permissions,omitempty"` + IncludedPermissions []string `protobuf:"bytes,7,rep,name=included_permissions,json=includedPermissions,proto3" json:"included_permissions,omitempty"` // The current launch stage of the role. - Stage Role_RoleLaunchStage `protobuf:"varint,8,opt,name=stage,enum=google.iam.admin.v1.Role_RoleLaunchStage" json:"stage,omitempty"` + Stage Role_RoleLaunchStage `protobuf:"varint,8,opt,name=stage,proto3,enum=google.iam.admin.v1.Role_RoleLaunchStage" json:"stage,omitempty"` // Used to perform a consistent read-modify-write. Etag []byte `protobuf:"bytes,9,opt,name=etag,proto3" json:"etag,omitempty"` // The current deleted state of the role. This field is read only. // It will be ignored in calls to CreateRole and UpdateRole. - Deleted bool `protobuf:"varint,11,opt,name=deleted" json:"deleted,omitempty"` + Deleted bool `protobuf:"varint,11,opt,name=deleted,proto3" json:"deleted,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1367,13 +1367,13 @@ type QueryGrantableRolesRequest struct { // The name follows the Google Cloud Platform resource format. // For example, a Cloud Platform project with id `my-project` will be named // `//cloudresourcemanager.googleapis.com/projects/my-project`. - FullResourceName string `protobuf:"bytes,1,opt,name=full_resource_name,json=fullResourceName" json:"full_resource_name,omitempty"` - View RoleView `protobuf:"varint,2,opt,name=view,enum=google.iam.admin.v1.RoleView" json:"view,omitempty"` + FullResourceName string `protobuf:"bytes,1,opt,name=full_resource_name,json=fullResourceName,proto3" json:"full_resource_name,omitempty"` + View RoleView `protobuf:"varint,2,opt,name=view,proto3,enum=google.iam.admin.v1.RoleView" json:"view,omitempty"` // Optional limit on the number of roles to include in the response. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional pagination token returned in an earlier // QueryGrantableRolesResponse. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1434,10 +1434,10 @@ func (m *QueryGrantableRolesRequest) GetPageToken() string { // The grantable role query response. type QueryGrantableRolesResponse struct { // The list of matching roles. - Roles []*Role `protobuf:"bytes,1,rep,name=roles" json:"roles,omitempty"` + Roles []*Role `protobuf:"bytes,1,rep,name=roles,proto3" json:"roles,omitempty"` // To retrieve the next page of results, set // `QueryGrantableRolesRequest.page_token` to this value. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1487,15 +1487,15 @@ type ListRolesRequest struct { // `` (empty string) -- this refers to curated roles. // `organizations/{ORGANIZATION_ID}` // `projects/{PROJECT_ID}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional limit on the number of roles to include in the response. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional pagination token returned in an earlier ListRolesResponse. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional view for the returned Role objects. - View RoleView `protobuf:"varint,4,opt,name=view,enum=google.iam.admin.v1.RoleView" json:"view,omitempty"` + View RoleView `protobuf:"varint,4,opt,name=view,proto3,enum=google.iam.admin.v1.RoleView" json:"view,omitempty"` // Include Roles that have been deleted. - ShowDeleted bool `protobuf:"varint,6,opt,name=show_deleted,json=showDeleted" json:"show_deleted,omitempty"` + ShowDeleted bool `protobuf:"varint,6,opt,name=show_deleted,json=showDeleted,proto3" json:"show_deleted,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1563,10 +1563,10 @@ func (m *ListRolesRequest) GetShowDeleted() bool { // The response containing the roles defined under a resource. type ListRolesResponse struct { // The Roles defined on this resource. - Roles []*Role `protobuf:"bytes,1,rep,name=roles" json:"roles,omitempty"` + Roles []*Role `protobuf:"bytes,1,rep,name=roles,proto3" json:"roles,omitempty"` // To retrieve the next page of results, set // `ListRolesRequest.page_token` to this value. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1616,7 +1616,7 @@ type GetRoleRequest struct { // `roles/{ROLE_NAME}` // `organizations/{ORGANIZATION_ID}/roles/{ROLE_NAME}` // `projects/{PROJECT_ID}/roles/{ROLE_NAME}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1658,11 +1658,11 @@ type CreateRoleRequest struct { // The resource name of the parent resource in one of the following formats: // `organizations/{ORGANIZATION_ID}` // `projects/{PROJECT_ID}` - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The role id to use for this role. - RoleId string `protobuf:"bytes,2,opt,name=role_id,json=roleId" json:"role_id,omitempty"` + RoleId string `protobuf:"bytes,2,opt,name=role_id,json=roleId,proto3" json:"role_id,omitempty"` // The Role resource to create. - Role *Role `protobuf:"bytes,3,opt,name=role" json:"role,omitempty"` + Role *Role `protobuf:"bytes,3,opt,name=role,proto3" json:"role,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1719,11 +1719,11 @@ type UpdateRoleRequest struct { // `roles/{ROLE_NAME}` // `organizations/{ORGANIZATION_ID}/roles/{ROLE_NAME}` // `projects/{PROJECT_ID}/roles/{ROLE_NAME}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The updated role. - Role *Role `protobuf:"bytes,2,opt,name=role" json:"role,omitempty"` + Role *Role `protobuf:"bytes,2,opt,name=role,proto3" json:"role,omitempty"` // A mask describing which fields in the Role have changed. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1779,7 +1779,7 @@ type DeleteRoleRequest struct { // The resource name of the role in one of the following formats: // `organizations/{ORGANIZATION_ID}/roles/{ROLE_NAME}` // `projects/{PROJECT_ID}/roles/{ROLE_NAME}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Used to perform a consistent read-modify-write. Etag []byte `protobuf:"bytes,2,opt,name=etag,proto3" json:"etag,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1830,7 +1830,7 @@ type UndeleteRoleRequest struct { // The resource name of the role in one of the following formats: // `organizations/{ORGANIZATION_ID}/roles/{ROLE_NAME}` // `projects/{PROJECT_ID}/roles/{ROLE_NAME}` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Used to perform a consistent read-modify-write. Etag []byte `protobuf:"bytes,2,opt,name=etag,proto3" json:"etag,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1879,17 +1879,17 @@ func (m *UndeleteRoleRequest) GetEtag() []byte { // A permission which can be included by a role. type Permission struct { // The name of this Permission. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The title of this Permission. - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` // A brief description of what this Permission is used for. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // This permission can ONLY be used in predefined roles. - OnlyInPredefinedRoles bool `protobuf:"varint,4,opt,name=only_in_predefined_roles,json=onlyInPredefinedRoles" json:"only_in_predefined_roles,omitempty"` + OnlyInPredefinedRoles bool `protobuf:"varint,4,opt,name=only_in_predefined_roles,json=onlyInPredefinedRoles,proto3" json:"only_in_predefined_roles,omitempty"` // The current launch stage of the permission. - Stage Permission_PermissionLaunchStage `protobuf:"varint,5,opt,name=stage,enum=google.iam.admin.v1.Permission_PermissionLaunchStage" json:"stage,omitempty"` + Stage Permission_PermissionLaunchStage `protobuf:"varint,5,opt,name=stage,proto3,enum=google.iam.admin.v1.Permission_PermissionLaunchStage" json:"stage,omitempty"` // The current custom role support level. - CustomRolesSupportLevel Permission_CustomRolesSupportLevel `protobuf:"varint,6,opt,name=custom_roles_support_level,json=customRolesSupportLevel,enum=google.iam.admin.v1.Permission_CustomRolesSupportLevel" json:"custom_roles_support_level,omitempty"` + CustomRolesSupportLevel Permission_CustomRolesSupportLevel `protobuf:"varint,6,opt,name=custom_roles_support_level,json=customRolesSupportLevel,proto3,enum=google.iam.admin.v1.Permission_CustomRolesSupportLevel" json:"custom_roles_support_level,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1969,12 +1969,12 @@ type QueryTestablePermissionsRequest struct { // The name follows the Google Cloud Platform resource format. // For example, a Cloud Platform project with id `my-project` will be named // `//cloudresourcemanager.googleapis.com/projects/my-project`. - FullResourceName string `protobuf:"bytes,1,opt,name=full_resource_name,json=fullResourceName" json:"full_resource_name,omitempty"` + FullResourceName string `protobuf:"bytes,1,opt,name=full_resource_name,json=fullResourceName,proto3" json:"full_resource_name,omitempty"` // Optional limit on the number of permissions to include in the response. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional pagination token returned in an earlier // QueryTestablePermissionsRequest. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2028,10 +2028,10 @@ func (m *QueryTestablePermissionsRequest) GetPageToken() string { // The response containing permissions which can be tested on a resource. type QueryTestablePermissionsResponse struct { // The Permissions testable on the requested resource. - Permissions []*Permission `protobuf:"bytes,1,rep,name=permissions" json:"permissions,omitempty"` + Permissions []*Permission `protobuf:"bytes,1,rep,name=permissions,proto3" json:"permissions,omitempty"` // To retrieve the next page of results, set // `QueryTestableRolesRequest.page_token` to this value. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2402,8 +2402,7 @@ func (c *iAMClient) QueryTestablePermissions(ctx context.Context, in *QueryTesta return out, nil } -// Server API for IAM service - +// IAMServer is the server API for IAM service. type IAMServer interface { // Lists [ServiceAccounts][google.iam.admin.v1.ServiceAccount] for a project. ListServiceAccounts(context.Context, *ListServiceAccountsRequest) (*ListServiceAccountsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/iam/v1/iam_policy.pb.go b/vendor/google.golang.org/genproto/googleapis/iam/v1/iam_policy.pb.go index 0450c83f7..7334ac2de 100644 --- a/vendor/google.golang.org/genproto/googleapis/iam/v1/iam_policy.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/iam/v1/iam_policy.pb.go @@ -29,12 +29,12 @@ type SetIamPolicyRequest struct { // REQUIRED: The resource for which the policy is being specified. // `resource` is usually specified as a path. For example, a Project // resource is specified as `projects/{project}`. - Resource string `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource string `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` // REQUIRED: The complete policy to be applied to the `resource`. The size of // the policy is limited to a few 10s of KB. An empty policy is a // valid policy but certain Cloud Platform services (such as Projects) // might reject them. - Policy *Policy `protobuf:"bytes,2,opt,name=policy" json:"policy,omitempty"` + Policy *Policy `protobuf:"bytes,2,opt,name=policy,proto3" json:"policy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -83,7 +83,7 @@ type GetIamPolicyRequest struct { // REQUIRED: The resource for which the policy is being requested. // `resource` is usually specified as a path. For example, a Project // resource is specified as `projects/{project}`. - Resource string `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource string `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -125,12 +125,12 @@ type TestIamPermissionsRequest struct { // REQUIRED: The resource for which the policy detail is being requested. // `resource` is usually specified as a path. For example, a Project // resource is specified as `projects/{project}`. - Resource string `protobuf:"bytes,1,opt,name=resource" json:"resource,omitempty"` + Resource string `protobuf:"bytes,1,opt,name=resource,proto3" json:"resource,omitempty"` // The set of permissions to check for the `resource`. Permissions with // wildcards (such as '*' or 'storage.*') are not allowed. For more // information see // [IAM Overview](https://cloud.google.com/iam/docs/overview#permissions). - Permissions []string `protobuf:"bytes,2,rep,name=permissions" json:"permissions,omitempty"` + Permissions []string `protobuf:"bytes,2,rep,name=permissions,proto3" json:"permissions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -178,7 +178,7 @@ func (m *TestIamPermissionsRequest) GetPermissions() []string { type TestIamPermissionsResponse struct { // A subset of `TestPermissionsRequest.permissions` that the caller is // allowed. - Permissions []string `protobuf:"bytes,1,rep,name=permissions" json:"permissions,omitempty"` + Permissions []string `protobuf:"bytes,1,rep,name=permissions,proto3" json:"permissions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -282,8 +282,7 @@ func (c *iAMPolicyClient) TestIamPermissions(ctx context.Context, in *TestIamPer return out, nil } -// Server API for IAMPolicy service - +// IAMPolicyServer is the server API for IAMPolicy service. type IAMPolicyServer interface { // Sets the access control policy on the specified resource. Replaces any // existing policy. diff --git a/vendor/google.golang.org/genproto/googleapis/iam/v1/logging/audit_data.pb.go b/vendor/google.golang.org/genproto/googleapis/iam/v1/logging/audit_data.pb.go index 9fc65c3a1..63b6a775e 100644 --- a/vendor/google.golang.org/genproto/googleapis/iam/v1/logging/audit_data.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/iam/v1/logging/audit_data.pb.go @@ -25,7 +25,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // `AuditLog` message. type AuditData struct { // Policy delta between the original policy and the newly set policy. - PolicyDelta *v1.PolicyDelta `protobuf:"bytes,2,opt,name=policy_delta,json=policyDelta" json:"policy_delta,omitempty"` + PolicyDelta *v1.PolicyDelta `protobuf:"bytes,2,opt,name=policy_delta,json=policyDelta,proto3" json:"policy_delta,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/iam/v1/policy.pb.go b/vendor/google.golang.org/genproto/googleapis/iam/v1/policy.pb.go index dd132dfc8..a4805ece2 100644 --- a/vendor/google.golang.org/genproto/googleapis/iam/v1/policy.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/iam/v1/policy.pb.go @@ -82,11 +82,11 @@ func (BindingDelta_Action) EnumDescriptor() ([]byte, []int) { // [IAM developer's guide](https://cloud.google.com/iam). type Policy struct { // Version of the `Policy`. The default version is 0. - Version int32 `protobuf:"varint,1,opt,name=version" json:"version,omitempty"` + Version int32 `protobuf:"varint,1,opt,name=version,proto3" json:"version,omitempty"` // Associates a list of `members` to a `role`. // Multiple `bindings` must not be specified for the same `role`. // `bindings` with no members will result in an error. - Bindings []*Binding `protobuf:"bytes,4,rep,name=bindings" json:"bindings,omitempty"` + Bindings []*Binding `protobuf:"bytes,4,rep,name=bindings,proto3" json:"bindings,omitempty"` // `etag` is used for optimistic concurrency control as a way to help // prevent simultaneous updates of a policy from overwriting each other. // It is strongly suggested that systems make use of the `etag` in the @@ -153,7 +153,7 @@ type Binding struct { // Role that is assigned to `members`. // For example, `roles/viewer`, `roles/editor`, or `roles/owner`. // Required - Role string `protobuf:"bytes,1,opt,name=role" json:"role,omitempty"` + Role string `protobuf:"bytes,1,opt,name=role,proto3" json:"role,omitempty"` // Specifies the identities requesting access for a Cloud Platform resource. // `members` can have the following values: // @@ -177,7 +177,7 @@ type Binding struct { // users of that domain. For example, `google.com` or `example.com`. // // - Members []string `protobuf:"bytes,2,rep,name=members" json:"members,omitempty"` + Members []string `protobuf:"bytes,2,rep,name=members,proto3" json:"members,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -224,7 +224,7 @@ func (m *Binding) GetMembers() []string { // The difference delta between two policies. type PolicyDelta struct { // The delta for Bindings between two policies. - BindingDeltas []*BindingDelta `protobuf:"bytes,1,rep,name=binding_deltas,json=bindingDeltas" json:"binding_deltas,omitempty"` + BindingDeltas []*BindingDelta `protobuf:"bytes,1,rep,name=binding_deltas,json=bindingDeltas,proto3" json:"binding_deltas,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -266,15 +266,15 @@ func (m *PolicyDelta) GetBindingDeltas() []*BindingDelta { type BindingDelta struct { // The action that was performed on a Binding. // Required - Action BindingDelta_Action `protobuf:"varint,1,opt,name=action,enum=google.iam.v1.BindingDelta_Action" json:"action,omitempty"` + Action BindingDelta_Action `protobuf:"varint,1,opt,name=action,proto3,enum=google.iam.v1.BindingDelta_Action" json:"action,omitempty"` // Role that is assigned to `members`. // For example, `roles/viewer`, `roles/editor`, or `roles/owner`. // Required - Role string `protobuf:"bytes,2,opt,name=role" json:"role,omitempty"` + Role string `protobuf:"bytes,2,opt,name=role,proto3" json:"role,omitempty"` // A single identity requesting access for a Cloud Platform resource. // Follows the same format of Binding.members. // Required - Member string `protobuf:"bytes,3,opt,name=member" json:"member,omitempty"` + Member string `protobuf:"bytes,3,opt,name=member,proto3" json:"member,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/logging/type/http_request.pb.go b/vendor/google.golang.org/genproto/googleapis/logging/type/http_request.pb.go index fb1d2cd47..43f27877e 100644 --- a/vendor/google.golang.org/genproto/googleapis/logging/type/http_request.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/logging/type/http_request.pb.go @@ -25,49 +25,49 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // information MUST be defined in a separate message. type HttpRequest struct { // The request method. Examples: `"GET"`, `"HEAD"`, `"PUT"`, `"POST"`. - RequestMethod string `protobuf:"bytes,1,opt,name=request_method,json=requestMethod" json:"request_method,omitempty"` + RequestMethod string `protobuf:"bytes,1,opt,name=request_method,json=requestMethod,proto3" json:"request_method,omitempty"` // The scheme (http, https), the host name, the path and the query // portion of the URL that was requested. // Example: `"http://example.com/some/info?color=red"`. - RequestUrl string `protobuf:"bytes,2,opt,name=request_url,json=requestUrl" json:"request_url,omitempty"` + RequestUrl string `protobuf:"bytes,2,opt,name=request_url,json=requestUrl,proto3" json:"request_url,omitempty"` // The size of the HTTP request message in bytes, including the request // headers and the request body. - RequestSize int64 `protobuf:"varint,3,opt,name=request_size,json=requestSize" json:"request_size,omitempty"` + RequestSize int64 `protobuf:"varint,3,opt,name=request_size,json=requestSize,proto3" json:"request_size,omitempty"` // The response code indicating the status of response. // Examples: 200, 404. - Status int32 `protobuf:"varint,4,opt,name=status" json:"status,omitempty"` + Status int32 `protobuf:"varint,4,opt,name=status,proto3" json:"status,omitempty"` // The size of the HTTP response message sent back to the client, in bytes, // including the response headers and the response body. - ResponseSize int64 `protobuf:"varint,5,opt,name=response_size,json=responseSize" json:"response_size,omitempty"` + ResponseSize int64 `protobuf:"varint,5,opt,name=response_size,json=responseSize,proto3" json:"response_size,omitempty"` // The user agent sent by the client. Example: // `"Mozilla/4.0 (compatible; MSIE 6.0; Windows 98; Q312461; .NET CLR 1.0.3705)"`. - UserAgent string `protobuf:"bytes,6,opt,name=user_agent,json=userAgent" json:"user_agent,omitempty"` + UserAgent string `protobuf:"bytes,6,opt,name=user_agent,json=userAgent,proto3" json:"user_agent,omitempty"` // The IP address (IPv4 or IPv6) of the client that issued the HTTP // request. Examples: `"192.168.1.1"`, `"FE80::0202:B3FF:FE1E:8329"`. - RemoteIp string `protobuf:"bytes,7,opt,name=remote_ip,json=remoteIp" json:"remote_ip,omitempty"` + RemoteIp string `protobuf:"bytes,7,opt,name=remote_ip,json=remoteIp,proto3" json:"remote_ip,omitempty"` // The IP address (IPv4 or IPv6) of the origin server that the request was // sent to. - ServerIp string `protobuf:"bytes,13,opt,name=server_ip,json=serverIp" json:"server_ip,omitempty"` + ServerIp string `protobuf:"bytes,13,opt,name=server_ip,json=serverIp,proto3" json:"server_ip,omitempty"` // The referer URL of the request, as defined in // [HTTP/1.1 Header Field Definitions](http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html). - Referer string `protobuf:"bytes,8,opt,name=referer" json:"referer,omitempty"` + Referer string `protobuf:"bytes,8,opt,name=referer,proto3" json:"referer,omitempty"` // The request processing latency on the server, from the time the request was // received until the response was sent. - Latency *duration.Duration `protobuf:"bytes,14,opt,name=latency" json:"latency,omitempty"` + Latency *duration.Duration `protobuf:"bytes,14,opt,name=latency,proto3" json:"latency,omitempty"` // Whether or not a cache lookup was attempted. - CacheLookup bool `protobuf:"varint,11,opt,name=cache_lookup,json=cacheLookup" json:"cache_lookup,omitempty"` + CacheLookup bool `protobuf:"varint,11,opt,name=cache_lookup,json=cacheLookup,proto3" json:"cache_lookup,omitempty"` // Whether or not an entity was served from cache // (with or without validation). - CacheHit bool `protobuf:"varint,9,opt,name=cache_hit,json=cacheHit" json:"cache_hit,omitempty"` + CacheHit bool `protobuf:"varint,9,opt,name=cache_hit,json=cacheHit,proto3" json:"cache_hit,omitempty"` // Whether or not the response was validated with the origin server before // being served from cache. This field is only meaningful if `cache_hit` is // True. - CacheValidatedWithOriginServer bool `protobuf:"varint,10,opt,name=cache_validated_with_origin_server,json=cacheValidatedWithOriginServer" json:"cache_validated_with_origin_server,omitempty"` + CacheValidatedWithOriginServer bool `protobuf:"varint,10,opt,name=cache_validated_with_origin_server,json=cacheValidatedWithOriginServer,proto3" json:"cache_validated_with_origin_server,omitempty"` // The number of HTTP response bytes inserted into cache. Set only when a // cache fill was attempted. - CacheFillBytes int64 `protobuf:"varint,12,opt,name=cache_fill_bytes,json=cacheFillBytes" json:"cache_fill_bytes,omitempty"` + CacheFillBytes int64 `protobuf:"varint,12,opt,name=cache_fill_bytes,json=cacheFillBytes,proto3" json:"cache_fill_bytes,omitempty"` // Protocol used for the request. Examples: "HTTP/1.1", "HTTP/2", "websocket" - Protocol string `protobuf:"bytes,15,opt,name=protocol" json:"protocol,omitempty"` + Protocol string `protobuf:"bytes,15,opt,name=protocol,proto3" json:"protocol,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -77,7 +77,7 @@ func (m *HttpRequest) Reset() { *m = HttpRequest{} } func (m *HttpRequest) String() string { return proto.CompactTextString(m) } func (*HttpRequest) ProtoMessage() {} func (*HttpRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_http_request_512b09169cdf78ff, []int{0} + return fileDescriptor_http_request_e718ea9739c965d6, []int{0} } func (m *HttpRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_HttpRequest.Unmarshal(m, b) @@ -207,10 +207,10 @@ func init() { } func init() { - proto.RegisterFile("google/logging/type/http_request.proto", fileDescriptor_http_request_512b09169cdf78ff) + proto.RegisterFile("google/logging/type/http_request.proto", fileDescriptor_http_request_e718ea9739c965d6) } -var fileDescriptor_http_request_512b09169cdf78ff = []byte{ +var fileDescriptor_http_request_e718ea9739c965d6 = []byte{ // 511 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x6c, 0x92, 0x5b, 0x6b, 0x14, 0x31, 0x18, 0x86, 0x99, 0x1e, 0xf6, 0x90, 0x3d, 0x58, 0x22, 0x68, 0xba, 0x6a, 0x5d, 0x2b, 0xca, 0x5c, diff --git a/vendor/google.golang.org/genproto/googleapis/logging/type/log_severity.pb.go b/vendor/google.golang.org/genproto/googleapis/logging/type/log_severity.pb.go index 374f50281..aea2530e9 100644 --- a/vendor/google.golang.org/genproto/googleapis/logging/type/log_severity.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/logging/type/log_severity.pb.go @@ -85,7 +85,7 @@ func (x LogSeverity) String() string { return proto.EnumName(LogSeverity_name, int32(x)) } func (LogSeverity) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_log_severity_6ec6ee905a000c23, []int{0} + return fileDescriptor_log_severity_bd22c162c6e9611a, []int{0} } func init() { @@ -93,10 +93,10 @@ func init() { } func init() { - proto.RegisterFile("google/logging/type/log_severity.proto", fileDescriptor_log_severity_6ec6ee905a000c23) + proto.RegisterFile("google/logging/type/log_severity.proto", fileDescriptor_log_severity_bd22c162c6e9611a) } -var fileDescriptor_log_severity_6ec6ee905a000c23 = []byte{ +var fileDescriptor_log_severity_bd22c162c6e9611a = []byte{ // 309 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x52, 0x4b, 0xcf, 0xcf, 0x4f, 0xcf, 0x49, 0xd5, 0xcf, 0xc9, 0x4f, 0x4f, 0xcf, 0xcc, 0x4b, 0xd7, 0x2f, 0xa9, 0x2c, 0x00, 0x73, diff --git a/vendor/google.golang.org/genproto/googleapis/logging/v2/log_entry.pb.go b/vendor/google.golang.org/genproto/googleapis/logging/v2/log_entry.pb.go index bdda1ec5d..fc989d469 100644 --- a/vendor/google.golang.org/genproto/googleapis/logging/v2/log_entry.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/logging/v2/log_entry.pb.go @@ -48,12 +48,12 @@ type LogEntry struct { // forward-slash is removed. Listing the log entry will not show the leading // slash and filtering for a log name with a leading slash will never return // any results. - LogName string `protobuf:"bytes,12,opt,name=log_name,json=logName" json:"log_name,omitempty"` - // Required. The monitored resource associated with this log entry. + LogName string `protobuf:"bytes,12,opt,name=log_name,json=logName,proto3" json:"log_name,omitempty"` + // Required. The primary monitored resource associated with this log entry. // Example: a log entry that reports a database error would be // associated with the monitored resource designating the particular // database that reported the error. - Resource *monitoredres.MonitoredResource `protobuf:"bytes,8,opt,name=resource" json:"resource,omitempty"` + Resource *monitoredres.MonitoredResource `protobuf:"bytes,8,opt,name=resource,proto3" json:"resource,omitempty"` // Optional. The log entry payload, which can be one of multiple types. // // Types that are valid to be assigned to Payload: @@ -65,47 +65,54 @@ type LogEntry struct { // This time is used to compute the log entry's age and to enforce // the logs retention period. If this field is omitted in a new log // entry, then Stackdriver Logging assigns it the current time. + // Timestamps have nanosecond accuracy, but trailing zeros in the fractional + // seconds might be omitted when the timestamp is displayed. // // Incoming log entries should have timestamps that are no more than - // the [logs retention period](/logging/quota-policy) in the past, - // and no more than 24 hours in the future. - // See the `entries.write` API method for more information. - Timestamp *timestamp.Timestamp `protobuf:"bytes,9,opt,name=timestamp" json:"timestamp,omitempty"` + // the [logs retention period](/logging/quotas) in the past, + // and no more than 24 hours in the future. Log entries outside those time + // boundaries will not be available when calling `entries.list`, but + // those log entries can still be exported with + // [LogSinks](/logging/docs/api/tasks/exporting-logs). + Timestamp *timestamp.Timestamp `protobuf:"bytes,9,opt,name=timestamp,proto3" json:"timestamp,omitempty"` // Output only. The time the log entry was received by Stackdriver Logging. - ReceiveTimestamp *timestamp.Timestamp `protobuf:"bytes,24,opt,name=receive_timestamp,json=receiveTimestamp" json:"receive_timestamp,omitempty"` + ReceiveTimestamp *timestamp.Timestamp `protobuf:"bytes,24,opt,name=receive_timestamp,json=receiveTimestamp,proto3" json:"receive_timestamp,omitempty"` // Optional. The severity of the log entry. The default value is // `LogSeverity.DEFAULT`. - Severity _type.LogSeverity `protobuf:"varint,10,opt,name=severity,enum=google.logging.type.LogSeverity" json:"severity,omitempty"` + Severity _type.LogSeverity `protobuf:"varint,10,opt,name=severity,proto3,enum=google.logging.type.LogSeverity" json:"severity,omitempty"` // Optional. A unique identifier for the log entry. If you provide a value, // then Stackdriver Logging considers other log entries in the same project, // with the same `timestamp`, and with the same `insert_id` to be duplicates // which can be removed. If omitted in new log entries, then Stackdriver // Logging assigns its own unique identifier. The `insert_id` is also used // to order log entries that have the same `timestamp` value. - InsertId string `protobuf:"bytes,4,opt,name=insert_id,json=insertId" json:"insert_id,omitempty"` + InsertId string `protobuf:"bytes,4,opt,name=insert_id,json=insertId,proto3" json:"insert_id,omitempty"` // Optional. Information about the HTTP request associated with this // log entry, if applicable. - HttpRequest *_type.HttpRequest `protobuf:"bytes,7,opt,name=http_request,json=httpRequest" json:"http_request,omitempty"` + HttpRequest *_type.HttpRequest `protobuf:"bytes,7,opt,name=http_request,json=httpRequest,proto3" json:"http_request,omitempty"` // Optional. A set of user-defined (key, value) data that provides additional // information about the log entry. - Labels map[string]string `protobuf:"bytes,11,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,11,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + // Output only. Additional metadata about the monitored resource. + // Only `k8s_container`, `k8s_pod`, and `k8s_node` MonitoredResources have + // this field populated. + Metadata *monitoredres.MonitoredResourceMetadata `protobuf:"bytes,25,opt,name=metadata,proto3" json:"metadata,omitempty"` // Optional. Information about an operation associated with the log entry, if // applicable. - Operation *LogEntryOperation `protobuf:"bytes,15,opt,name=operation" json:"operation,omitempty"` + Operation *LogEntryOperation `protobuf:"bytes,15,opt,name=operation,proto3" json:"operation,omitempty"` // Optional. Resource name of the trace associated with the log entry, if any. // If it contains a relative resource name, the name is assumed to be relative // to `//tracing.googleapis.com`. Example: // `projects/my-projectid/traces/06796866738c859f2f19b7cfb3214824` - Trace string `protobuf:"bytes,22,opt,name=trace" json:"trace,omitempty"` - // Optional. Id of the span within the trace associated with the log entry. - // e.g. "0000000000000042" - // For Stackdriver trace spans, this is the same format that the Stackdriver - // trace API uses. - // The ID is a 16-character hexadecimal encoding of an 8-byte array. - SpanId string `protobuf:"bytes,27,opt,name=span_id,json=spanId" json:"span_id,omitempty"` + Trace string `protobuf:"bytes,22,opt,name=trace,proto3" json:"trace,omitempty"` + // Optional. The span ID within the trace associated with the log entry. For + // Stackdriver Trace spans, this is the same format that the Stackdriver Trace + // API v2 uses: a 16-character hexadecimal encoding of an 8-byte array, such + // as "000000000000004a". + SpanId string `protobuf:"bytes,27,opt,name=span_id,json=spanId,proto3" json:"span_id,omitempty"` // Optional. Source code location information associated with the log entry, // if any. - SourceLocation *LogEntrySourceLocation `protobuf:"bytes,23,opt,name=source_location,json=sourceLocation" json:"source_location,omitempty"` + SourceLocation *LogEntrySourceLocation `protobuf:"bytes,23,opt,name=source_location,json=sourceLocation,proto3" json:"source_location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -115,7 +122,7 @@ func (m *LogEntry) Reset() { *m = LogEntry{} } func (m *LogEntry) String() string { return proto.CompactTextString(m) } func (*LogEntry) ProtoMessage() {} func (*LogEntry) Descriptor() ([]byte, []int) { - return fileDescriptor_log_entry_186c4d608b380092, []int{0} + return fileDescriptor_log_entry_63c24caaf3e31aeb, []int{0} } func (m *LogEntry) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LogEntry.Unmarshal(m, b) @@ -140,13 +147,13 @@ type isLogEntry_Payload interface { } type LogEntry_ProtoPayload struct { - ProtoPayload *any.Any `protobuf:"bytes,2,opt,name=proto_payload,json=protoPayload,oneof"` + ProtoPayload *any.Any `protobuf:"bytes,2,opt,name=proto_payload,json=protoPayload,proto3,oneof"` } type LogEntry_TextPayload struct { - TextPayload string `protobuf:"bytes,3,opt,name=text_payload,json=textPayload,oneof"` + TextPayload string `protobuf:"bytes,3,opt,name=text_payload,json=textPayload,proto3,oneof"` } type LogEntry_JsonPayload struct { - JsonPayload *_struct.Struct `protobuf:"bytes,6,opt,name=json_payload,json=jsonPayload,oneof"` + JsonPayload *_struct.Struct `protobuf:"bytes,6,opt,name=json_payload,json=jsonPayload,proto3,oneof"` } func (*LogEntry_ProtoPayload) isLogEntry_Payload() {} @@ -237,6 +244,13 @@ func (m *LogEntry) GetLabels() map[string]string { return nil } +func (m *LogEntry) GetMetadata() *monitoredres.MonitoredResourceMetadata { + if m != nil { + return m.Metadata + } + return nil +} + func (m *LogEntry) GetOperation() *LogEntryOperation { if m != nil { return m.Operation @@ -359,15 +373,15 @@ func _LogEntry_OneofSizer(msg proto.Message) (n int) { type LogEntryOperation struct { // Optional. An arbitrary operation identifier. Log entries with the // same identifier are assumed to be part of the same operation. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` // Optional. An arbitrary producer identifier. The combination of // `id` and `producer` must be globally unique. Examples for `producer`: // `"MyDivision.MyBigCompany.com"`, `"github.com/MyProject/MyApplication"`. - Producer string `protobuf:"bytes,2,opt,name=producer" json:"producer,omitempty"` + Producer string `protobuf:"bytes,2,opt,name=producer,proto3" json:"producer,omitempty"` // Optional. Set this to True if this is the first log entry in the operation. - First bool `protobuf:"varint,3,opt,name=first" json:"first,omitempty"` + First bool `protobuf:"varint,3,opt,name=first,proto3" json:"first,omitempty"` // Optional. Set this to True if this is the last log entry in the operation. - Last bool `protobuf:"varint,4,opt,name=last" json:"last,omitempty"` + Last bool `protobuf:"varint,4,opt,name=last,proto3" json:"last,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -377,7 +391,7 @@ func (m *LogEntryOperation) Reset() { *m = LogEntryOperation{} } func (m *LogEntryOperation) String() string { return proto.CompactTextString(m) } func (*LogEntryOperation) ProtoMessage() {} func (*LogEntryOperation) Descriptor() ([]byte, []int) { - return fileDescriptor_log_entry_186c4d608b380092, []int{1} + return fileDescriptor_log_entry_63c24caaf3e31aeb, []int{1} } func (m *LogEntryOperation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LogEntryOperation.Unmarshal(m, b) @@ -430,17 +444,17 @@ func (m *LogEntryOperation) GetLast() bool { type LogEntrySourceLocation struct { // Optional. Source file name. Depending on the runtime environment, this // might be a simple name or a fully-qualified name. - File string `protobuf:"bytes,1,opt,name=file" json:"file,omitempty"` + File string `protobuf:"bytes,1,opt,name=file,proto3" json:"file,omitempty"` // Optional. Line within the source file. 1-based; 0 indicates no line number // available. - Line int64 `protobuf:"varint,2,opt,name=line" json:"line,omitempty"` + Line int64 `protobuf:"varint,2,opt,name=line,proto3" json:"line,omitempty"` // Optional. Human-readable name of the function or method being invoked, with // optional context such as the class or package name. This information may be // used in contexts such as the logs viewer, where a file and line number are // less meaningful. The format can vary by language. For example: // `qual.if.ied.Class.method` (Java), `dir/package.func` (Go), `function` // (Python). - Function string `protobuf:"bytes,3,opt,name=function" json:"function,omitempty"` + Function string `protobuf:"bytes,3,opt,name=function,proto3" json:"function,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -450,7 +464,7 @@ func (m *LogEntrySourceLocation) Reset() { *m = LogEntrySourceLocation{} func (m *LogEntrySourceLocation) String() string { return proto.CompactTextString(m) } func (*LogEntrySourceLocation) ProtoMessage() {} func (*LogEntrySourceLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_log_entry_186c4d608b380092, []int{2} + return fileDescriptor_log_entry_63c24caaf3e31aeb, []int{2} } func (m *LogEntrySourceLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LogEntrySourceLocation.Unmarshal(m, b) @@ -499,55 +513,56 @@ func init() { } func init() { - proto.RegisterFile("google/logging/v2/log_entry.proto", fileDescriptor_log_entry_186c4d608b380092) + proto.RegisterFile("google/logging/v2/log_entry.proto", fileDescriptor_log_entry_63c24caaf3e31aeb) } -var fileDescriptor_log_entry_186c4d608b380092 = []byte{ - // 729 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x54, 0xcd, 0x6e, 0xdb, 0x46, - 0x10, 0x36, 0x25, 0x57, 0xa2, 0x56, 0xf2, 0xdf, 0xc2, 0xb5, 0x68, 0xd9, 0x45, 0x55, 0xbb, 0x68, - 0xd5, 0x0b, 0x05, 0xa8, 0x17, 0xbb, 0x36, 0x50, 0x54, 0x46, 0x61, 0x1b, 0x50, 0x5b, 0x63, 0x5d, - 0xf8, 0x10, 0x08, 0x10, 0xd6, 0xe4, 0x8a, 0xde, 0x84, 0xda, 0x65, 0x96, 0x4b, 0x21, 0x7a, 0x94, - 0xbc, 0x42, 0x1e, 0x25, 0x6f, 0x93, 0x5b, 0x8e, 0xc1, 0x0e, 0x97, 0x92, 0x22, 0x19, 0xce, 0x6d, - 0x66, 0xe7, 0xfb, 0xe6, 0x9b, 0x19, 0xce, 0x10, 0xfd, 0x14, 0x49, 0x19, 0xc5, 0xac, 0x1b, 0xcb, - 0x28, 0xe2, 0x22, 0xea, 0x4e, 0x7b, 0xc6, 0x1c, 0x31, 0xa1, 0xd5, 0xcc, 0x4f, 0x94, 0xd4, 0x12, - 0xef, 0xe5, 0x10, 0xdf, 0x42, 0xfc, 0x69, 0xaf, 0x75, 0x6c, 0x59, 0x34, 0xe1, 0x5d, 0x2a, 0x84, - 0xd4, 0x54, 0x73, 0x29, 0xd2, 0x9c, 0xd0, 0x3a, 0x5d, 0x8a, 0x4e, 0xa4, 0xe0, 0x5a, 0x2a, 0x16, - 0x8e, 0x14, 0x4b, 0x65, 0xa6, 0x02, 0x66, 0x41, 0xbf, 0xac, 0x08, 0xeb, 0x59, 0xc2, 0xba, 0x4f, - 0x5a, 0x27, 0x23, 0xc5, 0xde, 0x66, 0x2c, 0xd5, 0x2f, 0xe1, 0x4c, 0x89, 0x29, 0x9b, 0x32, 0xc5, - 0xb5, 0xad, 0xb2, 0x75, 0x68, 0x71, 0xe0, 0x3d, 0x66, 0xe3, 0x2e, 0x15, 0x45, 0xe8, 0x78, 0x35, - 0x94, 0x6a, 0x95, 0x05, 0x85, 0xc0, 0x8f, 0xab, 0x51, 0xcd, 0x27, 0x2c, 0xd5, 0x74, 0x92, 0xe4, - 0x80, 0x93, 0x4f, 0x15, 0xe4, 0x0e, 0x64, 0xf4, 0xb7, 0x19, 0x09, 0x3e, 0x44, 0xae, 0x11, 0x17, - 0x74, 0xc2, 0xbc, 0x46, 0xdb, 0xe9, 0xd4, 0x48, 0x35, 0x96, 0xd1, 0xbf, 0x74, 0xc2, 0xf0, 0x39, - 0x72, 0x8b, 0x1e, 0x3d, 0xb7, 0xed, 0x74, 0xea, 0xbd, 0x1f, 0x7c, 0x3b, 0x3a, 0x9a, 0x70, 0xff, - 0x9f, 0x62, 0x12, 0xc4, 0x82, 0xc8, 0x1c, 0x8e, 0x2f, 0xd0, 0x16, 0x68, 0x8d, 0x12, 0x3a, 0x8b, - 0x25, 0x0d, 0xbd, 0x12, 0xf0, 0xf7, 0x0b, 0x7e, 0x51, 0x9b, 0xff, 0x97, 0x98, 0xdd, 0x6c, 0x90, - 0x06, 0xf8, 0x77, 0x39, 0x16, 0x9f, 0xa2, 0x86, 0x66, 0xef, 0xf4, 0x9c, 0x5b, 0x36, 0x65, 0xdd, - 0x6c, 0x90, 0xba, 0x79, 0x2d, 0x40, 0x97, 0xa8, 0xf1, 0x3a, 0x95, 0x62, 0x0e, 0xaa, 0x80, 0x40, - 0x73, 0x4d, 0xe0, 0x1e, 0x46, 0x63, 0xd8, 0x06, 0x5e, 0xb0, 0xcf, 0x50, 0x6d, 0x3e, 0x15, 0xaf, - 0x06, 0xd4, 0xd6, 0x1a, 0xf5, 0xff, 0x02, 0x41, 0x16, 0x60, 0x7c, 0x8d, 0xf6, 0x14, 0x0b, 0x18, - 0x9f, 0xb2, 0xd1, 0x22, 0x83, 0xf7, 0xcd, 0x0c, 0xbb, 0x96, 0x34, 0x7f, 0xc1, 0x97, 0xc8, 0x2d, - 0xbe, 0xb8, 0x87, 0xda, 0x4e, 0x67, 0xbb, 0xd7, 0xf6, 0x57, 0x16, 0xd3, 0xac, 0x86, 0x3f, 0x90, - 0xd1, 0xbd, 0xc5, 0x91, 0x39, 0x03, 0x1f, 0xa1, 0x1a, 0x17, 0x29, 0x53, 0x7a, 0xc4, 0x43, 0x6f, - 0x13, 0xbe, 0x9b, 0x9b, 0x3f, 0xdc, 0x86, 0xf8, 0x0a, 0x35, 0x96, 0x17, 0xcf, 0xab, 0x42, 0x79, - 0xcf, 0xa7, 0xbf, 0xd1, 0x3a, 0x21, 0x39, 0x8e, 0xd4, 0x9f, 0x16, 0x0e, 0xfe, 0x13, 0x55, 0x62, - 0xfa, 0xc8, 0xe2, 0xd4, 0xab, 0xb7, 0xcb, 0x9d, 0x7a, 0xef, 0x57, 0x7f, 0xed, 0x6c, 0xfc, 0x62, - 0x8b, 0xfc, 0x01, 0x20, 0xc1, 0x26, 0x96, 0x86, 0xfb, 0xa8, 0x26, 0x13, 0xa6, 0xe0, 0x92, 0xbc, - 0x1d, 0x28, 0xe1, 0xe7, 0x17, 0x72, 0xfc, 0x57, 0x60, 0xc9, 0x82, 0x86, 0xf7, 0xd1, 0x77, 0x5a, - 0xd1, 0x80, 0x79, 0x07, 0xd0, 0x62, 0xee, 0xe0, 0x26, 0xaa, 0xa6, 0x09, 0x15, 0xa6, 0xf5, 0x23, - 0x78, 0xaf, 0x18, 0xf7, 0x36, 0xc4, 0x04, 0xed, 0xe4, 0x0b, 0x38, 0x8a, 0x65, 0x90, 0x0b, 0x37, - 0x41, 0xf8, 0xb7, 0x17, 0x84, 0xef, 0x81, 0x31, 0xb0, 0x04, 0xb2, 0x9d, 0x7e, 0xe5, 0xb7, 0xce, - 0x51, 0x7d, 0xa9, 0x3b, 0xbc, 0x8b, 0xca, 0x6f, 0xd8, 0xcc, 0x73, 0x40, 0xd7, 0x98, 0xa6, 0xc6, - 0x29, 0x8d, 0x33, 0x06, 0x3b, 0x5e, 0x23, 0xb9, 0xf3, 0x47, 0xe9, 0xcc, 0xe9, 0xd7, 0x50, 0xd5, - 0xae, 0xe7, 0x09, 0x47, 0x7b, 0x6b, 0x8d, 0xe2, 0x6d, 0x54, 0xe2, 0xa1, 0x4d, 0x55, 0xe2, 0x21, - 0x6e, 0x21, 0x37, 0x51, 0x32, 0xcc, 0x02, 0xa6, 0x6c, 0xb2, 0xb9, 0x6f, 0x54, 0xc6, 0x5c, 0xa5, - 0x1a, 0xae, 0xc1, 0x25, 0xb9, 0x83, 0x31, 0xda, 0x8c, 0x69, 0xaa, 0x61, 0x03, 0x5c, 0x02, 0xf6, - 0xc9, 0x10, 0x1d, 0x3c, 0xdf, 0x9a, 0x41, 0x8f, 0x79, 0xcc, 0xac, 0x22, 0xd8, 0x90, 0x81, 0x8b, - 0xbc, 0xf8, 0x32, 0x01, 0xdb, 0xd4, 0x31, 0xce, 0x44, 0x00, 0xf3, 0x2b, 0xe7, 0x75, 0x14, 0x7e, - 0xff, 0xbd, 0x83, 0xbe, 0x0f, 0xe4, 0x64, 0x7d, 0x9e, 0xfd, 0xad, 0x42, 0xf5, 0x0e, 0x8e, 0xd9, - 0x79, 0x75, 0x66, 0x31, 0x91, 0x8c, 0xa9, 0x88, 0x7c, 0xa9, 0xa2, 0x6e, 0xc4, 0x04, 0x1c, 0x47, - 0x37, 0x0f, 0xd1, 0x84, 0xa7, 0x4b, 0x7f, 0xea, 0x0b, 0x6b, 0x7e, 0x76, 0x9c, 0x0f, 0xa5, 0xe6, - 0x75, 0xce, 0xbe, 0x8a, 0x65, 0x16, 0x9a, 0x8f, 0x05, 0x3a, 0x0f, 0xbd, 0x8f, 0x45, 0x64, 0x08, - 0x91, 0xa1, 0x8d, 0x0c, 0x1f, 0x7a, 0x8f, 0x15, 0xc8, 0xfd, 0xfb, 0x97, 0x00, 0x00, 0x00, 0xff, - 0xff, 0x80, 0x53, 0xd3, 0xff, 0x04, 0x06, 0x00, 0x00, +var fileDescriptor_log_entry_63c24caaf3e31aeb = []byte{ + // 748 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x54, 0x5d, 0x6f, 0xf3, 0x34, + 0x14, 0x7e, 0xd3, 0xbe, 0xb4, 0xa9, 0xdb, 0x77, 0x1f, 0xd6, 0x58, 0xb3, 0x6e, 0x88, 0xb2, 0xf1, + 0x51, 0x6e, 0x52, 0xa9, 0xdc, 0x6c, 0x6c, 0x12, 0x5a, 0x27, 0xb4, 0x4d, 0xea, 0x60, 0xf2, 0xd0, + 0x2e, 0x50, 0xa5, 0xca, 0x4b, 0xdc, 0xcc, 0x90, 0xda, 0xc1, 0x71, 0x2a, 0xfa, 0x53, 0xf8, 0x0b, + 0x5c, 0xf3, 0x2b, 0xf8, 0x45, 0x5c, 0x22, 0x9f, 0x38, 0x69, 0x69, 0xa7, 0x72, 0x77, 0x8e, 0xcf, + 0xf3, 0x9c, 0xe7, 0xc9, 0xf1, 0x71, 0xd0, 0x67, 0x91, 0x94, 0x51, 0xcc, 0xfa, 0xb1, 0x8c, 0x22, + 0x2e, 0xa2, 0xfe, 0x7c, 0x60, 0xc2, 0x09, 0x13, 0x5a, 0x2d, 0xfc, 0x44, 0x49, 0x2d, 0xf1, 0x7e, + 0x0e, 0xf1, 0x2d, 0xc4, 0x9f, 0x0f, 0x3a, 0x27, 0x96, 0x45, 0x13, 0xde, 0xa7, 0x42, 0x48, 0x4d, + 0x35, 0x97, 0x22, 0xcd, 0x09, 0x9d, 0xb3, 0x95, 0xea, 0x4c, 0x0a, 0xae, 0xa5, 0x62, 0xe1, 0x44, + 0xb1, 0x54, 0x66, 0x2a, 0x60, 0x16, 0xf4, 0xe5, 0x9a, 0xb0, 0x5e, 0x24, 0xac, 0xff, 0xaa, 0x75, + 0x32, 0x51, 0xec, 0xb7, 0x8c, 0xa5, 0x7a, 0x1b, 0xce, 0x58, 0x4c, 0xd9, 0x9c, 0x29, 0xae, 0xad, + 0xcb, 0xce, 0x91, 0xc5, 0x41, 0xf6, 0x92, 0x4d, 0xfb, 0x54, 0x14, 0xa5, 0x93, 0xf5, 0x52, 0xaa, + 0x55, 0x16, 0x14, 0x02, 0x9f, 0xae, 0x57, 0x35, 0x9f, 0xb1, 0x54, 0xd3, 0x59, 0x92, 0x03, 0x4e, + 0xff, 0xaa, 0x23, 0x77, 0x24, 0xa3, 0xef, 0xcd, 0x48, 0xf0, 0x11, 0x72, 0x8d, 0xb8, 0xa0, 0x33, + 0xe6, 0xb5, 0xba, 0x4e, 0xaf, 0x41, 0xea, 0xb1, 0x8c, 0x7e, 0xa0, 0x33, 0x86, 0x2f, 0x90, 0x5b, + 0x7c, 0xa3, 0xe7, 0x76, 0x9d, 0x5e, 0x73, 0xf0, 0x89, 0x6f, 0x47, 0x47, 0x13, 0xee, 0x3f, 0x14, + 0x93, 0x20, 0x16, 0x44, 0x4a, 0x38, 0xbe, 0x44, 0x1f, 0x40, 0x6b, 0x92, 0xd0, 0x45, 0x2c, 0x69, + 0xe8, 0x55, 0x80, 0x7f, 0x50, 0xf0, 0x0b, 0x6f, 0xfe, 0xb5, 0x58, 0xdc, 0xbd, 0x23, 0x2d, 0xc8, + 0x1f, 0x73, 0x2c, 0x3e, 0x43, 0x2d, 0xcd, 0x7e, 0xd7, 0x25, 0xb7, 0x6a, 0x6c, 0xdd, 0xbd, 0x23, + 0x4d, 0x73, 0x5a, 0x80, 0xae, 0x50, 0xeb, 0x97, 0x54, 0x8a, 0x12, 0x54, 0x03, 0x81, 0xf6, 0x86, + 0xc0, 0x13, 0x8c, 0xc6, 0xb0, 0x0d, 0xbc, 0x60, 0x9f, 0xa3, 0x46, 0x39, 0x15, 0xaf, 0x01, 0xd4, + 0xce, 0x06, 0xf5, 0xa7, 0x02, 0x41, 0x96, 0x60, 0x7c, 0x8b, 0xf6, 0x15, 0x0b, 0x18, 0x9f, 0xb3, + 0xc9, 0xb2, 0x83, 0xf7, 0xbf, 0x1d, 0xf6, 0x2c, 0xa9, 0x3c, 0xc1, 0x57, 0xc8, 0x2d, 0x6e, 0xdc, + 0x43, 0x5d, 0xa7, 0xb7, 0x33, 0xe8, 0xfa, 0x6b, 0x8b, 0x69, 0x56, 0xc3, 0x1f, 0xc9, 0xe8, 0xc9, + 0xe2, 0x48, 0xc9, 0xc0, 0xc7, 0xa8, 0xc1, 0x45, 0xca, 0x94, 0x9e, 0xf0, 0xd0, 0x7b, 0x0f, 0xf7, + 0xe6, 0xe6, 0x07, 0xf7, 0x21, 0xbe, 0x41, 0xad, 0xd5, 0xc5, 0xf3, 0xea, 0x60, 0xef, 0xed, 0xf6, + 0x77, 0x5a, 0x27, 0x24, 0xc7, 0x91, 0xe6, 0xeb, 0x32, 0xc1, 0xdf, 0xa1, 0x5a, 0x4c, 0x5f, 0x58, + 0x9c, 0x7a, 0xcd, 0x6e, 0xb5, 0xd7, 0x1c, 0x7c, 0xe5, 0x6f, 0x3c, 0x1b, 0xbf, 0xd8, 0x22, 0x7f, + 0x04, 0x48, 0x88, 0x89, 0xa5, 0xe1, 0x6b, 0xe4, 0xce, 0x98, 0xa6, 0x21, 0xd5, 0xd4, 0x3b, 0x02, + 0x07, 0x5f, 0x6c, 0x5d, 0x9f, 0x07, 0x0b, 0x26, 0x25, 0x0d, 0x0f, 0x51, 0x43, 0x26, 0x4c, 0xc1, + 0x63, 0xf4, 0x76, 0xa1, 0xc7, 0xe7, 0x5b, 0x6c, 0xfc, 0x58, 0x60, 0xc9, 0x92, 0x86, 0x0f, 0xd0, + 0x47, 0x5a, 0xd1, 0x80, 0x79, 0x87, 0x30, 0xa5, 0x3c, 0xc1, 0x6d, 0x54, 0x4f, 0x13, 0x2a, 0xcc, + 0xf4, 0x8e, 0xe1, 0xbc, 0x66, 0xd2, 0xfb, 0x10, 0x13, 0xb4, 0x9b, 0xdb, 0x99, 0xc4, 0x32, 0xc8, + 0x85, 0xdb, 0x20, 0xfc, 0xf5, 0x16, 0xe1, 0x27, 0x60, 0x8c, 0x2c, 0x81, 0xec, 0xa4, 0xff, 0xc9, + 0x3b, 0x17, 0xa8, 0xb9, 0x32, 0x20, 0xbc, 0x87, 0xaa, 0xbf, 0xb2, 0x85, 0xe7, 0x80, 0xae, 0x09, + 0x8d, 0xc7, 0x39, 0x8d, 0x33, 0x06, 0xcf, 0xa4, 0x41, 0xf2, 0xe4, 0xdb, 0xca, 0xb9, 0x33, 0x6c, + 0xa0, 0xba, 0xdd, 0xf0, 0x53, 0x8e, 0xf6, 0x37, 0x3e, 0x14, 0xef, 0xa0, 0x0a, 0x0f, 0x6d, 0xab, + 0x0a, 0x0f, 0x71, 0x07, 0xb9, 0x89, 0x92, 0x61, 0x16, 0x30, 0x65, 0x9b, 0x95, 0xb9, 0x51, 0x99, + 0x72, 0x95, 0x6a, 0x78, 0x50, 0x2e, 0xc9, 0x13, 0x8c, 0xd1, 0xfb, 0x98, 0xa6, 0x1a, 0x96, 0xc8, + 0x25, 0x10, 0x9f, 0x8e, 0xd1, 0xe1, 0xdb, 0x9f, 0x66, 0xd0, 0x53, 0x1e, 0x33, 0xab, 0x08, 0x31, + 0x74, 0xe0, 0x22, 0x37, 0x5f, 0x25, 0x10, 0x1b, 0x1f, 0xd3, 0x4c, 0x04, 0x30, 0xbf, 0x6a, 0xee, + 0xa3, 0xc8, 0x87, 0x7f, 0x38, 0xe8, 0xe3, 0x40, 0xce, 0x36, 0xe7, 0x39, 0xfc, 0x50, 0xa8, 0x3e, + 0xc2, 0xff, 0xc0, 0xf9, 0xf9, 0xdc, 0x62, 0x22, 0x19, 0x53, 0x11, 0xf9, 0x52, 0x45, 0xfd, 0x88, + 0x09, 0x78, 0x5f, 0xfd, 0xbc, 0x44, 0x13, 0x9e, 0xae, 0xfc, 0xec, 0x2f, 0x6d, 0xf8, 0x8f, 0xe3, + 0xfc, 0x59, 0x69, 0xdf, 0xe6, 0xec, 0x9b, 0x58, 0x66, 0xa1, 0xb9, 0x2c, 0xd0, 0x79, 0x1e, 0xfc, + 0x5d, 0x54, 0xc6, 0x50, 0x19, 0xdb, 0xca, 0xf8, 0x79, 0xf0, 0x52, 0x83, 0xde, 0xdf, 0xfc, 0x1b, + 0x00, 0x00, 0xff, 0xff, 0xc7, 0x91, 0xad, 0xed, 0x47, 0x06, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/logging/v2/logging.pb.go b/vendor/google.golang.org/genproto/googleapis/logging/v2/logging.pb.go index a94f1f8de..ce8c746cc 100644 --- a/vendor/google.golang.org/genproto/googleapis/logging/v2/logging.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/logging/v2/logging.pb.go @@ -43,7 +43,7 @@ type DeleteLogRequest struct { // `"organizations/1234567890/logs/cloudresourcemanager.googleapis.com%2Factivity"`. // For more information about log names, see // [LogEntry][google.logging.v2.LogEntry]. - LogName string `protobuf:"bytes,1,opt,name=log_name,json=logName" json:"log_name,omitempty"` + LogName string `protobuf:"bytes,1,opt,name=log_name,json=logName,proto3" json:"log_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -53,7 +53,7 @@ func (m *DeleteLogRequest) Reset() { *m = DeleteLogRequest{} } func (m *DeleteLogRequest) String() string { return proto.CompactTextString(m) } func (*DeleteLogRequest) ProtoMessage() {} func (*DeleteLogRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{0} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{0} } func (m *DeleteLogRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteLogRequest.Unmarshal(m, b) @@ -95,7 +95,7 @@ type WriteLogEntriesRequest struct { // `"organizations/1234567890/logs/cloudresourcemanager.googleapis.com%2Factivity"`. // For more information about log names, see // [LogEntry][google.logging.v2.LogEntry]. - LogName string `protobuf:"bytes,1,opt,name=log_name,json=logName" json:"log_name,omitempty"` + LogName string `protobuf:"bytes,1,opt,name=log_name,json=logName,proto3" json:"log_name,omitempty"` // Optional. A default monitored resource object that is assigned to all log // entries in `entries` that do not specify a value for `resource`. Example: // @@ -104,17 +104,18 @@ type WriteLogEntriesRequest struct { // "zone": "us-central1-a", "instance_id": "00000000000000000000" }} // // See [LogEntry][google.logging.v2.LogEntry]. - Resource *monitoredres.MonitoredResource `protobuf:"bytes,2,opt,name=resource" json:"resource,omitempty"` + Resource *monitoredres.MonitoredResource `protobuf:"bytes,2,opt,name=resource,proto3" json:"resource,omitempty"` // Optional. Default labels that are added to the `labels` field of all log // entries in `entries`. If a log entry already has a label with the same key // as a label in this parameter, then the log entry's label is not changed. // See [LogEntry][google.logging.v2.LogEntry]. - Labels map[string]string `protobuf:"bytes,3,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Required. The log entries to send to Stackdriver Logging. The order of log // entries in this list does not matter. Values supplied in this method's // `log_name`, `resource`, and `labels` fields are copied into those log // entries in this list that do not include values for their corresponding - // fields. For more information, see the [LogEntry][google.logging.v2.LogEntry] type. + // fields. For more information, see the + // [LogEntry][google.logging.v2.LogEntry] type. // // If the `timestamp` or `insert_id` fields are missing in log entries, then // this method supplies the current time or a unique identifier, respectively. @@ -124,20 +125,25 @@ type WriteLogEntriesRequest struct { // // Log entries with timestamps that are more than the // [logs retention period](/logging/quota-policy) in the past or more than - // 24 hours in the future might be discarded. Discarding does not return - // an error. + // 24 hours in the future will not be available when calling `entries.list`. + // However, those log entries can still be exported with + // [LogSinks](/logging/docs/api/tasks/exporting-logs). // // To improve throughput and to avoid exceeding the // [quota limit](/logging/quota-policy) for calls to `entries.write`, // you should try to include several log entries in this list, // rather than calling this method for each individual log entry. - Entries []*LogEntry `protobuf:"bytes,4,rep,name=entries" json:"entries,omitempty"` + Entries []*LogEntry `protobuf:"bytes,4,rep,name=entries,proto3" json:"entries,omitempty"` // Optional. Whether valid entries should be written even if some other // entries fail due to INVALID_ARGUMENT or PERMISSION_DENIED errors. If any // entry is not written, then the response status is the error associated // with one of the failed entries and the response includes error details // keyed by the entries' zero-based index in the `entries.write` method. - PartialSuccess bool `protobuf:"varint,5,opt,name=partial_success,json=partialSuccess" json:"partial_success,omitempty"` + PartialSuccess bool `protobuf:"varint,5,opt,name=partial_success,json=partialSuccess,proto3" json:"partial_success,omitempty"` + // Optional. If true, the request should expect normal response, but the + // entries won't be persisted nor exported. Useful for checking whether the + // logging API endpoints are working properly before sending valuable data. + DryRun bool `protobuf:"varint,6,opt,name=dry_run,json=dryRun,proto3" json:"dry_run,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -147,7 +153,7 @@ func (m *WriteLogEntriesRequest) Reset() { *m = WriteLogEntriesRequest{} func (m *WriteLogEntriesRequest) String() string { return proto.CompactTextString(m) } func (*WriteLogEntriesRequest) ProtoMessage() {} func (*WriteLogEntriesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{1} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{1} } func (m *WriteLogEntriesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WriteLogEntriesRequest.Unmarshal(m, b) @@ -202,6 +208,13 @@ func (m *WriteLogEntriesRequest) GetPartialSuccess() bool { return false } +func (m *WriteLogEntriesRequest) GetDryRun() bool { + if m != nil { + return m.DryRun + } + return false +} + // Result returned from WriteLogEntries. // empty type WriteLogEntriesResponse struct { @@ -214,7 +227,7 @@ func (m *WriteLogEntriesResponse) Reset() { *m = WriteLogEntriesResponse func (m *WriteLogEntriesResponse) String() string { return proto.CompactTextString(m) } func (*WriteLogEntriesResponse) ProtoMessage() {} func (*WriteLogEntriesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{2} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{2} } func (m *WriteLogEntriesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WriteLogEntriesResponse.Unmarshal(m, b) @@ -242,7 +255,7 @@ type WriteLogEntriesPartialErrors struct { // // Failed requests for which no entries are written will not include // per-entry errors. - LogEntryErrors map[int32]*status.Status `protobuf:"bytes,1,rep,name=log_entry_errors,json=logEntryErrors" json:"log_entry_errors,omitempty" protobuf_key:"varint,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + LogEntryErrors map[int32]*status.Status `protobuf:"bytes,1,rep,name=log_entry_errors,json=logEntryErrors,proto3" json:"log_entry_errors,omitempty" protobuf_key:"varint,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -252,7 +265,7 @@ func (m *WriteLogEntriesPartialErrors) Reset() { *m = WriteLogEntriesPar func (m *WriteLogEntriesPartialErrors) String() string { return proto.CompactTextString(m) } func (*WriteLogEntriesPartialErrors) ProtoMessage() {} func (*WriteLogEntriesPartialErrors) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{3} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{3} } func (m *WriteLogEntriesPartialErrors) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_WriteLogEntriesPartialErrors.Unmarshal(m, b) @@ -286,7 +299,7 @@ type ListLogEntriesRequest struct { // `"my-project-1A"`. If present, these project identifiers are converted to // resource name format and added to the list of resources in // `resource_names`. - ProjectIds []string `protobuf:"bytes,1,rep,name=project_ids,json=projectIds" json:"project_ids,omitempty"` + ProjectIds []string `protobuf:"bytes,1,rep,name=project_ids,json=projectIds,proto3" json:"project_ids,omitempty"` // Deprecated: Do not use. // Required. Names of one or more parent resources from which to // retrieve log entries: // @@ -296,7 +309,7 @@ type ListLogEntriesRequest struct { // "folders/[FOLDER_ID]" // // Projects listed in the `project_ids` field are added to this list. - ResourceNames []string `protobuf:"bytes,8,rep,name=resource_names,json=resourceNames" json:"resource_names,omitempty"` + ResourceNames []string `protobuf:"bytes,8,rep,name=resource_names,json=resourceNames,proto3" json:"resource_names,omitempty"` // Optional. A filter that chooses which log entries to return. See [Advanced // Logs Filters](/logging/docs/view/advanced_filters). Only log entries that // match the filter are returned. An empty filter matches all log entries in @@ -304,23 +317,23 @@ type ListLogEntriesRequest struct { // that is not listed in `resource_names` will cause the filter to return no // results. // The maximum length of the filter is 20000 characters. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Optional. How the results should be sorted. Presently, the only permitted // values are `"timestamp asc"` (default) and `"timestamp desc"`. The first // option returns entries in order of increasing values of // `LogEntry.timestamp` (oldest first), and the second option returns entries // in order of decreasing timestamps (newest first). Entries with equal // timestamps are returned in order of their `insert_id` values. - OrderBy string `protobuf:"bytes,3,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy string `protobuf:"bytes,3,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` // Optional. The maximum number of results to return from this request. // Non-positive values are ignored. The presence of `next_page_token` in the // response indicates that more results might be available. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. If present, then retrieve the next batch of results from the // preceding call to this method. `page_token` must be the value of // `next_page_token` from the previous response. The values of other method // parameters should be identical to those in the previous call. - PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -330,7 +343,7 @@ func (m *ListLogEntriesRequest) Reset() { *m = ListLogEntriesRequest{} } func (m *ListLogEntriesRequest) String() string { return proto.CompactTextString(m) } func (*ListLogEntriesRequest) ProtoMessage() {} func (*ListLogEntriesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{4} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{4} } func (m *ListLogEntriesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListLogEntriesRequest.Unmarshal(m, b) @@ -350,6 +363,7 @@ func (m *ListLogEntriesRequest) XXX_DiscardUnknown() { var xxx_messageInfo_ListLogEntriesRequest proto.InternalMessageInfo +// Deprecated: Do not use. func (m *ListLogEntriesRequest) GetProjectIds() []string { if m != nil { return m.ProjectIds @@ -397,7 +411,7 @@ type ListLogEntriesResponse struct { // A list of log entries. If `entries` is empty, `nextPageToken` may still be // returned, indicating that more entries may exist. See `nextPageToken` for // more information. - Entries []*LogEntry `protobuf:"bytes,1,rep,name=entries" json:"entries,omitempty"` + Entries []*LogEntry `protobuf:"bytes,1,rep,name=entries,proto3" json:"entries,omitempty"` // If there might be more results than those appearing in this response, then // `nextPageToken` is included. To get the next set of results, call this // method again using the value of `nextPageToken` as `pageToken`. @@ -408,7 +422,7 @@ type ListLogEntriesResponse struct { // value for `page_token` to continue the search. Alternatively, consider // speeding up the search by changing your filter to specify a single log name // or resource type, or to narrow the time range of the search. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -418,7 +432,7 @@ func (m *ListLogEntriesResponse) Reset() { *m = ListLogEntriesResponse{} func (m *ListLogEntriesResponse) String() string { return proto.CompactTextString(m) } func (*ListLogEntriesResponse) ProtoMessage() {} func (*ListLogEntriesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{5} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{5} } func (m *ListLogEntriesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListLogEntriesResponse.Unmarshal(m, b) @@ -457,12 +471,12 @@ type ListMonitoredResourceDescriptorsRequest struct { // Optional. The maximum number of results to return from this request. // Non-positive values are ignored. The presence of `nextPageToken` in the // response indicates that more results might be available. - PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,1,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. If present, then retrieve the next batch of results from the // preceding call to this method. `pageToken` must be the value of // `nextPageToken` from the previous response. The values of other method // parameters should be identical to those in the previous call. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -474,7 +488,7 @@ func (m *ListMonitoredResourceDescriptorsRequest) Reset() { func (m *ListMonitoredResourceDescriptorsRequest) String() string { return proto.CompactTextString(m) } func (*ListMonitoredResourceDescriptorsRequest) ProtoMessage() {} func (*ListMonitoredResourceDescriptorsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{6} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{6} } func (m *ListMonitoredResourceDescriptorsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListMonitoredResourceDescriptorsRequest.Unmarshal(m, b) @@ -511,11 +525,11 @@ func (m *ListMonitoredResourceDescriptorsRequest) GetPageToken() string { // Result returned from ListMonitoredResourceDescriptors. type ListMonitoredResourceDescriptorsResponse struct { // A list of resource descriptors. - ResourceDescriptors []*monitoredres.MonitoredResourceDescriptor `protobuf:"bytes,1,rep,name=resource_descriptors,json=resourceDescriptors" json:"resource_descriptors,omitempty"` + ResourceDescriptors []*monitoredres.MonitoredResourceDescriptor `protobuf:"bytes,1,rep,name=resource_descriptors,json=resourceDescriptors,proto3" json:"resource_descriptors,omitempty"` // If there might be more results than those appearing in this response, then // `nextPageToken` is included. To get the next set of results, call this // method again using the value of `nextPageToken` as `pageToken`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -527,7 +541,7 @@ func (m *ListMonitoredResourceDescriptorsResponse) Reset() { func (m *ListMonitoredResourceDescriptorsResponse) String() string { return proto.CompactTextString(m) } func (*ListMonitoredResourceDescriptorsResponse) ProtoMessage() {} func (*ListMonitoredResourceDescriptorsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{7} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{7} } func (m *ListMonitoredResourceDescriptorsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListMonitoredResourceDescriptorsResponse.Unmarshal(m, b) @@ -569,16 +583,16 @@ type ListLogsRequest struct { // "organizations/[ORGANIZATION_ID]" // "billingAccounts/[BILLING_ACCOUNT_ID]" // "folders/[FOLDER_ID]" - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. The maximum number of results to return from this request. // Non-positive values are ignored. The presence of `nextPageToken` in the // response indicates that more results might be available. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional. If present, then retrieve the next batch of results from the // preceding call to this method. `pageToken` must be the value of // `nextPageToken` from the previous response. The values of other method // parameters should be identical to those in the previous call. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -588,7 +602,7 @@ func (m *ListLogsRequest) Reset() { *m = ListLogsRequest{} } func (m *ListLogsRequest) String() string { return proto.CompactTextString(m) } func (*ListLogsRequest) ProtoMessage() {} func (*ListLogsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{8} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{8} } func (m *ListLogsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListLogsRequest.Unmarshal(m, b) @@ -634,11 +648,11 @@ type ListLogsResponse struct { // A list of log names. For example, // `"projects/my-project/syslog"` or // `"organizations/123/cloudresourcemanager.googleapis.com%2Factivity"`. - LogNames []string `protobuf:"bytes,3,rep,name=log_names,json=logNames" json:"log_names,omitempty"` + LogNames []string `protobuf:"bytes,3,rep,name=log_names,json=logNames,proto3" json:"log_names,omitempty"` // If there might be more results than those appearing in this response, then // `nextPageToken` is included. To get the next set of results, call this // method again using the value of `nextPageToken` as `pageToken`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -648,7 +662,7 @@ func (m *ListLogsResponse) Reset() { *m = ListLogsResponse{} } func (m *ListLogsResponse) String() string { return proto.CompactTextString(m) } func (*ListLogsResponse) ProtoMessage() {} func (*ListLogsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_968d0a5616b17899, []int{9} + return fileDescriptor_logging_38da0f0e8dbb26ee, []int{9} } func (m *ListLogsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListLogsResponse.Unmarshal(m, b) @@ -714,13 +728,14 @@ type LoggingServiceV2Client interface { // Log entries written shortly before the delete operation might not be // deleted. DeleteLog(ctx context.Context, in *DeleteLogRequest, opts ...grpc.CallOption) (*empty.Empty, error) - // ## Log entry resources - // // Writes log entries to Stackdriver Logging. This API method is the // only way to send log entries to Stackdriver Logging. This method // is used, directly or indirectly, by the Stackdriver Logging agent // (fluentd) and all logging libraries configured to use Stackdriver // Logging. + // A single request may contain log entries for a maximum of 1000 + // different resources (projects, organizations, billing accounts or + // folders) WriteLogEntries(ctx context.Context, in *WriteLogEntriesRequest, opts ...grpc.CallOption) (*WriteLogEntriesResponse, error) // Lists log entries. Use this method to retrieve log entries from // Stackdriver Logging. For ways to export log entries, see @@ -787,21 +802,21 @@ func (c *loggingServiceV2Client) ListLogs(ctx context.Context, in *ListLogsReque return out, nil } -// Server API for LoggingServiceV2 service - +// LoggingServiceV2Server is the server API for LoggingServiceV2 service. type LoggingServiceV2Server interface { // Deletes all the log entries in a log. // The log reappears if it receives new entries. // Log entries written shortly before the delete operation might not be // deleted. DeleteLog(context.Context, *DeleteLogRequest) (*empty.Empty, error) - // ## Log entry resources - // // Writes log entries to Stackdriver Logging. This API method is the // only way to send log entries to Stackdriver Logging. This method // is used, directly or indirectly, by the Stackdriver Logging agent // (fluentd) and all logging libraries configured to use Stackdriver // Logging. + // A single request may contain log entries for a maximum of 1000 + // different resources (projects, organizations, billing accounts or + // folders) WriteLogEntries(context.Context, *WriteLogEntriesRequest) (*WriteLogEntriesResponse, error) // Lists log entries. Use this method to retrieve log entries from // Stackdriver Logging. For ways to export log entries, see @@ -939,71 +954,78 @@ var _LoggingServiceV2_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/logging/v2/logging.proto", fileDescriptor_logging_968d0a5616b17899) + proto.RegisterFile("google/logging/v2/logging.proto", fileDescriptor_logging_38da0f0e8dbb26ee) } -var fileDescriptor_logging_968d0a5616b17899 = []byte{ - // 991 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x56, 0xcd, 0x6e, 0xdb, 0x46, - 0x10, 0xc6, 0x4a, 0xb1, 0x23, 0x8d, 0x1a, 0x5b, 0xd9, 0xc4, 0xb2, 0x22, 0xd9, 0xb5, 0x4a, 0x23, - 0xb5, 0x22, 0x20, 0x24, 0xca, 0x22, 0x40, 0xe2, 0x20, 0x17, 0x27, 0x46, 0x51, 0xc0, 0x29, 0x0c, - 0xba, 0x75, 0x80, 0xc0, 0x80, 0x40, 0x49, 0x1b, 0x62, 0x1b, 0x8a, 0xcb, 0xee, 0xae, 0xe4, 0x2a, - 0x41, 0x2e, 0x39, 0xf4, 0x05, 0x7a, 0xe9, 0x33, 0xf4, 0xd0, 0xb7, 0xe8, 0xa5, 0x87, 0x5e, 0x8a, - 0x02, 0x7d, 0x80, 0x3c, 0x44, 0x8f, 0x05, 0x77, 0x97, 0x32, 0xf5, 0x13, 0x59, 0xee, 0x4d, 0x3b, - 0xf3, 0xed, 0xcc, 0x7c, 0xc3, 0x6f, 0x66, 0x05, 0x3b, 0x01, 0x63, 0x41, 0x48, 0x9c, 0x90, 0x05, - 0x01, 0x8d, 0x02, 0x67, 0xe8, 0xa6, 0x3f, 0xed, 0x98, 0x33, 0xc9, 0xf0, 0x4d, 0x0d, 0xb0, 0x53, - 0xeb, 0xd0, 0xad, 0x6d, 0x99, 0x3b, 0x7e, 0x4c, 0x1d, 0x3f, 0x8a, 0x98, 0xf4, 0x25, 0x65, 0x91, - 0xd0, 0x17, 0x6a, 0xbb, 0x19, 0x6f, 0x9f, 0x45, 0x54, 0x32, 0x4e, 0x7a, 0x6d, 0x4e, 0x04, 0x1b, - 0xf0, 0x2e, 0x31, 0xa0, 0xcf, 0xe6, 0xa6, 0x6d, 0x93, 0x48, 0xf2, 0x91, 0x81, 0x7c, 0x6a, 0x20, - 0xea, 0xd4, 0x19, 0xbc, 0x72, 0x7a, 0x03, 0xae, 0x12, 0x19, 0x7f, 0x7d, 0xda, 0x4f, 0xfa, 0xb1, - 0x4c, 0x2f, 0xef, 0x4c, 0x3b, 0x25, 0xed, 0x13, 0x21, 0xfd, 0x7e, 0x6c, 0x00, 0x9b, 0x06, 0xc0, - 0xe3, 0xae, 0x23, 0xa4, 0x2f, 0x07, 0xa6, 0x7c, 0xeb, 0x3e, 0x94, 0x9f, 0x91, 0x90, 0x48, 0x72, - 0xc4, 0x02, 0x8f, 0xfc, 0x30, 0x20, 0x42, 0xe2, 0x3b, 0x50, 0x48, 0xaa, 0x8b, 0xfc, 0x3e, 0xa9, - 0xa2, 0x06, 0x6a, 0x16, 0xbd, 0xeb, 0x21, 0x0b, 0xbe, 0xf1, 0xfb, 0xc4, 0xfa, 0x27, 0x07, 0x95, - 0x17, 0x9c, 0x2a, 0xf8, 0x61, 0x24, 0x39, 0x25, 0xe2, 0xf2, 0x5b, 0xf8, 0x11, 0x14, 0xd2, 0x86, - 0x54, 0x73, 0x0d, 0xd4, 0x2c, 0xb9, 0xdb, 0xb6, 0xe9, 0xb3, 0x1f, 0x53, 0xfb, 0x79, 0xda, 0x36, - 0xcf, 0x80, 0xbc, 0x31, 0x1c, 0x3f, 0x87, 0xd5, 0xd0, 0xef, 0x90, 0x50, 0x54, 0xf3, 0x8d, 0x7c, - 0xb3, 0xe4, 0x3e, 0xb0, 0x67, 0x3e, 0x90, 0x3d, 0xbf, 0x20, 0xfb, 0x48, 0xdd, 0x4b, 0x8c, 0x23, - 0xcf, 0x04, 0xc1, 0x0f, 0xe0, 0x3a, 0xd1, 0xa8, 0xea, 0x35, 0x15, 0xaf, 0x3e, 0x27, 0x9e, 0x09, - 0x35, 0xf2, 0x52, 0x2c, 0xde, 0x83, 0xf5, 0xd8, 0xe7, 0x92, 0xfa, 0x61, 0x5b, 0x0c, 0xba, 0x5d, - 0x22, 0x44, 0x75, 0xa5, 0x81, 0x9a, 0x05, 0x6f, 0xcd, 0x98, 0x4f, 0xb4, 0xb5, 0xf6, 0x08, 0x4a, - 0x99, 0xb4, 0xb8, 0x0c, 0xf9, 0xd7, 0x64, 0x64, 0xda, 0x91, 0xfc, 0xc4, 0xb7, 0x61, 0x65, 0xe8, - 0x87, 0x03, 0xdd, 0x87, 0xa2, 0xa7, 0x0f, 0xfb, 0xb9, 0x87, 0xc8, 0xba, 0x03, 0x9b, 0x33, 0x44, - 0x44, 0xcc, 0x22, 0x41, 0xac, 0x0f, 0x08, 0xb6, 0xa6, 0x7c, 0xc7, 0x3a, 0xef, 0x21, 0xe7, 0x8c, - 0x0b, 0xdc, 0x87, 0xf2, 0x58, 0x4f, 0x6d, 0xa2, 0x6c, 0x55, 0xa4, 0xf8, 0x3d, 0xbd, 0xbc, 0x5f, - 0x13, 0xa1, 0xc6, 0xe4, 0xf5, 0x51, 0xf7, 0x61, 0x2d, 0x9c, 0x30, 0xd6, 0xbe, 0x83, 0x5b, 0x73, - 0x60, 0x59, 0xb6, 0x2b, 0x9a, 0x6d, 0x33, 0xcb, 0xb6, 0xe4, 0xe2, 0xb4, 0x18, 0x1e, 0x77, 0xed, - 0x13, 0x25, 0xc3, 0x6c, 0x07, 0xfe, 0x44, 0xb0, 0x71, 0x44, 0x85, 0x9c, 0xd5, 0xd6, 0x0e, 0x94, - 0x62, 0xce, 0xbe, 0x27, 0x5d, 0xd9, 0xa6, 0x3d, 0x4d, 0xad, 0xe8, 0x81, 0x31, 0x7d, 0xdd, 0x13, - 0xf8, 0x2e, 0xac, 0xa5, 0x92, 0x51, 0x0a, 0x14, 0xd5, 0x82, 0xc2, 0xdc, 0x48, 0xad, 0x89, 0x0e, - 0x05, 0xae, 0xc0, 0xea, 0x2b, 0x1a, 0x4a, 0xc2, 0x4d, 0xfb, 0xcd, 0x29, 0xd1, 0x2e, 0xe3, 0x3d, - 0xc2, 0xdb, 0x9d, 0x51, 0x35, 0xaf, 0xb5, 0xab, 0xce, 0x07, 0x23, 0x5c, 0x87, 0x62, 0xec, 0x07, - 0xa4, 0x2d, 0xe8, 0x1b, 0x52, 0xbd, 0xa6, 0xa8, 0x15, 0x12, 0xc3, 0x09, 0x7d, 0x43, 0xf0, 0x36, - 0x80, 0x72, 0x4a, 0xf6, 0x9a, 0x44, 0x4a, 0x12, 0x45, 0x4f, 0xc1, 0xbf, 0x4d, 0x0c, 0xd6, 0x39, - 0x54, 0xa6, 0xf9, 0xe8, 0x2f, 0x9a, 0xd5, 0x21, 0xba, 0x82, 0x0e, 0x3f, 0x87, 0xf5, 0x88, 0xfc, - 0x28, 0xdb, 0x99, 0xa4, 0x9a, 0xc8, 0x8d, 0xc4, 0x7c, 0x3c, 0x4e, 0x4c, 0x60, 0x2f, 0x49, 0x3c, - 0x33, 0x58, 0xcf, 0x88, 0xe8, 0x72, 0x1a, 0x4b, 0xc6, 0xc7, 0xad, 0x9d, 0xe0, 0x87, 0x16, 0xf2, - 0xcb, 0x4d, 0xf3, 0xfb, 0x0d, 0x41, 0xf3, 0xf2, 0x3c, 0x86, 0xf2, 0x4b, 0xb8, 0x3d, 0xfe, 0x44, - 0xbd, 0x0b, 0xbf, 0xe1, 0xbf, 0xb7, 0x70, 0x21, 0x5c, 0xc4, 0xf3, 0x6e, 0xf1, 0xd9, 0x1c, 0x57, - 0xe8, 0xcb, 0xba, 0xf9, 0x20, 0x63, 0xfe, 0x15, 0x58, 0x8d, 0x7d, 0x4e, 0x22, 0x69, 0xa6, 0xd4, - 0x9c, 0x26, 0xfb, 0x92, 0x5b, 0xd8, 0x97, 0xfc, 0x74, 0x5f, 0x5e, 0x40, 0xf9, 0x22, 0x8d, 0xa1, - 0x5f, 0x87, 0x62, 0xba, 0x1e, 0xf5, 0x2e, 0x2b, 0x7a, 0x05, 0xb3, 0x1f, 0x97, 0xae, 0xdf, 0xfd, - 0x7b, 0x05, 0xca, 0x47, 0x5a, 0x20, 0x27, 0x84, 0x0f, 0x69, 0x97, 0x9c, 0xba, 0xf8, 0x1c, 0x8a, - 0xe3, 0x15, 0x8e, 0x77, 0xe7, 0xe8, 0x68, 0x7a, 0xc1, 0xd7, 0x2a, 0x29, 0x28, 0x7d, 0x2f, 0xec, - 0xc3, 0xe4, 0x31, 0xb1, 0xee, 0xbf, 0xff, 0xeb, 0xc3, 0xcf, 0xb9, 0xbd, 0xd6, 0x5d, 0x67, 0xe8, - 0x76, 0x88, 0xf4, 0xbf, 0x70, 0xde, 0xa6, 0x35, 0x3f, 0x31, 0xc3, 0x26, 0x9c, 0x56, 0xf2, 0x74, - 0x09, 0xa7, 0xf5, 0x0e, 0xff, 0x84, 0x60, 0x7d, 0x6a, 0x97, 0xe0, 0x7b, 0x4b, 0xef, 0xe7, 0x5a, - 0x6b, 0x19, 0xa8, 0xd9, 0x80, 0x5b, 0xaa, 0xb2, 0x8a, 0x75, 0x33, 0x79, 0x3a, 0xcd, 0x34, 0xec, - 0x9f, 0x27, 0xe0, 0x7d, 0xd4, 0xc2, 0xef, 0x11, 0xac, 0x4d, 0x0e, 0x1a, 0x6e, 0xce, 0x9b, 0xa7, - 0x79, 0xbb, 0xa5, 0x76, 0x6f, 0x09, 0xa4, 0xa9, 0xa2, 0xae, 0xaa, 0xd8, 0xb0, 0xca, 0xd9, 0x2a, - 0x42, 0x2a, 0x64, 0x52, 0xc4, 0xef, 0x08, 0x1a, 0x97, 0x0d, 0x03, 0xde, 0xff, 0x48, 0xb2, 0x25, - 0x26, 0xb5, 0xf6, 0xf8, 0x7f, 0xdd, 0x35, 0xa5, 0x37, 0x55, 0xe9, 0x16, 0x6e, 0x24, 0xa5, 0xf7, - 0x17, 0x95, 0xc8, 0xa1, 0x90, 0x8a, 0x17, 0x5b, 0x1f, 0xef, 0xcd, 0xb8, 0xac, 0xdd, 0x85, 0x18, - 0x93, 0x7e, 0x5b, 0xa5, 0xdf, 0xc4, 0x1b, 0x49, 0xfa, 0xb7, 0x7a, 0xc4, 0x9e, 0xb4, 0x9c, 0xd6, - 0x3b, 0x25, 0xa6, 0x83, 0x5f, 0x10, 0x6c, 0x74, 0x59, 0x7f, 0x36, 0xd2, 0xc1, 0x27, 0x46, 0xee, - 0xc7, 0x89, 0x52, 0x8f, 0xd1, 0xcb, 0x87, 0x06, 0x12, 0xb0, 0xd0, 0x8f, 0x02, 0x9b, 0xf1, 0xc0, - 0x09, 0x48, 0xa4, 0x74, 0xec, 0x68, 0x97, 0x1f, 0x53, 0x91, 0xf9, 0xa3, 0xf5, 0xd8, 0xfc, 0xfc, - 0x17, 0xa1, 0x5f, 0x73, 0x9b, 0x5f, 0xe9, 0xdb, 0x4f, 0x43, 0x36, 0xe8, 0xd9, 0x26, 0xb4, 0x7d, - 0xea, 0xfe, 0x91, 0x7a, 0xce, 0x94, 0xe7, 0xcc, 0x78, 0xce, 0x4e, 0xdd, 0xce, 0xaa, 0x8a, 0xfd, - 0xe5, 0x7f, 0x01, 0x00, 0x00, 0xff, 0xff, 0xfd, 0x96, 0xc5, 0x3e, 0x3a, 0x0a, 0x00, 0x00, +var fileDescriptor_logging_38da0f0e8dbb26ee = []byte{ + // 1093 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x56, 0x4f, 0x6f, 0xdc, 0x44, + 0x14, 0xd7, 0x6c, 0x9a, 0xcd, 0xee, 0x84, 0x26, 0xdb, 0x69, 0xb3, 0x71, 0x77, 0x93, 0x66, 0xb3, + 0xab, 0x90, 0x8d, 0x25, 0x6c, 0x69, 0x51, 0xa5, 0x36, 0x55, 0x0f, 0xa4, 0x8d, 0x10, 0x52, 0x8a, + 0x22, 0x07, 0x52, 0x29, 0x8a, 0xb4, 0x72, 0xec, 0x89, 0x35, 0xd4, 0xeb, 0x31, 0x33, 0xe3, 0x94, + 0x0d, 0xea, 0xa5, 0x07, 0xbe, 0x00, 0x17, 0xb8, 0x21, 0x6e, 0x70, 0xe0, 0x5b, 0x70, 0xe1, 0x08, + 0x17, 0x3e, 0x40, 0x6f, 0x9c, 0xb8, 0x71, 0x03, 0x79, 0x66, 0xec, 0x78, 0xff, 0x34, 0x49, 0xb9, + 0x79, 0xde, 0xfb, 0xcd, 0xbc, 0xf7, 0x7b, 0xf3, 0x7b, 0x6f, 0x0c, 0xd7, 0x02, 0x4a, 0x83, 0x10, + 0xdb, 0x21, 0x0d, 0x02, 0x12, 0x05, 0xf6, 0x59, 0x2f, 0xfb, 0xb4, 0x62, 0x46, 0x05, 0x45, 0xb7, + 0x14, 0xc0, 0xca, 0xac, 0x67, 0xbd, 0xc6, 0x8a, 0xde, 0xe3, 0xc6, 0xc4, 0x76, 0xa3, 0x88, 0x0a, + 0x57, 0x10, 0x1a, 0x71, 0xb5, 0xa1, 0xd1, 0x29, 0x78, 0x07, 0x34, 0x22, 0x82, 0x32, 0xec, 0xf7, + 0x19, 0xe6, 0x34, 0x61, 0x1e, 0xd6, 0xa0, 0xf5, 0xa9, 0x61, 0xfb, 0x38, 0x12, 0x6c, 0xa8, 0x21, + 0xf7, 0x34, 0x44, 0xae, 0x4e, 0x92, 0x53, 0xdb, 0x4f, 0x98, 0x0c, 0xa4, 0xfd, 0xcd, 0x71, 0x3f, + 0x1e, 0xc4, 0x22, 0xdb, 0xbc, 0x36, 0xee, 0x14, 0x64, 0x80, 0xb9, 0x70, 0x07, 0xb1, 0x06, 0x2c, + 0x6b, 0x00, 0x8b, 0x3d, 0x9b, 0x0b, 0x57, 0x24, 0x3a, 0xfd, 0xf6, 0x07, 0xb0, 0xf6, 0x14, 0x87, + 0x58, 0xe0, 0x3d, 0x1a, 0x38, 0xf8, 0xcb, 0x04, 0x73, 0x81, 0xee, 0xc2, 0x4a, 0x9a, 0x5d, 0xe4, + 0x0e, 0xb0, 0x01, 0x5a, 0xa0, 0x5b, 0x75, 0xe6, 0x42, 0x1a, 0x7c, 0xea, 0x0e, 0x70, 0xfb, 0xef, + 0x12, 0xac, 0x3f, 0x67, 0x44, 0xc2, 0x77, 0x23, 0xc1, 0x08, 0xe6, 0x57, 0xef, 0x42, 0x0f, 0x61, + 0x25, 0x2b, 0x88, 0x51, 0x6a, 0x81, 0xee, 0x7c, 0x6f, 0xd5, 0xd2, 0x75, 0x76, 0x63, 0x62, 0x3d, + 0xcb, 0xca, 0xe6, 0x68, 0x90, 0x93, 0xc3, 0xd1, 0x33, 0x58, 0x0e, 0xdd, 0x13, 0x1c, 0x72, 0x63, + 0xa6, 0x35, 0xd3, 0x9d, 0xef, 0xdd, 0xb7, 0x26, 0x2e, 0xc8, 0x9a, 0x9e, 0x90, 0xb5, 0x27, 0xf7, + 0xa5, 0xc6, 0xa1, 0xa3, 0x0f, 0x41, 0xf7, 0xe1, 0x1c, 0x56, 0x28, 0xe3, 0x86, 0x3c, 0xaf, 0x39, + 0xe5, 0x3c, 0x7d, 0xd4, 0xd0, 0xc9, 0xb0, 0x68, 0x13, 0x2e, 0xc6, 0x2e, 0x13, 0xc4, 0x0d, 0xfb, + 0x3c, 0xf1, 0x3c, 0xcc, 0xb9, 0x31, 0xdb, 0x02, 0xdd, 0x8a, 0xb3, 0xa0, 0xcd, 0x07, 0xca, 0x8a, + 0x96, 0xe1, 0x9c, 0xcf, 0x86, 0x7d, 0x96, 0x44, 0x46, 0x59, 0x02, 0xca, 0x3e, 0x1b, 0x3a, 0x49, + 0xd4, 0x78, 0x08, 0xe7, 0x0b, 0xf9, 0xa0, 0x1a, 0x9c, 0x79, 0x81, 0x87, 0xba, 0x4e, 0xe9, 0x27, + 0xba, 0x03, 0x67, 0xcf, 0xdc, 0x30, 0x51, 0x05, 0xaa, 0x3a, 0x6a, 0xb1, 0x5d, 0x7a, 0x00, 0xda, + 0x77, 0xe1, 0xf2, 0x04, 0x43, 0x1e, 0xd3, 0x88, 0xe3, 0xf6, 0x1b, 0x00, 0x57, 0xc6, 0x7c, 0xfb, + 0x2a, 0xa1, 0x5d, 0xc6, 0x28, 0xe3, 0x68, 0x00, 0x6b, 0xb9, 0xd0, 0xfa, 0x58, 0xda, 0x0c, 0x20, + 0x89, 0x3f, 0xb9, 0xba, 0x90, 0x23, 0x47, 0xe5, 0x55, 0x51, 0x4b, 0x55, 0xa0, 0x85, 0x70, 0xc4, + 0xd8, 0xf8, 0x1c, 0xde, 0x9e, 0x02, 0x2b, 0xb2, 0x9d, 0x55, 0x6c, 0xbb, 0x45, 0xb6, 0xf3, 0x3d, + 0x94, 0x25, 0xc3, 0x62, 0xcf, 0x3a, 0x90, 0xfa, 0x2c, 0x56, 0xe0, 0x77, 0x00, 0x97, 0xf6, 0x08, + 0x17, 0x93, 0xa2, 0xeb, 0xc0, 0xf9, 0x98, 0xd1, 0x2f, 0xb0, 0x27, 0xfa, 0xc4, 0x57, 0xd4, 0xaa, + 0x3b, 0x25, 0x03, 0x38, 0x50, 0x9b, 0x3f, 0xf1, 0x39, 0xda, 0x80, 0x0b, 0x99, 0x9e, 0xa4, 0x3c, + 0xb9, 0x51, 0x49, 0x71, 0xce, 0xcd, 0xcc, 0x9a, 0x8a, 0x94, 0xa3, 0x3a, 0x2c, 0x9f, 0x92, 0x50, + 0x60, 0xa6, 0xaf, 0x40, 0xaf, 0x52, 0x61, 0x53, 0xe6, 0x63, 0xd6, 0x3f, 0x19, 0x1a, 0x33, 0x4a, + 0xd8, 0x72, 0xbd, 0x33, 0x44, 0x4d, 0x58, 0x8d, 0xdd, 0x00, 0xf7, 0x39, 0x39, 0xc7, 0xc6, 0x0d, + 0x49, 0xaf, 0x92, 0x1a, 0x0e, 0xc8, 0x39, 0x46, 0xab, 0x10, 0x4a, 0xa7, 0xa0, 0x2f, 0x70, 0x24, + 0xf5, 0x52, 0x75, 0x24, 0xfc, 0xb3, 0xd4, 0xd0, 0x7e, 0x09, 0xeb, 0xe3, 0x9c, 0xd4, 0xad, 0x16, + 0x45, 0x0a, 0xde, 0x41, 0xa4, 0xef, 0xc3, 0xc5, 0x08, 0x7f, 0x25, 0xfa, 0x85, 0xa0, 0x8a, 0xc8, + 0xcd, 0xd4, 0xbc, 0x9f, 0x07, 0xc6, 0x70, 0x33, 0x0d, 0x3c, 0xd1, 0x75, 0x4f, 0x31, 0xf7, 0x18, + 0x89, 0x05, 0x65, 0x79, 0x79, 0x47, 0xf8, 0x81, 0x4b, 0xf9, 0x95, 0xc6, 0xf9, 0xfd, 0x02, 0x60, + 0xf7, 0xea, 0x38, 0x9a, 0xf2, 0x11, 0xbc, 0x93, 0x5f, 0x91, 0x7f, 0xe1, 0xd7, 0xfc, 0x37, 0x2f, + 0x9d, 0x16, 0x17, 0xe7, 0x39, 0xb7, 0xd9, 0x64, 0x8c, 0x77, 0xa8, 0xcb, 0xa2, 0xbe, 0x90, 0x9c, + 0x7f, 0x1d, 0x96, 0x63, 0x97, 0xe1, 0x48, 0xe8, 0x4e, 0xd5, 0xab, 0xd1, 0xba, 0x94, 0x2e, 0xad, + 0xcb, 0xcc, 0x78, 0x5d, 0x9e, 0xc3, 0xda, 0x45, 0x18, 0x4d, 0xbf, 0x09, 0xab, 0xd9, 0xec, 0x54, + 0x83, 0xae, 0xea, 0x54, 0xf4, 0xf0, 0xbc, 0x76, 0xfe, 0xbd, 0x3f, 0xe7, 0x60, 0x6d, 0x4f, 0x09, + 0xe4, 0x00, 0xb3, 0x33, 0xe2, 0xe1, 0xc3, 0x1e, 0xfa, 0x0b, 0xc0, 0x6a, 0x3e, 0xe0, 0x51, 0x67, + 0x8a, 0x90, 0xc6, 0xc7, 0x7f, 0xa3, 0x9e, 0x81, 0xb2, 0xd7, 0xc4, 0xda, 0x4d, 0x9f, 0x9a, 0xf6, + 0x0f, 0xe0, 0xf5, 0x1f, 0x6f, 0xbe, 0x2d, 0x7d, 0x0f, 0xcc, 0x56, 0xfa, 0x80, 0x7d, 0x9d, 0x65, + 0xfc, 0x58, 0xb7, 0x1a, 0xb7, 0xcd, 0xf4, 0x55, 0xe3, 0xb6, 0xf9, 0xea, 0x68, 0xd3, 0xdc, 0x18, + 0xc5, 0x50, 0x16, 0xb8, 0x11, 0x39, 0x57, 0x4f, 0x67, 0x01, 0xb8, 0x6e, 0xae, 0x8d, 0x02, 0x4f, + 0x69, 0xe8, 0x63, 0x56, 0x84, 0x6c, 0x99, 0x9b, 0xa3, 0x90, 0x13, 0x12, 0x86, 0x24, 0x0a, 0x3e, + 0xf2, 0x3c, 0x9a, 0x44, 0xc5, 0xb0, 0xe8, 0x1b, 0x00, 0x17, 0xc7, 0x86, 0x18, 0xda, 0xba, 0xf6, + 0x8b, 0xd1, 0x30, 0xaf, 0x03, 0xd5, 0xa3, 0x77, 0x45, 0x16, 0xa3, 0xde, 0xbe, 0x95, 0xe6, 0xa6, + 0x5b, 0x70, 0xfb, 0x65, 0x0a, 0xde, 0x06, 0x26, 0x7a, 0x0d, 0xe0, 0xc2, 0x68, 0x77, 0xa3, 0xee, + 0xb4, 0x26, 0x9e, 0x36, 0xd4, 0x1a, 0x5b, 0xd7, 0x40, 0xea, 0x2c, 0x9a, 0x32, 0x8b, 0xa5, 0x76, + 0xad, 0x98, 0x45, 0x48, 0xb8, 0x48, 0x93, 0xf8, 0x15, 0xc0, 0xd6, 0x55, 0x1d, 0x88, 0xb6, 0xdf, + 0x12, 0xec, 0x1a, 0xe3, 0xa1, 0xf1, 0xe8, 0x7f, 0xed, 0xd5, 0xa9, 0x77, 0x65, 0xea, 0x6d, 0x24, + 0xc5, 0x34, 0xb8, 0x2c, 0xc5, 0x7f, 0x01, 0xac, 0x64, 0x2d, 0x83, 0xda, 0x6f, 0x2f, 0x4e, 0x9e, + 0x57, 0xe7, 0x52, 0x8c, 0x8e, 0xff, 0x93, 0x92, 0xf3, 0x8f, 0x00, 0x2d, 0x49, 0x79, 0xa9, 0xd6, + 0x7e, 0x6c, 0xda, 0xe6, 0x2b, 0xa9, 0xa7, 0xa3, 0x7b, 0x68, 0xa5, 0xe8, 0xb8, 0x50, 0xb9, 0xf6, + 0x77, 0xd0, 0x7a, 0xd1, 0x3f, 0xa6, 0x70, 0x0d, 0x5a, 0x45, 0xcd, 0x22, 0x28, 0x57, 0xb7, 0x76, + 0x6f, 0xa0, 0x4e, 0xd1, 0x3d, 0xa1, 0x6c, 0x05, 0xdb, 0xf9, 0x0e, 0xc0, 0x25, 0x8f, 0x0e, 0x26, + 0x69, 0xed, 0xbc, 0xa7, 0x3b, 0x7e, 0x3f, 0xed, 0xd5, 0x7d, 0x70, 0xf4, 0x40, 0x43, 0x02, 0x1a, + 0xba, 0x51, 0x60, 0x51, 0x16, 0xd8, 0x01, 0x8e, 0x64, 0x27, 0xdb, 0xca, 0xe5, 0xc6, 0x84, 0x17, + 0x7e, 0x44, 0x1f, 0xe9, 0xcf, 0x7f, 0x00, 0xf8, 0xb9, 0xb4, 0xfc, 0xb1, 0xda, 0xfd, 0x24, 0xa4, + 0x89, 0x6f, 0xe9, 0xa3, 0xad, 0xc3, 0xde, 0x6f, 0x99, 0xe7, 0x58, 0x7a, 0x8e, 0xb5, 0xe7, 0xf8, + 0xb0, 0x77, 0x52, 0x96, 0x67, 0x7f, 0xf8, 0x5f, 0x00, 0x00, 0x00, 0xff, 0xff, 0xc7, 0xec, 0x91, + 0x2d, 0x5a, 0x0b, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/logging/v2/logging_config.pb.go b/vendor/google.golang.org/genproto/googleapis/logging/v2/logging_config.pb.go index 7360c461e..8f55753c9 100644 --- a/vendor/google.golang.org/genproto/googleapis/logging/v2/logging_config.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/logging/v2/logging_config.pb.go @@ -56,7 +56,7 @@ func (x LogSink_VersionFormat) String() string { return proto.EnumName(LogSink_VersionFormat_name, int32(x)) } func (LogSink_VersionFormat) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{0, 0} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{0, 0} } // Describes a sink used to export log entries to one of the following @@ -70,7 +70,7 @@ type LogSink struct { // limited to 100 characters and can include only the following characters: // upper and lower-case alphanumeric characters, underscores, hyphens, and // periods. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The export destination: // // "storage.googleapis.com/[GCS_BUCKET]" @@ -81,17 +81,17 @@ type LogSink struct { // have permission to write to the destination or else the log // entries are not exported. For more information, see // [Exporting Logs With Sinks](/logging/docs/api/tasks/exporting-logs). - Destination string `protobuf:"bytes,3,opt,name=destination" json:"destination,omitempty"` + Destination string `protobuf:"bytes,3,opt,name=destination,proto3" json:"destination,omitempty"` // Optional. // An [advanced logs filter](/logging/docs/view/advanced_filters). The only // exported log entries are those that are in the resource owning the sink and // that match the filter. For example: // // logName="projects/[PROJECT_ID]/logs/[LOG_ID]" AND severity>=ERROR - Filter string `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // Deprecated. The log entry format to use for this sink's exported log // entries. The v2 format is used by default and cannot be changed. - OutputVersionFormat LogSink_VersionFormat `protobuf:"varint,6,opt,name=output_version_format,json=outputVersionFormat,enum=google.logging.v2.LogSink_VersionFormat" json:"output_version_format,omitempty"` + OutputVersionFormat LogSink_VersionFormat `protobuf:"varint,6,opt,name=output_version_format,json=outputVersionFormat,proto3,enum=google.logging.v2.LogSink_VersionFormat" json:"output_version_format,omitempty"` // Deprecated: Do not use. // Output only. An IAM identity—a service account or group—under // which Stackdriver Logging writes the exported log entries to the sink's // destination. This field is set by @@ -106,7 +106,7 @@ type LogSink struct { // resource](/iam/docs/granting-roles-to-service-accounts#granting_access_to_a_service_account_for_a_resource). // Consult the destination service's documentation to determine the // appropriate IAM roles to assign to the identity. - WriterIdentity string `protobuf:"bytes,8,opt,name=writer_identity,json=writerIdentity" json:"writer_identity,omitempty"` + WriterIdentity string `protobuf:"bytes,8,opt,name=writer_identity,json=writerIdentity,proto3" json:"writer_identity,omitempty"` // Optional. This field applies only to sinks owned by organizations and // folders. If the field is false, the default, only the logs owned by the // sink's parent resource are available for export. If the field is true, then @@ -120,11 +120,11 @@ type LogSink struct { // // logName:("projects/test-project1/" OR "projects/test-project2/") AND // resource.type=gce_instance - IncludeChildren bool `protobuf:"varint,9,opt,name=include_children,json=includeChildren" json:"include_children,omitempty"` + IncludeChildren bool `protobuf:"varint,9,opt,name=include_children,json=includeChildren,proto3" json:"include_children,omitempty"` // Deprecated. This field is ignored when creating or updating sinks. - StartTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Deprecated: Do not use. // Deprecated. This field is ignored when creating or updating sinks. - EndTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,11,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Deprecated: Do not use. XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -134,7 +134,7 @@ func (m *LogSink) Reset() { *m = LogSink{} } func (m *LogSink) String() string { return proto.CompactTextString(m) } func (*LogSink) ProtoMessage() {} func (*LogSink) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{0} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{0} } func (m *LogSink) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LogSink.Unmarshal(m, b) @@ -175,6 +175,7 @@ func (m *LogSink) GetFilter() string { return "" } +// Deprecated: Do not use. func (m *LogSink) GetOutputVersionFormat() LogSink_VersionFormat { if m != nil { return m.OutputVersionFormat @@ -196,6 +197,7 @@ func (m *LogSink) GetIncludeChildren() bool { return false } +// Deprecated: Do not use. func (m *LogSink) GetStartTime() *timestamp.Timestamp { if m != nil { return m.StartTime @@ -203,6 +205,7 @@ func (m *LogSink) GetStartTime() *timestamp.Timestamp { return nil } +// Deprecated: Do not use. func (m *LogSink) GetEndTime() *timestamp.Timestamp { if m != nil { return m.EndTime @@ -218,16 +221,16 @@ type ListSinksRequest struct { // "organizations/[ORGANIZATION_ID]" // "billingAccounts/[BILLING_ACCOUNT_ID]" // "folders/[FOLDER_ID]" - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. If present, then retrieve the next batch of results from the // preceding call to this method. `pageToken` must be the value of // `nextPageToken` from the previous response. The values of other method // parameters should be identical to those in the previous call. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. The maximum number of results to return from this request. // Non-positive values are ignored. The presence of `nextPageToken` in the // response indicates that more results might be available. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -237,7 +240,7 @@ func (m *ListSinksRequest) Reset() { *m = ListSinksRequest{} } func (m *ListSinksRequest) String() string { return proto.CompactTextString(m) } func (*ListSinksRequest) ProtoMessage() {} func (*ListSinksRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{1} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{1} } func (m *ListSinksRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSinksRequest.Unmarshal(m, b) @@ -281,11 +284,11 @@ func (m *ListSinksRequest) GetPageSize() int32 { // Result returned from `ListSinks`. type ListSinksResponse struct { // A list of sinks. - Sinks []*LogSink `protobuf:"bytes,1,rep,name=sinks" json:"sinks,omitempty"` + Sinks []*LogSink `protobuf:"bytes,1,rep,name=sinks,proto3" json:"sinks,omitempty"` // If there might be more results than appear in this response, then // `nextPageToken` is included. To get the next set of results, call the same // method again using the value of `nextPageToken` as `pageToken`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -295,7 +298,7 @@ func (m *ListSinksResponse) Reset() { *m = ListSinksResponse{} } func (m *ListSinksResponse) String() string { return proto.CompactTextString(m) } func (*ListSinksResponse) ProtoMessage() {} func (*ListSinksResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{2} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{2} } func (m *ListSinksResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSinksResponse.Unmarshal(m, b) @@ -339,7 +342,7 @@ type GetSinkRequest struct { // "folders/[FOLDER_ID]/sinks/[SINK_ID]" // // Example: `"projects/my-project-id/sinks/my-sink-id"`. - SinkName string `protobuf:"bytes,1,opt,name=sink_name,json=sinkName" json:"sink_name,omitempty"` + SinkName string `protobuf:"bytes,1,opt,name=sink_name,json=sinkName,proto3" json:"sink_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -349,7 +352,7 @@ func (m *GetSinkRequest) Reset() { *m = GetSinkRequest{} } func (m *GetSinkRequest) String() string { return proto.CompactTextString(m) } func (*GetSinkRequest) ProtoMessage() {} func (*GetSinkRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{3} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{3} } func (m *GetSinkRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSinkRequest.Unmarshal(m, b) @@ -386,10 +389,10 @@ type CreateSinkRequest struct { // "folders/[FOLDER_ID]" // // Examples: `"projects/my-logging-project"`, `"organizations/123456789"`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The new sink, whose `name` parameter is a sink identifier that // is not already in use. - Sink *LogSink `protobuf:"bytes,2,opt,name=sink" json:"sink,omitempty"` + Sink *LogSink `protobuf:"bytes,2,opt,name=sink,proto3" json:"sink,omitempty"` // Optional. Determines the kind of IAM identity returned as `writer_identity` // in the new sink. If this value is omitted or set to false, and if the // sink's parent is a project, then the value returned as `writer_identity` is @@ -401,7 +404,7 @@ type CreateSinkRequest struct { // resource such as an organization, then the value of `writer_identity` will // be a unique service account used only for exports from the new sink. For // more information, see `writer_identity` in [LogSink][google.logging.v2.LogSink]. - UniqueWriterIdentity bool `protobuf:"varint,3,opt,name=unique_writer_identity,json=uniqueWriterIdentity" json:"unique_writer_identity,omitempty"` + UniqueWriterIdentity bool `protobuf:"varint,3,opt,name=unique_writer_identity,json=uniqueWriterIdentity,proto3" json:"unique_writer_identity,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -411,7 +414,7 @@ func (m *CreateSinkRequest) Reset() { *m = CreateSinkRequest{} } func (m *CreateSinkRequest) String() string { return proto.CompactTextString(m) } func (*CreateSinkRequest) ProtoMessage() {} func (*CreateSinkRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{4} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{4} } func (m *CreateSinkRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateSinkRequest.Unmarshal(m, b) @@ -463,10 +466,10 @@ type UpdateSinkRequest struct { // "folders/[FOLDER_ID]/sinks/[SINK_ID]" // // Example: `"projects/my-project-id/sinks/my-sink-id"`. - SinkName string `protobuf:"bytes,1,opt,name=sink_name,json=sinkName" json:"sink_name,omitempty"` + SinkName string `protobuf:"bytes,1,opt,name=sink_name,json=sinkName,proto3" json:"sink_name,omitempty"` // Required. The updated sink, whose name is the same identifier that appears // as part of `sink_name`. - Sink *LogSink `protobuf:"bytes,2,opt,name=sink" json:"sink,omitempty"` + Sink *LogSink `protobuf:"bytes,2,opt,name=sink,proto3" json:"sink,omitempty"` // Optional. See // [sinks.create](/logging/docs/api/reference/rest/v2/projects.sinks/create) // for a description of this field. When updating a sink, the effect of this @@ -479,7 +482,7 @@ type UpdateSinkRequest struct { // `writer_identity` is changed to a unique service account. // + It is an error if the old value is true and the new value is // set to false or defaulted to false. - UniqueWriterIdentity bool `protobuf:"varint,3,opt,name=unique_writer_identity,json=uniqueWriterIdentity" json:"unique_writer_identity,omitempty"` + UniqueWriterIdentity bool `protobuf:"varint,3,opt,name=unique_writer_identity,json=uniqueWriterIdentity,proto3" json:"unique_writer_identity,omitempty"` // Optional. Field mask that specifies the fields in `sink` that need // an update. A sink field will be overwritten if, and only if, it is // in the update mask. `name` and output only fields cannot be updated. @@ -494,7 +497,7 @@ type UpdateSinkRequest struct { // https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#fieldmask // // Example: `updateMask=filter`. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -504,7 +507,7 @@ func (m *UpdateSinkRequest) Reset() { *m = UpdateSinkRequest{} } func (m *UpdateSinkRequest) String() string { return proto.CompactTextString(m) } func (*UpdateSinkRequest) ProtoMessage() {} func (*UpdateSinkRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{5} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{5} } func (m *UpdateSinkRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateSinkRequest.Unmarshal(m, b) @@ -563,7 +566,7 @@ type DeleteSinkRequest struct { // "folders/[FOLDER_ID]/sinks/[SINK_ID]" // // Example: `"projects/my-project-id/sinks/my-sink-id"`. - SinkName string `protobuf:"bytes,1,opt,name=sink_name,json=sinkName" json:"sink_name,omitempty"` + SinkName string `protobuf:"bytes,1,opt,name=sink_name,json=sinkName,proto3" json:"sink_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -573,7 +576,7 @@ func (m *DeleteSinkRequest) Reset() { *m = DeleteSinkRequest{} } func (m *DeleteSinkRequest) String() string { return proto.CompactTextString(m) } func (*DeleteSinkRequest) ProtoMessage() {} func (*DeleteSinkRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_config_e3432177b05142c4, []int{6} + return fileDescriptor_logging_config_1517a6dec11632fb, []int{6} } func (m *DeleteSinkRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteSinkRequest.Unmarshal(m, b) @@ -610,9 +613,9 @@ type LogExclusion struct { // Required. A client-assigned identifier, such as // `"load-balancer-exclusion"`. Identifiers are limited to 100 characters and // can include only letters, digits, underscores, hyphens, and periods. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Optional. A description of this exclusion. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Required. // An [advanced logs filter](/logging/docs/view/advanced_filters) // that matches the log entries to be excluded. By using the @@ -621,13 +624,13 @@ type LogExclusion struct { // For example, the following filter matches 99% of low-severity log // entries from load balancers: // - // "resource.type=http_load_balancer severity=ERROR" // // The maximum length of the filter is 20000 characters. - Filter string `protobuf:"bytes,3,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,3,opt,name=filter,proto3" json:"filter,omitempty"` // Optional. The metric descriptor associated with the logs-based metric. // If unspecified, it uses a default metric descriptor with a DELTA metric // kind, INT64 value type, with no labels and a unit of "1". Such a metric @@ -109,7 +109,7 @@ type LogMetric struct { // be updated once initially configured. New labels can be added in the // `metric_descriptor`, but existing labels cannot be modified except for // their description. - MetricDescriptor *metric.MetricDescriptor `protobuf:"bytes,5,opt,name=metric_descriptor,json=metricDescriptor" json:"metric_descriptor,omitempty"` + MetricDescriptor *metric.MetricDescriptor `protobuf:"bytes,5,opt,name=metric_descriptor,json=metricDescriptor,proto3" json:"metric_descriptor,omitempty"` // Optional. A `value_extractor` is required when using a distribution // logs-based metric to extract the values to record from a log entry. // Two functions are supported for value extraction: `EXTRACT(field)` or @@ -129,7 +129,7 @@ type LogMetric struct { // distribution. // // Example: `REGEXP_EXTRACT(jsonPayload.request, ".*quantity=(\d+).*")` - ValueExtractor string `protobuf:"bytes,6,opt,name=value_extractor,json=valueExtractor" json:"value_extractor,omitempty"` + ValueExtractor string `protobuf:"bytes,6,opt,name=value_extractor,json=valueExtractor,proto3" json:"value_extractor,omitempty"` // Optional. A map from a label key string to an extractor expression which is // used to extract data from a log entry field and assign as the label value. // Each label key specified in the LabelDescriptor must have an associated @@ -144,14 +144,14 @@ type LogMetric struct { // // Note that there are upper bounds on the maximum number of labels and the // number of active time series that are allowed in a project. - LabelExtractors map[string]string `protobuf:"bytes,7,rep,name=label_extractors,json=labelExtractors" json:"label_extractors,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + LabelExtractors map[string]string `protobuf:"bytes,7,rep,name=label_extractors,json=labelExtractors,proto3" json:"label_extractors,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Optional. The `bucket_options` are required when the logs-based metric is // using a DISTRIBUTION value type and it describes the bucket boundaries // used to create a histogram of the extracted values. - BucketOptions *distribution.Distribution_BucketOptions `protobuf:"bytes,8,opt,name=bucket_options,json=bucketOptions" json:"bucket_options,omitempty"` + BucketOptions *distribution.Distribution_BucketOptions `protobuf:"bytes,8,opt,name=bucket_options,json=bucketOptions,proto3" json:"bucket_options,omitempty"` // Deprecated. The API version that created or updated this metric. // The v2 format is used by default and cannot be changed. - Version LogMetric_ApiVersion `protobuf:"varint,4,opt,name=version,enum=google.logging.v2.LogMetric_ApiVersion" json:"version,omitempty"` + Version LogMetric_ApiVersion `protobuf:"varint,4,opt,name=version,proto3,enum=google.logging.v2.LogMetric_ApiVersion" json:"version,omitempty"` // Deprecated: Do not use. XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -161,7 +161,7 @@ func (m *LogMetric) Reset() { *m = LogMetric{} } func (m *LogMetric) String() string { return proto.CompactTextString(m) } func (*LogMetric) ProtoMessage() {} func (*LogMetric) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{0} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{0} } func (m *LogMetric) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LogMetric.Unmarshal(m, b) @@ -230,6 +230,7 @@ func (m *LogMetric) GetBucketOptions() *distribution.Distribution_BucketOptions return nil } +// Deprecated: Do not use. func (m *LogMetric) GetVersion() LogMetric_ApiVersion { if m != nil { return m.Version @@ -242,16 +243,16 @@ type ListLogMetricsRequest struct { // Required. The name of the project containing the metrics: // // "projects/[PROJECT_ID]" - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. If present, then retrieve the next batch of results from the // preceding call to this method. `pageToken` must be the value of // `nextPageToken` from the previous response. The values of other method // parameters should be identical to those in the previous call. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional. The maximum number of results to return from this request. // Non-positive values are ignored. The presence of `nextPageToken` in the // response indicates that more results might be available. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -261,7 +262,7 @@ func (m *ListLogMetricsRequest) Reset() { *m = ListLogMetricsRequest{} } func (m *ListLogMetricsRequest) String() string { return proto.CompactTextString(m) } func (*ListLogMetricsRequest) ProtoMessage() {} func (*ListLogMetricsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{1} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{1} } func (m *ListLogMetricsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListLogMetricsRequest.Unmarshal(m, b) @@ -305,11 +306,11 @@ func (m *ListLogMetricsRequest) GetPageSize() int32 { // Result returned from ListLogMetrics. type ListLogMetricsResponse struct { // A list of logs-based metrics. - Metrics []*LogMetric `protobuf:"bytes,1,rep,name=metrics" json:"metrics,omitempty"` + Metrics []*LogMetric `protobuf:"bytes,1,rep,name=metrics,proto3" json:"metrics,omitempty"` // If there might be more results than appear in this response, then // `nextPageToken` is included. To get the next set of results, call this // method again using the value of `nextPageToken` as `pageToken`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -319,7 +320,7 @@ func (m *ListLogMetricsResponse) Reset() { *m = ListLogMetricsResponse{} func (m *ListLogMetricsResponse) String() string { return proto.CompactTextString(m) } func (*ListLogMetricsResponse) ProtoMessage() {} func (*ListLogMetricsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{2} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{2} } func (m *ListLogMetricsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListLogMetricsResponse.Unmarshal(m, b) @@ -358,7 +359,7 @@ type GetLogMetricRequest struct { // The resource name of the desired metric: // // "projects/[PROJECT_ID]/metrics/[METRIC_ID]" - MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName" json:"metric_name,omitempty"` + MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName,proto3" json:"metric_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -368,7 +369,7 @@ func (m *GetLogMetricRequest) Reset() { *m = GetLogMetricRequest{} } func (m *GetLogMetricRequest) String() string { return proto.CompactTextString(m) } func (*GetLogMetricRequest) ProtoMessage() {} func (*GetLogMetricRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{3} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{3} } func (m *GetLogMetricRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetLogMetricRequest.Unmarshal(m, b) @@ -402,10 +403,10 @@ type CreateLogMetricRequest struct { // "projects/[PROJECT_ID]" // // The new metric must be provided in the request. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The new logs-based metric, which must not have an identifier that // already exists. - Metric *LogMetric `protobuf:"bytes,2,opt,name=metric" json:"metric,omitempty"` + Metric *LogMetric `protobuf:"bytes,2,opt,name=metric,proto3" json:"metric,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -415,7 +416,7 @@ func (m *CreateLogMetricRequest) Reset() { *m = CreateLogMetricRequest{} func (m *CreateLogMetricRequest) String() string { return proto.CompactTextString(m) } func (*CreateLogMetricRequest) ProtoMessage() {} func (*CreateLogMetricRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{4} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{4} } func (m *CreateLogMetricRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateLogMetricRequest.Unmarshal(m, b) @@ -458,9 +459,9 @@ type UpdateLogMetricRequest struct { // The updated metric must be provided in the request and it's // `name` field must be the same as `[METRIC_ID]` If the metric // does not exist in `[PROJECT_ID]`, then a new metric is created. - MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName" json:"metric_name,omitempty"` + MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName,proto3" json:"metric_name,omitempty"` // The updated metric. - Metric *LogMetric `protobuf:"bytes,2,opt,name=metric" json:"metric,omitempty"` + Metric *LogMetric `protobuf:"bytes,2,opt,name=metric,proto3" json:"metric,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -470,7 +471,7 @@ func (m *UpdateLogMetricRequest) Reset() { *m = UpdateLogMetricRequest{} func (m *UpdateLogMetricRequest) String() string { return proto.CompactTextString(m) } func (*UpdateLogMetricRequest) ProtoMessage() {} func (*UpdateLogMetricRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{5} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{5} } func (m *UpdateLogMetricRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateLogMetricRequest.Unmarshal(m, b) @@ -509,7 +510,7 @@ type DeleteLogMetricRequest struct { // The resource name of the metric to delete: // // "projects/[PROJECT_ID]/metrics/[METRIC_ID]" - MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName" json:"metric_name,omitempty"` + MetricName string `protobuf:"bytes,1,opt,name=metric_name,json=metricName,proto3" json:"metric_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -519,7 +520,7 @@ func (m *DeleteLogMetricRequest) Reset() { *m = DeleteLogMetricRequest{} func (m *DeleteLogMetricRequest) String() string { return proto.CompactTextString(m) } func (*DeleteLogMetricRequest) ProtoMessage() {} func (*DeleteLogMetricRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_logging_metrics_ec8e3ff261e4c5c2, []int{6} + return fileDescriptor_logging_metrics_2af73e5e02c8b3e5, []int{6} } func (m *DeleteLogMetricRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteLogMetricRequest.Unmarshal(m, b) @@ -635,8 +636,7 @@ func (c *metricsServiceV2Client) DeleteLogMetric(ctx context.Context, in *Delete return out, nil } -// Server API for MetricsServiceV2 service - +// MetricsServiceV2Server is the server API for MetricsServiceV2 service. type MetricsServiceV2Server interface { // Lists logs-based metrics. ListLogMetrics(context.Context, *ListLogMetricsRequest) (*ListLogMetricsResponse, error) @@ -774,62 +774,63 @@ var _MetricsServiceV2_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/logging/v2/logging_metrics.proto", fileDescriptor_logging_metrics_ec8e3ff261e4c5c2) + proto.RegisterFile("google/logging/v2/logging_metrics.proto", fileDescriptor_logging_metrics_2af73e5e02c8b3e5) } -var fileDescriptor_logging_metrics_ec8e3ff261e4c5c2 = []byte{ - // 846 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x56, 0xc1, 0x6e, 0xdb, 0x46, - 0x10, 0xed, 0xca, 0xb6, 0x1c, 0x8f, 0x1b, 0x4b, 0xd9, 0x24, 0x8a, 0xa0, 0x38, 0x88, 0xca, 0x83, - 0xa5, 0xf8, 0x40, 0x36, 0x6c, 0x61, 0xa4, 0x29, 0x7a, 0x88, 0x62, 0x23, 0x28, 0xa0, 0xb4, 0x06, - 0xd3, 0xea, 0x50, 0x08, 0x20, 0x28, 0x69, 0x44, 0x6c, 0x45, 0x71, 0x59, 0xee, 0x4a, 0xb0, 0x53, - 0xe4, 0x52, 0xe4, 0x56, 0xa0, 0x87, 0xf6, 0x03, 0x72, 0xef, 0xa7, 0xf4, 0xda, 0x7e, 0x42, 0x3f, - 0xa2, 0xc7, 0x82, 0xcb, 0xa5, 0xcc, 0x48, 0x8c, 0x65, 0xf8, 0xe4, 0xdd, 0x79, 0x33, 0xfb, 0xde, - 0xcc, 0x3c, 0x59, 0x82, 0x96, 0xcf, 0xb9, 0x1f, 0xa0, 0x15, 0x70, 0xdf, 0x67, 0xa1, 0x6f, 0xcd, - 0xed, 0xec, 0xe8, 0x4e, 0x51, 0xc6, 0x6c, 0x28, 0xcc, 0x28, 0xe6, 0x92, 0xd3, 0x5b, 0x69, 0xa2, - 0xa9, 0x51, 0x73, 0x6e, 0x37, 0xf6, 0x75, 0xad, 0x17, 0x31, 0xcb, 0x0b, 0x43, 0x2e, 0x3d, 0xc9, - 0x78, 0xa8, 0x0b, 0x1a, 0x0f, 0x72, 0xe8, 0x88, 0x09, 0x19, 0xb3, 0xc1, 0x2c, 0xc1, 0x35, 0x7c, - 0x2f, 0x07, 0xa7, 0x4c, 0x1a, 0xb8, 0xaf, 0x01, 0x75, 0x1b, 0xcc, 0xc6, 0x16, 0x4e, 0x23, 0x79, - 0xae, 0xc1, 0xe6, 0x32, 0x38, 0x66, 0x18, 0x8c, 0xdc, 0xa9, 0x27, 0x26, 0x69, 0x86, 0xf1, 0x6e, - 0x13, 0x76, 0xba, 0xdc, 0x7f, 0xa9, 0x9e, 0xa4, 0x14, 0x36, 0x43, 0x6f, 0x8a, 0x75, 0xd2, 0x24, - 0xed, 0x1d, 0x47, 0x9d, 0x69, 0x13, 0x76, 0x47, 0x28, 0x86, 0x31, 0x8b, 0x12, 0x39, 0xf5, 0x92, - 0x82, 0xf2, 0x21, 0x5a, 0x83, 0xf2, 0x98, 0x05, 0x12, 0xe3, 0xfa, 0x86, 0x02, 0xf5, 0x8d, 0x7e, - 0x0d, 0xb7, 0x52, 0xa9, 0x6e, 0x96, 0xcd, 0xe3, 0xfa, 0x56, 0x93, 0xb4, 0x77, 0xed, 0x7d, 0x53, - 0xcf, 0xc7, 0x8b, 0x98, 0x99, 0x92, 0x1f, 0x2f, 0x72, 0x9c, 0xea, 0x74, 0x29, 0x42, 0x5b, 0x50, - 0x99, 0x7b, 0xc1, 0x0c, 0x5d, 0x3c, 0x93, 0xb1, 0x37, 0x4c, 0x1e, 0x2a, 0x2b, 0xae, 0x3d, 0x15, - 0x3e, 0xc9, 0xa2, 0xb4, 0x0f, 0xd5, 0xc0, 0x1b, 0x60, 0x70, 0x91, 0x28, 0xea, 0xdb, 0xcd, 0x8d, - 0xf6, 0xae, 0xfd, 0xd8, 0x5c, 0x59, 0x89, 0xb9, 0xe8, 0xdc, 0xec, 0x26, 0x45, 0x8b, 0x67, 0xc4, - 0x49, 0x28, 0xe3, 0x73, 0xa7, 0x12, 0xbc, 0x1f, 0xa5, 0x2f, 0x61, 0x6f, 0x30, 0x1b, 0x4e, 0x50, - 0xba, 0x5c, 0xb5, 0x2e, 0xea, 0x37, 0x54, 0x3b, 0x07, 0xf9, 0x76, 0x8e, 0xf3, 0xdb, 0xeb, 0xa8, - 0xf4, 0x6f, 0xd3, 0x6c, 0xe7, 0xe6, 0x20, 0x7f, 0xa5, 0xcf, 0x60, 0x7b, 0x8e, 0xb1, 0x48, 0xc6, - 0xba, 0xd9, 0x24, 0xed, 0x3d, 0xbb, 0x75, 0xa9, 0xc6, 0x67, 0x11, 0xeb, 0xa5, 0xe9, 0x4e, 0x56, - 0xd7, 0xe8, 0xc0, 0x9d, 0x22, 0xe9, 0xb4, 0x0a, 0x1b, 0x13, 0x3c, 0xd7, 0x8b, 0x4c, 0x8e, 0xf4, - 0x0e, 0x6c, 0xa9, 0x59, 0xe9, 0x0d, 0xa6, 0x97, 0xa7, 0xa5, 0x27, 0xc4, 0xd8, 0x07, 0xb8, 0x78, - 0x9a, 0x96, 0xa1, 0xd4, 0xb3, 0xab, 0x1f, 0xa9, 0xbf, 0x8f, 0xab, 0xc4, 0x98, 0xc0, 0xdd, 0x2e, - 0x13, 0x72, 0x21, 0x43, 0x38, 0xf8, 0xd3, 0x0c, 0x85, 0x4c, 0xd6, 0x1e, 0x79, 0x31, 0x86, 0x52, - 0xb3, 0xe8, 0x1b, 0x7d, 0x00, 0x10, 0x79, 0x3e, 0xba, 0x92, 0x4f, 0x30, 0xf3, 0xcb, 0x4e, 0x12, - 0xf9, 0x2e, 0x09, 0xd0, 0xfb, 0xa0, 0x2e, 0xae, 0x60, 0xaf, 0x51, 0x19, 0x66, 0xcb, 0xb9, 0x91, - 0x04, 0x5e, 0xb1, 0xd7, 0x68, 0x9c, 0x41, 0x6d, 0x99, 0x4c, 0x44, 0x3c, 0x14, 0x48, 0x8f, 0x60, - 0x5b, 0x7f, 0xc2, 0xea, 0x44, 0xed, 0x73, 0xff, 0xb2, 0x59, 0x39, 0x59, 0x32, 0x3d, 0x80, 0x4a, - 0x88, 0x67, 0xd2, 0x5d, 0x91, 0x74, 0x33, 0x09, 0x9f, 0x66, 0xb2, 0x8c, 0x23, 0xb8, 0xfd, 0x02, - 0x2f, 0x88, 0xb3, 0x26, 0x1f, 0xc2, 0xae, 0xf6, 0x70, 0xee, 0x83, 0x01, 0x69, 0xe8, 0x1b, 0x6f, - 0x8a, 0xc6, 0x18, 0x6a, 0xcf, 0x63, 0xf4, 0x24, 0xae, 0x94, 0x7e, 0x68, 0x3e, 0x9f, 0x43, 0x39, - 0xad, 0x57, 0x42, 0xd6, 0x35, 0xa2, 0x73, 0x0d, 0x0e, 0xb5, 0xef, 0xa3, 0x51, 0x11, 0xcf, 0x3a, - 0x89, 0xd7, 0x24, 0xfc, 0x02, 0x6a, 0xc7, 0x18, 0xe0, 0x35, 0x08, 0xed, 0x7f, 0xb6, 0xa0, 0xaa, - 0xf7, 0xf7, 0x0a, 0xe3, 0x39, 0x1b, 0x62, 0xcf, 0xa6, 0xbf, 0x11, 0xd8, 0x7b, 0x7f, 0xb7, 0xb4, - 0x5d, 0x24, 0xa4, 0xc8, 0x6b, 0x8d, 0x47, 0x57, 0xc8, 0x4c, 0x8d, 0x62, 0xb4, 0x7e, 0xf9, 0xfb, - 0xdf, 0x3f, 0x4a, 0x9f, 0xd0, 0x87, 0xc9, 0x3f, 0xe7, 0x9f, 0xd3, 0x99, 0x7f, 0x15, 0xc5, 0xfc, - 0x47, 0x1c, 0x4a, 0x61, 0x1d, 0xbe, 0xb1, 0x32, 0x67, 0xbc, 0x25, 0xf0, 0x71, 0x7e, 0xe5, 0xf4, - 0xa0, 0x80, 0xa4, 0xc0, 0x13, 0x8d, 0x4b, 0xe7, 0x67, 0x98, 0x8a, 0xbf, 0x4d, 0x0f, 0x14, 0x7f, - 0x6e, 0x50, 0x39, 0x11, 0x99, 0x06, 0xeb, 0xf0, 0x0d, 0xfd, 0x95, 0x40, 0x65, 0xc9, 0x41, 0xb4, - 0xa8, 0xdd, 0x62, 0x97, 0xad, 0x11, 0x63, 0x29, 0x31, 0x8f, 0x8c, 0x75, 0xc3, 0x78, 0xaa, 0xb7, - 0x4e, 0x7f, 0x27, 0x50, 0x59, 0xf2, 0x59, 0xa1, 0x9a, 0x62, 0x2f, 0xae, 0x51, 0x73, 0xa4, 0xd4, - 0x7c, 0xda, 0xb8, 0xe2, 0x68, 0x16, 0xa2, 0xde, 0x12, 0xa8, 0x2c, 0x79, 0xb1, 0x50, 0x54, 0xb1, - 0x5f, 0x1b, 0xb5, 0x2c, 0x35, 0xfb, 0x1a, 0x34, 0x4f, 0x92, 0xef, 0xc8, 0x6c, 0x53, 0x87, 0x57, - 0x94, 0xd3, 0x79, 0x47, 0xe0, 0xee, 0x90, 0x4f, 0x57, 0x89, 0x3b, 0xb7, 0xbb, 0xe9, 0x59, 0x7b, - 0xf1, 0x34, 0xe1, 0x39, 0x25, 0x3f, 0x3c, 0xd1, 0x99, 0x3e, 0x0f, 0xbc, 0xd0, 0x37, 0x79, 0xec, - 0x5b, 0x3e, 0x86, 0x4a, 0x85, 0x95, 0x42, 0x5e, 0xc4, 0x44, 0xee, 0xc7, 0xc4, 0x97, 0xfa, 0xf8, - 0x1f, 0x21, 0x7f, 0x96, 0xee, 0xbd, 0x48, 0xab, 0x9f, 0x07, 0x7c, 0x36, 0x32, 0x35, 0x83, 0xd9, - 0xb3, 0xff, 0xca, 0x90, 0xbe, 0x42, 0xfa, 0x1a, 0xe9, 0xf7, 0xec, 0x41, 0x59, 0xbd, 0xfd, 0xd9, - 0xff, 0x01, 0x00, 0x00, 0xff, 0xff, 0xf7, 0xb9, 0xb3, 0xe4, 0xa7, 0x08, 0x00, 0x00, +var fileDescriptor_logging_metrics_2af73e5e02c8b3e5 = []byte{ + // 850 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x9c, 0x56, 0x41, 0x6f, 0x1b, 0x45, + 0x14, 0x66, 0x9c, 0xc4, 0x69, 0x5e, 0x68, 0xec, 0x4e, 0x5b, 0x77, 0xe5, 0xa6, 0xaa, 0xd9, 0x43, + 0xe2, 0xe6, 0xb0, 0x4b, 0x17, 0x14, 0x95, 0x22, 0x0e, 0xb8, 0x89, 0x2a, 0x24, 0x17, 0xa2, 0x2d, + 0xf8, 0x80, 0x2c, 0xad, 0xd6, 0xf6, 0xf3, 0x6a, 0xf0, 0x7a, 0x67, 0xd9, 0x19, 0x5b, 0x49, 0x51, + 0x2f, 0xa8, 0x37, 0x24, 0x0e, 0xf0, 0x03, 0x38, 0x70, 0xe3, 0xa7, 0x70, 0x85, 0x9f, 0xc0, 0x8f, + 0xe0, 0x88, 0x76, 0x76, 0xd6, 0xd9, 0xda, 0x4b, 0x1c, 0xf5, 0x94, 0x99, 0xf7, 0xbd, 0x37, 0xdf, + 0xf7, 0xde, 0xfb, 0x1c, 0x1b, 0x0e, 0x03, 0xce, 0x83, 0x10, 0xed, 0x90, 0x07, 0x01, 0x8b, 0x02, + 0x7b, 0xee, 0xe4, 0x47, 0x6f, 0x8a, 0x32, 0x61, 0x43, 0x61, 0xc5, 0x09, 0x97, 0x9c, 0xde, 0xca, + 0x12, 0x2d, 0x8d, 0x5a, 0x73, 0xa7, 0xb9, 0xaf, 0x6b, 0xfd, 0x98, 0xd9, 0x7e, 0x14, 0x71, 0xe9, + 0x4b, 0xc6, 0x23, 0x5d, 0xd0, 0x7c, 0x50, 0x40, 0x47, 0x4c, 0xc8, 0x84, 0x0d, 0x66, 0x29, 0xae, + 0xe1, 0x7b, 0x05, 0x38, 0x63, 0xd2, 0xc0, 0x7d, 0x0d, 0xa8, 0xdb, 0x60, 0x36, 0xb6, 0x71, 0x1a, + 0xcb, 0x0b, 0x0d, 0xb6, 0x96, 0xc1, 0x31, 0xc3, 0x70, 0xe4, 0x4d, 0x7d, 0x31, 0xc9, 0x32, 0xcc, + 0xdf, 0x37, 0x61, 0xa7, 0xcb, 0x83, 0x17, 0xea, 0x49, 0x4a, 0x61, 0x33, 0xf2, 0xa7, 0x68, 0x90, + 0x16, 0x69, 0xef, 0xb8, 0xea, 0x4c, 0x5b, 0xb0, 0x3b, 0x42, 0x31, 0x4c, 0x58, 0x9c, 0xca, 0x31, + 0x2a, 0x0a, 0x2a, 0x86, 0x68, 0x03, 0xaa, 0x63, 0x16, 0x4a, 0x4c, 0x8c, 0x0d, 0x05, 0xea, 0x1b, + 0xfd, 0x02, 0x6e, 0x65, 0x52, 0xbd, 0x3c, 0x9b, 0x27, 0xc6, 0x56, 0x8b, 0xb4, 0x77, 0x9d, 0x7d, + 0x4b, 0xcf, 0xc7, 0x8f, 0x99, 0x95, 0x91, 0x9f, 0x2c, 0x72, 0xdc, 0xfa, 0x74, 0x29, 0x42, 0x0f, + 0xa1, 0x36, 0xf7, 0xc3, 0x19, 0x7a, 0x78, 0x2e, 0x13, 0x7f, 0x98, 0x3e, 0x54, 0x55, 0x5c, 0x7b, + 0x2a, 0x7c, 0x9a, 0x47, 0x69, 0x1f, 0xea, 0xa1, 0x3f, 0xc0, 0xf0, 0x32, 0x51, 0x18, 0xdb, 0xad, + 0x8d, 0xf6, 0xae, 0xf3, 0xd8, 0x5a, 0x59, 0x89, 0xb5, 0xe8, 0xdc, 0xea, 0xa6, 0x45, 0x8b, 0x67, + 0xc4, 0x69, 0x24, 0x93, 0x0b, 0xb7, 0x16, 0xbe, 0x1d, 0xa5, 0x2f, 0x60, 0x6f, 0x30, 0x1b, 0x4e, + 0x50, 0x7a, 0x5c, 0xb5, 0x2e, 0x8c, 0x1b, 0xaa, 0x9d, 0x83, 0x62, 0x3b, 0x27, 0xc5, 0xed, 0x75, + 0x54, 0xfa, 0x57, 0x59, 0xb6, 0x7b, 0x73, 0x50, 0xbc, 0xd2, 0x53, 0xd8, 0x9e, 0x63, 0x22, 0xd2, + 0xb1, 0x6e, 0xb6, 0x48, 0x7b, 0xcf, 0x39, 0xbc, 0x52, 0xe3, 0xe7, 0x31, 0xeb, 0x65, 0xe9, 0x9d, + 0x8a, 0x41, 0xdc, 0xbc, 0xb6, 0xd9, 0x81, 0x3b, 0x65, 0xf2, 0x69, 0x1d, 0x36, 0x26, 0x78, 0xa1, + 0x97, 0x99, 0x1e, 0xe9, 0x1d, 0xd8, 0x52, 0xf3, 0xd2, 0x5b, 0xcc, 0x2e, 0x4f, 0x2b, 0x4f, 0x88, + 0xb9, 0x0f, 0x70, 0xf9, 0x3c, 0xad, 0x42, 0xa5, 0xe7, 0xd4, 0xdf, 0x53, 0x7f, 0x1f, 0xd7, 0x89, + 0x39, 0x81, 0xbb, 0x5d, 0x26, 0xe4, 0x42, 0x8a, 0x70, 0xf1, 0xfb, 0x19, 0x0a, 0x99, 0xae, 0x3e, + 0xf6, 0x13, 0x8c, 0xa4, 0x66, 0xd1, 0x37, 0xfa, 0x00, 0x20, 0xf6, 0x03, 0xf4, 0x24, 0x9f, 0x60, + 0xee, 0x99, 0x9d, 0x34, 0xf2, 0x75, 0x1a, 0xa0, 0xf7, 0x41, 0x5d, 0x3c, 0xc1, 0x5e, 0xa1, 0x32, + 0xcd, 0x96, 0x7b, 0x23, 0x0d, 0xbc, 0x64, 0xaf, 0xd0, 0x3c, 0x87, 0xc6, 0x32, 0x99, 0x88, 0x79, + 0x24, 0x90, 0x1e, 0xc3, 0xb6, 0xfe, 0x94, 0x19, 0x44, 0xed, 0x74, 0xff, 0xaa, 0x79, 0xb9, 0x79, + 0x32, 0x3d, 0x80, 0x5a, 0x84, 0xe7, 0xd2, 0x5b, 0x91, 0x74, 0x33, 0x0d, 0x9f, 0xe5, 0xb2, 0xcc, + 0x63, 0xb8, 0xfd, 0x1c, 0x2f, 0x89, 0xf3, 0x26, 0x1f, 0xc2, 0xae, 0xf6, 0x71, 0xe1, 0xc3, 0x01, + 0x59, 0xe8, 0x4b, 0x7f, 0x8a, 0xe6, 0x18, 0x1a, 0xcf, 0x12, 0xf4, 0x25, 0xae, 0x94, 0xfe, 0xdf, + 0x7c, 0x3e, 0x86, 0x6a, 0x56, 0xaf, 0x84, 0xac, 0x6b, 0x44, 0xe7, 0x9a, 0x1c, 0x1a, 0xdf, 0xc4, + 0xa3, 0x32, 0x9e, 0x75, 0x12, 0xdf, 0x91, 0xf0, 0x13, 0x68, 0x9c, 0x60, 0x88, 0xef, 0x40, 0xe8, + 0xfc, 0xbd, 0x05, 0x75, 0xbd, 0xbf, 0x97, 0x98, 0xcc, 0xd9, 0x10, 0x7b, 0x0e, 0xfd, 0x99, 0xc0, + 0xde, 0xdb, 0xbb, 0xa5, 0xed, 0x32, 0x21, 0x65, 0x5e, 0x6b, 0x3e, 0xba, 0x46, 0x66, 0x66, 0x14, + 0xf3, 0xf0, 0xc7, 0xbf, 0xfe, 0xf9, 0xb5, 0xf2, 0x01, 0x7d, 0x98, 0xfe, 0x83, 0xfe, 0x21, 0x9b, + 0xf9, 0x67, 0x71, 0xc2, 0xbf, 0xc3, 0xa1, 0x14, 0xf6, 0xd1, 0x6b, 0x3b, 0x77, 0xc6, 0x1b, 0x02, + 0xef, 0x17, 0x57, 0x4e, 0x0f, 0x4a, 0x48, 0x4a, 0x3c, 0xd1, 0xbc, 0x72, 0x7e, 0xa6, 0xa5, 0xf8, + 0xdb, 0xf4, 0x40, 0xf1, 0x17, 0x06, 0x55, 0x10, 0x91, 0x6b, 0xb0, 0x8f, 0x5e, 0xd3, 0x9f, 0x08, + 0xd4, 0x96, 0x1c, 0x44, 0xcb, 0xda, 0x2d, 0x77, 0xd9, 0x1a, 0x31, 0xb6, 0x12, 0xf3, 0xc8, 0x5c, + 0x37, 0x8c, 0xa7, 0x7a, 0xeb, 0xf4, 0x17, 0x02, 0xb5, 0x25, 0x9f, 0x95, 0xaa, 0x29, 0xf7, 0xe2, + 0x1a, 0x35, 0xc7, 0x4a, 0xcd, 0x87, 0xcd, 0x6b, 0x8e, 0x66, 0x21, 0xea, 0x0d, 0x81, 0xda, 0x92, + 0x17, 0x4b, 0x45, 0x95, 0xfb, 0xb5, 0xd9, 0xc8, 0x53, 0xf3, 0xaf, 0x42, 0xeb, 0x34, 0xfd, 0x9e, + 0xcc, 0x37, 0x75, 0x74, 0x4d, 0x39, 0x9d, 0xdf, 0x08, 0xdc, 0x1d, 0xf2, 0xe9, 0x2a, 0x71, 0xe7, + 0x76, 0x37, 0x3b, 0x6b, 0x2f, 0x9e, 0xa5, 0x3c, 0x67, 0xe4, 0xdb, 0x27, 0x3a, 0x33, 0xe0, 0xa1, + 0x1f, 0x05, 0x16, 0x4f, 0x02, 0x3b, 0xc0, 0x48, 0xa9, 0xb0, 0x33, 0xc8, 0x8f, 0x99, 0x28, 0xfc, + 0xa0, 0xf8, 0x54, 0x1f, 0xff, 0x25, 0xe4, 0x8f, 0xca, 0xbd, 0xe7, 0x59, 0xf5, 0xb3, 0x90, 0xcf, + 0x46, 0x96, 0x66, 0xb0, 0x7a, 0xce, 0x9f, 0x39, 0xd2, 0x57, 0x48, 0x5f, 0x23, 0xfd, 0x9e, 0x33, + 0xa8, 0xaa, 0xb7, 0x3f, 0xfa, 0x2f, 0x00, 0x00, 0xff, 0xff, 0xdc, 0x79, 0xf6, 0xb5, 0xab, 0x08, + 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/longrunning/operations.pb.go b/vendor/google.golang.org/genproto/googleapis/longrunning/operations.pb.go index 3828ec51f..7944499fb 100644 --- a/vendor/google.golang.org/genproto/googleapis/longrunning/operations.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/longrunning/operations.pb.go @@ -33,16 +33,16 @@ type Operation struct { // The server-assigned name, which is only unique within the same service that // originally returns it. If you use the default HTTP mapping, the // `name` should have the format of `operations/some/unique/name`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Service-specific metadata associated with the operation. It typically // contains progress information and common metadata such as create time. // Some services might not provide such metadata. Any method that returns a // long-running operation should document the metadata type, if any. - Metadata *any.Any `protobuf:"bytes,2,opt,name=metadata" json:"metadata,omitempty"` + Metadata *any.Any `protobuf:"bytes,2,opt,name=metadata,proto3" json:"metadata,omitempty"` // If the value is `false`, it means the operation is still in progress. // If true, the operation is completed, and either `error` or `response` is // available. - Done bool `protobuf:"varint,3,opt,name=done" json:"done,omitempty"` + Done bool `protobuf:"varint,3,opt,name=done,proto3" json:"done,omitempty"` // The operation result, which can be either an `error` or a valid `response`. // If `done` == `false`, neither `error` nor `response` is set. // If `done` == `true`, exactly one of `error` or `response` is set. @@ -85,10 +85,10 @@ type isOperation_Result interface { } type Operation_Error struct { - Error *status.Status `protobuf:"bytes,4,opt,name=error,oneof"` + Error *status.Status `protobuf:"bytes,4,opt,name=error,proto3,oneof"` } type Operation_Response struct { - Response *any.Any `protobuf:"bytes,5,opt,name=response,oneof"` + Response *any.Any `protobuf:"bytes,5,opt,name=response,proto3,oneof"` } func (*Operation_Error) isOperation_Result() {} @@ -213,7 +213,7 @@ func _Operation_OneofSizer(msg proto.Message) (n int) { // The request message for [Operations.GetOperation][google.longrunning.Operations.GetOperation]. type GetOperationRequest struct { // The name of the operation resource. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -253,13 +253,13 @@ func (m *GetOperationRequest) GetName() string { // The request message for [Operations.ListOperations][google.longrunning.Operations.ListOperations]. type ListOperationsRequest struct { // The name of the operation collection. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // The standard list filter. - Filter string `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,1,opt,name=filter,proto3" json:"filter,omitempty"` // The standard list page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The standard list page token. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -320,9 +320,9 @@ func (m *ListOperationsRequest) GetPageToken() string { // The response message for [Operations.ListOperations][google.longrunning.Operations.ListOperations]. type ListOperationsResponse struct { // A list of operations that matches the specified filter in the request. - Operations []*Operation `protobuf:"bytes,1,rep,name=operations" json:"operations,omitempty"` + Operations []*Operation `protobuf:"bytes,1,rep,name=operations,proto3" json:"operations,omitempty"` // The standard List next-page token. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -369,7 +369,7 @@ func (m *ListOperationsResponse) GetNextPageToken() string { // The request message for [Operations.CancelOperation][google.longrunning.Operations.CancelOperation]. type CancelOperationRequest struct { // The name of the operation resource to be cancelled. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -409,7 +409,7 @@ func (m *CancelOperationRequest) GetName() string { // The request message for [Operations.DeleteOperation][google.longrunning.Operations.DeleteOperation]. type DeleteOperationRequest struct { // The name of the operation resource to be deleted. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -539,8 +539,7 @@ func (c *operationsClient) CancelOperation(ctx context.Context, in *CancelOperat return out, nil } -// Server API for Operations service - +// OperationsServer is the server API for Operations service. type OperationsServer interface { // Lists operations that match the specified filter in the request. If the // server doesn't support this method, it returns `UNIMPLEMENTED`. diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert.pb.go index 000e91878..7b320e1bf 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert.pb.go @@ -76,18 +76,18 @@ type AlertPolicy struct { // [alertPolicies.create][google.monitoring.v3.AlertPolicyService.CreateAlertPolicy] // method, do not include the `name` field in the alerting policy passed as // part of the request. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A short name or phrase used to identify the policy in dashboards, // notifications, and incidents. To avoid confusion, don't use the same // display name for multiple policies in the same project. The name is // limited to 512 Unicode characters. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Documentation that is included with notifications and incidents related to // this policy. Best practice is for the documentation to include information // to help responders understand, mitigate, escalate, and correct the // underlying problems detected by the alerting policy. Notification channels // that have limited capacity might not show this documentation. - Documentation *AlertPolicy_Documentation `protobuf:"bytes,13,opt,name=documentation" json:"documentation,omitempty"` + Documentation *AlertPolicy_Documentation `protobuf:"bytes,13,opt,name=documentation,proto3" json:"documentation,omitempty"` // User-supplied key/value data to be used for organizing and // identifying the `AlertPolicy` objects. // @@ -95,21 +95,21 @@ type AlertPolicy struct { // 63 Unicode characters or 128 bytes, whichever is smaller. Labels and // values can contain only lowercase letters, numerals, underscores, and // dashes. Keys must begin with a letter. - UserLabels map[string]string `protobuf:"bytes,16,rep,name=user_labels,json=userLabels" json:"user_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + UserLabels map[string]string `protobuf:"bytes,16,rep,name=user_labels,json=userLabels,proto3" json:"user_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // A list of conditions for the policy. The conditions are combined by AND or // OR according to the `combiner` field. If the combined conditions evaluate // to true, then an incident is created. A policy can have from one to six // conditions. - Conditions []*AlertPolicy_Condition `protobuf:"bytes,12,rep,name=conditions" json:"conditions,omitempty"` + Conditions []*AlertPolicy_Condition `protobuf:"bytes,12,rep,name=conditions,proto3" json:"conditions,omitempty"` // How to combine the results of multiple conditions // to determine if an incident should be opened. - Combiner AlertPolicy_ConditionCombinerType `protobuf:"varint,6,opt,name=combiner,enum=google.monitoring.v3.AlertPolicy_ConditionCombinerType" json:"combiner,omitempty"` + Combiner AlertPolicy_ConditionCombinerType `protobuf:"varint,6,opt,name=combiner,proto3,enum=google.monitoring.v3.AlertPolicy_ConditionCombinerType" json:"combiner,omitempty"` // Whether or not the policy is enabled. On write, the default interpretation // if unset is that the policy is enabled. On read, clients should not make // any assumption about the state if it has not been populated. The // field should always be populated on List and Get operations, unless // a field projection has been specified that strips it out. - Enabled *wrappers.BoolValue `protobuf:"bytes,17,opt,name=enabled" json:"enabled,omitempty"` + Enabled *wrappers.BoolValue `protobuf:"bytes,17,opt,name=enabled,proto3" json:"enabled,omitempty"` // Identifies the notification channels to which notifications should be sent // when incidents are opened or closed or when new violations occur on // an already opened incident. Each element of this array corresponds to @@ -120,13 +120,13 @@ type AlertPolicy struct { // method. The syntax of the entries in this field is: // // projects/[PROJECT_ID]/notificationChannels/[CHANNEL_ID] - NotificationChannels []string `protobuf:"bytes,14,rep,name=notification_channels,json=notificationChannels" json:"notification_channels,omitempty"` + NotificationChannels []string `protobuf:"bytes,14,rep,name=notification_channels,json=notificationChannels,proto3" json:"notification_channels,omitempty"` // A read-only record of the creation of the alerting policy. If provided // in a call to create or update, this field will be ignored. - CreationRecord *MutationRecord `protobuf:"bytes,10,opt,name=creation_record,json=creationRecord" json:"creation_record,omitempty"` + CreationRecord *MutationRecord `protobuf:"bytes,10,opt,name=creation_record,json=creationRecord,proto3" json:"creation_record,omitempty"` // A read-only record of the most recent change to the alerting policy. If // provided in a call to create or update, this field will be ignored. - MutationRecord *MutationRecord `protobuf:"bytes,11,opt,name=mutation_record,json=mutationRecord" json:"mutation_record,omitempty"` + MutationRecord *MutationRecord `protobuf:"bytes,11,opt,name=mutation_record,json=mutationRecord,proto3" json:"mutation_record,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -233,11 +233,11 @@ type AlertPolicy_Documentation struct { // The content may not exceed 8,192 Unicode characters and may not exceed // more than 10,240 bytes when encoded in UTF-8 format, whichever is // smaller. - Content string `protobuf:"bytes,1,opt,name=content" json:"content,omitempty"` + Content string `protobuf:"bytes,1,opt,name=content,proto3" json:"content,omitempty"` // The format of the `content` field. Presently, only the value // `"text/markdown"` is supported. See // [Markdown](https://en.wikipedia.org/wiki/Markdown) for more information. - MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType" json:"mime_type,omitempty"` + MimeType string `protobuf:"bytes,2,opt,name=mime_type,json=mimeType,proto3" json:"mime_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -310,11 +310,11 @@ type AlertPolicy_Condition struct { // changes, such as those to condition thresholds, durations, or trigger // values. Otherwise, treat the change as a new condition and let the // existing condition be deleted. - Name string `protobuf:"bytes,12,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,12,opt,name=name,proto3" json:"name,omitempty"` // A short name or phrase used to identify the condition in dashboards, // notifications, and incidents. To avoid confusion, don't use the same // display name for multiple conditions in the same policy. - DisplayName string `protobuf:"bytes,6,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,6,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Only one of the following condition types will be specified. // // Types that are valid to be assigned to Condition: @@ -355,10 +355,10 @@ type isAlertPolicy_Condition_Condition interface { } type AlertPolicy_Condition_ConditionThreshold struct { - ConditionThreshold *AlertPolicy_Condition_MetricThreshold `protobuf:"bytes,1,opt,name=condition_threshold,json=conditionThreshold,oneof"` + ConditionThreshold *AlertPolicy_Condition_MetricThreshold `protobuf:"bytes,1,opt,name=condition_threshold,json=conditionThreshold,proto3,oneof"` } type AlertPolicy_Condition_ConditionAbsent struct { - ConditionAbsent *AlertPolicy_Condition_MetricAbsence `protobuf:"bytes,2,opt,name=condition_absent,json=conditionAbsent,oneof"` + ConditionAbsent *AlertPolicy_Condition_MetricAbsence `protobuf:"bytes,2,opt,name=condition_absent,json=conditionAbsent,proto3,oneof"` } func (*AlertPolicy_Condition_ConditionThreshold) isAlertPolicy_Condition_Condition() {} @@ -516,10 +516,10 @@ type isAlertPolicy_Condition_Trigger_Type interface { } type AlertPolicy_Condition_Trigger_Count struct { - Count int32 `protobuf:"varint,1,opt,name=count,oneof"` + Count int32 `protobuf:"varint,1,opt,name=count,proto3,oneof"` } type AlertPolicy_Condition_Trigger_Percent struct { - Percent float64 `protobuf:"fixed64,2,opt,name=percent,oneof"` + Percent float64 `protobuf:"fixed64,2,opt,name=percent,proto3,oneof"` } func (*AlertPolicy_Condition_Trigger_Count) isAlertPolicy_Condition_Trigger_Type() {} @@ -623,7 +623,7 @@ type AlertPolicy_Condition_MetricThreshold struct { // processed) and must specify the metric type and optionally may contain // restrictions on resource type, resource labels, and metric labels. // This field may not exceed 2048 Unicode characters in length. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // Specifies the alignment of data points in individual time series as // well as how to combine the retrieved time series together (such as // when aggregating multiple streams on each resource to a single @@ -634,7 +634,7 @@ type AlertPolicy_Condition_MetricThreshold struct { // This field is similar to the one in the // [`MetricService.ListTimeSeries` request](/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list). // It is advisable to use the `ListTimeSeries` method when debugging this field. - Aggregations []*Aggregation `protobuf:"bytes,8,rep,name=aggregations" json:"aggregations,omitempty"` + Aggregations []*Aggregation `protobuf:"bytes,8,rep,name=aggregations,proto3" json:"aggregations,omitempty"` // A [filter](/monitoring/api/v3/filters) that identifies a time // series that should be used as the denominator of a ratio that will be // compared with the threshold. If a `denominator_filter` is specified, @@ -648,7 +648,7 @@ type AlertPolicy_Condition_MetricThreshold struct { // processed) and must specify the metric type and optionally may contain // restrictions on resource type, resource labels, and metric labels. // This field may not exceed 2048 Unicode characters in length. - DenominatorFilter string `protobuf:"bytes,9,opt,name=denominator_filter,json=denominatorFilter" json:"denominator_filter,omitempty"` + DenominatorFilter string `protobuf:"bytes,9,opt,name=denominator_filter,json=denominatorFilter,proto3" json:"denominator_filter,omitempty"` // Specifies the alignment of data points in individual time series // selected by `denominatorFilter` as // well as how to combine the retrieved time series together (such as @@ -665,16 +665,16 @@ type AlertPolicy_Condition_MetricThreshold struct { // request](/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list). It // is advisable to use the `ListTimeSeries` method when debugging this // field. - DenominatorAggregations []*Aggregation `protobuf:"bytes,10,rep,name=denominator_aggregations,json=denominatorAggregations" json:"denominator_aggregations,omitempty"` + DenominatorAggregations []*Aggregation `protobuf:"bytes,10,rep,name=denominator_aggregations,json=denominatorAggregations,proto3" json:"denominator_aggregations,omitempty"` // The comparison to apply between the time series (indicated by `filter` // and `aggregation`) and the threshold (indicated by `threshold_value`). // The comparison is applied on each time series, with the time series // on the left-hand side and the threshold on the right-hand side. // // Only `COMPARISON_LT` and `COMPARISON_GT` are supported currently. - Comparison ComparisonType `protobuf:"varint,4,opt,name=comparison,enum=google.monitoring.v3.ComparisonType" json:"comparison,omitempty"` + Comparison ComparisonType `protobuf:"varint,4,opt,name=comparison,proto3,enum=google.monitoring.v3.ComparisonType" json:"comparison,omitempty"` // A value against which to compare the time series. - ThresholdValue float64 `protobuf:"fixed64,5,opt,name=threshold_value,json=thresholdValue" json:"threshold_value,omitempty"` + ThresholdValue float64 `protobuf:"fixed64,5,opt,name=threshold_value,json=thresholdValue,proto3" json:"threshold_value,omitempty"` // The amount of time that a time series must violate the // threshold to be considered failing. Currently, only values // that are a multiple of a minute--e.g. 60, 120, or 300 @@ -686,14 +686,14 @@ type AlertPolicy_Condition_MetricThreshold struct { // a good duration is long enough so that a single outlier does not // generate spurious alerts, but short enough that unhealthy states // are detected and alerted on quickly. - Duration *duration.Duration `protobuf:"bytes,6,opt,name=duration" json:"duration,omitempty"` + Duration *duration.Duration `protobuf:"bytes,6,opt,name=duration,proto3" json:"duration,omitempty"` // The number/percent of time series for which the comparison must hold // in order for the condition to trigger. If unspecified, then the // condition will trigger if the comparison is true for any of the // time series that have been identified by `filter` and `aggregations`, // or by the ratio, if `denominator_filter` and `denominator_aggregations` // are specified. - Trigger *AlertPolicy_Condition_Trigger `protobuf:"bytes,7,opt,name=trigger" json:"trigger,omitempty"` + Trigger *AlertPolicy_Condition_Trigger `protobuf:"bytes,7,opt,name=trigger,proto3" json:"trigger,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -795,7 +795,7 @@ type AlertPolicy_Condition_MetricAbsence struct { // processed) and must specify the metric type and optionally may contain // restrictions on resource type, resource labels, and metric labels. // This field may not exceed 2048 Unicode characters in length. - Filter string `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,1,opt,name=filter,proto3" json:"filter,omitempty"` // Specifies the alignment of data points in individual time series as // well as how to combine the retrieved time series together (such as // when aggregating multiple streams on each resource to a single @@ -806,19 +806,19 @@ type AlertPolicy_Condition_MetricAbsence struct { // This field is similar to the // one in the [`MetricService.ListTimeSeries` request](/monitoring/api/ref_v3/rest/v3/projects.timeSeries/list). // It is advisable to use the `ListTimeSeries` method when debugging this field. - Aggregations []*Aggregation `protobuf:"bytes,5,rep,name=aggregations" json:"aggregations,omitempty"` + Aggregations []*Aggregation `protobuf:"bytes,5,rep,name=aggregations,proto3" json:"aggregations,omitempty"` // The amount of time that a time series must fail to report new // data to be considered failing. Currently, only values that // are a multiple of a minute--e.g. 60, 120, or 300 // seconds--are supported. If an invalid value is given, an // error will be returned. The `Duration.nanos` field is // ignored. - Duration *duration.Duration `protobuf:"bytes,2,opt,name=duration" json:"duration,omitempty"` + Duration *duration.Duration `protobuf:"bytes,2,opt,name=duration,proto3" json:"duration,omitempty"` // The number/percent of time series for which the comparison must hold // in order for the condition to trigger. If unspecified, then the // condition will trigger if the comparison is true for any of the // time series that have been identified by `filter` and `aggregations`. - Trigger *AlertPolicy_Condition_Trigger `protobuf:"bytes,3,opt,name=trigger" json:"trigger,omitempty"` + Trigger *AlertPolicy_Condition_Trigger `protobuf:"bytes,3,opt,name=trigger,proto3" json:"trigger,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert_service.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert_service.pb.go index 3f65f8c21..e7551b31c 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/alert_service.pb.go @@ -36,11 +36,11 @@ type CreateAlertPolicyRequest struct { // policy that is returned will have a name that contains a normalized // representation of this name as a prefix but adds a suffix of the form // `/alertPolicies/[POLICY_ID]`, identifying the policy in the container. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The requested alerting policy. You should omit the `name` field in this // policy. The name will be returned in the new policy, including // a new [ALERT_POLICY_ID] value. - AlertPolicy *AlertPolicy `protobuf:"bytes,2,opt,name=alert_policy,json=alertPolicy" json:"alert_policy,omitempty"` + AlertPolicy *AlertPolicy `protobuf:"bytes,2,opt,name=alert_policy,json=alertPolicy,proto3" json:"alert_policy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -89,7 +89,7 @@ type GetAlertPolicyRequest struct { // The alerting policy to retrieve. The format is // // projects/[PROJECT_ID]/alertPolicies/[ALERT_POLICY_ID] - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -137,26 +137,26 @@ type ListAlertPoliciesRequest struct { // by name, use the // [GetAlertPolicy][google.monitoring.v3.AlertPolicyService.GetAlertPolicy] // operation, instead. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // If provided, this field specifies the criteria that must be met by // alert policies to be included in the response. // // For more details, see [sorting and // filtering](/monitoring/api/v3/sorting-and-filtering). - Filter string `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // A comma-separated list of fields by which to sort the result. Supports // the same set of field references as the `filter` field. Entries can be // prefixed with a minus sign to sort by the field in descending order. // // For more details, see [sorting and // filtering](/monitoring/api/v3/sorting-and-filtering). - OrderBy string `protobuf:"bytes,6,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy string `protobuf:"bytes,6,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` // The maximum number of results to return in a single response. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return more results from the previous method call. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -224,11 +224,11 @@ func (m *ListAlertPoliciesRequest) GetPageToken() string { // The protocol for the `ListAlertPolicies` response. type ListAlertPoliciesResponse struct { // The returned alert policies. - AlertPolicies []*AlertPolicy `protobuf:"bytes,3,rep,name=alert_policies,json=alertPolicies" json:"alert_policies,omitempty"` + AlertPolicies []*AlertPolicy `protobuf:"bytes,3,rep,name=alert_policies,json=alertPolicies,proto3" json:"alert_policies,omitempty"` // If there might be more results than were returned, then this field is set // to a non-empty value. To see the additional results, // use that value as `pageToken` in the next call to this method. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -295,12 +295,12 @@ type UpdateAlertPolicyRequest struct { // the supplied condition includes the `name` field with that // `[CONDITION_ID]`. If the supplied condition omits the `name` field, // then a new `[CONDITION_ID]` is created. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Required. The updated alerting policy or the updated values for the // fields listed in `update_mask`. // If `update_mask` is not empty, any fields in this policy that are // not in `update_mask` are ignored. - AlertPolicy *AlertPolicy `protobuf:"bytes,3,opt,name=alert_policy,json=alertPolicy" json:"alert_policy,omitempty"` + AlertPolicy *AlertPolicy `protobuf:"bytes,3,opt,name=alert_policy,json=alertPolicy,proto3" json:"alert_policy,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -351,7 +351,7 @@ type DeleteAlertPolicyRequest struct { // projects/[PROJECT_ID]/alertPolicies/[ALERT_POLICY_ID] // // For more information, see [AlertPolicy][google.monitoring.v3.AlertPolicy]. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -477,8 +477,7 @@ func (c *alertPolicyServiceClient) UpdateAlertPolicy(ctx context.Context, in *Up return out, nil } -// Server API for AlertPolicyService service - +// AlertPolicyServiceServer is the server API for AlertPolicyService service. type AlertPolicyServiceServer interface { // Lists the existing alerting policies for the project. ListAlertPolicies(context.Context, *ListAlertPoliciesRequest) (*ListAlertPoliciesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/common.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/common.pb.go index 1daf7f768..d4d0e7528 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/common.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/common.pb.go @@ -457,19 +457,19 @@ type isTypedValue_Value interface { } type TypedValue_BoolValue struct { - BoolValue bool `protobuf:"varint,1,opt,name=bool_value,json=boolValue,oneof"` + BoolValue bool `protobuf:"varint,1,opt,name=bool_value,json=boolValue,proto3,oneof"` } type TypedValue_Int64Value struct { - Int64Value int64 `protobuf:"varint,2,opt,name=int64_value,json=int64Value,oneof"` + Int64Value int64 `protobuf:"varint,2,opt,name=int64_value,json=int64Value,proto3,oneof"` } type TypedValue_DoubleValue struct { - DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,oneof"` + DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,proto3,oneof"` } type TypedValue_StringValue struct { - StringValue string `protobuf:"bytes,4,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,4,opt,name=string_value,json=stringValue,proto3,oneof"` } type TypedValue_DistributionValue struct { - DistributionValue *distribution.Distribution `protobuf:"bytes,5,opt,name=distribution_value,json=distributionValue,oneof"` + DistributionValue *distribution.Distribution `protobuf:"bytes,5,opt,name=distribution_value,json=distributionValue,proto3,oneof"` } func (*TypedValue_BoolValue) isTypedValue_Value() {} @@ -641,11 +641,11 @@ func _TypedValue_OneofSizer(msg proto.Message) (n int) { // represents a single point in time. type TimeInterval struct { // Required. The end of the time interval. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Optional. The beginning of the time interval. The default value // for the start time is the end time. The start time must not be // later than the end time. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -704,7 +704,7 @@ type Aggregation struct { // this field is ignored. If `perSeriesAligner` is specified and // does not equal `ALIGN_NONE`, then this field must be defined; // otherwise an error is returned. - AlignmentPeriod *duration.Duration `protobuf:"bytes,1,opt,name=alignment_period,json=alignmentPeriod" json:"alignment_period,omitempty"` + AlignmentPeriod *duration.Duration `protobuf:"bytes,1,opt,name=alignment_period,json=alignmentPeriod,proto3" json:"alignment_period,omitempty"` // The approach to be used to align individual time series. Not all // alignment functions may be applied to all time series, depending // on the metric type and value type of the original time @@ -716,7 +716,7 @@ type Aggregation struct { // `perSeriesAligner` must be specified and not equal `ALIGN_NONE` // and `alignmentPeriod` must be specified; otherwise, an error is // returned. - PerSeriesAligner Aggregation_Aligner `protobuf:"varint,2,opt,name=per_series_aligner,json=perSeriesAligner,enum=google.monitoring.v3.Aggregation_Aligner" json:"per_series_aligner,omitempty"` + PerSeriesAligner Aggregation_Aligner `protobuf:"varint,2,opt,name=per_series_aligner,json=perSeriesAligner,proto3,enum=google.monitoring.v3.Aggregation_Aligner" json:"per_series_aligner,omitempty"` // The approach to be used to combine time series. Not all reducer // functions may be applied to all time series, depending on the // metric type and the value type of the original time @@ -728,7 +728,7 @@ type Aggregation struct { // `perSeriesAligner` must be specified and not equal `ALIGN_NONE` // and `alignmentPeriod` must be specified; otherwise, an error is // returned. - CrossSeriesReducer Aggregation_Reducer `protobuf:"varint,4,opt,name=cross_series_reducer,json=crossSeriesReducer,enum=google.monitoring.v3.Aggregation_Reducer" json:"cross_series_reducer,omitempty"` + CrossSeriesReducer Aggregation_Reducer `protobuf:"varint,4,opt,name=cross_series_reducer,json=crossSeriesReducer,proto3,enum=google.monitoring.v3.Aggregation_Reducer" json:"cross_series_reducer,omitempty"` // The set of fields to preserve when `crossSeriesReducer` is // specified. The `groupByFields` determine how the time series are // partitioned into subsets prior to applying the aggregation @@ -743,7 +743,7 @@ type Aggregation struct { // the same resource type, then the time series are aggregated into // a single output time series. If `crossSeriesReducer` is not // defined, this field is ignored. - GroupByFields []string `protobuf:"bytes,5,rep,name=group_by_fields,json=groupByFields" json:"group_by_fields,omitempty"` + GroupByFields []string `protobuf:"bytes,5,rep,name=group_by_fields,json=groupByFields,proto3" json:"group_by_fields,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group.pb.go index 064dc52e0..9ae9f756d 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group.pb.go @@ -50,18 +50,18 @@ type Group struct { // When creating a group, this field is ignored and a new name is created // consisting of the project specified in the call to `CreateGroup` // and a unique `{group_id}` that is generated automatically. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A user-assigned name for this group, used only for display purposes. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The name of the group's parent, if it has one. // The format is `"projects/{project_id_or_number}/groups/{group_id}"`. // For groups with no parent, `parentName` is the empty string, `""`. - ParentName string `protobuf:"bytes,3,opt,name=parent_name,json=parentName" json:"parent_name,omitempty"` + ParentName string `protobuf:"bytes,3,opt,name=parent_name,json=parentName,proto3" json:"parent_name,omitempty"` // The filter used to determine which monitored resources belong to this group. - Filter string `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // If true, the members of this group are considered to be a cluster. // The system can perform additional analysis on groups that are clusters. - IsCluster bool `protobuf:"varint,6,opt,name=is_cluster,json=isCluster" json:"is_cluster,omitempty"` + IsCluster bool `protobuf:"varint,6,opt,name=is_cluster,json=isCluster,proto3" json:"is_cluster,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group_service.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group_service.pb.go index 99cd3a30d..57d11175d 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/group_service.pb.go @@ -30,7 +30,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type ListGroupsRequest struct { // The project whose groups are to be listed. The format is // `"projects/{project_id_or_number}"`. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` // An optional filter consisting of a single group name. The filters limit the // groups returned based on their parent-child relationship with the specified // group. If no filter is specified, all groups are returned. @@ -41,11 +41,11 @@ type ListGroupsRequest struct { // *ListGroupsRequest_DescendantsOfGroup Filter isListGroupsRequest_Filter `protobuf_oneof:"filter"` // A positive number that is the maximum number of results to return. - PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,5,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return additional results from the previous method call. - PageToken string `protobuf:"bytes,6,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,6,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -80,13 +80,13 @@ type isListGroupsRequest_Filter interface { } type ListGroupsRequest_ChildrenOfGroup struct { - ChildrenOfGroup string `protobuf:"bytes,2,opt,name=children_of_group,json=childrenOfGroup,oneof"` + ChildrenOfGroup string `protobuf:"bytes,2,opt,name=children_of_group,json=childrenOfGroup,proto3,oneof"` } type ListGroupsRequest_AncestorsOfGroup struct { - AncestorsOfGroup string `protobuf:"bytes,3,opt,name=ancestors_of_group,json=ancestorsOfGroup,oneof"` + AncestorsOfGroup string `protobuf:"bytes,3,opt,name=ancestors_of_group,json=ancestorsOfGroup,proto3,oneof"` } type ListGroupsRequest_DescendantsOfGroup struct { - DescendantsOfGroup string `protobuf:"bytes,4,opt,name=descendants_of_group,json=descendantsOfGroup,oneof"` + DescendantsOfGroup string `protobuf:"bytes,4,opt,name=descendants_of_group,json=descendantsOfGroup,proto3,oneof"` } func (*ListGroupsRequest_ChildrenOfGroup) isListGroupsRequest_Filter() {} @@ -226,11 +226,11 @@ func _ListGroupsRequest_OneofSizer(msg proto.Message) (n int) { // The `ListGroups` response. type ListGroupsResponse struct { // The groups that match the specified filters. - Group []*Group `protobuf:"bytes,1,rep,name=group" json:"group,omitempty"` + Group []*Group `protobuf:"bytes,1,rep,name=group,proto3" json:"group,omitempty"` // If there are more results than have been returned, then this field is set // to a non-empty value. To see the additional results, // use that value as `pageToken` in the next call to this method. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -278,7 +278,7 @@ func (m *ListGroupsResponse) GetNextPageToken() string { type GetGroupRequest struct { // The group to retrieve. The format is // `"projects/{project_id_or_number}/groups/{group_id}"`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -319,12 +319,12 @@ func (m *GetGroupRequest) GetName() string { type CreateGroupRequest struct { // The project in which to create the group. The format is // `"projects/{project_id_or_number}"`. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // A group definition. It is an error to define the `name` field because // the system assigns the name. - Group *Group `protobuf:"bytes,2,opt,name=group" json:"group,omitempty"` + Group *Group `protobuf:"bytes,2,opt,name=group,proto3" json:"group,omitempty"` // If true, validate this request but do not create the group. - ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly" json:"validate_only,omitempty"` + ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly,proto3" json:"validate_only,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -379,9 +379,9 @@ func (m *CreateGroupRequest) GetValidateOnly() bool { type UpdateGroupRequest struct { // The new definition of the group. All fields of the existing group, // excepting `name`, are replaced with the corresponding fields of this group. - Group *Group `protobuf:"bytes,2,opt,name=group" json:"group,omitempty"` + Group *Group `protobuf:"bytes,2,opt,name=group,proto3" json:"group,omitempty"` // If true, validate this request but do not update the existing group. - ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly" json:"validate_only,omitempty"` + ValidateOnly bool `protobuf:"varint,3,opt,name=validate_only,json=validateOnly,proto3" json:"validate_only,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -429,7 +429,7 @@ func (m *UpdateGroupRequest) GetValidateOnly() bool { type DeleteGroupRequest struct { // The group to delete. The format is // `"projects/{project_id_or_number}/groups/{group_id}"`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -470,13 +470,13 @@ func (m *DeleteGroupRequest) GetName() string { type ListGroupMembersRequest struct { // The group whose members are listed. The format is // `"projects/{project_id_or_number}/groups/{group_id}"`. - Name string `protobuf:"bytes,7,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,7,opt,name=name,proto3" json:"name,omitempty"` // A positive number that is the maximum number of results to return. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return additional results from the previous method call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // An optional [list filter](/monitoring/api/learn_more#filtering) describing // the members to be returned. The filter may reference the type, labels, and // metadata of monitored resources that comprise the group. @@ -484,12 +484,12 @@ type ListGroupMembersRequest struct { // instances, use this filter: // // resource.type = "gce_instance" - Filter string `protobuf:"bytes,5,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,5,opt,name=filter,proto3" json:"filter,omitempty"` // An optional time interval for which results should be returned. Only // members that were part of the group during the specified interval are // included in the response. If no interval is provided then the group // membership over the last minute is returned. - Interval *TimeInterval `protobuf:"bytes,6,opt,name=interval" json:"interval,omitempty"` + Interval *TimeInterval `protobuf:"bytes,6,opt,name=interval,proto3" json:"interval,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -557,13 +557,13 @@ func (m *ListGroupMembersRequest) GetInterval() *TimeInterval { // The `ListGroupMembers` response. type ListGroupMembersResponse struct { // A set of monitored resources in the group. - Members []*monitoredres.MonitoredResource `protobuf:"bytes,1,rep,name=members" json:"members,omitempty"` + Members []*monitoredres.MonitoredResource `protobuf:"bytes,1,rep,name=members,proto3" json:"members,omitempty"` // If there are more results than have been returned, then this field is // set to a non-empty value. To see the additional results, use that value as // `pageToken` in the next call to this method. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` // The total number of elements matching this request. - TotalSize int32 `protobuf:"varint,3,opt,name=total_size,json=totalSize" json:"total_size,omitempty"` + TotalSize int32 `protobuf:"varint,3,opt,name=total_size,json=totalSize,proto3" json:"total_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -714,8 +714,7 @@ func (c *groupServiceClient) ListGroupMembers(ctx context.Context, in *ListGroup return out, nil } -// Server API for GroupService service - +// GroupServiceServer is the server API for GroupService service. type GroupServiceServer interface { // Lists the existing groups. ListGroups(context.Context, *ListGroupsRequest) (*ListGroupsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric.pb.go index e9d726fcf..b448a2ceb 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric.pb.go @@ -30,9 +30,9 @@ type Point struct { // points specifying the same start time and increasing end times, until an // event resets the cumulative value to zero and sets a new start time for the // following points. - Interval *TimeInterval `protobuf:"bytes,1,opt,name=interval" json:"interval,omitempty"` + Interval *TimeInterval `protobuf:"bytes,1,opt,name=interval,proto3" json:"interval,omitempty"` // The value of the data point. - Value *TypedValue `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value *TypedValue `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -83,10 +83,10 @@ func (m *Point) GetValue() *TypedValue { type TimeSeries struct { // The associated metric. A fully-specified metric used to identify the time // series. - Metric *metric.Metric `protobuf:"bytes,1,opt,name=metric" json:"metric,omitempty"` + Metric *metric.Metric `protobuf:"bytes,1,opt,name=metric,proto3" json:"metric,omitempty"` // The associated resource. A fully-specified monitored resource used to // identify the time series. - Resource *monitoredres.MonitoredResource `protobuf:"bytes,2,opt,name=resource" json:"resource,omitempty"` + Resource *monitoredres.MonitoredResource `protobuf:"bytes,2,opt,name=resource,proto3" json:"resource,omitempty"` // The metric kind of the time series. When listing time series, this metric // kind might be different from the metric kind of the associated metric if // this time series is an alignment or reduction of other time series. @@ -96,14 +96,14 @@ type TimeSeries struct { // metric's descriptor must be auto-created, then this field specifies the // metric kind of the new descriptor and must be either `GAUGE` (the default) // or `CUMULATIVE`. - MetricKind metric.MetricDescriptor_MetricKind `protobuf:"varint,3,opt,name=metric_kind,json=metricKind,enum=google.api.MetricDescriptor_MetricKind" json:"metric_kind,omitempty"` + MetricKind metric.MetricDescriptor_MetricKind `protobuf:"varint,3,opt,name=metric_kind,json=metricKind,proto3,enum=google.api.MetricDescriptor_MetricKind" json:"metric_kind,omitempty"` // The value type of the time series. When listing time series, this value // type might be different from the value type of the associated metric if // this time series is an alignment or reduction of other time series. // // When creating a time series, this field is optional. If present, it must be // the same as the type of the data in the `points` field. - ValueType metric.MetricDescriptor_ValueType `protobuf:"varint,4,opt,name=value_type,json=valueType,enum=google.api.MetricDescriptor_ValueType" json:"value_type,omitempty"` + ValueType metric.MetricDescriptor_ValueType `protobuf:"varint,4,opt,name=value_type,json=valueType,proto3,enum=google.api.MetricDescriptor_ValueType" json:"value_type,omitempty"` // The data points of this time series. When listing time series, the order of // the points is specified by the list method. // @@ -112,7 +112,7 @@ type TimeSeries struct { // metric. If the associated metric's descriptor must be auto-created, then // the value type of the descriptor is determined by the point's type, which // must be `BOOL`, `INT64`, `DOUBLE`, or `DISTRIBUTION`. - Points []*Point `protobuf:"bytes,5,rep,name=points" json:"points,omitempty"` + Points []*Point `protobuf:"bytes,5,rep,name=points,proto3" json:"points,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric_service.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric_service.pb.go index 881d92971..8a538ab46 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/metric_service.pb.go @@ -60,7 +60,7 @@ func (ListTimeSeriesRequest_TimeSeriesView) EnumDescriptor() ([]byte, []int) { type ListMonitoredResourceDescriptorsRequest struct { // The project on which to execute the request. The format is // `"projects/{project_id_or_number}"`. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` // An optional [filter](/monitoring/api/v3/filters) describing // the descriptors to be returned. The filter can reference // the descriptor's type and labels. For example, the @@ -68,13 +68,13 @@ type ListMonitoredResourceDescriptorsRequest struct { // that have an `id` label: // // resource.type = starts_with("gce_") AND resource.label:id - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // A positive number that is the maximum number of results to return. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return additional results from the previous method call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -138,11 +138,11 @@ func (m *ListMonitoredResourceDescriptorsRequest) GetPageToken() string { type ListMonitoredResourceDescriptorsResponse struct { // The monitored resource descriptors that are available to this project // and that match `filter`, if present. - ResourceDescriptors []*monitoredres.MonitoredResourceDescriptor `protobuf:"bytes,1,rep,name=resource_descriptors,json=resourceDescriptors" json:"resource_descriptors,omitempty"` + ResourceDescriptors []*monitoredres.MonitoredResourceDescriptor `protobuf:"bytes,1,rep,name=resource_descriptors,json=resourceDescriptors,proto3" json:"resource_descriptors,omitempty"` // If there are more results than have been returned, then this field is set // to a non-empty value. To see the additional results, // use that value as `pageToken` in the next call to this method. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -194,7 +194,7 @@ type GetMonitoredResourceDescriptorRequest struct { // `"projects/{project_id_or_number}/monitoredResourceDescriptors/{resource_type}"`. // The `{resource_type}` is a predefined type, such as // `cloudsql_database`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -235,7 +235,7 @@ func (m *GetMonitoredResourceDescriptorRequest) GetName() string { type ListMetricDescriptorsRequest struct { // The project on which to execute the request. The format is // `"projects/{project_id_or_number}"`. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` // If this field is empty, all custom and // system-defined metric descriptors are returned. // Otherwise, the [filter](/monitoring/api/v3/filters) @@ -244,13 +244,13 @@ type ListMetricDescriptorsRequest struct { // [custom metrics](/monitoring/custom-metrics): // // metric.type = starts_with("custom.googleapis.com/") - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // A positive number that is the maximum number of results to return. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return additional results from the previous method call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -312,11 +312,11 @@ func (m *ListMetricDescriptorsRequest) GetPageToken() string { type ListMetricDescriptorsResponse struct { // The metric descriptors that are available to the project // and that match the value of `filter`, if present. - MetricDescriptors []*metric.MetricDescriptor `protobuf:"bytes,1,rep,name=metric_descriptors,json=metricDescriptors" json:"metric_descriptors,omitempty"` + MetricDescriptors []*metric.MetricDescriptor `protobuf:"bytes,1,rep,name=metric_descriptors,json=metricDescriptors,proto3" json:"metric_descriptors,omitempty"` // If there are more results than have been returned, then this field is set // to a non-empty value. To see the additional results, // use that value as `pageToken` in the next call to this method. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -366,7 +366,7 @@ type GetMetricDescriptorRequest struct { // `"projects/{project_id_or_number}/metricDescriptors/{metric_id}"`. // An example value of `{metric_id}` is // `"compute.googleapis.com/instance/disk/read_bytes_count"`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -407,10 +407,10 @@ func (m *GetMetricDescriptorRequest) GetName() string { type CreateMetricDescriptorRequest struct { // The project on which to execute the request. The format is // `"projects/{project_id_or_number}"`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The new [custom metric](/monitoring/custom-metrics) // descriptor. - MetricDescriptor *metric.MetricDescriptor `protobuf:"bytes,2,opt,name=metric_descriptor,json=metricDescriptor" json:"metric_descriptor,omitempty"` + MetricDescriptor *metric.MetricDescriptor `protobuf:"bytes,2,opt,name=metric_descriptor,json=metricDescriptor,proto3" json:"metric_descriptor,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -460,7 +460,7 @@ type DeleteMetricDescriptorRequest struct { // `"projects/{project_id_or_number}/metricDescriptors/{metric_id}"`. // An example of `{metric_id}` is: // `"custom.googleapis.com/my_test_metric"`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -501,7 +501,7 @@ func (m *DeleteMetricDescriptorRequest) GetName() string { type ListTimeSeriesRequest struct { // The project on which to execute the request. The format is // "projects/{project_id_or_number}". - Name string `protobuf:"bytes,10,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,10,opt,name=name,proto3" json:"name,omitempty"` // A [monitoring filter](/monitoring/api/v3/filters) that specifies which time // series should be returned. The filter must specify a single metric type, // and can additionally specify metric labels and other information. For @@ -509,30 +509,30 @@ type ListTimeSeriesRequest struct { // // metric.type = "compute.googleapis.com/instance/cpu/usage_time" AND // metric.label.instance_name = "my-instance-name" - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` // The time interval for which results should be returned. Only time series // that contain data points in the specified interval are included // in the response. - Interval *TimeInterval `protobuf:"bytes,4,opt,name=interval" json:"interval,omitempty"` + Interval *TimeInterval `protobuf:"bytes,4,opt,name=interval,proto3" json:"interval,omitempty"` // By default, the raw time series data is returned. // Use this field to combine multiple time series for different // views of the data. - Aggregation *Aggregation `protobuf:"bytes,5,opt,name=aggregation" json:"aggregation,omitempty"` + Aggregation *Aggregation `protobuf:"bytes,5,opt,name=aggregation,proto3" json:"aggregation,omitempty"` // Specifies the order in which the points of the time series should // be returned. By default, results are not ordered. Currently, // this field must be left blank. - OrderBy string `protobuf:"bytes,6,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy string `protobuf:"bytes,6,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` // Specifies which information is returned about the time series. - View ListTimeSeriesRequest_TimeSeriesView `protobuf:"varint,7,opt,name=view,enum=google.monitoring.v3.ListTimeSeriesRequest_TimeSeriesView" json:"view,omitempty"` + View ListTimeSeriesRequest_TimeSeriesView `protobuf:"varint,7,opt,name=view,proto3,enum=google.monitoring.v3.ListTimeSeriesRequest_TimeSeriesView" json:"view,omitempty"` // A positive number that is the maximum number of results to return. // When `view` field sets to `FULL`, it limits the number of `Points` server // will return; if `view` field is `HEADERS`, it limits the number of // `TimeSeries` server will return. - PageSize int32 `protobuf:"varint,8,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,8,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return additional results from the previous method call. - PageToken string `protobuf:"bytes,9,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,9,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -621,11 +621,11 @@ func (m *ListTimeSeriesRequest) GetPageToken() string { // The `ListTimeSeries` response. type ListTimeSeriesResponse struct { // One or more time series that match the filter included in the request. - TimeSeries []*TimeSeries `protobuf:"bytes,1,rep,name=time_series,json=timeSeries" json:"time_series,omitempty"` + TimeSeries []*TimeSeries `protobuf:"bytes,1,rep,name=time_series,json=timeSeries,proto3" json:"time_series,omitempty"` // If there are more results than have been returned, then this field is set // to a non-empty value. To see the additional results, // use that value as `pageToken` in the next call to this method. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -673,13 +673,13 @@ func (m *ListTimeSeriesResponse) GetNextPageToken() string { type CreateTimeSeriesRequest struct { // The project on which to execute the request. The format is // `"projects/{project_id_or_number}"`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The new data to be added to a list of time series. // Adds at most one data point to each of several time series. The new data // point must be more recent than any other point in its time series. Each // `TimeSeries` value must fully specify a unique time series by supplying // all label values for the metric and the monitored resource. - TimeSeries []*TimeSeries `protobuf:"bytes,2,rep,name=time_series,json=timeSeries" json:"time_series,omitempty"` + TimeSeries []*TimeSeries `protobuf:"bytes,2,rep,name=time_series,json=timeSeries,proto3" json:"time_series,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -729,9 +729,9 @@ type CreateTimeSeriesError struct { // and `Point`s (including timestamp and value) that resulted // in the error. This field provides all of the context that // would be needed to retry the operation. - TimeSeries *TimeSeries `protobuf:"bytes,1,opt,name=time_series,json=timeSeries" json:"time_series,omitempty"` + TimeSeries *TimeSeries `protobuf:"bytes,1,opt,name=time_series,json=timeSeries,proto3" json:"time_series,omitempty"` // The status of the requested write operation. - Status *status.Status `protobuf:"bytes,2,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,2,opt,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -907,8 +907,7 @@ func (c *metricServiceClient) CreateTimeSeries(ctx context.Context, in *CreateTi return out, nil } -// Server API for MetricService service - +// MetricServiceServer is the server API for MetricService service. type MetricServiceServer interface { // Lists monitored resource descriptors that match a filter. This method does not require a Stackdriver account. ListMonitoredResourceDescriptors(context.Context, *ListMonitoredResourceDescriptorsRequest) (*ListMonitoredResourceDescriptorsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/mutation_record.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/mutation_record.pb.go index d2952f124..f59cba8af 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/mutation_record.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/mutation_record.pb.go @@ -22,9 +22,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Describes a change made to a configuration. type MutationRecord struct { // When the change occurred. - MutateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=mutate_time,json=mutateTime" json:"mutate_time,omitempty"` + MutateTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=mutate_time,json=mutateTime,proto3" json:"mutate_time,omitempty"` // The email address of the user making the change. - MutatedBy string `protobuf:"bytes,2,opt,name=mutated_by,json=mutatedBy" json:"mutated_by,omitempty"` + MutatedBy string `protobuf:"bytes,2,opt,name=mutated_by,json=mutatedBy,proto3" json:"mutated_by,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification.pb.go index d13bb8dc8..8895769ad 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification.pb.go @@ -72,24 +72,24 @@ type NotificationChannelDescriptor struct { // projects/[PROJECT_ID]/notificationChannelDescriptors/[TYPE] // // In the above, `[TYPE]` is the value of the `type` field. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` // The type of notification channel, such as "email", "sms", etc. // Notification channel types are globally unique. - Type string `protobuf:"bytes,1,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` // A human-readable name for the notification channel type. This // form of the name is suitable for a user interface. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // A human-readable description of the notification channel // type. The description may include a description of the properties // of the channel and pointers to external documentation. - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The set of labels that must be defined to identify a particular // channel of the corresponding type. Each label includes a // description for how that field should be populated. - Labels []*label.LabelDescriptor `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty"` + Labels []*label.LabelDescriptor `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty"` // The tiers that support this notification channel; the project service tier // must be one of the supported_tiers. - SupportedTiers []ServiceTier `protobuf:"varint,5,rep,packed,name=supported_tiers,json=supportedTiers,enum=google.monitoring.v3.ServiceTier" json:"supported_tiers,omitempty"` + SupportedTiers []ServiceTier `protobuf:"varint,5,rep,packed,name=supported_tiers,json=supportedTiers,proto3,enum=google.monitoring.v3.ServiceTier" json:"supported_tiers,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,27 +169,27 @@ func (m *NotificationChannelDescriptor) GetSupportedTiers() []ServiceTier { type NotificationChannel struct { // The type of the notification channel. This field matches the // value of the [NotificationChannelDescriptor.type][google.monitoring.v3.NotificationChannelDescriptor.type] field. - Type string `protobuf:"bytes,1,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` // The full REST resource name for this channel. The syntax is: // // projects/[PROJECT_ID]/notificationChannels/[CHANNEL_ID] // // The `[CHANNEL_ID]` is automatically assigned by the server on creation. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` // An optional human-readable name for this notification channel. It is // recommended that you specify a non-empty and unique name in order to // make it easier to identify the channels in your project, though this is // not enforced. The display name is limited to 512 Unicode characters. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // An optional human-readable description of this notification channel. This // description may provide additional details, beyond the display // name, for the channel. This may not exceeed 1024 Unicode characters. - Description string `protobuf:"bytes,4,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` // Configuration fields that define the channel and its behavior. The // permissible and required labels are specified in the // [NotificationChannelDescriptor.labels][google.monitoring.v3.NotificationChannelDescriptor.labels] of the // `NotificationChannelDescriptor` corresponding to the `type` field. - Labels map[string]string `protobuf:"bytes,5,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,5,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // User-supplied key/value data that does not need to conform to // the corresponding `NotificationChannelDescriptor`'s schema, unlike // the `labels` field. This field is intended to be used for organizing @@ -199,7 +199,7 @@ type NotificationChannel struct { // 63 Unicode characters or 128 bytes, whichever is smaller. Labels and // values can contain only lowercase letters, numerals, underscores, and // dashes. Keys must begin with a letter. - UserLabels map[string]string `protobuf:"bytes,8,rep,name=user_labels,json=userLabels" json:"user_labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + UserLabels map[string]string `protobuf:"bytes,8,rep,name=user_labels,json=userLabels,proto3" json:"user_labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Indicates whether this channel has been verified or not. On a // [`ListNotificationChannels`][google.monitoring.v3.NotificationChannelService.ListNotificationChannels] // or @@ -219,14 +219,14 @@ type NotificationChannel struct { // [`UpdateNotificationChannel`][google.monitoring.v3.NotificationChannelService.UpdateNotificationChannel] // operation. To change the value of this field, you must call // [`VerifyNotificationChannel`][google.monitoring.v3.NotificationChannelService.VerifyNotificationChannel]. - VerificationStatus NotificationChannel_VerificationStatus `protobuf:"varint,9,opt,name=verification_status,json=verificationStatus,enum=google.monitoring.v3.NotificationChannel_VerificationStatus" json:"verification_status,omitempty"` + VerificationStatus NotificationChannel_VerificationStatus `protobuf:"varint,9,opt,name=verification_status,json=verificationStatus,proto3,enum=google.monitoring.v3.NotificationChannel_VerificationStatus" json:"verification_status,omitempty"` // Whether notifications are forwarded to the described channel. This makes // it possible to disable delivery of notifications to a particular channel // without removing the channel from all alerting policies that reference // the channel. This is a more convenient approach when the change is // temporary and you want to receive notifications from the same set // of alerting policies on the channel at some point in the future. - Enabled *wrappers.BoolValue `protobuf:"bytes,11,opt,name=enabled" json:"enabled,omitempty"` + Enabled *wrappers.BoolValue `protobuf:"bytes,11,opt,name=enabled,proto3" json:"enabled,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification_service.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification_service.pb.go index 6d333a161..7a14f9aca 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/notification_service.pb.go @@ -38,15 +38,15 @@ type ListNotificationChannelDescriptorsRequest struct { // descriptors; to retrieve a single descriptor by name, use the // [GetNotificationChannelDescriptor][google.monitoring.v3.NotificationChannelService.GetNotificationChannelDescriptor] // operation, instead. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // The maximum number of results to return in a single response. If // not set to a positive number, a reasonable value will be chosen by the // service. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If non-empty, `page_token` must contain a value returned as the // `next_page_token` in a previous response to request the next set // of results. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -103,12 +103,12 @@ func (m *ListNotificationChannelDescriptorsRequest) GetPageToken() string { type ListNotificationChannelDescriptorsResponse struct { // The monitored resource descriptors supported for the specified // project, optionally filtered. - ChannelDescriptors []*NotificationChannelDescriptor `protobuf:"bytes,1,rep,name=channel_descriptors,json=channelDescriptors" json:"channel_descriptors,omitempty"` + ChannelDescriptors []*NotificationChannelDescriptor `protobuf:"bytes,1,rep,name=channel_descriptors,json=channelDescriptors,proto3" json:"channel_descriptors,omitempty"` // If not empty, indicates that there may be more results that match // the request. Use the value in the `page_token` field in a // subsequent request to fetch the next set of results. If empty, // all results have been returned. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -160,7 +160,7 @@ func (m *ListNotificationChannelDescriptorsResponse) GetNextPageToken() string { type GetNotificationChannelDescriptorRequest struct { // The channel type for which to execute the request. The format is // `projects/[PROJECT_ID]/notificationChannelDescriptors/{channel_type}`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -209,9 +209,9 @@ type CreateNotificationChannelRequest struct { // written. This does not name the newly created channel. The resulting // channel's name will have a normalized version of this field as a prefix, // but will add `/notificationChannels/[CHANNEL_ID]` to identify the channel. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // The definition of the `NotificationChannel` to create. - NotificationChannel *NotificationChannel `protobuf:"bytes,2,opt,name=notification_channel,json=notificationChannel" json:"notification_channel,omitempty"` + NotificationChannel *NotificationChannel `protobuf:"bytes,2,opt,name=notification_channel,json=notificationChannel,proto3" json:"notification_channel,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -263,28 +263,28 @@ type ListNotificationChannelsRequest struct { // specific channel. To query a specific channel by REST resource name, use // the // [`GetNotificationChannel`][google.monitoring.v3.NotificationChannelService.GetNotificationChannel] operation. - Name string `protobuf:"bytes,5,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,5,opt,name=name,proto3" json:"name,omitempty"` // If provided, this field specifies the criteria that must be met by // notification channels to be included in the response. // // For more details, see [sorting and // filtering](/monitoring/api/v3/sorting-and-filtering). - Filter string `protobuf:"bytes,6,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,6,opt,name=filter,proto3" json:"filter,omitempty"` // A comma-separated list of fields by which to sort the result. Supports // the same set of fields as in `filter`. Entries can be prefixed with // a minus sign to sort in descending rather than ascending order. // // For more details, see [sorting and // filtering](/monitoring/api/v3/sorting-and-filtering). - OrderBy string `protobuf:"bytes,7,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + OrderBy string `protobuf:"bytes,7,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` // The maximum number of results to return in a single response. If // not set to a positive number, a reasonable value will be chosen by the // service. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If non-empty, `page_token` must contain a value returned as the // `next_page_token` in a previous response to request the next set // of results. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -352,12 +352,12 @@ func (m *ListNotificationChannelsRequest) GetPageToken() string { // The `ListNotificationChannels` response. type ListNotificationChannelsResponse struct { // The notification channels defined for the specified project. - NotificationChannels []*NotificationChannel `protobuf:"bytes,3,rep,name=notification_channels,json=notificationChannels" json:"notification_channels,omitempty"` + NotificationChannels []*NotificationChannel `protobuf:"bytes,3,rep,name=notification_channels,json=notificationChannels,proto3" json:"notification_channels,omitempty"` // If not empty, indicates that there may be more results that match // the request. Use the value in the `page_token` field in a // subsequent request to fetch the next set of results. If empty, // all results have been returned. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -405,7 +405,7 @@ func (m *ListNotificationChannelsResponse) GetNextPageToken() string { type GetNotificationChannelRequest struct { // The channel for which to execute the request. The format is // `projects/[PROJECT_ID]/notificationChannels/[CHANNEL_ID]`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -445,12 +445,12 @@ func (m *GetNotificationChannelRequest) GetName() string { // The `UpdateNotificationChannel` request. type UpdateNotificationChannelRequest struct { // The fields to update. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // A description of the changes to be applied to the specified // notification channel. The description must provide a definition for // fields to be updated; the names of these fields should also be // included in the `update_mask`. - NotificationChannel *NotificationChannel `protobuf:"bytes,3,opt,name=notification_channel,json=notificationChannel" json:"notification_channel,omitempty"` + NotificationChannel *NotificationChannel `protobuf:"bytes,3,opt,name=notification_channel,json=notificationChannel,proto3" json:"notification_channel,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -498,12 +498,12 @@ func (m *UpdateNotificationChannelRequest) GetNotificationChannel() *Notificatio type DeleteNotificationChannelRequest struct { // The channel for which to execute the request. The format is // `projects/[PROJECT_ID]/notificationChannels/[CHANNEL_ID]`. - Name string `protobuf:"bytes,3,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,3,opt,name=name,proto3" json:"name,omitempty"` // If true, the notification channel will be deleted regardless of its // use in alert policies (the policies will be updated to remove the // channel). If false, channels that are still referenced by an existing // alerting policy will fail to be deleted in a delete operation. - Force bool `protobuf:"varint,5,opt,name=force" json:"force,omitempty"` + Force bool `protobuf:"varint,5,opt,name=force,proto3" json:"force,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -550,7 +550,7 @@ func (m *DeleteNotificationChannelRequest) GetForce() bool { // The `SendNotificationChannelVerificationCode` request. type SendNotificationChannelVerificationCodeRequest struct { // The notification channel to which to send a verification code. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -596,7 +596,7 @@ type GetNotificationChannelVerificationCodeRequest struct { // The notification channel for which a verification code is to be generated // and retrieved. This must name a channel that is already verified; if // the specified channel is not verified, the request will fail. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The desired expiration time. If specified, the API will guarantee that // the returned code will not be valid after the specified timestamp; // however, the API cannot guarantee that the returned code will be @@ -606,7 +606,7 @@ type GetNotificationChannelVerificationCodeRequest struct { // permissible expiration (so specifying an expiration may extend the // code's lifetime over omitting an expiration, even though the API does // impose an upper limit on the maximum expiration that is permitted). - ExpireTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -660,11 +660,11 @@ type GetNotificationChannelVerificationCodeResponse struct { // that have an equivalent identity (i.e. other channels of the same // type with the same fingerprint such as other email channels with // the same email address or other sms channels with the same number). - Code string `protobuf:"bytes,1,opt,name=code" json:"code,omitempty"` + Code string `protobuf:"bytes,1,opt,name=code,proto3" json:"code,omitempty"` // The expiration time associated with the code that was returned. If // an expiration was provided in the request, this is the minimum of the // requested expiration in the request and the max permitted expiration. - ExpireTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + ExpireTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -715,7 +715,7 @@ func (m *GetNotificationChannelVerificationCodeResponse) GetExpireTime() *timest // The `VerifyNotificationChannel` request. type VerifyNotificationChannelRequest struct { // The notification channel to verify. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The verification code that was delivered to the channel as // a result of invoking the `SendNotificationChannelVerificationCode` API // method or that was retrieved from a verified channel via @@ -723,7 +723,7 @@ type VerifyNotificationChannelRequest struct { // "G-123456" or "TKNZGhhd2EyN3I1MnRnMjRv" (in general, one is only // guaranteed that the code is valid UTF-8; one should not // make any assumptions regarding the structure or format of the code). - Code string `protobuf:"bytes,2,opt,name=code" json:"code,omitempty"` + Code string `protobuf:"bytes,2,opt,name=code,proto3" json:"code,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -946,8 +946,7 @@ func (c *notificationChannelServiceClient) VerifyNotificationChannel(ctx context return out, nil } -// Server API for NotificationChannelService service - +// NotificationChannelServiceServer is the server API for NotificationChannelService service. type NotificationChannelServiceServer interface { // Lists the descriptors for supported channel types. The use of descriptors // makes it possible for new channel types to be dynamically added. diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime.pb.go index cb85cbe05..cb1cf7ee4 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime.pb.go @@ -104,11 +104,11 @@ type UptimeCheckConfig struct { // This field should be omitted when creating the uptime check configuration; // on create, the resource name is assigned by the server and included in the // response. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A human-friendly name for the uptime check configuration. The display name // should be unique within a Stackdriver Account in order to make it easier // to identify; however, uniqueness is not enforced. Required. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // The resource the check is checking. Required. // // Types that are valid to be assigned to Resource: @@ -123,24 +123,24 @@ type UptimeCheckConfig struct { CheckRequestType isUptimeCheckConfig_CheckRequestType `protobuf_oneof:"check_request_type"` // How often the uptime check is performed. // Currently, only 1, 5, 10, and 15 minutes are supported. Required. - Period *duration.Duration `protobuf:"bytes,7,opt,name=period" json:"period,omitempty"` + Period *duration.Duration `protobuf:"bytes,7,opt,name=period,proto3" json:"period,omitempty"` // The maximum amount of time to wait for the request to complete (must be // between 1 and 60 seconds). Required. - Timeout *duration.Duration `protobuf:"bytes,8,opt,name=timeout" json:"timeout,omitempty"` + Timeout *duration.Duration `protobuf:"bytes,8,opt,name=timeout,proto3" json:"timeout,omitempty"` // The expected content on the page the check is run against. // Currently, only the first entry in the list is supported, and other entries // will be ignored. The server will look for an exact match of the string in // the page response's content. This field is optional and should only be // specified if a content match is required. - ContentMatchers []*UptimeCheckConfig_ContentMatcher `protobuf:"bytes,9,rep,name=content_matchers,json=contentMatchers" json:"content_matchers,omitempty"` + ContentMatchers []*UptimeCheckConfig_ContentMatcher `protobuf:"bytes,9,rep,name=content_matchers,json=contentMatchers,proto3" json:"content_matchers,omitempty"` // The list of regions from which the check will be run. // If this field is specified, enough regions to include a minimum of // 3 locations must be provided, or an error message is returned. // Not specifying this field will result in uptime checks running from all // regions. - SelectedRegions []UptimeCheckRegion `protobuf:"varint,10,rep,packed,name=selected_regions,json=selectedRegions,enum=google.monitoring.v3.UptimeCheckRegion" json:"selected_regions,omitempty"` + SelectedRegions []UptimeCheckRegion `protobuf:"varint,10,rep,packed,name=selected_regions,json=selectedRegions,proto3,enum=google.monitoring.v3.UptimeCheckRegion" json:"selected_regions,omitempty"` // The internal checkers that this check will egress from. - InternalCheckers []*UptimeCheckConfig_InternalChecker `protobuf:"bytes,14,rep,name=internal_checkers,json=internalCheckers" json:"internal_checkers,omitempty"` + InternalCheckers []*UptimeCheckConfig_InternalChecker `protobuf:"bytes,14,rep,name=internal_checkers,json=internalCheckers,proto3" json:"internal_checkers,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -178,16 +178,16 @@ type isUptimeCheckConfig_CheckRequestType interface { } type UptimeCheckConfig_MonitoredResource struct { - MonitoredResource *monitoredres.MonitoredResource `protobuf:"bytes,3,opt,name=monitored_resource,json=monitoredResource,oneof"` + MonitoredResource *monitoredres.MonitoredResource `protobuf:"bytes,3,opt,name=monitored_resource,json=monitoredResource,proto3,oneof"` } type UptimeCheckConfig_ResourceGroup_ struct { - ResourceGroup *UptimeCheckConfig_ResourceGroup `protobuf:"bytes,4,opt,name=resource_group,json=resourceGroup,oneof"` + ResourceGroup *UptimeCheckConfig_ResourceGroup `protobuf:"bytes,4,opt,name=resource_group,json=resourceGroup,proto3,oneof"` } type UptimeCheckConfig_HttpCheck_ struct { - HttpCheck *UptimeCheckConfig_HttpCheck `protobuf:"bytes,5,opt,name=http_check,json=httpCheck,oneof"` + HttpCheck *UptimeCheckConfig_HttpCheck `protobuf:"bytes,5,opt,name=http_check,json=httpCheck,proto3,oneof"` } type UptimeCheckConfig_TcpCheck_ struct { - TcpCheck *UptimeCheckConfig_TcpCheck `protobuf:"bytes,6,opt,name=tcp_check,json=tcpCheck,oneof"` + TcpCheck *UptimeCheckConfig_TcpCheck `protobuf:"bytes,6,opt,name=tcp_check,json=tcpCheck,proto3,oneof"` } func (*UptimeCheckConfig_MonitoredResource) isUptimeCheckConfig_Resource() {} @@ -414,9 +414,9 @@ func _UptimeCheckConfig_OneofSizer(msg proto.Message) (n int) { type UptimeCheckConfig_ResourceGroup struct { // The group of resources being monitored. Should be only the // group_id, not projects//groups/. - GroupId string `protobuf:"bytes,1,opt,name=group_id,json=groupId" json:"group_id,omitempty"` + GroupId string `protobuf:"bytes,1,opt,name=group_id,json=groupId,proto3" json:"group_id,omitempty"` // The resource type of the group members. - ResourceType GroupResourceType `protobuf:"varint,2,opt,name=resource_type,json=resourceType,enum=google.monitoring.v3.GroupResourceType" json:"resource_type,omitempty"` + ResourceType GroupResourceType `protobuf:"varint,2,opt,name=resource_type,json=resourceType,proto3,enum=google.monitoring.v3.GroupResourceType" json:"resource_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -463,25 +463,25 @@ func (m *UptimeCheckConfig_ResourceGroup) GetResourceType() GroupResourceType { // Information involved in an HTTP/HTTPS uptime check request. type UptimeCheckConfig_HttpCheck struct { // If true, use HTTPS instead of HTTP to run the check. - UseSsl bool `protobuf:"varint,1,opt,name=use_ssl,json=useSsl" json:"use_ssl,omitempty"` + UseSsl bool `protobuf:"varint,1,opt,name=use_ssl,json=useSsl,proto3" json:"use_ssl,omitempty"` // The path to the page to run the check against. Will be combined with the // host (specified within the MonitoredResource) and port to construct the // full URL. Optional (defaults to "/"). - Path string `protobuf:"bytes,2,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,2,opt,name=path,proto3" json:"path,omitempty"` // The port to the page to run the check against. Will be combined with host // (specified within the MonitoredResource) and path to construct the full // URL. Optional (defaults to 80 without SSL, or 443 with SSL). - Port int32 `protobuf:"varint,3,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,3,opt,name=port,proto3" json:"port,omitempty"` // The authentication information. Optional when creating an HTTP check; // defaults to empty. - AuthInfo *UptimeCheckConfig_HttpCheck_BasicAuthentication `protobuf:"bytes,4,opt,name=auth_info,json=authInfo" json:"auth_info,omitempty"` + AuthInfo *UptimeCheckConfig_HttpCheck_BasicAuthentication `protobuf:"bytes,4,opt,name=auth_info,json=authInfo,proto3" json:"auth_info,omitempty"` // Boolean specifiying whether to encrypt the header information. // Encryption should be specified for any headers related to authentication // that you do not wish to be seen when retrieving the configuration. The // server will be responsible for encrypting the headers. // On Get/List calls, if mask_headers is set to True then the headers // will be obscured with ******. - MaskHeaders bool `protobuf:"varint,5,opt,name=mask_headers,json=maskHeaders" json:"mask_headers,omitempty"` + MaskHeaders bool `protobuf:"varint,5,opt,name=mask_headers,json=maskHeaders,proto3" json:"mask_headers,omitempty"` // The list of headers to send as part of the uptime check request. // If two headers have the same key and different values, they should // be entered as a single header, with the value being a comma-separated @@ -489,7 +489,7 @@ type UptimeCheckConfig_HttpCheck struct { // https://www.w3.org/Protocols/rfc2616/rfc2616.txt (page 31). // Entering two separate headers with the same key in a Create call will // cause the first to be overwritten by the second. - Headers map[string]string `protobuf:"bytes,6,rep,name=headers" json:"headers,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Headers map[string]string `protobuf:"bytes,6,rep,name=headers,proto3" json:"headers,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -566,9 +566,9 @@ func (m *UptimeCheckConfig_HttpCheck) GetHeaders() map[string]string { // Currently, only Basic authentication is supported in Uptime Monitoring. type UptimeCheckConfig_HttpCheck_BasicAuthentication struct { // The username to authenticate. - Username string `protobuf:"bytes,1,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,1,opt,name=username,proto3" json:"username,omitempty"` // The password to authenticate. - Password string `protobuf:"bytes,2,opt,name=password" json:"password,omitempty"` + Password string `protobuf:"bytes,2,opt,name=password,proto3" json:"password,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -621,7 +621,7 @@ type UptimeCheckConfig_TcpCheck struct { // The port to the page to run the check against. Will be combined with host // (specified within the MonitoredResource) to construct the full URL. // Required. - Port int32 `protobuf:"varint,1,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,1,opt,name=port,proto3" json:"port,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -663,7 +663,7 @@ func (m *UptimeCheckConfig_TcpCheck) GetPort() int32 { // and more complex matching. type UptimeCheckConfig_ContentMatcher struct { // String content to match - Content string `protobuf:"bytes,1,opt,name=content" json:"content,omitempty"` + Content string `protobuf:"bytes,1,opt,name=content,proto3" json:"content,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -703,16 +703,16 @@ func (m *UptimeCheckConfig_ContentMatcher) GetContent() string { // Nimbus InternalCheckers. type UptimeCheckConfig_InternalChecker struct { // The GCP project ID. Not necessarily the same as the project_id for the config. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The internal network to perform this uptime check on. - Network string `protobuf:"bytes,2,opt,name=network" json:"network,omitempty"` + Network string `protobuf:"bytes,2,opt,name=network,proto3" json:"network,omitempty"` // The GCP zone the uptime check should egress from. Only respected for // internal uptime checks, where internal_network is specified. - GcpZone string `protobuf:"bytes,3,opt,name=gcp_zone,json=gcpZone" json:"gcp_zone,omitempty"` + GcpZone string `protobuf:"bytes,3,opt,name=gcp_zone,json=gcpZone,proto3" json:"gcp_zone,omitempty"` // The checker ID. - CheckerId string `protobuf:"bytes,4,opt,name=checker_id,json=checkerId" json:"checker_id,omitempty"` + CheckerId string `protobuf:"bytes,4,opt,name=checker_id,json=checkerId,proto3" json:"checker_id,omitempty"` // The checker's human-readable name. - DisplayName string `protobuf:"bytes,5,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,5,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -781,17 +781,17 @@ func (m *UptimeCheckConfig_InternalChecker) GetDisplayName() string { // addresses where checkers in the location run from. type UptimeCheckIp struct { // A broad region category in which the IP address is located. - Region UptimeCheckRegion `protobuf:"varint,1,opt,name=region,enum=google.monitoring.v3.UptimeCheckRegion" json:"region,omitempty"` + Region UptimeCheckRegion `protobuf:"varint,1,opt,name=region,proto3,enum=google.monitoring.v3.UptimeCheckRegion" json:"region,omitempty"` // A more specific location within the region that typically encodes // a particular city/town/metro (and its containing state/province or country) // within the broader umbrella region category. - Location string `protobuf:"bytes,2,opt,name=location" json:"location,omitempty"` + Location string `protobuf:"bytes,2,opt,name=location,proto3" json:"location,omitempty"` // The IP address from which the uptime check originates. This is a full // IP address (not an IP address range). Most IP addresses, as of this // publication, are in IPv4 format; however, one should not rely on the // IP addresses being in IPv4 format indefinitely and should support // interpreting this field in either IPv4 or IPv6 format. - IpAddress string `protobuf:"bytes,3,opt,name=ip_address,json=ipAddress" json:"ip_address,omitempty"` + IpAddress string `protobuf:"bytes,3,opt,name=ip_address,json=ipAddress,proto3" json:"ip_address,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime_service.pb.go b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime_service.pb.go index cdf8e1a7e..23fe5ef45 100644 --- a/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime_service.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/monitoring/v3/uptime_service.pb.go @@ -31,16 +31,16 @@ type ListUptimeCheckConfigsRequest struct { // The project whose uptime check configurations are listed. The format is // // `projects/[PROJECT_ID]`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The maximum number of results to return in a single response. The server // may further constrain the maximum number of results returned in a single // page. If the page_size is <=0, the server will decide the number of results // to be returned. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return more results from the previous method call. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -94,13 +94,13 @@ func (m *ListUptimeCheckConfigsRequest) GetPageToken() string { // The protocol for the `ListUptimeCheckConfigs` response. type ListUptimeCheckConfigsResponse struct { // The returned uptime check configurations. - UptimeCheckConfigs []*UptimeCheckConfig `protobuf:"bytes,1,rep,name=uptime_check_configs,json=uptimeCheckConfigs" json:"uptime_check_configs,omitempty"` + UptimeCheckConfigs []*UptimeCheckConfig `protobuf:"bytes,1,rep,name=uptime_check_configs,json=uptimeCheckConfigs,proto3" json:"uptime_check_configs,omitempty"` // This field represents the pagination token to retrieve the next page of // results. If the value is empty, it means no further results for the // request. To retrieve the next page of results, the value of the // next_page_token is passed to the subsequent List method call (in the // request message's page_token field). - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -149,7 +149,7 @@ type GetUptimeCheckConfigRequest struct { // The uptime check configuration to retrieve. The format is // // `projects/[PROJECT_ID]/uptimeCheckConfigs/[UPTIME_CHECK_ID]`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -191,9 +191,9 @@ type CreateUptimeCheckConfigRequest struct { // The project in which to create the uptime check. The format is: // // `projects/[PROJECT_ID]`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The new uptime check configuration. - UptimeCheckConfig *UptimeCheckConfig `protobuf:"bytes,2,opt,name=uptime_check_config,json=uptimeCheckConfig" json:"uptime_check_config,omitempty"` + UptimeCheckConfig *UptimeCheckConfig `protobuf:"bytes,2,opt,name=uptime_check_config,json=uptimeCheckConfig,proto3" json:"uptime_check_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -243,7 +243,7 @@ type UpdateUptimeCheckConfigRequest struct { // configuration are updated with values from the new configuration. If this // field is empty, then the current configuration is completely replaced with // the new configuration. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` // Required. If an `"updateMask"` has been specified, this field gives // the values for the set of fields mentioned in the `"updateMask"`. If an // `"updateMask"` has not been given, this uptime check configuration replaces @@ -251,7 +251,7 @@ type UpdateUptimeCheckConfigRequest struct { // the corresonding field is omitted in this partial uptime check // configuration, it has the effect of deleting/clearing the field from the // configuration on the server. - UptimeCheckConfig *UptimeCheckConfig `protobuf:"bytes,3,opt,name=uptime_check_config,json=uptimeCheckConfig" json:"uptime_check_config,omitempty"` + UptimeCheckConfig *UptimeCheckConfig `protobuf:"bytes,3,opt,name=uptime_check_config,json=uptimeCheckConfig,proto3" json:"uptime_check_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -300,7 +300,7 @@ type DeleteUptimeCheckConfigRequest struct { // The uptime check configuration to delete. The format is // // `projects/[PROJECT_ID]/uptimeCheckConfigs/[UPTIME_CHECK_ID]`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -344,12 +344,12 @@ type ListUptimeCheckIpsRequest struct { // page. If the page_size is <=0, the server will decide the number of results // to be returned. // NOTE: this field is not yet implemented - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If this field is not empty then it must contain the `nextPageToken` value // returned by a previous call to this method. Using this field causes the // method to return more results from the previous method call. // NOTE: this field is not yet implemented - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -397,14 +397,14 @@ func (m *ListUptimeCheckIpsRequest) GetPageToken() string { type ListUptimeCheckIpsResponse struct { // The returned list of IP addresses (including region and location) that the // checkers run from. - UptimeCheckIps []*UptimeCheckIp `protobuf:"bytes,1,rep,name=uptime_check_ips,json=uptimeCheckIps" json:"uptime_check_ips,omitempty"` + UptimeCheckIps []*UptimeCheckIp `protobuf:"bytes,1,rep,name=uptime_check_ips,json=uptimeCheckIps,proto3" json:"uptime_check_ips,omitempty"` // This field represents the pagination token to retrieve the next page of // results. If the value is empty, it means no further results for the // request. To retrieve the next page of results, the value of the // next_page_token is passed to the subsequent List method call (in the // request message's page_token field). // NOTE: this field is not yet implemented - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -553,8 +553,7 @@ func (c *uptimeCheckServiceClient) ListUptimeCheckIps(ctx context.Context, in *L return out, nil } -// Server API for UptimeCheckService service - +// UptimeCheckServiceServer is the server API for UptimeCheckService service. type UptimeCheckServiceServer interface { // Lists the existing valid uptime check configurations for the project, // leaving out any invalid configurations. diff --git a/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/dlp.pb.go b/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/dlp.pb.go index 59a865fe1..234993e2e 100644 --- a/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/dlp.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/dlp.pb.go @@ -59,7 +59,7 @@ func (x ContentOption) String() string { return proto.EnumName(ContentOption_name, int32(x)) } func (ContentOption) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{0} + return fileDescriptor_dlp_536657b46c30097f, []int{0} } // Parts of the APIs which use certain infoTypes. @@ -88,7 +88,7 @@ func (x InfoTypeSupportedBy) String() string { return proto.EnumName(InfoTypeSupportedBy_name, int32(x)) } func (InfoTypeSupportedBy) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{1} + return fileDescriptor_dlp_536657b46c30097f, []int{1} } // Operators available for comparing the value of fields. @@ -137,7 +137,7 @@ func (x RelationalOperator) String() string { return proto.EnumName(RelationalOperator_name, int32(x)) } func (RelationalOperator) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{2} + return fileDescriptor_dlp_536657b46c30097f, []int{2} } // An enum to represent the various type of DLP jobs. @@ -166,7 +166,7 @@ func (x DlpJobType) String() string { return proto.EnumName(DlpJobType_name, int32(x)) } func (DlpJobType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{3} + return fileDescriptor_dlp_536657b46c30097f, []int{3} } type ByteContentItem_BytesType int32 @@ -204,7 +204,7 @@ func (x ByteContentItem_BytesType) String() string { return proto.EnumName(ByteContentItem_BytesType_name, int32(x)) } func (ByteContentItem_BytesType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{1, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{1, 0} } // Predefined schemas for storing findings. @@ -246,7 +246,7 @@ func (x OutputStorageConfig_OutputSchema) String() string { return proto.EnumName(OutputStorageConfig_OutputSchema_name, int32(x)) } func (OutputStorageConfig_OutputSchema) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{23, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{23, 0} } type TimePartConfig_TimePart int32 @@ -290,7 +290,7 @@ func (x TimePartConfig_TimePart) String() string { return proto.EnumName(TimePartConfig_TimePart_name, int32(x)) } func (TimePartConfig_TimePart) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{38, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{40, 0} } type CharsToIgnore_CommonCharsToIgnore int32 @@ -330,7 +330,7 @@ func (x CharsToIgnore_CommonCharsToIgnore) String() string { return proto.EnumName(CharsToIgnore_CommonCharsToIgnore_name, int32(x)) } func (CharsToIgnore_CommonCharsToIgnore) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{43, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{45, 0} } // These are commonly used subsets of the alphabet that the FFX mode @@ -369,7 +369,7 @@ func (x CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet) String() string { return proto.EnumName(CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet_name, int32(x)) } func (CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{47, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{49, 0} } type RecordCondition_Expressions_LogicalOperator int32 @@ -392,7 +392,7 @@ func (x RecordCondition_Expressions_LogicalOperator) String() string { return proto.EnumName(RecordCondition_Expressions_LogicalOperator_name, int32(x)) } func (RecordCondition_Expressions_LogicalOperator) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{57, 2, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{59, 2, 0} } // Possible outcomes of transformations. @@ -419,7 +419,7 @@ func (x TransformationSummary_TransformationResultCode) String() string { return proto.EnumName(TransformationSummary_TransformationResultCode_name, int32(x)) } func (TransformationSummary_TransformationResultCode) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{59, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{61, 0} } // Whether the trigger is currently active. If PAUSED or CANCELLED, no jobs @@ -455,7 +455,7 @@ func (x JobTrigger_Status) String() string { return proto.EnumName(JobTrigger_Status_name, int32(x)) } func (JobTrigger_Status) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{64, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{66, 0} } type DlpJob_JobState int32 @@ -495,7 +495,7 @@ func (x DlpJob_JobState) String() string { return proto.EnumName(DlpJob_JobState_name, int32(x)) } func (DlpJob_JobState) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{80, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{82, 0} } // Configuration description of the scanning process. @@ -503,22 +503,29 @@ func (DlpJob_JobState) EnumDescriptor() ([]byte, []int) { // used. type InspectConfig struct { // Restricts what info_types to look for. The values must correspond to - // InfoType values returned by ListInfoTypes or found in documentation. - InfoTypes []*InfoType `protobuf:"bytes,1,rep,name=info_types,json=infoTypes" json:"info_types,omitempty"` + // InfoType values returned by ListInfoTypes or listed at + // https://cloud.google.com/dlp/docs/infotypes-reference. + // + // When no InfoTypes or CustomInfoTypes are specified in a request, the + // system may automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. + InfoTypes []*InfoType `protobuf:"bytes,1,rep,name=info_types,json=infoTypes,proto3" json:"info_types,omitempty"` // Only returns findings equal or above this threshold. The default is // POSSIBLE. - MinLikelihood Likelihood `protobuf:"varint,2,opt,name=min_likelihood,json=minLikelihood,enum=google.privacy.dlp.v2.Likelihood" json:"min_likelihood,omitempty"` - Limits *InspectConfig_FindingLimits `protobuf:"bytes,3,opt,name=limits" json:"limits,omitempty"` + // See https://cloud.google.com/dlp/docs/likelihood to learn more. + MinLikelihood Likelihood `protobuf:"varint,2,opt,name=min_likelihood,json=minLikelihood,proto3,enum=google.privacy.dlp.v2.Likelihood" json:"min_likelihood,omitempty"` + Limits *InspectConfig_FindingLimits `protobuf:"bytes,3,opt,name=limits,proto3" json:"limits,omitempty"` // When true, a contextual quote from the data that triggered a finding is // included in the response; see Finding.quote. - IncludeQuote bool `protobuf:"varint,4,opt,name=include_quote,json=includeQuote" json:"include_quote,omitempty"` + IncludeQuote bool `protobuf:"varint,4,opt,name=include_quote,json=includeQuote,proto3" json:"include_quote,omitempty"` // When true, excludes type information of the findings. - ExcludeInfoTypes bool `protobuf:"varint,5,opt,name=exclude_info_types,json=excludeInfoTypes" json:"exclude_info_types,omitempty"` - // Custom infoTypes provided by the user. - CustomInfoTypes []*CustomInfoType `protobuf:"bytes,6,rep,name=custom_info_types,json=customInfoTypes" json:"custom_info_types,omitempty"` + ExcludeInfoTypes bool `protobuf:"varint,5,opt,name=exclude_info_types,json=excludeInfoTypes,proto3" json:"exclude_info_types,omitempty"` + // CustomInfoTypes provided by the user. See + // https://cloud.google.com/dlp/docs/creating-custom-infotypes to learn more. + CustomInfoTypes []*CustomInfoType `protobuf:"bytes,6,rep,name=custom_info_types,json=customInfoTypes,proto3" json:"custom_info_types,omitempty"` // List of options defining data content to scan. // If empty, text, images, and other content will be included. - ContentOptions []ContentOption `protobuf:"varint,8,rep,packed,name=content_options,json=contentOptions,enum=google.privacy.dlp.v2.ContentOption" json:"content_options,omitempty"` + ContentOptions []ContentOption `protobuf:"varint,8,rep,packed,name=content_options,json=contentOptions,proto3,enum=google.privacy.dlp.v2.ContentOption" json:"content_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -528,7 +535,7 @@ func (m *InspectConfig) Reset() { *m = InspectConfig{} } func (m *InspectConfig) String() string { return proto.CompactTextString(m) } func (*InspectConfig) ProtoMessage() {} func (*InspectConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{0} + return fileDescriptor_dlp_536657b46c30097f, []int{0} } func (m *InspectConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectConfig.Unmarshal(m, b) @@ -602,13 +609,13 @@ type InspectConfig_FindingLimits struct { // When set within `InspectDataSourceRequest`, // the maximum returned is 1000 regardless if this is set higher. // When set within `InspectContentRequest`, this field is ignored. - MaxFindingsPerItem int32 `protobuf:"varint,1,opt,name=max_findings_per_item,json=maxFindingsPerItem" json:"max_findings_per_item,omitempty"` + MaxFindingsPerItem int32 `protobuf:"varint,1,opt,name=max_findings_per_item,json=maxFindingsPerItem,proto3" json:"max_findings_per_item,omitempty"` // Max number of findings that will be returned per request/job. // When set within `InspectContentRequest`, the maximum returned is 1000 // regardless if this is set higher. - MaxFindingsPerRequest int32 `protobuf:"varint,2,opt,name=max_findings_per_request,json=maxFindingsPerRequest" json:"max_findings_per_request,omitempty"` + MaxFindingsPerRequest int32 `protobuf:"varint,2,opt,name=max_findings_per_request,json=maxFindingsPerRequest,proto3" json:"max_findings_per_request,omitempty"` // Configuration of findings limit given for specified infoTypes. - MaxFindingsPerInfoType []*InspectConfig_FindingLimits_InfoTypeLimit `protobuf:"bytes,3,rep,name=max_findings_per_info_type,json=maxFindingsPerInfoType" json:"max_findings_per_info_type,omitempty"` + MaxFindingsPerInfoType []*InspectConfig_FindingLimits_InfoTypeLimit `protobuf:"bytes,3,rep,name=max_findings_per_info_type,json=maxFindingsPerInfoType,proto3" json:"max_findings_per_info_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -618,7 +625,7 @@ func (m *InspectConfig_FindingLimits) Reset() { *m = InspectConfig_Findi func (m *InspectConfig_FindingLimits) String() string { return proto.CompactTextString(m) } func (*InspectConfig_FindingLimits) ProtoMessage() {} func (*InspectConfig_FindingLimits) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{0, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{0, 0} } func (m *InspectConfig_FindingLimits) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectConfig_FindingLimits.Unmarshal(m, b) @@ -666,9 +673,9 @@ type InspectConfig_FindingLimits_InfoTypeLimit struct { // info_type should be provided. If InfoTypeLimit does not have an // info_type, the DLP API applies the limit against all info_types that // are found but not specified in another InfoTypeLimit. - InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType" json:"info_type,omitempty"` + InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType,proto3" json:"info_type,omitempty"` // Max findings limit for the given infoType. - MaxFindings int32 `protobuf:"varint,2,opt,name=max_findings,json=maxFindings" json:"max_findings,omitempty"` + MaxFindings int32 `protobuf:"varint,2,opt,name=max_findings,json=maxFindings,proto3" json:"max_findings,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -680,7 +687,7 @@ func (m *InspectConfig_FindingLimits_InfoTypeLimit) Reset() { func (m *InspectConfig_FindingLimits_InfoTypeLimit) String() string { return proto.CompactTextString(m) } func (*InspectConfig_FindingLimits_InfoTypeLimit) ProtoMessage() {} func (*InspectConfig_FindingLimits_InfoTypeLimit) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{0, 0, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{0, 0, 0} } func (m *InspectConfig_FindingLimits_InfoTypeLimit) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectConfig_FindingLimits_InfoTypeLimit.Unmarshal(m, b) @@ -717,7 +724,7 @@ func (m *InspectConfig_FindingLimits_InfoTypeLimit) GetMaxFindings() int32 { // Container for bytes to inspect or redact. type ByteContentItem struct { // The type of data stored in the bytes string. Default will be TEXT_UTF8. - Type ByteContentItem_BytesType `protobuf:"varint,1,opt,name=type,enum=google.privacy.dlp.v2.ByteContentItem_BytesType" json:"type,omitempty"` + Type ByteContentItem_BytesType `protobuf:"varint,1,opt,name=type,proto3,enum=google.privacy.dlp.v2.ByteContentItem_BytesType" json:"type,omitempty"` // Content data to inspect or redact. Data []byte `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -729,7 +736,7 @@ func (m *ByteContentItem) Reset() { *m = ByteContentItem{} } func (m *ByteContentItem) String() string { return proto.CompactTextString(m) } func (*ByteContentItem) ProtoMessage() {} func (*ByteContentItem) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{1} + return fileDescriptor_dlp_536657b46c30097f, []int{1} } func (m *ByteContentItem) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ByteContentItem.Unmarshal(m, b) @@ -781,7 +788,7 @@ func (m *ContentItem) Reset() { *m = ContentItem{} } func (m *ContentItem) String() string { return proto.CompactTextString(m) } func (*ContentItem) ProtoMessage() {} func (*ContentItem) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{2} + return fileDescriptor_dlp_536657b46c30097f, []int{2} } func (m *ContentItem) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ContentItem.Unmarshal(m, b) @@ -806,13 +813,13 @@ type isContentItem_DataItem interface { } type ContentItem_Value struct { - Value string `protobuf:"bytes,3,opt,name=value,oneof"` + Value string `protobuf:"bytes,3,opt,name=value,proto3,oneof"` } type ContentItem_Table struct { - Table *Table `protobuf:"bytes,4,opt,name=table,oneof"` + Table *Table `protobuf:"bytes,4,opt,name=table,proto3,oneof"` } type ContentItem_ByteItem struct { - ByteItem *ByteContentItem `protobuf:"bytes,5,opt,name=byte_item,json=byteItem,oneof"` + ByteItem *ByteContentItem `protobuf:"bytes,5,opt,name=byte_item,json=byteItem,proto3,oneof"` } func (*ContentItem_Value) isContentItem_DataItem() {} @@ -937,9 +944,11 @@ func _ContentItem_OneofSizer(msg proto.Message) (n int) { } // Structured content to inspect. Up to 50,000 `Value`s per request allowed. +// See https://cloud.google.com/dlp/docs/inspecting-text#inspecting_a_table to +// learn more. type Table struct { - Headers []*FieldId `protobuf:"bytes,1,rep,name=headers" json:"headers,omitempty"` - Rows []*Table_Row `protobuf:"bytes,2,rep,name=rows" json:"rows,omitempty"` + Headers []*FieldId `protobuf:"bytes,1,rep,name=headers,proto3" json:"headers,omitempty"` + Rows []*Table_Row `protobuf:"bytes,2,rep,name=rows,proto3" json:"rows,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -949,7 +958,7 @@ func (m *Table) Reset() { *m = Table{} } func (m *Table) String() string { return proto.CompactTextString(m) } func (*Table) ProtoMessage() {} func (*Table) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{3} + return fileDescriptor_dlp_536657b46c30097f, []int{3} } func (m *Table) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Table.Unmarshal(m, b) @@ -984,7 +993,7 @@ func (m *Table) GetRows() []*Table_Row { } type Table_Row struct { - Values []*Value `protobuf:"bytes,1,rep,name=values" json:"values,omitempty"` + Values []*Value `protobuf:"bytes,1,rep,name=values,proto3" json:"values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -994,7 +1003,7 @@ func (m *Table_Row) Reset() { *m = Table_Row{} } func (m *Table_Row) String() string { return proto.CompactTextString(m) } func (*Table_Row) ProtoMessage() {} func (*Table_Row) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{3, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{3, 0} } func (m *Table_Row) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Table_Row.Unmarshal(m, b) @@ -1024,14 +1033,14 @@ func (m *Table_Row) GetValues() []*Value { // All the findings for a single scanned item. type InspectResult struct { // List of findings for an item. - Findings []*Finding `protobuf:"bytes,1,rep,name=findings" json:"findings,omitempty"` + Findings []*Finding `protobuf:"bytes,1,rep,name=findings,proto3" json:"findings,omitempty"` // If true, then this item might have more findings than were returned, // and the findings returned are an arbitrary subset of all findings. // The findings list might be truncated because the input items were too // large, or because the server reached the maximum amount of resources // allowed for a single API call. For best results, divide the input into // smaller batches. - FindingsTruncated bool `protobuf:"varint,2,opt,name=findings_truncated,json=findingsTruncated" json:"findings_truncated,omitempty"` + FindingsTruncated bool `protobuf:"varint,2,opt,name=findings_truncated,json=findingsTruncated,proto3" json:"findings_truncated,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1041,7 +1050,7 @@ func (m *InspectResult) Reset() { *m = InspectResult{} } func (m *InspectResult) String() string { return proto.CompactTextString(m) } func (*InspectResult) ProtoMessage() {} func (*InspectResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{4} + return fileDescriptor_dlp_536657b46c30097f, []int{4} } func (m *InspectResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectResult.Unmarshal(m, b) @@ -1079,23 +1088,23 @@ func (m *InspectResult) GetFindingsTruncated() bool { type Finding struct { // The content that was found. Even if the content is not textual, it // may be converted to a textual representation here. - // Provided if requested by the `InspectConfig` and the finding is + // Provided if `include_quote` is true and the finding is // less than or equal to 4096 bytes long. If the finding exceeds 4096 bytes // in length, the quote may be omitted. - Quote string `protobuf:"bytes,1,opt,name=quote" json:"quote,omitempty"` + Quote string `protobuf:"bytes,1,opt,name=quote,proto3" json:"quote,omitempty"` // The type of content that might have been found. - // Provided if requested by the `InspectConfig`. - InfoType *InfoType `protobuf:"bytes,2,opt,name=info_type,json=infoType" json:"info_type,omitempty"` - // Estimate of how likely it is that the `info_type` is correct. - Likelihood Likelihood `protobuf:"varint,3,opt,name=likelihood,enum=google.privacy.dlp.v2.Likelihood" json:"likelihood,omitempty"` + // Provided if `excluded_types` is false. + InfoType *InfoType `protobuf:"bytes,2,opt,name=info_type,json=infoType,proto3" json:"info_type,omitempty"` + // Confidence of how likely it is that the `info_type` is correct. + Likelihood Likelihood `protobuf:"varint,3,opt,name=likelihood,proto3,enum=google.privacy.dlp.v2.Likelihood" json:"likelihood,omitempty"` // Where the content was found. - Location *Location `protobuf:"bytes,4,opt,name=location" json:"location,omitempty"` + Location *Location `protobuf:"bytes,4,opt,name=location,proto3" json:"location,omitempty"` // Timestamp when finding was detected. - CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Contains data parsed from quotes. Only populated if include_quote was set // to true and a supported infoType was requested. Currently supported // infoTypes: DATE, DATE_OF_BIRTH and TIME. - QuoteInfo *QuoteInfo `protobuf:"bytes,7,opt,name=quote_info,json=quoteInfo" json:"quote_info,omitempty"` + QuoteInfo *QuoteInfo `protobuf:"bytes,7,opt,name=quote_info,json=quoteInfo,proto3" json:"quote_info,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1105,7 +1114,7 @@ func (m *Finding) Reset() { *m = Finding{} } func (m *Finding) String() string { return proto.CompactTextString(m) } func (*Finding) ProtoMessage() {} func (*Finding) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{5} + return fileDescriptor_dlp_536657b46c30097f, []int{5} } func (m *Finding) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Finding.Unmarshal(m, b) @@ -1174,14 +1183,14 @@ type Location struct { // Note that when the content is not textual, this references // the UTF-8 encoded textual representation of the content. // Omitted if content is an image. - ByteRange *Range `protobuf:"bytes,1,opt,name=byte_range,json=byteRange" json:"byte_range,omitempty"` + ByteRange *Range `protobuf:"bytes,1,opt,name=byte_range,json=byteRange,proto3" json:"byte_range,omitempty"` // Unicode character offsets delimiting the finding. // These are relative to the finding's containing element. // Provided when the content is text. - CodepointRange *Range `protobuf:"bytes,2,opt,name=codepoint_range,json=codepointRange" json:"codepoint_range,omitempty"` + CodepointRange *Range `protobuf:"bytes,2,opt,name=codepoint_range,json=codepointRange,proto3" json:"codepoint_range,omitempty"` // List of nested objects pointing to the precise location of the finding // within the file or record. - ContentLocations []*ContentLocation `protobuf:"bytes,7,rep,name=content_locations,json=contentLocations" json:"content_locations,omitempty"` + ContentLocations []*ContentLocation `protobuf:"bytes,7,rep,name=content_locations,json=contentLocations,proto3" json:"content_locations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1191,7 +1200,7 @@ func (m *Location) Reset() { *m = Location{} } func (m *Location) String() string { return proto.CompactTextString(m) } func (*Location) ProtoMessage() {} func (*Location) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{6} + return fileDescriptor_dlp_536657b46c30097f, []int{6} } func (m *Location) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Location.Unmarshal(m, b) @@ -1238,7 +1247,7 @@ type ContentLocation struct { // The top level name is the source file name or table name. Nested names // could be absent if the embedded object has no string identifier // (for an example an image contained within a document). - ContainerName string `protobuf:"bytes,1,opt,name=container_name,json=containerName" json:"container_name,omitempty"` + ContainerName string `protobuf:"bytes,1,opt,name=container_name,json=containerName,proto3" json:"container_name,omitempty"` // Type of the container within the file with location of the finding. // // Types that are valid to be assigned to Location: @@ -1250,10 +1259,10 @@ type ContentLocation struct { // For Google Cloud Storage contains last file modification timestamp. // For BigQuery table contains last_modified_time property. // For Datastore - not populated. - ContainerTimestamp *timestamp.Timestamp `protobuf:"bytes,6,opt,name=container_timestamp,json=containerTimestamp" json:"container_timestamp,omitempty"` + ContainerTimestamp *timestamp.Timestamp `protobuf:"bytes,6,opt,name=container_timestamp,json=containerTimestamp,proto3" json:"container_timestamp,omitempty"` // Findings container version, if available // ("generation" for Google Cloud Storage). - ContainerVersion string `protobuf:"bytes,7,opt,name=container_version,json=containerVersion" json:"container_version,omitempty"` + ContainerVersion string `protobuf:"bytes,7,opt,name=container_version,json=containerVersion,proto3" json:"container_version,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1263,7 +1272,7 @@ func (m *ContentLocation) Reset() { *m = ContentLocation{} } func (m *ContentLocation) String() string { return proto.CompactTextString(m) } func (*ContentLocation) ProtoMessage() {} func (*ContentLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{7} + return fileDescriptor_dlp_536657b46c30097f, []int{7} } func (m *ContentLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ContentLocation.Unmarshal(m, b) @@ -1288,13 +1297,13 @@ type isContentLocation_Location interface { } type ContentLocation_RecordLocation struct { - RecordLocation *RecordLocation `protobuf:"bytes,2,opt,name=record_location,json=recordLocation,oneof"` + RecordLocation *RecordLocation `protobuf:"bytes,2,opt,name=record_location,json=recordLocation,proto3,oneof"` } type ContentLocation_ImageLocation struct { - ImageLocation *ImageLocation `protobuf:"bytes,3,opt,name=image_location,json=imageLocation,oneof"` + ImageLocation *ImageLocation `protobuf:"bytes,3,opt,name=image_location,json=imageLocation,proto3,oneof"` } type ContentLocation_DocumentLocation struct { - DocumentLocation *DocumentLocation `protobuf:"bytes,5,opt,name=document_location,json=documentLocation,oneof"` + DocumentLocation *DocumentLocation `protobuf:"bytes,5,opt,name=document_location,json=documentLocation,proto3,oneof"` } func (*ContentLocation_RecordLocation) isContentLocation_Location() {} @@ -1447,7 +1456,7 @@ func _ContentLocation_OneofSizer(msg proto.Message) (n int) { type DocumentLocation struct { // Offset of the line, from the beginning of the file, where the finding // is located. - FileOffset int64 `protobuf:"varint,1,opt,name=file_offset,json=fileOffset" json:"file_offset,omitempty"` + FileOffset int64 `protobuf:"varint,1,opt,name=file_offset,json=fileOffset,proto3" json:"file_offset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1457,7 +1466,7 @@ func (m *DocumentLocation) Reset() { *m = DocumentLocation{} } func (m *DocumentLocation) String() string { return proto.CompactTextString(m) } func (*DocumentLocation) ProtoMessage() {} func (*DocumentLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{8} + return fileDescriptor_dlp_536657b46c30097f, []int{8} } func (m *DocumentLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DocumentLocation.Unmarshal(m, b) @@ -1487,11 +1496,11 @@ func (m *DocumentLocation) GetFileOffset() int64 { // Location of a finding within a row or record. type RecordLocation struct { // Key of the finding. - RecordKey *RecordKey `protobuf:"bytes,1,opt,name=record_key,json=recordKey" json:"record_key,omitempty"` + RecordKey *RecordKey `protobuf:"bytes,1,opt,name=record_key,json=recordKey,proto3" json:"record_key,omitempty"` // Field id of the field containing the finding. - FieldId *FieldId `protobuf:"bytes,2,opt,name=field_id,json=fieldId" json:"field_id,omitempty"` + FieldId *FieldId `protobuf:"bytes,2,opt,name=field_id,json=fieldId,proto3" json:"field_id,omitempty"` // Location within a `ContentItem.Table`. - TableLocation *TableLocation `protobuf:"bytes,3,opt,name=table_location,json=tableLocation" json:"table_location,omitempty"` + TableLocation *TableLocation `protobuf:"bytes,3,opt,name=table_location,json=tableLocation,proto3" json:"table_location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1501,7 +1510,7 @@ func (m *RecordLocation) Reset() { *m = RecordLocation{} } func (m *RecordLocation) String() string { return proto.CompactTextString(m) } func (*RecordLocation) ProtoMessage() {} func (*RecordLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{9} + return fileDescriptor_dlp_536657b46c30097f, []int{9} } func (m *RecordLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordLocation.Unmarshal(m, b) @@ -1545,7 +1554,7 @@ func (m *RecordLocation) GetTableLocation() *TableLocation { // Location of a finding within a table. type TableLocation struct { // The zero-based index of the row where the finding is located. - RowIndex int64 `protobuf:"varint,1,opt,name=row_index,json=rowIndex" json:"row_index,omitempty"` + RowIndex int64 `protobuf:"varint,1,opt,name=row_index,json=rowIndex,proto3" json:"row_index,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1555,7 +1564,7 @@ func (m *TableLocation) Reset() { *m = TableLocation{} } func (m *TableLocation) String() string { return proto.CompactTextString(m) } func (*TableLocation) ProtoMessage() {} func (*TableLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{10} + return fileDescriptor_dlp_536657b46c30097f, []int{10} } func (m *TableLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TableLocation.Unmarshal(m, b) @@ -1585,9 +1594,9 @@ func (m *TableLocation) GetRowIndex() int64 { // Generic half-open interval [start, end) type Range struct { // Index of the first character of the range (inclusive). - Start int64 `protobuf:"varint,1,opt,name=start" json:"start,omitempty"` + Start int64 `protobuf:"varint,1,opt,name=start,proto3" json:"start,omitempty"` // Index of the last character of the range (exclusive). - End int64 `protobuf:"varint,2,opt,name=end" json:"end,omitempty"` + End int64 `protobuf:"varint,2,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1597,7 +1606,7 @@ func (m *Range) Reset() { *m = Range{} } func (m *Range) String() string { return proto.CompactTextString(m) } func (*Range) ProtoMessage() {} func (*Range) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{11} + return fileDescriptor_dlp_536657b46c30097f, []int{11} } func (m *Range) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Range.Unmarshal(m, b) @@ -1634,7 +1643,7 @@ func (m *Range) GetEnd() int64 { // Location of the finding within an image. type ImageLocation struct { // Bounding boxes locating the pixels within the image containing the finding. - BoundingBoxes []*BoundingBox `protobuf:"bytes,1,rep,name=bounding_boxes,json=boundingBoxes" json:"bounding_boxes,omitempty"` + BoundingBoxes []*BoundingBox `protobuf:"bytes,1,rep,name=bounding_boxes,json=boundingBoxes,proto3" json:"bounding_boxes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1644,7 +1653,7 @@ func (m *ImageLocation) Reset() { *m = ImageLocation{} } func (m *ImageLocation) String() string { return proto.CompactTextString(m) } func (*ImageLocation) ProtoMessage() {} func (*ImageLocation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{12} + return fileDescriptor_dlp_536657b46c30097f, []int{12} } func (m *ImageLocation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ImageLocation.Unmarshal(m, b) @@ -1674,13 +1683,13 @@ func (m *ImageLocation) GetBoundingBoxes() []*BoundingBox { // Bounding box encompassing detected text within an image. type BoundingBox struct { // Top coordinate of the bounding box. (0,0) is upper left. - Top int32 `protobuf:"varint,1,opt,name=top" json:"top,omitempty"` + Top int32 `protobuf:"varint,1,opt,name=top,proto3" json:"top,omitempty"` // Left coordinate of the bounding box. (0,0) is upper left. - Left int32 `protobuf:"varint,2,opt,name=left" json:"left,omitempty"` + Left int32 `protobuf:"varint,2,opt,name=left,proto3" json:"left,omitempty"` // Width of the bounding box in pixels. - Width int32 `protobuf:"varint,3,opt,name=width" json:"width,omitempty"` + Width int32 `protobuf:"varint,3,opt,name=width,proto3" json:"width,omitempty"` // Height of the bounding box in pixels. - Height int32 `protobuf:"varint,4,opt,name=height" json:"height,omitempty"` + Height int32 `protobuf:"varint,4,opt,name=height,proto3" json:"height,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1690,7 +1699,7 @@ func (m *BoundingBox) Reset() { *m = BoundingBox{} } func (m *BoundingBox) String() string { return proto.CompactTextString(m) } func (*BoundingBox) ProtoMessage() {} func (*BoundingBox) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{13} + return fileDescriptor_dlp_536657b46c30097f, []int{13} } func (m *BoundingBox) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BoundingBox.Unmarshal(m, b) @@ -1738,17 +1747,20 @@ func (m *BoundingBox) GetHeight() int32 { return 0 } -// Request to search for potentially sensitive info in a list of items -// and replace it with a default or provided content. +// Request to search for potentially sensitive info in an image and redact it +// by covering it with a colored rectangle. type RedactImageRequest struct { // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Configuration for the inspector. - InspectConfig *InspectConfig `protobuf:"bytes,2,opt,name=inspect_config,json=inspectConfig" json:"inspect_config,omitempty"` + InspectConfig *InspectConfig `protobuf:"bytes,2,opt,name=inspect_config,json=inspectConfig,proto3" json:"inspect_config,omitempty"` // The configuration for specifying what content to redact from images. - ImageRedactionConfigs []*RedactImageRequest_ImageRedactionConfig `protobuf:"bytes,5,rep,name=image_redaction_configs,json=imageRedactionConfigs" json:"image_redaction_configs,omitempty"` + ImageRedactionConfigs []*RedactImageRequest_ImageRedactionConfig `protobuf:"bytes,5,rep,name=image_redaction_configs,json=imageRedactionConfigs,proto3" json:"image_redaction_configs,omitempty"` + // Whether the response should include findings along with the redacted + // image. + IncludeFindings bool `protobuf:"varint,6,opt,name=include_findings,json=includeFindings,proto3" json:"include_findings,omitempty"` // The content must be PNG, JPEG, SVG or BMP. - ByteItem *ByteContentItem `protobuf:"bytes,7,opt,name=byte_item,json=byteItem" json:"byte_item,omitempty"` + ByteItem *ByteContentItem `protobuf:"bytes,7,opt,name=byte_item,json=byteItem,proto3" json:"byte_item,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1758,7 +1770,7 @@ func (m *RedactImageRequest) Reset() { *m = RedactImageRequest{} } func (m *RedactImageRequest) String() string { return proto.CompactTextString(m) } func (*RedactImageRequest) ProtoMessage() {} func (*RedactImageRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{14} + return fileDescriptor_dlp_536657b46c30097f, []int{14} } func (m *RedactImageRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RedactImageRequest.Unmarshal(m, b) @@ -1799,6 +1811,13 @@ func (m *RedactImageRequest) GetImageRedactionConfigs() []*RedactImageRequest_Im return nil } +func (m *RedactImageRequest) GetIncludeFindings() bool { + if m != nil { + return m.IncludeFindings + } + return false +} + func (m *RedactImageRequest) GetByteItem() *ByteContentItem { if m != nil { return m.ByteItem @@ -1816,7 +1835,7 @@ type RedactImageRequest_ImageRedactionConfig struct { Target isRedactImageRequest_ImageRedactionConfig_Target `protobuf_oneof:"target"` // The color to use when redacting content from an image. If not specified, // the default is black. - RedactionColor *Color `protobuf:"bytes,3,opt,name=redaction_color,json=redactionColor" json:"redaction_color,omitempty"` + RedactionColor *Color `protobuf:"bytes,3,opt,name=redaction_color,json=redactionColor,proto3" json:"redaction_color,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1828,7 +1847,7 @@ func (m *RedactImageRequest_ImageRedactionConfig) Reset() { func (m *RedactImageRequest_ImageRedactionConfig) String() string { return proto.CompactTextString(m) } func (*RedactImageRequest_ImageRedactionConfig) ProtoMessage() {} func (*RedactImageRequest_ImageRedactionConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{14, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{14, 0} } func (m *RedactImageRequest_ImageRedactionConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RedactImageRequest_ImageRedactionConfig.Unmarshal(m, b) @@ -1853,10 +1872,10 @@ type isRedactImageRequest_ImageRedactionConfig_Target interface { } type RedactImageRequest_ImageRedactionConfig_InfoType struct { - InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType,oneof"` + InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType,proto3,oneof"` } type RedactImageRequest_ImageRedactionConfig_RedactAllText struct { - RedactAllText bool `protobuf:"varint,2,opt,name=redact_all_text,json=redactAllText,oneof"` + RedactAllText bool `protobuf:"varint,2,opt,name=redact_all_text,json=redactAllText,proto3,oneof"` } func (*RedactImageRequest_ImageRedactionConfig_InfoType) isRedactImageRequest_ImageRedactionConfig_Target() { @@ -1968,11 +1987,11 @@ func _RedactImageRequest_ImageRedactionConfig_OneofSizer(msg proto.Message) (n i // Represents a color in the RGB color space. type Color struct { // The amount of red in the color as a value in the interval [0, 1]. - Red float32 `protobuf:"fixed32,1,opt,name=red" json:"red,omitempty"` + Red float32 `protobuf:"fixed32,1,opt,name=red,proto3" json:"red,omitempty"` // The amount of green in the color as a value in the interval [0, 1]. - Green float32 `protobuf:"fixed32,2,opt,name=green" json:"green,omitempty"` + Green float32 `protobuf:"fixed32,2,opt,name=green,proto3" json:"green,omitempty"` // The amount of blue in the color as a value in the interval [0, 1]. - Blue float32 `protobuf:"fixed32,3,opt,name=blue" json:"blue,omitempty"` + Blue float32 `protobuf:"fixed32,3,opt,name=blue,proto3" json:"blue,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1982,7 +2001,7 @@ func (m *Color) Reset() { *m = Color{} } func (m *Color) String() string { return proto.CompactTextString(m) } func (*Color) ProtoMessage() {} func (*Color) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{15} + return fileDescriptor_dlp_536657b46c30097f, []int{15} } func (m *Color) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Color.Unmarshal(m, b) @@ -2030,17 +2049,19 @@ type RedactImageResponse struct { // If an image was being inspected and the InspectConfig's include_quote was // set to true, then this field will include all text, if any, that was found // in the image. - ExtractedText string `protobuf:"bytes,2,opt,name=extracted_text,json=extractedText" json:"extracted_text,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + ExtractedText string `protobuf:"bytes,2,opt,name=extracted_text,json=extractedText,proto3" json:"extracted_text,omitempty"` + // The findings. Populated when include_findings in the request is true. + InspectResult *InspectResult `protobuf:"bytes,3,opt,name=inspect_result,json=inspectResult,proto3" json:"inspect_result,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *RedactImageResponse) Reset() { *m = RedactImageResponse{} } func (m *RedactImageResponse) String() string { return proto.CompactTextString(m) } func (*RedactImageResponse) ProtoMessage() {} func (*RedactImageResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{16} + return fileDescriptor_dlp_536657b46c30097f, []int{16} } func (m *RedactImageResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RedactImageResponse.Unmarshal(m, b) @@ -2074,32 +2095,39 @@ func (m *RedactImageResponse) GetExtractedText() string { return "" } +func (m *RedactImageResponse) GetInspectResult() *InspectResult { + if m != nil { + return m.InspectResult + } + return nil +} + // Request to de-identify a list of items. type DeidentifyContentRequest struct { // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Configuration for the de-identification of the content item. // Items specified here will override the template referenced by the // deidentify_template_name argument. - DeidentifyConfig *DeidentifyConfig `protobuf:"bytes,2,opt,name=deidentify_config,json=deidentifyConfig" json:"deidentify_config,omitempty"` + DeidentifyConfig *DeidentifyConfig `protobuf:"bytes,2,opt,name=deidentify_config,json=deidentifyConfig,proto3" json:"deidentify_config,omitempty"` // Configuration for the inspector. // Items specified here will override the template referenced by the // inspect_template_name argument. - InspectConfig *InspectConfig `protobuf:"bytes,3,opt,name=inspect_config,json=inspectConfig" json:"inspect_config,omitempty"` + InspectConfig *InspectConfig `protobuf:"bytes,3,opt,name=inspect_config,json=inspectConfig,proto3" json:"inspect_config,omitempty"` // The item to de-identify. Will be treated as text. - Item *ContentItem `protobuf:"bytes,4,opt,name=item" json:"item,omitempty"` + Item *ContentItem `protobuf:"bytes,4,opt,name=item,proto3" json:"item,omitempty"` // Optional template to use. Any configuration directly specified in // inspect_config will override those set in the template. Singular fields // that are set in this request will replace their corresponding fields in the // template. Repeated fields are appended. Singular sub-messages and groups // are recursively merged. - InspectTemplateName string `protobuf:"bytes,5,opt,name=inspect_template_name,json=inspectTemplateName" json:"inspect_template_name,omitempty"` + InspectTemplateName string `protobuf:"bytes,5,opt,name=inspect_template_name,json=inspectTemplateName,proto3" json:"inspect_template_name,omitempty"` // Optional template to use. Any configuration directly specified in // deidentify_config will override those set in the template. Singular fields // that are set in this request will replace their corresponding fields in the // template. Repeated fields are appended. Singular sub-messages and groups // are recursively merged. - DeidentifyTemplateName string `protobuf:"bytes,6,opt,name=deidentify_template_name,json=deidentifyTemplateName" json:"deidentify_template_name,omitempty"` + DeidentifyTemplateName string `protobuf:"bytes,6,opt,name=deidentify_template_name,json=deidentifyTemplateName,proto3" json:"deidentify_template_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2109,7 +2137,7 @@ func (m *DeidentifyContentRequest) Reset() { *m = DeidentifyContentReque func (m *DeidentifyContentRequest) String() string { return proto.CompactTextString(m) } func (*DeidentifyContentRequest) ProtoMessage() {} func (*DeidentifyContentRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{17} + return fileDescriptor_dlp_536657b46c30097f, []int{17} } func (m *DeidentifyContentRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeidentifyContentRequest.Unmarshal(m, b) @@ -2174,9 +2202,9 @@ func (m *DeidentifyContentRequest) GetDeidentifyTemplateName() string { // Results of de-identifying a ContentItem. type DeidentifyContentResponse struct { // The de-identified item. - Item *ContentItem `protobuf:"bytes,1,opt,name=item" json:"item,omitempty"` + Item *ContentItem `protobuf:"bytes,1,opt,name=item,proto3" json:"item,omitempty"` // An overview of the changes that were made on the `item`. - Overview *TransformationOverview `protobuf:"bytes,2,opt,name=overview" json:"overview,omitempty"` + Overview *TransformationOverview `protobuf:"bytes,2,opt,name=overview,proto3" json:"overview,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2186,7 +2214,7 @@ func (m *DeidentifyContentResponse) Reset() { *m = DeidentifyContentResp func (m *DeidentifyContentResponse) String() string { return proto.CompactTextString(m) } func (*DeidentifyContentResponse) ProtoMessage() {} func (*DeidentifyContentResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{18} + return fileDescriptor_dlp_536657b46c30097f, []int{18} } func (m *DeidentifyContentResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeidentifyContentResponse.Unmarshal(m, b) @@ -2223,7 +2251,7 @@ func (m *DeidentifyContentResponse) GetOverview() *TransformationOverview { // Request to re-identify an item. type ReidentifyContentRequest struct { // The parent resource name. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Configuration for the re-identification of the content item. // This field shares the same proto message type that is used for // de-identification, however its usage here is for the reversal of the @@ -2233,24 +2261,24 @@ type ReidentifyContentRequest struct { // be provided here. The reversible transformations are: // // - `CryptoReplaceFfxFpeConfig` - ReidentifyConfig *DeidentifyConfig `protobuf:"bytes,2,opt,name=reidentify_config,json=reidentifyConfig" json:"reidentify_config,omitempty"` + ReidentifyConfig *DeidentifyConfig `protobuf:"bytes,2,opt,name=reidentify_config,json=reidentifyConfig,proto3" json:"reidentify_config,omitempty"` // Configuration for the inspector. - InspectConfig *InspectConfig `protobuf:"bytes,3,opt,name=inspect_config,json=inspectConfig" json:"inspect_config,omitempty"` + InspectConfig *InspectConfig `protobuf:"bytes,3,opt,name=inspect_config,json=inspectConfig,proto3" json:"inspect_config,omitempty"` // The item to re-identify. Will be treated as text. - Item *ContentItem `protobuf:"bytes,4,opt,name=item" json:"item,omitempty"` + Item *ContentItem `protobuf:"bytes,4,opt,name=item,proto3" json:"item,omitempty"` // Optional template to use. Any configuration directly specified in // `inspect_config` will override those set in the template. Singular fields // that are set in this request will replace their corresponding fields in the // template. Repeated fields are appended. Singular sub-messages and groups // are recursively merged. - InspectTemplateName string `protobuf:"bytes,5,opt,name=inspect_template_name,json=inspectTemplateName" json:"inspect_template_name,omitempty"` + InspectTemplateName string `protobuf:"bytes,5,opt,name=inspect_template_name,json=inspectTemplateName,proto3" json:"inspect_template_name,omitempty"` // Optional template to use. References an instance of `DeidentifyTemplate`. // Any configuration directly specified in `reidentify_config` or // `inspect_config` will override those set in the template. Singular fields // that are set in this request will replace their corresponding fields in the // template. Repeated fields are appended. Singular sub-messages and groups // are recursively merged. - ReidentifyTemplateName string `protobuf:"bytes,6,opt,name=reidentify_template_name,json=reidentifyTemplateName" json:"reidentify_template_name,omitempty"` + ReidentifyTemplateName string `protobuf:"bytes,6,opt,name=reidentify_template_name,json=reidentifyTemplateName,proto3" json:"reidentify_template_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2260,7 +2288,7 @@ func (m *ReidentifyContentRequest) Reset() { *m = ReidentifyContentReque func (m *ReidentifyContentRequest) String() string { return proto.CompactTextString(m) } func (*ReidentifyContentRequest) ProtoMessage() {} func (*ReidentifyContentRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{19} + return fileDescriptor_dlp_536657b46c30097f, []int{19} } func (m *ReidentifyContentRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReidentifyContentRequest.Unmarshal(m, b) @@ -2325,9 +2353,9 @@ func (m *ReidentifyContentRequest) GetReidentifyTemplateName() string { // Results of re-identifying a item. type ReidentifyContentResponse struct { // The re-identified item. - Item *ContentItem `protobuf:"bytes,1,opt,name=item" json:"item,omitempty"` + Item *ContentItem `protobuf:"bytes,1,opt,name=item,proto3" json:"item,omitempty"` // An overview of the changes that were made to the `item`. - Overview *TransformationOverview `protobuf:"bytes,2,opt,name=overview" json:"overview,omitempty"` + Overview *TransformationOverview `protobuf:"bytes,2,opt,name=overview,proto3" json:"overview,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2337,7 +2365,7 @@ func (m *ReidentifyContentResponse) Reset() { *m = ReidentifyContentResp func (m *ReidentifyContentResponse) String() string { return proto.CompactTextString(m) } func (*ReidentifyContentResponse) ProtoMessage() {} func (*ReidentifyContentResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{20} + return fileDescriptor_dlp_536657b46c30097f, []int{20} } func (m *ReidentifyContentResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReidentifyContentResponse.Unmarshal(m, b) @@ -2374,18 +2402,18 @@ func (m *ReidentifyContentResponse) GetOverview() *TransformationOverview { // Request to search for potentially sensitive info in a ContentItem. type InspectContentRequest struct { // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Configuration for the inspector. What specified here will override // the template referenced by the inspect_template_name argument. - InspectConfig *InspectConfig `protobuf:"bytes,2,opt,name=inspect_config,json=inspectConfig" json:"inspect_config,omitempty"` + InspectConfig *InspectConfig `protobuf:"bytes,2,opt,name=inspect_config,json=inspectConfig,proto3" json:"inspect_config,omitempty"` // The item to inspect. - Item *ContentItem `protobuf:"bytes,3,opt,name=item" json:"item,omitempty"` + Item *ContentItem `protobuf:"bytes,3,opt,name=item,proto3" json:"item,omitempty"` // Optional template to use. Any configuration directly specified in // inspect_config will override those set in the template. Singular fields // that are set in this request will replace their corresponding fields in the // template. Repeated fields are appended. Singular sub-messages and groups // are recursively merged. - InspectTemplateName string `protobuf:"bytes,4,opt,name=inspect_template_name,json=inspectTemplateName" json:"inspect_template_name,omitempty"` + InspectTemplateName string `protobuf:"bytes,4,opt,name=inspect_template_name,json=inspectTemplateName,proto3" json:"inspect_template_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2395,7 +2423,7 @@ func (m *InspectContentRequest) Reset() { *m = InspectContentRequest{} } func (m *InspectContentRequest) String() string { return proto.CompactTextString(m) } func (*InspectContentRequest) ProtoMessage() {} func (*InspectContentRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{21} + return fileDescriptor_dlp_536657b46c30097f, []int{21} } func (m *InspectContentRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectContentRequest.Unmarshal(m, b) @@ -2446,7 +2474,7 @@ func (m *InspectContentRequest) GetInspectTemplateName() string { // Results of inspecting an item. type InspectContentResponse struct { // The findings. - Result *InspectResult `protobuf:"bytes,1,opt,name=result" json:"result,omitempty"` + Result *InspectResult `protobuf:"bytes,1,opt,name=result,proto3" json:"result,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2456,7 +2484,7 @@ func (m *InspectContentResponse) Reset() { *m = InspectContentResponse{} func (m *InspectContentResponse) String() string { return proto.CompactTextString(m) } func (*InspectContentResponse) ProtoMessage() {} func (*InspectContentResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{22} + return fileDescriptor_dlp_536657b46c30097f, []int{22} } func (m *InspectContentResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectContentResponse.Unmarshal(m, b) @@ -2488,14 +2516,15 @@ type OutputStorageConfig struct { // Types that are valid to be assigned to Type: // *OutputStorageConfig_Table Type isOutputStorageConfig_Type `protobuf_oneof:"type"` - // Schema used for writing the findings. Columns are derived from the - // `Finding` object. If appending to an existing table, any columns from the - // predefined schema that are missing will be added. No columns in the - // existing table will be deleted. + // Schema used for writing the findings for Inspect jobs. This field is only + // used for Inspect and must be unspecified for Risk jobs. Columns are derived + // from the `Finding` object. If appending to an existing table, any columns + // from the predefined schema that are missing will be added. No columns in + // the existing table will be deleted. // // If unspecified, then all available columns will be used for a new table, // and no changes will be made to an existing table. - OutputSchema OutputStorageConfig_OutputSchema `protobuf:"varint,3,opt,name=output_schema,json=outputSchema,enum=google.privacy.dlp.v2.OutputStorageConfig_OutputSchema" json:"output_schema,omitempty"` + OutputSchema OutputStorageConfig_OutputSchema `protobuf:"varint,3,opt,name=output_schema,json=outputSchema,proto3,enum=google.privacy.dlp.v2.OutputStorageConfig_OutputSchema" json:"output_schema,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2505,7 +2534,7 @@ func (m *OutputStorageConfig) Reset() { *m = OutputStorageConfig{} } func (m *OutputStorageConfig) String() string { return proto.CompactTextString(m) } func (*OutputStorageConfig) ProtoMessage() {} func (*OutputStorageConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{23} + return fileDescriptor_dlp_536657b46c30097f, []int{23} } func (m *OutputStorageConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_OutputStorageConfig.Unmarshal(m, b) @@ -2530,7 +2559,7 @@ type isOutputStorageConfig_Type interface { } type OutputStorageConfig_Table struct { - Table *BigQueryTable `protobuf:"bytes,1,opt,name=table,oneof"` + Table *BigQueryTable `protobuf:"bytes,1,opt,name=table,proto3,oneof"` } func (*OutputStorageConfig_Table) isOutputStorageConfig_Type() {} @@ -2614,9 +2643,9 @@ func _OutputStorageConfig_OneofSizer(msg proto.Message) (n int) { // Statistics regarding a specific InfoType. type InfoTypeStats struct { // The type of finding this stat is for. - InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType" json:"info_type,omitempty"` + InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType,proto3" json:"info_type,omitempty"` // Number of findings for this infoType. - Count int64 `protobuf:"varint,2,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,2,opt,name=count,proto3" json:"count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2626,7 +2655,7 @@ func (m *InfoTypeStats) Reset() { *m = InfoTypeStats{} } func (m *InfoTypeStats) String() string { return proto.CompactTextString(m) } func (*InfoTypeStats) ProtoMessage() {} func (*InfoTypeStats) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{24} + return fileDescriptor_dlp_536657b46c30097f, []int{24} } func (m *InfoTypeStats) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InfoTypeStats.Unmarshal(m, b) @@ -2663,9 +2692,9 @@ func (m *InfoTypeStats) GetCount() int64 { // The results of an inspect DataSource job. type InspectDataSourceDetails struct { // The configuration used for this job. - RequestedOptions *InspectDataSourceDetails_RequestedOptions `protobuf:"bytes,2,opt,name=requested_options,json=requestedOptions" json:"requested_options,omitempty"` + RequestedOptions *InspectDataSourceDetails_RequestedOptions `protobuf:"bytes,2,opt,name=requested_options,json=requestedOptions,proto3" json:"requested_options,omitempty"` // A summary of the outcome of this inspect job. - Result *InspectDataSourceDetails_Result `protobuf:"bytes,3,opt,name=result" json:"result,omitempty"` + Result *InspectDataSourceDetails_Result `protobuf:"bytes,3,opt,name=result,proto3" json:"result,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2675,7 +2704,7 @@ func (m *InspectDataSourceDetails) Reset() { *m = InspectDataSourceDetai func (m *InspectDataSourceDetails) String() string { return proto.CompactTextString(m) } func (*InspectDataSourceDetails) ProtoMessage() {} func (*InspectDataSourceDetails) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{25} + return fileDescriptor_dlp_536657b46c30097f, []int{25} } func (m *InspectDataSourceDetails) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectDataSourceDetails.Unmarshal(m, b) @@ -2712,8 +2741,8 @@ func (m *InspectDataSourceDetails) GetResult() *InspectDataSourceDetails_Result type InspectDataSourceDetails_RequestedOptions struct { // If run with an InspectTemplate, a snapshot of its state at the time of // this run. - SnapshotInspectTemplate *InspectTemplate `protobuf:"bytes,1,opt,name=snapshot_inspect_template,json=snapshotInspectTemplate" json:"snapshot_inspect_template,omitempty"` - JobConfig *InspectJobConfig `protobuf:"bytes,3,opt,name=job_config,json=jobConfig" json:"job_config,omitempty"` + SnapshotInspectTemplate *InspectTemplate `protobuf:"bytes,1,opt,name=snapshot_inspect_template,json=snapshotInspectTemplate,proto3" json:"snapshot_inspect_template,omitempty"` + JobConfig *InspectJobConfig `protobuf:"bytes,3,opt,name=job_config,json=jobConfig,proto3" json:"job_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2725,7 +2754,7 @@ func (m *InspectDataSourceDetails_RequestedOptions) Reset() { func (m *InspectDataSourceDetails_RequestedOptions) String() string { return proto.CompactTextString(m) } func (*InspectDataSourceDetails_RequestedOptions) ProtoMessage() {} func (*InspectDataSourceDetails_RequestedOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{25, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{25, 0} } func (m *InspectDataSourceDetails_RequestedOptions) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectDataSourceDetails_RequestedOptions.Unmarshal(m, b) @@ -2761,12 +2790,12 @@ func (m *InspectDataSourceDetails_RequestedOptions) GetJobConfig() *InspectJobCo type InspectDataSourceDetails_Result struct { // Total size in bytes that were processed. - ProcessedBytes int64 `protobuf:"varint,1,opt,name=processed_bytes,json=processedBytes" json:"processed_bytes,omitempty"` + ProcessedBytes int64 `protobuf:"varint,1,opt,name=processed_bytes,json=processedBytes,proto3" json:"processed_bytes,omitempty"` // Estimate of the number of bytes to process. - TotalEstimatedBytes int64 `protobuf:"varint,2,opt,name=total_estimated_bytes,json=totalEstimatedBytes" json:"total_estimated_bytes,omitempty"` + TotalEstimatedBytes int64 `protobuf:"varint,2,opt,name=total_estimated_bytes,json=totalEstimatedBytes,proto3" json:"total_estimated_bytes,omitempty"` // Statistics of how many instances of each info type were found during // inspect job. - InfoTypeStats []*InfoTypeStats `protobuf:"bytes,3,rep,name=info_type_stats,json=infoTypeStats" json:"info_type_stats,omitempty"` + InfoTypeStats []*InfoTypeStats `protobuf:"bytes,3,rep,name=info_type_stats,json=infoTypeStats,proto3" json:"info_type_stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2776,7 +2805,7 @@ func (m *InspectDataSourceDetails_Result) Reset() { *m = InspectDataSour func (m *InspectDataSourceDetails_Result) String() string { return proto.CompactTextString(m) } func (*InspectDataSourceDetails_Result) ProtoMessage() {} func (*InspectDataSourceDetails_Result) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{25, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{25, 1} } func (m *InspectDataSourceDetails_Result) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectDataSourceDetails_Result.Unmarshal(m, b) @@ -2820,11 +2849,11 @@ func (m *InspectDataSourceDetails_Result) GetInfoTypeStats() []*InfoTypeStats { // InfoType description. type InfoTypeDescription struct { // Internal name of the infoType. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Human readable form of the infoType name. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Which parts of the API supports this InfoType. - SupportedBy []InfoTypeSupportedBy `protobuf:"varint,3,rep,packed,name=supported_by,json=supportedBy,enum=google.privacy.dlp.v2.InfoTypeSupportedBy" json:"supported_by,omitempty"` + SupportedBy []InfoTypeSupportedBy `protobuf:"varint,3,rep,packed,name=supported_by,json=supportedBy,proto3,enum=google.privacy.dlp.v2.InfoTypeSupportedBy" json:"supported_by,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2834,7 +2863,7 @@ func (m *InfoTypeDescription) Reset() { *m = InfoTypeDescription{} } func (m *InfoTypeDescription) String() string { return proto.CompactTextString(m) } func (*InfoTypeDescription) ProtoMessage() {} func (*InfoTypeDescription) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{26} + return fileDescriptor_dlp_536657b46c30097f, []int{26} } func (m *InfoTypeDescription) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InfoTypeDescription.Unmarshal(m, b) @@ -2880,10 +2909,10 @@ type ListInfoTypesRequest struct { // Optional BCP-47 language code for localized infoType friendly // names. If omitted, or if localized strings are not available, // en-US strings will be returned. - LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional filter to only return infoTypes supported by certain parts of the // API. Defaults to supported_by=INSPECT. - Filter string `protobuf:"bytes,2,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,2,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2893,7 +2922,7 @@ func (m *ListInfoTypesRequest) Reset() { *m = ListInfoTypesRequest{} } func (m *ListInfoTypesRequest) String() string { return proto.CompactTextString(m) } func (*ListInfoTypesRequest) ProtoMessage() {} func (*ListInfoTypesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{27} + return fileDescriptor_dlp_536657b46c30097f, []int{27} } func (m *ListInfoTypesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListInfoTypesRequest.Unmarshal(m, b) @@ -2930,7 +2959,7 @@ func (m *ListInfoTypesRequest) GetFilter() string { // Response to the ListInfoTypes request. type ListInfoTypesResponse struct { // Set of sensitive infoTypes. - InfoTypes []*InfoTypeDescription `protobuf:"bytes,1,rep,name=info_types,json=infoTypes" json:"info_types,omitempty"` + InfoTypes []*InfoTypeDescription `protobuf:"bytes,1,rep,name=info_types,json=infoTypes,proto3" json:"info_types,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2940,7 +2969,7 @@ func (m *ListInfoTypesResponse) Reset() { *m = ListInfoTypesResponse{} } func (m *ListInfoTypesResponse) String() string { return proto.CompactTextString(m) } func (*ListInfoTypesResponse) ProtoMessage() {} func (*ListInfoTypesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{28} + return fileDescriptor_dlp_536657b46c30097f, []int{28} } func (m *ListInfoTypesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListInfoTypesResponse.Unmarshal(m, b) @@ -2967,15 +2996,16 @@ func (m *ListInfoTypesResponse) GetInfoTypes() []*InfoTypeDescription { return nil } -// Configuration for a risk analysis job. +// Configuration for a risk analysis job. See +// https://cloud.google.com/dlp/docs/concepts-risk-analysis to learn more. type RiskAnalysisJobConfig struct { // Privacy metric to compute. - PrivacyMetric *PrivacyMetric `protobuf:"bytes,1,opt,name=privacy_metric,json=privacyMetric" json:"privacy_metric,omitempty"` + PrivacyMetric *PrivacyMetric `protobuf:"bytes,1,opt,name=privacy_metric,json=privacyMetric,proto3" json:"privacy_metric,omitempty"` // Input dataset to compute metrics over. - SourceTable *BigQueryTable `protobuf:"bytes,2,opt,name=source_table,json=sourceTable" json:"source_table,omitempty"` + SourceTable *BigQueryTable `protobuf:"bytes,2,opt,name=source_table,json=sourceTable,proto3" json:"source_table,omitempty"` // Actions to execute at the completion of the job. Are executed in the order // provided. - Actions []*Action `protobuf:"bytes,3,rep,name=actions" json:"actions,omitempty"` + Actions []*Action `protobuf:"bytes,3,rep,name=actions,proto3" json:"actions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2985,7 +3015,7 @@ func (m *RiskAnalysisJobConfig) Reset() { *m = RiskAnalysisJobConfig{} } func (m *RiskAnalysisJobConfig) String() string { return proto.CompactTextString(m) } func (*RiskAnalysisJobConfig) ProtoMessage() {} func (*RiskAnalysisJobConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{29} + return fileDescriptor_dlp_536657b46c30097f, []int{29} } func (m *RiskAnalysisJobConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RiskAnalysisJobConfig.Unmarshal(m, b) @@ -3026,6 +3056,304 @@ func (m *RiskAnalysisJobConfig) GetActions() []*Action { return nil } +// A column with a semantic tag attached. +type QuasiId struct { + // Identifies the column. [required] + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` + // Semantic tag that identifies what a column contains, to determine which + // statistical model to use to estimate the reidentifiability of each + // value. [required] + // + // Types that are valid to be assigned to Tag: + // *QuasiId_InfoType + // *QuasiId_CustomTag + // *QuasiId_Inferred + Tag isQuasiId_Tag `protobuf_oneof:"tag"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *QuasiId) Reset() { *m = QuasiId{} } +func (m *QuasiId) String() string { return proto.CompactTextString(m) } +func (*QuasiId) ProtoMessage() {} +func (*QuasiId) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{30} +} +func (m *QuasiId) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_QuasiId.Unmarshal(m, b) +} +func (m *QuasiId) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_QuasiId.Marshal(b, m, deterministic) +} +func (dst *QuasiId) XXX_Merge(src proto.Message) { + xxx_messageInfo_QuasiId.Merge(dst, src) +} +func (m *QuasiId) XXX_Size() int { + return xxx_messageInfo_QuasiId.Size(m) +} +func (m *QuasiId) XXX_DiscardUnknown() { + xxx_messageInfo_QuasiId.DiscardUnknown(m) +} + +var xxx_messageInfo_QuasiId proto.InternalMessageInfo + +type isQuasiId_Tag interface { + isQuasiId_Tag() +} + +type QuasiId_InfoType struct { + InfoType *InfoType `protobuf:"bytes,2,opt,name=info_type,json=infoType,proto3,oneof"` +} +type QuasiId_CustomTag struct { + CustomTag string `protobuf:"bytes,3,opt,name=custom_tag,json=customTag,proto3,oneof"` +} +type QuasiId_Inferred struct { + Inferred *empty.Empty `protobuf:"bytes,4,opt,name=inferred,proto3,oneof"` +} + +func (*QuasiId_InfoType) isQuasiId_Tag() {} +func (*QuasiId_CustomTag) isQuasiId_Tag() {} +func (*QuasiId_Inferred) isQuasiId_Tag() {} + +func (m *QuasiId) GetTag() isQuasiId_Tag { + if m != nil { + return m.Tag + } + return nil +} + +func (m *QuasiId) GetField() *FieldId { + if m != nil { + return m.Field + } + return nil +} + +func (m *QuasiId) GetInfoType() *InfoType { + if x, ok := m.GetTag().(*QuasiId_InfoType); ok { + return x.InfoType + } + return nil +} + +func (m *QuasiId) GetCustomTag() string { + if x, ok := m.GetTag().(*QuasiId_CustomTag); ok { + return x.CustomTag + } + return "" +} + +func (m *QuasiId) GetInferred() *empty.Empty { + if x, ok := m.GetTag().(*QuasiId_Inferred); ok { + return x.Inferred + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*QuasiId) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _QuasiId_OneofMarshaler, _QuasiId_OneofUnmarshaler, _QuasiId_OneofSizer, []interface{}{ + (*QuasiId_InfoType)(nil), + (*QuasiId_CustomTag)(nil), + (*QuasiId_Inferred)(nil), + } +} + +func _QuasiId_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*QuasiId) + // tag + switch x := m.Tag.(type) { + case *QuasiId_InfoType: + b.EncodeVarint(2<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.InfoType); err != nil { + return err + } + case *QuasiId_CustomTag: + b.EncodeVarint(3<<3 | proto.WireBytes) + b.EncodeStringBytes(x.CustomTag) + case *QuasiId_Inferred: + b.EncodeVarint(4<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.Inferred); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("QuasiId.Tag has unexpected type %T", x) + } + return nil +} + +func _QuasiId_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*QuasiId) + switch tag { + case 2: // tag.info_type + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(InfoType) + err := b.DecodeMessage(msg) + m.Tag = &QuasiId_InfoType{msg} + return true, err + case 3: // tag.custom_tag + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + x, err := b.DecodeStringBytes() + m.Tag = &QuasiId_CustomTag{x} + return true, err + case 4: // tag.inferred + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(empty.Empty) + err := b.DecodeMessage(msg) + m.Tag = &QuasiId_Inferred{msg} + return true, err + default: + return false, nil + } +} + +func _QuasiId_OneofSizer(msg proto.Message) (n int) { + m := msg.(*QuasiId) + // tag + switch x := m.Tag.(type) { + case *QuasiId_InfoType: + s := proto.Size(x.InfoType) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *QuasiId_CustomTag: + n += 1 // tag and wire + n += proto.SizeVarint(uint64(len(x.CustomTag))) + n += len(x.CustomTag) + case *QuasiId_Inferred: + s := proto.Size(x.Inferred) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +// An auxiliary table containing statistical information on the relative +// frequency of different quasi-identifiers values. It has one or several +// quasi-identifiers columns, and one column that indicates the relative +// frequency of each quasi-identifier tuple. +// If a tuple is present in the data but not in the auxiliary table, the +// corresponding relative frequency is assumed to be zero (and thus, the +// tuple is highly reidentifiable). +type StatisticalTable struct { + // Auxiliary table location. [required] + Table *BigQueryTable `protobuf:"bytes,3,opt,name=table,proto3" json:"table,omitempty"` + // Quasi-identifier columns. [required] + QuasiIds []*StatisticalTable_QuasiIdentifierField `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds,proto3" json:"quasi_ids,omitempty"` + // The relative frequency column must contain a floating-point number + // between 0 and 1 (inclusive). Null values are assumed to be zero. + // [required] + RelativeFrequency *FieldId `protobuf:"bytes,2,opt,name=relative_frequency,json=relativeFrequency,proto3" json:"relative_frequency,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StatisticalTable) Reset() { *m = StatisticalTable{} } +func (m *StatisticalTable) String() string { return proto.CompactTextString(m) } +func (*StatisticalTable) ProtoMessage() {} +func (*StatisticalTable) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{31} +} +func (m *StatisticalTable) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StatisticalTable.Unmarshal(m, b) +} +func (m *StatisticalTable) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StatisticalTable.Marshal(b, m, deterministic) +} +func (dst *StatisticalTable) XXX_Merge(src proto.Message) { + xxx_messageInfo_StatisticalTable.Merge(dst, src) +} +func (m *StatisticalTable) XXX_Size() int { + return xxx_messageInfo_StatisticalTable.Size(m) +} +func (m *StatisticalTable) XXX_DiscardUnknown() { + xxx_messageInfo_StatisticalTable.DiscardUnknown(m) +} + +var xxx_messageInfo_StatisticalTable proto.InternalMessageInfo + +func (m *StatisticalTable) GetTable() *BigQueryTable { + if m != nil { + return m.Table + } + return nil +} + +func (m *StatisticalTable) GetQuasiIds() []*StatisticalTable_QuasiIdentifierField { + if m != nil { + return m.QuasiIds + } + return nil +} + +func (m *StatisticalTable) GetRelativeFrequency() *FieldId { + if m != nil { + return m.RelativeFrequency + } + return nil +} + +// A quasi-identifier column has a custom_tag, used to know which column +// in the data corresponds to which column in the statistical model. +type StatisticalTable_QuasiIdentifierField struct { + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` + CustomTag string `protobuf:"bytes,2,opt,name=custom_tag,json=customTag,proto3" json:"custom_tag,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *StatisticalTable_QuasiIdentifierField) Reset() { *m = StatisticalTable_QuasiIdentifierField{} } +func (m *StatisticalTable_QuasiIdentifierField) String() string { return proto.CompactTextString(m) } +func (*StatisticalTable_QuasiIdentifierField) ProtoMessage() {} +func (*StatisticalTable_QuasiIdentifierField) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{31, 0} +} +func (m *StatisticalTable_QuasiIdentifierField) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StatisticalTable_QuasiIdentifierField.Unmarshal(m, b) +} +func (m *StatisticalTable_QuasiIdentifierField) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StatisticalTable_QuasiIdentifierField.Marshal(b, m, deterministic) +} +func (dst *StatisticalTable_QuasiIdentifierField) XXX_Merge(src proto.Message) { + xxx_messageInfo_StatisticalTable_QuasiIdentifierField.Merge(dst, src) +} +func (m *StatisticalTable_QuasiIdentifierField) XXX_Size() int { + return xxx_messageInfo_StatisticalTable_QuasiIdentifierField.Size(m) +} +func (m *StatisticalTable_QuasiIdentifierField) XXX_DiscardUnknown() { + xxx_messageInfo_StatisticalTable_QuasiIdentifierField.DiscardUnknown(m) +} + +var xxx_messageInfo_StatisticalTable_QuasiIdentifierField proto.InternalMessageInfo + +func (m *StatisticalTable_QuasiIdentifierField) GetField() *FieldId { + if m != nil { + return m.Field + } + return nil +} + +func (m *StatisticalTable_QuasiIdentifierField) GetCustomTag() string { + if m != nil { + return m.CustomTag + } + return "" +} + // Privacy metric to compute for reidentification risk analysis. type PrivacyMetric struct { // Types that are valid to be assigned to Type: @@ -3034,6 +3362,7 @@ type PrivacyMetric struct { // *PrivacyMetric_KAnonymityConfig_ // *PrivacyMetric_LDiversityConfig_ // *PrivacyMetric_KMapEstimationConfig_ + // *PrivacyMetric_DeltaPresenceEstimationConfig_ Type isPrivacyMetric_Type `protobuf_oneof:"type"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` @@ -3044,7 +3373,7 @@ func (m *PrivacyMetric) Reset() { *m = PrivacyMetric{} } func (m *PrivacyMetric) String() string { return proto.CompactTextString(m) } func (*PrivacyMetric) ProtoMessage() {} func (*PrivacyMetric) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30} + return fileDescriptor_dlp_536657b46c30097f, []int{32} } func (m *PrivacyMetric) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric.Unmarshal(m, b) @@ -3069,26 +3398,30 @@ type isPrivacyMetric_Type interface { } type PrivacyMetric_NumericalStatsConfig_ struct { - NumericalStatsConfig *PrivacyMetric_NumericalStatsConfig `protobuf:"bytes,1,opt,name=numerical_stats_config,json=numericalStatsConfig,oneof"` + NumericalStatsConfig *PrivacyMetric_NumericalStatsConfig `protobuf:"bytes,1,opt,name=numerical_stats_config,json=numericalStatsConfig,proto3,oneof"` } type PrivacyMetric_CategoricalStatsConfig_ struct { - CategoricalStatsConfig *PrivacyMetric_CategoricalStatsConfig `protobuf:"bytes,2,opt,name=categorical_stats_config,json=categoricalStatsConfig,oneof"` + CategoricalStatsConfig *PrivacyMetric_CategoricalStatsConfig `protobuf:"bytes,2,opt,name=categorical_stats_config,json=categoricalStatsConfig,proto3,oneof"` } type PrivacyMetric_KAnonymityConfig_ struct { - KAnonymityConfig *PrivacyMetric_KAnonymityConfig `protobuf:"bytes,3,opt,name=k_anonymity_config,json=kAnonymityConfig,oneof"` + KAnonymityConfig *PrivacyMetric_KAnonymityConfig `protobuf:"bytes,3,opt,name=k_anonymity_config,json=kAnonymityConfig,proto3,oneof"` } type PrivacyMetric_LDiversityConfig_ struct { - LDiversityConfig *PrivacyMetric_LDiversityConfig `protobuf:"bytes,4,opt,name=l_diversity_config,json=lDiversityConfig,oneof"` + LDiversityConfig *PrivacyMetric_LDiversityConfig `protobuf:"bytes,4,opt,name=l_diversity_config,json=lDiversityConfig,proto3,oneof"` } type PrivacyMetric_KMapEstimationConfig_ struct { - KMapEstimationConfig *PrivacyMetric_KMapEstimationConfig `protobuf:"bytes,5,opt,name=k_map_estimation_config,json=kMapEstimationConfig,oneof"` + KMapEstimationConfig *PrivacyMetric_KMapEstimationConfig `protobuf:"bytes,5,opt,name=k_map_estimation_config,json=kMapEstimationConfig,proto3,oneof"` +} +type PrivacyMetric_DeltaPresenceEstimationConfig_ struct { + DeltaPresenceEstimationConfig *PrivacyMetric_DeltaPresenceEstimationConfig `protobuf:"bytes,6,opt,name=delta_presence_estimation_config,json=deltaPresenceEstimationConfig,proto3,oneof"` } -func (*PrivacyMetric_NumericalStatsConfig_) isPrivacyMetric_Type() {} -func (*PrivacyMetric_CategoricalStatsConfig_) isPrivacyMetric_Type() {} -func (*PrivacyMetric_KAnonymityConfig_) isPrivacyMetric_Type() {} -func (*PrivacyMetric_LDiversityConfig_) isPrivacyMetric_Type() {} -func (*PrivacyMetric_KMapEstimationConfig_) isPrivacyMetric_Type() {} +func (*PrivacyMetric_NumericalStatsConfig_) isPrivacyMetric_Type() {} +func (*PrivacyMetric_CategoricalStatsConfig_) isPrivacyMetric_Type() {} +func (*PrivacyMetric_KAnonymityConfig_) isPrivacyMetric_Type() {} +func (*PrivacyMetric_LDiversityConfig_) isPrivacyMetric_Type() {} +func (*PrivacyMetric_KMapEstimationConfig_) isPrivacyMetric_Type() {} +func (*PrivacyMetric_DeltaPresenceEstimationConfig_) isPrivacyMetric_Type() {} func (m *PrivacyMetric) GetType() isPrivacyMetric_Type { if m != nil { @@ -3132,6 +3465,13 @@ func (m *PrivacyMetric) GetKMapEstimationConfig() *PrivacyMetric_KMapEstimationC return nil } +func (m *PrivacyMetric) GetDeltaPresenceEstimationConfig() *PrivacyMetric_DeltaPresenceEstimationConfig { + if x, ok := m.GetType().(*PrivacyMetric_DeltaPresenceEstimationConfig_); ok { + return x.DeltaPresenceEstimationConfig + } + return nil +} + // XXX_OneofFuncs is for the internal use of the proto package. func (*PrivacyMetric) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { return _PrivacyMetric_OneofMarshaler, _PrivacyMetric_OneofUnmarshaler, _PrivacyMetric_OneofSizer, []interface{}{ @@ -3140,6 +3480,7 @@ func (*PrivacyMetric) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) (*PrivacyMetric_KAnonymityConfig_)(nil), (*PrivacyMetric_LDiversityConfig_)(nil), (*PrivacyMetric_KMapEstimationConfig_)(nil), + (*PrivacyMetric_DeltaPresenceEstimationConfig_)(nil), } } @@ -3172,6 +3513,11 @@ func _PrivacyMetric_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { if err := b.EncodeMessage(x.KMapEstimationConfig); err != nil { return err } + case *PrivacyMetric_DeltaPresenceEstimationConfig_: + b.EncodeVarint(6<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.DeltaPresenceEstimationConfig); err != nil { + return err + } case nil: default: return fmt.Errorf("PrivacyMetric.Type has unexpected type %T", x) @@ -3222,6 +3568,14 @@ func _PrivacyMetric_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto. err := b.DecodeMessage(msg) m.Type = &PrivacyMetric_KMapEstimationConfig_{msg} return true, err + case 6: // type.delta_presence_estimation_config + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(PrivacyMetric_DeltaPresenceEstimationConfig) + err := b.DecodeMessage(msg) + m.Type = &PrivacyMetric_DeltaPresenceEstimationConfig_{msg} + return true, err default: return false, nil } @@ -3256,6 +3610,11 @@ func _PrivacyMetric_OneofSizer(msg proto.Message) (n int) { n += 1 // tag and wire n += proto.SizeVarint(uint64(s)) n += s + case *PrivacyMetric_DeltaPresenceEstimationConfig_: + s := proto.Size(x.DeltaPresenceEstimationConfig) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s case nil: default: panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) @@ -3268,7 +3627,7 @@ func _PrivacyMetric_OneofSizer(msg proto.Message) (n int) { type PrivacyMetric_NumericalStatsConfig struct { // Field to compute numerical stats on. Supported types are // integer, float, date, datetime, timestamp, time. - Field *FieldId `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3278,7 +3637,7 @@ func (m *PrivacyMetric_NumericalStatsConfig) Reset() { *m = PrivacyMetri func (m *PrivacyMetric_NumericalStatsConfig) String() string { return proto.CompactTextString(m) } func (*PrivacyMetric_NumericalStatsConfig) ProtoMessage() {} func (*PrivacyMetric_NumericalStatsConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 0} } func (m *PrivacyMetric_NumericalStatsConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_NumericalStatsConfig.Unmarshal(m, b) @@ -3312,7 +3671,7 @@ type PrivacyMetric_CategoricalStatsConfig struct { // supported except for arrays and structs. However, it may be more // informative to use NumericalStats when the field type is supported, // depending on the data. - Field *FieldId `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3322,7 +3681,7 @@ func (m *PrivacyMetric_CategoricalStatsConfig) Reset() { *m = PrivacyMet func (m *PrivacyMetric_CategoricalStatsConfig) String() string { return proto.CompactTextString(m) } func (*PrivacyMetric_CategoricalStatsConfig) ProtoMessage() {} func (*PrivacyMetric_CategoricalStatsConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 1} } func (m *PrivacyMetric_CategoricalStatsConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_CategoricalStatsConfig.Unmarshal(m, b) @@ -3356,7 +3715,7 @@ type PrivacyMetric_KAnonymityConfig struct { // repeated data types are not supported; however, nested fields are // supported so long as they are not structs themselves or nested within // a repeated field. - QuasiIds []*FieldId `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds" json:"quasi_ids,omitempty"` + QuasiIds []*FieldId `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds,proto3" json:"quasi_ids,omitempty"` // Optional message indicating that multiple rows might be associated to a // single individual. If the same entity_id is associated to multiple // quasi-identifier tuples over distict rows, we consider the entire @@ -3367,7 +3726,7 @@ type PrivacyMetric_KAnonymityConfig struct { // Important note: a maximum of 1000 rows can be associated to a single // entity ID. If more rows are associated with the same entity ID, some // might be ignored. - EntityId *EntityId `protobuf:"bytes,2,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"` + EntityId *EntityId `protobuf:"bytes,2,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3377,7 +3736,7 @@ func (m *PrivacyMetric_KAnonymityConfig) Reset() { *m = PrivacyMetric_KA func (m *PrivacyMetric_KAnonymityConfig) String() string { return proto.CompactTextString(m) } func (*PrivacyMetric_KAnonymityConfig) ProtoMessage() {} func (*PrivacyMetric_KAnonymityConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 2} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 2} } func (m *PrivacyMetric_KAnonymityConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_KAnonymityConfig.Unmarshal(m, b) @@ -3416,9 +3775,9 @@ type PrivacyMetric_LDiversityConfig struct { // Set of quasi-identifiers indicating how equivalence classes are // defined for the l-diversity computation. When multiple fields are // specified, they are considered a single composite key. - QuasiIds []*FieldId `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds" json:"quasi_ids,omitempty"` + QuasiIds []*FieldId `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds,proto3" json:"quasi_ids,omitempty"` // Sensitive field for computing the l-value. - SensitiveAttribute *FieldId `protobuf:"bytes,2,opt,name=sensitive_attribute,json=sensitiveAttribute" json:"sensitive_attribute,omitempty"` + SensitiveAttribute *FieldId `protobuf:"bytes,2,opt,name=sensitive_attribute,json=sensitiveAttribute,proto3" json:"sensitive_attribute,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3428,7 +3787,7 @@ func (m *PrivacyMetric_LDiversityConfig) Reset() { *m = PrivacyMetric_LD func (m *PrivacyMetric_LDiversityConfig) String() string { return proto.CompactTextString(m) } func (*PrivacyMetric_LDiversityConfig) ProtoMessage() {} func (*PrivacyMetric_LDiversityConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 3} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 3} } func (m *PrivacyMetric_LDiversityConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_LDiversityConfig.Unmarshal(m, b) @@ -3468,18 +3827,19 @@ func (m *PrivacyMetric_LDiversityConfig) GetSensitiveAttribute() *FieldId { // using publicly available data (like the US Census), or using a custom // statistical model (indicated as one or several BigQuery tables), or by // extrapolating from the distribution of values in the input dataset. +// A column with a semantic tag attached. type PrivacyMetric_KMapEstimationConfig struct { // Fields considered to be quasi-identifiers. No two columns can have the // same tag. [required] - QuasiIds []*PrivacyMetric_KMapEstimationConfig_TaggedField `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds" json:"quasi_ids,omitempty"` + QuasiIds []*PrivacyMetric_KMapEstimationConfig_TaggedField `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds,proto3" json:"quasi_ids,omitempty"` // ISO 3166-1 alpha-2 region code to use in the statistical modeling. // Required if no column is tagged with a region-specific InfoType (like // US_ZIP_5) or a region code. - RegionCode string `protobuf:"bytes,2,opt,name=region_code,json=regionCode" json:"region_code,omitempty"` + RegionCode string `protobuf:"bytes,2,opt,name=region_code,json=regionCode,proto3" json:"region_code,omitempty"` // Several auxiliary tables can be used in the analysis. Each custom_tag // used to tag a quasi-identifiers column must appear in exactly one column // of one auxiliary table. - AuxiliaryTables []*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable `protobuf:"bytes,3,rep,name=auxiliary_tables,json=auxiliaryTables" json:"auxiliary_tables,omitempty"` + AuxiliaryTables []*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable `protobuf:"bytes,3,rep,name=auxiliary_tables,json=auxiliaryTables,proto3" json:"auxiliary_tables,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3489,7 +3849,7 @@ func (m *PrivacyMetric_KMapEstimationConfig) Reset() { *m = PrivacyMetri func (m *PrivacyMetric_KMapEstimationConfig) String() string { return proto.CompactTextString(m) } func (*PrivacyMetric_KMapEstimationConfig) ProtoMessage() {} func (*PrivacyMetric_KMapEstimationConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 4} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 4} } func (m *PrivacyMetric_KMapEstimationConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_KMapEstimationConfig.Unmarshal(m, b) @@ -3530,10 +3890,9 @@ func (m *PrivacyMetric_KMapEstimationConfig) GetAuxiliaryTables() []*PrivacyMetr return nil } -// A column with a semantic tag attached. type PrivacyMetric_KMapEstimationConfig_TaggedField struct { // Identifies the column. [required] - Field *FieldId `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` // Semantic tag that identifies what a column contains, to determine which // statistical model to use to estimate the reidentifiability of each // value. [required] @@ -3556,7 +3915,7 @@ func (m *PrivacyMetric_KMapEstimationConfig_TaggedField) String() string { } func (*PrivacyMetric_KMapEstimationConfig_TaggedField) ProtoMessage() {} func (*PrivacyMetric_KMapEstimationConfig_TaggedField) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 4, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 4, 0} } func (m *PrivacyMetric_KMapEstimationConfig_TaggedField) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_KMapEstimationConfig_TaggedField.Unmarshal(m, b) @@ -3581,13 +3940,13 @@ type isPrivacyMetric_KMapEstimationConfig_TaggedField_Tag interface { } type PrivacyMetric_KMapEstimationConfig_TaggedField_InfoType struct { - InfoType *InfoType `protobuf:"bytes,2,opt,name=info_type,json=infoType,oneof"` + InfoType *InfoType `protobuf:"bytes,2,opt,name=info_type,json=infoType,proto3,oneof"` } type PrivacyMetric_KMapEstimationConfig_TaggedField_CustomTag struct { - CustomTag string `protobuf:"bytes,3,opt,name=custom_tag,json=customTag,oneof"` + CustomTag string `protobuf:"bytes,3,opt,name=custom_tag,json=customTag,proto3,oneof"` } type PrivacyMetric_KMapEstimationConfig_TaggedField_Inferred struct { - Inferred *empty.Empty `protobuf:"bytes,4,opt,name=inferred,oneof"` + Inferred *empty.Empty `protobuf:"bytes,4,opt,name=inferred,proto3,oneof"` } func (*PrivacyMetric_KMapEstimationConfig_TaggedField_InfoType) isPrivacyMetric_KMapEstimationConfig_TaggedField_Tag() { @@ -3730,13 +4089,13 @@ func _PrivacyMetric_KMapEstimationConfig_TaggedField_OneofSizer(msg proto.Messag // tuple is highly reidentifiable). type PrivacyMetric_KMapEstimationConfig_AuxiliaryTable struct { // Auxiliary table location. [required] - Table *BigQueryTable `protobuf:"bytes,3,opt,name=table" json:"table,omitempty"` + Table *BigQueryTable `protobuf:"bytes,3,opt,name=table,proto3" json:"table,omitempty"` // Quasi-identifier columns. [required] - QuasiIds []*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds" json:"quasi_ids,omitempty"` + QuasiIds []*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds,proto3" json:"quasi_ids,omitempty"` // The relative frequency column must contain a floating-point number // between 0 and 1 (inclusive). Null values are assumed to be zero. // [required] - RelativeFrequency *FieldId `protobuf:"bytes,2,opt,name=relative_frequency,json=relativeFrequency" json:"relative_frequency,omitempty"` + RelativeFrequency *FieldId `protobuf:"bytes,2,opt,name=relative_frequency,json=relativeFrequency,proto3" json:"relative_frequency,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3750,7 +4109,7 @@ func (m *PrivacyMetric_KMapEstimationConfig_AuxiliaryTable) String() string { } func (*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable) ProtoMessage() {} func (*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 4, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 4, 1} } func (m *PrivacyMetric_KMapEstimationConfig_AuxiliaryTable) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_KMapEstimationConfig_AuxiliaryTable.Unmarshal(m, b) @@ -3794,8 +4153,8 @@ func (m *PrivacyMetric_KMapEstimationConfig_AuxiliaryTable) GetRelativeFrequency // A quasi-identifier column has a custom_tag, used to know which column // in the data corresponds to which column in the statistical model. type PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField struct { - Field *FieldId `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` - CustomTag string `protobuf:"bytes,2,opt,name=custom_tag,json=customTag" json:"custom_tag,omitempty"` + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` + CustomTag string `protobuf:"bytes,2,opt,name=custom_tag,json=customTag,proto3" json:"custom_tag,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -3809,7 +4168,7 @@ func (m *PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField) String( } func (*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField) ProtoMessage() {} func (*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{30, 4, 1, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{32, 4, 1, 0} } func (m *PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField.Unmarshal(m, b) @@ -3843,12 +4202,82 @@ func (m *PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField) GetCust return "" } +// δ-presence metric, used to estimate how likely it is for an attacker to +// figure out that one given individual appears in a de-identified dataset. +// Similarly to the k-map metric, we cannot compute δ-presence exactly without +// knowing the attack dataset, so we use a statistical model instead. +type PrivacyMetric_DeltaPresenceEstimationConfig struct { + // Fields considered to be quasi-identifiers. No two fields can have the + // same tag. [required] + QuasiIds []*QuasiId `protobuf:"bytes,1,rep,name=quasi_ids,json=quasiIds,proto3" json:"quasi_ids,omitempty"` + // ISO 3166-1 alpha-2 region code to use in the statistical modeling. + // Required if no column is tagged with a region-specific InfoType (like + // US_ZIP_5) or a region code. + RegionCode string `protobuf:"bytes,2,opt,name=region_code,json=regionCode,proto3" json:"region_code,omitempty"` + // Several auxiliary tables can be used in the analysis. Each custom_tag + // used to tag a quasi-identifiers field must appear in exactly one + // field of one auxiliary table. + AuxiliaryTables []*StatisticalTable `protobuf:"bytes,3,rep,name=auxiliary_tables,json=auxiliaryTables,proto3" json:"auxiliary_tables,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) Reset() { + *m = PrivacyMetric_DeltaPresenceEstimationConfig{} +} +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) String() string { + return proto.CompactTextString(m) +} +func (*PrivacyMetric_DeltaPresenceEstimationConfig) ProtoMessage() {} +func (*PrivacyMetric_DeltaPresenceEstimationConfig) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{32, 5} +} +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_PrivacyMetric_DeltaPresenceEstimationConfig.Unmarshal(m, b) +} +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_PrivacyMetric_DeltaPresenceEstimationConfig.Marshal(b, m, deterministic) +} +func (dst *PrivacyMetric_DeltaPresenceEstimationConfig) XXX_Merge(src proto.Message) { + xxx_messageInfo_PrivacyMetric_DeltaPresenceEstimationConfig.Merge(dst, src) +} +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) XXX_Size() int { + return xxx_messageInfo_PrivacyMetric_DeltaPresenceEstimationConfig.Size(m) +} +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) XXX_DiscardUnknown() { + xxx_messageInfo_PrivacyMetric_DeltaPresenceEstimationConfig.DiscardUnknown(m) +} + +var xxx_messageInfo_PrivacyMetric_DeltaPresenceEstimationConfig proto.InternalMessageInfo + +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) GetQuasiIds() []*QuasiId { + if m != nil { + return m.QuasiIds + } + return nil +} + +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) GetRegionCode() string { + if m != nil { + return m.RegionCode + } + return "" +} + +func (m *PrivacyMetric_DeltaPresenceEstimationConfig) GetAuxiliaryTables() []*StatisticalTable { + if m != nil { + return m.AuxiliaryTables + } + return nil +} + // Result of a risk analysis operation request. type AnalyzeDataSourceRiskDetails struct { // Privacy metric to compute. - RequestedPrivacyMetric *PrivacyMetric `protobuf:"bytes,1,opt,name=requested_privacy_metric,json=requestedPrivacyMetric" json:"requested_privacy_metric,omitempty"` + RequestedPrivacyMetric *PrivacyMetric `protobuf:"bytes,1,opt,name=requested_privacy_metric,json=requestedPrivacyMetric,proto3" json:"requested_privacy_metric,omitempty"` // Input dataset to compute metrics over. - RequestedSourceTable *BigQueryTable `protobuf:"bytes,2,opt,name=requested_source_table,json=requestedSourceTable" json:"requested_source_table,omitempty"` + RequestedSourceTable *BigQueryTable `protobuf:"bytes,2,opt,name=requested_source_table,json=requestedSourceTable,proto3" json:"requested_source_table,omitempty"` // Values associated with this metric. // // Types that are valid to be assigned to Result: @@ -3857,6 +4286,7 @@ type AnalyzeDataSourceRiskDetails struct { // *AnalyzeDataSourceRiskDetails_KAnonymityResult_ // *AnalyzeDataSourceRiskDetails_LDiversityResult_ // *AnalyzeDataSourceRiskDetails_KMapEstimationResult_ + // *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_ Result isAnalyzeDataSourceRiskDetails_Result `protobuf_oneof:"result"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` @@ -3867,7 +4297,7 @@ func (m *AnalyzeDataSourceRiskDetails) Reset() { *m = AnalyzeDataSourceR func (m *AnalyzeDataSourceRiskDetails) String() string { return proto.CompactTextString(m) } func (*AnalyzeDataSourceRiskDetails) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31} + return fileDescriptor_dlp_536657b46c30097f, []int{33} } func (m *AnalyzeDataSourceRiskDetails) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails.Unmarshal(m, b) @@ -3892,19 +4322,22 @@ type isAnalyzeDataSourceRiskDetails_Result interface { } type AnalyzeDataSourceRiskDetails_NumericalStatsResult_ struct { - NumericalStatsResult *AnalyzeDataSourceRiskDetails_NumericalStatsResult `protobuf:"bytes,3,opt,name=numerical_stats_result,json=numericalStatsResult,oneof"` + NumericalStatsResult *AnalyzeDataSourceRiskDetails_NumericalStatsResult `protobuf:"bytes,3,opt,name=numerical_stats_result,json=numericalStatsResult,proto3,oneof"` } type AnalyzeDataSourceRiskDetails_CategoricalStatsResult_ struct { - CategoricalStatsResult *AnalyzeDataSourceRiskDetails_CategoricalStatsResult `protobuf:"bytes,4,opt,name=categorical_stats_result,json=categoricalStatsResult,oneof"` + CategoricalStatsResult *AnalyzeDataSourceRiskDetails_CategoricalStatsResult `protobuf:"bytes,4,opt,name=categorical_stats_result,json=categoricalStatsResult,proto3,oneof"` } type AnalyzeDataSourceRiskDetails_KAnonymityResult_ struct { - KAnonymityResult *AnalyzeDataSourceRiskDetails_KAnonymityResult `protobuf:"bytes,5,opt,name=k_anonymity_result,json=kAnonymityResult,oneof"` + KAnonymityResult *AnalyzeDataSourceRiskDetails_KAnonymityResult `protobuf:"bytes,5,opt,name=k_anonymity_result,json=kAnonymityResult,proto3,oneof"` } type AnalyzeDataSourceRiskDetails_LDiversityResult_ struct { - LDiversityResult *AnalyzeDataSourceRiskDetails_LDiversityResult `protobuf:"bytes,6,opt,name=l_diversity_result,json=lDiversityResult,oneof"` + LDiversityResult *AnalyzeDataSourceRiskDetails_LDiversityResult `protobuf:"bytes,6,opt,name=l_diversity_result,json=lDiversityResult,proto3,oneof"` } type AnalyzeDataSourceRiskDetails_KMapEstimationResult_ struct { - KMapEstimationResult *AnalyzeDataSourceRiskDetails_KMapEstimationResult `protobuf:"bytes,7,opt,name=k_map_estimation_result,json=kMapEstimationResult,oneof"` + KMapEstimationResult *AnalyzeDataSourceRiskDetails_KMapEstimationResult `protobuf:"bytes,7,opt,name=k_map_estimation_result,json=kMapEstimationResult,proto3,oneof"` +} +type AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_ struct { + DeltaPresenceEstimationResult *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult `protobuf:"bytes,9,opt,name=delta_presence_estimation_result,json=deltaPresenceEstimationResult,proto3,oneof"` } func (*AnalyzeDataSourceRiskDetails_NumericalStatsResult_) isAnalyzeDataSourceRiskDetails_Result() {} @@ -3912,6 +4345,8 @@ func (*AnalyzeDataSourceRiskDetails_CategoricalStatsResult_) isAnalyzeDataSource func (*AnalyzeDataSourceRiskDetails_KAnonymityResult_) isAnalyzeDataSourceRiskDetails_Result() {} func (*AnalyzeDataSourceRiskDetails_LDiversityResult_) isAnalyzeDataSourceRiskDetails_Result() {} func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult_) isAnalyzeDataSourceRiskDetails_Result() {} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_) isAnalyzeDataSourceRiskDetails_Result() { +} func (m *AnalyzeDataSourceRiskDetails) GetResult() isAnalyzeDataSourceRiskDetails_Result { if m != nil { @@ -3969,6 +4404,13 @@ func (m *AnalyzeDataSourceRiskDetails) GetKMapEstimationResult() *AnalyzeDataSou return nil } +func (m *AnalyzeDataSourceRiskDetails) GetDeltaPresenceEstimationResult() *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult { + if x, ok := m.GetResult().(*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_); ok { + return x.DeltaPresenceEstimationResult + } + return nil +} + // XXX_OneofFuncs is for the internal use of the proto package. func (*AnalyzeDataSourceRiskDetails) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { return _AnalyzeDataSourceRiskDetails_OneofMarshaler, _AnalyzeDataSourceRiskDetails_OneofUnmarshaler, _AnalyzeDataSourceRiskDetails_OneofSizer, []interface{}{ @@ -3977,6 +4419,7 @@ func (*AnalyzeDataSourceRiskDetails) XXX_OneofFuncs() (func(msg proto.Message, b (*AnalyzeDataSourceRiskDetails_KAnonymityResult_)(nil), (*AnalyzeDataSourceRiskDetails_LDiversityResult_)(nil), (*AnalyzeDataSourceRiskDetails_KMapEstimationResult_)(nil), + (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_)(nil), } } @@ -4009,6 +4452,11 @@ func _AnalyzeDataSourceRiskDetails_OneofMarshaler(msg proto.Message, b *proto.Bu if err := b.EncodeMessage(x.KMapEstimationResult); err != nil { return err } + case *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_: + b.EncodeVarint(9<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.DeltaPresenceEstimationResult); err != nil { + return err + } case nil: default: return fmt.Errorf("AnalyzeDataSourceRiskDetails.Result has unexpected type %T", x) @@ -4059,6 +4507,14 @@ func _AnalyzeDataSourceRiskDetails_OneofUnmarshaler(msg proto.Message, tag, wire err := b.DecodeMessage(msg) m.Result = &AnalyzeDataSourceRiskDetails_KMapEstimationResult_{msg} return true, err + case 9: // result.delta_presence_estimation_result + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) + err := b.DecodeMessage(msg) + m.Result = &AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_{msg} + return true, err default: return false, nil } @@ -4093,6 +4549,11 @@ func _AnalyzeDataSourceRiskDetails_OneofSizer(msg proto.Message) (n int) { n += 1 // tag and wire n += proto.SizeVarint(uint64(s)) n += s + case *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_: + s := proto.Size(x.DeltaPresenceEstimationResult) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s case nil: default: panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) @@ -4103,12 +4564,12 @@ func _AnalyzeDataSourceRiskDetails_OneofSizer(msg proto.Message) (n int) { // Result of the numerical stats computation. type AnalyzeDataSourceRiskDetails_NumericalStatsResult struct { // Minimum value appearing in the column. - MinValue *Value `protobuf:"bytes,1,opt,name=min_value,json=minValue" json:"min_value,omitempty"` + MinValue *Value `protobuf:"bytes,1,opt,name=min_value,json=minValue,proto3" json:"min_value,omitempty"` // Maximum value appearing in the column. - MaxValue *Value `protobuf:"bytes,2,opt,name=max_value,json=maxValue" json:"max_value,omitempty"` + MaxValue *Value `protobuf:"bytes,2,opt,name=max_value,json=maxValue,proto3" json:"max_value,omitempty"` // List of 99 values that partition the set of field values into 100 equal // sized buckets. - QuantileValues []*Value `protobuf:"bytes,4,rep,name=quantile_values,json=quantileValues" json:"quantile_values,omitempty"` + QuantileValues []*Value `protobuf:"bytes,4,rep,name=quantile_values,json=quantileValues,proto3" json:"quantile_values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4122,7 +4583,7 @@ func (m *AnalyzeDataSourceRiskDetails_NumericalStatsResult) String() string { } func (*AnalyzeDataSourceRiskDetails_NumericalStatsResult) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_NumericalStatsResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 0} } func (m *AnalyzeDataSourceRiskDetails_NumericalStatsResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_NumericalStatsResult.Unmarshal(m, b) @@ -4166,7 +4627,7 @@ func (m *AnalyzeDataSourceRiskDetails_NumericalStatsResult) GetQuantileValues() // Result of the categorical stats computation. type AnalyzeDataSourceRiskDetails_CategoricalStatsResult struct { // Histogram of value frequencies in the column. - ValueFrequencyHistogramBuckets []*AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket `protobuf:"bytes,5,rep,name=value_frequency_histogram_buckets,json=valueFrequencyHistogramBuckets" json:"value_frequency_histogram_buckets,omitempty"` + ValueFrequencyHistogramBuckets []*AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket `protobuf:"bytes,5,rep,name=value_frequency_histogram_buckets,json=valueFrequencyHistogramBuckets,proto3" json:"value_frequency_histogram_buckets,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4180,7 +4641,7 @@ func (m *AnalyzeDataSourceRiskDetails_CategoricalStatsResult) String() string { } func (*AnalyzeDataSourceRiskDetails_CategoricalStatsResult) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_CategoricalStatsResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 1} } func (m *AnalyzeDataSourceRiskDetails_CategoricalStatsResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_CategoricalStatsResult.Unmarshal(m, b) @@ -4209,16 +4670,16 @@ func (m *AnalyzeDataSourceRiskDetails_CategoricalStatsResult) GetValueFrequencyH type AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket struct { // Lower bound on the value frequency of the values in this bucket. - ValueFrequencyLowerBound int64 `protobuf:"varint,1,opt,name=value_frequency_lower_bound,json=valueFrequencyLowerBound" json:"value_frequency_lower_bound,omitempty"` + ValueFrequencyLowerBound int64 `protobuf:"varint,1,opt,name=value_frequency_lower_bound,json=valueFrequencyLowerBound,proto3" json:"value_frequency_lower_bound,omitempty"` // Upper bound on the value frequency of the values in this bucket. - ValueFrequencyUpperBound int64 `protobuf:"varint,2,opt,name=value_frequency_upper_bound,json=valueFrequencyUpperBound" json:"value_frequency_upper_bound,omitempty"` + ValueFrequencyUpperBound int64 `protobuf:"varint,2,opt,name=value_frequency_upper_bound,json=valueFrequencyUpperBound,proto3" json:"value_frequency_upper_bound,omitempty"` // Total number of values in this bucket. - BucketSize int64 `protobuf:"varint,3,opt,name=bucket_size,json=bucketSize" json:"bucket_size,omitempty"` + BucketSize int64 `protobuf:"varint,3,opt,name=bucket_size,json=bucketSize,proto3" json:"bucket_size,omitempty"` // Sample of value frequencies in this bucket. The total number of // values returned per bucket is capped at 20. - BucketValues []*ValueFrequency `protobuf:"bytes,4,rep,name=bucket_values,json=bucketValues" json:"bucket_values,omitempty"` + BucketValues []*ValueFrequency `protobuf:"bytes,4,rep,name=bucket_values,json=bucketValues,proto3" json:"bucket_values,omitempty"` // Total number of distinct values in this bucket. - BucketValueCount int64 `protobuf:"varint,5,opt,name=bucket_value_count,json=bucketValueCount" json:"bucket_value_count,omitempty"` + BucketValueCount int64 `protobuf:"varint,5,opt,name=bucket_value_count,json=bucketValueCount,proto3" json:"bucket_value_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4233,7 +4694,7 @@ func (m *AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHis func (*AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket) ProtoMessage() { } func (*AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 1, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 1, 0} } func (m *AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHistogramBucket.Unmarshal(m, b) @@ -4291,7 +4752,7 @@ func (m *AnalyzeDataSourceRiskDetails_CategoricalStatsResult_CategoricalStatsHis // Result of the k-anonymity computation. type AnalyzeDataSourceRiskDetails_KAnonymityResult struct { // Histogram of k-anonymity equivalence classes. - EquivalenceClassHistogramBuckets []*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket `protobuf:"bytes,5,rep,name=equivalence_class_histogram_buckets,json=equivalenceClassHistogramBuckets" json:"equivalence_class_histogram_buckets,omitempty"` + EquivalenceClassHistogramBuckets []*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket `protobuf:"bytes,5,rep,name=equivalence_class_histogram_buckets,json=equivalenceClassHistogramBuckets,proto3" json:"equivalence_class_histogram_buckets,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4305,7 +4766,7 @@ func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult) String() string { } func (*AnalyzeDataSourceRiskDetails_KAnonymityResult) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_KAnonymityResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 2} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 2} } func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_KAnonymityResult.Unmarshal(m, b) @@ -4337,10 +4798,10 @@ type AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass st // Set of values defining the equivalence class. One value per // quasi-identifier column in the original KAnonymity metric message. // The order is always the same as the original request. - QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues" json:"quasi_ids_values,omitempty"` + QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues,proto3" json:"quasi_ids_values,omitempty"` // Size of the equivalence class, for example number of rows with the // above set of values. - EquivalenceClassSize int64 `protobuf:"varint,2,opt,name=equivalence_class_size,json=equivalenceClassSize" json:"equivalence_class_size,omitempty"` + EquivalenceClassSize int64 `protobuf:"varint,2,opt,name=equivalence_class_size,json=equivalenceClassSize,proto3" json:"equivalence_class_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4354,7 +4815,7 @@ func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClas } func (*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 2, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 2, 0} } func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass.Unmarshal(m, b) @@ -4390,16 +4851,16 @@ func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClas type AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket struct { // Lower bound on the size of the equivalence classes in this bucket. - EquivalenceClassSizeLowerBound int64 `protobuf:"varint,1,opt,name=equivalence_class_size_lower_bound,json=equivalenceClassSizeLowerBound" json:"equivalence_class_size_lower_bound,omitempty"` + EquivalenceClassSizeLowerBound int64 `protobuf:"varint,1,opt,name=equivalence_class_size_lower_bound,json=equivalenceClassSizeLowerBound,proto3" json:"equivalence_class_size_lower_bound,omitempty"` // Upper bound on the size of the equivalence classes in this bucket. - EquivalenceClassSizeUpperBound int64 `protobuf:"varint,2,opt,name=equivalence_class_size_upper_bound,json=equivalenceClassSizeUpperBound" json:"equivalence_class_size_upper_bound,omitempty"` + EquivalenceClassSizeUpperBound int64 `protobuf:"varint,2,opt,name=equivalence_class_size_upper_bound,json=equivalenceClassSizeUpperBound,proto3" json:"equivalence_class_size_upper_bound,omitempty"` // Total number of equivalence classes in this bucket. - BucketSize int64 `protobuf:"varint,3,opt,name=bucket_size,json=bucketSize" json:"bucket_size,omitempty"` + BucketSize int64 `protobuf:"varint,3,opt,name=bucket_size,json=bucketSize,proto3" json:"bucket_size,omitempty"` // Sample of equivalence classes in this bucket. The total number of // classes returned per bucket is capped at 20. - BucketValues []*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass `protobuf:"bytes,4,rep,name=bucket_values,json=bucketValues" json:"bucket_values,omitempty"` + BucketValues []*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityEquivalenceClass `protobuf:"bytes,4,rep,name=bucket_values,json=bucketValues,proto3" json:"bucket_values,omitempty"` // Total number of distinct equivalence classes in this bucket. - BucketValueCount int64 `protobuf:"varint,5,opt,name=bucket_value_count,json=bucketValueCount" json:"bucket_value_count,omitempty"` + BucketValueCount int64 `protobuf:"varint,5,opt,name=bucket_value_count,json=bucketValueCount,proto3" json:"bucket_value_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4413,7 +4874,7 @@ func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket } func (*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 2, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 2, 1} } func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket.Unmarshal(m, b) @@ -4471,7 +4932,7 @@ func (m *AnalyzeDataSourceRiskDetails_KAnonymityResult_KAnonymityHistogramBucket // Result of the l-diversity computation. type AnalyzeDataSourceRiskDetails_LDiversityResult struct { // Histogram of l-diversity equivalence class sensitive value frequencies. - SensitiveValueFrequencyHistogramBuckets []*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket `protobuf:"bytes,5,rep,name=sensitive_value_frequency_histogram_buckets,json=sensitiveValueFrequencyHistogramBuckets" json:"sensitive_value_frequency_histogram_buckets,omitempty"` + SensitiveValueFrequencyHistogramBuckets []*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket `protobuf:"bytes,5,rep,name=sensitive_value_frequency_histogram_buckets,json=sensitiveValueFrequencyHistogramBuckets,proto3" json:"sensitive_value_frequency_histogram_buckets,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4485,7 +4946,7 @@ func (m *AnalyzeDataSourceRiskDetails_LDiversityResult) String() string { } func (*AnalyzeDataSourceRiskDetails_LDiversityResult) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_LDiversityResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 3} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 3} } func (m *AnalyzeDataSourceRiskDetails_LDiversityResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_LDiversityResult.Unmarshal(m, b) @@ -4516,13 +4977,13 @@ func (m *AnalyzeDataSourceRiskDetails_LDiversityResult) GetSensitiveValueFrequen type AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass struct { // Quasi-identifier values defining the k-anonymity equivalence // class. The order is always the same as the original request. - QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues" json:"quasi_ids_values,omitempty"` + QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues,proto3" json:"quasi_ids_values,omitempty"` // Size of the k-anonymity equivalence class. - EquivalenceClassSize int64 `protobuf:"varint,2,opt,name=equivalence_class_size,json=equivalenceClassSize" json:"equivalence_class_size,omitempty"` + EquivalenceClassSize int64 `protobuf:"varint,2,opt,name=equivalence_class_size,json=equivalenceClassSize,proto3" json:"equivalence_class_size,omitempty"` // Number of distinct sensitive values in this equivalence class. - NumDistinctSensitiveValues int64 `protobuf:"varint,3,opt,name=num_distinct_sensitive_values,json=numDistinctSensitiveValues" json:"num_distinct_sensitive_values,omitempty"` + NumDistinctSensitiveValues int64 `protobuf:"varint,3,opt,name=num_distinct_sensitive_values,json=numDistinctSensitiveValues,proto3" json:"num_distinct_sensitive_values,omitempty"` // Estimated frequencies of top sensitive values. - TopSensitiveValues []*ValueFrequency `protobuf:"bytes,4,rep,name=top_sensitive_values,json=topSensitiveValues" json:"top_sensitive_values,omitempty"` + TopSensitiveValues []*ValueFrequency `protobuf:"bytes,4,rep,name=top_sensitive_values,json=topSensitiveValues,proto3" json:"top_sensitive_values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4536,7 +4997,7 @@ func (m *AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClas } func (*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 3, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 3, 0} } func (m *AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass.Unmarshal(m, b) @@ -4587,17 +5048,17 @@ func (m *AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClas type AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket struct { // Lower bound on the sensitive value frequencies of the equivalence // classes in this bucket. - SensitiveValueFrequencyLowerBound int64 `protobuf:"varint,1,opt,name=sensitive_value_frequency_lower_bound,json=sensitiveValueFrequencyLowerBound" json:"sensitive_value_frequency_lower_bound,omitempty"` + SensitiveValueFrequencyLowerBound int64 `protobuf:"varint,1,opt,name=sensitive_value_frequency_lower_bound,json=sensitiveValueFrequencyLowerBound,proto3" json:"sensitive_value_frequency_lower_bound,omitempty"` // Upper bound on the sensitive value frequencies of the equivalence // classes in this bucket. - SensitiveValueFrequencyUpperBound int64 `protobuf:"varint,2,opt,name=sensitive_value_frequency_upper_bound,json=sensitiveValueFrequencyUpperBound" json:"sensitive_value_frequency_upper_bound,omitempty"` + SensitiveValueFrequencyUpperBound int64 `protobuf:"varint,2,opt,name=sensitive_value_frequency_upper_bound,json=sensitiveValueFrequencyUpperBound,proto3" json:"sensitive_value_frequency_upper_bound,omitempty"` // Total number of equivalence classes in this bucket. - BucketSize int64 `protobuf:"varint,3,opt,name=bucket_size,json=bucketSize" json:"bucket_size,omitempty"` + BucketSize int64 `protobuf:"varint,3,opt,name=bucket_size,json=bucketSize,proto3" json:"bucket_size,omitempty"` // Sample of equivalence classes in this bucket. The total number of // classes returned per bucket is capped at 20. - BucketValues []*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass `protobuf:"bytes,4,rep,name=bucket_values,json=bucketValues" json:"bucket_values,omitempty"` + BucketValues []*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityEquivalenceClass `protobuf:"bytes,4,rep,name=bucket_values,json=bucketValues,proto3" json:"bucket_values,omitempty"` // Total number of distinct equivalence classes in this bucket. - BucketValueCount int64 `protobuf:"varint,5,opt,name=bucket_value_count,json=bucketValueCount" json:"bucket_value_count,omitempty"` + BucketValueCount int64 `protobuf:"varint,5,opt,name=bucket_value_count,json=bucketValueCount,proto3" json:"bucket_value_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4611,7 +5072,7 @@ func (m *AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket } func (*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 3, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 3, 1} } func (m *AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_LDiversityResult_LDiversityHistogramBucket.Unmarshal(m, b) @@ -4677,7 +5138,7 @@ type AnalyzeDataSourceRiskDetails_KMapEstimationResult struct { // {min_anonymity: 5, max_anonymity: 10, frequency: 99} // mean that there are no record with an estimated anonymity of 4, 5, or // larger than 10. - KMapEstimationHistogram []*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket `protobuf:"bytes,1,rep,name=k_map_estimation_histogram,json=kMapEstimationHistogram" json:"k_map_estimation_histogram,omitempty"` + KMapEstimationHistogram []*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket `protobuf:"bytes,1,rep,name=k_map_estimation_histogram,json=kMapEstimationHistogram,proto3" json:"k_map_estimation_histogram,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4691,7 +5152,7 @@ func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult) String() string { } func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 4} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 4} } func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_KMapEstimationResult.Unmarshal(m, b) @@ -4721,9 +5182,9 @@ func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult) GetKMapEstimationHis // A tuple of values for the quasi-identifier columns. type AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues struct { // The quasi-identifier values. - QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues" json:"quasi_ids_values,omitempty"` + QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues,proto3" json:"quasi_ids_values,omitempty"` // The estimated anonymity for these quasi-identifier values. - EstimatedAnonymity int64 `protobuf:"varint,2,opt,name=estimated_anonymity,json=estimatedAnonymity" json:"estimated_anonymity,omitempty"` + EstimatedAnonymity int64 `protobuf:"varint,2,opt,name=estimated_anonymity,json=estimatedAnonymity,proto3" json:"estimated_anonymity,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4737,7 +5198,7 @@ func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiId } func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues) ProtoMessage() {} func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 4, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 4, 0} } func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues.Unmarshal(m, b) @@ -4781,16 +5242,16 @@ func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiId // corresponds to the number of uniquely identifiable records. type AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket struct { // Always positive. - MinAnonymity int64 `protobuf:"varint,1,opt,name=min_anonymity,json=minAnonymity" json:"min_anonymity,omitempty"` + MinAnonymity int64 `protobuf:"varint,1,opt,name=min_anonymity,json=minAnonymity,proto3" json:"min_anonymity,omitempty"` // Always greater than or equal to min_anonymity. - MaxAnonymity int64 `protobuf:"varint,2,opt,name=max_anonymity,json=maxAnonymity" json:"max_anonymity,omitempty"` + MaxAnonymity int64 `protobuf:"varint,2,opt,name=max_anonymity,json=maxAnonymity,proto3" json:"max_anonymity,omitempty"` // Number of records within these anonymity bounds. - BucketSize int64 `protobuf:"varint,5,opt,name=bucket_size,json=bucketSize" json:"bucket_size,omitempty"` + BucketSize int64 `protobuf:"varint,5,opt,name=bucket_size,json=bucketSize,proto3" json:"bucket_size,omitempty"` // Sample of quasi-identifier tuple values in this bucket. The total // number of classes returned per bucket is capped at 20. - BucketValues []*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues `protobuf:"bytes,6,rep,name=bucket_values,json=bucketValues" json:"bucket_values,omitempty"` + BucketValues []*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues `protobuf:"bytes,6,rep,name=bucket_values,json=bucketValues,proto3" json:"bucket_values,omitempty"` // Total number of distinct quasi-identifier tuple values in this bucket. - BucketValueCount int64 `protobuf:"varint,7,opt,name=bucket_value_count,json=bucketValueCount" json:"bucket_value_count,omitempty"` + BucketValueCount int64 `protobuf:"varint,7,opt,name=bucket_value_count,json=bucketValueCount,proto3" json:"bucket_value_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4805,7 +5266,7 @@ func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogr func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket) ProtoMessage() { } func (*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{31, 4, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{33, 4, 1} } func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket.Unmarshal(m, b) @@ -4860,12 +5321,215 @@ func (m *AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogr return 0 } +// Result of the δ-presence computation. Note that these results are an +// estimation, not exact values. +type AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult struct { + // The intervals [min_probability, max_probability) do not overlap. If a + // value doesn't correspond to any such interval, the associated frequency + // is zero. For example, the following records: + // {min_probability: 0, max_probability: 0.1, frequency: 17} + // {min_probability: 0.2, max_probability: 0.3, frequency: 42} + // {min_probability: 0.3, max_probability: 0.4, frequency: 99} + // mean that there are no record with an estimated probability in [0.1, 0.2) + // nor larger or equal to 0.4. + DeltaPresenceEstimationHistogram []*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket `protobuf:"bytes,1,rep,name=delta_presence_estimation_histogram,json=deltaPresenceEstimationHistogram,proto3" json:"delta_presence_estimation_histogram,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) Reset() { + *m = AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult{} +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) String() string { + return proto.CompactTextString(m) +} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) ProtoMessage() {} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{33, 5} +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult.Unmarshal(m, b) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult.Marshal(b, m, deterministic) +} +func (dst *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) XXX_Merge(src proto.Message) { + xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult.Merge(dst, src) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) XXX_Size() int { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult.Size(m) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) XXX_DiscardUnknown() { + xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult.DiscardUnknown(m) +} + +var xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult proto.InternalMessageInfo + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult) GetDeltaPresenceEstimationHistogram() []*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket { + if m != nil { + return m.DeltaPresenceEstimationHistogram + } + return nil +} + +// A tuple of values for the quasi-identifier columns. +type AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues struct { + // The quasi-identifier values. + QuasiIdsValues []*Value `protobuf:"bytes,1,rep,name=quasi_ids_values,json=quasiIdsValues,proto3" json:"quasi_ids_values,omitempty"` + // The estimated probability that a given individual sharing these + // quasi-identifier values is in the dataset. This value, typically called + // δ, is the ratio between the number of records in the dataset with these + // quasi-identifier values, and the total number of individuals (inside + // *and* outside the dataset) with these quasi-identifier values. + // For example, if there are 15 individuals in the dataset who share the + // same quasi-identifier values, and an estimated 100 people in the entire + // population with these values, then δ is 0.15. + EstimatedProbability float64 `protobuf:"fixed64,2,opt,name=estimated_probability,json=estimatedProbability,proto3" json:"estimated_probability,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) Reset() { + *m = AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues{} +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) String() string { + return proto.CompactTextString(m) +} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) ProtoMessage() { +} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{33, 5, 0} +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues.Unmarshal(m, b) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues.Marshal(b, m, deterministic) +} +func (dst *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) XXX_Merge(src proto.Message) { + xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues.Merge(dst, src) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) XXX_Size() int { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues.Size(m) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) XXX_DiscardUnknown() { + xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues.DiscardUnknown(m) +} + +var xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues proto.InternalMessageInfo + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) GetQuasiIdsValues() []*Value { + if m != nil { + return m.QuasiIdsValues + } + return nil +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues) GetEstimatedProbability() float64 { + if m != nil { + return m.EstimatedProbability + } + return 0 +} + +// A DeltaPresenceEstimationHistogramBucket message with the following +// values: +// min_probability: 0.1 +// max_probability: 0.2 +// frequency: 42 +// means that there are 42 records for which δ is in [0.1, 0.2). An +// important particular case is when min_probability = max_probability = 1: +// then, every individual who shares this quasi-identifier combination is in +// the dataset. +type AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket struct { + // Between 0 and 1. + MinProbability float64 `protobuf:"fixed64,1,opt,name=min_probability,json=minProbability,proto3" json:"min_probability,omitempty"` + // Always greater than or equal to min_probability. + MaxProbability float64 `protobuf:"fixed64,2,opt,name=max_probability,json=maxProbability,proto3" json:"max_probability,omitempty"` + // Number of records within these probability bounds. + BucketSize int64 `protobuf:"varint,5,opt,name=bucket_size,json=bucketSize,proto3" json:"bucket_size,omitempty"` + // Sample of quasi-identifier tuple values in this bucket. The total + // number of classes returned per bucket is capped at 20. + BucketValues []*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues `protobuf:"bytes,6,rep,name=bucket_values,json=bucketValues,proto3" json:"bucket_values,omitempty"` + // Total number of distinct quasi-identifier tuple values in this bucket. + BucketValueCount int64 `protobuf:"varint,7,opt,name=bucket_value_count,json=bucketValueCount,proto3" json:"bucket_value_count,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) Reset() { + *m = AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket{} +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) String() string { + return proto.CompactTextString(m) +} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) ProtoMessage() { +} +func (*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) Descriptor() ([]byte, []int) { + return fileDescriptor_dlp_536657b46c30097f, []int{33, 5, 1} +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket.Unmarshal(m, b) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket.Marshal(b, m, deterministic) +} +func (dst *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) XXX_Merge(src proto.Message) { + xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket.Merge(dst, src) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) XXX_Size() int { + return xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket.Size(m) +} +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) XXX_DiscardUnknown() { + xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket.DiscardUnknown(m) +} + +var xxx_messageInfo_AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket proto.InternalMessageInfo + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) GetMinProbability() float64 { + if m != nil { + return m.MinProbability + } + return 0 +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) GetMaxProbability() float64 { + if m != nil { + return m.MaxProbability + } + return 0 +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) GetBucketSize() int64 { + if m != nil { + return m.BucketSize + } + return 0 +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) GetBucketValues() []*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues { + if m != nil { + return m.BucketValues + } + return nil +} + +func (m *AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket) GetBucketValueCount() int64 { + if m != nil { + return m.BucketValueCount + } + return 0 +} + // A value of a field, including its frequency. type ValueFrequency struct { // A value contained in the field in question. - Value *Value `protobuf:"bytes,1,opt,name=value" json:"value,omitempty"` + Value *Value `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` // How many times the value is contained in the field. - Count int64 `protobuf:"varint,2,opt,name=count" json:"count,omitempty"` + Count int64 `protobuf:"varint,2,opt,name=count,proto3" json:"count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -4875,7 +5539,7 @@ func (m *ValueFrequency) Reset() { *m = ValueFrequency{} } func (m *ValueFrequency) String() string { return proto.CompactTextString(m) } func (*ValueFrequency) ProtoMessage() {} func (*ValueFrequency) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{32} + return fileDescriptor_dlp_536657b46c30097f, []int{34} } func (m *ValueFrequency) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ValueFrequency.Unmarshal(m, b) @@ -4935,7 +5599,7 @@ func (m *Value) Reset() { *m = Value{} } func (m *Value) String() string { return proto.CompactTextString(m) } func (*Value) ProtoMessage() {} func (*Value) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{33} + return fileDescriptor_dlp_536657b46c30097f, []int{35} } func (m *Value) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Value.Unmarshal(m, b) @@ -4960,28 +5624,28 @@ type isValue_Type interface { } type Value_IntegerValue struct { - IntegerValue int64 `protobuf:"varint,1,opt,name=integer_value,json=integerValue,oneof"` + IntegerValue int64 `protobuf:"varint,1,opt,name=integer_value,json=integerValue,proto3,oneof"` } type Value_FloatValue struct { - FloatValue float64 `protobuf:"fixed64,2,opt,name=float_value,json=floatValue,oneof"` + FloatValue float64 `protobuf:"fixed64,2,opt,name=float_value,json=floatValue,proto3,oneof"` } type Value_StringValue struct { - StringValue string `protobuf:"bytes,3,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,3,opt,name=string_value,json=stringValue,proto3,oneof"` } type Value_BooleanValue struct { - BooleanValue bool `protobuf:"varint,4,opt,name=boolean_value,json=booleanValue,oneof"` + BooleanValue bool `protobuf:"varint,4,opt,name=boolean_value,json=booleanValue,proto3,oneof"` } type Value_TimestampValue struct { - TimestampValue *timestamp.Timestamp `protobuf:"bytes,5,opt,name=timestamp_value,json=timestampValue,oneof"` + TimestampValue *timestamp.Timestamp `protobuf:"bytes,5,opt,name=timestamp_value,json=timestampValue,proto3,oneof"` } type Value_TimeValue struct { - TimeValue *timeofday.TimeOfDay `protobuf:"bytes,6,opt,name=time_value,json=timeValue,oneof"` + TimeValue *timeofday.TimeOfDay `protobuf:"bytes,6,opt,name=time_value,json=timeValue,proto3,oneof"` } type Value_DateValue struct { - DateValue *date.Date `protobuf:"bytes,7,opt,name=date_value,json=dateValue,oneof"` + DateValue *date.Date `protobuf:"bytes,7,opt,name=date_value,json=dateValue,proto3,oneof"` } type Value_DayOfWeekValue struct { - DayOfWeekValue dayofweek.DayOfWeek `protobuf:"varint,8,opt,name=day_of_week_value,json=dayOfWeekValue,enum=google.type.DayOfWeek,oneof"` + DayOfWeekValue dayofweek.DayOfWeek `protobuf:"varint,8,opt,name=day_of_week_value,json=dayOfWeekValue,proto3,enum=google.type.DayOfWeek,oneof"` } func (*Value_IntegerValue) isValue_Type() {} @@ -5240,7 +5904,7 @@ func (m *QuoteInfo) Reset() { *m = QuoteInfo{} } func (m *QuoteInfo) String() string { return proto.CompactTextString(m) } func (*QuoteInfo) ProtoMessage() {} func (*QuoteInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{34} + return fileDescriptor_dlp_536657b46c30097f, []int{36} } func (m *QuoteInfo) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_QuoteInfo.Unmarshal(m, b) @@ -5265,7 +5929,7 @@ type isQuoteInfo_ParsedQuote interface { } type QuoteInfo_DateTime struct { - DateTime *DateTime `protobuf:"bytes,2,opt,name=date_time,json=dateTime,oneof"` + DateTime *DateTime `protobuf:"bytes,2,opt,name=date_time,json=dateTime,proto3,oneof"` } func (*QuoteInfo_DateTime) isQuoteInfo_ParsedQuote() {} @@ -5343,10 +6007,10 @@ func _QuoteInfo_OneofSizer(msg proto.Message) (n int) { type DateTime struct { // One or more of the following must be set. All fields are optional, but // when set must be valid date or time values. - Date *date.Date `protobuf:"bytes,1,opt,name=date" json:"date,omitempty"` - DayOfWeek dayofweek.DayOfWeek `protobuf:"varint,2,opt,name=day_of_week,json=dayOfWeek,enum=google.type.DayOfWeek" json:"day_of_week,omitempty"` - Time *timeofday.TimeOfDay `protobuf:"bytes,3,opt,name=time" json:"time,omitempty"` - TimeZone *DateTime_TimeZone `protobuf:"bytes,4,opt,name=time_zone,json=timeZone" json:"time_zone,omitempty"` + Date *date.Date `protobuf:"bytes,1,opt,name=date,proto3" json:"date,omitempty"` + DayOfWeek dayofweek.DayOfWeek `protobuf:"varint,2,opt,name=day_of_week,json=dayOfWeek,proto3,enum=google.type.DayOfWeek" json:"day_of_week,omitempty"` + Time *timeofday.TimeOfDay `protobuf:"bytes,3,opt,name=time,proto3" json:"time,omitempty"` + TimeZone *DateTime_TimeZone `protobuf:"bytes,4,opt,name=time_zone,json=timeZone,proto3" json:"time_zone,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5356,7 +6020,7 @@ func (m *DateTime) Reset() { *m = DateTime{} } func (m *DateTime) String() string { return proto.CompactTextString(m) } func (*DateTime) ProtoMessage() {} func (*DateTime) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{35} + return fileDescriptor_dlp_536657b46c30097f, []int{37} } func (m *DateTime) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DateTime.Unmarshal(m, b) @@ -5407,7 +6071,7 @@ func (m *DateTime) GetTimeZone() *DateTime_TimeZone { type DateTime_TimeZone struct { // Set only if the offset can be determined. Positive for time ahead of UTC. // E.g. For "UTC-9", this value is -540. - OffsetMinutes int32 `protobuf:"varint,1,opt,name=offset_minutes,json=offsetMinutes" json:"offset_minutes,omitempty"` + OffsetMinutes int32 `protobuf:"varint,1,opt,name=offset_minutes,json=offsetMinutes,proto3" json:"offset_minutes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5417,7 +6081,7 @@ func (m *DateTime_TimeZone) Reset() { *m = DateTime_TimeZone{} } func (m *DateTime_TimeZone) String() string { return proto.CompactTextString(m) } func (*DateTime_TimeZone) ProtoMessage() {} func (*DateTime_TimeZone) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{35, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{37, 0} } func (m *DateTime_TimeZone) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DateTime_TimeZone.Unmarshal(m, b) @@ -5459,7 +6123,7 @@ func (m *DeidentifyConfig) Reset() { *m = DeidentifyConfig{} } func (m *DeidentifyConfig) String() string { return proto.CompactTextString(m) } func (*DeidentifyConfig) ProtoMessage() {} func (*DeidentifyConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{36} + return fileDescriptor_dlp_536657b46c30097f, []int{38} } func (m *DeidentifyConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeidentifyConfig.Unmarshal(m, b) @@ -5484,10 +6148,10 @@ type isDeidentifyConfig_Transformation interface { } type DeidentifyConfig_InfoTypeTransformations struct { - InfoTypeTransformations *InfoTypeTransformations `protobuf:"bytes,1,opt,name=info_type_transformations,json=infoTypeTransformations,oneof"` + InfoTypeTransformations *InfoTypeTransformations `protobuf:"bytes,1,opt,name=info_type_transformations,json=infoTypeTransformations,proto3,oneof"` } type DeidentifyConfig_RecordTransformations struct { - RecordTransformations *RecordTransformations `protobuf:"bytes,2,opt,name=record_transformations,json=recordTransformations,oneof"` + RecordTransformations *RecordTransformations `protobuf:"bytes,2,opt,name=record_transformations,json=recordTransformations,proto3,oneof"` } func (*DeidentifyConfig_InfoTypeTransformations) isDeidentifyConfig_Transformation() {} @@ -5611,7 +6275,7 @@ func (m *PrimitiveTransformation) Reset() { *m = PrimitiveTransformation func (m *PrimitiveTransformation) String() string { return proto.CompactTextString(m) } func (*PrimitiveTransformation) ProtoMessage() {} func (*PrimitiveTransformation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{37} + return fileDescriptor_dlp_536657b46c30097f, []int{39} } func (m *PrimitiveTransformation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PrimitiveTransformation.Unmarshal(m, b) @@ -5636,34 +6300,34 @@ type isPrimitiveTransformation_Transformation interface { } type PrimitiveTransformation_ReplaceConfig struct { - ReplaceConfig *ReplaceValueConfig `protobuf:"bytes,1,opt,name=replace_config,json=replaceConfig,oneof"` + ReplaceConfig *ReplaceValueConfig `protobuf:"bytes,1,opt,name=replace_config,json=replaceConfig,proto3,oneof"` } type PrimitiveTransformation_RedactConfig struct { - RedactConfig *RedactConfig `protobuf:"bytes,2,opt,name=redact_config,json=redactConfig,oneof"` + RedactConfig *RedactConfig `protobuf:"bytes,2,opt,name=redact_config,json=redactConfig,proto3,oneof"` } type PrimitiveTransformation_CharacterMaskConfig struct { - CharacterMaskConfig *CharacterMaskConfig `protobuf:"bytes,3,opt,name=character_mask_config,json=characterMaskConfig,oneof"` + CharacterMaskConfig *CharacterMaskConfig `protobuf:"bytes,3,opt,name=character_mask_config,json=characterMaskConfig,proto3,oneof"` } type PrimitiveTransformation_CryptoReplaceFfxFpeConfig struct { - CryptoReplaceFfxFpeConfig *CryptoReplaceFfxFpeConfig `protobuf:"bytes,4,opt,name=crypto_replace_ffx_fpe_config,json=cryptoReplaceFfxFpeConfig,oneof"` + CryptoReplaceFfxFpeConfig *CryptoReplaceFfxFpeConfig `protobuf:"bytes,4,opt,name=crypto_replace_ffx_fpe_config,json=cryptoReplaceFfxFpeConfig,proto3,oneof"` } type PrimitiveTransformation_FixedSizeBucketingConfig struct { - FixedSizeBucketingConfig *FixedSizeBucketingConfig `protobuf:"bytes,5,opt,name=fixed_size_bucketing_config,json=fixedSizeBucketingConfig,oneof"` + FixedSizeBucketingConfig *FixedSizeBucketingConfig `protobuf:"bytes,5,opt,name=fixed_size_bucketing_config,json=fixedSizeBucketingConfig,proto3,oneof"` } type PrimitiveTransformation_BucketingConfig struct { - BucketingConfig *BucketingConfig `protobuf:"bytes,6,opt,name=bucketing_config,json=bucketingConfig,oneof"` + BucketingConfig *BucketingConfig `protobuf:"bytes,6,opt,name=bucketing_config,json=bucketingConfig,proto3,oneof"` } type PrimitiveTransformation_ReplaceWithInfoTypeConfig struct { - ReplaceWithInfoTypeConfig *ReplaceWithInfoTypeConfig `protobuf:"bytes,7,opt,name=replace_with_info_type_config,json=replaceWithInfoTypeConfig,oneof"` + ReplaceWithInfoTypeConfig *ReplaceWithInfoTypeConfig `protobuf:"bytes,7,opt,name=replace_with_info_type_config,json=replaceWithInfoTypeConfig,proto3,oneof"` } type PrimitiveTransformation_TimePartConfig struct { - TimePartConfig *TimePartConfig `protobuf:"bytes,8,opt,name=time_part_config,json=timePartConfig,oneof"` + TimePartConfig *TimePartConfig `protobuf:"bytes,8,opt,name=time_part_config,json=timePartConfig,proto3,oneof"` } type PrimitiveTransformation_CryptoHashConfig struct { - CryptoHashConfig *CryptoHashConfig `protobuf:"bytes,9,opt,name=crypto_hash_config,json=cryptoHashConfig,oneof"` + CryptoHashConfig *CryptoHashConfig `protobuf:"bytes,9,opt,name=crypto_hash_config,json=cryptoHashConfig,proto3,oneof"` } type PrimitiveTransformation_DateShiftConfig struct { - DateShiftConfig *DateShiftConfig `protobuf:"bytes,11,opt,name=date_shift_config,json=dateShiftConfig,oneof"` + DateShiftConfig *DateShiftConfig `protobuf:"bytes,11,opt,name=date_shift_config,json=dateShiftConfig,proto3,oneof"` } func (*PrimitiveTransformation_ReplaceConfig) isPrimitiveTransformation_Transformation() {} @@ -5983,7 +6647,7 @@ func _PrimitiveTransformation_OneofSizer(msg proto.Message) (n int) { // For use with `Date`, `Timestamp`, and `TimeOfDay`, extract or preserve a // portion of the value. type TimePartConfig struct { - PartToExtract TimePartConfig_TimePart `protobuf:"varint,1,opt,name=part_to_extract,json=partToExtract,enum=google.privacy.dlp.v2.TimePartConfig_TimePart" json:"part_to_extract,omitempty"` + PartToExtract TimePartConfig_TimePart `protobuf:"varint,1,opt,name=part_to_extract,json=partToExtract,proto3,enum=google.privacy.dlp.v2.TimePartConfig_TimePart" json:"part_to_extract,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -5993,7 +6657,7 @@ func (m *TimePartConfig) Reset() { *m = TimePartConfig{} } func (m *TimePartConfig) String() string { return proto.CompactTextString(m) } func (*TimePartConfig) ProtoMessage() {} func (*TimePartConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{38} + return fileDescriptor_dlp_536657b46c30097f, []int{40} } func (m *TimePartConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TimePartConfig.Unmarshal(m, b) @@ -6028,7 +6692,7 @@ func (m *TimePartConfig) GetPartToExtract() TimePartConfig_TimePart { // Currently, only string and integer values can be hashed. type CryptoHashConfig struct { // The key used by the hash function. - CryptoKey *CryptoKey `protobuf:"bytes,1,opt,name=crypto_key,json=cryptoKey" json:"crypto_key,omitempty"` + CryptoKey *CryptoKey `protobuf:"bytes,1,opt,name=crypto_key,json=cryptoKey,proto3" json:"crypto_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6038,7 +6702,7 @@ func (m *CryptoHashConfig) Reset() { *m = CryptoHashConfig{} } func (m *CryptoHashConfig) String() string { return proto.CompactTextString(m) } func (*CryptoHashConfig) ProtoMessage() {} func (*CryptoHashConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{39} + return fileDescriptor_dlp_536657b46c30097f, []int{41} } func (m *CryptoHashConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CryptoHashConfig.Unmarshal(m, b) @@ -6068,7 +6732,7 @@ func (m *CryptoHashConfig) GetCryptoKey() *CryptoKey { // Replace each input value with a given `Value`. type ReplaceValueConfig struct { // Value to replace it with. - NewValue *Value `protobuf:"bytes,1,opt,name=new_value,json=newValue" json:"new_value,omitempty"` + NewValue *Value `protobuf:"bytes,1,opt,name=new_value,json=newValue,proto3" json:"new_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6078,7 +6742,7 @@ func (m *ReplaceValueConfig) Reset() { *m = ReplaceValueConfig{} } func (m *ReplaceValueConfig) String() string { return proto.CompactTextString(m) } func (*ReplaceValueConfig) ProtoMessage() {} func (*ReplaceValueConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{40} + return fileDescriptor_dlp_536657b46c30097f, []int{42} } func (m *ReplaceValueConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReplaceValueConfig.Unmarshal(m, b) @@ -6116,7 +6780,7 @@ func (m *ReplaceWithInfoTypeConfig) Reset() { *m = ReplaceWithInfoTypeCo func (m *ReplaceWithInfoTypeConfig) String() string { return proto.CompactTextString(m) } func (*ReplaceWithInfoTypeConfig) ProtoMessage() {} func (*ReplaceWithInfoTypeConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{41} + return fileDescriptor_dlp_536657b46c30097f, []int{43} } func (m *ReplaceWithInfoTypeConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReplaceWithInfoTypeConfig.Unmarshal(m, b) @@ -6149,7 +6813,7 @@ func (m *RedactConfig) Reset() { *m = RedactConfig{} } func (m *RedactConfig) String() string { return proto.CompactTextString(m) } func (*RedactConfig) ProtoMessage() {} func (*RedactConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{42} + return fileDescriptor_dlp_536657b46c30097f, []int{44} } func (m *RedactConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RedactConfig.Unmarshal(m, b) @@ -6185,7 +6849,7 @@ func (m *CharsToIgnore) Reset() { *m = CharsToIgnore{} } func (m *CharsToIgnore) String() string { return proto.CompactTextString(m) } func (*CharsToIgnore) ProtoMessage() {} func (*CharsToIgnore) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{43} + return fileDescriptor_dlp_536657b46c30097f, []int{45} } func (m *CharsToIgnore) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CharsToIgnore.Unmarshal(m, b) @@ -6210,10 +6874,10 @@ type isCharsToIgnore_Characters interface { } type CharsToIgnore_CharactersToSkip struct { - CharactersToSkip string `protobuf:"bytes,1,opt,name=characters_to_skip,json=charactersToSkip,oneof"` + CharactersToSkip string `protobuf:"bytes,1,opt,name=characters_to_skip,json=charactersToSkip,proto3,oneof"` } type CharsToIgnore_CommonCharactersToIgnore struct { - CommonCharactersToIgnore CharsToIgnore_CommonCharsToIgnore `protobuf:"varint,2,opt,name=common_characters_to_ignore,json=commonCharactersToIgnore,enum=google.privacy.dlp.v2.CharsToIgnore_CommonCharsToIgnore,oneof"` + CommonCharactersToIgnore CharsToIgnore_CommonCharsToIgnore `protobuf:"varint,2,opt,name=common_characters_to_ignore,json=commonCharactersToIgnore,proto3,enum=google.privacy.dlp.v2.CharsToIgnore_CommonCharsToIgnore,oneof"` } func (*CharsToIgnore_CharactersToSkip) isCharsToIgnore_Characters() {} @@ -6316,20 +6980,20 @@ type CharacterMaskConfig struct { // alphabetic string such as name, or "0" for a numeric string such as ZIP // code or credit card number. String must have length 1. If not supplied, we // will default to "*" for strings, 0 for digits. - MaskingCharacter string `protobuf:"bytes,1,opt,name=masking_character,json=maskingCharacter" json:"masking_character,omitempty"` + MaskingCharacter string `protobuf:"bytes,1,opt,name=masking_character,json=maskingCharacter,proto3" json:"masking_character,omitempty"` // Number of characters to mask. If not set, all matching chars will be // masked. Skipped characters do not count towards this tally. - NumberToMask int32 `protobuf:"varint,2,opt,name=number_to_mask,json=numberToMask" json:"number_to_mask,omitempty"` + NumberToMask int32 `protobuf:"varint,2,opt,name=number_to_mask,json=numberToMask,proto3" json:"number_to_mask,omitempty"` // Mask characters in reverse order. For example, if `masking_character` is // '0', number_to_mask is 14, and `reverse_order` is false, then // 1234-5678-9012-3456 -> 00000000000000-3456 // If `masking_character` is '*', `number_to_mask` is 3, and `reverse_order` // is true, then 12345 -> 12*** - ReverseOrder bool `protobuf:"varint,3,opt,name=reverse_order,json=reverseOrder" json:"reverse_order,omitempty"` + ReverseOrder bool `protobuf:"varint,3,opt,name=reverse_order,json=reverseOrder,proto3" json:"reverse_order,omitempty"` // When masking a string, items in this list will be skipped when replacing. // For example, if your string is 555-555-5555 and you ask us to skip `-` and // mask 5 chars with * we would produce ***-*55-5555. - CharactersToIgnore []*CharsToIgnore `protobuf:"bytes,4,rep,name=characters_to_ignore,json=charactersToIgnore" json:"characters_to_ignore,omitempty"` + CharactersToIgnore []*CharsToIgnore `protobuf:"bytes,4,rep,name=characters_to_ignore,json=charactersToIgnore,proto3" json:"characters_to_ignore,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6339,7 +7003,7 @@ func (m *CharacterMaskConfig) Reset() { *m = CharacterMaskConfig{} } func (m *CharacterMaskConfig) String() string { return proto.CompactTextString(m) } func (*CharacterMaskConfig) ProtoMessage() {} func (*CharacterMaskConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{44} + return fileDescriptor_dlp_536657b46c30097f, []int{46} } func (m *CharacterMaskConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CharacterMaskConfig.Unmarshal(m, b) @@ -6401,21 +7065,23 @@ func (m *CharacterMaskConfig) GetCharactersToIgnore() []*CharsToIgnore { // If the bound Value type differs from the type of data // being transformed, we will first attempt converting the type of the data to // be transformed to match the type of the bound before comparing. +// +// See https://cloud.google.com/dlp/docs/concepts-bucketing to learn more. type FixedSizeBucketingConfig struct { // Lower bound value of buckets. All values less than `lower_bound` are // grouped together into a single bucket; for example if `lower_bound` = 10, // then all values less than 10 are replaced with the value “-10”. [Required]. - LowerBound *Value `protobuf:"bytes,1,opt,name=lower_bound,json=lowerBound" json:"lower_bound,omitempty"` + LowerBound *Value `protobuf:"bytes,1,opt,name=lower_bound,json=lowerBound,proto3" json:"lower_bound,omitempty"` // Upper bound value of buckets. All values greater than upper_bound are // grouped together into a single bucket; for example if `upper_bound` = 89, // then all values greater than 89 are replaced with the value “89+”. // [Required]. - UpperBound *Value `protobuf:"bytes,2,opt,name=upper_bound,json=upperBound" json:"upper_bound,omitempty"` + UpperBound *Value `protobuf:"bytes,2,opt,name=upper_bound,json=upperBound,proto3" json:"upper_bound,omitempty"` // Size of each bucket (except for minimum and maximum buckets). So if // `lower_bound` = 10, `upper_bound` = 89, and `bucket_size` = 10, then the // following buckets would be used: -10, 10-20, 20-30, 30-40, 40-50, 50-60, // 60-70, 70-80, 80-89, 89+. Precision up to 2 decimals works. [Required]. - BucketSize float64 `protobuf:"fixed64,3,opt,name=bucket_size,json=bucketSize" json:"bucket_size,omitempty"` + BucketSize float64 `protobuf:"fixed64,3,opt,name=bucket_size,json=bucketSize,proto3" json:"bucket_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6425,7 +7091,7 @@ func (m *FixedSizeBucketingConfig) Reset() { *m = FixedSizeBucketingConf func (m *FixedSizeBucketingConfig) String() string { return proto.CompactTextString(m) } func (*FixedSizeBucketingConfig) ProtoMessage() {} func (*FixedSizeBucketingConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{45} + return fileDescriptor_dlp_536657b46c30097f, []int{47} } func (m *FixedSizeBucketingConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FixedSizeBucketingConfig.Unmarshal(m, b) @@ -6474,9 +7140,10 @@ func (m *FixedSizeBucketingConfig) GetBucketSize() float64 { // If the bound `Value` type differs from the type of data being transformed, we // will first attempt converting the type of the data to be transformed to match // the type of the bound before comparing. +// See https://cloud.google.com/dlp/docs/concepts-bucketing to learn more. type BucketingConfig struct { // Set of buckets. Ranges must be non-overlapping. - Buckets []*BucketingConfig_Bucket `protobuf:"bytes,1,rep,name=buckets" json:"buckets,omitempty"` + Buckets []*BucketingConfig_Bucket `protobuf:"bytes,1,rep,name=buckets,proto3" json:"buckets,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6486,7 +7153,7 @@ func (m *BucketingConfig) Reset() { *m = BucketingConfig{} } func (m *BucketingConfig) String() string { return proto.CompactTextString(m) } func (*BucketingConfig) ProtoMessage() {} func (*BucketingConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{46} + return fileDescriptor_dlp_536657b46c30097f, []int{48} } func (m *BucketingConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BucketingConfig.Unmarshal(m, b) @@ -6517,12 +7184,12 @@ func (m *BucketingConfig) GetBuckets() []*BucketingConfig_Bucket { type BucketingConfig_Bucket struct { // Lower bound of the range, inclusive. Type should be the same as max if // used. - Min *Value `protobuf:"bytes,1,opt,name=min" json:"min,omitempty"` + Min *Value `protobuf:"bytes,1,opt,name=min,proto3" json:"min,omitempty"` // Upper bound of the range, exclusive; type must match min. - Max *Value `protobuf:"bytes,2,opt,name=max" json:"max,omitempty"` + Max *Value `protobuf:"bytes,2,opt,name=max,proto3" json:"max,omitempty"` // Replacement value for this bucket. If not provided // the default behavior will be to hyphenate the min-max range. - ReplacementValue *Value `protobuf:"bytes,3,opt,name=replacement_value,json=replacementValue" json:"replacement_value,omitempty"` + ReplacementValue *Value `protobuf:"bytes,3,opt,name=replacement_value,json=replacementValue,proto3" json:"replacement_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6532,7 +7199,7 @@ func (m *BucketingConfig_Bucket) Reset() { *m = BucketingConfig_Bucket{} func (m *BucketingConfig_Bucket) String() string { return proto.CompactTextString(m) } func (*BucketingConfig_Bucket) ProtoMessage() {} func (*BucketingConfig_Bucket) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{46, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{48, 0} } func (m *BucketingConfig_Bucket) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BucketingConfig_Bucket.Unmarshal(m, b) @@ -6582,10 +7249,10 @@ func (m *BucketingConfig_Bucket) GetReplacementValue() *Value { // replaced with the same surrogate. // Identifiers must be at least two characters long. // In the case that the identifier is the empty string, it will be skipped. -// See [Pseudonymization](/dlp/docs/pseudonymization) for example usage. +// See https://cloud.google.com/dlp/docs/pseudonymization to learn more. type CryptoReplaceFfxFpeConfig struct { // The key used by the encryption algorithm. [required] - CryptoKey *CryptoKey `protobuf:"bytes,1,opt,name=crypto_key,json=cryptoKey" json:"crypto_key,omitempty"` + CryptoKey *CryptoKey `protobuf:"bytes,1,opt,name=crypto_key,json=cryptoKey,proto3" json:"crypto_key,omitempty"` // The 'tweak', a context may be used for higher security since the same // identifier in two different contexts won't be given the same surrogate. If // the context is not set, a default tweak will be used. @@ -6605,9 +7272,8 @@ type CryptoReplaceFfxFpeConfig struct { // such that: // // - a 64 bit integer is encoded followed by a single byte of value 1 - // - a string is encoded in UTF-8 format followed by a single byte of value - // å 2 - Context *FieldId `protobuf:"bytes,2,opt,name=context" json:"context,omitempty"` + // - a string is encoded in UTF-8 format followed by a single byte of value 2 + Context *FieldId `protobuf:"bytes,2,opt,name=context,proto3" json:"context,omitempty"` // Types that are valid to be assigned to Alphabet: // *CryptoReplaceFfxFpeConfig_CommonAlphabet // *CryptoReplaceFfxFpeConfig_CustomAlphabet @@ -6638,7 +7304,7 @@ type CryptoReplaceFfxFpeConfig struct { // For example, assuming your data is entered from a regular ASCII keyboard, // the symbol with the hex code point 29DD might be used like so: // ⧝MY_TOKEN_TYPE - SurrogateInfoType *InfoType `protobuf:"bytes,8,opt,name=surrogate_info_type,json=surrogateInfoType" json:"surrogate_info_type,omitempty"` + SurrogateInfoType *InfoType `protobuf:"bytes,8,opt,name=surrogate_info_type,json=surrogateInfoType,proto3" json:"surrogate_info_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -6648,7 +7314,7 @@ func (m *CryptoReplaceFfxFpeConfig) Reset() { *m = CryptoReplaceFfxFpeCo func (m *CryptoReplaceFfxFpeConfig) String() string { return proto.CompactTextString(m) } func (*CryptoReplaceFfxFpeConfig) ProtoMessage() {} func (*CryptoReplaceFfxFpeConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{47} + return fileDescriptor_dlp_536657b46c30097f, []int{49} } func (m *CryptoReplaceFfxFpeConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CryptoReplaceFfxFpeConfig.Unmarshal(m, b) @@ -6673,13 +7339,13 @@ type isCryptoReplaceFfxFpeConfig_Alphabet interface { } type CryptoReplaceFfxFpeConfig_CommonAlphabet struct { - CommonAlphabet CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet `protobuf:"varint,4,opt,name=common_alphabet,json=commonAlphabet,enum=google.privacy.dlp.v2.CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet,oneof"` + CommonAlphabet CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet `protobuf:"varint,4,opt,name=common_alphabet,json=commonAlphabet,proto3,enum=google.privacy.dlp.v2.CryptoReplaceFfxFpeConfig_FfxCommonNativeAlphabet,oneof"` } type CryptoReplaceFfxFpeConfig_CustomAlphabet struct { - CustomAlphabet string `protobuf:"bytes,5,opt,name=custom_alphabet,json=customAlphabet,oneof"` + CustomAlphabet string `protobuf:"bytes,5,opt,name=custom_alphabet,json=customAlphabet,proto3,oneof"` } type CryptoReplaceFfxFpeConfig_Radix struct { - Radix int32 `protobuf:"varint,6,opt,name=radix,oneof"` + Radix int32 `protobuf:"varint,6,opt,name=radix,proto3,oneof"` } func (*CryptoReplaceFfxFpeConfig_CommonAlphabet) isCryptoReplaceFfxFpeConfig_Alphabet() {} @@ -6834,7 +7500,7 @@ func (m *CryptoKey) Reset() { *m = CryptoKey{} } func (m *CryptoKey) String() string { return proto.CompactTextString(m) } func (*CryptoKey) ProtoMessage() {} func (*CryptoKey) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{48} + return fileDescriptor_dlp_536657b46c30097f, []int{50} } func (m *CryptoKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CryptoKey.Unmarshal(m, b) @@ -6859,13 +7525,13 @@ type isCryptoKey_Source interface { } type CryptoKey_Transient struct { - Transient *TransientCryptoKey `protobuf:"bytes,1,opt,name=transient,oneof"` + Transient *TransientCryptoKey `protobuf:"bytes,1,opt,name=transient,proto3,oneof"` } type CryptoKey_Unwrapped struct { - Unwrapped *UnwrappedCryptoKey `protobuf:"bytes,2,opt,name=unwrapped,oneof"` + Unwrapped *UnwrappedCryptoKey `protobuf:"bytes,2,opt,name=unwrapped,proto3,oneof"` } type CryptoKey_KmsWrapped struct { - KmsWrapped *KmsWrappedCryptoKey `protobuf:"bytes,3,opt,name=kms_wrapped,json=kmsWrapped,oneof"` + KmsWrapped *KmsWrappedCryptoKey `protobuf:"bytes,3,opt,name=kms_wrapped,json=kmsWrapped,proto3,oneof"` } func (*CryptoKey_Transient) isCryptoKey_Source() {} @@ -7002,7 +7668,7 @@ type TransientCryptoKey struct { // protos share the same generated key if their names are the same. // When the data crypto key is generated, this name is not used in any way // (repeating the api call will result in a different key being generated). - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7012,7 +7678,7 @@ func (m *TransientCryptoKey) Reset() { *m = TransientCryptoKey{} } func (m *TransientCryptoKey) String() string { return proto.CompactTextString(m) } func (*TransientCryptoKey) ProtoMessage() {} func (*TransientCryptoKey) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{49} + return fileDescriptor_dlp_536657b46c30097f, []int{51} } func (m *TransientCryptoKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TransientCryptoKey.Unmarshal(m, b) @@ -7053,7 +7719,7 @@ func (m *UnwrappedCryptoKey) Reset() { *m = UnwrappedCryptoKey{} } func (m *UnwrappedCryptoKey) String() string { return proto.CompactTextString(m) } func (*UnwrappedCryptoKey) ProtoMessage() {} func (*UnwrappedCryptoKey) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{50} + return fileDescriptor_dlp_536657b46c30097f, []int{52} } func (m *UnwrappedCryptoKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UnwrappedCryptoKey.Unmarshal(m, b) @@ -7088,7 +7754,7 @@ type KmsWrappedCryptoKey struct { // The wrapped data crypto key. [required] WrappedKey []byte `protobuf:"bytes,1,opt,name=wrapped_key,json=wrappedKey,proto3" json:"wrapped_key,omitempty"` // The resource name of the KMS CryptoKey to use for unwrapping. [required] - CryptoKeyName string `protobuf:"bytes,2,opt,name=crypto_key_name,json=cryptoKeyName" json:"crypto_key_name,omitempty"` + CryptoKeyName string `protobuf:"bytes,2,opt,name=crypto_key_name,json=cryptoKeyName,proto3" json:"crypto_key_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7098,7 +7764,7 @@ func (m *KmsWrappedCryptoKey) Reset() { *m = KmsWrappedCryptoKey{} } func (m *KmsWrappedCryptoKey) String() string { return proto.CompactTextString(m) } func (*KmsWrappedCryptoKey) ProtoMessage() {} func (*KmsWrappedCryptoKey) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{51} + return fileDescriptor_dlp_536657b46c30097f, []int{53} } func (m *KmsWrappedCryptoKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_KmsWrappedCryptoKey.Unmarshal(m, b) @@ -7133,7 +7799,8 @@ func (m *KmsWrappedCryptoKey) GetCryptoKeyName() string { } // Shifts dates by random number of days, with option to be consistent for the -// same context. +// same context. See https://cloud.google.com/dlp/docs/concepts-date-shifting +// to learn more. type DateShiftConfig struct { // Range of shift in days. Actual shift will be selected at random within this // range (inclusive ends). Negative means shift to earlier in time. Must not @@ -7141,14 +7808,14 @@ type DateShiftConfig struct { // // For example, 3 means shift date to at most 3 days into the future. // [Required] - UpperBoundDays int32 `protobuf:"varint,1,opt,name=upper_bound_days,json=upperBoundDays" json:"upper_bound_days,omitempty"` + UpperBoundDays int32 `protobuf:"varint,1,opt,name=upper_bound_days,json=upperBoundDays,proto3" json:"upper_bound_days,omitempty"` // For example, -5 means shift date to at most 5 days back in the past. // [Required] - LowerBoundDays int32 `protobuf:"varint,2,opt,name=lower_bound_days,json=lowerBoundDays" json:"lower_bound_days,omitempty"` + LowerBoundDays int32 `protobuf:"varint,2,opt,name=lower_bound_days,json=lowerBoundDays,proto3" json:"lower_bound_days,omitempty"` // Points to the field that contains the context, for example, an entity id. // If set, must also set method. If set, shift will be consistent for the // given context. - Context *FieldId `protobuf:"bytes,3,opt,name=context" json:"context,omitempty"` + Context *FieldId `protobuf:"bytes,3,opt,name=context,proto3" json:"context,omitempty"` // Method for calculating shift that takes context into consideration. If // set, must also set context. Can only be applied to table items. // @@ -7164,7 +7831,7 @@ func (m *DateShiftConfig) Reset() { *m = DateShiftConfig{} } func (m *DateShiftConfig) String() string { return proto.CompactTextString(m) } func (*DateShiftConfig) ProtoMessage() {} func (*DateShiftConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{52} + return fileDescriptor_dlp_536657b46c30097f, []int{54} } func (m *DateShiftConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DateShiftConfig.Unmarshal(m, b) @@ -7189,7 +7856,7 @@ type isDateShiftConfig_Method interface { } type DateShiftConfig_CryptoKey struct { - CryptoKey *CryptoKey `protobuf:"bytes,4,opt,name=crypto_key,json=cryptoKey,oneof"` + CryptoKey *CryptoKey `protobuf:"bytes,4,opt,name=crypto_key,json=cryptoKey,proto3,oneof"` } func (*DateShiftConfig_CryptoKey) isDateShiftConfig_Method() {} @@ -7291,7 +7958,7 @@ func _DateShiftConfig_OneofSizer(msg proto.Message) (n int) { type InfoTypeTransformations struct { // Transformation for each infoType. Cannot specify more than one // for a given infoType. [required] - Transformations []*InfoTypeTransformations_InfoTypeTransformation `protobuf:"bytes,1,rep,name=transformations" json:"transformations,omitempty"` + Transformations []*InfoTypeTransformations_InfoTypeTransformation `protobuf:"bytes,1,rep,name=transformations,proto3" json:"transformations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7301,7 +7968,7 @@ func (m *InfoTypeTransformations) Reset() { *m = InfoTypeTransformations func (m *InfoTypeTransformations) String() string { return proto.CompactTextString(m) } func (*InfoTypeTransformations) ProtoMessage() {} func (*InfoTypeTransformations) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{53} + return fileDescriptor_dlp_536657b46c30097f, []int{55} } func (m *InfoTypeTransformations) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InfoTypeTransformations.Unmarshal(m, b) @@ -7331,11 +7998,12 @@ func (m *InfoTypeTransformations) GetTransformations() []*InfoTypeTransformation // A transformation to apply to text that is identified as a specific // info_type. type InfoTypeTransformations_InfoTypeTransformation struct { - // InfoTypes to apply the transformation to. Empty list will match all - // available infoTypes for this transformation. - InfoTypes []*InfoType `protobuf:"bytes,1,rep,name=info_types,json=infoTypes" json:"info_types,omitempty"` + // InfoTypes to apply the transformation to. An empty list will cause + // this transformation to apply to all findings that correspond to + // infoTypes that were requested in `InspectConfig`. + InfoTypes []*InfoType `protobuf:"bytes,1,rep,name=info_types,json=infoTypes,proto3" json:"info_types,omitempty"` // Primitive transformation to apply to the infoType. [required] - PrimitiveTransformation *PrimitiveTransformation `protobuf:"bytes,2,opt,name=primitive_transformation,json=primitiveTransformation" json:"primitive_transformation,omitempty"` + PrimitiveTransformation *PrimitiveTransformation `protobuf:"bytes,2,opt,name=primitive_transformation,json=primitiveTransformation,proto3" json:"primitive_transformation,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7349,7 +8017,7 @@ func (m *InfoTypeTransformations_InfoTypeTransformation) String() string { } func (*InfoTypeTransformations_InfoTypeTransformation) ProtoMessage() {} func (*InfoTypeTransformations_InfoTypeTransformation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{53, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{55, 0} } func (m *InfoTypeTransformations_InfoTypeTransformation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InfoTypeTransformations_InfoTypeTransformation.Unmarshal(m, b) @@ -7386,7 +8054,7 @@ func (m *InfoTypeTransformations_InfoTypeTransformation) GetPrimitiveTransformat // The transformation to apply to the field. type FieldTransformation struct { // Input field(s) to apply the transformation to. [required] - Fields []*FieldId `protobuf:"bytes,1,rep,name=fields" json:"fields,omitempty"` + Fields []*FieldId `protobuf:"bytes,1,rep,name=fields,proto3" json:"fields,omitempty"` // Only apply the transformation if the condition evaluates to true for the // given `RecordCondition`. The conditions are allowed to reference fields // that are not used in the actual transformation. [optional] @@ -7396,7 +8064,7 @@ type FieldTransformation struct { // - Apply a different bucket transformation to an age column if the zip code // column for the same record is within a specific range. // - Redact a field if the date of birth field is greater than 85. - Condition *RecordCondition `protobuf:"bytes,3,opt,name=condition" json:"condition,omitempty"` + Condition *RecordCondition `protobuf:"bytes,3,opt,name=condition,proto3" json:"condition,omitempty"` // Transformation to apply. [required] // // Types that are valid to be assigned to Transformation: @@ -7412,7 +8080,7 @@ func (m *FieldTransformation) Reset() { *m = FieldTransformation{} } func (m *FieldTransformation) String() string { return proto.CompactTextString(m) } func (*FieldTransformation) ProtoMessage() {} func (*FieldTransformation) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{54} + return fileDescriptor_dlp_536657b46c30097f, []int{56} } func (m *FieldTransformation) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FieldTransformation.Unmarshal(m, b) @@ -7437,10 +8105,10 @@ type isFieldTransformation_Transformation interface { } type FieldTransformation_PrimitiveTransformation struct { - PrimitiveTransformation *PrimitiveTransformation `protobuf:"bytes,4,opt,name=primitive_transformation,json=primitiveTransformation,oneof"` + PrimitiveTransformation *PrimitiveTransformation `protobuf:"bytes,4,opt,name=primitive_transformation,json=primitiveTransformation,proto3,oneof"` } type FieldTransformation_InfoTypeTransformations struct { - InfoTypeTransformations *InfoTypeTransformations `protobuf:"bytes,5,opt,name=info_type_transformations,json=infoTypeTransformations,oneof"` + InfoTypeTransformations *InfoTypeTransformations `protobuf:"bytes,5,opt,name=info_type_transformations,json=infoTypeTransformations,proto3,oneof"` } func (*FieldTransformation_PrimitiveTransformation) isFieldTransformation_Transformation() {} @@ -7559,10 +8227,10 @@ func _FieldTransformation_OneofSizer(msg proto.Message) (n int) { // table. type RecordTransformations struct { // Transform the record by applying various field transformations. - FieldTransformations []*FieldTransformation `protobuf:"bytes,1,rep,name=field_transformations,json=fieldTransformations" json:"field_transformations,omitempty"` + FieldTransformations []*FieldTransformation `protobuf:"bytes,1,rep,name=field_transformations,json=fieldTransformations,proto3" json:"field_transformations,omitempty"` // Configuration defining which records get suppressed entirely. Records that // match any suppression rule are omitted from the output [optional]. - RecordSuppressions []*RecordSuppression `protobuf:"bytes,2,rep,name=record_suppressions,json=recordSuppressions" json:"record_suppressions,omitempty"` + RecordSuppressions []*RecordSuppression `protobuf:"bytes,2,rep,name=record_suppressions,json=recordSuppressions,proto3" json:"record_suppressions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7572,7 +8240,7 @@ func (m *RecordTransformations) Reset() { *m = RecordTransformations{} } func (m *RecordTransformations) String() string { return proto.CompactTextString(m) } func (*RecordTransformations) ProtoMessage() {} func (*RecordTransformations) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{55} + return fileDescriptor_dlp_536657b46c30097f, []int{57} } func (m *RecordTransformations) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordTransformations.Unmarshal(m, b) @@ -7611,7 +8279,7 @@ func (m *RecordTransformations) GetRecordSuppressions() []*RecordSuppression { type RecordSuppression struct { // A condition that when it evaluates to true will result in the record being // evaluated to be suppressed from the transformed content. - Condition *RecordCondition `protobuf:"bytes,1,opt,name=condition" json:"condition,omitempty"` + Condition *RecordCondition `protobuf:"bytes,1,opt,name=condition,proto3" json:"condition,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7621,7 +8289,7 @@ func (m *RecordSuppression) Reset() { *m = RecordSuppression{} } func (m *RecordSuppression) String() string { return proto.CompactTextString(m) } func (*RecordSuppression) ProtoMessage() {} func (*RecordSuppression) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{56} + return fileDescriptor_dlp_536657b46c30097f, []int{58} } func (m *RecordSuppression) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordSuppression.Unmarshal(m, b) @@ -7652,7 +8320,7 @@ func (m *RecordSuppression) GetCondition() *RecordCondition { // a field. type RecordCondition struct { // An expression. - Expressions *RecordCondition_Expressions `protobuf:"bytes,3,opt,name=expressions" json:"expressions,omitempty"` + Expressions *RecordCondition_Expressions `protobuf:"bytes,3,opt,name=expressions,proto3" json:"expressions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7662,7 +8330,7 @@ func (m *RecordCondition) Reset() { *m = RecordCondition{} } func (m *RecordCondition) String() string { return proto.CompactTextString(m) } func (*RecordCondition) ProtoMessage() {} func (*RecordCondition) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{57} + return fileDescriptor_dlp_536657b46c30097f, []int{59} } func (m *RecordCondition) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordCondition.Unmarshal(m, b) @@ -7709,11 +8377,11 @@ func (m *RecordCondition) GetExpressions() *RecordCondition_Expressions { // the condition will evaluate to false. type RecordCondition_Condition struct { // Field within the record this condition is evaluated against. [required] - Field *FieldId `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` // Operator used to compare the field or infoType to the value. [required] - Operator RelationalOperator `protobuf:"varint,3,opt,name=operator,enum=google.privacy.dlp.v2.RelationalOperator" json:"operator,omitempty"` + Operator RelationalOperator `protobuf:"varint,3,opt,name=operator,proto3,enum=google.privacy.dlp.v2.RelationalOperator" json:"operator,omitempty"` // Value to compare against. [Required, except for `EXISTS` tests.] - Value *Value `protobuf:"bytes,4,opt,name=value" json:"value,omitempty"` + Value *Value `protobuf:"bytes,4,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7723,7 +8391,7 @@ func (m *RecordCondition_Condition) Reset() { *m = RecordCondition_Condi func (m *RecordCondition_Condition) String() string { return proto.CompactTextString(m) } func (*RecordCondition_Condition) ProtoMessage() {} func (*RecordCondition_Condition) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{57, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{59, 0} } func (m *RecordCondition_Condition) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordCondition_Condition.Unmarshal(m, b) @@ -7766,7 +8434,7 @@ func (m *RecordCondition_Condition) GetValue() *Value { // A collection of conditions. type RecordCondition_Conditions struct { - Conditions []*RecordCondition_Condition `protobuf:"bytes,1,rep,name=conditions" json:"conditions,omitempty"` + Conditions []*RecordCondition_Condition `protobuf:"bytes,1,rep,name=conditions,proto3" json:"conditions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7776,7 +8444,7 @@ func (m *RecordCondition_Conditions) Reset() { *m = RecordCondition_Cond func (m *RecordCondition_Conditions) String() string { return proto.CompactTextString(m) } func (*RecordCondition_Conditions) ProtoMessage() {} func (*RecordCondition_Conditions) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{57, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{59, 1} } func (m *RecordCondition_Conditions) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordCondition_Conditions.Unmarshal(m, b) @@ -7807,7 +8475,7 @@ func (m *RecordCondition_Conditions) GetConditions() []*RecordCondition_Conditio type RecordCondition_Expressions struct { // The operator to apply to the result of conditions. Default and currently // only supported value is `AND`. - LogicalOperator RecordCondition_Expressions_LogicalOperator `protobuf:"varint,1,opt,name=logical_operator,json=logicalOperator,enum=google.privacy.dlp.v2.RecordCondition_Expressions_LogicalOperator" json:"logical_operator,omitempty"` + LogicalOperator RecordCondition_Expressions_LogicalOperator `protobuf:"varint,1,opt,name=logical_operator,json=logicalOperator,proto3,enum=google.privacy.dlp.v2.RecordCondition_Expressions_LogicalOperator" json:"logical_operator,omitempty"` // Types that are valid to be assigned to Type: // *RecordCondition_Expressions_Conditions Type isRecordCondition_Expressions_Type `protobuf_oneof:"type"` @@ -7820,7 +8488,7 @@ func (m *RecordCondition_Expressions) Reset() { *m = RecordCondition_Exp func (m *RecordCondition_Expressions) String() string { return proto.CompactTextString(m) } func (*RecordCondition_Expressions) ProtoMessage() {} func (*RecordCondition_Expressions) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{57, 2} + return fileDescriptor_dlp_536657b46c30097f, []int{59, 2} } func (m *RecordCondition_Expressions) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordCondition_Expressions.Unmarshal(m, b) @@ -7845,7 +8513,7 @@ type isRecordCondition_Expressions_Type interface { } type RecordCondition_Expressions_Conditions struct { - Conditions *RecordCondition_Conditions `protobuf:"bytes,3,opt,name=conditions,oneof"` + Conditions *RecordCondition_Conditions `protobuf:"bytes,3,opt,name=conditions,proto3,oneof"` } func (*RecordCondition_Expressions_Conditions) isRecordCondition_Expressions_Type() {} @@ -7929,9 +8597,9 @@ func _RecordCondition_Expressions_OneofSizer(msg proto.Message) (n int) { // Overview of the modifications that occurred. type TransformationOverview struct { // Total size in bytes that were transformed in some way. - TransformedBytes int64 `protobuf:"varint,2,opt,name=transformed_bytes,json=transformedBytes" json:"transformed_bytes,omitempty"` + TransformedBytes int64 `protobuf:"varint,2,opt,name=transformed_bytes,json=transformedBytes,proto3" json:"transformed_bytes,omitempty"` // Transformations applied to the dataset. - TransformationSummaries []*TransformationSummary `protobuf:"bytes,3,rep,name=transformation_summaries,json=transformationSummaries" json:"transformation_summaries,omitempty"` + TransformationSummaries []*TransformationSummary `protobuf:"bytes,3,rep,name=transformation_summaries,json=transformationSummaries,proto3" json:"transformation_summaries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -7941,7 +8609,7 @@ func (m *TransformationOverview) Reset() { *m = TransformationOverview{} func (m *TransformationOverview) String() string { return proto.CompactTextString(m) } func (*TransformationOverview) ProtoMessage() {} func (*TransformationOverview) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{58} + return fileDescriptor_dlp_536657b46c30097f, []int{60} } func (m *TransformationOverview) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TransformationOverview.Unmarshal(m, b) @@ -7980,20 +8648,20 @@ func (m *TransformationOverview) GetTransformationSummaries() []*TransformationS // will be set. type TransformationSummary struct { // Set if the transformation was limited to a specific info_type. - InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType" json:"info_type,omitempty"` + InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType,proto3" json:"info_type,omitempty"` // Set if the transformation was limited to a specific FieldId. - Field *FieldId `protobuf:"bytes,2,opt,name=field" json:"field,omitempty"` + Field *FieldId `protobuf:"bytes,2,opt,name=field,proto3" json:"field,omitempty"` // The specific transformation these stats apply to. - Transformation *PrimitiveTransformation `protobuf:"bytes,3,opt,name=transformation" json:"transformation,omitempty"` + Transformation *PrimitiveTransformation `protobuf:"bytes,3,opt,name=transformation,proto3" json:"transformation,omitempty"` // The field transformation that was applied. // If multiple field transformations are requested for a single field, // this list will contain all of them; otherwise, only one is supplied. - FieldTransformations []*FieldTransformation `protobuf:"bytes,5,rep,name=field_transformations,json=fieldTransformations" json:"field_transformations,omitempty"` + FieldTransformations []*FieldTransformation `protobuf:"bytes,5,rep,name=field_transformations,json=fieldTransformations,proto3" json:"field_transformations,omitempty"` // The specific suppression option these stats apply to. - RecordSuppress *RecordSuppression `protobuf:"bytes,6,opt,name=record_suppress,json=recordSuppress" json:"record_suppress,omitempty"` - Results []*TransformationSummary_SummaryResult `protobuf:"bytes,4,rep,name=results" json:"results,omitempty"` + RecordSuppress *RecordSuppression `protobuf:"bytes,6,opt,name=record_suppress,json=recordSuppress,proto3" json:"record_suppress,omitempty"` + Results []*TransformationSummary_SummaryResult `protobuf:"bytes,4,rep,name=results,proto3" json:"results,omitempty"` // Total size in bytes that were transformed in some way. - TransformedBytes int64 `protobuf:"varint,7,opt,name=transformed_bytes,json=transformedBytes" json:"transformed_bytes,omitempty"` + TransformedBytes int64 `protobuf:"varint,7,opt,name=transformed_bytes,json=transformedBytes,proto3" json:"transformed_bytes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8003,7 +8671,7 @@ func (m *TransformationSummary) Reset() { *m = TransformationSummary{} } func (m *TransformationSummary) String() string { return proto.CompactTextString(m) } func (*TransformationSummary) ProtoMessage() {} func (*TransformationSummary) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{59} + return fileDescriptor_dlp_536657b46c30097f, []int{61} } func (m *TransformationSummary) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TransformationSummary.Unmarshal(m, b) @@ -8075,11 +8743,11 @@ func (m *TransformationSummary) GetTransformedBytes() int64 { // A collection that informs the user the number of times a particular // `TransformationResultCode` and error details occurred. type TransformationSummary_SummaryResult struct { - Count int64 `protobuf:"varint,1,opt,name=count" json:"count,omitempty"` - Code TransformationSummary_TransformationResultCode `protobuf:"varint,2,opt,name=code,enum=google.privacy.dlp.v2.TransformationSummary_TransformationResultCode" json:"code,omitempty"` + Count int64 `protobuf:"varint,1,opt,name=count,proto3" json:"count,omitempty"` + Code TransformationSummary_TransformationResultCode `protobuf:"varint,2,opt,name=code,proto3,enum=google.privacy.dlp.v2.TransformationSummary_TransformationResultCode" json:"code,omitempty"` // A place for warnings or errors to show up if a transformation didn't // work as expected. - Details string `protobuf:"bytes,3,opt,name=details" json:"details,omitempty"` + Details string `protobuf:"bytes,3,opt,name=details,proto3" json:"details,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8089,7 +8757,7 @@ func (m *TransformationSummary_SummaryResult) Reset() { *m = Transformat func (m *TransformationSummary_SummaryResult) String() string { return proto.CompactTextString(m) } func (*TransformationSummary_SummaryResult) ProtoMessage() {} func (*TransformationSummary_SummaryResult) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{59, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{61, 0} } func (m *TransformationSummary_SummaryResult) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TransformationSummary_SummaryResult.Unmarshal(m, b) @@ -8144,7 +8812,7 @@ func (m *Schedule) Reset() { *m = Schedule{} } func (m *Schedule) String() string { return proto.CompactTextString(m) } func (*Schedule) ProtoMessage() {} func (*Schedule) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{60} + return fileDescriptor_dlp_536657b46c30097f, []int{62} } func (m *Schedule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Schedule.Unmarshal(m, b) @@ -8169,7 +8837,7 @@ type isSchedule_Option interface { } type Schedule_RecurrencePeriodDuration struct { - RecurrencePeriodDuration *duration.Duration `protobuf:"bytes,1,opt,name=recurrence_period_duration,json=recurrencePeriodDuration,oneof"` + RecurrencePeriodDuration *duration.Duration `protobuf:"bytes,1,opt,name=recurrence_period_duration,json=recurrencePeriodDuration,proto3,oneof"` } func (*Schedule_RecurrencePeriodDuration) isSchedule_Option() {} @@ -8245,24 +8913,25 @@ func _Schedule_OneofSizer(msg proto.Message) (n int) { // The inspectTemplate contains a configuration (set of types of sensitive data // to be detected) to be used anywhere you otherwise would normally specify -// InspectConfig. +// InspectConfig. See https://cloud.google.com/dlp/docs/concepts-templates +// to learn more. type InspectTemplate struct { // The template name. Output only. // // The template will have one of the following formats: // `projects/PROJECT_ID/inspectTemplates/TEMPLATE_ID` OR // `organizations/ORGANIZATION_ID/inspectTemplates/TEMPLATE_ID` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Display name (max 256 chars). - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Short description (max 256 chars). - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The creation timestamp of a inspectTemplate, output only field. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The last update timestamp of a inspectTemplate, output only field. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // The core content of the template. Configuration of the scanning process. - InspectConfig *InspectConfig `protobuf:"bytes,6,opt,name=inspect_config,json=inspectConfig" json:"inspect_config,omitempty"` + InspectConfig *InspectConfig `protobuf:"bytes,6,opt,name=inspect_config,json=inspectConfig,proto3" json:"inspect_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8272,7 +8941,7 @@ func (m *InspectTemplate) Reset() { *m = InspectTemplate{} } func (m *InspectTemplate) String() string { return proto.CompactTextString(m) } func (*InspectTemplate) ProtoMessage() {} func (*InspectTemplate) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{61} + return fileDescriptor_dlp_536657b46c30097f, []int{63} } func (m *InspectTemplate) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectTemplate.Unmarshal(m, b) @@ -8335,23 +9004,24 @@ func (m *InspectTemplate) GetInspectConfig() *InspectConfig { } // The DeidentifyTemplates contains instructions on how to deidentify content. +// See https://cloud.google.com/dlp/docs/concepts-templates to learn more. type DeidentifyTemplate struct { // The template name. Output only. // // The template will have one of the following formats: // `projects/PROJECT_ID/deidentifyTemplates/TEMPLATE_ID` OR // `organizations/ORGANIZATION_ID/deidentifyTemplates/TEMPLATE_ID` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Display name (max 256 chars). - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Short description (max 256 chars). - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The creation timestamp of a inspectTemplate, output only field. - CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The last update timestamp of a inspectTemplate, output only field. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // ///////////// // The core content of the template // /////////////// - DeidentifyConfig *DeidentifyConfig `protobuf:"bytes,6,opt,name=deidentify_config,json=deidentifyConfig" json:"deidentify_config,omitempty"` + DeidentifyConfig *DeidentifyConfig `protobuf:"bytes,6,opt,name=deidentify_config,json=deidentifyConfig,proto3" json:"deidentify_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8361,7 +9031,7 @@ func (m *DeidentifyTemplate) Reset() { *m = DeidentifyTemplate{} } func (m *DeidentifyTemplate) String() string { return proto.CompactTextString(m) } func (*DeidentifyTemplate) ProtoMessage() {} func (*DeidentifyTemplate) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{62} + return fileDescriptor_dlp_536657b46c30097f, []int{64} } func (m *DeidentifyTemplate) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeidentifyTemplate.Unmarshal(m, b) @@ -8427,9 +9097,9 @@ func (m *DeidentifyTemplate) GetDeidentifyConfig() *DeidentifyConfig { // the results of an unsuccessful activation of the JobTrigger. // Output only field. type Error struct { - Details *status.Status `protobuf:"bytes,1,opt,name=details" json:"details,omitempty"` + Details *status.Status `protobuf:"bytes,1,opt,name=details,proto3" json:"details,omitempty"` // The times the error occurred. - Timestamps []*timestamp.Timestamp `protobuf:"bytes,2,rep,name=timestamps" json:"timestamps,omitempty"` + Timestamps []*timestamp.Timestamp `protobuf:"bytes,2,rep,name=timestamps,proto3" json:"timestamps,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8439,7 +9109,7 @@ func (m *Error) Reset() { *m = Error{} } func (m *Error) String() string { return proto.CompactTextString(m) } func (*Error) ProtoMessage() {} func (*Error) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{63} + return fileDescriptor_dlp_536657b46c30097f, []int{65} } func (m *Error) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Error.Unmarshal(m, b) @@ -8474,15 +9144,16 @@ func (m *Error) GetTimestamps() []*timestamp.Timestamp { } // Contains a configuration to make dlp api calls on a repeating basis. +// See https://cloud.google.com/dlp/docs/concepts-job-triggers to learn more. type JobTrigger struct { // Unique resource name for the triggeredJob, assigned by the service when the // triggeredJob is created, for example // `projects/dlp-test-project/triggeredJobs/53234423`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Display name (max 100 chars) - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // User provided description (max 256 chars) - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` // The configuration details for the specific type of job to run. // // Types that are valid to be assigned to Job: @@ -8491,20 +9162,20 @@ type JobTrigger struct { // A list of triggers which will be OR'ed together. Only one in the list // needs to trigger for a job to be started. The list may contain only // a single Schedule trigger and must have at least one object. - Triggers []*JobTrigger_Trigger `protobuf:"bytes,5,rep,name=triggers" json:"triggers,omitempty"` + Triggers []*JobTrigger_Trigger `protobuf:"bytes,5,rep,name=triggers,proto3" json:"triggers,omitempty"` // A stream of errors encountered when the trigger was activated. Repeated // errors may result in the JobTrigger automaticaly being paused. // Will return the last 100 errors. Whenever the JobTrigger is modified // this list will be cleared. Output only field. - Errors []*Error `protobuf:"bytes,6,rep,name=errors" json:"errors,omitempty"` + Errors []*Error `protobuf:"bytes,6,rep,name=errors,proto3" json:"errors,omitempty"` // The creation timestamp of a triggeredJob, output only field. - CreateTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // The last update timestamp of a triggeredJob, output only field. - UpdateTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=update_time,json=updateTime" json:"update_time,omitempty"` + UpdateTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // The timestamp of the last time this trigger executed, output only field. - LastRunTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=last_run_time,json=lastRunTime" json:"last_run_time,omitempty"` + LastRunTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=last_run_time,json=lastRunTime,proto3" json:"last_run_time,omitempty"` // A status for this trigger. [required] - Status JobTrigger_Status `protobuf:"varint,10,opt,name=status,enum=google.privacy.dlp.v2.JobTrigger_Status" json:"status,omitempty"` + Status JobTrigger_Status `protobuf:"varint,10,opt,name=status,proto3,enum=google.privacy.dlp.v2.JobTrigger_Status" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8514,7 +9185,7 @@ func (m *JobTrigger) Reset() { *m = JobTrigger{} } func (m *JobTrigger) String() string { return proto.CompactTextString(m) } func (*JobTrigger) ProtoMessage() {} func (*JobTrigger) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{64} + return fileDescriptor_dlp_536657b46c30097f, []int{66} } func (m *JobTrigger) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_JobTrigger.Unmarshal(m, b) @@ -8539,7 +9210,7 @@ type isJobTrigger_Job interface { } type JobTrigger_InspectJob struct { - InspectJob *InspectJobConfig `protobuf:"bytes,4,opt,name=inspect_job,json=inspectJob,oneof"` + InspectJob *InspectJobConfig `protobuf:"bytes,4,opt,name=inspect_job,json=inspectJob,proto3,oneof"` } func (*JobTrigger_InspectJob) isJobTrigger_Job() {} @@ -8690,7 +9361,7 @@ func (m *JobTrigger_Trigger) Reset() { *m = JobTrigger_Trigger{} } func (m *JobTrigger_Trigger) String() string { return proto.CompactTextString(m) } func (*JobTrigger_Trigger) ProtoMessage() {} func (*JobTrigger_Trigger) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{64, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{66, 0} } func (m *JobTrigger_Trigger) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_JobTrigger_Trigger.Unmarshal(m, b) @@ -8715,7 +9386,7 @@ type isJobTrigger_Trigger_Trigger interface { } type JobTrigger_Trigger_Schedule struct { - Schedule *Schedule `protobuf:"bytes,1,opt,name=schedule,oneof"` + Schedule *Schedule `protobuf:"bytes,1,opt,name=schedule,proto3,oneof"` } func (*JobTrigger_Trigger_Schedule) isJobTrigger_Trigger_Trigger() {} @@ -8790,6 +9461,7 @@ func _JobTrigger_Trigger_OneofSizer(msg proto.Message) (n int) { } // A task to execute on the completion of a job. +// See https://cloud.google.com/dlp/docs/concepts-actions to learn more. type Action struct { // Types that are valid to be assigned to Action: // *Action_SaveFindings_ @@ -8805,7 +9477,7 @@ func (m *Action) Reset() { *m = Action{} } func (m *Action) String() string { return proto.CompactTextString(m) } func (*Action) ProtoMessage() {} func (*Action) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{65} + return fileDescriptor_dlp_536657b46c30097f, []int{67} } func (m *Action) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Action.Unmarshal(m, b) @@ -8830,13 +9502,13 @@ type isAction_Action interface { } type Action_SaveFindings_ struct { - SaveFindings *Action_SaveFindings `protobuf:"bytes,1,opt,name=save_findings,json=saveFindings,oneof"` + SaveFindings *Action_SaveFindings `protobuf:"bytes,1,opt,name=save_findings,json=saveFindings,proto3,oneof"` } type Action_PubSub struct { - PubSub *Action_PublishToPubSub `protobuf:"bytes,2,opt,name=pub_sub,json=pubSub,oneof"` + PubSub *Action_PublishToPubSub `protobuf:"bytes,2,opt,name=pub_sub,json=pubSub,proto3,oneof"` } type Action_PublishSummaryToCscc_ struct { - PublishSummaryToCscc *Action_PublishSummaryToCscc `protobuf:"bytes,3,opt,name=publish_summary_to_cscc,json=publishSummaryToCscc,oneof"` + PublishSummaryToCscc *Action_PublishSummaryToCscc `protobuf:"bytes,3,opt,name=publish_summary_to_cscc,json=publishSummaryToCscc,proto3,oneof"` } func (*Action_SaveFindings_) isAction_Action() {} @@ -8967,9 +9639,9 @@ func _Action_OneofSizer(msg proto.Message) (n int) { // If set, the detailed findings will be persisted to the specified // OutputStorageConfig. Only a single instance of this action can be // specified. -// Compatible with: Inspect +// Compatible with: Inspect, Risk type Action_SaveFindings struct { - OutputConfig *OutputStorageConfig `protobuf:"bytes,1,opt,name=output_config,json=outputConfig" json:"output_config,omitempty"` + OutputConfig *OutputStorageConfig `protobuf:"bytes,1,opt,name=output_config,json=outputConfig,proto3" json:"output_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -8979,7 +9651,7 @@ func (m *Action_SaveFindings) Reset() { *m = Action_SaveFindings{} } func (m *Action_SaveFindings) String() string { return proto.CompactTextString(m) } func (*Action_SaveFindings) ProtoMessage() {} func (*Action_SaveFindings) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{65, 0} + return fileDescriptor_dlp_536657b46c30097f, []int{67, 0} } func (m *Action_SaveFindings) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Action_SaveFindings.Unmarshal(m, b) @@ -9013,7 +9685,7 @@ type Action_PublishToPubSub struct { // publishing access rights to the DLP API service account executing // the long running DlpJob sending the notifications. // Format is projects/{project}/topics/{topic}. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9023,7 +9695,7 @@ func (m *Action_PublishToPubSub) Reset() { *m = Action_PublishToPubSub{} func (m *Action_PublishToPubSub) String() string { return proto.CompactTextString(m) } func (*Action_PublishToPubSub) ProtoMessage() {} func (*Action_PublishToPubSub) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{65, 1} + return fileDescriptor_dlp_536657b46c30097f, []int{67, 1} } func (m *Action_PublishToPubSub) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Action_PublishToPubSub.Unmarshal(m, b) @@ -9070,7 +9742,7 @@ func (m *Action_PublishSummaryToCscc) Reset() { *m = Action_PublishSumma func (m *Action_PublishSummaryToCscc) String() string { return proto.CompactTextString(m) } func (*Action_PublishSummaryToCscc) ProtoMessage() {} func (*Action_PublishSummaryToCscc) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{65, 2} + return fileDescriptor_dlp_536657b46c30097f, []int{67, 2} } func (m *Action_PublishSummaryToCscc) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Action_PublishSummaryToCscc.Unmarshal(m, b) @@ -9094,14 +9766,14 @@ var xxx_messageInfo_Action_PublishSummaryToCscc proto.InternalMessageInfo type CreateInspectTemplateRequest struct { // The parent resource name, for example projects/my-project-id or // organizations/my-org-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The InspectTemplate to create. - InspectTemplate *InspectTemplate `protobuf:"bytes,2,opt,name=inspect_template,json=inspectTemplate" json:"inspect_template,omitempty"` + InspectTemplate *InspectTemplate `protobuf:"bytes,2,opt,name=inspect_template,json=inspectTemplate,proto3" json:"inspect_template,omitempty"` // The template id can contain uppercase and lowercase letters, // numbers, and hyphens; that is, it must match the regular // expression: `[a-zA-Z\\d-]+`. The maximum length is 100 // characters. Can be empty to allow the system to generate one. - TemplateId string `protobuf:"bytes,3,opt,name=template_id,json=templateId" json:"template_id,omitempty"` + TemplateId string `protobuf:"bytes,3,opt,name=template_id,json=templateId,proto3" json:"template_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9111,7 +9783,7 @@ func (m *CreateInspectTemplateRequest) Reset() { *m = CreateInspectTempl func (m *CreateInspectTemplateRequest) String() string { return proto.CompactTextString(m) } func (*CreateInspectTemplateRequest) ProtoMessage() {} func (*CreateInspectTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{66} + return fileDescriptor_dlp_536657b46c30097f, []int{68} } func (m *CreateInspectTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateInspectTemplateRequest.Unmarshal(m, b) @@ -9157,11 +9829,11 @@ type UpdateInspectTemplateRequest struct { // Resource name of organization and inspectTemplate to be updated, for // example `organizations/433245324/inspectTemplates/432452342` or // projects/project-id/inspectTemplates/432452342. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // New InspectTemplate value. - InspectTemplate *InspectTemplate `protobuf:"bytes,2,opt,name=inspect_template,json=inspectTemplate" json:"inspect_template,omitempty"` + InspectTemplate *InspectTemplate `protobuf:"bytes,2,opt,name=inspect_template,json=inspectTemplate,proto3" json:"inspect_template,omitempty"` // Mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9171,7 +9843,7 @@ func (m *UpdateInspectTemplateRequest) Reset() { *m = UpdateInspectTempl func (m *UpdateInspectTemplateRequest) String() string { return proto.CompactTextString(m) } func (*UpdateInspectTemplateRequest) ProtoMessage() {} func (*UpdateInspectTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{67} + return fileDescriptor_dlp_536657b46c30097f, []int{69} } func (m *UpdateInspectTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateInspectTemplateRequest.Unmarshal(m, b) @@ -9217,7 +9889,7 @@ type GetInspectTemplateRequest struct { // Resource name of the organization and inspectTemplate to be read, for // example `organizations/433245324/inspectTemplates/432452342` or // projects/project-id/inspectTemplates/432452342. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9227,7 +9899,7 @@ func (m *GetInspectTemplateRequest) Reset() { *m = GetInspectTemplateReq func (m *GetInspectTemplateRequest) String() string { return proto.CompactTextString(m) } func (*GetInspectTemplateRequest) ProtoMessage() {} func (*GetInspectTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{68} + return fileDescriptor_dlp_536657b46c30097f, []int{70} } func (m *GetInspectTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetInspectTemplateRequest.Unmarshal(m, b) @@ -9258,13 +9930,13 @@ func (m *GetInspectTemplateRequest) GetName() string { type ListInspectTemplatesRequest struct { // The parent resource name, for example projects/my-project-id or // organizations/my-org-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional page token to continue retrieval. Comes from previous call // to `ListInspectTemplates`. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional size of the page, can be limited by server. If zero server returns // a page of max size 100. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9274,7 +9946,7 @@ func (m *ListInspectTemplatesRequest) Reset() { *m = ListInspectTemplate func (m *ListInspectTemplatesRequest) String() string { return proto.CompactTextString(m) } func (*ListInspectTemplatesRequest) ProtoMessage() {} func (*ListInspectTemplatesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{69} + return fileDescriptor_dlp_536657b46c30097f, []int{71} } func (m *ListInspectTemplatesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListInspectTemplatesRequest.Unmarshal(m, b) @@ -9318,10 +9990,10 @@ func (m *ListInspectTemplatesRequest) GetPageSize() int32 { // Response message for ListInspectTemplates. type ListInspectTemplatesResponse struct { // List of inspectTemplates, up to page_size in ListInspectTemplatesRequest. - InspectTemplates []*InspectTemplate `protobuf:"bytes,1,rep,name=inspect_templates,json=inspectTemplates" json:"inspect_templates,omitempty"` + InspectTemplates []*InspectTemplate `protobuf:"bytes,1,rep,name=inspect_templates,json=inspectTemplates,proto3" json:"inspect_templates,omitempty"` // If the next page is available then the next page token to be used // in following ListInspectTemplates request. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9331,7 +10003,7 @@ func (m *ListInspectTemplatesResponse) Reset() { *m = ListInspectTemplat func (m *ListInspectTemplatesResponse) String() string { return proto.CompactTextString(m) } func (*ListInspectTemplatesResponse) ProtoMessage() {} func (*ListInspectTemplatesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{70} + return fileDescriptor_dlp_536657b46c30097f, []int{72} } func (m *ListInspectTemplatesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListInspectTemplatesResponse.Unmarshal(m, b) @@ -9370,7 +10042,7 @@ type DeleteInspectTemplateRequest struct { // Resource name of the organization and inspectTemplate to be deleted, for // example `organizations/433245324/inspectTemplates/432452342` or // projects/project-id/inspectTemplates/432452342. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9380,7 +10052,7 @@ func (m *DeleteInspectTemplateRequest) Reset() { *m = DeleteInspectTempl func (m *DeleteInspectTemplateRequest) String() string { return proto.CompactTextString(m) } func (*DeleteInspectTemplateRequest) ProtoMessage() {} func (*DeleteInspectTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{71} + return fileDescriptor_dlp_536657b46c30097f, []int{73} } func (m *DeleteInspectTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteInspectTemplateRequest.Unmarshal(m, b) @@ -9410,14 +10082,14 @@ func (m *DeleteInspectTemplateRequest) GetName() string { // Request message for CreateJobTrigger. type CreateJobTriggerRequest struct { // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The JobTrigger to create. - JobTrigger *JobTrigger `protobuf:"bytes,2,opt,name=job_trigger,json=jobTrigger" json:"job_trigger,omitempty"` + JobTrigger *JobTrigger `protobuf:"bytes,2,opt,name=job_trigger,json=jobTrigger,proto3" json:"job_trigger,omitempty"` // The trigger id can contain uppercase and lowercase letters, // numbers, and hyphens; that is, it must match the regular // expression: `[a-zA-Z\\d-]+`. The maximum length is 100 // characters. Can be empty to allow the system to generate one. - TriggerId string `protobuf:"bytes,3,opt,name=trigger_id,json=triggerId" json:"trigger_id,omitempty"` + TriggerId string `protobuf:"bytes,3,opt,name=trigger_id,json=triggerId,proto3" json:"trigger_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9427,7 +10099,7 @@ func (m *CreateJobTriggerRequest) Reset() { *m = CreateJobTriggerRequest func (m *CreateJobTriggerRequest) String() string { return proto.CompactTextString(m) } func (*CreateJobTriggerRequest) ProtoMessage() {} func (*CreateJobTriggerRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{72} + return fileDescriptor_dlp_536657b46c30097f, []int{74} } func (m *CreateJobTriggerRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateJobTriggerRequest.Unmarshal(m, b) @@ -9472,11 +10144,11 @@ func (m *CreateJobTriggerRequest) GetTriggerId() string { type UpdateJobTriggerRequest struct { // Resource name of the project and the triggeredJob, for example // `projects/dlp-test-project/jobTriggers/53234423`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // New JobTrigger value. - JobTrigger *JobTrigger `protobuf:"bytes,2,opt,name=job_trigger,json=jobTrigger" json:"job_trigger,omitempty"` + JobTrigger *JobTrigger `protobuf:"bytes,2,opt,name=job_trigger,json=jobTrigger,proto3" json:"job_trigger,omitempty"` // Mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9486,7 +10158,7 @@ func (m *UpdateJobTriggerRequest) Reset() { *m = UpdateJobTriggerRequest func (m *UpdateJobTriggerRequest) String() string { return proto.CompactTextString(m) } func (*UpdateJobTriggerRequest) ProtoMessage() {} func (*UpdateJobTriggerRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{73} + return fileDescriptor_dlp_536657b46c30097f, []int{75} } func (m *UpdateJobTriggerRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateJobTriggerRequest.Unmarshal(m, b) @@ -9531,7 +10203,7 @@ func (m *UpdateJobTriggerRequest) GetUpdateMask() *field_mask.FieldMask { type GetJobTriggerRequest struct { // Resource name of the project and the triggeredJob, for example // `projects/dlp-test-project/jobTriggers/53234423`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9541,7 +10213,7 @@ func (m *GetJobTriggerRequest) Reset() { *m = GetJobTriggerRequest{} } func (m *GetJobTriggerRequest) String() string { return proto.CompactTextString(m) } func (*GetJobTriggerRequest) ProtoMessage() {} func (*GetJobTriggerRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{74} + return fileDescriptor_dlp_536657b46c30097f, []int{76} } func (m *GetJobTriggerRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetJobTriggerRequest.Unmarshal(m, b) @@ -9573,7 +10245,7 @@ func (m *GetJobTriggerRequest) GetName() string { // Storage. type CreateDlpJobRequest struct { // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The configuration details for the specific type of job to run. // // Types that are valid to be assigned to Job: @@ -9584,7 +10256,7 @@ type CreateDlpJobRequest struct { // numbers, and hyphens; that is, it must match the regular // expression: `[a-zA-Z\\d-]+`. The maximum length is 100 // characters. Can be empty to allow the system to generate one. - JobId string `protobuf:"bytes,4,opt,name=job_id,json=jobId" json:"job_id,omitempty"` + JobId string `protobuf:"bytes,4,opt,name=job_id,json=jobId,proto3" json:"job_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9594,7 +10266,7 @@ func (m *CreateDlpJobRequest) Reset() { *m = CreateDlpJobRequest{} } func (m *CreateDlpJobRequest) String() string { return proto.CompactTextString(m) } func (*CreateDlpJobRequest) ProtoMessage() {} func (*CreateDlpJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{75} + return fileDescriptor_dlp_536657b46c30097f, []int{77} } func (m *CreateDlpJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateDlpJobRequest.Unmarshal(m, b) @@ -9619,10 +10291,10 @@ type isCreateDlpJobRequest_Job interface { } type CreateDlpJobRequest_InspectJob struct { - InspectJob *InspectJobConfig `protobuf:"bytes,2,opt,name=inspect_job,json=inspectJob,oneof"` + InspectJob *InspectJobConfig `protobuf:"bytes,2,opt,name=inspect_job,json=inspectJob,proto3,oneof"` } type CreateDlpJobRequest_RiskJob struct { - RiskJob *RiskAnalysisJobConfig `protobuf:"bytes,3,opt,name=risk_job,json=riskJob,oneof"` + RiskJob *RiskAnalysisJobConfig `protobuf:"bytes,3,opt,name=risk_job,json=riskJob,proto3,oneof"` } func (*CreateDlpJobRequest_InspectJob) isCreateDlpJobRequest_Job() {} @@ -9739,28 +10411,27 @@ func _CreateDlpJobRequest_OneofSizer(msg proto.Message) (n int) { // Request message for ListJobTriggers. type ListJobTriggersRequest struct { - // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + // The parent resource name, for example `projects/my-project-id`. + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional page token to continue retrieval. Comes from previous call - // to ListJobTriggers. `order_by` and `filter` should not change for - // subsequent calls, but can be omitted if token is specified. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + // to ListJobTriggers. `order_by` field must not + // change for subsequent calls. + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional size of the page, can be limited by a server. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // Optional comma separated list of triggeredJob fields to order by, - // followed by 'asc/desc' postfix, i.e. - // `"create_time asc,name desc,schedule_mode asc"`. This list is - // case-insensitive. + // followed by `asc` or `desc` postfix. This list is case-insensitive, + // default sorting order is ascending, redundant space characters are + // insignificant. // - // Example: `"name asc,schedule_mode desc, status desc"` + // Example: `name asc,update_time, create_time desc` // - // Supported filters keys and values are: + // Supported fields are: // // - `create_time`: corresponds to time the triggeredJob was created. // - `update_time`: corresponds to time the triggeredJob was last updated. - // - `name`: corresponds to JobTrigger's display name. - // - `status`: corresponds to the triggeredJob status. - OrderBy string `protobuf:"bytes,4,opt,name=order_by,json=orderBy" json:"order_by,omitempty"` + // - `name`: corresponds to JobTrigger's name. + OrderBy string `protobuf:"bytes,4,opt,name=order_by,json=orderBy,proto3" json:"order_by,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9770,7 +10441,7 @@ func (m *ListJobTriggersRequest) Reset() { *m = ListJobTriggersRequest{} func (m *ListJobTriggersRequest) String() string { return proto.CompactTextString(m) } func (*ListJobTriggersRequest) ProtoMessage() {} func (*ListJobTriggersRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{76} + return fileDescriptor_dlp_536657b46c30097f, []int{78} } func (m *ListJobTriggersRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListJobTriggersRequest.Unmarshal(m, b) @@ -9821,10 +10492,10 @@ func (m *ListJobTriggersRequest) GetOrderBy() string { // Response message for ListJobTriggers. type ListJobTriggersResponse struct { // List of triggeredJobs, up to page_size in ListJobTriggersRequest. - JobTriggers []*JobTrigger `protobuf:"bytes,1,rep,name=job_triggers,json=jobTriggers" json:"job_triggers,omitempty"` + JobTriggers []*JobTrigger `protobuf:"bytes,1,rep,name=job_triggers,json=jobTriggers,proto3" json:"job_triggers,omitempty"` // If the next page is available then the next page token to be used // in following ListJobTriggers request. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9834,7 +10505,7 @@ func (m *ListJobTriggersResponse) Reset() { *m = ListJobTriggersResponse func (m *ListJobTriggersResponse) String() string { return proto.CompactTextString(m) } func (*ListJobTriggersResponse) ProtoMessage() {} func (*ListJobTriggersResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{77} + return fileDescriptor_dlp_536657b46c30097f, []int{79} } func (m *ListJobTriggersResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListJobTriggersResponse.Unmarshal(m, b) @@ -9872,7 +10543,7 @@ func (m *ListJobTriggersResponse) GetNextPageToken() string { type DeleteJobTriggerRequest struct { // Resource name of the project and the triggeredJob, for example // `projects/dlp-test-project/jobTriggers/53234423`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9882,7 +10553,7 @@ func (m *DeleteJobTriggerRequest) Reset() { *m = DeleteJobTriggerRequest func (m *DeleteJobTriggerRequest) String() string { return proto.CompactTextString(m) } func (*DeleteJobTriggerRequest) ProtoMessage() {} func (*DeleteJobTriggerRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{78} + return fileDescriptor_dlp_536657b46c30097f, []int{80} } func (m *DeleteJobTriggerRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteJobTriggerRequest.Unmarshal(m, b) @@ -9911,16 +10582,16 @@ func (m *DeleteJobTriggerRequest) GetName() string { type InspectJobConfig struct { // The data to scan. - StorageConfig *StorageConfig `protobuf:"bytes,1,opt,name=storage_config,json=storageConfig" json:"storage_config,omitempty"` + StorageConfig *StorageConfig `protobuf:"bytes,1,opt,name=storage_config,json=storageConfig,proto3" json:"storage_config,omitempty"` // How and what to scan for. - InspectConfig *InspectConfig `protobuf:"bytes,2,opt,name=inspect_config,json=inspectConfig" json:"inspect_config,omitempty"` + InspectConfig *InspectConfig `protobuf:"bytes,2,opt,name=inspect_config,json=inspectConfig,proto3" json:"inspect_config,omitempty"` // If provided, will be used as the default for all values in InspectConfig. // `inspect_config` will be merged into the values persisted as part of the // template. - InspectTemplateName string `protobuf:"bytes,3,opt,name=inspect_template_name,json=inspectTemplateName" json:"inspect_template_name,omitempty"` + InspectTemplateName string `protobuf:"bytes,3,opt,name=inspect_template_name,json=inspectTemplateName,proto3" json:"inspect_template_name,omitempty"` // Actions to execute at the completion of the job. Are executed in the order // provided. - Actions []*Action `protobuf:"bytes,4,rep,name=actions" json:"actions,omitempty"` + Actions []*Action `protobuf:"bytes,4,rep,name=actions,proto3" json:"actions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -9930,7 +10601,7 @@ func (m *InspectJobConfig) Reset() { *m = InspectJobConfig{} } func (m *InspectJobConfig) String() string { return proto.CompactTextString(m) } func (*InspectJobConfig) ProtoMessage() {} func (*InspectJobConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{79} + return fileDescriptor_dlp_536657b46c30097f, []int{81} } func (m *InspectJobConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InspectJobConfig.Unmarshal(m, b) @@ -9981,26 +10652,26 @@ func (m *InspectJobConfig) GetActions() []*Action { // Combines all of the information about a DLP job. type DlpJob struct { // The server-assigned name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The type of job. - Type DlpJobType `protobuf:"varint,2,opt,name=type,enum=google.privacy.dlp.v2.DlpJobType" json:"type,omitempty"` + Type DlpJobType `protobuf:"varint,2,opt,name=type,proto3,enum=google.privacy.dlp.v2.DlpJobType" json:"type,omitempty"` // State of a job. - State DlpJob_JobState `protobuf:"varint,3,opt,name=state,enum=google.privacy.dlp.v2.DlpJob_JobState" json:"state,omitempty"` + State DlpJob_JobState `protobuf:"varint,3,opt,name=state,proto3,enum=google.privacy.dlp.v2.DlpJob_JobState" json:"state,omitempty"` // Types that are valid to be assigned to Details: // *DlpJob_RiskDetails // *DlpJob_InspectDetails Details isDlpJob_Details `protobuf_oneof:"details"` // Time when the job was created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Time when the job started. - StartTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time when the job finished. - EndTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // If created by a job trigger, the resource name of the trigger that // instantiated the job. - JobTriggerName string `protobuf:"bytes,10,opt,name=job_trigger_name,json=jobTriggerName" json:"job_trigger_name,omitempty"` + JobTriggerName string `protobuf:"bytes,10,opt,name=job_trigger_name,json=jobTriggerName,proto3" json:"job_trigger_name,omitempty"` // A stream of errors encountered running the job. - Errors []*Error `protobuf:"bytes,11,rep,name=errors" json:"errors,omitempty"` + Errors []*Error `protobuf:"bytes,11,rep,name=errors,proto3" json:"errors,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10010,7 +10681,7 @@ func (m *DlpJob) Reset() { *m = DlpJob{} } func (m *DlpJob) String() string { return proto.CompactTextString(m) } func (*DlpJob) ProtoMessage() {} func (*DlpJob) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{80} + return fileDescriptor_dlp_536657b46c30097f, []int{82} } func (m *DlpJob) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DlpJob.Unmarshal(m, b) @@ -10035,10 +10706,10 @@ type isDlpJob_Details interface { } type DlpJob_RiskDetails struct { - RiskDetails *AnalyzeDataSourceRiskDetails `protobuf:"bytes,4,opt,name=risk_details,json=riskDetails,oneof"` + RiskDetails *AnalyzeDataSourceRiskDetails `protobuf:"bytes,4,opt,name=risk_details,json=riskDetails,proto3,oneof"` } type DlpJob_InspectDetails struct { - InspectDetails *InspectDataSourceDetails `protobuf:"bytes,5,opt,name=inspect_details,json=inspectDetails,oneof"` + InspectDetails *InspectDataSourceDetails `protobuf:"bytes,5,opt,name=inspect_details,json=inspectDetails,proto3,oneof"` } func (*DlpJob_RiskDetails) isDlpJob_Details() {} @@ -10198,7 +10869,7 @@ func _DlpJob_OneofSizer(msg proto.Message) (n int) { // The request message for [DlpJobs.GetDlpJob][]. type GetDlpJobRequest struct { // The name of the DlpJob resource. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10208,7 +10879,7 @@ func (m *GetDlpJobRequest) Reset() { *m = GetDlpJobRequest{} } func (m *GetDlpJobRequest) String() string { return proto.CompactTextString(m) } func (*GetDlpJobRequest) ProtoMessage() {} func (*GetDlpJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{81} + return fileDescriptor_dlp_536657b46c30097f, []int{83} } func (m *GetDlpJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetDlpJobRequest.Unmarshal(m, b) @@ -10238,7 +10909,7 @@ func (m *GetDlpJobRequest) GetName() string { // The request message for listing DLP jobs. type ListDlpJobsRequest struct { // The parent resource name, for example projects/my-project-id. - Parent string `protobuf:"bytes,4,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,4,opt,name=parent,proto3" json:"parent,omitempty"` // Optional. Allows filtering. // // Supported syntax: @@ -10262,13 +10933,13 @@ type ListDlpJobsRequest struct { // * inspected_storage = cloud_storage AND (state = done OR state = canceled) // // The length of this field should be no more than 500 characters. - Filter string `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,1,opt,name=filter,proto3" json:"filter,omitempty"` // The standard list page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The standard list page token. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The type of job. Defaults to `DlpJobType.INSPECT` - Type DlpJobType `protobuf:"varint,5,opt,name=type,enum=google.privacy.dlp.v2.DlpJobType" json:"type,omitempty"` + Type DlpJobType `protobuf:"varint,5,opt,name=type,proto3,enum=google.privacy.dlp.v2.DlpJobType" json:"type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10278,7 +10949,7 @@ func (m *ListDlpJobsRequest) Reset() { *m = ListDlpJobsRequest{} } func (m *ListDlpJobsRequest) String() string { return proto.CompactTextString(m) } func (*ListDlpJobsRequest) ProtoMessage() {} func (*ListDlpJobsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{82} + return fileDescriptor_dlp_536657b46c30097f, []int{84} } func (m *ListDlpJobsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListDlpJobsRequest.Unmarshal(m, b) @@ -10336,9 +11007,9 @@ func (m *ListDlpJobsRequest) GetType() DlpJobType { // The response message for listing DLP jobs. type ListDlpJobsResponse struct { // A list of DlpJobs that matches the specified filter in the request. - Jobs []*DlpJob `protobuf:"bytes,1,rep,name=jobs" json:"jobs,omitempty"` + Jobs []*DlpJob `protobuf:"bytes,1,rep,name=jobs,proto3" json:"jobs,omitempty"` // The standard List next-page token. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10348,7 +11019,7 @@ func (m *ListDlpJobsResponse) Reset() { *m = ListDlpJobsResponse{} } func (m *ListDlpJobsResponse) String() string { return proto.CompactTextString(m) } func (*ListDlpJobsResponse) ProtoMessage() {} func (*ListDlpJobsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{83} + return fileDescriptor_dlp_536657b46c30097f, []int{85} } func (m *ListDlpJobsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListDlpJobsResponse.Unmarshal(m, b) @@ -10385,7 +11056,7 @@ func (m *ListDlpJobsResponse) GetNextPageToken() string { // The request message for canceling a DLP job. type CancelDlpJobRequest struct { // The name of the DlpJob resource to be cancelled. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10395,7 +11066,7 @@ func (m *CancelDlpJobRequest) Reset() { *m = CancelDlpJobRequest{} } func (m *CancelDlpJobRequest) String() string { return proto.CompactTextString(m) } func (*CancelDlpJobRequest) ProtoMessage() {} func (*CancelDlpJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{84} + return fileDescriptor_dlp_536657b46c30097f, []int{86} } func (m *CancelDlpJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CancelDlpJobRequest.Unmarshal(m, b) @@ -10425,7 +11096,7 @@ func (m *CancelDlpJobRequest) GetName() string { // The request message for deleting a DLP job. type DeleteDlpJobRequest struct { // The name of the DlpJob resource to be deleted. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10435,7 +11106,7 @@ func (m *DeleteDlpJobRequest) Reset() { *m = DeleteDlpJobRequest{} } func (m *DeleteDlpJobRequest) String() string { return proto.CompactTextString(m) } func (*DeleteDlpJobRequest) ProtoMessage() {} func (*DeleteDlpJobRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{85} + return fileDescriptor_dlp_536657b46c30097f, []int{87} } func (m *DeleteDlpJobRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteDlpJobRequest.Unmarshal(m, b) @@ -10466,14 +11137,14 @@ func (m *DeleteDlpJobRequest) GetName() string { type CreateDeidentifyTemplateRequest struct { // The parent resource name, for example projects/my-project-id or // organizations/my-org-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // The DeidentifyTemplate to create. - DeidentifyTemplate *DeidentifyTemplate `protobuf:"bytes,2,opt,name=deidentify_template,json=deidentifyTemplate" json:"deidentify_template,omitempty"` + DeidentifyTemplate *DeidentifyTemplate `protobuf:"bytes,2,opt,name=deidentify_template,json=deidentifyTemplate,proto3" json:"deidentify_template,omitempty"` // The template id can contain uppercase and lowercase letters, // numbers, and hyphens; that is, it must match the regular // expression: `[a-zA-Z\\d-]+`. The maximum length is 100 // characters. Can be empty to allow the system to generate one. - TemplateId string `protobuf:"bytes,3,opt,name=template_id,json=templateId" json:"template_id,omitempty"` + TemplateId string `protobuf:"bytes,3,opt,name=template_id,json=templateId,proto3" json:"template_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10483,7 +11154,7 @@ func (m *CreateDeidentifyTemplateRequest) Reset() { *m = CreateDeidentif func (m *CreateDeidentifyTemplateRequest) String() string { return proto.CompactTextString(m) } func (*CreateDeidentifyTemplateRequest) ProtoMessage() {} func (*CreateDeidentifyTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{86} + return fileDescriptor_dlp_536657b46c30097f, []int{88} } func (m *CreateDeidentifyTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateDeidentifyTemplateRequest.Unmarshal(m, b) @@ -10529,11 +11200,11 @@ type UpdateDeidentifyTemplateRequest struct { // Resource name of organization and deidentify template to be updated, for // example `organizations/433245324/deidentifyTemplates/432452342` or // projects/project-id/deidentifyTemplates/432452342. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // New DeidentifyTemplate value. - DeidentifyTemplate *DeidentifyTemplate `protobuf:"bytes,2,opt,name=deidentify_template,json=deidentifyTemplate" json:"deidentify_template,omitempty"` + DeidentifyTemplate *DeidentifyTemplate `protobuf:"bytes,2,opt,name=deidentify_template,json=deidentifyTemplate,proto3" json:"deidentify_template,omitempty"` // Mask to control which fields get updated. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,3,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10543,7 +11214,7 @@ func (m *UpdateDeidentifyTemplateRequest) Reset() { *m = UpdateDeidentif func (m *UpdateDeidentifyTemplateRequest) String() string { return proto.CompactTextString(m) } func (*UpdateDeidentifyTemplateRequest) ProtoMessage() {} func (*UpdateDeidentifyTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{87} + return fileDescriptor_dlp_536657b46c30097f, []int{89} } func (m *UpdateDeidentifyTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateDeidentifyTemplateRequest.Unmarshal(m, b) @@ -10589,7 +11260,7 @@ type GetDeidentifyTemplateRequest struct { // Resource name of the organization and deidentify template to be read, for // example `organizations/433245324/deidentifyTemplates/432452342` or // projects/project-id/deidentifyTemplates/432452342. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10599,7 +11270,7 @@ func (m *GetDeidentifyTemplateRequest) Reset() { *m = GetDeidentifyTempl func (m *GetDeidentifyTemplateRequest) String() string { return proto.CompactTextString(m) } func (*GetDeidentifyTemplateRequest) ProtoMessage() {} func (*GetDeidentifyTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{88} + return fileDescriptor_dlp_536657b46c30097f, []int{90} } func (m *GetDeidentifyTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetDeidentifyTemplateRequest.Unmarshal(m, b) @@ -10630,13 +11301,13 @@ func (m *GetDeidentifyTemplateRequest) GetName() string { type ListDeidentifyTemplatesRequest struct { // The parent resource name, for example projects/my-project-id or // organizations/my-org-id. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Optional page token to continue retrieval. Comes from previous call // to `ListDeidentifyTemplates`. - PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,2,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // Optional size of the page, can be limited by server. If zero server returns // a page of max size 100. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10646,7 +11317,7 @@ func (m *ListDeidentifyTemplatesRequest) Reset() { *m = ListDeidentifyTe func (m *ListDeidentifyTemplatesRequest) String() string { return proto.CompactTextString(m) } func (*ListDeidentifyTemplatesRequest) ProtoMessage() {} func (*ListDeidentifyTemplatesRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{89} + return fileDescriptor_dlp_536657b46c30097f, []int{91} } func (m *ListDeidentifyTemplatesRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListDeidentifyTemplatesRequest.Unmarshal(m, b) @@ -10691,10 +11362,10 @@ func (m *ListDeidentifyTemplatesRequest) GetPageSize() int32 { type ListDeidentifyTemplatesResponse struct { // List of deidentify templates, up to page_size in // ListDeidentifyTemplatesRequest. - DeidentifyTemplates []*DeidentifyTemplate `protobuf:"bytes,1,rep,name=deidentify_templates,json=deidentifyTemplates" json:"deidentify_templates,omitempty"` + DeidentifyTemplates []*DeidentifyTemplate `protobuf:"bytes,1,rep,name=deidentify_templates,json=deidentifyTemplates,proto3" json:"deidentify_templates,omitempty"` // If the next page is available then the next page token to be used // in following ListDeidentifyTemplates request. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10704,7 +11375,7 @@ func (m *ListDeidentifyTemplatesResponse) Reset() { *m = ListDeidentifyT func (m *ListDeidentifyTemplatesResponse) String() string { return proto.CompactTextString(m) } func (*ListDeidentifyTemplatesResponse) ProtoMessage() {} func (*ListDeidentifyTemplatesResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{90} + return fileDescriptor_dlp_536657b46c30097f, []int{92} } func (m *ListDeidentifyTemplatesResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListDeidentifyTemplatesResponse.Unmarshal(m, b) @@ -10743,7 +11414,7 @@ type DeleteDeidentifyTemplateRequest struct { // Resource name of the organization and deidentify template to be deleted, // for example `organizations/433245324/deidentifyTemplates/432452342` or // projects/project-id/deidentifyTemplates/432452342. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -10753,7 +11424,7 @@ func (m *DeleteDeidentifyTemplateRequest) Reset() { *m = DeleteDeidentif func (m *DeleteDeidentifyTemplateRequest) String() string { return proto.CompactTextString(m) } func (*DeleteDeidentifyTemplateRequest) ProtoMessage() {} func (*DeleteDeidentifyTemplateRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_dlp_beb477e7efffa071, []int{91} + return fileDescriptor_dlp_536657b46c30097f, []int{93} } func (m *DeleteDeidentifyTemplateRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteDeidentifyTemplateRequest.Unmarshal(m, b) @@ -10817,6 +11488,9 @@ func init() { proto.RegisterType((*ListInfoTypesRequest)(nil), "google.privacy.dlp.v2.ListInfoTypesRequest") proto.RegisterType((*ListInfoTypesResponse)(nil), "google.privacy.dlp.v2.ListInfoTypesResponse") proto.RegisterType((*RiskAnalysisJobConfig)(nil), "google.privacy.dlp.v2.RiskAnalysisJobConfig") + proto.RegisterType((*QuasiId)(nil), "google.privacy.dlp.v2.QuasiId") + proto.RegisterType((*StatisticalTable)(nil), "google.privacy.dlp.v2.StatisticalTable") + proto.RegisterType((*StatisticalTable_QuasiIdentifierField)(nil), "google.privacy.dlp.v2.StatisticalTable.QuasiIdentifierField") proto.RegisterType((*PrivacyMetric)(nil), "google.privacy.dlp.v2.PrivacyMetric") proto.RegisterType((*PrivacyMetric_NumericalStatsConfig)(nil), "google.privacy.dlp.v2.PrivacyMetric.NumericalStatsConfig") proto.RegisterType((*PrivacyMetric_CategoricalStatsConfig)(nil), "google.privacy.dlp.v2.PrivacyMetric.CategoricalStatsConfig") @@ -10826,6 +11500,7 @@ func init() { proto.RegisterType((*PrivacyMetric_KMapEstimationConfig_TaggedField)(nil), "google.privacy.dlp.v2.PrivacyMetric.KMapEstimationConfig.TaggedField") proto.RegisterType((*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable)(nil), "google.privacy.dlp.v2.PrivacyMetric.KMapEstimationConfig.AuxiliaryTable") proto.RegisterType((*PrivacyMetric_KMapEstimationConfig_AuxiliaryTable_QuasiIdField)(nil), "google.privacy.dlp.v2.PrivacyMetric.KMapEstimationConfig.AuxiliaryTable.QuasiIdField") + proto.RegisterType((*PrivacyMetric_DeltaPresenceEstimationConfig)(nil), "google.privacy.dlp.v2.PrivacyMetric.DeltaPresenceEstimationConfig") proto.RegisterType((*AnalyzeDataSourceRiskDetails)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails") proto.RegisterType((*AnalyzeDataSourceRiskDetails_NumericalStatsResult)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.NumericalStatsResult") proto.RegisterType((*AnalyzeDataSourceRiskDetails_CategoricalStatsResult)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.CategoricalStatsResult") @@ -10839,6 +11514,9 @@ func init() { proto.RegisterType((*AnalyzeDataSourceRiskDetails_KMapEstimationResult)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.KMapEstimationResult") proto.RegisterType((*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationQuasiIdValues)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.KMapEstimationResult.KMapEstimationQuasiIdValues") proto.RegisterType((*AnalyzeDataSourceRiskDetails_KMapEstimationResult_KMapEstimationHistogramBucket)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.KMapEstimationResult.KMapEstimationHistogramBucket") + proto.RegisterType((*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.DeltaPresenceEstimationResult") + proto.RegisterType((*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationQuasiIdValues)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.DeltaPresenceEstimationResult.DeltaPresenceEstimationQuasiIdValues") + proto.RegisterType((*AnalyzeDataSourceRiskDetails_DeltaPresenceEstimationResult_DeltaPresenceEstimationHistogramBucket)(nil), "google.privacy.dlp.v2.AnalyzeDataSourceRiskDetails.DeltaPresenceEstimationResult.DeltaPresenceEstimationHistogramBucket") proto.RegisterType((*ValueFrequency)(nil), "google.privacy.dlp.v2.ValueFrequency") proto.RegisterType((*Value)(nil), "google.privacy.dlp.v2.Value") proto.RegisterType((*QuoteInfo)(nil), "google.privacy.dlp.v2.QuoteInfo") @@ -10939,70 +11617,121 @@ const _ = grpc.SupportPackageIsVersion4 type DlpServiceClient interface { // Finds potentially sensitive info in content. // This method has limits on input size, processing time, and output size. - // [How-to guide for text](/dlp/docs/inspecting-text), [How-to guide for - // images](/dlp/docs/inspecting-images) + // + // When no InfoTypes or CustomInfoTypes are specified in this request, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. + // + // For how to guides, see https://cloud.google.com/dlp/docs/inspecting-images + // and https://cloud.google.com/dlp/docs/inspecting-text, InspectContent(ctx context.Context, in *InspectContentRequest, opts ...grpc.CallOption) (*InspectContentResponse, error) // Redacts potentially sensitive info from an image. // This method has limits on input size, processing time, and output size. - // [How-to guide](/dlp/docs/redacting-sensitive-data-images) + // See https://cloud.google.com/dlp/docs/redacting-sensitive-data-images to + // learn more. + // + // When no InfoTypes or CustomInfoTypes are specified in this request, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. RedactImage(ctx context.Context, in *RedactImageRequest, opts ...grpc.CallOption) (*RedactImageResponse, error) // De-identifies potentially sensitive info from a ContentItem. // This method has limits on input size and output size. - // [How-to guide](/dlp/docs/deidentify-sensitive-data) + // See https://cloud.google.com/dlp/docs/deidentify-sensitive-data to + // learn more. + // + // When no InfoTypes or CustomInfoTypes are specified in this request, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. DeidentifyContent(ctx context.Context, in *DeidentifyContentRequest, opts ...grpc.CallOption) (*DeidentifyContentResponse, error) // Re-identifies content that has been de-identified. + // See + // https://cloud.google.com/dlp/docs/pseudonymization#re-identification_in_free_text_code_example + // to learn more. ReidentifyContent(ctx context.Context, in *ReidentifyContentRequest, opts ...grpc.CallOption) (*ReidentifyContentResponse, error) // Returns a list of the sensitive information types that the DLP API - // supports. For more information, see [Listing supported predefined - // infoTypes](/dlp/docs/listing-infotypes). + // supports. See https://cloud.google.com/dlp/docs/infotypes-reference to + // learn more. ListInfoTypes(ctx context.Context, in *ListInfoTypesRequest, opts ...grpc.CallOption) (*ListInfoTypesResponse, error) // Creates an InspectTemplate for re-using frequently used configuration // for inspecting content, images, and storage. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. CreateInspectTemplate(ctx context.Context, in *CreateInspectTemplateRequest, opts ...grpc.CallOption) (*InspectTemplate, error) // Updates the InspectTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. UpdateInspectTemplate(ctx context.Context, in *UpdateInspectTemplateRequest, opts ...grpc.CallOption) (*InspectTemplate, error) // Gets an InspectTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. GetInspectTemplate(ctx context.Context, in *GetInspectTemplateRequest, opts ...grpc.CallOption) (*InspectTemplate, error) // Lists InspectTemplates. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. ListInspectTemplates(ctx context.Context, in *ListInspectTemplatesRequest, opts ...grpc.CallOption) (*ListInspectTemplatesResponse, error) // Deletes an InspectTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. DeleteInspectTemplate(ctx context.Context, in *DeleteInspectTemplateRequest, opts ...grpc.CallOption) (*empty.Empty, error) // Creates a DeidentifyTemplate for re-using frequently used configuration // for de-identifying content, images, and storage. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. CreateDeidentifyTemplate(ctx context.Context, in *CreateDeidentifyTemplateRequest, opts ...grpc.CallOption) (*DeidentifyTemplate, error) // Updates the DeidentifyTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. UpdateDeidentifyTemplate(ctx context.Context, in *UpdateDeidentifyTemplateRequest, opts ...grpc.CallOption) (*DeidentifyTemplate, error) // Gets a DeidentifyTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. GetDeidentifyTemplate(ctx context.Context, in *GetDeidentifyTemplateRequest, opts ...grpc.CallOption) (*DeidentifyTemplate, error) // Lists DeidentifyTemplates. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. ListDeidentifyTemplates(ctx context.Context, in *ListDeidentifyTemplatesRequest, opts ...grpc.CallOption) (*ListDeidentifyTemplatesResponse, error) // Deletes a DeidentifyTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. DeleteDeidentifyTemplate(ctx context.Context, in *DeleteDeidentifyTemplateRequest, opts ...grpc.CallOption) (*empty.Empty, error) // Creates a job trigger to run DLP actions such as scanning storage for // sensitive information on a set schedule. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. CreateJobTrigger(ctx context.Context, in *CreateJobTriggerRequest, opts ...grpc.CallOption) (*JobTrigger, error) // Updates a job trigger. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. UpdateJobTrigger(ctx context.Context, in *UpdateJobTriggerRequest, opts ...grpc.CallOption) (*JobTrigger, error) // Gets a job trigger. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. GetJobTrigger(ctx context.Context, in *GetJobTriggerRequest, opts ...grpc.CallOption) (*JobTrigger, error) // Lists job triggers. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. ListJobTriggers(ctx context.Context, in *ListJobTriggersRequest, opts ...grpc.CallOption) (*ListJobTriggersResponse, error) // Deletes a job trigger. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. DeleteJobTrigger(ctx context.Context, in *DeleteJobTriggerRequest, opts ...grpc.CallOption) (*empty.Empty, error) // Creates a new job to inspect storage or calculate risk metrics. - // [How-to guide](/dlp/docs/compute-risk-analysis). + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. + // + // When no InfoTypes or CustomInfoTypes are specified in inspect jobs, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. CreateDlpJob(ctx context.Context, in *CreateDlpJobRequest, opts ...grpc.CallOption) (*DlpJob, error) // Lists DlpJobs that match the specified filter in the request. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. ListDlpJobs(ctx context.Context, in *ListDlpJobsRequest, opts ...grpc.CallOption) (*ListDlpJobsResponse, error) // Gets the latest state of a long-running DlpJob. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. GetDlpJob(ctx context.Context, in *GetDlpJobRequest, opts ...grpc.CallOption) (*DlpJob, error) // Deletes a long-running DlpJob. This method indicates that the client is // no longer interested in the DlpJob result. The job will be cancelled if // possible. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. DeleteDlpJob(ctx context.Context, in *DeleteDlpJobRequest, opts ...grpc.CallOption) (*empty.Empty, error) // Starts asynchronous cancellation on a long-running DlpJob. The server // makes a best effort to cancel the DlpJob, but success is not // guaranteed. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. CancelDlpJob(ctx context.Context, in *CancelDlpJobRequest, opts ...grpc.CallOption) (*empty.Empty, error) } @@ -11239,75 +11968,125 @@ func (c *dlpServiceClient) CancelDlpJob(ctx context.Context, in *CancelDlpJobReq return out, nil } -// Server API for DlpService service - +// DlpServiceServer is the server API for DlpService service. type DlpServiceServer interface { // Finds potentially sensitive info in content. // This method has limits on input size, processing time, and output size. - // [How-to guide for text](/dlp/docs/inspecting-text), [How-to guide for - // images](/dlp/docs/inspecting-images) + // + // When no InfoTypes or CustomInfoTypes are specified in this request, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. + // + // For how to guides, see https://cloud.google.com/dlp/docs/inspecting-images + // and https://cloud.google.com/dlp/docs/inspecting-text, InspectContent(context.Context, *InspectContentRequest) (*InspectContentResponse, error) // Redacts potentially sensitive info from an image. // This method has limits on input size, processing time, and output size. - // [How-to guide](/dlp/docs/redacting-sensitive-data-images) + // See https://cloud.google.com/dlp/docs/redacting-sensitive-data-images to + // learn more. + // + // When no InfoTypes or CustomInfoTypes are specified in this request, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. RedactImage(context.Context, *RedactImageRequest) (*RedactImageResponse, error) // De-identifies potentially sensitive info from a ContentItem. // This method has limits on input size and output size. - // [How-to guide](/dlp/docs/deidentify-sensitive-data) + // See https://cloud.google.com/dlp/docs/deidentify-sensitive-data to + // learn more. + // + // When no InfoTypes or CustomInfoTypes are specified in this request, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. DeidentifyContent(context.Context, *DeidentifyContentRequest) (*DeidentifyContentResponse, error) // Re-identifies content that has been de-identified. + // See + // https://cloud.google.com/dlp/docs/pseudonymization#re-identification_in_free_text_code_example + // to learn more. ReidentifyContent(context.Context, *ReidentifyContentRequest) (*ReidentifyContentResponse, error) // Returns a list of the sensitive information types that the DLP API - // supports. For more information, see [Listing supported predefined - // infoTypes](/dlp/docs/listing-infotypes). + // supports. See https://cloud.google.com/dlp/docs/infotypes-reference to + // learn more. ListInfoTypes(context.Context, *ListInfoTypesRequest) (*ListInfoTypesResponse, error) // Creates an InspectTemplate for re-using frequently used configuration // for inspecting content, images, and storage. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. CreateInspectTemplate(context.Context, *CreateInspectTemplateRequest) (*InspectTemplate, error) // Updates the InspectTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. UpdateInspectTemplate(context.Context, *UpdateInspectTemplateRequest) (*InspectTemplate, error) // Gets an InspectTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. GetInspectTemplate(context.Context, *GetInspectTemplateRequest) (*InspectTemplate, error) // Lists InspectTemplates. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. ListInspectTemplates(context.Context, *ListInspectTemplatesRequest) (*ListInspectTemplatesResponse, error) // Deletes an InspectTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates to learn more. DeleteInspectTemplate(context.Context, *DeleteInspectTemplateRequest) (*empty.Empty, error) // Creates a DeidentifyTemplate for re-using frequently used configuration // for de-identifying content, images, and storage. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. CreateDeidentifyTemplate(context.Context, *CreateDeidentifyTemplateRequest) (*DeidentifyTemplate, error) // Updates the DeidentifyTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. UpdateDeidentifyTemplate(context.Context, *UpdateDeidentifyTemplateRequest) (*DeidentifyTemplate, error) // Gets a DeidentifyTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. GetDeidentifyTemplate(context.Context, *GetDeidentifyTemplateRequest) (*DeidentifyTemplate, error) // Lists DeidentifyTemplates. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. ListDeidentifyTemplates(context.Context, *ListDeidentifyTemplatesRequest) (*ListDeidentifyTemplatesResponse, error) // Deletes a DeidentifyTemplate. + // See https://cloud.google.com/dlp/docs/creating-templates-deid to learn + // more. DeleteDeidentifyTemplate(context.Context, *DeleteDeidentifyTemplateRequest) (*empty.Empty, error) // Creates a job trigger to run DLP actions such as scanning storage for // sensitive information on a set schedule. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. CreateJobTrigger(context.Context, *CreateJobTriggerRequest) (*JobTrigger, error) // Updates a job trigger. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. UpdateJobTrigger(context.Context, *UpdateJobTriggerRequest) (*JobTrigger, error) // Gets a job trigger. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. GetJobTrigger(context.Context, *GetJobTriggerRequest) (*JobTrigger, error) // Lists job triggers. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. ListJobTriggers(context.Context, *ListJobTriggersRequest) (*ListJobTriggersResponse, error) // Deletes a job trigger. + // See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. DeleteJobTrigger(context.Context, *DeleteJobTriggerRequest) (*empty.Empty, error) // Creates a new job to inspect storage or calculate risk metrics. - // [How-to guide](/dlp/docs/compute-risk-analysis). + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. + // + // When no InfoTypes or CustomInfoTypes are specified in inspect jobs, the + // system will automatically choose what detectors to run. By default this may + // be all types, but may change over time as detectors are updated. CreateDlpJob(context.Context, *CreateDlpJobRequest) (*DlpJob, error) // Lists DlpJobs that match the specified filter in the request. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. ListDlpJobs(context.Context, *ListDlpJobsRequest) (*ListDlpJobsResponse, error) // Gets the latest state of a long-running DlpJob. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. GetDlpJob(context.Context, *GetDlpJobRequest) (*DlpJob, error) // Deletes a long-running DlpJob. This method indicates that the client is // no longer interested in the DlpJob result. The job will be cancelled if // possible. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. DeleteDlpJob(context.Context, *DeleteDlpJobRequest) (*empty.Empty, error) // Starts asynchronous cancellation on a long-running DlpJob. The server // makes a best effort to cancel the DlpJob, but success is not // guaranteed. + // See https://cloud.google.com/dlp/docs/inspecting-storage and + // https://cloud.google.com/dlp/docs/compute-risk-analysis to learn more. CancelDlpJob(context.Context, *CancelDlpJobRequest) (*empty.Empty, error) } @@ -11875,512 +12654,531 @@ var _DlpService_serviceDesc = grpc.ServiceDesc{ } func init() { - proto.RegisterFile("google/privacy/dlp/v2/dlp.proto", fileDescriptor_dlp_beb477e7efffa071) + proto.RegisterFile("google/privacy/dlp/v2/dlp.proto", fileDescriptor_dlp_536657b46c30097f) } -var fileDescriptor_dlp_beb477e7efffa071 = []byte{ - // 8047 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xdc, 0x7d, 0x6b, 0x6c, 0x1b, 0x59, - 0x77, 0x98, 0x86, 0x14, 0x25, 0xf2, 0xf0, 0x21, 0xea, 0xea, 0x61, 0x99, 0xb6, 0xd7, 0xde, 0xf1, - 0xae, 0xd7, 0xab, 0xb5, 0xa5, 0x5d, 0xed, 0xe3, 0xdb, 0xf5, 0xee, 0xb7, 0x09, 0x45, 0xd2, 0x26, - 0xbd, 0x92, 0x28, 0x0f, 0x29, 0xef, 0xda, 0x59, 0xec, 0x74, 0x44, 0x5e, 0x51, 0x63, 0x91, 0x33, - 0xf4, 0xcc, 0xd0, 0x96, 0x36, 0x49, 0xf1, 0xa5, 0x48, 0x1b, 0xa4, 0xe8, 0x87, 0x06, 0x68, 0xbf, - 0x7e, 0x68, 0xd2, 0x06, 0x69, 0xfb, 0xa3, 0x40, 0x80, 0xb4, 0x28, 0xda, 0x26, 0x68, 0x9b, 0xe6, - 0x47, 0x5a, 0xb4, 0x45, 0x51, 0x14, 0x5f, 0xf0, 0xfd, 0x2a, 0x8a, 0x02, 0xed, 0xd7, 0x07, 0x0a, - 0x6c, 0xfe, 0x14, 0xe8, 0x8f, 0x3e, 0x80, 0xa2, 0xb8, 0x8f, 0x79, 0x72, 0x86, 0x0f, 0xd9, 0x8b, - 0x04, 0xf9, 0x25, 0xce, 0xb9, 0xe7, 0x9c, 0x7b, 0xee, 0x39, 0xe7, 0x9e, 0x7b, 0xef, 0xb9, 0x0f, - 0xc1, 0xd5, 0x8e, 0xae, 0x77, 0xba, 0x78, 0xb3, 0x6f, 0xa8, 0xcf, 0x94, 0xd6, 0xd9, 0x66, 0xbb, - 0xdb, 0xdf, 0x7c, 0xb6, 0x45, 0xfe, 0x6c, 0xf4, 0x0d, 0xdd, 0xd2, 0xd1, 0x0a, 0x43, 0xd8, 0xe0, - 0x08, 0x1b, 0xa4, 0xe4, 0xd9, 0x56, 0xe1, 0x32, 0xa7, 0x53, 0xfa, 0xea, 0xa6, 0xa2, 0x69, 0xba, - 0xa5, 0x58, 0xaa, 0xae, 0x99, 0x8c, 0xa8, 0x70, 0x3d, 0x9c, 0xab, 0x69, 0xe9, 0x86, 0xd2, 0xc1, - 0x1c, 0xe9, 0x15, 0x07, 0x49, 0xb7, 0xf4, 0xc3, 0xc1, 0xd1, 0x66, 0x7b, 0x60, 0x50, 0x2e, 0xbc, - 0xfc, 0x52, 0xb0, 0x1c, 0xf7, 0xfa, 0xd6, 0x19, 0x2f, 0xbc, 0x16, 0x2c, 0x3c, 0x52, 0x71, 0xb7, - 0x2d, 0xf7, 0x14, 0xf3, 0x84, 0x63, 0x5c, 0x0d, 0x62, 0x58, 0x6a, 0x0f, 0x9b, 0x96, 0xd2, 0xe3, - 0x2d, 0x2b, 0x5c, 0xe0, 0x08, 0x46, 0xbf, 0xb5, 0x69, 0x5a, 0x8a, 0x35, 0xb0, 0xa5, 0x5f, 0xe5, - 0x05, 0xd6, 0x59, 0x1f, 0x6f, 0xb6, 0x15, 0x0b, 0x07, 0x04, 0xe2, 0xf0, 0x33, 0xfd, 0xe8, 0x39, - 0xc6, 0x27, 0x61, 0x85, 0xa4, 0x2a, 0xfd, 0xa8, 0xad, 0x70, 0x69, 0xc5, 0x7f, 0x35, 0x07, 0xd9, - 0x9a, 0x66, 0xf6, 0x71, 0xcb, 0x2a, 0xe9, 0xda, 0x91, 0xda, 0x41, 0x9f, 0x02, 0xa8, 0xda, 0x91, - 0x2e, 0x13, 0x74, 0x73, 0x4d, 0xb8, 0x16, 0xbf, 0x99, 0xde, 0xba, 0xba, 0x11, 0xaa, 0xeb, 0x8d, - 0x9a, 0x76, 0xa4, 0x37, 0xcf, 0xfa, 0x58, 0x4a, 0xa9, 0xfc, 0x97, 0x89, 0xaa, 0x90, 0xeb, 0xa9, - 0x9a, 0xdc, 0x55, 0x4f, 0x70, 0x57, 0x3d, 0xd6, 0xf5, 0xf6, 0x5a, 0xec, 0x9a, 0x70, 0x33, 0xb7, - 0xf5, 0x6a, 0x04, 0x8f, 0x1d, 0x07, 0x51, 0xca, 0xf6, 0x54, 0xcd, 0xfd, 0x44, 0xf7, 0x61, 0xae, - 0xab, 0xf6, 0x54, 0xcb, 0x5c, 0x8b, 0x5f, 0x13, 0x6e, 0xa6, 0xb7, 0xb6, 0x22, 0xa5, 0xf0, 0xc8, - 0xbf, 0x71, 0x57, 0xd5, 0xda, 0xaa, 0xd6, 0xd9, 0xa1, 0x94, 0x12, 0xe7, 0x80, 0xae, 0x43, 0x56, - 0xd5, 0x5a, 0xdd, 0x41, 0x1b, 0xcb, 0x4f, 0x07, 0xba, 0x85, 0xd7, 0x66, 0xaf, 0x09, 0x37, 0x93, - 0x52, 0x86, 0x03, 0x1f, 0x10, 0x18, 0xba, 0x05, 0x08, 0x9f, 0x32, 0x24, 0x8f, 0x0a, 0x12, 0x14, - 0x33, 0xcf, 0x4b, 0x6a, 0x4e, 0x43, 0x1f, 0xc0, 0x62, 0x6b, 0x60, 0x5a, 0x7a, 0xcf, 0x8b, 0x3c, - 0x47, 0xf5, 0xf5, 0x7a, 0x84, 0xa4, 0x25, 0x8a, 0xef, 0x68, 0x6d, 0xa1, 0xe5, 0xfb, 0x36, 0xd1, - 0x2e, 0x2c, 0xb4, 0x74, 0xcd, 0xc2, 0x9a, 0x25, 0xeb, 0x7d, 0xea, 0xb6, 0x6b, 0xc9, 0x6b, 0xf1, - 0x9b, 0xb9, 0xad, 0xd7, 0xa2, 0x18, 0x32, 0xec, 0x3a, 0x45, 0x96, 0x72, 0x2d, 0xef, 0xa7, 0x59, - 0xf8, 0x26, 0x06, 0x59, 0x9f, 0x3a, 0xd0, 0x3b, 0xb0, 0xd2, 0x53, 0x4e, 0xe5, 0x23, 0x06, 0x34, - 0xe5, 0x3e, 0x36, 0x64, 0xd5, 0xc2, 0xbd, 0x35, 0xe1, 0x9a, 0x70, 0x33, 0x21, 0xa1, 0x9e, 0x72, - 0xca, 0x09, 0xcc, 0x7d, 0x6c, 0xd4, 0x2c, 0xdc, 0x43, 0xdf, 0x81, 0xb5, 0x21, 0x12, 0x03, 0x3f, - 0x1d, 0x60, 0xd3, 0xa2, 0x96, 0x4d, 0x48, 0x2b, 0x7e, 0x2a, 0x89, 0x15, 0xa2, 0x9f, 0x83, 0xc2, - 0x70, 0x5d, 0xb6, 0xa6, 0xd6, 0xe2, 0x54, 0x51, 0x3f, 0x3d, 0xbd, 0x49, 0x1d, 0xa7, 0xa3, 0x9f, - 0xd2, 0x6a, 0x40, 0x64, 0x5e, 0x58, 0xe8, 0x13, 0xbf, 0xf6, 0x20, 0xa2, 0x4f, 0x20, 0xe5, 0xd6, - 0x2e, 0x50, 0x87, 0x1a, 0xeb, 0xd6, 0x49, 0xdb, 0xad, 0xd1, 0xab, 0x90, 0xf1, 0x36, 0x86, 0xb7, - 0x3c, 0xed, 0xa9, 0x5c, 0xfc, 0x46, 0x80, 0x85, 0xed, 0x33, 0x0b, 0x73, 0x9b, 0x50, 0xe5, 0x95, - 0x61, 0xd6, 0xa9, 0x2f, 0xb7, 0xf5, 0x76, 0x44, 0x7d, 0x01, 0x2a, 0xfa, 0x6d, 0x52, 0x01, 0x28, - 0x35, 0x42, 0x30, 0xdb, 0x56, 0x2c, 0x85, 0x56, 0x9a, 0x91, 0xe8, 0x6f, 0xf1, 0x4f, 0x43, 0xca, - 0x41, 0x43, 0x05, 0x58, 0xdd, 0x7e, 0xd4, 0xac, 0x34, 0xe4, 0xe6, 0xa3, 0xfd, 0x8a, 0x7c, 0xb0, - 0xd7, 0xd8, 0xaf, 0x94, 0x6a, 0x77, 0x6b, 0x95, 0x72, 0x7e, 0x06, 0xa5, 0x20, 0x51, 0xdb, 0x2d, - 0xde, 0xab, 0xe4, 0xe7, 0x50, 0x0e, 0x80, 0xfe, 0x94, 0xef, 0xef, 0x57, 0xee, 0xe5, 0x05, 0x94, - 0x85, 0x14, 0xfb, 0xde, 0xde, 0xdd, 0xcf, 0xc7, 0xdc, 0xcf, 0xfd, 0xbd, 0x7b, 0xf9, 0xb8, 0xfb, - 0xd9, 0x78, 0x78, 0x2f, 0x3f, 0x4b, 0x3e, 0x9b, 0x95, 0x2f, 0x9a, 0xf2, 0x41, 0xf3, 0xee, 0x87, - 0xf9, 0x84, 0xf8, 0x77, 0x05, 0x48, 0x7b, 0x5b, 0xba, 0x0a, 0x89, 0x67, 0x4a, 0x77, 0x80, 0x69, - 0x5f, 0x4d, 0x55, 0x67, 0x24, 0xf6, 0x89, 0xde, 0x83, 0x84, 0xa5, 0x1c, 0x76, 0x59, 0x87, 0x4b, - 0x6f, 0x5d, 0x8e, 0x50, 0x41, 0x93, 0xe0, 0x10, 0x2a, 0x8a, 0x8c, 0x2a, 0x90, 0x3a, 0x3c, 0xb3, - 0x30, 0xf3, 0xcd, 0x04, 0xa5, 0xbc, 0x31, 0x99, 0xf2, 0xaa, 0x33, 0x52, 0x92, 0x90, 0x92, 0xdf, - 0xdb, 0x69, 0x48, 0x11, 0x65, 0x51, 0x36, 0xe2, 0x3f, 0x10, 0x20, 0x41, 0xab, 0x41, 0x1f, 0xc2, - 0xfc, 0x31, 0x56, 0xda, 0xd8, 0xb0, 0xe3, 0xdb, 0x2b, 0x11, 0xbc, 0xef, 0x92, 0xd0, 0x5d, 0x6b, - 0x4b, 0x36, 0x3a, 0x7a, 0x0f, 0x66, 0x0d, 0xfd, 0x39, 0x31, 0x3f, 0x21, 0xbb, 0x36, 0xaa, 0x31, - 0x1b, 0x92, 0xfe, 0x5c, 0xa2, 0xd8, 0x85, 0x8f, 0x21, 0x2e, 0xe9, 0xcf, 0xd1, 0x7b, 0x30, 0x47, - 0x75, 0x62, 0xd7, 0x1a, 0xa5, 0x8b, 0x87, 0x04, 0x49, 0xe2, 0xb8, 0xe2, 0xd7, 0x4e, 0x80, 0x96, - 0xb0, 0x39, 0xe8, 0x5a, 0xe8, 0x0e, 0x24, 0x1d, 0x37, 0x1c, 0x27, 0x3e, 0x45, 0x93, 0x1c, 0x7c, - 0x74, 0x1b, 0x90, 0xd3, 0x1f, 0x2d, 0x63, 0xa0, 0xb5, 0x14, 0x0b, 0xb3, 0x00, 0x9d, 0x94, 0x16, - 0xed, 0x92, 0xa6, 0x5d, 0x20, 0xfe, 0xfb, 0x18, 0xcc, 0x73, 0x26, 0x68, 0x19, 0x12, 0x2c, 0x72, - 0x12, 0x5f, 0x4e, 0x49, 0xec, 0xc3, 0xdf, 0xab, 0x62, 0xd3, 0xf6, 0xaa, 0x22, 0x80, 0x67, 0x9c, - 0x88, 0x4f, 0x3a, 0x4e, 0x78, 0x88, 0xd0, 0xc7, 0x90, 0xec, 0xea, 0x2d, 0x3a, 0x3a, 0x73, 0x17, - 0x8b, 0xaa, 0x7f, 0x87, 0xa3, 0x49, 0x0e, 0x01, 0xfa, 0x18, 0xd2, 0x2d, 0x03, 0x2b, 0x16, 0x96, - 0xc9, 0xb8, 0xb8, 0x36, 0x47, 0xe9, 0x0b, 0x2e, 0x3d, 0x1b, 0x9f, 0x37, 0x9a, 0xf6, 0xf8, 0x2c, - 0x01, 0x43, 0x27, 0x00, 0xf4, 0x53, 0x00, 0x54, 0x07, 0x34, 0xa8, 0xad, 0xcd, 0x53, 0xda, 0x28, - 0x8f, 0xa0, 0xe3, 0x0b, 0x51, 0x80, 0x94, 0x7a, 0x6a, 0xff, 0x14, 0xff, 0xb3, 0x00, 0xc9, 0x1d, - 0x57, 0x14, 0xa0, 0x1e, 0x6f, 0x28, 0x5a, 0xc7, 0x8e, 0x4f, 0x51, 0x0e, 0x22, 0x11, 0x1c, 0x89, - 0xf6, 0x10, 0xfa, 0x13, 0x55, 0xc8, 0xb8, 0xd1, 0xc6, 0x7d, 0x5d, 0xd5, 0x2c, 0xce, 0x21, 0x36, - 0x01, 0x87, 0x9c, 0x43, 0xc4, 0xd8, 0x34, 0x60, 0xd1, 0x1e, 0x7e, 0x6c, 0x15, 0x99, 0x6b, 0xf3, - 0xd4, 0xc5, 0x6e, 0x8c, 0x1e, 0x80, 0x1c, 0xdd, 0xe6, 0x5b, 0x7e, 0x80, 0x29, 0xfe, 0x4e, 0x1c, - 0x16, 0x02, 0x58, 0xe8, 0x75, 0xa0, 0x43, 0x95, 0xa2, 0x6a, 0xd8, 0x90, 0x35, 0xa5, 0x67, 0x3b, - 0x55, 0xd6, 0x81, 0xee, 0x29, 0x3d, 0x8c, 0xf6, 0x61, 0xc1, 0xc0, 0x2d, 0xdd, 0x68, 0x3b, 0xe2, - 0xf0, 0x66, 0x45, 0x8d, 0xaf, 0x12, 0xc5, 0xb6, 0xab, 0xa9, 0xce, 0x48, 0x39, 0xc3, 0x07, 0x41, - 0xbb, 0x90, 0x53, 0x7b, 0x4a, 0x07, 0xbb, 0x0c, 0xd9, 0xd4, 0x22, 0x6a, 0x7c, 0xad, 0x11, 0x64, - 0x0f, 0xbf, 0xac, 0xea, 0x05, 0xa0, 0x87, 0xb0, 0xd8, 0xd6, 0x5b, 0x83, 0x9e, 0x57, 0x63, 0x3c, - 0x5c, 0xbd, 0x11, 0xc1, 0xb1, 0xcc, 0xf1, 0x3d, 0x4c, 0xf3, 0xed, 0x00, 0x0c, 0x7d, 0x06, 0x4b, - 0xae, 0x7e, 0x9c, 0xd9, 0xe1, 0x04, 0xfe, 0x89, 0x1c, 0x32, 0x07, 0x86, 0xde, 0x62, 0x56, 0x65, - 0xcc, 0x9e, 0x61, 0xc3, 0x24, 0x42, 0xce, 0x53, 0x7d, 0xe7, 0x9d, 0x82, 0x87, 0x0c, 0xbe, 0x0d, - 0x6e, 0x77, 0x12, 0xdf, 0x85, 0x7c, 0x50, 0x5a, 0x74, 0x15, 0xd2, 0x47, 0x6a, 0x17, 0xcb, 0xfa, - 0xd1, 0x91, 0x89, 0x2d, 0x6a, 0xb6, 0xb8, 0x04, 0x04, 0x54, 0xa7, 0x10, 0xf1, 0xdf, 0x09, 0x90, - 0xf3, 0x9b, 0x81, 0x74, 0x14, 0x6e, 0xc6, 0x13, 0x7c, 0xc6, 0x5d, 0xfb, 0xda, 0x48, 0x0b, 0x7e, - 0x86, 0xcf, 0xa4, 0x94, 0x61, 0xff, 0x44, 0x1f, 0x91, 0x88, 0x47, 0x26, 0xd1, 0x6a, 0x9b, 0x3b, - 0xc0, 0xd8, 0x80, 0x7d, 0xc4, 0x7e, 0xa0, 0xcf, 0x20, 0x47, 0x47, 0x94, 0x49, 0x0d, 0x4e, 0x43, - 0xb7, 0xe3, 0xcd, 0x59, 0xcb, 0xfb, 0x29, 0xde, 0x82, 0xac, 0xaf, 0x1c, 0x5d, 0x82, 0x94, 0xa1, - 0x3f, 0x97, 0x55, 0xad, 0x8d, 0x4f, 0xb9, 0x2e, 0x92, 0x86, 0xfe, 0xbc, 0x46, 0xbe, 0xc5, 0x4d, - 0x48, 0xb0, 0x6e, 0xb5, 0x0c, 0x09, 0xd3, 0x52, 0x0c, 0x5b, 0x5b, 0xec, 0x03, 0xe5, 0x21, 0x8e, - 0x35, 0xd6, 0x9e, 0xb8, 0x44, 0x7e, 0x8a, 0x8f, 0x21, 0xeb, 0xf3, 0x37, 0x54, 0x83, 0xdc, 0xa1, - 0x3e, 0xa0, 0xe1, 0x57, 0x3e, 0xd4, 0x4f, 0x9d, 0x81, 0x43, 0x8c, 0x1a, 0x0a, 0x39, 0xf2, 0xb6, - 0x7e, 0x2a, 0x65, 0x0f, 0xdd, 0x0f, 0x6c, 0x8a, 0x0a, 0xa4, 0x3d, 0xa5, 0xa4, 0x72, 0x4b, 0xef, - 0xf3, 0x59, 0x1f, 0xf9, 0x49, 0xe6, 0x18, 0x5d, 0x7c, 0x64, 0x4f, 0xe9, 0xe8, 0x6f, 0x22, 0xf8, - 0x73, 0xb5, 0x6d, 0x1d, 0x53, 0x9d, 0x25, 0x24, 0xf6, 0x81, 0x56, 0x61, 0xee, 0x18, 0xab, 0x9d, - 0x63, 0x8b, 0xc6, 0xdb, 0x84, 0xc4, 0xbf, 0xc4, 0x3f, 0x3f, 0x0b, 0x48, 0xc2, 0x6d, 0xa5, 0x65, - 0xd1, 0x56, 0xd8, 0xd3, 0xc0, 0x55, 0x98, 0xeb, 0x2b, 0x06, 0xd6, 0x2c, 0xde, 0xc7, 0xf9, 0x17, - 0xb1, 0x8c, 0xca, 0xc6, 0x35, 0xb9, 0x45, 0xa7, 0x79, 0xdc, 0xb4, 0xaf, 0x4d, 0x32, 0x25, 0x94, - 0xb2, 0xaa, 0x6f, 0xd1, 0xf2, 0x0c, 0x2e, 0xb0, 0x7e, 0x6d, 0x50, 0x01, 0x54, 0x5d, 0xe3, 0x4c, - 0xc9, 0xf4, 0x9d, 0xa8, 0xec, 0xd3, 0x48, 0x7f, 0x0b, 0x0a, 0xbc, 0xc1, 0x3f, 0x38, 0x1f, 0x5e, - 0xdf, 0x8a, 0x1a, 0x02, 0x35, 0x51, 0xc9, 0x3b, 0x4f, 0x99, 0x9f, 0x66, 0x9e, 0xe2, 0xce, 0x52, - 0x0a, 0x7f, 0x20, 0xc0, 0x72, 0x58, 0xa5, 0xe8, 0xd3, 0xe9, 0xa7, 0xac, 0x64, 0xfa, 0xe3, 0x0c, - 0xaf, 0x37, 0x49, 0xfc, 0x24, 0x2c, 0x65, 0xa5, 0xdb, 0x95, 0x2d, 0x7c, 0xca, 0xcc, 0x9b, 0x24, - 0x81, 0x8c, 0x15, 0x14, 0xbb, 0xdd, 0x26, 0x3e, 0xb5, 0xc8, 0x00, 0xe2, 0xd5, 0x5c, 0x57, 0x37, - 0x78, 0x3f, 0xb9, 0x1c, 0x19, 0xf7, 0xbb, 0xba, 0x41, 0xc2, 0xab, 0x23, 0x71, 0x57, 0x37, 0xb6, - 0x93, 0x30, 0x67, 0x29, 0x46, 0x07, 0x5b, 0x62, 0x09, 0x12, 0x14, 0x44, 0x3c, 0xcd, 0xc0, 0x6d, - 0x2a, 0x7d, 0x4c, 0x22, 0x3f, 0x89, 0x57, 0x75, 0x0c, 0x8c, 0x59, 0x2c, 0x8f, 0x49, 0xec, 0x83, - 0xf8, 0xdf, 0xa1, 0x3d, 0x7d, 0x8c, 0x49, 0xf4, 0xb7, 0xd8, 0x82, 0x25, 0x9f, 0x7d, 0xcc, 0xbe, - 0xae, 0x99, 0x98, 0x8c, 0x1e, 0xac, 0x5e, 0xdc, 0x96, 0xa9, 0x59, 0x28, 0xf7, 0x8c, 0xdd, 0x26, - 0xdc, 0xa6, 0xe8, 0x04, 0x0d, 0x9f, 0x5a, 0x06, 0xc3, 0x73, 0x1a, 0x9f, 0x92, 0xb2, 0x0e, 0x94, - 0x34, 0x5d, 0xfc, 0xbf, 0x31, 0x58, 0x2b, 0x63, 0xb5, 0x8d, 0x35, 0x4b, 0x3d, 0x3a, 0xe3, 0x16, - 0x1a, 0xe7, 0xbc, 0x4d, 0x58, 0x6c, 0x3b, 0x34, 0x7e, 0xff, 0x8d, 0x0c, 0xfc, 0xde, 0x3a, 0x88, - 0x4b, 0xe5, 0xdb, 0x01, 0x48, 0x48, 0x97, 0x88, 0x9f, 0xbf, 0x4b, 0x7c, 0x00, 0xb3, 0xd4, 0x2b, - 0xd9, 0xa4, 0x48, 0x1c, 0x3d, 0x7e, 0x53, 0x8f, 0xa4, 0xf8, 0x68, 0x0b, 0x56, 0x6c, 0x21, 0x2c, - 0xdc, 0xeb, 0x77, 0xc9, 0xec, 0x88, 0x0e, 0xd1, 0x09, 0xaa, 0x81, 0x25, 0x5e, 0xd8, 0xe4, 0x65, - 0x74, 0xa0, 0xfe, 0x10, 0xd6, 0x3c, 0xea, 0xf0, 0x93, 0xcd, 0x51, 0xb2, 0x55, 0xb7, 0xdc, 0x4b, - 0x29, 0xfe, 0xba, 0x00, 0x17, 0x43, 0xb4, 0xcf, 0x2d, 0x6d, 0xb7, 0x41, 0x98, 0xb2, 0x0d, 0x35, - 0x48, 0xea, 0xcf, 0xb0, 0xf1, 0x4c, 0xc5, 0xcf, 0xb9, 0x55, 0x6e, 0x47, 0xc5, 0x7b, 0x43, 0xd1, - 0xcc, 0x23, 0xdd, 0xe8, 0xd1, 0x88, 0x5b, 0xe7, 0x44, 0x92, 0x43, 0x4e, 0xdd, 0x43, 0x3a, 0x87, - 0x7b, 0x18, 0x2f, 0xec, 0x1e, 0xc6, 0x9f, 0x24, 0xf7, 0x30, 0xc6, 0xb8, 0x87, 0x11, 0xed, 0x1e, - 0xd2, 0x1f, 0x67, 0xf7, 0xf8, 0xef, 0x02, 0xac, 0xb8, 0x7a, 0x9e, 0xc4, 0x37, 0x5e, 0xea, 0xb8, - 0x67, 0x6b, 0x20, 0xfe, 0xb2, 0xac, 0x38, 0x1b, 0x69, 0x45, 0xf1, 0x21, 0xac, 0x06, 0x5b, 0xca, - 0xed, 0xf0, 0x09, 0xcc, 0x19, 0x74, 0x6d, 0xca, 0x2d, 0x31, 0xa6, 0x29, 0x6c, 0x1d, 0x2b, 0x71, - 0x1a, 0xf1, 0xdf, 0xc4, 0x60, 0xa9, 0x3e, 0xb0, 0xfa, 0x03, 0xab, 0xc1, 0xb2, 0xb0, 0xbc, 0x6d, - 0x9f, 0xd8, 0x99, 0x83, 0xd1, 0x4c, 0xb7, 0xd5, 0xce, 0x83, 0x01, 0x36, 0xce, 0x02, 0x19, 0x84, - 0x2f, 0x21, 0xab, 0x53, 0xa6, 0xb2, 0xd9, 0x3a, 0xc6, 0x3d, 0x85, 0xaf, 0x2e, 0xbf, 0x13, 0xc1, - 0x25, 0x44, 0x00, 0x1b, 0x46, 0xc9, 0xa5, 0x8c, 0xee, 0xf9, 0x12, 0x7f, 0x45, 0x80, 0x8c, 0xb7, - 0x18, 0x5d, 0x81, 0x8b, 0xf5, 0x83, 0xe6, 0xfe, 0x41, 0x53, 0x6e, 0x94, 0xaa, 0x95, 0xdd, 0x62, - 0x20, 0x09, 0xb3, 0x08, 0xd9, 0xed, 0x62, 0xa3, 0x56, 0x92, 0x4b, 0xf5, 0x9d, 0x83, 0xdd, 0xbd, - 0x46, 0x5e, 0x40, 0x0b, 0x90, 0xbe, 0x57, 0x6a, 0x38, 0x80, 0x18, 0x5a, 0x81, 0xc5, 0x72, 0xb1, - 0x59, 0x6c, 0x34, 0xeb, 0x52, 0xc5, 0x01, 0xc7, 0x09, 0x78, 0xbb, 0x76, 0x4f, 0x7e, 0x70, 0x50, - 0x91, 0x1e, 0x39, 0xe0, 0x59, 0x42, 0x5e, 0xdc, 0xd9, 0x71, 0x00, 0x89, 0xed, 0x39, 0x96, 0x6a, - 0x12, 0x5b, 0x6e, 0xe2, 0xab, 0x61, 0x29, 0x96, 0xf9, 0x82, 0x89, 0xaf, 0x65, 0x48, 0xb4, 0xf4, - 0x81, 0x66, 0xf1, 0x89, 0x2a, 0xfb, 0x10, 0x7f, 0x34, 0x0b, 0x6b, 0xdc, 0x9a, 0x65, 0xc5, 0x52, - 0x1a, 0xfa, 0xc0, 0x68, 0xe1, 0x32, 0xb6, 0x14, 0xb5, 0x6b, 0xa2, 0x1e, 0x89, 0x7e, 0xb4, 0x13, - 0xe0, 0xb6, 0x93, 0xc7, 0x64, 0x4e, 0x3e, 0x26, 0xdf, 0x37, 0xc4, 0x6b, 0x43, 0xb2, 0x19, 0xf1, - 0x9c, 0x26, 0x09, 0x8b, 0x7e, 0x08, 0xda, 0x73, 0xbc, 0x8f, 0xf5, 0x82, 0x0f, 0xa6, 0xaf, 0xc3, - 0xeb, 0x8f, 0x85, 0xdf, 0x13, 0x20, 0x1f, 0xac, 0x16, 0x1d, 0xc2, 0x45, 0x53, 0x53, 0xfa, 0xe6, - 0xb1, 0x6e, 0xc9, 0xc1, 0x9e, 0xc3, 0x95, 0x7a, 0x63, 0x74, 0xbd, 0x76, 0x5f, 0x92, 0x2e, 0xd8, - 0x8c, 0x02, 0x05, 0xe8, 0x2e, 0xc0, 0x13, 0xfd, 0xd0, 0x1f, 0xdb, 0xdf, 0x18, 0xcd, 0xf4, 0xbe, - 0x7e, 0xc8, 0x03, 0x43, 0xea, 0x89, 0xfd, 0xb3, 0xf0, 0xf7, 0x05, 0x98, 0xe3, 0xb9, 0xa2, 0x37, - 0x60, 0xa1, 0x6f, 0xe8, 0x2d, 0x6c, 0x9a, 0xb8, 0x2d, 0x93, 0x09, 0xa7, 0xc9, 0x17, 0x21, 0x39, - 0x07, 0x4c, 0xb3, 0x88, 0x24, 0x20, 0x58, 0xba, 0xa5, 0x74, 0x65, 0x6c, 0x5a, 0x6a, 0x4f, 0xb1, - 0x1c, 0x74, 0x66, 0xf6, 0x25, 0x5a, 0x58, 0xb1, 0xcb, 0x18, 0xcd, 0x0e, 0x2c, 0x38, 0x8e, 0x25, - 0x9b, 0xc4, 0xd7, 0x78, 0x56, 0xf7, 0xb5, 0x31, 0xee, 0x45, 0xfd, 0x92, 0x84, 0x32, 0xcf, 0xa7, - 0xf8, 0x6b, 0x02, 0x2c, 0xd9, 0x08, 0x65, 0x6c, 0xb6, 0x0c, 0x95, 0xaa, 0x9e, 0x4c, 0x0c, 0x3d, - 0x19, 0x02, 0xfa, 0x1b, 0xbd, 0x0a, 0x99, 0xb6, 0x6a, 0xf6, 0xbb, 0xca, 0x19, 0x8b, 0x5a, 0x6c, - 0x62, 0x97, 0xe6, 0x30, 0x3a, 0xe6, 0xec, 0x42, 0xc6, 0x1c, 0xf4, 0xfb, 0xba, 0xc1, 0x9a, 0x42, - 0x25, 0xcb, 0x6d, 0xad, 0x8f, 0x93, 0xcc, 0x26, 0xd9, 0x3e, 0x93, 0xd2, 0xa6, 0xfb, 0x21, 0x36, - 0x60, 0x79, 0x47, 0x35, 0x2d, 0x27, 0x55, 0x6f, 0x47, 0xf9, 0xeb, 0x90, 0xed, 0x2a, 0x5a, 0x67, - 0x40, 0xd6, 0x1e, 0x2d, 0xbd, 0x6d, 0x8b, 0x99, 0xb1, 0x81, 0x25, 0xbd, 0x8d, 0xc9, 0x50, 0x70, - 0xa4, 0x76, 0x2d, 0x6c, 0x70, 0x41, 0xf9, 0x97, 0x78, 0x08, 0x2b, 0x01, 0xa6, 0x3c, 0xa0, 0xd6, - 0x42, 0xf6, 0x60, 0xc6, 0x89, 0xee, 0xd1, 0x99, 0x67, 0x3b, 0x46, 0xfc, 0x6f, 0x02, 0xac, 0x48, - 0xaa, 0x79, 0x52, 0xd4, 0x94, 0xee, 0x99, 0xa9, 0x9a, 0x8e, 0xc7, 0x90, 0x81, 0x88, 0xb3, 0x92, - 0x7b, 0xd8, 0x32, 0xd4, 0xd6, 0x98, 0x40, 0xbb, 0xcf, 0x3e, 0x77, 0x29, 0xae, 0x94, 0xed, 0x7b, - 0x3f, 0xd1, 0x3d, 0xc8, 0x98, 0xb4, 0x53, 0xc9, 0x2c, 0x66, 0xc7, 0x26, 0x8f, 0xd9, 0x52, 0x9a, - 0x51, 0xb2, 0xdc, 0xec, 0x77, 0x60, 0x9e, 0xad, 0x28, 0x6c, 0x67, 0xba, 0x12, 0xc1, 0xa3, 0x48, - 0xb1, 0x24, 0x1b, 0x5b, 0xfc, 0xc5, 0x05, 0xc8, 0xfa, 0x44, 0x44, 0x4f, 0x61, 0x55, 0x1b, 0xf4, - 0xb0, 0xa1, 0xb6, 0x94, 0x2e, 0xf3, 0x4f, 0xbb, 0x6f, 0xb1, 0x86, 0x7e, 0x34, 0x49, 0x43, 0x37, - 0xf6, 0x6c, 0x16, 0xd4, 0x4d, 0x99, 0xee, 0xaa, 0x33, 0xd2, 0xb2, 0x16, 0x02, 0x47, 0xcf, 0x61, - 0xad, 0xa5, 0x58, 0xb8, 0xa3, 0x87, 0x54, 0xca, 0x54, 0xf2, 0xf1, 0x44, 0x95, 0x96, 0x5c, 0x26, - 0xfe, 0x6a, 0x57, 0x5b, 0xa1, 0x25, 0x08, 0x03, 0x3a, 0x91, 0x15, 0x4d, 0xd7, 0xce, 0x7a, 0xaa, - 0x75, 0xe6, 0x8f, 0x21, 0xef, 0x4f, 0x54, 0xe5, 0x67, 0x45, 0x9b, 0xda, 0xa9, 0x2c, 0x7f, 0x12, - 0x80, 0x91, 0x6a, 0xba, 0x72, 0x5b, 0xa5, 0x59, 0x24, 0xb7, 0x9a, 0xd9, 0x29, 0xaa, 0xd9, 0x29, - 0xdb, 0xd4, 0x6e, 0x35, 0xdd, 0x00, 0x0c, 0x19, 0x70, 0xe1, 0x44, 0xee, 0x29, 0x7d, 0x3b, 0x1a, - 0xb9, 0xeb, 0x79, 0x9e, 0x5d, 0x9b, 0xcc, 0x74, 0x9f, 0xed, 0x2a, 0xfd, 0x8a, 0xc3, 0xc1, 0x35, - 0xdd, 0x49, 0x08, 0xbc, 0xb0, 0x03, 0xcb, 0x61, 0xa6, 0x46, 0xef, 0x41, 0x82, 0x26, 0x93, 0xb8, - 0xd3, 0x8c, 0xcb, 0x3c, 0x31, 0xe4, 0xc2, 0x1e, 0xac, 0x86, 0xdb, 0xf0, 0x9c, 0xfc, 0xbe, 0x2f, - 0x40, 0x3e, 0x68, 0x21, 0xf4, 0x31, 0xa4, 0x9e, 0x0e, 0x14, 0x53, 0x95, 0xd5, 0xf6, 0xa4, 0x3b, - 0x19, 0x49, 0x4a, 0x50, 0x6b, 0xd3, 0x69, 0x01, 0x99, 0x55, 0x5b, 0x67, 0x6e, 0x56, 0x2d, 0x6a, - 0x5a, 0x50, 0xa1, 0x78, 0x84, 0x1a, 0xf3, 0x5f, 0x85, 0xdf, 0x10, 0x20, 0x1f, 0x34, 0xe5, 0x8b, - 0xc9, 0x53, 0x87, 0x25, 0x13, 0x6b, 0xa6, 0x6a, 0xa9, 0xcf, 0xb0, 0xac, 0x58, 0x96, 0xa1, 0x1e, - 0x0e, 0x2c, 0x3c, 0x61, 0xbe, 0x0f, 0x39, 0xa4, 0x45, 0x9b, 0xb2, 0xf0, 0xcd, 0x1c, 0x2c, 0x87, - 0x79, 0x00, 0x3a, 0x1c, 0x16, 0xb3, 0x72, 0x6e, 0x7f, 0xda, 0x68, 0x2a, 0x9d, 0x0e, 0x6e, 0x53, - 0x41, 0x3c, 0xad, 0xb9, 0x0a, 0x69, 0x03, 0x77, 0x98, 0xdf, 0xb6, 0xed, 0x01, 0x0a, 0x18, 0x88, - 0x8e, 0x09, 0x26, 0xe4, 0x95, 0xc1, 0xa9, 0xda, 0x55, 0x15, 0xe3, 0x8c, 0xc5, 0x4c, 0x3b, 0xe0, - 0x55, 0xcf, 0x2f, 0x4b, 0xd1, 0xe6, 0xc8, 0x02, 0xeb, 0x82, 0xe2, 0xfb, 0x36, 0x0b, 0xff, 0x49, - 0x80, 0xb4, 0x47, 0xde, 0xf3, 0xf9, 0xa2, 0x3f, 0x2d, 0x15, 0x9b, 0x3e, 0x2d, 0x75, 0x15, 0x80, - 0x6f, 0x9c, 0x5b, 0x4a, 0xc7, 0xd9, 0x2f, 0x4c, 0x31, 0x58, 0x53, 0x21, 0x5d, 0x84, 0x20, 0x63, - 0xc3, 0xc0, 0x6d, 0x1e, 0x5b, 0x56, 0x87, 0x72, 0xde, 0x95, 0x5e, 0xdf, 0x3a, 0xe3, 0x6c, 0x29, - 0xe6, 0x76, 0x02, 0xe2, 0x96, 0xd2, 0x29, 0xfc, 0xcf, 0x18, 0xe4, 0xfc, 0x7a, 0x40, 0x77, 0xec, - 0x95, 0x44, 0x7c, 0x8a, 0x51, 0x89, 0xaf, 0x23, 0x8c, 0x61, 0x67, 0x39, 0x78, 0x59, 0x06, 0xda, - 0x78, 0xc0, 0xdc, 0x25, 0xe8, 0x3c, 0xbb, 0x80, 0x0c, 0xdc, 0x55, 0x68, 0x4f, 0x38, 0xa2, 0xf3, - 0x5d, 0xad, 0x75, 0x36, 0x61, 0x4f, 0x58, 0xb4, 0x29, 0xef, 0xda, 0x84, 0x85, 0x16, 0x64, 0xbc, - 0x15, 0x9d, 0xd3, 0xea, 0x57, 0x7c, 0x56, 0x63, 0x0e, 0xed, 0xda, 0xcc, 0x59, 0x7e, 0xfc, 0xaf, - 0x57, 0xe0, 0x32, 0x9d, 0x6b, 0x7c, 0x8d, 0xdd, 0x99, 0x36, 0x99, 0x80, 0xd8, 0xab, 0x83, 0xaf, - 0x60, 0xcd, 0x5d, 0x1d, 0xbc, 0xc0, 0x04, 0x64, 0xd5, 0xe1, 0xe2, 0x1f, 0xf5, 0x1f, 0x83, 0x5b, - 0x22, 0x9f, 0x7b, 0x4e, 0xb2, 0xec, 0xf0, 0x68, 0x78, 0x26, 0x27, 0xdf, 0x13, 0x86, 0xa7, 0x14, - 0xbe, 0xb5, 0x47, 0x54, 0xdf, 0x1d, 0xa5, 0x91, 0xc0, 0x0c, 0x83, 0xcd, 0xdc, 0x87, 0x67, 0x18, - 0x7c, 0x46, 0xff, 0xe7, 0x84, 0xb0, 0x29, 0x06, 0x17, 0x82, 0x75, 0x96, 0xfb, 0xe7, 0x11, 0x22, - 0x38, 0x5a, 0x39, 0x62, 0x0c, 0xcd, 0x38, 0xb8, 0x20, 0x96, 0x7f, 0xc6, 0xc1, 0x25, 0x60, 0xc3, - 0x73, 0xf9, 0x3c, 0x12, 0xb8, 0xc3, 0x9b, 0x53, 0xb7, 0x67, 0x02, 0xe2, 0xd6, 0xea, 0x9d, 0x80, - 0xf0, 0x5a, 0xe7, 0xce, 0x5f, 0xab, 0x3b, 0x88, 0xb9, 0xb5, 0x76, 0x03, 0x30, 0xf4, 0x0b, 0x42, - 0xc8, 0x84, 0x84, 0xd7, 0x3d, 0x7f, 0x7e, 0xc3, 0xfb, 0x43, 0x84, 0x6b, 0xf8, 0x93, 0x10, 0x78, - 0xe1, 0xc7, 0x42, 0x70, 0x82, 0xc2, 0x85, 0xfb, 0x08, 0x52, 0x3d, 0x55, 0x93, 0xd9, 0xe9, 0x8b, - 0xd1, 0x1b, 0xc7, 0xec, 0x64, 0x41, 0xb2, 0xa7, 0x6a, 0xf4, 0x17, 0x25, 0x55, 0x4e, 0x39, 0x69, - 0x6c, 0x22, 0x52, 0xe5, 0x94, 0x91, 0x56, 0x60, 0xe1, 0xe9, 0x40, 0xd1, 0x2c, 0xb5, 0x8b, 0x65, - 0x7e, 0xaa, 0x61, 0x76, 0x82, 0x53, 0x0d, 0x39, 0x9b, 0x88, 0x7e, 0x9a, 0x85, 0xef, 0xcf, 0x0e, - 0x4f, 0x94, 0x78, 0xbb, 0xfe, 0xa1, 0x00, 0xaf, 0x52, 0xce, 0x6e, 0x0c, 0x94, 0x8f, 0x55, 0xd3, - 0xd2, 0x3b, 0x86, 0xd2, 0x93, 0x0f, 0x07, 0xad, 0x13, 0x6c, 0xd9, 0xdb, 0x3b, 0x4f, 0x5e, 0x9e, - 0xcb, 0x0f, 0x81, 0xab, 0x76, 0x9d, 0xdb, 0xb4, 0x4a, 0xe9, 0x15, 0x2a, 0x94, 0x13, 0x5e, 0x03, - 0xc5, 0x66, 0xe1, 0x1f, 0xc7, 0xe0, 0xea, 0x18, 0x1e, 0xe8, 0xbb, 0x70, 0x29, 0xd8, 0xb4, 0xae, - 0xfe, 0x1c, 0x1b, 0x32, 0xdd, 0xb5, 0xe3, 0x6b, 0xf4, 0x35, 0x7f, 0x45, 0x3b, 0x04, 0x81, 0x6e, - 0xe2, 0x85, 0x91, 0x0f, 0xfa, 0x7d, 0x87, 0x3c, 0x16, 0x46, 0x7e, 0x40, 0x10, 0x18, 0xf9, 0x55, - 0x48, 0x33, 0xf5, 0xc9, 0xa6, 0xfa, 0x35, 0x1b, 0x15, 0xe3, 0x12, 0x30, 0x50, 0x43, 0xfd, 0x1a, - 0xa3, 0xfb, 0x90, 0xe5, 0x08, 0x3e, 0xd3, 0xbe, 0x3e, 0xca, 0xb4, 0x4e, 0x45, 0x52, 0x86, 0xd1, - 0x32, 0x0b, 0xa3, 0x5b, 0x80, 0xbc, 0xbc, 0x64, 0x96, 0x4d, 0x4a, 0xd0, 0x3a, 0xf3, 0x1e, 0xcc, - 0x12, 0x81, 0x17, 0xbe, 0x49, 0x78, 0xe7, 0xb9, 0xdc, 0x13, 0x7e, 0x4b, 0x80, 0xeb, 0xf8, 0xe9, - 0x40, 0x7d, 0xa6, 0x74, 0xb1, 0xd6, 0xc2, 0x72, 0xab, 0xab, 0x98, 0x66, 0xa4, 0x2f, 0x7c, 0xf5, - 0x32, 0x82, 0x8f, 0x07, 0x10, 0xb4, 0xff, 0x35, 0x8f, 0x28, 0x25, 0x22, 0xc9, 0x90, 0x07, 0xfc, - 0xaa, 0x00, 0x05, 0x97, 0xbe, 0x12, 0x40, 0x47, 0x77, 0x21, 0xef, 0x4c, 0x29, 0xe4, 0x29, 0x4e, - 0x04, 0xe5, 0xec, 0x09, 0x02, 0xd7, 0xec, 0x7b, 0xb0, 0x3a, 0xac, 0x15, 0x6a, 0x51, 0xe6, 0x00, - 0xcb, 0x41, 0x41, 0x89, 0x6d, 0x0b, 0xbf, 0x1c, 0x87, 0x8b, 0x91, 0x8d, 0x43, 0xf7, 0x41, 0x0c, - 0xe7, 0x19, 0xe2, 0x9f, 0xaf, 0x84, 0xf1, 0xf7, 0x78, 0x69, 0x34, 0xaf, 0x61, 0x67, 0x0d, 0xe5, - 0x35, 0x8d, 0xcb, 0xfe, 0xa2, 0x10, 0xee, 0xb3, 0xf2, 0x4b, 0xf6, 0x86, 0xa0, 0x35, 0x5f, 0xc8, - 0xdb, 0xff, 0xe6, 0xbc, 0x77, 0x15, 0xc5, 0xbd, 0xfd, 0x77, 0x05, 0x78, 0xcb, 0x5d, 0x09, 0x4d, - 0x1a, 0x01, 0xbf, 0x7a, 0x19, 0x83, 0x9f, 0x07, 0x10, 0xf4, 0xfa, 0x37, 0x1c, 0x91, 0x1e, 0x8e, - 0x0e, 0x7f, 0xbf, 0x1d, 0x83, 0x82, 0xcb, 0xe6, 0x8f, 0x97, 0xf3, 0xa3, 0x22, 0x5c, 0xd1, 0x06, - 0x3d, 0xb9, 0xad, 0x9a, 0x96, 0xaa, 0xb5, 0x2c, 0x39, 0xa0, 0x67, 0x93, 0x3b, 0x56, 0x41, 0x1b, - 0xf4, 0xca, 0x1c, 0xa7, 0xe1, 0x6b, 0xb7, 0x89, 0x3e, 0x87, 0x65, 0x4b, 0xef, 0x0f, 0x53, 0x4e, - 0x15, 0x22, 0x91, 0xa5, 0xf7, 0x03, 0x8c, 0x0b, 0x3f, 0x88, 0xc3, 0xc5, 0x48, 0xfd, 0xa3, 0x7d, - 0x78, 0x3d, 0xda, 0x29, 0x86, 0xfb, 0xe6, 0xab, 0x11, 0xe6, 0xf2, 0x74, 0xcf, 0x91, 0x1c, 0x87, - 0x7b, 0x68, 0x14, 0xc7, 0x3f, 0xb2, 0x4e, 0x3a, 0xc2, 0x79, 0x5f, 0x6a, 0x27, 0xfd, 0x6b, 0x89, - 0x60, 0x1e, 0x81, 0x77, 0xd4, 0xbf, 0x25, 0x40, 0x61, 0x68, 0x56, 0xe8, 0xf4, 0x4f, 0xee, 0xd5, - 0x47, 0x2f, 0x6b, 0x62, 0x18, 0x00, 0x06, 0xfb, 0xe7, 0x85, 0x93, 0xf0, 0xe2, 0xc2, 0x5f, 0x11, - 0xe0, 0x92, 0x9f, 0x94, 0x2f, 0x06, 0xb9, 0x32, 0x5e, 0x56, 0x87, 0xdc, 0x84, 0x25, 0x77, 0xef, - 0xc0, 0x59, 0x17, 0x70, 0xe7, 0x41, 0x4e, 0x91, 0x13, 0x48, 0x0b, 0xff, 0x3c, 0x06, 0x57, 0x46, - 0xb6, 0x09, 0x5d, 0x87, 0x2c, 0x99, 0xd9, 0xba, 0xcc, 0x98, 0x6f, 0x67, 0x7a, 0xaa, 0xe6, 0xb0, - 0xa1, 0x48, 0xca, 0xe9, 0x50, 0x8d, 0x99, 0x9e, 0x72, 0xea, 0x22, 0x05, 0x3c, 0x33, 0x31, 0xe4, - 0x99, 0xbf, 0x3c, 0xe4, 0x99, 0xec, 0x24, 0x7f, 0xfb, 0x5b, 0x32, 0x9f, 0xcf, 0x06, 0x13, 0xb9, - 0xe7, 0x7c, 0xb8, 0x7b, 0x6e, 0x27, 0xed, 0xed, 0x2f, 0xf1, 0x31, 0xe4, 0xfc, 0x5d, 0x13, 0x6d, - 0xd9, 0x87, 0xb2, 0x27, 0x59, 0x16, 0xf0, 0x03, 0xdb, 0xe1, 0x1b, 0x7e, 0xbf, 0x1e, 0x87, 0x04, - 0x9b, 0xf8, 0xbf, 0x0e, 0x59, 0x55, 0xb3, 0x70, 0x07, 0x1b, 0x9e, 0x25, 0x47, 0xbc, 0x3a, 0x23, - 0x65, 0x38, 0x98, 0xa1, 0xbd, 0x0a, 0xe9, 0xa3, 0xae, 0xae, 0x58, 0x9e, 0xc5, 0x85, 0x50, 0x9d, - 0x91, 0x80, 0x02, 0x19, 0xca, 0x75, 0xc8, 0x98, 0x96, 0xa1, 0x6a, 0x1d, 0xd9, 0x7f, 0x72, 0x3c, - 0xcd, 0xa0, 0x4e, 0x75, 0x87, 0xba, 0xde, 0xc5, 0x8a, 0xbd, 0xc2, 0x99, 0xe5, 0x27, 0x98, 0x32, - 0x1c, 0xec, 0x2c, 0x47, 0x9c, 0x73, 0x92, 0x1c, 0x31, 0x31, 0xee, 0xb4, 0x64, 0x75, 0x46, 0xca, - 0x39, 0x44, 0x8c, 0xcd, 0x77, 0x00, 0x08, 0x84, 0x73, 0x98, 0xf3, 0xe7, 0x9e, 0xac, 0xb3, 0x3e, - 0xa6, 0xd4, 0xf5, 0xa3, 0xb2, 0x72, 0x56, 0x9d, 0x91, 0x52, 0x04, 0x97, 0x11, 0x6e, 0x01, 0xb4, - 0x15, 0xcb, 0x26, 0x64, 0x6b, 0xc2, 0x45, 0x1f, 0x61, 0x59, 0xb1, 0x30, 0xa1, 0x21, 0x68, 0x8c, - 0xa6, 0x04, 0x8b, 0x6d, 0xe5, 0x4c, 0xd6, 0x8f, 0xe4, 0xe7, 0x18, 0x9f, 0x70, 0xd2, 0x24, 0xdd, - 0xa6, 0x5e, 0x0d, 0x90, 0x9e, 0xd5, 0x8f, 0x3e, 0xc7, 0xf8, 0x84, 0x48, 0xdc, 0xb6, 0x3f, 0x28, - 0x13, 0x27, 0xef, 0xf2, 0x33, 0x90, 0x72, 0x0e, 0x19, 0xa3, 0x4f, 0xe9, 0xb9, 0x77, 0x7e, 0xaa, - 0x79, 0x74, 0x86, 0xae, 0xcc, 0x8f, 0x33, 0x57, 0x67, 0xa4, 0x64, 0x9b, 0xff, 0xde, 0xce, 0x41, - 0xa6, 0xaf, 0x18, 0x26, 0x6e, 0xb3, 0xcb, 0x32, 0xe2, 0x5f, 0x8c, 0x41, 0xd2, 0x46, 0x44, 0xaf, - 0xd3, 0xdb, 0x08, 0xb6, 0x4f, 0x0d, 0x37, 0x92, 0x5e, 0x50, 0xc0, 0xe8, 0x03, 0x48, 0x7b, 0x5a, - 0xc7, 0x2f, 0x01, 0x45, 0xb4, 0x8b, 0x68, 0x85, 0xff, 0x44, 0xeb, 0x30, 0x4b, 0xc5, 0x8e, 0x8f, - 0x52, 0xbe, 0x44, 0x71, 0x50, 0x05, 0xa8, 0x09, 0xe4, 0xaf, 0x75, 0xcd, 0xbe, 0x60, 0x70, 0x73, - 0x4c, 0x3b, 0x29, 0x8f, 0xc7, 0xba, 0x86, 0xa5, 0xa4, 0xc5, 0x7f, 0x15, 0xde, 0x81, 0xa4, 0x0d, - 0x45, 0xaf, 0x43, 0x8e, 0x9d, 0x6d, 0x95, 0x7b, 0xaa, 0x36, 0xb0, 0x37, 0x4c, 0x13, 0x52, 0x96, - 0x41, 0x77, 0x19, 0x50, 0xfc, 0xdf, 0x02, 0xe4, 0x83, 0x47, 0x76, 0x50, 0x17, 0x2e, 0xba, 0x1b, - 0xa2, 0x96, 0xef, 0xe8, 0x88, 0xc9, 0xd5, 0xb5, 0x31, 0x26, 0x51, 0xea, 0x3f, 0x70, 0x62, 0x56, - 0x67, 0xa4, 0x0b, 0x6a, 0x78, 0x11, 0xc2, 0xb0, 0xca, 0x8f, 0xd5, 0x06, 0xab, 0x62, 0x16, 0xbf, - 0x35, 0xf2, 0x88, 0xed, 0x70, 0x45, 0x2b, 0x46, 0x58, 0xc1, 0x76, 0x1e, 0x72, 0x7e, 0xfe, 0xe2, - 0x4f, 0xe6, 0xe1, 0xc2, 0xbe, 0xa1, 0xf6, 0xe8, 0x64, 0xc0, 0x8f, 0x8e, 0x24, 0xc8, 0x19, 0xb8, - 0xdf, 0x55, 0xc8, 0x94, 0xcc, 0xbb, 0xd7, 0xf6, 0x66, 0xa4, 0x30, 0x14, 0x99, 0xc7, 0x33, 0xbe, - 0x41, 0x93, 0xe5, 0x2c, 0xb8, 0x5a, 0xef, 0x03, 0x3f, 0xd9, 0xe7, 0xdf, 0x49, 0xbb, 0x3e, 0xf2, - 0x48, 0xa7, 0xc3, 0x2c, 0x63, 0x78, 0xbe, 0xd1, 0x9f, 0x82, 0x95, 0xd6, 0xb1, 0x42, 0x8f, 0xff, - 0x19, 0xf4, 0x4e, 0x9e, 0x7f, 0xab, 0x2c, 0x6a, 0x93, 0xb5, 0x64, 0xd3, 0xec, 0x2a, 0xe6, 0x89, - 0xc3, 0x7a, 0xa9, 0x35, 0x0c, 0x46, 0x16, 0x5c, 0x69, 0x19, 0x67, 0x7d, 0x4b, 0x97, 0x6d, 0x45, - 0x1c, 0x1d, 0x9d, 0xca, 0x47, 0x7d, 0xec, 0xdf, 0x2d, 0x8b, 0xba, 0x0b, 0x54, 0xa2, 0xb4, 0x5c, - 0x2d, 0x77, 0x8f, 0x4e, 0xef, 0xf6, 0x5d, 0xbd, 0x5c, 0x6c, 0x45, 0x15, 0xa2, 0x3e, 0x5c, 0x3a, - 0x52, 0x4f, 0x71, 0x9b, 0xad, 0xaf, 0xd8, 0x20, 0x41, 0x22, 0xab, 0x6f, 0xd7, 0x6c, 0x33, 0x32, - 0xd3, 0x7b, 0x8a, 0xdb, 0x64, 0x18, 0xdc, 0xb6, 0xe9, 0x9c, 0x2a, 0xd7, 0x8e, 0x22, 0xca, 0x50, - 0x03, 0xf2, 0x43, 0xd5, 0xcc, 0x8d, 0x3e, 0x01, 0x3b, 0xc4, 0x7d, 0xe1, 0x30, 0xc0, 0xd4, 0x82, - 0x2b, 0xb6, 0xd6, 0x9e, 0xab, 0xd6, 0xb1, 0x7b, 0x5f, 0xcc, 0xae, 0x61, 0x7e, 0xa4, 0xf2, 0xb8, - 0x66, 0x3e, 0x57, 0xad, 0x63, 0xbb, 0x43, 0xb9, 0xca, 0x33, 0xa2, 0x0a, 0xd1, 0x03, 0xc8, 0xd3, - 0x30, 0xd2, 0x57, 0x0c, 0xc7, 0xc7, 0x92, 0x23, 0x2f, 0x1a, 0x90, 0x70, 0xb1, 0xaf, 0x18, 0xae, - 0x97, 0xd1, 0x81, 0xc4, 0x85, 0xa0, 0xcf, 0x01, 0x71, 0x2f, 0x38, 0x56, 0xcc, 0x63, 0x9b, 0x69, - 0x6a, 0xe4, 0x99, 0x0e, 0x66, 0xfa, 0xaa, 0x62, 0x1e, 0xbb, 0x5b, 0xa3, 0xad, 0x00, 0x8c, 0x9e, - 0x3c, 0x25, 0xa1, 0xdd, 0x3c, 0x56, 0x8f, 0x1c, 0x61, 0xd3, 0x23, 0xf5, 0x4e, 0x42, 0x5f, 0x83, - 0xa0, 0xbb, 0x7a, 0x6f, 0xfb, 0x41, 0x21, 0x9d, 0xfc, 0x1b, 0x01, 0x72, 0xfe, 0x56, 0xa2, 0x87, - 0xb0, 0x40, 0x35, 0x64, 0xe9, 0x32, 0x3f, 0x42, 0xcb, 0xef, 0xb5, 0x6d, 0x4c, 0xa4, 0x25, 0xe7, - 0x53, 0xca, 0x12, 0x36, 0x4d, 0xbd, 0xc2, 0x98, 0x88, 0xdf, 0x13, 0x58, 0xfc, 0x25, 0x65, 0xe8, - 0x22, 0xac, 0x34, 0x6b, 0xbb, 0x15, 0x79, 0xbf, 0x28, 0x35, 0x03, 0x87, 0xa8, 0x92, 0x30, 0xfb, - 0xa8, 0x52, 0x94, 0xf2, 0x02, 0x4a, 0x41, 0x62, 0xb7, 0xbe, 0xd7, 0xac, 0xe6, 0x63, 0x28, 0x0f, - 0x99, 0x72, 0xf1, 0x91, 0x5c, 0xbf, 0x2b, 0x33, 0x48, 0x1c, 0x2d, 0x40, 0x9a, 0x43, 0x3e, 0xaf, - 0x54, 0x3e, 0xcb, 0xcf, 0x12, 0x14, 0xf2, 0x8b, 0x40, 0x28, 0x7d, 0x82, 0xa0, 0x54, 0xeb, 0x07, - 0x12, 0x81, 0x94, 0x8b, 0x8f, 0xf2, 0x73, 0x62, 0x03, 0xf2, 0x41, 0xed, 0xa3, 0x9f, 0x02, 0xe0, - 0x26, 0x1c, 0x7f, 0x6d, 0x81, 0x11, 0xd3, 0x6b, 0x0b, 0x2d, 0xfb, 0xa7, 0x58, 0x07, 0x34, 0x1c, - 0xde, 0xd0, 0x47, 0x90, 0xd2, 0xf0, 0xf3, 0x69, 0xd2, 0xb5, 0x1a, 0x7e, 0x4e, 0x7f, 0x89, 0x97, - 0xe0, 0x62, 0xa4, 0x87, 0x8b, 0x39, 0xc8, 0x78, 0x23, 0x9f, 0xf8, 0x93, 0x18, 0x64, 0x49, 0xd8, - 0x32, 0x9b, 0x7a, 0xad, 0xa3, 0xe9, 0x06, 0x46, 0x1b, 0x80, 0x9c, 0x80, 0x65, 0x12, 0x2b, 0x9a, - 0x27, 0x2a, 0xbb, 0x03, 0x90, 0xa2, 0xae, 0xe6, 0x94, 0x35, 0xf5, 0xc6, 0x89, 0xda, 0x47, 0x67, - 0x70, 0xa9, 0xa5, 0xf7, 0x7a, 0xba, 0x26, 0xfb, 0xc9, 0x54, 0xca, 0x8e, 0x8f, 0xe8, 0x1f, 0x8e, - 0x88, 0x98, 0x4e, 0xd5, 0x1b, 0x25, 0xca, 0xc7, 0x07, 0x23, 0xc1, 0xa5, 0xe5, 0x80, 0xed, 0x8a, - 0x59, 0x99, 0xf8, 0x43, 0x01, 0x96, 0x42, 0x68, 0xd0, 0x0d, 0x10, 0x4b, 0xf5, 0xdd, 0xdd, 0xfa, - 0x9e, 0x5c, 0xaa, 0x16, 0xa5, 0x86, 0xdc, 0xac, 0xcb, 0xb5, 0x7b, 0x7b, 0x75, 0x29, 0x78, 0xe9, - 0x31, 0x0d, 0xf3, 0x7b, 0x07, 0xbb, 0x15, 0xa9, 0x56, 0xca, 0x0b, 0x68, 0x19, 0xf2, 0xc5, 0x9d, - 0xfd, 0x6a, 0x51, 0x3e, 0xd8, 0xdf, 0xaf, 0x48, 0x72, 0xa9, 0xd8, 0xa8, 0xe4, 0x63, 0x2e, 0x74, - 0xa7, 0xfe, 0xb9, 0x0d, 0xa5, 0xce, 0xb3, 0x7f, 0xb0, 0x57, 0x6a, 0x1e, 0x14, 0x9b, 0xb5, 0xfa, - 0x5e, 0x7e, 0x16, 0xe5, 0x00, 0x3e, 0xaf, 0xd6, 0x9a, 0x95, 0xc6, 0x7e, 0xb1, 0x54, 0xc9, 0x27, - 0xb6, 0x33, 0x00, 0xae, 0x36, 0xc4, 0xff, 0x4a, 0xe4, 0x0c, 0x19, 0x04, 0xde, 0x82, 0x45, 0x32, - 0xb8, 0xd0, 0xd0, 0x68, 0x17, 0xf3, 0xa3, 0x41, 0x79, 0x5e, 0xe0, 0x90, 0xa1, 0xd7, 0x20, 0xa7, - 0x0d, 0x7a, 0x87, 0xd8, 0x20, 0xca, 0x25, 0xa5, 0xfc, 0x12, 0x46, 0x86, 0x41, 0x9b, 0x3a, 0x61, - 0x4c, 0xd6, 0x39, 0x06, 0x26, 0xeb, 0x5b, 0x2c, 0xeb, 0x46, 0x1b, 0xb3, 0x03, 0xfa, 0x49, 0x32, - 0xbc, 0x51, 0x60, 0x9d, 0xc0, 0xd0, 0x43, 0x58, 0x0e, 0xb5, 0xd5, 0xec, 0xc8, 0x73, 0x59, 0x3e, - 0x1d, 0x4b, 0xa8, 0x35, 0x6c, 0x8f, 0x7f, 0x24, 0xc0, 0x5a, 0xd4, 0x28, 0x81, 0xbe, 0x0b, 0xe9, - 0x60, 0x02, 0x62, 0x9c, 0x4f, 0x43, 0xd7, 0x9b, 0xcc, 0x4e, 0x07, 0xb3, 0x0d, 0x63, 0xc9, 0x07, - 0x23, 0x93, 0x0e, 0x82, 0x77, 0x69, 0x27, 0xfe, 0x4a, 0x0c, 0x16, 0x82, 0x22, 0xdf, 0x83, 0x79, - 0x3b, 0x7d, 0xc6, 0xd6, 0xba, 0xb7, 0x27, 0x1b, 0xb3, 0xf8, 0xb7, 0x64, 0x53, 0xd3, 0xb3, 0x76, - 0x7c, 0xb5, 0xba, 0x01, 0xf1, 0x9e, 0xaa, 0x4d, 0xd4, 0x7c, 0x82, 0x48, 0xf1, 0x95, 0xd3, 0x89, - 0xda, 0x4b, 0x10, 0x51, 0x0d, 0x16, 0xf9, 0x10, 0x46, 0xef, 0x9b, 0xb9, 0x6b, 0xa6, 0x71, 0xd4, - 0x79, 0x0f, 0x19, 0x0b, 0x24, 0xbf, 0x3f, 0x0b, 0x17, 0x23, 0x27, 0x1a, 0x2f, 0x1c, 0xf8, 0xd0, - 0x87, 0x30, 0x4f, 0xaf, 0x01, 0xf2, 0x2b, 0x17, 0x13, 0x5c, 0xd7, 0xe2, 0xe8, 0xc8, 0x84, 0x05, - 0x1e, 0x72, 0x94, 0x6e, 0xff, 0x58, 0x39, 0xc4, 0x6c, 0x4f, 0x33, 0x17, 0xb9, 0xbf, 0x16, 0xd9, - 0x8a, 0x8d, 0xbb, 0x47, 0xa7, 0x2c, 0x82, 0xec, 0xd1, 0xed, 0xf0, 0x22, 0xe7, 0x47, 0xc6, 0x6a, - 0x56, 0x85, 0x0d, 0x41, 0x6f, 0x02, 0xbf, 0x88, 0xef, 0x56, 0x9a, 0xe0, 0x41, 0x31, 0xc7, 0x0a, - 0x1c, 0xd4, 0x55, 0x48, 0x18, 0x4a, 0x5b, 0x3d, 0xa5, 0x33, 0x9d, 0x44, 0x75, 0x46, 0x62, 0x9f, - 0xf4, 0xf0, 0xca, 0xc0, 0x30, 0xf4, 0x8e, 0x62, 0x79, 0xde, 0x0e, 0xe0, 0x93, 0x88, 0xb1, 0xa7, - 0x6d, 0x17, 0x1d, 0x5a, 0x1b, 0x24, 0xfe, 0x65, 0x01, 0x2e, 0x44, 0xb4, 0x00, 0xad, 0xc3, 0x8d, - 0xbb, 0x77, 0xbf, 0x90, 0x79, 0x20, 0xdc, 0x2b, 0x36, 0x6b, 0x0f, 0x2b, 0x32, 0x8d, 0x65, 0xdb, - 0x95, 0xe6, 0xa8, 0x40, 0x48, 0x86, 0xbd, 0xca, 0x17, 0xc5, 0x72, 0xa5, 0x54, 0xdb, 0x2d, 0xee, - 0xe4, 0x63, 0xe8, 0x32, 0xac, 0xb9, 0x31, 0x91, 0xb1, 0x90, 0x6d, 0xf4, 0x38, 0x5a, 0x84, 0xac, - 0x1f, 0x34, 0xbb, 0x0d, 0x90, 0xb4, 0x75, 0x24, 0xfe, 0x1f, 0x01, 0x52, 0x8e, 0xf9, 0x51, 0x0d, - 0x52, 0x74, 0x06, 0xa1, 0xda, 0x27, 0xde, 0xa3, 0xe7, 0xfc, 0x4d, 0x1b, 0xcf, 0xa1, 0xa6, 0x6b, - 0x69, 0x1b, 0x4a, 0x58, 0x0d, 0xb4, 0xe7, 0x86, 0xd2, 0xef, 0x63, 0x3b, 0x1c, 0x44, 0xb1, 0x3a, - 0xb0, 0xf1, 0x7c, 0xac, 0x1c, 0x6a, 0xb4, 0x0b, 0xe9, 0x93, 0x9e, 0x29, 0xdb, 0xcc, 0x46, 0x4f, - 0xf2, 0x3f, 0xeb, 0x99, 0x9f, 0x0f, 0x73, 0x83, 0x13, 0x07, 0xbc, 0x9d, 0x84, 0x39, 0x76, 0xa2, - 0x40, 0xbc, 0x09, 0x68, 0xb8, 0x19, 0x61, 0x67, 0x55, 0xc5, 0x1b, 0x80, 0x86, 0xa5, 0x44, 0x79, - 0x88, 0xdb, 0x9d, 0x2b, 0x23, 0x91, 0x9f, 0xe2, 0x57, 0xb0, 0x14, 0x22, 0x00, 0x09, 0x6f, 0x9c, - 0x58, 0x76, 0x09, 0x80, 0x83, 0x08, 0xc2, 0x0d, 0x58, 0x70, 0x7b, 0xab, 0xf7, 0x38, 0x6c, 0xd6, - 0xe9, 0x90, 0xf4, 0xf8, 0xfe, 0x1f, 0x0a, 0xb0, 0x10, 0x98, 0x09, 0xa2, 0x9b, 0x90, 0xf7, 0x84, - 0x5e, 0xb9, 0xad, 0x9c, 0xd9, 0xcb, 0xdd, 0x9c, 0x1b, 0x61, 0xcb, 0xca, 0x99, 0x49, 0x30, 0x3d, - 0x31, 0x9e, 0x61, 0xb2, 0x51, 0x2a, 0xe7, 0x86, 0x72, 0x8a, 0xe9, 0xe9, 0xfc, 0xf1, 0xe9, 0x3a, - 0x7f, 0xd1, 0x17, 0x77, 0x66, 0x27, 0x8b, 0x3b, 0xf4, 0xe4, 0x90, 0xfd, 0x41, 0x0c, 0xd4, 0xc3, - 0xd6, 0xb1, 0xde, 0x16, 0x7f, 0x14, 0x83, 0x0b, 0x11, 0x8b, 0x6a, 0xa4, 0xc3, 0xc2, 0xf0, 0xea, - 0x7c, 0xd4, 0x31, 0xb0, 0x08, 0x46, 0x11, 0x70, 0x29, 0xc8, 0xbd, 0xf0, 0xcf, 0x04, 0x58, 0x0d, - 0xc7, 0x7d, 0xe1, 0xe7, 0x56, 0x54, 0x58, 0xeb, 0xdb, 0x6b, 0xf1, 0x40, 0x22, 0x80, 0xf7, 0x9d, - 0x8d, 0xe8, 0xe3, 0x34, 0x61, 0x4b, 0x78, 0xe9, 0x42, 0x3f, 0xbc, 0x40, 0xfc, 0x5e, 0x1c, 0x96, - 0xa8, 0xd1, 0x02, 0x4d, 0xf8, 0x00, 0xe6, 0xe8, 0x11, 0xa1, 0x49, 0xcf, 0xfc, 0x71, 0x6c, 0x54, - 0x86, 0x54, 0x4b, 0xd7, 0xda, 0xaa, 0xe7, 0x5a, 0xee, 0x8d, 0x91, 0x39, 0x8b, 0x92, 0x8d, 0x2d, - 0xb9, 0x84, 0xe8, 0x64, 0x84, 0x02, 0x66, 0xcf, 0xa3, 0x80, 0xea, 0x4c, 0xa4, 0x0a, 0x46, 0x67, - 0x78, 0x12, 0x2f, 0x39, 0xc3, 0x13, 0xb2, 0x2a, 0xfb, 0xb1, 0x00, 0x2b, 0xa1, 0xf9, 0x1b, 0x24, - 0xc3, 0x0a, 0xbb, 0x23, 0x1d, 0xee, 0xd9, 0xeb, 0xa3, 0x6c, 0x12, 0x70, 0x80, 0xe5, 0xa3, 0x61, - 0xa0, 0x89, 0x1e, 0xc1, 0x12, 0x4f, 0x37, 0x99, 0x83, 0x7e, 0xdf, 0xc0, 0xa6, 0xc9, 0x73, 0x4d, - 0xf1, 0x11, 0x59, 0x37, 0x26, 0x6b, 0xc3, 0x25, 0x90, 0x90, 0x11, 0x04, 0x99, 0xe2, 0x23, 0x58, - 0x1c, 0x42, 0xf4, 0x7b, 0x87, 0x70, 0x4e, 0xef, 0x10, 0x7f, 0x23, 0x01, 0x0b, 0x81, 0x62, 0xd4, - 0x84, 0x34, 0x3e, 0x75, 0x5b, 0x30, 0xfa, 0x71, 0xa1, 0x00, 0xf1, 0x46, 0xc5, 0xa5, 0x94, 0xbc, - 0x6c, 0x0a, 0xbf, 0x4b, 0x86, 0x43, 0xa7, 0x8e, 0xf3, 0x9d, 0xb1, 0xab, 0x40, 0x52, 0xef, 0x63, - 0x43, 0xb1, 0xf8, 0xfd, 0xdb, 0xdc, 0x88, 0xbc, 0x59, 0x97, 0xda, 0x45, 0xe9, 0xd6, 0x39, 0x81, - 0xe4, 0x90, 0xba, 0x69, 0xff, 0xd9, 0x89, 0xd3, 0xfe, 0x85, 0xaf, 0x00, 0x1c, 0xe9, 0x4d, 0xb4, - 0x0f, 0xe0, 0xe8, 0xd0, 0x76, 0xa1, 0xb7, 0x27, 0xd4, 0x90, 0x6b, 0x07, 0x0f, 0x8f, 0xc2, 0x0f, - 0x63, 0x90, 0xf6, 0xe8, 0x0e, 0xf5, 0xc8, 0x80, 0xd2, 0xa1, 0x47, 0xd8, 0x9c, 0x26, 0xb3, 0x6c, - 0xc2, 0xf6, 0xf4, 0x96, 0xd8, 0xd8, 0x61, 0xac, 0x1c, 0x5d, 0x2c, 0x74, 0xfd, 0x00, 0xd4, 0xf0, - 0x35, 0x88, 0x99, 0xfc, 0x9d, 0x69, 0x1b, 0x44, 0x3a, 0xab, 0x87, 0x8d, 0xf8, 0x09, 0x2c, 0x04, - 0x2a, 0x46, 0xd7, 0xe0, 0xf2, 0x4e, 0xfd, 0x5e, 0xad, 0x54, 0xdc, 0x91, 0xeb, 0xfb, 0x15, 0xa9, - 0xd8, 0xac, 0x4b, 0x81, 0x19, 0xd9, 0x3c, 0xc4, 0x8b, 0x7b, 0xe5, 0xbc, 0xe0, 0x64, 0xee, 0xff, - 0x8e, 0x00, 0xab, 0xe1, 0xf7, 0x0c, 0xc9, 0x32, 0xd2, 0xe9, 0xce, 0x81, 0x1b, 0x39, 0x79, 0x4f, - 0x01, 0xbb, 0x8e, 0xd3, 0x81, 0x35, 0x7f, 0xdf, 0x97, 0xcd, 0x41, 0xaf, 0xa7, 0x18, 0xaa, 0x73, - 0xb2, 0xf8, 0xd6, 0x44, 0xb7, 0x1c, 0x1b, 0x94, 0xea, 0x4c, 0xba, 0x60, 0x85, 0x80, 0x55, 0x6c, - 0x8a, 0x3f, 0x9c, 0x83, 0x95, 0x50, 0x92, 0x17, 0xbc, 0x6a, 0xe6, 0xf4, 0x99, 0xd8, 0x34, 0x7d, - 0xe6, 0x61, 0x30, 0x48, 0x72, 0xeb, 0x4e, 0x3b, 0xec, 0x05, 0xb8, 0x44, 0x07, 0xd4, 0xc4, 0x4b, - 0x0a, 0xa8, 0x0f, 0x9c, 0xd7, 0x4d, 0xec, 0x80, 0xca, 0xf3, 0xa7, 0x93, 0x07, 0xd3, 0x9c, 0x3f, - 0x98, 0xa2, 0x26, 0xcc, 0xb3, 0xbd, 0x40, 0x7b, 0x63, 0xfd, 0xce, 0x34, 0x16, 0xdf, 0xb0, 0x2d, - 0xcf, 0xee, 0xc3, 0xd9, 0xac, 0xc2, 0xbd, 0x70, 0x3e, 0xdc, 0x0b, 0x0b, 0xbf, 0x2a, 0x40, 0xd6, - 0xc7, 0xc7, 0xdd, 0x50, 0x14, 0x3c, 0x1b, 0x8a, 0xe8, 0x11, 0xcc, 0x3a, 0x27, 0xe3, 0x73, 0x91, - 0x13, 0xaf, 0x70, 0x39, 0x03, 0xea, 0xa5, 0xd5, 0x94, 0xf4, 0x36, 0x96, 0x28, 0x4b, 0xb4, 0x06, - 0xf3, 0x6d, 0xb6, 0x13, 0xcb, 0xb6, 0x14, 0x25, 0xfb, 0x53, 0xfc, 0x0a, 0xd6, 0xa2, 0x68, 0xc9, - 0xaa, 0xaa, 0x29, 0x15, 0xf7, 0x1a, 0x77, 0xeb, 0xd2, 0x2e, 0x4d, 0xfe, 0xc8, 0x52, 0xa5, 0x71, - 0xb0, 0xd3, 0x94, 0x4b, 0xf5, 0x72, 0x48, 0x7a, 0xa9, 0x71, 0x50, 0x2a, 0x55, 0x1a, 0x0d, 0x96, - 0x8c, 0xac, 0x48, 0x52, 0x5d, 0xca, 0xc7, 0x44, 0x1d, 0x92, 0x8d, 0xd6, 0x31, 0x6e, 0x0f, 0xba, - 0x18, 0x3d, 0x82, 0x82, 0x81, 0x5b, 0x03, 0xc3, 0xa0, 0xe7, 0x53, 0xfa, 0xd8, 0x50, 0xf5, 0xb6, - 0x6c, 0x3f, 0x24, 0xc8, 0x3b, 0xc7, 0xc5, 0xa1, 0xcd, 0xc9, 0x32, 0x47, 0xa8, 0xce, 0x48, 0x6b, - 0x2e, 0xf9, 0x3e, 0xa5, 0xb6, 0xcb, 0xc8, 0x2c, 0x97, 0x5d, 0xab, 0x14, 0xff, 0x5e, 0x0c, 0x16, - 0x82, 0xd7, 0x08, 0xcf, 0x79, 0x61, 0xee, 0x1a, 0xa4, 0xdb, 0xee, 0x15, 0x32, 0xae, 0x39, 0x2f, - 0x28, 0xf8, 0x5a, 0xd2, 0xec, 0x54, 0xaf, 0x25, 0x7d, 0x0c, 0xe9, 0x41, 0xdf, 0xdd, 0x94, 0x4c, - 0x8c, 0x27, 0x66, 0xe8, 0x94, 0x78, 0xf8, 0xce, 0xf4, 0xdc, 0xb9, 0xef, 0x4c, 0x8b, 0xff, 0x34, - 0x06, 0xa8, 0x3c, 0x74, 0xdd, 0xfc, 0x4f, 0xa2, 0xda, 0x42, 0x5f, 0xa9, 0x98, 0x7b, 0xc1, 0x57, - 0x2a, 0xc4, 0xa7, 0x90, 0xa8, 0x18, 0x86, 0x6e, 0xa0, 0x5b, 0x6e, 0x3f, 0x63, 0xee, 0x8c, 0x6c, - 0xa6, 0x46, 0xbf, 0xb5, 0xd1, 0xa0, 0x0f, 0x57, 0x3a, 0x7d, 0x0f, 0xdd, 0x61, 0x5b, 0xeb, 0x54, - 0x4a, 0x7b, 0xda, 0x38, 0xb2, 0x21, 0x2e, 0xb6, 0xf8, 0xa3, 0x04, 0xc0, 0x7d, 0xfd, 0xb0, 0x69, - 0xa8, 0x9d, 0x0e, 0x36, 0xbe, 0x3d, 0x53, 0xdd, 0x87, 0xb4, 0xed, 0x67, 0x4f, 0xf4, 0x43, 0x6e, - 0xaa, 0x49, 0xaf, 0xe0, 0x92, 0xc9, 0x81, 0xea, 0xc0, 0xc8, 0x5c, 0xce, 0x62, 0xf2, 0xda, 0x43, - 0x46, 0xd4, 0x5c, 0xce, 0x6d, 0xd9, 0x06, 0xff, 0x2b, 0x39, 0xa4, 0xe8, 0x3d, 0x98, 0xc3, 0x44, - 0xdb, 0xf6, 0x81, 0x94, 0xa8, 0xc9, 0x1c, 0x35, 0x89, 0xc4, 0x71, 0x83, 0x3e, 0x37, 0xff, 0x22, - 0x3e, 0x97, 0x9c, 0xca, 0xe7, 0x3e, 0x85, 0x6c, 0x57, 0x31, 0x2d, 0xd9, 0x18, 0x68, 0x8c, 0x3c, - 0x35, 0x96, 0x3c, 0x4d, 0x08, 0xa4, 0x81, 0x46, 0xe9, 0x7f, 0x1a, 0xe6, 0xd8, 0x93, 0xa7, 0x6b, - 0x40, 0x47, 0x86, 0x9b, 0xe3, 0x95, 0xc6, 0x3d, 0x8d, 0xd3, 0x15, 0x1a, 0x30, 0x6f, 0x3b, 0xca, - 0x77, 0x21, 0x69, 0xf2, 0x78, 0x3c, 0x66, 0x3a, 0x62, 0x87, 0xed, 0xea, 0x8c, 0xe4, 0x90, 0x6c, - 0xa7, 0x60, 0x9e, 0xdb, 0x41, 0xac, 0xc2, 0x1c, 0xab, 0x06, 0xad, 0x02, 0x6a, 0x34, 0x8b, 0xcd, - 0x83, 0xc6, 0xf0, 0x98, 0x50, 0xad, 0x14, 0x77, 0x9a, 0xd5, 0x47, 0x79, 0x01, 0x01, 0xcc, 0xed, - 0x17, 0x0f, 0x1a, 0x95, 0x32, 0x7b, 0x56, 0xb1, 0x54, 0xdc, 0x2b, 0x55, 0x76, 0x76, 0x2a, 0xe5, - 0x7c, 0x7c, 0x3b, 0x01, 0xf1, 0x27, 0xfa, 0xa1, 0xf8, 0x7b, 0x71, 0x98, 0x63, 0x57, 0x58, 0xd1, - 0x03, 0xc8, 0x9a, 0xca, 0x33, 0x2c, 0x7b, 0x5e, 0xf5, 0x1b, 0x95, 0xa6, 0x62, 0x54, 0x1b, 0x0d, - 0xe5, 0x19, 0xb6, 0xdf, 0x9e, 0xac, 0xce, 0x48, 0x19, 0xd3, 0xf3, 0x8d, 0xaa, 0x30, 0xdf, 0x1f, - 0x1c, 0xca, 0xe6, 0xe0, 0x70, 0xcc, 0x03, 0x17, 0x9c, 0xd9, 0xfe, 0xe0, 0xb0, 0xab, 0x9a, 0xc7, - 0x4d, 0x7d, 0x7f, 0x70, 0xd8, 0x18, 0x1c, 0x56, 0x67, 0xa4, 0xb9, 0x3e, 0xfd, 0x85, 0x4e, 0xe0, - 0x42, 0x9f, 0x15, 0xf2, 0xe9, 0xe4, 0x99, 0x6c, 0xe9, 0x72, 0xcb, 0x6c, 0xb5, 0xc6, 0x2c, 0x9c, - 0xfc, 0x9c, 0xf9, 0xd0, 0xdd, 0xd4, 0x4b, 0x66, 0xab, 0x55, 0x9d, 0x91, 0x96, 0xfb, 0x21, 0xf0, - 0x82, 0x0c, 0x19, 0x6f, 0xb3, 0x50, 0xdd, 0x79, 0xc4, 0xc1, 0x77, 0x98, 0x60, 0x7d, 0xf2, 0x47, - 0x1c, 0xec, 0x77, 0x1b, 0xf8, 0x25, 0xcf, 0x37, 0x60, 0x21, 0xd0, 0x54, 0x32, 0x3d, 0xb1, 0xf4, - 0x3e, 0xbf, 0x7c, 0x94, 0x92, 0xd8, 0x47, 0x61, 0x15, 0x96, 0xc3, 0x24, 0x27, 0x43, 0x2f, 0xbb, - 0x70, 0x2c, 0xfe, 0xa6, 0x00, 0x97, 0x4b, 0xb4, 0xd3, 0x04, 0x2f, 0xf8, 0x8f, 0x79, 0x00, 0xe4, - 0x01, 0xe4, 0x87, 0x5e, 0x10, 0x88, 0x4d, 0xf5, 0x82, 0xc0, 0x42, 0xe0, 0x79, 0x0e, 0x74, 0x15, - 0xd2, 0xce, 0x33, 0x1e, 0x6a, 0x9b, 0x47, 0x36, 0xb0, 0x41, 0xb5, 0xb6, 0xf8, 0xfb, 0x02, 0x5c, - 0x3e, 0xa0, 0x9d, 0x34, 0x42, 0xd8, 0xb0, 0x90, 0xfa, 0x2d, 0x08, 0xea, 0x86, 0x16, 0xba, 0xcf, - 0x15, 0x8f, 0x88, 0x0d, 0x74, 0x12, 0xbd, 0xab, 0x98, 0x27, 0x76, 0x68, 0x21, 0xbf, 0xc5, 0x4d, - 0xb8, 0x78, 0x0f, 0x5b, 0x93, 0x37, 0x40, 0x7c, 0x0a, 0x97, 0xd8, 0xfd, 0x7a, 0x1f, 0x85, 0x39, - 0xce, 0x40, 0x57, 0x00, 0xfa, 0x4a, 0x07, 0xcb, 0x96, 0x7e, 0xc2, 0x5f, 0xa9, 0x4a, 0x49, 0x29, - 0x02, 0x69, 0x12, 0x00, 0xba, 0x04, 0xf4, 0xc3, 0xdd, 0x6e, 0x4a, 0x48, 0x49, 0x02, 0xa0, 0x9b, - 0x4d, 0xbf, 0x26, 0xc0, 0xe5, 0xf0, 0x3a, 0xf9, 0xd5, 0xfe, 0x06, 0x2c, 0x06, 0x95, 0x6a, 0x2f, - 0xb0, 0x27, 0xd5, 0x6a, 0x3e, 0xa0, 0x55, 0x13, 0xdd, 0x80, 0x05, 0x0d, 0x9f, 0x5a, 0xf2, 0x90, - 0xd8, 0x59, 0x02, 0xde, 0xb7, 0x45, 0x17, 0xb7, 0xe0, 0x72, 0x19, 0x77, 0xf1, 0x34, 0x5e, 0x40, - 0x77, 0x22, 0x98, 0x9f, 0xbb, 0x21, 0x77, 0x9c, 0x06, 0xb7, 0x21, 0xfd, 0x44, 0x3f, 0x94, 0x79, - 0xf0, 0xe4, 0x4e, 0xf3, 0xea, 0xd8, 0x48, 0x2e, 0xc1, 0x13, 0x77, 0x90, 0xbf, 0x02, 0xc0, 0xe9, - 0x5d, 0x97, 0x4e, 0x71, 0x48, 0xad, 0x4d, 0x56, 0xcd, 0x17, 0x98, 0x47, 0x0f, 0x8b, 0x15, 0xe6, - 0xcc, 0x2f, 0x43, 0xa4, 0x17, 0xf2, 0xde, 0x75, 0x58, 0xbe, 0x87, 0xad, 0x89, 0x84, 0x15, 0x7f, - 0x22, 0xc0, 0x12, 0xd3, 0x79, 0xb9, 0xdb, 0xbf, 0xaf, 0x1f, 0x8e, 0xd3, 0x77, 0x60, 0xde, 0x12, - 0x7b, 0x91, 0x79, 0x4b, 0x0d, 0x92, 0x86, 0x6a, 0x9e, 0x50, 0x46, 0xf1, 0xd1, 0x07, 0xc9, 0xc2, - 0x9e, 0x95, 0xa8, 0xce, 0x48, 0xf3, 0x84, 0x9e, 0xb0, 0x5a, 0x81, 0x39, 0xa2, 0x73, 0xb5, 0xcd, - 0x9f, 0x15, 0x4a, 0x3c, 0xd1, 0x0f, 0x6b, 0x6d, 0x7b, 0x04, 0xfc, 0x25, 0x01, 0x56, 0x49, 0x57, - 0x71, 0x55, 0xf2, 0x6d, 0xf6, 0x4c, 0x74, 0x11, 0x92, 0x74, 0xdf, 0x5c, 0x3e, 0x3c, 0xe3, 0xe2, - 0xcc, 0xd3, 0xef, 0xed, 0x33, 0x22, 0xc9, 0x85, 0x21, 0x49, 0x78, 0x7f, 0x2d, 0x43, 0xc6, 0xe3, - 0x37, 0x76, 0x57, 0x9d, 0xc0, 0x71, 0xd2, 0xae, 0xe3, 0x4c, 0xde, 0x41, 0x6f, 0xc3, 0x05, 0xd6, - 0x41, 0x27, 0xf3, 0x93, 0xbf, 0x1a, 0x83, 0x7c, 0xd0, 0x9c, 0x64, 0xb1, 0xc4, 0x9f, 0xb3, 0xf7, - 0x8f, 0x9a, 0x51, 0x8b, 0x25, 0xff, 0x78, 0x99, 0x35, 0x7d, 0x8f, 0x30, 0xbd, 0xd4, 0xd7, 0xaa, - 0x22, 0x5f, 0x9d, 0x8a, 0x47, 0xbf, 0x1d, 0xe6, 0x79, 0x0f, 0x64, 0x76, 0xaa, 0xf7, 0x40, 0xfe, - 0x45, 0x02, 0xe6, 0x58, 0xef, 0x09, 0x8d, 0x07, 0xef, 0xf3, 0x97, 0xb9, 0x47, 0x3f, 0x4e, 0xcf, - 0x18, 0x78, 0x9e, 0xe2, 0xfe, 0x84, 0xbe, 0xe5, 0x69, 0x61, 0x9e, 0x9e, 0xbd, 0x31, 0x92, 0x8e, - 0xb8, 0x03, 0x99, 0x35, 0x62, 0x89, 0x11, 0xa1, 0x2f, 0x20, 0x43, 0xfb, 0x96, 0xbd, 0x6c, 0x62, - 0x0b, 0x8c, 0x77, 0xcf, 0x71, 0xc6, 0xbc, 0x3a, 0x23, 0xa5, 0x0d, 0xcf, 0xad, 0xea, 0xc7, 0x60, - 0x8f, 0xb5, 0x0e, 0xf3, 0xd1, 0x67, 0xfe, 0xa2, 0x5e, 0x43, 0xaa, 0xce, 0x48, 0xb6, 0xc5, 0x6d, - 0xde, 0x2f, 0xf4, 0x4a, 0xf2, 0x47, 0x00, 0xf4, 0xbd, 0xd3, 0x49, 0x17, 0x22, 0x29, 0x8a, 0x4d, - 0x49, 0xdf, 0x87, 0x24, 0xd6, 0xda, 0x93, 0x2e, 0x42, 0xe6, 0xb1, 0xd6, 0xa6, 0x64, 0x37, 0x21, - 0xef, 0xe9, 0xb1, 0xcc, 0xc1, 0x80, 0x5a, 0x3e, 0xe7, 0x76, 0x49, 0xea, 0x5b, 0xee, 0xda, 0x2a, - 0x3d, 0xf9, 0xda, 0x4a, 0x6c, 0x41, 0xd2, 0xb6, 0x2b, 0xba, 0x08, 0x2b, 0xf7, 0xeb, 0xdb, 0x32, - 0x59, 0x10, 0x84, 0xe4, 0x88, 0xf6, 0x2b, 0x7b, 0xe5, 0xda, 0xde, 0xbd, 0xbc, 0x40, 0x3e, 0xa4, - 0x83, 0xbd, 0x3d, 0xf2, 0x11, 0x43, 0x49, 0x98, 0x2d, 0xd7, 0xf7, 0x2a, 0xf9, 0x38, 0xca, 0x40, - 0x92, 0x2d, 0x0d, 0x2a, 0xe5, 0xfc, 0x2c, 0x59, 0x34, 0xdc, 0x2d, 0xd6, 0xc8, 0xef, 0x04, 0x59, - 0x7a, 0xd8, 0x49, 0xab, 0x1b, 0x90, 0xbf, 0x87, 0x2d, 0xff, 0x40, 0x10, 0x16, 0x0c, 0x7e, 0x47, - 0x00, 0x44, 0xa2, 0x18, 0xc3, 0x0c, 0x89, 0xa5, 0xb3, 0xbe, 0x58, 0xea, 0x3e, 0x4a, 0x24, 0x78, - 0x1f, 0x25, 0xf2, 0x07, 0xd1, 0x58, 0x20, 0x88, 0xfa, 0x03, 0x70, 0x3c, 0x18, 0x80, 0xed, 0x4e, - 0x95, 0x98, 0xaa, 0x53, 0x89, 0x7d, 0x58, 0xf2, 0x09, 0xce, 0x43, 0xef, 0x3b, 0x30, 0xfb, 0x44, - 0x3f, 0xb4, 0x43, 0xee, 0x95, 0x91, 0xdc, 0x24, 0x8a, 0x3a, 0x71, 0x9c, 0x7d, 0x13, 0x96, 0x4a, - 0x8a, 0xd6, 0xc2, 0xdd, 0xf1, 0x6a, 0x7d, 0x13, 0x96, 0x58, 0x48, 0x1e, 0x8f, 0xfa, 0xdb, 0x02, - 0x5c, 0xe5, 0xc3, 0xf6, 0x50, 0x7e, 0x69, 0xdc, 0xd0, 0xf6, 0x18, 0x96, 0x42, 0x9e, 0xd0, 0x1c, - 0x73, 0xfc, 0x21, 0xa4, 0x1a, 0x34, 0xfc, 0xd0, 0xe6, 0xf8, 0xe5, 0xc1, 0xbf, 0x15, 0xe0, 0x2a, - 0x9b, 0x4c, 0x45, 0x0b, 0x1e, 0x16, 0x44, 0xbf, 0x4d, 0xa1, 0x5f, 0x68, 0xb2, 0xb5, 0x05, 0x97, - 0x49, 0x9f, 0x99, 0xa6, 0x31, 0xa2, 0x05, 0xaf, 0x50, 0x2f, 0x1c, 0x22, 0xfa, 0x56, 0x17, 0x0c, - 0x7f, 0x5b, 0x80, 0xab, 0x91, 0xd5, 0xf2, 0x8e, 0xf0, 0x25, 0x2c, 0x87, 0xa8, 0xd9, 0xee, 0x18, - 0x53, 0xe8, 0x79, 0x69, 0x58, 0xcf, 0x93, 0xf7, 0x99, 0xf7, 0xe1, 0x2a, 0xef, 0x08, 0xd3, 0xa8, - 0x75, 0x7d, 0x17, 0xb2, 0xbe, 0xff, 0x52, 0x82, 0x2e, 0xc0, 0x52, 0xa9, 0xbe, 0xd7, 0xac, 0xec, - 0x05, 0xcf, 0x2a, 0xe5, 0x21, 0x63, 0x17, 0x34, 0x2b, 0x5f, 0x34, 0xf3, 0x02, 0x5a, 0x84, 0xac, - 0x0d, 0x61, 0xff, 0xc3, 0x22, 0xb6, 0xbe, 0xef, 0xbe, 0x12, 0xe7, 0x79, 0xac, 0x8d, 0x04, 0xe2, - 0xca, 0xde, 0xc1, 0x6e, 0xd8, 0x3f, 0xc0, 0x48, 0xc3, 0x7c, 0x8d, 0x02, 0x38, 0x47, 0xa9, 0xd6, - 0xf8, 0x4c, 0x2e, 0xee, 0x15, 0x77, 0x1e, 0x35, 0x6a, 0x8d, 0x7c, 0x6c, 0xfd, 0x9f, 0x08, 0x80, - 0x86, 0xb7, 0x53, 0xd1, 0x75, 0xb8, 0x2a, 0x55, 0x76, 0xe8, 0x4e, 0x40, 0xf4, 0x66, 0x5e, 0x06, - 0x92, 0x95, 0x07, 0x07, 0xc5, 0x1d, 0xb9, 0x59, 0xcf, 0x0b, 0xa4, 0x01, 0x7b, 0xf5, 0xa6, 0xec, - 0x40, 0xe8, 0xe9, 0xe4, 0x7b, 0x52, 0xa5, 0xd8, 0xac, 0x48, 0x72, 0xb3, 0x5a, 0xdc, 0x63, 0xff, - 0x55, 0x63, 0xa7, 0xd2, 0x68, 0xb0, 0xcf, 0x59, 0x54, 0x80, 0x55, 0x2f, 0x82, 0x5c, 0x97, 0x18, - 0x79, 0x23, 0x9f, 0x20, 0x8a, 0x72, 0x50, 0x3d, 0x05, 0x73, 0x64, 0xa0, 0xa8, 0x7c, 0x51, 0x6b, - 0x34, 0x1b, 0xf9, 0xf9, 0x75, 0x09, 0xc0, 0x8d, 0xa7, 0xe8, 0x32, 0xac, 0x95, 0x77, 0xf6, 0x65, - 0x32, 0x26, 0x85, 0x68, 0x62, 0x01, 0xd2, 0x5c, 0x13, 0x04, 0x23, 0x2f, 0xa0, 0x15, 0x58, 0xf4, - 0x69, 0x83, 0x82, 0x63, 0x5b, 0xff, 0x4f, 0xa4, 0x4c, 0x1b, 0xd8, 0x78, 0xa6, 0xb6, 0x30, 0xfa, - 0x1b, 0x02, 0xe4, 0xfc, 0x2f, 0x7f, 0xa2, 0x5b, 0x63, 0xa7, 0x7f, 0x9e, 0xa7, 0x50, 0x0b, 0xb7, - 0x27, 0xc4, 0x66, 0xee, 0x2e, 0x6e, 0xfd, 0x99, 0x1f, 0xff, 0x97, 0xbf, 0x14, 0xbb, 0x25, 0xbe, - 0xb1, 0xf9, 0x6c, 0x6b, 0xf3, 0x67, 0x59, 0x1f, 0xfb, 0x6e, 0xdf, 0xd0, 0x9f, 0xe0, 0x96, 0x65, - 0x6e, 0xae, 0xff, 0xfc, 0x26, 0xff, 0x1f, 0x03, 0x77, 0xf8, 0x3c, 0xe5, 0x8e, 0xb0, 0x8e, 0x7e, - 0x20, 0x40, 0xda, 0xf3, 0x56, 0x34, 0x7a, 0x73, 0xe2, 0xf7, 0xbe, 0x0b, 0xeb, 0x93, 0xa0, 0x72, - 0xd1, 0x36, 0xa9, 0x68, 0x6f, 0x8a, 0xaf, 0x45, 0x89, 0x46, 0xdf, 0xa3, 0xbe, 0xc3, 0xae, 0x9c, - 0x10, 0xb9, 0x7e, 0x4b, 0x80, 0xc5, 0xa1, 0xf7, 0x8d, 0xd1, 0xe6, 0x24, 0xd9, 0x77, 0xaf, 0x06, - 0xdf, 0x9e, 0x9c, 0x80, 0x4b, 0xfa, 0x3e, 0x95, 0x74, 0x53, 0x5c, 0x1f, 0xa7, 0x44, 0x37, 0x24, - 0xd8, 0xf2, 0x4a, 0x13, 0xcb, 0x2b, 0x4d, 0x2b, 0xaf, 0xf4, 0xe2, 0xf2, 0x1a, 0x3e, 0x79, 0x7f, - 0x41, 0x80, 0xac, 0xef, 0x0d, 0x45, 0xf4, 0x56, 0xe4, 0xff, 0x0f, 0x19, 0x7e, 0xbe, 0xb1, 0x70, - 0x6b, 0x32, 0x64, 0x2e, 0xe3, 0x0a, 0x95, 0x71, 0x01, 0x65, 0x89, 0x8c, 0xee, 0x11, 0xac, 0xff, - 0x20, 0xc0, 0x4a, 0x68, 0x26, 0x10, 0xbd, 0x1b, 0x79, 0x7a, 0x2d, 0x3a, 0x6f, 0x58, 0x98, 0x30, - 0x0d, 0x24, 0x76, 0xa8, 0x34, 0x8a, 0x78, 0xdb, 0xab, 0x31, 0xdd, 0xe8, 0x28, 0x9a, 0xfa, 0x35, - 0xdb, 0x62, 0xa6, 0x0e, 0x19, 0xc8, 0x15, 0xdd, 0x11, 0xd6, 0x1f, 0xdf, 0x16, 0x6f, 0x46, 0xfa, - 0xef, 0x30, 0x3a, 0x6d, 0x5f, 0x68, 0xf2, 0x30, 0xb2, 0x7d, 0xa3, 0x52, 0x8d, 0xd3, 0xb6, 0x6f, - 0x8b, 0xb5, 0x8f, 0x8c, 0x25, 0xc1, 0xd6, 0x0d, 0x49, 0xbb, 0xb9, 0xfe, 0xf3, 0xb4, 0x7d, 0x5b, - 0x37, 0x5d, 0x1a, 0xb7, 0x75, 0x11, 0xe8, 0xe8, 0x0f, 0x04, 0x40, 0xc3, 0x89, 0x45, 0x14, 0xe5, - 0xc3, 0x91, 0x39, 0xc8, 0x89, 0x5b, 0xa6, 0xd0, 0x96, 0xfd, 0x0c, 0x9a, 0xae, 0x65, 0x8f, 0xd7, - 0xd1, 0xc4, 0xcd, 0x42, 0x3f, 0x11, 0xec, 0x07, 0x4b, 0x03, 0xa9, 0xc2, 0xad, 0x91, 0x1e, 0x1f, - 0x9a, 0x28, 0x2d, 0xbc, 0x3b, 0x15, 0x0d, 0xef, 0x2c, 0xfe, 0x46, 0x4e, 0xea, 0x9e, 0x4e, 0x23, - 0x27, 0xf0, 0x4d, 0xf4, 0x2f, 0x05, 0x58, 0x09, 0xcd, 0x67, 0x46, 0x3a, 0xe6, 0xa8, 0xec, 0x67, - 0x21, 0xe2, 0x91, 0x39, 0xbb, 0x25, 0xeb, 0xd3, 0x9a, 0x6b, 0x7d, 0x72, 0x73, 0xfd, 0xa1, 0x00, - 0x6b, 0x51, 0x2b, 0x07, 0xf4, 0xc1, 0xc8, 0x28, 0x12, 0x39, 0x1b, 0x2b, 0x4c, 0x3e, 0x31, 0x14, - 0x7b, 0xb4, 0x89, 0x1d, 0xf1, 0xed, 0x91, 0xc6, 0x0a, 0x99, 0x3d, 0x92, 0xee, 0xf6, 0xb6, 0xf8, - 0x56, 0x94, 0xc9, 0xc2, 0x29, 0x68, 0x73, 0xa3, 0xd6, 0x1b, 0x91, 0xcd, 0x1d, 0xb3, 0x40, 0x39, - 0x47, 0x73, 0xb7, 0xde, 0x8e, 0xb6, 0x68, 0x88, 0xe4, 0x3c, 0xba, 0xbc, 0xbd, 0xf5, 0x56, 0xa8, - 0x5d, 0x23, 0x29, 0xd0, 0x7f, 0x14, 0x60, 0x25, 0x74, 0x39, 0x12, 0xe9, 0xa7, 0xa3, 0x16, 0x2f, - 0xd3, 0x34, 0x94, 0xc7, 0x50, 0x34, 0x75, 0x43, 0x1f, 0xdf, 0x46, 0xd3, 0xb4, 0x12, 0xfd, 0x0f, - 0x9e, 0x42, 0x0d, 0x59, 0xc6, 0xa0, 0xf7, 0x47, 0x84, 0x8f, 0xe8, 0xd5, 0x56, 0xe1, 0x83, 0x69, - 0xc9, 0x78, 0xe0, 0xf1, 0xb7, 0x79, 0x0a, 0x5f, 0x76, 0xda, 0x3c, 0x99, 0x23, 0xa3, 0x1f, 0x0b, - 0xb0, 0x16, 0xb5, 0x22, 0x8a, 0xf4, 0xe2, 0x31, 0x4b, 0xa8, 0xc8, 0x20, 0xc4, 0x5b, 0xb5, 0x7e, - 0x0e, 0x4b, 0xae, 0x4f, 0x65, 0xc9, 0x1f, 0x08, 0x90, 0x0f, 0xee, 0xf7, 0xa0, 0x8d, 0x91, 0x21, - 0x68, 0x28, 0x59, 0x5d, 0x18, 0x9f, 0x1f, 0x17, 0x37, 0x68, 0x83, 0x6e, 0x8a, 0xd7, 0xa3, 0x54, - 0xee, 0xc9, 0x9f, 0xf3, 0x19, 0x7e, 0x3e, 0xb8, 0xe1, 0x13, 0x29, 0x57, 0xc4, 0xce, 0xd0, 0x14, - 0x72, 0x6d, 0x5d, 0x0f, 0x55, 0x9a, 0x47, 0x28, 0xde, 0xb9, 0xff, 0x82, 0x00, 0x59, 0xdf, 0xc6, - 0x4e, 0xe4, 0x0c, 0x34, 0x6c, 0xfb, 0x67, 0x12, 0x89, 0xde, 0xa2, 0x12, 0xbd, 0x8e, 0x26, 0x91, - 0x08, 0xfd, 0x75, 0x01, 0x16, 0x02, 0x7b, 0x19, 0xe8, 0xf6, 0x88, 0x9e, 0x34, 0xbc, 0xfb, 0x52, - 0xd8, 0x98, 0x14, 0x9d, 0x77, 0x38, 0xbf, 0x7c, 0xa3, 0x2d, 0x89, 0x7e, 0x89, 0x3e, 0x9c, 0xe0, - 0xdf, 0xe2, 0x88, 0x34, 0x63, 0xc4, 0x5e, 0x48, 0x64, 0x27, 0xe1, 0x92, 0xac, 0x4f, 0xa4, 0xa9, - 0x3f, 0x2b, 0x40, 0xc6, 0xbb, 0xc9, 0x86, 0xd6, 0x47, 0x8f, 0xb3, 0xde, 0xf4, 0x5f, 0x61, 0x74, - 0x36, 0x52, 0x5c, 0xa7, 0x82, 0xbc, 0x26, 0x5e, 0x8d, 0x8c, 0x27, 0x2c, 0xd7, 0x49, 0x1c, 0xe8, - 0xfb, 0x02, 0xa4, 0x3d, 0xe9, 0xcf, 0xc8, 0xa5, 0xeb, 0x70, 0x6e, 0x37, 0x72, 0xe9, 0x1a, 0x92, - 0x4d, 0x15, 0xdf, 0xa0, 0x22, 0xbd, 0x8a, 0xc6, 0x89, 0x84, 0x7e, 0x0e, 0x52, 0x4e, 0xbe, 0x19, - 0xbd, 0x31, 0x62, 0x80, 0x9a, 0x46, 0x21, 0xfe, 0xda, 0x87, 0x42, 0x11, 0xab, 0x9a, 0x58, 0xe5, - 0x67, 0x21, 0xe3, 0x4d, 0xb7, 0x46, 0x1a, 0x25, 0x24, 0x27, 0x1b, 0xe9, 0x16, 0xbc, 0xf2, 0xf5, - 0xb1, 0x95, 0x53, 0x97, 0xf0, 0xe4, 0x85, 0xa3, 0x5d, 0x62, 0x38, 0x79, 0x1c, 0x59, 0xfb, 0x3b, - 0xb4, 0xf6, 0xb7, 0xc4, 0x1b, 0x63, 0x6a, 0xbf, 0xd3, 0xa2, 0x4c, 0xef, 0x08, 0xeb, 0xdb, 0xdf, - 0x17, 0xe0, 0x62, 0x4b, 0xef, 0x85, 0x57, 0xbe, 0x9d, 0x2c, 0x77, 0xfb, 0xfb, 0xa4, 0x8e, 0x7d, - 0xe1, 0xf1, 0x87, 0x1c, 0xa5, 0xa3, 0x77, 0x15, 0xad, 0xb3, 0xa1, 0x1b, 0x9d, 0xcd, 0x0e, 0xd6, - 0xa8, 0x04, 0x9b, 0xac, 0x48, 0xe9, 0xab, 0x66, 0xe0, 0xdf, 0x5d, 0x7f, 0xdc, 0xee, 0xf6, 0x7f, - 0x33, 0xb6, 0x74, 0x8f, 0x91, 0x96, 0xba, 0xfa, 0xa0, 0x4d, 0xec, 0xb4, 0xf1, 0x70, 0xeb, 0x5f, - 0xdb, 0xd0, 0x2f, 0x29, 0xf4, 0xcb, 0x72, 0xb7, 0xff, 0xe5, 0xc3, 0xad, 0xc3, 0x39, 0xca, 0xf0, - 0xdd, 0xff, 0x1f, 0x00, 0x00, 0xff, 0xff, 0x68, 0x82, 0x86, 0x5c, 0x94, 0x7b, 0x00, 0x00, +var fileDescriptor_dlp_536657b46c30097f = []byte{ + // 8339 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xdc, 0x7d, 0x6d, 0x6c, 0x1b, 0xd9, + 0x76, 0x98, 0x86, 0x14, 0x29, 0xf2, 0x50, 0xa4, 0xa8, 0xab, 0x0f, 0xcb, 0xb4, 0xbd, 0xf6, 0x8e, + 0x77, 0x6d, 0xaf, 0xd6, 0x96, 0x76, 0xb5, 0x1f, 0x6f, 0xd7, 0xbb, 0x6f, 0x13, 0x8a, 0xa4, 0x4d, + 0x7a, 0x25, 0x51, 0x1e, 0x52, 0xde, 0xb5, 0xb3, 0xd8, 0xe9, 0x90, 0xbc, 0xa2, 0xc6, 0x22, 0x39, + 0xf4, 0xcc, 0xd0, 0x96, 0x36, 0x49, 0x91, 0x14, 0x6d, 0x83, 0x00, 0x5d, 0x34, 0x40, 0xfa, 0xfa, + 0xd0, 0xbc, 0x34, 0x48, 0x9b, 0x1f, 0x05, 0x02, 0xa4, 0x1f, 0x68, 0x93, 0xa0, 0x6d, 0x9a, 0x1f, + 0x69, 0xd1, 0x14, 0x45, 0x51, 0xbc, 0x87, 0xf7, 0xab, 0x28, 0x02, 0xb4, 0xaf, 0x4d, 0x1b, 0x60, + 0xf3, 0xa7, 0x40, 0xff, 0xa4, 0x05, 0x8a, 0xe2, 0x7e, 0xcc, 0x27, 0x67, 0xf8, 0x21, 0xdb, 0x6d, + 0xf0, 0x7e, 0x89, 0xf7, 0xdc, 0x73, 0xce, 0x3d, 0xf7, 0xdc, 0x73, 0xcf, 0x3d, 0xf7, 0xcc, 0xbd, + 0x57, 0x70, 0xb9, 0xad, 0x69, 0xed, 0x0e, 0xde, 0xec, 0xeb, 0xea, 0x53, 0xa5, 0x79, 0xba, 0xd9, + 0xea, 0xf4, 0x37, 0x9f, 0x6e, 0x91, 0x3f, 0x1b, 0x7d, 0x5d, 0x33, 0x35, 0xb4, 0xc2, 0x10, 0x36, + 0x38, 0xc2, 0x06, 0xa9, 0x79, 0xba, 0x95, 0xbb, 0xc8, 0xe9, 0x94, 0xbe, 0xba, 0xa9, 0xf4, 0x7a, + 0x9a, 0xa9, 0x98, 0xaa, 0xd6, 0x33, 0x18, 0x51, 0xee, 0x6a, 0x30, 0x57, 0xc3, 0xd4, 0x74, 0xa5, + 0x8d, 0x39, 0xd2, 0x2b, 0x36, 0x92, 0x66, 0x6a, 0x8d, 0xc1, 0xe1, 0x66, 0x6b, 0xa0, 0x53, 0x2e, + 0xbc, 0xfe, 0x82, 0xbf, 0x1e, 0x77, 0xfb, 0xe6, 0x29, 0xaf, 0xbc, 0xe2, 0xaf, 0x3c, 0x54, 0x71, + 0xa7, 0x25, 0x77, 0x15, 0xe3, 0x98, 0x63, 0x5c, 0xf6, 0x63, 0x98, 0x6a, 0x17, 0x1b, 0xa6, 0xd2, + 0xe5, 0x3d, 0xcb, 0x9d, 0xe3, 0x08, 0x7a, 0xbf, 0xb9, 0x69, 0x98, 0x8a, 0x39, 0xb0, 0xa4, 0x5f, + 0xe5, 0x15, 0xe6, 0x69, 0x1f, 0x6f, 0xb6, 0x14, 0x13, 0xfb, 0x04, 0xe2, 0xf0, 0x53, 0xed, 0xf0, + 0x19, 0xc6, 0xc7, 0x41, 0x95, 0xa4, 0x29, 0xed, 0xb0, 0xa5, 0x70, 0x69, 0xc5, 0x7f, 0x13, 0x87, + 0x74, 0xa5, 0x67, 0xf4, 0x71, 0xd3, 0x2c, 0x68, 0xbd, 0x43, 0xb5, 0x8d, 0x3e, 0x01, 0x50, 0x7b, + 0x87, 0x9a, 0x4c, 0xd0, 0x8d, 0x35, 0xe1, 0x4a, 0xf4, 0x46, 0x6a, 0xeb, 0xf2, 0x46, 0xa0, 0xae, + 0x37, 0x2a, 0xbd, 0x43, 0xad, 0x7e, 0xda, 0xc7, 0x52, 0x52, 0xe5, 0xbf, 0x0c, 0x54, 0x86, 0x4c, + 0x57, 0xed, 0xc9, 0x1d, 0xf5, 0x18, 0x77, 0xd4, 0x23, 0x4d, 0x6b, 0xad, 0x45, 0xae, 0x08, 0x37, + 0x32, 0x5b, 0xaf, 0x86, 0xf0, 0xd8, 0xb1, 0x11, 0xa5, 0x74, 0x57, 0xed, 0x39, 0x45, 0x74, 0x0f, + 0xe2, 0x1d, 0xb5, 0xab, 0x9a, 0xc6, 0x5a, 0xf4, 0x8a, 0x70, 0x23, 0xb5, 0xb5, 0x15, 0x2a, 0x85, + 0x4b, 0xfe, 0x8d, 0x3b, 0x6a, 0xaf, 0xa5, 0xf6, 0xda, 0x3b, 0x94, 0x52, 0xe2, 0x1c, 0xd0, 0x55, + 0x48, 0xab, 0xbd, 0x66, 0x67, 0xd0, 0xc2, 0xf2, 0x93, 0x81, 0x66, 0xe2, 0xb5, 0xd9, 0x2b, 0xc2, + 0x8d, 0x84, 0x34, 0xcf, 0x81, 0xf7, 0x09, 0x0c, 0xdd, 0x04, 0x84, 0x4f, 0x18, 0x92, 0x4b, 0x05, + 0x31, 0x8a, 0x99, 0xe5, 0x35, 0x15, 0xbb, 0xa3, 0xf7, 0x61, 0xb1, 0x39, 0x30, 0x4c, 0xad, 0xeb, + 0x46, 0x8e, 0x53, 0x7d, 0xbd, 0x1e, 0x22, 0x69, 0x81, 0xe2, 0xdb, 0x5a, 0x5b, 0x68, 0x7a, 0xca, + 0x06, 0xda, 0x85, 0x85, 0xa6, 0xd6, 0x33, 0x71, 0xcf, 0x94, 0xb5, 0x3e, 0x35, 0xdb, 0xb5, 0xc4, + 0x95, 0xe8, 0x8d, 0xcc, 0xd6, 0x6b, 0x61, 0x0c, 0x19, 0x76, 0x95, 0x22, 0x4b, 0x99, 0xa6, 0xbb, + 0x68, 0xe4, 0xbe, 0x89, 0x40, 0xda, 0xa3, 0x0e, 0xf4, 0x36, 0xac, 0x74, 0x95, 0x13, 0xf9, 0x90, + 0x01, 0x0d, 0xb9, 0x8f, 0x75, 0x59, 0x35, 0x71, 0x77, 0x4d, 0xb8, 0x22, 0xdc, 0x88, 0x49, 0xa8, + 0xab, 0x9c, 0x70, 0x02, 0x63, 0x1f, 0xeb, 0x15, 0x13, 0x77, 0xd1, 0xb7, 0x60, 0x6d, 0x88, 0x44, + 0xc7, 0x4f, 0x06, 0xd8, 0x30, 0xe9, 0xc8, 0xc6, 0xa4, 0x15, 0x2f, 0x95, 0xc4, 0x2a, 0xd1, 0xcf, + 0x40, 0x6e, 0xb8, 0x2d, 0x4b, 0x53, 0x6b, 0x51, 0xaa, 0xa8, 0x9f, 0x9c, 0x7e, 0x48, 0x6d, 0xa3, + 0xa3, 0x45, 0x69, 0xd5, 0x27, 0x32, 0xaf, 0xcc, 0xf5, 0x89, 0x5d, 0xbb, 0x10, 0xd1, 0xc7, 0x90, + 0x74, 0x5a, 0x17, 0xa8, 0x41, 0x8d, 0x35, 0xeb, 0x84, 0x65, 0xd6, 0xe8, 0x55, 0x98, 0x77, 0x77, + 0x86, 0xf7, 0x3c, 0xe5, 0x6a, 0x5c, 0xfc, 0x46, 0x80, 0x85, 0xed, 0x53, 0x13, 0xf3, 0x31, 0xa1, + 0xca, 0x2b, 0xc2, 0xac, 0xdd, 0x5e, 0x66, 0xeb, 0xad, 0x90, 0xf6, 0x7c, 0x54, 0xb4, 0x6c, 0x50, + 0x01, 0x28, 0x35, 0x42, 0x30, 0xdb, 0x52, 0x4c, 0x85, 0x36, 0x3a, 0x2f, 0xd1, 0xdf, 0xe2, 0x5f, + 0x84, 0xa4, 0x8d, 0x86, 0x72, 0xb0, 0xba, 0xfd, 0xb0, 0x5e, 0xaa, 0xc9, 0xf5, 0x87, 0xfb, 0x25, + 0xf9, 0x60, 0xaf, 0xb6, 0x5f, 0x2a, 0x54, 0xee, 0x54, 0x4a, 0xc5, 0xec, 0x0c, 0x4a, 0x42, 0xac, + 0xb2, 0x9b, 0xbf, 0x5b, 0xca, 0xc6, 0x51, 0x06, 0x80, 0xfe, 0x94, 0xef, 0xed, 0x97, 0xee, 0x66, + 0x05, 0x94, 0x86, 0x24, 0x2b, 0x6f, 0xef, 0xee, 0x67, 0x23, 0x4e, 0x71, 0x7f, 0xef, 0x6e, 0x36, + 0xea, 0x14, 0x6b, 0x0f, 0xee, 0x66, 0x67, 0x49, 0xb1, 0x5e, 0xfa, 0xbc, 0x2e, 0x1f, 0xd4, 0xef, + 0x7c, 0x90, 0x8d, 0x89, 0xff, 0x40, 0x80, 0x94, 0xbb, 0xa7, 0xab, 0x10, 0x7b, 0xaa, 0x74, 0x06, + 0x98, 0xce, 0xd5, 0x64, 0x79, 0x46, 0x62, 0x45, 0xf4, 0x2e, 0xc4, 0x4c, 0xa5, 0xd1, 0x61, 0x13, + 0x2e, 0xb5, 0x75, 0x31, 0x44, 0x05, 0x75, 0x82, 0x43, 0xa8, 0x28, 0x32, 0x2a, 0x41, 0xb2, 0x71, + 0x6a, 0x62, 0x66, 0x9b, 0x31, 0x4a, 0x79, 0x6d, 0x32, 0xe5, 0x95, 0x67, 0xa4, 0x04, 0x21, 0x25, + 0xbf, 0xb7, 0x53, 0x90, 0x24, 0xca, 0xa2, 0x6c, 0xc4, 0x7f, 0x22, 0x40, 0x8c, 0x36, 0x83, 0x3e, + 0x80, 0xb9, 0x23, 0xac, 0xb4, 0xb0, 0x6e, 0xf9, 0xb7, 0x57, 0x42, 0x78, 0xdf, 0x21, 0xae, 0xbb, + 0xd2, 0x92, 0x2c, 0x74, 0xf4, 0x2e, 0xcc, 0xea, 0xda, 0x33, 0x32, 0xfc, 0x84, 0xec, 0xca, 0xa8, + 0xce, 0x6c, 0x48, 0xda, 0x33, 0x89, 0x62, 0xe7, 0x3e, 0x82, 0xa8, 0xa4, 0x3d, 0x43, 0xef, 0x42, + 0x9c, 0xea, 0xc4, 0x6a, 0x35, 0x4c, 0x17, 0x0f, 0x08, 0x92, 0xc4, 0x71, 0xc5, 0xaf, 0x6c, 0x07, + 0x2d, 0x61, 0x63, 0xd0, 0x31, 0xd1, 0x6d, 0x48, 0xd8, 0x66, 0x38, 0x4e, 0x7c, 0x8a, 0x26, 0xd9, + 0xf8, 0xe8, 0x16, 0x20, 0x7b, 0x3e, 0x9a, 0xfa, 0xa0, 0xd7, 0x54, 0x4c, 0xcc, 0x1c, 0x74, 0x42, + 0x5a, 0xb4, 0x6a, 0xea, 0x56, 0x85, 0xf8, 0x1f, 0x23, 0x30, 0xc7, 0x99, 0xa0, 0x65, 0x88, 0x31, + 0xcf, 0x49, 0x6c, 0x39, 0x29, 0xb1, 0x82, 0x77, 0x56, 0x45, 0xa6, 0x9d, 0x55, 0x79, 0x00, 0xd7, + 0x3a, 0x11, 0x9d, 0x74, 0x9d, 0x70, 0x11, 0xa1, 0x8f, 0x20, 0xd1, 0xd1, 0x9a, 0x74, 0x75, 0xe6, + 0x26, 0x16, 0xd6, 0xfe, 0x0e, 0x47, 0x93, 0x6c, 0x02, 0xf4, 0x11, 0xa4, 0x9a, 0x3a, 0x56, 0x4c, + 0x2c, 0x93, 0x75, 0x71, 0x2d, 0x4e, 0xe9, 0x73, 0x0e, 0x3d, 0x5b, 0x9f, 0x37, 0xea, 0xd6, 0xfa, + 0x2c, 0x01, 0x43, 0x27, 0x00, 0xf4, 0x13, 0x00, 0x54, 0x07, 0xd4, 0xa9, 0xad, 0xcd, 0x51, 0xda, + 0x30, 0x8b, 0xa0, 0xeb, 0x0b, 0x51, 0x80, 0x94, 0x7c, 0x62, 0xfd, 0x14, 0xff, 0x8b, 0x00, 0x89, + 0x1d, 0x47, 0x14, 0xa0, 0x16, 0xaf, 0x2b, 0xbd, 0xb6, 0xe5, 0x9f, 0xc2, 0x0c, 0x44, 0x22, 0x38, + 0x12, 0x9d, 0x21, 0xf4, 0x27, 0x2a, 0x91, 0x75, 0xa3, 0x85, 0xfb, 0x9a, 0xda, 0x33, 0x39, 0x87, + 0xc8, 0x04, 0x1c, 0x32, 0x36, 0x11, 0x63, 0x53, 0x83, 0x45, 0x6b, 0xf9, 0xb1, 0x54, 0x64, 0xac, + 0xcd, 0x51, 0x13, 0xbb, 0x36, 0x7a, 0x01, 0xb2, 0x75, 0x9b, 0x6d, 0x7a, 0x01, 0x86, 0xf8, 0xbb, + 0x51, 0x58, 0xf0, 0x61, 0xa1, 0xd7, 0x81, 0x2e, 0x55, 0x8a, 0xda, 0xc3, 0xba, 0xdc, 0x53, 0xba, + 0x96, 0x51, 0xa5, 0x6d, 0xe8, 0x9e, 0xd2, 0xc5, 0x68, 0x1f, 0x16, 0x74, 0xdc, 0xd4, 0xf4, 0x96, + 0x2d, 0x0e, 0xef, 0x56, 0xd8, 0xfa, 0x2a, 0x51, 0x6c, 0xab, 0x99, 0xf2, 0x8c, 0x94, 0xd1, 0x3d, + 0x10, 0xb4, 0x0b, 0x19, 0xb5, 0xab, 0xb4, 0xb1, 0xc3, 0x90, 0x85, 0x16, 0x61, 0xeb, 0x6b, 0x85, + 0x20, 0xbb, 0xf8, 0xa5, 0x55, 0x37, 0x00, 0x3d, 0x80, 0xc5, 0x96, 0xd6, 0x1c, 0x74, 0xdd, 0x1a, + 0xe3, 0xee, 0xea, 0x7a, 0x08, 0xc7, 0x22, 0xc7, 0x77, 0x31, 0xcd, 0xb6, 0x7c, 0x30, 0xf4, 0x29, + 0x2c, 0x39, 0xfa, 0xb1, 0xa3, 0xc3, 0x09, 0xec, 0x13, 0xd9, 0x64, 0x36, 0x0c, 0xbd, 0xc9, 0x46, + 0x95, 0x31, 0x7b, 0x8a, 0x75, 0x83, 0x08, 0x39, 0x47, 0xf5, 0x9d, 0xb5, 0x2b, 0x1e, 0x30, 0xf8, + 0x36, 0x38, 0xd3, 0x49, 0x7c, 0x07, 0xb2, 0x7e, 0x69, 0xd1, 0x65, 0x48, 0x1d, 0xaa, 0x1d, 0x2c, + 0x6b, 0x87, 0x87, 0x06, 0x36, 0xe9, 0xb0, 0x45, 0x25, 0x20, 0xa0, 0x2a, 0x85, 0x88, 0xff, 0x41, + 0x80, 0x8c, 0x77, 0x18, 0xc8, 0x44, 0xe1, 0xc3, 0x78, 0x8c, 0x4f, 0xb9, 0x69, 0x5f, 0x19, 0x39, + 0x82, 0x9f, 0xe2, 0x53, 0x29, 0xa9, 0x5b, 0x3f, 0xd1, 0x87, 0xc4, 0xe3, 0x91, 0x20, 0x5a, 0x6d, + 0x71, 0x03, 0x18, 0xeb, 0xb0, 0x0f, 0xd9, 0x0f, 0xf4, 0x29, 0x64, 0xe8, 0x8a, 0x32, 0xe9, 0x80, + 0x53, 0xd7, 0x6d, 0x5b, 0x73, 0xda, 0x74, 0x17, 0xc5, 0x9b, 0x90, 0xf6, 0xd4, 0xa3, 0x0b, 0x90, + 0xd4, 0xb5, 0x67, 0xb2, 0xda, 0x6b, 0xe1, 0x13, 0xae, 0x8b, 0x84, 0xae, 0x3d, 0xab, 0x90, 0xb2, + 0xb8, 0x09, 0x31, 0x36, 0xad, 0x96, 0x21, 0x66, 0x98, 0x8a, 0x6e, 0x69, 0x8b, 0x15, 0x50, 0x16, + 0xa2, 0xb8, 0xc7, 0xfa, 0x13, 0x95, 0xc8, 0x4f, 0xf1, 0x11, 0xa4, 0x3d, 0xf6, 0x86, 0x2a, 0x90, + 0x69, 0x68, 0x03, 0xea, 0x7e, 0xe5, 0x86, 0x76, 0x62, 0x2f, 0x1c, 0x62, 0xd8, 0x52, 0xc8, 0x91, + 0xb7, 0xb5, 0x13, 0x29, 0xdd, 0x70, 0x0a, 0xd8, 0x10, 0x15, 0x48, 0xb9, 0x6a, 0x49, 0xe3, 0xa6, + 0xd6, 0xe7, 0x51, 0x1f, 0xf9, 0x49, 0x62, 0x8c, 0x0e, 0x3e, 0xb4, 0x42, 0x3a, 0xfa, 0x9b, 0x08, + 0xfe, 0x4c, 0x6d, 0x99, 0x47, 0x54, 0x67, 0x31, 0x89, 0x15, 0xd0, 0x2a, 0xc4, 0x8f, 0xb0, 0xda, + 0x3e, 0x32, 0xa9, 0xbf, 0x8d, 0x49, 0xbc, 0x24, 0xfe, 0xf6, 0x2c, 0x20, 0x09, 0xb7, 0x94, 0xa6, + 0x49, 0x7b, 0x61, 0x85, 0x81, 0xab, 0x10, 0xef, 0x2b, 0x3a, 0xee, 0x99, 0x7c, 0x8e, 0xf3, 0x12, + 0x19, 0x19, 0x95, 0xad, 0x6b, 0x72, 0x93, 0x86, 0x79, 0x7c, 0x68, 0x5f, 0x9b, 0x24, 0x24, 0x94, + 0xd2, 0xaa, 0x67, 0xd3, 0xf2, 0x14, 0xce, 0xb1, 0x79, 0xad, 0x53, 0x01, 0x54, 0xad, 0xc7, 0x99, + 0x92, 0xf0, 0x9d, 0xa8, 0xec, 0x93, 0x50, 0x7b, 0xf3, 0x0b, 0xbc, 0xc1, 0x0b, 0x9c, 0x0f, 0x6f, + 0x6f, 0x45, 0x0d, 0x80, 0x1a, 0xe8, 0x0d, 0xc8, 0x5a, 0xdb, 0x0a, 0x7b, 0x4d, 0x8e, 0xd3, 0xd5, + 0x74, 0x81, 0xc3, 0xad, 0xf0, 0x10, 0x15, 0xdc, 0x21, 0xcd, 0xdc, 0x34, 0x21, 0x8d, 0x13, 0xd0, + 0xe4, 0x7e, 0x20, 0xc0, 0x72, 0x90, 0x7c, 0xe8, 0x93, 0xe9, 0xa3, 0x5b, 0x12, 0x29, 0xd9, 0x2b, + 0xf1, 0x0d, 0xe2, 0x6a, 0x09, 0x4b, 0x59, 0xe9, 0x74, 0x64, 0x13, 0x9f, 0x30, 0x4b, 0x48, 0x10, + 0x9f, 0xc7, 0x2a, 0xf2, 0x9d, 0x4e, 0x1d, 0x9f, 0x98, 0x64, 0xad, 0x71, 0x2b, 0xb9, 0xa3, 0xe9, + 0x7c, 0x4a, 0x5d, 0x0c, 0x5d, 0x22, 0x3a, 0x9a, 0x4e, 0x3c, 0xb1, 0x2d, 0x71, 0x47, 0xd3, 0xb7, + 0x13, 0x10, 0x37, 0x15, 0xbd, 0x8d, 0x4d, 0xb1, 0x00, 0x31, 0x0a, 0x22, 0x46, 0xa9, 0xe3, 0x16, + 0x95, 0x3e, 0x22, 0x91, 0x9f, 0xc4, 0x00, 0xdb, 0x3a, 0xc6, 0xcc, 0xed, 0x47, 0x24, 0x56, 0x20, + 0xa6, 0xda, 0xb0, 0x22, 0xcd, 0x88, 0x44, 0x7f, 0x8b, 0xff, 0x50, 0x80, 0x25, 0xcf, 0x58, 0x1a, + 0x7d, 0xad, 0x67, 0x60, 0xb2, 0xd2, 0xb0, 0x86, 0x71, 0x4b, 0xa6, 0x43, 0x48, 0xd9, 0xcf, 0x5b, + 0x9d, 0xc2, 0x2d, 0x8a, 0x4e, 0xd0, 0xf0, 0x89, 0xa9, 0x33, 0x3c, 0xbb, 0xf7, 0x49, 0x29, 0x6d, + 0x43, 0x69, 0xdf, 0x5d, 0x36, 0xab, 0xd3, 0x60, 0x6c, 0xdc, 0xf2, 0xe1, 0x0e, 0xdc, 0x6c, 0x9b, + 0x65, 0x45, 0xf1, 0x7f, 0x47, 0x60, 0xad, 0x88, 0xd5, 0x16, 0xee, 0x99, 0xea, 0xe1, 0x29, 0x1f, + 0xef, 0x71, 0xb3, 0xa6, 0x0e, 0x8b, 0x2d, 0x9b, 0xc6, 0x3b, 0x71, 0x42, 0x57, 0x1c, 0x77, 0x1b, + 0xc4, 0x96, 0xb3, 0x2d, 0x1f, 0x24, 0x60, 0x2e, 0x46, 0xcf, 0x3e, 0x17, 0xdf, 0x87, 0x59, 0x6a, + 0xe3, 0x2c, 0x1a, 0x13, 0x47, 0x07, 0x0e, 0xd4, 0xbe, 0x29, 0x3e, 0xda, 0x82, 0x15, 0x4b, 0x08, + 0x13, 0x77, 0xfb, 0x1d, 0x12, 0x96, 0xd1, 0xd8, 0x20, 0x46, 0x35, 0xb0, 0xc4, 0x2b, 0xeb, 0xbc, + 0x8e, 0x46, 0x08, 0x1f, 0xc0, 0x9a, 0x4b, 0x1d, 0x5e, 0xb2, 0x38, 0x25, 0x5b, 0x75, 0xea, 0xdd, + 0x94, 0xe2, 0xaf, 0x09, 0x70, 0x3e, 0x40, 0xfb, 0xdc, 0x6c, 0xac, 0x3e, 0x08, 0x53, 0xf6, 0xa1, + 0x02, 0x09, 0xed, 0x29, 0xd6, 0x9f, 0xaa, 0xf8, 0x19, 0x1f, 0x95, 0x5b, 0x61, 0x0b, 0x8d, 0xae, + 0xf4, 0x8c, 0x43, 0x4d, 0xef, 0x52, 0x57, 0x5f, 0xe5, 0x44, 0x92, 0x4d, 0x4e, 0xcd, 0x43, 0x3a, + 0x83, 0x79, 0xe8, 0xcf, 0x6d, 0x1e, 0xfa, 0x8f, 0x93, 0x79, 0xe8, 0x63, 0xcc, 0x43, 0x0f, 0x37, + 0x0f, 0xe9, 0xcf, 0xb3, 0x79, 0xfc, 0x89, 0x00, 0x2b, 0x8e, 0x9e, 0x27, 0xb1, 0x8d, 0x17, 0xba, + 0xe0, 0x5a, 0x1a, 0x88, 0xbe, 0xa8, 0x51, 0x9c, 0x0d, 0x1d, 0x45, 0xf1, 0x01, 0xac, 0xfa, 0x7b, + 0xca, 0xc7, 0xe1, 0x63, 0x88, 0x73, 0x3f, 0x2c, 0x4c, 0xe1, 0x87, 0x39, 0x8d, 0xf8, 0xef, 0x22, + 0xb0, 0x54, 0x1d, 0x98, 0xfd, 0x81, 0x59, 0x63, 0xe9, 0x5f, 0xde, 0xb7, 0x8f, 0xad, 0x94, 0xc5, + 0x68, 0xa6, 0xdb, 0x6a, 0xfb, 0xfe, 0x00, 0xeb, 0xa7, 0xbe, 0xd4, 0xc5, 0x17, 0x90, 0xd6, 0x28, + 0x53, 0xd9, 0x68, 0x1e, 0xe1, 0xae, 0xc2, 0xb7, 0xb5, 0xdf, 0x0a, 0xe1, 0x12, 0x20, 0x80, 0x05, + 0xa3, 0xe4, 0xd2, 0xbc, 0xe6, 0x2a, 0x89, 0xbf, 0x24, 0xc0, 0xbc, 0xbb, 0x1a, 0x5d, 0x82, 0xf3, + 0xd5, 0x83, 0xfa, 0xfe, 0x41, 0x5d, 0xae, 0x15, 0xca, 0xa5, 0xdd, 0xbc, 0x2f, 0xfb, 0xb3, 0x08, + 0xe9, 0xed, 0x7c, 0xad, 0x52, 0x90, 0x0b, 0xd5, 0x9d, 0x83, 0xdd, 0xbd, 0x5a, 0x56, 0x40, 0x0b, + 0x90, 0xba, 0x5b, 0xa8, 0xd9, 0x80, 0x08, 0x5a, 0x81, 0xc5, 0x62, 0xbe, 0x9e, 0xaf, 0xd5, 0xab, + 0x52, 0xc9, 0x06, 0x47, 0x09, 0x78, 0xbb, 0x72, 0x57, 0xbe, 0x7f, 0x50, 0x92, 0x1e, 0xda, 0xe0, + 0x59, 0x42, 0x9e, 0xdf, 0xd9, 0xb1, 0x01, 0xb1, 0xed, 0x38, 0xcb, 0x71, 0x89, 0x4d, 0x27, 0xe3, + 0x56, 0x33, 0x15, 0xd3, 0x78, 0xce, 0x8c, 0xdb, 0x32, 0xc4, 0x9a, 0xda, 0xa0, 0x67, 0xf2, 0x08, + 0x99, 0x15, 0xc4, 0xef, 0xcf, 0xc2, 0x1a, 0x1f, 0xcd, 0xa2, 0x62, 0x2a, 0x35, 0x6d, 0xa0, 0x37, + 0x71, 0x11, 0x9b, 0x8a, 0xda, 0x31, 0x50, 0x97, 0x78, 0x3f, 0x3a, 0x09, 0x70, 0xcb, 0x4e, 0xa0, + 0x32, 0x23, 0x1f, 0x93, 0x68, 0x1c, 0xe2, 0xb5, 0x21, 0x59, 0x8c, 0x78, 0x32, 0x95, 0xb8, 0x45, + 0x2f, 0x04, 0xed, 0xd9, 0xd6, 0xc7, 0x66, 0xc1, 0xfb, 0xd3, 0xb7, 0xe1, 0xb6, 0xc7, 0xdc, 0xef, + 0x0b, 0x90, 0xf5, 0x37, 0x8b, 0x1a, 0x70, 0xde, 0xe8, 0x29, 0x7d, 0xe3, 0x48, 0x33, 0x65, 0xff, + 0xcc, 0xe1, 0x4a, 0xbd, 0x36, 0xba, 0x5d, 0x6b, 0x2e, 0x49, 0xe7, 0x2c, 0x46, 0xbe, 0x0a, 0x74, + 0x07, 0xe0, 0xb1, 0xd6, 0xf0, 0xfa, 0xf6, 0xeb, 0xa3, 0x99, 0xde, 0xd3, 0x1a, 0xdc, 0x31, 0x24, + 0x1f, 0x5b, 0x3f, 0x73, 0xff, 0x58, 0x80, 0x38, 0x4f, 0x52, 0x5d, 0x87, 0x85, 0xbe, 0xae, 0x35, + 0xb1, 0x61, 0xe0, 0x96, 0x4c, 0xc2, 0x57, 0x83, 0xef, 0x7e, 0x32, 0x36, 0x98, 0xa6, 0x2f, 0x89, + 0x43, 0x30, 0x35, 0x53, 0xe9, 0xc8, 0xd8, 0x30, 0xd5, 0xae, 0x62, 0xda, 0xe8, 0x6c, 0xd8, 0x97, + 0x68, 0x65, 0xc9, 0xaa, 0x63, 0x34, 0x3b, 0xb0, 0x60, 0x1b, 0x96, 0x6c, 0x10, 0x5b, 0xe3, 0xe9, + 0xe4, 0xd7, 0xc6, 0x98, 0x17, 0xb5, 0x4b, 0xe2, 0xca, 0x5c, 0x45, 0xf1, 0x7b, 0x02, 0x2c, 0x59, + 0x08, 0x45, 0x6c, 0x34, 0x75, 0x95, 0xaa, 0x9e, 0x84, 0x99, 0xae, 0xd4, 0x04, 0xfd, 0x8d, 0x5e, + 0x85, 0xf9, 0x96, 0x6a, 0xf4, 0x3b, 0xca, 0x29, 0xf3, 0x5a, 0x2c, 0x4a, 0x4c, 0x71, 0x18, 0x5d, + 0x73, 0x76, 0x61, 0xde, 0x18, 0xf4, 0xfb, 0x9a, 0xce, 0xba, 0x42, 0x25, 0xcb, 0x6c, 0xad, 0x8f, + 0x93, 0xcc, 0x22, 0xd9, 0x3e, 0x95, 0x52, 0x86, 0x53, 0x10, 0x6b, 0xb0, 0xbc, 0xa3, 0x1a, 0xa6, + 0xfd, 0x8d, 0xc0, 0xf2, 0xf2, 0x57, 0x21, 0xdd, 0x51, 0x7a, 0xed, 0x01, 0xd9, 0xf4, 0x34, 0xb5, + 0x96, 0x25, 0xe6, 0xbc, 0x05, 0x2c, 0x68, 0x2d, 0x4c, 0x96, 0x82, 0x43, 0xb5, 0x63, 0x62, 0x9d, + 0x0b, 0xca, 0x4b, 0x62, 0x03, 0x56, 0x7c, 0x4c, 0xb9, 0x43, 0xad, 0x04, 0x7c, 0xfc, 0x19, 0x27, + 0xba, 0x4b, 0x67, 0xae, 0xef, 0x40, 0xe2, 0x7f, 0x13, 0x60, 0x45, 0x52, 0x8d, 0xe3, 0x7c, 0x4f, + 0xe9, 0x9c, 0x1a, 0xaa, 0x61, 0x5b, 0x0c, 0x59, 0x88, 0x38, 0x2b, 0xb9, 0x8b, 0x4d, 0x5d, 0x6d, + 0x8e, 0x71, 0xb4, 0xfb, 0xac, 0xb8, 0x4b, 0x71, 0xa5, 0x74, 0xdf, 0x5d, 0x44, 0x77, 0x61, 0xde, + 0xa0, 0x93, 0x4a, 0x66, 0x3e, 0x3b, 0x32, 0xb9, 0xcf, 0x96, 0x52, 0x8c, 0x92, 0x25, 0x85, 0xbf, + 0x05, 0x73, 0x6c, 0x7f, 0x62, 0x19, 0xd3, 0xa5, 0x10, 0x1e, 0x79, 0x8a, 0x25, 0x59, 0xd8, 0xe2, + 0x1f, 0x09, 0x30, 0x77, 0x7f, 0xa0, 0x18, 0x6a, 0xa5, 0x85, 0xde, 0x85, 0x18, 0xcd, 0x3c, 0xf0, + 0x1e, 0x8d, 0x4b, 0x53, 0x30, 0x64, 0xef, 0xe6, 0x2d, 0x32, 0xfd, 0xe6, 0xed, 0x32, 0x00, 0xff, + 0x12, 0x65, 0x2a, 0x6d, 0x3b, 0x01, 0x9f, 0x64, 0xb0, 0xba, 0xd2, 0x46, 0xef, 0x02, 0x41, 0xc6, + 0x3a, 0xd9, 0x5e, 0xb1, 0xb8, 0x6b, 0x75, 0x28, 0x89, 0x54, 0xea, 0xf6, 0xcd, 0x53, 0xce, 0x96, + 0x62, 0x6e, 0xc7, 0x20, 0x6a, 0x2a, 0x6d, 0xf1, 0x8f, 0x23, 0x90, 0x25, 0x33, 0x45, 0x35, 0x4c, + 0xb5, 0xa9, 0x74, 0x98, 0xb6, 0x6e, 0x5b, 0x6b, 0x64, 0x74, 0x0a, 0x7d, 0xf3, 0x15, 0xf2, 0x21, + 0x24, 0x9f, 0x10, 0x7d, 0xc9, 0x6a, 0xcb, 0xb2, 0xb1, 0x8f, 0x43, 0xe8, 0xfd, 0xed, 0x6e, 0x70, + 0x45, 0xd3, 0xb0, 0x4c, 0xc5, 0x3a, 0x55, 0xa4, 0x94, 0x78, 0xc2, 0xa0, 0x06, 0xda, 0x05, 0xa4, + 0xe3, 0x8e, 0x62, 0xaa, 0x4f, 0xb1, 0x7c, 0x48, 0x1d, 0x76, 0xaf, 0x79, 0x3a, 0x61, 0xce, 0x68, + 0xd1, 0xa2, 0xbc, 0x63, 0x11, 0xe6, 0x8e, 0x61, 0x39, 0xa8, 0xc1, 0x33, 0x0e, 0xf3, 0x25, 0xcf, + 0x30, 0xb1, 0x19, 0xe9, 0x0c, 0x92, 0xf8, 0x27, 0x8b, 0x90, 0xf6, 0x98, 0x3a, 0x7a, 0x02, 0xab, + 0xbd, 0x41, 0x17, 0xeb, 0xa4, 0xfb, 0xcc, 0xcf, 0x59, 0x3e, 0x9a, 0xb5, 0xfb, 0xe1, 0x24, 0x13, + 0x66, 0x63, 0xcf, 0x62, 0x41, 0xdd, 0x1d, 0x9b, 0x83, 0xe5, 0x19, 0x69, 0xb9, 0x17, 0x00, 0x47, + 0xcf, 0x60, 0xad, 0xa9, 0x98, 0xb8, 0xad, 0x05, 0x34, 0xca, 0xd4, 0xf8, 0xd1, 0x44, 0x8d, 0x16, + 0x1c, 0x26, 0xde, 0x66, 0x57, 0x9b, 0x81, 0x35, 0x08, 0x03, 0x3a, 0x96, 0x95, 0x9e, 0xd6, 0x3b, + 0xed, 0xaa, 0xe6, 0xa9, 0x77, 0x2d, 0x7a, 0x6f, 0xa2, 0x26, 0x3f, 0xcd, 0x5b, 0xd4, 0x76, 0x63, + 0xd9, 0x63, 0x1f, 0x8c, 0x34, 0xd3, 0x91, 0x5b, 0x2a, 0x4d, 0x83, 0x3a, 0xcd, 0xcc, 0x4e, 0xd1, + 0xcc, 0x4e, 0xd1, 0xa2, 0x76, 0x9a, 0xe9, 0xf8, 0x60, 0x48, 0x87, 0x73, 0xc7, 0x72, 0x57, 0xe9, + 0x5b, 0xab, 0x9a, 0x93, 0x90, 0xe2, 0xe9, 0xe1, 0xc9, 0x86, 0xee, 0xd3, 0x5d, 0xa5, 0x5f, 0xb2, + 0x39, 0x38, 0x43, 0x77, 0x1c, 0x00, 0x47, 0x5f, 0x0b, 0x70, 0xa5, 0x85, 0x3b, 0xa6, 0x22, 0xf7, + 0x75, 0x6c, 0xe0, 0x5e, 0x13, 0x07, 0xb4, 0xce, 0x52, 0xc8, 0xdb, 0x13, 0xb5, 0x5e, 0x24, 0xcc, + 0xf6, 0x39, 0xaf, 0x00, 0x31, 0x2e, 0xb5, 0x46, 0x21, 0xe4, 0x76, 0x60, 0x39, 0xc8, 0xf4, 0xce, + 0x36, 0x79, 0x72, 0x7b, 0xb0, 0x1a, 0x6c, 0x53, 0x67, 0xe4, 0xf7, 0xb5, 0x00, 0x59, 0xbf, 0xc5, + 0xa0, 0x8f, 0x86, 0x3d, 0xd3, 0x38, 0x76, 0x8e, 0xef, 0xf9, 0x18, 0x92, 0xc4, 0x4b, 0x98, 0xa7, + 0x4e, 0x9a, 0x3a, 0xcc, 0x8b, 0x97, 0x28, 0x1e, 0xa1, 0xc6, 0xfc, 0x57, 0xee, 0xd7, 0x05, 0xc8, + 0xfa, 0x4d, 0xeb, 0xf9, 0xe4, 0xa9, 0xc2, 0x92, 0x81, 0x7b, 0x86, 0x4a, 0x9d, 0xa1, 0x62, 0x9a, + 0xba, 0xda, 0x18, 0x98, 0x78, 0x42, 0x67, 0x88, 0x6c, 0xd2, 0xbc, 0x45, 0x99, 0xfb, 0x26, 0x0e, + 0xcb, 0x41, 0x16, 0x89, 0x1a, 0xc3, 0x62, 0x96, 0xce, 0x6c, 0xdf, 0x1b, 0x75, 0xa5, 0xdd, 0xc6, + 0x2d, 0xbf, 0x67, 0xbf, 0x0c, 0x29, 0x1d, 0xb7, 0x99, 0x25, 0xb7, 0xac, 0xc0, 0x0b, 0x18, 0x88, + 0xc6, 0x3a, 0x06, 0x64, 0x95, 0xc1, 0x89, 0xda, 0x51, 0x15, 0xfd, 0x94, 0xc5, 0x02, 0xd6, 0x42, + 0x5e, 0x3e, 0xbb, 0x2c, 0x79, 0x8b, 0x23, 0x5b, 0xc0, 0x16, 0x14, 0x4f, 0xd9, 0xc8, 0xfd, 0x67, + 0x01, 0x52, 0x2e, 0x79, 0x7f, 0x1c, 0xd7, 0xff, 0xdc, 0xff, 0x8c, 0x40, 0xc6, 0xab, 0x87, 0xe7, + 0x5a, 0xfd, 0xf5, 0x61, 0x63, 0x39, 0x78, 0x51, 0x03, 0x64, 0x85, 0x07, 0x2f, 0x39, 0x2c, 0x68, + 0xc2, 0xbc, 0xbb, 0xa1, 0x97, 0x12, 0x0e, 0xe4, 0x7e, 0x20, 0xc0, 0xa5, 0x91, 0x1e, 0x78, 0x1a, + 0xef, 0xc0, 0xc5, 0x9d, 0x66, 0x3e, 0x49, 0xa1, 0xf3, 0xe9, 0xfa, 0x84, 0xc1, 0xda, 0xd0, 0x74, + 0xb1, 0x53, 0x05, 0xbf, 0x7c, 0x0d, 0x2e, 0xd2, 0x7d, 0xc1, 0x57, 0xd8, 0xd9, 0x15, 0x93, 0xcd, + 0x82, 0xb5, 0x93, 0xff, 0x12, 0xd6, 0x9c, 0x9d, 0xfc, 0x73, 0x6c, 0x16, 0x56, 0x6d, 0x2e, 0xde, + 0xc8, 0xea, 0x11, 0x38, 0x35, 0xf2, 0x99, 0xf7, 0x0f, 0xcb, 0x36, 0x8f, 0x9a, 0x6b, 0x23, 0xf1, + 0x73, 0xc2, 0x70, 0xd8, 0xe6, 0xc9, 0x13, 0x84, 0xf9, 0xa3, 0x51, 0x1a, 0xf1, 0x45, 0x71, 0x6c, + 0x97, 0x3d, 0x1c, 0xc5, 0xf1, 0xdd, 0xf7, 0x5f, 0x15, 0x82, 0xc2, 0x38, 0x2e, 0x04, 0x73, 0x00, + 0xf7, 0xce, 0x22, 0x84, 0x7f, 0x05, 0xb6, 0xc5, 0x18, 0x8a, 0xea, 0xb8, 0x20, 0xa6, 0x37, 0xaa, + 0xe3, 0x12, 0xb0, 0x10, 0xa8, 0x78, 0x16, 0x09, 0x9c, 0x25, 0xdb, 0x6e, 0xdb, 0x15, 0xe4, 0x39, + 0xad, 0xba, 0x83, 0x3c, 0xde, 0x6a, 0xfc, 0xec, 0xad, 0x3a, 0x0b, 0xb3, 0xd3, 0x6a, 0xc7, 0x07, + 0x43, 0x3f, 0x2f, 0x04, 0x04, 0x7d, 0xbc, 0xed, 0xb9, 0xb3, 0x0f, 0xbc, 0xd7, 0xed, 0x39, 0x03, + 0x7f, 0x1c, 0x00, 0x47, 0xbf, 0x3a, 0x32, 0x06, 0xe4, 0xc2, 0x24, 0xa9, 0x30, 0xf7, 0xcf, 0x22, + 0x4c, 0x88, 0x43, 0xb2, 0xa5, 0x0a, 0x0b, 0x09, 0x19, 0x42, 0xee, 0x87, 0x82, 0x3f, 0x26, 0xe4, + 0x72, 0x7f, 0x08, 0xc9, 0xae, 0xda, 0x93, 0xd9, 0x09, 0xb2, 0xd1, 0x87, 0x5f, 0xd8, 0xe9, 0xa8, + 0x44, 0x57, 0xed, 0xd1, 0x5f, 0x94, 0x54, 0x39, 0xe1, 0xa4, 0x91, 0x89, 0x48, 0x95, 0x13, 0x46, + 0x5a, 0x82, 0x85, 0x27, 0x03, 0xa5, 0x67, 0xaa, 0x1d, 0x2c, 0xf3, 0x93, 0x59, 0xb3, 0x13, 0x9c, + 0xcc, 0xca, 0x58, 0x44, 0xb4, 0x68, 0xe4, 0xbe, 0x9e, 0x1d, 0x8e, 0x4d, 0x79, 0xbf, 0x7e, 0x5b, + 0x80, 0x57, 0x29, 0x67, 0x67, 0xd9, 0x91, 0x8f, 0x54, 0xc3, 0xd4, 0xda, 0xba, 0xd2, 0x95, 0x1b, + 0x83, 0xe6, 0x31, 0x36, 0xad, 0x4f, 0xd4, 0x8f, 0x5f, 0xdc, 0x8c, 0x1c, 0x02, 0x97, 0xad, 0x36, + 0xb7, 0x69, 0x93, 0xd2, 0x2b, 0x54, 0x28, 0x7b, 0x45, 0xf3, 0x55, 0x1b, 0xb9, 0x7f, 0x16, 0x81, + 0xcb, 0x63, 0x78, 0xa0, 0x6f, 0xc3, 0x05, 0x7f, 0xd7, 0x3a, 0xda, 0x33, 0xac, 0xcb, 0xf4, 0xe4, + 0x01, 0x4f, 0xf7, 0xad, 0x79, 0x1b, 0xda, 0x21, 0x08, 0xf4, 0x20, 0x42, 0x10, 0xf9, 0xa0, 0xdf, + 0xb7, 0xc9, 0x23, 0x41, 0xe4, 0x07, 0x04, 0x81, 0x91, 0x5f, 0x86, 0x14, 0x53, 0x9f, 0x6c, 0xa8, + 0x5f, 0xb1, 0x40, 0x24, 0x2a, 0x01, 0x03, 0xd5, 0xd4, 0xaf, 0x30, 0xba, 0x07, 0x69, 0x8e, 0xe0, + 0x19, 0xda, 0xd7, 0x47, 0x0d, 0xad, 0xdd, 0x90, 0x34, 0xcf, 0x68, 0xd9, 0x08, 0xa3, 0x9b, 0x80, + 0xdc, 0xbc, 0x64, 0x96, 0x98, 0x8e, 0xd1, 0x36, 0xb3, 0x2e, 0xcc, 0x02, 0x81, 0xe7, 0xbe, 0x89, + 0xb9, 0xb7, 0x16, 0xdc, 0x12, 0x7e, 0x4b, 0x80, 0xab, 0xf8, 0xc9, 0x40, 0x7d, 0xaa, 0x74, 0xe8, + 0xb4, 0x6c, 0x76, 0x14, 0xc3, 0x08, 0xb5, 0x85, 0x2f, 0x5f, 0x84, 0x6f, 0x74, 0x01, 0xfc, 0xe3, + 0x7f, 0xc5, 0x25, 0x4a, 0x81, 0x48, 0x32, 0x64, 0x01, 0xbf, 0x22, 0x40, 0xce, 0xa1, 0x2f, 0xf9, + 0xd0, 0xd1, 0x1d, 0xc8, 0xda, 0xb1, 0x87, 0x3c, 0xc5, 0xa9, 0xc6, 0x8c, 0x15, 0x80, 0x70, 0xcd, + 0xbe, 0x0b, 0xab, 0xc3, 0x5a, 0xa1, 0x23, 0xca, 0x0c, 0x60, 0xd9, 0x2f, 0x28, 0x19, 0xdb, 0xdc, + 0x2f, 0x46, 0xe1, 0x7c, 0x68, 0xe7, 0xd0, 0x3d, 0x10, 0x83, 0x79, 0x06, 0xd8, 0xe7, 0x2b, 0x41, + 0xfc, 0x5d, 0x56, 0x1a, 0xce, 0x6b, 0xd8, 0x58, 0x03, 0x79, 0x4d, 0x63, 0xb2, 0x7f, 0x59, 0x08, + 0xb6, 0x59, 0xf9, 0x05, 0x5b, 0x83, 0x7f, 0x34, 0x9f, 0xcb, 0xda, 0xff, 0xee, 0x9c, 0x7b, 0xe3, + 0xca, 0xad, 0xfd, 0xf7, 0x04, 0x78, 0xd3, 0xd9, 0x7c, 0x4e, 0xea, 0x01, 0xbf, 0x7c, 0x11, 0x6b, + 0xb3, 0x0b, 0xe0, 0xb7, 0xfa, 0xeb, 0xb6, 0x48, 0x0f, 0x46, 0xbb, 0xbf, 0xdf, 0x89, 0x40, 0xce, + 0x61, 0xf3, 0xe7, 0xcb, 0xf8, 0x51, 0x1e, 0x2e, 0xf5, 0x06, 0x5d, 0xb9, 0x45, 0x42, 0xed, 0x5e, + 0xd3, 0x94, 0x7d, 0x7a, 0x36, 0xb8, 0x61, 0xe5, 0x7a, 0x83, 0x6e, 0x91, 0xe3, 0xd4, 0x3c, 0xfd, + 0x36, 0xd0, 0x67, 0xb0, 0x6c, 0x6a, 0xfd, 0x61, 0xca, 0xa9, 0x5c, 0x24, 0x32, 0xb5, 0xbe, 0x8f, + 0x71, 0xee, 0x3b, 0x51, 0x38, 0x1f, 0xaa, 0x7f, 0xb4, 0x0f, 0xaf, 0x87, 0x1b, 0xc5, 0xf0, 0xdc, + 0x7c, 0x35, 0x64, 0xb8, 0x5c, 0xd3, 0x73, 0x24, 0xc7, 0xe1, 0x19, 0x1a, 0xc6, 0xf1, 0xff, 0xdb, + 0x24, 0x1d, 0x61, 0xbc, 0x2f, 0x74, 0x92, 0xfe, 0x6a, 0xcc, 0x9f, 0xba, 0xe1, 0x13, 0xf5, 0x37, + 0x04, 0xc8, 0x0d, 0x05, 0xad, 0xf6, 0xfc, 0xe4, 0x56, 0x7d, 0xf8, 0xa2, 0xe2, 0x56, 0x1f, 0xd0, + 0x3f, 0x3f, 0xcf, 0x1d, 0x07, 0x57, 0xe7, 0xfe, 0xa6, 0x00, 0x17, 0xbc, 0xa4, 0x7c, 0x43, 0xcb, + 0x95, 0xf1, 0xa2, 0x26, 0xe4, 0x26, 0x2c, 0x39, 0x9f, 0x21, 0xed, 0x6d, 0x0b, 0x37, 0x1e, 0x64, + 0x57, 0xd9, 0x8e, 0x34, 0xf7, 0xaf, 0x22, 0x70, 0x69, 0x64, 0x9f, 0xd0, 0x55, 0x48, 0x93, 0xc8, + 0xd6, 0x61, 0xc6, 0x6c, 0x7b, 0xbe, 0xab, 0xf6, 0x6c, 0x36, 0x14, 0x49, 0x39, 0x19, 0x6a, 0x71, + 0xbe, 0xab, 0x9c, 0x38, 0x48, 0x3e, 0xcb, 0x8c, 0x0d, 0x59, 0xe6, 0x2f, 0x0e, 0x59, 0x26, 0xbb, + 0x8d, 0xd4, 0x7a, 0x49, 0xc3, 0xe7, 0x19, 0x83, 0x89, 0xcc, 0x73, 0x2e, 0xc4, 0x3c, 0xff, 0x2c, + 0x16, 0x9a, 0xeb, 0xe0, 0x76, 0xfa, 0x87, 0x02, 0x5c, 0x0d, 0xdf, 0xd8, 0xf8, 0x0d, 0xf6, 0xe4, + 0x85, 0xef, 0x6d, 0xc2, 0x6a, 0x87, 0x02, 0xab, 0xd6, 0x18, 0xbc, 0xdc, 0x6f, 0x08, 0xf0, 0x5a, + 0x08, 0xb3, 0x97, 0x63, 0xd4, 0xef, 0xc0, 0x8a, 0x63, 0xd4, 0x7d, 0x5d, 0x6b, 0x28, 0x0d, 0xb5, + 0x63, 0x19, 0x99, 0x20, 0x2d, 0xdb, 0x95, 0xfb, 0x4e, 0x5d, 0xee, 0xbf, 0x47, 0xe0, 0xda, 0x64, + 0x5d, 0x46, 0xd7, 0x61, 0x81, 0x58, 0xb8, 0x9b, 0xb3, 0x40, 0x39, 0x67, 0xba, 0x6a, 0xcf, 0xc5, + 0x93, 0x22, 0x2a, 0x27, 0x01, 0x22, 0x64, 0xba, 0xca, 0x89, 0x1b, 0x71, 0xac, 0xa5, 0x7f, 0x2f, + 0xc4, 0xd2, 0x9f, 0xfe, 0x3f, 0x1b, 0xf7, 0x17, 0x66, 0xfb, 0xdb, 0x09, 0xeb, 0x14, 0x89, 0xf8, + 0x08, 0x32, 0xde, 0x65, 0x09, 0x6d, 0x59, 0x97, 0xaa, 0x26, 0xd9, 0x12, 0xf3, 0x0b, 0x57, 0xc1, + 0xe7, 0x66, 0x7e, 0x2d, 0x0a, 0x31, 0xb6, 0xe9, 0x7d, 0x1d, 0xd2, 0x6a, 0xcf, 0xc4, 0x6d, 0xac, + 0xbb, 0xb6, 0xdb, 0xd1, 0xf2, 0x8c, 0x34, 0xcf, 0xc1, 0x0c, 0xed, 0x55, 0x48, 0x1d, 0x76, 0x34, + 0xc5, 0x74, 0x6d, 0xac, 0x85, 0xf2, 0x8c, 0x04, 0x14, 0xc8, 0x50, 0xae, 0xc2, 0xbc, 0x61, 0xea, + 0x6a, 0xaf, 0x2d, 0x7b, 0x6f, 0x7e, 0xa5, 0x18, 0xd4, 0x6e, 0xae, 0xa1, 0x69, 0x1d, 0xac, 0x58, + 0xbb, 0xfb, 0x59, 0x7e, 0xac, 0x78, 0x9e, 0x83, 0xed, 0xad, 0xb8, 0x7d, 0xcf, 0x81, 0x23, 0xc6, + 0xc6, 0xdd, 0x76, 0x28, 0xcf, 0x48, 0x19, 0x9b, 0x88, 0xb1, 0xf9, 0x16, 0x00, 0x81, 0x70, 0x0e, + 0x71, 0x6f, 0xaa, 0xdb, 0x3c, 0xed, 0x63, 0x4a, 0x5d, 0x3d, 0x2c, 0x2a, 0xa7, 0xe5, 0x19, 0x29, + 0x49, 0x70, 0x19, 0xe1, 0x16, 0x40, 0x4b, 0x31, 0x2d, 0x42, 0x96, 0xae, 0x59, 0xf4, 0x10, 0x16, + 0x15, 0x13, 0x13, 0x1a, 0x82, 0xc6, 0x68, 0x0a, 0xb0, 0xd8, 0x52, 0x4e, 0x65, 0xed, 0x50, 0x7e, + 0x86, 0xf1, 0x31, 0x27, 0x4d, 0xd0, 0xd3, 0x5e, 0xab, 0x3e, 0xd2, 0xd3, 0xea, 0xe1, 0x67, 0x18, + 0x1f, 0x13, 0x89, 0x5b, 0x56, 0x81, 0x32, 0xb1, 0x53, 0xa2, 0x3f, 0x05, 0x49, 0xfb, 0x92, 0x10, + 0xfa, 0x84, 0xde, 0x5b, 0xe3, 0xb7, 0x92, 0x46, 0x7f, 0x10, 0x28, 0xf2, 0xeb, 0x48, 0xe5, 0x19, + 0x29, 0xd1, 0xe2, 0xbf, 0xb7, 0x33, 0x30, 0xdf, 0x57, 0x74, 0x03, 0xb7, 0xd8, 0x65, 0x57, 0xf1, + 0xaf, 0x47, 0x20, 0x61, 0x21, 0xa2, 0xd7, 0xe9, 0x6d, 0x42, 0xcb, 0xa6, 0x86, 0x3b, 0x49, 0x2f, + 0x18, 0x62, 0xf4, 0x3e, 0xa4, 0x5c, 0xbd, 0xe3, 0x97, 0x78, 0x43, 0xfa, 0x45, 0xb4, 0xc2, 0x7f, + 0xa2, 0x75, 0x98, 0xa5, 0x62, 0x47, 0x47, 0x29, 0x5f, 0xa2, 0x38, 0xa8, 0x04, 0x74, 0x08, 0xe4, + 0xaf, 0xb4, 0x9e, 0x75, 0x41, 0xf0, 0xc6, 0x98, 0x7e, 0x52, 0x1e, 0x8f, 0xb4, 0x1e, 0x96, 0x12, + 0x26, 0xff, 0x95, 0x7b, 0x1b, 0x12, 0x16, 0x14, 0xbd, 0x0e, 0x19, 0x76, 0x37, 0x45, 0xee, 0xaa, + 0xbd, 0x81, 0x75, 0xee, 0x28, 0x26, 0xa5, 0x19, 0x74, 0x97, 0x01, 0xc5, 0x3f, 0x13, 0x20, 0xeb, + 0x3f, 0xf9, 0x8a, 0x3a, 0x70, 0xde, 0x39, 0x57, 0x64, 0x7a, 0x4e, 0x60, 0x1a, 0x5c, 0x5d, 0x1b, + 0x63, 0xbe, 0xcb, 0x78, 0xcf, 0x6d, 0x1a, 0xe5, 0x19, 0xe9, 0x9c, 0x1a, 0x5c, 0x85, 0x30, 0xac, + 0xf2, 0x6b, 0x31, 0xfe, 0xa6, 0xd8, 0x88, 0xdf, 0x1c, 0x79, 0x45, 0x66, 0xb8, 0xa1, 0x15, 0x3d, + 0xa8, 0x62, 0x3b, 0x0b, 0x19, 0x2f, 0x7f, 0xf1, 0x47, 0x73, 0x70, 0x6e, 0x5f, 0x57, 0xbb, 0x34, + 0x10, 0xf6, 0xa2, 0x23, 0x09, 0x32, 0x3a, 0xee, 0x77, 0x14, 0xb2, 0x1d, 0x71, 0x1f, 0x35, 0x78, + 0x23, 0x54, 0x18, 0x8a, 0xcc, 0xfd, 0x19, 0xff, 0x30, 0x9c, 0xe6, 0x2c, 0xb8, 0x5a, 0xef, 0x01, + 0x3f, 0x6d, 0xef, 0x3d, 0x48, 0x70, 0x75, 0xe4, 0x95, 0x0c, 0x9b, 0xd9, 0xbc, 0xee, 0x2a, 0xa3, + 0xbf, 0x00, 0x2b, 0xcd, 0x23, 0x85, 0x1e, 0xc9, 0xd7, 0xe9, 0x9d, 0x7a, 0xef, 0x49, 0x81, 0xb0, + 0xb3, 0x4a, 0x05, 0x8b, 0x66, 0x57, 0x31, 0x8e, 0x6d, 0xd6, 0x4b, 0xcd, 0x61, 0x30, 0x32, 0xe1, + 0x52, 0x53, 0x3f, 0xed, 0x9b, 0x9a, 0x6c, 0x29, 0xe2, 0xf0, 0xf0, 0x44, 0x3e, 0xec, 0x63, 0xef, + 0x61, 0x81, 0xb0, 0xbb, 0xbc, 0x05, 0x4a, 0xcb, 0xd5, 0x72, 0xe7, 0xf0, 0xe4, 0x4e, 0xdf, 0xd1, + 0xcb, 0xf9, 0x66, 0x58, 0x25, 0xea, 0xc3, 0x85, 0x43, 0xf5, 0x04, 0xb7, 0x58, 0x6e, 0x81, 0x2d, + 0x12, 0xc4, 0xb3, 0x7a, 0x0e, 0x0d, 0x6c, 0x86, 0x7e, 0x58, 0x3a, 0xc1, 0x2d, 0xb2, 0x30, 0x6e, + 0x5b, 0x74, 0x76, 0x93, 0x6b, 0x87, 0x21, 0x75, 0xa8, 0x06, 0xd9, 0xa1, 0x66, 0xe2, 0xa3, 0xaf, + 0xa5, 0x0c, 0x71, 0x5f, 0x68, 0xf8, 0x98, 0x9a, 0x70, 0xc9, 0xd2, 0xda, 0x33, 0xd5, 0x3c, 0x72, + 0xee, 0x7b, 0x5b, 0x2d, 0xcc, 0x8d, 0x54, 0x1e, 0xd7, 0xcc, 0x67, 0xaa, 0x79, 0x64, 0x4d, 0x28, + 0x47, 0x79, 0x7a, 0x58, 0x25, 0xba, 0x0f, 0x59, 0xea, 0x46, 0xfa, 0x8a, 0x6e, 0xdb, 0x58, 0x62, + 0xe4, 0x45, 0x41, 0xe2, 0x2e, 0xf6, 0x15, 0xdd, 0xb1, 0x32, 0xba, 0x90, 0x38, 0x10, 0xf4, 0x19, + 0x20, 0x6e, 0x05, 0x47, 0x8a, 0x71, 0x64, 0x31, 0x4d, 0x8e, 0x3c, 0x1a, 0xc9, 0x86, 0xbe, 0xac, + 0x18, 0x47, 0xce, 0xc9, 0x90, 0xa6, 0x0f, 0x46, 0x2f, 0x70, 0x10, 0xd7, 0x6e, 0x1c, 0xa9, 0x87, + 0xb6, 0xb0, 0xa9, 0x91, 0x7a, 0x27, 0xae, 0xaf, 0x46, 0xd0, 0x1d, 0xbd, 0xb7, 0xbc, 0xa0, 0x80, + 0x49, 0xfe, 0x8d, 0x00, 0x19, 0x6f, 0x2f, 0xd1, 0x03, 0x58, 0xa0, 0x1a, 0x32, 0x35, 0x99, 0x5f, + 0x6b, 0xe1, 0xf7, 0xd2, 0x37, 0x26, 0xd2, 0x92, 0x5d, 0x94, 0xd2, 0x84, 0x4d, 0x5d, 0x2b, 0x31, + 0x26, 0xe2, 0xcf, 0x09, 0xcc, 0xff, 0x92, 0x3a, 0x74, 0x1e, 0x56, 0xea, 0x95, 0xdd, 0x92, 0xbc, + 0x9f, 0x97, 0xea, 0xbe, 0xb3, 0xc8, 0x09, 0x98, 0x7d, 0x58, 0xca, 0x4b, 0x59, 0x01, 0x25, 0x21, + 0xb6, 0x5b, 0xdd, 0xab, 0x97, 0xb3, 0x11, 0x94, 0x85, 0xf9, 0x62, 0xfe, 0xa1, 0x5c, 0xbd, 0x23, + 0x33, 0x48, 0x14, 0x2d, 0x40, 0x8a, 0x43, 0x3e, 0x2b, 0x95, 0x3e, 0xcd, 0xce, 0x12, 0x14, 0xf2, + 0x8b, 0x40, 0x28, 0x7d, 0x8c, 0xa0, 0x94, 0xab, 0x07, 0x12, 0x81, 0x14, 0xf3, 0x0f, 0xb3, 0x71, + 0xb1, 0x06, 0x59, 0xbf, 0xf6, 0xd1, 0x4f, 0x00, 0xf0, 0x21, 0x1c, 0x7f, 0xed, 0x90, 0x11, 0xd3, + 0x6b, 0x87, 0x4d, 0xeb, 0xa7, 0x58, 0x05, 0x34, 0xec, 0xde, 0xd0, 0x87, 0x90, 0xec, 0xe1, 0x67, + 0xd3, 0x7c, 0xaa, 0xe8, 0xe1, 0x67, 0xf4, 0x97, 0x78, 0x01, 0xce, 0x87, 0x5a, 0xb8, 0x98, 0x81, + 0x79, 0xb7, 0xe7, 0x13, 0x7f, 0x14, 0x81, 0x34, 0x71, 0x5b, 0x46, 0x5d, 0xab, 0xb4, 0x7b, 0x9a, + 0x8e, 0xd1, 0x06, 0x20, 0xdb, 0x61, 0x19, 0x64, 0x14, 0x8d, 0x63, 0x95, 0xdd, 0xe1, 0x4b, 0x52, + 0x53, 0xb3, 0xeb, 0xea, 0x5a, 0xed, 0x58, 0xed, 0xa3, 0x53, 0xb8, 0xd0, 0xd4, 0xba, 0x5d, 0xad, + 0x27, 0x7b, 0xc9, 0x54, 0xca, 0x8e, 0xaf, 0xe8, 0x1f, 0x8c, 0xf0, 0x98, 0x76, 0xd3, 0x1b, 0x05, + 0xca, 0xc7, 0x03, 0x23, 0xce, 0xa5, 0x69, 0x83, 0xad, 0x86, 0x59, 0x9d, 0xf8, 0x5d, 0x01, 0x96, + 0x02, 0x68, 0xd0, 0x35, 0x10, 0x0b, 0xd5, 0xdd, 0xdd, 0xea, 0x9e, 0x5c, 0x28, 0xe7, 0xa5, 0x9a, + 0x5c, 0xaf, 0xca, 0x95, 0xbb, 0x7b, 0x55, 0xc9, 0xff, 0x68, 0x41, 0x0a, 0xe6, 0xf6, 0x0e, 0x76, + 0x4b, 0x52, 0xa5, 0x90, 0x15, 0xd0, 0x32, 0x64, 0xf3, 0x3b, 0xfb, 0xe5, 0xbc, 0x7c, 0xb0, 0xbf, + 0x5f, 0x92, 0xe4, 0x42, 0xbe, 0x56, 0xca, 0x46, 0x1c, 0xe8, 0x4e, 0xf5, 0x33, 0x0b, 0x4a, 0x8d, + 0x67, 0xff, 0x60, 0xaf, 0x50, 0x3f, 0xc8, 0xd7, 0x2b, 0xd5, 0xbd, 0xec, 0x2c, 0xca, 0x00, 0x7c, + 0x56, 0xae, 0xd4, 0x4b, 0xb5, 0xfd, 0x7c, 0xa1, 0x94, 0x8d, 0x6d, 0xcf, 0x03, 0x38, 0xda, 0x10, + 0xff, 0x98, 0xc8, 0x19, 0xb0, 0x08, 0xbc, 0x09, 0x8b, 0x64, 0x71, 0xa1, 0xae, 0xd1, 0xaa, 0xe6, + 0x27, 0x6c, 0xb3, 0xbc, 0xc2, 0x26, 0x43, 0xaf, 0x41, 0xa6, 0x37, 0xe8, 0x36, 0xb0, 0x4e, 0x94, + 0x4b, 0x6a, 0xf9, 0x25, 0xca, 0x79, 0x06, 0xad, 0x6b, 0x84, 0x31, 0xd9, 0xe3, 0xeb, 0xf8, 0x29, + 0xd6, 0x0d, 0x2c, 0x6b, 0x7a, 0x0b, 0xb3, 0x5b, 0x73, 0x09, 0xb2, 0xbc, 0x51, 0x60, 0x95, 0xc0, + 0xd0, 0x03, 0x58, 0x0e, 0x1c, 0xab, 0xd9, 0x91, 0xc7, 0x9b, 0x3d, 0x3a, 0x96, 0x50, 0x73, 0x78, + 0x3c, 0xfe, 0xa9, 0x00, 0x6b, 0x61, 0xab, 0x04, 0xfa, 0x36, 0xa4, 0xfc, 0xc9, 0xb7, 0x71, 0x36, + 0x0d, 0x1d, 0xf7, 0x87, 0x9c, 0x94, 0x3f, 0xd3, 0x36, 0x96, 0x7c, 0x30, 0x32, 0xe1, 0x26, 0xb8, + 0x37, 0x7b, 0xe2, 0x2f, 0x45, 0x60, 0xc1, 0x2f, 0xf2, 0x5d, 0x98, 0xb3, 0x52, 0xc7, 0x6c, 0x4b, + 0x7c, 0x6b, 0xb2, 0x35, 0x8b, 0x97, 0x25, 0x8b, 0x9a, 0x1e, 0x59, 0xe7, 0xfb, 0xd8, 0x0d, 0x88, + 0x76, 0xd5, 0xde, 0x44, 0xdd, 0x27, 0x88, 0x14, 0x5f, 0x39, 0x99, 0xa8, 0xbf, 0x04, 0x11, 0x55, + 0x60, 0x91, 0x2f, 0x61, 0xf4, 0xbe, 0xb8, 0xb3, 0x67, 0x1a, 0x47, 0x9d, 0x75, 0x91, 0x31, 0x47, + 0xf2, 0x07, 0xb3, 0x70, 0x3e, 0x34, 0xd0, 0x78, 0x6e, 0xc7, 0x87, 0x3e, 0x80, 0x39, 0x7a, 0x8d, + 0x9f, 0x5f, 0x83, 0x9c, 0xe0, 0xba, 0x35, 0x47, 0x47, 0x06, 0x2c, 0x70, 0x97, 0xa3, 0x74, 0xfa, + 0x47, 0x4a, 0x03, 0xb3, 0xe3, 0x06, 0x99, 0xd0, 0x4f, 0xdf, 0xa1, 0xbd, 0xd8, 0xb8, 0x73, 0x78, + 0xc2, 0x3c, 0xc8, 0x1e, 0x3d, 0x7d, 0x93, 0xe7, 0xfc, 0xc8, 0x5a, 0xcd, 0x9a, 0xb0, 0x20, 0xe8, + 0x0d, 0xe0, 0x0f, 0xe9, 0x38, 0x8d, 0xc6, 0xb8, 0x53, 0xcc, 0xb0, 0x0a, 0x1b, 0x75, 0x15, 0x62, + 0xba, 0xd2, 0x52, 0x4f, 0x68, 0xa4, 0x13, 0x2b, 0xcf, 0x48, 0xac, 0x48, 0xcf, 0xca, 0x0d, 0x74, + 0x5d, 0x6b, 0x2b, 0xa6, 0xeb, 0xed, 0x1f, 0x1e, 0x44, 0x8c, 0xbd, 0xb4, 0xb2, 0x68, 0xd3, 0x5a, + 0x20, 0xf1, 0x6f, 0x08, 0x70, 0x2e, 0xa4, 0x07, 0x68, 0x1d, 0xae, 0xdd, 0xb9, 0xf3, 0xb9, 0xcc, + 0x1d, 0xe1, 0x5e, 0xbe, 0x5e, 0x79, 0x50, 0x92, 0xa9, 0x2f, 0xdb, 0x2e, 0xd5, 0x47, 0x39, 0x42, + 0xb2, 0xec, 0x95, 0x3e, 0xcf, 0x17, 0x4b, 0x85, 0xca, 0x6e, 0x7e, 0x27, 0x1b, 0x41, 0x17, 0x61, + 0xcd, 0xf1, 0x89, 0x8c, 0x85, 0x6c, 0xa1, 0x47, 0xd1, 0x22, 0xa4, 0xbd, 0xa0, 0xd9, 0x6d, 0x80, + 0x84, 0xa5, 0x23, 0xf1, 0x7f, 0x09, 0x90, 0xb4, 0x87, 0x1f, 0x55, 0x20, 0x49, 0x23, 0x08, 0xd5, + 0xba, 0x38, 0x16, 0x1e, 0xf3, 0xd7, 0x2d, 0x3c, 0x9b, 0x9a, 0xee, 0xa5, 0x2d, 0x28, 0x61, 0x35, + 0xe8, 0x3d, 0xd3, 0x95, 0x7e, 0x1f, 0x5b, 0xee, 0x20, 0x8c, 0xd5, 0x81, 0x85, 0xe7, 0x61, 0x65, + 0x53, 0xa3, 0x5d, 0x48, 0x1d, 0x77, 0x0d, 0xd9, 0x62, 0x36, 0x3a, 0xc8, 0xff, 0xb4, 0x6b, 0x7c, + 0x36, 0xcc, 0x0d, 0x8e, 0x6d, 0xf0, 0x76, 0x02, 0xe2, 0xec, 0xb0, 0x8f, 0x78, 0x03, 0xd0, 0x70, + 0x37, 0x82, 0xae, 0x7c, 0x88, 0xd7, 0x00, 0x0d, 0x4b, 0x89, 0xb2, 0x10, 0xb5, 0x26, 0xd7, 0xbc, + 0x44, 0x7e, 0x8a, 0x5f, 0xc2, 0x52, 0x80, 0x00, 0xc4, 0xbd, 0x71, 0x62, 0xd9, 0x21, 0x00, 0x0e, + 0x22, 0x08, 0xd7, 0x60, 0xc1, 0x99, 0xad, 0xee, 0x5b, 0x25, 0x69, 0x7b, 0x42, 0xd2, 0x5b, 0x70, + 0x7f, 0x2a, 0xc0, 0x82, 0x2f, 0x12, 0x44, 0x37, 0x20, 0xeb, 0x72, 0xbd, 0x72, 0x4b, 0x39, 0xb5, + 0xb6, 0xbb, 0x19, 0xc7, 0xc3, 0x16, 0x95, 0x53, 0x83, 0x60, 0xba, 0x7c, 0x3c, 0xc3, 0x64, 0xab, + 0x54, 0xc6, 0x71, 0xe5, 0x14, 0xd3, 0x35, 0xf9, 0xa3, 0xd3, 0x4d, 0xfe, 0xbc, 0xc7, 0xef, 0xcc, + 0x4e, 0xe6, 0x77, 0xe8, 0x41, 0x45, 0xab, 0x40, 0x06, 0xa8, 0x8b, 0xcd, 0x23, 0xad, 0x25, 0x7e, + 0x3f, 0x02, 0xe7, 0x42, 0x36, 0xd5, 0x48, 0x83, 0x85, 0xe1, 0xdd, 0xf9, 0xa8, 0x53, 0xa7, 0x21, + 0x8c, 0x42, 0xe0, 0x92, 0x9f, 0x7b, 0xee, 0x5f, 0x0a, 0xb0, 0x1a, 0x8c, 0xfb, 0xdc, 0xcf, 0xa5, + 0xa9, 0xb0, 0xd6, 0xb7, 0xf6, 0xe2, 0xbe, 0x44, 0x00, 0x9f, 0x3b, 0x1b, 0xe1, 0x27, 0xdd, 0x82, + 0xb6, 0xf0, 0xd2, 0xb9, 0x7e, 0x70, 0x85, 0xf8, 0x73, 0x51, 0x58, 0xa2, 0x83, 0xe6, 0xeb, 0xc2, + 0xfb, 0x10, 0xa7, 0x27, 0x12, 0x27, 0x3d, 0x62, 0xcc, 0xb1, 0x51, 0x11, 0x92, 0x4d, 0xad, 0xd7, + 0x52, 0x5d, 0xcf, 0x6a, 0x5c, 0x1b, 0x99, 0xb3, 0x28, 0x58, 0xd8, 0x92, 0x43, 0x88, 0x8e, 0x47, + 0x28, 0x60, 0xf6, 0x2c, 0x0a, 0x28, 0xcf, 0x84, 0xaa, 0x60, 0x74, 0x86, 0x27, 0xf6, 0x82, 0x33, + 0x3c, 0x01, 0xbb, 0xb2, 0x1f, 0x0a, 0xb0, 0x12, 0x98, 0xbf, 0x41, 0x32, 0xac, 0xb0, 0x37, 0x4e, + 0x82, 0x2d, 0x7b, 0x7d, 0xd4, 0x98, 0xf8, 0x0c, 0x60, 0xf9, 0x70, 0x18, 0x68, 0xa0, 0x87, 0xb0, + 0xc4, 0xd3, 0x4d, 0xc6, 0xa0, 0xdf, 0xd7, 0xb1, 0x61, 0xf0, 0x5c, 0x53, 0x74, 0x44, 0xd6, 0x8d, + 0xc9, 0x5a, 0x73, 0x08, 0x24, 0xa4, 0xfb, 0x41, 0x86, 0xf8, 0x10, 0x16, 0x87, 0x10, 0xbd, 0xd6, + 0x21, 0x9c, 0xd1, 0x3a, 0xc4, 0x5f, 0x8f, 0xc1, 0x82, 0xaf, 0x1a, 0xd5, 0x21, 0x85, 0x4f, 0x9c, + 0x1e, 0x8c, 0x7e, 0x1c, 0xd0, 0x47, 0xbc, 0x51, 0x72, 0x28, 0x25, 0x37, 0x9b, 0xdc, 0xef, 0x91, + 0xe5, 0xd0, 0x6e, 0xe3, 0x6c, 0x47, 0x7a, 0x4b, 0x90, 0xd0, 0xfa, 0x58, 0x57, 0x4c, 0xfe, 0x28, + 0x46, 0x66, 0x44, 0xde, 0xac, 0x43, 0xc7, 0x45, 0xe9, 0x54, 0x39, 0x81, 0x64, 0x93, 0x3a, 0x69, + 0xff, 0xd9, 0x89, 0xd3, 0xfe, 0xb9, 0x2f, 0x01, 0x6c, 0xe9, 0x0d, 0xb4, 0x0f, 0x60, 0xeb, 0xd0, + 0x32, 0xa1, 0xb7, 0x26, 0xd4, 0x90, 0x33, 0x0e, 0x2e, 0x1e, 0xb9, 0xef, 0x46, 0x20, 0xe5, 0xd2, + 0x1d, 0xea, 0x92, 0x05, 0xa5, 0x4d, 0x4f, 0x97, 0xda, 0x5d, 0x66, 0xd9, 0x84, 0xed, 0xe9, 0x47, + 0x62, 0x63, 0x87, 0xb1, 0xb2, 0x75, 0xb1, 0xd0, 0xf1, 0x02, 0x50, 0xcd, 0xd3, 0x21, 0x36, 0xe4, + 0x6f, 0x4f, 0xdb, 0x21, 0x32, 0x59, 0x5d, 0x6c, 0xc4, 0x8f, 0x61, 0xc1, 0xd7, 0x30, 0xba, 0x02, + 0x17, 0x77, 0xaa, 0x77, 0x2b, 0x85, 0xfc, 0x8e, 0x5c, 0xdd, 0x2f, 0x49, 0xf9, 0x7a, 0x55, 0xf2, + 0x45, 0x64, 0x73, 0x10, 0xcd, 0xef, 0x15, 0xb3, 0x82, 0x9d, 0xb9, 0xff, 0xfb, 0x02, 0xac, 0x06, + 0x5f, 0xd7, 0x27, 0xdb, 0x48, 0x7b, 0x3a, 0xfb, 0x2e, 0xb6, 0x66, 0x5d, 0x15, 0xec, 0x56, 0x6b, + 0x1b, 0xd6, 0xbc, 0x73, 0x5f, 0x36, 0x06, 0xdd, 0xae, 0xa2, 0xab, 0xf6, 0xc1, 0xeb, 0x9b, 0x13, + 0x3d, 0x16, 0x50, 0xa3, 0x54, 0xa7, 0xd2, 0x39, 0x33, 0x00, 0xac, 0x62, 0x43, 0xfc, 0x6e, 0x1c, + 0x56, 0x02, 0x49, 0x9e, 0xf3, 0xc6, 0xb6, 0x3d, 0x67, 0x22, 0xd3, 0xcc, 0x99, 0x07, 0x7e, 0x27, + 0xc9, 0x47, 0x77, 0xda, 0x65, 0xcf, 0xc7, 0x25, 0xdc, 0xa1, 0xc6, 0x5e, 0x90, 0x43, 0xbd, 0x6f, + 0xbf, 0x4e, 0x66, 0x39, 0x54, 0x9e, 0x3f, 0x9d, 0xdc, 0x99, 0x66, 0xbc, 0xce, 0x14, 0xd5, 0x61, + 0x8e, 0x7d, 0x0b, 0xb4, 0x0e, 0x95, 0xdc, 0x9e, 0x66, 0xc4, 0x37, 0xac, 0x91, 0x67, 0xd7, 0xca, + 0x2d, 0x56, 0xc1, 0x56, 0x38, 0x17, 0x6c, 0x85, 0xb9, 0x5f, 0x11, 0x20, 0xed, 0xe1, 0xe3, 0x7c, + 0x50, 0x14, 0x5c, 0x1f, 0x14, 0xd1, 0x43, 0x98, 0xb5, 0x2f, 0x0e, 0x64, 0x42, 0x03, 0xaf, 0x60, + 0x39, 0x7d, 0xea, 0xa5, 0xcd, 0x14, 0xb4, 0x16, 0x96, 0x28, 0x4b, 0xb4, 0x06, 0x73, 0x2d, 0xf6, + 0x6d, 0x96, 0x7d, 0x52, 0x94, 0xac, 0xa2, 0xf8, 0x25, 0xac, 0x85, 0xd1, 0x92, 0x5d, 0x55, 0x5d, + 0xca, 0xef, 0xd5, 0xee, 0x54, 0xa5, 0x5d, 0x9a, 0xfc, 0x91, 0xa5, 0x52, 0xed, 0x60, 0xa7, 0x2e, + 0x17, 0xaa, 0xc5, 0x80, 0xf4, 0x52, 0xed, 0xa0, 0x50, 0x28, 0xd5, 0x6a, 0x2c, 0x19, 0x59, 0x92, + 0xa4, 0xaa, 0x94, 0x8d, 0x88, 0x1a, 0x24, 0x6a, 0xcd, 0x23, 0xdc, 0x1a, 0xd0, 0x5b, 0xaa, 0x39, + 0x1d, 0x37, 0x07, 0xba, 0x4e, 0x0f, 0x1b, 0xf4, 0xb1, 0xae, 0x6a, 0x2d, 0xd9, 0x7a, 0x08, 0x98, + 0x4f, 0x8e, 0xf3, 0x43, 0x1f, 0x27, 0x8b, 0x1c, 0xa1, 0x3c, 0x23, 0xad, 0x39, 0xe4, 0xfb, 0x94, + 0xda, 0xaa, 0x23, 0x51, 0x2e, 0x7b, 0x9d, 0x40, 0xfc, 0x47, 0x11, 0x58, 0xf0, 0xdf, 0xc6, 0x3f, + 0xe3, 0xbd, 0xf3, 0x2b, 0x90, 0x6a, 0x39, 0x37, 0xb1, 0xb9, 0xe6, 0xdc, 0x20, 0xff, 0x6b, 0x87, + 0xb3, 0x53, 0xbd, 0x76, 0xf8, 0x11, 0xa4, 0x06, 0x7d, 0xe7, 0xa3, 0x64, 0x6c, 0x3c, 0x31, 0x43, + 0xa7, 0xc4, 0xc3, 0x4f, 0x8f, 0xc4, 0xcf, 0xfc, 0xf4, 0x88, 0xf8, 0x2f, 0x22, 0x80, 0x8a, 0x43, + 0xaf, 0xb6, 0xfc, 0x38, 0xaa, 0x2d, 0xf0, 0xb1, 0xa7, 0xf8, 0x73, 0x3e, 0xf6, 0x24, 0x3e, 0x81, + 0x58, 0x49, 0xd7, 0x35, 0x1d, 0xdd, 0x74, 0xe6, 0x19, 0x33, 0x67, 0x64, 0x31, 0xd5, 0xfb, 0x4d, + 0x7a, 0x9b, 0x67, 0x60, 0xd8, 0x73, 0x0f, 0xdd, 0x66, 0x9f, 0xd6, 0xa9, 0x94, 0x56, 0xd8, 0x38, + 0xb2, 0x23, 0x0e, 0xb6, 0xf8, 0xfd, 0x18, 0xc0, 0x3d, 0xad, 0x51, 0xd7, 0xd5, 0x76, 0x1b, 0xeb, + 0x2f, 0x6f, 0xa8, 0xee, 0x41, 0xca, 0xb2, 0xb3, 0xc7, 0x5a, 0x83, 0x0f, 0xd5, 0xa4, 0x2f, 0x59, + 0x90, 0xe0, 0x40, 0xb5, 0x61, 0x24, 0x96, 0x33, 0x99, 0xbc, 0xd6, 0x92, 0x11, 0x16, 0xcb, 0x39, + 0x3d, 0xdb, 0xe0, 0x7f, 0x25, 0x9b, 0x14, 0xbd, 0x0b, 0x71, 0x4c, 0xb4, 0x6d, 0x1d, 0x51, 0x09, + 0x0b, 0xe6, 0xe8, 0x90, 0x48, 0x1c, 0xd7, 0x6f, 0x73, 0x73, 0xcf, 0x63, 0x73, 0x89, 0xa9, 0x6c, + 0xee, 0x13, 0x48, 0x77, 0x14, 0xc3, 0x94, 0xf5, 0x41, 0x8f, 0x91, 0x27, 0xc7, 0x92, 0xa7, 0x08, + 0x81, 0x34, 0xe8, 0x51, 0xfa, 0x9f, 0x84, 0x38, 0x7b, 0xb2, 0x7c, 0x0d, 0xe8, 0xca, 0x70, 0x63, + 0xbc, 0xd2, 0xb8, 0xa5, 0x71, 0xba, 0x5c, 0x0d, 0xe6, 0x2c, 0x43, 0xf9, 0x36, 0x24, 0x0c, 0xee, + 0x8f, 0xc7, 0x84, 0x23, 0x96, 0xdb, 0x2e, 0xcf, 0x48, 0x36, 0xc9, 0x76, 0x12, 0xe6, 0xf8, 0x38, + 0x88, 0x65, 0x88, 0xb3, 0x66, 0xd0, 0x2a, 0xa0, 0x5a, 0x3d, 0x5f, 0x3f, 0xa8, 0x0d, 0xaf, 0x09, + 0xe5, 0x52, 0x7e, 0xa7, 0x5e, 0x7e, 0x98, 0x15, 0x10, 0x40, 0x7c, 0x3f, 0x7f, 0x50, 0x2b, 0x15, + 0xd9, 0xb3, 0xc8, 0x85, 0xfc, 0x5e, 0xa1, 0xb4, 0xb3, 0x53, 0x2a, 0x66, 0xa3, 0xdb, 0x31, 0x88, + 0x3e, 0xd6, 0x1a, 0xe2, 0xef, 0x47, 0x21, 0xce, 0x5e, 0x82, 0x40, 0xf7, 0x21, 0x6d, 0x28, 0x4f, + 0x5d, 0x2f, 0x00, 0x0a, 0x23, 0xd3, 0x54, 0x8c, 0x6a, 0xa3, 0xa6, 0x3c, 0xb5, 0x1f, 0x07, 0x2c, + 0xcf, 0x48, 0xf3, 0x86, 0xab, 0x8c, 0xca, 0x30, 0xd7, 0x1f, 0x34, 0x64, 0x63, 0xd0, 0x18, 0xf3, + 0x4e, 0x14, 0x67, 0xb6, 0x3f, 0x68, 0x74, 0x54, 0xe3, 0xa8, 0xae, 0xed, 0x0f, 0x1a, 0xb5, 0x41, + 0xa3, 0x3c, 0x23, 0xc5, 0xfb, 0xf4, 0x17, 0x3a, 0x86, 0x73, 0x7d, 0x56, 0xc9, 0xc3, 0xc9, 0x53, + 0xd9, 0xd4, 0xe4, 0xa6, 0xd1, 0x6c, 0x8e, 0xd9, 0x38, 0x79, 0x39, 0xf3, 0xa5, 0xbb, 0xae, 0x15, + 0x8c, 0x66, 0xb3, 0x3c, 0x23, 0x2d, 0xf7, 0x03, 0xe0, 0x39, 0x19, 0xe6, 0xdd, 0xdd, 0x42, 0x55, + 0xfb, 0x2d, 0x24, 0xcf, 0x61, 0x82, 0xf5, 0xc9, 0xdf, 0x42, 0xb2, 0x9e, 0x3f, 0xe2, 0x77, 0xca, + 0xaf, 0xc3, 0x82, 0xaf, 0xab, 0x24, 0x3c, 0x31, 0xb5, 0x3e, 0xbf, 0x17, 0x98, 0x94, 0x58, 0x21, + 0xb7, 0x0a, 0xcb, 0x41, 0x92, 0x93, 0xa5, 0x97, 0xbd, 0xdb, 0x21, 0xfe, 0xa6, 0x00, 0x17, 0x0b, + 0x74, 0xd2, 0xf8, 0xdf, 0xc9, 0x19, 0xf3, 0x8e, 0xd6, 0x7d, 0xc8, 0x0e, 0x3d, 0xc4, 0x13, 0x99, + 0xea, 0x21, 0x9e, 0x05, 0xdf, 0x2b, 0x57, 0xe8, 0x32, 0xa4, 0xec, 0xd7, 0xb0, 0xd4, 0x16, 0xf7, + 0x6c, 0x60, 0x81, 0x2a, 0x2d, 0xf1, 0x0f, 0x04, 0xb8, 0x78, 0x40, 0x27, 0x69, 0x88, 0xb0, 0x41, + 0x2e, 0xf5, 0x25, 0x08, 0xea, 0xb8, 0x16, 0xfa, 0x9d, 0x2b, 0x1a, 0xe2, 0x1b, 0x68, 0x10, 0xbd, + 0xab, 0x18, 0xc7, 0x96, 0x6b, 0x21, 0xbf, 0xc5, 0x4d, 0x38, 0x7f, 0x17, 0x9b, 0x93, 0x77, 0x40, + 0x7c, 0x02, 0x17, 0xd8, 0x33, 0x35, 0x1e, 0x0a, 0x63, 0xdc, 0x00, 0x5d, 0x02, 0xe8, 0x2b, 0x6d, + 0x2c, 0x9b, 0xda, 0x31, 0x7f, 0x3a, 0x32, 0x29, 0x25, 0x09, 0xa4, 0x4e, 0x00, 0xe8, 0x02, 0xd0, + 0x82, 0xf3, 0xb9, 0x29, 0x26, 0x25, 0x08, 0x80, 0x7e, 0x6c, 0xfa, 0x9e, 0x00, 0x17, 0x83, 0xdb, + 0xe4, 0x2f, 0xe4, 0xd4, 0x60, 0xd1, 0xaf, 0x54, 0x6b, 0x83, 0x3d, 0xa9, 0x56, 0xb3, 0x3e, 0xad, + 0x1a, 0xe8, 0x1a, 0x2c, 0xf4, 0xf0, 0x89, 0x29, 0x0f, 0x89, 0x9d, 0x26, 0xe0, 0x7d, 0x4b, 0x74, + 0x71, 0x0b, 0x2e, 0x16, 0x71, 0x07, 0x4f, 0x63, 0x05, 0xf4, 0x4b, 0x04, 0xb3, 0x73, 0xc7, 0xe5, + 0x8e, 0xd3, 0xe0, 0x36, 0xa4, 0x1e, 0x6b, 0x0d, 0x99, 0x3b, 0x4f, 0x6e, 0x34, 0xaf, 0x8e, 0xf5, + 0xe4, 0x12, 0x3c, 0x76, 0x16, 0xf9, 0x4b, 0x00, 0x9c, 0xde, 0x31, 0xe9, 0x24, 0x87, 0x54, 0x5a, + 0x64, 0xd7, 0x7c, 0x8e, 0x59, 0xf4, 0xb0, 0x58, 0x41, 0xc6, 0xfc, 0x22, 0x44, 0x7a, 0x2e, 0xeb, + 0x5d, 0x87, 0xe5, 0xbb, 0xd8, 0x9c, 0x48, 0x58, 0xf1, 0x47, 0x02, 0x2c, 0x31, 0x9d, 0x17, 0x3b, + 0xfd, 0x7b, 0x5a, 0x63, 0x9c, 0xbe, 0x7d, 0x71, 0x4b, 0xe4, 0x79, 0xe2, 0x96, 0x0a, 0x24, 0x74, + 0xd5, 0x38, 0xa6, 0x8c, 0xa2, 0xa3, 0x0f, 0x92, 0x05, 0xbd, 0xce, 0x54, 0x9e, 0x91, 0xe6, 0x08, + 0x3d, 0x61, 0xb5, 0x02, 0x71, 0xa2, 0x73, 0xb5, 0xc5, 0x5f, 0xe7, 0x8b, 0x3d, 0xd6, 0x1a, 0x95, + 0x96, 0xb5, 0x02, 0xfe, 0x82, 0x00, 0xab, 0x64, 0xaa, 0x38, 0x2a, 0x79, 0x99, 0x33, 0x13, 0x9d, + 0x87, 0x04, 0xfd, 0x6e, 0x2e, 0x37, 0x4e, 0xb9, 0x38, 0x73, 0xb4, 0xbc, 0x7d, 0x4a, 0x24, 0x39, + 0x37, 0x24, 0x09, 0x9f, 0xaf, 0x45, 0x98, 0x77, 0xd9, 0x8d, 0x35, 0x55, 0x27, 0x30, 0x9c, 0x94, + 0x63, 0x38, 0x93, 0x4f, 0xd0, 0x5b, 0x70, 0x8e, 0x4d, 0xd0, 0xc9, 0xec, 0xe4, 0x6f, 0x45, 0x20, + 0xeb, 0x1f, 0x4e, 0xb2, 0x59, 0xe2, 0xff, 0x8e, 0xc6, 0xbb, 0x6a, 0xbe, 0x16, 0x7a, 0xed, 0xde, + 0xbd, 0x5e, 0xa6, 0x0d, 0xcf, 0x5b, 0x86, 0x2f, 0xf4, 0xd1, 0xc7, 0xd0, 0xc7, 0x1b, 0xa3, 0xe1, + 0x4f, 0x70, 0xba, 0x9e, 0xd5, 0x9a, 0x9d, 0xea, 0x59, 0xad, 0x7f, 0x1d, 0x83, 0x38, 0x9b, 0x3d, + 0x81, 0xfe, 0xe0, 0x3d, 0xfe, 0x9f, 0x35, 0x46, 0xff, 0x73, 0x19, 0xc6, 0xc0, 0xf5, 0xaf, 0x34, + 0x3e, 0xa6, 0x6f, 0x71, 0x9b, 0x98, 0xa7, 0x67, 0xaf, 0x8d, 0xa4, 0x23, 0xe6, 0x40, 0xa2, 0x46, + 0x2c, 0x31, 0x22, 0xf4, 0x39, 0xcc, 0xd3, 0xb9, 0x65, 0x6d, 0x9b, 0xd8, 0x06, 0xe3, 0x9d, 0x33, + 0x9c, 0x3a, 0x2f, 0xcf, 0x48, 0x29, 0xdd, 0xf5, 0xe0, 0xc1, 0x23, 0xb0, 0xd6, 0x5a, 0x9b, 0xf9, + 0xe8, 0x33, 0x7f, 0x61, 0x8f, 0x0a, 0x96, 0x67, 0x24, 0x6b, 0xc4, 0x2d, 0xde, 0xcf, 0xf5, 0x5f, + 0x0e, 0x3e, 0x04, 0xa0, 0xef, 0x95, 0x4f, 0xba, 0x11, 0x49, 0x52, 0x6c, 0x4a, 0xfa, 0x1e, 0x24, + 0x70, 0xaf, 0x35, 0xe9, 0x26, 0x64, 0x0e, 0xf7, 0x5a, 0x94, 0xec, 0x06, 0x64, 0x5d, 0x33, 0x96, + 0x19, 0x18, 0xd0, 0x91, 0xcf, 0x38, 0x53, 0x92, 0xda, 0x96, 0xb3, 0xb7, 0x4a, 0x4d, 0xbe, 0xb7, + 0x12, 0x9b, 0x90, 0xb0, 0xc6, 0x15, 0x9d, 0x87, 0x95, 0x7b, 0xd5, 0x6d, 0x99, 0x6c, 0x08, 0x02, + 0x72, 0x44, 0xfb, 0xa5, 0xbd, 0x62, 0x65, 0xef, 0x6e, 0x56, 0x20, 0x05, 0xe9, 0x60, 0x6f, 0x8f, + 0x14, 0x22, 0x28, 0x01, 0xb3, 0xc5, 0xea, 0x5e, 0x29, 0x1b, 0x45, 0xf3, 0x90, 0x60, 0x5b, 0x83, + 0x52, 0x31, 0x3b, 0x4b, 0x36, 0x0d, 0x77, 0xf2, 0x15, 0xf2, 0x3b, 0x46, 0xb6, 0x1e, 0x56, 0xd2, + 0xea, 0x1a, 0x64, 0xef, 0x62, 0xd3, 0xbb, 0x10, 0x04, 0x39, 0x83, 0xdf, 0x15, 0x00, 0x11, 0x2f, + 0xc6, 0x30, 0x03, 0x7c, 0xe9, 0xac, 0xc7, 0x97, 0x3a, 0x6f, 0xfb, 0x09, 0xee, 0xb7, 0xfd, 0xbc, + 0x4e, 0x34, 0xe2, 0x73, 0xa2, 0x5e, 0x07, 0x1c, 0xf5, 0x3b, 0x60, 0x6b, 0x52, 0xc5, 0xa6, 0x9a, + 0x54, 0x62, 0x1f, 0x96, 0x3c, 0x82, 0x73, 0xd7, 0xfb, 0x36, 0xcc, 0x3e, 0xd6, 0x1a, 0x96, 0xcb, + 0xbd, 0x34, 0x92, 0x9b, 0x44, 0x51, 0x27, 0xf6, 0xb3, 0x6f, 0xc0, 0x52, 0x41, 0xe9, 0x35, 0x71, + 0x67, 0xbc, 0x5a, 0xdf, 0x80, 0x25, 0xe6, 0x92, 0xc7, 0xa3, 0xfe, 0x8e, 0x00, 0x97, 0xf9, 0xb2, + 0x3d, 0x94, 0x5f, 0x1a, 0xb7, 0xb4, 0x3d, 0x82, 0xa5, 0x80, 0x97, 0xa8, 0xc7, 0x1c, 0x7f, 0x08, + 0x68, 0x06, 0x0d, 0xbf, 0x57, 0x3d, 0x7e, 0x7b, 0xf0, 0xef, 0x05, 0xb8, 0xcc, 0x82, 0xa9, 0x70, + 0xc1, 0x83, 0x9c, 0xe8, 0xcb, 0x14, 0xfa, 0xb9, 0x82, 0xad, 0x2d, 0xb8, 0x48, 0xe6, 0xcc, 0x34, + 0x9d, 0x11, 0x4d, 0x78, 0x85, 0x5a, 0xe1, 0x10, 0xd1, 0x4b, 0xdd, 0x30, 0xfc, 0x3d, 0x01, 0x2e, + 0x87, 0x36, 0xcb, 0x27, 0xc2, 0x17, 0xb0, 0x1c, 0xa0, 0x66, 0x6b, 0x62, 0x4c, 0xa1, 0xe7, 0xa5, + 0x61, 0x3d, 0x4f, 0x3e, 0x67, 0xde, 0x83, 0xcb, 0x7c, 0x22, 0x4c, 0xa3, 0xd6, 0xf5, 0x5d, 0x48, + 0x7b, 0xfe, 0xcb, 0x18, 0x3a, 0x07, 0x4b, 0x85, 0xea, 0x5e, 0xbd, 0xb4, 0xe7, 0x3f, 0xab, 0x94, + 0x85, 0x79, 0xab, 0xa2, 0x5e, 0xfa, 0xbc, 0x9e, 0x15, 0xd0, 0x22, 0xa4, 0x2d, 0x08, 0xfb, 0x1f, + 0x54, 0x91, 0xf5, 0x7d, 0xe7, 0xb1, 0x55, 0xd7, 0x9b, 0xa7, 0xc4, 0x11, 0x97, 0xf6, 0x0e, 0x76, + 0x83, 0xfe, 0x81, 0x55, 0x0a, 0xe6, 0x2a, 0x14, 0xc0, 0x39, 0x4a, 0x95, 0xda, 0xa7, 0x72, 0x7e, + 0x2f, 0xbf, 0xf3, 0xb0, 0x56, 0xa9, 0x65, 0x23, 0xeb, 0xff, 0x5c, 0x00, 0x34, 0xfc, 0x39, 0x15, + 0x5d, 0x85, 0xcb, 0x52, 0x69, 0x87, 0x7e, 0x09, 0x08, 0xff, 0x98, 0x37, 0x0f, 0x89, 0xd2, 0xfd, + 0x83, 0xfc, 0x8e, 0x5c, 0xaf, 0x66, 0x05, 0xd2, 0x81, 0xbd, 0x6a, 0x5d, 0xb6, 0x21, 0xf4, 0x74, + 0xf2, 0x5d, 0xa9, 0x94, 0xaf, 0x97, 0x24, 0xb9, 0x5e, 0xce, 0xef, 0xb1, 0xff, 0x8a, 0xb5, 0x53, + 0xaa, 0xd5, 0x58, 0x71, 0x16, 0xe5, 0x60, 0xd5, 0x8d, 0x20, 0x57, 0x25, 0x46, 0x5e, 0xcb, 0xc6, + 0x88, 0xa2, 0x6c, 0x54, 0x57, 0x45, 0x9c, 0x2c, 0x14, 0xa5, 0xcf, 0x2b, 0xb5, 0x7a, 0x2d, 0x3b, + 0xb7, 0x2e, 0x01, 0x38, 0xfe, 0x14, 0x5d, 0x84, 0xb5, 0xe2, 0xce, 0xbe, 0x4c, 0xd6, 0xa4, 0x00, + 0x4d, 0x2c, 0x40, 0x8a, 0x6b, 0x82, 0x60, 0x64, 0x05, 0xb4, 0x02, 0x8b, 0x1e, 0x6d, 0x50, 0x70, + 0x64, 0xeb, 0xff, 0x88, 0x94, 0x69, 0x0d, 0xeb, 0x4f, 0xd5, 0x26, 0x46, 0x7f, 0x47, 0x80, 0x8c, + 0xf7, 0x01, 0x6d, 0x74, 0x73, 0x6c, 0xf8, 0xe7, 0x7a, 0x51, 0x3c, 0x77, 0x6b, 0x42, 0x6c, 0x66, + 0xee, 0xe2, 0xd6, 0x5f, 0xfa, 0xe1, 0x7f, 0xfd, 0xe5, 0xc8, 0x4d, 0xf1, 0xfa, 0xe6, 0xd3, 0xad, + 0xcd, 0x9f, 0x66, 0x73, 0xec, 0xdb, 0x7d, 0x5d, 0x7b, 0x8c, 0x9b, 0xa6, 0xb1, 0xb9, 0xfe, 0xb3, + 0x9b, 0xfc, 0x7f, 0x04, 0xdd, 0xe6, 0x71, 0xca, 0x6d, 0x61, 0x1d, 0x7d, 0x47, 0x80, 0x94, 0xeb, + 0xff, 0x37, 0xa0, 0x37, 0x26, 0xfe, 0x7f, 0x1d, 0xb9, 0xf5, 0x49, 0x50, 0xb9, 0x68, 0x9b, 0x54, + 0xb4, 0x37, 0xc4, 0xd7, 0xc2, 0x44, 0xa3, 0xff, 0x23, 0xe2, 0x36, 0xbb, 0x72, 0x42, 0xe4, 0xfa, + 0x2d, 0x01, 0x16, 0x87, 0xfe, 0x4d, 0x00, 0xda, 0x9c, 0x24, 0xfb, 0xee, 0xd6, 0xe0, 0x5b, 0x93, + 0x13, 0x70, 0x49, 0xdf, 0xa3, 0x92, 0x6e, 0x8a, 0xeb, 0xe3, 0x94, 0xe8, 0xb8, 0x04, 0x4b, 0x5e, + 0x69, 0x62, 0x79, 0xa5, 0x69, 0xe5, 0x95, 0x9e, 0x5f, 0x5e, 0xdd, 0x23, 0xef, 0xcf, 0x0b, 0x90, + 0xf6, 0x3c, 0x45, 0x8c, 0xde, 0x0c, 0xfd, 0xff, 0x5f, 0xc3, 0xaf, 0x20, 0xe7, 0x6e, 0x4e, 0x86, + 0xcc, 0x65, 0x5c, 0xa1, 0x32, 0x2e, 0xa0, 0x34, 0x91, 0xd1, 0x39, 0x82, 0xf5, 0x47, 0x02, 0xac, + 0x04, 0x66, 0x02, 0xd1, 0x3b, 0xa1, 0xa7, 0xd7, 0xc2, 0xf3, 0x86, 0xb9, 0x09, 0xd3, 0x40, 0x62, + 0x9b, 0x4a, 0xa3, 0x88, 0xb7, 0xdc, 0x1a, 0xd3, 0xf4, 0xb6, 0xd2, 0x53, 0xbf, 0x62, 0x9f, 0x98, + 0xa9, 0x41, 0xfa, 0x72, 0x45, 0xb7, 0x85, 0xf5, 0x47, 0xb7, 0xc4, 0x1b, 0xa1, 0xf6, 0x3b, 0x8c, + 0x4e, 0xfb, 0x17, 0x98, 0x3c, 0x0c, 0xed, 0xdf, 0xa8, 0x54, 0xe3, 0xb4, 0xfd, 0xdb, 0x62, 0xfd, + 0x23, 0x6b, 0x89, 0xbf, 0x77, 0x43, 0xd2, 0x6e, 0xae, 0xff, 0x2c, 0xed, 0xdf, 0xd6, 0x0d, 0x87, + 0xc6, 0xe9, 0x5d, 0x08, 0x3a, 0xfa, 0x81, 0x00, 0x68, 0x38, 0xb1, 0x88, 0xc2, 0x6c, 0x38, 0x34, + 0x07, 0x39, 0x71, 0xcf, 0x14, 0xda, 0xb3, 0x9f, 0x42, 0xd3, 0xf5, 0xec, 0xd1, 0x3a, 0x9a, 0xb8, + 0x5b, 0xe8, 0x47, 0x82, 0xf5, 0xee, 0xb7, 0x2f, 0x55, 0xb8, 0x35, 0xd2, 0xe2, 0x03, 0x13, 0xa5, + 0xb9, 0x77, 0xa6, 0xa2, 0xe1, 0x93, 0xc5, 0xdb, 0xc9, 0x49, 0xcd, 0xd3, 0xee, 0xe4, 0x04, 0xb6, + 0x89, 0xfe, 0x50, 0x80, 0x95, 0xc0, 0x7c, 0x66, 0xa8, 0x61, 0x8e, 0xca, 0x7e, 0xe6, 0x42, 0xde, + 0xb4, 0xb4, 0x7a, 0xb2, 0x3e, 0xed, 0x70, 0xad, 0x4f, 0x3e, 0x5c, 0x7f, 0x2a, 0xc0, 0x5a, 0xd8, + 0xce, 0x01, 0xbd, 0x3f, 0xd2, 0x8b, 0x84, 0x46, 0x63, 0xb9, 0xc9, 0x03, 0x43, 0xb1, 0x4b, 0xbb, + 0xd8, 0x16, 0xdf, 0x1a, 0x39, 0x58, 0x01, 0xd1, 0x23, 0x99, 0x6e, 0x6f, 0x89, 0x6f, 0x86, 0x0d, + 0x59, 0x30, 0x05, 0xed, 0x6e, 0xd8, 0x7e, 0x23, 0xb4, 0xbb, 0x63, 0x36, 0x28, 0x67, 0xe8, 0xee, + 0xd6, 0x5b, 0xe1, 0x23, 0x1a, 0x20, 0x39, 0xf7, 0x2e, 0x6f, 0x6d, 0xbd, 0x19, 0x38, 0xae, 0xa1, + 0x14, 0xe8, 0x3f, 0x09, 0xb0, 0x12, 0xb8, 0x1d, 0x09, 0xb5, 0xd3, 0x51, 0x9b, 0x97, 0x69, 0x3a, + 0xca, 0x7d, 0x28, 0x9a, 0xba, 0xa3, 0x8f, 0x6e, 0xa1, 0x69, 0x7a, 0x89, 0xfe, 0x07, 0x4f, 0xa1, + 0x06, 0x6c, 0x63, 0xd0, 0x7b, 0x23, 0xdc, 0x47, 0xf8, 0x6e, 0x2b, 0xf7, 0xfe, 0xb4, 0x64, 0xdc, + 0xf1, 0x78, 0xfb, 0x3c, 0x85, 0x2d, 0xdb, 0x7d, 0x9e, 0xcc, 0x90, 0xd1, 0x0f, 0x05, 0x58, 0x0b, + 0xdb, 0x11, 0x85, 0x5a, 0xf1, 0x98, 0x2d, 0x54, 0xa8, 0x13, 0xe2, 0xbd, 0x5a, 0x3f, 0xc3, 0x48, + 0xae, 0x4f, 0x35, 0x92, 0xdf, 0x11, 0x20, 0xeb, 0xff, 0xde, 0x83, 0x36, 0x46, 0xba, 0xa0, 0xa1, + 0x64, 0x75, 0x6e, 0x7c, 0x7e, 0x5c, 0xdc, 0xa0, 0x1d, 0xba, 0x21, 0x5e, 0x0d, 0x53, 0xb9, 0x2b, + 0x7f, 0xce, 0x23, 0xfc, 0xac, 0xff, 0x83, 0x4f, 0xa8, 0x5c, 0x21, 0x5f, 0x86, 0xa6, 0x90, 0x6b, + 0xeb, 0x6a, 0xa0, 0xd2, 0x5c, 0x42, 0xf1, 0xc9, 0xfd, 0xd7, 0x04, 0x48, 0x7b, 0x3e, 0xec, 0x84, + 0x46, 0xa0, 0x41, 0x9f, 0x7f, 0x26, 0x91, 0xe8, 0x4d, 0x2a, 0xd1, 0xeb, 0x68, 0x12, 0x89, 0xd0, + 0xdf, 0x16, 0x60, 0xc1, 0xf7, 0x2d, 0x03, 0xdd, 0x1a, 0x31, 0x93, 0x86, 0xbf, 0xbe, 0xe4, 0x36, + 0x26, 0x45, 0xe7, 0x13, 0xce, 0x2b, 0xdf, 0xe8, 0x91, 0x44, 0xbf, 0x40, 0x1f, 0x4e, 0xf0, 0x7e, + 0xe2, 0x08, 0x1d, 0xc6, 0x90, 0x6f, 0x21, 0xa1, 0x93, 0x84, 0x4b, 0xb2, 0x3e, 0x91, 0xa6, 0xfe, + 0x8a, 0x00, 0xf3, 0xee, 0x8f, 0x6c, 0x68, 0x7d, 0xf4, 0x3a, 0xeb, 0x4e, 0xff, 0xe5, 0x46, 0x67, + 0x23, 0xc5, 0x75, 0x2a, 0xc8, 0x6b, 0xe2, 0xe5, 0x50, 0x7f, 0xc2, 0x72, 0x9d, 0xc4, 0x80, 0xbe, + 0x16, 0x20, 0xe5, 0x4a, 0x7f, 0x86, 0x6e, 0x5d, 0x87, 0x73, 0xbb, 0xa1, 0x5b, 0xd7, 0x80, 0x6c, + 0xaa, 0x78, 0x9d, 0x8a, 0xf4, 0x2a, 0x1a, 0x27, 0x12, 0xfa, 0x19, 0x48, 0xda, 0xf9, 0x66, 0x74, + 0x7d, 0xc4, 0x02, 0x35, 0x8d, 0x42, 0xbc, 0xad, 0x0f, 0xb9, 0x22, 0xd6, 0x34, 0x19, 0x95, 0x9f, + 0x86, 0x79, 0x77, 0xba, 0x35, 0x74, 0x50, 0x02, 0x72, 0xb2, 0xa1, 0x66, 0xc1, 0x1b, 0x5f, 0x1f, + 0xdb, 0x38, 0x35, 0x09, 0x57, 0x5e, 0x38, 0xdc, 0x24, 0x86, 0x93, 0xc7, 0xa1, 0xad, 0xbf, 0x4d, + 0x5b, 0x7f, 0x53, 0xbc, 0x36, 0xa6, 0xf5, 0xdb, 0x4d, 0xca, 0xf4, 0xb6, 0xb0, 0xbe, 0xfd, 0xb5, + 0x00, 0xe7, 0x9b, 0x5a, 0x37, 0xb8, 0xf1, 0xed, 0x44, 0xb1, 0xd3, 0xdf, 0x27, 0x6d, 0xec, 0x0b, + 0x8f, 0x3e, 0xe0, 0x28, 0x6d, 0xad, 0xa3, 0xf4, 0xda, 0x1b, 0x9a, 0xde, 0xde, 0x6c, 0xe3, 0x1e, + 0x95, 0x60, 0x93, 0x55, 0x29, 0x7d, 0xd5, 0xd8, 0xe4, 0x1c, 0x48, 0x6b, 0x9b, 0x4f, 0xb7, 0x3e, + 0x6a, 0x75, 0xfa, 0xbf, 0x19, 0x59, 0xba, 0xcb, 0x48, 0x0b, 0x1d, 0x6d, 0xd0, 0x22, 0xe3, 0xb4, + 0xf1, 0x60, 0xeb, 0xdf, 0x5a, 0xd0, 0x2f, 0x28, 0xf4, 0x8b, 0x62, 0xa7, 0xff, 0xc5, 0x83, 0xad, + 0x46, 0x9c, 0x32, 0x7c, 0xe7, 0xff, 0x06, 0x00, 0x00, 0xff, 0xff, 0x69, 0x25, 0xf9, 0x3d, 0x54, + 0x83, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/storage.pb.go b/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/storage.pb.go index 9d2dbae90..470aa610b 100644 --- a/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/storage.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/privacy/dlp/v2/storage.pb.go @@ -25,7 +25,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Likelihood int32 const ( - // Default value; information with all likelihoods is included. + // Default value; same as POSSIBLE. Likelihood_LIKELIHOOD_UNSPECIFIED Likelihood = 0 // Few matching elements. Likelihood_VERY_UNLIKELY Likelihood = 1 @@ -58,7 +58,7 @@ func (x Likelihood) String() string { return proto.EnumName(Likelihood_name, int32(x)) } func (Likelihood) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{0} } // Definitions of file type groups to scan. @@ -93,13 +93,79 @@ func (x FileType) String() string { return proto.EnumName(FileType_name, int32(x)) } func (FileType) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1} +} + +// How to sample bytes if not all bytes are scanned. Meaningful only when used +// in conjunction with bytes_limit_per_file. If not specified, scanning would +// start from the top. +type CloudStorageOptions_SampleMethod int32 + +const ( + CloudStorageOptions_SAMPLE_METHOD_UNSPECIFIED CloudStorageOptions_SampleMethod = 0 + // Scan from the top (default). + CloudStorageOptions_TOP CloudStorageOptions_SampleMethod = 1 + // For each file larger than bytes_limit_per_file, randomly pick the offset + // to start scanning. The scanned bytes are contiguous. + CloudStorageOptions_RANDOM_START CloudStorageOptions_SampleMethod = 2 +) + +var CloudStorageOptions_SampleMethod_name = map[int32]string{ + 0: "SAMPLE_METHOD_UNSPECIFIED", + 1: "TOP", + 2: "RANDOM_START", +} +var CloudStorageOptions_SampleMethod_value = map[string]int32{ + "SAMPLE_METHOD_UNSPECIFIED": 0, + "TOP": 1, + "RANDOM_START": 2, +} + +func (x CloudStorageOptions_SampleMethod) String() string { + return proto.EnumName(CloudStorageOptions_SampleMethod_name, int32(x)) +} +func (CloudStorageOptions_SampleMethod) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_storage_fb7f209e00c99d61, []int{6, 0} +} + +// How to sample rows if not all rows are scanned. Meaningful only when used +// in conjunction with rows_limit. If not specified, scanning would start +// from the top. +type BigQueryOptions_SampleMethod int32 + +const ( + BigQueryOptions_SAMPLE_METHOD_UNSPECIFIED BigQueryOptions_SampleMethod = 0 + // Scan from the top (default). + BigQueryOptions_TOP BigQueryOptions_SampleMethod = 1 + // Randomly pick the row to start scanning. The scanned rows are contiguous. + BigQueryOptions_RANDOM_START BigQueryOptions_SampleMethod = 2 +) + +var BigQueryOptions_SampleMethod_name = map[int32]string{ + 0: "SAMPLE_METHOD_UNSPECIFIED", + 1: "TOP", + 2: "RANDOM_START", +} +var BigQueryOptions_SampleMethod_value = map[string]int32{ + "SAMPLE_METHOD_UNSPECIFIED": 0, + "TOP": 1, + "RANDOM_START": 2, +} + +func (x BigQueryOptions_SampleMethod) String() string { + return proto.EnumName(BigQueryOptions_SampleMethod_name, int32(x)) +} +func (BigQueryOptions_SampleMethod) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_storage_fb7f209e00c99d61, []int{8, 0} } // Type of information detected by the API. type InfoType struct { - // Name of the information type. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + // Name of the information type. Either a name of your choosing when + // creating a CustomInfoType, or one of the names listed + // at https://cloud.google.com/dlp/docs/infotypes-reference when specifying + // a built-in type. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -109,7 +175,7 @@ func (m *InfoType) Reset() { *m = InfoType{} } func (m *InfoType) String() string { return proto.CompactTextString(m) } func (*InfoType) ProtoMessage() {} func (*InfoType) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{0} } func (m *InfoType) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InfoType.Unmarshal(m, b) @@ -139,22 +205,22 @@ func (m *InfoType) GetName() string { // Custom information type provided by the user. Used to find domain-specific // sensitive information configurable to the data in question. type CustomInfoType struct { - // Info type configuration. All custom info types must have configurations - // that do not conflict with built-in info types or other custom info types. - InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType" json:"info_type,omitempty"` - // Likelihood to return for this custom info type. This base value can be + // All CustomInfoTypes must have a name + // that does not conflict with built-in InfoTypes or other CustomInfoTypes. + InfoType *InfoType `protobuf:"bytes,1,opt,name=info_type,json=infoType,proto3" json:"info_type,omitempty"` + // Likelihood to return for this CustomInfoType. This base value can be // altered by a detection rule if the finding meets the criteria specified by // the rule. Defaults to `VERY_LIKELY` if not specified. - Likelihood Likelihood `protobuf:"varint,6,opt,name=likelihood,enum=google.privacy.dlp.v2.Likelihood" json:"likelihood,omitempty"` + Likelihood Likelihood `protobuf:"varint,6,opt,name=likelihood,proto3,enum=google.privacy.dlp.v2.Likelihood" json:"likelihood,omitempty"` // Types that are valid to be assigned to Type: // *CustomInfoType_Dictionary_ // *CustomInfoType_Regex_ // *CustomInfoType_SurrogateType_ Type isCustomInfoType_Type `protobuf_oneof:"type"` - // Set of detection rules to apply to all findings of this custom info type. + // Set of detection rules to apply to all findings of this CustomInfoType. // Rules are applied in order that they are specified. Not supported for the - // `surrogate_type` custom info type. - DetectionRules []*CustomInfoType_DetectionRule `protobuf:"bytes,7,rep,name=detection_rules,json=detectionRules" json:"detection_rules,omitempty"` + // `surrogate_type` CustomInfoType. + DetectionRules []*CustomInfoType_DetectionRule `protobuf:"bytes,7,rep,name=detection_rules,json=detectionRules,proto3" json:"detection_rules,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -164,7 +230,7 @@ func (m *CustomInfoType) Reset() { *m = CustomInfoType{} } func (m *CustomInfoType) String() string { return proto.CompactTextString(m) } func (*CustomInfoType) ProtoMessage() {} func (*CustomInfoType) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1} } func (m *CustomInfoType) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType.Unmarshal(m, b) @@ -189,13 +255,13 @@ type isCustomInfoType_Type interface { } type CustomInfoType_Dictionary_ struct { - Dictionary *CustomInfoType_Dictionary `protobuf:"bytes,2,opt,name=dictionary,oneof"` + Dictionary *CustomInfoType_Dictionary `protobuf:"bytes,2,opt,name=dictionary,proto3,oneof"` } type CustomInfoType_Regex_ struct { - Regex *CustomInfoType_Regex `protobuf:"bytes,3,opt,name=regex,oneof"` + Regex *CustomInfoType_Regex `protobuf:"bytes,3,opt,name=regex,proto3,oneof"` } type CustomInfoType_SurrogateType_ struct { - SurrogateType *CustomInfoType_SurrogateType `protobuf:"bytes,4,opt,name=surrogate_type,json=surrogateType,oneof"` + SurrogateType *CustomInfoType_SurrogateType `protobuf:"bytes,4,opt,name=surrogate_type,json=surrogateType,proto3,oneof"` } func (*CustomInfoType_Dictionary_) isCustomInfoType_Type() {} @@ -377,7 +443,7 @@ func (m *CustomInfoType_Dictionary) Reset() { *m = CustomInfoType_Dictio func (m *CustomInfoType_Dictionary) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_Dictionary) ProtoMessage() {} func (*CustomInfoType_Dictionary) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 0} } func (m *CustomInfoType_Dictionary) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_Dictionary.Unmarshal(m, b) @@ -402,10 +468,10 @@ type isCustomInfoType_Dictionary_Source interface { } type CustomInfoType_Dictionary_WordList_ struct { - WordList *CustomInfoType_Dictionary_WordList `protobuf:"bytes,1,opt,name=word_list,json=wordList,oneof"` + WordList *CustomInfoType_Dictionary_WordList `protobuf:"bytes,1,opt,name=word_list,json=wordList,proto3,oneof"` } type CustomInfoType_Dictionary_CloudStoragePath struct { - CloudStoragePath *CloudStoragePath `protobuf:"bytes,3,opt,name=cloud_storage_path,json=cloudStoragePath,oneof"` + CloudStoragePath *CloudStoragePath `protobuf:"bytes,3,opt,name=cloud_storage_path,json=cloudStoragePath,proto3,oneof"` } func (*CustomInfoType_Dictionary_WordList_) isCustomInfoType_Dictionary_Source() {} @@ -511,7 +577,7 @@ type CustomInfoType_Dictionary_WordList struct { // Words or phrases defining the dictionary. The dictionary must contain // at least one phrase and every phrase must contain at least 2 characters // that are letters or digits. [required] - Words []string `protobuf:"bytes,1,rep,name=words" json:"words,omitempty"` + Words []string `protobuf:"bytes,1,rep,name=words,proto3" json:"words,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -521,7 +587,7 @@ func (m *CustomInfoType_Dictionary_WordList) Reset() { *m = CustomInfoTy func (m *CustomInfoType_Dictionary_WordList) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_Dictionary_WordList) ProtoMessage() {} func (*CustomInfoType_Dictionary_WordList) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 0, 0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 0, 0} } func (m *CustomInfoType_Dictionary_WordList) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_Dictionary_WordList.Unmarshal(m, b) @@ -551,7 +617,7 @@ func (m *CustomInfoType_Dictionary_WordList) GetWords() []string { // Message defining a custom regular expression. type CustomInfoType_Regex struct { // Pattern defining the regular expression. - Pattern string `protobuf:"bytes,1,opt,name=pattern" json:"pattern,omitempty"` + Pattern string `protobuf:"bytes,1,opt,name=pattern,proto3" json:"pattern,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -561,7 +627,7 @@ func (m *CustomInfoType_Regex) Reset() { *m = CustomInfoType_Regex{} } func (m *CustomInfoType_Regex) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_Regex) ProtoMessage() {} func (*CustomInfoType_Regex) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 1} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 1} } func (m *CustomInfoType_Regex) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_Regex.Unmarshal(m, b) @@ -594,7 +660,7 @@ func (m *CustomInfoType_Regex) GetPattern() string { // These types of transformations are // those that perform pseudonymization, thereby producing a "surrogate" as // output. This should be used in conjunction with a field on the -// transformation such as `surrogate_info_type`. This custom info type does +// transformation such as `surrogate_info_type`. This CustomInfoType does // not support the use of `detection_rules`. type CustomInfoType_SurrogateType struct { XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -606,7 +672,7 @@ func (m *CustomInfoType_SurrogateType) Reset() { *m = CustomInfoType_Sur func (m *CustomInfoType_SurrogateType) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_SurrogateType) ProtoMessage() {} func (*CustomInfoType_SurrogateType) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 2} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 2} } func (m *CustomInfoType_SurrogateType) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_SurrogateType.Unmarshal(m, b) @@ -626,7 +692,7 @@ func (m *CustomInfoType_SurrogateType) XXX_DiscardUnknown() { var xxx_messageInfo_CustomInfoType_SurrogateType proto.InternalMessageInfo -// Rule for modifying a custom info type to alter behavior under certain +// Rule for modifying a CustomInfoType to alter behavior under certain // circumstances, depending on the specific details of the rule. Not supported // for the `surrogate_type` custom info type. type CustomInfoType_DetectionRule struct { @@ -642,7 +708,7 @@ func (m *CustomInfoType_DetectionRule) Reset() { *m = CustomInfoType_Det func (m *CustomInfoType_DetectionRule) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_DetectionRule) ProtoMessage() {} func (*CustomInfoType_DetectionRule) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 3} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 3} } func (m *CustomInfoType_DetectionRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_DetectionRule.Unmarshal(m, b) @@ -667,7 +733,7 @@ type isCustomInfoType_DetectionRule_Type interface { } type CustomInfoType_DetectionRule_HotwordRule_ struct { - HotwordRule *CustomInfoType_DetectionRule_HotwordRule `protobuf:"bytes,1,opt,name=hotword_rule,json=hotwordRule,oneof"` + HotwordRule *CustomInfoType_DetectionRule_HotwordRule `protobuf:"bytes,1,opt,name=hotword_rule,json=hotwordRule,proto3,oneof"` } func (*CustomInfoType_DetectionRule_HotwordRule_) isCustomInfoType_DetectionRule_Type() {} @@ -745,9 +811,9 @@ func _CustomInfoType_DetectionRule_OneofSizer(msg proto.Message) (n int) { // rule. type CustomInfoType_DetectionRule_Proximity struct { // Number of characters before the finding to consider. - WindowBefore int32 `protobuf:"varint,1,opt,name=window_before,json=windowBefore" json:"window_before,omitempty"` + WindowBefore int32 `protobuf:"varint,1,opt,name=window_before,json=windowBefore,proto3" json:"window_before,omitempty"` // Number of characters after the finding to consider. - WindowAfter int32 `protobuf:"varint,2,opt,name=window_after,json=windowAfter" json:"window_after,omitempty"` + WindowAfter int32 `protobuf:"varint,2,opt,name=window_after,json=windowAfter,proto3" json:"window_after,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -759,7 +825,7 @@ func (m *CustomInfoType_DetectionRule_Proximity) Reset() { func (m *CustomInfoType_DetectionRule_Proximity) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_DetectionRule_Proximity) ProtoMessage() {} func (*CustomInfoType_DetectionRule_Proximity) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 3, 0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 3, 0} } func (m *CustomInfoType_DetectionRule_Proximity) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_DetectionRule_Proximity.Unmarshal(m, b) @@ -813,7 +879,7 @@ func (m *CustomInfoType_DetectionRule_LikelihoodAdjustment) String() string { } func (*CustomInfoType_DetectionRule_LikelihoodAdjustment) ProtoMessage() {} func (*CustomInfoType_DetectionRule_LikelihoodAdjustment) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 3, 1} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 3, 1} } func (m *CustomInfoType_DetectionRule_LikelihoodAdjustment) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_DetectionRule_LikelihoodAdjustment.Unmarshal(m, b) @@ -838,10 +904,10 @@ type isCustomInfoType_DetectionRule_LikelihoodAdjustment_Adjustment interface { } type CustomInfoType_DetectionRule_LikelihoodAdjustment_FixedLikelihood struct { - FixedLikelihood Likelihood `protobuf:"varint,1,opt,name=fixed_likelihood,json=fixedLikelihood,enum=google.privacy.dlp.v2.Likelihood,oneof"` + FixedLikelihood Likelihood `protobuf:"varint,1,opt,name=fixed_likelihood,json=fixedLikelihood,proto3,enum=google.privacy.dlp.v2.Likelihood,oneof"` } type CustomInfoType_DetectionRule_LikelihoodAdjustment_RelativeLikelihood struct { - RelativeLikelihood int32 `protobuf:"varint,2,opt,name=relative_likelihood,json=relativeLikelihood,oneof"` + RelativeLikelihood int32 `protobuf:"varint,2,opt,name=relative_likelihood,json=relativeLikelihood,proto3,oneof"` } func (*CustomInfoType_DetectionRule_LikelihoodAdjustment_FixedLikelihood) isCustomInfoType_DetectionRule_LikelihoodAdjustment_Adjustment() { @@ -934,11 +1000,11 @@ func _CustomInfoType_DetectionRule_LikelihoodAdjustment_OneofSizer(msg proto.Mes return n } -// Detection rule that adjusts the likelihood of findings within a certain +// The rule that adjusts the likelihood of findings within a certain // proximity of hotwords. type CustomInfoType_DetectionRule_HotwordRule struct { - // Regex pattern defining what qualifies as a hotword. - HotwordRegex *CustomInfoType_Regex `protobuf:"bytes,1,opt,name=hotword_regex,json=hotwordRegex" json:"hotword_regex,omitempty"` + // Regular expression pattern defining what qualifies as a hotword. + HotwordRegex *CustomInfoType_Regex `protobuf:"bytes,1,opt,name=hotword_regex,json=hotwordRegex,proto3" json:"hotword_regex,omitempty"` // Proximity of the finding within which the entire hotword must reside. // The total length of the window cannot exceed 1000 characters. Note that // the finding itself will be included in the window, so that hotwords may @@ -947,9 +1013,9 @@ type CustomInfoType_DetectionRule_HotwordRule struct { // adjusted upwards if the area code is known to be the local area code of // a company office using the hotword regex "\(xxx\)", where "xxx" // is the area code in question. - Proximity *CustomInfoType_DetectionRule_Proximity `protobuf:"bytes,2,opt,name=proximity" json:"proximity,omitempty"` + Proximity *CustomInfoType_DetectionRule_Proximity `protobuf:"bytes,2,opt,name=proximity,proto3" json:"proximity,omitempty"` // Likelihood adjustment to apply to all matching findings. - LikelihoodAdjustment *CustomInfoType_DetectionRule_LikelihoodAdjustment `protobuf:"bytes,3,opt,name=likelihood_adjustment,json=likelihoodAdjustment" json:"likelihood_adjustment,omitempty"` + LikelihoodAdjustment *CustomInfoType_DetectionRule_LikelihoodAdjustment `protobuf:"bytes,3,opt,name=likelihood_adjustment,json=likelihoodAdjustment,proto3" json:"likelihood_adjustment,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -961,7 +1027,7 @@ func (m *CustomInfoType_DetectionRule_HotwordRule) Reset() { func (m *CustomInfoType_DetectionRule_HotwordRule) String() string { return proto.CompactTextString(m) } func (*CustomInfoType_DetectionRule_HotwordRule) ProtoMessage() {} func (*CustomInfoType_DetectionRule_HotwordRule) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{1, 3, 2} + return fileDescriptor_storage_fb7f209e00c99d61, []int{1, 3, 2} } func (m *CustomInfoType_DetectionRule_HotwordRule) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CustomInfoType_DetectionRule_HotwordRule.Unmarshal(m, b) @@ -1005,7 +1071,7 @@ func (m *CustomInfoType_DetectionRule_HotwordRule) GetLikelihoodAdjustment() *Cu // General identifier of a data field in a storage service. type FieldId struct { // Name describing the field. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1015,7 +1081,7 @@ func (m *FieldId) Reset() { *m = FieldId{} } func (m *FieldId) String() string { return proto.CompactTextString(m) } func (*FieldId) ProtoMessage() {} func (*FieldId) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{2} + return fileDescriptor_storage_fb7f209e00c99d61, []int{2} } func (m *FieldId) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_FieldId.Unmarshal(m, b) @@ -1050,9 +1116,9 @@ func (m *FieldId) GetName() string { // project ID and namespace ID. type PartitionId struct { // The ID of the project to which the entities belong. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // If not empty, the ID of the namespace to which the entities belong. - NamespaceId string `protobuf:"bytes,4,opt,name=namespace_id,json=namespaceId" json:"namespace_id,omitempty"` + NamespaceId string `protobuf:"bytes,4,opt,name=namespace_id,json=namespaceId,proto3" json:"namespace_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1062,7 +1128,7 @@ func (m *PartitionId) Reset() { *m = PartitionId{} } func (m *PartitionId) String() string { return proto.CompactTextString(m) } func (*PartitionId) ProtoMessage() {} func (*PartitionId) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{3} + return fileDescriptor_storage_fb7f209e00c99d61, []int{3} } func (m *PartitionId) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PartitionId.Unmarshal(m, b) @@ -1099,7 +1165,7 @@ func (m *PartitionId) GetNamespaceId() string { // A representation of a Datastore kind. type KindExpression struct { // The name of the kind. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1109,7 +1175,7 @@ func (m *KindExpression) Reset() { *m = KindExpression{} } func (m *KindExpression) String() string { return proto.CompactTextString(m) } func (*KindExpression) ProtoMessage() {} func (*KindExpression) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{4} + return fileDescriptor_storage_fb7f209e00c99d61, []int{4} } func (m *KindExpression) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_KindExpression.Unmarshal(m, b) @@ -1140,9 +1206,9 @@ func (m *KindExpression) GetName() string { type DatastoreOptions struct { // A partition ID identifies a grouping of entities. The grouping is always // by project and namespace, however the namespace ID may be empty. - PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId" json:"partition_id,omitempty"` + PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId,proto3" json:"partition_id,omitempty"` // The kind to process. - Kind *KindExpression `protobuf:"bytes,2,opt,name=kind" json:"kind,omitempty"` + Kind *KindExpression `protobuf:"bytes,2,opt,name=kind,proto3" json:"kind,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1152,7 +1218,7 @@ func (m *DatastoreOptions) Reset() { *m = DatastoreOptions{} } func (m *DatastoreOptions) String() string { return proto.CompactTextString(m) } func (*DatastoreOptions) ProtoMessage() {} func (*DatastoreOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{5} + return fileDescriptor_storage_fb7f209e00c99d61, []int{5} } func (m *DatastoreOptions) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DatastoreOptions.Unmarshal(m, b) @@ -1189,24 +1255,29 @@ func (m *DatastoreOptions) GetKind() *KindExpression { // Options defining a file or a set of files (path ending with *) within // a Google Cloud Storage bucket. type CloudStorageOptions struct { - FileSet *CloudStorageOptions_FileSet `protobuf:"bytes,1,opt,name=file_set,json=fileSet" json:"file_set,omitempty"` + FileSet *CloudStorageOptions_FileSet `protobuf:"bytes,1,opt,name=file_set,json=fileSet,proto3" json:"file_set,omitempty"` // Max number of bytes to scan from a file. If a scanned file's size is bigger // than this value then the rest of the bytes are omitted. - BytesLimitPerFile int64 `protobuf:"varint,4,opt,name=bytes_limit_per_file,json=bytesLimitPerFile" json:"bytes_limit_per_file,omitempty"` + BytesLimitPerFile int64 `protobuf:"varint,4,opt,name=bytes_limit_per_file,json=bytesLimitPerFile,proto3" json:"bytes_limit_per_file,omitempty"` // List of file type groups to include in the scan. // If empty, all files are scanned and available data format processors // are applied. - FileTypes []FileType `protobuf:"varint,5,rep,packed,name=file_types,json=fileTypes,enum=google.privacy.dlp.v2.FileType" json:"file_types,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + FileTypes []FileType `protobuf:"varint,5,rep,packed,name=file_types,json=fileTypes,proto3,enum=google.privacy.dlp.v2.FileType" json:"file_types,omitempty"` + SampleMethod CloudStorageOptions_SampleMethod `protobuf:"varint,6,opt,name=sample_method,json=sampleMethod,proto3,enum=google.privacy.dlp.v2.CloudStorageOptions_SampleMethod" json:"sample_method,omitempty"` + // Limits the number of files to scan to this percentage of the input FileSet. + // Number of files scanned is rounded down. Must be between 0 and 100, + // inclusively. Both 0 and 100 means no limit. Defaults to 0. + FilesLimitPercent int32 `protobuf:"varint,7,opt,name=files_limit_percent,json=filesLimitPercent,proto3" json:"files_limit_percent,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *CloudStorageOptions) Reset() { *m = CloudStorageOptions{} } func (m *CloudStorageOptions) String() string { return proto.CompactTextString(m) } func (*CloudStorageOptions) ProtoMessage() {} func (*CloudStorageOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{6} + return fileDescriptor_storage_fb7f209e00c99d61, []int{6} } func (m *CloudStorageOptions) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CloudStorageOptions.Unmarshal(m, b) @@ -1247,11 +1318,25 @@ func (m *CloudStorageOptions) GetFileTypes() []FileType { return nil } +func (m *CloudStorageOptions) GetSampleMethod() CloudStorageOptions_SampleMethod { + if m != nil { + return m.SampleMethod + } + return CloudStorageOptions_SAMPLE_METHOD_UNSPECIFIED +} + +func (m *CloudStorageOptions) GetFilesLimitPercent() int32 { + if m != nil { + return m.FilesLimitPercent + } + return 0 +} + // Set of files to scan. type CloudStorageOptions_FileSet struct { // The url, in the format `gs:///`. Trailing wildcard in the // path is allowed. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1261,7 +1346,7 @@ func (m *CloudStorageOptions_FileSet) Reset() { *m = CloudStorageOptions func (m *CloudStorageOptions_FileSet) String() string { return proto.CompactTextString(m) } func (*CloudStorageOptions_FileSet) ProtoMessage() {} func (*CloudStorageOptions_FileSet) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{6, 0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{6, 0} } func (m *CloudStorageOptions_FileSet) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CloudStorageOptions_FileSet.Unmarshal(m, b) @@ -1288,11 +1373,11 @@ func (m *CloudStorageOptions_FileSet) GetUrl() string { return "" } -// Message representing a path in Cloud Storage. +// Message representing a single file or path in Cloud Storage. type CloudStoragePath struct { // A url representing a file or path (no wildcards) in Cloud Storage. // Example: gs://[BUCKET_NAME]/dictionary.txt - Path string `protobuf:"bytes,1,opt,name=path" json:"path,omitempty"` + Path string `protobuf:"bytes,1,opt,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1302,7 +1387,7 @@ func (m *CloudStoragePath) Reset() { *m = CloudStoragePath{} } func (m *CloudStoragePath) String() string { return proto.CompactTextString(m) } func (*CloudStoragePath) ProtoMessage() {} func (*CloudStoragePath) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{7} + return fileDescriptor_storage_fb7f209e00c99d61, []int{7} } func (m *CloudStoragePath) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CloudStoragePath.Unmarshal(m, b) @@ -1332,24 +1417,25 @@ func (m *CloudStoragePath) GetPath() string { // Options defining BigQuery table and row identifiers. type BigQueryOptions struct { // Complete BigQuery table reference. - TableReference *BigQueryTable `protobuf:"bytes,1,opt,name=table_reference,json=tableReference" json:"table_reference,omitempty"` + TableReference *BigQueryTable `protobuf:"bytes,1,opt,name=table_reference,json=tableReference,proto3" json:"table_reference,omitempty"` // References to fields uniquely identifying rows within the table. // Nested fields in the format, like `person.birthdate.year`, are allowed. - IdentifyingFields []*FieldId `protobuf:"bytes,2,rep,name=identifying_fields,json=identifyingFields" json:"identifying_fields,omitempty"` + IdentifyingFields []*FieldId `protobuf:"bytes,2,rep,name=identifying_fields,json=identifyingFields,proto3" json:"identifying_fields,omitempty"` // Max number of rows to scan. If the table has more rows than this value, the // rest of the rows are omitted. If not set, or if set to 0, all rows will be // scanned. Cannot be used in conjunction with TimespanConfig. - RowsLimit int64 `protobuf:"varint,3,opt,name=rows_limit,json=rowsLimit" json:"rows_limit,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + RowsLimit int64 `protobuf:"varint,3,opt,name=rows_limit,json=rowsLimit,proto3" json:"rows_limit,omitempty"` + SampleMethod BigQueryOptions_SampleMethod `protobuf:"varint,4,opt,name=sample_method,json=sampleMethod,proto3,enum=google.privacy.dlp.v2.BigQueryOptions_SampleMethod" json:"sample_method,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *BigQueryOptions) Reset() { *m = BigQueryOptions{} } func (m *BigQueryOptions) String() string { return proto.CompactTextString(m) } func (*BigQueryOptions) ProtoMessage() {} func (*BigQueryOptions) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{8} + return fileDescriptor_storage_fb7f209e00c99d61, []int{8} } func (m *BigQueryOptions) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BigQueryOptions.Unmarshal(m, b) @@ -1390,6 +1476,13 @@ func (m *BigQueryOptions) GetRowsLimit() int64 { return 0 } +func (m *BigQueryOptions) GetSampleMethod() BigQueryOptions_SampleMethod { + if m != nil { + return m.SampleMethod + } + return BigQueryOptions_SAMPLE_METHOD_UNSPECIFIED +} + // Shared message indicating Cloud storage type. type StorageConfig struct { // Types that are valid to be assigned to Type: @@ -1397,7 +1490,7 @@ type StorageConfig struct { // *StorageConfig_CloudStorageOptions // *StorageConfig_BigQueryOptions Type isStorageConfig_Type `protobuf_oneof:"type"` - TimespanConfig *StorageConfig_TimespanConfig `protobuf:"bytes,6,opt,name=timespan_config,json=timespanConfig" json:"timespan_config,omitempty"` + TimespanConfig *StorageConfig_TimespanConfig `protobuf:"bytes,6,opt,name=timespan_config,json=timespanConfig,proto3" json:"timespan_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1407,7 +1500,7 @@ func (m *StorageConfig) Reset() { *m = StorageConfig{} } func (m *StorageConfig) String() string { return proto.CompactTextString(m) } func (*StorageConfig) ProtoMessage() {} func (*StorageConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{9} + return fileDescriptor_storage_fb7f209e00c99d61, []int{9} } func (m *StorageConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StorageConfig.Unmarshal(m, b) @@ -1432,13 +1525,13 @@ type isStorageConfig_Type interface { } type StorageConfig_DatastoreOptions struct { - DatastoreOptions *DatastoreOptions `protobuf:"bytes,2,opt,name=datastore_options,json=datastoreOptions,oneof"` + DatastoreOptions *DatastoreOptions `protobuf:"bytes,2,opt,name=datastore_options,json=datastoreOptions,proto3,oneof"` } type StorageConfig_CloudStorageOptions struct { - CloudStorageOptions *CloudStorageOptions `protobuf:"bytes,3,opt,name=cloud_storage_options,json=cloudStorageOptions,oneof"` + CloudStorageOptions *CloudStorageOptions `protobuf:"bytes,3,opt,name=cloud_storage_options,json=cloudStorageOptions,proto3,oneof"` } type StorageConfig_BigQueryOptions struct { - BigQueryOptions *BigQueryOptions `protobuf:"bytes,4,opt,name=big_query_options,json=bigQueryOptions,oneof"` + BigQueryOptions *BigQueryOptions `protobuf:"bytes,4,opt,name=big_query_options,json=bigQueryOptions,proto3,oneof"` } func (*StorageConfig_DatastoreOptions) isStorageConfig_Type() {} @@ -1577,10 +1670,10 @@ func _StorageConfig_OneofSizer(msg proto.Message) (n int) { // Currently only supported when inspecting Google Cloud Storage and BigQuery. type StorageConfig_TimespanConfig struct { // Exclude files older than this value. - StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Exclude files newer than this value. // If set to zero, no upper time limit is applied. - EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Specification of the field containing the timestamp of scanned items. // Required for data sources like Datastore or BigQuery. // The valid data types of the timestamp field are: @@ -1588,12 +1681,12 @@ type StorageConfig_TimespanConfig struct { // for Datastore - timestamp. // Datastore entity will be scanned if the timestamp property does not exist // or its value is empty or invalid. - TimestampField *FieldId `protobuf:"bytes,3,opt,name=timestamp_field,json=timestampField" json:"timestamp_field,omitempty"` + TimestampField *FieldId `protobuf:"bytes,3,opt,name=timestamp_field,json=timestampField,proto3" json:"timestamp_field,omitempty"` // When the job is started by a JobTrigger we will automatically figure out // a valid start_time to avoid scanning files that have not been modified // since the last time the JobTrigger executed. This will be based on the // time of the execution of the last run of the JobTrigger. - EnableAutoPopulationOfTimespanConfig bool `protobuf:"varint,4,opt,name=enable_auto_population_of_timespan_config,json=enableAutoPopulationOfTimespanConfig" json:"enable_auto_population_of_timespan_config,omitempty"` + EnableAutoPopulationOfTimespanConfig bool `protobuf:"varint,4,opt,name=enable_auto_population_of_timespan_config,json=enableAutoPopulationOfTimespanConfig,proto3" json:"enable_auto_population_of_timespan_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1603,7 +1696,7 @@ func (m *StorageConfig_TimespanConfig) Reset() { *m = StorageConfig_Time func (m *StorageConfig_TimespanConfig) String() string { return proto.CompactTextString(m) } func (*StorageConfig_TimespanConfig) ProtoMessage() {} func (*StorageConfig_TimespanConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{9, 0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{9, 0} } func (m *StorageConfig_TimespanConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StorageConfig_TimespanConfig.Unmarshal(m, b) @@ -1654,10 +1747,10 @@ func (m *StorageConfig_TimespanConfig) GetEnableAutoPopulationOfTimespanConfig() // Row key for identifying a record in BigQuery table. type BigQueryKey struct { // Complete BigQuery table reference. - TableReference *BigQueryTable `protobuf:"bytes,1,opt,name=table_reference,json=tableReference" json:"table_reference,omitempty"` + TableReference *BigQueryTable `protobuf:"bytes,1,opt,name=table_reference,json=tableReference,proto3" json:"table_reference,omitempty"` // Absolute number of the row from the beginning of the table at the time // of scanning. - RowNumber int64 `protobuf:"varint,2,opt,name=row_number,json=rowNumber" json:"row_number,omitempty"` + RowNumber int64 `protobuf:"varint,2,opt,name=row_number,json=rowNumber,proto3" json:"row_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1667,7 +1760,7 @@ func (m *BigQueryKey) Reset() { *m = BigQueryKey{} } func (m *BigQueryKey) String() string { return proto.CompactTextString(m) } func (*BigQueryKey) ProtoMessage() {} func (*BigQueryKey) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{10} + return fileDescriptor_storage_fb7f209e00c99d61, []int{10} } func (m *BigQueryKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BigQueryKey.Unmarshal(m, b) @@ -1704,7 +1797,7 @@ func (m *BigQueryKey) GetRowNumber() int64 { // Record key for a finding in Cloud Datastore. type DatastoreKey struct { // Datastore entity key. - EntityKey *Key `protobuf:"bytes,1,opt,name=entity_key,json=entityKey" json:"entity_key,omitempty"` + EntityKey *Key `protobuf:"bytes,1,opt,name=entity_key,json=entityKey,proto3" json:"entity_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1714,7 +1807,7 @@ func (m *DatastoreKey) Reset() { *m = DatastoreKey{} } func (m *DatastoreKey) String() string { return proto.CompactTextString(m) } func (*DatastoreKey) ProtoMessage() {} func (*DatastoreKey) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{11} + return fileDescriptor_storage_fb7f209e00c99d61, []int{11} } func (m *DatastoreKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DatastoreKey.Unmarshal(m, b) @@ -1749,7 +1842,7 @@ type Key struct { // Entities are partitioned into subsets, currently identified by a project // ID and namespace ID. // Queries are scoped to a single partition. - PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId" json:"partition_id,omitempty"` + PartitionId *PartitionId `protobuf:"bytes,1,opt,name=partition_id,json=partitionId,proto3" json:"partition_id,omitempty"` // The entity path. // An entity path consists of one or more elements composed of a kind and a // string or numerical identifier, which identify entities. The first @@ -1759,7 +1852,7 @@ type Key struct { // the path are called the element's _ancestors_. // // A path can never be empty, and a path can have at most 100 elements. - Path []*Key_PathElement `protobuf:"bytes,2,rep,name=path" json:"path,omitempty"` + Path []*Key_PathElement `protobuf:"bytes,2,rep,name=path,proto3" json:"path,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1769,7 +1862,7 @@ func (m *Key) Reset() { *m = Key{} } func (m *Key) String() string { return proto.CompactTextString(m) } func (*Key) ProtoMessage() {} func (*Key) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{12} + return fileDescriptor_storage_fb7f209e00c99d61, []int{12} } func (m *Key) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Key.Unmarshal(m, b) @@ -1812,7 +1905,7 @@ type Key_PathElement struct { // A kind matching regex `__.*__` is reserved/read-only. // A kind must not contain more than 1500 bytes when UTF-8 encoded. // Cannot be `""`. - Kind string `protobuf:"bytes,1,opt,name=kind" json:"kind,omitempty"` + Kind string `protobuf:"bytes,1,opt,name=kind,proto3" json:"kind,omitempty"` // The type of ID. // // Types that are valid to be assigned to IdType: @@ -1828,7 +1921,7 @@ func (m *Key_PathElement) Reset() { *m = Key_PathElement{} } func (m *Key_PathElement) String() string { return proto.CompactTextString(m) } func (*Key_PathElement) ProtoMessage() {} func (*Key_PathElement) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{12, 0} + return fileDescriptor_storage_fb7f209e00c99d61, []int{12, 0} } func (m *Key_PathElement) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Key_PathElement.Unmarshal(m, b) @@ -1853,10 +1946,10 @@ type isKey_PathElement_IdType interface { } type Key_PathElement_Id struct { - Id int64 `protobuf:"varint,2,opt,name=id,oneof"` + Id int64 `protobuf:"varint,2,opt,name=id,proto3,oneof"` } type Key_PathElement_Name struct { - Name string `protobuf:"bytes,3,opt,name=name,oneof"` + Name string `protobuf:"bytes,3,opt,name=name,proto3,oneof"` } func (*Key_PathElement_Id) isKey_PathElement_IdType() {} @@ -1970,7 +2063,7 @@ func (m *RecordKey) Reset() { *m = RecordKey{} } func (m *RecordKey) String() string { return proto.CompactTextString(m) } func (*RecordKey) ProtoMessage() {} func (*RecordKey) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{13} + return fileDescriptor_storage_fb7f209e00c99d61, []int{13} } func (m *RecordKey) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_RecordKey.Unmarshal(m, b) @@ -1995,10 +2088,10 @@ type isRecordKey_Type interface { } type RecordKey_DatastoreKey struct { - DatastoreKey *DatastoreKey `protobuf:"bytes,2,opt,name=datastore_key,json=datastoreKey,oneof"` + DatastoreKey *DatastoreKey `protobuf:"bytes,2,opt,name=datastore_key,json=datastoreKey,proto3,oneof"` } type RecordKey_BigQueryKey struct { - BigQueryKey *BigQueryKey `protobuf:"bytes,3,opt,name=big_query_key,json=bigQueryKey,oneof"` + BigQueryKey *BigQueryKey `protobuf:"bytes,3,opt,name=big_query_key,json=bigQueryKey,proto3,oneof"` } func (*RecordKey_DatastoreKey) isRecordKey_Type() {} @@ -2107,11 +2200,11 @@ func _RecordKey_OneofSizer(msg proto.Message) (n int) { type BigQueryTable struct { // The Google Cloud Platform project ID of the project containing the table. // If omitted, project ID is inferred from the API call. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Dataset ID of the table. - DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId" json:"dataset_id,omitempty"` + DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"` // Name of the table. - TableId string `protobuf:"bytes,3,opt,name=table_id,json=tableId" json:"table_id,omitempty"` + TableId string `protobuf:"bytes,3,opt,name=table_id,json=tableId,proto3" json:"table_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2121,7 +2214,7 @@ func (m *BigQueryTable) Reset() { *m = BigQueryTable{} } func (m *BigQueryTable) String() string { return proto.CompactTextString(m) } func (*BigQueryTable) ProtoMessage() {} func (*BigQueryTable) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{14} + return fileDescriptor_storage_fb7f209e00c99d61, []int{14} } func (m *BigQueryTable) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_BigQueryTable.Unmarshal(m, b) @@ -2169,7 +2262,7 @@ func (m *BigQueryTable) GetTableId() string { // into account that multiple rows correspond to the same entity. type EntityId struct { // Composite key indicating which field contains the entity identifier. - Field *FieldId `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field *FieldId `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2179,7 +2272,7 @@ func (m *EntityId) Reset() { *m = EntityId{} } func (m *EntityId) String() string { return proto.CompactTextString(m) } func (*EntityId) ProtoMessage() {} func (*EntityId) Descriptor() ([]byte, []int) { - return fileDescriptor_storage_38ffc711e7cad19e, []int{15} + return fileDescriptor_storage_fb7f209e00c99d61, []int{15} } func (m *EntityId) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_EntityId.Unmarshal(m, b) @@ -2236,115 +2329,124 @@ func init() { proto.RegisterType((*EntityId)(nil), "google.privacy.dlp.v2.EntityId") proto.RegisterEnum("google.privacy.dlp.v2.Likelihood", Likelihood_name, Likelihood_value) proto.RegisterEnum("google.privacy.dlp.v2.FileType", FileType_name, FileType_value) + proto.RegisterEnum("google.privacy.dlp.v2.CloudStorageOptions_SampleMethod", CloudStorageOptions_SampleMethod_name, CloudStorageOptions_SampleMethod_value) + proto.RegisterEnum("google.privacy.dlp.v2.BigQueryOptions_SampleMethod", BigQueryOptions_SampleMethod_name, BigQueryOptions_SampleMethod_value) } func init() { - proto.RegisterFile("google/privacy/dlp/v2/storage.proto", fileDescriptor_storage_38ffc711e7cad19e) + proto.RegisterFile("google/privacy/dlp/v2/storage.proto", fileDescriptor_storage_fb7f209e00c99d61) } -var fileDescriptor_storage_38ffc711e7cad19e = []byte{ - // 1646 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x58, 0xfd, 0x72, 0x1b, 0x49, - 0x11, 0xd7, 0x4a, 0x96, 0xad, 0x6d, 0x59, 0x1f, 0x9e, 0x38, 0x94, 0x22, 0xc8, 0x9d, 0xa3, 0x84, - 0x10, 0x42, 0x95, 0x04, 0x3e, 0xa8, 0x22, 0x7c, 0x5b, 0x91, 0x72, 0x5a, 0xec, 0xb3, 0xc5, 0x58, - 0x97, 0x5c, 0xc0, 0x55, 0xcb, 0x4a, 0x33, 0x92, 0xe7, 0xb2, 0xda, 0x59, 0x76, 0x47, 0x71, 0x54, - 0x14, 0xcf, 0xc0, 0x3f, 0x3c, 0x02, 0x55, 0x14, 0x14, 0x4f, 0x00, 0x6f, 0x40, 0x15, 0x3c, 0x02, - 0x2f, 0xc0, 0x7f, 0x3c, 0x01, 0x35, 0x1f, 0xbb, 0xfa, 0xc0, 0xba, 0xf8, 0x52, 0xf9, 0xcb, 0xd3, - 0xbd, 0xfd, 0xeb, 0x99, 0xe9, 0xfe, 0x75, 0xf7, 0x58, 0x70, 0x7f, 0xc2, 0xf9, 0xc4, 0xa7, 0xad, - 0x30, 0x62, 0xaf, 0xbd, 0xd1, 0xbc, 0x45, 0xfc, 0xb0, 0xf5, 0xfa, 0xb0, 0x15, 0x0b, 0x1e, 0x79, - 0x13, 0xda, 0x0c, 0x23, 0x2e, 0x38, 0xba, 0xad, 0x8d, 0x9a, 0xc6, 0xa8, 0x49, 0xfc, 0xb0, 0xf9, - 0xfa, 0xb0, 0xfe, 0x35, 0x83, 0xf5, 0x42, 0xd6, 0xf2, 0x82, 0x80, 0x0b, 0x4f, 0x30, 0x1e, 0xc4, - 0x1a, 0x54, 0xff, 0x30, 0xf5, 0xcc, 0x05, 0x1f, 0xce, 0xc6, 0x2d, 0xc1, 0xa6, 0x34, 0x16, 0xde, - 0x34, 0xd4, 0x06, 0x8d, 0x0f, 0xa0, 0xe0, 0x04, 0x63, 0x3e, 0x98, 0x87, 0x14, 0x21, 0xd8, 0x0a, - 0xbc, 0x29, 0xad, 0x59, 0x07, 0xd6, 0x23, 0x1b, 0xab, 0x75, 0xe3, 0x9f, 0x45, 0x28, 0x3f, 0x9d, - 0xc5, 0x82, 0x4f, 0x53, 0xb3, 0x1f, 0x81, 0xcd, 0x82, 0x31, 0x77, 0xc5, 0x3c, 0xd4, 0xb6, 0xc5, - 0xc3, 0x0f, 0x9b, 0xd7, 0x1e, 0xae, 0x99, 0x60, 0x70, 0x81, 0x25, 0xe8, 0x23, 0x00, 0x9f, 0xbd, - 0xa2, 0x3e, 0xbb, 0xe4, 0x9c, 0xd4, 0xb6, 0x0f, 0xac, 0x47, 0xe5, 0xc3, 0x7b, 0x1b, 0xe0, 0x27, - 0xa9, 0x21, 0x5e, 0x02, 0x21, 0x0c, 0x40, 0xd8, 0x48, 0x5e, 0xd3, 0x8b, 0xe6, 0xb5, 0xac, 0x3a, - 0xc1, 0xb7, 0x37, 0xb8, 0x58, 0x3d, 0x7b, 0xb3, 0x93, 0xe2, 0x7a, 0x19, 0xbc, 0xe4, 0x05, 0x3d, - 0x85, 0x7c, 0x44, 0x27, 0xf4, 0x4d, 0x2d, 0xa7, 0xdc, 0x7d, 0xeb, 0x66, 0xee, 0xb0, 0x84, 0xf4, - 0x32, 0x58, 0x63, 0xd1, 0x05, 0x94, 0xe3, 0x59, 0x14, 0xf1, 0x89, 0x27, 0xa8, 0x0e, 0xcf, 0x96, - 0xf2, 0xf6, 0xd1, 0xcd, 0xbc, 0x9d, 0x27, 0x58, 0x29, 0xf5, 0x32, 0xb8, 0x14, 0x2f, 0x2b, 0xd0, - 0x05, 0x54, 0x08, 0x15, 0x54, 0x1d, 0xd9, 0x8d, 0x66, 0x3e, 0x8d, 0x6b, 0x3b, 0x07, 0xb9, 0x9b, - 0xbb, 0xef, 0x24, 0x60, 0x3c, 0xf3, 0x29, 0x2e, 0x93, 0x65, 0x31, 0xae, 0xff, 0xc7, 0x02, 0x58, - 0x44, 0x07, 0x7d, 0x06, 0xf6, 0x15, 0x8f, 0x88, 0xeb, 0xb3, 0x58, 0x98, 0x24, 0x3f, 0xf9, 0xb2, - 0x21, 0x6e, 0xbe, 0xe0, 0x11, 0x39, 0x61, 0xb1, 0xe8, 0x65, 0x70, 0xe1, 0xca, 0xac, 0xd1, 0x0b, - 0x40, 0x23, 0x9f, 0xcf, 0x88, 0x6b, 0xe8, 0xed, 0x86, 0x9e, 0xb8, 0x34, 0x61, 0xff, 0xc6, 0xa6, - 0x2d, 0x24, 0xe0, 0x5c, 0xdb, 0xf7, 0x3d, 0x71, 0xd9, 0xcb, 0xe0, 0xea, 0x68, 0x4d, 0x57, 0x3f, - 0x80, 0x42, 0xb2, 0x21, 0xda, 0x87, 0xbc, 0xdc, 0x30, 0xae, 0x59, 0x07, 0xb9, 0x47, 0x36, 0xd6, - 0x42, 0xbb, 0x00, 0xdb, 0x31, 0x9f, 0x45, 0x23, 0x5a, 0xbf, 0x07, 0x79, 0x95, 0x3b, 0x54, 0x83, - 0x9d, 0xd0, 0x13, 0x82, 0x46, 0x81, 0xa1, 0x7d, 0x22, 0xd6, 0x2b, 0x50, 0x5a, 0x49, 0x48, 0xfd, - 0x6f, 0x79, 0x28, 0xad, 0xc4, 0x10, 0x11, 0xd8, 0xbd, 0xe4, 0x42, 0xc5, 0x49, 0xe6, 0xc3, 0xc4, - 0xe9, 0xa7, 0xef, 0x90, 0x8e, 0x66, 0x4f, 0xfb, 0x91, 0xeb, 0x5e, 0x06, 0x17, 0x2f, 0x17, 0x62, - 0xfd, 0x1c, 0xec, 0x7e, 0xc4, 0xdf, 0xb0, 0x29, 0x13, 0x73, 0x74, 0x1f, 0x4a, 0x57, 0x2c, 0x20, - 0xfc, 0xca, 0x1d, 0xd2, 0x31, 0x8f, 0xf4, 0x9e, 0x79, 0xbc, 0xab, 0x95, 0x6d, 0xa5, 0x43, 0xf7, - 0xc0, 0xc8, 0xae, 0x37, 0x16, 0x34, 0x52, 0x25, 0x92, 0xc7, 0x45, 0xad, 0x3b, 0x92, 0xaa, 0xfa, - 0x9f, 0x2c, 0xd8, 0x5f, 0x94, 0xd7, 0x11, 0xf9, 0x7c, 0x16, 0x8b, 0x29, 0x0d, 0x04, 0x3a, 0x85, - 0xea, 0x98, 0xbd, 0xa1, 0x32, 0xf3, 0x69, 0x95, 0x5a, 0x37, 0xac, 0xd2, 0x5e, 0x06, 0x57, 0x14, - 0x78, 0xa1, 0x42, 0xdf, 0x81, 0x5b, 0x11, 0xf5, 0x3d, 0xc1, 0x5e, 0xd3, 0x65, 0x97, 0xea, 0x48, - 0xbd, 0x0c, 0x46, 0xc9, 0xc7, 0x05, 0xa4, 0xbd, 0x0b, 0xe0, 0xa5, 0x07, 0xaa, 0xff, 0x3d, 0x0b, - 0xc5, 0xa5, 0xe8, 0xa0, 0x3e, 0x94, 0xd2, 0xa0, 0xab, 0x8a, 0xb5, 0xbe, 0x74, 0xc5, 0xe2, 0x24, - 0x6d, 0x9a, 0x03, 0xbf, 0x02, 0x3b, 0x4c, 0x02, 0x6c, 0xda, 0xc9, 0x8f, 0xdf, 0x25, 0x87, 0x69, - 0x96, 0xf0, 0xc2, 0x1f, 0xfa, 0x1d, 0xdc, 0x5e, 0x5c, 0xdb, 0x5d, 0xdc, 0xcb, 0x30, 0xbe, 0xf7, - 0x2e, 0x1b, 0x5d, 0x97, 0x38, 0xbc, 0xef, 0x5f, 0xa3, 0x6d, 0x6f, 0xc3, 0x96, 0x6c, 0x44, 0xc9, - 0xdf, 0xc6, 0x5d, 0xd8, 0x79, 0xc6, 0xa8, 0x4f, 0x1c, 0x72, 0x6d, 0xbb, 0x3f, 0x83, 0x62, 0xdf, - 0x8b, 0x04, 0x93, 0x5b, 0x39, 0x04, 0xdd, 0x05, 0x08, 0x23, 0xfe, 0x39, 0x1d, 0x09, 0x97, 0xe9, - 0x9c, 0xd9, 0xea, 0x6e, 0x52, 0xe3, 0x10, 0xc9, 0x33, 0x89, 0x8a, 0x43, 0x6f, 0x44, 0xa5, 0xc1, - 0x96, 0x32, 0x28, 0xa6, 0x3a, 0x87, 0x34, 0x1e, 0x40, 0xf9, 0x98, 0x05, 0xa4, 0xfb, 0x26, 0x8c, - 0x68, 0x1c, 0x33, 0x1e, 0x5c, 0xbb, 0xed, 0x1f, 0x2c, 0xa8, 0x76, 0x3c, 0xe1, 0xc9, 0x96, 0x40, - 0xcf, 0x42, 0x35, 0xc1, 0x50, 0x17, 0x76, 0xc3, 0xe4, 0x2c, 0xd2, 0xbb, 0xce, 0x73, 0x63, 0x43, - 0xc0, 0x96, 0x8e, 0x8d, 0x8b, 0xe1, 0xd2, 0x1d, 0x9e, 0xc0, 0xd6, 0x2b, 0x16, 0x10, 0x93, 0xd8, - 0xaf, 0x6f, 0x80, 0xaf, 0x1e, 0x12, 0x2b, 0x48, 0xe3, 0xbf, 0x16, 0xdc, 0x5a, 0x6e, 0x3d, 0xc9, - 0xc9, 0x3e, 0x81, 0xc2, 0x98, 0xf9, 0xd4, 0x8d, 0x69, 0xd2, 0x1b, 0x0f, 0x6f, 0xd0, 0xb8, 0x0c, - 0xba, 0xf9, 0x8c, 0xf9, 0xf4, 0x9c, 0x0a, 0xbc, 0x33, 0xd6, 0x0b, 0xd4, 0x82, 0xfd, 0xe1, 0x5c, - 0xd0, 0xd8, 0xf5, 0x25, 0x65, 0xdc, 0x90, 0x46, 0xae, 0xfc, 0xa4, 0xc2, 0x99, 0xc3, 0x7b, 0xea, - 0xdb, 0x89, 0xfc, 0xd4, 0xa7, 0x91, 0x04, 0xa3, 0x9f, 0x00, 0xa8, 0xfd, 0x65, 0x46, 0xe3, 0x5a, - 0xfe, 0x20, 0xf7, 0xa8, 0xbc, 0x71, 0x04, 0x4b, 0x80, 0x1a, 0xc1, 0xf6, 0xd8, 0xac, 0xe2, 0xfa, - 0x57, 0x25, 0x09, 0xf4, 0xde, 0x55, 0xc8, 0xcd, 0x22, 0xdf, 0x24, 0x43, 0x2e, 0x1b, 0x0f, 0xa1, - 0xba, 0xde, 0x6e, 0x65, 0xce, 0x54, 0x97, 0x36, 0x39, 0x93, 0xeb, 0xc6, 0xbf, 0x2c, 0xa8, 0xb4, - 0xd9, 0xe4, 0x17, 0x33, 0x1a, 0xcd, 0x17, 0x81, 0xa9, 0x08, 0x6f, 0xe8, 0x53, 0x37, 0xa2, 0x63, - 0x1a, 0xd1, 0x60, 0x94, 0xf4, 0xc4, 0x07, 0x1b, 0x4e, 0x97, 0x38, 0x18, 0x48, 0x14, 0x2e, 0x2b, - 0x30, 0x4e, 0xb0, 0xe8, 0x13, 0x40, 0x8c, 0xd0, 0x40, 0xb0, 0xf1, 0x9c, 0x05, 0x13, 0x77, 0x2c, - 0x89, 0x1b, 0xd7, 0xb2, 0x6a, 0xe8, 0x7d, 0xb0, 0xf1, 0xbe, 0x8a, 0xdd, 0x78, 0x6f, 0x09, 0xa9, - 0x74, 0xb1, 0x64, 0x73, 0xc4, 0xaf, 0x4c, 0x98, 0x55, 0xfd, 0xe5, 0xb0, 0x2d, 0x35, 0x2a, 0xb8, - 0x8d, 0xbf, 0xe6, 0xa1, 0x64, 0x2e, 0xfd, 0x94, 0x07, 0x63, 0x36, 0x41, 0xcf, 0x61, 0x8f, 0x24, - 0xac, 0x74, 0xb9, 0xbe, 0xa3, 0xe1, 0xd1, 0xa6, 0x49, 0xb5, 0xce, 0x62, 0x39, 0xa9, 0xc8, 0x3a, - 0xb3, 0x7f, 0x0d, 0xb7, 0x57, 0x47, 0x60, 0xe2, 0x5b, 0xf7, 0x84, 0xc7, 0x37, 0x27, 0x53, 0x2f, - 0x83, 0x6f, 0x8d, 0xae, 0x61, 0xe8, 0x00, 0xf6, 0x86, 0x6c, 0xe2, 0xfe, 0x46, 0xc6, 0x36, 0xf5, - 0xae, 0x1f, 0x23, 0x0f, 0xdf, 0x92, 0x8a, 0x85, 0xe7, 0xca, 0x70, 0x2d, 0xbd, 0x17, 0x50, 0x51, - 0xef, 0xc7, 0xd0, 0x0b, 0xdc, 0x91, 0x0a, 0x91, 0x7a, 0xc0, 0x6d, 0x7e, 0x81, 0xac, 0x84, 0xb3, - 0x39, 0x30, 0x58, 0x2d, 0xe2, 0xb2, 0x58, 0x91, 0xeb, 0x7f, 0xcc, 0x42, 0x79, 0xd5, 0x04, 0x3d, - 0x01, 0x88, 0x85, 0x17, 0x09, 0x57, 0x9a, 0x1a, 0x2a, 0xd5, 0x17, 0x7b, 0xe9, 0x37, 0xad, 0xf6, - 0x2b, 0xdf, 0xb4, 0xd8, 0x56, 0xd6, 0x52, 0x46, 0xdf, 0x83, 0x02, 0x0d, 0x88, 0x06, 0x66, 0xdf, - 0x0a, 0xdc, 0xa1, 0x01, 0x51, 0xb0, 0x8f, 0xcd, 0x15, 0xa5, 0x56, 0x13, 0xce, 0x24, 0xe5, 0x6d, - 0x7c, 0x2b, 0xa7, 0x30, 0xa5, 0x41, 0x2f, 0xe0, 0x9b, 0x34, 0x50, 0xb5, 0xe0, 0xcd, 0x04, 0x77, - 0x43, 0x1e, 0xce, 0x7c, 0xf5, 0x34, 0x77, 0xf9, 0xd8, 0x5d, 0x8f, 0xa2, 0xcc, 0x4c, 0x01, 0x3f, - 0xd0, 0x80, 0xa3, 0x99, 0xe0, 0xfd, 0xd4, 0xfc, 0x6c, 0xbc, 0x1a, 0x93, 0xb4, 0x93, 0xff, 0x16, - 0x8a, 0x49, 0xca, 0x8e, 0xe9, 0xfc, 0x7d, 0x97, 0x9e, 0xae, 0x15, 0x37, 0x98, 0x4d, 0x87, 0xe6, - 0x01, 0xa1, 0x6b, 0xe5, 0x54, 0x29, 0x1a, 0x0e, 0xec, 0xa6, 0x4c, 0x97, 0xbb, 0x3f, 0x01, 0x90, - 0xd5, 0x26, 0xe6, 0xee, 0x2b, 0x3a, 0xff, 0xff, 0x44, 0xad, 0xb6, 0x5a, 0x3a, 0xc7, 0xb6, 0xb6, - 0x3e, 0xa6, 0xf3, 0xc6, 0xbf, 0x2d, 0xc8, 0x49, 0x17, 0xef, 0xa9, 0xdd, 0xff, 0xc0, 0xb4, 0x2a, - 0xdd, 0x25, 0x1e, 0x6e, 0x3e, 0x43, 0x53, 0x76, 0xb6, 0xae, 0x4f, 0xd5, 0xf0, 0x54, 0x98, 0xfa, - 0x40, 0x4e, 0xbf, 0x54, 0x29, 0xbb, 0x9e, 0x9a, 0x1c, 0xa6, 0xeb, 0xc9, 0x35, 0xaa, 0x42, 0xd6, - 0x4c, 0xc2, 0x5c, 0x2f, 0x83, 0xb3, 0x8c, 0xa0, 0x7d, 0x33, 0xcf, 0x24, 0x4d, 0xec, 0x5e, 0x46, - 0x4f, 0xb4, 0xb6, 0x0d, 0x3b, 0x8c, 0xa8, 0x06, 0xdd, 0xf8, 0xb3, 0x05, 0x36, 0xa6, 0x23, 0x1e, - 0x11, 0x79, 0xcd, 0x9f, 0x43, 0x69, 0xd1, 0x53, 0x64, 0xb0, 0x34, 0x39, 0xef, 0xbf, 0xad, 0x9f, - 0x1c, 0x53, 0xf9, 0x2f, 0xcb, 0x2e, 0x59, 0x8e, 0x7a, 0x0f, 0x4a, 0x8b, 0x2a, 0x97, 0xbe, 0x72, - 0x5f, 0x18, 0xb3, 0x25, 0xba, 0xc8, 0x37, 0xe6, 0x70, 0x21, 0xa6, 0xa4, 0xba, 0x84, 0xd2, 0x0a, - 0x2f, 0xd6, 0x5e, 0x00, 0xd6, 0xfa, 0x0b, 0xe0, 0x2e, 0x80, 0x3a, 0x11, 0x5d, 0x7e, 0x20, 0x18, - 0x8d, 0x43, 0xd0, 0x1d, 0x28, 0x68, 0x52, 0x32, 0x5d, 0x46, 0x36, 0xde, 0x51, 0xb2, 0x43, 0x1a, - 0x3f, 0x83, 0x42, 0x57, 0x71, 0xc0, 0x21, 0xe8, 0xbb, 0x90, 0xd7, 0xa5, 0x66, 0xdd, 0xa8, 0xd4, - 0xb4, 0xf1, 0x63, 0x01, 0xb0, 0xf4, 0xce, 0xac, 0xc3, 0x57, 0x4e, 0x9c, 0xe3, 0xee, 0x89, 0xd3, - 0x3b, 0x3b, 0xeb, 0xb8, 0x9f, 0x9e, 0x9e, 0xf7, 0xbb, 0x4f, 0x9d, 0x67, 0x4e, 0xb7, 0x53, 0xcd, - 0xa0, 0x3d, 0x28, 0x3d, 0xef, 0xe2, 0x97, 0xee, 0xa7, 0xa7, 0xca, 0xe4, 0x65, 0xd5, 0x42, 0xbb, - 0x50, 0x48, 0xa5, 0xac, 0x94, 0xfa, 0x67, 0xe7, 0xe7, 0x4e, 0xfb, 0xa4, 0x5b, 0xcd, 0x21, 0x80, - 0x6d, 0xf3, 0x65, 0x0b, 0x55, 0xa0, 0xa8, 0xa0, 0x46, 0x91, 0x7f, 0xdc, 0x85, 0x42, 0x32, 0x52, - 0xd1, 0x1d, 0xb8, 0xfd, 0xcc, 0x39, 0xe9, 0xba, 0x83, 0x97, 0xfd, 0xee, 0xda, 0x96, 0x15, 0x28, - 0xb6, 0x9d, 0xd3, 0x23, 0xfc, 0xd2, 0x95, 0x16, 0x55, 0x0b, 0x95, 0xc0, 0x1e, 0x74, 0x3f, 0x1b, - 0x68, 0x31, 0xdb, 0xfe, 0xbd, 0x05, 0x77, 0x46, 0x7c, 0x7a, 0xfd, 0x4d, 0xdb, 0xd0, 0xf1, 0xc3, - 0x64, 0xfe, 0x5a, 0xbf, 0xfc, 0xbe, 0x31, 0x9a, 0x70, 0xdf, 0x0b, 0x26, 0x4d, 0x1e, 0x4d, 0x5a, - 0x13, 0x1a, 0xa8, 0x2e, 0xd6, 0xd2, 0x9f, 0xbc, 0x90, 0xc5, 0x6b, 0xbf, 0x1e, 0xfc, 0x90, 0xf8, - 0xe1, 0x5f, 0xb2, 0xb7, 0x3e, 0xd6, 0x50, 0x35, 0x3a, 0x9a, 0x1d, 0x3f, 0x6c, 0x3e, 0x3f, 0xfc, - 0x47, 0xa2, 0xbd, 0x50, 0xda, 0x8b, 0x8e, 0x1f, 0x5e, 0x3c, 0x3f, 0x1c, 0x6e, 0x2b, 0x87, 0x1f, - 0xfd, 0x2f, 0x00, 0x00, 0xff, 0xff, 0x2c, 0x19, 0xec, 0xdb, 0x8d, 0x10, 0x00, 0x00, +var fileDescriptor_storage_fb7f209e00c99d61 = []byte{ + // 1753 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x58, 0xfd, 0x92, 0x1a, 0x4b, + 0x15, 0x67, 0x60, 0x59, 0xe0, 0xf0, 0xb1, 0x6c, 0xef, 0xc6, 0x22, 0x68, 0xee, 0xdd, 0x90, 0x18, + 0x63, 0xac, 0x02, 0xdd, 0x68, 0x69, 0xfc, 0x5e, 0x02, 0xb9, 0x33, 0xee, 0x17, 0x36, 0xdc, 0x7c, + 0xe8, 0x56, 0x8d, 0x03, 0xd3, 0xb0, 0x7d, 0x33, 0x4c, 0x8f, 0x33, 0x4d, 0x36, 0x94, 0xe5, 0x33, + 0xf8, 0x8f, 0x8f, 0x60, 0x95, 0xa5, 0xe5, 0x13, 0xe8, 0x1b, 0xdc, 0x2a, 0x5f, 0xc1, 0x17, 0xf0, + 0x3f, 0x9f, 0xc0, 0xea, 0xee, 0x99, 0x61, 0x20, 0xcb, 0x5d, 0x92, 0xca, 0x5f, 0x3b, 0xe7, 0xf4, + 0xf9, 0x9d, 0x3e, 0x7d, 0xbe, 0x59, 0xb8, 0x37, 0x61, 0x6c, 0xe2, 0x90, 0x96, 0xe7, 0xd3, 0x37, + 0xd6, 0x68, 0xde, 0xb2, 0x1d, 0xaf, 0xf5, 0xe6, 0xb0, 0x15, 0x70, 0xe6, 0x5b, 0x13, 0xd2, 0xf4, + 0x7c, 0xc6, 0x19, 0xba, 0xa5, 0x84, 0x9a, 0xa1, 0x50, 0xd3, 0x76, 0xbc, 0xe6, 0x9b, 0xc3, 0xfa, + 0x37, 0x42, 0xac, 0xe5, 0xd1, 0x96, 0xe5, 0xba, 0x8c, 0x5b, 0x9c, 0x32, 0x37, 0x50, 0xa0, 0xfa, + 0xa7, 0xb1, 0x66, 0xc6, 0xd9, 0x70, 0x36, 0x6e, 0x71, 0x3a, 0x25, 0x01, 0xb7, 0xa6, 0x9e, 0x12, + 0x68, 0x7c, 0x02, 0x79, 0xc3, 0x1d, 0xb3, 0xc1, 0xdc, 0x23, 0x08, 0xc1, 0x96, 0x6b, 0x4d, 0x49, + 0x4d, 0x3b, 0xd0, 0x1e, 0x16, 0xb0, 0xfc, 0x6e, 0xfc, 0xbb, 0x08, 0x95, 0xa7, 0xb3, 0x80, 0xb3, + 0x69, 0x2c, 0xf6, 0x53, 0x28, 0x50, 0x77, 0xcc, 0x4c, 0x3e, 0xf7, 0x94, 0x6c, 0xf1, 0xf0, 0xd3, + 0xe6, 0xb5, 0xc6, 0x35, 0x23, 0x0c, 0xce, 0xd3, 0x08, 0x7d, 0x04, 0xe0, 0xd0, 0xd7, 0xc4, 0xa1, + 0x97, 0x8c, 0xd9, 0xb5, 0xed, 0x03, 0xed, 0x61, 0xe5, 0xf0, 0xee, 0x1a, 0xf8, 0x49, 0x2c, 0x88, + 0x13, 0x20, 0x84, 0x01, 0x6c, 0x3a, 0x12, 0xcf, 0xb4, 0xfc, 0x79, 0x2d, 0x2d, 0x2d, 0xf8, 0xee, + 0x1a, 0x15, 0xcb, 0xb6, 0x37, 0x3b, 0x31, 0x4e, 0x4f, 0xe1, 0x84, 0x16, 0xf4, 0x14, 0xb2, 0x3e, + 0x99, 0x90, 0xb7, 0xb5, 0x8c, 0x54, 0xf7, 0x9d, 0xcd, 0xd4, 0x61, 0x01, 0xd1, 0x53, 0x58, 0x61, + 0xd1, 0x05, 0x54, 0x82, 0x99, 0xef, 0xb3, 0x89, 0xc5, 0x89, 0x72, 0xcf, 0x96, 0xd4, 0xf6, 0x78, + 0x33, 0x6d, 0xfd, 0x08, 0x2b, 0x28, 0x3d, 0x85, 0xcb, 0x41, 0x92, 0x81, 0x2e, 0x60, 0xc7, 0x26, + 0x9c, 0x48, 0x93, 0x4d, 0x7f, 0xe6, 0x90, 0xa0, 0x96, 0x3b, 0xc8, 0x6c, 0xae, 0xbe, 0x13, 0x81, + 0xf1, 0xcc, 0x21, 0xb8, 0x62, 0x27, 0xc9, 0xa0, 0xfe, 0x5f, 0x0d, 0x60, 0xe1, 0x1d, 0xf4, 0x12, + 0x0a, 0x57, 0xcc, 0xb7, 0x4d, 0x87, 0x06, 0x3c, 0x0c, 0xf2, 0x93, 0xf7, 0x75, 0x71, 0xf3, 0x05, + 0xf3, 0xed, 0x13, 0x1a, 0x70, 0x3d, 0x85, 0xf3, 0x57, 0xe1, 0x37, 0x7a, 0x01, 0x68, 0xe4, 0xb0, + 0x99, 0x6d, 0x86, 0xe9, 0x6d, 0x7a, 0x16, 0xbf, 0x0c, 0xdd, 0xfe, 0xad, 0x75, 0x57, 0x08, 0x40, + 0x5f, 0xc9, 0xf7, 0x2c, 0x7e, 0xa9, 0xa7, 0x70, 0x75, 0xb4, 0xc2, 0xab, 0x1f, 0x40, 0x3e, 0xba, + 0x10, 0xed, 0x43, 0x56, 0x5c, 0x18, 0xd4, 0xb4, 0x83, 0xcc, 0xc3, 0x02, 0x56, 0x44, 0x3b, 0x0f, + 0xdb, 0x01, 0x9b, 0xf9, 0x23, 0x52, 0xbf, 0x0b, 0x59, 0x19, 0x3b, 0x54, 0x83, 0x9c, 0x67, 0x71, + 0x4e, 0x7c, 0x37, 0x4c, 0xfb, 0x88, 0xac, 0xef, 0x40, 0x79, 0x29, 0x20, 0xf5, 0x7f, 0x66, 0xa1, + 0xbc, 0xe4, 0x43, 0x64, 0x43, 0xe9, 0x92, 0x71, 0xe9, 0x27, 0x11, 0x8f, 0xd0, 0x4f, 0xbf, 0xf8, + 0x80, 0x70, 0x34, 0x75, 0xa5, 0x47, 0x7c, 0xeb, 0x29, 0x5c, 0xbc, 0x5c, 0x90, 0xf5, 0x3e, 0x14, + 0x7a, 0x3e, 0x7b, 0x4b, 0xa7, 0x94, 0xcf, 0xd1, 0x3d, 0x28, 0x5f, 0x51, 0xd7, 0x66, 0x57, 0xe6, + 0x90, 0x8c, 0x99, 0xaf, 0xee, 0xcc, 0xe2, 0x92, 0x62, 0xb6, 0x25, 0x0f, 0xdd, 0x85, 0x90, 0x36, + 0xad, 0x31, 0x27, 0xbe, 0x2c, 0x91, 0x2c, 0x2e, 0x2a, 0xde, 0x91, 0x60, 0xd5, 0xff, 0xaa, 0xc1, + 0xfe, 0xa2, 0xbc, 0x8e, 0xec, 0x2f, 0x66, 0x01, 0x9f, 0x12, 0x97, 0xa3, 0x33, 0xa8, 0x8e, 0xe9, + 0x5b, 0x22, 0x22, 0x1f, 0x57, 0xa9, 0xb6, 0x61, 0x95, 0xea, 0x29, 0xbc, 0x23, 0xc1, 0x0b, 0x16, + 0xfa, 0x1e, 0xec, 0xf9, 0xc4, 0xb1, 0x38, 0x7d, 0x43, 0x92, 0x2a, 0xa5, 0x49, 0x7a, 0x0a, 0xa3, + 0xe8, 0x70, 0x01, 0x69, 0x97, 0x00, 0xac, 0xd8, 0xa0, 0xfa, 0xbf, 0xd2, 0x50, 0x4c, 0x78, 0x07, + 0xf5, 0xa0, 0x1c, 0x3b, 0x5d, 0x56, 0xac, 0xf6, 0xde, 0x15, 0x8b, 0xa3, 0xb0, 0xa9, 0x1c, 0xf8, + 0x2d, 0x14, 0xbc, 0xc8, 0xc1, 0x61, 0x3b, 0xf9, 0xd9, 0x87, 0xc4, 0x30, 0x8e, 0x12, 0x5e, 0xe8, + 0x43, 0x7f, 0x84, 0x5b, 0x8b, 0x67, 0x9b, 0x8b, 0x77, 0x85, 0x19, 0xaf, 0x7f, 0xc8, 0x45, 0xd7, + 0x05, 0x0e, 0xef, 0x3b, 0xd7, 0x70, 0xdb, 0xdb, 0xb0, 0x25, 0x1a, 0x51, 0xf4, 0xb7, 0x71, 0x07, + 0x72, 0xcf, 0x28, 0x71, 0x6c, 0xc3, 0xbe, 0xb6, 0xdd, 0x9f, 0x43, 0xb1, 0x67, 0xf9, 0x9c, 0x8a, + 0xab, 0x0c, 0x1b, 0xdd, 0x01, 0xf0, 0x7c, 0xf6, 0x05, 0x19, 0x71, 0x93, 0xaa, 0x98, 0x15, 0xe4, + 0xdb, 0x04, 0xc7, 0xb0, 0x45, 0x9e, 0x09, 0x54, 0xe0, 0x59, 0x23, 0x22, 0x04, 0xb6, 0xa4, 0x40, + 0x31, 0xe6, 0x19, 0x76, 0xe3, 0x3e, 0x54, 0x8e, 0xa9, 0x6b, 0x77, 0xdf, 0x7a, 0x3e, 0x09, 0x02, + 0xca, 0xdc, 0x6b, 0xaf, 0xfd, 0xb3, 0x06, 0xd5, 0x8e, 0xc5, 0x2d, 0xd1, 0x12, 0xc8, 0xb9, 0x27, + 0x27, 0x18, 0xea, 0x42, 0xc9, 0x8b, 0x6c, 0x11, 0xda, 0x55, 0x9c, 0x1b, 0x6b, 0x1c, 0x96, 0x30, + 0x1b, 0x17, 0xbd, 0xc4, 0x1b, 0x9e, 0xc0, 0xd6, 0x6b, 0xea, 0xda, 0x61, 0x60, 0xbf, 0xb9, 0x06, + 0xbe, 0x6c, 0x24, 0x96, 0x90, 0xc6, 0x97, 0x19, 0xd8, 0x4b, 0xb6, 0x9e, 0xc8, 0xb2, 0x53, 0xc8, + 0x8f, 0xa9, 0x43, 0xcc, 0x80, 0x44, 0xbd, 0xf1, 0x70, 0x83, 0xc6, 0x15, 0xa2, 0x9b, 0xcf, 0xa8, + 0x43, 0xfa, 0x84, 0xe3, 0xdc, 0x58, 0x7d, 0xa0, 0x16, 0xec, 0x0f, 0xe7, 0x9c, 0x04, 0xa6, 0x23, + 0x52, 0xc6, 0xf4, 0x88, 0x6f, 0x8a, 0x23, 0xe9, 0xce, 0x0c, 0xde, 0x95, 0x67, 0x27, 0xe2, 0xa8, + 0x47, 0x7c, 0x01, 0x46, 0x3f, 0x07, 0x90, 0xf7, 0x8b, 0x88, 0x06, 0xb5, 0xec, 0x41, 0xe6, 0x61, + 0x65, 0xed, 0x08, 0x16, 0x00, 0x39, 0x82, 0x0b, 0xe3, 0xf0, 0x2b, 0x40, 0x17, 0x50, 0x0e, 0xac, + 0xa9, 0xe7, 0x10, 0x73, 0x4a, 0xf8, 0x65, 0x3c, 0x86, 0x7f, 0xf8, 0x1e, 0x8f, 0xe8, 0x4b, 0xfc, + 0xa9, 0x84, 0xe3, 0x52, 0x90, 0xa0, 0x50, 0x13, 0xf6, 0xc4, 0x55, 0x89, 0xe7, 0x8c, 0x44, 0xbe, + 0xe7, 0x64, 0x13, 0xda, 0x95, 0x47, 0xd1, 0x6b, 0xc4, 0x41, 0xfd, 0xeb, 0x22, 0x25, 0x95, 0x27, + 0xaa, 0x90, 0x99, 0xf9, 0x4e, 0x98, 0x1a, 0xe2, 0xb3, 0xa1, 0x43, 0x29, 0x79, 0x15, 0xba, 0x03, + 0xb7, 0xfb, 0x47, 0xa7, 0xbd, 0x93, 0xae, 0x79, 0xda, 0x1d, 0xe8, 0xe7, 0x1d, 0xf3, 0xf3, 0xb3, + 0x7e, 0xaf, 0xfb, 0xd4, 0x78, 0x66, 0x74, 0x3b, 0xd5, 0x14, 0xca, 0x41, 0x66, 0x70, 0xde, 0xab, + 0x6a, 0xa8, 0x0a, 0x25, 0x7c, 0x74, 0xd6, 0x39, 0x3f, 0x35, 0xfb, 0x83, 0x23, 0x3c, 0xa8, 0xa6, + 0x1b, 0x0f, 0xa0, 0xba, 0x3a, 0x46, 0x44, 0x2e, 0xca, 0xe9, 0x13, 0xe6, 0xa2, 0xf8, 0x6e, 0xfc, + 0x2f, 0x0d, 0x3b, 0x6d, 0x3a, 0xf9, 0xf5, 0x8c, 0xf8, 0xf3, 0x45, 0xc0, 0x77, 0xb8, 0x35, 0x74, + 0x88, 0xe9, 0x93, 0x31, 0xf1, 0x89, 0x3b, 0x8a, 0x7a, 0xfd, 0xfd, 0x35, 0x2e, 0x8b, 0x14, 0x0c, + 0x04, 0x0a, 0x57, 0x24, 0x18, 0x47, 0x58, 0x74, 0x0a, 0x88, 0xda, 0xc4, 0xe5, 0x74, 0x3c, 0xa7, + 0xee, 0xc4, 0x1c, 0x8b, 0x82, 0x0c, 0x6a, 0x69, 0x39, 0xcc, 0x3f, 0x59, 0x1b, 0x47, 0x59, 0xb5, + 0x78, 0x37, 0x81, 0x94, 0xbc, 0x40, 0x54, 0xa9, 0xcf, 0xae, 0x42, 0x7f, 0xcb, 0xbe, 0x92, 0xc1, + 0x05, 0xc1, 0x91, 0x6e, 0x46, 0x2f, 0x57, 0xa3, 0xbd, 0x25, 0xa3, 0xfd, 0xf8, 0x06, 0xd3, 0x6f, + 0x8e, 0xf4, 0x47, 0x0c, 0xce, 0x3f, 0xb2, 0x50, 0x0e, 0x03, 0xf3, 0x94, 0xb9, 0x63, 0x3a, 0x41, + 0xcf, 0x61, 0xd7, 0x8e, 0x3a, 0x82, 0xc9, 0x94, 0x2d, 0x61, 0x0d, 0xaf, 0xdb, 0x12, 0x56, 0x3b, + 0x88, 0xd8, 0x12, 0xec, 0xd5, 0xae, 0xf2, 0x3b, 0xb8, 0xb5, 0xbc, 0x7e, 0x44, 0xba, 0x55, 0x3f, + 0x7e, 0xb4, 0x79, 0x0d, 0xe8, 0x29, 0xbc, 0x37, 0xba, 0xa6, 0x3b, 0x0c, 0x60, 0x77, 0x48, 0x27, + 0xe6, 0xef, 0x85, 0x13, 0x63, 0xed, 0x6a, 0x11, 0x7c, 0xb0, 0x99, 0xcf, 0xc5, 0x1c, 0x1d, 0xae, + 0xa4, 0xe0, 0x05, 0xec, 0xc8, 0xdd, 0xdd, 0xb3, 0x5c, 0x73, 0x24, 0x5d, 0x24, 0xab, 0x76, 0xfd, + 0xf6, 0xb7, 0xe4, 0xce, 0xe6, 0x20, 0xc4, 0x2a, 0x12, 0x57, 0xf8, 0x12, 0x5d, 0xff, 0x4b, 0x1a, + 0x2a, 0xcb, 0x22, 0xe8, 0x09, 0x40, 0xc0, 0x2d, 0x9f, 0x9b, 0x42, 0x34, 0x4c, 0xf7, 0xfa, 0xe2, + 0x2e, 0xf5, 0x7b, 0x42, 0xe9, 0x15, 0xbf, 0x27, 0x70, 0x41, 0x4a, 0x0b, 0x1a, 0xfd, 0x00, 0xf2, + 0xc4, 0xb5, 0x15, 0x30, 0x7d, 0x23, 0x30, 0x47, 0x5c, 0x5b, 0xc2, 0x3e, 0x0b, 0x9f, 0x28, 0xb8, + 0xaa, 0x28, 0xc2, 0xa0, 0xdc, 0x54, 0x13, 0x95, 0x18, 0x26, 0x39, 0xe8, 0x05, 0x7c, 0x9b, 0xb8, + 0xb2, 0x5e, 0xad, 0x19, 0x67, 0xa6, 0xc7, 0xbc, 0x99, 0x23, 0x7f, 0x16, 0x99, 0x6c, 0x6c, 0xae, + 0x7a, 0x51, 0x44, 0x26, 0x8f, 0xef, 0x2b, 0xc0, 0xd1, 0x8c, 0xb3, 0x5e, 0x2c, 0x7e, 0x3e, 0x5e, + 0xf6, 0x49, 0x3c, 0x45, 0xff, 0x00, 0xc5, 0x28, 0x64, 0xc7, 0x64, 0xfe, 0xb1, 0xdb, 0x83, 0xaa, + 0x67, 0xd3, 0x9d, 0x4d, 0x87, 0xe1, 0xf2, 0xa6, 0xea, 0xf9, 0x4c, 0x32, 0x1a, 0x06, 0x94, 0xe2, + 0x4c, 0x17, 0xb7, 0x3f, 0x01, 0x10, 0x1d, 0x81, 0xcf, 0xcd, 0xd7, 0x64, 0xfe, 0x6e, 0xa0, 0x96, + 0xc7, 0x1c, 0x99, 0xe3, 0x82, 0x92, 0x3e, 0x26, 0xf3, 0xc6, 0x7f, 0x34, 0xc8, 0x08, 0x15, 0x1f, + 0x69, 0xd4, 0xfe, 0x38, 0x6c, 0xa7, 0xaa, 0x93, 0x3d, 0x58, 0x6f, 0x43, 0x53, 0x74, 0xdf, 0xae, + 0x43, 0xe4, 0xe2, 0x22, 0x31, 0xf5, 0x81, 0xd8, 0x3c, 0x62, 0xa6, 0xe8, 0xcc, 0x72, 0x6a, 0x87, + 0x9d, 0x59, 0x7c, 0xa3, 0x2a, 0xa4, 0xc3, 0x2d, 0x24, 0xa3, 0xa7, 0x70, 0x9a, 0xda, 0x68, 0x3f, + 0xdc, 0x25, 0x44, 0x9a, 0x14, 0xf4, 0x94, 0xda, 0x26, 0xda, 0x05, 0xc8, 0x51, 0x5b, 0x0e, 0xc7, + 0xc6, 0xdf, 0x34, 0x28, 0x60, 0x32, 0x62, 0xbe, 0x2d, 0x9e, 0xf9, 0x2b, 0x28, 0x2f, 0x7a, 0x8a, + 0x70, 0x96, 0x4a, 0xce, 0x7b, 0x37, 0xf5, 0x93, 0x63, 0x22, 0x7e, 0x2e, 0x96, 0xec, 0xa4, 0xd7, + 0x75, 0x28, 0x2f, 0xaa, 0x5c, 0xe8, 0xca, 0x7c, 0xa5, 0xcf, 0x12, 0xe9, 0x22, 0xf6, 0xfb, 0xe1, + 0x82, 0x8c, 0x93, 0xea, 0x12, 0xca, 0x4b, 0x79, 0xb1, 0xb2, 0x7d, 0x69, 0xab, 0xdb, 0xd7, 0x1d, + 0x00, 0x69, 0x11, 0x49, 0x2e, 0x67, 0x21, 0xc7, 0xb0, 0xd1, 0x6d, 0xc8, 0xab, 0xa4, 0xa4, 0xaa, + 0x8c, 0x0a, 0x38, 0x27, 0x69, 0xc3, 0x6e, 0xfc, 0x12, 0xf2, 0x5d, 0x99, 0x03, 0x86, 0x8d, 0xbe, + 0x0f, 0x59, 0x55, 0x6a, 0xda, 0x46, 0xa5, 0xa6, 0x84, 0x1f, 0x71, 0x80, 0xc4, 0x8e, 0x5f, 0x87, + 0xaf, 0x9d, 0x18, 0xc7, 0xdd, 0x13, 0x43, 0x3f, 0x7f, 0xa7, 0xe9, 0xef, 0x42, 0xf9, 0x79, 0x17, + 0xbf, 0x32, 0x3f, 0x3f, 0x93, 0x22, 0xaf, 0xaa, 0x1a, 0x2a, 0x41, 0x3e, 0xa6, 0xd2, 0x82, 0xea, + 0x9d, 0xf7, 0xfb, 0x46, 0xfb, 0xa4, 0x5b, 0xcd, 0x20, 0x80, 0xed, 0xf0, 0x64, 0x0b, 0xed, 0x40, + 0x51, 0x42, 0x43, 0x46, 0xf6, 0x51, 0x17, 0xf2, 0xd1, 0x3a, 0x83, 0x6e, 0xc3, 0xad, 0x67, 0xc6, + 0x49, 0xd7, 0x1c, 0xbc, 0xea, 0x75, 0x57, 0xae, 0xdc, 0x81, 0x62, 0xdb, 0x38, 0x3b, 0xc2, 0xaf, + 0x4c, 0x21, 0x51, 0xd5, 0x50, 0x19, 0x0a, 0x83, 0xee, 0xcb, 0x81, 0x22, 0xd3, 0xed, 0x3f, 0x69, + 0x70, 0x7b, 0xc4, 0xa6, 0xd7, 0xbf, 0xb4, 0x0d, 0x1d, 0xc7, 0x8b, 0x76, 0x04, 0xed, 0x37, 0x3f, + 0x0a, 0x85, 0x26, 0xcc, 0xb1, 0xdc, 0x49, 0x93, 0xf9, 0x93, 0xd6, 0x84, 0xb8, 0xb2, 0x8b, 0xb5, + 0xd4, 0x91, 0xe5, 0xd1, 0x60, 0xe5, 0x3f, 0x37, 0x3f, 0xb1, 0x1d, 0xef, 0xef, 0xe9, 0xbd, 0xcf, + 0x14, 0x54, 0x8e, 0x8e, 0x66, 0xc7, 0xf1, 0x9a, 0xcf, 0x0f, 0xbf, 0x8c, 0xb8, 0x17, 0x92, 0x7b, + 0xd1, 0x71, 0xbc, 0x8b, 0xe7, 0x87, 0xc3, 0x6d, 0xa9, 0xf0, 0xf1, 0xff, 0x03, 0x00, 0x00, 0xff, + 0xff, 0x69, 0x7c, 0xa2, 0x7a, 0x09, 0x12, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/pubsub/v1/pubsub.pb.go b/vendor/google.golang.org/genproto/googleapis/pubsub/v1/pubsub.pb.go index 96690c7f4..4b6cf03a6 100644 --- a/vendor/google.golang.org/genproto/googleapis/pubsub/v1/pubsub.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/pubsub/v1/pubsub.pb.go @@ -28,6 +28,50 @@ var _ = math.Inf // proto package needs to be updated. const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package +type MessageStoragePolicy struct { + // The list of GCP regions where messages that are published to the topic may + // be persisted in storage. Messages published by publishers running in + // non-allowed GCP regions (or running outside of GCP altogether) will be + // routed for storage in one of the allowed regions. An empty list indicates a + // misconfiguration at the project or organization level, which will result in + // all Publish operations failing. + AllowedPersistenceRegions []string `protobuf:"bytes,1,rep,name=allowed_persistence_regions,json=allowedPersistenceRegions,proto3" json:"allowed_persistence_regions,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *MessageStoragePolicy) Reset() { *m = MessageStoragePolicy{} } +func (m *MessageStoragePolicy) String() string { return proto.CompactTextString(m) } +func (*MessageStoragePolicy) ProtoMessage() {} +func (*MessageStoragePolicy) Descriptor() ([]byte, []int) { + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{0} +} +func (m *MessageStoragePolicy) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_MessageStoragePolicy.Unmarshal(m, b) +} +func (m *MessageStoragePolicy) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_MessageStoragePolicy.Marshal(b, m, deterministic) +} +func (dst *MessageStoragePolicy) XXX_Merge(src proto.Message) { + xxx_messageInfo_MessageStoragePolicy.Merge(dst, src) +} +func (m *MessageStoragePolicy) XXX_Size() int { + return xxx_messageInfo_MessageStoragePolicy.Size(m) +} +func (m *MessageStoragePolicy) XXX_DiscardUnknown() { + xxx_messageInfo_MessageStoragePolicy.DiscardUnknown(m) +} + +var xxx_messageInfo_MessageStoragePolicy proto.InternalMessageInfo + +func (m *MessageStoragePolicy) GetAllowedPersistenceRegions() []string { + if m != nil { + return m.AllowedPersistenceRegions + } + return nil +} + // A topic resource. type Topic struct { // The name of the topic. It must have the format @@ -36,19 +80,26 @@ type Topic struct { // underscores (`_`), periods (`.`), tildes (`~`), plus (`+`) or percent // signs (`%`). It must be between 3 and 255 characters in length, and it // must not start with `"goog"`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // User labels. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + // Policy constraining how messages published to the topic may be stored. It + // is determined when the topic is created based on the policy configured at + // the project level. It must not be set by the caller in the request to + // CreateTopic or to UpdateTopic. This field will be populated in the + // responses for GetTopic, CreateTopic, and UpdateTopic: if not present in the + // response, then no constraints are in effect. + MessageStoragePolicy *MessageStoragePolicy `protobuf:"bytes,3,opt,name=message_storage_policy,json=messageStoragePolicy,proto3" json:"message_storage_policy,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *Topic) Reset() { *m = Topic{} } func (m *Topic) String() string { return proto.CompactTextString(m) } func (*Topic) ProtoMessage() {} func (*Topic) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{0} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{1} } func (m *Topic) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Topic.Unmarshal(m, b) @@ -82,22 +133,29 @@ func (m *Topic) GetLabels() map[string]string { return nil } +func (m *Topic) GetMessageStoragePolicy() *MessageStoragePolicy { + if m != nil { + return m.MessageStoragePolicy + } + return nil +} + // A message data and its attributes. The message payload must not be empty; // it must contain either a non-empty data field, or at least one attribute. type PubsubMessage struct { // The message payload. Data []byte `protobuf:"bytes,1,opt,name=data,proto3" json:"data,omitempty"` // Optional attributes for this message. - Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes,proto3" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // ID of this message, assigned by the server when the message is published. // Guaranteed to be unique within the topic. This value may be read by a // subscriber that receives a `PubsubMessage` via a `Pull` call or a push // delivery. It must not be populated by the publisher in a `Publish` call. - MessageId string `protobuf:"bytes,3,opt,name=message_id,json=messageId" json:"message_id,omitempty"` + MessageId string `protobuf:"bytes,3,opt,name=message_id,json=messageId,proto3" json:"message_id,omitempty"` // The time at which the message was published, populated by the server when // it receives the `Publish` call. It must not be populated by the // publisher in a `Publish` call. - PublishTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=publish_time,json=publishTime" json:"publish_time,omitempty"` + PublishTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=publish_time,json=publishTime,proto3" json:"publish_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -107,7 +165,7 @@ func (m *PubsubMessage) Reset() { *m = PubsubMessage{} } func (m *PubsubMessage) String() string { return proto.CompactTextString(m) } func (*PubsubMessage) ProtoMessage() {} func (*PubsubMessage) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{1} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{2} } func (m *PubsubMessage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PubsubMessage.Unmarshal(m, b) @@ -159,7 +217,7 @@ func (m *PubsubMessage) GetPublishTime() *timestamp.Timestamp { type GetTopicRequest struct { // The name of the topic to get. // Format is `projects/{project}/topics/{topic}`. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,7 +227,7 @@ func (m *GetTopicRequest) Reset() { *m = GetTopicRequest{} } func (m *GetTopicRequest) String() string { return proto.CompactTextString(m) } func (*GetTopicRequest) ProtoMessage() {} func (*GetTopicRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{2} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{3} } func (m *GetTopicRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetTopicRequest.Unmarshal(m, b) @@ -198,11 +256,14 @@ func (m *GetTopicRequest) GetTopic() string { // Request for the UpdateTopic method. type UpdateTopicRequest struct { - // The topic to update. - Topic *Topic `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` - // Indicates which fields in the provided topic to update. - // Must be specified and non-empty. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + // The updated topic object. + Topic *Topic `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` + // Indicates which fields in the provided topic to update. Must be specified + // and non-empty. Note that if `update_mask` contains + // "message_storage_policy" then the new value will be determined based on the + // policy configured at the project or organization level. The + // `message_storage_policy` must not be set in the `topic` provided above. + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -212,7 +273,7 @@ func (m *UpdateTopicRequest) Reset() { *m = UpdateTopicRequest{} } func (m *UpdateTopicRequest) String() string { return proto.CompactTextString(m) } func (*UpdateTopicRequest) ProtoMessage() {} func (*UpdateTopicRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{3} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{4} } func (m *UpdateTopicRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateTopicRequest.Unmarshal(m, b) @@ -250,9 +311,9 @@ func (m *UpdateTopicRequest) GetUpdateMask() *field_mask.FieldMask { type PublishRequest struct { // The messages in the request will be published on this topic. // Format is `projects/{project}/topics/{topic}`. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` // The messages to publish. - Messages []*PubsubMessage `protobuf:"bytes,2,rep,name=messages" json:"messages,omitempty"` + Messages []*PubsubMessage `protobuf:"bytes,2,rep,name=messages,proto3" json:"messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,7 +323,7 @@ func (m *PublishRequest) Reset() { *m = PublishRequest{} } func (m *PublishRequest) String() string { return proto.CompactTextString(m) } func (*PublishRequest) ProtoMessage() {} func (*PublishRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{4} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{5} } func (m *PublishRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PublishRequest.Unmarshal(m, b) @@ -301,7 +362,7 @@ type PublishResponse struct { // The server-assigned ID of each published message, in the same order as // the messages in the request. IDs are guaranteed to be unique within // the topic. - MessageIds []string `protobuf:"bytes,1,rep,name=message_ids,json=messageIds" json:"message_ids,omitempty"` + MessageIds []string `protobuf:"bytes,1,rep,name=message_ids,json=messageIds,proto3" json:"message_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -311,7 +372,7 @@ func (m *PublishResponse) Reset() { *m = PublishResponse{} } func (m *PublishResponse) String() string { return proto.CompactTextString(m) } func (*PublishResponse) ProtoMessage() {} func (*PublishResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{5} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{6} } func (m *PublishResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PublishResponse.Unmarshal(m, b) @@ -342,13 +403,13 @@ func (m *PublishResponse) GetMessageIds() []string { type ListTopicsRequest struct { // The name of the cloud project that topics belong to. // Format is `projects/{project}`. - Project string `protobuf:"bytes,1,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,1,opt,name=project,proto3" json:"project,omitempty"` // Maximum number of topics to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last `ListTopicsResponse`; indicates that this is // a continuation of a prior `ListTopics` call, and that the system should // return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -358,7 +419,7 @@ func (m *ListTopicsRequest) Reset() { *m = ListTopicsRequest{} } func (m *ListTopicsRequest) String() string { return proto.CompactTextString(m) } func (*ListTopicsRequest) ProtoMessage() {} func (*ListTopicsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{6} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{7} } func (m *ListTopicsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListTopicsRequest.Unmarshal(m, b) @@ -402,10 +463,10 @@ func (m *ListTopicsRequest) GetPageToken() string { // Response for the `ListTopics` method. type ListTopicsResponse struct { // The resulting topics. - Topics []*Topic `protobuf:"bytes,1,rep,name=topics" json:"topics,omitempty"` + Topics []*Topic `protobuf:"bytes,1,rep,name=topics,proto3" json:"topics,omitempty"` // If not empty, indicates that there may be more topics that match the // request; this value should be passed in a new `ListTopicsRequest`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -415,7 +476,7 @@ func (m *ListTopicsResponse) Reset() { *m = ListTopicsResponse{} } func (m *ListTopicsResponse) String() string { return proto.CompactTextString(m) } func (*ListTopicsResponse) ProtoMessage() {} func (*ListTopicsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{7} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{8} } func (m *ListTopicsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListTopicsResponse.Unmarshal(m, b) @@ -453,13 +514,13 @@ func (m *ListTopicsResponse) GetNextPageToken() string { type ListTopicSubscriptionsRequest struct { // The name of the topic that subscriptions are attached to. // Format is `projects/{project}/topics/{topic}`. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` // Maximum number of subscription names to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last `ListTopicSubscriptionsResponse`; indicates // that this is a continuation of a prior `ListTopicSubscriptions` call, and // that the system should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -469,7 +530,7 @@ func (m *ListTopicSubscriptionsRequest) Reset() { *m = ListTopicSubscrip func (m *ListTopicSubscriptionsRequest) String() string { return proto.CompactTextString(m) } func (*ListTopicSubscriptionsRequest) ProtoMessage() {} func (*ListTopicSubscriptionsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{8} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{9} } func (m *ListTopicSubscriptionsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListTopicSubscriptionsRequest.Unmarshal(m, b) @@ -513,11 +574,11 @@ func (m *ListTopicSubscriptionsRequest) GetPageToken() string { // Response for the `ListTopicSubscriptions` method. type ListTopicSubscriptionsResponse struct { // The names of the subscriptions that match the request. - Subscriptions []string `protobuf:"bytes,1,rep,name=subscriptions" json:"subscriptions,omitempty"` + Subscriptions []string `protobuf:"bytes,1,rep,name=subscriptions,proto3" json:"subscriptions,omitempty"` // If not empty, indicates that there may be more subscriptions that match // the request; this value should be passed in a new // `ListTopicSubscriptionsRequest` to get more subscriptions. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -527,7 +588,7 @@ func (m *ListTopicSubscriptionsResponse) Reset() { *m = ListTopicSubscri func (m *ListTopicSubscriptionsResponse) String() string { return proto.CompactTextString(m) } func (*ListTopicSubscriptionsResponse) ProtoMessage() {} func (*ListTopicSubscriptionsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{9} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{10} } func (m *ListTopicSubscriptionsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListTopicSubscriptionsResponse.Unmarshal(m, b) @@ -561,11 +622,129 @@ func (m *ListTopicSubscriptionsResponse) GetNextPageToken() string { return "" } +// Request for the `ListTopicSnapshots` method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. +type ListTopicSnapshotsRequest struct { + // The name of the topic that snapshots are attached to. + // Format is `projects/{project}/topics/{topic}`. + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` + // Maximum number of snapshot names to return. + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` + // The value returned by the last `ListTopicSnapshotsResponse`; indicates + // that this is a continuation of a prior `ListTopicSnapshots` call, and + // that the system should return the next page of data. + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListTopicSnapshotsRequest) Reset() { *m = ListTopicSnapshotsRequest{} } +func (m *ListTopicSnapshotsRequest) String() string { return proto.CompactTextString(m) } +func (*ListTopicSnapshotsRequest) ProtoMessage() {} +func (*ListTopicSnapshotsRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{11} +} +func (m *ListTopicSnapshotsRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListTopicSnapshotsRequest.Unmarshal(m, b) +} +func (m *ListTopicSnapshotsRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListTopicSnapshotsRequest.Marshal(b, m, deterministic) +} +func (dst *ListTopicSnapshotsRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListTopicSnapshotsRequest.Merge(dst, src) +} +func (m *ListTopicSnapshotsRequest) XXX_Size() int { + return xxx_messageInfo_ListTopicSnapshotsRequest.Size(m) +} +func (m *ListTopicSnapshotsRequest) XXX_DiscardUnknown() { + xxx_messageInfo_ListTopicSnapshotsRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_ListTopicSnapshotsRequest proto.InternalMessageInfo + +func (m *ListTopicSnapshotsRequest) GetTopic() string { + if m != nil { + return m.Topic + } + return "" +} + +func (m *ListTopicSnapshotsRequest) GetPageSize() int32 { + if m != nil { + return m.PageSize + } + return 0 +} + +func (m *ListTopicSnapshotsRequest) GetPageToken() string { + if m != nil { + return m.PageToken + } + return "" +} + +// Response for the `ListTopicSnapshots` method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. +type ListTopicSnapshotsResponse struct { + // The names of the snapshots that match the request. + Snapshots []string `protobuf:"bytes,1,rep,name=snapshots,proto3" json:"snapshots,omitempty"` + // If not empty, indicates that there may be more snapshots that match + // the request; this value should be passed in a new + // `ListTopicSnapshotsRequest` to get more snapshots. + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ListTopicSnapshotsResponse) Reset() { *m = ListTopicSnapshotsResponse{} } +func (m *ListTopicSnapshotsResponse) String() string { return proto.CompactTextString(m) } +func (*ListTopicSnapshotsResponse) ProtoMessage() {} +func (*ListTopicSnapshotsResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{12} +} +func (m *ListTopicSnapshotsResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ListTopicSnapshotsResponse.Unmarshal(m, b) +} +func (m *ListTopicSnapshotsResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ListTopicSnapshotsResponse.Marshal(b, m, deterministic) +} +func (dst *ListTopicSnapshotsResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_ListTopicSnapshotsResponse.Merge(dst, src) +} +func (m *ListTopicSnapshotsResponse) XXX_Size() int { + return xxx_messageInfo_ListTopicSnapshotsResponse.Size(m) +} +func (m *ListTopicSnapshotsResponse) XXX_DiscardUnknown() { + xxx_messageInfo_ListTopicSnapshotsResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_ListTopicSnapshotsResponse proto.InternalMessageInfo + +func (m *ListTopicSnapshotsResponse) GetSnapshots() []string { + if m != nil { + return m.Snapshots + } + return nil +} + +func (m *ListTopicSnapshotsResponse) GetNextPageToken() string { + if m != nil { + return m.NextPageToken + } + return "" +} + // Request for the `DeleteTopic` method. type DeleteTopicRequest struct { // Name of the topic to delete. // Format is `projects/{project}/topics/{topic}`. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -575,7 +754,7 @@ func (m *DeleteTopicRequest) Reset() { *m = DeleteTopicRequest{} } func (m *DeleteTopicRequest) String() string { return proto.CompactTextString(m) } func (*DeleteTopicRequest) ProtoMessage() {} func (*DeleteTopicRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{10} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{13} } func (m *DeleteTopicRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteTopicRequest.Unmarshal(m, b) @@ -610,16 +789,16 @@ type Subscription struct { // (`[0-9]`), dashes (`-`), underscores (`_`), periods (`.`), tildes (`~`), // plus (`+`) or percent signs (`%`). It must be between 3 and 255 characters // in length, and it must not start with `"goog"`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the topic from which this subscription is receiving messages. // Format is `projects/{project}/topics/{topic}`. // The value of this field will be `_deleted-topic_` if the topic has been // deleted. - Topic string `protobuf:"bytes,2,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,2,opt,name=topic,proto3" json:"topic,omitempty"` // If push delivery is used with this subscription, this field is // used to configure it. An empty `pushConfig` signifies that the subscriber // will pull and ack messages using API methods. - PushConfig *PushConfig `protobuf:"bytes,4,opt,name=push_config,json=pushConfig" json:"push_config,omitempty"` + PushConfig *PushConfig `protobuf:"bytes,4,opt,name=push_config,json=pushConfig,proto3" json:"push_config,omitempty"` // This value is the maximum time after a subscriber receives a message // before the subscriber should acknowledge the message. After message // delivery but before the ack deadline expires and before the message is @@ -629,7 +808,8 @@ type Subscription struct { // For pull subscriptions, this value is used as the initial value for the ack // deadline. To override this value for a given message, call // `ModifyAckDeadline` with the corresponding `ack_id` if using - // pull. + // non-streaming pull or send the `ack_id` in a + // `StreamingModifyAckDeadlineRequest` if using streaming pull. // The minimum custom deadline you can specify is 10 seconds. // The maximum custom deadline you can specify is 600 seconds (10 minutes). // If this parameter is 0, a default value of 10 seconds is used. @@ -639,21 +819,27 @@ type Subscription struct { // // If the subscriber never acknowledges the message, the Pub/Sub // system will eventually redeliver the message. - AckDeadlineSeconds int32 `protobuf:"varint,5,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds" json:"ack_deadline_seconds,omitempty"` + AckDeadlineSeconds int32 `protobuf:"varint,5,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds,proto3" json:"ack_deadline_seconds,omitempty"` // Indicates whether to retain acknowledged messages. If true, then // messages are not expunged from the subscription's backlog, even if they are // acknowledged, until they fall out of the `message_retention_duration` - // window. - RetainAckedMessages bool `protobuf:"varint,7,opt,name=retain_acked_messages,json=retainAckedMessages" json:"retain_acked_messages,omitempty"` + // window.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + RetainAckedMessages bool `protobuf:"varint,7,opt,name=retain_acked_messages,json=retainAckedMessages,proto3" json:"retain_acked_messages,omitempty"` // How long to retain unacknowledged messages in the subscription's backlog, // from the moment a message is published. // If `retain_acked_messages` is true, then this also configures the retention // of acknowledged messages, and thus configures how far back in time a `Seek` // can be done. Defaults to 7 days. Cannot be more than 7 days or less than 10 - // minutes. - MessageRetentionDuration *duration.Duration `protobuf:"bytes,8,opt,name=message_retention_duration,json=messageRetentionDuration" json:"message_retention_duration,omitempty"` + // minutes.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + MessageRetentionDuration *duration.Duration `protobuf:"bytes,8,opt,name=message_retention_duration,json=messageRetentionDuration,proto3" json:"message_retention_duration,omitempty"` // User labels. - Labels map[string]string `protobuf:"bytes,9,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,9,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -663,7 +849,7 @@ func (m *Subscription) Reset() { *m = Subscription{} } func (m *Subscription) String() string { return proto.CompactTextString(m) } func (*Subscription) ProtoMessage() {} func (*Subscription) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{11} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{14} } func (m *Subscription) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Subscription.Unmarshal(m, b) @@ -736,7 +922,7 @@ func (m *Subscription) GetLabels() map[string]string { type PushConfig struct { // A URL locating the endpoint to which messages should be pushed. // For example, a Webhook endpoint might use "https://example.com/push". - PushEndpoint string `protobuf:"bytes,1,opt,name=push_endpoint,json=pushEndpoint" json:"push_endpoint,omitempty"` + PushEndpoint string `protobuf:"bytes,1,opt,name=push_endpoint,json=pushEndpoint,proto3" json:"push_endpoint,omitempty"` // Endpoint configuration attributes. // // Every endpoint has a set of API supported attributes that can be used to @@ -758,7 +944,7 @@ type PushConfig struct { // // * `v1beta1`: uses the push format defined in the v1beta1 Pub/Sub API. // * `v1` or `v1beta2`: uses the push format defined in the v1 Pub/Sub API. - Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes,proto3" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -768,7 +954,7 @@ func (m *PushConfig) Reset() { *m = PushConfig{} } func (m *PushConfig) String() string { return proto.CompactTextString(m) } func (*PushConfig) ProtoMessage() {} func (*PushConfig) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{12} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{15} } func (m *PushConfig) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PushConfig.Unmarshal(m, b) @@ -805,9 +991,9 @@ func (m *PushConfig) GetAttributes() map[string]string { // A message and its corresponding acknowledgment ID. type ReceivedMessage struct { // This ID can be used to acknowledge the received message. - AckId string `protobuf:"bytes,1,opt,name=ack_id,json=ackId" json:"ack_id,omitempty"` + AckId string `protobuf:"bytes,1,opt,name=ack_id,json=ackId,proto3" json:"ack_id,omitempty"` // The message. - Message *PubsubMessage `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Message *PubsubMessage `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -817,7 +1003,7 @@ func (m *ReceivedMessage) Reset() { *m = ReceivedMessage{} } func (m *ReceivedMessage) String() string { return proto.CompactTextString(m) } func (*ReceivedMessage) ProtoMessage() {} func (*ReceivedMessage) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{13} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{16} } func (m *ReceivedMessage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ReceivedMessage.Unmarshal(m, b) @@ -855,7 +1041,7 @@ func (m *ReceivedMessage) GetMessage() *PubsubMessage { type GetSubscriptionRequest struct { // The name of the subscription to get. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -865,7 +1051,7 @@ func (m *GetSubscriptionRequest) Reset() { *m = GetSubscriptionRequest{} func (m *GetSubscriptionRequest) String() string { return proto.CompactTextString(m) } func (*GetSubscriptionRequest) ProtoMessage() {} func (*GetSubscriptionRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{14} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{17} } func (m *GetSubscriptionRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSubscriptionRequest.Unmarshal(m, b) @@ -895,10 +1081,10 @@ func (m *GetSubscriptionRequest) GetSubscription() string { // Request for the UpdateSubscription method. type UpdateSubscriptionRequest struct { // The updated subscription object. - Subscription *Subscription `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription *Subscription `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // Indicates which fields in the provided subscription to update. // Must be specified and non-empty. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -908,7 +1094,7 @@ func (m *UpdateSubscriptionRequest) Reset() { *m = UpdateSubscriptionReq func (m *UpdateSubscriptionRequest) String() string { return proto.CompactTextString(m) } func (*UpdateSubscriptionRequest) ProtoMessage() {} func (*UpdateSubscriptionRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{15} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{18} } func (m *UpdateSubscriptionRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateSubscriptionRequest.Unmarshal(m, b) @@ -946,13 +1132,13 @@ func (m *UpdateSubscriptionRequest) GetUpdateMask() *field_mask.FieldMask { type ListSubscriptionsRequest struct { // The name of the cloud project that subscriptions belong to. // Format is `projects/{project}`. - Project string `protobuf:"bytes,1,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,1,opt,name=project,proto3" json:"project,omitempty"` // Maximum number of subscriptions to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last `ListSubscriptionsResponse`; indicates that // this is a continuation of a prior `ListSubscriptions` call, and that the // system should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -962,7 +1148,7 @@ func (m *ListSubscriptionsRequest) Reset() { *m = ListSubscriptionsReque func (m *ListSubscriptionsRequest) String() string { return proto.CompactTextString(m) } func (*ListSubscriptionsRequest) ProtoMessage() {} func (*ListSubscriptionsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{16} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{19} } func (m *ListSubscriptionsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSubscriptionsRequest.Unmarshal(m, b) @@ -1006,11 +1192,11 @@ func (m *ListSubscriptionsRequest) GetPageToken() string { // Response for the `ListSubscriptions` method. type ListSubscriptionsResponse struct { // The subscriptions that match the request. - Subscriptions []*Subscription `protobuf:"bytes,1,rep,name=subscriptions" json:"subscriptions,omitempty"` + Subscriptions []*Subscription `protobuf:"bytes,1,rep,name=subscriptions,proto3" json:"subscriptions,omitempty"` // If not empty, indicates that there may be more subscriptions that match // the request; this value should be passed in a new // `ListSubscriptionsRequest` to get more subscriptions. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1020,7 +1206,7 @@ func (m *ListSubscriptionsResponse) Reset() { *m = ListSubscriptionsResp func (m *ListSubscriptionsResponse) String() string { return proto.CompactTextString(m) } func (*ListSubscriptionsResponse) ProtoMessage() {} func (*ListSubscriptionsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{17} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{20} } func (m *ListSubscriptionsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSubscriptionsResponse.Unmarshal(m, b) @@ -1058,7 +1244,7 @@ func (m *ListSubscriptionsResponse) GetNextPageToken() string { type DeleteSubscriptionRequest struct { // The subscription to delete. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1068,7 +1254,7 @@ func (m *DeleteSubscriptionRequest) Reset() { *m = DeleteSubscriptionReq func (m *DeleteSubscriptionRequest) String() string { return proto.CompactTextString(m) } func (*DeleteSubscriptionRequest) ProtoMessage() {} func (*DeleteSubscriptionRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{18} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{21} } func (m *DeleteSubscriptionRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteSubscriptionRequest.Unmarshal(m, b) @@ -1099,14 +1285,14 @@ func (m *DeleteSubscriptionRequest) GetSubscription() string { type ModifyPushConfigRequest struct { // The name of the subscription. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // The push configuration for future deliveries. // // An empty `pushConfig` indicates that the Pub/Sub system should // stop pushing messages from the given subscription and allow // messages to be pulled and acknowledged - effectively pausing - // the subscription if `Pull` is not called. - PushConfig *PushConfig `protobuf:"bytes,2,opt,name=push_config,json=pushConfig" json:"push_config,omitempty"` + // the subscription if `Pull` or `StreamingPull` is not called. + PushConfig *PushConfig `protobuf:"bytes,2,opt,name=push_config,json=pushConfig,proto3" json:"push_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1116,7 +1302,7 @@ func (m *ModifyPushConfigRequest) Reset() { *m = ModifyPushConfigRequest func (m *ModifyPushConfigRequest) String() string { return proto.CompactTextString(m) } func (*ModifyPushConfigRequest) ProtoMessage() {} func (*ModifyPushConfigRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{19} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{22} } func (m *ModifyPushConfigRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ModifyPushConfigRequest.Unmarshal(m, b) @@ -1154,17 +1340,17 @@ func (m *ModifyPushConfigRequest) GetPushConfig() *PushConfig { type PullRequest struct { // The subscription from which messages should be pulled. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // If this field set to true, the system will respond immediately even if // it there are no messages available to return in the `Pull` response. // Otherwise, the system may wait (for a bounded amount of time) until at // least one message is available, rather than returning no messages. The // client may cancel the request if it does not wish to wait any longer for // the response. - ReturnImmediately bool `protobuf:"varint,2,opt,name=return_immediately,json=returnImmediately" json:"return_immediately,omitempty"` + ReturnImmediately bool `protobuf:"varint,2,opt,name=return_immediately,json=returnImmediately,proto3" json:"return_immediately,omitempty"` // The maximum number of messages returned for this request. The Pub/Sub // system may return fewer than the number specified. - MaxMessages int32 `protobuf:"varint,3,opt,name=max_messages,json=maxMessages" json:"max_messages,omitempty"` + MaxMessages int32 `protobuf:"varint,3,opt,name=max_messages,json=maxMessages,proto3" json:"max_messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1174,7 +1360,7 @@ func (m *PullRequest) Reset() { *m = PullRequest{} } func (m *PullRequest) String() string { return proto.CompactTextString(m) } func (*PullRequest) ProtoMessage() {} func (*PullRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{20} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{23} } func (m *PullRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PullRequest.Unmarshal(m, b) @@ -1221,7 +1407,7 @@ type PullResponse struct { // there are no more available in the backlog. The Pub/Sub system may return // fewer than the `maxMessages` requested even if there are more messages // available in the backlog. - ReceivedMessages []*ReceivedMessage `protobuf:"bytes,1,rep,name=received_messages,json=receivedMessages" json:"received_messages,omitempty"` + ReceivedMessages []*ReceivedMessage `protobuf:"bytes,1,rep,name=received_messages,json=receivedMessages,proto3" json:"received_messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1231,7 +1417,7 @@ func (m *PullResponse) Reset() { *m = PullResponse{} } func (m *PullResponse) String() string { return proto.CompactTextString(m) } func (*PullResponse) ProtoMessage() {} func (*PullResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{21} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{24} } func (m *PullResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_PullResponse.Unmarshal(m, b) @@ -1262,9 +1448,9 @@ func (m *PullResponse) GetReceivedMessages() []*ReceivedMessage { type ModifyAckDeadlineRequest struct { // The name of the subscription. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // List of acknowledgment IDs. - AckIds []string `protobuf:"bytes,4,rep,name=ack_ids,json=ackIds" json:"ack_ids,omitempty"` + AckIds []string `protobuf:"bytes,4,rep,name=ack_ids,json=ackIds,proto3" json:"ack_ids,omitempty"` // The new ack deadline with respect to the time this request was sent to // the Pub/Sub system. For example, if the value is 10, the new // ack deadline will expire 10 seconds after the `ModifyAckDeadline` call @@ -1272,7 +1458,7 @@ type ModifyAckDeadlineRequest struct { // another pull request. // The minimum deadline you can specify is 0 seconds. // The maximum deadline you can specify is 600 seconds (10 minutes). - AckDeadlineSeconds int32 `protobuf:"varint,3,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds" json:"ack_deadline_seconds,omitempty"` + AckDeadlineSeconds int32 `protobuf:"varint,3,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds,proto3" json:"ack_deadline_seconds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1282,7 +1468,7 @@ func (m *ModifyAckDeadlineRequest) Reset() { *m = ModifyAckDeadlineReque func (m *ModifyAckDeadlineRequest) String() string { return proto.CompactTextString(m) } func (*ModifyAckDeadlineRequest) ProtoMessage() {} func (*ModifyAckDeadlineRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{22} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{25} } func (m *ModifyAckDeadlineRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ModifyAckDeadlineRequest.Unmarshal(m, b) @@ -1327,10 +1513,10 @@ func (m *ModifyAckDeadlineRequest) GetAckDeadlineSeconds() int32 { type AcknowledgeRequest struct { // The subscription whose message is being acknowledged. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // The acknowledgment ID for the messages being acknowledged that was returned // by the Pub/Sub system in the `Pull` response. Must not be empty. - AckIds []string `protobuf:"bytes,2,rep,name=ack_ids,json=ackIds" json:"ack_ids,omitempty"` + AckIds []string `protobuf:"bytes,2,rep,name=ack_ids,json=ackIds,proto3" json:"ack_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1340,7 +1526,7 @@ func (m *AcknowledgeRequest) Reset() { *m = AcknowledgeRequest{} } func (m *AcknowledgeRequest) String() string { return proto.CompactTextString(m) } func (*AcknowledgeRequest) ProtoMessage() {} func (*AcknowledgeRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{23} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{26} } func (m *AcknowledgeRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_AcknowledgeRequest.Unmarshal(m, b) @@ -1382,13 +1568,13 @@ type StreamingPullRequest struct { // provided in the first request on the stream, and must not be set in // subsequent requests from client to server. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // List of acknowledgement IDs for acknowledging previously received messages // (received on this stream or a different stream). If an ack ID has expired, // the corresponding message may be redelivered later. Acknowledging a message // more than once will not result in an error. If the acknowledgement ID is // malformed, the stream will be aborted with status `INVALID_ARGUMENT`. - AckIds []string `protobuf:"bytes,2,rep,name=ack_ids,json=ackIds" json:"ack_ids,omitempty"` + AckIds []string `protobuf:"bytes,2,rep,name=ack_ids,json=ackIds,proto3" json:"ack_ids,omitempty"` // The list of new ack deadlines for the IDs listed in // `modify_deadline_ack_ids`. The size of this list must be the same as the // size of `modify_deadline_ack_ids`. If it differs the stream will be aborted @@ -1400,18 +1586,18 @@ type StreamingPullRequest struct { // the message is immediately made available for another streaming or // non-streaming pull request. If the value is < 0 (an error), the stream will // be aborted with status `INVALID_ARGUMENT`. - ModifyDeadlineSeconds []int32 `protobuf:"varint,3,rep,packed,name=modify_deadline_seconds,json=modifyDeadlineSeconds" json:"modify_deadline_seconds,omitempty"` + ModifyDeadlineSeconds []int32 `protobuf:"varint,3,rep,packed,name=modify_deadline_seconds,json=modifyDeadlineSeconds,proto3" json:"modify_deadline_seconds,omitempty"` // List of acknowledgement IDs whose deadline will be modified based on the // corresponding element in `modify_deadline_seconds`. This field can be used // to indicate that more time is needed to process a message by the // subscriber, or to make the message available for redelivery if the // processing was interrupted. - ModifyDeadlineAckIds []string `protobuf:"bytes,4,rep,name=modify_deadline_ack_ids,json=modifyDeadlineAckIds" json:"modify_deadline_ack_ids,omitempty"` + ModifyDeadlineAckIds []string `protobuf:"bytes,4,rep,name=modify_deadline_ack_ids,json=modifyDeadlineAckIds,proto3" json:"modify_deadline_ack_ids,omitempty"` // The ack deadline to use for the stream. This must be provided in the // first request on the stream, but it can also be updated on subsequent // requests from client to server. The minimum deadline you can specify is 10 // seconds. The maximum deadline you can specify is 600 seconds (10 minutes). - StreamAckDeadlineSeconds int32 `protobuf:"varint,5,opt,name=stream_ack_deadline_seconds,json=streamAckDeadlineSeconds" json:"stream_ack_deadline_seconds,omitempty"` + StreamAckDeadlineSeconds int32 `protobuf:"varint,5,opt,name=stream_ack_deadline_seconds,json=streamAckDeadlineSeconds,proto3" json:"stream_ack_deadline_seconds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1421,7 +1607,7 @@ func (m *StreamingPullRequest) Reset() { *m = StreamingPullRequest{} } func (m *StreamingPullRequest) String() string { return proto.CompactTextString(m) } func (*StreamingPullRequest) ProtoMessage() {} func (*StreamingPullRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{24} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{27} } func (m *StreamingPullRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StreamingPullRequest.Unmarshal(m, b) @@ -1480,7 +1666,7 @@ func (m *StreamingPullRequest) GetStreamAckDeadlineSeconds() int32 { // messages from the server to the client. type StreamingPullResponse struct { // Received Pub/Sub messages. This will not be empty. - ReceivedMessages []*ReceivedMessage `protobuf:"bytes,1,rep,name=received_messages,json=receivedMessages" json:"received_messages,omitempty"` + ReceivedMessages []*ReceivedMessage `protobuf:"bytes,1,rep,name=received_messages,json=receivedMessages,proto3" json:"received_messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1490,7 +1676,7 @@ func (m *StreamingPullResponse) Reset() { *m = StreamingPullResponse{} } func (m *StreamingPullResponse) String() string { return proto.CompactTextString(m) } func (*StreamingPullResponse) ProtoMessage() {} func (*StreamingPullResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{25} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{28} } func (m *StreamingPullResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_StreamingPullResponse.Unmarshal(m, b) @@ -1517,14 +1703,17 @@ func (m *StreamingPullResponse) GetReceivedMessages() []*ReceivedMessage { return nil } -// Request for the `CreateSnapshot` method. +// Request for the `CreateSnapshot` method.

+// ALPHA: This feature is part of an alpha release. This API might be changed in +// backward-incompatible ways and is not recommended for production use. +// It is not subject to any SLA or deprecation policy. type CreateSnapshotRequest struct { // Optional user-provided name for this snapshot. // If the name is not provided in the request, the server will assign a random // name for this snapshot on the same project as the subscription. // Note that for REST API requests, you must specify a name. // Format is `projects/{project}/snapshots/{snap}`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The subscription whose backlog the snapshot retains. // Specifically, the created snapshot is guaranteed to retain: // (a) The existing backlog on the subscription. More precisely, this is @@ -1534,17 +1723,19 @@ type CreateSnapshotRequest struct { // (b) Any messages published to the subscription's topic following the // successful completion of the CreateSnapshot request. // Format is `projects/{project}/subscriptions/{sub}`. - Subscription string `protobuf:"bytes,2,opt,name=subscription" json:"subscription,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + Subscription string `protobuf:"bytes,2,opt,name=subscription,proto3" json:"subscription,omitempty"` + // User labels. + Labels map[string]string `protobuf:"bytes,3,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *CreateSnapshotRequest) Reset() { *m = CreateSnapshotRequest{} } func (m *CreateSnapshotRequest) String() string { return proto.CompactTextString(m) } func (*CreateSnapshotRequest) ProtoMessage() {} func (*CreateSnapshotRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{26} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{29} } func (m *CreateSnapshotRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_CreateSnapshotRequest.Unmarshal(m, b) @@ -1578,13 +1769,23 @@ func (m *CreateSnapshotRequest) GetSubscription() string { return "" } -// Request for the UpdateSnapshot method. +func (m *CreateSnapshotRequest) GetLabels() map[string]string { + if m != nil { + return m.Labels + } + return nil +} + +// Request for the UpdateSnapshot method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. type UpdateSnapshotRequest struct { - // The updated snpashot object. - Snapshot *Snapshot `protobuf:"bytes,1,opt,name=snapshot" json:"snapshot,omitempty"` + // The updated snapshot object. + Snapshot *Snapshot `protobuf:"bytes,1,opt,name=snapshot,proto3" json:"snapshot,omitempty"` // Indicates which fields in the provided snapshot to update. // Must be specified and non-empty. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1594,7 +1795,7 @@ func (m *UpdateSnapshotRequest) Reset() { *m = UpdateSnapshotRequest{} } func (m *UpdateSnapshotRequest) String() string { return proto.CompactTextString(m) } func (*UpdateSnapshotRequest) ProtoMessage() {} func (*UpdateSnapshotRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{27} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{30} } func (m *UpdateSnapshotRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_UpdateSnapshotRequest.Unmarshal(m, b) @@ -1628,12 +1829,15 @@ func (m *UpdateSnapshotRequest) GetUpdateMask() *field_mask.FieldMask { return nil } -// A snapshot resource. +// A snapshot resource.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. type Snapshot struct { // The name of the snapshot. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the topic from which this snapshot is retaining messages. - Topic string `protobuf:"bytes,2,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,2,opt,name=topic,proto3" json:"topic,omitempty"` // The snapshot is guaranteed to exist up until this time. // A newly-created snapshot expires no later than 7 days from the time of its // creation. Its exact lifetime is determined at creation by the existing @@ -1642,10 +1846,11 @@ type Snapshot struct { // For example, consider a subscription whose oldest unacked message is 3 days // old. If a snapshot is created from this subscription, the snapshot -- which // will always capture this 3-day-old backlog as long as the snapshot - // exists -- will expire in 4 days. - ExpireTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=expire_time,json=expireTime" json:"expire_time,omitempty"` + // exists -- will expire in 4 days. The service will refuse to create a + // snapshot that would expire in less than 1 hour after creation. + ExpireTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=expire_time,json=expireTime,proto3" json:"expire_time,omitempty"` // User labels. - Labels map[string]string `protobuf:"bytes,4,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,4,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1655,7 +1860,7 @@ func (m *Snapshot) Reset() { *m = Snapshot{} } func (m *Snapshot) String() string { return proto.CompactTextString(m) } func (*Snapshot) ProtoMessage() {} func (*Snapshot) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{28} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{31} } func (m *Snapshot) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Snapshot.Unmarshal(m, b) @@ -1703,17 +1908,64 @@ func (m *Snapshot) GetLabels() map[string]string { return nil } -// Request for the `ListSnapshots` method. +// Request for the GetSnapshot method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. +type GetSnapshotRequest struct { + // The name of the snapshot to get. + // Format is `projects/{project}/snapshots/{snap}`. + Snapshot string `protobuf:"bytes,1,opt,name=snapshot,proto3" json:"snapshot,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *GetSnapshotRequest) Reset() { *m = GetSnapshotRequest{} } +func (m *GetSnapshotRequest) String() string { return proto.CompactTextString(m) } +func (*GetSnapshotRequest) ProtoMessage() {} +func (*GetSnapshotRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{32} +} +func (m *GetSnapshotRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_GetSnapshotRequest.Unmarshal(m, b) +} +func (m *GetSnapshotRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_GetSnapshotRequest.Marshal(b, m, deterministic) +} +func (dst *GetSnapshotRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_GetSnapshotRequest.Merge(dst, src) +} +func (m *GetSnapshotRequest) XXX_Size() int { + return xxx_messageInfo_GetSnapshotRequest.Size(m) +} +func (m *GetSnapshotRequest) XXX_DiscardUnknown() { + xxx_messageInfo_GetSnapshotRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_GetSnapshotRequest proto.InternalMessageInfo + +func (m *GetSnapshotRequest) GetSnapshot() string { + if m != nil { + return m.Snapshot + } + return "" +} + +// Request for the `ListSnapshots` method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. type ListSnapshotsRequest struct { // The name of the cloud project that snapshots belong to. // Format is `projects/{project}`. - Project string `protobuf:"bytes,1,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,1,opt,name=project,proto3" json:"project,omitempty"` // Maximum number of snapshots to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last `ListSnapshotsResponse`; indicates that this // is a continuation of a prior `ListSnapshots` call, and that the system // should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1723,7 +1975,7 @@ func (m *ListSnapshotsRequest) Reset() { *m = ListSnapshotsRequest{} } func (m *ListSnapshotsRequest) String() string { return proto.CompactTextString(m) } func (*ListSnapshotsRequest) ProtoMessage() {} func (*ListSnapshotsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{29} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{33} } func (m *ListSnapshotsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSnapshotsRequest.Unmarshal(m, b) @@ -1764,13 +2016,16 @@ func (m *ListSnapshotsRequest) GetPageToken() string { return "" } -// Response for the `ListSnapshots` method. +// Response for the `ListSnapshots` method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. type ListSnapshotsResponse struct { // The resulting snapshots. - Snapshots []*Snapshot `protobuf:"bytes,1,rep,name=snapshots" json:"snapshots,omitempty"` + Snapshots []*Snapshot `protobuf:"bytes,1,rep,name=snapshots,proto3" json:"snapshots,omitempty"` // If not empty, indicates that there may be more snapshot that match the // request; this value should be passed in a new `ListSnapshotsRequest`. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1780,7 +2035,7 @@ func (m *ListSnapshotsResponse) Reset() { *m = ListSnapshotsResponse{} } func (m *ListSnapshotsResponse) String() string { return proto.CompactTextString(m) } func (*ListSnapshotsResponse) ProtoMessage() {} func (*ListSnapshotsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{30} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{34} } func (m *ListSnapshotsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ListSnapshotsResponse.Unmarshal(m, b) @@ -1814,11 +2069,14 @@ func (m *ListSnapshotsResponse) GetNextPageToken() string { return "" } -// Request for the `DeleteSnapshot` method. +// Request for the `DeleteSnapshot` method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. type DeleteSnapshotRequest struct { // The name of the snapshot to delete. // Format is `projects/{project}/snapshots/{snap}`. - Snapshot string `protobuf:"bytes,1,opt,name=snapshot" json:"snapshot,omitempty"` + Snapshot string `protobuf:"bytes,1,opt,name=snapshot,proto3" json:"snapshot,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1828,7 +2086,7 @@ func (m *DeleteSnapshotRequest) Reset() { *m = DeleteSnapshotRequest{} } func (m *DeleteSnapshotRequest) String() string { return proto.CompactTextString(m) } func (*DeleteSnapshotRequest) ProtoMessage() {} func (*DeleteSnapshotRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{31} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{35} } func (m *DeleteSnapshotRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_DeleteSnapshotRequest.Unmarshal(m, b) @@ -1855,10 +2113,13 @@ func (m *DeleteSnapshotRequest) GetSnapshot() string { return "" } -// Request for the `Seek` method. +// Request for the `Seek` method.

+// ALPHA: This feature is part of an alpha release. This API might be +// changed in backward-incompatible ways and is not recommended for production +// use. It is not subject to any SLA or deprecation policy. type SeekRequest struct { // The subscription to affect. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // Types that are valid to be assigned to Target: // *SeekRequest_Time // *SeekRequest_Snapshot @@ -1872,7 +2133,7 @@ func (m *SeekRequest) Reset() { *m = SeekRequest{} } func (m *SeekRequest) String() string { return proto.CompactTextString(m) } func (*SeekRequest) ProtoMessage() {} func (*SeekRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{32} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{36} } func (m *SeekRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SeekRequest.Unmarshal(m, b) @@ -1897,10 +2158,10 @@ type isSeekRequest_Target interface { } type SeekRequest_Time struct { - Time *timestamp.Timestamp `protobuf:"bytes,2,opt,name=time,oneof"` + Time *timestamp.Timestamp `protobuf:"bytes,2,opt,name=time,proto3,oneof"` } type SeekRequest_Snapshot struct { - Snapshot string `protobuf:"bytes,3,opt,name=snapshot,oneof"` + Snapshot string `protobuf:"bytes,3,opt,name=snapshot,proto3,oneof"` } func (*SeekRequest_Time) isSeekRequest_Target() {} @@ -2014,7 +2275,7 @@ func (m *SeekResponse) Reset() { *m = SeekResponse{} } func (m *SeekResponse) String() string { return proto.CompactTextString(m) } func (*SeekResponse) ProtoMessage() {} func (*SeekResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_pubsub_2c3603090b805ba2, []int{33} + return fileDescriptor_pubsub_ccb5b83827ebb62b, []int{37} } func (m *SeekResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SeekResponse.Unmarshal(m, b) @@ -2035,6 +2296,7 @@ func (m *SeekResponse) XXX_DiscardUnknown() { var xxx_messageInfo_SeekResponse proto.InternalMessageInfo func init() { + proto.RegisterType((*MessageStoragePolicy)(nil), "google.pubsub.v1.MessageStoragePolicy") proto.RegisterType((*Topic)(nil), "google.pubsub.v1.Topic") proto.RegisterMapType((map[string]string)(nil), "google.pubsub.v1.Topic.LabelsEntry") proto.RegisterType((*PubsubMessage)(nil), "google.pubsub.v1.PubsubMessage") @@ -2047,6 +2309,8 @@ func init() { proto.RegisterType((*ListTopicsResponse)(nil), "google.pubsub.v1.ListTopicsResponse") proto.RegisterType((*ListTopicSubscriptionsRequest)(nil), "google.pubsub.v1.ListTopicSubscriptionsRequest") proto.RegisterType((*ListTopicSubscriptionsResponse)(nil), "google.pubsub.v1.ListTopicSubscriptionsResponse") + proto.RegisterType((*ListTopicSnapshotsRequest)(nil), "google.pubsub.v1.ListTopicSnapshotsRequest") + proto.RegisterType((*ListTopicSnapshotsResponse)(nil), "google.pubsub.v1.ListTopicSnapshotsResponse") proto.RegisterType((*DeleteTopicRequest)(nil), "google.pubsub.v1.DeleteTopicRequest") proto.RegisterType((*Subscription)(nil), "google.pubsub.v1.Subscription") proto.RegisterMapType((map[string]string)(nil), "google.pubsub.v1.Subscription.LabelsEntry") @@ -2066,9 +2330,11 @@ func init() { proto.RegisterType((*StreamingPullRequest)(nil), "google.pubsub.v1.StreamingPullRequest") proto.RegisterType((*StreamingPullResponse)(nil), "google.pubsub.v1.StreamingPullResponse") proto.RegisterType((*CreateSnapshotRequest)(nil), "google.pubsub.v1.CreateSnapshotRequest") + proto.RegisterMapType((map[string]string)(nil), "google.pubsub.v1.CreateSnapshotRequest.LabelsEntry") proto.RegisterType((*UpdateSnapshotRequest)(nil), "google.pubsub.v1.UpdateSnapshotRequest") proto.RegisterType((*Snapshot)(nil), "google.pubsub.v1.Snapshot") proto.RegisterMapType((map[string]string)(nil), "google.pubsub.v1.Snapshot.LabelsEntry") + proto.RegisterType((*GetSnapshotRequest)(nil), "google.pubsub.v1.GetSnapshotRequest") proto.RegisterType((*ListSnapshotsRequest)(nil), "google.pubsub.v1.ListSnapshotsRequest") proto.RegisterType((*ListSnapshotsResponse)(nil), "google.pubsub.v1.ListSnapshotsResponse") proto.RegisterType((*DeleteSnapshotRequest)(nil), "google.pubsub.v1.DeleteSnapshotRequest") @@ -2084,11 +2350,345 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 +// PublisherClient is the client API for Publisher service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type PublisherClient interface { + // Creates the given topic with the given name. See the + // resource name rules. + CreateTopic(ctx context.Context, in *Topic, opts ...grpc.CallOption) (*Topic, error) + // Updates an existing topic. Note that certain properties of a + // topic are not modifiable. + UpdateTopic(ctx context.Context, in *UpdateTopicRequest, opts ...grpc.CallOption) (*Topic, error) + // Adds one or more messages to the topic. Returns `NOT_FOUND` if the topic + // does not exist. The message payload must not be empty; it must contain + // either a non-empty data field, or at least one attribute. + Publish(ctx context.Context, in *PublishRequest, opts ...grpc.CallOption) (*PublishResponse, error) + // Gets the configuration of a topic. + GetTopic(ctx context.Context, in *GetTopicRequest, opts ...grpc.CallOption) (*Topic, error) + // Lists matching topics. + ListTopics(ctx context.Context, in *ListTopicsRequest, opts ...grpc.CallOption) (*ListTopicsResponse, error) + // Lists the names of the subscriptions on this topic. + ListTopicSubscriptions(ctx context.Context, in *ListTopicSubscriptionsRequest, opts ...grpc.CallOption) (*ListTopicSubscriptionsResponse, error) + // Lists the names of the snapshots on this topic.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + ListTopicSnapshots(ctx context.Context, in *ListTopicSnapshotsRequest, opts ...grpc.CallOption) (*ListTopicSnapshotsResponse, error) + // Deletes the topic with the given name. Returns `NOT_FOUND` if the topic + // does not exist. After a topic is deleted, a new topic may be created with + // the same name; this is an entirely new topic with none of the old + // configuration or subscriptions. Existing subscriptions to this topic are + // not deleted, but their `topic` field is set to `_deleted-topic_`. + DeleteTopic(ctx context.Context, in *DeleteTopicRequest, opts ...grpc.CallOption) (*empty.Empty, error) +} + +type publisherClient struct { + cc *grpc.ClientConn +} + +func NewPublisherClient(cc *grpc.ClientConn) PublisherClient { + return &publisherClient{cc} +} + +func (c *publisherClient) CreateTopic(ctx context.Context, in *Topic, opts ...grpc.CallOption) (*Topic, error) { + out := new(Topic) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/CreateTopic", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) UpdateTopic(ctx context.Context, in *UpdateTopicRequest, opts ...grpc.CallOption) (*Topic, error) { + out := new(Topic) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/UpdateTopic", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) Publish(ctx context.Context, in *PublishRequest, opts ...grpc.CallOption) (*PublishResponse, error) { + out := new(PublishResponse) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/Publish", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) GetTopic(ctx context.Context, in *GetTopicRequest, opts ...grpc.CallOption) (*Topic, error) { + out := new(Topic) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/GetTopic", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) ListTopics(ctx context.Context, in *ListTopicsRequest, opts ...grpc.CallOption) (*ListTopicsResponse, error) { + out := new(ListTopicsResponse) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/ListTopics", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) ListTopicSubscriptions(ctx context.Context, in *ListTopicSubscriptionsRequest, opts ...grpc.CallOption) (*ListTopicSubscriptionsResponse, error) { + out := new(ListTopicSubscriptionsResponse) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/ListTopicSubscriptions", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) ListTopicSnapshots(ctx context.Context, in *ListTopicSnapshotsRequest, opts ...grpc.CallOption) (*ListTopicSnapshotsResponse, error) { + out := new(ListTopicSnapshotsResponse) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/ListTopicSnapshots", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +func (c *publisherClient) DeleteTopic(ctx context.Context, in *DeleteTopicRequest, opts ...grpc.CallOption) (*empty.Empty, error) { + out := new(empty.Empty) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/DeleteTopic", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + +// PublisherServer is the server API for Publisher service. +type PublisherServer interface { + // Creates the given topic with the given name. See the + // resource name rules. + CreateTopic(context.Context, *Topic) (*Topic, error) + // Updates an existing topic. Note that certain properties of a + // topic are not modifiable. + UpdateTopic(context.Context, *UpdateTopicRequest) (*Topic, error) + // Adds one or more messages to the topic. Returns `NOT_FOUND` if the topic + // does not exist. The message payload must not be empty; it must contain + // either a non-empty data field, or at least one attribute. + Publish(context.Context, *PublishRequest) (*PublishResponse, error) + // Gets the configuration of a topic. + GetTopic(context.Context, *GetTopicRequest) (*Topic, error) + // Lists matching topics. + ListTopics(context.Context, *ListTopicsRequest) (*ListTopicsResponse, error) + // Lists the names of the subscriptions on this topic. + ListTopicSubscriptions(context.Context, *ListTopicSubscriptionsRequest) (*ListTopicSubscriptionsResponse, error) + // Lists the names of the snapshots on this topic.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + ListTopicSnapshots(context.Context, *ListTopicSnapshotsRequest) (*ListTopicSnapshotsResponse, error) + // Deletes the topic with the given name. Returns `NOT_FOUND` if the topic + // does not exist. After a topic is deleted, a new topic may be created with + // the same name; this is an entirely new topic with none of the old + // configuration or subscriptions. Existing subscriptions to this topic are + // not deleted, but their `topic` field is set to `_deleted-topic_`. + DeleteTopic(context.Context, *DeleteTopicRequest) (*empty.Empty, error) +} + +func RegisterPublisherServer(s *grpc.Server, srv PublisherServer) { + s.RegisterService(&_Publisher_serviceDesc, srv) +} + +func _Publisher_CreateTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(Topic) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).CreateTopic(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/CreateTopic", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).CreateTopic(ctx, req.(*Topic)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_UpdateTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(UpdateTopicRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).UpdateTopic(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/UpdateTopic", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).UpdateTopic(ctx, req.(*UpdateTopicRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_Publish_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(PublishRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).Publish(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/Publish", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).Publish(ctx, req.(*PublishRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_GetTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetTopicRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).GetTopic(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/GetTopic", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).GetTopic(ctx, req.(*GetTopicRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_ListTopics_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListTopicsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).ListTopics(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/ListTopics", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).ListTopics(ctx, req.(*ListTopicsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_ListTopicSubscriptions_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListTopicSubscriptionsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).ListTopicSubscriptions(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/ListTopicSubscriptions", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).ListTopicSubscriptions(ctx, req.(*ListTopicSubscriptionsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_ListTopicSnapshots_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(ListTopicSnapshotsRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).ListTopicSnapshots(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/ListTopicSnapshots", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).ListTopicSnapshots(ctx, req.(*ListTopicSnapshotsRequest)) + } + return interceptor(ctx, in, info, handler) +} + +func _Publisher_DeleteTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(DeleteTopicRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(PublisherServer).DeleteTopic(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Publisher/DeleteTopic", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(PublisherServer).DeleteTopic(ctx, req.(*DeleteTopicRequest)) + } + return interceptor(ctx, in, info, handler) +} + +var _Publisher_serviceDesc = grpc.ServiceDesc{ + ServiceName: "google.pubsub.v1.Publisher", + HandlerType: (*PublisherServer)(nil), + Methods: []grpc.MethodDesc{ + { + MethodName: "CreateTopic", + Handler: _Publisher_CreateTopic_Handler, + }, + { + MethodName: "UpdateTopic", + Handler: _Publisher_UpdateTopic_Handler, + }, + { + MethodName: "Publish", + Handler: _Publisher_Publish_Handler, + }, + { + MethodName: "GetTopic", + Handler: _Publisher_GetTopic_Handler, + }, + { + MethodName: "ListTopics", + Handler: _Publisher_ListTopics_Handler, + }, + { + MethodName: "ListTopicSubscriptions", + Handler: _Publisher_ListTopicSubscriptions_Handler, + }, + { + MethodName: "ListTopicSnapshots", + Handler: _Publisher_ListTopicSnapshots_Handler, + }, + { + MethodName: "DeleteTopic", + Handler: _Publisher_DeleteTopic_Handler, + }, + }, + Streams: []grpc.StreamDesc{}, + Metadata: "google/pubsub/v1/pubsub.proto", +} + // SubscriberClient is the client API for Subscriber service. // // For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type SubscriberClient interface { - // Creates a subscription to a given topic. + // Creates a subscription to a given topic. See the + // resource name rules. // If the subscription already exists, returns `ALREADY_EXISTS`. // If the corresponding topic doesn't exist, returns `NOT_FOUND`. // @@ -2103,10 +2703,6 @@ type SubscriberClient interface { GetSubscription(ctx context.Context, in *GetSubscriptionRequest, opts ...grpc.CallOption) (*Subscription, error) // Updates an existing subscription. Note that certain properties of a // subscription, such as its topic, are not modifiable. - // NOTE: The style guide requires body: "subscription" instead of body: "*". - // Keeping the latter for internal consistency in V1, however it should be - // corrected in V2. See - // https://cloud.google.com/apis/design/standard_methods#update for details. UpdateSubscription(ctx context.Context, in *UpdateSubscriptionRequest, opts ...grpc.CallOption) (*Subscription, error) // Lists matching subscriptions. ListSubscriptions(ctx context.Context, in *ListSubscriptionsRequest, opts ...grpc.CallOption) (*ListSubscriptionsResponse, error) @@ -2135,18 +2731,13 @@ type SubscriberClient interface { // there are too many concurrent pull requests pending for the given // subscription. Pull(ctx context.Context, in *PullRequest, opts ...grpc.CallOption) (*PullResponse, error) - // (EXPERIMENTAL) StreamingPull is an experimental feature. This RPC will - // respond with UNIMPLEMENTED errors unless you have been invited to test - // this feature. Contact cloud-pubsub@google.com with any questions. - // // Establishes a stream with the server, which sends messages down to the // client. The client streams acknowledgements and ack deadline modifications // back to the server. The server will close the stream and return the status - // on any error. The server may close the stream with status `OK` to reassign - // server-side resources, in which case, the client should re-establish the - // stream. `UNAVAILABLE` may also be returned in the case of a transient error - // (e.g., a server restart). These should also be retried by the client. Flow - // control can be achieved by configuring the underlying RPC channel. + // on any error. The server may close the stream with status `UNAVAILABLE` to + // reassign server-side resources, in which case, the client should + // re-establish the stream. Flow control can be achieved by configuring the + // underlying RPC channel. StreamingPull(ctx context.Context, opts ...grpc.CallOption) (Subscriber_StreamingPullClient, error) // Modifies the `PushConfig` for a specified subscription. // @@ -2155,33 +2746,52 @@ type SubscriberClient interface { // attributes of a push subscription. Messages will accumulate for delivery // continuously through the call regardless of changes to the `PushConfig`. ModifyPushConfig(ctx context.Context, in *ModifyPushConfigRequest, opts ...grpc.CallOption) (*empty.Empty, error) - // Lists the existing snapshots. + // Gets the configuration details of a snapshot.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + GetSnapshot(ctx context.Context, in *GetSnapshotRequest, opts ...grpc.CallOption) (*Snapshot, error) + // Lists the existing snapshots.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. ListSnapshots(ctx context.Context, in *ListSnapshotsRequest, opts ...grpc.CallOption) (*ListSnapshotsResponse, error) - // Creates a snapshot from the requested subscription. + // Creates a snapshot from the requested subscription.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. // If the snapshot already exists, returns `ALREADY_EXISTS`. // If the requested subscription doesn't exist, returns `NOT_FOUND`. - // - // If the name is not provided in the request, the server will assign a random + // If the backlog in the subscription is too old -- and the resulting snapshot + // would expire in less than 1 hour -- then `FAILED_PRECONDITION` is returned. + // See also the `Snapshot.expire_time` field. If the name is not provided in + // the request, the server will assign a random // name for this snapshot on the same project as the subscription, conforming - // to the - // [resource name format](https://cloud.google.com/pubsub/docs/overview#names). - // The generated name is populated in the returned Snapshot object. - // Note that for REST API requests, you must specify a name in the request. + // to the [resource name format](https://cloud.google.com/pubsub/docs/overview#names). + // The generated + // name is populated in the returned Snapshot object. Note that for REST API + // requests, you must specify a name in the request. CreateSnapshot(ctx context.Context, in *CreateSnapshotRequest, opts ...grpc.CallOption) (*Snapshot, error) - // Updates an existing snapshot. Note that certain properties of a snapshot - // are not modifiable. - // NOTE: The style guide requires body: "snapshot" instead of body: "*". - // Keeping the latter for internal consistency in V1, however it should be - // corrected in V2. See - // https://cloud.google.com/apis/design/standard_methods#update for details. + // Updates an existing snapshot.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + // Note that certain properties of a snapshot are not modifiable. UpdateSnapshot(ctx context.Context, in *UpdateSnapshotRequest, opts ...grpc.CallOption) (*Snapshot, error) - // Removes an existing snapshot. All messages retained in the snapshot + // Removes an existing snapshot.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + // When the snapshot is deleted, all messages retained in the snapshot // are immediately dropped. After a snapshot is deleted, a new one may be // created with the same name, but the new one has no association with the old // snapshot or its subscription, unless the same subscription is specified. DeleteSnapshot(ctx context.Context, in *DeleteSnapshotRequest, opts ...grpc.CallOption) (*empty.Empty, error) // Seeks an existing subscription to a point in time or to a given snapshot, - // whichever is provided in the request. + // whichever is provided in the request.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. Seek(ctx context.Context, in *SeekRequest, opts ...grpc.CallOption) (*SeekResponse, error) } @@ -2305,6 +2915,15 @@ func (c *subscriberClient) ModifyPushConfig(ctx context.Context, in *ModifyPushC return out, nil } +func (c *subscriberClient) GetSnapshot(ctx context.Context, in *GetSnapshotRequest, opts ...grpc.CallOption) (*Snapshot, error) { + out := new(Snapshot) + err := c.cc.Invoke(ctx, "/google.pubsub.v1.Subscriber/GetSnapshot", in, out, opts...) + if err != nil { + return nil, err + } + return out, nil +} + func (c *subscriberClient) ListSnapshots(ctx context.Context, in *ListSnapshotsRequest, opts ...grpc.CallOption) (*ListSnapshotsResponse, error) { out := new(ListSnapshotsResponse) err := c.cc.Invoke(ctx, "/google.pubsub.v1.Subscriber/ListSnapshots", in, out, opts...) @@ -2350,10 +2969,10 @@ func (c *subscriberClient) Seek(ctx context.Context, in *SeekRequest, opts ...gr return out, nil } -// Server API for Subscriber service - +// SubscriberServer is the server API for Subscriber service. type SubscriberServer interface { - // Creates a subscription to a given topic. + // Creates a subscription to a given topic. See the + // resource name rules. // If the subscription already exists, returns `ALREADY_EXISTS`. // If the corresponding topic doesn't exist, returns `NOT_FOUND`. // @@ -2368,10 +2987,6 @@ type SubscriberServer interface { GetSubscription(context.Context, *GetSubscriptionRequest) (*Subscription, error) // Updates an existing subscription. Note that certain properties of a // subscription, such as its topic, are not modifiable. - // NOTE: The style guide requires body: "subscription" instead of body: "*". - // Keeping the latter for internal consistency in V1, however it should be - // corrected in V2. See - // https://cloud.google.com/apis/design/standard_methods#update for details. UpdateSubscription(context.Context, *UpdateSubscriptionRequest) (*Subscription, error) // Lists matching subscriptions. ListSubscriptions(context.Context, *ListSubscriptionsRequest) (*ListSubscriptionsResponse, error) @@ -2400,18 +3015,13 @@ type SubscriberServer interface { // there are too many concurrent pull requests pending for the given // subscription. Pull(context.Context, *PullRequest) (*PullResponse, error) - // (EXPERIMENTAL) StreamingPull is an experimental feature. This RPC will - // respond with UNIMPLEMENTED errors unless you have been invited to test - // this feature. Contact cloud-pubsub@google.com with any questions. - // // Establishes a stream with the server, which sends messages down to the // client. The client streams acknowledgements and ack deadline modifications // back to the server. The server will close the stream and return the status - // on any error. The server may close the stream with status `OK` to reassign - // server-side resources, in which case, the client should re-establish the - // stream. `UNAVAILABLE` may also be returned in the case of a transient error - // (e.g., a server restart). These should also be retried by the client. Flow - // control can be achieved by configuring the underlying RPC channel. + // on any error. The server may close the stream with status `UNAVAILABLE` to + // reassign server-side resources, in which case, the client should + // re-establish the stream. Flow control can be achieved by configuring the + // underlying RPC channel. StreamingPull(Subscriber_StreamingPullServer) error // Modifies the `PushConfig` for a specified subscription. // @@ -2420,33 +3030,52 @@ type SubscriberServer interface { // attributes of a push subscription. Messages will accumulate for delivery // continuously through the call regardless of changes to the `PushConfig`. ModifyPushConfig(context.Context, *ModifyPushConfigRequest) (*empty.Empty, error) - // Lists the existing snapshots. + // Gets the configuration details of a snapshot.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + GetSnapshot(context.Context, *GetSnapshotRequest) (*Snapshot, error) + // Lists the existing snapshots.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. ListSnapshots(context.Context, *ListSnapshotsRequest) (*ListSnapshotsResponse, error) - // Creates a snapshot from the requested subscription. + // Creates a snapshot from the requested subscription.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. // If the snapshot already exists, returns `ALREADY_EXISTS`. // If the requested subscription doesn't exist, returns `NOT_FOUND`. - // - // If the name is not provided in the request, the server will assign a random + // If the backlog in the subscription is too old -- and the resulting snapshot + // would expire in less than 1 hour -- then `FAILED_PRECONDITION` is returned. + // See also the `Snapshot.expire_time` field. If the name is not provided in + // the request, the server will assign a random // name for this snapshot on the same project as the subscription, conforming - // to the - // [resource name format](https://cloud.google.com/pubsub/docs/overview#names). - // The generated name is populated in the returned Snapshot object. - // Note that for REST API requests, you must specify a name in the request. + // to the [resource name format](https://cloud.google.com/pubsub/docs/overview#names). + // The generated + // name is populated in the returned Snapshot object. Note that for REST API + // requests, you must specify a name in the request. CreateSnapshot(context.Context, *CreateSnapshotRequest) (*Snapshot, error) - // Updates an existing snapshot. Note that certain properties of a snapshot - // are not modifiable. - // NOTE: The style guide requires body: "snapshot" instead of body: "*". - // Keeping the latter for internal consistency in V1, however it should be - // corrected in V2. See - // https://cloud.google.com/apis/design/standard_methods#update for details. + // Updates an existing snapshot.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + // Note that certain properties of a snapshot are not modifiable. UpdateSnapshot(context.Context, *UpdateSnapshotRequest) (*Snapshot, error) - // Removes an existing snapshot. All messages retained in the snapshot + // Removes an existing snapshot.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. + // When the snapshot is deleted, all messages retained in the snapshot // are immediately dropped. After a snapshot is deleted, a new one may be // created with the same name, but the new one has no association with the old // snapshot or its subscription, unless the same subscription is specified. DeleteSnapshot(context.Context, *DeleteSnapshotRequest) (*empty.Empty, error) // Seeks an existing subscription to a point in time or to a given snapshot, - // whichever is provided in the request. + // whichever is provided in the request.

+ // ALPHA: This feature is part of an alpha release. This API might be + // changed in backward-incompatible ways and is not recommended for production + // use. It is not subject to any SLA or deprecation policy. Seek(context.Context, *SeekRequest) (*SeekResponse, error) } @@ -2642,6 +3271,24 @@ func _Subscriber_ModifyPushConfig_Handler(srv interface{}, ctx context.Context, return interceptor(ctx, in, info, handler) } +func _Subscriber_GetSnapshot_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { + in := new(GetSnapshotRequest) + if err := dec(in); err != nil { + return nil, err + } + if interceptor == nil { + return srv.(SubscriberServer).GetSnapshot(ctx, in) + } + info := &grpc.UnaryServerInfo{ + Server: srv, + FullMethod: "/google.pubsub.v1.Subscriber/GetSnapshot", + } + handler := func(ctx context.Context, req interface{}) (interface{}, error) { + return srv.(SubscriberServer).GetSnapshot(ctx, req.(*GetSnapshotRequest)) + } + return interceptor(ctx, in, info, handler) +} + func _Subscriber_ListSnapshots_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { in := new(ListSnapshotsRequest) if err := dec(in); err != nil { @@ -2772,6 +3419,10 @@ var _Subscriber_serviceDesc = grpc.ServiceDesc{ MethodName: "ModifyPushConfig", Handler: _Subscriber_ModifyPushConfig_Handler, }, + { + MethodName: "GetSnapshot", + Handler: _Subscriber_GetSnapshot_Handler, + }, { MethodName: "ListSnapshots", Handler: _Subscriber_ListSnapshots_Handler, @@ -2804,436 +3455,148 @@ var _Subscriber_serviceDesc = grpc.ServiceDesc{ Metadata: "google/pubsub/v1/pubsub.proto", } -// PublisherClient is the client API for Publisher service. -// -// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. -type PublisherClient interface { - // Creates the given topic with the given name. - CreateTopic(ctx context.Context, in *Topic, opts ...grpc.CallOption) (*Topic, error) - // Updates an existing topic. Note that certain properties of a topic are not - // modifiable. Options settings follow the style guide: - // NOTE: The style guide requires body: "topic" instead of body: "*". - // Keeping the latter for internal consistency in V1, however it should be - // corrected in V2. See - // https://cloud.google.com/apis/design/standard_methods#update for details. - UpdateTopic(ctx context.Context, in *UpdateTopicRequest, opts ...grpc.CallOption) (*Topic, error) - // Adds one or more messages to the topic. Returns `NOT_FOUND` if the topic - // does not exist. The message payload must not be empty; it must contain - // either a non-empty data field, or at least one attribute. - Publish(ctx context.Context, in *PublishRequest, opts ...grpc.CallOption) (*PublishResponse, error) - // Gets the configuration of a topic. - GetTopic(ctx context.Context, in *GetTopicRequest, opts ...grpc.CallOption) (*Topic, error) - // Lists matching topics. - ListTopics(ctx context.Context, in *ListTopicsRequest, opts ...grpc.CallOption) (*ListTopicsResponse, error) - // Lists the name of the subscriptions for this topic. - ListTopicSubscriptions(ctx context.Context, in *ListTopicSubscriptionsRequest, opts ...grpc.CallOption) (*ListTopicSubscriptionsResponse, error) - // Deletes the topic with the given name. Returns `NOT_FOUND` if the topic - // does not exist. After a topic is deleted, a new topic may be created with - // the same name; this is an entirely new topic with none of the old - // configuration or subscriptions. Existing subscriptions to this topic are - // not deleted, but their `topic` field is set to `_deleted-topic_`. - DeleteTopic(ctx context.Context, in *DeleteTopicRequest, opts ...grpc.CallOption) (*empty.Empty, error) -} - -type publisherClient struct { - cc *grpc.ClientConn -} - -func NewPublisherClient(cc *grpc.ClientConn) PublisherClient { - return &publisherClient{cc} -} - -func (c *publisherClient) CreateTopic(ctx context.Context, in *Topic, opts ...grpc.CallOption) (*Topic, error) { - out := new(Topic) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/CreateTopic", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *publisherClient) UpdateTopic(ctx context.Context, in *UpdateTopicRequest, opts ...grpc.CallOption) (*Topic, error) { - out := new(Topic) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/UpdateTopic", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *publisherClient) Publish(ctx context.Context, in *PublishRequest, opts ...grpc.CallOption) (*PublishResponse, error) { - out := new(PublishResponse) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/Publish", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *publisherClient) GetTopic(ctx context.Context, in *GetTopicRequest, opts ...grpc.CallOption) (*Topic, error) { - out := new(Topic) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/GetTopic", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *publisherClient) ListTopics(ctx context.Context, in *ListTopicsRequest, opts ...grpc.CallOption) (*ListTopicsResponse, error) { - out := new(ListTopicsResponse) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/ListTopics", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *publisherClient) ListTopicSubscriptions(ctx context.Context, in *ListTopicSubscriptionsRequest, opts ...grpc.CallOption) (*ListTopicSubscriptionsResponse, error) { - out := new(ListTopicSubscriptionsResponse) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/ListTopicSubscriptions", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -func (c *publisherClient) DeleteTopic(ctx context.Context, in *DeleteTopicRequest, opts ...grpc.CallOption) (*empty.Empty, error) { - out := new(empty.Empty) - err := c.cc.Invoke(ctx, "/google.pubsub.v1.Publisher/DeleteTopic", in, out, opts...) - if err != nil { - return nil, err - } - return out, nil -} - -// Server API for Publisher service - -type PublisherServer interface { - // Creates the given topic with the given name. - CreateTopic(context.Context, *Topic) (*Topic, error) - // Updates an existing topic. Note that certain properties of a topic are not - // modifiable. Options settings follow the style guide: - // NOTE: The style guide requires body: "topic" instead of body: "*". - // Keeping the latter for internal consistency in V1, however it should be - // corrected in V2. See - // https://cloud.google.com/apis/design/standard_methods#update for details. - UpdateTopic(context.Context, *UpdateTopicRequest) (*Topic, error) - // Adds one or more messages to the topic. Returns `NOT_FOUND` if the topic - // does not exist. The message payload must not be empty; it must contain - // either a non-empty data field, or at least one attribute. - Publish(context.Context, *PublishRequest) (*PublishResponse, error) - // Gets the configuration of a topic. - GetTopic(context.Context, *GetTopicRequest) (*Topic, error) - // Lists matching topics. - ListTopics(context.Context, *ListTopicsRequest) (*ListTopicsResponse, error) - // Lists the name of the subscriptions for this topic. - ListTopicSubscriptions(context.Context, *ListTopicSubscriptionsRequest) (*ListTopicSubscriptionsResponse, error) - // Deletes the topic with the given name. Returns `NOT_FOUND` if the topic - // does not exist. After a topic is deleted, a new topic may be created with - // the same name; this is an entirely new topic with none of the old - // configuration or subscriptions. Existing subscriptions to this topic are - // not deleted, but their `topic` field is set to `_deleted-topic_`. - DeleteTopic(context.Context, *DeleteTopicRequest) (*empty.Empty, error) -} - -func RegisterPublisherServer(s *grpc.Server, srv PublisherServer) { - s.RegisterService(&_Publisher_serviceDesc, srv) -} - -func _Publisher_CreateTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(Topic) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).CreateTopic(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/CreateTopic", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).CreateTopic(ctx, req.(*Topic)) - } - return interceptor(ctx, in, info, handler) -} - -func _Publisher_UpdateTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(UpdateTopicRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).UpdateTopic(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/UpdateTopic", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).UpdateTopic(ctx, req.(*UpdateTopicRequest)) - } - return interceptor(ctx, in, info, handler) -} - -func _Publisher_Publish_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(PublishRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).Publish(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/Publish", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).Publish(ctx, req.(*PublishRequest)) - } - return interceptor(ctx, in, info, handler) -} - -func _Publisher_GetTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(GetTopicRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).GetTopic(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/GetTopic", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).GetTopic(ctx, req.(*GetTopicRequest)) - } - return interceptor(ctx, in, info, handler) -} - -func _Publisher_ListTopics_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(ListTopicsRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).ListTopics(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/ListTopics", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).ListTopics(ctx, req.(*ListTopicsRequest)) - } - return interceptor(ctx, in, info, handler) -} - -func _Publisher_ListTopicSubscriptions_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(ListTopicSubscriptionsRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).ListTopicSubscriptions(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/ListTopicSubscriptions", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).ListTopicSubscriptions(ctx, req.(*ListTopicSubscriptionsRequest)) - } - return interceptor(ctx, in, info, handler) -} - -func _Publisher_DeleteTopic_Handler(srv interface{}, ctx context.Context, dec func(interface{}) error, interceptor grpc.UnaryServerInterceptor) (interface{}, error) { - in := new(DeleteTopicRequest) - if err := dec(in); err != nil { - return nil, err - } - if interceptor == nil { - return srv.(PublisherServer).DeleteTopic(ctx, in) - } - info := &grpc.UnaryServerInfo{ - Server: srv, - FullMethod: "/google.pubsub.v1.Publisher/DeleteTopic", - } - handler := func(ctx context.Context, req interface{}) (interface{}, error) { - return srv.(PublisherServer).DeleteTopic(ctx, req.(*DeleteTopicRequest)) - } - return interceptor(ctx, in, info, handler) -} - -var _Publisher_serviceDesc = grpc.ServiceDesc{ - ServiceName: "google.pubsub.v1.Publisher", - HandlerType: (*PublisherServer)(nil), - Methods: []grpc.MethodDesc{ - { - MethodName: "CreateTopic", - Handler: _Publisher_CreateTopic_Handler, - }, - { - MethodName: "UpdateTopic", - Handler: _Publisher_UpdateTopic_Handler, - }, - { - MethodName: "Publish", - Handler: _Publisher_Publish_Handler, - }, - { - MethodName: "GetTopic", - Handler: _Publisher_GetTopic_Handler, - }, - { - MethodName: "ListTopics", - Handler: _Publisher_ListTopics_Handler, - }, - { - MethodName: "ListTopicSubscriptions", - Handler: _Publisher_ListTopicSubscriptions_Handler, - }, - { - MethodName: "DeleteTopic", - Handler: _Publisher_DeleteTopic_Handler, - }, - }, - Streams: []grpc.StreamDesc{}, - Metadata: "google/pubsub/v1/pubsub.proto", -} - func init() { - proto.RegisterFile("google/pubsub/v1/pubsub.proto", fileDescriptor_pubsub_2c3603090b805ba2) + proto.RegisterFile("google/pubsub/v1/pubsub.proto", fileDescriptor_pubsub_ccb5b83827ebb62b) } -var fileDescriptor_pubsub_2c3603090b805ba2 = []byte{ - // 2026 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x59, 0xdb, 0x6f, 0xdb, 0xd6, - 0x19, 0xef, 0x91, 0x6f, 0xf2, 0x47, 0xdf, 0x72, 0x66, 0x27, 0x0a, 0x73, 0xb3, 0x19, 0x37, 0x56, - 0xd4, 0x44, 0xb2, 0x55, 0x34, 0x6b, 0xe2, 0x39, 0x85, 0x1d, 0x67, 0x69, 0x86, 0x64, 0xf5, 0xe8, - 0xac, 0x03, 0x86, 0x60, 0x02, 0x25, 0x9e, 0x28, 0xac, 0x24, 0x92, 0x25, 0x29, 0x2f, 0xee, 0x16, - 0xa0, 0x6b, 0x87, 0x01, 0xc3, 0xf2, 0xb0, 0xb6, 0x6f, 0x43, 0x1f, 0x06, 0xec, 0x6d, 0x8f, 0x03, - 0xf6, 0xba, 0x3f, 0x60, 0xaf, 0xfb, 0x17, 0xf6, 0xb8, 0xf7, 0xed, 0x71, 0x38, 0x17, 0x52, 0xbc, - 0x1c, 0x4a, 0x96, 0xd3, 0xbc, 0x91, 0xe7, 0xfb, 0xce, 0xf9, 0x7e, 0xdf, 0xfd, 0x23, 0x0f, 0x5c, - 0x6a, 0x3b, 0x4e, 0xbb, 0x4b, 0x6a, 0x6e, 0xbf, 0xe9, 0xf7, 0x9b, 0xb5, 0xa3, 0x2d, 0xf1, 0x54, - 0x75, 0x3d, 0x27, 0x70, 0xf0, 0x12, 0x27, 0x57, 0xc5, 0xe2, 0xd1, 0x96, 0x7a, 0x51, 0x6c, 0x30, - 0x5c, 0xab, 0x66, 0xd8, 0xb6, 0x13, 0x18, 0x81, 0xe5, 0xd8, 0x3e, 0xe7, 0x57, 0x2f, 0x87, 0xc7, - 0xd1, 0xb7, 0x66, 0xff, 0x59, 0xcd, 0xec, 0x7b, 0x8c, 0x41, 0xd0, 0x2f, 0xa4, 0xe9, 0xa4, 0xe7, - 0x06, 0xc7, 0x82, 0xb8, 0x9a, 0x26, 0x3e, 0xb3, 0x48, 0xd7, 0x6c, 0xf4, 0x0c, 0xbf, 0x23, 0x38, - 0xae, 0xa4, 0x39, 0x02, 0xab, 0x47, 0xfc, 0xc0, 0xe8, 0xb9, 0x9c, 0x41, 0xfb, 0x06, 0xc1, 0xd4, - 0x13, 0xc7, 0xb5, 0x5a, 0x18, 0xc3, 0xa4, 0x6d, 0xf4, 0x48, 0x09, 0xad, 0xa2, 0xf2, 0xac, 0xce, - 0x9e, 0xf1, 0x36, 0x4c, 0x77, 0x8d, 0x26, 0xe9, 0xfa, 0xa5, 0xc2, 0xea, 0x44, 0x59, 0xa9, 0x5f, - 0xad, 0xa6, 0xd5, 0xab, 0xb2, 0xcd, 0xd5, 0x47, 0x8c, 0xeb, 0xbe, 0x1d, 0x78, 0xc7, 0xba, 0xd8, - 0xa2, 0xde, 0x06, 0x25, 0xb6, 0x8c, 0x97, 0x60, 0xa2, 0x43, 0x8e, 0xc5, 0xf1, 0xf4, 0x11, 0x2f, - 0xc3, 0xd4, 0x91, 0xd1, 0xed, 0x93, 0x52, 0x81, 0xad, 0xf1, 0x97, 0x3b, 0x85, 0xf7, 0x91, 0xf6, - 0x55, 0x01, 0xe6, 0x0f, 0x98, 0x88, 0xc7, 0xc4, 0xf7, 0x8d, 0x36, 0xa1, 0xe8, 0x4c, 0x23, 0x30, - 0xd8, 0xf6, 0x39, 0x9d, 0x3d, 0xe3, 0x8f, 0x00, 0x8c, 0x20, 0xf0, 0xac, 0x66, 0x3f, 0x20, 0x21, - 0xc2, 0x5a, 0x16, 0x61, 0xe2, 0xa0, 0xea, 0x6e, 0xb4, 0x83, 0xa3, 0x8d, 0x1d, 0x81, 0x2f, 0x01, - 0xf4, 0x38, 0x5b, 0xc3, 0x32, 0x4b, 0x13, 0x0c, 0xd5, 0xac, 0x58, 0x79, 0x68, 0xe2, 0x1d, 0x98, - 0x73, 0xfb, 0xcd, 0xae, 0xe5, 0x3f, 0x6f, 0x50, 0x33, 0x96, 0x26, 0x57, 0x51, 0x59, 0xa9, 0xab, - 0x91, 0x44, 0x61, 0xe3, 0xea, 0x93, 0xd0, 0xc6, 0xba, 0x22, 0xf8, 0xe9, 0x8a, 0xba, 0x03, 0x8b, - 0x29, 0xe1, 0x63, 0xd9, 0x64, 0x03, 0x16, 0x1f, 0x90, 0x80, 0x99, 0x5b, 0x27, 0x9f, 0xf6, 0x89, - 0x1f, 0x50, 0xe6, 0x80, 0xbe, 0x8b, 0x03, 0xf8, 0x8b, 0xf6, 0x39, 0x02, 0xfc, 0x53, 0xd7, 0x34, - 0x02, 0x92, 0x60, 0xbe, 0x19, 0x67, 0x56, 0xea, 0xe7, 0x72, 0x5c, 0x29, 0x4e, 0xc1, 0xdb, 0xa0, - 0xf4, 0xd9, 0x21, 0x2c, 0x9c, 0x18, 0x1c, 0x99, 0xae, 0x3f, 0xa4, 0x11, 0xf7, 0xd8, 0xf0, 0x3b, - 0x3a, 0x70, 0x76, 0xfa, 0xac, 0xb5, 0x60, 0xe1, 0x80, 0x6b, 0x3e, 0x14, 0x2a, 0xde, 0x86, 0xa2, - 0x30, 0x6f, 0xe8, 0xbf, 0x2b, 0x23, 0xfc, 0xa7, 0x47, 0x1b, 0xb4, 0x3a, 0x2c, 0x46, 0x42, 0x7c, - 0xd7, 0xb1, 0x7d, 0x82, 0xaf, 0x80, 0x32, 0x70, 0xa0, 0x5f, 0x42, 0xab, 0x13, 0xe5, 0x59, 0x1d, - 0x22, 0x0f, 0xfa, 0x9a, 0x05, 0x67, 0x1e, 0x59, 0x3e, 0xb7, 0xa2, 0x1f, 0x62, 0x2b, 0xc1, 0x8c, - 0xeb, 0x39, 0x9f, 0x90, 0x56, 0x20, 0xd0, 0x85, 0xaf, 0xf8, 0x02, 0xcc, 0xba, 0xf4, 0x30, 0xdf, - 0xfa, 0x8c, 0x7b, 0x64, 0x4a, 0x2f, 0xd2, 0x85, 0x43, 0xeb, 0x33, 0x42, 0xa3, 0x85, 0x11, 0x03, - 0xa7, 0x43, 0xec, 0x30, 0x5a, 0xe8, 0xca, 0x13, 0xba, 0xa0, 0xf5, 0x00, 0xc7, 0x45, 0x09, 0x84, - 0x35, 0x98, 0x66, 0xaa, 0x73, 0x70, 0x43, 0xdc, 0x20, 0xd8, 0xf0, 0x35, 0x58, 0xb4, 0xc9, 0x8b, - 0xa0, 0x11, 0x13, 0xc5, 0x43, 0x63, 0x9e, 0x2e, 0x1f, 0x44, 0xe2, 0x3e, 0x85, 0x4b, 0x91, 0xb8, - 0xc3, 0x7e, 0xd3, 0x6f, 0x79, 0x96, 0xcb, 0x0a, 0xcd, 0x70, 0x0f, 0xbc, 0x8e, 0x86, 0x36, 0x5c, - 0xce, 0x13, 0x29, 0xb4, 0x5d, 0x87, 0x79, 0x3f, 0x4e, 0x10, 0x1e, 0x49, 0x2e, 0x9e, 0x58, 0xc5, - 0x0a, 0xe0, 0x7d, 0xd2, 0x25, 0xa9, 0xb8, 0x96, 0x27, 0xc1, 0xdf, 0x27, 0x60, 0x2e, 0x8e, 0x49, - 0x5a, 0xde, 0xa2, 0xad, 0x85, 0xb8, 0x49, 0x76, 0x40, 0x71, 0xfb, 0xfe, 0xf3, 0x46, 0xcb, 0xb1, - 0x9f, 0x59, 0x6d, 0x91, 0xe5, 0x17, 0x65, 0x71, 0xe9, 0x3f, 0xbf, 0xc7, 0x78, 0x74, 0x70, 0xa3, - 0x67, 0xbc, 0x09, 0xcb, 0x46, 0xab, 0xd3, 0x30, 0x89, 0x61, 0x76, 0x2d, 0x9b, 0x34, 0x7c, 0xd2, - 0x72, 0x6c, 0xd3, 0x2f, 0x4d, 0x31, 0xe3, 0x62, 0xa3, 0xd5, 0xd9, 0x17, 0xa4, 0x43, 0x4e, 0xc1, - 0x75, 0x58, 0xf1, 0x48, 0x60, 0x58, 0x76, 0xc3, 0x68, 0x75, 0x88, 0xd9, 0x88, 0x52, 0x62, 0x66, - 0x15, 0x95, 0x8b, 0xfa, 0xf7, 0x38, 0x71, 0x97, 0xd2, 0x44, 0x16, 0xf8, 0xf8, 0x67, 0xa0, 0x86, - 0x91, 0xee, 0x91, 0x80, 0xd8, 0x54, 0xc7, 0x46, 0xd8, 0x3b, 0x4a, 0x45, 0x86, 0xf9, 0x7c, 0x26, - 0x5b, 0xf7, 0x05, 0x83, 0x5e, 0x12, 0x9b, 0xf5, 0x70, 0x6f, 0x48, 0xc1, 0x7b, 0x51, 0xc9, 0x9f, - 0x65, 0x01, 0x5a, 0xc9, 0x2a, 0x1e, 0xb7, 0xeb, 0x77, 0x5d, 0xf9, 0xff, 0x81, 0x00, 0x06, 0x86, - 0xc5, 0x57, 0x61, 0x9e, 0xf9, 0x82, 0xd8, 0xa6, 0xeb, 0x58, 0x76, 0x98, 0xa0, 0x73, 0x74, 0xf1, - 0xbe, 0x58, 0xc3, 0x8f, 0x24, 0x7d, 0xe0, 0xc6, 0x30, 0x7f, 0x0d, 0x6b, 0x02, 0xaf, 0x5b, 0xa6, - 0x5b, 0xb0, 0xa8, 0x93, 0x16, 0xb1, 0x8e, 0x22, 0x67, 0xe1, 0x15, 0x98, 0xa6, 0x11, 0x61, 0x99, - 0x61, 0x88, 0x1a, 0xad, 0xce, 0x43, 0x13, 0xdf, 0x86, 0x19, 0xe1, 0x05, 0x51, 0x5d, 0x47, 0xd6, - 0xbe, 0x90, 0x5f, 0xfb, 0x01, 0x9c, 0x7d, 0x40, 0x82, 0xb8, 0x1f, 0xc2, 0x6c, 0xd0, 0x60, 0x2e, - 0x9e, 0x5c, 0xa1, 0xbd, 0xe2, 0x6b, 0xda, 0xb7, 0x08, 0xce, 0xf3, 0x06, 0x21, 0x3b, 0x61, 0x4f, - 0x72, 0x82, 0x52, 0xbf, 0x3c, 0x3c, 0x0c, 0x92, 0x12, 0x5e, 0xaf, 0x79, 0xb8, 0x50, 0xa2, 0x65, - 0x45, 0x5a, 0xc4, 0xde, 0x4c, 0xa9, 0xfe, 0x3d, 0x82, 0xf3, 0x12, 0x91, 0xa2, 0x88, 0xed, 0xcb, - 0x8a, 0xd8, 0x68, 0x8b, 0x9c, 0xb2, 0xc8, 0x7d, 0x00, 0xe7, 0x79, 0x91, 0x3b, 0xad, 0x77, 0x7f, - 0x0d, 0xe7, 0x1e, 0x3b, 0xa6, 0xf5, 0xec, 0x38, 0x56, 0x9f, 0x4e, 0xbe, 0x3d, 0x5d, 0xfd, 0x0a, - 0xe3, 0x55, 0x3f, 0xed, 0x4b, 0x04, 0xca, 0x41, 0xbf, 0xdb, 0x1d, 0x47, 0xe4, 0x4d, 0xc0, 0x1e, - 0x09, 0xfa, 0x9e, 0xdd, 0xb0, 0x7a, 0x3d, 0x62, 0x5a, 0x46, 0x40, 0xba, 0xc7, 0x4c, 0x72, 0x51, - 0x3f, 0xc3, 0x29, 0x0f, 0x07, 0x04, 0xbc, 0x06, 0x73, 0x3d, 0xe3, 0xc5, 0xa0, 0x4a, 0x4e, 0x30, - 0x67, 0x2b, 0x3d, 0xe3, 0x45, 0x58, 0x1d, 0xb5, 0x5f, 0xc0, 0x1c, 0x07, 0x21, 0x5c, 0xf8, 0x63, - 0x38, 0xe3, 0x89, 0xa4, 0x1c, 0xec, 0xe3, 0x6e, 0x5c, 0xcb, 0xaa, 0x96, 0xca, 0x5f, 0x7d, 0xc9, - 0x4b, 0x2e, 0xf8, 0x34, 0x60, 0x4a, 0xdc, 0xc8, 0xbb, 0x83, 0x72, 0x3e, 0x8e, 0xca, 0xe7, 0x60, - 0x86, 0x97, 0x04, 0xbf, 0x34, 0xc9, 0x5a, 0xe2, 0x34, 0xab, 0x09, 0x7e, 0x6e, 0xf7, 0x98, 0xc8, - 0xeb, 0x1e, 0xda, 0x4f, 0x00, 0xef, 0xb6, 0x3a, 0xb6, 0xf3, 0xcb, 0x2e, 0x31, 0xdb, 0xa7, 0x05, - 0x51, 0x88, 0x83, 0xd0, 0x7e, 0x53, 0x80, 0xe5, 0xc3, 0xc0, 0x23, 0x46, 0xcf, 0xb2, 0xdb, 0xe3, - 0x7a, 0x33, 0xef, 0x54, 0x7c, 0x0b, 0xce, 0xf5, 0x98, 0xcd, 0x64, 0xda, 0x4d, 0x94, 0xa7, 0xf4, - 0x15, 0x4e, 0x4e, 0xb7, 0xc7, 0xf7, 0xb2, 0xfb, 0x92, 0xb6, 0x5b, 0x4e, 0xee, 0xdb, 0xe5, 0xe2, - 0x76, 0xe0, 0x82, 0xcf, 0x74, 0x68, 0x0c, 0x69, 0xc7, 0x25, 0xce, 0xb2, 0x9b, 0x35, 0x6b, 0x1b, - 0x56, 0x52, 0x26, 0x78, 0x43, 0xb1, 0xf4, 0x11, 0xac, 0xdc, 0xf3, 0x08, 0x2d, 0xc6, 0xb6, 0xe1, - 0xfa, 0xcf, 0x9d, 0x20, 0x34, 0xb6, 0x6c, 0x62, 0x49, 0x3b, 0xa0, 0x20, 0x29, 0x00, 0xaf, 0x10, - 0xac, 0x88, 0xf2, 0x9e, 0x3a, 0xf1, 0x16, 0x14, 0x7d, 0xb1, 0x24, 0xca, 0xba, 0x2a, 0x29, 0x62, - 0xe1, 0xa6, 0x88, 0xf7, 0xf5, 0xca, 0xf9, 0x7f, 0x10, 0x14, 0xc3, 0x33, 0xc7, 0x98, 0xc2, 0xb6, - 0x41, 0x21, 0x2f, 0x5c, 0xcb, 0x23, 0xfc, 0x5b, 0x6b, 0x62, 0xe4, 0xb7, 0x16, 0x70, 0x76, 0xba, - 0x80, 0xef, 0x46, 0x43, 0xcc, 0x24, 0x73, 0xcc, 0xb5, 0x7c, 0x35, 0xbf, 0xeb, 0x01, 0xa6, 0x0b, - 0xcb, 0xac, 0x95, 0x88, 0xe3, 0xdf, 0x70, 0xe7, 0x3a, 0x86, 0x95, 0x94, 0x34, 0x11, 0xa5, 0xef, - 0xc3, 0x6c, 0xe8, 0xbe, 0x30, 0x3a, 0x87, 0xf9, 0x7a, 0xc0, 0x7c, 0xe2, 0x46, 0xf5, 0x2e, 0xac, - 0x88, 0x46, 0x95, 0x8a, 0x32, 0x35, 0x15, 0x65, 0xb3, 0x83, 0x48, 0xd2, 0xfe, 0x80, 0x40, 0x39, - 0x24, 0xa4, 0x33, 0x4e, 0x41, 0xd9, 0x84, 0x49, 0x16, 0x02, 0x85, 0x51, 0x21, 0xf0, 0xe1, 0x5b, - 0x3a, 0xe3, 0xc4, 0x17, 0x63, 0x08, 0x98, 0xc9, 0x3e, 0x7c, 0x6b, 0x80, 0x61, 0xaf, 0x08, 0xd3, - 0x81, 0xe1, 0xb5, 0x49, 0xa0, 0x2d, 0xc0, 0x1c, 0x07, 0xc3, 0x8d, 0x56, 0xff, 0xef, 0x12, 0x80, - 0x68, 0xbb, 0x4d, 0xe2, 0xe1, 0xdf, 0x21, 0xc0, 0x22, 0x35, 0xe3, 0x78, 0x46, 0x34, 0x7e, 0x75, - 0x04, 0x5d, 0xdb, 0xfc, 0xe2, 0x5f, 0xff, 0xfe, 0xa6, 0x50, 0x51, 0xdf, 0xae, 0x1d, 0x6d, 0xd5, - 0x7e, 0x45, 0x53, 0x60, 0x47, 0x84, 0x82, 0x5f, 0xab, 0xd4, 0x12, 0x53, 0x43, 0xad, 0xf2, 0xf2, - 0x0e, 0xaa, 0xe0, 0xaf, 0x11, 0xfb, 0xf6, 0x4f, 0xa0, 0x28, 0x67, 0xa5, 0xc8, 0x47, 0xc2, 0x91, - 0x78, 0xde, 0x63, 0x78, 0x6a, 0xf8, 0x26, 0xc3, 0x13, 0x97, 0x3f, 0x0c, 0x17, 0xfe, 0x73, 0xf4, - 0x9b, 0x21, 0x81, 0xeb, 0x9d, 0xac, 0xb4, 0xdc, 0x59, 0x73, 0x24, 0xb4, 0x1d, 0x06, 0xed, 0xfb, - 0xf5, 0x7a, 0x06, 0x5a, 0xf5, 0x24, 0x76, 0xfb, 0x16, 0xf1, 0xcf, 0xfd, 0xc4, 0x5c, 0x87, 0x25, - 0x5f, 0x34, 0x79, 0xf3, 0xa6, 0xfa, 0xce, 0x89, 0x78, 0x79, 0xf8, 0x68, 0x55, 0x86, 0xb6, 0x8c, - 0xaf, 0x31, 0xb4, 0x02, 0x5b, 0x0c, 0xe3, 0xcb, 0x24, 0x48, 0xfc, 0x47, 0x14, 0x7e, 0xd0, 0x8e, - 0xb2, 0x60, 0xee, 0x44, 0xa8, 0x9e, 0xcd, 0xa4, 0xc3, 0xfd, 0x9e, 0x1b, 0x1c, 0x87, 0x4e, 0xad, - 0x8c, 0xe9, 0xd4, 0xbf, 0x20, 0x38, 0x93, 0x19, 0x6c, 0x64, 0x16, 0xcb, 0x9b, 0x7e, 0x72, 0x01, - 0xfd, 0x88, 0x01, 0xda, 0xd7, 0x3e, 0x18, 0x0b, 0xd0, 0x9d, 0x5e, 0x5a, 0x0e, 0xf5, 0xeb, 0x57, - 0x08, 0x94, 0xd8, 0xcc, 0x83, 0xd7, 0xb3, 0xf8, 0xb2, 0x23, 0x51, 0x2e, 0xb2, 0x7d, 0x86, 0xec, - 0xae, 0x76, 0x7b, 0x3c, 0x64, 0xc6, 0x40, 0x02, 0xc5, 0xf4, 0x5b, 0x04, 0x93, 0x74, 0x4e, 0xc0, - 0x97, 0x64, 0xb3, 0x72, 0x34, 0x42, 0xc9, 0x42, 0x3e, 0x3e, 0x5e, 0x84, 0x21, 0xaf, 0xd5, 0xc7, - 0x43, 0xe3, 0xf6, 0xbb, 0x5d, 0x0a, 0xc3, 0x84, 0xf9, 0xc4, 0xd8, 0x82, 0x65, 0xad, 0x4f, 0x32, - 0xda, 0xa9, 0x1b, 0x23, 0xf9, 0x38, 0xc0, 0x32, 0xda, 0x44, 0x34, 0xf7, 0x97, 0xd2, 0x1f, 0x19, - 0xf8, 0x7a, 0x5e, 0x94, 0x64, 0x3e, 0x44, 0x72, 0x5d, 0xf1, 0x90, 0x29, 0x7f, 0x4f, 0xbb, 0x7b, - 0x9a, 0x20, 0x19, 0x88, 0xa1, 0x86, 0x78, 0x85, 0x60, 0x3e, 0xd1, 0x1a, 0x65, 0x96, 0x90, 0x75, - 0x6a, 0x99, 0x25, 0xa4, 0x3d, 0x56, 0xab, 0x30, 0xb4, 0xeb, 0x58, 0xcb, 0xcf, 0xf7, 0x48, 0xf8, - 0x97, 0x08, 0x16, 0x92, 0x63, 0x1e, 0x96, 0xc8, 0x91, 0x0e, 0x82, 0xea, 0x90, 0xc6, 0xad, 0xdd, - 0x60, 0x18, 0xae, 0xa9, 0x6b, 0xf2, 0x66, 0x12, 0xca, 0x17, 0x05, 0xf1, 0x15, 0x82, 0x85, 0xe4, - 0x68, 0x28, 0x43, 0x21, 0x1d, 0x1e, 0x87, 0xa2, 0x10, 0xd5, 0xa6, 0x5e, 0xe1, 0x7e, 0x0b, 0x47, - 0xab, 0x51, 0x70, 0x3e, 0x47, 0xb0, 0x90, 0x9c, 0x21, 0x64, 0x70, 0xa4, 0x53, 0x46, 0x6e, 0x08, - 0xdd, 0x64, 0x50, 0x36, 0x2a, 0x6f, 0x27, 0xa0, 0xe4, 0xa1, 0x60, 0x69, 0x4b, 0x67, 0x00, 0x59, - 0xda, 0xc6, 0x06, 0x15, 0x69, 0xa7, 0x8a, 0x8d, 0x0e, 0xa7, 0x4d, 0x5b, 0x9f, 0x90, 0xce, 0x1d, - 0x54, 0xa9, 0xff, 0x69, 0x06, 0x66, 0xc5, 0xcf, 0x6c, 0xe2, 0xe1, 0x4f, 0x40, 0xe1, 0x91, 0xc0, - 0x6f, 0x66, 0xf2, 0xfe, 0x11, 0xab, 0x79, 0x04, 0xed, 0x3a, 0x43, 0x73, 0x55, 0xbd, 0x2c, 0x8d, - 0x0a, 0xfe, 0x67, 0x59, 0xf8, 0xe0, 0x25, 0x28, 0xb1, 0xcb, 0x02, 0x59, 0x29, 0xcd, 0xde, 0x25, - 0xe4, 0x0b, 0xae, 0x31, 0xc1, 0xd7, 0xeb, 0xeb, 0x4c, 0x30, 0x13, 0x54, 0x1d, 0x2a, 0xfe, 0x0b, - 0x04, 0x33, 0x42, 0x71, 0xbc, 0x2a, 0xfd, 0xff, 0x15, 0xbb, 0x45, 0x50, 0xd7, 0x86, 0x70, 0x08, - 0x47, 0xd4, 0x19, 0x82, 0x1b, 0xda, 0xc6, 0x00, 0x81, 0x5c, 0xb8, 0xb8, 0x9a, 0xa1, 0x20, 0x1c, - 0x28, 0x86, 0x57, 0x2b, 0x78, 0x4d, 0x3a, 0x57, 0x9d, 0x4c, 0xfb, 0x0d, 0x26, 0x7b, 0x0d, 0x5f, - 0x19, 0x21, 0x9b, 0x06, 0x3e, 0x0c, 0x2e, 0x07, 0xf0, 0x55, 0x79, 0xc5, 0x49, 0xdc, 0x52, 0xa8, - 0xeb, 0xc3, 0x99, 0x84, 0xfa, 0x49, 0x08, 0xb2, 0x9a, 0x24, 0xee, 0x15, 0xfe, 0x86, 0xe0, 0xac, - 0xfc, 0xef, 0x3d, 0xae, 0x0d, 0x91, 0x24, 0x9d, 0x92, 0x36, 0x4f, 0xbe, 0x41, 0xc0, 0x4c, 0xce, - 0x9c, 0xf9, 0x96, 0x4a, 0x4d, 0x4c, 0x01, 0x28, 0xb1, 0x1b, 0x00, 0x59, 0xb0, 0x66, 0x2f, 0x08, - 0x72, 0x2b, 0x85, 0x30, 0x55, 0x65, 0x94, 0xb7, 0xf6, 0xbe, 0x46, 0xb0, 0xdc, 0x72, 0x7a, 0x19, - 0x61, 0x7b, 0x0a, 0xff, 0x3d, 0x7b, 0x40, 0xcf, 0x3d, 0x40, 0x3f, 0xbf, 0x25, 0x18, 0xda, 0x4e, - 0xd7, 0xb0, 0xdb, 0x55, 0xc7, 0x6b, 0xd7, 0xda, 0xc4, 0x66, 0x52, 0x6b, 0x9c, 0x64, 0xb8, 0x96, - 0x3f, 0xb8, 0x39, 0xde, 0xe6, 0x4f, 0xff, 0x43, 0xe8, 0xaf, 0x85, 0xb3, 0x0f, 0xf8, 0xde, 0x7b, - 0x5d, 0xa7, 0x6f, 0xd2, 0xa0, 0x3e, 0xec, 0x37, 0xab, 0x1f, 0x6f, 0xfd, 0x33, 0x24, 0x3c, 0x65, - 0x84, 0xa7, 0x9c, 0xf0, 0xf4, 0xe3, 0xad, 0xe6, 0x34, 0x3b, 0xf7, 0xdd, 0xff, 0x07, 0x00, 0x00, - 0xff, 0xff, 0xcb, 0xe1, 0xb8, 0xca, 0x90, 0x1e, 0x00, 0x00, +var fileDescriptor_pubsub_ccb5b83827ebb62b = []byte{ + // 2199 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x5a, 0xcd, 0x73, 0x1b, 0x49, + 0x15, 0xdf, 0x96, 0x1c, 0x47, 0x7e, 0xb2, 0xe3, 0xb8, 0xb1, 0x1d, 0x79, 0xf2, 0x65, 0x4f, 0x8c, + 0xed, 0x28, 0x89, 0x64, 0x2b, 0xb5, 0x61, 0x13, 0xe3, 0x6c, 0xd9, 0x71, 0xc8, 0x06, 0x12, 0x30, + 0xe3, 0x10, 0xaa, 0xa8, 0x14, 0xaa, 0x91, 0xd4, 0x51, 0x66, 0x35, 0x9a, 0x99, 0x9d, 0x19, 0x79, + 0xe3, 0x85, 0x50, 0x61, 0x97, 0xa2, 0x8a, 0x22, 0x07, 0x76, 0xb9, 0xee, 0x81, 0x82, 0x1b, 0x47, + 0xaa, 0xb8, 0xc2, 0x9d, 0x2b, 0x07, 0xfe, 0x01, 0x8e, 0x5c, 0x38, 0xc1, 0x71, 0xab, 0x3f, 0x66, + 0x34, 0x1f, 0x3d, 0x92, 0x65, 0x6f, 0x6e, 0xa3, 0xee, 0xd7, 0xfd, 0x7e, 0xef, 0xbb, 0xdf, 0x2b, + 0xc1, 0xc5, 0xb6, 0x6d, 0xb7, 0x4d, 0x52, 0x75, 0x7a, 0x0d, 0xaf, 0xd7, 0xa8, 0x1e, 0x6c, 0x88, + 0xaf, 0x8a, 0xe3, 0xda, 0xbe, 0x8d, 0xcf, 0xf2, 0xed, 0x8a, 0x58, 0x3c, 0xd8, 0x50, 0x2e, 0x88, + 0x03, 0xba, 0x63, 0x54, 0x75, 0xcb, 0xb2, 0x7d, 0xdd, 0x37, 0x6c, 0xcb, 0xe3, 0xf4, 0xca, 0xa5, + 0xe0, 0x3a, 0xfa, 0xab, 0xd1, 0x7b, 0x5e, 0x6d, 0xf5, 0x5c, 0x46, 0x20, 0xf6, 0xcf, 0x27, 0xf7, + 0x49, 0xd7, 0xf1, 0x0f, 0xc5, 0xe6, 0x62, 0x72, 0xf3, 0xb9, 0x41, 0xcc, 0x56, 0xbd, 0xab, 0x7b, + 0x1d, 0x41, 0x71, 0x39, 0x49, 0xe1, 0x1b, 0x5d, 0xe2, 0xf9, 0x7a, 0xd7, 0xe1, 0x04, 0xea, 0x53, + 0x98, 0x7d, 0x4c, 0x3c, 0x4f, 0x6f, 0x93, 0x7d, 0xdf, 0x76, 0xf5, 0x36, 0xd9, 0xb3, 0x4d, 0xa3, + 0x79, 0x88, 0xef, 0xc2, 0x79, 0xdd, 0x34, 0xed, 0x8f, 0x49, 0xab, 0xee, 0x10, 0xd7, 0x33, 0x3c, + 0x9f, 0x58, 0x4d, 0x52, 0x77, 0x49, 0x9b, 0x82, 0x2f, 0xa1, 0xc5, 0xfc, 0xda, 0x84, 0xb6, 0x20, + 0x48, 0xf6, 0xfa, 0x14, 0x1a, 0x27, 0x50, 0xff, 0x8b, 0xe0, 0xd4, 0x13, 0xdb, 0x31, 0x9a, 0x18, + 0xc3, 0x98, 0xa5, 0x77, 0x49, 0x09, 0x2d, 0xa2, 0xb5, 0x09, 0x8d, 0x7d, 0xe3, 0x4d, 0x18, 0x37, + 0xf5, 0x06, 0x31, 0xbd, 0x52, 0x6e, 0x31, 0xbf, 0x56, 0xac, 0x5d, 0xa9, 0x24, 0xd5, 0x56, 0x61, + 0x87, 0x2b, 0x8f, 0x18, 0xd5, 0x7d, 0xcb, 0x77, 0x0f, 0x35, 0x71, 0x04, 0x3f, 0x83, 0xf9, 0x2e, + 0x87, 0x5c, 0xf7, 0x38, 0xe6, 0xba, 0xc3, 0x40, 0x97, 0xf2, 0x8b, 0x68, 0xad, 0x58, 0x5b, 0x49, + 0x5f, 0x26, 0x13, 0x51, 0x9b, 0xed, 0x4a, 0x56, 0x95, 0xdb, 0x50, 0x8c, 0x30, 0xc5, 0x67, 0x21, + 0xdf, 0x21, 0x87, 0x02, 0x3c, 0xfd, 0xc4, 0xb3, 0x70, 0xea, 0x40, 0x37, 0x7b, 0xa4, 0x94, 0x63, + 0x6b, 0xfc, 0xc7, 0x9d, 0xdc, 0x7b, 0x48, 0xfd, 0x3c, 0x07, 0x53, 0x7b, 0x8c, 0xa7, 0xe0, 0x47, + 0x65, 0x6f, 0xe9, 0xbe, 0xce, 0x8e, 0x4f, 0x6a, 0xec, 0x1b, 0xff, 0x00, 0x40, 0xf7, 0x7d, 0xd7, + 0x68, 0xf4, 0x7c, 0x12, 0xc8, 0x5f, 0x4d, 0x43, 0x8e, 0x5d, 0x54, 0xd9, 0x0e, 0x4f, 0x70, 0x5d, + 0x44, 0xae, 0xc0, 0x17, 0x01, 0x02, 0x7d, 0x18, 0x2d, 0xa6, 0x83, 0x09, 0x6d, 0x42, 0xac, 0x3c, + 0x6c, 0xe1, 0x2d, 0x98, 0x74, 0x7a, 0x0d, 0xd3, 0xf0, 0x5e, 0xd4, 0xa9, 0xf1, 0x4b, 0x63, 0x4c, + 0x49, 0x4a, 0xc8, 0x51, 0x78, 0x46, 0xe5, 0x49, 0xe0, 0x19, 0x5a, 0x51, 0xd0, 0xd3, 0x15, 0x65, + 0x0b, 0xa6, 0x13, 0xcc, 0x47, 0xd2, 0xc9, 0x2a, 0x4c, 0x3f, 0x20, 0x3e, 0x33, 0xa6, 0x46, 0x3e, + 0xea, 0x11, 0xcf, 0xa7, 0xc4, 0x3e, 0xfd, 0x2d, 0x2e, 0xe0, 0x3f, 0xd4, 0xd7, 0x08, 0xf0, 0x8f, + 0x9c, 0x96, 0xee, 0x93, 0x18, 0xf1, 0x8d, 0x28, 0x71, 0xb1, 0x76, 0x2e, 0xc3, 0x51, 0xc4, 0x2d, + 0x78, 0x13, 0x8a, 0x3d, 0x76, 0x09, 0x0b, 0x02, 0x06, 0x47, 0x26, 0xeb, 0x77, 0x68, 0x9c, 0x3c, + 0xd6, 0xbd, 0x8e, 0x06, 0x9c, 0x9c, 0x7e, 0xab, 0x4d, 0x38, 0xb3, 0xc7, 0x25, 0x1f, 0x08, 0x15, + 0x6f, 0x42, 0x41, 0xa8, 0x37, 0xb0, 0xdf, 0xe5, 0x21, 0xf6, 0xd3, 0xc2, 0x03, 0x6a, 0x0d, 0xa6, + 0x43, 0x26, 0x9e, 0x63, 0x5b, 0x1e, 0xc1, 0x97, 0xa1, 0xd8, 0x37, 0x60, 0x10, 0x5b, 0x10, 0x5a, + 0xd0, 0x53, 0x0d, 0x98, 0x79, 0x64, 0x78, 0x5c, 0x8b, 0x5e, 0x80, 0xad, 0x04, 0xa7, 0x1d, 0xd7, + 0xfe, 0x90, 0x34, 0x7d, 0x81, 0x2e, 0xf8, 0x89, 0xcf, 0xc3, 0x84, 0xc3, 0xa2, 0xc3, 0xf8, 0x84, + 0x5b, 0xe4, 0x94, 0x56, 0xa0, 0x0b, 0xfb, 0xc6, 0x27, 0x84, 0x7a, 0x0b, 0xdb, 0xf4, 0xed, 0x0e, + 0xb1, 0x02, 0x6f, 0xa1, 0x2b, 0x4f, 0xe8, 0x82, 0xda, 0x05, 0x1c, 0x65, 0x25, 0x10, 0x56, 0x61, + 0x9c, 0x89, 0xce, 0xc1, 0x0d, 0x30, 0x83, 0x20, 0xc3, 0x2b, 0x30, 0x6d, 0x91, 0x97, 0x7e, 0x3d, + 0xc2, 0x8a, 0xbb, 0xc6, 0x14, 0x5d, 0xde, 0x0b, 0xd9, 0x7d, 0x04, 0x17, 0x43, 0x76, 0xfb, 0xbd, + 0x86, 0xd7, 0x74, 0x0d, 0x87, 0xa5, 0xc7, 0xc1, 0x16, 0x38, 0x89, 0x84, 0x16, 0x5c, 0xca, 0x62, + 0x29, 0xa4, 0x5d, 0x86, 0x29, 0x2f, 0xba, 0x21, 0x2c, 0x12, 0x5f, 0x3c, 0xb2, 0x88, 0x5d, 0x58, + 0xe8, 0xf3, 0xb3, 0x74, 0xc7, 0x7b, 0x61, 0xfb, 0x6f, 0x51, 0xbc, 0x06, 0x28, 0x32, 0x76, 0x42, + 0xb4, 0x0b, 0x30, 0xe1, 0x05, 0x8b, 0x42, 0xac, 0xfe, 0xc2, 0x91, 0x45, 0x2a, 0x03, 0xde, 0x25, + 0x26, 0x49, 0x84, 0xaa, 0x3c, 0xae, 0xff, 0x9a, 0x87, 0xc9, 0xa8, 0x9a, 0xa5, 0xf5, 0x20, 0x3c, + 0x9a, 0x8b, 0xaa, 0x61, 0x0b, 0x8a, 0x4e, 0xcf, 0x7b, 0x51, 0x6f, 0xda, 0xd6, 0x73, 0xa3, 0x2d, + 0x12, 0xd7, 0x05, 0x59, 0xa8, 0x79, 0x2f, 0xee, 0x31, 0x1a, 0x0d, 0x9c, 0xf0, 0x1b, 0xaf, 0xc3, + 0xac, 0xde, 0xec, 0xd4, 0x5b, 0x44, 0x6f, 0x99, 0x86, 0x45, 0xea, 0x1e, 0x69, 0xda, 0x56, 0xcb, + 0x2b, 0x9d, 0x62, 0x0a, 0xc5, 0x7a, 0xb3, 0xb3, 0x2b, 0xb6, 0xf6, 0xf9, 0x0e, 0xae, 0xc1, 0x9c, + 0x4b, 0x7c, 0xdd, 0xb0, 0xea, 0x7a, 0xb3, 0x43, 0x5a, 0xf5, 0x30, 0xca, 0x4f, 0x2f, 0xa2, 0xb5, + 0x82, 0xf6, 0x0d, 0xbe, 0xb9, 0x4d, 0xf7, 0x44, 0x60, 0x7b, 0xf8, 0xc7, 0xa0, 0x04, 0xc1, 0xeb, + 0x12, 0x9f, 0x58, 0x54, 0xc6, 0x7a, 0x50, 0xc4, 0x4b, 0x05, 0x86, 0x79, 0x21, 0x95, 0x80, 0x76, + 0x05, 0x81, 0x56, 0x12, 0x87, 0xb5, 0xe0, 0x6c, 0xb0, 0x83, 0x77, 0xc2, 0x1a, 0x39, 0xc1, 0x62, + 0xae, 0x9c, 0x16, 0x3c, 0xaa, 0x57, 0x59, 0xa9, 0x3c, 0x49, 0x31, 0xfb, 0x1b, 0x02, 0xe8, 0x2b, + 0x16, 0x5f, 0x81, 0x29, 0x66, 0x0b, 0x62, 0xb5, 0x1c, 0xdb, 0xb0, 0x82, 0x9c, 0x33, 0x49, 0x17, + 0xef, 0x8b, 0x35, 0xfc, 0x48, 0x52, 0xda, 0xae, 0x0f, 0xb2, 0xd7, 0xa0, 0xba, 0x76, 0xd2, 0xca, + 0xd3, 0x84, 0x69, 0x8d, 0x34, 0x89, 0x71, 0x10, 0x1a, 0x0b, 0xcf, 0xc1, 0x38, 0xf5, 0x08, 0xa3, + 0x15, 0xb8, 0xa8, 0xde, 0xec, 0x3c, 0x6c, 0xe1, 0xdb, 0x70, 0x5a, 0x58, 0x41, 0x14, 0x8c, 0xa1, + 0xe9, 0x3c, 0xa0, 0x57, 0xbf, 0x0d, 0xf3, 0x0f, 0x88, 0x1f, 0xb5, 0x43, 0x10, 0x0d, 0x2a, 0x4c, + 0x46, 0xf3, 0x45, 0xa0, 0xaf, 0xe8, 0x9a, 0xfa, 0x25, 0x82, 0x05, 0x5e, 0xf3, 0x64, 0x37, 0xec, + 0x48, 0x6e, 0x28, 0xd6, 0x2e, 0x0d, 0x76, 0x83, 0x38, 0x87, 0x93, 0xd5, 0x43, 0x07, 0x4a, 0x34, + 0x95, 0x48, 0xf3, 0xf2, 0xdb, 0xa9, 0x3e, 0xbf, 0x41, 0x3c, 0x59, 0xca, 0xf3, 0xf2, 0xae, 0x2c, + 0x2f, 0x0f, 0xd7, 0xc8, 0x31, 0xf3, 0xf6, 0xfb, 0xb0, 0xc0, 0x93, 0xdc, 0x71, 0xad, 0xfb, 0x73, + 0x38, 0xf7, 0xd8, 0x6e, 0x19, 0xcf, 0x0f, 0x23, 0xf9, 0xe9, 0xe8, 0xc7, 0x93, 0xd9, 0x2f, 0x37, + 0x5a, 0xf6, 0x53, 0x3f, 0x43, 0x50, 0xdc, 0xeb, 0x99, 0xe6, 0x28, 0x2c, 0x6f, 0x00, 0x76, 0x89, + 0xdf, 0x73, 0xad, 0xba, 0xd1, 0xed, 0x92, 0x96, 0xa1, 0xfb, 0xc4, 0x3c, 0x64, 0x9c, 0x0b, 0xda, + 0x0c, 0xdf, 0x79, 0xd8, 0xdf, 0xc0, 0x4b, 0x30, 0xd9, 0xd5, 0x5f, 0xf6, 0xb3, 0x64, 0x9e, 0x19, + 0xbb, 0xd8, 0xd5, 0x5f, 0x06, 0xd9, 0x51, 0xfd, 0x29, 0x4c, 0x72, 0x10, 0xc2, 0x84, 0xdf, 0x87, + 0x19, 0x57, 0x04, 0x65, 0xff, 0x1c, 0x37, 0xe3, 0x52, 0x5a, 0xb4, 0x44, 0xfc, 0x6a, 0x67, 0xdd, + 0xf8, 0x82, 0x47, 0x1d, 0xa6, 0xc4, 0x95, 0xbc, 0xdd, 0x4f, 0xe7, 0xa3, 0x88, 0x7c, 0x0e, 0x4e, + 0xf3, 0x94, 0xe0, 0x95, 0xc6, 0x58, 0x39, 0x1c, 0x67, 0x39, 0xc1, 0xcb, 0xac, 0x1e, 0xf9, 0xac, + 0xea, 0xa1, 0xfe, 0x10, 0xf0, 0x76, 0xb3, 0x63, 0xd9, 0x1f, 0x9b, 0xa4, 0xd5, 0x3e, 0x2e, 0x88, + 0x5c, 0x14, 0x84, 0xfa, 0xcb, 0x1c, 0xcc, 0xee, 0xfb, 0x2e, 0xd1, 0xbb, 0x86, 0xd5, 0x1e, 0xd5, + 0x9a, 0x59, 0xb7, 0xe2, 0x5b, 0x70, 0xae, 0xcb, 0x74, 0x26, 0x93, 0x2e, 0xbf, 0x76, 0x4a, 0x9b, + 0xe3, 0xdb, 0xc9, 0xf2, 0xf8, 0x6e, 0xfa, 0x5c, 0x5c, 0x77, 0xb3, 0xf1, 0x73, 0xdb, 0x9c, 0xdd, + 0x16, 0x9c, 0xf7, 0x98, 0x0c, 0xf5, 0x01, 0xe5, 0xb8, 0xc4, 0x49, 0xb6, 0xd3, 0x6a, 0x6d, 0xc3, + 0x5c, 0x42, 0x05, 0x6f, 0xc9, 0x97, 0xfe, 0x85, 0x60, 0xee, 0x9e, 0x4b, 0x68, 0x36, 0x16, 0x2f, + 0xa2, 0x40, 0xdb, 0xb2, 0x27, 0x4b, 0xd2, 0x02, 0x39, 0x89, 0x05, 0xbe, 0x17, 0x96, 0xf0, 0x3c, + 0x83, 0x75, 0x33, 0x0d, 0x4b, 0xca, 0xf0, 0xeb, 0xae, 0xe5, 0x6f, 0x10, 0xcc, 0x89, 0x3a, 0x93, + 0x90, 0xec, 0x16, 0x14, 0x82, 0xe7, 0x9f, 0xa8, 0x2f, 0x8a, 0x24, 0x9b, 0x06, 0x87, 0x42, 0xda, + 0x93, 0xd5, 0x95, 0xff, 0x20, 0x28, 0x04, 0x77, 0x8e, 0xf0, 0x1c, 0xdc, 0x84, 0x22, 0x79, 0xe9, + 0x18, 0x2e, 0xe1, 0x7d, 0x6c, 0x7e, 0x68, 0x1f, 0x0b, 0x9c, 0x9c, 0x2e, 0xe0, 0xbb, 0xa1, 0x29, + 0xc6, 0x98, 0x29, 0x56, 0xb2, 0xc5, 0xfc, 0xba, 0xb5, 0xbf, 0x0e, 0x98, 0xbe, 0x11, 0x12, 0x9a, + 0x57, 0x12, 0x9a, 0x9f, 0xe8, 0x6b, 0x57, 0x35, 0x61, 0x96, 0x55, 0xc1, 0x64, 0xb7, 0xf0, 0x76, + 0x8a, 0xee, 0x21, 0xcc, 0x25, 0xb8, 0x89, 0x00, 0x7b, 0x2f, 0xd9, 0x2c, 0x0c, 0xf6, 0x8e, 0x63, + 0x34, 0x12, 0x37, 0x61, 0x4e, 0xd4, 0xd8, 0x11, 0xb4, 0xf3, 0x5b, 0x04, 0xc5, 0x7d, 0x42, 0x3a, + 0xa3, 0xe4, 0xc2, 0x75, 0x18, 0x63, 0x4e, 0x93, 0x1b, 0xe6, 0x34, 0x1f, 0xbc, 0xa3, 0x31, 0x4a, + 0x7c, 0x21, 0x82, 0x80, 0xa9, 0xec, 0x83, 0x77, 0xfa, 0x18, 0x76, 0x0a, 0x30, 0xee, 0xeb, 0x6e, + 0x9b, 0xf8, 0xea, 0x19, 0x98, 0xe4, 0x60, 0xb8, 0xd2, 0x6a, 0x7f, 0x2f, 0xc0, 0x84, 0x68, 0xf0, + 0x89, 0x8b, 0x3f, 0x84, 0x22, 0x8f, 0x70, 0x3e, 0x0b, 0xcb, 0xea, 0x9b, 0x95, 0xac, 0x0d, 0xf5, + 0xea, 0xa7, 0xff, 0xfc, 0xf7, 0xef, 0x73, 0x57, 0x94, 0x4b, 0xd5, 0x83, 0x8d, 0xea, 0xcf, 0x68, + 0x74, 0x6c, 0x09, 0x9b, 0x7b, 0xd5, 0x72, 0x95, 0x77, 0xdb, 0xd5, 0xf2, 0xab, 0x3b, 0xa8, 0x8c, + 0x5f, 0x41, 0x31, 0x32, 0x40, 0xc1, 0xcb, 0xe9, 0x2b, 0xd3, 0xf3, 0x95, 0x6c, 0xc6, 0x55, 0xc6, + 0xf8, 0x6a, 0x6d, 0x99, 0x31, 0x66, 0x8c, 0x2a, 0x03, 0xd9, 0x7f, 0x8a, 0xe0, 0xb4, 0x10, 0x1c, + 0x2f, 0x4a, 0x1f, 0xd0, 0x91, 0xc9, 0x8a, 0xb2, 0x34, 0x80, 0x82, 0x6b, 0x52, 0xad, 0x31, 0x04, + 0xd7, 0xd5, 0xd5, 0x3e, 0x02, 0x39, 0x73, 0x31, 0xae, 0xa2, 0x20, 0x6c, 0x28, 0x04, 0xe3, 0x26, + 0x2c, 0x61, 0x91, 0x18, 0x45, 0x65, 0x4b, 0xbf, 0xca, 0x78, 0x2f, 0xe1, 0xcb, 0x43, 0x78, 0xe3, + 0xd7, 0x08, 0xa0, 0x3f, 0x30, 0xc1, 0x92, 0x41, 0x66, 0x6a, 0x72, 0xa3, 0x2c, 0x0f, 0x26, 0x12, + 0xe2, 0xc7, 0x21, 0x08, 0xe6, 0x11, 0x10, 0xaf, 0x04, 0x0a, 0xfc, 0x17, 0x04, 0xf3, 0xf2, 0x89, + 0x06, 0xae, 0x0e, 0xe0, 0x24, 0x7b, 0xd6, 0x2b, 0xeb, 0x47, 0x3f, 0x20, 0x60, 0xbe, 0xcb, 0x60, + 0x56, 0xf1, 0x8d, 0x21, 0x9a, 0xaa, 0xc6, 0x5f, 0xe1, 0x7f, 0x44, 0x91, 0x41, 0x53, 0x98, 0x7a, + 0xf0, 0xb5, 0x41, 0xfc, 0x13, 0xe9, 0x50, 0xb9, 0x7e, 0x34, 0x62, 0x01, 0x74, 0x83, 0x01, 0xbd, + 0x86, 0xaf, 0x0e, 0x05, 0x1a, 0xa2, 0xf1, 0xa1, 0x18, 0x99, 0x73, 0xc8, 0x22, 0x2a, 0x3d, 0x06, + 0x51, 0xe6, 0x53, 0xc9, 0xe5, 0x7e, 0xd7, 0xf1, 0x0f, 0x03, 0x7b, 0x96, 0x87, 0xb9, 0x54, 0xed, + 0x7f, 0x33, 0x00, 0x42, 0xd7, 0x0d, 0xe2, 0xe2, 0x5f, 0x23, 0xc0, 0xe2, 0x95, 0x10, 0xcd, 0x68, + 0x43, 0xba, 0x1e, 0x65, 0xc8, 0xbe, 0xba, 0xce, 0xe0, 0x94, 0x95, 0x6f, 0x4a, 0x13, 0x4b, 0xcc, + 0x58, 0x22, 0xc0, 0xbf, 0x40, 0x6c, 0x96, 0x1b, 0x43, 0xb1, 0x26, 0x8d, 0x31, 0x49, 0xc7, 0x34, + 0x14, 0x4f, 0xdc, 0x8f, 0xa2, 0xfc, 0x07, 0xe1, 0xc2, 0x7f, 0x08, 0xc7, 0xc6, 0x31, 0x5c, 0xd7, + 0xb2, 0x92, 0xdf, 0x71, 0xa0, 0x6d, 0x31, 0x68, 0xdf, 0xaa, 0xd5, 0x52, 0xd0, 0x2a, 0x47, 0xd1, + 0xdb, 0x97, 0x88, 0x8f, 0x6f, 0xe3, 0xa1, 0x59, 0x96, 0x3b, 0xaf, 0x34, 0x2a, 0xaf, 0x1d, 0x89, + 0x56, 0xf8, 0x79, 0x85, 0xa1, 0x5d, 0xc3, 0x2b, 0x99, 0x79, 0x23, 0x1e, 0x89, 0xbf, 0x43, 0xc1, + 0x34, 0x6f, 0x98, 0x06, 0x33, 0xdb, 0xe1, 0x4c, 0x9f, 0x17, 0x46, 0x2d, 0x8f, 0x68, 0xd4, 0x3f, + 0x21, 0x98, 0x49, 0x75, 0x75, 0x32, 0x8d, 0x65, 0xb5, 0x7e, 0x99, 0x80, 0xbe, 0xcb, 0x00, 0xed, + 0xaa, 0xef, 0x8f, 0x04, 0xe8, 0x4e, 0x37, 0xc9, 0x87, 0xda, 0xf5, 0x73, 0x04, 0xc5, 0x48, 0xc3, + 0x27, 0x4b, 0x0f, 0xe9, 0x7e, 0x30, 0x13, 0xd9, 0x2e, 0x43, 0x76, 0x57, 0xbd, 0x3d, 0x1a, 0x32, + 0xbd, 0xcf, 0x81, 0x62, 0xfa, 0x15, 0x82, 0x31, 0xda, 0x24, 0xe1, 0x8b, 0xb2, 0xfa, 0x1a, 0xf6, + 0x8f, 0x32, 0x97, 0x8f, 0xf6, 0x56, 0x81, 0xcb, 0xab, 0xb5, 0xd1, 0xd0, 0x38, 0x3d, 0xd3, 0xa4, + 0x30, 0x5a, 0x30, 0x15, 0xeb, 0xd9, 0xb0, 0xec, 0xb9, 0x2d, 0xe9, 0x6b, 0x95, 0xd5, 0xa1, 0x74, + 0x1c, 0xe0, 0x1a, 0x5a, 0x47, 0x34, 0xf6, 0xcf, 0x26, 0x27, 0x2c, 0xf8, 0x6a, 0x96, 0x97, 0xa4, + 0xa6, 0x30, 0x99, 0xa6, 0x78, 0xc8, 0x84, 0xbf, 0xa7, 0xde, 0x3d, 0x8e, 0x93, 0xf4, 0xd9, 0x50, + 0x45, 0xfc, 0x02, 0x8a, 0x91, 0xb7, 0xbf, 0xcc, 0x45, 0xd2, 0xad, 0x81, 0x32, 0xe0, 0x91, 0xad, + 0xde, 0x60, 0xd8, 0x56, 0x31, 0x4f, 0xdb, 0x41, 0xa9, 0x8a, 0xe1, 0x0a, 0xca, 0x17, 0x8d, 0xa4, + 0x37, 0x08, 0xa6, 0x62, 0x8f, 0x7b, 0x99, 0x25, 0x64, 0xbd, 0x86, 0xcc, 0x12, 0xd2, 0x2e, 0x41, + 0x2d, 0x33, 0x44, 0xcb, 0x58, 0xcd, 0xce, 0x37, 0x21, 0xf3, 0xcf, 0x10, 0x9c, 0x89, 0x77, 0xbc, + 0x78, 0xf5, 0x88, 0x3d, 0xf1, 0x40, 0xad, 0x5c, 0x67, 0x18, 0x56, 0x94, 0x25, 0x79, 0x31, 0x8b, + 0x68, 0x84, 0x1a, 0xe5, 0x0d, 0x82, 0x33, 0xf1, 0x76, 0x58, 0x86, 0x42, 0xda, 0x30, 0x0f, 0x44, + 0x21, 0xb2, 0x5d, 0xad, 0x1c, 0xb3, 0x4d, 0x65, 0x18, 0x9c, 0xd7, 0x08, 0xce, 0xc4, 0xbb, 0x20, + 0x19, 0x1c, 0x69, 0x9f, 0x94, 0xe9, 0xc2, 0xc2, 0x4d, 0xca, 0x47, 0x74, 0x13, 0x9a, 0x36, 0x68, + 0x17, 0x23, 0x4b, 0x1b, 0x91, 0x56, 0x4b, 0x5a, 0x29, 0x23, 0xcd, 0xcf, 0x71, 0xd3, 0x86, 0x47, + 0x48, 0xe7, 0x0e, 0x2a, 0xef, 0x7c, 0x81, 0x60, 0xb6, 0x69, 0x77, 0x53, 0x4c, 0x76, 0x8a, 0x7c, + 0xfc, 0xbe, 0x47, 0x85, 0xdc, 0x43, 0x3f, 0xb9, 0x25, 0x08, 0xda, 0xb6, 0xa9, 0x5b, 0xed, 0x8a, + 0xed, 0xb6, 0xab, 0x6d, 0x62, 0x31, 0x15, 0x54, 0xf9, 0x96, 0xee, 0x18, 0x5e, 0xff, 0x2f, 0x1a, + 0x9b, 0xfc, 0xeb, 0xff, 0x08, 0xfd, 0x39, 0x37, 0xff, 0x80, 0x9f, 0xbd, 0x67, 0xda, 0xbd, 0x16, + 0xed, 0x39, 0xf6, 0x7b, 0x8d, 0xca, 0xd3, 0x8d, 0x7f, 0x04, 0x1b, 0xcf, 0xd8, 0xc6, 0x33, 0xbe, + 0xf1, 0xec, 0xe9, 0x46, 0x63, 0x9c, 0xdd, 0x7b, 0xf3, 0xab, 0x00, 0x00, 0x00, 0xff, 0xff, 0x52, + 0x93, 0x85, 0x71, 0xf9, 0x21, 0x00, 0x00, } diff --git a/vendor/google.golang.org/genproto/googleapis/pubsub/v1beta2/pubsub.pb.go b/vendor/google.golang.org/genproto/googleapis/pubsub/v1beta2/pubsub.pb.go index 7e020490d..d68a598b9 100644 --- a/vendor/google.golang.org/genproto/googleapis/pubsub/v1beta2/pubsub.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/pubsub/v1beta2/pubsub.pb.go @@ -27,7 +27,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // A topic resource. type Topic struct { // Name of the topic. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -70,12 +70,12 @@ type PubsubMessage struct { // base64-encoded. Data []byte `protobuf:"bytes,1,opt,name=data,proto3" json:"data,omitempty"` // Optional attributes for this message. - Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes,proto3" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // ID of this message assigned by the server at publication time. Guaranteed // to be unique within the topic. This value may be read by a subscriber // that receives a PubsubMessage via a Pull call or a push delivery. It must // not be populated by a publisher in a Publish call. - MessageId string `protobuf:"bytes,3,opt,name=message_id,json=messageId" json:"message_id,omitempty"` + MessageId string `protobuf:"bytes,3,opt,name=message_id,json=messageId,proto3" json:"message_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -129,7 +129,7 @@ func (m *PubsubMessage) GetMessageId() string { // Request for the GetTopic method. type GetTopicRequest struct { // The name of the topic to get. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -169,9 +169,9 @@ func (m *GetTopicRequest) GetTopic() string { // Request for the Publish method. type PublishRequest struct { // The messages in the request will be published on this topic. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` // The messages to publish. - Messages []*PubsubMessage `protobuf:"bytes,2,rep,name=messages" json:"messages,omitempty"` + Messages []*PubsubMessage `protobuf:"bytes,2,rep,name=messages,proto3" json:"messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -220,7 +220,7 @@ type PublishResponse struct { // The server-assigned ID of each published message, in the same order as // the messages in the request. IDs are guaranteed to be unique within // the topic. - MessageIds []string `protobuf:"bytes,1,rep,name=message_ids,json=messageIds" json:"message_ids,omitempty"` + MessageIds []string `protobuf:"bytes,1,rep,name=message_ids,json=messageIds,proto3" json:"message_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -260,13 +260,13 @@ func (m *PublishResponse) GetMessageIds() []string { // Request for the ListTopics method. type ListTopicsRequest struct { // The name of the cloud project that topics belong to. - Project string `protobuf:"bytes,1,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,1,opt,name=project,proto3" json:"project,omitempty"` // Maximum number of topics to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last ListTopicsResponse; indicates that this is // a continuation of a prior ListTopics call, and that the system should // return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -320,10 +320,10 @@ func (m *ListTopicsRequest) GetPageToken() string { // Response for the ListTopics method. type ListTopicsResponse struct { // The resulting topics. - Topics []*Topic `protobuf:"bytes,1,rep,name=topics" json:"topics,omitempty"` + Topics []*Topic `protobuf:"bytes,1,rep,name=topics,proto3" json:"topics,omitempty"` // If not empty, indicates that there may be more topics that match the // request; this value should be passed in a new ListTopicsRequest. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -370,13 +370,13 @@ func (m *ListTopicsResponse) GetNextPageToken() string { // Request for the ListTopicSubscriptions method. type ListTopicSubscriptionsRequest struct { // The name of the topic that subscriptions are attached to. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` // Maximum number of subscription names to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last ListTopicSubscriptionsResponse; indicates // that this is a continuation of a prior ListTopicSubscriptions call, and // that the system should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -430,11 +430,11 @@ func (m *ListTopicSubscriptionsRequest) GetPageToken() string { // Response for the ListTopicSubscriptions method. type ListTopicSubscriptionsResponse struct { // The names of the subscriptions that match the request. - Subscriptions []string `protobuf:"bytes,1,rep,name=subscriptions" json:"subscriptions,omitempty"` + Subscriptions []string `protobuf:"bytes,1,rep,name=subscriptions,proto3" json:"subscriptions,omitempty"` // If not empty, indicates that there may be more subscriptions that match // the request; this value should be passed in a new // ListTopicSubscriptionsRequest to get more subscriptions. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -481,7 +481,7 @@ func (m *ListTopicSubscriptionsResponse) GetNextPageToken() string { // Request for the DeleteTopic method. type DeleteTopicRequest struct { // Name of the topic to delete. - Topic string `protobuf:"bytes,1,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,1,opt,name=topic,proto3" json:"topic,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -521,15 +521,15 @@ func (m *DeleteTopicRequest) GetTopic() string { // A subscription resource. type Subscription struct { // Name of the subscription. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of the topic from which this subscription is receiving messages. // This will be present if and only if the subscription has not been detached // from its topic. - Topic string `protobuf:"bytes,2,opt,name=topic" json:"topic,omitempty"` + Topic string `protobuf:"bytes,2,opt,name=topic,proto3" json:"topic,omitempty"` // If push delivery is used with this subscription, this field is // used to configure it. An empty pushConfig signifies that the subscriber // will pull and ack messages using API methods. - PushConfig *PushConfig `protobuf:"bytes,4,opt,name=push_config,json=pushConfig" json:"push_config,omitempty"` + PushConfig *PushConfig `protobuf:"bytes,4,opt,name=push_config,json=pushConfig,proto3" json:"push_config,omitempty"` // This value is the maximum time after a subscriber receives a message // before the subscriber should acknowledge the message. After message // delivery but before the ack deadline expires and before the message is @@ -545,7 +545,7 @@ type Subscription struct { // // If the subscriber never acknowledges the message, the Pub/Sub // system will eventually redeliver the message. - AckDeadlineSeconds int32 `protobuf:"varint,5,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds" json:"ack_deadline_seconds,omitempty"` + AckDeadlineSeconds int32 `protobuf:"varint,5,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds,proto3" json:"ack_deadline_seconds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -607,7 +607,7 @@ func (m *Subscription) GetAckDeadlineSeconds() int32 { type PushConfig struct { // A URL locating the endpoint to which messages should be pushed. // For example, a Webhook endpoint might use "https://example.com/push". - PushEndpoint string `protobuf:"bytes,1,opt,name=push_endpoint,json=pushEndpoint" json:"push_endpoint,omitempty"` + PushEndpoint string `protobuf:"bytes,1,opt,name=push_endpoint,json=pushEndpoint,proto3" json:"push_endpoint,omitempty"` // Endpoint configuration attributes. // // Every endpoint has a set of API supported attributes that can be used to @@ -631,7 +631,7 @@ type PushConfig struct { // * `v1beta1`: uses the push format defined in the v1beta1 Pub/Sub API. // * `v1beta2`: uses the push format defined in the v1beta2 Pub/Sub API. // - Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Attributes map[string]string `protobuf:"bytes,2,rep,name=attributes,proto3" json:"attributes,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -678,9 +678,9 @@ func (m *PushConfig) GetAttributes() map[string]string { // A message and its corresponding acknowledgment ID. type ReceivedMessage struct { // This ID can be used to acknowledge the received message. - AckId string `protobuf:"bytes,1,opt,name=ack_id,json=ackId" json:"ack_id,omitempty"` + AckId string `protobuf:"bytes,1,opt,name=ack_id,json=ackId,proto3" json:"ack_id,omitempty"` // The message. - Message *PubsubMessage `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Message *PubsubMessage `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -727,7 +727,7 @@ func (m *ReceivedMessage) GetMessage() *PubsubMessage { // Request for the GetSubscription method. type GetSubscriptionRequest struct { // The name of the subscription to get. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -767,13 +767,13 @@ func (m *GetSubscriptionRequest) GetSubscription() string { // Request for the ListSubscriptions method. type ListSubscriptionsRequest struct { // The name of the cloud project that subscriptions belong to. - Project string `protobuf:"bytes,1,opt,name=project" json:"project,omitempty"` + Project string `protobuf:"bytes,1,opt,name=project,proto3" json:"project,omitempty"` // Maximum number of subscriptions to return. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The value returned by the last ListSubscriptionsResponse; indicates that // this is a continuation of a prior ListSubscriptions call, and that the // system should return the next page of data. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -827,11 +827,11 @@ func (m *ListSubscriptionsRequest) GetPageToken() string { // Response for the ListSubscriptions method. type ListSubscriptionsResponse struct { // The subscriptions that match the request. - Subscriptions []*Subscription `protobuf:"bytes,1,rep,name=subscriptions" json:"subscriptions,omitempty"` + Subscriptions []*Subscription `protobuf:"bytes,1,rep,name=subscriptions,proto3" json:"subscriptions,omitempty"` // If not empty, indicates that there may be more subscriptions that match // the request; this value should be passed in a new ListSubscriptionsRequest // to get more subscriptions. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -878,7 +878,7 @@ func (m *ListSubscriptionsResponse) GetNextPageToken() string { // Request for the DeleteSubscription method. type DeleteSubscriptionRequest struct { // The subscription to delete. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -918,14 +918,14 @@ func (m *DeleteSubscriptionRequest) GetSubscription() string { // Request for the ModifyPushConfig method. type ModifyPushConfigRequest struct { // The name of the subscription. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // The push configuration for future deliveries. // // An empty pushConfig indicates that the Pub/Sub system should // stop pushing messages from the given subscription and allow // messages to be pulled and acknowledged - effectively pausing // the subscription if Pull is not called. - PushConfig *PushConfig `protobuf:"bytes,2,opt,name=push_config,json=pushConfig" json:"push_config,omitempty"` + PushConfig *PushConfig `protobuf:"bytes,2,opt,name=push_config,json=pushConfig,proto3" json:"push_config,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -972,16 +972,16 @@ func (m *ModifyPushConfigRequest) GetPushConfig() *PushConfig { // Request for the Pull method. type PullRequest struct { // The subscription from which messages should be pulled. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // If this is specified as true the system will respond immediately even if // it is not able to return a message in the Pull response. Otherwise the // system is allowed to wait until at least one message is available rather // than returning no messages. The client may cancel the request if it does // not wish to wait any longer for the response. - ReturnImmediately bool `protobuf:"varint,2,opt,name=return_immediately,json=returnImmediately" json:"return_immediately,omitempty"` + ReturnImmediately bool `protobuf:"varint,2,opt,name=return_immediately,json=returnImmediately,proto3" json:"return_immediately,omitempty"` // The maximum number of messages returned for this request. The Pub/Sub // system may return fewer than the number specified. - MaxMessages int32 `protobuf:"varint,3,opt,name=max_messages,json=maxMessages" json:"max_messages,omitempty"` + MaxMessages int32 `protobuf:"varint,3,opt,name=max_messages,json=maxMessages,proto3" json:"max_messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1038,7 +1038,7 @@ type PullResponse struct { // there are no more available in the backlog. The Pub/Sub system may return // fewer than the maxMessages requested even if there are more messages // available in the backlog. - ReceivedMessages []*ReceivedMessage `protobuf:"bytes,1,rep,name=received_messages,json=receivedMessages" json:"received_messages,omitempty"` + ReceivedMessages []*ReceivedMessage `protobuf:"bytes,1,rep,name=received_messages,json=receivedMessages,proto3" json:"received_messages,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1078,15 +1078,15 @@ func (m *PullResponse) GetReceivedMessages() []*ReceivedMessage { // Request for the ModifyAckDeadline method. type ModifyAckDeadlineRequest struct { // The name of the subscription. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // The acknowledgment ID. - AckId string `protobuf:"bytes,2,opt,name=ack_id,json=ackId" json:"ack_id,omitempty"` + AckId string `protobuf:"bytes,2,opt,name=ack_id,json=ackId,proto3" json:"ack_id,omitempty"` // The new ack deadline with respect to the time this request was sent to the // Pub/Sub system. Must be >= 0. For example, if the value is 10, the new ack // deadline will expire 10 seconds after the ModifyAckDeadline call was made. // Specifying zero may immediately make the message available for another pull // request. - AckDeadlineSeconds int32 `protobuf:"varint,3,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds" json:"ack_deadline_seconds,omitempty"` + AckDeadlineSeconds int32 `protobuf:"varint,3,opt,name=ack_deadline_seconds,json=ackDeadlineSeconds,proto3" json:"ack_deadline_seconds,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1140,10 +1140,10 @@ func (m *ModifyAckDeadlineRequest) GetAckDeadlineSeconds() int32 { // Request for the Acknowledge method. type AcknowledgeRequest struct { // The subscription whose message is being acknowledged. - Subscription string `protobuf:"bytes,1,opt,name=subscription" json:"subscription,omitempty"` + Subscription string `protobuf:"bytes,1,opt,name=subscription,proto3" json:"subscription,omitempty"` // The acknowledgment ID for the messages being acknowledged that was returned // by the Pub/Sub system in the Pull response. Must not be empty. - AckIds []string `protobuf:"bytes,2,rep,name=ack_ids,json=ackIds" json:"ack_ids,omitempty"` + AckIds []string `protobuf:"bytes,2,rep,name=ack_ids,json=ackIds,proto3" json:"ack_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1351,8 +1351,7 @@ func (c *subscriberClient) ModifyPushConfig(ctx context.Context, in *ModifyPushC return out, nil } -// Server API for Subscriber service - +// SubscriberServer is the server API for Subscriber service. type SubscriberServer interface { // Creates a subscription to a given topic for a given subscriber. // If the subscription already exists, returns ALREADY_EXISTS. @@ -1673,8 +1672,7 @@ func (c *publisherClient) DeleteTopic(ctx context.Context, in *DeleteTopicReques return out, nil } -// Server API for Publisher service - +// PublisherServer is the server API for Publisher service. type PublisherServer interface { // Creates the given topic with the given name. CreateTopic(context.Context, *Topic) (*Topic, error) diff --git a/vendor/google.golang.org/genproto/googleapis/rpc/errdetails/error_details.pb.go b/vendor/google.golang.org/genproto/googleapis/rpc/errdetails/error_details.pb.go index 00a8c6d12..c65f8b40f 100644 --- a/vendor/google.golang.org/genproto/googleapis/rpc/errdetails/error_details.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/rpc/errdetails/error_details.pb.go @@ -34,7 +34,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // reached. type RetryInfo struct { // Clients should wait at least this long between retrying the same request. - RetryDelay *duration.Duration `protobuf:"bytes,1,opt,name=retry_delay,json=retryDelay" json:"retry_delay,omitempty"` + RetryDelay *duration.Duration `protobuf:"bytes,1,opt,name=retry_delay,json=retryDelay,proto3" json:"retry_delay,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -74,9 +74,9 @@ func (m *RetryInfo) GetRetryDelay() *duration.Duration { // Describes additional debugging info. type DebugInfo struct { // The stack trace entries indicating where the error occurred. - StackEntries []string `protobuf:"bytes,1,rep,name=stack_entries,json=stackEntries" json:"stack_entries,omitempty"` + StackEntries []string `protobuf:"bytes,1,rep,name=stack_entries,json=stackEntries,proto3" json:"stack_entries,omitempty"` // Additional debugging information provided by the server. - Detail string `protobuf:"bytes,2,opt,name=detail" json:"detail,omitempty"` + Detail string `protobuf:"bytes,2,opt,name=detail,proto3" json:"detail,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -133,7 +133,7 @@ func (m *DebugInfo) GetDetail() string { // quota failure. type QuotaFailure struct { // Describes all quota violations. - Violations []*QuotaFailure_Violation `protobuf:"bytes,1,rep,name=violations" json:"violations,omitempty"` + Violations []*QuotaFailure_Violation `protobuf:"bytes,1,rep,name=violations,proto3" json:"violations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -176,7 +176,7 @@ type QuotaFailure_Violation struct { // The subject on which the quota check failed. // For example, "clientip:" or "project:". - Subject string `protobuf:"bytes,1,opt,name=subject" json:"subject,omitempty"` + Subject string `protobuf:"bytes,1,opt,name=subject,proto3" json:"subject,omitempty"` // A description of how the quota check failed. Clients can use this // description to find more about the quota configuration in the service's // public documentation, or find the relevant quota limit to adjust through @@ -184,7 +184,7 @@ type QuotaFailure_Violation struct { // // For example: "Service disabled" or "Daily Limit for read operations // exceeded". - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -235,7 +235,7 @@ func (m *QuotaFailure_Violation) GetDescription() string { // PreconditionFailure message. type PreconditionFailure struct { // Describes all precondition violations. - Violations []*PreconditionFailure_Violation `protobuf:"bytes,1,rep,name=violations" json:"violations,omitempty"` + Violations []*PreconditionFailure_Violation `protobuf:"bytes,1,rep,name=violations,proto3" json:"violations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -277,16 +277,16 @@ type PreconditionFailure_Violation struct { // The type of PreconditionFailure. We recommend using a service-specific // enum type to define the supported precondition violation types. For // example, "TOS" for "Terms of Service violation". - Type string `protobuf:"bytes,1,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` // The subject, relative to the type, that failed. // For example, "google.com/cloud" relative to the "TOS" type would // indicate which terms of service is being referenced. - Subject string `protobuf:"bytes,2,opt,name=subject" json:"subject,omitempty"` + Subject string `protobuf:"bytes,2,opt,name=subject,proto3" json:"subject,omitempty"` // A description of how the precondition failed. Developers can use this // description to understand how to fix the failure. // // For example: "Terms of service not accepted". - Description string `protobuf:"bytes,3,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,3,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -341,7 +341,7 @@ func (m *PreconditionFailure_Violation) GetDescription() string { // syntactic aspects of the request. type BadRequest struct { // Describes all violations in a client request. - FieldViolations []*BadRequest_FieldViolation `protobuf:"bytes,1,rep,name=field_violations,json=fieldViolations" json:"field_violations,omitempty"` + FieldViolations []*BadRequest_FieldViolation `protobuf:"bytes,1,rep,name=field_violations,json=fieldViolations,proto3" json:"field_violations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -383,9 +383,9 @@ type BadRequest_FieldViolation struct { // A path leading to a field in the request body. The value will be a // sequence of dot-separated identifiers that identify a protocol buffer // field. E.g., "field_violations.field" would identify this field. - Field string `protobuf:"bytes,1,opt,name=field" json:"field,omitempty"` + Field string `protobuf:"bytes,1,opt,name=field,proto3" json:"field,omitempty"` // A description of why the request element is bad. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -434,10 +434,10 @@ func (m *BadRequest_FieldViolation) GetDescription() string { type RequestInfo struct { // An opaque string that should only be interpreted by the service generating // it. For example, it can be used to identify requests in the service's logs. - RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId" json:"request_id,omitempty"` + RequestId string `protobuf:"bytes,1,opt,name=request_id,json=requestId,proto3" json:"request_id,omitempty"` // Any data that was used to serve this request. For example, an encrypted // stack trace that can be sent back to the service provider for debugging. - ServingData string `protobuf:"bytes,2,opt,name=serving_data,json=servingData" json:"serving_data,omitempty"` + ServingData string `protobuf:"bytes,2,opt,name=serving_data,json=servingData,proto3" json:"serving_data,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -486,19 +486,19 @@ type ResourceInfo struct { // A name for the type of resource being accessed, e.g. "sql table", // "cloud storage bucket", "file", "Google calendar"; or the type URL // of the resource: e.g. "type.googleapis.com/google.pubsub.v1.Topic". - ResourceType string `protobuf:"bytes,1,opt,name=resource_type,json=resourceType" json:"resource_type,omitempty"` + ResourceType string `protobuf:"bytes,1,opt,name=resource_type,json=resourceType,proto3" json:"resource_type,omitempty"` // The name of the resource being accessed. For example, a shared calendar // name: "example.com_4fghdhgsrgh@group.calendar.google.com", if the current // error is [google.rpc.Code.PERMISSION_DENIED][google.rpc.Code.PERMISSION_DENIED]. - ResourceName string `protobuf:"bytes,2,opt,name=resource_name,json=resourceName" json:"resource_name,omitempty"` + ResourceName string `protobuf:"bytes,2,opt,name=resource_name,json=resourceName,proto3" json:"resource_name,omitempty"` // The owner of the resource (optional). // For example, "user:" or "project:". - Owner string `protobuf:"bytes,3,opt,name=owner" json:"owner,omitempty"` + Owner string `protobuf:"bytes,3,opt,name=owner,proto3" json:"owner,omitempty"` // Describes what error is encountered when accessing this resource. // For example, updating a cloud project may require the `writer` permission // on the developer console project. - Description string `protobuf:"bytes,4,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,4,opt,name=description,proto3" json:"description,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -563,7 +563,7 @@ func (m *ResourceInfo) GetDescription() string { // directly to the right place in the developer console to flip the bit. type Help struct { // URL(s) pointing to additional information on handling the current error. - Links []*Help_Link `protobuf:"bytes,1,rep,name=links" json:"links,omitempty"` + Links []*Help_Link `protobuf:"bytes,1,rep,name=links,proto3" json:"links,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -603,9 +603,9 @@ func (m *Help) GetLinks() []*Help_Link { // Describes a URL link. type Help_Link struct { // Describes what the link offers. - Description string `protobuf:"bytes,1,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` // The URL of the link. - Url string `protobuf:"bytes,2,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,2,opt,name=url,proto3" json:"url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -655,9 +655,9 @@ type LocalizedMessage struct { // The locale used following the specification defined at // http://www.rfc-editor.org/rfc/bcp/bcp47.txt. // Examples are: "en-US", "fr-CH", "es-MX" - Locale string `protobuf:"bytes,1,opt,name=locale" json:"locale,omitempty"` + Locale string `protobuf:"bytes,1,opt,name=locale,proto3" json:"locale,omitempty"` // The localized error message in the above locale. - Message string `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go b/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go index fb104e6df..3b07a25b7 100644 --- a/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/rpc/status/status.pb.go @@ -73,14 +73,14 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // be used directly after any stripping needed for security/privacy reasons. type Status struct { // The status code, which should be an enum value of [google.rpc.Code][google.rpc.Code]. - Code int32 `protobuf:"varint,1,opt,name=code" json:"code,omitempty"` + Code int32 `protobuf:"varint,1,opt,name=code,proto3" json:"code,omitempty"` // A developer-facing error message, which should be in English. Any // user-facing error message should be localized and sent in the // [google.rpc.Status.details][google.rpc.Status.details] field, or localized by the client. - Message string `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` // A list of messages that carry the error details. There is a common set of // message types for APIs to use. - Details []*any.Any `protobuf:"bytes,3,rep,name=details" json:"details,omitempty"` + Details []*any.Any `protobuf:"bytes,3,rep,name=details,proto3" json:"details,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/admin/database/v1/spanner_database_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/admin/database/v1/spanner_database_admin.pb.go index eb867e79c..f381072fc 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/admin/database/v1/spanner_database_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/admin/database/v1/spanner_database_admin.pb.go @@ -66,9 +66,9 @@ type Database struct { // where `` is as specified in the `CREATE DATABASE` // statement. This name can be passed to other API methods to // identify the database. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Output only. The current database state. - State Database_State `protobuf:"varint,2,opt,name=state,enum=google.spanner.admin.database.v1.Database_State" json:"state,omitempty"` + State Database_State `protobuf:"varint,2,opt,name=state,proto3,enum=google.spanner.admin.database.v1.Database_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -116,14 +116,14 @@ func (m *Database) GetState() Database_State { type ListDatabasesRequest struct { // Required. The instance whose databases should be listed. // Values are of the form `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Number of databases to be returned in the response. If 0 or less, // defaults to the server's maximum allowed page size. - PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,3,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If non-empty, `page_token` should contain a // [next_page_token][google.spanner.admin.database.v1.ListDatabasesResponse.next_page_token] from a // previous [ListDatabasesResponse][google.spanner.admin.database.v1.ListDatabasesResponse]. - PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,4,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -177,11 +177,11 @@ func (m *ListDatabasesRequest) GetPageToken() string { // The response for [ListDatabases][google.spanner.admin.database.v1.DatabaseAdmin.ListDatabases]. type ListDatabasesResponse struct { // Databases that matched the request. - Databases []*Database `protobuf:"bytes,1,rep,name=databases" json:"databases,omitempty"` + Databases []*Database `protobuf:"bytes,1,rep,name=databases,proto3" json:"databases,omitempty"` // `next_page_token` can be sent in a subsequent // [ListDatabases][google.spanner.admin.database.v1.DatabaseAdmin.ListDatabases] call to fetch more // of the matching databases. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -229,18 +229,18 @@ func (m *ListDatabasesResponse) GetNextPageToken() string { type CreateDatabaseRequest struct { // Required. The name of the instance that will serve the new database. // Values are of the form `projects//instances/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. A `CREATE DATABASE` statement, which specifies the ID of the // new database. The database ID must conform to the regular expression // `[a-z][a-z0-9_\-]*[a-z0-9]` and be between 2 and 30 characters in length. // If the database ID is a reserved word or if it contains a hyphen, the // database ID must be enclosed in backticks (`` ` ``). - CreateStatement string `protobuf:"bytes,2,opt,name=create_statement,json=createStatement" json:"create_statement,omitempty"` + CreateStatement string `protobuf:"bytes,2,opt,name=create_statement,json=createStatement,proto3" json:"create_statement,omitempty"` // An optional list of DDL statements to run inside the newly created // database. Statements can create tables, indexes, etc. These // statements execute atomically with the creation of the database: // if there is an error in any statement, the database is not created. - ExtraStatements []string `protobuf:"bytes,3,rep,name=extra_statements,json=extraStatements" json:"extra_statements,omitempty"` + ExtraStatements []string `protobuf:"bytes,3,rep,name=extra_statements,json=extraStatements,proto3" json:"extra_statements,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -295,7 +295,7 @@ func (m *CreateDatabaseRequest) GetExtraStatements() []string { // [CreateDatabase][google.spanner.admin.database.v1.DatabaseAdmin.CreateDatabase]. type CreateDatabaseMetadata struct { // The database being created. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -336,7 +336,7 @@ func (m *CreateDatabaseMetadata) GetDatabase() string { type GetDatabaseRequest struct { // Required. The name of the requested database. Values are of the form // `projects//instances//databases/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -391,9 +391,9 @@ func (m *GetDatabaseRequest) GetName() string { // details. type UpdateDatabaseDdlRequest struct { // Required. The database to update. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // DDL statements to be applied to the database. - Statements []string `protobuf:"bytes,2,rep,name=statements" json:"statements,omitempty"` + Statements []string `protobuf:"bytes,2,rep,name=statements,proto3" json:"statements,omitempty"` // If empty, the new update request is assigned an // automatically-generated operation ID. Otherwise, `operation_id` // is used to construct the name of the resulting @@ -413,7 +413,7 @@ type UpdateDatabaseDdlRequest struct { // underscore. If the named operation already exists, // [UpdateDatabaseDdl][google.spanner.admin.database.v1.DatabaseAdmin.UpdateDatabaseDdl] returns // `ALREADY_EXISTS`. - OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId" json:"operation_id,omitempty"` + OperationId string `protobuf:"bytes,3,opt,name=operation_id,json=operationId,proto3" json:"operation_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -468,14 +468,14 @@ func (m *UpdateDatabaseDdlRequest) GetOperationId() string { // [UpdateDatabaseDdl][google.spanner.admin.database.v1.DatabaseAdmin.UpdateDatabaseDdl]. type UpdateDatabaseDdlMetadata struct { // The database being modified. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // For an update this list contains all the statements. For an // individual statement, this list contains only that statement. - Statements []string `protobuf:"bytes,2,rep,name=statements" json:"statements,omitempty"` + Statements []string `protobuf:"bytes,2,rep,name=statements,proto3" json:"statements,omitempty"` // Reports the commit timestamps of all statements that have // succeeded so far, where `commit_timestamps[i]` is the commit // timestamp for the statement `statements[i]`. - CommitTimestamps []*timestamp.Timestamp `protobuf:"bytes,3,rep,name=commit_timestamps,json=commitTimestamps" json:"commit_timestamps,omitempty"` + CommitTimestamps []*timestamp.Timestamp `protobuf:"bytes,3,rep,name=commit_timestamps,json=commitTimestamps,proto3" json:"commit_timestamps,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -529,7 +529,7 @@ func (m *UpdateDatabaseDdlMetadata) GetCommitTimestamps() []*timestamp.Timestamp // The request for [DropDatabase][google.spanner.admin.database.v1.DatabaseAdmin.DropDatabase]. type DropDatabaseRequest struct { // Required. The database to be dropped. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -569,7 +569,7 @@ func (m *DropDatabaseRequest) GetDatabase() string { // The request for [GetDatabaseDdl][google.spanner.admin.database.v1.DatabaseAdmin.GetDatabaseDdl]. type GetDatabaseDdlRequest struct { // Required. The database whose schema we wish to get. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -610,7 +610,7 @@ func (m *GetDatabaseDdlRequest) GetDatabase() string { type GetDatabaseDdlResponse struct { // A list of formatted DDL statements defining the schema of the database // specified in the request. - Statements []string `protobuf:"bytes,1,rep,name=statements" json:"statements,omitempty"` + Statements []string `protobuf:"bytes,1,rep,name=statements,proto3" json:"statements,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -811,8 +811,7 @@ func (c *databaseAdminClient) TestIamPermissions(ctx context.Context, in *v1.Tes return out, nil } -// Server API for DatabaseAdmin service - +// DatabaseAdminServer is the server API for DatabaseAdmin service. type DatabaseAdminServer interface { // Lists Cloud Spanner databases. ListDatabases(context.Context, *ListDatabasesRequest) (*ListDatabasesResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/admin/instance/v1/spanner_instance_admin.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/admin/instance/v1/spanner_instance_admin.pb.go index 1708a8365..56dfcb240 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/admin/instance/v1/spanner_instance_admin.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/admin/instance/v1/spanner_instance_admin.pb.go @@ -68,9 +68,9 @@ type InstanceConfig struct { // A unique identifier for the instance configuration. Values // are of the form // `projects//instanceConfigs/[a-z][-a-z0-9]*` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The name of this instance configuration as it appears in UIs. - DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,2,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -120,27 +120,27 @@ type Instance struct { // after the instance is created. Values are of the form // `projects//instances/[a-z][-a-z0-9]*[a-z0-9]`. The final // segment of the name must be between 6 and 30 characters in length. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Required. The name of the instance's configuration. Values are of the form // `projects//instanceConfigs/`. See // also [InstanceConfig][google.spanner.admin.instance.v1.InstanceConfig] and // [ListInstanceConfigs][google.spanner.admin.instance.v1.InstanceAdmin.ListInstanceConfigs]. - Config string `protobuf:"bytes,2,opt,name=config" json:"config,omitempty"` + Config string `protobuf:"bytes,2,opt,name=config,proto3" json:"config,omitempty"` // Required. The descriptive name for this instance as it appears in UIs. // Must be unique per project and between 4 and 30 characters in length. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // Required. The number of nodes allocated to this instance. This may be zero // in API responses for instances that are not yet in state `READY`. // // See [the documentation](https://cloud.google.com/spanner/docs/instances#node_count) // for more information about nodes. - NodeCount int32 `protobuf:"varint,5,opt,name=node_count,json=nodeCount" json:"node_count,omitempty"` + NodeCount int32 `protobuf:"varint,5,opt,name=node_count,json=nodeCount,proto3" json:"node_count,omitempty"` // Output only. The current instance state. For // [CreateInstance][google.spanner.admin.instance.v1.InstanceAdmin.CreateInstance], the state must be // either omitted or set to `CREATING`. For // [UpdateInstance][google.spanner.admin.instance.v1.InstanceAdmin.UpdateInstance], the state must be // either omitted or set to `READY`. - State Instance_State `protobuf:"varint,6,opt,name=state,enum=google.spanner.admin.instance.v1.Instance_State" json:"state,omitempty"` + State Instance_State `protobuf:"varint,6,opt,name=state,proto3,enum=google.spanner.admin.instance.v1.Instance_State" json:"state,omitempty"` // Cloud Labels are a flexible and lightweight mechanism for organizing cloud // resources into groups that reflect a customer's organizational needs and // deployment strategies. Cloud Labels can be used to filter collections of @@ -162,7 +162,7 @@ type Instance struct { // specific characters being disallowed. For example, representing labels // as the string: name + "_" + value would prove problematic if we were to // allow "_" in a future release. - Labels map[string]string `protobuf:"bytes,7,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,7,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -239,14 +239,14 @@ type ListInstanceConfigsRequest struct { // Required. The name of the project for which a list of supported instance // configurations is requested. Values are of the form // `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Number of instance configurations to be returned in the response. If 0 or // less, defaults to the server's maximum allowed page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If non-empty, `page_token` should contain a // [next_page_token][google.spanner.admin.instance.v1.ListInstanceConfigsResponse.next_page_token] // from a previous [ListInstanceConfigsResponse][google.spanner.admin.instance.v1.ListInstanceConfigsResponse]. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -300,11 +300,11 @@ func (m *ListInstanceConfigsRequest) GetPageToken() string { // The response for [ListInstanceConfigs][google.spanner.admin.instance.v1.InstanceAdmin.ListInstanceConfigs]. type ListInstanceConfigsResponse struct { // The list of requested instance configurations. - InstanceConfigs []*InstanceConfig `protobuf:"bytes,1,rep,name=instance_configs,json=instanceConfigs" json:"instance_configs,omitempty"` + InstanceConfigs []*InstanceConfig `protobuf:"bytes,1,rep,name=instance_configs,json=instanceConfigs,proto3" json:"instance_configs,omitempty"` // `next_page_token` can be sent in a subsequent // [ListInstanceConfigs][google.spanner.admin.instance.v1.InstanceAdmin.ListInstanceConfigs] call to // fetch more of the matching instance configurations. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -353,7 +353,7 @@ func (m *ListInstanceConfigsResponse) GetNextPageToken() string { type GetInstanceConfigRequest struct { // Required. The name of the requested instance configuration. Values are of // the form `projects//instanceConfigs/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -394,7 +394,7 @@ func (m *GetInstanceConfigRequest) GetName() string { type GetInstanceRequest struct { // Required. The name of the requested instance. Values are of the form // `projects//instances/`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -435,14 +435,14 @@ func (m *GetInstanceRequest) GetName() string { type CreateInstanceRequest struct { // Required. The name of the project in which to create the instance. Values // are of the form `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Required. The ID of the instance to create. Valid identifiers are of the // form `[a-z][-a-z0-9]*[a-z0-9]` and must be between 6 and 30 characters in // length. - InstanceId string `protobuf:"bytes,2,opt,name=instance_id,json=instanceId" json:"instance_id,omitempty"` + InstanceId string `protobuf:"bytes,2,opt,name=instance_id,json=instanceId,proto3" json:"instance_id,omitempty"` // Required. The instance to create. The name may be omitted, but if // specified must be `/instances/`. - Instance *Instance `protobuf:"bytes,3,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,3,opt,name=instance,proto3" json:"instance,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -497,14 +497,14 @@ func (m *CreateInstanceRequest) GetInstance() *Instance { type ListInstancesRequest struct { // Required. The name of the project for which a list of instances is // requested. Values are of the form `projects/`. - Parent string `protobuf:"bytes,1,opt,name=parent" json:"parent,omitempty"` + Parent string `protobuf:"bytes,1,opt,name=parent,proto3" json:"parent,omitempty"` // Number of instances to be returned in the response. If 0 or less, defaults // to the server's maximum allowed page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If non-empty, `page_token` should contain a // [next_page_token][google.spanner.admin.instance.v1.ListInstancesResponse.next_page_token] from a // previous [ListInstancesResponse][google.spanner.admin.instance.v1.ListInstancesResponse]. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // An expression for filtering the results of the request. Filter rules are // case insensitive. The fields eligible for filtering are: // @@ -524,7 +524,7 @@ type ListInstancesRequest struct { // * `name:howl labels.env:dev` --> The instance's name contains "howl" and // it has the label "env" with its value // containing "dev". - Filter string `protobuf:"bytes,4,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,4,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -585,11 +585,11 @@ func (m *ListInstancesRequest) GetFilter() string { // The response for [ListInstances][google.spanner.admin.instance.v1.InstanceAdmin.ListInstances]. type ListInstancesResponse struct { // The list of requested instances. - Instances []*Instance `protobuf:"bytes,1,rep,name=instances" json:"instances,omitempty"` + Instances []*Instance `protobuf:"bytes,1,rep,name=instances,proto3" json:"instances,omitempty"` // `next_page_token` can be sent in a subsequent // [ListInstances][google.spanner.admin.instance.v1.InstanceAdmin.ListInstances] call to fetch more // of the matching instances. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -637,12 +637,12 @@ func (m *ListInstancesResponse) GetNextPageToken() string { type UpdateInstanceRequest struct { // Required. The instance to update, which must always include the instance // name. Otherwise, only fields mentioned in [][google.spanner.admin.instance.v1.UpdateInstanceRequest.field_mask] need be included. - Instance *Instance `protobuf:"bytes,1,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,1,opt,name=instance,proto3" json:"instance,omitempty"` // Required. A mask specifying which fields in [][google.spanner.admin.instance.v1.UpdateInstanceRequest.instance] should be updated. // The field mask must always be specified; this prevents any future fields in // [][google.spanner.admin.instance.v1.Instance] from being erased accidentally by clients that do not know // about them. - FieldMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=field_mask,json=fieldMask" json:"field_mask,omitempty"` + FieldMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=field_mask,json=fieldMask,proto3" json:"field_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -690,7 +690,7 @@ func (m *UpdateInstanceRequest) GetFieldMask() *field_mask.FieldMask { type DeleteInstanceRequest struct { // Required. The name of the instance to be deleted. Values are of the form // `projects//instances/` - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -731,17 +731,17 @@ func (m *DeleteInstanceRequest) GetName() string { // [CreateInstance][google.spanner.admin.instance.v1.InstanceAdmin.CreateInstance]. type CreateInstanceMetadata struct { // The instance being created. - Instance *Instance `protobuf:"bytes,1,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,1,opt,name=instance,proto3" json:"instance,omitempty"` // The time at which the // [CreateInstance][google.spanner.admin.instance.v1.InstanceAdmin.CreateInstance] request was // received. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time at which this operation was cancelled. If set, this operation is // in the process of undoing itself (which is guaranteed to succeed) and // cannot be cancelled again. - CancelTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=cancel_time,json=cancelTime" json:"cancel_time,omitempty"` + CancelTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=cancel_time,json=cancelTime,proto3" json:"cancel_time,omitempty"` // The time at which this operation failed or was completed successfully. - EndTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -803,16 +803,16 @@ func (m *CreateInstanceMetadata) GetEndTime() *timestamp.Timestamp { // [UpdateInstance][google.spanner.admin.instance.v1.InstanceAdmin.UpdateInstance]. type UpdateInstanceMetadata struct { // The desired end state of the update. - Instance *Instance `protobuf:"bytes,1,opt,name=instance" json:"instance,omitempty"` + Instance *Instance `protobuf:"bytes,1,opt,name=instance,proto3" json:"instance,omitempty"` // The time at which [UpdateInstance][google.spanner.admin.instance.v1.InstanceAdmin.UpdateInstance] // request was received. - StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // The time at which this operation was cancelled. If set, this operation is // in the process of undoing itself (which is guaranteed to succeed) and // cannot be cancelled again. - CancelTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=cancel_time,json=cancelTime" json:"cancel_time,omitempty"` + CancelTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=cancel_time,json=cancelTime,proto3" json:"cancel_time,omitempty"` // The time at which this operation failed or was completed successfully. - EndTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1115,8 +1115,7 @@ func (c *instanceAdminClient) TestIamPermissions(ctx context.Context, in *v1.Tes return out, nil } -// Server API for InstanceAdmin service - +// InstanceAdminServer is the server API for InstanceAdmin service. type InstanceAdminServer interface { // Lists the supported instance configurations for a given project. ListInstanceConfigs(context.Context, *ListInstanceConfigsRequest) (*ListInstanceConfigsResponse, error) diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/keys.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/keys.pb.go index a3e365db3..c93e58987 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/keys.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/keys.pb.go @@ -157,16 +157,16 @@ type isKeyRange_EndKeyType interface { } type KeyRange_StartClosed struct { - StartClosed *_struct.ListValue `protobuf:"bytes,1,opt,name=start_closed,json=startClosed,oneof"` + StartClosed *_struct.ListValue `protobuf:"bytes,1,opt,name=start_closed,json=startClosed,proto3,oneof"` } type KeyRange_StartOpen struct { - StartOpen *_struct.ListValue `protobuf:"bytes,2,opt,name=start_open,json=startOpen,oneof"` + StartOpen *_struct.ListValue `protobuf:"bytes,2,opt,name=start_open,json=startOpen,proto3,oneof"` } type KeyRange_EndClosed struct { - EndClosed *_struct.ListValue `protobuf:"bytes,3,opt,name=end_closed,json=endClosed,oneof"` + EndClosed *_struct.ListValue `protobuf:"bytes,3,opt,name=end_closed,json=endClosed,proto3,oneof"` } type KeyRange_EndOpen struct { - EndOpen *_struct.ListValue `protobuf:"bytes,4,opt,name=end_open,json=endOpen,oneof"` + EndOpen *_struct.ListValue `protobuf:"bytes,4,opt,name=end_open,json=endOpen,proto3,oneof"` } func (*KeyRange_StartClosed) isKeyRange_StartKeyType() {} @@ -351,14 +351,14 @@ type KeySet struct { // many elements as there are columns in the primary or index key // with which this `KeySet` is used. Individual key values are // encoded as described [here][google.spanner.v1.TypeCode]. - Keys []*_struct.ListValue `protobuf:"bytes,1,rep,name=keys" json:"keys,omitempty"` + Keys []*_struct.ListValue `protobuf:"bytes,1,rep,name=keys,proto3" json:"keys,omitempty"` // A list of key ranges. See [KeyRange][google.spanner.v1.KeyRange] for more information about // key range specifications. - Ranges []*KeyRange `protobuf:"bytes,2,rep,name=ranges" json:"ranges,omitempty"` + Ranges []*KeyRange `protobuf:"bytes,2,rep,name=ranges,proto3" json:"ranges,omitempty"` // For convenience `all` can be set to `true` to indicate that this // `KeySet` matches all keys in the table or index. Note that any keys // specified in `keys` or `ranges` are only yielded once. - All bool `protobuf:"varint,3,opt,name=all" json:"all,omitempty"` + All bool `protobuf:"varint,3,opt,name=all,proto3" json:"all,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/mutation.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/mutation.pb.go index fb3bdbb82..e6564cf5d 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/mutation.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/mutation.pb.go @@ -67,19 +67,19 @@ type isMutation_Operation interface { } type Mutation_Insert struct { - Insert *Mutation_Write `protobuf:"bytes,1,opt,name=insert,oneof"` + Insert *Mutation_Write `protobuf:"bytes,1,opt,name=insert,proto3,oneof"` } type Mutation_Update struct { - Update *Mutation_Write `protobuf:"bytes,2,opt,name=update,oneof"` + Update *Mutation_Write `protobuf:"bytes,2,opt,name=update,proto3,oneof"` } type Mutation_InsertOrUpdate struct { - InsertOrUpdate *Mutation_Write `protobuf:"bytes,3,opt,name=insert_or_update,json=insertOrUpdate,oneof"` + InsertOrUpdate *Mutation_Write `protobuf:"bytes,3,opt,name=insert_or_update,json=insertOrUpdate,proto3,oneof"` } type Mutation_Replace struct { - Replace *Mutation_Write `protobuf:"bytes,4,opt,name=replace,oneof"` + Replace *Mutation_Write `protobuf:"bytes,4,opt,name=replace,proto3,oneof"` } type Mutation_Delete_ struct { - Delete *Mutation_Delete `protobuf:"bytes,5,opt,name=delete,oneof"` + Delete *Mutation_Delete `protobuf:"bytes,5,opt,name=delete,proto3,oneof"` } func (*Mutation_Insert) isMutation_Operation() {} @@ -265,13 +265,13 @@ func _Mutation_OneofSizer(msg proto.Message) (n int) { // [replace][google.spanner.v1.Mutation.replace] operations. type Mutation_Write struct { // Required. The table whose rows will be written. - Table string `protobuf:"bytes,1,opt,name=table" json:"table,omitempty"` + Table string `protobuf:"bytes,1,opt,name=table,proto3" json:"table,omitempty"` // The names of the columns in [table][google.spanner.v1.Mutation.Write.table] to be written. // // The list of columns must contain enough columns to allow // Cloud Spanner to derive values for all primary key columns in the // row(s) to be modified. - Columns []string `protobuf:"bytes,2,rep,name=columns" json:"columns,omitempty"` + Columns []string `protobuf:"bytes,2,rep,name=columns,proto3" json:"columns,omitempty"` // The values to be written. `values` can contain more than one // list of values. If it does, then multiple rows are written, one // for each entry in `values`. Each list in `values` must have @@ -280,7 +280,7 @@ type Mutation_Write struct { // `Mutation`s, each containing one `values` entry and repeating // [table][google.spanner.v1.Mutation.Write.table] and [columns][google.spanner.v1.Mutation.Write.columns]. Individual values in each list are // encoded as described [here][google.spanner.v1.TypeCode]. - Values []*_struct.ListValue `protobuf:"bytes,3,rep,name=values" json:"values,omitempty"` + Values []*_struct.ListValue `protobuf:"bytes,3,rep,name=values,proto3" json:"values,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -334,11 +334,11 @@ func (m *Mutation_Write) GetValues() []*_struct.ListValue { // Arguments to [delete][google.spanner.v1.Mutation.delete] operations. type Mutation_Delete struct { // Required. The table whose rows will be deleted. - Table string `protobuf:"bytes,1,opt,name=table" json:"table,omitempty"` + Table string `protobuf:"bytes,1,opt,name=table,proto3" json:"table,omitempty"` // Required. The primary keys of the rows within [table][google.spanner.v1.Mutation.Delete.table] to delete. // Delete is idempotent. The transaction will succeed even if some or all // rows do not exist. - KeySet *KeySet `protobuf:"bytes,2,opt,name=key_set,json=keySet" json:"key_set,omitempty"` + KeySet *KeySet `protobuf:"bytes,2,opt,name=key_set,json=keySet,proto3" json:"key_set,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/query_plan.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/query_plan.pb.go index 063b270c7..cc528a7d6 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/query_plan.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/query_plan.pb.go @@ -59,19 +59,19 @@ func (PlanNode_Kind) EnumDescriptor() ([]byte, []int) { // Node information for nodes appearing in a [QueryPlan.plan_nodes][google.spanner.v1.QueryPlan.plan_nodes]. type PlanNode struct { // The `PlanNode`'s index in [node list][google.spanner.v1.QueryPlan.plan_nodes]. - Index int32 `protobuf:"varint,1,opt,name=index" json:"index,omitempty"` + Index int32 `protobuf:"varint,1,opt,name=index,proto3" json:"index,omitempty"` // Used to determine the type of node. May be needed for visualizing // different kinds of nodes differently. For example, If the node is a // [SCALAR][google.spanner.v1.PlanNode.Kind.SCALAR] node, it will have a condensed representation // which can be used to directly embed a description of the node in its // parent. - Kind PlanNode_Kind `protobuf:"varint,2,opt,name=kind,enum=google.spanner.v1.PlanNode_Kind" json:"kind,omitempty"` + Kind PlanNode_Kind `protobuf:"varint,2,opt,name=kind,proto3,enum=google.spanner.v1.PlanNode_Kind" json:"kind,omitempty"` // The display name for the node. - DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"` + DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` // List of child node `index`es and their relationship to this parent. - ChildLinks []*PlanNode_ChildLink `protobuf:"bytes,4,rep,name=child_links,json=childLinks" json:"child_links,omitempty"` + ChildLinks []*PlanNode_ChildLink `protobuf:"bytes,4,rep,name=child_links,json=childLinks,proto3" json:"child_links,omitempty"` // Condensed representation for [SCALAR][google.spanner.v1.PlanNode.Kind.SCALAR] nodes. - ShortRepresentation *PlanNode_ShortRepresentation `protobuf:"bytes,5,opt,name=short_representation,json=shortRepresentation" json:"short_representation,omitempty"` + ShortRepresentation *PlanNode_ShortRepresentation `protobuf:"bytes,5,opt,name=short_representation,json=shortRepresentation,proto3" json:"short_representation,omitempty"` // Attributes relevant to the node contained in a group of key-value pairs. // For example, a Parameter Reference node could have the following // information in its metadata: @@ -80,12 +80,12 @@ type PlanNode struct { // "parameter_reference": "param1", // "parameter_type": "array" // } - Metadata *_struct.Struct `protobuf:"bytes,6,opt,name=metadata" json:"metadata,omitempty"` + Metadata *_struct.Struct `protobuf:"bytes,6,opt,name=metadata,proto3" json:"metadata,omitempty"` // The execution statistics associated with the node, contained in a group of // key-value pairs. Only present if the plan was returned as a result of a // profile query. For example, number of executions, number of rows/time per // execution etc. - ExecutionStats *_struct.Struct `protobuf:"bytes,7,opt,name=execution_stats,json=executionStats" json:"execution_stats,omitempty"` + ExecutionStats *_struct.Struct `protobuf:"bytes,7,opt,name=execution_stats,json=executionStats,proto3" json:"execution_stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -168,12 +168,12 @@ func (m *PlanNode) GetExecutionStats() *_struct.Struct { // [PlanNode][google.spanner.v1.PlanNode]. type PlanNode_ChildLink struct { // The node to which the link points. - ChildIndex int32 `protobuf:"varint,1,opt,name=child_index,json=childIndex" json:"child_index,omitempty"` + ChildIndex int32 `protobuf:"varint,1,opt,name=child_index,json=childIndex,proto3" json:"child_index,omitempty"` // The type of the link. For example, in Hash Joins this could be used to // distinguish between the build child and the probe child, or in the case // of the child being an output variable, to represent the tag associated // with the output variable. - Type string `protobuf:"bytes,2,opt,name=type" json:"type,omitempty"` + Type string `protobuf:"bytes,2,opt,name=type,proto3" json:"type,omitempty"` // Only present if the child node is [SCALAR][google.spanner.v1.PlanNode.Kind.SCALAR] and corresponds // to an output variable of the parent node. The field carries the name of // the output variable. @@ -182,7 +182,7 @@ type PlanNode_ChildLink struct { // created for each column that is read by the operator. The corresponding // `variable` fields will be set to the variable names assigned to the // columns. - Variable string `protobuf:"bytes,3,opt,name=variable" json:"variable,omitempty"` + Variable string `protobuf:"bytes,3,opt,name=variable,proto3" json:"variable,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -237,13 +237,13 @@ func (m *PlanNode_ChildLink) GetVariable() string { // `SCALAR` [PlanNode(s)][google.spanner.v1.PlanNode]. type PlanNode_ShortRepresentation struct { // A string representation of the expression subtree rooted at this node. - Description string `protobuf:"bytes,1,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` // A mapping of (subquery variable name) -> (subquery node id) for cases // where the `description` string of this node references a `SCALAR` // subquery contained in the expression subtree rooted at this node. The // referenced `SCALAR` subquery may not necessarily be a direct child of // this node. - Subqueries map[string]int32 `protobuf:"bytes,2,rep,name=subqueries" json:"subqueries,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"varint,2,opt,name=value"` + Subqueries map[string]int32 `protobuf:"bytes,2,rep,name=subqueries,proto3" json:"subqueries,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"varint,2,opt,name=value,proto3"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -292,7 +292,7 @@ type QueryPlan struct { // The nodes in the query plan. Plan nodes are returned in pre-order starting // with the plan root. Each [PlanNode][google.spanner.v1.PlanNode]'s `id` corresponds to its index in // `plan_nodes`. - PlanNodes []*PlanNode `protobuf:"bytes,1,rep,name=plan_nodes,json=planNodes" json:"plan_nodes,omitempty"` + PlanNodes []*PlanNode `protobuf:"bytes,1,rep,name=plan_nodes,json=planNodes,proto3" json:"plan_nodes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/result_set.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/result_set.pb.go index 64aed86cf..50f498fff 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/result_set.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/result_set.pb.go @@ -24,18 +24,18 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // [ExecuteSql][google.spanner.v1.Spanner.ExecuteSql]. type ResultSet struct { // Metadata about the result set, such as row type information. - Metadata *ResultSetMetadata `protobuf:"bytes,1,opt,name=metadata" json:"metadata,omitempty"` + Metadata *ResultSetMetadata `protobuf:"bytes,1,opt,name=metadata,proto3" json:"metadata,omitempty"` // Each element in `rows` is a row whose format is defined by // [metadata.row_type][google.spanner.v1.ResultSetMetadata.row_type]. The ith element // in each row matches the ith field in // [metadata.row_type][google.spanner.v1.ResultSetMetadata.row_type]. Elements are // encoded based on type as described // [here][google.spanner.v1.TypeCode]. - Rows []*_struct.ListValue `protobuf:"bytes,2,rep,name=rows" json:"rows,omitempty"` + Rows []*_struct.ListValue `protobuf:"bytes,2,rep,name=rows,proto3" json:"rows,omitempty"` // Query plan and execution statistics for the query that produced this // result set. These can be requested by setting // [ExecuteSqlRequest.query_mode][google.spanner.v1.ExecuteSqlRequest.query_mode]. - Stats *ResultSetStats `protobuf:"bytes,3,opt,name=stats" json:"stats,omitempty"` + Stats *ResultSetStats `protobuf:"bytes,3,opt,name=stats,proto3" json:"stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -92,7 +92,7 @@ func (m *ResultSet) GetStats() *ResultSetStats { type PartialResultSet struct { // Metadata about the result set, such as row type information. // Only present in the first response. - Metadata *ResultSetMetadata `protobuf:"bytes,1,opt,name=metadata" json:"metadata,omitempty"` + Metadata *ResultSetMetadata `protobuf:"bytes,1,opt,name=metadata,proto3" json:"metadata,omitempty"` // A streamed result set consists of a stream of values, which might // be split into many `PartialResultSet` messages to accommodate // large rows and/or large values. Every N complete values defines a @@ -166,11 +166,11 @@ type PartialResultSet struct { // This sequence of `PartialResultSet`s encodes two rows, one // containing the field value `"Hello"`, and a second containing the // field value `"World" = "W" + "orl" + "d"`. - Values []*_struct.Value `protobuf:"bytes,2,rep,name=values" json:"values,omitempty"` + Values []*_struct.Value `protobuf:"bytes,2,rep,name=values,proto3" json:"values,omitempty"` // If true, then the final value in [values][google.spanner.v1.PartialResultSet.values] is chunked, and must // be combined with more values from subsequent `PartialResultSet`s // to obtain a complete field value. - ChunkedValue bool `protobuf:"varint,3,opt,name=chunked_value,json=chunkedValue" json:"chunked_value,omitempty"` + ChunkedValue bool `protobuf:"varint,3,opt,name=chunked_value,json=chunkedValue,proto3" json:"chunked_value,omitempty"` // Streaming calls might be interrupted for a variety of reasons, such // as TCP connection loss. If this occurs, the stream of results can // be resumed by re-sending the original request and including @@ -181,7 +181,7 @@ type PartialResultSet struct { // streaming result set. These can be requested by setting // [ExecuteSqlRequest.query_mode][google.spanner.v1.ExecuteSqlRequest.query_mode] and are sent // only once with the last response in the stream. - Stats *ResultSetStats `protobuf:"bytes,5,opt,name=stats" json:"stats,omitempty"` + Stats *ResultSetStats `protobuf:"bytes,5,opt,name=stats,proto3" json:"stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -256,10 +256,10 @@ type ResultSetMetadata struct { // { "name": "UserId", "type": { "code": "INT64" } }, // { "name": "UserName", "type": { "code": "STRING" } }, // ] - RowType *StructType `protobuf:"bytes,1,opt,name=row_type,json=rowType" json:"row_type,omitempty"` + RowType *StructType `protobuf:"bytes,1,opt,name=row_type,json=rowType,proto3" json:"row_type,omitempty"` // If the read or SQL query began a transaction as a side-effect, the // information about the new transaction is yielded here. - Transaction *Transaction `protobuf:"bytes,2,opt,name=transaction" json:"transaction,omitempty"` + Transaction *Transaction `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -306,7 +306,7 @@ func (m *ResultSetMetadata) GetTransaction() *Transaction { // Additional statistics about a [ResultSet][google.spanner.v1.ResultSet] or [PartialResultSet][google.spanner.v1.PartialResultSet]. type ResultSetStats struct { // [QueryPlan][google.spanner.v1.QueryPlan] for the query associated with this result. - QueryPlan *QueryPlan `protobuf:"bytes,1,opt,name=query_plan,json=queryPlan" json:"query_plan,omitempty"` + QueryPlan *QueryPlan `protobuf:"bytes,1,opt,name=query_plan,json=queryPlan,proto3" json:"query_plan,omitempty"` // Aggregated statistics from the execution of the query. Only present when // the query is profiled. For example, a query could return the statistics as // follows: @@ -316,7 +316,7 @@ type ResultSetStats struct { // "elapsed_time": "1.22 secs", // "cpu_time": "1.19 secs" // } - QueryStats *_struct.Struct `protobuf:"bytes,2,opt,name=query_stats,json=queryStats" json:"query_stats,omitempty"` + QueryStats *_struct.Struct `protobuf:"bytes,2,opt,name=query_stats,json=queryStats,proto3" json:"query_stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/spanner.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/spanner.pb.go index 74bf3f98e..59adfb7b4 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/spanner.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/spanner.pb.go @@ -63,9 +63,9 @@ func (ExecuteSqlRequest_QueryMode) EnumDescriptor() ([]byte, []int) { // The request for [CreateSession][google.spanner.v1.Spanner.CreateSession]. type CreateSessionRequest struct { // Required. The database in which the new session is created. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // The session to create. - Session *Session `protobuf:"bytes,2,opt,name=session" json:"session,omitempty"` + Session *Session `protobuf:"bytes,2,opt,name=session,proto3" json:"session,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -113,7 +113,7 @@ func (m *CreateSessionRequest) GetSession() *Session { type Session struct { // The name of the session. This is always system-assigned; values provided // when creating a session are ignored. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The labels for the session. // // * Label keys must be between 1 and 63 characters long and must conform to @@ -123,12 +123,12 @@ type Session struct { // * No more than 64 labels can be associated with a given session. // // See https://goo.gl/xmQnxf for more information on and examples of labels. - Labels map[string]string `protobuf:"bytes,2,rep,name=labels" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + Labels map[string]string `protobuf:"bytes,2,rep,name=labels,proto3" json:"labels,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Output only. The timestamp when the session is created. - CreateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=create_time,json=createTime" json:"create_time,omitempty"` + CreateTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=create_time,json=createTime,proto3" json:"create_time,omitempty"` // Output only. The approximate timestamp when the session is last used. It is // typically earlier than the actual last use time. - ApproximateLastUseTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=approximate_last_use_time,json=approximateLastUseTime" json:"approximate_last_use_time,omitempty"` + ApproximateLastUseTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=approximate_last_use_time,json=approximateLastUseTime,proto3" json:"approximate_last_use_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -189,7 +189,7 @@ func (m *Session) GetApproximateLastUseTime() *timestamp.Timestamp { // The request for [GetSession][google.spanner.v1.Spanner.GetSession]. type GetSessionRequest struct { // Required. The name of the session to retrieve. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -229,14 +229,14 @@ func (m *GetSessionRequest) GetName() string { // The request for [ListSessions][google.spanner.v1.Spanner.ListSessions]. type ListSessionsRequest struct { // Required. The database in which to list sessions. - Database string `protobuf:"bytes,1,opt,name=database" json:"database,omitempty"` + Database string `protobuf:"bytes,1,opt,name=database,proto3" json:"database,omitempty"` // Number of sessions to be returned in the response. If 0 or less, defaults // to the server's maximum allowed page size. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // If non-empty, `page_token` should contain a // [next_page_token][google.spanner.v1.ListSessionsResponse.next_page_token] from a previous // [ListSessionsResponse][google.spanner.v1.ListSessionsResponse]. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // An expression for filtering the results of the request. Filter rules are // case insensitive. The fields eligible for filtering are: // @@ -247,7 +247,7 @@ type ListSessionsRequest struct { // * `labels.env:*` --> The session has the label "env". // * `labels.env:dev` --> The session has the label "env" and the value of // the label contains the string "dev". - Filter string `protobuf:"bytes,4,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,4,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -308,11 +308,11 @@ func (m *ListSessionsRequest) GetFilter() string { // The response for [ListSessions][google.spanner.v1.Spanner.ListSessions]. type ListSessionsResponse struct { // The list of requested sessions. - Sessions []*Session `protobuf:"bytes,1,rep,name=sessions" json:"sessions,omitempty"` + Sessions []*Session `protobuf:"bytes,1,rep,name=sessions,proto3" json:"sessions,omitempty"` // `next_page_token` can be sent in a subsequent // [ListSessions][google.spanner.v1.Spanner.ListSessions] call to fetch more of the matching // sessions. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -359,7 +359,7 @@ func (m *ListSessionsResponse) GetNextPageToken() string { // The request for [DeleteSession][google.spanner.v1.Spanner.DeleteSession]. type DeleteSessionRequest struct { // Required. The name of the session to delete. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -400,12 +400,12 @@ func (m *DeleteSessionRequest) GetName() string { // [ExecuteStreamingSql][google.spanner.v1.Spanner.ExecuteStreamingSql]. type ExecuteSqlRequest struct { // Required. The session in which the SQL query should be performed. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // The transaction to use. If none is provided, the default is a // temporary read-only transaction with strong concurrency. - Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction" json:"transaction,omitempty"` + Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` // Required. The SQL query string. - Sql string `protobuf:"bytes,3,opt,name=sql" json:"sql,omitempty"` + Sql string `protobuf:"bytes,3,opt,name=sql,proto3" json:"sql,omitempty"` // The SQL query string can contain parameter placeholders. A parameter // placeholder consists of `'@'` followed by the parameter // name. Parameter names consist of any combination of letters, @@ -420,7 +420,7 @@ type ExecuteSqlRequest struct { // Parameter values are specified using `params`, which is a JSON // object whose keys are parameter names, and whose values are the // corresponding parameter values. - Params *_struct.Struct `protobuf:"bytes,4,opt,name=params" json:"params,omitempty"` + Params *_struct.Struct `protobuf:"bytes,4,opt,name=params,proto3" json:"params,omitempty"` // It is not always possible for Cloud Spanner to infer the right SQL type // from a JSON value. For example, values of type `BYTES` and values // of type `STRING` both appear in [params][google.spanner.v1.ExecuteSqlRequest.params] as JSON strings. @@ -429,7 +429,7 @@ type ExecuteSqlRequest struct { // SQL type for some or all of the SQL query parameters. See the // definition of [Type][google.spanner.v1.Type] for more information // about SQL types. - ParamTypes map[string]*Type `protobuf:"bytes,5,rep,name=param_types,json=paramTypes" json:"param_types,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ParamTypes map[string]*Type `protobuf:"bytes,5,rep,name=param_types,json=paramTypes,proto3" json:"param_types,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // If this request is resuming a previously interrupted SQL query // execution, `resume_token` should be copied from the last // [PartialResultSet][google.spanner.v1.PartialResultSet] yielded before the interruption. Doing this @@ -440,7 +440,7 @@ type ExecuteSqlRequest struct { // Used to control the amount of debugging information returned in // [ResultSetStats][google.spanner.v1.ResultSetStats]. If [partition_token][google.spanner.v1.ExecuteSqlRequest.partition_token] is set, [query_mode][google.spanner.v1.ExecuteSqlRequest.query_mode] can only // be set to [QueryMode.NORMAL][google.spanner.v1.ExecuteSqlRequest.QueryMode.NORMAL]. - QueryMode ExecuteSqlRequest_QueryMode `protobuf:"varint,7,opt,name=query_mode,json=queryMode,enum=google.spanner.v1.ExecuteSqlRequest_QueryMode" json:"query_mode,omitempty"` + QueryMode ExecuteSqlRequest_QueryMode `protobuf:"varint,7,opt,name=query_mode,json=queryMode,proto3,enum=google.spanner.v1.ExecuteSqlRequest_QueryMode" json:"query_mode,omitempty"` // If present, results will be restricted to the specified partition // previously created using PartitionQuery(). There must be an exact // match for the values of fields common to this message and the @@ -540,7 +540,7 @@ type PartitionOptions struct { // The desired data size for each partition generated. The default for this // option is currently 1 GiB. This is only a hint. The actual size of each // partition may be smaller or larger than this size request. - PartitionSizeBytes int64 `protobuf:"varint,1,opt,name=partition_size_bytes,json=partitionSizeBytes" json:"partition_size_bytes,omitempty"` + PartitionSizeBytes int64 `protobuf:"varint,1,opt,name=partition_size_bytes,json=partitionSizeBytes,proto3" json:"partition_size_bytes,omitempty"` // **Note:** This hint is currently ignored by PartitionQuery and // PartitionRead requests. // @@ -549,7 +549,7 @@ type PartitionOptions struct { // is currently 10,000. The maximum value is currently 200,000. This is only // a hint. The actual number of partitions returned may be smaller or larger // than this maximum count request. - MaxPartitions int64 `protobuf:"varint,2,opt,name=max_partitions,json=maxPartitions" json:"max_partitions,omitempty"` + MaxPartitions int64 `protobuf:"varint,2,opt,name=max_partitions,json=maxPartitions,proto3" json:"max_partitions,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -596,17 +596,17 @@ func (m *PartitionOptions) GetMaxPartitions() int64 { // The request for [PartitionQuery][google.spanner.v1.Spanner.PartitionQuery] type PartitionQueryRequest struct { // Required. The session used to create the partitions. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Read only snapshot transactions are supported, read/write and single use // transactions are not. - Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction" json:"transaction,omitempty"` + Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` // The query request to generate partitions for. The request will fail if // the query is not root partitionable. The query plan of a root // partitionable query has a single distributed union operator. A distributed // union operator conceptually divides one or more tables into multiple // splits, remotely evaluates a subquery independently on each split, and // then unions all results. - Sql string `protobuf:"bytes,3,opt,name=sql" json:"sql,omitempty"` + Sql string `protobuf:"bytes,3,opt,name=sql,proto3" json:"sql,omitempty"` // The SQL query string can contain parameter placeholders. A parameter // placeholder consists of `'@'` followed by the parameter // name. Parameter names consist of any combination of letters, @@ -621,7 +621,7 @@ type PartitionQueryRequest struct { // Parameter values are specified using `params`, which is a JSON // object whose keys are parameter names, and whose values are the // corresponding parameter values. - Params *_struct.Struct `protobuf:"bytes,4,opt,name=params" json:"params,omitempty"` + Params *_struct.Struct `protobuf:"bytes,4,opt,name=params,proto3" json:"params,omitempty"` // It is not always possible for Cloud Spanner to infer the right SQL type // from a JSON value. For example, values of type `BYTES` and values // of type `STRING` both appear in [params][google.spanner.v1.PartitionQueryRequest.params] as JSON strings. @@ -630,9 +630,9 @@ type PartitionQueryRequest struct { // SQL type for some or all of the SQL query parameters. See the // definition of [Type][google.spanner.v1.Type] for more information // about SQL types. - ParamTypes map[string]*Type `protobuf:"bytes,5,rep,name=param_types,json=paramTypes" json:"param_types,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + ParamTypes map[string]*Type `protobuf:"bytes,5,rep,name=param_types,json=paramTypes,proto3" json:"param_types,omitempty" protobuf_key:"bytes,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Additional options that affect how many partitions are created. - PartitionOptions *PartitionOptions `protobuf:"bytes,6,opt,name=partition_options,json=partitionOptions" json:"partition_options,omitempty"` + PartitionOptions *PartitionOptions `protobuf:"bytes,6,opt,name=partition_options,json=partitionOptions,proto3" json:"partition_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -707,19 +707,19 @@ func (m *PartitionQueryRequest) GetPartitionOptions() *PartitionOptions { // The request for [PartitionRead][google.spanner.v1.Spanner.PartitionRead] type PartitionReadRequest struct { // Required. The session used to create the partitions. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Read only snapshot transactions are supported, read/write and single use // transactions are not. - Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction" json:"transaction,omitempty"` + Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` // Required. The name of the table in the database to be read. - Table string `protobuf:"bytes,3,opt,name=table" json:"table,omitempty"` + Table string `protobuf:"bytes,3,opt,name=table,proto3" json:"table,omitempty"` // If non-empty, the name of an index on [table][google.spanner.v1.PartitionReadRequest.table]. This index is // used instead of the table primary key when interpreting [key_set][google.spanner.v1.PartitionReadRequest.key_set] // and sorting result rows. See [key_set][google.spanner.v1.PartitionReadRequest.key_set] for further information. - Index string `protobuf:"bytes,4,opt,name=index" json:"index,omitempty"` + Index string `protobuf:"bytes,4,opt,name=index,proto3" json:"index,omitempty"` // The columns of [table][google.spanner.v1.PartitionReadRequest.table] to be returned for each row matching // this request. - Columns []string `protobuf:"bytes,5,rep,name=columns" json:"columns,omitempty"` + Columns []string `protobuf:"bytes,5,rep,name=columns,proto3" json:"columns,omitempty"` // Required. `key_set` identifies the rows to be yielded. `key_set` names the // primary keys of the rows in [table][google.spanner.v1.PartitionReadRequest.table] to be yielded, unless [index][google.spanner.v1.PartitionReadRequest.index] // is present. If [index][google.spanner.v1.PartitionReadRequest.index] is present, then [key_set][google.spanner.v1.PartitionReadRequest.key_set] instead names @@ -727,9 +727,9 @@ type PartitionReadRequest struct { // // It is not an error for the `key_set` to name rows that do not // exist in the database. Read yields nothing for nonexistent rows. - KeySet *KeySet `protobuf:"bytes,6,opt,name=key_set,json=keySet" json:"key_set,omitempty"` + KeySet *KeySet `protobuf:"bytes,6,opt,name=key_set,json=keySet,proto3" json:"key_set,omitempty"` // Additional options that affect how many partitions are created. - PartitionOptions *PartitionOptions `protobuf:"bytes,9,opt,name=partition_options,json=partitionOptions" json:"partition_options,omitempty"` + PartitionOptions *PartitionOptions `protobuf:"bytes,9,opt,name=partition_options,json=partitionOptions,proto3" json:"partition_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -855,9 +855,9 @@ func (m *Partition) GetPartitionToken() []byte { // or [PartitionRead][google.spanner.v1.Spanner.PartitionRead] type PartitionResponse struct { // Partitions created by this request. - Partitions []*Partition `protobuf:"bytes,1,rep,name=partitions" json:"partitions,omitempty"` + Partitions []*Partition `protobuf:"bytes,1,rep,name=partitions,proto3" json:"partitions,omitempty"` // Transaction created by this request. - Transaction *Transaction `protobuf:"bytes,2,opt,name=transaction" json:"transaction,omitempty"` + Transaction *Transaction `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -905,19 +905,19 @@ func (m *PartitionResponse) GetTransaction() *Transaction { // [StreamingRead][google.spanner.v1.Spanner.StreamingRead]. type ReadRequest struct { // Required. The session in which the read should be performed. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // The transaction to use. If none is provided, the default is a // temporary read-only transaction with strong concurrency. - Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction" json:"transaction,omitempty"` + Transaction *TransactionSelector `protobuf:"bytes,2,opt,name=transaction,proto3" json:"transaction,omitempty"` // Required. The name of the table in the database to be read. - Table string `protobuf:"bytes,3,opt,name=table" json:"table,omitempty"` + Table string `protobuf:"bytes,3,opt,name=table,proto3" json:"table,omitempty"` // If non-empty, the name of an index on [table][google.spanner.v1.ReadRequest.table]. This index is // used instead of the table primary key when interpreting [key_set][google.spanner.v1.ReadRequest.key_set] // and sorting result rows. See [key_set][google.spanner.v1.ReadRequest.key_set] for further information. - Index string `protobuf:"bytes,4,opt,name=index" json:"index,omitempty"` + Index string `protobuf:"bytes,4,opt,name=index,proto3" json:"index,omitempty"` // The columns of [table][google.spanner.v1.ReadRequest.table] to be returned for each row matching // this request. - Columns []string `protobuf:"bytes,5,rep,name=columns" json:"columns,omitempty"` + Columns []string `protobuf:"bytes,5,rep,name=columns,proto3" json:"columns,omitempty"` // Required. `key_set` identifies the rows to be yielded. `key_set` names the // primary keys of the rows in [table][google.spanner.v1.ReadRequest.table] to be yielded, unless [index][google.spanner.v1.ReadRequest.index] // is present. If [index][google.spanner.v1.ReadRequest.index] is present, then [key_set][google.spanner.v1.ReadRequest.key_set] instead names @@ -930,11 +930,11 @@ type ReadRequest struct { // // It is not an error for the `key_set` to name rows that do not // exist in the database. Read yields nothing for nonexistent rows. - KeySet *KeySet `protobuf:"bytes,6,opt,name=key_set,json=keySet" json:"key_set,omitempty"` + KeySet *KeySet `protobuf:"bytes,6,opt,name=key_set,json=keySet,proto3" json:"key_set,omitempty"` // If greater than zero, only the first `limit` rows are yielded. If `limit` // is zero, the default is no limit. A limit cannot be specified if // `partition_token` is set. - Limit int64 `protobuf:"varint,8,opt,name=limit" json:"limit,omitempty"` + Limit int64 `protobuf:"varint,8,opt,name=limit,proto3" json:"limit,omitempty"` // If this request is resuming a previously interrupted read, // `resume_token` should be copied from the last // [PartialResultSet][google.spanner.v1.PartialResultSet] yielded before the interruption. Doing this @@ -1042,9 +1042,9 @@ func (m *ReadRequest) GetPartitionToken() []byte { // The request for [BeginTransaction][google.spanner.v1.Spanner.BeginTransaction]. type BeginTransactionRequest struct { // Required. The session in which the transaction runs. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Required. Options for the new transaction. - Options *TransactionOptions `protobuf:"bytes,2,opt,name=options" json:"options,omitempty"` + Options *TransactionOptions `protobuf:"bytes,2,opt,name=options,proto3" json:"options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1091,7 +1091,7 @@ func (m *BeginTransactionRequest) GetOptions() *TransactionOptions { // The request for [Commit][google.spanner.v1.Spanner.Commit]. type CommitRequest struct { // Required. The session in which the transaction to be committed is running. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Required. The transaction in which to commit. // // Types that are valid to be assigned to Transaction: @@ -1101,7 +1101,7 @@ type CommitRequest struct { // The mutations to be executed when this transaction commits. All // mutations are applied atomically, in the order they appear in // this list. - Mutations []*Mutation `protobuf:"bytes,4,rep,name=mutations" json:"mutations,omitempty"` + Mutations []*Mutation `protobuf:"bytes,4,rep,name=mutations,proto3" json:"mutations,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1139,7 +1139,7 @@ type CommitRequest_TransactionId struct { TransactionId []byte `protobuf:"bytes,2,opt,name=transaction_id,json=transactionId,proto3,oneof"` } type CommitRequest_SingleUseTransaction struct { - SingleUseTransaction *TransactionOptions `protobuf:"bytes,3,opt,name=single_use_transaction,json=singleUseTransaction,oneof"` + SingleUseTransaction *TransactionOptions `protobuf:"bytes,3,opt,name=single_use_transaction,json=singleUseTransaction,proto3,oneof"` } func (*CommitRequest_TransactionId) isCommitRequest_Transaction() {} @@ -1253,7 +1253,7 @@ func _CommitRequest_OneofSizer(msg proto.Message) (n int) { // The response for [Commit][google.spanner.v1.Spanner.Commit]. type CommitResponse struct { // The Cloud Spanner timestamp at which the transaction committed. - CommitTimestamp *timestamp.Timestamp `protobuf:"bytes,1,opt,name=commit_timestamp,json=commitTimestamp" json:"commit_timestamp,omitempty"` + CommitTimestamp *timestamp.Timestamp `protobuf:"bytes,1,opt,name=commit_timestamp,json=commitTimestamp,proto3" json:"commit_timestamp,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1293,7 +1293,7 @@ func (m *CommitResponse) GetCommitTimestamp() *timestamp.Timestamp { // The request for [Rollback][google.spanner.v1.Spanner.Rollback]. type RollbackRequest struct { // Required. The session in which the transaction to roll back is running. - Session string `protobuf:"bytes,1,opt,name=session" json:"session,omitempty"` + Session string `protobuf:"bytes,1,opt,name=session,proto3" json:"session,omitempty"` // Required. The transaction to roll back. TransactionId []byte `protobuf:"bytes,2,opt,name=transaction_id,json=transactionId,proto3" json:"transaction_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -1655,8 +1655,7 @@ func (c *spannerClient) PartitionRead(ctx context.Context, in *PartitionReadRequ return out, nil } -// Server API for Spanner service - +// SpannerServer is the server API for Spanner service. type SpannerServer interface { // Creates a new session. A session can be used to perform // transactions that read and/or modify data in a Cloud Spanner database. diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/transaction.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/transaction.pb.go index 5d0b0bba0..18e92887c 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/transaction.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/transaction.pb.go @@ -279,10 +279,10 @@ type isTransactionOptions_Mode interface { } type TransactionOptions_ReadWrite_ struct { - ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,1,opt,name=read_write,json=readWrite,oneof"` + ReadWrite *TransactionOptions_ReadWrite `protobuf:"bytes,1,opt,name=read_write,json=readWrite,proto3,oneof"` } type TransactionOptions_ReadOnly_ struct { - ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,oneof"` + ReadOnly *TransactionOptions_ReadOnly `protobuf:"bytes,2,opt,name=read_only,json=readOnly,proto3,oneof"` } func (*TransactionOptions_ReadWrite_) isTransactionOptions_Mode() {} @@ -428,7 +428,7 @@ type TransactionOptions_ReadOnly struct { TimestampBound isTransactionOptions_ReadOnly_TimestampBound `protobuf_oneof:"timestamp_bound"` // If true, the Cloud Spanner-selected read timestamp is included in // the [Transaction][google.spanner.v1.Transaction] message that describes the transaction. - ReturnReadTimestamp bool `protobuf:"varint,6,opt,name=return_read_timestamp,json=returnReadTimestamp" json:"return_read_timestamp,omitempty"` + ReturnReadTimestamp bool `protobuf:"varint,6,opt,name=return_read_timestamp,json=returnReadTimestamp,proto3" json:"return_read_timestamp,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -463,19 +463,19 @@ type isTransactionOptions_ReadOnly_TimestampBound interface { } type TransactionOptions_ReadOnly_Strong struct { - Strong bool `protobuf:"varint,1,opt,name=strong,oneof"` + Strong bool `protobuf:"varint,1,opt,name=strong,proto3,oneof"` } type TransactionOptions_ReadOnly_MinReadTimestamp struct { - MinReadTimestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=min_read_timestamp,json=minReadTimestamp,oneof"` + MinReadTimestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=min_read_timestamp,json=minReadTimestamp,proto3,oneof"` } type TransactionOptions_ReadOnly_MaxStaleness struct { - MaxStaleness *duration.Duration `protobuf:"bytes,3,opt,name=max_staleness,json=maxStaleness,oneof"` + MaxStaleness *duration.Duration `protobuf:"bytes,3,opt,name=max_staleness,json=maxStaleness,proto3,oneof"` } type TransactionOptions_ReadOnly_ReadTimestamp struct { - ReadTimestamp *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_timestamp,json=readTimestamp,oneof"` + ReadTimestamp *timestamp.Timestamp `protobuf:"bytes,4,opt,name=read_timestamp,json=readTimestamp,proto3,oneof"` } type TransactionOptions_ReadOnly_ExactStaleness struct { - ExactStaleness *duration.Duration `protobuf:"bytes,5,opt,name=exact_staleness,json=exactStaleness,oneof"` + ExactStaleness *duration.Duration `protobuf:"bytes,5,opt,name=exact_staleness,json=exactStaleness,proto3,oneof"` } func (*TransactionOptions_ReadOnly_Strong) isTransactionOptions_ReadOnly_TimestampBound() {} @@ -680,7 +680,7 @@ type Transaction struct { // // A timestamp in RFC3339 UTC \"Zulu\" format, accurate to nanoseconds. // Example: `"2014-10-02T15:01:23.045123456Z"`. - ReadTimestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=read_timestamp,json=readTimestamp" json:"read_timestamp,omitempty"` + ReadTimestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=read_timestamp,json=readTimestamp,proto3" json:"read_timestamp,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -772,13 +772,13 @@ type isTransactionSelector_Selector interface { } type TransactionSelector_SingleUse struct { - SingleUse *TransactionOptions `protobuf:"bytes,1,opt,name=single_use,json=singleUse,oneof"` + SingleUse *TransactionOptions `protobuf:"bytes,1,opt,name=single_use,json=singleUse,proto3,oneof"` } type TransactionSelector_Id struct { Id []byte `protobuf:"bytes,2,opt,name=id,proto3,oneof"` } type TransactionSelector_Begin struct { - Begin *TransactionOptions `protobuf:"bytes,3,opt,name=begin,oneof"` + Begin *TransactionOptions `protobuf:"bytes,3,opt,name=begin,proto3,oneof"` } func (*TransactionSelector_SingleUse) isTransactionSelector_Selector() {} diff --git a/vendor/google.golang.org/genproto/googleapis/spanner/v1/type.pb.go b/vendor/google.golang.org/genproto/googleapis/spanner/v1/type.pb.go index e97e2767b..688c94174 100644 --- a/vendor/google.golang.org/genproto/googleapis/spanner/v1/type.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/spanner/v1/type.pb.go @@ -98,13 +98,13 @@ func (TypeCode) EnumDescriptor() ([]byte, []int) { // table cell or returned from an SQL query. type Type struct { // Required. The [TypeCode][google.spanner.v1.TypeCode] for this type. - Code TypeCode `protobuf:"varint,1,opt,name=code,enum=google.spanner.v1.TypeCode" json:"code,omitempty"` + Code TypeCode `protobuf:"varint,1,opt,name=code,proto3,enum=google.spanner.v1.TypeCode" json:"code,omitempty"` // If [code][google.spanner.v1.Type.code] == [ARRAY][google.spanner.v1.TypeCode.ARRAY], then `array_element_type` // is the type of the array elements. - ArrayElementType *Type `protobuf:"bytes,2,opt,name=array_element_type,json=arrayElementType" json:"array_element_type,omitempty"` + ArrayElementType *Type `protobuf:"bytes,2,opt,name=array_element_type,json=arrayElementType,proto3" json:"array_element_type,omitempty"` // If [code][google.spanner.v1.Type.code] == [STRUCT][google.spanner.v1.TypeCode.STRUCT], then `struct_type` // provides type information for the struct's fields. - StructType *StructType `protobuf:"bytes,3,opt,name=struct_type,json=structType" json:"struct_type,omitempty"` + StructType *StructType `protobuf:"bytes,3,opt,name=struct_type,json=structType,proto3" json:"struct_type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -163,7 +163,7 @@ type StructType struct { // fields in the [StructType][google.spanner.v1.StructType]. In turn, the order of fields // matches the order of columns in a read request, or the order of // fields in the `SELECT` clause of a query. - Fields []*StructType_Field `protobuf:"bytes,1,rep,name=fields" json:"fields,omitempty"` + Fields []*StructType_Field `protobuf:"bytes,1,rep,name=fields,proto3" json:"fields,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -209,9 +209,9 @@ type StructType_Field struct { // columns might have an empty name (e.g., !"SELECT // UPPER(ColName)"`). Note that a query result can contain // multiple fields with the same name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The type of the field. - Type *Type `protobuf:"bytes,2,opt,name=type" json:"type,omitempty"` + Type *Type `protobuf:"bytes,2,opt,name=type,proto3" json:"type,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer.pb.go b/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer.pb.go index 3d6499fb8..ea86ef7f4 100644 --- a/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer.pb.go @@ -31,7 +31,7 @@ type GetGoogleServiceAccountRequest struct { // The ID of the Google Cloud Platform Console project that the Google service // account is associated with. // Required. - ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -72,7 +72,7 @@ func (m *GetGoogleServiceAccountRequest) GetProjectId() string { type CreateTransferJobRequest struct { // The job to create. // Required. - TransferJob *TransferJob `protobuf:"bytes,1,opt,name=transfer_job,json=transferJob" json:"transfer_job,omitempty"` + TransferJob *TransferJob `protobuf:"bytes,1,opt,name=transfer_job,json=transferJob,proto3" json:"transfer_job,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -113,23 +113,23 @@ func (m *CreateTransferJobRequest) GetTransferJob() *TransferJob { type UpdateTransferJobRequest struct { // The name of job to update. // Required. - JobName string `protobuf:"bytes,1,opt,name=job_name,json=jobName" json:"job_name,omitempty"` + JobName string `protobuf:"bytes,1,opt,name=job_name,json=jobName,proto3" json:"job_name,omitempty"` // The ID of the Google Cloud Platform Console project that owns the job. // Required. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // The job to update. `transferJob` is expected to specify only three fields: // `description`, `transferSpec`, and `status`. An UpdateTransferJobRequest // that specifies other fields will be rejected with an error // `INVALID_ARGUMENT`. // Required. - TransferJob *TransferJob `protobuf:"bytes,3,opt,name=transfer_job,json=transferJob" json:"transfer_job,omitempty"` + TransferJob *TransferJob `protobuf:"bytes,3,opt,name=transfer_job,json=transferJob,proto3" json:"transfer_job,omitempty"` // The field mask of the fields in `transferJob` that are to be updated in // this request. Fields in `transferJob` that can be updated are: // `description`, `transferSpec`, and `status`. To update the `transferSpec` // of the job, a complete transfer specification has to be provided. An // incomplete specification which misses any required fields will be rejected // with the error `INVALID_ARGUMENT`. - UpdateTransferJobFieldMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_transfer_job_field_mask,json=updateTransferJobFieldMask" json:"update_transfer_job_field_mask,omitempty"` + UpdateTransferJobFieldMask *field_mask.FieldMask `protobuf:"bytes,4,opt,name=update_transfer_job_field_mask,json=updateTransferJobFieldMask,proto3" json:"update_transfer_job_field_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -191,10 +191,10 @@ func (m *UpdateTransferJobRequest) GetUpdateTransferJobFieldMask() *field_mask.F type GetTransferJobRequest struct { // The job to get. // Required. - JobName string `protobuf:"bytes,1,opt,name=job_name,json=jobName" json:"job_name,omitempty"` + JobName string `protobuf:"bytes,1,opt,name=job_name,json=jobName,proto3" json:"job_name,omitempty"` // The ID of the Google Cloud Platform Console project that owns the job. // Required. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -249,11 +249,11 @@ type ListTransferJobsRequest struct { // must be specified with array notation. `project_id` is required. `job_names` // and `job_statuses` are optional. The valid values for `job_statuses` are // case-insensitive: `ENABLED`, `DISABLED`, and `DELETED`. - Filter string `protobuf:"bytes,1,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,1,opt,name=filter,proto3" json:"filter,omitempty"` // The list page size. The max allowed value is 256. - PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,4,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The list page token. - PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,5,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -307,9 +307,9 @@ func (m *ListTransferJobsRequest) GetPageToken() string { // Response from ListTransferJobs. type ListTransferJobsResponse struct { // A list of transfer jobs. - TransferJobs []*TransferJob `protobuf:"bytes,1,rep,name=transfer_jobs,json=transferJobs" json:"transfer_jobs,omitempty"` + TransferJobs []*TransferJob `protobuf:"bytes,1,rep,name=transfer_jobs,json=transferJobs,proto3" json:"transfer_jobs,omitempty"` // The list next page token. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -357,7 +357,7 @@ func (m *ListTransferJobsResponse) GetNextPageToken() string { type PauseTransferOperationRequest struct { // The name of the transfer operation. // Required. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -398,7 +398,7 @@ func (m *PauseTransferOperationRequest) GetName() string { type ResumeTransferOperationRequest struct { // The name of the transfer operation. // Required. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -554,8 +554,7 @@ func (c *storageTransferServiceClient) ResumeTransferOperation(ctx context.Conte return out, nil } -// Server API for StorageTransferService service - +// StorageTransferServiceServer is the server API for StorageTransferService service. type StorageTransferServiceServer interface { // Returns the Google service account that is used by Storage Transfer // Service to access buckets in the project where transfers diff --git a/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer_types.pb.go b/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer_types.pb.go index 2bdf861d4..3c1d3aa28 100644 --- a/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer_types.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/storagetransfer/v1/transfer_types.pb.go @@ -105,7 +105,7 @@ func (TransferOperation_Status) EnumDescriptor() ([]byte, []int) { // Google service account type GoogleServiceAccount struct { // Required. - AccountEmail string `protobuf:"bytes,1,opt,name=account_email,json=accountEmail" json:"account_email,omitempty"` + AccountEmail string `protobuf:"bytes,1,opt,name=account_email,json=accountEmail,proto3" json:"account_email,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -147,10 +147,10 @@ func (m *GoogleServiceAccount) GetAccountEmail() string { type AwsAccessKey struct { // AWS access key ID. // Required. - AccessKeyId string `protobuf:"bytes,1,opt,name=access_key_id,json=accessKeyId" json:"access_key_id,omitempty"` + AccessKeyId string `protobuf:"bytes,1,opt,name=access_key_id,json=accessKeyId,proto3" json:"access_key_id,omitempty"` // AWS secret access key. This field is not returned in RPC responses. // Required. - SecretAccessKey string `protobuf:"bytes,2,opt,name=secret_access_key,json=secretAccessKey" json:"secret_access_key,omitempty"` + SecretAccessKey string `protobuf:"bytes,2,opt,name=secret_access_key,json=secretAccessKey,proto3" json:"secret_access_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -203,10 +203,10 @@ type ObjectConditions struct { // `NOW` - `maxTimeElapsedSinceLastModification` and less than // `NOW` - `minTimeElapsedSinceLastModification`, or not have a // `lastModificationTime`. - MinTimeElapsedSinceLastModification *duration.Duration `protobuf:"bytes,1,opt,name=min_time_elapsed_since_last_modification,json=minTimeElapsedSinceLastModification" json:"min_time_elapsed_since_last_modification,omitempty"` + MinTimeElapsedSinceLastModification *duration.Duration `protobuf:"bytes,1,opt,name=min_time_elapsed_since_last_modification,json=minTimeElapsedSinceLastModification,proto3" json:"min_time_elapsed_since_last_modification,omitempty"` // `maxTimeElapsedSinceLastModification` is the complement to // `minTimeElapsedSinceLastModification`. - MaxTimeElapsedSinceLastModification *duration.Duration `protobuf:"bytes,2,opt,name=max_time_elapsed_since_last_modification,json=maxTimeElapsedSinceLastModification" json:"max_time_elapsed_since_last_modification,omitempty"` + MaxTimeElapsedSinceLastModification *duration.Duration `protobuf:"bytes,2,opt,name=max_time_elapsed_since_last_modification,json=maxTimeElapsedSinceLastModification,proto3" json:"max_time_elapsed_since_last_modification,omitempty"` // If `includePrefixes` is specified, objects that satisfy the object // conditions must have names that start with one of the `includePrefixes` // and that do not start with any of the `excludePrefixes`. If `includePrefixes` @@ -240,12 +240,12 @@ type ObjectConditions struct { // with the value of a path explicitly included by `includePrefixes`. // // The max size of `includePrefixes` is 1000. - IncludePrefixes []string `protobuf:"bytes,3,rep,name=include_prefixes,json=includePrefixes" json:"include_prefixes,omitempty"` + IncludePrefixes []string `protobuf:"bytes,3,rep,name=include_prefixes,json=includePrefixes,proto3" json:"include_prefixes,omitempty"` // `excludePrefixes` must follow the requirements described for // `includePrefixes`. // // The max size of `excludePrefixes` is 1000. - ExcludePrefixes []string `protobuf:"bytes,4,rep,name=exclude_prefixes,json=excludePrefixes" json:"exclude_prefixes,omitempty"` + ExcludePrefixes []string `protobuf:"bytes,4,rep,name=exclude_prefixes,json=excludePrefixes,proto3" json:"exclude_prefixes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -310,7 +310,7 @@ type GcsData struct { // Google Cloud Storage bucket name (see // [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). // Required. - BucketName string `protobuf:"bytes,1,opt,name=bucket_name,json=bucketName" json:"bucket_name,omitempty"` + BucketName string `protobuf:"bytes,1,opt,name=bucket_name,json=bucketName,proto3" json:"bucket_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -353,12 +353,12 @@ type AwsS3Data struct { // S3 Bucket name (see // [Creating a bucket](http://docs.aws.amazon.com/AmazonS3/latest/dev/create-bucket-get-location-example.html)). // Required. - BucketName string `protobuf:"bytes,1,opt,name=bucket_name,json=bucketName" json:"bucket_name,omitempty"` + BucketName string `protobuf:"bytes,1,opt,name=bucket_name,json=bucketName,proto3" json:"bucket_name,omitempty"` // AWS access key used to sign the API requests to the AWS S3 bucket. // Permissions on the bucket must be granted to the access ID of the // AWS access key. // Required. - AwsAccessKey *AwsAccessKey `protobuf:"bytes,2,opt,name=aws_access_key,json=awsAccessKey" json:"aws_access_key,omitempty"` + AwsAccessKey *AwsAccessKey `protobuf:"bytes,2,opt,name=aws_access_key,json=awsAccessKey,proto3" json:"aws_access_key,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -448,7 +448,7 @@ type HttpData struct { // This file must allow public access. Currently, only URLs with HTTP and // HTTPS schemes are supported. // Required. - ListUrl string `protobuf:"bytes,1,opt,name=list_url,json=listUrl" json:"list_url,omitempty"` + ListUrl string `protobuf:"bytes,1,opt,name=list_url,json=listUrl,proto3" json:"list_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -489,15 +489,15 @@ func (m *HttpData) GetListUrl() string { // to be performed on objects in a transfer. type TransferOptions struct { // Whether overwriting objects that already exist in the sink is allowed. - OverwriteObjectsAlreadyExistingInSink bool `protobuf:"varint,1,opt,name=overwrite_objects_already_existing_in_sink,json=overwriteObjectsAlreadyExistingInSink" json:"overwrite_objects_already_existing_in_sink,omitempty"` + OverwriteObjectsAlreadyExistingInSink bool `protobuf:"varint,1,opt,name=overwrite_objects_already_existing_in_sink,json=overwriteObjectsAlreadyExistingInSink,proto3" json:"overwrite_objects_already_existing_in_sink,omitempty"` // Whether objects that exist only in the sink should be deleted. Note that // this option and `deleteObjectsFromSourceAfterTransfer` are mutually // exclusive. - DeleteObjectsUniqueInSink bool `protobuf:"varint,2,opt,name=delete_objects_unique_in_sink,json=deleteObjectsUniqueInSink" json:"delete_objects_unique_in_sink,omitempty"` + DeleteObjectsUniqueInSink bool `protobuf:"varint,2,opt,name=delete_objects_unique_in_sink,json=deleteObjectsUniqueInSink,proto3" json:"delete_objects_unique_in_sink,omitempty"` // Whether objects should be deleted from the source after they are // transferred to the sink. Note that this option and // `deleteObjectsUniqueInSink` are mutually exclusive. - DeleteObjectsFromSourceAfterTransfer bool `protobuf:"varint,3,opt,name=delete_objects_from_source_after_transfer,json=deleteObjectsFromSourceAfterTransfer" json:"delete_objects_from_source_after_transfer,omitempty"` + DeleteObjectsFromSourceAfterTransfer bool `protobuf:"varint,3,opt,name=delete_objects_from_source_after_transfer,json=deleteObjectsFromSourceAfterTransfer,proto3" json:"delete_objects_from_source_after_transfer,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -565,11 +565,11 @@ type TransferSpec struct { // Only objects that satisfy these object conditions are included in the set // of data source and data sink objects. Object conditions based on // objects' `lastModificationTime` do not exclude objects in a data sink. - ObjectConditions *ObjectConditions `protobuf:"bytes,5,opt,name=object_conditions,json=objectConditions" json:"object_conditions,omitempty"` + ObjectConditions *ObjectConditions `protobuf:"bytes,5,opt,name=object_conditions,json=objectConditions,proto3" json:"object_conditions,omitempty"` // If the option `deleteObjectsUniqueInSink` is `true`, object conditions // based on objects' `lastModificationTime` are ignored and do not exclude // objects in a data source or a data sink. - TransferOptions *TransferOptions `protobuf:"bytes,6,opt,name=transfer_options,json=transferOptions" json:"transfer_options,omitempty"` + TransferOptions *TransferOptions `protobuf:"bytes,6,opt,name=transfer_options,json=transferOptions,proto3" json:"transfer_options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -607,16 +607,16 @@ type isTransferSpec_DataSink interface { } type TransferSpec_GcsDataSource struct { - GcsDataSource *GcsData `protobuf:"bytes,1,opt,name=gcs_data_source,json=gcsDataSource,oneof"` + GcsDataSource *GcsData `protobuf:"bytes,1,opt,name=gcs_data_source,json=gcsDataSource,proto3,oneof"` } type TransferSpec_AwsS3DataSource struct { - AwsS3DataSource *AwsS3Data `protobuf:"bytes,2,opt,name=aws_s3_data_source,json=awsS3DataSource,oneof"` + AwsS3DataSource *AwsS3Data `protobuf:"bytes,2,opt,name=aws_s3_data_source,json=awsS3DataSource,proto3,oneof"` } type TransferSpec_HttpDataSource struct { - HttpDataSource *HttpData `protobuf:"bytes,3,opt,name=http_data_source,json=httpDataSource,oneof"` + HttpDataSource *HttpData `protobuf:"bytes,3,opt,name=http_data_source,json=httpDataSource,proto3,oneof"` } type TransferSpec_GcsDataSink struct { - GcsDataSink *GcsData `protobuf:"bytes,4,opt,name=gcs_data_sink,json=gcsDataSink,oneof"` + GcsDataSink *GcsData `protobuf:"bytes,4,opt,name=gcs_data_sink,json=gcsDataSink,proto3,oneof"` } func (*TransferSpec_GcsDataSource) isTransferSpec_DataSource() {} @@ -809,11 +809,11 @@ type Schedule struct { // `scheduleStartDate` is in the past, the transfer will run for the first // time on the following day. // Required. - ScheduleStartDate *date.Date `protobuf:"bytes,1,opt,name=schedule_start_date,json=scheduleStartDate" json:"schedule_start_date,omitempty"` + ScheduleStartDate *date.Date `protobuf:"bytes,1,opt,name=schedule_start_date,json=scheduleStartDate,proto3" json:"schedule_start_date,omitempty"` // The last day the recurring transfer will be run. If `scheduleEndDate` // is the same as `scheduleStartDate`, the transfer will be executed only // once. - ScheduleEndDate *date.Date `protobuf:"bytes,2,opt,name=schedule_end_date,json=scheduleEndDate" json:"schedule_end_date,omitempty"` + ScheduleEndDate *date.Date `protobuf:"bytes,2,opt,name=schedule_end_date,json=scheduleEndDate,proto3" json:"schedule_end_date,omitempty"` // The time in UTC at which the transfer will be scheduled to start in a day. // Transfers may start later than this time. If not specified, recurring and // one-time transfers that are scheduled to run today will run immediately; @@ -821,7 +821,7 @@ type Schedule struct { // at approximately midnight UTC on that date. Note that when configuring a // transfer with the Cloud Platform Console, the transfer's start time in a // day is specified in your local timezone. - StartTimeOfDay *timeofday.TimeOfDay `protobuf:"bytes,3,opt,name=start_time_of_day,json=startTimeOfDay" json:"start_time_of_day,omitempty"` + StartTimeOfDay *timeofday.TimeOfDay `protobuf:"bytes,3,opt,name=start_time_of_day,json=startTimeOfDay,proto3" json:"start_time_of_day,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -879,16 +879,16 @@ type TransferJob struct { // job is created. This field should be left empty in requests to create a new // transfer job; otherwise, the requests result in an `INVALID_ARGUMENT` // error. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A description provided by the user for the job. Its max length is 1024 // bytes when Unicode-encoded. - Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"` + Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // The ID of the Google Cloud Platform Console project that owns the job. - ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,3,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Transfer specification. - TransferSpec *TransferSpec `protobuf:"bytes,4,opt,name=transfer_spec,json=transferSpec" json:"transfer_spec,omitempty"` + TransferSpec *TransferSpec `protobuf:"bytes,4,opt,name=transfer_spec,json=transferSpec,proto3" json:"transfer_spec,omitempty"` // Schedule specification. - Schedule *Schedule `protobuf:"bytes,5,opt,name=schedule" json:"schedule,omitempty"` + Schedule *Schedule `protobuf:"bytes,5,opt,name=schedule,proto3" json:"schedule,omitempty"` // Status of the job. This value MUST be specified for // `CreateTransferJobRequests`. // @@ -896,13 +896,13 @@ type TransferJob struct { // run. For example, if you change the job status from `ENABLED` to // `DISABLED`, and an operation spawned by the transfer is running, the status // change would not affect the current operation. - Status TransferJob_Status `protobuf:"varint,6,opt,name=status,enum=google.storagetransfer.v1.TransferJob_Status" json:"status,omitempty"` + Status TransferJob_Status `protobuf:"varint,6,opt,name=status,proto3,enum=google.storagetransfer.v1.TransferJob_Status" json:"status,omitempty"` // This field cannot be changed by user requests. - CreationTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=creation_time,json=creationTime" json:"creation_time,omitempty"` + CreationTime *timestamp.Timestamp `protobuf:"bytes,7,opt,name=creation_time,json=creationTime,proto3" json:"creation_time,omitempty"` // This field cannot be changed by user requests. - LastModificationTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=last_modification_time,json=lastModificationTime" json:"last_modification_time,omitempty"` + LastModificationTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=last_modification_time,json=lastModificationTime,proto3" json:"last_modification_time,omitempty"` // This field cannot be changed by user requests. - DeletionTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=deletion_time,json=deletionTime" json:"deletion_time,omitempty"` + DeletionTime *timestamp.Timestamp `protobuf:"bytes,9,opt,name=deletion_time,json=deletionTime,proto3" json:"deletion_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1000,9 +1000,9 @@ type ErrorLogEntry struct { // A URL that refers to the target (a data source, a data sink, // or an object) with which the error is associated. // Required. - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` // A list of messages that carry the error details. - ErrorDetails []string `protobuf:"bytes,3,rep,name=error_details,json=errorDetails" json:"error_details,omitempty"` + ErrorDetails []string `protobuf:"bytes,3,rep,name=error_details,json=errorDetails,proto3" json:"error_details,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1050,12 +1050,12 @@ func (m *ErrorLogEntry) GetErrorDetails() []string { // entries. type ErrorSummary struct { // Required. - ErrorCode code.Code `protobuf:"varint,1,opt,name=error_code,json=errorCode,enum=google.rpc.Code" json:"error_code,omitempty"` + ErrorCode code.Code `protobuf:"varint,1,opt,name=error_code,json=errorCode,proto3,enum=google.rpc.Code" json:"error_code,omitempty"` // Count of this type of error. // Required. - ErrorCount int64 `protobuf:"varint,2,opt,name=error_count,json=errorCount" json:"error_count,omitempty"` + ErrorCount int64 `protobuf:"varint,2,opt,name=error_count,json=errorCount,proto3" json:"error_count,omitempty"` // Error samples. - ErrorLogEntries []*ErrorLogEntry `protobuf:"bytes,3,rep,name=error_log_entries,json=errorLogEntries" json:"error_log_entries,omitempty"` + ErrorLogEntries []*ErrorLogEntry `protobuf:"bytes,3,rep,name=error_log_entries,json=errorLogEntries,proto3" json:"error_log_entries,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1111,41 +1111,41 @@ type TransferCounters struct { // Objects found in the data source that are scheduled to be transferred, // excluding any that are filtered based on object conditions or skipped due // to sync. - ObjectsFoundFromSource int64 `protobuf:"varint,1,opt,name=objects_found_from_source,json=objectsFoundFromSource" json:"objects_found_from_source,omitempty"` + ObjectsFoundFromSource int64 `protobuf:"varint,1,opt,name=objects_found_from_source,json=objectsFoundFromSource,proto3" json:"objects_found_from_source,omitempty"` // Bytes found in the data source that are scheduled to be transferred, // excluding any that are filtered based on object conditions or skipped due // to sync. - BytesFoundFromSource int64 `protobuf:"varint,2,opt,name=bytes_found_from_source,json=bytesFoundFromSource" json:"bytes_found_from_source,omitempty"` + BytesFoundFromSource int64 `protobuf:"varint,2,opt,name=bytes_found_from_source,json=bytesFoundFromSource,proto3" json:"bytes_found_from_source,omitempty"` // Objects found only in the data sink that are scheduled to be deleted. - ObjectsFoundOnlyFromSink int64 `protobuf:"varint,3,opt,name=objects_found_only_from_sink,json=objectsFoundOnlyFromSink" json:"objects_found_only_from_sink,omitempty"` + ObjectsFoundOnlyFromSink int64 `protobuf:"varint,3,opt,name=objects_found_only_from_sink,json=objectsFoundOnlyFromSink,proto3" json:"objects_found_only_from_sink,omitempty"` // Bytes found only in the data sink that are scheduled to be deleted. - BytesFoundOnlyFromSink int64 `protobuf:"varint,4,opt,name=bytes_found_only_from_sink,json=bytesFoundOnlyFromSink" json:"bytes_found_only_from_sink,omitempty"` + BytesFoundOnlyFromSink int64 `protobuf:"varint,4,opt,name=bytes_found_only_from_sink,json=bytesFoundOnlyFromSink,proto3" json:"bytes_found_only_from_sink,omitempty"` // Objects in the data source that are not transferred because they already // exist in the data sink. - ObjectsFromSourceSkippedBySync int64 `protobuf:"varint,5,opt,name=objects_from_source_skipped_by_sync,json=objectsFromSourceSkippedBySync" json:"objects_from_source_skipped_by_sync,omitempty"` + ObjectsFromSourceSkippedBySync int64 `protobuf:"varint,5,opt,name=objects_from_source_skipped_by_sync,json=objectsFromSourceSkippedBySync,proto3" json:"objects_from_source_skipped_by_sync,omitempty"` // Bytes in the data source that are not transferred because they already // exist in the data sink. - BytesFromSourceSkippedBySync int64 `protobuf:"varint,6,opt,name=bytes_from_source_skipped_by_sync,json=bytesFromSourceSkippedBySync" json:"bytes_from_source_skipped_by_sync,omitempty"` + BytesFromSourceSkippedBySync int64 `protobuf:"varint,6,opt,name=bytes_from_source_skipped_by_sync,json=bytesFromSourceSkippedBySync,proto3" json:"bytes_from_source_skipped_by_sync,omitempty"` // Objects that are copied to the data sink. - ObjectsCopiedToSink int64 `protobuf:"varint,7,opt,name=objects_copied_to_sink,json=objectsCopiedToSink" json:"objects_copied_to_sink,omitempty"` + ObjectsCopiedToSink int64 `protobuf:"varint,7,opt,name=objects_copied_to_sink,json=objectsCopiedToSink,proto3" json:"objects_copied_to_sink,omitempty"` // Bytes that are copied to the data sink. - BytesCopiedToSink int64 `protobuf:"varint,8,opt,name=bytes_copied_to_sink,json=bytesCopiedToSink" json:"bytes_copied_to_sink,omitempty"` + BytesCopiedToSink int64 `protobuf:"varint,8,opt,name=bytes_copied_to_sink,json=bytesCopiedToSink,proto3" json:"bytes_copied_to_sink,omitempty"` // Objects that are deleted from the data source. - ObjectsDeletedFromSource int64 `protobuf:"varint,9,opt,name=objects_deleted_from_source,json=objectsDeletedFromSource" json:"objects_deleted_from_source,omitempty"` + ObjectsDeletedFromSource int64 `protobuf:"varint,9,opt,name=objects_deleted_from_source,json=objectsDeletedFromSource,proto3" json:"objects_deleted_from_source,omitempty"` // Bytes that are deleted from the data source. - BytesDeletedFromSource int64 `protobuf:"varint,10,opt,name=bytes_deleted_from_source,json=bytesDeletedFromSource" json:"bytes_deleted_from_source,omitempty"` + BytesDeletedFromSource int64 `protobuf:"varint,10,opt,name=bytes_deleted_from_source,json=bytesDeletedFromSource,proto3" json:"bytes_deleted_from_source,omitempty"` // Objects that are deleted from the data sink. - ObjectsDeletedFromSink int64 `protobuf:"varint,11,opt,name=objects_deleted_from_sink,json=objectsDeletedFromSink" json:"objects_deleted_from_sink,omitempty"` + ObjectsDeletedFromSink int64 `protobuf:"varint,11,opt,name=objects_deleted_from_sink,json=objectsDeletedFromSink,proto3" json:"objects_deleted_from_sink,omitempty"` // Bytes that are deleted from the data sink. - BytesDeletedFromSink int64 `protobuf:"varint,12,opt,name=bytes_deleted_from_sink,json=bytesDeletedFromSink" json:"bytes_deleted_from_sink,omitempty"` + BytesDeletedFromSink int64 `protobuf:"varint,12,opt,name=bytes_deleted_from_sink,json=bytesDeletedFromSink,proto3" json:"bytes_deleted_from_sink,omitempty"` // Objects in the data source that failed during the transfer. - ObjectsFromSourceFailed int64 `protobuf:"varint,13,opt,name=objects_from_source_failed,json=objectsFromSourceFailed" json:"objects_from_source_failed,omitempty"` + ObjectsFromSourceFailed int64 `protobuf:"varint,13,opt,name=objects_from_source_failed,json=objectsFromSourceFailed,proto3" json:"objects_from_source_failed,omitempty"` // Bytes in the data source that failed during the transfer. - BytesFromSourceFailed int64 `protobuf:"varint,14,opt,name=bytes_from_source_failed,json=bytesFromSourceFailed" json:"bytes_from_source_failed,omitempty"` + BytesFromSourceFailed int64 `protobuf:"varint,14,opt,name=bytes_from_source_failed,json=bytesFromSourceFailed,proto3" json:"bytes_from_source_failed,omitempty"` // Objects that failed to be deleted from the data sink. - ObjectsFailedToDeleteFromSink int64 `protobuf:"varint,15,opt,name=objects_failed_to_delete_from_sink,json=objectsFailedToDeleteFromSink" json:"objects_failed_to_delete_from_sink,omitempty"` + ObjectsFailedToDeleteFromSink int64 `protobuf:"varint,15,opt,name=objects_failed_to_delete_from_sink,json=objectsFailedToDeleteFromSink,proto3" json:"objects_failed_to_delete_from_sink,omitempty"` // Bytes that failed to be deleted from the data sink. - BytesFailedToDeleteFromSink int64 `protobuf:"varint,16,opt,name=bytes_failed_to_delete_from_sink,json=bytesFailedToDeleteFromSink" json:"bytes_failed_to_delete_from_sink,omitempty"` + BytesFailedToDeleteFromSink int64 `protobuf:"varint,16,opt,name=bytes_failed_to_delete_from_sink,json=bytesFailedToDeleteFromSink,proto3" json:"bytes_failed_to_delete_from_sink,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1290,25 +1290,25 @@ func (m *TransferCounters) GetBytesFailedToDeleteFromSink() int64 { // A description of the execution of a transfer. type TransferOperation struct { // A globally unique ID assigned by the system. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The ID of the Google Cloud Platform Console project that owns the operation. // Required. - ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId" json:"project_id,omitempty"` + ProjectId string `protobuf:"bytes,2,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"` // Transfer specification. // Required. - TransferSpec *TransferSpec `protobuf:"bytes,3,opt,name=transfer_spec,json=transferSpec" json:"transfer_spec,omitempty"` + TransferSpec *TransferSpec `protobuf:"bytes,3,opt,name=transfer_spec,json=transferSpec,proto3" json:"transfer_spec,omitempty"` // Start time of this transfer execution. - StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"` + StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // End time of this transfer execution. - EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"` + EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"` // Status of the transfer operation. - Status TransferOperation_Status `protobuf:"varint,6,opt,name=status,enum=google.storagetransfer.v1.TransferOperation_Status" json:"status,omitempty"` + Status TransferOperation_Status `protobuf:"varint,6,opt,name=status,proto3,enum=google.storagetransfer.v1.TransferOperation_Status" json:"status,omitempty"` // Information about the progress of the transfer operation. - Counters *TransferCounters `protobuf:"bytes,7,opt,name=counters" json:"counters,omitempty"` + Counters *TransferCounters `protobuf:"bytes,7,opt,name=counters,proto3" json:"counters,omitempty"` // Summarizes errors encountered with sample error log entries. - ErrorBreakdowns []*ErrorSummary `protobuf:"bytes,8,rep,name=error_breakdowns,json=errorBreakdowns" json:"error_breakdowns,omitempty"` + ErrorBreakdowns []*ErrorSummary `protobuf:"bytes,8,rep,name=error_breakdowns,json=errorBreakdowns,proto3" json:"error_breakdowns,omitempty"` // The name of the transfer job that triggers this transfer operation. - TransferJobName string `protobuf:"bytes,9,opt,name=transfer_job_name,json=transferJobName" json:"transfer_job_name,omitempty"` + TransferJobName string `protobuf:"bytes,9,opt,name=transfer_job_name,json=transferJobName,proto3" json:"transfer_job_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/resources.pb.go b/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/resources.pb.go index 317d9c3bc..78c8f9f0b 100644 --- a/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/resources.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/resources.pb.go @@ -26,7 +26,7 @@ type UploadRef struct { // Required. An upload reference should be unique for each user. It follows // the form: // "https://streetviewpublish.googleapis.com/media/user//photo/" - UploadUrl string `protobuf:"bytes,1,opt,name=upload_url,json=uploadUrl" json:"upload_url,omitempty"` + UploadUrl string `protobuf:"bytes,1,opt,name=upload_url,json=uploadUrl,proto3" json:"upload_url,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -66,7 +66,7 @@ func (m *UploadRef) GetUploadUrl() string { // Identifier for a photo. type PhotoId struct { // Required. A base64 encoded identifier. - Id string `protobuf:"bytes,1,opt,name=id" json:"id,omitempty"` + Id string `protobuf:"bytes,1,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -108,11 +108,11 @@ type Level struct { // Floor number, used for ordering. 0 indicates the ground level, 1 indicates // the first level above ground level, -1 indicates the first level under // ground level. Non-integer values are OK. - Number float64 `protobuf:"fixed64,1,opt,name=number" json:"number,omitempty"` + Number float64 `protobuf:"fixed64,1,opt,name=number,proto3" json:"number,omitempty"` // Required. A name assigned to this Level, restricted to 3 characters. // Consider how the elevator buttons would be labeled for this level if there // was an elevator. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -164,25 +164,25 @@ type Pose struct { // here, the geolocation from the exif header will be used. // If the latitude and longitude pair is not provided and cannot be found in // the exif header, the create photo process will fail. - LatLngPair *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng_pair,json=latLngPair" json:"lat_lng_pair,omitempty"` + LatLngPair *latlng.LatLng `protobuf:"bytes,1,opt,name=lat_lng_pair,json=latLngPair,proto3" json:"lat_lng_pair,omitempty"` // Altitude of the pose in meters above ground level (as defined by WGS84). // NaN indicates an unmeasured quantity. - Altitude float64 `protobuf:"fixed64,2,opt,name=altitude" json:"altitude,omitempty"` + Altitude float64 `protobuf:"fixed64,2,opt,name=altitude,proto3" json:"altitude,omitempty"` // Compass heading, measured at the center of the photo in degrees clockwise // from North. Value must be >=0 and <360. // NaN indicates an unmeasured quantity. - Heading float64 `protobuf:"fixed64,3,opt,name=heading" json:"heading,omitempty"` + Heading float64 `protobuf:"fixed64,3,opt,name=heading,proto3" json:"heading,omitempty"` // Pitch, measured at the center of the photo in degrees. Value must be >=-90 // and <= 90. A value of -90 means looking directly down, and a value of 90 // means looking directly up. // NaN indicates an unmeasured quantity. - Pitch float64 `protobuf:"fixed64,4,opt,name=pitch" json:"pitch,omitempty"` + Pitch float64 `protobuf:"fixed64,4,opt,name=pitch,proto3" json:"pitch,omitempty"` // Roll, measured in degrees. Value must be >= 0 and <360. A value of 0 // means level with the horizon. // NaN indicates an unmeasured quantity. - Roll float64 `protobuf:"fixed64,5,opt,name=roll" json:"roll,omitempty"` + Roll float64 `protobuf:"fixed64,5,opt,name=roll,proto3" json:"roll,omitempty"` // Level (the floor in a building) used to configure vertical navigation. - Level *Level `protobuf:"bytes,7,opt,name=level" json:"level,omitempty"` + Level *Level `protobuf:"bytes,7,opt,name=level,proto3" json:"level,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -258,7 +258,7 @@ func (m *Pose) GetLevel() *Level { type Place struct { // Required. Place identifier, as described in // https://developers.google.com/places/place-id. - PlaceId string `protobuf:"bytes,1,opt,name=place_id,json=placeId" json:"place_id,omitempty"` + PlaceId string `protobuf:"bytes,1,opt,name=place_id,json=placeId,proto3" json:"place_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -299,7 +299,7 @@ func (m *Place) GetPlaceId() string { type Connection struct { // Required. The destination of the connection from the containing photo to // another photo. - Target *PhotoId `protobuf:"bytes,1,opt,name=target" json:"target,omitempty"` + Target *PhotoId `protobuf:"bytes,1,opt,name=target,proto3" json:"target,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -340,31 +340,31 @@ func (m *Connection) GetTarget() *PhotoId { type Photo struct { // Output only. Identifier for the photo, which is unique among all photos in // Google. - PhotoId *PhotoId `protobuf:"bytes,1,opt,name=photo_id,json=photoId" json:"photo_id,omitempty"` + PhotoId *PhotoId `protobuf:"bytes,1,opt,name=photo_id,json=photoId,proto3" json:"photo_id,omitempty"` // Required (when creating photo). Input only. The resource URL where the // photo bytes are uploaded to. - UploadReference *UploadRef `protobuf:"bytes,2,opt,name=upload_reference,json=uploadReference" json:"upload_reference,omitempty"` + UploadReference *UploadRef `protobuf:"bytes,2,opt,name=upload_reference,json=uploadReference,proto3" json:"upload_reference,omitempty"` // Output only. The download URL for the photo bytes. This field is set only // when the `view` parameter in a `GetPhotoRequest` is set to // `INCLUDE_DOWNLOAD_URL`. - DownloadUrl string `protobuf:"bytes,3,opt,name=download_url,json=downloadUrl" json:"download_url,omitempty"` + DownloadUrl string `protobuf:"bytes,3,opt,name=download_url,json=downloadUrl,proto3" json:"download_url,omitempty"` // Output only. The thumbnail URL for showing a preview of the given photo. - ThumbnailUrl string `protobuf:"bytes,9,opt,name=thumbnail_url,json=thumbnailUrl" json:"thumbnail_url,omitempty"` + ThumbnailUrl string `protobuf:"bytes,9,opt,name=thumbnail_url,json=thumbnailUrl,proto3" json:"thumbnail_url,omitempty"` // Output only. The share link for the photo. - ShareLink string `protobuf:"bytes,11,opt,name=share_link,json=shareLink" json:"share_link,omitempty"` + ShareLink string `protobuf:"bytes,11,opt,name=share_link,json=shareLink,proto3" json:"share_link,omitempty"` // Pose of the photo. - Pose *Pose `protobuf:"bytes,4,opt,name=pose" json:"pose,omitempty"` + Pose *Pose `protobuf:"bytes,4,opt,name=pose,proto3" json:"pose,omitempty"` // Connections to other photos. A connection represents the link from this // photo to another photo. - Connections []*Connection `protobuf:"bytes,5,rep,name=connections" json:"connections,omitempty"` + Connections []*Connection `protobuf:"bytes,5,rep,name=connections,proto3" json:"connections,omitempty"` // Absolute time when the photo was captured. // When the photo has no exif timestamp, this is used to set a timestamp in // the photo metadata. - CaptureTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=capture_time,json=captureTime" json:"capture_time,omitempty"` + CaptureTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=capture_time,json=captureTime,proto3" json:"capture_time,omitempty"` // Places where this photo belongs. - Places []*Place `protobuf:"bytes,7,rep,name=places" json:"places,omitempty"` + Places []*Place `protobuf:"bytes,7,rep,name=places,proto3" json:"places,omitempty"` // Output only. View count of the photo. - ViewCount int64 `protobuf:"varint,10,opt,name=view_count,json=viewCount" json:"view_count,omitempty"` + ViewCount int64 `protobuf:"varint,10,opt,name=view_count,json=viewCount,proto3" json:"view_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/rpcmessages.pb.go b/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/rpcmessages.pb.go index 012f5bd55..12056a0ea 100644 --- a/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/rpcmessages.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/rpcmessages.pb.go @@ -50,7 +50,7 @@ func (PhotoView) EnumDescriptor() ([]byte, []int) { // Request to create a photo. type CreatePhotoRequest struct { // Required. Photo to create. - Photo *Photo `protobuf:"bytes,1,opt,name=photo" json:"photo,omitempty"` + Photo *Photo `protobuf:"bytes,1,opt,name=photo,proto3" json:"photo,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -96,10 +96,10 @@ func (m *CreatePhotoRequest) GetPhoto() *Photo { // - 'view' controls if the download URL for the photo bytes will be returned. type GetPhotoRequest struct { // Required. ID of the photo. - PhotoId string `protobuf:"bytes,1,opt,name=photo_id,json=photoId" json:"photo_id,omitempty"` + PhotoId string `protobuf:"bytes,1,opt,name=photo_id,json=photoId,proto3" json:"photo_id,omitempty"` // Specifies if a download URL for the photo bytes should be returned in the // Photo response. - View PhotoView `protobuf:"varint,2,opt,name=view,enum=google.streetview.publish.v1.PhotoView" json:"view,omitempty"` + View PhotoView `protobuf:"varint,2,opt,name=view,proto3,enum=google.streetview.publish.v1.PhotoView" json:"view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -151,10 +151,10 @@ func (m *GetPhotoRequest) GetView() PhotoView { // - 'view' controls if the download URL for the photo bytes will be returned. type BatchGetPhotosRequest struct { // Required. IDs of the photos. - PhotoIds []string `protobuf:"bytes,1,rep,name=photo_ids,json=photoIds" json:"photo_ids,omitempty"` + PhotoIds []string `protobuf:"bytes,1,rep,name=photo_ids,json=photoIds,proto3" json:"photo_ids,omitempty"` // Specifies if a download URL for the photo bytes should be returned in the // Photo response. - View PhotoView `protobuf:"varint,2,opt,name=view,enum=google.streetview.publish.v1.PhotoView" json:"view,omitempty"` + View PhotoView `protobuf:"varint,2,opt,name=view,proto3,enum=google.streetview.publish.v1.PhotoView" json:"view,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -202,7 +202,7 @@ func (m *BatchGetPhotosRequest) GetView() PhotoView { type BatchGetPhotosResponse struct { // List of results for each individual photo requested, in the same order as // the request. - Results []*PhotoResponse `protobuf:"bytes,1,rep,name=results" json:"results,omitempty"` + Results []*PhotoResponse `protobuf:"bytes,1,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -244,9 +244,9 @@ func (m *BatchGetPhotosResponse) GetResults() []*PhotoResponse { type PhotoResponse struct { // The status for the operation to get or update a single photo in the batch // request. - Status *status.Status `protobuf:"bytes,1,opt,name=status" json:"status,omitempty"` + Status *status.Status `protobuf:"bytes,1,opt,name=status,proto3" json:"status,omitempty"` // The photo resource, if the request was successful. - Photo *Photo `protobuf:"bytes,2,opt,name=photo" json:"photo,omitempty"` + Photo *Photo `protobuf:"bytes,2,opt,name=photo,proto3" json:"photo,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -303,18 +303,18 @@ func (m *PhotoResponse) GetPhoto() *Photo { type ListPhotosRequest struct { // Specifies if a download URL for the photos bytes should be returned in the // Photos response. - View PhotoView `protobuf:"varint,1,opt,name=view,enum=google.streetview.publish.v1.PhotoView" json:"view,omitempty"` + View PhotoView `protobuf:"varint,1,opt,name=view,proto3,enum=google.streetview.publish.v1.PhotoView" json:"view,omitempty"` // The maximum number of photos to return. // `page_size` must be non-negative. If `page_size` is zero or is not // provided, the default page size of 100 will be used. // The number of photos returned in the response may be less than `page_size` // if the number of photos that belong to the user is less than `page_size`. - PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize" json:"page_size,omitempty"` + PageSize int32 `protobuf:"varint,2,opt,name=page_size,json=pageSize,proto3" json:"page_size,omitempty"` // The next_page_token value returned from a previous List request, if any. - PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken" json:"page_token,omitempty"` + PageToken string `protobuf:"bytes,3,opt,name=page_token,json=pageToken,proto3" json:"page_token,omitempty"` // The filter expression. // Example: `placeId=ChIJj61dQgK6j4AR4GeTYWZsKWw` - Filter string `protobuf:"bytes,4,opt,name=filter" json:"filter,omitempty"` + Filter string `protobuf:"bytes,4,opt,name=filter,proto3" json:"filter,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -376,10 +376,10 @@ func (m *ListPhotosRequest) GetFilter() string { type ListPhotosResponse struct { // List of photos. There will be a maximum number of items returned based on // the page_size field in the request. - Photos []*Photo `protobuf:"bytes,1,rep,name=photos" json:"photos,omitempty"` + Photos []*Photo `protobuf:"bytes,1,rep,name=photos,proto3" json:"photos,omitempty"` // Token to retrieve the next page of results, or empty if there are no // more results in the list. - NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken" json:"next_page_token,omitempty"` + NextPageToken string `protobuf:"bytes,2,opt,name=next_page_token,json=nextPageToken,proto3" json:"next_page_token,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -431,7 +431,7 @@ type UpdatePhotoRequest struct { // update applies to all fields. // **Note:** To update `pose.altitude`, `pose.latlngpair` has to be filled as // well. Otherwise, the request will fail. - Photo *Photo `protobuf:"bytes,1,opt,name=photo" json:"photo,omitempty"` + Photo *Photo `protobuf:"bytes,1,opt,name=photo,proto3" json:"photo,omitempty"` // Mask that identifies fields on the photo metadata to update. // If not present, the old Photo metadata will be entirely replaced with the // new Photo metadata in this request. The update fails if invalid fields are @@ -454,7 +454,7 @@ type UpdatePhotoRequest struct { // `UpdatePhotoRequest.photo.update_mask` contains `connections` and // `UpdatePhotoRequest.photo.connections` is empty, all connections will be // removed. - UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask" json:"update_mask,omitempty"` + UpdateMask *field_mask.FieldMask `protobuf:"bytes,2,opt,name=update_mask,json=updateMask,proto3" json:"update_mask,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -502,7 +502,7 @@ func (m *UpdatePhotoRequest) GetUpdateMask() *field_mask.FieldMask { // Updating the pixels of photos is not supported. type BatchUpdatePhotosRequest struct { // Required. List of update photo requests. - UpdatePhotoRequests []*UpdatePhotoRequest `protobuf:"bytes,1,rep,name=update_photo_requests,json=updatePhotoRequests" json:"update_photo_requests,omitempty"` + UpdatePhotoRequests []*UpdatePhotoRequest `protobuf:"bytes,1,rep,name=update_photo_requests,json=updatePhotoRequests,proto3" json:"update_photo_requests,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -543,7 +543,7 @@ func (m *BatchUpdatePhotosRequest) GetUpdatePhotoRequests() []*UpdatePhotoReques type BatchUpdatePhotosResponse struct { // List of results for each individual photo updated, in the same order as // the request. - Results []*PhotoResponse `protobuf:"bytes,1,rep,name=results" json:"results,omitempty"` + Results []*PhotoResponse `protobuf:"bytes,1,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -583,7 +583,7 @@ func (m *BatchUpdatePhotosResponse) GetResults() []*PhotoResponse { // Request to delete a photo. type DeletePhotoRequest struct { // Required. ID of the photo. - PhotoId string `protobuf:"bytes,1,opt,name=photo_id,json=photoId" json:"photo_id,omitempty"` + PhotoId string `protobuf:"bytes,1,opt,name=photo_id,json=photoId,proto3" json:"photo_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -623,7 +623,7 @@ func (m *DeletePhotoRequest) GetPhotoId() string { // Request to delete multiple photos. type BatchDeletePhotosRequest struct { // Required. List of delete photo requests. - PhotoIds []string `protobuf:"bytes,1,rep,name=photo_ids,json=photoIds" json:"photo_ids,omitempty"` + PhotoIds []string `protobuf:"bytes,1,rep,name=photo_ids,json=photoIds,proto3" json:"photo_ids,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -663,7 +663,7 @@ func (m *BatchDeletePhotosRequest) GetPhotoIds() []string { // Response to batch delete of one or more photos. type BatchDeletePhotosResponse struct { // The status for the operation to delete a single photo in the batch request. - Status []*status.Status `protobuf:"bytes,1,rep,name=status" json:"status,omitempty"` + Status []*status.Status `protobuf:"bytes,1,rep,name=status,proto3" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/streetview_publish.pb.go b/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/streetview_publish.pb.go index 4a5766416..e8ced687b 100644 --- a/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/streetview_publish.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/streetview/publish/v1/streetview_publish.pb.go @@ -200,8 +200,7 @@ func (c *streetViewPublishServiceClient) BatchDeletePhotos(ctx context.Context, return out, nil } -// Server API for StreetViewPublishService service - +// StreetViewPublishServiceServer is the server API for StreetViewPublishService service. type StreetViewPublishServiceServer interface { // Creates an upload session to start uploading photo data. The upload URL of // the returned `UploadRef` is used to upload the data for the photo. diff --git a/vendor/google.golang.org/genproto/googleapis/type/color/color.pb.go b/vendor/google.golang.org/genproto/googleapis/type/color/color.pb.go index 79f42865c..d55342f6a 100644 --- a/vendor/google.golang.org/genproto/googleapis/type/color/color.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/type/color/color.pb.go @@ -135,11 +135,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // // ... type Color struct { // The amount of red in the color as a value in the interval [0, 1]. - Red float32 `protobuf:"fixed32,1,opt,name=red" json:"red,omitempty"` + Red float32 `protobuf:"fixed32,1,opt,name=red,proto3" json:"red,omitempty"` // The amount of green in the color as a value in the interval [0, 1]. - Green float32 `protobuf:"fixed32,2,opt,name=green" json:"green,omitempty"` + Green float32 `protobuf:"fixed32,2,opt,name=green,proto3" json:"green,omitempty"` // The amount of blue in the color as a value in the interval [0, 1]. - Blue float32 `protobuf:"fixed32,3,opt,name=blue" json:"blue,omitempty"` + Blue float32 `protobuf:"fixed32,3,opt,name=blue,proto3" json:"blue,omitempty"` // The fraction of this color that should be applied to the pixel. That is, // the final pixel color is defined by the equation: // @@ -151,7 +151,7 @@ type Color struct { // possible to distinguish between a default value and the value being unset. // If omitted, this color object is to be rendered as a solid color // (as if the alpha value had been explicitly given with a value of 1.0). - Alpha *wrappers.FloatValue `protobuf:"bytes,4,opt,name=alpha" json:"alpha,omitempty"` + Alpha *wrappers.FloatValue `protobuf:"bytes,4,opt,name=alpha,proto3" json:"alpha,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/type/date/date.pb.go b/vendor/google.golang.org/genproto/googleapis/type/date/date.pb.go index 996b12784..feb17fe42 100644 --- a/vendor/google.golang.org/genproto/googleapis/type/date/date.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/type/date/date.pb.go @@ -28,12 +28,12 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Date struct { // Year of date. Must be from 1 to 9999, or 0 if specifying a date without // a year. - Year int32 `protobuf:"varint,1,opt,name=year" json:"year,omitempty"` + Year int32 `protobuf:"varint,1,opt,name=year,proto3" json:"year,omitempty"` // Month of year. Must be from 1 to 12. - Month int32 `protobuf:"varint,2,opt,name=month" json:"month,omitempty"` + Month int32 `protobuf:"varint,2,opt,name=month,proto3" json:"month,omitempty"` // Day of month. Must be from 1 to 31 and valid for the year and month, or 0 // if specifying a year/month where the day is not significant. - Day int32 `protobuf:"varint,3,opt,name=day" json:"day,omitempty"` + Day int32 `protobuf:"varint,3,opt,name=day,proto3" json:"day,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/type/latlng/latlng.pb.go b/vendor/google.golang.org/genproto/googleapis/type/latlng/latlng.pb.go index 144327c57..2fd90d437 100644 --- a/vendor/google.golang.org/genproto/googleapis/type/latlng/latlng.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/type/latlng/latlng.pb.go @@ -59,9 +59,9 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0) type LatLng struct { // The latitude in degrees. It must be in the range [-90.0, +90.0]. - Latitude float64 `protobuf:"fixed64,1,opt,name=latitude" json:"latitude,omitempty"` + Latitude float64 `protobuf:"fixed64,1,opt,name=latitude,proto3" json:"latitude,omitempty"` // The longitude in degrees. It must be in the range [-180.0, +180.0]. - Longitude float64 `protobuf:"fixed64,2,opt,name=longitude" json:"longitude,omitempty"` + Longitude float64 `protobuf:"fixed64,2,opt,name=longitude,proto3" json:"longitude,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/type/money/money.pb.go b/vendor/google.golang.org/genproto/googleapis/type/money/money.pb.go index 1aa64f3e6..9187be69c 100644 --- a/vendor/google.golang.org/genproto/googleapis/type/money/money.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/type/money/money.pb.go @@ -21,17 +21,17 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Represents an amount of money with its currency type. type Money struct { // The 3-letter currency code defined in ISO 4217. - CurrencyCode string `protobuf:"bytes,1,opt,name=currency_code,json=currencyCode" json:"currency_code,omitempty"` + CurrencyCode string `protobuf:"bytes,1,opt,name=currency_code,json=currencyCode,proto3" json:"currency_code,omitempty"` // The whole units of the amount. // For example if `currencyCode` is `"USD"`, then 1 unit is one US dollar. - Units int64 `protobuf:"varint,2,opt,name=units" json:"units,omitempty"` + Units int64 `protobuf:"varint,2,opt,name=units,proto3" json:"units,omitempty"` // Number of nano (10^-9) units of the amount. // The value must be between -999,999,999 and +999,999,999 inclusive. // If `units` is positive, `nanos` must be positive or zero. // If `units` is zero, `nanos` can be positive, zero, or negative. // If `units` is negative, `nanos` must be negative or zero. // For example $-1.75 is represented as `units`=-1 and `nanos`=-750,000,000. - Nanos int32 `protobuf:"varint,3,opt,name=nanos" json:"nanos,omitempty"` + Nanos int32 `protobuf:"varint,3,opt,name=nanos,proto3" json:"nanos,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/type/postaladdress/postal_address.pb.go b/vendor/google.golang.org/genproto/googleapis/type/postaladdress/postal_address.pb.go index ceb909da8..68ff4a2b7 100644 --- a/vendor/google.golang.org/genproto/googleapis/type/postaladdress/postal_address.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/type/postaladdress/postal_address.pb.go @@ -38,13 +38,13 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type PostalAddress struct { // The schema revision of the `PostalAddress`. // All new revisions **must** be backward compatible with old revisions. - Revision int32 `protobuf:"varint,1,opt,name=revision" json:"revision,omitempty"` + Revision int32 `protobuf:"varint,1,opt,name=revision,proto3" json:"revision,omitempty"` // Required. CLDR region code of the country/region of the address. This // is never inferred and it is up to the user to ensure the value is // correct. See http://cldr.unicode.org/ and // http://www.unicode.org/cldr/charts/30/supplemental/territory_information.html // for details. Example: "CH" for Switzerland. - RegionCode string `protobuf:"bytes,2,opt,name=region_code,json=regionCode" json:"region_code,omitempty"` + RegionCode string `protobuf:"bytes,2,opt,name=region_code,json=regionCode,proto3" json:"region_code,omitempty"` // Optional. BCP-47 language code of the contents of this address (if // known). This is often the UI language of the input form or is expected // to match one of the languages used in the address' country/region, or their @@ -57,18 +57,18 @@ type PostalAddress struct { // possibly incorrect default). // // Examples: "zh-Hant", "ja", "ja-Latn", "en". - LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` + LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // Optional. Postal code of the address. Not all countries use or require // postal codes to be present, but where they are used, they may trigger // additional validation with other parts of the address (e.g. state/zip // validation in the U.S.A.). - PostalCode string `protobuf:"bytes,4,opt,name=postal_code,json=postalCode" json:"postal_code,omitempty"` + PostalCode string `protobuf:"bytes,4,opt,name=postal_code,json=postalCode,proto3" json:"postal_code,omitempty"` // Optional. Additional, country-specific, sorting code. This is not used // in most regions. Where it is used, the value is either a string like // "CEDEX", optionally followed by a number (e.g. "CEDEX 7"), or just a number // alone, representing the "sector code" (Jamaica), "delivery area indicator" // (Malawi) or "post office indicator" (e.g. Côte d'Ivoire). - SortingCode string `protobuf:"bytes,5,opt,name=sorting_code,json=sortingCode" json:"sorting_code,omitempty"` + SortingCode string `protobuf:"bytes,5,opt,name=sorting_code,json=sortingCode,proto3" json:"sorting_code,omitempty"` // Optional. Highest administrative subdivision which is used for postal // addresses of a country or region. // For example, this can be a state, a province, an oblast, or a prefecture. @@ -76,15 +76,15 @@ type PostalAddress struct { // community (e.g. "Barcelona" and not "Catalonia"). // Many countries don't use an administrative area in postal addresses. E.g. // in Switzerland this should be left unpopulated. - AdministrativeArea string `protobuf:"bytes,6,opt,name=administrative_area,json=administrativeArea" json:"administrative_area,omitempty"` + AdministrativeArea string `protobuf:"bytes,6,opt,name=administrative_area,json=administrativeArea,proto3" json:"administrative_area,omitempty"` // Optional. Generally refers to the city/town portion of the address. // Examples: US city, IT comune, UK post town. // In regions of the world where localities are not well defined or do not fit // into this structure well, leave locality empty and use address_lines. - Locality string `protobuf:"bytes,7,opt,name=locality" json:"locality,omitempty"` + Locality string `protobuf:"bytes,7,opt,name=locality,proto3" json:"locality,omitempty"` // Optional. Sublocality of the address. // For example, this can be neighborhoods, boroughs, districts. - Sublocality string `protobuf:"bytes,8,opt,name=sublocality" json:"sublocality,omitempty"` + Sublocality string `protobuf:"bytes,8,opt,name=sublocality,proto3" json:"sublocality,omitempty"` // Unstructured address lines describing the lower levels of an address. // // Because values in address_lines do not have type information and may @@ -107,13 +107,13 @@ type PostalAddress struct { // then geocoding is the recommended way to handle completely unstructured // addresses (as opposed to guessing which parts of the address should be // localities or administrative areas). - AddressLines []string `protobuf:"bytes,9,rep,name=address_lines,json=addressLines" json:"address_lines,omitempty"` + AddressLines []string `protobuf:"bytes,9,rep,name=address_lines,json=addressLines,proto3" json:"address_lines,omitempty"` // Optional. The recipient at the address. // This field may, under certain circumstances, contain multiline information. // For example, it might contain "care of" information. - Recipients []string `protobuf:"bytes,10,rep,name=recipients" json:"recipients,omitempty"` + Recipients []string `protobuf:"bytes,10,rep,name=recipients,proto3" json:"recipients,omitempty"` // Optional. The name of the organization at the address. - Organization string `protobuf:"bytes,11,opt,name=organization" json:"organization,omitempty"` + Organization string `protobuf:"bytes,11,opt,name=organization,proto3" json:"organization,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/type/timeofday/timeofday.pb.go b/vendor/google.golang.org/genproto/googleapis/type/timeofday/timeofday.pb.go index 2d1d6ae99..7a2122b25 100644 --- a/vendor/google.golang.org/genproto/googleapis/type/timeofday/timeofday.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/type/timeofday/timeofday.pb.go @@ -24,14 +24,14 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type TimeOfDay struct { // Hours of day in 24 hour format. Should be from 0 to 23. An API may choose // to allow the value "24:00:00" for scenarios like business closing time. - Hours int32 `protobuf:"varint,1,opt,name=hours" json:"hours,omitempty"` + Hours int32 `protobuf:"varint,1,opt,name=hours,proto3" json:"hours,omitempty"` // Minutes of hour of day. Must be from 0 to 59. - Minutes int32 `protobuf:"varint,2,opt,name=minutes" json:"minutes,omitempty"` + Minutes int32 `protobuf:"varint,2,opt,name=minutes,proto3" json:"minutes,omitempty"` // Seconds of minutes of the time. Must normally be from 0 to 59. An API may // allow the value 60 if it allows leap-seconds. - Seconds int32 `protobuf:"varint,3,opt,name=seconds" json:"seconds,omitempty"` + Seconds int32 `protobuf:"varint,3,opt,name=seconds,proto3" json:"seconds,omitempty"` // Fractions of seconds in nanoseconds. Must be from 0 to 999,999,999. - Nanos int32 `protobuf:"varint,4,opt,name=nanos" json:"nanos,omitempty"` + Nanos int32 `protobuf:"varint,4,opt,name=nanos,proto3" json:"nanos,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/googleapis/watcher/v1/watch.pb.go b/vendor/google.golang.org/genproto/googleapis/watcher/v1/watch.pb.go index 2c13f3570..2babad420 100644 --- a/vendor/google.golang.org/genproto/googleapis/watcher/v1/watch.pb.go +++ b/vendor/google.golang.org/genproto/googleapis/watcher/v1/watch.pb.go @@ -90,7 +90,7 @@ type Request struct { // contains a special character, it must be %-encoded. We recommend that // clients and servers use their runtime's URL library to produce and consume // target values. - Target string `protobuf:"bytes,1,opt,name=target" json:"target,omitempty"` + Target string `protobuf:"bytes,1,opt,name=target,proto3" json:"target,omitempty"` // The `resume_marker` specifies how much of the existing underlying state is // delivered to the client when the watch request is received by the // system. The client can set this marker in one of the following ways to get @@ -167,7 +167,7 @@ func (m *Request) GetResumeMarker() []byte { // A batch of Change messages. type ChangeBatch struct { // A list of Change messages. - Changes []*Change `protobuf:"bytes,1,rep,name=changes" json:"changes,omitempty"` + Changes []*Change `protobuf:"bytes,1,rep,name=changes,proto3" json:"changes,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -209,13 +209,13 @@ type Change struct { // Name of the element, interpreted relative to the entity's actual // name. "" refers to the entity itself. The element name is a valid // UTF-8 string. - Element string `protobuf:"bytes,1,opt,name=element" json:"element,omitempty"` + Element string `protobuf:"bytes,1,opt,name=element,proto3" json:"element,omitempty"` // The state of the `element`. - State Change_State `protobuf:"varint,2,opt,name=state,enum=google.watcher.v1.Change_State" json:"state,omitempty"` + State Change_State `protobuf:"varint,2,opt,name=state,proto3,enum=google.watcher.v1.Change_State" json:"state,omitempty"` // The actual change data. This field is present only when `state() == EXISTS` // or `state() == ERROR`. Please see [google.protobuf.Any][google.protobuf.Any] about how to use // the Any type. - Data *any.Any `protobuf:"bytes,6,opt,name=data" json:"data,omitempty"` + Data *any.Any `protobuf:"bytes,6,opt,name=data,proto3" json:"data,omitempty"` // If present, provides a compact representation of all the messages that have // been received by the caller for the given entity, e.g., it could be a // sequence number or a multi-part timestamp/version vector. This marker can @@ -224,7 +224,7 @@ type Change struct { ResumeMarker []byte `protobuf:"bytes,4,opt,name=resume_marker,json=resumeMarker,proto3" json:"resume_marker,omitempty"` // If true, this Change is followed by more Changes that are in the same group // as this Change. - Continued bool `protobuf:"varint,5,opt,name=continued" json:"continued,omitempty"` + Continued bool `protobuf:"varint,5,opt,name=continued,proto3" json:"continued,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -352,8 +352,7 @@ func (x *watcherWatchClient) Recv() (*ChangeBatch, error) { return m, nil } -// Server API for Watcher service - +// WatcherServer is the server API for Watcher service. type WatcherServer interface { // Start a streaming RPC to get watch information from the server. Watch(*Request, Watcher_WatchServer) error diff --git a/vendor/google.golang.org/genproto/protobuf/api/api.pb.go b/vendor/google.golang.org/genproto/protobuf/api/api.pb.go index cf51a1495..617934549 100644 --- a/vendor/google.golang.org/genproto/protobuf/api/api.pb.go +++ b/vendor/google.golang.org/genproto/protobuf/api/api.pb.go @@ -32,11 +32,11 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type Api struct { // The fully qualified name of this interface, including package name // followed by the interface's simple name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The methods of this interface, in unspecified order. - Methods []*Method `protobuf:"bytes,2,rep,name=methods" json:"methods,omitempty"` + Methods []*Method `protobuf:"bytes,2,rep,name=methods,proto3" json:"methods,omitempty"` // Any metadata attached to the interface. - Options []*ptype.Option `protobuf:"bytes,3,rep,name=options" json:"options,omitempty"` + Options []*ptype.Option `protobuf:"bytes,3,rep,name=options,proto3" json:"options,omitempty"` // A version string for this interface. If specified, must have the form // `major-version.minor-version`, as in `1.10`. If the minor version is // omitted, it defaults to zero. If the entire version field is empty, the @@ -58,14 +58,14 @@ type Api struct { // experimental, non-GA interfaces. // // - Version string `protobuf:"bytes,4,opt,name=version" json:"version,omitempty"` + Version string `protobuf:"bytes,4,opt,name=version,proto3" json:"version,omitempty"` // Source context for the protocol buffer service represented by this // message. - SourceContext *source_context.SourceContext `protobuf:"bytes,5,opt,name=source_context,json=sourceContext" json:"source_context,omitempty"` + SourceContext *source_context.SourceContext `protobuf:"bytes,5,opt,name=source_context,json=sourceContext,proto3" json:"source_context,omitempty"` // Included interfaces. See [Mixin][]. - Mixins []*Mixin `protobuf:"bytes,6,rep,name=mixins" json:"mixins,omitempty"` + Mixins []*Mixin `protobuf:"bytes,6,rep,name=mixins,proto3" json:"mixins,omitempty"` // The source syntax of the service. - Syntax ptype.Syntax `protobuf:"varint,7,opt,name=syntax,enum=google.protobuf.Syntax" json:"syntax,omitempty"` + Syntax ptype.Syntax `protobuf:"varint,7,opt,name=syntax,proto3,enum=google.protobuf.Syntax" json:"syntax,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -147,19 +147,19 @@ func (m *Api) GetSyntax() ptype.Syntax { // Method represents a method of an API interface. type Method struct { // The simple name of this method. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // A URL of the input message type. - RequestTypeUrl string `protobuf:"bytes,2,opt,name=request_type_url,json=requestTypeUrl" json:"request_type_url,omitempty"` + RequestTypeUrl string `protobuf:"bytes,2,opt,name=request_type_url,json=requestTypeUrl,proto3" json:"request_type_url,omitempty"` // If true, the request is streamed. - RequestStreaming bool `protobuf:"varint,3,opt,name=request_streaming,json=requestStreaming" json:"request_streaming,omitempty"` + RequestStreaming bool `protobuf:"varint,3,opt,name=request_streaming,json=requestStreaming,proto3" json:"request_streaming,omitempty"` // The URL of the output message type. - ResponseTypeUrl string `protobuf:"bytes,4,opt,name=response_type_url,json=responseTypeUrl" json:"response_type_url,omitempty"` + ResponseTypeUrl string `protobuf:"bytes,4,opt,name=response_type_url,json=responseTypeUrl,proto3" json:"response_type_url,omitempty"` // If true, the response is streamed. - ResponseStreaming bool `protobuf:"varint,5,opt,name=response_streaming,json=responseStreaming" json:"response_streaming,omitempty"` + ResponseStreaming bool `protobuf:"varint,5,opt,name=response_streaming,json=responseStreaming,proto3" json:"response_streaming,omitempty"` // Any metadata attached to the method. - Options []*ptype.Option `protobuf:"bytes,6,rep,name=options" json:"options,omitempty"` + Options []*ptype.Option `protobuf:"bytes,6,rep,name=options,proto3" json:"options,omitempty"` // The source syntax of this method. - Syntax ptype.Syntax `protobuf:"varint,7,opt,name=syntax,enum=google.protobuf.Syntax" json:"syntax,omitempty"` + Syntax ptype.Syntax `protobuf:"varint,7,opt,name=syntax,proto3,enum=google.protobuf.Syntax" json:"syntax,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -318,10 +318,10 @@ func (m *Method) GetSyntax() ptype.Syntax { // } type Mixin struct { // The fully qualified name of the interface which is included. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // If non-empty specifies a path under which inherited HTTP paths // are rooted. - Root string `protobuf:"bytes,2,opt,name=root" json:"root,omitempty"` + Root string `protobuf:"bytes,2,opt,name=root,proto3" json:"root,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/protobuf/field_mask/field_mask.pb.go b/vendor/google.golang.org/genproto/protobuf/field_mask/field_mask.pb.go index ace751f68..f7059620b 100644 --- a/vendor/google.golang.org/genproto/protobuf/field_mask/field_mask.pb.go +++ b/vendor/google.golang.org/genproto/protobuf/field_mask/field_mask.pb.go @@ -227,7 +227,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // `INVALID_ARGUMENT` error if any path is duplicated or unmappable. type FieldMask struct { // The set of field mask paths. - Paths []string `protobuf:"bytes,1,rep,name=paths" json:"paths,omitempty"` + Paths []string `protobuf:"bytes,1,rep,name=paths,proto3" json:"paths,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/protobuf/ptype/type.pb.go b/vendor/google.golang.org/genproto/protobuf/ptype/type.pb.go index 52dfb5e30..6a88e1d8c 100644 --- a/vendor/google.golang.org/genproto/protobuf/ptype/type.pb.go +++ b/vendor/google.golang.org/genproto/protobuf/ptype/type.pb.go @@ -177,17 +177,17 @@ func (Field_Cardinality) EnumDescriptor() ([]byte, []int) { // A protocol buffer message type. type Type struct { // The fully qualified message name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The list of fields. - Fields []*Field `protobuf:"bytes,2,rep,name=fields" json:"fields,omitempty"` + Fields []*Field `protobuf:"bytes,2,rep,name=fields,proto3" json:"fields,omitempty"` // The list of types appearing in `oneof` definitions in this type. - Oneofs []string `protobuf:"bytes,3,rep,name=oneofs" json:"oneofs,omitempty"` + Oneofs []string `protobuf:"bytes,3,rep,name=oneofs,proto3" json:"oneofs,omitempty"` // The protocol buffer options. - Options []*Option `protobuf:"bytes,4,rep,name=options" json:"options,omitempty"` + Options []*Option `protobuf:"bytes,4,rep,name=options,proto3" json:"options,omitempty"` // The source context. - SourceContext *source_context.SourceContext `protobuf:"bytes,5,opt,name=source_context,json=sourceContext" json:"source_context,omitempty"` + SourceContext *source_context.SourceContext `protobuf:"bytes,5,opt,name=source_context,json=sourceContext,proto3" json:"source_context,omitempty"` // The source syntax. - Syntax Syntax `protobuf:"varint,6,opt,name=syntax,enum=google.protobuf.Syntax" json:"syntax,omitempty"` + Syntax Syntax `protobuf:"varint,6,opt,name=syntax,proto3,enum=google.protobuf.Syntax" json:"syntax,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -262,27 +262,27 @@ func (m *Type) GetSyntax() Syntax { // A single field of a message type. type Field struct { // The field type. - Kind Field_Kind `protobuf:"varint,1,opt,name=kind,enum=google.protobuf.Field_Kind" json:"kind,omitempty"` + Kind Field_Kind `protobuf:"varint,1,opt,name=kind,proto3,enum=google.protobuf.Field_Kind" json:"kind,omitempty"` // The field cardinality. - Cardinality Field_Cardinality `protobuf:"varint,2,opt,name=cardinality,enum=google.protobuf.Field_Cardinality" json:"cardinality,omitempty"` + Cardinality Field_Cardinality `protobuf:"varint,2,opt,name=cardinality,proto3,enum=google.protobuf.Field_Cardinality" json:"cardinality,omitempty"` // The field number. - Number int32 `protobuf:"varint,3,opt,name=number" json:"number,omitempty"` + Number int32 `protobuf:"varint,3,opt,name=number,proto3" json:"number,omitempty"` // The field name. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` // The field type URL, without the scheme, for message or enumeration // types. Example: `"type.googleapis.com/google.protobuf.Timestamp"`. - TypeUrl string `protobuf:"bytes,6,opt,name=type_url,json=typeUrl" json:"type_url,omitempty"` + TypeUrl string `protobuf:"bytes,6,opt,name=type_url,json=typeUrl,proto3" json:"type_url,omitempty"` // The index of the field type in `Type.oneofs`, for message or enumeration // types. The first type has index 1; zero means the type is not in the list. - OneofIndex int32 `protobuf:"varint,7,opt,name=oneof_index,json=oneofIndex" json:"oneof_index,omitempty"` + OneofIndex int32 `protobuf:"varint,7,opt,name=oneof_index,json=oneofIndex,proto3" json:"oneof_index,omitempty"` // Whether to use alternative packed wire representation. - Packed bool `protobuf:"varint,8,opt,name=packed" json:"packed,omitempty"` + Packed bool `protobuf:"varint,8,opt,name=packed,proto3" json:"packed,omitempty"` // The protocol buffer options. - Options []*Option `protobuf:"bytes,9,rep,name=options" json:"options,omitempty"` + Options []*Option `protobuf:"bytes,9,rep,name=options,proto3" json:"options,omitempty"` // The field JSON name. - JsonName string `protobuf:"bytes,10,opt,name=json_name,json=jsonName" json:"json_name,omitempty"` + JsonName string `protobuf:"bytes,10,opt,name=json_name,json=jsonName,proto3" json:"json_name,omitempty"` // The string value of the default value of this field. Proto2 syntax only. - DefaultValue string `protobuf:"bytes,11,opt,name=default_value,json=defaultValue" json:"default_value,omitempty"` + DefaultValue string `protobuf:"bytes,11,opt,name=default_value,json=defaultValue,proto3" json:"default_value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -385,15 +385,15 @@ func (m *Field) GetDefaultValue() string { // Enum type definition. type Enum struct { // Enum type name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Enum value definitions. - Enumvalue []*EnumValue `protobuf:"bytes,2,rep,name=enumvalue" json:"enumvalue,omitempty"` + Enumvalue []*EnumValue `protobuf:"bytes,2,rep,name=enumvalue,proto3" json:"enumvalue,omitempty"` // Protocol buffer options. - Options []*Option `protobuf:"bytes,3,rep,name=options" json:"options,omitempty"` + Options []*Option `protobuf:"bytes,3,rep,name=options,proto3" json:"options,omitempty"` // The source context. - SourceContext *source_context.SourceContext `protobuf:"bytes,4,opt,name=source_context,json=sourceContext" json:"source_context,omitempty"` + SourceContext *source_context.SourceContext `protobuf:"bytes,4,opt,name=source_context,json=sourceContext,proto3" json:"source_context,omitempty"` // The source syntax. - Syntax Syntax `protobuf:"varint,5,opt,name=syntax,enum=google.protobuf.Syntax" json:"syntax,omitempty"` + Syntax Syntax `protobuf:"varint,5,opt,name=syntax,proto3,enum=google.protobuf.Syntax" json:"syntax,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -461,11 +461,11 @@ func (m *Enum) GetSyntax() Syntax { // Enum value definition. type EnumValue struct { // Enum value name. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Enum value number. - Number int32 `protobuf:"varint,2,opt,name=number" json:"number,omitempty"` + Number int32 `protobuf:"varint,2,opt,name=number,proto3" json:"number,omitempty"` // Protocol buffer options. - Options []*Option `protobuf:"bytes,3,rep,name=options" json:"options,omitempty"` + Options []*Option `protobuf:"bytes,3,rep,name=options,proto3" json:"options,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -523,12 +523,12 @@ type Option struct { // descriptor.proto), this is the short name. For example, `"map_entry"`. // For custom options, it should be the fully-qualified name. For example, // `"google.api.http"`. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The option's value packed in an Any message. If the value is a primitive, // the corresponding wrapper type defined in google/protobuf/wrappers.proto // should be used. If the value is an enum, it should be stored as an int32 // value using the google.protobuf.Int32Value type. - Value *any.Any `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value *any.Any `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/genproto/protobuf/source_context/source_context.pb.go b/vendor/google.golang.org/genproto/protobuf/source_context/source_context.pb.go index c4e03335e..67a8873c1 100644 --- a/vendor/google.golang.org/genproto/protobuf/source_context/source_context.pb.go +++ b/vendor/google.golang.org/genproto/protobuf/source_context/source_context.pb.go @@ -23,7 +23,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type SourceContext struct { // The path-qualified name of the .proto file that contained the associated // protobuf element. For example: `"google/protobuf/source_context.proto"`. - FileName string `protobuf:"bytes,1,opt,name=file_name,json=fileName" json:"file_name,omitempty"` + FileName string `protobuf:"bytes,1,opt,name=file_name,json=fileName,proto3" json:"file_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/grpc/Documentation/concurrency.md b/vendor/google.golang.org/grpc/Documentation/concurrency.md new file mode 100644 index 000000000..d0c1142a0 --- /dev/null +++ b/vendor/google.golang.org/grpc/Documentation/concurrency.md @@ -0,0 +1,33 @@ +# Concurrency + +In general, gRPC-go provides a concurrency-friendly API. What follows are some +guidelines. + +## Clients + +A [ClientConn][client-conn] can safely be accessed concurrently. Using +[helloworld][helloworld] as an example, one could share the `ClientConn` across +multiple goroutines to create multiple `GreeterClient` types. In this case, RPCs +would be sent in parallel. + +## Streams + +When using streams, one must take care to avoid calling either `SendMsg` or +`RecvMsg` multiple times against the same [Stream][stream] from different +goroutines. In other words, it's safe to have a goroutine calling `SendMsg` and +another goroutine calling `RecvMsg` on the same stream at the same time. But it +is not safe to call `SendMsg` on the same stream in different goroutines, or to +call `RecvMsg` on the same stream in different goroutines. + +## Servers + +Each RPC handler attached to a registered server will be invoked in its own +goroutine. For example, [SayHello][say-hello] will be invoked in its own +goroutine. The same is true for service handlers for streaming RPCs, as seen +in the route guide example [here][route-guide-stream]. + +[helloworld]: https://github.com/grpc/grpc-go/blob/master/examples/helloworld/greeter_client/main.go#L43 +[client-conn]: https://godoc.org/google.golang.org/grpc#ClientConn +[stream]: https://godoc.org/google.golang.org/grpc#Stream +[say-hello]: https://github.com/grpc/grpc-go/blob/master/examples/helloworld/greeter_server/main.go#L41 +[route-guide-stream]: https://github.com/grpc/grpc-go/blob/master/examples/route_guide/server/server.go#L126 diff --git a/vendor/google.golang.org/grpc/Documentation/log_levels.md b/vendor/google.golang.org/grpc/Documentation/log_levels.md new file mode 100644 index 000000000..6839f6d09 --- /dev/null +++ b/vendor/google.golang.org/grpc/Documentation/log_levels.md @@ -0,0 +1,49 @@ +# Log Levels + +This document describes the different log levels supported by the grpc-go +library, and under what conditions they should be used. + +### Info + +Info messages are for informational purposes and may aid in the debugging of +applications or the gRPC library. + +Examples: +- The name resolver received an update. +- The balancer updated its picker. +- Significant gRPC state is changing. + +At verbosity of 0 (the default), any single info message should not be output +more than once every 5 minutes under normal operation. + +### Warning + +Warning messages indicate problems that are non-fatal for the application, but +could lead to unexpected behavior or subsequent errors. + +Examples: +- Resolver could not resolve target name. +- Error received while connecting to a server. +- Lost or corrupt connection with remote endpoint. + +### Error + +Error messages represent errors in the usage of gRPC that cannot be returned to +the application as errors, or internal gRPC-Go errors that are recoverable. + +Internal errors are detected during gRPC tests and will result in test failures. + +Examples: +- Invalid arguments passed to a function that cannot return an error. +- An internal error that cannot be returned or would be inappropriate to return + to the user. + +### Fatal + +Fatal errors are severe internal errors that are unrecoverable. These lead +directly to panics, and are avoided as much as possible. + +Example: +- Internal invariant was violated. +- User attempted an action that cannot return an error gracefully, but would + lead to an invalid state if performed. diff --git a/vendor/google.golang.org/grpc/backoff.go b/vendor/google.golang.org/grpc/backoff.go index c40facce5..fa31565fd 100644 --- a/vendor/google.golang.org/grpc/backoff.go +++ b/vendor/google.golang.org/grpc/backoff.go @@ -16,81 +16,23 @@ * */ +// See internal/backoff package for the backoff implementation. This file is +// kept for the exported types and API backward compatility. + package grpc import ( - "math/rand" "time" ) // DefaultBackoffConfig uses values specified for backoff in // https://github.com/grpc/grpc/blob/master/doc/connection-backoff.md. var DefaultBackoffConfig = BackoffConfig{ - MaxDelay: 120 * time.Second, - baseDelay: 1.0 * time.Second, - factor: 1.6, - jitter: 0.2, -} - -// backoffStrategy defines the methodology for backing off after a grpc -// connection failure. -// -// This is unexported until the gRPC project decides whether or not to allow -// alternative backoff strategies. Once a decision is made, this type and its -// method may be exported. -type backoffStrategy interface { - // backoff returns the amount of time to wait before the next retry given - // the number of consecutive failures. - backoff(retries int) time.Duration + MaxDelay: 120 * time.Second, } // BackoffConfig defines the parameters for the default gRPC backoff strategy. type BackoffConfig struct { // MaxDelay is the upper bound of backoff delay. MaxDelay time.Duration - - // TODO(stevvooe): The following fields are not exported, as allowing - // changes would violate the current gRPC specification for backoff. If - // gRPC decides to allow more interesting backoff strategies, these fields - // may be opened up in the future. - - // baseDelay is the amount of time to wait before retrying after the first - // failure. - baseDelay time.Duration - - // factor is applied to the backoff after each retry. - factor float64 - - // jitter provides a range to randomize backoff delays. - jitter float64 -} - -func setDefaults(bc *BackoffConfig) { - md := bc.MaxDelay - *bc = DefaultBackoffConfig - - if md > 0 { - bc.MaxDelay = md - } -} - -func (bc BackoffConfig) backoff(retries int) time.Duration { - if retries == 0 { - return bc.baseDelay - } - backoff, max := float64(bc.baseDelay), float64(bc.MaxDelay) - for backoff < max && retries > 0 { - backoff *= bc.factor - retries-- - } - if backoff > max { - backoff = max - } - // Randomize backoff delays so that if a cluster of requests start at - // the same time, they won't operate in lockstep. - backoff *= 1 + bc.jitter*(rand.Float64()*2-1) - if backoff < 0 { - return 0 - } - return time.Duration(backoff) } diff --git a/vendor/google.golang.org/grpc/balancer/balancer.go b/vendor/google.golang.org/grpc/balancer/balancer.go index 63b8d7137..f9d83c2f3 100644 --- a/vendor/google.golang.org/grpc/balancer/balancer.go +++ b/vendor/google.golang.org/grpc/balancer/balancer.go @@ -226,3 +226,45 @@ type Balancer interface { // ClientConn.RemoveSubConn for its existing SubConns. Close() } + +// ConnectivityStateEvaluator takes the connectivity states of multiple SubConns +// and returns one aggregated connectivity state. +// +// It's not thread safe. +type ConnectivityStateEvaluator struct { + numReady uint64 // Number of addrConns in ready state. + numConnecting uint64 // Number of addrConns in connecting state. + numTransientFailure uint64 // Number of addrConns in transientFailure. +} + +// RecordTransition records state change happening in subConn and based on that +// it evaluates what aggregated state should be. +// +// - If at least one SubConn in Ready, the aggregated state is Ready; +// - Else if at least one SubConn in Connecting, the aggregated state is Connecting; +// - Else the aggregated state is TransientFailure. +// +// Idle and Shutdown are not considered. +func (cse *ConnectivityStateEvaluator) RecordTransition(oldState, newState connectivity.State) connectivity.State { + // Update counters. + for idx, state := range []connectivity.State{oldState, newState} { + updateVal := 2*uint64(idx) - 1 // -1 for oldState and +1 for new. + switch state { + case connectivity.Ready: + cse.numReady += updateVal + case connectivity.Connecting: + cse.numConnecting += updateVal + case connectivity.TransientFailure: + cse.numTransientFailure += updateVal + } + } + + // Evaluate. + if cse.numReady > 0 { + return connectivity.Ready + } + if cse.numConnecting > 0 { + return connectivity.Connecting + } + return connectivity.TransientFailure +} diff --git a/vendor/google.golang.org/grpc/balancer/grpclb/grpc_lb_v1/load_balancer.pb.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpc_lb_v1/load_balancer.pb.go new file mode 100644 index 000000000..d08bffc47 --- /dev/null +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpc_lb_v1/load_balancer.pb.go @@ -0,0 +1,832 @@ +// Code generated by protoc-gen-go. DO NOT EDIT. +// source: grpc/lb/v1/load_balancer.proto + +package grpc_lb_v1 // import "google.golang.org/grpc/balancer/grpclb/grpc_lb_v1" + +import proto "github.com/golang/protobuf/proto" +import fmt "fmt" +import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" + +import ( + context "golang.org/x/net/context" + grpc "google.golang.org/grpc" +) + +// Reference imports to suppress errors if they are not otherwise used. +var _ = proto.Marshal +var _ = fmt.Errorf +var _ = math.Inf + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + +type LoadBalanceRequest struct { + // Types that are valid to be assigned to LoadBalanceRequestType: + // *LoadBalanceRequest_InitialRequest + // *LoadBalanceRequest_ClientStats + LoadBalanceRequestType isLoadBalanceRequest_LoadBalanceRequestType `protobuf_oneof:"load_balance_request_type"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LoadBalanceRequest) Reset() { *m = LoadBalanceRequest{} } +func (m *LoadBalanceRequest) String() string { return proto.CompactTextString(m) } +func (*LoadBalanceRequest) ProtoMessage() {} +func (*LoadBalanceRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{0} +} +func (m *LoadBalanceRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LoadBalanceRequest.Unmarshal(m, b) +} +func (m *LoadBalanceRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LoadBalanceRequest.Marshal(b, m, deterministic) +} +func (dst *LoadBalanceRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_LoadBalanceRequest.Merge(dst, src) +} +func (m *LoadBalanceRequest) XXX_Size() int { + return xxx_messageInfo_LoadBalanceRequest.Size(m) +} +func (m *LoadBalanceRequest) XXX_DiscardUnknown() { + xxx_messageInfo_LoadBalanceRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_LoadBalanceRequest proto.InternalMessageInfo + +type isLoadBalanceRequest_LoadBalanceRequestType interface { + isLoadBalanceRequest_LoadBalanceRequestType() +} + +type LoadBalanceRequest_InitialRequest struct { + InitialRequest *InitialLoadBalanceRequest `protobuf:"bytes,1,opt,name=initial_request,json=initialRequest,proto3,oneof"` +} +type LoadBalanceRequest_ClientStats struct { + ClientStats *ClientStats `protobuf:"bytes,2,opt,name=client_stats,json=clientStats,proto3,oneof"` +} + +func (*LoadBalanceRequest_InitialRequest) isLoadBalanceRequest_LoadBalanceRequestType() {} +func (*LoadBalanceRequest_ClientStats) isLoadBalanceRequest_LoadBalanceRequestType() {} + +func (m *LoadBalanceRequest) GetLoadBalanceRequestType() isLoadBalanceRequest_LoadBalanceRequestType { + if m != nil { + return m.LoadBalanceRequestType + } + return nil +} + +func (m *LoadBalanceRequest) GetInitialRequest() *InitialLoadBalanceRequest { + if x, ok := m.GetLoadBalanceRequestType().(*LoadBalanceRequest_InitialRequest); ok { + return x.InitialRequest + } + return nil +} + +func (m *LoadBalanceRequest) GetClientStats() *ClientStats { + if x, ok := m.GetLoadBalanceRequestType().(*LoadBalanceRequest_ClientStats); ok { + return x.ClientStats + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*LoadBalanceRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _LoadBalanceRequest_OneofMarshaler, _LoadBalanceRequest_OneofUnmarshaler, _LoadBalanceRequest_OneofSizer, []interface{}{ + (*LoadBalanceRequest_InitialRequest)(nil), + (*LoadBalanceRequest_ClientStats)(nil), + } +} + +func _LoadBalanceRequest_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*LoadBalanceRequest) + // load_balance_request_type + switch x := m.LoadBalanceRequestType.(type) { + case *LoadBalanceRequest_InitialRequest: + b.EncodeVarint(1<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.InitialRequest); err != nil { + return err + } + case *LoadBalanceRequest_ClientStats: + b.EncodeVarint(2<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.ClientStats); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("LoadBalanceRequest.LoadBalanceRequestType has unexpected type %T", x) + } + return nil +} + +func _LoadBalanceRequest_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*LoadBalanceRequest) + switch tag { + case 1: // load_balance_request_type.initial_request + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(InitialLoadBalanceRequest) + err := b.DecodeMessage(msg) + m.LoadBalanceRequestType = &LoadBalanceRequest_InitialRequest{msg} + return true, err + case 2: // load_balance_request_type.client_stats + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(ClientStats) + err := b.DecodeMessage(msg) + m.LoadBalanceRequestType = &LoadBalanceRequest_ClientStats{msg} + return true, err + default: + return false, nil + } +} + +func _LoadBalanceRequest_OneofSizer(msg proto.Message) (n int) { + m := msg.(*LoadBalanceRequest) + // load_balance_request_type + switch x := m.LoadBalanceRequestType.(type) { + case *LoadBalanceRequest_InitialRequest: + s := proto.Size(x.InitialRequest) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *LoadBalanceRequest_ClientStats: + s := proto.Size(x.ClientStats) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +type InitialLoadBalanceRequest struct { + // Name of load balanced service (IE, service.googleapis.com). Its + // length should be less than 256 bytes. + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *InitialLoadBalanceRequest) Reset() { *m = InitialLoadBalanceRequest{} } +func (m *InitialLoadBalanceRequest) String() string { return proto.CompactTextString(m) } +func (*InitialLoadBalanceRequest) ProtoMessage() {} +func (*InitialLoadBalanceRequest) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{1} +} +func (m *InitialLoadBalanceRequest) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_InitialLoadBalanceRequest.Unmarshal(m, b) +} +func (m *InitialLoadBalanceRequest) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_InitialLoadBalanceRequest.Marshal(b, m, deterministic) +} +func (dst *InitialLoadBalanceRequest) XXX_Merge(src proto.Message) { + xxx_messageInfo_InitialLoadBalanceRequest.Merge(dst, src) +} +func (m *InitialLoadBalanceRequest) XXX_Size() int { + return xxx_messageInfo_InitialLoadBalanceRequest.Size(m) +} +func (m *InitialLoadBalanceRequest) XXX_DiscardUnknown() { + xxx_messageInfo_InitialLoadBalanceRequest.DiscardUnknown(m) +} + +var xxx_messageInfo_InitialLoadBalanceRequest proto.InternalMessageInfo + +func (m *InitialLoadBalanceRequest) GetName() string { + if m != nil { + return m.Name + } + return "" +} + +// Contains the number of calls finished for a particular load balance token. +type ClientStatsPerToken struct { + // See Server.load_balance_token. + LoadBalanceToken string `protobuf:"bytes,1,opt,name=load_balance_token,json=loadBalanceToken,proto3" json:"load_balance_token,omitempty"` + // The total number of RPCs that finished associated with the token. + NumCalls int64 `protobuf:"varint,2,opt,name=num_calls,json=numCalls,proto3" json:"num_calls,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ClientStatsPerToken) Reset() { *m = ClientStatsPerToken{} } +func (m *ClientStatsPerToken) String() string { return proto.CompactTextString(m) } +func (*ClientStatsPerToken) ProtoMessage() {} +func (*ClientStatsPerToken) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{2} +} +func (m *ClientStatsPerToken) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ClientStatsPerToken.Unmarshal(m, b) +} +func (m *ClientStatsPerToken) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ClientStatsPerToken.Marshal(b, m, deterministic) +} +func (dst *ClientStatsPerToken) XXX_Merge(src proto.Message) { + xxx_messageInfo_ClientStatsPerToken.Merge(dst, src) +} +func (m *ClientStatsPerToken) XXX_Size() int { + return xxx_messageInfo_ClientStatsPerToken.Size(m) +} +func (m *ClientStatsPerToken) XXX_DiscardUnknown() { + xxx_messageInfo_ClientStatsPerToken.DiscardUnknown(m) +} + +var xxx_messageInfo_ClientStatsPerToken proto.InternalMessageInfo + +func (m *ClientStatsPerToken) GetLoadBalanceToken() string { + if m != nil { + return m.LoadBalanceToken + } + return "" +} + +func (m *ClientStatsPerToken) GetNumCalls() int64 { + if m != nil { + return m.NumCalls + } + return 0 +} + +// Contains client level statistics that are useful to load balancing. Each +// count except the timestamp should be reset to zero after reporting the stats. +type ClientStats struct { + // The timestamp of generating the report. + Timestamp *timestamp.Timestamp `protobuf:"bytes,1,opt,name=timestamp,proto3" json:"timestamp,omitempty"` + // The total number of RPCs that started. + NumCallsStarted int64 `protobuf:"varint,2,opt,name=num_calls_started,json=numCallsStarted,proto3" json:"num_calls_started,omitempty"` + // The total number of RPCs that finished. + NumCallsFinished int64 `protobuf:"varint,3,opt,name=num_calls_finished,json=numCallsFinished,proto3" json:"num_calls_finished,omitempty"` + // The total number of RPCs that failed to reach a server except dropped RPCs. + NumCallsFinishedWithClientFailedToSend int64 `protobuf:"varint,6,opt,name=num_calls_finished_with_client_failed_to_send,json=numCallsFinishedWithClientFailedToSend,proto3" json:"num_calls_finished_with_client_failed_to_send,omitempty"` + // The total number of RPCs that finished and are known to have been received + // by a server. + NumCallsFinishedKnownReceived int64 `protobuf:"varint,7,opt,name=num_calls_finished_known_received,json=numCallsFinishedKnownReceived,proto3" json:"num_calls_finished_known_received,omitempty"` + // The list of dropped calls. + CallsFinishedWithDrop []*ClientStatsPerToken `protobuf:"bytes,8,rep,name=calls_finished_with_drop,json=callsFinishedWithDrop,proto3" json:"calls_finished_with_drop,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ClientStats) Reset() { *m = ClientStats{} } +func (m *ClientStats) String() string { return proto.CompactTextString(m) } +func (*ClientStats) ProtoMessage() {} +func (*ClientStats) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{3} +} +func (m *ClientStats) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ClientStats.Unmarshal(m, b) +} +func (m *ClientStats) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ClientStats.Marshal(b, m, deterministic) +} +func (dst *ClientStats) XXX_Merge(src proto.Message) { + xxx_messageInfo_ClientStats.Merge(dst, src) +} +func (m *ClientStats) XXX_Size() int { + return xxx_messageInfo_ClientStats.Size(m) +} +func (m *ClientStats) XXX_DiscardUnknown() { + xxx_messageInfo_ClientStats.DiscardUnknown(m) +} + +var xxx_messageInfo_ClientStats proto.InternalMessageInfo + +func (m *ClientStats) GetTimestamp() *timestamp.Timestamp { + if m != nil { + return m.Timestamp + } + return nil +} + +func (m *ClientStats) GetNumCallsStarted() int64 { + if m != nil { + return m.NumCallsStarted + } + return 0 +} + +func (m *ClientStats) GetNumCallsFinished() int64 { + if m != nil { + return m.NumCallsFinished + } + return 0 +} + +func (m *ClientStats) GetNumCallsFinishedWithClientFailedToSend() int64 { + if m != nil { + return m.NumCallsFinishedWithClientFailedToSend + } + return 0 +} + +func (m *ClientStats) GetNumCallsFinishedKnownReceived() int64 { + if m != nil { + return m.NumCallsFinishedKnownReceived + } + return 0 +} + +func (m *ClientStats) GetCallsFinishedWithDrop() []*ClientStatsPerToken { + if m != nil { + return m.CallsFinishedWithDrop + } + return nil +} + +type LoadBalanceResponse struct { + // Types that are valid to be assigned to LoadBalanceResponseType: + // *LoadBalanceResponse_InitialResponse + // *LoadBalanceResponse_ServerList + LoadBalanceResponseType isLoadBalanceResponse_LoadBalanceResponseType `protobuf_oneof:"load_balance_response_type"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *LoadBalanceResponse) Reset() { *m = LoadBalanceResponse{} } +func (m *LoadBalanceResponse) String() string { return proto.CompactTextString(m) } +func (*LoadBalanceResponse) ProtoMessage() {} +func (*LoadBalanceResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{4} +} +func (m *LoadBalanceResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_LoadBalanceResponse.Unmarshal(m, b) +} +func (m *LoadBalanceResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_LoadBalanceResponse.Marshal(b, m, deterministic) +} +func (dst *LoadBalanceResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_LoadBalanceResponse.Merge(dst, src) +} +func (m *LoadBalanceResponse) XXX_Size() int { + return xxx_messageInfo_LoadBalanceResponse.Size(m) +} +func (m *LoadBalanceResponse) XXX_DiscardUnknown() { + xxx_messageInfo_LoadBalanceResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_LoadBalanceResponse proto.InternalMessageInfo + +type isLoadBalanceResponse_LoadBalanceResponseType interface { + isLoadBalanceResponse_LoadBalanceResponseType() +} + +type LoadBalanceResponse_InitialResponse struct { + InitialResponse *InitialLoadBalanceResponse `protobuf:"bytes,1,opt,name=initial_response,json=initialResponse,proto3,oneof"` +} +type LoadBalanceResponse_ServerList struct { + ServerList *ServerList `protobuf:"bytes,2,opt,name=server_list,json=serverList,proto3,oneof"` +} + +func (*LoadBalanceResponse_InitialResponse) isLoadBalanceResponse_LoadBalanceResponseType() {} +func (*LoadBalanceResponse_ServerList) isLoadBalanceResponse_LoadBalanceResponseType() {} + +func (m *LoadBalanceResponse) GetLoadBalanceResponseType() isLoadBalanceResponse_LoadBalanceResponseType { + if m != nil { + return m.LoadBalanceResponseType + } + return nil +} + +func (m *LoadBalanceResponse) GetInitialResponse() *InitialLoadBalanceResponse { + if x, ok := m.GetLoadBalanceResponseType().(*LoadBalanceResponse_InitialResponse); ok { + return x.InitialResponse + } + return nil +} + +func (m *LoadBalanceResponse) GetServerList() *ServerList { + if x, ok := m.GetLoadBalanceResponseType().(*LoadBalanceResponse_ServerList); ok { + return x.ServerList + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*LoadBalanceResponse) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _LoadBalanceResponse_OneofMarshaler, _LoadBalanceResponse_OneofUnmarshaler, _LoadBalanceResponse_OneofSizer, []interface{}{ + (*LoadBalanceResponse_InitialResponse)(nil), + (*LoadBalanceResponse_ServerList)(nil), + } +} + +func _LoadBalanceResponse_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*LoadBalanceResponse) + // load_balance_response_type + switch x := m.LoadBalanceResponseType.(type) { + case *LoadBalanceResponse_InitialResponse: + b.EncodeVarint(1<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.InitialResponse); err != nil { + return err + } + case *LoadBalanceResponse_ServerList: + b.EncodeVarint(2<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.ServerList); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("LoadBalanceResponse.LoadBalanceResponseType has unexpected type %T", x) + } + return nil +} + +func _LoadBalanceResponse_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*LoadBalanceResponse) + switch tag { + case 1: // load_balance_response_type.initial_response + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(InitialLoadBalanceResponse) + err := b.DecodeMessage(msg) + m.LoadBalanceResponseType = &LoadBalanceResponse_InitialResponse{msg} + return true, err + case 2: // load_balance_response_type.server_list + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(ServerList) + err := b.DecodeMessage(msg) + m.LoadBalanceResponseType = &LoadBalanceResponse_ServerList{msg} + return true, err + default: + return false, nil + } +} + +func _LoadBalanceResponse_OneofSizer(msg proto.Message) (n int) { + m := msg.(*LoadBalanceResponse) + // load_balance_response_type + switch x := m.LoadBalanceResponseType.(type) { + case *LoadBalanceResponse_InitialResponse: + s := proto.Size(x.InitialResponse) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *LoadBalanceResponse_ServerList: + s := proto.Size(x.ServerList) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + +type InitialLoadBalanceResponse struct { + // This is an application layer redirect that indicates the client should use + // the specified server for load balancing. When this field is non-empty in + // the response, the client should open a separate connection to the + // load_balancer_delegate and call the BalanceLoad method. Its length should + // be less than 64 bytes. + LoadBalancerDelegate string `protobuf:"bytes,1,opt,name=load_balancer_delegate,json=loadBalancerDelegate,proto3" json:"load_balancer_delegate,omitempty"` + // This interval defines how often the client should send the client stats + // to the load balancer. Stats should only be reported when the duration is + // positive. + ClientStatsReportInterval *duration.Duration `protobuf:"bytes,2,opt,name=client_stats_report_interval,json=clientStatsReportInterval,proto3" json:"client_stats_report_interval,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *InitialLoadBalanceResponse) Reset() { *m = InitialLoadBalanceResponse{} } +func (m *InitialLoadBalanceResponse) String() string { return proto.CompactTextString(m) } +func (*InitialLoadBalanceResponse) ProtoMessage() {} +func (*InitialLoadBalanceResponse) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{5} +} +func (m *InitialLoadBalanceResponse) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_InitialLoadBalanceResponse.Unmarshal(m, b) +} +func (m *InitialLoadBalanceResponse) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_InitialLoadBalanceResponse.Marshal(b, m, deterministic) +} +func (dst *InitialLoadBalanceResponse) XXX_Merge(src proto.Message) { + xxx_messageInfo_InitialLoadBalanceResponse.Merge(dst, src) +} +func (m *InitialLoadBalanceResponse) XXX_Size() int { + return xxx_messageInfo_InitialLoadBalanceResponse.Size(m) +} +func (m *InitialLoadBalanceResponse) XXX_DiscardUnknown() { + xxx_messageInfo_InitialLoadBalanceResponse.DiscardUnknown(m) +} + +var xxx_messageInfo_InitialLoadBalanceResponse proto.InternalMessageInfo + +func (m *InitialLoadBalanceResponse) GetLoadBalancerDelegate() string { + if m != nil { + return m.LoadBalancerDelegate + } + return "" +} + +func (m *InitialLoadBalanceResponse) GetClientStatsReportInterval() *duration.Duration { + if m != nil { + return m.ClientStatsReportInterval + } + return nil +} + +type ServerList struct { + // Contains a list of servers selected by the load balancer. The list will + // be updated when server resolutions change or as needed to balance load + // across more servers. The client should consume the server list in order + // unless instructed otherwise via the client_config. + Servers []*Server `protobuf:"bytes,1,rep,name=servers,proto3" json:"servers,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ServerList) Reset() { *m = ServerList{} } +func (m *ServerList) String() string { return proto.CompactTextString(m) } +func (*ServerList) ProtoMessage() {} +func (*ServerList) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{6} +} +func (m *ServerList) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ServerList.Unmarshal(m, b) +} +func (m *ServerList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ServerList.Marshal(b, m, deterministic) +} +func (dst *ServerList) XXX_Merge(src proto.Message) { + xxx_messageInfo_ServerList.Merge(dst, src) +} +func (m *ServerList) XXX_Size() int { + return xxx_messageInfo_ServerList.Size(m) +} +func (m *ServerList) XXX_DiscardUnknown() { + xxx_messageInfo_ServerList.DiscardUnknown(m) +} + +var xxx_messageInfo_ServerList proto.InternalMessageInfo + +func (m *ServerList) GetServers() []*Server { + if m != nil { + return m.Servers + } + return nil +} + +// Contains server information. When the drop field is not true, use the other +// fields. +type Server struct { + // A resolved address for the server, serialized in network-byte-order. It may + // either be an IPv4 or IPv6 address. + IpAddress []byte `protobuf:"bytes,1,opt,name=ip_address,json=ipAddress,proto3" json:"ip_address,omitempty"` + // A resolved port number for the server. + Port int32 `protobuf:"varint,2,opt,name=port,proto3" json:"port,omitempty"` + // An opaque but printable token given to the frontend for each pick. All + // frontend requests for that pick must include the token in its initial + // metadata. The token is used by the backend to verify the request and to + // allow the backend to report load to the gRPC LB system. The token is also + // used in client stats for reporting dropped calls. + LoadBalanceToken string `protobuf:"bytes,3,opt,name=load_balance_token,json=loadBalanceToken,proto3" json:"load_balance_token,omitempty"` + // Indicates whether this particular request should be dropped by the client. + // If the request is dropped, there will be a corresponding entry in + // ClientStats.calls_finished_with_drop. + Drop bool `protobuf:"varint,4,opt,name=drop,proto3" json:"drop,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *Server) Reset() { *m = Server{} } +func (m *Server) String() string { return proto.CompactTextString(m) } +func (*Server) ProtoMessage() {} +func (*Server) Descriptor() ([]byte, []int) { + return fileDescriptor_load_balancer_077502ef49b0d1b1, []int{7} +} +func (m *Server) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Server.Unmarshal(m, b) +} +func (m *Server) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Server.Marshal(b, m, deterministic) +} +func (dst *Server) XXX_Merge(src proto.Message) { + xxx_messageInfo_Server.Merge(dst, src) +} +func (m *Server) XXX_Size() int { + return xxx_messageInfo_Server.Size(m) +} +func (m *Server) XXX_DiscardUnknown() { + xxx_messageInfo_Server.DiscardUnknown(m) +} + +var xxx_messageInfo_Server proto.InternalMessageInfo + +func (m *Server) GetIpAddress() []byte { + if m != nil { + return m.IpAddress + } + return nil +} + +func (m *Server) GetPort() int32 { + if m != nil { + return m.Port + } + return 0 +} + +func (m *Server) GetLoadBalanceToken() string { + if m != nil { + return m.LoadBalanceToken + } + return "" +} + +func (m *Server) GetDrop() bool { + if m != nil { + return m.Drop + } + return false +} + +func init() { + proto.RegisterType((*LoadBalanceRequest)(nil), "grpc.lb.v1.LoadBalanceRequest") + proto.RegisterType((*InitialLoadBalanceRequest)(nil), "grpc.lb.v1.InitialLoadBalanceRequest") + proto.RegisterType((*ClientStatsPerToken)(nil), "grpc.lb.v1.ClientStatsPerToken") + proto.RegisterType((*ClientStats)(nil), "grpc.lb.v1.ClientStats") + proto.RegisterType((*LoadBalanceResponse)(nil), "grpc.lb.v1.LoadBalanceResponse") + proto.RegisterType((*InitialLoadBalanceResponse)(nil), "grpc.lb.v1.InitialLoadBalanceResponse") + proto.RegisterType((*ServerList)(nil), "grpc.lb.v1.ServerList") + proto.RegisterType((*Server)(nil), "grpc.lb.v1.Server") +} + +// Reference imports to suppress errors if they are not otherwise used. +var _ context.Context +var _ grpc.ClientConn + +// This is a compile-time assertion to ensure that this generated file +// is compatible with the grpc package it is being compiled against. +const _ = grpc.SupportPackageIsVersion4 + +// LoadBalancerClient is the client API for LoadBalancer service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. +type LoadBalancerClient interface { + // Bidirectional rpc to get a list of servers. + BalanceLoad(ctx context.Context, opts ...grpc.CallOption) (LoadBalancer_BalanceLoadClient, error) +} + +type loadBalancerClient struct { + cc *grpc.ClientConn +} + +func NewLoadBalancerClient(cc *grpc.ClientConn) LoadBalancerClient { + return &loadBalancerClient{cc} +} + +func (c *loadBalancerClient) BalanceLoad(ctx context.Context, opts ...grpc.CallOption) (LoadBalancer_BalanceLoadClient, error) { + stream, err := c.cc.NewStream(ctx, &_LoadBalancer_serviceDesc.Streams[0], "/grpc.lb.v1.LoadBalancer/BalanceLoad", opts...) + if err != nil { + return nil, err + } + x := &loadBalancerBalanceLoadClient{stream} + return x, nil +} + +type LoadBalancer_BalanceLoadClient interface { + Send(*LoadBalanceRequest) error + Recv() (*LoadBalanceResponse, error) + grpc.ClientStream +} + +type loadBalancerBalanceLoadClient struct { + grpc.ClientStream +} + +func (x *loadBalancerBalanceLoadClient) Send(m *LoadBalanceRequest) error { + return x.ClientStream.SendMsg(m) +} + +func (x *loadBalancerBalanceLoadClient) Recv() (*LoadBalanceResponse, error) { + m := new(LoadBalanceResponse) + if err := x.ClientStream.RecvMsg(m); err != nil { + return nil, err + } + return m, nil +} + +// LoadBalancerServer is the server API for LoadBalancer service. +type LoadBalancerServer interface { + // Bidirectional rpc to get a list of servers. + BalanceLoad(LoadBalancer_BalanceLoadServer) error +} + +func RegisterLoadBalancerServer(s *grpc.Server, srv LoadBalancerServer) { + s.RegisterService(&_LoadBalancer_serviceDesc, srv) +} + +func _LoadBalancer_BalanceLoad_Handler(srv interface{}, stream grpc.ServerStream) error { + return srv.(LoadBalancerServer).BalanceLoad(&loadBalancerBalanceLoadServer{stream}) +} + +type LoadBalancer_BalanceLoadServer interface { + Send(*LoadBalanceResponse) error + Recv() (*LoadBalanceRequest, error) + grpc.ServerStream +} + +type loadBalancerBalanceLoadServer struct { + grpc.ServerStream +} + +func (x *loadBalancerBalanceLoadServer) Send(m *LoadBalanceResponse) error { + return x.ServerStream.SendMsg(m) +} + +func (x *loadBalancerBalanceLoadServer) Recv() (*LoadBalanceRequest, error) { + m := new(LoadBalanceRequest) + if err := x.ServerStream.RecvMsg(m); err != nil { + return nil, err + } + return m, nil +} + +var _LoadBalancer_serviceDesc = grpc.ServiceDesc{ + ServiceName: "grpc.lb.v1.LoadBalancer", + HandlerType: (*LoadBalancerServer)(nil), + Methods: []grpc.MethodDesc{}, + Streams: []grpc.StreamDesc{ + { + StreamName: "BalanceLoad", + Handler: _LoadBalancer_BalanceLoad_Handler, + ServerStreams: true, + ClientStreams: true, + }, + }, + Metadata: "grpc/lb/v1/load_balancer.proto", +} + +func init() { + proto.RegisterFile("grpc/lb/v1/load_balancer.proto", fileDescriptor_load_balancer_077502ef49b0d1b1) +} + +var fileDescriptor_load_balancer_077502ef49b0d1b1 = []byte{ + // 756 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x55, 0xdd, 0x6e, 0x23, 0x35, + 0x14, 0xee, 0x90, 0x69, 0x36, 0x39, 0xa9, 0xb6, 0x59, 0x2f, 0x2c, 0x93, 0xec, 0x6e, 0x5b, 0x22, + 0xb1, 0x8a, 0xd0, 0x32, 0x43, 0x0a, 0x17, 0x20, 0x71, 0x01, 0xd9, 0x6a, 0x95, 0x2d, 0xbd, 0x88, + 0x9c, 0x4a, 0x54, 0x95, 0x90, 0x99, 0xc9, 0xb8, 0xa9, 0x55, 0xc7, 0x1e, 0x3c, 0x4e, 0x2a, 0xae, + 0x79, 0x1f, 0xc4, 0x2b, 0x20, 0x5e, 0x0c, 0x8d, 0xed, 0x49, 0xa6, 0x49, 0xa3, 0xbd, 0x69, 0x3d, + 0xe7, 0x7c, 0xfe, 0xce, 0xef, 0xe7, 0xc0, 0xd1, 0x4c, 0x65, 0xd3, 0x88, 0x27, 0xd1, 0x72, 0x10, + 0x71, 0x19, 0xa7, 0x24, 0x89, 0x79, 0x2c, 0xa6, 0x54, 0x85, 0x99, 0x92, 0x5a, 0x22, 0x28, 0xfc, + 0x21, 0x4f, 0xc2, 0xe5, 0xa0, 0x7b, 0x34, 0x93, 0x72, 0xc6, 0x69, 0x64, 0x3c, 0xc9, 0xe2, 0x26, + 0x4a, 0x17, 0x2a, 0xd6, 0x4c, 0x0a, 0x8b, 0xed, 0x1e, 0x6f, 0xfa, 0x35, 0x9b, 0xd3, 0x5c, 0xc7, + 0xf3, 0xcc, 0x02, 0x7a, 0xff, 0x7a, 0x80, 0x2e, 0x64, 0x9c, 0x0e, 0x6d, 0x0c, 0x4c, 0xff, 0x58, + 0xd0, 0x5c, 0xa3, 0x31, 0x1c, 0x32, 0xc1, 0x34, 0x8b, 0x39, 0x51, 0xd6, 0x14, 0x78, 0x27, 0x5e, + 0xbf, 0x75, 0xfa, 0x65, 0xb8, 0x8e, 0x1e, 0x7e, 0xb0, 0x90, 0xed, 0xfb, 0xa3, 0x3d, 0xfc, 0xd4, + 0xdd, 0x2f, 0x19, 0x7f, 0x84, 0x83, 0x29, 0x67, 0x54, 0x68, 0x92, 0xeb, 0x58, 0xe7, 0xc1, 0x27, + 0x86, 0xee, 0xf3, 0x2a, 0xdd, 0x3b, 0xe3, 0x9f, 0x14, 0xee, 0xd1, 0x1e, 0x6e, 0x4d, 0xd7, 0x9f, + 0xc3, 0x97, 0xd0, 0xa9, 0xb6, 0xa2, 0x4c, 0x8a, 0xe8, 0x3f, 0x33, 0xda, 0x8b, 0xa0, 0xb3, 0x33, + 0x13, 0x84, 0xc0, 0x17, 0xf1, 0x9c, 0x9a, 0xf4, 0x9b, 0xd8, 0x9c, 0x7b, 0xbf, 0xc3, 0xf3, 0x4a, + 0xac, 0x31, 0x55, 0x97, 0xf2, 0x8e, 0x0a, 0xf4, 0x16, 0xd0, 0x83, 0x20, 0xba, 0xb0, 0xba, 0x8b, + 0x6d, 0xbe, 0xa6, 0xb6, 0xe8, 0x97, 0xd0, 0x14, 0x8b, 0x39, 0x99, 0xc6, 0x9c, 0xdb, 0x6a, 0x6a, + 0xb8, 0x21, 0x16, 0xf3, 0x77, 0xc5, 0x77, 0xef, 0x9f, 0x1a, 0xb4, 0x2a, 0x21, 0xd0, 0xf7, 0xd0, + 0x5c, 0x75, 0xde, 0x75, 0xb2, 0x1b, 0xda, 0xd9, 0x84, 0xe5, 0x6c, 0xc2, 0xcb, 0x12, 0x81, 0xd7, + 0x60, 0xf4, 0x15, 0x3c, 0x5b, 0x85, 0x29, 0x5a, 0xa7, 0x34, 0x4d, 0x5d, 0xb8, 0xc3, 0x32, 0xdc, + 0xc4, 0x9a, 0x8b, 0x02, 0xd6, 0xd8, 0x1b, 0x26, 0x58, 0x7e, 0x4b, 0xd3, 0xa0, 0x66, 0xc0, 0xed, + 0x12, 0xfc, 0xde, 0xd9, 0xd1, 0x6f, 0xf0, 0xf5, 0x36, 0x9a, 0xdc, 0x33, 0x7d, 0x4b, 0xdc, 0xa4, + 0x6e, 0x62, 0xc6, 0x69, 0x4a, 0xb4, 0x24, 0x39, 0x15, 0x69, 0x50, 0x37, 0x44, 0x6f, 0x36, 0x89, + 0x7e, 0x65, 0xfa, 0xd6, 0xd6, 0xfa, 0xde, 0xe0, 0x2f, 0xe5, 0x84, 0x8a, 0x14, 0x8d, 0xe0, 0x8b, + 0x47, 0xe8, 0xef, 0x84, 0xbc, 0x17, 0x44, 0xd1, 0x29, 0x65, 0x4b, 0x9a, 0x06, 0x4f, 0x0c, 0xe5, + 0xeb, 0x4d, 0xca, 0x5f, 0x0a, 0x14, 0x76, 0x20, 0x74, 0x05, 0xc1, 0x63, 0x49, 0xa6, 0x4a, 0x66, + 0x41, 0xe3, 0xa4, 0xd6, 0x6f, 0x9d, 0x1e, 0xef, 0x58, 0xa3, 0x72, 0xb4, 0xf8, 0xb3, 0xe9, 0x66, + 0xc6, 0x67, 0x4a, 0x66, 0xe7, 0x7e, 0xc3, 0x6f, 0xef, 0x9f, 0xfb, 0x8d, 0xfd, 0x76, 0xbd, 0xf7, + 0x9f, 0x07, 0xcf, 0x1f, 0xec, 0x4f, 0x9e, 0x49, 0x91, 0x53, 0x34, 0x81, 0xf6, 0x5a, 0x0a, 0xd6, + 0xe6, 0x26, 0xf8, 0xe6, 0x63, 0x5a, 0xb0, 0xe8, 0xd1, 0x1e, 0x3e, 0x5c, 0x89, 0xc1, 0x91, 0xfe, + 0x00, 0xad, 0x9c, 0xaa, 0x25, 0x55, 0x84, 0xb3, 0x5c, 0x3b, 0x31, 0xbc, 0xa8, 0xf2, 0x4d, 0x8c, + 0xfb, 0x82, 0x19, 0x31, 0x41, 0xbe, 0xfa, 0x1a, 0xbe, 0x82, 0xee, 0x86, 0x14, 0x2c, 0xa7, 0xd5, + 0xc2, 0xdf, 0x1e, 0x74, 0x77, 0xa7, 0x82, 0xbe, 0x83, 0x17, 0x0f, 0x9e, 0x14, 0x92, 0x52, 0x4e, + 0x67, 0xb1, 0x2e, 0xf5, 0xf1, 0x69, 0x65, 0xcd, 0xd5, 0x99, 0xf3, 0xa1, 0x6b, 0x78, 0x55, 0xd5, + 0x2e, 0x51, 0x34, 0x93, 0x4a, 0x13, 0x26, 0x34, 0x55, 0xcb, 0x98, 0xbb, 0xf4, 0x3b, 0x5b, 0x0b, + 0x7d, 0xe6, 0x1e, 0x23, 0xdc, 0xa9, 0x68, 0x19, 0x9b, 0xcb, 0x1f, 0xdc, 0xdd, 0xde, 0x4f, 0x00, + 0xeb, 0x52, 0xd1, 0x5b, 0x78, 0x62, 0x4b, 0xcd, 0x03, 0xcf, 0x4c, 0x16, 0x6d, 0xf7, 0x04, 0x97, + 0x90, 0x73, 0xbf, 0x51, 0x6b, 0xfb, 0xbd, 0xbf, 0x3c, 0xa8, 0x5b, 0x0f, 0x7a, 0x0d, 0xc0, 0x32, + 0x12, 0xa7, 0xa9, 0xa2, 0x79, 0x6e, 0x4a, 0x3a, 0xc0, 0x4d, 0x96, 0xfd, 0x6c, 0x0d, 0xc5, 0x5b, + 0x50, 0xc4, 0x36, 0xf9, 0xee, 0x63, 0x73, 0xde, 0x21, 0xfa, 0xda, 0x0e, 0xd1, 0x23, 0xf0, 0xcd, + 0xda, 0xf9, 0x27, 0x5e, 0xbf, 0x81, 0xcd, 0xd9, 0xae, 0xcf, 0x69, 0x02, 0x07, 0x95, 0x86, 0x2b, + 0x84, 0xa1, 0xe5, 0xce, 0x85, 0x19, 0x1d, 0x55, 0xeb, 0xd8, 0x7e, 0xa6, 0xba, 0xc7, 0x3b, 0xfd, + 0x76, 0x72, 0x7d, 0xef, 0x1b, 0x6f, 0x78, 0x05, 0x4f, 0x99, 0xb4, 0xc0, 0xe2, 0x0f, 0x4f, 0x86, + 0xcf, 0xaa, 0x31, 0xc7, 0x45, 0xdf, 0xc7, 0xde, 0xf5, 0xc0, 0xcd, 0x61, 0x26, 0x79, 0x2c, 0x66, + 0xa1, 0x54, 0xb3, 0xc8, 0xfc, 0xa4, 0x94, 0x43, 0x8f, 0xec, 0x65, 0xf3, 0x8f, 0xf0, 0x84, 0x2c, + 0x07, 0x49, 0xdd, 0xcc, 0xec, 0xdb, 0xff, 0x03, 0x00, 0x00, 0xff, 0xff, 0x5f, 0xc0, 0x96, 0xfd, + 0x7c, 0x06, 0x00, 0x00, +} diff --git a/vendor/google.golang.org/grpc/grpclb.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb.go similarity index 86% rename from vendor/google.golang.org/grpc/grpclb.go rename to vendor/google.golang.org/grpc/balancer/grpclb/grpclb.go index bc2b44525..0c46ae32f 100644 --- a/vendor/google.golang.org/grpc/grpclb.go +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb.go @@ -16,19 +16,29 @@ * */ -package grpc +//go:generate ./regenerate.sh + +// Package grpclb defines a grpclb balancer. +// +// To install grpclb balancer, import this package as: +// import _ "google.golang.org/grpc/balancer/grpclb" +package grpclb import ( + "fmt" "strconv" "strings" "sync" "time" + durationpb "github.com/golang/protobuf/ptypes/duration" "golang.org/x/net/context" + "google.golang.org/grpc" "google.golang.org/grpc/balancer" + lbpb "google.golang.org/grpc/balancer/grpclb/grpc_lb_v1" "google.golang.org/grpc/connectivity" - lbpb "google.golang.org/grpc/grpclb/grpc_lb_v1/messages" "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal/backoff" "google.golang.org/grpc/resolver" ) @@ -38,7 +48,21 @@ const ( grpclbName = "grpclb" ) -func convertDuration(d *lbpb.Duration) time.Duration { +var ( + // defaultBackoffConfig configures the backoff strategy that's used when the + // init handshake in the RPC is unsuccessful. It's not for the clientconn + // reconnect backoff. + // + // It has the same value as the default grpc.DefaultBackoffConfig. + // + // TODO: make backoff configurable. + defaultBackoffConfig = backoff.Exponential{ + MaxDelay: 120 * time.Second, + } + errServerTerminatedConnection = fmt.Errorf("grpclb: failed to recv server list: server terminated connection") +) + +func convertDuration(d *durationpb.Duration) time.Duration { if d == nil { return 0 } @@ -49,11 +73,11 @@ func convertDuration(d *lbpb.Duration) time.Duration { // Mostly copied from generated pb.go file. // To avoid circular dependency. type loadBalancerClient struct { - cc *ClientConn + cc *grpc.ClientConn } -func (c *loadBalancerClient) BalanceLoad(ctx context.Context, opts ...CallOption) (*balanceLoadClientStream, error) { - desc := &StreamDesc{ +func (c *loadBalancerClient) BalanceLoad(ctx context.Context, opts ...grpc.CallOption) (*balanceLoadClientStream, error) { + desc := &grpc.StreamDesc{ StreamName: "BalanceLoad", ServerStreams: true, ClientStreams: true, @@ -67,7 +91,7 @@ func (c *loadBalancerClient) BalanceLoad(ctx context.Context, opts ...CallOption } type balanceLoadClientStream struct { - ClientStream + grpc.ClientStream } func (x *balanceLoadClientStream) Send(m *lbpb.LoadBalanceRequest) error { @@ -88,16 +112,16 @@ func init() { // newLBBuilder creates a builder for grpclb. func newLBBuilder() balancer.Builder { - return NewLBBuilderWithFallbackTimeout(defaultFallbackTimeout) + return newLBBuilderWithFallbackTimeout(defaultFallbackTimeout) } -// NewLBBuilderWithFallbackTimeout creates a grpclb builder with the given +// newLBBuilderWithFallbackTimeout creates a grpclb builder with the given // fallbackTimeout. If no response is received from the remote balancer within // fallbackTimeout, the backend addresses from the resolved address list will be // used. // // Only call this function when a non-default fallback timeout is needed. -func NewLBBuilderWithFallbackTimeout(fallbackTimeout time.Duration) balancer.Builder { +func newLBBuilderWithFallbackTimeout(fallbackTimeout time.Duration) balancer.Builder { return &lbBuilder{ fallbackTimeout: fallbackTimeout, } @@ -134,11 +158,12 @@ func (b *lbBuilder) Build(cc balancer.ClientConn, opt balancer.BuildOptions) bal doneCh: make(chan struct{}), manualResolver: r, - csEvltr: &connectivityStateEvaluator{}, + csEvltr: &balancer.ConnectivityStateEvaluator{}, subConns: make(map[resolver.Address]balancer.SubConn), scStates: make(map[balancer.SubConn]connectivity.State), picker: &errPicker{err: balancer.ErrNoSubConnAvailable}, - clientStats: &rpcStats{}, + clientStats: newRPCStats(), + backoff: defaultBackoffConfig, // TODO: make backoff configurable. } return lb @@ -156,7 +181,9 @@ type lbBalancer struct { // send to remote LB ClientConn through this resolver. manualResolver *lbManualResolver // The ClientConn to talk to the remote balancer. - ccRemoteLB *ClientConn + ccRemoteLB *grpc.ClientConn + // backoff for calling remote balancer. + backoff backoff.Strategy // Support client side load reporting. Each picker gets a reference to this, // and will update its content. @@ -173,7 +200,7 @@ type lbBalancer struct { // but with only READY SCs will be gerenated. backendAddrs []resolver.Address // Roundrobin functionalities. - csEvltr *connectivityStateEvaluator + csEvltr *balancer.ConnectivityStateEvaluator state connectivity.State subConns map[resolver.Address]balancer.SubConn // Used to new/remove SubConn. scStates map[balancer.SubConn]connectivity.State // Used to filter READY SubConns. @@ -243,7 +270,7 @@ func (lb *lbBalancer) HandleSubConnStateChange(sc balancer.SubConn, s connectivi } oldAggrState := lb.state - lb.state = lb.csEvltr.recordTransition(oldS, s) + lb.state = lb.csEvltr.RecordTransition(oldS, s) // Regenerate picker when one of the following happens: // - this sc became ready from not-ready diff --git a/vendor/google.golang.org/grpc/grpclb_picker.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_picker.go similarity index 65% rename from vendor/google.golang.org/grpc/grpclb_picker.go rename to vendor/google.golang.org/grpc/balancer/grpclb/grpclb_picker.go index 872c7ccea..d77af401b 100644 --- a/vendor/google.golang.org/grpc/grpclb_picker.go +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_picker.go @@ -16,7 +16,7 @@ * */ -package grpc +package grpclb import ( "sync" @@ -24,55 +24,70 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/balancer" + lbpb "google.golang.org/grpc/balancer/grpclb/grpc_lb_v1" "google.golang.org/grpc/codes" - lbpb "google.golang.org/grpc/grpclb/grpc_lb_v1/messages" "google.golang.org/grpc/status" ) +// rpcStats is same as lbmpb.ClientStats, except that numCallsDropped is a map +// instead of a slice. type rpcStats struct { - NumCallsStarted int64 - NumCallsFinished int64 - NumCallsFinishedWithDropForRateLimiting int64 - NumCallsFinishedWithDropForLoadBalancing int64 - NumCallsFinishedWithClientFailedToSend int64 - NumCallsFinishedKnownReceived int64 + // Only access the following fields atomically. + numCallsStarted int64 + numCallsFinished int64 + numCallsFinishedWithClientFailedToSend int64 + numCallsFinishedKnownReceived int64 + + mu sync.Mutex + // map load_balance_token -> num_calls_dropped + numCallsDropped map[string]int64 +} + +func newRPCStats() *rpcStats { + return &rpcStats{ + numCallsDropped: make(map[string]int64), + } } // toClientStats converts rpcStats to lbpb.ClientStats, and clears rpcStats. func (s *rpcStats) toClientStats() *lbpb.ClientStats { stats := &lbpb.ClientStats{ - NumCallsStarted: atomic.SwapInt64(&s.NumCallsStarted, 0), - NumCallsFinished: atomic.SwapInt64(&s.NumCallsFinished, 0), - NumCallsFinishedWithDropForRateLimiting: atomic.SwapInt64(&s.NumCallsFinishedWithDropForRateLimiting, 0), - NumCallsFinishedWithDropForLoadBalancing: atomic.SwapInt64(&s.NumCallsFinishedWithDropForLoadBalancing, 0), - NumCallsFinishedWithClientFailedToSend: atomic.SwapInt64(&s.NumCallsFinishedWithClientFailedToSend, 0), - NumCallsFinishedKnownReceived: atomic.SwapInt64(&s.NumCallsFinishedKnownReceived, 0), + NumCallsStarted: atomic.SwapInt64(&s.numCallsStarted, 0), + NumCallsFinished: atomic.SwapInt64(&s.numCallsFinished, 0), + NumCallsFinishedWithClientFailedToSend: atomic.SwapInt64(&s.numCallsFinishedWithClientFailedToSend, 0), + NumCallsFinishedKnownReceived: atomic.SwapInt64(&s.numCallsFinishedKnownReceived, 0), + } + s.mu.Lock() + dropped := s.numCallsDropped + s.numCallsDropped = make(map[string]int64) + s.mu.Unlock() + for token, count := range dropped { + stats.CallsFinishedWithDrop = append(stats.CallsFinishedWithDrop, &lbpb.ClientStatsPerToken{ + LoadBalanceToken: token, + NumCalls: count, + }) } return stats } -func (s *rpcStats) dropForRateLimiting() { - atomic.AddInt64(&s.NumCallsStarted, 1) - atomic.AddInt64(&s.NumCallsFinishedWithDropForRateLimiting, 1) - atomic.AddInt64(&s.NumCallsFinished, 1) -} - -func (s *rpcStats) dropForLoadBalancing() { - atomic.AddInt64(&s.NumCallsStarted, 1) - atomic.AddInt64(&s.NumCallsFinishedWithDropForLoadBalancing, 1) - atomic.AddInt64(&s.NumCallsFinished, 1) +func (s *rpcStats) drop(token string) { + atomic.AddInt64(&s.numCallsStarted, 1) + s.mu.Lock() + s.numCallsDropped[token]++ + s.mu.Unlock() + atomic.AddInt64(&s.numCallsFinished, 1) } func (s *rpcStats) failedToSend() { - atomic.AddInt64(&s.NumCallsStarted, 1) - atomic.AddInt64(&s.NumCallsFinishedWithClientFailedToSend, 1) - atomic.AddInt64(&s.NumCallsFinished, 1) + atomic.AddInt64(&s.numCallsStarted, 1) + atomic.AddInt64(&s.numCallsFinishedWithClientFailedToSend, 1) + atomic.AddInt64(&s.numCallsFinished, 1) } func (s *rpcStats) knownReceived() { - atomic.AddInt64(&s.NumCallsStarted, 1) - atomic.AddInt64(&s.NumCallsFinishedKnownReceived, 1) - atomic.AddInt64(&s.NumCallsFinished, 1) + atomic.AddInt64(&s.numCallsStarted, 1) + atomic.AddInt64(&s.numCallsFinishedKnownReceived, 1) + atomic.AddInt64(&s.numCallsFinished, 1) } type errPicker struct { @@ -131,12 +146,8 @@ func (p *lbPicker) Pick(ctx context.Context, opts balancer.PickOptions) (balance p.serverListNext = (p.serverListNext + 1) % len(p.serverList) // If it's a drop, return an error and fail the RPC. - if s.DropForRateLimiting { - p.stats.dropForRateLimiting() - return nil, nil, status.Errorf(codes.Unavailable, "request dropped by grpclb") - } - if s.DropForLoadBalancing { - p.stats.dropForLoadBalancing() + if s.Drop { + p.stats.drop(s.LoadBalanceToken) return nil, nil, status.Errorf(codes.Unavailable, "request dropped by grpclb") } diff --git a/vendor/google.golang.org/grpc/grpclb_remote_balancer.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_remote_balancer.go similarity index 74% rename from vendor/google.golang.org/grpc/grpclb_remote_balancer.go rename to vendor/google.golang.org/grpc/balancer/grpclb/grpclb_remote_balancer.go index b8dd4f18c..d0ecd68be 100644 --- a/vendor/google.golang.org/grpc/grpclb_remote_balancer.go +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_remote_balancer.go @@ -16,21 +16,24 @@ * */ -package grpc +package grpclb import ( "fmt" + "io" "net" "reflect" "time" + timestamppb "github.com/golang/protobuf/ptypes/timestamp" "golang.org/x/net/context" + "google.golang.org/grpc" "google.golang.org/grpc/balancer" - "google.golang.org/grpc/channelz" - + lbpb "google.golang.org/grpc/balancer/grpclb/grpc_lb_v1" "google.golang.org/grpc/connectivity" - lbpb "google.golang.org/grpc/grpclb/grpc_lb_v1/messages" "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal" + "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/metadata" "google.golang.org/grpc/resolver" ) @@ -54,8 +57,8 @@ func (lb *lbBalancer) processServerList(l *lbpb.ServerList) { lb.fullServerList = l.Servers var backendAddrs []resolver.Address - for _, s := range l.Servers { - if s.DropForLoadBalancing || s.DropForRateLimiting { + for i, s := range l.Servers { + if s.Drop { continue } @@ -71,7 +74,8 @@ func (lb *lbBalancer) processServerList(l *lbpb.ServerList) { Addr: fmt.Sprintf("%s:%d", ipStr, s.Port), Metadata: &md, } - + grpclog.Infof("lbBalancer: server list entry[%d]: ipStr:|%s|, port:|%d|, load balancer token:|%v|", + i, ipStr, s.Port, s.LoadBalanceToken) backendAddrs = append(backendAddrs, addr) } @@ -143,6 +147,9 @@ func (lb *lbBalancer) readServerList(s *balanceLoadClientStream) error { for { reply, err := s.Recv() if err != nil { + if err == io.EOF { + return errServerTerminatedConnection + } return fmt.Errorf("grpclb: failed to recv server list: %v", err) } if serverList := reply.GetServerList(); serverList != nil { @@ -162,7 +169,7 @@ func (lb *lbBalancer) sendLoadReport(s *balanceLoadClientStream, interval time.D } stats := lb.clientStats.toClientStats() t := time.Now() - stats.Timestamp = &lbpb.Timestamp{ + stats.Timestamp = ×tamppb.Timestamp{ Seconds: t.Unix(), Nanos: int32(t.Nanosecond()), } @@ -176,13 +183,13 @@ func (lb *lbBalancer) sendLoadReport(s *balanceLoadClientStream, interval time.D } } -func (lb *lbBalancer) callRemoteBalancer() error { +func (lb *lbBalancer) callRemoteBalancer() (backoff bool, _ error) { lbClient := &loadBalancerClient{cc: lb.ccRemoteLB} ctx, cancel := context.WithCancel(context.Background()) defer cancel() - stream, err := lbClient.BalanceLoad(ctx, FailFast(false)) + stream, err := lbClient.BalanceLoad(ctx, grpc.FailFast(false)) if err != nil { - return fmt.Errorf("grpclb: failed to perform RPC to the remote balancer %v", err) + return true, fmt.Errorf("grpclb: failed to perform RPC to the remote balancer %v", err) } // grpclb handshake on the stream. @@ -194,18 +201,18 @@ func (lb *lbBalancer) callRemoteBalancer() error { }, } if err := stream.Send(initReq); err != nil { - return fmt.Errorf("grpclb: failed to send init request: %v", err) + return true, fmt.Errorf("grpclb: failed to send init request: %v", err) } reply, err := stream.Recv() if err != nil { - return fmt.Errorf("grpclb: failed to recv init response: %v", err) + return true, fmt.Errorf("grpclb: failed to recv init response: %v", err) } initResp := reply.GetInitialResponse() if initResp == nil { - return fmt.Errorf("grpclb: reply from remote balancer did not include initial response") + return true, fmt.Errorf("grpclb: reply from remote balancer did not include initial response") } if initResp.LoadBalancerDelegate != "" { - return fmt.Errorf("grpclb: Delegation is not supported") + return true, fmt.Errorf("grpclb: Delegation is not supported") } go func() { @@ -213,51 +220,72 @@ func (lb *lbBalancer) callRemoteBalancer() error { lb.sendLoadReport(stream, d) } }() - return lb.readServerList(stream) + // No backoff if init req/resp handshake was successful. + return false, lb.readServerList(stream) } func (lb *lbBalancer) watchRemoteBalancer() { + var retryCount int for { - err := lb.callRemoteBalancer() + doBackoff, err := lb.callRemoteBalancer() select { case <-lb.doneCh: return default: if err != nil { - grpclog.Error(err) + if err == errServerTerminatedConnection { + grpclog.Info(err) + } else { + grpclog.Error(err) + } } } + if !doBackoff { + retryCount = 0 + continue + } + + timer := time.NewTimer(lb.backoff.Backoff(retryCount)) + select { + case <-timer.C: + case <-lb.doneCh: + timer.Stop() + return + } + retryCount++ } } func (lb *lbBalancer) dialRemoteLB(remoteLBName string) { - var dopts []DialOption + var dopts []grpc.DialOption if creds := lb.opt.DialCreds; creds != nil { if err := creds.OverrideServerName(remoteLBName); err == nil { - dopts = append(dopts, WithTransportCredentials(creds)) + dopts = append(dopts, grpc.WithTransportCredentials(creds)) } else { grpclog.Warningf("grpclb: failed to override the server name in the credentials: %v, using Insecure", err) - dopts = append(dopts, WithInsecure()) + dopts = append(dopts, grpc.WithInsecure()) } } else { - dopts = append(dopts, WithInsecure()) + dopts = append(dopts, grpc.WithInsecure()) } if lb.opt.Dialer != nil { // WithDialer takes a different type of function, so we instead use a // special DialOption here. - dopts = append(dopts, withContextDialer(lb.opt.Dialer)) + wcd := internal.WithContextDialer.(func(func(context.Context, string) (net.Conn, error)) grpc.DialOption) + dopts = append(dopts, wcd(lb.opt.Dialer)) } // Explicitly set pickfirst as the balancer. - dopts = append(dopts, WithBalancerName(PickFirstBalancerName)) - dopts = append(dopts, withResolverBuilder(lb.manualResolver)) + dopts = append(dopts, grpc.WithBalancerName(grpc.PickFirstBalancerName)) + wrb := internal.WithResolverBuilder.(func(resolver.Builder) grpc.DialOption) + dopts = append(dopts, wrb(lb.manualResolver)) if channelz.IsOn() { - dopts = append(dopts, WithChannelzParentID(lb.opt.ChannelzParentID)) + dopts = append(dopts, grpc.WithChannelzParentID(lb.opt.ChannelzParentID)) } // DialContext using manualResolver.Scheme, which is a random scheme generated // when init grpclb. The target name is not important. - cc, err := DialContext(context.Background(), "grpclb:///grpclb.server", dopts...) + cc, err := grpc.DialContext(context.Background(), "grpclb:///grpclb.server", dopts...) if err != nil { grpclog.Fatalf("failed to dial: %v", err) } diff --git a/vendor/google.golang.org/grpc/grpclb/grpclb_test.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_test.go similarity index 75% rename from vendor/google.golang.org/grpc/grpclb/grpclb_test.go rename to vendor/google.golang.org/grpc/balancer/grpclb/grpclb_test.go index a62d5903b..7e6ca76ed 100644 --- a/vendor/google.golang.org/grpc/grpclb/grpclb_test.go +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_test.go @@ -16,11 +16,7 @@ * */ -//go:generate protoc --go_out=plugins=:$GOPATH/src grpc_lb_v1/messages/messages.proto -//go:generate protoc --go_out=plugins=grpc:$GOPATH/src grpc_lb_v1/service/service.proto - -// Package grpclb_test is currently used only for grpclb testing. -package grpclb_test +package grpclb import ( "errors" @@ -30,25 +26,26 @@ import ( "strconv" "strings" "sync" + "sync/atomic" "testing" "time" - "github.com/golang/protobuf/proto" + durationpb "github.com/golang/protobuf/ptypes/duration" "golang.org/x/net/context" "google.golang.org/grpc" "google.golang.org/grpc/balancer" + lbgrpc "google.golang.org/grpc/balancer/grpclb/grpc_lb_v1" + lbpb "google.golang.org/grpc/balancer/grpclb/grpc_lb_v1" "google.golang.org/grpc/codes" "google.golang.org/grpc/credentials" - lbmpb "google.golang.org/grpc/grpclb/grpc_lb_v1/messages" - lbspb "google.golang.org/grpc/grpclb/grpc_lb_v1/service" _ "google.golang.org/grpc/grpclog/glogger" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" "google.golang.org/grpc/status" testpb "google.golang.org/grpc/test/grpc_testing" - "google.golang.org/grpc/test/leakcheck" ) var ( @@ -125,18 +122,75 @@ func fakeNameDialer(addr string, timeout time.Duration) (net.Conn, error) { return net.DialTimeout("tcp", addr, timeout) } +// merge merges the new client stats into current stats. +// +// It's a test-only method. rpcStats is defined in grpclb_picker. +func (s *rpcStats) merge(new *lbpb.ClientStats) { + atomic.AddInt64(&s.numCallsStarted, new.NumCallsStarted) + atomic.AddInt64(&s.numCallsFinished, new.NumCallsFinished) + atomic.AddInt64(&s.numCallsFinishedWithClientFailedToSend, new.NumCallsFinishedWithClientFailedToSend) + atomic.AddInt64(&s.numCallsFinishedKnownReceived, new.NumCallsFinishedKnownReceived) + s.mu.Lock() + for _, perToken := range new.CallsFinishedWithDrop { + s.numCallsDropped[perToken.LoadBalanceToken] += perToken.NumCalls + } + s.mu.Unlock() +} + +func mapsEqual(a, b map[string]int64) bool { + if len(a) != len(b) { + return false + } + for k, v1 := range a { + if v2, ok := b[k]; !ok || v1 != v2 { + return false + } + } + return true +} + +func atomicEqual(a, b *int64) bool { + return atomic.LoadInt64(a) == atomic.LoadInt64(b) +} + +// equal compares two rpcStats. +// +// It's a test-only method. rpcStats is defined in grpclb_picker. +func (s *rpcStats) equal(new *rpcStats) bool { + if !atomicEqual(&s.numCallsStarted, &new.numCallsStarted) { + return false + } + if !atomicEqual(&s.numCallsFinished, &new.numCallsFinished) { + return false + } + if !atomicEqual(&s.numCallsFinishedWithClientFailedToSend, &new.numCallsFinishedWithClientFailedToSend) { + return false + } + if !atomicEqual(&s.numCallsFinishedKnownReceived, &new.numCallsFinishedKnownReceived) { + return false + } + s.mu.Lock() + defer s.mu.Unlock() + new.mu.Lock() + defer new.mu.Unlock() + if !mapsEqual(s.numCallsDropped, new.numCallsDropped) { + return false + } + return true +} + type remoteBalancer struct { - sls chan *lbmpb.ServerList + sls chan *lbpb.ServerList statsDura time.Duration done chan struct{} - mu sync.Mutex - stats lbmpb.ClientStats + stats *rpcStats } func newRemoteBalancer(intervals []time.Duration) *remoteBalancer { return &remoteBalancer{ - sls: make(chan *lbmpb.ServerList, 1), - done: make(chan struct{}), + sls: make(chan *lbpb.ServerList, 1), + done: make(chan struct{}), + stats: newRPCStats(), } } @@ -145,7 +199,7 @@ func (b *remoteBalancer) stop() { close(b.done) } -func (b *remoteBalancer) BalanceLoad(stream lbspb.LoadBalancer_BalanceLoadServer) error { +func (b *remoteBalancer) BalanceLoad(stream lbgrpc.LoadBalancer_BalanceLoadServer) error { req, err := stream.Recv() if err != nil { return err @@ -154,10 +208,10 @@ func (b *remoteBalancer) BalanceLoad(stream lbspb.LoadBalancer_BalanceLoadServer if initReq.Name != beServerName { return status.Errorf(codes.InvalidArgument, "invalid service name: %v", initReq.Name) } - resp := &lbmpb.LoadBalanceResponse{ - LoadBalanceResponseType: &lbmpb.LoadBalanceResponse_InitialResponse{ - InitialResponse: &lbmpb.InitialLoadBalanceResponse{ - ClientStatsReportInterval: &lbmpb.Duration{ + resp := &lbpb.LoadBalanceResponse{ + LoadBalanceResponseType: &lbpb.LoadBalanceResponse_InitialResponse{ + InitialResponse: &lbpb.InitialLoadBalanceResponse{ + ClientStatsReportInterval: &durationpb.Duration{ Seconds: int64(b.statsDura.Seconds()), Nanos: int32(b.statsDura.Nanoseconds() - int64(b.statsDura.Seconds())*1e9), }, @@ -170,25 +224,18 @@ func (b *remoteBalancer) BalanceLoad(stream lbspb.LoadBalancer_BalanceLoadServer go func() { for { var ( - req *lbmpb.LoadBalanceRequest + req *lbpb.LoadBalanceRequest err error ) if req, err = stream.Recv(); err != nil { return } - b.mu.Lock() - b.stats.NumCallsStarted += req.GetClientStats().NumCallsStarted - b.stats.NumCallsFinished += req.GetClientStats().NumCallsFinished - b.stats.NumCallsFinishedWithDropForRateLimiting += req.GetClientStats().NumCallsFinishedWithDropForRateLimiting - b.stats.NumCallsFinishedWithDropForLoadBalancing += req.GetClientStats().NumCallsFinishedWithDropForLoadBalancing - b.stats.NumCallsFinishedWithClientFailedToSend += req.GetClientStats().NumCallsFinishedWithClientFailedToSend - b.stats.NumCallsFinishedKnownReceived += req.GetClientStats().NumCallsFinishedKnownReceived - b.mu.Unlock() + b.stats.merge(req.GetClientStats()) } }() for v := range b.sls { - resp = &lbmpb.LoadBalanceResponse{ - LoadBalanceResponseType: &lbmpb.LoadBalanceResponse_ServerList{ + resp = &lbpb.LoadBalanceResponse{ + LoadBalanceResponseType: &lbpb.LoadBalanceResponse_ServerList{ ServerList: v, }, } @@ -286,12 +333,8 @@ func newLoadBalancer(numberOfBackends int) (tss *testServers, cleanup func(), er sn: lbServerName, } lb = grpc.NewServer(grpc.Creds(lbCreds)) - if err != nil { - err = fmt.Errorf("Failed to generate the port number %v", err) - return - } ls = newRemoteBalancer(nil) - lbspb.RegisterLoadBalancerServer(lb, ls) + lbgrpc.RegisterLoadBalancerServer(lb, ls) go func() { lb.Serve(lbLis) }() @@ -325,14 +368,14 @@ func TestGRPCLB(t *testing.T) { } defer cleanup() - be := &lbmpb.Server{ + be := &lbpb.Server{ IpAddress: tss.beIPs[0], Port: int32(tss.bePorts[0]), LoadBalanceToken: lbToken, } - var bes []*lbmpb.Server + var bes []*lbpb.Server bes = append(bes, be) - sl := &lbmpb.ServerList{ + sl := &lbpb.ServerList{ Servers: bes, } tss.ls.sls <- sl @@ -373,7 +416,7 @@ func TestGRPCLBWeighted(t *testing.T) { } defer cleanup() - beServers := []*lbmpb.Server{{ + beServers := []*lbpb.Server{{ IpAddress: tss.beIPs[0], Port: int32(tss.bePorts[0]), LoadBalanceToken: lbToken, @@ -409,14 +452,14 @@ func TestGRPCLBWeighted(t *testing.T) { sequences := []string{"00101", "00011"} for _, seq := range sequences { var ( - bes []*lbmpb.Server + bes []*lbpb.Server p peer.Peer result string ) for _, s := range seq { bes = append(bes, beServers[s-'0']) } - tss.ls.sls <- &lbmpb.ServerList{Servers: bes} + tss.ls.sls <- &lbpb.ServerList{Servers: bes} for i := 0; i < 1000; i++ { if _, err := testC.EmptyCall(context.Background(), &testpb.Empty{}, grpc.FailFast(false), grpc.Peer(&p)); err != nil { @@ -442,14 +485,14 @@ func TestDropRequest(t *testing.T) { t.Fatalf("failed to create new load balancer: %v", err) } defer cleanup() - tss.ls.sls <- &lbmpb.ServerList{ - Servers: []*lbmpb.Server{{ - IpAddress: tss.beIPs[0], - Port: int32(tss.bePorts[0]), - LoadBalanceToken: lbToken, - DropForLoadBalancing: false, + tss.ls.sls <- &lbpb.ServerList{ + Servers: []*lbpb.Server{{ + IpAddress: tss.beIPs[0], + Port: int32(tss.bePorts[0]), + LoadBalanceToken: lbToken, + Drop: false, }, { - DropForLoadBalancing: true, + Drop: true, }}, } creds := serverNameCheckCreds{ @@ -476,7 +519,7 @@ func TestDropRequest(t *testing.T) { // to true. var i int for i = 0; i < 1000; i++ { - if _, err := testC.EmptyCall(context.Background(), &testpb.Empty{}, grpc.FailFast(false)); err == nil { + if _, err := testC.EmptyCall(ctx, &testpb.Empty{}, grpc.FailFast(false)); err == nil { break } time.Sleep(time.Millisecond) @@ -484,6 +527,11 @@ func TestDropRequest(t *testing.T) { if i >= 1000 { t.Fatalf("%v.SayHello(_, _) = _, %v, want _, ", testC, err) } + select { + case <-ctx.Done(): + t.Fatal("timed out", ctx.Err()) + default: + } for _, failfast := range []bool{true, false} { for i := 0; i < 3; i++ { // Even RPCs should fail, because the 2st backend has @@ -518,14 +566,14 @@ func TestBalancerDisconnects(t *testing.T) { } defer cleanup() - be := &lbmpb.Server{ + be := &lbpb.Server{ IpAddress: tss.beIPs[0], Port: int32(tss.bePorts[0]), LoadBalanceToken: lbToken, } - var bes []*lbmpb.Server + var bes []*lbpb.Server bes = append(bes, be) - sl := &lbmpb.ServerList{ + sl := &lbpb.ServerList{ Servers: bes, } tss.ls.sls <- sl @@ -593,7 +641,7 @@ const grpclbCustomFallbackName = "grpclb_with_custom_fallback_timeout" func init() { balancer.Register(&customGRPCLBBuilder{ - Builder: grpc.NewLBBuilderWithFallbackTimeout(100 * time.Millisecond), + Builder: newLBBuilderWithFallbackTimeout(100 * time.Millisecond), name: grpclbCustomFallbackName, }) } @@ -619,14 +667,14 @@ func TestFallback(t *testing.T) { standaloneBEs := startBackends(beServerName, true, beLis) defer stopBackends(standaloneBEs) - be := &lbmpb.Server{ + be := &lbpb.Server{ IpAddress: tss.beIPs[0], Port: int32(tss.bePorts[0]), LoadBalanceToken: lbToken, } - var bes []*lbmpb.Server + var bes []*lbpb.Server bes = append(bes, be) - sl := &lbmpb.ServerList{ + sl := &lbpb.ServerList{ Servers: bes, } tss.ls.sls <- sl @@ -697,14 +745,14 @@ func (failPreRPCCred) RequireTransportSecurity() bool { return false } -func checkStats(stats *lbmpb.ClientStats, expected *lbmpb.ClientStats) error { - if !proto.Equal(stats, expected) { +func checkStats(stats, expected *rpcStats) error { + if !stats.equal(expected) { return fmt.Errorf("stats not equal: got %+v, want %+v", stats, expected) } return nil } -func runAndGetStats(t *testing.T, dropForLoadBalancing, dropForRateLimiting bool, runRPCs func(*grpc.ClientConn)) lbmpb.ClientStats { +func runAndGetStats(t *testing.T, drop bool, runRPCs func(*grpc.ClientConn)) *rpcStats { defer leakcheck.Check(t) r, cleanup := manual.GenerateAndRegisterManualResolver() @@ -715,13 +763,12 @@ func runAndGetStats(t *testing.T, dropForLoadBalancing, dropForRateLimiting bool t.Fatalf("failed to create new load balancer: %v", err) } defer cleanup() - tss.ls.sls <- &lbmpb.ServerList{ - Servers: []*lbmpb.Server{{ - IpAddress: tss.beIPs[0], - Port: int32(tss.bePorts[0]), - LoadBalanceToken: lbToken, - DropForLoadBalancing: dropForLoadBalancing, - DropForRateLimiting: dropForRateLimiting, + tss.ls.sls <- &lbpb.ServerList{ + Servers: []*lbpb.Server{{ + IpAddress: tss.beIPs[0], + Port: int32(tss.bePorts[0]), + LoadBalanceToken: lbToken, + Drop: drop, }}, } tss.ls.statsDura = 100 * time.Millisecond @@ -746,9 +793,7 @@ func runAndGetStats(t *testing.T, dropForLoadBalancing, dropForRateLimiting bool runRPCs(cc) time.Sleep(1 * time.Second) - tss.ls.mu.Lock() stats := tss.ls.stats - tss.ls.mu.Unlock() return stats } @@ -760,7 +805,7 @@ const ( func TestGRPCLBStatsUnarySuccess(t *testing.T) { defer leakcheck.Check(t) - stats := runAndGetStats(t, false, false, func(cc *grpc.ClientConn) { + stats := runAndGetStats(t, false, func(cc *grpc.ClientConn) { testC := testpb.NewTestServiceClient(cc) // The first non-failfast RPC succeeds, all connections are up. if _, err := testC.EmptyCall(context.Background(), &testpb.Empty{}, grpc.FailFast(false)); err != nil { @@ -771,19 +816,19 @@ func TestGRPCLBStatsUnarySuccess(t *testing.T) { } }) - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC), - NumCallsFinished: int64(countRPC), - NumCallsFinishedKnownReceived: int64(countRPC), + if err := checkStats(stats, &rpcStats{ + numCallsStarted: int64(countRPC), + numCallsFinished: int64(countRPC), + numCallsFinishedKnownReceived: int64(countRPC), }); err != nil { t.Fatal(err) } } -func TestGRPCLBStatsUnaryDropLoadBalancing(t *testing.T) { +func TestGRPCLBStatsUnaryDrop(t *testing.T) { defer leakcheck.Check(t) c := 0 - stats := runAndGetStats(t, true, false, func(cc *grpc.ClientConn) { + stats := runAndGetStats(t, true, func(cc *grpc.ClientConn) { testC := testpb.NewTestServiceClient(cc) for { c++ @@ -798,39 +843,11 @@ func TestGRPCLBStatsUnaryDropLoadBalancing(t *testing.T) { } }) - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC + c), - NumCallsFinished: int64(countRPC + c), - NumCallsFinishedWithDropForLoadBalancing: int64(countRPC + 1), - NumCallsFinishedWithClientFailedToSend: int64(c - 1), - }); err != nil { - t.Fatal(err) - } -} - -func TestGRPCLBStatsUnaryDropRateLimiting(t *testing.T) { - defer leakcheck.Check(t) - c := 0 - stats := runAndGetStats(t, false, true, func(cc *grpc.ClientConn) { - testC := testpb.NewTestServiceClient(cc) - for { - c++ - if _, err := testC.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { - if strings.Contains(err.Error(), dropErrDesc) { - break - } - } - } - for i := 0; i < countRPC; i++ { - testC.EmptyCall(context.Background(), &testpb.Empty{}) - } - }) - - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC + c), - NumCallsFinished: int64(countRPC + c), - NumCallsFinishedWithDropForRateLimiting: int64(countRPC + 1), - NumCallsFinishedWithClientFailedToSend: int64(c - 1), + if err := checkStats(stats, &rpcStats{ + numCallsStarted: int64(countRPC + c), + numCallsFinished: int64(countRPC + c), + numCallsFinishedWithClientFailedToSend: int64(c - 1), + numCallsDropped: map[string]int64{lbToken: int64(countRPC + 1)}, }); err != nil { t.Fatal(err) } @@ -838,7 +855,7 @@ func TestGRPCLBStatsUnaryDropRateLimiting(t *testing.T) { func TestGRPCLBStatsUnaryFailedToSend(t *testing.T) { defer leakcheck.Check(t) - stats := runAndGetStats(t, false, false, func(cc *grpc.ClientConn) { + stats := runAndGetStats(t, false, func(cc *grpc.ClientConn) { testC := testpb.NewTestServiceClient(cc) // The first non-failfast RPC succeeds, all connections are up. if _, err := testC.EmptyCall(context.Background(), &testpb.Empty{}, grpc.FailFast(false)); err != nil { @@ -849,11 +866,11 @@ func TestGRPCLBStatsUnaryFailedToSend(t *testing.T) { } }) - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC), - NumCallsFinished: int64(countRPC), - NumCallsFinishedWithClientFailedToSend: int64(countRPC - 1), - NumCallsFinishedKnownReceived: 1, + if err := checkStats(stats, &rpcStats{ + numCallsStarted: int64(countRPC), + numCallsFinished: int64(countRPC), + numCallsFinishedWithClientFailedToSend: int64(countRPC - 1), + numCallsFinishedKnownReceived: 1, }); err != nil { t.Fatal(err) } @@ -861,7 +878,7 @@ func TestGRPCLBStatsUnaryFailedToSend(t *testing.T) { func TestGRPCLBStatsStreamingSuccess(t *testing.T) { defer leakcheck.Check(t) - stats := runAndGetStats(t, false, false, func(cc *grpc.ClientConn) { + stats := runAndGetStats(t, false, func(cc *grpc.ClientConn) { testC := testpb.NewTestServiceClient(cc) // The first non-failfast RPC succeeds, all connections are up. stream, err := testC.FullDuplexCall(context.Background(), grpc.FailFast(false)) @@ -886,19 +903,19 @@ func TestGRPCLBStatsStreamingSuccess(t *testing.T) { } }) - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC), - NumCallsFinished: int64(countRPC), - NumCallsFinishedKnownReceived: int64(countRPC), + if err := checkStats(stats, &rpcStats{ + numCallsStarted: int64(countRPC), + numCallsFinished: int64(countRPC), + numCallsFinishedKnownReceived: int64(countRPC), }); err != nil { t.Fatal(err) } } -func TestGRPCLBStatsStreamingDropLoadBalancing(t *testing.T) { +func TestGRPCLBStatsStreamingDrop(t *testing.T) { defer leakcheck.Check(t) c := 0 - stats := runAndGetStats(t, true, false, func(cc *grpc.ClientConn) { + stats := runAndGetStats(t, true, func(cc *grpc.ClientConn) { testC := testpb.NewTestServiceClient(cc) for { c++ @@ -913,39 +930,11 @@ func TestGRPCLBStatsStreamingDropLoadBalancing(t *testing.T) { } }) - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC + c), - NumCallsFinished: int64(countRPC + c), - NumCallsFinishedWithDropForLoadBalancing: int64(countRPC + 1), - NumCallsFinishedWithClientFailedToSend: int64(c - 1), - }); err != nil { - t.Fatal(err) - } -} - -func TestGRPCLBStatsStreamingDropRateLimiting(t *testing.T) { - defer leakcheck.Check(t) - c := 0 - stats := runAndGetStats(t, false, true, func(cc *grpc.ClientConn) { - testC := testpb.NewTestServiceClient(cc) - for { - c++ - if _, err := testC.EmptyCall(context.Background(), &testpb.Empty{}); err != nil { - if strings.Contains(err.Error(), dropErrDesc) { - break - } - } - } - for i := 0; i < countRPC; i++ { - testC.FullDuplexCall(context.Background()) - } - }) - - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC + c), - NumCallsFinished: int64(countRPC + c), - NumCallsFinishedWithDropForRateLimiting: int64(countRPC + 1), - NumCallsFinishedWithClientFailedToSend: int64(c - 1), + if err := checkStats(stats, &rpcStats{ + numCallsStarted: int64(countRPC + c), + numCallsFinished: int64(countRPC + c), + numCallsFinishedWithClientFailedToSend: int64(c - 1), + numCallsDropped: map[string]int64{lbToken: int64(countRPC + 1)}, }); err != nil { t.Fatal(err) } @@ -953,7 +942,7 @@ func TestGRPCLBStatsStreamingDropRateLimiting(t *testing.T) { func TestGRPCLBStatsStreamingFailedToSend(t *testing.T) { defer leakcheck.Check(t) - stats := runAndGetStats(t, false, false, func(cc *grpc.ClientConn) { + stats := runAndGetStats(t, false, func(cc *grpc.ClientConn) { testC := testpb.NewTestServiceClient(cc) // The first non-failfast RPC succeeds, all connections are up. stream, err := testC.FullDuplexCall(context.Background(), grpc.FailFast(false)) @@ -970,11 +959,11 @@ func TestGRPCLBStatsStreamingFailedToSend(t *testing.T) { } }) - if err := checkStats(&stats, &lbmpb.ClientStats{ - NumCallsStarted: int64(countRPC), - NumCallsFinished: int64(countRPC), - NumCallsFinishedWithClientFailedToSend: int64(countRPC - 1), - NumCallsFinishedKnownReceived: 1, + if err := checkStats(stats, &rpcStats{ + numCallsStarted: int64(countRPC), + numCallsFinished: int64(countRPC), + numCallsFinishedWithClientFailedToSend: int64(countRPC - 1), + numCallsFinishedKnownReceived: 1, }); err != nil { t.Fatal(err) } diff --git a/vendor/google.golang.org/grpc/grpclb_util.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_util.go similarity index 99% rename from vendor/google.golang.org/grpc/grpclb_util.go rename to vendor/google.golang.org/grpc/balancer/grpclb/grpclb_util.go index 063ba9d85..3194c0707 100644 --- a/vendor/google.golang.org/grpc/grpclb_util.go +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_util.go @@ -16,7 +16,7 @@ * */ -package grpc +package grpclb import ( "fmt" diff --git a/vendor/google.golang.org/grpc/grpclb_util_test.go b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_util_test.go similarity index 99% rename from vendor/google.golang.org/grpc/grpclb_util_test.go rename to vendor/google.golang.org/grpc/balancer/grpclb/grpclb_util_test.go index 2773fae74..39d422a70 100644 --- a/vendor/google.golang.org/grpc/grpclb_util_test.go +++ b/vendor/google.golang.org/grpc/balancer/grpclb/grpclb_util_test.go @@ -16,7 +16,7 @@ * */ -package grpc +package grpclb import ( "fmt" diff --git a/vendor/google.golang.org/grpc/balancer/grpclb/regenerate.sh b/vendor/google.golang.org/grpc/balancer/grpclb/regenerate.sh new file mode 100755 index 000000000..b8978e11b --- /dev/null +++ b/vendor/google.golang.org/grpc/balancer/grpclb/regenerate.sh @@ -0,0 +1,33 @@ +#!/bin/bash +# Copyright 2018 gRPC authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +set -eux -o pipefail + +TMP=$(mktemp -d) + +function finish { + rm -rf "$TMP" +} +trap finish EXIT + +pushd "$TMP" +mkdir -p grpc/lb/v1 +curl https://raw.githubusercontent.com/grpc/grpc-proto/master/grpc/lb/v1/load_balancer.proto > grpc/lb/v1/load_balancer.proto + +protoc --go_out=plugins=grpc,paths=source_relative:. -I. grpc/lb/v1/*.proto +popd +rm -f grpc_lb_v1/*.pb.go +cp "$TMP"/grpc/lb/v1/*.pb.go grpc_lb_v1/ + diff --git a/vendor/google.golang.org/grpc/balancer/roundrobin/roundrobin_test.go b/vendor/google.golang.org/grpc/balancer/roundrobin/roundrobin_test.go index 59cac4b1b..70d8af716 100644 --- a/vendor/google.golang.org/grpc/balancer/roundrobin/roundrobin_test.go +++ b/vendor/google.golang.org/grpc/balancer/roundrobin/roundrobin_test.go @@ -30,12 +30,12 @@ import ( "google.golang.org/grpc/balancer/roundrobin" "google.golang.org/grpc/codes" _ "google.golang.org/grpc/grpclog/glogger" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/peer" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" "google.golang.org/grpc/status" testpb "google.golang.org/grpc/test/grpc_testing" - "google.golang.org/grpc/test/leakcheck" ) type testServer struct { diff --git a/vendor/google.golang.org/grpc/balancer_switching_test.go b/vendor/google.golang.org/grpc/balancer_switching_test.go index e44198e66..0dcd9165c 100644 --- a/vendor/google.golang.org/grpc/balancer_switching_test.go +++ b/vendor/google.golang.org/grpc/balancer_switching_test.go @@ -25,13 +25,39 @@ import ( "time" "golang.org/x/net/context" + "google.golang.org/grpc/balancer" "google.golang.org/grpc/balancer/roundrobin" + "google.golang.org/grpc/connectivity" _ "google.golang.org/grpc/grpclog/glogger" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" - "google.golang.org/grpc/test/leakcheck" ) +var _ balancer.Builder = &magicalLB{} +var _ balancer.Balancer = &magicalLB{} + +// magicalLB is a ringer for grpclb. It is used to avoid circular dependencies on the grpclb package +type magicalLB struct{} + +func (b *magicalLB) Name() string { + return "grpclb" +} + +func (b *magicalLB) Build(cc balancer.ClientConn, opts balancer.BuildOptions) balancer.Balancer { + return b +} + +func (b *magicalLB) HandleSubConnStateChange(balancer.SubConn, connectivity.State) {} + +func (b *magicalLB) HandleResolvedAddrs([]resolver.Address, error) {} + +func (b *magicalLB) Close() {} + +func init() { + balancer.Register(&magicalLB{}) +} + func checkPickFirst(cc *ClientConn, servers []*server) error { var ( req = "port" diff --git a/vendor/google.golang.org/grpc/balancer_test.go b/vendor/google.golang.org/grpc/balancer_test.go index cc7e46dd7..8675e7d5e 100644 --- a/vendor/google.golang.org/grpc/balancer_test.go +++ b/vendor/google.golang.org/grpc/balancer_test.go @@ -29,9 +29,9 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/codes" _ "google.golang.org/grpc/grpclog/glogger" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/naming" "google.golang.org/grpc/status" - "google.golang.org/grpc/test/leakcheck" // V1 balancer tests use passthrough resolver instead of dns. // TODO(bar) remove this when removing v1 balaner entirely. diff --git a/vendor/google.golang.org/grpc/balancer_v1_wrapper.go b/vendor/google.golang.org/grpc/balancer_v1_wrapper.go index b7abc6b74..e0ce32cfb 100644 --- a/vendor/google.golang.org/grpc/balancer_v1_wrapper.go +++ b/vendor/google.golang.org/grpc/balancer_v1_wrapper.go @@ -55,7 +55,7 @@ func (bwb *balancerWrapperBuilder) Build(cc balancer.ClientConn, opts balancer.B startCh: make(chan struct{}), conns: make(map[resolver.Address]balancer.SubConn), connSt: make(map[balancer.SubConn]*scState), - csEvltr: &connectivityStateEvaluator{}, + csEvltr: &balancer.ConnectivityStateEvaluator{}, state: connectivity.Idle, } cc.UpdateBalancerState(connectivity.Idle, bw) @@ -80,10 +80,6 @@ type balancerWrapper struct { cc balancer.ClientConn targetAddr string // Target without the scheme. - // To aggregate the connectivity state. - csEvltr *connectivityStateEvaluator - state connectivity.State - mu sync.Mutex conns map[resolver.Address]balancer.SubConn connSt map[balancer.SubConn]*scState @@ -92,6 +88,10 @@ type balancerWrapper struct { // - NewSubConn is created, cc wants to notify balancer of state changes; // - Build hasn't return, cc doesn't have access to balancer. startCh chan struct{} + + // To aggregate the connectivity state. + csEvltr *balancer.ConnectivityStateEvaluator + state connectivity.State } // lbWatcher watches the Notify channel of the balancer and manages @@ -248,7 +248,7 @@ func (bw *balancerWrapper) HandleSubConnStateChange(sc balancer.SubConn, s conne scSt.down(errConnClosing) } } - sa := bw.csEvltr.recordTransition(oldS, s) + sa := bw.csEvltr.RecordTransition(oldS, s) if bw.state != sa { bw.state = sa } @@ -326,47 +326,3 @@ func (bw *balancerWrapper) Pick(ctx context.Context, opts balancer.PickOptions) return sc, done, nil } - -// connectivityStateEvaluator gets updated by addrConns when their -// states transition, based on which it evaluates the state of -// ClientConn. -type connectivityStateEvaluator struct { - mu sync.Mutex - numReady uint64 // Number of addrConns in ready state. - numConnecting uint64 // Number of addrConns in connecting state. - numTransientFailure uint64 // Number of addrConns in transientFailure. -} - -// recordTransition records state change happening in every subConn and based on -// that it evaluates what aggregated state should be. -// It can only transition between Ready, Connecting and TransientFailure. Other states, -// Idle and Shutdown are transitioned into by ClientConn; in the beginning of the connection -// before any subConn is created ClientConn is in idle state. In the end when ClientConn -// closes it is in Shutdown state. -// TODO Note that in later releases, a ClientConn with no activity will be put into an Idle state. -func (cse *connectivityStateEvaluator) recordTransition(oldState, newState connectivity.State) connectivity.State { - cse.mu.Lock() - defer cse.mu.Unlock() - - // Update counters. - for idx, state := range []connectivity.State{oldState, newState} { - updateVal := 2*uint64(idx) - 1 // -1 for oldState and +1 for new. - switch state { - case connectivity.Ready: - cse.numReady += updateVal - case connectivity.Connecting: - cse.numConnecting += updateVal - case connectivity.TransientFailure: - cse.numTransientFailure += updateVal - } - } - - // Evaluate. - if cse.numReady > 0 { - return connectivity.Ready - } - if cse.numConnecting > 0 { - return connectivity.Connecting - } - return connectivity.TransientFailure -} diff --git a/vendor/google.golang.org/grpc/benchmark/benchmark.go b/vendor/google.golang.org/grpc/benchmark/benchmark.go index dc5d97afb..ec9b50b2f 100644 --- a/vendor/google.golang.org/grpc/benchmark/benchmark.go +++ b/vendor/google.golang.org/grpc/benchmark/benchmark.go @@ -226,9 +226,14 @@ func DoByteBufStreamingRoundTrip(stream testpb.BenchmarkService_StreamingCallCli // NewClientConn creates a gRPC client connection to addr. func NewClientConn(addr string, opts ...grpc.DialOption) *grpc.ClientConn { + return NewClientConnWithContext(context.Background(), addr, opts...) +} + +// NewClientConnWithContext creates a gRPC client connection to addr using ctx. +func NewClientConnWithContext(ctx context.Context, addr string, opts ...grpc.DialOption) *grpc.ClientConn { opts = append(opts, grpc.WithWriteBufferSize(128*1024)) opts = append(opts, grpc.WithReadBufferSize(128*1024)) - conn, err := grpc.Dial(addr, opts...) + conn, err := grpc.DialContext(ctx, addr, opts...) if err != nil { grpclog.Fatalf("NewClientConn(%q) failed to create a ClientConn %v", addr, err) } diff --git a/vendor/google.golang.org/grpc/benchmark/client/main.go b/vendor/google.golang.org/grpc/benchmark/client/main.go index 0984b6fe2..13a5fb3e6 100644 --- a/vendor/google.golang.org/grpc/benchmark/client/main.go +++ b/vendor/google.golang.org/grpc/benchmark/client/main.go @@ -20,14 +20,15 @@ package main import ( "flag" - "math" - "net" - "net/http" - _ "net/http/pprof" + "fmt" + "os" + "runtime" + "runtime/pprof" "sync" "time" "golang.org/x/net/context" + "golang.org/x/sys/unix" "google.golang.org/grpc" "google.golang.org/grpc/benchmark" testpb "google.golang.org/grpc/benchmark/grpc_testing" @@ -36,145 +37,159 @@ import ( ) var ( - server = flag.String("server", "", "The server address") - maxConcurrentRPCs = flag.Int("max_concurrent_rpcs", 1, "The max number of concurrent RPCs") - duration = flag.Int("duration", math.MaxInt32, "The duration in seconds to run the benchmark client") - trace = flag.Bool("trace", true, "Whether tracing is on") - rpcType = flag.Int("rpc_type", 0, + port = flag.String("port", "50051", "Localhost port to connect to.") + numRPC = flag.Int("r", 1, "The number of concurrent RPCs on each connection.") + numConn = flag.Int("c", 1, "The number of parallel connections.") + warmupDur = flag.Int("w", 10, "Warm-up duration in seconds") + duration = flag.Int("d", 60, "Benchmark duration in seconds") + rqSize = flag.Int("req", 1, "Request message size in bytes.") + rspSize = flag.Int("resp", 1, "Response message size in bytes.") + rpcType = flag.String("rpc_type", "unary", `Configure different client rpc type. Valid options are: - 0 : unary call; - 1 : streaming call.`) + unary; + streaming.`) + testName = flag.String("test_name", "", "Name of the test used for creating profiles.") + wg sync.WaitGroup + hopts = stats.HistogramOptions{ + NumBuckets: 2495, + GrowthFactor: .01, + } + mu sync.Mutex + hists []*stats.Histogram ) -func unaryCaller(client testpb.BenchmarkServiceClient) { - benchmark.DoUnaryCall(client, 1, 1) -} - -func streamCaller(stream testpb.BenchmarkService_StreamingCallClient) { - benchmark.DoStreamingRoundTrip(stream, 1, 1) -} - -func buildConnection() (s *stats.Stats, conn *grpc.ClientConn, tc testpb.BenchmarkServiceClient) { - s = stats.NewStats(256) - conn = benchmark.NewClientConn(*server) - tc = testpb.NewBenchmarkServiceClient(conn) - return s, conn, tc -} - -func closeLoopUnary() { - s, conn, tc := buildConnection() - - for i := 0; i < 100; i++ { - unaryCaller(tc) - } - ch := make(chan int, *maxConcurrentRPCs*4) - var ( - mu sync.Mutex - wg sync.WaitGroup - ) - wg.Add(*maxConcurrentRPCs) - - for i := 0; i < *maxConcurrentRPCs; i++ { - go func() { - for range ch { - start := time.Now() - unaryCaller(tc) - elapse := time.Since(start) - mu.Lock() - s.Add(elapse) - mu.Unlock() - } - wg.Done() - }() - } - // Stop the client when time is up. - done := make(chan struct{}) - go func() { - <-time.After(time.Duration(*duration) * time.Second) - close(done) - }() - ok := true - for ok { - select { - case ch <- 0: - case <-done: - ok = false - } - } - close(ch) - wg.Wait() - conn.Close() - grpclog.Infoln(s.String()) - -} - -func closeLoopStream() { - s, conn, tc := buildConnection() - ch := make(chan int, *maxConcurrentRPCs*4) - var ( - mu sync.Mutex - wg sync.WaitGroup - ) - wg.Add(*maxConcurrentRPCs) - // Distribute RPCs over maxConcurrentCalls workers. - for i := 0; i < *maxConcurrentRPCs; i++ { - go func() { - stream, err := tc.StreamingCall(context.Background()) - if err != nil { - grpclog.Fatalf("%v.StreamingCall(_) = _, %v", tc, err) - } - // Do some warm up. - for i := 0; i < 100; i++ { - streamCaller(stream) - } - for range ch { - start := time.Now() - streamCaller(stream) - elapse := time.Since(start) - mu.Lock() - s.Add(elapse) - mu.Unlock() - } - wg.Done() - }() - } - // Stop the client when time is up. - done := make(chan struct{}) - go func() { - <-time.After(time.Duration(*duration) * time.Second) - close(done) - }() - ok := true - for ok { - select { - case ch <- 0: - case <-done: - ok = false - } - } - close(ch) - wg.Wait() - conn.Close() - grpclog.Infoln(s.String()) -} - func main() { flag.Parse() - grpc.EnableTracing = *trace - go func() { - lis, err := net.Listen("tcp", ":0") - if err != nil { - grpclog.Fatalf("Failed to listen: %v", err) - } - grpclog.Infoln("Client profiling address: ", lis.Addr().String()) - if err := http.Serve(lis, nil); err != nil { - grpclog.Fatalf("Failed to serve: %v", err) - } - }() - switch *rpcType { - case 0: - closeLoopUnary() - case 1: - closeLoopStream() + if *testName == "" { + grpclog.Fatalf("test_name not set") + } + req := &testpb.SimpleRequest{ + ResponseType: testpb.PayloadType_COMPRESSABLE, + ResponseSize: int32(*rspSize), + Payload: &testpb.Payload{ + Type: testpb.PayloadType_COMPRESSABLE, + Body: make([]byte, *rqSize), + }, + } + connectCtx, connectCancel := context.WithDeadline(context.Background(), time.Now().Add(5*time.Second)) + defer connectCancel() + ccs := buildConnections(connectCtx) + warmDeadline := time.Now().Add(time.Duration(*warmupDur) * time.Second) + endDeadline := warmDeadline.Add(time.Duration(*duration) * time.Second) + cf, err := os.Create("/tmp/" + *testName + ".cpu") + if err != nil { + grpclog.Fatalf("Error creating file: %v", err) + } + defer cf.Close() + pprof.StartCPUProfile(cf) + cpuBeg := getCPUTime() + for _, cc := range ccs { + runWithConn(cc, req, warmDeadline, endDeadline) + } + wg.Wait() + cpu := time.Duration(getCPUTime() - cpuBeg) + pprof.StopCPUProfile() + mf, err := os.Create("/tmp/" + *testName + ".mem") + if err != nil { + grpclog.Fatalf("Error creating file: %v", err) + } + defer mf.Close() + runtime.GC() // materialize all statistics + if err := pprof.WriteHeapProfile(mf); err != nil { + grpclog.Fatalf("Error writing memory profile: %v", err) + } + hist := stats.NewHistogram(hopts) + for _, h := range hists { + hist.Merge(h) + } + parseHist(hist) + fmt.Println("Client CPU utilization:", cpu) + fmt.Println("Client CPU profile:", cf.Name()) + fmt.Println("Client Mem Profile:", mf.Name()) +} + +func buildConnections(ctx context.Context) []*grpc.ClientConn { + ccs := make([]*grpc.ClientConn, *numConn) + for i := range ccs { + ccs[i] = benchmark.NewClientConnWithContext(ctx, "localhost:"+*port, grpc.WithInsecure(), grpc.WithBlock()) + } + return ccs +} + +func runWithConn(cc *grpc.ClientConn, req *testpb.SimpleRequest, warmDeadline, endDeadline time.Time) { + for i := 0; i < *numRPC; i++ { + wg.Add(1) + go func() { + defer wg.Done() + caller := makeCaller(cc, req) + hist := stats.NewHistogram(hopts) + for { + start := time.Now() + if start.After(endDeadline) { + mu.Lock() + hists = append(hists, hist) + mu.Unlock() + return + } + caller() + elapsed := time.Since(start) + if start.After(warmDeadline) { + hist.Add(elapsed.Nanoseconds()) + } + } + }() } } + +func makeCaller(cc *grpc.ClientConn, req *testpb.SimpleRequest) func() { + client := testpb.NewBenchmarkServiceClient(cc) + if *rpcType == "unary" { + return func() { + if _, err := client.UnaryCall(context.Background(), req); err != nil { + grpclog.Fatalf("RPC failed: %v", err) + } + } + } + stream, err := client.StreamingCall(context.Background()) + if err != nil { + grpclog.Fatalf("RPC failed: %v", err) + } + return func() { + if err := stream.Send(req); err != nil { + grpclog.Fatalf("Streaming RPC failed to send: %v", err) + } + if _, err := stream.Recv(); err != nil { + grpclog.Fatalf("Streaming RPC failed to read: %v", err) + } + } +} + +func parseHist(hist *stats.Histogram) { + fmt.Println("qps:", float64(hist.Count)/float64(*duration)) + fmt.Printf("Latency: (50/90/99 %%ile): %v/%v/%v\n", + time.Duration(median(.5, hist)), + time.Duration(median(.9, hist)), + time.Duration(median(.99, hist))) +} + +func median(percentile float64, h *stats.Histogram) int64 { + need := int64(float64(h.Count) * percentile) + have := int64(0) + for _, bucket := range h.Buckets { + count := bucket.Count + if have+count >= need { + percent := float64(need-have) / float64(count) + return int64((1.0-percent)*bucket.LowBound + percent*bucket.LowBound*(1.0+hopts.GrowthFactor)) + } + have += bucket.Count + } + panic("should have found a bound") +} + +func getCPUTime() int64 { + var ts unix.Timespec + if err := unix.ClockGettime(unix.CLOCK_PROCESS_CPUTIME_ID, &ts); err != nil { + grpclog.Fatal(err) + } + return ts.Nano() +} diff --git a/vendor/google.golang.org/grpc/benchmark/grpc_testing/control.pb.go b/vendor/google.golang.org/grpc/benchmark/grpc_testing/control.pb.go index a6f286531..22134427a 100644 --- a/vendor/google.golang.org/grpc/benchmark/grpc_testing/control.pb.go +++ b/vendor/google.golang.org/grpc/benchmark/grpc_testing/control.pb.go @@ -94,7 +94,7 @@ func (RpcType) EnumDescriptor() ([]byte, []int) { // of activity coming in from independent identical stationary sources. type PoissonParams struct { // The rate of arrivals (a.k.a. lambda parameter of the exp distribution). - OfferedLoad float64 `protobuf:"fixed64,1,opt,name=offered_load,json=offeredLoad" json:"offered_load,omitempty"` + OfferedLoad float64 `protobuf:"fixed64,1,opt,name=offered_load,json=offeredLoad,proto3" json:"offered_load,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -132,8 +132,8 @@ func (m *PoissonParams) GetOfferedLoad() float64 { } type UniformParams struct { - InterarrivalLo float64 `protobuf:"fixed64,1,opt,name=interarrival_lo,json=interarrivalLo" json:"interarrival_lo,omitempty"` - InterarrivalHi float64 `protobuf:"fixed64,2,opt,name=interarrival_hi,json=interarrivalHi" json:"interarrival_hi,omitempty"` + InterarrivalLo float64 `protobuf:"fixed64,1,opt,name=interarrival_lo,json=interarrivalLo,proto3" json:"interarrival_lo,omitempty"` + InterarrivalHi float64 `protobuf:"fixed64,2,opt,name=interarrival_hi,json=interarrivalHi,proto3" json:"interarrival_hi,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -178,7 +178,7 @@ func (m *UniformParams) GetInterarrivalHi() float64 { } type DeterministicParams struct { - OfferedLoad float64 `protobuf:"fixed64,1,opt,name=offered_load,json=offeredLoad" json:"offered_load,omitempty"` + OfferedLoad float64 `protobuf:"fixed64,1,opt,name=offered_load,json=offeredLoad,proto3" json:"offered_load,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -216,8 +216,8 @@ func (m *DeterministicParams) GetOfferedLoad() float64 { } type ParetoParams struct { - InterarrivalBase float64 `protobuf:"fixed64,1,opt,name=interarrival_base,json=interarrivalBase" json:"interarrival_base,omitempty"` - Alpha float64 `protobuf:"fixed64,2,opt,name=alpha" json:"alpha,omitempty"` + InterarrivalBase float64 `protobuf:"fixed64,1,opt,name=interarrival_base,json=interarrivalBase,proto3" json:"interarrival_base,omitempty"` + Alpha float64 `protobuf:"fixed64,2,opt,name=alpha,proto3" json:"alpha,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -335,19 +335,19 @@ type isLoadParams_Load interface { } type LoadParams_ClosedLoop struct { - ClosedLoop *ClosedLoopParams `protobuf:"bytes,1,opt,name=closed_loop,json=closedLoop,oneof"` + ClosedLoop *ClosedLoopParams `protobuf:"bytes,1,opt,name=closed_loop,json=closedLoop,proto3,oneof"` } type LoadParams_Poisson struct { - Poisson *PoissonParams `protobuf:"bytes,2,opt,name=poisson,oneof"` + Poisson *PoissonParams `protobuf:"bytes,2,opt,name=poisson,proto3,oneof"` } type LoadParams_Uniform struct { - Uniform *UniformParams `protobuf:"bytes,3,opt,name=uniform,oneof"` + Uniform *UniformParams `protobuf:"bytes,3,opt,name=uniform,proto3,oneof"` } type LoadParams_Determ struct { - Determ *DeterministicParams `protobuf:"bytes,4,opt,name=determ,oneof"` + Determ *DeterministicParams `protobuf:"bytes,4,opt,name=determ,proto3,oneof"` } type LoadParams_Pareto struct { - Pareto *ParetoParams `protobuf:"bytes,5,opt,name=pareto,oneof"` + Pareto *ParetoParams `protobuf:"bytes,5,opt,name=pareto,proto3,oneof"` } func (*LoadParams_ClosedLoop) isLoadParams_Load() {} @@ -531,8 +531,8 @@ func _LoadParams_OneofSizer(msg proto.Message) (n int) { // presence of SecurityParams implies use of TLS type SecurityParams struct { - UseTestCa bool `protobuf:"varint,1,opt,name=use_test_ca,json=useTestCa" json:"use_test_ca,omitempty"` - ServerHostOverride string `protobuf:"bytes,2,opt,name=server_host_override,json=serverHostOverride" json:"server_host_override,omitempty"` + UseTestCa bool `protobuf:"varint,1,opt,name=use_test_ca,json=useTestCa,proto3" json:"use_test_ca,omitempty"` + ServerHostOverride string `protobuf:"bytes,2,opt,name=server_host_override,json=serverHostOverride,proto3" json:"server_host_override,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -578,25 +578,25 @@ func (m *SecurityParams) GetServerHostOverride() string { type ClientConfig struct { // List of targets to connect to. At least one target needs to be specified. - ServerTargets []string `protobuf:"bytes,1,rep,name=server_targets,json=serverTargets" json:"server_targets,omitempty"` - ClientType ClientType `protobuf:"varint,2,opt,name=client_type,json=clientType,enum=grpc.testing.ClientType" json:"client_type,omitempty"` - SecurityParams *SecurityParams `protobuf:"bytes,3,opt,name=security_params,json=securityParams" json:"security_params,omitempty"` + ServerTargets []string `protobuf:"bytes,1,rep,name=server_targets,json=serverTargets,proto3" json:"server_targets,omitempty"` + ClientType ClientType `protobuf:"varint,2,opt,name=client_type,json=clientType,proto3,enum=grpc.testing.ClientType" json:"client_type,omitempty"` + SecurityParams *SecurityParams `protobuf:"bytes,3,opt,name=security_params,json=securityParams,proto3" json:"security_params,omitempty"` // How many concurrent RPCs to start for each channel. // For synchronous client, use a separate thread for each outstanding RPC. - OutstandingRpcsPerChannel int32 `protobuf:"varint,4,opt,name=outstanding_rpcs_per_channel,json=outstandingRpcsPerChannel" json:"outstanding_rpcs_per_channel,omitempty"` + OutstandingRpcsPerChannel int32 `protobuf:"varint,4,opt,name=outstanding_rpcs_per_channel,json=outstandingRpcsPerChannel,proto3" json:"outstanding_rpcs_per_channel,omitempty"` // Number of independent client channels to create. // i-th channel will connect to server_target[i % server_targets.size()] - ClientChannels int32 `protobuf:"varint,5,opt,name=client_channels,json=clientChannels" json:"client_channels,omitempty"` + ClientChannels int32 `protobuf:"varint,5,opt,name=client_channels,json=clientChannels,proto3" json:"client_channels,omitempty"` // Only for async client. Number of threads to use to start/manage RPCs. - AsyncClientThreads int32 `protobuf:"varint,7,opt,name=async_client_threads,json=asyncClientThreads" json:"async_client_threads,omitempty"` - RpcType RpcType `protobuf:"varint,8,opt,name=rpc_type,json=rpcType,enum=grpc.testing.RpcType" json:"rpc_type,omitempty"` + AsyncClientThreads int32 `protobuf:"varint,7,opt,name=async_client_threads,json=asyncClientThreads,proto3" json:"async_client_threads,omitempty"` + RpcType RpcType `protobuf:"varint,8,opt,name=rpc_type,json=rpcType,proto3,enum=grpc.testing.RpcType" json:"rpc_type,omitempty"` // The requested load for the entire client (aggregated over all the threads). - LoadParams *LoadParams `protobuf:"bytes,10,opt,name=load_params,json=loadParams" json:"load_params,omitempty"` - PayloadConfig *PayloadConfig `protobuf:"bytes,11,opt,name=payload_config,json=payloadConfig" json:"payload_config,omitempty"` - HistogramParams *HistogramParams `protobuf:"bytes,12,opt,name=histogram_params,json=histogramParams" json:"histogram_params,omitempty"` + LoadParams *LoadParams `protobuf:"bytes,10,opt,name=load_params,json=loadParams,proto3" json:"load_params,omitempty"` + PayloadConfig *PayloadConfig `protobuf:"bytes,11,opt,name=payload_config,json=payloadConfig,proto3" json:"payload_config,omitempty"` + HistogramParams *HistogramParams `protobuf:"bytes,12,opt,name=histogram_params,json=histogramParams,proto3" json:"histogram_params,omitempty"` // Specify the cores we should run the client on, if desired - CoreList []int32 `protobuf:"varint,13,rep,packed,name=core_list,json=coreList" json:"core_list,omitempty"` - CoreLimit int32 `protobuf:"varint,14,opt,name=core_limit,json=coreLimit" json:"core_limit,omitempty"` + CoreList []int32 `protobuf:"varint,13,rep,packed,name=core_list,json=coreList,proto3" json:"core_list,omitempty"` + CoreLimit int32 `protobuf:"varint,14,opt,name=core_limit,json=coreLimit,proto3" json:"core_limit,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -711,7 +711,7 @@ func (m *ClientConfig) GetCoreLimit() int32 { } type ClientStatus struct { - Stats *ClientStats `protobuf:"bytes,1,opt,name=stats" json:"stats,omitempty"` + Stats *ClientStats `protobuf:"bytes,1,opt,name=stats,proto3" json:"stats,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -751,7 +751,7 @@ func (m *ClientStatus) GetStats() *ClientStats { // Request current stats type Mark struct { // if true, the stats will be reset after taking their snapshot. - Reset_ bool `protobuf:"varint,1,opt,name=reset" json:"reset,omitempty"` + Reset_ bool `protobuf:"varint,1,opt,name=reset,proto3" json:"reset,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -827,10 +827,10 @@ type isClientArgs_Argtype interface { } type ClientArgs_Setup struct { - Setup *ClientConfig `protobuf:"bytes,1,opt,name=setup,oneof"` + Setup *ClientConfig `protobuf:"bytes,1,opt,name=setup,proto3,oneof"` } type ClientArgs_Mark struct { - Mark *Mark `protobuf:"bytes,2,opt,name=mark,oneof"` + Mark *Mark `protobuf:"bytes,2,opt,name=mark,proto3,oneof"` } func (*ClientArgs_Setup) isClientArgs_Argtype() {} @@ -932,18 +932,18 @@ func _ClientArgs_OneofSizer(msg proto.Message) (n int) { } type ServerConfig struct { - ServerType ServerType `protobuf:"varint,1,opt,name=server_type,json=serverType,enum=grpc.testing.ServerType" json:"server_type,omitempty"` - SecurityParams *SecurityParams `protobuf:"bytes,2,opt,name=security_params,json=securityParams" json:"security_params,omitempty"` + ServerType ServerType `protobuf:"varint,1,opt,name=server_type,json=serverType,proto3,enum=grpc.testing.ServerType" json:"server_type,omitempty"` + SecurityParams *SecurityParams `protobuf:"bytes,2,opt,name=security_params,json=securityParams,proto3" json:"security_params,omitempty"` // Port on which to listen. Zero means pick unused port. - Port int32 `protobuf:"varint,4,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,4,opt,name=port,proto3" json:"port,omitempty"` // Only for async server. Number of threads used to serve the requests. - AsyncServerThreads int32 `protobuf:"varint,7,opt,name=async_server_threads,json=asyncServerThreads" json:"async_server_threads,omitempty"` + AsyncServerThreads int32 `protobuf:"varint,7,opt,name=async_server_threads,json=asyncServerThreads,proto3" json:"async_server_threads,omitempty"` // Specify the number of cores to limit server to, if desired - CoreLimit int32 `protobuf:"varint,8,opt,name=core_limit,json=coreLimit" json:"core_limit,omitempty"` + CoreLimit int32 `protobuf:"varint,8,opt,name=core_limit,json=coreLimit,proto3" json:"core_limit,omitempty"` // payload config, used in generic server - PayloadConfig *PayloadConfig `protobuf:"bytes,9,opt,name=payload_config,json=payloadConfig" json:"payload_config,omitempty"` + PayloadConfig *PayloadConfig `protobuf:"bytes,9,opt,name=payload_config,json=payloadConfig,proto3" json:"payload_config,omitempty"` // Specify the cores we should run the server on, if desired - CoreList []int32 `protobuf:"varint,10,rep,packed,name=core_list,json=coreList" json:"core_list,omitempty"` + CoreList []int32 `protobuf:"varint,10,rep,packed,name=core_list,json=coreList,proto3" json:"core_list,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1061,10 +1061,10 @@ type isServerArgs_Argtype interface { } type ServerArgs_Setup struct { - Setup *ServerConfig `protobuf:"bytes,1,opt,name=setup,oneof"` + Setup *ServerConfig `protobuf:"bytes,1,opt,name=setup,proto3,oneof"` } type ServerArgs_Mark struct { - Mark *Mark `protobuf:"bytes,2,opt,name=mark,oneof"` + Mark *Mark `protobuf:"bytes,2,opt,name=mark,proto3,oneof"` } func (*ServerArgs_Setup) isServerArgs_Argtype() {} @@ -1166,11 +1166,11 @@ func _ServerArgs_OneofSizer(msg proto.Message) (n int) { } type ServerStatus struct { - Stats *ServerStats `protobuf:"bytes,1,opt,name=stats" json:"stats,omitempty"` + Stats *ServerStats `protobuf:"bytes,1,opt,name=stats,proto3" json:"stats,omitempty"` // the port bound by the server - Port int32 `protobuf:"varint,2,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,2,opt,name=port,proto3" json:"port,omitempty"` // Number of cores available to the server - Cores int32 `protobuf:"varint,3,opt,name=cores" json:"cores,omitempty"` + Cores int32 `protobuf:"varint,3,opt,name=cores,proto3" json:"cores,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1253,7 +1253,7 @@ var xxx_messageInfo_CoreRequest proto.InternalMessageInfo type CoreResponse struct { // Number of cores available on the server - Cores int32 `protobuf:"varint,1,opt,name=cores" json:"cores,omitempty"` + Cores int32 `protobuf:"varint,1,opt,name=cores,proto3" json:"cores,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1323,21 +1323,21 @@ var xxx_messageInfo_Void proto.InternalMessageInfo // A single performance scenario: input to qps_json_driver type Scenario struct { // Human readable name for this scenario - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Client configuration - ClientConfig *ClientConfig `protobuf:"bytes,2,opt,name=client_config,json=clientConfig" json:"client_config,omitempty"` + ClientConfig *ClientConfig `protobuf:"bytes,2,opt,name=client_config,json=clientConfig,proto3" json:"client_config,omitempty"` // Number of clients to start for the test - NumClients int32 `protobuf:"varint,3,opt,name=num_clients,json=numClients" json:"num_clients,omitempty"` + NumClients int32 `protobuf:"varint,3,opt,name=num_clients,json=numClients,proto3" json:"num_clients,omitempty"` // Server configuration - ServerConfig *ServerConfig `protobuf:"bytes,4,opt,name=server_config,json=serverConfig" json:"server_config,omitempty"` + ServerConfig *ServerConfig `protobuf:"bytes,4,opt,name=server_config,json=serverConfig,proto3" json:"server_config,omitempty"` // Number of servers to start for the test - NumServers int32 `protobuf:"varint,5,opt,name=num_servers,json=numServers" json:"num_servers,omitempty"` + NumServers int32 `protobuf:"varint,5,opt,name=num_servers,json=numServers,proto3" json:"num_servers,omitempty"` // Warmup period, in seconds - WarmupSeconds int32 `protobuf:"varint,6,opt,name=warmup_seconds,json=warmupSeconds" json:"warmup_seconds,omitempty"` + WarmupSeconds int32 `protobuf:"varint,6,opt,name=warmup_seconds,json=warmupSeconds,proto3" json:"warmup_seconds,omitempty"` // Benchmark time, in seconds - BenchmarkSeconds int32 `protobuf:"varint,7,opt,name=benchmark_seconds,json=benchmarkSeconds" json:"benchmark_seconds,omitempty"` + BenchmarkSeconds int32 `protobuf:"varint,7,opt,name=benchmark_seconds,json=benchmarkSeconds,proto3" json:"benchmark_seconds,omitempty"` // Number of workers to spawn locally (usually zero) - SpawnLocalWorkerCount int32 `protobuf:"varint,8,opt,name=spawn_local_worker_count,json=spawnLocalWorkerCount" json:"spawn_local_worker_count,omitempty"` + SpawnLocalWorkerCount int32 `protobuf:"varint,8,opt,name=spawn_local_worker_count,json=spawnLocalWorkerCount,proto3" json:"spawn_local_worker_count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1425,7 +1425,7 @@ func (m *Scenario) GetSpawnLocalWorkerCount() int32 { // A set of scenarios to be run with qps_json_driver type Scenarios struct { - Scenarios []*Scenario `protobuf:"bytes,1,rep,name=scenarios" json:"scenarios,omitempty"` + Scenarios []*Scenario `protobuf:"bytes,1,rep,name=scenarios,proto3" json:"scenarios,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/grpc/benchmark/grpc_testing/messages.pb.go b/vendor/google.golang.org/grpc/benchmark/grpc_testing/messages.pb.go index 72cd5f77a..9354eeaa0 100644 --- a/vendor/google.golang.org/grpc/benchmark/grpc_testing/messages.pb.go +++ b/vendor/google.golang.org/grpc/benchmark/grpc_testing/messages.pb.go @@ -79,7 +79,7 @@ func (CompressionType) EnumDescriptor() ([]byte, []int) { // A block of data, to simply increase gRPC message size. type Payload struct { // The type of data in body. - Type PayloadType `protobuf:"varint,1,opt,name=type,enum=grpc.testing.PayloadType" json:"type,omitempty"` + Type PayloadType `protobuf:"varint,1,opt,name=type,proto3,enum=grpc.testing.PayloadType" json:"type,omitempty"` // Primary contents of payload. Body []byte `protobuf:"bytes,2,opt,name=body,proto3" json:"body,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -128,8 +128,8 @@ func (m *Payload) GetBody() []byte { // A protobuf representation for grpc status. This is used by test // clients to specify a status that the server should attempt to return. type EchoStatus struct { - Code int32 `protobuf:"varint,1,opt,name=code" json:"code,omitempty"` - Message string `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Code int32 `protobuf:"varint,1,opt,name=code,proto3" json:"code,omitempty"` + Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -177,20 +177,20 @@ func (m *EchoStatus) GetMessage() string { type SimpleRequest struct { // Desired payload type in the response from the server. // If response_type is RANDOM, server randomly chooses one from other formats. - ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` + ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,proto3,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` // Desired payload size in the response from the server. // If response_type is COMPRESSABLE, this denotes the size before compression. - ResponseSize int32 `protobuf:"varint,2,opt,name=response_size,json=responseSize" json:"response_size,omitempty"` + ResponseSize int32 `protobuf:"varint,2,opt,name=response_size,json=responseSize,proto3" json:"response_size,omitempty"` // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` // Whether SimpleResponse should include username. - FillUsername bool `protobuf:"varint,4,opt,name=fill_username,json=fillUsername" json:"fill_username,omitempty"` + FillUsername bool `protobuf:"varint,4,opt,name=fill_username,json=fillUsername,proto3" json:"fill_username,omitempty"` // Whether SimpleResponse should include OAuth scope. - FillOauthScope bool `protobuf:"varint,5,opt,name=fill_oauth_scope,json=fillOauthScope" json:"fill_oauth_scope,omitempty"` + FillOauthScope bool `protobuf:"varint,5,opt,name=fill_oauth_scope,json=fillOauthScope,proto3" json:"fill_oauth_scope,omitempty"` // Compression algorithm to be used by the server for the response (stream) - ResponseCompression CompressionType `protobuf:"varint,6,opt,name=response_compression,json=responseCompression,enum=grpc.testing.CompressionType" json:"response_compression,omitempty"` + ResponseCompression CompressionType `protobuf:"varint,6,opt,name=response_compression,json=responseCompression,proto3,enum=grpc.testing.CompressionType" json:"response_compression,omitempty"` // Whether server should return a given status - ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus" json:"response_status,omitempty"` + ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus,proto3" json:"response_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -272,12 +272,12 @@ func (m *SimpleRequest) GetResponseStatus() *EchoStatus { // Unary response, as configured by the request. type SimpleResponse struct { // Payload to increase message size. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` // The user the request came from, for verifying authentication was // successful when the client expected it. - Username string `protobuf:"bytes,2,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,2,opt,name=username,proto3" json:"username,omitempty"` // OAuth scope. - OauthScope string `protobuf:"bytes,3,opt,name=oauth_scope,json=oauthScope" json:"oauth_scope,omitempty"` + OauthScope string `protobuf:"bytes,3,opt,name=oauth_scope,json=oauthScope,proto3" json:"oauth_scope,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -331,7 +331,7 @@ func (m *SimpleResponse) GetOauthScope() string { // Client-streaming request. type StreamingInputCallRequest struct { // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -371,7 +371,7 @@ func (m *StreamingInputCallRequest) GetPayload() *Payload { // Client-streaming response. type StreamingInputCallResponse struct { // Aggregated size of payloads received from the client. - AggregatedPayloadSize int32 `protobuf:"varint,1,opt,name=aggregated_payload_size,json=aggregatedPayloadSize" json:"aggregated_payload_size,omitempty"` + AggregatedPayloadSize int32 `protobuf:"varint,1,opt,name=aggregated_payload_size,json=aggregatedPayloadSize,proto3" json:"aggregated_payload_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -412,10 +412,10 @@ func (m *StreamingInputCallResponse) GetAggregatedPayloadSize() int32 { type ResponseParameters struct { // Desired payload sizes in responses from the server. // If response_type is COMPRESSABLE, this denotes the size before compression. - Size int32 `protobuf:"varint,1,opt,name=size" json:"size,omitempty"` + Size int32 `protobuf:"varint,1,opt,name=size,proto3" json:"size,omitempty"` // Desired interval between consecutive responses in the response stream in // microseconds. - IntervalUs int32 `protobuf:"varint,2,opt,name=interval_us,json=intervalUs" json:"interval_us,omitempty"` + IntervalUs int32 `protobuf:"varint,2,opt,name=interval_us,json=intervalUs,proto3" json:"interval_us,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -465,15 +465,15 @@ type StreamingOutputCallRequest struct { // If response_type is RANDOM, the payload from each response in the stream // might be of different types. This is to simulate a mixed type of payload // stream. - ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` + ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,proto3,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` // Configuration for each expected response message. - ResponseParameters []*ResponseParameters `protobuf:"bytes,2,rep,name=response_parameters,json=responseParameters" json:"response_parameters,omitempty"` + ResponseParameters []*ResponseParameters `protobuf:"bytes,2,rep,name=response_parameters,json=responseParameters,proto3" json:"response_parameters,omitempty"` // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` // Compression algorithm to be used by the server for the response (stream) - ResponseCompression CompressionType `protobuf:"varint,6,opt,name=response_compression,json=responseCompression,enum=grpc.testing.CompressionType" json:"response_compression,omitempty"` + ResponseCompression CompressionType `protobuf:"varint,6,opt,name=response_compression,json=responseCompression,proto3,enum=grpc.testing.CompressionType" json:"response_compression,omitempty"` // Whether server should return a given status - ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus" json:"response_status,omitempty"` + ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus,proto3" json:"response_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -541,7 +541,7 @@ func (m *StreamingOutputCallRequest) GetResponseStatus() *EchoStatus { // Server-streaming response, as configured by the request and parameters. type StreamingOutputCallResponse struct { // Payload to increase response size. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -581,7 +581,7 @@ func (m *StreamingOutputCallResponse) GetPayload() *Payload { // For reconnect interop test only. // Client tells server what reconnection parameters it used. type ReconnectParams struct { - MaxReconnectBackoffMs int32 `protobuf:"varint,1,opt,name=max_reconnect_backoff_ms,json=maxReconnectBackoffMs" json:"max_reconnect_backoff_ms,omitempty"` + MaxReconnectBackoffMs int32 `protobuf:"varint,1,opt,name=max_reconnect_backoff_ms,json=maxReconnectBackoffMs,proto3" json:"max_reconnect_backoff_ms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -622,8 +622,8 @@ func (m *ReconnectParams) GetMaxReconnectBackoffMs() int32 { // Server tells client whether its reconnects are following the spec and the // reconnect backoffs it saw. type ReconnectInfo struct { - Passed bool `protobuf:"varint,1,opt,name=passed" json:"passed,omitempty"` - BackoffMs []int32 `protobuf:"varint,2,rep,packed,name=backoff_ms,json=backoffMs" json:"backoff_ms,omitempty"` + Passed bool `protobuf:"varint,1,opt,name=passed,proto3" json:"passed,omitempty"` + BackoffMs []int32 `protobuf:"varint,2,rep,packed,name=backoff_ms,json=backoffMs,proto3" json:"backoff_ms,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/grpc/benchmark/grpc_testing/payloads.pb.go b/vendor/google.golang.org/grpc/benchmark/grpc_testing/payloads.pb.go index 080d7c659..1882fe7e2 100644 --- a/vendor/google.golang.org/grpc/benchmark/grpc_testing/payloads.pb.go +++ b/vendor/google.golang.org/grpc/benchmark/grpc_testing/payloads.pb.go @@ -19,8 +19,8 @@ var _ = math.Inf const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type ByteBufferParams struct { - ReqSize int32 `protobuf:"varint,1,opt,name=req_size,json=reqSize" json:"req_size,omitempty"` - RespSize int32 `protobuf:"varint,2,opt,name=resp_size,json=respSize" json:"resp_size,omitempty"` + ReqSize int32 `protobuf:"varint,1,opt,name=req_size,json=reqSize,proto3" json:"req_size,omitempty"` + RespSize int32 `protobuf:"varint,2,opt,name=resp_size,json=respSize,proto3" json:"resp_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -65,8 +65,8 @@ func (m *ByteBufferParams) GetRespSize() int32 { } type SimpleProtoParams struct { - ReqSize int32 `protobuf:"varint,1,opt,name=req_size,json=reqSize" json:"req_size,omitempty"` - RespSize int32 `protobuf:"varint,2,opt,name=resp_size,json=respSize" json:"resp_size,omitempty"` + ReqSize int32 `protobuf:"varint,1,opt,name=req_size,json=reqSize,proto3" json:"req_size,omitempty"` + RespSize int32 `protobuf:"varint,2,opt,name=resp_size,json=respSize,proto3" json:"resp_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -180,13 +180,13 @@ type isPayloadConfig_Payload interface { } type PayloadConfig_BytebufParams struct { - BytebufParams *ByteBufferParams `protobuf:"bytes,1,opt,name=bytebuf_params,json=bytebufParams,oneof"` + BytebufParams *ByteBufferParams `protobuf:"bytes,1,opt,name=bytebuf_params,json=bytebufParams,proto3,oneof"` } type PayloadConfig_SimpleParams struct { - SimpleParams *SimpleProtoParams `protobuf:"bytes,2,opt,name=simple_params,json=simpleParams,oneof"` + SimpleParams *SimpleProtoParams `protobuf:"bytes,2,opt,name=simple_params,json=simpleParams,proto3,oneof"` } type PayloadConfig_ComplexParams struct { - ComplexParams *ComplexProtoParams `protobuf:"bytes,3,opt,name=complex_params,json=complexParams,oneof"` + ComplexParams *ComplexProtoParams `protobuf:"bytes,3,opt,name=complex_params,json=complexParams,proto3,oneof"` } func (*PayloadConfig_BytebufParams) isPayloadConfig_Payload() {} diff --git a/vendor/google.golang.org/grpc/benchmark/grpc_testing/services.pb.go b/vendor/google.golang.org/grpc/benchmark/grpc_testing/services.pb.go index bd602c1cf..5156ae733 100644 --- a/vendor/google.golang.org/grpc/benchmark/grpc_testing/services.pb.go +++ b/vendor/google.golang.org/grpc/benchmark/grpc_testing/services.pb.go @@ -31,8 +31,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for BenchmarkService service - +// BenchmarkServiceClient is the client API for BenchmarkService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type BenchmarkServiceClient interface { // One request followed by one response. // The server returns the client payload as-is. @@ -52,7 +53,7 @@ func NewBenchmarkServiceClient(cc *grpc.ClientConn) BenchmarkServiceClient { func (c *benchmarkServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, opts ...grpc.CallOption) (*SimpleResponse, error) { out := new(SimpleResponse) - err := grpc.Invoke(ctx, "/grpc.testing.BenchmarkService/UnaryCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.BenchmarkService/UnaryCall", in, out, opts...) if err != nil { return nil, err } @@ -60,7 +61,7 @@ func (c *benchmarkServiceClient) UnaryCall(ctx context.Context, in *SimpleReques } func (c *benchmarkServiceClient) StreamingCall(ctx context.Context, opts ...grpc.CallOption) (BenchmarkService_StreamingCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_BenchmarkService_serviceDesc.Streams[0], c.cc, "/grpc.testing.BenchmarkService/StreamingCall", opts...) + stream, err := c.cc.NewStream(ctx, &_BenchmarkService_serviceDesc.Streams[0], "/grpc.testing.BenchmarkService/StreamingCall", opts...) if err != nil { return nil, err } @@ -90,8 +91,7 @@ func (x *benchmarkServiceStreamingCallClient) Recv() (*SimpleResponse, error) { return m, nil } -// Server API for BenchmarkService service - +// BenchmarkServiceServer is the server API for BenchmarkService service. type BenchmarkServiceServer interface { // One request followed by one response. // The server returns the client payload as-is. @@ -169,8 +169,9 @@ var _BenchmarkService_serviceDesc = grpc.ServiceDesc{ Metadata: "services.proto", } -// Client API for WorkerService service - +// WorkerServiceClient is the client API for WorkerService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type WorkerServiceClient interface { // Start server with specified workload. // First request sent specifies the ServerConfig followed by ServerStatus @@ -201,7 +202,7 @@ func NewWorkerServiceClient(cc *grpc.ClientConn) WorkerServiceClient { } func (c *workerServiceClient) RunServer(ctx context.Context, opts ...grpc.CallOption) (WorkerService_RunServerClient, error) { - stream, err := grpc.NewClientStream(ctx, &_WorkerService_serviceDesc.Streams[0], c.cc, "/grpc.testing.WorkerService/RunServer", opts...) + stream, err := c.cc.NewStream(ctx, &_WorkerService_serviceDesc.Streams[0], "/grpc.testing.WorkerService/RunServer", opts...) if err != nil { return nil, err } @@ -232,7 +233,7 @@ func (x *workerServiceRunServerClient) Recv() (*ServerStatus, error) { } func (c *workerServiceClient) RunClient(ctx context.Context, opts ...grpc.CallOption) (WorkerService_RunClientClient, error) { - stream, err := grpc.NewClientStream(ctx, &_WorkerService_serviceDesc.Streams[1], c.cc, "/grpc.testing.WorkerService/RunClient", opts...) + stream, err := c.cc.NewStream(ctx, &_WorkerService_serviceDesc.Streams[1], "/grpc.testing.WorkerService/RunClient", opts...) if err != nil { return nil, err } @@ -264,7 +265,7 @@ func (x *workerServiceRunClientClient) Recv() (*ClientStatus, error) { func (c *workerServiceClient) CoreCount(ctx context.Context, in *CoreRequest, opts ...grpc.CallOption) (*CoreResponse, error) { out := new(CoreResponse) - err := grpc.Invoke(ctx, "/grpc.testing.WorkerService/CoreCount", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.WorkerService/CoreCount", in, out, opts...) if err != nil { return nil, err } @@ -273,15 +274,14 @@ func (c *workerServiceClient) CoreCount(ctx context.Context, in *CoreRequest, op func (c *workerServiceClient) QuitWorker(ctx context.Context, in *Void, opts ...grpc.CallOption) (*Void, error) { out := new(Void) - err := grpc.Invoke(ctx, "/grpc.testing.WorkerService/QuitWorker", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.WorkerService/QuitWorker", in, out, opts...) if err != nil { return nil, err } return out, nil } -// Server API for WorkerService service - +// WorkerServiceServer is the server API for WorkerService service. type WorkerServiceServer interface { // Start server with specified workload. // First request sent specifies the ServerConfig followed by ServerStatus diff --git a/vendor/google.golang.org/grpc/benchmark/grpc_testing/stats.pb.go b/vendor/google.golang.org/grpc/benchmark/grpc_testing/stats.pb.go index 91d188702..abc2aea76 100644 --- a/vendor/google.golang.org/grpc/benchmark/grpc_testing/stats.pb.go +++ b/vendor/google.golang.org/grpc/benchmark/grpc_testing/stats.pb.go @@ -20,12 +20,12 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type ServerStats struct { // wall clock time change in seconds since last reset - TimeElapsed float64 `protobuf:"fixed64,1,opt,name=time_elapsed,json=timeElapsed" json:"time_elapsed,omitempty"` + TimeElapsed float64 `protobuf:"fixed64,1,opt,name=time_elapsed,json=timeElapsed,proto3" json:"time_elapsed,omitempty"` // change in user time (in seconds) used by the server since last reset - TimeUser float64 `protobuf:"fixed64,2,opt,name=time_user,json=timeUser" json:"time_user,omitempty"` + TimeUser float64 `protobuf:"fixed64,2,opt,name=time_user,json=timeUser,proto3" json:"time_user,omitempty"` // change in server time (in seconds) used by the server process and all // threads since last reset - TimeSystem float64 `protobuf:"fixed64,3,opt,name=time_system,json=timeSystem" json:"time_system,omitempty"` + TimeSystem float64 `protobuf:"fixed64,3,opt,name=time_system,json=timeSystem,proto3" json:"time_system,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -78,8 +78,8 @@ func (m *ServerStats) GetTimeSystem() float64 { // Histogram params based on grpc/support/histogram.c type HistogramParams struct { - Resolution float64 `protobuf:"fixed64,1,opt,name=resolution" json:"resolution,omitempty"` - MaxPossible float64 `protobuf:"fixed64,2,opt,name=max_possible,json=maxPossible" json:"max_possible,omitempty"` + Resolution float64 `protobuf:"fixed64,1,opt,name=resolution,proto3" json:"resolution,omitempty"` + MaxPossible float64 `protobuf:"fixed64,2,opt,name=max_possible,json=maxPossible,proto3" json:"max_possible,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -125,12 +125,12 @@ func (m *HistogramParams) GetMaxPossible() float64 { // Histogram data based on grpc/support/histogram.c type HistogramData struct { - Bucket []uint32 `protobuf:"varint,1,rep,packed,name=bucket" json:"bucket,omitempty"` - MinSeen float64 `protobuf:"fixed64,2,opt,name=min_seen,json=minSeen" json:"min_seen,omitempty"` - MaxSeen float64 `protobuf:"fixed64,3,opt,name=max_seen,json=maxSeen" json:"max_seen,omitempty"` - Sum float64 `protobuf:"fixed64,4,opt,name=sum" json:"sum,omitempty"` - SumOfSquares float64 `protobuf:"fixed64,5,opt,name=sum_of_squares,json=sumOfSquares" json:"sum_of_squares,omitempty"` - Count float64 `protobuf:"fixed64,6,opt,name=count" json:"count,omitempty"` + Bucket []uint32 `protobuf:"varint,1,rep,packed,name=bucket,proto3" json:"bucket,omitempty"` + MinSeen float64 `protobuf:"fixed64,2,opt,name=min_seen,json=minSeen,proto3" json:"min_seen,omitempty"` + MaxSeen float64 `protobuf:"fixed64,3,opt,name=max_seen,json=maxSeen,proto3" json:"max_seen,omitempty"` + Sum float64 `protobuf:"fixed64,4,opt,name=sum,proto3" json:"sum,omitempty"` + SumOfSquares float64 `protobuf:"fixed64,5,opt,name=sum_of_squares,json=sumOfSquares,proto3" json:"sum_of_squares,omitempty"` + Count float64 `protobuf:"fixed64,6,opt,name=count,proto3" json:"count,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -204,11 +204,11 @@ func (m *HistogramData) GetCount() float64 { type ClientStats struct { // Latency histogram. Data points are in nanoseconds. - Latencies *HistogramData `protobuf:"bytes,1,opt,name=latencies" json:"latencies,omitempty"` + Latencies *HistogramData `protobuf:"bytes,1,opt,name=latencies,proto3" json:"latencies,omitempty"` // See ServerStats for details. - TimeElapsed float64 `protobuf:"fixed64,2,opt,name=time_elapsed,json=timeElapsed" json:"time_elapsed,omitempty"` - TimeUser float64 `protobuf:"fixed64,3,opt,name=time_user,json=timeUser" json:"time_user,omitempty"` - TimeSystem float64 `protobuf:"fixed64,4,opt,name=time_system,json=timeSystem" json:"time_system,omitempty"` + TimeElapsed float64 `protobuf:"fixed64,2,opt,name=time_elapsed,json=timeElapsed,proto3" json:"time_elapsed,omitempty"` + TimeUser float64 `protobuf:"fixed64,3,opt,name=time_user,json=timeUser,proto3" json:"time_user,omitempty"` + TimeSystem float64 `protobuf:"fixed64,4,opt,name=time_system,json=timeSystem,proto3" json:"time_system,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` diff --git a/vendor/google.golang.org/grpc/benchmark/run_bench.sh b/vendor/google.golang.org/grpc/benchmark/run_bench.sh new file mode 100755 index 000000000..045de1602 --- /dev/null +++ b/vendor/google.golang.org/grpc/benchmark/run_bench.sh @@ -0,0 +1,187 @@ +#!/bin/bash + +rpcs=(1) +conns=(1) +warmup=10 +dur=10 +reqs=(1) +resps=(1) +rpc_types=(unary) + +# idx[0] = idx value for rpcs +# idx[1] = idx value for conns +# idx[2] = idx value for reqs +# idx[3] = idx value for resps +# idx[4] = idx value for rpc_types +idx=(0 0 0 0 0) +idx_max=(1 1 1 1 1) + +inc() +{ + for i in $(seq $((${#idx[@]}-1)) -1 0); do + idx[${i}]=$((${idx[${i}]}+1)) + if [ ${idx[${i}]} == ${idx_max[${i}]} ]; then + idx[${i}]=0 + else + break + fi + done + local fin + fin=1 + # Check to see if we have looped back to the beginning. + for v in ${idx[@]}; do + if [ ${v} != 0 ]; then + fin=0 + break + fi + done + if [ ${fin} == 1 ]; then + rm -Rf ${out_dir} + clean_and_die 0 + fi +} + +clean_and_die() { + rm -Rf ${out_dir} + exit $1 +} + +run(){ + local nr + nr=${rpcs[${idx[0]}]} + local nc + nc=${conns[${idx[1]}]} + req_sz=${reqs[${idx[2]}]} + resp_sz=${resps[${idx[3]}]} + r_type=${rpc_types[${idx[4]}]} + # Following runs one benchmark + base_port=50051 + delta=0 + test_name="r_"${nr}"_c_"${nc}"_req_"${req_sz}"_resp_"${resp_sz}"_"${r_type}"_"$(date +%s) + echo "================================================================================" + echo ${test_name} + while : + do + port=$((${base_port}+${delta})) + + # Launch the server in background + ${out_dir}/server --port=${port} --test_name="Server_"${test_name}& + server_pid=$(echo $!) + + # Launch the client + ${out_dir}/client --port=${port} --d=${dur} --w=${warmup} --r=${nr} --c=${nc} --req=${req_sz} --resp=${resp_sz} --rpc_type=${r_type} --test_name="client_"${test_name} + client_status=$(echo $?) + + kill ${server_pid} + wait ${server_pid} + + if [ ${client_status} == 0 ]; then + break + fi + + delta=$((${delta}+1)) + if [ ${delta} == 10 ]; then + echo "Continuous 10 failed runs. Exiting now." + rm -Rf ${out_dir} + clean_and_die 1 + fi + done + +} + +set_param(){ + local argname=$1 + shift + local idx=$1 + shift + if [ $# -eq 0 ]; then + echo "${argname} not specified" + exit 1 + fi + PARAM=($(echo $1 | sed 's/,/ /g')) + if [ ${idx} -lt 0 ]; then + return + fi + idx_max[${idx}]=${#PARAM[@]} +} + +while [ $# -gt 0 ]; do + case "$1" in + -r) + shift + set_param "number of rpcs" 0 $1 + rpcs=(${PARAM[@]}) + shift + ;; + -c) + shift + set_param "number of connections" 1 $1 + conns=(${PARAM[@]}) + shift + ;; + -w) + shift + set_param "warm-up period" -1 $1 + warmup=${PARAM} + shift + ;; + -d) + shift + set_param "duration" -1 $1 + dur=${PARAM} + shift + ;; + -req) + shift + set_param "request size" 2 $1 + reqs=(${PARAM[@]}) + shift + ;; + -resp) + shift + set_param "response size" 3 $1 + resps=(${PARAM[@]}) + shift + ;; + -rpc_type) + shift + set_param "rpc type" 4 $1 + rpc_types=(${PARAM[@]}) + shift + ;; + -h|--help) + echo "Following are valid options:" + echo + echo "-h, --help show brief help" + echo "-w warm-up duration in seconds, default value is 10" + echo "-d benchmark duration in seconds, default value is 60" + echo "" + echo "Each of the following can have multiple comma separated values." + echo "" + echo "-r number of RPCs, default value is 1" + echo "-c number of Connections, default value is 1" + echo "-req req size in bytes, default value is 1" + echo "-resp resp size in bytes, default value is 1" + echo "-rpc_type valid values are unary|streaming, default is unary" + ;; + *) + echo "Incorrect option $1" + exit 1 + ;; + esac +done + +# Build server and client +out_dir=$(mktemp -d oss_benchXXX) + +go build -o ${out_dir}/server $GOPATH/src/google.golang.org/grpc/benchmark/server/main.go && go build -o ${out_dir}/client $GOPATH/src/google.golang.org/grpc/benchmark/client/main.go +if [ $? != 0 ]; then + clean_and_die 1 +fi + + +while : +do + run + inc +done diff --git a/vendor/google.golang.org/grpc/benchmark/server/main.go b/vendor/google.golang.org/grpc/benchmark/server/main.go index 6bcfe71f0..253657e0e 100644 --- a/vendor/google.golang.org/grpc/benchmark/server/main.go +++ b/vendor/google.golang.org/grpc/benchmark/server/main.go @@ -20,37 +20,71 @@ package main import ( "flag" - "math" + "fmt" "net" - "net/http" _ "net/http/pprof" + "os" + "os/signal" + "runtime" + "runtime/pprof" + "syscall" "time" + "golang.org/x/sys/unix" "google.golang.org/grpc/benchmark" "google.golang.org/grpc/grpclog" ) -var duration = flag.Int("duration", math.MaxInt32, "The duration in seconds to run the benchmark server") +var ( + port = flag.String("port", "50051", "Localhost port to listen on.") + testName = flag.String("test_name", "", "Name of the test used for creating profiles.") +) func main() { flag.Parse() - go func() { - lis, err := net.Listen("tcp", ":0") - if err != nil { - grpclog.Fatalf("Failed to listen: %v", err) - } - grpclog.Infoln("Server profiling address: ", lis.Addr().String()) - if err := http.Serve(lis, nil); err != nil { - grpclog.Fatalf("Failed to serve: %v", err) - } - }() - lis, err := net.Listen("tcp", "localhost:0") + if *testName == "" { + grpclog.Fatalf("test name not set") + } + lis, err := net.Listen("tcp", ":"+*port) if err != nil { grpclog.Fatalf("Failed to listen: %v", err) } - addr := lis.Addr().String() - stopper := benchmark.StartServer(benchmark.ServerInfo{Type: "protobuf", Listener: lis}) // listen on all interfaces - grpclog.Infoln("Server Address: ", addr) - <-time.After(time.Duration(*duration) * time.Second) - stopper() + defer lis.Close() + + cf, err := os.Create("/tmp/" + *testName + ".cpu") + if err != nil { + grpclog.Fatalf("Failed to create file: %v", err) + } + defer cf.Close() + pprof.StartCPUProfile(cf) + cpuBeg := getCPUTime() + // Launch server in a separate goroutine. + stop := benchmark.StartServer(benchmark.ServerInfo{Type: "protobuf", Listener: lis}) + // Wait on OS terminate signal. + ch := make(chan os.Signal, 1) + signal.Notify(ch, syscall.SIGTERM) + <-ch + cpu := time.Duration(getCPUTime() - cpuBeg) + stop() + pprof.StopCPUProfile() + mf, err := os.Create("/tmp/" + *testName + ".mem") + if err != nil { + grpclog.Fatalf("Failed to create file: %v", err) + } + defer mf.Close() + runtime.GC() // materialize all statistics + if err := pprof.WriteHeapProfile(mf); err != nil { + grpclog.Fatalf("Failed to write memory profile: %v", err) + } + fmt.Println("Server CPU utilization:", cpu) + fmt.Println("Server CPU profile:", cf.Name()) + fmt.Println("Server Mem Profile:", mf.Name()) +} + +func getCPUTime() int64 { + var ts unix.Timespec + if err := unix.ClockGettime(unix.CLOCK_PROCESS_CPUTIME_ID, &ts); err != nil { + grpclog.Fatal(err) + } + return ts.Nano() } diff --git a/vendor/google.golang.org/grpc/benchmark/worker/benchmark_server.go b/vendor/google.golang.org/grpc/benchmark/worker/benchmark_server.go index ab311e1ba..9b1f676fa 100644 --- a/vendor/google.golang.org/grpc/benchmark/worker/benchmark_server.go +++ b/vendor/google.golang.org/grpc/benchmark/worker/benchmark_server.go @@ -112,7 +112,7 @@ func startBenchmarkServer(config *testpb.ServerConfig, serverPort int) (*benchma if port == 0 { port = serverPort } - lis, err := net.Listen("tcp", fmt.Sprintf("localhost:%d", port)) + lis, err := net.Listen("tcp", fmt.Sprintf(":%d", port)) if err != nil { grpclog.Fatalf("Failed to listen: %v", err) } diff --git a/vendor/google.golang.org/grpc/call_test.go b/vendor/google.golang.org/grpc/call_test.go index 2bcea807f..d57d01e4a 100644 --- a/vendor/google.golang.org/grpc/call_test.go +++ b/vendor/google.golang.org/grpc/call_test.go @@ -31,8 +31,8 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/codes" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/status" - "google.golang.org/grpc/test/leakcheck" "google.golang.org/grpc/transport" ) @@ -105,12 +105,13 @@ func (h *testStreamHandler) handleStream(t *testing.T, s *transport.Stream) { } } // send a response back to end the stream. - hdr, data, err := encode(testCodec{}, &expectedResponse, nil, nil, nil) + data, err := encode(testCodec{}, &expectedResponse) if err != nil { t.Errorf("Failed to encode the response: %v", err) return } - h.t.Write(s, hdr, data, &transport.Options{}) + hdr, payload := msgHeader(data, nil) + h.t.Write(s, hdr, payload, &transport.Options{}) h.t.WriteStatus(s, status.New(codes.OK, "")) } diff --git a/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.pb.go b/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.pb.go index 4722c0bc3..747b8f833 100644 --- a/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.pb.go +++ b/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.pb.go @@ -1,5 +1,5 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: channelz.proto +// source: grpc/channelz/v1/channelz.proto package grpc_channelz_v1 // import "google.golang.org/grpc/channelz/grpc_channelz_v1" @@ -59,26 +59,56 @@ func (x ChannelConnectivityState_State) String() string { return proto.EnumName(ChannelConnectivityState_State_name, int32(x)) } func (ChannelConnectivityState_State) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{2, 0} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{2, 0} +} + +// The supported severity levels of trace events. +type ChannelTraceEvent_Severity int32 + +const ( + ChannelTraceEvent_CT_UNKNOWN ChannelTraceEvent_Severity = 0 + ChannelTraceEvent_CT_INFO ChannelTraceEvent_Severity = 1 + ChannelTraceEvent_CT_WARNING ChannelTraceEvent_Severity = 2 + ChannelTraceEvent_CT_ERROR ChannelTraceEvent_Severity = 3 +) + +var ChannelTraceEvent_Severity_name = map[int32]string{ + 0: "CT_UNKNOWN", + 1: "CT_INFO", + 2: "CT_WARNING", + 3: "CT_ERROR", +} +var ChannelTraceEvent_Severity_value = map[string]int32{ + "CT_UNKNOWN": 0, + "CT_INFO": 1, + "CT_WARNING": 2, + "CT_ERROR": 3, +} + +func (x ChannelTraceEvent_Severity) String() string { + return proto.EnumName(ChannelTraceEvent_Severity_name, int32(x)) +} +func (ChannelTraceEvent_Severity) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{4, 0} } // Channel is a logical grouping of channels, subchannels, and sockets. type Channel struct { // The identifier for this channel. This should bet set. - Ref *ChannelRef `protobuf:"bytes,1,opt,name=ref" json:"ref,omitempty"` + Ref *ChannelRef `protobuf:"bytes,1,opt,name=ref,proto3" json:"ref,omitempty"` // Data specific to this channel. - Data *ChannelData `protobuf:"bytes,2,opt,name=data" json:"data,omitempty"` + Data *ChannelData `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` // There are no ordering guarantees on the order of channel refs. // There may not be cycles in the ref graph. // A channel ref may be present in more than one channel or subchannel. - ChannelRef []*ChannelRef `protobuf:"bytes,3,rep,name=channel_ref,json=channelRef" json:"channel_ref,omitempty"` + ChannelRef []*ChannelRef `protobuf:"bytes,3,rep,name=channel_ref,json=channelRef,proto3" json:"channel_ref,omitempty"` // At most one of 'channel_ref+subchannel_ref' and 'socket' is set. // There are no ordering guarantees on the order of subchannel refs. // There may not be cycles in the ref graph. // A sub channel ref may be present in more than one channel or subchannel. - SubchannelRef []*SubchannelRef `protobuf:"bytes,4,rep,name=subchannel_ref,json=subchannelRef" json:"subchannel_ref,omitempty"` + SubchannelRef []*SubchannelRef `protobuf:"bytes,4,rep,name=subchannel_ref,json=subchannelRef,proto3" json:"subchannel_ref,omitempty"` // There are no ordering guarantees on the order of sockets. - SocketRef []*SocketRef `protobuf:"bytes,5,rep,name=socket_ref,json=socketRef" json:"socket_ref,omitempty"` + SocketRef []*SocketRef `protobuf:"bytes,5,rep,name=socket_ref,json=socketRef,proto3" json:"socket_ref,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -88,7 +118,7 @@ func (m *Channel) Reset() { *m = Channel{} } func (m *Channel) String() string { return proto.CompactTextString(m) } func (*Channel) ProtoMessage() {} func (*Channel) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{0} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{0} } func (m *Channel) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Channel.Unmarshal(m, b) @@ -147,20 +177,20 @@ func (m *Channel) GetSocketRef() []*SocketRef { // A subchannel is load balanced over by it's ancestor type Subchannel struct { // The identifier for this channel. - Ref *SubchannelRef `protobuf:"bytes,1,opt,name=ref" json:"ref,omitempty"` + Ref *SubchannelRef `protobuf:"bytes,1,opt,name=ref,proto3" json:"ref,omitempty"` // Data specific to this channel. - Data *ChannelData `protobuf:"bytes,2,opt,name=data" json:"data,omitempty"` + Data *ChannelData `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` // There are no ordering guarantees on the order of channel refs. // There may not be cycles in the ref graph. // A channel ref may be present in more than one channel or subchannel. - ChannelRef []*ChannelRef `protobuf:"bytes,3,rep,name=channel_ref,json=channelRef" json:"channel_ref,omitempty"` + ChannelRef []*ChannelRef `protobuf:"bytes,3,rep,name=channel_ref,json=channelRef,proto3" json:"channel_ref,omitempty"` // At most one of 'channel_ref+subchannel_ref' and 'socket' is set. // There are no ordering guarantees on the order of subchannel refs. // There may not be cycles in the ref graph. // A sub channel ref may be present in more than one channel or subchannel. - SubchannelRef []*SubchannelRef `protobuf:"bytes,4,rep,name=subchannel_ref,json=subchannelRef" json:"subchannel_ref,omitempty"` + SubchannelRef []*SubchannelRef `protobuf:"bytes,4,rep,name=subchannel_ref,json=subchannelRef,proto3" json:"subchannel_ref,omitempty"` // There are no ordering guarantees on the order of sockets. - SocketRef []*SocketRef `protobuf:"bytes,5,rep,name=socket_ref,json=socketRef" json:"socket_ref,omitempty"` + SocketRef []*SocketRef `protobuf:"bytes,5,rep,name=socket_ref,json=socketRef,proto3" json:"socket_ref,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -170,7 +200,7 @@ func (m *Subchannel) Reset() { *m = Subchannel{} } func (m *Subchannel) String() string { return proto.CompactTextString(m) } func (*Subchannel) ProtoMessage() {} func (*Subchannel) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{1} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{1} } func (m *Subchannel) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Subchannel.Unmarshal(m, b) @@ -228,7 +258,7 @@ func (m *Subchannel) GetSocketRef() []*SocketRef { // These come from the specified states in this document: // https://github.com/grpc/grpc/blob/master/doc/connectivity-semantics-and-api.md type ChannelConnectivityState struct { - State ChannelConnectivityState_State `protobuf:"varint,1,opt,name=state,enum=grpc.channelz.v1.ChannelConnectivityState_State" json:"state,omitempty"` + State ChannelConnectivityState_State `protobuf:"varint,1,opt,name=state,proto3,enum=grpc.channelz.v1.ChannelConnectivityState_State" json:"state,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -238,7 +268,7 @@ func (m *ChannelConnectivityState) Reset() { *m = ChannelConnectivitySta func (m *ChannelConnectivityState) String() string { return proto.CompactTextString(m) } func (*ChannelConnectivityState) ProtoMessage() {} func (*ChannelConnectivityState) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{2} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{2} } func (m *ChannelConnectivityState) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ChannelConnectivityState.Unmarshal(m, b) @@ -269,19 +299,19 @@ func (m *ChannelConnectivityState) GetState() ChannelConnectivityState_State { type ChannelData struct { // The connectivity state of the channel or subchannel. Implementations // should always set this. - State *ChannelConnectivityState `protobuf:"bytes,1,opt,name=state" json:"state,omitempty"` + State *ChannelConnectivityState `protobuf:"bytes,1,opt,name=state,proto3" json:"state,omitempty"` // The target this channel originally tried to connect to. May be absent - Target string `protobuf:"bytes,2,opt,name=target" json:"target,omitempty"` + Target string `protobuf:"bytes,2,opt,name=target,proto3" json:"target,omitempty"` // A trace of recent events on the channel. May be absent. - Trace *ChannelTrace `protobuf:"bytes,3,opt,name=trace" json:"trace,omitempty"` + Trace *ChannelTrace `protobuf:"bytes,3,opt,name=trace,proto3" json:"trace,omitempty"` // The number of calls started on the channel - CallsStarted int64 `protobuf:"varint,4,opt,name=calls_started,json=callsStarted" json:"calls_started,omitempty"` + CallsStarted int64 `protobuf:"varint,4,opt,name=calls_started,json=callsStarted,proto3" json:"calls_started,omitempty"` // The number of calls that have completed with an OK status - CallsSucceeded int64 `protobuf:"varint,5,opt,name=calls_succeeded,json=callsSucceeded" json:"calls_succeeded,omitempty"` + CallsSucceeded int64 `protobuf:"varint,5,opt,name=calls_succeeded,json=callsSucceeded,proto3" json:"calls_succeeded,omitempty"` // The number of calls that have completed with a non-OK status - CallsFailed int64 `protobuf:"varint,6,opt,name=calls_failed,json=callsFailed" json:"calls_failed,omitempty"` + CallsFailed int64 `protobuf:"varint,6,opt,name=calls_failed,json=callsFailed,proto3" json:"calls_failed,omitempty"` // The last time a call was started on the channel. - LastCallStartedTimestamp *timestamp.Timestamp `protobuf:"bytes,7,opt,name=last_call_started_timestamp,json=lastCallStartedTimestamp" json:"last_call_started_timestamp,omitempty"` + LastCallStartedTimestamp *timestamp.Timestamp `protobuf:"bytes,7,opt,name=last_call_started_timestamp,json=lastCallStartedTimestamp,proto3" json:"last_call_started_timestamp,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -291,7 +321,7 @@ func (m *ChannelData) Reset() { *m = ChannelData{} } func (m *ChannelData) String() string { return proto.CompactTextString(m) } func (*ChannelData) ProtoMessage() {} func (*ChannelData) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{3} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{3} } func (m *ChannelData) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ChannelData.Unmarshal(m, b) @@ -360,18 +390,203 @@ func (m *ChannelData) GetLastCallStartedTimestamp() *timestamp.Timestamp { return nil } +// A trace event is an interesting thing that happened to a channel or +// subchannel, such as creation, address resolution, subchannel creation, etc. +type ChannelTraceEvent struct { + // High level description of the event. + Description string `protobuf:"bytes,1,opt,name=description,proto3" json:"description,omitempty"` + // the severity of the trace event + Severity ChannelTraceEvent_Severity `protobuf:"varint,2,opt,name=severity,proto3,enum=grpc.channelz.v1.ChannelTraceEvent_Severity" json:"severity,omitempty"` + // When this event occurred. + Timestamp *timestamp.Timestamp `protobuf:"bytes,3,opt,name=timestamp,proto3" json:"timestamp,omitempty"` + // ref of referenced channel or subchannel. + // Optional, only present if this event refers to a child object. For example, + // this field would be filled if this trace event was for a subchannel being + // created. + // + // Types that are valid to be assigned to ChildRef: + // *ChannelTraceEvent_ChannelRef + // *ChannelTraceEvent_SubchannelRef + ChildRef isChannelTraceEvent_ChildRef `protobuf_oneof:"child_ref"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ChannelTraceEvent) Reset() { *m = ChannelTraceEvent{} } +func (m *ChannelTraceEvent) String() string { return proto.CompactTextString(m) } +func (*ChannelTraceEvent) ProtoMessage() {} +func (*ChannelTraceEvent) Descriptor() ([]byte, []int) { + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{4} +} +func (m *ChannelTraceEvent) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ChannelTraceEvent.Unmarshal(m, b) +} +func (m *ChannelTraceEvent) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ChannelTraceEvent.Marshal(b, m, deterministic) +} +func (dst *ChannelTraceEvent) XXX_Merge(src proto.Message) { + xxx_messageInfo_ChannelTraceEvent.Merge(dst, src) +} +func (m *ChannelTraceEvent) XXX_Size() int { + return xxx_messageInfo_ChannelTraceEvent.Size(m) +} +func (m *ChannelTraceEvent) XXX_DiscardUnknown() { + xxx_messageInfo_ChannelTraceEvent.DiscardUnknown(m) +} + +var xxx_messageInfo_ChannelTraceEvent proto.InternalMessageInfo + +type isChannelTraceEvent_ChildRef interface { + isChannelTraceEvent_ChildRef() +} + +type ChannelTraceEvent_ChannelRef struct { + ChannelRef *ChannelRef `protobuf:"bytes,4,opt,name=channel_ref,json=channelRef,proto3,oneof"` +} +type ChannelTraceEvent_SubchannelRef struct { + SubchannelRef *SubchannelRef `protobuf:"bytes,5,opt,name=subchannel_ref,json=subchannelRef,proto3,oneof"` +} + +func (*ChannelTraceEvent_ChannelRef) isChannelTraceEvent_ChildRef() {} +func (*ChannelTraceEvent_SubchannelRef) isChannelTraceEvent_ChildRef() {} + +func (m *ChannelTraceEvent) GetChildRef() isChannelTraceEvent_ChildRef { + if m != nil { + return m.ChildRef + } + return nil +} + +func (m *ChannelTraceEvent) GetDescription() string { + if m != nil { + return m.Description + } + return "" +} + +func (m *ChannelTraceEvent) GetSeverity() ChannelTraceEvent_Severity { + if m != nil { + return m.Severity + } + return ChannelTraceEvent_CT_UNKNOWN +} + +func (m *ChannelTraceEvent) GetTimestamp() *timestamp.Timestamp { + if m != nil { + return m.Timestamp + } + return nil +} + +func (m *ChannelTraceEvent) GetChannelRef() *ChannelRef { + if x, ok := m.GetChildRef().(*ChannelTraceEvent_ChannelRef); ok { + return x.ChannelRef + } + return nil +} + +func (m *ChannelTraceEvent) GetSubchannelRef() *SubchannelRef { + if x, ok := m.GetChildRef().(*ChannelTraceEvent_SubchannelRef); ok { + return x.SubchannelRef + } + return nil +} + +// XXX_OneofFuncs is for the internal use of the proto package. +func (*ChannelTraceEvent) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{}) { + return _ChannelTraceEvent_OneofMarshaler, _ChannelTraceEvent_OneofUnmarshaler, _ChannelTraceEvent_OneofSizer, []interface{}{ + (*ChannelTraceEvent_ChannelRef)(nil), + (*ChannelTraceEvent_SubchannelRef)(nil), + } +} + +func _ChannelTraceEvent_OneofMarshaler(msg proto.Message, b *proto.Buffer) error { + m := msg.(*ChannelTraceEvent) + // child_ref + switch x := m.ChildRef.(type) { + case *ChannelTraceEvent_ChannelRef: + b.EncodeVarint(4<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.ChannelRef); err != nil { + return err + } + case *ChannelTraceEvent_SubchannelRef: + b.EncodeVarint(5<<3 | proto.WireBytes) + if err := b.EncodeMessage(x.SubchannelRef); err != nil { + return err + } + case nil: + default: + return fmt.Errorf("ChannelTraceEvent.ChildRef has unexpected type %T", x) + } + return nil +} + +func _ChannelTraceEvent_OneofUnmarshaler(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error) { + m := msg.(*ChannelTraceEvent) + switch tag { + case 4: // child_ref.channel_ref + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(ChannelRef) + err := b.DecodeMessage(msg) + m.ChildRef = &ChannelTraceEvent_ChannelRef{msg} + return true, err + case 5: // child_ref.subchannel_ref + if wire != proto.WireBytes { + return true, proto.ErrInternalBadWireType + } + msg := new(SubchannelRef) + err := b.DecodeMessage(msg) + m.ChildRef = &ChannelTraceEvent_SubchannelRef{msg} + return true, err + default: + return false, nil + } +} + +func _ChannelTraceEvent_OneofSizer(msg proto.Message) (n int) { + m := msg.(*ChannelTraceEvent) + // child_ref + switch x := m.ChildRef.(type) { + case *ChannelTraceEvent_ChannelRef: + s := proto.Size(x.ChannelRef) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case *ChannelTraceEvent_SubchannelRef: + s := proto.Size(x.SubchannelRef) + n += 1 // tag and wire + n += proto.SizeVarint(uint64(s)) + n += s + case nil: + default: + panic(fmt.Sprintf("proto: unexpected type %T in oneof", x)) + } + return n +} + // ChannelTrace represents the recent events that have occurred on the channel. type ChannelTrace struct { - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + // Number of events ever logged in this tracing object. This can differ from + // events.size() because events can be overwritten or garbage collected by + // implementations. + NumEventsLogged int64 `protobuf:"varint,1,opt,name=num_events_logged,json=numEventsLogged,proto3" json:"num_events_logged,omitempty"` + // Time that this channel was created. + CreationTimestamp *timestamp.Timestamp `protobuf:"bytes,2,opt,name=creation_timestamp,json=creationTimestamp,proto3" json:"creation_timestamp,omitempty"` + // List of events that have occurred on this channel. + Events []*ChannelTraceEvent `protobuf:"bytes,3,rep,name=events,proto3" json:"events,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *ChannelTrace) Reset() { *m = ChannelTrace{} } func (m *ChannelTrace) String() string { return proto.CompactTextString(m) } func (*ChannelTrace) ProtoMessage() {} func (*ChannelTrace) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{4} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{5} } func (m *ChannelTrace) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ChannelTrace.Unmarshal(m, b) @@ -391,12 +606,33 @@ func (m *ChannelTrace) XXX_DiscardUnknown() { var xxx_messageInfo_ChannelTrace proto.InternalMessageInfo +func (m *ChannelTrace) GetNumEventsLogged() int64 { + if m != nil { + return m.NumEventsLogged + } + return 0 +} + +func (m *ChannelTrace) GetCreationTimestamp() *timestamp.Timestamp { + if m != nil { + return m.CreationTimestamp + } + return nil +} + +func (m *ChannelTrace) GetEvents() []*ChannelTraceEvent { + if m != nil { + return m.Events + } + return nil +} + // ChannelRef is a reference to a Channel. type ChannelRef struct { // The globally unique id for this channel. Must be a positive number. - ChannelId int64 `protobuf:"varint,1,opt,name=channel_id,json=channelId" json:"channel_id,omitempty"` + ChannelId int64 `protobuf:"varint,1,opt,name=channel_id,json=channelId,proto3" json:"channel_id,omitempty"` // An optional name associated with the channel. - Name string `protobuf:"bytes,2,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,2,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -406,7 +642,7 @@ func (m *ChannelRef) Reset() { *m = ChannelRef{} } func (m *ChannelRef) String() string { return proto.CompactTextString(m) } func (*ChannelRef) ProtoMessage() {} func (*ChannelRef) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{5} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{6} } func (m *ChannelRef) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ChannelRef.Unmarshal(m, b) @@ -443,9 +679,9 @@ func (m *ChannelRef) GetName() string { // ChannelRef is a reference to a Subchannel. type SubchannelRef struct { // The globally unique id for this subchannel. Must be a positive number. - SubchannelId int64 `protobuf:"varint,7,opt,name=subchannel_id,json=subchannelId" json:"subchannel_id,omitempty"` + SubchannelId int64 `protobuf:"varint,7,opt,name=subchannel_id,json=subchannelId,proto3" json:"subchannel_id,omitempty"` // An optional name associated with the subchannel. - Name string `protobuf:"bytes,8,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,8,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -455,7 +691,7 @@ func (m *SubchannelRef) Reset() { *m = SubchannelRef{} } func (m *SubchannelRef) String() string { return proto.CompactTextString(m) } func (*SubchannelRef) ProtoMessage() {} func (*SubchannelRef) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{6} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{7} } func (m *SubchannelRef) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SubchannelRef.Unmarshal(m, b) @@ -491,9 +727,9 @@ func (m *SubchannelRef) GetName() string { // SocketRef is a reference to a Socket. type SocketRef struct { - SocketId int64 `protobuf:"varint,3,opt,name=socket_id,json=socketId" json:"socket_id,omitempty"` + SocketId int64 `protobuf:"varint,3,opt,name=socket_id,json=socketId,proto3" json:"socket_id,omitempty"` // An optional name associated with the socket. - Name string `protobuf:"bytes,4,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,4,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -503,7 +739,7 @@ func (m *SocketRef) Reset() { *m = SocketRef{} } func (m *SocketRef) String() string { return proto.CompactTextString(m) } func (*SocketRef) ProtoMessage() {} func (*SocketRef) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{7} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{8} } func (m *SocketRef) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SocketRef.Unmarshal(m, b) @@ -540,9 +776,9 @@ func (m *SocketRef) GetName() string { // ServerRef is a reference to a Server. type ServerRef struct { // A globally unique identifier for this server. Must be a positive number. - ServerId int64 `protobuf:"varint,5,opt,name=server_id,json=serverId" json:"server_id,omitempty"` + ServerId int64 `protobuf:"varint,5,opt,name=server_id,json=serverId,proto3" json:"server_id,omitempty"` // An optional name associated with the server. - Name string `protobuf:"bytes,6,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,6,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -552,7 +788,7 @@ func (m *ServerRef) Reset() { *m = ServerRef{} } func (m *ServerRef) String() string { return proto.CompactTextString(m) } func (*ServerRef) ProtoMessage() {} func (*ServerRef) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{8} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{9} } func (m *ServerRef) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ServerRef.Unmarshal(m, b) @@ -590,12 +826,12 @@ func (m *ServerRef) GetName() string { // program. type Server struct { // The identifier for a Server. This should be set. - Ref *ServerRef `protobuf:"bytes,1,opt,name=ref" json:"ref,omitempty"` + Ref *ServerRef `protobuf:"bytes,1,opt,name=ref,proto3" json:"ref,omitempty"` // The associated data of the Server. - Data *ServerData `protobuf:"bytes,2,opt,name=data" json:"data,omitempty"` + Data *ServerData `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` // The sockets that the server is listening on. There are no ordering // guarantees. This may be absent. - ListenSocket []*SocketRef `protobuf:"bytes,3,rep,name=listen_socket,json=listenSocket" json:"listen_socket,omitempty"` + ListenSocket []*SocketRef `protobuf:"bytes,3,rep,name=listen_socket,json=listenSocket,proto3" json:"listen_socket,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -605,7 +841,7 @@ func (m *Server) Reset() { *m = Server{} } func (m *Server) String() string { return proto.CompactTextString(m) } func (*Server) ProtoMessage() {} func (*Server) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{9} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{10} } func (m *Server) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Server.Unmarshal(m, b) @@ -649,15 +885,15 @@ func (m *Server) GetListenSocket() []*SocketRef { // ServerData is data for a specific Server. type ServerData struct { // A trace of recent events on the server. May be absent. - Trace *ChannelTrace `protobuf:"bytes,1,opt,name=trace" json:"trace,omitempty"` + Trace *ChannelTrace `protobuf:"bytes,1,opt,name=trace,proto3" json:"trace,omitempty"` // The number of incoming calls started on the server - CallsStarted int64 `protobuf:"varint,2,opt,name=calls_started,json=callsStarted" json:"calls_started,omitempty"` + CallsStarted int64 `protobuf:"varint,2,opt,name=calls_started,json=callsStarted,proto3" json:"calls_started,omitempty"` // The number of incoming calls that have completed with an OK status - CallsSucceeded int64 `protobuf:"varint,3,opt,name=calls_succeeded,json=callsSucceeded" json:"calls_succeeded,omitempty"` + CallsSucceeded int64 `protobuf:"varint,3,opt,name=calls_succeeded,json=callsSucceeded,proto3" json:"calls_succeeded,omitempty"` // The number of incoming calls that have a completed with a non-OK status - CallsFailed int64 `protobuf:"varint,4,opt,name=calls_failed,json=callsFailed" json:"calls_failed,omitempty"` + CallsFailed int64 `protobuf:"varint,4,opt,name=calls_failed,json=callsFailed,proto3" json:"calls_failed,omitempty"` // The last time a call was started on the server. - LastCallStartedTimestamp *timestamp.Timestamp `protobuf:"bytes,5,opt,name=last_call_started_timestamp,json=lastCallStartedTimestamp" json:"last_call_started_timestamp,omitempty"` + LastCallStartedTimestamp *timestamp.Timestamp `protobuf:"bytes,5,opt,name=last_call_started_timestamp,json=lastCallStartedTimestamp,proto3" json:"last_call_started_timestamp,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -667,7 +903,7 @@ func (m *ServerData) Reset() { *m = ServerData{} } func (m *ServerData) String() string { return proto.CompactTextString(m) } func (*ServerData) ProtoMessage() {} func (*ServerData) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{10} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{11} } func (m *ServerData) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ServerData.Unmarshal(m, b) @@ -725,19 +961,19 @@ func (m *ServerData) GetLastCallStartedTimestamp() *timestamp.Timestamp { // Information about an actual connection. Pronounced "sock-ay". type Socket struct { // The identifier for the Socket. - Ref *SocketRef `protobuf:"bytes,1,opt,name=ref" json:"ref,omitempty"` + Ref *SocketRef `protobuf:"bytes,1,opt,name=ref,proto3" json:"ref,omitempty"` // Data specific to this Socket. - Data *SocketData `protobuf:"bytes,2,opt,name=data" json:"data,omitempty"` + Data *SocketData `protobuf:"bytes,2,opt,name=data,proto3" json:"data,omitempty"` // The locally bound address. - Local *Address `protobuf:"bytes,3,opt,name=local" json:"local,omitempty"` + Local *Address `protobuf:"bytes,3,opt,name=local,proto3" json:"local,omitempty"` // The remote bound address. May be absent. - Remote *Address `protobuf:"bytes,4,opt,name=remote" json:"remote,omitempty"` + Remote *Address `protobuf:"bytes,4,opt,name=remote,proto3" json:"remote,omitempty"` // Security details for this socket. May be absent if not available, or // there is no security on the socket. - Security *Security `protobuf:"bytes,5,opt,name=security" json:"security,omitempty"` + Security *Security `protobuf:"bytes,5,opt,name=security,proto3" json:"security,omitempty"` // Optional, represents the name of the remote endpoint, if different than // the original target name. - RemoteName string `protobuf:"bytes,6,opt,name=remote_name,json=remoteName" json:"remote_name,omitempty"` + RemoteName string `protobuf:"bytes,6,opt,name=remote_name,json=remoteName,proto3" json:"remote_name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -747,7 +983,7 @@ func (m *Socket) Reset() { *m = Socket{} } func (m *Socket) String() string { return proto.CompactTextString(m) } func (*Socket) ProtoMessage() {} func (*Socket) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{11} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{12} } func (m *Socket) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Socket.Unmarshal(m, b) @@ -814,42 +1050,42 @@ func (m *Socket) GetRemoteName() string { // the semantics. (e.g. flow control windows) type SocketData struct { // The number of streams that have been started. - StreamsStarted int64 `protobuf:"varint,1,opt,name=streams_started,json=streamsStarted" json:"streams_started,omitempty"` + StreamsStarted int64 `protobuf:"varint,1,opt,name=streams_started,json=streamsStarted,proto3" json:"streams_started,omitempty"` // The number of streams that have ended successfully: // On client side, received frame with eos bit set; // On server side, sent frame with eos bit set. - StreamsSucceeded int64 `protobuf:"varint,2,opt,name=streams_succeeded,json=streamsSucceeded" json:"streams_succeeded,omitempty"` + StreamsSucceeded int64 `protobuf:"varint,2,opt,name=streams_succeeded,json=streamsSucceeded,proto3" json:"streams_succeeded,omitempty"` // The number of streams that have ended unsuccessfully: // On client side, ended without receiving frame with eos bit set; // On server side, ended without sending frame with eos bit set. - StreamsFailed int64 `protobuf:"varint,3,opt,name=streams_failed,json=streamsFailed" json:"streams_failed,omitempty"` + StreamsFailed int64 `protobuf:"varint,3,opt,name=streams_failed,json=streamsFailed,proto3" json:"streams_failed,omitempty"` // The number of grpc messages successfully sent on this socket. - MessagesSent int64 `protobuf:"varint,4,opt,name=messages_sent,json=messagesSent" json:"messages_sent,omitempty"` + MessagesSent int64 `protobuf:"varint,4,opt,name=messages_sent,json=messagesSent,proto3" json:"messages_sent,omitempty"` // The number of grpc messages received on this socket. - MessagesReceived int64 `protobuf:"varint,5,opt,name=messages_received,json=messagesReceived" json:"messages_received,omitempty"` + MessagesReceived int64 `protobuf:"varint,5,opt,name=messages_received,json=messagesReceived,proto3" json:"messages_received,omitempty"` // The number of keep alives sent. This is typically implemented with HTTP/2 // ping messages. - KeepAlivesSent int64 `protobuf:"varint,6,opt,name=keep_alives_sent,json=keepAlivesSent" json:"keep_alives_sent,omitempty"` + KeepAlivesSent int64 `protobuf:"varint,6,opt,name=keep_alives_sent,json=keepAlivesSent,proto3" json:"keep_alives_sent,omitempty"` // The last time a stream was created by this endpoint. Usually unset for // servers. - LastLocalStreamCreatedTimestamp *timestamp.Timestamp `protobuf:"bytes,7,opt,name=last_local_stream_created_timestamp,json=lastLocalStreamCreatedTimestamp" json:"last_local_stream_created_timestamp,omitempty"` + LastLocalStreamCreatedTimestamp *timestamp.Timestamp `protobuf:"bytes,7,opt,name=last_local_stream_created_timestamp,json=lastLocalStreamCreatedTimestamp,proto3" json:"last_local_stream_created_timestamp,omitempty"` // The last time a stream was created by the remote endpoint. Usually unset // for clients. - LastRemoteStreamCreatedTimestamp *timestamp.Timestamp `protobuf:"bytes,8,opt,name=last_remote_stream_created_timestamp,json=lastRemoteStreamCreatedTimestamp" json:"last_remote_stream_created_timestamp,omitempty"` + LastRemoteStreamCreatedTimestamp *timestamp.Timestamp `protobuf:"bytes,8,opt,name=last_remote_stream_created_timestamp,json=lastRemoteStreamCreatedTimestamp,proto3" json:"last_remote_stream_created_timestamp,omitempty"` // The last time a message was sent by this endpoint. - LastMessageSentTimestamp *timestamp.Timestamp `protobuf:"bytes,9,opt,name=last_message_sent_timestamp,json=lastMessageSentTimestamp" json:"last_message_sent_timestamp,omitempty"` + LastMessageSentTimestamp *timestamp.Timestamp `protobuf:"bytes,9,opt,name=last_message_sent_timestamp,json=lastMessageSentTimestamp,proto3" json:"last_message_sent_timestamp,omitempty"` // The last time a message was received by this endpoint. - LastMessageReceivedTimestamp *timestamp.Timestamp `protobuf:"bytes,10,opt,name=last_message_received_timestamp,json=lastMessageReceivedTimestamp" json:"last_message_received_timestamp,omitempty"` + LastMessageReceivedTimestamp *timestamp.Timestamp `protobuf:"bytes,10,opt,name=last_message_received_timestamp,json=lastMessageReceivedTimestamp,proto3" json:"last_message_received_timestamp,omitempty"` // The amount of window, granted to the local endpoint by the remote endpoint. // This may be slightly out of date due to network latency. This does NOT // include stream level or TCP level flow control info. - LocalFlowControlWindow *wrappers.Int64Value `protobuf:"bytes,11,opt,name=local_flow_control_window,json=localFlowControlWindow" json:"local_flow_control_window,omitempty"` + LocalFlowControlWindow *wrappers.Int64Value `protobuf:"bytes,11,opt,name=local_flow_control_window,json=localFlowControlWindow,proto3" json:"local_flow_control_window,omitempty"` // The amount of window, granted to the remote endpoint by the local endpoint. // This may be slightly out of date due to network latency. This does NOT // include stream level or TCP level flow control info. - RemoteFlowControlWindow *wrappers.Int64Value `protobuf:"bytes,12,opt,name=remote_flow_control_window,json=remoteFlowControlWindow" json:"remote_flow_control_window,omitempty"` + RemoteFlowControlWindow *wrappers.Int64Value `protobuf:"bytes,12,opt,name=remote_flow_control_window,json=remoteFlowControlWindow,proto3" json:"remote_flow_control_window,omitempty"` // Socket options set on this socket. May be absent. - Option []*SocketOption `protobuf:"bytes,13,rep,name=option" json:"option,omitempty"` + Option []*SocketOption `protobuf:"bytes,13,rep,name=option,proto3" json:"option,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -859,7 +1095,7 @@ func (m *SocketData) Reset() { *m = SocketData{} } func (m *SocketData) String() string { return proto.CompactTextString(m) } func (*SocketData) ProtoMessage() {} func (*SocketData) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{12} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{13} } func (m *SocketData) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SocketData.Unmarshal(m, b) @@ -986,7 +1222,7 @@ func (m *Address) Reset() { *m = Address{} } func (m *Address) String() string { return proto.CompactTextString(m) } func (*Address) ProtoMessage() {} func (*Address) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{13} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{14} } func (m *Address) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Address.Unmarshal(m, b) @@ -1011,13 +1247,13 @@ type isAddress_Address interface { } type Address_TcpipAddress struct { - TcpipAddress *Address_TcpIpAddress `protobuf:"bytes,1,opt,name=tcpip_address,json=tcpipAddress,oneof"` + TcpipAddress *Address_TcpIpAddress `protobuf:"bytes,1,opt,name=tcpip_address,json=tcpipAddress,proto3,oneof"` } type Address_UdsAddress_ struct { - UdsAddress *Address_UdsAddress `protobuf:"bytes,2,opt,name=uds_address,json=udsAddress,oneof"` + UdsAddress *Address_UdsAddress `protobuf:"bytes,2,opt,name=uds_address,json=udsAddress,proto3,oneof"` } type Address_OtherAddress_ struct { - OtherAddress *Address_OtherAddress `protobuf:"bytes,3,opt,name=other_address,json=otherAddress,oneof"` + OtherAddress *Address_OtherAddress `protobuf:"bytes,3,opt,name=other_address,json=otherAddress,proto3,oneof"` } func (*Address_TcpipAddress) isAddress_Address() {} @@ -1150,7 +1386,7 @@ type Address_TcpIpAddress struct { // bytes in length. IpAddress []byte `protobuf:"bytes,1,opt,name=ip_address,json=ipAddress,proto3" json:"ip_address,omitempty"` // 0-64k, or -1 if not appropriate. - Port int32 `protobuf:"varint,2,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,2,opt,name=port,proto3" json:"port,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1160,7 +1396,7 @@ func (m *Address_TcpIpAddress) Reset() { *m = Address_TcpIpAddress{} } func (m *Address_TcpIpAddress) String() string { return proto.CompactTextString(m) } func (*Address_TcpIpAddress) ProtoMessage() {} func (*Address_TcpIpAddress) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{13, 0} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{14, 0} } func (m *Address_TcpIpAddress) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Address_TcpIpAddress.Unmarshal(m, b) @@ -1196,7 +1432,7 @@ func (m *Address_TcpIpAddress) GetPort() int32 { // A Unix Domain Socket address. type Address_UdsAddress struct { - Filename string `protobuf:"bytes,1,opt,name=filename" json:"filename,omitempty"` + Filename string `protobuf:"bytes,1,opt,name=filename,proto3" json:"filename,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1206,7 +1442,7 @@ func (m *Address_UdsAddress) Reset() { *m = Address_UdsAddress{} } func (m *Address_UdsAddress) String() string { return proto.CompactTextString(m) } func (*Address_UdsAddress) ProtoMessage() {} func (*Address_UdsAddress) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{13, 1} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{14, 1} } func (m *Address_UdsAddress) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Address_UdsAddress.Unmarshal(m, b) @@ -1236,9 +1472,9 @@ func (m *Address_UdsAddress) GetFilename() string { // An address type not included above. type Address_OtherAddress struct { // The human readable version of the value. This value should be set. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The actual address message. - Value *any.Any `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value *any.Any `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1248,7 +1484,7 @@ func (m *Address_OtherAddress) Reset() { *m = Address_OtherAddress{} } func (m *Address_OtherAddress) String() string { return proto.CompactTextString(m) } func (*Address_OtherAddress) ProtoMessage() {} func (*Address_OtherAddress) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{13, 2} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{14, 2} } func (m *Address_OtherAddress) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Address_OtherAddress.Unmarshal(m, b) @@ -1297,7 +1533,7 @@ func (m *Security) Reset() { *m = Security{} } func (m *Security) String() string { return proto.CompactTextString(m) } func (*Security) ProtoMessage() {} func (*Security) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{14} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{15} } func (m *Security) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Security.Unmarshal(m, b) @@ -1322,10 +1558,10 @@ type isSecurity_Model interface { } type Security_Tls_ struct { - Tls *Security_Tls `protobuf:"bytes,1,opt,name=tls,oneof"` + Tls *Security_Tls `protobuf:"bytes,1,opt,name=tls,proto3,oneof"` } type Security_Other struct { - Other *Security_OtherSecurity `protobuf:"bytes,2,opt,name=other,oneof"` + Other *Security_OtherSecurity `protobuf:"bytes,2,opt,name=other,proto3,oneof"` } func (*Security_Tls_) isSecurity_Model() {} @@ -1444,7 +1680,7 @@ func (m *Security_Tls) Reset() { *m = Security_Tls{} } func (m *Security_Tls) String() string { return proto.CompactTextString(m) } func (*Security_Tls) ProtoMessage() {} func (*Security_Tls) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{14, 0} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{15, 0} } func (m *Security_Tls) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Security_Tls.Unmarshal(m, b) @@ -1469,10 +1705,10 @@ type isSecurity_Tls_CipherSuite interface { } type Security_Tls_StandardName struct { - StandardName string `protobuf:"bytes,1,opt,name=standard_name,json=standardName,oneof"` + StandardName string `protobuf:"bytes,1,opt,name=standard_name,json=standardName,proto3,oneof"` } type Security_Tls_OtherName struct { - OtherName string `protobuf:"bytes,2,opt,name=other_name,json=otherName,oneof"` + OtherName string `protobuf:"bytes,2,opt,name=other_name,json=otherName,proto3,oneof"` } func (*Security_Tls_StandardName) isSecurity_Tls_CipherSuite() {} @@ -1581,9 +1817,9 @@ func _Security_Tls_OneofSizer(msg proto.Message) (n int) { type Security_OtherSecurity struct { // The human readable version of the value. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The actual security details message. - Value *any.Any `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value *any.Any `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1593,7 +1829,7 @@ func (m *Security_OtherSecurity) Reset() { *m = Security_OtherSecurity{} func (m *Security_OtherSecurity) String() string { return proto.CompactTextString(m) } func (*Security_OtherSecurity) ProtoMessage() {} func (*Security_OtherSecurity) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{14, 1} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{15, 1} } func (m *Security_OtherSecurity) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Security_OtherSecurity.Unmarshal(m, b) @@ -1632,13 +1868,13 @@ func (m *Security_OtherSecurity) GetValue() *any.Any { type SocketOption struct { // The full name of the socket option. Typically this will be the upper case // name, such as "SO_REUSEPORT". - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The human readable value of this socket option. At least one of value or // additional will be set. - Value string `protobuf:"bytes,2,opt,name=value" json:"value,omitempty"` + Value string `protobuf:"bytes,2,opt,name=value,proto3" json:"value,omitempty"` // Additional data associated with the socket option. At least one of value // or additional will be set. - Additional *any.Any `protobuf:"bytes,3,opt,name=additional" json:"additional,omitempty"` + Additional *any.Any `protobuf:"bytes,3,opt,name=additional,proto3" json:"additional,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1648,7 +1884,7 @@ func (m *SocketOption) Reset() { *m = SocketOption{} } func (m *SocketOption) String() string { return proto.CompactTextString(m) } func (*SocketOption) ProtoMessage() {} func (*SocketOption) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{15} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{16} } func (m *SocketOption) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SocketOption.Unmarshal(m, b) @@ -1692,7 +1928,7 @@ func (m *SocketOption) GetAdditional() *any.Any { // For use with SocketOption's additional field. This is primarily used for // SO_RCVTIMEO and SO_SNDTIMEO type SocketOptionTimeout struct { - Duration *duration.Duration `protobuf:"bytes,1,opt,name=duration" json:"duration,omitempty"` + Duration *duration.Duration `protobuf:"bytes,1,opt,name=duration,proto3" json:"duration,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1702,7 +1938,7 @@ func (m *SocketOptionTimeout) Reset() { *m = SocketOptionTimeout{} } func (m *SocketOptionTimeout) String() string { return proto.CompactTextString(m) } func (*SocketOptionTimeout) ProtoMessage() {} func (*SocketOptionTimeout) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{16} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{17} } func (m *SocketOptionTimeout) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SocketOptionTimeout.Unmarshal(m, b) @@ -1733,9 +1969,9 @@ func (m *SocketOptionTimeout) GetDuration() *duration.Duration { // SO_LINGER. type SocketOptionLinger struct { // active maps to `struct linger.l_onoff` - Active bool `protobuf:"varint,1,opt,name=active" json:"active,omitempty"` + Active bool `protobuf:"varint,1,opt,name=active,proto3" json:"active,omitempty"` // duration maps to `struct linger.l_linger` - Duration *duration.Duration `protobuf:"bytes,2,opt,name=duration" json:"duration,omitempty"` + Duration *duration.Duration `protobuf:"bytes,2,opt,name=duration,proto3" json:"duration,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1745,7 +1981,7 @@ func (m *SocketOptionLinger) Reset() { *m = SocketOptionLinger{} } func (m *SocketOptionLinger) String() string { return proto.CompactTextString(m) } func (*SocketOptionLinger) ProtoMessage() {} func (*SocketOptionLinger) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{17} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{18} } func (m *SocketOptionLinger) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SocketOptionLinger.Unmarshal(m, b) @@ -1782,35 +2018,35 @@ func (m *SocketOptionLinger) GetDuration() *duration.Duration { // For use with SocketOption's additional field. Tcp info for // SOL_TCP and TCP_INFO. type SocketOptionTcpInfo struct { - TcpiState uint32 `protobuf:"varint,1,opt,name=tcpi_state,json=tcpiState" json:"tcpi_state,omitempty"` - TcpiCaState uint32 `protobuf:"varint,2,opt,name=tcpi_ca_state,json=tcpiCaState" json:"tcpi_ca_state,omitempty"` - TcpiRetransmits uint32 `protobuf:"varint,3,opt,name=tcpi_retransmits,json=tcpiRetransmits" json:"tcpi_retransmits,omitempty"` - TcpiProbes uint32 `protobuf:"varint,4,opt,name=tcpi_probes,json=tcpiProbes" json:"tcpi_probes,omitempty"` - TcpiBackoff uint32 `protobuf:"varint,5,opt,name=tcpi_backoff,json=tcpiBackoff" json:"tcpi_backoff,omitempty"` - TcpiOptions uint32 `protobuf:"varint,6,opt,name=tcpi_options,json=tcpiOptions" json:"tcpi_options,omitempty"` - TcpiSndWscale uint32 `protobuf:"varint,7,opt,name=tcpi_snd_wscale,json=tcpiSndWscale" json:"tcpi_snd_wscale,omitempty"` - TcpiRcvWscale uint32 `protobuf:"varint,8,opt,name=tcpi_rcv_wscale,json=tcpiRcvWscale" json:"tcpi_rcv_wscale,omitempty"` - TcpiRto uint32 `protobuf:"varint,9,opt,name=tcpi_rto,json=tcpiRto" json:"tcpi_rto,omitempty"` - TcpiAto uint32 `protobuf:"varint,10,opt,name=tcpi_ato,json=tcpiAto" json:"tcpi_ato,omitempty"` - TcpiSndMss uint32 `protobuf:"varint,11,opt,name=tcpi_snd_mss,json=tcpiSndMss" json:"tcpi_snd_mss,omitempty"` - TcpiRcvMss uint32 `protobuf:"varint,12,opt,name=tcpi_rcv_mss,json=tcpiRcvMss" json:"tcpi_rcv_mss,omitempty"` - TcpiUnacked uint32 `protobuf:"varint,13,opt,name=tcpi_unacked,json=tcpiUnacked" json:"tcpi_unacked,omitempty"` - TcpiSacked uint32 `protobuf:"varint,14,opt,name=tcpi_sacked,json=tcpiSacked" json:"tcpi_sacked,omitempty"` - TcpiLost uint32 `protobuf:"varint,15,opt,name=tcpi_lost,json=tcpiLost" json:"tcpi_lost,omitempty"` - TcpiRetrans uint32 `protobuf:"varint,16,opt,name=tcpi_retrans,json=tcpiRetrans" json:"tcpi_retrans,omitempty"` - TcpiFackets uint32 `protobuf:"varint,17,opt,name=tcpi_fackets,json=tcpiFackets" json:"tcpi_fackets,omitempty"` - TcpiLastDataSent uint32 `protobuf:"varint,18,opt,name=tcpi_last_data_sent,json=tcpiLastDataSent" json:"tcpi_last_data_sent,omitempty"` - TcpiLastAckSent uint32 `protobuf:"varint,19,opt,name=tcpi_last_ack_sent,json=tcpiLastAckSent" json:"tcpi_last_ack_sent,omitempty"` - TcpiLastDataRecv uint32 `protobuf:"varint,20,opt,name=tcpi_last_data_recv,json=tcpiLastDataRecv" json:"tcpi_last_data_recv,omitempty"` - TcpiLastAckRecv uint32 `protobuf:"varint,21,opt,name=tcpi_last_ack_recv,json=tcpiLastAckRecv" json:"tcpi_last_ack_recv,omitempty"` - TcpiPmtu uint32 `protobuf:"varint,22,opt,name=tcpi_pmtu,json=tcpiPmtu" json:"tcpi_pmtu,omitempty"` - TcpiRcvSsthresh uint32 `protobuf:"varint,23,opt,name=tcpi_rcv_ssthresh,json=tcpiRcvSsthresh" json:"tcpi_rcv_ssthresh,omitempty"` - TcpiRtt uint32 `protobuf:"varint,24,opt,name=tcpi_rtt,json=tcpiRtt" json:"tcpi_rtt,omitempty"` - TcpiRttvar uint32 `protobuf:"varint,25,opt,name=tcpi_rttvar,json=tcpiRttvar" json:"tcpi_rttvar,omitempty"` - TcpiSndSsthresh uint32 `protobuf:"varint,26,opt,name=tcpi_snd_ssthresh,json=tcpiSndSsthresh" json:"tcpi_snd_ssthresh,omitempty"` - TcpiSndCwnd uint32 `protobuf:"varint,27,opt,name=tcpi_snd_cwnd,json=tcpiSndCwnd" json:"tcpi_snd_cwnd,omitempty"` - TcpiAdvmss uint32 `protobuf:"varint,28,opt,name=tcpi_advmss,json=tcpiAdvmss" json:"tcpi_advmss,omitempty"` - TcpiReordering uint32 `protobuf:"varint,29,opt,name=tcpi_reordering,json=tcpiReordering" json:"tcpi_reordering,omitempty"` + TcpiState uint32 `protobuf:"varint,1,opt,name=tcpi_state,json=tcpiState,proto3" json:"tcpi_state,omitempty"` + TcpiCaState uint32 `protobuf:"varint,2,opt,name=tcpi_ca_state,json=tcpiCaState,proto3" json:"tcpi_ca_state,omitempty"` + TcpiRetransmits uint32 `protobuf:"varint,3,opt,name=tcpi_retransmits,json=tcpiRetransmits,proto3" json:"tcpi_retransmits,omitempty"` + TcpiProbes uint32 `protobuf:"varint,4,opt,name=tcpi_probes,json=tcpiProbes,proto3" json:"tcpi_probes,omitempty"` + TcpiBackoff uint32 `protobuf:"varint,5,opt,name=tcpi_backoff,json=tcpiBackoff,proto3" json:"tcpi_backoff,omitempty"` + TcpiOptions uint32 `protobuf:"varint,6,opt,name=tcpi_options,json=tcpiOptions,proto3" json:"tcpi_options,omitempty"` + TcpiSndWscale uint32 `protobuf:"varint,7,opt,name=tcpi_snd_wscale,json=tcpiSndWscale,proto3" json:"tcpi_snd_wscale,omitempty"` + TcpiRcvWscale uint32 `protobuf:"varint,8,opt,name=tcpi_rcv_wscale,json=tcpiRcvWscale,proto3" json:"tcpi_rcv_wscale,omitempty"` + TcpiRto uint32 `protobuf:"varint,9,opt,name=tcpi_rto,json=tcpiRto,proto3" json:"tcpi_rto,omitempty"` + TcpiAto uint32 `protobuf:"varint,10,opt,name=tcpi_ato,json=tcpiAto,proto3" json:"tcpi_ato,omitempty"` + TcpiSndMss uint32 `protobuf:"varint,11,opt,name=tcpi_snd_mss,json=tcpiSndMss,proto3" json:"tcpi_snd_mss,omitempty"` + TcpiRcvMss uint32 `protobuf:"varint,12,opt,name=tcpi_rcv_mss,json=tcpiRcvMss,proto3" json:"tcpi_rcv_mss,omitempty"` + TcpiUnacked uint32 `protobuf:"varint,13,opt,name=tcpi_unacked,json=tcpiUnacked,proto3" json:"tcpi_unacked,omitempty"` + TcpiSacked uint32 `protobuf:"varint,14,opt,name=tcpi_sacked,json=tcpiSacked,proto3" json:"tcpi_sacked,omitempty"` + TcpiLost uint32 `protobuf:"varint,15,opt,name=tcpi_lost,json=tcpiLost,proto3" json:"tcpi_lost,omitempty"` + TcpiRetrans uint32 `protobuf:"varint,16,opt,name=tcpi_retrans,json=tcpiRetrans,proto3" json:"tcpi_retrans,omitempty"` + TcpiFackets uint32 `protobuf:"varint,17,opt,name=tcpi_fackets,json=tcpiFackets,proto3" json:"tcpi_fackets,omitempty"` + TcpiLastDataSent uint32 `protobuf:"varint,18,opt,name=tcpi_last_data_sent,json=tcpiLastDataSent,proto3" json:"tcpi_last_data_sent,omitempty"` + TcpiLastAckSent uint32 `protobuf:"varint,19,opt,name=tcpi_last_ack_sent,json=tcpiLastAckSent,proto3" json:"tcpi_last_ack_sent,omitempty"` + TcpiLastDataRecv uint32 `protobuf:"varint,20,opt,name=tcpi_last_data_recv,json=tcpiLastDataRecv,proto3" json:"tcpi_last_data_recv,omitempty"` + TcpiLastAckRecv uint32 `protobuf:"varint,21,opt,name=tcpi_last_ack_recv,json=tcpiLastAckRecv,proto3" json:"tcpi_last_ack_recv,omitempty"` + TcpiPmtu uint32 `protobuf:"varint,22,opt,name=tcpi_pmtu,json=tcpiPmtu,proto3" json:"tcpi_pmtu,omitempty"` + TcpiRcvSsthresh uint32 `protobuf:"varint,23,opt,name=tcpi_rcv_ssthresh,json=tcpiRcvSsthresh,proto3" json:"tcpi_rcv_ssthresh,omitempty"` + TcpiRtt uint32 `protobuf:"varint,24,opt,name=tcpi_rtt,json=tcpiRtt,proto3" json:"tcpi_rtt,omitempty"` + TcpiRttvar uint32 `protobuf:"varint,25,opt,name=tcpi_rttvar,json=tcpiRttvar,proto3" json:"tcpi_rttvar,omitempty"` + TcpiSndSsthresh uint32 `protobuf:"varint,26,opt,name=tcpi_snd_ssthresh,json=tcpiSndSsthresh,proto3" json:"tcpi_snd_ssthresh,omitempty"` + TcpiSndCwnd uint32 `protobuf:"varint,27,opt,name=tcpi_snd_cwnd,json=tcpiSndCwnd,proto3" json:"tcpi_snd_cwnd,omitempty"` + TcpiAdvmss uint32 `protobuf:"varint,28,opt,name=tcpi_advmss,json=tcpiAdvmss,proto3" json:"tcpi_advmss,omitempty"` + TcpiReordering uint32 `protobuf:"varint,29,opt,name=tcpi_reordering,json=tcpiReordering,proto3" json:"tcpi_reordering,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -1820,7 +2056,7 @@ func (m *SocketOptionTcpInfo) Reset() { *m = SocketOptionTcpInfo{} } func (m *SocketOptionTcpInfo) String() string { return proto.CompactTextString(m) } func (*SocketOptionTcpInfo) ProtoMessage() {} func (*SocketOptionTcpInfo) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{18} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{19} } func (m *SocketOptionTcpInfo) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_SocketOptionTcpInfo.Unmarshal(m, b) @@ -2046,7 +2282,7 @@ func (m *SocketOptionTcpInfo) GetTcpiReordering() uint32 { type GetTopChannelsRequest struct { // start_channel_id indicates that only channels at or above this id should be // included in the results. - StartChannelId int64 `protobuf:"varint,1,opt,name=start_channel_id,json=startChannelId" json:"start_channel_id,omitempty"` + StartChannelId int64 `protobuf:"varint,1,opt,name=start_channel_id,json=startChannelId,proto3" json:"start_channel_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2056,7 +2292,7 @@ func (m *GetTopChannelsRequest) Reset() { *m = GetTopChannelsRequest{} } func (m *GetTopChannelsRequest) String() string { return proto.CompactTextString(m) } func (*GetTopChannelsRequest) ProtoMessage() {} func (*GetTopChannelsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{19} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{20} } func (m *GetTopChannelsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetTopChannelsRequest.Unmarshal(m, b) @@ -2086,11 +2322,11 @@ func (m *GetTopChannelsRequest) GetStartChannelId() int64 { type GetTopChannelsResponse struct { // list of channels that the connection detail service knows about. Sorted in // ascending channel_id order. - Channel []*Channel `protobuf:"bytes,1,rep,name=channel" json:"channel,omitempty"` + Channel []*Channel `protobuf:"bytes,1,rep,name=channel,proto3" json:"channel,omitempty"` // If set, indicates that the list of channels is the final list. Requesting // more channels can only return more if they are created after this RPC // completes. - End bool `protobuf:"varint,2,opt,name=end" json:"end,omitempty"` + End bool `protobuf:"varint,2,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2100,7 +2336,7 @@ func (m *GetTopChannelsResponse) Reset() { *m = GetTopChannelsResponse{} func (m *GetTopChannelsResponse) String() string { return proto.CompactTextString(m) } func (*GetTopChannelsResponse) ProtoMessage() {} func (*GetTopChannelsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{20} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{21} } func (m *GetTopChannelsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetTopChannelsResponse.Unmarshal(m, b) @@ -2137,7 +2373,7 @@ func (m *GetTopChannelsResponse) GetEnd() bool { type GetServersRequest struct { // start_server_id indicates that only servers at or above this id should be // included in the results. - StartServerId int64 `protobuf:"varint,1,opt,name=start_server_id,json=startServerId" json:"start_server_id,omitempty"` + StartServerId int64 `protobuf:"varint,1,opt,name=start_server_id,json=startServerId,proto3" json:"start_server_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2147,7 +2383,7 @@ func (m *GetServersRequest) Reset() { *m = GetServersRequest{} } func (m *GetServersRequest) String() string { return proto.CompactTextString(m) } func (*GetServersRequest) ProtoMessage() {} func (*GetServersRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{21} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{22} } func (m *GetServersRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetServersRequest.Unmarshal(m, b) @@ -2177,11 +2413,11 @@ func (m *GetServersRequest) GetStartServerId() int64 { type GetServersResponse struct { // list of servers that the connection detail service knows about. Sorted in // ascending server_id order. - Server []*Server `protobuf:"bytes,1,rep,name=server" json:"server,omitempty"` + Server []*Server `protobuf:"bytes,1,rep,name=server,proto3" json:"server,omitempty"` // If set, indicates that the list of servers is the final list. Requesting // more servers will only return more if they are created after this RPC // completes. - End bool `protobuf:"varint,2,opt,name=end" json:"end,omitempty"` + End bool `protobuf:"varint,2,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2191,7 +2427,7 @@ func (m *GetServersResponse) Reset() { *m = GetServersResponse{} } func (m *GetServersResponse) String() string { return proto.CompactTextString(m) } func (*GetServersResponse) ProtoMessage() {} func (*GetServersResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{22} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{23} } func (m *GetServersResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetServersResponse.Unmarshal(m, b) @@ -2226,10 +2462,10 @@ func (m *GetServersResponse) GetEnd() bool { } type GetServerSocketsRequest struct { - ServerId int64 `protobuf:"varint,1,opt,name=server_id,json=serverId" json:"server_id,omitempty"` + ServerId int64 `protobuf:"varint,1,opt,name=server_id,json=serverId,proto3" json:"server_id,omitempty"` // start_socket_id indicates that only sockets at or above this id should be // included in the results. - StartSocketId int64 `protobuf:"varint,2,opt,name=start_socket_id,json=startSocketId" json:"start_socket_id,omitempty"` + StartSocketId int64 `protobuf:"varint,2,opt,name=start_socket_id,json=startSocketId,proto3" json:"start_socket_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2239,7 +2475,7 @@ func (m *GetServerSocketsRequest) Reset() { *m = GetServerSocketsRequest func (m *GetServerSocketsRequest) String() string { return proto.CompactTextString(m) } func (*GetServerSocketsRequest) ProtoMessage() {} func (*GetServerSocketsRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{23} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{24} } func (m *GetServerSocketsRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetServerSocketsRequest.Unmarshal(m, b) @@ -2276,11 +2512,11 @@ func (m *GetServerSocketsRequest) GetStartSocketId() int64 { type GetServerSocketsResponse struct { // list of socket refs that the connection detail service knows about. Sorted in // ascending socket_id order. - SocketRef []*SocketRef `protobuf:"bytes,1,rep,name=socket_ref,json=socketRef" json:"socket_ref,omitempty"` + SocketRef []*SocketRef `protobuf:"bytes,1,rep,name=socket_ref,json=socketRef,proto3" json:"socket_ref,omitempty"` // If set, indicates that the list of sockets is the final list. Requesting // more sockets will only return more if they are created after this RPC // completes. - End bool `protobuf:"varint,2,opt,name=end" json:"end,omitempty"` + End bool `protobuf:"varint,2,opt,name=end,proto3" json:"end,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2290,7 +2526,7 @@ func (m *GetServerSocketsResponse) Reset() { *m = GetServerSocketsRespon func (m *GetServerSocketsResponse) String() string { return proto.CompactTextString(m) } func (*GetServerSocketsResponse) ProtoMessage() {} func (*GetServerSocketsResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{24} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{25} } func (m *GetServerSocketsResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetServerSocketsResponse.Unmarshal(m, b) @@ -2326,7 +2562,7 @@ func (m *GetServerSocketsResponse) GetEnd() bool { type GetChannelRequest struct { // channel_id is the identifier of the specific channel to get. - ChannelId int64 `protobuf:"varint,1,opt,name=channel_id,json=channelId" json:"channel_id,omitempty"` + ChannelId int64 `protobuf:"varint,1,opt,name=channel_id,json=channelId,proto3" json:"channel_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2336,7 +2572,7 @@ func (m *GetChannelRequest) Reset() { *m = GetChannelRequest{} } func (m *GetChannelRequest) String() string { return proto.CompactTextString(m) } func (*GetChannelRequest) ProtoMessage() {} func (*GetChannelRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{25} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{26} } func (m *GetChannelRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetChannelRequest.Unmarshal(m, b) @@ -2366,7 +2602,7 @@ func (m *GetChannelRequest) GetChannelId() int64 { type GetChannelResponse struct { // The Channel that corresponds to the requested channel_id. This field // should be set. - Channel *Channel `protobuf:"bytes,1,opt,name=channel" json:"channel,omitempty"` + Channel *Channel `protobuf:"bytes,1,opt,name=channel,proto3" json:"channel,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2376,7 +2612,7 @@ func (m *GetChannelResponse) Reset() { *m = GetChannelResponse{} } func (m *GetChannelResponse) String() string { return proto.CompactTextString(m) } func (*GetChannelResponse) ProtoMessage() {} func (*GetChannelResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{26} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{27} } func (m *GetChannelResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetChannelResponse.Unmarshal(m, b) @@ -2405,7 +2641,7 @@ func (m *GetChannelResponse) GetChannel() *Channel { type GetSubchannelRequest struct { // subchannel_id is the identifier of the specific subchannel to get. - SubchannelId int64 `protobuf:"varint,1,opt,name=subchannel_id,json=subchannelId" json:"subchannel_id,omitempty"` + SubchannelId int64 `protobuf:"varint,1,opt,name=subchannel_id,json=subchannelId,proto3" json:"subchannel_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2415,7 +2651,7 @@ func (m *GetSubchannelRequest) Reset() { *m = GetSubchannelRequest{} } func (m *GetSubchannelRequest) String() string { return proto.CompactTextString(m) } func (*GetSubchannelRequest) ProtoMessage() {} func (*GetSubchannelRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{27} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{28} } func (m *GetSubchannelRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSubchannelRequest.Unmarshal(m, b) @@ -2445,7 +2681,7 @@ func (m *GetSubchannelRequest) GetSubchannelId() int64 { type GetSubchannelResponse struct { // The Subchannel that corresponds to the requested subchannel_id. This // field should be set. - Subchannel *Subchannel `protobuf:"bytes,1,opt,name=subchannel" json:"subchannel,omitempty"` + Subchannel *Subchannel `protobuf:"bytes,1,opt,name=subchannel,proto3" json:"subchannel,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2455,7 +2691,7 @@ func (m *GetSubchannelResponse) Reset() { *m = GetSubchannelResponse{} } func (m *GetSubchannelResponse) String() string { return proto.CompactTextString(m) } func (*GetSubchannelResponse) ProtoMessage() {} func (*GetSubchannelResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{28} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{29} } func (m *GetSubchannelResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSubchannelResponse.Unmarshal(m, b) @@ -2484,7 +2720,7 @@ func (m *GetSubchannelResponse) GetSubchannel() *Subchannel { type GetSocketRequest struct { // socket_id is the identifier of the specific socket to get. - SocketId int64 `protobuf:"varint,1,opt,name=socket_id,json=socketId" json:"socket_id,omitempty"` + SocketId int64 `protobuf:"varint,1,opt,name=socket_id,json=socketId,proto3" json:"socket_id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2494,7 +2730,7 @@ func (m *GetSocketRequest) Reset() { *m = GetSocketRequest{} } func (m *GetSocketRequest) String() string { return proto.CompactTextString(m) } func (*GetSocketRequest) ProtoMessage() {} func (*GetSocketRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{29} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{30} } func (m *GetSocketRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSocketRequest.Unmarshal(m, b) @@ -2524,7 +2760,7 @@ func (m *GetSocketRequest) GetSocketId() int64 { type GetSocketResponse struct { // The Socket that corresponds to the requested socket_id. This field // should be set. - Socket *Socket `protobuf:"bytes,1,opt,name=socket" json:"socket,omitempty"` + Socket *Socket `protobuf:"bytes,1,opt,name=socket,proto3" json:"socket,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -2534,7 +2770,7 @@ func (m *GetSocketResponse) Reset() { *m = GetSocketResponse{} } func (m *GetSocketResponse) String() string { return proto.CompactTextString(m) } func (*GetSocketResponse) ProtoMessage() {} func (*GetSocketResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_channelz_544d5edb3d49afb2, []int{30} + return fileDescriptor_channelz_ce3ed45d08251f2f, []int{31} } func (m *GetSocketResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_GetSocketResponse.Unmarshal(m, b) @@ -2566,6 +2802,7 @@ func init() { proto.RegisterType((*Subchannel)(nil), "grpc.channelz.v1.Subchannel") proto.RegisterType((*ChannelConnectivityState)(nil), "grpc.channelz.v1.ChannelConnectivityState") proto.RegisterType((*ChannelData)(nil), "grpc.channelz.v1.ChannelData") + proto.RegisterType((*ChannelTraceEvent)(nil), "grpc.channelz.v1.ChannelTraceEvent") proto.RegisterType((*ChannelTrace)(nil), "grpc.channelz.v1.ChannelTrace") proto.RegisterType((*ChannelRef)(nil), "grpc.channelz.v1.ChannelRef") proto.RegisterType((*SubchannelRef)(nil), "grpc.channelz.v1.SubchannelRef") @@ -2599,6 +2836,7 @@ func init() { proto.RegisterType((*GetSocketRequest)(nil), "grpc.channelz.v1.GetSocketRequest") proto.RegisterType((*GetSocketResponse)(nil), "grpc.channelz.v1.GetSocketResponse") proto.RegisterEnum("grpc.channelz.v1.ChannelConnectivityState_State", ChannelConnectivityState_State_name, ChannelConnectivityState_State_value) + proto.RegisterEnum("grpc.channelz.v1.ChannelTraceEvent_Severity", ChannelTraceEvent_Severity_name, ChannelTraceEvent_Severity_value) } // Reference imports to suppress errors if they are not otherwise used. @@ -2609,8 +2847,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for Channelz service - +// ChannelzClient is the client API for Channelz service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type ChannelzClient interface { // Gets all root channels (i.e. channels the application has directly // created). This does not include subchannels nor non-top level channels. @@ -2637,7 +2876,7 @@ func NewChannelzClient(cc *grpc.ClientConn) ChannelzClient { func (c *channelzClient) GetTopChannels(ctx context.Context, in *GetTopChannelsRequest, opts ...grpc.CallOption) (*GetTopChannelsResponse, error) { out := new(GetTopChannelsResponse) - err := grpc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetTopChannels", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetTopChannels", in, out, opts...) if err != nil { return nil, err } @@ -2646,7 +2885,7 @@ func (c *channelzClient) GetTopChannels(ctx context.Context, in *GetTopChannelsR func (c *channelzClient) GetServers(ctx context.Context, in *GetServersRequest, opts ...grpc.CallOption) (*GetServersResponse, error) { out := new(GetServersResponse) - err := grpc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetServers", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetServers", in, out, opts...) if err != nil { return nil, err } @@ -2655,7 +2894,7 @@ func (c *channelzClient) GetServers(ctx context.Context, in *GetServersRequest, func (c *channelzClient) GetServerSockets(ctx context.Context, in *GetServerSocketsRequest, opts ...grpc.CallOption) (*GetServerSocketsResponse, error) { out := new(GetServerSocketsResponse) - err := grpc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetServerSockets", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetServerSockets", in, out, opts...) if err != nil { return nil, err } @@ -2664,7 +2903,7 @@ func (c *channelzClient) GetServerSockets(ctx context.Context, in *GetServerSock func (c *channelzClient) GetChannel(ctx context.Context, in *GetChannelRequest, opts ...grpc.CallOption) (*GetChannelResponse, error) { out := new(GetChannelResponse) - err := grpc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetChannel", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetChannel", in, out, opts...) if err != nil { return nil, err } @@ -2673,7 +2912,7 @@ func (c *channelzClient) GetChannel(ctx context.Context, in *GetChannelRequest, func (c *channelzClient) GetSubchannel(ctx context.Context, in *GetSubchannelRequest, opts ...grpc.CallOption) (*GetSubchannelResponse, error) { out := new(GetSubchannelResponse) - err := grpc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetSubchannel", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetSubchannel", in, out, opts...) if err != nil { return nil, err } @@ -2682,15 +2921,14 @@ func (c *channelzClient) GetSubchannel(ctx context.Context, in *GetSubchannelReq func (c *channelzClient) GetSocket(ctx context.Context, in *GetSocketRequest, opts ...grpc.CallOption) (*GetSocketResponse, error) { out := new(GetSocketResponse) - err := grpc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetSocket", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.channelz.v1.Channelz/GetSocket", in, out, opts...) if err != nil { return nil, err } return out, nil } -// Server API for Channelz service - +// ChannelzServer is the server API for Channelz service. type ChannelzServer interface { // Gets all root channels (i.e. channels the application has directly // created). This does not include subchannels nor non-top level channels. @@ -2849,156 +3087,171 @@ var _Channelz_serviceDesc = grpc.ServiceDesc{ }, }, Streams: []grpc.StreamDesc{}, - Metadata: "channelz.proto", + Metadata: "grpc/channelz/v1/channelz.proto", } -func init() { proto.RegisterFile("channelz.proto", fileDescriptor_channelz_544d5edb3d49afb2) } - -var fileDescriptor_channelz_544d5edb3d49afb2 = []byte{ - // 2317 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe4, 0x59, 0xcd, 0x6f, 0xdb, 0xc8, - 0x15, 0x8f, 0xbe, 0x2c, 0xea, 0x59, 0x52, 0xe8, 0x49, 0x36, 0x4b, 0x2b, 0x1f, 0x4e, 0x99, 0x6c, - 0xe2, 0xcd, 0x22, 0x72, 0xe2, 0x4d, 0x73, 0xe8, 0xb6, 0xc0, 0x2a, 0x8a, 0x9d, 0xd8, 0x75, 0x9c, - 0x80, 0x92, 0x37, 0xd9, 0x1e, 0x96, 0x3b, 0x21, 0x47, 0x36, 0x6b, 0x8a, 0x54, 0x39, 0x23, 0x19, - 0xc9, 0xb5, 0xf7, 0x1e, 0xfb, 0x27, 0xf4, 0xd2, 0x43, 0x81, 0x02, 0x05, 0xda, 0x6b, 0xd1, 0x7f, - 0xa6, 0xff, 0x45, 0x31, 0x1f, 0x24, 0x47, 0x96, 0x64, 0x3b, 0xc8, 0xb1, 0x17, 0x43, 0x7c, 0xfc, - 0xbd, 0xdf, 0x7b, 0xf3, 0xbe, 0x38, 0x33, 0x86, 0xa6, 0x77, 0x84, 0xa3, 0x88, 0x84, 0x1f, 0xdb, - 0xa3, 0x24, 0x66, 0x31, 0x32, 0x0f, 0x93, 0x91, 0xd7, 0xce, 0x84, 0x93, 0xc7, 0xad, 0xd5, 0xc3, - 0x38, 0x3e, 0x0c, 0xc9, 0x86, 0x78, 0xff, 0x7e, 0x3c, 0xd8, 0xc0, 0xd1, 0x07, 0x09, 0x6e, 0xdd, - 0x3a, 0xfd, 0xca, 0x1f, 0x27, 0x98, 0x05, 0x71, 0xa4, 0xde, 0xaf, 0x9d, 0x7e, 0xcf, 0x82, 0x21, - 0xa1, 0x0c, 0x0f, 0x47, 0x8b, 0x08, 0x4e, 0x12, 0x3c, 0x1a, 0x91, 0x84, 0xca, 0xf7, 0xf6, 0xdf, - 0x8a, 0x50, 0xed, 0x4a, 0x5f, 0x50, 0x1b, 0x4a, 0x09, 0x19, 0x58, 0x85, 0xdb, 0x85, 0xf5, 0xe5, - 0xcd, 0x1b, 0xed, 0xd3, 0x7e, 0xb6, 0x15, 0xce, 0x21, 0x03, 0x87, 0x03, 0xd1, 0x63, 0x28, 0xfb, - 0x98, 0x61, 0xab, 0x28, 0x14, 0x6e, 0x2e, 0x54, 0x78, 0x8e, 0x19, 0x76, 0x04, 0x14, 0xfd, 0x06, - 0x96, 0x15, 0xc0, 0xe5, 0xa6, 0x4a, 0xb7, 0x4b, 0xe7, 0x9a, 0x02, 0x2f, 0xfb, 0x8d, 0xb6, 0xa1, - 0x49, 0xc7, 0xef, 0x75, 0x86, 0xb2, 0x60, 0x58, 0x9b, 0x65, 0xe8, 0x65, 0x38, 0x4e, 0xd2, 0xa0, - 0xfa, 0x23, 0xfa, 0x15, 0x00, 0x8d, 0xbd, 0x63, 0xc2, 0x04, 0x47, 0x45, 0x70, 0x5c, 0x9f, 0xc3, - 0x21, 0x30, 0x5c, 0xbf, 0x46, 0xd3, 0x9f, 0xf6, 0x3f, 0x8a, 0x00, 0x39, 0x39, 0x7a, 0xac, 0x07, - 0xed, 0x5c, 0x3f, 0xfe, 0x8f, 0xe3, 0xf6, 0xef, 0x02, 0x58, 0xca, 0xbd, 0x6e, 0x1c, 0x45, 0xc4, - 0x63, 0xc1, 0x24, 0x60, 0x1f, 0x7a, 0x0c, 0x33, 0x82, 0xb6, 0xa1, 0x42, 0xf9, 0x0f, 0x11, 0xc7, - 0xe6, 0xe6, 0xa3, 0x85, 0x2b, 0x9b, 0x51, 0x6d, 0x8b, 0xbf, 0x8e, 0x54, 0xb7, 0x7f, 0x82, 0x8a, - 0x24, 0x5c, 0x86, 0xea, 0xc1, 0xfe, 0x6f, 0xf7, 0x5f, 0xbf, 0xdd, 0x37, 0x2f, 0x21, 0x03, 0xca, - 0x3b, 0xcf, 0xf7, 0xb6, 0xcc, 0x02, 0x6a, 0x02, 0x74, 0x5f, 0xef, 0xef, 0x6f, 0x75, 0xfb, 0x3b, - 0xfb, 0x2f, 0xcc, 0x22, 0xaa, 0x41, 0xc5, 0xd9, 0xea, 0x3c, 0xff, 0xd1, 0x2c, 0xa1, 0x2f, 0x60, - 0xa5, 0xef, 0x74, 0xf6, 0x7b, 0x3b, 0x5b, 0xfb, 0x7d, 0x77, 0xbb, 0xb3, 0xb3, 0x77, 0xe0, 0x6c, - 0x99, 0x65, 0x54, 0x07, 0xa3, 0xf7, 0xf2, 0xa0, 0xff, 0x9c, 0x33, 0x55, 0xec, 0xff, 0x16, 0x61, - 0x59, 0xcb, 0x0e, 0xfa, 0x5e, 0xf7, 0x7b, 0x79, 0xf3, 0xc1, 0xc5, 0xfd, 0x56, 0x1e, 0xa3, 0x6b, - 0xb0, 0xc4, 0x70, 0x72, 0x48, 0x98, 0x28, 0x87, 0x9a, 0xa3, 0x9e, 0xd0, 0x13, 0xa8, 0xb0, 0x04, - 0x7b, 0xc4, 0x2a, 0x09, 0xe6, 0x5b, 0x0b, 0x99, 0xfb, 0x1c, 0xe5, 0x48, 0x30, 0xba, 0x03, 0x0d, - 0x0f, 0x87, 0x21, 0x75, 0x29, 0xc3, 0x09, 0x23, 0xbe, 0x55, 0xbe, 0x5d, 0x58, 0x2f, 0x39, 0x75, - 0x21, 0xec, 0x49, 0x19, 0xba, 0x0f, 0x97, 0x15, 0x68, 0xec, 0x79, 0x84, 0xf8, 0xc4, 0xb7, 0x2a, - 0x02, 0xd6, 0x94, 0xb0, 0x54, 0x8a, 0x7e, 0x01, 0x52, 0xd1, 0x1d, 0xe0, 0x20, 0x24, 0xbe, 0xb5, - 0x24, 0x50, 0xcb, 0x42, 0xb6, 0x2d, 0x44, 0xe8, 0x47, 0xb8, 0x1e, 0x62, 0xca, 0x5c, 0x2e, 0x4b, - 0x8d, 0xba, 0xd9, 0x10, 0xb2, 0xaa, 0xc2, 0xf9, 0x56, 0x5b, 0x4e, 0xa1, 0x76, 0x3a, 0x85, 0xda, - 0xfd, 0x14, 0xe1, 0x58, 0x5c, 0xbd, 0x8b, 0xc3, 0x50, 0x79, 0x97, 0xbd, 0xb1, 0x9b, 0x50, 0xd7, - 0x97, 0x68, 0x7b, 0x00, 0x79, 0x79, 0xa3, 0x9b, 0x90, 0x16, 0xb8, 0x1b, 0xf8, 0x22, 0xfc, 0x25, - 0xa7, 0xa6, 0x24, 0x3b, 0x3e, 0x42, 0x50, 0x8e, 0xf0, 0x90, 0xa8, 0xa0, 0x8a, 0xdf, 0xbb, 0x65, - 0xa3, 0x64, 0x96, 0x77, 0xcb, 0x46, 0xd9, 0xac, 0xec, 0x96, 0x8d, 0x8a, 0xb9, 0xb4, 0x5b, 0x36, - 0x96, 0xcc, 0xea, 0x6e, 0xd9, 0xa8, 0x9a, 0xc6, 0x6e, 0xd9, 0x30, 0xcc, 0x9a, 0x1d, 0x42, 0x63, - 0xaa, 0x03, 0x78, 0x44, 0xb5, 0xd6, 0x09, 0x7c, 0xb1, 0xa4, 0x92, 0x53, 0xcf, 0x85, 0x9a, 0x35, - 0x63, 0xca, 0x5a, 0xc1, 0x2c, 0xee, 0x96, 0x8d, 0xa2, 0x59, 0x5a, 0x64, 0xd9, 0xfe, 0x19, 0x6a, - 0x59, 0xaf, 0xa0, 0xeb, 0xa0, 0xba, 0x85, 0x5b, 0x29, 0x09, 0x2b, 0x86, 0x14, 0x68, 0x16, 0xca, - 0x0b, 0x2d, 0xcc, 0x5f, 0x0f, 0xb7, 0x40, 0x92, 0x09, 0x49, 0x52, 0x0b, 0xe2, 0x81, 0x5b, 0xa8, - 0x28, 0x0b, 0x42, 0xa0, 0x59, 0x58, 0xba, 0xd0, 0x1a, 0x72, 0x0b, 0x7f, 0x2d, 0xc0, 0x92, 0x34, - 0x81, 0x1e, 0xea, 0xb3, 0x70, 0xde, 0x5c, 0x48, 0x3d, 0x91, 0x73, 0xf0, 0xd1, 0xd4, 0x1c, 0xbc, - 0xb1, 0x08, 0xaf, 0x8d, 0xc1, 0xef, 0xa1, 0x11, 0x06, 0x94, 0x91, 0xc8, 0x95, 0x81, 0x51, 0x83, - 0xf0, 0xcc, 0x11, 0x54, 0x97, 0x1a, 0x52, 0x60, 0xff, 0x89, 0x4f, 0xef, 0x8c, 0x36, 0xef, 0xb2, - 0xc2, 0x67, 0x75, 0x59, 0xf1, 0x62, 0x5d, 0x56, 0xba, 0x50, 0x97, 0x95, 0x3f, 0xb9, 0xcb, 0x2a, - 0x9f, 0xd1, 0x65, 0x7f, 0x29, 0xc2, 0x92, 0x8c, 0xcd, 0xf9, 0xe9, 0xcb, 0x62, 0x7a, 0xc1, 0xf4, - 0x09, 0xbc, 0x96, 0xbe, 0x0d, 0xa8, 0x84, 0xb1, 0x87, 0x43, 0x35, 0xd3, 0x56, 0x67, 0x55, 0x3a, - 0xbe, 0x9f, 0x10, 0x4a, 0x1d, 0x89, 0x43, 0x8f, 0x61, 0x29, 0x21, 0xc3, 0x98, 0xc9, 0xba, 0x3f, - 0x53, 0x43, 0x01, 0xd1, 0x53, 0x30, 0x28, 0xf1, 0xc6, 0x49, 0xc0, 0x3e, 0xe4, 0x71, 0x99, 0x2d, - 0x2c, 0x89, 0x70, 0x32, 0x2c, 0x5a, 0x83, 0x65, 0xc9, 0xe0, 0x6a, 0x5d, 0x00, 0x52, 0xb4, 0x8f, - 0x87, 0xc4, 0xfe, 0x63, 0x15, 0x20, 0x5f, 0x11, 0x4f, 0x2f, 0x65, 0x09, 0xc1, 0xc3, 0xbc, 0x0a, - 0xe4, 0x10, 0x6a, 0x2a, 0x71, 0x5a, 0x07, 0xdf, 0xc0, 0x4a, 0x06, 0xcc, 0x2a, 0x41, 0x16, 0x8c, - 0x99, 0x42, 0xb3, 0x5a, 0xf8, 0x0a, 0x52, 0xf5, 0xb4, 0x1a, 0x64, 0xcd, 0x34, 0x94, 0x54, 0xd5, - 0xc3, 0x1d, 0x68, 0x0c, 0x09, 0xa5, 0xf8, 0x90, 0x50, 0x97, 0x92, 0x88, 0xa5, 0x63, 0x3e, 0x15, - 0xf6, 0x48, 0xc4, 0xb8, 0xe1, 0x0c, 0x94, 0x10, 0x8f, 0x04, 0x93, 0x6c, 0xd0, 0x9b, 0xe9, 0x0b, - 0x47, 0xc9, 0xd1, 0x3a, 0x98, 0xc7, 0x84, 0x8c, 0x5c, 0x1c, 0x06, 0x93, 0x94, 0x54, 0x8e, 0xfb, - 0x26, 0x97, 0x77, 0x84, 0x58, 0xd0, 0x1e, 0xc1, 0x1d, 0x51, 0x8b, 0x22, 0x43, 0xae, 0xf4, 0xcb, - 0xf5, 0x12, 0x82, 0x3f, 0x75, 0xf2, 0xaf, 0x71, 0x9a, 0x3d, 0xce, 0xd2, 0x13, 0x24, 0x5d, 0xc9, - 0x91, 0x01, 0xd0, 0xef, 0xe1, 0xae, 0xb0, 0xa4, 0xf2, 0xb2, 0xd0, 0x94, 0x71, 0xae, 0xa9, 0xdb, - 0x9c, 0xc7, 0x11, 0x34, 0x0b, 0x6c, 0xa5, 0x1d, 0xa6, 0x02, 0x23, 0x02, 0xa0, 0x99, 0xa8, 0x5d, - 0xac, 0xc3, 0x5e, 0x49, 0x6d, 0x1e, 0xa7, 0x9c, 0x1a, 0xc3, 0xda, 0x14, 0x75, 0x9a, 0x0b, 0x8d, - 0x1e, 0xce, 0xa5, 0xbf, 0xa1, 0xd1, 0xa7, 0x49, 0xcb, 0x4d, 0xfc, 0x00, 0xab, 0x32, 0x1d, 0x83, - 0x30, 0x3e, 0x71, 0xbd, 0x38, 0x62, 0x49, 0x1c, 0xba, 0x27, 0x41, 0xe4, 0xc7, 0x27, 0xd6, 0x72, - 0xda, 0xcf, 0xa7, 0xc8, 0x77, 0x22, 0xf6, 0xf4, 0xc9, 0x0f, 0x38, 0x1c, 0x13, 0xe7, 0x9a, 0xd0, - 0xde, 0x0e, 0xe3, 0x93, 0xae, 0xd4, 0x7d, 0x2b, 0x54, 0xd1, 0x3b, 0x68, 0xa9, 0xe0, 0xcf, 0x23, - 0xae, 0x9f, 0x4f, 0xfc, 0xa5, 0x54, 0x9f, 0x65, 0x7e, 0x0a, 0x4b, 0xf1, 0x88, 0x1f, 0x64, 0xac, - 0x86, 0x18, 0xe1, 0xb7, 0x16, 0x8d, 0x8f, 0xd7, 0x02, 0xe5, 0x28, 0xb4, 0xfd, 0xcf, 0x12, 0x54, - 0x55, 0xcb, 0xa3, 0x57, 0xd0, 0x60, 0xde, 0x28, 0x18, 0xb9, 0x58, 0x0a, 0xd4, 0xe4, 0xba, 0xb7, - 0x70, 0x48, 0xb4, 0xfb, 0xde, 0x68, 0x67, 0xa4, 0x1e, 0x5e, 0x5e, 0x72, 0xea, 0x42, 0x3d, 0xa5, - 0x7b, 0x01, 0xcb, 0x63, 0x9f, 0x66, 0x64, 0x72, 0xac, 0xdd, 0x5d, 0x4c, 0x76, 0xe0, 0xd3, 0x9c, - 0x0a, 0xc6, 0xd9, 0x13, 0xf7, 0x2b, 0x66, 0x47, 0x24, 0xc9, 0xa8, 0x4a, 0xe7, 0xf9, 0xf5, 0x9a, - 0xc3, 0x35, 0xbf, 0x62, 0xed, 0xb9, 0xd5, 0x81, 0xba, 0xee, 0x37, 0xdf, 0xf9, 0x9c, 0x5a, 0x73, - 0xdd, 0xa9, 0xe5, 0xcb, 0x40, 0x50, 0x1e, 0xc5, 0x89, 0xdc, 0x4e, 0x56, 0x1c, 0xf1, 0xbb, 0xb5, - 0x0e, 0x90, 0x7b, 0x8b, 0x5a, 0x60, 0x0c, 0x82, 0x90, 0x88, 0x39, 0x57, 0x10, 0x73, 0x2e, 0x7b, - 0x6e, 0xed, 0x43, 0x5d, 0x77, 0x26, 0xdb, 0x15, 0x14, 0xf2, 0x5d, 0x01, 0x7a, 0x00, 0x95, 0x09, - 0xcf, 0xae, 0x0a, 0xd1, 0xd5, 0x99, 0x02, 0xe8, 0x44, 0x1f, 0x1c, 0x09, 0x79, 0x56, 0x83, 0xaa, - 0xf2, 0xd4, 0xfe, 0x73, 0x09, 0x8c, 0x74, 0xf0, 0xa2, 0x4d, 0x28, 0xb1, 0x90, 0x2e, 0xfe, 0xec, - 0xa6, 0xc0, 0x76, 0x3f, 0xe4, 0x11, 0xe1, 0x60, 0xbe, 0xd9, 0x16, 0x81, 0x51, 0x76, 0xd7, 0xcf, - 0xd0, 0x12, 0x6b, 0x48, 0x9f, 0x5e, 0x5e, 0x72, 0xa4, 0x62, 0xeb, 0x5f, 0x05, 0x28, 0xf5, 0x43, - 0x8a, 0xbe, 0x82, 0x06, 0x65, 0x38, 0xf2, 0x71, 0xe2, 0xbb, 0xf9, 0xf2, 0x78, 0xe4, 0x53, 0x31, - 0x1f, 0xf9, 0x68, 0x0d, 0x40, 0x26, 0x32, 0xdf, 0x4a, 0xbe, 0xbc, 0xe4, 0xd4, 0x84, 0x4c, 0x00, - 0xbe, 0x81, 0x15, 0xd9, 0x77, 0x1e, 0x49, 0x58, 0x30, 0x08, 0x3c, 0x7e, 0x14, 0x28, 0x89, 0x8c, - 0x98, 0xe2, 0x45, 0x37, 0x97, 0xa3, 0x87, 0x80, 0x54, 0x33, 0xe9, 0xe8, 0xb2, 0x40, 0xaf, 0xc8, - 0x37, 0x1a, 0xfc, 0x59, 0x13, 0xea, 0x5e, 0x30, 0xe2, 0xd6, 0xe9, 0x38, 0x60, 0xa4, 0xf5, 0x1a, - 0x1a, 0x53, 0xab, 0xfa, 0xec, 0xd4, 0x54, 0xa1, 0x32, 0x8c, 0x7d, 0x12, 0xda, 0x11, 0xd4, 0xf5, - 0x5e, 0x9b, 0x4b, 0x7c, 0x55, 0x27, 0xae, 0x29, 0x0a, 0xf4, 0x04, 0x00, 0xfb, 0x7e, 0xc0, 0xb5, - 0xb2, 0xaf, 0xfa, 0x7c, 0x9b, 0x1a, 0xce, 0xde, 0x83, 0x2b, 0xba, 0x3d, 0x3e, 0xc6, 0xe2, 0x31, - 0x43, 0xbf, 0x04, 0x23, 0xbd, 0xdd, 0x50, 0x75, 0xb1, 0x3a, 0x43, 0xf5, 0x5c, 0x01, 0x9c, 0x0c, - 0x6a, 0x7b, 0x80, 0x74, 0xb6, 0xbd, 0x20, 0x3a, 0x24, 0x09, 0x3f, 0x56, 0x61, 0x7e, 0xdc, 0x92, - 0xab, 0x30, 0x1c, 0xf5, 0x34, 0x65, 0xa4, 0x78, 0x71, 0x23, 0x7f, 0x37, 0x4e, 0xf9, 0xec, 0x8d, - 0x76, 0xa2, 0x41, 0xcc, 0x7b, 0x91, 0xcf, 0x10, 0x37, 0x3f, 0x04, 0x36, 0x9c, 0x1a, 0x97, 0xc8, - 0x53, 0xa8, 0x2d, 0x27, 0x94, 0xeb, 0x61, 0x85, 0x28, 0x0a, 0xc4, 0x32, 0x17, 0x76, 0xb1, 0xc4, - 0x7c, 0x0d, 0xa6, 0xc0, 0x24, 0x84, 0x25, 0x38, 0xa2, 0xc3, 0x80, 0xc9, 0x81, 0xd1, 0x70, 0x2e, - 0x73, 0xb9, 0x93, 0x8b, 0xf9, 0x1e, 0x45, 0x40, 0x47, 0x49, 0xfc, 0x9e, 0x50, 0x51, 0x3a, 0x0d, - 0x47, 0x38, 0xf0, 0x46, 0x48, 0xf8, 0x56, 0x52, 0x00, 0xde, 0x63, 0xef, 0x38, 0x1e, 0x0c, 0xc4, - 0xd7, 0x5e, 0x99, 0x7b, 0x26, 0x45, 0x19, 0x44, 0xce, 0x53, 0x2a, 0x3e, 0xf2, 0x0a, 0x22, 0x97, - 0x46, 0xd1, 0x3d, 0xb8, 0x2c, 0x17, 0x15, 0xf9, 0xee, 0x09, 0xf5, 0x70, 0x48, 0xc4, 0xd7, 0xbc, - 0xe1, 0x88, 0xc5, 0xf4, 0x22, 0xff, 0xad, 0x10, 0x66, 0xb8, 0xc4, 0x9b, 0xa4, 0x38, 0x23, 0xc7, - 0x39, 0xde, 0x44, 0xe1, 0x56, 0xc1, 0x90, 0x38, 0x16, 0x8b, 0x0f, 0x69, 0xc3, 0xa9, 0x0a, 0x00, - 0x8b, 0xb3, 0x57, 0x98, 0xc5, 0xe2, 0x23, 0xa8, 0x5e, 0x75, 0x58, 0x8c, 0x6e, 0x2b, 0x47, 0xb9, - 0x17, 0x43, 0x4a, 0xc5, 0x67, 0x4c, 0xad, 0xb6, 0x17, 0xf9, 0xaf, 0x28, 0xcd, 0x10, 0xdc, 0x3e, - 0x47, 0xd4, 0x73, 0x84, 0xe3, 0x4d, 0x38, 0x22, 0x5d, 0xec, 0x38, 0xc2, 0xde, 0x31, 0xf1, 0xad, - 0x46, 0xbe, 0xd8, 0x03, 0x29, 0xca, 0x62, 0x4a, 0x25, 0xa2, 0xa9, 0x59, 0x91, 0x80, 0xeb, 0x20, - 0x12, 0xea, 0x86, 0x31, 0x65, 0xd6, 0x65, 0xf1, 0x5a, 0xf8, 0xbc, 0x17, 0x53, 0x96, 0x19, 0x50, - 0xc9, 0xb3, 0xcc, 0xdc, 0x80, 0x4a, 0x5c, 0x06, 0x19, 0x70, 0x3a, 0x46, 0xad, 0x95, 0x1c, 0xb2, - 0x2d, 0x45, 0xe8, 0x21, 0x5c, 0x91, 0x26, 0xf8, 0x36, 0x81, 0xef, 0x94, 0xe5, 0xfe, 0x0b, 0x09, - 0xa4, 0xa8, 0x8e, 0x3d, 0x4c, 0xc5, 0xb6, 0x53, 0x6d, 0xec, 0x50, 0x0e, 0xc7, 0xde, 0xb1, 0x44, - 0x5f, 0xc9, 0x6b, 0x86, 0xa3, 0x3b, 0xde, 0xb1, 0x00, 0xcf, 0x72, 0x27, 0xc4, 0x9b, 0x58, 0x57, - 0x67, 0xb9, 0x1d, 0xe2, 0x4d, 0x66, 0xb9, 0x05, 0xfa, 0x8b, 0x19, 0x6e, 0x01, 0x4e, 0x43, 0x33, - 0x1a, 0xb2, 0xb1, 0x75, 0x2d, 0x0f, 0xcd, 0x9b, 0x21, 0x1b, 0xa3, 0x07, 0xb0, 0x92, 0x65, 0x87, - 0x52, 0x76, 0x94, 0x10, 0x7a, 0x64, 0x7d, 0xa9, 0x15, 0xb6, 0x37, 0xe9, 0x29, 0xb1, 0x56, 0x21, - 0xcc, 0xb2, 0xf4, 0x0a, 0x61, 0x59, 0x7e, 0x12, 0xc6, 0x26, 0x38, 0xb1, 0x56, 0xb5, 0x1c, 0x0b, - 0x49, 0x66, 0x87, 0xd7, 0x49, 0x66, 0xa7, 0x95, 0xdb, 0xe9, 0x45, 0x7e, 0x66, 0x27, 0xed, 0x47, - 0x8e, 0xf5, 0x4e, 0x22, 0xdf, 0xba, 0x9e, 0x27, 0xa3, 0x17, 0xf9, 0xdd, 0x93, 0x28, 0x2f, 0x08, - 0xec, 0x4f, 0x78, 0x51, 0xdd, 0xc8, 0x0d, 0x76, 0x84, 0x84, 0xef, 0xfc, 0x55, 0xce, 0xe3, 0xc4, - 0x27, 0x49, 0x10, 0x1d, 0x5a, 0x37, 0x05, 0xa8, 0x29, 0xd3, 0x9e, 0x4a, 0xed, 0x0e, 0x7c, 0xf1, - 0x82, 0xb0, 0x7e, 0x3c, 0x52, 0x67, 0x48, 0xea, 0x90, 0x3f, 0x8c, 0x09, 0x65, 0x7c, 0xb3, 0x2d, - 0xce, 0x0c, 0xee, 0xcc, 0x0d, 0x46, 0x53, 0xc8, 0xbb, 0xe9, 0xc5, 0x82, 0xed, 0xc2, 0xb5, 0xd3, - 0x14, 0x74, 0x14, 0x47, 0x94, 0xa0, 0x6f, 0xa1, 0xaa, 0xb4, 0xad, 0x82, 0xd8, 0x41, 0xad, 0x2e, - 0xbe, 0x0d, 0x4c, 0x91, 0xc8, 0x84, 0x12, 0x89, 0xe4, 0xe9, 0xc3, 0x70, 0xf8, 0x4f, 0xfb, 0x3b, - 0x58, 0x79, 0x41, 0x98, 0x3c, 0x11, 0x67, 0xfe, 0xdd, 0xe3, 0x67, 0x1b, 0xee, 0x5f, 0x7e, 0x5b, - 0x50, 0x48, 0x8f, 0x21, 0x38, 0x51, 0xe8, 0x1d, 0xdf, 0x7e, 0x07, 0x48, 0x57, 0x56, 0x9e, 0x3d, - 0x82, 0x25, 0xa9, 0xa7, 0x1c, 0xb3, 0x16, 0x5e, 0x04, 0x28, 0xdc, 0x1c, 0xb7, 0x7e, 0x82, 0x2f, - 0x33, 0x66, 0x39, 0x77, 0x33, 0xe7, 0xa6, 0x2e, 0x31, 0x0a, 0xa7, 0x2e, 0x31, 0x72, 0xcf, 0xb3, - 0x9b, 0x94, 0xa2, 0xee, 0xb9, 0xba, 0x4e, 0xb1, 0x8f, 0xc0, 0x9a, 0xe5, 0x57, 0xfe, 0x4f, 0x5f, - 0x72, 0x16, 0x3e, 0xe5, 0x92, 0x73, 0xce, 0x4a, 0x36, 0x45, 0x80, 0xb3, 0x8b, 0x2b, 0xb9, 0x86, - 0xb3, 0x2f, 0xaf, 0xec, 0x1d, 0x11, 0xd7, 0x4c, 0x67, 0x5e, 0xc6, 0x0b, 0x17, 0xcb, 0xb8, 0xfd, - 0x1d, 0x5c, 0xe5, 0x0b, 0xd5, 0xae, 0xb4, 0xa4, 0x07, 0x33, 0xd7, 0x5a, 0x85, 0xd9, 0x6b, 0x2d, - 0xfb, 0x40, 0x14, 0xb0, 0xae, 0xac, 0x5c, 0xf9, 0x35, 0x40, 0x0e, 0x5c, 0xfc, 0x0f, 0x03, 0x4d, - 0x53, 0xc3, 0xdb, 0x1b, 0x60, 0x72, 0x5a, 0x15, 0xb4, 0x3c, 0xab, 0x59, 0xca, 0x0a, 0xd3, 0x97, - 0x5f, 0xf6, 0x96, 0x2c, 0x52, 0xa5, 0xa0, 0x95, 0x99, 0xbc, 0x04, 0x92, 0xf6, 0xad, 0x85, 0x29, - 0x52, 0xb8, 0xcd, 0xff, 0x94, 0xc1, 0x50, 0x01, 0xfa, 0x88, 0x3c, 0x68, 0x4e, 0x77, 0x16, 0xba, - 0x3f, 0x4b, 0x30, 0xb7, 0x7d, 0x5b, 0xeb, 0xe7, 0x03, 0x95, 0x8f, 0x6f, 0x01, 0xf2, 0x06, 0x41, - 0x77, 0xe6, 0xea, 0x4d, 0xf7, 0x5e, 0xeb, 0xee, 0xd9, 0x20, 0x45, 0x1c, 0xc8, 0x10, 0xea, 0xf5, - 0x8b, 0xbe, 0x3e, 0x43, 0x73, 0xba, 0x87, 0x5a, 0x0f, 0x2e, 0x02, 0x9d, 0x5a, 0x43, 0xfa, 0x3f, - 0xa2, 0xf9, 0x6b, 0x98, 0x2e, 0xef, 0x05, 0x6b, 0x38, 0x5d, 0xcf, 0x3f, 0x43, 0x63, 0xaa, 0xba, - 0xd0, 0xbd, 0xf9, 0x5e, 0x9d, 0xae, 0xdd, 0xd6, 0xfd, 0x73, 0x71, 0xca, 0x42, 0x1f, 0x6a, 0x59, - 0xdd, 0x20, 0x7b, 0xbe, 0x96, 0x5e, 0x85, 0xad, 0x3b, 0x67, 0x62, 0x24, 0xeb, 0xb3, 0x77, 0x70, - 0x25, 0x88, 0x67, 0x80, 0xcf, 0x1a, 0x69, 0x69, 0xbd, 0xe1, 0xfb, 0xc8, 0x37, 0x85, 0xdf, 0x3d, - 0x52, 0xfb, 0xca, 0xc3, 0x38, 0xc4, 0xd1, 0x61, 0x3b, 0x4e, 0x0e, 0x37, 0xb8, 0xd2, 0x46, 0xaa, - 0x24, 0x9e, 0xd2, 0x4f, 0xc0, 0x47, 0x77, 0xf2, 0xf8, 0xfd, 0x92, 0xd8, 0x82, 0x7e, 0xfb, 0xbf, - 0x00, 0x00, 0x00, 0xff, 0xff, 0xce, 0xbe, 0x6e, 0x8a, 0x33, 0x1c, 0x00, 0x00, +func init() { + proto.RegisterFile("grpc/channelz/v1/channelz.proto", fileDescriptor_channelz_ce3ed45d08251f2f) +} + +var fileDescriptor_channelz_ce3ed45d08251f2f = []byte{ + // 2515 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe4, 0x59, 0x4b, 0x6f, 0xdb, 0xca, + 0x15, 0xb6, 0xde, 0xd4, 0xd1, 0x23, 0xf2, 0xe4, 0x45, 0x2b, 0x0f, 0xbb, 0x74, 0x6e, 0xe2, 0x9b, + 0x34, 0x72, 0xec, 0x9b, 0x06, 0x45, 0xd3, 0xa2, 0x57, 0x56, 0xec, 0x58, 0xae, 0x23, 0x07, 0x94, + 0x7c, 0x93, 0xdb, 0xc5, 0xe5, 0x65, 0xc8, 0xb1, 0xcc, 0x9a, 0x22, 0x55, 0xce, 0x48, 0x46, 0xb2, + 0xed, 0xbe, 0xcb, 0xa2, 0xbf, 0xa0, 0x9b, 0x2e, 0x0a, 0x14, 0x28, 0xd0, 0x6e, 0x8b, 0xfe, 0x99, + 0xfe, 0x8b, 0x62, 0x1e, 0x7c, 0xc8, 0x92, 0x2c, 0x07, 0x59, 0x76, 0x63, 0x88, 0x87, 0xdf, 0xf9, + 0xce, 0x99, 0xf3, 0x9a, 0xe1, 0x18, 0x56, 0xfb, 0xc1, 0xd0, 0xda, 0xb4, 0x4e, 0x4d, 0xcf, 0xc3, + 0xee, 0xa7, 0xcd, 0xf1, 0x56, 0xf4, 0xbb, 0x31, 0x0c, 0x7c, 0xea, 0xa3, 0x1a, 0x03, 0x34, 0x22, + 0xe1, 0x78, 0xab, 0xbe, 0xd2, 0xf7, 0xfd, 0xbe, 0x8b, 0x37, 0xf9, 0xfb, 0x0f, 0xa3, 0x93, 0x4d, + 0xd3, 0xfb, 0x28, 0xc0, 0xf5, 0xfb, 0x17, 0x5f, 0xd9, 0xa3, 0xc0, 0xa4, 0x8e, 0xef, 0xc9, 0xf7, + 0xab, 0x17, 0xdf, 0x53, 0x67, 0x80, 0x09, 0x35, 0x07, 0xc3, 0x79, 0x04, 0xe7, 0x81, 0x39, 0x1c, + 0xe2, 0x80, 0x88, 0xf7, 0xda, 0xdf, 0xd2, 0x50, 0x68, 0x09, 0x5f, 0x50, 0x03, 0x32, 0x01, 0x3e, + 0x51, 0x53, 0x6b, 0xa9, 0x8d, 0xd2, 0xf6, 0xdd, 0xc6, 0x45, 0x3f, 0x1b, 0x12, 0xa7, 0xe3, 0x13, + 0x9d, 0x01, 0xd1, 0x16, 0x64, 0x6d, 0x93, 0x9a, 0x6a, 0x9a, 0x2b, 0xdc, 0x9b, 0xab, 0xf0, 0xca, + 0xa4, 0xa6, 0xce, 0xa1, 0xe8, 0x57, 0x50, 0x92, 0x00, 0x83, 0x99, 0xca, 0xac, 0x65, 0x16, 0x9a, + 0x02, 0x2b, 0xfa, 0x8d, 0xf6, 0xa0, 0x4a, 0x46, 0x1f, 0x92, 0x0c, 0x59, 0xce, 0xb0, 0x3a, 0xcd, + 0xd0, 0x8d, 0x70, 0x8c, 0xa4, 0x42, 0x92, 0x8f, 0xe8, 0x17, 0x00, 0xc4, 0xb7, 0xce, 0x30, 0xe5, + 0x1c, 0x39, 0xce, 0x71, 0x67, 0x06, 0x07, 0xc7, 0x30, 0xfd, 0x22, 0x09, 0x7f, 0x6a, 0xff, 0x48, + 0x03, 0xc4, 0xe4, 0x68, 0x2b, 0x19, 0xb4, 0x85, 0x7e, 0xfc, 0x1f, 0xc7, 0xed, 0xdf, 0x29, 0x50, + 0xa5, 0x7b, 0x2d, 0xdf, 0xf3, 0xb0, 0x45, 0x9d, 0xb1, 0x43, 0x3f, 0x76, 0xa9, 0x49, 0x31, 0xda, + 0x83, 0x1c, 0x61, 0x3f, 0x78, 0x1c, 0xab, 0xdb, 0xcf, 0xe6, 0xae, 0x6c, 0x4a, 0xb5, 0xc1, 0xff, + 0xea, 0x42, 0x5d, 0xfb, 0x01, 0x72, 0x82, 0xb0, 0x04, 0x85, 0xe3, 0xce, 0x6f, 0x3a, 0x47, 0xef, + 0x3a, 0xb5, 0x25, 0xa4, 0x40, 0xb6, 0xfd, 0xea, 0x70, 0xb7, 0x96, 0x42, 0x55, 0x80, 0xd6, 0x51, + 0xa7, 0xb3, 0xdb, 0xea, 0xb5, 0x3b, 0xaf, 0x6b, 0x69, 0x54, 0x84, 0x9c, 0xbe, 0xdb, 0x7c, 0xf5, + 0x7d, 0x2d, 0x83, 0x6e, 0xc2, 0x72, 0x4f, 0x6f, 0x76, 0xba, 0xed, 0xdd, 0x4e, 0xcf, 0xd8, 0x6b, + 0xb6, 0x0f, 0x8f, 0xf5, 0xdd, 0x5a, 0x16, 0x95, 0x41, 0xe9, 0xee, 0x1f, 0xf7, 0x5e, 0x31, 0xa6, + 0x9c, 0xf6, 0xdf, 0x34, 0x94, 0x12, 0xd9, 0x41, 0xdf, 0x26, 0xfd, 0x2e, 0x6d, 0x3f, 0xbe, 0xba, + 0xdf, 0xd2, 0x63, 0x74, 0x0b, 0xf2, 0xd4, 0x0c, 0xfa, 0x98, 0xf2, 0x72, 0x28, 0xea, 0xf2, 0x09, + 0x3d, 0x87, 0x1c, 0x0d, 0x4c, 0x0b, 0xab, 0x19, 0xce, 0x7c, 0x7f, 0x2e, 0x73, 0x8f, 0xa1, 0x74, + 0x01, 0x46, 0xeb, 0x50, 0xb1, 0x4c, 0xd7, 0x25, 0x06, 0xa1, 0x66, 0x40, 0xb1, 0xad, 0x66, 0xd7, + 0x52, 0x1b, 0x19, 0xbd, 0xcc, 0x85, 0x5d, 0x21, 0x43, 0x8f, 0xe0, 0x9a, 0x04, 0x8d, 0x2c, 0x0b, + 0x63, 0x1b, 0xdb, 0x6a, 0x8e, 0xc3, 0xaa, 0x02, 0x16, 0x4a, 0xd1, 0x4f, 0x40, 0x28, 0x1a, 0x27, + 0xa6, 0xe3, 0x62, 0x5b, 0xcd, 0x73, 0x54, 0x89, 0xcb, 0xf6, 0xb8, 0x08, 0x7d, 0x0f, 0x77, 0x5c, + 0x93, 0x50, 0x83, 0xc9, 0x42, 0xa3, 0x46, 0x34, 0x84, 0xd4, 0x02, 0x77, 0xbe, 0xde, 0x10, 0x53, + 0xa8, 0x11, 0x4e, 0xa1, 0x46, 0x2f, 0x44, 0xe8, 0x2a, 0x53, 0x6f, 0x99, 0xae, 0x2b, 0xbd, 0x8b, + 0xde, 0x68, 0x7f, 0xce, 0xc0, 0x72, 0x72, 0x8d, 0xbb, 0x63, 0xec, 0x51, 0xb4, 0x06, 0x25, 0x1b, + 0x13, 0x2b, 0x70, 0x86, 0x6c, 0x0c, 0xf2, 0xb8, 0x17, 0xf5, 0xa4, 0x08, 0xed, 0x83, 0x42, 0xf0, + 0x18, 0x07, 0x0e, 0xfd, 0xc8, 0x63, 0x5a, 0xdd, 0xfe, 0xe9, 0xe5, 0xc1, 0xe3, 0xc4, 0x8d, 0xae, + 0xd4, 0xd1, 0x23, 0x6d, 0xf4, 0x73, 0x28, 0xc6, 0x4b, 0xc9, 0x2c, 0x5c, 0x4a, 0x0c, 0x46, 0xbf, + 0x9e, 0xec, 0xd7, 0xec, 0xe2, 0x91, 0xba, 0xbf, 0x34, 0xd1, 0xb1, 0xfb, 0x53, 0x1d, 0x9b, 0xbb, + 0xd2, 0x84, 0xd9, 0x5f, 0xba, 0xd0, 0xb3, 0xda, 0x2e, 0x28, 0xe1, 0xd2, 0x78, 0xf9, 0xf7, 0x8c, + 0xb8, 0x31, 0x4a, 0x50, 0x68, 0xf5, 0x8c, 0x76, 0x67, 0xef, 0x48, 0xf6, 0x46, 0xcf, 0x78, 0xd7, + 0xd4, 0x3b, 0xa2, 0x37, 0xca, 0xa0, 0xb4, 0x7a, 0xc6, 0xae, 0xae, 0x1f, 0xe9, 0xb5, 0xcc, 0x4e, + 0x09, 0x8a, 0xd6, 0xa9, 0xe3, 0xda, 0xcc, 0x17, 0xd6, 0xcb, 0xe5, 0x64, 0x04, 0xd1, 0x63, 0x58, + 0xf6, 0x46, 0x03, 0x03, 0xb3, 0x48, 0x12, 0xc3, 0xf5, 0xfb, 0x7d, 0x6c, 0xf3, 0xdc, 0x64, 0xf4, + 0x6b, 0xde, 0x68, 0xc0, 0x23, 0x4c, 0x0e, 0xb9, 0x18, 0xb5, 0x01, 0x59, 0x01, 0xe6, 0xbb, 0x58, + 0xa2, 0x52, 0xd2, 0x0b, 0xc3, 0xbb, 0x1c, 0x6a, 0x45, 0x22, 0xf4, 0x12, 0xf2, 0xc2, 0xa4, 0x9c, + 0x88, 0xeb, 0x57, 0x48, 0xb4, 0x2e, 0x55, 0x34, 0x0b, 0x20, 0x0e, 0x3f, 0xba, 0x07, 0x61, 0xf8, + 0x0d, 0x27, 0x74, 0xbd, 0x28, 0x25, 0x6d, 0x1b, 0x21, 0xc8, 0x7a, 0xe6, 0x00, 0xcb, 0x26, 0xe5, + 0xbf, 0x0f, 0xb2, 0x4a, 0xa6, 0x96, 0x3d, 0xc8, 0x2a, 0xd9, 0x5a, 0xee, 0x20, 0xab, 0xe4, 0x6a, + 0xf9, 0x83, 0xac, 0x92, 0xaf, 0x15, 0x0e, 0xb2, 0x4a, 0xa1, 0xa6, 0x1c, 0x64, 0x15, 0xa5, 0x56, + 0xd4, 0x5c, 0xa8, 0x4c, 0xe4, 0x87, 0x75, 0x68, 0x22, 0xb1, 0x8e, 0xcd, 0x5b, 0x24, 0xa3, 0x97, + 0x63, 0x61, 0xc2, 0x9a, 0x32, 0x61, 0x2d, 0x55, 0x4b, 0x1f, 0x64, 0x95, 0x74, 0x2d, 0x33, 0xcf, + 0xb2, 0xf6, 0x23, 0x14, 0xa3, 0xd9, 0x8b, 0xee, 0x80, 0x9c, 0xbe, 0xcc, 0x4a, 0x86, 0x5b, 0x51, + 0x84, 0x20, 0x61, 0x21, 0x3b, 0xd7, 0xc2, 0xec, 0xf5, 0x30, 0x0b, 0x38, 0x18, 0xe3, 0x20, 0xb4, + 0xc0, 0x1f, 0x98, 0x85, 0x9c, 0xb4, 0xc0, 0x05, 0x09, 0x0b, 0xf9, 0x2b, 0xad, 0x21, 0xb6, 0xf0, + 0xd7, 0x14, 0xe4, 0x85, 0x09, 0xf4, 0x34, 0xb9, 0xb7, 0xce, 0xda, 0x67, 0x42, 0x4f, 0xc4, 0xbe, + 0xfa, 0x6c, 0x62, 0x5f, 0xbd, 0x3b, 0x0f, 0x9f, 0xd8, 0x56, 0xbf, 0x85, 0x8a, 0xeb, 0x10, 0x8a, + 0x3d, 0x43, 0x04, 0x46, 0x96, 0xd1, 0xa5, 0x5b, 0x5a, 0x59, 0x68, 0x08, 0x81, 0xf6, 0x47, 0x76, + 0x1a, 0x88, 0x68, 0xe3, 0xa9, 0x9d, 0xfa, 0xa2, 0xa9, 0x9d, 0xbe, 0xda, 0xd4, 0xce, 0x5c, 0x69, + 0x6a, 0x67, 0x3f, 0x7b, 0x6a, 0xe7, 0xbe, 0x60, 0x6a, 0xff, 0x25, 0x0d, 0x79, 0x11, 0x9b, 0xc5, + 0xe9, 0x8b, 0x62, 0x7a, 0xc5, 0xf4, 0x71, 0x7c, 0x22, 0x7d, 0x9b, 0x90, 0x73, 0x7d, 0xcb, 0x74, + 0xe5, 0x6c, 0x5e, 0x99, 0x56, 0x69, 0xda, 0x76, 0x80, 0x09, 0xd1, 0x05, 0x0e, 0x6d, 0x41, 0x3e, + 0xc0, 0x03, 0x9f, 0x62, 0x39, 0x91, 0x2f, 0xd1, 0x90, 0x40, 0xf4, 0x82, 0xed, 0x26, 0xd6, 0x88, + 0xef, 0x26, 0x51, 0x5c, 0xa6, 0x0b, 0x4b, 0x20, 0xf4, 0x08, 0x8b, 0x56, 0xa1, 0x24, 0x18, 0x8c, + 0x44, 0x17, 0x80, 0x10, 0x75, 0xcc, 0x01, 0xd6, 0xfe, 0x50, 0x00, 0x88, 0x57, 0xc4, 0xd2, 0x4b, + 0x68, 0x80, 0xcd, 0x41, 0x5c, 0x05, 0x62, 0x08, 0x55, 0xa5, 0x38, 0xac, 0x83, 0x27, 0xb0, 0x1c, + 0x01, 0xa3, 0x4a, 0x10, 0x05, 0x53, 0x0b, 0xa1, 0x51, 0x2d, 0x7c, 0x05, 0xa1, 0x7a, 0x58, 0x0d, + 0xa2, 0x66, 0x2a, 0x52, 0x2a, 0xeb, 0x61, 0x1d, 0x2a, 0x03, 0x4c, 0x88, 0xd9, 0xc7, 0xc4, 0x20, + 0xd8, 0xa3, 0xe1, 0xb1, 0x21, 0x14, 0x76, 0xd9, 0xce, 0xfb, 0x04, 0x96, 0x23, 0x50, 0x80, 0x2d, + 0xec, 0x8c, 0xa3, 0x83, 0x43, 0x2d, 0x7c, 0xa1, 0x4b, 0x39, 0xda, 0x80, 0xda, 0x19, 0xc6, 0x43, + 0xc3, 0x74, 0x9d, 0x71, 0x48, 0x2a, 0x8e, 0x0f, 0x55, 0x26, 0x6f, 0x72, 0x31, 0xa7, 0x3d, 0x85, + 0x75, 0x5e, 0x8b, 0x3c, 0x43, 0x86, 0xf0, 0xcb, 0xe0, 0xa3, 0xfe, 0x33, 0x4f, 0x12, 0xab, 0x8c, + 0xe6, 0x90, 0xb1, 0x74, 0x39, 0x49, 0x4b, 0x70, 0xc4, 0xbb, 0xc5, 0xef, 0xe0, 0x01, 0xb7, 0x24, + 0xf3, 0x32, 0xd7, 0x94, 0xb2, 0xd0, 0xd4, 0x1a, 0xe3, 0xd1, 0x39, 0xcd, 0x1c, 0x5b, 0x61, 0x87, + 0xc9, 0xc0, 0xf0, 0x00, 0x24, 0x4c, 0x14, 0xaf, 0xd6, 0x61, 0x6f, 0x84, 0x36, 0x8b, 0x53, 0x4c, + 0x6d, 0xc2, 0xea, 0x04, 0x75, 0x98, 0x8b, 0x04, 0x3d, 0x2c, 0xa4, 0xbf, 0x9b, 0xa0, 0x0f, 0x93, + 0x16, 0x9b, 0xf8, 0x0e, 0x56, 0x44, 0x3a, 0x4e, 0x5c, 0xff, 0xdc, 0xb0, 0x7c, 0x8f, 0x06, 0xbe, + 0x6b, 0x9c, 0x3b, 0x9e, 0xed, 0x9f, 0xab, 0xa5, 0xb0, 0x9f, 0x2f, 0x90, 0xb7, 0x3d, 0xfa, 0xe2, + 0xf9, 0x77, 0xa6, 0x3b, 0xc2, 0xfa, 0x2d, 0xae, 0xbd, 0xe7, 0xfa, 0xe7, 0x2d, 0xa1, 0xfb, 0x8e, + 0xab, 0xa2, 0xf7, 0x50, 0x97, 0xc1, 0x9f, 0x45, 0x5c, 0x5e, 0x4c, 0x7c, 0x5b, 0xa8, 0x4f, 0x33, + 0xbf, 0x80, 0xbc, 0x2f, 0x4e, 0x84, 0x15, 0x3e, 0xc2, 0xef, 0xcf, 0x1b, 0x1f, 0x47, 0x1c, 0xa5, + 0x4b, 0xb4, 0xf6, 0xcf, 0x0c, 0x14, 0x64, 0xcb, 0xa3, 0x37, 0x50, 0xa1, 0xd6, 0xd0, 0x19, 0x1a, + 0xa6, 0x10, 0xc8, 0xc9, 0xf5, 0x70, 0xee, 0x90, 0x68, 0xf4, 0xac, 0x61, 0x7b, 0x28, 0x1f, 0xf6, + 0x97, 0xf4, 0x32, 0x57, 0x0f, 0xe9, 0x5e, 0x43, 0x69, 0x64, 0x93, 0x88, 0x4c, 0x8c, 0xb5, 0x07, + 0xf3, 0xc9, 0x8e, 0x6d, 0x12, 0x53, 0xc1, 0x28, 0x7a, 0x62, 0x7e, 0xf9, 0xf4, 0x14, 0x07, 0x11, + 0x55, 0x66, 0x91, 0x5f, 0x47, 0x0c, 0x9e, 0xf0, 0xcb, 0x4f, 0x3c, 0xd7, 0x9b, 0x50, 0x4e, 0xfa, + 0xcd, 0x4e, 0x3e, 0x17, 0xd6, 0x5c, 0xd6, 0x8b, 0xf1, 0x32, 0x10, 0x64, 0x87, 0x7e, 0x20, 0x3e, + 0x4f, 0x72, 0x3a, 0xff, 0x5d, 0xdf, 0x00, 0x88, 0xbd, 0x45, 0x75, 0x50, 0x4e, 0x1c, 0x17, 0xf3, + 0x39, 0x27, 0xce, 0xe3, 0xd1, 0x73, 0xbd, 0x03, 0xe5, 0xa4, 0x33, 0xd1, 0xa9, 0x20, 0x15, 0x9f, + 0x0a, 0xd0, 0x63, 0xc8, 0x8d, 0x59, 0x76, 0x65, 0x88, 0x6e, 0x4c, 0x15, 0x40, 0xd3, 0xfb, 0xa8, + 0x0b, 0xc8, 0x4e, 0x11, 0x0a, 0xd2, 0x53, 0xed, 0x4f, 0x19, 0x76, 0xb2, 0x95, 0xe3, 0x76, 0x1b, + 0x32, 0xd4, 0x25, 0xf3, 0xb7, 0xdd, 0x10, 0xd8, 0xe8, 0xb9, 0x2c, 0x22, 0x0c, 0xcc, 0x3e, 0xde, + 0x78, 0x60, 0xa4, 0xdd, 0x8d, 0x4b, 0xb4, 0xf8, 0x1a, 0xc2, 0xa7, 0xfd, 0x25, 0x5d, 0x28, 0xd6, + 0xff, 0x95, 0x82, 0x4c, 0xcf, 0x25, 0xe8, 0x2b, 0xa8, 0x10, 0x6a, 0x7a, 0xb6, 0x19, 0xd8, 0x46, + 0xbc, 0x3c, 0x16, 0xf9, 0x50, 0xcc, 0x46, 0x3e, 0x5a, 0x05, 0x10, 0x89, 0x8c, 0x8f, 0x92, 0xfb, + 0x4b, 0x7a, 0x91, 0xcb, 0x38, 0xe0, 0x09, 0x2c, 0x8b, 0xbe, 0xb3, 0x70, 0x40, 0x9d, 0x13, 0xc7, + 0x62, 0x9f, 0x96, 0x19, 0x9e, 0x91, 0x1a, 0x7f, 0xd1, 0x8a, 0xe5, 0xe8, 0x29, 0x20, 0xd9, 0x4c, + 0x49, 0x74, 0x96, 0xa3, 0x97, 0xc5, 0x9b, 0x04, 0x7c, 0xa7, 0x0a, 0x65, 0xcb, 0x19, 0x32, 0xeb, + 0x64, 0xe4, 0x50, 0x5c, 0x3f, 0x82, 0xca, 0xc4, 0xaa, 0xbe, 0x38, 0x35, 0x05, 0xc8, 0x0d, 0x7c, + 0x1b, 0xbb, 0x9a, 0x07, 0xe5, 0x64, 0xaf, 0xcd, 0x24, 0xbe, 0x91, 0x24, 0x2e, 0x4a, 0x0a, 0xf4, + 0x1c, 0xc0, 0xb4, 0x6d, 0x87, 0x69, 0x45, 0xbb, 0xfa, 0x6c, 0x9b, 0x09, 0x9c, 0x76, 0x08, 0xd7, + 0x93, 0xf6, 0xd8, 0x18, 0xf3, 0x47, 0x14, 0xfd, 0x0c, 0x94, 0xf0, 0xb6, 0x4c, 0xd6, 0xc5, 0xca, + 0x14, 0xd5, 0x2b, 0x09, 0xd0, 0x23, 0xa8, 0x66, 0x01, 0x4a, 0xb2, 0x1d, 0x3a, 0x5e, 0x1f, 0x07, + 0xec, 0x33, 0xdd, 0x64, 0x9f, 0xef, 0x62, 0x15, 0x8a, 0x2e, 0x9f, 0x26, 0x8c, 0xa4, 0xaf, 0x6e, + 0xe4, 0xef, 0xca, 0x05, 0x9f, 0xad, 0x61, 0xdb, 0x3b, 0xf1, 0x59, 0x2f, 0xb2, 0x19, 0x62, 0xc4, + 0x97, 0x0a, 0x15, 0xbd, 0xc8, 0x24, 0xe2, 0x56, 0x43, 0x13, 0x13, 0xca, 0xb0, 0x4c, 0x89, 0x48, + 0x73, 0x44, 0x89, 0x09, 0x5b, 0xa6, 0xc0, 0x7c, 0x0d, 0x35, 0x8e, 0x09, 0x30, 0x0d, 0x4c, 0x8f, + 0x0c, 0x1c, 0x2a, 0x06, 0x46, 0x45, 0xbf, 0xc6, 0xe4, 0x7a, 0x2c, 0x66, 0x67, 0x14, 0x0e, 0x1d, + 0x06, 0xfe, 0x07, 0x4c, 0x78, 0xe9, 0x54, 0x74, 0xee, 0xc0, 0x5b, 0x2e, 0x61, 0x47, 0x49, 0x0e, + 0xf8, 0x60, 0x5a, 0x67, 0xfe, 0x89, 0xf8, 0x06, 0x95, 0xe6, 0x76, 0x84, 0x28, 0x82, 0x88, 0x79, + 0x4a, 0xf8, 0x26, 0x2f, 0x21, 0x62, 0x69, 0x04, 0x3d, 0x84, 0x6b, 0x62, 0x51, 0x9e, 0x6d, 0x9c, + 0x13, 0xcb, 0x74, 0x31, 0xdf, 0xcd, 0x2b, 0x3a, 0x5f, 0x4c, 0xd7, 0xb3, 0xdf, 0x71, 0x61, 0x84, + 0x0b, 0xac, 0x71, 0x88, 0x53, 0x62, 0x9c, 0x6e, 0x8d, 0x25, 0x6e, 0x05, 0x14, 0x81, 0xa3, 0x3e, + 0xdf, 0x48, 0x2b, 0x7a, 0x81, 0x03, 0xa8, 0x1f, 0xbd, 0x32, 0xa9, 0xcf, 0x37, 0x41, 0xf9, 0xaa, + 0x49, 0x7d, 0xb4, 0x26, 0x1d, 0x65, 0x5e, 0x0c, 0x08, 0xe1, 0xdb, 0x98, 0x5c, 0x6d, 0xd7, 0xb3, + 0xdf, 0x10, 0x12, 0x21, 0x98, 0x7d, 0x86, 0x28, 0xc7, 0x08, 0xdd, 0x1a, 0x33, 0x44, 0xb8, 0xd8, + 0x91, 0x67, 0x5a, 0x67, 0xd8, 0x56, 0x2b, 0xf1, 0x62, 0x8f, 0x85, 0x28, 0x8a, 0x29, 0x11, 0x88, + 0x6a, 0xc2, 0x8a, 0x00, 0xdc, 0x01, 0x9e, 0x50, 0xc3, 0xf5, 0x09, 0x55, 0xaf, 0xf1, 0xd7, 0xdc, + 0xe7, 0x43, 0x9f, 0xd0, 0xc8, 0x80, 0x4c, 0x9e, 0x5a, 0x8b, 0x0d, 0xc8, 0xc4, 0x45, 0x90, 0x13, + 0x46, 0x47, 0x89, 0xba, 0x1c, 0x43, 0xf6, 0x84, 0x08, 0x3d, 0x85, 0xeb, 0xc2, 0x04, 0x3b, 0x26, + 0xb0, 0x93, 0xb2, 0x38, 0x7f, 0x21, 0x8e, 0xe4, 0xd5, 0x71, 0x68, 0x12, 0x7e, 0xec, 0x94, 0x07, + 0x3b, 0x14, 0xc3, 0x4d, 0xeb, 0x4c, 0xa0, 0xaf, 0xc7, 0x35, 0xc3, 0xd0, 0x4d, 0xeb, 0x8c, 0x83, + 0xa7, 0xb9, 0x03, 0x6c, 0x8d, 0xd5, 0x1b, 0xd3, 0xdc, 0x3a, 0xb6, 0xc6, 0xd3, 0xdc, 0x1c, 0x7d, + 0x73, 0x8a, 0x9b, 0x83, 0xc3, 0xd0, 0x0c, 0x07, 0x74, 0xa4, 0xde, 0x8a, 0x43, 0xf3, 0x76, 0x40, + 0x47, 0xe8, 0x31, 0x2c, 0x47, 0xd9, 0x21, 0x84, 0x9e, 0x06, 0x98, 0x9c, 0xaa, 0xb7, 0x13, 0x85, + 0x6d, 0x8d, 0xbb, 0x52, 0x9c, 0xa8, 0x10, 0xaa, 0xaa, 0xc9, 0x0a, 0xa1, 0x51, 0x7e, 0x02, 0x4a, + 0xc7, 0x66, 0xa0, 0xae, 0x24, 0x72, 0xcc, 0x25, 0x91, 0x1d, 0x56, 0x27, 0x91, 0x9d, 0x7a, 0x6c, + 0xa7, 0xeb, 0xd9, 0x91, 0x9d, 0xb0, 0x1f, 0x19, 0xd6, 0x3a, 0xf7, 0x6c, 0xf5, 0x4e, 0x9c, 0x8c, + 0xae, 0x67, 0xb7, 0xce, 0xbd, 0xb8, 0x20, 0x4c, 0x7b, 0xcc, 0x8a, 0xea, 0x6e, 0x6c, 0xb0, 0xc9, + 0x25, 0xec, 0xe4, 0x2f, 0x73, 0xee, 0x07, 0x36, 0x0e, 0x1c, 0xaf, 0xaf, 0xde, 0xe3, 0xa0, 0xaa, + 0x48, 0x7b, 0x28, 0xd5, 0x9a, 0x70, 0xf3, 0x35, 0xa6, 0x3d, 0x7f, 0x28, 0xbf, 0x21, 0x89, 0x8e, + 0x7f, 0x3f, 0xc2, 0x84, 0xb2, 0xc3, 0x36, 0xff, 0x66, 0x30, 0xa6, 0x6e, 0x30, 0xaa, 0x5c, 0xde, + 0x0a, 0x2f, 0x16, 0x34, 0x03, 0x6e, 0x5d, 0xa4, 0x20, 0x43, 0xdf, 0x23, 0x18, 0x7d, 0x03, 0x05, + 0xa9, 0xad, 0xa6, 0xf8, 0x09, 0x6a, 0x65, 0xfe, 0x6d, 0x55, 0x88, 0x44, 0x35, 0xc8, 0x60, 0x4f, + 0x7c, 0x7d, 0x28, 0x3a, 0xfb, 0xa9, 0xbd, 0x84, 0xe5, 0xd7, 0x98, 0x8a, 0x2f, 0xe2, 0xc8, 0xbf, + 0x87, 0xec, 0xdb, 0x86, 0xf9, 0x17, 0xdf, 0x16, 0xa4, 0xc2, 0xcf, 0x10, 0x33, 0x90, 0xe8, 0xb6, + 0xad, 0xbd, 0x07, 0x94, 0x54, 0x96, 0x9e, 0x3d, 0x83, 0xbc, 0xd0, 0x93, 0x8e, 0xa9, 0x73, 0x2f, + 0x02, 0x24, 0x6e, 0x86, 0x5b, 0x3f, 0xc0, 0xed, 0x88, 0x59, 0xcc, 0xdd, 0xc8, 0xb9, 0x89, 0x4b, + 0x8c, 0xd4, 0x85, 0x4b, 0x8c, 0xd8, 0xf3, 0xe8, 0x26, 0x25, 0x9d, 0xf4, 0x5c, 0x5e, 0xa7, 0x68, + 0xa7, 0xa0, 0x4e, 0xf3, 0x4b, 0xff, 0x27, 0x2f, 0xcd, 0x53, 0x9f, 0x73, 0x69, 0x3e, 0x63, 0x25, + 0xdb, 0x3c, 0xc0, 0xd1, 0xc5, 0x95, 0x58, 0xc3, 0xe5, 0x97, 0x57, 0x5a, 0x9b, 0xc7, 0x35, 0xd2, + 0x99, 0x95, 0xf1, 0xd4, 0xd5, 0x32, 0xae, 0xbd, 0x84, 0x1b, 0x6c, 0xa1, 0x89, 0x2b, 0x2d, 0xe1, + 0xc1, 0xd4, 0xb5, 0x56, 0x6a, 0xfa, 0x5a, 0x4b, 0x3b, 0xe6, 0x05, 0x9c, 0x54, 0x96, 0xae, 0xfc, + 0x12, 0x20, 0x06, 0xce, 0xff, 0x07, 0x54, 0x42, 0x33, 0x81, 0xd7, 0x36, 0xa1, 0xc6, 0x68, 0x65, + 0xd0, 0xe2, 0xac, 0x46, 0x29, 0x4b, 0x4d, 0x5e, 0x7e, 0x69, 0xbb, 0xa2, 0x48, 0xa5, 0x42, 0xa2, + 0xcc, 0xc4, 0x25, 0x90, 0xb0, 0xaf, 0xce, 0x4d, 0x91, 0xc4, 0x6d, 0xff, 0x27, 0x0b, 0x8a, 0x0c, + 0xd0, 0x27, 0x64, 0x41, 0x75, 0xb2, 0xb3, 0xd0, 0xa3, 0x69, 0x82, 0x99, 0xed, 0x5b, 0xdf, 0x58, + 0x0c, 0x94, 0x3e, 0xbe, 0x03, 0x88, 0x1b, 0x04, 0xad, 0xcf, 0xd4, 0x9b, 0xec, 0xbd, 0xfa, 0x83, + 0xcb, 0x41, 0x92, 0xd8, 0x11, 0x21, 0x4c, 0xd6, 0x2f, 0xfa, 0xfa, 0x12, 0xcd, 0xc9, 0x1e, 0xaa, + 0x3f, 0xbe, 0x0a, 0x74, 0x62, 0x0d, 0xe1, 0xff, 0x1c, 0x67, 0xaf, 0x61, 0xb2, 0xbc, 0xe7, 0xac, + 0xe1, 0x62, 0x3d, 0xff, 0x08, 0x95, 0x89, 0xea, 0x42, 0x0f, 0x67, 0x7b, 0x75, 0xb1, 0x76, 0xeb, + 0x8f, 0x16, 0xe2, 0xa4, 0x85, 0x1e, 0x14, 0xa3, 0xba, 0x41, 0xda, 0x6c, 0xad, 0x64, 0x15, 0xd6, + 0xd7, 0x2f, 0xc5, 0x08, 0xd6, 0x9d, 0xf7, 0x70, 0xdd, 0xf1, 0xa7, 0x80, 0x3b, 0x95, 0xb0, 0xb4, + 0xde, 0xb2, 0x73, 0xe4, 0xdb, 0xd4, 0x6f, 0x9f, 0xc9, 0x73, 0x65, 0xdf, 0x77, 0x4d, 0xaf, 0xdf, + 0xf0, 0x83, 0xfe, 0xe6, 0xe4, 0xff, 0x9a, 0xd9, 0x53, 0xb8, 0x05, 0x7c, 0x32, 0xc6, 0x5b, 0x1f, + 0xf2, 0xfc, 0x08, 0xfa, 0xcd, 0xff, 0x02, 0x00, 0x00, 0xff, 0xff, 0xd3, 0x7a, 0xae, 0xc5, 0x94, + 0x1e, 0x00, 0x00, } diff --git a/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.proto b/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.proto deleted file mode 100644 index de6c6ef2a..000000000 --- a/vendor/google.golang.org/grpc/channelz/grpc_channelz_v1/channelz.proto +++ /dev/null @@ -1,485 +0,0 @@ -// Copyright 2018 gRPC authors. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// This file defines an interface for exporting monitoring information -// out of gRPC servers. See the full design at -// https://github.com/grpc/proposal/blob/master/A14-channelz.md -// -// The canonical version of this proto can be found at -// https://github.com/grpc/grpc-proto/blob/master/grpc/channelz/v1/channelz.proto - -syntax = "proto3"; - -package grpc.channelz.v1; - -import "google/protobuf/any.proto"; -import "google/protobuf/duration.proto"; -import "google/protobuf/timestamp.proto"; -import "google/protobuf/wrappers.proto"; - -option go_package = "google.golang.org/grpc/channelz/grpc_channelz_v1"; -option java_multiple_files = true; -option java_package = "io.grpc.channelz.v1"; -option java_outer_classname = "ChannelzProto"; - -// Channel is a logical grouping of channels, subchannels, and sockets. -message Channel { - // The identifier for this channel. This should bet set. - ChannelRef ref = 1; - // Data specific to this channel. - ChannelData data = 2; - // At most one of 'channel_ref+subchannel_ref' and 'socket' is set. - - // There are no ordering guarantees on the order of channel refs. - // There may not be cycles in the ref graph. - // A channel ref may be present in more than one channel or subchannel. - repeated ChannelRef channel_ref = 3; - - // At most one of 'channel_ref+subchannel_ref' and 'socket' is set. - // There are no ordering guarantees on the order of subchannel refs. - // There may not be cycles in the ref graph. - // A sub channel ref may be present in more than one channel or subchannel. - repeated SubchannelRef subchannel_ref = 4; - - // There are no ordering guarantees on the order of sockets. - repeated SocketRef socket_ref = 5; -} - -// Subchannel is a logical grouping of channels, subchannels, and sockets. -// A subchannel is load balanced over by it's ancestor -message Subchannel { - // The identifier for this channel. - SubchannelRef ref = 1; - // Data specific to this channel. - ChannelData data = 2; - // At most one of 'channel_ref+subchannel_ref' and 'socket' is set. - - // There are no ordering guarantees on the order of channel refs. - // There may not be cycles in the ref graph. - // A channel ref may be present in more than one channel or subchannel. - repeated ChannelRef channel_ref = 3; - - // At most one of 'channel_ref+subchannel_ref' and 'socket' is set. - // There are no ordering guarantees on the order of subchannel refs. - // There may not be cycles in the ref graph. - // A sub channel ref may be present in more than one channel or subchannel. - repeated SubchannelRef subchannel_ref = 4; - - // There are no ordering guarantees on the order of sockets. - repeated SocketRef socket_ref = 5; -} - -// These come from the specified states in this document: -// https://github.com/grpc/grpc/blob/master/doc/connectivity-semantics-and-api.md -message ChannelConnectivityState { - enum State { - UNKNOWN = 0; - IDLE = 1; - CONNECTING = 2; - READY = 3; - TRANSIENT_FAILURE = 4; - SHUTDOWN = 5; - } - State state = 1; -} - -// Channel data is data related to a specific Channel or Subchannel. -message ChannelData { - // The connectivity state of the channel or subchannel. Implementations - // should always set this. - ChannelConnectivityState state = 1; - - // The target this channel originally tried to connect to. May be absent - string target = 2; - - // A trace of recent events on the channel. May be absent. - ChannelTrace trace = 3; - - // The number of calls started on the channel - int64 calls_started = 4; - // The number of calls that have completed with an OK status - int64 calls_succeeded = 5; - // The number of calls that have completed with a non-OK status - int64 calls_failed = 6; - - // The last time a call was started on the channel. - google.protobuf.Timestamp last_call_started_timestamp = 7; -} - -// ChannelTrace represents the recent events that have occurred on the channel. -message ChannelTrace { - // see the proto in the gRFC for channel tracing: - // A3-channel-tracing.md -} - -// ChannelRef is a reference to a Channel. -message ChannelRef { - // The globally unique id for this channel. Must be a positive number. - int64 channel_id = 1; - // An optional name associated with the channel. - string name = 2; - // Intentionally don't use field numbers from other refs. - reserved 3, 4, 5, 6, 7, 8; -} - -// ChannelRef is a reference to a Subchannel. -message SubchannelRef { - // The globally unique id for this subchannel. Must be a positive number. - int64 subchannel_id = 7; - // An optional name associated with the subchannel. - string name = 8; - // Intentionally don't use field numbers from other refs. - reserved 1, 2, 3, 4, 5, 6; -} - -// SocketRef is a reference to a Socket. -message SocketRef { - int64 socket_id = 3; - // An optional name associated with the socket. - string name = 4; - // Intentionally don't use field numbers from other refs. - reserved 1, 2, 5, 6, 7, 8; -} - -// ServerRef is a reference to a Server. -message ServerRef { - // A globally unique identifier for this server. Must be a positive number. - int64 server_id = 5; - // An optional name associated with the server. - string name = 6; - // Intentionally don't use field numbers from other refs. - reserved 1, 2, 3, 4, 7, 8; -} - -// Server represents a single server. There may be multiple servers in a single -// program. -message Server { - // The identifier for a Server. This should be set. - ServerRef ref = 1; - // The associated data of the Server. - ServerData data = 2; - - // The sockets that the server is listening on. There are no ordering - // guarantees. This may be absent. - repeated SocketRef listen_socket = 3; -} - -// ServerData is data for a specific Server. -message ServerData { - // A trace of recent events on the server. May be absent. - ChannelTrace trace = 1; - - // The number of incoming calls started on the server - int64 calls_started = 2; - // The number of incoming calls that have completed with an OK status - int64 calls_succeeded = 3; - // The number of incoming calls that have a completed with a non-OK status - int64 calls_failed = 4; - - // The last time a call was started on the server. - google.protobuf.Timestamp last_call_started_timestamp = 5; -} - -// Information about an actual connection. Pronounced "sock-ay". -message Socket { - // The identifier for the Socket. - SocketRef ref = 1; - - // Data specific to this Socket. - SocketData data = 2; - // The locally bound address. - Address local = 3; - // The remote bound address. May be absent. - Address remote = 4; - // Security details for this socket. May be absent if not available, or - // there is no security on the socket. - Security security = 5; - - // Optional, represents the name of the remote endpoint, if different than - // the original target name. - string remote_name = 6; -} - -// SocketData is data associated for a specific Socket. The fields present -// are specific to the implementation, so there may be minor differences in -// the semantics. (e.g. flow control windows) -message SocketData { - // The number of streams that have been started. - int64 streams_started = 1; - // The number of streams that have ended successfully: - // On client side, received frame with eos bit set; - // On server side, sent frame with eos bit set. - int64 streams_succeeded = 2; - // The number of streams that have ended unsuccessfully: - // On client side, ended without receiving frame with eos bit set; - // On server side, ended without sending frame with eos bit set. - int64 streams_failed = 3; - // The number of grpc messages successfully sent on this socket. - int64 messages_sent = 4; - // The number of grpc messages received on this socket. - int64 messages_received = 5; - - // The number of keep alives sent. This is typically implemented with HTTP/2 - // ping messages. - int64 keep_alives_sent = 6; - - // The last time a stream was created by this endpoint. Usually unset for - // servers. - google.protobuf.Timestamp last_local_stream_created_timestamp = 7; - // The last time a stream was created by the remote endpoint. Usually unset - // for clients. - google.protobuf.Timestamp last_remote_stream_created_timestamp = 8; - - // The last time a message was sent by this endpoint. - google.protobuf.Timestamp last_message_sent_timestamp = 9; - // The last time a message was received by this endpoint. - google.protobuf.Timestamp last_message_received_timestamp = 10; - - // The amount of window, granted to the local endpoint by the remote endpoint. - // This may be slightly out of date due to network latency. This does NOT - // include stream level or TCP level flow control info. - google.protobuf.Int64Value local_flow_control_window = 11; - - // The amount of window, granted to the remote endpoint by the local endpoint. - // This may be slightly out of date due to network latency. This does NOT - // include stream level or TCP level flow control info. - google.protobuf.Int64Value remote_flow_control_window = 12; - - // Socket options set on this socket. May be absent. - repeated SocketOption option = 13; -} - -// Address represents the address used to create the socket. -message Address { - message TcpIpAddress { - // Either the IPv4 or IPv6 address in bytes. Will be either 4 bytes or 16 - // bytes in length. - bytes ip_address = 1; - // 0-64k, or -1 if not appropriate. - int32 port = 2; - } - // A Unix Domain Socket address. - message UdsAddress { - string filename = 1; - } - // An address type not included above. - message OtherAddress { - // The human readable version of the value. This value should be set. - string name = 1; - // The actual address message. - google.protobuf.Any value = 2; - } - - oneof address { - TcpIpAddress tcpip_address = 1; - UdsAddress uds_address = 2; - OtherAddress other_address = 3; - } -} - -// Security represents details about how secure the socket is. -message Security { - message Tls { - oneof cipher_suite { - // The cipher suite name in the RFC 4346 format: - // https://tools.ietf.org/html/rfc4346#appendix-C - string standard_name = 1; - // Some other way to describe the cipher suite if - // the RFC 4346 name is not available. - string other_name = 2; - } - // the certificate used by this endpoint. - bytes local_certificate = 3; - // the certificate used by the remote endpoint. - bytes remote_certificate = 4; - } - message OtherSecurity { - // The human readable version of the value. - string name = 1; - // The actual security details message. - google.protobuf.Any value = 2; - } - oneof model { - Tls tls = 1; - OtherSecurity other = 2; - } -} - -// SocketOption represents socket options for a socket. Specifically, these -// are the options returned by getsockopt(). -message SocketOption { - // The full name of the socket option. Typically this will be the upper case - // name, such as "SO_REUSEPORT". - string name = 1; - // The human readable value of this socket option. At least one of value or - // additional will be set. - string value = 2; - // Additional data associated with the socket option. At least one of value - // or additional will be set. - google.protobuf.Any additional = 3; -} - -// For use with SocketOption's additional field. This is primarily used for -// SO_RCVTIMEO and SO_SNDTIMEO -message SocketOptionTimeout { - google.protobuf.Duration duration = 1; -} - -// For use with SocketOption's additional field. This is primarily used for -// SO_LINGER. -message SocketOptionLinger { - // active maps to `struct linger.l_onoff` - bool active = 1; - // duration maps to `struct linger.l_linger` - google.protobuf.Duration duration = 2; -} - -// For use with SocketOption's additional field. Tcp info for -// SOL_TCP and TCP_INFO. -message SocketOptionTcpInfo { - uint32 tcpi_state = 1; - - uint32 tcpi_ca_state = 2; - uint32 tcpi_retransmits = 3; - uint32 tcpi_probes = 4; - uint32 tcpi_backoff = 5; - uint32 tcpi_options = 6; - uint32 tcpi_snd_wscale = 7; - uint32 tcpi_rcv_wscale = 8; - - uint32 tcpi_rto = 9; - uint32 tcpi_ato = 10; - uint32 tcpi_snd_mss = 11; - uint32 tcpi_rcv_mss = 12; - - uint32 tcpi_unacked = 13; - uint32 tcpi_sacked = 14; - uint32 tcpi_lost = 15; - uint32 tcpi_retrans = 16; - uint32 tcpi_fackets = 17; - - uint32 tcpi_last_data_sent = 18; - uint32 tcpi_last_ack_sent = 19; - uint32 tcpi_last_data_recv = 20; - uint32 tcpi_last_ack_recv = 21; - - uint32 tcpi_pmtu = 22; - uint32 tcpi_rcv_ssthresh = 23; - uint32 tcpi_rtt = 24; - uint32 tcpi_rttvar = 25; - uint32 tcpi_snd_ssthresh = 26; - uint32 tcpi_snd_cwnd = 27; - uint32 tcpi_advmss = 28; - uint32 tcpi_reordering = 29; -} - -// Channelz is a service exposed by gRPC servers that provides detailed debug -// information. -service Channelz { - // Gets all root channels (i.e. channels the application has directly - // created). This does not include subchannels nor non-top level channels. - rpc GetTopChannels(GetTopChannelsRequest) returns (GetTopChannelsResponse); - // Gets all servers that exist in the process. - rpc GetServers(GetServersRequest) returns (GetServersResponse); - // Gets all server sockets that exist in the process. - rpc GetServerSockets(GetServerSocketsRequest) returns (GetServerSocketsResponse); - // Returns a single Channel, or else a NOT_FOUND code. - rpc GetChannel(GetChannelRequest) returns (GetChannelResponse); - // Returns a single Subchannel, or else a NOT_FOUND code. - rpc GetSubchannel(GetSubchannelRequest) returns (GetSubchannelResponse); - // Returns a single Socket or else a NOT_FOUND code. - rpc GetSocket(GetSocketRequest) returns (GetSocketResponse); -} - -message GetTopChannelsRequest { - // start_channel_id indicates that only channels at or above this id should be - // included in the results. - int64 start_channel_id = 1; -} - -message GetTopChannelsResponse { - // list of channels that the connection detail service knows about. Sorted in - // ascending channel_id order. - repeated Channel channel = 1; - // If set, indicates that the list of channels is the final list. Requesting - // more channels can only return more if they are created after this RPC - // completes. - bool end = 2; -} - -message GetServersRequest { - // start_server_id indicates that only servers at or above this id should be - // included in the results. - int64 start_server_id = 1; -} - -message GetServersResponse { - // list of servers that the connection detail service knows about. Sorted in - // ascending server_id order. - repeated Server server = 1; - // If set, indicates that the list of servers is the final list. Requesting - // more servers will only return more if they are created after this RPC - // completes. - bool end = 2; -} - -message GetServerSocketsRequest { - int64 server_id = 1; - // start_socket_id indicates that only sockets at or above this id should be - // included in the results. - int64 start_socket_id = 2; -} - -message GetServerSocketsResponse { - // list of socket refs that the connection detail service knows about. Sorted in - // ascending socket_id order. - repeated SocketRef socket_ref = 1; - // If set, indicates that the list of sockets is the final list. Requesting - // more sockets will only return more if they are created after this RPC - // completes. - bool end = 2; -} - -message GetChannelRequest { - // channel_id is the identifier of the specific channel to get. - int64 channel_id = 1; -} - -message GetChannelResponse { - // The Channel that corresponds to the requested channel_id. This field - // should be set. - Channel channel = 1; -} - -message GetSubchannelRequest { - // subchannel_id is the identifier of the specific subchannel to get. - int64 subchannel_id = 1; -} - -message GetSubchannelResponse { - // The Subchannel that corresponds to the requested subchannel_id. This - // field should be set. - Subchannel subchannel = 1; -} - -message GetSocketRequest { - // socket_id is the identifier of the specific socket to get. - int64 socket_id = 1; -} - -message GetSocketResponse { - // The Socket that corresponds to the requested socket_id. This field - // should be set. - Socket socket = 1; -} diff --git a/vendor/google.golang.org/grpc/channelz/service/regenerate.sh b/vendor/google.golang.org/grpc/channelz/service/regenerate.sh new file mode 100755 index 000000000..9c255b12c --- /dev/null +++ b/vendor/google.golang.org/grpc/channelz/service/regenerate.sh @@ -0,0 +1,33 @@ +#!/bin/bash +# Copyright 2018 gRPC authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +set -eux -o pipefail + +TMP=$(mktemp -d) + +function finish { + rm -rf "$TMP" +} +trap finish EXIT + +pushd "$TMP" +mkdir -p grpc/channelz/v1 +curl https://raw.githubusercontent.com/grpc/grpc-proto/master/grpc/channelz/v1/channelz.proto > grpc/channelz/v1/channelz.proto + +protoc --go_out=plugins=grpc,paths=source_relative:. -I. grpc/channelz/v1/*.proto +popd +rm -f ../grpc_channelz_v1/*.pb.go +cp "$TMP"/grpc/channelz/v1/*.pb.go ../grpc_channelz_v1/ + diff --git a/vendor/google.golang.org/grpc/channelz/service/service.go b/vendor/google.golang.org/grpc/channelz/service/service.go index 16f69eb89..19c016bc7 100644 --- a/vendor/google.golang.org/grpc/channelz/service/service.go +++ b/vendor/google.golang.org/grpc/channelz/service/service.go @@ -16,7 +16,7 @@ * */ -//go:generate protoc -I ../grpc_channelz_v1 --go_out=plugins=grpc,paths=source_relative:../grpc_channelz_v1 ../grpc_channelz_v1/channelz.proto +//go:generate ./regenerate.sh // Package service provides an implementation for channelz service server. package service @@ -28,44 +28,45 @@ import ( wrpb "github.com/golang/protobuf/ptypes/wrappers" "golang.org/x/net/context" "google.golang.org/grpc" - "google.golang.org/grpc/channelz" - pb "google.golang.org/grpc/channelz/grpc_channelz_v1" + channelzgrpc "google.golang.org/grpc/channelz/grpc_channelz_v1" + channelzpb "google.golang.org/grpc/channelz/grpc_channelz_v1" "google.golang.org/grpc/connectivity" + "google.golang.org/grpc/internal/channelz" ) // RegisterChannelzServiceToServer registers the channelz service to the given server. func RegisterChannelzServiceToServer(s *grpc.Server) { - pb.RegisterChannelzServer(s, &serverImpl{}) + channelzgrpc.RegisterChannelzServer(s, &serverImpl{}) } -func newCZServer() pb.ChannelzServer { +func newCZServer() channelzgrpc.ChannelzServer { return &serverImpl{} } type serverImpl struct{} -func connectivityStateToProto(s connectivity.State) *pb.ChannelConnectivityState { +func connectivityStateToProto(s connectivity.State) *channelzpb.ChannelConnectivityState { switch s { case connectivity.Idle: - return &pb.ChannelConnectivityState{State: pb.ChannelConnectivityState_IDLE} + return &channelzpb.ChannelConnectivityState{State: channelzpb.ChannelConnectivityState_IDLE} case connectivity.Connecting: - return &pb.ChannelConnectivityState{State: pb.ChannelConnectivityState_CONNECTING} + return &channelzpb.ChannelConnectivityState{State: channelzpb.ChannelConnectivityState_CONNECTING} case connectivity.Ready: - return &pb.ChannelConnectivityState{State: pb.ChannelConnectivityState_READY} + return &channelzpb.ChannelConnectivityState{State: channelzpb.ChannelConnectivityState_READY} case connectivity.TransientFailure: - return &pb.ChannelConnectivityState{State: pb.ChannelConnectivityState_TRANSIENT_FAILURE} + return &channelzpb.ChannelConnectivityState{State: channelzpb.ChannelConnectivityState_TRANSIENT_FAILURE} case connectivity.Shutdown: - return &pb.ChannelConnectivityState{State: pb.ChannelConnectivityState_SHUTDOWN} + return &channelzpb.ChannelConnectivityState{State: channelzpb.ChannelConnectivityState_SHUTDOWN} default: - return &pb.ChannelConnectivityState{State: pb.ChannelConnectivityState_UNKNOWN} + return &channelzpb.ChannelConnectivityState{State: channelzpb.ChannelConnectivityState_UNKNOWN} } } -func channelMetricToProto(cm *channelz.ChannelMetric) *pb.Channel { - c := &pb.Channel{} - c.Ref = &pb.ChannelRef{ChannelId: cm.ID, Name: cm.RefName} +func channelMetricToProto(cm *channelz.ChannelMetric) *channelzpb.Channel { + c := &channelzpb.Channel{} + c.Ref = &channelzpb.ChannelRef{ChannelId: cm.ID, Name: cm.RefName} - c.Data = &pb.ChannelData{ + c.Data = &channelzpb.ChannelData{ State: connectivityStateToProto(cm.ChannelData.State), Target: cm.ChannelData.Target, CallsStarted: cm.ChannelData.CallsStarted, @@ -75,31 +76,31 @@ func channelMetricToProto(cm *channelz.ChannelMetric) *pb.Channel { if ts, err := ptypes.TimestampProto(cm.ChannelData.LastCallStartedTimestamp); err == nil { c.Data.LastCallStartedTimestamp = ts } - nestedChans := make([]*pb.ChannelRef, 0, len(cm.NestedChans)) + nestedChans := make([]*channelzpb.ChannelRef, 0, len(cm.NestedChans)) for id, ref := range cm.NestedChans { - nestedChans = append(nestedChans, &pb.ChannelRef{ChannelId: id, Name: ref}) + nestedChans = append(nestedChans, &channelzpb.ChannelRef{ChannelId: id, Name: ref}) } c.ChannelRef = nestedChans - subChans := make([]*pb.SubchannelRef, 0, len(cm.SubChans)) + subChans := make([]*channelzpb.SubchannelRef, 0, len(cm.SubChans)) for id, ref := range cm.SubChans { - subChans = append(subChans, &pb.SubchannelRef{SubchannelId: id, Name: ref}) + subChans = append(subChans, &channelzpb.SubchannelRef{SubchannelId: id, Name: ref}) } c.SubchannelRef = subChans - sockets := make([]*pb.SocketRef, 0, len(cm.Sockets)) + sockets := make([]*channelzpb.SocketRef, 0, len(cm.Sockets)) for id, ref := range cm.Sockets { - sockets = append(sockets, &pb.SocketRef{SocketId: id, Name: ref}) + sockets = append(sockets, &channelzpb.SocketRef{SocketId: id, Name: ref}) } c.SocketRef = sockets return c } -func subChannelMetricToProto(cm *channelz.SubChannelMetric) *pb.Subchannel { - sc := &pb.Subchannel{} - sc.Ref = &pb.SubchannelRef{SubchannelId: cm.ID, Name: cm.RefName} +func subChannelMetricToProto(cm *channelz.SubChannelMetric) *channelzpb.Subchannel { + sc := &channelzpb.Subchannel{} + sc.Ref = &channelzpb.SubchannelRef{SubchannelId: cm.ID, Name: cm.RefName} - sc.Data = &pb.ChannelData{ + sc.Data = &channelzpb.ChannelData{ State: connectivityStateToProto(cm.ChannelData.State), Target: cm.ChannelData.Target, CallsStarted: cm.ChannelData.CallsStarted, @@ -109,51 +110,51 @@ func subChannelMetricToProto(cm *channelz.SubChannelMetric) *pb.Subchannel { if ts, err := ptypes.TimestampProto(cm.ChannelData.LastCallStartedTimestamp); err == nil { sc.Data.LastCallStartedTimestamp = ts } - nestedChans := make([]*pb.ChannelRef, 0, len(cm.NestedChans)) + nestedChans := make([]*channelzpb.ChannelRef, 0, len(cm.NestedChans)) for id, ref := range cm.NestedChans { - nestedChans = append(nestedChans, &pb.ChannelRef{ChannelId: id, Name: ref}) + nestedChans = append(nestedChans, &channelzpb.ChannelRef{ChannelId: id, Name: ref}) } sc.ChannelRef = nestedChans - subChans := make([]*pb.SubchannelRef, 0, len(cm.SubChans)) + subChans := make([]*channelzpb.SubchannelRef, 0, len(cm.SubChans)) for id, ref := range cm.SubChans { - subChans = append(subChans, &pb.SubchannelRef{SubchannelId: id, Name: ref}) + subChans = append(subChans, &channelzpb.SubchannelRef{SubchannelId: id, Name: ref}) } sc.SubchannelRef = subChans - sockets := make([]*pb.SocketRef, 0, len(cm.Sockets)) + sockets := make([]*channelzpb.SocketRef, 0, len(cm.Sockets)) for id, ref := range cm.Sockets { - sockets = append(sockets, &pb.SocketRef{SocketId: id, Name: ref}) + sockets = append(sockets, &channelzpb.SocketRef{SocketId: id, Name: ref}) } sc.SocketRef = sockets return sc } -func addrToProto(a net.Addr) *pb.Address { +func addrToProto(a net.Addr) *channelzpb.Address { switch a.Network() { case "udp": // TODO: Address_OtherAddress{}. Need proto def for Value. case "ip": // Note zone info is discarded through the conversion. - return &pb.Address{Address: &pb.Address_TcpipAddress{TcpipAddress: &pb.Address_TcpIpAddress{IpAddress: a.(*net.IPAddr).IP}}} + return &channelzpb.Address{Address: &channelzpb.Address_TcpipAddress{TcpipAddress: &channelzpb.Address_TcpIpAddress{IpAddress: a.(*net.IPAddr).IP}}} case "ip+net": // Note mask info is discarded through the conversion. - return &pb.Address{Address: &pb.Address_TcpipAddress{TcpipAddress: &pb.Address_TcpIpAddress{IpAddress: a.(*net.IPNet).IP}}} + return &channelzpb.Address{Address: &channelzpb.Address_TcpipAddress{TcpipAddress: &channelzpb.Address_TcpIpAddress{IpAddress: a.(*net.IPNet).IP}}} case "tcp": // Note zone info is discarded through the conversion. - return &pb.Address{Address: &pb.Address_TcpipAddress{TcpipAddress: &pb.Address_TcpIpAddress{IpAddress: a.(*net.TCPAddr).IP, Port: int32(a.(*net.TCPAddr).Port)}}} + return &channelzpb.Address{Address: &channelzpb.Address_TcpipAddress{TcpipAddress: &channelzpb.Address_TcpIpAddress{IpAddress: a.(*net.TCPAddr).IP, Port: int32(a.(*net.TCPAddr).Port)}}} case "unix", "unixgram", "unixpacket": - return &pb.Address{Address: &pb.Address_UdsAddress_{UdsAddress: &pb.Address_UdsAddress{Filename: a.String()}}} + return &channelzpb.Address{Address: &channelzpb.Address_UdsAddress_{UdsAddress: &channelzpb.Address_UdsAddress{Filename: a.String()}}} default: } - return &pb.Address{} + return &channelzpb.Address{} } -func socketMetricToProto(sm *channelz.SocketMetric) *pb.Socket { - s := &pb.Socket{} - s.Ref = &pb.SocketRef{SocketId: sm.ID, Name: sm.RefName} +func socketMetricToProto(sm *channelz.SocketMetric) *channelzpb.Socket { + s := &channelzpb.Socket{} + s.Ref = &channelzpb.SocketRef{SocketId: sm.ID, Name: sm.RefName} - s.Data = &pb.SocketData{ + s.Data = &channelzpb.SocketData{ StreamsStarted: sm.SocketData.StreamsStarted, StreamsSucceeded: sm.SocketData.StreamsSucceeded, StreamsFailed: sm.SocketData.StreamsFailed, @@ -186,9 +187,9 @@ func socketMetricToProto(sm *channelz.SocketMetric) *pb.Socket { return s } -func (s *serverImpl) GetTopChannels(ctx context.Context, req *pb.GetTopChannelsRequest) (*pb.GetTopChannelsResponse, error) { +func (s *serverImpl) GetTopChannels(ctx context.Context, req *channelzpb.GetTopChannelsRequest) (*channelzpb.GetTopChannelsResponse, error) { metrics, end := channelz.GetTopChannels(req.GetStartChannelId()) - resp := &pb.GetTopChannelsResponse{} + resp := &channelzpb.GetTopChannelsResponse{} for _, m := range metrics { resp.Channel = append(resp.Channel, channelMetricToProto(m)) } @@ -196,11 +197,11 @@ func (s *serverImpl) GetTopChannels(ctx context.Context, req *pb.GetTopChannelsR return resp, nil } -func serverMetricToProto(sm *channelz.ServerMetric) *pb.Server { - s := &pb.Server{} - s.Ref = &pb.ServerRef{ServerId: sm.ID, Name: sm.RefName} +func serverMetricToProto(sm *channelz.ServerMetric) *channelzpb.Server { + s := &channelzpb.Server{} + s.Ref = &channelzpb.ServerRef{ServerId: sm.ID, Name: sm.RefName} - s.Data = &pb.ServerData{ + s.Data = &channelzpb.ServerData{ CallsStarted: sm.ServerData.CallsStarted, CallsSucceeded: sm.ServerData.CallsSucceeded, CallsFailed: sm.ServerData.CallsFailed, @@ -209,17 +210,17 @@ func serverMetricToProto(sm *channelz.ServerMetric) *pb.Server { if ts, err := ptypes.TimestampProto(sm.ServerData.LastCallStartedTimestamp); err == nil { s.Data.LastCallStartedTimestamp = ts } - sockets := make([]*pb.SocketRef, 0, len(sm.ListenSockets)) + sockets := make([]*channelzpb.SocketRef, 0, len(sm.ListenSockets)) for id, ref := range sm.ListenSockets { - sockets = append(sockets, &pb.SocketRef{SocketId: id, Name: ref}) + sockets = append(sockets, &channelzpb.SocketRef{SocketId: id, Name: ref}) } s.ListenSocket = sockets return s } -func (s *serverImpl) GetServers(ctx context.Context, req *pb.GetServersRequest) (*pb.GetServersResponse, error) { +func (s *serverImpl) GetServers(ctx context.Context, req *channelzpb.GetServersRequest) (*channelzpb.GetServersResponse, error) { metrics, end := channelz.GetServers(req.GetStartServerId()) - resp := &pb.GetServersResponse{} + resp := &channelzpb.GetServersResponse{} for _, m := range metrics { resp.Server = append(resp.Server, serverMetricToProto(m)) } @@ -227,39 +228,39 @@ func (s *serverImpl) GetServers(ctx context.Context, req *pb.GetServersRequest) return resp, nil } -func (s *serverImpl) GetServerSockets(ctx context.Context, req *pb.GetServerSocketsRequest) (*pb.GetServerSocketsResponse, error) { +func (s *serverImpl) GetServerSockets(ctx context.Context, req *channelzpb.GetServerSocketsRequest) (*channelzpb.GetServerSocketsResponse, error) { metrics, end := channelz.GetServerSockets(req.GetServerId(), req.GetStartSocketId()) - resp := &pb.GetServerSocketsResponse{} + resp := &channelzpb.GetServerSocketsResponse{} for _, m := range metrics { - resp.SocketRef = append(resp.SocketRef, &pb.SocketRef{SocketId: m.ID, Name: m.RefName}) + resp.SocketRef = append(resp.SocketRef, &channelzpb.SocketRef{SocketId: m.ID, Name: m.RefName}) } resp.End = end return resp, nil } -func (s *serverImpl) GetChannel(ctx context.Context, req *pb.GetChannelRequest) (*pb.GetChannelResponse, error) { +func (s *serverImpl) GetChannel(ctx context.Context, req *channelzpb.GetChannelRequest) (*channelzpb.GetChannelResponse, error) { var metric *channelz.ChannelMetric if metric = channelz.GetChannel(req.GetChannelId()); metric == nil { - return &pb.GetChannelResponse{}, nil + return &channelzpb.GetChannelResponse{}, nil } - resp := &pb.GetChannelResponse{Channel: channelMetricToProto(metric)} + resp := &channelzpb.GetChannelResponse{Channel: channelMetricToProto(metric)} return resp, nil } -func (s *serverImpl) GetSubchannel(ctx context.Context, req *pb.GetSubchannelRequest) (*pb.GetSubchannelResponse, error) { +func (s *serverImpl) GetSubchannel(ctx context.Context, req *channelzpb.GetSubchannelRequest) (*channelzpb.GetSubchannelResponse, error) { var metric *channelz.SubChannelMetric if metric = channelz.GetSubChannel(req.GetSubchannelId()); metric == nil { - return &pb.GetSubchannelResponse{}, nil + return &channelzpb.GetSubchannelResponse{}, nil } - resp := &pb.GetSubchannelResponse{Subchannel: subChannelMetricToProto(metric)} + resp := &channelzpb.GetSubchannelResponse{Subchannel: subChannelMetricToProto(metric)} return resp, nil } -func (s *serverImpl) GetSocket(ctx context.Context, req *pb.GetSocketRequest) (*pb.GetSocketResponse, error) { +func (s *serverImpl) GetSocket(ctx context.Context, req *channelzpb.GetSocketRequest) (*channelzpb.GetSocketResponse, error) { var metric *channelz.SocketMetric if metric = channelz.GetSocket(req.GetSocketId()); metric == nil { - return &pb.GetSocketResponse{}, nil + return &channelzpb.GetSocketResponse{}, nil } - resp := &pb.GetSocketResponse{Socket: socketMetricToProto(metric)} + resp := &channelzpb.GetSocketResponse{Socket: socketMetricToProto(metric)} return resp, nil } diff --git a/vendor/google.golang.org/grpc/channelz/service/service_test.go b/vendor/google.golang.org/grpc/channelz/service/service_test.go index 9503b7e90..398cd6c97 100644 --- a/vendor/google.golang.org/grpc/channelz/service/service_test.go +++ b/vendor/google.golang.org/grpc/channelz/service/service_test.go @@ -27,9 +27,9 @@ import ( "github.com/golang/protobuf/ptypes" "golang.org/x/net/context" - "google.golang.org/grpc/channelz" - pb "google.golang.org/grpc/channelz/grpc_channelz_v1" + channelzpb "google.golang.org/grpc/channelz/grpc_channelz_v1" "google.golang.org/grpc/connectivity" + "google.golang.org/grpc/internal/channelz" ) func init() { @@ -121,21 +121,21 @@ func (d *dummySocket) ChannelzMetric() *channelz.SocketInternalMetric { } } -func channelProtoToStruct(c *pb.Channel) *dummyChannel { +func channelProtoToStruct(c *channelzpb.Channel) *dummyChannel { dc := &dummyChannel{} pdata := c.GetData() switch pdata.GetState().GetState() { - case pb.ChannelConnectivityState_UNKNOWN: + case channelzpb.ChannelConnectivityState_UNKNOWN: // TODO: what should we set here? - case pb.ChannelConnectivityState_IDLE: + case channelzpb.ChannelConnectivityState_IDLE: dc.state = connectivity.Idle - case pb.ChannelConnectivityState_CONNECTING: + case channelzpb.ChannelConnectivityState_CONNECTING: dc.state = connectivity.Connecting - case pb.ChannelConnectivityState_READY: + case channelzpb.ChannelConnectivityState_READY: dc.state = connectivity.Ready - case pb.ChannelConnectivityState_TRANSIENT_FAILURE: + case channelzpb.ChannelConnectivityState_TRANSIENT_FAILURE: dc.state = connectivity.TransientFailure - case pb.ChannelConnectivityState_SHUTDOWN: + case channelzpb.ChannelConnectivityState_SHUTDOWN: dc.state = connectivity.Shutdown } dc.target = pdata.GetTarget() @@ -150,7 +150,7 @@ func channelProtoToStruct(c *pb.Channel) *dummyChannel { return dc } -func serverProtoToStruct(s *pb.Server) *dummyServer { +func serverProtoToStruct(s *channelzpb.Server) *dummyServer { ds := &dummyServer{} pdata := s.GetData() ds.callsStarted = pdata.CallsStarted @@ -164,22 +164,22 @@ func serverProtoToStruct(s *pb.Server) *dummyServer { return ds } -func protoToAddr(a *pb.Address) net.Addr { +func protoToAddr(a *channelzpb.Address) net.Addr { switch v := a.Address.(type) { - case *pb.Address_TcpipAddress: + case *channelzpb.Address_TcpipAddress: if port := v.TcpipAddress.GetPort(); port != 0 { return &net.TCPAddr{IP: v.TcpipAddress.GetIpAddress(), Port: int(port)} } return &net.IPAddr{IP: v.TcpipAddress.GetIpAddress()} - case *pb.Address_UdsAddress_: + case *channelzpb.Address_UdsAddress_: return &net.UnixAddr{Name: v.UdsAddress.GetFilename(), Net: "unix"} - case *pb.Address_OtherAddress_: + case *channelzpb.Address_OtherAddress_: // TODO: } return nil } -func socketProtoToStruct(s *pb.Socket) *dummySocket { +func socketProtoToStruct(s *channelzpb.Socket) *dummySocket { ds := &dummySocket{} pdata := s.GetData() ds.streamsStarted = pdata.GetStreamsStarted() @@ -224,7 +224,7 @@ func socketProtoToStruct(s *pb.Socket) *dummySocket { return ds } -func convertSocketRefSliceToMap(sktRefs []*pb.SocketRef) map[int64]string { +func convertSocketRefSliceToMap(sktRefs []*channelzpb.SocketRef) map[int64]string { m := make(map[int64]string) for _, sr := range sktRefs { m[sr.SocketId] = sr.Name @@ -264,7 +264,7 @@ func TestGetTopChannels(t *testing.T) { channelz.RegisterChannel(c, 0, "") } s := newCZServer() - resp, _ := s.GetTopChannels(context.Background(), &pb.GetTopChannelsRequest{StartChannelId: 0}) + resp, _ := s.GetTopChannels(context.Background(), &channelzpb.GetTopChannelsRequest{StartChannelId: 0}) if !resp.GetEnd() { t.Fatalf("resp.GetEnd() want true, got %v", resp.GetEnd()) } @@ -276,7 +276,7 @@ func TestGetTopChannels(t *testing.T) { for i := 0; i < 50; i++ { channelz.RegisterChannel(tcs[0], 0, "") } - resp, _ = s.GetTopChannels(context.Background(), &pb.GetTopChannelsRequest{StartChannelId: 0}) + resp, _ = s.GetTopChannels(context.Background(), &channelzpb.GetTopChannelsRequest{StartChannelId: 0}) if resp.GetEnd() { t.Fatalf("resp.GetEnd() want false, got %v", resp.GetEnd()) } @@ -308,7 +308,7 @@ func TestGetServers(t *testing.T) { channelz.RegisterServer(s, "") } svr := newCZServer() - resp, _ := svr.GetServers(context.Background(), &pb.GetServersRequest{StartServerId: 0}) + resp, _ := svr.GetServers(context.Background(), &channelzpb.GetServersRequest{StartServerId: 0}) if !resp.GetEnd() { t.Fatalf("resp.GetEnd() want true, got %v", resp.GetEnd()) } @@ -320,7 +320,7 @@ func TestGetServers(t *testing.T) { for i := 0; i < 50; i++ { channelz.RegisterServer(ss[0], "") } - resp, _ = svr.GetServers(context.Background(), &pb.GetServersRequest{StartServerId: 0}) + resp, _ = svr.GetServers(context.Background(), &channelzpb.GetServersRequest{StartServerId: 0}) if resp.GetEnd() { t.Fatalf("resp.GetEnd() want false, got %v", resp.GetEnd()) } @@ -335,7 +335,7 @@ func TestGetServerSockets(t *testing.T) { ids[1] = channelz.RegisterNormalSocket(&dummySocket{}, svrID, refNames[1]) ids[2] = channelz.RegisterNormalSocket(&dummySocket{}, svrID, refNames[2]) svr := newCZServer() - resp, _ := svr.GetServerSockets(context.Background(), &pb.GetServerSocketsRequest{ServerId: svrID, StartSocketId: 0}) + resp, _ := svr.GetServerSockets(context.Background(), &channelzpb.GetServerSocketsRequest{ServerId: svrID, StartSocketId: 0}) if !resp.GetEnd() { t.Fatalf("resp.GetEnd() want: true, got: %v", resp.GetEnd()) } @@ -351,7 +351,7 @@ func TestGetServerSockets(t *testing.T) { for i := 0; i < 50; i++ { channelz.RegisterNormalSocket(&dummySocket{}, svrID, "") } - resp, _ = svr.GetServerSockets(context.Background(), &pb.GetServerSocketsRequest{ServerId: svrID, StartSocketId: 0}) + resp, _ = svr.GetServerSockets(context.Background(), &channelzpb.GetServerSocketsRequest{ServerId: svrID, StartSocketId: 0}) if resp.GetEnd() { t.Fatalf("resp.GetEnd() want false, got %v", resp.GetEnd()) } @@ -366,22 +366,22 @@ func TestGetChannel(t *testing.T) { ids[2] = channelz.RegisterSubChannel(&dummyChannel{}, ids[0], refNames[2]) ids[3] = channelz.RegisterChannel(&dummyChannel{}, ids[1], refNames[3]) svr := newCZServer() - resp, _ := svr.GetChannel(context.Background(), &pb.GetChannelRequest{ChannelId: ids[0]}) + resp, _ := svr.GetChannel(context.Background(), &channelzpb.GetChannelRequest{ChannelId: ids[0]}) metrics := resp.GetChannel() subChans := metrics.GetSubchannelRef() if len(subChans) != 1 || subChans[0].GetName() != refNames[2] || subChans[0].GetSubchannelId() != ids[2] { - t.Fatalf("GetSubChannelRef() want %#v, got %#v", []*pb.SubchannelRef{{SubchannelId: ids[2], Name: refNames[2]}}, subChans) + t.Fatalf("GetSubChannelRef() want %#v, got %#v", []*channelzpb.SubchannelRef{{SubchannelId: ids[2], Name: refNames[2]}}, subChans) } nestedChans := metrics.GetChannelRef() if len(nestedChans) != 1 || nestedChans[0].GetName() != refNames[1] || nestedChans[0].GetChannelId() != ids[1] { - t.Fatalf("GetChannelRef() want %#v, got %#v", []*pb.ChannelRef{{ChannelId: ids[1], Name: refNames[1]}}, nestedChans) + t.Fatalf("GetChannelRef() want %#v, got %#v", []*channelzpb.ChannelRef{{ChannelId: ids[1], Name: refNames[1]}}, nestedChans) } - resp, _ = svr.GetChannel(context.Background(), &pb.GetChannelRequest{ChannelId: ids[1]}) + resp, _ = svr.GetChannel(context.Background(), &channelzpb.GetChannelRequest{ChannelId: ids[1]}) metrics = resp.GetChannel() nestedChans = metrics.GetChannelRef() if len(nestedChans) != 1 || nestedChans[0].GetName() != refNames[3] || nestedChans[0].GetChannelId() != ids[3] { - t.Fatalf("GetChannelRef() want %#v, got %#v", []*pb.ChannelRef{{ChannelId: ids[3], Name: refNames[3]}}, nestedChans) + t.Fatalf("GetChannelRef() want %#v, got %#v", []*channelzpb.ChannelRef{{ChannelId: ids[3], Name: refNames[3]}}, nestedChans) } } @@ -394,7 +394,7 @@ func TestGetSubChannel(t *testing.T) { ids[2] = channelz.RegisterNormalSocket(&dummySocket{}, ids[1], refNames[2]) ids[3] = channelz.RegisterNormalSocket(&dummySocket{}, ids[1], refNames[3]) svr := newCZServer() - resp, _ := svr.GetSubchannel(context.Background(), &pb.GetSubchannelRequest{SubchannelId: ids[1]}) + resp, _ := svr.GetSubchannel(context.Background(), &channelzpb.GetSubchannelRequest{SubchannelId: ids[1]}) metrics := resp.GetSubchannel() want := map[int64]string{ ids[2]: refNames[2], @@ -468,10 +468,10 @@ func TestGetSocket(t *testing.T) { ids[i] = channelz.RegisterNormalSocket(s, svrID, strconv.Itoa(i)) } for i, s := range ss { - resp, _ := svr.GetSocket(context.Background(), &pb.GetSocketRequest{SocketId: ids[i]}) + resp, _ := svr.GetSocket(context.Background(), &channelzpb.GetSocketRequest{SocketId: ids[i]}) metrics := resp.GetSocket() - if !reflect.DeepEqual(metrics.GetRef(), &pb.SocketRef{SocketId: ids[i], Name: strconv.Itoa(i)}) || !reflect.DeepEqual(socketProtoToStruct(metrics), s) { - t.Fatalf("resp.GetSocket() want: metrics.GetRef() = %#v and %#v, got: metrics.GetRef() = %#v and %#v", &pb.SocketRef{SocketId: ids[i], Name: strconv.Itoa(i)}, s, metrics.GetRef(), socketProtoToStruct(metrics)) + if !reflect.DeepEqual(metrics.GetRef(), &channelzpb.SocketRef{SocketId: ids[i], Name: strconv.Itoa(i)}) || !reflect.DeepEqual(socketProtoToStruct(metrics), s) { + t.Fatalf("resp.GetSocket() want: metrics.GetRef() = %#v and %#v, got: metrics.GetRef() = %#v and %#v", &channelzpb.SocketRef{SocketId: ids[i], Name: strconv.Itoa(i)}, s, metrics.GetRef(), socketProtoToStruct(metrics)) } } } diff --git a/vendor/google.golang.org/grpc/clientconn.go b/vendor/google.golang.org/grpc/clientconn.go index e8d95b43b..84ba9e5ad 100644 --- a/vendor/google.golang.org/grpc/clientconn.go +++ b/vendor/google.golang.org/grpc/clientconn.go @@ -32,11 +32,13 @@ import ( "golang.org/x/net/trace" "google.golang.org/grpc/balancer" _ "google.golang.org/grpc/balancer/roundrobin" // To register roundrobin. - "google.golang.org/grpc/channelz" "google.golang.org/grpc/codes" "google.golang.org/grpc/connectivity" "google.golang.org/grpc/credentials" "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal" + "google.golang.org/grpc/internal/backoff" + "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/resolver" _ "google.golang.org/grpc/resolver/dns" // To register dns resolver. @@ -49,6 +51,8 @@ import ( const ( // minimum time to give a connection to complete minConnectTimeout = 20 * time.Second + // must match grpclbName in grpclb/grpclb.go + grpclbName = "grpclb" ) var ( @@ -97,7 +101,7 @@ type dialOptions struct { streamInt StreamClientInterceptor cp Compressor dc Decompressor - bs backoffStrategy + bs backoff.Strategy block bool insecure bool timeout time.Duration @@ -275,17 +279,17 @@ func WithBackoffMaxDelay(md time.Duration) DialOption { // Use WithBackoffMaxDelay until more parameters on BackoffConfig are opened up // for use. func WithBackoffConfig(b BackoffConfig) DialOption { - // Set defaults to ensure that provided BackoffConfig is valid and - // unexported fields get default values. - setDefaults(&b) - return withBackoff(b) + + return withBackoff(backoff.Exponential{ + MaxDelay: b.MaxDelay, + }) } // withBackoff sets the backoff strategy used for connectRetryNum after a // failed connection attempt. // // This can be exported if arbitrary backoff strategies are allowed by gRPC. -func withBackoff(bs backoffStrategy) DialOption { +func withBackoff(bs backoff.Strategy) DialOption { return func(o *dialOptions) { o.bs = bs } @@ -340,6 +344,11 @@ func withContextDialer(f func(context.Context, string) (net.Conn, error)) DialOp } } +func init() { + internal.WithContextDialer = withContextDialer + internal.WithResolverBuilder = withResolverBuilder +} + // WithDialer returns a DialOption that specifies a function to use for dialing network addresses. // If FailOnNonTempDialError() is set to true, and an error is returned by f, gRPC checks the error's // Temporary() method to decide if it should try to reconnect to the network address. @@ -532,7 +541,9 @@ func DialContext(ctx context.Context, target string, opts ...DialOption) (conn * } } if cc.dopts.bs == nil { - cc.dopts.bs = DefaultBackoffConfig + cc.dopts.bs = backoff.Exponential{ + MaxDelay: DefaultBackoffConfig.MaxDelay, + } } if cc.dopts.resolverBuilder == nil { // Only try to parse target when resolver builder is not already set. @@ -1052,9 +1063,9 @@ func (cc *ClientConn) handleServiceConfig(js string) error { } func (cc *ClientConn) resolveNow(o resolver.ResolveNowOption) { - cc.mu.Lock() + cc.mu.RLock() r := cc.resolverWrapper - cc.mu.Unlock() + cc.mu.RUnlock() if r == nil { return } @@ -1203,7 +1214,7 @@ func (ac *addrConn) resetTransport() error { // This means either a successful HTTP2 connection was established // or this is the first time this addrConn is trying to establish a // connection. - backoffFor := ac.dopts.bs.backoff(connectRetryNum) // time.Duration. + backoffFor := ac.dopts.bs.Backoff(connectRetryNum) // time.Duration. // This will be the duration that dial gets to finish. dialDuration := getMinConnectTimeout() if backoffFor > dialDuration { diff --git a/vendor/google.golang.org/grpc/clientconn_test.go b/vendor/google.golang.org/grpc/clientconn_test.go index 4c5bf8bc8..fd239d36d 100644 --- a/vendor/google.golang.org/grpc/clientconn_test.go +++ b/vendor/google.golang.org/grpc/clientconn_test.go @@ -27,15 +27,15 @@ import ( "golang.org/x/net/context" "golang.org/x/net/http2" - "google.golang.org/grpc/connectivity" "google.golang.org/grpc/credentials" + "google.golang.org/grpc/internal/backoff" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/naming" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" _ "google.golang.org/grpc/resolver/passthrough" - "google.golang.org/grpc/test/leakcheck" "google.golang.org/grpc/testdata" ) @@ -504,7 +504,6 @@ func TestWithBackoffConfig(t *testing.T) { defer leakcheck.Check(t) b := BackoffConfig{MaxDelay: DefaultBackoffConfig.MaxDelay / 2} expected := b - setDefaults(&expected) // defaults should be set testBackoffConfigSet(t, &expected, WithBackoffConfig(b)) } @@ -512,7 +511,6 @@ func TestWithBackoffMaxDelay(t *testing.T) { defer leakcheck.Check(t) md := DefaultBackoffConfig.MaxDelay / 2 expected := BackoffConfig{MaxDelay: md} - setDefaults(&expected) testBackoffConfigSet(t, &expected, WithBackoffMaxDelay(md)) } @@ -528,12 +526,15 @@ func testBackoffConfigSet(t *testing.T, expected *BackoffConfig, opts ...DialOpt t.Fatalf("backoff config not set") } - actual, ok := conn.dopts.bs.(BackoffConfig) + actual, ok := conn.dopts.bs.(backoff.Exponential) if !ok { t.Fatalf("unexpected type of backoff config: %#v", conn.dopts.bs) } - if actual != *expected { + expectedValue := backoff.Exponential{ + MaxDelay: expected.MaxDelay, + } + if actual != expectedValue { t.Fatalf("unexpected backoff config on connection: %v, want %v", actual, expected) } } diff --git a/vendor/google.golang.org/grpc/codes/codes.go b/vendor/google.golang.org/grpc/codes/codes.go index a8280ae66..d9b9d5782 100644 --- a/vendor/google.golang.org/grpc/codes/codes.go +++ b/vendor/google.golang.org/grpc/codes/codes.go @@ -22,6 +22,7 @@ package codes // import "google.golang.org/grpc/codes" import ( "fmt" + "strconv" ) // A Code is an unsigned 32-bit error code as defined in the gRPC spec. @@ -143,6 +144,8 @@ const ( // Unauthenticated indicates the request does not have valid // authentication credentials for the operation. Unauthenticated Code = 16 + + _maxCode = 17 ) var strToCode = map[string]Code{ @@ -176,6 +179,16 @@ func (c *Code) UnmarshalJSON(b []byte) error { if c == nil { return fmt.Errorf("nil receiver passed to UnmarshalJSON") } + + if ci, err := strconv.ParseUint(string(b), 10, 32); err == nil { + if ci >= _maxCode { + return fmt.Errorf("invalid code: %q", ci) + } + + *c = Code(ci) + return nil + } + if jc, ok := strToCode[string(b)]; ok { *c = jc return nil diff --git a/vendor/google.golang.org/grpc/codes/codes_test.go b/vendor/google.golang.org/grpc/codes/codes_test.go index 1e3b99184..d3e32d229 100644 --- a/vendor/google.golang.org/grpc/codes/codes_test.go +++ b/vendor/google.golang.org/grpc/codes/codes_test.go @@ -62,3 +62,23 @@ func TestUnmarshalJSON_UnknownInput(t *testing.T) { } } } + +func TestUnmarshalJSON_MarshalUnmarshal(t *testing.T) { + for i := 0; i < _maxCode; i++ { + var cUnMarshaled Code + c := Code(i) + + cJSON, err := json.Marshal(c) + if err != nil { + t.Errorf("marshalling %q failed: %v", c, err) + } + + if err := json.Unmarshal(cJSON, &cUnMarshaled); err != nil { + t.Errorf("unmarshalling code failed: %s", err) + } + + if c != cUnMarshaled { + t.Errorf("code is %q after marshalling/unmarshalling, expected %q", cUnMarshaled, c) + } + } +} diff --git a/vendor/google.golang.org/grpc/credentials/alts/alts.go b/vendor/google.golang.org/grpc/credentials/alts/alts.go index c24c5faf3..940056543 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/alts.go +++ b/vendor/google.golang.org/grpc/credentials/alts/alts.go @@ -131,7 +131,6 @@ func DefaultServerOptions() *ServerOptions { // It implements credentials.TransportCredentials interface. type altsTC struct { info *credentials.ProtocolInfo - hsAddr string side core.Side accounts []string hsAddress string @@ -269,8 +268,16 @@ func (g *altsTC) Info() credentials.ProtocolInfo { func (g *altsTC) Clone() credentials.TransportCredentials { info := *g.info + var accounts []string + if g.accounts != nil { + accounts = make([]string, len(g.accounts)) + copy(accounts, g.accounts) + } return &altsTC{ - info: &info, + info: &info, + side: g.side, + hsAddress: g.hsAddress, + accounts: accounts, } } diff --git a/vendor/google.golang.org/grpc/credentials/alts/alts_test.go b/vendor/google.golang.org/grpc/credentials/alts/alts_test.go index 35fae87be..6b041d999 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/alts_test.go +++ b/vendor/google.golang.org/grpc/credentials/alts/alts_test.go @@ -19,6 +19,7 @@ package alts import ( + "reflect" "testing" "github.com/golang/protobuf/proto" @@ -45,10 +46,40 @@ func TestOverrideServerName(t *testing.T) { } } -func TestClone(t *testing.T) { +func TestCloneClient(t *testing.T) { + wantServerName := "server.name" + opt := DefaultClientOptions() + opt.TargetServiceAccounts = []string{"not", "empty"} + c := NewClientCreds(opt) + c.OverrideServerName(wantServerName) + cc := c.Clone() + if got, want := cc.Info().ServerName, wantServerName; got != want { + t.Fatalf("cc.Info().ServerName = %v, want %v", got, want) + } + cc.OverrideServerName("") + if got, want := c.Info().ServerName, wantServerName; got != want { + t.Fatalf("Change in clone should not affect the original, c.Info().ServerName = %v, want %v", got, want) + } + if got, want := cc.Info().ServerName, ""; got != want { + t.Fatalf("cc.Info().ServerName = %v, want %v", got, want) + } + + ct := c.(*altsTC) + cct := cc.(*altsTC) + + if ct.side != cct.side { + t.Errorf("cc.side = %q, want %q", cct.side, ct.side) + } + if ct.hsAddress != cct.hsAddress { + t.Errorf("cc.hsAddress = %q, want %q", cct.hsAddress, ct.hsAddress) + } + if !reflect.DeepEqual(ct.accounts, cct.accounts) { + t.Errorf("cc.accounts = %q, want %q", cct.accounts, ct.accounts) + } +} + +func TestCloneServer(t *testing.T) { wantServerName := "server.name" - // This is not testing any handshaker functionality, so it's fine to only - // use NewServerCreds and not NewClientCreds. c := NewServerCreds(DefaultServerOptions()) c.OverrideServerName(wantServerName) cc := c.Clone() @@ -62,6 +93,19 @@ func TestClone(t *testing.T) { if got, want := cc.Info().ServerName, ""; got != want { t.Fatalf("cc.Info().ServerName = %v, want %v", got, want) } + + ct := c.(*altsTC) + cct := cc.(*altsTC) + + if ct.side != cct.side { + t.Errorf("cc.side = %q, want %q", cct.side, ct.side) + } + if ct.hsAddress != cct.hsAddress { + t.Errorf("cc.hsAddress = %q, want %q", cct.hsAddress, ct.hsAddress) + } + if !reflect.DeepEqual(ct.accounts, cct.accounts) { + t.Errorf("cc.accounts = %q, want %q", cct.accounts, ct.accounts) + } } func TestInfo(t *testing.T) { diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/common.go b/vendor/google.golang.org/grpc/credentials/alts/core/common.go index 0112d51cd..b7baf77ac 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/core/common.go +++ b/vendor/google.golang.org/grpc/credentials/alts/core/common.go @@ -16,6 +16,8 @@ * */ +//go:generate ./regenerate.sh + // Package core contains common core functionality for ALTS. // Disclaimer: users should NEVER reference this package directly. package core diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.pb.go b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.pb.go index 4375ed1fd..cd4fe79bb 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.pb.go +++ b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.pb.go @@ -1,29 +1,7 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: altscontext.proto +// source: grpc/gcp/altscontext.proto -/* -Package grpc_gcp is a generated protocol buffer package. - -It is generated from these files: - altscontext.proto - handshaker.proto - transport_security_common.proto - -It has these top-level messages: - AltsContext - Endpoint - Identity - StartClientHandshakeReq - ServerHandshakeParameters - StartServerHandshakeReq - NextHandshakeMessageReq - HandshakerReq - HandshakerResult - HandshakerStatus - HandshakerResp - RpcProtocolVersions -*/ -package grpc_gcp +package grpc_gcp // import "google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp" import proto "github.com/golang/protobuf/proto" import fmt "fmt" @@ -42,23 +20,45 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type AltsContext struct { // The application protocol negotiated for this connection. - ApplicationProtocol string `protobuf:"bytes,1,opt,name=application_protocol,json=applicationProtocol" json:"application_protocol,omitempty"` + ApplicationProtocol string `protobuf:"bytes,1,opt,name=application_protocol,json=applicationProtocol,proto3" json:"application_protocol,omitempty"` // The record protocol negotiated for this connection. - RecordProtocol string `protobuf:"bytes,2,opt,name=record_protocol,json=recordProtocol" json:"record_protocol,omitempty"` + RecordProtocol string `protobuf:"bytes,2,opt,name=record_protocol,json=recordProtocol,proto3" json:"record_protocol,omitempty"` // The security level of the created secure channel. - SecurityLevel SecurityLevel `protobuf:"varint,3,opt,name=security_level,json=securityLevel,enum=grpc.gcp.SecurityLevel" json:"security_level,omitempty"` + SecurityLevel SecurityLevel `protobuf:"varint,3,opt,name=security_level,json=securityLevel,proto3,enum=grpc.gcp.SecurityLevel" json:"security_level,omitempty"` // The peer service account. - PeerServiceAccount string `protobuf:"bytes,4,opt,name=peer_service_account,json=peerServiceAccount" json:"peer_service_account,omitempty"` + PeerServiceAccount string `protobuf:"bytes,4,opt,name=peer_service_account,json=peerServiceAccount,proto3" json:"peer_service_account,omitempty"` // The local service account. - LocalServiceAccount string `protobuf:"bytes,5,opt,name=local_service_account,json=localServiceAccount" json:"local_service_account,omitempty"` + LocalServiceAccount string `protobuf:"bytes,5,opt,name=local_service_account,json=localServiceAccount,proto3" json:"local_service_account,omitempty"` // The RPC protocol versions supported by the peer. - PeerRpcVersions *RpcProtocolVersions `protobuf:"bytes,6,opt,name=peer_rpc_versions,json=peerRpcVersions" json:"peer_rpc_versions,omitempty"` + PeerRpcVersions *RpcProtocolVersions `protobuf:"bytes,6,opt,name=peer_rpc_versions,json=peerRpcVersions,proto3" json:"peer_rpc_versions,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *AltsContext) Reset() { *m = AltsContext{} } -func (m *AltsContext) String() string { return proto.CompactTextString(m) } -func (*AltsContext) ProtoMessage() {} -func (*AltsContext) Descriptor() ([]byte, []int) { return fileDescriptor0, []int{0} } +func (m *AltsContext) Reset() { *m = AltsContext{} } +func (m *AltsContext) String() string { return proto.CompactTextString(m) } +func (*AltsContext) ProtoMessage() {} +func (*AltsContext) Descriptor() ([]byte, []int) { + return fileDescriptor_altscontext_2f63c0ac7e856743, []int{0} +} +func (m *AltsContext) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_AltsContext.Unmarshal(m, b) +} +func (m *AltsContext) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_AltsContext.Marshal(b, m, deterministic) +} +func (dst *AltsContext) XXX_Merge(src proto.Message) { + xxx_messageInfo_AltsContext.Merge(dst, src) +} +func (m *AltsContext) XXX_Size() int { + return xxx_messageInfo_AltsContext.Size(m) +} +func (m *AltsContext) XXX_DiscardUnknown() { + xxx_messageInfo_AltsContext.DiscardUnknown(m) +} + +var xxx_messageInfo_AltsContext proto.InternalMessageInfo func (m *AltsContext) GetApplicationProtocol() string { if m != nil { @@ -106,26 +106,32 @@ func init() { proto.RegisterType((*AltsContext)(nil), "grpc.gcp.AltsContext") } -func init() { proto.RegisterFile("altscontext.proto", fileDescriptor0) } - -var fileDescriptor0 = []byte{ - // 284 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x64, 0x91, 0xc1, 0x4a, 0x03, 0x31, - 0x10, 0x86, 0xd9, 0xaa, 0x45, 0x53, 0x6c, 0x69, 0x6c, 0xe9, 0x22, 0x88, 0xc5, 0x8b, 0x3d, 0x2d, - 0xba, 0xde, 0x85, 0xea, 0x49, 0xf0, 0x20, 0x5b, 0xf0, 0x1a, 0xe2, 0x18, 0x4a, 0x20, 0xcd, 0x84, - 0x49, 0xba, 0xe8, 0xab, 0xfa, 0x34, 0xb2, 0xc9, 0x6e, 0x5b, 0xf4, 0x98, 0xf9, 0xbe, 0x3f, 0x33, - 0x93, 0xb0, 0xb1, 0x34, 0xc1, 0x03, 0xda, 0xa0, 0xbe, 0x42, 0xe1, 0x08, 0x03, 0xf2, 0xd3, 0x35, - 0x39, 0x28, 0xd6, 0xe0, 0x2e, 0xaf, 0x03, 0x49, 0xeb, 0x1d, 0x52, 0x10, 0x5e, 0xc1, 0x96, 0x74, - 0xf8, 0x16, 0x80, 0x9b, 0x0d, 0xda, 0xa4, 0xde, 0xfc, 0xf4, 0xd8, 0x60, 0x69, 0x82, 0x7f, 0x4e, - 0x17, 0xf0, 0x7b, 0x36, 0x91, 0xce, 0x19, 0x0d, 0x32, 0x68, 0xb4, 0x22, 0x4a, 0x80, 0x26, 0xcf, - 0xe6, 0xd9, 0xe2, 0xac, 0xba, 0x38, 0x60, 0x6f, 0x2d, 0xe2, 0xb7, 0x6c, 0x44, 0x0a, 0x90, 0x3e, - 0xf7, 0x76, 0x2f, 0xda, 0xc3, 0x54, 0xde, 0x89, 0x8f, 0x6c, 0xb8, 0x1b, 0xc2, 0xa8, 0x5a, 0x99, - 0xfc, 0x68, 0x9e, 0x2d, 0x86, 0xe5, 0xac, 0xe8, 0xe6, 0x2d, 0x56, 0x2d, 0x7f, 0x6d, 0x70, 0x75, - 0xee, 0x0f, 0x8f, 0xfc, 0x8e, 0x4d, 0x9c, 0x52, 0x24, 0xbc, 0xa2, 0x5a, 0x83, 0x12, 0x12, 0x00, - 0xb7, 0x36, 0xe4, 0xc7, 0xb1, 0x1b, 0x6f, 0xd8, 0x2a, 0xa1, 0x65, 0x22, 0xbc, 0x64, 0x53, 0x83, - 0x20, 0xcd, 0xbf, 0xc8, 0x49, 0x5a, 0x27, 0xc2, 0x3f, 0x99, 0x17, 0x36, 0x8e, 0x5d, 0xc8, 0x81, - 0xa8, 0x15, 0x79, 0x8d, 0xd6, 0xe7, 0xfd, 0x79, 0xb6, 0x18, 0x94, 0x57, 0xfb, 0x41, 0x2b, 0x07, - 0xdd, 0x5e, 0xef, 0xad, 0x54, 0x8d, 0x9a, 0x5c, 0xe5, 0xa0, 0x2b, 0x3c, 0xcd, 0xd8, 0x54, 0x63, - 0xca, 0x34, 0x9f, 0x54, 0x68, 0x1b, 0x14, 0x59, 0x69, 0x3e, 0xfa, 0xf1, 0xa5, 0x1e, 0x7e, 0x03, - 0x00, 0x00, 0xff, 0xff, 0x36, 0xd0, 0xe1, 0x63, 0xbc, 0x01, 0x00, 0x00, +func init() { + proto.RegisterFile("grpc/gcp/altscontext.proto", fileDescriptor_altscontext_2f63c0ac7e856743) +} + +var fileDescriptor_altscontext_2f63c0ac7e856743 = []byte{ + // 338 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x64, 0x91, 0x4d, 0x4b, 0x23, 0x41, + 0x10, 0x86, 0x99, 0xec, 0x6e, 0xd8, 0xed, 0xb0, 0xc9, 0xee, 0x6c, 0xc2, 0x0e, 0x01, 0x21, 0x78, + 0x71, 0x4e, 0x33, 0x1a, 0x8f, 0x82, 0x90, 0x78, 0x12, 0x3c, 0x84, 0x09, 0x78, 0xf0, 0x32, 0xb4, + 0x95, 0xa2, 0x6d, 0xe8, 0x74, 0x35, 0xd5, 0x9d, 0xa0, 0x7f, 0xd5, 0x5f, 0x23, 0xd3, 0x93, 0x2f, + 0xf4, 0x58, 0xf5, 0xbc, 0x6f, 0x7d, 0x8a, 0xb1, 0x62, 0x07, 0xa5, 0x02, 0x57, 0x4a, 0x13, 0x3c, + 0x90, 0x0d, 0xf8, 0x1a, 0x0a, 0xc7, 0x14, 0x28, 0xfd, 0xd9, 0xb0, 0x42, 0x81, 0x1b, 0xe7, 0x07, + 0x55, 0x60, 0x69, 0xbd, 0x23, 0x0e, 0xb5, 0x47, 0xd8, 0xb0, 0x0e, 0x6f, 0x35, 0xd0, 0x7a, 0x4d, + 0xb6, 0xf5, 0x9c, 0xbf, 0x77, 0x44, 0x6f, 0x66, 0x82, 0xbf, 0x6b, 0x2b, 0xa5, 0x57, 0x62, 0x28, + 0x9d, 0x33, 0x1a, 0x64, 0xd0, 0x64, 0xeb, 0x28, 0x02, 0x32, 0x59, 0x32, 0x49, 0xf2, 0x5f, 0xd5, + 0xbf, 0x13, 0xb6, 0xd8, 0xa1, 0xf4, 0x42, 0x0c, 0x18, 0x81, 0x78, 0x75, 0x54, 0x77, 0xa2, 0xba, + 0xdf, 0xa6, 0x0f, 0xc2, 0x5b, 0xd1, 0x3f, 0x0c, 0x61, 0x70, 0x8b, 0x26, 0xfb, 0x36, 0x49, 0xf2, + 0xfe, 0xf4, 0x7f, 0xb1, 0x1f, 0xbc, 0x58, 0xee, 0xf8, 0x43, 0x83, 0xab, 0xdf, 0xfe, 0x34, 0x4c, + 0x2f, 0xc5, 0xd0, 0x21, 0x72, 0xed, 0x91, 0xb7, 0x1a, 0xb0, 0x96, 0x00, 0xb4, 0xb1, 0x21, 0xfb, + 0x1e, 0xbb, 0xa5, 0x0d, 0x5b, 0xb6, 0x68, 0xd6, 0x92, 0x74, 0x2a, 0x46, 0x86, 0x40, 0x9a, 0x2f, + 0x96, 0x1f, 0xed, 0x3a, 0x11, 0x7e, 0xf2, 0xdc, 0x8b, 0xbf, 0xb1, 0x0b, 0x3b, 0xa8, 0xb7, 0xc8, + 0x5e, 0x93, 0xf5, 0x59, 0x77, 0x92, 0xe4, 0xbd, 0xe9, 0xd9, 0x71, 0xd0, 0xca, 0xc1, 0x7e, 0xaf, + 0xc7, 0x9d, 0xa8, 0x1a, 0x34, 0xbe, 0xca, 0xc1, 0x3e, 0x31, 0x7f, 0x11, 0x23, 0x4d, 0xad, 0xa7, + 0xf9, 0x56, 0xa1, 0x6d, 0x40, 0xb6, 0xd2, 0xcc, 0xff, 0x9c, 0x9c, 0x3c, 0x96, 0x59, 0x24, 0x4f, + 0x37, 0x8a, 0x48, 0x19, 0x2c, 0x14, 0x19, 0x69, 0x55, 0x41, 0xac, 0xca, 0xf8, 0x45, 0x60, 0x5c, + 0xa1, 0x0d, 0x5a, 0x1a, 0x1f, 0x7f, 0x5e, 0x02, 0x31, 0x96, 0xf1, 0xd4, 0x51, 0x50, 0x2b, 0x70, + 0xcf, 0xdd, 0x18, 0x5f, 0x7f, 0x04, 0x00, 0x00, 0xff, 0xff, 0xe8, 0x41, 0xe3, 0x52, 0x1f, 0x02, + 0x00, 0x00, } diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.proto b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.proto deleted file mode 100644 index 9a1dad5f5..000000000 --- a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/altscontext.proto +++ /dev/null @@ -1,41 +0,0 @@ -// Copyright 2018 gRPC authors. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -syntax = "proto3"; - -import "transport_security_common.proto"; - -package grpc.gcp; - -option java_package = "io.grpc.alts.internal"; - -message AltsContext { - // The application protocol negotiated for this connection. - string application_protocol = 1; - - // The record protocol negotiated for this connection. - string record_protocol = 2; - - // The security level of the created secure channel. - SecurityLevel security_level = 3; - - // The peer service account. - string peer_service_account = 4; - - // The local service account. - string local_service_account = 5; - - // The RPC protocol versions supported by the peer. - RpcProtocolVersions peer_rpc_versions = 6; -} diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.pb.go b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.pb.go index 9c3e86070..01e87194b 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.pb.go +++ b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.pb.go @@ -1,7 +1,7 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: handshaker.proto +// source: grpc/gcp/handshaker.proto -package grpc_gcp +package grpc_gcp // import "google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp" import proto "github.com/golang/protobuf/proto" import fmt "fmt" @@ -17,6 +17,12 @@ var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + type HandshakeProtocol int32 const ( @@ -42,7 +48,9 @@ var HandshakeProtocol_value = map[string]int32{ func (x HandshakeProtocol) String() string { return proto.EnumName(HandshakeProtocol_name, int32(x)) } -func (HandshakeProtocol) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{0} } +func (HandshakeProtocol) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{0} +} type NetworkProtocol int32 @@ -66,22 +74,46 @@ var NetworkProtocol_value = map[string]int32{ func (x NetworkProtocol) String() string { return proto.EnumName(NetworkProtocol_name, int32(x)) } -func (NetworkProtocol) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{1} } +func (NetworkProtocol) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{1} +} type Endpoint struct { // IP address. It should contain an IPv4 or IPv6 string literal, e.g. // "192.168.0.1" or "2001:db8::1". - IpAddress string `protobuf:"bytes,1,opt,name=ip_address,json=ipAddress" json:"ip_address,omitempty"` + IpAddress string `protobuf:"bytes,1,opt,name=ip_address,json=ipAddress,proto3" json:"ip_address,omitempty"` // Port number. - Port int32 `protobuf:"varint,2,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,2,opt,name=port,proto3" json:"port,omitempty"` // Network protocol (e.g., TCP, UDP) associated with this endpoint. - Protocol NetworkProtocol `protobuf:"varint,3,opt,name=protocol,enum=grpc.gcp.NetworkProtocol" json:"protocol,omitempty"` + Protocol NetworkProtocol `protobuf:"varint,3,opt,name=protocol,proto3,enum=grpc.gcp.NetworkProtocol" json:"protocol,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *Endpoint) Reset() { *m = Endpoint{} } -func (m *Endpoint) String() string { return proto.CompactTextString(m) } -func (*Endpoint) ProtoMessage() {} -func (*Endpoint) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{0} } +func (m *Endpoint) Reset() { *m = Endpoint{} } +func (m *Endpoint) String() string { return proto.CompactTextString(m) } +func (*Endpoint) ProtoMessage() {} +func (*Endpoint) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{0} +} +func (m *Endpoint) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Endpoint.Unmarshal(m, b) +} +func (m *Endpoint) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Endpoint.Marshal(b, m, deterministic) +} +func (dst *Endpoint) XXX_Merge(src proto.Message) { + xxx_messageInfo_Endpoint.Merge(dst, src) +} +func (m *Endpoint) XXX_Size() int { + return xxx_messageInfo_Endpoint.Size(m) +} +func (m *Endpoint) XXX_DiscardUnknown() { + xxx_messageInfo_Endpoint.DiscardUnknown(m) +} + +var xxx_messageInfo_Endpoint proto.InternalMessageInfo func (m *Endpoint) GetIpAddress() string { if m != nil { @@ -108,23 +140,45 @@ type Identity struct { // Types that are valid to be assigned to IdentityOneof: // *Identity_ServiceAccount // *Identity_Hostname - IdentityOneof isIdentity_IdentityOneof `protobuf_oneof:"identity_oneof"` + IdentityOneof isIdentity_IdentityOneof `protobuf_oneof:"identity_oneof"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *Identity) Reset() { *m = Identity{} } -func (m *Identity) String() string { return proto.CompactTextString(m) } -func (*Identity) ProtoMessage() {} -func (*Identity) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{1} } +func (m *Identity) Reset() { *m = Identity{} } +func (m *Identity) String() string { return proto.CompactTextString(m) } +func (*Identity) ProtoMessage() {} +func (*Identity) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{1} +} +func (m *Identity) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_Identity.Unmarshal(m, b) +} +func (m *Identity) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_Identity.Marshal(b, m, deterministic) +} +func (dst *Identity) XXX_Merge(src proto.Message) { + xxx_messageInfo_Identity.Merge(dst, src) +} +func (m *Identity) XXX_Size() int { + return xxx_messageInfo_Identity.Size(m) +} +func (m *Identity) XXX_DiscardUnknown() { + xxx_messageInfo_Identity.DiscardUnknown(m) +} + +var xxx_messageInfo_Identity proto.InternalMessageInfo type isIdentity_IdentityOneof interface { isIdentity_IdentityOneof() } type Identity_ServiceAccount struct { - ServiceAccount string `protobuf:"bytes,1,opt,name=service_account,json=serviceAccount,oneof"` + ServiceAccount string `protobuf:"bytes,1,opt,name=service_account,json=serviceAccount,proto3,oneof"` } type Identity_Hostname struct { - Hostname string `protobuf:"bytes,2,opt,name=hostname,oneof"` + Hostname string `protobuf:"bytes,2,opt,name=hostname,proto3,oneof"` } func (*Identity_ServiceAccount) isIdentity_IdentityOneof() {} @@ -203,11 +257,11 @@ func _Identity_OneofSizer(msg proto.Message) (n int) { // identity_oneof switch x := m.IdentityOneof.(type) { case *Identity_ServiceAccount: - n += proto.SizeVarint(1<<3 | proto.WireBytes) + n += 1 // tag and wire n += proto.SizeVarint(uint64(len(x.ServiceAccount))) n += len(x.ServiceAccount) case *Identity_Hostname: - n += proto.SizeVarint(2<<3 | proto.WireBytes) + n += 1 // tag and wire n += proto.SizeVarint(uint64(len(x.Hostname))) n += len(x.Hostname) case nil: @@ -219,38 +273,60 @@ func _Identity_OneofSizer(msg proto.Message) (n int) { type StartClientHandshakeReq struct { // Handshake security protocol requested by the client. - HandshakeSecurityProtocol HandshakeProtocol `protobuf:"varint,1,opt,name=handshake_security_protocol,json=handshakeSecurityProtocol,enum=grpc.gcp.HandshakeProtocol" json:"handshake_security_protocol,omitempty"` + HandshakeSecurityProtocol HandshakeProtocol `protobuf:"varint,1,opt,name=handshake_security_protocol,json=handshakeSecurityProtocol,proto3,enum=grpc.gcp.HandshakeProtocol" json:"handshake_security_protocol,omitempty"` // The application protocols supported by the client, e.g., "h2" (for http2), // "grpc". - ApplicationProtocols []string `protobuf:"bytes,2,rep,name=application_protocols,json=applicationProtocols" json:"application_protocols,omitempty"` + ApplicationProtocols []string `protobuf:"bytes,2,rep,name=application_protocols,json=applicationProtocols,proto3" json:"application_protocols,omitempty"` // The record protocols supported by the client, e.g., // "ALTSRP_GCM_AES128". - RecordProtocols []string `protobuf:"bytes,3,rep,name=record_protocols,json=recordProtocols" json:"record_protocols,omitempty"` + RecordProtocols []string `protobuf:"bytes,3,rep,name=record_protocols,json=recordProtocols,proto3" json:"record_protocols,omitempty"` // (Optional) Describes which server identities are acceptable by the client. // If target identities are provided and none of them matches the peer // identity of the server, handshake will fail. - TargetIdentities []*Identity `protobuf:"bytes,4,rep,name=target_identities,json=targetIdentities" json:"target_identities,omitempty"` + TargetIdentities []*Identity `protobuf:"bytes,4,rep,name=target_identities,json=targetIdentities,proto3" json:"target_identities,omitempty"` // (Optional) Application may specify a local identity. Otherwise, the // handshaker chooses a default local identity. - LocalIdentity *Identity `protobuf:"bytes,5,opt,name=local_identity,json=localIdentity" json:"local_identity,omitempty"` + LocalIdentity *Identity `protobuf:"bytes,5,opt,name=local_identity,json=localIdentity,proto3" json:"local_identity,omitempty"` // (Optional) Local endpoint information of the connection to the server, // such as local IP address, port number, and network protocol. - LocalEndpoint *Endpoint `protobuf:"bytes,6,opt,name=local_endpoint,json=localEndpoint" json:"local_endpoint,omitempty"` + LocalEndpoint *Endpoint `protobuf:"bytes,6,opt,name=local_endpoint,json=localEndpoint,proto3" json:"local_endpoint,omitempty"` // (Optional) Endpoint information of the remote server, such as IP address, // port number, and network protocol. - RemoteEndpoint *Endpoint `protobuf:"bytes,7,opt,name=remote_endpoint,json=remoteEndpoint" json:"remote_endpoint,omitempty"` + RemoteEndpoint *Endpoint `protobuf:"bytes,7,opt,name=remote_endpoint,json=remoteEndpoint,proto3" json:"remote_endpoint,omitempty"` // (Optional) If target name is provided, a secure naming check is performed // to verify that the peer authenticated identity is indeed authorized to run // the target name. - TargetName string `protobuf:"bytes,8,opt,name=target_name,json=targetName" json:"target_name,omitempty"` + TargetName string `protobuf:"bytes,8,opt,name=target_name,json=targetName,proto3" json:"target_name,omitempty"` // (Optional) RPC protocol versions supported by the client. - RpcVersions *RpcProtocolVersions `protobuf:"bytes,9,opt,name=rpc_versions,json=rpcVersions" json:"rpc_versions,omitempty"` + RpcVersions *RpcProtocolVersions `protobuf:"bytes,9,opt,name=rpc_versions,json=rpcVersions,proto3" json:"rpc_versions,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *StartClientHandshakeReq) Reset() { *m = StartClientHandshakeReq{} } -func (m *StartClientHandshakeReq) String() string { return proto.CompactTextString(m) } -func (*StartClientHandshakeReq) ProtoMessage() {} -func (*StartClientHandshakeReq) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{2} } +func (m *StartClientHandshakeReq) Reset() { *m = StartClientHandshakeReq{} } +func (m *StartClientHandshakeReq) String() string { return proto.CompactTextString(m) } +func (*StartClientHandshakeReq) ProtoMessage() {} +func (*StartClientHandshakeReq) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{2} +} +func (m *StartClientHandshakeReq) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StartClientHandshakeReq.Unmarshal(m, b) +} +func (m *StartClientHandshakeReq) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StartClientHandshakeReq.Marshal(b, m, deterministic) +} +func (dst *StartClientHandshakeReq) XXX_Merge(src proto.Message) { + xxx_messageInfo_StartClientHandshakeReq.Merge(dst, src) +} +func (m *StartClientHandshakeReq) XXX_Size() int { + return xxx_messageInfo_StartClientHandshakeReq.Size(m) +} +func (m *StartClientHandshakeReq) XXX_DiscardUnknown() { + xxx_messageInfo_StartClientHandshakeReq.DiscardUnknown(m) +} + +var xxx_messageInfo_StartClientHandshakeReq proto.InternalMessageInfo func (m *StartClientHandshakeReq) GetHandshakeSecurityProtocol() HandshakeProtocol { if m != nil { @@ -318,16 +394,38 @@ func (m *StartClientHandshakeReq) GetRpcVersions() *RpcProtocolVersions { type ServerHandshakeParameters struct { // The record protocols supported by the server, e.g., // "ALTSRP_GCM_AES128". - RecordProtocols []string `protobuf:"bytes,1,rep,name=record_protocols,json=recordProtocols" json:"record_protocols,omitempty"` + RecordProtocols []string `protobuf:"bytes,1,rep,name=record_protocols,json=recordProtocols,proto3" json:"record_protocols,omitempty"` // (Optional) A list of local identities supported by the server, if // specified. Otherwise, the handshaker chooses a default local identity. - LocalIdentities []*Identity `protobuf:"bytes,2,rep,name=local_identities,json=localIdentities" json:"local_identities,omitempty"` + LocalIdentities []*Identity `protobuf:"bytes,2,rep,name=local_identities,json=localIdentities,proto3" json:"local_identities,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *ServerHandshakeParameters) Reset() { *m = ServerHandshakeParameters{} } -func (m *ServerHandshakeParameters) String() string { return proto.CompactTextString(m) } -func (*ServerHandshakeParameters) ProtoMessage() {} -func (*ServerHandshakeParameters) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{3} } +func (m *ServerHandshakeParameters) Reset() { *m = ServerHandshakeParameters{} } +func (m *ServerHandshakeParameters) String() string { return proto.CompactTextString(m) } +func (*ServerHandshakeParameters) ProtoMessage() {} +func (*ServerHandshakeParameters) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{3} +} +func (m *ServerHandshakeParameters) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ServerHandshakeParameters.Unmarshal(m, b) +} +func (m *ServerHandshakeParameters) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ServerHandshakeParameters.Marshal(b, m, deterministic) +} +func (dst *ServerHandshakeParameters) XXX_Merge(src proto.Message) { + xxx_messageInfo_ServerHandshakeParameters.Merge(dst, src) +} +func (m *ServerHandshakeParameters) XXX_Size() int { + return xxx_messageInfo_ServerHandshakeParameters.Size(m) +} +func (m *ServerHandshakeParameters) XXX_DiscardUnknown() { + xxx_messageInfo_ServerHandshakeParameters.DiscardUnknown(m) +} + +var xxx_messageInfo_ServerHandshakeParameters proto.InternalMessageInfo func (m *ServerHandshakeParameters) GetRecordProtocols() []string { if m != nil { @@ -346,30 +444,52 @@ func (m *ServerHandshakeParameters) GetLocalIdentities() []*Identity { type StartServerHandshakeReq struct { // The application protocols supported by the server, e.g., "h2" (for http2), // "grpc". - ApplicationProtocols []string `protobuf:"bytes,1,rep,name=application_protocols,json=applicationProtocols" json:"application_protocols,omitempty"` + ApplicationProtocols []string `protobuf:"bytes,1,rep,name=application_protocols,json=applicationProtocols,proto3" json:"application_protocols,omitempty"` // Handshake parameters (record protocols and local identities supported by // the server) mapped by the handshake protocol. Each handshake security // protocol (e.g., TLS or ALTS) has its own set of record protocols and local // identities. Since protobuf does not support enum as key to the map, the key // to handshake_parameters is the integer value of HandshakeProtocol enum. - HandshakeParameters map[int32]*ServerHandshakeParameters `protobuf:"bytes,2,rep,name=handshake_parameters,json=handshakeParameters" json:"handshake_parameters,omitempty" protobuf_key:"varint,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` + HandshakeParameters map[int32]*ServerHandshakeParameters `protobuf:"bytes,2,rep,name=handshake_parameters,json=handshakeParameters,proto3" json:"handshake_parameters,omitempty" protobuf_key:"varint,1,opt,name=key,proto3" protobuf_val:"bytes,2,opt,name=value,proto3"` // Bytes in out_frames returned from the peer's HandshakerResp. It is possible // that the peer's out_frames are split into multiple HandshakReq messages. InBytes []byte `protobuf:"bytes,3,opt,name=in_bytes,json=inBytes,proto3" json:"in_bytes,omitempty"` // (Optional) Local endpoint information of the connection to the client, // such as local IP address, port number, and network protocol. - LocalEndpoint *Endpoint `protobuf:"bytes,4,opt,name=local_endpoint,json=localEndpoint" json:"local_endpoint,omitempty"` + LocalEndpoint *Endpoint `protobuf:"bytes,4,opt,name=local_endpoint,json=localEndpoint,proto3" json:"local_endpoint,omitempty"` // (Optional) Endpoint information of the remote client, such as IP address, // port number, and network protocol. - RemoteEndpoint *Endpoint `protobuf:"bytes,5,opt,name=remote_endpoint,json=remoteEndpoint" json:"remote_endpoint,omitempty"` + RemoteEndpoint *Endpoint `protobuf:"bytes,5,opt,name=remote_endpoint,json=remoteEndpoint,proto3" json:"remote_endpoint,omitempty"` // (Optional) RPC protocol versions supported by the server. - RpcVersions *RpcProtocolVersions `protobuf:"bytes,6,opt,name=rpc_versions,json=rpcVersions" json:"rpc_versions,omitempty"` + RpcVersions *RpcProtocolVersions `protobuf:"bytes,6,opt,name=rpc_versions,json=rpcVersions,proto3" json:"rpc_versions,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *StartServerHandshakeReq) Reset() { *m = StartServerHandshakeReq{} } -func (m *StartServerHandshakeReq) String() string { return proto.CompactTextString(m) } -func (*StartServerHandshakeReq) ProtoMessage() {} -func (*StartServerHandshakeReq) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{4} } +func (m *StartServerHandshakeReq) Reset() { *m = StartServerHandshakeReq{} } +func (m *StartServerHandshakeReq) String() string { return proto.CompactTextString(m) } +func (*StartServerHandshakeReq) ProtoMessage() {} +func (*StartServerHandshakeReq) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{4} +} +func (m *StartServerHandshakeReq) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_StartServerHandshakeReq.Unmarshal(m, b) +} +func (m *StartServerHandshakeReq) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_StartServerHandshakeReq.Marshal(b, m, deterministic) +} +func (dst *StartServerHandshakeReq) XXX_Merge(src proto.Message) { + xxx_messageInfo_StartServerHandshakeReq.Merge(dst, src) +} +func (m *StartServerHandshakeReq) XXX_Size() int { + return xxx_messageInfo_StartServerHandshakeReq.Size(m) +} +func (m *StartServerHandshakeReq) XXX_DiscardUnknown() { + xxx_messageInfo_StartServerHandshakeReq.DiscardUnknown(m) +} + +var xxx_messageInfo_StartServerHandshakeReq proto.InternalMessageInfo func (m *StartServerHandshakeReq) GetApplicationProtocols() []string { if m != nil { @@ -417,13 +537,35 @@ type NextHandshakeMessageReq struct { // Bytes in out_frames returned from the peer's HandshakerResp. It is possible // that the peer's out_frames are split into multiple NextHandshakerMessageReq // messages. - InBytes []byte `protobuf:"bytes,1,opt,name=in_bytes,json=inBytes,proto3" json:"in_bytes,omitempty"` + InBytes []byte `protobuf:"bytes,1,opt,name=in_bytes,json=inBytes,proto3" json:"in_bytes,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *NextHandshakeMessageReq) Reset() { *m = NextHandshakeMessageReq{} } -func (m *NextHandshakeMessageReq) String() string { return proto.CompactTextString(m) } -func (*NextHandshakeMessageReq) ProtoMessage() {} -func (*NextHandshakeMessageReq) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{5} } +func (m *NextHandshakeMessageReq) Reset() { *m = NextHandshakeMessageReq{} } +func (m *NextHandshakeMessageReq) String() string { return proto.CompactTextString(m) } +func (*NextHandshakeMessageReq) ProtoMessage() {} +func (*NextHandshakeMessageReq) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{5} +} +func (m *NextHandshakeMessageReq) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_NextHandshakeMessageReq.Unmarshal(m, b) +} +func (m *NextHandshakeMessageReq) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_NextHandshakeMessageReq.Marshal(b, m, deterministic) +} +func (dst *NextHandshakeMessageReq) XXX_Merge(src proto.Message) { + xxx_messageInfo_NextHandshakeMessageReq.Merge(dst, src) +} +func (m *NextHandshakeMessageReq) XXX_Size() int { + return xxx_messageInfo_NextHandshakeMessageReq.Size(m) +} +func (m *NextHandshakeMessageReq) XXX_DiscardUnknown() { + xxx_messageInfo_NextHandshakeMessageReq.DiscardUnknown(m) +} + +var xxx_messageInfo_NextHandshakeMessageReq proto.InternalMessageInfo func (m *NextHandshakeMessageReq) GetInBytes() []byte { if m != nil { @@ -437,26 +579,48 @@ type HandshakerReq struct { // *HandshakerReq_ClientStart // *HandshakerReq_ServerStart // *HandshakerReq_Next - ReqOneof isHandshakerReq_ReqOneof `protobuf_oneof:"req_oneof"` + ReqOneof isHandshakerReq_ReqOneof `protobuf_oneof:"req_oneof"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *HandshakerReq) Reset() { *m = HandshakerReq{} } -func (m *HandshakerReq) String() string { return proto.CompactTextString(m) } -func (*HandshakerReq) ProtoMessage() {} -func (*HandshakerReq) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{6} } +func (m *HandshakerReq) Reset() { *m = HandshakerReq{} } +func (m *HandshakerReq) String() string { return proto.CompactTextString(m) } +func (*HandshakerReq) ProtoMessage() {} +func (*HandshakerReq) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{6} +} +func (m *HandshakerReq) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_HandshakerReq.Unmarshal(m, b) +} +func (m *HandshakerReq) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_HandshakerReq.Marshal(b, m, deterministic) +} +func (dst *HandshakerReq) XXX_Merge(src proto.Message) { + xxx_messageInfo_HandshakerReq.Merge(dst, src) +} +func (m *HandshakerReq) XXX_Size() int { + return xxx_messageInfo_HandshakerReq.Size(m) +} +func (m *HandshakerReq) XXX_DiscardUnknown() { + xxx_messageInfo_HandshakerReq.DiscardUnknown(m) +} + +var xxx_messageInfo_HandshakerReq proto.InternalMessageInfo type isHandshakerReq_ReqOneof interface { isHandshakerReq_ReqOneof() } type HandshakerReq_ClientStart struct { - ClientStart *StartClientHandshakeReq `protobuf:"bytes,1,opt,name=client_start,json=clientStart,oneof"` + ClientStart *StartClientHandshakeReq `protobuf:"bytes,1,opt,name=client_start,json=clientStart,proto3,oneof"` } type HandshakerReq_ServerStart struct { - ServerStart *StartServerHandshakeReq `protobuf:"bytes,2,opt,name=server_start,json=serverStart,oneof"` + ServerStart *StartServerHandshakeReq `protobuf:"bytes,2,opt,name=server_start,json=serverStart,proto3,oneof"` } type HandshakerReq_Next struct { - Next *NextHandshakeMessageReq `protobuf:"bytes,3,opt,name=next,oneof"` + Next *NextHandshakeMessageReq `protobuf:"bytes,3,opt,name=next,proto3,oneof"` } func (*HandshakerReq_ClientStart) isHandshakerReq_ReqOneof() {} @@ -564,17 +728,17 @@ func _HandshakerReq_OneofSizer(msg proto.Message) (n int) { switch x := m.ReqOneof.(type) { case *HandshakerReq_ClientStart: s := proto.Size(x.ClientStart) - n += proto.SizeVarint(1<<3 | proto.WireBytes) + n += 1 // tag and wire n += proto.SizeVarint(uint64(s)) n += s case *HandshakerReq_ServerStart: s := proto.Size(x.ServerStart) - n += proto.SizeVarint(2<<3 | proto.WireBytes) + n += 1 // tag and wire n += proto.SizeVarint(uint64(s)) n += s case *HandshakerReq_Next: s := proto.Size(x.Next) - n += proto.SizeVarint(3<<3 | proto.WireBytes) + n += 1 // tag and wire n += proto.SizeVarint(uint64(s)) n += s case nil: @@ -586,29 +750,51 @@ func _HandshakerReq_OneofSizer(msg proto.Message) (n int) { type HandshakerResult struct { // The application protocol negotiated for this connection. - ApplicationProtocol string `protobuf:"bytes,1,opt,name=application_protocol,json=applicationProtocol" json:"application_protocol,omitempty"` + ApplicationProtocol string `protobuf:"bytes,1,opt,name=application_protocol,json=applicationProtocol,proto3" json:"application_protocol,omitempty"` // The record protocol negotiated for this connection. - RecordProtocol string `protobuf:"bytes,2,opt,name=record_protocol,json=recordProtocol" json:"record_protocol,omitempty"` + RecordProtocol string `protobuf:"bytes,2,opt,name=record_protocol,json=recordProtocol,proto3" json:"record_protocol,omitempty"` // Cryptographic key data. The key data may be more than the key length // required for the record protocol, thus the client of the handshaker // service needs to truncate the key data into the right key length. KeyData []byte `protobuf:"bytes,3,opt,name=key_data,json=keyData,proto3" json:"key_data,omitempty"` // The authenticated identity of the peer. - PeerIdentity *Identity `protobuf:"bytes,4,opt,name=peer_identity,json=peerIdentity" json:"peer_identity,omitempty"` + PeerIdentity *Identity `protobuf:"bytes,4,opt,name=peer_identity,json=peerIdentity,proto3" json:"peer_identity,omitempty"` // The local identity used in the handshake. - LocalIdentity *Identity `protobuf:"bytes,5,opt,name=local_identity,json=localIdentity" json:"local_identity,omitempty"` + LocalIdentity *Identity `protobuf:"bytes,5,opt,name=local_identity,json=localIdentity,proto3" json:"local_identity,omitempty"` // Indicate whether the handshaker service client should keep the channel // between the handshaker service open, e.g., in order to handle // post-handshake messages in the future. - KeepChannelOpen bool `protobuf:"varint,6,opt,name=keep_channel_open,json=keepChannelOpen" json:"keep_channel_open,omitempty"` + KeepChannelOpen bool `protobuf:"varint,6,opt,name=keep_channel_open,json=keepChannelOpen,proto3" json:"keep_channel_open,omitempty"` // The RPC protocol versions supported by the peer. - PeerRpcVersions *RpcProtocolVersions `protobuf:"bytes,7,opt,name=peer_rpc_versions,json=peerRpcVersions" json:"peer_rpc_versions,omitempty"` + PeerRpcVersions *RpcProtocolVersions `protobuf:"bytes,7,opt,name=peer_rpc_versions,json=peerRpcVersions,proto3" json:"peer_rpc_versions,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *HandshakerResult) Reset() { *m = HandshakerResult{} } -func (m *HandshakerResult) String() string { return proto.CompactTextString(m) } -func (*HandshakerResult) ProtoMessage() {} -func (*HandshakerResult) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{7} } +func (m *HandshakerResult) Reset() { *m = HandshakerResult{} } +func (m *HandshakerResult) String() string { return proto.CompactTextString(m) } +func (*HandshakerResult) ProtoMessage() {} +func (*HandshakerResult) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{7} +} +func (m *HandshakerResult) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_HandshakerResult.Unmarshal(m, b) +} +func (m *HandshakerResult) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_HandshakerResult.Marshal(b, m, deterministic) +} +func (dst *HandshakerResult) XXX_Merge(src proto.Message) { + xxx_messageInfo_HandshakerResult.Merge(dst, src) +} +func (m *HandshakerResult) XXX_Size() int { + return xxx_messageInfo_HandshakerResult.Size(m) +} +func (m *HandshakerResult) XXX_DiscardUnknown() { + xxx_messageInfo_HandshakerResult.DiscardUnknown(m) +} + +var xxx_messageInfo_HandshakerResult proto.InternalMessageInfo func (m *HandshakerResult) GetApplicationProtocol() string { if m != nil { @@ -661,15 +847,37 @@ func (m *HandshakerResult) GetPeerRpcVersions() *RpcProtocolVersions { type HandshakerStatus struct { // The status code. This could be the gRPC status code. - Code uint32 `protobuf:"varint,1,opt,name=code" json:"code,omitempty"` + Code uint32 `protobuf:"varint,1,opt,name=code,proto3" json:"code,omitempty"` // The status details. - Details string `protobuf:"bytes,2,opt,name=details" json:"details,omitempty"` + Details string `protobuf:"bytes,2,opt,name=details,proto3" json:"details,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *HandshakerStatus) Reset() { *m = HandshakerStatus{} } -func (m *HandshakerStatus) String() string { return proto.CompactTextString(m) } -func (*HandshakerStatus) ProtoMessage() {} -func (*HandshakerStatus) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{8} } +func (m *HandshakerStatus) Reset() { *m = HandshakerStatus{} } +func (m *HandshakerStatus) String() string { return proto.CompactTextString(m) } +func (*HandshakerStatus) ProtoMessage() {} +func (*HandshakerStatus) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{8} +} +func (m *HandshakerStatus) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_HandshakerStatus.Unmarshal(m, b) +} +func (m *HandshakerStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_HandshakerStatus.Marshal(b, m, deterministic) +} +func (dst *HandshakerStatus) XXX_Merge(src proto.Message) { + xxx_messageInfo_HandshakerStatus.Merge(dst, src) +} +func (m *HandshakerStatus) XXX_Size() int { + return xxx_messageInfo_HandshakerStatus.Size(m) +} +func (m *HandshakerStatus) XXX_DiscardUnknown() { + xxx_messageInfo_HandshakerStatus.DiscardUnknown(m) +} + +var xxx_messageInfo_HandshakerStatus proto.InternalMessageInfo func (m *HandshakerStatus) GetCode() uint32 { if m != nil { @@ -695,18 +903,40 @@ type HandshakerResp struct { // Number of bytes in the in_bytes consumed by the handshaker. It is possible // that part of in_bytes in HandshakerReq was unrelated to the handshake // process. - BytesConsumed uint32 `protobuf:"varint,2,opt,name=bytes_consumed,json=bytesConsumed" json:"bytes_consumed,omitempty"` + BytesConsumed uint32 `protobuf:"varint,2,opt,name=bytes_consumed,json=bytesConsumed,proto3" json:"bytes_consumed,omitempty"` // This is set iff the handshake was successful. out_frames may still be set // to frames that needs to be forwarded to the peer. - Result *HandshakerResult `protobuf:"bytes,3,opt,name=result" json:"result,omitempty"` + Result *HandshakerResult `protobuf:"bytes,3,opt,name=result,proto3" json:"result,omitempty"` // Status of the handshaker. - Status *HandshakerStatus `protobuf:"bytes,4,opt,name=status" json:"status,omitempty"` + Status *HandshakerStatus `protobuf:"bytes,4,opt,name=status,proto3" json:"status,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *HandshakerResp) Reset() { *m = HandshakerResp{} } -func (m *HandshakerResp) String() string { return proto.CompactTextString(m) } -func (*HandshakerResp) ProtoMessage() {} -func (*HandshakerResp) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{9} } +func (m *HandshakerResp) Reset() { *m = HandshakerResp{} } +func (m *HandshakerResp) String() string { return proto.CompactTextString(m) } +func (*HandshakerResp) ProtoMessage() {} +func (*HandshakerResp) Descriptor() ([]byte, []int) { + return fileDescriptor_handshaker_b28e45bdd1661054, []int{9} +} +func (m *HandshakerResp) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_HandshakerResp.Unmarshal(m, b) +} +func (m *HandshakerResp) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_HandshakerResp.Marshal(b, m, deterministic) +} +func (dst *HandshakerResp) XXX_Merge(src proto.Message) { + xxx_messageInfo_HandshakerResp.Merge(dst, src) +} +func (m *HandshakerResp) XXX_Size() int { + return xxx_messageInfo_HandshakerResp.Size(m) +} +func (m *HandshakerResp) XXX_DiscardUnknown() { + xxx_messageInfo_HandshakerResp.DiscardUnknown(m) +} + +var xxx_messageInfo_HandshakerResp proto.InternalMessageInfo func (m *HandshakerResp) GetOutFrames() []byte { if m != nil { @@ -742,6 +972,7 @@ func init() { proto.RegisterType((*StartClientHandshakeReq)(nil), "grpc.gcp.StartClientHandshakeReq") proto.RegisterType((*ServerHandshakeParameters)(nil), "grpc.gcp.ServerHandshakeParameters") proto.RegisterType((*StartServerHandshakeReq)(nil), "grpc.gcp.StartServerHandshakeReq") + proto.RegisterMapType((map[int32]*ServerHandshakeParameters)(nil), "grpc.gcp.StartServerHandshakeReq.HandshakeParametersEntry") proto.RegisterType((*NextHandshakeMessageReq)(nil), "grpc.gcp.NextHandshakeMessageReq") proto.RegisterType((*HandshakerReq)(nil), "grpc.gcp.HandshakerReq") proto.RegisterType((*HandshakerResult)(nil), "grpc.gcp.HandshakerResult") @@ -759,8 +990,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for HandshakerService service - +// HandshakerServiceClient is the client API for HandshakerService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type HandshakerServiceClient interface { // Handshaker service accepts a stream of handshaker request, returning a // stream of handshaker response. Client is expected to send exactly one @@ -780,7 +1012,7 @@ func NewHandshakerServiceClient(cc *grpc.ClientConn) HandshakerServiceClient { } func (c *handshakerServiceClient) DoHandshake(ctx context.Context, opts ...grpc.CallOption) (HandshakerService_DoHandshakeClient, error) { - stream, err := grpc.NewClientStream(ctx, &_HandshakerService_serviceDesc.Streams[0], c.cc, "/grpc.gcp.HandshakerService/DoHandshake", opts...) + stream, err := c.cc.NewStream(ctx, &_HandshakerService_serviceDesc.Streams[0], "/grpc.gcp.HandshakerService/DoHandshake", opts...) if err != nil { return nil, err } @@ -810,8 +1042,7 @@ func (x *handshakerServiceDoHandshakeClient) Recv() (*HandshakerResp, error) { return m, nil } -// Server API for HandshakerService service - +// HandshakerServiceServer is the server API for HandshakerService service. type HandshakerServiceServer interface { // Handshaker service accepts a stream of handshaker request, returning a // stream of handshaker response. Client is expected to send exactly one @@ -864,79 +1095,84 @@ var _HandshakerService_serviceDesc = grpc.ServiceDesc{ ClientStreams: true, }, }, - Metadata: "handshaker.proto", + Metadata: "grpc/gcp/handshaker.proto", } -func init() { proto.RegisterFile("handshaker.proto", fileDescriptor1) } - -var fileDescriptor1 = []byte{ - // 1073 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x56, 0xdd, 0x6e, 0x1a, 0x47, - 0x14, 0xf6, 0x02, 0xb6, 0xe1, 0x60, 0x60, 0x3d, 0x49, 0x64, 0xec, 0x24, 0x0d, 0xa5, 0xaa, 0x4a, - 0x72, 0x81, 0x5a, 0xd2, 0x2a, 0x4d, 0xaa, 0xaa, 0xb1, 0x31, 0x16, 0x6e, 0x5c, 0x6c, 0x0d, 0x4e, - 0x7b, 0x91, 0x8b, 0xd5, 0x64, 0x39, 0xb1, 0x57, 0xc0, 0xcc, 0x7a, 0x66, 0x70, 0xc3, 0x03, 0xf4, - 0x71, 0xfa, 0x0a, 0x7d, 0x9b, 0xbe, 0x41, 0xef, 0x5b, 0xed, 0xec, 0x1f, 0xc6, 0x4b, 0x94, 0xa8, - 0x77, 0xbb, 0x67, 0xbf, 0xef, 0xec, 0x9c, 0xef, 0x7c, 0x73, 0x66, 0xc0, 0xbe, 0x64, 0x7c, 0xa4, - 0x2e, 0xd9, 0x18, 0x65, 0xdb, 0x97, 0x42, 0x0b, 0x52, 0xbc, 0x90, 0xbe, 0xdb, 0xbe, 0x70, 0xfd, - 0xbd, 0x47, 0x5a, 0x32, 0xae, 0x7c, 0x21, 0xb5, 0xa3, 0xd0, 0x9d, 0x49, 0x4f, 0xcf, 0x1d, 0x57, - 0x4c, 0xa7, 0x82, 0x87, 0xd0, 0xa6, 0x86, 0x62, 0x8f, 0x8f, 0x7c, 0xe1, 0x71, 0x4d, 0x1e, 0x02, - 0x78, 0xbe, 0xc3, 0x46, 0x23, 0x89, 0x4a, 0xd5, 0xad, 0x86, 0xd5, 0x2a, 0xd1, 0x92, 0xe7, 0xef, - 0x87, 0x01, 0x42, 0xa0, 0x10, 0x24, 0xaa, 0xe7, 0x1a, 0x56, 0x6b, 0x9d, 0x9a, 0x67, 0xf2, 0x1d, - 0x14, 0x4d, 0x1e, 0x57, 0x4c, 0xea, 0xf9, 0x86, 0xd5, 0xaa, 0x76, 0x76, 0xdb, 0xf1, 0xcf, 0xdb, - 0x03, 0xd4, 0xbf, 0x0b, 0x39, 0x3e, 0x8b, 0x00, 0x34, 0x81, 0x36, 0x11, 0x8a, 0xc7, 0x23, 0xe4, - 0xda, 0xd3, 0x73, 0xf2, 0x18, 0x6a, 0x0a, 0xe5, 0xb5, 0xe7, 0xa2, 0xc3, 0x5c, 0x57, 0xcc, 0xb8, - 0x0e, 0x7f, 0xdd, 0x5f, 0xa3, 0xd5, 0xe8, 0xc3, 0x7e, 0x18, 0x27, 0x0f, 0xa0, 0x78, 0x29, 0x94, - 0xe6, 0x6c, 0x8a, 0x66, 0x15, 0x01, 0x26, 0x89, 0x1c, 0xd8, 0x50, 0xf5, 0xa2, 0xa4, 0x8e, 0xe0, - 0x28, 0xde, 0x35, 0xff, 0x2c, 0xc0, 0xce, 0x50, 0x33, 0xa9, 0xbb, 0x13, 0x0f, 0xb9, 0xee, 0xc7, - 0x3a, 0x51, 0xbc, 0x22, 0x6f, 0xe0, 0x7e, 0xa2, 0x5b, 0xaa, 0x4d, 0x52, 0x8c, 0x65, 0x8a, 0xb9, - 0x9f, 0x16, 0x93, 0x90, 0x93, 0x72, 0x76, 0x13, 0xfe, 0x30, 0xa2, 0xc7, 0x9f, 0xc8, 0x53, 0xb8, - 0xc7, 0x7c, 0x7f, 0xe2, 0xb9, 0x4c, 0x7b, 0x82, 0x27, 0x59, 0x55, 0x3d, 0xd7, 0xc8, 0xb7, 0x4a, - 0xf4, 0xee, 0xc2, 0xc7, 0x98, 0xa3, 0xc8, 0x63, 0xb0, 0x25, 0xba, 0x42, 0x8e, 0x16, 0xf0, 0x79, - 0x83, 0xaf, 0x85, 0xf1, 0x14, 0xfa, 0x13, 0x6c, 0x6b, 0x26, 0x2f, 0x50, 0x3b, 0x51, 0xc5, 0x1e, - 0xaa, 0x7a, 0xa1, 0x91, 0x6f, 0x95, 0x3b, 0x24, 0x5d, 0x72, 0x2c, 0x31, 0xb5, 0x43, 0xf0, 0x71, - 0x82, 0x25, 0xcf, 0xa1, 0x3a, 0x11, 0x2e, 0x9b, 0xc4, 0xfc, 0x79, 0x7d, 0xbd, 0x61, 0xad, 0x60, - 0x57, 0x0c, 0x32, 0xe9, 0x57, 0x42, 0xc5, 0xc8, 0x37, 0xf5, 0x8d, 0x65, 0x6a, 0xec, 0xa8, 0x88, - 0x9a, 0x18, 0xec, 0x07, 0xa8, 0x49, 0x9c, 0x0a, 0x8d, 0x29, 0x77, 0x73, 0x25, 0xb7, 0x1a, 0x42, - 0x13, 0xf2, 0x23, 0x28, 0x47, 0x35, 0x9b, 0xfe, 0x17, 0x8d, 0x3d, 0x21, 0x0c, 0x0d, 0xd8, 0x14, - 0xc9, 0x4b, 0xd8, 0x92, 0xbe, 0xeb, 0x5c, 0xa3, 0x54, 0x9e, 0xe0, 0xaa, 0x5e, 0x32, 0xa9, 0x1f, - 0xa6, 0xa9, 0xa9, 0xef, 0xc6, 0x12, 0xfe, 0x1a, 0x81, 0x68, 0x59, 0xfa, 0x6e, 0xfc, 0xd2, 0xfc, - 0xc3, 0x82, 0xdd, 0x21, 0xca, 0x6b, 0x94, 0x69, 0xb7, 0x99, 0x64, 0x53, 0xd4, 0x28, 0xb3, 0xfb, - 0x63, 0x65, 0xf7, 0xe7, 0x47, 0xb0, 0x6f, 0xc8, 0x1b, 0xb4, 0x27, 0xb7, 0xb2, 0x3d, 0xb5, 0x45, - 0x81, 0x3d, 0x54, 0xcd, 0x7f, 0xf3, 0x91, 0x6f, 0x97, 0x16, 0x13, 0xf8, 0x76, 0xa5, 0xb5, 0xac, - 0x0f, 0x58, 0x6b, 0x0a, 0x77, 0x53, 0xb3, 0xfb, 0x49, 0x49, 0xd1, 0x9a, 0x5e, 0xa4, 0x6b, 0x5a, - 0xf1, 0xd7, 0x76, 0x86, 0x1e, 0x3d, 0xae, 0xe5, 0x9c, 0xde, 0xb9, 0xcc, 0x50, 0x6a, 0x17, 0x8a, - 0x1e, 0x77, 0xde, 0xce, 0x35, 0x2a, 0x33, 0x15, 0xb6, 0xe8, 0xa6, 0xc7, 0x0f, 0x82, 0xd7, 0x0c, - 0xf7, 0x14, 0xfe, 0x87, 0x7b, 0xd6, 0x3f, 0xda, 0x3d, 0xcb, 0xe6, 0xd8, 0xf8, 0x54, 0x73, 0xec, - 0x8d, 0xa1, 0xbe, 0x4a, 0x05, 0x62, 0x43, 0x7e, 0x8c, 0x73, 0x33, 0x34, 0xd6, 0x69, 0xf0, 0x48, - 0x9e, 0xc3, 0xfa, 0x35, 0x9b, 0xcc, 0xc2, 0x39, 0x55, 0xee, 0x7c, 0xb1, 0x20, 0xf1, 0x2a, 0x83, - 0xd1, 0x90, 0xf1, 0x22, 0xf7, 0xbd, 0xd5, 0xfc, 0x16, 0x76, 0x06, 0xf8, 0x3e, 0x9d, 0x58, 0xbf, - 0xa0, 0x52, 0xec, 0xc2, 0x18, 0x60, 0x51, 0x5c, 0xeb, 0x86, 0xb8, 0xcd, 0xbf, 0x2d, 0xa8, 0x24, - 0x14, 0x19, 0x80, 0x8f, 0x60, 0xcb, 0x35, 0xb3, 0xcf, 0x51, 0x41, 0x67, 0x0d, 0xa1, 0xdc, 0xf9, - 0x7c, 0xa9, 0xe1, 0xb7, 0xc7, 0x63, 0x7f, 0x8d, 0x96, 0x43, 0xa2, 0x01, 0x04, 0x79, 0x94, 0x59, - 0x77, 0x94, 0x27, 0x97, 0x99, 0xe7, 0xb6, 0x71, 0x82, 0x3c, 0x21, 0x31, 0xcc, 0xf3, 0x0c, 0x0a, - 0x1c, 0xdf, 0x6b, 0xe3, 0x8a, 0x1b, 0xfc, 0x15, 0xd5, 0xf6, 0xd7, 0xa8, 0x21, 0x1c, 0x94, 0xa1, - 0x24, 0xf1, 0x2a, 0x9a, 0xeb, 0xff, 0xe4, 0xc0, 0x5e, 0xac, 0x53, 0xcd, 0x26, 0x9a, 0x7c, 0x03, - 0x77, 0xb3, 0x36, 0x46, 0x74, 0x8e, 0xdd, 0xc9, 0xd8, 0x17, 0xe4, 0x2b, 0xa8, 0x2d, 0xed, 0xe8, - 0xf0, 0x58, 0x09, 0xdc, 0xb3, 0xb8, 0xa1, 0x03, 0xcd, 0xc7, 0x38, 0x77, 0x46, 0x4c, 0xb3, 0xd8, - 0xd0, 0x63, 0x9c, 0x1f, 0x32, 0xcd, 0xc8, 0x33, 0xa8, 0xf8, 0x88, 0x32, 0x1d, 0xa4, 0x85, 0x95, - 0x83, 0x74, 0x2b, 0x00, 0xde, 0x9e, 0xa3, 0x9f, 0x3e, 0x82, 0x9f, 0xc0, 0xf6, 0x18, 0xd1, 0x77, - 0xdc, 0x4b, 0xc6, 0x39, 0x4e, 0x1c, 0xe1, 0x23, 0x37, 0x8e, 0x2e, 0xd2, 0x5a, 0xf0, 0xa1, 0x1b, - 0xc6, 0x4f, 0x7d, 0xe4, 0xe4, 0x18, 0xb6, 0xcd, 0xfa, 0x6e, 0xb8, 0x7f, 0xf3, 0x63, 0xdc, 0x5f, - 0x0b, 0x78, 0x74, 0x61, 0x3c, 0xbe, 0x5c, 0x54, 0x7d, 0xa8, 0x99, 0x9e, 0x99, 0x4b, 0x81, 0x2b, - 0x46, 0x68, 0x54, 0xae, 0x50, 0xf3, 0x4c, 0xea, 0xb0, 0x39, 0x42, 0xcd, 0x3c, 0x73, 0xde, 0x05, - 0x72, 0xc6, 0xaf, 0xcd, 0xbf, 0x2c, 0xa8, 0xde, 0x68, 0x9c, 0x1f, 0x5c, 0x3a, 0xc4, 0x4c, 0x3b, - 0xef, 0x82, 0x5d, 0x10, 0x1b, 0xba, 0x24, 0x66, 0xfa, 0xc8, 0x04, 0xc8, 0x97, 0x50, 0x35, 0x56, - 0x77, 0x5c, 0xc1, 0xd5, 0x6c, 0x8a, 0x23, 0x93, 0xb2, 0x42, 0x2b, 0x26, 0xda, 0x8d, 0x82, 0xa4, - 0x03, 0x1b, 0xd2, 0xd8, 0x20, 0x72, 0xd6, 0x5e, 0xc6, 0xc1, 0x1d, 0x19, 0x85, 0x46, 0xc8, 0x80, - 0xa3, 0x4c, 0x11, 0x51, 0xcb, 0x32, 0x39, 0x61, 0x99, 0x34, 0x42, 0x3e, 0xf9, 0x19, 0xb6, 0x6f, - 0x5d, 0x04, 0x48, 0x13, 0x3e, 0xeb, 0xef, 0x0f, 0x0e, 0x87, 0xfd, 0xfd, 0x57, 0x3d, 0xe7, 0x8c, - 0x9e, 0x9e, 0x9f, 0x76, 0x4f, 0x4f, 0x9c, 0xd7, 0x83, 0xe1, 0x59, 0xaf, 0x7b, 0x7c, 0x74, 0xdc, - 0x3b, 0xb4, 0xd7, 0xc8, 0x26, 0xe4, 0xcf, 0x4f, 0x86, 0xb6, 0x45, 0x8a, 0x50, 0xd8, 0x3f, 0x39, - 0x1f, 0xda, 0xb9, 0x27, 0x3d, 0xa8, 0x2d, 0xdd, 0x90, 0x48, 0x03, 0x1e, 0x0c, 0x7a, 0xe7, 0xbf, - 0x9d, 0xd2, 0x57, 0x1f, 0xca, 0xd3, 0x3d, 0xb3, 0xad, 0xe0, 0xe1, 0xf5, 0xe1, 0x99, 0x9d, 0xeb, - 0xbc, 0x59, 0x58, 0x92, 0x1c, 0x86, 0x17, 0x26, 0x72, 0x04, 0xe5, 0x43, 0x91, 0x84, 0xc9, 0x4e, - 0xb6, 0x1c, 0x57, 0x7b, 0xf5, 0x15, 0x3a, 0xf9, 0xcd, 0xb5, 0x96, 0xf5, 0xb5, 0x75, 0xb0, 0x03, - 0xf7, 0x3c, 0x11, 0x62, 0xd8, 0x44, 0xab, 0xb6, 0xc7, 0x35, 0x4a, 0xce, 0x26, 0x6f, 0x37, 0xcc, - 0x8e, 0x79, 0xfa, 0x5f, 0x00, 0x00, 0x00, 0xff, 0xff, 0x76, 0x65, 0x6f, 0x3a, 0x7d, 0x0a, 0x00, - 0x00, +func init() { + proto.RegisterFile("grpc/gcp/handshaker.proto", fileDescriptor_handshaker_b28e45bdd1661054) +} + +var fileDescriptor_handshaker_b28e45bdd1661054 = []byte{ + // 1128 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x56, 0xdf, 0x6e, 0x1a, 0xc7, + 0x17, 0xf6, 0x02, 0xb6, 0xe1, 0x60, 0xfe, 0x78, 0x92, 0x28, 0xd8, 0x49, 0x7e, 0x3f, 0x4a, 0x55, + 0x95, 0xf8, 0x02, 0x5a, 0xd2, 0x2a, 0x4d, 0xa2, 0xaa, 0xb1, 0x31, 0x16, 0x6e, 0x5c, 0x8c, 0x16, + 0xa7, 0x95, 0x9a, 0x8b, 0xd5, 0x64, 0x39, 0x59, 0xaf, 0x58, 0x66, 0xd6, 0x33, 0x83, 0x1b, 0x1e, + 0xa0, 0x8f, 0xd3, 0x57, 0xe8, 0xdb, 0xf4, 0x0d, 0x7a, 0xdf, 0x6a, 0x67, 0xff, 0x61, 0x0c, 0x51, + 0xa2, 0xde, 0xed, 0xce, 0x7c, 0xdf, 0xd9, 0x73, 0xbe, 0xf3, 0xcd, 0xd9, 0x81, 0x3d, 0x47, 0xf8, + 0x76, 0xdb, 0xb1, 0xfd, 0xf6, 0x25, 0x65, 0x63, 0x79, 0x49, 0x27, 0x28, 0x5a, 0xbe, 0xe0, 0x8a, + 0x93, 0x7c, 0xb0, 0xd5, 0x72, 0x6c, 0x7f, 0xbf, 0x99, 0x80, 0x94, 0xa0, 0x4c, 0xfa, 0x5c, 0x28, + 0x4b, 0xa2, 0x3d, 0x13, 0xae, 0x9a, 0x5b, 0x36, 0x9f, 0x4e, 0x39, 0x0b, 0x39, 0x0d, 0x05, 0xf9, + 0x1e, 0x1b, 0xfb, 0xdc, 0x65, 0x8a, 0x3c, 0x02, 0x70, 0x7d, 0x8b, 0x8e, 0xc7, 0x02, 0xa5, 0xac, + 0x19, 0x75, 0xa3, 0x59, 0x30, 0x0b, 0xae, 0x7f, 0x18, 0x2e, 0x10, 0x02, 0xb9, 0x20, 0x50, 0x2d, + 0x53, 0x37, 0x9a, 0x9b, 0xa6, 0x7e, 0x26, 0xdf, 0x42, 0x5e, 0xc7, 0xb1, 0xb9, 0x57, 0xcb, 0xd6, + 0x8d, 0x66, 0xb9, 0xb3, 0xd7, 0x8a, 0xb3, 0x68, 0x0d, 0x50, 0xfd, 0xc6, 0xc5, 0x64, 0x18, 0x01, + 0xcc, 0x04, 0xda, 0x40, 0xc8, 0x9f, 0x8e, 0x91, 0x29, 0x57, 0xcd, 0xc9, 0x63, 0xa8, 0x48, 0x14, + 0xd7, 0xae, 0x8d, 0x16, 0xb5, 0x6d, 0x3e, 0x63, 0x2a, 0xfc, 0x74, 0x7f, 0xc3, 0x2c, 0x47, 0x1b, + 0x87, 0xe1, 0x3a, 0x79, 0x08, 0xf9, 0x4b, 0x2e, 0x15, 0xa3, 0x53, 0xd4, 0x59, 0x04, 0x98, 0x64, + 0xe5, 0xa8, 0x0a, 0x65, 0x37, 0x0a, 0x6a, 0x71, 0x86, 0xfc, 0x5d, 0xe3, 0x8f, 0x1c, 0xdc, 0x1f, + 0x29, 0x2a, 0x54, 0xd7, 0x73, 0x91, 0xa9, 0x7e, 0x2c, 0x98, 0x89, 0x57, 0xe4, 0x0d, 0x3c, 0x48, + 0x04, 0x4c, 0xb5, 0x49, 0x8a, 0x31, 0x74, 0x31, 0x0f, 0xd2, 0x62, 0x12, 0x72, 0x52, 0xce, 0x5e, + 0xc2, 0x1f, 0x45, 0xf4, 0x78, 0x8b, 0x3c, 0x81, 0x7b, 0xd4, 0xf7, 0x3d, 0xd7, 0xa6, 0xca, 0xe5, + 0x2c, 0x89, 0x2a, 0x6b, 0x99, 0x7a, 0xb6, 0x59, 0x30, 0xef, 0x2e, 0x6c, 0xc6, 0x1c, 0x49, 0x1e, + 0x43, 0x55, 0xa0, 0xcd, 0xc5, 0x78, 0x01, 0x9f, 0xd5, 0xf8, 0x4a, 0xb8, 0x9e, 0x42, 0x7f, 0x80, + 0x5d, 0x45, 0x85, 0x83, 0xca, 0x8a, 0x2a, 0x76, 0x51, 0xd6, 0x72, 0xf5, 0x6c, 0xb3, 0xd8, 0x21, + 0x69, 0xca, 0xb1, 0xc4, 0x66, 0x35, 0x04, 0x9f, 0x26, 0x58, 0xf2, 0x0c, 0xca, 0x1e, 0xb7, 0xa9, + 0x17, 0xf3, 0xe7, 0xb5, 0xcd, 0xba, 0xb1, 0x86, 0x5d, 0xd2, 0xc8, 0xa4, 0x5f, 0x09, 0x15, 0x23, + 0xdf, 0xd4, 0xb6, 0x96, 0xa9, 0xb1, 0xa3, 0x22, 0x6a, 0x62, 0xb0, 0x17, 0x50, 0x11, 0x38, 0xe5, + 0x0a, 0x53, 0xee, 0xf6, 0x5a, 0x6e, 0x39, 0x84, 0x26, 0xe4, 0xff, 0x43, 0x31, 0xaa, 0x59, 0xf7, + 0x3f, 0xaf, 0xed, 0x09, 0xe1, 0xd2, 0x80, 0x4e, 0x91, 0xbc, 0x84, 0x1d, 0xe1, 0xdb, 0xd6, 0x35, + 0x0a, 0xe9, 0x72, 0x26, 0x6b, 0x05, 0x1d, 0xfa, 0x51, 0x1a, 0xda, 0xf4, 0xed, 0x58, 0xc2, 0x9f, + 0x23, 0x90, 0x59, 0x14, 0xbe, 0x1d, 0xbf, 0x34, 0x7e, 0x37, 0x60, 0x6f, 0x84, 0xe2, 0x1a, 0x45, + 0xda, 0x6d, 0x2a, 0xe8, 0x14, 0x15, 0x8a, 0xd5, 0xfd, 0x31, 0x56, 0xf7, 0xe7, 0x7b, 0xa8, 0xde, + 0x90, 0x37, 0x68, 0x4f, 0x66, 0x6d, 0x7b, 0x2a, 0x8b, 0x02, 0xbb, 0x28, 0x1b, 0xff, 0x64, 0x23, + 0xdf, 0x2e, 0x25, 0x13, 0xf8, 0x76, 0xad, 0xb5, 0x8c, 0x0f, 0x58, 0x6b, 0x0a, 0x77, 0x53, 0xb3, + 0xfb, 0x49, 0x49, 0x51, 0x4e, 0xcf, 0xd3, 0x9c, 0xd6, 0x7c, 0xb5, 0xb5, 0x42, 0x8f, 0x1e, 0x53, + 0x62, 0x6e, 0xde, 0xb9, 0x5c, 0xa1, 0xd4, 0x1e, 0xe4, 0x5d, 0x66, 0xbd, 0x9d, 0x2b, 0x94, 0x7a, + 0x2a, 0xec, 0x98, 0xdb, 0x2e, 0x3b, 0x0a, 0x5e, 0x57, 0xb8, 0x27, 0xf7, 0x1f, 0xdc, 0xb3, 0xf9, + 0xd1, 0xee, 0x59, 0x36, 0xc7, 0xd6, 0xa7, 0x9a, 0x63, 0x7f, 0x02, 0xb5, 0x75, 0x2a, 0x90, 0x2a, + 0x64, 0x27, 0x38, 0xd7, 0x43, 0x63, 0xd3, 0x0c, 0x1e, 0xc9, 0x33, 0xd8, 0xbc, 0xa6, 0xde, 0x2c, + 0x9c, 0x53, 0xc5, 0xce, 0xe7, 0x0b, 0x12, 0xaf, 0x33, 0x98, 0x19, 0x32, 0x9e, 0x67, 0xbe, 0x33, + 0x1a, 0xdf, 0xc0, 0xfd, 0x01, 0xbe, 0x4f, 0x27, 0xd6, 0x4f, 0x28, 0x25, 0x75, 0xb4, 0x01, 0x16, + 0xc5, 0x35, 0x6e, 0x88, 0xdb, 0xf8, 0xcb, 0x80, 0x52, 0x42, 0x11, 0x01, 0xf8, 0x04, 0x76, 0x6c, + 0x3d, 0xfb, 0x2c, 0x19, 0x74, 0x56, 0x13, 0x8a, 0x9d, 0xcf, 0x96, 0x1a, 0x7e, 0x7b, 0x3c, 0xf6, + 0x37, 0xcc, 0x62, 0x48, 0xd4, 0x80, 0x20, 0x8e, 0xd4, 0x79, 0x47, 0x71, 0x32, 0x2b, 0xe3, 0xdc, + 0x36, 0x4e, 0x10, 0x27, 0x24, 0x86, 0x71, 0x9e, 0x42, 0x8e, 0xe1, 0x7b, 0xa5, 0x5d, 0x71, 0x83, + 0xbf, 0xa6, 0xda, 0xfe, 0x86, 0xa9, 0x09, 0x47, 0x45, 0x28, 0x08, 0xbc, 0x8a, 0xe6, 0xfa, 0xdf, + 0x19, 0xa8, 0x2e, 0xd6, 0x29, 0x67, 0x9e, 0x22, 0x5f, 0xc3, 0xdd, 0x55, 0x07, 0x23, 0xfa, 0x8f, + 0xdd, 0x59, 0x71, 0x2e, 0xc8, 0x97, 0x50, 0x59, 0x3a, 0xd1, 0xe1, 0x6f, 0x25, 0x70, 0xcf, 0xe2, + 0x81, 0x0e, 0x34, 0x9f, 0xe0, 0xdc, 0x1a, 0x53, 0x45, 0x63, 0x43, 0x4f, 0x70, 0x7e, 0x4c, 0x15, + 0x25, 0x4f, 0xa1, 0xe4, 0x23, 0x8a, 0x74, 0x90, 0xe6, 0xd6, 0x0e, 0xd2, 0x9d, 0x00, 0x78, 0x7b, + 0x8e, 0x7e, 0xfa, 0x08, 0x3e, 0x80, 0xdd, 0x09, 0xa2, 0x6f, 0xd9, 0x97, 0x94, 0x31, 0xf4, 0x2c, + 0xee, 0x23, 0xd3, 0x8e, 0xce, 0x9b, 0x95, 0x60, 0xa3, 0x1b, 0xae, 0x9f, 0xfb, 0xc8, 0xc8, 0x29, + 0xec, 0xea, 0xfc, 0x6e, 0xb8, 0x7f, 0xfb, 0x63, 0xdc, 0x5f, 0x09, 0x78, 0xe6, 0xc2, 0x78, 0x7c, + 0xb9, 0xa8, 0xfa, 0x48, 0x51, 0x35, 0xd3, 0x97, 0x02, 0x9b, 0x8f, 0x51, 0xab, 0x5c, 0x32, 0xf5, + 0x33, 0xa9, 0xc1, 0xf6, 0x18, 0x15, 0x75, 0xf5, 0xff, 0x2e, 0x90, 0x33, 0x7e, 0x6d, 0xfc, 0x69, + 0x40, 0xf9, 0x46, 0xe3, 0xfc, 0xe0, 0xd2, 0xc1, 0x67, 0xca, 0x7a, 0x17, 0x9c, 0x82, 0xd8, 0xd0, + 0x05, 0x3e, 0x53, 0x27, 0x7a, 0x81, 0x7c, 0x01, 0x65, 0x6d, 0x75, 0xcb, 0xe6, 0x4c, 0xce, 0xa6, + 0x38, 0xd6, 0x21, 0x4b, 0x66, 0x49, 0xaf, 0x76, 0xa3, 0x45, 0xd2, 0x81, 0x2d, 0xa1, 0x6d, 0x10, + 0x39, 0x6b, 0x7f, 0xc5, 0x8f, 0x3b, 0x32, 0x8a, 0x19, 0x21, 0x03, 0x8e, 0xd4, 0x45, 0x44, 0x2d, + 0x5b, 0xc9, 0x09, 0xcb, 0x34, 0x23, 0xe4, 0xc1, 0x8f, 0xb0, 0x7b, 0xeb, 0x22, 0x40, 0x1a, 0xf0, + 0xbf, 0xfe, 0xe1, 0xe0, 0x78, 0xd4, 0x3f, 0x7c, 0xd5, 0xb3, 0x86, 0xe6, 0xf9, 0xc5, 0x79, 0xf7, + 0xfc, 0xcc, 0x7a, 0x3d, 0x18, 0x0d, 0x7b, 0xdd, 0xd3, 0x93, 0xd3, 0xde, 0x71, 0x75, 0x83, 0x6c, + 0x43, 0xf6, 0xe2, 0x6c, 0x54, 0x35, 0x48, 0x1e, 0x72, 0x87, 0x67, 0x17, 0xa3, 0x6a, 0xe6, 0xa0, + 0x07, 0x95, 0xa5, 0x1b, 0x12, 0xa9, 0xc3, 0xc3, 0x41, 0xef, 0xe2, 0x97, 0x73, 0xf3, 0xd5, 0x87, + 0xe2, 0x74, 0x87, 0x55, 0x23, 0x78, 0x78, 0x7d, 0x3c, 0xac, 0x66, 0x3a, 0x6f, 0x16, 0x52, 0x12, + 0xa3, 0xf0, 0xc2, 0x44, 0x4e, 0xa0, 0x78, 0xcc, 0x93, 0x65, 0x72, 0x7f, 0xb5, 0x1c, 0x57, 0xfb, + 0xb5, 0x35, 0x3a, 0xf9, 0x8d, 0x8d, 0xa6, 0xf1, 0x95, 0x71, 0xe4, 0xc0, 0x3d, 0x97, 0x87, 0x18, + 0xea, 0x29, 0xd9, 0x72, 0x99, 0x42, 0xc1, 0xa8, 0x77, 0x54, 0x49, 0xe1, 0x3a, 0xfb, 0xa1, 0xf1, + 0xeb, 0x0b, 0x87, 0x73, 0xc7, 0xc3, 0x96, 0xc3, 0x3d, 0xca, 0x9c, 0x16, 0x17, 0x4e, 0x5b, 0x5f, + 0x43, 0x6d, 0x81, 0xda, 0xb8, 0xd4, 0x93, 0xed, 0x20, 0x48, 0xdb, 0xe6, 0x02, 0xdb, 0xfa, 0xc4, + 0x69, 0x80, 0xe5, 0xd8, 0xfe, 0xdb, 0x2d, 0xfd, 0xfe, 0xe4, 0xdf, 0x00, 0x00, 0x00, 0xff, 0xff, + 0xfa, 0xd2, 0x42, 0x7f, 0xdf, 0x0a, 0x00, 0x00, } diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.proto b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.proto deleted file mode 100644 index 84a4153b7..000000000 --- a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/handshaker.proto +++ /dev/null @@ -1,224 +0,0 @@ -// Copyright 2018 gRPC authors. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -syntax = "proto3"; - -import "transport_security_common.proto"; - -package grpc.gcp; - -option java_package = "io.grpc.alts.internal"; - -enum HandshakeProtocol { - // Default value. - HANDSHAKE_PROTOCOL_UNSPECIFIED = 0; - - // TLS handshake protocol. - TLS = 1; - - // Application Layer Transport Security handshake protocol. - ALTS = 2; -} - -enum NetworkProtocol { - NETWORK_PROTOCOL_UNSPECIFIED = 0; - TCP = 1; - UDP = 2; -} - -message Endpoint { - // IP address. It should contain an IPv4 or IPv6 string literal, e.g. - // "192.168.0.1" or "2001:db8::1". - string ip_address = 1; - - // Port number. - int32 port = 2; - - // Network protocol (e.g., TCP, UDP) associated with this endpoint. - NetworkProtocol protocol = 3; -} - -message Identity { - oneof identity_oneof { - // Service account of a connection endpoint. - string service_account = 1; - - // Hostname of a connection endpoint. - string hostname = 2; - } -} - -message StartClientHandshakeReq { - // Handshake security protocol requested by the client. - HandshakeProtocol handshake_security_protocol = 1; - - // The application protocols supported by the client, e.g., "h2" (for http2), - // "grpc". - repeated string application_protocols = 2; - - // The record protocols supported by the client, e.g., - // "ALTSRP_GCM_AES128". - repeated string record_protocols = 3; - - // (Optional) Describes which server identities are acceptable by the client. - // If target identities are provided and none of them matches the peer - // identity of the server, handshake will fail. - repeated Identity target_identities = 4; - - // (Optional) Application may specify a local identity. Otherwise, the - // handshaker chooses a default local identity. - Identity local_identity = 5; - - // (Optional) Local endpoint information of the connection to the server, - // such as local IP address, port number, and network protocol. - Endpoint local_endpoint = 6; - - // (Optional) Endpoint information of the remote server, such as IP address, - // port number, and network protocol. - Endpoint remote_endpoint = 7; - - // (Optional) If target name is provided, a secure naming check is performed - // to verify that the peer authenticated identity is indeed authorized to run - // the target name. - string target_name = 8; - - // (Optional) RPC protocol versions supported by the client. - RpcProtocolVersions rpc_versions = 9; -} - -message ServerHandshakeParameters { - // The record protocols supported by the server, e.g., - // "ALTSRP_GCM_AES128". - repeated string record_protocols = 1; - - // (Optional) A list of local identities supported by the server, if - // specified. Otherwise, the handshaker chooses a default local identity. - repeated Identity local_identities = 2; -} - -message StartServerHandshakeReq { - // The application protocols supported by the server, e.g., "h2" (for http2), - // "grpc". - repeated string application_protocols = 1; - - // Handshake parameters (record protocols and local identities supported by - // the server) mapped by the handshake protocol. Each handshake security - // protocol (e.g., TLS or ALTS) has its own set of record protocols and local - // identities. Since protobuf does not support enum as key to the map, the key - // to handshake_parameters is the integer value of HandshakeProtocol enum. - map handshake_parameters = 2; - - // Bytes in out_frames returned from the peer's HandshakerResp. It is possible - // that the peer's out_frames are split into multiple HandshakReq messages. - bytes in_bytes = 3; - - // (Optional) Local endpoint information of the connection to the client, - // such as local IP address, port number, and network protocol. - Endpoint local_endpoint = 4; - - // (Optional) Endpoint information of the remote client, such as IP address, - // port number, and network protocol. - Endpoint remote_endpoint = 5; - - // (Optional) RPC protocol versions supported by the server. - RpcProtocolVersions rpc_versions = 6; -} - -message NextHandshakeMessageReq { - // Bytes in out_frames returned from the peer's HandshakerResp. It is possible - // that the peer's out_frames are split into multiple NextHandshakerMessageReq - // messages. - bytes in_bytes = 1; -} - -message HandshakerReq { - oneof req_oneof { - // The start client handshake request message. - StartClientHandshakeReq client_start = 1; - - // The start server handshake request message. - StartServerHandshakeReq server_start = 2; - - // The next handshake request message. - NextHandshakeMessageReq next = 3; - } -} - -message HandshakerResult { - // The application protocol negotiated for this connection. - string application_protocol = 1; - - // The record protocol negotiated for this connection. - string record_protocol = 2; - - // Cryptographic key data. The key data may be more than the key length - // required for the record protocol, thus the client of the handshaker - // service needs to truncate the key data into the right key length. - bytes key_data = 3; - - // The authenticated identity of the peer. - Identity peer_identity = 4; - - // The local identity used in the handshake. - Identity local_identity = 5; - - // Indicate whether the handshaker service client should keep the channel - // between the handshaker service open, e.g., in order to handle - // post-handshake messages in the future. - bool keep_channel_open = 6; - - // The RPC protocol versions supported by the peer. - RpcProtocolVersions peer_rpc_versions = 7; -} - -message HandshakerStatus { - // The status code. This could be the gRPC status code. - uint32 code = 1; - - // The status details. - string details = 2; -} - -message HandshakerResp { - // Frames to be given to the peer for the NextHandshakeMessageReq. May be - // empty if no out_frames have to be sent to the peer or if in_bytes in the - // HandshakerReq are incomplete. All the non-empty out frames must be sent to - // the peer even if the handshaker status is not OK as these frames may - // contain the alert frames. - bytes out_frames = 1; - - // Number of bytes in the in_bytes consumed by the handshaker. It is possible - // that part of in_bytes in HandshakerReq was unrelated to the handshake - // process. - uint32 bytes_consumed = 2; - - // This is set iff the handshake was successful. out_frames may still be set - // to frames that needs to be forwarded to the peer. - HandshakerResult result = 3; - - // Status of the handshaker. - HandshakerStatus status = 4; -} - -service HandshakerService { - // Handshaker service accepts a stream of handshaker request, returning a - // stream of handshaker response. Client is expected to send exactly one - // message with either client_start or server_start followed by one or more - // messages with next. Each time client sends a request, the handshaker - // service expects to respond. Client does not have to wait for service's - // response before sending next request. - rpc DoHandshake(stream HandshakerReq) - returns (stream HandshakerResp) { - } -} diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.pb.go b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.pb.go index 4e5f2e1b9..120700eaa 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.pb.go +++ b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.pb.go @@ -1,7 +1,7 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: transport_security_common.proto +// source: grpc/gcp/transport_security_common.proto -package grpc_gcp +package grpc_gcp // import "google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp" import proto "github.com/golang/protobuf/proto" import fmt "fmt" @@ -12,6 +12,12 @@ var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf +// This is a compile-time assertion to ensure that this generated file +// is compatible with the proto package it is being compiled against. +// A compilation error at this line likely means your copy of the +// proto package needs to be updated. +const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package + // The security level of the created channel. The list is sorted in increasing // level of security. This order must always be maintained. type SecurityLevel int32 @@ -36,20 +42,44 @@ var SecurityLevel_value = map[string]int32{ func (x SecurityLevel) String() string { return proto.EnumName(SecurityLevel_name, int32(x)) } -func (SecurityLevel) EnumDescriptor() ([]byte, []int) { return fileDescriptor2, []int{0} } +func (SecurityLevel) EnumDescriptor() ([]byte, []int) { + return fileDescriptor_transport_security_common_6489ab913bf26255, []int{0} +} // Max and min supported RPC protocol versions. type RpcProtocolVersions struct { // Maximum supported RPC version. - MaxRpcVersion *RpcProtocolVersions_Version `protobuf:"bytes,1,opt,name=max_rpc_version,json=maxRpcVersion" json:"max_rpc_version,omitempty"` + MaxRpcVersion *RpcProtocolVersions_Version `protobuf:"bytes,1,opt,name=max_rpc_version,json=maxRpcVersion,proto3" json:"max_rpc_version,omitempty"` // Minimum supported RPC version. - MinRpcVersion *RpcProtocolVersions_Version `protobuf:"bytes,2,opt,name=min_rpc_version,json=minRpcVersion" json:"min_rpc_version,omitempty"` + MinRpcVersion *RpcProtocolVersions_Version `protobuf:"bytes,2,opt,name=min_rpc_version,json=minRpcVersion,proto3" json:"min_rpc_version,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *RpcProtocolVersions) Reset() { *m = RpcProtocolVersions{} } -func (m *RpcProtocolVersions) String() string { return proto.CompactTextString(m) } -func (*RpcProtocolVersions) ProtoMessage() {} -func (*RpcProtocolVersions) Descriptor() ([]byte, []int) { return fileDescriptor2, []int{0} } +func (m *RpcProtocolVersions) Reset() { *m = RpcProtocolVersions{} } +func (m *RpcProtocolVersions) String() string { return proto.CompactTextString(m) } +func (*RpcProtocolVersions) ProtoMessage() {} +func (*RpcProtocolVersions) Descriptor() ([]byte, []int) { + return fileDescriptor_transport_security_common_6489ab913bf26255, []int{0} +} +func (m *RpcProtocolVersions) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RpcProtocolVersions.Unmarshal(m, b) +} +func (m *RpcProtocolVersions) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RpcProtocolVersions.Marshal(b, m, deterministic) +} +func (dst *RpcProtocolVersions) XXX_Merge(src proto.Message) { + xxx_messageInfo_RpcProtocolVersions.Merge(dst, src) +} +func (m *RpcProtocolVersions) XXX_Size() int { + return xxx_messageInfo_RpcProtocolVersions.Size(m) +} +func (m *RpcProtocolVersions) XXX_DiscardUnknown() { + xxx_messageInfo_RpcProtocolVersions.DiscardUnknown(m) +} + +var xxx_messageInfo_RpcProtocolVersions proto.InternalMessageInfo func (m *RpcProtocolVersions) GetMaxRpcVersion() *RpcProtocolVersions_Version { if m != nil { @@ -67,14 +97,36 @@ func (m *RpcProtocolVersions) GetMinRpcVersion() *RpcProtocolVersions_Version { // RPC version contains a major version and a minor version. type RpcProtocolVersions_Version struct { - Major uint32 `protobuf:"varint,1,opt,name=major" json:"major,omitempty"` - Minor uint32 `protobuf:"varint,2,opt,name=minor" json:"minor,omitempty"` + Major uint32 `protobuf:"varint,1,opt,name=major,proto3" json:"major,omitempty"` + Minor uint32 `protobuf:"varint,2,opt,name=minor,proto3" json:"minor,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } -func (m *RpcProtocolVersions_Version) Reset() { *m = RpcProtocolVersions_Version{} } -func (m *RpcProtocolVersions_Version) String() string { return proto.CompactTextString(m) } -func (*RpcProtocolVersions_Version) ProtoMessage() {} -func (*RpcProtocolVersions_Version) Descriptor() ([]byte, []int) { return fileDescriptor2, []int{0, 0} } +func (m *RpcProtocolVersions_Version) Reset() { *m = RpcProtocolVersions_Version{} } +func (m *RpcProtocolVersions_Version) String() string { return proto.CompactTextString(m) } +func (*RpcProtocolVersions_Version) ProtoMessage() {} +func (*RpcProtocolVersions_Version) Descriptor() ([]byte, []int) { + return fileDescriptor_transport_security_common_6489ab913bf26255, []int{0, 0} +} +func (m *RpcProtocolVersions_Version) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_RpcProtocolVersions_Version.Unmarshal(m, b) +} +func (m *RpcProtocolVersions_Version) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_RpcProtocolVersions_Version.Marshal(b, m, deterministic) +} +func (dst *RpcProtocolVersions_Version) XXX_Merge(src proto.Message) { + xxx_messageInfo_RpcProtocolVersions_Version.Merge(dst, src) +} +func (m *RpcProtocolVersions_Version) XXX_Size() int { + return xxx_messageInfo_RpcProtocolVersions_Version.Size(m) +} +func (m *RpcProtocolVersions_Version) XXX_DiscardUnknown() { + xxx_messageInfo_RpcProtocolVersions_Version.DiscardUnknown(m) +} + +var xxx_messageInfo_RpcProtocolVersions_Version proto.InternalMessageInfo func (m *RpcProtocolVersions_Version) GetMajor() uint32 { if m != nil { @@ -96,25 +148,31 @@ func init() { proto.RegisterEnum("grpc.gcp.SecurityLevel", SecurityLevel_name, SecurityLevel_value) } -func init() { proto.RegisterFile("transport_security_common.proto", fileDescriptor2) } - -var fileDescriptor2 = []byte{ - // 269 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x90, 0xdf, 0x4a, 0xf3, 0x30, - 0x18, 0xc6, 0xbf, 0x0e, 0x3e, 0x95, 0x48, 0xb5, 0x46, 0x87, 0x7f, 0x4e, 0x14, 0x41, 0x10, 0x0f, - 0x72, 0xa0, 0x78, 0x01, 0x73, 0x16, 0x29, 0xcc, 0x6e, 0x66, 0x73, 0xd0, 0xa3, 0x10, 0x43, 0x19, - 0x91, 0x36, 0x6f, 0x78, 0x13, 0xc7, 0xbc, 0x65, 0xaf, 0x42, 0x9a, 0x76, 0x0c, 0xc1, 0x13, 0xcf, - 0xf2, 0xfc, 0xe0, 0xf9, 0x25, 0x79, 0xc8, 0xb9, 0x47, 0x69, 0x9c, 0x05, 0xf4, 0xc2, 0x95, 0xea, - 0x03, 0xb5, 0xff, 0x14, 0x0a, 0xea, 0x1a, 0x0c, 0xb3, 0x08, 0x1e, 0xe8, 0xce, 0x02, 0xad, 0x62, - 0x0b, 0x65, 0x2f, 0xbf, 0x22, 0x72, 0xc8, 0xad, 0x9a, 0x34, 0x58, 0x41, 0x35, 0x2f, 0xd1, 0x69, - 0x30, 0x8e, 0x3e, 0x93, 0xfd, 0x5a, 0xae, 0x04, 0x5a, 0x25, 0x96, 0x2d, 0x3b, 0x89, 0x2e, 0xa2, - 0xeb, 0xdd, 0xdb, 0x2b, 0xb6, 0xee, 0xb2, 0x5f, 0x7a, 0xac, 0x3b, 0xf0, 0xb8, 0x96, 0x2b, 0x6e, - 0x55, 0x17, 0x83, 0x4e, 0x9b, 0x1f, 0xba, 0xde, 0xdf, 0x74, 0xda, 0x6c, 0x74, 0x67, 0xf7, 0x64, - 0x7b, 0x6d, 0x3e, 0x22, 0xff, 0x6b, 0xf9, 0x0e, 0x18, 0x9e, 0x17, 0xf3, 0x36, 0x04, 0xaa, 0x0d, - 0x60, 0xb8, 0xa5, 0xa1, 0x4d, 0xb8, 0x79, 0x21, 0xf1, 0xb4, 0xdb, 0x63, 0x54, 0x2e, 0xcb, 0x8a, - 0x1e, 0x90, 0x78, 0x9a, 0x0e, 0x5f, 0x79, 0x36, 0x2b, 0x44, 0x3e, 0xce, 0xd3, 0xe4, 0x1f, 0xa5, - 0x64, 0x2f, 0xcb, 0x67, 0xe9, 0x53, 0x60, 0xe3, 0x7c, 0x54, 0x24, 0x11, 0x3d, 0x25, 0xfd, 0x0d, - 0x1b, 0xe4, 0x8f, 0x62, 0xc2, 0xb3, 0xf9, 0x60, 0x58, 0x24, 0xbd, 0x87, 0x63, 0xd2, 0xd7, 0xd0, - 0xfe, 0x41, 0x56, 0xde, 0x31, 0x6d, 0x7c, 0x89, 0x46, 0x56, 0x6f, 0x5b, 0x61, 0xe9, 0xbb, 0xef, - 0x00, 0x00, 0x00, 0xff, 0xff, 0x1f, 0x4d, 0xc2, 0xf0, 0x8c, 0x01, 0x00, 0x00, +func init() { + proto.RegisterFile("grpc/gcp/transport_security_common.proto", fileDescriptor_transport_security_common_6489ab913bf26255) +} + +var fileDescriptor_transport_security_common_6489ab913bf26255 = []byte{ + // 324 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x91, 0x41, 0x4b, 0x3b, 0x31, + 0x10, 0xc5, 0xff, 0x5b, 0xf8, 0xab, 0x44, 0x56, 0xeb, 0x6a, 0x41, 0xc5, 0x83, 0x08, 0x42, 0xf1, + 0x90, 0x05, 0xc5, 0x93, 0xa7, 0xb6, 0x16, 0x29, 0xd4, 0x6d, 0xdd, 0xd6, 0x42, 0xbd, 0x84, 0x18, + 0x43, 0x88, 0x64, 0x33, 0x61, 0x36, 0x96, 0xfa, 0x95, 0xfd, 0x14, 0xb2, 0x69, 0x97, 0x22, 0x78, + 0xf1, 0x96, 0xf7, 0x98, 0xf9, 0x4d, 0x66, 0x1e, 0x69, 0x2b, 0x74, 0x22, 0x55, 0xc2, 0xa5, 0x1e, + 0xb9, 0x2d, 0x1d, 0xa0, 0x67, 0xa5, 0x14, 0x1f, 0xa8, 0xfd, 0x27, 0x13, 0x50, 0x14, 0x60, 0xa9, + 0x43, 0xf0, 0x90, 0xec, 0x54, 0x95, 0x54, 0x09, 0x77, 0xf1, 0x15, 0x91, 0xc3, 0xdc, 0x89, 0x71, + 0x65, 0x0b, 0x30, 0x33, 0x89, 0xa5, 0x06, 0x5b, 0x26, 0x8f, 0x64, 0xbf, 0xe0, 0x4b, 0x86, 0x4e, + 0xb0, 0xc5, 0xca, 0x3b, 0x8e, 0xce, 0xa3, 0xf6, 0xee, 0xf5, 0x25, 0xad, 0x7b, 0xe9, 0x2f, 0x7d, + 0x74, 0xfd, 0xc8, 0xe3, 0x82, 0x2f, 0x73, 0x27, 0xd6, 0x32, 0xe0, 0xb4, 0xfd, 0x81, 0x6b, 0xfc, + 0x0d, 0xa7, 0xed, 0x06, 0x77, 0x7a, 0x4b, 0xb6, 0x6b, 0xf2, 0x11, 0xf9, 0x5f, 0xf0, 0x77, 0xc0, + 0xf0, 0xbd, 0x38, 0x5f, 0x89, 0xe0, 0x6a, 0x0b, 0x18, 0xa6, 0x54, 0x6e, 0x25, 0xae, 0x9e, 0x48, + 0x3c, 0x59, 0xdf, 0x63, 0x28, 0x17, 0xd2, 0x24, 0x07, 0x24, 0x9e, 0xf4, 0x7b, 0xcf, 0xf9, 0x60, + 0x3a, 0x67, 0xd9, 0x28, 0xeb, 0x37, 0xff, 0x25, 0x09, 0xd9, 0x1b, 0x64, 0xd3, 0xfe, 0x43, 0xf0, + 0x46, 0xd9, 0x70, 0xde, 0x8c, 0x92, 0x13, 0xd2, 0xda, 0x78, 0x9d, 0xec, 0x9e, 0x8d, 0xf3, 0xc1, + 0xac, 0xd3, 0x9b, 0x37, 0x1b, 0x5d, 0x4f, 0x5a, 0x1a, 0x56, 0x3b, 0x70, 0xe3, 0x4b, 0xaa, 0xad, + 0x97, 0x68, 0xb9, 0xe9, 0x9e, 0x4d, 0xeb, 0x0c, 0xea, 0x91, 0xbd, 0x90, 0x40, 0x58, 0x71, 0x1c, + 0xbd, 0xdc, 0x29, 0x00, 0x65, 0x24, 0x55, 0x60, 0xb8, 0x55, 0x14, 0x50, 0xa5, 0x21, 0x3e, 0x81, + 0xf2, 0x4d, 0x5a, 0xaf, 0xb9, 0x29, 0xd3, 0x8a, 0x98, 0x0a, 0x40, 0x99, 0x86, 0xd8, 0x42, 0x01, + 0x53, 0xc2, 0xbd, 0x6e, 0x05, 0x7d, 0xf3, 0x1d, 0x00, 0x00, 0xff, 0xff, 0xa8, 0xc8, 0x38, 0x9d, + 0xf2, 0x01, 0x00, 0x00, } diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.proto b/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.proto deleted file mode 100644 index d0f861e64..000000000 --- a/vendor/google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp/transport_security_common.proto +++ /dev/null @@ -1,40 +0,0 @@ -// Copyright 2018 gRPC authors. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -syntax = "proto3"; - -package grpc.gcp; - -option java_package = "io.grpc.alts.internal"; - -// The security level of the created channel. The list is sorted in increasing -// level of security. This order must always be maintained. -enum SecurityLevel { - SECURITY_NONE = 0; - INTEGRITY_ONLY = 1; - INTEGRITY_AND_PRIVACY = 2; -} - -// Max and min supported RPC protocol versions. -message RpcProtocolVersions { - // RPC version contains a major version and a minor version. - message Version { - uint32 major = 1; - uint32 minor = 2; - } - // Maximum supported RPC version. - Version max_rpc_version = 1; - // Minimum supported RPC version. - Version min_rpc_version = 2; -} diff --git a/vendor/google.golang.org/grpc/credentials/alts/core/regenerate.sh b/vendor/google.golang.org/grpc/credentials/alts/core/regenerate.sh new file mode 100755 index 000000000..a79c4201b --- /dev/null +++ b/vendor/google.golang.org/grpc/credentials/alts/core/regenerate.sh @@ -0,0 +1,35 @@ +#!/bin/bash +# Copyright 2018 gRPC authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +set -eux -o pipefail + +TMP=$(mktemp -d) + +function finish { + rm -rf "$TMP" +} +trap finish EXIT + +pushd "$TMP" +mkdir -p grpc/gcp +curl https://raw.githubusercontent.com/grpc/grpc-proto/master/grpc/gcp/altscontext.proto > grpc/gcp/altscontext.proto +curl https://raw.githubusercontent.com/grpc/grpc-proto/master/grpc/gcp/handshaker.proto > grpc/gcp/handshaker.proto +curl https://raw.githubusercontent.com/grpc/grpc-proto/master/grpc/gcp/transport_security_common.proto > grpc/gcp/transport_security_common.proto + +protoc --go_out=plugins=grpc,paths=source_relative:. -I. grpc/gcp/*.proto +popd +rm -f proto/grpc_gcp/*.pb.go +cp "$TMP"/grpc/gcp/*.pb.go proto/grpc_gcp/ + diff --git a/vendor/google.golang.org/grpc/credentials/alts/utils.go b/vendor/google.golang.org/grpc/credentials/alts/utils.go index 86d0a2137..9f92eb1dd 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/utils.go +++ b/vendor/google.golang.org/grpc/credentials/alts/utils.go @@ -30,6 +30,9 @@ import ( "runtime" "strings" "time" + + "golang.org/x/net/context" + "google.golang.org/grpc/peer" ) const ( @@ -115,3 +118,19 @@ func readManufacturer() ([]byte, error) { } return manufacturer, nil } + +// AuthInfoFromContext extracts the alts.AuthInfo object from the given context, +// if it exists. This API should be used by gRPC server RPC handlers to get +// information about the communicating peer. For client-side, use grpc.Peer() +// CallOption. +func AuthInfoFromContext(ctx context.Context) (AuthInfo, error) { + peer, ok := peer.FromContext(ctx) + if !ok { + return nil, errors.New("no Peer found in Context") + } + altsAuthInfo, ok := peer.AuthInfo.(AuthInfo) + if !ok { + return nil, errors.New("no alts.AuthInfo found in Context") + } + return altsAuthInfo, nil +} diff --git a/vendor/google.golang.org/grpc/credentials/alts/utils_test.go b/vendor/google.golang.org/grpc/credentials/alts/utils_test.go index 32c5e1bf4..4724103f6 100644 --- a/vendor/google.golang.org/grpc/credentials/alts/utils_test.go +++ b/vendor/google.golang.org/grpc/credentials/alts/utils_test.go @@ -22,6 +22,10 @@ import ( "io" "strings" "testing" + + "golang.org/x/net/context" + altspb "google.golang.org/grpc/credentials/alts/core/proto/grpc_gcp" + "google.golang.org/grpc/peer" ) func TestIsRunningOnGCP(t *testing.T) { @@ -64,3 +68,44 @@ func setup(testOS string, testReader io.Reader) func() { manufacturerReader = tmpReader } } + +func TestAuthInfoFromContext(t *testing.T) { + ctx := context.Background() + altsAuthInfo := &fakeALTSAuthInfo{} + p := &peer.Peer{ + AuthInfo: altsAuthInfo, + } + for _, tc := range []struct { + desc string + ctx context.Context + success bool + out AuthInfo + }{ + { + "working case", + peer.NewContext(ctx, p), + true, + altsAuthInfo, + }, + } { + authInfo, err := AuthInfoFromContext(tc.ctx) + if got, want := (err == nil), tc.success; got != want { + t.Errorf("%v: AuthInfoFromContext(_)=(err=nil)=%v, want %v", tc.desc, got, want) + } + if got, want := authInfo, tc.out; got != want { + t.Errorf("%v:, AuthInfoFromContext(_)=(%v, _), want (%v, _)", tc.desc, got, want) + } + } +} + +type fakeALTSAuthInfo struct{} + +func (*fakeALTSAuthInfo) AuthType() string { return "" } +func (*fakeALTSAuthInfo) ApplicationProtocol() string { return "" } +func (*fakeALTSAuthInfo) RecordProtocol() string { return "" } +func (*fakeALTSAuthInfo) SecurityLevel() altspb.SecurityLevel { + return altspb.SecurityLevel_SECURITY_NONE +} +func (*fakeALTSAuthInfo) PeerServiceAccount() string { return "" } +func (*fakeALTSAuthInfo) LocalServiceAccount() string { return "" } +func (*fakeALTSAuthInfo) PeerRPCVersions() *altspb.RpcProtocolVersions { return nil } diff --git a/vendor/google.golang.org/grpc/examples/helloworld/helloworld/helloworld.pb.go b/vendor/google.golang.org/grpc/examples/helloworld/helloworld/helloworld.pb.go index f504382f0..11383d740 100644 --- a/vendor/google.golang.org/grpc/examples/helloworld/helloworld/helloworld.pb.go +++ b/vendor/google.golang.org/grpc/examples/helloworld/helloworld/helloworld.pb.go @@ -25,7 +25,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // The request message containing the user's name. type HelloRequest struct { - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -64,7 +64,7 @@ func (m *HelloRequest) GetName() string { // The response message containing the greetings type HelloReply struct { - Message string `protobuf:"bytes,1,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,1,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -114,8 +114,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for Greeter service - +// GreeterClient is the client API for Greeter service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type GreeterClient interface { // Sends a greeting SayHello(ctx context.Context, in *HelloRequest, opts ...grpc.CallOption) (*HelloReply, error) @@ -131,15 +132,14 @@ func NewGreeterClient(cc *grpc.ClientConn) GreeterClient { func (c *greeterClient) SayHello(ctx context.Context, in *HelloRequest, opts ...grpc.CallOption) (*HelloReply, error) { out := new(HelloReply) - err := grpc.Invoke(ctx, "/helloworld.Greeter/SayHello", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/helloworld.Greeter/SayHello", in, out, opts...) if err != nil { return nil, err } return out, nil } -// Server API for Greeter service - +// GreeterServer is the server API for Greeter service. type GreeterServer interface { // Sends a greeting SayHello(context.Context, *HelloRequest) (*HelloReply, error) diff --git a/vendor/google.golang.org/grpc/examples/route_guide/routeguide/route_guide.pb.go b/vendor/google.golang.org/grpc/examples/route_guide/routeguide/route_guide.pb.go index d7934c7c6..199a2ee05 100644 --- a/vendor/google.golang.org/grpc/examples/route_guide/routeguide/route_guide.pb.go +++ b/vendor/google.golang.org/grpc/examples/route_guide/routeguide/route_guide.pb.go @@ -28,8 +28,8 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Latitudes should be in the range +/- 90 degrees and longitude should be in // the range +/- 180 degrees (inclusive). type Point struct { - Latitude int32 `protobuf:"varint,1,opt,name=latitude" json:"latitude,omitempty"` - Longitude int32 `protobuf:"varint,2,opt,name=longitude" json:"longitude,omitempty"` + Latitude int32 `protobuf:"varint,1,opt,name=latitude,proto3" json:"latitude,omitempty"` + Longitude int32 `protobuf:"varint,2,opt,name=longitude,proto3" json:"longitude,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -77,9 +77,9 @@ func (m *Point) GetLongitude() int32 { // points "lo" and "hi". type Rectangle struct { // One corner of the rectangle. - Lo *Point `protobuf:"bytes,1,opt,name=lo" json:"lo,omitempty"` + Lo *Point `protobuf:"bytes,1,opt,name=lo,proto3" json:"lo,omitempty"` // The other corner of the rectangle. - Hi *Point `protobuf:"bytes,2,opt,name=hi" json:"hi,omitempty"` + Hi *Point `protobuf:"bytes,2,opt,name=hi,proto3" json:"hi,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -128,9 +128,9 @@ func (m *Rectangle) GetHi() *Point { // If a feature could not be named, the name is empty. type Feature struct { // The name of the feature. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // The point where the feature is detected. - Location *Point `protobuf:"bytes,2,opt,name=location" json:"location,omitempty"` + Location *Point `protobuf:"bytes,2,opt,name=location,proto3" json:"location,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -177,9 +177,9 @@ func (m *Feature) GetLocation() *Point { // A RouteNote is a message sent while at a given point. type RouteNote struct { // The location from which the message is sent. - Location *Point `protobuf:"bytes,1,opt,name=location" json:"location,omitempty"` + Location *Point `protobuf:"bytes,1,opt,name=location,proto3" json:"location,omitempty"` // The message to be sent. - Message string `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -230,13 +230,13 @@ func (m *RouteNote) GetMessage() string { // the distance between each point. type RouteSummary struct { // The number of points received. - PointCount int32 `protobuf:"varint,1,opt,name=point_count,json=pointCount" json:"point_count,omitempty"` + PointCount int32 `protobuf:"varint,1,opt,name=point_count,json=pointCount,proto3" json:"point_count,omitempty"` // The number of known features passed while traversing the route. - FeatureCount int32 `protobuf:"varint,2,opt,name=feature_count,json=featureCount" json:"feature_count,omitempty"` + FeatureCount int32 `protobuf:"varint,2,opt,name=feature_count,json=featureCount,proto3" json:"feature_count,omitempty"` // The distance covered in metres. - Distance int32 `protobuf:"varint,3,opt,name=distance" json:"distance,omitempty"` + Distance int32 `protobuf:"varint,3,opt,name=distance,proto3" json:"distance,omitempty"` // The duration of the traversal in seconds. - ElapsedTime int32 `protobuf:"varint,4,opt,name=elapsed_time,json=elapsedTime" json:"elapsed_time,omitempty"` + ElapsedTime int32 `protobuf:"varint,4,opt,name=elapsed_time,json=elapsedTime,proto3" json:"elapsed_time,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -310,8 +310,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for RouteGuide service - +// RouteGuideClient is the client API for RouteGuide service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type RouteGuideClient interface { // A simple RPC. // @@ -349,7 +350,7 @@ func NewRouteGuideClient(cc *grpc.ClientConn) RouteGuideClient { func (c *routeGuideClient) GetFeature(ctx context.Context, in *Point, opts ...grpc.CallOption) (*Feature, error) { out := new(Feature) - err := grpc.Invoke(ctx, "/routeguide.RouteGuide/GetFeature", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/routeguide.RouteGuide/GetFeature", in, out, opts...) if err != nil { return nil, err } @@ -357,7 +358,7 @@ func (c *routeGuideClient) GetFeature(ctx context.Context, in *Point, opts ...gr } func (c *routeGuideClient) ListFeatures(ctx context.Context, in *Rectangle, opts ...grpc.CallOption) (RouteGuide_ListFeaturesClient, error) { - stream, err := grpc.NewClientStream(ctx, &_RouteGuide_serviceDesc.Streams[0], c.cc, "/routeguide.RouteGuide/ListFeatures", opts...) + stream, err := c.cc.NewStream(ctx, &_RouteGuide_serviceDesc.Streams[0], "/routeguide.RouteGuide/ListFeatures", opts...) if err != nil { return nil, err } @@ -389,7 +390,7 @@ func (x *routeGuideListFeaturesClient) Recv() (*Feature, error) { } func (c *routeGuideClient) RecordRoute(ctx context.Context, opts ...grpc.CallOption) (RouteGuide_RecordRouteClient, error) { - stream, err := grpc.NewClientStream(ctx, &_RouteGuide_serviceDesc.Streams[1], c.cc, "/routeguide.RouteGuide/RecordRoute", opts...) + stream, err := c.cc.NewStream(ctx, &_RouteGuide_serviceDesc.Streams[1], "/routeguide.RouteGuide/RecordRoute", opts...) if err != nil { return nil, err } @@ -423,7 +424,7 @@ func (x *routeGuideRecordRouteClient) CloseAndRecv() (*RouteSummary, error) { } func (c *routeGuideClient) RouteChat(ctx context.Context, opts ...grpc.CallOption) (RouteGuide_RouteChatClient, error) { - stream, err := grpc.NewClientStream(ctx, &_RouteGuide_serviceDesc.Streams[2], c.cc, "/routeguide.RouteGuide/RouteChat", opts...) + stream, err := c.cc.NewStream(ctx, &_RouteGuide_serviceDesc.Streams[2], "/routeguide.RouteGuide/RouteChat", opts...) if err != nil { return nil, err } @@ -453,8 +454,7 @@ func (x *routeGuideRouteChatClient) Recv() (*RouteNote, error) { return m, nil } -// Server API for RouteGuide service - +// RouteGuideServer is the server API for RouteGuide service. type RouteGuideServer interface { // A simple RPC. // diff --git a/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.pb.go b/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.pb.go index b3b32b48e..59b1ed4ca 100644 --- a/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.pb.go +++ b/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.pb.go @@ -6,6 +6,8 @@ package messages // import "google.golang.org/grpc/grpclb/grpc_lb_v1/messages" import proto "github.com/golang/protobuf/proto" import fmt "fmt" import math "math" +import duration "github.com/golang/protobuf/ptypes/duration" +import timestamp "github.com/golang/protobuf/ptypes/timestamp" // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal @@ -18,113 +20,6 @@ var _ = math.Inf // proto package needs to be updated. const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package -type Duration struct { - // Signed seconds of the span of time. Must be from -315,576,000,000 - // to +315,576,000,000 inclusive. - Seconds int64 `protobuf:"varint,1,opt,name=seconds" json:"seconds,omitempty"` - // Signed fractions of a second at nanosecond resolution of the span - // of time. Durations less than one second are represented with a 0 - // `seconds` field and a positive or negative `nanos` field. For durations - // of one second or more, a non-zero value for the `nanos` field must be - // of the same sign as the `seconds` field. Must be from -999,999,999 - // to +999,999,999 inclusive. - Nanos int32 `protobuf:"varint,2,opt,name=nanos" json:"nanos,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Duration) Reset() { *m = Duration{} } -func (m *Duration) String() string { return proto.CompactTextString(m) } -func (*Duration) ProtoMessage() {} -func (*Duration) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{0} -} -func (m *Duration) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Duration.Unmarshal(m, b) -} -func (m *Duration) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Duration.Marshal(b, m, deterministic) -} -func (dst *Duration) XXX_Merge(src proto.Message) { - xxx_messageInfo_Duration.Merge(dst, src) -} -func (m *Duration) XXX_Size() int { - return xxx_messageInfo_Duration.Size(m) -} -func (m *Duration) XXX_DiscardUnknown() { - xxx_messageInfo_Duration.DiscardUnknown(m) -} - -var xxx_messageInfo_Duration proto.InternalMessageInfo - -func (m *Duration) GetSeconds() int64 { - if m != nil { - return m.Seconds - } - return 0 -} - -func (m *Duration) GetNanos() int32 { - if m != nil { - return m.Nanos - } - return 0 -} - -type Timestamp struct { - // Represents seconds of UTC time since Unix epoch - // 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to - // 9999-12-31T23:59:59Z inclusive. - Seconds int64 `protobuf:"varint,1,opt,name=seconds" json:"seconds,omitempty"` - // Non-negative fractions of a second at nanosecond resolution. Negative - // second values with fractions must still have non-negative nanos values - // that count forward in time. Must be from 0 to 999,999,999 - // inclusive. - Nanos int32 `protobuf:"varint,2,opt,name=nanos" json:"nanos,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` -} - -func (m *Timestamp) Reset() { *m = Timestamp{} } -func (m *Timestamp) String() string { return proto.CompactTextString(m) } -func (*Timestamp) ProtoMessage() {} -func (*Timestamp) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{1} -} -func (m *Timestamp) XXX_Unmarshal(b []byte) error { - return xxx_messageInfo_Timestamp.Unmarshal(m, b) -} -func (m *Timestamp) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { - return xxx_messageInfo_Timestamp.Marshal(b, m, deterministic) -} -func (dst *Timestamp) XXX_Merge(src proto.Message) { - xxx_messageInfo_Timestamp.Merge(dst, src) -} -func (m *Timestamp) XXX_Size() int { - return xxx_messageInfo_Timestamp.Size(m) -} -func (m *Timestamp) XXX_DiscardUnknown() { - xxx_messageInfo_Timestamp.DiscardUnknown(m) -} - -var xxx_messageInfo_Timestamp proto.InternalMessageInfo - -func (m *Timestamp) GetSeconds() int64 { - if m != nil { - return m.Seconds - } - return 0 -} - -func (m *Timestamp) GetNanos() int32 { - if m != nil { - return m.Nanos - } - return 0 -} - type LoadBalanceRequest struct { // Types that are valid to be assigned to LoadBalanceRequestType: // *LoadBalanceRequest_InitialRequest @@ -139,7 +34,7 @@ func (m *LoadBalanceRequest) Reset() { *m = LoadBalanceRequest{} } func (m *LoadBalanceRequest) String() string { return proto.CompactTextString(m) } func (*LoadBalanceRequest) ProtoMessage() {} func (*LoadBalanceRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{2} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{0} } func (m *LoadBalanceRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LoadBalanceRequest.Unmarshal(m, b) @@ -164,10 +59,10 @@ type isLoadBalanceRequest_LoadBalanceRequestType interface { } type LoadBalanceRequest_InitialRequest struct { - InitialRequest *InitialLoadBalanceRequest `protobuf:"bytes,1,opt,name=initial_request,json=initialRequest,oneof"` + InitialRequest *InitialLoadBalanceRequest `protobuf:"bytes,1,opt,name=initial_request,json=initialRequest,proto3,oneof"` } type LoadBalanceRequest_ClientStats struct { - ClientStats *ClientStats `protobuf:"bytes,2,opt,name=client_stats,json=clientStats,oneof"` + ClientStats *ClientStats `protobuf:"bytes,2,opt,name=client_stats,json=clientStats,proto3,oneof"` } func (*LoadBalanceRequest_InitialRequest) isLoadBalanceRequest_LoadBalanceRequestType() {} @@ -269,9 +164,9 @@ func _LoadBalanceRequest_OneofSizer(msg proto.Message) (n int) { } type InitialLoadBalanceRequest struct { - // Name of load balanced service (IE, balancer.service.com) + // Name of load balanced service (IE, balancer.service.com). Its // length should be less than 256 bytes. - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -281,7 +176,7 @@ func (m *InitialLoadBalanceRequest) Reset() { *m = InitialLoadBalanceReq func (m *InitialLoadBalanceRequest) String() string { return proto.CompactTextString(m) } func (*InitialLoadBalanceRequest) ProtoMessage() {} func (*InitialLoadBalanceRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{3} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{1} } func (m *InitialLoadBalanceRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InitialLoadBalanceRequest.Unmarshal(m, b) @@ -308,36 +203,81 @@ func (m *InitialLoadBalanceRequest) GetName() string { return "" } +// Contains the number of calls finished for a particular load balance token. +type ClientStatsPerToken struct { + // See Server.load_balance_token. + LoadBalanceToken string `protobuf:"bytes,1,opt,name=load_balance_token,json=loadBalanceToken,proto3" json:"load_balance_token,omitempty"` + // The total number of RPCs that finished associated with the token. + NumCalls int64 `protobuf:"varint,2,opt,name=num_calls,json=numCalls,proto3" json:"num_calls,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` +} + +func (m *ClientStatsPerToken) Reset() { *m = ClientStatsPerToken{} } +func (m *ClientStatsPerToken) String() string { return proto.CompactTextString(m) } +func (*ClientStatsPerToken) ProtoMessage() {} +func (*ClientStatsPerToken) Descriptor() ([]byte, []int) { + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{2} +} +func (m *ClientStatsPerToken) XXX_Unmarshal(b []byte) error { + return xxx_messageInfo_ClientStatsPerToken.Unmarshal(m, b) +} +func (m *ClientStatsPerToken) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { + return xxx_messageInfo_ClientStatsPerToken.Marshal(b, m, deterministic) +} +func (dst *ClientStatsPerToken) XXX_Merge(src proto.Message) { + xxx_messageInfo_ClientStatsPerToken.Merge(dst, src) +} +func (m *ClientStatsPerToken) XXX_Size() int { + return xxx_messageInfo_ClientStatsPerToken.Size(m) +} +func (m *ClientStatsPerToken) XXX_DiscardUnknown() { + xxx_messageInfo_ClientStatsPerToken.DiscardUnknown(m) +} + +var xxx_messageInfo_ClientStatsPerToken proto.InternalMessageInfo + +func (m *ClientStatsPerToken) GetLoadBalanceToken() string { + if m != nil { + return m.LoadBalanceToken + } + return "" +} + +func (m *ClientStatsPerToken) GetNumCalls() int64 { + if m != nil { + return m.NumCalls + } + return 0 +} + // Contains client level statistics that are useful to load balancing. Each // count except the timestamp should be reset to zero after reporting the stats. type ClientStats struct { // The timestamp of generating the report. - Timestamp *Timestamp `protobuf:"bytes,1,opt,name=timestamp" json:"timestamp,omitempty"` + Timestamp *timestamp.Timestamp `protobuf:"bytes,1,opt,name=timestamp,proto3" json:"timestamp,omitempty"` // The total number of RPCs that started. - NumCallsStarted int64 `protobuf:"varint,2,opt,name=num_calls_started,json=numCallsStarted" json:"num_calls_started,omitempty"` + NumCallsStarted int64 `protobuf:"varint,2,opt,name=num_calls_started,json=numCallsStarted,proto3" json:"num_calls_started,omitempty"` // The total number of RPCs that finished. - NumCallsFinished int64 `protobuf:"varint,3,opt,name=num_calls_finished,json=numCallsFinished" json:"num_calls_finished,omitempty"` - // The total number of RPCs that were dropped by the client because of rate - // limiting. - NumCallsFinishedWithDropForRateLimiting int64 `protobuf:"varint,4,opt,name=num_calls_finished_with_drop_for_rate_limiting,json=numCallsFinishedWithDropForRateLimiting" json:"num_calls_finished_with_drop_for_rate_limiting,omitempty"` - // The total number of RPCs that were dropped by the client because of load - // balancing. - NumCallsFinishedWithDropForLoadBalancing int64 `protobuf:"varint,5,opt,name=num_calls_finished_with_drop_for_load_balancing,json=numCallsFinishedWithDropForLoadBalancing" json:"num_calls_finished_with_drop_for_load_balancing,omitempty"` + NumCallsFinished int64 `protobuf:"varint,3,opt,name=num_calls_finished,json=numCallsFinished,proto3" json:"num_calls_finished,omitempty"` // The total number of RPCs that failed to reach a server except dropped RPCs. - NumCallsFinishedWithClientFailedToSend int64 `protobuf:"varint,6,opt,name=num_calls_finished_with_client_failed_to_send,json=numCallsFinishedWithClientFailedToSend" json:"num_calls_finished_with_client_failed_to_send,omitempty"` + NumCallsFinishedWithClientFailedToSend int64 `protobuf:"varint,6,opt,name=num_calls_finished_with_client_failed_to_send,json=numCallsFinishedWithClientFailedToSend,proto3" json:"num_calls_finished_with_client_failed_to_send,omitempty"` // The total number of RPCs that finished and are known to have been received // by a server. - NumCallsFinishedKnownReceived int64 `protobuf:"varint,7,opt,name=num_calls_finished_known_received,json=numCallsFinishedKnownReceived" json:"num_calls_finished_known_received,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + NumCallsFinishedKnownReceived int64 `protobuf:"varint,7,opt,name=num_calls_finished_known_received,json=numCallsFinishedKnownReceived,proto3" json:"num_calls_finished_known_received,omitempty"` + // The list of dropped calls. + CallsFinishedWithDrop []*ClientStatsPerToken `protobuf:"bytes,8,rep,name=calls_finished_with_drop,json=callsFinishedWithDrop,proto3" json:"calls_finished_with_drop,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *ClientStats) Reset() { *m = ClientStats{} } func (m *ClientStats) String() string { return proto.CompactTextString(m) } func (*ClientStats) ProtoMessage() {} func (*ClientStats) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{4} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{3} } func (m *ClientStats) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ClientStats.Unmarshal(m, b) @@ -357,7 +297,7 @@ func (m *ClientStats) XXX_DiscardUnknown() { var xxx_messageInfo_ClientStats proto.InternalMessageInfo -func (m *ClientStats) GetTimestamp() *Timestamp { +func (m *ClientStats) GetTimestamp() *timestamp.Timestamp { if m != nil { return m.Timestamp } @@ -378,20 +318,6 @@ func (m *ClientStats) GetNumCallsFinished() int64 { return 0 } -func (m *ClientStats) GetNumCallsFinishedWithDropForRateLimiting() int64 { - if m != nil { - return m.NumCallsFinishedWithDropForRateLimiting - } - return 0 -} - -func (m *ClientStats) GetNumCallsFinishedWithDropForLoadBalancing() int64 { - if m != nil { - return m.NumCallsFinishedWithDropForLoadBalancing - } - return 0 -} - func (m *ClientStats) GetNumCallsFinishedWithClientFailedToSend() int64 { if m != nil { return m.NumCallsFinishedWithClientFailedToSend @@ -406,6 +332,13 @@ func (m *ClientStats) GetNumCallsFinishedKnownReceived() int64 { return 0 } +func (m *ClientStats) GetCallsFinishedWithDrop() []*ClientStatsPerToken { + if m != nil { + return m.CallsFinishedWithDrop + } + return nil +} + type LoadBalanceResponse struct { // Types that are valid to be assigned to LoadBalanceResponseType: // *LoadBalanceResponse_InitialResponse @@ -420,7 +353,7 @@ func (m *LoadBalanceResponse) Reset() { *m = LoadBalanceResponse{} } func (m *LoadBalanceResponse) String() string { return proto.CompactTextString(m) } func (*LoadBalanceResponse) ProtoMessage() {} func (*LoadBalanceResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{5} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{4} } func (m *LoadBalanceResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_LoadBalanceResponse.Unmarshal(m, b) @@ -445,10 +378,10 @@ type isLoadBalanceResponse_LoadBalanceResponseType interface { } type LoadBalanceResponse_InitialResponse struct { - InitialResponse *InitialLoadBalanceResponse `protobuf:"bytes,1,opt,name=initial_response,json=initialResponse,oneof"` + InitialResponse *InitialLoadBalanceResponse `protobuf:"bytes,1,opt,name=initial_response,json=initialResponse,proto3,oneof"` } type LoadBalanceResponse_ServerList struct { - ServerList *ServerList `protobuf:"bytes,2,opt,name=server_list,json=serverList,oneof"` + ServerList *ServerList `protobuf:"bytes,2,opt,name=server_list,json=serverList,proto3,oneof"` } func (*LoadBalanceResponse_InitialResponse) isLoadBalanceResponse_LoadBalanceResponseType() {} @@ -555,21 +488,21 @@ type InitialLoadBalanceResponse struct { // the response, the client should open a separate connection to the // load_balancer_delegate and call the BalanceLoad method. Its length should // be less than 64 bytes. - LoadBalancerDelegate string `protobuf:"bytes,1,opt,name=load_balancer_delegate,json=loadBalancerDelegate" json:"load_balancer_delegate,omitempty"` + LoadBalancerDelegate string `protobuf:"bytes,1,opt,name=load_balancer_delegate,json=loadBalancerDelegate,proto3" json:"load_balancer_delegate,omitempty"` // This interval defines how often the client should send the client stats // to the load balancer. Stats should only be reported when the duration is // positive. - ClientStatsReportInterval *Duration `protobuf:"bytes,2,opt,name=client_stats_report_interval,json=clientStatsReportInterval" json:"client_stats_report_interval,omitempty"` - XXX_NoUnkeyedLiteral struct{} `json:"-"` - XXX_unrecognized []byte `json:"-"` - XXX_sizecache int32 `json:"-"` + ClientStatsReportInterval *duration.Duration `protobuf:"bytes,2,opt,name=client_stats_report_interval,json=clientStatsReportInterval,proto3" json:"client_stats_report_interval,omitempty"` + XXX_NoUnkeyedLiteral struct{} `json:"-"` + XXX_unrecognized []byte `json:"-"` + XXX_sizecache int32 `json:"-"` } func (m *InitialLoadBalanceResponse) Reset() { *m = InitialLoadBalanceResponse{} } func (m *InitialLoadBalanceResponse) String() string { return proto.CompactTextString(m) } func (*InitialLoadBalanceResponse) ProtoMessage() {} func (*InitialLoadBalanceResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{6} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{5} } func (m *InitialLoadBalanceResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_InitialLoadBalanceResponse.Unmarshal(m, b) @@ -596,7 +529,7 @@ func (m *InitialLoadBalanceResponse) GetLoadBalancerDelegate() string { return "" } -func (m *InitialLoadBalanceResponse) GetClientStatsReportInterval() *Duration { +func (m *InitialLoadBalanceResponse) GetClientStatsReportInterval() *duration.Duration { if m != nil { return m.ClientStatsReportInterval } @@ -608,7 +541,7 @@ type ServerList struct { // be updated when server resolutions change or as needed to balance load // across more servers. The client should consume the server list in order // unless instructed otherwise via the client_config. - Servers []*Server `protobuf:"bytes,1,rep,name=servers" json:"servers,omitempty"` + Servers []*Server `protobuf:"bytes,1,rep,name=servers,proto3" json:"servers,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -618,7 +551,7 @@ func (m *ServerList) Reset() { *m = ServerList{} } func (m *ServerList) String() string { return proto.CompactTextString(m) } func (*ServerList) ProtoMessage() {} func (*ServerList) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{7} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{6} } func (m *ServerList) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_ServerList.Unmarshal(m, b) @@ -645,29 +578,24 @@ func (m *ServerList) GetServers() []*Server { return nil } -// Contains server information. When none of the [drop_for_*] fields are true, -// use the other fields. When drop_for_rate_limiting is true, ignore all other -// fields. Use drop_for_load_balancing only when it is true and -// drop_for_rate_limiting is false. +// Contains server information. When the drop field is not true, use the other +// fields. type Server struct { // A resolved address for the server, serialized in network-byte-order. It may // either be an IPv4 or IPv6 address. IpAddress []byte `protobuf:"bytes,1,opt,name=ip_address,json=ipAddress,proto3" json:"ip_address,omitempty"` // A resolved port number for the server. - Port int32 `protobuf:"varint,2,opt,name=port" json:"port,omitempty"` + Port int32 `protobuf:"varint,2,opt,name=port,proto3" json:"port,omitempty"` // An opaque but printable token given to the frontend for each pick. All // frontend requests for that pick must include the token in its initial // metadata. The token is used by the backend to verify the request and to - // allow the backend to report load to the gRPC LB system. - // - // Its length is variable but less than 50 bytes. - LoadBalanceToken string `protobuf:"bytes,3,opt,name=load_balance_token,json=loadBalanceToken" json:"load_balance_token,omitempty"` - // Indicates whether this particular request should be dropped by the client - // for rate limiting. - DropForRateLimiting bool `protobuf:"varint,4,opt,name=drop_for_rate_limiting,json=dropForRateLimiting" json:"drop_for_rate_limiting,omitempty"` - // Indicates whether this particular request should be dropped by the client - // for load balancing. - DropForLoadBalancing bool `protobuf:"varint,5,opt,name=drop_for_load_balancing,json=dropForLoadBalancing" json:"drop_for_load_balancing,omitempty"` + // allow the backend to report load to the gRPC LB system. The token is also + // used in client stats for reporting dropped calls. + LoadBalanceToken string `protobuf:"bytes,3,opt,name=load_balance_token,json=loadBalanceToken,proto3" json:"load_balance_token,omitempty"` + // Indicates whether this particular request should be dropped by the client. + // If the request is dropped, there will be a corresponding entry in + // ClientStats.calls_finished_with_drop. + Drop bool `protobuf:"varint,4,opt,name=drop,proto3" json:"drop,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -677,7 +605,7 @@ func (m *Server) Reset() { *m = Server{} } func (m *Server) String() string { return proto.CompactTextString(m) } func (*Server) ProtoMessage() {} func (*Server) Descriptor() ([]byte, []int) { - return fileDescriptor_messages_b81c731f0e83edbd, []int{8} + return fileDescriptor_messages_b3d89fcb5aa158f8, []int{7} } func (m *Server) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_Server.Unmarshal(m, b) @@ -718,25 +646,17 @@ func (m *Server) GetLoadBalanceToken() string { return "" } -func (m *Server) GetDropForRateLimiting() bool { +func (m *Server) GetDrop() bool { if m != nil { - return m.DropForRateLimiting - } - return false -} - -func (m *Server) GetDropForLoadBalancing() bool { - if m != nil { - return m.DropForLoadBalancing + return m.Drop } return false } func init() { - proto.RegisterType((*Duration)(nil), "grpc.lb.v1.Duration") - proto.RegisterType((*Timestamp)(nil), "grpc.lb.v1.Timestamp") proto.RegisterType((*LoadBalanceRequest)(nil), "grpc.lb.v1.LoadBalanceRequest") proto.RegisterType((*InitialLoadBalanceRequest)(nil), "grpc.lb.v1.InitialLoadBalanceRequest") + proto.RegisterType((*ClientStatsPerToken)(nil), "grpc.lb.v1.ClientStatsPerToken") proto.RegisterType((*ClientStats)(nil), "grpc.lb.v1.ClientStats") proto.RegisterType((*LoadBalanceResponse)(nil), "grpc.lb.v1.LoadBalanceResponse") proto.RegisterType((*InitialLoadBalanceResponse)(nil), "grpc.lb.v1.InitialLoadBalanceResponse") @@ -745,55 +665,54 @@ func init() { } func init() { - proto.RegisterFile("grpc_lb_v1/messages/messages.proto", fileDescriptor_messages_b81c731f0e83edbd) + proto.RegisterFile("grpc_lb_v1/messages/messages.proto", fileDescriptor_messages_b3d89fcb5aa158f8) } -var fileDescriptor_messages_b81c731f0e83edbd = []byte{ - // 731 bytes of a gzipped FileDescriptorProto - 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x94, 0x55, 0xdd, 0x4e, 0x1b, 0x39, - 0x14, 0x26, 0x9b, 0x00, 0xc9, 0x09, 0x5a, 0xb2, 0x26, 0x0b, 0x81, 0x05, 0x89, 0x1d, 0x69, 0xd9, - 0x68, 0xc5, 0x4e, 0x04, 0xd9, 0xbd, 0xe8, 0xcf, 0x45, 0x1b, 0x10, 0x0a, 0x2d, 0x17, 0x95, 0x43, - 0x55, 0xa9, 0x52, 0x65, 0x39, 0x19, 0x33, 0x58, 0x38, 0xf6, 0xd4, 0x76, 0x82, 0xfa, 0x08, 0x7d, - 0x94, 0x3e, 0x46, 0xd5, 0x67, 0xe8, 0xfb, 0x54, 0xe3, 0x99, 0xc9, 0x0c, 0x10, 0x40, 0xbd, 0x89, - 0xec, 0xe3, 0xef, 0x7c, 0xdf, 0xf1, 0x89, 0xbf, 0x33, 0xe0, 0x85, 0x3a, 0x1a, 0x11, 0x31, 0x24, - 0xd3, 0x83, 0xce, 0x98, 0x19, 0x43, 0x43, 0x66, 0x66, 0x0b, 0x3f, 0xd2, 0xca, 0x2a, 0x04, 0x31, - 0xc6, 0x17, 0x43, 0x7f, 0x7a, 0xe0, 0x3d, 0x85, 0xea, 0xf1, 0x44, 0x53, 0xcb, 0x95, 0x44, 0x2d, - 0x58, 0x36, 0x6c, 0xa4, 0x64, 0x60, 0x5a, 0xa5, 0xdd, 0x52, 0xbb, 0x8c, 0xb3, 0x2d, 0x6a, 0xc2, - 0xa2, 0xa4, 0x52, 0x99, 0xd6, 0x2f, 0xbb, 0xa5, 0xf6, 0x22, 0x4e, 0x36, 0xde, 0x33, 0xa8, 0x9d, - 0xf3, 0x31, 0x33, 0x96, 0x8e, 0xa3, 0x9f, 0x4e, 0xfe, 0x5a, 0x02, 0x74, 0xa6, 0x68, 0xd0, 0xa3, - 0x82, 0xca, 0x11, 0xc3, 0xec, 0xe3, 0x84, 0x19, 0x8b, 0xde, 0xc0, 0x2a, 0x97, 0xdc, 0x72, 0x2a, - 0x88, 0x4e, 0x42, 0x8e, 0xae, 0x7e, 0xf8, 0x97, 0x9f, 0x57, 0xed, 0x9f, 0x26, 0x90, 0xbb, 0xf9, - 0xfd, 0x05, 0xfc, 0x6b, 0x9a, 0x9f, 0x31, 0x3e, 0x87, 0x95, 0x91, 0xe0, 0x4c, 0x5a, 0x62, 0x2c, - 0xb5, 0x49, 0x15, 0xf5, 0xc3, 0x8d, 0x22, 0xdd, 0x91, 0x3b, 0x1f, 0xc4, 0xc7, 0xfd, 0x05, 0x5c, - 0x1f, 0xe5, 0xdb, 0xde, 0x1f, 0xb0, 0x29, 0x14, 0x0d, 0xc8, 0x30, 0x91, 0xc9, 0x8a, 0x22, 0xf6, - 0x53, 0xc4, 0xbc, 0x0e, 0x6c, 0xde, 0x5b, 0x09, 0x42, 0x50, 0x91, 0x74, 0xcc, 0x5c, 0xf9, 0x35, - 0xec, 0xd6, 0xde, 0xe7, 0x0a, 0xd4, 0x0b, 0x62, 0xa8, 0x0b, 0x35, 0x9b, 0x75, 0x30, 0xbd, 0xe7, - 0xef, 0xc5, 0xc2, 0x66, 0xed, 0xc5, 0x39, 0x0e, 0xfd, 0x03, 0xbf, 0xc9, 0xc9, 0x98, 0x8c, 0xa8, - 0x10, 0x26, 0xbe, 0x93, 0xb6, 0x2c, 0x70, 0xb7, 0x2a, 0xe3, 0x55, 0x39, 0x19, 0x1f, 0xc5, 0xf1, - 0x41, 0x12, 0x46, 0xfb, 0x80, 0x72, 0xec, 0x05, 0x97, 0xdc, 0x5c, 0xb2, 0xa0, 0x55, 0x76, 0xe0, - 0x46, 0x06, 0x3e, 0x49, 0xe3, 0x88, 0x80, 0x7f, 0x17, 0x4d, 0xae, 0xb9, 0xbd, 0x24, 0x81, 0x56, - 0x11, 0xb9, 0x50, 0x9a, 0x68, 0x6a, 0x19, 0x11, 0x7c, 0xcc, 0x2d, 0x97, 0x61, 0xab, 0xe2, 0x98, - 0xfe, 0xbe, 0xcd, 0xf4, 0x8e, 0xdb, 0xcb, 0x63, 0xad, 0xa2, 0x13, 0xa5, 0x31, 0xb5, 0xec, 0x2c, - 0x85, 0x23, 0x0a, 0x9d, 0x47, 0x05, 0x0a, 0xed, 0x8e, 0x15, 0x16, 0x9d, 0x42, 0xfb, 0x01, 0x85, - 0xbc, 0xf7, 0xb1, 0xc4, 0x07, 0xf8, 0xf7, 0x3e, 0x89, 0xf4, 0x19, 0x5c, 0x50, 0x2e, 0x58, 0x40, - 0xac, 0x22, 0x86, 0xc9, 0xa0, 0xb5, 0xe4, 0x04, 0xf6, 0xe6, 0x09, 0x24, 0x7f, 0xd5, 0x89, 0xc3, - 0x9f, 0xab, 0x01, 0x93, 0x01, 0xea, 0xc3, 0x9f, 0x73, 0xe8, 0xaf, 0xa4, 0xba, 0x96, 0x44, 0xb3, - 0x11, 0xe3, 0x53, 0x16, 0xb4, 0x96, 0x1d, 0xe5, 0xce, 0x6d, 0xca, 0xd7, 0x31, 0x0a, 0xa7, 0x20, - 0xef, 0x5b, 0x09, 0xd6, 0x6e, 0x3c, 0x1b, 0x13, 0x29, 0x69, 0x18, 0x1a, 0x40, 0x23, 0x77, 0x40, - 0x12, 0x4b, 0x9f, 0xc6, 0xde, 0x63, 0x16, 0x48, 0xd0, 0xfd, 0x05, 0xbc, 0x3a, 0xf3, 0x40, 0x4a, - 0xfa, 0x04, 0xea, 0x86, 0xe9, 0x29, 0xd3, 0x44, 0x70, 0x63, 0x53, 0x0f, 0xac, 0x17, 0xf9, 0x06, - 0xee, 0xf8, 0x8c, 0x3b, 0x0f, 0x81, 0x99, 0xed, 0x7a, 0xdb, 0xb0, 0x75, 0xcb, 0x01, 0x09, 0x67, - 0x62, 0x81, 0x2f, 0x25, 0xd8, 0xba, 0xbf, 0x14, 0xf4, 0x1f, 0xac, 0x17, 0x93, 0x35, 0x09, 0x98, - 0x60, 0x21, 0xb5, 0x99, 0x2d, 0x9a, 0x22, 0x4f, 0xd2, 0xc7, 0xe9, 0x19, 0x7a, 0x0b, 0xdb, 0x45, - 0xcb, 0x12, 0xcd, 0x22, 0xa5, 0x2d, 0xe1, 0xd2, 0x32, 0x3d, 0xa5, 0x22, 0x2d, 0xbf, 0x59, 0x2c, - 0x3f, 0x1b, 0x62, 0x78, 0xb3, 0xe0, 0x5e, 0xec, 0xf2, 0x4e, 0xd3, 0x34, 0xef, 0x05, 0x40, 0x7e, - 0x4b, 0xb4, 0x1f, 0x0f, 0xac, 0x78, 0x17, 0x0f, 0xac, 0x72, 0xbb, 0x7e, 0x88, 0xee, 0xb6, 0x03, - 0x67, 0x90, 0x57, 0x95, 0x6a, 0xb9, 0x51, 0xf1, 0xbe, 0x97, 0x60, 0x29, 0x39, 0x41, 0x3b, 0x00, - 0x3c, 0x22, 0x34, 0x08, 0x34, 0x33, 0xc9, 0xc8, 0x5b, 0xc1, 0x35, 0x1e, 0xbd, 0x4c, 0x02, 0xb1, - 0xfb, 0x63, 0xed, 0x74, 0xe6, 0xb9, 0x75, 0x6c, 0xc6, 0x1b, 0x9d, 0xb4, 0xea, 0x8a, 0x49, 0x67, - 0xc6, 0x1a, 0x6e, 0x14, 0x1a, 0x71, 0x1e, 0xc7, 0x51, 0x17, 0xd6, 0x1f, 0x30, 0x5d, 0x15, 0xaf, - 0x05, 0x73, 0x0c, 0xf6, 0x3f, 0x6c, 0x3c, 0x64, 0xa4, 0x2a, 0x6e, 0x06, 0x73, 0x4c, 0xd3, 0xeb, - 0xbe, 0x3f, 0x08, 0x95, 0x0a, 0x05, 0xf3, 0x43, 0x25, 0xa8, 0x0c, 0x7d, 0xa5, 0xc3, 0x4e, 0xdc, - 0x0d, 0xf7, 0x23, 0x86, 0x9d, 0x39, 0x5f, 0x95, 0xe1, 0x92, 0xfb, 0x9a, 0x74, 0x7f, 0x04, 0x00, - 0x00, 0xff, 0xff, 0x8e, 0xd0, 0x70, 0xb7, 0x73, 0x06, 0x00, 0x00, +var fileDescriptor_messages_b3d89fcb5aa158f8 = []byte{ + // 708 bytes of a gzipped FileDescriptorProto + 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x54, 0x61, 0x4f, 0xf3, 0x36, + 0x10, 0x26, 0x6b, 0xe8, 0xdb, 0x5e, 0x5f, 0x8d, 0xce, 0x6c, 0x2c, 0x2d, 0x30, 0x58, 0xa4, 0x21, + 0x34, 0xb1, 0x54, 0xc0, 0x3e, 0x6c, 0xd2, 0x3e, 0x6c, 0x05, 0xa1, 0xc2, 0xf8, 0x80, 0x52, 0xa4, + 0x4d, 0x48, 0x93, 0xe7, 0x36, 0x26, 0x58, 0xb8, 0x76, 0x66, 0xbb, 0x45, 0xfb, 0xbc, 0xff, 0x33, + 0xed, 0x2f, 0x4c, 0xfb, 0x63, 0x53, 0xec, 0xa4, 0x0d, 0x2d, 0xd5, 0xfb, 0x25, 0x72, 0xee, 0x9e, + 0x7b, 0xee, 0xce, 0x77, 0x8f, 0x21, 0x4c, 0x55, 0x36, 0xc6, 0x7c, 0x84, 0x67, 0xa7, 0xbd, 0x09, + 0xd5, 0x9a, 0xa4, 0x54, 0xcf, 0x0f, 0x51, 0xa6, 0xa4, 0x91, 0x08, 0x72, 0x4c, 0xc4, 0x47, 0xd1, + 0xec, 0xb4, 0xfb, 0x45, 0x2a, 0x65, 0xca, 0x69, 0xcf, 0x7a, 0x46, 0xd3, 0xc7, 0x5e, 0x32, 0x55, + 0xc4, 0x30, 0x29, 0x1c, 0xb6, 0x7b, 0xb0, 0xec, 0x37, 0x6c, 0x42, 0xb5, 0x21, 0x93, 0xcc, 0x01, + 0xc2, 0x7f, 0x3d, 0x40, 0xb7, 0x92, 0x24, 0x7d, 0xc2, 0x89, 0x18, 0xd3, 0x98, 0xfe, 0x31, 0xa5, + 0xda, 0xa0, 0x3b, 0xd8, 0x62, 0x82, 0x19, 0x46, 0x38, 0x56, 0xce, 0x14, 0x78, 0x87, 0xde, 0x71, + 0xeb, 0xec, 0xab, 0x68, 0x91, 0x3d, 0xba, 0x76, 0x90, 0xd5, 0xf8, 0xc1, 0x46, 0xfc, 0x71, 0x11, + 0x5f, 0x32, 0xfe, 0x00, 0xef, 0xc7, 0x9c, 0x51, 0x61, 0xb0, 0x36, 0xc4, 0xe8, 0xe0, 0x23, 0x4b, + 0xf7, 0x79, 0x95, 0xee, 0xc2, 0xfa, 0x87, 0xb9, 0x7b, 0xb0, 0x11, 0xb7, 0xc6, 0x8b, 0xdf, 0xfe, + 0x2e, 0x74, 0xb8, 0x24, 0x09, 0x1e, 0xb9, 0x34, 0x65, 0x51, 0xd8, 0xfc, 0x99, 0xd1, 0xb0, 0x07, + 0x9d, 0xb5, 0x95, 0x20, 0x04, 0xbe, 0x20, 0x13, 0x6a, 0xcb, 0x6f, 0xc6, 0xf6, 0x1c, 0xfe, 0x0e, + 0xdb, 0x95, 0x5c, 0x77, 0x54, 0xdd, 0xcb, 0x67, 0x2a, 0xd0, 0x09, 0xa0, 0x57, 0x49, 0x4c, 0x6e, + 0x2d, 0x02, 0xdb, 0x7c, 0x41, 0xed, 0xd0, 0xbb, 0xd0, 0x14, 0xd3, 0x09, 0x1e, 0x13, 0xce, 0x5d, + 0x37, 0xb5, 0xb8, 0x21, 0xa6, 0x93, 0x8b, 0xfc, 0x3f, 0xfc, 0xa7, 0x06, 0xad, 0x4a, 0x0a, 0xf4, + 0x1d, 0x34, 0xe7, 0x37, 0x5f, 0xdc, 0x64, 0x37, 0x72, 0xb3, 0x89, 0xca, 0xd9, 0x44, 0xf7, 0x25, + 0x22, 0x5e, 0x80, 0xd1, 0xd7, 0xf0, 0xc9, 0x3c, 0x4d, 0x7e, 0x75, 0xca, 0xd0, 0xa4, 0x48, 0xb7, + 0x55, 0xa6, 0x1b, 0x3a, 0x73, 0xde, 0xc0, 0x02, 0xfb, 0xc8, 0x04, 0xd3, 0x4f, 0x34, 0x09, 0x6a, + 0x16, 0xdc, 0x2e, 0xc1, 0x57, 0x85, 0x1d, 0xfd, 0x06, 0xdf, 0xac, 0xa2, 0xf1, 0x0b, 0x33, 0x4f, + 0xb8, 0x98, 0xd4, 0x23, 0x61, 0x9c, 0x26, 0xd8, 0x48, 0xac, 0xa9, 0x48, 0x82, 0xba, 0x25, 0x3a, + 0x5a, 0x26, 0xfa, 0x85, 0x99, 0x27, 0xd7, 0xeb, 0x95, 0xc5, 0xdf, 0xcb, 0x21, 0x15, 0x09, 0x1a, + 0xc0, 0x97, 0x6f, 0xd0, 0x3f, 0x0b, 0xf9, 0x22, 0xb0, 0xa2, 0x63, 0xca, 0x66, 0x34, 0x09, 0xde, + 0x59, 0xca, 0xfd, 0x65, 0xca, 0x9f, 0x73, 0x54, 0x5c, 0x80, 0xd0, 0xaf, 0x10, 0xbc, 0x55, 0x64, + 0xa2, 0x64, 0x16, 0x34, 0x0e, 0x6b, 0xc7, 0xad, 0xb3, 0x83, 0x35, 0x6b, 0x54, 0x8e, 0x36, 0xfe, + 0x6c, 0xbc, 0x5c, 0xf1, 0xa5, 0x92, 0xd9, 0x8d, 0xdf, 0xf0, 0xdb, 0x9b, 0x37, 0x7e, 0x63, 0xb3, + 0x5d, 0x0f, 0xff, 0xf3, 0x60, 0xfb, 0xd5, 0xfe, 0xe8, 0x4c, 0x0a, 0x4d, 0xd1, 0x10, 0xda, 0x0b, + 0x29, 0x38, 0x5b, 0x31, 0xc1, 0xa3, 0x0f, 0x69, 0xc1, 0xa1, 0x07, 0x1b, 0xf1, 0xd6, 0x5c, 0x0c, + 0x05, 0xe9, 0xf7, 0xd0, 0xd2, 0x54, 0xcd, 0xa8, 0xc2, 0x9c, 0x69, 0x53, 0x88, 0x61, 0xa7, 0xca, + 0x37, 0xb4, 0xee, 0x5b, 0x66, 0xc5, 0x04, 0x7a, 0xfe, 0xd7, 0xdf, 0x83, 0xee, 0x92, 0x14, 0x1c, + 0xa7, 0xd3, 0xc2, 0xdf, 0x1e, 0x74, 0xd7, 0x97, 0x82, 0xbe, 0x85, 0x9d, 0x6a, 0xb0, 0xc2, 0x09, + 0xe5, 0x34, 0x25, 0xa6, 0xd4, 0xc7, 0xa7, 0x95, 0x35, 0x57, 0x97, 0x85, 0x0f, 0x3d, 0xc0, 0x5e, + 0x55, 0xbb, 0x58, 0xd1, 0x4c, 0x2a, 0x83, 0x99, 0x30, 0x54, 0xcd, 0x08, 0x2f, 0xca, 0xef, 0xac, + 0x2c, 0xf4, 0x65, 0xf1, 0x18, 0xc5, 0x9d, 0x8a, 0x96, 0x63, 0x1b, 0x7c, 0x5d, 0xc4, 0x86, 0x3f, + 0x02, 0x2c, 0x5a, 0x45, 0x27, 0xf0, 0xce, 0xb5, 0xaa, 0x03, 0xcf, 0x4e, 0x16, 0xad, 0xde, 0x49, + 0x5c, 0x42, 0x6e, 0xfc, 0x46, 0xad, 0xed, 0x87, 0x7f, 0x79, 0x50, 0x77, 0x1e, 0xb4, 0x0f, 0xc0, + 0x32, 0x4c, 0x92, 0x44, 0x51, 0xad, 0x6d, 0x4b, 0xef, 0xe3, 0x26, 0xcb, 0x7e, 0x72, 0x86, 0xfc, + 0x2d, 0xc8, 0x73, 0xdb, 0x7a, 0x37, 0x63, 0x7b, 0x5e, 0x23, 0xfa, 0xda, 0x1a, 0xd1, 0x23, 0xf0, + 0xed, 0xda, 0xf9, 0x87, 0xde, 0x71, 0x23, 0xb6, 0x67, 0xb7, 0x3e, 0xfd, 0xf3, 0x87, 0xd3, 0xa2, + 0xfd, 0x54, 0x72, 0x22, 0xd2, 0x48, 0xaa, 0xb4, 0x97, 0xd7, 0x6e, 0x3f, 0x7c, 0xd4, 0x7b, 0xe3, + 0x65, 0x1f, 0xd5, 0xed, 0x55, 0x9d, 0xff, 0x1f, 0x00, 0x00, 0xff, 0xff, 0xc8, 0x88, 0xe6, 0xf4, + 0xf7, 0x05, 0x00, 0x00, } diff --git a/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.proto b/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.proto index 42d99c109..56d767e54 100644 --- a/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.proto +++ b/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/messages/messages.proto @@ -14,36 +14,12 @@ syntax = "proto3"; +import "google/protobuf/duration.proto"; +import "google/protobuf/timestamp.proto"; + package grpc.lb.v1; option go_package = "google.golang.org/grpc/grpclb/grpc_lb_v1/messages"; -message Duration { - // Signed seconds of the span of time. Must be from -315,576,000,000 - // to +315,576,000,000 inclusive. - int64 seconds = 1; - - // Signed fractions of a second at nanosecond resolution of the span - // of time. Durations less than one second are represented with a 0 - // `seconds` field and a positive or negative `nanos` field. For durations - // of one second or more, a non-zero value for the `nanos` field must be - // of the same sign as the `seconds` field. Must be from -999,999,999 - // to +999,999,999 inclusive. - int32 nanos = 2; -} - -message Timestamp { - // Represents seconds of UTC time since Unix epoch - // 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to - // 9999-12-31T23:59:59Z inclusive. - int64 seconds = 1; - - // Non-negative fractions of a second at nanosecond resolution. Negative - // second values with fractions must still have non-negative nanos values - // that count forward in time. Must be from 0 to 999,999,999 - // inclusive. - int32 nanos = 2; -} - message LoadBalanceRequest { oneof load_balance_request_type { // This message should be sent on the first request to the load balancer. @@ -56,16 +32,25 @@ message LoadBalanceRequest { } message InitialLoadBalanceRequest { - // Name of load balanced service (IE, balancer.service.com) + // Name of load balanced service (IE, balancer.service.com). Its // length should be less than 256 bytes. string name = 1; } +// Contains the number of calls finished for a particular load balance token. +message ClientStatsPerToken { + // See Server.load_balance_token. + string load_balance_token = 1; + + // The total number of RPCs that finished associated with the token. + int64 num_calls = 2; +} + // Contains client level statistics that are useful to load balancing. Each // count except the timestamp should be reset to zero after reporting the stats. message ClientStats { // The timestamp of generating the report. - Timestamp timestamp = 1; + google.protobuf.Timestamp timestamp = 1; // The total number of RPCs that started. int64 num_calls_started = 2; @@ -73,20 +58,17 @@ message ClientStats { // The total number of RPCs that finished. int64 num_calls_finished = 3; - // The total number of RPCs that were dropped by the client because of rate - // limiting. - int64 num_calls_finished_with_drop_for_rate_limiting = 4; - - // The total number of RPCs that were dropped by the client because of load - // balancing. - int64 num_calls_finished_with_drop_for_load_balancing = 5; - // The total number of RPCs that failed to reach a server except dropped RPCs. int64 num_calls_finished_with_client_failed_to_send = 6; // The total number of RPCs that finished and are known to have been received // by a server. int64 num_calls_finished_known_received = 7; + + // The list of dropped calls. + repeated ClientStatsPerToken calls_finished_with_drop = 8; + + reserved 4, 5; } message LoadBalanceResponse { @@ -111,7 +93,7 @@ message InitialLoadBalanceResponse { // This interval defines how often the client should send the client stats // to the load balancer. Stats should only be reported when the duration is // positive. - Duration client_stats_report_interval = 2; + google.protobuf.Duration client_stats_report_interval = 2; } message ServerList { @@ -125,10 +107,8 @@ message ServerList { reserved 3; } -// Contains server information. When none of the [drop_for_*] fields are true, -// use the other fields. When drop_for_rate_limiting is true, ignore all other -// fields. Use drop_for_load_balancing only when it is true and -// drop_for_rate_limiting is false. +// Contains server information. When the drop field is not true, use the other +// fields. message Server { // A resolved address for the server, serialized in network-byte-order. It may // either be an IPv4 or IPv6 address. @@ -140,16 +120,14 @@ message Server { // An opaque but printable token given to the frontend for each pick. All // frontend requests for that pick must include the token in its initial // metadata. The token is used by the backend to verify the request and to - // allow the backend to report load to the gRPC LB system. - // - // Its length is variable but less than 50 bytes. + // allow the backend to report load to the gRPC LB system. The token is also + // used in client stats for reporting dropped calls. string load_balance_token = 3; - // Indicates whether this particular request should be dropped by the client - // for rate limiting. - bool drop_for_rate_limiting = 4; + // Indicates whether this particular request should be dropped by the client. + // If the request is dropped, there will be a corresponding entry in + // ClientStats.calls_finished_with_drop. + bool drop = 4; - // Indicates whether this particular request should be dropped by the client - // for load balancing. - bool drop_for_load_balancing = 5; + reserved 5; } diff --git a/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/service/service.pb.go b/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/service/service.pb.go index 245b8a65a..3e4a97bfc 100644 --- a/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/service/service.pb.go +++ b/vendor/google.golang.org/grpc/grpclb/grpc_lb_v1/service/service.pb.go @@ -32,8 +32,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for LoadBalancer service - +// LoadBalancerClient is the client API for LoadBalancer service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type LoadBalancerClient interface { // Bidirectional rpc to get a list of servers. BalanceLoad(ctx context.Context, opts ...grpc.CallOption) (LoadBalancer_BalanceLoadClient, error) @@ -48,7 +49,7 @@ func NewLoadBalancerClient(cc *grpc.ClientConn) LoadBalancerClient { } func (c *loadBalancerClient) BalanceLoad(ctx context.Context, opts ...grpc.CallOption) (LoadBalancer_BalanceLoadClient, error) { - stream, err := grpc.NewClientStream(ctx, &_LoadBalancer_serviceDesc.Streams[0], c.cc, "/grpc.lb.v1.LoadBalancer/BalanceLoad", opts...) + stream, err := c.cc.NewStream(ctx, &_LoadBalancer_serviceDesc.Streams[0], "/grpc.lb.v1.LoadBalancer/BalanceLoad", opts...) if err != nil { return nil, err } @@ -78,8 +79,7 @@ func (x *loadBalancerBalanceLoadClient) Recv() (*messages.LoadBalanceResponse, e return m, nil } -// Server API for LoadBalancer service - +// LoadBalancerServer is the server API for LoadBalancer service. type LoadBalancerServer interface { // Bidirectional rpc to get a list of servers. BalanceLoad(LoadBalancer_BalanceLoadServer) error diff --git a/vendor/google.golang.org/grpc/grpclb/noimport.go b/vendor/google.golang.org/grpc/grpclb/noimport.go new file mode 100644 index 000000000..b1b24298d --- /dev/null +++ b/vendor/google.golang.org/grpc/grpclb/noimport.go @@ -0,0 +1,27 @@ +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +//go:generate protoc --go_out=plugins=:$GOPATH/src grpc_lb_v1/messages/messages.proto +//go:generate protoc --go_out=plugins=grpc:$GOPATH/src grpc_lb_v1/service/service.proto + +// Package grpclb is a dummy package for generating code. Look at balancer/grpclb instead. +package grpclb + +func init() { + panic("Don't import this package. For grpclb, see package google.golang.org/grpc/balancer/grpclb") +} diff --git a/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go b/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go index e5906de7d..a1fda2801 100644 --- a/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go +++ b/vendor/google.golang.org/grpc/health/grpc_health_v1/health.pb.go @@ -1,5 +1,5 @@ // Code generated by protoc-gen-go. DO NOT EDIT. -// source: grpc_health_v1/health.proto +// source: grpc/health/v1/health.proto package grpc_health_v1 // import "google.golang.org/grpc/health/grpc_health_v1" @@ -46,11 +46,11 @@ func (x HealthCheckResponse_ServingStatus) String() string { return proto.EnumName(HealthCheckResponse_ServingStatus_name, int32(x)) } func (HealthCheckResponse_ServingStatus) EnumDescriptor() ([]byte, []int) { - return fileDescriptor_health_8e5b8a3074428511, []int{1, 0} + return fileDescriptor_health_85731b6c49265086, []int{1, 0} } type HealthCheckRequest struct { - Service string `protobuf:"bytes,1,opt,name=service" json:"service,omitempty"` + Service string `protobuf:"bytes,1,opt,name=service,proto3" json:"service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -60,7 +60,7 @@ func (m *HealthCheckRequest) Reset() { *m = HealthCheckRequest{} } func (m *HealthCheckRequest) String() string { return proto.CompactTextString(m) } func (*HealthCheckRequest) ProtoMessage() {} func (*HealthCheckRequest) Descriptor() ([]byte, []int) { - return fileDescriptor_health_8e5b8a3074428511, []int{0} + return fileDescriptor_health_85731b6c49265086, []int{0} } func (m *HealthCheckRequest) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_HealthCheckRequest.Unmarshal(m, b) @@ -88,7 +88,7 @@ func (m *HealthCheckRequest) GetService() string { } type HealthCheckResponse struct { - Status HealthCheckResponse_ServingStatus `protobuf:"varint,1,opt,name=status,enum=grpc.health.v1.HealthCheckResponse_ServingStatus" json:"status,omitempty"` + Status HealthCheckResponse_ServingStatus `protobuf:"varint,1,opt,name=status,proto3,enum=grpc.health.v1.HealthCheckResponse_ServingStatus" json:"status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -98,7 +98,7 @@ func (m *HealthCheckResponse) Reset() { *m = HealthCheckResponse{} } func (m *HealthCheckResponse) String() string { return proto.CompactTextString(m) } func (*HealthCheckResponse) ProtoMessage() {} func (*HealthCheckResponse) Descriptor() ([]byte, []int) { - return fileDescriptor_health_8e5b8a3074428511, []int{1} + return fileDescriptor_health_85731b6c49265086, []int{1} } func (m *HealthCheckResponse) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_HealthCheckResponse.Unmarshal(m, b) @@ -139,8 +139,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for Health service - +// HealthClient is the client API for Health service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type HealthClient interface { Check(ctx context.Context, in *HealthCheckRequest, opts ...grpc.CallOption) (*HealthCheckResponse, error) } @@ -155,15 +156,14 @@ func NewHealthClient(cc *grpc.ClientConn) HealthClient { func (c *healthClient) Check(ctx context.Context, in *HealthCheckRequest, opts ...grpc.CallOption) (*HealthCheckResponse, error) { out := new(HealthCheckResponse) - err := grpc.Invoke(ctx, "/grpc.health.v1.Health/Check", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.health.v1.Health/Check", in, out, opts...) if err != nil { return nil, err } return out, nil } -// Server API for Health service - +// HealthServer is the server API for Health service. type HealthServer interface { Check(context.Context, *HealthCheckRequest) (*HealthCheckResponse, error) } @@ -200,28 +200,28 @@ var _Health_serviceDesc = grpc.ServiceDesc{ }, }, Streams: []grpc.StreamDesc{}, - Metadata: "grpc_health_v1/health.proto", + Metadata: "grpc/health/v1/health.proto", } -func init() { proto.RegisterFile("grpc_health_v1/health.proto", fileDescriptor_health_8e5b8a3074428511) } +func init() { proto.RegisterFile("grpc/health/v1/health.proto", fileDescriptor_health_85731b6c49265086) } -var fileDescriptor_health_8e5b8a3074428511 = []byte{ - // 269 bytes of a gzipped FileDescriptorProto +var fileDescriptor_health_85731b6c49265086 = []byte{ + // 271 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xe2, 0x92, 0x4e, 0x2f, 0x2a, 0x48, - 0x8e, 0xcf, 0x48, 0x4d, 0xcc, 0x29, 0xc9, 0x88, 0x2f, 0x33, 0xd4, 0x87, 0xb0, 0xf4, 0x0a, 0x8a, - 0xf2, 0x4b, 0xf2, 0x85, 0xf8, 0x40, 0x92, 0x7a, 0x50, 0xa1, 0x32, 0x43, 0x25, 0x3d, 0x2e, 0x21, - 0x0f, 0x30, 0xc7, 0x39, 0x23, 0x35, 0x39, 0x3b, 0x28, 0xb5, 0xb0, 0x34, 0xb5, 0xb8, 0x44, 0x48, - 0x82, 0x8b, 0xbd, 0x38, 0xb5, 0xa8, 0x2c, 0x33, 0x39, 0x55, 0x82, 0x51, 0x81, 0x51, 0x83, 0x33, - 0x08, 0xc6, 0x55, 0x9a, 0xc3, 0xc8, 0x25, 0x8c, 0xa2, 0xa1, 0xb8, 0x20, 0x3f, 0xaf, 0x38, 0x55, - 0xc8, 0x93, 0x8b, 0xad, 0xb8, 0x24, 0xb1, 0xa4, 0xb4, 0x18, 0xac, 0x81, 0xcf, 0xc8, 0x50, 0x0f, - 0xd5, 0x22, 0x3d, 0x2c, 0x9a, 0xf4, 0x82, 0x41, 0x86, 0xe6, 0xa5, 0x07, 0x83, 0x35, 0x06, 0x41, - 0x0d, 0x50, 0xb2, 0xe2, 0xe2, 0x45, 0x91, 0x10, 0xe2, 0xe6, 0x62, 0x0f, 0xf5, 0xf3, 0xf6, 0xf3, - 0x0f, 0xf7, 0x13, 0x60, 0x00, 0x71, 0x82, 0x5d, 0x83, 0xc2, 0x3c, 0xfd, 0xdc, 0x05, 0x18, 0x85, - 0xf8, 0xb9, 0xb8, 0xfd, 0xfc, 0x43, 0xe2, 0x61, 0x02, 0x4c, 0x46, 0x51, 0x5c, 0x6c, 0x10, 0x8b, - 0x84, 0x02, 0xb8, 0x58, 0xc1, 0x96, 0x09, 0x29, 0xe1, 0x75, 0x09, 0xd8, 0xbf, 0x52, 0xca, 0x44, - 0xb8, 0xd6, 0x29, 0x91, 0x4b, 0x30, 0x33, 0x1f, 0x4d, 0xa1, 0x13, 0x37, 0x44, 0x65, 0x00, 0x28, - 0x70, 0x03, 0x18, 0xa3, 0x74, 0xd2, 0xf3, 0xf3, 0xd3, 0x73, 0x52, 0xf5, 0xd2, 0xf3, 0x73, 0x12, - 0xf3, 0xd2, 0xf5, 0xf2, 0x8b, 0xd2, 0xf5, 0x41, 0x1a, 0xa0, 0x71, 0xa0, 0x8f, 0x1a, 0x33, 0xab, - 0x98, 0xf8, 0xdc, 0x41, 0xa6, 0x41, 0x8c, 0xd0, 0x0b, 0x33, 0x4c, 0x62, 0x03, 0x47, 0x92, 0x31, - 0x20, 0x00, 0x00, 0xff, 0xff, 0xb7, 0x70, 0xc4, 0xa7, 0xc3, 0x01, 0x00, 0x00, + 0xd6, 0xcf, 0x48, 0x4d, 0xcc, 0x29, 0xc9, 0xd0, 0x2f, 0x33, 0x84, 0xb2, 0xf4, 0x0a, 0x8a, 0xf2, + 0x4b, 0xf2, 0x85, 0xf8, 0x40, 0x92, 0x7a, 0x50, 0xa1, 0x32, 0x43, 0x25, 0x3d, 0x2e, 0x21, 0x0f, + 0x30, 0xc7, 0x39, 0x23, 0x35, 0x39, 0x3b, 0x28, 0xb5, 0xb0, 0x34, 0xb5, 0xb8, 0x44, 0x48, 0x82, + 0x8b, 0xbd, 0x38, 0xb5, 0xa8, 0x2c, 0x33, 0x39, 0x55, 0x82, 0x51, 0x81, 0x51, 0x83, 0x33, 0x08, + 0xc6, 0x55, 0x9a, 0xc3, 0xc8, 0x25, 0x8c, 0xa2, 0xa1, 0xb8, 0x20, 0x3f, 0xaf, 0x38, 0x55, 0xc8, + 0x93, 0x8b, 0xad, 0xb8, 0x24, 0xb1, 0xa4, 0xb4, 0x18, 0xac, 0x81, 0xcf, 0xc8, 0x50, 0x0f, 0xd5, + 0x22, 0x3d, 0x2c, 0x9a, 0xf4, 0x82, 0x41, 0x86, 0xe6, 0xa5, 0x07, 0x83, 0x35, 0x06, 0x41, 0x0d, + 0x50, 0xb2, 0xe2, 0xe2, 0x45, 0x91, 0x10, 0xe2, 0xe6, 0x62, 0x0f, 0xf5, 0xf3, 0xf6, 0xf3, 0x0f, + 0xf7, 0x13, 0x60, 0x00, 0x71, 0x82, 0x5d, 0x83, 0xc2, 0x3c, 0xfd, 0xdc, 0x05, 0x18, 0x85, 0xf8, + 0xb9, 0xb8, 0xfd, 0xfc, 0x43, 0xe2, 0x61, 0x02, 0x4c, 0x46, 0x51, 0x5c, 0x6c, 0x10, 0x8b, 0x84, + 0x02, 0xb8, 0x58, 0xc1, 0x96, 0x09, 0x29, 0xe1, 0x75, 0x09, 0xd8, 0xbf, 0x52, 0xca, 0x44, 0xb8, + 0xd6, 0x29, 0x91, 0x4b, 0x30, 0x33, 0x1f, 0x4d, 0xa1, 0x13, 0x37, 0x44, 0x65, 0x00, 0x28, 0x70, + 0x03, 0x18, 0xa3, 0x74, 0xd2, 0xf3, 0xf3, 0xd3, 0x73, 0x52, 0xf5, 0xd2, 0xf3, 0x73, 0x12, 0xf3, + 0xd2, 0xf5, 0xf2, 0x8b, 0xd2, 0xf5, 0x91, 0x63, 0x03, 0xc4, 0x8e, 0x87, 0xb0, 0xe3, 0xcb, 0x0c, + 0x57, 0x31, 0xf1, 0xb9, 0x83, 0x4c, 0x83, 0x18, 0xa1, 0x17, 0x66, 0x98, 0xc4, 0x06, 0x8e, 0x24, + 0x63, 0x40, 0x00, 0x00, 0x00, 0xff, 0xff, 0xec, 0x66, 0x81, 0xcb, 0xc3, 0x01, 0x00, 0x00, } diff --git a/vendor/google.golang.org/grpc/health/grpc_health_v1/health.proto b/vendor/google.golang.org/grpc/health/grpc_health_v1/health.proto deleted file mode 100644 index bcc02f8ac..000000000 --- a/vendor/google.golang.org/grpc/health/grpc_health_v1/health.proto +++ /dev/null @@ -1,44 +0,0 @@ -// Copyright 2015, gRPC Authors -// All rights reserved. -// -// Licensed under the Apache License, Version 2.0 (the "License"); -// you may not use this file except in compliance with the License. -// You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, software -// distributed under the License is distributed on an "AS IS" BASIS, -// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -// See the License for the specific language governing permissions and -// limitations under the License. - -// The canonical version of this proto can be found at -// https://github.com/grpc/grpc-proto/blob/master/grpc/health/v1/health.proto - -syntax = "proto3"; - -package grpc.health.v1; - -option csharp_namespace = "Grpc.Health.V1"; -option go_package = "google.golang.org/grpc/health/grpc_health_v1"; -option java_multiple_files = true; -option java_outer_classname = "HealthProto"; -option java_package = "io.grpc.health.v1"; - -message HealthCheckRequest { - string service = 1; -} - -message HealthCheckResponse { - enum ServingStatus { - UNKNOWN = 0; - SERVING = 1; - NOT_SERVING = 2; - } - ServingStatus status = 1; -} - -service Health { - rpc Check(HealthCheckRequest) returns (HealthCheckResponse); -} diff --git a/vendor/google.golang.org/grpc/health/health.go b/vendor/google.golang.org/grpc/health/health.go index de7f9ba79..c2588867e 100644 --- a/vendor/google.golang.org/grpc/health/health.go +++ b/vendor/google.golang.org/grpc/health/health.go @@ -16,7 +16,7 @@ * */ -//go:generate protoc --go_out=plugins=grpc,paths=source_relative:. grpc_health_v1/health.proto +//go:generate ./regenerate.sh // Package health provides some utility functions to health-check a server. The implementation // is based on protobuf. Users need to write their own implementations if other IDLs are used. diff --git a/vendor/google.golang.org/grpc/health/regenerate.sh b/vendor/google.golang.org/grpc/health/regenerate.sh new file mode 100755 index 000000000..b11eccb29 --- /dev/null +++ b/vendor/google.golang.org/grpc/health/regenerate.sh @@ -0,0 +1,33 @@ +#!/bin/bash +# Copyright 2018 gRPC authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +set -eux -o pipefail + +TMP=$(mktemp -d) + +function finish { + rm -rf "$TMP" +} +trap finish EXIT + +pushd "$TMP" +mkdir -p grpc/health/v1 +curl https://raw.githubusercontent.com/grpc/grpc-proto/master/grpc/health/v1/health.proto > grpc/health/v1/health.proto + +protoc --go_out=plugins=grpc,paths=source_relative:. -I. grpc/health/v1/*.proto +popd +rm -f grpc_health_v1/*.pb.go +cp "$TMP"/grpc/health/v1/*.pb.go grpc_health_v1/ + diff --git a/vendor/google.golang.org/grpc/internal/backoff/backoff.go b/vendor/google.golang.org/grpc/internal/backoff/backoff.go new file mode 100644 index 000000000..1bd0cce5a --- /dev/null +++ b/vendor/google.golang.org/grpc/internal/backoff/backoff.go @@ -0,0 +1,78 @@ +/* + * + * Copyright 2017 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +// Package backoff implement the backoff strategy for gRPC. +// +// This is kept in internal until the gRPC project decides whether or not to +// allow alternative backoff strategies. +package backoff + +import ( + "time" + + "google.golang.org/grpc/internal/grpcrand" +) + +// Strategy defines the methodology for backing off after a grpc connection +// failure. +// +type Strategy interface { + // Backoff returns the amount of time to wait before the next retry given + // the number of consecutive failures. + Backoff(retries int) time.Duration +} + +const ( + // baseDelay is the amount of time to wait before retrying after the first + // failure. + baseDelay = 1.0 * time.Second + // factor is applied to the backoff after each retry. + factor = 1.6 + // jitter provides a range to randomize backoff delays. + jitter = 0.2 +) + +// Exponential implements exponential backoff algorithm as defined in +// https://github.com/grpc/grpc/blob/master/doc/connection-backoff.md. +type Exponential struct { + // MaxDelay is the upper bound of backoff delay. + MaxDelay time.Duration +} + +// Backoff returns the amount of time to wait before the next retry given the +// number of retries. +func (bc Exponential) Backoff(retries int) time.Duration { + if retries == 0 { + return baseDelay + } + backoff, max := float64(baseDelay), float64(bc.MaxDelay) + for backoff < max && retries > 0 { + backoff *= factor + retries-- + } + if backoff > max { + backoff = max + } + // Randomize backoff delays so that if a cluster of requests start at + // the same time, they won't operate in lockstep. + backoff *= 1 + jitter*(grpcrand.Float64()*2-1) + if backoff < 0 { + return 0 + } + return time.Duration(backoff) +} diff --git a/vendor/google.golang.org/grpc/channelz/funcs.go b/vendor/google.golang.org/grpc/internal/channelz/funcs.go similarity index 100% rename from vendor/google.golang.org/grpc/channelz/funcs.go rename to vendor/google.golang.org/grpc/internal/channelz/funcs.go diff --git a/vendor/google.golang.org/grpc/channelz/types.go b/vendor/google.golang.org/grpc/internal/channelz/types.go similarity index 100% rename from vendor/google.golang.org/grpc/channelz/types.go rename to vendor/google.golang.org/grpc/internal/channelz/types.go diff --git a/vendor/google.golang.org/grpc/internal/grpcrand/grpcrand.go b/vendor/google.golang.org/grpc/internal/grpcrand/grpcrand.go new file mode 100644 index 000000000..200b115ca --- /dev/null +++ b/vendor/google.golang.org/grpc/internal/grpcrand/grpcrand.go @@ -0,0 +1,56 @@ +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +// Package grpcrand implements math/rand functions in a concurrent-safe way +// with a global random source, independent of math/rand's global source. +package grpcrand + +import ( + "math/rand" + "sync" + "time" +) + +var ( + r = rand.New(rand.NewSource(time.Now().UnixNano())) + mu sync.Mutex +) + +// Int63n implements rand.Int63n on the grpcrand global source. +func Int63n(n int64) int64 { + mu.Lock() + res := r.Int63n(n) + mu.Unlock() + return res +} + +// Intn implements rand.Intn on the grpcrand global source. +func Intn(n int) int { + mu.Lock() + res := r.Intn(n) + mu.Unlock() + return res +} + +// Float64 implements rand.Float64 on the grpcrand global source. +func Float64() float64 { + mu.Lock() + res := r.Float64() + mu.Unlock() + return res +} diff --git a/vendor/google.golang.org/grpc/internal/internal.go b/vendor/google.golang.org/grpc/internal/internal.go index 53f177520..cd34267f7 100644 --- a/vendor/google.golang.org/grpc/internal/internal.go +++ b/vendor/google.golang.org/grpc/internal/internal.go @@ -15,13 +15,22 @@ * */ -// Package internal contains gRPC-internal code for testing, to avoid polluting -// the godoc of the top-level grpc package. +// Package internal contains gRPC-internal code, to avoid polluting +// the godoc of the top-level grpc package. It must not import any grpc +// symbols to avoid circular dependencies. package internal -// TestingUseHandlerImpl enables the http.Handler-based server implementation. -// It must be called before Serve and requires TLS credentials. -// -// The provided grpcServer must be of type *grpc.Server. It is untyped -// for circular dependency reasons. -var TestingUseHandlerImpl func(grpcServer interface{}) +var ( + + // TestingUseHandlerImpl enables the http.Handler-based server implementation. + // It must be called before Serve and requires TLS credentials. + // + // The provided grpcServer must be of type *grpc.Server. It is untyped + // for circular dependency reasons. + TestingUseHandlerImpl func(grpcServer interface{}) + + // WithContextDialer is exported by clientconn.go + WithContextDialer interface{} // func(context.Context, string) (net.Conn, error) grpc.DialOption + // WithResolverBuilder is exported by clientconn.go + WithResolverBuilder interface{} // func (resolver.Builder) grpc.DialOption +) diff --git a/vendor/google.golang.org/grpc/test/leakcheck/leakcheck.go b/vendor/google.golang.org/grpc/internal/leakcheck/leakcheck.go similarity index 100% rename from vendor/google.golang.org/grpc/test/leakcheck/leakcheck.go rename to vendor/google.golang.org/grpc/internal/leakcheck/leakcheck.go diff --git a/vendor/google.golang.org/grpc/test/leakcheck/leakcheck_test.go b/vendor/google.golang.org/grpc/internal/leakcheck/leakcheck_test.go similarity index 100% rename from vendor/google.golang.org/grpc/test/leakcheck/leakcheck_test.go rename to vendor/google.golang.org/grpc/internal/leakcheck/leakcheck_test.go diff --git a/vendor/google.golang.org/grpc/interop/client/client.go b/vendor/google.golang.org/grpc/interop/client/client.go index 1e0a23911..c8acae94b 100644 --- a/vendor/google.golang.org/grpc/interop/client/client.go +++ b/vendor/google.golang.org/grpc/interop/client/client.go @@ -24,6 +24,7 @@ import ( "strconv" "google.golang.org/grpc" + _ "google.golang.org/grpc/balancer/grpclb" "google.golang.org/grpc/credentials" "google.golang.org/grpc/credentials/alts" "google.golang.org/grpc/credentials/oauth" diff --git a/vendor/google.golang.org/grpc/interop/grpc_testing/test.pb.go b/vendor/google.golang.org/grpc/interop/grpc_testing/test.pb.go index 264bb064f..bb5e2eeb3 100644 --- a/vendor/google.golang.org/grpc/interop/grpc_testing/test.pb.go +++ b/vendor/google.golang.org/grpc/interop/grpc_testing/test.pb.go @@ -86,7 +86,7 @@ var xxx_messageInfo_Empty proto.InternalMessageInfo // A block of data, to simply increase gRPC message size. type Payload struct { // The type of data in body. - Type PayloadType `protobuf:"varint,1,opt,name=type,enum=grpc.testing.PayloadType" json:"type,omitempty"` + Type PayloadType `protobuf:"varint,1,opt,name=type,proto3,enum=grpc.testing.PayloadType" json:"type,omitempty"` // Primary contents of payload. Body []byte `protobuf:"bytes,2,opt,name=body,proto3" json:"body,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -135,8 +135,8 @@ func (m *Payload) GetBody() []byte { // A protobuf representation for grpc status. This is used by test // clients to specify a status that the server should attempt to return. type EchoStatus struct { - Code int32 `protobuf:"varint,1,opt,name=code" json:"code,omitempty"` - Message string `protobuf:"bytes,2,opt,name=message" json:"message,omitempty"` + Code int32 `protobuf:"varint,1,opt,name=code,proto3" json:"code,omitempty"` + Message string `protobuf:"bytes,2,opt,name=message,proto3" json:"message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -184,18 +184,18 @@ func (m *EchoStatus) GetMessage() string { type SimpleRequest struct { // Desired payload type in the response from the server. // If response_type is RANDOM, server randomly chooses one from other formats. - ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` + ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,proto3,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` // Desired payload size in the response from the server. // If response_type is COMPRESSABLE, this denotes the size before compression. - ResponseSize int32 `protobuf:"varint,2,opt,name=response_size,json=responseSize" json:"response_size,omitempty"` + ResponseSize int32 `protobuf:"varint,2,opt,name=response_size,json=responseSize,proto3" json:"response_size,omitempty"` // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` // Whether SimpleResponse should include username. - FillUsername bool `protobuf:"varint,4,opt,name=fill_username,json=fillUsername" json:"fill_username,omitempty"` + FillUsername bool `protobuf:"varint,4,opt,name=fill_username,json=fillUsername,proto3" json:"fill_username,omitempty"` // Whether SimpleResponse should include OAuth scope. - FillOauthScope bool `protobuf:"varint,5,opt,name=fill_oauth_scope,json=fillOauthScope" json:"fill_oauth_scope,omitempty"` + FillOauthScope bool `protobuf:"varint,5,opt,name=fill_oauth_scope,json=fillOauthScope,proto3" json:"fill_oauth_scope,omitempty"` // Whether server should return a given status - ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus" json:"response_status,omitempty"` + ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus,proto3" json:"response_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -270,12 +270,12 @@ func (m *SimpleRequest) GetResponseStatus() *EchoStatus { // Unary response, as configured by the request. type SimpleResponse struct { // Payload to increase message size. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` // The user the request came from, for verifying authentication was // successful when the client expected it. - Username string `protobuf:"bytes,2,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,2,opt,name=username,proto3" json:"username,omitempty"` // OAuth scope. - OauthScope string `protobuf:"bytes,3,opt,name=oauth_scope,json=oauthScope" json:"oauth_scope,omitempty"` + OauthScope string `protobuf:"bytes,3,opt,name=oauth_scope,json=oauthScope,proto3" json:"oauth_scope,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -329,7 +329,7 @@ func (m *SimpleResponse) GetOauthScope() string { // Client-streaming request. type StreamingInputCallRequest struct { // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -369,7 +369,7 @@ func (m *StreamingInputCallRequest) GetPayload() *Payload { // Client-streaming response. type StreamingInputCallResponse struct { // Aggregated size of payloads received from the client. - AggregatedPayloadSize int32 `protobuf:"varint,1,opt,name=aggregated_payload_size,json=aggregatedPayloadSize" json:"aggregated_payload_size,omitempty"` + AggregatedPayloadSize int32 `protobuf:"varint,1,opt,name=aggregated_payload_size,json=aggregatedPayloadSize,proto3" json:"aggregated_payload_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -410,10 +410,10 @@ func (m *StreamingInputCallResponse) GetAggregatedPayloadSize() int32 { type ResponseParameters struct { // Desired payload sizes in responses from the server. // If response_type is COMPRESSABLE, this denotes the size before compression. - Size int32 `protobuf:"varint,1,opt,name=size" json:"size,omitempty"` + Size int32 `protobuf:"varint,1,opt,name=size,proto3" json:"size,omitempty"` // Desired interval between consecutive responses in the response stream in // microseconds. - IntervalUs int32 `protobuf:"varint,2,opt,name=interval_us,json=intervalUs" json:"interval_us,omitempty"` + IntervalUs int32 `protobuf:"varint,2,opt,name=interval_us,json=intervalUs,proto3" json:"interval_us,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -463,13 +463,13 @@ type StreamingOutputCallRequest struct { // If response_type is RANDOM, the payload from each response in the stream // might be of different types. This is to simulate a mixed type of payload // stream. - ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` + ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,proto3,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` // Configuration for each expected response message. - ResponseParameters []*ResponseParameters `protobuf:"bytes,2,rep,name=response_parameters,json=responseParameters" json:"response_parameters,omitempty"` + ResponseParameters []*ResponseParameters `protobuf:"bytes,2,rep,name=response_parameters,json=responseParameters,proto3" json:"response_parameters,omitempty"` // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` // Whether server should return a given status - ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus" json:"response_status,omitempty"` + ResponseStatus *EchoStatus `protobuf:"bytes,7,opt,name=response_status,json=responseStatus,proto3" json:"response_status,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -530,7 +530,7 @@ func (m *StreamingOutputCallRequest) GetResponseStatus() *EchoStatus { // Server-streaming response, as configured by the request and parameters. type StreamingOutputCallResponse struct { // Payload to increase response size. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -589,8 +589,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for TestService service - +// TestServiceClient is the client API for TestService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type TestServiceClient interface { // One empty request followed by one empty response. EmptyCall(ctx context.Context, in *Empty, opts ...grpc.CallOption) (*Empty, error) @@ -624,7 +625,7 @@ func NewTestServiceClient(cc *grpc.ClientConn) TestServiceClient { func (c *testServiceClient) EmptyCall(ctx context.Context, in *Empty, opts ...grpc.CallOption) (*Empty, error) { out := new(Empty) - err := grpc.Invoke(ctx, "/grpc.testing.TestService/EmptyCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.TestService/EmptyCall", in, out, opts...) if err != nil { return nil, err } @@ -633,7 +634,7 @@ func (c *testServiceClient) EmptyCall(ctx context.Context, in *Empty, opts ...gr func (c *testServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, opts ...grpc.CallOption) (*SimpleResponse, error) { out := new(SimpleResponse) - err := grpc.Invoke(ctx, "/grpc.testing.TestService/UnaryCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.TestService/UnaryCall", in, out, opts...) if err != nil { return nil, err } @@ -641,7 +642,7 @@ func (c *testServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, op } func (c *testServiceClient) StreamingOutputCall(ctx context.Context, in *StreamingOutputCallRequest, opts ...grpc.CallOption) (TestService_StreamingOutputCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[0], c.cc, "/grpc.testing.TestService/StreamingOutputCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[0], "/grpc.testing.TestService/StreamingOutputCall", opts...) if err != nil { return nil, err } @@ -673,7 +674,7 @@ func (x *testServiceStreamingOutputCallClient) Recv() (*StreamingOutputCallRespo } func (c *testServiceClient) StreamingInputCall(ctx context.Context, opts ...grpc.CallOption) (TestService_StreamingInputCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[1], c.cc, "/grpc.testing.TestService/StreamingInputCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[1], "/grpc.testing.TestService/StreamingInputCall", opts...) if err != nil { return nil, err } @@ -707,7 +708,7 @@ func (x *testServiceStreamingInputCallClient) CloseAndRecv() (*StreamingInputCal } func (c *testServiceClient) FullDuplexCall(ctx context.Context, opts ...grpc.CallOption) (TestService_FullDuplexCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[2], c.cc, "/grpc.testing.TestService/FullDuplexCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[2], "/grpc.testing.TestService/FullDuplexCall", opts...) if err != nil { return nil, err } @@ -738,7 +739,7 @@ func (x *testServiceFullDuplexCallClient) Recv() (*StreamingOutputCallResponse, } func (c *testServiceClient) HalfDuplexCall(ctx context.Context, opts ...grpc.CallOption) (TestService_HalfDuplexCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[3], c.cc, "/grpc.testing.TestService/HalfDuplexCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[3], "/grpc.testing.TestService/HalfDuplexCall", opts...) if err != nil { return nil, err } @@ -768,8 +769,7 @@ func (x *testServiceHalfDuplexCallClient) Recv() (*StreamingOutputCallResponse, return m, nil } -// Server API for TestService service - +// TestServiceServer is the server API for TestService service. type TestServiceServer interface { // One empty request followed by one empty response. EmptyCall(context.Context, *Empty) (*Empty, error) @@ -972,8 +972,9 @@ var _TestService_serviceDesc = grpc.ServiceDesc{ Metadata: "grpc_testing/test.proto", } -// Client API for UnimplementedService service - +// UnimplementedServiceClient is the client API for UnimplementedService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type UnimplementedServiceClient interface { // A call that no server should implement UnimplementedCall(ctx context.Context, in *Empty, opts ...grpc.CallOption) (*Empty, error) @@ -989,15 +990,14 @@ func NewUnimplementedServiceClient(cc *grpc.ClientConn) UnimplementedServiceClie func (c *unimplementedServiceClient) UnimplementedCall(ctx context.Context, in *Empty, opts ...grpc.CallOption) (*Empty, error) { out := new(Empty) - err := grpc.Invoke(ctx, "/grpc.testing.UnimplementedService/UnimplementedCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.UnimplementedService/UnimplementedCall", in, out, opts...) if err != nil { return nil, err } return out, nil } -// Server API for UnimplementedService service - +// UnimplementedServiceServer is the server API for UnimplementedService service. type UnimplementedServiceServer interface { // A call that no server should implement UnimplementedCall(context.Context, *Empty) (*Empty, error) diff --git a/vendor/google.golang.org/grpc/picker_wrapper.go b/vendor/google.golang.org/grpc/picker_wrapper.go index 0a984e6c8..019e65800 100644 --- a/vendor/google.golang.org/grpc/picker_wrapper.go +++ b/vendor/google.golang.org/grpc/picker_wrapper.go @@ -25,9 +25,9 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/balancer" - "google.golang.org/grpc/channelz" "google.golang.org/grpc/codes" "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/metadata" "google.golang.org/grpc/resolver" "google.golang.org/grpc/status" @@ -233,6 +233,8 @@ func (bp *pickerWrapper) close() { close(bp.blockingCh) } +const stickinessKeyCountLimit = 1000 + type stickyStoreEntry struct { acw *acBalancerWrapper addr resolver.Address @@ -243,12 +245,12 @@ type stickyStore struct { // curMDKey is check before every get/put to avoid races. The operation will // abort immediately when the given mdKey is different from the curMDKey. curMDKey string - store map[string]*stickyStoreEntry + store *linkedMap } func newStickyStore() *stickyStore { return &stickyStore{ - store: make(map[string]*stickyStoreEntry), + store: newLinkedMap(), } } @@ -256,7 +258,7 @@ func newStickyStore() *stickyStore { func (ss *stickyStore) reset(newMDKey string) { ss.mu.Lock() ss.curMDKey = newMDKey - ss.store = make(map[string]*stickyStoreEntry) + ss.store.clear() ss.mu.Unlock() } @@ -269,9 +271,12 @@ func (ss *stickyStore) put(mdKey, stickyKey string, acw *acBalancerWrapper) { return } // TODO(stickiness): limit the total number of entries. - ss.store[stickyKey] = &stickyStoreEntry{ + ss.store.put(stickyKey, &stickyStoreEntry{ acw: acw, addr: acw.getAddrConn().getCurAddr(), + }) + if ss.store.len() > stickinessKeyCountLimit { + ss.store.removeOldest() } } @@ -283,18 +288,18 @@ func (ss *stickyStore) get(mdKey, stickyKey string) (transport.ClientTransport, if mdKey != ss.curMDKey { return nil, false } - entry, ok := ss.store[stickyKey] + entry, ok := ss.store.get(stickyKey) if !ok { return nil, false } ac := entry.acw.getAddrConn() if ac.getCurAddr() != entry.addr { - delete(ss.store, stickyKey) + ss.store.remove(stickyKey) return nil, false } t, ok := ac.getReadyTransport() if !ok { - delete(ss.store, stickyKey) + ss.store.remove(stickyKey) return nil, false } return t, true diff --git a/vendor/google.golang.org/grpc/picker_wrapper_test.go b/vendor/google.golang.org/grpc/picker_wrapper_test.go index 37dffa9e9..d455286d3 100644 --- a/vendor/google.golang.org/grpc/picker_wrapper_test.go +++ b/vendor/google.golang.org/grpc/picker_wrapper_test.go @@ -28,7 +28,7 @@ import ( "google.golang.org/grpc/balancer" "google.golang.org/grpc/connectivity" _ "google.golang.org/grpc/grpclog/glogger" - "google.golang.org/grpc/test/leakcheck" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/transport" ) diff --git a/vendor/google.golang.org/grpc/pickfirst_test.go b/vendor/google.golang.org/grpc/pickfirst_test.go index 9e1fd03df..7f8169f19 100644 --- a/vendor/google.golang.org/grpc/pickfirst_test.go +++ b/vendor/google.golang.org/grpc/pickfirst_test.go @@ -26,10 +26,10 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/codes" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" "google.golang.org/grpc/status" - "google.golang.org/grpc/test/leakcheck" ) func errorDesc(err error) string { diff --git a/vendor/google.golang.org/grpc/proxy_test.go b/vendor/google.golang.org/grpc/proxy_test.go index 8466676c1..7183ba342 100644 --- a/vendor/google.golang.org/grpc/proxy_test.go +++ b/vendor/google.golang.org/grpc/proxy_test.go @@ -30,7 +30,7 @@ import ( "time" "golang.org/x/net/context" - "google.golang.org/grpc/test/leakcheck" + "google.golang.org/grpc/internal/leakcheck" ) const ( diff --git a/vendor/google.golang.org/grpc/reflection/grpc_reflection_v1alpha/reflection.pb.go b/vendor/google.golang.org/grpc/reflection/grpc_reflection_v1alpha/reflection.pb.go index ab5b18893..4bc5404ae 100644 --- a/vendor/google.golang.org/grpc/reflection/grpc_reflection_v1alpha/reflection.pb.go +++ b/vendor/google.golang.org/grpc/reflection/grpc_reflection_v1alpha/reflection.pb.go @@ -25,7 +25,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // The message sent by the client when calling ServerReflectionInfo method. type ServerReflectionRequest struct { - Host string `protobuf:"bytes,1,opt,name=host" json:"host,omitempty"` + Host string `protobuf:"bytes,1,opt,name=host,proto3" json:"host,omitempty"` // To use reflection service, the client should set one of the following // fields in message_request. The server distinguishes requests by their // defined field and then handles them using corresponding methods. @@ -71,19 +71,19 @@ type isServerReflectionRequest_MessageRequest interface { } type ServerReflectionRequest_FileByFilename struct { - FileByFilename string `protobuf:"bytes,3,opt,name=file_by_filename,json=fileByFilename,oneof"` + FileByFilename string `protobuf:"bytes,3,opt,name=file_by_filename,json=fileByFilename,proto3,oneof"` } type ServerReflectionRequest_FileContainingSymbol struct { - FileContainingSymbol string `protobuf:"bytes,4,opt,name=file_containing_symbol,json=fileContainingSymbol,oneof"` + FileContainingSymbol string `protobuf:"bytes,4,opt,name=file_containing_symbol,json=fileContainingSymbol,proto3,oneof"` } type ServerReflectionRequest_FileContainingExtension struct { - FileContainingExtension *ExtensionRequest `protobuf:"bytes,5,opt,name=file_containing_extension,json=fileContainingExtension,oneof"` + FileContainingExtension *ExtensionRequest `protobuf:"bytes,5,opt,name=file_containing_extension,json=fileContainingExtension,proto3,oneof"` } type ServerReflectionRequest_AllExtensionNumbersOfType struct { - AllExtensionNumbersOfType string `protobuf:"bytes,6,opt,name=all_extension_numbers_of_type,json=allExtensionNumbersOfType,oneof"` + AllExtensionNumbersOfType string `protobuf:"bytes,6,opt,name=all_extension_numbers_of_type,json=allExtensionNumbersOfType,proto3,oneof"` } type ServerReflectionRequest_ListServices struct { - ListServices string `protobuf:"bytes,7,opt,name=list_services,json=listServices,oneof"` + ListServices string `protobuf:"bytes,7,opt,name=list_services,json=listServices,proto3,oneof"` } func (*ServerReflectionRequest_FileByFilename) isServerReflectionRequest_MessageRequest() {} @@ -260,8 +260,8 @@ func _ServerReflectionRequest_OneofSizer(msg proto.Message) (n int) { // file_containing_extension. type ExtensionRequest struct { // Fully-qualified type name. The format should be . - ContainingType string `protobuf:"bytes,1,opt,name=containing_type,json=containingType" json:"containing_type,omitempty"` - ExtensionNumber int32 `protobuf:"varint,2,opt,name=extension_number,json=extensionNumber" json:"extension_number,omitempty"` + ContainingType string `protobuf:"bytes,1,opt,name=containing_type,json=containingType,proto3" json:"containing_type,omitempty"` + ExtensionNumber int32 `protobuf:"varint,2,opt,name=extension_number,json=extensionNumber,proto3" json:"extension_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -307,8 +307,8 @@ func (m *ExtensionRequest) GetExtensionNumber() int32 { // The message sent by the server to answer ServerReflectionInfo method. type ServerReflectionResponse struct { - ValidHost string `protobuf:"bytes,1,opt,name=valid_host,json=validHost" json:"valid_host,omitempty"` - OriginalRequest *ServerReflectionRequest `protobuf:"bytes,2,opt,name=original_request,json=originalRequest" json:"original_request,omitempty"` + ValidHost string `protobuf:"bytes,1,opt,name=valid_host,json=validHost,proto3" json:"valid_host,omitempty"` + OriginalRequest *ServerReflectionRequest `protobuf:"bytes,2,opt,name=original_request,json=originalRequest,proto3" json:"original_request,omitempty"` // The server sets one of the following fields according to the // message_request in the request. // @@ -352,16 +352,16 @@ type isServerReflectionResponse_MessageResponse interface { } type ServerReflectionResponse_FileDescriptorResponse struct { - FileDescriptorResponse *FileDescriptorResponse `protobuf:"bytes,4,opt,name=file_descriptor_response,json=fileDescriptorResponse,oneof"` + FileDescriptorResponse *FileDescriptorResponse `protobuf:"bytes,4,opt,name=file_descriptor_response,json=fileDescriptorResponse,proto3,oneof"` } type ServerReflectionResponse_AllExtensionNumbersResponse struct { - AllExtensionNumbersResponse *ExtensionNumberResponse `protobuf:"bytes,5,opt,name=all_extension_numbers_response,json=allExtensionNumbersResponse,oneof"` + AllExtensionNumbersResponse *ExtensionNumberResponse `protobuf:"bytes,5,opt,name=all_extension_numbers_response,json=allExtensionNumbersResponse,proto3,oneof"` } type ServerReflectionResponse_ListServicesResponse struct { - ListServicesResponse *ListServiceResponse `protobuf:"bytes,6,opt,name=list_services_response,json=listServicesResponse,oneof"` + ListServicesResponse *ListServiceResponse `protobuf:"bytes,6,opt,name=list_services_response,json=listServicesResponse,proto3,oneof"` } type ServerReflectionResponse_ErrorResponse struct { - ErrorResponse *ErrorResponse `protobuf:"bytes,7,opt,name=error_response,json=errorResponse,oneof"` + ErrorResponse *ErrorResponse `protobuf:"bytes,7,opt,name=error_response,json=errorResponse,proto3,oneof"` } func (*ServerReflectionResponse_FileDescriptorResponse) isServerReflectionResponse_MessageResponse() {} @@ -580,8 +580,8 @@ func (m *FileDescriptorResponse) GetFileDescriptorProto() [][]byte { type ExtensionNumberResponse struct { // Full name of the base type, including the package name. The format // is . - BaseTypeName string `protobuf:"bytes,1,opt,name=base_type_name,json=baseTypeName" json:"base_type_name,omitempty"` - ExtensionNumber []int32 `protobuf:"varint,2,rep,packed,name=extension_number,json=extensionNumber" json:"extension_number,omitempty"` + BaseTypeName string `protobuf:"bytes,1,opt,name=base_type_name,json=baseTypeName,proto3" json:"base_type_name,omitempty"` + ExtensionNumber []int32 `protobuf:"varint,2,rep,packed,name=extension_number,json=extensionNumber,proto3" json:"extension_number,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -629,7 +629,7 @@ func (m *ExtensionNumberResponse) GetExtensionNumber() []int32 { type ListServiceResponse struct { // The information of each service may be expanded in the future, so we use // ServiceResponse message to encapsulate it. - Service []*ServiceResponse `protobuf:"bytes,1,rep,name=service" json:"service,omitempty"` + Service []*ServiceResponse `protobuf:"bytes,1,rep,name=service,proto3" json:"service,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -671,7 +671,7 @@ func (m *ListServiceResponse) GetService() []*ServiceResponse { type ServiceResponse struct { // Full name of a registered service, including its package name. The format // is . - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -711,8 +711,8 @@ func (m *ServiceResponse) GetName() string { // The error code and error message sent by the server when an error occurs. type ErrorResponse struct { // This field uses the error codes defined in grpc::StatusCode. - ErrorCode int32 `protobuf:"varint,1,opt,name=error_code,json=errorCode" json:"error_code,omitempty"` - ErrorMessage string `protobuf:"bytes,2,opt,name=error_message,json=errorMessage" json:"error_message,omitempty"` + ErrorCode int32 `protobuf:"varint,1,opt,name=error_code,json=errorCode,proto3" json:"error_code,omitempty"` + ErrorMessage string `protobuf:"bytes,2,opt,name=error_message,json=errorMessage,proto3" json:"error_message,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -775,8 +775,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for ServerReflection service - +// ServerReflectionClient is the client API for ServerReflection service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type ServerReflectionClient interface { // The reflection service is structured as a bidirectional stream, ensuring // all related requests go to a single server. @@ -792,7 +793,7 @@ func NewServerReflectionClient(cc *grpc.ClientConn) ServerReflectionClient { } func (c *serverReflectionClient) ServerReflectionInfo(ctx context.Context, opts ...grpc.CallOption) (ServerReflection_ServerReflectionInfoClient, error) { - stream, err := grpc.NewClientStream(ctx, &_ServerReflection_serviceDesc.Streams[0], c.cc, "/grpc.reflection.v1alpha.ServerReflection/ServerReflectionInfo", opts...) + stream, err := c.cc.NewStream(ctx, &_ServerReflection_serviceDesc.Streams[0], "/grpc.reflection.v1alpha.ServerReflection/ServerReflectionInfo", opts...) if err != nil { return nil, err } @@ -822,8 +823,7 @@ func (x *serverReflectionServerReflectionInfoClient) Recv() (*ServerReflectionRe return m, nil } -// Server API for ServerReflection service - +// ServerReflectionServer is the server API for ServerReflection service. type ServerReflectionServer interface { // The reflection service is structured as a bidirectional stream, ensuring // all related requests go to a single server. diff --git a/vendor/google.golang.org/grpc/reflection/grpc_testing/test.pb.go b/vendor/google.golang.org/grpc/reflection/grpc_testing/test.pb.go index de08dccac..2566a49f9 100644 --- a/vendor/google.golang.org/grpc/reflection/grpc_testing/test.pb.go +++ b/vendor/google.golang.org/grpc/reflection/grpc_testing/test.pb.go @@ -24,7 +24,7 @@ var _ = math.Inf const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type SearchResponse struct { - Results []*SearchResponse_Result `protobuf:"bytes,1,rep,name=results" json:"results,omitempty"` + Results []*SearchResponse_Result `protobuf:"bytes,1,rep,name=results,proto3" json:"results,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -62,9 +62,9 @@ func (m *SearchResponse) GetResults() []*SearchResponse_Result { } type SearchResponse_Result struct { - Url string `protobuf:"bytes,1,opt,name=url" json:"url,omitempty"` - Title string `protobuf:"bytes,2,opt,name=title" json:"title,omitempty"` - Snippets []string `protobuf:"bytes,3,rep,name=snippets" json:"snippets,omitempty"` + Url string `protobuf:"bytes,1,opt,name=url,proto3" json:"url,omitempty"` + Title string `protobuf:"bytes,2,opt,name=title,proto3" json:"title,omitempty"` + Snippets []string `protobuf:"bytes,3,rep,name=snippets,proto3" json:"snippets,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -116,7 +116,7 @@ func (m *SearchResponse_Result) GetSnippets() []string { } type SearchRequest struct { - Query string `protobuf:"bytes,1,opt,name=query" json:"query,omitempty"` + Query string `protobuf:"bytes,1,opt,name=query,proto3" json:"query,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -167,8 +167,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for SearchService service - +// SearchServiceClient is the client API for SearchService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type SearchServiceClient interface { Search(ctx context.Context, in *SearchRequest, opts ...grpc.CallOption) (*SearchResponse, error) StreamingSearch(ctx context.Context, opts ...grpc.CallOption) (SearchService_StreamingSearchClient, error) @@ -184,7 +185,7 @@ func NewSearchServiceClient(cc *grpc.ClientConn) SearchServiceClient { func (c *searchServiceClient) Search(ctx context.Context, in *SearchRequest, opts ...grpc.CallOption) (*SearchResponse, error) { out := new(SearchResponse) - err := grpc.Invoke(ctx, "/grpc.testing.SearchService/Search", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.SearchService/Search", in, out, opts...) if err != nil { return nil, err } @@ -192,7 +193,7 @@ func (c *searchServiceClient) Search(ctx context.Context, in *SearchRequest, opt } func (c *searchServiceClient) StreamingSearch(ctx context.Context, opts ...grpc.CallOption) (SearchService_StreamingSearchClient, error) { - stream, err := grpc.NewClientStream(ctx, &_SearchService_serviceDesc.Streams[0], c.cc, "/grpc.testing.SearchService/StreamingSearch", opts...) + stream, err := c.cc.NewStream(ctx, &_SearchService_serviceDesc.Streams[0], "/grpc.testing.SearchService/StreamingSearch", opts...) if err != nil { return nil, err } @@ -222,8 +223,7 @@ func (x *searchServiceStreamingSearchClient) Recv() (*SearchResponse, error) { return m, nil } -// Server API for SearchService service - +// SearchServiceServer is the server API for SearchService service. type SearchServiceServer interface { Search(context.Context, *SearchRequest) (*SearchResponse, error) StreamingSearch(SearchService_StreamingSearchServer) error diff --git a/vendor/google.golang.org/grpc/resolver/dns/dns_resolver.go b/vendor/google.golang.org/grpc/resolver/dns/dns_resolver.go index c1cabfc99..048fde67d 100644 --- a/vendor/google.golang.org/grpc/resolver/dns/dns_resolver.go +++ b/vendor/google.golang.org/grpc/resolver/dns/dns_resolver.go @@ -24,7 +24,6 @@ import ( "encoding/json" "errors" "fmt" - "math/rand" "net" "os" "strconv" @@ -34,6 +33,7 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal/grpcrand" "google.golang.org/grpc/resolver" ) @@ -52,7 +52,6 @@ const ( var ( errMissingAddr = errors.New("missing address") - randomGen = rand.New(rand.NewSource(time.Now().UnixNano())) ) // NewBuilder creates a dnsBuilder which is used to factory DNS resolvers. @@ -67,6 +66,9 @@ type dnsBuilder struct { // Build creates and starts a DNS resolver that watches the name resolution of the target. func (b *dnsBuilder) Build(target resolver.Target, cc resolver.ClientConn, opts resolver.BuildOption) (resolver.Resolver, error) { + if target.Authority != "" { + return nil, fmt.Errorf("Default DNS resolver does not support custom DNS server") + } host, port, err := parseTarget(target.Endpoint) if err != nil { return nil, err @@ -346,7 +348,7 @@ func chosenByPercentage(a *int) bool { if a == nil { return true } - return randomGen.Intn(100)+1 <= *a + return grpcrand.Intn(100)+1 <= *a } func canaryingSC(js string) string { diff --git a/vendor/google.golang.org/grpc/resolver/dns/dns_resolver_test.go b/vendor/google.golang.org/grpc/resolver/dns/dns_resolver_test.go index 3e82db677..66c9735f7 100644 --- a/vendor/google.golang.org/grpc/resolver/dns/dns_resolver_test.go +++ b/vendor/google.golang.org/grpc/resolver/dns/dns_resolver_test.go @@ -27,8 +27,8 @@ import ( "testing" "time" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/resolver" - "google.golang.org/grpc/test/leakcheck" ) func TestMain(m *testing.M) { diff --git a/vendor/google.golang.org/grpc/resolver_conn_wrapper.go b/vendor/google.golang.org/grpc/resolver_conn_wrapper.go index 1b493db2e..494d6931e 100644 --- a/vendor/google.golang.org/grpc/resolver_conn_wrapper.go +++ b/vendor/google.golang.org/grpc/resolver_conn_wrapper.go @@ -65,8 +65,8 @@ func parseTarget(target string) (ret resolver.Target) { } // newCCResolverWrapper parses cc.target for scheme and gets the resolver -// builder for this scheme. It then builds the resolver and starts the -// monitoring goroutine for it. +// builder for this scheme and builds the resolver. The monitoring goroutine +// for it is not started yet and can be created by calling start(). // // If withResolverBuilder dial option is set, the specified resolver will be // used instead. @@ -148,7 +148,7 @@ func (ccr *ccResolverWrapper) NewAddress(addrs []resolver.Address) { } // NewServiceConfig is called by the resolver implemenetion to send service -// configs to gPRC. +// configs to gRPC. func (ccr *ccResolverWrapper) NewServiceConfig(sc string) { select { case <-ccr.scCh: diff --git a/vendor/google.golang.org/grpc/rpc_util.go b/vendor/google.golang.org/grpc/rpc_util.go index 69ef3c0b5..033801f34 100644 --- a/vendor/google.golang.org/grpc/rpc_util.go +++ b/vendor/google.golang.org/grpc/rpc_util.go @@ -419,8 +419,8 @@ func (o CustomCodecCallOption) after(c *callInfo) {} type payloadFormat uint8 const ( - compressionNone payloadFormat = iota // no compression - compressionMade + compressionNone payloadFormat = 0 // no compression + compressionMade payloadFormat = 1 // compressed ) // parser reads complete gRPC messages from the underlying reader. @@ -477,65 +477,82 @@ func (p *parser) recvMsg(maxReceiveMessageSize int) (pf payloadFormat, msg []byt return pf, msg, nil } -// encode serializes msg and returns a buffer of message header and a buffer of msg. -// If msg is nil, it generates the message header and an empty msg buffer. -// TODO(ddyihai): eliminate extra Compressor parameter. -func encode(c baseCodec, msg interface{}, cp Compressor, outPayload *stats.OutPayload, compressor encoding.Compressor) ([]byte, []byte, error) { - var ( - b []byte - cbuf *bytes.Buffer - ) - const ( - payloadLen = 1 - sizeLen = 4 - ) - if msg != nil { - var err error - b, err = c.Marshal(msg) - if err != nil { - return nil, nil, status.Errorf(codes.Internal, "grpc: error while marshaling: %v", err.Error()) - } - if outPayload != nil { - outPayload.Payload = msg - // TODO truncate large payload. - outPayload.Data = b - outPayload.Length = len(b) - } - if compressor != nil || cp != nil { - cbuf = new(bytes.Buffer) - // Has compressor, check Compressor is set by UseCompressor first. - if compressor != nil { - z, _ := compressor.Compress(cbuf) - if _, err := z.Write(b); err != nil { - return nil, nil, status.Errorf(codes.Internal, "grpc: error while compressing: %v", err.Error()) - } - z.Close() - } else { - // If Compressor is not set by UseCompressor, use default Compressor - if err := cp.Do(cbuf, b); err != nil { - return nil, nil, status.Errorf(codes.Internal, "grpc: error while compressing: %v", err.Error()) - } - } - b = cbuf.Bytes() - } +// encode serializes msg and returns a buffer containing the message, or an +// error if it is too large to be transmitted by grpc. If msg is nil, it +// generates an empty message. +func encode(c baseCodec, msg interface{}) ([]byte, error) { + if msg == nil { // NOTE: typed nils will not be caught by this check + return nil, nil + } + b, err := c.Marshal(msg) + if err != nil { + return nil, status.Errorf(codes.Internal, "grpc: error while marshaling: %v", err.Error()) } if uint(len(b)) > math.MaxUint32 { - return nil, nil, status.Errorf(codes.ResourceExhausted, "grpc: message too large (%d bytes)", len(b)) + return nil, status.Errorf(codes.ResourceExhausted, "grpc: message too large (%d bytes)", len(b)) } + return b, nil +} - bufHeader := make([]byte, payloadLen+sizeLen) - if compressor != nil || cp != nil { - bufHeader[0] = byte(compressionMade) +// compress returns the input bytes compressed by compressor or cp. If both +// compressors are nil, returns nil. +// +// TODO(dfawley): eliminate cp parameter by wrapping Compressor in an encoding.Compressor. +func compress(in []byte, cp Compressor, compressor encoding.Compressor) ([]byte, error) { + if compressor == nil && cp == nil { + return nil, nil + } + wrapErr := func(err error) error { + return status.Errorf(codes.Internal, "grpc: error while compressing: %v", err.Error()) + } + cbuf := &bytes.Buffer{} + if compressor != nil { + z, _ := compressor.Compress(cbuf) + if _, err := z.Write(in); err != nil { + return nil, wrapErr(err) + } + if err := z.Close(); err != nil { + return nil, wrapErr(err) + } } else { - bufHeader[0] = byte(compressionNone) + if err := cp.Do(cbuf, in); err != nil { + return nil, wrapErr(err) + } + } + return cbuf.Bytes(), nil +} + +const ( + payloadLen = 1 + sizeLen = 4 + headerLen = payloadLen + sizeLen +) + +// msgHeader returns a 5-byte header for the message being transmitted and the +// payload, which is compData if non-nil or data otherwise. +func msgHeader(data, compData []byte) (hdr []byte, payload []byte) { + hdr = make([]byte, headerLen) + if compData != nil { + hdr[0] = byte(compressionMade) + data = compData + } else { + hdr[0] = byte(compressionNone) } - // Write length of b into buf - binary.BigEndian.PutUint32(bufHeader[payloadLen:], uint32(len(b))) - if outPayload != nil { - outPayload.WireLength = payloadLen + sizeLen + len(b) + // Write length of payload into buf + binary.BigEndian.PutUint32(hdr[payloadLen:], uint32(len(data))) + return hdr, data +} + +func outPayload(client bool, msg interface{}, data, payload []byte, t time.Time) *stats.OutPayload { + return &stats.OutPayload{ + Client: client, + Payload: msg, + Data: data, + Length: len(data), + WireLength: len(payload) + headerLen, + SentTime: t, } - return bufHeader, b, nil } func checkRecvPayload(pf payloadFormat, recvCompress string, haveCompressor bool) *status.Status { @@ -721,7 +738,4 @@ const ( SupportPackageIsVersion5 = true ) -// Version is the current grpc version. -const Version = "1.12.0" - const grpcUA = "grpc-go/" + Version diff --git a/vendor/google.golang.org/grpc/rpc_util_test.go b/vendor/google.golang.org/grpc/rpc_util_test.go index 770e850c6..f28cff233 100644 --- a/vendor/google.golang.org/grpc/rpc_util_test.go +++ b/vendor/google.golang.org/grpc/rpc_util_test.go @@ -105,23 +105,25 @@ func TestEncode(t *testing.T) { for _, test := range []struct { // input msg proto.Message - cp Compressor // outputs hdr []byte data []byte err error }{ - {nil, nil, []byte{0, 0, 0, 0, 0}, []byte{}, nil}, + {nil, []byte{0, 0, 0, 0, 0}, []byte{}, nil}, } { - hdr, data, err := encode(encoding.GetCodec(protoenc.Name), test.msg, nil, nil, nil) - if err != test.err || !bytes.Equal(hdr, test.hdr) || !bytes.Equal(data, test.data) { - t.Fatalf("encode(_, _, %v, _) = %v, %v, %v\nwant %v, %v, %v", test.cp, hdr, data, err, test.hdr, test.data, test.err) + data, err := encode(encoding.GetCodec(protoenc.Name), test.msg) + if err != test.err || !bytes.Equal(data, test.data) { + t.Errorf("encode(_, %v) = %v, %v; want %v, %v", test.msg, data, err, test.data, test.err) + continue + } + if hdr, _ := msgHeader(data, nil); !bytes.Equal(hdr, test.hdr) { + t.Errorf("msgHeader(%v, false) = %v; want %v", data, hdr, test.hdr) } } } func TestCompress(t *testing.T) { - bestCompressor, err := NewGZIPCompressorWithLevel(gzip.BestCompression) if err != nil { t.Fatalf("Could not initialize gzip compressor with best compression.") @@ -214,12 +216,12 @@ func TestParseDialTarget(t *testing.T) { func bmEncode(b *testing.B, mSize int) { cdc := encoding.GetCodec(protoenc.Name) msg := &perfpb.Buffer{Body: make([]byte, mSize)} - encodeHdr, encodeData, _ := encode(cdc, msg, nil, nil, nil) - encodedSz := int64(len(encodeHdr) + len(encodeData)) + encodeData, _ := encode(cdc, msg) + encodedSz := int64(len(encodeData)) b.ReportAllocs() b.ResetTimer() for i := 0; i < b.N; i++ { - encode(cdc, msg, nil, nil, nil) + encode(cdc, msg) } b.SetBytes(encodedSz) } diff --git a/vendor/google.golang.org/grpc/server.go b/vendor/google.golang.org/grpc/server.go index 4969331cb..014c72b3f 100644 --- a/vendor/google.golang.org/grpc/server.go +++ b/vendor/google.golang.org/grpc/server.go @@ -38,13 +38,13 @@ import ( "golang.org/x/net/http2" "golang.org/x/net/trace" - "google.golang.org/grpc/channelz" "google.golang.org/grpc/codes" "google.golang.org/grpc/credentials" "google.golang.org/grpc/encoding" "google.golang.org/grpc/encoding/proto" "google.golang.org/grpc/grpclog" "google.golang.org/grpc/internal" + "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/metadata" "google.golang.org/grpc/stats" @@ -827,24 +827,24 @@ func (s *Server) incrCallsFailed() { } func (s *Server) sendResponse(t transport.ServerTransport, stream *transport.Stream, msg interface{}, cp Compressor, opts *transport.Options, comp encoding.Compressor) error { - var ( - outPayload *stats.OutPayload - ) - if s.opts.statsHandler != nil { - outPayload = &stats.OutPayload{} - } - hdr, data, err := encode(s.getCodec(stream.ContentSubtype()), msg, cp, outPayload, comp) + data, err := encode(s.getCodec(stream.ContentSubtype()), msg) if err != nil { grpclog.Errorln("grpc: server failed to encode response: ", err) return err } - if len(data) > s.opts.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "grpc: trying to send message larger than max (%d vs. %d)", len(data), s.opts.maxSendMessageSize) + compData, err := compress(data, cp, comp) + if err != nil { + grpclog.Errorln("grpc: server failed to compress response: ", err) + return err } - err = t.Write(stream, hdr, data, opts) - if err == nil && outPayload != nil { - outPayload.SentTime = time.Now() - s.opts.statsHandler.HandleRPC(stream.Context(), outPayload) + hdr, payload := msgHeader(data, compData) + // TODO(dfawley): should we be checking len(data) instead? + if len(payload) > s.opts.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "grpc: trying to send message larger than max (%d vs. %d)", len(payload), s.opts.maxSendMessageSize) + } + err = t.Write(stream, hdr, payload, opts) + if err == nil && s.opts.statsHandler != nil { + s.opts.statsHandler.HandleRPC(stream.Context(), outPayload(false, msg, data, payload, time.Now())) } return err } diff --git a/vendor/google.golang.org/grpc/server_test.go b/vendor/google.golang.org/grpc/server_test.go index 2b379f372..77e40b029 100644 --- a/vendor/google.golang.org/grpc/server_test.go +++ b/vendor/google.golang.org/grpc/server_test.go @@ -26,7 +26,7 @@ import ( "time" "golang.org/x/net/context" - "google.golang.org/grpc/test/leakcheck" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/transport" ) diff --git a/vendor/google.golang.org/grpc/stats/grpc_testing/test.pb.go b/vendor/google.golang.org/grpc/stats/grpc_testing/test.pb.go index e0b96bbe2..9db10f815 100644 --- a/vendor/google.golang.org/grpc/stats/grpc_testing/test.pb.go +++ b/vendor/google.golang.org/grpc/stats/grpc_testing/test.pb.go @@ -24,7 +24,7 @@ var _ = math.Inf const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package type SimpleRequest struct { - Id int32 `protobuf:"varint,2,opt,name=id" json:"id,omitempty"` + Id int32 `protobuf:"varint,2,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -62,7 +62,7 @@ func (m *SimpleRequest) GetId() int32 { } type SimpleResponse struct { - Id int32 `protobuf:"varint,3,opt,name=id" json:"id,omitempty"` + Id int32 `protobuf:"varint,3,opt,name=id,proto3" json:"id,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -112,8 +112,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for TestService service - +// TestServiceClient is the client API for TestService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type TestServiceClient interface { // One request followed by one response. // The server returns the client id as-is. @@ -138,7 +139,7 @@ func NewTestServiceClient(cc *grpc.ClientConn) TestServiceClient { func (c *testServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, opts ...grpc.CallOption) (*SimpleResponse, error) { out := new(SimpleResponse) - err := grpc.Invoke(ctx, "/grpc.testing.TestService/UnaryCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.TestService/UnaryCall", in, out, opts...) if err != nil { return nil, err } @@ -146,7 +147,7 @@ func (c *testServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, op } func (c *testServiceClient) FullDuplexCall(ctx context.Context, opts ...grpc.CallOption) (TestService_FullDuplexCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[0], c.cc, "/grpc.testing.TestService/FullDuplexCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[0], "/grpc.testing.TestService/FullDuplexCall", opts...) if err != nil { return nil, err } @@ -177,7 +178,7 @@ func (x *testServiceFullDuplexCallClient) Recv() (*SimpleResponse, error) { } func (c *testServiceClient) ClientStreamCall(ctx context.Context, opts ...grpc.CallOption) (TestService_ClientStreamCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[1], c.cc, "/grpc.testing.TestService/ClientStreamCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[1], "/grpc.testing.TestService/ClientStreamCall", opts...) if err != nil { return nil, err } @@ -211,7 +212,7 @@ func (x *testServiceClientStreamCallClient) CloseAndRecv() (*SimpleResponse, err } func (c *testServiceClient) ServerStreamCall(ctx context.Context, in *SimpleRequest, opts ...grpc.CallOption) (TestService_ServerStreamCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[2], c.cc, "/grpc.testing.TestService/ServerStreamCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[2], "/grpc.testing.TestService/ServerStreamCall", opts...) if err != nil { return nil, err } @@ -242,8 +243,7 @@ func (x *testServiceServerStreamCallClient) Recv() (*SimpleResponse, error) { return m, nil } -// Server API for TestService service - +// TestServiceServer is the server API for TestService service. type TestServiceServer interface { // One request followed by one response. // The server returns the client id as-is. diff --git a/vendor/google.golang.org/grpc/status/go16.go b/vendor/google.golang.org/grpc/status/go16.go new file mode 100644 index 000000000..e59b53e82 --- /dev/null +++ b/vendor/google.golang.org/grpc/status/go16.go @@ -0,0 +1,42 @@ +// +build go1.6,!go1.7 + +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package status + +import ( + "golang.org/x/net/context" + "google.golang.org/grpc/codes" +) + +// FromContextError converts a context error into a Status. It returns a +// Status with codes.OK if err is nil, or a Status with codes.Unknown if err is +// non-nil and not a context error. +func FromContextError(err error) *Status { + switch err { + case nil: + return New(codes.OK, "") + case context.DeadlineExceeded: + return New(codes.DeadlineExceeded, err.Error()) + case context.Canceled: + return New(codes.Canceled, err.Error()) + default: + return New(codes.Unknown, err.Error()) + } +} diff --git a/vendor/google.golang.org/grpc/status/go17.go b/vendor/google.golang.org/grpc/status/go17.go new file mode 100644 index 000000000..090215149 --- /dev/null +++ b/vendor/google.golang.org/grpc/status/go17.go @@ -0,0 +1,44 @@ +// +build go1.7 + +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package status + +import ( + "context" + + netctx "golang.org/x/net/context" + "google.golang.org/grpc/codes" +) + +// FromContextError converts a context error into a Status. It returns a +// Status with codes.OK if err is nil, or a Status with codes.Unknown if err is +// non-nil and not a context error. +func FromContextError(err error) *Status { + switch err { + case nil: + return New(codes.OK, "") + case context.DeadlineExceeded, netctx.DeadlineExceeded: + return New(codes.DeadlineExceeded, err.Error()) + case context.Canceled, netctx.Canceled: + return New(codes.Canceled, err.Error()) + default: + return New(codes.Unknown, err.Error()) + } +} diff --git a/vendor/google.golang.org/grpc/status/go17_test.go b/vendor/google.golang.org/grpc/status/go17_test.go new file mode 100644 index 000000000..2dc5add94 --- /dev/null +++ b/vendor/google.golang.org/grpc/status/go17_test.go @@ -0,0 +1,44 @@ +// +build go1.7 + +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package status + +import ( + "context" + "testing" + + "google.golang.org/grpc/codes" +) + +func TestFromStdContextError(t *testing.T) { + testCases := []struct { + in error + want *Status + }{ + {in: context.DeadlineExceeded, want: New(codes.DeadlineExceeded, context.DeadlineExceeded.Error())}, + {in: context.Canceled, want: New(codes.Canceled, context.Canceled.Error())}, + } + for _, tc := range testCases { + got := FromContextError(tc.in) + if got.Code() != tc.want.Code() || got.Message() != tc.want.Message() { + t.Errorf("FromContextError(%v) = %v; want %v", tc.in, got, tc.want) + } + } +} diff --git a/vendor/google.golang.org/grpc/status/status_test.go b/vendor/google.golang.org/grpc/status/status_test.go index f97fadd75..1eb132093 100644 --- a/vendor/google.golang.org/grpc/status/status_test.go +++ b/vendor/google.golang.org/grpc/status/status_test.go @@ -28,6 +28,7 @@ import ( "github.com/golang/protobuf/ptypes" apb "github.com/golang/protobuf/ptypes/any" dpb "github.com/golang/protobuf/ptypes/duration" + "golang.org/x/net/context" cpb "google.golang.org/genproto/googleapis/rpc/code" epb "google.golang.org/genproto/googleapis/rpc/errdetails" spb "google.golang.org/genproto/googleapis/rpc/status" @@ -327,3 +328,21 @@ func mustMarshalAny(msg proto.Message) *apb.Any { } return any } + +func TestFromContextError(t *testing.T) { + testCases := []struct { + in error + want *Status + }{ + {in: nil, want: New(codes.OK, "")}, + {in: context.DeadlineExceeded, want: New(codes.DeadlineExceeded, context.DeadlineExceeded.Error())}, + {in: context.Canceled, want: New(codes.Canceled, context.Canceled.Error())}, + {in: errors.New("other"), want: New(codes.Unknown, "other")}, + } + for _, tc := range testCases { + got := FromContextError(tc.in) + if got.Code() != tc.want.Code() || got.Message() != tc.want.Message() { + t.Errorf("FromContextError(%v) = %v; want %v", tc.in, got, tc.want) + } + } +} diff --git a/vendor/google.golang.org/grpc/stickiness_linkedmap.go b/vendor/google.golang.org/grpc/stickiness_linkedmap.go new file mode 100644 index 000000000..1c726af16 --- /dev/null +++ b/vendor/google.golang.org/grpc/stickiness_linkedmap.go @@ -0,0 +1,97 @@ +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package grpc + +import ( + "container/list" +) + +type linkedMapKVPair struct { + key string + value *stickyStoreEntry +} + +// linkedMap is an implementation of a map that supports removing the oldest +// entry. +// +// linkedMap is NOT thread safe. +// +// It's for use of stickiness only! +type linkedMap struct { + m map[string]*list.Element + l *list.List // Head of the list is the oldest element. +} + +// newLinkedMap returns a new LinkedMap. +func newLinkedMap() *linkedMap { + return &linkedMap{ + m: make(map[string]*list.Element), + l: list.New(), + } +} + +// put adds entry (key, value) to the map. Existing key will be overridden. +func (m *linkedMap) put(key string, value *stickyStoreEntry) { + if oldE, ok := m.m[key]; ok { + // Remove existing entry. + m.l.Remove(oldE) + } + e := m.l.PushBack(&linkedMapKVPair{key: key, value: value}) + m.m[key] = e +} + +// get returns the value of the given key. +func (m *linkedMap) get(key string) (*stickyStoreEntry, bool) { + e, ok := m.m[key] + if !ok { + return nil, false + } + m.l.MoveToBack(e) + return e.Value.(*linkedMapKVPair).value, true +} + +// remove removes key from the map, and returns the value. The map is not +// modified if key is not in the map. +func (m *linkedMap) remove(key string) (*stickyStoreEntry, bool) { + e, ok := m.m[key] + if !ok { + return nil, false + } + delete(m.m, key) + m.l.Remove(e) + return e.Value.(*linkedMapKVPair).value, true +} + +// len returns the len of the map. +func (m *linkedMap) len() int { + return len(m.m) +} + +// clear removes all elements from the map. +func (m *linkedMap) clear() { + m.m = make(map[string]*list.Element) + m.l = list.New() +} + +// removeOldest removes the oldest key from the map. +func (m *linkedMap) removeOldest() { + e := m.l.Front() + m.l.Remove(e) + delete(m.m, e.Value.(*linkedMapKVPair).key) +} diff --git a/vendor/google.golang.org/grpc/stickiness_linkedmap_test.go b/vendor/google.golang.org/grpc/stickiness_linkedmap_test.go new file mode 100644 index 000000000..e2d763555 --- /dev/null +++ b/vendor/google.golang.org/grpc/stickiness_linkedmap_test.go @@ -0,0 +1,186 @@ +/* + * + * Copyright 2018 gRPC authors. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + */ + +package grpc + +import ( + "container/list" + "fmt" + "reflect" + "testing" +) + +var linkedMapTestData = make([]*stickyStoreEntry, 5) + +func TestLinkedMapPutGet(t *testing.T) { + m := newLinkedMap() + m.put("one", linkedMapTestData[0]) + if got, ok := m.get("one"); !ok || got != linkedMapTestData[0] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 1, got, ok, "one") + } + m.put("two", linkedMapTestData[1]) + if got, ok := m.get("two"); !ok || got != linkedMapTestData[1] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 2, got, ok, "two") + } + m.put("oneone", linkedMapTestData[4]) + if got, ok := m.get("one"); !ok || got != linkedMapTestData[4] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 1, got, ok, "oneone") + } +} + +func TestLinkedMapRemove(t *testing.T) { + m := newLinkedMap() + m.put("one", linkedMapTestData[0]) + if got, ok := m.get("one"); !ok || got != linkedMapTestData[0] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 1, got, ok, "one") + } + m.put("two", linkedMapTestData[1]) + if got, ok := m.get("two"); !ok || got != linkedMapTestData[1] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 2, got, ok, "two") + } + + if got, ok := m.remove("one"); !ok || got != linkedMapTestData[0] { + t.Errorf("m.Remove(%v) = %v, %v, want %v, true", 1, got, ok, "one") + } + if got, ok := m.get("one"); ok { + t.Errorf("m.Get(%v) = %v, %v, want _, false", 1, got, ok) + } + // 2 should still in the map. + if got, ok := m.get("two"); !ok || got != linkedMapTestData[1] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 2, got, ok, "two") + } +} + +func TestLinkedMapLen(t *testing.T) { + m := newLinkedMap() + if got := m.len(); got != 0 { + t.Errorf("m.Len() = %v, want %v", got, 0) + } + m.put("one", linkedMapTestData[0]) + if got := m.len(); got != 1 { + t.Errorf("m.Len() = %v, want %v", got, 1) + } + m.put("two", linkedMapTestData[1]) + if got := m.len(); got != 2 { + t.Errorf("m.Len() = %v, want %v", got, 2) + } + m.put("one", linkedMapTestData[4]) + if got := m.len(); got != 2 { + t.Errorf("m.Len() = %v, want %v", got, 2) + } + + // Internal checks. + if got := len(m.m); got != 2 { + t.Errorf("len(m.m) = %v, want %v", got, 2) + } + if got := m.l.Len(); got != 2 { + t.Errorf("m.l.Len() = %v, want %v", got, 2) + } +} + +func TestLinkedMapClear(t *testing.T) { + m := newLinkedMap() + m.put("one", linkedMapTestData[0]) + if got, ok := m.get("one"); !ok || got != linkedMapTestData[0] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 1, got, ok, "one") + } + m.put("two", linkedMapTestData[1]) + if got, ok := m.get("two"); !ok || got != linkedMapTestData[1] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 2, got, ok, "two") + } + + m.clear() + if got, ok := m.get("one"); ok { + t.Errorf("m.Get(%v) = %v, %v, want _, false", 1, got, ok) + } + if got, ok := m.get("two"); ok { + t.Errorf("m.Get(%v) = %v, %v, want _, false", 2, got, ok) + } + if got := m.len(); got != 0 { + t.Errorf("m.Len() = %v, want %v", got, 0) + } +} + +func TestLinkedMapRemoveOldest(t *testing.T) { + m := newLinkedMap() + m.put("one", linkedMapTestData[0]) + m.put("two", linkedMapTestData[1]) + m.put("three", linkedMapTestData[2]) + m.put("four", linkedMapTestData[3]) + if got, ok := m.get("one"); !ok || got != linkedMapTestData[0] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 1, got, ok, "one") + } + if got, ok := m.get("two"); !ok || got != linkedMapTestData[1] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 2, got, ok, "two") + } + if got, ok := m.get("three"); !ok || got != linkedMapTestData[2] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 3, got, ok, "three") + } + if got, ok := m.get("four"); !ok || got != linkedMapTestData[3] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 4, got, ok, "four") + } + + if err := checkListOrdered(m.l, []string{"one", "two", "three", "four"}); err != nil { + t.Fatalf("m.l is not expected: %v", err) + } + + m.put("three", linkedMapTestData[2]) + if err := checkListOrdered(m.l, []string{"one", "two", "four", "three"}); err != nil { + t.Fatalf("m.l is not expected: %v", err) + } + m.put("four", linkedMapTestData[3]) + if err := checkListOrdered(m.l, []string{"one", "two", "three", "four"}); err != nil { + t.Fatalf("m.l is not expected: %v", err) + } + + m.removeOldest() + if got, ok := m.get("one"); ok { + t.Errorf("m.Get(%v) = %v, %v, want _, false", 1, got, ok) + } + if err := checkListOrdered(m.l, []string{"two", "three", "four"}); err != nil { + t.Fatalf("m.l is not expected: %v", err) + } + + m.get("two") // 2 is refreshed, 3 becomes the oldest + if err := checkListOrdered(m.l, []string{"three", "four", "two"}); err != nil { + t.Fatalf("m.l is not expected: %v", err) + } + + m.removeOldest() + if got, ok := m.get("three"); ok { + t.Errorf("m.Get(%v) = %v, %v, want _, false", 3, got, ok) + } + // 2 and 4 are still in map. + if got, ok := m.get("two"); !ok || got != linkedMapTestData[1] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 2, got, ok, "two") + } + if got, ok := m.get("four"); !ok || got != linkedMapTestData[3] { + t.Errorf("m.Get(%v) = %v, %v, want %v, true", 4, got, ok, "four") + } +} + +func checkListOrdered(l *list.List, want []string) error { + got := make([]string, 0, len(want)) + for p := l.Front(); p != nil; p = p.Next() { + got = append(got, p.Value.(*linkedMapKVPair).key) + } + if !reflect.DeepEqual(got, want) { + return fmt.Errorf("list elements: %v, want %v", got, want) + } + return nil +} diff --git a/vendor/google.golang.org/grpc/stickiness_test.go b/vendor/google.golang.org/grpc/stickiness_test.go index 10786faf2..f625d680b 100644 --- a/vendor/google.golang.org/grpc/stickiness_test.go +++ b/vendor/google.golang.org/grpc/stickiness_test.go @@ -27,10 +27,10 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc/balancer/roundrobin" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/metadata" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" - "google.golang.org/grpc/test/leakcheck" ) func TestStickyKeyFromContext(t *testing.T) { @@ -207,7 +207,7 @@ func TestStickinessChangeMDKey(t *testing.T) { } cc.blockingpicker.stickiness.mu.Lock() - mapLen := len(cc.blockingpicker.stickiness.store) + mapLen := cc.blockingpicker.stickiness.store.len() cc.blockingpicker.stickiness.mu.Unlock() if mapLen != 1 { t.Fatalf("length of stickiness map is %v, want 1", mapLen) @@ -218,7 +218,7 @@ func TestStickinessChangeMDKey(t *testing.T) { var i int for i = 0; i < 1000; i++ { cc.blockingpicker.stickiness.mu.Lock() - mapLen = len(cc.blockingpicker.stickiness.store) + mapLen = cc.blockingpicker.stickiness.store.len() cc.blockingpicker.stickiness.mu.Unlock() if mapLen == 0 { break @@ -262,7 +262,7 @@ func TestStickinessSwitchingBalancer(t *testing.T) { } cc.blockingpicker.stickiness.mu.Lock() - mapLen := len(cc.blockingpicker.stickiness.store) + mapLen := cc.blockingpicker.stickiness.store.len() cc.blockingpicker.stickiness.mu.Unlock() if mapLen != 1 { t.Fatalf("length of stickiness map is %v, want 1", mapLen) @@ -275,7 +275,7 @@ func TestStickinessSwitchingBalancer(t *testing.T) { var i int for i = 0; i < 1000; i++ { cc.blockingpicker.stickiness.mu.Lock() - mapLen = len(cc.blockingpicker.stickiness.store) + mapLen = cc.blockingpicker.stickiness.store.len() cc.blockingpicker.stickiness.mu.Unlock() if mapLen == 0 { break diff --git a/vendor/google.golang.org/grpc/stream.go b/vendor/google.golang.org/grpc/stream.go index 82921a15a..152d9eccd 100644 --- a/vendor/google.golang.org/grpc/stream.go +++ b/vendor/google.golang.org/grpc/stream.go @@ -27,9 +27,9 @@ import ( "golang.org/x/net/context" "golang.org/x/net/trace" "google.golang.org/grpc/balancer" - "google.golang.org/grpc/channelz" "google.golang.org/grpc/codes" "google.golang.org/grpc/encoding" + "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/metadata" "google.golang.org/grpc/stats" "google.golang.org/grpc/status" @@ -101,7 +101,21 @@ type ClientStream interface { } // NewStream creates a new Stream for the client side. This is typically -// called by generated code. +// called by generated code. ctx is used for the lifetime of the stream. +// +// To ensure resources are not leaked due to the stream returned, one of the following +// actions must be performed: +// +// 1. Call Close on the ClientConn. +// 2. Cancel the context provided. +// 3. Call RecvMsg until a non-nil error is returned. A protobuf-generated +// client-streaming RPC, for instance, might use the helper function +// CloseAndRecv (note that CloseSend does not Recv, therefore is not +// guaranteed to release all resources). +// 4. Receive a non-nil, non-io.EOF error from Header or SendMsg. +// +// If none of the above happen, a goroutine and a context will be leaked, and grpc +// will not call the optionally-configured stats handler with a stats.End message. func (cc *ClientConn) NewStream(ctx context.Context, desc *StreamDesc, method string, opts ...CallOption) (ClientStream, error) { // allow interceptor to see all applicable call options, which means those // configured as defaults from dial option as well as per-call options @@ -113,8 +127,7 @@ func (cc *ClientConn) NewStream(ctx context.Context, desc *StreamDesc, method st return newClientStream(ctx, desc, cc, method, opts...) } -// NewClientStream creates a new Stream for the client side. This is typically -// called by generated code. +// NewClientStream is a wrapper for ClientConn.NewStream. // // DEPRECATED: Use ClientConn.NewStream instead. func NewClientStream(ctx context.Context, desc *StreamDesc, cc *ClientConn, method string, opts ...CallOption) (ClientStream, error) { @@ -466,27 +479,27 @@ func (a *csAttempt) sendMsg(m interface{}) (err error) { } a.mu.Unlock() } - var outPayload *stats.OutPayload - if a.statsHandler != nil { - outPayload = &stats.OutPayload{ - Client: true, - } - } - hdr, data, err := encode(cs.codec, m, cs.cp, outPayload, cs.comp) + data, err := encode(cs.codec, m) if err != nil { return err } - if len(data) > *cs.c.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(data), *cs.c.maxSendMessageSize) + compData, err := compress(data, cs.cp, cs.comp) + if err != nil { + return err } + hdr, payload := msgHeader(data, compData) + // TODO(dfawley): should we be checking len(data) instead? + if len(payload) > *cs.c.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(payload), *cs.c.maxSendMessageSize) + } + if !cs.desc.ClientStreams { cs.sentLast = true } - err = a.t.Write(a.s, hdr, data, &transport.Options{Last: !cs.desc.ClientStreams}) + err = a.t.Write(a.s, hdr, payload, &transport.Options{Last: !cs.desc.ClientStreams}) if err == nil { - if outPayload != nil { - outPayload.SentTime = time.Now() - a.statsHandler.HandleRPC(a.ctx, outPayload) + if a.statsHandler != nil { + a.statsHandler.HandleRPC(a.ctx, outPayload(true, m, data, payload, time.Now())) } if channelz.IsOn() { a.t.IncrMsgSent() @@ -696,23 +709,24 @@ func (ss *serverStream) SendMsg(m interface{}) (err error) { ss.t.IncrMsgSent() } }() - var outPayload *stats.OutPayload - if ss.statsHandler != nil { - outPayload = &stats.OutPayload{} - } - hdr, data, err := encode(ss.codec, m, ss.cp, outPayload, ss.comp) + data, err := encode(ss.codec, m) if err != nil { return err } - if len(data) > ss.maxSendMessageSize { - return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(data), ss.maxSendMessageSize) + compData, err := compress(data, ss.cp, ss.comp) + if err != nil { + return err } - if err := ss.t.Write(ss.s, hdr, data, &transport.Options{Last: false}); err != nil { + hdr, payload := msgHeader(data, compData) + // TODO(dfawley): should we be checking len(data) instead? + if len(payload) > ss.maxSendMessageSize { + return status.Errorf(codes.ResourceExhausted, "trying to send message larger than max (%d vs. %d)", len(payload), ss.maxSendMessageSize) + } + if err := ss.t.Write(ss.s, hdr, payload, &transport.Options{Last: false}); err != nil { return toRPCErr(err) } - if outPayload != nil { - outPayload.SentTime = time.Now() - ss.statsHandler.HandleRPC(ss.s.Context(), outPayload) + if ss.statsHandler != nil { + ss.statsHandler.HandleRPC(ss.s.Context(), outPayload(false, m, data, payload, time.Now())) } return nil } diff --git a/vendor/google.golang.org/grpc/stress/grpc_testing/metrics.pb.go b/vendor/google.golang.org/grpc/stress/grpc_testing/metrics.pb.go index 0440d3dab..31e95d2b8 100644 --- a/vendor/google.golang.org/grpc/stress/grpc_testing/metrics.pb.go +++ b/vendor/google.golang.org/grpc/stress/grpc_testing/metrics.pb.go @@ -25,7 +25,7 @@ const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package // Response message containing the gauge name and value type GaugeResponse struct { - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` // Types that are valid to be assigned to Value: // *GaugeResponse_LongValue // *GaugeResponse_DoubleValue @@ -65,13 +65,13 @@ type isGaugeResponse_Value interface { } type GaugeResponse_LongValue struct { - LongValue int64 `protobuf:"varint,2,opt,name=long_value,json=longValue,oneof"` + LongValue int64 `protobuf:"varint,2,opt,name=long_value,json=longValue,proto3,oneof"` } type GaugeResponse_DoubleValue struct { - DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,oneof"` + DoubleValue float64 `protobuf:"fixed64,3,opt,name=double_value,json=doubleValue,proto3,oneof"` } type GaugeResponse_StringValue struct { - StringValue string `protobuf:"bytes,4,opt,name=string_value,json=stringValue,oneof"` + StringValue string `protobuf:"bytes,4,opt,name=string_value,json=stringValue,proto3,oneof"` } func (*GaugeResponse_LongValue) isGaugeResponse_Value() {} @@ -194,7 +194,7 @@ func _GaugeResponse_OneofSizer(msg proto.Message) (n int) { // Request message containing the gauge name type GaugeRequest struct { - Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"` + Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -275,8 +275,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for MetricsService service - +// MetricsServiceClient is the client API for MetricsService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type MetricsServiceClient interface { // Returns the values of all the gauges that are currently being maintained by // the service @@ -294,7 +295,7 @@ func NewMetricsServiceClient(cc *grpc.ClientConn) MetricsServiceClient { } func (c *metricsServiceClient) GetAllGauges(ctx context.Context, in *EmptyMessage, opts ...grpc.CallOption) (MetricsService_GetAllGaugesClient, error) { - stream, err := grpc.NewClientStream(ctx, &_MetricsService_serviceDesc.Streams[0], c.cc, "/grpc.testing.MetricsService/GetAllGauges", opts...) + stream, err := c.cc.NewStream(ctx, &_MetricsService_serviceDesc.Streams[0], "/grpc.testing.MetricsService/GetAllGauges", opts...) if err != nil { return nil, err } @@ -327,15 +328,14 @@ func (x *metricsServiceGetAllGaugesClient) Recv() (*GaugeResponse, error) { func (c *metricsServiceClient) GetGauge(ctx context.Context, in *GaugeRequest, opts ...grpc.CallOption) (*GaugeResponse, error) { out := new(GaugeResponse) - err := grpc.Invoke(ctx, "/grpc.testing.MetricsService/GetGauge", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.MetricsService/GetGauge", in, out, opts...) if err != nil { return nil, err } return out, nil } -// Server API for MetricsService service - +// MetricsServiceServer is the server API for MetricsService service. type MetricsServiceServer interface { // Returns the values of all the gauges that are currently being maintained by // the service diff --git a/vendor/google.golang.org/grpc/test/channelz_test.go b/vendor/google.golang.org/grpc/test/channelz_test.go index ffee50b2e..eaa657c4b 100644 --- a/vendor/google.golang.org/grpc/test/channelz_test.go +++ b/vendor/google.golang.org/grpc/test/channelz_test.go @@ -28,14 +28,15 @@ import ( "golang.org/x/net/context" "golang.org/x/net/http2" "google.golang.org/grpc" - "google.golang.org/grpc/channelz" + _ "google.golang.org/grpc/balancer/grpclb" "google.golang.org/grpc/codes" + "google.golang.org/grpc/internal/channelz" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/resolver" "google.golang.org/grpc/resolver/manual" "google.golang.org/grpc/status" testpb "google.golang.org/grpc/test/grpc_testing" - "google.golang.org/grpc/test/leakcheck" ) func init() { diff --git a/vendor/google.golang.org/grpc/test/end2end_test.go b/vendor/google.golang.org/grpc/test/end2end_test.go index ace57cb72..6574a64fc 100644 --- a/vendor/google.golang.org/grpc/test/end2end_test.go +++ b/vendor/google.golang.org/grpc/test/end2end_test.go @@ -53,8 +53,10 @@ import ( _ "google.golang.org/grpc/encoding/gzip" _ "google.golang.org/grpc/grpclog/glogger" "google.golang.org/grpc/health" + healthgrpc "google.golang.org/grpc/health/grpc_health_v1" healthpb "google.golang.org/grpc/health/grpc_health_v1" "google.golang.org/grpc/internal" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" @@ -65,7 +67,6 @@ import ( "google.golang.org/grpc/status" "google.golang.org/grpc/tap" testpb "google.golang.org/grpc/test/grpc_testing" - "google.golang.org/grpc/test/leakcheck" "google.golang.org/grpc/testdata" ) @@ -601,7 +602,7 @@ func (te *test) listenAndServe(ts testpb.TestServiceServer, listen func(network, internal.TestingUseHandlerImpl(s) } if te.healthServer != nil { - healthpb.RegisterHealthServer(s, te.healthServer) + healthgrpc.RegisterHealthServer(s, te.healthServer) } if te.testServer != nil { testpb.RegisterTestServiceServer(s, te.testServer) @@ -2252,7 +2253,7 @@ func testTap(t *testing.T, e env) { func healthCheck(d time.Duration, cc *grpc.ClientConn, serviceName string) (*healthpb.HealthCheckResponse, error) { ctx, cancel := context.WithTimeout(context.Background(), d) defer cancel() - hc := healthpb.NewHealthClient(cc) + hc := healthgrpc.NewHealthClient(cc) req := &healthpb.HealthCheckRequest{ Service: serviceName, } @@ -6142,6 +6143,73 @@ func TestServeExitsWhenListenerClosed(t *testing.T) { } } +// Service handler returns status with invalid utf8 message. +func TestStatusInvalidUTF8Message(t *testing.T) { + defer leakcheck.Check(t) + + var ( + origMsg = string([]byte{0xff, 0xfe, 0xfd}) + wantMsg = "���" + ) + + ss := &stubServer{ + emptyCall: func(ctx context.Context, in *testpb.Empty) (*testpb.Empty, error) { + return nil, status.Errorf(codes.Internal, origMsg) + }, + } + if err := ss.Start(nil); err != nil { + t.Fatalf("Error starting endpoint server: %v", err) + } + defer ss.Stop() + + ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second) + defer cancel() + + if _, err := ss.client.EmptyCall(ctx, &testpb.Empty{}); status.Convert(err).Message() != wantMsg { + t.Fatalf("ss.client.EmptyCall(_, _) = _, %v (msg %q); want _, err with msg %q", err, status.Convert(err).Message(), wantMsg) + } +} + +// Service handler returns status with details and invalid utf8 message. Proto +// will fail to marshal the status because of the invalid utf8 message. Details +// will be dropped when sending. +func TestStatusInvalidUTF8Details(t *testing.T) { + defer leakcheck.Check(t) + + var ( + origMsg = string([]byte{0xff, 0xfe, 0xfd}) + wantMsg = "���" + ) + + ss := &stubServer{ + emptyCall: func(ctx context.Context, in *testpb.Empty) (*testpb.Empty, error) { + st := status.New(codes.Internal, origMsg) + st, err := st.WithDetails(&testpb.Empty{}) + if err != nil { + return nil, err + } + return nil, st.Err() + }, + } + if err := ss.Start(nil); err != nil { + t.Fatalf("Error starting endpoint server: %v", err) + } + defer ss.Stop() + + ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second) + defer cancel() + + _, err := ss.client.EmptyCall(ctx, &testpb.Empty{}) + st := status.Convert(err) + if st.Message() != wantMsg { + t.Fatalf("ss.client.EmptyCall(_, _) = _, %v (msg %q); want _, err with msg %q", err, st.Message(), wantMsg) + } + if len(st.Details()) != 0 { + // Details should be dropped on the server side. + t.Fatalf("RPC status contain details: %v, want no details", st.Details()) + } +} + func TestClientDoesntDeadlockWhileWritingErrornousLargeMessages(t *testing.T) { defer leakcheck.Check(t) for _, e := range listTestEnv() { diff --git a/vendor/google.golang.org/grpc/test/gracefulstop_test.go b/vendor/google.golang.org/grpc/test/gracefulstop_test.go index 7ac12b09b..560235859 100644 --- a/vendor/google.golang.org/grpc/test/gracefulstop_test.go +++ b/vendor/google.golang.org/grpc/test/gracefulstop_test.go @@ -20,6 +20,7 @@ package test import ( "fmt" + "io" "net" "sync" "testing" @@ -27,7 +28,7 @@ import ( "golang.org/x/net/context" "google.golang.org/grpc" - "google.golang.org/grpc/test/leakcheck" + "google.golang.org/grpc/internal/leakcheck" testpb "google.golang.org/grpc/test/grpc_testing" ) @@ -50,7 +51,12 @@ func (d *delayListener) Accept() (net.Conn, error) { return nil, fmt.Errorf("listener is closed") default: close(d.acceptCalled) - return d.Listener.Accept() + conn, err := d.Listener.Accept() + // Allow closing of listener only after accept. + // Note: Dial can return successfully, yet Accept + // might now have finished. + d.allowClose() + return conn, err } } @@ -86,39 +92,19 @@ func (d *delayListener) Dial(to time.Duration) (net.Conn, error) { return d.cc, nil } -func (d *delayListener) clientWriteCalledChan() <-chan struct{} { - return d.cc.writeCalledChan() -} - type delayConn struct { net.Conn - blockRead chan struct{} - mu sync.Mutex - writeCalled chan struct{} + blockRead chan struct{} } -func (d *delayConn) writeCalledChan() <-chan struct{} { - d.mu.Lock() - defer d.mu.Unlock() - d.writeCalled = make(chan struct{}) - return d.writeCalled -} func (d *delayConn) allowRead() { close(d.blockRead) } + func (d *delayConn) Read(b []byte) (n int, err error) { <-d.blockRead return d.Conn.Read(b) } -func (d *delayConn) Write(b []byte) (n int, err error) { - d.mu.Lock() - if d.writeCalled != nil { - close(d.writeCalled) - d.writeCalled = nil - } - d.mu.Unlock() - return d.Conn.Write(b) -} func TestGracefulStop(t *testing.T) { defer leakcheck.Check(t) @@ -148,10 +134,16 @@ func TestGracefulStop(t *testing.T) { allowCloseCh: make(chan struct{}), } d := func(_ string, to time.Duration) (net.Conn, error) { return dlis.Dial(to) } + serverGotReq := make(chan struct{}) ss := &stubServer{ - emptyCall: func(ctx context.Context, in *testpb.Empty) (*testpb.Empty, error) { - return &testpb.Empty{}, nil + fullDuplexCall: func(stream testpb.TestService_FullDuplexCallServer) error { + close(serverGotReq) + _, err := stream.Recv() + if err != nil { + return err + } + return stream.Send(&testpb.StreamingOutputCallResponse{}) }, } s := grpc.NewServer() @@ -182,33 +174,38 @@ func TestGracefulStop(t *testing.T) { // Now dial. The listener's Accept method will return a valid connection, // even though GracefulStop has closed the listener. - ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second) + ctx, dialCancel := context.WithTimeout(context.Background(), 5*time.Second) + defer dialCancel() cc, err := grpc.DialContext(ctx, "", grpc.WithInsecure(), grpc.WithBlock(), grpc.WithDialer(d)) if err != nil { t.Fatalf("grpc.Dial(%q) = %v", lis.Addr().String(), err) } - cancel() client := testpb.NewTestServiceClient(cc) defer cc.Close() - dlis.allowClose() - - wcch := dlis.clientWriteCalledChan() - go func() { - // 5. Allow the client to read the GoAway. The RPC should complete - // successfully. - <-wcch - dlis.allowClientRead() - }() - // 4. Send an RPC on the new connection. // The server would send a GOAWAY first, but we are delaying the server's // writes for now until the client writes more than the preface. - ctx, cancel = context.WithTimeout(context.Background(), 5*time.Second) - if _, err := client.EmptyCall(ctx, &testpb.Empty{}); err != nil { - t.Fatalf("EmptyCall() = %v; want ", err) + ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second) + stream, err := client.FullDuplexCall(ctx) + if err != nil { + t.Fatalf("FullDuplexCall= _, %v; want _, ", err) + } + go func() { + // 5. Allow the client to read the GoAway. The RPC should complete + // successfully. + <-serverGotReq + dlis.allowClientRead() + }() + if err := stream.Send(&testpb.StreamingOutputCallRequest{}); err != nil { + t.Fatalf("stream.Send(_) = %v, want ", err) + } + if _, err := stream.Recv(); err != nil { + t.Fatalf("stream.Recv() = _, %v, want _, ", err) + } + if _, err := stream.Recv(); err != io.EOF { + t.Fatalf("stream.Recv() = _, %v, want _, io.EOF", err) } - // 5. happens above, then we finish the call. cancel() wg.Wait() diff --git a/vendor/google.golang.org/grpc/test/grpc_testing/test.pb.go b/vendor/google.golang.org/grpc/test/grpc_testing/test.pb.go index 290492a4e..871f7cc98 100644 --- a/vendor/google.golang.org/grpc/test/grpc_testing/test.pb.go +++ b/vendor/google.golang.org/grpc/test/grpc_testing/test.pb.go @@ -86,7 +86,7 @@ var xxx_messageInfo_Empty proto.InternalMessageInfo // A block of data, to simply increase gRPC message size. type Payload struct { // The type of data in body. - Type PayloadType `protobuf:"varint,1,opt,name=type,enum=grpc.testing.PayloadType" json:"type,omitempty"` + Type PayloadType `protobuf:"varint,1,opt,name=type,proto3,enum=grpc.testing.PayloadType" json:"type,omitempty"` // Primary contents of payload. Body []byte `protobuf:"bytes,2,opt,name=body,proto3" json:"body,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` @@ -136,16 +136,16 @@ func (m *Payload) GetBody() []byte { type SimpleRequest struct { // Desired payload type in the response from the server. // If response_type is RANDOM, server randomly chooses one from other formats. - ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` + ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,proto3,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` // Desired payload size in the response from the server. // If response_type is COMPRESSABLE, this denotes the size before compression. - ResponseSize int32 `protobuf:"varint,2,opt,name=response_size,json=responseSize" json:"response_size,omitempty"` + ResponseSize int32 `protobuf:"varint,2,opt,name=response_size,json=responseSize,proto3" json:"response_size,omitempty"` // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` // Whether SimpleResponse should include username. - FillUsername bool `protobuf:"varint,4,opt,name=fill_username,json=fillUsername" json:"fill_username,omitempty"` + FillUsername bool `protobuf:"varint,4,opt,name=fill_username,json=fillUsername,proto3" json:"fill_username,omitempty"` // Whether SimpleResponse should include OAuth scope. - FillOauthScope bool `protobuf:"varint,5,opt,name=fill_oauth_scope,json=fillOauthScope" json:"fill_oauth_scope,omitempty"` + FillOauthScope bool `protobuf:"varint,5,opt,name=fill_oauth_scope,json=fillOauthScope,proto3" json:"fill_oauth_scope,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -213,12 +213,12 @@ func (m *SimpleRequest) GetFillOauthScope() bool { // Unary response, as configured by the request. type SimpleResponse struct { // Payload to increase message size. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` // The user the request came from, for verifying authentication was // successful when the client expected it. - Username string `protobuf:"bytes,2,opt,name=username" json:"username,omitempty"` + Username string `protobuf:"bytes,2,opt,name=username,proto3" json:"username,omitempty"` // OAuth scope. - OauthScope string `protobuf:"bytes,3,opt,name=oauth_scope,json=oauthScope" json:"oauth_scope,omitempty"` + OauthScope string `protobuf:"bytes,3,opt,name=oauth_scope,json=oauthScope,proto3" json:"oauth_scope,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -272,7 +272,7 @@ func (m *SimpleResponse) GetOauthScope() string { // Client-streaming request. type StreamingInputCallRequest struct { // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -312,7 +312,7 @@ func (m *StreamingInputCallRequest) GetPayload() *Payload { // Client-streaming response. type StreamingInputCallResponse struct { // Aggregated size of payloads received from the client. - AggregatedPayloadSize int32 `protobuf:"varint,1,opt,name=aggregated_payload_size,json=aggregatedPayloadSize" json:"aggregated_payload_size,omitempty"` + AggregatedPayloadSize int32 `protobuf:"varint,1,opt,name=aggregated_payload_size,json=aggregatedPayloadSize,proto3" json:"aggregated_payload_size,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -353,10 +353,10 @@ func (m *StreamingInputCallResponse) GetAggregatedPayloadSize() int32 { type ResponseParameters struct { // Desired payload sizes in responses from the server. // If response_type is COMPRESSABLE, this denotes the size before compression. - Size int32 `protobuf:"varint,1,opt,name=size" json:"size,omitempty"` + Size int32 `protobuf:"varint,1,opt,name=size,proto3" json:"size,omitempty"` // Desired interval between consecutive responses in the response stream in // microseconds. - IntervalUs int32 `protobuf:"varint,2,opt,name=interval_us,json=intervalUs" json:"interval_us,omitempty"` + IntervalUs int32 `protobuf:"varint,2,opt,name=interval_us,json=intervalUs,proto3" json:"interval_us,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -406,11 +406,11 @@ type StreamingOutputCallRequest struct { // If response_type is RANDOM, the payload from each response in the stream // might be of different types. This is to simulate a mixed type of payload // stream. - ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` + ResponseType PayloadType `protobuf:"varint,1,opt,name=response_type,json=responseType,proto3,enum=grpc.testing.PayloadType" json:"response_type,omitempty"` // Configuration for each expected response message. - ResponseParameters []*ResponseParameters `protobuf:"bytes,2,rep,name=response_parameters,json=responseParameters" json:"response_parameters,omitempty"` + ResponseParameters []*ResponseParameters `protobuf:"bytes,2,rep,name=response_parameters,json=responseParameters,proto3" json:"response_parameters,omitempty"` // Optional input payload sent along with the request. - Payload *Payload `protobuf:"bytes,3,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,3,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -464,7 +464,7 @@ func (m *StreamingOutputCallRequest) GetPayload() *Payload { // Server-streaming response, as configured by the request and parameters. type StreamingOutputCallResponse struct { // Payload to increase response size. - Payload *Payload `protobuf:"bytes,1,opt,name=payload" json:"payload,omitempty"` + Payload *Payload `protobuf:"bytes,1,opt,name=payload,proto3" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` @@ -522,8 +522,9 @@ var _ grpc.ClientConn // is compatible with the grpc package it is being compiled against. const _ = grpc.SupportPackageIsVersion4 -// Client API for TestService service - +// TestServiceClient is the client API for TestService service. +// +// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream. type TestServiceClient interface { // One empty request followed by one empty response. EmptyCall(ctx context.Context, in *Empty, opts ...grpc.CallOption) (*Empty, error) @@ -557,7 +558,7 @@ func NewTestServiceClient(cc *grpc.ClientConn) TestServiceClient { func (c *testServiceClient) EmptyCall(ctx context.Context, in *Empty, opts ...grpc.CallOption) (*Empty, error) { out := new(Empty) - err := grpc.Invoke(ctx, "/grpc.testing.TestService/EmptyCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.TestService/EmptyCall", in, out, opts...) if err != nil { return nil, err } @@ -566,7 +567,7 @@ func (c *testServiceClient) EmptyCall(ctx context.Context, in *Empty, opts ...gr func (c *testServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, opts ...grpc.CallOption) (*SimpleResponse, error) { out := new(SimpleResponse) - err := grpc.Invoke(ctx, "/grpc.testing.TestService/UnaryCall", in, out, c.cc, opts...) + err := c.cc.Invoke(ctx, "/grpc.testing.TestService/UnaryCall", in, out, opts...) if err != nil { return nil, err } @@ -574,7 +575,7 @@ func (c *testServiceClient) UnaryCall(ctx context.Context, in *SimpleRequest, op } func (c *testServiceClient) StreamingOutputCall(ctx context.Context, in *StreamingOutputCallRequest, opts ...grpc.CallOption) (TestService_StreamingOutputCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[0], c.cc, "/grpc.testing.TestService/StreamingOutputCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[0], "/grpc.testing.TestService/StreamingOutputCall", opts...) if err != nil { return nil, err } @@ -606,7 +607,7 @@ func (x *testServiceStreamingOutputCallClient) Recv() (*StreamingOutputCallRespo } func (c *testServiceClient) StreamingInputCall(ctx context.Context, opts ...grpc.CallOption) (TestService_StreamingInputCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[1], c.cc, "/grpc.testing.TestService/StreamingInputCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[1], "/grpc.testing.TestService/StreamingInputCall", opts...) if err != nil { return nil, err } @@ -640,7 +641,7 @@ func (x *testServiceStreamingInputCallClient) CloseAndRecv() (*StreamingInputCal } func (c *testServiceClient) FullDuplexCall(ctx context.Context, opts ...grpc.CallOption) (TestService_FullDuplexCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[2], c.cc, "/grpc.testing.TestService/FullDuplexCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[2], "/grpc.testing.TestService/FullDuplexCall", opts...) if err != nil { return nil, err } @@ -671,7 +672,7 @@ func (x *testServiceFullDuplexCallClient) Recv() (*StreamingOutputCallResponse, } func (c *testServiceClient) HalfDuplexCall(ctx context.Context, opts ...grpc.CallOption) (TestService_HalfDuplexCallClient, error) { - stream, err := grpc.NewClientStream(ctx, &_TestService_serviceDesc.Streams[3], c.cc, "/grpc.testing.TestService/HalfDuplexCall", opts...) + stream, err := c.cc.NewStream(ctx, &_TestService_serviceDesc.Streams[3], "/grpc.testing.TestService/HalfDuplexCall", opts...) if err != nil { return nil, err } @@ -701,8 +702,7 @@ func (x *testServiceHalfDuplexCallClient) Recv() (*StreamingOutputCallResponse, return m, nil } -// Server API for TestService service - +// TestServiceServer is the server API for TestService service. type TestServiceServer interface { // One empty request followed by one empty response. EmptyCall(context.Context, *Empty) (*Empty, error) diff --git a/vendor/google.golang.org/grpc/transport/controlbuf.go b/vendor/google.golang.org/grpc/transport/controlbuf.go index e147cd51b..5c5891a11 100644 --- a/vendor/google.golang.org/grpc/transport/controlbuf.go +++ b/vendor/google.golang.org/grpc/transport/controlbuf.go @@ -28,6 +28,10 @@ import ( "golang.org/x/net/http2/hpack" ) +var updateHeaderTblSize = func(e *hpack.Encoder, v uint32) { + e.SetMaxDynamicTableSizeLimit(v) +} + type itemNode struct { it interface{} next *itemNode @@ -80,6 +84,13 @@ func (il *itemList) isEmpty() bool { // the control buffer of transport. They represent different aspects of // control tasks, e.g., flow control, settings, streaming resetting, etc. +// registerStream is used to register an incoming stream with loopy writer. +type registerStream struct { + streamID uint32 + wq *writeQuota +} + +// headerFrame is also used to register stream on the client-side. type headerFrame struct { streamID uint32 hf []hpack.HeaderField @@ -361,44 +372,47 @@ func newLoopyWriter(s side, fr *framer, cbuf *controlBuffer, bdpEst *bdpEstimato const minBatchSize = 1000 // run should be run in a separate goroutine. -func (l *loopyWriter) run() { - var ( - it interface{} - err error - isEmpty bool - ) +func (l *loopyWriter) run() (err error) { defer func() { - errorf("transport: loopyWriter.run returning. Err: %v", err) + if err == ErrConnClosing { + // Don't log ErrConnClosing as error since it happens + // 1. When the connection is closed by some other known issue. + // 2. User closed the connection. + // 3. A graceful close of connection. + infof("transport: loopyWriter.run returning. %v", err) + err = nil + } }() for { - it, err = l.cbuf.get(true) + it, err := l.cbuf.get(true) if err != nil { - return + return err } if err = l.handle(it); err != nil { - return + return err } if _, err = l.processData(); err != nil { - return + return err } gosched := true hasdata: for { - it, err = l.cbuf.get(false) + it, err := l.cbuf.get(false) if err != nil { - return + return err } if it != nil { if err = l.handle(it); err != nil { - return + return err } if _, err = l.processData(); err != nil { - return + return err } continue hasdata } - if isEmpty, err = l.processData(); err != nil { - return + isEmpty, err := l.processData() + if err != nil { + return err } if !isEmpty { continue hasdata @@ -450,30 +464,39 @@ func (l *loopyWriter) incomingSettingsHandler(s *incomingSettings) error { return l.framer.fr.WriteSettingsAck() } +func (l *loopyWriter) registerStreamHandler(h *registerStream) error { + str := &outStream{ + id: h.streamID, + state: empty, + itl: &itemList{}, + wq: h.wq, + } + l.estdStreams[h.streamID] = str + return nil +} + func (l *loopyWriter) headerHandler(h *headerFrame) error { if l.side == serverSide { - if h.endStream { // Case 1.A: Server wants to close stream. - // Make sure it's not a trailers only response. - if str, ok := l.estdStreams[h.streamID]; ok { - if str.state != empty { // either active or waiting on stream quota. - // add it str's list of items. - str.itl.enqueue(h) - return nil - } - } - if err := l.writeHeader(h.streamID, h.endStream, h.hf, h.onWrite); err != nil { - return err - } - return l.cleanupStreamHandler(h.cleanup) + str, ok := l.estdStreams[h.streamID] + if !ok { + warningf("transport: loopy doesn't recognize the stream: %d", h.streamID) + return nil } - // Case 1.B: Server is responding back with headers. - str := &outStream{ - state: empty, - itl: &itemList{}, - wq: h.wq, + // Case 1.A: Server is responding back with headers. + if !h.endStream { + return l.writeHeader(h.streamID, h.endStream, h.hf, h.onWrite) } - l.estdStreams[h.streamID] = str - return l.writeHeader(h.streamID, h.endStream, h.hf, h.onWrite) + // else: Case 1.B: Server wants to close stream. + + if str.state != empty { // either active or waiting on stream quota. + // add it str's list of items. + str.itl.enqueue(h) + return nil + } + if err := l.writeHeader(h.streamID, h.endStream, h.hf, h.onWrite); err != nil { + return err + } + return l.cleanupStreamHandler(h.cleanup) } // Case 2: Client wants to originate stream. str := &outStream{ @@ -632,6 +655,8 @@ func (l *loopyWriter) handle(i interface{}) error { return l.outgoingSettingsHandler(i) case *headerFrame: return l.headerHandler(i) + case *registerStream: + return l.registerStreamHandler(i) case *cleanupStream: return l.cleanupStreamHandler(i) case *incomingGoAway: @@ -664,6 +689,8 @@ func (l *loopyWriter) applySettings(ss []http2.Setting) error { } } } + case http2.SettingHeaderTableSize: + updateHeaderTblSize(l.hEnc, s.Val) } } return nil diff --git a/vendor/google.golang.org/grpc/transport/flowcontrol.go b/vendor/google.golang.org/grpc/transport/flowcontrol.go index 378f5c450..bbf98b6f5 100644 --- a/vendor/google.golang.org/grpc/transport/flowcontrol.go +++ b/vendor/google.golang.org/grpc/transport/flowcontrol.go @@ -58,14 +58,20 @@ type writeQuota struct { ch chan struct{} // done is triggered in error case. done <-chan struct{} + // replenish is called by loopyWriter to give quota back to. + // It is implemented as a field so that it can be updated + // by tests. + replenish func(n int) } func newWriteQuota(sz int32, done <-chan struct{}) *writeQuota { - return &writeQuota{ + w := &writeQuota{ quota: sz, ch: make(chan struct{}, 1), done: done, } + w.replenish = w.realReplenish + return w } func (w *writeQuota) get(sz int32) error { @@ -83,7 +89,7 @@ func (w *writeQuota) get(sz int32) error { } } -func (w *writeQuota) replenish(n int) { +func (w *writeQuota) realReplenish(n int) { sz := int32(n) a := atomic.AddInt32(&w.quota, sz) b := a - sz diff --git a/vendor/google.golang.org/grpc/transport/http2_client.go b/vendor/google.golang.org/grpc/transport/http2_client.go index 1fdabd954..eaf007eb0 100644 --- a/vendor/google.golang.org/grpc/transport/http2_client.go +++ b/vendor/google.golang.org/grpc/transport/http2_client.go @@ -31,9 +31,9 @@ import ( "golang.org/x/net/http2" "golang.org/x/net/http2/hpack" - "google.golang.org/grpc/channelz" "google.golang.org/grpc/codes" "google.golang.org/grpc/credentials" + "google.golang.org/grpc/internal/channelz" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" @@ -76,8 +76,9 @@ type http2Client struct { // Boolean to keep track of reading activity on transport. // 1 is true and 0 is false. - activity uint32 // Accessed atomically. - kp keepalive.ClientParameters + activity uint32 // Accessed atomically. + kp keepalive.ClientParameters + keepaliveEnabled bool statsHandler stats.Handler @@ -259,6 +260,10 @@ func newHTTP2Client(connectCtx, ctx context.Context, addr TargetInfo, opts Conne if channelz.IsOn() { t.channelzID = channelz.RegisterNormalSocket(t, opts.ChannelzParentID, "") } + if t.kp.Time != infinity { + t.keepaliveEnabled = true + go t.keepalive() + } // Start the reader goroutine for incoming message. Each transport has // a dedicated goroutine which reads HTTP2 frame from network. Then it // dispatches the frame to the corresponding stream entity. @@ -295,13 +300,17 @@ func newHTTP2Client(connectCtx, ctx context.Context, addr TargetInfo, opts Conne t.framer.writer.Flush() go func() { t.loopy = newLoopyWriter(clientSide, t.framer, t.controlBuf, t.bdpEst) - t.loopy.run() - t.conn.Close() + err := t.loopy.run() + if err != nil { + errorf("transport: loopyWriter.run returning. Err: %v", err) + } + // If it's a connection error, let reader goroutine handle it + // since there might be data in the buffers. + if _, ok := err.(net.Error); !ok { + t.conn.Close() + } close(t.writerDone) }() - if t.kp.Time != infinity { - go t.keepalive() - } return t, nil } @@ -537,7 +546,7 @@ func (t *http2Client) NewStream(ctx context.Context, callHdr *CallHdr) (_ *Strea var sendPing bool // If the number of active streams change from 0 to 1, then check if keepalive // has gone dormant. If so, wake it up. - if len(t.activeStreams) == 1 { + if len(t.activeStreams) == 1 && t.keepaliveEnabled { select { case t.awakenKeepalive <- struct{}{}: sendPing = true @@ -735,6 +744,7 @@ func (t *http2Client) GracefulClose() error { if active == 0 { return t.Close() } + t.controlBuf.put(&incomingGoAway{}) return nil } @@ -1109,7 +1119,9 @@ func (t *http2Client) reader() { t.Close() return } - atomic.CompareAndSwapUint32(&t.activity, 0, 1) + if t.keepaliveEnabled { + atomic.CompareAndSwapUint32(&t.activity, 0, 1) + } sf, ok := frame.(*http2.SettingsFrame) if !ok { t.Close() @@ -1121,7 +1133,9 @@ func (t *http2Client) reader() { // loop to keep reading incoming messages on this transport. for { frame, err := t.framer.fr.ReadFrame() - atomic.CompareAndSwapUint32(&t.activity, 0, 1) + if t.keepaliveEnabled { + atomic.CompareAndSwapUint32(&t.activity, 0, 1) + } if err != nil { // Abort an active stream if the http2.Framer returns a // http2.StreamError. This can happen only if the server's response diff --git a/vendor/google.golang.org/grpc/transport/http2_server.go b/vendor/google.golang.org/grpc/transport/http2_server.go index 8b93e222e..19acedb2b 100644 --- a/vendor/google.golang.org/grpc/transport/http2_server.go +++ b/vendor/google.golang.org/grpc/transport/http2_server.go @@ -24,7 +24,6 @@ import ( "fmt" "io" "math" - "math/rand" "net" "strconv" "sync" @@ -36,9 +35,11 @@ import ( "golang.org/x/net/http2" "golang.org/x/net/http2/hpack" - "google.golang.org/grpc/channelz" "google.golang.org/grpc/codes" "google.golang.org/grpc/credentials" + "google.golang.org/grpc/grpclog" + "google.golang.org/grpc/internal/channelz" + "google.golang.org/grpc/internal/grpcrand" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/metadata" "google.golang.org/grpc/peer" @@ -273,7 +274,9 @@ func newHTTP2Server(conn net.Conn, config *ServerConfig) (_ ServerTransport, err go func() { t.loopy = newLoopyWriter(serverSide, t.framer, t.controlBuf, t.bdpEst) t.loopy.ssGoAwayHandler = t.outgoingGoAwayHandler - t.loopy.run() + if err := t.loopy.run(); err != nil { + errorf("transport: loopyWriter.run returning. Err: %v", err) + } t.conn.Close() close(t.writerDone) }() @@ -413,6 +416,11 @@ func (t *http2Server) operateHeaders(frame *http2.MetaHeadersFrame, handle func( t.updateWindow(s, uint32(n)) }, } + // Register the stream with loopy. + t.controlBuf.put(®isterStream{ + streamID: s.id, + wq: s.wq, + }) handle(s) return } @@ -682,28 +690,7 @@ func (t *http2Server) handleWindowUpdate(f *http2.WindowUpdateFrame) { }) } -// WriteHeader sends the header metedata md back to the client. -func (t *http2Server) WriteHeader(s *Stream, md metadata.MD) error { - if s.headerOk || s.getState() == streamDone { - return ErrIllegalHeaderWrite - } - s.headerOk = true - if md.Len() > 0 { - if s.header.Len() > 0 { - s.header = metadata.Join(s.header, md) - } else { - s.header = md - } - } - md = s.header - // TODO(mmukhi): Benchmark if the performance gets better if count the metadata and other header fields - // first and create a slice of that exact size. - headerFields := make([]hpack.HeaderField, 0, 2) // at least :status, content-type will be there if none else. - headerFields = append(headerFields, hpack.HeaderField{Name: ":status", Value: "200"}) - headerFields = append(headerFields, hpack.HeaderField{Name: "content-type", Value: contentType(s.contentSubtype)}) - if s.sendCompress != "" { - headerFields = append(headerFields, hpack.HeaderField{Name: "grpc-encoding", Value: s.sendCompress}) - } +func appendHeaderFieldsFromMD(headerFields []hpack.HeaderField, md metadata.MD) []hpack.HeaderField { for k, vv := range md { if isReservedHeader(k) { // Clients don't tolerate reading restricted headers after some non restricted ones were sent. @@ -713,6 +700,37 @@ func (t *http2Server) WriteHeader(s *Stream, md metadata.MD) error { headerFields = append(headerFields, hpack.HeaderField{Name: k, Value: encodeMetadataHeader(k, v)}) } } + return headerFields +} + +// WriteHeader sends the header metedata md back to the client. +func (t *http2Server) WriteHeader(s *Stream, md metadata.MD) error { + if s.updateHeaderSent() || s.getState() == streamDone { + return ErrIllegalHeaderWrite + } + s.hdrMu.Lock() + if md.Len() > 0 { + if s.header.Len() > 0 { + s.header = metadata.Join(s.header, md) + } else { + s.header = md + } + } + t.writeHeaderLocked(s) + s.hdrMu.Unlock() + return nil +} + +func (t *http2Server) writeHeaderLocked(s *Stream) { + // TODO(mmukhi): Benchmark if the performance gets better if count the metadata and other header fields + // first and create a slice of that exact size. + headerFields := make([]hpack.HeaderField, 0, 2) // at least :status, content-type will be there if none else. + headerFields = append(headerFields, hpack.HeaderField{Name: ":status", Value: "200"}) + headerFields = append(headerFields, hpack.HeaderField{Name: "content-type", Value: contentType(s.contentSubtype)}) + if s.sendCompress != "" { + headerFields = append(headerFields, hpack.HeaderField{Name: "grpc-encoding", Value: s.sendCompress}) + } + headerFields = appendHeaderFieldsFromMD(headerFields, s.header) t.controlBuf.put(&headerFrame{ streamID: s.id, hf: headerFields, @@ -720,7 +738,6 @@ func (t *http2Server) WriteHeader(s *Stream, md metadata.MD) error { onWrite: func() { atomic.StoreUint32(&t.resetPingStrikes, 1) }, - wq: s.wq, }) if t.stats != nil { // Note: WireLength is not set in outHeader. @@ -728,7 +745,6 @@ func (t *http2Server) WriteHeader(s *Stream, md metadata.MD) error { outHeader := &stats.OutHeader{} t.stats.HandleRPC(s.Context(), outHeader) } - return nil } // WriteStatus sends stream status to the client and terminates the stream. @@ -736,21 +752,20 @@ func (t *http2Server) WriteHeader(s *Stream, md metadata.MD) error { // TODO(zhaoq): Now it indicates the end of entire stream. Revisit if early // OK is adopted. func (t *http2Server) WriteStatus(s *Stream, st *status.Status) error { - if !s.headerOk && s.header.Len() > 0 { - if err := t.WriteHeader(s, nil); err != nil { - return err - } - } else { - if s.getState() == streamDone { - return nil - } + if s.getState() == streamDone { + return nil } + s.hdrMu.Lock() // TODO(mmukhi): Benchmark if the performance gets better if count the metadata and other header fields // first and create a slice of that exact size. headerFields := make([]hpack.HeaderField, 0, 2) // grpc-status and grpc-message will be there if none else. - if !s.headerOk { - headerFields = append(headerFields, hpack.HeaderField{Name: ":status", Value: "200"}) - headerFields = append(headerFields, hpack.HeaderField{Name: "content-type", Value: contentType(s.contentSubtype)}) + if !s.updateHeaderSent() { // No headers have been sent. + if len(s.header) > 0 { // Send a separate header frame. + t.writeHeaderLocked(s) + } else { // Send a trailer only response. + headerFields = append(headerFields, hpack.HeaderField{Name: ":status", Value: "200"}) + headerFields = append(headerFields, hpack.HeaderField{Name: "content-type", Value: contentType(s.contentSubtype)}) + } } headerFields = append(headerFields, hpack.HeaderField{Name: "grpc-status", Value: strconv.Itoa(int(st.Code()))}) headerFields = append(headerFields, hpack.HeaderField{Name: "grpc-message", Value: encodeGrpcMessage(st.Message())}) @@ -759,23 +774,15 @@ func (t *http2Server) WriteStatus(s *Stream, st *status.Status) error { stBytes, err := proto.Marshal(p) if err != nil { // TODO: return error instead, when callers are able to handle it. - panic(err) + grpclog.Errorf("transport: failed to marshal rpc status: %v, error: %v", p, err) + } else { + headerFields = append(headerFields, hpack.HeaderField{Name: "grpc-status-details-bin", Value: encodeBinHeader(stBytes)}) } - - headerFields = append(headerFields, hpack.HeaderField{Name: "grpc-status-details-bin", Value: encodeBinHeader(stBytes)}) } // Attach the trailer metadata. - for k, vv := range s.trailer { - // Clients don't tolerate reading restricted headers after some non restricted ones were sent. - if isReservedHeader(k) { - continue - } - for _, v := range vv { - headerFields = append(headerFields, hpack.HeaderField{Name: k, Value: encodeMetadataHeader(k, v)}) - } - } - trailer := &headerFrame{ + headerFields = appendHeaderFieldsFromMD(headerFields, s.trailer) + trailingHeader := &headerFrame{ streamID: s.id, hf: headerFields, endStream: true, @@ -783,7 +790,8 @@ func (t *http2Server) WriteStatus(s *Stream, st *status.Status) error { atomic.StoreUint32(&t.resetPingStrikes, 1) }, } - t.closeStream(s, false, 0, trailer, true) + s.hdrMu.Unlock() + t.closeStream(s, false, 0, trailingHeader, true) if t.stats != nil { t.stats.HandleRPC(s.Context(), &stats.OutTrailer{}) } @@ -793,7 +801,7 @@ func (t *http2Server) WriteStatus(s *Stream, st *status.Status) error { // Write converts the data into HTTP2 data frame and sends it out. Non-nil error // is returns if it fails (e.g., framing error, transport error). func (t *http2Server) Write(s *Stream, hdr []byte, data []byte, opts *Options) error { - if !s.headerOk { // Headers haven't been written yet. + if !s.isHeaderSent() { // Headers haven't been written yet. if err := t.WriteHeader(s, nil); err != nil { // TODO(mmukhi, dfawley): Make sure this is the right code to return. return streamErrorf(codes.Internal, "transport: %v", err) @@ -1123,14 +1131,12 @@ func (t *http2Server) getOutFlowWindow() int64 { } } -var rgen = rand.New(rand.NewSource(time.Now().UnixNano())) - func getJitter(v time.Duration) time.Duration { if v == infinity { return 0 } // Generate a jitter between +/- 10% of the value. r := int64(v / 10) - j := rgen.Int63n(2*r) - r + j := grpcrand.Int63n(2*r) - r return time.Duration(j) } diff --git a/vendor/google.golang.org/grpc/transport/http_util.go b/vendor/google.golang.org/grpc/transport/http_util.go index a35586608..7d15c7d74 100644 --- a/vendor/google.golang.org/grpc/transport/http_util.go +++ b/vendor/google.golang.org/grpc/transport/http_util.go @@ -28,6 +28,7 @@ import ( "strconv" "strings" "time" + "unicode/utf8" "github.com/golang/protobuf/proto" "golang.org/x/net/http2" @@ -437,16 +438,17 @@ func decodeTimeout(s string) (time.Duration, error) { const ( spaceByte = ' ' - tildaByte = '~' + tildeByte = '~' percentByte = '%' ) // encodeGrpcMessage is used to encode status code in header field -// "grpc-message". -// It checks to see if each individual byte in msg is an -// allowable byte, and then either percent encoding or passing it through. -// When percent encoding, the byte is converted into hexadecimal notation -// with a '%' prepended. +// "grpc-message". It does percent encoding and also replaces invalid utf-8 +// characters with Unicode replacement character. +// +// It checks to see if each individual byte in msg is an allowable byte, and +// then either percent encoding or passing it through. When percent encoding, +// the byte is converted into hexadecimal notation with a '%' prepended. func encodeGrpcMessage(msg string) string { if msg == "" { return "" @@ -454,7 +456,7 @@ func encodeGrpcMessage(msg string) string { lenMsg := len(msg) for i := 0; i < lenMsg; i++ { c := msg[i] - if !(c >= spaceByte && c < tildaByte && c != percentByte) { + if !(c >= spaceByte && c <= tildeByte && c != percentByte) { return encodeGrpcMessageUnchecked(msg) } } @@ -463,14 +465,26 @@ func encodeGrpcMessage(msg string) string { func encodeGrpcMessageUnchecked(msg string) string { var buf bytes.Buffer - lenMsg := len(msg) - for i := 0; i < lenMsg; i++ { - c := msg[i] - if c >= spaceByte && c < tildaByte && c != percentByte { - buf.WriteByte(c) - } else { - buf.WriteString(fmt.Sprintf("%%%02X", c)) + for len(msg) > 0 { + r, size := utf8.DecodeRuneInString(msg) + for _, b := range []byte(string(r)) { + if size > 1 { + // If size > 1, r is not ascii. Always do percent encoding. + buf.WriteString(fmt.Sprintf("%%%02X", b)) + continue + } + + // The for loop is necessary even if size == 1. r could be + // utf8.RuneError. + // + // fmt.Sprintf("%%%02X", utf8.RuneError) gives "%FFFD". + if b >= spaceByte && b <= tildeByte && b != percentByte { + buf.WriteByte(b) + } else { + buf.WriteString(fmt.Sprintf("%%%02X", b)) + } } + msg = msg[size:] } return buf.String() } @@ -531,10 +545,14 @@ func (w *bufWriter) Write(b []byte) (n int, err error) { if w.err != nil { return 0, w.err } - n = copy(w.buf[w.offset:], b) - w.offset += n - if w.offset >= w.batchSize { - err = w.Flush() + for len(b) > 0 { + nn := copy(w.buf[w.offset:], b) + b = b[nn:] + w.offset += nn + n += nn + if w.offset >= w.batchSize { + err = w.Flush() + } } return n, err } diff --git a/vendor/google.golang.org/grpc/transport/http_util_test.go b/vendor/google.golang.org/grpc/transport/http_util_test.go index c3754781d..44169b653 100644 --- a/vendor/google.golang.org/grpc/transport/http_util_test.go +++ b/vendor/google.golang.org/grpc/transport/http_util_test.go @@ -102,12 +102,31 @@ func TestEncodeGrpcMessage(t *testing.T) { }{ {"", ""}, {"Hello", "Hello"}, - {"my favorite character is \u0000", "my favorite character is %00"}, - {"my favorite character is %", "my favorite character is %25"}, + {"\u0000", "%00"}, + {"%", "%25"}, + {"系统", "%E7%B3%BB%E7%BB%9F"}, + {string([]byte{0xff, 0xfe, 0xfd}), "%EF%BF%BD%EF%BF%BD%EF%BF%BD"}, } { actual := encodeGrpcMessage(tt.input) if tt.expected != actual { - t.Errorf("encodeGrpcMessage(%v) = %v, want %v", tt.input, actual, tt.expected) + t.Errorf("encodeGrpcMessage(%q) = %q, want %q", tt.input, actual, tt.expected) + } + } + + // make sure that all the visible ASCII chars except '%' are not percent encoded. + for i := ' '; i <= '~' && i != '%'; i++ { + output := encodeGrpcMessage(string(i)) + if output != string(i) { + t.Errorf("encodeGrpcMessage(%v) = %v, want %v", string(i), output, string(i)) + } + } + + // make sure that all the invisible ASCII chars and '%' are percent encoded. + for i := rune(0); i == '%' || (i >= rune(0) && i < ' ') || (i > '~' && i <= rune(127)); i++ { + output := encodeGrpcMessage(string(i)) + expected := fmt.Sprintf("%%%02X", i) + if output != expected { + t.Errorf("encodeGrpcMessage(%v) = %v, want %v", string(i), output, expected) } } } @@ -123,10 +142,52 @@ func TestDecodeGrpcMessage(t *testing.T) { {"H%6", "H%6"}, {"%G0", "%G0"}, {"%E7%B3%BB%E7%BB%9F", "系统"}, + {"%EF%BF%BD", "�"}, } { actual := decodeGrpcMessage(tt.input) if tt.expected != actual { - t.Errorf("dncodeGrpcMessage(%v) = %v, want %v", tt.input, actual, tt.expected) + t.Errorf("decodeGrpcMessage(%q) = %q, want %q", tt.input, actual, tt.expected) + } + } + + // make sure that all the visible ASCII chars except '%' are not percent decoded. + for i := ' '; i <= '~' && i != '%'; i++ { + output := decodeGrpcMessage(string(i)) + if output != string(i) { + t.Errorf("decodeGrpcMessage(%v) = %v, want %v", string(i), output, string(i)) + } + } + + // make sure that all the invisible ASCII chars and '%' are percent decoded. + for i := rune(0); i == '%' || (i >= rune(0) && i < ' ') || (i > '~' && i <= rune(127)); i++ { + output := decodeGrpcMessage(fmt.Sprintf("%%%02X", i)) + if output != string(i) { + t.Errorf("decodeGrpcMessage(%v) = %v, want %v", fmt.Sprintf("%%%02X", i), output, string(i)) + } + } +} + +// Decode an encoded string should get the same thing back, except for invalid +// utf8 chars. +func TestDecodeEncodeGrpcMessage(t *testing.T) { + testCases := []struct { + orig string + want string + }{ + {"", ""}, + {"hello", "hello"}, + {"h%6", "h%6"}, + {"%G0", "%G0"}, + {"系统", "系统"}, + {"Hello, 世界", "Hello, 世界"}, + + {string([]byte{0xff, 0xfe, 0xfd}), "���"}, + {string([]byte{0xff}) + "Hello" + string([]byte{0xfe}) + "世界" + string([]byte{0xfd}), "�Hello�世界�"}, + } + for _, tC := range testCases { + got := decodeGrpcMessage(encodeGrpcMessage(tC.orig)) + if got != tC.want { + t.Errorf("decodeGrpcMessage(encodeGrpcMessage(%q)) = %q, want %q", tC.orig, got, tC.want) } } } diff --git a/vendor/google.golang.org/grpc/transport/transport.go b/vendor/google.golang.org/grpc/transport/transport.go index 2f643a3d0..f51f87888 100644 --- a/vendor/google.golang.org/grpc/transport/transport.go +++ b/vendor/google.golang.org/grpc/transport/transport.go @@ -185,13 +185,20 @@ type Stream struct { headerChan chan struct{} // closed to indicate the end of header metadata. headerDone uint32 // set when headerChan is closed. Used to avoid closing headerChan multiple times. - header metadata.MD // the received header metadata. - trailer metadata.MD // the key-value map of trailer metadata. - headerOk bool // becomes true from the first header is about to send - state streamState + // hdrMu protects header and trailer metadata on the server-side. + hdrMu sync.Mutex + header metadata.MD // the received header metadata. + trailer metadata.MD // the key-value map of trailer metadata. - status *status.Status // the status error received from the server + // On the server-side, headerSent is atomically set to 1 when the headers are sent out. + headerSent uint32 + + state streamState + + // On client-side it is the status error received from the server. + // On server-side it is unused. + status *status.Status bytesReceived uint32 // indicates whether any bytes have been received on this stream unprocessed uint32 // set if the server sends a refused stream or GOAWAY including this stream @@ -201,6 +208,17 @@ type Stream struct { contentSubtype string } +// isHeaderSent is only valid on the server-side. +func (s *Stream) isHeaderSent() bool { + return atomic.LoadUint32(&s.headerSent) == 1 +} + +// updateHeaderSent updates headerSent and returns true +// if it was alreay set. It is valid only on server-side. +func (s *Stream) updateHeaderSent() bool { + return atomic.SwapUint32(&s.headerSent, 1) == 1 +} + func (s *Stream) swapState(st streamState) streamState { return streamState(atomic.SwapUint32((*uint32)(&s.state), uint32(st))) } @@ -313,10 +331,12 @@ func (s *Stream) SetHeader(md metadata.MD) error { if md.Len() == 0 { return nil } - if s.headerOk || atomic.LoadUint32((*uint32)(&s.state)) == uint32(streamDone) { + if s.isHeaderSent() || s.getState() == streamDone { return ErrIllegalHeaderWrite } + s.hdrMu.Lock() s.header = metadata.Join(s.header, md) + s.hdrMu.Unlock() return nil } @@ -335,7 +355,12 @@ func (s *Stream) SetTrailer(md metadata.MD) error { if md.Len() == 0 { return nil } + if s.getState() == streamDone { + return ErrIllegalHeaderWrite + } + s.hdrMu.Lock() s.trailer = metadata.Join(s.trailer, md) + s.hdrMu.Unlock() return nil } diff --git a/vendor/google.golang.org/grpc/transport/transport_test.go b/vendor/google.golang.org/grpc/transport/transport_test.go index f25de2924..b6add205d 100644 --- a/vendor/google.golang.org/grpc/transport/transport_test.go +++ b/vendor/google.golang.org/grpc/transport/transport_test.go @@ -40,6 +40,7 @@ import ( "golang.org/x/net/http2" "golang.org/x/net/http2/hpack" "google.golang.org/grpc/codes" + "google.golang.org/grpc/internal/leakcheck" "google.golang.org/grpc/keepalive" "google.golang.org/grpc/status" ) @@ -51,6 +52,7 @@ type server struct { mu sync.Mutex conns map[ServerTransport]bool h *testStreamHandler + ready chan struct{} } var ( @@ -62,8 +64,9 @@ var ( ) type testStreamHandler struct { - t *http2Server - notify chan struct{} + t *http2Server + notify chan struct{} + getNotified chan struct{} } type hType int @@ -76,7 +79,6 @@ const ( encodingRequiredStatus invalidHeaderField delayRead - delayWrite pingpong ) @@ -182,6 +184,10 @@ func (h *testStreamHandler) handleStreamInvalidHeaderField(t *testing.T, s *Stre }) } +// handleStreamDelayRead delays reads so that the other side has to halt on +// stream-level flow control. +// This handler assumes dynamic flow control is turned off and assumes window +// sizes to be set to defaultWindowSize. func (h *testStreamHandler) handleStreamDelayRead(t *testing.T, s *Stream) { req := expectedRequest resp := expectedResponse @@ -189,11 +195,52 @@ func (h *testStreamHandler) handleStreamDelayRead(t *testing.T, s *Stream) { req = expectedRequestLarge resp = expectedResponseLarge } + var ( + mu sync.Mutex + total int + ) + s.wq.replenish = func(n int) { + mu.Lock() + total += n + mu.Unlock() + s.wq.realReplenish(n) + } + getTotal := func() int { + mu.Lock() + defer mu.Unlock() + return total + } + done := make(chan struct{}) + defer close(done) + go func() { + for { + select { + // Prevent goroutine from leaking. + case <-done: + return + default: + } + if getTotal() == defaultWindowSize { + // Signal the client to start reading and + // thereby send window update. + close(h.notify) + return + } + runtime.Gosched() + } + }() p := make([]byte, len(req)) - // Wait before reading. Give time to client to start sending - // before server starts reading. - time.Sleep(2 * time.Second) + // Let the other side run out of stream-level window before + // starting to read and thereby sending a window update. + timer := time.NewTimer(time.Second * 10) + select { + case <-h.getNotified: + timer.Stop() + case <-timer.C: + t.Errorf("Server timed-out.") + return + } _, err := s.Read(p) if err != nil { t.Errorf("s.Read(_) = _, %v, want _, ", err) @@ -204,41 +251,19 @@ func (h *testStreamHandler) handleStreamDelayRead(t *testing.T, s *Stream) { t.Errorf("handleStream got %v, want %v", p, req) return } - // send a response back to the client. + // This write will cause server to run out of stream level, + // flow control and the other side won't send a window update + // until that happens. if err := h.t.Write(s, nil, resp, &Options{}); err != nil { t.Errorf("server Write got %v, want ", err) return } - // send the trailer to end the stream. - if err := h.t.WriteStatus(s, status.New(codes.OK, "")); err != nil { - t.Errorf("server WriteStatus got %v, want ", err) - return - } -} - -func (h *testStreamHandler) handleStreamDelayWrite(t *testing.T, s *Stream) { - req := expectedRequest - resp := expectedResponse - if s.Method() == "foo.Large" { - req = expectedRequestLarge - resp = expectedResponseLarge - } - p := make([]byte, len(req)) - _, err := s.Read(p) + // Read one more time to ensure that everything remains fine and + // that the goroutine, that we launched earlier to signal client + // to read, gets enough time to process. + _, err = s.Read(p) if err != nil { - t.Errorf("s.Read(_) = _, %v, want _, ", err) - return - } - if !bytes.Equal(p, req) { - t.Errorf("handleStream got %v, want %v", p, req) - return - } - - // Wait before sending. Give time to client to start reading - // before server starts sending. - time.Sleep(2 * time.Second) - if err := h.t.Write(s, nil, resp, &Options{}); err != nil { - t.Errorf("server Write got %v, want ", err) + t.Errorf("s.Read(_) = _, %v, want _, nil", err) return } // send the trailer to end the stream. @@ -317,17 +342,16 @@ func (s *server) start(t *testing.T, port int, serverConfig *ServerConfig, ht hT return ctx }) case delayRead: + h.notify = make(chan struct{}) + h.getNotified = make(chan struct{}) + s.mu.Lock() + close(s.ready) + s.mu.Unlock() go transport.HandleStreams(func(s *Stream) { go h.handleStreamDelayRead(t, s) }, func(ctx context.Context, method string) context.Context { return ctx }) - case delayWrite: - go transport.HandleStreams(func(s *Stream) { - go h.handleStreamDelayWrite(t, s) - }, func(ctx context.Context, method string) context.Context { - return ctx - }) case pingpong: go transport.HandleStreams(func(s *Stream) { go h.handleStreamPingPong(t, s) @@ -366,7 +390,7 @@ func (s *server) stop() { } func setUpServerOnly(t *testing.T, port int, serverConfig *ServerConfig, ht hType) *server { - server := &server{startedErr: make(chan error, 1)} + server := &server{startedErr: make(chan error, 1), ready: make(chan struct{})} go server.start(t, port, serverConfig, ht) server.wait(t, 2*time.Second) return server @@ -957,89 +981,113 @@ func TestLargeMessage(t *testing.T) { } func TestLargeMessageWithDelayRead(t *testing.T) { - server, ct := setUp(t, 0, math.MaxUint32, delayRead) + // Disable dynamic flow control. + sc := &ServerConfig{ + InitialWindowSize: defaultWindowSize, + InitialConnWindowSize: defaultWindowSize, + } + co := ConnectOptions{ + InitialWindowSize: defaultWindowSize, + InitialConnWindowSize: defaultWindowSize, + } + server, ct := setUpWithOptions(t, 0, sc, delayRead, co, func() {}) + defer server.stop() + defer ct.Close() + server.mu.Lock() + ready := server.ready + server.mu.Unlock() callHdr := &CallHdr{ Host: "localhost", Method: "foo.Large", } - var wg sync.WaitGroup - for i := 0; i < 2; i++ { - wg.Add(1) - go func() { - defer wg.Done() - ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(time.Second*10)) - defer cancel() - s, err := ct.NewStream(ctx, callHdr) - if err != nil { - t.Errorf("%v.NewStream(_, _) = _, %v, want _, ", ct, err) - return - } - if err := ct.Write(s, []byte{}, expectedRequestLarge, &Options{Last: true, Delay: false}); err != nil { - t.Errorf("%v.Write(_, _, _) = %v, want ", ct, err) - return - } - p := make([]byte, len(expectedResponseLarge)) - - // Give time to server to begin sending before client starts reading. - time.Sleep(2 * time.Second) - if _, err := s.Read(p); err != nil || !bytes.Equal(p, expectedResponseLarge) { - t.Errorf("s.Read(_) = _, %v, want _, ", err) - return - } - if _, err = s.Read(p); err != io.EOF { - t.Errorf("Failed to complete the stream %v; want ", err) - } - }() + ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(time.Second*10)) + defer cancel() + s, err := ct.NewStream(ctx, callHdr) + if err != nil { + t.Fatalf("%v.NewStream(_, _) = _, %v, want _, ", ct, err) + return } - wg.Wait() - ct.Close() - server.stop() -} - -func TestLargeMessageDelayWrite(t *testing.T) { - server, ct := setUp(t, 0, math.MaxUint32, delayWrite) - callHdr := &CallHdr{ - Host: "localhost", - Method: "foo.Large", + // Wait for server's handerler to be initialized + select { + case <-ready: + case <-ctx.Done(): + t.Fatalf("Client timed out waiting for server handler to be initialized.") } - var wg sync.WaitGroup - for i := 0; i < 2; i++ { - wg.Add(1) - go func() { - defer wg.Done() - ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(time.Second*10)) - defer cancel() - s, err := ct.NewStream(ctx, callHdr) - if err != nil { - t.Errorf("%v.NewStream(_, _) = _, %v, want _, ", ct, err) - return - } - - // Give time to server to start reading before client starts sending. - time.Sleep(2 * time.Second) - if err := ct.Write(s, []byte{}, expectedRequestLarge, &Options{Last: true, Delay: false}); err != nil { - t.Errorf("%v.Write(_, _, _) = %v, want ", ct, err) - return - } - p := make([]byte, len(expectedResponseLarge)) - if _, err := s.Read(p); err != nil || !bytes.Equal(p, expectedResponseLarge) { - t.Errorf("io.ReadFull(%v) = _, %v, want %v, ", err, p, expectedResponse) - return - } - if _, err = s.Read(p); err != io.EOF { - t.Errorf("Failed to complete the stream %v; want ", err) - } - }() + server.mu.Lock() + serviceHandler := server.h + server.mu.Unlock() + var ( + mu sync.Mutex + total int + ) + s.wq.replenish = func(n int) { + mu.Lock() + total += n + mu.Unlock() + s.wq.realReplenish(n) + } + getTotal := func() int { + mu.Lock() + defer mu.Unlock() + return total + } + done := make(chan struct{}) + defer close(done) + go func() { + for { + select { + // Prevent goroutine from leaking in case of error. + case <-done: + return + default: + } + if getTotal() == defaultWindowSize { + // unblock server to be able to read and + // thereby send stream level window update. + close(serviceHandler.getNotified) + return + } + runtime.Gosched() + } + }() + // This write will cause client to run out of stream level, + // flow control and the other side won't send a window update + // until that happens. + if err := ct.Write(s, []byte{}, expectedRequestLarge, &Options{}); err != nil { + t.Fatalf("write(_, _, _) = %v, want ", err) + } + p := make([]byte, len(expectedResponseLarge)) + + // Wait for the other side to run out of stream level flow control before + // reading and thereby sending a window update. + select { + case <-serviceHandler.notify: + case <-ctx.Done(): + t.Fatalf("Client timed out") + } + if _, err := s.Read(p); err != nil || !bytes.Equal(p, expectedResponseLarge) { + t.Fatalf("s.Read(_) = _, %v, want _, ", err) + } + if err := ct.Write(s, []byte{}, expectedRequestLarge, &Options{Last: true}); err != nil { + t.Fatalf("Write(_, _, _) = %v, want ", err) + } + if _, err = s.Read(p); err != io.EOF { + t.Fatalf("Failed to complete the stream %v; want ", err) } - wg.Wait() - ct.Close() - server.stop() } func TestGracefulClose(t *testing.T) { server, ct := setUp(t, 0, math.MaxUint32, pingpong) - defer server.stop() - defer ct.Close() + defer func() { + // Stop the server's listener to make the server's goroutines terminate + // (after the last active stream is done). + server.lis.Close() + // Check for goroutine leaks (i.e. GracefulClose with an active stream + // doesn't eventually close the connection when that stream completes). + leakcheck.Check(t) + // Correctly clean up the server + server.stop() + }() ctx, cancel := context.WithDeadline(context.Background(), time.Now().Add(time.Second*10)) defer cancel() s, err := ct.NewStream(ctx, &CallHdr{}) @@ -1067,17 +1115,20 @@ func TestGracefulClose(t *testing.T) { } var wg sync.WaitGroup // Expect the failure for all the follow-up streams because ct has been closed gracefully. - for i := 0; i < 100; i++ { + for i := 0; i < 200; i++ { wg.Add(1) go func() { defer wg.Done() str, err := ct.NewStream(context.Background(), &CallHdr{}) - if err == errStreamDrain { + if err == ErrConnClosing { + return + } else if err != nil { + t.Errorf("_.NewStream(_, _) = _, %v, want _, %v", err, ErrConnClosing) return } ct.Write(str, nil, nil, &Options{Last: true}) - if _, err := str.Read(make([]byte, 8)); err != errStreamDrain { - t.Errorf("_.NewStream(_, _) = _, %v, want _, %v", err, errStreamDrain) + if _, err := str.Read(make([]byte, 8)); err != errStreamDrain && err != ErrConnClosing { + t.Errorf("_.Read(_) = _, %v, want _, %v or %v", err, errStreamDrain, ErrConnClosing) } }() } @@ -2149,3 +2200,113 @@ func runPingPongTest(t *testing.T, msgSize int) { } } } + +type tableSizeLimit struct { + mu sync.Mutex + limits []uint32 +} + +func (t *tableSizeLimit) add(limit uint32) { + t.mu.Lock() + t.limits = append(t.limits, limit) + t.mu.Unlock() +} + +func (t *tableSizeLimit) getLen() int { + t.mu.Lock() + defer t.mu.Unlock() + return len(t.limits) +} + +func (t *tableSizeLimit) getIndex(i int) uint32 { + t.mu.Lock() + defer t.mu.Unlock() + return t.limits[i] +} + +func TestHeaderTblSize(t *testing.T) { + limits := &tableSizeLimit{} + updateHeaderTblSize = func(e *hpack.Encoder, v uint32) { + e.SetMaxDynamicTableSizeLimit(v) + limits.add(v) + } + defer func() { + updateHeaderTblSize = func(e *hpack.Encoder, v uint32) { + e.SetMaxDynamicTableSizeLimit(v) + } + }() + + server, ct := setUp(t, 0, math.MaxUint32, normal) + defer ct.Close() + defer server.stop() + _, err := ct.NewStream(context.Background(), &CallHdr{}) + if err != nil { + t.Fatalf("failed to open stream: %v", err) + } + + var svrTransport ServerTransport + var i int + for i = 0; i < 1000; i++ { + server.mu.Lock() + if len(server.conns) != 0 { + server.mu.Unlock() + break + } + server.mu.Unlock() + time.Sleep(10 * time.Millisecond) + continue + } + if i == 1000 { + t.Fatalf("unable to create any server transport after 10s") + } + + for st := range server.conns { + svrTransport = st + break + } + svrTransport.(*http2Server).controlBuf.put(&outgoingSettings{ + ss: []http2.Setting{ + { + ID: http2.SettingHeaderTableSize, + Val: uint32(100), + }, + }, + }) + + for i = 0; i < 1000; i++ { + if limits.getLen() != 1 { + time.Sleep(10 * time.Millisecond) + continue + } + if val := limits.getIndex(0); val != uint32(100) { + t.Fatalf("expected limits[0] = 100, got %d", val) + } + break + } + if i == 1000 { + t.Fatalf("expected len(limits) = 1 within 10s, got != 1") + } + + ct.(*http2Client).controlBuf.put(&outgoingSettings{ + ss: []http2.Setting{ + { + ID: http2.SettingHeaderTableSize, + Val: uint32(200), + }, + }, + }) + + for i := 0; i < 1000; i++ { + if limits.getLen() != 2 { + time.Sleep(10 * time.Millisecond) + continue + } + if val := limits.getIndex(1); val != uint32(200) { + t.Fatalf("expected limits[1] = 200, got %d", val) + } + break + } + if i == 1000 { + t.Fatalf("expected len(limits) = 2 within 10s, got != 2") + } +} diff --git a/vendor/google.golang.org/grpc/backoff_test.go b/vendor/google.golang.org/grpc/version.go similarity index 68% rename from vendor/google.golang.org/grpc/backoff_test.go rename to vendor/google.golang.org/grpc/version.go index 37e8e3f62..7f124fbd5 100644 --- a/vendor/google.golang.org/grpc/backoff_test.go +++ b/vendor/google.golang.org/grpc/version.go @@ -1,6 +1,6 @@ /* * - * Copyright 2017 gRPC authors. + * Copyright 2018 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -18,12 +18,5 @@ package grpc -import "testing" - -func TestBackoffConfigDefaults(t *testing.T) { - b := BackoffConfig{} - setDefaults(&b) - if b != DefaultBackoffConfig { - t.Fatalf("expected BackoffConfig to pickup default parameters: %v != %v", b, DefaultBackoffConfig) - } -} +// Version is the current grpc version. +const Version = "1.13.0" diff --git a/vendor/google.golang.org/grpc/vet.sh b/vendor/google.golang.org/grpc/vet.sh index 6126ab64b..079bc2896 100755 --- a/vendor/google.golang.org/grpc/vet.sh +++ b/vendor/google.golang.org/grpc/vet.sh @@ -54,7 +54,8 @@ if git status --porcelain | read; then fi git ls-files "*.go" | xargs grep -L "\(Copyright [0-9]\{4,\} gRPC authors\)\|DO NOT EDIT" 2>&1 | tee /dev/stderr | (! read) -git ls-files "*.go" | xargs grep -l "\"unsafe\"" 2>&1 | (! grep -v '_test.go') | tee /dev/stderr | (! read) +git ls-files "*.go" | xargs grep -l '"unsafe"' 2>&1 | (! grep -v '_test.go') | tee /dev/stderr | (! read) +git ls-files "*.go" | xargs grep -l '"math/rand"' 2>&1 | (! grep -v '^examples\|^stress\|grpcrand') | tee /dev/stderr | (! read) gofmt -s -d -l . 2>&1 | tee /dev/stderr | (! read) goimports -l . 2>&1 | tee /dev/stderr | (! read) golint ./... 2>&1 | (grep -vE "(_mock|\.pb)\.go:" || true) | tee /dev/stderr | (! read)